{ "best_global_step": 38160, "best_metric": 0.4603464603424072, "best_model_checkpoint": "saves_multiple/ia3/llama-3-8b-instruct/train_codealpacapy_456_1765330670/checkpoint-38160", "epoch": 20.0, "eval_steps": 1908, "global_step": 38160, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.002620545073375262, "grad_norm": 0.32511261105537415, "learning_rate": 5.241090146750525e-08, "loss": 1.0086, "num_input_tokens_seen": 3808, "step": 5 }, { "epoch": 0.005241090146750524, "grad_norm": 0.5838137865066528, "learning_rate": 1.1792452830188679e-07, "loss": 1.4042, "num_input_tokens_seen": 6432, "step": 10 }, { "epoch": 0.007861635220125786, "grad_norm": 0.7283626794815063, "learning_rate": 1.8343815513626837e-07, "loss": 1.4188, "num_input_tokens_seen": 9120, "step": 15 }, { "epoch": 0.010482180293501049, "grad_norm": 0.5501592755317688, "learning_rate": 2.489517819706499e-07, "loss": 1.225, "num_input_tokens_seen": 11776, "step": 20 }, { "epoch": 0.01310272536687631, "grad_norm": 0.4131910502910614, "learning_rate": 3.144654088050315e-07, "loss": 1.0907, "num_input_tokens_seen": 15968, "step": 25 }, { "epoch": 0.015723270440251572, "grad_norm": 0.3226127326488495, "learning_rate": 3.79979035639413e-07, "loss": 0.9587, "num_input_tokens_seen": 20704, "step": 30 }, { "epoch": 0.018343815513626835, "grad_norm": 0.6408722400665283, "learning_rate": 4.454926624737945e-07, "loss": 1.3352, "num_input_tokens_seen": 23264, "step": 35 }, { "epoch": 0.020964360587002098, "grad_norm": 1.6195334196090698, "learning_rate": 5.110062893081761e-07, "loss": 1.5693, "num_input_tokens_seen": 26784, "step": 40 }, { "epoch": 0.02358490566037736, "grad_norm": 0.4692218005657196, "learning_rate": 5.765199161425577e-07, "loss": 1.0837, "num_input_tokens_seen": 30144, "step": 45 }, { "epoch": 0.02620545073375262, "grad_norm": 0.3905942440032959, "learning_rate": 6.420335429769392e-07, "loss": 1.1225, "num_input_tokens_seen": 33216, "step": 50 }, { "epoch": 0.028825995807127882, "grad_norm": 0.42883849143981934, "learning_rate": 7.075471698113208e-07, "loss": 1.0569, "num_input_tokens_seen": 36768, "step": 55 }, { "epoch": 0.031446540880503145, "grad_norm": 0.48735204339027405, "learning_rate": 7.730607966457024e-07, "loss": 1.265, "num_input_tokens_seen": 40288, "step": 60 }, { "epoch": 0.034067085953878404, "grad_norm": 0.6352704167366028, "learning_rate": 8.38574423480084e-07, "loss": 1.2034, "num_input_tokens_seen": 43520, "step": 65 }, { "epoch": 0.03668763102725367, "grad_norm": 0.6355311274528503, "learning_rate": 9.040880503144654e-07, "loss": 1.0534, "num_input_tokens_seen": 46816, "step": 70 }, { "epoch": 0.03930817610062893, "grad_norm": 0.3665657937526703, "learning_rate": 9.69601677148847e-07, "loss": 1.0422, "num_input_tokens_seen": 49664, "step": 75 }, { "epoch": 0.041928721174004195, "grad_norm": 0.3607203960418701, "learning_rate": 1.0351153039832286e-06, "loss": 1.1798, "num_input_tokens_seen": 53088, "step": 80 }, { "epoch": 0.044549266247379454, "grad_norm": 0.4366300702095032, "learning_rate": 1.10062893081761e-06, "loss": 1.0457, "num_input_tokens_seen": 56608, "step": 85 }, { "epoch": 0.04716981132075472, "grad_norm": 0.6762767434120178, "learning_rate": 1.1661425576519918e-06, "loss": 1.247, "num_input_tokens_seen": 59904, "step": 90 }, { "epoch": 0.04979035639412998, "grad_norm": 0.4409782588481903, "learning_rate": 1.2316561844863732e-06, "loss": 1.0009, "num_input_tokens_seen": 63392, "step": 95 }, { "epoch": 0.05241090146750524, "grad_norm": 0.5663008689880371, "learning_rate": 1.2971698113207548e-06, "loss": 1.2967, "num_input_tokens_seen": 65952, "step": 100 }, { "epoch": 0.055031446540880505, "grad_norm": 0.6701657772064209, "learning_rate": 1.3626834381551364e-06, "loss": 1.1392, "num_input_tokens_seen": 68672, "step": 105 }, { "epoch": 0.057651991614255764, "grad_norm": 0.6033591628074646, "learning_rate": 1.4281970649895177e-06, "loss": 1.2221, "num_input_tokens_seen": 71712, "step": 110 }, { "epoch": 0.06027253668763103, "grad_norm": 0.6005898118019104, "learning_rate": 1.4937106918238995e-06, "loss": 1.451, "num_input_tokens_seen": 74656, "step": 115 }, { "epoch": 0.06289308176100629, "grad_norm": 0.6134551167488098, "learning_rate": 1.559224318658281e-06, "loss": 1.0721, "num_input_tokens_seen": 78208, "step": 120 }, { "epoch": 0.06551362683438156, "grad_norm": 0.6200802326202393, "learning_rate": 1.6247379454926625e-06, "loss": 1.1707, "num_input_tokens_seen": 81376, "step": 125 }, { "epoch": 0.06813417190775681, "grad_norm": 0.4053666591644287, "learning_rate": 1.690251572327044e-06, "loss": 1.4091, "num_input_tokens_seen": 84288, "step": 130 }, { "epoch": 0.07075471698113207, "grad_norm": 0.8910534977912903, "learning_rate": 1.7557651991614257e-06, "loss": 1.496, "num_input_tokens_seen": 86976, "step": 135 }, { "epoch": 0.07337526205450734, "grad_norm": 0.5155412554740906, "learning_rate": 1.821278825995807e-06, "loss": 1.1917, "num_input_tokens_seen": 89952, "step": 140 }, { "epoch": 0.0759958071278826, "grad_norm": 0.34066665172576904, "learning_rate": 1.8867924528301887e-06, "loss": 1.043, "num_input_tokens_seen": 93120, "step": 145 }, { "epoch": 0.07861635220125786, "grad_norm": 0.9253441095352173, "learning_rate": 1.9523060796645703e-06, "loss": 1.2901, "num_input_tokens_seen": 96640, "step": 150 }, { "epoch": 0.08123689727463312, "grad_norm": 0.43350499868392944, "learning_rate": 2.017819706498952e-06, "loss": 1.071, "num_input_tokens_seen": 99744, "step": 155 }, { "epoch": 0.08385744234800839, "grad_norm": 0.4996618628501892, "learning_rate": 2.0833333333333334e-06, "loss": 1.1505, "num_input_tokens_seen": 102720, "step": 160 }, { "epoch": 0.08647798742138364, "grad_norm": 0.34180599451065063, "learning_rate": 2.148846960167715e-06, "loss": 1.3087, "num_input_tokens_seen": 105344, "step": 165 }, { "epoch": 0.08909853249475891, "grad_norm": 0.7076641321182251, "learning_rate": 2.2143605870020966e-06, "loss": 1.3892, "num_input_tokens_seen": 108576, "step": 170 }, { "epoch": 0.09171907756813417, "grad_norm": 0.595442533493042, "learning_rate": 2.279874213836478e-06, "loss": 1.1036, "num_input_tokens_seen": 111520, "step": 175 }, { "epoch": 0.09433962264150944, "grad_norm": 0.5885025262832642, "learning_rate": 2.34538784067086e-06, "loss": 1.5123, "num_input_tokens_seen": 114400, "step": 180 }, { "epoch": 0.09696016771488469, "grad_norm": 0.567248523235321, "learning_rate": 2.410901467505241e-06, "loss": 0.9889, "num_input_tokens_seen": 118464, "step": 185 }, { "epoch": 0.09958071278825996, "grad_norm": 0.4108067452907562, "learning_rate": 2.476415094339623e-06, "loss": 1.3216, "num_input_tokens_seen": 121376, "step": 190 }, { "epoch": 0.10220125786163523, "grad_norm": 0.6779152750968933, "learning_rate": 2.5419287211740044e-06, "loss": 1.2849, "num_input_tokens_seen": 124800, "step": 195 }, { "epoch": 0.10482180293501048, "grad_norm": 0.32327407598495483, "learning_rate": 2.6074423480083857e-06, "loss": 1.1513, "num_input_tokens_seen": 127840, "step": 200 }, { "epoch": 0.10744234800838574, "grad_norm": 0.4359290599822998, "learning_rate": 2.6729559748427675e-06, "loss": 1.2899, "num_input_tokens_seen": 131296, "step": 205 }, { "epoch": 0.11006289308176101, "grad_norm": 0.6619760990142822, "learning_rate": 2.738469601677149e-06, "loss": 1.3977, "num_input_tokens_seen": 133760, "step": 210 }, { "epoch": 0.11268343815513626, "grad_norm": 0.437395840883255, "learning_rate": 2.8039832285115303e-06, "loss": 1.191, "num_input_tokens_seen": 137536, "step": 215 }, { "epoch": 0.11530398322851153, "grad_norm": 0.32848072052001953, "learning_rate": 2.869496855345912e-06, "loss": 1.126, "num_input_tokens_seen": 141568, "step": 220 }, { "epoch": 0.1179245283018868, "grad_norm": 0.23541362583637238, "learning_rate": 2.935010482180294e-06, "loss": 1.0802, "num_input_tokens_seen": 145248, "step": 225 }, { "epoch": 0.12054507337526206, "grad_norm": 0.5061608552932739, "learning_rate": 3.0005241090146753e-06, "loss": 1.1806, "num_input_tokens_seen": 148576, "step": 230 }, { "epoch": 0.12316561844863731, "grad_norm": 0.6239566802978516, "learning_rate": 3.0660377358490567e-06, "loss": 1.3176, "num_input_tokens_seen": 151360, "step": 235 }, { "epoch": 0.12578616352201258, "grad_norm": 0.819707989692688, "learning_rate": 3.131551362683438e-06, "loss": 1.1372, "num_input_tokens_seen": 154144, "step": 240 }, { "epoch": 0.12840670859538783, "grad_norm": 0.27636849880218506, "learning_rate": 3.1970649895178203e-06, "loss": 1.1045, "num_input_tokens_seen": 157600, "step": 245 }, { "epoch": 0.1310272536687631, "grad_norm": 0.4789278507232666, "learning_rate": 3.2625786163522012e-06, "loss": 1.0092, "num_input_tokens_seen": 160832, "step": 250 }, { "epoch": 0.13364779874213836, "grad_norm": 0.6146583557128906, "learning_rate": 3.3280922431865826e-06, "loss": 1.2429, "num_input_tokens_seen": 163968, "step": 255 }, { "epoch": 0.13626834381551362, "grad_norm": 0.3992033898830414, "learning_rate": 3.393605870020965e-06, "loss": 1.0797, "num_input_tokens_seen": 168992, "step": 260 }, { "epoch": 0.1388888888888889, "grad_norm": 0.39298102259635925, "learning_rate": 3.459119496855346e-06, "loss": 1.1785, "num_input_tokens_seen": 171616, "step": 265 }, { "epoch": 0.14150943396226415, "grad_norm": 0.5552573204040527, "learning_rate": 3.524633123689727e-06, "loss": 1.0495, "num_input_tokens_seen": 175680, "step": 270 }, { "epoch": 0.1441299790356394, "grad_norm": 0.36026835441589355, "learning_rate": 3.5901467505241094e-06, "loss": 1.0382, "num_input_tokens_seen": 178880, "step": 275 }, { "epoch": 0.14675052410901468, "grad_norm": 0.4963723421096802, "learning_rate": 3.6556603773584908e-06, "loss": 1.0857, "num_input_tokens_seen": 182368, "step": 280 }, { "epoch": 0.14937106918238993, "grad_norm": 0.5430764555931091, "learning_rate": 3.7211740041928726e-06, "loss": 1.0728, "num_input_tokens_seen": 185792, "step": 285 }, { "epoch": 0.1519916142557652, "grad_norm": 0.6508251428604126, "learning_rate": 3.786687631027254e-06, "loss": 1.0193, "num_input_tokens_seen": 189664, "step": 290 }, { "epoch": 0.15461215932914046, "grad_norm": 0.42841723561286926, "learning_rate": 3.852201257861635e-06, "loss": 1.2485, "num_input_tokens_seen": 193632, "step": 295 }, { "epoch": 0.15723270440251572, "grad_norm": 0.35972529649734497, "learning_rate": 3.917714884696017e-06, "loss": 1.3002, "num_input_tokens_seen": 196800, "step": 300 }, { "epoch": 0.159853249475891, "grad_norm": 0.8098708987236023, "learning_rate": 3.9832285115303985e-06, "loss": 1.1288, "num_input_tokens_seen": 199744, "step": 305 }, { "epoch": 0.16247379454926625, "grad_norm": 0.5123339891433716, "learning_rate": 4.04874213836478e-06, "loss": 1.1971, "num_input_tokens_seen": 202624, "step": 310 }, { "epoch": 0.1650943396226415, "grad_norm": 0.5645805597305298, "learning_rate": 4.114255765199162e-06, "loss": 1.3005, "num_input_tokens_seen": 205632, "step": 315 }, { "epoch": 0.16771488469601678, "grad_norm": 0.5561589598655701, "learning_rate": 4.1797693920335435e-06, "loss": 1.0358, "num_input_tokens_seen": 209280, "step": 320 }, { "epoch": 0.17033542976939203, "grad_norm": 0.25697800517082214, "learning_rate": 4.245283018867925e-06, "loss": 1.1406, "num_input_tokens_seen": 213344, "step": 325 }, { "epoch": 0.17295597484276728, "grad_norm": 0.8325145840644836, "learning_rate": 4.310796645702306e-06, "loss": 1.0689, "num_input_tokens_seen": 216448, "step": 330 }, { "epoch": 0.17557651991614256, "grad_norm": 0.5855342149734497, "learning_rate": 4.376310272536688e-06, "loss": 1.3425, "num_input_tokens_seen": 218976, "step": 335 }, { "epoch": 0.17819706498951782, "grad_norm": 0.5904995799064636, "learning_rate": 4.441823899371069e-06, "loss": 1.0107, "num_input_tokens_seen": 222080, "step": 340 }, { "epoch": 0.18081761006289307, "grad_norm": 0.4911738932132721, "learning_rate": 4.507337526205451e-06, "loss": 1.2853, "num_input_tokens_seen": 225120, "step": 345 }, { "epoch": 0.18343815513626835, "grad_norm": 0.4988720715045929, "learning_rate": 4.572851153039833e-06, "loss": 1.521, "num_input_tokens_seen": 227872, "step": 350 }, { "epoch": 0.1860587002096436, "grad_norm": 0.3966246247291565, "learning_rate": 4.638364779874214e-06, "loss": 1.1673, "num_input_tokens_seen": 232320, "step": 355 }, { "epoch": 0.18867924528301888, "grad_norm": 0.6134082078933716, "learning_rate": 4.703878406708595e-06, "loss": 1.2071, "num_input_tokens_seen": 235104, "step": 360 }, { "epoch": 0.19129979035639413, "grad_norm": 0.5700008869171143, "learning_rate": 4.769392033542977e-06, "loss": 1.4113, "num_input_tokens_seen": 237792, "step": 365 }, { "epoch": 0.19392033542976939, "grad_norm": 0.5238808989524841, "learning_rate": 4.834905660377359e-06, "loss": 1.0206, "num_input_tokens_seen": 241728, "step": 370 }, { "epoch": 0.19654088050314467, "grad_norm": 1.1372665166854858, "learning_rate": 4.90041928721174e-06, "loss": 1.1741, "num_input_tokens_seen": 245696, "step": 375 }, { "epoch": 0.19916142557651992, "grad_norm": 0.3099195957183838, "learning_rate": 4.965932914046122e-06, "loss": 1.0009, "num_input_tokens_seen": 249408, "step": 380 }, { "epoch": 0.20178197064989517, "grad_norm": 0.7075557708740234, "learning_rate": 5.031446540880504e-06, "loss": 1.279, "num_input_tokens_seen": 252160, "step": 385 }, { "epoch": 0.20440251572327045, "grad_norm": 0.6453000903129578, "learning_rate": 5.096960167714885e-06, "loss": 1.3687, "num_input_tokens_seen": 255552, "step": 390 }, { "epoch": 0.2070230607966457, "grad_norm": 0.6561174392700195, "learning_rate": 5.162473794549266e-06, "loss": 1.2093, "num_input_tokens_seen": 259072, "step": 395 }, { "epoch": 0.20964360587002095, "grad_norm": 0.458575576543808, "learning_rate": 5.227987421383648e-06, "loss": 1.0637, "num_input_tokens_seen": 263008, "step": 400 }, { "epoch": 0.21226415094339623, "grad_norm": 0.49078989028930664, "learning_rate": 5.2935010482180295e-06, "loss": 1.1409, "num_input_tokens_seen": 267424, "step": 405 }, { "epoch": 0.2148846960167715, "grad_norm": 0.3482169210910797, "learning_rate": 5.359014675052411e-06, "loss": 1.4963, "num_input_tokens_seen": 270720, "step": 410 }, { "epoch": 0.21750524109014674, "grad_norm": 1.12503182888031, "learning_rate": 5.424528301886793e-06, "loss": 1.4239, "num_input_tokens_seen": 273568, "step": 415 }, { "epoch": 0.22012578616352202, "grad_norm": 0.5627071261405945, "learning_rate": 5.4900419287211744e-06, "loss": 1.2563, "num_input_tokens_seen": 276224, "step": 420 }, { "epoch": 0.22274633123689727, "grad_norm": 0.9075502157211304, "learning_rate": 5.555555555555556e-06, "loss": 1.4742, "num_input_tokens_seen": 279040, "step": 425 }, { "epoch": 0.22536687631027252, "grad_norm": 0.487382173538208, "learning_rate": 5.621069182389937e-06, "loss": 1.2592, "num_input_tokens_seen": 282048, "step": 430 }, { "epoch": 0.2279874213836478, "grad_norm": 0.5097913146018982, "learning_rate": 5.686582809224319e-06, "loss": 1.0576, "num_input_tokens_seen": 285184, "step": 435 }, { "epoch": 0.23060796645702306, "grad_norm": 0.5013033747673035, "learning_rate": 5.752096436058701e-06, "loss": 1.0936, "num_input_tokens_seen": 289152, "step": 440 }, { "epoch": 0.23322851153039834, "grad_norm": 0.3392641544342041, "learning_rate": 5.817610062893082e-06, "loss": 1.2126, "num_input_tokens_seen": 292768, "step": 445 }, { "epoch": 0.2358490566037736, "grad_norm": 0.25614792108535767, "learning_rate": 5.8831236897274636e-06, "loss": 1.0918, "num_input_tokens_seen": 296128, "step": 450 }, { "epoch": 0.23846960167714884, "grad_norm": 0.6609297394752502, "learning_rate": 5.948637316561846e-06, "loss": 0.985, "num_input_tokens_seen": 300544, "step": 455 }, { "epoch": 0.24109014675052412, "grad_norm": 0.5884218811988831, "learning_rate": 6.014150943396227e-06, "loss": 1.1768, "num_input_tokens_seen": 303200, "step": 460 }, { "epoch": 0.24371069182389937, "grad_norm": 0.3160429000854492, "learning_rate": 6.079664570230608e-06, "loss": 1.1467, "num_input_tokens_seen": 306368, "step": 465 }, { "epoch": 0.24633123689727462, "grad_norm": 0.4476754367351532, "learning_rate": 6.14517819706499e-06, "loss": 1.1799, "num_input_tokens_seen": 309952, "step": 470 }, { "epoch": 0.2489517819706499, "grad_norm": 0.5567960739135742, "learning_rate": 6.210691823899371e-06, "loss": 1.1701, "num_input_tokens_seen": 312928, "step": 475 }, { "epoch": 0.25157232704402516, "grad_norm": 0.41406768560409546, "learning_rate": 6.2762054507337535e-06, "loss": 1.2423, "num_input_tokens_seen": 316160, "step": 480 }, { "epoch": 0.25419287211740044, "grad_norm": 0.4446223974227905, "learning_rate": 6.341719077568134e-06, "loss": 1.0526, "num_input_tokens_seen": 319136, "step": 485 }, { "epoch": 0.25681341719077566, "grad_norm": 0.3747099041938782, "learning_rate": 6.407232704402516e-06, "loss": 1.1479, "num_input_tokens_seen": 321824, "step": 490 }, { "epoch": 0.25943396226415094, "grad_norm": 1.5601915121078491, "learning_rate": 6.472746331236898e-06, "loss": 1.7964, "num_input_tokens_seen": 324864, "step": 495 }, { "epoch": 0.2620545073375262, "grad_norm": 0.9058985114097595, "learning_rate": 6.538259958071278e-06, "loss": 1.5172, "num_input_tokens_seen": 327264, "step": 500 }, { "epoch": 0.26467505241090145, "grad_norm": 0.3456466495990753, "learning_rate": 6.60377358490566e-06, "loss": 1.3266, "num_input_tokens_seen": 330656, "step": 505 }, { "epoch": 0.2672955974842767, "grad_norm": 0.3264319896697998, "learning_rate": 6.669287211740043e-06, "loss": 0.8761, "num_input_tokens_seen": 335776, "step": 510 }, { "epoch": 0.269916142557652, "grad_norm": 0.6381968259811401, "learning_rate": 6.734800838574423e-06, "loss": 1.5016, "num_input_tokens_seen": 339392, "step": 515 }, { "epoch": 0.27253668763102723, "grad_norm": 0.42570415139198303, "learning_rate": 6.800314465408805e-06, "loss": 1.2501, "num_input_tokens_seen": 342848, "step": 520 }, { "epoch": 0.2751572327044025, "grad_norm": 0.4439488351345062, "learning_rate": 6.865828092243188e-06, "loss": 1.2811, "num_input_tokens_seen": 345600, "step": 525 }, { "epoch": 0.2777777777777778, "grad_norm": 0.34627828001976013, "learning_rate": 6.931341719077568e-06, "loss": 1.3279, "num_input_tokens_seen": 348960, "step": 530 }, { "epoch": 0.280398322851153, "grad_norm": 0.6748606562614441, "learning_rate": 6.9968553459119495e-06, "loss": 1.2226, "num_input_tokens_seen": 351520, "step": 535 }, { "epoch": 0.2830188679245283, "grad_norm": 0.6524113416671753, "learning_rate": 7.062368972746332e-06, "loss": 1.4534, "num_input_tokens_seen": 355456, "step": 540 }, { "epoch": 0.2856394129979036, "grad_norm": 0.50235915184021, "learning_rate": 7.127882599580712e-06, "loss": 1.7858, "num_input_tokens_seen": 358144, "step": 545 }, { "epoch": 0.2882599580712788, "grad_norm": 0.4828290343284607, "learning_rate": 7.1933962264150945e-06, "loss": 0.8638, "num_input_tokens_seen": 363104, "step": 550 }, { "epoch": 0.2908805031446541, "grad_norm": 0.5383928418159485, "learning_rate": 7.258909853249477e-06, "loss": 1.0793, "num_input_tokens_seen": 367200, "step": 555 }, { "epoch": 0.29350104821802936, "grad_norm": 0.3656754791736603, "learning_rate": 7.324423480083858e-06, "loss": 1.3601, "num_input_tokens_seen": 370176, "step": 560 }, { "epoch": 0.29612159329140464, "grad_norm": 0.8277395963668823, "learning_rate": 7.3899371069182395e-06, "loss": 1.599, "num_input_tokens_seen": 372832, "step": 565 }, { "epoch": 0.29874213836477986, "grad_norm": 0.34222114086151123, "learning_rate": 7.455450733752621e-06, "loss": 1.3142, "num_input_tokens_seen": 375296, "step": 570 }, { "epoch": 0.30136268343815514, "grad_norm": 0.44216784834861755, "learning_rate": 7.520964360587003e-06, "loss": 1.069, "num_input_tokens_seen": 378464, "step": 575 }, { "epoch": 0.3039832285115304, "grad_norm": 0.7378522753715515, "learning_rate": 7.586477987421384e-06, "loss": 1.1891, "num_input_tokens_seen": 381664, "step": 580 }, { "epoch": 0.30660377358490565, "grad_norm": 0.557349681854248, "learning_rate": 7.651991614255765e-06, "loss": 1.1521, "num_input_tokens_seen": 385248, "step": 585 }, { "epoch": 0.30922431865828093, "grad_norm": 0.6604116559028625, "learning_rate": 7.717505241090148e-06, "loss": 1.3146, "num_input_tokens_seen": 387840, "step": 590 }, { "epoch": 0.3118448637316562, "grad_norm": 0.594268798828125, "learning_rate": 7.783018867924528e-06, "loss": 1.1369, "num_input_tokens_seen": 391424, "step": 595 }, { "epoch": 0.31446540880503143, "grad_norm": 0.7351045608520508, "learning_rate": 7.848532494758911e-06, "loss": 1.1442, "num_input_tokens_seen": 394656, "step": 600 }, { "epoch": 0.3170859538784067, "grad_norm": 0.5761642456054688, "learning_rate": 7.914046121593292e-06, "loss": 1.3099, "num_input_tokens_seen": 397824, "step": 605 }, { "epoch": 0.319706498951782, "grad_norm": 0.6263448596000671, "learning_rate": 7.979559748427674e-06, "loss": 1.277, "num_input_tokens_seen": 400544, "step": 610 }, { "epoch": 0.3223270440251572, "grad_norm": 0.5970672965049744, "learning_rate": 8.045073375262055e-06, "loss": 1.2009, "num_input_tokens_seen": 403552, "step": 615 }, { "epoch": 0.3249475890985325, "grad_norm": 0.8647836446762085, "learning_rate": 8.110587002096436e-06, "loss": 1.2886, "num_input_tokens_seen": 406464, "step": 620 }, { "epoch": 0.3275681341719078, "grad_norm": 0.6966333985328674, "learning_rate": 8.176100628930818e-06, "loss": 1.012, "num_input_tokens_seen": 412032, "step": 625 }, { "epoch": 0.330188679245283, "grad_norm": 0.41781243681907654, "learning_rate": 8.241614255765199e-06, "loss": 1.1257, "num_input_tokens_seen": 415104, "step": 630 }, { "epoch": 0.3328092243186583, "grad_norm": 0.5355955958366394, "learning_rate": 8.307127882599582e-06, "loss": 1.101, "num_input_tokens_seen": 418368, "step": 635 }, { "epoch": 0.33542976939203356, "grad_norm": 0.6937329173088074, "learning_rate": 8.372641509433962e-06, "loss": 1.5057, "num_input_tokens_seen": 420928, "step": 640 }, { "epoch": 0.3380503144654088, "grad_norm": 0.6481752991676331, "learning_rate": 8.438155136268345e-06, "loss": 1.0869, "num_input_tokens_seen": 423840, "step": 645 }, { "epoch": 0.34067085953878407, "grad_norm": 0.34509727358818054, "learning_rate": 8.503668763102726e-06, "loss": 1.0405, "num_input_tokens_seen": 427904, "step": 650 }, { "epoch": 0.34329140461215935, "grad_norm": 0.40071260929107666, "learning_rate": 8.569182389937106e-06, "loss": 1.5802, "num_input_tokens_seen": 430304, "step": 655 }, { "epoch": 0.34591194968553457, "grad_norm": 0.7012487053871155, "learning_rate": 8.634696016771489e-06, "loss": 1.3793, "num_input_tokens_seen": 433088, "step": 660 }, { "epoch": 0.34853249475890985, "grad_norm": 0.4679982364177704, "learning_rate": 8.70020964360587e-06, "loss": 1.0778, "num_input_tokens_seen": 437120, "step": 665 }, { "epoch": 0.35115303983228513, "grad_norm": 0.37659478187561035, "learning_rate": 8.765723270440252e-06, "loss": 0.8481, "num_input_tokens_seen": 441440, "step": 670 }, { "epoch": 0.35377358490566035, "grad_norm": 0.7007670402526855, "learning_rate": 8.831236897274633e-06, "loss": 1.3285, "num_input_tokens_seen": 444000, "step": 675 }, { "epoch": 0.35639412997903563, "grad_norm": 0.4591682553291321, "learning_rate": 8.896750524109015e-06, "loss": 1.3002, "num_input_tokens_seen": 447232, "step": 680 }, { "epoch": 0.3590146750524109, "grad_norm": 0.49725407361984253, "learning_rate": 8.962264150943396e-06, "loss": 1.0864, "num_input_tokens_seen": 450432, "step": 685 }, { "epoch": 0.36163522012578614, "grad_norm": 0.5663197040557861, "learning_rate": 9.027777777777777e-06, "loss": 1.067, "num_input_tokens_seen": 453440, "step": 690 }, { "epoch": 0.3642557651991614, "grad_norm": 0.545314371585846, "learning_rate": 9.09329140461216e-06, "loss": 1.1065, "num_input_tokens_seen": 456352, "step": 695 }, { "epoch": 0.3668763102725367, "grad_norm": 0.5855574607849121, "learning_rate": 9.15880503144654e-06, "loss": 1.1021, "num_input_tokens_seen": 459712, "step": 700 }, { "epoch": 0.3694968553459119, "grad_norm": 0.46587902307510376, "learning_rate": 9.224318658280923e-06, "loss": 0.976, "num_input_tokens_seen": 463168, "step": 705 }, { "epoch": 0.3721174004192872, "grad_norm": 0.748126208782196, "learning_rate": 9.289832285115305e-06, "loss": 1.4023, "num_input_tokens_seen": 465824, "step": 710 }, { "epoch": 0.3747379454926625, "grad_norm": 0.8102132678031921, "learning_rate": 9.355345911949686e-06, "loss": 1.5295, "num_input_tokens_seen": 468032, "step": 715 }, { "epoch": 0.37735849056603776, "grad_norm": 0.4549533426761627, "learning_rate": 9.420859538784067e-06, "loss": 0.9746, "num_input_tokens_seen": 471392, "step": 720 }, { "epoch": 0.379979035639413, "grad_norm": 0.708053708076477, "learning_rate": 9.486373165618449e-06, "loss": 1.1771, "num_input_tokens_seen": 474400, "step": 725 }, { "epoch": 0.38259958071278827, "grad_norm": 0.6716095209121704, "learning_rate": 9.551886792452832e-06, "loss": 1.2147, "num_input_tokens_seen": 477792, "step": 730 }, { "epoch": 0.38522012578616355, "grad_norm": 0.48681798577308655, "learning_rate": 9.617400419287211e-06, "loss": 1.2624, "num_input_tokens_seen": 481216, "step": 735 }, { "epoch": 0.38784067085953877, "grad_norm": 0.3495166301727295, "learning_rate": 9.682914046121595e-06, "loss": 1.2311, "num_input_tokens_seen": 484928, "step": 740 }, { "epoch": 0.39046121593291405, "grad_norm": 0.6146240830421448, "learning_rate": 9.748427672955976e-06, "loss": 1.1001, "num_input_tokens_seen": 488224, "step": 745 }, { "epoch": 0.39308176100628933, "grad_norm": 0.4001065194606781, "learning_rate": 9.813941299790357e-06, "loss": 1.1227, "num_input_tokens_seen": 491840, "step": 750 }, { "epoch": 0.39570230607966456, "grad_norm": 1.0418214797973633, "learning_rate": 9.879454926624739e-06, "loss": 1.2834, "num_input_tokens_seen": 495232, "step": 755 }, { "epoch": 0.39832285115303984, "grad_norm": 0.3469036817550659, "learning_rate": 9.94496855345912e-06, "loss": 1.0718, "num_input_tokens_seen": 498336, "step": 760 }, { "epoch": 0.4009433962264151, "grad_norm": 0.5466294884681702, "learning_rate": 1.0010482180293501e-05, "loss": 1.4996, "num_input_tokens_seen": 500448, "step": 765 }, { "epoch": 0.40356394129979034, "grad_norm": 0.4380440413951874, "learning_rate": 1.0075995807127883e-05, "loss": 1.0784, "num_input_tokens_seen": 503520, "step": 770 }, { "epoch": 0.4061844863731656, "grad_norm": 0.49803292751312256, "learning_rate": 1.0141509433962266e-05, "loss": 1.2442, "num_input_tokens_seen": 506208, "step": 775 }, { "epoch": 0.4088050314465409, "grad_norm": 0.6512178778648376, "learning_rate": 1.0207023060796646e-05, "loss": 1.1087, "num_input_tokens_seen": 509152, "step": 780 }, { "epoch": 0.4114255765199161, "grad_norm": 0.4601573348045349, "learning_rate": 1.0272536687631027e-05, "loss": 1.1616, "num_input_tokens_seen": 513184, "step": 785 }, { "epoch": 0.4140461215932914, "grad_norm": 0.625270664691925, "learning_rate": 1.033805031446541e-05, "loss": 1.4134, "num_input_tokens_seen": 516128, "step": 790 }, { "epoch": 0.4166666666666667, "grad_norm": 0.3560948669910431, "learning_rate": 1.040356394129979e-05, "loss": 0.9256, "num_input_tokens_seen": 519264, "step": 795 }, { "epoch": 0.4192872117400419, "grad_norm": 0.639944851398468, "learning_rate": 1.0469077568134173e-05, "loss": 1.426, "num_input_tokens_seen": 522304, "step": 800 }, { "epoch": 0.4219077568134172, "grad_norm": 0.7252195477485657, "learning_rate": 1.0534591194968554e-05, "loss": 0.9788, "num_input_tokens_seen": 525600, "step": 805 }, { "epoch": 0.42452830188679247, "grad_norm": 0.6154329776763916, "learning_rate": 1.0600104821802936e-05, "loss": 1.2652, "num_input_tokens_seen": 528608, "step": 810 }, { "epoch": 0.4271488469601677, "grad_norm": 0.5136966109275818, "learning_rate": 1.0665618448637317e-05, "loss": 1.0793, "num_input_tokens_seen": 531232, "step": 815 }, { "epoch": 0.429769392033543, "grad_norm": 0.6089774370193481, "learning_rate": 1.0731132075471698e-05, "loss": 0.977, "num_input_tokens_seen": 534208, "step": 820 }, { "epoch": 0.43238993710691825, "grad_norm": 0.5699732303619385, "learning_rate": 1.079664570230608e-05, "loss": 1.0358, "num_input_tokens_seen": 537792, "step": 825 }, { "epoch": 0.4350104821802935, "grad_norm": 0.36522939801216125, "learning_rate": 1.0862159329140461e-05, "loss": 1.2924, "num_input_tokens_seen": 540448, "step": 830 }, { "epoch": 0.43763102725366876, "grad_norm": 0.4640253484249115, "learning_rate": 1.0927672955974844e-05, "loss": 1.0561, "num_input_tokens_seen": 543072, "step": 835 }, { "epoch": 0.44025157232704404, "grad_norm": 0.24464552104473114, "learning_rate": 1.0993186582809224e-05, "loss": 0.931, "num_input_tokens_seen": 547456, "step": 840 }, { "epoch": 0.44287211740041926, "grad_norm": 0.358370840549469, "learning_rate": 1.1058700209643607e-05, "loss": 1.0121, "num_input_tokens_seen": 551968, "step": 845 }, { "epoch": 0.44549266247379454, "grad_norm": 0.7485343813896179, "learning_rate": 1.1124213836477988e-05, "loss": 1.2608, "num_input_tokens_seen": 554400, "step": 850 }, { "epoch": 0.4481132075471698, "grad_norm": 0.4224436283111572, "learning_rate": 1.118972746331237e-05, "loss": 1.0251, "num_input_tokens_seen": 557184, "step": 855 }, { "epoch": 0.45073375262054505, "grad_norm": 0.4095068871974945, "learning_rate": 1.1255241090146751e-05, "loss": 0.9941, "num_input_tokens_seen": 559840, "step": 860 }, { "epoch": 0.4533542976939203, "grad_norm": 0.5308467745780945, "learning_rate": 1.1320754716981132e-05, "loss": 1.0647, "num_input_tokens_seen": 562624, "step": 865 }, { "epoch": 0.4559748427672956, "grad_norm": 0.6067817211151123, "learning_rate": 1.1386268343815514e-05, "loss": 0.9817, "num_input_tokens_seen": 565888, "step": 870 }, { "epoch": 0.4585953878406709, "grad_norm": 0.6586860418319702, "learning_rate": 1.1451781970649895e-05, "loss": 0.9499, "num_input_tokens_seen": 569088, "step": 875 }, { "epoch": 0.4612159329140461, "grad_norm": 0.4403952658176422, "learning_rate": 1.1517295597484278e-05, "loss": 1.0602, "num_input_tokens_seen": 571904, "step": 880 }, { "epoch": 0.4638364779874214, "grad_norm": 0.5454542636871338, "learning_rate": 1.158280922431866e-05, "loss": 0.8931, "num_input_tokens_seen": 575232, "step": 885 }, { "epoch": 0.46645702306079667, "grad_norm": 0.8617185950279236, "learning_rate": 1.164832285115304e-05, "loss": 1.1701, "num_input_tokens_seen": 578432, "step": 890 }, { "epoch": 0.4690775681341719, "grad_norm": 0.8187503218650818, "learning_rate": 1.1713836477987422e-05, "loss": 1.4088, "num_input_tokens_seen": 582112, "step": 895 }, { "epoch": 0.4716981132075472, "grad_norm": 0.4108867645263672, "learning_rate": 1.1779350104821804e-05, "loss": 1.4216, "num_input_tokens_seen": 584608, "step": 900 }, { "epoch": 0.47431865828092246, "grad_norm": 0.4324195384979248, "learning_rate": 1.1844863731656185e-05, "loss": 0.834, "num_input_tokens_seen": 587808, "step": 905 }, { "epoch": 0.4769392033542977, "grad_norm": 0.37321436405181885, "learning_rate": 1.1910377358490566e-05, "loss": 0.9578, "num_input_tokens_seen": 590848, "step": 910 }, { "epoch": 0.47955974842767296, "grad_norm": 0.4321545362472534, "learning_rate": 1.197589098532495e-05, "loss": 0.9366, "num_input_tokens_seen": 595232, "step": 915 }, { "epoch": 0.48218029350104824, "grad_norm": 0.32816100120544434, "learning_rate": 1.204140461215933e-05, "loss": 0.9047, "num_input_tokens_seen": 598560, "step": 920 }, { "epoch": 0.48480083857442346, "grad_norm": 0.4909590482711792, "learning_rate": 1.210691823899371e-05, "loss": 0.8344, "num_input_tokens_seen": 606336, "step": 925 }, { "epoch": 0.48742138364779874, "grad_norm": 0.4354963004589081, "learning_rate": 1.2172431865828094e-05, "loss": 1.0788, "num_input_tokens_seen": 609728, "step": 930 }, { "epoch": 0.490041928721174, "grad_norm": 0.7200040221214294, "learning_rate": 1.2237945492662473e-05, "loss": 1.1361, "num_input_tokens_seen": 612256, "step": 935 }, { "epoch": 0.49266247379454925, "grad_norm": 0.4401719272136688, "learning_rate": 1.2303459119496856e-05, "loss": 1.1144, "num_input_tokens_seen": 615520, "step": 940 }, { "epoch": 0.49528301886792453, "grad_norm": 0.5191959738731384, "learning_rate": 1.2368972746331238e-05, "loss": 0.9684, "num_input_tokens_seen": 618880, "step": 945 }, { "epoch": 0.4979035639412998, "grad_norm": 0.2928885221481323, "learning_rate": 1.243448637316562e-05, "loss": 1.0479, "num_input_tokens_seen": 621664, "step": 950 }, { "epoch": 0.500524109014675, "grad_norm": 0.43005189299583435, "learning_rate": 1.25e-05, "loss": 1.1356, "num_input_tokens_seen": 624736, "step": 955 }, { "epoch": 0.5031446540880503, "grad_norm": 0.6799668073654175, "learning_rate": 1.2565513626834382e-05, "loss": 1.0124, "num_input_tokens_seen": 628576, "step": 960 }, { "epoch": 0.5057651991614256, "grad_norm": 0.46261468529701233, "learning_rate": 1.2631027253668765e-05, "loss": 0.94, "num_input_tokens_seen": 632288, "step": 965 }, { "epoch": 0.5083857442348009, "grad_norm": 0.737153947353363, "learning_rate": 1.2696540880503146e-05, "loss": 1.0611, "num_input_tokens_seen": 635008, "step": 970 }, { "epoch": 0.5110062893081762, "grad_norm": 0.42279648780822754, "learning_rate": 1.2762054507337526e-05, "loss": 0.8459, "num_input_tokens_seen": 638720, "step": 975 }, { "epoch": 0.5136268343815513, "grad_norm": 0.4849759638309479, "learning_rate": 1.2827568134171907e-05, "loss": 1.251, "num_input_tokens_seen": 641344, "step": 980 }, { "epoch": 0.5162473794549266, "grad_norm": 0.5231432318687439, "learning_rate": 1.289308176100629e-05, "loss": 0.882, "num_input_tokens_seen": 645600, "step": 985 }, { "epoch": 0.5188679245283019, "grad_norm": 0.4112459123134613, "learning_rate": 1.2958595387840672e-05, "loss": 1.2927, "num_input_tokens_seen": 648672, "step": 990 }, { "epoch": 0.5214884696016772, "grad_norm": 0.7864604592323303, "learning_rate": 1.3024109014675053e-05, "loss": 1.0567, "num_input_tokens_seen": 652416, "step": 995 }, { "epoch": 0.5241090146750524, "grad_norm": 0.4505275785923004, "learning_rate": 1.3089622641509436e-05, "loss": 1.0525, "num_input_tokens_seen": 655712, "step": 1000 }, { "epoch": 0.5267295597484277, "grad_norm": 0.6822474002838135, "learning_rate": 1.3155136268343814e-05, "loss": 1.1334, "num_input_tokens_seen": 657760, "step": 1005 }, { "epoch": 0.5293501048218029, "grad_norm": 0.4999997913837433, "learning_rate": 1.3220649895178197e-05, "loss": 1.2484, "num_input_tokens_seen": 660640, "step": 1010 }, { "epoch": 0.5319706498951782, "grad_norm": 0.8899758458137512, "learning_rate": 1.3286163522012579e-05, "loss": 1.0999, "num_input_tokens_seen": 666496, "step": 1015 }, { "epoch": 0.5345911949685535, "grad_norm": 0.4820021092891693, "learning_rate": 1.3351677148846962e-05, "loss": 1.0349, "num_input_tokens_seen": 669088, "step": 1020 }, { "epoch": 0.5372117400419287, "grad_norm": 0.4598056674003601, "learning_rate": 1.3417190775681343e-05, "loss": 1.0595, "num_input_tokens_seen": 672192, "step": 1025 }, { "epoch": 0.539832285115304, "grad_norm": 0.4982270896434784, "learning_rate": 1.3482704402515725e-05, "loss": 1.0575, "num_input_tokens_seen": 675104, "step": 1030 }, { "epoch": 0.5424528301886793, "grad_norm": 0.40331143140792847, "learning_rate": 1.3548218029350104e-05, "loss": 0.8921, "num_input_tokens_seen": 678368, "step": 1035 }, { "epoch": 0.5450733752620545, "grad_norm": 0.34076985716819763, "learning_rate": 1.3613731656184486e-05, "loss": 1.0186, "num_input_tokens_seen": 681600, "step": 1040 }, { "epoch": 0.5476939203354297, "grad_norm": 0.4976576864719391, "learning_rate": 1.3679245283018869e-05, "loss": 0.952, "num_input_tokens_seen": 684576, "step": 1045 }, { "epoch": 0.550314465408805, "grad_norm": 0.6217771768569946, "learning_rate": 1.374475890985325e-05, "loss": 1.0397, "num_input_tokens_seen": 688832, "step": 1050 }, { "epoch": 0.5529350104821803, "grad_norm": 0.46819010376930237, "learning_rate": 1.3810272536687632e-05, "loss": 1.1627, "num_input_tokens_seen": 691520, "step": 1055 }, { "epoch": 0.5555555555555556, "grad_norm": 0.4716970920562744, "learning_rate": 1.3875786163522015e-05, "loss": 0.9268, "num_input_tokens_seen": 694336, "step": 1060 }, { "epoch": 0.5581761006289309, "grad_norm": 0.3118034601211548, "learning_rate": 1.3941299790356394e-05, "loss": 0.8139, "num_input_tokens_seen": 698016, "step": 1065 }, { "epoch": 0.560796645702306, "grad_norm": 0.49243253469467163, "learning_rate": 1.4006813417190776e-05, "loss": 1.0292, "num_input_tokens_seen": 700896, "step": 1070 }, { "epoch": 0.5634171907756813, "grad_norm": 0.3799131214618683, "learning_rate": 1.4072327044025157e-05, "loss": 0.7465, "num_input_tokens_seen": 704416, "step": 1075 }, { "epoch": 0.5660377358490566, "grad_norm": 0.57357257604599, "learning_rate": 1.413784067085954e-05, "loss": 1.1116, "num_input_tokens_seen": 707648, "step": 1080 }, { "epoch": 0.5686582809224319, "grad_norm": 0.5370060801506042, "learning_rate": 1.4203354297693922e-05, "loss": 0.9998, "num_input_tokens_seen": 711200, "step": 1085 }, { "epoch": 0.5712788259958071, "grad_norm": 0.34883683919906616, "learning_rate": 1.4268867924528303e-05, "loss": 0.8635, "num_input_tokens_seen": 715200, "step": 1090 }, { "epoch": 0.5738993710691824, "grad_norm": 0.4455840587615967, "learning_rate": 1.4334381551362683e-05, "loss": 0.9765, "num_input_tokens_seen": 718080, "step": 1095 }, { "epoch": 0.5765199161425576, "grad_norm": 0.38782423734664917, "learning_rate": 1.4399895178197064e-05, "loss": 0.9889, "num_input_tokens_seen": 722112, "step": 1100 }, { "epoch": 0.5791404612159329, "grad_norm": 0.26700231432914734, "learning_rate": 1.4465408805031447e-05, "loss": 1.0783, "num_input_tokens_seen": 725696, "step": 1105 }, { "epoch": 0.5817610062893082, "grad_norm": 0.37282243371009827, "learning_rate": 1.4530922431865828e-05, "loss": 0.8782, "num_input_tokens_seen": 729824, "step": 1110 }, { "epoch": 0.5843815513626834, "grad_norm": 0.4320385456085205, "learning_rate": 1.4596436058700211e-05, "loss": 0.8559, "num_input_tokens_seen": 734528, "step": 1115 }, { "epoch": 0.5870020964360587, "grad_norm": 0.3818330764770508, "learning_rate": 1.4661949685534593e-05, "loss": 0.8806, "num_input_tokens_seen": 737568, "step": 1120 }, { "epoch": 0.589622641509434, "grad_norm": 0.5440534949302673, "learning_rate": 1.4727463312368974e-05, "loss": 1.088, "num_input_tokens_seen": 740256, "step": 1125 }, { "epoch": 0.5922431865828093, "grad_norm": 0.3373180329799652, "learning_rate": 1.4792976939203354e-05, "loss": 0.8085, "num_input_tokens_seen": 743968, "step": 1130 }, { "epoch": 0.5948637316561844, "grad_norm": 0.5914024710655212, "learning_rate": 1.4858490566037735e-05, "loss": 1.1824, "num_input_tokens_seen": 746528, "step": 1135 }, { "epoch": 0.5974842767295597, "grad_norm": 0.3965953588485718, "learning_rate": 1.4924004192872118e-05, "loss": 1.0398, "num_input_tokens_seen": 749632, "step": 1140 }, { "epoch": 0.600104821802935, "grad_norm": 0.5121103525161743, "learning_rate": 1.49895178197065e-05, "loss": 0.9965, "num_input_tokens_seen": 752832, "step": 1145 }, { "epoch": 0.6027253668763103, "grad_norm": 0.36844202876091003, "learning_rate": 1.5055031446540883e-05, "loss": 0.9593, "num_input_tokens_seen": 756032, "step": 1150 }, { "epoch": 0.6053459119496856, "grad_norm": 0.5528642535209656, "learning_rate": 1.5120545073375264e-05, "loss": 0.9972, "num_input_tokens_seen": 759488, "step": 1155 }, { "epoch": 0.6079664570230608, "grad_norm": 0.6574069857597351, "learning_rate": 1.5186058700209644e-05, "loss": 0.9106, "num_input_tokens_seen": 762304, "step": 1160 }, { "epoch": 0.610587002096436, "grad_norm": 0.3896234929561615, "learning_rate": 1.5251572327044025e-05, "loss": 0.9155, "num_input_tokens_seen": 765152, "step": 1165 }, { "epoch": 0.6132075471698113, "grad_norm": 0.6227007508277893, "learning_rate": 1.5317085953878407e-05, "loss": 0.9203, "num_input_tokens_seen": 768608, "step": 1170 }, { "epoch": 0.6158280922431866, "grad_norm": 0.4743615388870239, "learning_rate": 1.538259958071279e-05, "loss": 0.9321, "num_input_tokens_seen": 771328, "step": 1175 }, { "epoch": 0.6184486373165619, "grad_norm": 0.3214273154735565, "learning_rate": 1.5448113207547173e-05, "loss": 1.2016, "num_input_tokens_seen": 774464, "step": 1180 }, { "epoch": 0.6210691823899371, "grad_norm": 0.7018495798110962, "learning_rate": 1.5513626834381552e-05, "loss": 0.9923, "num_input_tokens_seen": 777824, "step": 1185 }, { "epoch": 0.6236897274633124, "grad_norm": 0.27893126010894775, "learning_rate": 1.5579140461215932e-05, "loss": 0.7013, "num_input_tokens_seen": 782464, "step": 1190 }, { "epoch": 0.6263102725366876, "grad_norm": 1.215631365776062, "learning_rate": 1.5644654088050315e-05, "loss": 1.2285, "num_input_tokens_seen": 785632, "step": 1195 }, { "epoch": 0.6289308176100629, "grad_norm": 0.24758245050907135, "learning_rate": 1.5710167714884695e-05, "loss": 0.9581, "num_input_tokens_seen": 789344, "step": 1200 }, { "epoch": 0.6315513626834381, "grad_norm": 0.5190907716751099, "learning_rate": 1.5775681341719078e-05, "loss": 0.9286, "num_input_tokens_seen": 792928, "step": 1205 }, { "epoch": 0.6341719077568134, "grad_norm": 0.9622239470481873, "learning_rate": 1.584119496855346e-05, "loss": 0.9638, "num_input_tokens_seen": 796032, "step": 1210 }, { "epoch": 0.6367924528301887, "grad_norm": 0.33297741413116455, "learning_rate": 1.5906708595387844e-05, "loss": 1.1693, "num_input_tokens_seen": 798464, "step": 1215 }, { "epoch": 0.639412997903564, "grad_norm": 0.4518563151359558, "learning_rate": 1.597222222222222e-05, "loss": 1.1496, "num_input_tokens_seen": 801088, "step": 1220 }, { "epoch": 0.6420335429769392, "grad_norm": 0.783501923084259, "learning_rate": 1.6037735849056604e-05, "loss": 1.044, "num_input_tokens_seen": 803872, "step": 1225 }, { "epoch": 0.6446540880503144, "grad_norm": 0.6379638910293579, "learning_rate": 1.6103249475890987e-05, "loss": 0.9689, "num_input_tokens_seen": 806784, "step": 1230 }, { "epoch": 0.6472746331236897, "grad_norm": 0.47743454575538635, "learning_rate": 1.6168763102725366e-05, "loss": 0.6881, "num_input_tokens_seen": 810656, "step": 1235 }, { "epoch": 0.649895178197065, "grad_norm": 0.3239179849624634, "learning_rate": 1.623427672955975e-05, "loss": 0.9392, "num_input_tokens_seen": 813568, "step": 1240 }, { "epoch": 0.6525157232704403, "grad_norm": 0.45313066244125366, "learning_rate": 1.6299790356394132e-05, "loss": 1.1925, "num_input_tokens_seen": 816800, "step": 1245 }, { "epoch": 0.6551362683438156, "grad_norm": 0.34106430411338806, "learning_rate": 1.6365303983228512e-05, "loss": 1.1154, "num_input_tokens_seen": 820512, "step": 1250 }, { "epoch": 0.6577568134171907, "grad_norm": 0.5667005181312561, "learning_rate": 1.6430817610062892e-05, "loss": 0.7996, "num_input_tokens_seen": 824096, "step": 1255 }, { "epoch": 0.660377358490566, "grad_norm": 0.4254443347454071, "learning_rate": 1.6496331236897275e-05, "loss": 0.9403, "num_input_tokens_seen": 827744, "step": 1260 }, { "epoch": 0.6629979035639413, "grad_norm": 0.5989857316017151, "learning_rate": 1.6561844863731658e-05, "loss": 1.0719, "num_input_tokens_seen": 830944, "step": 1265 }, { "epoch": 0.6656184486373166, "grad_norm": 0.4716537892818451, "learning_rate": 1.6627358490566038e-05, "loss": 0.9056, "num_input_tokens_seen": 834176, "step": 1270 }, { "epoch": 0.6682389937106918, "grad_norm": 0.563991904258728, "learning_rate": 1.669287211740042e-05, "loss": 1.321, "num_input_tokens_seen": 837664, "step": 1275 }, { "epoch": 0.6708595387840671, "grad_norm": 0.7641096115112305, "learning_rate": 1.6758385744234804e-05, "loss": 0.984, "num_input_tokens_seen": 840864, "step": 1280 }, { "epoch": 0.6734800838574424, "grad_norm": 0.337348610162735, "learning_rate": 1.6823899371069183e-05, "loss": 0.9292, "num_input_tokens_seen": 844576, "step": 1285 }, { "epoch": 0.6761006289308176, "grad_norm": 0.3837306499481201, "learning_rate": 1.6889412997903563e-05, "loss": 1.056, "num_input_tokens_seen": 847328, "step": 1290 }, { "epoch": 0.6787211740041929, "grad_norm": 0.3825112283229828, "learning_rate": 1.6954926624737946e-05, "loss": 0.9225, "num_input_tokens_seen": 850048, "step": 1295 }, { "epoch": 0.6813417190775681, "grad_norm": 0.2233346402645111, "learning_rate": 1.702044025157233e-05, "loss": 1.0493, "num_input_tokens_seen": 853568, "step": 1300 }, { "epoch": 0.6839622641509434, "grad_norm": 0.7766002416610718, "learning_rate": 1.708595387840671e-05, "loss": 0.7993, "num_input_tokens_seen": 857024, "step": 1305 }, { "epoch": 0.6865828092243187, "grad_norm": 0.29439103603363037, "learning_rate": 1.7151467505241092e-05, "loss": 0.8289, "num_input_tokens_seen": 861568, "step": 1310 }, { "epoch": 0.689203354297694, "grad_norm": 0.8092571496963501, "learning_rate": 1.7216981132075472e-05, "loss": 1.0857, "num_input_tokens_seen": 864384, "step": 1315 }, { "epoch": 0.6918238993710691, "grad_norm": 0.4487694203853607, "learning_rate": 1.7282494758909855e-05, "loss": 0.8725, "num_input_tokens_seen": 867136, "step": 1320 }, { "epoch": 0.6944444444444444, "grad_norm": 0.6550716161727905, "learning_rate": 1.7348008385744234e-05, "loss": 0.9912, "num_input_tokens_seen": 870080, "step": 1325 }, { "epoch": 0.6970649895178197, "grad_norm": 0.8939180374145508, "learning_rate": 1.7413522012578618e-05, "loss": 1.1101, "num_input_tokens_seen": 873568, "step": 1330 }, { "epoch": 0.699685534591195, "grad_norm": 0.5780777335166931, "learning_rate": 1.7479035639413e-05, "loss": 1.0898, "num_input_tokens_seen": 876192, "step": 1335 }, { "epoch": 0.7023060796645703, "grad_norm": 0.6398524641990662, "learning_rate": 1.754454926624738e-05, "loss": 0.9351, "num_input_tokens_seen": 878784, "step": 1340 }, { "epoch": 0.7049266247379455, "grad_norm": 0.34129148721694946, "learning_rate": 1.761006289308176e-05, "loss": 0.9526, "num_input_tokens_seen": 882304, "step": 1345 }, { "epoch": 0.7075471698113207, "grad_norm": 0.5889203548431396, "learning_rate": 1.7675576519916143e-05, "loss": 0.8282, "num_input_tokens_seen": 885792, "step": 1350 }, { "epoch": 0.710167714884696, "grad_norm": 0.7708578109741211, "learning_rate": 1.7741090146750526e-05, "loss": 1.0284, "num_input_tokens_seen": 887840, "step": 1355 }, { "epoch": 0.7127882599580713, "grad_norm": 0.6336182951927185, "learning_rate": 1.7806603773584906e-05, "loss": 0.9486, "num_input_tokens_seen": 890848, "step": 1360 }, { "epoch": 0.7154088050314465, "grad_norm": 0.35924410820007324, "learning_rate": 1.787211740041929e-05, "loss": 0.9969, "num_input_tokens_seen": 894304, "step": 1365 }, { "epoch": 0.7180293501048218, "grad_norm": 0.5656065344810486, "learning_rate": 1.7937631027253672e-05, "loss": 0.9716, "num_input_tokens_seen": 897600, "step": 1370 }, { "epoch": 0.7206498951781971, "grad_norm": 1.1483134031295776, "learning_rate": 1.8003144654088048e-05, "loss": 0.8744, "num_input_tokens_seen": 900928, "step": 1375 }, { "epoch": 0.7232704402515723, "grad_norm": 0.5080873966217041, "learning_rate": 1.806865828092243e-05, "loss": 0.8824, "num_input_tokens_seen": 903872, "step": 1380 }, { "epoch": 0.7258909853249476, "grad_norm": 0.48686087131500244, "learning_rate": 1.8134171907756814e-05, "loss": 0.6725, "num_input_tokens_seen": 907456, "step": 1385 }, { "epoch": 0.7285115303983228, "grad_norm": 0.291760116815567, "learning_rate": 1.8199685534591197e-05, "loss": 0.7021, "num_input_tokens_seen": 911072, "step": 1390 }, { "epoch": 0.7311320754716981, "grad_norm": 0.4804026186466217, "learning_rate": 1.8265199161425577e-05, "loss": 0.831, "num_input_tokens_seen": 914112, "step": 1395 }, { "epoch": 0.7337526205450734, "grad_norm": 1.4271572828292847, "learning_rate": 1.833071278825996e-05, "loss": 0.9639, "num_input_tokens_seen": 917280, "step": 1400 }, { "epoch": 0.7363731656184487, "grad_norm": 0.9816332459449768, "learning_rate": 1.839622641509434e-05, "loss": 0.9236, "num_input_tokens_seen": 919936, "step": 1405 }, { "epoch": 0.7389937106918238, "grad_norm": 0.6196041107177734, "learning_rate": 1.846174004192872e-05, "loss": 1.1059, "num_input_tokens_seen": 922144, "step": 1410 }, { "epoch": 0.7416142557651991, "grad_norm": 0.4911401867866516, "learning_rate": 1.8527253668763103e-05, "loss": 0.9887, "num_input_tokens_seen": 925184, "step": 1415 }, { "epoch": 0.7442348008385744, "grad_norm": 0.5584189295768738, "learning_rate": 1.8592767295597486e-05, "loss": 0.9843, "num_input_tokens_seen": 928096, "step": 1420 }, { "epoch": 0.7468553459119497, "grad_norm": 0.7233555316925049, "learning_rate": 1.865828092243187e-05, "loss": 0.9368, "num_input_tokens_seen": 930976, "step": 1425 }, { "epoch": 0.749475890985325, "grad_norm": 0.8225273489952087, "learning_rate": 1.872379454926625e-05, "loss": 0.8152, "num_input_tokens_seen": 934336, "step": 1430 }, { "epoch": 0.7520964360587002, "grad_norm": 0.5971591472625732, "learning_rate": 1.878930817610063e-05, "loss": 0.9969, "num_input_tokens_seen": 936544, "step": 1435 }, { "epoch": 0.7547169811320755, "grad_norm": 0.29172787070274353, "learning_rate": 1.885482180293501e-05, "loss": 0.738, "num_input_tokens_seen": 940160, "step": 1440 }, { "epoch": 0.7573375262054507, "grad_norm": 0.5099011659622192, "learning_rate": 1.892033542976939e-05, "loss": 1.0572, "num_input_tokens_seen": 942592, "step": 1445 }, { "epoch": 0.759958071278826, "grad_norm": 0.3030194044113159, "learning_rate": 1.8985849056603774e-05, "loss": 0.8284, "num_input_tokens_seen": 945856, "step": 1450 }, { "epoch": 0.7625786163522013, "grad_norm": 0.714146614074707, "learning_rate": 1.9051362683438157e-05, "loss": 0.7328, "num_input_tokens_seen": 948288, "step": 1455 }, { "epoch": 0.7651991614255765, "grad_norm": 0.664987325668335, "learning_rate": 1.9116876310272537e-05, "loss": 1.0493, "num_input_tokens_seen": 950720, "step": 1460 }, { "epoch": 0.7678197064989518, "grad_norm": 0.4322815239429474, "learning_rate": 1.918238993710692e-05, "loss": 0.7303, "num_input_tokens_seen": 953760, "step": 1465 }, { "epoch": 0.7704402515723271, "grad_norm": 0.7262795567512512, "learning_rate": 1.92479035639413e-05, "loss": 0.9635, "num_input_tokens_seen": 956864, "step": 1470 }, { "epoch": 0.7730607966457023, "grad_norm": 0.4004043638706207, "learning_rate": 1.9313417190775683e-05, "loss": 0.7111, "num_input_tokens_seen": 959584, "step": 1475 }, { "epoch": 0.7756813417190775, "grad_norm": 0.3396696448326111, "learning_rate": 1.9378930817610062e-05, "loss": 0.6404, "num_input_tokens_seen": 962720, "step": 1480 }, { "epoch": 0.7783018867924528, "grad_norm": 0.5989900231361389, "learning_rate": 1.9444444444444445e-05, "loss": 0.9015, "num_input_tokens_seen": 965568, "step": 1485 }, { "epoch": 0.7809224318658281, "grad_norm": 0.43028324842453003, "learning_rate": 1.950995807127883e-05, "loss": 0.8277, "num_input_tokens_seen": 969088, "step": 1490 }, { "epoch": 0.7835429769392034, "grad_norm": 0.36684831976890564, "learning_rate": 1.9575471698113208e-05, "loss": 0.729, "num_input_tokens_seen": 972960, "step": 1495 }, { "epoch": 0.7861635220125787, "grad_norm": 0.38986101746559143, "learning_rate": 1.9640985324947588e-05, "loss": 0.7884, "num_input_tokens_seen": 976256, "step": 1500 }, { "epoch": 0.7887840670859538, "grad_norm": 0.532055139541626, "learning_rate": 1.970649895178197e-05, "loss": 0.6954, "num_input_tokens_seen": 979584, "step": 1505 }, { "epoch": 0.7914046121593291, "grad_norm": 1.2182199954986572, "learning_rate": 1.9772012578616354e-05, "loss": 1.0161, "num_input_tokens_seen": 982208, "step": 1510 }, { "epoch": 0.7940251572327044, "grad_norm": 0.7234908938407898, "learning_rate": 1.9837526205450734e-05, "loss": 0.7893, "num_input_tokens_seen": 986624, "step": 1515 }, { "epoch": 0.7966457023060797, "grad_norm": 0.3492211103439331, "learning_rate": 1.9903039832285117e-05, "loss": 0.7887, "num_input_tokens_seen": 989856, "step": 1520 }, { "epoch": 0.799266247379455, "grad_norm": 0.5257292985916138, "learning_rate": 1.99685534591195e-05, "loss": 0.9212, "num_input_tokens_seen": 992384, "step": 1525 }, { "epoch": 0.8018867924528302, "grad_norm": 0.19761735200881958, "learning_rate": 2.003406708595388e-05, "loss": 0.6713, "num_input_tokens_seen": 996096, "step": 1530 }, { "epoch": 0.8045073375262054, "grad_norm": 0.684801459312439, "learning_rate": 2.009958071278826e-05, "loss": 0.8314, "num_input_tokens_seen": 998880, "step": 1535 }, { "epoch": 0.8071278825995807, "grad_norm": 0.3790067434310913, "learning_rate": 2.0165094339622642e-05, "loss": 0.9398, "num_input_tokens_seen": 1002336, "step": 1540 }, { "epoch": 0.809748427672956, "grad_norm": 0.3127588629722595, "learning_rate": 2.0230607966457025e-05, "loss": 0.748, "num_input_tokens_seen": 1005888, "step": 1545 }, { "epoch": 0.8123689727463312, "grad_norm": 0.1547544002532959, "learning_rate": 2.0296121593291405e-05, "loss": 0.5974, "num_input_tokens_seen": 1010496, "step": 1550 }, { "epoch": 0.8149895178197065, "grad_norm": 0.4011535346508026, "learning_rate": 2.0361635220125788e-05, "loss": 0.8424, "num_input_tokens_seen": 1014848, "step": 1555 }, { "epoch": 0.8176100628930818, "grad_norm": 0.39039596915245056, "learning_rate": 2.0427148846960168e-05, "loss": 0.6992, "num_input_tokens_seen": 1018176, "step": 1560 }, { "epoch": 0.820230607966457, "grad_norm": 0.4862308204174042, "learning_rate": 2.049266247379455e-05, "loss": 0.7848, "num_input_tokens_seen": 1021056, "step": 1565 }, { "epoch": 0.8228511530398323, "grad_norm": 0.6426750421524048, "learning_rate": 2.055817610062893e-05, "loss": 0.8247, "num_input_tokens_seen": 1024096, "step": 1570 }, { "epoch": 0.8254716981132075, "grad_norm": 0.5668312311172485, "learning_rate": 2.0623689727463314e-05, "loss": 0.934, "num_input_tokens_seen": 1027072, "step": 1575 }, { "epoch": 0.8280922431865828, "grad_norm": 0.4457755982875824, "learning_rate": 2.0689203354297697e-05, "loss": 0.8288, "num_input_tokens_seen": 1029536, "step": 1580 }, { "epoch": 0.8307127882599581, "grad_norm": 0.691275954246521, "learning_rate": 2.0754716981132076e-05, "loss": 0.7418, "num_input_tokens_seen": 1032320, "step": 1585 }, { "epoch": 0.8333333333333334, "grad_norm": 0.35151493549346924, "learning_rate": 2.0820230607966456e-05, "loss": 0.9496, "num_input_tokens_seen": 1035168, "step": 1590 }, { "epoch": 0.8359538784067087, "grad_norm": 0.3204285800457001, "learning_rate": 2.088574423480084e-05, "loss": 0.766, "num_input_tokens_seen": 1038720, "step": 1595 }, { "epoch": 0.8385744234800838, "grad_norm": 0.7847768068313599, "learning_rate": 2.0951257861635222e-05, "loss": 0.8134, "num_input_tokens_seen": 1041344, "step": 1600 }, { "epoch": 0.8411949685534591, "grad_norm": 0.834126889705658, "learning_rate": 2.1016771488469602e-05, "loss": 1.2451, "num_input_tokens_seen": 1043808, "step": 1605 }, { "epoch": 0.8438155136268344, "grad_norm": 0.4900798797607422, "learning_rate": 2.1082285115303985e-05, "loss": 0.7236, "num_input_tokens_seen": 1046752, "step": 1610 }, { "epoch": 0.8464360587002097, "grad_norm": 0.48539531230926514, "learning_rate": 2.1147798742138368e-05, "loss": 0.8098, "num_input_tokens_seen": 1049664, "step": 1615 }, { "epoch": 0.8490566037735849, "grad_norm": 0.4462023973464966, "learning_rate": 2.1213312368972748e-05, "loss": 0.8091, "num_input_tokens_seen": 1052128, "step": 1620 }, { "epoch": 0.8516771488469602, "grad_norm": 0.5551913380622864, "learning_rate": 2.1278825995807127e-05, "loss": 0.758, "num_input_tokens_seen": 1055008, "step": 1625 }, { "epoch": 0.8542976939203354, "grad_norm": 0.3642597794532776, "learning_rate": 2.134433962264151e-05, "loss": 0.831, "num_input_tokens_seen": 1058368, "step": 1630 }, { "epoch": 0.8569182389937107, "grad_norm": 0.37824809551239014, "learning_rate": 2.1409853249475894e-05, "loss": 0.7044, "num_input_tokens_seen": 1061888, "step": 1635 }, { "epoch": 0.859538784067086, "grad_norm": 0.8155788779258728, "learning_rate": 2.1475366876310273e-05, "loss": 0.6927, "num_input_tokens_seen": 1065344, "step": 1640 }, { "epoch": 0.8621593291404612, "grad_norm": 0.3406650424003601, "learning_rate": 2.1540880503144656e-05, "loss": 0.7076, "num_input_tokens_seen": 1068288, "step": 1645 }, { "epoch": 0.8647798742138365, "grad_norm": 0.23601289093494415, "learning_rate": 2.160639412997904e-05, "loss": 0.5829, "num_input_tokens_seen": 1072512, "step": 1650 }, { "epoch": 0.8674004192872118, "grad_norm": 0.31351742148399353, "learning_rate": 2.1671907756813416e-05, "loss": 0.6406, "num_input_tokens_seen": 1077280, "step": 1655 }, { "epoch": 0.870020964360587, "grad_norm": 0.466521292924881, "learning_rate": 2.17374213836478e-05, "loss": 0.738, "num_input_tokens_seen": 1081632, "step": 1660 }, { "epoch": 0.8726415094339622, "grad_norm": 0.5902673006057739, "learning_rate": 2.1802935010482182e-05, "loss": 0.5911, "num_input_tokens_seen": 1084768, "step": 1665 }, { "epoch": 0.8752620545073375, "grad_norm": 0.33705171942710876, "learning_rate": 2.186844863731656e-05, "loss": 0.7908, "num_input_tokens_seen": 1088864, "step": 1670 }, { "epoch": 0.8778825995807128, "grad_norm": 0.36558791995048523, "learning_rate": 2.1933962264150945e-05, "loss": 0.6608, "num_input_tokens_seen": 1091840, "step": 1675 }, { "epoch": 0.8805031446540881, "grad_norm": 0.46693360805511475, "learning_rate": 2.1999475890985328e-05, "loss": 0.7088, "num_input_tokens_seen": 1095136, "step": 1680 }, { "epoch": 0.8831236897274634, "grad_norm": 0.44779929518699646, "learning_rate": 2.2064989517819707e-05, "loss": 0.8189, "num_input_tokens_seen": 1098848, "step": 1685 }, { "epoch": 0.8857442348008385, "grad_norm": 0.2979402542114258, "learning_rate": 2.2130503144654087e-05, "loss": 0.7426, "num_input_tokens_seen": 1101920, "step": 1690 }, { "epoch": 0.8883647798742138, "grad_norm": 0.5179060697555542, "learning_rate": 2.219601677148847e-05, "loss": 0.7227, "num_input_tokens_seen": 1105152, "step": 1695 }, { "epoch": 0.8909853249475891, "grad_norm": 0.4442940652370453, "learning_rate": 2.2261530398322853e-05, "loss": 0.6254, "num_input_tokens_seen": 1111328, "step": 1700 }, { "epoch": 0.8936058700209644, "grad_norm": 1.0133228302001953, "learning_rate": 2.2327044025157233e-05, "loss": 0.7912, "num_input_tokens_seen": 1114240, "step": 1705 }, { "epoch": 0.8962264150943396, "grad_norm": 0.929100751876831, "learning_rate": 2.2392557651991616e-05, "loss": 0.794, "num_input_tokens_seen": 1118912, "step": 1710 }, { "epoch": 0.8988469601677149, "grad_norm": 0.309295654296875, "learning_rate": 2.2458071278825996e-05, "loss": 0.9025, "num_input_tokens_seen": 1122368, "step": 1715 }, { "epoch": 0.9014675052410901, "grad_norm": 0.3463698923587799, "learning_rate": 2.252358490566038e-05, "loss": 0.6468, "num_input_tokens_seen": 1125184, "step": 1720 }, { "epoch": 0.9040880503144654, "grad_norm": 1.1549592018127441, "learning_rate": 2.258909853249476e-05, "loss": 0.8717, "num_input_tokens_seen": 1128992, "step": 1725 }, { "epoch": 0.9067085953878407, "grad_norm": 0.24084386229515076, "learning_rate": 2.265461215932914e-05, "loss": 0.8756, "num_input_tokens_seen": 1131872, "step": 1730 }, { "epoch": 0.9093291404612159, "grad_norm": 0.31425735354423523, "learning_rate": 2.2720125786163524e-05, "loss": 0.695, "num_input_tokens_seen": 1135904, "step": 1735 }, { "epoch": 0.9119496855345912, "grad_norm": 0.8004541993141174, "learning_rate": 2.2785639412997904e-05, "loss": 0.7277, "num_input_tokens_seen": 1139520, "step": 1740 }, { "epoch": 0.9145702306079665, "grad_norm": 0.6635422110557556, "learning_rate": 2.2851153039832284e-05, "loss": 0.7486, "num_input_tokens_seen": 1142304, "step": 1745 }, { "epoch": 0.9171907756813418, "grad_norm": 0.6221281886100769, "learning_rate": 2.2916666666666667e-05, "loss": 0.6912, "num_input_tokens_seen": 1145280, "step": 1750 }, { "epoch": 0.9198113207547169, "grad_norm": 0.545012891292572, "learning_rate": 2.298218029350105e-05, "loss": 0.7626, "num_input_tokens_seen": 1147904, "step": 1755 }, { "epoch": 0.9224318658280922, "grad_norm": 0.8043617606163025, "learning_rate": 2.304769392033543e-05, "loss": 0.9046, "num_input_tokens_seen": 1151200, "step": 1760 }, { "epoch": 0.9250524109014675, "grad_norm": 0.4599638283252716, "learning_rate": 2.3113207547169813e-05, "loss": 0.8729, "num_input_tokens_seen": 1154528, "step": 1765 }, { "epoch": 0.9276729559748428, "grad_norm": 0.18969237804412842, "learning_rate": 2.3178721174004196e-05, "loss": 0.7641, "num_input_tokens_seen": 1157632, "step": 1770 }, { "epoch": 0.9302935010482181, "grad_norm": 0.3828127682209015, "learning_rate": 2.3244234800838576e-05, "loss": 0.5901, "num_input_tokens_seen": 1161280, "step": 1775 }, { "epoch": 0.9329140461215933, "grad_norm": 0.373897910118103, "learning_rate": 2.3309748427672955e-05, "loss": 0.7001, "num_input_tokens_seen": 1164512, "step": 1780 }, { "epoch": 0.9355345911949685, "grad_norm": 0.30474942922592163, "learning_rate": 2.3375262054507338e-05, "loss": 0.6885, "num_input_tokens_seen": 1167808, "step": 1785 }, { "epoch": 0.9381551362683438, "grad_norm": 0.43462759256362915, "learning_rate": 2.344077568134172e-05, "loss": 0.6951, "num_input_tokens_seen": 1170432, "step": 1790 }, { "epoch": 0.9407756813417191, "grad_norm": 0.27664604783058167, "learning_rate": 2.35062893081761e-05, "loss": 0.5621, "num_input_tokens_seen": 1173088, "step": 1795 }, { "epoch": 0.9433962264150944, "grad_norm": 0.3423472046852112, "learning_rate": 2.3571802935010484e-05, "loss": 0.6362, "num_input_tokens_seen": 1176448, "step": 1800 }, { "epoch": 0.9460167714884696, "grad_norm": 0.2998783588409424, "learning_rate": 2.3637316561844867e-05, "loss": 0.7188, "num_input_tokens_seen": 1179296, "step": 1805 }, { "epoch": 0.9486373165618449, "grad_norm": 0.21704374253749847, "learning_rate": 2.3702830188679247e-05, "loss": 0.7471, "num_input_tokens_seen": 1186976, "step": 1810 }, { "epoch": 0.9512578616352201, "grad_norm": 0.4522440731525421, "learning_rate": 2.3768343815513627e-05, "loss": 0.8194, "num_input_tokens_seen": 1189248, "step": 1815 }, { "epoch": 0.9538784067085954, "grad_norm": 0.3389008343219757, "learning_rate": 2.383385744234801e-05, "loss": 0.8359, "num_input_tokens_seen": 1191840, "step": 1820 }, { "epoch": 0.9564989517819706, "grad_norm": 0.4361843466758728, "learning_rate": 2.3899371069182393e-05, "loss": 0.7496, "num_input_tokens_seen": 1194720, "step": 1825 }, { "epoch": 0.9591194968553459, "grad_norm": 0.30437028408050537, "learning_rate": 2.3964884696016772e-05, "loss": 0.665, "num_input_tokens_seen": 1198592, "step": 1830 }, { "epoch": 0.9617400419287212, "grad_norm": 0.34752097725868225, "learning_rate": 2.4030398322851155e-05, "loss": 0.9559, "num_input_tokens_seen": 1201216, "step": 1835 }, { "epoch": 0.9643605870020965, "grad_norm": 0.3937646746635437, "learning_rate": 2.4095911949685535e-05, "loss": 0.7153, "num_input_tokens_seen": 1204096, "step": 1840 }, { "epoch": 0.9669811320754716, "grad_norm": 0.3298949599266052, "learning_rate": 2.4161425576519918e-05, "loss": 0.5717, "num_input_tokens_seen": 1206784, "step": 1845 }, { "epoch": 0.9696016771488469, "grad_norm": 0.7534866333007812, "learning_rate": 2.4226939203354298e-05, "loss": 0.6683, "num_input_tokens_seen": 1209632, "step": 1850 }, { "epoch": 0.9722222222222222, "grad_norm": 0.3640762269496918, "learning_rate": 2.429245283018868e-05, "loss": 0.7436, "num_input_tokens_seen": 1212512, "step": 1855 }, { "epoch": 0.9748427672955975, "grad_norm": 0.2949434816837311, "learning_rate": 2.4357966457023064e-05, "loss": 0.7502, "num_input_tokens_seen": 1216096, "step": 1860 }, { "epoch": 0.9774633123689728, "grad_norm": 0.2670477330684662, "learning_rate": 2.4423480083857444e-05, "loss": 0.683, "num_input_tokens_seen": 1219264, "step": 1865 }, { "epoch": 0.980083857442348, "grad_norm": 0.35899728536605835, "learning_rate": 2.4488993710691823e-05, "loss": 0.6064, "num_input_tokens_seen": 1222080, "step": 1870 }, { "epoch": 0.9827044025157232, "grad_norm": 0.43462249636650085, "learning_rate": 2.4554507337526206e-05, "loss": 0.7838, "num_input_tokens_seen": 1224288, "step": 1875 }, { "epoch": 0.9853249475890985, "grad_norm": 0.2671926021575928, "learning_rate": 2.4620020964360586e-05, "loss": 0.776, "num_input_tokens_seen": 1227168, "step": 1880 }, { "epoch": 0.9879454926624738, "grad_norm": 0.5238470435142517, "learning_rate": 2.468553459119497e-05, "loss": 0.6254, "num_input_tokens_seen": 1230656, "step": 1885 }, { "epoch": 0.9905660377358491, "grad_norm": 0.2777676582336426, "learning_rate": 2.4751048218029352e-05, "loss": 0.8385, "num_input_tokens_seen": 1233312, "step": 1890 }, { "epoch": 0.9931865828092243, "grad_norm": 1.028541088104248, "learning_rate": 2.4816561844863735e-05, "loss": 0.6871, "num_input_tokens_seen": 1239040, "step": 1895 }, { "epoch": 0.9958071278825996, "grad_norm": 0.159436896443367, "learning_rate": 2.488207547169811e-05, "loss": 0.6769, "num_input_tokens_seen": 1242720, "step": 1900 }, { "epoch": 0.9984276729559748, "grad_norm": 0.29392337799072266, "learning_rate": 2.4947589098532495e-05, "loss": 0.7179, "num_input_tokens_seen": 1245664, "step": 1905 }, { "epoch": 1.0, "eval_loss": 0.6613768339157104, "eval_runtime": 9.8968, "eval_samples_per_second": 85.685, "eval_steps_per_second": 21.421, "num_input_tokens_seen": 1246832, "step": 1908 }, { "epoch": 1.00104821802935, "grad_norm": 0.6122419238090515, "learning_rate": 2.5013102725366878e-05, "loss": 1.1753, "num_input_tokens_seen": 1248016, "step": 1910 }, { "epoch": 1.0036687631027255, "grad_norm": 0.47024866938591003, "learning_rate": 2.507861635220126e-05, "loss": 0.5791, "num_input_tokens_seen": 1251600, "step": 1915 }, { "epoch": 1.0062893081761006, "grad_norm": 0.25523918867111206, "learning_rate": 2.514412997903564e-05, "loss": 0.8244, "num_input_tokens_seen": 1254640, "step": 1920 }, { "epoch": 1.0089098532494758, "grad_norm": 0.1756059229373932, "learning_rate": 2.520964360587002e-05, "loss": 0.569, "num_input_tokens_seen": 1258576, "step": 1925 }, { "epoch": 1.0115303983228512, "grad_norm": 0.43867772817611694, "learning_rate": 2.5275157232704407e-05, "loss": 0.7575, "num_input_tokens_seen": 1261488, "step": 1930 }, { "epoch": 1.0141509433962264, "grad_norm": 0.37392210960388184, "learning_rate": 2.5340670859538783e-05, "loss": 0.7375, "num_input_tokens_seen": 1264688, "step": 1935 }, { "epoch": 1.0167714884696017, "grad_norm": 0.37488695979118347, "learning_rate": 2.540618448637317e-05, "loss": 0.6312, "num_input_tokens_seen": 1267216, "step": 1940 }, { "epoch": 1.019392033542977, "grad_norm": 0.4318466782569885, "learning_rate": 2.547169811320755e-05, "loss": 0.788, "num_input_tokens_seen": 1270000, "step": 1945 }, { "epoch": 1.0220125786163523, "grad_norm": 0.34776419401168823, "learning_rate": 2.553721174004193e-05, "loss": 0.6742, "num_input_tokens_seen": 1275920, "step": 1950 }, { "epoch": 1.0246331236897275, "grad_norm": 0.19965307414531708, "learning_rate": 2.5602725366876312e-05, "loss": 0.4329, "num_input_tokens_seen": 1279280, "step": 1955 }, { "epoch": 1.0272536687631026, "grad_norm": 0.3467235565185547, "learning_rate": 2.566823899371069e-05, "loss": 0.8417, "num_input_tokens_seen": 1282608, "step": 1960 }, { "epoch": 1.029874213836478, "grad_norm": 0.6614015698432922, "learning_rate": 2.5733752620545075e-05, "loss": 0.6246, "num_input_tokens_seen": 1285136, "step": 1965 }, { "epoch": 1.0324947589098532, "grad_norm": 0.291385293006897, "learning_rate": 2.5799266247379454e-05, "loss": 0.8423, "num_input_tokens_seen": 1288336, "step": 1970 }, { "epoch": 1.0351153039832286, "grad_norm": 0.4053424894809723, "learning_rate": 2.586477987421384e-05, "loss": 0.6302, "num_input_tokens_seen": 1290928, "step": 1975 }, { "epoch": 1.0377358490566038, "grad_norm": 0.6025903820991516, "learning_rate": 2.593029350104822e-05, "loss": 0.6541, "num_input_tokens_seen": 1294960, "step": 1980 }, { "epoch": 1.040356394129979, "grad_norm": 0.29726001620292664, "learning_rate": 2.59958071278826e-05, "loss": 0.7155, "num_input_tokens_seen": 1298960, "step": 1985 }, { "epoch": 1.0429769392033543, "grad_norm": 0.46489307284355164, "learning_rate": 2.6061320754716983e-05, "loss": 0.6942, "num_input_tokens_seen": 1302288, "step": 1990 }, { "epoch": 1.0455974842767295, "grad_norm": 0.2609984278678894, "learning_rate": 2.6126834381551363e-05, "loss": 0.6082, "num_input_tokens_seen": 1305872, "step": 1995 }, { "epoch": 1.0482180293501049, "grad_norm": 0.3798237144947052, "learning_rate": 2.6192348008385746e-05, "loss": 0.5637, "num_input_tokens_seen": 1309808, "step": 2000 }, { "epoch": 1.05083857442348, "grad_norm": 0.38269245624542236, "learning_rate": 2.6257861635220126e-05, "loss": 0.7043, "num_input_tokens_seen": 1312400, "step": 2005 }, { "epoch": 1.0534591194968554, "grad_norm": 0.3544243276119232, "learning_rate": 2.6323375262054505e-05, "loss": 0.6971, "num_input_tokens_seen": 1314896, "step": 2010 }, { "epoch": 1.0560796645702306, "grad_norm": 0.28190937638282776, "learning_rate": 2.6388888888888892e-05, "loss": 0.6664, "num_input_tokens_seen": 1317392, "step": 2015 }, { "epoch": 1.0587002096436058, "grad_norm": 0.38849908113479614, "learning_rate": 2.645440251572327e-05, "loss": 0.8714, "num_input_tokens_seen": 1319920, "step": 2020 }, { "epoch": 1.0613207547169812, "grad_norm": 0.23536847531795502, "learning_rate": 2.6519916142557655e-05, "loss": 0.6758, "num_input_tokens_seen": 1323408, "step": 2025 }, { "epoch": 1.0639412997903563, "grad_norm": 0.4768660068511963, "learning_rate": 2.6585429769392034e-05, "loss": 0.7889, "num_input_tokens_seen": 1326192, "step": 2030 }, { "epoch": 1.0665618448637317, "grad_norm": 0.21417176723480225, "learning_rate": 2.6650943396226417e-05, "loss": 0.6247, "num_input_tokens_seen": 1329648, "step": 2035 }, { "epoch": 1.069182389937107, "grad_norm": 0.2657839357852936, "learning_rate": 2.6716457023060797e-05, "loss": 0.769, "num_input_tokens_seen": 1332720, "step": 2040 }, { "epoch": 1.0718029350104823, "grad_norm": 0.37625008821487427, "learning_rate": 2.6781970649895177e-05, "loss": 0.5256, "num_input_tokens_seen": 1338288, "step": 2045 }, { "epoch": 1.0744234800838575, "grad_norm": 0.3238316476345062, "learning_rate": 2.6847484276729563e-05, "loss": 0.7645, "num_input_tokens_seen": 1341040, "step": 2050 }, { "epoch": 1.0770440251572326, "grad_norm": 0.5525323748588562, "learning_rate": 2.6912997903563943e-05, "loss": 0.7426, "num_input_tokens_seen": 1343920, "step": 2055 }, { "epoch": 1.079664570230608, "grad_norm": 0.39935553073883057, "learning_rate": 2.6978511530398326e-05, "loss": 0.6006, "num_input_tokens_seen": 1346832, "step": 2060 }, { "epoch": 1.0822851153039832, "grad_norm": 0.3150446116924286, "learning_rate": 2.7044025157232706e-05, "loss": 0.5697, "num_input_tokens_seen": 1350160, "step": 2065 }, { "epoch": 1.0849056603773586, "grad_norm": 0.19441048800945282, "learning_rate": 2.710953878406709e-05, "loss": 0.6878, "num_input_tokens_seen": 1354128, "step": 2070 }, { "epoch": 1.0875262054507338, "grad_norm": 0.25787535309791565, "learning_rate": 2.717505241090147e-05, "loss": 0.6064, "num_input_tokens_seen": 1357840, "step": 2075 }, { "epoch": 1.090146750524109, "grad_norm": 0.34337562322616577, "learning_rate": 2.7240566037735848e-05, "loss": 0.7988, "num_input_tokens_seen": 1360560, "step": 2080 }, { "epoch": 1.0927672955974843, "grad_norm": 0.21316586434841156, "learning_rate": 2.7306079664570235e-05, "loss": 0.5535, "num_input_tokens_seen": 1364176, "step": 2085 }, { "epoch": 1.0953878406708595, "grad_norm": 0.30268073081970215, "learning_rate": 2.737159329140461e-05, "loss": 0.7221, "num_input_tokens_seen": 1367120, "step": 2090 }, { "epoch": 1.0980083857442349, "grad_norm": 0.28144538402557373, "learning_rate": 2.7437106918238997e-05, "loss": 0.8736, "num_input_tokens_seen": 1370800, "step": 2095 }, { "epoch": 1.10062893081761, "grad_norm": 0.2091626226902008, "learning_rate": 2.7502620545073377e-05, "loss": 0.5809, "num_input_tokens_seen": 1374640, "step": 2100 }, { "epoch": 1.1032494758909852, "grad_norm": 0.19347092509269714, "learning_rate": 2.7568134171907757e-05, "loss": 0.6106, "num_input_tokens_seen": 1378000, "step": 2105 }, { "epoch": 1.1058700209643606, "grad_norm": 0.19551637768745422, "learning_rate": 2.763364779874214e-05, "loss": 0.6455, "num_input_tokens_seen": 1381328, "step": 2110 }, { "epoch": 1.1084905660377358, "grad_norm": 0.3988804519176483, "learning_rate": 2.769916142557652e-05, "loss": 0.679, "num_input_tokens_seen": 1383600, "step": 2115 }, { "epoch": 1.1111111111111112, "grad_norm": 0.23758788406848907, "learning_rate": 2.7764675052410906e-05, "loss": 0.6398, "num_input_tokens_seen": 1387024, "step": 2120 }, { "epoch": 1.1137316561844863, "grad_norm": 0.17966684699058533, "learning_rate": 2.7830188679245282e-05, "loss": 0.5495, "num_input_tokens_seen": 1390352, "step": 2125 }, { "epoch": 1.1163522012578617, "grad_norm": 0.32790687680244446, "learning_rate": 2.789570230607967e-05, "loss": 0.6156, "num_input_tokens_seen": 1393328, "step": 2130 }, { "epoch": 1.118972746331237, "grad_norm": 0.19655565917491913, "learning_rate": 2.796121593291405e-05, "loss": 0.5756, "num_input_tokens_seen": 1396688, "step": 2135 }, { "epoch": 1.121593291404612, "grad_norm": 0.2972959876060486, "learning_rate": 2.8026729559748428e-05, "loss": 0.7605, "num_input_tokens_seen": 1400400, "step": 2140 }, { "epoch": 1.1242138364779874, "grad_norm": 0.2685047388076782, "learning_rate": 2.809224318658281e-05, "loss": 0.5697, "num_input_tokens_seen": 1404144, "step": 2145 }, { "epoch": 1.1268343815513626, "grad_norm": 0.5370845794677734, "learning_rate": 2.815775681341719e-05, "loss": 0.731, "num_input_tokens_seen": 1406960, "step": 2150 }, { "epoch": 1.129454926624738, "grad_norm": 0.3426693379878998, "learning_rate": 2.8223270440251577e-05, "loss": 0.5676, "num_input_tokens_seen": 1409840, "step": 2155 }, { "epoch": 1.1320754716981132, "grad_norm": 0.24801179766654968, "learning_rate": 2.8288784067085954e-05, "loss": 0.7884, "num_input_tokens_seen": 1412944, "step": 2160 }, { "epoch": 1.1346960167714886, "grad_norm": 0.26438549160957336, "learning_rate": 2.8354297693920333e-05, "loss": 0.5373, "num_input_tokens_seen": 1416272, "step": 2165 }, { "epoch": 1.1373165618448637, "grad_norm": 0.20893613994121552, "learning_rate": 2.841981132075472e-05, "loss": 0.8679, "num_input_tokens_seen": 1419248, "step": 2170 }, { "epoch": 1.139937106918239, "grad_norm": 0.21153946220874786, "learning_rate": 2.84853249475891e-05, "loss": 0.6393, "num_input_tokens_seen": 1422352, "step": 2175 }, { "epoch": 1.1425576519916143, "grad_norm": 0.1401282250881195, "learning_rate": 2.8550838574423482e-05, "loss": 0.5208, "num_input_tokens_seen": 1425808, "step": 2180 }, { "epoch": 1.1451781970649895, "grad_norm": 0.2824084758758545, "learning_rate": 2.8616352201257862e-05, "loss": 0.8309, "num_input_tokens_seen": 1429392, "step": 2185 }, { "epoch": 1.1477987421383649, "grad_norm": 0.5169875025749207, "learning_rate": 2.868186582809225e-05, "loss": 0.6613, "num_input_tokens_seen": 1432208, "step": 2190 }, { "epoch": 1.15041928721174, "grad_norm": 0.6439647674560547, "learning_rate": 2.8747379454926625e-05, "loss": 0.5621, "num_input_tokens_seen": 1434928, "step": 2195 }, { "epoch": 1.1530398322851152, "grad_norm": 0.30019670724868774, "learning_rate": 2.8812893081761005e-05, "loss": 0.5787, "num_input_tokens_seen": 1438192, "step": 2200 }, { "epoch": 1.1556603773584906, "grad_norm": 0.2540460526943207, "learning_rate": 2.887840670859539e-05, "loss": 0.6387, "num_input_tokens_seen": 1441136, "step": 2205 }, { "epoch": 1.1582809224318658, "grad_norm": 0.2732018828392029, "learning_rate": 2.894392033542977e-05, "loss": 0.709, "num_input_tokens_seen": 1444528, "step": 2210 }, { "epoch": 1.1609014675052411, "grad_norm": 0.173203706741333, "learning_rate": 2.9009433962264154e-05, "loss": 0.8773, "num_input_tokens_seen": 1448240, "step": 2215 }, { "epoch": 1.1635220125786163, "grad_norm": 0.2936672866344452, "learning_rate": 2.9074947589098533e-05, "loss": 0.9397, "num_input_tokens_seen": 1451792, "step": 2220 }, { "epoch": 1.1661425576519917, "grad_norm": 0.18681783974170685, "learning_rate": 2.9140461215932913e-05, "loss": 0.6559, "num_input_tokens_seen": 1453904, "step": 2225 }, { "epoch": 1.1687631027253669, "grad_norm": 0.35677191615104675, "learning_rate": 2.9205974842767296e-05, "loss": 0.6208, "num_input_tokens_seen": 1457200, "step": 2230 }, { "epoch": 1.171383647798742, "grad_norm": 0.23783445358276367, "learning_rate": 2.9271488469601676e-05, "loss": 0.5424, "num_input_tokens_seen": 1461488, "step": 2235 }, { "epoch": 1.1740041928721174, "grad_norm": 0.3113991320133209, "learning_rate": 2.9337002096436062e-05, "loss": 0.5597, "num_input_tokens_seen": 1464080, "step": 2240 }, { "epoch": 1.1766247379454926, "grad_norm": 0.28469130396842957, "learning_rate": 2.9402515723270442e-05, "loss": 0.6859, "num_input_tokens_seen": 1466832, "step": 2245 }, { "epoch": 1.179245283018868, "grad_norm": 0.5367029905319214, "learning_rate": 2.9468029350104825e-05, "loss": 0.6028, "num_input_tokens_seen": 1469520, "step": 2250 }, { "epoch": 1.1818658280922432, "grad_norm": 0.2141764909029007, "learning_rate": 2.9533542976939205e-05, "loss": 0.7119, "num_input_tokens_seen": 1476336, "step": 2255 }, { "epoch": 1.1844863731656186, "grad_norm": 0.15714547038078308, "learning_rate": 2.9599056603773584e-05, "loss": 0.6099, "num_input_tokens_seen": 1479984, "step": 2260 }, { "epoch": 1.1871069182389937, "grad_norm": 0.1907777339220047, "learning_rate": 2.9664570230607968e-05, "loss": 0.6104, "num_input_tokens_seen": 1483056, "step": 2265 }, { "epoch": 1.189727463312369, "grad_norm": 0.688449501991272, "learning_rate": 2.9730083857442347e-05, "loss": 0.7511, "num_input_tokens_seen": 1485936, "step": 2270 }, { "epoch": 1.1923480083857443, "grad_norm": 0.19415561854839325, "learning_rate": 2.9795597484276734e-05, "loss": 0.6652, "num_input_tokens_seen": 1488688, "step": 2275 }, { "epoch": 1.1949685534591195, "grad_norm": 0.18324783444404602, "learning_rate": 2.9861111111111113e-05, "loss": 0.6922, "num_input_tokens_seen": 1492272, "step": 2280 }, { "epoch": 1.1975890985324948, "grad_norm": 0.22802279889583588, "learning_rate": 2.9926624737945496e-05, "loss": 0.459, "num_input_tokens_seen": 1495504, "step": 2285 }, { "epoch": 1.20020964360587, "grad_norm": 0.2002985179424286, "learning_rate": 2.9992138364779876e-05, "loss": 0.692, "num_input_tokens_seen": 1498736, "step": 2290 }, { "epoch": 1.2028301886792452, "grad_norm": 0.18018528819084167, "learning_rate": 3.0057651991614256e-05, "loss": 0.6063, "num_input_tokens_seen": 1501616, "step": 2295 }, { "epoch": 1.2054507337526206, "grad_norm": 0.18194885551929474, "learning_rate": 3.012316561844864e-05, "loss": 0.7636, "num_input_tokens_seen": 1504560, "step": 2300 }, { "epoch": 1.2080712788259957, "grad_norm": 0.3560081124305725, "learning_rate": 3.018867924528302e-05, "loss": 0.6401, "num_input_tokens_seen": 1508080, "step": 2305 }, { "epoch": 1.2106918238993711, "grad_norm": 0.1869676560163498, "learning_rate": 3.0254192872117405e-05, "loss": 0.5126, "num_input_tokens_seen": 1511824, "step": 2310 }, { "epoch": 1.2133123689727463, "grad_norm": 0.28423959016799927, "learning_rate": 3.0319706498951785e-05, "loss": 0.4747, "num_input_tokens_seen": 1515088, "step": 2315 }, { "epoch": 1.2159329140461215, "grad_norm": 0.28452739119529724, "learning_rate": 3.038522012578616e-05, "loss": 0.6183, "num_input_tokens_seen": 1517744, "step": 2320 }, { "epoch": 1.2185534591194969, "grad_norm": 0.6607798337936401, "learning_rate": 3.0450733752620547e-05, "loss": 0.6618, "num_input_tokens_seen": 1521808, "step": 2325 }, { "epoch": 1.221174004192872, "grad_norm": 0.15628759562969208, "learning_rate": 3.0516247379454927e-05, "loss": 0.605, "num_input_tokens_seen": 1525296, "step": 2330 }, { "epoch": 1.2237945492662474, "grad_norm": 0.4290388226509094, "learning_rate": 3.0581761006289314e-05, "loss": 0.7281, "num_input_tokens_seen": 1527856, "step": 2335 }, { "epoch": 1.2264150943396226, "grad_norm": 0.31068310141563416, "learning_rate": 3.064727463312369e-05, "loss": 0.6273, "num_input_tokens_seen": 1530384, "step": 2340 }, { "epoch": 1.229035639412998, "grad_norm": 0.16550907492637634, "learning_rate": 3.071278825995807e-05, "loss": 0.3915, "num_input_tokens_seen": 1534224, "step": 2345 }, { "epoch": 1.2316561844863732, "grad_norm": 0.25000837445259094, "learning_rate": 3.0778301886792456e-05, "loss": 0.5876, "num_input_tokens_seen": 1536656, "step": 2350 }, { "epoch": 1.2342767295597485, "grad_norm": 0.19482456147670746, "learning_rate": 3.084381551362683e-05, "loss": 0.6171, "num_input_tokens_seen": 1541168, "step": 2355 }, { "epoch": 1.2368972746331237, "grad_norm": 0.40615540742874146, "learning_rate": 3.0909329140461215e-05, "loss": 0.6725, "num_input_tokens_seen": 1543632, "step": 2360 }, { "epoch": 1.2395178197064989, "grad_norm": 0.2545541226863861, "learning_rate": 3.09748427672956e-05, "loss": 0.6145, "num_input_tokens_seen": 1548816, "step": 2365 }, { "epoch": 1.2421383647798743, "grad_norm": 0.3879792094230652, "learning_rate": 3.104035639412998e-05, "loss": 0.6112, "num_input_tokens_seen": 1552368, "step": 2370 }, { "epoch": 1.2447589098532494, "grad_norm": 0.5250260829925537, "learning_rate": 3.1105870020964365e-05, "loss": 0.7271, "num_input_tokens_seen": 1555504, "step": 2375 }, { "epoch": 1.2473794549266248, "grad_norm": 0.4029460549354553, "learning_rate": 3.117138364779874e-05, "loss": 0.6361, "num_input_tokens_seen": 1558096, "step": 2380 }, { "epoch": 1.25, "grad_norm": 0.8854926824569702, "learning_rate": 3.1236897274633124e-05, "loss": 0.8037, "num_input_tokens_seen": 1561008, "step": 2385 }, { "epoch": 1.2526205450733752, "grad_norm": 0.20423895120620728, "learning_rate": 3.130241090146751e-05, "loss": 0.7051, "num_input_tokens_seen": 1564176, "step": 2390 }, { "epoch": 1.2552410901467506, "grad_norm": 0.22609776258468628, "learning_rate": 3.136792452830189e-05, "loss": 0.5263, "num_input_tokens_seen": 1566992, "step": 2395 }, { "epoch": 1.2578616352201257, "grad_norm": 0.18181540071964264, "learning_rate": 3.1433438155136266e-05, "loss": 0.5647, "num_input_tokens_seen": 1570352, "step": 2400 }, { "epoch": 1.2604821802935011, "grad_norm": 0.1922295242547989, "learning_rate": 3.1498951781970656e-05, "loss": 0.5936, "num_input_tokens_seen": 1573040, "step": 2405 }, { "epoch": 1.2631027253668763, "grad_norm": 0.3215290606021881, "learning_rate": 3.156446540880503e-05, "loss": 0.5894, "num_input_tokens_seen": 1576944, "step": 2410 }, { "epoch": 1.2657232704402515, "grad_norm": 0.29304492473602295, "learning_rate": 3.162997903563941e-05, "loss": 0.5961, "num_input_tokens_seen": 1580080, "step": 2415 }, { "epoch": 1.2683438155136268, "grad_norm": 0.3777753710746765, "learning_rate": 3.16954926624738e-05, "loss": 0.5568, "num_input_tokens_seen": 1582512, "step": 2420 }, { "epoch": 1.270964360587002, "grad_norm": 0.3450731039047241, "learning_rate": 3.1761006289308175e-05, "loss": 0.6211, "num_input_tokens_seen": 1586512, "step": 2425 }, { "epoch": 1.2735849056603774, "grad_norm": 0.1598760187625885, "learning_rate": 3.182651991614256e-05, "loss": 0.6713, "num_input_tokens_seen": 1589296, "step": 2430 }, { "epoch": 1.2762054507337526, "grad_norm": 0.5808752179145813, "learning_rate": 3.189203354297694e-05, "loss": 0.6776, "num_input_tokens_seen": 1592656, "step": 2435 }, { "epoch": 1.2788259958071277, "grad_norm": 0.2132626473903656, "learning_rate": 3.1957547169811324e-05, "loss": 0.699, "num_input_tokens_seen": 1596368, "step": 2440 }, { "epoch": 1.2814465408805031, "grad_norm": 0.22274774312973022, "learning_rate": 3.202306079664571e-05, "loss": 0.6016, "num_input_tokens_seen": 1598832, "step": 2445 }, { "epoch": 1.2840670859538785, "grad_norm": 0.25016283988952637, "learning_rate": 3.2088574423480084e-05, "loss": 0.5581, "num_input_tokens_seen": 1602992, "step": 2450 }, { "epoch": 1.2866876310272537, "grad_norm": 0.26781418919563293, "learning_rate": 3.215408805031447e-05, "loss": 0.6392, "num_input_tokens_seen": 1607344, "step": 2455 }, { "epoch": 1.2893081761006289, "grad_norm": 0.27561861276626587, "learning_rate": 3.221960167714885e-05, "loss": 0.7252, "num_input_tokens_seen": 1610992, "step": 2460 }, { "epoch": 1.2919287211740043, "grad_norm": 0.3324461281299591, "learning_rate": 3.228511530398323e-05, "loss": 0.6228, "num_input_tokens_seen": 1614000, "step": 2465 }, { "epoch": 1.2945492662473794, "grad_norm": 0.18591269850730896, "learning_rate": 3.235062893081761e-05, "loss": 0.67, "num_input_tokens_seen": 1616752, "step": 2470 }, { "epoch": 1.2971698113207548, "grad_norm": 0.14549697935581207, "learning_rate": 3.241614255765199e-05, "loss": 0.9394, "num_input_tokens_seen": 1620144, "step": 2475 }, { "epoch": 1.29979035639413, "grad_norm": 0.18827985227108002, "learning_rate": 3.2481656184486375e-05, "loss": 0.5475, "num_input_tokens_seen": 1624880, "step": 2480 }, { "epoch": 1.3024109014675052, "grad_norm": 0.1559191793203354, "learning_rate": 3.254716981132075e-05, "loss": 0.6561, "num_input_tokens_seen": 1627984, "step": 2485 }, { "epoch": 1.3050314465408805, "grad_norm": 0.1516018509864807, "learning_rate": 3.261268343815514e-05, "loss": 0.7829, "num_input_tokens_seen": 1631408, "step": 2490 }, { "epoch": 1.3076519916142557, "grad_norm": 0.1936991661787033, "learning_rate": 3.267819706498952e-05, "loss": 0.7598, "num_input_tokens_seen": 1635248, "step": 2495 }, { "epoch": 1.310272536687631, "grad_norm": 0.22420626878738403, "learning_rate": 3.27437106918239e-05, "loss": 0.649, "num_input_tokens_seen": 1637904, "step": 2500 }, { "epoch": 1.3128930817610063, "grad_norm": 0.1975494921207428, "learning_rate": 3.2809224318658284e-05, "loss": 0.5217, "num_input_tokens_seen": 1643216, "step": 2505 }, { "epoch": 1.3155136268343814, "grad_norm": 0.15046566724777222, "learning_rate": 3.287473794549266e-05, "loss": 0.7847, "num_input_tokens_seen": 1646320, "step": 2510 }, { "epoch": 1.3181341719077568, "grad_norm": 0.2220218926668167, "learning_rate": 3.294025157232704e-05, "loss": 0.5829, "num_input_tokens_seen": 1650192, "step": 2515 }, { "epoch": 1.320754716981132, "grad_norm": 0.2575998306274414, "learning_rate": 3.3005765199161426e-05, "loss": 0.5999, "num_input_tokens_seen": 1653840, "step": 2520 }, { "epoch": 1.3233752620545074, "grad_norm": 0.14756865799427032, "learning_rate": 3.307127882599581e-05, "loss": 0.6025, "num_input_tokens_seen": 1656944, "step": 2525 }, { "epoch": 1.3259958071278826, "grad_norm": 0.2310672253370285, "learning_rate": 3.313679245283019e-05, "loss": 0.662, "num_input_tokens_seen": 1661424, "step": 2530 }, { "epoch": 1.3286163522012577, "grad_norm": 0.22280137240886688, "learning_rate": 3.320230607966457e-05, "loss": 0.5718, "num_input_tokens_seen": 1664976, "step": 2535 }, { "epoch": 1.3312368972746331, "grad_norm": 0.2912602126598358, "learning_rate": 3.326781970649895e-05, "loss": 0.6184, "num_input_tokens_seen": 1668528, "step": 2540 }, { "epoch": 1.3338574423480085, "grad_norm": 0.19936062395572662, "learning_rate": 3.3333333333333335e-05, "loss": 0.5439, "num_input_tokens_seen": 1671888, "step": 2545 }, { "epoch": 1.3364779874213837, "grad_norm": 0.31007376313209534, "learning_rate": 3.339884696016772e-05, "loss": 0.5284, "num_input_tokens_seen": 1675280, "step": 2550 }, { "epoch": 1.3390985324947589, "grad_norm": 0.27189871668815613, "learning_rate": 3.3464360587002094e-05, "loss": 0.623, "num_input_tokens_seen": 1678256, "step": 2555 }, { "epoch": 1.3417190775681342, "grad_norm": 0.40850508213043213, "learning_rate": 3.3529874213836484e-05, "loss": 0.7301, "num_input_tokens_seen": 1681360, "step": 2560 }, { "epoch": 1.3443396226415094, "grad_norm": 0.20549780130386353, "learning_rate": 3.359538784067086e-05, "loss": 0.6989, "num_input_tokens_seen": 1684752, "step": 2565 }, { "epoch": 1.3469601677148848, "grad_norm": 0.19499731063842773, "learning_rate": 3.3660901467505244e-05, "loss": 0.5251, "num_input_tokens_seen": 1689072, "step": 2570 }, { "epoch": 1.34958071278826, "grad_norm": 0.23889148235321045, "learning_rate": 3.3726415094339627e-05, "loss": 0.6109, "num_input_tokens_seen": 1692048, "step": 2575 }, { "epoch": 1.3522012578616351, "grad_norm": 0.3749021887779236, "learning_rate": 3.3791928721174e-05, "loss": 0.5959, "num_input_tokens_seen": 1694800, "step": 2580 }, { "epoch": 1.3548218029350105, "grad_norm": 0.19697485864162445, "learning_rate": 3.3857442348008386e-05, "loss": 0.582, "num_input_tokens_seen": 1698768, "step": 2585 }, { "epoch": 1.3574423480083857, "grad_norm": 0.5020652413368225, "learning_rate": 3.392295597484277e-05, "loss": 0.6606, "num_input_tokens_seen": 1701616, "step": 2590 }, { "epoch": 1.360062893081761, "grad_norm": 0.24250811338424683, "learning_rate": 3.398846960167715e-05, "loss": 0.6502, "num_input_tokens_seen": 1704528, "step": 2595 }, { "epoch": 1.3626834381551363, "grad_norm": 0.25278162956237793, "learning_rate": 3.4053983228511535e-05, "loss": 0.6043, "num_input_tokens_seen": 1708112, "step": 2600 }, { "epoch": 1.3653039832285114, "grad_norm": 0.3500746786594391, "learning_rate": 3.411949685534591e-05, "loss": 0.4679, "num_input_tokens_seen": 1710864, "step": 2605 }, { "epoch": 1.3679245283018868, "grad_norm": 0.21429443359375, "learning_rate": 3.4185010482180295e-05, "loss": 0.4589, "num_input_tokens_seen": 1713648, "step": 2610 }, { "epoch": 1.370545073375262, "grad_norm": 0.2970775067806244, "learning_rate": 3.425052410901468e-05, "loss": 0.7335, "num_input_tokens_seen": 1716176, "step": 2615 }, { "epoch": 1.3731656184486374, "grad_norm": 0.4238505959510803, "learning_rate": 3.431603773584906e-05, "loss": 0.548, "num_input_tokens_seen": 1719344, "step": 2620 }, { "epoch": 1.3757861635220126, "grad_norm": 0.29425492882728577, "learning_rate": 3.438155136268344e-05, "loss": 0.6246, "num_input_tokens_seen": 1722800, "step": 2625 }, { "epoch": 1.3784067085953877, "grad_norm": 0.31119921803474426, "learning_rate": 3.444706498951782e-05, "loss": 0.757, "num_input_tokens_seen": 1725840, "step": 2630 }, { "epoch": 1.381027253668763, "grad_norm": 0.20143043994903564, "learning_rate": 3.45125786163522e-05, "loss": 0.5875, "num_input_tokens_seen": 1729424, "step": 2635 }, { "epoch": 1.3836477987421385, "grad_norm": 0.22049292922019958, "learning_rate": 3.457809224318658e-05, "loss": 0.6369, "num_input_tokens_seen": 1734128, "step": 2640 }, { "epoch": 1.3862683438155137, "grad_norm": 0.3310697078704834, "learning_rate": 3.464360587002097e-05, "loss": 0.686, "num_input_tokens_seen": 1737136, "step": 2645 }, { "epoch": 1.3888888888888888, "grad_norm": 0.5504927635192871, "learning_rate": 3.4709119496855346e-05, "loss": 0.7388, "num_input_tokens_seen": 1739696, "step": 2650 }, { "epoch": 1.3915094339622642, "grad_norm": 0.17181190848350525, "learning_rate": 3.477463312368973e-05, "loss": 0.6005, "num_input_tokens_seen": 1742608, "step": 2655 }, { "epoch": 1.3941299790356394, "grad_norm": 0.17898516356945038, "learning_rate": 3.484014675052411e-05, "loss": 0.5841, "num_input_tokens_seen": 1745584, "step": 2660 }, { "epoch": 1.3967505241090148, "grad_norm": 0.14770860970020294, "learning_rate": 3.490566037735849e-05, "loss": 0.6312, "num_input_tokens_seen": 1748976, "step": 2665 }, { "epoch": 1.39937106918239, "grad_norm": 0.28357067704200745, "learning_rate": 3.497117400419288e-05, "loss": 1.0031, "num_input_tokens_seen": 1751696, "step": 2670 }, { "epoch": 1.4019916142557651, "grad_norm": 0.26138630509376526, "learning_rate": 3.5036687631027254e-05, "loss": 0.7126, "num_input_tokens_seen": 1754640, "step": 2675 }, { "epoch": 1.4046121593291405, "grad_norm": 0.5562939047813416, "learning_rate": 3.510220125786164e-05, "loss": 0.7424, "num_input_tokens_seen": 1757488, "step": 2680 }, { "epoch": 1.4072327044025157, "grad_norm": 0.38364964723587036, "learning_rate": 3.516771488469602e-05, "loss": 0.6623, "num_input_tokens_seen": 1760368, "step": 2685 }, { "epoch": 1.409853249475891, "grad_norm": 0.1165279746055603, "learning_rate": 3.5233228511530397e-05, "loss": 0.559, "num_input_tokens_seen": 1763184, "step": 2690 }, { "epoch": 1.4124737945492662, "grad_norm": 0.25218483805656433, "learning_rate": 3.529874213836478e-05, "loss": 0.5175, "num_input_tokens_seen": 1766352, "step": 2695 }, { "epoch": 1.4150943396226414, "grad_norm": 0.2748885452747345, "learning_rate": 3.536425576519916e-05, "loss": 0.6983, "num_input_tokens_seen": 1769936, "step": 2700 }, { "epoch": 1.4177148846960168, "grad_norm": 0.24015778303146362, "learning_rate": 3.5429769392033546e-05, "loss": 0.727, "num_input_tokens_seen": 1773040, "step": 2705 }, { "epoch": 1.420335429769392, "grad_norm": 0.17782634496688843, "learning_rate": 3.549528301886792e-05, "loss": 0.5668, "num_input_tokens_seen": 1776688, "step": 2710 }, { "epoch": 1.4229559748427674, "grad_norm": 0.2327415496110916, "learning_rate": 3.556079664570231e-05, "loss": 0.6152, "num_input_tokens_seen": 1779728, "step": 2715 }, { "epoch": 1.4255765199161425, "grad_norm": 0.2211238592863083, "learning_rate": 3.562631027253669e-05, "loss": 0.5169, "num_input_tokens_seen": 1783056, "step": 2720 }, { "epoch": 1.4281970649895177, "grad_norm": 0.33798107504844666, "learning_rate": 3.569182389937107e-05, "loss": 0.6716, "num_input_tokens_seen": 1785840, "step": 2725 }, { "epoch": 1.430817610062893, "grad_norm": 0.33073678612709045, "learning_rate": 3.5757337526205454e-05, "loss": 0.7095, "num_input_tokens_seen": 1788080, "step": 2730 }, { "epoch": 1.4334381551362683, "grad_norm": 0.2837933897972107, "learning_rate": 3.582285115303983e-05, "loss": 0.7459, "num_input_tokens_seen": 1791600, "step": 2735 }, { "epoch": 1.4360587002096437, "grad_norm": 0.18412280082702637, "learning_rate": 3.588836477987422e-05, "loss": 0.6351, "num_input_tokens_seen": 1794832, "step": 2740 }, { "epoch": 1.4386792452830188, "grad_norm": 0.21222054958343506, "learning_rate": 3.59538784067086e-05, "loss": 0.6768, "num_input_tokens_seen": 1798256, "step": 2745 }, { "epoch": 1.441299790356394, "grad_norm": 0.18440885841846466, "learning_rate": 3.601939203354298e-05, "loss": 0.774, "num_input_tokens_seen": 1802224, "step": 2750 }, { "epoch": 1.4439203354297694, "grad_norm": 0.1859092116355896, "learning_rate": 3.608490566037736e-05, "loss": 0.5391, "num_input_tokens_seen": 1805648, "step": 2755 }, { "epoch": 1.4465408805031448, "grad_norm": 0.24502365291118622, "learning_rate": 3.615041928721174e-05, "loss": 0.5456, "num_input_tokens_seen": 1808528, "step": 2760 }, { "epoch": 1.44916142557652, "grad_norm": 0.2408255934715271, "learning_rate": 3.621593291404612e-05, "loss": 0.6701, "num_input_tokens_seen": 1811888, "step": 2765 }, { "epoch": 1.4517819706498951, "grad_norm": 0.37274160981178284, "learning_rate": 3.6281446540880505e-05, "loss": 0.6405, "num_input_tokens_seen": 1814672, "step": 2770 }, { "epoch": 1.4544025157232705, "grad_norm": 0.3257506787776947, "learning_rate": 3.634696016771489e-05, "loss": 0.5928, "num_input_tokens_seen": 1817552, "step": 2775 }, { "epoch": 1.4570230607966457, "grad_norm": 0.21404534578323364, "learning_rate": 3.6412473794549265e-05, "loss": 0.611, "num_input_tokens_seen": 1820464, "step": 2780 }, { "epoch": 1.459643605870021, "grad_norm": 0.25751906633377075, "learning_rate": 3.647798742138365e-05, "loss": 0.5734, "num_input_tokens_seen": 1823152, "step": 2785 }, { "epoch": 1.4622641509433962, "grad_norm": 0.15449781715869904, "learning_rate": 3.654350104821803e-05, "loss": 0.5791, "num_input_tokens_seen": 1826320, "step": 2790 }, { "epoch": 1.4648846960167714, "grad_norm": 0.24271880090236664, "learning_rate": 3.6609014675052414e-05, "loss": 0.7565, "num_input_tokens_seen": 1830640, "step": 2795 }, { "epoch": 1.4675052410901468, "grad_norm": 0.12682276964187622, "learning_rate": 3.66745283018868e-05, "loss": 0.5302, "num_input_tokens_seen": 1833712, "step": 2800 }, { "epoch": 1.470125786163522, "grad_norm": 0.25358846783638, "learning_rate": 3.6740041928721173e-05, "loss": 0.6807, "num_input_tokens_seen": 1837616, "step": 2805 }, { "epoch": 1.4727463312368974, "grad_norm": 0.21814404428005219, "learning_rate": 3.6805555555555556e-05, "loss": 0.8516, "num_input_tokens_seen": 1841264, "step": 2810 }, { "epoch": 1.4753668763102725, "grad_norm": 0.29055267572402954, "learning_rate": 3.687106918238994e-05, "loss": 0.5484, "num_input_tokens_seen": 1843888, "step": 2815 }, { "epoch": 1.4779874213836477, "grad_norm": 0.14371107518672943, "learning_rate": 3.6936582809224316e-05, "loss": 0.6357, "num_input_tokens_seen": 1847440, "step": 2820 }, { "epoch": 1.480607966457023, "grad_norm": 0.7214068174362183, "learning_rate": 3.7002096436058706e-05, "loss": 0.7798, "num_input_tokens_seen": 1850576, "step": 2825 }, { "epoch": 1.4832285115303983, "grad_norm": 0.21192702651023865, "learning_rate": 3.706761006289308e-05, "loss": 0.5422, "num_input_tokens_seen": 1853200, "step": 2830 }, { "epoch": 1.4858490566037736, "grad_norm": 1.5926295518875122, "learning_rate": 3.7133123689727465e-05, "loss": 0.9536, "num_input_tokens_seen": 1856784, "step": 2835 }, { "epoch": 1.4884696016771488, "grad_norm": 0.23497509956359863, "learning_rate": 3.719863731656185e-05, "loss": 0.7067, "num_input_tokens_seen": 1859984, "step": 2840 }, { "epoch": 1.491090146750524, "grad_norm": 0.17007172107696533, "learning_rate": 3.7264150943396224e-05, "loss": 0.588, "num_input_tokens_seen": 1863920, "step": 2845 }, { "epoch": 1.4937106918238994, "grad_norm": 0.17280055582523346, "learning_rate": 3.732966457023061e-05, "loss": 0.725, "num_input_tokens_seen": 1867664, "step": 2850 }, { "epoch": 1.4963312368972748, "grad_norm": 0.22589550912380219, "learning_rate": 3.739517819706499e-05, "loss": 0.5967, "num_input_tokens_seen": 1870352, "step": 2855 }, { "epoch": 1.49895178197065, "grad_norm": 0.22272615134716034, "learning_rate": 3.7460691823899374e-05, "loss": 0.6786, "num_input_tokens_seen": 1873488, "step": 2860 }, { "epoch": 1.501572327044025, "grad_norm": 0.3178577423095703, "learning_rate": 3.752620545073376e-05, "loss": 0.5538, "num_input_tokens_seen": 1876496, "step": 2865 }, { "epoch": 1.5041928721174003, "grad_norm": 0.2934839725494385, "learning_rate": 3.759171907756814e-05, "loss": 0.7355, "num_input_tokens_seen": 1880304, "step": 2870 }, { "epoch": 1.5068134171907757, "grad_norm": 0.24912990629673004, "learning_rate": 3.7657232704402516e-05, "loss": 0.5586, "num_input_tokens_seen": 1882960, "step": 2875 }, { "epoch": 1.509433962264151, "grad_norm": 0.12079175561666489, "learning_rate": 3.77227463312369e-05, "loss": 0.5201, "num_input_tokens_seen": 1886288, "step": 2880 }, { "epoch": 1.5120545073375262, "grad_norm": 0.21943499147891998, "learning_rate": 3.778825995807128e-05, "loss": 0.4916, "num_input_tokens_seen": 1890416, "step": 2885 }, { "epoch": 1.5146750524109014, "grad_norm": 0.16202165186405182, "learning_rate": 3.785377358490566e-05, "loss": 0.6207, "num_input_tokens_seen": 1893648, "step": 2890 }, { "epoch": 1.5172955974842768, "grad_norm": 0.18375547230243683, "learning_rate": 3.791928721174005e-05, "loss": 0.5126, "num_input_tokens_seen": 1897296, "step": 2895 }, { "epoch": 1.519916142557652, "grad_norm": 0.20829921960830688, "learning_rate": 3.7984800838574425e-05, "loss": 0.5783, "num_input_tokens_seen": 1899856, "step": 2900 }, { "epoch": 1.5225366876310273, "grad_norm": 0.19792994856834412, "learning_rate": 3.805031446540881e-05, "loss": 0.4536, "num_input_tokens_seen": 1906544, "step": 2905 }, { "epoch": 1.5251572327044025, "grad_norm": 0.18968787789344788, "learning_rate": 3.811582809224319e-05, "loss": 0.5356, "num_input_tokens_seen": 1909680, "step": 2910 }, { "epoch": 1.5277777777777777, "grad_norm": 0.1174970269203186, "learning_rate": 3.818134171907757e-05, "loss": 0.5505, "num_input_tokens_seen": 1913008, "step": 2915 }, { "epoch": 1.530398322851153, "grad_norm": 0.22536718845367432, "learning_rate": 3.824685534591195e-05, "loss": 0.7015, "num_input_tokens_seen": 1916080, "step": 2920 }, { "epoch": 1.5330188679245285, "grad_norm": 0.2557680904865265, "learning_rate": 3.831236897274633e-05, "loss": 0.644, "num_input_tokens_seen": 1919504, "step": 2925 }, { "epoch": 1.5356394129979036, "grad_norm": 0.27001768350601196, "learning_rate": 3.8377882599580716e-05, "loss": 0.6364, "num_input_tokens_seen": 1922736, "step": 2930 }, { "epoch": 1.5382599580712788, "grad_norm": 0.2200794517993927, "learning_rate": 3.844339622641509e-05, "loss": 0.4651, "num_input_tokens_seen": 1925456, "step": 2935 }, { "epoch": 1.540880503144654, "grad_norm": 0.16220983862876892, "learning_rate": 3.8508909853249476e-05, "loss": 0.4921, "num_input_tokens_seen": 1929936, "step": 2940 }, { "epoch": 1.5435010482180294, "grad_norm": 0.20675507187843323, "learning_rate": 3.857442348008386e-05, "loss": 0.5357, "num_input_tokens_seen": 1933232, "step": 2945 }, { "epoch": 1.5461215932914047, "grad_norm": 0.1499308943748474, "learning_rate": 3.863993710691824e-05, "loss": 0.6412, "num_input_tokens_seen": 1938288, "step": 2950 }, { "epoch": 1.54874213836478, "grad_norm": 0.28373947739601135, "learning_rate": 3.8705450733752625e-05, "loss": 0.574, "num_input_tokens_seen": 1941104, "step": 2955 }, { "epoch": 1.551362683438155, "grad_norm": 0.20363861322402954, "learning_rate": 3.8770964360587e-05, "loss": 0.4824, "num_input_tokens_seen": 1943632, "step": 2960 }, { "epoch": 1.5539832285115303, "grad_norm": 0.25152280926704407, "learning_rate": 3.883647798742139e-05, "loss": 0.7726, "num_input_tokens_seen": 1946288, "step": 2965 }, { "epoch": 1.5566037735849056, "grad_norm": 0.23374368250370026, "learning_rate": 3.890199161425577e-05, "loss": 0.5844, "num_input_tokens_seen": 1949264, "step": 2970 }, { "epoch": 1.559224318658281, "grad_norm": 0.2312510758638382, "learning_rate": 3.8967505241090144e-05, "loss": 0.7608, "num_input_tokens_seen": 1952016, "step": 2975 }, { "epoch": 1.5618448637316562, "grad_norm": 0.1337854117155075, "learning_rate": 3.9033018867924534e-05, "loss": 0.7161, "num_input_tokens_seen": 1956112, "step": 2980 }, { "epoch": 1.5644654088050314, "grad_norm": 0.279483437538147, "learning_rate": 3.909853249475891e-05, "loss": 0.7289, "num_input_tokens_seen": 1959248, "step": 2985 }, { "epoch": 1.5670859538784065, "grad_norm": 0.29173046350479126, "learning_rate": 3.916404612159329e-05, "loss": 0.5179, "num_input_tokens_seen": 1961616, "step": 2990 }, { "epoch": 1.569706498951782, "grad_norm": 0.18946458399295807, "learning_rate": 3.9229559748427676e-05, "loss": 0.5878, "num_input_tokens_seen": 1964720, "step": 2995 }, { "epoch": 1.5723270440251573, "grad_norm": 0.21837005019187927, "learning_rate": 3.929507337526205e-05, "loss": 0.5238, "num_input_tokens_seen": 1967280, "step": 3000 }, { "epoch": 1.5749475890985325, "grad_norm": 0.15026918053627014, "learning_rate": 3.9360587002096435e-05, "loss": 0.6278, "num_input_tokens_seen": 1970992, "step": 3005 }, { "epoch": 1.5775681341719077, "grad_norm": 0.2054615616798401, "learning_rate": 3.942610062893082e-05, "loss": 0.5401, "num_input_tokens_seen": 1974320, "step": 3010 }, { "epoch": 1.580188679245283, "grad_norm": 0.2727251946926117, "learning_rate": 3.94916142557652e-05, "loss": 0.4871, "num_input_tokens_seen": 1977680, "step": 3015 }, { "epoch": 1.5828092243186582, "grad_norm": 0.5051704049110413, "learning_rate": 3.9557127882599585e-05, "loss": 0.588, "num_input_tokens_seen": 1980688, "step": 3020 }, { "epoch": 1.5854297693920336, "grad_norm": 0.36313748359680176, "learning_rate": 3.962264150943397e-05, "loss": 0.5527, "num_input_tokens_seen": 1983728, "step": 3025 }, { "epoch": 1.5880503144654088, "grad_norm": 0.1814219057559967, "learning_rate": 3.9688155136268344e-05, "loss": 0.5746, "num_input_tokens_seen": 1986992, "step": 3030 }, { "epoch": 1.590670859538784, "grad_norm": 0.2367706447839737, "learning_rate": 3.975366876310273e-05, "loss": 0.5847, "num_input_tokens_seen": 1990032, "step": 3035 }, { "epoch": 1.5932914046121593, "grad_norm": 0.46186015009880066, "learning_rate": 3.981918238993711e-05, "loss": 0.5612, "num_input_tokens_seen": 1992912, "step": 3040 }, { "epoch": 1.5959119496855347, "grad_norm": 0.2575664520263672, "learning_rate": 3.9884696016771486e-05, "loss": 0.5585, "num_input_tokens_seen": 1995664, "step": 3045 }, { "epoch": 1.59853249475891, "grad_norm": 0.13381458818912506, "learning_rate": 3.9950209643605876e-05, "loss": 0.5399, "num_input_tokens_seen": 1999696, "step": 3050 }, { "epoch": 1.601153039832285, "grad_norm": 0.2541233003139496, "learning_rate": 4.001572327044025e-05, "loss": 0.5396, "num_input_tokens_seen": 2002128, "step": 3055 }, { "epoch": 1.6037735849056602, "grad_norm": 0.18262629210948944, "learning_rate": 4.0081236897274636e-05, "loss": 0.5568, "num_input_tokens_seen": 2005712, "step": 3060 }, { "epoch": 1.6063941299790356, "grad_norm": 0.21714042127132416, "learning_rate": 4.014675052410902e-05, "loss": 0.539, "num_input_tokens_seen": 2008720, "step": 3065 }, { "epoch": 1.609014675052411, "grad_norm": 0.18231244385242462, "learning_rate": 4.0212264150943395e-05, "loss": 0.6673, "num_input_tokens_seen": 2012176, "step": 3070 }, { "epoch": 1.6116352201257862, "grad_norm": 0.26008257269859314, "learning_rate": 4.027777777777778e-05, "loss": 0.5886, "num_input_tokens_seen": 2014448, "step": 3075 }, { "epoch": 1.6142557651991614, "grad_norm": 0.12999586760997772, "learning_rate": 4.034329140461216e-05, "loss": 0.5979, "num_input_tokens_seen": 2017488, "step": 3080 }, { "epoch": 1.6168763102725365, "grad_norm": 0.3802458941936493, "learning_rate": 4.0408805031446544e-05, "loss": 0.5024, "num_input_tokens_seen": 2020400, "step": 3085 }, { "epoch": 1.619496855345912, "grad_norm": 0.2818225622177124, "learning_rate": 4.047431865828093e-05, "loss": 0.5699, "num_input_tokens_seen": 2023216, "step": 3090 }, { "epoch": 1.6221174004192873, "grad_norm": 0.18837189674377441, "learning_rate": 4.0539832285115304e-05, "loss": 0.5944, "num_input_tokens_seen": 2028368, "step": 3095 }, { "epoch": 1.6247379454926625, "grad_norm": 0.4090529680252075, "learning_rate": 4.0605345911949687e-05, "loss": 0.6435, "num_input_tokens_seen": 2030896, "step": 3100 }, { "epoch": 1.6273584905660377, "grad_norm": 0.2689136266708374, "learning_rate": 4.067085953878407e-05, "loss": 0.614, "num_input_tokens_seen": 2034128, "step": 3105 }, { "epoch": 1.629979035639413, "grad_norm": 0.20486722886562347, "learning_rate": 4.073637316561845e-05, "loss": 0.6262, "num_input_tokens_seen": 2037584, "step": 3110 }, { "epoch": 1.6325995807127882, "grad_norm": 0.1934918314218521, "learning_rate": 4.080188679245283e-05, "loss": 0.5424, "num_input_tokens_seen": 2040368, "step": 3115 }, { "epoch": 1.6352201257861636, "grad_norm": 0.18099544942378998, "learning_rate": 4.086740041928722e-05, "loss": 0.5855, "num_input_tokens_seen": 2043440, "step": 3120 }, { "epoch": 1.6378406708595388, "grad_norm": 0.2458605170249939, "learning_rate": 4.0932914046121595e-05, "loss": 0.725, "num_input_tokens_seen": 2046224, "step": 3125 }, { "epoch": 1.640461215932914, "grad_norm": 0.20443691313266754, "learning_rate": 4.099842767295597e-05, "loss": 0.5286, "num_input_tokens_seen": 2049264, "step": 3130 }, { "epoch": 1.6430817610062893, "grad_norm": 0.20552867650985718, "learning_rate": 4.106394129979036e-05, "loss": 0.5251, "num_input_tokens_seen": 2053232, "step": 3135 }, { "epoch": 1.6457023060796647, "grad_norm": 0.19627033174037933, "learning_rate": 4.112945492662474e-05, "loss": 0.5298, "num_input_tokens_seen": 2056592, "step": 3140 }, { "epoch": 1.64832285115304, "grad_norm": 0.18907597661018372, "learning_rate": 4.119496855345912e-05, "loss": 0.6018, "num_input_tokens_seen": 2060688, "step": 3145 }, { "epoch": 1.650943396226415, "grad_norm": 0.18543770909309387, "learning_rate": 4.1260482180293504e-05, "loss": 0.5453, "num_input_tokens_seen": 2064752, "step": 3150 }, { "epoch": 1.6535639412997902, "grad_norm": 0.22132088243961334, "learning_rate": 4.132599580712788e-05, "loss": 0.5796, "num_input_tokens_seen": 2067280, "step": 3155 }, { "epoch": 1.6561844863731656, "grad_norm": 0.32024165987968445, "learning_rate": 4.139150943396227e-05, "loss": 0.606, "num_input_tokens_seen": 2070608, "step": 3160 }, { "epoch": 1.658805031446541, "grad_norm": 0.17991171777248383, "learning_rate": 4.1457023060796646e-05, "loss": 0.5619, "num_input_tokens_seen": 2074160, "step": 3165 }, { "epoch": 1.6614255765199162, "grad_norm": 0.21489085257053375, "learning_rate": 4.152253668763103e-05, "loss": 0.5012, "num_input_tokens_seen": 2077168, "step": 3170 }, { "epoch": 1.6640461215932913, "grad_norm": 0.2878616452217102, "learning_rate": 4.158805031446541e-05, "loss": 0.5708, "num_input_tokens_seen": 2080784, "step": 3175 }, { "epoch": 1.6666666666666665, "grad_norm": 0.328632652759552, "learning_rate": 4.1653563941299795e-05, "loss": 0.6148, "num_input_tokens_seen": 2084400, "step": 3180 }, { "epoch": 1.669287211740042, "grad_norm": 0.1996048539876938, "learning_rate": 4.171907756813417e-05, "loss": 0.6625, "num_input_tokens_seen": 2087632, "step": 3185 }, { "epoch": 1.6719077568134173, "grad_norm": 0.21520909667015076, "learning_rate": 4.1784591194968555e-05, "loss": 0.6558, "num_input_tokens_seen": 2091280, "step": 3190 }, { "epoch": 1.6745283018867925, "grad_norm": 0.1942380666732788, "learning_rate": 4.185010482180294e-05, "loss": 0.5613, "num_input_tokens_seen": 2094640, "step": 3195 }, { "epoch": 1.6771488469601676, "grad_norm": 0.2263624668121338, "learning_rate": 4.1915618448637314e-05, "loss": 0.7612, "num_input_tokens_seen": 2098064, "step": 3200 }, { "epoch": 1.679769392033543, "grad_norm": 0.20416384935379028, "learning_rate": 4.1981132075471704e-05, "loss": 0.5737, "num_input_tokens_seen": 2102064, "step": 3205 }, { "epoch": 1.6823899371069182, "grad_norm": 0.1799529492855072, "learning_rate": 4.204664570230608e-05, "loss": 0.745, "num_input_tokens_seen": 2105296, "step": 3210 }, { "epoch": 1.6850104821802936, "grad_norm": 0.2072146087884903, "learning_rate": 4.2112159329140463e-05, "loss": 0.53, "num_input_tokens_seen": 2108592, "step": 3215 }, { "epoch": 1.6876310272536688, "grad_norm": 0.19064772129058838, "learning_rate": 4.2177672955974846e-05, "loss": 0.4173, "num_input_tokens_seen": 2111632, "step": 3220 }, { "epoch": 1.690251572327044, "grad_norm": 0.17999328672885895, "learning_rate": 4.224318658280922e-05, "loss": 0.7948, "num_input_tokens_seen": 2115792, "step": 3225 }, { "epoch": 1.6928721174004193, "grad_norm": 0.1693917214870453, "learning_rate": 4.2308700209643606e-05, "loss": 0.528, "num_input_tokens_seen": 2118832, "step": 3230 }, { "epoch": 1.6954926624737947, "grad_norm": 0.15523962676525116, "learning_rate": 4.237421383647799e-05, "loss": 0.5528, "num_input_tokens_seen": 2122032, "step": 3235 }, { "epoch": 1.6981132075471699, "grad_norm": 0.2699625492095947, "learning_rate": 4.243972746331237e-05, "loss": 0.5631, "num_input_tokens_seen": 2124400, "step": 3240 }, { "epoch": 1.700733752620545, "grad_norm": 0.20973248779773712, "learning_rate": 4.2505241090146755e-05, "loss": 0.6475, "num_input_tokens_seen": 2126928, "step": 3245 }, { "epoch": 1.7033542976939202, "grad_norm": 0.1583462804555893, "learning_rate": 4.257075471698113e-05, "loss": 0.5262, "num_input_tokens_seen": 2130128, "step": 3250 }, { "epoch": 1.7059748427672956, "grad_norm": 0.12714610993862152, "learning_rate": 4.2636268343815514e-05, "loss": 0.6751, "num_input_tokens_seen": 2133360, "step": 3255 }, { "epoch": 1.708595387840671, "grad_norm": 0.24097836017608643, "learning_rate": 4.27017819706499e-05, "loss": 0.7103, "num_input_tokens_seen": 2136464, "step": 3260 }, { "epoch": 1.7112159329140462, "grad_norm": 0.14087195694446564, "learning_rate": 4.276729559748428e-05, "loss": 0.7588, "num_input_tokens_seen": 2140080, "step": 3265 }, { "epoch": 1.7138364779874213, "grad_norm": 0.8079196214675903, "learning_rate": 4.283280922431866e-05, "loss": 0.7255, "num_input_tokens_seen": 2142960, "step": 3270 }, { "epoch": 1.7164570230607965, "grad_norm": 0.15594889223575592, "learning_rate": 4.289832285115305e-05, "loss": 0.6001, "num_input_tokens_seen": 2146224, "step": 3275 }, { "epoch": 1.719077568134172, "grad_norm": 0.16321516036987305, "learning_rate": 4.296383647798742e-05, "loss": 0.5417, "num_input_tokens_seen": 2150992, "step": 3280 }, { "epoch": 1.7216981132075473, "grad_norm": 0.21281692385673523, "learning_rate": 4.3029350104821806e-05, "loss": 0.7187, "num_input_tokens_seen": 2154448, "step": 3285 }, { "epoch": 1.7243186582809225, "grad_norm": 0.3205283582210541, "learning_rate": 4.309486373165619e-05, "loss": 0.4592, "num_input_tokens_seen": 2156752, "step": 3290 }, { "epoch": 1.7269392033542976, "grad_norm": 0.1727067530155182, "learning_rate": 4.3160377358490565e-05, "loss": 0.6469, "num_input_tokens_seen": 2159568, "step": 3295 }, { "epoch": 1.7295597484276728, "grad_norm": 0.2298641800880432, "learning_rate": 4.322589098532495e-05, "loss": 0.6064, "num_input_tokens_seen": 2162736, "step": 3300 }, { "epoch": 1.7321802935010482, "grad_norm": 0.12352199852466583, "learning_rate": 4.329140461215933e-05, "loss": 0.6863, "num_input_tokens_seen": 2166288, "step": 3305 }, { "epoch": 1.7348008385744236, "grad_norm": 0.24256537854671478, "learning_rate": 4.335691823899371e-05, "loss": 0.6109, "num_input_tokens_seen": 2169584, "step": 3310 }, { "epoch": 1.7374213836477987, "grad_norm": 0.21830366551876068, "learning_rate": 4.34224318658281e-05, "loss": 0.5803, "num_input_tokens_seen": 2172560, "step": 3315 }, { "epoch": 1.740041928721174, "grad_norm": 0.45943182706832886, "learning_rate": 4.3487945492662474e-05, "loss": 0.56, "num_input_tokens_seen": 2175600, "step": 3320 }, { "epoch": 1.7426624737945493, "grad_norm": 0.18340912461280823, "learning_rate": 4.355345911949686e-05, "loss": 0.5566, "num_input_tokens_seen": 2178800, "step": 3325 }, { "epoch": 1.7452830188679245, "grad_norm": 0.1355971246957779, "learning_rate": 4.361897274633124e-05, "loss": 0.5937, "num_input_tokens_seen": 2181744, "step": 3330 }, { "epoch": 1.7479035639412999, "grad_norm": 0.1768263727426529, "learning_rate": 4.368448637316562e-05, "loss": 0.5453, "num_input_tokens_seen": 2184720, "step": 3335 }, { "epoch": 1.750524109014675, "grad_norm": 0.3153042495250702, "learning_rate": 4.375e-05, "loss": 0.57, "num_input_tokens_seen": 2186992, "step": 3340 }, { "epoch": 1.7531446540880502, "grad_norm": 0.17497038841247559, "learning_rate": 4.381551362683438e-05, "loss": 0.5508, "num_input_tokens_seen": 2190192, "step": 3345 }, { "epoch": 1.7557651991614256, "grad_norm": 0.15422682464122772, "learning_rate": 4.3881027253668766e-05, "loss": 0.4875, "num_input_tokens_seen": 2193488, "step": 3350 }, { "epoch": 1.758385744234801, "grad_norm": 0.2769602835178375, "learning_rate": 4.394654088050314e-05, "loss": 0.5288, "num_input_tokens_seen": 2196528, "step": 3355 }, { "epoch": 1.7610062893081762, "grad_norm": 0.1679302304983139, "learning_rate": 4.401205450733753e-05, "loss": 0.4702, "num_input_tokens_seen": 2202224, "step": 3360 }, { "epoch": 1.7636268343815513, "grad_norm": 0.10773921757936478, "learning_rate": 4.407756813417191e-05, "loss": 0.4935, "num_input_tokens_seen": 2206160, "step": 3365 }, { "epoch": 1.7662473794549265, "grad_norm": 0.26973652839660645, "learning_rate": 4.414308176100629e-05, "loss": 0.5364, "num_input_tokens_seen": 2209296, "step": 3370 }, { "epoch": 1.7688679245283019, "grad_norm": 0.2381792813539505, "learning_rate": 4.4208595387840674e-05, "loss": 0.5552, "num_input_tokens_seen": 2212368, "step": 3375 }, { "epoch": 1.7714884696016773, "grad_norm": 0.2729581594467163, "learning_rate": 4.427410901467505e-05, "loss": 0.5884, "num_input_tokens_seen": 2215760, "step": 3380 }, { "epoch": 1.7741090146750524, "grad_norm": 0.16684648394584656, "learning_rate": 4.433962264150944e-05, "loss": 0.6569, "num_input_tokens_seen": 2219312, "step": 3385 }, { "epoch": 1.7767295597484276, "grad_norm": 0.34679675102233887, "learning_rate": 4.440513626834382e-05, "loss": 0.3757, "num_input_tokens_seen": 2221712, "step": 3390 }, { "epoch": 1.7793501048218028, "grad_norm": 0.18991594016551971, "learning_rate": 4.44706498951782e-05, "loss": 0.6461, "num_input_tokens_seen": 2224816, "step": 3395 }, { "epoch": 1.7819706498951782, "grad_norm": 0.195577472448349, "learning_rate": 4.453616352201258e-05, "loss": 0.4747, "num_input_tokens_seen": 2228464, "step": 3400 }, { "epoch": 1.7845911949685536, "grad_norm": 0.3036997616291046, "learning_rate": 4.460167714884696e-05, "loss": 0.5872, "num_input_tokens_seen": 2232592, "step": 3405 }, { "epoch": 1.7872117400419287, "grad_norm": 0.18560630083084106, "learning_rate": 4.466719077568134e-05, "loss": 0.5272, "num_input_tokens_seen": 2235728, "step": 3410 }, { "epoch": 1.789832285115304, "grad_norm": 0.17864814400672913, "learning_rate": 4.4732704402515725e-05, "loss": 0.3893, "num_input_tokens_seen": 2239216, "step": 3415 }, { "epoch": 1.7924528301886793, "grad_norm": 0.2816377580165863, "learning_rate": 4.479821802935011e-05, "loss": 0.5941, "num_input_tokens_seen": 2242256, "step": 3420 }, { "epoch": 1.7950733752620545, "grad_norm": 0.22289489209651947, "learning_rate": 4.4863731656184485e-05, "loss": 0.5504, "num_input_tokens_seen": 2245744, "step": 3425 }, { "epoch": 1.7976939203354299, "grad_norm": 0.3106836974620819, "learning_rate": 4.4929245283018875e-05, "loss": 0.5939, "num_input_tokens_seen": 2249808, "step": 3430 }, { "epoch": 1.800314465408805, "grad_norm": 0.2522095739841461, "learning_rate": 4.499475890985325e-05, "loss": 0.5445, "num_input_tokens_seen": 2252976, "step": 3435 }, { "epoch": 1.8029350104821802, "grad_norm": 0.2629152834415436, "learning_rate": 4.5060272536687634e-05, "loss": 0.5932, "num_input_tokens_seen": 2256368, "step": 3440 }, { "epoch": 1.8055555555555556, "grad_norm": 0.3554117977619171, "learning_rate": 4.512578616352202e-05, "loss": 0.5733, "num_input_tokens_seen": 2259536, "step": 3445 }, { "epoch": 1.808176100628931, "grad_norm": 0.198189377784729, "learning_rate": 4.519129979035639e-05, "loss": 0.55, "num_input_tokens_seen": 2262928, "step": 3450 }, { "epoch": 1.8107966457023061, "grad_norm": 0.6317784786224365, "learning_rate": 4.525681341719078e-05, "loss": 0.9707, "num_input_tokens_seen": 2265520, "step": 3455 }, { "epoch": 1.8134171907756813, "grad_norm": 0.23048336803913116, "learning_rate": 4.532232704402516e-05, "loss": 0.8058, "num_input_tokens_seen": 2268464, "step": 3460 }, { "epoch": 1.8160377358490565, "grad_norm": 0.17267677187919617, "learning_rate": 4.5387840670859536e-05, "loss": 0.5998, "num_input_tokens_seen": 2271504, "step": 3465 }, { "epoch": 1.8186582809224319, "grad_norm": 0.23854751884937286, "learning_rate": 4.5453354297693926e-05, "loss": 0.5479, "num_input_tokens_seen": 2274384, "step": 3470 }, { "epoch": 1.8212788259958073, "grad_norm": 0.1372595876455307, "learning_rate": 4.55188679245283e-05, "loss": 0.4305, "num_input_tokens_seen": 2277648, "step": 3475 }, { "epoch": 1.8238993710691824, "grad_norm": 0.43666180968284607, "learning_rate": 4.5584381551362685e-05, "loss": 0.7089, "num_input_tokens_seen": 2280528, "step": 3480 }, { "epoch": 1.8265199161425576, "grad_norm": 0.12473513931035995, "learning_rate": 4.564989517819707e-05, "loss": 0.6258, "num_input_tokens_seen": 2283792, "step": 3485 }, { "epoch": 1.8291404612159328, "grad_norm": 0.21573007106781006, "learning_rate": 4.571540880503145e-05, "loss": 0.7994, "num_input_tokens_seen": 2286832, "step": 3490 }, { "epoch": 1.8317610062893082, "grad_norm": 0.19326449930667877, "learning_rate": 4.578092243186583e-05, "loss": 0.4835, "num_input_tokens_seen": 2291312, "step": 3495 }, { "epoch": 1.8343815513626835, "grad_norm": 0.2758347988128662, "learning_rate": 4.584643605870021e-05, "loss": 0.559, "num_input_tokens_seen": 2294384, "step": 3500 }, { "epoch": 1.8370020964360587, "grad_norm": 0.16250452399253845, "learning_rate": 4.5911949685534594e-05, "loss": 0.6761, "num_input_tokens_seen": 2297232, "step": 3505 }, { "epoch": 1.8396226415094339, "grad_norm": 0.15205439925193787, "learning_rate": 4.5977463312368977e-05, "loss": 0.5783, "num_input_tokens_seen": 2300560, "step": 3510 }, { "epoch": 1.8422431865828093, "grad_norm": 0.3055301010608673, "learning_rate": 4.604297693920336e-05, "loss": 0.4187, "num_input_tokens_seen": 2303312, "step": 3515 }, { "epoch": 1.8448637316561844, "grad_norm": 0.17208430171012878, "learning_rate": 4.6108490566037736e-05, "loss": 0.6121, "num_input_tokens_seen": 2306096, "step": 3520 }, { "epoch": 1.8474842767295598, "grad_norm": 0.16713188588619232, "learning_rate": 4.617400419287212e-05, "loss": 0.5904, "num_input_tokens_seen": 2310096, "step": 3525 }, { "epoch": 1.850104821802935, "grad_norm": 0.190532386302948, "learning_rate": 4.62395178197065e-05, "loss": 0.5412, "num_input_tokens_seen": 2313072, "step": 3530 }, { "epoch": 1.8527253668763102, "grad_norm": 0.13549678027629852, "learning_rate": 4.630503144654088e-05, "loss": 0.5586, "num_input_tokens_seen": 2316304, "step": 3535 }, { "epoch": 1.8553459119496856, "grad_norm": 0.30521732568740845, "learning_rate": 4.637054507337527e-05, "loss": 0.6128, "num_input_tokens_seen": 2318832, "step": 3540 }, { "epoch": 1.857966457023061, "grad_norm": 0.2014981508255005, "learning_rate": 4.6436058700209645e-05, "loss": 0.577, "num_input_tokens_seen": 2321840, "step": 3545 }, { "epoch": 1.8605870020964361, "grad_norm": 0.21024784445762634, "learning_rate": 4.650157232704403e-05, "loss": 0.6608, "num_input_tokens_seen": 2324400, "step": 3550 }, { "epoch": 1.8632075471698113, "grad_norm": 0.2199363261461258, "learning_rate": 4.656708595387841e-05, "loss": 0.5086, "num_input_tokens_seen": 2327376, "step": 3555 }, { "epoch": 1.8658280922431865, "grad_norm": 0.09344957023859024, "learning_rate": 4.663259958071279e-05, "loss": 0.4928, "num_input_tokens_seen": 2330832, "step": 3560 }, { "epoch": 1.8684486373165619, "grad_norm": 0.21702177822589874, "learning_rate": 4.669811320754717e-05, "loss": 0.5659, "num_input_tokens_seen": 2335376, "step": 3565 }, { "epoch": 1.8710691823899372, "grad_norm": 0.2864896357059479, "learning_rate": 4.676362683438155e-05, "loss": 0.6069, "num_input_tokens_seen": 2338128, "step": 3570 }, { "epoch": 1.8736897274633124, "grad_norm": 0.15906767547130585, "learning_rate": 4.6829140461215936e-05, "loss": 0.4805, "num_input_tokens_seen": 2341552, "step": 3575 }, { "epoch": 1.8763102725366876, "grad_norm": 0.21418766677379608, "learning_rate": 4.689465408805032e-05, "loss": 0.6071, "num_input_tokens_seen": 2344304, "step": 3580 }, { "epoch": 1.8789308176100628, "grad_norm": 0.23137494921684265, "learning_rate": 4.69601677148847e-05, "loss": 0.4065, "num_input_tokens_seen": 2347216, "step": 3585 }, { "epoch": 1.8815513626834381, "grad_norm": 0.15868371725082397, "learning_rate": 4.702568134171908e-05, "loss": 0.6439, "num_input_tokens_seen": 2350640, "step": 3590 }, { "epoch": 1.8841719077568135, "grad_norm": 0.21329136192798615, "learning_rate": 4.709119496855346e-05, "loss": 0.5826, "num_input_tokens_seen": 2354224, "step": 3595 }, { "epoch": 1.8867924528301887, "grad_norm": 0.186593160033226, "learning_rate": 4.7156708595387845e-05, "loss": 0.5964, "num_input_tokens_seen": 2356688, "step": 3600 }, { "epoch": 1.8894129979035639, "grad_norm": 0.3092027008533478, "learning_rate": 4.722222222222222e-05, "loss": 0.5711, "num_input_tokens_seen": 2359568, "step": 3605 }, { "epoch": 1.892033542976939, "grad_norm": 0.17208540439605713, "learning_rate": 4.728773584905661e-05, "loss": 0.6434, "num_input_tokens_seen": 2363280, "step": 3610 }, { "epoch": 1.8946540880503144, "grad_norm": 0.2077549546957016, "learning_rate": 4.735324947589099e-05, "loss": 0.607, "num_input_tokens_seen": 2366480, "step": 3615 }, { "epoch": 1.8972746331236898, "grad_norm": 0.17551659047603607, "learning_rate": 4.7418763102725364e-05, "loss": 0.7127, "num_input_tokens_seen": 2369200, "step": 3620 }, { "epoch": 1.899895178197065, "grad_norm": 0.11749785393476486, "learning_rate": 4.7484276729559753e-05, "loss": 0.5464, "num_input_tokens_seen": 2373296, "step": 3625 }, { "epoch": 1.9025157232704402, "grad_norm": 0.18727520108222961, "learning_rate": 4.754979035639413e-05, "loss": 0.571, "num_input_tokens_seen": 2376432, "step": 3630 }, { "epoch": 1.9051362683438156, "grad_norm": 0.20027457177639008, "learning_rate": 4.761530398322851e-05, "loss": 0.5334, "num_input_tokens_seen": 2379344, "step": 3635 }, { "epoch": 1.9077568134171907, "grad_norm": 0.15296773612499237, "learning_rate": 4.7680817610062896e-05, "loss": 0.6542, "num_input_tokens_seen": 2382960, "step": 3640 }, { "epoch": 1.9103773584905661, "grad_norm": 0.19560664892196655, "learning_rate": 4.774633123689728e-05, "loss": 0.5611, "num_input_tokens_seen": 2386064, "step": 3645 }, { "epoch": 1.9129979035639413, "grad_norm": 0.1797192543745041, "learning_rate": 4.7811844863731655e-05, "loss": 0.5474, "num_input_tokens_seen": 2390384, "step": 3650 }, { "epoch": 1.9156184486373165, "grad_norm": 0.1460302621126175, "learning_rate": 4.787735849056604e-05, "loss": 0.6514, "num_input_tokens_seen": 2394384, "step": 3655 }, { "epoch": 1.9182389937106918, "grad_norm": 0.5466549396514893, "learning_rate": 4.794287211740042e-05, "loss": 0.5489, "num_input_tokens_seen": 2397584, "step": 3660 }, { "epoch": 1.9208595387840672, "grad_norm": 0.3442513048648834, "learning_rate": 4.8008385744234804e-05, "loss": 0.6445, "num_input_tokens_seen": 2400560, "step": 3665 }, { "epoch": 1.9234800838574424, "grad_norm": 0.28926196694374084, "learning_rate": 4.807389937106919e-05, "loss": 0.5802, "num_input_tokens_seen": 2403280, "step": 3670 }, { "epoch": 1.9261006289308176, "grad_norm": 0.413724809885025, "learning_rate": 4.8139412997903564e-05, "loss": 0.7479, "num_input_tokens_seen": 2406192, "step": 3675 }, { "epoch": 1.9287211740041927, "grad_norm": 0.43823668360710144, "learning_rate": 4.8204926624737954e-05, "loss": 0.5572, "num_input_tokens_seen": 2409584, "step": 3680 }, { "epoch": 1.9313417190775681, "grad_norm": 0.3475315272808075, "learning_rate": 4.827044025157233e-05, "loss": 0.5505, "num_input_tokens_seen": 2412400, "step": 3685 }, { "epoch": 1.9339622641509435, "grad_norm": 0.4618608057498932, "learning_rate": 4.8335953878406706e-05, "loss": 0.5425, "num_input_tokens_seen": 2415504, "step": 3690 }, { "epoch": 1.9365828092243187, "grad_norm": 0.1325797587633133, "learning_rate": 4.8401467505241096e-05, "loss": 0.6048, "num_input_tokens_seen": 2418256, "step": 3695 }, { "epoch": 1.9392033542976939, "grad_norm": 0.29447582364082336, "learning_rate": 4.846698113207547e-05, "loss": 0.5022, "num_input_tokens_seen": 2421776, "step": 3700 }, { "epoch": 1.941823899371069, "grad_norm": 0.18047355115413666, "learning_rate": 4.8532494758909855e-05, "loss": 0.7064, "num_input_tokens_seen": 2425808, "step": 3705 }, { "epoch": 1.9444444444444444, "grad_norm": 0.15473617613315582, "learning_rate": 4.859800838574424e-05, "loss": 0.4573, "num_input_tokens_seen": 2428880, "step": 3710 }, { "epoch": 1.9470649895178198, "grad_norm": 0.2272174060344696, "learning_rate": 4.8663522012578615e-05, "loss": 0.5806, "num_input_tokens_seen": 2432112, "step": 3715 }, { "epoch": 1.949685534591195, "grad_norm": 0.23630790412425995, "learning_rate": 4.8729035639413e-05, "loss": 0.4654, "num_input_tokens_seen": 2435120, "step": 3720 }, { "epoch": 1.9523060796645701, "grad_norm": 0.23766541481018066, "learning_rate": 4.879454926624738e-05, "loss": 0.6678, "num_input_tokens_seen": 2437392, "step": 3725 }, { "epoch": 1.9549266247379455, "grad_norm": 0.18895606696605682, "learning_rate": 4.8860062893081764e-05, "loss": 0.5676, "num_input_tokens_seen": 2439920, "step": 3730 }, { "epoch": 1.9575471698113207, "grad_norm": 0.17838552594184875, "learning_rate": 4.892557651991615e-05, "loss": 0.4493, "num_input_tokens_seen": 2443536, "step": 3735 }, { "epoch": 1.960167714884696, "grad_norm": 0.17039629817008972, "learning_rate": 4.899109014675053e-05, "loss": 0.5944, "num_input_tokens_seen": 2446096, "step": 3740 }, { "epoch": 1.9627882599580713, "grad_norm": 0.3845389187335968, "learning_rate": 4.9056603773584906e-05, "loss": 0.6742, "num_input_tokens_seen": 2448976, "step": 3745 }, { "epoch": 1.9654088050314464, "grad_norm": 0.16591879725456238, "learning_rate": 4.912211740041929e-05, "loss": 0.6762, "num_input_tokens_seen": 2451632, "step": 3750 }, { "epoch": 1.9680293501048218, "grad_norm": 0.143702894449234, "learning_rate": 4.918763102725367e-05, "loss": 0.7087, "num_input_tokens_seen": 2454928, "step": 3755 }, { "epoch": 1.9706498951781972, "grad_norm": 0.3096771240234375, "learning_rate": 4.925314465408805e-05, "loss": 0.5712, "num_input_tokens_seen": 2457232, "step": 3760 }, { "epoch": 1.9732704402515724, "grad_norm": 0.24432596564292908, "learning_rate": 4.931865828092244e-05, "loss": 0.6811, "num_input_tokens_seen": 2460176, "step": 3765 }, { "epoch": 1.9758909853249476, "grad_norm": 0.40825605392456055, "learning_rate": 4.9384171907756815e-05, "loss": 0.6855, "num_input_tokens_seen": 2464432, "step": 3770 }, { "epoch": 1.9785115303983227, "grad_norm": 0.2629067897796631, "learning_rate": 4.944968553459119e-05, "loss": 0.6145, "num_input_tokens_seen": 2469136, "step": 3775 }, { "epoch": 1.9811320754716981, "grad_norm": 0.1592777818441391, "learning_rate": 4.951519916142558e-05, "loss": 0.518, "num_input_tokens_seen": 2472400, "step": 3780 }, { "epoch": 1.9837526205450735, "grad_norm": 0.2126634567975998, "learning_rate": 4.958071278825996e-05, "loss": 0.4949, "num_input_tokens_seen": 2477328, "step": 3785 }, { "epoch": 1.9863731656184487, "grad_norm": 0.20206141471862793, "learning_rate": 4.964622641509434e-05, "loss": 0.5722, "num_input_tokens_seen": 2480656, "step": 3790 }, { "epoch": 1.9889937106918238, "grad_norm": 0.25675299763679504, "learning_rate": 4.9711740041928724e-05, "loss": 0.575, "num_input_tokens_seen": 2484816, "step": 3795 }, { "epoch": 1.991614255765199, "grad_norm": 0.12050213664770126, "learning_rate": 4.977725366876311e-05, "loss": 0.6054, "num_input_tokens_seen": 2488752, "step": 3800 }, { "epoch": 1.9942348008385744, "grad_norm": 0.1698211133480072, "learning_rate": 4.984276729559749e-05, "loss": 0.4971, "num_input_tokens_seen": 2491952, "step": 3805 }, { "epoch": 1.9968553459119498, "grad_norm": 0.12132832407951355, "learning_rate": 4.9908280922431866e-05, "loss": 0.6606, "num_input_tokens_seen": 2495248, "step": 3810 }, { "epoch": 1.999475890985325, "grad_norm": 0.12374335527420044, "learning_rate": 4.997379454926625e-05, "loss": 0.6218, "num_input_tokens_seen": 2497840, "step": 3815 }, { "epoch": 2.0, "eval_loss": 0.5374822616577148, "eval_runtime": 9.885, "eval_samples_per_second": 85.786, "eval_steps_per_second": 21.447, "num_input_tokens_seen": 2497936, "step": 3816 }, { "epoch": 2.0020964360587, "grad_norm": 0.13086098432540894, "learning_rate": 4.999999905865133e-05, "loss": 0.4828, "num_input_tokens_seen": 2501104, "step": 3820 }, { "epoch": 2.0047169811320753, "grad_norm": 0.2809651494026184, "learning_rate": 4.999999330596526e-05, "loss": 0.6308, "num_input_tokens_seen": 2504624, "step": 3825 }, { "epoch": 2.007337526205451, "grad_norm": 0.2623366117477417, "learning_rate": 4.9999982323565806e-05, "loss": 0.6411, "num_input_tokens_seen": 2507024, "step": 3830 }, { "epoch": 2.009958071278826, "grad_norm": 0.30508658289909363, "learning_rate": 4.999996611145527e-05, "loss": 0.556, "num_input_tokens_seen": 2509808, "step": 3835 }, { "epoch": 2.0125786163522013, "grad_norm": 0.17165575921535492, "learning_rate": 4.9999944669637024e-05, "loss": 0.5264, "num_input_tokens_seen": 2512944, "step": 3840 }, { "epoch": 2.0151991614255764, "grad_norm": 0.16510958969593048, "learning_rate": 4.999991799811557e-05, "loss": 0.5306, "num_input_tokens_seen": 2516496, "step": 3845 }, { "epoch": 2.0178197064989516, "grad_norm": 0.2168509066104889, "learning_rate": 4.99998860968965e-05, "loss": 0.4239, "num_input_tokens_seen": 2519664, "step": 3850 }, { "epoch": 2.020440251572327, "grad_norm": 0.1834137886762619, "learning_rate": 4.999984896598646e-05, "loss": 0.4726, "num_input_tokens_seen": 2522992, "step": 3855 }, { "epoch": 2.0230607966457024, "grad_norm": 0.14069640636444092, "learning_rate": 4.999980660539323e-05, "loss": 0.6379, "num_input_tokens_seen": 2526416, "step": 3860 }, { "epoch": 2.0256813417190775, "grad_norm": 0.28794190287590027, "learning_rate": 4.999975901512567e-05, "loss": 0.5137, "num_input_tokens_seen": 2529200, "step": 3865 }, { "epoch": 2.0283018867924527, "grad_norm": 0.21581949293613434, "learning_rate": 4.999970619519374e-05, "loss": 0.6117, "num_input_tokens_seen": 2532336, "step": 3870 }, { "epoch": 2.030922431865828, "grad_norm": 0.14010371267795563, "learning_rate": 4.999964814560849e-05, "loss": 0.5959, "num_input_tokens_seen": 2536784, "step": 3875 }, { "epoch": 2.0335429769392035, "grad_norm": 0.2234334647655487, "learning_rate": 4.999958486638205e-05, "loss": 0.5511, "num_input_tokens_seen": 2539824, "step": 3880 }, { "epoch": 2.0361635220125787, "grad_norm": 0.13270999491214752, "learning_rate": 4.999951635752767e-05, "loss": 0.6422, "num_input_tokens_seen": 2543664, "step": 3885 }, { "epoch": 2.038784067085954, "grad_norm": 0.1777062863111496, "learning_rate": 4.999944261905968e-05, "loss": 0.5134, "num_input_tokens_seen": 2547632, "step": 3890 }, { "epoch": 2.041404612159329, "grad_norm": 0.1400701403617859, "learning_rate": 4.99993636509935e-05, "loss": 0.5997, "num_input_tokens_seen": 2551312, "step": 3895 }, { "epoch": 2.0440251572327046, "grad_norm": 0.6236603856086731, "learning_rate": 4.999927945334565e-05, "loss": 0.7016, "num_input_tokens_seen": 2554640, "step": 3900 }, { "epoch": 2.04664570230608, "grad_norm": 0.28222858905792236, "learning_rate": 4.9999190026133745e-05, "loss": 0.4345, "num_input_tokens_seen": 2557744, "step": 3905 }, { "epoch": 2.049266247379455, "grad_norm": 0.2000940442085266, "learning_rate": 4.9999095369376506e-05, "loss": 0.5667, "num_input_tokens_seen": 2560816, "step": 3910 }, { "epoch": 2.05188679245283, "grad_norm": 0.30030766129493713, "learning_rate": 4.99989954830937e-05, "loss": 0.6866, "num_input_tokens_seen": 2563376, "step": 3915 }, { "epoch": 2.0545073375262053, "grad_norm": 0.25501978397369385, "learning_rate": 4.9998890367306255e-05, "loss": 0.5095, "num_input_tokens_seen": 2566640, "step": 3920 }, { "epoch": 2.057127882599581, "grad_norm": 0.13933733105659485, "learning_rate": 4.999878002203614e-05, "loss": 0.4974, "num_input_tokens_seen": 2570000, "step": 3925 }, { "epoch": 2.059748427672956, "grad_norm": 0.16438564658164978, "learning_rate": 4.999866444730645e-05, "loss": 0.6062, "num_input_tokens_seen": 2572912, "step": 3930 }, { "epoch": 2.0623689727463312, "grad_norm": 0.1519940346479416, "learning_rate": 4.999854364314136e-05, "loss": 0.5465, "num_input_tokens_seen": 2576208, "step": 3935 }, { "epoch": 2.0649895178197064, "grad_norm": 0.1867966204881668, "learning_rate": 4.999841760956614e-05, "loss": 0.6944, "num_input_tokens_seen": 2580176, "step": 3940 }, { "epoch": 2.0676100628930816, "grad_norm": 0.2625948488712311, "learning_rate": 4.999828634660715e-05, "loss": 0.5281, "num_input_tokens_seen": 2583504, "step": 3945 }, { "epoch": 2.070230607966457, "grad_norm": 0.3069300055503845, "learning_rate": 4.9998149854291856e-05, "loss": 0.6393, "num_input_tokens_seen": 2586928, "step": 3950 }, { "epoch": 2.0728511530398324, "grad_norm": 0.14245550334453583, "learning_rate": 4.99980081326488e-05, "loss": 0.5722, "num_input_tokens_seen": 2589808, "step": 3955 }, { "epoch": 2.0754716981132075, "grad_norm": 0.18039728701114655, "learning_rate": 4.999786118170763e-05, "loss": 0.4737, "num_input_tokens_seen": 2592720, "step": 3960 }, { "epoch": 2.0780922431865827, "grad_norm": 0.18257492780685425, "learning_rate": 4.99977090014991e-05, "loss": 0.6454, "num_input_tokens_seen": 2595056, "step": 3965 }, { "epoch": 2.080712788259958, "grad_norm": 0.37329792976379395, "learning_rate": 4.9997551592055034e-05, "loss": 0.4653, "num_input_tokens_seen": 2598256, "step": 3970 }, { "epoch": 2.0833333333333335, "grad_norm": 0.20323015749454498, "learning_rate": 4.999738895340836e-05, "loss": 0.6497, "num_input_tokens_seen": 2602064, "step": 3975 }, { "epoch": 2.0859538784067087, "grad_norm": 0.16574394702911377, "learning_rate": 4.999722108559311e-05, "loss": 0.6155, "num_input_tokens_seen": 2605232, "step": 3980 }, { "epoch": 2.088574423480084, "grad_norm": 0.14614775776863098, "learning_rate": 4.9997047988644385e-05, "loss": 0.6312, "num_input_tokens_seen": 2608336, "step": 3985 }, { "epoch": 2.091194968553459, "grad_norm": 0.1278282254934311, "learning_rate": 4.99968696625984e-05, "loss": 0.4997, "num_input_tokens_seen": 2611472, "step": 3990 }, { "epoch": 2.0938155136268346, "grad_norm": 0.13896393775939941, "learning_rate": 4.999668610749246e-05, "loss": 0.5191, "num_input_tokens_seen": 2614704, "step": 3995 }, { "epoch": 2.0964360587002098, "grad_norm": 0.1560625582933426, "learning_rate": 4.999649732336498e-05, "loss": 0.5782, "num_input_tokens_seen": 2617328, "step": 4000 }, { "epoch": 2.099056603773585, "grad_norm": 0.15637166798114777, "learning_rate": 4.999630331025542e-05, "loss": 0.4302, "num_input_tokens_seen": 2620656, "step": 4005 }, { "epoch": 2.10167714884696, "grad_norm": 0.21642085909843445, "learning_rate": 4.9996104068204385e-05, "loss": 0.6478, "num_input_tokens_seen": 2624464, "step": 4010 }, { "epoch": 2.1042976939203353, "grad_norm": 0.20042669773101807, "learning_rate": 4.999589959725355e-05, "loss": 0.4545, "num_input_tokens_seen": 2629296, "step": 4015 }, { "epoch": 2.106918238993711, "grad_norm": 0.25912314653396606, "learning_rate": 4.999568989744568e-05, "loss": 0.5431, "num_input_tokens_seen": 2631792, "step": 4020 }, { "epoch": 2.109538784067086, "grad_norm": 0.15030521154403687, "learning_rate": 4.999547496882466e-05, "loss": 0.4666, "num_input_tokens_seen": 2635024, "step": 4025 }, { "epoch": 2.1121593291404612, "grad_norm": 0.15522396564483643, "learning_rate": 4.999525481143543e-05, "loss": 0.4897, "num_input_tokens_seen": 2638800, "step": 4030 }, { "epoch": 2.1147798742138364, "grad_norm": 0.09860366582870483, "learning_rate": 4.999502942532406e-05, "loss": 0.4704, "num_input_tokens_seen": 2642800, "step": 4035 }, { "epoch": 2.1174004192872116, "grad_norm": 0.18715867400169373, "learning_rate": 4.999479881053769e-05, "loss": 0.5032, "num_input_tokens_seen": 2645104, "step": 4040 }, { "epoch": 2.120020964360587, "grad_norm": 0.23163415491580963, "learning_rate": 4.999456296712457e-05, "loss": 0.5329, "num_input_tokens_seen": 2648144, "step": 4045 }, { "epoch": 2.1226415094339623, "grad_norm": 0.21227684617042542, "learning_rate": 4.9994321895134025e-05, "loss": 0.6123, "num_input_tokens_seen": 2651152, "step": 4050 }, { "epoch": 2.1252620545073375, "grad_norm": 0.2167496383190155, "learning_rate": 4.999407559461649e-05, "loss": 0.5407, "num_input_tokens_seen": 2657232, "step": 4055 }, { "epoch": 2.1278825995807127, "grad_norm": 0.15625211596488953, "learning_rate": 4.999382406562349e-05, "loss": 0.4365, "num_input_tokens_seen": 2660112, "step": 4060 }, { "epoch": 2.130503144654088, "grad_norm": 0.2631852328777313, "learning_rate": 4.999356730820765e-05, "loss": 0.8111, "num_input_tokens_seen": 2664400, "step": 4065 }, { "epoch": 2.1331236897274635, "grad_norm": 0.2058153599500656, "learning_rate": 4.999330532242266e-05, "loss": 0.6008, "num_input_tokens_seen": 2667632, "step": 4070 }, { "epoch": 2.1357442348008386, "grad_norm": 0.5059956312179565, "learning_rate": 4.999303810832333e-05, "loss": 0.5172, "num_input_tokens_seen": 2671504, "step": 4075 }, { "epoch": 2.138364779874214, "grad_norm": 0.13830457627773285, "learning_rate": 4.999276566596557e-05, "loss": 0.5166, "num_input_tokens_seen": 2674384, "step": 4080 }, { "epoch": 2.140985324947589, "grad_norm": 0.1458139419555664, "learning_rate": 4.9992487995406364e-05, "loss": 0.5361, "num_input_tokens_seen": 2677392, "step": 4085 }, { "epoch": 2.1436058700209646, "grad_norm": 0.22739483416080475, "learning_rate": 4.9992205096703806e-05, "loss": 0.6904, "num_input_tokens_seen": 2680240, "step": 4090 }, { "epoch": 2.1462264150943398, "grad_norm": 0.13776329159736633, "learning_rate": 4.9991916969917054e-05, "loss": 0.5863, "num_input_tokens_seen": 2683888, "step": 4095 }, { "epoch": 2.148846960167715, "grad_norm": 0.56193608045578, "learning_rate": 4.99916236151064e-05, "loss": 0.6747, "num_input_tokens_seen": 2686352, "step": 4100 }, { "epoch": 2.15146750524109, "grad_norm": 0.21303346753120422, "learning_rate": 4.999132503233321e-05, "loss": 0.6205, "num_input_tokens_seen": 2688560, "step": 4105 }, { "epoch": 2.1540880503144653, "grad_norm": 0.2031724750995636, "learning_rate": 4.999102122165994e-05, "loss": 0.4673, "num_input_tokens_seen": 2691056, "step": 4110 }, { "epoch": 2.156708595387841, "grad_norm": 0.1737007051706314, "learning_rate": 4.9990712183150135e-05, "loss": 0.4702, "num_input_tokens_seen": 2694032, "step": 4115 }, { "epoch": 2.159329140461216, "grad_norm": 0.5315463542938232, "learning_rate": 4.9990397916868456e-05, "loss": 0.6294, "num_input_tokens_seen": 2696944, "step": 4120 }, { "epoch": 2.161949685534591, "grad_norm": 0.26436373591423035, "learning_rate": 4.999007842288064e-05, "loss": 0.6523, "num_input_tokens_seen": 2699632, "step": 4125 }, { "epoch": 2.1645702306079664, "grad_norm": 0.1558794230222702, "learning_rate": 4.9989753701253514e-05, "loss": 0.5152, "num_input_tokens_seen": 2703312, "step": 4130 }, { "epoch": 2.1671907756813416, "grad_norm": 0.20301185548305511, "learning_rate": 4.998942375205502e-05, "loss": 0.5793, "num_input_tokens_seen": 2706928, "step": 4135 }, { "epoch": 2.169811320754717, "grad_norm": 0.25498420000076294, "learning_rate": 4.998908857535416e-05, "loss": 0.6706, "num_input_tokens_seen": 2710992, "step": 4140 }, { "epoch": 2.1724318658280923, "grad_norm": 0.14253626763820648, "learning_rate": 4.9988748171221065e-05, "loss": 0.5372, "num_input_tokens_seen": 2714544, "step": 4145 }, { "epoch": 2.1750524109014675, "grad_norm": 0.2277221381664276, "learning_rate": 4.998840253972695e-05, "loss": 0.7163, "num_input_tokens_seen": 2718064, "step": 4150 }, { "epoch": 2.1776729559748427, "grad_norm": 0.29232344031333923, "learning_rate": 4.9988051680944094e-05, "loss": 0.6856, "num_input_tokens_seen": 2721648, "step": 4155 }, { "epoch": 2.180293501048218, "grad_norm": 0.20303790271282196, "learning_rate": 4.998769559494591e-05, "loss": 0.5406, "num_input_tokens_seen": 2725200, "step": 4160 }, { "epoch": 2.1829140461215935, "grad_norm": 0.16311804950237274, "learning_rate": 4.998733428180689e-05, "loss": 0.5485, "num_input_tokens_seen": 2728432, "step": 4165 }, { "epoch": 2.1855345911949686, "grad_norm": 0.22295813262462616, "learning_rate": 4.998696774160259e-05, "loss": 0.4551, "num_input_tokens_seen": 2732688, "step": 4170 }, { "epoch": 2.188155136268344, "grad_norm": 0.17030946910381317, "learning_rate": 4.9986595974409714e-05, "loss": 0.606, "num_input_tokens_seen": 2738448, "step": 4175 }, { "epoch": 2.190775681341719, "grad_norm": 0.20434220135211945, "learning_rate": 4.998621898030603e-05, "loss": 0.516, "num_input_tokens_seen": 2741552, "step": 4180 }, { "epoch": 2.1933962264150946, "grad_norm": 0.2004912793636322, "learning_rate": 4.998583675937039e-05, "loss": 0.5495, "num_input_tokens_seen": 2744816, "step": 4185 }, { "epoch": 2.1960167714884697, "grad_norm": 0.1919853836297989, "learning_rate": 4.998544931168274e-05, "loss": 0.6517, "num_input_tokens_seen": 2748912, "step": 4190 }, { "epoch": 2.198637316561845, "grad_norm": 0.22539077699184418, "learning_rate": 4.9985056637324154e-05, "loss": 0.6186, "num_input_tokens_seen": 2752240, "step": 4195 }, { "epoch": 2.20125786163522, "grad_norm": 0.18293409049510956, "learning_rate": 4.998465873637677e-05, "loss": 0.5871, "num_input_tokens_seen": 2755216, "step": 4200 }, { "epoch": 2.2038784067085953, "grad_norm": 0.27540019154548645, "learning_rate": 4.9984255608923806e-05, "loss": 0.6554, "num_input_tokens_seen": 2758224, "step": 4205 }, { "epoch": 2.2064989517819704, "grad_norm": 0.13564831018447876, "learning_rate": 4.998384725504961e-05, "loss": 0.6825, "num_input_tokens_seen": 2761680, "step": 4210 }, { "epoch": 2.209119496855346, "grad_norm": 0.14327743649482727, "learning_rate": 4.9983433674839596e-05, "loss": 0.5707, "num_input_tokens_seen": 2764464, "step": 4215 }, { "epoch": 2.211740041928721, "grad_norm": 0.17653632164001465, "learning_rate": 4.998301486838028e-05, "loss": 0.5695, "num_input_tokens_seen": 2767600, "step": 4220 }, { "epoch": 2.2143605870020964, "grad_norm": 0.2342243790626526, "learning_rate": 4.998259083575929e-05, "loss": 0.5402, "num_input_tokens_seen": 2770928, "step": 4225 }, { "epoch": 2.2169811320754715, "grad_norm": 0.15445928275585175, "learning_rate": 4.99821615770653e-05, "loss": 0.4874, "num_input_tokens_seen": 2774544, "step": 4230 }, { "epoch": 2.219601677148847, "grad_norm": 0.23813143372535706, "learning_rate": 4.998172709238813e-05, "loss": 0.6531, "num_input_tokens_seen": 2777392, "step": 4235 }, { "epoch": 2.2222222222222223, "grad_norm": 0.13173402845859528, "learning_rate": 4.9981287381818654e-05, "loss": 0.6188, "num_input_tokens_seen": 2780272, "step": 4240 }, { "epoch": 2.2248427672955975, "grad_norm": 0.2701840102672577, "learning_rate": 4.9980842445448856e-05, "loss": 0.5845, "num_input_tokens_seen": 2783824, "step": 4245 }, { "epoch": 2.2274633123689727, "grad_norm": 0.20214858651161194, "learning_rate": 4.9980392283371815e-05, "loss": 0.5586, "num_input_tokens_seen": 2786992, "step": 4250 }, { "epoch": 2.230083857442348, "grad_norm": 0.11600081622600555, "learning_rate": 4.99799368956817e-05, "loss": 0.4445, "num_input_tokens_seen": 2789840, "step": 4255 }, { "epoch": 2.2327044025157234, "grad_norm": 0.2936851680278778, "learning_rate": 4.997947628247378e-05, "loss": 0.5065, "num_input_tokens_seen": 2792432, "step": 4260 }, { "epoch": 2.2353249475890986, "grad_norm": 0.26479193568229675, "learning_rate": 4.9979010443844395e-05, "loss": 0.6466, "num_input_tokens_seen": 2795056, "step": 4265 }, { "epoch": 2.237945492662474, "grad_norm": 0.1779606193304062, "learning_rate": 4.9978539379891e-05, "loss": 0.586, "num_input_tokens_seen": 2798416, "step": 4270 }, { "epoch": 2.240566037735849, "grad_norm": 0.18693752586841583, "learning_rate": 4.997806309071214e-05, "loss": 0.479, "num_input_tokens_seen": 2801040, "step": 4275 }, { "epoch": 2.243186582809224, "grad_norm": 0.19342748820781708, "learning_rate": 4.997758157640745e-05, "loss": 0.5375, "num_input_tokens_seen": 2804880, "step": 4280 }, { "epoch": 2.2458071278825997, "grad_norm": 0.24772149324417114, "learning_rate": 4.997709483707765e-05, "loss": 0.4992, "num_input_tokens_seen": 2810864, "step": 4285 }, { "epoch": 2.248427672955975, "grad_norm": 0.35823798179626465, "learning_rate": 4.997660287282456e-05, "loss": 0.562, "num_input_tokens_seen": 2813328, "step": 4290 }, { "epoch": 2.25104821802935, "grad_norm": 0.19866208732128143, "learning_rate": 4.99761056837511e-05, "loss": 0.5845, "num_input_tokens_seen": 2816016, "step": 4295 }, { "epoch": 2.2536687631027252, "grad_norm": 0.15719416737556458, "learning_rate": 4.997560326996128e-05, "loss": 0.6861, "num_input_tokens_seen": 2818512, "step": 4300 }, { "epoch": 2.2562893081761004, "grad_norm": 0.15976253151893616, "learning_rate": 4.997509563156018e-05, "loss": 0.7202, "num_input_tokens_seen": 2821776, "step": 4305 }, { "epoch": 2.258909853249476, "grad_norm": 0.24739189445972443, "learning_rate": 4.997458276865401e-05, "loss": 0.5116, "num_input_tokens_seen": 2825232, "step": 4310 }, { "epoch": 2.261530398322851, "grad_norm": 0.22648650407791138, "learning_rate": 4.997406468135005e-05, "loss": 0.6034, "num_input_tokens_seen": 2828592, "step": 4315 }, { "epoch": 2.2641509433962264, "grad_norm": 0.22106531262397766, "learning_rate": 4.9973541369756686e-05, "loss": 0.5852, "num_input_tokens_seen": 2831472, "step": 4320 }, { "epoch": 2.2667714884696015, "grad_norm": 0.12407079339027405, "learning_rate": 4.9973012833983365e-05, "loss": 0.4376, "num_input_tokens_seen": 2834640, "step": 4325 }, { "epoch": 2.269392033542977, "grad_norm": 0.23743316531181335, "learning_rate": 4.9972479074140684e-05, "loss": 0.4326, "num_input_tokens_seen": 2838160, "step": 4330 }, { "epoch": 2.2720125786163523, "grad_norm": 0.17704394459724426, "learning_rate": 4.9971940090340276e-05, "loss": 0.3804, "num_input_tokens_seen": 2842160, "step": 4335 }, { "epoch": 2.2746331236897275, "grad_norm": 0.21844340860843658, "learning_rate": 4.997139588269489e-05, "loss": 0.4767, "num_input_tokens_seen": 2845712, "step": 4340 }, { "epoch": 2.2772536687631026, "grad_norm": 0.16022326052188873, "learning_rate": 4.997084645131838e-05, "loss": 0.5343, "num_input_tokens_seen": 2848976, "step": 4345 }, { "epoch": 2.279874213836478, "grad_norm": 0.22903423011302948, "learning_rate": 4.9970291796325666e-05, "loss": 0.4669, "num_input_tokens_seen": 2852304, "step": 4350 }, { "epoch": 2.2824947589098534, "grad_norm": 0.18783067166805267, "learning_rate": 4.996973191783279e-05, "loss": 0.7802, "num_input_tokens_seen": 2855152, "step": 4355 }, { "epoch": 2.2851153039832286, "grad_norm": 0.22177651524543762, "learning_rate": 4.996916681595687e-05, "loss": 0.5651, "num_input_tokens_seen": 2857968, "step": 4360 }, { "epoch": 2.2877358490566038, "grad_norm": 0.16380523145198822, "learning_rate": 4.996859649081611e-05, "loss": 0.4743, "num_input_tokens_seen": 2862032, "step": 4365 }, { "epoch": 2.290356394129979, "grad_norm": 0.17763295769691467, "learning_rate": 4.996802094252983e-05, "loss": 0.59, "num_input_tokens_seen": 2864624, "step": 4370 }, { "epoch": 2.2929769392033545, "grad_norm": 0.31996774673461914, "learning_rate": 4.996744017121841e-05, "loss": 0.6069, "num_input_tokens_seen": 2868208, "step": 4375 }, { "epoch": 2.2955974842767297, "grad_norm": 0.14829476177692413, "learning_rate": 4.9966854177003355e-05, "loss": 0.7635, "num_input_tokens_seen": 2871056, "step": 4380 }, { "epoch": 2.298218029350105, "grad_norm": 0.2692261338233948, "learning_rate": 4.996626296000724e-05, "loss": 0.5695, "num_input_tokens_seen": 2873904, "step": 4385 }, { "epoch": 2.30083857442348, "grad_norm": 0.17317479848861694, "learning_rate": 4.996566652035374e-05, "loss": 0.5405, "num_input_tokens_seen": 2877040, "step": 4390 }, { "epoch": 2.3034591194968552, "grad_norm": 0.20414823293685913, "learning_rate": 4.996506485816763e-05, "loss": 0.6494, "num_input_tokens_seen": 2880560, "step": 4395 }, { "epoch": 2.3060796645702304, "grad_norm": 0.25176388025283813, "learning_rate": 4.996445797357477e-05, "loss": 0.5933, "num_input_tokens_seen": 2883920, "step": 4400 }, { "epoch": 2.308700209643606, "grad_norm": 0.23262138664722443, "learning_rate": 4.996384586670211e-05, "loss": 0.5432, "num_input_tokens_seen": 2886448, "step": 4405 }, { "epoch": 2.311320754716981, "grad_norm": 0.33206412196159363, "learning_rate": 4.99632285376777e-05, "loss": 0.529, "num_input_tokens_seen": 2889520, "step": 4410 }, { "epoch": 2.3139412997903563, "grad_norm": 0.12054730206727982, "learning_rate": 4.9962605986630674e-05, "loss": 0.5596, "num_input_tokens_seen": 2893040, "step": 4415 }, { "epoch": 2.3165618448637315, "grad_norm": 0.10672920942306519, "learning_rate": 4.996197821369126e-05, "loss": 0.4387, "num_input_tokens_seen": 2896880, "step": 4420 }, { "epoch": 2.319182389937107, "grad_norm": 0.22946777939796448, "learning_rate": 4.9961345218990796e-05, "loss": 0.4626, "num_input_tokens_seen": 2899600, "step": 4425 }, { "epoch": 2.3218029350104823, "grad_norm": 0.13585729897022247, "learning_rate": 4.996070700266168e-05, "loss": 0.5063, "num_input_tokens_seen": 2902608, "step": 4430 }, { "epoch": 2.3244234800838575, "grad_norm": 0.10406824946403503, "learning_rate": 4.9960063564837424e-05, "loss": 0.5728, "num_input_tokens_seen": 2906224, "step": 4435 }, { "epoch": 2.3270440251572326, "grad_norm": 0.13714540004730225, "learning_rate": 4.995941490565262e-05, "loss": 0.472, "num_input_tokens_seen": 2909232, "step": 4440 }, { "epoch": 2.329664570230608, "grad_norm": 0.3230399191379547, "learning_rate": 4.995876102524299e-05, "loss": 0.5056, "num_input_tokens_seen": 2916848, "step": 4445 }, { "epoch": 2.3322851153039834, "grad_norm": 0.09846612066030502, "learning_rate": 4.995810192374529e-05, "loss": 0.6203, "num_input_tokens_seen": 2920560, "step": 4450 }, { "epoch": 2.3349056603773586, "grad_norm": 0.1781528741121292, "learning_rate": 4.99574376012974e-05, "loss": 0.6121, "num_input_tokens_seen": 2923408, "step": 4455 }, { "epoch": 2.3375262054507338, "grad_norm": 0.15243582427501678, "learning_rate": 4.995676805803829e-05, "loss": 0.5367, "num_input_tokens_seen": 2927280, "step": 4460 }, { "epoch": 2.340146750524109, "grad_norm": 0.21580342948436737, "learning_rate": 4.995609329410804e-05, "loss": 0.7355, "num_input_tokens_seen": 2930800, "step": 4465 }, { "epoch": 2.342767295597484, "grad_norm": 0.21518898010253906, "learning_rate": 4.9955413309647776e-05, "loss": 0.5047, "num_input_tokens_seen": 2934448, "step": 4470 }, { "epoch": 2.3453878406708597, "grad_norm": 0.254954993724823, "learning_rate": 4.9954728104799756e-05, "loss": 0.747, "num_input_tokens_seen": 2937104, "step": 4475 }, { "epoch": 2.348008385744235, "grad_norm": 0.12656615674495697, "learning_rate": 4.995403767970731e-05, "loss": 0.5243, "num_input_tokens_seen": 2940368, "step": 4480 }, { "epoch": 2.35062893081761, "grad_norm": 0.22941888868808746, "learning_rate": 4.995334203451488e-05, "loss": 0.4766, "num_input_tokens_seen": 2942960, "step": 4485 }, { "epoch": 2.353249475890985, "grad_norm": 0.5842645764350891, "learning_rate": 4.995264116936798e-05, "loss": 0.618, "num_input_tokens_seen": 2946704, "step": 4490 }, { "epoch": 2.3558700209643604, "grad_norm": 0.3936944007873535, "learning_rate": 4.995193508441322e-05, "loss": 0.4689, "num_input_tokens_seen": 2950608, "step": 4495 }, { "epoch": 2.358490566037736, "grad_norm": 0.16511759161949158, "learning_rate": 4.99512237797983e-05, "loss": 0.5213, "num_input_tokens_seen": 2954416, "step": 4500 }, { "epoch": 2.361111111111111, "grad_norm": 0.14974555373191833, "learning_rate": 4.995050725567203e-05, "loss": 0.6123, "num_input_tokens_seen": 2957424, "step": 4505 }, { "epoch": 2.3637316561844863, "grad_norm": 0.30041229724884033, "learning_rate": 4.994978551218429e-05, "loss": 0.5933, "num_input_tokens_seen": 2960400, "step": 4510 }, { "epoch": 2.3663522012578615, "grad_norm": 0.2129075974225998, "learning_rate": 4.9949058549486066e-05, "loss": 0.6029, "num_input_tokens_seen": 2963568, "step": 4515 }, { "epoch": 2.368972746331237, "grad_norm": 0.16783744096755981, "learning_rate": 4.994832636772943e-05, "loss": 0.7527, "num_input_tokens_seen": 2966384, "step": 4520 }, { "epoch": 2.3715932914046123, "grad_norm": 0.14484146237373352, "learning_rate": 4.994758896706754e-05, "loss": 0.396, "num_input_tokens_seen": 2969360, "step": 4525 }, { "epoch": 2.3742138364779874, "grad_norm": 0.16276223957538605, "learning_rate": 4.994684634765465e-05, "loss": 0.712, "num_input_tokens_seen": 2972560, "step": 4530 }, { "epoch": 2.3768343815513626, "grad_norm": 0.2728906571865082, "learning_rate": 4.9946098509646125e-05, "loss": 0.7124, "num_input_tokens_seen": 2975472, "step": 4535 }, { "epoch": 2.379454926624738, "grad_norm": 0.16985811293125153, "learning_rate": 4.994534545319838e-05, "loss": 0.6078, "num_input_tokens_seen": 2978288, "step": 4540 }, { "epoch": 2.3820754716981134, "grad_norm": 0.1554822325706482, "learning_rate": 4.994458717846896e-05, "loss": 0.6455, "num_input_tokens_seen": 2981840, "step": 4545 }, { "epoch": 2.3846960167714886, "grad_norm": 0.19884051382541656, "learning_rate": 4.9943823685616485e-05, "loss": 0.5673, "num_input_tokens_seen": 2985104, "step": 4550 }, { "epoch": 2.3873165618448637, "grad_norm": 0.14943942427635193, "learning_rate": 4.994305497480067e-05, "loss": 0.5692, "num_input_tokens_seen": 2989360, "step": 4555 }, { "epoch": 2.389937106918239, "grad_norm": 0.18387967348098755, "learning_rate": 4.994228104618233e-05, "loss": 0.5464, "num_input_tokens_seen": 2993136, "step": 4560 }, { "epoch": 2.392557651991614, "grad_norm": 0.18045680224895477, "learning_rate": 4.994150189992334e-05, "loss": 0.5989, "num_input_tokens_seen": 2996848, "step": 4565 }, { "epoch": 2.3951781970649897, "grad_norm": 0.23925292491912842, "learning_rate": 4.99407175361867e-05, "loss": 0.5845, "num_input_tokens_seen": 2999952, "step": 4570 }, { "epoch": 2.397798742138365, "grad_norm": 0.13225212693214417, "learning_rate": 4.99399279551365e-05, "loss": 0.6234, "num_input_tokens_seen": 3002896, "step": 4575 }, { "epoch": 2.40041928721174, "grad_norm": 0.16064894199371338, "learning_rate": 4.9939133156937886e-05, "loss": 0.5405, "num_input_tokens_seen": 3005936, "step": 4580 }, { "epoch": 2.403039832285115, "grad_norm": 0.15873406827449799, "learning_rate": 4.993833314175716e-05, "loss": 0.4653, "num_input_tokens_seen": 3009488, "step": 4585 }, { "epoch": 2.4056603773584904, "grad_norm": 0.2606850862503052, "learning_rate": 4.993752790976163e-05, "loss": 0.5389, "num_input_tokens_seen": 3012560, "step": 4590 }, { "epoch": 2.408280922431866, "grad_norm": 0.14862705767154694, "learning_rate": 4.993671746111977e-05, "loss": 0.4396, "num_input_tokens_seen": 3015440, "step": 4595 }, { "epoch": 2.410901467505241, "grad_norm": 0.16494037210941315, "learning_rate": 4.993590179600111e-05, "loss": 0.6918, "num_input_tokens_seen": 3019088, "step": 4600 }, { "epoch": 2.4135220125786163, "grad_norm": 0.3325177729129791, "learning_rate": 4.9935080914576284e-05, "loss": 0.5304, "num_input_tokens_seen": 3022384, "step": 4605 }, { "epoch": 2.4161425576519915, "grad_norm": 0.13783885538578033, "learning_rate": 4.9934254817016995e-05, "loss": 0.4216, "num_input_tokens_seen": 3026672, "step": 4610 }, { "epoch": 2.418763102725367, "grad_norm": 0.16845177114009857, "learning_rate": 4.993342350349607e-05, "loss": 0.5416, "num_input_tokens_seen": 3029936, "step": 4615 }, { "epoch": 2.4213836477987423, "grad_norm": 0.17050932347774506, "learning_rate": 4.9932586974187404e-05, "loss": 0.7436, "num_input_tokens_seen": 3032752, "step": 4620 }, { "epoch": 2.4240041928721174, "grad_norm": 0.342423677444458, "learning_rate": 4.9931745229265986e-05, "loss": 0.6066, "num_input_tokens_seen": 3036400, "step": 4625 }, { "epoch": 2.4266247379454926, "grad_norm": 0.1857626885175705, "learning_rate": 4.99308982689079e-05, "loss": 0.559, "num_input_tokens_seen": 3039152, "step": 4630 }, { "epoch": 2.4292452830188678, "grad_norm": 0.11911116540431976, "learning_rate": 4.993004609329033e-05, "loss": 0.4389, "num_input_tokens_seen": 3042096, "step": 4635 }, { "epoch": 2.431865828092243, "grad_norm": 0.1707780361175537, "learning_rate": 4.992918870259153e-05, "loss": 0.3741, "num_input_tokens_seen": 3045136, "step": 4640 }, { "epoch": 2.4344863731656186, "grad_norm": 0.2792665958404541, "learning_rate": 4.992832609699086e-05, "loss": 0.5146, "num_input_tokens_seen": 3047344, "step": 4645 }, { "epoch": 2.4371069182389937, "grad_norm": 0.15304146707057953, "learning_rate": 4.992745827666877e-05, "loss": 0.5748, "num_input_tokens_seen": 3049968, "step": 4650 }, { "epoch": 2.439727463312369, "grad_norm": 0.1608443260192871, "learning_rate": 4.99265852418068e-05, "loss": 0.6211, "num_input_tokens_seen": 3052784, "step": 4655 }, { "epoch": 2.442348008385744, "grad_norm": 0.10021214187145233, "learning_rate": 4.992570699258757e-05, "loss": 0.5347, "num_input_tokens_seen": 3056752, "step": 4660 }, { "epoch": 2.4449685534591197, "grad_norm": 0.14150433242321014, "learning_rate": 4.99248235291948e-05, "loss": 0.5033, "num_input_tokens_seen": 3059696, "step": 4665 }, { "epoch": 2.447589098532495, "grad_norm": 0.27907711267471313, "learning_rate": 4.9923934851813314e-05, "loss": 0.5036, "num_input_tokens_seen": 3062608, "step": 4670 }, { "epoch": 2.45020964360587, "grad_norm": 0.12568055093288422, "learning_rate": 4.9923040960628996e-05, "loss": 0.524, "num_input_tokens_seen": 3065712, "step": 4675 }, { "epoch": 2.452830188679245, "grad_norm": 0.11141189187765121, "learning_rate": 4.9922141855828856e-05, "loss": 0.674, "num_input_tokens_seen": 3069936, "step": 4680 }, { "epoch": 2.4554507337526204, "grad_norm": 0.17689824104309082, "learning_rate": 4.992123753760096e-05, "loss": 0.5066, "num_input_tokens_seen": 3073072, "step": 4685 }, { "epoch": 2.458071278825996, "grad_norm": 0.10977219045162201, "learning_rate": 4.9920328006134484e-05, "loss": 0.5563, "num_input_tokens_seen": 3076368, "step": 4690 }, { "epoch": 2.460691823899371, "grad_norm": 0.3411492705345154, "learning_rate": 4.991941326161971e-05, "loss": 0.5597, "num_input_tokens_seen": 3079024, "step": 4695 }, { "epoch": 2.4633123689727463, "grad_norm": 0.11794164031744003, "learning_rate": 4.991849330424796e-05, "loss": 0.5197, "num_input_tokens_seen": 3085616, "step": 4700 }, { "epoch": 2.4659329140461215, "grad_norm": 0.23346975445747375, "learning_rate": 4.9917568134211703e-05, "loss": 0.4816, "num_input_tokens_seen": 3088720, "step": 4705 }, { "epoch": 2.468553459119497, "grad_norm": 0.2236947864294052, "learning_rate": 4.991663775170447e-05, "loss": 0.701, "num_input_tokens_seen": 3091632, "step": 4710 }, { "epoch": 2.4711740041928723, "grad_norm": 0.132043719291687, "learning_rate": 4.991570215692088e-05, "loss": 0.4949, "num_input_tokens_seen": 3094960, "step": 4715 }, { "epoch": 2.4737945492662474, "grad_norm": 0.33724525570869446, "learning_rate": 4.991476135005665e-05, "loss": 0.5954, "num_input_tokens_seen": 3097744, "step": 4720 }, { "epoch": 2.4764150943396226, "grad_norm": 0.15162405371665955, "learning_rate": 4.991381533130859e-05, "loss": 0.5902, "num_input_tokens_seen": 3100656, "step": 4725 }, { "epoch": 2.4790356394129978, "grad_norm": 0.2361920177936554, "learning_rate": 4.99128641008746e-05, "loss": 0.6146, "num_input_tokens_seen": 3103632, "step": 4730 }, { "epoch": 2.481656184486373, "grad_norm": 0.1788526326417923, "learning_rate": 4.9911907658953664e-05, "loss": 0.6369, "num_input_tokens_seen": 3107216, "step": 4735 }, { "epoch": 2.4842767295597485, "grad_norm": 0.3506932854652405, "learning_rate": 4.991094600574584e-05, "loss": 0.7225, "num_input_tokens_seen": 3109680, "step": 4740 }, { "epoch": 2.4868972746331237, "grad_norm": 0.15994901955127716, "learning_rate": 4.9909979141452324e-05, "loss": 0.4353, "num_input_tokens_seen": 3112624, "step": 4745 }, { "epoch": 2.489517819706499, "grad_norm": 0.20392894744873047, "learning_rate": 4.9909007066275346e-05, "loss": 0.6286, "num_input_tokens_seen": 3116048, "step": 4750 }, { "epoch": 2.492138364779874, "grad_norm": 0.22202695906162262, "learning_rate": 4.990802978041828e-05, "loss": 0.615, "num_input_tokens_seen": 3118864, "step": 4755 }, { "epoch": 2.4947589098532497, "grad_norm": 0.17285925149917603, "learning_rate": 4.990704728408555e-05, "loss": 0.5817, "num_input_tokens_seen": 3122192, "step": 4760 }, { "epoch": 2.497379454926625, "grad_norm": 0.14082235097885132, "learning_rate": 4.990605957748268e-05, "loss": 0.4201, "num_input_tokens_seen": 3128016, "step": 4765 }, { "epoch": 2.5, "grad_norm": 0.6697519421577454, "learning_rate": 4.990506666081628e-05, "loss": 0.52, "num_input_tokens_seen": 3130736, "step": 4770 }, { "epoch": 2.502620545073375, "grad_norm": 0.19741064310073853, "learning_rate": 4.990406853429408e-05, "loss": 0.5865, "num_input_tokens_seen": 3133616, "step": 4775 }, { "epoch": 2.5052410901467503, "grad_norm": 0.3677177131175995, "learning_rate": 4.990306519812485e-05, "loss": 0.6622, "num_input_tokens_seen": 3136848, "step": 4780 }, { "epoch": 2.507861635220126, "grad_norm": 0.21049603819847107, "learning_rate": 4.99020566525185e-05, "loss": 0.5876, "num_input_tokens_seen": 3139600, "step": 4785 }, { "epoch": 2.510482180293501, "grad_norm": 0.1921415776014328, "learning_rate": 4.9901042897685996e-05, "loss": 0.5583, "num_input_tokens_seen": 3142896, "step": 4790 }, { "epoch": 2.5131027253668763, "grad_norm": 0.12551099061965942, "learning_rate": 4.99000239338394e-05, "loss": 0.4635, "num_input_tokens_seen": 3147248, "step": 4795 }, { "epoch": 2.5157232704402515, "grad_norm": 0.20350982248783112, "learning_rate": 4.989899976119187e-05, "loss": 0.5521, "num_input_tokens_seen": 3150288, "step": 4800 }, { "epoch": 2.518343815513627, "grad_norm": 0.08696465939283371, "learning_rate": 4.989797037995766e-05, "loss": 0.5227, "num_input_tokens_seen": 3153808, "step": 4805 }, { "epoch": 2.5209643605870022, "grad_norm": 0.16912715137004852, "learning_rate": 4.989693579035209e-05, "loss": 0.6446, "num_input_tokens_seen": 3156784, "step": 4810 }, { "epoch": 2.5235849056603774, "grad_norm": 0.22655139863491058, "learning_rate": 4.989589599259159e-05, "loss": 0.6384, "num_input_tokens_seen": 3160176, "step": 4815 }, { "epoch": 2.5262054507337526, "grad_norm": 0.4252983331680298, "learning_rate": 4.9894850986893685e-05, "loss": 0.6566, "num_input_tokens_seen": 3162608, "step": 4820 }, { "epoch": 2.5288259958071277, "grad_norm": 0.1409195065498352, "learning_rate": 4.9893800773476965e-05, "loss": 0.5906, "num_input_tokens_seen": 3166576, "step": 4825 }, { "epoch": 2.531446540880503, "grad_norm": 0.19560925662517548, "learning_rate": 4.9892745352561134e-05, "loss": 0.4987, "num_input_tokens_seen": 3169552, "step": 4830 }, { "epoch": 2.5340670859538785, "grad_norm": 0.08912453055381775, "learning_rate": 4.9891684724366955e-05, "loss": 0.5292, "num_input_tokens_seen": 3173616, "step": 4835 }, { "epoch": 2.5366876310272537, "grad_norm": 0.14760558307170868, "learning_rate": 4.989061888911631e-05, "loss": 0.5096, "num_input_tokens_seen": 3176816, "step": 4840 }, { "epoch": 2.539308176100629, "grad_norm": 0.17017915844917297, "learning_rate": 4.988954784703217e-05, "loss": 0.536, "num_input_tokens_seen": 3179824, "step": 4845 }, { "epoch": 2.541928721174004, "grad_norm": 0.16363470256328583, "learning_rate": 4.9888471598338574e-05, "loss": 0.4067, "num_input_tokens_seen": 3183056, "step": 4850 }, { "epoch": 2.5445492662473796, "grad_norm": 0.16528671979904175, "learning_rate": 4.988739014326066e-05, "loss": 0.4805, "num_input_tokens_seen": 3187216, "step": 4855 }, { "epoch": 2.547169811320755, "grad_norm": 0.1658063679933548, "learning_rate": 4.9886303482024656e-05, "loss": 0.5048, "num_input_tokens_seen": 3190576, "step": 4860 }, { "epoch": 2.54979035639413, "grad_norm": 0.16162027418613434, "learning_rate": 4.988521161485789e-05, "loss": 0.6379, "num_input_tokens_seen": 3194000, "step": 4865 }, { "epoch": 2.552410901467505, "grad_norm": 0.10817413777112961, "learning_rate": 4.9884114541988747e-05, "loss": 0.4687, "num_input_tokens_seen": 3197232, "step": 4870 }, { "epoch": 2.5550314465408803, "grad_norm": 0.22008657455444336, "learning_rate": 4.988301226364674e-05, "loss": 0.493, "num_input_tokens_seen": 3200592, "step": 4875 }, { "epoch": 2.5576519916142555, "grad_norm": 0.16919411718845367, "learning_rate": 4.988190478006245e-05, "loss": 0.6771, "num_input_tokens_seen": 3204240, "step": 4880 }, { "epoch": 2.560272536687631, "grad_norm": 0.15217538177967072, "learning_rate": 4.988079209146754e-05, "loss": 0.4803, "num_input_tokens_seen": 3207216, "step": 4885 }, { "epoch": 2.5628930817610063, "grad_norm": 0.15952779352664948, "learning_rate": 4.9879674198094794e-05, "loss": 0.5754, "num_input_tokens_seen": 3210096, "step": 4890 }, { "epoch": 2.5655136268343814, "grad_norm": 0.30469608306884766, "learning_rate": 4.987855110017804e-05, "loss": 0.7259, "num_input_tokens_seen": 3213104, "step": 4895 }, { "epoch": 2.568134171907757, "grad_norm": 0.1402035802602768, "learning_rate": 4.9877422797952224e-05, "loss": 0.5132, "num_input_tokens_seen": 3216432, "step": 4900 }, { "epoch": 2.5707547169811322, "grad_norm": 0.1277998685836792, "learning_rate": 4.987628929165338e-05, "loss": 0.6007, "num_input_tokens_seen": 3219696, "step": 4905 }, { "epoch": 2.5733752620545074, "grad_norm": 0.2665533423423767, "learning_rate": 4.9875150581518615e-05, "loss": 0.5902, "num_input_tokens_seen": 3222448, "step": 4910 }, { "epoch": 2.5759958071278826, "grad_norm": 0.09823666512966156, "learning_rate": 4.987400666778614e-05, "loss": 0.5099, "num_input_tokens_seen": 3226128, "step": 4915 }, { "epoch": 2.5786163522012577, "grad_norm": 0.43018895387649536, "learning_rate": 4.987285755069524e-05, "loss": 0.7971, "num_input_tokens_seen": 3229232, "step": 4920 }, { "epoch": 2.581236897274633, "grad_norm": 0.13372404873371124, "learning_rate": 4.987170323048632e-05, "loss": 0.5706, "num_input_tokens_seen": 3232048, "step": 4925 }, { "epoch": 2.5838574423480085, "grad_norm": 0.28916195034980774, "learning_rate": 4.987054370740083e-05, "loss": 0.7094, "num_input_tokens_seen": 3235120, "step": 4930 }, { "epoch": 2.5864779874213837, "grad_norm": 0.3967157006263733, "learning_rate": 4.986937898168134e-05, "loss": 0.5419, "num_input_tokens_seen": 3238224, "step": 4935 }, { "epoch": 2.589098532494759, "grad_norm": 0.2137679159641266, "learning_rate": 4.986820905357148e-05, "loss": 0.6154, "num_input_tokens_seen": 3241744, "step": 4940 }, { "epoch": 2.591719077568134, "grad_norm": 0.3887445628643036, "learning_rate": 4.9867033923315995e-05, "loss": 0.6335, "num_input_tokens_seen": 3244528, "step": 4945 }, { "epoch": 2.5943396226415096, "grad_norm": 0.20718608796596527, "learning_rate": 4.9865853591160724e-05, "loss": 0.625, "num_input_tokens_seen": 3247952, "step": 4950 }, { "epoch": 2.596960167714885, "grad_norm": 0.12409727275371552, "learning_rate": 4.9864668057352556e-05, "loss": 0.6561, "num_input_tokens_seen": 3251120, "step": 4955 }, { "epoch": 2.59958071278826, "grad_norm": 0.33868515491485596, "learning_rate": 4.9863477322139517e-05, "loss": 0.7112, "num_input_tokens_seen": 3255408, "step": 4960 }, { "epoch": 2.602201257861635, "grad_norm": 0.12937267124652863, "learning_rate": 4.986228138577067e-05, "loss": 0.4968, "num_input_tokens_seen": 3258832, "step": 4965 }, { "epoch": 2.6048218029350103, "grad_norm": 0.2918730676174164, "learning_rate": 4.986108024849619e-05, "loss": 0.5806, "num_input_tokens_seen": 3261072, "step": 4970 }, { "epoch": 2.6074423480083855, "grad_norm": 0.330819308757782, "learning_rate": 4.985987391056737e-05, "loss": 0.6759, "num_input_tokens_seen": 3263760, "step": 4975 }, { "epoch": 2.610062893081761, "grad_norm": 0.22892169654369354, "learning_rate": 4.9858662372236544e-05, "loss": 0.6599, "num_input_tokens_seen": 3266192, "step": 4980 }, { "epoch": 2.6126834381551363, "grad_norm": 0.3367346525192261, "learning_rate": 4.985744563375714e-05, "loss": 0.5458, "num_input_tokens_seen": 3268656, "step": 4985 }, { "epoch": 2.6153039832285114, "grad_norm": 0.2144167423248291, "learning_rate": 4.985622369538371e-05, "loss": 0.5704, "num_input_tokens_seen": 3272112, "step": 4990 }, { "epoch": 2.617924528301887, "grad_norm": 0.27761608362197876, "learning_rate": 4.985499655737185e-05, "loss": 0.6216, "num_input_tokens_seen": 3275056, "step": 4995 }, { "epoch": 2.620545073375262, "grad_norm": 0.21549130976200104, "learning_rate": 4.985376421997827e-05, "loss": 0.4678, "num_input_tokens_seen": 3278416, "step": 5000 }, { "epoch": 2.6231656184486374, "grad_norm": 0.31103479862213135, "learning_rate": 4.985252668346076e-05, "loss": 0.5471, "num_input_tokens_seen": 3281616, "step": 5005 }, { "epoch": 2.6257861635220126, "grad_norm": 0.14042524993419647, "learning_rate": 4.9851283948078206e-05, "loss": 0.566, "num_input_tokens_seen": 3285008, "step": 5010 }, { "epoch": 2.6284067085953877, "grad_norm": 0.2879348695278168, "learning_rate": 4.985003601409057e-05, "loss": 0.5557, "num_input_tokens_seen": 3288112, "step": 5015 }, { "epoch": 2.631027253668763, "grad_norm": 0.7752573490142822, "learning_rate": 4.9848782881758906e-05, "loss": 0.6524, "num_input_tokens_seen": 3291408, "step": 5020 }, { "epoch": 2.6336477987421385, "grad_norm": 0.15874098241329193, "learning_rate": 4.9847524551345345e-05, "loss": 0.5823, "num_input_tokens_seen": 3294128, "step": 5025 }, { "epoch": 2.6362683438155137, "grad_norm": 0.3083614110946655, "learning_rate": 4.984626102311313e-05, "loss": 0.5108, "num_input_tokens_seen": 3296688, "step": 5030 }, { "epoch": 2.638888888888889, "grad_norm": 0.26682689785957336, "learning_rate": 4.984499229732656e-05, "loss": 0.363, "num_input_tokens_seen": 3300528, "step": 5035 }, { "epoch": 2.641509433962264, "grad_norm": 0.17870168387889862, "learning_rate": 4.984371837425106e-05, "loss": 0.5452, "num_input_tokens_seen": 3304368, "step": 5040 }, { "epoch": 2.6441299790356396, "grad_norm": 0.32515397667884827, "learning_rate": 4.9842439254153106e-05, "loss": 0.8335, "num_input_tokens_seen": 3307504, "step": 5045 }, { "epoch": 2.646750524109015, "grad_norm": 0.1726343035697937, "learning_rate": 4.984115493730027e-05, "loss": 0.6718, "num_input_tokens_seen": 3309840, "step": 5050 }, { "epoch": 2.64937106918239, "grad_norm": 0.24130657315254211, "learning_rate": 4.983986542396123e-05, "loss": 0.429, "num_input_tokens_seen": 3313264, "step": 5055 }, { "epoch": 2.651991614255765, "grad_norm": 0.21467691659927368, "learning_rate": 4.983857071440573e-05, "loss": 0.5599, "num_input_tokens_seen": 3316368, "step": 5060 }, { "epoch": 2.6546121593291403, "grad_norm": 0.2360885590314865, "learning_rate": 4.9837270808904616e-05, "loss": 0.5925, "num_input_tokens_seen": 3319504, "step": 5065 }, { "epoch": 2.6572327044025155, "grad_norm": 0.29329603910446167, "learning_rate": 4.9835965707729807e-05, "loss": 0.5581, "num_input_tokens_seen": 3322096, "step": 5070 }, { "epoch": 2.659853249475891, "grad_norm": 0.13427485525608063, "learning_rate": 4.983465541115431e-05, "loss": 0.5974, "num_input_tokens_seen": 3325072, "step": 5075 }, { "epoch": 2.6624737945492662, "grad_norm": 0.14588138461112976, "learning_rate": 4.983333991945223e-05, "loss": 0.5808, "num_input_tokens_seen": 3329456, "step": 5080 }, { "epoch": 2.6650943396226414, "grad_norm": 0.2696894705295563, "learning_rate": 4.983201923289875e-05, "loss": 0.6117, "num_input_tokens_seen": 3332592, "step": 5085 }, { "epoch": 2.667714884696017, "grad_norm": 0.15300951898097992, "learning_rate": 4.983069335177014e-05, "loss": 0.7157, "num_input_tokens_seen": 3335152, "step": 5090 }, { "epoch": 2.670335429769392, "grad_norm": 0.13655985891819, "learning_rate": 4.982936227634378e-05, "loss": 0.4888, "num_input_tokens_seen": 3338384, "step": 5095 }, { "epoch": 2.6729559748427674, "grad_norm": 0.3365355134010315, "learning_rate": 4.982802600689809e-05, "loss": 0.5548, "num_input_tokens_seen": 3341840, "step": 5100 }, { "epoch": 2.6755765199161425, "grad_norm": 0.18954254686832428, "learning_rate": 4.982668454371261e-05, "loss": 0.7001, "num_input_tokens_seen": 3345776, "step": 5105 }, { "epoch": 2.6781970649895177, "grad_norm": 0.2567233741283417, "learning_rate": 4.9825337887067966e-05, "loss": 0.4544, "num_input_tokens_seen": 3348656, "step": 5110 }, { "epoch": 2.680817610062893, "grad_norm": 0.15969006717205048, "learning_rate": 4.982398603724586e-05, "loss": 0.5286, "num_input_tokens_seen": 3351344, "step": 5115 }, { "epoch": 2.6834381551362685, "grad_norm": 0.196327343583107, "learning_rate": 4.982262899452909e-05, "loss": 0.4599, "num_input_tokens_seen": 3354192, "step": 5120 }, { "epoch": 2.6860587002096437, "grad_norm": 0.22279706597328186, "learning_rate": 4.982126675920151e-05, "loss": 0.4275, "num_input_tokens_seen": 3358960, "step": 5125 }, { "epoch": 2.688679245283019, "grad_norm": 0.13365770876407623, "learning_rate": 4.98198993315481e-05, "loss": 0.393, "num_input_tokens_seen": 3361744, "step": 5130 }, { "epoch": 2.691299790356394, "grad_norm": 0.1896636039018631, "learning_rate": 4.981852671185492e-05, "loss": 0.5225, "num_input_tokens_seen": 3364464, "step": 5135 }, { "epoch": 2.6939203354297696, "grad_norm": 0.13453565537929535, "learning_rate": 4.9817148900409093e-05, "loss": 0.4391, "num_input_tokens_seen": 3367728, "step": 5140 }, { "epoch": 2.6965408805031448, "grad_norm": 0.22290951013565063, "learning_rate": 4.981576589749885e-05, "loss": 0.5717, "num_input_tokens_seen": 3371056, "step": 5145 }, { "epoch": 2.69916142557652, "grad_norm": 0.35489147901535034, "learning_rate": 4.981437770341348e-05, "loss": 0.5258, "num_input_tokens_seen": 3374256, "step": 5150 }, { "epoch": 2.701781970649895, "grad_norm": 0.12749020755290985, "learning_rate": 4.9812984318443414e-05, "loss": 0.5968, "num_input_tokens_seen": 3377776, "step": 5155 }, { "epoch": 2.7044025157232703, "grad_norm": 0.2627698481082916, "learning_rate": 4.98115857428801e-05, "loss": 0.7888, "num_input_tokens_seen": 3381200, "step": 5160 }, { "epoch": 2.7070230607966455, "grad_norm": 0.22409479320049286, "learning_rate": 4.9810181977016104e-05, "loss": 0.5465, "num_input_tokens_seen": 3384048, "step": 5165 }, { "epoch": 2.709643605870021, "grad_norm": 0.1601647287607193, "learning_rate": 4.9808773021145096e-05, "loss": 0.4446, "num_input_tokens_seen": 3386800, "step": 5170 }, { "epoch": 2.7122641509433962, "grad_norm": 0.24969790875911713, "learning_rate": 4.98073588755618e-05, "loss": 0.5712, "num_input_tokens_seen": 3389840, "step": 5175 }, { "epoch": 2.7148846960167714, "grad_norm": 0.2875523567199707, "learning_rate": 4.980593954056205e-05, "loss": 0.5887, "num_input_tokens_seen": 3393264, "step": 5180 }, { "epoch": 2.717505241090147, "grad_norm": 0.3089263141155243, "learning_rate": 4.9804515016442754e-05, "loss": 0.5538, "num_input_tokens_seen": 3396720, "step": 5185 }, { "epoch": 2.720125786163522, "grad_norm": 0.14085298776626587, "learning_rate": 4.9803085303501896e-05, "loss": 0.6882, "num_input_tokens_seen": 3399792, "step": 5190 }, { "epoch": 2.7227463312368974, "grad_norm": 0.29548531770706177, "learning_rate": 4.9801650402038555e-05, "loss": 0.5258, "num_input_tokens_seen": 3402576, "step": 5195 }, { "epoch": 2.7253668763102725, "grad_norm": 0.12694799900054932, "learning_rate": 4.980021031235291e-05, "loss": 0.5, "num_input_tokens_seen": 3405648, "step": 5200 }, { "epoch": 2.7279874213836477, "grad_norm": 0.13086187839508057, "learning_rate": 4.9798765034746206e-05, "loss": 0.5067, "num_input_tokens_seen": 3408816, "step": 5205 }, { "epoch": 2.730607966457023, "grad_norm": 0.14181433618068695, "learning_rate": 4.979731456952077e-05, "loss": 0.4744, "num_input_tokens_seen": 3412400, "step": 5210 }, { "epoch": 2.7332285115303985, "grad_norm": 0.149038165807724, "learning_rate": 4.9795858916980034e-05, "loss": 0.4384, "num_input_tokens_seen": 3415440, "step": 5215 }, { "epoch": 2.7358490566037736, "grad_norm": 0.19792252779006958, "learning_rate": 4.9794398077428495e-05, "loss": 0.505, "num_input_tokens_seen": 3418736, "step": 5220 }, { "epoch": 2.738469601677149, "grad_norm": 0.23303468525409698, "learning_rate": 4.979293205117175e-05, "loss": 0.555, "num_input_tokens_seen": 3421456, "step": 5225 }, { "epoch": 2.741090146750524, "grad_norm": 0.3136582374572754, "learning_rate": 4.979146083851647e-05, "loss": 0.6581, "num_input_tokens_seen": 3424432, "step": 5230 }, { "epoch": 2.7437106918238996, "grad_norm": 0.186526820063591, "learning_rate": 4.978998443977042e-05, "loss": 0.465, "num_input_tokens_seen": 3427504, "step": 5235 }, { "epoch": 2.7463312368972748, "grad_norm": 0.2528383135795593, "learning_rate": 4.9788502855242436e-05, "loss": 0.4005, "num_input_tokens_seen": 3431568, "step": 5240 }, { "epoch": 2.74895178197065, "grad_norm": 0.14081180095672607, "learning_rate": 4.978701608524247e-05, "loss": 0.6867, "num_input_tokens_seen": 3434736, "step": 5245 }, { "epoch": 2.751572327044025, "grad_norm": 0.34970030188560486, "learning_rate": 4.978552413008152e-05, "loss": 0.5488, "num_input_tokens_seen": 3437360, "step": 5250 }, { "epoch": 2.7541928721174003, "grad_norm": 0.21212415397167206, "learning_rate": 4.9784026990071687e-05, "loss": 0.4616, "num_input_tokens_seen": 3441008, "step": 5255 }, { "epoch": 2.7568134171907754, "grad_norm": 0.10721341520547867, "learning_rate": 4.978252466552615e-05, "loss": 0.438, "num_input_tokens_seen": 3445072, "step": 5260 }, { "epoch": 2.759433962264151, "grad_norm": 0.13100561499595642, "learning_rate": 4.9781017156759205e-05, "loss": 0.5465, "num_input_tokens_seen": 3447952, "step": 5265 }, { "epoch": 2.762054507337526, "grad_norm": 0.14856506884098053, "learning_rate": 4.977950446408618e-05, "loss": 0.5217, "num_input_tokens_seen": 3450640, "step": 5270 }, { "epoch": 2.7646750524109014, "grad_norm": 0.2542225420475006, "learning_rate": 4.9777986587823523e-05, "loss": 0.7004, "num_input_tokens_seen": 3454256, "step": 5275 }, { "epoch": 2.767295597484277, "grad_norm": 0.4187909960746765, "learning_rate": 4.977646352828875e-05, "loss": 0.6303, "num_input_tokens_seen": 3457104, "step": 5280 }, { "epoch": 2.769916142557652, "grad_norm": 0.2818000018596649, "learning_rate": 4.9774935285800476e-05, "loss": 0.409, "num_input_tokens_seen": 3460016, "step": 5285 }, { "epoch": 2.7725366876310273, "grad_norm": 0.11063932627439499, "learning_rate": 4.977340186067838e-05, "loss": 0.6487, "num_input_tokens_seen": 3464464, "step": 5290 }, { "epoch": 2.7751572327044025, "grad_norm": 0.15119335055351257, "learning_rate": 4.977186325324325e-05, "loss": 0.4548, "num_input_tokens_seen": 3468048, "step": 5295 }, { "epoch": 2.7777777777777777, "grad_norm": 0.17056240141391754, "learning_rate": 4.977031946381694e-05, "loss": 0.4869, "num_input_tokens_seen": 3470736, "step": 5300 }, { "epoch": 2.780398322851153, "grad_norm": 0.1343960464000702, "learning_rate": 4.9768770492722394e-05, "loss": 0.737, "num_input_tokens_seen": 3473424, "step": 5305 }, { "epoch": 2.7830188679245285, "grad_norm": 0.30415618419647217, "learning_rate": 4.976721634028363e-05, "loss": 0.4888, "num_input_tokens_seen": 3476816, "step": 5310 }, { "epoch": 2.7856394129979036, "grad_norm": 0.15892189741134644, "learning_rate": 4.976565700682578e-05, "loss": 0.6632, "num_input_tokens_seen": 3480016, "step": 5315 }, { "epoch": 2.788259958071279, "grad_norm": 0.2979678511619568, "learning_rate": 4.976409249267502e-05, "loss": 0.4957, "num_input_tokens_seen": 3483824, "step": 5320 }, { "epoch": 2.790880503144654, "grad_norm": 0.18669888377189636, "learning_rate": 4.976252279815863e-05, "loss": 0.4828, "num_input_tokens_seen": 3486416, "step": 5325 }, { "epoch": 2.7935010482180296, "grad_norm": 0.16954150795936584, "learning_rate": 4.976094792360498e-05, "loss": 0.5403, "num_input_tokens_seen": 3490384, "step": 5330 }, { "epoch": 2.7961215932914047, "grad_norm": 0.17366717755794525, "learning_rate": 4.975936786934351e-05, "loss": 0.4683, "num_input_tokens_seen": 3493136, "step": 5335 }, { "epoch": 2.79874213836478, "grad_norm": 0.10864986479282379, "learning_rate": 4.975778263570475e-05, "loss": 0.5477, "num_input_tokens_seen": 3496112, "step": 5340 }, { "epoch": 2.801362683438155, "grad_norm": 0.4345795214176178, "learning_rate": 4.975619222302032e-05, "loss": 0.8126, "num_input_tokens_seen": 3498864, "step": 5345 }, { "epoch": 2.8039832285115303, "grad_norm": 0.2465321570634842, "learning_rate": 4.975459663162291e-05, "loss": 0.4436, "num_input_tokens_seen": 3502224, "step": 5350 }, { "epoch": 2.8066037735849054, "grad_norm": 0.22107917070388794, "learning_rate": 4.9752995861846296e-05, "loss": 0.5052, "num_input_tokens_seen": 3505488, "step": 5355 }, { "epoch": 2.809224318658281, "grad_norm": 0.20938080549240112, "learning_rate": 4.975138991402535e-05, "loss": 0.504, "num_input_tokens_seen": 3508560, "step": 5360 }, { "epoch": 2.811844863731656, "grad_norm": 0.12362340837717056, "learning_rate": 4.974977878849601e-05, "loss": 0.4117, "num_input_tokens_seen": 3512944, "step": 5365 }, { "epoch": 2.8144654088050314, "grad_norm": 0.22885026037693024, "learning_rate": 4.9748162485595303e-05, "loss": 0.5124, "num_input_tokens_seen": 3517232, "step": 5370 }, { "epoch": 2.8170859538784065, "grad_norm": 0.24870546162128448, "learning_rate": 4.974654100566135e-05, "loss": 0.5786, "num_input_tokens_seen": 3520304, "step": 5375 }, { "epoch": 2.819706498951782, "grad_norm": 0.2939617931842804, "learning_rate": 4.974491434903335e-05, "loss": 0.7651, "num_input_tokens_seen": 3523472, "step": 5380 }, { "epoch": 2.8223270440251573, "grad_norm": 0.22764652967453003, "learning_rate": 4.9743282516051556e-05, "loss": 0.5715, "num_input_tokens_seen": 3528336, "step": 5385 }, { "epoch": 2.8249475890985325, "grad_norm": 0.28692400455474854, "learning_rate": 4.974164550705736e-05, "loss": 0.4684, "num_input_tokens_seen": 3533680, "step": 5390 }, { "epoch": 2.8275681341719077, "grad_norm": 0.1741792857646942, "learning_rate": 4.974000332239319e-05, "loss": 0.5592, "num_input_tokens_seen": 3536656, "step": 5395 }, { "epoch": 2.830188679245283, "grad_norm": 0.2683621048927307, "learning_rate": 4.973835596240257e-05, "loss": 0.6631, "num_input_tokens_seen": 3539280, "step": 5400 }, { "epoch": 2.832809224318658, "grad_norm": 0.18990464508533478, "learning_rate": 4.9736703427430114e-05, "loss": 0.5545, "num_input_tokens_seen": 3542000, "step": 5405 }, { "epoch": 2.8354297693920336, "grad_norm": 0.2891353964805603, "learning_rate": 4.9735045717821514e-05, "loss": 0.5165, "num_input_tokens_seen": 3544624, "step": 5410 }, { "epoch": 2.838050314465409, "grad_norm": 0.1248798742890358, "learning_rate": 4.973338283392354e-05, "loss": 0.4257, "num_input_tokens_seen": 3547568, "step": 5415 }, { "epoch": 2.840670859538784, "grad_norm": 0.16655662655830383, "learning_rate": 4.9731714776084046e-05, "loss": 0.6255, "num_input_tokens_seen": 3550992, "step": 5420 }, { "epoch": 2.8432914046121596, "grad_norm": 0.15949925780296326, "learning_rate": 4.973004154465198e-05, "loss": 0.5478, "num_input_tokens_seen": 3554896, "step": 5425 }, { "epoch": 2.8459119496855347, "grad_norm": 0.16688132286071777, "learning_rate": 4.972836313997736e-05, "loss": 0.5639, "num_input_tokens_seen": 3557744, "step": 5430 }, { "epoch": 2.84853249475891, "grad_norm": 0.17079992592334747, "learning_rate": 4.9726679562411285e-05, "loss": 0.5468, "num_input_tokens_seen": 3561136, "step": 5435 }, { "epoch": 2.851153039832285, "grad_norm": 0.159438356757164, "learning_rate": 4.972499081230595e-05, "loss": 0.5168, "num_input_tokens_seen": 3563984, "step": 5440 }, { "epoch": 2.8537735849056602, "grad_norm": 0.2933115065097809, "learning_rate": 4.9723296890014605e-05, "loss": 0.4703, "num_input_tokens_seen": 3567184, "step": 5445 }, { "epoch": 2.8563941299790354, "grad_norm": 0.1295139640569687, "learning_rate": 4.9721597795891605e-05, "loss": 0.4689, "num_input_tokens_seen": 3571120, "step": 5450 }, { "epoch": 2.859014675052411, "grad_norm": 0.2538633644580841, "learning_rate": 4.9719893530292396e-05, "loss": 0.4802, "num_input_tokens_seen": 3574000, "step": 5455 }, { "epoch": 2.861635220125786, "grad_norm": 0.29781031608581543, "learning_rate": 4.9718184093573475e-05, "loss": 0.6666, "num_input_tokens_seen": 3577296, "step": 5460 }, { "epoch": 2.8642557651991614, "grad_norm": 0.24026481807231903, "learning_rate": 4.9716469486092435e-05, "loss": 0.6307, "num_input_tokens_seen": 3580368, "step": 5465 }, { "epoch": 2.8668763102725365, "grad_norm": 0.8172718286514282, "learning_rate": 4.971474970820797e-05, "loss": 0.4013, "num_input_tokens_seen": 3583984, "step": 5470 }, { "epoch": 2.869496855345912, "grad_norm": 0.09404393285512924, "learning_rate": 4.971302476027982e-05, "loss": 0.5307, "num_input_tokens_seen": 3588048, "step": 5475 }, { "epoch": 2.8721174004192873, "grad_norm": 0.23864704370498657, "learning_rate": 4.971129464266884e-05, "loss": 0.5384, "num_input_tokens_seen": 3590576, "step": 5480 }, { "epoch": 2.8747379454926625, "grad_norm": 0.19157728552818298, "learning_rate": 4.970955935573693e-05, "loss": 0.6362, "num_input_tokens_seen": 3593072, "step": 5485 }, { "epoch": 2.8773584905660377, "grad_norm": 0.3187257647514343, "learning_rate": 4.9707818899847106e-05, "loss": 0.5394, "num_input_tokens_seen": 3596144, "step": 5490 }, { "epoch": 2.879979035639413, "grad_norm": 0.16766957938671112, "learning_rate": 4.970607327536346e-05, "loss": 0.6128, "num_input_tokens_seen": 3599952, "step": 5495 }, { "epoch": 2.882599580712788, "grad_norm": 0.19441542029380798, "learning_rate": 4.970432248265114e-05, "loss": 0.5896, "num_input_tokens_seen": 3603152, "step": 5500 }, { "epoch": 2.8852201257861636, "grad_norm": 0.13031412661075592, "learning_rate": 4.97025665220764e-05, "loss": 0.5603, "num_input_tokens_seen": 3607504, "step": 5505 }, { "epoch": 2.8878406708595388, "grad_norm": 0.19971339404582977, "learning_rate": 4.970080539400656e-05, "loss": 0.4397, "num_input_tokens_seen": 3611440, "step": 5510 }, { "epoch": 2.890461215932914, "grad_norm": 0.302946001291275, "learning_rate": 4.969903909881004e-05, "loss": 0.5503, "num_input_tokens_seen": 3614128, "step": 5515 }, { "epoch": 2.8930817610062896, "grad_norm": 0.18259304761886597, "learning_rate": 4.9697267636856317e-05, "loss": 0.4307, "num_input_tokens_seen": 3618256, "step": 5520 }, { "epoch": 2.8957023060796647, "grad_norm": 0.23514221608638763, "learning_rate": 4.969549100851597e-05, "loss": 0.5564, "num_input_tokens_seen": 3621360, "step": 5525 }, { "epoch": 2.89832285115304, "grad_norm": 0.1874706894159317, "learning_rate": 4.9693709214160645e-05, "loss": 0.4156, "num_input_tokens_seen": 3624432, "step": 5530 }, { "epoch": 2.900943396226415, "grad_norm": 0.18365566432476044, "learning_rate": 4.969192225416307e-05, "loss": 0.6069, "num_input_tokens_seen": 3626800, "step": 5535 }, { "epoch": 2.9035639412997902, "grad_norm": 0.2616543769836426, "learning_rate": 4.969013012889706e-05, "loss": 0.6516, "num_input_tokens_seen": 3629424, "step": 5540 }, { "epoch": 2.9061844863731654, "grad_norm": 0.1264335960149765, "learning_rate": 4.9688332838737504e-05, "loss": 0.6457, "num_input_tokens_seen": 3632496, "step": 5545 }, { "epoch": 2.908805031446541, "grad_norm": 0.11053810268640518, "learning_rate": 4.9686530384060386e-05, "loss": 0.4958, "num_input_tokens_seen": 3635568, "step": 5550 }, { "epoch": 2.911425576519916, "grad_norm": 0.22973570227622986, "learning_rate": 4.968472276524275e-05, "loss": 0.6048, "num_input_tokens_seen": 3638704, "step": 5555 }, { "epoch": 2.9140461215932913, "grad_norm": 0.2640734314918518, "learning_rate": 4.9682909982662725e-05, "loss": 0.5279, "num_input_tokens_seen": 3641904, "step": 5560 }, { "epoch": 2.9166666666666665, "grad_norm": 0.23688867688179016, "learning_rate": 4.968109203669953e-05, "loss": 0.4569, "num_input_tokens_seen": 3645296, "step": 5565 }, { "epoch": 2.919287211740042, "grad_norm": 0.16980625689029694, "learning_rate": 4.967926892773346e-05, "loss": 0.6876, "num_input_tokens_seen": 3648400, "step": 5570 }, { "epoch": 2.9219077568134173, "grad_norm": 0.35554566979408264, "learning_rate": 4.967744065614588e-05, "loss": 0.4659, "num_input_tokens_seen": 3651600, "step": 5575 }, { "epoch": 2.9245283018867925, "grad_norm": 0.12490098923444748, "learning_rate": 4.9675607222319255e-05, "loss": 0.5653, "num_input_tokens_seen": 3655280, "step": 5580 }, { "epoch": 2.9271488469601676, "grad_norm": 0.17698180675506592, "learning_rate": 4.967376862663711e-05, "loss": 0.5014, "num_input_tokens_seen": 3657584, "step": 5585 }, { "epoch": 2.929769392033543, "grad_norm": 0.15276603400707245, "learning_rate": 4.9671924869484064e-05, "loss": 0.4341, "num_input_tokens_seen": 3660592, "step": 5590 }, { "epoch": 2.932389937106918, "grad_norm": 0.35519710183143616, "learning_rate": 4.967007595124581e-05, "loss": 0.5319, "num_input_tokens_seen": 3663120, "step": 5595 }, { "epoch": 2.9350104821802936, "grad_norm": 0.26053714752197266, "learning_rate": 4.966822187230911e-05, "loss": 0.4525, "num_input_tokens_seen": 3665456, "step": 5600 }, { "epoch": 2.9376310272536688, "grad_norm": 0.3700225353240967, "learning_rate": 4.966636263306183e-05, "loss": 0.5013, "num_input_tokens_seen": 3668080, "step": 5605 }, { "epoch": 2.940251572327044, "grad_norm": 0.2649376392364502, "learning_rate": 4.9664498233892895e-05, "loss": 0.5097, "num_input_tokens_seen": 3670672, "step": 5610 }, { "epoch": 2.9428721174004195, "grad_norm": 0.2168603390455246, "learning_rate": 4.966262867519232e-05, "loss": 0.5406, "num_input_tokens_seen": 3675312, "step": 5615 }, { "epoch": 2.9454926624737947, "grad_norm": 0.17792876064777374, "learning_rate": 4.966075395735119e-05, "loss": 0.385, "num_input_tokens_seen": 3678864, "step": 5620 }, { "epoch": 2.94811320754717, "grad_norm": 0.28188273310661316, "learning_rate": 4.965887408076167e-05, "loss": 0.5578, "num_input_tokens_seen": 3681808, "step": 5625 }, { "epoch": 2.950733752620545, "grad_norm": 0.20976987481117249, "learning_rate": 4.965698904581703e-05, "loss": 0.4498, "num_input_tokens_seen": 3684720, "step": 5630 }, { "epoch": 2.95335429769392, "grad_norm": 0.17660434544086456, "learning_rate": 4.965509885291157e-05, "loss": 0.5854, "num_input_tokens_seen": 3687920, "step": 5635 }, { "epoch": 2.9559748427672954, "grad_norm": 0.14918024837970734, "learning_rate": 4.9653203502440716e-05, "loss": 0.5421, "num_input_tokens_seen": 3691152, "step": 5640 }, { "epoch": 2.958595387840671, "grad_norm": 0.17916081845760345, "learning_rate": 4.965130299480095e-05, "loss": 0.5877, "num_input_tokens_seen": 3694000, "step": 5645 }, { "epoch": 2.961215932914046, "grad_norm": 0.21818600594997406, "learning_rate": 4.9649397330389834e-05, "loss": 0.5073, "num_input_tokens_seen": 3696816, "step": 5650 }, { "epoch": 2.9638364779874213, "grad_norm": 0.21475355327129364, "learning_rate": 4.964748650960601e-05, "loss": 0.5565, "num_input_tokens_seen": 3699184, "step": 5655 }, { "epoch": 2.9664570230607965, "grad_norm": 0.2018856257200241, "learning_rate": 4.96455705328492e-05, "loss": 0.6339, "num_input_tokens_seen": 3701552, "step": 5660 }, { "epoch": 2.969077568134172, "grad_norm": 0.13611695170402527, "learning_rate": 4.964364940052021e-05, "loss": 0.5614, "num_input_tokens_seen": 3704784, "step": 5665 }, { "epoch": 2.9716981132075473, "grad_norm": 0.3111140727996826, "learning_rate": 4.964172311302091e-05, "loss": 0.4385, "num_input_tokens_seen": 3707760, "step": 5670 }, { "epoch": 2.9743186582809225, "grad_norm": 0.1356106400489807, "learning_rate": 4.9639791670754265e-05, "loss": 0.5875, "num_input_tokens_seen": 3711536, "step": 5675 }, { "epoch": 2.9769392033542976, "grad_norm": 0.12069517374038696, "learning_rate": 4.9637855074124305e-05, "loss": 0.4321, "num_input_tokens_seen": 3714224, "step": 5680 }, { "epoch": 2.979559748427673, "grad_norm": 0.1718376725912094, "learning_rate": 4.963591332353615e-05, "loss": 0.5015, "num_input_tokens_seen": 3717264, "step": 5685 }, { "epoch": 2.982180293501048, "grad_norm": 0.19077089428901672, "learning_rate": 4.9633966419396e-05, "loss": 0.5946, "num_input_tokens_seen": 3720496, "step": 5690 }, { "epoch": 2.9848008385744236, "grad_norm": 0.171136736869812, "learning_rate": 4.96320143621111e-05, "loss": 0.534, "num_input_tokens_seen": 3723280, "step": 5695 }, { "epoch": 2.9874213836477987, "grad_norm": 0.11390677839517593, "learning_rate": 4.963005715208981e-05, "loss": 0.7201, "num_input_tokens_seen": 3727888, "step": 5700 }, { "epoch": 2.990041928721174, "grad_norm": 0.18853069841861725, "learning_rate": 4.962809478974157e-05, "loss": 0.4642, "num_input_tokens_seen": 3731248, "step": 5705 }, { "epoch": 2.9926624737945495, "grad_norm": 0.2618752121925354, "learning_rate": 4.962612727547686e-05, "loss": 0.5342, "num_input_tokens_seen": 3734032, "step": 5710 }, { "epoch": 2.9952830188679247, "grad_norm": 0.6298772692680359, "learning_rate": 4.962415460970728e-05, "loss": 0.4816, "num_input_tokens_seen": 3737872, "step": 5715 }, { "epoch": 2.9979035639413, "grad_norm": 0.135541170835495, "learning_rate": 4.962217679284548e-05, "loss": 0.5057, "num_input_tokens_seen": 3741360, "step": 5720 }, { "epoch": 3.0, "eval_loss": 0.5058302283287048, "eval_runtime": 9.8581, "eval_samples_per_second": 86.021, "eval_steps_per_second": 21.505, "num_input_tokens_seen": 3743760, "step": 5724 }, { "epoch": 3.000524109014675, "grad_norm": 0.1788816899061203, "learning_rate": 4.962019382530521e-05, "loss": 0.518, "num_input_tokens_seen": 3744240, "step": 5725 }, { "epoch": 3.00314465408805, "grad_norm": 0.2593192458152771, "learning_rate": 4.9618205707501256e-05, "loss": 0.4953, "num_input_tokens_seen": 3746832, "step": 5730 }, { "epoch": 3.0057651991614254, "grad_norm": 0.1916285753250122, "learning_rate": 4.961621243984954e-05, "loss": 0.4781, "num_input_tokens_seen": 3750608, "step": 5735 }, { "epoch": 3.008385744234801, "grad_norm": 0.5093003511428833, "learning_rate": 4.9614214022767014e-05, "loss": 0.535, "num_input_tokens_seen": 3753360, "step": 5740 }, { "epoch": 3.011006289308176, "grad_norm": 0.264464795589447, "learning_rate": 4.9612210456671735e-05, "loss": 0.5684, "num_input_tokens_seen": 3755824, "step": 5745 }, { "epoch": 3.0136268343815513, "grad_norm": 0.18695400655269623, "learning_rate": 4.961020174198281e-05, "loss": 0.4239, "num_input_tokens_seen": 3758640, "step": 5750 }, { "epoch": 3.0162473794549265, "grad_norm": 0.23660707473754883, "learning_rate": 4.9608187879120454e-05, "loss": 0.5466, "num_input_tokens_seen": 3761712, "step": 5755 }, { "epoch": 3.018867924528302, "grad_norm": 0.34738245606422424, "learning_rate": 4.9606168868505936e-05, "loss": 0.4783, "num_input_tokens_seen": 3764368, "step": 5760 }, { "epoch": 3.0214884696016773, "grad_norm": 0.10861897468566895, "learning_rate": 4.9604144710561616e-05, "loss": 0.6407, "num_input_tokens_seen": 3768112, "step": 5765 }, { "epoch": 3.0241090146750524, "grad_norm": 0.11129090934991837, "learning_rate": 4.960211540571092e-05, "loss": 0.5215, "num_input_tokens_seen": 3773200, "step": 5770 }, { "epoch": 3.0267295597484276, "grad_norm": 0.18277011811733246, "learning_rate": 4.960008095437836e-05, "loss": 0.6512, "num_input_tokens_seen": 3777360, "step": 5775 }, { "epoch": 3.029350104821803, "grad_norm": 0.30373942852020264, "learning_rate": 4.959804135698951e-05, "loss": 0.6471, "num_input_tokens_seen": 3780240, "step": 5780 }, { "epoch": 3.0319706498951784, "grad_norm": 0.3350612223148346, "learning_rate": 4.959599661397104e-05, "loss": 0.4896, "num_input_tokens_seen": 3784528, "step": 5785 }, { "epoch": 3.0345911949685536, "grad_norm": 0.15285207331180573, "learning_rate": 4.959394672575068e-05, "loss": 0.5805, "num_input_tokens_seen": 3787376, "step": 5790 }, { "epoch": 3.0372117400419287, "grad_norm": 0.13998788595199585, "learning_rate": 4.959189169275725e-05, "loss": 0.5186, "num_input_tokens_seen": 3790512, "step": 5795 }, { "epoch": 3.039832285115304, "grad_norm": 0.14898698031902313, "learning_rate": 4.958983151542065e-05, "loss": 0.5245, "num_input_tokens_seen": 3793744, "step": 5800 }, { "epoch": 3.042452830188679, "grad_norm": 0.15143398940563202, "learning_rate": 4.958776619417182e-05, "loss": 0.5882, "num_input_tokens_seen": 3796560, "step": 5805 }, { "epoch": 3.0450733752620547, "grad_norm": 0.16342133283615112, "learning_rate": 4.9585695729442814e-05, "loss": 0.4276, "num_input_tokens_seen": 3799440, "step": 5810 }, { "epoch": 3.04769392033543, "grad_norm": 0.1395065039396286, "learning_rate": 4.9583620121666754e-05, "loss": 0.5824, "num_input_tokens_seen": 3802320, "step": 5815 }, { "epoch": 3.050314465408805, "grad_norm": 0.17403574287891388, "learning_rate": 4.958153937127783e-05, "loss": 0.5062, "num_input_tokens_seen": 3805552, "step": 5820 }, { "epoch": 3.05293501048218, "grad_norm": 0.2162361592054367, "learning_rate": 4.95794534787113e-05, "loss": 0.5491, "num_input_tokens_seen": 3808688, "step": 5825 }, { "epoch": 3.0555555555555554, "grad_norm": 0.13335402309894562, "learning_rate": 4.957736244440353e-05, "loss": 0.4627, "num_input_tokens_seen": 3811952, "step": 5830 }, { "epoch": 3.058176100628931, "grad_norm": 0.16390761733055115, "learning_rate": 4.957526626879193e-05, "loss": 0.5199, "num_input_tokens_seen": 3818960, "step": 5835 }, { "epoch": 3.060796645702306, "grad_norm": 0.161413311958313, "learning_rate": 4.957316495231499e-05, "loss": 0.4762, "num_input_tokens_seen": 3822384, "step": 5840 }, { "epoch": 3.0634171907756813, "grad_norm": 0.13913187384605408, "learning_rate": 4.9571058495412294e-05, "loss": 0.5984, "num_input_tokens_seen": 3825136, "step": 5845 }, { "epoch": 3.0660377358490565, "grad_norm": 0.12196864187717438, "learning_rate": 4.9568946898524485e-05, "loss": 0.4826, "num_input_tokens_seen": 3829680, "step": 5850 }, { "epoch": 3.068658280922432, "grad_norm": 0.18105950951576233, "learning_rate": 4.956683016209327e-05, "loss": 0.4949, "num_input_tokens_seen": 3832688, "step": 5855 }, { "epoch": 3.0712788259958073, "grad_norm": 0.1713331639766693, "learning_rate": 4.956470828656147e-05, "loss": 0.4999, "num_input_tokens_seen": 3835472, "step": 5860 }, { "epoch": 3.0738993710691824, "grad_norm": 0.14383885264396667, "learning_rate": 4.956258127237294e-05, "loss": 0.5362, "num_input_tokens_seen": 3838192, "step": 5865 }, { "epoch": 3.0765199161425576, "grad_norm": 0.28230586647987366, "learning_rate": 4.956044911997263e-05, "loss": 0.5857, "num_input_tokens_seen": 3841584, "step": 5870 }, { "epoch": 3.0791404612159328, "grad_norm": 0.13348372280597687, "learning_rate": 4.955831182980657e-05, "loss": 0.5381, "num_input_tokens_seen": 3844400, "step": 5875 }, { "epoch": 3.0817610062893084, "grad_norm": 0.04090273752808571, "learning_rate": 4.955616940232185e-05, "loss": 0.3975, "num_input_tokens_seen": 3851632, "step": 5880 }, { "epoch": 3.0843815513626835, "grad_norm": 0.2720227837562561, "learning_rate": 4.9554021837966636e-05, "loss": 0.5076, "num_input_tokens_seen": 3854448, "step": 5885 }, { "epoch": 3.0870020964360587, "grad_norm": 0.12987609207630157, "learning_rate": 4.955186913719019e-05, "loss": 0.6074, "num_input_tokens_seen": 3858128, "step": 5890 }, { "epoch": 3.089622641509434, "grad_norm": 0.19692157208919525, "learning_rate": 4.954971130044281e-05, "loss": 0.4904, "num_input_tokens_seen": 3861104, "step": 5895 }, { "epoch": 3.092243186582809, "grad_norm": 0.1600429117679596, "learning_rate": 4.954754832817592e-05, "loss": 0.4844, "num_input_tokens_seen": 3864368, "step": 5900 }, { "epoch": 3.0948637316561847, "grad_norm": 0.15146519243717194, "learning_rate": 4.954538022084196e-05, "loss": 0.5627, "num_input_tokens_seen": 3867184, "step": 5905 }, { "epoch": 3.09748427672956, "grad_norm": 0.16642014682292938, "learning_rate": 4.9543206978894496e-05, "loss": 0.5487, "num_input_tokens_seen": 3869776, "step": 5910 }, { "epoch": 3.100104821802935, "grad_norm": 0.9207198619842529, "learning_rate": 4.954102860278813e-05, "loss": 0.7673, "num_input_tokens_seen": 3871952, "step": 5915 }, { "epoch": 3.10272536687631, "grad_norm": 0.1340644508600235, "learning_rate": 4.9538845092978567e-05, "loss": 0.4956, "num_input_tokens_seen": 3875408, "step": 5920 }, { "epoch": 3.1053459119496853, "grad_norm": 0.21609468758106232, "learning_rate": 4.953665644992256e-05, "loss": 0.7023, "num_input_tokens_seen": 3879664, "step": 5925 }, { "epoch": 3.107966457023061, "grad_norm": 0.19554205238819122, "learning_rate": 4.953446267407795e-05, "loss": 0.6694, "num_input_tokens_seen": 3882928, "step": 5930 }, { "epoch": 3.110587002096436, "grad_norm": 0.1826043426990509, "learning_rate": 4.953226376590365e-05, "loss": 0.6001, "num_input_tokens_seen": 3886288, "step": 5935 }, { "epoch": 3.1132075471698113, "grad_norm": 0.14658674597740173, "learning_rate": 4.953005972585965e-05, "loss": 0.6836, "num_input_tokens_seen": 3889232, "step": 5940 }, { "epoch": 3.1158280922431865, "grad_norm": 0.2134912759065628, "learning_rate": 4.952785055440701e-05, "loss": 0.5631, "num_input_tokens_seen": 3892528, "step": 5945 }, { "epoch": 3.1184486373165616, "grad_norm": 0.16181334853172302, "learning_rate": 4.9525636252007864e-05, "loss": 0.5772, "num_input_tokens_seen": 3894992, "step": 5950 }, { "epoch": 3.1210691823899372, "grad_norm": 0.5151065587997437, "learning_rate": 4.9523416819125425e-05, "loss": 0.5191, "num_input_tokens_seen": 3897744, "step": 5955 }, { "epoch": 3.1236897274633124, "grad_norm": 0.21084779500961304, "learning_rate": 4.952119225622396e-05, "loss": 0.5016, "num_input_tokens_seen": 3899888, "step": 5960 }, { "epoch": 3.1263102725366876, "grad_norm": 0.1897493153810501, "learning_rate": 4.951896256376882e-05, "loss": 0.6188, "num_input_tokens_seen": 3905904, "step": 5965 }, { "epoch": 3.1289308176100628, "grad_norm": 0.17686010897159576, "learning_rate": 4.9516727742226444e-05, "loss": 0.6594, "num_input_tokens_seen": 3908688, "step": 5970 }, { "epoch": 3.131551362683438, "grad_norm": 0.19645628333091736, "learning_rate": 4.9514487792064323e-05, "loss": 0.6408, "num_input_tokens_seen": 3912048, "step": 5975 }, { "epoch": 3.1341719077568135, "grad_norm": 0.18819044530391693, "learning_rate": 4.951224271375104e-05, "loss": 0.5938, "num_input_tokens_seen": 3915344, "step": 5980 }, { "epoch": 3.1367924528301887, "grad_norm": 0.131166473031044, "learning_rate": 4.950999250775623e-05, "loss": 0.5912, "num_input_tokens_seen": 3918448, "step": 5985 }, { "epoch": 3.139412997903564, "grad_norm": 0.3381202518939972, "learning_rate": 4.950773717455061e-05, "loss": 0.6167, "num_input_tokens_seen": 3921040, "step": 5990 }, { "epoch": 3.142033542976939, "grad_norm": 0.2380809783935547, "learning_rate": 4.950547671460596e-05, "loss": 0.6077, "num_input_tokens_seen": 3924560, "step": 5995 }, { "epoch": 3.1446540880503147, "grad_norm": 0.1294398009777069, "learning_rate": 4.950321112839517e-05, "loss": 0.5523, "num_input_tokens_seen": 3927056, "step": 6000 }, { "epoch": 3.14727463312369, "grad_norm": 0.15933100879192352, "learning_rate": 4.9500940416392156e-05, "loss": 0.5117, "num_input_tokens_seen": 3930032, "step": 6005 }, { "epoch": 3.149895178197065, "grad_norm": 0.17581728100776672, "learning_rate": 4.949866457907193e-05, "loss": 0.456, "num_input_tokens_seen": 3933488, "step": 6010 }, { "epoch": 3.15251572327044, "grad_norm": 0.235385000705719, "learning_rate": 4.9496383616910565e-05, "loss": 0.5242, "num_input_tokens_seen": 3937488, "step": 6015 }, { "epoch": 3.1551362683438153, "grad_norm": 0.1360219120979309, "learning_rate": 4.949409753038522e-05, "loss": 0.6058, "num_input_tokens_seen": 3940528, "step": 6020 }, { "epoch": 3.157756813417191, "grad_norm": 0.31153154373168945, "learning_rate": 4.949180631997411e-05, "loss": 0.4585, "num_input_tokens_seen": 3944112, "step": 6025 }, { "epoch": 3.160377358490566, "grad_norm": 0.1988576501607895, "learning_rate": 4.948950998615654e-05, "loss": 0.5272, "num_input_tokens_seen": 3947568, "step": 6030 }, { "epoch": 3.1629979035639413, "grad_norm": 0.17619948089122772, "learning_rate": 4.9487208529412876e-05, "loss": 0.4649, "num_input_tokens_seen": 3950896, "step": 6035 }, { "epoch": 3.1656184486373165, "grad_norm": 0.2164357602596283, "learning_rate": 4.9484901950224546e-05, "loss": 0.5258, "num_input_tokens_seen": 3954256, "step": 6040 }, { "epoch": 3.1682389937106916, "grad_norm": 0.2204127162694931, "learning_rate": 4.948259024907407e-05, "loss": 0.5439, "num_input_tokens_seen": 3958640, "step": 6045 }, { "epoch": 3.1708595387840672, "grad_norm": 0.4122595191001892, "learning_rate": 4.948027342644502e-05, "loss": 0.6914, "num_input_tokens_seen": 3961072, "step": 6050 }, { "epoch": 3.1734800838574424, "grad_norm": 0.20805153250694275, "learning_rate": 4.947795148282206e-05, "loss": 0.4171, "num_input_tokens_seen": 3964816, "step": 6055 }, { "epoch": 3.1761006289308176, "grad_norm": 0.1806689351797104, "learning_rate": 4.947562441869091e-05, "loss": 0.5593, "num_input_tokens_seen": 3967824, "step": 6060 }, { "epoch": 3.1787211740041927, "grad_norm": 0.23762236535549164, "learning_rate": 4.947329223453836e-05, "loss": 0.5167, "num_input_tokens_seen": 3970960, "step": 6065 }, { "epoch": 3.181341719077568, "grad_norm": 0.18815521895885468, "learning_rate": 4.947095493085229e-05, "loss": 0.5002, "num_input_tokens_seen": 3973872, "step": 6070 }, { "epoch": 3.1839622641509435, "grad_norm": 0.2343703657388687, "learning_rate": 4.9468612508121616e-05, "loss": 0.6652, "num_input_tokens_seen": 3976496, "step": 6075 }, { "epoch": 3.1865828092243187, "grad_norm": 0.23769082129001617, "learning_rate": 4.9466264966836365e-05, "loss": 0.5292, "num_input_tokens_seen": 3979248, "step": 6080 }, { "epoch": 3.189203354297694, "grad_norm": 0.1475287228822708, "learning_rate": 4.94639123074876e-05, "loss": 0.5301, "num_input_tokens_seen": 3986352, "step": 6085 }, { "epoch": 3.191823899371069, "grad_norm": 0.25068575143814087, "learning_rate": 4.946155453056749e-05, "loss": 0.7008, "num_input_tokens_seen": 3989808, "step": 6090 }, { "epoch": 3.1944444444444446, "grad_norm": 0.15945814549922943, "learning_rate": 4.945919163656923e-05, "loss": 0.6853, "num_input_tokens_seen": 3993104, "step": 6095 }, { "epoch": 3.19706498951782, "grad_norm": 0.11211786419153214, "learning_rate": 4.945682362598715e-05, "loss": 0.4792, "num_input_tokens_seen": 3996240, "step": 6100 }, { "epoch": 3.199685534591195, "grad_norm": 0.14322447776794434, "learning_rate": 4.945445049931656e-05, "loss": 0.385, "num_input_tokens_seen": 4000336, "step": 6105 }, { "epoch": 3.20230607966457, "grad_norm": 0.15655742585659027, "learning_rate": 4.945207225705393e-05, "loss": 0.5303, "num_input_tokens_seen": 4004400, "step": 6110 }, { "epoch": 3.2049266247379453, "grad_norm": 0.1808287352323532, "learning_rate": 4.944968889969674e-05, "loss": 0.5172, "num_input_tokens_seen": 4007920, "step": 6115 }, { "epoch": 3.207547169811321, "grad_norm": 0.22319148480892181, "learning_rate": 4.944730042774358e-05, "loss": 0.4696, "num_input_tokens_seen": 4011152, "step": 6120 }, { "epoch": 3.210167714884696, "grad_norm": 0.14600522816181183, "learning_rate": 4.944490684169408e-05, "loss": 0.5038, "num_input_tokens_seen": 4013968, "step": 6125 }, { "epoch": 3.2127882599580713, "grad_norm": 0.20737262070178986, "learning_rate": 4.9442508142048945e-05, "loss": 0.5706, "num_input_tokens_seen": 4017072, "step": 6130 }, { "epoch": 3.2154088050314464, "grad_norm": 0.12533661723136902, "learning_rate": 4.9440104329309956e-05, "loss": 0.5921, "num_input_tokens_seen": 4020976, "step": 6135 }, { "epoch": 3.2180293501048216, "grad_norm": 0.12594811618328094, "learning_rate": 4.943769540397998e-05, "loss": 0.63, "num_input_tokens_seen": 4025008, "step": 6140 }, { "epoch": 3.220649895178197, "grad_norm": 0.13036051392555237, "learning_rate": 4.943528136656292e-05, "loss": 0.4965, "num_input_tokens_seen": 4029744, "step": 6145 }, { "epoch": 3.2232704402515724, "grad_norm": 0.1913880556821823, "learning_rate": 4.943286221756377e-05, "loss": 0.4233, "num_input_tokens_seen": 4033904, "step": 6150 }, { "epoch": 3.2258909853249476, "grad_norm": 0.15347784757614136, "learning_rate": 4.9430437957488586e-05, "loss": 0.5251, "num_input_tokens_seen": 4037392, "step": 6155 }, { "epoch": 3.2285115303983227, "grad_norm": 0.6124194264411926, "learning_rate": 4.94280085868445e-05, "loss": 0.6759, "num_input_tokens_seen": 4040336, "step": 6160 }, { "epoch": 3.231132075471698, "grad_norm": 0.2027629166841507, "learning_rate": 4.942557410613972e-05, "loss": 0.5819, "num_input_tokens_seen": 4044016, "step": 6165 }, { "epoch": 3.2337526205450735, "grad_norm": 0.21678680181503296, "learning_rate": 4.942313451588348e-05, "loss": 0.5222, "num_input_tokens_seen": 4046832, "step": 6170 }, { "epoch": 3.2363731656184487, "grad_norm": 0.2499133199453354, "learning_rate": 4.942068981658614e-05, "loss": 0.5441, "num_input_tokens_seen": 4048816, "step": 6175 }, { "epoch": 3.238993710691824, "grad_norm": 0.12625981867313385, "learning_rate": 4.9418240008759097e-05, "loss": 0.4946, "num_input_tokens_seen": 4051728, "step": 6180 }, { "epoch": 3.241614255765199, "grad_norm": 0.14127413928508759, "learning_rate": 4.9415785092914826e-05, "loss": 0.4354, "num_input_tokens_seen": 4055472, "step": 6185 }, { "epoch": 3.2442348008385746, "grad_norm": 0.1235988587141037, "learning_rate": 4.941332506956686e-05, "loss": 0.6532, "num_input_tokens_seen": 4059056, "step": 6190 }, { "epoch": 3.24685534591195, "grad_norm": 0.18893137574195862, "learning_rate": 4.941085993922981e-05, "loss": 0.4265, "num_input_tokens_seen": 4061840, "step": 6195 }, { "epoch": 3.249475890985325, "grad_norm": 0.17179112136363983, "learning_rate": 4.9408389702419354e-05, "loss": 0.5458, "num_input_tokens_seen": 4064656, "step": 6200 }, { "epoch": 3.2520964360587, "grad_norm": 0.21498030424118042, "learning_rate": 4.940591435965224e-05, "loss": 0.684, "num_input_tokens_seen": 4068368, "step": 6205 }, { "epoch": 3.2547169811320753, "grad_norm": 0.31825345754623413, "learning_rate": 4.9403433911446285e-05, "loss": 0.5913, "num_input_tokens_seen": 4071312, "step": 6210 }, { "epoch": 3.257337526205451, "grad_norm": 0.18158796429634094, "learning_rate": 4.9400948358320356e-05, "loss": 0.5387, "num_input_tokens_seen": 4074736, "step": 6215 }, { "epoch": 3.259958071278826, "grad_norm": 0.1005588248372078, "learning_rate": 4.939845770079442e-05, "loss": 0.4944, "num_input_tokens_seen": 4078960, "step": 6220 }, { "epoch": 3.2625786163522013, "grad_norm": 0.19782739877700806, "learning_rate": 4.9395961939389475e-05, "loss": 0.5476, "num_input_tokens_seen": 4082512, "step": 6225 }, { "epoch": 3.2651991614255764, "grad_norm": 0.17669731378555298, "learning_rate": 4.939346107462762e-05, "loss": 0.4564, "num_input_tokens_seen": 4086352, "step": 6230 }, { "epoch": 3.2678197064989516, "grad_norm": 0.5438869595527649, "learning_rate": 4.9390955107032e-05, "loss": 0.7189, "num_input_tokens_seen": 4090832, "step": 6235 }, { "epoch": 3.270440251572327, "grad_norm": 0.14589034020900726, "learning_rate": 4.938844403712685e-05, "loss": 0.5107, "num_input_tokens_seen": 4093936, "step": 6240 }, { "epoch": 3.2730607966457024, "grad_norm": 0.18991948664188385, "learning_rate": 4.938592786543743e-05, "loss": 0.6364, "num_input_tokens_seen": 4097136, "step": 6245 }, { "epoch": 3.2756813417190775, "grad_norm": 0.27368974685668945, "learning_rate": 4.938340659249012e-05, "loss": 0.7798, "num_input_tokens_seen": 4099632, "step": 6250 }, { "epoch": 3.2783018867924527, "grad_norm": 0.19494320452213287, "learning_rate": 4.938088021881233e-05, "loss": 0.5444, "num_input_tokens_seen": 4102576, "step": 6255 }, { "epoch": 3.280922431865828, "grad_norm": 0.13804641366004944, "learning_rate": 4.937834874493255e-05, "loss": 0.4646, "num_input_tokens_seen": 4106928, "step": 6260 }, { "epoch": 3.2835429769392035, "grad_norm": 0.26858049631118774, "learning_rate": 4.9375812171380334e-05, "loss": 0.6422, "num_input_tokens_seen": 4111152, "step": 6265 }, { "epoch": 3.2861635220125787, "grad_norm": 0.3031182289123535, "learning_rate": 4.9373270498686306e-05, "loss": 0.5305, "num_input_tokens_seen": 4113712, "step": 6270 }, { "epoch": 3.288784067085954, "grad_norm": 0.24239744246006012, "learning_rate": 4.937072372738216e-05, "loss": 0.7565, "num_input_tokens_seen": 4116944, "step": 6275 }, { "epoch": 3.291404612159329, "grad_norm": 0.25100407004356384, "learning_rate": 4.9368171858000635e-05, "loss": 0.4521, "num_input_tokens_seen": 4122128, "step": 6280 }, { "epoch": 3.2940251572327046, "grad_norm": 0.19467051327228546, "learning_rate": 4.9365614891075576e-05, "loss": 0.5934, "num_input_tokens_seen": 4125584, "step": 6285 }, { "epoch": 3.29664570230608, "grad_norm": 0.2884593605995178, "learning_rate": 4.936305282714185e-05, "loss": 0.4074, "num_input_tokens_seen": 4128784, "step": 6290 }, { "epoch": 3.299266247379455, "grad_norm": 0.261331170797348, "learning_rate": 4.936048566673542e-05, "loss": 0.4049, "num_input_tokens_seen": 4131120, "step": 6295 }, { "epoch": 3.30188679245283, "grad_norm": 0.13835158944129944, "learning_rate": 4.935791341039332e-05, "loss": 0.4788, "num_input_tokens_seen": 4133904, "step": 6300 }, { "epoch": 3.3045073375262053, "grad_norm": 0.10739225149154663, "learning_rate": 4.935533605865361e-05, "loss": 0.4269, "num_input_tokens_seen": 4137456, "step": 6305 }, { "epoch": 3.307127882599581, "grad_norm": 0.22614890336990356, "learning_rate": 4.935275361205547e-05, "loss": 0.6025, "num_input_tokens_seen": 4140272, "step": 6310 }, { "epoch": 3.309748427672956, "grad_norm": 0.14196191728115082, "learning_rate": 4.93501660711391e-05, "loss": 0.5531, "num_input_tokens_seen": 4143600, "step": 6315 }, { "epoch": 3.3123689727463312, "grad_norm": 0.3298580050468445, "learning_rate": 4.934757343644578e-05, "loss": 0.4546, "num_input_tokens_seen": 4146896, "step": 6320 }, { "epoch": 3.3149895178197064, "grad_norm": 0.13474802672863007, "learning_rate": 4.934497570851788e-05, "loss": 0.5182, "num_input_tokens_seen": 4150448, "step": 6325 }, { "epoch": 3.3176100628930816, "grad_norm": 0.1383526772260666, "learning_rate": 4.9342372887898804e-05, "loss": 0.4295, "num_input_tokens_seen": 4155504, "step": 6330 }, { "epoch": 3.320230607966457, "grad_norm": 0.09064556658267975, "learning_rate": 4.9339764975133026e-05, "loss": 0.4444, "num_input_tokens_seen": 4158800, "step": 6335 }, { "epoch": 3.3228511530398324, "grad_norm": 0.11276407539844513, "learning_rate": 4.93371519707661e-05, "loss": 0.4439, "num_input_tokens_seen": 4162832, "step": 6340 }, { "epoch": 3.3254716981132075, "grad_norm": 0.15875302255153656, "learning_rate": 4.933453387534464e-05, "loss": 0.5076, "num_input_tokens_seen": 4166544, "step": 6345 }, { "epoch": 3.3280922431865827, "grad_norm": 0.19022630155086517, "learning_rate": 4.933191068941631e-05, "loss": 0.7106, "num_input_tokens_seen": 4170448, "step": 6350 }, { "epoch": 3.330712788259958, "grad_norm": 0.17708256840705872, "learning_rate": 4.932928241352987e-05, "loss": 0.5677, "num_input_tokens_seen": 4174160, "step": 6355 }, { "epoch": 3.3333333333333335, "grad_norm": 0.2514781355857849, "learning_rate": 4.9326649048235095e-05, "loss": 0.512, "num_input_tokens_seen": 4177584, "step": 6360 }, { "epoch": 3.3359538784067087, "grad_norm": 0.220132514834404, "learning_rate": 4.932401059408287e-05, "loss": 0.4015, "num_input_tokens_seen": 4179472, "step": 6365 }, { "epoch": 3.338574423480084, "grad_norm": 0.25650718808174133, "learning_rate": 4.932136705162514e-05, "loss": 0.6051, "num_input_tokens_seen": 4182192, "step": 6370 }, { "epoch": 3.341194968553459, "grad_norm": 0.24722304940223694, "learning_rate": 4.93187184214149e-05, "loss": 0.3907, "num_input_tokens_seen": 4185040, "step": 6375 }, { "epoch": 3.3438155136268346, "grad_norm": 0.18112453818321228, "learning_rate": 4.9316064704006205e-05, "loss": 0.5653, "num_input_tokens_seen": 4187952, "step": 6380 }, { "epoch": 3.3464360587002098, "grad_norm": 0.17602728307247162, "learning_rate": 4.931340589995419e-05, "loss": 0.5396, "num_input_tokens_seen": 4191184, "step": 6385 }, { "epoch": 3.349056603773585, "grad_norm": 0.17230212688446045, "learning_rate": 4.9310742009815035e-05, "loss": 0.5686, "num_input_tokens_seen": 4194064, "step": 6390 }, { "epoch": 3.35167714884696, "grad_norm": 0.11852315813302994, "learning_rate": 4.930807303414601e-05, "loss": 0.4454, "num_input_tokens_seen": 4196688, "step": 6395 }, { "epoch": 3.3542976939203353, "grad_norm": 0.15718930959701538, "learning_rate": 4.930539897350542e-05, "loss": 0.4341, "num_input_tokens_seen": 4199952, "step": 6400 }, { "epoch": 3.3569182389937104, "grad_norm": 0.1827602982521057, "learning_rate": 4.930271982845266e-05, "loss": 0.4737, "num_input_tokens_seen": 4203792, "step": 6405 }, { "epoch": 3.359538784067086, "grad_norm": 0.18453091382980347, "learning_rate": 4.930003559954817e-05, "loss": 0.4085, "num_input_tokens_seen": 4206736, "step": 6410 }, { "epoch": 3.3621593291404612, "grad_norm": 0.15242624282836914, "learning_rate": 4.929734628735346e-05, "loss": 0.5988, "num_input_tokens_seen": 4210032, "step": 6415 }, { "epoch": 3.3647798742138364, "grad_norm": 0.23421572148799896, "learning_rate": 4.9294651892431115e-05, "loss": 0.5283, "num_input_tokens_seen": 4212848, "step": 6420 }, { "epoch": 3.3674004192872116, "grad_norm": 0.4140131175518036, "learning_rate": 4.929195241534476e-05, "loss": 0.6558, "num_input_tokens_seen": 4215376, "step": 6425 }, { "epoch": 3.370020964360587, "grad_norm": 0.2540588974952698, "learning_rate": 4.9289247856659095e-05, "loss": 0.5646, "num_input_tokens_seen": 4218288, "step": 6430 }, { "epoch": 3.3726415094339623, "grad_norm": 0.17372798919677734, "learning_rate": 4.9286538216939883e-05, "loss": 0.44, "num_input_tokens_seen": 4221488, "step": 6435 }, { "epoch": 3.3752620545073375, "grad_norm": 0.1353875994682312, "learning_rate": 4.9283823496753944e-05, "loss": 0.859, "num_input_tokens_seen": 4224432, "step": 6440 }, { "epoch": 3.3778825995807127, "grad_norm": 0.28958743810653687, "learning_rate": 4.928110369666918e-05, "loss": 0.3533, "num_input_tokens_seen": 4229296, "step": 6445 }, { "epoch": 3.380503144654088, "grad_norm": 0.21435360610485077, "learning_rate": 4.927837881725455e-05, "loss": 0.5922, "num_input_tokens_seen": 4232016, "step": 6450 }, { "epoch": 3.3831236897274635, "grad_norm": 0.19161035120487213, "learning_rate": 4.927564885908003e-05, "loss": 0.4292, "num_input_tokens_seen": 4235248, "step": 6455 }, { "epoch": 3.3857442348008386, "grad_norm": 0.18078097701072693, "learning_rate": 4.9272913822716735e-05, "loss": 0.5347, "num_input_tokens_seen": 4238544, "step": 6460 }, { "epoch": 3.388364779874214, "grad_norm": 0.28528016805648804, "learning_rate": 4.927017370873678e-05, "loss": 0.5847, "num_input_tokens_seen": 4241648, "step": 6465 }, { "epoch": 3.390985324947589, "grad_norm": 0.11260097473859787, "learning_rate": 4.926742851771338e-05, "loss": 0.5618, "num_input_tokens_seen": 4245520, "step": 6470 }, { "epoch": 3.3936058700209646, "grad_norm": 0.21692979335784912, "learning_rate": 4.926467825022079e-05, "loss": 0.5538, "num_input_tokens_seen": 4248400, "step": 6475 }, { "epoch": 3.3962264150943398, "grad_norm": 0.17743760347366333, "learning_rate": 4.926192290683434e-05, "loss": 0.4764, "num_input_tokens_seen": 4251568, "step": 6480 }, { "epoch": 3.398846960167715, "grad_norm": 0.1113213449716568, "learning_rate": 4.9259162488130404e-05, "loss": 0.4776, "num_input_tokens_seen": 4255888, "step": 6485 }, { "epoch": 3.40146750524109, "grad_norm": 0.18133027851581573, "learning_rate": 4.925639699468645e-05, "loss": 0.6169, "num_input_tokens_seen": 4258960, "step": 6490 }, { "epoch": 3.4040880503144653, "grad_norm": 0.2783883213996887, "learning_rate": 4.925362642708097e-05, "loss": 0.511, "num_input_tokens_seen": 4262256, "step": 6495 }, { "epoch": 3.4067085953878404, "grad_norm": 0.142732173204422, "learning_rate": 4.9250850785893536e-05, "loss": 0.4594, "num_input_tokens_seen": 4266352, "step": 6500 }, { "epoch": 3.409329140461216, "grad_norm": 0.13237152993679047, "learning_rate": 4.924807007170479e-05, "loss": 0.4787, "num_input_tokens_seen": 4269136, "step": 6505 }, { "epoch": 3.411949685534591, "grad_norm": 0.1519063413143158, "learning_rate": 4.924528428509643e-05, "loss": 0.4544, "num_input_tokens_seen": 4271536, "step": 6510 }, { "epoch": 3.4145702306079664, "grad_norm": 0.21016152203083038, "learning_rate": 4.9242493426651196e-05, "loss": 0.4638, "num_input_tokens_seen": 4275248, "step": 6515 }, { "epoch": 3.4171907756813416, "grad_norm": 0.17191898822784424, "learning_rate": 4.92396974969529e-05, "loss": 0.4982, "num_input_tokens_seen": 4278608, "step": 6520 }, { "epoch": 3.419811320754717, "grad_norm": 0.18220838904380798, "learning_rate": 4.923689649658644e-05, "loss": 0.6417, "num_input_tokens_seen": 4282064, "step": 6525 }, { "epoch": 3.4224318658280923, "grad_norm": 0.27501389384269714, "learning_rate": 4.923409042613774e-05, "loss": 0.4987, "num_input_tokens_seen": 4285872, "step": 6530 }, { "epoch": 3.4250524109014675, "grad_norm": 0.10501659661531448, "learning_rate": 4.92312792861938e-05, "loss": 0.4557, "num_input_tokens_seen": 4288816, "step": 6535 }, { "epoch": 3.4276729559748427, "grad_norm": 0.1728868931531906, "learning_rate": 4.922846307734268e-05, "loss": 0.5479, "num_input_tokens_seen": 4292304, "step": 6540 }, { "epoch": 3.430293501048218, "grad_norm": 0.15741406381130219, "learning_rate": 4.9225641800173496e-05, "loss": 0.5451, "num_input_tokens_seen": 4295504, "step": 6545 }, { "epoch": 3.4329140461215935, "grad_norm": 0.25156885385513306, "learning_rate": 4.922281545527643e-05, "loss": 0.4774, "num_input_tokens_seen": 4298160, "step": 6550 }, { "epoch": 3.4355345911949686, "grad_norm": 0.2612077593803406, "learning_rate": 4.9219984043242715e-05, "loss": 0.5056, "num_input_tokens_seen": 4300944, "step": 6555 }, { "epoch": 3.438155136268344, "grad_norm": 0.25148406624794006, "learning_rate": 4.921714756466466e-05, "loss": 0.55, "num_input_tokens_seen": 4303600, "step": 6560 }, { "epoch": 3.440775681341719, "grad_norm": 0.3354698121547699, "learning_rate": 4.921430602013561e-05, "loss": 0.4434, "num_input_tokens_seen": 4306256, "step": 6565 }, { "epoch": 3.4433962264150946, "grad_norm": 0.17189595103263855, "learning_rate": 4.9211459410249995e-05, "loss": 0.4823, "num_input_tokens_seen": 4309264, "step": 6570 }, { "epoch": 3.4460167714884697, "grad_norm": 0.22063276171684265, "learning_rate": 4.92086077356033e-05, "loss": 0.5772, "num_input_tokens_seen": 4312176, "step": 6575 }, { "epoch": 3.448637316561845, "grad_norm": 0.13574501872062683, "learning_rate": 4.920575099679204e-05, "loss": 0.6004, "num_input_tokens_seen": 4317488, "step": 6580 }, { "epoch": 3.45125786163522, "grad_norm": 0.1872745156288147, "learning_rate": 4.920288919441383e-05, "loss": 0.4307, "num_input_tokens_seen": 4320592, "step": 6585 }, { "epoch": 3.4538784067085953, "grad_norm": 0.18471281230449677, "learning_rate": 4.9200022329067316e-05, "loss": 0.5445, "num_input_tokens_seen": 4323632, "step": 6590 }, { "epoch": 3.4564989517819704, "grad_norm": 0.1400659829378128, "learning_rate": 4.9197150401352224e-05, "loss": 0.374, "num_input_tokens_seen": 4326576, "step": 6595 }, { "epoch": 3.459119496855346, "grad_norm": 0.44161519408226013, "learning_rate": 4.919427341186933e-05, "loss": 0.5773, "num_input_tokens_seen": 4329424, "step": 6600 }, { "epoch": 3.461740041928721, "grad_norm": 0.15890827775001526, "learning_rate": 4.919139136122045e-05, "loss": 0.7118, "num_input_tokens_seen": 4332656, "step": 6605 }, { "epoch": 3.4643605870020964, "grad_norm": 0.28854241967201233, "learning_rate": 4.918850425000849e-05, "loss": 0.5664, "num_input_tokens_seen": 4335536, "step": 6610 }, { "epoch": 3.4669811320754715, "grad_norm": 0.10695409029722214, "learning_rate": 4.918561207883741e-05, "loss": 0.584, "num_input_tokens_seen": 4339248, "step": 6615 }, { "epoch": 3.469601677148847, "grad_norm": 0.1584293693304062, "learning_rate": 4.9182714848312195e-05, "loss": 0.5398, "num_input_tokens_seen": 4342320, "step": 6620 }, { "epoch": 3.4722222222222223, "grad_norm": 0.16300374269485474, "learning_rate": 4.917981255903893e-05, "loss": 0.5125, "num_input_tokens_seen": 4346256, "step": 6625 }, { "epoch": 3.4748427672955975, "grad_norm": 0.19719192385673523, "learning_rate": 4.917690521162474e-05, "loss": 0.6419, "num_input_tokens_seen": 4349872, "step": 6630 }, { "epoch": 3.4774633123689727, "grad_norm": 0.1810423731803894, "learning_rate": 4.9173992806677796e-05, "loss": 0.5119, "num_input_tokens_seen": 4352368, "step": 6635 }, { "epoch": 3.480083857442348, "grad_norm": 0.16418714821338654, "learning_rate": 4.917107534480735e-05, "loss": 0.513, "num_input_tokens_seen": 4356880, "step": 6640 }, { "epoch": 3.4827044025157234, "grad_norm": 0.18621289730072021, "learning_rate": 4.916815282662369e-05, "loss": 0.3418, "num_input_tokens_seen": 4360688, "step": 6645 }, { "epoch": 3.4853249475890986, "grad_norm": 0.16049689054489136, "learning_rate": 4.91652252527382e-05, "loss": 0.4521, "num_input_tokens_seen": 4363856, "step": 6650 }, { "epoch": 3.487945492662474, "grad_norm": 0.10946711152791977, "learning_rate": 4.9162292623763275e-05, "loss": 0.5098, "num_input_tokens_seen": 4366640, "step": 6655 }, { "epoch": 3.490566037735849, "grad_norm": 0.19659699499607086, "learning_rate": 4.9159354940312386e-05, "loss": 0.43, "num_input_tokens_seen": 4369584, "step": 6660 }, { "epoch": 3.4931865828092246, "grad_norm": 0.19369108974933624, "learning_rate": 4.915641220300007e-05, "loss": 0.5728, "num_input_tokens_seen": 4372752, "step": 6665 }, { "epoch": 3.4958071278825997, "grad_norm": 0.15275268256664276, "learning_rate": 4.9153464412441915e-05, "loss": 0.5241, "num_input_tokens_seen": 4375760, "step": 6670 }, { "epoch": 3.498427672955975, "grad_norm": 0.16321265697479248, "learning_rate": 4.915051156925456e-05, "loss": 0.4931, "num_input_tokens_seen": 4379248, "step": 6675 }, { "epoch": 3.50104821802935, "grad_norm": 0.16619443893432617, "learning_rate": 4.91475536740557e-05, "loss": 0.7103, "num_input_tokens_seen": 4382352, "step": 6680 }, { "epoch": 3.5036687631027252, "grad_norm": 0.21931982040405273, "learning_rate": 4.914459072746411e-05, "loss": 0.6701, "num_input_tokens_seen": 4385968, "step": 6685 }, { "epoch": 3.5062893081761004, "grad_norm": 0.2839338779449463, "learning_rate": 4.91416227300996e-05, "loss": 0.5686, "num_input_tokens_seen": 4388720, "step": 6690 }, { "epoch": 3.508909853249476, "grad_norm": 0.1938534826040268, "learning_rate": 4.913864968258303e-05, "loss": 0.539, "num_input_tokens_seen": 4391568, "step": 6695 }, { "epoch": 3.511530398322851, "grad_norm": 0.19524478912353516, "learning_rate": 4.913567158553634e-05, "loss": 0.5622, "num_input_tokens_seen": 4394256, "step": 6700 }, { "epoch": 3.5141509433962264, "grad_norm": 0.2788836359977722, "learning_rate": 4.9132688439582505e-05, "loss": 0.5147, "num_input_tokens_seen": 4396816, "step": 6705 }, { "epoch": 3.5167714884696015, "grad_norm": 0.3361208140850067, "learning_rate": 4.9129700245345565e-05, "loss": 0.5874, "num_input_tokens_seen": 4399952, "step": 6710 }, { "epoch": 3.519392033542977, "grad_norm": 0.12347142398357391, "learning_rate": 4.912670700345063e-05, "loss": 0.5189, "num_input_tokens_seen": 4403440, "step": 6715 }, { "epoch": 3.5220125786163523, "grad_norm": 0.1158863827586174, "learning_rate": 4.912370871452383e-05, "loss": 0.5886, "num_input_tokens_seen": 4406736, "step": 6720 }, { "epoch": 3.5246331236897275, "grad_norm": 0.11572165787220001, "learning_rate": 4.91207053791924e-05, "loss": 0.4903, "num_input_tokens_seen": 4410640, "step": 6725 }, { "epoch": 3.5272536687631026, "grad_norm": 0.2893217206001282, "learning_rate": 4.911769699808457e-05, "loss": 0.4471, "num_input_tokens_seen": 4414096, "step": 6730 }, { "epoch": 3.529874213836478, "grad_norm": 0.15841157734394073, "learning_rate": 4.911468357182969e-05, "loss": 0.4857, "num_input_tokens_seen": 4416976, "step": 6735 }, { "epoch": 3.532494758909853, "grad_norm": 0.11231611669063568, "learning_rate": 4.9111665101058116e-05, "loss": 0.4935, "num_input_tokens_seen": 4419792, "step": 6740 }, { "epoch": 3.5351153039832286, "grad_norm": 0.14839808642864227, "learning_rate": 4.910864158640128e-05, "loss": 0.5212, "num_input_tokens_seen": 4422928, "step": 6745 }, { "epoch": 3.5377358490566038, "grad_norm": 0.19096986949443817, "learning_rate": 4.910561302849168e-05, "loss": 0.6728, "num_input_tokens_seen": 4426576, "step": 6750 }, { "epoch": 3.540356394129979, "grad_norm": 0.16021305322647095, "learning_rate": 4.910257942796284e-05, "loss": 0.56, "num_input_tokens_seen": 4431408, "step": 6755 }, { "epoch": 3.5429769392033545, "grad_norm": 0.1747269481420517, "learning_rate": 4.909954078544936e-05, "loss": 0.6306, "num_input_tokens_seen": 4434352, "step": 6760 }, { "epoch": 3.5455974842767297, "grad_norm": 0.12563489377498627, "learning_rate": 4.909649710158689e-05, "loss": 0.5012, "num_input_tokens_seen": 4437488, "step": 6765 }, { "epoch": 3.548218029350105, "grad_norm": 0.2712559998035431, "learning_rate": 4.909344837701214e-05, "loss": 0.4452, "num_input_tokens_seen": 4440208, "step": 6770 }, { "epoch": 3.55083857442348, "grad_norm": 0.21342730522155762, "learning_rate": 4.909039461236285e-05, "loss": 0.5397, "num_input_tokens_seen": 4443024, "step": 6775 }, { "epoch": 3.5534591194968552, "grad_norm": 0.1755785048007965, "learning_rate": 4.908733580827786e-05, "loss": 0.6428, "num_input_tokens_seen": 4446512, "step": 6780 }, { "epoch": 3.5560796645702304, "grad_norm": 0.22737132012844086, "learning_rate": 4.9084271965397014e-05, "loss": 0.4715, "num_input_tokens_seen": 4449360, "step": 6785 }, { "epoch": 3.558700209643606, "grad_norm": 0.10897070914506912, "learning_rate": 4.908120308436124e-05, "loss": 0.5253, "num_input_tokens_seen": 4451728, "step": 6790 }, { "epoch": 3.561320754716981, "grad_norm": 0.143552765250206, "learning_rate": 4.907812916581251e-05, "loss": 0.4494, "num_input_tokens_seen": 4455568, "step": 6795 }, { "epoch": 3.5639412997903563, "grad_norm": 0.24465201795101166, "learning_rate": 4.907505021039386e-05, "loss": 0.5052, "num_input_tokens_seen": 4458768, "step": 6800 }, { "epoch": 3.5665618448637315, "grad_norm": 0.14982566237449646, "learning_rate": 4.907196621874937e-05, "loss": 0.5642, "num_input_tokens_seen": 4461648, "step": 6805 }, { "epoch": 3.569182389937107, "grad_norm": 0.21367554366588593, "learning_rate": 4.9068877191524176e-05, "loss": 0.5624, "num_input_tokens_seen": 4465232, "step": 6810 }, { "epoch": 3.5718029350104823, "grad_norm": 0.1830551028251648, "learning_rate": 4.906578312936446e-05, "loss": 0.5194, "num_input_tokens_seen": 4467984, "step": 6815 }, { "epoch": 3.5744234800838575, "grad_norm": 0.16858543455600739, "learning_rate": 4.906268403291747e-05, "loss": 0.4052, "num_input_tokens_seen": 4471088, "step": 6820 }, { "epoch": 3.5770440251572326, "grad_norm": 0.17479045689105988, "learning_rate": 4.905957990283151e-05, "loss": 0.6267, "num_input_tokens_seen": 4474128, "step": 6825 }, { "epoch": 3.579664570230608, "grad_norm": 0.16074928641319275, "learning_rate": 4.905647073975592e-05, "loss": 0.72, "num_input_tokens_seen": 4478160, "step": 6830 }, { "epoch": 3.582285115303983, "grad_norm": 0.16217604279518127, "learning_rate": 4.9053356544341095e-05, "loss": 0.5512, "num_input_tokens_seen": 4480784, "step": 6835 }, { "epoch": 3.5849056603773586, "grad_norm": 0.205156147480011, "learning_rate": 4.90502373172385e-05, "loss": 0.6445, "num_input_tokens_seen": 4483632, "step": 6840 }, { "epoch": 3.5875262054507338, "grad_norm": 0.1598014235496521, "learning_rate": 4.9047113059100635e-05, "loss": 0.5302, "num_input_tokens_seen": 4486448, "step": 6845 }, { "epoch": 3.590146750524109, "grad_norm": 0.16489174962043762, "learning_rate": 4.904398377058106e-05, "loss": 0.3978, "num_input_tokens_seen": 4490192, "step": 6850 }, { "epoch": 3.5927672955974845, "grad_norm": 0.36596688628196716, "learning_rate": 4.90408494523344e-05, "loss": 0.5978, "num_input_tokens_seen": 4492880, "step": 6855 }, { "epoch": 3.5953878406708597, "grad_norm": 0.22714775800704956, "learning_rate": 4.90377101050163e-05, "loss": 0.5041, "num_input_tokens_seen": 4495952, "step": 6860 }, { "epoch": 3.598008385744235, "grad_norm": 0.16810445487499237, "learning_rate": 4.903456572928348e-05, "loss": 0.554, "num_input_tokens_seen": 4498512, "step": 6865 }, { "epoch": 3.60062893081761, "grad_norm": 0.1552681028842926, "learning_rate": 4.903141632579371e-05, "loss": 0.4867, "num_input_tokens_seen": 4503152, "step": 6870 }, { "epoch": 3.603249475890985, "grad_norm": 0.36869046092033386, "learning_rate": 4.902826189520581e-05, "loss": 0.5748, "num_input_tokens_seen": 4506128, "step": 6875 }, { "epoch": 3.6058700209643604, "grad_norm": 0.2956083416938782, "learning_rate": 4.9025102438179656e-05, "loss": 0.5297, "num_input_tokens_seen": 4509776, "step": 6880 }, { "epoch": 3.608490566037736, "grad_norm": 0.2014244794845581, "learning_rate": 4.902193795537616e-05, "loss": 0.6293, "num_input_tokens_seen": 4513136, "step": 6885 }, { "epoch": 3.611111111111111, "grad_norm": 0.141885906457901, "learning_rate": 4.90187684474573e-05, "loss": 0.497, "num_input_tokens_seen": 4516176, "step": 6890 }, { "epoch": 3.6137316561844863, "grad_norm": 0.19754330813884735, "learning_rate": 4.901559391508611e-05, "loss": 0.5928, "num_input_tokens_seen": 4519120, "step": 6895 }, { "epoch": 3.6163522012578615, "grad_norm": 0.24283690750598907, "learning_rate": 4.9012414358926644e-05, "loss": 0.4835, "num_input_tokens_seen": 4522352, "step": 6900 }, { "epoch": 3.618972746331237, "grad_norm": 0.16481268405914307, "learning_rate": 4.900922977964405e-05, "loss": 0.5619, "num_input_tokens_seen": 4525104, "step": 6905 }, { "epoch": 3.6215932914046123, "grad_norm": 0.13499976694583893, "learning_rate": 4.90060401779045e-05, "loss": 0.5469, "num_input_tokens_seen": 4528528, "step": 6910 }, { "epoch": 3.6242138364779874, "grad_norm": 0.20782619714736938, "learning_rate": 4.900284555437521e-05, "loss": 0.4006, "num_input_tokens_seen": 4531312, "step": 6915 }, { "epoch": 3.6268343815513626, "grad_norm": 0.377023845911026, "learning_rate": 4.899964590972448e-05, "loss": 0.6774, "num_input_tokens_seen": 4534128, "step": 6920 }, { "epoch": 3.629454926624738, "grad_norm": 0.1387571543455124, "learning_rate": 4.899644124462163e-05, "loss": 0.593, "num_input_tokens_seen": 4537168, "step": 6925 }, { "epoch": 3.632075471698113, "grad_norm": 0.18171268701553345, "learning_rate": 4.8993231559737026e-05, "loss": 0.7911, "num_input_tokens_seen": 4540464, "step": 6930 }, { "epoch": 3.6346960167714886, "grad_norm": 0.14784489572048187, "learning_rate": 4.8990016855742114e-05, "loss": 0.4713, "num_input_tokens_seen": 4544208, "step": 6935 }, { "epoch": 3.6373165618448637, "grad_norm": 0.6113256812095642, "learning_rate": 4.898679713330937e-05, "loss": 0.4821, "num_input_tokens_seen": 4547440, "step": 6940 }, { "epoch": 3.639937106918239, "grad_norm": 0.5736700296401978, "learning_rate": 4.898357239311232e-05, "loss": 0.5423, "num_input_tokens_seen": 4550512, "step": 6945 }, { "epoch": 3.6425576519916145, "grad_norm": 0.24286064505577087, "learning_rate": 4.898034263582554e-05, "loss": 0.7466, "num_input_tokens_seen": 4553168, "step": 6950 }, { "epoch": 3.6451781970649897, "grad_norm": 0.1540730744600296, "learning_rate": 4.897710786212467e-05, "loss": 0.526, "num_input_tokens_seen": 4556720, "step": 6955 }, { "epoch": 3.647798742138365, "grad_norm": 0.1567969024181366, "learning_rate": 4.897386807268638e-05, "loss": 0.7147, "num_input_tokens_seen": 4560432, "step": 6960 }, { "epoch": 3.65041928721174, "grad_norm": 0.1460934579372406, "learning_rate": 4.897062326818839e-05, "loss": 0.6005, "num_input_tokens_seen": 4563792, "step": 6965 }, { "epoch": 3.653039832285115, "grad_norm": 0.15087725222110748, "learning_rate": 4.8967373449309495e-05, "loss": 0.5059, "num_input_tokens_seen": 4568144, "step": 6970 }, { "epoch": 3.6556603773584904, "grad_norm": 0.23211058974266052, "learning_rate": 4.89641186167295e-05, "loss": 0.4913, "num_input_tokens_seen": 4570448, "step": 6975 }, { "epoch": 3.658280922431866, "grad_norm": 0.25098755955696106, "learning_rate": 4.8960858771129294e-05, "loss": 0.4574, "num_input_tokens_seen": 4573232, "step": 6980 }, { "epoch": 3.660901467505241, "grad_norm": 0.38766220211982727, "learning_rate": 4.895759391319079e-05, "loss": 0.6105, "num_input_tokens_seen": 4576784, "step": 6985 }, { "epoch": 3.6635220125786163, "grad_norm": 0.12841305136680603, "learning_rate": 4.8954324043596954e-05, "loss": 0.5901, "num_input_tokens_seen": 4580592, "step": 6990 }, { "epoch": 3.6661425576519915, "grad_norm": 0.18340517580509186, "learning_rate": 4.8951049163031826e-05, "loss": 0.5281, "num_input_tokens_seen": 4583792, "step": 6995 }, { "epoch": 3.668763102725367, "grad_norm": 0.6166594624519348, "learning_rate": 4.894776927218045e-05, "loss": 0.7025, "num_input_tokens_seen": 4586896, "step": 7000 }, { "epoch": 3.6713836477987423, "grad_norm": 0.10219574719667435, "learning_rate": 4.894448437172896e-05, "loss": 0.4507, "num_input_tokens_seen": 4590096, "step": 7005 }, { "epoch": 3.6740041928721174, "grad_norm": 0.1190553680062294, "learning_rate": 4.894119446236452e-05, "loss": 0.3979, "num_input_tokens_seen": 4592976, "step": 7010 }, { "epoch": 3.6766247379454926, "grad_norm": 0.14898599684238434, "learning_rate": 4.893789954477532e-05, "loss": 0.5829, "num_input_tokens_seen": 4595664, "step": 7015 }, { "epoch": 3.6792452830188678, "grad_norm": 0.19838397204875946, "learning_rate": 4.893459961965063e-05, "loss": 0.6363, "num_input_tokens_seen": 4598320, "step": 7020 }, { "epoch": 3.681865828092243, "grad_norm": 0.16015270352363586, "learning_rate": 4.8931294687680775e-05, "loss": 0.3708, "num_input_tokens_seen": 4604976, "step": 7025 }, { "epoch": 3.6844863731656186, "grad_norm": 0.127739816904068, "learning_rate": 4.8927984749557086e-05, "loss": 0.4948, "num_input_tokens_seen": 4608528, "step": 7030 }, { "epoch": 3.6871069182389937, "grad_norm": 0.16333191096782684, "learning_rate": 4.892466980597198e-05, "loss": 0.4712, "num_input_tokens_seen": 4611536, "step": 7035 }, { "epoch": 3.689727463312369, "grad_norm": 0.14546242356300354, "learning_rate": 4.892134985761889e-05, "loss": 0.5172, "num_input_tokens_seen": 4614544, "step": 7040 }, { "epoch": 3.6923480083857445, "grad_norm": 0.17633415758609772, "learning_rate": 4.891802490519232e-05, "loss": 0.4896, "num_input_tokens_seen": 4617424, "step": 7045 }, { "epoch": 3.6949685534591197, "grad_norm": 0.18194855749607086, "learning_rate": 4.891469494938781e-05, "loss": 0.4204, "num_input_tokens_seen": 4620112, "step": 7050 }, { "epoch": 3.697589098532495, "grad_norm": 0.10707549005746841, "learning_rate": 4.891135999090195e-05, "loss": 0.575, "num_input_tokens_seen": 4623888, "step": 7055 }, { "epoch": 3.70020964360587, "grad_norm": 0.1660548448562622, "learning_rate": 4.890802003043238e-05, "loss": 0.5273, "num_input_tokens_seen": 4627696, "step": 7060 }, { "epoch": 3.702830188679245, "grad_norm": 0.09169525653123856, "learning_rate": 4.890467506867777e-05, "loss": 0.4902, "num_input_tokens_seen": 4630288, "step": 7065 }, { "epoch": 3.7054507337526204, "grad_norm": 0.1615528017282486, "learning_rate": 4.890132510633786e-05, "loss": 0.566, "num_input_tokens_seen": 4633168, "step": 7070 }, { "epoch": 3.708071278825996, "grad_norm": 0.27896541357040405, "learning_rate": 4.8897970144113416e-05, "loss": 0.5993, "num_input_tokens_seen": 4635984, "step": 7075 }, { "epoch": 3.710691823899371, "grad_norm": 0.34270530939102173, "learning_rate": 4.889461018270626e-05, "loss": 0.6194, "num_input_tokens_seen": 4639312, "step": 7080 }, { "epoch": 3.7133123689727463, "grad_norm": 0.1265009641647339, "learning_rate": 4.889124522281926e-05, "loss": 0.5604, "num_input_tokens_seen": 4642320, "step": 7085 }, { "epoch": 3.7159329140461215, "grad_norm": 0.11997382342815399, "learning_rate": 4.888787526515632e-05, "loss": 0.561, "num_input_tokens_seen": 4645040, "step": 7090 }, { "epoch": 3.718553459119497, "grad_norm": 0.20635554194450378, "learning_rate": 4.8884500310422406e-05, "loss": 0.5448, "num_input_tokens_seen": 4647856, "step": 7095 }, { "epoch": 3.7211740041928723, "grad_norm": 0.6667801141738892, "learning_rate": 4.8881120359323516e-05, "loss": 0.5051, "num_input_tokens_seen": 4651280, "step": 7100 }, { "epoch": 3.7237945492662474, "grad_norm": 0.16490492224693298, "learning_rate": 4.887773541256669e-05, "loss": 0.438, "num_input_tokens_seen": 4654512, "step": 7105 }, { "epoch": 3.7264150943396226, "grad_norm": 0.2622659206390381, "learning_rate": 4.887434547086004e-05, "loss": 0.4782, "num_input_tokens_seen": 4657456, "step": 7110 }, { "epoch": 3.7290356394129978, "grad_norm": 0.2613312900066376, "learning_rate": 4.887095053491269e-05, "loss": 0.3789, "num_input_tokens_seen": 4660560, "step": 7115 }, { "epoch": 3.731656184486373, "grad_norm": 0.20466020703315735, "learning_rate": 4.8867550605434806e-05, "loss": 0.6596, "num_input_tokens_seen": 4663344, "step": 7120 }, { "epoch": 3.7342767295597485, "grad_norm": 0.225444495677948, "learning_rate": 4.8864145683137644e-05, "loss": 0.5643, "num_input_tokens_seen": 4666384, "step": 7125 }, { "epoch": 3.7368972746331237, "grad_norm": 0.1826229840517044, "learning_rate": 4.886073576873346e-05, "loss": 0.4556, "num_input_tokens_seen": 4668880, "step": 7130 }, { "epoch": 3.739517819706499, "grad_norm": 0.1017305850982666, "learning_rate": 4.885732086293556e-05, "loss": 0.6114, "num_input_tokens_seen": 4672368, "step": 7135 }, { "epoch": 3.742138364779874, "grad_norm": 0.17886823415756226, "learning_rate": 4.8853900966458335e-05, "loss": 0.5257, "num_input_tokens_seen": 4676368, "step": 7140 }, { "epoch": 3.7447589098532497, "grad_norm": 0.17012429237365723, "learning_rate": 4.8850476080017146e-05, "loss": 0.4669, "num_input_tokens_seen": 4679632, "step": 7145 }, { "epoch": 3.747379454926625, "grad_norm": 0.17926327884197235, "learning_rate": 4.884704620432848e-05, "loss": 0.4549, "num_input_tokens_seen": 4683120, "step": 7150 }, { "epoch": 3.75, "grad_norm": 0.13483396172523499, "learning_rate": 4.8843611340109794e-05, "loss": 0.6198, "num_input_tokens_seen": 4687504, "step": 7155 }, { "epoch": 3.752620545073375, "grad_norm": 0.09934893995523453, "learning_rate": 4.884017148807965e-05, "loss": 0.4557, "num_input_tokens_seen": 4690448, "step": 7160 }, { "epoch": 3.7552410901467503, "grad_norm": 0.18604789674282074, "learning_rate": 4.883672664895761e-05, "loss": 0.5352, "num_input_tokens_seen": 4693104, "step": 7165 }, { "epoch": 3.757861635220126, "grad_norm": 0.148334801197052, "learning_rate": 4.8833276823464296e-05, "loss": 0.3596, "num_input_tokens_seen": 4696080, "step": 7170 }, { "epoch": 3.760482180293501, "grad_norm": 0.14081011712551117, "learning_rate": 4.882982201232138e-05, "loss": 0.5582, "num_input_tokens_seen": 4698416, "step": 7175 }, { "epoch": 3.7631027253668763, "grad_norm": 0.18319545686244965, "learning_rate": 4.882636221625156e-05, "loss": 0.5192, "num_input_tokens_seen": 4701072, "step": 7180 }, { "epoch": 3.7657232704402515, "grad_norm": 0.16744889318943024, "learning_rate": 4.8822897435978586e-05, "loss": 0.594, "num_input_tokens_seen": 4703696, "step": 7185 }, { "epoch": 3.768343815513627, "grad_norm": 0.3140692114830017, "learning_rate": 4.8819427672227264e-05, "loss": 0.443, "num_input_tokens_seen": 4706288, "step": 7190 }, { "epoch": 3.7709643605870022, "grad_norm": 0.19568617641925812, "learning_rate": 4.881595292572341e-05, "loss": 0.5796, "num_input_tokens_seen": 4709040, "step": 7195 }, { "epoch": 3.7735849056603774, "grad_norm": 0.16747692227363586, "learning_rate": 4.881247319719392e-05, "loss": 0.4872, "num_input_tokens_seen": 4712944, "step": 7200 }, { "epoch": 3.7762054507337526, "grad_norm": 0.21050460636615753, "learning_rate": 4.88089884873667e-05, "loss": 0.4609, "num_input_tokens_seen": 4715664, "step": 7205 }, { "epoch": 3.7788259958071277, "grad_norm": 0.444245845079422, "learning_rate": 4.880549879697071e-05, "loss": 0.4807, "num_input_tokens_seen": 4718544, "step": 7210 }, { "epoch": 3.781446540880503, "grad_norm": 0.1564105600118637, "learning_rate": 4.8802004126735965e-05, "loss": 0.5112, "num_input_tokens_seen": 4721584, "step": 7215 }, { "epoch": 3.7840670859538785, "grad_norm": 0.20267651975154877, "learning_rate": 4.87985044773935e-05, "loss": 0.6359, "num_input_tokens_seen": 4724720, "step": 7220 }, { "epoch": 3.7866876310272537, "grad_norm": 0.37565478682518005, "learning_rate": 4.879499984967541e-05, "loss": 0.5378, "num_input_tokens_seen": 4727696, "step": 7225 }, { "epoch": 3.789308176100629, "grad_norm": 0.2214682698249817, "learning_rate": 4.8791490244314824e-05, "loss": 0.5435, "num_input_tokens_seen": 4730832, "step": 7230 }, { "epoch": 3.791928721174004, "grad_norm": 0.15963034331798553, "learning_rate": 4.8787975662045904e-05, "loss": 0.5765, "num_input_tokens_seen": 4733456, "step": 7235 }, { "epoch": 3.7945492662473796, "grad_norm": 0.13075217604637146, "learning_rate": 4.878445610360386e-05, "loss": 0.5536, "num_input_tokens_seen": 4737008, "step": 7240 }, { "epoch": 3.797169811320755, "grad_norm": 0.15781599283218384, "learning_rate": 4.878093156972494e-05, "loss": 0.8207, "num_input_tokens_seen": 4740560, "step": 7245 }, { "epoch": 3.79979035639413, "grad_norm": 0.12160549312829971, "learning_rate": 4.877740206114645e-05, "loss": 0.4804, "num_input_tokens_seen": 4743536, "step": 7250 }, { "epoch": 3.802410901467505, "grad_norm": 0.08091877400875092, "learning_rate": 4.8773867578606716e-05, "loss": 0.427, "num_input_tokens_seen": 4748144, "step": 7255 }, { "epoch": 3.8050314465408803, "grad_norm": 0.16079390048980713, "learning_rate": 4.877032812284511e-05, "loss": 0.5927, "num_input_tokens_seen": 4751248, "step": 7260 }, { "epoch": 3.8076519916142555, "grad_norm": 0.14591935276985168, "learning_rate": 4.876678369460205e-05, "loss": 0.537, "num_input_tokens_seen": 4754832, "step": 7265 }, { "epoch": 3.810272536687631, "grad_norm": 0.11574709415435791, "learning_rate": 4.8763234294618985e-05, "loss": 0.687, "num_input_tokens_seen": 4758224, "step": 7270 }, { "epoch": 3.8128930817610063, "grad_norm": 0.0931241512298584, "learning_rate": 4.875967992363841e-05, "loss": 0.5547, "num_input_tokens_seen": 4761040, "step": 7275 }, { "epoch": 3.8155136268343814, "grad_norm": 0.21460729837417603, "learning_rate": 4.875612058240386e-05, "loss": 0.5396, "num_input_tokens_seen": 4763728, "step": 7280 }, { "epoch": 3.818134171907757, "grad_norm": 0.18321985006332397, "learning_rate": 4.875255627165991e-05, "loss": 0.4326, "num_input_tokens_seen": 4766864, "step": 7285 }, { "epoch": 3.8207547169811322, "grad_norm": 0.3439599275588989, "learning_rate": 4.874898699215217e-05, "loss": 0.6233, "num_input_tokens_seen": 4769872, "step": 7290 }, { "epoch": 3.8233752620545074, "grad_norm": 0.13429929316043854, "learning_rate": 4.8745412744627296e-05, "loss": 0.5261, "num_input_tokens_seen": 4772816, "step": 7295 }, { "epoch": 3.8259958071278826, "grad_norm": 0.1509678214788437, "learning_rate": 4.874183352983297e-05, "loss": 0.578, "num_input_tokens_seen": 4776336, "step": 7300 }, { "epoch": 3.8286163522012577, "grad_norm": 0.12960785627365112, "learning_rate": 4.873824934851794e-05, "loss": 0.6522, "num_input_tokens_seen": 4780240, "step": 7305 }, { "epoch": 3.831236897274633, "grad_norm": 0.20225058495998383, "learning_rate": 4.873466020143196e-05, "loss": 0.7534, "num_input_tokens_seen": 4782544, "step": 7310 }, { "epoch": 3.8338574423480085, "grad_norm": 0.1348191648721695, "learning_rate": 4.873106608932585e-05, "loss": 0.5077, "num_input_tokens_seen": 4785104, "step": 7315 }, { "epoch": 3.8364779874213837, "grad_norm": 0.13636450469493866, "learning_rate": 4.872746701295145e-05, "loss": 0.467, "num_input_tokens_seen": 4788016, "step": 7320 }, { "epoch": 3.839098532494759, "grad_norm": 0.19259211421012878, "learning_rate": 4.872386297306165e-05, "loss": 0.5358, "num_input_tokens_seen": 4790640, "step": 7325 }, { "epoch": 3.841719077568134, "grad_norm": 0.21972183883190155, "learning_rate": 4.872025397041037e-05, "loss": 0.675, "num_input_tokens_seen": 4793232, "step": 7330 }, { "epoch": 3.8443396226415096, "grad_norm": 0.2335413247346878, "learning_rate": 4.871664000575258e-05, "loss": 0.5742, "num_input_tokens_seen": 4797136, "step": 7335 }, { "epoch": 3.846960167714885, "grad_norm": 0.22358478605747223, "learning_rate": 4.871302107984427e-05, "loss": 0.8159, "num_input_tokens_seen": 4800784, "step": 7340 }, { "epoch": 3.84958071278826, "grad_norm": 0.10092558711767197, "learning_rate": 4.8709397193442474e-05, "loss": 0.4287, "num_input_tokens_seen": 4804432, "step": 7345 }, { "epoch": 3.852201257861635, "grad_norm": 0.13189749419689178, "learning_rate": 4.870576834730529e-05, "loss": 0.59, "num_input_tokens_seen": 4808208, "step": 7350 }, { "epoch": 3.8548218029350103, "grad_norm": 0.21778327226638794, "learning_rate": 4.8702134542191816e-05, "loss": 0.6224, "num_input_tokens_seen": 4811248, "step": 7355 }, { "epoch": 3.8574423480083855, "grad_norm": 0.2465164065361023, "learning_rate": 4.86984957788622e-05, "loss": 0.5928, "num_input_tokens_seen": 4814352, "step": 7360 }, { "epoch": 3.860062893081761, "grad_norm": 0.1283256560564041, "learning_rate": 4.869485205807763e-05, "loss": 0.527, "num_input_tokens_seen": 4819152, "step": 7365 }, { "epoch": 3.8626834381551363, "grad_norm": 0.20652063190937042, "learning_rate": 4.869120338060034e-05, "loss": 0.4697, "num_input_tokens_seen": 4822224, "step": 7370 }, { "epoch": 3.8653039832285114, "grad_norm": 0.1605382263660431, "learning_rate": 4.8687549747193585e-05, "loss": 0.564, "num_input_tokens_seen": 4825648, "step": 7375 }, { "epoch": 3.867924528301887, "grad_norm": 0.1081993356347084, "learning_rate": 4.868389115862166e-05, "loss": 0.3966, "num_input_tokens_seen": 4828816, "step": 7380 }, { "epoch": 3.870545073375262, "grad_norm": 0.2240595668554306, "learning_rate": 4.868022761564991e-05, "loss": 0.5028, "num_input_tokens_seen": 4831408, "step": 7385 }, { "epoch": 3.8731656184486374, "grad_norm": 0.08755694329738617, "learning_rate": 4.8676559119044696e-05, "loss": 0.4679, "num_input_tokens_seen": 4834864, "step": 7390 }, { "epoch": 3.8757861635220126, "grad_norm": 0.15246446430683136, "learning_rate": 4.867288566957344e-05, "loss": 0.5327, "num_input_tokens_seen": 4837552, "step": 7395 }, { "epoch": 3.8784067085953877, "grad_norm": 0.22047170996665955, "learning_rate": 4.866920726800457e-05, "loss": 0.5226, "num_input_tokens_seen": 4840336, "step": 7400 }, { "epoch": 3.881027253668763, "grad_norm": 0.14238673448562622, "learning_rate": 4.8665523915107576e-05, "loss": 0.5666, "num_input_tokens_seen": 4843184, "step": 7405 }, { "epoch": 3.8836477987421385, "grad_norm": 0.13360993564128876, "learning_rate": 4.866183561165296e-05, "loss": 0.6088, "num_input_tokens_seen": 4846576, "step": 7410 }, { "epoch": 3.8862683438155137, "grad_norm": 0.16810192167758942, "learning_rate": 4.86581423584123e-05, "loss": 0.5068, "num_input_tokens_seen": 4849392, "step": 7415 }, { "epoch": 3.888888888888889, "grad_norm": 0.1973574012517929, "learning_rate": 4.8654444156158154e-05, "loss": 0.4003, "num_input_tokens_seen": 4852752, "step": 7420 }, { "epoch": 3.891509433962264, "grad_norm": 0.10583242028951645, "learning_rate": 4.865074100566415e-05, "loss": 0.4949, "num_input_tokens_seen": 4856880, "step": 7425 }, { "epoch": 3.8941299790356396, "grad_norm": 0.16714659333229065, "learning_rate": 4.864703290770495e-05, "loss": 0.5861, "num_input_tokens_seen": 4859984, "step": 7430 }, { "epoch": 3.896750524109015, "grad_norm": 0.17890898883342743, "learning_rate": 4.864331986305626e-05, "loss": 0.4422, "num_input_tokens_seen": 4862640, "step": 7435 }, { "epoch": 3.89937106918239, "grad_norm": 0.41902729868888855, "learning_rate": 4.8639601872494775e-05, "loss": 0.5935, "num_input_tokens_seen": 4865168, "step": 7440 }, { "epoch": 3.901991614255765, "grad_norm": 0.1570514738559723, "learning_rate": 4.8635878936798287e-05, "loss": 0.5805, "num_input_tokens_seen": 4869520, "step": 7445 }, { "epoch": 3.9046121593291403, "grad_norm": 0.12997007369995117, "learning_rate": 4.863215105674557e-05, "loss": 0.5095, "num_input_tokens_seen": 4872944, "step": 7450 }, { "epoch": 3.9072327044025155, "grad_norm": 0.16980616748332977, "learning_rate": 4.8628418233116457e-05, "loss": 0.4719, "num_input_tokens_seen": 4876624, "step": 7455 }, { "epoch": 3.909853249475891, "grad_norm": 0.19586284458637238, "learning_rate": 4.8624680466691824e-05, "loss": 0.5434, "num_input_tokens_seen": 4879152, "step": 7460 }, { "epoch": 3.9124737945492662, "grad_norm": 0.2193787395954132, "learning_rate": 4.862093775825356e-05, "loss": 0.445, "num_input_tokens_seen": 4883760, "step": 7465 }, { "epoch": 3.9150943396226414, "grad_norm": 0.1682325005531311, "learning_rate": 4.86171901085846e-05, "loss": 0.507, "num_input_tokens_seen": 4886640, "step": 7470 }, { "epoch": 3.917714884696017, "grad_norm": 0.21924655139446259, "learning_rate": 4.861343751846891e-05, "loss": 0.4898, "num_input_tokens_seen": 4890896, "step": 7475 }, { "epoch": 3.920335429769392, "grad_norm": 0.2445659339427948, "learning_rate": 4.860967998869148e-05, "loss": 0.5918, "num_input_tokens_seen": 4893744, "step": 7480 }, { "epoch": 3.9229559748427674, "grad_norm": 0.17669913172721863, "learning_rate": 4.860591752003836e-05, "loss": 0.7207, "num_input_tokens_seen": 4896240, "step": 7485 }, { "epoch": 3.9255765199161425, "grad_norm": 0.22758688032627106, "learning_rate": 4.8602150113296596e-05, "loss": 0.5796, "num_input_tokens_seen": 4899856, "step": 7490 }, { "epoch": 3.9281970649895177, "grad_norm": 0.13190172612667084, "learning_rate": 4.8598377769254305e-05, "loss": 0.485, "num_input_tokens_seen": 4903312, "step": 7495 }, { "epoch": 3.930817610062893, "grad_norm": 0.20298121869564056, "learning_rate": 4.85946004887006e-05, "loss": 0.5021, "num_input_tokens_seen": 4905744, "step": 7500 }, { "epoch": 3.9334381551362685, "grad_norm": 0.24637871980667114, "learning_rate": 4.859081827242566e-05, "loss": 0.5832, "num_input_tokens_seen": 4908368, "step": 7505 }, { "epoch": 3.9360587002096437, "grad_norm": 0.18384523689746857, "learning_rate": 4.858703112122068e-05, "loss": 0.4263, "num_input_tokens_seen": 4912112, "step": 7510 }, { "epoch": 3.938679245283019, "grad_norm": 0.1366911679506302, "learning_rate": 4.8583239035877876e-05, "loss": 0.5163, "num_input_tokens_seen": 4915216, "step": 7515 }, { "epoch": 3.941299790356394, "grad_norm": 0.12645173072814941, "learning_rate": 4.8579442017190526e-05, "loss": 0.4983, "num_input_tokens_seen": 4918320, "step": 7520 }, { "epoch": 3.9439203354297696, "grad_norm": 0.19554533064365387, "learning_rate": 4.8575640065952905e-05, "loss": 0.4485, "num_input_tokens_seen": 4922384, "step": 7525 }, { "epoch": 3.9465408805031448, "grad_norm": 0.16662931442260742, "learning_rate": 4.857183318296035e-05, "loss": 0.5099, "num_input_tokens_seen": 4925904, "step": 7530 }, { "epoch": 3.94916142557652, "grad_norm": 0.32427749037742615, "learning_rate": 4.856802136900922e-05, "loss": 0.494, "num_input_tokens_seen": 4928112, "step": 7535 }, { "epoch": 3.951781970649895, "grad_norm": 0.23178265988826752, "learning_rate": 4.856420462489689e-05, "loss": 0.5705, "num_input_tokens_seen": 4931536, "step": 7540 }, { "epoch": 3.9544025157232703, "grad_norm": 0.10432711243629456, "learning_rate": 4.856038295142179e-05, "loss": 0.4345, "num_input_tokens_seen": 4935184, "step": 7545 }, { "epoch": 3.9570230607966455, "grad_norm": 0.16416871547698975, "learning_rate": 4.855655634938337e-05, "loss": 0.4042, "num_input_tokens_seen": 4938704, "step": 7550 }, { "epoch": 3.959643605870021, "grad_norm": 0.18754276633262634, "learning_rate": 4.8552724819582106e-05, "loss": 0.5306, "num_input_tokens_seen": 4941360, "step": 7555 }, { "epoch": 3.9622641509433962, "grad_norm": 0.15844283998012543, "learning_rate": 4.8548888362819515e-05, "loss": 0.5934, "num_input_tokens_seen": 4944656, "step": 7560 }, { "epoch": 3.9648846960167714, "grad_norm": 0.343977689743042, "learning_rate": 4.854504697989814e-05, "loss": 0.4293, "num_input_tokens_seen": 4947248, "step": 7565 }, { "epoch": 3.967505241090147, "grad_norm": 0.1386493444442749, "learning_rate": 4.8541200671621554e-05, "loss": 0.4944, "num_input_tokens_seen": 4950160, "step": 7570 }, { "epoch": 3.970125786163522, "grad_norm": 0.24422569572925568, "learning_rate": 4.8537349438794354e-05, "loss": 0.5913, "num_input_tokens_seen": 4953264, "step": 7575 }, { "epoch": 3.9727463312368974, "grad_norm": 0.19063551723957062, "learning_rate": 4.853349328222219e-05, "loss": 0.5451, "num_input_tokens_seen": 4955760, "step": 7580 }, { "epoch": 3.9753668763102725, "grad_norm": 0.2014787644147873, "learning_rate": 4.8529632202711706e-05, "loss": 0.5147, "num_input_tokens_seen": 4961616, "step": 7585 }, { "epoch": 3.9779874213836477, "grad_norm": 0.44874829053878784, "learning_rate": 4.852576620107061e-05, "loss": 0.3681, "num_input_tokens_seen": 4964240, "step": 7590 }, { "epoch": 3.980607966457023, "grad_norm": 0.4917041063308716, "learning_rate": 4.8521895278107615e-05, "loss": 0.4742, "num_input_tokens_seen": 4968016, "step": 7595 }, { "epoch": 3.9832285115303985, "grad_norm": 0.09454305469989777, "learning_rate": 4.8518019434632484e-05, "loss": 0.5204, "num_input_tokens_seen": 4971952, "step": 7600 }, { "epoch": 3.9858490566037736, "grad_norm": 0.20275399088859558, "learning_rate": 4.8514138671455995e-05, "loss": 0.5377, "num_input_tokens_seen": 4975984, "step": 7605 }, { "epoch": 3.988469601677149, "grad_norm": 0.18291284143924713, "learning_rate": 4.851025298938996e-05, "loss": 0.5693, "num_input_tokens_seen": 4978832, "step": 7610 }, { "epoch": 3.991090146750524, "grad_norm": 0.1762763112783432, "learning_rate": 4.850636238924722e-05, "loss": 0.4613, "num_input_tokens_seen": 4981808, "step": 7615 }, { "epoch": 3.9937106918238996, "grad_norm": 0.1565980315208435, "learning_rate": 4.850246687184163e-05, "loss": 0.6783, "num_input_tokens_seen": 4985072, "step": 7620 }, { "epoch": 3.9963312368972748, "grad_norm": 0.18792738020420074, "learning_rate": 4.849856643798812e-05, "loss": 0.4569, "num_input_tokens_seen": 4987920, "step": 7625 }, { "epoch": 3.99895178197065, "grad_norm": 0.11361822485923767, "learning_rate": 4.849466108850259e-05, "loss": 0.4925, "num_input_tokens_seen": 4990768, "step": 7630 }, { "epoch": 4.0, "eval_loss": 0.4908377230167389, "eval_runtime": 9.8803, "eval_samples_per_second": 85.827, "eval_steps_per_second": 21.457, "num_input_tokens_seen": 4991472, "step": 7632 }, { "epoch": 4.001572327044025, "grad_norm": 0.15878821909427643, "learning_rate": 4.8490750824202e-05, "loss": 0.6176, "num_input_tokens_seen": 4993136, "step": 7635 }, { "epoch": 4.0041928721174, "grad_norm": 0.22556297481060028, "learning_rate": 4.8486835645904324e-05, "loss": 0.4654, "num_input_tokens_seen": 4997680, "step": 7640 }, { "epoch": 4.006813417190775, "grad_norm": 0.14587482810020447, "learning_rate": 4.84829155544286e-05, "loss": 0.5015, "num_input_tokens_seen": 5000976, "step": 7645 }, { "epoch": 4.009433962264151, "grad_norm": 0.2598191499710083, "learning_rate": 4.847899055059484e-05, "loss": 0.4672, "num_input_tokens_seen": 5003664, "step": 7650 }, { "epoch": 4.012054507337526, "grad_norm": 0.1021958589553833, "learning_rate": 4.847506063522412e-05, "loss": 0.5287, "num_input_tokens_seen": 5006832, "step": 7655 }, { "epoch": 4.014675052410902, "grad_norm": 0.15084144473075867, "learning_rate": 4.847112580913854e-05, "loss": 0.4865, "num_input_tokens_seen": 5010512, "step": 7660 }, { "epoch": 4.017295597484277, "grad_norm": 0.1423945426940918, "learning_rate": 4.846718607316121e-05, "loss": 0.6099, "num_input_tokens_seen": 5014352, "step": 7665 }, { "epoch": 4.019916142557652, "grad_norm": 0.17493513226509094, "learning_rate": 4.846324142811628e-05, "loss": 0.4973, "num_input_tokens_seen": 5017072, "step": 7670 }, { "epoch": 4.022536687631027, "grad_norm": 0.13842937350273132, "learning_rate": 4.8459291874828924e-05, "loss": 0.5454, "num_input_tokens_seen": 5019824, "step": 7675 }, { "epoch": 4.0251572327044025, "grad_norm": 0.15597917139530182, "learning_rate": 4.845533741412534e-05, "loss": 0.5845, "num_input_tokens_seen": 5022800, "step": 7680 }, { "epoch": 4.027777777777778, "grad_norm": 0.16233429312705994, "learning_rate": 4.845137804683277e-05, "loss": 0.4646, "num_input_tokens_seen": 5025424, "step": 7685 }, { "epoch": 4.030398322851153, "grad_norm": 0.14904600381851196, "learning_rate": 4.844741377377946e-05, "loss": 0.5124, "num_input_tokens_seen": 5028144, "step": 7690 }, { "epoch": 4.033018867924528, "grad_norm": 0.26634347438812256, "learning_rate": 4.844344459579469e-05, "loss": 0.5165, "num_input_tokens_seen": 5031568, "step": 7695 }, { "epoch": 4.035639412997903, "grad_norm": 0.21002498269081116, "learning_rate": 4.8439470513708766e-05, "loss": 0.5565, "num_input_tokens_seen": 5034960, "step": 7700 }, { "epoch": 4.038259958071279, "grad_norm": 0.3004229664802551, "learning_rate": 4.8435491528353026e-05, "loss": 0.3928, "num_input_tokens_seen": 5037552, "step": 7705 }, { "epoch": 4.040880503144654, "grad_norm": 0.20615629851818085, "learning_rate": 4.843150764055982e-05, "loss": 0.5029, "num_input_tokens_seen": 5041136, "step": 7710 }, { "epoch": 4.04350104821803, "grad_norm": 0.17020943760871887, "learning_rate": 4.842751885116254e-05, "loss": 0.5588, "num_input_tokens_seen": 5044240, "step": 7715 }, { "epoch": 4.046121593291405, "grad_norm": 0.13329093158245087, "learning_rate": 4.842352516099559e-05, "loss": 0.6352, "num_input_tokens_seen": 5047248, "step": 7720 }, { "epoch": 4.04874213836478, "grad_norm": 0.13495826721191406, "learning_rate": 4.84195265708944e-05, "loss": 0.4588, "num_input_tokens_seen": 5050128, "step": 7725 }, { "epoch": 4.051362683438155, "grad_norm": 0.14840881526470184, "learning_rate": 4.841552308169545e-05, "loss": 0.5462, "num_input_tokens_seen": 5053744, "step": 7730 }, { "epoch": 4.05398322851153, "grad_norm": 0.14680668711662292, "learning_rate": 4.841151469423619e-05, "loss": 0.6065, "num_input_tokens_seen": 5056912, "step": 7735 }, { "epoch": 4.056603773584905, "grad_norm": 0.1556388884782791, "learning_rate": 4.840750140935516e-05, "loss": 0.5222, "num_input_tokens_seen": 5060880, "step": 7740 }, { "epoch": 4.059224318658281, "grad_norm": 0.17371873557567596, "learning_rate": 4.840348322789189e-05, "loss": 0.6691, "num_input_tokens_seen": 5063888, "step": 7745 }, { "epoch": 4.061844863731656, "grad_norm": 0.12317319214344025, "learning_rate": 4.839946015068692e-05, "loss": 0.5116, "num_input_tokens_seen": 5066960, "step": 7750 }, { "epoch": 4.064465408805032, "grad_norm": 0.458268940448761, "learning_rate": 4.839543217858184e-05, "loss": 0.5621, "num_input_tokens_seen": 5070352, "step": 7755 }, { "epoch": 4.067085953878407, "grad_norm": 0.16563990712165833, "learning_rate": 4.839139931241926e-05, "loss": 0.6604, "num_input_tokens_seen": 5073488, "step": 7760 }, { "epoch": 4.069706498951782, "grad_norm": 0.1594584733247757, "learning_rate": 4.838736155304281e-05, "loss": 0.4381, "num_input_tokens_seen": 5076720, "step": 7765 }, { "epoch": 4.072327044025157, "grad_norm": 0.1944187730550766, "learning_rate": 4.8383318901297135e-05, "loss": 0.5826, "num_input_tokens_seen": 5080080, "step": 7770 }, { "epoch": 4.0749475890985325, "grad_norm": 0.1978784203529358, "learning_rate": 4.837927135802792e-05, "loss": 0.5314, "num_input_tokens_seen": 5083056, "step": 7775 }, { "epoch": 4.077568134171908, "grad_norm": 0.1473170518875122, "learning_rate": 4.8375218924081856e-05, "loss": 0.5847, "num_input_tokens_seen": 5085904, "step": 7780 }, { "epoch": 4.080188679245283, "grad_norm": 0.1454402208328247, "learning_rate": 4.837116160030668e-05, "loss": 0.5555, "num_input_tokens_seen": 5088816, "step": 7785 }, { "epoch": 4.082809224318658, "grad_norm": 0.13890576362609863, "learning_rate": 4.8367099387551126e-05, "loss": 0.5781, "num_input_tokens_seen": 5092112, "step": 7790 }, { "epoch": 4.085429769392033, "grad_norm": 0.27515092492103577, "learning_rate": 4.8363032286664964e-05, "loss": 0.4408, "num_input_tokens_seen": 5095536, "step": 7795 }, { "epoch": 4.088050314465409, "grad_norm": 0.15890467166900635, "learning_rate": 4.835896029849899e-05, "loss": 0.6095, "num_input_tokens_seen": 5099216, "step": 7800 }, { "epoch": 4.090670859538784, "grad_norm": 0.10959023982286453, "learning_rate": 4.835488342390502e-05, "loss": 0.4351, "num_input_tokens_seen": 5102288, "step": 7805 }, { "epoch": 4.09329140461216, "grad_norm": 0.1854366958141327, "learning_rate": 4.835080166373587e-05, "loss": 0.5513, "num_input_tokens_seen": 5105232, "step": 7810 }, { "epoch": 4.095911949685535, "grad_norm": 0.14861205220222473, "learning_rate": 4.834671501884542e-05, "loss": 0.6057, "num_input_tokens_seen": 5109840, "step": 7815 }, { "epoch": 4.09853249475891, "grad_norm": 0.3127140402793884, "learning_rate": 4.8342623490088543e-05, "loss": 0.5507, "num_input_tokens_seen": 5112240, "step": 7820 }, { "epoch": 4.101153039832285, "grad_norm": 0.09993305057287216, "learning_rate": 4.833852707832114e-05, "loss": 0.5181, "num_input_tokens_seen": 5115536, "step": 7825 }, { "epoch": 4.10377358490566, "grad_norm": 0.22768619656562805, "learning_rate": 4.833442578440013e-05, "loss": 0.5179, "num_input_tokens_seen": 5118160, "step": 7830 }, { "epoch": 4.106394129979035, "grad_norm": 0.25962650775909424, "learning_rate": 4.833031960918346e-05, "loss": 0.4706, "num_input_tokens_seen": 5120720, "step": 7835 }, { "epoch": 4.109014675052411, "grad_norm": 0.132898211479187, "learning_rate": 4.8326208553530095e-05, "loss": 0.5599, "num_input_tokens_seen": 5123632, "step": 7840 }, { "epoch": 4.111635220125786, "grad_norm": 0.15385600924491882, "learning_rate": 4.832209261830002e-05, "loss": 0.5485, "num_input_tokens_seen": 5126288, "step": 7845 }, { "epoch": 4.114255765199162, "grad_norm": 0.5516005158424377, "learning_rate": 4.831797180435423e-05, "loss": 0.8101, "num_input_tokens_seen": 5128752, "step": 7850 }, { "epoch": 4.116876310272537, "grad_norm": 0.17081983387470245, "learning_rate": 4.831384611255478e-05, "loss": 0.5936, "num_input_tokens_seen": 5132048, "step": 7855 }, { "epoch": 4.119496855345912, "grad_norm": 0.20109213888645172, "learning_rate": 4.8309715543764697e-05, "loss": 0.5069, "num_input_tokens_seen": 5135024, "step": 7860 }, { "epoch": 4.122117400419287, "grad_norm": 0.14650942385196686, "learning_rate": 4.8305580098848046e-05, "loss": 0.5074, "num_input_tokens_seen": 5138512, "step": 7865 }, { "epoch": 4.1247379454926625, "grad_norm": 0.19173626601696014, "learning_rate": 4.830143977866993e-05, "loss": 0.4335, "num_input_tokens_seen": 5141872, "step": 7870 }, { "epoch": 4.127358490566038, "grad_norm": 0.22699272632598877, "learning_rate": 4.829729458409645e-05, "loss": 0.5424, "num_input_tokens_seen": 5147664, "step": 7875 }, { "epoch": 4.129979035639413, "grad_norm": 0.3253679573535919, "learning_rate": 4.8293144515994725e-05, "loss": 0.4887, "num_input_tokens_seen": 5151696, "step": 7880 }, { "epoch": 4.132599580712788, "grad_norm": 0.14402587711811066, "learning_rate": 4.828898957523292e-05, "loss": 0.5163, "num_input_tokens_seen": 5154800, "step": 7885 }, { "epoch": 4.135220125786163, "grad_norm": 0.13866421580314636, "learning_rate": 4.8284829762680183e-05, "loss": 0.5785, "num_input_tokens_seen": 5157744, "step": 7890 }, { "epoch": 4.137840670859539, "grad_norm": 0.14359353482723236, "learning_rate": 4.828066507920671e-05, "loss": 0.4172, "num_input_tokens_seen": 5160944, "step": 7895 }, { "epoch": 4.140461215932914, "grad_norm": 0.15606079995632172, "learning_rate": 4.82764955256837e-05, "loss": 0.4486, "num_input_tokens_seen": 5163920, "step": 7900 }, { "epoch": 4.1430817610062896, "grad_norm": 0.26186153292655945, "learning_rate": 4.827232110298338e-05, "loss": 0.6342, "num_input_tokens_seen": 5167760, "step": 7905 }, { "epoch": 4.145702306079665, "grad_norm": 0.23258405923843384, "learning_rate": 4.8268141811978995e-05, "loss": 0.5596, "num_input_tokens_seen": 5170800, "step": 7910 }, { "epoch": 4.14832285115304, "grad_norm": 0.14586281776428223, "learning_rate": 4.826395765354479e-05, "loss": 0.4619, "num_input_tokens_seen": 5173488, "step": 7915 }, { "epoch": 4.150943396226415, "grad_norm": 0.17272591590881348, "learning_rate": 4.825976862855606e-05, "loss": 0.5807, "num_input_tokens_seen": 5176272, "step": 7920 }, { "epoch": 4.15356394129979, "grad_norm": 0.19079826772212982, "learning_rate": 4.8255574737889094e-05, "loss": 0.4706, "num_input_tokens_seen": 5178896, "step": 7925 }, { "epoch": 4.156184486373165, "grad_norm": 0.19401545822620392, "learning_rate": 4.82513759824212e-05, "loss": 0.494, "num_input_tokens_seen": 5182192, "step": 7930 }, { "epoch": 4.158805031446541, "grad_norm": 0.1703919768333435, "learning_rate": 4.824717236303072e-05, "loss": 0.8132, "num_input_tokens_seen": 5185232, "step": 7935 }, { "epoch": 4.161425576519916, "grad_norm": 0.12876929342746735, "learning_rate": 4.8242963880597005e-05, "loss": 0.688, "num_input_tokens_seen": 5188176, "step": 7940 }, { "epoch": 4.164046121593292, "grad_norm": 0.1525936722755432, "learning_rate": 4.823875053600041e-05, "loss": 0.6514, "num_input_tokens_seen": 5190768, "step": 7945 }, { "epoch": 4.166666666666667, "grad_norm": 0.26591724157333374, "learning_rate": 4.823453233012233e-05, "loss": 0.4773, "num_input_tokens_seen": 5193584, "step": 7950 }, { "epoch": 4.169287211740042, "grad_norm": 0.2196301519870758, "learning_rate": 4.8230309263845155e-05, "loss": 0.6289, "num_input_tokens_seen": 5196592, "step": 7955 }, { "epoch": 4.171907756813417, "grad_norm": 0.09658001363277435, "learning_rate": 4.822608133805231e-05, "loss": 0.5384, "num_input_tokens_seen": 5199024, "step": 7960 }, { "epoch": 4.1745283018867925, "grad_norm": 0.1250506341457367, "learning_rate": 4.8221848553628226e-05, "loss": 0.4627, "num_input_tokens_seen": 5202288, "step": 7965 }, { "epoch": 4.177148846960168, "grad_norm": 0.1346990019083023, "learning_rate": 4.8217610911458346e-05, "loss": 0.6336, "num_input_tokens_seen": 5205296, "step": 7970 }, { "epoch": 4.179769392033543, "grad_norm": 0.1217217892408371, "learning_rate": 4.821336841242915e-05, "loss": 0.4716, "num_input_tokens_seen": 5208912, "step": 7975 }, { "epoch": 4.182389937106918, "grad_norm": 0.27707692980766296, "learning_rate": 4.820912105742811e-05, "loss": 0.5284, "num_input_tokens_seen": 5212048, "step": 7980 }, { "epoch": 4.185010482180293, "grad_norm": 0.17288652062416077, "learning_rate": 4.8204868847343736e-05, "loss": 0.4973, "num_input_tokens_seen": 5214800, "step": 7985 }, { "epoch": 4.187631027253669, "grad_norm": 0.23903529345989227, "learning_rate": 4.820061178306553e-05, "loss": 0.3636, "num_input_tokens_seen": 5217744, "step": 7990 }, { "epoch": 4.190251572327044, "grad_norm": 0.2409410923719406, "learning_rate": 4.819634986548403e-05, "loss": 0.666, "num_input_tokens_seen": 5220784, "step": 7995 }, { "epoch": 4.1928721174004195, "grad_norm": 0.18600499629974365, "learning_rate": 4.819208309549077e-05, "loss": 0.4732, "num_input_tokens_seen": 5223504, "step": 8000 }, { "epoch": 4.195492662473795, "grad_norm": 0.18439403176307678, "learning_rate": 4.8187811473978316e-05, "loss": 0.4842, "num_input_tokens_seen": 5226672, "step": 8005 }, { "epoch": 4.19811320754717, "grad_norm": 0.20469433069229126, "learning_rate": 4.818353500184025e-05, "loss": 0.3918, "num_input_tokens_seen": 5229808, "step": 8010 }, { "epoch": 4.200733752620545, "grad_norm": 0.28837230801582336, "learning_rate": 4.817925367997115e-05, "loss": 0.5078, "num_input_tokens_seen": 5233328, "step": 8015 }, { "epoch": 4.20335429769392, "grad_norm": 0.10371947288513184, "learning_rate": 4.8174967509266615e-05, "loss": 0.4445, "num_input_tokens_seen": 5238512, "step": 8020 }, { "epoch": 4.205974842767295, "grad_norm": 0.18361540138721466, "learning_rate": 4.817067649062328e-05, "loss": 0.5106, "num_input_tokens_seen": 5242256, "step": 8025 }, { "epoch": 4.2085953878406706, "grad_norm": 0.31290242075920105, "learning_rate": 4.816638062493876e-05, "loss": 0.4577, "num_input_tokens_seen": 5245040, "step": 8030 }, { "epoch": 4.211215932914046, "grad_norm": 0.16128574311733246, "learning_rate": 4.8162079913111716e-05, "loss": 0.4937, "num_input_tokens_seen": 5248112, "step": 8035 }, { "epoch": 4.213836477987422, "grad_norm": 0.15192972123622894, "learning_rate": 4.81577743560418e-05, "loss": 0.5902, "num_input_tokens_seen": 5251792, "step": 8040 }, { "epoch": 4.216457023060797, "grad_norm": 0.20443060994148254, "learning_rate": 4.8153463954629684e-05, "loss": 0.6401, "num_input_tokens_seen": 5255120, "step": 8045 }, { "epoch": 4.219077568134172, "grad_norm": 0.16960462927818298, "learning_rate": 4.814914870977706e-05, "loss": 0.6723, "num_input_tokens_seen": 5259248, "step": 8050 }, { "epoch": 4.221698113207547, "grad_norm": 0.1665789783000946, "learning_rate": 4.8144828622386627e-05, "loss": 0.5097, "num_input_tokens_seen": 5262832, "step": 8055 }, { "epoch": 4.2243186582809225, "grad_norm": 0.2331022471189499, "learning_rate": 4.8140503693362095e-05, "loss": 0.6002, "num_input_tokens_seen": 5266064, "step": 8060 }, { "epoch": 4.226939203354298, "grad_norm": 0.1612444967031479, "learning_rate": 4.813617392360818e-05, "loss": 0.5559, "num_input_tokens_seen": 5269008, "step": 8065 }, { "epoch": 4.229559748427673, "grad_norm": 0.185243159532547, "learning_rate": 4.813183931403065e-05, "loss": 0.5864, "num_input_tokens_seen": 5272176, "step": 8070 }, { "epoch": 4.232180293501048, "grad_norm": 0.1125757098197937, "learning_rate": 4.8127499865536226e-05, "loss": 0.5734, "num_input_tokens_seen": 5276016, "step": 8075 }, { "epoch": 4.234800838574423, "grad_norm": 0.14975985884666443, "learning_rate": 4.812315557903269e-05, "loss": 0.5699, "num_input_tokens_seen": 5279024, "step": 8080 }, { "epoch": 4.237421383647799, "grad_norm": 0.15765529870986938, "learning_rate": 4.8118806455428806e-05, "loss": 0.5491, "num_input_tokens_seen": 5283088, "step": 8085 }, { "epoch": 4.240041928721174, "grad_norm": 0.32928845286369324, "learning_rate": 4.811445249563437e-05, "loss": 0.732, "num_input_tokens_seen": 5285936, "step": 8090 }, { "epoch": 4.2426624737945495, "grad_norm": 0.1698331981897354, "learning_rate": 4.811009370056017e-05, "loss": 0.4214, "num_input_tokens_seen": 5288912, "step": 8095 }, { "epoch": 4.245283018867925, "grad_norm": 0.24624907970428467, "learning_rate": 4.8105730071118016e-05, "loss": 0.6641, "num_input_tokens_seen": 5292912, "step": 8100 }, { "epoch": 4.2479035639413, "grad_norm": 0.14313150942325592, "learning_rate": 4.810136160822074e-05, "loss": 0.4528, "num_input_tokens_seen": 5296368, "step": 8105 }, { "epoch": 4.250524109014675, "grad_norm": 0.28506284952163696, "learning_rate": 4.8096988312782174e-05, "loss": 0.5853, "num_input_tokens_seen": 5299408, "step": 8110 }, { "epoch": 4.25314465408805, "grad_norm": 0.24160702526569366, "learning_rate": 4.809261018571715e-05, "loss": 0.475, "num_input_tokens_seen": 5302160, "step": 8115 }, { "epoch": 4.255765199161425, "grad_norm": 0.29476919770240784, "learning_rate": 4.808822722794153e-05, "loss": 0.5363, "num_input_tokens_seen": 5305296, "step": 8120 }, { "epoch": 4.2583857442348005, "grad_norm": 0.18649610877037048, "learning_rate": 4.8083839440372194e-05, "loss": 0.4646, "num_input_tokens_seen": 5307888, "step": 8125 }, { "epoch": 4.261006289308176, "grad_norm": 0.14749589562416077, "learning_rate": 4.8079446823926984e-05, "loss": 0.4578, "num_input_tokens_seen": 5311024, "step": 8130 }, { "epoch": 4.263626834381552, "grad_norm": 0.15114399790763855, "learning_rate": 4.8075049379524805e-05, "loss": 0.4931, "num_input_tokens_seen": 5314512, "step": 8135 }, { "epoch": 4.266247379454927, "grad_norm": 0.25857052206993103, "learning_rate": 4.807064710808555e-05, "loss": 0.5156, "num_input_tokens_seen": 5317744, "step": 8140 }, { "epoch": 4.268867924528302, "grad_norm": 0.2720852792263031, "learning_rate": 4.806624001053013e-05, "loss": 0.4636, "num_input_tokens_seen": 5322640, "step": 8145 }, { "epoch": 4.271488469601677, "grad_norm": 0.4169777035713196, "learning_rate": 4.8061828087780455e-05, "loss": 0.5743, "num_input_tokens_seen": 5326192, "step": 8150 }, { "epoch": 4.274109014675052, "grad_norm": 0.10089343786239624, "learning_rate": 4.805741134075944e-05, "loss": 0.5424, "num_input_tokens_seen": 5329104, "step": 8155 }, { "epoch": 4.276729559748428, "grad_norm": 0.13971886038780212, "learning_rate": 4.8052989770391026e-05, "loss": 0.4752, "num_input_tokens_seen": 5332368, "step": 8160 }, { "epoch": 4.279350104821803, "grad_norm": 0.6348995566368103, "learning_rate": 4.804856337760016e-05, "loss": 0.573, "num_input_tokens_seen": 5336240, "step": 8165 }, { "epoch": 4.281970649895178, "grad_norm": 0.09561752527952194, "learning_rate": 4.804413216331278e-05, "loss": 0.4825, "num_input_tokens_seen": 5341104, "step": 8170 }, { "epoch": 4.284591194968553, "grad_norm": 0.1348913162946701, "learning_rate": 4.8039696128455856e-05, "loss": 0.5536, "num_input_tokens_seen": 5343440, "step": 8175 }, { "epoch": 4.287211740041929, "grad_norm": 0.1118089109659195, "learning_rate": 4.803525527395736e-05, "loss": 0.4798, "num_input_tokens_seen": 5347280, "step": 8180 }, { "epoch": 4.289832285115304, "grad_norm": 0.17975324392318726, "learning_rate": 4.803080960074625e-05, "loss": 0.4893, "num_input_tokens_seen": 5354512, "step": 8185 }, { "epoch": 4.2924528301886795, "grad_norm": 0.12125525623559952, "learning_rate": 4.8026359109752525e-05, "loss": 0.515, "num_input_tokens_seen": 5357840, "step": 8190 }, { "epoch": 4.295073375262055, "grad_norm": 0.1586335301399231, "learning_rate": 4.8021903801907175e-05, "loss": 0.6104, "num_input_tokens_seen": 5362096, "step": 8195 }, { "epoch": 4.29769392033543, "grad_norm": 0.15079206228256226, "learning_rate": 4.801744367814219e-05, "loss": 0.5987, "num_input_tokens_seen": 5364784, "step": 8200 }, { "epoch": 4.300314465408805, "grad_norm": 0.21473419666290283, "learning_rate": 4.801297873939059e-05, "loss": 0.4559, "num_input_tokens_seen": 5367504, "step": 8205 }, { "epoch": 4.30293501048218, "grad_norm": 0.22658488154411316, "learning_rate": 4.8008508986586384e-05, "loss": 0.5859, "num_input_tokens_seen": 5370384, "step": 8210 }, { "epoch": 4.305555555555555, "grad_norm": 0.13325174152851105, "learning_rate": 4.800403442066459e-05, "loss": 0.5309, "num_input_tokens_seen": 5373264, "step": 8215 }, { "epoch": 4.3081761006289305, "grad_norm": 0.19321578741073608, "learning_rate": 4.799955504256125e-05, "loss": 0.4879, "num_input_tokens_seen": 5377040, "step": 8220 }, { "epoch": 4.310796645702306, "grad_norm": 0.32035520672798157, "learning_rate": 4.7995070853213376e-05, "loss": 0.5676, "num_input_tokens_seen": 5379888, "step": 8225 }, { "epoch": 4.313417190775682, "grad_norm": 0.34862571954727173, "learning_rate": 4.799058185355903e-05, "loss": 0.6276, "num_input_tokens_seen": 5382864, "step": 8230 }, { "epoch": 4.316037735849057, "grad_norm": 0.18387484550476074, "learning_rate": 4.7986088044537235e-05, "loss": 0.4868, "num_input_tokens_seen": 5386224, "step": 8235 }, { "epoch": 4.318658280922432, "grad_norm": 0.15370525419712067, "learning_rate": 4.798158942708807e-05, "loss": 0.5126, "num_input_tokens_seen": 5389104, "step": 8240 }, { "epoch": 4.321278825995807, "grad_norm": 0.16490939259529114, "learning_rate": 4.797708600215258e-05, "loss": 0.4492, "num_input_tokens_seen": 5391376, "step": 8245 }, { "epoch": 4.323899371069182, "grad_norm": 0.2232048213481903, "learning_rate": 4.7972577770672845e-05, "loss": 0.4219, "num_input_tokens_seen": 5394672, "step": 8250 }, { "epoch": 4.326519916142558, "grad_norm": 0.10944756120443344, "learning_rate": 4.796806473359192e-05, "loss": 0.5299, "num_input_tokens_seen": 5398096, "step": 8255 }, { "epoch": 4.329140461215933, "grad_norm": 0.25805437564849854, "learning_rate": 4.796354689185389e-05, "loss": 0.6388, "num_input_tokens_seen": 5400880, "step": 8260 }, { "epoch": 4.331761006289308, "grad_norm": 0.1735619157552719, "learning_rate": 4.795902424640383e-05, "loss": 0.6437, "num_input_tokens_seen": 5404336, "step": 8265 }, { "epoch": 4.334381551362683, "grad_norm": 0.21445070207118988, "learning_rate": 4.795449679818782e-05, "loss": 0.4485, "num_input_tokens_seen": 5408272, "step": 8270 }, { "epoch": 4.337002096436059, "grad_norm": 0.2250184416770935, "learning_rate": 4.794996454815297e-05, "loss": 0.5957, "num_input_tokens_seen": 5411824, "step": 8275 }, { "epoch": 4.339622641509434, "grad_norm": 0.40745624899864197, "learning_rate": 4.7945427497247355e-05, "loss": 0.4964, "num_input_tokens_seen": 5415344, "step": 8280 }, { "epoch": 4.3422431865828095, "grad_norm": 0.18648147583007812, "learning_rate": 4.7940885646420084e-05, "loss": 0.4815, "num_input_tokens_seen": 5418384, "step": 8285 }, { "epoch": 4.344863731656185, "grad_norm": 0.17554941773414612, "learning_rate": 4.7936338996621264e-05, "loss": 0.4886, "num_input_tokens_seen": 5422192, "step": 8290 }, { "epoch": 4.34748427672956, "grad_norm": 0.12489905208349228, "learning_rate": 4.7931787548801996e-05, "loss": 0.5542, "num_input_tokens_seen": 5426224, "step": 8295 }, { "epoch": 4.350104821802935, "grad_norm": 0.20721203088760376, "learning_rate": 4.7927231303914394e-05, "loss": 0.4834, "num_input_tokens_seen": 5428816, "step": 8300 }, { "epoch": 4.35272536687631, "grad_norm": 0.36742159724235535, "learning_rate": 4.7922670262911565e-05, "loss": 0.4959, "num_input_tokens_seen": 5433712, "step": 8305 }, { "epoch": 4.355345911949685, "grad_norm": 0.14740106463432312, "learning_rate": 4.791810442674764e-05, "loss": 0.6141, "num_input_tokens_seen": 5436080, "step": 8310 }, { "epoch": 4.3579664570230605, "grad_norm": 0.2353379726409912, "learning_rate": 4.791353379637773e-05, "loss": 0.5359, "num_input_tokens_seen": 5438864, "step": 8315 }, { "epoch": 4.360587002096436, "grad_norm": 0.18602685630321503, "learning_rate": 4.790895837275796e-05, "loss": 0.372, "num_input_tokens_seen": 5441488, "step": 8320 }, { "epoch": 4.363207547169811, "grad_norm": 0.12529075145721436, "learning_rate": 4.790437815684545e-05, "loss": 0.5227, "num_input_tokens_seen": 5444752, "step": 8325 }, { "epoch": 4.365828092243187, "grad_norm": 0.25596383213996887, "learning_rate": 4.789979314959835e-05, "loss": 0.5735, "num_input_tokens_seen": 5447600, "step": 8330 }, { "epoch": 4.368448637316562, "grad_norm": 0.24679426848888397, "learning_rate": 4.789520335197577e-05, "loss": 0.5133, "num_input_tokens_seen": 5450928, "step": 8335 }, { "epoch": 4.371069182389937, "grad_norm": 0.12791113555431366, "learning_rate": 4.789060876493785e-05, "loss": 0.5469, "num_input_tokens_seen": 5454192, "step": 8340 }, { "epoch": 4.373689727463312, "grad_norm": 0.13964205980300903, "learning_rate": 4.788600938944573e-05, "loss": 0.524, "num_input_tokens_seen": 5457200, "step": 8345 }, { "epoch": 4.376310272536688, "grad_norm": 0.14499019086360931, "learning_rate": 4.788140522646155e-05, "loss": 0.4904, "num_input_tokens_seen": 5460816, "step": 8350 }, { "epoch": 4.378930817610063, "grad_norm": 0.13311070203781128, "learning_rate": 4.787679627694843e-05, "loss": 0.5668, "num_input_tokens_seen": 5466256, "step": 8355 }, { "epoch": 4.381551362683438, "grad_norm": 0.18442972004413605, "learning_rate": 4.787218254187053e-05, "loss": 0.5744, "num_input_tokens_seen": 5469520, "step": 8360 }, { "epoch": 4.384171907756813, "grad_norm": 0.1829148381948471, "learning_rate": 4.7867564022192965e-05, "loss": 0.5264, "num_input_tokens_seen": 5473200, "step": 8365 }, { "epoch": 4.386792452830189, "grad_norm": 0.2564718723297119, "learning_rate": 4.786294071888191e-05, "loss": 0.4216, "num_input_tokens_seen": 5475888, "step": 8370 }, { "epoch": 4.389412997903564, "grad_norm": 0.18181300163269043, "learning_rate": 4.785831263290449e-05, "loss": 0.4325, "num_input_tokens_seen": 5478224, "step": 8375 }, { "epoch": 4.3920335429769395, "grad_norm": 0.12016533315181732, "learning_rate": 4.7853679765228845e-05, "loss": 0.5218, "num_input_tokens_seen": 5481264, "step": 8380 }, { "epoch": 4.394654088050315, "grad_norm": 0.10976345092058182, "learning_rate": 4.7849042116824116e-05, "loss": 0.5572, "num_input_tokens_seen": 5485040, "step": 8385 }, { "epoch": 4.39727463312369, "grad_norm": 0.253701776266098, "learning_rate": 4.7844399688660456e-05, "loss": 0.4875, "num_input_tokens_seen": 5488848, "step": 8390 }, { "epoch": 4.399895178197065, "grad_norm": 0.21017417311668396, "learning_rate": 4.783975248170901e-05, "loss": 0.5393, "num_input_tokens_seen": 5491504, "step": 8395 }, { "epoch": 4.40251572327044, "grad_norm": 0.17777849733829498, "learning_rate": 4.783510049694191e-05, "loss": 0.6791, "num_input_tokens_seen": 5494384, "step": 8400 }, { "epoch": 4.405136268343815, "grad_norm": 0.2837103605270386, "learning_rate": 4.78304437353323e-05, "loss": 0.7088, "num_input_tokens_seen": 5497712, "step": 8405 }, { "epoch": 4.4077568134171905, "grad_norm": 0.1905815750360489, "learning_rate": 4.782578219785433e-05, "loss": 0.4522, "num_input_tokens_seen": 5500464, "step": 8410 }, { "epoch": 4.410377358490566, "grad_norm": 0.14522621035575867, "learning_rate": 4.782111588548312e-05, "loss": 0.3889, "num_input_tokens_seen": 5503632, "step": 8415 }, { "epoch": 4.412997903563941, "grad_norm": 0.11701123416423798, "learning_rate": 4.781644479919484e-05, "loss": 0.547, "num_input_tokens_seen": 5506512, "step": 8420 }, { "epoch": 4.415618448637317, "grad_norm": 0.30342918634414673, "learning_rate": 4.7811768939966605e-05, "loss": 0.4031, "num_input_tokens_seen": 5509648, "step": 8425 }, { "epoch": 4.418238993710692, "grad_norm": 0.20972658693790436, "learning_rate": 4.780708830877656e-05, "loss": 0.5365, "num_input_tokens_seen": 5512144, "step": 8430 }, { "epoch": 4.420859538784067, "grad_norm": 0.17847931385040283, "learning_rate": 4.780240290660384e-05, "loss": 0.5612, "num_input_tokens_seen": 5515312, "step": 8435 }, { "epoch": 4.423480083857442, "grad_norm": 0.12098278850317001, "learning_rate": 4.7797712734428575e-05, "loss": 0.533, "num_input_tokens_seen": 5518000, "step": 8440 }, { "epoch": 4.426100628930818, "grad_norm": 0.1055857390165329, "learning_rate": 4.779301779323189e-05, "loss": 0.3889, "num_input_tokens_seen": 5521200, "step": 8445 }, { "epoch": 4.428721174004193, "grad_norm": 0.17864249646663666, "learning_rate": 4.778831808399593e-05, "loss": 0.6377, "num_input_tokens_seen": 5524336, "step": 8450 }, { "epoch": 4.431341719077568, "grad_norm": 0.2518327236175537, "learning_rate": 4.77836136077038e-05, "loss": 0.5636, "num_input_tokens_seen": 5527344, "step": 8455 }, { "epoch": 4.433962264150943, "grad_norm": 0.15578748285770416, "learning_rate": 4.777890436533963e-05, "loss": 0.3679, "num_input_tokens_seen": 5534256, "step": 8460 }, { "epoch": 4.436582809224318, "grad_norm": 0.11953968554735184, "learning_rate": 4.777419035788855e-05, "loss": 0.3932, "num_input_tokens_seen": 5537296, "step": 8465 }, { "epoch": 4.439203354297694, "grad_norm": 0.2232620120048523, "learning_rate": 4.776947158633667e-05, "loss": 0.6074, "num_input_tokens_seen": 5540656, "step": 8470 }, { "epoch": 4.4418238993710695, "grad_norm": 0.17401544749736786, "learning_rate": 4.7764748051671095e-05, "loss": 0.4586, "num_input_tokens_seen": 5543632, "step": 8475 }, { "epoch": 4.444444444444445, "grad_norm": 0.17711450159549713, "learning_rate": 4.7760019754879946e-05, "loss": 0.594, "num_input_tokens_seen": 5546544, "step": 8480 }, { "epoch": 4.44706498951782, "grad_norm": 0.2758691906929016, "learning_rate": 4.7755286696952324e-05, "loss": 0.5534, "num_input_tokens_seen": 5549168, "step": 8485 }, { "epoch": 4.449685534591195, "grad_norm": 0.09557351469993591, "learning_rate": 4.775054887887833e-05, "loss": 0.3805, "num_input_tokens_seen": 5553264, "step": 8490 }, { "epoch": 4.45230607966457, "grad_norm": 0.1266026645898819, "learning_rate": 4.774580630164906e-05, "loss": 0.5298, "num_input_tokens_seen": 5556144, "step": 8495 }, { "epoch": 4.454926624737945, "grad_norm": 0.20664772391319275, "learning_rate": 4.774105896625661e-05, "loss": 0.3842, "num_input_tokens_seen": 5558832, "step": 8500 }, { "epoch": 4.4575471698113205, "grad_norm": 0.13238638639450073, "learning_rate": 4.773630687369407e-05, "loss": 0.525, "num_input_tokens_seen": 5561968, "step": 8505 }, { "epoch": 4.460167714884696, "grad_norm": 0.1663808375597, "learning_rate": 4.773155002495552e-05, "loss": 0.4619, "num_input_tokens_seen": 5565456, "step": 8510 }, { "epoch": 4.462788259958071, "grad_norm": 0.16212333738803864, "learning_rate": 4.772678842103605e-05, "loss": 0.5504, "num_input_tokens_seen": 5568144, "step": 8515 }, { "epoch": 4.465408805031447, "grad_norm": 0.1284119337797165, "learning_rate": 4.772202206293172e-05, "loss": 0.5769, "num_input_tokens_seen": 5572080, "step": 8520 }, { "epoch": 4.468029350104822, "grad_norm": 0.17127011716365814, "learning_rate": 4.7717250951639586e-05, "loss": 0.4972, "num_input_tokens_seen": 5574992, "step": 8525 }, { "epoch": 4.470649895178197, "grad_norm": 0.2016131728887558, "learning_rate": 4.771247508815774e-05, "loss": 0.4196, "num_input_tokens_seen": 5578000, "step": 8530 }, { "epoch": 4.473270440251572, "grad_norm": 0.13106851279735565, "learning_rate": 4.770769447348522e-05, "loss": 0.4758, "num_input_tokens_seen": 5581200, "step": 8535 }, { "epoch": 4.475890985324948, "grad_norm": 0.2066706269979477, "learning_rate": 4.7702909108622074e-05, "loss": 0.4502, "num_input_tokens_seen": 5584336, "step": 8540 }, { "epoch": 4.478511530398323, "grad_norm": 0.1798993945121765, "learning_rate": 4.7698118994569355e-05, "loss": 0.5017, "num_input_tokens_seen": 5587440, "step": 8545 }, { "epoch": 4.481132075471698, "grad_norm": 0.3416654169559479, "learning_rate": 4.769332413232909e-05, "loss": 0.6981, "num_input_tokens_seen": 5590192, "step": 8550 }, { "epoch": 4.483752620545073, "grad_norm": 0.1284240484237671, "learning_rate": 4.768852452290432e-05, "loss": 0.4667, "num_input_tokens_seen": 5593712, "step": 8555 }, { "epoch": 4.486373165618448, "grad_norm": 0.16205941140651703, "learning_rate": 4.768372016729907e-05, "loss": 0.5636, "num_input_tokens_seen": 5596624, "step": 8560 }, { "epoch": 4.488993710691824, "grad_norm": 0.24822786450386047, "learning_rate": 4.767891106651834e-05, "loss": 0.495, "num_input_tokens_seen": 5599536, "step": 8565 }, { "epoch": 4.4916142557651995, "grad_norm": 0.11181289702653885, "learning_rate": 4.767409722156816e-05, "loss": 0.4904, "num_input_tokens_seen": 5603056, "step": 8570 }, { "epoch": 4.494234800838575, "grad_norm": 0.2968381941318512, "learning_rate": 4.7669278633455515e-05, "loss": 0.5409, "num_input_tokens_seen": 5605968, "step": 8575 }, { "epoch": 4.49685534591195, "grad_norm": 0.2509845197200775, "learning_rate": 4.7664455303188404e-05, "loss": 0.6121, "num_input_tokens_seen": 5608656, "step": 8580 }, { "epoch": 4.499475890985325, "grad_norm": 0.2507673501968384, "learning_rate": 4.7659627231775815e-05, "loss": 0.5487, "num_input_tokens_seen": 5612176, "step": 8585 }, { "epoch": 4.5020964360587, "grad_norm": 0.15434707701206207, "learning_rate": 4.765479442022773e-05, "loss": 0.426, "num_input_tokens_seen": 5615728, "step": 8590 }, { "epoch": 4.504716981132075, "grad_norm": 0.20543338358402252, "learning_rate": 4.76499568695551e-05, "loss": 0.5866, "num_input_tokens_seen": 5618448, "step": 8595 }, { "epoch": 4.5073375262054505, "grad_norm": 0.10491448640823364, "learning_rate": 4.764511458076991e-05, "loss": 0.4899, "num_input_tokens_seen": 5622256, "step": 8600 }, { "epoch": 4.509958071278826, "grad_norm": 0.1479307860136032, "learning_rate": 4.7640267554885085e-05, "loss": 0.5322, "num_input_tokens_seen": 5625264, "step": 8605 }, { "epoch": 4.512578616352201, "grad_norm": 0.17601263523101807, "learning_rate": 4.763541579291459e-05, "loss": 0.5597, "num_input_tokens_seen": 5627888, "step": 8610 }, { "epoch": 4.515199161425577, "grad_norm": 0.11380011588335037, "learning_rate": 4.763055929587336e-05, "loss": 0.4744, "num_input_tokens_seen": 5632208, "step": 8615 }, { "epoch": 4.517819706498952, "grad_norm": 0.12662474811077118, "learning_rate": 4.7625698064777294e-05, "loss": 0.4203, "num_input_tokens_seen": 5634384, "step": 8620 }, { "epoch": 4.520440251572327, "grad_norm": 0.168138325214386, "learning_rate": 4.762083210064332e-05, "loss": 0.5364, "num_input_tokens_seen": 5636944, "step": 8625 }, { "epoch": 4.523060796645702, "grad_norm": 0.10112447291612625, "learning_rate": 4.761596140448934e-05, "loss": 0.5805, "num_input_tokens_seen": 5640080, "step": 8630 }, { "epoch": 4.5256813417190775, "grad_norm": 0.17251616716384888, "learning_rate": 4.7611085977334256e-05, "loss": 0.5636, "num_input_tokens_seen": 5643408, "step": 8635 }, { "epoch": 4.528301886792453, "grad_norm": 0.3658086657524109, "learning_rate": 4.7606205820197946e-05, "loss": 0.5888, "num_input_tokens_seen": 5647216, "step": 8640 }, { "epoch": 4.530922431865828, "grad_norm": 0.2908863127231598, "learning_rate": 4.760132093410129e-05, "loss": 0.5212, "num_input_tokens_seen": 5651024, "step": 8645 }, { "epoch": 4.533542976939203, "grad_norm": 0.21741941571235657, "learning_rate": 4.759643132006614e-05, "loss": 0.4667, "num_input_tokens_seen": 5654032, "step": 8650 }, { "epoch": 4.536163522012579, "grad_norm": 0.17726731300354004, "learning_rate": 4.759153697911535e-05, "loss": 0.4225, "num_input_tokens_seen": 5656720, "step": 8655 }, { "epoch": 4.538784067085954, "grad_norm": 0.17564880847930908, "learning_rate": 4.758663791227276e-05, "loss": 0.5077, "num_input_tokens_seen": 5658992, "step": 8660 }, { "epoch": 4.5414046121593294, "grad_norm": 0.14647305011749268, "learning_rate": 4.7581734120563206e-05, "loss": 0.329, "num_input_tokens_seen": 5662448, "step": 8665 }, { "epoch": 4.544025157232705, "grad_norm": 0.16365182399749756, "learning_rate": 4.757682560501249e-05, "loss": 0.6602, "num_input_tokens_seen": 5665648, "step": 8670 }, { "epoch": 4.54664570230608, "grad_norm": 0.3085756301879883, "learning_rate": 4.757191236664744e-05, "loss": 0.5858, "num_input_tokens_seen": 5668976, "step": 8675 }, { "epoch": 4.549266247379455, "grad_norm": 0.10711446404457092, "learning_rate": 4.7566994406495825e-05, "loss": 0.3864, "num_input_tokens_seen": 5672272, "step": 8680 }, { "epoch": 4.55188679245283, "grad_norm": 0.20481131970882416, "learning_rate": 4.756207172558645e-05, "loss": 0.5304, "num_input_tokens_seen": 5675056, "step": 8685 }, { "epoch": 4.554507337526205, "grad_norm": 0.13378289341926575, "learning_rate": 4.7557144324949065e-05, "loss": 0.4682, "num_input_tokens_seen": 5677808, "step": 8690 }, { "epoch": 4.5571278825995805, "grad_norm": 0.14981448650360107, "learning_rate": 4.755221220561443e-05, "loss": 0.5036, "num_input_tokens_seen": 5681808, "step": 8695 }, { "epoch": 4.559748427672956, "grad_norm": 0.11927811056375504, "learning_rate": 4.7547275368614295e-05, "loss": 0.5943, "num_input_tokens_seen": 5684688, "step": 8700 }, { "epoch": 4.562368972746331, "grad_norm": 0.13957710564136505, "learning_rate": 4.754233381498139e-05, "loss": 0.4785, "num_input_tokens_seen": 5687792, "step": 8705 }, { "epoch": 4.564989517819707, "grad_norm": 0.2154967337846756, "learning_rate": 4.753738754574942e-05, "loss": 0.5959, "num_input_tokens_seen": 5690352, "step": 8710 }, { "epoch": 4.567610062893082, "grad_norm": 0.16437053680419922, "learning_rate": 4.75324365619531e-05, "loss": 0.6029, "num_input_tokens_seen": 5693232, "step": 8715 }, { "epoch": 4.570230607966457, "grad_norm": 0.19539186358451843, "learning_rate": 4.752748086462811e-05, "loss": 0.4506, "num_input_tokens_seen": 5695792, "step": 8720 }, { "epoch": 4.572851153039832, "grad_norm": 0.25052207708358765, "learning_rate": 4.752252045481113e-05, "loss": 0.4219, "num_input_tokens_seen": 5699664, "step": 8725 }, { "epoch": 4.5754716981132075, "grad_norm": 0.2199891209602356, "learning_rate": 4.7517555333539834e-05, "loss": 0.4223, "num_input_tokens_seen": 5702832, "step": 8730 }, { "epoch": 4.578092243186583, "grad_norm": 0.1430269181728363, "learning_rate": 4.751258550185284e-05, "loss": 0.448, "num_input_tokens_seen": 5705520, "step": 8735 }, { "epoch": 4.580712788259958, "grad_norm": 0.1257135570049286, "learning_rate": 4.7507610960789804e-05, "loss": 0.4987, "num_input_tokens_seen": 5709584, "step": 8740 }, { "epoch": 4.583333333333333, "grad_norm": 0.23275426030158997, "learning_rate": 4.7502631711391334e-05, "loss": 0.4648, "num_input_tokens_seen": 5713840, "step": 8745 }, { "epoch": 4.585953878406709, "grad_norm": 0.14841264486312866, "learning_rate": 4.749764775469904e-05, "loss": 0.6504, "num_input_tokens_seen": 5716016, "step": 8750 }, { "epoch": 4.588574423480084, "grad_norm": 0.13824674487113953, "learning_rate": 4.749265909175549e-05, "loss": 0.4557, "num_input_tokens_seen": 5718640, "step": 8755 }, { "epoch": 4.591194968553459, "grad_norm": 0.1506960242986679, "learning_rate": 4.7487665723604275e-05, "loss": 0.3981, "num_input_tokens_seen": 5722448, "step": 8760 }, { "epoch": 4.593815513626835, "grad_norm": 0.16912853717803955, "learning_rate": 4.7482667651289944e-05, "loss": 0.4671, "num_input_tokens_seen": 5725520, "step": 8765 }, { "epoch": 4.59643605870021, "grad_norm": 0.2081112116575241, "learning_rate": 4.747766487585803e-05, "loss": 0.5412, "num_input_tokens_seen": 5728144, "step": 8770 }, { "epoch": 4.599056603773585, "grad_norm": 0.17807720601558685, "learning_rate": 4.747265739835507e-05, "loss": 0.5996, "num_input_tokens_seen": 5731280, "step": 8775 }, { "epoch": 4.60167714884696, "grad_norm": 0.15503950417041779, "learning_rate": 4.746764521982856e-05, "loss": 0.6485, "num_input_tokens_seen": 5734096, "step": 8780 }, { "epoch": 4.604297693920335, "grad_norm": 0.1496698558330536, "learning_rate": 4.7462628341326995e-05, "loss": 0.4942, "num_input_tokens_seen": 5737360, "step": 8785 }, { "epoch": 4.6069182389937104, "grad_norm": 0.35189545154571533, "learning_rate": 4.7457606763899856e-05, "loss": 0.5089, "num_input_tokens_seen": 5740432, "step": 8790 }, { "epoch": 4.609538784067086, "grad_norm": 0.14790815114974976, "learning_rate": 4.745258048859758e-05, "loss": 0.4997, "num_input_tokens_seen": 5745552, "step": 8795 }, { "epoch": 4.612159329140461, "grad_norm": 0.2473706305027008, "learning_rate": 4.744754951647162e-05, "loss": 0.5122, "num_input_tokens_seen": 5749712, "step": 8800 }, { "epoch": 4.614779874213837, "grad_norm": 0.1379374861717224, "learning_rate": 4.74425138485744e-05, "loss": 0.4537, "num_input_tokens_seen": 5753648, "step": 8805 }, { "epoch": 4.617400419287212, "grad_norm": 0.10483356565237045, "learning_rate": 4.743747348595932e-05, "loss": 0.5938, "num_input_tokens_seen": 5757008, "step": 8810 }, { "epoch": 4.620020964360587, "grad_norm": 0.16254450380802155, "learning_rate": 4.743242842968077e-05, "loss": 0.4408, "num_input_tokens_seen": 5760816, "step": 8815 }, { "epoch": 4.622641509433962, "grad_norm": 0.13436710834503174, "learning_rate": 4.742737868079411e-05, "loss": 0.777, "num_input_tokens_seen": 5764080, "step": 8820 }, { "epoch": 4.6252620545073375, "grad_norm": 0.13573332130908966, "learning_rate": 4.7422324240355696e-05, "loss": 0.6307, "num_input_tokens_seen": 5767728, "step": 8825 }, { "epoch": 4.627882599580713, "grad_norm": 0.1600760519504547, "learning_rate": 4.741726510942286e-05, "loss": 0.5781, "num_input_tokens_seen": 5771568, "step": 8830 }, { "epoch": 4.630503144654088, "grad_norm": 0.17022773623466492, "learning_rate": 4.741220128905391e-05, "loss": 0.51, "num_input_tokens_seen": 5774576, "step": 8835 }, { "epoch": 4.633123689727463, "grad_norm": 0.20092076063156128, "learning_rate": 4.740713278030815e-05, "loss": 0.4131, "num_input_tokens_seen": 5778480, "step": 8840 }, { "epoch": 4.635744234800838, "grad_norm": 0.1397802084684372, "learning_rate": 4.740205958424584e-05, "loss": 0.4559, "num_input_tokens_seen": 5782160, "step": 8845 }, { "epoch": 4.638364779874214, "grad_norm": 0.30626580119132996, "learning_rate": 4.739698170192824e-05, "loss": 0.6309, "num_input_tokens_seen": 5784560, "step": 8850 }, { "epoch": 4.640985324947589, "grad_norm": 0.24539963901042938, "learning_rate": 4.7391899134417596e-05, "loss": 0.5448, "num_input_tokens_seen": 5787984, "step": 8855 }, { "epoch": 4.643605870020965, "grad_norm": 0.18217353522777557, "learning_rate": 4.73868118827771e-05, "loss": 0.6346, "num_input_tokens_seen": 5791056, "step": 8860 }, { "epoch": 4.64622641509434, "grad_norm": 0.15574966371059418, "learning_rate": 4.738171994807097e-05, "loss": 0.4129, "num_input_tokens_seen": 5793456, "step": 8865 }, { "epoch": 4.648846960167715, "grad_norm": 0.2884071171283722, "learning_rate": 4.7376623331364364e-05, "loss": 0.4609, "num_input_tokens_seen": 5797104, "step": 8870 }, { "epoch": 4.65146750524109, "grad_norm": 0.18289954960346222, "learning_rate": 4.737152203372345e-05, "loss": 0.5037, "num_input_tokens_seen": 5800048, "step": 8875 }, { "epoch": 4.654088050314465, "grad_norm": 0.2435384839773178, "learning_rate": 4.7366416056215345e-05, "loss": 0.5485, "num_input_tokens_seen": 5803600, "step": 8880 }, { "epoch": 4.65670859538784, "grad_norm": 0.19409117102622986, "learning_rate": 4.7361305399908185e-05, "loss": 0.7751, "num_input_tokens_seen": 5805776, "step": 8885 }, { "epoch": 4.659329140461216, "grad_norm": 0.2297433316707611, "learning_rate": 4.735619006587103e-05, "loss": 0.5018, "num_input_tokens_seen": 5808560, "step": 8890 }, { "epoch": 4.661949685534591, "grad_norm": 0.16509903967380524, "learning_rate": 4.735107005517398e-05, "loss": 0.4992, "num_input_tokens_seen": 5812048, "step": 8895 }, { "epoch": 4.664570230607967, "grad_norm": 0.14198417961597443, "learning_rate": 4.734594536888806e-05, "loss": 0.536, "num_input_tokens_seen": 5815856, "step": 8900 }, { "epoch": 4.667190775681342, "grad_norm": 0.1401142030954361, "learning_rate": 4.734081600808531e-05, "loss": 0.4188, "num_input_tokens_seen": 5820208, "step": 8905 }, { "epoch": 4.669811320754717, "grad_norm": 0.11767906695604324, "learning_rate": 4.733568197383872e-05, "loss": 0.5734, "num_input_tokens_seen": 5823888, "step": 8910 }, { "epoch": 4.672431865828092, "grad_norm": 0.1459094136953354, "learning_rate": 4.7330543267222294e-05, "loss": 0.3915, "num_input_tokens_seen": 5826832, "step": 8915 }, { "epoch": 4.6750524109014675, "grad_norm": 0.2761420011520386, "learning_rate": 4.732539988931097e-05, "loss": 0.4548, "num_input_tokens_seen": 5831344, "step": 8920 }, { "epoch": 4.677672955974843, "grad_norm": 0.24260462820529938, "learning_rate": 4.732025184118068e-05, "loss": 0.5494, "num_input_tokens_seen": 5835824, "step": 8925 }, { "epoch": 4.680293501048218, "grad_norm": 0.2430175542831421, "learning_rate": 4.731509912390837e-05, "loss": 0.6883, "num_input_tokens_seen": 5838192, "step": 8930 }, { "epoch": 4.682914046121593, "grad_norm": 0.18117107450962067, "learning_rate": 4.730994173857189e-05, "loss": 0.6362, "num_input_tokens_seen": 5841008, "step": 8935 }, { "epoch": 4.685534591194968, "grad_norm": 0.32926303148269653, "learning_rate": 4.7304779686250126e-05, "loss": 0.4928, "num_input_tokens_seen": 5844592, "step": 8940 }, { "epoch": 4.688155136268344, "grad_norm": 0.1796501874923706, "learning_rate": 4.729961296802292e-05, "loss": 0.4665, "num_input_tokens_seen": 5850256, "step": 8945 }, { "epoch": 4.690775681341719, "grad_norm": 0.2720605432987213, "learning_rate": 4.7294441584971095e-05, "loss": 0.4372, "num_input_tokens_seen": 5853232, "step": 8950 }, { "epoch": 4.693396226415095, "grad_norm": 0.13119274377822876, "learning_rate": 4.728926553817643e-05, "loss": 0.662, "num_input_tokens_seen": 5857584, "step": 8955 }, { "epoch": 4.69601677148847, "grad_norm": 0.2909463346004486, "learning_rate": 4.728408482872171e-05, "loss": 0.5466, "num_input_tokens_seen": 5860592, "step": 8960 }, { "epoch": 4.698637316561845, "grad_norm": 0.14501099288463593, "learning_rate": 4.727889945769067e-05, "loss": 0.6457, "num_input_tokens_seen": 5864208, "step": 8965 }, { "epoch": 4.70125786163522, "grad_norm": 0.27564990520477295, "learning_rate": 4.727370942616803e-05, "loss": 0.5072, "num_input_tokens_seen": 5867728, "step": 8970 }, { "epoch": 4.703878406708595, "grad_norm": 0.23916709423065186, "learning_rate": 4.72685147352395e-05, "loss": 0.6673, "num_input_tokens_seen": 5871472, "step": 8975 }, { "epoch": 4.70649895178197, "grad_norm": 0.1321743130683899, "learning_rate": 4.7263315385991735e-05, "loss": 0.3718, "num_input_tokens_seen": 5875152, "step": 8980 }, { "epoch": 4.709119496855346, "grad_norm": 0.15464574098587036, "learning_rate": 4.7258111379512384e-05, "loss": 0.5236, "num_input_tokens_seen": 5878192, "step": 8985 }, { "epoch": 4.711740041928721, "grad_norm": 0.11797003448009491, "learning_rate": 4.725290271689007e-05, "loss": 0.5474, "num_input_tokens_seen": 5881904, "step": 8990 }, { "epoch": 4.714360587002097, "grad_norm": 0.2504371702671051, "learning_rate": 4.724768939921438e-05, "loss": 0.543, "num_input_tokens_seen": 5884560, "step": 8995 }, { "epoch": 4.716981132075472, "grad_norm": 0.23778843879699707, "learning_rate": 4.7242471427575884e-05, "loss": 0.5194, "num_input_tokens_seen": 5887344, "step": 9000 }, { "epoch": 4.719601677148847, "grad_norm": 0.15934647619724274, "learning_rate": 4.723724880306612e-05, "loss": 0.5974, "num_input_tokens_seen": 5891088, "step": 9005 }, { "epoch": 4.722222222222222, "grad_norm": 0.1070263460278511, "learning_rate": 4.72320215267776e-05, "loss": 0.5255, "num_input_tokens_seen": 5894576, "step": 9010 }, { "epoch": 4.7248427672955975, "grad_norm": 0.20852166414260864, "learning_rate": 4.722678959980381e-05, "loss": 0.5443, "num_input_tokens_seen": 5898128, "step": 9015 }, { "epoch": 4.727463312368973, "grad_norm": 0.1820257008075714, "learning_rate": 4.7221553023239216e-05, "loss": 0.4617, "num_input_tokens_seen": 5901328, "step": 9020 }, { "epoch": 4.730083857442348, "grad_norm": 0.1267513632774353, "learning_rate": 4.721631179817925e-05, "loss": 0.3982, "num_input_tokens_seen": 5904656, "step": 9025 }, { "epoch": 4.732704402515723, "grad_norm": 0.13277624547481537, "learning_rate": 4.721106592572031e-05, "loss": 0.4855, "num_input_tokens_seen": 5907728, "step": 9030 }, { "epoch": 4.735324947589098, "grad_norm": 0.2087268829345703, "learning_rate": 4.7205815406959774e-05, "loss": 0.4607, "num_input_tokens_seen": 5910544, "step": 9035 }, { "epoch": 4.737945492662474, "grad_norm": 0.17674820125102997, "learning_rate": 4.7200560242996e-05, "loss": 0.5711, "num_input_tokens_seen": 5914096, "step": 9040 }, { "epoch": 4.740566037735849, "grad_norm": 0.1256900578737259, "learning_rate": 4.719530043492829e-05, "loss": 0.4279, "num_input_tokens_seen": 5916688, "step": 9045 }, { "epoch": 4.743186582809225, "grad_norm": 0.18522748351097107, "learning_rate": 4.7190035983856954e-05, "loss": 0.4898, "num_input_tokens_seen": 5919632, "step": 9050 }, { "epoch": 4.7458071278826, "grad_norm": 0.21085265278816223, "learning_rate": 4.718476689088325e-05, "loss": 0.5525, "num_input_tokens_seen": 5923280, "step": 9055 }, { "epoch": 4.748427672955975, "grad_norm": 0.16248635947704315, "learning_rate": 4.71794931571094e-05, "loss": 0.5441, "num_input_tokens_seen": 5926384, "step": 9060 }, { "epoch": 4.75104821802935, "grad_norm": 0.27476903796195984, "learning_rate": 4.717421478363862e-05, "loss": 0.5566, "num_input_tokens_seen": 5929168, "step": 9065 }, { "epoch": 4.753668763102725, "grad_norm": 0.4293088912963867, "learning_rate": 4.7168931771575085e-05, "loss": 0.5734, "num_input_tokens_seen": 5931632, "step": 9070 }, { "epoch": 4.7562893081761, "grad_norm": 0.2154930979013443, "learning_rate": 4.716364412202394e-05, "loss": 0.5478, "num_input_tokens_seen": 5934736, "step": 9075 }, { "epoch": 4.758909853249476, "grad_norm": 0.1679600477218628, "learning_rate": 4.7158351836091306e-05, "loss": 0.4448, "num_input_tokens_seen": 5937616, "step": 9080 }, { "epoch": 4.761530398322851, "grad_norm": 0.18701086938381195, "learning_rate": 4.7153054914884266e-05, "loss": 0.4949, "num_input_tokens_seen": 5940592, "step": 9085 }, { "epoch": 4.764150943396227, "grad_norm": 0.1614922434091568, "learning_rate": 4.714775335951086e-05, "loss": 0.4435, "num_input_tokens_seen": 5943568, "step": 9090 }, { "epoch": 4.766771488469602, "grad_norm": 0.1439923346042633, "learning_rate": 4.7142447171080137e-05, "loss": 0.4503, "num_input_tokens_seen": 5946960, "step": 9095 }, { "epoch": 4.769392033542977, "grad_norm": 0.17868570983409882, "learning_rate": 4.713713635070208e-05, "loss": 0.4603, "num_input_tokens_seen": 5950512, "step": 9100 }, { "epoch": 4.772012578616352, "grad_norm": 0.1982487291097641, "learning_rate": 4.713182089948765e-05, "loss": 0.582, "num_input_tokens_seen": 5954192, "step": 9105 }, { "epoch": 4.7746331236897275, "grad_norm": 0.16465625166893005, "learning_rate": 4.712650081854878e-05, "loss": 0.5786, "num_input_tokens_seen": 5957584, "step": 9110 }, { "epoch": 4.777253668763103, "grad_norm": 0.17883914709091187, "learning_rate": 4.712117610899837e-05, "loss": 0.5516, "num_input_tokens_seen": 5960432, "step": 9115 }, { "epoch": 4.779874213836478, "grad_norm": 0.15803292393684387, "learning_rate": 4.7115846771950286e-05, "loss": 0.4583, "num_input_tokens_seen": 5963984, "step": 9120 }, { "epoch": 4.782494758909853, "grad_norm": 0.1304355412721634, "learning_rate": 4.711051280851938e-05, "loss": 0.3786, "num_input_tokens_seen": 5967600, "step": 9125 }, { "epoch": 4.785115303983228, "grad_norm": 0.23233631253242493, "learning_rate": 4.7105174219821435e-05, "loss": 0.5916, "num_input_tokens_seen": 5970608, "step": 9130 }, { "epoch": 4.787735849056604, "grad_norm": 0.15291424095630646, "learning_rate": 4.709983100697323e-05, "loss": 0.4414, "num_input_tokens_seen": 5973840, "step": 9135 }, { "epoch": 4.790356394129979, "grad_norm": 0.22190088033676147, "learning_rate": 4.709448317109252e-05, "loss": 0.4585, "num_input_tokens_seen": 5977168, "step": 9140 }, { "epoch": 4.7929769392033545, "grad_norm": 0.20638687908649445, "learning_rate": 4.708913071329798e-05, "loss": 0.5107, "num_input_tokens_seen": 5980528, "step": 9145 }, { "epoch": 4.79559748427673, "grad_norm": 0.10107110440731049, "learning_rate": 4.708377363470932e-05, "loss": 0.6272, "num_input_tokens_seen": 5984368, "step": 9150 }, { "epoch": 4.798218029350105, "grad_norm": 0.16931375861167908, "learning_rate": 4.7078411936447144e-05, "loss": 0.4485, "num_input_tokens_seen": 5987568, "step": 9155 }, { "epoch": 4.80083857442348, "grad_norm": 0.22523802518844604, "learning_rate": 4.707304561963308e-05, "loss": 0.5299, "num_input_tokens_seen": 5990256, "step": 9160 }, { "epoch": 4.803459119496855, "grad_norm": 0.1692517250776291, "learning_rate": 4.70676746853897e-05, "loss": 0.579, "num_input_tokens_seen": 5993680, "step": 9165 }, { "epoch": 4.80607966457023, "grad_norm": 0.16865094006061554, "learning_rate": 4.706229913484054e-05, "loss": 0.5275, "num_input_tokens_seen": 6001392, "step": 9170 }, { "epoch": 4.808700209643606, "grad_norm": 0.22808411717414856, "learning_rate": 4.705691896911009e-05, "loss": 0.4951, "num_input_tokens_seen": 6004240, "step": 9175 }, { "epoch": 4.811320754716981, "grad_norm": 0.45261043310165405, "learning_rate": 4.705153418932383e-05, "loss": 0.5839, "num_input_tokens_seen": 6008304, "step": 9180 }, { "epoch": 4.813941299790356, "grad_norm": 0.24883124232292175, "learning_rate": 4.7046144796608206e-05, "loss": 0.5689, "num_input_tokens_seen": 6011504, "step": 9185 }, { "epoch": 4.816561844863732, "grad_norm": 0.18457211554050446, "learning_rate": 4.70407507920906e-05, "loss": 0.5083, "num_input_tokens_seen": 6014192, "step": 9190 }, { "epoch": 4.819182389937107, "grad_norm": 0.1768738329410553, "learning_rate": 4.7035352176899384e-05, "loss": 0.6265, "num_input_tokens_seen": 6016560, "step": 9195 }, { "epoch": 4.821802935010482, "grad_norm": 0.24951227009296417, "learning_rate": 4.7029948952163885e-05, "loss": 0.614, "num_input_tokens_seen": 6019120, "step": 9200 }, { "epoch": 4.8244234800838575, "grad_norm": 0.1990058571100235, "learning_rate": 4.7024541119014386e-05, "loss": 0.5214, "num_input_tokens_seen": 6022672, "step": 9205 }, { "epoch": 4.827044025157233, "grad_norm": 0.2451958805322647, "learning_rate": 4.701912867858217e-05, "loss": 0.3663, "num_input_tokens_seen": 6025520, "step": 9210 }, { "epoch": 4.829664570230608, "grad_norm": 0.21158400177955627, "learning_rate": 4.7013711631999434e-05, "loss": 0.4966, "num_input_tokens_seen": 6028208, "step": 9215 }, { "epoch": 4.832285115303983, "grad_norm": 0.21260811388492584, "learning_rate": 4.700828998039937e-05, "loss": 0.5107, "num_input_tokens_seen": 6031760, "step": 9220 }, { "epoch": 4.834905660377358, "grad_norm": 0.11618166416883469, "learning_rate": 4.700286372491612e-05, "loss": 0.5064, "num_input_tokens_seen": 6034128, "step": 9225 }, { "epoch": 4.837526205450734, "grad_norm": 0.19998283684253693, "learning_rate": 4.6997432866684816e-05, "loss": 0.5736, "num_input_tokens_seen": 6036976, "step": 9230 }, { "epoch": 4.840146750524109, "grad_norm": 0.14399781823158264, "learning_rate": 4.6991997406841504e-05, "loss": 0.6463, "num_input_tokens_seen": 6040272, "step": 9235 }, { "epoch": 4.8427672955974845, "grad_norm": 0.1480903923511505, "learning_rate": 4.698655734652324e-05, "loss": 0.5588, "num_input_tokens_seen": 6042768, "step": 9240 }, { "epoch": 4.84538784067086, "grad_norm": 0.2671760022640228, "learning_rate": 4.698111268686801e-05, "loss": 0.4107, "num_input_tokens_seen": 6046672, "step": 9245 }, { "epoch": 4.848008385744235, "grad_norm": 0.14680549502372742, "learning_rate": 4.6975663429014785e-05, "loss": 0.4487, "num_input_tokens_seen": 6049328, "step": 9250 }, { "epoch": 4.85062893081761, "grad_norm": 0.2799640893936157, "learning_rate": 4.697020957410347e-05, "loss": 0.4893, "num_input_tokens_seen": 6053168, "step": 9255 }, { "epoch": 4.853249475890985, "grad_norm": 0.14470618963241577, "learning_rate": 4.6964751123274974e-05, "loss": 0.5616, "num_input_tokens_seen": 6056560, "step": 9260 }, { "epoch": 4.85587002096436, "grad_norm": 0.24804317951202393, "learning_rate": 4.695928807767113e-05, "loss": 0.4239, "num_input_tokens_seen": 6059856, "step": 9265 }, { "epoch": 4.8584905660377355, "grad_norm": 0.1435040384531021, "learning_rate": 4.695382043843474e-05, "loss": 0.6215, "num_input_tokens_seen": 6063280, "step": 9270 }, { "epoch": 4.861111111111111, "grad_norm": 0.21815411746501923, "learning_rate": 4.694834820670958e-05, "loss": 0.5627, "num_input_tokens_seen": 6066256, "step": 9275 }, { "epoch": 4.863731656184486, "grad_norm": 0.3228498697280884, "learning_rate": 4.694287138364037e-05, "loss": 0.5475, "num_input_tokens_seen": 6069008, "step": 9280 }, { "epoch": 4.866352201257862, "grad_norm": 0.17624033987522125, "learning_rate": 4.6937389970372815e-05, "loss": 0.4926, "num_input_tokens_seen": 6072848, "step": 9285 }, { "epoch": 4.868972746331237, "grad_norm": 0.4041382968425751, "learning_rate": 4.693190396805354e-05, "loss": 0.5063, "num_input_tokens_seen": 6075536, "step": 9290 }, { "epoch": 4.871593291404612, "grad_norm": 0.12785518169403076, "learning_rate": 4.692641337783017e-05, "loss": 0.4741, "num_input_tokens_seen": 6079376, "step": 9295 }, { "epoch": 4.8742138364779874, "grad_norm": 0.18109245598316193, "learning_rate": 4.6920918200851274e-05, "loss": 0.402, "num_input_tokens_seen": 6083056, "step": 9300 }, { "epoch": 4.876834381551363, "grad_norm": 0.2368442565202713, "learning_rate": 4.691541843826638e-05, "loss": 0.53, "num_input_tokens_seen": 6085520, "step": 9305 }, { "epoch": 4.879454926624738, "grad_norm": 0.2664293646812439, "learning_rate": 4.690991409122597e-05, "loss": 0.5973, "num_input_tokens_seen": 6089040, "step": 9310 }, { "epoch": 4.882075471698113, "grad_norm": 0.1998872607946396, "learning_rate": 4.690440516088149e-05, "loss": 0.3917, "num_input_tokens_seen": 6092016, "step": 9315 }, { "epoch": 4.884696016771488, "grad_norm": 0.15335306525230408, "learning_rate": 4.689889164838536e-05, "loss": 0.5895, "num_input_tokens_seen": 6094608, "step": 9320 }, { "epoch": 4.887316561844864, "grad_norm": 0.17329491674900055, "learning_rate": 4.6893373554890925e-05, "loss": 0.3598, "num_input_tokens_seen": 6097520, "step": 9325 }, { "epoch": 4.889937106918239, "grad_norm": 0.22645528614521027, "learning_rate": 4.688785088155251e-05, "loss": 0.4119, "num_input_tokens_seen": 6100464, "step": 9330 }, { "epoch": 4.8925576519916145, "grad_norm": 0.13302022218704224, "learning_rate": 4.688232362952541e-05, "loss": 0.5648, "num_input_tokens_seen": 6104304, "step": 9335 }, { "epoch": 4.89517819706499, "grad_norm": 0.2066720873117447, "learning_rate": 4.687679179996585e-05, "loss": 0.4542, "num_input_tokens_seen": 6107856, "step": 9340 }, { "epoch": 4.897798742138365, "grad_norm": 0.2723369002342224, "learning_rate": 4.687125539403103e-05, "loss": 0.528, "num_input_tokens_seen": 6110832, "step": 9345 }, { "epoch": 4.90041928721174, "grad_norm": 0.22003987431526184, "learning_rate": 4.68657144128791e-05, "loss": 0.4331, "num_input_tokens_seen": 6114576, "step": 9350 }, { "epoch": 4.903039832285115, "grad_norm": 0.12941321730613708, "learning_rate": 4.6860168857669175e-05, "loss": 0.4477, "num_input_tokens_seen": 6118512, "step": 9355 }, { "epoch": 4.90566037735849, "grad_norm": 0.1652555912733078, "learning_rate": 4.6854618729561315e-05, "loss": 0.4551, "num_input_tokens_seen": 6121872, "step": 9360 }, { "epoch": 4.9082809224318655, "grad_norm": 0.14919321238994598, "learning_rate": 4.684906402971655e-05, "loss": 0.6181, "num_input_tokens_seen": 6125520, "step": 9365 }, { "epoch": 4.910901467505241, "grad_norm": 0.0998726412653923, "learning_rate": 4.684350475929685e-05, "loss": 0.4933, "num_input_tokens_seen": 6129936, "step": 9370 }, { "epoch": 4.913522012578616, "grad_norm": 0.22411106526851654, "learning_rate": 4.683794091946517e-05, "loss": 0.7359, "num_input_tokens_seen": 6132656, "step": 9375 }, { "epoch": 4.916142557651992, "grad_norm": 0.27685487270355225, "learning_rate": 4.683237251138538e-05, "loss": 0.4813, "num_input_tokens_seen": 6136528, "step": 9380 }, { "epoch": 4.918763102725367, "grad_norm": 0.14340268075466156, "learning_rate": 4.6826799536222346e-05, "loss": 0.447, "num_input_tokens_seen": 6139248, "step": 9385 }, { "epoch": 4.921383647798742, "grad_norm": 0.1611669510602951, "learning_rate": 4.6821221995141865e-05, "loss": 0.5106, "num_input_tokens_seen": 6142864, "step": 9390 }, { "epoch": 4.924004192872117, "grad_norm": 0.14887657761573792, "learning_rate": 4.681563988931068e-05, "loss": 0.7154, "num_input_tokens_seen": 6146128, "step": 9395 }, { "epoch": 4.926624737945493, "grad_norm": 0.19895455241203308, "learning_rate": 4.681005321989652e-05, "loss": 0.5687, "num_input_tokens_seen": 6148976, "step": 9400 }, { "epoch": 4.929245283018868, "grad_norm": 0.16089852154254913, "learning_rate": 4.6804461988068055e-05, "loss": 0.6775, "num_input_tokens_seen": 6151952, "step": 9405 }, { "epoch": 4.931865828092243, "grad_norm": 0.13928695023059845, "learning_rate": 4.6798866194994896e-05, "loss": 0.6372, "num_input_tokens_seen": 6155376, "step": 9410 }, { "epoch": 4.934486373165618, "grad_norm": 0.3727252781391144, "learning_rate": 4.6793265841847624e-05, "loss": 0.5638, "num_input_tokens_seen": 6157936, "step": 9415 }, { "epoch": 4.937106918238994, "grad_norm": 0.20987385511398315, "learning_rate": 4.6787660929797766e-05, "loss": 0.5082, "num_input_tokens_seen": 6163632, "step": 9420 }, { "epoch": 4.939727463312369, "grad_norm": 0.11305518448352814, "learning_rate": 4.678205146001781e-05, "loss": 0.4935, "num_input_tokens_seen": 6166800, "step": 9425 }, { "epoch": 4.9423480083857445, "grad_norm": 0.17359724640846252, "learning_rate": 4.67764374336812e-05, "loss": 0.518, "num_input_tokens_seen": 6170384, "step": 9430 }, { "epoch": 4.94496855345912, "grad_norm": 0.11237478256225586, "learning_rate": 4.6770818851962305e-05, "loss": 0.4393, "num_input_tokens_seen": 6173488, "step": 9435 }, { "epoch": 4.947589098532495, "grad_norm": 0.14282503724098206, "learning_rate": 4.676519571603648e-05, "loss": 0.4816, "num_input_tokens_seen": 6176560, "step": 9440 }, { "epoch": 4.95020964360587, "grad_norm": 0.09502990543842316, "learning_rate": 4.6759568027080034e-05, "loss": 0.5274, "num_input_tokens_seen": 6180496, "step": 9445 }, { "epoch": 4.952830188679245, "grad_norm": 0.24799016118049622, "learning_rate": 4.67539357862702e-05, "loss": 0.5594, "num_input_tokens_seen": 6183344, "step": 9450 }, { "epoch": 4.95545073375262, "grad_norm": 0.15387234091758728, "learning_rate": 4.6748298994785175e-05, "loss": 0.4639, "num_input_tokens_seen": 6186320, "step": 9455 }, { "epoch": 4.9580712788259955, "grad_norm": 0.16692492365837097, "learning_rate": 4.6742657653804116e-05, "loss": 0.4688, "num_input_tokens_seen": 6189904, "step": 9460 }, { "epoch": 4.960691823899371, "grad_norm": 0.1431969255208969, "learning_rate": 4.673701176450713e-05, "loss": 0.5915, "num_input_tokens_seen": 6194000, "step": 9465 }, { "epoch": 4.963312368972746, "grad_norm": 0.30738896131515503, "learning_rate": 4.6731361328075286e-05, "loss": 0.4781, "num_input_tokens_seen": 6197040, "step": 9470 }, { "epoch": 4.965932914046122, "grad_norm": 0.11317934095859528, "learning_rate": 4.672570634569056e-05, "loss": 0.4134, "num_input_tokens_seen": 6200944, "step": 9475 }, { "epoch": 4.968553459119497, "grad_norm": 0.2974034547805786, "learning_rate": 4.672004681853593e-05, "loss": 0.5173, "num_input_tokens_seen": 6203216, "step": 9480 }, { "epoch": 4.971174004192872, "grad_norm": 0.17745627462863922, "learning_rate": 4.6714382747795304e-05, "loss": 0.4392, "num_input_tokens_seen": 6206384, "step": 9485 }, { "epoch": 4.973794549266247, "grad_norm": 0.361564576625824, "learning_rate": 4.670871413465354e-05, "loss": 0.5477, "num_input_tokens_seen": 6209872, "step": 9490 }, { "epoch": 4.976415094339623, "grad_norm": 0.15212443470954895, "learning_rate": 4.670304098029645e-05, "loss": 0.5477, "num_input_tokens_seen": 6213328, "step": 9495 }, { "epoch": 4.979035639412998, "grad_norm": 0.17219682037830353, "learning_rate": 4.669736328591078e-05, "loss": 0.4855, "num_input_tokens_seen": 6216784, "step": 9500 }, { "epoch": 4.981656184486373, "grad_norm": 0.18073108792304993, "learning_rate": 4.669168105268424e-05, "loss": 0.5201, "num_input_tokens_seen": 6220016, "step": 9505 }, { "epoch": 4.984276729559748, "grad_norm": 0.1720736026763916, "learning_rate": 4.6685994281805514e-05, "loss": 0.4576, "num_input_tokens_seen": 6222672, "step": 9510 }, { "epoch": 4.986897274633124, "grad_norm": 0.16840416193008423, "learning_rate": 4.6680302974464175e-05, "loss": 0.5009, "num_input_tokens_seen": 6226224, "step": 9515 }, { "epoch": 4.989517819706499, "grad_norm": 0.26277557015419006, "learning_rate": 4.66746071318508e-05, "loss": 0.4537, "num_input_tokens_seen": 6228688, "step": 9520 }, { "epoch": 4.9921383647798745, "grad_norm": 0.14556457102298737, "learning_rate": 4.6668906755156896e-05, "loss": 0.4771, "num_input_tokens_seen": 6231440, "step": 9525 }, { "epoch": 4.99475890985325, "grad_norm": 0.2435116469860077, "learning_rate": 4.6663201845574906e-05, "loss": 0.6193, "num_input_tokens_seen": 6234160, "step": 9530 }, { "epoch": 4.997379454926625, "grad_norm": 0.17923542857170105, "learning_rate": 4.6657492404298234e-05, "loss": 0.3885, "num_input_tokens_seen": 6237232, "step": 9535 }, { "epoch": 5.0, "grad_norm": 6.79412317276001, "learning_rate": 4.665177843252124e-05, "loss": 1.4152, "num_input_tokens_seen": 6239608, "step": 9540 }, { "epoch": 5.0, "eval_loss": 0.48238223791122437, "eval_runtime": 10.0226, "eval_samples_per_second": 84.609, "eval_steps_per_second": 21.152, "num_input_tokens_seen": 6239608, "step": 9540 }, { "epoch": 5.002620545073375, "grad_norm": 0.1414533108472824, "learning_rate": 4.664605993143921e-05, "loss": 0.4897, "num_input_tokens_seen": 6242936, "step": 9545 }, { "epoch": 5.00524109014675, "grad_norm": 0.1792527586221695, "learning_rate": 4.6640336902248396e-05, "loss": 0.5858, "num_input_tokens_seen": 6246392, "step": 9550 }, { "epoch": 5.0078616352201255, "grad_norm": 0.1728523075580597, "learning_rate": 4.6634609346145984e-05, "loss": 0.6052, "num_input_tokens_seen": 6250008, "step": 9555 }, { "epoch": 5.010482180293501, "grad_norm": 0.1726694256067276, "learning_rate": 4.6628877264330116e-05, "loss": 0.4689, "num_input_tokens_seen": 6253336, "step": 9560 }, { "epoch": 5.013102725366877, "grad_norm": 0.13707417249679565, "learning_rate": 4.6623140657999886e-05, "loss": 0.4062, "num_input_tokens_seen": 6256888, "step": 9565 }, { "epoch": 5.015723270440252, "grad_norm": 0.22346174716949463, "learning_rate": 4.661739952835532e-05, "loss": 0.518, "num_input_tokens_seen": 6259768, "step": 9570 }, { "epoch": 5.018343815513627, "grad_norm": 0.6813058257102966, "learning_rate": 4.66116538765974e-05, "loss": 0.5345, "num_input_tokens_seen": 6262616, "step": 9575 }, { "epoch": 5.020964360587002, "grad_norm": 0.13666513562202454, "learning_rate": 4.6605903703928034e-05, "loss": 0.3851, "num_input_tokens_seen": 6265624, "step": 9580 }, { "epoch": 5.023584905660377, "grad_norm": 0.1287325769662857, "learning_rate": 4.6600149011550116e-05, "loss": 0.6543, "num_input_tokens_seen": 6269304, "step": 9585 }, { "epoch": 5.026205450733753, "grad_norm": 0.11767981946468353, "learning_rate": 4.659438980066745e-05, "loss": 0.5304, "num_input_tokens_seen": 6272312, "step": 9590 }, { "epoch": 5.028825995807128, "grad_norm": 0.22370851039886475, "learning_rate": 4.65886260724848e-05, "loss": 0.5392, "num_input_tokens_seen": 6274904, "step": 9595 }, { "epoch": 5.031446540880503, "grad_norm": 0.16702990233898163, "learning_rate": 4.658285782820787e-05, "loss": 0.4867, "num_input_tokens_seen": 6278104, "step": 9600 }, { "epoch": 5.034067085953878, "grad_norm": 0.14217080175876617, "learning_rate": 4.657708506904331e-05, "loss": 0.4448, "num_input_tokens_seen": 6281656, "step": 9605 }, { "epoch": 5.036687631027253, "grad_norm": 0.1500714123249054, "learning_rate": 4.6571307796198724e-05, "loss": 0.6137, "num_input_tokens_seen": 6285240, "step": 9610 }, { "epoch": 5.039308176100629, "grad_norm": 0.17939363420009613, "learning_rate": 4.656552601088265e-05, "loss": 0.5531, "num_input_tokens_seen": 6287768, "step": 9615 }, { "epoch": 5.0419287211740045, "grad_norm": 0.12126346677541733, "learning_rate": 4.655973971430456e-05, "loss": 0.4708, "num_input_tokens_seen": 6291128, "step": 9620 }, { "epoch": 5.04454926624738, "grad_norm": 0.1915358006954193, "learning_rate": 4.6553948907674895e-05, "loss": 0.3054, "num_input_tokens_seen": 6295416, "step": 9625 }, { "epoch": 5.047169811320755, "grad_norm": 0.14890128374099731, "learning_rate": 4.654815359220501e-05, "loss": 0.4894, "num_input_tokens_seen": 6299000, "step": 9630 }, { "epoch": 5.04979035639413, "grad_norm": 0.20116496086120605, "learning_rate": 4.6542353769107226e-05, "loss": 0.5252, "num_input_tokens_seen": 6302680, "step": 9635 }, { "epoch": 5.052410901467505, "grad_norm": 0.20006288588047028, "learning_rate": 4.653654943959481e-05, "loss": 0.5028, "num_input_tokens_seen": 6306744, "step": 9640 }, { "epoch": 5.05503144654088, "grad_norm": 0.20581437647342682, "learning_rate": 4.653074060488195e-05, "loss": 0.6636, "num_input_tokens_seen": 6309400, "step": 9645 }, { "epoch": 5.0576519916142555, "grad_norm": 0.16050978004932404, "learning_rate": 4.6524927266183796e-05, "loss": 0.4379, "num_input_tokens_seen": 6312728, "step": 9650 }, { "epoch": 5.060272536687631, "grad_norm": 0.15329931676387787, "learning_rate": 4.6519109424716426e-05, "loss": 0.6168, "num_input_tokens_seen": 6315960, "step": 9655 }, { "epoch": 5.062893081761007, "grad_norm": 0.08104902505874634, "learning_rate": 4.6513287081696855e-05, "loss": 0.436, "num_input_tokens_seen": 6319320, "step": 9660 }, { "epoch": 5.065513626834382, "grad_norm": 0.18715868890285492, "learning_rate": 4.6507460238343076e-05, "loss": 0.6108, "num_input_tokens_seen": 6321656, "step": 9665 }, { "epoch": 5.068134171907757, "grad_norm": 0.22844159603118896, "learning_rate": 4.6501628895873984e-05, "loss": 0.4498, "num_input_tokens_seen": 6324664, "step": 9670 }, { "epoch": 5.070754716981132, "grad_norm": 0.1486690789461136, "learning_rate": 4.649579305550942e-05, "loss": 0.5246, "num_input_tokens_seen": 6327192, "step": 9675 }, { "epoch": 5.073375262054507, "grad_norm": 0.22317923605442047, "learning_rate": 4.64899527184702e-05, "loss": 0.4672, "num_input_tokens_seen": 6330232, "step": 9680 }, { "epoch": 5.075995807127883, "grad_norm": 0.1668732613325119, "learning_rate": 4.6484107885978035e-05, "loss": 0.5559, "num_input_tokens_seen": 6332600, "step": 9685 }, { "epoch": 5.078616352201258, "grad_norm": 0.2155432254076004, "learning_rate": 4.64782585592556e-05, "loss": 0.5709, "num_input_tokens_seen": 6336600, "step": 9690 }, { "epoch": 5.081236897274633, "grad_norm": 0.19453033804893494, "learning_rate": 4.647240473952652e-05, "loss": 0.5166, "num_input_tokens_seen": 6339640, "step": 9695 }, { "epoch": 5.083857442348008, "grad_norm": 0.1711329072713852, "learning_rate": 4.6466546428015336e-05, "loss": 0.7143, "num_input_tokens_seen": 6342840, "step": 9700 }, { "epoch": 5.086477987421383, "grad_norm": 0.1702709197998047, "learning_rate": 4.646068362594753e-05, "loss": 0.5736, "num_input_tokens_seen": 6346008, "step": 9705 }, { "epoch": 5.089098532494759, "grad_norm": 0.15155255794525146, "learning_rate": 4.645481633454956e-05, "loss": 0.521, "num_input_tokens_seen": 6349016, "step": 9710 }, { "epoch": 5.0917190775681345, "grad_norm": 0.190457284450531, "learning_rate": 4.644894455504878e-05, "loss": 0.4615, "num_input_tokens_seen": 6352952, "step": 9715 }, { "epoch": 5.09433962264151, "grad_norm": 0.21649695932865143, "learning_rate": 4.64430682886735e-05, "loss": 0.4921, "num_input_tokens_seen": 6356504, "step": 9720 }, { "epoch": 5.096960167714885, "grad_norm": 0.1344349980354309, "learning_rate": 4.643718753665297e-05, "loss": 0.4644, "num_input_tokens_seen": 6359256, "step": 9725 }, { "epoch": 5.09958071278826, "grad_norm": 0.13223427534103394, "learning_rate": 4.643130230021737e-05, "loss": 0.5484, "num_input_tokens_seen": 6362520, "step": 9730 }, { "epoch": 5.102201257861635, "grad_norm": 0.11382966488599777, "learning_rate": 4.6425412580597836e-05, "loss": 0.5903, "num_input_tokens_seen": 6365368, "step": 9735 }, { "epoch": 5.10482180293501, "grad_norm": 0.21268533170223236, "learning_rate": 4.641951837902643e-05, "loss": 0.7818, "num_input_tokens_seen": 6367896, "step": 9740 }, { "epoch": 5.1074423480083855, "grad_norm": 0.30806106328964233, "learning_rate": 4.641361969673614e-05, "loss": 0.5132, "num_input_tokens_seen": 6371384, "step": 9745 }, { "epoch": 5.110062893081761, "grad_norm": 0.18379610776901245, "learning_rate": 4.640771653496091e-05, "loss": 0.4505, "num_input_tokens_seen": 6375128, "step": 9750 }, { "epoch": 5.112683438155136, "grad_norm": 0.12632650136947632, "learning_rate": 4.64018088949356e-05, "loss": 0.3975, "num_input_tokens_seen": 6377304, "step": 9755 }, { "epoch": 5.115303983228512, "grad_norm": 0.22453030943870544, "learning_rate": 4.6395896777896055e-05, "loss": 0.5258, "num_input_tokens_seen": 6379992, "step": 9760 }, { "epoch": 5.117924528301887, "grad_norm": 0.24078427255153656, "learning_rate": 4.638998018507899e-05, "loss": 0.5094, "num_input_tokens_seen": 6383000, "step": 9765 }, { "epoch": 5.120545073375262, "grad_norm": 0.16395017504692078, "learning_rate": 4.63840591177221e-05, "loss": 0.4744, "num_input_tokens_seen": 6386776, "step": 9770 }, { "epoch": 5.123165618448637, "grad_norm": 0.17077036201953888, "learning_rate": 4.637813357706401e-05, "loss": 0.4947, "num_input_tokens_seen": 6390424, "step": 9775 }, { "epoch": 5.1257861635220126, "grad_norm": 0.2844568192958832, "learning_rate": 4.6372203564344264e-05, "loss": 0.6833, "num_input_tokens_seen": 6392888, "step": 9780 }, { "epoch": 5.128406708595388, "grad_norm": 0.12131640315055847, "learning_rate": 4.6366269080803364e-05, "loss": 0.5037, "num_input_tokens_seen": 6396728, "step": 9785 }, { "epoch": 5.131027253668763, "grad_norm": 0.15692825615406036, "learning_rate": 4.636033012768273e-05, "loss": 0.5022, "num_input_tokens_seen": 6400088, "step": 9790 }, { "epoch": 5.133647798742138, "grad_norm": 0.13032971322536469, "learning_rate": 4.635438670622473e-05, "loss": 0.5143, "num_input_tokens_seen": 6403032, "step": 9795 }, { "epoch": 5.136268343815513, "grad_norm": 0.14246422052383423, "learning_rate": 4.6348438817672646e-05, "loss": 0.5063, "num_input_tokens_seen": 6405944, "step": 9800 }, { "epoch": 5.138888888888889, "grad_norm": 0.1818077564239502, "learning_rate": 4.634248646327072e-05, "loss": 0.4967, "num_input_tokens_seen": 6409048, "step": 9805 }, { "epoch": 5.1415094339622645, "grad_norm": 0.16944465041160583, "learning_rate": 4.6336529644264114e-05, "loss": 0.4283, "num_input_tokens_seen": 6412920, "step": 9810 }, { "epoch": 5.14412997903564, "grad_norm": 0.18349815905094147, "learning_rate": 4.633056836189893e-05, "loss": 0.5044, "num_input_tokens_seen": 6416120, "step": 9815 }, { "epoch": 5.146750524109015, "grad_norm": 0.14899174869060516, "learning_rate": 4.632460261742219e-05, "loss": 0.6511, "num_input_tokens_seen": 6419896, "step": 9820 }, { "epoch": 5.14937106918239, "grad_norm": 0.17638444900512695, "learning_rate": 4.631863241208187e-05, "loss": 0.4894, "num_input_tokens_seen": 6423096, "step": 9825 }, { "epoch": 5.151991614255765, "grad_norm": 0.23766474425792694, "learning_rate": 4.6312657747126865e-05, "loss": 0.5511, "num_input_tokens_seen": 6425592, "step": 9830 }, { "epoch": 5.15461215932914, "grad_norm": 0.2367236465215683, "learning_rate": 4.630667862380701e-05, "loss": 0.4782, "num_input_tokens_seen": 6429304, "step": 9835 }, { "epoch": 5.1572327044025155, "grad_norm": 0.3339790999889374, "learning_rate": 4.630069504337306e-05, "loss": 0.4526, "num_input_tokens_seen": 6432984, "step": 9840 }, { "epoch": 5.159853249475891, "grad_norm": 0.13635480403900146, "learning_rate": 4.629470700707672e-05, "loss": 0.4489, "num_input_tokens_seen": 6436376, "step": 9845 }, { "epoch": 5.162473794549266, "grad_norm": 0.25275370478630066, "learning_rate": 4.628871451617062e-05, "loss": 0.5536, "num_input_tokens_seen": 6439544, "step": 9850 }, { "epoch": 5.165094339622642, "grad_norm": 0.17315857112407684, "learning_rate": 4.628271757190832e-05, "loss": 0.5043, "num_input_tokens_seen": 6443128, "step": 9855 }, { "epoch": 5.167714884696017, "grad_norm": 0.15191875398159027, "learning_rate": 4.62767161755443e-05, "loss": 0.5849, "num_input_tokens_seen": 6446456, "step": 9860 }, { "epoch": 5.170335429769392, "grad_norm": 0.25175201892852783, "learning_rate": 4.6270710328334004e-05, "loss": 0.5774, "num_input_tokens_seen": 6449272, "step": 9865 }, { "epoch": 5.172955974842767, "grad_norm": 0.18411865830421448, "learning_rate": 4.626470003153377e-05, "loss": 0.6258, "num_input_tokens_seen": 6452632, "step": 9870 }, { "epoch": 5.1755765199161425, "grad_norm": 0.15364713966846466, "learning_rate": 4.625868528640089e-05, "loss": 0.4149, "num_input_tokens_seen": 6454936, "step": 9875 }, { "epoch": 5.178197064989518, "grad_norm": 0.16797658801078796, "learning_rate": 4.6252666094193586e-05, "loss": 0.4604, "num_input_tokens_seen": 6458168, "step": 9880 }, { "epoch": 5.180817610062893, "grad_norm": 0.18545696139335632, "learning_rate": 4.6246642456171e-05, "loss": 0.5829, "num_input_tokens_seen": 6462392, "step": 9885 }, { "epoch": 5.183438155136268, "grad_norm": 0.13667841255664825, "learning_rate": 4.6240614373593196e-05, "loss": 0.5699, "num_input_tokens_seen": 6470968, "step": 9890 }, { "epoch": 5.186058700209643, "grad_norm": 0.20201030373573303, "learning_rate": 4.623458184772119e-05, "loss": 0.5199, "num_input_tokens_seen": 6473880, "step": 9895 }, { "epoch": 5.188679245283019, "grad_norm": 0.1769799143075943, "learning_rate": 4.622854487981693e-05, "loss": 0.5151, "num_input_tokens_seen": 6476504, "step": 9900 }, { "epoch": 5.191299790356394, "grad_norm": 0.28989559412002563, "learning_rate": 4.622250347114326e-05, "loss": 0.4956, "num_input_tokens_seen": 6479032, "step": 9905 }, { "epoch": 5.19392033542977, "grad_norm": 0.18808841705322266, "learning_rate": 4.621645762296399e-05, "loss": 0.4368, "num_input_tokens_seen": 6482392, "step": 9910 }, { "epoch": 5.196540880503145, "grad_norm": 0.1710859090089798, "learning_rate": 4.6210407336543835e-05, "loss": 0.5719, "num_input_tokens_seen": 6485816, "step": 9915 }, { "epoch": 5.19916142557652, "grad_norm": 0.15431638062000275, "learning_rate": 4.620435261314844e-05, "loss": 0.4607, "num_input_tokens_seen": 6489112, "step": 9920 }, { "epoch": 5.201781970649895, "grad_norm": 0.2632644474506378, "learning_rate": 4.61982934540444e-05, "loss": 0.5276, "num_input_tokens_seen": 6492248, "step": 9925 }, { "epoch": 5.20440251572327, "grad_norm": 0.198106050491333, "learning_rate": 4.6192229860499206e-05, "loss": 0.583, "num_input_tokens_seen": 6496600, "step": 9930 }, { "epoch": 5.2070230607966455, "grad_norm": 0.14604173600673676, "learning_rate": 4.6186161833781295e-05, "loss": 0.3811, "num_input_tokens_seen": 6499256, "step": 9935 }, { "epoch": 5.209643605870021, "grad_norm": 0.24379447102546692, "learning_rate": 4.618008937516003e-05, "loss": 0.5421, "num_input_tokens_seen": 6502360, "step": 9940 }, { "epoch": 5.212264150943396, "grad_norm": 0.22440455853939056, "learning_rate": 4.617401248590572e-05, "loss": 0.3813, "num_input_tokens_seen": 6504792, "step": 9945 }, { "epoch": 5.214884696016772, "grad_norm": 0.1616443246603012, "learning_rate": 4.616793116728955e-05, "loss": 0.5031, "num_input_tokens_seen": 6507992, "step": 9950 }, { "epoch": 5.217505241090147, "grad_norm": 0.11511270701885223, "learning_rate": 4.6161845420583685e-05, "loss": 0.395, "num_input_tokens_seen": 6511128, "step": 9955 }, { "epoch": 5.220125786163522, "grad_norm": 0.11628727614879608, "learning_rate": 4.6155755247061174e-05, "loss": 0.626, "num_input_tokens_seen": 6514872, "step": 9960 }, { "epoch": 5.222746331236897, "grad_norm": 0.26936542987823486, "learning_rate": 4.614966064799603e-05, "loss": 0.4794, "num_input_tokens_seen": 6517464, "step": 9965 }, { "epoch": 5.2253668763102725, "grad_norm": 0.15273840725421906, "learning_rate": 4.6143561624663166e-05, "loss": 0.4622, "num_input_tokens_seen": 6520056, "step": 9970 }, { "epoch": 5.227987421383648, "grad_norm": 0.16955386102199554, "learning_rate": 4.6137458178338425e-05, "loss": 0.4737, "num_input_tokens_seen": 6522840, "step": 9975 }, { "epoch": 5.230607966457023, "grad_norm": 0.10506811738014221, "learning_rate": 4.613135031029859e-05, "loss": 0.4798, "num_input_tokens_seen": 6526424, "step": 9980 }, { "epoch": 5.233228511530398, "grad_norm": 0.11328978091478348, "learning_rate": 4.612523802182134e-05, "loss": 0.4974, "num_input_tokens_seen": 6530008, "step": 9985 }, { "epoch": 5.235849056603773, "grad_norm": 0.15588824450969696, "learning_rate": 4.611912131418531e-05, "loss": 0.4448, "num_input_tokens_seen": 6532760, "step": 9990 }, { "epoch": 5.238469601677149, "grad_norm": 0.1365135759115219, "learning_rate": 4.611300018867004e-05, "loss": 0.5401, "num_input_tokens_seen": 6536984, "step": 9995 }, { "epoch": 5.241090146750524, "grad_norm": 0.15266314148902893, "learning_rate": 4.6106874646556e-05, "loss": 0.5163, "num_input_tokens_seen": 6540696, "step": 10000 }, { "epoch": 5.2437106918239, "grad_norm": 0.12624570727348328, "learning_rate": 4.6100744689124584e-05, "loss": 0.4269, "num_input_tokens_seen": 6543960, "step": 10005 }, { "epoch": 5.246331236897275, "grad_norm": 0.43992263078689575, "learning_rate": 4.6094610317658116e-05, "loss": 0.4882, "num_input_tokens_seen": 6546712, "step": 10010 }, { "epoch": 5.24895178197065, "grad_norm": 0.1734180599451065, "learning_rate": 4.608847153343982e-05, "loss": 0.4728, "num_input_tokens_seen": 6549944, "step": 10015 }, { "epoch": 5.251572327044025, "grad_norm": 0.18282045423984528, "learning_rate": 4.6082328337753874e-05, "loss": 0.5675, "num_input_tokens_seen": 6552792, "step": 10020 }, { "epoch": 5.2541928721174, "grad_norm": 0.19092977046966553, "learning_rate": 4.6076180731885353e-05, "loss": 0.4578, "num_input_tokens_seen": 6555288, "step": 10025 }, { "epoch": 5.256813417190775, "grad_norm": 0.1468929499387741, "learning_rate": 4.6070028717120286e-05, "loss": 0.4057, "num_input_tokens_seen": 6558232, "step": 10030 }, { "epoch": 5.259433962264151, "grad_norm": 0.12811078131198883, "learning_rate": 4.606387229474558e-05, "loss": 0.5958, "num_input_tokens_seen": 6562040, "step": 10035 }, { "epoch": 5.262054507337526, "grad_norm": 0.5249856114387512, "learning_rate": 4.60577114660491e-05, "loss": 0.6229, "num_input_tokens_seen": 6564664, "step": 10040 }, { "epoch": 5.264675052410902, "grad_norm": 0.1384889781475067, "learning_rate": 4.605154623231962e-05, "loss": 0.5426, "num_input_tokens_seen": 6567608, "step": 10045 }, { "epoch": 5.267295597484277, "grad_norm": 0.15070946514606476, "learning_rate": 4.604537659484684e-05, "loss": 0.5983, "num_input_tokens_seen": 6570424, "step": 10050 }, { "epoch": 5.269916142557652, "grad_norm": 0.38419607281684875, "learning_rate": 4.603920255492138e-05, "loss": 0.5984, "num_input_tokens_seen": 6573624, "step": 10055 }, { "epoch": 5.272536687631027, "grad_norm": 0.16509857773780823, "learning_rate": 4.603302411383477e-05, "loss": 0.4978, "num_input_tokens_seen": 6576920, "step": 10060 }, { "epoch": 5.2751572327044025, "grad_norm": 0.11484166234731674, "learning_rate": 4.602684127287946e-05, "loss": 0.4356, "num_input_tokens_seen": 6580440, "step": 10065 }, { "epoch": 5.277777777777778, "grad_norm": 0.1577739715576172, "learning_rate": 4.602065403334885e-05, "loss": 0.4265, "num_input_tokens_seen": 6583064, "step": 10070 }, { "epoch": 5.280398322851153, "grad_norm": 0.18037310242652893, "learning_rate": 4.601446239653724e-05, "loss": 0.5066, "num_input_tokens_seen": 6586840, "step": 10075 }, { "epoch": 5.283018867924528, "grad_norm": 0.1870613843202591, "learning_rate": 4.600826636373984e-05, "loss": 0.6049, "num_input_tokens_seen": 6590296, "step": 10080 }, { "epoch": 5.285639412997903, "grad_norm": 0.1982107162475586, "learning_rate": 4.600206593625279e-05, "loss": 0.529, "num_input_tokens_seen": 6593368, "step": 10085 }, { "epoch": 5.288259958071279, "grad_norm": 0.1564379185438156, "learning_rate": 4.599586111537315e-05, "loss": 0.521, "num_input_tokens_seen": 6596856, "step": 10090 }, { "epoch": 5.290880503144654, "grad_norm": 0.26510658860206604, "learning_rate": 4.59896519023989e-05, "loss": 0.498, "num_input_tokens_seen": 6600184, "step": 10095 }, { "epoch": 5.29350104821803, "grad_norm": 0.1859801858663559, "learning_rate": 4.5983438298628935e-05, "loss": 0.5367, "num_input_tokens_seen": 6603608, "step": 10100 }, { "epoch": 5.296121593291405, "grad_norm": 0.25243693590164185, "learning_rate": 4.597722030536306e-05, "loss": 0.4905, "num_input_tokens_seen": 6606712, "step": 10105 }, { "epoch": 5.29874213836478, "grad_norm": 0.13552285730838776, "learning_rate": 4.597099792390202e-05, "loss": 0.4289, "num_input_tokens_seen": 6609624, "step": 10110 }, { "epoch": 5.301362683438155, "grad_norm": 0.12842519581317902, "learning_rate": 4.596477115554748e-05, "loss": 0.4377, "num_input_tokens_seen": 6613336, "step": 10115 }, { "epoch": 5.30398322851153, "grad_norm": 0.20815463364124298, "learning_rate": 4.5958540001601976e-05, "loss": 0.5592, "num_input_tokens_seen": 6617560, "step": 10120 }, { "epoch": 5.306603773584905, "grad_norm": 0.13553020358085632, "learning_rate": 4.595230446336902e-05, "loss": 0.5162, "num_input_tokens_seen": 6621496, "step": 10125 }, { "epoch": 5.309224318658281, "grad_norm": 0.257035493850708, "learning_rate": 4.5946064542153005e-05, "loss": 0.4141, "num_input_tokens_seen": 6625656, "step": 10130 }, { "epoch": 5.311844863731656, "grad_norm": 0.1313299983739853, "learning_rate": 4.593982023925926e-05, "loss": 0.3536, "num_input_tokens_seen": 6629240, "step": 10135 }, { "epoch": 5.314465408805032, "grad_norm": 0.139663964509964, "learning_rate": 4.5933571555994e-05, "loss": 0.5967, "num_input_tokens_seen": 6633784, "step": 10140 }, { "epoch": 5.317085953878407, "grad_norm": 0.15610426664352417, "learning_rate": 4.59273184936644e-05, "loss": 0.5869, "num_input_tokens_seen": 6636600, "step": 10145 }, { "epoch": 5.319706498951782, "grad_norm": 0.27288565039634705, "learning_rate": 4.5921061053578523e-05, "loss": 0.4762, "num_input_tokens_seen": 6640088, "step": 10150 }, { "epoch": 5.322327044025157, "grad_norm": 0.1770601123571396, "learning_rate": 4.5914799237045355e-05, "loss": 0.4922, "num_input_tokens_seen": 6642840, "step": 10155 }, { "epoch": 5.3249475890985325, "grad_norm": 0.11865950375795364, "learning_rate": 4.5908533045374796e-05, "loss": 0.4347, "num_input_tokens_seen": 6646136, "step": 10160 }, { "epoch": 5.327568134171908, "grad_norm": 0.3114774227142334, "learning_rate": 4.590226247987766e-05, "loss": 0.4262, "num_input_tokens_seen": 6649560, "step": 10165 }, { "epoch": 5.330188679245283, "grad_norm": 0.12900707125663757, "learning_rate": 4.589598754186568e-05, "loss": 0.4174, "num_input_tokens_seen": 6652888, "step": 10170 }, { "epoch": 5.332809224318658, "grad_norm": 0.1929120570421219, "learning_rate": 4.5889708232651494e-05, "loss": 0.4325, "num_input_tokens_seen": 6655576, "step": 10175 }, { "epoch": 5.335429769392033, "grad_norm": 0.2826760411262512, "learning_rate": 4.588342455354867e-05, "loss": 0.5104, "num_input_tokens_seen": 6659288, "step": 10180 }, { "epoch": 5.338050314465409, "grad_norm": 0.16771776974201202, "learning_rate": 4.5877136505871685e-05, "loss": 0.4639, "num_input_tokens_seen": 6662360, "step": 10185 }, { "epoch": 5.340670859538784, "grad_norm": 0.1050444096326828, "learning_rate": 4.5870844090935914e-05, "loss": 0.5072, "num_input_tokens_seen": 6666584, "step": 10190 }, { "epoch": 5.34329140461216, "grad_norm": 0.23618443310260773, "learning_rate": 4.586454731005768e-05, "loss": 0.7023, "num_input_tokens_seen": 6669176, "step": 10195 }, { "epoch": 5.345911949685535, "grad_norm": 0.1411769986152649, "learning_rate": 4.5858246164554175e-05, "loss": 0.6046, "num_input_tokens_seen": 6672728, "step": 10200 }, { "epoch": 5.34853249475891, "grad_norm": 0.12487757205963135, "learning_rate": 4.5851940655743536e-05, "loss": 0.3871, "num_input_tokens_seen": 6676024, "step": 10205 }, { "epoch": 5.351153039832285, "grad_norm": 0.15393942594528198, "learning_rate": 4.5845630784944804e-05, "loss": 0.6001, "num_input_tokens_seen": 6679224, "step": 10210 }, { "epoch": 5.35377358490566, "grad_norm": 0.3472099304199219, "learning_rate": 4.583931655347793e-05, "loss": 0.5478, "num_input_tokens_seen": 6681528, "step": 10215 }, { "epoch": 5.356394129979035, "grad_norm": 0.17008699476718903, "learning_rate": 4.5832997962663785e-05, "loss": 0.4204, "num_input_tokens_seen": 6683992, "step": 10220 }, { "epoch": 5.359014675052411, "grad_norm": 0.17320074141025543, "learning_rate": 4.582667501382414e-05, "loss": 0.4283, "num_input_tokens_seen": 6687576, "step": 10225 }, { "epoch": 5.361635220125786, "grad_norm": 0.14687423408031464, "learning_rate": 4.582034770828169e-05, "loss": 0.5675, "num_input_tokens_seen": 6690264, "step": 10230 }, { "epoch": 5.364255765199162, "grad_norm": 0.24608147144317627, "learning_rate": 4.581401604736002e-05, "loss": 0.4418, "num_input_tokens_seen": 6693400, "step": 10235 }, { "epoch": 5.366876310272537, "grad_norm": 0.1748395413160324, "learning_rate": 4.580768003238365e-05, "loss": 0.5478, "num_input_tokens_seen": 6695960, "step": 10240 }, { "epoch": 5.369496855345912, "grad_norm": 0.2110033929347992, "learning_rate": 4.5801339664678014e-05, "loss": 0.4856, "num_input_tokens_seen": 6699064, "step": 10245 }, { "epoch": 5.372117400419287, "grad_norm": 0.11077331006526947, "learning_rate": 4.579499494556943e-05, "loss": 0.4359, "num_input_tokens_seen": 6702008, "step": 10250 }, { "epoch": 5.3747379454926625, "grad_norm": 0.21186435222625732, "learning_rate": 4.5788645876385147e-05, "loss": 0.7307, "num_input_tokens_seen": 6705560, "step": 10255 }, { "epoch": 5.377358490566038, "grad_norm": 0.1978662759065628, "learning_rate": 4.578229245845331e-05, "loss": 0.6234, "num_input_tokens_seen": 6709016, "step": 10260 }, { "epoch": 5.379979035639413, "grad_norm": 0.2088550478219986, "learning_rate": 4.5775934693102994e-05, "loss": 0.5299, "num_input_tokens_seen": 6711736, "step": 10265 }, { "epoch": 5.382599580712788, "grad_norm": 0.1403336524963379, "learning_rate": 4.5769572581664165e-05, "loss": 0.479, "num_input_tokens_seen": 6714104, "step": 10270 }, { "epoch": 5.385220125786163, "grad_norm": 0.2589082419872284, "learning_rate": 4.57632061254677e-05, "loss": 0.5064, "num_input_tokens_seen": 6717720, "step": 10275 }, { "epoch": 5.387840670859539, "grad_norm": 0.19932442903518677, "learning_rate": 4.5756835325845394e-05, "loss": 0.5703, "num_input_tokens_seen": 6720728, "step": 10280 }, { "epoch": 5.390461215932914, "grad_norm": 0.18601903319358826, "learning_rate": 4.575046018412995e-05, "loss": 0.4941, "num_input_tokens_seen": 6724344, "step": 10285 }, { "epoch": 5.3930817610062896, "grad_norm": 0.13705000281333923, "learning_rate": 4.5744080701654955e-05, "loss": 0.5123, "num_input_tokens_seen": 6728280, "step": 10290 }, { "epoch": 5.395702306079665, "grad_norm": 0.15377770364284515, "learning_rate": 4.573769687975495e-05, "loss": 0.5048, "num_input_tokens_seen": 6730936, "step": 10295 }, { "epoch": 5.39832285115304, "grad_norm": 0.133610799908638, "learning_rate": 4.573130871976534e-05, "loss": 0.4829, "num_input_tokens_seen": 6733624, "step": 10300 }, { "epoch": 5.400943396226415, "grad_norm": 0.18409352004528046, "learning_rate": 4.572491622302247e-05, "loss": 0.5368, "num_input_tokens_seen": 6736184, "step": 10305 }, { "epoch": 5.40356394129979, "grad_norm": 0.13021080195903778, "learning_rate": 4.571851939086356e-05, "loss": 0.527, "num_input_tokens_seen": 6739192, "step": 10310 }, { "epoch": 5.406184486373165, "grad_norm": 0.16440518200397491, "learning_rate": 4.571211822462676e-05, "loss": 0.3871, "num_input_tokens_seen": 6742360, "step": 10315 }, { "epoch": 5.408805031446541, "grad_norm": 0.13657905161380768, "learning_rate": 4.570571272565114e-05, "loss": 0.5061, "num_input_tokens_seen": 6746200, "step": 10320 }, { "epoch": 5.411425576519916, "grad_norm": 0.10855061560869217, "learning_rate": 4.569930289527663e-05, "loss": 0.5217, "num_input_tokens_seen": 6750616, "step": 10325 }, { "epoch": 5.414046121593292, "grad_norm": 0.21888120472431183, "learning_rate": 4.56928887348441e-05, "loss": 0.5131, "num_input_tokens_seen": 6753688, "step": 10330 }, { "epoch": 5.416666666666667, "grad_norm": 0.09131351113319397, "learning_rate": 4.568647024569533e-05, "loss": 0.3945, "num_input_tokens_seen": 6756920, "step": 10335 }, { "epoch": 5.419287211740042, "grad_norm": 0.1576610803604126, "learning_rate": 4.568004742917299e-05, "loss": 0.4646, "num_input_tokens_seen": 6761048, "step": 10340 }, { "epoch": 5.421907756813417, "grad_norm": 0.13230213522911072, "learning_rate": 4.567362028662065e-05, "loss": 0.4003, "num_input_tokens_seen": 6764280, "step": 10345 }, { "epoch": 5.4245283018867925, "grad_norm": 0.3978637456893921, "learning_rate": 4.566718881938281e-05, "loss": 0.6592, "num_input_tokens_seen": 6767704, "step": 10350 }, { "epoch": 5.427148846960168, "grad_norm": 0.0980151817202568, "learning_rate": 4.5660753028804846e-05, "loss": 0.4627, "num_input_tokens_seen": 6772280, "step": 10355 }, { "epoch": 5.429769392033543, "grad_norm": 0.1820092499256134, "learning_rate": 4.5654312916233066e-05, "loss": 0.4149, "num_input_tokens_seen": 6775704, "step": 10360 }, { "epoch": 5.432389937106918, "grad_norm": 0.12473656237125397, "learning_rate": 4.564786848301466e-05, "loss": 0.5028, "num_input_tokens_seen": 6779288, "step": 10365 }, { "epoch": 5.435010482180293, "grad_norm": 0.23806484043598175, "learning_rate": 4.564141973049773e-05, "loss": 0.5033, "num_input_tokens_seen": 6781848, "step": 10370 }, { "epoch": 5.437631027253669, "grad_norm": 0.17681723833084106, "learning_rate": 4.563496666003128e-05, "loss": 0.4794, "num_input_tokens_seen": 6784920, "step": 10375 }, { "epoch": 5.440251572327044, "grad_norm": 0.1846752017736435, "learning_rate": 4.562850927296523e-05, "loss": 0.6015, "num_input_tokens_seen": 6788344, "step": 10380 }, { "epoch": 5.4428721174004195, "grad_norm": 0.20332631468772888, "learning_rate": 4.5622047570650365e-05, "loss": 0.6657, "num_input_tokens_seen": 6791000, "step": 10385 }, { "epoch": 5.445492662473795, "grad_norm": 0.16480864584445953, "learning_rate": 4.561558155443843e-05, "loss": 0.4991, "num_input_tokens_seen": 6793560, "step": 10390 }, { "epoch": 5.44811320754717, "grad_norm": 0.160554900765419, "learning_rate": 4.560911122568202e-05, "loss": 0.41, "num_input_tokens_seen": 6795896, "step": 10395 }, { "epoch": 5.450733752620545, "grad_norm": 0.2665640413761139, "learning_rate": 4.560263658573467e-05, "loss": 0.3528, "num_input_tokens_seen": 6798584, "step": 10400 }, { "epoch": 5.45335429769392, "grad_norm": 0.1999226212501526, "learning_rate": 4.559615763595079e-05, "loss": 0.437, "num_input_tokens_seen": 6801816, "step": 10405 }, { "epoch": 5.455974842767295, "grad_norm": 0.147107794880867, "learning_rate": 4.558967437768571e-05, "loss": 0.4737, "num_input_tokens_seen": 6805880, "step": 10410 }, { "epoch": 5.4585953878406706, "grad_norm": 0.14533217251300812, "learning_rate": 4.558318681229565e-05, "loss": 0.7521, "num_input_tokens_seen": 6809240, "step": 10415 }, { "epoch": 5.461215932914046, "grad_norm": 0.1396358758211136, "learning_rate": 4.557669494113773e-05, "loss": 0.4085, "num_input_tokens_seen": 6812344, "step": 10420 }, { "epoch": 5.463836477987422, "grad_norm": 0.10543348640203476, "learning_rate": 4.557019876556998e-05, "loss": 0.5719, "num_input_tokens_seen": 6815800, "step": 10425 }, { "epoch": 5.466457023060797, "grad_norm": 0.16079600155353546, "learning_rate": 4.556369828695133e-05, "loss": 0.4673, "num_input_tokens_seen": 6823736, "step": 10430 }, { "epoch": 5.469077568134172, "grad_norm": 0.2051958590745926, "learning_rate": 4.5557193506641596e-05, "loss": 0.4593, "num_input_tokens_seen": 6826520, "step": 10435 }, { "epoch": 5.471698113207547, "grad_norm": 0.2731103301048279, "learning_rate": 4.555068442600151e-05, "loss": 0.4662, "num_input_tokens_seen": 6829880, "step": 10440 }, { "epoch": 5.4743186582809225, "grad_norm": 0.19540175795555115, "learning_rate": 4.5544171046392703e-05, "loss": 0.5571, "num_input_tokens_seen": 6833752, "step": 10445 }, { "epoch": 5.476939203354298, "grad_norm": 0.26408642530441284, "learning_rate": 4.5537653369177685e-05, "loss": 0.637, "num_input_tokens_seen": 6837208, "step": 10450 }, { "epoch": 5.479559748427673, "grad_norm": 0.12026722729206085, "learning_rate": 4.5531131395719895e-05, "loss": 0.4984, "num_input_tokens_seen": 6843544, "step": 10455 }, { "epoch": 5.482180293501048, "grad_norm": 0.18307290971279144, "learning_rate": 4.552460512738364e-05, "loss": 0.6674, "num_input_tokens_seen": 6846488, "step": 10460 }, { "epoch": 5.484800838574423, "grad_norm": 0.17858189344406128, "learning_rate": 4.551807456553416e-05, "loss": 0.4638, "num_input_tokens_seen": 6850136, "step": 10465 }, { "epoch": 5.487421383647799, "grad_norm": 0.20453815162181854, "learning_rate": 4.551153971153755e-05, "loss": 0.5236, "num_input_tokens_seen": 6853016, "step": 10470 }, { "epoch": 5.490041928721174, "grad_norm": 0.1908484399318695, "learning_rate": 4.550500056676085e-05, "loss": 0.5186, "num_input_tokens_seen": 6856088, "step": 10475 }, { "epoch": 5.4926624737945495, "grad_norm": 0.15756160020828247, "learning_rate": 4.5498457132571966e-05, "loss": 0.5998, "num_input_tokens_seen": 6859288, "step": 10480 }, { "epoch": 5.495283018867925, "grad_norm": 0.14919088780879974, "learning_rate": 4.5491909410339706e-05, "loss": 0.4391, "num_input_tokens_seen": 6863480, "step": 10485 }, { "epoch": 5.4979035639413, "grad_norm": 0.20413687825202942, "learning_rate": 4.548535740143378e-05, "loss": 0.2786, "num_input_tokens_seen": 6869656, "step": 10490 }, { "epoch": 5.500524109014675, "grad_norm": 0.14886616170406342, "learning_rate": 4.54788011072248e-05, "loss": 0.4104, "num_input_tokens_seen": 6872344, "step": 10495 }, { "epoch": 5.50314465408805, "grad_norm": 0.15369276702404022, "learning_rate": 4.5472240529084254e-05, "loss": 0.4683, "num_input_tokens_seen": 6876216, "step": 10500 }, { "epoch": 5.505765199161425, "grad_norm": 0.1274702250957489, "learning_rate": 4.5465675668384557e-05, "loss": 0.3933, "num_input_tokens_seen": 6879384, "step": 10505 }, { "epoch": 5.5083857442348005, "grad_norm": 0.48897361755371094, "learning_rate": 4.545910652649899e-05, "loss": 0.6102, "num_input_tokens_seen": 6882232, "step": 10510 }, { "epoch": 5.511006289308176, "grad_norm": 0.20078061521053314, "learning_rate": 4.5452533104801756e-05, "loss": 0.5581, "num_input_tokens_seen": 6886072, "step": 10515 }, { "epoch": 5.513626834381551, "grad_norm": 0.12325750291347504, "learning_rate": 4.544595540466792e-05, "loss": 0.4228, "num_input_tokens_seen": 6889272, "step": 10520 }, { "epoch": 5.516247379454927, "grad_norm": 0.14811429381370544, "learning_rate": 4.5439373427473476e-05, "loss": 0.5862, "num_input_tokens_seen": 6892824, "step": 10525 }, { "epoch": 5.518867924528302, "grad_norm": 0.17595842480659485, "learning_rate": 4.54327871745953e-05, "loss": 0.5147, "num_input_tokens_seen": 6896248, "step": 10530 }, { "epoch": 5.521488469601677, "grad_norm": 0.1360546052455902, "learning_rate": 4.542619664741116e-05, "loss": 0.5401, "num_input_tokens_seen": 6899640, "step": 10535 }, { "epoch": 5.524109014675052, "grad_norm": 0.14320242404937744, "learning_rate": 4.5419601847299706e-05, "loss": 0.4668, "num_input_tokens_seen": 6903896, "step": 10540 }, { "epoch": 5.526729559748428, "grad_norm": 0.18358854949474335, "learning_rate": 4.5413002775640503e-05, "loss": 0.5838, "num_input_tokens_seen": 6907448, "step": 10545 }, { "epoch": 5.529350104821803, "grad_norm": 0.11615972965955734, "learning_rate": 4.540639943381401e-05, "loss": 0.509, "num_input_tokens_seen": 6910840, "step": 10550 }, { "epoch": 5.531970649895178, "grad_norm": 0.17269988358020782, "learning_rate": 4.539979182320157e-05, "loss": 0.4354, "num_input_tokens_seen": 6913528, "step": 10555 }, { "epoch": 5.534591194968553, "grad_norm": 0.2299223244190216, "learning_rate": 4.5393179945185404e-05, "loss": 0.5391, "num_input_tokens_seen": 6916312, "step": 10560 }, { "epoch": 5.537211740041929, "grad_norm": 0.1266108900308609, "learning_rate": 4.538656380114865e-05, "loss": 0.477, "num_input_tokens_seen": 6919384, "step": 10565 }, { "epoch": 5.539832285115304, "grad_norm": 0.2156437635421753, "learning_rate": 4.537994339247533e-05, "loss": 0.6026, "num_input_tokens_seen": 6921816, "step": 10570 }, { "epoch": 5.5424528301886795, "grad_norm": 0.19732166826725006, "learning_rate": 4.5373318720550366e-05, "loss": 0.5863, "num_input_tokens_seen": 6924760, "step": 10575 }, { "epoch": 5.545073375262055, "grad_norm": 0.09109058231115341, "learning_rate": 4.536668978675956e-05, "loss": 0.4513, "num_input_tokens_seen": 6928600, "step": 10580 }, { "epoch": 5.54769392033543, "grad_norm": 0.2788042724132538, "learning_rate": 4.53600565924896e-05, "loss": 0.4084, "num_input_tokens_seen": 6932120, "step": 10585 }, { "epoch": 5.550314465408805, "grad_norm": 0.19856038689613342, "learning_rate": 4.535341913912808e-05, "loss": 0.5041, "num_input_tokens_seen": 6934968, "step": 10590 }, { "epoch": 5.55293501048218, "grad_norm": 0.14854338765144348, "learning_rate": 4.534677742806348e-05, "loss": 0.5258, "num_input_tokens_seen": 6938776, "step": 10595 }, { "epoch": 5.555555555555555, "grad_norm": 0.26841866970062256, "learning_rate": 4.5340131460685175e-05, "loss": 0.4871, "num_input_tokens_seen": 6942200, "step": 10600 }, { "epoch": 5.5581761006289305, "grad_norm": 0.16471359133720398, "learning_rate": 4.533348123838341e-05, "loss": 0.7498, "num_input_tokens_seen": 6944888, "step": 10605 }, { "epoch": 5.560796645702306, "grad_norm": 0.23227669298648834, "learning_rate": 4.532682676254936e-05, "loss": 0.4788, "num_input_tokens_seen": 6947896, "step": 10610 }, { "epoch": 5.563417190775681, "grad_norm": 0.22750505805015564, "learning_rate": 4.5320168034575044e-05, "loss": 0.6592, "num_input_tokens_seen": 6950712, "step": 10615 }, { "epoch": 5.566037735849057, "grad_norm": 0.20132549107074738, "learning_rate": 4.53135050558534e-05, "loss": 0.6602, "num_input_tokens_seen": 6953496, "step": 10620 }, { "epoch": 5.568658280922432, "grad_norm": 0.15224090218544006, "learning_rate": 4.530683782777825e-05, "loss": 0.5023, "num_input_tokens_seen": 6956152, "step": 10625 }, { "epoch": 5.571278825995807, "grad_norm": 0.18288487195968628, "learning_rate": 4.53001663517443e-05, "loss": 0.4594, "num_input_tokens_seen": 6959032, "step": 10630 }, { "epoch": 5.573899371069182, "grad_norm": 0.10997270792722702, "learning_rate": 4.529349062914714e-05, "loss": 0.4102, "num_input_tokens_seen": 6962296, "step": 10635 }, { "epoch": 5.576519916142558, "grad_norm": 0.17225684225559235, "learning_rate": 4.528681066138326e-05, "loss": 0.5477, "num_input_tokens_seen": 6965784, "step": 10640 }, { "epoch": 5.579140461215933, "grad_norm": 0.36137527227401733, "learning_rate": 4.528012644985005e-05, "loss": 0.5458, "num_input_tokens_seen": 6969112, "step": 10645 }, { "epoch": 5.581761006289308, "grad_norm": 0.1760893017053604, "learning_rate": 4.527343799594574e-05, "loss": 0.5162, "num_input_tokens_seen": 6972216, "step": 10650 }, { "epoch": 5.584381551362683, "grad_norm": 0.11777137219905853, "learning_rate": 4.5266745301069494e-05, "loss": 0.6176, "num_input_tokens_seen": 6975064, "step": 10655 }, { "epoch": 5.587002096436059, "grad_norm": 0.1899840235710144, "learning_rate": 4.526004836662135e-05, "loss": 0.5265, "num_input_tokens_seen": 6977688, "step": 10660 }, { "epoch": 5.589622641509434, "grad_norm": 0.18767337501049042, "learning_rate": 4.525334719400223e-05, "loss": 0.4785, "num_input_tokens_seen": 6981208, "step": 10665 }, { "epoch": 5.5922431865828095, "grad_norm": 0.3260478675365448, "learning_rate": 4.524664178461393e-05, "loss": 0.5432, "num_input_tokens_seen": 6983960, "step": 10670 }, { "epoch": 5.594863731656185, "grad_norm": 0.209536612033844, "learning_rate": 4.5239932139859154e-05, "loss": 0.5577, "num_input_tokens_seen": 6987320, "step": 10675 }, { "epoch": 5.59748427672956, "grad_norm": 0.1559479981660843, "learning_rate": 4.523321826114149e-05, "loss": 0.5027, "num_input_tokens_seen": 6991064, "step": 10680 }, { "epoch": 5.600104821802935, "grad_norm": 0.20212718844413757, "learning_rate": 4.5226500149865395e-05, "loss": 0.5316, "num_input_tokens_seen": 6993624, "step": 10685 }, { "epoch": 5.60272536687631, "grad_norm": 0.15839935839176178, "learning_rate": 4.5219777807436224e-05, "loss": 0.5352, "num_input_tokens_seen": 6996344, "step": 10690 }, { "epoch": 5.605345911949685, "grad_norm": 0.20607173442840576, "learning_rate": 4.5213051235260225e-05, "loss": 0.5041, "num_input_tokens_seen": 6999576, "step": 10695 }, { "epoch": 5.6079664570230605, "grad_norm": 0.2038048654794693, "learning_rate": 4.52063204347445e-05, "loss": 0.5435, "num_input_tokens_seen": 7003096, "step": 10700 }, { "epoch": 5.610587002096436, "grad_norm": 0.1617356687784195, "learning_rate": 4.519958540729707e-05, "loss": 0.4796, "num_input_tokens_seen": 7006328, "step": 10705 }, { "epoch": 5.613207547169811, "grad_norm": 0.16499543190002441, "learning_rate": 4.519284615432682e-05, "loss": 0.5731, "num_input_tokens_seen": 7009560, "step": 10710 }, { "epoch": 5.615828092243187, "grad_norm": 0.13968172669410706, "learning_rate": 4.518610267724352e-05, "loss": 0.5764, "num_input_tokens_seen": 7012216, "step": 10715 }, { "epoch": 5.618448637316562, "grad_norm": 0.1931256502866745, "learning_rate": 4.517935497745786e-05, "loss": 0.5239, "num_input_tokens_seen": 7015224, "step": 10720 }, { "epoch": 5.621069182389937, "grad_norm": 0.16791445016860962, "learning_rate": 4.517260305638133e-05, "loss": 0.5304, "num_input_tokens_seen": 7018296, "step": 10725 }, { "epoch": 5.623689727463312, "grad_norm": 0.11502692103385925, "learning_rate": 4.516584691542639e-05, "loss": 0.4543, "num_input_tokens_seen": 7022424, "step": 10730 }, { "epoch": 5.626310272536688, "grad_norm": 0.13992194831371307, "learning_rate": 4.515908655600634e-05, "loss": 0.4254, "num_input_tokens_seen": 7026456, "step": 10735 }, { "epoch": 5.628930817610063, "grad_norm": 0.16721810400485992, "learning_rate": 4.515232197953537e-05, "loss": 0.6392, "num_input_tokens_seen": 7029880, "step": 10740 }, { "epoch": 5.631551362683438, "grad_norm": 0.1641012728214264, "learning_rate": 4.514555318742856e-05, "loss": 0.4696, "num_input_tokens_seen": 7033080, "step": 10745 }, { "epoch": 5.634171907756813, "grad_norm": 0.15451303124427795, "learning_rate": 4.513878018110184e-05, "loss": 0.5378, "num_input_tokens_seen": 7036760, "step": 10750 }, { "epoch": 5.636792452830189, "grad_norm": 0.21633991599082947, "learning_rate": 4.513200296197207e-05, "loss": 0.5367, "num_input_tokens_seen": 7040920, "step": 10755 }, { "epoch": 5.639412997903564, "grad_norm": 0.14701679348945618, "learning_rate": 4.512522153145696e-05, "loss": 0.4921, "num_input_tokens_seen": 7043416, "step": 10760 }, { "epoch": 5.6420335429769395, "grad_norm": 0.2829729914665222, "learning_rate": 4.5118435890975106e-05, "loss": 0.5183, "num_input_tokens_seen": 7046296, "step": 10765 }, { "epoch": 5.644654088050315, "grad_norm": 0.20145703852176666, "learning_rate": 4.511164604194598e-05, "loss": 0.3701, "num_input_tokens_seen": 7050008, "step": 10770 }, { "epoch": 5.64727463312369, "grad_norm": 0.14993228018283844, "learning_rate": 4.5104851985789956e-05, "loss": 0.4907, "num_input_tokens_seen": 7053240, "step": 10775 }, { "epoch": 5.649895178197065, "grad_norm": 0.218993678689003, "learning_rate": 4.5098053723928254e-05, "loss": 0.5672, "num_input_tokens_seen": 7056312, "step": 10780 }, { "epoch": 5.65251572327044, "grad_norm": 0.11743247509002686, "learning_rate": 4.509125125778302e-05, "loss": 0.3844, "num_input_tokens_seen": 7059960, "step": 10785 }, { "epoch": 5.655136268343815, "grad_norm": 0.2018202245235443, "learning_rate": 4.508444458877722e-05, "loss": 0.5522, "num_input_tokens_seen": 7062936, "step": 10790 }, { "epoch": 5.6577568134171905, "grad_norm": 0.23967745900154114, "learning_rate": 4.5077633718334755e-05, "loss": 0.5523, "num_input_tokens_seen": 7065368, "step": 10795 }, { "epoch": 5.660377358490566, "grad_norm": 0.15395264327526093, "learning_rate": 4.5070818647880375e-05, "loss": 0.6102, "num_input_tokens_seen": 7069464, "step": 10800 }, { "epoch": 5.662997903563941, "grad_norm": 0.23040500283241272, "learning_rate": 4.506399937883972e-05, "loss": 0.4237, "num_input_tokens_seen": 7071768, "step": 10805 }, { "epoch": 5.665618448637317, "grad_norm": 0.1369907110929489, "learning_rate": 4.5057175912639286e-05, "loss": 0.4107, "num_input_tokens_seen": 7075704, "step": 10810 }, { "epoch": 5.668238993710692, "grad_norm": 0.18406960368156433, "learning_rate": 4.5050348250706475e-05, "loss": 0.5054, "num_input_tokens_seen": 7078264, "step": 10815 }, { "epoch": 5.670859538784067, "grad_norm": 0.2562721371650696, "learning_rate": 4.5043516394469566e-05, "loss": 0.5297, "num_input_tokens_seen": 7082424, "step": 10820 }, { "epoch": 5.673480083857442, "grad_norm": 0.1869906634092331, "learning_rate": 4.503668034535768e-05, "loss": 0.4728, "num_input_tokens_seen": 7085624, "step": 10825 }, { "epoch": 5.676100628930818, "grad_norm": 0.12192525714635849, "learning_rate": 4.5029840104800876e-05, "loss": 0.4521, "num_input_tokens_seen": 7089112, "step": 10830 }, { "epoch": 5.678721174004193, "grad_norm": 0.13927003741264343, "learning_rate": 4.5022995674230025e-05, "loss": 0.4175, "num_input_tokens_seen": 7092728, "step": 10835 }, { "epoch": 5.681341719077568, "grad_norm": 0.0999540239572525, "learning_rate": 4.5016147055076914e-05, "loss": 0.6568, "num_input_tokens_seen": 7095544, "step": 10840 }, { "epoch": 5.683962264150943, "grad_norm": 0.1720608025789261, "learning_rate": 4.500929424877419e-05, "loss": 0.4749, "num_input_tokens_seen": 7098264, "step": 10845 }, { "epoch": 5.686582809224319, "grad_norm": 0.14291222393512726, "learning_rate": 4.5002437256755394e-05, "loss": 0.4968, "num_input_tokens_seen": 7101464, "step": 10850 }, { "epoch": 5.689203354297694, "grad_norm": 0.24606171250343323, "learning_rate": 4.4995576080454924e-05, "loss": 0.5845, "num_input_tokens_seen": 7105080, "step": 10855 }, { "epoch": 5.6918238993710695, "grad_norm": 0.20437973737716675, "learning_rate": 4.4988710721308056e-05, "loss": 0.533, "num_input_tokens_seen": 7107800, "step": 10860 }, { "epoch": 5.694444444444445, "grad_norm": 0.20553013682365417, "learning_rate": 4.498184118075095e-05, "loss": 0.4852, "num_input_tokens_seen": 7110712, "step": 10865 }, { "epoch": 5.69706498951782, "grad_norm": 0.16137807071208954, "learning_rate": 4.497496746022063e-05, "loss": 0.7166, "num_input_tokens_seen": 7114520, "step": 10870 }, { "epoch": 5.699685534591195, "grad_norm": 0.2749122381210327, "learning_rate": 4.4968089561155015e-05, "loss": 0.3836, "num_input_tokens_seen": 7118232, "step": 10875 }, { "epoch": 5.70230607966457, "grad_norm": 0.16062834858894348, "learning_rate": 4.496120748499286e-05, "loss": 0.5952, "num_input_tokens_seen": 7121080, "step": 10880 }, { "epoch": 5.704926624737945, "grad_norm": 0.371945858001709, "learning_rate": 4.4954321233173833e-05, "loss": 0.5478, "num_input_tokens_seen": 7124792, "step": 10885 }, { "epoch": 5.7075471698113205, "grad_norm": 0.170213982462883, "learning_rate": 4.494743080713846e-05, "loss": 0.3878, "num_input_tokens_seen": 7127640, "step": 10890 }, { "epoch": 5.710167714884696, "grad_norm": 0.18140314519405365, "learning_rate": 4.494053620832812e-05, "loss": 0.486, "num_input_tokens_seen": 7130392, "step": 10895 }, { "epoch": 5.712788259958071, "grad_norm": 0.11120877414941788, "learning_rate": 4.493363743818511e-05, "loss": 0.4954, "num_input_tokens_seen": 7133912, "step": 10900 }, { "epoch": 5.715408805031447, "grad_norm": 0.13414078950881958, "learning_rate": 4.492673449815257e-05, "loss": 0.4455, "num_input_tokens_seen": 7137272, "step": 10905 }, { "epoch": 5.718029350104822, "grad_norm": 0.2607403099536896, "learning_rate": 4.4919827389674496e-05, "loss": 0.6494, "num_input_tokens_seen": 7140152, "step": 10910 }, { "epoch": 5.720649895178197, "grad_norm": 0.14843067526817322, "learning_rate": 4.491291611419579e-05, "loss": 0.4975, "num_input_tokens_seen": 7143256, "step": 10915 }, { "epoch": 5.723270440251572, "grad_norm": 0.1445314884185791, "learning_rate": 4.4906000673162216e-05, "loss": 0.5469, "num_input_tokens_seen": 7146456, "step": 10920 }, { "epoch": 5.725890985324948, "grad_norm": 0.17769216001033783, "learning_rate": 4.48990810680204e-05, "loss": 0.5325, "num_input_tokens_seen": 7149432, "step": 10925 }, { "epoch": 5.728511530398323, "grad_norm": 0.20952920615673065, "learning_rate": 4.4892157300217844e-05, "loss": 0.4874, "num_input_tokens_seen": 7155640, "step": 10930 }, { "epoch": 5.731132075471698, "grad_norm": 0.1300675868988037, "learning_rate": 4.488522937120293e-05, "loss": 0.5076, "num_input_tokens_seen": 7158136, "step": 10935 }, { "epoch": 5.733752620545073, "grad_norm": 0.39438578486442566, "learning_rate": 4.4878297282424886e-05, "loss": 0.6467, "num_input_tokens_seen": 7161208, "step": 10940 }, { "epoch": 5.736373165618449, "grad_norm": 0.08311542868614197, "learning_rate": 4.4871361035333836e-05, "loss": 0.338, "num_input_tokens_seen": 7167160, "step": 10945 }, { "epoch": 5.738993710691824, "grad_norm": 0.20214472711086273, "learning_rate": 4.486442063138076e-05, "loss": 0.4495, "num_input_tokens_seen": 7170072, "step": 10950 }, { "epoch": 5.7416142557651995, "grad_norm": 0.15260422229766846, "learning_rate": 4.4857476072017516e-05, "loss": 0.4779, "num_input_tokens_seen": 7173368, "step": 10955 }, { "epoch": 5.744234800838575, "grad_norm": 0.14090484380722046, "learning_rate": 4.485052735869682e-05, "loss": 0.5701, "num_input_tokens_seen": 7176504, "step": 10960 }, { "epoch": 5.74685534591195, "grad_norm": 0.20678363740444183, "learning_rate": 4.484357449287228e-05, "loss": 0.6435, "num_input_tokens_seen": 7179672, "step": 10965 }, { "epoch": 5.749475890985325, "grad_norm": 0.16448800265789032, "learning_rate": 4.483661747599833e-05, "loss": 0.3774, "num_input_tokens_seen": 7182360, "step": 10970 }, { "epoch": 5.7520964360587, "grad_norm": 0.14259226620197296, "learning_rate": 4.482965630953031e-05, "loss": 0.4969, "num_input_tokens_seen": 7186040, "step": 10975 }, { "epoch": 5.754716981132075, "grad_norm": 0.1528768539428711, "learning_rate": 4.4822690994924425e-05, "loss": 0.45, "num_input_tokens_seen": 7189368, "step": 10980 }, { "epoch": 5.7573375262054505, "grad_norm": 0.25765761733055115, "learning_rate": 4.481572153363773e-05, "loss": 0.4766, "num_input_tokens_seen": 7192664, "step": 10985 }, { "epoch": 5.759958071278826, "grad_norm": 0.156350776553154, "learning_rate": 4.480874792712817e-05, "loss": 0.4631, "num_input_tokens_seen": 7196120, "step": 10990 }, { "epoch": 5.762578616352201, "grad_norm": 0.1566162407398224, "learning_rate": 4.480177017685452e-05, "loss": 0.4914, "num_input_tokens_seen": 7199992, "step": 10995 }, { "epoch": 5.765199161425577, "grad_norm": 0.23489083349704742, "learning_rate": 4.479478828427647e-05, "loss": 0.6791, "num_input_tokens_seen": 7203096, "step": 11000 }, { "epoch": 5.767819706498952, "grad_norm": 0.12786711752414703, "learning_rate": 4.478780225085454e-05, "loss": 0.5648, "num_input_tokens_seen": 7206200, "step": 11005 }, { "epoch": 5.770440251572327, "grad_norm": 0.30498364567756653, "learning_rate": 4.4780812078050125e-05, "loss": 0.5862, "num_input_tokens_seen": 7209656, "step": 11010 }, { "epoch": 5.773060796645702, "grad_norm": 0.2001933753490448, "learning_rate": 4.47738177673255e-05, "loss": 0.6734, "num_input_tokens_seen": 7213272, "step": 11015 }, { "epoch": 5.7756813417190775, "grad_norm": 0.15024511516094208, "learning_rate": 4.4766819320143785e-05, "loss": 0.5036, "num_input_tokens_seen": 7216728, "step": 11020 }, { "epoch": 5.778301886792453, "grad_norm": 0.29093196988105774, "learning_rate": 4.475981673796899e-05, "loss": 0.5153, "num_input_tokens_seen": 7219480, "step": 11025 }, { "epoch": 5.780922431865828, "grad_norm": 0.10862193256616592, "learning_rate": 4.475281002226596e-05, "loss": 0.482, "num_input_tokens_seen": 7224376, "step": 11030 }, { "epoch": 5.783542976939203, "grad_norm": 0.12296116352081299, "learning_rate": 4.4745799174500414e-05, "loss": 0.3916, "num_input_tokens_seen": 7227704, "step": 11035 }, { "epoch": 5.786163522012579, "grad_norm": 0.12756168842315674, "learning_rate": 4.473878419613896e-05, "loss": 0.4356, "num_input_tokens_seen": 7231928, "step": 11040 }, { "epoch": 5.788784067085954, "grad_norm": 0.18253350257873535, "learning_rate": 4.473176508864905e-05, "loss": 0.4409, "num_input_tokens_seen": 7236984, "step": 11045 }, { "epoch": 5.7914046121593294, "grad_norm": 0.1658063679933548, "learning_rate": 4.4724741853498995e-05, "loss": 0.5078, "num_input_tokens_seen": 7239704, "step": 11050 }, { "epoch": 5.794025157232705, "grad_norm": 0.2056249976158142, "learning_rate": 4.471771449215797e-05, "loss": 0.5163, "num_input_tokens_seen": 7242936, "step": 11055 }, { "epoch": 5.79664570230608, "grad_norm": 0.141619011759758, "learning_rate": 4.471068300609603e-05, "loss": 0.5629, "num_input_tokens_seen": 7245976, "step": 11060 }, { "epoch": 5.799266247379455, "grad_norm": 0.3069094717502594, "learning_rate": 4.4703647396784065e-05, "loss": 0.5576, "num_input_tokens_seen": 7249304, "step": 11065 }, { "epoch": 5.80188679245283, "grad_norm": 0.1781919002532959, "learning_rate": 4.469660766569387e-05, "loss": 0.5903, "num_input_tokens_seen": 7252344, "step": 11070 }, { "epoch": 5.804507337526205, "grad_norm": 0.13561341166496277, "learning_rate": 4.4689563814298057e-05, "loss": 0.5449, "num_input_tokens_seen": 7257144, "step": 11075 }, { "epoch": 5.8071278825995805, "grad_norm": 0.1738152951002121, "learning_rate": 4.468251584407012e-05, "loss": 0.7037, "num_input_tokens_seen": 7260376, "step": 11080 }, { "epoch": 5.809748427672956, "grad_norm": 0.18517114222049713, "learning_rate": 4.467546375648443e-05, "loss": 0.4875, "num_input_tokens_seen": 7263256, "step": 11085 }, { "epoch": 5.812368972746331, "grad_norm": 0.20889228582382202, "learning_rate": 4.4668407553016176e-05, "loss": 0.5537, "num_input_tokens_seen": 7265688, "step": 11090 }, { "epoch": 5.814989517819707, "grad_norm": 0.12142499536275864, "learning_rate": 4.466134723514146e-05, "loss": 0.5381, "num_input_tokens_seen": 7268280, "step": 11095 }, { "epoch": 5.817610062893082, "grad_norm": 0.18679741024971008, "learning_rate": 4.465428280433721e-05, "loss": 0.4425, "num_input_tokens_seen": 7271640, "step": 11100 }, { "epoch": 5.820230607966457, "grad_norm": 0.1657831072807312, "learning_rate": 4.464721426208123e-05, "loss": 0.3197, "num_input_tokens_seen": 7274040, "step": 11105 }, { "epoch": 5.822851153039832, "grad_norm": 0.14869533479213715, "learning_rate": 4.464014160985217e-05, "loss": 0.5976, "num_input_tokens_seen": 7276888, "step": 11110 }, { "epoch": 5.8254716981132075, "grad_norm": 0.18206246197223663, "learning_rate": 4.463306484912955e-05, "loss": 0.6108, "num_input_tokens_seen": 7280280, "step": 11115 }, { "epoch": 5.828092243186583, "grad_norm": 0.1180608943104744, "learning_rate": 4.462598398139375e-05, "loss": 0.6787, "num_input_tokens_seen": 7283448, "step": 11120 }, { "epoch": 5.830712788259958, "grad_norm": 0.17519712448120117, "learning_rate": 4.4618899008126005e-05, "loss": 0.7819, "num_input_tokens_seen": 7286168, "step": 11125 }, { "epoch": 5.833333333333333, "grad_norm": 0.16950999200344086, "learning_rate": 4.461180993080841e-05, "loss": 0.5676, "num_input_tokens_seen": 7289688, "step": 11130 }, { "epoch": 5.835953878406709, "grad_norm": 0.15144756436347961, "learning_rate": 4.460471675092393e-05, "loss": 0.5688, "num_input_tokens_seen": 7292664, "step": 11135 }, { "epoch": 5.838574423480084, "grad_norm": 0.1842367947101593, "learning_rate": 4.4597619469956356e-05, "loss": 0.4882, "num_input_tokens_seen": 7295928, "step": 11140 }, { "epoch": 5.841194968553459, "grad_norm": 0.15578626096248627, "learning_rate": 4.459051808939038e-05, "loss": 0.5561, "num_input_tokens_seen": 7298424, "step": 11145 }, { "epoch": 5.843815513626835, "grad_norm": 0.24954372644424438, "learning_rate": 4.458341261071151e-05, "loss": 0.5056, "num_input_tokens_seen": 7301624, "step": 11150 }, { "epoch": 5.84643605870021, "grad_norm": 0.19061343371868134, "learning_rate": 4.457630303540615e-05, "loss": 0.7277, "num_input_tokens_seen": 7304536, "step": 11155 }, { "epoch": 5.849056603773585, "grad_norm": 0.2395208477973938, "learning_rate": 4.456918936496154e-05, "loss": 0.5694, "num_input_tokens_seen": 7306904, "step": 11160 }, { "epoch": 5.85167714884696, "grad_norm": 0.1787465214729309, "learning_rate": 4.456207160086575e-05, "loss": 0.51, "num_input_tokens_seen": 7308952, "step": 11165 }, { "epoch": 5.854297693920335, "grad_norm": 0.128655806183815, "learning_rate": 4.4554949744607775e-05, "loss": 0.6605, "num_input_tokens_seen": 7312600, "step": 11170 }, { "epoch": 5.8569182389937104, "grad_norm": 0.2096102237701416, "learning_rate": 4.4547823797677405e-05, "loss": 0.5605, "num_input_tokens_seen": 7315416, "step": 11175 }, { "epoch": 5.859538784067086, "grad_norm": 0.1621171236038208, "learning_rate": 4.45406937615653e-05, "loss": 0.5583, "num_input_tokens_seen": 7318296, "step": 11180 }, { "epoch": 5.862159329140461, "grad_norm": 0.2589348256587982, "learning_rate": 4.4533559637763006e-05, "loss": 0.4577, "num_input_tokens_seen": 7320728, "step": 11185 }, { "epoch": 5.864779874213837, "grad_norm": 0.12121694535017014, "learning_rate": 4.452642142776288e-05, "loss": 0.5068, "num_input_tokens_seen": 7323928, "step": 11190 }, { "epoch": 5.867400419287212, "grad_norm": 0.17511257529258728, "learning_rate": 4.451927913305816e-05, "loss": 0.5527, "num_input_tokens_seen": 7327064, "step": 11195 }, { "epoch": 5.870020964360587, "grad_norm": 0.1713535636663437, "learning_rate": 4.4512132755142935e-05, "loss": 0.5084, "num_input_tokens_seen": 7329656, "step": 11200 }, { "epoch": 5.872641509433962, "grad_norm": 0.15955789387226105, "learning_rate": 4.450498229551214e-05, "loss": 0.4768, "num_input_tokens_seen": 7332760, "step": 11205 }, { "epoch": 5.8752620545073375, "grad_norm": 0.20219109952449799, "learning_rate": 4.449782775566158e-05, "loss": 0.6195, "num_input_tokens_seen": 7336344, "step": 11210 }, { "epoch": 5.877882599580713, "grad_norm": 0.16227561235427856, "learning_rate": 4.449066913708789e-05, "loss": 0.5986, "num_input_tokens_seen": 7339064, "step": 11215 }, { "epoch": 5.880503144654088, "grad_norm": 0.2512722313404083, "learning_rate": 4.448350644128858e-05, "loss": 0.7139, "num_input_tokens_seen": 7341848, "step": 11220 }, { "epoch": 5.883123689727463, "grad_norm": 0.1542041152715683, "learning_rate": 4.4476339669762e-05, "loss": 0.6234, "num_input_tokens_seen": 7344632, "step": 11225 }, { "epoch": 5.885744234800838, "grad_norm": 0.15732184052467346, "learning_rate": 4.446916882400737e-05, "loss": 0.4255, "num_input_tokens_seen": 7347512, "step": 11230 }, { "epoch": 5.888364779874214, "grad_norm": 0.23985548317432404, "learning_rate": 4.446199390552472e-05, "loss": 0.5715, "num_input_tokens_seen": 7350424, "step": 11235 }, { "epoch": 5.890985324947589, "grad_norm": 0.19300268590450287, "learning_rate": 4.445481491581499e-05, "loss": 0.5492, "num_input_tokens_seen": 7354232, "step": 11240 }, { "epoch": 5.893605870020965, "grad_norm": 0.2683989405632019, "learning_rate": 4.444763185637993e-05, "loss": 0.4989, "num_input_tokens_seen": 7356632, "step": 11245 }, { "epoch": 5.89622641509434, "grad_norm": 0.18279924988746643, "learning_rate": 4.4440444728722166e-05, "loss": 0.4534, "num_input_tokens_seen": 7359256, "step": 11250 }, { "epoch": 5.898846960167715, "grad_norm": 0.17846371233463287, "learning_rate": 4.4433253534345136e-05, "loss": 0.3898, "num_input_tokens_seen": 7362200, "step": 11255 }, { "epoch": 5.90146750524109, "grad_norm": 0.13723330199718475, "learning_rate": 4.442605827475318e-05, "loss": 0.3704, "num_input_tokens_seen": 7364664, "step": 11260 }, { "epoch": 5.904088050314465, "grad_norm": 0.23109902441501617, "learning_rate": 4.4418858951451457e-05, "loss": 0.5114, "num_input_tokens_seen": 7368248, "step": 11265 }, { "epoch": 5.90670859538784, "grad_norm": 0.15017539262771606, "learning_rate": 4.441165556594599e-05, "loss": 0.4495, "num_input_tokens_seen": 7371064, "step": 11270 }, { "epoch": 5.909329140461216, "grad_norm": 0.16830645501613617, "learning_rate": 4.440444811974363e-05, "loss": 0.5007, "num_input_tokens_seen": 7374168, "step": 11275 }, { "epoch": 5.911949685534591, "grad_norm": 0.1211148127913475, "learning_rate": 4.43972366143521e-05, "loss": 0.5665, "num_input_tokens_seen": 7377336, "step": 11280 }, { "epoch": 5.914570230607967, "grad_norm": 0.4020238220691681, "learning_rate": 4.4390021051279975e-05, "loss": 0.5575, "num_input_tokens_seen": 7380280, "step": 11285 }, { "epoch": 5.917190775681342, "grad_norm": 0.16897796094417572, "learning_rate": 4.438280143203665e-05, "loss": 0.4673, "num_input_tokens_seen": 7383864, "step": 11290 }, { "epoch": 5.919811320754717, "grad_norm": 0.13902094960212708, "learning_rate": 4.43755777581324e-05, "loss": 0.3562, "num_input_tokens_seen": 7387256, "step": 11295 }, { "epoch": 5.922431865828092, "grad_norm": 0.11478561162948608, "learning_rate": 4.4368350031078334e-05, "loss": 0.3581, "num_input_tokens_seen": 7391032, "step": 11300 }, { "epoch": 5.9250524109014675, "grad_norm": 0.10810239613056183, "learning_rate": 4.43611182523864e-05, "loss": 0.4569, "num_input_tokens_seen": 7393848, "step": 11305 }, { "epoch": 5.927672955974843, "grad_norm": 0.29757973551750183, "learning_rate": 4.4353882423569416e-05, "loss": 0.5191, "num_input_tokens_seen": 7397688, "step": 11310 }, { "epoch": 5.930293501048218, "grad_norm": 0.1950012892484665, "learning_rate": 4.434664254614103e-05, "loss": 0.512, "num_input_tokens_seen": 7400664, "step": 11315 }, { "epoch": 5.932914046121593, "grad_norm": 0.12082367390394211, "learning_rate": 4.4339398621615734e-05, "loss": 0.5398, "num_input_tokens_seen": 7403832, "step": 11320 }, { "epoch": 5.935534591194968, "grad_norm": 0.18026284873485565, "learning_rate": 4.433215065150889e-05, "loss": 0.476, "num_input_tokens_seen": 7406744, "step": 11325 }, { "epoch": 5.938155136268344, "grad_norm": 0.34048840403556824, "learning_rate": 4.4324898637336674e-05, "loss": 0.624, "num_input_tokens_seen": 7410232, "step": 11330 }, { "epoch": 5.940775681341719, "grad_norm": 0.20928767323493958, "learning_rate": 4.431764258061615e-05, "loss": 0.5517, "num_input_tokens_seen": 7413688, "step": 11335 }, { "epoch": 5.943396226415095, "grad_norm": 0.1360371708869934, "learning_rate": 4.431038248286517e-05, "loss": 0.4789, "num_input_tokens_seen": 7417112, "step": 11340 }, { "epoch": 5.94601677148847, "grad_norm": 0.13519161939620972, "learning_rate": 4.430311834560248e-05, "loss": 0.6571, "num_input_tokens_seen": 7419576, "step": 11345 }, { "epoch": 5.948637316561845, "grad_norm": 0.17993970215320587, "learning_rate": 4.4295850170347664e-05, "loss": 0.4634, "num_input_tokens_seen": 7423256, "step": 11350 }, { "epoch": 5.95125786163522, "grad_norm": 0.2713152766227722, "learning_rate": 4.428857795862113e-05, "loss": 0.3851, "num_input_tokens_seen": 7428792, "step": 11355 }, { "epoch": 5.953878406708595, "grad_norm": 0.15568839013576508, "learning_rate": 4.428130171194414e-05, "loss": 0.4622, "num_input_tokens_seen": 7432376, "step": 11360 }, { "epoch": 5.95649895178197, "grad_norm": 0.1173866018652916, "learning_rate": 4.427402143183879e-05, "loss": 0.6075, "num_input_tokens_seen": 7435992, "step": 11365 }, { "epoch": 5.959119496855346, "grad_norm": 0.24553918838500977, "learning_rate": 4.426673711982806e-05, "loss": 0.4687, "num_input_tokens_seen": 7438872, "step": 11370 }, { "epoch": 5.961740041928721, "grad_norm": 0.23716294765472412, "learning_rate": 4.425944877743573e-05, "loss": 0.4955, "num_input_tokens_seen": 7442008, "step": 11375 }, { "epoch": 5.964360587002097, "grad_norm": 0.425423264503479, "learning_rate": 4.425215640618644e-05, "loss": 0.5288, "num_input_tokens_seen": 7444408, "step": 11380 }, { "epoch": 5.966981132075472, "grad_norm": 0.18455636501312256, "learning_rate": 4.424486000760567e-05, "loss": 0.4054, "num_input_tokens_seen": 7448440, "step": 11385 }, { "epoch": 5.969601677148847, "grad_norm": 0.37885582447052, "learning_rate": 4.423755958321974e-05, "loss": 0.4876, "num_input_tokens_seen": 7451480, "step": 11390 }, { "epoch": 5.972222222222222, "grad_norm": 0.12834122776985168, "learning_rate": 4.423025513455582e-05, "loss": 0.4854, "num_input_tokens_seen": 7454424, "step": 11395 }, { "epoch": 5.9748427672955975, "grad_norm": 0.2643417716026306, "learning_rate": 4.422294666314192e-05, "loss": 0.5347, "num_input_tokens_seen": 7456824, "step": 11400 }, { "epoch": 5.977463312368973, "grad_norm": 0.12966272234916687, "learning_rate": 4.421563417050688e-05, "loss": 0.468, "num_input_tokens_seen": 7459608, "step": 11405 }, { "epoch": 5.980083857442348, "grad_norm": 0.20405662059783936, "learning_rate": 4.42083176581804e-05, "loss": 0.5488, "num_input_tokens_seen": 7462840, "step": 11410 }, { "epoch": 5.982704402515723, "grad_norm": 0.10766397416591644, "learning_rate": 4.4200997127693e-05, "loss": 0.4744, "num_input_tokens_seen": 7466456, "step": 11415 }, { "epoch": 5.985324947589098, "grad_norm": 0.13777439296245575, "learning_rate": 4.419367258057606e-05, "loss": 0.3546, "num_input_tokens_seen": 7469304, "step": 11420 }, { "epoch": 5.987945492662474, "grad_norm": 0.33718255162239075, "learning_rate": 4.418634401836179e-05, "loss": 0.5452, "num_input_tokens_seen": 7472248, "step": 11425 }, { "epoch": 5.990566037735849, "grad_norm": 0.2606658935546875, "learning_rate": 4.417901144258324e-05, "loss": 0.581, "num_input_tokens_seen": 7475032, "step": 11430 }, { "epoch": 5.993186582809225, "grad_norm": 0.12483527511358261, "learning_rate": 4.4171674854774305e-05, "loss": 0.4505, "num_input_tokens_seen": 7478008, "step": 11435 }, { "epoch": 5.9958071278826, "grad_norm": 0.15271735191345215, "learning_rate": 4.416433425646971e-05, "loss": 0.454, "num_input_tokens_seen": 7481112, "step": 11440 }, { "epoch": 5.998427672955975, "grad_norm": 0.12641170620918274, "learning_rate": 4.415698964920503e-05, "loss": 0.4578, "num_input_tokens_seen": 7484024, "step": 11445 }, { "epoch": 6.0, "eval_loss": 0.4763221740722656, "eval_runtime": 9.961, "eval_samples_per_second": 85.132, "eval_steps_per_second": 21.283, "num_input_tokens_seen": 7485248, "step": 11448 }, { "epoch": 6.00104821802935, "grad_norm": 0.1290351301431656, "learning_rate": 4.414964103451667e-05, "loss": 0.4283, "num_input_tokens_seen": 7486304, "step": 11450 }, { "epoch": 6.003668763102725, "grad_norm": 0.13221462070941925, "learning_rate": 4.414228841394187e-05, "loss": 0.57, "num_input_tokens_seen": 7489504, "step": 11455 }, { "epoch": 6.0062893081761, "grad_norm": 0.18622368574142456, "learning_rate": 4.413493178901872e-05, "loss": 0.3908, "num_input_tokens_seen": 7492896, "step": 11460 }, { "epoch": 6.008909853249476, "grad_norm": 0.14689159393310547, "learning_rate": 4.412757116128614e-05, "loss": 0.5235, "num_input_tokens_seen": 7495584, "step": 11465 }, { "epoch": 6.011530398322851, "grad_norm": 0.17514526844024658, "learning_rate": 4.41202065322839e-05, "loss": 0.5709, "num_input_tokens_seen": 7498208, "step": 11470 }, { "epoch": 6.014150943396227, "grad_norm": 0.09896177798509598, "learning_rate": 4.411283790355259e-05, "loss": 0.4314, "num_input_tokens_seen": 7501664, "step": 11475 }, { "epoch": 6.016771488469602, "grad_norm": 0.13907405734062195, "learning_rate": 4.410546527663364e-05, "loss": 0.4589, "num_input_tokens_seen": 7504416, "step": 11480 }, { "epoch": 6.019392033542977, "grad_norm": 0.18141017854213715, "learning_rate": 4.409808865306931e-05, "loss": 0.5195, "num_input_tokens_seen": 7507264, "step": 11485 }, { "epoch": 6.022012578616352, "grad_norm": 0.21647945046424866, "learning_rate": 4.409070803440273e-05, "loss": 0.6025, "num_input_tokens_seen": 7510432, "step": 11490 }, { "epoch": 6.0246331236897275, "grad_norm": 0.17540955543518066, "learning_rate": 4.4083323422177824e-05, "loss": 0.5805, "num_input_tokens_seen": 7513856, "step": 11495 }, { "epoch": 6.027253668763103, "grad_norm": 0.2042359560728073, "learning_rate": 4.407593481793937e-05, "loss": 0.6564, "num_input_tokens_seen": 7516736, "step": 11500 }, { "epoch": 6.029874213836478, "grad_norm": 0.20423613488674164, "learning_rate": 4.406854222323298e-05, "loss": 0.4827, "num_input_tokens_seen": 7519968, "step": 11505 }, { "epoch": 6.032494758909853, "grad_norm": 0.07881099730730057, "learning_rate": 4.406114563960511e-05, "loss": 0.5327, "num_input_tokens_seen": 7523040, "step": 11510 }, { "epoch": 6.035115303983228, "grad_norm": 0.1358974725008011, "learning_rate": 4.4053745068603026e-05, "loss": 0.5083, "num_input_tokens_seen": 7526464, "step": 11515 }, { "epoch": 6.037735849056604, "grad_norm": 0.13038763403892517, "learning_rate": 4.404634051177485e-05, "loss": 0.374, "num_input_tokens_seen": 7529376, "step": 11520 }, { "epoch": 6.040356394129979, "grad_norm": 0.11849001795053482, "learning_rate": 4.403893197066953e-05, "loss": 0.4316, "num_input_tokens_seen": 7535904, "step": 11525 }, { "epoch": 6.0429769392033545, "grad_norm": 0.1431630402803421, "learning_rate": 4.4031519446836855e-05, "loss": 0.3627, "num_input_tokens_seen": 7538784, "step": 11530 }, { "epoch": 6.04559748427673, "grad_norm": 0.23423703014850616, "learning_rate": 4.402410294182743e-05, "loss": 0.5703, "num_input_tokens_seen": 7542368, "step": 11535 }, { "epoch": 6.048218029350105, "grad_norm": 0.12714408338069916, "learning_rate": 4.4016682457192714e-05, "loss": 0.3603, "num_input_tokens_seen": 7545248, "step": 11540 }, { "epoch": 6.05083857442348, "grad_norm": 0.1486148089170456, "learning_rate": 4.400925799448497e-05, "loss": 0.4515, "num_input_tokens_seen": 7548160, "step": 11545 }, { "epoch": 6.053459119496855, "grad_norm": 0.11837663501501083, "learning_rate": 4.400182955525733e-05, "loss": 0.4377, "num_input_tokens_seen": 7551008, "step": 11550 }, { "epoch": 6.05607966457023, "grad_norm": 0.2007007896900177, "learning_rate": 4.3994397141063734e-05, "loss": 0.5978, "num_input_tokens_seen": 7554016, "step": 11555 }, { "epoch": 6.058700209643606, "grad_norm": 0.09920802712440491, "learning_rate": 4.398696075345896e-05, "loss": 0.5143, "num_input_tokens_seen": 7557376, "step": 11560 }, { "epoch": 6.061320754716981, "grad_norm": 0.17175619304180145, "learning_rate": 4.39795203939986e-05, "loss": 0.4851, "num_input_tokens_seen": 7561184, "step": 11565 }, { "epoch": 6.063941299790357, "grad_norm": 0.16847862303256989, "learning_rate": 4.397207606423911e-05, "loss": 0.4419, "num_input_tokens_seen": 7564832, "step": 11570 }, { "epoch": 6.066561844863732, "grad_norm": 0.1347966343164444, "learning_rate": 4.396462776573775e-05, "loss": 0.4794, "num_input_tokens_seen": 7569152, "step": 11575 }, { "epoch": 6.069182389937107, "grad_norm": 0.18671198189258575, "learning_rate": 4.395717550005262e-05, "loss": 0.3681, "num_input_tokens_seen": 7572352, "step": 11580 }, { "epoch": 6.071802935010482, "grad_norm": 0.23179523646831512, "learning_rate": 4.394971926874265e-05, "loss": 0.4393, "num_input_tokens_seen": 7574944, "step": 11585 }, { "epoch": 6.0744234800838575, "grad_norm": 0.1615445911884308, "learning_rate": 4.3942259073367604e-05, "loss": 0.4663, "num_input_tokens_seen": 7577632, "step": 11590 }, { "epoch": 6.077044025157233, "grad_norm": 0.1557636857032776, "learning_rate": 4.3934794915488065e-05, "loss": 0.4532, "num_input_tokens_seen": 7580512, "step": 11595 }, { "epoch": 6.079664570230608, "grad_norm": 0.16593918204307556, "learning_rate": 4.392732679666544e-05, "loss": 0.5139, "num_input_tokens_seen": 7583584, "step": 11600 }, { "epoch": 6.082285115303983, "grad_norm": 0.17720434069633484, "learning_rate": 4.3919854718461986e-05, "loss": 0.5664, "num_input_tokens_seen": 7586688, "step": 11605 }, { "epoch": 6.084905660377358, "grad_norm": 0.2618885338306427, "learning_rate": 4.391237868244078e-05, "loss": 0.4767, "num_input_tokens_seen": 7589792, "step": 11610 }, { "epoch": 6.087526205450734, "grad_norm": 0.13884806632995605, "learning_rate": 4.390489869016571e-05, "loss": 0.4844, "num_input_tokens_seen": 7593440, "step": 11615 }, { "epoch": 6.090146750524109, "grad_norm": 0.1500093638896942, "learning_rate": 4.389741474320152e-05, "loss": 0.6357, "num_input_tokens_seen": 7596544, "step": 11620 }, { "epoch": 6.0927672955974845, "grad_norm": 0.23375901579856873, "learning_rate": 4.388992684311376e-05, "loss": 0.4423, "num_input_tokens_seen": 7602816, "step": 11625 }, { "epoch": 6.09538784067086, "grad_norm": 0.2550595700740814, "learning_rate": 4.388243499146881e-05, "loss": 0.5533, "num_input_tokens_seen": 7605248, "step": 11630 }, { "epoch": 6.098008385744235, "grad_norm": 0.12116231769323349, "learning_rate": 4.3874939189833876e-05, "loss": 0.592, "num_input_tokens_seen": 7608640, "step": 11635 }, { "epoch": 6.10062893081761, "grad_norm": 0.1318509429693222, "learning_rate": 4.3867439439777005e-05, "loss": 0.5033, "num_input_tokens_seen": 7612224, "step": 11640 }, { "epoch": 6.103249475890985, "grad_norm": 0.2512860894203186, "learning_rate": 4.385993574286705e-05, "loss": 0.5446, "num_input_tokens_seen": 7614688, "step": 11645 }, { "epoch": 6.10587002096436, "grad_norm": 0.13669823110103607, "learning_rate": 4.3852428100673705e-05, "loss": 0.3895, "num_input_tokens_seen": 7618624, "step": 11650 }, { "epoch": 6.1084905660377355, "grad_norm": 0.19781634211540222, "learning_rate": 4.3844916514767484e-05, "loss": 0.4822, "num_input_tokens_seen": 7621568, "step": 11655 }, { "epoch": 6.111111111111111, "grad_norm": 0.47287824749946594, "learning_rate": 4.383740098671972e-05, "loss": 0.6562, "num_input_tokens_seen": 7623712, "step": 11660 }, { "epoch": 6.113731656184487, "grad_norm": 0.14640162885189056, "learning_rate": 4.3829881518102576e-05, "loss": 0.5627, "num_input_tokens_seen": 7626944, "step": 11665 }, { "epoch": 6.116352201257862, "grad_norm": 0.1820712834596634, "learning_rate": 4.3822358110489045e-05, "loss": 0.5557, "num_input_tokens_seen": 7629504, "step": 11670 }, { "epoch": 6.118972746331237, "grad_norm": 0.27662327885627747, "learning_rate": 4.381483076545294e-05, "loss": 0.3946, "num_input_tokens_seen": 7632320, "step": 11675 }, { "epoch": 6.121593291404612, "grad_norm": 0.2096313089132309, "learning_rate": 4.380729948456888e-05, "loss": 0.4414, "num_input_tokens_seen": 7634944, "step": 11680 }, { "epoch": 6.1242138364779874, "grad_norm": 0.14411447942256927, "learning_rate": 4.3799764269412336e-05, "loss": 0.431, "num_input_tokens_seen": 7638752, "step": 11685 }, { "epoch": 6.126834381551363, "grad_norm": 0.1921835094690323, "learning_rate": 4.379222512155958e-05, "loss": 0.5106, "num_input_tokens_seen": 7641920, "step": 11690 }, { "epoch": 6.129454926624738, "grad_norm": 0.1757781058549881, "learning_rate": 4.378468204258773e-05, "loss": 0.553, "num_input_tokens_seen": 7645248, "step": 11695 }, { "epoch": 6.132075471698113, "grad_norm": 0.19130082428455353, "learning_rate": 4.37771350340747e-05, "loss": 0.6264, "num_input_tokens_seen": 7652256, "step": 11700 }, { "epoch": 6.134696016771488, "grad_norm": 0.18420001864433289, "learning_rate": 4.376958409759925e-05, "loss": 0.4091, "num_input_tokens_seen": 7654976, "step": 11705 }, { "epoch": 6.137316561844864, "grad_norm": 0.08866284787654877, "learning_rate": 4.376202923474093e-05, "loss": 0.5127, "num_input_tokens_seen": 7658944, "step": 11710 }, { "epoch": 6.139937106918239, "grad_norm": 0.13966037333011627, "learning_rate": 4.3754470447080145e-05, "loss": 0.524, "num_input_tokens_seen": 7662112, "step": 11715 }, { "epoch": 6.1425576519916145, "grad_norm": 0.17045287787914276, "learning_rate": 4.374690773619811e-05, "loss": 0.5632, "num_input_tokens_seen": 7664480, "step": 11720 }, { "epoch": 6.14517819706499, "grad_norm": 0.09477995336055756, "learning_rate": 4.373934110367685e-05, "loss": 0.4769, "num_input_tokens_seen": 7667520, "step": 11725 }, { "epoch": 6.147798742138365, "grad_norm": 0.11087172478437424, "learning_rate": 4.373177055109922e-05, "loss": 0.52, "num_input_tokens_seen": 7670880, "step": 11730 }, { "epoch": 6.15041928721174, "grad_norm": 0.23499971628189087, "learning_rate": 4.3724196080048896e-05, "loss": 0.4931, "num_input_tokens_seen": 7673856, "step": 11735 }, { "epoch": 6.153039832285115, "grad_norm": 0.17648470401763916, "learning_rate": 4.3716617692110364e-05, "loss": 0.5508, "num_input_tokens_seen": 7676864, "step": 11740 }, { "epoch": 6.15566037735849, "grad_norm": 0.1428879052400589, "learning_rate": 4.3709035388868943e-05, "loss": 0.507, "num_input_tokens_seen": 7679296, "step": 11745 }, { "epoch": 6.1582809224318655, "grad_norm": 0.28309890627861023, "learning_rate": 4.3701449171910767e-05, "loss": 0.5156, "num_input_tokens_seen": 7682912, "step": 11750 }, { "epoch": 6.160901467505241, "grad_norm": 0.10813245922327042, "learning_rate": 4.3693859042822774e-05, "loss": 0.4489, "num_input_tokens_seen": 7686592, "step": 11755 }, { "epoch": 6.163522012578617, "grad_norm": 0.17337773740291595, "learning_rate": 4.368626500319275e-05, "loss": 0.502, "num_input_tokens_seen": 7689440, "step": 11760 }, { "epoch": 6.166142557651992, "grad_norm": 0.09630488604307175, "learning_rate": 4.367866705460927e-05, "loss": 0.5799, "num_input_tokens_seen": 7693184, "step": 11765 }, { "epoch": 6.168763102725367, "grad_norm": 0.16989681124687195, "learning_rate": 4.367106519866173e-05, "loss": 0.5021, "num_input_tokens_seen": 7696672, "step": 11770 }, { "epoch": 6.171383647798742, "grad_norm": 0.16426460444927216, "learning_rate": 4.366345943694037e-05, "loss": 0.4644, "num_input_tokens_seen": 7700352, "step": 11775 }, { "epoch": 6.174004192872117, "grad_norm": 0.20710337162017822, "learning_rate": 4.365584977103622e-05, "loss": 0.6134, "num_input_tokens_seen": 7703008, "step": 11780 }, { "epoch": 6.176624737945493, "grad_norm": 0.18721358478069305, "learning_rate": 4.364823620254113e-05, "loss": 0.5108, "num_input_tokens_seen": 7705824, "step": 11785 }, { "epoch": 6.179245283018868, "grad_norm": 0.16254578530788422, "learning_rate": 4.364061873304778e-05, "loss": 0.42, "num_input_tokens_seen": 7708672, "step": 11790 }, { "epoch": 6.181865828092243, "grad_norm": 0.12512357532978058, "learning_rate": 4.363299736414965e-05, "loss": 0.4359, "num_input_tokens_seen": 7712480, "step": 11795 }, { "epoch": 6.184486373165618, "grad_norm": 0.18259486556053162, "learning_rate": 4.3625372097441045e-05, "loss": 0.3567, "num_input_tokens_seen": 7715296, "step": 11800 }, { "epoch": 6.187106918238993, "grad_norm": 0.13502924144268036, "learning_rate": 4.3617742934517094e-05, "loss": 0.5174, "num_input_tokens_seen": 7719808, "step": 11805 }, { "epoch": 6.189727463312369, "grad_norm": 0.15323741734027863, "learning_rate": 4.361010987697373e-05, "loss": 0.5439, "num_input_tokens_seen": 7722208, "step": 11810 }, { "epoch": 6.1923480083857445, "grad_norm": 0.40502431988716125, "learning_rate": 4.360247292640768e-05, "loss": 0.3448, "num_input_tokens_seen": 7726816, "step": 11815 }, { "epoch": 6.19496855345912, "grad_norm": 0.18971723318099976, "learning_rate": 4.359483208441653e-05, "loss": 0.5517, "num_input_tokens_seen": 7729504, "step": 11820 }, { "epoch": 6.197589098532495, "grad_norm": 0.2104429453611374, "learning_rate": 4.3587187352598646e-05, "loss": 0.5055, "num_input_tokens_seen": 7732096, "step": 11825 }, { "epoch": 6.20020964360587, "grad_norm": 0.16133466362953186, "learning_rate": 4.357953873255323e-05, "loss": 0.3976, "num_input_tokens_seen": 7735008, "step": 11830 }, { "epoch": 6.202830188679245, "grad_norm": 0.16824081540107727, "learning_rate": 4.357188622588027e-05, "loss": 0.4877, "num_input_tokens_seen": 7738432, "step": 11835 }, { "epoch": 6.20545073375262, "grad_norm": 0.1825469434261322, "learning_rate": 4.356422983418059e-05, "loss": 0.5146, "num_input_tokens_seen": 7741280, "step": 11840 }, { "epoch": 6.2080712788259955, "grad_norm": 0.13741223514080048, "learning_rate": 4.3556569559055824e-05, "loss": 0.5408, "num_input_tokens_seen": 7744512, "step": 11845 }, { "epoch": 6.210691823899371, "grad_norm": 0.13513752818107605, "learning_rate": 4.3548905402108406e-05, "loss": 0.5177, "num_input_tokens_seen": 7747904, "step": 11850 }, { "epoch": 6.213312368972747, "grad_norm": 0.1822509765625, "learning_rate": 4.35412373649416e-05, "loss": 0.5144, "num_input_tokens_seen": 7751168, "step": 11855 }, { "epoch": 6.215932914046122, "grad_norm": 0.3332875669002533, "learning_rate": 4.353356544915947e-05, "loss": 0.4319, "num_input_tokens_seen": 7754368, "step": 11860 }, { "epoch": 6.218553459119497, "grad_norm": 0.21028706431388855, "learning_rate": 4.352588965636689e-05, "loss": 0.5795, "num_input_tokens_seen": 7757056, "step": 11865 }, { "epoch": 6.221174004192872, "grad_norm": 0.09010639041662216, "learning_rate": 4.351820998816954e-05, "loss": 0.3506, "num_input_tokens_seen": 7759904, "step": 11870 }, { "epoch": 6.223794549266247, "grad_norm": 0.16188254952430725, "learning_rate": 4.3510526446173935e-05, "loss": 0.6304, "num_input_tokens_seen": 7764768, "step": 11875 }, { "epoch": 6.226415094339623, "grad_norm": 0.14442692697048187, "learning_rate": 4.350283903198737e-05, "loss": 0.5315, "num_input_tokens_seen": 7768416, "step": 11880 }, { "epoch": 6.229035639412998, "grad_norm": 0.1574459969997406, "learning_rate": 4.349514774721797e-05, "loss": 0.4323, "num_input_tokens_seen": 7770912, "step": 11885 }, { "epoch": 6.231656184486373, "grad_norm": 0.1409386396408081, "learning_rate": 4.348745259347467e-05, "loss": 0.4494, "num_input_tokens_seen": 7774240, "step": 11890 }, { "epoch": 6.234276729559748, "grad_norm": 0.13870395720005035, "learning_rate": 4.347975357236721e-05, "loss": 0.6347, "num_input_tokens_seen": 7777728, "step": 11895 }, { "epoch": 6.236897274633123, "grad_norm": 0.23391225934028625, "learning_rate": 4.347205068550612e-05, "loss": 0.4216, "num_input_tokens_seen": 7780512, "step": 11900 }, { "epoch": 6.239517819706499, "grad_norm": 0.21734118461608887, "learning_rate": 4.3464343934502774e-05, "loss": 0.8201, "num_input_tokens_seen": 7783104, "step": 11905 }, { "epoch": 6.2421383647798745, "grad_norm": 0.14388148486614227, "learning_rate": 4.345663332096932e-05, "loss": 0.4558, "num_input_tokens_seen": 7785920, "step": 11910 }, { "epoch": 6.24475890985325, "grad_norm": 0.2571527361869812, "learning_rate": 4.344891884651875e-05, "loss": 0.5467, "num_input_tokens_seen": 7789024, "step": 11915 }, { "epoch": 6.247379454926625, "grad_norm": 0.1370486170053482, "learning_rate": 4.344120051276483e-05, "loss": 0.5057, "num_input_tokens_seen": 7792640, "step": 11920 }, { "epoch": 6.25, "grad_norm": 0.09126509726047516, "learning_rate": 4.3433478321322155e-05, "loss": 0.4677, "num_input_tokens_seen": 7796576, "step": 11925 }, { "epoch": 6.252620545073375, "grad_norm": 0.13402456045150757, "learning_rate": 4.3425752273806106e-05, "loss": 0.4785, "num_input_tokens_seen": 7799264, "step": 11930 }, { "epoch": 6.25524109014675, "grad_norm": 0.16279010474681854, "learning_rate": 4.34180223718329e-05, "loss": 0.4606, "num_input_tokens_seen": 7802016, "step": 11935 }, { "epoch": 6.2578616352201255, "grad_norm": 0.208334282040596, "learning_rate": 4.341028861701953e-05, "loss": 0.5648, "num_input_tokens_seen": 7804320, "step": 11940 }, { "epoch": 6.260482180293501, "grad_norm": 0.10687897354364395, "learning_rate": 4.3402551010983825e-05, "loss": 0.578, "num_input_tokens_seen": 7807648, "step": 11945 }, { "epoch": 6.263102725366876, "grad_norm": 0.2201777994632721, "learning_rate": 4.33948095553444e-05, "loss": 0.3672, "num_input_tokens_seen": 7815680, "step": 11950 }, { "epoch": 6.265723270440252, "grad_norm": 0.1944229155778885, "learning_rate": 4.338706425172067e-05, "loss": 0.4496, "num_input_tokens_seen": 7818528, "step": 11955 }, { "epoch": 6.268343815513627, "grad_norm": 0.15265057981014252, "learning_rate": 4.3379315101732866e-05, "loss": 0.3807, "num_input_tokens_seen": 7821728, "step": 11960 }, { "epoch": 6.270964360587002, "grad_norm": 0.21535171568393707, "learning_rate": 4.3371562107002026e-05, "loss": 0.5246, "num_input_tokens_seen": 7825440, "step": 11965 }, { "epoch": 6.273584905660377, "grad_norm": 0.16502150893211365, "learning_rate": 4.3363805269149984e-05, "loss": 0.4429, "num_input_tokens_seen": 7828192, "step": 11970 }, { "epoch": 6.276205450733753, "grad_norm": 0.4873487055301666, "learning_rate": 4.335604458979939e-05, "loss": 0.4577, "num_input_tokens_seen": 7831040, "step": 11975 }, { "epoch": 6.278825995807128, "grad_norm": 0.20301435887813568, "learning_rate": 4.3348280070573686e-05, "loss": 0.6308, "num_input_tokens_seen": 7833888, "step": 11980 }, { "epoch": 6.281446540880503, "grad_norm": 0.2467399388551712, "learning_rate": 4.334051171309712e-05, "loss": 0.4624, "num_input_tokens_seen": 7836928, "step": 11985 }, { "epoch": 6.284067085953878, "grad_norm": 0.1373889446258545, "learning_rate": 4.333273951899474e-05, "loss": 0.4679, "num_input_tokens_seen": 7839808, "step": 11990 }, { "epoch": 6.286687631027253, "grad_norm": 0.21847262978553772, "learning_rate": 4.33249634898924e-05, "loss": 0.6073, "num_input_tokens_seen": 7842656, "step": 11995 }, { "epoch": 6.289308176100629, "grad_norm": 0.19308964908123016, "learning_rate": 4.331718362741677e-05, "loss": 0.5049, "num_input_tokens_seen": 7845536, "step": 12000 }, { "epoch": 6.2919287211740045, "grad_norm": 0.14887629449367523, "learning_rate": 4.330939993319529e-05, "loss": 0.5333, "num_input_tokens_seen": 7849632, "step": 12005 }, { "epoch": 6.29454926624738, "grad_norm": 0.09158012270927429, "learning_rate": 4.330161240885623e-05, "loss": 0.4203, "num_input_tokens_seen": 7852480, "step": 12010 }, { "epoch": 6.297169811320755, "grad_norm": 0.18845470249652863, "learning_rate": 4.3293821056028654e-05, "loss": 0.4491, "num_input_tokens_seen": 7856192, "step": 12015 }, { "epoch": 6.29979035639413, "grad_norm": 0.14539919793605804, "learning_rate": 4.3286025876342415e-05, "loss": 0.4928, "num_input_tokens_seen": 7859744, "step": 12020 }, { "epoch": 6.302410901467505, "grad_norm": 0.16067785024642944, "learning_rate": 4.327822687142819e-05, "loss": 0.5853, "num_input_tokens_seen": 7862848, "step": 12025 }, { "epoch": 6.30503144654088, "grad_norm": 0.2134600430727005, "learning_rate": 4.327042404291742e-05, "loss": 0.5907, "num_input_tokens_seen": 7865408, "step": 12030 }, { "epoch": 6.3076519916142555, "grad_norm": 0.12623339891433716, "learning_rate": 4.326261739244239e-05, "loss": 0.3829, "num_input_tokens_seen": 7868064, "step": 12035 }, { "epoch": 6.310272536687631, "grad_norm": 0.2116221934556961, "learning_rate": 4.325480692163615e-05, "loss": 0.5505, "num_input_tokens_seen": 7870400, "step": 12040 }, { "epoch": 6.312893081761006, "grad_norm": 0.13279268145561218, "learning_rate": 4.324699263213256e-05, "loss": 0.4037, "num_input_tokens_seen": 7875008, "step": 12045 }, { "epoch": 6.315513626834382, "grad_norm": 0.13764511048793793, "learning_rate": 4.3239174525566285e-05, "loss": 0.5251, "num_input_tokens_seen": 7878560, "step": 12050 }, { "epoch": 6.318134171907757, "grad_norm": 0.13351625204086304, "learning_rate": 4.3231352603572784e-05, "loss": 0.4332, "num_input_tokens_seen": 7881984, "step": 12055 }, { "epoch": 6.320754716981132, "grad_norm": 0.1224580779671669, "learning_rate": 4.322352686778832e-05, "loss": 0.4859, "num_input_tokens_seen": 7885056, "step": 12060 }, { "epoch": 6.323375262054507, "grad_norm": 0.10255450755357742, "learning_rate": 4.3215697319849926e-05, "loss": 0.5623, "num_input_tokens_seen": 7888928, "step": 12065 }, { "epoch": 6.325995807127883, "grad_norm": 0.1686377078294754, "learning_rate": 4.3207863961395476e-05, "loss": 0.5322, "num_input_tokens_seen": 7891648, "step": 12070 }, { "epoch": 6.328616352201258, "grad_norm": 0.192306786775589, "learning_rate": 4.320002679406361e-05, "loss": 0.4347, "num_input_tokens_seen": 7894432, "step": 12075 }, { "epoch": 6.331236897274633, "grad_norm": 0.1860085427761078, "learning_rate": 4.319218581949378e-05, "loss": 0.4627, "num_input_tokens_seen": 7896608, "step": 12080 }, { "epoch": 6.333857442348008, "grad_norm": 0.21173079311847687, "learning_rate": 4.318434103932622e-05, "loss": 0.5629, "num_input_tokens_seen": 7900000, "step": 12085 }, { "epoch": 6.336477987421383, "grad_norm": 0.3142346739768982, "learning_rate": 4.317649245520197e-05, "loss": 0.4527, "num_input_tokens_seen": 7902848, "step": 12090 }, { "epoch": 6.339098532494759, "grad_norm": 0.13291886448860168, "learning_rate": 4.3168640068762875e-05, "loss": 0.3714, "num_input_tokens_seen": 7905760, "step": 12095 }, { "epoch": 6.3417190775681345, "grad_norm": 0.16358762979507446, "learning_rate": 4.316078388165155e-05, "loss": 0.5555, "num_input_tokens_seen": 7909408, "step": 12100 }, { "epoch": 6.34433962264151, "grad_norm": 0.16026678681373596, "learning_rate": 4.315292389551143e-05, "loss": 0.4371, "num_input_tokens_seen": 7912768, "step": 12105 }, { "epoch": 6.346960167714885, "grad_norm": 0.2580249011516571, "learning_rate": 4.3145060111986726e-05, "loss": 0.5788, "num_input_tokens_seen": 7915488, "step": 12110 }, { "epoch": 6.34958071278826, "grad_norm": 0.1439979523420334, "learning_rate": 4.313719253272246e-05, "loss": 0.5043, "num_input_tokens_seen": 7918592, "step": 12115 }, { "epoch": 6.352201257861635, "grad_norm": 0.18259041011333466, "learning_rate": 4.3129321159364436e-05, "loss": 0.4954, "num_input_tokens_seen": 7921952, "step": 12120 }, { "epoch": 6.35482180293501, "grad_norm": 0.12381583452224731, "learning_rate": 4.312144599355925e-05, "loss": 0.4899, "num_input_tokens_seen": 7925632, "step": 12125 }, { "epoch": 6.3574423480083855, "grad_norm": 0.14010339975357056, "learning_rate": 4.31135670369543e-05, "loss": 0.4859, "num_input_tokens_seen": 7928896, "step": 12130 }, { "epoch": 6.360062893081761, "grad_norm": 0.1070903092622757, "learning_rate": 4.310568429119779e-05, "loss": 0.5125, "num_input_tokens_seen": 7932064, "step": 12135 }, { "epoch": 6.362683438155136, "grad_norm": 0.1830310970544815, "learning_rate": 4.309779775793868e-05, "loss": 0.5628, "num_input_tokens_seen": 7934688, "step": 12140 }, { "epoch": 6.365303983228512, "grad_norm": 0.22427250444889069, "learning_rate": 4.308990743882675e-05, "loss": 0.5808, "num_input_tokens_seen": 7937536, "step": 12145 }, { "epoch": 6.367924528301887, "grad_norm": 0.1954992711544037, "learning_rate": 4.308201333551256e-05, "loss": 0.3961, "num_input_tokens_seen": 7940832, "step": 12150 }, { "epoch": 6.370545073375262, "grad_norm": 0.19854456186294556, "learning_rate": 4.3074115449647476e-05, "loss": 0.4063, "num_input_tokens_seen": 7943392, "step": 12155 }, { "epoch": 6.373165618448637, "grad_norm": 0.22114019095897675, "learning_rate": 4.306621378288364e-05, "loss": 0.4486, "num_input_tokens_seen": 7946688, "step": 12160 }, { "epoch": 6.3757861635220126, "grad_norm": 0.20539125800132751, "learning_rate": 4.305830833687399e-05, "loss": 0.6179, "num_input_tokens_seen": 7948960, "step": 12165 }, { "epoch": 6.378406708595388, "grad_norm": 0.24442777037620544, "learning_rate": 4.305039911327225e-05, "loss": 0.4166, "num_input_tokens_seen": 7952832, "step": 12170 }, { "epoch": 6.381027253668763, "grad_norm": 0.16849276423454285, "learning_rate": 4.304248611373296e-05, "loss": 0.478, "num_input_tokens_seen": 7955616, "step": 12175 }, { "epoch": 6.383647798742138, "grad_norm": 0.11493067443370819, "learning_rate": 4.30345693399114e-05, "loss": 0.5225, "num_input_tokens_seen": 7959456, "step": 12180 }, { "epoch": 6.386268343815513, "grad_norm": 0.24448050558567047, "learning_rate": 4.30266487934637e-05, "loss": 0.7412, "num_input_tokens_seen": 7964128, "step": 12185 }, { "epoch": 6.388888888888889, "grad_norm": 0.20427003502845764, "learning_rate": 4.3018724476046714e-05, "loss": 0.4702, "num_input_tokens_seen": 7966816, "step": 12190 }, { "epoch": 6.3915094339622645, "grad_norm": 0.15204615890979767, "learning_rate": 4.301079638931814e-05, "loss": 0.687, "num_input_tokens_seen": 7969600, "step": 12195 }, { "epoch": 6.39412997903564, "grad_norm": 0.1443682610988617, "learning_rate": 4.3002864534936444e-05, "loss": 0.4883, "num_input_tokens_seen": 7972672, "step": 12200 }, { "epoch": 6.396750524109015, "grad_norm": 0.22211810946464539, "learning_rate": 4.2994928914560874e-05, "loss": 0.533, "num_input_tokens_seen": 7975744, "step": 12205 }, { "epoch": 6.39937106918239, "grad_norm": 0.19269739091396332, "learning_rate": 4.298698952985147e-05, "loss": 0.585, "num_input_tokens_seen": 7978816, "step": 12210 }, { "epoch": 6.401991614255765, "grad_norm": 0.26624777913093567, "learning_rate": 4.297904638246906e-05, "loss": 0.5669, "num_input_tokens_seen": 7981504, "step": 12215 }, { "epoch": 6.40461215932914, "grad_norm": 0.1555182784795761, "learning_rate": 4.297109947407527e-05, "loss": 0.5572, "num_input_tokens_seen": 7984000, "step": 12220 }, { "epoch": 6.4072327044025155, "grad_norm": 0.14509959518909454, "learning_rate": 4.2963148806332486e-05, "loss": 0.4643, "num_input_tokens_seen": 7986624, "step": 12225 }, { "epoch": 6.409853249475891, "grad_norm": 0.19081731140613556, "learning_rate": 4.2955194380903915e-05, "loss": 0.4268, "num_input_tokens_seen": 7989728, "step": 12230 }, { "epoch": 6.412473794549266, "grad_norm": 0.21568170189857483, "learning_rate": 4.2947236199453524e-05, "loss": 0.5835, "num_input_tokens_seen": 7993696, "step": 12235 }, { "epoch": 6.415094339622642, "grad_norm": 0.16910181939601898, "learning_rate": 4.2939274263646065e-05, "loss": 0.4491, "num_input_tokens_seen": 7996192, "step": 12240 }, { "epoch": 6.417714884696017, "grad_norm": 0.17725038528442383, "learning_rate": 4.29313085751471e-05, "loss": 0.5298, "num_input_tokens_seen": 8000192, "step": 12245 }, { "epoch": 6.420335429769392, "grad_norm": 0.1714532971382141, "learning_rate": 4.292333913562294e-05, "loss": 0.5966, "num_input_tokens_seen": 8004256, "step": 12250 }, { "epoch": 6.422955974842767, "grad_norm": 0.24433664977550507, "learning_rate": 4.291536594674073e-05, "loss": 0.6769, "num_input_tokens_seen": 8007200, "step": 12255 }, { "epoch": 6.4255765199161425, "grad_norm": 0.3195359706878662, "learning_rate": 4.2907389010168344e-05, "loss": 0.501, "num_input_tokens_seen": 8011328, "step": 12260 }, { "epoch": 6.428197064989518, "grad_norm": 0.25967249274253845, "learning_rate": 4.2899408327574475e-05, "loss": 0.5532, "num_input_tokens_seen": 8014240, "step": 12265 }, { "epoch": 6.430817610062893, "grad_norm": 0.14597417414188385, "learning_rate": 4.289142390062859e-05, "loss": 0.5515, "num_input_tokens_seen": 8017600, "step": 12270 }, { "epoch": 6.433438155136268, "grad_norm": 0.1715100258588791, "learning_rate": 4.2883435731000946e-05, "loss": 0.5284, "num_input_tokens_seen": 8020416, "step": 12275 }, { "epoch": 6.436058700209643, "grad_norm": 0.12077756226062775, "learning_rate": 4.287544382036256e-05, "loss": 0.5282, "num_input_tokens_seen": 8023328, "step": 12280 }, { "epoch": 6.438679245283019, "grad_norm": 0.13302046060562134, "learning_rate": 4.286744817038527e-05, "loss": 0.6327, "num_input_tokens_seen": 8026336, "step": 12285 }, { "epoch": 6.441299790356394, "grad_norm": 0.22302281856536865, "learning_rate": 4.2859448782741666e-05, "loss": 0.606, "num_input_tokens_seen": 8029728, "step": 12290 }, { "epoch": 6.44392033542977, "grad_norm": 0.17114831507205963, "learning_rate": 4.2851445659105126e-05, "loss": 0.5105, "num_input_tokens_seen": 8033088, "step": 12295 }, { "epoch": 6.446540880503145, "grad_norm": 0.14359945058822632, "learning_rate": 4.284343880114982e-05, "loss": 0.4611, "num_input_tokens_seen": 8036352, "step": 12300 }, { "epoch": 6.44916142557652, "grad_norm": 0.15612851083278656, "learning_rate": 4.2835428210550676e-05, "loss": 0.5099, "num_input_tokens_seen": 8039616, "step": 12305 }, { "epoch": 6.451781970649895, "grad_norm": 0.4911758303642273, "learning_rate": 4.282741388898344e-05, "loss": 0.5864, "num_input_tokens_seen": 8042240, "step": 12310 }, { "epoch": 6.45440251572327, "grad_norm": 0.15371038019657135, "learning_rate": 4.281939583812459e-05, "loss": 0.5487, "num_input_tokens_seen": 8046816, "step": 12315 }, { "epoch": 6.4570230607966455, "grad_norm": 0.09846498817205429, "learning_rate": 4.281137405965143e-05, "loss": 0.5116, "num_input_tokens_seen": 8050528, "step": 12320 }, { "epoch": 6.459643605870021, "grad_norm": 0.21378779411315918, "learning_rate": 4.280334855524203e-05, "loss": 0.6202, "num_input_tokens_seen": 8054336, "step": 12325 }, { "epoch": 6.462264150943396, "grad_norm": 0.20343729853630066, "learning_rate": 4.279531932657521e-05, "loss": 0.5839, "num_input_tokens_seen": 8056832, "step": 12330 }, { "epoch": 6.464884696016772, "grad_norm": 0.30580058693885803, "learning_rate": 4.278728637533062e-05, "loss": 0.5374, "num_input_tokens_seen": 8059616, "step": 12335 }, { "epoch": 6.467505241090147, "grad_norm": 0.16580164432525635, "learning_rate": 4.2779249703188626e-05, "loss": 0.522, "num_input_tokens_seen": 8062400, "step": 12340 }, { "epoch": 6.470125786163522, "grad_norm": 0.18671880662441254, "learning_rate": 4.277120931183045e-05, "loss": 0.5165, "num_input_tokens_seen": 8065824, "step": 12345 }, { "epoch": 6.472746331236897, "grad_norm": 0.28498339653015137, "learning_rate": 4.276316520293803e-05, "loss": 0.3892, "num_input_tokens_seen": 8068544, "step": 12350 }, { "epoch": 6.4753668763102725, "grad_norm": 0.21995116770267487, "learning_rate": 4.27551173781941e-05, "loss": 0.541, "num_input_tokens_seen": 8071712, "step": 12355 }, { "epoch": 6.477987421383648, "grad_norm": 0.18886452913284302, "learning_rate": 4.274706583928216e-05, "loss": 0.5363, "num_input_tokens_seen": 8075360, "step": 12360 }, { "epoch": 6.480607966457023, "grad_norm": 0.22041231393814087, "learning_rate": 4.273901058788654e-05, "loss": 0.4865, "num_input_tokens_seen": 8078720, "step": 12365 }, { "epoch": 6.483228511530398, "grad_norm": 0.13887007534503937, "learning_rate": 4.2730951625692275e-05, "loss": 0.4701, "num_input_tokens_seen": 8081504, "step": 12370 }, { "epoch": 6.485849056603773, "grad_norm": 0.18640181422233582, "learning_rate": 4.27228889543852e-05, "loss": 0.5037, "num_input_tokens_seen": 8084608, "step": 12375 }, { "epoch": 6.488469601677149, "grad_norm": 0.16133733093738556, "learning_rate": 4.2714822575651966e-05, "loss": 0.5727, "num_input_tokens_seen": 8087936, "step": 12380 }, { "epoch": 6.491090146750524, "grad_norm": 0.17551861703395844, "learning_rate": 4.270675249117994e-05, "loss": 0.3389, "num_input_tokens_seen": 8091648, "step": 12385 }, { "epoch": 6.4937106918239, "grad_norm": 0.10022825002670288, "learning_rate": 4.269867870265731e-05, "loss": 0.5408, "num_input_tokens_seen": 8094848, "step": 12390 }, { "epoch": 6.496331236897275, "grad_norm": 0.12044268846511841, "learning_rate": 4.269060121177301e-05, "loss": 0.5323, "num_input_tokens_seen": 8097952, "step": 12395 }, { "epoch": 6.49895178197065, "grad_norm": 0.20049147307872772, "learning_rate": 4.2682520020216764e-05, "loss": 0.525, "num_input_tokens_seen": 8102016, "step": 12400 }, { "epoch": 6.501572327044025, "grad_norm": 0.14905498921871185, "learning_rate": 4.267443512967906e-05, "loss": 0.7449, "num_input_tokens_seen": 8105568, "step": 12405 }, { "epoch": 6.5041928721174, "grad_norm": 0.16550099849700928, "learning_rate": 4.2666346541851165e-05, "loss": 0.5253, "num_input_tokens_seen": 8108864, "step": 12410 }, { "epoch": 6.506813417190775, "grad_norm": 0.1394747644662857, "learning_rate": 4.265825425842512e-05, "loss": 0.4343, "num_input_tokens_seen": 8112032, "step": 12415 }, { "epoch": 6.509433962264151, "grad_norm": 0.16624246537685394, "learning_rate": 4.265015828109375e-05, "loss": 0.7927, "num_input_tokens_seen": 8115136, "step": 12420 }, { "epoch": 6.512054507337526, "grad_norm": 0.12681756913661957, "learning_rate": 4.264205861155063e-05, "loss": 0.5532, "num_input_tokens_seen": 8120704, "step": 12425 }, { "epoch": 6.514675052410902, "grad_norm": 0.2048020213842392, "learning_rate": 4.263395525149011e-05, "loss": 0.4535, "num_input_tokens_seen": 8123904, "step": 12430 }, { "epoch": 6.517295597484277, "grad_norm": 0.11400195956230164, "learning_rate": 4.262584820260734e-05, "loss": 0.5891, "num_input_tokens_seen": 8127264, "step": 12435 }, { "epoch": 6.519916142557652, "grad_norm": 0.16718140244483948, "learning_rate": 4.261773746659821e-05, "loss": 0.5887, "num_input_tokens_seen": 8130048, "step": 12440 }, { "epoch": 6.522536687631027, "grad_norm": 0.16702859103679657, "learning_rate": 4.260962304515939e-05, "loss": 0.4334, "num_input_tokens_seen": 8133440, "step": 12445 }, { "epoch": 6.5251572327044025, "grad_norm": 0.26188233494758606, "learning_rate": 4.2601504939988335e-05, "loss": 0.581, "num_input_tokens_seen": 8137216, "step": 12450 }, { "epoch": 6.527777777777778, "grad_norm": 0.1381075531244278, "learning_rate": 4.259338315278325e-05, "loss": 0.6354, "num_input_tokens_seen": 8141408, "step": 12455 }, { "epoch": 6.530398322851153, "grad_norm": 0.17050430178642273, "learning_rate": 4.258525768524313e-05, "loss": 0.5677, "num_input_tokens_seen": 8144832, "step": 12460 }, { "epoch": 6.533018867924528, "grad_norm": 0.1819848269224167, "learning_rate": 4.2577128539067725e-05, "loss": 0.4707, "num_input_tokens_seen": 8147584, "step": 12465 }, { "epoch": 6.535639412997903, "grad_norm": 0.19319011270999908, "learning_rate": 4.256899571595756e-05, "loss": 0.5303, "num_input_tokens_seen": 8150528, "step": 12470 }, { "epoch": 6.538259958071279, "grad_norm": 0.11286619305610657, "learning_rate": 4.256085921761393e-05, "loss": 0.5274, "num_input_tokens_seen": 8154304, "step": 12475 }, { "epoch": 6.540880503144654, "grad_norm": 0.1804678738117218, "learning_rate": 4.25527190457389e-05, "loss": 0.6431, "num_input_tokens_seen": 8156832, "step": 12480 }, { "epoch": 6.54350104821803, "grad_norm": 0.2071605920791626, "learning_rate": 4.2544575202035285e-05, "loss": 0.4836, "num_input_tokens_seen": 8159456, "step": 12485 }, { "epoch": 6.546121593291405, "grad_norm": 0.28746497631073, "learning_rate": 4.253642768820671e-05, "loss": 0.5496, "num_input_tokens_seen": 8162720, "step": 12490 }, { "epoch": 6.54874213836478, "grad_norm": 0.11505286395549774, "learning_rate": 4.252827650595752e-05, "loss": 0.4609, "num_input_tokens_seen": 8165824, "step": 12495 }, { "epoch": 6.551362683438155, "grad_norm": 0.16751410067081451, "learning_rate": 4.252012165699286e-05, "loss": 0.5409, "num_input_tokens_seen": 8168928, "step": 12500 }, { "epoch": 6.55398322851153, "grad_norm": 0.16734671592712402, "learning_rate": 4.2511963143018634e-05, "loss": 0.4697, "num_input_tokens_seen": 8172320, "step": 12505 }, { "epoch": 6.556603773584905, "grad_norm": 0.13466250896453857, "learning_rate": 4.25038009657415e-05, "loss": 0.4197, "num_input_tokens_seen": 8175360, "step": 12510 }, { "epoch": 6.559224318658281, "grad_norm": 0.13211305439472198, "learning_rate": 4.24956351268689e-05, "loss": 0.6771, "num_input_tokens_seen": 8178464, "step": 12515 }, { "epoch": 6.561844863731656, "grad_norm": 0.24107934534549713, "learning_rate": 4.2487465628109036e-05, "loss": 0.4711, "num_input_tokens_seen": 8181536, "step": 12520 }, { "epoch": 6.564465408805032, "grad_norm": 0.2058018147945404, "learning_rate": 4.247929247117087e-05, "loss": 0.6295, "num_input_tokens_seen": 8187328, "step": 12525 }, { "epoch": 6.567085953878407, "grad_norm": 0.1877116709947586, "learning_rate": 4.247111565776413e-05, "loss": 0.5452, "num_input_tokens_seen": 8190400, "step": 12530 }, { "epoch": 6.569706498951782, "grad_norm": 0.17218144237995148, "learning_rate": 4.246293518959932e-05, "loss": 0.4047, "num_input_tokens_seen": 8193568, "step": 12535 }, { "epoch": 6.572327044025157, "grad_norm": 0.1632687747478485, "learning_rate": 4.245475106838769e-05, "loss": 0.6138, "num_input_tokens_seen": 8196672, "step": 12540 }, { "epoch": 6.5749475890985325, "grad_norm": 0.20983824133872986, "learning_rate": 4.2446563295841266e-05, "loss": 0.6188, "num_input_tokens_seen": 8199744, "step": 12545 }, { "epoch": 6.577568134171908, "grad_norm": 0.1428271383047104, "learning_rate": 4.243837187367285e-05, "loss": 0.4848, "num_input_tokens_seen": 8201984, "step": 12550 }, { "epoch": 6.580188679245283, "grad_norm": 0.1822759509086609, "learning_rate": 4.2430176803595987e-05, "loss": 0.4525, "num_input_tokens_seen": 8205344, "step": 12555 }, { "epoch": 6.582809224318658, "grad_norm": 0.1442473977804184, "learning_rate": 4.2421978087324985e-05, "loss": 0.4612, "num_input_tokens_seen": 8208736, "step": 12560 }, { "epoch": 6.585429769392033, "grad_norm": 0.16457919776439667, "learning_rate": 4.241377572657493e-05, "loss": 0.5892, "num_input_tokens_seen": 8212832, "step": 12565 }, { "epoch": 6.588050314465409, "grad_norm": 0.14117252826690674, "learning_rate": 4.2405569723061654e-05, "loss": 0.4964, "num_input_tokens_seen": 8216512, "step": 12570 }, { "epoch": 6.590670859538784, "grad_norm": 0.2202054113149643, "learning_rate": 4.239736007850178e-05, "loss": 0.5908, "num_input_tokens_seen": 8220000, "step": 12575 }, { "epoch": 6.59329140461216, "grad_norm": 0.1999366730451584, "learning_rate": 4.238914679461264e-05, "loss": 0.3979, "num_input_tokens_seen": 8223840, "step": 12580 }, { "epoch": 6.595911949685535, "grad_norm": 0.26717185974121094, "learning_rate": 4.2380929873112386e-05, "loss": 0.5071, "num_input_tokens_seen": 8226688, "step": 12585 }, { "epoch": 6.59853249475891, "grad_norm": 0.09945899993181229, "learning_rate": 4.2372709315719896e-05, "loss": 0.4403, "num_input_tokens_seen": 8230208, "step": 12590 }, { "epoch": 6.601153039832285, "grad_norm": 0.2046777606010437, "learning_rate": 4.2364485124154806e-05, "loss": 0.4216, "num_input_tokens_seen": 8232448, "step": 12595 }, { "epoch": 6.60377358490566, "grad_norm": 0.223697692155838, "learning_rate": 4.235625730013754e-05, "loss": 0.4839, "num_input_tokens_seen": 8235136, "step": 12600 }, { "epoch": 6.606394129979035, "grad_norm": 0.20619577169418335, "learning_rate": 4.234802584538926e-05, "loss": 0.4983, "num_input_tokens_seen": 8237856, "step": 12605 }, { "epoch": 6.609014675052411, "grad_norm": 0.21365010738372803, "learning_rate": 4.2339790761631884e-05, "loss": 0.4316, "num_input_tokens_seen": 8241472, "step": 12610 }, { "epoch": 6.611635220125786, "grad_norm": 0.16126932203769684, "learning_rate": 4.233155205058811e-05, "loss": 0.5754, "num_input_tokens_seen": 8245504, "step": 12615 }, { "epoch": 6.614255765199162, "grad_norm": 0.1656590849161148, "learning_rate": 4.232330971398137e-05, "loss": 0.5547, "num_input_tokens_seen": 8248480, "step": 12620 }, { "epoch": 6.616876310272537, "grad_norm": 0.17624910175800323, "learning_rate": 4.2315063753535875e-05, "loss": 0.6328, "num_input_tokens_seen": 8251968, "step": 12625 }, { "epoch": 6.619496855345912, "grad_norm": 0.21525079011917114, "learning_rate": 4.230681417097658e-05, "loss": 0.5204, "num_input_tokens_seen": 8254496, "step": 12630 }, { "epoch": 6.622117400419287, "grad_norm": 0.1978466957807541, "learning_rate": 4.229856096802921e-05, "loss": 0.7011, "num_input_tokens_seen": 8257472, "step": 12635 }, { "epoch": 6.6247379454926625, "grad_norm": 0.1400349885225296, "learning_rate": 4.229030414642024e-05, "loss": 0.5694, "num_input_tokens_seen": 8260768, "step": 12640 }, { "epoch": 6.627358490566038, "grad_norm": 0.14750991761684418, "learning_rate": 4.2282043707876896e-05, "loss": 0.4552, "num_input_tokens_seen": 8264672, "step": 12645 }, { "epoch": 6.629979035639413, "grad_norm": 0.13248591125011444, "learning_rate": 4.2273779654127175e-05, "loss": 0.4267, "num_input_tokens_seen": 8267424, "step": 12650 }, { "epoch": 6.632599580712788, "grad_norm": 0.18271145224571228, "learning_rate": 4.226551198689982e-05, "loss": 0.5779, "num_input_tokens_seen": 8270912, "step": 12655 }, { "epoch": 6.635220125786163, "grad_norm": 0.16535751521587372, "learning_rate": 4.225724070792434e-05, "loss": 0.5074, "num_input_tokens_seen": 8274080, "step": 12660 }, { "epoch": 6.637840670859539, "grad_norm": 0.19834907352924347, "learning_rate": 4.2248965818930975e-05, "loss": 0.4072, "num_input_tokens_seen": 8278656, "step": 12665 }, { "epoch": 6.640461215932914, "grad_norm": 0.174844428896904, "learning_rate": 4.224068732165075e-05, "loss": 0.3743, "num_input_tokens_seen": 8282496, "step": 12670 }, { "epoch": 6.6430817610062896, "grad_norm": 0.26887112855911255, "learning_rate": 4.223240521781543e-05, "loss": 0.5079, "num_input_tokens_seen": 8285504, "step": 12675 }, { "epoch": 6.645702306079665, "grad_norm": 0.20504358410835266, "learning_rate": 4.2224119509157535e-05, "loss": 0.5601, "num_input_tokens_seen": 8288448, "step": 12680 }, { "epoch": 6.64832285115304, "grad_norm": 0.3294432461261749, "learning_rate": 4.221583019741033e-05, "loss": 0.4199, "num_input_tokens_seen": 8291072, "step": 12685 }, { "epoch": 6.650943396226415, "grad_norm": 0.18771310150623322, "learning_rate": 4.2207537284307876e-05, "loss": 0.5539, "num_input_tokens_seen": 8294144, "step": 12690 }, { "epoch": 6.65356394129979, "grad_norm": 0.1142912432551384, "learning_rate": 4.219924077158493e-05, "loss": 0.6546, "num_input_tokens_seen": 8297408, "step": 12695 }, { "epoch": 6.656184486373165, "grad_norm": 0.17742645740509033, "learning_rate": 4.219094066097703e-05, "loss": 0.4883, "num_input_tokens_seen": 8300768, "step": 12700 }, { "epoch": 6.658805031446541, "grad_norm": 0.14655937254428864, "learning_rate": 4.218263695422047e-05, "loss": 0.4691, "num_input_tokens_seen": 8303616, "step": 12705 }, { "epoch": 6.661425576519916, "grad_norm": 0.11852780729532242, "learning_rate": 4.217432965305229e-05, "loss": 0.5934, "num_input_tokens_seen": 8307424, "step": 12710 }, { "epoch": 6.664046121593291, "grad_norm": 0.14909179508686066, "learning_rate": 4.2166018759210286e-05, "loss": 0.487, "num_input_tokens_seen": 8310560, "step": 12715 }, { "epoch": 6.666666666666667, "grad_norm": 0.2040916532278061, "learning_rate": 4.215770427443299e-05, "loss": 0.5239, "num_input_tokens_seen": 8312992, "step": 12720 }, { "epoch": 6.669287211740042, "grad_norm": 0.1900099366903305, "learning_rate": 4.214938620045971e-05, "loss": 0.4046, "num_input_tokens_seen": 8316544, "step": 12725 }, { "epoch": 6.671907756813417, "grad_norm": 0.23045536875724792, "learning_rate": 4.214106453903048e-05, "loss": 0.4464, "num_input_tokens_seen": 8319520, "step": 12730 }, { "epoch": 6.6745283018867925, "grad_norm": 0.1522996425628662, "learning_rate": 4.2132739291886114e-05, "loss": 0.5046, "num_input_tokens_seen": 8323872, "step": 12735 }, { "epoch": 6.677148846960168, "grad_norm": 0.15685689449310303, "learning_rate": 4.212441046076814e-05, "loss": 0.4751, "num_input_tokens_seen": 8326976, "step": 12740 }, { "epoch": 6.679769392033543, "grad_norm": 0.13096651434898376, "learning_rate": 4.211607804741887e-05, "loss": 0.4198, "num_input_tokens_seen": 8330304, "step": 12745 }, { "epoch": 6.682389937106918, "grad_norm": 0.2050398588180542, "learning_rate": 4.2107742053581344e-05, "loss": 0.4227, "num_input_tokens_seen": 8333632, "step": 12750 }, { "epoch": 6.685010482180293, "grad_norm": 0.2503328025341034, "learning_rate": 4.209940248099935e-05, "loss": 0.4733, "num_input_tokens_seen": 8336896, "step": 12755 }, { "epoch": 6.687631027253669, "grad_norm": 0.09203818440437317, "learning_rate": 4.2091059331417445e-05, "loss": 0.4827, "num_input_tokens_seen": 8340512, "step": 12760 }, { "epoch": 6.690251572327044, "grad_norm": 0.13133111596107483, "learning_rate": 4.208271260658091e-05, "loss": 0.5324, "num_input_tokens_seen": 8343680, "step": 12765 }, { "epoch": 6.6928721174004195, "grad_norm": 0.10995439440011978, "learning_rate": 4.2074362308235785e-05, "loss": 0.498, "num_input_tokens_seen": 8347456, "step": 12770 }, { "epoch": 6.695492662473795, "grad_norm": 0.11152401566505432, "learning_rate": 4.206600843812886e-05, "loss": 0.591, "num_input_tokens_seen": 8351200, "step": 12775 }, { "epoch": 6.69811320754717, "grad_norm": 0.22496774792671204, "learning_rate": 4.205765099800767e-05, "loss": 0.5164, "num_input_tokens_seen": 8354240, "step": 12780 }, { "epoch": 6.700733752620545, "grad_norm": 0.14278273284435272, "learning_rate": 4.204928998962049e-05, "loss": 0.5396, "num_input_tokens_seen": 8357856, "step": 12785 }, { "epoch": 6.70335429769392, "grad_norm": 0.14029526710510254, "learning_rate": 4.204092541471636e-05, "loss": 0.4818, "num_input_tokens_seen": 8361184, "step": 12790 }, { "epoch": 6.705974842767295, "grad_norm": 0.18524816632270813, "learning_rate": 4.203255727504504e-05, "loss": 0.497, "num_input_tokens_seen": 8364256, "step": 12795 }, { "epoch": 6.7085953878406706, "grad_norm": 0.24610882997512817, "learning_rate": 4.202418557235706e-05, "loss": 0.4934, "num_input_tokens_seen": 8367232, "step": 12800 }, { "epoch": 6.711215932914046, "grad_norm": 0.17927923798561096, "learning_rate": 4.201581030840368e-05, "loss": 0.6039, "num_input_tokens_seen": 8370688, "step": 12805 }, { "epoch": 6.713836477987421, "grad_norm": 0.11052736639976501, "learning_rate": 4.200743148493691e-05, "loss": 0.5228, "num_input_tokens_seen": 8374464, "step": 12810 }, { "epoch": 6.716457023060797, "grad_norm": 0.16694149374961853, "learning_rate": 4.199904910370951e-05, "loss": 0.4762, "num_input_tokens_seen": 8382368, "step": 12815 }, { "epoch": 6.719077568134172, "grad_norm": 0.15948790311813354, "learning_rate": 4.1990663166474964e-05, "loss": 0.5359, "num_input_tokens_seen": 8385248, "step": 12820 }, { "epoch": 6.721698113207547, "grad_norm": 0.21202434599399567, "learning_rate": 4.1982273674987524e-05, "loss": 0.4921, "num_input_tokens_seen": 8388128, "step": 12825 }, { "epoch": 6.7243186582809225, "grad_norm": 0.12559984624385834, "learning_rate": 4.197388063100218e-05, "loss": 0.5002, "num_input_tokens_seen": 8391008, "step": 12830 }, { "epoch": 6.726939203354298, "grad_norm": 0.11874645948410034, "learning_rate": 4.196548403627465e-05, "loss": 0.3944, "num_input_tokens_seen": 8393952, "step": 12835 }, { "epoch": 6.729559748427673, "grad_norm": 0.2070641815662384, "learning_rate": 4.195708389256142e-05, "loss": 0.488, "num_input_tokens_seen": 8396704, "step": 12840 }, { "epoch": 6.732180293501048, "grad_norm": 0.16755233705043793, "learning_rate": 4.194868020161969e-05, "loss": 0.5156, "num_input_tokens_seen": 8400160, "step": 12845 }, { "epoch": 6.734800838574423, "grad_norm": 0.12562990188598633, "learning_rate": 4.1940272965207425e-05, "loss": 0.4467, "num_input_tokens_seen": 8403168, "step": 12850 }, { "epoch": 6.737421383647799, "grad_norm": 0.13870179653167725, "learning_rate": 4.1931862185083326e-05, "loss": 0.3451, "num_input_tokens_seen": 8405952, "step": 12855 }, { "epoch": 6.740041928721174, "grad_norm": 0.2444201409816742, "learning_rate": 4.192344786300682e-05, "loss": 0.4642, "num_input_tokens_seen": 8408192, "step": 12860 }, { "epoch": 6.7426624737945495, "grad_norm": 0.17544297873973846, "learning_rate": 4.19150300007381e-05, "loss": 0.4311, "num_input_tokens_seen": 8410688, "step": 12865 }, { "epoch": 6.745283018867925, "grad_norm": 0.15068429708480835, "learning_rate": 4.190660860003806e-05, "loss": 0.4376, "num_input_tokens_seen": 8414688, "step": 12870 }, { "epoch": 6.7479035639413, "grad_norm": 0.1430456042289734, "learning_rate": 4.189818366266841e-05, "loss": 0.4331, "num_input_tokens_seen": 8418240, "step": 12875 }, { "epoch": 6.750524109014675, "grad_norm": 0.25422459840774536, "learning_rate": 4.188975519039151e-05, "loss": 0.5659, "num_input_tokens_seen": 8421888, "step": 12880 }, { "epoch": 6.75314465408805, "grad_norm": 0.1691594421863556, "learning_rate": 4.188132318497051e-05, "loss": 0.6041, "num_input_tokens_seen": 8424768, "step": 12885 }, { "epoch": 6.755765199161425, "grad_norm": 0.2543218731880188, "learning_rate": 4.1872887648169304e-05, "loss": 0.3673, "num_input_tokens_seen": 8428576, "step": 12890 }, { "epoch": 6.7583857442348005, "grad_norm": 0.1610824167728424, "learning_rate": 4.186444858175248e-05, "loss": 0.5221, "num_input_tokens_seen": 8431744, "step": 12895 }, { "epoch": 6.761006289308176, "grad_norm": 0.28346139192581177, "learning_rate": 4.1856005987485424e-05, "loss": 0.4234, "num_input_tokens_seen": 8435712, "step": 12900 }, { "epoch": 6.763626834381551, "grad_norm": 0.1707966923713684, "learning_rate": 4.1847559867134215e-05, "loss": 0.4353, "num_input_tokens_seen": 8439072, "step": 12905 }, { "epoch": 6.766247379454927, "grad_norm": 0.14330388605594635, "learning_rate": 4.1839110222465693e-05, "loss": 0.3587, "num_input_tokens_seen": 8443392, "step": 12910 }, { "epoch": 6.768867924528302, "grad_norm": 0.10074298828840256, "learning_rate": 4.183065705524742e-05, "loss": 0.4231, "num_input_tokens_seen": 8446592, "step": 12915 }, { "epoch": 6.771488469601677, "grad_norm": 0.19588272273540497, "learning_rate": 4.1822200367247696e-05, "loss": 0.4817, "num_input_tokens_seen": 8449632, "step": 12920 }, { "epoch": 6.774109014675052, "grad_norm": 0.1735985279083252, "learning_rate": 4.181374016023559e-05, "loss": 0.5216, "num_input_tokens_seen": 8452608, "step": 12925 }, { "epoch": 6.776729559748428, "grad_norm": 0.16348572075366974, "learning_rate": 4.1805276435980854e-05, "loss": 0.4787, "num_input_tokens_seen": 8454880, "step": 12930 }, { "epoch": 6.779350104821803, "grad_norm": 0.20207622647285461, "learning_rate": 4.1796809196254014e-05, "loss": 0.5265, "num_input_tokens_seen": 8457696, "step": 12935 }, { "epoch": 6.781970649895178, "grad_norm": 0.09540349245071411, "learning_rate": 4.178833844282632e-05, "loss": 0.5683, "num_input_tokens_seen": 8461088, "step": 12940 }, { "epoch": 6.784591194968553, "grad_norm": 0.3410591781139374, "learning_rate": 4.177986417746976e-05, "loss": 0.4995, "num_input_tokens_seen": 8464416, "step": 12945 }, { "epoch": 6.787211740041929, "grad_norm": 0.09995166212320328, "learning_rate": 4.177138640195704e-05, "loss": 0.5893, "num_input_tokens_seen": 8468512, "step": 12950 }, { "epoch": 6.789832285115304, "grad_norm": 0.09113252907991409, "learning_rate": 4.176290511806162e-05, "loss": 0.4327, "num_input_tokens_seen": 8472256, "step": 12955 }, { "epoch": 6.7924528301886795, "grad_norm": 0.22021909058094025, "learning_rate": 4.1754420327557706e-05, "loss": 0.4478, "num_input_tokens_seen": 8476416, "step": 12960 }, { "epoch": 6.795073375262055, "grad_norm": 0.13365072011947632, "learning_rate": 4.1745932032220195e-05, "loss": 0.4051, "num_input_tokens_seen": 8481184, "step": 12965 }, { "epoch": 6.79769392033543, "grad_norm": 0.20342355966567993, "learning_rate": 4.173744023382474e-05, "loss": 0.4547, "num_input_tokens_seen": 8485472, "step": 12970 }, { "epoch": 6.800314465408805, "grad_norm": 0.16175180673599243, "learning_rate": 4.172894493414776e-05, "loss": 0.5879, "num_input_tokens_seen": 8488544, "step": 12975 }, { "epoch": 6.80293501048218, "grad_norm": 0.24304533004760742, "learning_rate": 4.172044613496634e-05, "loss": 0.5207, "num_input_tokens_seen": 8491712, "step": 12980 }, { "epoch": 6.805555555555555, "grad_norm": 0.14198055863380432, "learning_rate": 4.171194383805835e-05, "loss": 0.5034, "num_input_tokens_seen": 8495008, "step": 12985 }, { "epoch": 6.8081761006289305, "grad_norm": 0.11630260944366455, "learning_rate": 4.170343804520237e-05, "loss": 0.3974, "num_input_tokens_seen": 8498464, "step": 12990 }, { "epoch": 6.810796645702306, "grad_norm": 0.15789054334163666, "learning_rate": 4.169492875817771e-05, "loss": 0.4824, "num_input_tokens_seen": 8501696, "step": 12995 }, { "epoch": 6.813417190775681, "grad_norm": 0.16548091173171997, "learning_rate": 4.1686415978764415e-05, "loss": 0.5554, "num_input_tokens_seen": 8504288, "step": 13000 }, { "epoch": 6.816037735849057, "grad_norm": 0.11296375840902328, "learning_rate": 4.167789970874327e-05, "loss": 0.4938, "num_input_tokens_seen": 8507840, "step": 13005 }, { "epoch": 6.818658280922432, "grad_norm": 0.2041979879140854, "learning_rate": 4.166937994989578e-05, "loss": 0.4969, "num_input_tokens_seen": 8510944, "step": 13010 }, { "epoch": 6.821278825995807, "grad_norm": 0.15539225935935974, "learning_rate": 4.166085670400418e-05, "loss": 0.5724, "num_input_tokens_seen": 8514944, "step": 13015 }, { "epoch": 6.823899371069182, "grad_norm": 0.13019445538520813, "learning_rate": 4.165232997285143e-05, "loss": 0.6488, "num_input_tokens_seen": 8517920, "step": 13020 }, { "epoch": 6.826519916142558, "grad_norm": 0.15526768565177917, "learning_rate": 4.164379975822124e-05, "loss": 0.5367, "num_input_tokens_seen": 8520448, "step": 13025 }, { "epoch": 6.829140461215933, "grad_norm": 0.13403759896755219, "learning_rate": 4.1635266061898015e-05, "loss": 0.5058, "num_input_tokens_seen": 8524000, "step": 13030 }, { "epoch": 6.831761006289308, "grad_norm": 0.1533069908618927, "learning_rate": 4.162672888566691e-05, "loss": 0.4429, "num_input_tokens_seen": 8526624, "step": 13035 }, { "epoch": 6.834381551362683, "grad_norm": 0.10413361340761185, "learning_rate": 4.1618188231313816e-05, "loss": 0.5064, "num_input_tokens_seen": 8529920, "step": 13040 }, { "epoch": 6.837002096436059, "grad_norm": 0.19178873300552368, "learning_rate": 4.1609644100625335e-05, "loss": 0.4807, "num_input_tokens_seen": 8532672, "step": 13045 }, { "epoch": 6.839622641509434, "grad_norm": 0.16523073613643646, "learning_rate": 4.16010964953888e-05, "loss": 0.438, "num_input_tokens_seen": 8535520, "step": 13050 }, { "epoch": 6.8422431865828095, "grad_norm": 0.20423811674118042, "learning_rate": 4.159254541739227e-05, "loss": 0.5905, "num_input_tokens_seen": 8537920, "step": 13055 }, { "epoch": 6.844863731656185, "grad_norm": 0.1491660326719284, "learning_rate": 4.158399086842454e-05, "loss": 0.4912, "num_input_tokens_seen": 8541984, "step": 13060 }, { "epoch": 6.84748427672956, "grad_norm": 0.17137792706489563, "learning_rate": 4.157543285027512e-05, "loss": 0.4955, "num_input_tokens_seen": 8546272, "step": 13065 }, { "epoch": 6.850104821802935, "grad_norm": 0.3257315158843994, "learning_rate": 4.156687136473425e-05, "loss": 0.6332, "num_input_tokens_seen": 8549280, "step": 13070 }, { "epoch": 6.85272536687631, "grad_norm": 0.23933809995651245, "learning_rate": 4.1558306413592894e-05, "loss": 0.4129, "num_input_tokens_seen": 8552064, "step": 13075 }, { "epoch": 6.855345911949685, "grad_norm": 0.27703365683555603, "learning_rate": 4.1549737998642745e-05, "loss": 0.4503, "num_input_tokens_seen": 8555712, "step": 13080 }, { "epoch": 6.8579664570230605, "grad_norm": 0.12901850044727325, "learning_rate": 4.154116612167622e-05, "loss": 0.5374, "num_input_tokens_seen": 8559328, "step": 13085 }, { "epoch": 6.860587002096436, "grad_norm": 0.14364415407180786, "learning_rate": 4.1532590784486444e-05, "loss": 0.6235, "num_input_tokens_seen": 8561856, "step": 13090 }, { "epoch": 6.863207547169811, "grad_norm": 0.2262519747018814, "learning_rate": 4.1524011988867294e-05, "loss": 0.6023, "num_input_tokens_seen": 8564960, "step": 13095 }, { "epoch": 6.865828092243187, "grad_norm": 0.12965404987335205, "learning_rate": 4.151542973661336e-05, "loss": 0.5736, "num_input_tokens_seen": 8568640, "step": 13100 }, { "epoch": 6.868448637316562, "grad_norm": 0.17774206399917603, "learning_rate": 4.150684402951994e-05, "loss": 0.4787, "num_input_tokens_seen": 8572064, "step": 13105 }, { "epoch": 6.871069182389937, "grad_norm": 0.16007982194423676, "learning_rate": 4.149825486938306e-05, "loss": 0.4124, "num_input_tokens_seen": 8575104, "step": 13110 }, { "epoch": 6.873689727463312, "grad_norm": 0.2506212294101715, "learning_rate": 4.148966225799949e-05, "loss": 0.6318, "num_input_tokens_seen": 8577888, "step": 13115 }, { "epoch": 6.876310272536688, "grad_norm": 0.12200082093477249, "learning_rate": 4.14810661971667e-05, "loss": 0.5723, "num_input_tokens_seen": 8580896, "step": 13120 }, { "epoch": 6.878930817610063, "grad_norm": 0.16545836627483368, "learning_rate": 4.147246668868289e-05, "loss": 0.5427, "num_input_tokens_seen": 8584864, "step": 13125 }, { "epoch": 6.881551362683438, "grad_norm": 0.14540544152259827, "learning_rate": 4.1463863734346964e-05, "loss": 0.7392, "num_input_tokens_seen": 8587936, "step": 13130 }, { "epoch": 6.884171907756813, "grad_norm": 0.14554719626903534, "learning_rate": 4.145525733595859e-05, "loss": 0.4508, "num_input_tokens_seen": 8592064, "step": 13135 }, { "epoch": 6.886792452830189, "grad_norm": 0.3711107671260834, "learning_rate": 4.14466474953181e-05, "loss": 0.5517, "num_input_tokens_seen": 8595136, "step": 13140 }, { "epoch": 6.889412997903564, "grad_norm": 0.12946993112564087, "learning_rate": 4.14380342142266e-05, "loss": 0.4846, "num_input_tokens_seen": 8597472, "step": 13145 }, { "epoch": 6.8920335429769395, "grad_norm": 0.27516332268714905, "learning_rate": 4.142941749448587e-05, "loss": 0.5773, "num_input_tokens_seen": 8600000, "step": 13150 }, { "epoch": 6.894654088050315, "grad_norm": 0.14532175660133362, "learning_rate": 4.142079733789843e-05, "loss": 0.3979, "num_input_tokens_seen": 8604960, "step": 13155 }, { "epoch": 6.89727463312369, "grad_norm": 0.11712977290153503, "learning_rate": 4.141217374626753e-05, "loss": 0.4877, "num_input_tokens_seen": 8607776, "step": 13160 }, { "epoch": 6.899895178197065, "grad_norm": 0.19712094962596893, "learning_rate": 4.140354672139713e-05, "loss": 0.4495, "num_input_tokens_seen": 8610944, "step": 13165 }, { "epoch": 6.90251572327044, "grad_norm": 0.2108861654996872, "learning_rate": 4.1394916265091886e-05, "loss": 0.582, "num_input_tokens_seen": 8614368, "step": 13170 }, { "epoch": 6.905136268343815, "grad_norm": 0.13918933272361755, "learning_rate": 4.1386282379157214e-05, "loss": 0.4733, "num_input_tokens_seen": 8618272, "step": 13175 }, { "epoch": 6.9077568134171905, "grad_norm": 0.2264728993177414, "learning_rate": 4.137764506539921e-05, "loss": 0.3312, "num_input_tokens_seen": 8620640, "step": 13180 }, { "epoch": 6.910377358490566, "grad_norm": 0.16766870021820068, "learning_rate": 4.136900432562471e-05, "loss": 0.4392, "num_input_tokens_seen": 8624192, "step": 13185 }, { "epoch": 6.912997903563941, "grad_norm": 0.27769768238067627, "learning_rate": 4.136036016164123e-05, "loss": 0.3882, "num_input_tokens_seen": 8627072, "step": 13190 }, { "epoch": 6.915618448637317, "grad_norm": 0.19721555709838867, "learning_rate": 4.135171257525707e-05, "loss": 0.4656, "num_input_tokens_seen": 8631232, "step": 13195 }, { "epoch": 6.918238993710692, "grad_norm": 0.18050406873226166, "learning_rate": 4.1343061568281195e-05, "loss": 0.6429, "num_input_tokens_seen": 8633696, "step": 13200 }, { "epoch": 6.920859538784067, "grad_norm": 0.2335425317287445, "learning_rate": 4.133440714252328e-05, "loss": 0.3686, "num_input_tokens_seen": 8636800, "step": 13205 }, { "epoch": 6.923480083857442, "grad_norm": 0.15424203872680664, "learning_rate": 4.132574929979375e-05, "loss": 0.4749, "num_input_tokens_seen": 8639968, "step": 13210 }, { "epoch": 6.926100628930818, "grad_norm": 0.10713745653629303, "learning_rate": 4.1317088041903714e-05, "loss": 0.5005, "num_input_tokens_seen": 8643328, "step": 13215 }, { "epoch": 6.928721174004193, "grad_norm": 0.14033763110637665, "learning_rate": 4.1308423370665015e-05, "loss": 0.4398, "num_input_tokens_seen": 8646144, "step": 13220 }, { "epoch": 6.931341719077568, "grad_norm": 0.13052356243133545, "learning_rate": 4.1299755287890206e-05, "loss": 0.5301, "num_input_tokens_seen": 8649536, "step": 13225 }, { "epoch": 6.933962264150943, "grad_norm": 0.13127104938030243, "learning_rate": 4.129108379539255e-05, "loss": 0.3544, "num_input_tokens_seen": 8652992, "step": 13230 }, { "epoch": 6.936582809224319, "grad_norm": 0.22042790055274963, "learning_rate": 4.128240889498602e-05, "loss": 0.4546, "num_input_tokens_seen": 8656608, "step": 13235 }, { "epoch": 6.939203354297694, "grad_norm": 0.18781857192516327, "learning_rate": 4.127373058848531e-05, "loss": 0.6844, "num_input_tokens_seen": 8659744, "step": 13240 }, { "epoch": 6.9418238993710695, "grad_norm": 0.13905784487724304, "learning_rate": 4.126504887770582e-05, "loss": 0.5376, "num_input_tokens_seen": 8662176, "step": 13245 }, { "epoch": 6.944444444444445, "grad_norm": 0.16005392372608185, "learning_rate": 4.1256363764463666e-05, "loss": 0.481, "num_input_tokens_seen": 8665184, "step": 13250 }, { "epoch": 6.94706498951782, "grad_norm": 0.3833385705947876, "learning_rate": 4.1247675250575676e-05, "loss": 0.4746, "num_input_tokens_seen": 8668096, "step": 13255 }, { "epoch": 6.949685534591195, "grad_norm": 0.1050202026963234, "learning_rate": 4.1238983337859396e-05, "loss": 0.458, "num_input_tokens_seen": 8671392, "step": 13260 }, { "epoch": 6.95230607966457, "grad_norm": 0.158879354596138, "learning_rate": 4.123028802813305e-05, "loss": 0.7045, "num_input_tokens_seen": 8675360, "step": 13265 }, { "epoch": 6.954926624737945, "grad_norm": 0.15302838385105133, "learning_rate": 4.1221589323215625e-05, "loss": 0.3864, "num_input_tokens_seen": 8678880, "step": 13270 }, { "epoch": 6.9575471698113205, "grad_norm": 0.3239055275917053, "learning_rate": 4.121288722492678e-05, "loss": 0.3845, "num_input_tokens_seen": 8682176, "step": 13275 }, { "epoch": 6.960167714884696, "grad_norm": 0.3086394965648651, "learning_rate": 4.120418173508688e-05, "loss": 0.4101, "num_input_tokens_seen": 8685024, "step": 13280 }, { "epoch": 6.962788259958071, "grad_norm": 0.20263689756393433, "learning_rate": 4.1195472855517045e-05, "loss": 0.66, "num_input_tokens_seen": 8688768, "step": 13285 }, { "epoch": 6.965408805031447, "grad_norm": 0.12979233264923096, "learning_rate": 4.1186760588039045e-05, "loss": 0.4242, "num_input_tokens_seen": 8692032, "step": 13290 }, { "epoch": 6.968029350104822, "grad_norm": 0.13221241533756256, "learning_rate": 4.11780449344754e-05, "loss": 0.539, "num_input_tokens_seen": 8695744, "step": 13295 }, { "epoch": 6.970649895178197, "grad_norm": 0.13633382320404053, "learning_rate": 4.1169325896649326e-05, "loss": 0.5389, "num_input_tokens_seen": 8699680, "step": 13300 }, { "epoch": 6.973270440251572, "grad_norm": 0.13702192902565002, "learning_rate": 4.116060347638474e-05, "loss": 0.4116, "num_input_tokens_seen": 8702816, "step": 13305 }, { "epoch": 6.975890985324948, "grad_norm": 0.11575435101985931, "learning_rate": 4.115187767550628e-05, "loss": 0.4482, "num_input_tokens_seen": 8705600, "step": 13310 }, { "epoch": 6.978511530398323, "grad_norm": 0.1478942483663559, "learning_rate": 4.114314849583928e-05, "loss": 0.4818, "num_input_tokens_seen": 8708064, "step": 13315 }, { "epoch": 6.981132075471698, "grad_norm": 0.17434315383434296, "learning_rate": 4.113441593920979e-05, "loss": 0.4098, "num_input_tokens_seen": 8710944, "step": 13320 }, { "epoch": 6.983752620545073, "grad_norm": 0.1468658298254013, "learning_rate": 4.112568000744454e-05, "loss": 0.6011, "num_input_tokens_seen": 8714368, "step": 13325 }, { "epoch": 6.986373165618449, "grad_norm": 0.13480983674526215, "learning_rate": 4.1116940702371025e-05, "loss": 0.5125, "num_input_tokens_seen": 8717824, "step": 13330 }, { "epoch": 6.988993710691824, "grad_norm": 0.1794731616973877, "learning_rate": 4.1108198025817376e-05, "loss": 0.4849, "num_input_tokens_seen": 8720928, "step": 13335 }, { "epoch": 6.9916142557651995, "grad_norm": 0.2956675887107849, "learning_rate": 4.1099451979612476e-05, "loss": 0.5278, "num_input_tokens_seen": 8724704, "step": 13340 }, { "epoch": 6.994234800838575, "grad_norm": 0.18579596281051636, "learning_rate": 4.109070256558589e-05, "loss": 0.4534, "num_input_tokens_seen": 8727232, "step": 13345 }, { "epoch": 6.99685534591195, "grad_norm": 0.18987992405891418, "learning_rate": 4.108194978556791e-05, "loss": 0.5187, "num_input_tokens_seen": 8729920, "step": 13350 }, { "epoch": 6.999475890985325, "grad_norm": 0.17475399374961853, "learning_rate": 4.1073193641389495e-05, "loss": 0.6984, "num_input_tokens_seen": 8732928, "step": 13355 }, { "epoch": 7.0, "eval_loss": 0.4722351133823395, "eval_runtime": 9.9651, "eval_samples_per_second": 85.097, "eval_steps_per_second": 21.274, "num_input_tokens_seen": 8733024, "step": 13356 }, { "epoch": 7.0020964360587, "grad_norm": 0.0961976870894432, "learning_rate": 4.1064434134882345e-05, "loss": 0.568, "num_input_tokens_seen": 8736160, "step": 13360 }, { "epoch": 7.004716981132075, "grad_norm": 0.1312410831451416, "learning_rate": 4.1055671267878847e-05, "loss": 0.5062, "num_input_tokens_seen": 8738656, "step": 13365 }, { "epoch": 7.0073375262054505, "grad_norm": 0.15876592695713043, "learning_rate": 4.1046905042212106e-05, "loss": 0.6979, "num_input_tokens_seen": 8742240, "step": 13370 }, { "epoch": 7.009958071278826, "grad_norm": 0.18076913058757782, "learning_rate": 4.103813545971589e-05, "loss": 0.5264, "num_input_tokens_seen": 8745312, "step": 13375 }, { "epoch": 7.012578616352202, "grad_norm": 0.21898153424263, "learning_rate": 4.10293625222247e-05, "loss": 0.4398, "num_input_tokens_seen": 8748032, "step": 13380 }, { "epoch": 7.015199161425577, "grad_norm": 0.20519129931926727, "learning_rate": 4.102058623157376e-05, "loss": 0.5864, "num_input_tokens_seen": 8751584, "step": 13385 }, { "epoch": 7.017819706498952, "grad_norm": 0.09349749982357025, "learning_rate": 4.101180658959894e-05, "loss": 0.4096, "num_input_tokens_seen": 8754432, "step": 13390 }, { "epoch": 7.020440251572327, "grad_norm": 0.2585204541683197, "learning_rate": 4.1003023598136866e-05, "loss": 0.5505, "num_input_tokens_seen": 8757344, "step": 13395 }, { "epoch": 7.023060796645702, "grad_norm": 0.023344865068793297, "learning_rate": 4.099423725902482e-05, "loss": 0.4103, "num_input_tokens_seen": 8764288, "step": 13400 }, { "epoch": 7.0256813417190775, "grad_norm": 0.1202884241938591, "learning_rate": 4.098544757410081e-05, "loss": 0.4786, "num_input_tokens_seen": 8768160, "step": 13405 }, { "epoch": 7.028301886792453, "grad_norm": 0.1324998289346695, "learning_rate": 4.097665454520354e-05, "loss": 0.4339, "num_input_tokens_seen": 8771712, "step": 13410 }, { "epoch": 7.030922431865828, "grad_norm": 0.15994355082511902, "learning_rate": 4.096785817417241e-05, "loss": 0.4896, "num_input_tokens_seen": 8775136, "step": 13415 }, { "epoch": 7.033542976939203, "grad_norm": 0.505622386932373, "learning_rate": 4.0959058462847524e-05, "loss": 0.545, "num_input_tokens_seen": 8778304, "step": 13420 }, { "epoch": 7.036163522012578, "grad_norm": 0.2215036004781723, "learning_rate": 4.095025541306968e-05, "loss": 0.5184, "num_input_tokens_seen": 8781024, "step": 13425 }, { "epoch": 7.038784067085954, "grad_norm": 0.1739261895418167, "learning_rate": 4.0941449026680374e-05, "loss": 0.6576, "num_input_tokens_seen": 8784000, "step": 13430 }, { "epoch": 7.0414046121593294, "grad_norm": 0.14212459325790405, "learning_rate": 4.09326393055218e-05, "loss": 0.6321, "num_input_tokens_seen": 8787424, "step": 13435 }, { "epoch": 7.044025157232705, "grad_norm": 0.12563517689704895, "learning_rate": 4.0923826251436845e-05, "loss": 0.4025, "num_input_tokens_seen": 8790816, "step": 13440 }, { "epoch": 7.04664570230608, "grad_norm": 0.4033562242984772, "learning_rate": 4.0915009866269115e-05, "loss": 0.4159, "num_input_tokens_seen": 8792992, "step": 13445 }, { "epoch": 7.049266247379455, "grad_norm": 0.17878033220767975, "learning_rate": 4.090619015186288e-05, "loss": 0.3472, "num_input_tokens_seen": 8795840, "step": 13450 }, { "epoch": 7.05188679245283, "grad_norm": 0.17209632694721222, "learning_rate": 4.0897367110063145e-05, "loss": 0.4581, "num_input_tokens_seen": 8798560, "step": 13455 }, { "epoch": 7.054507337526205, "grad_norm": 0.25595688819885254, "learning_rate": 4.088854074271557e-05, "loss": 0.7036, "num_input_tokens_seen": 8802240, "step": 13460 }, { "epoch": 7.0571278825995805, "grad_norm": 0.13718542456626892, "learning_rate": 4.0879711051666534e-05, "loss": 0.3996, "num_input_tokens_seen": 8804960, "step": 13465 }, { "epoch": 7.059748427672956, "grad_norm": 0.18972860276699066, "learning_rate": 4.087087803876313e-05, "loss": 0.4847, "num_input_tokens_seen": 8808000, "step": 13470 }, { "epoch": 7.062368972746331, "grad_norm": 0.17061245441436768, "learning_rate": 4.086204170585308e-05, "loss": 0.4261, "num_input_tokens_seen": 8811008, "step": 13475 }, { "epoch": 7.064989517819707, "grad_norm": 0.1315740942955017, "learning_rate": 4.085320205478488e-05, "loss": 0.4845, "num_input_tokens_seen": 8813728, "step": 13480 }, { "epoch": 7.067610062893082, "grad_norm": 0.18345309793949127, "learning_rate": 4.084435908740768e-05, "loss": 0.5439, "num_input_tokens_seen": 8817824, "step": 13485 }, { "epoch": 7.070230607966457, "grad_norm": 0.3348276615142822, "learning_rate": 4.0835512805571306e-05, "loss": 0.5088, "num_input_tokens_seen": 8822848, "step": 13490 }, { "epoch": 7.072851153039832, "grad_norm": 0.1766349971294403, "learning_rate": 4.082666321112633e-05, "loss": 0.516, "num_input_tokens_seen": 8826016, "step": 13495 }, { "epoch": 7.0754716981132075, "grad_norm": 0.19958585500717163, "learning_rate": 4.081781030592396e-05, "loss": 0.4481, "num_input_tokens_seen": 8829088, "step": 13500 }, { "epoch": 7.078092243186583, "grad_norm": 0.1951211839914322, "learning_rate": 4.080895409181613e-05, "loss": 0.5024, "num_input_tokens_seen": 8832448, "step": 13505 }, { "epoch": 7.080712788259958, "grad_norm": 0.22040502727031708, "learning_rate": 4.080009457065547e-05, "loss": 0.4611, "num_input_tokens_seen": 8835232, "step": 13510 }, { "epoch": 7.083333333333333, "grad_norm": 0.19740456342697144, "learning_rate": 4.0791231744295286e-05, "loss": 0.3712, "num_input_tokens_seen": 8838176, "step": 13515 }, { "epoch": 7.085953878406708, "grad_norm": 0.13132299482822418, "learning_rate": 4.078236561458957e-05, "loss": 0.4051, "num_input_tokens_seen": 8841120, "step": 13520 }, { "epoch": 7.088574423480084, "grad_norm": 0.27565309405326843, "learning_rate": 4.077349618339302e-05, "loss": 0.5626, "num_input_tokens_seen": 8843744, "step": 13525 }, { "epoch": 7.091194968553459, "grad_norm": 0.15427842736244202, "learning_rate": 4.076462345256103e-05, "loss": 0.5039, "num_input_tokens_seen": 8846944, "step": 13530 }, { "epoch": 7.093815513626835, "grad_norm": 0.9027252197265625, "learning_rate": 4.0755747423949655e-05, "loss": 0.5004, "num_input_tokens_seen": 8849888, "step": 13535 }, { "epoch": 7.09643605870021, "grad_norm": 0.2553872764110565, "learning_rate": 4.074686809941568e-05, "loss": 0.5228, "num_input_tokens_seen": 8852608, "step": 13540 }, { "epoch": 7.099056603773585, "grad_norm": 0.20431925356388092, "learning_rate": 4.0737985480816546e-05, "loss": 0.4843, "num_input_tokens_seen": 8855072, "step": 13545 }, { "epoch": 7.10167714884696, "grad_norm": 0.1306646466255188, "learning_rate": 4.0729099570010395e-05, "loss": 0.4091, "num_input_tokens_seen": 8858656, "step": 13550 }, { "epoch": 7.104297693920335, "grad_norm": 0.19586598873138428, "learning_rate": 4.072021036885607e-05, "loss": 0.6035, "num_input_tokens_seen": 8861856, "step": 13555 }, { "epoch": 7.1069182389937104, "grad_norm": 0.19119761884212494, "learning_rate": 4.0711317879213075e-05, "loss": 0.4997, "num_input_tokens_seen": 8864704, "step": 13560 }, { "epoch": 7.109538784067086, "grad_norm": 0.21263709664344788, "learning_rate": 4.070242210294162e-05, "loss": 0.4921, "num_input_tokens_seen": 8867776, "step": 13565 }, { "epoch": 7.112159329140461, "grad_norm": 0.18194006383419037, "learning_rate": 4.069352304190261e-05, "loss": 0.5364, "num_input_tokens_seen": 8870560, "step": 13570 }, { "epoch": 7.114779874213837, "grad_norm": 0.12066952884197235, "learning_rate": 4.068462069795761e-05, "loss": 0.6006, "num_input_tokens_seen": 8874112, "step": 13575 }, { "epoch": 7.117400419287212, "grad_norm": 0.16498492658138275, "learning_rate": 4.067571507296891e-05, "loss": 0.5206, "num_input_tokens_seen": 8877184, "step": 13580 }, { "epoch": 7.120020964360587, "grad_norm": 0.1886696219444275, "learning_rate": 4.0666806168799455e-05, "loss": 0.5887, "num_input_tokens_seen": 8880512, "step": 13585 }, { "epoch": 7.122641509433962, "grad_norm": 0.3000020980834961, "learning_rate": 4.0657893987312876e-05, "loss": 0.5415, "num_input_tokens_seen": 8883232, "step": 13590 }, { "epoch": 7.1252620545073375, "grad_norm": 0.17804685235023499, "learning_rate": 4.064897853037352e-05, "loss": 0.544, "num_input_tokens_seen": 8887008, "step": 13595 }, { "epoch": 7.127882599580713, "grad_norm": 0.26127371191978455, "learning_rate": 4.064005979984639e-05, "loss": 0.4925, "num_input_tokens_seen": 8889952, "step": 13600 }, { "epoch": 7.130503144654088, "grad_norm": 0.14342352747917175, "learning_rate": 4.0631137797597176e-05, "loss": 0.5122, "num_input_tokens_seen": 8893408, "step": 13605 }, { "epoch": 7.133123689727463, "grad_norm": 0.16227973997592926, "learning_rate": 4.0622212525492266e-05, "loss": 0.4614, "num_input_tokens_seen": 8897248, "step": 13610 }, { "epoch": 7.135744234800838, "grad_norm": 0.14494158327579498, "learning_rate": 4.061328398539872e-05, "loss": 0.5381, "num_input_tokens_seen": 8901152, "step": 13615 }, { "epoch": 7.138364779874214, "grad_norm": 0.1906096339225769, "learning_rate": 4.060435217918429e-05, "loss": 0.6041, "num_input_tokens_seen": 8904800, "step": 13620 }, { "epoch": 7.140985324947589, "grad_norm": 0.16073893010616302, "learning_rate": 4.0595417108717405e-05, "loss": 0.5194, "num_input_tokens_seen": 8907840, "step": 13625 }, { "epoch": 7.143605870020965, "grad_norm": 0.15736086666584015, "learning_rate": 4.0586478775867185e-05, "loss": 0.4286, "num_input_tokens_seen": 8910976, "step": 13630 }, { "epoch": 7.14622641509434, "grad_norm": 0.16380861401557922, "learning_rate": 4.057753718250343e-05, "loss": 0.5204, "num_input_tokens_seen": 8913728, "step": 13635 }, { "epoch": 7.148846960167715, "grad_norm": 0.10403230041265488, "learning_rate": 4.056859233049661e-05, "loss": 0.3818, "num_input_tokens_seen": 8916928, "step": 13640 }, { "epoch": 7.15146750524109, "grad_norm": 0.26284176111221313, "learning_rate": 4.05596442217179e-05, "loss": 0.5461, "num_input_tokens_seen": 8920128, "step": 13645 }, { "epoch": 7.154088050314465, "grad_norm": 0.10888973623514175, "learning_rate": 4.055069285803912e-05, "loss": 0.4719, "num_input_tokens_seen": 8923424, "step": 13650 }, { "epoch": 7.15670859538784, "grad_norm": 0.1930631548166275, "learning_rate": 4.0541738241332815e-05, "loss": 0.3994, "num_input_tokens_seen": 8925856, "step": 13655 }, { "epoch": 7.159329140461216, "grad_norm": 0.4647325575351715, "learning_rate": 4.053278037347217e-05, "loss": 0.4351, "num_input_tokens_seen": 8928544, "step": 13660 }, { "epoch": 7.161949685534591, "grad_norm": 0.23714886605739594, "learning_rate": 4.052381925633109e-05, "loss": 0.5338, "num_input_tokens_seen": 8931136, "step": 13665 }, { "epoch": 7.164570230607967, "grad_norm": 0.14909939467906952, "learning_rate": 4.051485489178412e-05, "loss": 0.4957, "num_input_tokens_seen": 8934336, "step": 13670 }, { "epoch": 7.167190775681342, "grad_norm": 0.22492793202400208, "learning_rate": 4.05058872817065e-05, "loss": 0.5936, "num_input_tokens_seen": 8936896, "step": 13675 }, { "epoch": 7.169811320754717, "grad_norm": 0.17333798110485077, "learning_rate": 4.0496916427974177e-05, "loss": 0.4388, "num_input_tokens_seen": 8940224, "step": 13680 }, { "epoch": 7.172431865828092, "grad_norm": 0.22248400747776031, "learning_rate": 4.048794233246373e-05, "loss": 0.513, "num_input_tokens_seen": 8943840, "step": 13685 }, { "epoch": 7.1750524109014675, "grad_norm": 0.29075831174850464, "learning_rate": 4.0478964997052435e-05, "loss": 0.5066, "num_input_tokens_seen": 8946624, "step": 13690 }, { "epoch": 7.177672955974843, "grad_norm": 0.3824380338191986, "learning_rate": 4.046998442361827e-05, "loss": 0.663, "num_input_tokens_seen": 8950272, "step": 13695 }, { "epoch": 7.180293501048218, "grad_norm": 0.15044522285461426, "learning_rate": 4.046100061403984e-05, "loss": 0.4577, "num_input_tokens_seen": 8952608, "step": 13700 }, { "epoch": 7.182914046121593, "grad_norm": 0.14976924657821655, "learning_rate": 4.045201357019648e-05, "loss": 0.6411, "num_input_tokens_seen": 8955744, "step": 13705 }, { "epoch": 7.185534591194968, "grad_norm": 0.13955937325954437, "learning_rate": 4.044302329396816e-05, "loss": 0.3437, "num_input_tokens_seen": 8958496, "step": 13710 }, { "epoch": 7.188155136268344, "grad_norm": 0.21905940771102905, "learning_rate": 4.0434029787235564e-05, "loss": 0.5194, "num_input_tokens_seen": 8961856, "step": 13715 }, { "epoch": 7.190775681341719, "grad_norm": 0.1021345779299736, "learning_rate": 4.042503305188e-05, "loss": 0.4573, "num_input_tokens_seen": 8964608, "step": 13720 }, { "epoch": 7.193396226415095, "grad_norm": 0.11569790542125702, "learning_rate": 4.041603308978351e-05, "loss": 0.4459, "num_input_tokens_seen": 8970016, "step": 13725 }, { "epoch": 7.19601677148847, "grad_norm": 0.21151131391525269, "learning_rate": 4.040702990282878e-05, "loss": 0.4439, "num_input_tokens_seen": 8973088, "step": 13730 }, { "epoch": 7.198637316561845, "grad_norm": 0.1218491792678833, "learning_rate": 4.039802349289914e-05, "loss": 0.5503, "num_input_tokens_seen": 8976416, "step": 13735 }, { "epoch": 7.20125786163522, "grad_norm": 0.15995877981185913, "learning_rate": 4.038901386187868e-05, "loss": 0.4441, "num_input_tokens_seen": 8979424, "step": 13740 }, { "epoch": 7.203878406708595, "grad_norm": 0.23196375370025635, "learning_rate": 4.038000101165208e-05, "loss": 0.5326, "num_input_tokens_seen": 8982944, "step": 13745 }, { "epoch": 7.20649895178197, "grad_norm": 0.20501157641410828, "learning_rate": 4.0370984944104725e-05, "loss": 0.4804, "num_input_tokens_seen": 8986144, "step": 13750 }, { "epoch": 7.209119496855346, "grad_norm": 0.17408768832683563, "learning_rate": 4.0361965661122695e-05, "loss": 0.4132, "num_input_tokens_seen": 8990048, "step": 13755 }, { "epoch": 7.211740041928721, "grad_norm": 0.1052561104297638, "learning_rate": 4.035294316459271e-05, "loss": 0.6008, "num_input_tokens_seen": 8992288, "step": 13760 }, { "epoch": 7.214360587002097, "grad_norm": 0.1787251979112625, "learning_rate": 4.034391745640215e-05, "loss": 0.4926, "num_input_tokens_seen": 8995200, "step": 13765 }, { "epoch": 7.216981132075472, "grad_norm": 0.1977422833442688, "learning_rate": 4.033488853843913e-05, "loss": 0.6074, "num_input_tokens_seen": 8998112, "step": 13770 }, { "epoch": 7.219601677148847, "grad_norm": 0.14295922219753265, "learning_rate": 4.0325856412592363e-05, "loss": 0.5881, "num_input_tokens_seen": 9002560, "step": 13775 }, { "epoch": 7.222222222222222, "grad_norm": 0.14160239696502686, "learning_rate": 4.031682108075129e-05, "loss": 0.4987, "num_input_tokens_seen": 9005536, "step": 13780 }, { "epoch": 7.2248427672955975, "grad_norm": 0.22362402081489563, "learning_rate": 4.030778254480599e-05, "loss": 0.5242, "num_input_tokens_seen": 9008480, "step": 13785 }, { "epoch": 7.227463312368973, "grad_norm": 0.42205655574798584, "learning_rate": 4.029874080664722e-05, "loss": 0.6343, "num_input_tokens_seen": 9011328, "step": 13790 }, { "epoch": 7.230083857442348, "grad_norm": 0.12367863953113556, "learning_rate": 4.0289695868166414e-05, "loss": 0.552, "num_input_tokens_seen": 9014304, "step": 13795 }, { "epoch": 7.232704402515723, "grad_norm": 0.16496595740318298, "learning_rate": 4.028064773125566e-05, "loss": 0.6794, "num_input_tokens_seen": 9016928, "step": 13800 }, { "epoch": 7.235324947589098, "grad_norm": 0.15276649594306946, "learning_rate": 4.0271596397807734e-05, "loss": 0.4735, "num_input_tokens_seen": 9019456, "step": 13805 }, { "epoch": 7.237945492662474, "grad_norm": 0.18514469265937805, "learning_rate": 4.026254186971607e-05, "loss": 0.5341, "num_input_tokens_seen": 9022816, "step": 13810 }, { "epoch": 7.240566037735849, "grad_norm": 0.20801901817321777, "learning_rate": 4.025348414887477e-05, "loss": 0.4322, "num_input_tokens_seen": 9025728, "step": 13815 }, { "epoch": 7.243186582809225, "grad_norm": 0.19035375118255615, "learning_rate": 4.024442323717861e-05, "loss": 0.5156, "num_input_tokens_seen": 9028864, "step": 13820 }, { "epoch": 7.2458071278826, "grad_norm": 0.41400787234306335, "learning_rate": 4.023535913652302e-05, "loss": 0.5585, "num_input_tokens_seen": 9032160, "step": 13825 }, { "epoch": 7.248427672955975, "grad_norm": 0.11925189197063446, "learning_rate": 4.022629184880412e-05, "loss": 0.2973, "num_input_tokens_seen": 9035872, "step": 13830 }, { "epoch": 7.25104821802935, "grad_norm": 0.09597589820623398, "learning_rate": 4.021722137591867e-05, "loss": 0.398, "num_input_tokens_seen": 9039456, "step": 13835 }, { "epoch": 7.253668763102725, "grad_norm": 0.1598869413137436, "learning_rate": 4.020814771976412e-05, "loss": 0.5503, "num_input_tokens_seen": 9042304, "step": 13840 }, { "epoch": 7.2562893081761, "grad_norm": 0.26468658447265625, "learning_rate": 4.019907088223857e-05, "loss": 0.6051, "num_input_tokens_seen": 9045760, "step": 13845 }, { "epoch": 7.258909853249476, "grad_norm": 0.10863026231527328, "learning_rate": 4.018999086524079e-05, "loss": 0.4805, "num_input_tokens_seen": 9048704, "step": 13850 }, { "epoch": 7.261530398322851, "grad_norm": 0.29179948568344116, "learning_rate": 4.018090767067022e-05, "loss": 0.503, "num_input_tokens_seen": 9052352, "step": 13855 }, { "epoch": 7.264150943396227, "grad_norm": 0.3373219668865204, "learning_rate": 4.017182130042696e-05, "loss": 0.4152, "num_input_tokens_seen": 9055136, "step": 13860 }, { "epoch": 7.266771488469602, "grad_norm": 0.16343148052692413, "learning_rate": 4.0162731756411764e-05, "loss": 0.5257, "num_input_tokens_seen": 9057632, "step": 13865 }, { "epoch": 7.269392033542977, "grad_norm": 0.08648253232240677, "learning_rate": 4.015363904052608e-05, "loss": 0.4295, "num_input_tokens_seen": 9061248, "step": 13870 }, { "epoch": 7.272012578616352, "grad_norm": 0.23141072690486908, "learning_rate": 4.014454315467199e-05, "loss": 0.463, "num_input_tokens_seen": 9064800, "step": 13875 }, { "epoch": 7.2746331236897275, "grad_norm": 0.262739360332489, "learning_rate": 4.013544410075224e-05, "loss": 0.4634, "num_input_tokens_seen": 9068928, "step": 13880 }, { "epoch": 7.277253668763103, "grad_norm": 0.1239750012755394, "learning_rate": 4.0126341880670273e-05, "loss": 0.3517, "num_input_tokens_seen": 9072160, "step": 13885 }, { "epoch": 7.279874213836478, "grad_norm": 0.24012258648872375, "learning_rate": 4.011723649633014e-05, "loss": 0.4753, "num_input_tokens_seen": 9074784, "step": 13890 }, { "epoch": 7.282494758909853, "grad_norm": 0.16672484576702118, "learning_rate": 4.0108127949636606e-05, "loss": 0.4433, "num_input_tokens_seen": 9078688, "step": 13895 }, { "epoch": 7.285115303983228, "grad_norm": 0.2819156348705292, "learning_rate": 4.009901624249507e-05, "loss": 0.5896, "num_input_tokens_seen": 9081184, "step": 13900 }, { "epoch": 7.287735849056604, "grad_norm": 0.1188502311706543, "learning_rate": 4.008990137681159e-05, "loss": 0.5555, "num_input_tokens_seen": 9085344, "step": 13905 }, { "epoch": 7.290356394129979, "grad_norm": 0.16448703408241272, "learning_rate": 4.00807833544929e-05, "loss": 0.451, "num_input_tokens_seen": 9088384, "step": 13910 }, { "epoch": 7.2929769392033545, "grad_norm": 0.17255771160125732, "learning_rate": 4.0071662177446376e-05, "loss": 0.4608, "num_input_tokens_seen": 9090848, "step": 13915 }, { "epoch": 7.29559748427673, "grad_norm": 0.15237459540367126, "learning_rate": 4.006253784758008e-05, "loss": 0.5048, "num_input_tokens_seen": 9095328, "step": 13920 }, { "epoch": 7.298218029350105, "grad_norm": 0.1982715129852295, "learning_rate": 4.0053410366802704e-05, "loss": 0.4724, "num_input_tokens_seen": 9098464, "step": 13925 }, { "epoch": 7.30083857442348, "grad_norm": 0.13076505064964294, "learning_rate": 4.004427973702362e-05, "loss": 0.3064, "num_input_tokens_seen": 9103360, "step": 13930 }, { "epoch": 7.303459119496855, "grad_norm": 0.1461620330810547, "learning_rate": 4.003514596015285e-05, "loss": 0.4216, "num_input_tokens_seen": 9107328, "step": 13935 }, { "epoch": 7.30607966457023, "grad_norm": 0.14281439781188965, "learning_rate": 4.002600903810108e-05, "loss": 0.4902, "num_input_tokens_seen": 9110848, "step": 13940 }, { "epoch": 7.308700209643606, "grad_norm": 0.16628603637218475, "learning_rate": 4.001686897277963e-05, "loss": 0.5633, "num_input_tokens_seen": 9114112, "step": 13945 }, { "epoch": 7.311320754716981, "grad_norm": 0.5332033038139343, "learning_rate": 4.000772576610052e-05, "loss": 0.5572, "num_input_tokens_seen": 9116512, "step": 13950 }, { "epoch": 7.313941299790357, "grad_norm": 0.12487241625785828, "learning_rate": 3.99985794199764e-05, "loss": 0.4798, "num_input_tokens_seen": 9119584, "step": 13955 }, { "epoch": 7.316561844863732, "grad_norm": 0.23388469219207764, "learning_rate": 3.998942993632058e-05, "loss": 0.6081, "num_input_tokens_seen": 9122784, "step": 13960 }, { "epoch": 7.319182389937107, "grad_norm": 0.1188168004155159, "learning_rate": 3.998027731704702e-05, "loss": 0.5263, "num_input_tokens_seen": 9126080, "step": 13965 }, { "epoch": 7.321802935010482, "grad_norm": 0.32218897342681885, "learning_rate": 3.9971121564070354e-05, "loss": 0.5158, "num_input_tokens_seen": 9128832, "step": 13970 }, { "epoch": 7.3244234800838575, "grad_norm": 0.16470946371555328, "learning_rate": 3.996196267930585e-05, "loss": 0.616, "num_input_tokens_seen": 9132256, "step": 13975 }, { "epoch": 7.327044025157233, "grad_norm": 0.26929154992103577, "learning_rate": 3.995280066466945e-05, "loss": 0.5103, "num_input_tokens_seen": 9135584, "step": 13980 }, { "epoch": 7.329664570230608, "grad_norm": 0.155954509973526, "learning_rate": 3.9943635522077746e-05, "loss": 0.5314, "num_input_tokens_seen": 9138368, "step": 13985 }, { "epoch": 7.332285115303983, "grad_norm": 0.18842969834804535, "learning_rate": 3.993446725344797e-05, "loss": 0.4763, "num_input_tokens_seen": 9141248, "step": 13990 }, { "epoch": 7.334905660377358, "grad_norm": 0.12152811139822006, "learning_rate": 3.992529586069803e-05, "loss": 0.5315, "num_input_tokens_seen": 9145536, "step": 13995 }, { "epoch": 7.337526205450734, "grad_norm": 0.16950088739395142, "learning_rate": 3.991612134574648e-05, "loss": 0.4146, "num_input_tokens_seen": 9148224, "step": 14000 }, { "epoch": 7.340146750524109, "grad_norm": 0.2826422452926636, "learning_rate": 3.99069437105125e-05, "loss": 0.4398, "num_input_tokens_seen": 9152096, "step": 14005 }, { "epoch": 7.3427672955974845, "grad_norm": 0.16922765970230103, "learning_rate": 3.9897762956915976e-05, "loss": 0.5626, "num_input_tokens_seen": 9155232, "step": 14010 }, { "epoch": 7.34538784067086, "grad_norm": 0.14686565101146698, "learning_rate": 3.9888579086877403e-05, "loss": 0.4381, "num_input_tokens_seen": 9158720, "step": 14015 }, { "epoch": 7.348008385744235, "grad_norm": 0.14136402308940887, "learning_rate": 3.987939210231794e-05, "loss": 0.3439, "num_input_tokens_seen": 9162560, "step": 14020 }, { "epoch": 7.35062893081761, "grad_norm": 0.0953274741768837, "learning_rate": 3.987020200515941e-05, "loss": 0.547, "num_input_tokens_seen": 9165504, "step": 14025 }, { "epoch": 7.353249475890985, "grad_norm": 0.14253856241703033, "learning_rate": 3.9861008797324255e-05, "loss": 0.506, "num_input_tokens_seen": 9168288, "step": 14030 }, { "epoch": 7.35587002096436, "grad_norm": 0.27515479922294617, "learning_rate": 3.9851812480735605e-05, "loss": 0.5145, "num_input_tokens_seen": 9171168, "step": 14035 }, { "epoch": 7.3584905660377355, "grad_norm": 0.15323425829410553, "learning_rate": 3.9842613057317235e-05, "loss": 0.4463, "num_input_tokens_seen": 9176960, "step": 14040 }, { "epoch": 7.361111111111111, "grad_norm": 0.12767046689987183, "learning_rate": 3.9833410528993534e-05, "loss": 0.4811, "num_input_tokens_seen": 9181760, "step": 14045 }, { "epoch": 7.363731656184487, "grad_norm": 0.21611224114894867, "learning_rate": 3.9824204897689585e-05, "loss": 0.4735, "num_input_tokens_seen": 9185280, "step": 14050 }, { "epoch": 7.366352201257862, "grad_norm": 0.1268635094165802, "learning_rate": 3.98149961653311e-05, "loss": 0.501, "num_input_tokens_seen": 9189184, "step": 14055 }, { "epoch": 7.368972746331237, "grad_norm": 0.12311476469039917, "learning_rate": 3.980578433384442e-05, "loss": 0.5561, "num_input_tokens_seen": 9191904, "step": 14060 }, { "epoch": 7.371593291404612, "grad_norm": 0.1754574477672577, "learning_rate": 3.979656940515659e-05, "loss": 0.3725, "num_input_tokens_seen": 9194496, "step": 14065 }, { "epoch": 7.3742138364779874, "grad_norm": 0.09987558424472809, "learning_rate": 3.9787351381195235e-05, "loss": 0.4228, "num_input_tokens_seen": 9197184, "step": 14070 }, { "epoch": 7.376834381551363, "grad_norm": 0.12765739858150482, "learning_rate": 3.977813026388868e-05, "loss": 0.5088, "num_input_tokens_seen": 9204128, "step": 14075 }, { "epoch": 7.379454926624738, "grad_norm": 0.13771605491638184, "learning_rate": 3.976890605516587e-05, "loss": 0.4129, "num_input_tokens_seen": 9206912, "step": 14080 }, { "epoch": 7.382075471698113, "grad_norm": 0.1493500918149948, "learning_rate": 3.97596787569564e-05, "loss": 0.4891, "num_input_tokens_seen": 9210144, "step": 14085 }, { "epoch": 7.384696016771488, "grad_norm": 0.13144336640834808, "learning_rate": 3.975044837119053e-05, "loss": 0.4742, "num_input_tokens_seen": 9214336, "step": 14090 }, { "epoch": 7.387316561844864, "grad_norm": 0.14960475265979767, "learning_rate": 3.974121489979914e-05, "loss": 0.5031, "num_input_tokens_seen": 9217728, "step": 14095 }, { "epoch": 7.389937106918239, "grad_norm": 0.16231058537960052, "learning_rate": 3.973197834471377e-05, "loss": 0.5437, "num_input_tokens_seen": 9221280, "step": 14100 }, { "epoch": 7.3925576519916145, "grad_norm": 0.1532057523727417, "learning_rate": 3.9722738707866595e-05, "loss": 0.5123, "num_input_tokens_seen": 9225376, "step": 14105 }, { "epoch": 7.39517819706499, "grad_norm": 0.12616726756095886, "learning_rate": 3.971349599119046e-05, "loss": 0.4493, "num_input_tokens_seen": 9229088, "step": 14110 }, { "epoch": 7.397798742138365, "grad_norm": 0.1683245152235031, "learning_rate": 3.970425019661881e-05, "loss": 0.4347, "num_input_tokens_seen": 9231840, "step": 14115 }, { "epoch": 7.40041928721174, "grad_norm": 0.21288150548934937, "learning_rate": 3.9695001326085785e-05, "loss": 0.5141, "num_input_tokens_seen": 9235264, "step": 14120 }, { "epoch": 7.403039832285115, "grad_norm": 0.1323375552892685, "learning_rate": 3.968574938152613e-05, "loss": 0.5446, "num_input_tokens_seen": 9238656, "step": 14125 }, { "epoch": 7.40566037735849, "grad_norm": 0.1372985541820526, "learning_rate": 3.9676494364875247e-05, "loss": 0.5268, "num_input_tokens_seen": 9242144, "step": 14130 }, { "epoch": 7.4082809224318655, "grad_norm": 0.2572752833366394, "learning_rate": 3.966723627806918e-05, "loss": 0.5506, "num_input_tokens_seen": 9245024, "step": 14135 }, { "epoch": 7.410901467505241, "grad_norm": 0.1657617688179016, "learning_rate": 3.9657975123044624e-05, "loss": 0.4693, "num_input_tokens_seen": 9248032, "step": 14140 }, { "epoch": 7.413522012578617, "grad_norm": 0.2581371068954468, "learning_rate": 3.964871090173888e-05, "loss": 0.5092, "num_input_tokens_seen": 9250240, "step": 14145 }, { "epoch": 7.416142557651992, "grad_norm": 0.15173877775669098, "learning_rate": 3.963944361608995e-05, "loss": 0.5414, "num_input_tokens_seen": 9253408, "step": 14150 }, { "epoch": 7.418763102725367, "grad_norm": 0.17893856763839722, "learning_rate": 3.963017326803644e-05, "loss": 0.4779, "num_input_tokens_seen": 9256320, "step": 14155 }, { "epoch": 7.421383647798742, "grad_norm": 0.07728017121553421, "learning_rate": 3.962089985951758e-05, "loss": 0.4826, "num_input_tokens_seen": 9261312, "step": 14160 }, { "epoch": 7.424004192872117, "grad_norm": 0.19387701153755188, "learning_rate": 3.961162339247327e-05, "loss": 0.5965, "num_input_tokens_seen": 9264544, "step": 14165 }, { "epoch": 7.426624737945493, "grad_norm": 0.17611782252788544, "learning_rate": 3.9602343868844055e-05, "loss": 0.5483, "num_input_tokens_seen": 9267584, "step": 14170 }, { "epoch": 7.429245283018868, "grad_norm": 0.1872110664844513, "learning_rate": 3.959306129057109e-05, "loss": 0.5157, "num_input_tokens_seen": 9270784, "step": 14175 }, { "epoch": 7.431865828092243, "grad_norm": 0.1238371804356575, "learning_rate": 3.958377565959619e-05, "loss": 0.4168, "num_input_tokens_seen": 9274048, "step": 14180 }, { "epoch": 7.434486373165618, "grad_norm": 0.09810732305049896, "learning_rate": 3.95744869778618e-05, "loss": 0.4057, "num_input_tokens_seen": 9277760, "step": 14185 }, { "epoch": 7.437106918238994, "grad_norm": 0.27112483978271484, "learning_rate": 3.956519524731101e-05, "loss": 0.5421, "num_input_tokens_seen": 9281088, "step": 14190 }, { "epoch": 7.439727463312369, "grad_norm": 0.15579280257225037, "learning_rate": 3.9555900469887555e-05, "loss": 0.5522, "num_input_tokens_seen": 9283968, "step": 14195 }, { "epoch": 7.4423480083857445, "grad_norm": 0.23153257369995117, "learning_rate": 3.954660264753577e-05, "loss": 0.5937, "num_input_tokens_seen": 9286816, "step": 14200 }, { "epoch": 7.44496855345912, "grad_norm": 0.158336341381073, "learning_rate": 3.953730178220067e-05, "loss": 0.5642, "num_input_tokens_seen": 9290272, "step": 14205 }, { "epoch": 7.447589098532495, "grad_norm": 0.1930699497461319, "learning_rate": 3.952799787582789e-05, "loss": 0.6111, "num_input_tokens_seen": 9293568, "step": 14210 }, { "epoch": 7.45020964360587, "grad_norm": 0.3164703845977783, "learning_rate": 3.9518690930363694e-05, "loss": 0.3467, "num_input_tokens_seen": 9295968, "step": 14215 }, { "epoch": 7.452830188679245, "grad_norm": 0.16151835024356842, "learning_rate": 3.9509380947754994e-05, "loss": 0.413, "num_input_tokens_seen": 9300128, "step": 14220 }, { "epoch": 7.45545073375262, "grad_norm": 0.3346072733402252, "learning_rate": 3.9500067929949336e-05, "loss": 0.4645, "num_input_tokens_seen": 9302816, "step": 14225 }, { "epoch": 7.4580712788259955, "grad_norm": 0.17907506227493286, "learning_rate": 3.9490751878894895e-05, "loss": 0.5788, "num_input_tokens_seen": 9305632, "step": 14230 }, { "epoch": 7.460691823899371, "grad_norm": 0.4327690005302429, "learning_rate": 3.948143279654047e-05, "loss": 0.6145, "num_input_tokens_seen": 9308224, "step": 14235 }, { "epoch": 7.463312368972747, "grad_norm": 0.14196446537971497, "learning_rate": 3.947211068483553e-05, "loss": 0.4461, "num_input_tokens_seen": 9311008, "step": 14240 }, { "epoch": 7.465932914046122, "grad_norm": 0.12552309036254883, "learning_rate": 3.946278554573013e-05, "loss": 0.4128, "num_input_tokens_seen": 9314272, "step": 14245 }, { "epoch": 7.468553459119497, "grad_norm": 0.1948448270559311, "learning_rate": 3.9453457381175e-05, "loss": 0.7637, "num_input_tokens_seen": 9316896, "step": 14250 }, { "epoch": 7.471174004192872, "grad_norm": 0.14961789548397064, "learning_rate": 3.9444126193121474e-05, "loss": 0.3535, "num_input_tokens_seen": 9320896, "step": 14255 }, { "epoch": 7.473794549266247, "grad_norm": 0.14087741076946259, "learning_rate": 3.943479198352153e-05, "loss": 0.5048, "num_input_tokens_seen": 9324832, "step": 14260 }, { "epoch": 7.476415094339623, "grad_norm": 0.20270265638828278, "learning_rate": 3.94254547543278e-05, "loss": 0.4344, "num_input_tokens_seen": 9328000, "step": 14265 }, { "epoch": 7.479035639412998, "grad_norm": 0.1844024360179901, "learning_rate": 3.9416114507493504e-05, "loss": 0.4666, "num_input_tokens_seen": 9330400, "step": 14270 }, { "epoch": 7.481656184486373, "grad_norm": 0.19031068682670593, "learning_rate": 3.940677124497252e-05, "loss": 0.4826, "num_input_tokens_seen": 9332992, "step": 14275 }, { "epoch": 7.484276729559748, "grad_norm": 0.17428340017795563, "learning_rate": 3.939742496871935e-05, "loss": 0.575, "num_input_tokens_seen": 9335968, "step": 14280 }, { "epoch": 7.486897274633124, "grad_norm": 0.15819412469863892, "learning_rate": 3.9388075680689136e-05, "loss": 0.3653, "num_input_tokens_seen": 9339456, "step": 14285 }, { "epoch": 7.489517819706499, "grad_norm": 0.16450068354606628, "learning_rate": 3.9378723382837634e-05, "loss": 0.4687, "num_input_tokens_seen": 9343520, "step": 14290 }, { "epoch": 7.4921383647798745, "grad_norm": 0.18025974929332733, "learning_rate": 3.936936807712124e-05, "loss": 0.542, "num_input_tokens_seen": 9346816, "step": 14295 }, { "epoch": 7.49475890985325, "grad_norm": 0.19138485193252563, "learning_rate": 3.936000976549698e-05, "loss": 0.4265, "num_input_tokens_seen": 9349856, "step": 14300 }, { "epoch": 7.497379454926625, "grad_norm": 0.33475059270858765, "learning_rate": 3.935064844992251e-05, "loss": 0.5527, "num_input_tokens_seen": 9353504, "step": 14305 }, { "epoch": 7.5, "grad_norm": 0.14006252586841583, "learning_rate": 3.93412841323561e-05, "loss": 0.5012, "num_input_tokens_seen": 9357504, "step": 14310 }, { "epoch": 7.502620545073375, "grad_norm": 0.2508019804954529, "learning_rate": 3.933191681475666e-05, "loss": 0.5713, "num_input_tokens_seen": 9361504, "step": 14315 }, { "epoch": 7.50524109014675, "grad_norm": 0.14001214504241943, "learning_rate": 3.932254649908373e-05, "loss": 0.4124, "num_input_tokens_seen": 9365760, "step": 14320 }, { "epoch": 7.5078616352201255, "grad_norm": 0.19536103308200836, "learning_rate": 3.931317318729748e-05, "loss": 0.5924, "num_input_tokens_seen": 9368384, "step": 14325 }, { "epoch": 7.510482180293501, "grad_norm": 0.26403218507766724, "learning_rate": 3.930379688135869e-05, "loss": 0.5583, "num_input_tokens_seen": 9371360, "step": 14330 }, { "epoch": 7.513102725366876, "grad_norm": 0.20075629651546478, "learning_rate": 3.929441758322878e-05, "loss": 0.5842, "num_input_tokens_seen": 9374624, "step": 14335 }, { "epoch": 7.515723270440252, "grad_norm": 0.10631829500198364, "learning_rate": 3.928503529486979e-05, "loss": 0.4537, "num_input_tokens_seen": 9378048, "step": 14340 }, { "epoch": 7.518343815513627, "grad_norm": 0.14907091856002808, "learning_rate": 3.927565001824438e-05, "loss": 0.5207, "num_input_tokens_seen": 9381120, "step": 14345 }, { "epoch": 7.520964360587002, "grad_norm": 0.16807331144809723, "learning_rate": 3.926626175531586e-05, "loss": 0.3979, "num_input_tokens_seen": 9384736, "step": 14350 }, { "epoch": 7.523584905660377, "grad_norm": 0.10082192718982697, "learning_rate": 3.925687050804814e-05, "loss": 0.3907, "num_input_tokens_seen": 9388192, "step": 14355 }, { "epoch": 7.526205450733753, "grad_norm": 0.1650088131427765, "learning_rate": 3.924747627840575e-05, "loss": 0.3825, "num_input_tokens_seen": 9391168, "step": 14360 }, { "epoch": 7.528825995807128, "grad_norm": 0.26225605607032776, "learning_rate": 3.923807906835388e-05, "loss": 0.5632, "num_input_tokens_seen": 9394528, "step": 14365 }, { "epoch": 7.531446540880503, "grad_norm": 0.15630190074443817, "learning_rate": 3.92286788798583e-05, "loss": 0.4729, "num_input_tokens_seen": 9398784, "step": 14370 }, { "epoch": 7.534067085953878, "grad_norm": 0.22800154983997345, "learning_rate": 3.921927571488543e-05, "loss": 0.6022, "num_input_tokens_seen": 9402176, "step": 14375 }, { "epoch": 7.536687631027254, "grad_norm": 0.28488868474960327, "learning_rate": 3.9209869575402296e-05, "loss": 0.4051, "num_input_tokens_seen": 9405344, "step": 14380 }, { "epoch": 7.539308176100629, "grad_norm": 0.1588655710220337, "learning_rate": 3.9200460463376566e-05, "loss": 0.4138, "num_input_tokens_seen": 9408384, "step": 14385 }, { "epoch": 7.5419287211740045, "grad_norm": 0.23322904109954834, "learning_rate": 3.9191048380776515e-05, "loss": 0.4649, "num_input_tokens_seen": 9410976, "step": 14390 }, { "epoch": 7.54454926624738, "grad_norm": 0.12083316594362259, "learning_rate": 3.918163332957104e-05, "loss": 0.5201, "num_input_tokens_seen": 9413632, "step": 14395 }, { "epoch": 7.547169811320755, "grad_norm": 0.22303391993045807, "learning_rate": 3.9172215311729665e-05, "loss": 0.5631, "num_input_tokens_seen": 9416640, "step": 14400 }, { "epoch": 7.54979035639413, "grad_norm": 0.1475543975830078, "learning_rate": 3.916279432922254e-05, "loss": 0.5563, "num_input_tokens_seen": 9419584, "step": 14405 }, { "epoch": 7.552410901467505, "grad_norm": 0.1974623203277588, "learning_rate": 3.915337038402041e-05, "loss": 0.4966, "num_input_tokens_seen": 9422624, "step": 14410 }, { "epoch": 7.55503144654088, "grad_norm": 0.20912273228168488, "learning_rate": 3.914394347809467e-05, "loss": 0.3467, "num_input_tokens_seen": 9424832, "step": 14415 }, { "epoch": 7.5576519916142555, "grad_norm": 0.17755328118801117, "learning_rate": 3.9134513613417315e-05, "loss": 0.7102, "num_input_tokens_seen": 9427360, "step": 14420 }, { "epoch": 7.560272536687631, "grad_norm": 0.19529730081558228, "learning_rate": 3.912508079196097e-05, "loss": 0.5785, "num_input_tokens_seen": 9430112, "step": 14425 }, { "epoch": 7.562893081761006, "grad_norm": 0.1519443243741989, "learning_rate": 3.9115645015698866e-05, "loss": 0.4757, "num_input_tokens_seen": 9434560, "step": 14430 }, { "epoch": 7.565513626834382, "grad_norm": 0.29769203066825867, "learning_rate": 3.9106206286604865e-05, "loss": 0.5308, "num_input_tokens_seen": 9437664, "step": 14435 }, { "epoch": 7.568134171907757, "grad_norm": 0.08699484914541245, "learning_rate": 3.909676460665345e-05, "loss": 0.2973, "num_input_tokens_seen": 9441472, "step": 14440 }, { "epoch": 7.570754716981132, "grad_norm": 0.18644575774669647, "learning_rate": 3.908731997781969e-05, "loss": 0.3732, "num_input_tokens_seen": 9444288, "step": 14445 }, { "epoch": 7.573375262054507, "grad_norm": 0.16095776855945587, "learning_rate": 3.907787240207932e-05, "loss": 0.4643, "num_input_tokens_seen": 9448256, "step": 14450 }, { "epoch": 7.575995807127883, "grad_norm": 0.09810902178287506, "learning_rate": 3.9068421881408645e-05, "loss": 0.4605, "num_input_tokens_seen": 9452768, "step": 14455 }, { "epoch": 7.578616352201258, "grad_norm": 0.16133344173431396, "learning_rate": 3.9058968417784614e-05, "loss": 0.5984, "num_input_tokens_seen": 9456288, "step": 14460 }, { "epoch": 7.581236897274633, "grad_norm": 0.24664883315563202, "learning_rate": 3.9049512013184785e-05, "loss": 0.5585, "num_input_tokens_seen": 9459104, "step": 14465 }, { "epoch": 7.583857442348008, "grad_norm": 0.1743166744709015, "learning_rate": 3.9040052669587325e-05, "loss": 0.4659, "num_input_tokens_seen": 9462688, "step": 14470 }, { "epoch": 7.586477987421384, "grad_norm": 0.13827836513519287, "learning_rate": 3.9030590388971026e-05, "loss": 0.4801, "num_input_tokens_seen": 9465440, "step": 14475 }, { "epoch": 7.589098532494759, "grad_norm": 0.21682168543338776, "learning_rate": 3.902112517331529e-05, "loss": 0.4374, "num_input_tokens_seen": 9469184, "step": 14480 }, { "epoch": 7.5917190775681345, "grad_norm": 0.12434698641300201, "learning_rate": 3.901165702460012e-05, "loss": 0.6061, "num_input_tokens_seen": 9472704, "step": 14485 }, { "epoch": 7.59433962264151, "grad_norm": 0.3063805103302002, "learning_rate": 3.900218594480616e-05, "loss": 0.6289, "num_input_tokens_seen": 9476224, "step": 14490 }, { "epoch": 7.596960167714885, "grad_norm": 0.16779868304729462, "learning_rate": 3.899271193591465e-05, "loss": 0.583, "num_input_tokens_seen": 9479264, "step": 14495 }, { "epoch": 7.59958071278826, "grad_norm": 0.20922785997390747, "learning_rate": 3.8983234999907435e-05, "loss": 0.5102, "num_input_tokens_seen": 9481824, "step": 14500 }, { "epoch": 7.602201257861635, "grad_norm": 0.21904338896274567, "learning_rate": 3.897375513876699e-05, "loss": 0.3792, "num_input_tokens_seen": 9485376, "step": 14505 }, { "epoch": 7.60482180293501, "grad_norm": 0.20374232530593872, "learning_rate": 3.8964272354476396e-05, "loss": 0.4109, "num_input_tokens_seen": 9488672, "step": 14510 }, { "epoch": 7.6074423480083855, "grad_norm": 0.12173565477132797, "learning_rate": 3.895478664901933e-05, "loss": 0.6096, "num_input_tokens_seen": 9491776, "step": 14515 }, { "epoch": 7.610062893081761, "grad_norm": 0.2522502839565277, "learning_rate": 3.89452980243801e-05, "loss": 0.4652, "num_input_tokens_seen": 9494624, "step": 14520 }, { "epoch": 7.612683438155136, "grad_norm": 0.19079053401947021, "learning_rate": 3.893580648254362e-05, "loss": 0.6137, "num_input_tokens_seen": 9497472, "step": 14525 }, { "epoch": 7.615303983228512, "grad_norm": 0.27154773473739624, "learning_rate": 3.8926312025495414e-05, "loss": 0.5043, "num_input_tokens_seen": 9500928, "step": 14530 }, { "epoch": 7.617924528301887, "grad_norm": 0.13398893177509308, "learning_rate": 3.891681465522161e-05, "loss": 0.4604, "num_input_tokens_seen": 9504384, "step": 14535 }, { "epoch": 7.620545073375262, "grad_norm": 0.1590726226568222, "learning_rate": 3.890731437370896e-05, "loss": 0.5053, "num_input_tokens_seen": 9506912, "step": 14540 }, { "epoch": 7.623165618448637, "grad_norm": 0.17667892575263977, "learning_rate": 3.88978111829448e-05, "loss": 0.3758, "num_input_tokens_seen": 9509632, "step": 14545 }, { "epoch": 7.6257861635220126, "grad_norm": 0.18912196159362793, "learning_rate": 3.88883050849171e-05, "loss": 0.5037, "num_input_tokens_seen": 9513024, "step": 14550 }, { "epoch": 7.628406708595388, "grad_norm": 0.19239941239356995, "learning_rate": 3.8878796081614417e-05, "loss": 0.4763, "num_input_tokens_seen": 9515936, "step": 14555 }, { "epoch": 7.631027253668763, "grad_norm": 0.15761873126029968, "learning_rate": 3.8869284175025935e-05, "loss": 0.4896, "num_input_tokens_seen": 9519744, "step": 14560 }, { "epoch": 7.633647798742138, "grad_norm": 0.30161014199256897, "learning_rate": 3.885976936714143e-05, "loss": 0.6169, "num_input_tokens_seen": 9522464, "step": 14565 }, { "epoch": 7.636268343815514, "grad_norm": 0.2282966673374176, "learning_rate": 3.885025165995129e-05, "loss": 0.5751, "num_input_tokens_seen": 9527200, "step": 14570 }, { "epoch": 7.638888888888889, "grad_norm": 0.19888482987880707, "learning_rate": 3.884073105544652e-05, "loss": 0.6545, "num_input_tokens_seen": 9530336, "step": 14575 }, { "epoch": 7.6415094339622645, "grad_norm": 0.15237286686897278, "learning_rate": 3.8831207555618723e-05, "loss": 0.4787, "num_input_tokens_seen": 9533728, "step": 14580 }, { "epoch": 7.64412997903564, "grad_norm": 0.19075828790664673, "learning_rate": 3.882168116246009e-05, "loss": 0.4427, "num_input_tokens_seen": 9536576, "step": 14585 }, { "epoch": 7.646750524109015, "grad_norm": 0.13985557854175568, "learning_rate": 3.881215187796344e-05, "loss": 0.3875, "num_input_tokens_seen": 9540160, "step": 14590 }, { "epoch": 7.64937106918239, "grad_norm": 0.18130654096603394, "learning_rate": 3.8802619704122204e-05, "loss": 0.5132, "num_input_tokens_seen": 9542976, "step": 14595 }, { "epoch": 7.651991614255765, "grad_norm": 0.12092968821525574, "learning_rate": 3.8793084642930387e-05, "loss": 0.4542, "num_input_tokens_seen": 9546240, "step": 14600 }, { "epoch": 7.65461215932914, "grad_norm": 0.18777166306972504, "learning_rate": 3.878354669638263e-05, "loss": 0.5599, "num_input_tokens_seen": 9549024, "step": 14605 }, { "epoch": 7.6572327044025155, "grad_norm": 0.1857977658510208, "learning_rate": 3.8774005866474155e-05, "loss": 0.6184, "num_input_tokens_seen": 9551328, "step": 14610 }, { "epoch": 7.659853249475891, "grad_norm": 0.10344067960977554, "learning_rate": 3.876446215520079e-05, "loss": 0.4117, "num_input_tokens_seen": 9554880, "step": 14615 }, { "epoch": 7.662473794549266, "grad_norm": 0.3599396347999573, "learning_rate": 3.875491556455898e-05, "loss": 0.5084, "num_input_tokens_seen": 9557536, "step": 14620 }, { "epoch": 7.665094339622642, "grad_norm": 0.14534905552864075, "learning_rate": 3.8745366096545754e-05, "loss": 0.3813, "num_input_tokens_seen": 9560544, "step": 14625 }, { "epoch": 7.667714884696017, "grad_norm": 0.19706493616104126, "learning_rate": 3.873581375315876e-05, "loss": 0.6924, "num_input_tokens_seen": 9563392, "step": 14630 }, { "epoch": 7.670335429769392, "grad_norm": 0.24055913090705872, "learning_rate": 3.872625853639623e-05, "loss": 0.4344, "num_input_tokens_seen": 9565984, "step": 14635 }, { "epoch": 7.672955974842767, "grad_norm": 0.19500155746936798, "learning_rate": 3.871670044825701e-05, "loss": 0.4188, "num_input_tokens_seen": 9569504, "step": 14640 }, { "epoch": 7.6755765199161425, "grad_norm": 0.16834092140197754, "learning_rate": 3.870713949074056e-05, "loss": 0.6484, "num_input_tokens_seen": 9572832, "step": 14645 }, { "epoch": 7.678197064989518, "grad_norm": 0.1793900430202484, "learning_rate": 3.86975756658469e-05, "loss": 0.5635, "num_input_tokens_seen": 9575488, "step": 14650 }, { "epoch": 7.680817610062893, "grad_norm": 0.19281922280788422, "learning_rate": 3.868800897557667e-05, "loss": 0.3876, "num_input_tokens_seen": 9578304, "step": 14655 }, { "epoch": 7.683438155136268, "grad_norm": 0.11423531174659729, "learning_rate": 3.867843942193113e-05, "loss": 0.4127, "num_input_tokens_seen": 9581440, "step": 14660 }, { "epoch": 7.686058700209644, "grad_norm": 0.16703541576862335, "learning_rate": 3.866886700691212e-05, "loss": 0.5278, "num_input_tokens_seen": 9584032, "step": 14665 }, { "epoch": 7.688679245283019, "grad_norm": 0.241520494222641, "learning_rate": 3.8659291732522075e-05, "loss": 0.4721, "num_input_tokens_seen": 9588384, "step": 14670 }, { "epoch": 7.691299790356394, "grad_norm": 0.09585876017808914, "learning_rate": 3.8649713600764026e-05, "loss": 0.4789, "num_input_tokens_seen": 9593120, "step": 14675 }, { "epoch": 7.69392033542977, "grad_norm": 0.10603068768978119, "learning_rate": 3.864013261364162e-05, "loss": 0.5304, "num_input_tokens_seen": 9596480, "step": 14680 }, { "epoch": 7.696540880503145, "grad_norm": 0.17293252050876617, "learning_rate": 3.8630548773159084e-05, "loss": 0.4909, "num_input_tokens_seen": 9599424, "step": 14685 }, { "epoch": 7.69916142557652, "grad_norm": 0.24727879464626312, "learning_rate": 3.862096208132125e-05, "loss": 0.4126, "num_input_tokens_seen": 9603648, "step": 14690 }, { "epoch": 7.701781970649895, "grad_norm": 0.1888655424118042, "learning_rate": 3.861137254013355e-05, "loss": 0.4113, "num_input_tokens_seen": 9606208, "step": 14695 }, { "epoch": 7.70440251572327, "grad_norm": 0.15373080968856812, "learning_rate": 3.8601780151601996e-05, "loss": 0.4593, "num_input_tokens_seen": 9610176, "step": 14700 }, { "epoch": 7.7070230607966455, "grad_norm": 0.2136748731136322, "learning_rate": 3.859218491773321e-05, "loss": 0.5645, "num_input_tokens_seen": 9613888, "step": 14705 }, { "epoch": 7.709643605870021, "grad_norm": 0.15315818786621094, "learning_rate": 3.858258684053441e-05, "loss": 0.3983, "num_input_tokens_seen": 9617792, "step": 14710 }, { "epoch": 7.712264150943396, "grad_norm": 0.189673513174057, "learning_rate": 3.8572985922013405e-05, "loss": 0.4377, "num_input_tokens_seen": 9620832, "step": 14715 }, { "epoch": 7.714884696016772, "grad_norm": 0.1379184126853943, "learning_rate": 3.856338216417859e-05, "loss": 0.5246, "num_input_tokens_seen": 9623744, "step": 14720 }, { "epoch": 7.717505241090147, "grad_norm": 0.2182503640651703, "learning_rate": 3.855377556903897e-05, "loss": 0.5906, "num_input_tokens_seen": 9627136, "step": 14725 }, { "epoch": 7.720125786163522, "grad_norm": 0.5031437277793884, "learning_rate": 3.8544166138604124e-05, "loss": 0.6093, "num_input_tokens_seen": 9630592, "step": 14730 }, { "epoch": 7.722746331236897, "grad_norm": 0.2543574571609497, "learning_rate": 3.8534553874884244e-05, "loss": 0.5169, "num_input_tokens_seen": 9634240, "step": 14735 }, { "epoch": 7.7253668763102725, "grad_norm": 0.2281436026096344, "learning_rate": 3.852493877989011e-05, "loss": 0.6471, "num_input_tokens_seen": 9637120, "step": 14740 }, { "epoch": 7.727987421383648, "grad_norm": 0.37153467535972595, "learning_rate": 3.851532085563308e-05, "loss": 0.4219, "num_input_tokens_seen": 9639808, "step": 14745 }, { "epoch": 7.730607966457023, "grad_norm": 0.11455971747636795, "learning_rate": 3.850570010412511e-05, "loss": 0.6485, "num_input_tokens_seen": 9644384, "step": 14750 }, { "epoch": 7.733228511530398, "grad_norm": 0.10495541989803314, "learning_rate": 3.8496076527378766e-05, "loss": 0.4754, "num_input_tokens_seen": 9648288, "step": 14755 }, { "epoch": 7.735849056603773, "grad_norm": 0.15305660665035248, "learning_rate": 3.848645012740718e-05, "loss": 0.5092, "num_input_tokens_seen": 9652032, "step": 14760 }, { "epoch": 7.738469601677149, "grad_norm": 0.1522543728351593, "learning_rate": 3.8476820906224096e-05, "loss": 0.5495, "num_input_tokens_seen": 9655296, "step": 14765 }, { "epoch": 7.741090146750524, "grad_norm": 0.19637569785118103, "learning_rate": 3.846718886584382e-05, "loss": 0.5185, "num_input_tokens_seen": 9657856, "step": 14770 }, { "epoch": 7.7437106918239, "grad_norm": 0.15583093464374542, "learning_rate": 3.845755400828128e-05, "loss": 0.4556, "num_input_tokens_seen": 9660576, "step": 14775 }, { "epoch": 7.746331236897275, "grad_norm": 0.21469031274318695, "learning_rate": 3.844791633555198e-05, "loss": 0.6532, "num_input_tokens_seen": 9663360, "step": 14780 }, { "epoch": 7.74895178197065, "grad_norm": 0.10573449730873108, "learning_rate": 3.8438275849671996e-05, "loss": 0.4538, "num_input_tokens_seen": 9667200, "step": 14785 }, { "epoch": 7.751572327044025, "grad_norm": 0.12607154250144958, "learning_rate": 3.842863255265802e-05, "loss": 0.5568, "num_input_tokens_seen": 9671136, "step": 14790 }, { "epoch": 7.7541928721174, "grad_norm": 0.2812707722187042, "learning_rate": 3.841898644652732e-05, "loss": 0.552, "num_input_tokens_seen": 9674208, "step": 14795 }, { "epoch": 7.756813417190775, "grad_norm": 0.19826796650886536, "learning_rate": 3.840933753329774e-05, "loss": 0.5022, "num_input_tokens_seen": 9677280, "step": 14800 }, { "epoch": 7.759433962264151, "grad_norm": 0.1839977651834488, "learning_rate": 3.839968581498773e-05, "loss": 0.5179, "num_input_tokens_seen": 9680320, "step": 14805 }, { "epoch": 7.762054507337526, "grad_norm": 0.18658088147640228, "learning_rate": 3.839003129361632e-05, "loss": 0.6735, "num_input_tokens_seen": 9682848, "step": 14810 }, { "epoch": 7.764675052410902, "grad_norm": 0.17580388486385345, "learning_rate": 3.8380373971203134e-05, "loss": 0.5243, "num_input_tokens_seen": 9686112, "step": 14815 }, { "epoch": 7.767295597484277, "grad_norm": 0.16100598871707916, "learning_rate": 3.8370713849768346e-05, "loss": 0.6065, "num_input_tokens_seen": 9689088, "step": 14820 }, { "epoch": 7.769916142557652, "grad_norm": 0.638170599937439, "learning_rate": 3.8361050931332774e-05, "loss": 0.569, "num_input_tokens_seen": 9693280, "step": 14825 }, { "epoch": 7.772536687631027, "grad_norm": 0.14427688717842102, "learning_rate": 3.835138521791778e-05, "loss": 0.4571, "num_input_tokens_seen": 9695872, "step": 14830 }, { "epoch": 7.7751572327044025, "grad_norm": 0.3505948781967163, "learning_rate": 3.8341716711545314e-05, "loss": 0.6183, "num_input_tokens_seen": 9698272, "step": 14835 }, { "epoch": 7.777777777777778, "grad_norm": 0.3483152985572815, "learning_rate": 3.833204541423793e-05, "loss": 0.5249, "num_input_tokens_seen": 9701568, "step": 14840 }, { "epoch": 7.780398322851153, "grad_norm": 0.11868907511234283, "learning_rate": 3.832237132801874e-05, "loss": 0.4593, "num_input_tokens_seen": 9704832, "step": 14845 }, { "epoch": 7.783018867924528, "grad_norm": 0.1867733597755432, "learning_rate": 3.831269445491146e-05, "loss": 0.6645, "num_input_tokens_seen": 9707616, "step": 14850 }, { "epoch": 7.785639412997903, "grad_norm": 0.18543067574501038, "learning_rate": 3.8303014796940375e-05, "loss": 0.5178, "num_input_tokens_seen": 9710560, "step": 14855 }, { "epoch": 7.788259958071279, "grad_norm": 0.1813698410987854, "learning_rate": 3.829333235613036e-05, "loss": 0.6177, "num_input_tokens_seen": 9713984, "step": 14860 }, { "epoch": 7.790880503144654, "grad_norm": 0.21677492558956146, "learning_rate": 3.828364713450688e-05, "loss": 0.3627, "num_input_tokens_seen": 9716832, "step": 14865 }, { "epoch": 7.79350104821803, "grad_norm": 0.17014051973819733, "learning_rate": 3.827395913409597e-05, "loss": 0.5092, "num_input_tokens_seen": 9720544, "step": 14870 }, { "epoch": 7.796121593291405, "grad_norm": 0.14353732764720917, "learning_rate": 3.826426835692424e-05, "loss": 0.5109, "num_input_tokens_seen": 9723392, "step": 14875 }, { "epoch": 7.79874213836478, "grad_norm": 0.15379874408245087, "learning_rate": 3.825457480501891e-05, "loss": 0.4517, "num_input_tokens_seen": 9726464, "step": 14880 }, { "epoch": 7.801362683438155, "grad_norm": 0.14748471975326538, "learning_rate": 3.824487848040773e-05, "loss": 0.4244, "num_input_tokens_seen": 9729952, "step": 14885 }, { "epoch": 7.80398322851153, "grad_norm": 0.17087087035179138, "learning_rate": 3.823517938511908e-05, "loss": 0.6151, "num_input_tokens_seen": 9732832, "step": 14890 }, { "epoch": 7.806603773584905, "grad_norm": 0.14109840989112854, "learning_rate": 3.8225477521181906e-05, "loss": 0.5578, "num_input_tokens_seen": 9736736, "step": 14895 }, { "epoch": 7.809224318658281, "grad_norm": 0.19836542010307312, "learning_rate": 3.821577289062571e-05, "loss": 0.4593, "num_input_tokens_seen": 9739968, "step": 14900 }, { "epoch": 7.811844863731656, "grad_norm": 0.14650839567184448, "learning_rate": 3.82060654954806e-05, "loss": 0.3842, "num_input_tokens_seen": 9742528, "step": 14905 }, { "epoch": 7.814465408805032, "grad_norm": 0.1467342972755432, "learning_rate": 3.819635533777726e-05, "loss": 0.4833, "num_input_tokens_seen": 9746752, "step": 14910 }, { "epoch": 7.817085953878407, "grad_norm": 0.18077492713928223, "learning_rate": 3.8186642419546924e-05, "loss": 0.548, "num_input_tokens_seen": 9750176, "step": 14915 }, { "epoch": 7.819706498951782, "grad_norm": 0.12811249494552612, "learning_rate": 3.817692674282145e-05, "loss": 0.564, "num_input_tokens_seen": 9752768, "step": 14920 }, { "epoch": 7.822327044025157, "grad_norm": 0.1396983116865158, "learning_rate": 3.816720830963322e-05, "loss": 0.4833, "num_input_tokens_seen": 9756352, "step": 14925 }, { "epoch": 7.8249475890985325, "grad_norm": 0.19879652559757233, "learning_rate": 3.815748712201523e-05, "loss": 0.4657, "num_input_tokens_seen": 9759264, "step": 14930 }, { "epoch": 7.827568134171908, "grad_norm": 0.24796327948570251, "learning_rate": 3.814776318200105e-05, "loss": 0.5503, "num_input_tokens_seen": 9766976, "step": 14935 }, { "epoch": 7.830188679245283, "grad_norm": 0.21440835297107697, "learning_rate": 3.8138036491624805e-05, "loss": 0.6727, "num_input_tokens_seen": 9769984, "step": 14940 }, { "epoch": 7.832809224318658, "grad_norm": 0.1508156806230545, "learning_rate": 3.812830705292122e-05, "loss": 0.6252, "num_input_tokens_seen": 9772864, "step": 14945 }, { "epoch": 7.835429769392033, "grad_norm": 0.17989102005958557, "learning_rate": 3.8118574867925564e-05, "loss": 0.4051, "num_input_tokens_seen": 9775264, "step": 14950 }, { "epoch": 7.838050314465409, "grad_norm": 0.11035338044166565, "learning_rate": 3.810883993867371e-05, "loss": 0.409, "num_input_tokens_seen": 9779584, "step": 14955 }, { "epoch": 7.840670859538784, "grad_norm": 0.10961179435253143, "learning_rate": 3.80991022672021e-05, "loss": 0.4928, "num_input_tokens_seen": 9782784, "step": 14960 }, { "epoch": 7.84329140461216, "grad_norm": 0.19629192352294922, "learning_rate": 3.8089361855547744e-05, "loss": 0.478, "num_input_tokens_seen": 9784960, "step": 14965 }, { "epoch": 7.845911949685535, "grad_norm": 0.24085868895053864, "learning_rate": 3.8079618705748214e-05, "loss": 0.5624, "num_input_tokens_seen": 9787904, "step": 14970 }, { "epoch": 7.84853249475891, "grad_norm": 0.2750745415687561, "learning_rate": 3.806987281984167e-05, "loss": 0.8525, "num_input_tokens_seen": 9790752, "step": 14975 }, { "epoch": 7.851153039832285, "grad_norm": 0.12574899196624756, "learning_rate": 3.806012419986684e-05, "loss": 0.5259, "num_input_tokens_seen": 9794624, "step": 14980 }, { "epoch": 7.85377358490566, "grad_norm": 0.13224323093891144, "learning_rate": 3.805037284786303e-05, "loss": 0.4457, "num_input_tokens_seen": 9797920, "step": 14985 }, { "epoch": 7.856394129979035, "grad_norm": 0.18236219882965088, "learning_rate": 3.804061876587009e-05, "loss": 0.6345, "num_input_tokens_seen": 9800896, "step": 14990 }, { "epoch": 7.859014675052411, "grad_norm": 0.2052154392004013, "learning_rate": 3.80308619559285e-05, "loss": 0.3471, "num_input_tokens_seen": 9803872, "step": 14995 }, { "epoch": 7.861635220125786, "grad_norm": 0.19307129085063934, "learning_rate": 3.802110242007925e-05, "loss": 0.4925, "num_input_tokens_seen": 9807136, "step": 15000 }, { "epoch": 7.864255765199162, "grad_norm": 0.15395231544971466, "learning_rate": 3.801134016036392e-05, "loss": 0.5322, "num_input_tokens_seen": 9810880, "step": 15005 }, { "epoch": 7.866876310272537, "grad_norm": 0.1418667584657669, "learning_rate": 3.8001575178824665e-05, "loss": 0.3565, "num_input_tokens_seen": 9814752, "step": 15010 }, { "epoch": 7.869496855345912, "grad_norm": 0.18538957834243774, "learning_rate": 3.799180747750422e-05, "loss": 0.541, "num_input_tokens_seen": 9818592, "step": 15015 }, { "epoch": 7.872117400419287, "grad_norm": 0.18429040908813477, "learning_rate": 3.798203705844586e-05, "loss": 0.5789, "num_input_tokens_seen": 9821152, "step": 15020 }, { "epoch": 7.8747379454926625, "grad_norm": 0.12736979126930237, "learning_rate": 3.7972263923693476e-05, "loss": 0.6398, "num_input_tokens_seen": 9824896, "step": 15025 }, { "epoch": 7.877358490566038, "grad_norm": 0.2241128534078598, "learning_rate": 3.796248807529145e-05, "loss": 0.4874, "num_input_tokens_seen": 9828000, "step": 15030 }, { "epoch": 7.879979035639413, "grad_norm": 0.2536473870277405, "learning_rate": 3.795270951528482e-05, "loss": 0.5907, "num_input_tokens_seen": 9830496, "step": 15035 }, { "epoch": 7.882599580712788, "grad_norm": 0.14025546610355377, "learning_rate": 3.7942928245719124e-05, "loss": 0.5197, "num_input_tokens_seen": 9833856, "step": 15040 }, { "epoch": 7.885220125786163, "grad_norm": 0.3653185963630676, "learning_rate": 3.7933144268640495e-05, "loss": 0.5755, "num_input_tokens_seen": 9836544, "step": 15045 }, { "epoch": 7.887840670859539, "grad_norm": 0.1880209594964981, "learning_rate": 3.7923357586095645e-05, "loss": 0.5175, "num_input_tokens_seen": 9839264, "step": 15050 }, { "epoch": 7.890461215932914, "grad_norm": 0.11952689290046692, "learning_rate": 3.7913568200131823e-05, "loss": 0.4883, "num_input_tokens_seen": 9845216, "step": 15055 }, { "epoch": 7.8930817610062896, "grad_norm": 0.15430906414985657, "learning_rate": 3.790377611279685e-05, "loss": 0.706, "num_input_tokens_seen": 9848256, "step": 15060 }, { "epoch": 7.895702306079665, "grad_norm": 0.11727526783943176, "learning_rate": 3.789398132613914e-05, "loss": 0.4404, "num_input_tokens_seen": 9851008, "step": 15065 }, { "epoch": 7.89832285115304, "grad_norm": 0.21370992064476013, "learning_rate": 3.788418384220763e-05, "loss": 0.5435, "num_input_tokens_seen": 9853920, "step": 15070 }, { "epoch": 7.900943396226415, "grad_norm": 0.14095529913902283, "learning_rate": 3.7874383663051846e-05, "loss": 0.5194, "num_input_tokens_seen": 9856736, "step": 15075 }, { "epoch": 7.90356394129979, "grad_norm": 0.21707381308078766, "learning_rate": 3.7864580790721894e-05, "loss": 0.4255, "num_input_tokens_seen": 9860896, "step": 15080 }, { "epoch": 7.906184486373165, "grad_norm": 0.180961474776268, "learning_rate": 3.785477522726839e-05, "loss": 0.4754, "num_input_tokens_seen": 9867296, "step": 15085 }, { "epoch": 7.908805031446541, "grad_norm": 0.1285092532634735, "learning_rate": 3.784496697474258e-05, "loss": 0.669, "num_input_tokens_seen": 9870816, "step": 15090 }, { "epoch": 7.911425576519916, "grad_norm": 0.12499154359102249, "learning_rate": 3.783515603519622e-05, "loss": 0.5063, "num_input_tokens_seen": 9873792, "step": 15095 }, { "epoch": 7.914046121593291, "grad_norm": 0.15342538058757782, "learning_rate": 3.782534241068164e-05, "loss": 0.4594, "num_input_tokens_seen": 9876544, "step": 15100 }, { "epoch": 7.916666666666667, "grad_norm": 0.15824450552463531, "learning_rate": 3.781552610325175e-05, "loss": 0.4551, "num_input_tokens_seen": 9880032, "step": 15105 }, { "epoch": 7.919287211740042, "grad_norm": 0.26280057430267334, "learning_rate": 3.780570711496001e-05, "loss": 0.514, "num_input_tokens_seen": 9883296, "step": 15110 }, { "epoch": 7.921907756813417, "grad_norm": 0.12820838391780853, "learning_rate": 3.779588544786044e-05, "loss": 0.4036, "num_input_tokens_seen": 9885824, "step": 15115 }, { "epoch": 7.9245283018867925, "grad_norm": 0.15991336107254028, "learning_rate": 3.778606110400761e-05, "loss": 0.5569, "num_input_tokens_seen": 9888896, "step": 15120 }, { "epoch": 7.927148846960168, "grad_norm": 0.143759623169899, "learning_rate": 3.7776234085456665e-05, "loss": 0.4309, "num_input_tokens_seen": 9891552, "step": 15125 }, { "epoch": 7.929769392033543, "grad_norm": 0.3107534945011139, "learning_rate": 3.776640439426332e-05, "loss": 0.4895, "num_input_tokens_seen": 9894400, "step": 15130 }, { "epoch": 7.932389937106918, "grad_norm": 0.1936374008655548, "learning_rate": 3.775657203248381e-05, "loss": 0.5017, "num_input_tokens_seen": 9897728, "step": 15135 }, { "epoch": 7.935010482180293, "grad_norm": 0.1682140827178955, "learning_rate": 3.774673700217497e-05, "loss": 0.4368, "num_input_tokens_seen": 9901248, "step": 15140 }, { "epoch": 7.937631027253669, "grad_norm": 0.14776547253131866, "learning_rate": 3.773689930539418e-05, "loss": 0.7915, "num_input_tokens_seen": 9904384, "step": 15145 }, { "epoch": 7.940251572327044, "grad_norm": 0.20092269778251648, "learning_rate": 3.7727058944199354e-05, "loss": 0.4372, "num_input_tokens_seen": 9907872, "step": 15150 }, { "epoch": 7.9428721174004195, "grad_norm": 0.134165421128273, "learning_rate": 3.7717215920648994e-05, "loss": 0.3709, "num_input_tokens_seen": 9911040, "step": 15155 }, { "epoch": 7.945492662473795, "grad_norm": 0.1535399854183197, "learning_rate": 3.770737023680215e-05, "loss": 0.4728, "num_input_tokens_seen": 9914208, "step": 15160 }, { "epoch": 7.94811320754717, "grad_norm": 0.15341323614120483, "learning_rate": 3.7697521894718436e-05, "loss": 0.5855, "num_input_tokens_seen": 9917856, "step": 15165 }, { "epoch": 7.950733752620545, "grad_norm": 0.15200738608837128, "learning_rate": 3.7687670896458e-05, "loss": 0.532, "num_input_tokens_seen": 9921280, "step": 15170 }, { "epoch": 7.95335429769392, "grad_norm": 0.13770191371440887, "learning_rate": 3.7677817244081556e-05, "loss": 0.4398, "num_input_tokens_seen": 9925376, "step": 15175 }, { "epoch": 7.955974842767295, "grad_norm": 0.11327125132083893, "learning_rate": 3.766796093965038e-05, "loss": 0.5682, "num_input_tokens_seen": 9928256, "step": 15180 }, { "epoch": 7.9585953878406706, "grad_norm": 0.21814455091953278, "learning_rate": 3.76581019852263e-05, "loss": 0.5905, "num_input_tokens_seen": 9931232, "step": 15185 }, { "epoch": 7.961215932914046, "grad_norm": 0.20447789132595062, "learning_rate": 3.7648240382871705e-05, "loss": 0.4562, "num_input_tokens_seen": 9934368, "step": 15190 }, { "epoch": 7.963836477987421, "grad_norm": 0.16620570421218872, "learning_rate": 3.763837613464951e-05, "loss": 0.4304, "num_input_tokens_seen": 9938432, "step": 15195 }, { "epoch": 7.966457023060797, "grad_norm": 0.25074824690818787, "learning_rate": 3.7628509242623215e-05, "loss": 0.3576, "num_input_tokens_seen": 9941024, "step": 15200 }, { "epoch": 7.969077568134172, "grad_norm": 0.16603504121303558, "learning_rate": 3.761863970885686e-05, "loss": 0.5063, "num_input_tokens_seen": 9944480, "step": 15205 }, { "epoch": 7.971698113207547, "grad_norm": 0.1709037870168686, "learning_rate": 3.760876753541504e-05, "loss": 0.4211, "num_input_tokens_seen": 9948192, "step": 15210 }, { "epoch": 7.9743186582809225, "grad_norm": 0.192095547914505, "learning_rate": 3.7598892724362907e-05, "loss": 0.5638, "num_input_tokens_seen": 9951168, "step": 15215 }, { "epoch": 7.976939203354298, "grad_norm": 0.22490732371807098, "learning_rate": 3.758901527776615e-05, "loss": 0.5788, "num_input_tokens_seen": 9955136, "step": 15220 }, { "epoch": 7.979559748427673, "grad_norm": 0.16162614524364471, "learning_rate": 3.7579135197691015e-05, "loss": 0.5001, "num_input_tokens_seen": 9958464, "step": 15225 }, { "epoch": 7.982180293501048, "grad_norm": 0.17420479655265808, "learning_rate": 3.7569252486204306e-05, "loss": 0.5431, "num_input_tokens_seen": 9961312, "step": 15230 }, { "epoch": 7.984800838574423, "grad_norm": 0.15660981833934784, "learning_rate": 3.755936714537338e-05, "loss": 0.4722, "num_input_tokens_seen": 9965280, "step": 15235 }, { "epoch": 7.987421383647799, "grad_norm": 0.17520256340503693, "learning_rate": 3.754947917726612e-05, "loss": 0.6605, "num_input_tokens_seen": 9969504, "step": 15240 }, { "epoch": 7.990041928721174, "grad_norm": 0.16268223524093628, "learning_rate": 3.7539588583951e-05, "loss": 0.3952, "num_input_tokens_seen": 9972960, "step": 15245 }, { "epoch": 7.9926624737945495, "grad_norm": 0.17927180230617523, "learning_rate": 3.752969536749701e-05, "loss": 0.5653, "num_input_tokens_seen": 9975712, "step": 15250 }, { "epoch": 7.995283018867925, "grad_norm": 0.13578401505947113, "learning_rate": 3.7519799529973685e-05, "loss": 0.5642, "num_input_tokens_seen": 9978752, "step": 15255 }, { "epoch": 7.9979035639413, "grad_norm": 0.15895937383174896, "learning_rate": 3.7509901073451134e-05, "loss": 0.668, "num_input_tokens_seen": 9981472, "step": 15260 }, { "epoch": 8.0, "eval_loss": 0.4691309332847595, "eval_runtime": 9.8429, "eval_samples_per_second": 86.154, "eval_steps_per_second": 21.538, "num_input_tokens_seen": 9983720, "step": 15264 }, { "epoch": 8.000524109014675, "grad_norm": 0.15816393494606018, "learning_rate": 3.7500000000000003e-05, "loss": 0.4025, "num_input_tokens_seen": 9984424, "step": 15265 }, { "epoch": 8.00314465408805, "grad_norm": 0.2640935480594635, "learning_rate": 3.749009631169147e-05, "loss": 0.5211, "num_input_tokens_seen": 9987528, "step": 15270 }, { "epoch": 8.005765199161425, "grad_norm": 0.1375468373298645, "learning_rate": 3.74801900105973e-05, "loss": 0.4858, "num_input_tokens_seen": 9990536, "step": 15275 }, { "epoch": 8.0083857442348, "grad_norm": 0.1618860363960266, "learning_rate": 3.747028109878975e-05, "loss": 0.4105, "num_input_tokens_seen": 9994504, "step": 15280 }, { "epoch": 8.011006289308176, "grad_norm": 0.18196429312229156, "learning_rate": 3.746036957834167e-05, "loss": 0.5002, "num_input_tokens_seen": 9997480, "step": 15285 }, { "epoch": 8.01362683438155, "grad_norm": 0.15480266511440277, "learning_rate": 3.7450455451326415e-05, "loss": 0.4334, "num_input_tokens_seen": 10000840, "step": 15290 }, { "epoch": 8.016247379454926, "grad_norm": 0.15899477899074554, "learning_rate": 3.7440538719817935e-05, "loss": 0.496, "num_input_tokens_seen": 10003304, "step": 15295 }, { "epoch": 8.018867924528301, "grad_norm": 0.17616455256938934, "learning_rate": 3.743061938589068e-05, "loss": 0.5457, "num_input_tokens_seen": 10006280, "step": 15300 }, { "epoch": 8.021488469601676, "grad_norm": 0.2238156944513321, "learning_rate": 3.742069745161966e-05, "loss": 0.5262, "num_input_tokens_seen": 10009800, "step": 15305 }, { "epoch": 8.024109014675052, "grad_norm": 0.2070647031068802, "learning_rate": 3.741077291908045e-05, "loss": 0.5477, "num_input_tokens_seen": 10012840, "step": 15310 }, { "epoch": 8.026729559748428, "grad_norm": 0.1411733478307724, "learning_rate": 3.740084579034913e-05, "loss": 0.4256, "num_input_tokens_seen": 10015656, "step": 15315 }, { "epoch": 8.029350104821804, "grad_norm": 0.22518350183963776, "learning_rate": 3.739091606750234e-05, "loss": 0.4858, "num_input_tokens_seen": 10018888, "step": 15320 }, { "epoch": 8.031970649895179, "grad_norm": 0.11873965710401535, "learning_rate": 3.7380983752617275e-05, "loss": 0.438, "num_input_tokens_seen": 10022280, "step": 15325 }, { "epoch": 8.034591194968554, "grad_norm": 0.24414432048797607, "learning_rate": 3.737104884777166e-05, "loss": 0.4292, "num_input_tokens_seen": 10025416, "step": 15330 }, { "epoch": 8.03721174004193, "grad_norm": 0.38468584418296814, "learning_rate": 3.736111135504376e-05, "loss": 0.5797, "num_input_tokens_seen": 10027944, "step": 15335 }, { "epoch": 8.039832285115304, "grad_norm": 0.1059034913778305, "learning_rate": 3.735117127651239e-05, "loss": 0.4556, "num_input_tokens_seen": 10030824, "step": 15340 }, { "epoch": 8.04245283018868, "grad_norm": 0.20211413502693176, "learning_rate": 3.734122861425689e-05, "loss": 0.3955, "num_input_tokens_seen": 10033128, "step": 15345 }, { "epoch": 8.045073375262055, "grad_norm": 0.16089044511318207, "learning_rate": 3.7331283370357173e-05, "loss": 0.4356, "num_input_tokens_seen": 10036488, "step": 15350 }, { "epoch": 8.04769392033543, "grad_norm": 0.20236001908779144, "learning_rate": 3.732133554689365e-05, "loss": 0.4407, "num_input_tokens_seen": 10039368, "step": 15355 }, { "epoch": 8.050314465408805, "grad_norm": 0.23129890859127045, "learning_rate": 3.73113851459473e-05, "loss": 0.5169, "num_input_tokens_seen": 10042472, "step": 15360 }, { "epoch": 8.05293501048218, "grad_norm": 0.18028169870376587, "learning_rate": 3.730143216959964e-05, "loss": 0.4127, "num_input_tokens_seen": 10045640, "step": 15365 }, { "epoch": 8.055555555555555, "grad_norm": 0.12740831077098846, "learning_rate": 3.729147661993271e-05, "loss": 0.4785, "num_input_tokens_seen": 10049864, "step": 15370 }, { "epoch": 8.05817610062893, "grad_norm": 0.1289902776479721, "learning_rate": 3.728151849902909e-05, "loss": 0.6227, "num_input_tokens_seen": 10053480, "step": 15375 }, { "epoch": 8.060796645702306, "grad_norm": 0.12689611315727234, "learning_rate": 3.727155780897193e-05, "loss": 0.5326, "num_input_tokens_seen": 10056872, "step": 15380 }, { "epoch": 8.06341719077568, "grad_norm": 0.10689475387334824, "learning_rate": 3.7261594551844865e-05, "loss": 0.5633, "num_input_tokens_seen": 10060264, "step": 15385 }, { "epoch": 8.066037735849056, "grad_norm": 0.18371650576591492, "learning_rate": 3.7251628729732116e-05, "loss": 0.3551, "num_input_tokens_seen": 10063240, "step": 15390 }, { "epoch": 8.068658280922431, "grad_norm": 0.1908610463142395, "learning_rate": 3.724166034471841e-05, "loss": 0.4515, "num_input_tokens_seen": 10065768, "step": 15395 }, { "epoch": 8.071278825995806, "grad_norm": 0.22910672426223755, "learning_rate": 3.723168939888901e-05, "loss": 0.5149, "num_input_tokens_seen": 10068872, "step": 15400 }, { "epoch": 8.073899371069182, "grad_norm": 0.15672726929187775, "learning_rate": 3.7221715894329747e-05, "loss": 0.5419, "num_input_tokens_seen": 10071496, "step": 15405 }, { "epoch": 8.076519916142558, "grad_norm": 0.1507989913225174, "learning_rate": 3.7211739833126945e-05, "loss": 0.6689, "num_input_tokens_seen": 10074632, "step": 15410 }, { "epoch": 8.079140461215934, "grad_norm": 0.24906474351882935, "learning_rate": 3.720176121736748e-05, "loss": 0.5122, "num_input_tokens_seen": 10077064, "step": 15415 }, { "epoch": 8.081761006289309, "grad_norm": 0.1492246687412262, "learning_rate": 3.719178004913879e-05, "loss": 0.4737, "num_input_tokens_seen": 10081288, "step": 15420 }, { "epoch": 8.084381551362684, "grad_norm": 0.22772864997386932, "learning_rate": 3.7181796330528796e-05, "loss": 0.4998, "num_input_tokens_seen": 10084392, "step": 15425 }, { "epoch": 8.08700209643606, "grad_norm": 0.15223637223243713, "learning_rate": 3.717181006362599e-05, "loss": 0.3479, "num_input_tokens_seen": 10088936, "step": 15430 }, { "epoch": 8.089622641509434, "grad_norm": 0.15120792388916016, "learning_rate": 3.7161821250519384e-05, "loss": 0.4761, "num_input_tokens_seen": 10091432, "step": 15435 }, { "epoch": 8.09224318658281, "grad_norm": 0.2429463267326355, "learning_rate": 3.7151829893298515e-05, "loss": 0.545, "num_input_tokens_seen": 10093704, "step": 15440 }, { "epoch": 8.094863731656185, "grad_norm": 0.12126114219427109, "learning_rate": 3.714183599405347e-05, "loss": 0.5114, "num_input_tokens_seen": 10098120, "step": 15445 }, { "epoch": 8.09748427672956, "grad_norm": 0.1640469878911972, "learning_rate": 3.713183955487487e-05, "loss": 0.5506, "num_input_tokens_seen": 10100872, "step": 15450 }, { "epoch": 8.100104821802935, "grad_norm": 0.14487016201019287, "learning_rate": 3.7121840577853815e-05, "loss": 0.502, "num_input_tokens_seen": 10104456, "step": 15455 }, { "epoch": 8.10272536687631, "grad_norm": 0.1891932338476181, "learning_rate": 3.7111839065082027e-05, "loss": 0.4816, "num_input_tokens_seen": 10107368, "step": 15460 }, { "epoch": 8.105345911949685, "grad_norm": 0.23308758437633514, "learning_rate": 3.710183501865168e-05, "loss": 0.4966, "num_input_tokens_seen": 10110344, "step": 15465 }, { "epoch": 8.10796645702306, "grad_norm": 0.14626823365688324, "learning_rate": 3.7091828440655514e-05, "loss": 0.5739, "num_input_tokens_seen": 10113224, "step": 15470 }, { "epoch": 8.110587002096436, "grad_norm": 0.19006715714931488, "learning_rate": 3.708181933318678e-05, "loss": 0.6748, "num_input_tokens_seen": 10117160, "step": 15475 }, { "epoch": 8.11320754716981, "grad_norm": 0.14605481922626495, "learning_rate": 3.707180769833929e-05, "loss": 0.5144, "num_input_tokens_seen": 10121064, "step": 15480 }, { "epoch": 8.115828092243186, "grad_norm": 0.19310541450977325, "learning_rate": 3.706179353820734e-05, "loss": 0.5268, "num_input_tokens_seen": 10124360, "step": 15485 }, { "epoch": 8.118448637316561, "grad_norm": 0.16694825887680054, "learning_rate": 3.7051776854885794e-05, "loss": 0.5825, "num_input_tokens_seen": 10127400, "step": 15490 }, { "epoch": 8.121069182389936, "grad_norm": 0.23392222821712494, "learning_rate": 3.7041757650470025e-05, "loss": 0.4965, "num_input_tokens_seen": 10129960, "step": 15495 }, { "epoch": 8.123689727463312, "grad_norm": 0.1471666842699051, "learning_rate": 3.703173592705594e-05, "loss": 0.4689, "num_input_tokens_seen": 10132936, "step": 15500 }, { "epoch": 8.126310272536688, "grad_norm": 0.09658295661211014, "learning_rate": 3.702171168673996e-05, "loss": 0.4647, "num_input_tokens_seen": 10136328, "step": 15505 }, { "epoch": 8.128930817610064, "grad_norm": 0.1958586871623993, "learning_rate": 3.701168493161905e-05, "loss": 0.5186, "num_input_tokens_seen": 10139688, "step": 15510 }, { "epoch": 8.131551362683439, "grad_norm": 0.13354116678237915, "learning_rate": 3.7001655663790676e-05, "loss": 0.4425, "num_input_tokens_seen": 10143016, "step": 15515 }, { "epoch": 8.134171907756814, "grad_norm": 0.08736623078584671, "learning_rate": 3.699162388535288e-05, "loss": 0.4667, "num_input_tokens_seen": 10147336, "step": 15520 }, { "epoch": 8.13679245283019, "grad_norm": 0.19851666688919067, "learning_rate": 3.698158959840416e-05, "loss": 0.4157, "num_input_tokens_seen": 10150504, "step": 15525 }, { "epoch": 8.139412997903564, "grad_norm": 0.20052160322666168, "learning_rate": 3.6971552805043585e-05, "loss": 0.648, "num_input_tokens_seen": 10153000, "step": 15530 }, { "epoch": 8.14203354297694, "grad_norm": 0.34189286828041077, "learning_rate": 3.6961513507370754e-05, "loss": 0.5049, "num_input_tokens_seen": 10155784, "step": 15535 }, { "epoch": 8.144654088050315, "grad_norm": 0.12456417828798294, "learning_rate": 3.695147170748576e-05, "loss": 0.97, "num_input_tokens_seen": 10159592, "step": 15540 }, { "epoch": 8.14727463312369, "grad_norm": 0.1821412444114685, "learning_rate": 3.694142740748923e-05, "loss": 0.4354, "num_input_tokens_seen": 10162600, "step": 15545 }, { "epoch": 8.149895178197065, "grad_norm": 0.27384546399116516, "learning_rate": 3.693138060948232e-05, "loss": 0.6612, "num_input_tokens_seen": 10164968, "step": 15550 }, { "epoch": 8.15251572327044, "grad_norm": 0.17603358626365662, "learning_rate": 3.692133131556671e-05, "loss": 0.6813, "num_input_tokens_seen": 10171912, "step": 15555 }, { "epoch": 8.155136268343815, "grad_norm": 0.23008984327316284, "learning_rate": 3.6911279527844596e-05, "loss": 0.6118, "num_input_tokens_seen": 10175112, "step": 15560 }, { "epoch": 8.15775681341719, "grad_norm": 0.2883392572402954, "learning_rate": 3.69012252484187e-05, "loss": 0.6068, "num_input_tokens_seen": 10178280, "step": 15565 }, { "epoch": 8.160377358490566, "grad_norm": 0.12276671081781387, "learning_rate": 3.6891168479392245e-05, "loss": 0.5178, "num_input_tokens_seen": 10181352, "step": 15570 }, { "epoch": 8.16299790356394, "grad_norm": 0.20428310334682465, "learning_rate": 3.688110922286901e-05, "loss": 0.524, "num_input_tokens_seen": 10185256, "step": 15575 }, { "epoch": 8.165618448637316, "grad_norm": 0.13603055477142334, "learning_rate": 3.687104748095328e-05, "loss": 0.5992, "num_input_tokens_seen": 10188872, "step": 15580 }, { "epoch": 8.168238993710691, "grad_norm": 0.17756468057632446, "learning_rate": 3.686098325574984e-05, "loss": 0.4834, "num_input_tokens_seen": 10192008, "step": 15585 }, { "epoch": 8.170859538784066, "grad_norm": 0.12172134965658188, "learning_rate": 3.685091654936403e-05, "loss": 0.3909, "num_input_tokens_seen": 10196808, "step": 15590 }, { "epoch": 8.173480083857442, "grad_norm": 0.17566367983818054, "learning_rate": 3.6840847363901675e-05, "loss": 0.6241, "num_input_tokens_seen": 10199848, "step": 15595 }, { "epoch": 8.176100628930818, "grad_norm": 0.1306140273809433, "learning_rate": 3.683077570146914e-05, "loss": 0.3352, "num_input_tokens_seen": 10203048, "step": 15600 }, { "epoch": 8.178721174004194, "grad_norm": 0.1918335109949112, "learning_rate": 3.68207015641733e-05, "loss": 0.4663, "num_input_tokens_seen": 10207144, "step": 15605 }, { "epoch": 8.181341719077569, "grad_norm": 0.22192621231079102, "learning_rate": 3.681062495412155e-05, "loss": 0.5449, "num_input_tokens_seen": 10209672, "step": 15610 }, { "epoch": 8.183962264150944, "grad_norm": 0.20159737765789032, "learning_rate": 3.68005458734218e-05, "loss": 0.4473, "num_input_tokens_seen": 10212520, "step": 15615 }, { "epoch": 8.18658280922432, "grad_norm": 0.1814924031496048, "learning_rate": 3.679046432418248e-05, "loss": 0.445, "num_input_tokens_seen": 10215048, "step": 15620 }, { "epoch": 8.189203354297694, "grad_norm": 0.16856320202350616, "learning_rate": 3.6780380308512543e-05, "loss": 0.5125, "num_input_tokens_seen": 10217352, "step": 15625 }, { "epoch": 8.19182389937107, "grad_norm": 0.16920651495456696, "learning_rate": 3.677029382852144e-05, "loss": 0.4767, "num_input_tokens_seen": 10221096, "step": 15630 }, { "epoch": 8.194444444444445, "grad_norm": 0.15483424067497253, "learning_rate": 3.676020488631915e-05, "loss": 0.4201, "num_input_tokens_seen": 10229608, "step": 15635 }, { "epoch": 8.19706498951782, "grad_norm": 0.14369195699691772, "learning_rate": 3.675011348401617e-05, "loss": 0.4145, "num_input_tokens_seen": 10233000, "step": 15640 }, { "epoch": 8.199685534591195, "grad_norm": 0.1503220945596695, "learning_rate": 3.6740019623723494e-05, "loss": 0.5533, "num_input_tokens_seen": 10236328, "step": 15645 }, { "epoch": 8.20230607966457, "grad_norm": 0.14924730360507965, "learning_rate": 3.672992330755265e-05, "loss": 0.52, "num_input_tokens_seen": 10239624, "step": 15650 }, { "epoch": 8.204926624737945, "grad_norm": 0.3101271688938141, "learning_rate": 3.671982453761568e-05, "loss": 0.609, "num_input_tokens_seen": 10242440, "step": 15655 }, { "epoch": 8.20754716981132, "grad_norm": 0.13908642530441284, "learning_rate": 3.6709723316025114e-05, "loss": 0.4439, "num_input_tokens_seen": 10245608, "step": 15660 }, { "epoch": 8.210167714884696, "grad_norm": 0.20700623095035553, "learning_rate": 3.669961964489403e-05, "loss": 0.517, "num_input_tokens_seen": 10247848, "step": 15665 }, { "epoch": 8.21278825995807, "grad_norm": 0.11780767142772675, "learning_rate": 3.668951352633599e-05, "loss": 0.5389, "num_input_tokens_seen": 10252616, "step": 15670 }, { "epoch": 8.215408805031446, "grad_norm": 0.3143024146556854, "learning_rate": 3.6679404962465084e-05, "loss": 0.6639, "num_input_tokens_seen": 10255880, "step": 15675 }, { "epoch": 8.218029350104821, "grad_norm": 0.12855030596256256, "learning_rate": 3.66692939553959e-05, "loss": 0.4609, "num_input_tokens_seen": 10258632, "step": 15680 }, { "epoch": 8.220649895178196, "grad_norm": 0.1413043737411499, "learning_rate": 3.6659180507243554e-05, "loss": 0.4745, "num_input_tokens_seen": 10262504, "step": 15685 }, { "epoch": 8.223270440251572, "grad_norm": 0.11268781870603561, "learning_rate": 3.664906462012366e-05, "loss": 0.46, "num_input_tokens_seen": 10265480, "step": 15690 }, { "epoch": 8.225890985324948, "grad_norm": 0.26892146468162537, "learning_rate": 3.663894629615234e-05, "loss": 0.5001, "num_input_tokens_seen": 10269000, "step": 15695 }, { "epoch": 8.228511530398324, "grad_norm": 0.13680273294448853, "learning_rate": 3.6628825537446246e-05, "loss": 0.3931, "num_input_tokens_seen": 10272072, "step": 15700 }, { "epoch": 8.231132075471699, "grad_norm": 0.1738513559103012, "learning_rate": 3.6618702346122525e-05, "loss": 0.5991, "num_input_tokens_seen": 10276008, "step": 15705 }, { "epoch": 8.233752620545074, "grad_norm": 0.20534496009349823, "learning_rate": 3.6608576724298817e-05, "loss": 0.475, "num_input_tokens_seen": 10279304, "step": 15710 }, { "epoch": 8.23637316561845, "grad_norm": 0.16839838027954102, "learning_rate": 3.65984486740933e-05, "loss": 0.5947, "num_input_tokens_seen": 10282856, "step": 15715 }, { "epoch": 8.238993710691824, "grad_norm": 0.15565000474452972, "learning_rate": 3.6588318197624644e-05, "loss": 0.4695, "num_input_tokens_seen": 10286376, "step": 15720 }, { "epoch": 8.2416142557652, "grad_norm": 0.17061999440193176, "learning_rate": 3.6578185297012015e-05, "loss": 0.5619, "num_input_tokens_seen": 10288904, "step": 15725 }, { "epoch": 8.244234800838575, "grad_norm": 0.1559523046016693, "learning_rate": 3.656804997437512e-05, "loss": 0.5622, "num_input_tokens_seen": 10294824, "step": 15730 }, { "epoch": 8.24685534591195, "grad_norm": 0.20682546496391296, "learning_rate": 3.6557912231834155e-05, "loss": 0.5034, "num_input_tokens_seen": 10299400, "step": 15735 }, { "epoch": 8.249475890985325, "grad_norm": 0.16375084221363068, "learning_rate": 3.65477720715098e-05, "loss": 0.3363, "num_input_tokens_seen": 10302312, "step": 15740 }, { "epoch": 8.2520964360587, "grad_norm": 0.3165960907936096, "learning_rate": 3.653762949552328e-05, "loss": 0.4506, "num_input_tokens_seen": 10305000, "step": 15745 }, { "epoch": 8.254716981132075, "grad_norm": 0.20339910686016083, "learning_rate": 3.6527484505996295e-05, "loss": 0.4935, "num_input_tokens_seen": 10308296, "step": 15750 }, { "epoch": 8.25733752620545, "grad_norm": 0.16837388277053833, "learning_rate": 3.651733710505106e-05, "loss": 0.4444, "num_input_tokens_seen": 10311464, "step": 15755 }, { "epoch": 8.259958071278826, "grad_norm": 0.15571175515651703, "learning_rate": 3.6507187294810305e-05, "loss": 0.4233, "num_input_tokens_seen": 10314920, "step": 15760 }, { "epoch": 8.2625786163522, "grad_norm": 0.1544743925333023, "learning_rate": 3.649703507739724e-05, "loss": 0.4134, "num_input_tokens_seen": 10317896, "step": 15765 }, { "epoch": 8.265199161425576, "grad_norm": 0.15698468685150146, "learning_rate": 3.6486880454935606e-05, "loss": 0.4877, "num_input_tokens_seen": 10320136, "step": 15770 }, { "epoch": 8.267819706498951, "grad_norm": 0.13586188852787018, "learning_rate": 3.647672342954963e-05, "loss": 0.3705, "num_input_tokens_seen": 10323496, "step": 15775 }, { "epoch": 8.270440251572326, "grad_norm": 0.1175263300538063, "learning_rate": 3.646656400336405e-05, "loss": 0.4175, "num_input_tokens_seen": 10327784, "step": 15780 }, { "epoch": 8.273060796645701, "grad_norm": 0.14834479987621307, "learning_rate": 3.6456402178504094e-05, "loss": 0.5098, "num_input_tokens_seen": 10330952, "step": 15785 }, { "epoch": 8.275681341719078, "grad_norm": 0.14919410645961761, "learning_rate": 3.6446237957095505e-05, "loss": 0.4252, "num_input_tokens_seen": 10334568, "step": 15790 }, { "epoch": 8.278301886792454, "grad_norm": 0.15029704570770264, "learning_rate": 3.643607134126452e-05, "loss": 0.4102, "num_input_tokens_seen": 10338184, "step": 15795 }, { "epoch": 8.280922431865829, "grad_norm": 0.15077577531337738, "learning_rate": 3.642590233313789e-05, "loss": 0.6195, "num_input_tokens_seen": 10341352, "step": 15800 }, { "epoch": 8.283542976939204, "grad_norm": 0.28504037857055664, "learning_rate": 3.6415730934842827e-05, "loss": 0.5952, "num_input_tokens_seen": 10344136, "step": 15805 }, { "epoch": 8.286163522012579, "grad_norm": 0.3558720052242279, "learning_rate": 3.64055571485071e-05, "loss": 0.535, "num_input_tokens_seen": 10346472, "step": 15810 }, { "epoch": 8.288784067085954, "grad_norm": 0.23742170631885529, "learning_rate": 3.639538097625894e-05, "loss": 0.4886, "num_input_tokens_seen": 10349928, "step": 15815 }, { "epoch": 8.29140461215933, "grad_norm": 0.12203843891620636, "learning_rate": 3.638520242022708e-05, "loss": 0.5324, "num_input_tokens_seen": 10352936, "step": 15820 }, { "epoch": 8.294025157232705, "grad_norm": 0.20210278034210205, "learning_rate": 3.637502148254077e-05, "loss": 0.4366, "num_input_tokens_seen": 10356040, "step": 15825 }, { "epoch": 8.29664570230608, "grad_norm": 0.33290305733680725, "learning_rate": 3.636483816532974e-05, "loss": 0.4192, "num_input_tokens_seen": 10359048, "step": 15830 }, { "epoch": 8.299266247379455, "grad_norm": 0.14838410913944244, "learning_rate": 3.6354652470724217e-05, "loss": 0.51, "num_input_tokens_seen": 10361832, "step": 15835 }, { "epoch": 8.30188679245283, "grad_norm": 0.1842259019613266, "learning_rate": 3.634446440085494e-05, "loss": 0.6061, "num_input_tokens_seen": 10364904, "step": 15840 }, { "epoch": 8.304507337526205, "grad_norm": 0.24394655227661133, "learning_rate": 3.633427395785313e-05, "loss": 0.6426, "num_input_tokens_seen": 10367816, "step": 15845 }, { "epoch": 8.30712788259958, "grad_norm": 0.16706646978855133, "learning_rate": 3.6324081143850516e-05, "loss": 0.5021, "num_input_tokens_seen": 10371080, "step": 15850 }, { "epoch": 8.309748427672956, "grad_norm": 0.15788643062114716, "learning_rate": 3.6313885960979326e-05, "loss": 0.419, "num_input_tokens_seen": 10373928, "step": 15855 }, { "epoch": 8.31236897274633, "grad_norm": 0.43387049436569214, "learning_rate": 3.630368841137226e-05, "loss": 0.5697, "num_input_tokens_seen": 10377384, "step": 15860 }, { "epoch": 8.314989517819706, "grad_norm": 0.27383342385292053, "learning_rate": 3.6293488497162554e-05, "loss": 0.6754, "num_input_tokens_seen": 10380392, "step": 15865 }, { "epoch": 8.317610062893081, "grad_norm": 0.18931466341018677, "learning_rate": 3.628328622048388e-05, "loss": 0.6079, "num_input_tokens_seen": 10382824, "step": 15870 }, { "epoch": 8.320230607966456, "grad_norm": 0.18277089297771454, "learning_rate": 3.627308158347046e-05, "loss": 0.4688, "num_input_tokens_seen": 10385192, "step": 15875 }, { "epoch": 8.322851153039831, "grad_norm": 0.14314119517803192, "learning_rate": 3.626287458825699e-05, "loss": 0.4707, "num_input_tokens_seen": 10389256, "step": 15880 }, { "epoch": 8.325471698113208, "grad_norm": 0.25444355607032776, "learning_rate": 3.625266523697864e-05, "loss": 0.497, "num_input_tokens_seen": 10392168, "step": 15885 }, { "epoch": 8.328092243186584, "grad_norm": 0.16986456513404846, "learning_rate": 3.62424535317711e-05, "loss": 0.3932, "num_input_tokens_seen": 10395240, "step": 15890 }, { "epoch": 8.330712788259959, "grad_norm": 0.15699222683906555, "learning_rate": 3.623223947477055e-05, "loss": 0.3697, "num_input_tokens_seen": 10398056, "step": 15895 }, { "epoch": 8.333333333333334, "grad_norm": 0.12251102179288864, "learning_rate": 3.6222023068113635e-05, "loss": 0.4852, "num_input_tokens_seen": 10401608, "step": 15900 }, { "epoch": 8.335953878406709, "grad_norm": 0.1663074791431427, "learning_rate": 3.621180431393754e-05, "loss": 0.4249, "num_input_tokens_seen": 10404424, "step": 15905 }, { "epoch": 8.338574423480084, "grad_norm": 0.2079453021287918, "learning_rate": 3.6201583214379874e-05, "loss": 0.3729, "num_input_tokens_seen": 10407016, "step": 15910 }, { "epoch": 8.34119496855346, "grad_norm": 0.21067269146442413, "learning_rate": 3.61913597715788e-05, "loss": 0.5731, "num_input_tokens_seen": 10410280, "step": 15915 }, { "epoch": 8.343815513626835, "grad_norm": 0.19084860384464264, "learning_rate": 3.6181133987672944e-05, "loss": 0.5741, "num_input_tokens_seen": 10413000, "step": 15920 }, { "epoch": 8.34643605870021, "grad_norm": 0.3138010501861572, "learning_rate": 3.617090586480142e-05, "loss": 0.4679, "num_input_tokens_seen": 10415336, "step": 15925 }, { "epoch": 8.349056603773585, "grad_norm": 0.1420566290616989, "learning_rate": 3.6160675405103826e-05, "loss": 0.4891, "num_input_tokens_seen": 10418536, "step": 15930 }, { "epoch": 8.35167714884696, "grad_norm": 0.2514569163322449, "learning_rate": 3.615044261072027e-05, "loss": 0.4405, "num_input_tokens_seen": 10421320, "step": 15935 }, { "epoch": 8.354297693920335, "grad_norm": 0.15664836764335632, "learning_rate": 3.614020748379133e-05, "loss": 0.4886, "num_input_tokens_seen": 10425224, "step": 15940 }, { "epoch": 8.35691823899371, "grad_norm": 0.19173215329647064, "learning_rate": 3.6129970026458076e-05, "loss": 0.4271, "num_input_tokens_seen": 10427912, "step": 15945 }, { "epoch": 8.359538784067086, "grad_norm": 0.19561663269996643, "learning_rate": 3.611973024086207e-05, "loss": 0.4116, "num_input_tokens_seen": 10430600, "step": 15950 }, { "epoch": 8.36215932914046, "grad_norm": 0.371480792760849, "learning_rate": 3.610948812914536e-05, "loss": 0.5109, "num_input_tokens_seen": 10433992, "step": 15955 }, { "epoch": 8.364779874213836, "grad_norm": 0.1697743535041809, "learning_rate": 3.609924369345048e-05, "loss": 0.3683, "num_input_tokens_seen": 10437832, "step": 15960 }, { "epoch": 8.367400419287211, "grad_norm": 0.18910092115402222, "learning_rate": 3.6088996935920444e-05, "loss": 0.4389, "num_input_tokens_seen": 10440840, "step": 15965 }, { "epoch": 8.370020964360586, "grad_norm": 0.21875527501106262, "learning_rate": 3.607874785869876e-05, "loss": 0.5772, "num_input_tokens_seen": 10443496, "step": 15970 }, { "epoch": 8.372641509433961, "grad_norm": 0.15060719847679138, "learning_rate": 3.6068496463929416e-05, "loss": 0.5473, "num_input_tokens_seen": 10446184, "step": 15975 }, { "epoch": 8.375262054507338, "grad_norm": 0.19355137646198273, "learning_rate": 3.605824275375689e-05, "loss": 0.4712, "num_input_tokens_seen": 10449064, "step": 15980 }, { "epoch": 8.377882599580714, "grad_norm": 0.1585671603679657, "learning_rate": 3.604798673032613e-05, "loss": 0.4591, "num_input_tokens_seen": 10452808, "step": 15985 }, { "epoch": 8.380503144654089, "grad_norm": 0.16180141270160675, "learning_rate": 3.60377283957826e-05, "loss": 0.5056, "num_input_tokens_seen": 10456808, "step": 15990 }, { "epoch": 8.383123689727464, "grad_norm": 0.1531788855791092, "learning_rate": 3.6027467752272203e-05, "loss": 0.4351, "num_input_tokens_seen": 10459880, "step": 15995 }, { "epoch": 8.385744234800839, "grad_norm": 0.10454148799180984, "learning_rate": 3.601720480194137e-05, "loss": 0.5262, "num_input_tokens_seen": 10463208, "step": 16000 }, { "epoch": 8.388364779874214, "grad_norm": 0.1766737699508667, "learning_rate": 3.600693954693698e-05, "loss": 0.5018, "num_input_tokens_seen": 10467368, "step": 16005 }, { "epoch": 8.39098532494759, "grad_norm": 0.1991826444864273, "learning_rate": 3.5996671989406414e-05, "loss": 0.464, "num_input_tokens_seen": 10470344, "step": 16010 }, { "epoch": 8.393605870020965, "grad_norm": 0.18337510526180267, "learning_rate": 3.598640213149752e-05, "loss": 0.4806, "num_input_tokens_seen": 10474312, "step": 16015 }, { "epoch": 8.39622641509434, "grad_norm": 0.1688084602355957, "learning_rate": 3.597612997535864e-05, "loss": 0.4929, "num_input_tokens_seen": 10477224, "step": 16020 }, { "epoch": 8.398846960167715, "grad_norm": 0.13049010932445526, "learning_rate": 3.596585552313859e-05, "loss": 0.462, "num_input_tokens_seen": 10481800, "step": 16025 }, { "epoch": 8.40146750524109, "grad_norm": 0.1347988098859787, "learning_rate": 3.595557877698666e-05, "loss": 0.4127, "num_input_tokens_seen": 10484584, "step": 16030 }, { "epoch": 8.404088050314465, "grad_norm": 0.20112687349319458, "learning_rate": 3.594529973905266e-05, "loss": 0.4226, "num_input_tokens_seen": 10487272, "step": 16035 }, { "epoch": 8.40670859538784, "grad_norm": 0.1016208827495575, "learning_rate": 3.593501841148681e-05, "loss": 0.3662, "num_input_tokens_seen": 10490856, "step": 16040 }, { "epoch": 8.409329140461216, "grad_norm": 0.15441849827766418, "learning_rate": 3.5924734796439864e-05, "loss": 0.6129, "num_input_tokens_seen": 10493800, "step": 16045 }, { "epoch": 8.41194968553459, "grad_norm": 0.24533717334270477, "learning_rate": 3.5914448896063044e-05, "loss": 0.5555, "num_input_tokens_seen": 10497512, "step": 16050 }, { "epoch": 8.414570230607966, "grad_norm": 0.09102064371109009, "learning_rate": 3.590416071250802e-05, "loss": 0.4765, "num_input_tokens_seen": 10501096, "step": 16055 }, { "epoch": 8.417190775681341, "grad_norm": 0.13966502249240875, "learning_rate": 3.589387024792699e-05, "loss": 0.4458, "num_input_tokens_seen": 10504648, "step": 16060 }, { "epoch": 8.419811320754716, "grad_norm": 0.1478097289800644, "learning_rate": 3.5883577504472576e-05, "loss": 0.3255, "num_input_tokens_seen": 10507016, "step": 16065 }, { "epoch": 8.422431865828091, "grad_norm": 0.08803272247314453, "learning_rate": 3.587328248429792e-05, "loss": 0.3969, "num_input_tokens_seen": 10510792, "step": 16070 }, { "epoch": 8.425052410901468, "grad_norm": 0.11946910619735718, "learning_rate": 3.586298518955661e-05, "loss": 0.5503, "num_input_tokens_seen": 10514696, "step": 16075 }, { "epoch": 8.427672955974844, "grad_norm": 0.10692142695188522, "learning_rate": 3.5852685622402746e-05, "loss": 0.4234, "num_input_tokens_seen": 10518152, "step": 16080 }, { "epoch": 8.430293501048219, "grad_norm": 0.1259082704782486, "learning_rate": 3.584238378499085e-05, "loss": 0.5783, "num_input_tokens_seen": 10521128, "step": 16085 }, { "epoch": 8.432914046121594, "grad_norm": 0.23891539871692657, "learning_rate": 3.5832079679475974e-05, "loss": 0.7402, "num_input_tokens_seen": 10524456, "step": 16090 }, { "epoch": 8.435534591194969, "grad_norm": 0.12815047800540924, "learning_rate": 3.582177330801359e-05, "loss": 0.4609, "num_input_tokens_seen": 10526920, "step": 16095 }, { "epoch": 8.438155136268344, "grad_norm": 0.13911709189414978, "learning_rate": 3.58114646727597e-05, "loss": 0.4462, "num_input_tokens_seen": 10529640, "step": 16100 }, { "epoch": 8.44077568134172, "grad_norm": 0.34607332944869995, "learning_rate": 3.580115377587075e-05, "loss": 0.4569, "num_input_tokens_seen": 10532968, "step": 16105 }, { "epoch": 8.443396226415095, "grad_norm": 0.1591119021177292, "learning_rate": 3.5790840619503654e-05, "loss": 0.5207, "num_input_tokens_seen": 10535944, "step": 16110 }, { "epoch": 8.44601677148847, "grad_norm": 0.11611481755971909, "learning_rate": 3.57805252058158e-05, "loss": 0.4162, "num_input_tokens_seen": 10539304, "step": 16115 }, { "epoch": 8.448637316561845, "grad_norm": 0.16982175409793854, "learning_rate": 3.577020753696507e-05, "loss": 0.3951, "num_input_tokens_seen": 10542120, "step": 16120 }, { "epoch": 8.45125786163522, "grad_norm": 0.11310254782438278, "learning_rate": 3.57598876151098e-05, "loss": 0.4637, "num_input_tokens_seen": 10544392, "step": 16125 }, { "epoch": 8.453878406708595, "grad_norm": 0.1909025013446808, "learning_rate": 3.5749565442408796e-05, "loss": 0.5351, "num_input_tokens_seen": 10547656, "step": 16130 }, { "epoch": 8.45649895178197, "grad_norm": 0.24056893587112427, "learning_rate": 3.573924102102133e-05, "loss": 0.5382, "num_input_tokens_seen": 10550984, "step": 16135 }, { "epoch": 8.459119496855346, "grad_norm": 0.15571998059749603, "learning_rate": 3.572891435310717e-05, "loss": 0.6427, "num_input_tokens_seen": 10554952, "step": 16140 }, { "epoch": 8.46174004192872, "grad_norm": 0.15592335164546967, "learning_rate": 3.5718585440826527e-05, "loss": 0.5688, "num_input_tokens_seen": 10557608, "step": 16145 }, { "epoch": 8.464360587002096, "grad_norm": 0.16313886642456055, "learning_rate": 3.570825428634009e-05, "loss": 0.5835, "num_input_tokens_seen": 10561352, "step": 16150 }, { "epoch": 8.466981132075471, "grad_norm": 0.22921383380889893, "learning_rate": 3.569792089180903e-05, "loss": 0.5478, "num_input_tokens_seen": 10565064, "step": 16155 }, { "epoch": 8.469601677148846, "grad_norm": 0.12326426804065704, "learning_rate": 3.568758525939497e-05, "loss": 0.4238, "num_input_tokens_seen": 10567976, "step": 16160 }, { "epoch": 8.472222222222221, "grad_norm": 0.12100467085838318, "learning_rate": 3.5677247391259995e-05, "loss": 0.5612, "num_input_tokens_seen": 10571528, "step": 16165 }, { "epoch": 8.474842767295598, "grad_norm": 0.16228331625461578, "learning_rate": 3.566690728956668e-05, "loss": 0.5358, "num_input_tokens_seen": 10574024, "step": 16170 }, { "epoch": 8.477463312368974, "grad_norm": 0.17740991711616516, "learning_rate": 3.5656564956478056e-05, "loss": 0.5519, "num_input_tokens_seen": 10578024, "step": 16175 }, { "epoch": 8.480083857442349, "grad_norm": 0.11300576478242874, "learning_rate": 3.5646220394157626e-05, "loss": 0.5003, "num_input_tokens_seen": 10581448, "step": 16180 }, { "epoch": 8.482704402515724, "grad_norm": 0.32739555835723877, "learning_rate": 3.563587360476933e-05, "loss": 0.4597, "num_input_tokens_seen": 10584328, "step": 16185 }, { "epoch": 8.485324947589099, "grad_norm": 0.19020713865756989, "learning_rate": 3.562552459047763e-05, "loss": 0.6395, "num_input_tokens_seen": 10587400, "step": 16190 }, { "epoch": 8.487945492662474, "grad_norm": 0.16598302125930786, "learning_rate": 3.561517335344741e-05, "loss": 0.5393, "num_input_tokens_seen": 10591368, "step": 16195 }, { "epoch": 8.49056603773585, "grad_norm": 0.12914863228797913, "learning_rate": 3.560481989584402e-05, "loss": 0.566, "num_input_tokens_seen": 10594856, "step": 16200 }, { "epoch": 8.493186582809225, "grad_norm": 0.21664807200431824, "learning_rate": 3.55944642198333e-05, "loss": 0.5809, "num_input_tokens_seen": 10597928, "step": 16205 }, { "epoch": 8.4958071278826, "grad_norm": 0.09369345009326935, "learning_rate": 3.5584106327581536e-05, "loss": 0.5139, "num_input_tokens_seen": 10601448, "step": 16210 }, { "epoch": 8.498427672955975, "grad_norm": 0.14841638505458832, "learning_rate": 3.557374622125547e-05, "loss": 0.4836, "num_input_tokens_seen": 10604520, "step": 16215 }, { "epoch": 8.50104821802935, "grad_norm": 0.13105148077011108, "learning_rate": 3.556338390302234e-05, "loss": 0.5474, "num_input_tokens_seen": 10607944, "step": 16220 }, { "epoch": 8.503668763102725, "grad_norm": 0.178222194314003, "learning_rate": 3.55530193750498e-05, "loss": 0.5368, "num_input_tokens_seen": 10611208, "step": 16225 }, { "epoch": 8.5062893081761, "grad_norm": 0.19959279894828796, "learning_rate": 3.5542652639505996e-05, "loss": 0.5647, "num_input_tokens_seen": 10613896, "step": 16230 }, { "epoch": 8.508909853249476, "grad_norm": 0.21597599983215332, "learning_rate": 3.5532283698559556e-05, "loss": 0.4333, "num_input_tokens_seen": 10616392, "step": 16235 }, { "epoch": 8.51153039832285, "grad_norm": 0.12707409262657166, "learning_rate": 3.55219125543795e-05, "loss": 0.5941, "num_input_tokens_seen": 10620232, "step": 16240 }, { "epoch": 8.514150943396226, "grad_norm": 0.20661510527133942, "learning_rate": 3.5511539209135396e-05, "loss": 0.4238, "num_input_tokens_seen": 10627272, "step": 16245 }, { "epoch": 8.516771488469601, "grad_norm": 0.15741515159606934, "learning_rate": 3.550116366499721e-05, "loss": 0.5532, "num_input_tokens_seen": 10630344, "step": 16250 }, { "epoch": 8.519392033542976, "grad_norm": 0.21227401494979858, "learning_rate": 3.549078592413538e-05, "loss": 0.4639, "num_input_tokens_seen": 10633480, "step": 16255 }, { "epoch": 8.522012578616351, "grad_norm": 0.24165202677249908, "learning_rate": 3.548040598872083e-05, "loss": 0.5535, "num_input_tokens_seen": 10636104, "step": 16260 }, { "epoch": 8.524633123689728, "grad_norm": 0.18500185012817383, "learning_rate": 3.5470023860924905e-05, "loss": 0.4446, "num_input_tokens_seen": 10638696, "step": 16265 }, { "epoch": 8.527253668763104, "grad_norm": 0.11793069541454315, "learning_rate": 3.5459639542919444e-05, "loss": 0.5062, "num_input_tokens_seen": 10641320, "step": 16270 }, { "epoch": 8.529874213836479, "grad_norm": 0.11295095831155777, "learning_rate": 3.5449253036876715e-05, "loss": 0.498, "num_input_tokens_seen": 10644360, "step": 16275 }, { "epoch": 8.532494758909854, "grad_norm": 0.18700554966926575, "learning_rate": 3.543886434496947e-05, "loss": 0.5897, "num_input_tokens_seen": 10647464, "step": 16280 }, { "epoch": 8.535115303983229, "grad_norm": 0.28969576954841614, "learning_rate": 3.542847346937089e-05, "loss": 0.4812, "num_input_tokens_seen": 10650920, "step": 16285 }, { "epoch": 8.537735849056604, "grad_norm": 0.15284641087055206, "learning_rate": 3.541808041225464e-05, "loss": 0.4978, "num_input_tokens_seen": 10654984, "step": 16290 }, { "epoch": 8.54035639412998, "grad_norm": 0.1693265736103058, "learning_rate": 3.540768517579481e-05, "loss": 0.4098, "num_input_tokens_seen": 10658216, "step": 16295 }, { "epoch": 8.542976939203355, "grad_norm": 0.18140016496181488, "learning_rate": 3.539728776216599e-05, "loss": 0.5983, "num_input_tokens_seen": 10660904, "step": 16300 }, { "epoch": 8.54559748427673, "grad_norm": 0.15214790403842926, "learning_rate": 3.5386888173543185e-05, "loss": 0.4935, "num_input_tokens_seen": 10664040, "step": 16305 }, { "epoch": 8.548218029350105, "grad_norm": 0.17718946933746338, "learning_rate": 3.5376486412101875e-05, "loss": 0.6005, "num_input_tokens_seen": 10668008, "step": 16310 }, { "epoch": 8.55083857442348, "grad_norm": 0.12290962040424347, "learning_rate": 3.5366082480017994e-05, "loss": 0.4365, "num_input_tokens_seen": 10673032, "step": 16315 }, { "epoch": 8.553459119496855, "grad_norm": 0.252908855676651, "learning_rate": 3.535567637946792e-05, "loss": 0.6749, "num_input_tokens_seen": 10676200, "step": 16320 }, { "epoch": 8.55607966457023, "grad_norm": 0.11855337023735046, "learning_rate": 3.534526811262848e-05, "loss": 0.5001, "num_input_tokens_seen": 10680168, "step": 16325 }, { "epoch": 8.558700209643606, "grad_norm": 0.17345285415649414, "learning_rate": 3.533485768167698e-05, "loss": 0.5597, "num_input_tokens_seen": 10683304, "step": 16330 }, { "epoch": 8.56132075471698, "grad_norm": 0.11531002819538116, "learning_rate": 3.532444508879117e-05, "loss": 0.5239, "num_input_tokens_seen": 10686856, "step": 16335 }, { "epoch": 8.563941299790356, "grad_norm": 0.1522371470928192, "learning_rate": 3.531403033614922e-05, "loss": 0.5301, "num_input_tokens_seen": 10691048, "step": 16340 }, { "epoch": 8.566561844863731, "grad_norm": 0.1834995597600937, "learning_rate": 3.530361342592981e-05, "loss": 0.5338, "num_input_tokens_seen": 10694504, "step": 16345 }, { "epoch": 8.569182389937106, "grad_norm": 0.18522822856903076, "learning_rate": 3.5293194360312006e-05, "loss": 0.4468, "num_input_tokens_seen": 10697576, "step": 16350 }, { "epoch": 8.571802935010481, "grad_norm": 0.23582595586776733, "learning_rate": 3.528277314147538e-05, "loss": 0.5047, "num_input_tokens_seen": 10700552, "step": 16355 }, { "epoch": 8.574423480083858, "grad_norm": 0.27517685294151306, "learning_rate": 3.5272349771599925e-05, "loss": 0.569, "num_input_tokens_seen": 10703240, "step": 16360 }, { "epoch": 8.577044025157234, "grad_norm": 0.7019517421722412, "learning_rate": 3.526192425286609e-05, "loss": 0.5154, "num_input_tokens_seen": 10705960, "step": 16365 }, { "epoch": 8.579664570230609, "grad_norm": 0.2014821618795395, "learning_rate": 3.525149658745476e-05, "loss": 0.4389, "num_input_tokens_seen": 10709672, "step": 16370 }, { "epoch": 8.582285115303984, "grad_norm": 0.1422538161277771, "learning_rate": 3.5241066777547315e-05, "loss": 0.5871, "num_input_tokens_seen": 10713096, "step": 16375 }, { "epoch": 8.584905660377359, "grad_norm": 0.16003714501857758, "learning_rate": 3.523063482532554e-05, "loss": 0.5566, "num_input_tokens_seen": 10716552, "step": 16380 }, { "epoch": 8.587526205450734, "grad_norm": 0.24212180078029633, "learning_rate": 3.5220200732971665e-05, "loss": 0.4437, "num_input_tokens_seen": 10719336, "step": 16385 }, { "epoch": 8.59014675052411, "grad_norm": 0.13385675847530365, "learning_rate": 3.52097645026684e-05, "loss": 0.3681, "num_input_tokens_seen": 10723528, "step": 16390 }, { "epoch": 8.592767295597485, "grad_norm": 0.18047313392162323, "learning_rate": 3.519932613659889e-05, "loss": 0.4374, "num_input_tokens_seen": 10726248, "step": 16395 }, { "epoch": 8.59538784067086, "grad_norm": 0.17800851166248322, "learning_rate": 3.5188885636946696e-05, "loss": 0.4551, "num_input_tokens_seen": 10729000, "step": 16400 }, { "epoch": 8.598008385744235, "grad_norm": 0.0993211418390274, "learning_rate": 3.5178443005895884e-05, "loss": 0.3838, "num_input_tokens_seen": 10732328, "step": 16405 }, { "epoch": 8.60062893081761, "grad_norm": 0.18055815994739532, "learning_rate": 3.51679982456309e-05, "loss": 0.4515, "num_input_tokens_seen": 10735112, "step": 16410 }, { "epoch": 8.603249475890985, "grad_norm": 0.12091477960348129, "learning_rate": 3.51575513583367e-05, "loss": 0.6212, "num_input_tokens_seen": 10739016, "step": 16415 }, { "epoch": 8.60587002096436, "grad_norm": 0.2571846842765808, "learning_rate": 3.5147102346198635e-05, "loss": 0.438, "num_input_tokens_seen": 10741992, "step": 16420 }, { "epoch": 8.608490566037736, "grad_norm": 0.12259909510612488, "learning_rate": 3.5136651211402515e-05, "loss": 0.4595, "num_input_tokens_seen": 10746120, "step": 16425 }, { "epoch": 8.61111111111111, "grad_norm": 0.2496204674243927, "learning_rate": 3.5126197956134614e-05, "loss": 0.4612, "num_input_tokens_seen": 10748584, "step": 16430 }, { "epoch": 8.613731656184486, "grad_norm": 0.15875357389450073, "learning_rate": 3.5115742582581626e-05, "loss": 0.4892, "num_input_tokens_seen": 10751400, "step": 16435 }, { "epoch": 8.616352201257861, "grad_norm": 0.19282831251621246, "learning_rate": 3.51052850929307e-05, "loss": 0.43, "num_input_tokens_seen": 10754024, "step": 16440 }, { "epoch": 8.618972746331236, "grad_norm": 0.11557531356811523, "learning_rate": 3.509482548936941e-05, "loss": 0.412, "num_input_tokens_seen": 10757608, "step": 16445 }, { "epoch": 8.621593291404611, "grad_norm": 0.15481412410736084, "learning_rate": 3.5084363774085804e-05, "loss": 0.6331, "num_input_tokens_seen": 10760232, "step": 16450 }, { "epoch": 8.624213836477988, "grad_norm": 0.17183847725391388, "learning_rate": 3.507389994926834e-05, "loss": 0.5272, "num_input_tokens_seen": 10763656, "step": 16455 }, { "epoch": 8.626834381551364, "grad_norm": 0.13367505371570587, "learning_rate": 3.5063434017105946e-05, "loss": 0.4044, "num_input_tokens_seen": 10768328, "step": 16460 }, { "epoch": 8.629454926624739, "grad_norm": 0.1429091840982437, "learning_rate": 3.505296597978795e-05, "loss": 0.6029, "num_input_tokens_seen": 10771112, "step": 16465 }, { "epoch": 8.632075471698114, "grad_norm": 0.13577070832252502, "learning_rate": 3.504249583950417e-05, "loss": 0.5838, "num_input_tokens_seen": 10774312, "step": 16470 }, { "epoch": 8.634696016771489, "grad_norm": 0.0907442569732666, "learning_rate": 3.503202359844483e-05, "loss": 0.4143, "num_input_tokens_seen": 10777160, "step": 16475 }, { "epoch": 8.637316561844864, "grad_norm": 0.1367044299840927, "learning_rate": 3.50215492588006e-05, "loss": 0.4022, "num_input_tokens_seen": 10781096, "step": 16480 }, { "epoch": 8.63993710691824, "grad_norm": 0.14516723155975342, "learning_rate": 3.5011072822762605e-05, "loss": 0.4905, "num_input_tokens_seen": 10784264, "step": 16485 }, { "epoch": 8.642557651991615, "grad_norm": 0.21043767035007477, "learning_rate": 3.500059429252238e-05, "loss": 0.4439, "num_input_tokens_seen": 10786856, "step": 16490 }, { "epoch": 8.64517819706499, "grad_norm": 0.13736112415790558, "learning_rate": 3.499011367027193e-05, "loss": 0.5037, "num_input_tokens_seen": 10789992, "step": 16495 }, { "epoch": 8.647798742138365, "grad_norm": 0.18684208393096924, "learning_rate": 3.4979630958203665e-05, "loss": 0.483, "num_input_tokens_seen": 10793064, "step": 16500 }, { "epoch": 8.65041928721174, "grad_norm": 0.15535737574100494, "learning_rate": 3.496914615851046e-05, "loss": 0.5502, "num_input_tokens_seen": 10795784, "step": 16505 }, { "epoch": 8.653039832285115, "grad_norm": 0.32583096623420715, "learning_rate": 3.495865927338562e-05, "loss": 0.519, "num_input_tokens_seen": 10798824, "step": 16510 }, { "epoch": 8.65566037735849, "grad_norm": 0.12302853912115097, "learning_rate": 3.494817030502286e-05, "loss": 0.5648, "num_input_tokens_seen": 10801960, "step": 16515 }, { "epoch": 8.658280922431866, "grad_norm": 0.1613321453332901, "learning_rate": 3.493767925561638e-05, "loss": 0.4577, "num_input_tokens_seen": 10805320, "step": 16520 }, { "epoch": 8.66090146750524, "grad_norm": 0.12103135138750076, "learning_rate": 3.492718612736077e-05, "loss": 0.3837, "num_input_tokens_seen": 10808520, "step": 16525 }, { "epoch": 8.663522012578616, "grad_norm": 0.197218656539917, "learning_rate": 3.491669092245108e-05, "loss": 0.542, "num_input_tokens_seen": 10811176, "step": 16530 }, { "epoch": 8.666142557651991, "grad_norm": 0.34042811393737793, "learning_rate": 3.4906193643082784e-05, "loss": 0.4961, "num_input_tokens_seen": 10815240, "step": 16535 }, { "epoch": 8.668763102725366, "grad_norm": 0.24823476374149323, "learning_rate": 3.4895694291451795e-05, "loss": 0.4422, "num_input_tokens_seen": 10818152, "step": 16540 }, { "epoch": 8.671383647798741, "grad_norm": 0.22349265217781067, "learning_rate": 3.4885192869754444e-05, "loss": 0.5298, "num_input_tokens_seen": 10821128, "step": 16545 }, { "epoch": 8.674004192872118, "grad_norm": 0.13720561563968658, "learning_rate": 3.4874689380187537e-05, "loss": 0.4787, "num_input_tokens_seen": 10824040, "step": 16550 }, { "epoch": 8.676624737945493, "grad_norm": 0.11451242864131927, "learning_rate": 3.486418382494826e-05, "loss": 0.5041, "num_input_tokens_seen": 10827304, "step": 16555 }, { "epoch": 8.679245283018869, "grad_norm": 0.2660675644874573, "learning_rate": 3.485367620623426e-05, "loss": 0.4862, "num_input_tokens_seen": 10831144, "step": 16560 }, { "epoch": 8.681865828092244, "grad_norm": 0.20562900602817535, "learning_rate": 3.484316652624362e-05, "loss": 0.4636, "num_input_tokens_seen": 10834696, "step": 16565 }, { "epoch": 8.684486373165619, "grad_norm": 0.18208293616771698, "learning_rate": 3.483265478717484e-05, "loss": 0.4032, "num_input_tokens_seen": 10837320, "step": 16570 }, { "epoch": 8.687106918238994, "grad_norm": 0.21014785766601562, "learning_rate": 3.482214099122686e-05, "loss": 0.7887, "num_input_tokens_seen": 10840008, "step": 16575 }, { "epoch": 8.68972746331237, "grad_norm": 0.21747104823589325, "learning_rate": 3.481162514059903e-05, "loss": 0.4577, "num_input_tokens_seen": 10844328, "step": 16580 }, { "epoch": 8.692348008385745, "grad_norm": 0.23862610757350922, "learning_rate": 3.480110723749116e-05, "loss": 0.6589, "num_input_tokens_seen": 10847400, "step": 16585 }, { "epoch": 8.69496855345912, "grad_norm": 0.13159437477588654, "learning_rate": 3.479058728410348e-05, "loss": 0.4253, "num_input_tokens_seen": 10850664, "step": 16590 }, { "epoch": 8.697589098532495, "grad_norm": 0.15726326406002045, "learning_rate": 3.478006528263663e-05, "loss": 0.5713, "num_input_tokens_seen": 10853864, "step": 16595 }, { "epoch": 8.70020964360587, "grad_norm": 0.29117727279663086, "learning_rate": 3.4769541235291704e-05, "loss": 0.53, "num_input_tokens_seen": 10856936, "step": 16600 }, { "epoch": 8.702830188679245, "grad_norm": 0.1436883509159088, "learning_rate": 3.475901514427021e-05, "loss": 0.5323, "num_input_tokens_seen": 10860488, "step": 16605 }, { "epoch": 8.70545073375262, "grad_norm": 0.20849394798278809, "learning_rate": 3.4748487011774075e-05, "loss": 0.4116, "num_input_tokens_seen": 10863592, "step": 16610 }, { "epoch": 8.708071278825996, "grad_norm": 0.11379147320985794, "learning_rate": 3.473795684000569e-05, "loss": 0.4992, "num_input_tokens_seen": 10866856, "step": 16615 }, { "epoch": 8.71069182389937, "grad_norm": 0.13103099167346954, "learning_rate": 3.4727424631167815e-05, "loss": 0.4415, "num_input_tokens_seen": 10869928, "step": 16620 }, { "epoch": 8.713312368972746, "grad_norm": 0.28902214765548706, "learning_rate": 3.471689038746369e-05, "loss": 0.8086, "num_input_tokens_seen": 10872456, "step": 16625 }, { "epoch": 8.715932914046121, "grad_norm": 0.14203138649463654, "learning_rate": 3.470635411109695e-05, "loss": 0.4277, "num_input_tokens_seen": 10876744, "step": 16630 }, { "epoch": 8.718553459119496, "grad_norm": 0.19026689231395721, "learning_rate": 3.469581580427167e-05, "loss": 0.3978, "num_input_tokens_seen": 10880072, "step": 16635 }, { "epoch": 8.721174004192871, "grad_norm": 0.2777983844280243, "learning_rate": 3.4685275469192335e-05, "loss": 0.5742, "num_input_tokens_seen": 10883112, "step": 16640 }, { "epoch": 8.723794549266248, "grad_norm": 0.1372445970773697, "learning_rate": 3.467473310806387e-05, "loss": 0.4159, "num_input_tokens_seen": 10886280, "step": 16645 }, { "epoch": 8.726415094339622, "grad_norm": 0.17772679030895233, "learning_rate": 3.466418872309161e-05, "loss": 0.5309, "num_input_tokens_seen": 10890280, "step": 16650 }, { "epoch": 8.729035639412999, "grad_norm": 0.2287197709083557, "learning_rate": 3.465364231648133e-05, "loss": 0.4311, "num_input_tokens_seen": 10893608, "step": 16655 }, { "epoch": 8.731656184486374, "grad_norm": 0.19974610209465027, "learning_rate": 3.464309389043921e-05, "loss": 0.509, "num_input_tokens_seen": 10896904, "step": 16660 }, { "epoch": 8.734276729559749, "grad_norm": 0.13167545199394226, "learning_rate": 3.4632543447171856e-05, "loss": 0.4693, "num_input_tokens_seen": 10899752, "step": 16665 }, { "epoch": 8.736897274633124, "grad_norm": 0.1572115272283554, "learning_rate": 3.4621990988886315e-05, "loss": 0.5351, "num_input_tokens_seen": 10903592, "step": 16670 }, { "epoch": 8.7395178197065, "grad_norm": 0.18377357721328735, "learning_rate": 3.461143651779003e-05, "loss": 0.4907, "num_input_tokens_seen": 10907752, "step": 16675 }, { "epoch": 8.742138364779874, "grad_norm": 0.253554105758667, "learning_rate": 3.4600880036090876e-05, "loss": 0.4124, "num_input_tokens_seen": 10910600, "step": 16680 }, { "epoch": 8.74475890985325, "grad_norm": 0.19288955628871918, "learning_rate": 3.459032154599715e-05, "loss": 0.4042, "num_input_tokens_seen": 10913800, "step": 16685 }, { "epoch": 8.747379454926625, "grad_norm": 0.15666654706001282, "learning_rate": 3.457976104971757e-05, "loss": 0.5233, "num_input_tokens_seen": 10916968, "step": 16690 }, { "epoch": 8.75, "grad_norm": 0.14639106392860413, "learning_rate": 3.4569198549461276e-05, "loss": 0.404, "num_input_tokens_seen": 10919944, "step": 16695 }, { "epoch": 8.752620545073375, "grad_norm": 0.18220800161361694, "learning_rate": 3.455863404743781e-05, "loss": 0.3892, "num_input_tokens_seen": 10923144, "step": 16700 }, { "epoch": 8.75524109014675, "grad_norm": 0.43260374665260315, "learning_rate": 3.454806754585716e-05, "loss": 0.7575, "num_input_tokens_seen": 10926088, "step": 16705 }, { "epoch": 8.757861635220126, "grad_norm": 0.14325669407844543, "learning_rate": 3.453749904692971e-05, "loss": 0.4445, "num_input_tokens_seen": 10929416, "step": 16710 }, { "epoch": 8.7604821802935, "grad_norm": 0.11442424356937408, "learning_rate": 3.452692855286626e-05, "loss": 0.5134, "num_input_tokens_seen": 10934088, "step": 16715 }, { "epoch": 8.763102725366876, "grad_norm": 0.14053696393966675, "learning_rate": 3.4516356065878055e-05, "loss": 0.5659, "num_input_tokens_seen": 10937256, "step": 16720 }, { "epoch": 8.765723270440251, "grad_norm": 0.3521753251552582, "learning_rate": 3.450578158817671e-05, "loss": 0.4963, "num_input_tokens_seen": 10940072, "step": 16725 }, { "epoch": 8.768343815513626, "grad_norm": 0.12869830429553986, "learning_rate": 3.4495205121974325e-05, "loss": 0.5819, "num_input_tokens_seen": 10943176, "step": 16730 }, { "epoch": 8.770964360587001, "grad_norm": 0.1782270222902298, "learning_rate": 3.448462666948335e-05, "loss": 0.5693, "num_input_tokens_seen": 10946312, "step": 16735 }, { "epoch": 8.773584905660378, "grad_norm": 0.19043920934200287, "learning_rate": 3.4474046232916674e-05, "loss": 0.3534, "num_input_tokens_seen": 10950792, "step": 16740 }, { "epoch": 8.776205450733752, "grad_norm": 0.13090845942497253, "learning_rate": 3.4463463814487616e-05, "loss": 0.4863, "num_input_tokens_seen": 10952936, "step": 16745 }, { "epoch": 8.778825995807129, "grad_norm": 0.15257728099822998, "learning_rate": 3.44528794164099e-05, "loss": 0.5565, "num_input_tokens_seen": 10957512, "step": 16750 }, { "epoch": 8.781446540880504, "grad_norm": 0.3877179026603699, "learning_rate": 3.4442293040897646e-05, "loss": 0.481, "num_input_tokens_seen": 10960328, "step": 16755 }, { "epoch": 8.784067085953879, "grad_norm": 0.13785621523857117, "learning_rate": 3.443170469016542e-05, "loss": 0.4345, "num_input_tokens_seen": 10963944, "step": 16760 }, { "epoch": 8.786687631027254, "grad_norm": 0.1599189043045044, "learning_rate": 3.442111436642817e-05, "loss": 0.4036, "num_input_tokens_seen": 10967272, "step": 16765 }, { "epoch": 8.78930817610063, "grad_norm": 0.2359660565853119, "learning_rate": 3.441052207190127e-05, "loss": 0.4413, "num_input_tokens_seen": 10970440, "step": 16770 }, { "epoch": 8.791928721174004, "grad_norm": 0.11551322042942047, "learning_rate": 3.4399927808800526e-05, "loss": 0.4041, "num_input_tokens_seen": 10974568, "step": 16775 }, { "epoch": 8.79454926624738, "grad_norm": 0.25131121277809143, "learning_rate": 3.438933157934211e-05, "loss": 0.5887, "num_input_tokens_seen": 10977096, "step": 16780 }, { "epoch": 8.797169811320755, "grad_norm": 0.13031768798828125, "learning_rate": 3.4378733385742653e-05, "loss": 0.3219, "num_input_tokens_seen": 10980200, "step": 16785 }, { "epoch": 8.79979035639413, "grad_norm": 0.16525200009346008, "learning_rate": 3.436813323021918e-05, "loss": 0.5207, "num_input_tokens_seen": 10983272, "step": 16790 }, { "epoch": 8.802410901467505, "grad_norm": 0.16417668759822845, "learning_rate": 3.43575311149891e-05, "loss": 0.6966, "num_input_tokens_seen": 10986952, "step": 16795 }, { "epoch": 8.80503144654088, "grad_norm": 0.09173838794231415, "learning_rate": 3.434692704227028e-05, "loss": 0.4027, "num_input_tokens_seen": 10990728, "step": 16800 }, { "epoch": 8.807651991614255, "grad_norm": 0.11314357072114944, "learning_rate": 3.4336321014280944e-05, "loss": 0.4945, "num_input_tokens_seen": 10994472, "step": 16805 }, { "epoch": 8.81027253668763, "grad_norm": 0.3428981900215149, "learning_rate": 3.432571303323977e-05, "loss": 0.5813, "num_input_tokens_seen": 11001224, "step": 16810 }, { "epoch": 8.812893081761006, "grad_norm": 0.12356404960155487, "learning_rate": 3.431510310136583e-05, "loss": 0.4681, "num_input_tokens_seen": 11005032, "step": 16815 }, { "epoch": 8.815513626834381, "grad_norm": 0.1754004955291748, "learning_rate": 3.4304491220878585e-05, "loss": 0.4737, "num_input_tokens_seen": 11007624, "step": 16820 }, { "epoch": 8.818134171907756, "grad_norm": 0.1953250616788864, "learning_rate": 3.429387739399793e-05, "loss": 0.5141, "num_input_tokens_seen": 11009704, "step": 16825 }, { "epoch": 8.820754716981131, "grad_norm": 0.1836714744567871, "learning_rate": 3.428326162294416e-05, "loss": 0.4522, "num_input_tokens_seen": 11013128, "step": 16830 }, { "epoch": 8.823375262054507, "grad_norm": 0.10371286422014236, "learning_rate": 3.427264390993795e-05, "loss": 0.4177, "num_input_tokens_seen": 11016328, "step": 16835 }, { "epoch": 8.825995807127882, "grad_norm": 0.12888462841510773, "learning_rate": 3.426202425720044e-05, "loss": 0.6869, "num_input_tokens_seen": 11019368, "step": 16840 }, { "epoch": 8.828616352201259, "grad_norm": 0.2183062732219696, "learning_rate": 3.425140266695312e-05, "loss": 0.5666, "num_input_tokens_seen": 11022696, "step": 16845 }, { "epoch": 8.831236897274634, "grad_norm": 0.23606693744659424, "learning_rate": 3.4240779141417905e-05, "loss": 0.4526, "num_input_tokens_seen": 11025992, "step": 16850 }, { "epoch": 8.833857442348009, "grad_norm": 0.11706769466400146, "learning_rate": 3.423015368281711e-05, "loss": 0.4467, "num_input_tokens_seen": 11028904, "step": 16855 }, { "epoch": 8.836477987421384, "grad_norm": 0.10807999968528748, "learning_rate": 3.421952629337348e-05, "loss": 0.5277, "num_input_tokens_seen": 11032712, "step": 16860 }, { "epoch": 8.83909853249476, "grad_norm": 0.1989871859550476, "learning_rate": 3.420889697531013e-05, "loss": 0.4198, "num_input_tokens_seen": 11036392, "step": 16865 }, { "epoch": 8.841719077568134, "grad_norm": 0.17974232137203217, "learning_rate": 3.419826573085058e-05, "loss": 0.5218, "num_input_tokens_seen": 11039368, "step": 16870 }, { "epoch": 8.84433962264151, "grad_norm": 0.17088617384433746, "learning_rate": 3.418763256221879e-05, "loss": 0.3926, "num_input_tokens_seen": 11042088, "step": 16875 }, { "epoch": 8.846960167714885, "grad_norm": 0.1954323649406433, "learning_rate": 3.4176997471639085e-05, "loss": 0.6123, "num_input_tokens_seen": 11045416, "step": 16880 }, { "epoch": 8.84958071278826, "grad_norm": 0.24029603600502014, "learning_rate": 3.4166360461336206e-05, "loss": 0.5245, "num_input_tokens_seen": 11048552, "step": 16885 }, { "epoch": 8.852201257861635, "grad_norm": 0.13218624889850616, "learning_rate": 3.41557215335353e-05, "loss": 0.3735, "num_input_tokens_seen": 11052072, "step": 16890 }, { "epoch": 8.85482180293501, "grad_norm": 0.1859077364206314, "learning_rate": 3.4145080690461887e-05, "loss": 0.5727, "num_input_tokens_seen": 11055016, "step": 16895 }, { "epoch": 8.857442348008385, "grad_norm": 0.09634093940258026, "learning_rate": 3.4134437934341936e-05, "loss": 0.3958, "num_input_tokens_seen": 11058376, "step": 16900 }, { "epoch": 8.86006289308176, "grad_norm": 0.12688107788562775, "learning_rate": 3.4123793267401784e-05, "loss": 0.4664, "num_input_tokens_seen": 11061384, "step": 16905 }, { "epoch": 8.862683438155136, "grad_norm": 0.17339327931404114, "learning_rate": 3.411314669186816e-05, "loss": 0.5984, "num_input_tokens_seen": 11064232, "step": 16910 }, { "epoch": 8.865303983228511, "grad_norm": 0.10479269921779633, "learning_rate": 3.410249820996823e-05, "loss": 0.4234, "num_input_tokens_seen": 11067368, "step": 16915 }, { "epoch": 8.867924528301886, "grad_norm": 0.15659846365451813, "learning_rate": 3.4091847823929513e-05, "loss": 0.4617, "num_input_tokens_seen": 11070056, "step": 16920 }, { "epoch": 8.870545073375261, "grad_norm": 0.1762099266052246, "learning_rate": 3.4081195535979946e-05, "loss": 0.4665, "num_input_tokens_seen": 11073768, "step": 16925 }, { "epoch": 8.873165618448636, "grad_norm": 0.17313766479492188, "learning_rate": 3.40705413483479e-05, "loss": 0.5124, "num_input_tokens_seen": 11076744, "step": 16930 }, { "epoch": 8.875786163522012, "grad_norm": 0.2244199812412262, "learning_rate": 3.405988526326207e-05, "loss": 0.4693, "num_input_tokens_seen": 11079848, "step": 16935 }, { "epoch": 8.878406708595389, "grad_norm": 0.15662437677383423, "learning_rate": 3.40492272829516e-05, "loss": 0.5204, "num_input_tokens_seen": 11082600, "step": 16940 }, { "epoch": 8.881027253668764, "grad_norm": 0.2287445068359375, "learning_rate": 3.403856740964604e-05, "loss": 0.5263, "num_input_tokens_seen": 11086280, "step": 16945 }, { "epoch": 8.883647798742139, "grad_norm": 0.1417286992073059, "learning_rate": 3.4027905645575275e-05, "loss": 0.5605, "num_input_tokens_seen": 11089352, "step": 16950 }, { "epoch": 8.886268343815514, "grad_norm": 0.24858392775058746, "learning_rate": 3.401724199296966e-05, "loss": 0.4647, "num_input_tokens_seen": 11092520, "step": 16955 }, { "epoch": 8.88888888888889, "grad_norm": 0.1769152283668518, "learning_rate": 3.400657645405989e-05, "loss": 0.5295, "num_input_tokens_seen": 11096136, "step": 16960 }, { "epoch": 8.891509433962264, "grad_norm": 0.3501614034175873, "learning_rate": 3.3995909031077065e-05, "loss": 0.583, "num_input_tokens_seen": 11099656, "step": 16965 }, { "epoch": 8.89412997903564, "grad_norm": 0.321979820728302, "learning_rate": 3.398523972625272e-05, "loss": 0.5842, "num_input_tokens_seen": 11102056, "step": 16970 }, { "epoch": 8.896750524109015, "grad_norm": 0.12525710463523865, "learning_rate": 3.3974568541818723e-05, "loss": 0.3609, "num_input_tokens_seen": 11104712, "step": 16975 }, { "epoch": 8.89937106918239, "grad_norm": 0.10827886313199997, "learning_rate": 3.396389548000738e-05, "loss": 0.3288, "num_input_tokens_seen": 11108616, "step": 16980 }, { "epoch": 8.901991614255765, "grad_norm": 0.2882954478263855, "learning_rate": 3.395322054305137e-05, "loss": 0.5481, "num_input_tokens_seen": 11111688, "step": 16985 }, { "epoch": 8.90461215932914, "grad_norm": 0.2204379141330719, "learning_rate": 3.394254373318376e-05, "loss": 0.4425, "num_input_tokens_seen": 11114632, "step": 16990 }, { "epoch": 8.907232704402515, "grad_norm": 0.16067726910114288, "learning_rate": 3.3931865052638026e-05, "loss": 0.417, "num_input_tokens_seen": 11117960, "step": 16995 }, { "epoch": 8.90985324947589, "grad_norm": 0.1355118751525879, "learning_rate": 3.392118450364802e-05, "loss": 0.5982, "num_input_tokens_seen": 11121384, "step": 17000 }, { "epoch": 8.912473794549266, "grad_norm": 0.1716524213552475, "learning_rate": 3.391050208844799e-05, "loss": 0.4855, "num_input_tokens_seen": 11124488, "step": 17005 }, { "epoch": 8.915094339622641, "grad_norm": 0.12038224190473557, "learning_rate": 3.389981780927259e-05, "loss": 0.4406, "num_input_tokens_seen": 11127656, "step": 17010 }, { "epoch": 8.917714884696016, "grad_norm": 0.17010405659675598, "learning_rate": 3.3889131668356834e-05, "loss": 0.4734, "num_input_tokens_seen": 11130024, "step": 17015 }, { "epoch": 8.920335429769391, "grad_norm": 0.19744136929512024, "learning_rate": 3.387844366793614e-05, "loss": 0.6142, "num_input_tokens_seen": 11132872, "step": 17020 }, { "epoch": 8.922955974842766, "grad_norm": 0.11736101657152176, "learning_rate": 3.3867753810246326e-05, "loss": 0.4724, "num_input_tokens_seen": 11136488, "step": 17025 }, { "epoch": 8.925576519916142, "grad_norm": 0.14517337083816528, "learning_rate": 3.3857062097523586e-05, "loss": 0.3229, "num_input_tokens_seen": 11140584, "step": 17030 }, { "epoch": 8.928197064989519, "grad_norm": 0.155693918466568, "learning_rate": 3.38463685320045e-05, "loss": 0.4913, "num_input_tokens_seen": 11143304, "step": 17035 }, { "epoch": 8.930817610062894, "grad_norm": 0.1260322481393814, "learning_rate": 3.3835673115926045e-05, "loss": 0.4349, "num_input_tokens_seen": 11146792, "step": 17040 }, { "epoch": 8.933438155136269, "grad_norm": 0.1336546242237091, "learning_rate": 3.382497585152557e-05, "loss": 0.3693, "num_input_tokens_seen": 11149832, "step": 17045 }, { "epoch": 8.936058700209644, "grad_norm": 0.1392393708229065, "learning_rate": 3.3814276741040835e-05, "loss": 0.3966, "num_input_tokens_seen": 11153192, "step": 17050 }, { "epoch": 8.93867924528302, "grad_norm": 0.14075002074241638, "learning_rate": 3.380357578670996e-05, "loss": 0.5119, "num_input_tokens_seen": 11155912, "step": 17055 }, { "epoch": 8.941299790356394, "grad_norm": 0.1788751631975174, "learning_rate": 3.379287299077146e-05, "loss": 0.4575, "num_input_tokens_seen": 11158472, "step": 17060 }, { "epoch": 8.94392033542977, "grad_norm": 0.1892213225364685, "learning_rate": 3.3782168355464263e-05, "loss": 0.5845, "num_input_tokens_seen": 11161192, "step": 17065 }, { "epoch": 8.946540880503145, "grad_norm": 0.15074525773525238, "learning_rate": 3.377146188302762e-05, "loss": 0.3903, "num_input_tokens_seen": 11163688, "step": 17070 }, { "epoch": 8.94916142557652, "grad_norm": 0.2995770573616028, "learning_rate": 3.376075357570124e-05, "loss": 0.5378, "num_input_tokens_seen": 11166408, "step": 17075 }, { "epoch": 8.951781970649895, "grad_norm": 0.2593911588191986, "learning_rate": 3.3750043435725144e-05, "loss": 0.5581, "num_input_tokens_seen": 11169640, "step": 17080 }, { "epoch": 8.95440251572327, "grad_norm": 0.26664116978645325, "learning_rate": 3.373933146533979e-05, "loss": 0.4403, "num_input_tokens_seen": 11172616, "step": 17085 }, { "epoch": 8.957023060796645, "grad_norm": 0.15803027153015137, "learning_rate": 3.3728617666786e-05, "loss": 0.7571, "num_input_tokens_seen": 11175880, "step": 17090 }, { "epoch": 8.95964360587002, "grad_norm": 0.1533956527709961, "learning_rate": 3.371790204230497e-05, "loss": 0.6241, "num_input_tokens_seen": 11178696, "step": 17095 }, { "epoch": 8.962264150943396, "grad_norm": 0.12230409681797028, "learning_rate": 3.3707184594138295e-05, "loss": 0.5416, "num_input_tokens_seen": 11182152, "step": 17100 }, { "epoch": 8.964884696016771, "grad_norm": 0.16772271692752838, "learning_rate": 3.369646532452795e-05, "loss": 0.5925, "num_input_tokens_seen": 11185416, "step": 17105 }, { "epoch": 8.967505241090146, "grad_norm": 0.22923369705677032, "learning_rate": 3.3685744235716254e-05, "loss": 0.5489, "num_input_tokens_seen": 11188392, "step": 17110 }, { "epoch": 8.970125786163521, "grad_norm": 0.12646342813968658, "learning_rate": 3.367502132994598e-05, "loss": 0.5072, "num_input_tokens_seen": 11191528, "step": 17115 }, { "epoch": 8.972746331236896, "grad_norm": 0.1282106637954712, "learning_rate": 3.366429660946019e-05, "loss": 0.4372, "num_input_tokens_seen": 11194792, "step": 17120 }, { "epoch": 8.975366876310272, "grad_norm": 0.10955532640218735, "learning_rate": 3.365357007650241e-05, "loss": 0.4969, "num_input_tokens_seen": 11198600, "step": 17125 }, { "epoch": 8.977987421383649, "grad_norm": 0.22728686034679413, "learning_rate": 3.364284173331649e-05, "loss": 0.3458, "num_input_tokens_seen": 11201448, "step": 17130 }, { "epoch": 8.980607966457024, "grad_norm": 0.16672579944133759, "learning_rate": 3.3632111582146686e-05, "loss": 0.3973, "num_input_tokens_seen": 11205160, "step": 17135 }, { "epoch": 8.983228511530399, "grad_norm": 0.17318671941757202, "learning_rate": 3.362137962523761e-05, "loss": 0.3294, "num_input_tokens_seen": 11208328, "step": 17140 }, { "epoch": 8.985849056603774, "grad_norm": 0.12747234106063843, "learning_rate": 3.361064586483429e-05, "loss": 0.4617, "num_input_tokens_seen": 11211848, "step": 17145 }, { "epoch": 8.98846960167715, "grad_norm": 0.10178764909505844, "learning_rate": 3.359991030318208e-05, "loss": 0.5917, "num_input_tokens_seen": 11214376, "step": 17150 }, { "epoch": 8.991090146750524, "grad_norm": 0.20022527873516083, "learning_rate": 3.358917294252675e-05, "loss": 0.413, "num_input_tokens_seen": 11218376, "step": 17155 }, { "epoch": 8.9937106918239, "grad_norm": 0.13950318098068237, "learning_rate": 3.357843378511443e-05, "loss": 0.464, "num_input_tokens_seen": 11221768, "step": 17160 }, { "epoch": 8.996331236897275, "grad_norm": 0.16856636106967926, "learning_rate": 3.3567692833191635e-05, "loss": 0.5023, "num_input_tokens_seen": 11224712, "step": 17165 }, { "epoch": 8.99895178197065, "grad_norm": 0.16649450361728668, "learning_rate": 3.355695008900525e-05, "loss": 0.3939, "num_input_tokens_seen": 11229096, "step": 17170 }, { "epoch": 9.0, "eval_loss": 0.46673670411109924, "eval_runtime": 9.9066, "eval_samples_per_second": 85.599, "eval_steps_per_second": 21.4, "num_input_tokens_seen": 11229792, "step": 17172 }, { "epoch": 9.001572327044025, "grad_norm": 0.15146976709365845, "learning_rate": 3.354620555480251e-05, "loss": 0.2781, "num_input_tokens_seen": 11232064, "step": 17175 }, { "epoch": 9.0041928721174, "grad_norm": 0.16347989439964294, "learning_rate": 3.353545923283108e-05, "loss": 0.666, "num_input_tokens_seen": 11236960, "step": 17180 }, { "epoch": 9.006813417190775, "grad_norm": 0.14803197979927063, "learning_rate": 3.352471112533895e-05, "loss": 0.5254, "num_input_tokens_seen": 11239840, "step": 17185 }, { "epoch": 9.00943396226415, "grad_norm": 0.22018694877624512, "learning_rate": 3.351396123457451e-05, "loss": 0.492, "num_input_tokens_seen": 11242432, "step": 17190 }, { "epoch": 9.012054507337526, "grad_norm": 0.24257108569145203, "learning_rate": 3.350320956278652e-05, "loss": 0.3616, "num_input_tokens_seen": 11245312, "step": 17195 }, { "epoch": 9.014675052410901, "grad_norm": 0.2963641583919525, "learning_rate": 3.349245611222409e-05, "loss": 0.4614, "num_input_tokens_seen": 11249152, "step": 17200 }, { "epoch": 9.017295597484276, "grad_norm": 0.2540249824523926, "learning_rate": 3.348170088513673e-05, "loss": 0.64, "num_input_tokens_seen": 11251968, "step": 17205 }, { "epoch": 9.019916142557651, "grad_norm": 0.20986230671405792, "learning_rate": 3.347094388377431e-05, "loss": 0.5965, "num_input_tokens_seen": 11256224, "step": 17210 }, { "epoch": 9.022536687631026, "grad_norm": 0.19120639562606812, "learning_rate": 3.3460185110387076e-05, "loss": 0.4997, "num_input_tokens_seen": 11258944, "step": 17215 }, { "epoch": 9.025157232704403, "grad_norm": 0.21835733950138092, "learning_rate": 3.344942456722563e-05, "loss": 0.4472, "num_input_tokens_seen": 11263488, "step": 17220 }, { "epoch": 9.027777777777779, "grad_norm": 0.13302555680274963, "learning_rate": 3.343866225654096e-05, "loss": 0.4481, "num_input_tokens_seen": 11266560, "step": 17225 }, { "epoch": 9.030398322851154, "grad_norm": 0.1345185935497284, "learning_rate": 3.342789818058443e-05, "loss": 0.489, "num_input_tokens_seen": 11270336, "step": 17230 }, { "epoch": 9.033018867924529, "grad_norm": 0.2043968290090561, "learning_rate": 3.341713234160774e-05, "loss": 0.5164, "num_input_tokens_seen": 11273536, "step": 17235 }, { "epoch": 9.035639412997904, "grad_norm": 0.13297532498836517, "learning_rate": 3.3406364741862986e-05, "loss": 0.441, "num_input_tokens_seen": 11277408, "step": 17240 }, { "epoch": 9.03825995807128, "grad_norm": 0.1453939825296402, "learning_rate": 3.3395595383602644e-05, "loss": 0.4301, "num_input_tokens_seen": 11280576, "step": 17245 }, { "epoch": 9.040880503144654, "grad_norm": 0.10255442559719086, "learning_rate": 3.338482426907953e-05, "loss": 0.5059, "num_input_tokens_seen": 11283904, "step": 17250 }, { "epoch": 9.04350104821803, "grad_norm": 0.12037003040313721, "learning_rate": 3.337405140054683e-05, "loss": 0.4466, "num_input_tokens_seen": 11286592, "step": 17255 }, { "epoch": 9.046121593291405, "grad_norm": 0.17044761776924133, "learning_rate": 3.336327678025813e-05, "loss": 0.413, "num_input_tokens_seen": 11289376, "step": 17260 }, { "epoch": 9.04874213836478, "grad_norm": 0.28462353348731995, "learning_rate": 3.3352500410467324e-05, "loss": 0.4278, "num_input_tokens_seen": 11292320, "step": 17265 }, { "epoch": 9.051362683438155, "grad_norm": 0.15532587468624115, "learning_rate": 3.3341722293428726e-05, "loss": 0.6792, "num_input_tokens_seen": 11295392, "step": 17270 }, { "epoch": 9.05398322851153, "grad_norm": 0.18874669075012207, "learning_rate": 3.3330942431397e-05, "loss": 0.5788, "num_input_tokens_seen": 11298656, "step": 17275 }, { "epoch": 9.056603773584905, "grad_norm": 0.15540967881679535, "learning_rate": 3.332016082662715e-05, "loss": 0.5068, "num_input_tokens_seen": 11301824, "step": 17280 }, { "epoch": 9.05922431865828, "grad_norm": 0.10799603909254074, "learning_rate": 3.3309377481374584e-05, "loss": 0.4961, "num_input_tokens_seen": 11304352, "step": 17285 }, { "epoch": 9.061844863731656, "grad_norm": 0.2471490502357483, "learning_rate": 3.3298592397895046e-05, "loss": 0.5106, "num_input_tokens_seen": 11306976, "step": 17290 }, { "epoch": 9.064465408805031, "grad_norm": 0.20339198410511017, "learning_rate": 3.3287805578444645e-05, "loss": 0.5038, "num_input_tokens_seen": 11310048, "step": 17295 }, { "epoch": 9.067085953878406, "grad_norm": 0.12589570879936218, "learning_rate": 3.327701702527989e-05, "loss": 0.4633, "num_input_tokens_seen": 11313984, "step": 17300 }, { "epoch": 9.069706498951781, "grad_norm": 0.17869296669960022, "learning_rate": 3.3266226740657584e-05, "loss": 0.3601, "num_input_tokens_seen": 11317536, "step": 17305 }, { "epoch": 9.072327044025156, "grad_norm": 0.11460200697183609, "learning_rate": 3.3255434726834954e-05, "loss": 0.4341, "num_input_tokens_seen": 11321152, "step": 17310 }, { "epoch": 9.074947589098532, "grad_norm": 0.1975567787885666, "learning_rate": 3.324464098606957e-05, "loss": 0.5653, "num_input_tokens_seen": 11323968, "step": 17315 }, { "epoch": 9.077568134171909, "grad_norm": 0.18212296068668365, "learning_rate": 3.323384552061934e-05, "loss": 0.4868, "num_input_tokens_seen": 11326496, "step": 17320 }, { "epoch": 9.080188679245284, "grad_norm": 0.18815737962722778, "learning_rate": 3.322304833274257e-05, "loss": 0.4565, "num_input_tokens_seen": 11329696, "step": 17325 }, { "epoch": 9.082809224318659, "grad_norm": 0.0888073742389679, "learning_rate": 3.32122494246979e-05, "loss": 0.4232, "num_input_tokens_seen": 11333120, "step": 17330 }, { "epoch": 9.085429769392034, "grad_norm": 0.10385490208864212, "learning_rate": 3.320144879874434e-05, "loss": 0.4581, "num_input_tokens_seen": 11336352, "step": 17335 }, { "epoch": 9.08805031446541, "grad_norm": 0.16565954685211182, "learning_rate": 3.319064645714126e-05, "loss": 0.4486, "num_input_tokens_seen": 11339968, "step": 17340 }, { "epoch": 9.090670859538784, "grad_norm": 0.20687896013259888, "learning_rate": 3.3179842402148386e-05, "loss": 0.4504, "num_input_tokens_seen": 11343296, "step": 17345 }, { "epoch": 9.09329140461216, "grad_norm": 0.12244468927383423, "learning_rate": 3.316903663602581e-05, "loss": 0.5981, "num_input_tokens_seen": 11347808, "step": 17350 }, { "epoch": 9.095911949685535, "grad_norm": 0.1351558417081833, "learning_rate": 3.315822916103395e-05, "loss": 0.6064, "num_input_tokens_seen": 11350880, "step": 17355 }, { "epoch": 9.09853249475891, "grad_norm": 0.09684723615646362, "learning_rate": 3.3147419979433635e-05, "loss": 0.4247, "num_input_tokens_seen": 11354016, "step": 17360 }, { "epoch": 9.101153039832285, "grad_norm": 0.1282985955476761, "learning_rate": 3.313660909348601e-05, "loss": 0.3826, "num_input_tokens_seen": 11357504, "step": 17365 }, { "epoch": 9.10377358490566, "grad_norm": 0.14748728275299072, "learning_rate": 3.312579650545259e-05, "loss": 0.5112, "num_input_tokens_seen": 11360512, "step": 17370 }, { "epoch": 9.106394129979035, "grad_norm": 0.18254822492599487, "learning_rate": 3.311498221759524e-05, "loss": 0.5684, "num_input_tokens_seen": 11363392, "step": 17375 }, { "epoch": 9.10901467505241, "grad_norm": 0.1276947408914566, "learning_rate": 3.310416623217619e-05, "loss": 0.4673, "num_input_tokens_seen": 11366400, "step": 17380 }, { "epoch": 9.111635220125786, "grad_norm": 0.25130516290664673, "learning_rate": 3.309334855145803e-05, "loss": 0.4669, "num_input_tokens_seen": 11369824, "step": 17385 }, { "epoch": 9.114255765199161, "grad_norm": 0.1573469489812851, "learning_rate": 3.3082529177703684e-05, "loss": 0.5436, "num_input_tokens_seen": 11373376, "step": 17390 }, { "epoch": 9.116876310272536, "grad_norm": 0.21871134638786316, "learning_rate": 3.3071708113176435e-05, "loss": 0.4844, "num_input_tokens_seen": 11376672, "step": 17395 }, { "epoch": 9.119496855345911, "grad_norm": 0.22312059998512268, "learning_rate": 3.306088536013995e-05, "loss": 0.5371, "num_input_tokens_seen": 11379456, "step": 17400 }, { "epoch": 9.122117400419286, "grad_norm": 0.1725604236125946, "learning_rate": 3.3050060920858206e-05, "loss": 0.5196, "num_input_tokens_seen": 11382560, "step": 17405 }, { "epoch": 9.124737945492662, "grad_norm": 0.11442624777555466, "learning_rate": 3.3039234797595544e-05, "loss": 0.4694, "num_input_tokens_seen": 11385792, "step": 17410 }, { "epoch": 9.127358490566039, "grad_norm": 0.1429189145565033, "learning_rate": 3.3028406992616695e-05, "loss": 0.446, "num_input_tokens_seen": 11391648, "step": 17415 }, { "epoch": 9.129979035639414, "grad_norm": 0.15204964578151703, "learning_rate": 3.301757750818669e-05, "loss": 0.3606, "num_input_tokens_seen": 11395744, "step": 17420 }, { "epoch": 9.132599580712789, "grad_norm": 0.13222473859786987, "learning_rate": 3.300674634657094e-05, "loss": 0.4458, "num_input_tokens_seen": 11398496, "step": 17425 }, { "epoch": 9.135220125786164, "grad_norm": 0.1545359343290329, "learning_rate": 3.2995913510035206e-05, "loss": 0.4219, "num_input_tokens_seen": 11401408, "step": 17430 }, { "epoch": 9.13784067085954, "grad_norm": 0.23713724315166473, "learning_rate": 3.298507900084558e-05, "loss": 0.567, "num_input_tokens_seen": 11404448, "step": 17435 }, { "epoch": 9.140461215932914, "grad_norm": 0.21730481088161469, "learning_rate": 3.2974242821268524e-05, "loss": 0.4471, "num_input_tokens_seen": 11407680, "step": 17440 }, { "epoch": 9.14308176100629, "grad_norm": 0.12783664464950562, "learning_rate": 3.2963404973570855e-05, "loss": 0.4949, "num_input_tokens_seen": 11410752, "step": 17445 }, { "epoch": 9.145702306079665, "grad_norm": 0.1344553679227829, "learning_rate": 3.2952565460019715e-05, "loss": 0.446, "num_input_tokens_seen": 11414432, "step": 17450 }, { "epoch": 9.14832285115304, "grad_norm": 0.2093842625617981, "learning_rate": 3.29417242828826e-05, "loss": 0.5769, "num_input_tokens_seen": 11417568, "step": 17455 }, { "epoch": 9.150943396226415, "grad_norm": 0.3795507848262787, "learning_rate": 3.2930881444427374e-05, "loss": 0.5647, "num_input_tokens_seen": 11421248, "step": 17460 }, { "epoch": 9.15356394129979, "grad_norm": 0.20141853392124176, "learning_rate": 3.292003694692223e-05, "loss": 0.4167, "num_input_tokens_seen": 11425472, "step": 17465 }, { "epoch": 9.156184486373165, "grad_norm": 0.11531771719455719, "learning_rate": 3.290919079263572e-05, "loss": 0.5271, "num_input_tokens_seen": 11428928, "step": 17470 }, { "epoch": 9.15880503144654, "grad_norm": 0.19516663253307343, "learning_rate": 3.2898342983836715e-05, "loss": 0.5524, "num_input_tokens_seen": 11433664, "step": 17475 }, { "epoch": 9.161425576519916, "grad_norm": 0.13423047959804535, "learning_rate": 3.288749352279447e-05, "loss": 0.7409, "num_input_tokens_seen": 11437024, "step": 17480 }, { "epoch": 9.164046121593291, "grad_norm": 0.14103834331035614, "learning_rate": 3.2876642411778573e-05, "loss": 0.4019, "num_input_tokens_seen": 11442304, "step": 17485 }, { "epoch": 9.166666666666666, "grad_norm": 0.15243102610111237, "learning_rate": 3.286578965305894e-05, "loss": 0.4458, "num_input_tokens_seen": 11445248, "step": 17490 }, { "epoch": 9.169287211740041, "grad_norm": 0.660611093044281, "learning_rate": 3.285493524890585e-05, "loss": 0.6562, "num_input_tokens_seen": 11448064, "step": 17495 }, { "epoch": 9.171907756813416, "grad_norm": 0.11565237492322922, "learning_rate": 3.284407920158992e-05, "loss": 0.3953, "num_input_tokens_seen": 11451232, "step": 17500 }, { "epoch": 9.174528301886792, "grad_norm": 0.13532905280590057, "learning_rate": 3.283322151338211e-05, "loss": 0.4391, "num_input_tokens_seen": 11455648, "step": 17505 }, { "epoch": 9.177148846960169, "grad_norm": 0.18588490784168243, "learning_rate": 3.282236218655373e-05, "loss": 0.5886, "num_input_tokens_seen": 11459296, "step": 17510 }, { "epoch": 9.179769392033544, "grad_norm": 0.17202994227409363, "learning_rate": 3.281150122337642e-05, "loss": 0.6745, "num_input_tokens_seen": 11462240, "step": 17515 }, { "epoch": 9.182389937106919, "grad_norm": 0.21749147772789001, "learning_rate": 3.280063862612216e-05, "loss": 0.5433, "num_input_tokens_seen": 11465600, "step": 17520 }, { "epoch": 9.185010482180294, "grad_norm": 0.1334819495677948, "learning_rate": 3.2789774397063314e-05, "loss": 0.4893, "num_input_tokens_seen": 11468896, "step": 17525 }, { "epoch": 9.18763102725367, "grad_norm": 0.22873789072036743, "learning_rate": 3.277890853847253e-05, "loss": 0.5869, "num_input_tokens_seen": 11471808, "step": 17530 }, { "epoch": 9.190251572327044, "grad_norm": 0.22919869422912598, "learning_rate": 3.2768041052622824e-05, "loss": 0.389, "num_input_tokens_seen": 11474752, "step": 17535 }, { "epoch": 9.19287211740042, "grad_norm": 0.17045582830905914, "learning_rate": 3.275717194178755e-05, "loss": 0.4116, "num_input_tokens_seen": 11477760, "step": 17540 }, { "epoch": 9.195492662473795, "grad_norm": 0.1445389837026596, "learning_rate": 3.27463012082404e-05, "loss": 0.4351, "num_input_tokens_seen": 11480864, "step": 17545 }, { "epoch": 9.19811320754717, "grad_norm": 0.29551318287849426, "learning_rate": 3.2735428854255416e-05, "loss": 0.3968, "num_input_tokens_seen": 11483776, "step": 17550 }, { "epoch": 9.200733752620545, "grad_norm": 0.1543920934200287, "learning_rate": 3.272455488210697e-05, "loss": 0.4989, "num_input_tokens_seen": 11487680, "step": 17555 }, { "epoch": 9.20335429769392, "grad_norm": 0.14461947977542877, "learning_rate": 3.271367929406977e-05, "loss": 0.5081, "num_input_tokens_seen": 11491072, "step": 17560 }, { "epoch": 9.205974842767295, "grad_norm": 0.1351877748966217, "learning_rate": 3.270280209241886e-05, "loss": 0.4465, "num_input_tokens_seen": 11494432, "step": 17565 }, { "epoch": 9.20859538784067, "grad_norm": 0.20829223096370697, "learning_rate": 3.2691923279429624e-05, "loss": 0.5522, "num_input_tokens_seen": 11498144, "step": 17570 }, { "epoch": 9.211215932914046, "grad_norm": 0.08800600469112396, "learning_rate": 3.2681042857377804e-05, "loss": 0.2985, "num_input_tokens_seen": 11501728, "step": 17575 }, { "epoch": 9.213836477987421, "grad_norm": 0.16197152435779572, "learning_rate": 3.2670160828539435e-05, "loss": 0.5271, "num_input_tokens_seen": 11505088, "step": 17580 }, { "epoch": 9.216457023060796, "grad_norm": 0.2005462944507599, "learning_rate": 3.2659277195190934e-05, "loss": 0.6093, "num_input_tokens_seen": 11507712, "step": 17585 }, { "epoch": 9.219077568134171, "grad_norm": 0.26857802271842957, "learning_rate": 3.2648391959609026e-05, "loss": 0.4343, "num_input_tokens_seen": 11510560, "step": 17590 }, { "epoch": 9.221698113207546, "grad_norm": 0.13920727372169495, "learning_rate": 3.263750512407078e-05, "loss": 0.6226, "num_input_tokens_seen": 11513760, "step": 17595 }, { "epoch": 9.224318658280922, "grad_norm": 0.14025738835334778, "learning_rate": 3.262661669085359e-05, "loss": 0.592, "num_input_tokens_seen": 11516288, "step": 17600 }, { "epoch": 9.226939203354299, "grad_norm": 0.14815840125083923, "learning_rate": 3.2615726662235205e-05, "loss": 0.3859, "num_input_tokens_seen": 11522496, "step": 17605 }, { "epoch": 9.229559748427674, "grad_norm": 0.24188172817230225, "learning_rate": 3.2604835040493676e-05, "loss": 0.7745, "num_input_tokens_seen": 11525216, "step": 17610 }, { "epoch": 9.232180293501049, "grad_norm": 0.17174385488033295, "learning_rate": 3.259394182790742e-05, "loss": 0.5246, "num_input_tokens_seen": 11527552, "step": 17615 }, { "epoch": 9.234800838574424, "grad_norm": 0.14265239238739014, "learning_rate": 3.2583047026755174e-05, "loss": 0.4789, "num_input_tokens_seen": 11530880, "step": 17620 }, { "epoch": 9.2374213836478, "grad_norm": 0.18762372434139252, "learning_rate": 3.2572150639316e-05, "loss": 0.5498, "num_input_tokens_seen": 11534208, "step": 17625 }, { "epoch": 9.240041928721174, "grad_norm": 0.11790566891431808, "learning_rate": 3.256125266786931e-05, "loss": 0.4543, "num_input_tokens_seen": 11537632, "step": 17630 }, { "epoch": 9.24266247379455, "grad_norm": 0.15744398534297943, "learning_rate": 3.255035311469481e-05, "loss": 0.4821, "num_input_tokens_seen": 11541024, "step": 17635 }, { "epoch": 9.245283018867925, "grad_norm": 0.21034066379070282, "learning_rate": 3.2539451982072586e-05, "loss": 0.6135, "num_input_tokens_seen": 11544192, "step": 17640 }, { "epoch": 9.2479035639413, "grad_norm": 0.2670862078666687, "learning_rate": 3.2528549272283016e-05, "loss": 0.5494, "num_input_tokens_seen": 11547840, "step": 17645 }, { "epoch": 9.250524109014675, "grad_norm": 0.20900960266590118, "learning_rate": 3.251764498760683e-05, "loss": 0.5008, "num_input_tokens_seen": 11551552, "step": 17650 }, { "epoch": 9.25314465408805, "grad_norm": 0.22400154173374176, "learning_rate": 3.2506739130325084e-05, "loss": 0.6487, "num_input_tokens_seen": 11554880, "step": 17655 }, { "epoch": 9.255765199161425, "grad_norm": 0.12769344449043274, "learning_rate": 3.2495831702719146e-05, "loss": 0.4322, "num_input_tokens_seen": 11557472, "step": 17660 }, { "epoch": 9.2583857442348, "grad_norm": 0.14543218910694122, "learning_rate": 3.248492270707074e-05, "loss": 0.4567, "num_input_tokens_seen": 11560992, "step": 17665 }, { "epoch": 9.261006289308176, "grad_norm": 0.1663162261247635, "learning_rate": 3.247401214566189e-05, "loss": 0.5164, "num_input_tokens_seen": 11564480, "step": 17670 }, { "epoch": 9.26362683438155, "grad_norm": 0.1427459716796875, "learning_rate": 3.2463100020774965e-05, "loss": 0.4135, "num_input_tokens_seen": 11567936, "step": 17675 }, { "epoch": 9.266247379454926, "grad_norm": 0.21729625761508942, "learning_rate": 3.245218633469267e-05, "loss": 0.4883, "num_input_tokens_seen": 11571360, "step": 17680 }, { "epoch": 9.268867924528301, "grad_norm": 0.2033582478761673, "learning_rate": 3.244127108969801e-05, "loss": 0.6501, "num_input_tokens_seen": 11574336, "step": 17685 }, { "epoch": 9.271488469601676, "grad_norm": 0.17222948372364044, "learning_rate": 3.2430354288074314e-05, "loss": 0.4577, "num_input_tokens_seen": 11576928, "step": 17690 }, { "epoch": 9.274109014675052, "grad_norm": 0.28403404355049133, "learning_rate": 3.241943593210529e-05, "loss": 0.4852, "num_input_tokens_seen": 11580096, "step": 17695 }, { "epoch": 9.276729559748428, "grad_norm": 0.19856539368629456, "learning_rate": 3.2408516024074916e-05, "loss": 0.4159, "num_input_tokens_seen": 11583872, "step": 17700 }, { "epoch": 9.279350104821804, "grad_norm": 0.20359210669994354, "learning_rate": 3.2397594566267505e-05, "loss": 0.5564, "num_input_tokens_seen": 11586496, "step": 17705 }, { "epoch": 9.281970649895179, "grad_norm": 0.19603127241134644, "learning_rate": 3.23866715609677e-05, "loss": 0.5619, "num_input_tokens_seen": 11589280, "step": 17710 }, { "epoch": 9.284591194968554, "grad_norm": 0.12407258152961731, "learning_rate": 3.237574701046049e-05, "loss": 0.5417, "num_input_tokens_seen": 11592832, "step": 17715 }, { "epoch": 9.28721174004193, "grad_norm": 0.11770105361938477, "learning_rate": 3.236482091703115e-05, "loss": 0.4168, "num_input_tokens_seen": 11595904, "step": 17720 }, { "epoch": 9.289832285115304, "grad_norm": 0.17488975822925568, "learning_rate": 3.235389328296528e-05, "loss": 0.4744, "num_input_tokens_seen": 11600672, "step": 17725 }, { "epoch": 9.29245283018868, "grad_norm": 0.15704675018787384, "learning_rate": 3.2342964110548845e-05, "loss": 0.4319, "num_input_tokens_seen": 11603968, "step": 17730 }, { "epoch": 9.295073375262055, "grad_norm": 0.14650321006774902, "learning_rate": 3.2332033402068093e-05, "loss": 0.4367, "num_input_tokens_seen": 11606944, "step": 17735 }, { "epoch": 9.29769392033543, "grad_norm": 0.16299933195114136, "learning_rate": 3.2321101159809605e-05, "loss": 0.4847, "num_input_tokens_seen": 11610368, "step": 17740 }, { "epoch": 9.300314465408805, "grad_norm": 0.18543827533721924, "learning_rate": 3.231016738606027e-05, "loss": 0.4324, "num_input_tokens_seen": 11613120, "step": 17745 }, { "epoch": 9.30293501048218, "grad_norm": 0.21446481347084045, "learning_rate": 3.2299232083107313e-05, "loss": 0.4201, "num_input_tokens_seen": 11616352, "step": 17750 }, { "epoch": 9.305555555555555, "grad_norm": 0.14411620795726776, "learning_rate": 3.228829525323829e-05, "loss": 0.4964, "num_input_tokens_seen": 11621536, "step": 17755 }, { "epoch": 9.30817610062893, "grad_norm": 0.1554349809885025, "learning_rate": 3.227735689874104e-05, "loss": 0.4489, "num_input_tokens_seen": 11624480, "step": 17760 }, { "epoch": 9.310796645702306, "grad_norm": 0.1954045593738556, "learning_rate": 3.226641702190375e-05, "loss": 0.4093, "num_input_tokens_seen": 11627360, "step": 17765 }, { "epoch": 9.31341719077568, "grad_norm": 0.14421656727790833, "learning_rate": 3.225547562501493e-05, "loss": 0.3746, "num_input_tokens_seen": 11631360, "step": 17770 }, { "epoch": 9.316037735849056, "grad_norm": 0.1312505602836609, "learning_rate": 3.2244532710363375e-05, "loss": 0.4714, "num_input_tokens_seen": 11634720, "step": 17775 }, { "epoch": 9.318658280922431, "grad_norm": 0.10971162468194962, "learning_rate": 3.2233588280238224e-05, "loss": 0.5055, "num_input_tokens_seen": 11638144, "step": 17780 }, { "epoch": 9.321278825995806, "grad_norm": 0.1811259388923645, "learning_rate": 3.2222642336928944e-05, "loss": 0.4763, "num_input_tokens_seen": 11640960, "step": 17785 }, { "epoch": 9.323899371069182, "grad_norm": 0.1602683961391449, "learning_rate": 3.2211694882725275e-05, "loss": 0.3601, "num_input_tokens_seen": 11644320, "step": 17790 }, { "epoch": 9.326519916142558, "grad_norm": 0.1271882802248001, "learning_rate": 3.220074591991731e-05, "loss": 0.4333, "num_input_tokens_seen": 11647008, "step": 17795 }, { "epoch": 9.329140461215934, "grad_norm": 0.19717787206172943, "learning_rate": 3.218979545079546e-05, "loss": 0.4474, "num_input_tokens_seen": 11650880, "step": 17800 }, { "epoch": 9.331761006289309, "grad_norm": 0.1246858462691307, "learning_rate": 3.217884347765042e-05, "loss": 0.5298, "num_input_tokens_seen": 11654336, "step": 17805 }, { "epoch": 9.334381551362684, "grad_norm": 0.2246944159269333, "learning_rate": 3.216789000277323e-05, "loss": 0.4559, "num_input_tokens_seen": 11657120, "step": 17810 }, { "epoch": 9.33700209643606, "grad_norm": 0.13605813682079315, "learning_rate": 3.2156935028455224e-05, "loss": 0.4646, "num_input_tokens_seen": 11660480, "step": 17815 }, { "epoch": 9.339622641509434, "grad_norm": 0.14304311573505402, "learning_rate": 3.2145978556988056e-05, "loss": 0.493, "num_input_tokens_seen": 11663264, "step": 17820 }, { "epoch": 9.34224318658281, "grad_norm": 0.14891335368156433, "learning_rate": 3.213502059066371e-05, "loss": 0.4517, "num_input_tokens_seen": 11665632, "step": 17825 }, { "epoch": 9.344863731656185, "grad_norm": 0.12504521012306213, "learning_rate": 3.212406113177445e-05, "loss": 0.5507, "num_input_tokens_seen": 11668544, "step": 17830 }, { "epoch": 9.34748427672956, "grad_norm": 0.1906566470861435, "learning_rate": 3.211310018261287e-05, "loss": 0.4546, "num_input_tokens_seen": 11672192, "step": 17835 }, { "epoch": 9.350104821802935, "grad_norm": 0.19612835347652435, "learning_rate": 3.21021377454719e-05, "loss": 0.5535, "num_input_tokens_seen": 11675104, "step": 17840 }, { "epoch": 9.35272536687631, "grad_norm": 0.13611073791980743, "learning_rate": 3.209117382264473e-05, "loss": 0.4479, "num_input_tokens_seen": 11678304, "step": 17845 }, { "epoch": 9.355345911949685, "grad_norm": 0.16775871813297272, "learning_rate": 3.208020841642489e-05, "loss": 0.4781, "num_input_tokens_seen": 11681760, "step": 17850 }, { "epoch": 9.35796645702306, "grad_norm": 0.26539430022239685, "learning_rate": 3.206924152910623e-05, "loss": 0.4057, "num_input_tokens_seen": 11684480, "step": 17855 }, { "epoch": 9.360587002096436, "grad_norm": 0.24734120070934296, "learning_rate": 3.205827316298289e-05, "loss": 0.5087, "num_input_tokens_seen": 11687488, "step": 17860 }, { "epoch": 9.36320754716981, "grad_norm": 0.20170904695987701, "learning_rate": 3.204730332034934e-05, "loss": 0.4518, "num_input_tokens_seen": 11691456, "step": 17865 }, { "epoch": 9.365828092243186, "grad_norm": 0.11672534048557281, "learning_rate": 3.2036332003500324e-05, "loss": 0.675, "num_input_tokens_seen": 11694016, "step": 17870 }, { "epoch": 9.368448637316561, "grad_norm": 0.17559002339839935, "learning_rate": 3.2025359214730924e-05, "loss": 0.5353, "num_input_tokens_seen": 11696960, "step": 17875 }, { "epoch": 9.371069182389936, "grad_norm": 0.23475764691829681, "learning_rate": 3.201438495633654e-05, "loss": 0.4339, "num_input_tokens_seen": 11700128, "step": 17880 }, { "epoch": 9.373689727463312, "grad_norm": 0.115696020424366, "learning_rate": 3.200340923061284e-05, "loss": 0.4258, "num_input_tokens_seen": 11704192, "step": 17885 }, { "epoch": 9.376310272536688, "grad_norm": 0.21370664238929749, "learning_rate": 3.199243203985583e-05, "loss": 0.5009, "num_input_tokens_seen": 11707552, "step": 17890 }, { "epoch": 9.378930817610064, "grad_norm": 0.33242127299308777, "learning_rate": 3.1981453386361804e-05, "loss": 0.4632, "num_input_tokens_seen": 11714304, "step": 17895 }, { "epoch": 9.381551362683439, "grad_norm": 0.186392679810524, "learning_rate": 3.1970473272427386e-05, "loss": 0.4613, "num_input_tokens_seen": 11716896, "step": 17900 }, { "epoch": 9.384171907756814, "grad_norm": 0.2016812562942505, "learning_rate": 3.195949170034948e-05, "loss": 0.4661, "num_input_tokens_seen": 11719680, "step": 17905 }, { "epoch": 9.38679245283019, "grad_norm": 0.12557363510131836, "learning_rate": 3.19485086724253e-05, "loss": 0.4172, "num_input_tokens_seen": 11723008, "step": 17910 }, { "epoch": 9.389412997903564, "grad_norm": 0.13334468007087708, "learning_rate": 3.193752419095239e-05, "loss": 0.3936, "num_input_tokens_seen": 11726272, "step": 17915 }, { "epoch": 9.39203354297694, "grad_norm": 0.10059771686792374, "learning_rate": 3.192653825822857e-05, "loss": 0.5426, "num_input_tokens_seen": 11729664, "step": 17920 }, { "epoch": 9.394654088050315, "grad_norm": 0.16813862323760986, "learning_rate": 3.1915550876551964e-05, "loss": 0.4747, "num_input_tokens_seen": 11733056, "step": 17925 }, { "epoch": 9.39727463312369, "grad_norm": 0.47025299072265625, "learning_rate": 3.190456204822101e-05, "loss": 0.4868, "num_input_tokens_seen": 11736128, "step": 17930 }, { "epoch": 9.399895178197065, "grad_norm": 0.2700330317020416, "learning_rate": 3.1893571775534445e-05, "loss": 0.6911, "num_input_tokens_seen": 11739456, "step": 17935 }, { "epoch": 9.40251572327044, "grad_norm": 0.15154461562633514, "learning_rate": 3.1882580060791315e-05, "loss": 0.4342, "num_input_tokens_seen": 11744256, "step": 17940 }, { "epoch": 9.405136268343815, "grad_norm": 0.15907247364521027, "learning_rate": 3.187158690629095e-05, "loss": 0.448, "num_input_tokens_seen": 11746976, "step": 17945 }, { "epoch": 9.40775681341719, "grad_norm": 0.09573126584291458, "learning_rate": 3.1860592314333e-05, "loss": 0.5582, "num_input_tokens_seen": 11749920, "step": 17950 }, { "epoch": 9.410377358490566, "grad_norm": 0.2761317193508148, "learning_rate": 3.184959628721742e-05, "loss": 0.5218, "num_input_tokens_seen": 11752896, "step": 17955 }, { "epoch": 9.41299790356394, "grad_norm": 0.22997459769248962, "learning_rate": 3.183859882724443e-05, "loss": 0.5531, "num_input_tokens_seen": 11755424, "step": 17960 }, { "epoch": 9.415618448637316, "grad_norm": 0.21361473202705383, "learning_rate": 3.1827599936714584e-05, "loss": 0.4861, "num_input_tokens_seen": 11758976, "step": 17965 }, { "epoch": 9.418238993710691, "grad_norm": 0.29493987560272217, "learning_rate": 3.1816599617928736e-05, "loss": 0.4156, "num_input_tokens_seen": 11761248, "step": 17970 }, { "epoch": 9.420859538784066, "grad_norm": 0.18308007717132568, "learning_rate": 3.1805597873188e-05, "loss": 0.4352, "num_input_tokens_seen": 11764288, "step": 17975 }, { "epoch": 9.423480083857442, "grad_norm": 0.13416120409965515, "learning_rate": 3.179459470479384e-05, "loss": 0.7006, "num_input_tokens_seen": 11767328, "step": 17980 }, { "epoch": 9.426100628930818, "grad_norm": 0.1325477510690689, "learning_rate": 3.1783590115047983e-05, "loss": 0.449, "num_input_tokens_seen": 11770656, "step": 17985 }, { "epoch": 9.428721174004194, "grad_norm": 0.17858995497226715, "learning_rate": 3.177258410625246e-05, "loss": 0.4808, "num_input_tokens_seen": 11773760, "step": 17990 }, { "epoch": 9.431341719077569, "grad_norm": 0.15805642306804657, "learning_rate": 3.176157668070962e-05, "loss": 0.4635, "num_input_tokens_seen": 11776480, "step": 17995 }, { "epoch": 9.433962264150944, "grad_norm": 0.1519612967967987, "learning_rate": 3.175056784072207e-05, "loss": 0.3889, "num_input_tokens_seen": 11780160, "step": 18000 }, { "epoch": 9.43658280922432, "grad_norm": 0.14684626460075378, "learning_rate": 3.173955758859274e-05, "loss": 0.4084, "num_input_tokens_seen": 11783200, "step": 18005 }, { "epoch": 9.439203354297694, "grad_norm": 0.1429576724767685, "learning_rate": 3.172854592662486e-05, "loss": 0.4542, "num_input_tokens_seen": 11786528, "step": 18010 }, { "epoch": 9.44182389937107, "grad_norm": 0.18347817659378052, "learning_rate": 3.1717532857121926e-05, "loss": 0.4681, "num_input_tokens_seen": 11789280, "step": 18015 }, { "epoch": 9.444444444444445, "grad_norm": 0.22017908096313477, "learning_rate": 3.170651838238776e-05, "loss": 0.5648, "num_input_tokens_seen": 11792608, "step": 18020 }, { "epoch": 9.44706498951782, "grad_norm": 0.17350079119205475, "learning_rate": 3.169550250472646e-05, "loss": 0.5737, "num_input_tokens_seen": 11794944, "step": 18025 }, { "epoch": 9.449685534591195, "grad_norm": 0.18109695613384247, "learning_rate": 3.168448522644243e-05, "loss": 0.4191, "num_input_tokens_seen": 11798208, "step": 18030 }, { "epoch": 9.45230607966457, "grad_norm": 0.10929538309574127, "learning_rate": 3.167346654984034e-05, "loss": 0.4127, "num_input_tokens_seen": 11801440, "step": 18035 }, { "epoch": 9.454926624737945, "grad_norm": 0.24438691139221191, "learning_rate": 3.166244647722519e-05, "loss": 0.502, "num_input_tokens_seen": 11804192, "step": 18040 }, { "epoch": 9.45754716981132, "grad_norm": 0.15916112065315247, "learning_rate": 3.1651425010902244e-05, "loss": 0.5587, "num_input_tokens_seen": 11807584, "step": 18045 }, { "epoch": 9.460167714884696, "grad_norm": 0.14363358914852142, "learning_rate": 3.164040215317706e-05, "loss": 0.4692, "num_input_tokens_seen": 11811392, "step": 18050 }, { "epoch": 9.46278825995807, "grad_norm": 0.11362944543361664, "learning_rate": 3.162937790635552e-05, "loss": 0.5009, "num_input_tokens_seen": 11815872, "step": 18055 }, { "epoch": 9.465408805031446, "grad_norm": 0.1103757694363594, "learning_rate": 3.1618352272743746e-05, "loss": 0.3863, "num_input_tokens_seen": 11818784, "step": 18060 }, { "epoch": 9.468029350104821, "grad_norm": 0.5418937802314758, "learning_rate": 3.160732525464818e-05, "loss": 0.6038, "num_input_tokens_seen": 11822624, "step": 18065 }, { "epoch": 9.470649895178196, "grad_norm": 0.14404605329036713, "learning_rate": 3.1596296854375554e-05, "loss": 0.5121, "num_input_tokens_seen": 11825312, "step": 18070 }, { "epoch": 9.473270440251572, "grad_norm": 0.3471035957336426, "learning_rate": 3.1585267074232884e-05, "loss": 0.5328, "num_input_tokens_seen": 11828224, "step": 18075 }, { "epoch": 9.475890985324948, "grad_norm": 0.1517089605331421, "learning_rate": 3.157423591652746e-05, "loss": 0.5719, "num_input_tokens_seen": 11831072, "step": 18080 }, { "epoch": 9.478511530398324, "grad_norm": 0.2700267434120178, "learning_rate": 3.1563203383566894e-05, "loss": 0.548, "num_input_tokens_seen": 11835168, "step": 18085 }, { "epoch": 9.481132075471699, "grad_norm": 0.17721349000930786, "learning_rate": 3.155216947765906e-05, "loss": 0.5535, "num_input_tokens_seen": 11838048, "step": 18090 }, { "epoch": 9.483752620545074, "grad_norm": 0.23866228759288788, "learning_rate": 3.154113420111211e-05, "loss": 0.6394, "num_input_tokens_seen": 11842048, "step": 18095 }, { "epoch": 9.48637316561845, "grad_norm": 0.13390323519706726, "learning_rate": 3.1530097556234526e-05, "loss": 0.5199, "num_input_tokens_seen": 11845856, "step": 18100 }, { "epoch": 9.488993710691824, "grad_norm": 0.1271452158689499, "learning_rate": 3.151905954533502e-05, "loss": 0.4584, "num_input_tokens_seen": 11849152, "step": 18105 }, { "epoch": 9.4916142557652, "grad_norm": 0.11141540110111237, "learning_rate": 3.150802017072263e-05, "loss": 0.4832, "num_input_tokens_seen": 11852096, "step": 18110 }, { "epoch": 9.494234800838575, "grad_norm": 0.15220776200294495, "learning_rate": 3.149697943470668e-05, "loss": 0.4333, "num_input_tokens_seen": 11854528, "step": 18115 }, { "epoch": 9.49685534591195, "grad_norm": 0.171585813164711, "learning_rate": 3.1485937339596736e-05, "loss": 0.5964, "num_input_tokens_seen": 11857440, "step": 18120 }, { "epoch": 9.499475890985325, "grad_norm": 0.23353970050811768, "learning_rate": 3.147489388770271e-05, "loss": 0.5766, "num_input_tokens_seen": 11860192, "step": 18125 }, { "epoch": 9.5020964360587, "grad_norm": 0.13613930344581604, "learning_rate": 3.146384908133475e-05, "loss": 0.5299, "num_input_tokens_seen": 11863104, "step": 18130 }, { "epoch": 9.504716981132075, "grad_norm": 0.23806078732013702, "learning_rate": 3.14528029228033e-05, "loss": 0.5505, "num_input_tokens_seen": 11865792, "step": 18135 }, { "epoch": 9.50733752620545, "grad_norm": 0.1639232188463211, "learning_rate": 3.144175541441911e-05, "loss": 0.4873, "num_input_tokens_seen": 11868384, "step": 18140 }, { "epoch": 9.509958071278826, "grad_norm": 0.10254448652267456, "learning_rate": 3.1430706558493166e-05, "loss": 0.434, "num_input_tokens_seen": 11871808, "step": 18145 }, { "epoch": 9.5125786163522, "grad_norm": 0.2557515799999237, "learning_rate": 3.141965635733677e-05, "loss": 0.6281, "num_input_tokens_seen": 11874368, "step": 18150 }, { "epoch": 9.515199161425576, "grad_norm": 0.18700630962848663, "learning_rate": 3.140860481326152e-05, "loss": 0.4301, "num_input_tokens_seen": 11877024, "step": 18155 }, { "epoch": 9.517819706498951, "grad_norm": 0.14286860823631287, "learning_rate": 3.139755192857925e-05, "loss": 0.467, "num_input_tokens_seen": 11880224, "step": 18160 }, { "epoch": 9.520440251572326, "grad_norm": 0.20511195063591003, "learning_rate": 3.1386497705602094e-05, "loss": 0.5065, "num_input_tokens_seen": 11883712, "step": 18165 }, { "epoch": 9.523060796645701, "grad_norm": 0.11856958270072937, "learning_rate": 3.137544214664249e-05, "loss": 0.3857, "num_input_tokens_seen": 11888384, "step": 18170 }, { "epoch": 9.525681341719078, "grad_norm": 0.15603771805763245, "learning_rate": 3.136438525401312e-05, "loss": 0.4884, "num_input_tokens_seen": 11891680, "step": 18175 }, { "epoch": 9.528301886792454, "grad_norm": 0.2615411877632141, "learning_rate": 3.135332703002696e-05, "loss": 0.4279, "num_input_tokens_seen": 11894592, "step": 18180 }, { "epoch": 9.530922431865829, "grad_norm": 0.1733647882938385, "learning_rate": 3.1342267476997276e-05, "loss": 0.5018, "num_input_tokens_seen": 11897536, "step": 18185 }, { "epoch": 9.533542976939204, "grad_norm": 0.11521989852190018, "learning_rate": 3.1331206597237584e-05, "loss": 0.4659, "num_input_tokens_seen": 11901440, "step": 18190 }, { "epoch": 9.536163522012579, "grad_norm": 0.1351686716079712, "learning_rate": 3.1320144393061715e-05, "loss": 0.5768, "num_input_tokens_seen": 11904736, "step": 18195 }, { "epoch": 9.538784067085954, "grad_norm": 0.11627396196126938, "learning_rate": 3.130908086678373e-05, "loss": 0.4297, "num_input_tokens_seen": 11907936, "step": 18200 }, { "epoch": 9.54140461215933, "grad_norm": 0.16890117526054382, "learning_rate": 3.1298016020718014e-05, "loss": 0.5244, "num_input_tokens_seen": 11911072, "step": 18205 }, { "epoch": 9.544025157232705, "grad_norm": 0.19962282478809357, "learning_rate": 3.128694985717919e-05, "loss": 0.4844, "num_input_tokens_seen": 11914432, "step": 18210 }, { "epoch": 9.54664570230608, "grad_norm": 0.44067031145095825, "learning_rate": 3.1275882378482185e-05, "loss": 0.5812, "num_input_tokens_seen": 11917344, "step": 18215 }, { "epoch": 9.549266247379455, "grad_norm": 0.10849950462579727, "learning_rate": 3.1264813586942185e-05, "loss": 0.4008, "num_input_tokens_seen": 11919968, "step": 18220 }, { "epoch": 9.55188679245283, "grad_norm": 0.14238779246807098, "learning_rate": 3.125374348487466e-05, "loss": 0.4055, "num_input_tokens_seen": 11923520, "step": 18225 }, { "epoch": 9.554507337526205, "grad_norm": 0.24649715423583984, "learning_rate": 3.1242672074595334e-05, "loss": 0.5196, "num_input_tokens_seen": 11926304, "step": 18230 }, { "epoch": 9.55712788259958, "grad_norm": 0.13153916597366333, "learning_rate": 3.123159935842024e-05, "loss": 0.6043, "num_input_tokens_seen": 11929248, "step": 18235 }, { "epoch": 9.559748427672956, "grad_norm": 0.16729454696178436, "learning_rate": 3.122052533866565e-05, "loss": 0.4897, "num_input_tokens_seen": 11934464, "step": 18240 }, { "epoch": 9.56236897274633, "grad_norm": 0.1412968933582306, "learning_rate": 3.120945001764813e-05, "loss": 0.4777, "num_input_tokens_seen": 11938560, "step": 18245 }, { "epoch": 9.564989517819706, "grad_norm": 0.13050934672355652, "learning_rate": 3.11983733976845e-05, "loss": 0.3826, "num_input_tokens_seen": 11941472, "step": 18250 }, { "epoch": 9.567610062893081, "grad_norm": 0.1527702361345291, "learning_rate": 3.1187295481091876e-05, "loss": 0.5549, "num_input_tokens_seen": 11944320, "step": 18255 }, { "epoch": 9.570230607966456, "grad_norm": 0.20430448651313782, "learning_rate": 3.1176216270187625e-05, "loss": 0.4224, "num_input_tokens_seen": 11947808, "step": 18260 }, { "epoch": 9.572851153039831, "grad_norm": 0.16569623351097107, "learning_rate": 3.116513576728939e-05, "loss": 0.5815, "num_input_tokens_seen": 11950944, "step": 18265 }, { "epoch": 9.575471698113208, "grad_norm": 0.19748207926750183, "learning_rate": 3.1154053974715085e-05, "loss": 0.5981, "num_input_tokens_seen": 11954208, "step": 18270 }, { "epoch": 9.578092243186584, "grad_norm": 0.09294426441192627, "learning_rate": 3.1142970894782906e-05, "loss": 0.5178, "num_input_tokens_seen": 11957728, "step": 18275 }, { "epoch": 9.580712788259959, "grad_norm": 0.13774453103542328, "learning_rate": 3.1131886529811296e-05, "loss": 0.7083, "num_input_tokens_seen": 11960544, "step": 18280 }, { "epoch": 9.583333333333334, "grad_norm": 0.144910529255867, "learning_rate": 3.1120800882118985e-05, "loss": 0.5275, "num_input_tokens_seen": 11963904, "step": 18285 }, { "epoch": 9.585953878406709, "grad_norm": 0.18886242806911469, "learning_rate": 3.1109713954024945e-05, "loss": 0.5135, "num_input_tokens_seen": 11966784, "step": 18290 }, { "epoch": 9.588574423480084, "grad_norm": 0.13164815306663513, "learning_rate": 3.109862574784846e-05, "loss": 0.417, "num_input_tokens_seen": 11970240, "step": 18295 }, { "epoch": 9.59119496855346, "grad_norm": 0.1373390406370163, "learning_rate": 3.108753626590905e-05, "loss": 0.4384, "num_input_tokens_seen": 11973760, "step": 18300 }, { "epoch": 9.593815513626835, "grad_norm": 0.13111820816993713, "learning_rate": 3.1076445510526486e-05, "loss": 0.4615, "num_input_tokens_seen": 11976512, "step": 18305 }, { "epoch": 9.59643605870021, "grad_norm": 0.13004004955291748, "learning_rate": 3.106535348402086e-05, "loss": 0.4893, "num_input_tokens_seen": 11979904, "step": 18310 }, { "epoch": 9.599056603773585, "grad_norm": 0.13395273685455322, "learning_rate": 3.105426018871247e-05, "loss": 0.5615, "num_input_tokens_seen": 11983232, "step": 18315 }, { "epoch": 9.60167714884696, "grad_norm": 0.2776253819465637, "learning_rate": 3.104316562692193e-05, "loss": 0.382, "num_input_tokens_seen": 11987392, "step": 18320 }, { "epoch": 9.604297693920335, "grad_norm": 0.15838408470153809, "learning_rate": 3.103206980097009e-05, "loss": 0.3787, "num_input_tokens_seen": 11991712, "step": 18325 }, { "epoch": 9.60691823899371, "grad_norm": 0.26149073243141174, "learning_rate": 3.102097271317805e-05, "loss": 0.4718, "num_input_tokens_seen": 11995040, "step": 18330 }, { "epoch": 9.609538784067086, "grad_norm": 0.16467513144016266, "learning_rate": 3.100987436586721e-05, "loss": 0.4611, "num_input_tokens_seen": 11998048, "step": 18335 }, { "epoch": 9.61215932914046, "grad_norm": 0.0353073850274086, "learning_rate": 3.099877476135923e-05, "loss": 0.3965, "num_input_tokens_seen": 12003200, "step": 18340 }, { "epoch": 9.614779874213836, "grad_norm": 0.180476114153862, "learning_rate": 3.0987673901976e-05, "loss": 0.4872, "num_input_tokens_seen": 12005728, "step": 18345 }, { "epoch": 9.617400419287211, "grad_norm": 0.11662006378173828, "learning_rate": 3.0976571790039697e-05, "loss": 0.4548, "num_input_tokens_seen": 12009280, "step": 18350 }, { "epoch": 9.620020964360586, "grad_norm": 0.15575933456420898, "learning_rate": 3.096546842787277e-05, "loss": 0.5501, "num_input_tokens_seen": 12013216, "step": 18355 }, { "epoch": 9.622641509433961, "grad_norm": 0.24128836393356323, "learning_rate": 3.095436381779789e-05, "loss": 0.4694, "num_input_tokens_seen": 12015712, "step": 18360 }, { "epoch": 9.625262054507338, "grad_norm": 0.1844276636838913, "learning_rate": 3.0943257962138045e-05, "loss": 0.5781, "num_input_tokens_seen": 12018688, "step": 18365 }, { "epoch": 9.627882599580714, "grad_norm": 0.2309160828590393, "learning_rate": 3.093215086321644e-05, "loss": 0.4316, "num_input_tokens_seen": 12022464, "step": 18370 }, { "epoch": 9.630503144654089, "grad_norm": 0.2203393131494522, "learning_rate": 3.092104252335654e-05, "loss": 0.464, "num_input_tokens_seen": 12025120, "step": 18375 }, { "epoch": 9.633123689727464, "grad_norm": 0.15983377397060394, "learning_rate": 3.09099329448821e-05, "loss": 0.5255, "num_input_tokens_seen": 12028064, "step": 18380 }, { "epoch": 9.635744234800839, "grad_norm": 0.17456254363059998, "learning_rate": 3.089882213011711e-05, "loss": 0.5497, "num_input_tokens_seen": 12031776, "step": 18385 }, { "epoch": 9.638364779874214, "grad_norm": 0.1519434154033661, "learning_rate": 3.088771008138583e-05, "loss": 0.5631, "num_input_tokens_seen": 12035008, "step": 18390 }, { "epoch": 9.64098532494759, "grad_norm": 0.12183224409818649, "learning_rate": 3.087659680101278e-05, "loss": 0.5234, "num_input_tokens_seen": 12039648, "step": 18395 }, { "epoch": 9.643605870020965, "grad_norm": 0.1626635193824768, "learning_rate": 3.086548229132271e-05, "loss": 0.6038, "num_input_tokens_seen": 12042304, "step": 18400 }, { "epoch": 9.64622641509434, "grad_norm": 0.17004752159118652, "learning_rate": 3.085436655464067e-05, "loss": 0.6048, "num_input_tokens_seen": 12045312, "step": 18405 }, { "epoch": 9.648846960167715, "grad_norm": 0.17253702878952026, "learning_rate": 3.084324959329193e-05, "loss": 0.4392, "num_input_tokens_seen": 12047936, "step": 18410 }, { "epoch": 9.65146750524109, "grad_norm": 0.1413055658340454, "learning_rate": 3.083213140960204e-05, "loss": 0.4093, "num_input_tokens_seen": 12050592, "step": 18415 }, { "epoch": 9.654088050314465, "grad_norm": 0.4203079640865326, "learning_rate": 3.08210120058968e-05, "loss": 0.6513, "num_input_tokens_seen": 12053312, "step": 18420 }, { "epoch": 9.65670859538784, "grad_norm": 0.1881040632724762, "learning_rate": 3.080989138450225e-05, "loss": 0.4469, "num_input_tokens_seen": 12055520, "step": 18425 }, { "epoch": 9.659329140461216, "grad_norm": 0.18506328761577606, "learning_rate": 3.0798769547744713e-05, "loss": 0.4712, "num_input_tokens_seen": 12058464, "step": 18430 }, { "epoch": 9.66194968553459, "grad_norm": 0.2033928483724594, "learning_rate": 3.0787646497950736e-05, "loss": 0.4884, "num_input_tokens_seen": 12061152, "step": 18435 }, { "epoch": 9.664570230607966, "grad_norm": 0.1764233261346817, "learning_rate": 3.0776522237447146e-05, "loss": 0.4548, "num_input_tokens_seen": 12063968, "step": 18440 }, { "epoch": 9.667190775681341, "grad_norm": 0.2313757687807083, "learning_rate": 3.076539676856101e-05, "loss": 0.4899, "num_input_tokens_seen": 12066816, "step": 18445 }, { "epoch": 9.669811320754716, "grad_norm": 0.19287535548210144, "learning_rate": 3.0754270093619634e-05, "loss": 0.5703, "num_input_tokens_seen": 12069536, "step": 18450 }, { "epoch": 9.672431865828091, "grad_norm": 0.45996254682540894, "learning_rate": 3.074314221495061e-05, "loss": 0.7478, "num_input_tokens_seen": 12072608, "step": 18455 }, { "epoch": 9.675052410901468, "grad_norm": 0.1629258096218109, "learning_rate": 3.0732013134881756e-05, "loss": 0.395, "num_input_tokens_seen": 12076576, "step": 18460 }, { "epoch": 9.677672955974844, "grad_norm": 0.3482743799686432, "learning_rate": 3.072088285574114e-05, "loss": 0.5385, "num_input_tokens_seen": 12079712, "step": 18465 }, { "epoch": 9.680293501048219, "grad_norm": 0.17115260660648346, "learning_rate": 3.070975137985711e-05, "loss": 0.4706, "num_input_tokens_seen": 12082944, "step": 18470 }, { "epoch": 9.682914046121594, "grad_norm": 0.17245972156524658, "learning_rate": 3.0698618709558226e-05, "loss": 0.5782, "num_input_tokens_seen": 12086752, "step": 18475 }, { "epoch": 9.685534591194969, "grad_norm": 0.17142458260059357, "learning_rate": 3.068748484717332e-05, "loss": 0.5049, "num_input_tokens_seen": 12089344, "step": 18480 }, { "epoch": 9.688155136268344, "grad_norm": 0.07913166284561157, "learning_rate": 3.067634979503147e-05, "loss": 0.5545, "num_input_tokens_seen": 12093024, "step": 18485 }, { "epoch": 9.69077568134172, "grad_norm": 0.10885442048311234, "learning_rate": 3.0665213555462e-05, "loss": 0.6292, "num_input_tokens_seen": 12096800, "step": 18490 }, { "epoch": 9.693396226415095, "grad_norm": 0.22749599814414978, "learning_rate": 3.0654076130794494e-05, "loss": 0.5106, "num_input_tokens_seen": 12099712, "step": 18495 }, { "epoch": 9.69601677148847, "grad_norm": 0.17744015157222748, "learning_rate": 3.064293752335876e-05, "loss": 0.5457, "num_input_tokens_seen": 12103328, "step": 18500 }, { "epoch": 9.698637316561845, "grad_norm": 0.1298147588968277, "learning_rate": 3.063179773548487e-05, "loss": 0.4045, "num_input_tokens_seen": 12106368, "step": 18505 }, { "epoch": 9.70125786163522, "grad_norm": 0.08531692624092102, "learning_rate": 3.0620656769503164e-05, "loss": 0.5173, "num_input_tokens_seen": 12110528, "step": 18510 }, { "epoch": 9.703878406708595, "grad_norm": 0.17893870174884796, "learning_rate": 3.060951462774417e-05, "loss": 0.4598, "num_input_tokens_seen": 12113312, "step": 18515 }, { "epoch": 9.70649895178197, "grad_norm": 0.2186276614665985, "learning_rate": 3.0598371312538714e-05, "loss": 0.3788, "num_input_tokens_seen": 12116736, "step": 18520 }, { "epoch": 9.709119496855346, "grad_norm": 0.2296259105205536, "learning_rate": 3.058722682621785e-05, "loss": 0.3978, "num_input_tokens_seen": 12120064, "step": 18525 }, { "epoch": 9.71174004192872, "grad_norm": 0.2001042515039444, "learning_rate": 3.0576081171112875e-05, "loss": 0.4965, "num_input_tokens_seen": 12123200, "step": 18530 }, { "epoch": 9.714360587002096, "grad_norm": 0.28133058547973633, "learning_rate": 3.056493434955534e-05, "loss": 0.6426, "num_input_tokens_seen": 12126784, "step": 18535 }, { "epoch": 9.716981132075471, "grad_norm": 0.45480185747146606, "learning_rate": 3.0553786363877027e-05, "loss": 0.4705, "num_input_tokens_seen": 12129280, "step": 18540 }, { "epoch": 9.719601677148846, "grad_norm": 0.1386568397283554, "learning_rate": 3.054263721640996e-05, "loss": 0.5601, "num_input_tokens_seen": 12132672, "step": 18545 }, { "epoch": 9.722222222222221, "grad_norm": 0.21915121376514435, "learning_rate": 3.053148690948642e-05, "loss": 0.4952, "num_input_tokens_seen": 12135584, "step": 18550 }, { "epoch": 9.724842767295598, "grad_norm": 0.18160030245780945, "learning_rate": 3.0520335445438935e-05, "loss": 0.522, "num_input_tokens_seen": 12137952, "step": 18555 }, { "epoch": 9.727463312368974, "grad_norm": 0.1461915820837021, "learning_rate": 3.0509182826600246e-05, "loss": 0.3872, "num_input_tokens_seen": 12141280, "step": 18560 }, { "epoch": 9.730083857442349, "grad_norm": 0.2564648985862732, "learning_rate": 3.049802905530337e-05, "loss": 0.4789, "num_input_tokens_seen": 12143872, "step": 18565 }, { "epoch": 9.732704402515724, "grad_norm": 0.17773008346557617, "learning_rate": 3.0486874133881533e-05, "loss": 0.4673, "num_input_tokens_seen": 12146912, "step": 18570 }, { "epoch": 9.735324947589099, "grad_norm": 0.19883403182029724, "learning_rate": 3.047571806466823e-05, "loss": 0.3967, "num_input_tokens_seen": 12150688, "step": 18575 }, { "epoch": 9.737945492662474, "grad_norm": 0.15580818057060242, "learning_rate": 3.0464560849997177e-05, "loss": 0.4901, "num_input_tokens_seen": 12154048, "step": 18580 }, { "epoch": 9.74056603773585, "grad_norm": 0.18707865476608276, "learning_rate": 3.0453402492202325e-05, "loss": 0.6271, "num_input_tokens_seen": 12157312, "step": 18585 }, { "epoch": 9.743186582809225, "grad_norm": 0.0947098582983017, "learning_rate": 3.0442242993617904e-05, "loss": 0.447, "num_input_tokens_seen": 12161472, "step": 18590 }, { "epoch": 9.7458071278826, "grad_norm": 0.18695496022701263, "learning_rate": 3.0431082356578334e-05, "loss": 0.5693, "num_input_tokens_seen": 12164384, "step": 18595 }, { "epoch": 9.748427672955975, "grad_norm": 0.13154295086860657, "learning_rate": 3.0419920583418294e-05, "loss": 0.4843, "num_input_tokens_seen": 12167872, "step": 18600 }, { "epoch": 9.75104821802935, "grad_norm": 0.11100075393915176, "learning_rate": 3.0408757676472694e-05, "loss": 0.5423, "num_input_tokens_seen": 12170752, "step": 18605 }, { "epoch": 9.753668763102725, "grad_norm": 0.21976789832115173, "learning_rate": 3.0397593638076705e-05, "loss": 0.5051, "num_input_tokens_seen": 12173728, "step": 18610 }, { "epoch": 9.7562893081761, "grad_norm": 0.13931356370449066, "learning_rate": 3.0386428470565702e-05, "loss": 0.4876, "num_input_tokens_seen": 12177024, "step": 18615 }, { "epoch": 9.758909853249476, "grad_norm": 0.18181212246418, "learning_rate": 3.0375262176275308e-05, "loss": 0.567, "num_input_tokens_seen": 12180192, "step": 18620 }, { "epoch": 9.76153039832285, "grad_norm": 0.07437620311975479, "learning_rate": 3.03640947575414e-05, "loss": 0.4593, "num_input_tokens_seen": 12183488, "step": 18625 }, { "epoch": 9.764150943396226, "grad_norm": 0.15119685232639313, "learning_rate": 3.0352926216700067e-05, "loss": 0.5784, "num_input_tokens_seen": 12186880, "step": 18630 }, { "epoch": 9.766771488469601, "grad_norm": 0.13659031689167023, "learning_rate": 3.034175655608763e-05, "loss": 0.3862, "num_input_tokens_seen": 12189376, "step": 18635 }, { "epoch": 9.769392033542976, "grad_norm": 0.24223193526268005, "learning_rate": 3.0330585778040677e-05, "loss": 0.5317, "num_input_tokens_seen": 12191616, "step": 18640 }, { "epoch": 9.772012578616351, "grad_norm": 0.12068372219800949, "learning_rate": 3.031941388489598e-05, "loss": 0.447, "num_input_tokens_seen": 12195744, "step": 18645 }, { "epoch": 9.774633123689728, "grad_norm": 0.17769107222557068, "learning_rate": 3.0308240878990584e-05, "loss": 0.5351, "num_input_tokens_seen": 12198784, "step": 18650 }, { "epoch": 9.777253668763104, "grad_norm": 0.19412094354629517, "learning_rate": 3.0297066762661773e-05, "loss": 0.5482, "num_input_tokens_seen": 12201408, "step": 18655 }, { "epoch": 9.779874213836479, "grad_norm": 0.15587934851646423, "learning_rate": 3.0285891538247015e-05, "loss": 0.5151, "num_input_tokens_seen": 12204448, "step": 18660 }, { "epoch": 9.782494758909854, "grad_norm": 0.17424632608890533, "learning_rate": 3.0274715208084053e-05, "loss": 0.4923, "num_input_tokens_seen": 12207712, "step": 18665 }, { "epoch": 9.785115303983229, "grad_norm": 0.28367912769317627, "learning_rate": 3.0263537774510847e-05, "loss": 0.4904, "num_input_tokens_seen": 12210624, "step": 18670 }, { "epoch": 9.787735849056604, "grad_norm": 0.1829339563846588, "learning_rate": 3.0252359239865585e-05, "loss": 0.5028, "num_input_tokens_seen": 12214112, "step": 18675 }, { "epoch": 9.79035639412998, "grad_norm": 0.11955905705690384, "learning_rate": 3.0241179606486704e-05, "loss": 0.5143, "num_input_tokens_seen": 12218656, "step": 18680 }, { "epoch": 9.792976939203355, "grad_norm": 0.14141148328781128, "learning_rate": 3.0229998876712827e-05, "loss": 0.4915, "num_input_tokens_seen": 12222112, "step": 18685 }, { "epoch": 9.79559748427673, "grad_norm": 0.18678267300128937, "learning_rate": 3.0218817052882858e-05, "loss": 0.4503, "num_input_tokens_seen": 12225216, "step": 18690 }, { "epoch": 9.798218029350105, "grad_norm": 0.17569991946220398, "learning_rate": 3.020763413733591e-05, "loss": 0.491, "num_input_tokens_seen": 12228192, "step": 18695 }, { "epoch": 9.80083857442348, "grad_norm": 0.12135322391986847, "learning_rate": 3.0196450132411302e-05, "loss": 0.3788, "num_input_tokens_seen": 12232512, "step": 18700 }, { "epoch": 9.803459119496855, "grad_norm": 0.20708492398262024, "learning_rate": 3.0185265040448616e-05, "loss": 0.5616, "num_input_tokens_seen": 12235584, "step": 18705 }, { "epoch": 9.80607966457023, "grad_norm": 0.33285391330718994, "learning_rate": 3.017407886378764e-05, "loss": 0.4613, "num_input_tokens_seen": 12237920, "step": 18710 }, { "epoch": 9.808700209643606, "grad_norm": 0.18893535435199738, "learning_rate": 3.016289160476839e-05, "loss": 0.515, "num_input_tokens_seen": 12240736, "step": 18715 }, { "epoch": 9.81132075471698, "grad_norm": 0.1684185266494751, "learning_rate": 3.0151703265731125e-05, "loss": 0.3509, "num_input_tokens_seen": 12243744, "step": 18720 }, { "epoch": 9.813941299790356, "grad_norm": 0.15545101463794708, "learning_rate": 3.0140513849016305e-05, "loss": 0.3877, "num_input_tokens_seen": 12248672, "step": 18725 }, { "epoch": 9.816561844863731, "grad_norm": 0.13172662258148193, "learning_rate": 3.0129323356964627e-05, "loss": 0.4667, "num_input_tokens_seen": 12252224, "step": 18730 }, { "epoch": 9.819182389937106, "grad_norm": 0.16972674429416656, "learning_rate": 3.0118131791917032e-05, "loss": 0.552, "num_input_tokens_seen": 12254624, "step": 18735 }, { "epoch": 9.821802935010481, "grad_norm": 0.15092642605304718, "learning_rate": 3.010693915621466e-05, "loss": 0.4115, "num_input_tokens_seen": 12258272, "step": 18740 }, { "epoch": 9.824423480083858, "grad_norm": 0.2674335241317749, "learning_rate": 3.009574545219887e-05, "loss": 0.5208, "num_input_tokens_seen": 12261088, "step": 18745 }, { "epoch": 9.827044025157234, "grad_norm": 0.17453783750534058, "learning_rate": 3.0084550682211272e-05, "loss": 0.546, "num_input_tokens_seen": 12264544, "step": 18750 }, { "epoch": 9.829664570230609, "grad_norm": 0.25600945949554443, "learning_rate": 3.0073354848593677e-05, "loss": 0.6203, "num_input_tokens_seen": 12267360, "step": 18755 }, { "epoch": 9.832285115303984, "grad_norm": 0.1745028793811798, "learning_rate": 3.0062157953688125e-05, "loss": 0.563, "num_input_tokens_seen": 12270720, "step": 18760 }, { "epoch": 9.834905660377359, "grad_norm": 0.14593848586082458, "learning_rate": 3.0050959999836887e-05, "loss": 0.4712, "num_input_tokens_seen": 12274112, "step": 18765 }, { "epoch": 9.837526205450734, "grad_norm": 0.24912337958812714, "learning_rate": 3.003976098938244e-05, "loss": 0.5615, "num_input_tokens_seen": 12277600, "step": 18770 }, { "epoch": 9.84014675052411, "grad_norm": 0.15050849318504333, "learning_rate": 3.0028560924667487e-05, "loss": 0.5319, "num_input_tokens_seen": 12280640, "step": 18775 }, { "epoch": 9.842767295597485, "grad_norm": 0.1903202086687088, "learning_rate": 3.001735980803496e-05, "loss": 0.6862, "num_input_tokens_seen": 12283776, "step": 18780 }, { "epoch": 9.84538784067086, "grad_norm": 0.1451442986726761, "learning_rate": 3.0006157641828004e-05, "loss": 0.3557, "num_input_tokens_seen": 12287392, "step": 18785 }, { "epoch": 9.848008385744235, "grad_norm": 0.1969505101442337, "learning_rate": 2.9994954428389977e-05, "loss": 0.6576, "num_input_tokens_seen": 12290208, "step": 18790 }, { "epoch": 9.85062893081761, "grad_norm": 0.16125395894050598, "learning_rate": 2.9983750170064472e-05, "loss": 0.5449, "num_input_tokens_seen": 12293120, "step": 18795 }, { "epoch": 9.853249475890985, "grad_norm": 0.1189185082912445, "learning_rate": 2.9972544869195292e-05, "loss": 0.4847, "num_input_tokens_seen": 12300096, "step": 18800 }, { "epoch": 9.85587002096436, "grad_norm": 0.1204293966293335, "learning_rate": 2.9961338528126447e-05, "loss": 0.4204, "num_input_tokens_seen": 12304864, "step": 18805 }, { "epoch": 9.858490566037736, "grad_norm": 0.14393797516822815, "learning_rate": 2.9950131149202192e-05, "loss": 0.3905, "num_input_tokens_seen": 12308512, "step": 18810 }, { "epoch": 9.86111111111111, "grad_norm": 0.19810114800930023, "learning_rate": 2.9938922734766973e-05, "loss": 0.4685, "num_input_tokens_seen": 12311232, "step": 18815 }, { "epoch": 9.863731656184486, "grad_norm": 0.12282344698905945, "learning_rate": 2.992771328716546e-05, "loss": 0.2962, "num_input_tokens_seen": 12314656, "step": 18820 }, { "epoch": 9.866352201257861, "grad_norm": 0.15197871625423431, "learning_rate": 2.991650280874255e-05, "loss": 0.5014, "num_input_tokens_seen": 12318336, "step": 18825 }, { "epoch": 9.868972746331236, "grad_norm": 0.23176704347133636, "learning_rate": 2.990529130184334e-05, "loss": 0.3523, "num_input_tokens_seen": 12321056, "step": 18830 }, { "epoch": 9.871593291404611, "grad_norm": 0.16061103343963623, "learning_rate": 2.989407876881316e-05, "loss": 0.4653, "num_input_tokens_seen": 12324640, "step": 18835 }, { "epoch": 9.874213836477988, "grad_norm": 0.2412039041519165, "learning_rate": 2.988286521199753e-05, "loss": 0.5319, "num_input_tokens_seen": 12327232, "step": 18840 }, { "epoch": 9.876834381551364, "grad_norm": 0.1411994993686676, "learning_rate": 2.9871650633742198e-05, "loss": 0.5911, "num_input_tokens_seen": 12330400, "step": 18845 }, { "epoch": 9.879454926624739, "grad_norm": 0.17094266414642334, "learning_rate": 2.9860435036393142e-05, "loss": 0.4735, "num_input_tokens_seen": 12333184, "step": 18850 }, { "epoch": 9.882075471698114, "grad_norm": 0.19151636958122253, "learning_rate": 2.9849218422296527e-05, "loss": 0.5651, "num_input_tokens_seen": 12335552, "step": 18855 }, { "epoch": 9.884696016771489, "grad_norm": 0.15950508415699005, "learning_rate": 2.983800079379873e-05, "loss": 0.5002, "num_input_tokens_seen": 12338848, "step": 18860 }, { "epoch": 9.887316561844864, "grad_norm": 0.17463871836662292, "learning_rate": 2.982678215324638e-05, "loss": 0.4602, "num_input_tokens_seen": 12341568, "step": 18865 }, { "epoch": 9.88993710691824, "grad_norm": 0.1958754062652588, "learning_rate": 2.9815562502986256e-05, "loss": 0.4718, "num_input_tokens_seen": 12344480, "step": 18870 }, { "epoch": 9.892557651991615, "grad_norm": 0.1489335596561432, "learning_rate": 2.9804341845365398e-05, "loss": 0.4862, "num_input_tokens_seen": 12347008, "step": 18875 }, { "epoch": 9.89517819706499, "grad_norm": 0.12246324121952057, "learning_rate": 2.9793120182731037e-05, "loss": 0.4487, "num_input_tokens_seen": 12350272, "step": 18880 }, { "epoch": 9.897798742138365, "grad_norm": 0.17289796471595764, "learning_rate": 2.9781897517430617e-05, "loss": 0.4428, "num_input_tokens_seen": 12352640, "step": 18885 }, { "epoch": 9.90041928721174, "grad_norm": 0.1927497833967209, "learning_rate": 2.977067385181179e-05, "loss": 0.5519, "num_input_tokens_seen": 12355488, "step": 18890 }, { "epoch": 9.903039832285115, "grad_norm": 0.17221969366073608, "learning_rate": 2.9759449188222426e-05, "loss": 0.4241, "num_input_tokens_seen": 12358336, "step": 18895 }, { "epoch": 9.90566037735849, "grad_norm": 0.2304305136203766, "learning_rate": 2.9748223529010582e-05, "loss": 0.5347, "num_input_tokens_seen": 12361952, "step": 18900 }, { "epoch": 9.908280922431866, "grad_norm": 0.3673957586288452, "learning_rate": 2.9736996876524546e-05, "loss": 0.5043, "num_input_tokens_seen": 12365472, "step": 18905 }, { "epoch": 9.91090146750524, "grad_norm": 0.21927949786186218, "learning_rate": 2.9725769233112816e-05, "loss": 0.4623, "num_input_tokens_seen": 12369792, "step": 18910 }, { "epoch": 9.913522012578616, "grad_norm": 0.20360572636127472, "learning_rate": 2.9714540601124074e-05, "loss": 0.4129, "num_input_tokens_seen": 12373248, "step": 18915 }, { "epoch": 9.916142557651991, "grad_norm": 0.1288377195596695, "learning_rate": 2.9703310982907218e-05, "loss": 0.4029, "num_input_tokens_seen": 12376224, "step": 18920 }, { "epoch": 9.918763102725366, "grad_norm": 0.2181953638792038, "learning_rate": 2.9692080380811365e-05, "loss": 0.5052, "num_input_tokens_seen": 12379424, "step": 18925 }, { "epoch": 9.921383647798741, "grad_norm": 0.1738995611667633, "learning_rate": 2.9680848797185828e-05, "loss": 0.4794, "num_input_tokens_seen": 12382784, "step": 18930 }, { "epoch": 9.924004192872118, "grad_norm": 0.15983127057552338, "learning_rate": 2.966961623438012e-05, "loss": 0.4348, "num_input_tokens_seen": 12386080, "step": 18935 }, { "epoch": 9.926624737945493, "grad_norm": 0.22121112048625946, "learning_rate": 2.9658382694743973e-05, "loss": 0.5759, "num_input_tokens_seen": 12389600, "step": 18940 }, { "epoch": 9.929245283018869, "grad_norm": 0.15716902911663055, "learning_rate": 2.9647148180627317e-05, "loss": 0.4311, "num_input_tokens_seen": 12392992, "step": 18945 }, { "epoch": 9.931865828092244, "grad_norm": 0.1552954614162445, "learning_rate": 2.9635912694380275e-05, "loss": 0.5007, "num_input_tokens_seen": 12396096, "step": 18950 }, { "epoch": 9.934486373165619, "grad_norm": 0.1782957911491394, "learning_rate": 2.962467623835319e-05, "loss": 0.4772, "num_input_tokens_seen": 12399488, "step": 18955 }, { "epoch": 9.937106918238994, "grad_norm": 0.18478456139564514, "learning_rate": 2.961343881489659e-05, "loss": 0.4423, "num_input_tokens_seen": 12403200, "step": 18960 }, { "epoch": 9.93972746331237, "grad_norm": 0.15630847215652466, "learning_rate": 2.9602200426361226e-05, "loss": 0.5806, "num_input_tokens_seen": 12406176, "step": 18965 }, { "epoch": 9.942348008385745, "grad_norm": 0.18225084245204926, "learning_rate": 2.9590961075098038e-05, "loss": 0.4835, "num_input_tokens_seen": 12408800, "step": 18970 }, { "epoch": 9.94496855345912, "grad_norm": 0.1322992593050003, "learning_rate": 2.957972076345817e-05, "loss": 0.4328, "num_input_tokens_seen": 12413216, "step": 18975 }, { "epoch": 9.947589098532495, "grad_norm": 0.22059595584869385, "learning_rate": 2.9568479493792967e-05, "loss": 0.4226, "num_input_tokens_seen": 12416352, "step": 18980 }, { "epoch": 9.95020964360587, "grad_norm": 0.14317141473293304, "learning_rate": 2.955723726845397e-05, "loss": 0.5065, "num_input_tokens_seen": 12419552, "step": 18985 }, { "epoch": 9.952830188679245, "grad_norm": 0.1699727475643158, "learning_rate": 2.9545994089792928e-05, "loss": 0.4988, "num_input_tokens_seen": 12423008, "step": 18990 }, { "epoch": 9.95545073375262, "grad_norm": 0.1593129187822342, "learning_rate": 2.9534749960161795e-05, "loss": 0.5844, "num_input_tokens_seen": 12426208, "step": 18995 }, { "epoch": 9.958071278825996, "grad_norm": 0.19632956385612488, "learning_rate": 2.9523504881912694e-05, "loss": 0.4057, "num_input_tokens_seen": 12429664, "step": 19000 }, { "epoch": 9.96069182389937, "grad_norm": 0.11522473394870758, "learning_rate": 2.9512258857397978e-05, "loss": 0.5103, "num_input_tokens_seen": 12432512, "step": 19005 }, { "epoch": 9.963312368972746, "grad_norm": 0.2200331836938858, "learning_rate": 2.9501011888970203e-05, "loss": 0.5054, "num_input_tokens_seen": 12435392, "step": 19010 }, { "epoch": 9.965932914046121, "grad_norm": 0.21567851305007935, "learning_rate": 2.9489763978982075e-05, "loss": 0.5839, "num_input_tokens_seen": 12437568, "step": 19015 }, { "epoch": 9.968553459119496, "grad_norm": 0.14456620812416077, "learning_rate": 2.9478515129786556e-05, "loss": 0.4056, "num_input_tokens_seen": 12440224, "step": 19020 }, { "epoch": 9.971174004192871, "grad_norm": 0.1764693707227707, "learning_rate": 2.946726534373676e-05, "loss": 0.583, "num_input_tokens_seen": 12443136, "step": 19025 }, { "epoch": 9.973794549266248, "grad_norm": 0.2780858874320984, "learning_rate": 2.945601462318601e-05, "loss": 0.4547, "num_input_tokens_seen": 12445888, "step": 19030 }, { "epoch": 9.976415094339622, "grad_norm": 0.15595413744449615, "learning_rate": 2.944476297048786e-05, "loss": 0.5732, "num_input_tokens_seen": 12448576, "step": 19035 }, { "epoch": 9.979035639412999, "grad_norm": 0.15689314901828766, "learning_rate": 2.9433510387995987e-05, "loss": 0.5455, "num_input_tokens_seen": 12451168, "step": 19040 }, { "epoch": 9.981656184486374, "grad_norm": 0.2119145542383194, "learning_rate": 2.9422256878064325e-05, "loss": 0.6951, "num_input_tokens_seen": 12456800, "step": 19045 }, { "epoch": 9.984276729559749, "grad_norm": 0.2196572721004486, "learning_rate": 2.9411002443046982e-05, "loss": 0.4125, "num_input_tokens_seen": 12460160, "step": 19050 }, { "epoch": 9.986897274633124, "grad_norm": 0.39056357741355896, "learning_rate": 2.9399747085298245e-05, "loss": 0.557, "num_input_tokens_seen": 12462272, "step": 19055 }, { "epoch": 9.9895178197065, "grad_norm": 0.17481574416160583, "learning_rate": 2.9388490807172614e-05, "loss": 0.4755, "num_input_tokens_seen": 12464992, "step": 19060 }, { "epoch": 9.992138364779874, "grad_norm": 0.2076602280139923, "learning_rate": 2.9377233611024775e-05, "loss": 0.4607, "num_input_tokens_seen": 12467744, "step": 19065 }, { "epoch": 9.99475890985325, "grad_norm": 0.18485233187675476, "learning_rate": 2.9365975499209596e-05, "loss": 0.4987, "num_input_tokens_seen": 12471456, "step": 19070 }, { "epoch": 9.997379454926625, "grad_norm": 0.14916668832302094, "learning_rate": 2.9354716474082154e-05, "loss": 0.5603, "num_input_tokens_seen": 12474304, "step": 19075 }, { "epoch": 10.0, "grad_norm": 0.4244973361492157, "learning_rate": 2.9343456537997714e-05, "loss": 0.4199, "num_input_tokens_seen": 12476552, "step": 19080 }, { "epoch": 10.0, "eval_loss": 0.4648926854133606, "eval_runtime": 9.9724, "eval_samples_per_second": 85.035, "eval_steps_per_second": 21.259, "num_input_tokens_seen": 12476552, "step": 19080 }, { "epoch": 10.002620545073375, "grad_norm": 0.19769296050071716, "learning_rate": 2.9332195693311714e-05, "loss": 0.4029, "num_input_tokens_seen": 12479208, "step": 19085 }, { "epoch": 10.00524109014675, "grad_norm": 0.17770768702030182, "learning_rate": 2.9320933942379803e-05, "loss": 0.5784, "num_input_tokens_seen": 12483080, "step": 19090 }, { "epoch": 10.007861635220126, "grad_norm": 0.14076808094978333, "learning_rate": 2.9309671287557805e-05, "loss": 0.7113, "num_input_tokens_seen": 12486696, "step": 19095 }, { "epoch": 10.0104821802935, "grad_norm": 0.12295357882976532, "learning_rate": 2.9298407731201737e-05, "loss": 0.4775, "num_input_tokens_seen": 12489224, "step": 19100 }, { "epoch": 10.013102725366876, "grad_norm": 0.11504720151424408, "learning_rate": 2.928714327566781e-05, "loss": 0.4783, "num_input_tokens_seen": 12492968, "step": 19105 }, { "epoch": 10.015723270440251, "grad_norm": 0.16370463371276855, "learning_rate": 2.9275877923312427e-05, "loss": 0.459, "num_input_tokens_seen": 12495976, "step": 19110 }, { "epoch": 10.018343815513626, "grad_norm": 0.1439773440361023, "learning_rate": 2.9264611676492165e-05, "loss": 0.3666, "num_input_tokens_seen": 12498696, "step": 19115 }, { "epoch": 10.020964360587001, "grad_norm": 0.18697179853916168, "learning_rate": 2.9253344537563782e-05, "loss": 0.4944, "num_input_tokens_seen": 12501224, "step": 19120 }, { "epoch": 10.023584905660377, "grad_norm": 0.12288036197423935, "learning_rate": 2.924207650888426e-05, "loss": 0.4524, "num_input_tokens_seen": 12504264, "step": 19125 }, { "epoch": 10.026205450733753, "grad_norm": 0.1540563553571701, "learning_rate": 2.9230807592810723e-05, "loss": 0.5011, "num_input_tokens_seen": 12507240, "step": 19130 }, { "epoch": 10.028825995807129, "grad_norm": 0.12625880539417267, "learning_rate": 2.92195377917005e-05, "loss": 0.3753, "num_input_tokens_seen": 12511304, "step": 19135 }, { "epoch": 10.031446540880504, "grad_norm": 0.14723476767539978, "learning_rate": 2.920826710791112e-05, "loss": 0.4377, "num_input_tokens_seen": 12513960, "step": 19140 }, { "epoch": 10.034067085953879, "grad_norm": 0.10365461558103561, "learning_rate": 2.919699554380026e-05, "loss": 0.4062, "num_input_tokens_seen": 12517704, "step": 19145 }, { "epoch": 10.036687631027254, "grad_norm": 0.15622730553150177, "learning_rate": 2.9185723101725825e-05, "loss": 0.3603, "num_input_tokens_seen": 12521320, "step": 19150 }, { "epoch": 10.03930817610063, "grad_norm": 0.08307527005672455, "learning_rate": 2.9174449784045867e-05, "loss": 0.4763, "num_input_tokens_seen": 12525448, "step": 19155 }, { "epoch": 10.041928721174004, "grad_norm": 0.1629151850938797, "learning_rate": 2.916317559311863e-05, "loss": 0.5316, "num_input_tokens_seen": 12529000, "step": 19160 }, { "epoch": 10.04454926624738, "grad_norm": 0.13206304609775543, "learning_rate": 2.9151900531302566e-05, "loss": 0.4653, "num_input_tokens_seen": 12531784, "step": 19165 }, { "epoch": 10.047169811320755, "grad_norm": 0.1318553388118744, "learning_rate": 2.9140624600956272e-05, "loss": 0.5635, "num_input_tokens_seen": 12535688, "step": 19170 }, { "epoch": 10.04979035639413, "grad_norm": 0.20037810504436493, "learning_rate": 2.912934780443855e-05, "loss": 0.4451, "num_input_tokens_seen": 12538472, "step": 19175 }, { "epoch": 10.052410901467505, "grad_norm": 0.24697774648666382, "learning_rate": 2.9118070144108388e-05, "loss": 0.7305, "num_input_tokens_seen": 12544936, "step": 19180 }, { "epoch": 10.05503144654088, "grad_norm": 0.22106987237930298, "learning_rate": 2.9106791622324924e-05, "loss": 0.4157, "num_input_tokens_seen": 12547816, "step": 19185 }, { "epoch": 10.057651991614255, "grad_norm": 0.13842222094535828, "learning_rate": 2.909551224144751e-05, "loss": 0.5071, "num_input_tokens_seen": 12551016, "step": 19190 }, { "epoch": 10.06027253668763, "grad_norm": 0.18285344541072845, "learning_rate": 2.9084232003835653e-05, "loss": 0.5646, "num_input_tokens_seen": 12554184, "step": 19195 }, { "epoch": 10.062893081761006, "grad_norm": 0.11307229101657867, "learning_rate": 2.9072950911849058e-05, "loss": 0.4972, "num_input_tokens_seen": 12557512, "step": 19200 }, { "epoch": 10.065513626834381, "grad_norm": 0.13850626349449158, "learning_rate": 2.90616689678476e-05, "loss": 0.6222, "num_input_tokens_seen": 12560904, "step": 19205 }, { "epoch": 10.068134171907756, "grad_norm": 0.2220844179391861, "learning_rate": 2.9050386174191337e-05, "loss": 0.5127, "num_input_tokens_seen": 12563368, "step": 19210 }, { "epoch": 10.070754716981131, "grad_norm": 0.1277497410774231, "learning_rate": 2.903910253324048e-05, "loss": 0.4241, "num_input_tokens_seen": 12566728, "step": 19215 }, { "epoch": 10.073375262054507, "grad_norm": 0.14897100627422333, "learning_rate": 2.9027818047355477e-05, "loss": 0.5225, "num_input_tokens_seen": 12570248, "step": 19220 }, { "epoch": 10.075995807127883, "grad_norm": 0.39054256677627563, "learning_rate": 2.9016532718896873e-05, "loss": 0.5908, "num_input_tokens_seen": 12572936, "step": 19225 }, { "epoch": 10.078616352201259, "grad_norm": 0.169655442237854, "learning_rate": 2.900524655022545e-05, "loss": 0.5829, "num_input_tokens_seen": 12576328, "step": 19230 }, { "epoch": 10.081236897274634, "grad_norm": 0.20015473663806915, "learning_rate": 2.899395954370215e-05, "loss": 0.4436, "num_input_tokens_seen": 12579208, "step": 19235 }, { "epoch": 10.083857442348009, "grad_norm": 0.1248365268111229, "learning_rate": 2.898267170168807e-05, "loss": 0.4671, "num_input_tokens_seen": 12581896, "step": 19240 }, { "epoch": 10.086477987421384, "grad_norm": 0.12770774960517883, "learning_rate": 2.8971383026544507e-05, "loss": 0.649, "num_input_tokens_seen": 12584840, "step": 19245 }, { "epoch": 10.08909853249476, "grad_norm": 0.1550632119178772, "learning_rate": 2.896009352063293e-05, "loss": 0.5609, "num_input_tokens_seen": 12588840, "step": 19250 }, { "epoch": 10.091719077568134, "grad_norm": 0.2158374786376953, "learning_rate": 2.894880318631496e-05, "loss": 0.5138, "num_input_tokens_seen": 12591752, "step": 19255 }, { "epoch": 10.09433962264151, "grad_norm": 0.27776050567626953, "learning_rate": 2.8937512025952417e-05, "loss": 0.5466, "num_input_tokens_seen": 12594504, "step": 19260 }, { "epoch": 10.096960167714885, "grad_norm": 0.0999152734875679, "learning_rate": 2.8926220041907283e-05, "loss": 0.4501, "num_input_tokens_seen": 12597544, "step": 19265 }, { "epoch": 10.09958071278826, "grad_norm": 0.13416196405887604, "learning_rate": 2.8914927236541706e-05, "loss": 0.4685, "num_input_tokens_seen": 12600488, "step": 19270 }, { "epoch": 10.102201257861635, "grad_norm": 0.18162333965301514, "learning_rate": 2.8903633612218012e-05, "loss": 0.4705, "num_input_tokens_seen": 12604040, "step": 19275 }, { "epoch": 10.10482180293501, "grad_norm": 0.17215414345264435, "learning_rate": 2.8892339171298703e-05, "loss": 0.6208, "num_input_tokens_seen": 12606984, "step": 19280 }, { "epoch": 10.107442348008385, "grad_norm": 0.18641020357608795, "learning_rate": 2.888104391614645e-05, "loss": 0.4621, "num_input_tokens_seen": 12609800, "step": 19285 }, { "epoch": 10.11006289308176, "grad_norm": 0.14242315292358398, "learning_rate": 2.886974784912408e-05, "loss": 0.4665, "num_input_tokens_seen": 12614024, "step": 19290 }, { "epoch": 10.112683438155136, "grad_norm": 0.24404028058052063, "learning_rate": 2.8858450972594612e-05, "loss": 0.5517, "num_input_tokens_seen": 12617480, "step": 19295 }, { "epoch": 10.115303983228511, "grad_norm": 0.23488038778305054, "learning_rate": 2.8847153288921226e-05, "loss": 0.5348, "num_input_tokens_seen": 12620904, "step": 19300 }, { "epoch": 10.117924528301886, "grad_norm": 0.1353810429573059, "learning_rate": 2.883585480046725e-05, "loss": 0.5166, "num_input_tokens_seen": 12623464, "step": 19305 }, { "epoch": 10.120545073375261, "grad_norm": 0.14900507032871246, "learning_rate": 2.8824555509596223e-05, "loss": 0.5988, "num_input_tokens_seen": 12626472, "step": 19310 }, { "epoch": 10.123165618448636, "grad_norm": 0.16454961895942688, "learning_rate": 2.8813255418671815e-05, "loss": 0.3125, "num_input_tokens_seen": 12629224, "step": 19315 }, { "epoch": 10.125786163522013, "grad_norm": 0.14865300059318542, "learning_rate": 2.8801954530057874e-05, "loss": 0.4963, "num_input_tokens_seen": 12633032, "step": 19320 }, { "epoch": 10.128406708595389, "grad_norm": 0.1538156419992447, "learning_rate": 2.879065284611842e-05, "loss": 0.4697, "num_input_tokens_seen": 12636072, "step": 19325 }, { "epoch": 10.131027253668764, "grad_norm": 0.16478131711483002, "learning_rate": 2.8779350369217634e-05, "loss": 0.4473, "num_input_tokens_seen": 12639400, "step": 19330 }, { "epoch": 10.133647798742139, "grad_norm": 0.5802121758460999, "learning_rate": 2.876804710171987e-05, "loss": 0.5515, "num_input_tokens_seen": 12642920, "step": 19335 }, { "epoch": 10.136268343815514, "grad_norm": 0.14125068485736847, "learning_rate": 2.8756743045989643e-05, "loss": 0.4387, "num_input_tokens_seen": 12646312, "step": 19340 }, { "epoch": 10.13888888888889, "grad_norm": 0.1681412309408188, "learning_rate": 2.8745438204391623e-05, "loss": 0.6012, "num_input_tokens_seen": 12648648, "step": 19345 }, { "epoch": 10.141509433962264, "grad_norm": 0.1401248276233673, "learning_rate": 2.8734132579290675e-05, "loss": 0.4527, "num_input_tokens_seen": 12651176, "step": 19350 }, { "epoch": 10.14412997903564, "grad_norm": 0.1631021499633789, "learning_rate": 2.8722826173051782e-05, "loss": 0.4679, "num_input_tokens_seen": 12653672, "step": 19355 }, { "epoch": 10.146750524109015, "grad_norm": 0.2611391544342041, "learning_rate": 2.871151898804012e-05, "loss": 0.5031, "num_input_tokens_seen": 12656744, "step": 19360 }, { "epoch": 10.14937106918239, "grad_norm": 0.14687974750995636, "learning_rate": 2.8700211026621043e-05, "loss": 0.5213, "num_input_tokens_seen": 12660168, "step": 19365 }, { "epoch": 10.151991614255765, "grad_norm": 0.12089991569519043, "learning_rate": 2.8688902291160025e-05, "loss": 0.6157, "num_input_tokens_seen": 12664424, "step": 19370 }, { "epoch": 10.15461215932914, "grad_norm": 0.18605884909629822, "learning_rate": 2.867759278402273e-05, "loss": 0.4366, "num_input_tokens_seen": 12667496, "step": 19375 }, { "epoch": 10.157232704402515, "grad_norm": 0.14896713197231293, "learning_rate": 2.8666282507574994e-05, "loss": 0.4716, "num_input_tokens_seen": 12671752, "step": 19380 }, { "epoch": 10.15985324947589, "grad_norm": 0.3527437746524811, "learning_rate": 2.8654971464182773e-05, "loss": 0.5259, "num_input_tokens_seen": 12674408, "step": 19385 }, { "epoch": 10.162473794549266, "grad_norm": 0.1721165031194687, "learning_rate": 2.8643659656212222e-05, "loss": 0.4182, "num_input_tokens_seen": 12676968, "step": 19390 }, { "epoch": 10.165094339622641, "grad_norm": 0.14092403650283813, "learning_rate": 2.8632347086029642e-05, "loss": 0.3523, "num_input_tokens_seen": 12680200, "step": 19395 }, { "epoch": 10.167714884696016, "grad_norm": 0.2181193083524704, "learning_rate": 2.8621033756001486e-05, "loss": 0.6862, "num_input_tokens_seen": 12684104, "step": 19400 }, { "epoch": 10.170335429769391, "grad_norm": 0.26787394285202026, "learning_rate": 2.8609719668494394e-05, "loss": 0.4264, "num_input_tokens_seen": 12687688, "step": 19405 }, { "epoch": 10.172955974842766, "grad_norm": 0.09258162975311279, "learning_rate": 2.8598404825875113e-05, "loss": 0.4244, "num_input_tokens_seen": 12690728, "step": 19410 }, { "epoch": 10.175576519916143, "grad_norm": 0.19886748492717743, "learning_rate": 2.8587089230510604e-05, "loss": 0.4257, "num_input_tokens_seen": 12694568, "step": 19415 }, { "epoch": 10.178197064989519, "grad_norm": 0.22861169278621674, "learning_rate": 2.857577288476795e-05, "loss": 0.5615, "num_input_tokens_seen": 12697768, "step": 19420 }, { "epoch": 10.180817610062894, "grad_norm": 0.14999458193778992, "learning_rate": 2.85644557910144e-05, "loss": 0.5038, "num_input_tokens_seen": 12700840, "step": 19425 }, { "epoch": 10.183438155136269, "grad_norm": 0.1624756157398224, "learning_rate": 2.8553137951617366e-05, "loss": 0.5074, "num_input_tokens_seen": 12703784, "step": 19430 }, { "epoch": 10.186058700209644, "grad_norm": 0.11689610034227371, "learning_rate": 2.854181936894441e-05, "loss": 0.4399, "num_input_tokens_seen": 12707240, "step": 19435 }, { "epoch": 10.18867924528302, "grad_norm": 0.14181287586688995, "learning_rate": 2.8530500045363247e-05, "loss": 0.3836, "num_input_tokens_seen": 12710312, "step": 19440 }, { "epoch": 10.191299790356394, "grad_norm": 0.1499454230070114, "learning_rate": 2.851917998324175e-05, "loss": 0.5941, "num_input_tokens_seen": 12713640, "step": 19445 }, { "epoch": 10.19392033542977, "grad_norm": 0.24802590906620026, "learning_rate": 2.8507859184947957e-05, "loss": 0.413, "num_input_tokens_seen": 12716616, "step": 19450 }, { "epoch": 10.196540880503145, "grad_norm": 0.14265204966068268, "learning_rate": 2.8496537652850043e-05, "loss": 0.4444, "num_input_tokens_seen": 12719528, "step": 19455 }, { "epoch": 10.19916142557652, "grad_norm": 0.22482001781463623, "learning_rate": 2.8485215389316333e-05, "loss": 0.4003, "num_input_tokens_seen": 12722952, "step": 19460 }, { "epoch": 10.201781970649895, "grad_norm": 0.09858239442110062, "learning_rate": 2.8473892396715328e-05, "loss": 0.4342, "num_input_tokens_seen": 12726440, "step": 19465 }, { "epoch": 10.20440251572327, "grad_norm": 0.19440080225467682, "learning_rate": 2.846256867741567e-05, "loss": 0.5395, "num_input_tokens_seen": 12729096, "step": 19470 }, { "epoch": 10.207023060796645, "grad_norm": 0.3096555471420288, "learning_rate": 2.845124423378614e-05, "loss": 0.5453, "num_input_tokens_seen": 12731976, "step": 19475 }, { "epoch": 10.20964360587002, "grad_norm": 0.24211490154266357, "learning_rate": 2.8439919068195697e-05, "loss": 0.4987, "num_input_tokens_seen": 12734696, "step": 19480 }, { "epoch": 10.212264150943396, "grad_norm": 0.18803369998931885, "learning_rate": 2.842859318301343e-05, "loss": 0.4963, "num_input_tokens_seen": 12738184, "step": 19485 }, { "epoch": 10.214884696016771, "grad_norm": 0.12356162816286087, "learning_rate": 2.841726658060858e-05, "loss": 0.3742, "num_input_tokens_seen": 12741224, "step": 19490 }, { "epoch": 10.217505241090146, "grad_norm": 0.1396147906780243, "learning_rate": 2.8405939263350555e-05, "loss": 0.4513, "num_input_tokens_seen": 12743976, "step": 19495 }, { "epoch": 10.220125786163521, "grad_norm": 0.17977742850780487, "learning_rate": 2.8394611233608886e-05, "loss": 0.566, "num_input_tokens_seen": 12747464, "step": 19500 }, { "epoch": 10.222746331236896, "grad_norm": 0.1900934875011444, "learning_rate": 2.8383282493753283e-05, "loss": 0.5199, "num_input_tokens_seen": 12750440, "step": 19505 }, { "epoch": 10.225366876310272, "grad_norm": 0.17050807178020477, "learning_rate": 2.8371953046153575e-05, "loss": 0.5568, "num_input_tokens_seen": 12753448, "step": 19510 }, { "epoch": 10.227987421383649, "grad_norm": 0.14606446027755737, "learning_rate": 2.8360622893179757e-05, "loss": 0.5, "num_input_tokens_seen": 12757352, "step": 19515 }, { "epoch": 10.230607966457024, "grad_norm": 0.09395245462656021, "learning_rate": 2.834929203720197e-05, "loss": 0.3564, "num_input_tokens_seen": 12760968, "step": 19520 }, { "epoch": 10.233228511530399, "grad_norm": 0.2876473069190979, "learning_rate": 2.83379604805905e-05, "loss": 0.5073, "num_input_tokens_seen": 12764008, "step": 19525 }, { "epoch": 10.235849056603774, "grad_norm": 0.1436389684677124, "learning_rate": 2.832662822571578e-05, "loss": 0.6503, "num_input_tokens_seen": 12766600, "step": 19530 }, { "epoch": 10.23846960167715, "grad_norm": 0.17449475824832916, "learning_rate": 2.8315295274948395e-05, "loss": 0.4294, "num_input_tokens_seen": 12769608, "step": 19535 }, { "epoch": 10.241090146750524, "grad_norm": 0.11428852379322052, "learning_rate": 2.8303961630659047e-05, "loss": 0.5443, "num_input_tokens_seen": 12773608, "step": 19540 }, { "epoch": 10.2437106918239, "grad_norm": 0.13318271934986115, "learning_rate": 2.8292627295218627e-05, "loss": 0.4359, "num_input_tokens_seen": 12776296, "step": 19545 }, { "epoch": 10.246331236897275, "grad_norm": 0.13153842091560364, "learning_rate": 2.828129227099815e-05, "loss": 0.436, "num_input_tokens_seen": 12778984, "step": 19550 }, { "epoch": 10.24895178197065, "grad_norm": 0.24838502705097198, "learning_rate": 2.826995656036875e-05, "loss": 0.5289, "num_input_tokens_seen": 12782248, "step": 19555 }, { "epoch": 10.251572327044025, "grad_norm": 0.14119571447372437, "learning_rate": 2.8258620165701747e-05, "loss": 0.4258, "num_input_tokens_seen": 12785032, "step": 19560 }, { "epoch": 10.2541928721174, "grad_norm": 0.1344161480665207, "learning_rate": 2.8247283089368586e-05, "loss": 0.5095, "num_input_tokens_seen": 12787848, "step": 19565 }, { "epoch": 10.256813417190775, "grad_norm": 0.24221540987491608, "learning_rate": 2.8235945333740842e-05, "loss": 0.4069, "num_input_tokens_seen": 12790440, "step": 19570 }, { "epoch": 10.25943396226415, "grad_norm": 0.2921071946620941, "learning_rate": 2.822460690119026e-05, "loss": 0.67, "num_input_tokens_seen": 12793032, "step": 19575 }, { "epoch": 10.262054507337526, "grad_norm": 0.10920301079750061, "learning_rate": 2.82132677940887e-05, "loss": 0.6253, "num_input_tokens_seen": 12796392, "step": 19580 }, { "epoch": 10.264675052410901, "grad_norm": 0.2298116534948349, "learning_rate": 2.820192801480817e-05, "loss": 0.5305, "num_input_tokens_seen": 12799240, "step": 19585 }, { "epoch": 10.267295597484276, "grad_norm": 0.18515080213546753, "learning_rate": 2.8190587565720842e-05, "loss": 0.477, "num_input_tokens_seen": 12803752, "step": 19590 }, { "epoch": 10.269916142557651, "grad_norm": 0.16510441899299622, "learning_rate": 2.8179246449198986e-05, "loss": 0.3922, "num_input_tokens_seen": 12806568, "step": 19595 }, { "epoch": 10.272536687631026, "grad_norm": 0.2725541293621063, "learning_rate": 2.8167904667615048e-05, "loss": 0.4018, "num_input_tokens_seen": 12811464, "step": 19600 }, { "epoch": 10.275157232704402, "grad_norm": 0.11741260439157486, "learning_rate": 2.8156562223341592e-05, "loss": 0.5647, "num_input_tokens_seen": 12815880, "step": 19605 }, { "epoch": 10.277777777777779, "grad_norm": 0.3453701436519623, "learning_rate": 2.814521911875133e-05, "loss": 0.5416, "num_input_tokens_seen": 12818632, "step": 19610 }, { "epoch": 10.280398322851154, "grad_norm": 0.0890776664018631, "learning_rate": 2.8133875356217114e-05, "loss": 0.3963, "num_input_tokens_seen": 12821736, "step": 19615 }, { "epoch": 10.283018867924529, "grad_norm": 0.13229449093341827, "learning_rate": 2.8122530938111923e-05, "loss": 0.5181, "num_input_tokens_seen": 12824584, "step": 19620 }, { "epoch": 10.285639412997904, "grad_norm": 0.17191079258918762, "learning_rate": 2.8111185866808888e-05, "loss": 0.5676, "num_input_tokens_seen": 12827624, "step": 19625 }, { "epoch": 10.28825995807128, "grad_norm": 0.17598839104175568, "learning_rate": 2.809984014468126e-05, "loss": 0.5046, "num_input_tokens_seen": 12830440, "step": 19630 }, { "epoch": 10.290880503144654, "grad_norm": 0.11463375389575958, "learning_rate": 2.8088493774102443e-05, "loss": 0.3941, "num_input_tokens_seen": 12836456, "step": 19635 }, { "epoch": 10.29350104821803, "grad_norm": 0.2222367525100708, "learning_rate": 2.807714675744596e-05, "loss": 0.6201, "num_input_tokens_seen": 12839432, "step": 19640 }, { "epoch": 10.296121593291405, "grad_norm": 0.21465131640434265, "learning_rate": 2.8065799097085475e-05, "loss": 0.3968, "num_input_tokens_seen": 12843624, "step": 19645 }, { "epoch": 10.29874213836478, "grad_norm": 0.2956957221031189, "learning_rate": 2.8054450795394804e-05, "loss": 0.45, "num_input_tokens_seen": 12845960, "step": 19650 }, { "epoch": 10.301362683438155, "grad_norm": 0.17816407978534698, "learning_rate": 2.8043101854747878e-05, "loss": 0.4719, "num_input_tokens_seen": 12850216, "step": 19655 }, { "epoch": 10.30398322851153, "grad_norm": 0.22935473918914795, "learning_rate": 2.8031752277518748e-05, "loss": 0.4562, "num_input_tokens_seen": 12853064, "step": 19660 }, { "epoch": 10.306603773584905, "grad_norm": 0.17933091521263123, "learning_rate": 2.802040206608164e-05, "loss": 0.439, "num_input_tokens_seen": 12855912, "step": 19665 }, { "epoch": 10.30922431865828, "grad_norm": 0.15540379285812378, "learning_rate": 2.8009051222810877e-05, "loss": 0.4698, "num_input_tokens_seen": 12858824, "step": 19670 }, { "epoch": 10.311844863731656, "grad_norm": 0.24578744173049927, "learning_rate": 2.7997699750080918e-05, "loss": 0.6463, "num_input_tokens_seen": 12861416, "step": 19675 }, { "epoch": 10.314465408805031, "grad_norm": 0.19936341047286987, "learning_rate": 2.7986347650266386e-05, "loss": 0.4884, "num_input_tokens_seen": 12864968, "step": 19680 }, { "epoch": 10.317085953878406, "grad_norm": 0.23463289439678192, "learning_rate": 2.797499492574199e-05, "loss": 0.474, "num_input_tokens_seen": 12868040, "step": 19685 }, { "epoch": 10.319706498951781, "grad_norm": 0.1201639398932457, "learning_rate": 2.7963641578882594e-05, "loss": 0.5062, "num_input_tokens_seen": 12871336, "step": 19690 }, { "epoch": 10.322327044025156, "grad_norm": 0.21156637370586395, "learning_rate": 2.7952287612063193e-05, "loss": 0.4344, "num_input_tokens_seen": 12875336, "step": 19695 }, { "epoch": 10.324947589098532, "grad_norm": 0.16661567986011505, "learning_rate": 2.79409330276589e-05, "loss": 0.4954, "num_input_tokens_seen": 12878696, "step": 19700 }, { "epoch": 10.327568134171909, "grad_norm": 0.07395590841770172, "learning_rate": 2.7929577828044974e-05, "loss": 0.5118, "num_input_tokens_seen": 12882600, "step": 19705 }, { "epoch": 10.330188679245284, "grad_norm": 0.1628849059343338, "learning_rate": 2.7918222015596784e-05, "loss": 0.4291, "num_input_tokens_seen": 12885192, "step": 19710 }, { "epoch": 10.332809224318659, "grad_norm": 0.24404531717300415, "learning_rate": 2.7906865592689847e-05, "loss": 0.3934, "num_input_tokens_seen": 12888040, "step": 19715 }, { "epoch": 10.335429769392034, "grad_norm": 0.2152109593153, "learning_rate": 2.789550856169979e-05, "loss": 0.55, "num_input_tokens_seen": 12890536, "step": 19720 }, { "epoch": 10.33805031446541, "grad_norm": 0.23128840327262878, "learning_rate": 2.7884150925002374e-05, "loss": 0.4805, "num_input_tokens_seen": 12892872, "step": 19725 }, { "epoch": 10.340670859538784, "grad_norm": 0.110042504966259, "learning_rate": 2.787279268497349e-05, "loss": 0.4389, "num_input_tokens_seen": 12896008, "step": 19730 }, { "epoch": 10.34329140461216, "grad_norm": 0.2272186577320099, "learning_rate": 2.786143384398915e-05, "loss": 0.3831, "num_input_tokens_seen": 12899464, "step": 19735 }, { "epoch": 10.345911949685535, "grad_norm": 0.1810617595911026, "learning_rate": 2.785007440442549e-05, "loss": 0.5034, "num_input_tokens_seen": 12902312, "step": 19740 }, { "epoch": 10.34853249475891, "grad_norm": 0.17343398928642273, "learning_rate": 2.7838714368658785e-05, "loss": 0.403, "num_input_tokens_seen": 12906760, "step": 19745 }, { "epoch": 10.351153039832285, "grad_norm": 0.21203581988811493, "learning_rate": 2.7827353739065427e-05, "loss": 0.5612, "num_input_tokens_seen": 12909576, "step": 19750 }, { "epoch": 10.35377358490566, "grad_norm": 0.1280868947505951, "learning_rate": 2.781599251802191e-05, "loss": 0.3713, "num_input_tokens_seen": 12912808, "step": 19755 }, { "epoch": 10.356394129979035, "grad_norm": 0.12328823655843735, "learning_rate": 2.780463070790489e-05, "loss": 0.343, "num_input_tokens_seen": 12916552, "step": 19760 }, { "epoch": 10.35901467505241, "grad_norm": 0.0756913423538208, "learning_rate": 2.7793268311091125e-05, "loss": 0.4821, "num_input_tokens_seen": 12919368, "step": 19765 }, { "epoch": 10.361635220125786, "grad_norm": 0.1651325523853302, "learning_rate": 2.7781905329957496e-05, "loss": 0.3405, "num_input_tokens_seen": 12923976, "step": 19770 }, { "epoch": 10.364255765199161, "grad_norm": 0.13729722797870636, "learning_rate": 2.7770541766881008e-05, "loss": 0.5078, "num_input_tokens_seen": 12926728, "step": 19775 }, { "epoch": 10.366876310272536, "grad_norm": 0.1553996056318283, "learning_rate": 2.7759177624238786e-05, "loss": 0.4265, "num_input_tokens_seen": 12931176, "step": 19780 }, { "epoch": 10.369496855345911, "grad_norm": 0.2361883521080017, "learning_rate": 2.7747812904408094e-05, "loss": 0.5819, "num_input_tokens_seen": 12934088, "step": 19785 }, { "epoch": 10.372117400419286, "grad_norm": 0.3972744643688202, "learning_rate": 2.7736447609766286e-05, "loss": 0.4199, "num_input_tokens_seen": 12936936, "step": 19790 }, { "epoch": 10.374737945492662, "grad_norm": 0.1746034175157547, "learning_rate": 2.772508174269085e-05, "loss": 0.4617, "num_input_tokens_seen": 12940968, "step": 19795 }, { "epoch": 10.377358490566039, "grad_norm": 0.26788437366485596, "learning_rate": 2.7713715305559414e-05, "loss": 0.5981, "num_input_tokens_seen": 12944840, "step": 19800 }, { "epoch": 10.379979035639414, "grad_norm": 0.14645753800868988, "learning_rate": 2.7702348300749692e-05, "loss": 0.4365, "num_input_tokens_seen": 12947848, "step": 19805 }, { "epoch": 10.382599580712789, "grad_norm": 0.12484229356050491, "learning_rate": 2.7690980730639533e-05, "loss": 0.4775, "num_input_tokens_seen": 12951496, "step": 19810 }, { "epoch": 10.385220125786164, "grad_norm": 0.12775260210037231, "learning_rate": 2.7679612597606902e-05, "loss": 0.414, "num_input_tokens_seen": 12955272, "step": 19815 }, { "epoch": 10.38784067085954, "grad_norm": 0.25702011585235596, "learning_rate": 2.7668243904029894e-05, "loss": 0.4205, "num_input_tokens_seen": 12958376, "step": 19820 }, { "epoch": 10.390461215932914, "grad_norm": 0.14706717431545258, "learning_rate": 2.76568746522867e-05, "loss": 0.4691, "num_input_tokens_seen": 12961192, "step": 19825 }, { "epoch": 10.39308176100629, "grad_norm": 0.12383750826120377, "learning_rate": 2.7645504844755637e-05, "loss": 0.4633, "num_input_tokens_seen": 12965256, "step": 19830 }, { "epoch": 10.395702306079665, "grad_norm": 0.18819458782672882, "learning_rate": 2.7634134483815143e-05, "loss": 0.6998, "num_input_tokens_seen": 12968456, "step": 19835 }, { "epoch": 10.39832285115304, "grad_norm": 0.18845538794994354, "learning_rate": 2.762276357184377e-05, "loss": 0.5293, "num_input_tokens_seen": 12972104, "step": 19840 }, { "epoch": 10.400943396226415, "grad_norm": 0.21809379756450653, "learning_rate": 2.7611392111220176e-05, "loss": 0.5545, "num_input_tokens_seen": 12975752, "step": 19845 }, { "epoch": 10.40356394129979, "grad_norm": 0.13032720983028412, "learning_rate": 2.7600020104323154e-05, "loss": 0.4835, "num_input_tokens_seen": 12979656, "step": 19850 }, { "epoch": 10.406184486373165, "grad_norm": 0.2084532082080841, "learning_rate": 2.7588647553531576e-05, "loss": 0.5563, "num_input_tokens_seen": 12982696, "step": 19855 }, { "epoch": 10.40880503144654, "grad_norm": 0.14895454049110413, "learning_rate": 2.7577274461224463e-05, "loss": 0.4383, "num_input_tokens_seen": 12985576, "step": 19860 }, { "epoch": 10.411425576519916, "grad_norm": 0.1494414359331131, "learning_rate": 2.756590082978095e-05, "loss": 0.4186, "num_input_tokens_seen": 12988904, "step": 19865 }, { "epoch": 10.414046121593291, "grad_norm": 0.29683858156204224, "learning_rate": 2.7554526661580243e-05, "loss": 0.3804, "num_input_tokens_seen": 12992520, "step": 19870 }, { "epoch": 10.416666666666666, "grad_norm": 0.1495324671268463, "learning_rate": 2.7543151959001708e-05, "loss": 0.411, "num_input_tokens_seen": 12996040, "step": 19875 }, { "epoch": 10.419287211740041, "grad_norm": 0.2140336036682129, "learning_rate": 2.7531776724424797e-05, "loss": 0.6532, "num_input_tokens_seen": 12999848, "step": 19880 }, { "epoch": 10.421907756813416, "grad_norm": 0.11727846413850784, "learning_rate": 2.752040096022907e-05, "loss": 0.5217, "num_input_tokens_seen": 13003144, "step": 19885 }, { "epoch": 10.424528301886792, "grad_norm": 0.18141017854213715, "learning_rate": 2.750902466879423e-05, "loss": 0.3811, "num_input_tokens_seen": 13005512, "step": 19890 }, { "epoch": 10.427148846960169, "grad_norm": 0.15814082324504852, "learning_rate": 2.7497647852500048e-05, "loss": 0.5806, "num_input_tokens_seen": 13008680, "step": 19895 }, { "epoch": 10.429769392033544, "grad_norm": 0.16864965856075287, "learning_rate": 2.7486270513726424e-05, "loss": 0.6091, "num_input_tokens_seen": 13011688, "step": 19900 }, { "epoch": 10.432389937106919, "grad_norm": 0.084623783826828, "learning_rate": 2.7474892654853384e-05, "loss": 0.4291, "num_input_tokens_seen": 13014664, "step": 19905 }, { "epoch": 10.435010482180294, "grad_norm": 0.1414581686258316, "learning_rate": 2.7463514278261034e-05, "loss": 0.4934, "num_input_tokens_seen": 13019112, "step": 19910 }, { "epoch": 10.43763102725367, "grad_norm": 0.29981157183647156, "learning_rate": 2.7452135386329596e-05, "loss": 0.6666, "num_input_tokens_seen": 13021832, "step": 19915 }, { "epoch": 10.440251572327044, "grad_norm": 0.1324511617422104, "learning_rate": 2.7440755981439414e-05, "loss": 0.5455, "num_input_tokens_seen": 13025128, "step": 19920 }, { "epoch": 10.44287211740042, "grad_norm": 0.1500314474105835, "learning_rate": 2.7429376065970924e-05, "loss": 0.5387, "num_input_tokens_seen": 13028072, "step": 19925 }, { "epoch": 10.445492662473795, "grad_norm": 0.19019319117069244, "learning_rate": 2.7417995642304678e-05, "loss": 0.5424, "num_input_tokens_seen": 13031272, "step": 19930 }, { "epoch": 10.44811320754717, "grad_norm": 0.17738443613052368, "learning_rate": 2.7406614712821333e-05, "loss": 0.3996, "num_input_tokens_seen": 13034120, "step": 19935 }, { "epoch": 10.450733752620545, "grad_norm": 0.13797920942306519, "learning_rate": 2.739523327990164e-05, "loss": 0.5483, "num_input_tokens_seen": 13037256, "step": 19940 }, { "epoch": 10.45335429769392, "grad_norm": 0.21351172029972076, "learning_rate": 2.7383851345926477e-05, "loss": 0.427, "num_input_tokens_seen": 13040008, "step": 19945 }, { "epoch": 10.455974842767295, "grad_norm": 0.10207636654376984, "learning_rate": 2.7372468913276804e-05, "loss": 0.4719, "num_input_tokens_seen": 13043944, "step": 19950 }, { "epoch": 10.45859538784067, "grad_norm": 0.13436844944953918, "learning_rate": 2.7361085984333705e-05, "loss": 0.5295, "num_input_tokens_seen": 13046408, "step": 19955 }, { "epoch": 10.461215932914046, "grad_norm": 0.2331111878156662, "learning_rate": 2.7349702561478356e-05, "loss": 0.4469, "num_input_tokens_seen": 13049832, "step": 19960 }, { "epoch": 10.463836477987421, "grad_norm": 0.10370539873838425, "learning_rate": 2.733831864709203e-05, "loss": 0.4857, "num_input_tokens_seen": 13053640, "step": 19965 }, { "epoch": 10.466457023060796, "grad_norm": 0.17445743083953857, "learning_rate": 2.732693424355613e-05, "loss": 0.4831, "num_input_tokens_seen": 13056680, "step": 19970 }, { "epoch": 10.469077568134171, "grad_norm": 0.16292612254619598, "learning_rate": 2.731554935325213e-05, "loss": 0.4476, "num_input_tokens_seen": 13059560, "step": 19975 }, { "epoch": 10.471698113207546, "grad_norm": 0.19376875460147858, "learning_rate": 2.7304163978561613e-05, "loss": 0.4957, "num_input_tokens_seen": 13062856, "step": 19980 }, { "epoch": 10.474318658280922, "grad_norm": 0.15358754992485046, "learning_rate": 2.729277812186629e-05, "loss": 0.376, "num_input_tokens_seen": 13068392, "step": 19985 }, { "epoch": 10.476939203354299, "grad_norm": 0.14304140210151672, "learning_rate": 2.7281391785547945e-05, "loss": 0.6002, "num_input_tokens_seen": 13071400, "step": 19990 }, { "epoch": 10.479559748427674, "grad_norm": 0.34877514839172363, "learning_rate": 2.7270004971988456e-05, "loss": 0.5828, "num_input_tokens_seen": 13074760, "step": 19995 }, { "epoch": 10.482180293501049, "grad_norm": 0.5554327368736267, "learning_rate": 2.7258617683569825e-05, "loss": 0.5173, "num_input_tokens_seen": 13078664, "step": 20000 }, { "epoch": 10.484800838574424, "grad_norm": 0.23681087791919708, "learning_rate": 2.7247229922674143e-05, "loss": 0.6548, "num_input_tokens_seen": 13081448, "step": 20005 }, { "epoch": 10.4874213836478, "grad_norm": 0.17114953696727753, "learning_rate": 2.7235841691683605e-05, "loss": 0.3934, "num_input_tokens_seen": 13085224, "step": 20010 }, { "epoch": 10.490041928721174, "grad_norm": 0.19741585850715637, "learning_rate": 2.7224452992980487e-05, "loss": 0.6448, "num_input_tokens_seen": 13088136, "step": 20015 }, { "epoch": 10.49266247379455, "grad_norm": 0.22446319460868835, "learning_rate": 2.7213063828947182e-05, "loss": 0.4517, "num_input_tokens_seen": 13091112, "step": 20020 }, { "epoch": 10.495283018867925, "grad_norm": 0.22285763919353485, "learning_rate": 2.7201674201966176e-05, "loss": 0.4417, "num_input_tokens_seen": 13094376, "step": 20025 }, { "epoch": 10.4979035639413, "grad_norm": 0.13147015869617462, "learning_rate": 2.7190284114420035e-05, "loss": 0.4846, "num_input_tokens_seen": 13097160, "step": 20030 }, { "epoch": 10.500524109014675, "grad_norm": 0.24162717163562775, "learning_rate": 2.717889356869146e-05, "loss": 0.4358, "num_input_tokens_seen": 13101128, "step": 20035 }, { "epoch": 10.50314465408805, "grad_norm": 0.18381400406360626, "learning_rate": 2.7167502567163194e-05, "loss": 0.5779, "num_input_tokens_seen": 13104552, "step": 20040 }, { "epoch": 10.505765199161425, "grad_norm": 0.1718340665102005, "learning_rate": 2.715611111221813e-05, "loss": 0.4981, "num_input_tokens_seen": 13106920, "step": 20045 }, { "epoch": 10.5083857442348, "grad_norm": 0.1640557497739792, "learning_rate": 2.7144719206239217e-05, "loss": 0.547, "num_input_tokens_seen": 13110408, "step": 20050 }, { "epoch": 10.511006289308176, "grad_norm": 0.16844971477985382, "learning_rate": 2.7133326851609508e-05, "loss": 0.5095, "num_input_tokens_seen": 13112808, "step": 20055 }, { "epoch": 10.51362683438155, "grad_norm": 0.12948693335056305, "learning_rate": 2.7121934050712166e-05, "loss": 0.4648, "num_input_tokens_seen": 13116680, "step": 20060 }, { "epoch": 10.516247379454926, "grad_norm": 0.23277297616004944, "learning_rate": 2.711054080593043e-05, "loss": 0.5469, "num_input_tokens_seen": 13119272, "step": 20065 }, { "epoch": 10.518867924528301, "grad_norm": 0.13664932548999786, "learning_rate": 2.7099147119647627e-05, "loss": 0.4699, "num_input_tokens_seen": 13122664, "step": 20070 }, { "epoch": 10.521488469601676, "grad_norm": 0.3032954931259155, "learning_rate": 2.7087752994247206e-05, "loss": 0.4745, "num_input_tokens_seen": 13125672, "step": 20075 }, { "epoch": 10.524109014675052, "grad_norm": 0.25679948925971985, "learning_rate": 2.707635843211267e-05, "loss": 0.5319, "num_input_tokens_seen": 13128392, "step": 20080 }, { "epoch": 10.526729559748428, "grad_norm": 0.21632808446884155, "learning_rate": 2.7064963435627637e-05, "loss": 0.4058, "num_input_tokens_seen": 13131624, "step": 20085 }, { "epoch": 10.529350104821804, "grad_norm": 0.13331007957458496, "learning_rate": 2.7053568007175823e-05, "loss": 0.5019, "num_input_tokens_seen": 13135240, "step": 20090 }, { "epoch": 10.531970649895179, "grad_norm": 0.10133788734674454, "learning_rate": 2.7042172149140994e-05, "loss": 0.6507, "num_input_tokens_seen": 13138152, "step": 20095 }, { "epoch": 10.534591194968554, "grad_norm": 0.19619937241077423, "learning_rate": 2.703077586390706e-05, "loss": 0.434, "num_input_tokens_seen": 13144040, "step": 20100 }, { "epoch": 10.53721174004193, "grad_norm": 0.16956692934036255, "learning_rate": 2.7019379153857982e-05, "loss": 0.5906, "num_input_tokens_seen": 13146664, "step": 20105 }, { "epoch": 10.539832285115304, "grad_norm": 0.17606830596923828, "learning_rate": 2.700798202137782e-05, "loss": 0.5765, "num_input_tokens_seen": 13151080, "step": 20110 }, { "epoch": 10.54245283018868, "grad_norm": 0.2623564898967743, "learning_rate": 2.6996584468850732e-05, "loss": 0.4893, "num_input_tokens_seen": 13154600, "step": 20115 }, { "epoch": 10.545073375262055, "grad_norm": 0.26657724380493164, "learning_rate": 2.6985186498660946e-05, "loss": 0.3486, "num_input_tokens_seen": 13157928, "step": 20120 }, { "epoch": 10.54769392033543, "grad_norm": 0.10576673597097397, "learning_rate": 2.69737881131928e-05, "loss": 0.4813, "num_input_tokens_seen": 13161000, "step": 20125 }, { "epoch": 10.550314465408805, "grad_norm": 0.20155318081378937, "learning_rate": 2.6962389314830692e-05, "loss": 0.5417, "num_input_tokens_seen": 13163880, "step": 20130 }, { "epoch": 10.55293501048218, "grad_norm": 0.136580690741539, "learning_rate": 2.6950990105959135e-05, "loss": 0.4882, "num_input_tokens_seen": 13166376, "step": 20135 }, { "epoch": 10.555555555555555, "grad_norm": 0.13653552532196045, "learning_rate": 2.6939590488962703e-05, "loss": 0.3822, "num_input_tokens_seen": 13170120, "step": 20140 }, { "epoch": 10.55817610062893, "grad_norm": 0.21231666207313538, "learning_rate": 2.6928190466226067e-05, "loss": 0.5769, "num_input_tokens_seen": 13172552, "step": 20145 }, { "epoch": 10.560796645702306, "grad_norm": 0.11859040707349777, "learning_rate": 2.691679004013398e-05, "loss": 0.4531, "num_input_tokens_seen": 13176456, "step": 20150 }, { "epoch": 10.56341719077568, "grad_norm": 0.16746032238006592, "learning_rate": 2.6905389213071285e-05, "loss": 0.5127, "num_input_tokens_seen": 13179496, "step": 20155 }, { "epoch": 10.566037735849056, "grad_norm": 0.14626477658748627, "learning_rate": 2.6893987987422907e-05, "loss": 0.5696, "num_input_tokens_seen": 13182504, "step": 20160 }, { "epoch": 10.568658280922431, "grad_norm": 0.06444206833839417, "learning_rate": 2.6882586365573854e-05, "loss": 0.4464, "num_input_tokens_seen": 13185960, "step": 20165 }, { "epoch": 10.571278825995806, "grad_norm": 0.1408083140850067, "learning_rate": 2.6871184349909206e-05, "loss": 0.4575, "num_input_tokens_seen": 13188968, "step": 20170 }, { "epoch": 10.573899371069182, "grad_norm": 0.1319417953491211, "learning_rate": 2.685978194281414e-05, "loss": 0.3435, "num_input_tokens_seen": 13191944, "step": 20175 }, { "epoch": 10.576519916142558, "grad_norm": 0.2180132269859314, "learning_rate": 2.6848379146673914e-05, "loss": 0.6522, "num_input_tokens_seen": 13195496, "step": 20180 }, { "epoch": 10.579140461215934, "grad_norm": 0.17216138541698456, "learning_rate": 2.6836975963873843e-05, "loss": 0.4901, "num_input_tokens_seen": 13198408, "step": 20185 }, { "epoch": 10.581761006289309, "grad_norm": 0.17572332918643951, "learning_rate": 2.6825572396799377e-05, "loss": 0.3646, "num_input_tokens_seen": 13202184, "step": 20190 }, { "epoch": 10.584381551362684, "grad_norm": 0.1947864443063736, "learning_rate": 2.6814168447835986e-05, "loss": 0.6217, "num_input_tokens_seen": 13204872, "step": 20195 }, { "epoch": 10.58700209643606, "grad_norm": 0.13201072812080383, "learning_rate": 2.6802764119369245e-05, "loss": 0.4129, "num_input_tokens_seen": 13208296, "step": 20200 }, { "epoch": 10.589622641509434, "grad_norm": 0.11836956441402435, "learning_rate": 2.6791359413784828e-05, "loss": 0.3804, "num_input_tokens_seen": 13211592, "step": 20205 }, { "epoch": 10.59224318658281, "grad_norm": 0.16751402616500854, "learning_rate": 2.677995433346845e-05, "loss": 0.5791, "num_input_tokens_seen": 13215048, "step": 20210 }, { "epoch": 10.594863731656185, "grad_norm": 0.29005834460258484, "learning_rate": 2.6768548880805934e-05, "loss": 0.6202, "num_input_tokens_seen": 13218760, "step": 20215 }, { "epoch": 10.59748427672956, "grad_norm": 0.15939413011074066, "learning_rate": 2.6757143058183177e-05, "loss": 0.4135, "num_input_tokens_seen": 13221800, "step": 20220 }, { "epoch": 10.600104821802935, "grad_norm": 0.1745874136686325, "learning_rate": 2.6745736867986132e-05, "loss": 0.5912, "num_input_tokens_seen": 13224872, "step": 20225 }, { "epoch": 10.60272536687631, "grad_norm": 0.2212914377450943, "learning_rate": 2.673433031260085e-05, "loss": 0.4287, "num_input_tokens_seen": 13227464, "step": 20230 }, { "epoch": 10.605345911949685, "grad_norm": 0.14011473953723907, "learning_rate": 2.6722923394413456e-05, "loss": 0.6432, "num_input_tokens_seen": 13230952, "step": 20235 }, { "epoch": 10.60796645702306, "grad_norm": 0.12394610792398453, "learning_rate": 2.6711516115810136e-05, "loss": 0.5239, "num_input_tokens_seen": 13235784, "step": 20240 }, { "epoch": 10.610587002096436, "grad_norm": 0.1594243198633194, "learning_rate": 2.6700108479177188e-05, "loss": 0.3416, "num_input_tokens_seen": 13238824, "step": 20245 }, { "epoch": 10.61320754716981, "grad_norm": 0.2500101327896118, "learning_rate": 2.6688700486900926e-05, "loss": 0.4497, "num_input_tokens_seen": 13241736, "step": 20250 }, { "epoch": 10.615828092243186, "grad_norm": 0.1434832215309143, "learning_rate": 2.6677292141367787e-05, "loss": 0.545, "num_input_tokens_seen": 13244520, "step": 20255 }, { "epoch": 10.618448637316561, "grad_norm": 0.3861081004142761, "learning_rate": 2.6665883444964278e-05, "loss": 0.4368, "num_input_tokens_seen": 13248520, "step": 20260 }, { "epoch": 10.621069182389936, "grad_norm": 0.12246726453304291, "learning_rate": 2.665447440007695e-05, "loss": 0.5142, "num_input_tokens_seen": 13251912, "step": 20265 }, { "epoch": 10.623689727463312, "grad_norm": 0.2381601333618164, "learning_rate": 2.664306500909246e-05, "loss": 0.5004, "num_input_tokens_seen": 13254280, "step": 20270 }, { "epoch": 10.626310272536688, "grad_norm": 0.16091005504131317, "learning_rate": 2.663165527439751e-05, "loss": 0.5464, "num_input_tokens_seen": 13257864, "step": 20275 }, { "epoch": 10.628930817610064, "grad_norm": 0.13572372496128082, "learning_rate": 2.662024519837889e-05, "loss": 0.4954, "num_input_tokens_seen": 13260520, "step": 20280 }, { "epoch": 10.631551362683439, "grad_norm": 0.22546501457691193, "learning_rate": 2.6608834783423465e-05, "loss": 0.5472, "num_input_tokens_seen": 13263432, "step": 20285 }, { "epoch": 10.634171907756814, "grad_norm": 0.20975710451602936, "learning_rate": 2.6597424031918156e-05, "loss": 0.5046, "num_input_tokens_seen": 13267112, "step": 20290 }, { "epoch": 10.63679245283019, "grad_norm": 0.3547961413860321, "learning_rate": 2.6586012946249954e-05, "loss": 0.4944, "num_input_tokens_seen": 13269736, "step": 20295 }, { "epoch": 10.639412997903564, "grad_norm": 0.2032892405986786, "learning_rate": 2.6574601528805953e-05, "loss": 0.6712, "num_input_tokens_seen": 13272520, "step": 20300 }, { "epoch": 10.64203354297694, "grad_norm": 0.32374581694602966, "learning_rate": 2.6563189781973268e-05, "loss": 0.582, "num_input_tokens_seen": 13275976, "step": 20305 }, { "epoch": 10.644654088050315, "grad_norm": 0.22612561285495758, "learning_rate": 2.655177770813912e-05, "loss": 0.5063, "num_input_tokens_seen": 13279304, "step": 20310 }, { "epoch": 10.64727463312369, "grad_norm": 0.20898860692977905, "learning_rate": 2.654036530969077e-05, "loss": 0.4992, "num_input_tokens_seen": 13282888, "step": 20315 }, { "epoch": 10.649895178197065, "grad_norm": 0.14643362164497375, "learning_rate": 2.6528952589015576e-05, "loss": 0.4122, "num_input_tokens_seen": 13285512, "step": 20320 }, { "epoch": 10.65251572327044, "grad_norm": 0.11785273998975754, "learning_rate": 2.6517539548500947e-05, "loss": 0.4542, "num_input_tokens_seen": 13288904, "step": 20325 }, { "epoch": 10.655136268343815, "grad_norm": 0.10132157802581787, "learning_rate": 2.6506126190534354e-05, "loss": 0.4139, "num_input_tokens_seen": 13291656, "step": 20330 }, { "epoch": 10.65775681341719, "grad_norm": 0.12937569618225098, "learning_rate": 2.649471251750334e-05, "loss": 0.4129, "num_input_tokens_seen": 13295816, "step": 20335 }, { "epoch": 10.660377358490566, "grad_norm": 0.1461445540189743, "learning_rate": 2.6483298531795527e-05, "loss": 0.5412, "num_input_tokens_seen": 13299112, "step": 20340 }, { "epoch": 10.66299790356394, "grad_norm": 0.14760318398475647, "learning_rate": 2.6471884235798583e-05, "loss": 0.4975, "num_input_tokens_seen": 13302696, "step": 20345 }, { "epoch": 10.665618448637316, "grad_norm": 0.10076563060283661, "learning_rate": 2.6460469631900247e-05, "loss": 0.4478, "num_input_tokens_seen": 13306408, "step": 20350 }, { "epoch": 10.668238993710691, "grad_norm": 0.16778846085071564, "learning_rate": 2.644905472248832e-05, "loss": 0.4275, "num_input_tokens_seen": 13309160, "step": 20355 }, { "epoch": 10.670859538784066, "grad_norm": 0.13496407866477966, "learning_rate": 2.6437639509950686e-05, "loss": 0.46, "num_input_tokens_seen": 13311752, "step": 20360 }, { "epoch": 10.673480083857442, "grad_norm": 0.10589100420475006, "learning_rate": 2.6426223996675264e-05, "loss": 0.4697, "num_input_tokens_seen": 13314792, "step": 20365 }, { "epoch": 10.676100628930818, "grad_norm": 0.1367257535457611, "learning_rate": 2.641480818505005e-05, "loss": 0.4623, "num_input_tokens_seen": 13317832, "step": 20370 }, { "epoch": 10.678721174004194, "grad_norm": 0.18507042527198792, "learning_rate": 2.6403392077463108e-05, "loss": 0.3409, "num_input_tokens_seen": 13321096, "step": 20375 }, { "epoch": 10.681341719077569, "grad_norm": 0.12476816028356552, "learning_rate": 2.639197567630256e-05, "loss": 0.475, "num_input_tokens_seen": 13324008, "step": 20380 }, { "epoch": 10.683962264150944, "grad_norm": 0.15348859131336212, "learning_rate": 2.638055898395657e-05, "loss": 0.4847, "num_input_tokens_seen": 13327400, "step": 20385 }, { "epoch": 10.68658280922432, "grad_norm": 0.23815815150737762, "learning_rate": 2.6369142002813403e-05, "loss": 0.4735, "num_input_tokens_seen": 13330504, "step": 20390 }, { "epoch": 10.689203354297694, "grad_norm": 0.16265583038330078, "learning_rate": 2.6357724735261335e-05, "loss": 0.5422, "num_input_tokens_seen": 13334152, "step": 20395 }, { "epoch": 10.69182389937107, "grad_norm": 0.1432827115058899, "learning_rate": 2.6346307183688752e-05, "loss": 0.4303, "num_input_tokens_seen": 13337960, "step": 20400 }, { "epoch": 10.694444444444445, "grad_norm": 0.15614697337150574, "learning_rate": 2.6334889350484064e-05, "loss": 0.4467, "num_input_tokens_seen": 13341096, "step": 20405 }, { "epoch": 10.69706498951782, "grad_norm": 0.2152642160654068, "learning_rate": 2.632347123803575e-05, "loss": 0.3748, "num_input_tokens_seen": 13344744, "step": 20410 }, { "epoch": 10.699685534591195, "grad_norm": 0.234037384390831, "learning_rate": 2.6312052848732353e-05, "loss": 0.5666, "num_input_tokens_seen": 13347752, "step": 20415 }, { "epoch": 10.70230607966457, "grad_norm": 0.12255359441041946, "learning_rate": 2.6300634184962475e-05, "loss": 0.5102, "num_input_tokens_seen": 13350696, "step": 20420 }, { "epoch": 10.704926624737945, "grad_norm": 0.287029504776001, "learning_rate": 2.628921524911475e-05, "loss": 0.6332, "num_input_tokens_seen": 13353032, "step": 20425 }, { "epoch": 10.70754716981132, "grad_norm": 0.16263385117053986, "learning_rate": 2.6277796043577924e-05, "loss": 0.4669, "num_input_tokens_seen": 13355784, "step": 20430 }, { "epoch": 10.710167714884696, "grad_norm": 0.21925219893455505, "learning_rate": 2.6266376570740725e-05, "loss": 0.4873, "num_input_tokens_seen": 13358408, "step": 20435 }, { "epoch": 10.71278825995807, "grad_norm": 0.11621072888374329, "learning_rate": 2.6254956832992005e-05, "loss": 0.433, "num_input_tokens_seen": 13365192, "step": 20440 }, { "epoch": 10.715408805031446, "grad_norm": 0.18952418863773346, "learning_rate": 2.6243536832720627e-05, "loss": 0.5104, "num_input_tokens_seen": 13368808, "step": 20445 }, { "epoch": 10.718029350104821, "grad_norm": 0.16000114381313324, "learning_rate": 2.623211657231553e-05, "loss": 0.4759, "num_input_tokens_seen": 13371880, "step": 20450 }, { "epoch": 10.720649895178196, "grad_norm": 0.25343993306159973, "learning_rate": 2.6220696054165704e-05, "loss": 0.8126, "num_input_tokens_seen": 13374632, "step": 20455 }, { "epoch": 10.723270440251572, "grad_norm": 0.14466100931167603, "learning_rate": 2.6209275280660194e-05, "loss": 0.5628, "num_input_tokens_seen": 13377960, "step": 20460 }, { "epoch": 10.725890985324948, "grad_norm": 0.15645425021648407, "learning_rate": 2.6197854254188085e-05, "loss": 0.4877, "num_input_tokens_seen": 13381864, "step": 20465 }, { "epoch": 10.728511530398324, "grad_norm": 0.22326532006263733, "learning_rate": 2.618643297713854e-05, "loss": 0.4619, "num_input_tokens_seen": 13384232, "step": 20470 }, { "epoch": 10.731132075471699, "grad_norm": 0.21755817532539368, "learning_rate": 2.617501145190075e-05, "loss": 0.4658, "num_input_tokens_seen": 13387208, "step": 20475 }, { "epoch": 10.733752620545074, "grad_norm": 0.30028194189071655, "learning_rate": 2.616358968086397e-05, "loss": 0.6224, "num_input_tokens_seen": 13389992, "step": 20480 }, { "epoch": 10.73637316561845, "grad_norm": 0.18321330845355988, "learning_rate": 2.6152167666417493e-05, "loss": 0.5422, "num_input_tokens_seen": 13393224, "step": 20485 }, { "epoch": 10.738993710691824, "grad_norm": 0.1579982191324234, "learning_rate": 2.6140745410950694e-05, "loss": 0.6668, "num_input_tokens_seen": 13395880, "step": 20490 }, { "epoch": 10.7416142557652, "grad_norm": 0.2534405589103699, "learning_rate": 2.6129322916852973e-05, "loss": 0.4606, "num_input_tokens_seen": 13399432, "step": 20495 }, { "epoch": 10.744234800838575, "grad_norm": 0.14386524260044098, "learning_rate": 2.611790018651377e-05, "loss": 0.554, "num_input_tokens_seen": 13402248, "step": 20500 }, { "epoch": 10.74685534591195, "grad_norm": 0.18465077877044678, "learning_rate": 2.6106477222322613e-05, "loss": 0.4334, "num_input_tokens_seen": 13404648, "step": 20505 }, { "epoch": 10.749475890985325, "grad_norm": 0.13814115524291992, "learning_rate": 2.6095054026669037e-05, "loss": 0.4999, "num_input_tokens_seen": 13408296, "step": 20510 }, { "epoch": 10.7520964360587, "grad_norm": 0.2048221379518509, "learning_rate": 2.608363060194265e-05, "loss": 0.4535, "num_input_tokens_seen": 13411560, "step": 20515 }, { "epoch": 10.754716981132075, "grad_norm": 0.22627407312393188, "learning_rate": 2.6072206950533117e-05, "loss": 0.4904, "num_input_tokens_seen": 13414536, "step": 20520 }, { "epoch": 10.75733752620545, "grad_norm": 0.14232735335826874, "learning_rate": 2.606078307483011e-05, "loss": 0.5446, "num_input_tokens_seen": 13417736, "step": 20525 }, { "epoch": 10.759958071278826, "grad_norm": 0.18134969472885132, "learning_rate": 2.6049358977223392e-05, "loss": 0.5303, "num_input_tokens_seen": 13420744, "step": 20530 }, { "epoch": 10.7625786163522, "grad_norm": 0.18924972414970398, "learning_rate": 2.6037934660102747e-05, "loss": 0.4713, "num_input_tokens_seen": 13425032, "step": 20535 }, { "epoch": 10.765199161425576, "grad_norm": 0.10463264584541321, "learning_rate": 2.6026510125858006e-05, "loss": 0.4148, "num_input_tokens_seen": 13429544, "step": 20540 }, { "epoch": 10.767819706498951, "grad_norm": 0.13768605887889862, "learning_rate": 2.6015085376879063e-05, "loss": 0.5479, "num_input_tokens_seen": 13434088, "step": 20545 }, { "epoch": 10.770440251572326, "grad_norm": 0.14350618422031403, "learning_rate": 2.600366041555584e-05, "loss": 0.4706, "num_input_tokens_seen": 13437096, "step": 20550 }, { "epoch": 10.773060796645701, "grad_norm": 0.19246801733970642, "learning_rate": 2.5992235244278302e-05, "loss": 0.4669, "num_input_tokens_seen": 13440680, "step": 20555 }, { "epoch": 10.775681341719078, "grad_norm": 0.14114950597286224, "learning_rate": 2.5980809865436485e-05, "loss": 0.3868, "num_input_tokens_seen": 13443496, "step": 20560 }, { "epoch": 10.778301886792454, "grad_norm": 0.10258933156728745, "learning_rate": 2.5969384281420424e-05, "loss": 0.572, "num_input_tokens_seen": 13446920, "step": 20565 }, { "epoch": 10.780922431865829, "grad_norm": 0.16027405858039856, "learning_rate": 2.5957958494620227e-05, "loss": 0.5789, "num_input_tokens_seen": 13449544, "step": 20570 }, { "epoch": 10.783542976939204, "grad_norm": 0.1625060737133026, "learning_rate": 2.594653250742605e-05, "loss": 0.5537, "num_input_tokens_seen": 13451848, "step": 20575 }, { "epoch": 10.786163522012579, "grad_norm": 0.1730729639530182, "learning_rate": 2.5935106322228065e-05, "loss": 0.5394, "num_input_tokens_seen": 13455272, "step": 20580 }, { "epoch": 10.788784067085954, "grad_norm": 0.15292586386203766, "learning_rate": 2.5923679941416514e-05, "loss": 0.5845, "num_input_tokens_seen": 13457992, "step": 20585 }, { "epoch": 10.79140461215933, "grad_norm": 0.10041964799165726, "learning_rate": 2.591225336738165e-05, "loss": 0.3679, "num_input_tokens_seen": 13461480, "step": 20590 }, { "epoch": 10.794025157232705, "grad_norm": 0.15690132975578308, "learning_rate": 2.5900826602513782e-05, "loss": 0.4885, "num_input_tokens_seen": 13464072, "step": 20595 }, { "epoch": 10.79664570230608, "grad_norm": 0.14101895689964294, "learning_rate": 2.5889399649203277e-05, "loss": 0.4596, "num_input_tokens_seen": 13466888, "step": 20600 }, { "epoch": 10.799266247379455, "grad_norm": 0.23775197565555573, "learning_rate": 2.5877972509840514e-05, "loss": 0.5148, "num_input_tokens_seen": 13469736, "step": 20605 }, { "epoch": 10.80188679245283, "grad_norm": 0.15639987587928772, "learning_rate": 2.5866545186815912e-05, "loss": 0.4641, "num_input_tokens_seen": 13472424, "step": 20610 }, { "epoch": 10.804507337526205, "grad_norm": 0.10850690305233002, "learning_rate": 2.5855117682519954e-05, "loss": 0.4039, "num_input_tokens_seen": 13476040, "step": 20615 }, { "epoch": 10.80712788259958, "grad_norm": 0.10281924903392792, "learning_rate": 2.584368999934312e-05, "loss": 0.4519, "num_input_tokens_seen": 13481160, "step": 20620 }, { "epoch": 10.809748427672956, "grad_norm": 0.14222806692123413, "learning_rate": 2.5832262139675966e-05, "loss": 0.4247, "num_input_tokens_seen": 13484904, "step": 20625 }, { "epoch": 10.81236897274633, "grad_norm": 0.1467220038175583, "learning_rate": 2.582083410590907e-05, "loss": 0.4474, "num_input_tokens_seen": 13487688, "step": 20630 }, { "epoch": 10.814989517819706, "grad_norm": 0.1318383514881134, "learning_rate": 2.5809405900433036e-05, "loss": 0.5094, "num_input_tokens_seen": 13491112, "step": 20635 }, { "epoch": 10.817610062893081, "grad_norm": 0.1964511275291443, "learning_rate": 2.579797752563853e-05, "loss": 0.4945, "num_input_tokens_seen": 13494248, "step": 20640 }, { "epoch": 10.820230607966456, "grad_norm": 0.14227360486984253, "learning_rate": 2.5786548983916232e-05, "loss": 0.5524, "num_input_tokens_seen": 13497960, "step": 20645 }, { "epoch": 10.822851153039831, "grad_norm": 0.1612814962863922, "learning_rate": 2.577512027765685e-05, "loss": 0.5228, "num_input_tokens_seen": 13501256, "step": 20650 }, { "epoch": 10.825471698113208, "grad_norm": 0.2343069612979889, "learning_rate": 2.5763691409251155e-05, "loss": 0.4614, "num_input_tokens_seen": 13503592, "step": 20655 }, { "epoch": 10.828092243186584, "grad_norm": 0.4383951723575592, "learning_rate": 2.5752262381089932e-05, "loss": 0.6575, "num_input_tokens_seen": 13505704, "step": 20660 }, { "epoch": 10.830712788259959, "grad_norm": 0.21621473133563995, "learning_rate": 2.5740833195563996e-05, "loss": 0.4612, "num_input_tokens_seen": 13508712, "step": 20665 }, { "epoch": 10.833333333333334, "grad_norm": 0.15647296607494354, "learning_rate": 2.572940385506421e-05, "loss": 0.5549, "num_input_tokens_seen": 13512136, "step": 20670 }, { "epoch": 10.835953878406709, "grad_norm": 0.1859930157661438, "learning_rate": 2.5717974361981457e-05, "loss": 0.5174, "num_input_tokens_seen": 13514760, "step": 20675 }, { "epoch": 10.838574423480084, "grad_norm": 0.1698906123638153, "learning_rate": 2.5706544718706665e-05, "loss": 0.5074, "num_input_tokens_seen": 13517384, "step": 20680 }, { "epoch": 10.84119496855346, "grad_norm": 0.2606038749217987, "learning_rate": 2.569511492763077e-05, "loss": 0.5567, "num_input_tokens_seen": 13519624, "step": 20685 }, { "epoch": 10.843815513626835, "grad_norm": 0.20053845643997192, "learning_rate": 2.5683684991144764e-05, "loss": 0.4152, "num_input_tokens_seen": 13522152, "step": 20690 }, { "epoch": 10.84643605870021, "grad_norm": 0.16223222017288208, "learning_rate": 2.5672254911639666e-05, "loss": 0.4816, "num_input_tokens_seen": 13525256, "step": 20695 }, { "epoch": 10.849056603773585, "grad_norm": 0.16723820567131042, "learning_rate": 2.5660824691506496e-05, "loss": 0.466, "num_input_tokens_seen": 13528744, "step": 20700 }, { "epoch": 10.85167714884696, "grad_norm": 0.2811715602874756, "learning_rate": 2.5649394333136356e-05, "loss": 0.5175, "num_input_tokens_seen": 13531304, "step": 20705 }, { "epoch": 10.854297693920335, "grad_norm": 0.1628895401954651, "learning_rate": 2.5637963838920316e-05, "loss": 0.5381, "num_input_tokens_seen": 13534696, "step": 20710 }, { "epoch": 10.85691823899371, "grad_norm": 0.1825730800628662, "learning_rate": 2.5626533211249525e-05, "loss": 0.5187, "num_input_tokens_seen": 13537672, "step": 20715 }, { "epoch": 10.859538784067086, "grad_norm": 0.16289901733398438, "learning_rate": 2.5615102452515138e-05, "loss": 0.4667, "num_input_tokens_seen": 13542504, "step": 20720 }, { "epoch": 10.86215932914046, "grad_norm": 0.19920310378074646, "learning_rate": 2.5603671565108328e-05, "loss": 0.3655, "num_input_tokens_seen": 13545544, "step": 20725 }, { "epoch": 10.864779874213836, "grad_norm": 0.17383462190628052, "learning_rate": 2.5592240551420317e-05, "loss": 0.5068, "num_input_tokens_seen": 13551336, "step": 20730 }, { "epoch": 10.867400419287211, "grad_norm": 0.2562277317047119, "learning_rate": 2.5580809413842343e-05, "loss": 0.4778, "num_input_tokens_seen": 13554952, "step": 20735 }, { "epoch": 10.870020964360586, "grad_norm": 0.17781129479408264, "learning_rate": 2.5569378154765655e-05, "loss": 0.4731, "num_input_tokens_seen": 13557864, "step": 20740 }, { "epoch": 10.872641509433961, "grad_norm": 0.15108820796012878, "learning_rate": 2.5557946776581572e-05, "loss": 0.597, "num_input_tokens_seen": 13561096, "step": 20745 }, { "epoch": 10.875262054507338, "grad_norm": 0.13504038751125336, "learning_rate": 2.554651528168137e-05, "loss": 0.5412, "num_input_tokens_seen": 13565000, "step": 20750 }, { "epoch": 10.877882599580714, "grad_norm": 0.15313416719436646, "learning_rate": 2.5535083672456407e-05, "loss": 0.41, "num_input_tokens_seen": 13568360, "step": 20755 }, { "epoch": 10.880503144654089, "grad_norm": 0.14450599253177643, "learning_rate": 2.5523651951298055e-05, "loss": 0.5269, "num_input_tokens_seen": 13572168, "step": 20760 }, { "epoch": 10.883123689727464, "grad_norm": 0.11512403190135956, "learning_rate": 2.5512220120597686e-05, "loss": 0.4184, "num_input_tokens_seen": 13575144, "step": 20765 }, { "epoch": 10.885744234800839, "grad_norm": 0.1516902893781662, "learning_rate": 2.550078818274671e-05, "loss": 0.4228, "num_input_tokens_seen": 13581832, "step": 20770 }, { "epoch": 10.888364779874214, "grad_norm": 0.13525372743606567, "learning_rate": 2.548935614013656e-05, "loss": 0.433, "num_input_tokens_seen": 13585096, "step": 20775 }, { "epoch": 10.89098532494759, "grad_norm": 0.40122190117836, "learning_rate": 2.5477923995158683e-05, "loss": 0.524, "num_input_tokens_seen": 13588328, "step": 20780 }, { "epoch": 10.893605870020965, "grad_norm": 0.18398401141166687, "learning_rate": 2.5466491750204562e-05, "loss": 0.4677, "num_input_tokens_seen": 13592744, "step": 20785 }, { "epoch": 10.89622641509434, "grad_norm": 0.2171357125043869, "learning_rate": 2.5455059407665693e-05, "loss": 0.4556, "num_input_tokens_seen": 13596232, "step": 20790 }, { "epoch": 10.898846960167715, "grad_norm": 0.3223632872104645, "learning_rate": 2.544362696993358e-05, "loss": 0.4947, "num_input_tokens_seen": 13599880, "step": 20795 }, { "epoch": 10.90146750524109, "grad_norm": 0.13065482676029205, "learning_rate": 2.5432194439399775e-05, "loss": 0.6155, "num_input_tokens_seen": 13603560, "step": 20800 }, { "epoch": 10.904088050314465, "grad_norm": 0.17609712481498718, "learning_rate": 2.542076181845582e-05, "loss": 0.4649, "num_input_tokens_seen": 13606152, "step": 20805 }, { "epoch": 10.90670859538784, "grad_norm": 0.169477641582489, "learning_rate": 2.540932910949329e-05, "loss": 0.4445, "num_input_tokens_seen": 13609864, "step": 20810 }, { "epoch": 10.909329140461216, "grad_norm": 0.15474626421928406, "learning_rate": 2.5397896314903785e-05, "loss": 0.5093, "num_input_tokens_seen": 13612424, "step": 20815 }, { "epoch": 10.91194968553459, "grad_norm": 0.21690493822097778, "learning_rate": 2.5386463437078905e-05, "loss": 0.6296, "num_input_tokens_seen": 13615496, "step": 20820 }, { "epoch": 10.914570230607966, "grad_norm": 0.1726013869047165, "learning_rate": 2.5375030478410282e-05, "loss": 0.6171, "num_input_tokens_seen": 13618248, "step": 20825 }, { "epoch": 10.917190775681341, "grad_norm": 0.11933504045009613, "learning_rate": 2.536359744128957e-05, "loss": 0.6851, "num_input_tokens_seen": 13622088, "step": 20830 }, { "epoch": 10.919811320754716, "grad_norm": 0.12467316538095474, "learning_rate": 2.5352164328108413e-05, "loss": 0.3964, "num_input_tokens_seen": 13625544, "step": 20835 }, { "epoch": 10.922431865828091, "grad_norm": 0.09355559945106506, "learning_rate": 2.53407311412585e-05, "loss": 0.4424, "num_input_tokens_seen": 13629896, "step": 20840 }, { "epoch": 10.925052410901468, "grad_norm": 0.11147888004779816, "learning_rate": 2.5329297883131526e-05, "loss": 0.5276, "num_input_tokens_seen": 13632936, "step": 20845 }, { "epoch": 10.927672955974844, "grad_norm": 0.197933629155159, "learning_rate": 2.5317864556119182e-05, "loss": 0.4338, "num_input_tokens_seen": 13635176, "step": 20850 }, { "epoch": 10.930293501048219, "grad_norm": 0.11961255222558975, "learning_rate": 2.53064311626132e-05, "loss": 0.4373, "num_input_tokens_seen": 13638376, "step": 20855 }, { "epoch": 10.932914046121594, "grad_norm": 0.33275821805000305, "learning_rate": 2.5294997705005318e-05, "loss": 0.4456, "num_input_tokens_seen": 13641832, "step": 20860 }, { "epoch": 10.935534591194969, "grad_norm": 0.19659872353076935, "learning_rate": 2.5283564185687282e-05, "loss": 0.4395, "num_input_tokens_seen": 13644872, "step": 20865 }, { "epoch": 10.938155136268344, "grad_norm": 0.1850927770137787, "learning_rate": 2.5272130607050848e-05, "loss": 0.4834, "num_input_tokens_seen": 13648776, "step": 20870 }, { "epoch": 10.94077568134172, "grad_norm": 0.30257126688957214, "learning_rate": 2.5260696971487802e-05, "loss": 0.5339, "num_input_tokens_seen": 13652648, "step": 20875 }, { "epoch": 10.943396226415095, "grad_norm": 0.16717040538787842, "learning_rate": 2.5249263281389923e-05, "loss": 0.4693, "num_input_tokens_seen": 13656040, "step": 20880 }, { "epoch": 10.94601677148847, "grad_norm": 0.2339249700307846, "learning_rate": 2.5237829539149004e-05, "loss": 0.5024, "num_input_tokens_seen": 13658888, "step": 20885 }, { "epoch": 10.948637316561845, "grad_norm": 0.11360990256071091, "learning_rate": 2.5226395747156866e-05, "loss": 0.3934, "num_input_tokens_seen": 13661736, "step": 20890 }, { "epoch": 10.95125786163522, "grad_norm": 0.14346875250339508, "learning_rate": 2.5214961907805317e-05, "loss": 0.4036, "num_input_tokens_seen": 13664872, "step": 20895 }, { "epoch": 10.953878406708595, "grad_norm": 0.258402019739151, "learning_rate": 2.520352802348619e-05, "loss": 0.4291, "num_input_tokens_seen": 13668488, "step": 20900 }, { "epoch": 10.95649895178197, "grad_norm": 0.16804222762584686, "learning_rate": 2.5192094096591323e-05, "loss": 0.433, "num_input_tokens_seen": 13671528, "step": 20905 }, { "epoch": 10.959119496855346, "grad_norm": 0.16612650454044342, "learning_rate": 2.5180660129512556e-05, "loss": 0.4371, "num_input_tokens_seen": 13674280, "step": 20910 }, { "epoch": 10.96174004192872, "grad_norm": 0.22638735175132751, "learning_rate": 2.516922612464176e-05, "loss": 0.3938, "num_input_tokens_seen": 13677832, "step": 20915 }, { "epoch": 10.964360587002096, "grad_norm": 0.17806121706962585, "learning_rate": 2.5157792084370783e-05, "loss": 0.4293, "num_input_tokens_seen": 13680392, "step": 20920 }, { "epoch": 10.966981132075471, "grad_norm": 0.13227806985378265, "learning_rate": 2.5146358011091502e-05, "loss": 0.6004, "num_input_tokens_seen": 13684584, "step": 20925 }, { "epoch": 10.969601677148846, "grad_norm": 0.1585206687450409, "learning_rate": 2.5134923907195805e-05, "loss": 0.4475, "num_input_tokens_seen": 13687592, "step": 20930 }, { "epoch": 10.972222222222221, "grad_norm": 0.18412713706493378, "learning_rate": 2.5123489775075558e-05, "loss": 0.5341, "num_input_tokens_seen": 13691816, "step": 20935 }, { "epoch": 10.974842767295598, "grad_norm": 0.13649329543113708, "learning_rate": 2.511205561712266e-05, "loss": 0.3804, "num_input_tokens_seen": 13695432, "step": 20940 }, { "epoch": 10.977463312368974, "grad_norm": 0.15715628862380981, "learning_rate": 2.5100621435729004e-05, "loss": 0.4806, "num_input_tokens_seen": 13699272, "step": 20945 }, { "epoch": 10.980083857442349, "grad_norm": 0.27289965748786926, "learning_rate": 2.5089187233286492e-05, "loss": 0.4971, "num_input_tokens_seen": 13702568, "step": 20950 }, { "epoch": 10.982704402515724, "grad_norm": 0.17994795739650726, "learning_rate": 2.5077753012187026e-05, "loss": 0.7296, "num_input_tokens_seen": 13706216, "step": 20955 }, { "epoch": 10.985324947589099, "grad_norm": 0.15242861211299896, "learning_rate": 2.506631877482252e-05, "loss": 0.4917, "num_input_tokens_seen": 13709832, "step": 20960 }, { "epoch": 10.987945492662474, "grad_norm": 0.24383670091629028, "learning_rate": 2.505488452358488e-05, "loss": 0.5149, "num_input_tokens_seen": 13713128, "step": 20965 }, { "epoch": 10.99056603773585, "grad_norm": 0.07845728099346161, "learning_rate": 2.504345026086603e-05, "loss": 0.5052, "num_input_tokens_seen": 13716008, "step": 20970 }, { "epoch": 10.993186582809225, "grad_norm": 0.4874379336833954, "learning_rate": 2.5032015989057877e-05, "loss": 0.6243, "num_input_tokens_seen": 13719304, "step": 20975 }, { "epoch": 10.9958071278826, "grad_norm": 0.16093169152736664, "learning_rate": 2.5020581710552348e-05, "loss": 0.7091, "num_input_tokens_seen": 13721704, "step": 20980 }, { "epoch": 10.998427672955975, "grad_norm": 0.15478090941905975, "learning_rate": 2.5009147427741358e-05, "loss": 0.4601, "num_input_tokens_seen": 13724456, "step": 20985 }, { "epoch": 11.0, "eval_loss": 0.4636939465999603, "eval_runtime": 9.9344, "eval_samples_per_second": 85.36, "eval_steps_per_second": 21.34, "num_input_tokens_seen": 13725560, "step": 20988 }, { "epoch": 11.00104821802935, "grad_norm": 0.18510037660598755, "learning_rate": 2.4997713143016822e-05, "loss": 0.3831, "num_input_tokens_seen": 13726872, "step": 20990 }, { "epoch": 11.003668763102725, "grad_norm": 0.11115235835313797, "learning_rate": 2.498627885877068e-05, "loss": 0.3916, "num_input_tokens_seen": 13729944, "step": 20995 }, { "epoch": 11.0062893081761, "grad_norm": 0.17004065215587616, "learning_rate": 2.4974844577394846e-05, "loss": 0.4604, "num_input_tokens_seen": 13732472, "step": 21000 }, { "epoch": 11.008909853249476, "grad_norm": 0.191814586520195, "learning_rate": 2.496341030128123e-05, "loss": 0.5769, "num_input_tokens_seen": 13735704, "step": 21005 }, { "epoch": 11.01153039832285, "grad_norm": 0.33770206570625305, "learning_rate": 2.495197603282177e-05, "loss": 0.5116, "num_input_tokens_seen": 13738392, "step": 21010 }, { "epoch": 11.014150943396226, "grad_norm": 0.16667409241199493, "learning_rate": 2.4940541774408368e-05, "loss": 0.4289, "num_input_tokens_seen": 13741336, "step": 21015 }, { "epoch": 11.016771488469601, "grad_norm": 0.18747112154960632, "learning_rate": 2.4929107528432953e-05, "loss": 0.4627, "num_input_tokens_seen": 13744888, "step": 21020 }, { "epoch": 11.019392033542976, "grad_norm": 0.280281126499176, "learning_rate": 2.4917673297287446e-05, "loss": 0.4807, "num_input_tokens_seen": 13747512, "step": 21025 }, { "epoch": 11.022012578616351, "grad_norm": 0.244634747505188, "learning_rate": 2.4906239083363728e-05, "loss": 0.402, "num_input_tokens_seen": 13750456, "step": 21030 }, { "epoch": 11.024633123689727, "grad_norm": 0.17075331509113312, "learning_rate": 2.4894804889053732e-05, "loss": 0.4491, "num_input_tokens_seen": 13753912, "step": 21035 }, { "epoch": 11.027253668763104, "grad_norm": 0.10027868300676346, "learning_rate": 2.4883370716749347e-05, "loss": 0.4226, "num_input_tokens_seen": 13758232, "step": 21040 }, { "epoch": 11.029874213836479, "grad_norm": 0.21635884046554565, "learning_rate": 2.4871936568842486e-05, "loss": 0.5427, "num_input_tokens_seen": 13760728, "step": 21045 }, { "epoch": 11.032494758909854, "grad_norm": 0.36504891514778137, "learning_rate": 2.4860502447725033e-05, "loss": 0.5095, "num_input_tokens_seen": 13763640, "step": 21050 }, { "epoch": 11.035115303983229, "grad_norm": 0.1650371253490448, "learning_rate": 2.4849068355788873e-05, "loss": 0.5875, "num_input_tokens_seen": 13766968, "step": 21055 }, { "epoch": 11.037735849056604, "grad_norm": 0.22074761986732483, "learning_rate": 2.4837634295425892e-05, "loss": 0.4665, "num_input_tokens_seen": 13770104, "step": 21060 }, { "epoch": 11.04035639412998, "grad_norm": 0.1451030969619751, "learning_rate": 2.4826200269027957e-05, "loss": 0.4744, "num_input_tokens_seen": 13774392, "step": 21065 }, { "epoch": 11.042976939203355, "grad_norm": 0.14035451412200928, "learning_rate": 2.481476627898695e-05, "loss": 0.4465, "num_input_tokens_seen": 13778360, "step": 21070 }, { "epoch": 11.04559748427673, "grad_norm": 0.10227753221988678, "learning_rate": 2.4803332327694724e-05, "loss": 0.3752, "num_input_tokens_seen": 13782360, "step": 21075 }, { "epoch": 11.048218029350105, "grad_norm": 0.08684162050485611, "learning_rate": 2.4791898417543136e-05, "loss": 0.3655, "num_input_tokens_seen": 13785848, "step": 21080 }, { "epoch": 11.05083857442348, "grad_norm": 0.2903635501861572, "learning_rate": 2.4780464550924014e-05, "loss": 0.4209, "num_input_tokens_seen": 13788504, "step": 21085 }, { "epoch": 11.053459119496855, "grad_norm": 0.11881358176469803, "learning_rate": 2.4769030730229213e-05, "loss": 0.5262, "num_input_tokens_seen": 13791064, "step": 21090 }, { "epoch": 11.05607966457023, "grad_norm": 0.2266547679901123, "learning_rate": 2.475759695785054e-05, "loss": 0.4629, "num_input_tokens_seen": 13794200, "step": 21095 }, { "epoch": 11.058700209643606, "grad_norm": 0.10952908545732498, "learning_rate": 2.4746163236179822e-05, "loss": 0.445, "num_input_tokens_seen": 13798232, "step": 21100 }, { "epoch": 11.06132075471698, "grad_norm": 0.1472010761499405, "learning_rate": 2.4734729567608858e-05, "loss": 0.5331, "num_input_tokens_seen": 13802040, "step": 21105 }, { "epoch": 11.063941299790356, "grad_norm": 0.12425341457128525, "learning_rate": 2.472329595452945e-05, "loss": 0.4224, "num_input_tokens_seen": 13805048, "step": 21110 }, { "epoch": 11.066561844863731, "grad_norm": 0.12372253835201263, "learning_rate": 2.4711862399333367e-05, "loss": 0.3979, "num_input_tokens_seen": 13807608, "step": 21115 }, { "epoch": 11.069182389937106, "grad_norm": 0.1928328275680542, "learning_rate": 2.4700428904412377e-05, "loss": 0.5278, "num_input_tokens_seen": 13811576, "step": 21120 }, { "epoch": 11.071802935010481, "grad_norm": 0.19575068354606628, "learning_rate": 2.4688995472158245e-05, "loss": 0.4895, "num_input_tokens_seen": 13814360, "step": 21125 }, { "epoch": 11.074423480083857, "grad_norm": 0.16002000868320465, "learning_rate": 2.4677562104962712e-05, "loss": 0.3692, "num_input_tokens_seen": 13817688, "step": 21130 }, { "epoch": 11.077044025157234, "grad_norm": 0.2407427728176117, "learning_rate": 2.4666128805217513e-05, "loss": 0.3605, "num_input_tokens_seen": 13821016, "step": 21135 }, { "epoch": 11.079664570230609, "grad_norm": 0.1604188233613968, "learning_rate": 2.465469557531436e-05, "loss": 0.4447, "num_input_tokens_seen": 13823640, "step": 21140 }, { "epoch": 11.082285115303984, "grad_norm": 0.19387482106685638, "learning_rate": 2.4643262417644944e-05, "loss": 0.5298, "num_input_tokens_seen": 13826328, "step": 21145 }, { "epoch": 11.084905660377359, "grad_norm": 0.12707947194576263, "learning_rate": 2.463182933460096e-05, "loss": 0.4917, "num_input_tokens_seen": 13829528, "step": 21150 }, { "epoch": 11.087526205450734, "grad_norm": 0.18444715440273285, "learning_rate": 2.4620396328574087e-05, "loss": 0.5422, "num_input_tokens_seen": 13832760, "step": 21155 }, { "epoch": 11.09014675052411, "grad_norm": 0.09682729840278625, "learning_rate": 2.4608963401955965e-05, "loss": 0.4193, "num_input_tokens_seen": 13836664, "step": 21160 }, { "epoch": 11.092767295597485, "grad_norm": 0.14963512122631073, "learning_rate": 2.4597530557138247e-05, "loss": 0.5644, "num_input_tokens_seen": 13840408, "step": 21165 }, { "epoch": 11.09538784067086, "grad_norm": 0.11391685158014297, "learning_rate": 2.4586097796512538e-05, "loss": 0.4512, "num_input_tokens_seen": 13842968, "step": 21170 }, { "epoch": 11.098008385744235, "grad_norm": 0.1860378533601761, "learning_rate": 2.457466512247045e-05, "loss": 0.489, "num_input_tokens_seen": 13845496, "step": 21175 }, { "epoch": 11.10062893081761, "grad_norm": 0.23317153751850128, "learning_rate": 2.4563232537403563e-05, "loss": 0.5236, "num_input_tokens_seen": 13848120, "step": 21180 }, { "epoch": 11.103249475890985, "grad_norm": 0.14242829382419586, "learning_rate": 2.4551800043703442e-05, "loss": 0.4724, "num_input_tokens_seen": 13851576, "step": 21185 }, { "epoch": 11.10587002096436, "grad_norm": 0.24603168666362762, "learning_rate": 2.454036764376165e-05, "loss": 0.4776, "num_input_tokens_seen": 13855064, "step": 21190 }, { "epoch": 11.108490566037736, "grad_norm": 0.1033875048160553, "learning_rate": 2.4528935339969705e-05, "loss": 0.3917, "num_input_tokens_seen": 13859864, "step": 21195 }, { "epoch": 11.11111111111111, "grad_norm": 0.12849397957324982, "learning_rate": 2.45175031347191e-05, "loss": 0.4151, "num_input_tokens_seen": 13863000, "step": 21200 }, { "epoch": 11.113731656184486, "grad_norm": 0.1201348528265953, "learning_rate": 2.4506071030401344e-05, "loss": 0.4895, "num_input_tokens_seen": 13868280, "step": 21205 }, { "epoch": 11.116352201257861, "grad_norm": 0.13450869917869568, "learning_rate": 2.4494639029407884e-05, "loss": 0.5615, "num_input_tokens_seen": 13871640, "step": 21210 }, { "epoch": 11.118972746331236, "grad_norm": 0.26881149411201477, "learning_rate": 2.448320713413018e-05, "loss": 0.5797, "num_input_tokens_seen": 13873752, "step": 21215 }, { "epoch": 11.121593291404611, "grad_norm": 0.22771573066711426, "learning_rate": 2.4471775346959654e-05, "loss": 0.4426, "num_input_tokens_seen": 13876888, "step": 21220 }, { "epoch": 11.124213836477987, "grad_norm": 0.1660434454679489, "learning_rate": 2.446034367028769e-05, "loss": 0.4938, "num_input_tokens_seen": 13879416, "step": 21225 }, { "epoch": 11.126834381551364, "grad_norm": 0.24472057819366455, "learning_rate": 2.444891210650567e-05, "loss": 0.4664, "num_input_tokens_seen": 13883736, "step": 21230 }, { "epoch": 11.129454926624739, "grad_norm": 0.18818287551403046, "learning_rate": 2.4437480658004958e-05, "loss": 0.538, "num_input_tokens_seen": 13886392, "step": 21235 }, { "epoch": 11.132075471698114, "grad_norm": 0.10699766129255295, "learning_rate": 2.4426049327176868e-05, "loss": 0.4416, "num_input_tokens_seen": 13890680, "step": 21240 }, { "epoch": 11.134696016771489, "grad_norm": 0.08981318771839142, "learning_rate": 2.4414618116412716e-05, "loss": 0.4143, "num_input_tokens_seen": 13894200, "step": 21245 }, { "epoch": 11.137316561844864, "grad_norm": 0.13330276310443878, "learning_rate": 2.440318702810378e-05, "loss": 0.3514, "num_input_tokens_seen": 13896728, "step": 21250 }, { "epoch": 11.13993710691824, "grad_norm": 0.19804145395755768, "learning_rate": 2.43917560646413e-05, "loss": 0.4384, "num_input_tokens_seen": 13899288, "step": 21255 }, { "epoch": 11.142557651991615, "grad_norm": 0.1602025032043457, "learning_rate": 2.4380325228416518e-05, "loss": 0.6323, "num_input_tokens_seen": 13903384, "step": 21260 }, { "epoch": 11.14517819706499, "grad_norm": 0.1459982544183731, "learning_rate": 2.4368894521820624e-05, "loss": 0.4664, "num_input_tokens_seen": 13906488, "step": 21265 }, { "epoch": 11.147798742138365, "grad_norm": 0.13488106429576874, "learning_rate": 2.43574639472448e-05, "loss": 0.4176, "num_input_tokens_seen": 13909720, "step": 21270 }, { "epoch": 11.15041928721174, "grad_norm": 0.17767736315727234, "learning_rate": 2.4346033507080178e-05, "loss": 0.5006, "num_input_tokens_seen": 13913496, "step": 21275 }, { "epoch": 11.153039832285115, "grad_norm": 0.25267693400382996, "learning_rate": 2.4334603203717903e-05, "loss": 0.6809, "num_input_tokens_seen": 13915928, "step": 21280 }, { "epoch": 11.15566037735849, "grad_norm": 0.22122390568256378, "learning_rate": 2.432317303954904e-05, "loss": 0.6203, "num_input_tokens_seen": 13919512, "step": 21285 }, { "epoch": 11.158280922431866, "grad_norm": 0.27381667494773865, "learning_rate": 2.4311743016964648e-05, "loss": 0.5336, "num_input_tokens_seen": 13922232, "step": 21290 }, { "epoch": 11.16090146750524, "grad_norm": 0.14895625412464142, "learning_rate": 2.4300313138355772e-05, "loss": 0.5596, "num_input_tokens_seen": 13925624, "step": 21295 }, { "epoch": 11.163522012578616, "grad_norm": 0.15509426593780518, "learning_rate": 2.4288883406113396e-05, "loss": 0.5703, "num_input_tokens_seen": 13929208, "step": 21300 }, { "epoch": 11.166142557651991, "grad_norm": 0.2171776443719864, "learning_rate": 2.4277453822628512e-05, "loss": 0.4332, "num_input_tokens_seen": 13932952, "step": 21305 }, { "epoch": 11.168763102725366, "grad_norm": 0.14853686094284058, "learning_rate": 2.4266024390292037e-05, "loss": 0.4388, "num_input_tokens_seen": 13935416, "step": 21310 }, { "epoch": 11.171383647798741, "grad_norm": 0.2421528398990631, "learning_rate": 2.425459511149489e-05, "loss": 0.4623, "num_input_tokens_seen": 13938424, "step": 21315 }, { "epoch": 11.174004192872117, "grad_norm": 0.18922606110572815, "learning_rate": 2.424316598862793e-05, "loss": 0.4285, "num_input_tokens_seen": 13941112, "step": 21320 }, { "epoch": 11.176624737945493, "grad_norm": 0.11433681100606918, "learning_rate": 2.4231737024082023e-05, "loss": 0.6063, "num_input_tokens_seen": 13944344, "step": 21325 }, { "epoch": 11.179245283018869, "grad_norm": 0.1431512087583542, "learning_rate": 2.4220308220247954e-05, "loss": 0.6038, "num_input_tokens_seen": 13947096, "step": 21330 }, { "epoch": 11.181865828092244, "grad_norm": 0.13810928165912628, "learning_rate": 2.4208879579516528e-05, "loss": 0.4321, "num_input_tokens_seen": 13950264, "step": 21335 }, { "epoch": 11.184486373165619, "grad_norm": 0.11532710492610931, "learning_rate": 2.419745110427846e-05, "loss": 0.4465, "num_input_tokens_seen": 13953656, "step": 21340 }, { "epoch": 11.187106918238994, "grad_norm": 0.26191574335098267, "learning_rate": 2.418602279692446e-05, "loss": 0.4585, "num_input_tokens_seen": 13956920, "step": 21345 }, { "epoch": 11.18972746331237, "grad_norm": 0.19295634329319, "learning_rate": 2.417459465984521e-05, "loss": 0.4501, "num_input_tokens_seen": 13961208, "step": 21350 }, { "epoch": 11.192348008385745, "grad_norm": 0.2115897536277771, "learning_rate": 2.4163166695431336e-05, "loss": 0.6916, "num_input_tokens_seen": 13964216, "step": 21355 }, { "epoch": 11.19496855345912, "grad_norm": 0.16748856008052826, "learning_rate": 2.4151738906073446e-05, "loss": 0.5464, "num_input_tokens_seen": 13967192, "step": 21360 }, { "epoch": 11.197589098532495, "grad_norm": 1.1042990684509277, "learning_rate": 2.4140311294162105e-05, "loss": 0.6194, "num_input_tokens_seen": 13969624, "step": 21365 }, { "epoch": 11.20020964360587, "grad_norm": 0.19865262508392334, "learning_rate": 2.412888386208782e-05, "loss": 0.4988, "num_input_tokens_seen": 13972312, "step": 21370 }, { "epoch": 11.202830188679245, "grad_norm": 0.14432352781295776, "learning_rate": 2.4117456612241103e-05, "loss": 0.3324, "num_input_tokens_seen": 13974712, "step": 21375 }, { "epoch": 11.20545073375262, "grad_norm": 0.1644279509782791, "learning_rate": 2.410602954701239e-05, "loss": 0.5451, "num_input_tokens_seen": 13978552, "step": 21380 }, { "epoch": 11.208071278825996, "grad_norm": 0.24508489668369293, "learning_rate": 2.409460266879209e-05, "loss": 0.532, "num_input_tokens_seen": 13982424, "step": 21385 }, { "epoch": 11.21069182389937, "grad_norm": 0.26109397411346436, "learning_rate": 2.4083175979970592e-05, "loss": 0.5063, "num_input_tokens_seen": 13985112, "step": 21390 }, { "epoch": 11.213312368972746, "grad_norm": 0.2868667542934418, "learning_rate": 2.4071749482938226e-05, "loss": 0.3548, "num_input_tokens_seen": 13987640, "step": 21395 }, { "epoch": 11.215932914046121, "grad_norm": 0.24591805040836334, "learning_rate": 2.4060323180085266e-05, "loss": 0.4496, "num_input_tokens_seen": 13990648, "step": 21400 }, { "epoch": 11.218553459119496, "grad_norm": 0.15200144052505493, "learning_rate": 2.404889707380198e-05, "loss": 0.5181, "num_input_tokens_seen": 13994360, "step": 21405 }, { "epoch": 11.221174004192871, "grad_norm": 0.1550896018743515, "learning_rate": 2.403747116647857e-05, "loss": 0.4398, "num_input_tokens_seen": 13997592, "step": 21410 }, { "epoch": 11.223794549266247, "grad_norm": 0.20638668537139893, "learning_rate": 2.4026045460505216e-05, "loss": 0.4519, "num_input_tokens_seen": 14000600, "step": 21415 }, { "epoch": 11.226415094339623, "grad_norm": 0.14038631319999695, "learning_rate": 2.4014619958272047e-05, "loss": 0.5033, "num_input_tokens_seen": 14005560, "step": 21420 }, { "epoch": 11.229035639412999, "grad_norm": 0.14183999598026276, "learning_rate": 2.4003194662169132e-05, "loss": 0.4861, "num_input_tokens_seen": 14008920, "step": 21425 }, { "epoch": 11.231656184486374, "grad_norm": 0.1611061692237854, "learning_rate": 2.3991769574586522e-05, "loss": 0.6009, "num_input_tokens_seen": 14011448, "step": 21430 }, { "epoch": 11.234276729559749, "grad_norm": 0.19053083658218384, "learning_rate": 2.3980344697914215e-05, "loss": 0.3996, "num_input_tokens_seen": 14015000, "step": 21435 }, { "epoch": 11.236897274633124, "grad_norm": 0.13759468495845795, "learning_rate": 2.3968920034542168e-05, "loss": 0.489, "num_input_tokens_seen": 14019640, "step": 21440 }, { "epoch": 11.2395178197065, "grad_norm": 0.1756979376077652, "learning_rate": 2.3957495586860278e-05, "loss": 0.554, "num_input_tokens_seen": 14023032, "step": 21445 }, { "epoch": 11.242138364779874, "grad_norm": 0.1773974448442459, "learning_rate": 2.394607135725843e-05, "loss": 0.5466, "num_input_tokens_seen": 14026040, "step": 21450 }, { "epoch": 11.24475890985325, "grad_norm": 0.1385597437620163, "learning_rate": 2.3934647348126426e-05, "loss": 0.4996, "num_input_tokens_seen": 14029400, "step": 21455 }, { "epoch": 11.247379454926625, "grad_norm": 0.20285353064537048, "learning_rate": 2.3923223561854035e-05, "loss": 0.5813, "num_input_tokens_seen": 14032664, "step": 21460 }, { "epoch": 11.25, "grad_norm": 0.19153602421283722, "learning_rate": 2.391180000083099e-05, "loss": 0.54, "num_input_tokens_seen": 14035320, "step": 21465 }, { "epoch": 11.252620545073375, "grad_norm": 0.11195818334817886, "learning_rate": 2.390037666744698e-05, "loss": 0.5728, "num_input_tokens_seen": 14038840, "step": 21470 }, { "epoch": 11.25524109014675, "grad_norm": 0.19331566989421844, "learning_rate": 2.388895356409162e-05, "loss": 0.5475, "num_input_tokens_seen": 14042872, "step": 21475 }, { "epoch": 11.257861635220126, "grad_norm": 0.17912817001342773, "learning_rate": 2.3877530693154506e-05, "loss": 0.5474, "num_input_tokens_seen": 14046200, "step": 21480 }, { "epoch": 11.2604821802935, "grad_norm": 0.13406343758106232, "learning_rate": 2.3866108057025162e-05, "loss": 0.4091, "num_input_tokens_seen": 14048984, "step": 21485 }, { "epoch": 11.263102725366876, "grad_norm": 0.12075307220220566, "learning_rate": 2.385468565809307e-05, "loss": 0.433, "num_input_tokens_seen": 14051704, "step": 21490 }, { "epoch": 11.265723270440251, "grad_norm": 0.15129661560058594, "learning_rate": 2.3843263498747675e-05, "loss": 0.4099, "num_input_tokens_seen": 14054264, "step": 21495 }, { "epoch": 11.268343815513626, "grad_norm": 0.12640540301799774, "learning_rate": 2.383184158137836e-05, "loss": 0.4957, "num_input_tokens_seen": 14057336, "step": 21500 }, { "epoch": 11.270964360587001, "grad_norm": 0.257713258266449, "learning_rate": 2.3820419908374465e-05, "loss": 0.4514, "num_input_tokens_seen": 14061272, "step": 21505 }, { "epoch": 11.273584905660377, "grad_norm": 0.2627350687980652, "learning_rate": 2.3808998482125265e-05, "loss": 0.4164, "num_input_tokens_seen": 14064824, "step": 21510 }, { "epoch": 11.276205450733753, "grad_norm": 0.18179142475128174, "learning_rate": 2.3797577305019987e-05, "loss": 0.4914, "num_input_tokens_seen": 14068280, "step": 21515 }, { "epoch": 11.278825995807129, "grad_norm": 0.17718322575092316, "learning_rate": 2.378615637944783e-05, "loss": 0.5951, "num_input_tokens_seen": 14071864, "step": 21520 }, { "epoch": 11.281446540880504, "grad_norm": 0.1638837456703186, "learning_rate": 2.3774735707797895e-05, "loss": 0.5651, "num_input_tokens_seen": 14075064, "step": 21525 }, { "epoch": 11.284067085953879, "grad_norm": 0.14980797469615936, "learning_rate": 2.376331529245928e-05, "loss": 0.5233, "num_input_tokens_seen": 14078360, "step": 21530 }, { "epoch": 11.286687631027254, "grad_norm": 0.1108330637216568, "learning_rate": 2.3751895135821006e-05, "loss": 0.478, "num_input_tokens_seen": 14081816, "step": 21535 }, { "epoch": 11.28930817610063, "grad_norm": 0.11438952386379242, "learning_rate": 2.374047524027202e-05, "loss": 0.5159, "num_input_tokens_seen": 14085176, "step": 21540 }, { "epoch": 11.291928721174004, "grad_norm": 0.21408583223819733, "learning_rate": 2.3729055608201243e-05, "loss": 0.3872, "num_input_tokens_seen": 14089624, "step": 21545 }, { "epoch": 11.29454926624738, "grad_norm": 0.37298595905303955, "learning_rate": 2.3717636241997537e-05, "loss": 0.4904, "num_input_tokens_seen": 14092088, "step": 21550 }, { "epoch": 11.297169811320755, "grad_norm": 0.15537506341934204, "learning_rate": 2.3706217144049696e-05, "loss": 0.3477, "num_input_tokens_seen": 14095512, "step": 21555 }, { "epoch": 11.29979035639413, "grad_norm": 0.13636955618858337, "learning_rate": 2.369479831674647e-05, "loss": 0.5082, "num_input_tokens_seen": 14099192, "step": 21560 }, { "epoch": 11.302410901467505, "grad_norm": 0.17644624412059784, "learning_rate": 2.3683379762476555e-05, "loss": 0.7481, "num_input_tokens_seen": 14102392, "step": 21565 }, { "epoch": 11.30503144654088, "grad_norm": 0.14398308098316193, "learning_rate": 2.3671961483628562e-05, "loss": 0.4363, "num_input_tokens_seen": 14106712, "step": 21570 }, { "epoch": 11.307651991614255, "grad_norm": 0.1538451910018921, "learning_rate": 2.3660543482591078e-05, "loss": 0.4449, "num_input_tokens_seen": 14110008, "step": 21575 }, { "epoch": 11.31027253668763, "grad_norm": 0.1880418211221695, "learning_rate": 2.364912576175261e-05, "loss": 0.4261, "num_input_tokens_seen": 14112632, "step": 21580 }, { "epoch": 11.312893081761006, "grad_norm": 0.12272528558969498, "learning_rate": 2.363770832350163e-05, "loss": 0.4047, "num_input_tokens_seen": 14117080, "step": 21585 }, { "epoch": 11.315513626834381, "grad_norm": 0.11595916002988815, "learning_rate": 2.362629117022653e-05, "loss": 0.4951, "num_input_tokens_seen": 14120568, "step": 21590 }, { "epoch": 11.318134171907756, "grad_norm": 0.12169615179300308, "learning_rate": 2.3614874304315633e-05, "loss": 0.5789, "num_input_tokens_seen": 14123448, "step": 21595 }, { "epoch": 11.320754716981131, "grad_norm": 0.14249879121780396, "learning_rate": 2.3603457728157237e-05, "loss": 0.607, "num_input_tokens_seen": 14128056, "step": 21600 }, { "epoch": 11.323375262054507, "grad_norm": 0.16734230518341064, "learning_rate": 2.3592041444139547e-05, "loss": 0.5583, "num_input_tokens_seen": 14130968, "step": 21605 }, { "epoch": 11.325995807127883, "grad_norm": 0.2460358440876007, "learning_rate": 2.358062545465073e-05, "loss": 0.5141, "num_input_tokens_seen": 14133400, "step": 21610 }, { "epoch": 11.328616352201259, "grad_norm": 0.12531879544258118, "learning_rate": 2.3569209762078866e-05, "loss": 0.4706, "num_input_tokens_seen": 14136056, "step": 21615 }, { "epoch": 11.331236897274634, "grad_norm": 0.15042977035045624, "learning_rate": 2.355779436881201e-05, "loss": 0.5061, "num_input_tokens_seen": 14139896, "step": 21620 }, { "epoch": 11.333857442348009, "grad_norm": 0.13616544008255005, "learning_rate": 2.3546379277238107e-05, "loss": 0.5698, "num_input_tokens_seen": 14142840, "step": 21625 }, { "epoch": 11.336477987421384, "grad_norm": 0.14750562608242035, "learning_rate": 2.353496448974508e-05, "loss": 0.5376, "num_input_tokens_seen": 14145688, "step": 21630 }, { "epoch": 11.33909853249476, "grad_norm": 0.18266768753528595, "learning_rate": 2.3523550008720763e-05, "loss": 0.5531, "num_input_tokens_seen": 14148376, "step": 21635 }, { "epoch": 11.341719077568134, "grad_norm": 0.23618078231811523, "learning_rate": 2.3512135836552945e-05, "loss": 0.491, "num_input_tokens_seen": 14150808, "step": 21640 }, { "epoch": 11.34433962264151, "grad_norm": 0.14116555452346802, "learning_rate": 2.350072197562933e-05, "loss": 0.4881, "num_input_tokens_seen": 14153848, "step": 21645 }, { "epoch": 11.346960167714885, "grad_norm": 0.16351528465747833, "learning_rate": 2.3489308428337586e-05, "loss": 0.4443, "num_input_tokens_seen": 14157016, "step": 21650 }, { "epoch": 11.34958071278826, "grad_norm": 0.12855075299739838, "learning_rate": 2.3477895197065275e-05, "loss": 0.3495, "num_input_tokens_seen": 14161208, "step": 21655 }, { "epoch": 11.352201257861635, "grad_norm": 0.153522789478302, "learning_rate": 2.346648228419992e-05, "loss": 0.4177, "num_input_tokens_seen": 14164344, "step": 21660 }, { "epoch": 11.35482180293501, "grad_norm": 0.18259389698505402, "learning_rate": 2.3455069692128982e-05, "loss": 0.3674, "num_input_tokens_seen": 14166680, "step": 21665 }, { "epoch": 11.357442348008385, "grad_norm": 0.10759323835372925, "learning_rate": 2.3443657423239832e-05, "loss": 0.2931, "num_input_tokens_seen": 14169304, "step": 21670 }, { "epoch": 11.36006289308176, "grad_norm": 0.1732158362865448, "learning_rate": 2.34322454799198e-05, "loss": 0.5499, "num_input_tokens_seen": 14172376, "step": 21675 }, { "epoch": 11.362683438155136, "grad_norm": 0.09113044291734695, "learning_rate": 2.3420833864556133e-05, "loss": 0.4951, "num_input_tokens_seen": 14178104, "step": 21680 }, { "epoch": 11.365303983228511, "grad_norm": 0.17027606070041656, "learning_rate": 2.3409422579535995e-05, "loss": 0.4044, "num_input_tokens_seen": 14181144, "step": 21685 }, { "epoch": 11.367924528301886, "grad_norm": 0.2996053993701935, "learning_rate": 2.339801162724651e-05, "loss": 0.4827, "num_input_tokens_seen": 14183416, "step": 21690 }, { "epoch": 11.370545073375261, "grad_norm": 0.44521400332450867, "learning_rate": 2.338660101007471e-05, "loss": 0.5519, "num_input_tokens_seen": 14186104, "step": 21695 }, { "epoch": 11.373165618448636, "grad_norm": 0.252375990152359, "learning_rate": 2.3375190730407578e-05, "loss": 0.4652, "num_input_tokens_seen": 14190168, "step": 21700 }, { "epoch": 11.375786163522013, "grad_norm": 0.16982871294021606, "learning_rate": 2.336378079063202e-05, "loss": 0.5084, "num_input_tokens_seen": 14192856, "step": 21705 }, { "epoch": 11.378406708595389, "grad_norm": 0.15845157206058502, "learning_rate": 2.335237119313484e-05, "loss": 0.3531, "num_input_tokens_seen": 14197112, "step": 21710 }, { "epoch": 11.381027253668764, "grad_norm": 0.13000766932964325, "learning_rate": 2.3340961940302807e-05, "loss": 0.3827, "num_input_tokens_seen": 14200024, "step": 21715 }, { "epoch": 11.383647798742139, "grad_norm": 0.24059195816516876, "learning_rate": 2.3329553034522616e-05, "loss": 0.6338, "num_input_tokens_seen": 14203800, "step": 21720 }, { "epoch": 11.386268343815514, "grad_norm": 0.1810581088066101, "learning_rate": 2.3318144478180866e-05, "loss": 0.6989, "num_input_tokens_seen": 14207128, "step": 21725 }, { "epoch": 11.38888888888889, "grad_norm": 0.14544031023979187, "learning_rate": 2.330673627366411e-05, "loss": 0.475, "num_input_tokens_seen": 14210168, "step": 21730 }, { "epoch": 11.391509433962264, "grad_norm": 0.2193608283996582, "learning_rate": 2.329532842335881e-05, "loss": 0.5201, "num_input_tokens_seen": 14216184, "step": 21735 }, { "epoch": 11.39412997903564, "grad_norm": 0.08527103066444397, "learning_rate": 2.328392092965135e-05, "loss": 0.4305, "num_input_tokens_seen": 14219576, "step": 21740 }, { "epoch": 11.396750524109015, "grad_norm": 0.1152419000864029, "learning_rate": 2.3272513794928054e-05, "loss": 0.5749, "num_input_tokens_seen": 14223544, "step": 21745 }, { "epoch": 11.39937106918239, "grad_norm": 0.16477854549884796, "learning_rate": 2.326110702157516e-05, "loss": 0.5087, "num_input_tokens_seen": 14227384, "step": 21750 }, { "epoch": 11.401991614255765, "grad_norm": 0.1552707403898239, "learning_rate": 2.3249700611978847e-05, "loss": 0.4614, "num_input_tokens_seen": 14229816, "step": 21755 }, { "epoch": 11.40461215932914, "grad_norm": 0.1301165670156479, "learning_rate": 2.323829456852519e-05, "loss": 0.6066, "num_input_tokens_seen": 14233176, "step": 21760 }, { "epoch": 11.407232704402515, "grad_norm": 0.1852768361568451, "learning_rate": 2.3226888893600223e-05, "loss": 0.5269, "num_input_tokens_seen": 14236728, "step": 21765 }, { "epoch": 11.40985324947589, "grad_norm": 0.09217103570699692, "learning_rate": 2.3215483589589865e-05, "loss": 0.3722, "num_input_tokens_seen": 14240184, "step": 21770 }, { "epoch": 11.412473794549266, "grad_norm": 0.2024373859167099, "learning_rate": 2.3204078658879974e-05, "loss": 0.4256, "num_input_tokens_seen": 14242808, "step": 21775 }, { "epoch": 11.415094339622641, "grad_norm": 0.27964791655540466, "learning_rate": 2.3192674103856337e-05, "loss": 0.5951, "num_input_tokens_seen": 14245400, "step": 21780 }, { "epoch": 11.417714884696016, "grad_norm": 0.18180054426193237, "learning_rate": 2.318126992690467e-05, "loss": 0.4599, "num_input_tokens_seen": 14247768, "step": 21785 }, { "epoch": 11.420335429769391, "grad_norm": 0.15316356718540192, "learning_rate": 2.3169866130410585e-05, "loss": 0.3737, "num_input_tokens_seen": 14254520, "step": 21790 }, { "epoch": 11.422955974842766, "grad_norm": 0.5314421057701111, "learning_rate": 2.315846271675962e-05, "loss": 0.4877, "num_input_tokens_seen": 14257880, "step": 21795 }, { "epoch": 11.425576519916142, "grad_norm": 0.12557241320610046, "learning_rate": 2.3147059688337244e-05, "loss": 0.39, "num_input_tokens_seen": 14261432, "step": 21800 }, { "epoch": 11.428197064989519, "grad_norm": 0.07313954085111618, "learning_rate": 2.313565704752884e-05, "loss": 0.3997, "num_input_tokens_seen": 14264856, "step": 21805 }, { "epoch": 11.430817610062894, "grad_norm": 0.17936433851718903, "learning_rate": 2.3124254796719718e-05, "loss": 0.413, "num_input_tokens_seen": 14267672, "step": 21810 }, { "epoch": 11.433438155136269, "grad_norm": 0.14103272557258606, "learning_rate": 2.3112852938295088e-05, "loss": 0.5093, "num_input_tokens_seen": 14271480, "step": 21815 }, { "epoch": 11.436058700209644, "grad_norm": 0.23804497718811035, "learning_rate": 2.3101451474640104e-05, "loss": 0.466, "num_input_tokens_seen": 14277400, "step": 21820 }, { "epoch": 11.43867924528302, "grad_norm": 0.14226099848747253, "learning_rate": 2.3090050408139808e-05, "loss": 0.4858, "num_input_tokens_seen": 14280664, "step": 21825 }, { "epoch": 11.441299790356394, "grad_norm": 0.27467700839042664, "learning_rate": 2.307864974117917e-05, "loss": 0.4426, "num_input_tokens_seen": 14284696, "step": 21830 }, { "epoch": 11.44392033542977, "grad_norm": 0.12470100075006485, "learning_rate": 2.3067249476143087e-05, "loss": 0.5584, "num_input_tokens_seen": 14287800, "step": 21835 }, { "epoch": 11.446540880503145, "grad_norm": 0.2318749874830246, "learning_rate": 2.3055849615416363e-05, "loss": 0.5623, "num_input_tokens_seen": 14289848, "step": 21840 }, { "epoch": 11.44916142557652, "grad_norm": 0.19759853184223175, "learning_rate": 2.304445016138372e-05, "loss": 0.4094, "num_input_tokens_seen": 14292920, "step": 21845 }, { "epoch": 11.451781970649895, "grad_norm": 0.30742841958999634, "learning_rate": 2.303305111642981e-05, "loss": 0.4336, "num_input_tokens_seen": 14296120, "step": 21850 }, { "epoch": 11.45440251572327, "grad_norm": 0.23212291300296783, "learning_rate": 2.3021652482939144e-05, "loss": 0.5016, "num_input_tokens_seen": 14299384, "step": 21855 }, { "epoch": 11.457023060796645, "grad_norm": 0.1764003038406372, "learning_rate": 2.3010254263296214e-05, "loss": 0.4474, "num_input_tokens_seen": 14302968, "step": 21860 }, { "epoch": 11.45964360587002, "grad_norm": 0.24561448395252228, "learning_rate": 2.2998856459885396e-05, "loss": 0.4275, "num_input_tokens_seen": 14305400, "step": 21865 }, { "epoch": 11.462264150943396, "grad_norm": 0.1992570161819458, "learning_rate": 2.298745907509097e-05, "loss": 0.4912, "num_input_tokens_seen": 14309368, "step": 21870 }, { "epoch": 11.464884696016771, "grad_norm": 0.1055312529206276, "learning_rate": 2.2976062111297156e-05, "loss": 0.4297, "num_input_tokens_seen": 14312632, "step": 21875 }, { "epoch": 11.467505241090146, "grad_norm": 0.1918456107378006, "learning_rate": 2.296466557088805e-05, "loss": 0.3961, "num_input_tokens_seen": 14315480, "step": 21880 }, { "epoch": 11.470125786163521, "grad_norm": 0.3834327161312103, "learning_rate": 2.2953269456247688e-05, "loss": 0.6002, "num_input_tokens_seen": 14318264, "step": 21885 }, { "epoch": 11.472746331236896, "grad_norm": 0.1522391140460968, "learning_rate": 2.2941873769760006e-05, "loss": 0.5331, "num_input_tokens_seen": 14320952, "step": 21890 }, { "epoch": 11.475366876310272, "grad_norm": 0.15098220109939575, "learning_rate": 2.2930478513808844e-05, "loss": 0.4521, "num_input_tokens_seen": 14324280, "step": 21895 }, { "epoch": 11.477987421383649, "grad_norm": 0.17332281172275543, "learning_rate": 2.2919083690777976e-05, "loss": 0.4168, "num_input_tokens_seen": 14326520, "step": 21900 }, { "epoch": 11.480607966457024, "grad_norm": 0.2070881724357605, "learning_rate": 2.2907689303051062e-05, "loss": 0.5032, "num_input_tokens_seen": 14329272, "step": 21905 }, { "epoch": 11.483228511530399, "grad_norm": 0.11970209330320358, "learning_rate": 2.289629535301167e-05, "loss": 0.4982, "num_input_tokens_seen": 14332312, "step": 21910 }, { "epoch": 11.485849056603774, "grad_norm": 0.2501019239425659, "learning_rate": 2.288490184304329e-05, "loss": 0.4859, "num_input_tokens_seen": 14334744, "step": 21915 }, { "epoch": 11.48846960167715, "grad_norm": 0.1488906890153885, "learning_rate": 2.2873508775529308e-05, "loss": 0.4806, "num_input_tokens_seen": 14338872, "step": 21920 }, { "epoch": 11.491090146750524, "grad_norm": 0.171982541680336, "learning_rate": 2.2862116152853034e-05, "loss": 0.4861, "num_input_tokens_seen": 14342232, "step": 21925 }, { "epoch": 11.4937106918239, "grad_norm": 0.19698907434940338, "learning_rate": 2.285072397739767e-05, "loss": 0.5153, "num_input_tokens_seen": 14345432, "step": 21930 }, { "epoch": 11.496331236897275, "grad_norm": 0.16185350716114044, "learning_rate": 2.283933225154634e-05, "loss": 0.6333, "num_input_tokens_seen": 14348792, "step": 21935 }, { "epoch": 11.49895178197065, "grad_norm": 0.16434748470783234, "learning_rate": 2.2827940977682045e-05, "loss": 0.4332, "num_input_tokens_seen": 14351736, "step": 21940 }, { "epoch": 11.501572327044025, "grad_norm": 0.3218834102153778, "learning_rate": 2.2816550158187717e-05, "loss": 0.5619, "num_input_tokens_seen": 14354680, "step": 21945 }, { "epoch": 11.5041928721174, "grad_norm": 0.23340272903442383, "learning_rate": 2.2805159795446182e-05, "loss": 0.5556, "num_input_tokens_seen": 14357432, "step": 21950 }, { "epoch": 11.506813417190775, "grad_norm": 0.11856938898563385, "learning_rate": 2.2793769891840184e-05, "loss": 0.4839, "num_input_tokens_seen": 14360728, "step": 21955 }, { "epoch": 11.50943396226415, "grad_norm": 0.18375322222709656, "learning_rate": 2.278238044975236e-05, "loss": 0.5735, "num_input_tokens_seen": 14364376, "step": 21960 }, { "epoch": 11.512054507337526, "grad_norm": 0.18602438271045685, "learning_rate": 2.2770991471565236e-05, "loss": 0.3909, "num_input_tokens_seen": 14367960, "step": 21965 }, { "epoch": 11.514675052410901, "grad_norm": 0.20001819729804993, "learning_rate": 2.2759602959661265e-05, "loss": 0.5426, "num_input_tokens_seen": 14372696, "step": 21970 }, { "epoch": 11.517295597484276, "grad_norm": 0.13907139003276825, "learning_rate": 2.2748214916422796e-05, "loss": 0.4258, "num_input_tokens_seen": 14376280, "step": 21975 }, { "epoch": 11.519916142557651, "grad_norm": 0.1351700872182846, "learning_rate": 2.2736827344232082e-05, "loss": 0.4008, "num_input_tokens_seen": 14379960, "step": 21980 }, { "epoch": 11.522536687631026, "grad_norm": 0.15796338021755219, "learning_rate": 2.272544024547126e-05, "loss": 0.4969, "num_input_tokens_seen": 14383000, "step": 21985 }, { "epoch": 11.525157232704402, "grad_norm": 0.16105039417743683, "learning_rate": 2.27140536225224e-05, "loss": 0.5165, "num_input_tokens_seen": 14386488, "step": 21990 }, { "epoch": 11.527777777777779, "grad_norm": 0.16235199570655823, "learning_rate": 2.270266747776744e-05, "loss": 0.3512, "num_input_tokens_seen": 14390264, "step": 21995 }, { "epoch": 11.530398322851154, "grad_norm": 0.16641949117183685, "learning_rate": 2.2691281813588226e-05, "loss": 0.4427, "num_input_tokens_seen": 14392920, "step": 22000 }, { "epoch": 11.533018867924529, "grad_norm": 0.16193807125091553, "learning_rate": 2.267989663236652e-05, "loss": 0.4614, "num_input_tokens_seen": 14396568, "step": 22005 }, { "epoch": 11.535639412997904, "grad_norm": 0.18489615619182587, "learning_rate": 2.2668511936483975e-05, "loss": 0.5123, "num_input_tokens_seen": 14400120, "step": 22010 }, { "epoch": 11.53825995807128, "grad_norm": 0.1400488317012787, "learning_rate": 2.2657127728322124e-05, "loss": 0.4997, "num_input_tokens_seen": 14403704, "step": 22015 }, { "epoch": 11.540880503144654, "grad_norm": 0.18897464871406555, "learning_rate": 2.264574401026244e-05, "loss": 0.3914, "num_input_tokens_seen": 14405976, "step": 22020 }, { "epoch": 11.54350104821803, "grad_norm": 0.1799473911523819, "learning_rate": 2.2634360784686242e-05, "loss": 0.4144, "num_input_tokens_seen": 14408600, "step": 22025 }, { "epoch": 11.546121593291405, "grad_norm": 0.15778985619544983, "learning_rate": 2.2622978053974776e-05, "loss": 0.3687, "num_input_tokens_seen": 14411352, "step": 22030 }, { "epoch": 11.54874213836478, "grad_norm": 0.29240554571151733, "learning_rate": 2.2611595820509186e-05, "loss": 0.6495, "num_input_tokens_seen": 14414552, "step": 22035 }, { "epoch": 11.551362683438155, "grad_norm": 0.1054544523358345, "learning_rate": 2.2600214086670494e-05, "loss": 0.4197, "num_input_tokens_seen": 14418680, "step": 22040 }, { "epoch": 11.55398322851153, "grad_norm": 0.199127197265625, "learning_rate": 2.2588832854839644e-05, "loss": 0.5549, "num_input_tokens_seen": 14421432, "step": 22045 }, { "epoch": 11.556603773584905, "grad_norm": 0.22129836678504944, "learning_rate": 2.2577452127397463e-05, "loss": 0.5731, "num_input_tokens_seen": 14424280, "step": 22050 }, { "epoch": 11.55922431865828, "grad_norm": 0.1870427280664444, "learning_rate": 2.256607190672464e-05, "loss": 0.5582, "num_input_tokens_seen": 14427064, "step": 22055 }, { "epoch": 11.561844863731656, "grad_norm": 0.14153428375720978, "learning_rate": 2.2554692195201817e-05, "loss": 0.4258, "num_input_tokens_seen": 14430584, "step": 22060 }, { "epoch": 11.564465408805031, "grad_norm": 0.217091903090477, "learning_rate": 2.254331299520948e-05, "loss": 0.5213, "num_input_tokens_seen": 14433496, "step": 22065 }, { "epoch": 11.567085953878406, "grad_norm": 0.2573617100715637, "learning_rate": 2.253193430912805e-05, "loss": 0.5713, "num_input_tokens_seen": 14436632, "step": 22070 }, { "epoch": 11.569706498951781, "grad_norm": 0.2905290126800537, "learning_rate": 2.2520556139337803e-05, "loss": 0.4269, "num_input_tokens_seen": 14439448, "step": 22075 }, { "epoch": 11.572327044025156, "grad_norm": 0.17183972895145416, "learning_rate": 2.250917848821892e-05, "loss": 0.5005, "num_input_tokens_seen": 14442328, "step": 22080 }, { "epoch": 11.574947589098532, "grad_norm": 0.18557879328727722, "learning_rate": 2.2497801358151472e-05, "loss": 0.4653, "num_input_tokens_seen": 14444984, "step": 22085 }, { "epoch": 11.577568134171909, "grad_norm": 0.10968386381864548, "learning_rate": 2.248642475151545e-05, "loss": 0.6287, "num_input_tokens_seen": 14448280, "step": 22090 }, { "epoch": 11.580188679245284, "grad_norm": 0.1679428517818451, "learning_rate": 2.2475048670690678e-05, "loss": 0.4467, "num_input_tokens_seen": 14451864, "step": 22095 }, { "epoch": 11.582809224318659, "grad_norm": 0.1946684569120407, "learning_rate": 2.2463673118056927e-05, "loss": 0.5006, "num_input_tokens_seen": 14454392, "step": 22100 }, { "epoch": 11.585429769392034, "grad_norm": 0.2132457196712494, "learning_rate": 2.245229809599383e-05, "loss": 0.4589, "num_input_tokens_seen": 14459928, "step": 22105 }, { "epoch": 11.58805031446541, "grad_norm": 0.13838738203048706, "learning_rate": 2.2440923606880896e-05, "loss": 0.4878, "num_input_tokens_seen": 14462616, "step": 22110 }, { "epoch": 11.590670859538784, "grad_norm": 0.1417401134967804, "learning_rate": 2.242954965309755e-05, "loss": 0.4664, "num_input_tokens_seen": 14466072, "step": 22115 }, { "epoch": 11.59329140461216, "grad_norm": 0.22381210327148438, "learning_rate": 2.2418176237023088e-05, "loss": 0.4224, "num_input_tokens_seen": 14468728, "step": 22120 }, { "epoch": 11.595911949685535, "grad_norm": 0.15893296897411346, "learning_rate": 2.2406803361036706e-05, "loss": 0.5151, "num_input_tokens_seen": 14471672, "step": 22125 }, { "epoch": 11.59853249475891, "grad_norm": 0.22705304622650146, "learning_rate": 2.2395431027517475e-05, "loss": 0.3812, "num_input_tokens_seen": 14474840, "step": 22130 }, { "epoch": 11.601153039832285, "grad_norm": 0.1406232863664627, "learning_rate": 2.2384059238844364e-05, "loss": 0.5543, "num_input_tokens_seen": 14478648, "step": 22135 }, { "epoch": 11.60377358490566, "grad_norm": 0.20888976752758026, "learning_rate": 2.2372687997396215e-05, "loss": 0.3667, "num_input_tokens_seen": 14482968, "step": 22140 }, { "epoch": 11.606394129979035, "grad_norm": 0.3092648386955261, "learning_rate": 2.2361317305551756e-05, "loss": 0.3719, "num_input_tokens_seen": 14486712, "step": 22145 }, { "epoch": 11.60901467505241, "grad_norm": 0.1693452000617981, "learning_rate": 2.2349947165689615e-05, "loss": 0.5268, "num_input_tokens_seen": 14489624, "step": 22150 }, { "epoch": 11.611635220125786, "grad_norm": 0.2347426563501358, "learning_rate": 2.2338577580188296e-05, "loss": 0.4824, "num_input_tokens_seen": 14493016, "step": 22155 }, { "epoch": 11.614255765199161, "grad_norm": 0.25804585218429565, "learning_rate": 2.2327208551426188e-05, "loss": 0.3605, "num_input_tokens_seen": 14495928, "step": 22160 }, { "epoch": 11.616876310272536, "grad_norm": 0.1732659786939621, "learning_rate": 2.2315840081781546e-05, "loss": 0.492, "num_input_tokens_seen": 14499416, "step": 22165 }, { "epoch": 11.619496855345911, "grad_norm": 0.1986047625541687, "learning_rate": 2.230447217363255e-05, "loss": 0.712, "num_input_tokens_seen": 14502232, "step": 22170 }, { "epoch": 11.622117400419286, "grad_norm": 0.16354870796203613, "learning_rate": 2.2293104829357212e-05, "loss": 0.6055, "num_input_tokens_seen": 14505144, "step": 22175 }, { "epoch": 11.624737945492662, "grad_norm": 0.11657439917325974, "learning_rate": 2.2281738051333467e-05, "loss": 0.5138, "num_input_tokens_seen": 14509208, "step": 22180 }, { "epoch": 11.627358490566039, "grad_norm": 0.1857132762670517, "learning_rate": 2.2270371841939105e-05, "loss": 0.454, "num_input_tokens_seen": 14512312, "step": 22185 }, { "epoch": 11.629979035639414, "grad_norm": 0.09682916104793549, "learning_rate": 2.2259006203551824e-05, "loss": 0.4959, "num_input_tokens_seen": 14515960, "step": 22190 }, { "epoch": 11.632599580712789, "grad_norm": 0.1470014899969101, "learning_rate": 2.2247641138549168e-05, "loss": 0.4413, "num_input_tokens_seen": 14518968, "step": 22195 }, { "epoch": 11.635220125786164, "grad_norm": 0.13704945147037506, "learning_rate": 2.2236276649308583e-05, "loss": 0.5696, "num_input_tokens_seen": 14522936, "step": 22200 }, { "epoch": 11.63784067085954, "grad_norm": 0.17508341372013092, "learning_rate": 2.2224912738207395e-05, "loss": 0.6494, "num_input_tokens_seen": 14525592, "step": 22205 }, { "epoch": 11.640461215932914, "grad_norm": 0.22643034160137177, "learning_rate": 2.2213549407622796e-05, "loss": 0.4902, "num_input_tokens_seen": 14529112, "step": 22210 }, { "epoch": 11.64308176100629, "grad_norm": 0.19315993785858154, "learning_rate": 2.220218665993188e-05, "loss": 0.4129, "num_input_tokens_seen": 14532536, "step": 22215 }, { "epoch": 11.645702306079665, "grad_norm": 0.2573029398918152, "learning_rate": 2.21908244975116e-05, "loss": 0.4468, "num_input_tokens_seen": 14535672, "step": 22220 }, { "epoch": 11.64832285115304, "grad_norm": 0.14962244033813477, "learning_rate": 2.2179462922738776e-05, "loss": 0.5178, "num_input_tokens_seen": 14538712, "step": 22225 }, { "epoch": 11.650943396226415, "grad_norm": 0.16011175513267517, "learning_rate": 2.2168101937990135e-05, "loss": 0.437, "num_input_tokens_seen": 14541624, "step": 22230 }, { "epoch": 11.65356394129979, "grad_norm": 0.11251433938741684, "learning_rate": 2.2156741545642253e-05, "loss": 0.4253, "num_input_tokens_seen": 14544600, "step": 22235 }, { "epoch": 11.656184486373165, "grad_norm": 0.15734180808067322, "learning_rate": 2.2145381748071606e-05, "loss": 0.3904, "num_input_tokens_seen": 14547800, "step": 22240 }, { "epoch": 11.65880503144654, "grad_norm": 0.17461998760700226, "learning_rate": 2.2134022547654538e-05, "loss": 0.6279, "num_input_tokens_seen": 14550072, "step": 22245 }, { "epoch": 11.661425576519916, "grad_norm": 0.19342483580112457, "learning_rate": 2.212266394676725e-05, "loss": 0.475, "num_input_tokens_seen": 14554584, "step": 22250 }, { "epoch": 11.664046121593291, "grad_norm": 0.12356709688901901, "learning_rate": 2.2111305947785836e-05, "loss": 0.4668, "num_input_tokens_seen": 14557912, "step": 22255 }, { "epoch": 11.666666666666666, "grad_norm": 0.13240869343280792, "learning_rate": 2.2099948553086263e-05, "loss": 0.4417, "num_input_tokens_seen": 14561304, "step": 22260 }, { "epoch": 11.669287211740041, "grad_norm": 0.15789422392845154, "learning_rate": 2.2088591765044365e-05, "loss": 0.5624, "num_input_tokens_seen": 14564856, "step": 22265 }, { "epoch": 11.671907756813416, "grad_norm": 0.17308995127677917, "learning_rate": 2.207723558603586e-05, "loss": 0.4699, "num_input_tokens_seen": 14567960, "step": 22270 }, { "epoch": 11.674528301886792, "grad_norm": 0.16439443826675415, "learning_rate": 2.206588001843633e-05, "loss": 0.4594, "num_input_tokens_seen": 14570840, "step": 22275 }, { "epoch": 11.677148846960169, "grad_norm": 0.16606895625591278, "learning_rate": 2.2054525064621216e-05, "loss": 0.4711, "num_input_tokens_seen": 14575128, "step": 22280 }, { "epoch": 11.679769392033544, "grad_norm": 0.21286816895008087, "learning_rate": 2.2043170726965858e-05, "loss": 0.5911, "num_input_tokens_seen": 14578680, "step": 22285 }, { "epoch": 11.682389937106919, "grad_norm": 0.2649058699607849, "learning_rate": 2.2031817007845447e-05, "loss": 0.5611, "num_input_tokens_seen": 14581400, "step": 22290 }, { "epoch": 11.685010482180294, "grad_norm": 0.26946499943733215, "learning_rate": 2.202046390963506e-05, "loss": 0.4638, "num_input_tokens_seen": 14584600, "step": 22295 }, { "epoch": 11.68763102725367, "grad_norm": 0.1541987806558609, "learning_rate": 2.2009111434709627e-05, "loss": 0.4603, "num_input_tokens_seen": 14587704, "step": 22300 }, { "epoch": 11.690251572327044, "grad_norm": 0.16330623626708984, "learning_rate": 2.1997759585443968e-05, "loss": 0.4565, "num_input_tokens_seen": 14590264, "step": 22305 }, { "epoch": 11.69287211740042, "grad_norm": 0.16121439635753632, "learning_rate": 2.198640836421275e-05, "loss": 0.4793, "num_input_tokens_seen": 14593528, "step": 22310 }, { "epoch": 11.695492662473795, "grad_norm": 0.14228959381580353, "learning_rate": 2.1975057773390515e-05, "loss": 0.5422, "num_input_tokens_seen": 14596568, "step": 22315 }, { "epoch": 11.69811320754717, "grad_norm": 0.3783232867717743, "learning_rate": 2.1963707815351683e-05, "loss": 0.6315, "num_input_tokens_seen": 14599288, "step": 22320 }, { "epoch": 11.700733752620545, "grad_norm": 0.17402192950248718, "learning_rate": 2.1952358492470544e-05, "loss": 0.5711, "num_input_tokens_seen": 14602648, "step": 22325 }, { "epoch": 11.70335429769392, "grad_norm": 0.1628820151090622, "learning_rate": 2.194100980712124e-05, "loss": 0.4259, "num_input_tokens_seen": 14605848, "step": 22330 }, { "epoch": 11.705974842767295, "grad_norm": 0.11587008088827133, "learning_rate": 2.1929661761677774e-05, "loss": 0.3711, "num_input_tokens_seen": 14609112, "step": 22335 }, { "epoch": 11.70859538784067, "grad_norm": 0.1633106768131256, "learning_rate": 2.191831435851405e-05, "loss": 0.7796, "num_input_tokens_seen": 14613208, "step": 22340 }, { "epoch": 11.711215932914046, "grad_norm": 0.10109000653028488, "learning_rate": 2.1906967600003793e-05, "loss": 0.4941, "num_input_tokens_seen": 14616152, "step": 22345 }, { "epoch": 11.713836477987421, "grad_norm": 0.12089098244905472, "learning_rate": 2.1895621488520636e-05, "loss": 0.4572, "num_input_tokens_seen": 14619128, "step": 22350 }, { "epoch": 11.716457023060796, "grad_norm": 0.21373271942138672, "learning_rate": 2.1884276026438044e-05, "loss": 0.5442, "num_input_tokens_seen": 14623192, "step": 22355 }, { "epoch": 11.719077568134171, "grad_norm": 0.16391417384147644, "learning_rate": 2.187293121612937e-05, "loss": 0.4966, "num_input_tokens_seen": 14626488, "step": 22360 }, { "epoch": 11.721698113207546, "grad_norm": 0.12046743184328079, "learning_rate": 2.1861587059967804e-05, "loss": 0.6051, "num_input_tokens_seen": 14629496, "step": 22365 }, { "epoch": 11.724318658280922, "grad_norm": 0.12235430628061295, "learning_rate": 2.1850243560326414e-05, "loss": 0.4502, "num_input_tokens_seen": 14632504, "step": 22370 }, { "epoch": 11.726939203354299, "grad_norm": 0.13539566099643707, "learning_rate": 2.1838900719578143e-05, "loss": 0.4828, "num_input_tokens_seen": 14635576, "step": 22375 }, { "epoch": 11.729559748427674, "grad_norm": 0.09032762795686722, "learning_rate": 2.1827558540095772e-05, "loss": 0.436, "num_input_tokens_seen": 14638488, "step": 22380 }, { "epoch": 11.732180293501049, "grad_norm": 0.11837539076805115, "learning_rate": 2.1816217024251966e-05, "loss": 0.5954, "num_input_tokens_seen": 14641720, "step": 22385 }, { "epoch": 11.734800838574424, "grad_norm": 0.14187392592430115, "learning_rate": 2.1804876174419242e-05, "loss": 0.5142, "num_input_tokens_seen": 14644792, "step": 22390 }, { "epoch": 11.7374213836478, "grad_norm": 0.13852113485336304, "learning_rate": 2.179353599296996e-05, "loss": 0.4963, "num_input_tokens_seen": 14647384, "step": 22395 }, { "epoch": 11.740041928721174, "grad_norm": 0.18794235587120056, "learning_rate": 2.1782196482276368e-05, "loss": 0.4399, "num_input_tokens_seen": 14650584, "step": 22400 }, { "epoch": 11.74266247379455, "grad_norm": 0.15611715614795685, "learning_rate": 2.1770857644710567e-05, "loss": 0.4438, "num_input_tokens_seen": 14653464, "step": 22405 }, { "epoch": 11.745283018867925, "grad_norm": 0.16789610683918, "learning_rate": 2.1759519482644495e-05, "loss": 0.4219, "num_input_tokens_seen": 14656376, "step": 22410 }, { "epoch": 11.7479035639413, "grad_norm": 0.23996834456920624, "learning_rate": 2.174818199844999e-05, "loss": 0.3398, "num_input_tokens_seen": 14659416, "step": 22415 }, { "epoch": 11.750524109014675, "grad_norm": 0.1719346046447754, "learning_rate": 2.173684519449872e-05, "loss": 0.4108, "num_input_tokens_seen": 14662232, "step": 22420 }, { "epoch": 11.75314465408805, "grad_norm": 0.13684561848640442, "learning_rate": 2.1725509073162188e-05, "loss": 0.487, "num_input_tokens_seen": 14664920, "step": 22425 }, { "epoch": 11.755765199161425, "grad_norm": 0.2587272524833679, "learning_rate": 2.171417363681181e-05, "loss": 0.5618, "num_input_tokens_seen": 14668056, "step": 22430 }, { "epoch": 11.7583857442348, "grad_norm": 0.14059069752693176, "learning_rate": 2.1702838887818816e-05, "loss": 0.4668, "num_input_tokens_seen": 14671032, "step": 22435 }, { "epoch": 11.761006289308176, "grad_norm": 0.10513318330049515, "learning_rate": 2.1691504828554315e-05, "loss": 0.4202, "num_input_tokens_seen": 14674360, "step": 22440 }, { "epoch": 11.76362683438155, "grad_norm": 0.16089388728141785, "learning_rate": 2.1680171461389265e-05, "loss": 0.5634, "num_input_tokens_seen": 14677720, "step": 22445 }, { "epoch": 11.766247379454926, "grad_norm": 0.164411261677742, "learning_rate": 2.166883878869446e-05, "loss": 0.4544, "num_input_tokens_seen": 14680216, "step": 22450 }, { "epoch": 11.768867924528301, "grad_norm": 0.16438163816928864, "learning_rate": 2.1657506812840582e-05, "loss": 0.5123, "num_input_tokens_seen": 14683608, "step": 22455 }, { "epoch": 11.771488469601676, "grad_norm": 0.18959060311317444, "learning_rate": 2.164617553619814e-05, "loss": 0.5406, "num_input_tokens_seen": 14686872, "step": 22460 }, { "epoch": 11.774109014675052, "grad_norm": 0.19450905919075012, "learning_rate": 2.1634844961137517e-05, "loss": 0.4571, "num_input_tokens_seen": 14688824, "step": 22465 }, { "epoch": 11.776729559748428, "grad_norm": 0.20776204764842987, "learning_rate": 2.162351509002893e-05, "loss": 0.4054, "num_input_tokens_seen": 14691192, "step": 22470 }, { "epoch": 11.779350104821804, "grad_norm": 0.20825593173503876, "learning_rate": 2.161218592524248e-05, "loss": 0.4723, "num_input_tokens_seen": 14694072, "step": 22475 }, { "epoch": 11.781970649895179, "grad_norm": 0.1535758227109909, "learning_rate": 2.160085746914807e-05, "loss": 0.5902, "num_input_tokens_seen": 14697976, "step": 22480 }, { "epoch": 11.784591194968554, "grad_norm": 0.26671189069747925, "learning_rate": 2.1589529724115503e-05, "loss": 0.5011, "num_input_tokens_seen": 14701368, "step": 22485 }, { "epoch": 11.78721174004193, "grad_norm": 0.23341183364391327, "learning_rate": 2.1578202692514407e-05, "loss": 0.5249, "num_input_tokens_seen": 14704888, "step": 22490 }, { "epoch": 11.789832285115304, "grad_norm": 0.175669327378273, "learning_rate": 2.1566876376714274e-05, "loss": 0.4349, "num_input_tokens_seen": 14707928, "step": 22495 }, { "epoch": 11.79245283018868, "grad_norm": 0.11076457053422928, "learning_rate": 2.155555077908443e-05, "loss": 0.6073, "num_input_tokens_seen": 14711800, "step": 22500 }, { "epoch": 11.795073375262055, "grad_norm": 0.12301322817802429, "learning_rate": 2.1544225901994077e-05, "loss": 0.5791, "num_input_tokens_seen": 14719000, "step": 22505 }, { "epoch": 11.79769392033543, "grad_norm": 0.1298670619726181, "learning_rate": 2.1532901747812234e-05, "loss": 0.3772, "num_input_tokens_seen": 14723192, "step": 22510 }, { "epoch": 11.800314465408805, "grad_norm": 0.22644038498401642, "learning_rate": 2.1521578318907787e-05, "loss": 0.5371, "num_input_tokens_seen": 14725496, "step": 22515 }, { "epoch": 11.80293501048218, "grad_norm": 0.11221186816692352, "learning_rate": 2.1510255617649476e-05, "loss": 0.4721, "num_input_tokens_seen": 14729272, "step": 22520 }, { "epoch": 11.805555555555555, "grad_norm": 0.1251053363084793, "learning_rate": 2.149893364640587e-05, "loss": 0.5761, "num_input_tokens_seen": 14732280, "step": 22525 }, { "epoch": 11.80817610062893, "grad_norm": 0.13190677762031555, "learning_rate": 2.1487612407545417e-05, "loss": 0.3847, "num_input_tokens_seen": 14737784, "step": 22530 }, { "epoch": 11.810796645702306, "grad_norm": 0.11336831748485565, "learning_rate": 2.147629190343637e-05, "loss": 0.5456, "num_input_tokens_seen": 14741048, "step": 22535 }, { "epoch": 11.81341719077568, "grad_norm": 0.14445358514785767, "learning_rate": 2.1464972136446852e-05, "loss": 0.434, "num_input_tokens_seen": 14744440, "step": 22540 }, { "epoch": 11.816037735849056, "grad_norm": 0.17346730828285217, "learning_rate": 2.1453653108944838e-05, "loss": 0.6658, "num_input_tokens_seen": 14747032, "step": 22545 }, { "epoch": 11.818658280922431, "grad_norm": 0.20022249221801758, "learning_rate": 2.1442334823298128e-05, "loss": 0.579, "num_input_tokens_seen": 14749528, "step": 22550 }, { "epoch": 11.821278825995806, "grad_norm": 0.13467979431152344, "learning_rate": 2.1431017281874386e-05, "loss": 0.558, "num_input_tokens_seen": 14752696, "step": 22555 }, { "epoch": 11.823899371069182, "grad_norm": 0.15727321803569794, "learning_rate": 2.1419700487041123e-05, "loss": 0.5239, "num_input_tokens_seen": 14755800, "step": 22560 }, { "epoch": 11.826519916142558, "grad_norm": 0.1465255469083786, "learning_rate": 2.140838444116566e-05, "loss": 0.3996, "num_input_tokens_seen": 14758264, "step": 22565 }, { "epoch": 11.829140461215934, "grad_norm": 0.13744384050369263, "learning_rate": 2.13970691466152e-05, "loss": 0.4222, "num_input_tokens_seen": 14760984, "step": 22570 }, { "epoch": 11.831761006289309, "grad_norm": 0.35945427417755127, "learning_rate": 2.1385754605756773e-05, "loss": 0.4462, "num_input_tokens_seen": 14764472, "step": 22575 }, { "epoch": 11.834381551362684, "grad_norm": 0.16531440615653992, "learning_rate": 2.137444082095724e-05, "loss": 0.425, "num_input_tokens_seen": 14767960, "step": 22580 }, { "epoch": 11.83700209643606, "grad_norm": 0.10731576383113861, "learning_rate": 2.1363127794583335e-05, "loss": 0.5159, "num_input_tokens_seen": 14770360, "step": 22585 }, { "epoch": 11.839622641509434, "grad_norm": 0.1506408154964447, "learning_rate": 2.1351815529001605e-05, "loss": 0.576, "num_input_tokens_seen": 14772920, "step": 22590 }, { "epoch": 11.84224318658281, "grad_norm": 0.12473611533641815, "learning_rate": 2.1340504026578438e-05, "loss": 0.4831, "num_input_tokens_seen": 14776024, "step": 22595 }, { "epoch": 11.844863731656185, "grad_norm": 0.15476500988006592, "learning_rate": 2.1329193289680084e-05, "loss": 0.4573, "num_input_tokens_seen": 14778776, "step": 22600 }, { "epoch": 11.84748427672956, "grad_norm": 0.2242746204137802, "learning_rate": 2.1317883320672608e-05, "loss": 0.4494, "num_input_tokens_seen": 14781176, "step": 22605 }, { "epoch": 11.850104821802935, "grad_norm": 0.18189869821071625, "learning_rate": 2.1306574121921942e-05, "loss": 0.5776, "num_input_tokens_seen": 14784280, "step": 22610 }, { "epoch": 11.85272536687631, "grad_norm": 0.11644015461206436, "learning_rate": 2.1295265695793836e-05, "loss": 0.435, "num_input_tokens_seen": 14788184, "step": 22615 }, { "epoch": 11.855345911949685, "grad_norm": 0.11740216612815857, "learning_rate": 2.128395804465387e-05, "loss": 0.5421, "num_input_tokens_seen": 14792472, "step": 22620 }, { "epoch": 11.85796645702306, "grad_norm": 0.1321968138217926, "learning_rate": 2.1272651170867487e-05, "loss": 0.517, "num_input_tokens_seen": 14795416, "step": 22625 }, { "epoch": 11.860587002096436, "grad_norm": 0.1303478181362152, "learning_rate": 2.1261345076799955e-05, "loss": 0.4871, "num_input_tokens_seen": 14799864, "step": 22630 }, { "epoch": 11.86320754716981, "grad_norm": 0.1970852166414261, "learning_rate": 2.125003976481638e-05, "loss": 0.4703, "num_input_tokens_seen": 14803384, "step": 22635 }, { "epoch": 11.865828092243186, "grad_norm": 0.24438489973545074, "learning_rate": 2.123873523728171e-05, "loss": 0.464, "num_input_tokens_seen": 14807224, "step": 22640 }, { "epoch": 11.868448637316561, "grad_norm": 0.14184385538101196, "learning_rate": 2.122743149656072e-05, "loss": 0.4837, "num_input_tokens_seen": 14811096, "step": 22645 }, { "epoch": 11.871069182389936, "grad_norm": 0.14216043055057526, "learning_rate": 2.1216128545018012e-05, "loss": 0.5532, "num_input_tokens_seen": 14814776, "step": 22650 }, { "epoch": 11.873689727463312, "grad_norm": 0.1435367465019226, "learning_rate": 2.1204826385018052e-05, "loss": 0.5541, "num_input_tokens_seen": 14818424, "step": 22655 }, { "epoch": 11.876310272536688, "grad_norm": 0.15161745250225067, "learning_rate": 2.1193525018925106e-05, "loss": 0.5473, "num_input_tokens_seen": 14822456, "step": 22660 }, { "epoch": 11.878930817610064, "grad_norm": 0.2770754098892212, "learning_rate": 2.1182224449103304e-05, "loss": 0.4755, "num_input_tokens_seen": 14824760, "step": 22665 }, { "epoch": 11.881551362683439, "grad_norm": 0.1599923074245453, "learning_rate": 2.117092467791659e-05, "loss": 0.3719, "num_input_tokens_seen": 14826968, "step": 22670 }, { "epoch": 11.884171907756814, "grad_norm": 0.2173297256231308, "learning_rate": 2.115962570772876e-05, "loss": 0.5295, "num_input_tokens_seen": 14830328, "step": 22675 }, { "epoch": 11.88679245283019, "grad_norm": 0.15946850180625916, "learning_rate": 2.1148327540903408e-05, "loss": 0.5292, "num_input_tokens_seen": 14833944, "step": 22680 }, { "epoch": 11.889412997903564, "grad_norm": 0.1745731234550476, "learning_rate": 2.113703017980399e-05, "loss": 0.536, "num_input_tokens_seen": 14840888, "step": 22685 }, { "epoch": 11.89203354297694, "grad_norm": 0.1550067663192749, "learning_rate": 2.1125733626793793e-05, "loss": 0.5355, "num_input_tokens_seen": 14843832, "step": 22690 }, { "epoch": 11.894654088050315, "grad_norm": 0.14178259670734406, "learning_rate": 2.1114437884235915e-05, "loss": 0.422, "num_input_tokens_seen": 14847352, "step": 22695 }, { "epoch": 11.89727463312369, "grad_norm": 0.10400397330522537, "learning_rate": 2.1103142954493313e-05, "loss": 0.3784, "num_input_tokens_seen": 14850712, "step": 22700 }, { "epoch": 11.899895178197065, "grad_norm": 0.34914299845695496, "learning_rate": 2.1091848839928745e-05, "loss": 0.5492, "num_input_tokens_seen": 14853432, "step": 22705 }, { "epoch": 11.90251572327044, "grad_norm": 0.23888856172561646, "learning_rate": 2.1080555542904805e-05, "loss": 0.541, "num_input_tokens_seen": 14856280, "step": 22710 }, { "epoch": 11.905136268343815, "grad_norm": 0.18745189905166626, "learning_rate": 2.106926306578393e-05, "loss": 0.569, "num_input_tokens_seen": 14858840, "step": 22715 }, { "epoch": 11.90775681341719, "grad_norm": 0.2486686110496521, "learning_rate": 2.1057971410928383e-05, "loss": 0.5068, "num_input_tokens_seen": 14861880, "step": 22720 }, { "epoch": 11.910377358490566, "grad_norm": 0.15837280452251434, "learning_rate": 2.104668058070024e-05, "loss": 0.5416, "num_input_tokens_seen": 14865016, "step": 22725 }, { "epoch": 11.91299790356394, "grad_norm": 0.11100167781114578, "learning_rate": 2.1035390577461427e-05, "loss": 0.5879, "num_input_tokens_seen": 14868760, "step": 22730 }, { "epoch": 11.915618448637316, "grad_norm": 0.13695451617240906, "learning_rate": 2.102410140357367e-05, "loss": 0.3964, "num_input_tokens_seen": 14872152, "step": 22735 }, { "epoch": 11.918238993710691, "grad_norm": 0.26855045557022095, "learning_rate": 2.1012813061398534e-05, "loss": 0.5434, "num_input_tokens_seen": 14875224, "step": 22740 }, { "epoch": 11.920859538784066, "grad_norm": 0.15764975547790527, "learning_rate": 2.1001525553297422e-05, "loss": 0.4427, "num_input_tokens_seen": 14878552, "step": 22745 }, { "epoch": 11.923480083857442, "grad_norm": 0.1885772943496704, "learning_rate": 2.099023888163154e-05, "loss": 0.4844, "num_input_tokens_seen": 14881304, "step": 22750 }, { "epoch": 11.926100628930818, "grad_norm": 0.2119583636522293, "learning_rate": 2.0978953048761947e-05, "loss": 0.644, "num_input_tokens_seen": 14884696, "step": 22755 }, { "epoch": 11.928721174004194, "grad_norm": 0.11540127545595169, "learning_rate": 2.0967668057049507e-05, "loss": 0.4554, "num_input_tokens_seen": 14887640, "step": 22760 }, { "epoch": 11.931341719077569, "grad_norm": 0.1068144366145134, "learning_rate": 2.0956383908854895e-05, "loss": 0.5115, "num_input_tokens_seen": 14890840, "step": 22765 }, { "epoch": 11.933962264150944, "grad_norm": 0.1827639490365982, "learning_rate": 2.0945100606538644e-05, "loss": 0.49, "num_input_tokens_seen": 14894712, "step": 22770 }, { "epoch": 11.93658280922432, "grad_norm": 0.3236503303050995, "learning_rate": 2.0933818152461077e-05, "loss": 0.561, "num_input_tokens_seen": 14897624, "step": 22775 }, { "epoch": 11.939203354297694, "grad_norm": 0.1422993391752243, "learning_rate": 2.092253654898237e-05, "loss": 0.4241, "num_input_tokens_seen": 14900632, "step": 22780 }, { "epoch": 11.94182389937107, "grad_norm": 0.1242656335234642, "learning_rate": 2.0911255798462494e-05, "loss": 0.471, "num_input_tokens_seen": 14903992, "step": 22785 }, { "epoch": 11.944444444444445, "grad_norm": 0.15586692094802856, "learning_rate": 2.0899975903261266e-05, "loss": 0.3916, "num_input_tokens_seen": 14907032, "step": 22790 }, { "epoch": 11.94706498951782, "grad_norm": 0.10999192297458649, "learning_rate": 2.0888696865738295e-05, "loss": 0.5071, "num_input_tokens_seen": 14909656, "step": 22795 }, { "epoch": 11.949685534591195, "grad_norm": 0.13971349596977234, "learning_rate": 2.087741868825304e-05, "loss": 0.2671, "num_input_tokens_seen": 14915608, "step": 22800 }, { "epoch": 11.95230607966457, "grad_norm": 0.2653124928474426, "learning_rate": 2.0866141373164758e-05, "loss": 0.6512, "num_input_tokens_seen": 14919256, "step": 22805 }, { "epoch": 11.954926624737945, "grad_norm": 0.23390422761440277, "learning_rate": 2.0854864922832544e-05, "loss": 0.7357, "num_input_tokens_seen": 14922520, "step": 22810 }, { "epoch": 11.95754716981132, "grad_norm": 0.38103610277175903, "learning_rate": 2.0843589339615304e-05, "loss": 0.4724, "num_input_tokens_seen": 14925336, "step": 22815 }, { "epoch": 11.960167714884696, "grad_norm": 0.10880255699157715, "learning_rate": 2.0832314625871742e-05, "loss": 0.459, "num_input_tokens_seen": 14928056, "step": 22820 }, { "epoch": 11.96278825995807, "grad_norm": 0.22605401277542114, "learning_rate": 2.0821040783960423e-05, "loss": 0.4931, "num_input_tokens_seen": 14931992, "step": 22825 }, { "epoch": 11.965408805031446, "grad_norm": 0.15405774116516113, "learning_rate": 2.0809767816239686e-05, "loss": 0.5377, "num_input_tokens_seen": 14935800, "step": 22830 }, { "epoch": 11.968029350104821, "grad_norm": 0.19240951538085938, "learning_rate": 2.0798495725067725e-05, "loss": 0.5269, "num_input_tokens_seen": 14938968, "step": 22835 }, { "epoch": 11.970649895178196, "grad_norm": 0.12250423431396484, "learning_rate": 2.078722451280252e-05, "loss": 0.4404, "num_input_tokens_seen": 14941400, "step": 22840 }, { "epoch": 11.973270440251572, "grad_norm": 0.15743716061115265, "learning_rate": 2.0775954181801893e-05, "loss": 0.4858, "num_input_tokens_seen": 14945080, "step": 22845 }, { "epoch": 11.975890985324948, "grad_norm": 0.2834348976612091, "learning_rate": 2.0764684734423455e-05, "loss": 0.4241, "num_input_tokens_seen": 14947448, "step": 22850 }, { "epoch": 11.978511530398324, "grad_norm": 0.14025752246379852, "learning_rate": 2.075341617302465e-05, "loss": 0.4828, "num_input_tokens_seen": 14950712, "step": 22855 }, { "epoch": 11.981132075471699, "grad_norm": 0.1854822039604187, "learning_rate": 2.074214849996274e-05, "loss": 0.4411, "num_input_tokens_seen": 14954072, "step": 22860 }, { "epoch": 11.983752620545074, "grad_norm": 0.12544915080070496, "learning_rate": 2.0730881717594776e-05, "loss": 0.4408, "num_input_tokens_seen": 14957592, "step": 22865 }, { "epoch": 11.98637316561845, "grad_norm": 0.13636991381645203, "learning_rate": 2.0719615828277657e-05, "loss": 0.4672, "num_input_tokens_seen": 14961080, "step": 22870 }, { "epoch": 11.988993710691824, "grad_norm": 0.3432421088218689, "learning_rate": 2.0708350834368083e-05, "loss": 0.5847, "num_input_tokens_seen": 14964248, "step": 22875 }, { "epoch": 11.9916142557652, "grad_norm": 0.14925675094127655, "learning_rate": 2.069708673822255e-05, "loss": 0.4978, "num_input_tokens_seen": 14967800, "step": 22880 }, { "epoch": 11.994234800838575, "grad_norm": 0.196987122297287, "learning_rate": 2.068582354219737e-05, "loss": 0.3994, "num_input_tokens_seen": 14971448, "step": 22885 }, { "epoch": 11.99685534591195, "grad_norm": 0.288332462310791, "learning_rate": 2.0674561248648697e-05, "loss": 0.5483, "num_input_tokens_seen": 14974840, "step": 22890 }, { "epoch": 11.999475890985325, "grad_norm": 0.1610766500234604, "learning_rate": 2.0663299859932456e-05, "loss": 0.4708, "num_input_tokens_seen": 14977912, "step": 22895 }, { "epoch": 12.0, "eval_loss": 0.4627397656440735, "eval_runtime": 16.0661, "eval_samples_per_second": 52.782, "eval_steps_per_second": 13.195, "num_input_tokens_seen": 14977976, "step": 22896 }, { "epoch": 12.0020964360587, "grad_norm": 0.30444246530532837, "learning_rate": 2.065203937840442e-05, "loss": 0.4373, "num_input_tokens_seen": 14979736, "step": 22900 }, { "epoch": 12.004716981132075, "grad_norm": 0.126534104347229, "learning_rate": 2.0640779806420136e-05, "loss": 0.4868, "num_input_tokens_seen": 14983864, "step": 22905 }, { "epoch": 12.00733752620545, "grad_norm": 0.14248061180114746, "learning_rate": 2.0629521146334978e-05, "loss": 0.5163, "num_input_tokens_seen": 14988152, "step": 22910 }, { "epoch": 12.009958071278826, "grad_norm": 0.13121768832206726, "learning_rate": 2.061826340050414e-05, "loss": 0.4713, "num_input_tokens_seen": 14991096, "step": 22915 }, { "epoch": 12.0125786163522, "grad_norm": 0.2134942263364792, "learning_rate": 2.0607006571282605e-05, "loss": 0.5662, "num_input_tokens_seen": 14993816, "step": 22920 }, { "epoch": 12.015199161425576, "grad_norm": 0.2586725652217865, "learning_rate": 2.0595750661025183e-05, "loss": 0.3925, "num_input_tokens_seen": 14996824, "step": 22925 }, { "epoch": 12.017819706498951, "grad_norm": 0.1981550008058548, "learning_rate": 2.0584495672086483e-05, "loss": 0.391, "num_input_tokens_seen": 14999544, "step": 22930 }, { "epoch": 12.020440251572326, "grad_norm": 0.15034885704517365, "learning_rate": 2.05732416068209e-05, "loss": 0.5622, "num_input_tokens_seen": 15002840, "step": 22935 }, { "epoch": 12.023060796645701, "grad_norm": 0.24613647162914276, "learning_rate": 2.0561988467582676e-05, "loss": 0.4909, "num_input_tokens_seen": 15005080, "step": 22940 }, { "epoch": 12.025681341719078, "grad_norm": 0.13768193125724792, "learning_rate": 2.0550736256725828e-05, "loss": 0.5306, "num_input_tokens_seen": 15008312, "step": 22945 }, { "epoch": 12.028301886792454, "grad_norm": 0.1816423237323761, "learning_rate": 2.0539484976604196e-05, "loss": 0.5595, "num_input_tokens_seen": 15011608, "step": 22950 }, { "epoch": 12.030922431865829, "grad_norm": 0.16996613144874573, "learning_rate": 2.052823462957142e-05, "loss": 0.478, "num_input_tokens_seen": 15014488, "step": 22955 }, { "epoch": 12.033542976939204, "grad_norm": 0.21101821959018707, "learning_rate": 2.0516985217980953e-05, "loss": 0.4582, "num_input_tokens_seen": 15017912, "step": 22960 }, { "epoch": 12.036163522012579, "grad_norm": 0.15061646699905396, "learning_rate": 2.050573674418602e-05, "loss": 0.504, "num_input_tokens_seen": 15020728, "step": 22965 }, { "epoch": 12.038784067085954, "grad_norm": 0.11441881954669952, "learning_rate": 2.0494489210539693e-05, "loss": 0.4472, "num_input_tokens_seen": 15024408, "step": 22970 }, { "epoch": 12.04140461215933, "grad_norm": 0.21091726422309875, "learning_rate": 2.0483242619394817e-05, "loss": 0.421, "num_input_tokens_seen": 15027928, "step": 22975 }, { "epoch": 12.044025157232705, "grad_norm": 0.14374950528144836, "learning_rate": 2.0471996973104063e-05, "loss": 0.4471, "num_input_tokens_seen": 15031480, "step": 22980 }, { "epoch": 12.04664570230608, "grad_norm": 0.1443459391593933, "learning_rate": 2.0460752274019886e-05, "loss": 0.5719, "num_input_tokens_seen": 15034040, "step": 22985 }, { "epoch": 12.049266247379455, "grad_norm": 0.11889398843050003, "learning_rate": 2.0449508524494537e-05, "loss": 0.3496, "num_input_tokens_seen": 15037464, "step": 22990 }, { "epoch": 12.05188679245283, "grad_norm": 0.16902917623519897, "learning_rate": 2.0438265726880092e-05, "loss": 0.4816, "num_input_tokens_seen": 15040728, "step": 22995 }, { "epoch": 12.054507337526205, "grad_norm": 0.13408373296260834, "learning_rate": 2.042702388352841e-05, "loss": 0.3922, "num_input_tokens_seen": 15044632, "step": 23000 }, { "epoch": 12.05712788259958, "grad_norm": 0.1323002129793167, "learning_rate": 2.041578299679117e-05, "loss": 0.4549, "num_input_tokens_seen": 15047352, "step": 23005 }, { "epoch": 12.059748427672956, "grad_norm": 0.15403670072555542, "learning_rate": 2.0404543069019818e-05, "loss": 0.4216, "num_input_tokens_seen": 15050520, "step": 23010 }, { "epoch": 12.06236897274633, "grad_norm": 0.14817678928375244, "learning_rate": 2.0393304102565642e-05, "loss": 0.4582, "num_input_tokens_seen": 15053208, "step": 23015 }, { "epoch": 12.064989517819706, "grad_norm": 0.2169240266084671, "learning_rate": 2.0382066099779685e-05, "loss": 0.5032, "num_input_tokens_seen": 15055736, "step": 23020 }, { "epoch": 12.067610062893081, "grad_norm": 0.17205709218978882, "learning_rate": 2.0370829063012808e-05, "loss": 0.3814, "num_input_tokens_seen": 15058648, "step": 23025 }, { "epoch": 12.070230607966456, "grad_norm": 0.26754969358444214, "learning_rate": 2.035959299461568e-05, "loss": 0.5066, "num_input_tokens_seen": 15061208, "step": 23030 }, { "epoch": 12.072851153039831, "grad_norm": 0.17158474028110504, "learning_rate": 2.0348357896938762e-05, "loss": 0.3817, "num_input_tokens_seen": 15065208, "step": 23035 }, { "epoch": 12.075471698113208, "grad_norm": 0.18276634812355042, "learning_rate": 2.03371237723323e-05, "loss": 0.5041, "num_input_tokens_seen": 15067928, "step": 23040 }, { "epoch": 12.078092243186584, "grad_norm": 0.09203033894300461, "learning_rate": 2.032589062314636e-05, "loss": 0.4636, "num_input_tokens_seen": 15071192, "step": 23045 }, { "epoch": 12.080712788259959, "grad_norm": 0.19379912316799164, "learning_rate": 2.0314658451730767e-05, "loss": 0.4787, "num_input_tokens_seen": 15074456, "step": 23050 }, { "epoch": 12.083333333333334, "grad_norm": 0.2008797824382782, "learning_rate": 2.030342726043517e-05, "loss": 0.6194, "num_input_tokens_seen": 15078232, "step": 23055 }, { "epoch": 12.085953878406709, "grad_norm": 0.1447623372077942, "learning_rate": 2.0292197051609013e-05, "loss": 0.488, "num_input_tokens_seen": 15081336, "step": 23060 }, { "epoch": 12.088574423480084, "grad_norm": 0.5159141421318054, "learning_rate": 2.0280967827601518e-05, "loss": 0.6007, "num_input_tokens_seen": 15085048, "step": 23065 }, { "epoch": 12.09119496855346, "grad_norm": 0.09871357679367065, "learning_rate": 2.0269739590761723e-05, "loss": 0.4988, "num_input_tokens_seen": 15092280, "step": 23070 }, { "epoch": 12.093815513626835, "grad_norm": 0.09457879513502121, "learning_rate": 2.0258512343438442e-05, "loss": 0.4331, "num_input_tokens_seen": 15096568, "step": 23075 }, { "epoch": 12.09643605870021, "grad_norm": 0.11432673782110214, "learning_rate": 2.0247286087980276e-05, "loss": 0.5213, "num_input_tokens_seen": 15099224, "step": 23080 }, { "epoch": 12.099056603773585, "grad_norm": 0.1324041485786438, "learning_rate": 2.0236060826735644e-05, "loss": 0.5008, "num_input_tokens_seen": 15103096, "step": 23085 }, { "epoch": 12.10167714884696, "grad_norm": 0.14108215272426605, "learning_rate": 2.0224836562052732e-05, "loss": 0.4777, "num_input_tokens_seen": 15105656, "step": 23090 }, { "epoch": 12.104297693920335, "grad_norm": 0.13782162964344025, "learning_rate": 2.0213613296279533e-05, "loss": 0.4146, "num_input_tokens_seen": 15108696, "step": 23095 }, { "epoch": 12.10691823899371, "grad_norm": 0.17984212934970856, "learning_rate": 2.0202391031763844e-05, "loss": 0.5264, "num_input_tokens_seen": 15111576, "step": 23100 }, { "epoch": 12.109538784067086, "grad_norm": 0.17420747876167297, "learning_rate": 2.0191169770853195e-05, "loss": 0.4397, "num_input_tokens_seen": 15115192, "step": 23105 }, { "epoch": 12.11215932914046, "grad_norm": 0.13067084550857544, "learning_rate": 2.0179949515894975e-05, "loss": 0.5396, "num_input_tokens_seen": 15119160, "step": 23110 }, { "epoch": 12.114779874213836, "grad_norm": 0.28067708015441895, "learning_rate": 2.016873026923633e-05, "loss": 0.3819, "num_input_tokens_seen": 15122136, "step": 23115 }, { "epoch": 12.117400419287211, "grad_norm": 0.19414004683494568, "learning_rate": 2.0157512033224187e-05, "loss": 0.4068, "num_input_tokens_seen": 15126328, "step": 23120 }, { "epoch": 12.120020964360586, "grad_norm": 0.1811971515417099, "learning_rate": 2.0146294810205287e-05, "loss": 0.3326, "num_input_tokens_seen": 15129304, "step": 23125 }, { "epoch": 12.122641509433961, "grad_norm": 0.11003687232732773, "learning_rate": 2.0135078602526146e-05, "loss": 0.4948, "num_input_tokens_seen": 15131768, "step": 23130 }, { "epoch": 12.125262054507338, "grad_norm": 0.1958695501089096, "learning_rate": 2.0123863412533046e-05, "loss": 0.4558, "num_input_tokens_seen": 15135128, "step": 23135 }, { "epoch": 12.127882599580714, "grad_norm": 0.18693731725215912, "learning_rate": 2.0112649242572095e-05, "loss": 0.4482, "num_input_tokens_seen": 15137752, "step": 23140 }, { "epoch": 12.130503144654089, "grad_norm": 0.16673485934734344, "learning_rate": 2.0101436094989164e-05, "loss": 0.353, "num_input_tokens_seen": 15141912, "step": 23145 }, { "epoch": 12.133123689727464, "grad_norm": 0.14607416093349457, "learning_rate": 2.0090223972129918e-05, "loss": 0.4114, "num_input_tokens_seen": 15144792, "step": 23150 }, { "epoch": 12.135744234800839, "grad_norm": 0.12014421820640564, "learning_rate": 2.0079012876339797e-05, "loss": 0.4903, "num_input_tokens_seen": 15149624, "step": 23155 }, { "epoch": 12.138364779874214, "grad_norm": 0.14654123783111572, "learning_rate": 2.0067802809964055e-05, "loss": 0.3954, "num_input_tokens_seen": 15152856, "step": 23160 }, { "epoch": 12.14098532494759, "grad_norm": 0.40070319175720215, "learning_rate": 2.0056593775347686e-05, "loss": 0.5924, "num_input_tokens_seen": 15155256, "step": 23165 }, { "epoch": 12.143605870020965, "grad_norm": 0.1424485445022583, "learning_rate": 2.00453857748355e-05, "loss": 0.5078, "num_input_tokens_seen": 15157976, "step": 23170 }, { "epoch": 12.14622641509434, "grad_norm": 0.14749892055988312, "learning_rate": 2.0034178810772082e-05, "loss": 0.6809, "num_input_tokens_seen": 15160344, "step": 23175 }, { "epoch": 12.148846960167715, "grad_norm": 0.15284407138824463, "learning_rate": 2.002297288550181e-05, "loss": 0.6111, "num_input_tokens_seen": 15163032, "step": 23180 }, { "epoch": 12.15146750524109, "grad_norm": 0.12897564470767975, "learning_rate": 2.0011768001368835e-05, "loss": 0.537, "num_input_tokens_seen": 15165784, "step": 23185 }, { "epoch": 12.154088050314465, "grad_norm": 0.1639043390750885, "learning_rate": 2.0000564160717076e-05, "loss": 0.5706, "num_input_tokens_seen": 15168664, "step": 23190 }, { "epoch": 12.15670859538784, "grad_norm": 0.13095036149024963, "learning_rate": 1.9989361365890263e-05, "loss": 0.5277, "num_input_tokens_seen": 15171576, "step": 23195 }, { "epoch": 12.159329140461216, "grad_norm": 0.12223461270332336, "learning_rate": 1.997815961923188e-05, "loss": 0.4077, "num_input_tokens_seen": 15175160, "step": 23200 }, { "epoch": 12.16194968553459, "grad_norm": 0.16124366223812103, "learning_rate": 1.996695892308522e-05, "loss": 0.5549, "num_input_tokens_seen": 15178744, "step": 23205 }, { "epoch": 12.164570230607966, "grad_norm": 0.1875586211681366, "learning_rate": 1.9955759279793327e-05, "loss": 0.4968, "num_input_tokens_seen": 15182264, "step": 23210 }, { "epoch": 12.167190775681341, "grad_norm": 0.22537177801132202, "learning_rate": 1.9944560691699057e-05, "loss": 0.5445, "num_input_tokens_seen": 15184984, "step": 23215 }, { "epoch": 12.169811320754716, "grad_norm": 0.23973248898983002, "learning_rate": 1.9933363161145006e-05, "loss": 0.531, "num_input_tokens_seen": 15187448, "step": 23220 }, { "epoch": 12.172431865828091, "grad_norm": 0.14689166843891144, "learning_rate": 1.9922166690473575e-05, "loss": 0.4844, "num_input_tokens_seen": 15190872, "step": 23225 }, { "epoch": 12.175052410901468, "grad_norm": 0.20885561406612396, "learning_rate": 1.991097128202695e-05, "loss": 0.5207, "num_input_tokens_seen": 15193400, "step": 23230 }, { "epoch": 12.177672955974844, "grad_norm": 0.17352545261383057, "learning_rate": 1.9899776938147065e-05, "loss": 0.4564, "num_input_tokens_seen": 15197336, "step": 23235 }, { "epoch": 12.180293501048219, "grad_norm": 0.23907381296157837, "learning_rate": 1.9888583661175665e-05, "loss": 0.5859, "num_input_tokens_seen": 15200504, "step": 23240 }, { "epoch": 12.182914046121594, "grad_norm": 0.15712232887744904, "learning_rate": 1.9877391453454253e-05, "loss": 0.5768, "num_input_tokens_seen": 15203640, "step": 23245 }, { "epoch": 12.185534591194969, "grad_norm": 0.15107563138008118, "learning_rate": 1.98662003173241e-05, "loss": 0.4666, "num_input_tokens_seen": 15206744, "step": 23250 }, { "epoch": 12.188155136268344, "grad_norm": 0.140220046043396, "learning_rate": 1.9855010255126266e-05, "loss": 0.382, "num_input_tokens_seen": 15211064, "step": 23255 }, { "epoch": 12.19077568134172, "grad_norm": 0.162740558385849, "learning_rate": 1.9843821269201598e-05, "loss": 0.5618, "num_input_tokens_seen": 15214040, "step": 23260 }, { "epoch": 12.193396226415095, "grad_norm": 0.09745647013187408, "learning_rate": 1.983263336189069e-05, "loss": 0.6515, "num_input_tokens_seen": 15216440, "step": 23265 }, { "epoch": 12.19601677148847, "grad_norm": 0.22260518372058868, "learning_rate": 1.9821446535533937e-05, "loss": 0.4884, "num_input_tokens_seen": 15219352, "step": 23270 }, { "epoch": 12.198637316561845, "grad_norm": 0.17029103636741638, "learning_rate": 1.9810260792471492e-05, "loss": 0.502, "num_input_tokens_seen": 15225432, "step": 23275 }, { "epoch": 12.20125786163522, "grad_norm": 0.10236583650112152, "learning_rate": 1.979907613504327e-05, "loss": 0.519, "num_input_tokens_seen": 15229528, "step": 23280 }, { "epoch": 12.203878406708595, "grad_norm": 0.211678609251976, "learning_rate": 1.978789256558899e-05, "loss": 0.5063, "num_input_tokens_seen": 15233208, "step": 23285 }, { "epoch": 12.20649895178197, "grad_norm": 0.1520906686782837, "learning_rate": 1.977671008644812e-05, "loss": 0.396, "num_input_tokens_seen": 15236504, "step": 23290 }, { "epoch": 12.209119496855346, "grad_norm": 0.11636386066675186, "learning_rate": 1.9765528699959912e-05, "loss": 0.4561, "num_input_tokens_seen": 15239768, "step": 23295 }, { "epoch": 12.21174004192872, "grad_norm": 0.23630346357822418, "learning_rate": 1.975434840846339e-05, "loss": 0.5857, "num_input_tokens_seen": 15243032, "step": 23300 }, { "epoch": 12.214360587002096, "grad_norm": 0.20770762860774994, "learning_rate": 1.9743169214297326e-05, "loss": 0.5129, "num_input_tokens_seen": 15245944, "step": 23305 }, { "epoch": 12.216981132075471, "grad_norm": 0.09781521558761597, "learning_rate": 1.973199111980029e-05, "loss": 0.4132, "num_input_tokens_seen": 15249880, "step": 23310 }, { "epoch": 12.219601677148846, "grad_norm": 0.2197549194097519, "learning_rate": 1.9720814127310608e-05, "loss": 0.4286, "num_input_tokens_seen": 15252984, "step": 23315 }, { "epoch": 12.222222222222221, "grad_norm": 0.2038782835006714, "learning_rate": 1.9709638239166386e-05, "loss": 0.4785, "num_input_tokens_seen": 15256024, "step": 23320 }, { "epoch": 12.224842767295598, "grad_norm": 0.14800967276096344, "learning_rate": 1.9698463457705485e-05, "loss": 0.3659, "num_input_tokens_seen": 15259608, "step": 23325 }, { "epoch": 12.227463312368974, "grad_norm": 0.37321674823760986, "learning_rate": 1.968728978526556e-05, "loss": 0.3729, "num_input_tokens_seen": 15262200, "step": 23330 }, { "epoch": 12.230083857442349, "grad_norm": 0.17628178000450134, "learning_rate": 1.9676117224183986e-05, "loss": 0.3894, "num_input_tokens_seen": 15265336, "step": 23335 }, { "epoch": 12.232704402515724, "grad_norm": 0.16798484325408936, "learning_rate": 1.9664945776797956e-05, "loss": 0.6017, "num_input_tokens_seen": 15268024, "step": 23340 }, { "epoch": 12.235324947589099, "grad_norm": 0.21527785062789917, "learning_rate": 1.96537754454444e-05, "loss": 0.5832, "num_input_tokens_seen": 15271480, "step": 23345 }, { "epoch": 12.237945492662474, "grad_norm": 0.15159614384174347, "learning_rate": 1.964260623246003e-05, "loss": 0.5298, "num_input_tokens_seen": 15274296, "step": 23350 }, { "epoch": 12.24056603773585, "grad_norm": 0.18082034587860107, "learning_rate": 1.9631438140181324e-05, "loss": 0.4379, "num_input_tokens_seen": 15276952, "step": 23355 }, { "epoch": 12.243186582809225, "grad_norm": 0.13593196868896484, "learning_rate": 1.9620271170944503e-05, "loss": 0.4208, "num_input_tokens_seen": 15280056, "step": 23360 }, { "epoch": 12.2458071278826, "grad_norm": 0.20169340074062347, "learning_rate": 1.960910532708558e-05, "loss": 0.5229, "num_input_tokens_seen": 15283256, "step": 23365 }, { "epoch": 12.248427672955975, "grad_norm": 0.26265832781791687, "learning_rate": 1.9597940610940315e-05, "loss": 0.4535, "num_input_tokens_seen": 15286808, "step": 23370 }, { "epoch": 12.25104821802935, "grad_norm": 0.13781286776065826, "learning_rate": 1.958677702484425e-05, "loss": 0.5364, "num_input_tokens_seen": 15291000, "step": 23375 }, { "epoch": 12.253668763102725, "grad_norm": 0.1332699954509735, "learning_rate": 1.957561457113267e-05, "loss": 0.5175, "num_input_tokens_seen": 15294744, "step": 23380 }, { "epoch": 12.2562893081761, "grad_norm": 0.28284692764282227, "learning_rate": 1.9564453252140648e-05, "loss": 0.4974, "num_input_tokens_seen": 15297752, "step": 23385 }, { "epoch": 12.258909853249476, "grad_norm": 0.12449435144662857, "learning_rate": 1.9553293070202984e-05, "loss": 0.4552, "num_input_tokens_seen": 15301208, "step": 23390 }, { "epoch": 12.26153039832285, "grad_norm": 0.11886993795633316, "learning_rate": 1.9542134027654267e-05, "loss": 0.5221, "num_input_tokens_seen": 15304984, "step": 23395 }, { "epoch": 12.264150943396226, "grad_norm": 0.14832979440689087, "learning_rate": 1.953097612682885e-05, "loss": 0.4337, "num_input_tokens_seen": 15307896, "step": 23400 }, { "epoch": 12.266771488469601, "grad_norm": 0.10678645968437195, "learning_rate": 1.9519819370060832e-05, "loss": 0.4458, "num_input_tokens_seen": 15311064, "step": 23405 }, { "epoch": 12.269392033542976, "grad_norm": 0.10242857038974762, "learning_rate": 1.9508663759684074e-05, "loss": 0.4549, "num_input_tokens_seen": 15314136, "step": 23410 }, { "epoch": 12.272012578616351, "grad_norm": 0.13573843240737915, "learning_rate": 1.9497509298032224e-05, "loss": 0.6611, "num_input_tokens_seen": 15317464, "step": 23415 }, { "epoch": 12.274633123689728, "grad_norm": 0.11820442229509354, "learning_rate": 1.9486355987438647e-05, "loss": 0.4686, "num_input_tokens_seen": 15320920, "step": 23420 }, { "epoch": 12.277253668763104, "grad_norm": 0.25859177112579346, "learning_rate": 1.947520383023649e-05, "loss": 0.3473, "num_input_tokens_seen": 15323800, "step": 23425 }, { "epoch": 12.279874213836479, "grad_norm": 0.10730060189962387, "learning_rate": 1.9464052828758666e-05, "loss": 0.5679, "num_input_tokens_seen": 15327512, "step": 23430 }, { "epoch": 12.282494758909854, "grad_norm": 0.18485669791698456, "learning_rate": 1.945290298533783e-05, "loss": 0.4784, "num_input_tokens_seen": 15331064, "step": 23435 }, { "epoch": 12.285115303983229, "grad_norm": 0.1529969722032547, "learning_rate": 1.9441754302306406e-05, "loss": 0.5189, "num_input_tokens_seen": 15334424, "step": 23440 }, { "epoch": 12.287735849056604, "grad_norm": 0.2092565894126892, "learning_rate": 1.9430606781996578e-05, "loss": 0.4537, "num_input_tokens_seen": 15337720, "step": 23445 }, { "epoch": 12.29035639412998, "grad_norm": 0.19184339046478271, "learning_rate": 1.941946042674026e-05, "loss": 0.5673, "num_input_tokens_seen": 15340216, "step": 23450 }, { "epoch": 12.292976939203355, "grad_norm": 0.19751344621181488, "learning_rate": 1.940831523886916e-05, "loss": 0.4172, "num_input_tokens_seen": 15343736, "step": 23455 }, { "epoch": 12.29559748427673, "grad_norm": 0.14844627678394318, "learning_rate": 1.9397171220714714e-05, "loss": 0.6111, "num_input_tokens_seen": 15347000, "step": 23460 }, { "epoch": 12.298218029350105, "grad_norm": 0.24565088748931885, "learning_rate": 1.9386028374608134e-05, "loss": 0.5489, "num_input_tokens_seen": 15349848, "step": 23465 }, { "epoch": 12.30083857442348, "grad_norm": 0.12237133085727692, "learning_rate": 1.9374886702880372e-05, "loss": 0.5153, "num_input_tokens_seen": 15353272, "step": 23470 }, { "epoch": 12.303459119496855, "grad_norm": 0.1638662964105606, "learning_rate": 1.936374620786213e-05, "loss": 0.3647, "num_input_tokens_seen": 15355960, "step": 23475 }, { "epoch": 12.30607966457023, "grad_norm": 0.19992685317993164, "learning_rate": 1.9352606891883885e-05, "loss": 0.5659, "num_input_tokens_seen": 15359800, "step": 23480 }, { "epoch": 12.308700209643606, "grad_norm": 0.14944390952587128, "learning_rate": 1.9341468757275844e-05, "loss": 0.3383, "num_input_tokens_seen": 15363640, "step": 23485 }, { "epoch": 12.31132075471698, "grad_norm": 0.09208056330680847, "learning_rate": 1.933033180636798e-05, "loss": 0.4403, "num_input_tokens_seen": 15367768, "step": 23490 }, { "epoch": 12.313941299790356, "grad_norm": 0.18386127054691315, "learning_rate": 1.9319196041490027e-05, "loss": 0.3672, "num_input_tokens_seen": 15370680, "step": 23495 }, { "epoch": 12.316561844863731, "grad_norm": 0.11438941955566406, "learning_rate": 1.930806146497146e-05, "loss": 0.4454, "num_input_tokens_seen": 15374008, "step": 23500 }, { "epoch": 12.319182389937106, "grad_norm": 0.18377825617790222, "learning_rate": 1.9296928079141485e-05, "loss": 0.5385, "num_input_tokens_seen": 15377976, "step": 23505 }, { "epoch": 12.321802935010481, "grad_norm": 0.14094635844230652, "learning_rate": 1.9285795886329093e-05, "loss": 0.5932, "num_input_tokens_seen": 15381144, "step": 23510 }, { "epoch": 12.324423480083858, "grad_norm": 0.20452436804771423, "learning_rate": 1.927466488886301e-05, "loss": 0.5302, "num_input_tokens_seen": 15385144, "step": 23515 }, { "epoch": 12.327044025157234, "grad_norm": 0.10934597998857498, "learning_rate": 1.926353508907172e-05, "loss": 0.5001, "num_input_tokens_seen": 15388504, "step": 23520 }, { "epoch": 12.329664570230609, "grad_norm": 0.4482428729534149, "learning_rate": 1.9252406489283437e-05, "loss": 0.5861, "num_input_tokens_seen": 15391448, "step": 23525 }, { "epoch": 12.332285115303984, "grad_norm": 0.19833751022815704, "learning_rate": 1.9241279091826155e-05, "loss": 0.4342, "num_input_tokens_seen": 15393816, "step": 23530 }, { "epoch": 12.334905660377359, "grad_norm": 0.15122805535793304, "learning_rate": 1.923015289902758e-05, "loss": 0.3951, "num_input_tokens_seen": 15396888, "step": 23535 }, { "epoch": 12.337526205450734, "grad_norm": 0.13522067666053772, "learning_rate": 1.9219027913215193e-05, "loss": 0.3926, "num_input_tokens_seen": 15399352, "step": 23540 }, { "epoch": 12.34014675052411, "grad_norm": 0.10597100853919983, "learning_rate": 1.9207904136716214e-05, "loss": 0.597, "num_input_tokens_seen": 15404152, "step": 23545 }, { "epoch": 12.342767295597485, "grad_norm": 0.1820746809244156, "learning_rate": 1.9196781571857608e-05, "loss": 0.4172, "num_input_tokens_seen": 15407000, "step": 23550 }, { "epoch": 12.34538784067086, "grad_norm": 0.15392529964447021, "learning_rate": 1.9185660220966097e-05, "loss": 0.4514, "num_input_tokens_seen": 15410360, "step": 23555 }, { "epoch": 12.348008385744235, "grad_norm": 0.1295483559370041, "learning_rate": 1.9174540086368135e-05, "loss": 0.4728, "num_input_tokens_seen": 15413560, "step": 23560 }, { "epoch": 12.35062893081761, "grad_norm": 0.11059284210205078, "learning_rate": 1.9163421170389918e-05, "loss": 0.457, "num_input_tokens_seen": 15417400, "step": 23565 }, { "epoch": 12.353249475890985, "grad_norm": 0.18023423850536346, "learning_rate": 1.9152303475357402e-05, "loss": 0.5184, "num_input_tokens_seen": 15420248, "step": 23570 }, { "epoch": 12.35587002096436, "grad_norm": 0.2910299599170685, "learning_rate": 1.9141187003596295e-05, "loss": 0.5596, "num_input_tokens_seen": 15423320, "step": 23575 }, { "epoch": 12.358490566037736, "grad_norm": 0.17047572135925293, "learning_rate": 1.913007175743202e-05, "loss": 0.4817, "num_input_tokens_seen": 15426840, "step": 23580 }, { "epoch": 12.36111111111111, "grad_norm": 0.11486426740884781, "learning_rate": 1.9118957739189773e-05, "loss": 0.4438, "num_input_tokens_seen": 15430488, "step": 23585 }, { "epoch": 12.363731656184486, "grad_norm": 0.18603703379631042, "learning_rate": 1.910784495119447e-05, "loss": 0.5964, "num_input_tokens_seen": 15433432, "step": 23590 }, { "epoch": 12.366352201257861, "grad_norm": 0.12247160822153091, "learning_rate": 1.9096733395770778e-05, "loss": 0.3781, "num_input_tokens_seen": 15437240, "step": 23595 }, { "epoch": 12.368972746331236, "grad_norm": 0.2393079698085785, "learning_rate": 1.9085623075243114e-05, "loss": 0.4454, "num_input_tokens_seen": 15440792, "step": 23600 }, { "epoch": 12.371593291404611, "grad_norm": 0.1417221873998642, "learning_rate": 1.9074513991935623e-05, "loss": 0.4381, "num_input_tokens_seen": 15443928, "step": 23605 }, { "epoch": 12.374213836477987, "grad_norm": 0.11147494614124298, "learning_rate": 1.906340614817221e-05, "loss": 0.5442, "num_input_tokens_seen": 15446936, "step": 23610 }, { "epoch": 12.376834381551364, "grad_norm": 0.22730520367622375, "learning_rate": 1.9052299546276506e-05, "loss": 0.4548, "num_input_tokens_seen": 15450168, "step": 23615 }, { "epoch": 12.379454926624739, "grad_norm": 0.14530499279499054, "learning_rate": 1.904119418857187e-05, "loss": 0.4971, "num_input_tokens_seen": 15453592, "step": 23620 }, { "epoch": 12.382075471698114, "grad_norm": 0.20291240513324738, "learning_rate": 1.903009007738144e-05, "loss": 0.558, "num_input_tokens_seen": 15460056, "step": 23625 }, { "epoch": 12.384696016771489, "grad_norm": 0.10771335661411285, "learning_rate": 1.9018987215028045e-05, "loss": 0.4478, "num_input_tokens_seen": 15463160, "step": 23630 }, { "epoch": 12.387316561844864, "grad_norm": 0.19873134791851044, "learning_rate": 1.9007885603834298e-05, "loss": 0.7424, "num_input_tokens_seen": 15466360, "step": 23635 }, { "epoch": 12.38993710691824, "grad_norm": 0.15873447060585022, "learning_rate": 1.8996785246122526e-05, "loss": 0.6555, "num_input_tokens_seen": 15469112, "step": 23640 }, { "epoch": 12.392557651991615, "grad_norm": 0.22629103064537048, "learning_rate": 1.898568614421478e-05, "loss": 0.4843, "num_input_tokens_seen": 15472088, "step": 23645 }, { "epoch": 12.39517819706499, "grad_norm": 0.10268712043762207, "learning_rate": 1.897458830043288e-05, "loss": 0.3798, "num_input_tokens_seen": 15475384, "step": 23650 }, { "epoch": 12.397798742138365, "grad_norm": 0.1430930346250534, "learning_rate": 1.8963491717098372e-05, "loss": 0.4676, "num_input_tokens_seen": 15477528, "step": 23655 }, { "epoch": 12.40041928721174, "grad_norm": 0.24858510494232178, "learning_rate": 1.8952396396532522e-05, "loss": 0.4692, "num_input_tokens_seen": 15479768, "step": 23660 }, { "epoch": 12.403039832285115, "grad_norm": 0.1535385549068451, "learning_rate": 1.8941302341056356e-05, "loss": 0.8727, "num_input_tokens_seen": 15482264, "step": 23665 }, { "epoch": 12.40566037735849, "grad_norm": 0.14376798272132874, "learning_rate": 1.8930209552990625e-05, "loss": 0.5245, "num_input_tokens_seen": 15485144, "step": 23670 }, { "epoch": 12.408280922431866, "grad_norm": 0.15956105291843414, "learning_rate": 1.8919118034655796e-05, "loss": 0.367, "num_input_tokens_seen": 15487480, "step": 23675 }, { "epoch": 12.41090146750524, "grad_norm": 0.12206342071294785, "learning_rate": 1.8908027788372107e-05, "loss": 0.5132, "num_input_tokens_seen": 15491576, "step": 23680 }, { "epoch": 12.413522012578616, "grad_norm": 0.21210278570652008, "learning_rate": 1.88969388164595e-05, "loss": 0.5669, "num_input_tokens_seen": 15494328, "step": 23685 }, { "epoch": 12.416142557651991, "grad_norm": 0.12989971041679382, "learning_rate": 1.8885851121237666e-05, "loss": 0.6187, "num_input_tokens_seen": 15498360, "step": 23690 }, { "epoch": 12.418763102725366, "grad_norm": 0.17313653230667114, "learning_rate": 1.8874764705026023e-05, "loss": 0.4189, "num_input_tokens_seen": 15501176, "step": 23695 }, { "epoch": 12.421383647798741, "grad_norm": 0.18390502035617828, "learning_rate": 1.8863679570143728e-05, "loss": 0.4378, "num_input_tokens_seen": 15504440, "step": 23700 }, { "epoch": 12.424004192872117, "grad_norm": 0.17085404694080353, "learning_rate": 1.885259571890966e-05, "loss": 0.4963, "num_input_tokens_seen": 15507512, "step": 23705 }, { "epoch": 12.426624737945493, "grad_norm": 0.21457162499427795, "learning_rate": 1.8841513153642427e-05, "loss": 0.5352, "num_input_tokens_seen": 15510840, "step": 23710 }, { "epoch": 12.429245283018869, "grad_norm": 0.24714960157871246, "learning_rate": 1.8830431876660388e-05, "loss": 0.5424, "num_input_tokens_seen": 15513656, "step": 23715 }, { "epoch": 12.431865828092244, "grad_norm": 0.15173617005348206, "learning_rate": 1.8819351890281607e-05, "loss": 0.4139, "num_input_tokens_seen": 15516952, "step": 23720 }, { "epoch": 12.434486373165619, "grad_norm": 0.21528734266757965, "learning_rate": 1.880827319682391e-05, "loss": 0.4645, "num_input_tokens_seen": 15519832, "step": 23725 }, { "epoch": 12.437106918238994, "grad_norm": 0.14384758472442627, "learning_rate": 1.879719579860481e-05, "loss": 0.5337, "num_input_tokens_seen": 15523480, "step": 23730 }, { "epoch": 12.43972746331237, "grad_norm": 0.30124256014823914, "learning_rate": 1.878611969794159e-05, "loss": 0.4949, "num_input_tokens_seen": 15526264, "step": 23735 }, { "epoch": 12.442348008385745, "grad_norm": 0.21616385877132416, "learning_rate": 1.8775044897151232e-05, "loss": 0.5644, "num_input_tokens_seen": 15529848, "step": 23740 }, { "epoch": 12.44496855345912, "grad_norm": 0.11920254677534103, "learning_rate": 1.876397139855047e-05, "loss": 0.3415, "num_input_tokens_seen": 15533368, "step": 23745 }, { "epoch": 12.447589098532495, "grad_norm": 0.09004934132099152, "learning_rate": 1.875289920445574e-05, "loss": 0.4099, "num_input_tokens_seen": 15537080, "step": 23750 }, { "epoch": 12.45020964360587, "grad_norm": 0.19531603157520294, "learning_rate": 1.8741828317183234e-05, "loss": 0.4007, "num_input_tokens_seen": 15540248, "step": 23755 }, { "epoch": 12.452830188679245, "grad_norm": 0.14179588854312897, "learning_rate": 1.873075873904885e-05, "loss": 0.5235, "num_input_tokens_seen": 15543064, "step": 23760 }, { "epoch": 12.45545073375262, "grad_norm": 0.1432265043258667, "learning_rate": 1.87196904723682e-05, "loss": 0.3669, "num_input_tokens_seen": 15546968, "step": 23765 }, { "epoch": 12.458071278825996, "grad_norm": 0.15717022120952606, "learning_rate": 1.8708623519456665e-05, "loss": 0.7497, "num_input_tokens_seen": 15550040, "step": 23770 }, { "epoch": 12.46069182389937, "grad_norm": 0.30891159176826477, "learning_rate": 1.8697557882629308e-05, "loss": 0.4287, "num_input_tokens_seen": 15552824, "step": 23775 }, { "epoch": 12.463312368972746, "grad_norm": 0.2624043822288513, "learning_rate": 1.868649356420094e-05, "loss": 0.4449, "num_input_tokens_seen": 15555448, "step": 23780 }, { "epoch": 12.465932914046121, "grad_norm": 0.15294304490089417, "learning_rate": 1.86754305664861e-05, "loss": 0.5035, "num_input_tokens_seen": 15559096, "step": 23785 }, { "epoch": 12.468553459119496, "grad_norm": 0.19350579380989075, "learning_rate": 1.866436889179902e-05, "loss": 0.685, "num_input_tokens_seen": 15562136, "step": 23790 }, { "epoch": 12.471174004192871, "grad_norm": 0.21672210097312927, "learning_rate": 1.865330854245369e-05, "loss": 0.3975, "num_input_tokens_seen": 15565528, "step": 23795 }, { "epoch": 12.473794549266247, "grad_norm": 0.24227021634578705, "learning_rate": 1.8642249520763804e-05, "loss": 0.6336, "num_input_tokens_seen": 15569048, "step": 23800 }, { "epoch": 12.476415094339623, "grad_norm": 0.126606285572052, "learning_rate": 1.863119182904278e-05, "loss": 0.4815, "num_input_tokens_seen": 15572088, "step": 23805 }, { "epoch": 12.479035639412999, "grad_norm": 0.20149372518062592, "learning_rate": 1.8620135469603777e-05, "loss": 0.4379, "num_input_tokens_seen": 15574488, "step": 23810 }, { "epoch": 12.481656184486374, "grad_norm": 0.17891666293144226, "learning_rate": 1.8609080444759653e-05, "loss": 0.3954, "num_input_tokens_seen": 15577176, "step": 23815 }, { "epoch": 12.484276729559749, "grad_norm": 0.18859067559242249, "learning_rate": 1.8598026756822978e-05, "loss": 0.4523, "num_input_tokens_seen": 15582168, "step": 23820 }, { "epoch": 12.486897274633124, "grad_norm": 0.13554716110229492, "learning_rate": 1.8586974408106074e-05, "loss": 0.5311, "num_input_tokens_seen": 15585880, "step": 23825 }, { "epoch": 12.4895178197065, "grad_norm": 0.18636630475521088, "learning_rate": 1.8575923400920954e-05, "loss": 0.5537, "num_input_tokens_seen": 15588920, "step": 23830 }, { "epoch": 12.492138364779874, "grad_norm": 0.10463033616542816, "learning_rate": 1.856487373757937e-05, "loss": 0.477, "num_input_tokens_seen": 15592312, "step": 23835 }, { "epoch": 12.49475890985325, "grad_norm": 0.17962782084941864, "learning_rate": 1.8553825420392792e-05, "loss": 0.5511, "num_input_tokens_seen": 15595512, "step": 23840 }, { "epoch": 12.497379454926625, "grad_norm": 0.19480551779270172, "learning_rate": 1.8542778451672385e-05, "loss": 0.4446, "num_input_tokens_seen": 15598040, "step": 23845 }, { "epoch": 12.5, "grad_norm": 0.18933051824569702, "learning_rate": 1.8531732833729063e-05, "loss": 0.3622, "num_input_tokens_seen": 15600344, "step": 23850 }, { "epoch": 12.502620545073375, "grad_norm": 0.15875768661499023, "learning_rate": 1.852068856887343e-05, "loss": 0.5762, "num_input_tokens_seen": 15603960, "step": 23855 }, { "epoch": 12.50524109014675, "grad_norm": 0.1409149318933487, "learning_rate": 1.8509645659415837e-05, "loss": 0.411, "num_input_tokens_seen": 15606328, "step": 23860 }, { "epoch": 12.507861635220126, "grad_norm": 0.17418928444385529, "learning_rate": 1.849860410766632e-05, "loss": 0.42, "num_input_tokens_seen": 15608760, "step": 23865 }, { "epoch": 12.5104821802935, "grad_norm": 0.1761978715658188, "learning_rate": 1.8487563915934664e-05, "loss": 0.4828, "num_input_tokens_seen": 15612088, "step": 23870 }, { "epoch": 12.513102725366876, "grad_norm": 0.2584642469882965, "learning_rate": 1.847652508653033e-05, "loss": 0.5663, "num_input_tokens_seen": 15615640, "step": 23875 }, { "epoch": 12.515723270440251, "grad_norm": 0.16892394423484802, "learning_rate": 1.846548762176252e-05, "loss": 0.3761, "num_input_tokens_seen": 15619000, "step": 23880 }, { "epoch": 12.518343815513626, "grad_norm": 0.25644221901893616, "learning_rate": 1.845445152394015e-05, "loss": 0.5494, "num_input_tokens_seen": 15621720, "step": 23885 }, { "epoch": 12.520964360587001, "grad_norm": 0.14744745194911957, "learning_rate": 1.8443416795371853e-05, "loss": 0.3792, "num_input_tokens_seen": 15624920, "step": 23890 }, { "epoch": 12.523584905660378, "grad_norm": 0.12979377806186676, "learning_rate": 1.8432383438365957e-05, "loss": 0.4976, "num_input_tokens_seen": 15628056, "step": 23895 }, { "epoch": 12.526205450733752, "grad_norm": 0.20150364935398102, "learning_rate": 1.8421351455230528e-05, "loss": 0.4643, "num_input_tokens_seen": 15630936, "step": 23900 }, { "epoch": 12.528825995807129, "grad_norm": 0.18537676334381104, "learning_rate": 1.8410320848273315e-05, "loss": 0.475, "num_input_tokens_seen": 15633816, "step": 23905 }, { "epoch": 12.531446540880504, "grad_norm": 0.16038674116134644, "learning_rate": 1.8399291619801794e-05, "loss": 0.3203, "num_input_tokens_seen": 15636312, "step": 23910 }, { "epoch": 12.534067085953879, "grad_norm": 0.22869648039340973, "learning_rate": 1.838826377212317e-05, "loss": 0.5416, "num_input_tokens_seen": 15638616, "step": 23915 }, { "epoch": 12.536687631027254, "grad_norm": 0.13283276557922363, "learning_rate": 1.8377237307544326e-05, "loss": 0.5234, "num_input_tokens_seen": 15642488, "step": 23920 }, { "epoch": 12.53930817610063, "grad_norm": 0.21182328462600708, "learning_rate": 1.8366212228371894e-05, "loss": 0.4264, "num_input_tokens_seen": 15645304, "step": 23925 }, { "epoch": 12.541928721174004, "grad_norm": 0.11057744920253754, "learning_rate": 1.8355188536912173e-05, "loss": 0.5585, "num_input_tokens_seen": 15649208, "step": 23930 }, { "epoch": 12.54454926624738, "grad_norm": 0.28566792607307434, "learning_rate": 1.83441662354712e-05, "loss": 0.5132, "num_input_tokens_seen": 15652120, "step": 23935 }, { "epoch": 12.547169811320755, "grad_norm": 0.5485523343086243, "learning_rate": 1.8333145326354722e-05, "loss": 0.7077, "num_input_tokens_seen": 15654968, "step": 23940 }, { "epoch": 12.54979035639413, "grad_norm": 0.16900664567947388, "learning_rate": 1.8322125811868175e-05, "loss": 0.4775, "num_input_tokens_seen": 15658456, "step": 23945 }, { "epoch": 12.552410901467505, "grad_norm": 0.33621522784233093, "learning_rate": 1.831110769431673e-05, "loss": 0.6473, "num_input_tokens_seen": 15661816, "step": 23950 }, { "epoch": 12.55503144654088, "grad_norm": 0.17987369000911713, "learning_rate": 1.8300090976005247e-05, "loss": 0.5257, "num_input_tokens_seen": 15664664, "step": 23955 }, { "epoch": 12.557651991614255, "grad_norm": 0.16575987637043, "learning_rate": 1.8289075659238285e-05, "loss": 0.5128, "num_input_tokens_seen": 15667448, "step": 23960 }, { "epoch": 12.56027253668763, "grad_norm": 0.16768239438533783, "learning_rate": 1.8278061746320134e-05, "loss": 0.4794, "num_input_tokens_seen": 15671512, "step": 23965 }, { "epoch": 12.562893081761006, "grad_norm": 0.11991017311811447, "learning_rate": 1.8267049239554786e-05, "loss": 0.3863, "num_input_tokens_seen": 15675832, "step": 23970 }, { "epoch": 12.565513626834381, "grad_norm": 0.18630348145961761, "learning_rate": 1.8256038141245912e-05, "loss": 0.4161, "num_input_tokens_seen": 15679096, "step": 23975 }, { "epoch": 12.568134171907756, "grad_norm": 0.12789234519004822, "learning_rate": 1.824502845369693e-05, "loss": 0.449, "num_input_tokens_seen": 15682200, "step": 23980 }, { "epoch": 12.570754716981131, "grad_norm": 0.20907916128635406, "learning_rate": 1.8234020179210938e-05, "loss": 0.5166, "num_input_tokens_seen": 15684952, "step": 23985 }, { "epoch": 12.573375262054507, "grad_norm": 0.1433137059211731, "learning_rate": 1.8223013320090715e-05, "loss": 0.3826, "num_input_tokens_seen": 15687896, "step": 23990 }, { "epoch": 12.575995807127882, "grad_norm": 0.18713712692260742, "learning_rate": 1.82120078786388e-05, "loss": 0.6068, "num_input_tokens_seen": 15690552, "step": 23995 }, { "epoch": 12.578616352201259, "grad_norm": 0.26136431097984314, "learning_rate": 1.820100385715739e-05, "loss": 0.3951, "num_input_tokens_seen": 15693528, "step": 24000 }, { "epoch": 12.581236897274634, "grad_norm": 0.1890772134065628, "learning_rate": 1.8190001257948414e-05, "loss": 0.3479, "num_input_tokens_seen": 15697464, "step": 24005 }, { "epoch": 12.583857442348009, "grad_norm": 0.21331943571567535, "learning_rate": 1.8179000083313483e-05, "loss": 0.4573, "num_input_tokens_seen": 15700792, "step": 24010 }, { "epoch": 12.586477987421384, "grad_norm": 0.17353399097919464, "learning_rate": 1.8168000335553907e-05, "loss": 0.5573, "num_input_tokens_seen": 15704600, "step": 24015 }, { "epoch": 12.58909853249476, "grad_norm": 0.15803422033786774, "learning_rate": 1.8157002016970723e-05, "loss": 0.5042, "num_input_tokens_seen": 15707800, "step": 24020 }, { "epoch": 12.591719077568134, "grad_norm": 0.11089042574167252, "learning_rate": 1.8146005129864644e-05, "loss": 0.467, "num_input_tokens_seen": 15710616, "step": 24025 }, { "epoch": 12.59433962264151, "grad_norm": 0.11620418727397919, "learning_rate": 1.81350096765361e-05, "loss": 0.3915, "num_input_tokens_seen": 15714264, "step": 24030 }, { "epoch": 12.596960167714885, "grad_norm": 0.12480755895376205, "learning_rate": 1.812401565928521e-05, "loss": 0.4213, "num_input_tokens_seen": 15717400, "step": 24035 }, { "epoch": 12.59958071278826, "grad_norm": 0.22879236936569214, "learning_rate": 1.8113023080411806e-05, "loss": 0.5112, "num_input_tokens_seen": 15719864, "step": 24040 }, { "epoch": 12.602201257861635, "grad_norm": 0.11848493665456772, "learning_rate": 1.810203194221539e-05, "loss": 0.5488, "num_input_tokens_seen": 15724088, "step": 24045 }, { "epoch": 12.60482180293501, "grad_norm": 0.15824677050113678, "learning_rate": 1.8091042246995206e-05, "loss": 0.4984, "num_input_tokens_seen": 15727128, "step": 24050 }, { "epoch": 12.607442348008385, "grad_norm": 0.2248276025056839, "learning_rate": 1.8080053997050154e-05, "loss": 0.6559, "num_input_tokens_seen": 15729688, "step": 24055 }, { "epoch": 12.61006289308176, "grad_norm": 0.18708321452140808, "learning_rate": 1.8069067194678867e-05, "loss": 0.4745, "num_input_tokens_seen": 15732696, "step": 24060 }, { "epoch": 12.612683438155136, "grad_norm": 0.09678550809621811, "learning_rate": 1.8058081842179642e-05, "loss": 0.4756, "num_input_tokens_seen": 15735608, "step": 24065 }, { "epoch": 12.615303983228511, "grad_norm": 0.12608611583709717, "learning_rate": 1.804709794185051e-05, "loss": 0.4565, "num_input_tokens_seen": 15739256, "step": 24070 }, { "epoch": 12.617924528301886, "grad_norm": 0.12194139510393143, "learning_rate": 1.8036115495989164e-05, "loss": 0.3929, "num_input_tokens_seen": 15741944, "step": 24075 }, { "epoch": 12.620545073375261, "grad_norm": 0.12691305577754974, "learning_rate": 1.8025134506892997e-05, "loss": 0.4942, "num_input_tokens_seen": 15744888, "step": 24080 }, { "epoch": 12.623165618448636, "grad_norm": 0.18865320086479187, "learning_rate": 1.8014154976859126e-05, "loss": 0.4394, "num_input_tokens_seen": 15747736, "step": 24085 }, { "epoch": 12.625786163522012, "grad_norm": 0.12691687047481537, "learning_rate": 1.800317690818433e-05, "loss": 0.5012, "num_input_tokens_seen": 15750616, "step": 24090 }, { "epoch": 12.628406708595389, "grad_norm": 0.16208820044994354, "learning_rate": 1.799220030316511e-05, "loss": 0.3693, "num_input_tokens_seen": 15754488, "step": 24095 }, { "epoch": 12.631027253668764, "grad_norm": 0.18204542994499207, "learning_rate": 1.7981225164097633e-05, "loss": 0.4328, "num_input_tokens_seen": 15757432, "step": 24100 }, { "epoch": 12.633647798742139, "grad_norm": 0.17030461132526398, "learning_rate": 1.7970251493277774e-05, "loss": 0.4343, "num_input_tokens_seen": 15761208, "step": 24105 }, { "epoch": 12.636268343815514, "grad_norm": 0.14591136574745178, "learning_rate": 1.7959279293001102e-05, "loss": 0.4121, "num_input_tokens_seen": 15764152, "step": 24110 }, { "epoch": 12.63888888888889, "grad_norm": 0.16551275551319122, "learning_rate": 1.7948308565562876e-05, "loss": 0.5995, "num_input_tokens_seen": 15767064, "step": 24115 }, { "epoch": 12.641509433962264, "grad_norm": 0.037276461720466614, "learning_rate": 1.793733931325805e-05, "loss": 0.4526, "num_input_tokens_seen": 15773560, "step": 24120 }, { "epoch": 12.64412997903564, "grad_norm": 0.18350151181221008, "learning_rate": 1.792637153838127e-05, "loss": 0.4344, "num_input_tokens_seen": 15777784, "step": 24125 }, { "epoch": 12.646750524109015, "grad_norm": 0.19890239834785461, "learning_rate": 1.7915405243226864e-05, "loss": 0.4563, "num_input_tokens_seen": 15780760, "step": 24130 }, { "epoch": 12.64937106918239, "grad_norm": 0.175607368350029, "learning_rate": 1.7904440430088847e-05, "loss": 0.4731, "num_input_tokens_seen": 15784312, "step": 24135 }, { "epoch": 12.651991614255765, "grad_norm": 0.19576406478881836, "learning_rate": 1.7893477101260947e-05, "loss": 0.4184, "num_input_tokens_seen": 15787000, "step": 24140 }, { "epoch": 12.65461215932914, "grad_norm": 0.11949757486581802, "learning_rate": 1.7882515259036558e-05, "loss": 0.5237, "num_input_tokens_seen": 15790104, "step": 24145 }, { "epoch": 12.657232704402515, "grad_norm": 0.14371705055236816, "learning_rate": 1.787155490570878e-05, "loss": 0.4972, "num_input_tokens_seen": 15793528, "step": 24150 }, { "epoch": 12.65985324947589, "grad_norm": 0.21866539120674133, "learning_rate": 1.7860596043570398e-05, "loss": 0.5276, "num_input_tokens_seen": 15795992, "step": 24155 }, { "epoch": 12.662473794549266, "grad_norm": 0.24060691893100739, "learning_rate": 1.7849638674913863e-05, "loss": 0.4703, "num_input_tokens_seen": 15798744, "step": 24160 }, { "epoch": 12.665094339622641, "grad_norm": 0.13969123363494873, "learning_rate": 1.7838682802031344e-05, "loss": 0.5101, "num_input_tokens_seen": 15801208, "step": 24165 }, { "epoch": 12.667714884696016, "grad_norm": 0.11220429092645645, "learning_rate": 1.782772842721468e-05, "loss": 0.4713, "num_input_tokens_seen": 15805176, "step": 24170 }, { "epoch": 12.670335429769391, "grad_norm": 0.4370885491371155, "learning_rate": 1.7816775552755406e-05, "loss": 0.4276, "num_input_tokens_seen": 15807704, "step": 24175 }, { "epoch": 12.672955974842766, "grad_norm": 0.18346263468265533, "learning_rate": 1.780582418094473e-05, "loss": 0.4305, "num_input_tokens_seen": 15810680, "step": 24180 }, { "epoch": 12.675576519916142, "grad_norm": 0.3076036870479584, "learning_rate": 1.7794874314073573e-05, "loss": 0.4562, "num_input_tokens_seen": 15813688, "step": 24185 }, { "epoch": 12.678197064989519, "grad_norm": 0.12700696289539337, "learning_rate": 1.7783925954432503e-05, "loss": 0.3888, "num_input_tokens_seen": 15817752, "step": 24190 }, { "epoch": 12.680817610062894, "grad_norm": 0.187956765294075, "learning_rate": 1.777297910431179e-05, "loss": 0.6287, "num_input_tokens_seen": 15820024, "step": 24195 }, { "epoch": 12.683438155136269, "grad_norm": 0.13724836707115173, "learning_rate": 1.7762033766001402e-05, "loss": 0.4189, "num_input_tokens_seen": 15825368, "step": 24200 }, { "epoch": 12.686058700209644, "grad_norm": 0.0896611362695694, "learning_rate": 1.7751089941790977e-05, "loss": 0.4952, "num_input_tokens_seen": 15829144, "step": 24205 }, { "epoch": 12.68867924528302, "grad_norm": 0.3056173622608185, "learning_rate": 1.7740147633969844e-05, "loss": 0.4855, "num_input_tokens_seen": 15832248, "step": 24210 }, { "epoch": 12.691299790356394, "grad_norm": 0.13881345093250275, "learning_rate": 1.772920684482699e-05, "loss": 0.4512, "num_input_tokens_seen": 15834840, "step": 24215 }, { "epoch": 12.69392033542977, "grad_norm": 0.1339552253484726, "learning_rate": 1.771826757665112e-05, "loss": 0.3832, "num_input_tokens_seen": 15837464, "step": 24220 }, { "epoch": 12.696540880503145, "grad_norm": 0.1482408493757248, "learning_rate": 1.770732983173059e-05, "loss": 0.4094, "num_input_tokens_seen": 15841752, "step": 24225 }, { "epoch": 12.69916142557652, "grad_norm": 0.20447301864624023, "learning_rate": 1.7696393612353467e-05, "loss": 0.4544, "num_input_tokens_seen": 15844600, "step": 24230 }, { "epoch": 12.701781970649895, "grad_norm": 0.14733144640922546, "learning_rate": 1.7685458920807474e-05, "loss": 0.4001, "num_input_tokens_seen": 15847640, "step": 24235 }, { "epoch": 12.70440251572327, "grad_norm": 0.21090607345104218, "learning_rate": 1.7674525759380033e-05, "loss": 0.5208, "num_input_tokens_seen": 15850904, "step": 24240 }, { "epoch": 12.707023060796645, "grad_norm": 0.1972992867231369, "learning_rate": 1.766359413035823e-05, "loss": 0.5523, "num_input_tokens_seen": 15853880, "step": 24245 }, { "epoch": 12.70964360587002, "grad_norm": 0.095140740275383, "learning_rate": 1.7652664036028827e-05, "loss": 0.5164, "num_input_tokens_seen": 15857016, "step": 24250 }, { "epoch": 12.712264150943396, "grad_norm": 0.16898813843727112, "learning_rate": 1.7641735478678288e-05, "loss": 0.5553, "num_input_tokens_seen": 15860888, "step": 24255 }, { "epoch": 12.714884696016771, "grad_norm": 0.22967684268951416, "learning_rate": 1.763080846059274e-05, "loss": 0.5116, "num_input_tokens_seen": 15863320, "step": 24260 }, { "epoch": 12.717505241090146, "grad_norm": 0.16601018607616425, "learning_rate": 1.7619882984057987e-05, "loss": 0.4817, "num_input_tokens_seen": 15867032, "step": 24265 }, { "epoch": 12.720125786163521, "grad_norm": 0.21106085181236267, "learning_rate": 1.760895905135953e-05, "loss": 0.3705, "num_input_tokens_seen": 15870552, "step": 24270 }, { "epoch": 12.722746331236896, "grad_norm": 0.19655834138393402, "learning_rate": 1.7598036664782508e-05, "loss": 0.4958, "num_input_tokens_seen": 15873976, "step": 24275 }, { "epoch": 12.725366876310272, "grad_norm": 0.23066797852516174, "learning_rate": 1.7587115826611766e-05, "loss": 0.6039, "num_input_tokens_seen": 15876664, "step": 24280 }, { "epoch": 12.727987421383649, "grad_norm": 0.30869540572166443, "learning_rate": 1.7576196539131828e-05, "loss": 0.5833, "num_input_tokens_seen": 15879736, "step": 24285 }, { "epoch": 12.730607966457024, "grad_norm": 0.21341270208358765, "learning_rate": 1.7565278804626872e-05, "loss": 0.4838, "num_input_tokens_seen": 15883192, "step": 24290 }, { "epoch": 12.733228511530399, "grad_norm": 0.2364722192287445, "learning_rate": 1.755436262538078e-05, "loss": 0.5385, "num_input_tokens_seen": 15886520, "step": 24295 }, { "epoch": 12.735849056603774, "grad_norm": 0.20503860712051392, "learning_rate": 1.7543448003677082e-05, "loss": 0.5268, "num_input_tokens_seen": 15890584, "step": 24300 }, { "epoch": 12.73846960167715, "grad_norm": 0.15421482920646667, "learning_rate": 1.7532534941798985e-05, "loss": 0.4166, "num_input_tokens_seen": 15894776, "step": 24305 }, { "epoch": 12.741090146750524, "grad_norm": 0.15552373230457306, "learning_rate": 1.752162344202939e-05, "loss": 0.5758, "num_input_tokens_seen": 15903000, "step": 24310 }, { "epoch": 12.7437106918239, "grad_norm": 0.16085317730903625, "learning_rate": 1.751071350665085e-05, "loss": 0.4397, "num_input_tokens_seen": 15905624, "step": 24315 }, { "epoch": 12.746331236897275, "grad_norm": 0.14643055200576782, "learning_rate": 1.7499805137945602e-05, "loss": 0.4728, "num_input_tokens_seen": 15909336, "step": 24320 }, { "epoch": 12.74895178197065, "grad_norm": 0.09944944828748703, "learning_rate": 1.7488898338195566e-05, "loss": 0.4703, "num_input_tokens_seen": 15912664, "step": 24325 }, { "epoch": 12.751572327044025, "grad_norm": 0.16209974884986877, "learning_rate": 1.7477993109682288e-05, "loss": 0.4982, "num_input_tokens_seen": 15915928, "step": 24330 }, { "epoch": 12.7541928721174, "grad_norm": 0.09590137749910355, "learning_rate": 1.7467089454687043e-05, "loss": 0.5403, "num_input_tokens_seen": 15918264, "step": 24335 }, { "epoch": 12.756813417190775, "grad_norm": 0.33026716113090515, "learning_rate": 1.745618737549074e-05, "loss": 0.5069, "num_input_tokens_seen": 15921112, "step": 24340 }, { "epoch": 12.75943396226415, "grad_norm": 0.2349838763475418, "learning_rate": 1.7445286874373974e-05, "loss": 0.5103, "num_input_tokens_seen": 15923512, "step": 24345 }, { "epoch": 12.762054507337526, "grad_norm": 0.12493868917226791, "learning_rate": 1.743438795361701e-05, "loss": 0.5622, "num_input_tokens_seen": 15926296, "step": 24350 }, { "epoch": 12.764675052410901, "grad_norm": 0.14334236085414886, "learning_rate": 1.742349061549978e-05, "loss": 0.4293, "num_input_tokens_seen": 15930520, "step": 24355 }, { "epoch": 12.767295597484276, "grad_norm": 0.18309292197227478, "learning_rate": 1.741259486230186e-05, "loss": 0.4297, "num_input_tokens_seen": 15933592, "step": 24360 }, { "epoch": 12.769916142557651, "grad_norm": 0.13083724677562714, "learning_rate": 1.7401700696302546e-05, "loss": 0.322, "num_input_tokens_seen": 15937016, "step": 24365 }, { "epoch": 12.772536687631026, "grad_norm": 0.16155874729156494, "learning_rate": 1.739080811978075e-05, "loss": 0.6679, "num_input_tokens_seen": 15940280, "step": 24370 }, { "epoch": 12.775157232704402, "grad_norm": 0.2061426192522049, "learning_rate": 1.7379917135015093e-05, "loss": 0.5348, "num_input_tokens_seen": 15943864, "step": 24375 }, { "epoch": 12.777777777777779, "grad_norm": 0.2605538070201874, "learning_rate": 1.736902774428384e-05, "loss": 0.6488, "num_input_tokens_seen": 15946904, "step": 24380 }, { "epoch": 12.780398322851154, "grad_norm": 0.18466302752494812, "learning_rate": 1.735813994986491e-05, "loss": 0.5373, "num_input_tokens_seen": 15949688, "step": 24385 }, { "epoch": 12.783018867924529, "grad_norm": 0.26417866349220276, "learning_rate": 1.7347253754035933e-05, "loss": 0.3797, "num_input_tokens_seen": 15953016, "step": 24390 }, { "epoch": 12.785639412997904, "grad_norm": 0.3000727891921997, "learning_rate": 1.7336369159074154e-05, "loss": 0.4924, "num_input_tokens_seen": 15955608, "step": 24395 }, { "epoch": 12.78825995807128, "grad_norm": 0.21653461456298828, "learning_rate": 1.7325486167256522e-05, "loss": 0.5405, "num_input_tokens_seen": 15959608, "step": 24400 }, { "epoch": 12.790880503144654, "grad_norm": 0.16040901839733124, "learning_rate": 1.7314604780859623e-05, "loss": 0.4754, "num_input_tokens_seen": 15963000, "step": 24405 }, { "epoch": 12.79350104821803, "grad_norm": 0.24862302839756012, "learning_rate": 1.730372500215973e-05, "loss": 0.5516, "num_input_tokens_seen": 15967928, "step": 24410 }, { "epoch": 12.796121593291405, "grad_norm": 0.26075777411460876, "learning_rate": 1.7292846833432763e-05, "loss": 0.3226, "num_input_tokens_seen": 15971416, "step": 24415 }, { "epoch": 12.79874213836478, "grad_norm": 0.1739499270915985, "learning_rate": 1.7281970276954305e-05, "loss": 0.5579, "num_input_tokens_seen": 15974104, "step": 24420 }, { "epoch": 12.801362683438155, "grad_norm": 0.12111770361661911, "learning_rate": 1.727109533499961e-05, "loss": 0.4302, "num_input_tokens_seen": 15978200, "step": 24425 }, { "epoch": 12.80398322851153, "grad_norm": 0.15302374958992004, "learning_rate": 1.72602220098436e-05, "loss": 0.4239, "num_input_tokens_seen": 15981624, "step": 24430 }, { "epoch": 12.806603773584905, "grad_norm": 0.16014450788497925, "learning_rate": 1.7249350303760846e-05, "loss": 0.5257, "num_input_tokens_seen": 15985016, "step": 24435 }, { "epoch": 12.80922431865828, "grad_norm": 0.1461511105298996, "learning_rate": 1.7238480219025593e-05, "loss": 0.6459, "num_input_tokens_seen": 15988856, "step": 24440 }, { "epoch": 12.811844863731656, "grad_norm": 0.1262458860874176, "learning_rate": 1.7227611757911723e-05, "loss": 0.5218, "num_input_tokens_seen": 15992568, "step": 24445 }, { "epoch": 12.814465408805031, "grad_norm": 0.1751302182674408, "learning_rate": 1.7216744922692794e-05, "loss": 0.3445, "num_input_tokens_seen": 15996216, "step": 24450 }, { "epoch": 12.817085953878406, "grad_norm": 0.2502143383026123, "learning_rate": 1.7205879715642037e-05, "loss": 0.5503, "num_input_tokens_seen": 15999224, "step": 24455 }, { "epoch": 12.819706498951781, "grad_norm": 0.13465306162834167, "learning_rate": 1.719501613903232e-05, "loss": 0.5206, "num_input_tokens_seen": 16002456, "step": 24460 }, { "epoch": 12.822327044025156, "grad_norm": 0.11411190032958984, "learning_rate": 1.7184154195136192e-05, "loss": 0.432, "num_input_tokens_seen": 16005240, "step": 24465 }, { "epoch": 12.824947589098532, "grad_norm": 0.11586292833089828, "learning_rate": 1.7173293886225837e-05, "loss": 0.399, "num_input_tokens_seen": 16009464, "step": 24470 }, { "epoch": 12.827568134171909, "grad_norm": 0.1333225816488266, "learning_rate": 1.71624352145731e-05, "loss": 0.5323, "num_input_tokens_seen": 16012888, "step": 24475 }, { "epoch": 12.830188679245284, "grad_norm": 0.2187608778476715, "learning_rate": 1.7151578182449508e-05, "loss": 0.6506, "num_input_tokens_seen": 16015960, "step": 24480 }, { "epoch": 12.832809224318659, "grad_norm": 0.21714359521865845, "learning_rate": 1.7140722792126212e-05, "loss": 0.4632, "num_input_tokens_seen": 16019704, "step": 24485 }, { "epoch": 12.835429769392034, "grad_norm": 0.1409853845834732, "learning_rate": 1.7129869045874053e-05, "loss": 0.509, "num_input_tokens_seen": 16022488, "step": 24490 }, { "epoch": 12.83805031446541, "grad_norm": 0.14476554095745087, "learning_rate": 1.7119016945963506e-05, "loss": 0.3353, "num_input_tokens_seen": 16025496, "step": 24495 }, { "epoch": 12.840670859538784, "grad_norm": 0.17668060958385468, "learning_rate": 1.7108166494664694e-05, "loss": 0.5218, "num_input_tokens_seen": 16028152, "step": 24500 }, { "epoch": 12.84329140461216, "grad_norm": 0.221660777926445, "learning_rate": 1.7097317694247417e-05, "loss": 0.3577, "num_input_tokens_seen": 16031352, "step": 24505 }, { "epoch": 12.845911949685535, "grad_norm": 0.1947484016418457, "learning_rate": 1.7086470546981122e-05, "loss": 0.4667, "num_input_tokens_seen": 16035736, "step": 24510 }, { "epoch": 12.84853249475891, "grad_norm": 0.18879228830337524, "learning_rate": 1.7075625055134904e-05, "loss": 0.4927, "num_input_tokens_seen": 16040536, "step": 24515 }, { "epoch": 12.851153039832285, "grad_norm": 0.1697598546743393, "learning_rate": 1.706478122097752e-05, "loss": 0.5831, "num_input_tokens_seen": 16043960, "step": 24520 }, { "epoch": 12.85377358490566, "grad_norm": 0.14819566905498505, "learning_rate": 1.7053939046777384e-05, "loss": 0.4931, "num_input_tokens_seen": 16047800, "step": 24525 }, { "epoch": 12.856394129979035, "grad_norm": 0.24038191139698029, "learning_rate": 1.7043098534802533e-05, "loss": 0.5753, "num_input_tokens_seen": 16050360, "step": 24530 }, { "epoch": 12.85901467505241, "grad_norm": 0.20173406600952148, "learning_rate": 1.7032259687320695e-05, "loss": 0.4926, "num_input_tokens_seen": 16053176, "step": 24535 }, { "epoch": 12.861635220125786, "grad_norm": 0.13545316457748413, "learning_rate": 1.7021422506599228e-05, "loss": 0.4211, "num_input_tokens_seen": 16055896, "step": 24540 }, { "epoch": 12.864255765199161, "grad_norm": 0.1685250997543335, "learning_rate": 1.7010586994905155e-05, "loss": 0.552, "num_input_tokens_seen": 16058968, "step": 24545 }, { "epoch": 12.866876310272536, "grad_norm": 0.1291782408952713, "learning_rate": 1.6999753154505123e-05, "loss": 0.3842, "num_input_tokens_seen": 16061912, "step": 24550 }, { "epoch": 12.869496855345911, "grad_norm": 0.11575187742710114, "learning_rate": 1.6988920987665476e-05, "loss": 0.5696, "num_input_tokens_seen": 16064504, "step": 24555 }, { "epoch": 12.872117400419286, "grad_norm": 0.21487310528755188, "learning_rate": 1.6978090496652156e-05, "loss": 0.5013, "num_input_tokens_seen": 16067320, "step": 24560 }, { "epoch": 12.874737945492662, "grad_norm": 0.23721763491630554, "learning_rate": 1.6967261683730778e-05, "loss": 0.4738, "num_input_tokens_seen": 16070968, "step": 24565 }, { "epoch": 12.877358490566039, "grad_norm": 0.11048106104135513, "learning_rate": 1.695643455116662e-05, "loss": 0.5414, "num_input_tokens_seen": 16074040, "step": 24570 }, { "epoch": 12.879979035639414, "grad_norm": 0.188421368598938, "learning_rate": 1.694560910122458e-05, "loss": 0.4824, "num_input_tokens_seen": 16076792, "step": 24575 }, { "epoch": 12.882599580712789, "grad_norm": 0.18734312057495117, "learning_rate": 1.693478533616924e-05, "loss": 0.4617, "num_input_tokens_seen": 16080376, "step": 24580 }, { "epoch": 12.885220125786164, "grad_norm": 0.09326382726430893, "learning_rate": 1.692396325826479e-05, "loss": 0.4159, "num_input_tokens_seen": 16083640, "step": 24585 }, { "epoch": 12.88784067085954, "grad_norm": 0.11195652186870575, "learning_rate": 1.6913142869775088e-05, "loss": 0.4294, "num_input_tokens_seen": 16087064, "step": 24590 }, { "epoch": 12.890461215932914, "grad_norm": 0.16351318359375, "learning_rate": 1.6902324172963635e-05, "loss": 0.6356, "num_input_tokens_seen": 16090584, "step": 24595 }, { "epoch": 12.89308176100629, "grad_norm": 0.1471259891986847, "learning_rate": 1.6891507170093586e-05, "loss": 0.6217, "num_input_tokens_seen": 16094200, "step": 24600 }, { "epoch": 12.895702306079665, "grad_norm": 0.23684853315353394, "learning_rate": 1.6880691863427725e-05, "loss": 0.6392, "num_input_tokens_seen": 16096952, "step": 24605 }, { "epoch": 12.89832285115304, "grad_norm": 0.13136552274227142, "learning_rate": 1.686987825522851e-05, "loss": 0.582, "num_input_tokens_seen": 16100024, "step": 24610 }, { "epoch": 12.900943396226415, "grad_norm": 0.12148097902536392, "learning_rate": 1.6859066347758002e-05, "loss": 0.5336, "num_input_tokens_seen": 16104824, "step": 24615 }, { "epoch": 12.90356394129979, "grad_norm": 0.2249123752117157, "learning_rate": 1.6848256143277933e-05, "loss": 0.3568, "num_input_tokens_seen": 16108536, "step": 24620 }, { "epoch": 12.906184486373165, "grad_norm": 0.18614323437213898, "learning_rate": 1.683744764404968e-05, "loss": 0.4689, "num_input_tokens_seen": 16111864, "step": 24625 }, { "epoch": 12.90880503144654, "grad_norm": 0.11275561898946762, "learning_rate": 1.6826640852334257e-05, "loss": 0.5079, "num_input_tokens_seen": 16115640, "step": 24630 }, { "epoch": 12.911425576519916, "grad_norm": 0.11483161896467209, "learning_rate": 1.6815835770392324e-05, "loss": 0.5217, "num_input_tokens_seen": 16119032, "step": 24635 }, { "epoch": 12.914046121593291, "grad_norm": 0.136033833026886, "learning_rate": 1.6805032400484184e-05, "loss": 0.3877, "num_input_tokens_seen": 16122200, "step": 24640 }, { "epoch": 12.916666666666666, "grad_norm": 0.19655457139015198, "learning_rate": 1.6794230744869767e-05, "loss": 0.4982, "num_input_tokens_seen": 16125080, "step": 24645 }, { "epoch": 12.919287211740041, "grad_norm": 0.16031484305858612, "learning_rate": 1.6783430805808658e-05, "loss": 0.5208, "num_input_tokens_seen": 16129304, "step": 24650 }, { "epoch": 12.921907756813416, "grad_norm": 0.15964604914188385, "learning_rate": 1.6772632585560084e-05, "loss": 0.6738, "num_input_tokens_seen": 16132664, "step": 24655 }, { "epoch": 12.924528301886792, "grad_norm": 0.16830852627754211, "learning_rate": 1.6761836086382908e-05, "loss": 0.422, "num_input_tokens_seen": 16135608, "step": 24660 }, { "epoch": 12.927148846960169, "grad_norm": 0.23906804621219635, "learning_rate": 1.6751041310535653e-05, "loss": 0.5671, "num_input_tokens_seen": 16138584, "step": 24665 }, { "epoch": 12.929769392033544, "grad_norm": 0.3248947262763977, "learning_rate": 1.6740248260276437e-05, "loss": 0.4807, "num_input_tokens_seen": 16141176, "step": 24670 }, { "epoch": 12.932389937106919, "grad_norm": 0.19824105501174927, "learning_rate": 1.6729456937863047e-05, "loss": 0.3946, "num_input_tokens_seen": 16144024, "step": 24675 }, { "epoch": 12.935010482180294, "grad_norm": 0.22897911071777344, "learning_rate": 1.6718667345552918e-05, "loss": 0.4085, "num_input_tokens_seen": 16148120, "step": 24680 }, { "epoch": 12.93763102725367, "grad_norm": 0.21169491112232208, "learning_rate": 1.6707879485603097e-05, "loss": 0.4851, "num_input_tokens_seen": 16151416, "step": 24685 }, { "epoch": 12.940251572327044, "grad_norm": 0.21044209599494934, "learning_rate": 1.669709336027029e-05, "loss": 0.668, "num_input_tokens_seen": 16154392, "step": 24690 }, { "epoch": 12.94287211740042, "grad_norm": 0.17176364362239838, "learning_rate": 1.6686308971810833e-05, "loss": 0.5663, "num_input_tokens_seen": 16157048, "step": 24695 }, { "epoch": 12.945492662473795, "grad_norm": 0.177351713180542, "learning_rate": 1.6675526322480677e-05, "loss": 0.6002, "num_input_tokens_seen": 16160152, "step": 24700 }, { "epoch": 12.94811320754717, "grad_norm": 0.19781257212162018, "learning_rate": 1.6664745414535454e-05, "loss": 0.4867, "num_input_tokens_seen": 16163352, "step": 24705 }, { "epoch": 12.950733752620545, "grad_norm": 0.1532069444656372, "learning_rate": 1.665396625023039e-05, "loss": 0.6049, "num_input_tokens_seen": 16166488, "step": 24710 }, { "epoch": 12.95335429769392, "grad_norm": 0.27584630250930786, "learning_rate": 1.6643188831820375e-05, "loss": 0.4794, "num_input_tokens_seen": 16169560, "step": 24715 }, { "epoch": 12.955974842767295, "grad_norm": 0.10071884840726852, "learning_rate": 1.6632413161559912e-05, "loss": 0.3943, "num_input_tokens_seen": 16172376, "step": 24720 }, { "epoch": 12.95859538784067, "grad_norm": 0.14502927660942078, "learning_rate": 1.6621639241703162e-05, "loss": 0.5015, "num_input_tokens_seen": 16175512, "step": 24725 }, { "epoch": 12.961215932914046, "grad_norm": 0.14810657501220703, "learning_rate": 1.6610867074503893e-05, "loss": 0.4614, "num_input_tokens_seen": 16178744, "step": 24730 }, { "epoch": 12.963836477987421, "grad_norm": 0.2434748262166977, "learning_rate": 1.660009666221552e-05, "loss": 0.5479, "num_input_tokens_seen": 16182936, "step": 24735 }, { "epoch": 12.966457023060796, "grad_norm": 0.2528781294822693, "learning_rate": 1.6589328007091093e-05, "loss": 0.497, "num_input_tokens_seen": 16185880, "step": 24740 }, { "epoch": 12.969077568134171, "grad_norm": 0.20238514244556427, "learning_rate": 1.65785611113833e-05, "loss": 0.4467, "num_input_tokens_seen": 16188792, "step": 24745 }, { "epoch": 12.971698113207546, "grad_norm": 0.17169040441513062, "learning_rate": 1.6567795977344447e-05, "loss": 0.3816, "num_input_tokens_seen": 16191480, "step": 24750 }, { "epoch": 12.974318658280922, "grad_norm": 0.17941711843013763, "learning_rate": 1.6557032607226466e-05, "loss": 0.3823, "num_input_tokens_seen": 16193784, "step": 24755 }, { "epoch": 12.976939203354299, "grad_norm": 0.09969950467348099, "learning_rate": 1.654627100328095e-05, "loss": 0.415, "num_input_tokens_seen": 16196312, "step": 24760 }, { "epoch": 12.979559748427674, "grad_norm": 0.20016653835773468, "learning_rate": 1.6535511167759085e-05, "loss": 0.4831, "num_input_tokens_seen": 16199352, "step": 24765 }, { "epoch": 12.982180293501049, "grad_norm": 0.17907321453094482, "learning_rate": 1.652475310291172e-05, "loss": 0.4688, "num_input_tokens_seen": 16202872, "step": 24770 }, { "epoch": 12.984800838574424, "grad_norm": 0.29018545150756836, "learning_rate": 1.6513996810989317e-05, "loss": 0.3955, "num_input_tokens_seen": 16205944, "step": 24775 }, { "epoch": 12.9874213836478, "grad_norm": 0.10273213684558868, "learning_rate": 1.6503242294241973e-05, "loss": 0.4552, "num_input_tokens_seen": 16208984, "step": 24780 }, { "epoch": 12.990041928721174, "grad_norm": 0.1127755269408226, "learning_rate": 1.6492489554919404e-05, "loss": 0.5489, "num_input_tokens_seen": 16212152, "step": 24785 }, { "epoch": 12.99266247379455, "grad_norm": 0.21201840043067932, "learning_rate": 1.6481738595270953e-05, "loss": 0.4188, "num_input_tokens_seen": 16216312, "step": 24790 }, { "epoch": 12.995283018867925, "grad_norm": 0.14370772242546082, "learning_rate": 1.6470989417545608e-05, "loss": 0.4134, "num_input_tokens_seen": 16220120, "step": 24795 }, { "epoch": 12.9979035639413, "grad_norm": 0.2231808453798294, "learning_rate": 1.646024202399197e-05, "loss": 0.4481, "num_input_tokens_seen": 16223384, "step": 24800 }, { "epoch": 13.0, "eval_loss": 0.46201664209365845, "eval_runtime": 9.9661, "eval_samples_per_second": 85.088, "eval_steps_per_second": 21.272, "num_input_tokens_seen": 16225896, "step": 24804 }, { "epoch": 13.000524109014675, "grad_norm": 0.20117688179016113, "learning_rate": 1.6449496416858284e-05, "loss": 0.5846, "num_input_tokens_seen": 16226312, "step": 24805 }, { "epoch": 13.00314465408805, "grad_norm": 0.22711077332496643, "learning_rate": 1.64387525983924e-05, "loss": 0.5089, "num_input_tokens_seen": 16230504, "step": 24810 }, { "epoch": 13.005765199161425, "grad_norm": 0.17673994600772858, "learning_rate": 1.6428010570841788e-05, "loss": 0.4574, "num_input_tokens_seen": 16234088, "step": 24815 }, { "epoch": 13.0083857442348, "grad_norm": 0.18623138964176178, "learning_rate": 1.6417270336453573e-05, "loss": 0.3506, "num_input_tokens_seen": 16237288, "step": 24820 }, { "epoch": 13.011006289308176, "grad_norm": 0.16826318204402924, "learning_rate": 1.6406531897474492e-05, "loss": 0.4056, "num_input_tokens_seen": 16240904, "step": 24825 }, { "epoch": 13.01362683438155, "grad_norm": 0.17235921323299408, "learning_rate": 1.639579525615089e-05, "loss": 0.4938, "num_input_tokens_seen": 16243752, "step": 24830 }, { "epoch": 13.016247379454926, "grad_norm": 0.1604394167661667, "learning_rate": 1.638506041472877e-05, "loss": 0.3889, "num_input_tokens_seen": 16247368, "step": 24835 }, { "epoch": 13.018867924528301, "grad_norm": 0.31625327467918396, "learning_rate": 1.6374327375453724e-05, "loss": 0.4689, "num_input_tokens_seen": 16250088, "step": 24840 }, { "epoch": 13.021488469601676, "grad_norm": 0.10242490470409393, "learning_rate": 1.6363596140570982e-05, "loss": 0.5234, "num_input_tokens_seen": 16254024, "step": 24845 }, { "epoch": 13.024109014675052, "grad_norm": 0.19808079302310944, "learning_rate": 1.6352866712325395e-05, "loss": 0.5271, "num_input_tokens_seen": 16256040, "step": 24850 }, { "epoch": 13.026729559748428, "grad_norm": 0.20602412521839142, "learning_rate": 1.634213909296144e-05, "loss": 0.4618, "num_input_tokens_seen": 16259016, "step": 24855 }, { "epoch": 13.029350104821804, "grad_norm": 0.11545455455780029, "learning_rate": 1.6331413284723217e-05, "loss": 0.507, "num_input_tokens_seen": 16261576, "step": 24860 }, { "epoch": 13.031970649895179, "grad_norm": 0.14355209469795227, "learning_rate": 1.6320689289854442e-05, "loss": 0.4418, "num_input_tokens_seen": 16265544, "step": 24865 }, { "epoch": 13.034591194968554, "grad_norm": 0.153884157538414, "learning_rate": 1.6309967110598438e-05, "loss": 0.4235, "num_input_tokens_seen": 16267944, "step": 24870 }, { "epoch": 13.03721174004193, "grad_norm": 0.23639293015003204, "learning_rate": 1.6299246749198183e-05, "loss": 0.6351, "num_input_tokens_seen": 16270856, "step": 24875 }, { "epoch": 13.039832285115304, "grad_norm": 0.16709885001182556, "learning_rate": 1.6288528207896236e-05, "loss": 0.3418, "num_input_tokens_seen": 16273480, "step": 24880 }, { "epoch": 13.04245283018868, "grad_norm": 0.1060730591416359, "learning_rate": 1.627781148893481e-05, "loss": 0.4515, "num_input_tokens_seen": 16276712, "step": 24885 }, { "epoch": 13.045073375262055, "grad_norm": 0.1264527142047882, "learning_rate": 1.6267096594555703e-05, "loss": 0.4834, "num_input_tokens_seen": 16279720, "step": 24890 }, { "epoch": 13.04769392033543, "grad_norm": 0.21930326521396637, "learning_rate": 1.6256383527000372e-05, "loss": 0.3914, "num_input_tokens_seen": 16285128, "step": 24895 }, { "epoch": 13.050314465408805, "grad_norm": 0.19070008397102356, "learning_rate": 1.6245672288509843e-05, "loss": 0.4229, "num_input_tokens_seen": 16287880, "step": 24900 }, { "epoch": 13.05293501048218, "grad_norm": 0.21918325126171112, "learning_rate": 1.62349628813248e-05, "loss": 0.497, "num_input_tokens_seen": 16290856, "step": 24905 }, { "epoch": 13.055555555555555, "grad_norm": 0.17453575134277344, "learning_rate": 1.6224255307685525e-05, "loss": 0.4559, "num_input_tokens_seen": 16293800, "step": 24910 }, { "epoch": 13.05817610062893, "grad_norm": 0.21807177364826202, "learning_rate": 1.6213549569831922e-05, "loss": 0.5245, "num_input_tokens_seen": 16296680, "step": 24915 }, { "epoch": 13.060796645702306, "grad_norm": 0.15960796177387238, "learning_rate": 1.62028456700035e-05, "loss": 0.4951, "num_input_tokens_seen": 16299432, "step": 24920 }, { "epoch": 13.06341719077568, "grad_norm": 0.11478348821401596, "learning_rate": 1.619214361043942e-05, "loss": 0.5062, "num_input_tokens_seen": 16302984, "step": 24925 }, { "epoch": 13.066037735849056, "grad_norm": 0.10003091394901276, "learning_rate": 1.61814433933784e-05, "loss": 0.4156, "num_input_tokens_seen": 16306280, "step": 24930 }, { "epoch": 13.068658280922431, "grad_norm": 0.17131918668746948, "learning_rate": 1.617074502105881e-05, "loss": 0.416, "num_input_tokens_seen": 16309736, "step": 24935 }, { "epoch": 13.071278825995806, "grad_norm": 0.1424102485179901, "learning_rate": 1.616004849571864e-05, "loss": 0.3558, "num_input_tokens_seen": 16313288, "step": 24940 }, { "epoch": 13.073899371069182, "grad_norm": 0.1505274623632431, "learning_rate": 1.6149353819595463e-05, "loss": 0.4821, "num_input_tokens_seen": 16315816, "step": 24945 }, { "epoch": 13.076519916142558, "grad_norm": 0.3009074926376343, "learning_rate": 1.6138660994926508e-05, "loss": 0.4256, "num_input_tokens_seen": 16319144, "step": 24950 }, { "epoch": 13.079140461215934, "grad_norm": 0.1232810989022255, "learning_rate": 1.612797002394857e-05, "loss": 0.5189, "num_input_tokens_seen": 16323464, "step": 24955 }, { "epoch": 13.081761006289309, "grad_norm": 0.24349315464496613, "learning_rate": 1.611728090889808e-05, "loss": 0.734, "num_input_tokens_seen": 16330280, "step": 24960 }, { "epoch": 13.084381551362684, "grad_norm": 0.16550205647945404, "learning_rate": 1.610659365201109e-05, "loss": 0.6278, "num_input_tokens_seen": 16333512, "step": 24965 }, { "epoch": 13.08700209643606, "grad_norm": 0.23299871385097504, "learning_rate": 1.609590825552324e-05, "loss": 0.4722, "num_input_tokens_seen": 16335880, "step": 24970 }, { "epoch": 13.089622641509434, "grad_norm": 0.1328705996274948, "learning_rate": 1.6085224721669807e-05, "loss": 0.404, "num_input_tokens_seen": 16338888, "step": 24975 }, { "epoch": 13.09224318658281, "grad_norm": 0.19097618758678436, "learning_rate": 1.607454305268566e-05, "loss": 0.3853, "num_input_tokens_seen": 16341928, "step": 24980 }, { "epoch": 13.094863731656185, "grad_norm": 0.0828530564904213, "learning_rate": 1.606386325080528e-05, "loss": 0.3644, "num_input_tokens_seen": 16346088, "step": 24985 }, { "epoch": 13.09748427672956, "grad_norm": 0.15729105472564697, "learning_rate": 1.6053185318262746e-05, "loss": 0.5709, "num_input_tokens_seen": 16349224, "step": 24990 }, { "epoch": 13.100104821802935, "grad_norm": 0.143016517162323, "learning_rate": 1.6042509257291783e-05, "loss": 0.5097, "num_input_tokens_seen": 16351656, "step": 24995 }, { "epoch": 13.10272536687631, "grad_norm": 0.17734938859939575, "learning_rate": 1.6031835070125685e-05, "loss": 0.3651, "num_input_tokens_seen": 16354952, "step": 25000 }, { "epoch": 13.105345911949685, "grad_norm": 0.09507877379655838, "learning_rate": 1.6021162758997382e-05, "loss": 0.4132, "num_input_tokens_seen": 16358504, "step": 25005 }, { "epoch": 13.10796645702306, "grad_norm": 0.11191652715206146, "learning_rate": 1.6010492326139398e-05, "loss": 0.4442, "num_input_tokens_seen": 16361544, "step": 25010 }, { "epoch": 13.110587002096436, "grad_norm": 0.19773496687412262, "learning_rate": 1.599982377378385e-05, "loss": 0.7793, "num_input_tokens_seen": 16364904, "step": 25015 }, { "epoch": 13.11320754716981, "grad_norm": 0.2418791651725769, "learning_rate": 1.5989157104162496e-05, "loss": 0.4772, "num_input_tokens_seen": 16368104, "step": 25020 }, { "epoch": 13.115828092243186, "grad_norm": 0.1498548835515976, "learning_rate": 1.5978492319506673e-05, "loss": 0.4023, "num_input_tokens_seen": 16370600, "step": 25025 }, { "epoch": 13.118448637316561, "grad_norm": 0.24587398767471313, "learning_rate": 1.596782942204733e-05, "loss": 0.5174, "num_input_tokens_seen": 16373672, "step": 25030 }, { "epoch": 13.121069182389936, "grad_norm": 0.17793087661266327, "learning_rate": 1.595716841401504e-05, "loss": 0.5401, "num_input_tokens_seen": 16376904, "step": 25035 }, { "epoch": 13.123689727463312, "grad_norm": 0.1144162267446518, "learning_rate": 1.594650929763994e-05, "loss": 0.557, "num_input_tokens_seen": 16379400, "step": 25040 }, { "epoch": 13.126310272536688, "grad_norm": 0.14993053674697876, "learning_rate": 1.5935852075151812e-05, "loss": 0.6742, "num_input_tokens_seen": 16382696, "step": 25045 }, { "epoch": 13.128930817610064, "grad_norm": 0.1484978348016739, "learning_rate": 1.5925196748780016e-05, "loss": 0.4083, "num_input_tokens_seen": 16385960, "step": 25050 }, { "epoch": 13.131551362683439, "grad_norm": 0.17351561784744263, "learning_rate": 1.5914543320753528e-05, "loss": 0.4991, "num_input_tokens_seen": 16389896, "step": 25055 }, { "epoch": 13.134171907756814, "grad_norm": 0.15964558720588684, "learning_rate": 1.5903891793300932e-05, "loss": 0.4107, "num_input_tokens_seen": 16393608, "step": 25060 }, { "epoch": 13.13679245283019, "grad_norm": 0.16925854980945587, "learning_rate": 1.5893242168650403e-05, "loss": 0.4327, "num_input_tokens_seen": 16397704, "step": 25065 }, { "epoch": 13.139412997903564, "grad_norm": 0.1621081680059433, "learning_rate": 1.5882594449029705e-05, "loss": 0.4735, "num_input_tokens_seen": 16402888, "step": 25070 }, { "epoch": 13.14203354297694, "grad_norm": 0.12274400144815445, "learning_rate": 1.587194863666624e-05, "loss": 0.4105, "num_input_tokens_seen": 16405480, "step": 25075 }, { "epoch": 13.144654088050315, "grad_norm": 0.24482043087482452, "learning_rate": 1.5861304733786974e-05, "loss": 0.5279, "num_input_tokens_seen": 16408648, "step": 25080 }, { "epoch": 13.14727463312369, "grad_norm": 0.12891288101673126, "learning_rate": 1.5850662742618505e-05, "loss": 0.4726, "num_input_tokens_seen": 16411688, "step": 25085 }, { "epoch": 13.149895178197065, "grad_norm": 0.2123415321111679, "learning_rate": 1.5840022665387004e-05, "loss": 0.512, "num_input_tokens_seen": 16414664, "step": 25090 }, { "epoch": 13.15251572327044, "grad_norm": 0.13825663924217224, "learning_rate": 1.5829384504318268e-05, "loss": 0.4296, "num_input_tokens_seen": 16417832, "step": 25095 }, { "epoch": 13.155136268343815, "grad_norm": 0.09523887187242508, "learning_rate": 1.5818748261637662e-05, "loss": 0.4606, "num_input_tokens_seen": 16421672, "step": 25100 }, { "epoch": 13.15775681341719, "grad_norm": 0.19032414257526398, "learning_rate": 1.5808113939570173e-05, "loss": 0.51, "num_input_tokens_seen": 16424520, "step": 25105 }, { "epoch": 13.160377358490566, "grad_norm": 0.21846790611743927, "learning_rate": 1.5797481540340386e-05, "loss": 0.5673, "num_input_tokens_seen": 16426760, "step": 25110 }, { "epoch": 13.16299790356394, "grad_norm": 0.23303347826004028, "learning_rate": 1.5786851066172466e-05, "loss": 0.4692, "num_input_tokens_seen": 16430536, "step": 25115 }, { "epoch": 13.165618448637316, "grad_norm": 0.1598367542028427, "learning_rate": 1.5776222519290208e-05, "loss": 0.4667, "num_input_tokens_seen": 16433384, "step": 25120 }, { "epoch": 13.168238993710691, "grad_norm": 0.15693329274654388, "learning_rate": 1.576559590191696e-05, "loss": 0.4668, "num_input_tokens_seen": 16436168, "step": 25125 }, { "epoch": 13.170859538784066, "grad_norm": 0.16417673230171204, "learning_rate": 1.5754971216275692e-05, "loss": 0.5255, "num_input_tokens_seen": 16438760, "step": 25130 }, { "epoch": 13.173480083857442, "grad_norm": 0.1333090364933014, "learning_rate": 1.5744348464588975e-05, "loss": 0.4983, "num_input_tokens_seen": 16441864, "step": 25135 }, { "epoch": 13.176100628930818, "grad_norm": 0.18216513097286224, "learning_rate": 1.5733727649078968e-05, "loss": 0.5048, "num_input_tokens_seen": 16445512, "step": 25140 }, { "epoch": 13.178721174004194, "grad_norm": 0.13881762325763702, "learning_rate": 1.5723108771967414e-05, "loss": 0.5135, "num_input_tokens_seen": 16448264, "step": 25145 }, { "epoch": 13.181341719077569, "grad_norm": 0.16405269503593445, "learning_rate": 1.5712491835475685e-05, "loss": 0.5403, "num_input_tokens_seen": 16451880, "step": 25150 }, { "epoch": 13.183962264150944, "grad_norm": 0.19977344572544098, "learning_rate": 1.5701876841824698e-05, "loss": 0.521, "num_input_tokens_seen": 16454728, "step": 25155 }, { "epoch": 13.18658280922432, "grad_norm": 0.14425259828567505, "learning_rate": 1.5691263793234993e-05, "loss": 0.3976, "num_input_tokens_seen": 16457480, "step": 25160 }, { "epoch": 13.189203354297694, "grad_norm": 0.13541674613952637, "learning_rate": 1.5680652691926706e-05, "loss": 0.5332, "num_input_tokens_seen": 16460232, "step": 25165 }, { "epoch": 13.19182389937107, "grad_norm": 0.16593030095100403, "learning_rate": 1.5670043540119553e-05, "loss": 0.3996, "num_input_tokens_seen": 16462856, "step": 25170 }, { "epoch": 13.194444444444445, "grad_norm": 0.15960940718650818, "learning_rate": 1.5659436340032855e-05, "loss": 0.3997, "num_input_tokens_seen": 16465640, "step": 25175 }, { "epoch": 13.19706498951782, "grad_norm": 0.15678851306438446, "learning_rate": 1.5648831093885515e-05, "loss": 0.41, "num_input_tokens_seen": 16468552, "step": 25180 }, { "epoch": 13.199685534591195, "grad_norm": 0.18309220671653748, "learning_rate": 1.5638227803896015e-05, "loss": 0.4418, "num_input_tokens_seen": 16471176, "step": 25185 }, { "epoch": 13.20230607966457, "grad_norm": 0.131991446018219, "learning_rate": 1.562762647228246e-05, "loss": 0.451, "num_input_tokens_seen": 16473896, "step": 25190 }, { "epoch": 13.204926624737945, "grad_norm": 0.07483674585819244, "learning_rate": 1.5617027101262515e-05, "loss": 0.4047, "num_input_tokens_seen": 16478600, "step": 25195 }, { "epoch": 13.20754716981132, "grad_norm": 0.24338217079639435, "learning_rate": 1.5606429693053463e-05, "loss": 0.4394, "num_input_tokens_seen": 16482152, "step": 25200 }, { "epoch": 13.210167714884696, "grad_norm": 0.15528236329555511, "learning_rate": 1.5595834249872145e-05, "loss": 0.5264, "num_input_tokens_seen": 16485128, "step": 25205 }, { "epoch": 13.21278825995807, "grad_norm": 0.20900540053844452, "learning_rate": 1.558524077393502e-05, "loss": 0.4697, "num_input_tokens_seen": 16488552, "step": 25210 }, { "epoch": 13.215408805031446, "grad_norm": 0.22037646174430847, "learning_rate": 1.557464926745811e-05, "loss": 0.4194, "num_input_tokens_seen": 16490984, "step": 25215 }, { "epoch": 13.218029350104821, "grad_norm": 0.25657936930656433, "learning_rate": 1.5564059732657045e-05, "loss": 0.5302, "num_input_tokens_seen": 16493928, "step": 25220 }, { "epoch": 13.220649895178196, "grad_norm": 0.15386617183685303, "learning_rate": 1.555347217174703e-05, "loss": 0.5893, "num_input_tokens_seen": 16496904, "step": 25225 }, { "epoch": 13.223270440251572, "grad_norm": 0.2243260145187378, "learning_rate": 1.554288658694287e-05, "loss": 0.4091, "num_input_tokens_seen": 16499464, "step": 25230 }, { "epoch": 13.225890985324948, "grad_norm": 0.14866039156913757, "learning_rate": 1.553230298045895e-05, "loss": 0.4496, "num_input_tokens_seen": 16502376, "step": 25235 }, { "epoch": 13.228511530398324, "grad_norm": 0.13937769830226898, "learning_rate": 1.5521721354509223e-05, "loss": 0.4522, "num_input_tokens_seen": 16505288, "step": 25240 }, { "epoch": 13.231132075471699, "grad_norm": 0.1457206755876541, "learning_rate": 1.5511141711307258e-05, "loss": 0.6387, "num_input_tokens_seen": 16507912, "step": 25245 }, { "epoch": 13.233752620545074, "grad_norm": 0.1618618369102478, "learning_rate": 1.5500564053066184e-05, "loss": 0.4946, "num_input_tokens_seen": 16510824, "step": 25250 }, { "epoch": 13.23637316561845, "grad_norm": 0.25686410069465637, "learning_rate": 1.5489988381998743e-05, "loss": 0.474, "num_input_tokens_seen": 16514408, "step": 25255 }, { "epoch": 13.238993710691824, "grad_norm": 0.17631842195987701, "learning_rate": 1.5479414700317237e-05, "loss": 0.4137, "num_input_tokens_seen": 16517480, "step": 25260 }, { "epoch": 13.2416142557652, "grad_norm": 0.14207735657691956, "learning_rate": 1.546884301023356e-05, "loss": 0.4107, "num_input_tokens_seen": 16520968, "step": 25265 }, { "epoch": 13.244234800838575, "grad_norm": 0.15313956141471863, "learning_rate": 1.5458273313959186e-05, "loss": 0.3948, "num_input_tokens_seen": 16523976, "step": 25270 }, { "epoch": 13.24685534591195, "grad_norm": 0.17677336931228638, "learning_rate": 1.5447705613705177e-05, "loss": 0.4675, "num_input_tokens_seen": 16527336, "step": 25275 }, { "epoch": 13.249475890985325, "grad_norm": 0.11959729343652725, "learning_rate": 1.5437139911682176e-05, "loss": 0.465, "num_input_tokens_seen": 16531368, "step": 25280 }, { "epoch": 13.2520964360587, "grad_norm": 0.09412724524736404, "learning_rate": 1.54265762101004e-05, "loss": 0.4888, "num_input_tokens_seen": 16535016, "step": 25285 }, { "epoch": 13.254716981132075, "grad_norm": 0.30841946601867676, "learning_rate": 1.5416014511169663e-05, "loss": 0.4879, "num_input_tokens_seen": 16539528, "step": 25290 }, { "epoch": 13.25733752620545, "grad_norm": 0.22477391362190247, "learning_rate": 1.5405454817099363e-05, "loss": 0.4972, "num_input_tokens_seen": 16543752, "step": 25295 }, { "epoch": 13.259958071278826, "grad_norm": 0.13571281731128693, "learning_rate": 1.5394897130098446e-05, "loss": 0.609, "num_input_tokens_seen": 16548040, "step": 25300 }, { "epoch": 13.2625786163522, "grad_norm": 0.15765489637851715, "learning_rate": 1.5384341452375465e-05, "loss": 0.3842, "num_input_tokens_seen": 16551496, "step": 25305 }, { "epoch": 13.265199161425576, "grad_norm": 0.274314820766449, "learning_rate": 1.5373787786138556e-05, "loss": 0.4859, "num_input_tokens_seen": 16554984, "step": 25310 }, { "epoch": 13.267819706498951, "grad_norm": 0.13145314157009125, "learning_rate": 1.5363236133595416e-05, "loss": 0.4708, "num_input_tokens_seen": 16558888, "step": 25315 }, { "epoch": 13.270440251572326, "grad_norm": 0.14354775846004486, "learning_rate": 1.5352686496953344e-05, "loss": 0.3868, "num_input_tokens_seen": 16562216, "step": 25320 }, { "epoch": 13.273060796645701, "grad_norm": 0.17076604068279266, "learning_rate": 1.534213887841919e-05, "loss": 0.5039, "num_input_tokens_seen": 16565320, "step": 25325 }, { "epoch": 13.275681341719078, "grad_norm": 0.14634065330028534, "learning_rate": 1.5331593280199395e-05, "loss": 0.5778, "num_input_tokens_seen": 16568072, "step": 25330 }, { "epoch": 13.278301886792454, "grad_norm": 0.15445010364055634, "learning_rate": 1.532104970449999e-05, "loss": 0.4402, "num_input_tokens_seen": 16571912, "step": 25335 }, { "epoch": 13.280922431865829, "grad_norm": 0.16267968714237213, "learning_rate": 1.5310508153526555e-05, "loss": 0.3917, "num_input_tokens_seen": 16577384, "step": 25340 }, { "epoch": 13.283542976939204, "grad_norm": 0.1527586579322815, "learning_rate": 1.5299968629484276e-05, "loss": 0.6011, "num_input_tokens_seen": 16581416, "step": 25345 }, { "epoch": 13.286163522012579, "grad_norm": 0.12896990776062012, "learning_rate": 1.5289431134577904e-05, "loss": 0.3929, "num_input_tokens_seen": 16583944, "step": 25350 }, { "epoch": 13.288784067085954, "grad_norm": 0.13544200360774994, "learning_rate": 1.527889567101174e-05, "loss": 0.4218, "num_input_tokens_seen": 16586888, "step": 25355 }, { "epoch": 13.29140461215933, "grad_norm": 0.11245688796043396, "learning_rate": 1.5268362240989707e-05, "loss": 0.4427, "num_input_tokens_seen": 16590376, "step": 25360 }, { "epoch": 13.294025157232705, "grad_norm": 0.13121873140335083, "learning_rate": 1.525783084671526e-05, "loss": 0.3945, "num_input_tokens_seen": 16593672, "step": 25365 }, { "epoch": 13.29664570230608, "grad_norm": 0.15990900993347168, "learning_rate": 1.5247301490391458e-05, "loss": 0.5534, "num_input_tokens_seen": 16596840, "step": 25370 }, { "epoch": 13.299266247379455, "grad_norm": 0.21737124025821686, "learning_rate": 1.5236774174220922e-05, "loss": 0.5642, "num_input_tokens_seen": 16599880, "step": 25375 }, { "epoch": 13.30188679245283, "grad_norm": 0.17971830070018768, "learning_rate": 1.5226248900405852e-05, "loss": 0.4673, "num_input_tokens_seen": 16602856, "step": 25380 }, { "epoch": 13.304507337526205, "grad_norm": 0.15958331525325775, "learning_rate": 1.5215725671147996e-05, "loss": 0.4903, "num_input_tokens_seen": 16606440, "step": 25385 }, { "epoch": 13.30712788259958, "grad_norm": 0.1632966250181198, "learning_rate": 1.5205204488648709e-05, "loss": 0.4533, "num_input_tokens_seen": 16610248, "step": 25390 }, { "epoch": 13.309748427672956, "grad_norm": 0.2718751132488251, "learning_rate": 1.5194685355108893e-05, "loss": 0.5322, "num_input_tokens_seen": 16612584, "step": 25395 }, { "epoch": 13.31236897274633, "grad_norm": 0.0970543846487999, "learning_rate": 1.518416827272904e-05, "loss": 0.4571, "num_input_tokens_seen": 16616168, "step": 25400 }, { "epoch": 13.314989517819706, "grad_norm": 0.18560703098773956, "learning_rate": 1.5173653243709207e-05, "loss": 0.5256, "num_input_tokens_seen": 16618984, "step": 25405 }, { "epoch": 13.317610062893081, "grad_norm": 0.15239395201206207, "learning_rate": 1.5163140270249e-05, "loss": 0.5262, "num_input_tokens_seen": 16623208, "step": 25410 }, { "epoch": 13.320230607966456, "grad_norm": 0.15201418101787567, "learning_rate": 1.5152629354547626e-05, "loss": 0.4273, "num_input_tokens_seen": 16627656, "step": 25415 }, { "epoch": 13.322851153039831, "grad_norm": 0.16122153401374817, "learning_rate": 1.5142120498803842e-05, "loss": 0.5342, "num_input_tokens_seen": 16630280, "step": 25420 }, { "epoch": 13.325471698113208, "grad_norm": 0.14074298739433289, "learning_rate": 1.5131613705215991e-05, "loss": 0.4539, "num_input_tokens_seen": 16632904, "step": 25425 }, { "epoch": 13.328092243186584, "grad_norm": 0.15298475325107574, "learning_rate": 1.5121108975981965e-05, "loss": 0.5475, "num_input_tokens_seen": 16635624, "step": 25430 }, { "epoch": 13.330712788259959, "grad_norm": 0.30907243490219116, "learning_rate": 1.5110606313299248e-05, "loss": 0.5534, "num_input_tokens_seen": 16638760, "step": 25435 }, { "epoch": 13.333333333333334, "grad_norm": 0.3377588391304016, "learning_rate": 1.5100105719364855e-05, "loss": 0.5871, "num_input_tokens_seen": 16641448, "step": 25440 }, { "epoch": 13.335953878406709, "grad_norm": 0.1637287735939026, "learning_rate": 1.50896071963754e-05, "loss": 0.4114, "num_input_tokens_seen": 16644520, "step": 25445 }, { "epoch": 13.338574423480084, "grad_norm": 0.10935789346694946, "learning_rate": 1.5079110746527055e-05, "loss": 0.4851, "num_input_tokens_seen": 16648584, "step": 25450 }, { "epoch": 13.34119496855346, "grad_norm": 0.15136775374412537, "learning_rate": 1.5068616372015565e-05, "loss": 0.495, "num_input_tokens_seen": 16651784, "step": 25455 }, { "epoch": 13.343815513626835, "grad_norm": 0.15917347371578217, "learning_rate": 1.5058124075036223e-05, "loss": 0.4382, "num_input_tokens_seen": 16654024, "step": 25460 }, { "epoch": 13.34643605870021, "grad_norm": 0.14298467338085175, "learning_rate": 1.5047633857783908e-05, "loss": 0.5003, "num_input_tokens_seen": 16657480, "step": 25465 }, { "epoch": 13.349056603773585, "grad_norm": 0.23316442966461182, "learning_rate": 1.5037145722453044e-05, "loss": 0.5272, "num_input_tokens_seen": 16661096, "step": 25470 }, { "epoch": 13.35167714884696, "grad_norm": 0.17774541676044464, "learning_rate": 1.5026659671237624e-05, "loss": 0.4013, "num_input_tokens_seen": 16663688, "step": 25475 }, { "epoch": 13.354297693920335, "grad_norm": 0.21350312232971191, "learning_rate": 1.501617570633123e-05, "loss": 0.5045, "num_input_tokens_seen": 16666024, "step": 25480 }, { "epoch": 13.35691823899371, "grad_norm": 0.11619263142347336, "learning_rate": 1.500569382992697e-05, "loss": 0.471, "num_input_tokens_seen": 16669608, "step": 25485 }, { "epoch": 13.359538784067086, "grad_norm": 0.2930254340171814, "learning_rate": 1.4995214044217547e-05, "loss": 0.5518, "num_input_tokens_seen": 16673480, "step": 25490 }, { "epoch": 13.36215932914046, "grad_norm": 0.12466705590486526, "learning_rate": 1.49847363513952e-05, "loss": 0.3309, "num_input_tokens_seen": 16677704, "step": 25495 }, { "epoch": 13.364779874213836, "grad_norm": 0.17052340507507324, "learning_rate": 1.4974260753651748e-05, "loss": 0.5739, "num_input_tokens_seen": 16680744, "step": 25500 }, { "epoch": 13.367400419287211, "grad_norm": 0.3015819191932678, "learning_rate": 1.4963787253178568e-05, "loss": 0.5425, "num_input_tokens_seen": 16683048, "step": 25505 }, { "epoch": 13.370020964360586, "grad_norm": 0.17022360861301422, "learning_rate": 1.4953315852166594e-05, "loss": 0.4472, "num_input_tokens_seen": 16685832, "step": 25510 }, { "epoch": 13.372641509433961, "grad_norm": 0.11459681391716003, "learning_rate": 1.4942846552806322e-05, "loss": 0.5326, "num_input_tokens_seen": 16688168, "step": 25515 }, { "epoch": 13.375262054507338, "grad_norm": 0.12159214913845062, "learning_rate": 1.4932379357287829e-05, "loss": 0.4603, "num_input_tokens_seen": 16690888, "step": 25520 }, { "epoch": 13.377882599580714, "grad_norm": 0.14898911118507385, "learning_rate": 1.49219142678007e-05, "loss": 0.4578, "num_input_tokens_seen": 16694280, "step": 25525 }, { "epoch": 13.380503144654089, "grad_norm": 0.170633465051651, "learning_rate": 1.4911451286534128e-05, "loss": 0.5112, "num_input_tokens_seen": 16697128, "step": 25530 }, { "epoch": 13.383123689727464, "grad_norm": 0.12593792378902435, "learning_rate": 1.490099041567686e-05, "loss": 0.4459, "num_input_tokens_seen": 16701192, "step": 25535 }, { "epoch": 13.385744234800839, "grad_norm": 0.16746100783348083, "learning_rate": 1.4890531657417176e-05, "loss": 0.4377, "num_input_tokens_seen": 16706280, "step": 25540 }, { "epoch": 13.388364779874214, "grad_norm": 0.22799813747406006, "learning_rate": 1.4880075013942938e-05, "loss": 0.3884, "num_input_tokens_seen": 16709448, "step": 25545 }, { "epoch": 13.39098532494759, "grad_norm": 0.10098676383495331, "learning_rate": 1.4869620487441558e-05, "loss": 0.6058, "num_input_tokens_seen": 16713256, "step": 25550 }, { "epoch": 13.393605870020965, "grad_norm": 0.18799985945224762, "learning_rate": 1.4859168080099988e-05, "loss": 0.5205, "num_input_tokens_seen": 16716296, "step": 25555 }, { "epoch": 13.39622641509434, "grad_norm": 0.1758178323507309, "learning_rate": 1.484871779410477e-05, "loss": 0.4826, "num_input_tokens_seen": 16719432, "step": 25560 }, { "epoch": 13.398846960167715, "grad_norm": 0.16322368383407593, "learning_rate": 1.4838269631641971e-05, "loss": 0.3713, "num_input_tokens_seen": 16722536, "step": 25565 }, { "epoch": 13.40146750524109, "grad_norm": 0.10516832023859024, "learning_rate": 1.482782359489724e-05, "loss": 0.5973, "num_input_tokens_seen": 16725704, "step": 25570 }, { "epoch": 13.404088050314465, "grad_norm": 0.15982036292552948, "learning_rate": 1.4817379686055754e-05, "loss": 0.515, "num_input_tokens_seen": 16729192, "step": 25575 }, { "epoch": 13.40670859538784, "grad_norm": 0.1429257094860077, "learning_rate": 1.4806937907302282e-05, "loss": 0.457, "num_input_tokens_seen": 16732168, "step": 25580 }, { "epoch": 13.409329140461216, "grad_norm": 0.1522061973810196, "learning_rate": 1.4796498260821103e-05, "loss": 0.4513, "num_input_tokens_seen": 16735496, "step": 25585 }, { "epoch": 13.41194968553459, "grad_norm": 0.16094522178173065, "learning_rate": 1.4786060748796076e-05, "loss": 0.5503, "num_input_tokens_seen": 16739176, "step": 25590 }, { "epoch": 13.414570230607966, "grad_norm": 0.1706455796957016, "learning_rate": 1.4775625373410617e-05, "loss": 0.4161, "num_input_tokens_seen": 16742344, "step": 25595 }, { "epoch": 13.417190775681341, "grad_norm": 0.130781352519989, "learning_rate": 1.4765192136847685e-05, "loss": 0.3811, "num_input_tokens_seen": 16745512, "step": 25600 }, { "epoch": 13.419811320754716, "grad_norm": 0.19151705503463745, "learning_rate": 1.47547610412898e-05, "loss": 0.4739, "num_input_tokens_seen": 16748584, "step": 25605 }, { "epoch": 13.422431865828091, "grad_norm": 0.09894141554832458, "learning_rate": 1.4744332088919011e-05, "loss": 0.5242, "num_input_tokens_seen": 16752168, "step": 25610 }, { "epoch": 13.425052410901468, "grad_norm": 0.10022597014904022, "learning_rate": 1.4733905281916954e-05, "loss": 0.4835, "num_input_tokens_seen": 16755688, "step": 25615 }, { "epoch": 13.427672955974844, "grad_norm": 0.11423389613628387, "learning_rate": 1.472348062246478e-05, "loss": 0.3839, "num_input_tokens_seen": 16759016, "step": 25620 }, { "epoch": 13.430293501048219, "grad_norm": 0.22573822736740112, "learning_rate": 1.4713058112743228e-05, "loss": 0.5785, "num_input_tokens_seen": 16762600, "step": 25625 }, { "epoch": 13.432914046121594, "grad_norm": 0.24840958416461945, "learning_rate": 1.4702637754932553e-05, "loss": 0.509, "num_input_tokens_seen": 16765768, "step": 25630 }, { "epoch": 13.435534591194969, "grad_norm": 0.21101097762584686, "learning_rate": 1.4692219551212587e-05, "loss": 0.5526, "num_input_tokens_seen": 16768584, "step": 25635 }, { "epoch": 13.438155136268344, "grad_norm": 0.25086671113967896, "learning_rate": 1.4681803503762695e-05, "loss": 0.4981, "num_input_tokens_seen": 16771240, "step": 25640 }, { "epoch": 13.44077568134172, "grad_norm": 0.22947560250759125, "learning_rate": 1.4671389614761788e-05, "loss": 0.4196, "num_input_tokens_seen": 16774792, "step": 25645 }, { "epoch": 13.443396226415095, "grad_norm": 0.17362813651561737, "learning_rate": 1.4660977886388339e-05, "loss": 0.6261, "num_input_tokens_seen": 16777864, "step": 25650 }, { "epoch": 13.44601677148847, "grad_norm": 0.10056629031896591, "learning_rate": 1.4650568320820362e-05, "loss": 0.3884, "num_input_tokens_seen": 16781768, "step": 25655 }, { "epoch": 13.448637316561845, "grad_norm": 0.16308419406414032, "learning_rate": 1.4640160920235423e-05, "loss": 0.4293, "num_input_tokens_seen": 16784648, "step": 25660 }, { "epoch": 13.45125786163522, "grad_norm": 0.1379965990781784, "learning_rate": 1.4629755686810631e-05, "loss": 0.5061, "num_input_tokens_seen": 16787368, "step": 25665 }, { "epoch": 13.453878406708595, "grad_norm": 0.14579634368419647, "learning_rate": 1.4619352622722632e-05, "loss": 0.4963, "num_input_tokens_seen": 16791048, "step": 25670 }, { "epoch": 13.45649895178197, "grad_norm": 0.1741400957107544, "learning_rate": 1.4608951730147635e-05, "loss": 0.5187, "num_input_tokens_seen": 16794504, "step": 25675 }, { "epoch": 13.459119496855346, "grad_norm": 0.08887236565351486, "learning_rate": 1.4598553011261395e-05, "loss": 0.3783, "num_input_tokens_seen": 16797960, "step": 25680 }, { "epoch": 13.46174004192872, "grad_norm": 0.15381723642349243, "learning_rate": 1.4588156468239192e-05, "loss": 0.4564, "num_input_tokens_seen": 16800456, "step": 25685 }, { "epoch": 13.464360587002096, "grad_norm": 0.20733986794948578, "learning_rate": 1.4577762103255882e-05, "loss": 0.4883, "num_input_tokens_seen": 16803048, "step": 25690 }, { "epoch": 13.466981132075471, "grad_norm": 0.1892932951450348, "learning_rate": 1.4567369918485828e-05, "loss": 0.6303, "num_input_tokens_seen": 16807272, "step": 25695 }, { "epoch": 13.469601677148846, "grad_norm": 0.12241131067276001, "learning_rate": 1.4556979916102976e-05, "loss": 0.3434, "num_input_tokens_seen": 16810792, "step": 25700 }, { "epoch": 13.472222222222221, "grad_norm": 0.13281656801700592, "learning_rate": 1.454659209828077e-05, "loss": 0.4921, "num_input_tokens_seen": 16814024, "step": 25705 }, { "epoch": 13.474842767295598, "grad_norm": 0.10373878479003906, "learning_rate": 1.4536206467192243e-05, "loss": 0.5031, "num_input_tokens_seen": 16817928, "step": 25710 }, { "epoch": 13.477463312368974, "grad_norm": 0.12395031005144119, "learning_rate": 1.4525823025009944e-05, "loss": 0.3972, "num_input_tokens_seen": 16820680, "step": 25715 }, { "epoch": 13.480083857442349, "grad_norm": 0.10663319379091263, "learning_rate": 1.4515441773905983e-05, "loss": 0.4333, "num_input_tokens_seen": 16824328, "step": 25720 }, { "epoch": 13.482704402515724, "grad_norm": 0.3018397390842438, "learning_rate": 1.4505062716051988e-05, "loss": 0.4998, "num_input_tokens_seen": 16827880, "step": 25725 }, { "epoch": 13.485324947589099, "grad_norm": 0.1921813040971756, "learning_rate": 1.4494685853619133e-05, "loss": 0.6843, "num_input_tokens_seen": 16830984, "step": 25730 }, { "epoch": 13.487945492662474, "grad_norm": 0.14386054873466492, "learning_rate": 1.4484311188778143e-05, "loss": 0.6235, "num_input_tokens_seen": 16835144, "step": 25735 }, { "epoch": 13.49056603773585, "grad_norm": 0.2217826247215271, "learning_rate": 1.4473938723699285e-05, "loss": 0.5094, "num_input_tokens_seen": 16838152, "step": 25740 }, { "epoch": 13.493186582809225, "grad_norm": 0.15540039539337158, "learning_rate": 1.4463568460552368e-05, "loss": 0.372, "num_input_tokens_seen": 16841480, "step": 25745 }, { "epoch": 13.4958071278826, "grad_norm": 0.12710057199001312, "learning_rate": 1.4453200401506723e-05, "loss": 0.4471, "num_input_tokens_seen": 16845288, "step": 25750 }, { "epoch": 13.498427672955975, "grad_norm": 0.2461223602294922, "learning_rate": 1.4442834548731218e-05, "loss": 0.6024, "num_input_tokens_seen": 16849096, "step": 25755 }, { "epoch": 13.50104821802935, "grad_norm": 0.17139790952205658, "learning_rate": 1.4432470904394285e-05, "loss": 0.5422, "num_input_tokens_seen": 16852232, "step": 25760 }, { "epoch": 13.503668763102725, "grad_norm": 0.38866180181503296, "learning_rate": 1.4422109470663878e-05, "loss": 0.6682, "num_input_tokens_seen": 16855176, "step": 25765 }, { "epoch": 13.5062893081761, "grad_norm": 0.14058370888233185, "learning_rate": 1.441175024970749e-05, "loss": 0.3602, "num_input_tokens_seen": 16858344, "step": 25770 }, { "epoch": 13.508909853249476, "grad_norm": 0.17326010763645172, "learning_rate": 1.4401393243692163e-05, "loss": 0.5731, "num_input_tokens_seen": 16861896, "step": 25775 }, { "epoch": 13.51153039832285, "grad_norm": 0.12765486538410187, "learning_rate": 1.4391038454784455e-05, "loss": 0.4756, "num_input_tokens_seen": 16865704, "step": 25780 }, { "epoch": 13.514150943396226, "grad_norm": 0.1463451385498047, "learning_rate": 1.4380685885150458e-05, "loss": 0.5728, "num_input_tokens_seen": 16868616, "step": 25785 }, { "epoch": 13.516771488469601, "grad_norm": 0.10169795900583267, "learning_rate": 1.437033553695582e-05, "loss": 0.4044, "num_input_tokens_seen": 16871816, "step": 25790 }, { "epoch": 13.519392033542976, "grad_norm": 0.23909702897071838, "learning_rate": 1.4359987412365721e-05, "loss": 0.4725, "num_input_tokens_seen": 16875016, "step": 25795 }, { "epoch": 13.522012578616351, "grad_norm": 0.1960229128599167, "learning_rate": 1.4349641513544876e-05, "loss": 0.5676, "num_input_tokens_seen": 16877928, "step": 25800 }, { "epoch": 13.524633123689728, "grad_norm": 0.1643824577331543, "learning_rate": 1.4339297842657523e-05, "loss": 0.419, "num_input_tokens_seen": 16881896, "step": 25805 }, { "epoch": 13.527253668763104, "grad_norm": 0.21644365787506104, "learning_rate": 1.4328956401867427e-05, "loss": 0.5041, "num_input_tokens_seen": 16884744, "step": 25810 }, { "epoch": 13.529874213836479, "grad_norm": 0.19535945355892181, "learning_rate": 1.4318617193337908e-05, "loss": 0.2685, "num_input_tokens_seen": 16890376, "step": 25815 }, { "epoch": 13.532494758909854, "grad_norm": 0.16188888251781464, "learning_rate": 1.4308280219231812e-05, "loss": 0.3956, "num_input_tokens_seen": 16892840, "step": 25820 }, { "epoch": 13.535115303983229, "grad_norm": 0.2348831593990326, "learning_rate": 1.4297945481711527e-05, "loss": 0.6045, "num_input_tokens_seen": 16896552, "step": 25825 }, { "epoch": 13.537735849056604, "grad_norm": 0.42342421412467957, "learning_rate": 1.4287612982938946e-05, "loss": 0.4621, "num_input_tokens_seen": 16899112, "step": 25830 }, { "epoch": 13.54035639412998, "grad_norm": 0.3372357487678528, "learning_rate": 1.4277282725075523e-05, "loss": 0.3065, "num_input_tokens_seen": 16902536, "step": 25835 }, { "epoch": 13.542976939203355, "grad_norm": 0.14643721282482147, "learning_rate": 1.426695471028221e-05, "loss": 0.5201, "num_input_tokens_seen": 16905576, "step": 25840 }, { "epoch": 13.54559748427673, "grad_norm": 0.14177779853343964, "learning_rate": 1.4256628940719524e-05, "loss": 0.63, "num_input_tokens_seen": 16909128, "step": 25845 }, { "epoch": 13.548218029350105, "grad_norm": 0.2443406879901886, "learning_rate": 1.42463054185475e-05, "loss": 0.4576, "num_input_tokens_seen": 16911592, "step": 25850 }, { "epoch": 13.55083857442348, "grad_norm": 0.28767508268356323, "learning_rate": 1.4235984145925705e-05, "loss": 0.5669, "num_input_tokens_seen": 16914536, "step": 25855 }, { "epoch": 13.553459119496855, "grad_norm": 0.1982164829969406, "learning_rate": 1.4225665125013224e-05, "loss": 0.6017, "num_input_tokens_seen": 16920936, "step": 25860 }, { "epoch": 13.55607966457023, "grad_norm": 0.17475199699401855, "learning_rate": 1.4215348357968669e-05, "loss": 0.4756, "num_input_tokens_seen": 16923464, "step": 25865 }, { "epoch": 13.558700209643606, "grad_norm": 0.14570246636867523, "learning_rate": 1.42050338469502e-05, "loss": 0.5016, "num_input_tokens_seen": 16926152, "step": 25870 }, { "epoch": 13.56132075471698, "grad_norm": 0.1447952836751938, "learning_rate": 1.4194721594115495e-05, "loss": 0.4323, "num_input_tokens_seen": 16929928, "step": 25875 }, { "epoch": 13.563941299790356, "grad_norm": 0.15875069797039032, "learning_rate": 1.4184411601621769e-05, "loss": 0.5634, "num_input_tokens_seen": 16933864, "step": 25880 }, { "epoch": 13.566561844863731, "grad_norm": 0.32425346970558167, "learning_rate": 1.4174103871625738e-05, "loss": 0.4104, "num_input_tokens_seen": 16936392, "step": 25885 }, { "epoch": 13.569182389937106, "grad_norm": 0.11312877386808395, "learning_rate": 1.4163798406283674e-05, "loss": 0.4906, "num_input_tokens_seen": 16940008, "step": 25890 }, { "epoch": 13.571802935010481, "grad_norm": 0.12762108445167542, "learning_rate": 1.415349520775135e-05, "loss": 0.4666, "num_input_tokens_seen": 16943176, "step": 25895 }, { "epoch": 13.574423480083858, "grad_norm": 0.1822395771741867, "learning_rate": 1.4143194278184086e-05, "loss": 0.5915, "num_input_tokens_seen": 16945992, "step": 25900 }, { "epoch": 13.577044025157234, "grad_norm": 0.11193490773439407, "learning_rate": 1.4132895619736724e-05, "loss": 0.6342, "num_input_tokens_seen": 16949352, "step": 25905 }, { "epoch": 13.579664570230609, "grad_norm": 0.1340796947479248, "learning_rate": 1.4122599234563613e-05, "loss": 0.5507, "num_input_tokens_seen": 16953000, "step": 25910 }, { "epoch": 13.582285115303984, "grad_norm": 0.10002245754003525, "learning_rate": 1.4112305124818647e-05, "loss": 0.4524, "num_input_tokens_seen": 16956232, "step": 25915 }, { "epoch": 13.584905660377359, "grad_norm": 0.16144110262393951, "learning_rate": 1.4102013292655242e-05, "loss": 0.5907, "num_input_tokens_seen": 16959464, "step": 25920 }, { "epoch": 13.587526205450734, "grad_norm": 0.2184913456439972, "learning_rate": 1.409172374022632e-05, "loss": 0.5233, "num_input_tokens_seen": 16962440, "step": 25925 }, { "epoch": 13.59014675052411, "grad_norm": 0.14896276593208313, "learning_rate": 1.4081436469684339e-05, "loss": 0.4776, "num_input_tokens_seen": 16965608, "step": 25930 }, { "epoch": 13.592767295597485, "grad_norm": 0.14108796417713165, "learning_rate": 1.4071151483181294e-05, "loss": 0.4791, "num_input_tokens_seen": 16969864, "step": 25935 }, { "epoch": 13.59538784067086, "grad_norm": 0.19485756754875183, "learning_rate": 1.4060868782868669e-05, "loss": 0.5164, "num_input_tokens_seen": 16972840, "step": 25940 }, { "epoch": 13.598008385744235, "grad_norm": 0.20099322497844696, "learning_rate": 1.405058837089749e-05, "loss": 0.5572, "num_input_tokens_seen": 16975816, "step": 25945 }, { "epoch": 13.60062893081761, "grad_norm": 0.15168426930904388, "learning_rate": 1.404031024941832e-05, "loss": 0.4171, "num_input_tokens_seen": 16979016, "step": 25950 }, { "epoch": 13.603249475890985, "grad_norm": 0.15440471470355988, "learning_rate": 1.40300344205812e-05, "loss": 0.4398, "num_input_tokens_seen": 16982056, "step": 25955 }, { "epoch": 13.60587002096436, "grad_norm": 0.22254128754138947, "learning_rate": 1.4019760886535738e-05, "loss": 0.4311, "num_input_tokens_seen": 16985032, "step": 25960 }, { "epoch": 13.608490566037736, "grad_norm": 0.16705256700515747, "learning_rate": 1.4009489649431018e-05, "loss": 0.5342, "num_input_tokens_seen": 16988008, "step": 25965 }, { "epoch": 13.61111111111111, "grad_norm": 0.14060494303703308, "learning_rate": 1.3999220711415675e-05, "loss": 0.5036, "num_input_tokens_seen": 16990696, "step": 25970 }, { "epoch": 13.613731656184486, "grad_norm": 0.13886329531669617, "learning_rate": 1.3988954074637867e-05, "loss": 0.5096, "num_input_tokens_seen": 16994088, "step": 25975 }, { "epoch": 13.616352201257861, "grad_norm": 0.13993747532367706, "learning_rate": 1.3978689741245238e-05, "loss": 0.5142, "num_input_tokens_seen": 16997000, "step": 25980 }, { "epoch": 13.618972746331236, "grad_norm": 0.20384258031845093, "learning_rate": 1.3968427713384986e-05, "loss": 0.5028, "num_input_tokens_seen": 16999912, "step": 25985 }, { "epoch": 13.621593291404611, "grad_norm": 0.17930233478546143, "learning_rate": 1.3958167993203791e-05, "loss": 0.513, "num_input_tokens_seen": 17003368, "step": 25990 }, { "epoch": 13.624213836477988, "grad_norm": 0.11752685904502869, "learning_rate": 1.3947910582847879e-05, "loss": 0.47, "num_input_tokens_seen": 17007336, "step": 25995 }, { "epoch": 13.626834381551364, "grad_norm": 0.18167687952518463, "learning_rate": 1.3937655484462985e-05, "loss": 0.4161, "num_input_tokens_seen": 17010056, "step": 26000 }, { "epoch": 13.629454926624739, "grad_norm": 0.20016075670719147, "learning_rate": 1.392740270019437e-05, "loss": 0.5048, "num_input_tokens_seen": 17012776, "step": 26005 }, { "epoch": 13.632075471698114, "grad_norm": 0.1234753355383873, "learning_rate": 1.3917152232186776e-05, "loss": 0.4268, "num_input_tokens_seen": 17016872, "step": 26010 }, { "epoch": 13.634696016771489, "grad_norm": 0.15329919755458832, "learning_rate": 1.3906904082584504e-05, "loss": 0.6125, "num_input_tokens_seen": 17020680, "step": 26015 }, { "epoch": 13.637316561844864, "grad_norm": 0.11262819916009903, "learning_rate": 1.3896658253531334e-05, "loss": 0.4613, "num_input_tokens_seen": 17024168, "step": 26020 }, { "epoch": 13.63993710691824, "grad_norm": 0.15743714570999146, "learning_rate": 1.3886414747170581e-05, "loss": 0.5445, "num_input_tokens_seen": 17027304, "step": 26025 }, { "epoch": 13.642557651991615, "grad_norm": 0.16307640075683594, "learning_rate": 1.3876173565645072e-05, "loss": 0.4106, "num_input_tokens_seen": 17030344, "step": 26030 }, { "epoch": 13.64517819706499, "grad_norm": 0.1522049605846405, "learning_rate": 1.3865934711097156e-05, "loss": 0.4861, "num_input_tokens_seen": 17034504, "step": 26035 }, { "epoch": 13.647798742138365, "grad_norm": 0.26864954829216003, "learning_rate": 1.3855698185668675e-05, "loss": 0.5383, "num_input_tokens_seen": 17038120, "step": 26040 }, { "epoch": 13.65041928721174, "grad_norm": 0.1734796017408371, "learning_rate": 1.3845463991500985e-05, "loss": 0.451, "num_input_tokens_seen": 17041128, "step": 26045 }, { "epoch": 13.653039832285115, "grad_norm": 0.12599454820156097, "learning_rate": 1.3835232130734968e-05, "loss": 0.8698, "num_input_tokens_seen": 17043464, "step": 26050 }, { "epoch": 13.65566037735849, "grad_norm": 0.16773243248462677, "learning_rate": 1.3825002605511017e-05, "loss": 0.6183, "num_input_tokens_seen": 17045960, "step": 26055 }, { "epoch": 13.658280922431866, "grad_norm": 0.20382414758205414, "learning_rate": 1.3814775417969037e-05, "loss": 0.5402, "num_input_tokens_seen": 17050280, "step": 26060 }, { "epoch": 13.66090146750524, "grad_norm": 0.16188621520996094, "learning_rate": 1.3804550570248431e-05, "loss": 0.5896, "num_input_tokens_seen": 17053352, "step": 26065 }, { "epoch": 13.663522012578616, "grad_norm": 0.16278275847434998, "learning_rate": 1.3794328064488116e-05, "loss": 0.602, "num_input_tokens_seen": 17055880, "step": 26070 }, { "epoch": 13.666142557651991, "grad_norm": 0.4636220633983612, "learning_rate": 1.3784107902826523e-05, "loss": 0.463, "num_input_tokens_seen": 17059368, "step": 26075 }, { "epoch": 13.668763102725366, "grad_norm": 0.13720083236694336, "learning_rate": 1.3773890087401603e-05, "loss": 0.4923, "num_input_tokens_seen": 17062184, "step": 26080 }, { "epoch": 13.671383647798741, "grad_norm": 0.14243602752685547, "learning_rate": 1.3763674620350802e-05, "loss": 0.5323, "num_input_tokens_seen": 17066504, "step": 26085 }, { "epoch": 13.674004192872118, "grad_norm": 0.17162643373012543, "learning_rate": 1.3753461503811089e-05, "loss": 0.5137, "num_input_tokens_seen": 17069512, "step": 26090 }, { "epoch": 13.676624737945493, "grad_norm": 0.15715433657169342, "learning_rate": 1.3743250739918922e-05, "loss": 0.5599, "num_input_tokens_seen": 17072040, "step": 26095 }, { "epoch": 13.679245283018869, "grad_norm": 0.17295020818710327, "learning_rate": 1.3733042330810265e-05, "loss": 0.4076, "num_input_tokens_seen": 17074536, "step": 26100 }, { "epoch": 13.681865828092244, "grad_norm": 0.13347753882408142, "learning_rate": 1.3722836278620616e-05, "loss": 0.3566, "num_input_tokens_seen": 17080168, "step": 26105 }, { "epoch": 13.684486373165619, "grad_norm": 0.23678332567214966, "learning_rate": 1.3712632585484957e-05, "loss": 0.4793, "num_input_tokens_seen": 17083208, "step": 26110 }, { "epoch": 13.687106918238994, "grad_norm": 0.19691051542758942, "learning_rate": 1.37024312535378e-05, "loss": 0.3936, "num_input_tokens_seen": 17086632, "step": 26115 }, { "epoch": 13.68972746331237, "grad_norm": 0.21851927042007446, "learning_rate": 1.3692232284913133e-05, "loss": 0.5482, "num_input_tokens_seen": 17090952, "step": 26120 }, { "epoch": 13.692348008385745, "grad_norm": 0.1479686051607132, "learning_rate": 1.3682035681744459e-05, "loss": 0.4912, "num_input_tokens_seen": 17093416, "step": 26125 }, { "epoch": 13.69496855345912, "grad_norm": 0.16144537925720215, "learning_rate": 1.3671841446164795e-05, "loss": 0.4727, "num_input_tokens_seen": 17097160, "step": 26130 }, { "epoch": 13.697589098532495, "grad_norm": 0.10727754980325699, "learning_rate": 1.3661649580306669e-05, "loss": 0.3764, "num_input_tokens_seen": 17100680, "step": 26135 }, { "epoch": 13.70020964360587, "grad_norm": 0.10671450197696686, "learning_rate": 1.3651460086302087e-05, "loss": 0.4281, "num_input_tokens_seen": 17104456, "step": 26140 }, { "epoch": 13.702830188679245, "grad_norm": 0.18502791225910187, "learning_rate": 1.3641272966282609e-05, "loss": 0.3853, "num_input_tokens_seen": 17107080, "step": 26145 }, { "epoch": 13.70545073375262, "grad_norm": 0.36871403455734253, "learning_rate": 1.3631088222379218e-05, "loss": 0.4915, "num_input_tokens_seen": 17113064, "step": 26150 }, { "epoch": 13.708071278825996, "grad_norm": 0.16515149176120758, "learning_rate": 1.3620905856722468e-05, "loss": 0.5161, "num_input_tokens_seen": 17116072, "step": 26155 }, { "epoch": 13.71069182389937, "grad_norm": 0.13725802302360535, "learning_rate": 1.3610725871442392e-05, "loss": 0.3958, "num_input_tokens_seen": 17119720, "step": 26160 }, { "epoch": 13.713312368972746, "grad_norm": 0.46176350116729736, "learning_rate": 1.3600548268668532e-05, "loss": 0.5113, "num_input_tokens_seen": 17122312, "step": 26165 }, { "epoch": 13.715932914046121, "grad_norm": 0.2594357132911682, "learning_rate": 1.3590373050529927e-05, "loss": 0.524, "num_input_tokens_seen": 17125544, "step": 26170 }, { "epoch": 13.718553459119496, "grad_norm": 0.13783466815948486, "learning_rate": 1.3580200219155116e-05, "loss": 0.5293, "num_input_tokens_seen": 17128968, "step": 26175 }, { "epoch": 13.721174004192871, "grad_norm": 0.0902448296546936, "learning_rate": 1.3570029776672122e-05, "loss": 0.5196, "num_input_tokens_seen": 17132712, "step": 26180 }, { "epoch": 13.723794549266248, "grad_norm": 0.16131289303302765, "learning_rate": 1.35598617252085e-05, "loss": 0.4452, "num_input_tokens_seen": 17136104, "step": 26185 }, { "epoch": 13.726415094339622, "grad_norm": 0.1152549609541893, "learning_rate": 1.354969606689129e-05, "loss": 0.3305, "num_input_tokens_seen": 17139240, "step": 26190 }, { "epoch": 13.729035639412999, "grad_norm": 0.1440429985523224, "learning_rate": 1.353953280384704e-05, "loss": 0.4852, "num_input_tokens_seen": 17142344, "step": 26195 }, { "epoch": 13.731656184486374, "grad_norm": 0.15379765629768372, "learning_rate": 1.352937193820177e-05, "loss": 0.5588, "num_input_tokens_seen": 17145992, "step": 26200 }, { "epoch": 13.734276729559749, "grad_norm": 0.1287083625793457, "learning_rate": 1.3519213472081038e-05, "loss": 0.5212, "num_input_tokens_seen": 17149544, "step": 26205 }, { "epoch": 13.736897274633124, "grad_norm": 0.14889799058437347, "learning_rate": 1.3509057407609854e-05, "loss": 0.4921, "num_input_tokens_seen": 17152616, "step": 26210 }, { "epoch": 13.7395178197065, "grad_norm": 0.1942049264907837, "learning_rate": 1.3498903746912769e-05, "loss": 0.6277, "num_input_tokens_seen": 17154888, "step": 26215 }, { "epoch": 13.742138364779874, "grad_norm": 0.22397659718990326, "learning_rate": 1.3488752492113807e-05, "loss": 0.4255, "num_input_tokens_seen": 17157960, "step": 26220 }, { "epoch": 13.74475890985325, "grad_norm": 0.12917017936706543, "learning_rate": 1.3478603645336508e-05, "loss": 0.5267, "num_input_tokens_seen": 17160936, "step": 26225 }, { "epoch": 13.747379454926625, "grad_norm": 0.2410639077425003, "learning_rate": 1.3468457208703877e-05, "loss": 0.3925, "num_input_tokens_seen": 17164136, "step": 26230 }, { "epoch": 13.75, "grad_norm": 0.17341139912605286, "learning_rate": 1.3458313184338437e-05, "loss": 0.5278, "num_input_tokens_seen": 17167176, "step": 26235 }, { "epoch": 13.752620545073375, "grad_norm": 0.10967779159545898, "learning_rate": 1.3448171574362201e-05, "loss": 0.5364, "num_input_tokens_seen": 17170504, "step": 26240 }, { "epoch": 13.75524109014675, "grad_norm": 0.13281439244747162, "learning_rate": 1.3438032380896681e-05, "loss": 0.358, "num_input_tokens_seen": 17173896, "step": 26245 }, { "epoch": 13.757861635220126, "grad_norm": 0.22676204144954681, "learning_rate": 1.342789560606289e-05, "loss": 0.5313, "num_input_tokens_seen": 17176904, "step": 26250 }, { "epoch": 13.7604821802935, "grad_norm": 0.16864895820617676, "learning_rate": 1.341776125198131e-05, "loss": 0.4048, "num_input_tokens_seen": 17179784, "step": 26255 }, { "epoch": 13.763102725366876, "grad_norm": 0.18684296309947968, "learning_rate": 1.3407629320771948e-05, "loss": 0.4397, "num_input_tokens_seen": 17182984, "step": 26260 }, { "epoch": 13.765723270440251, "grad_norm": 0.13236995041370392, "learning_rate": 1.3397499814554266e-05, "loss": 0.5518, "num_input_tokens_seen": 17185608, "step": 26265 }, { "epoch": 13.768343815513626, "grad_norm": 0.1651352196931839, "learning_rate": 1.3387372735447256e-05, "loss": 0.5062, "num_input_tokens_seen": 17189512, "step": 26270 }, { "epoch": 13.770964360587001, "grad_norm": 0.1687563955783844, "learning_rate": 1.3377248085569399e-05, "loss": 0.5759, "num_input_tokens_seen": 17192360, "step": 26275 }, { "epoch": 13.773584905660378, "grad_norm": 0.22285690903663635, "learning_rate": 1.336712586703863e-05, "loss": 0.5061, "num_input_tokens_seen": 17196296, "step": 26280 }, { "epoch": 13.776205450733752, "grad_norm": 0.187227264046669, "learning_rate": 1.3357006081972414e-05, "loss": 0.4992, "num_input_tokens_seen": 17199176, "step": 26285 }, { "epoch": 13.778825995807129, "grad_norm": 0.09156640619039536, "learning_rate": 1.3346888732487706e-05, "loss": 0.3888, "num_input_tokens_seen": 17203688, "step": 26290 }, { "epoch": 13.781446540880504, "grad_norm": 0.25751182436943054, "learning_rate": 1.3336773820700921e-05, "loss": 0.5408, "num_input_tokens_seen": 17206632, "step": 26295 }, { "epoch": 13.784067085953879, "grad_norm": 0.1309174746274948, "learning_rate": 1.3326661348727992e-05, "loss": 0.522, "num_input_tokens_seen": 17209992, "step": 26300 }, { "epoch": 13.786687631027254, "grad_norm": 0.2346443235874176, "learning_rate": 1.3316551318684342e-05, "loss": 0.4159, "num_input_tokens_seen": 17212968, "step": 26305 }, { "epoch": 13.78930817610063, "grad_norm": 0.5893359184265137, "learning_rate": 1.3306443732684854e-05, "loss": 0.4034, "num_input_tokens_seen": 17216200, "step": 26310 }, { "epoch": 13.791928721174004, "grad_norm": 0.1634870022535324, "learning_rate": 1.3296338592843929e-05, "loss": 0.5317, "num_input_tokens_seen": 17219144, "step": 26315 }, { "epoch": 13.79454926624738, "grad_norm": 0.14251822233200073, "learning_rate": 1.3286235901275457e-05, "loss": 0.3945, "num_input_tokens_seen": 17222792, "step": 26320 }, { "epoch": 13.797169811320755, "grad_norm": 0.14191314578056335, "learning_rate": 1.327613566009279e-05, "loss": 0.5457, "num_input_tokens_seen": 17226152, "step": 26325 }, { "epoch": 13.79979035639413, "grad_norm": 0.24071460962295532, "learning_rate": 1.3266037871408798e-05, "loss": 0.5551, "num_input_tokens_seen": 17228456, "step": 26330 }, { "epoch": 13.802410901467505, "grad_norm": 0.14466455578804016, "learning_rate": 1.3255942537335805e-05, "loss": 0.3893, "num_input_tokens_seen": 17231400, "step": 26335 }, { "epoch": 13.80503144654088, "grad_norm": 0.10616908967494965, "learning_rate": 1.3245849659985652e-05, "loss": 0.5472, "num_input_tokens_seen": 17235112, "step": 26340 }, { "epoch": 13.807651991614255, "grad_norm": 0.17470841109752655, "learning_rate": 1.3235759241469658e-05, "loss": 0.4802, "num_input_tokens_seen": 17238280, "step": 26345 }, { "epoch": 13.81027253668763, "grad_norm": 0.1401578187942505, "learning_rate": 1.322567128389861e-05, "loss": 0.5185, "num_input_tokens_seen": 17241192, "step": 26350 }, { "epoch": 13.812893081761006, "grad_norm": 0.40281638503074646, "learning_rate": 1.3215585789382812e-05, "loss": 0.6974, "num_input_tokens_seen": 17244712, "step": 26355 }, { "epoch": 13.815513626834381, "grad_norm": 0.09830518811941147, "learning_rate": 1.320550276003201e-05, "loss": 0.5127, "num_input_tokens_seen": 17248648, "step": 26360 }, { "epoch": 13.818134171907756, "grad_norm": 0.13922394812107086, "learning_rate": 1.3195422197955473e-05, "loss": 0.5382, "num_input_tokens_seen": 17252136, "step": 26365 }, { "epoch": 13.820754716981131, "grad_norm": 0.24267855286598206, "learning_rate": 1.3185344105261937e-05, "loss": 0.4324, "num_input_tokens_seen": 17254952, "step": 26370 }, { "epoch": 13.823375262054507, "grad_norm": 0.1943255066871643, "learning_rate": 1.3175268484059639e-05, "loss": 0.437, "num_input_tokens_seen": 17258280, "step": 26375 }, { "epoch": 13.825995807127882, "grad_norm": 0.2436133474111557, "learning_rate": 1.316519533645626e-05, "loss": 0.4845, "num_input_tokens_seen": 17260840, "step": 26380 }, { "epoch": 13.828616352201259, "grad_norm": 0.16621017456054688, "learning_rate": 1.3155124664559004e-05, "loss": 0.5285, "num_input_tokens_seen": 17264488, "step": 26385 }, { "epoch": 13.831236897274634, "grad_norm": 0.24583397805690765, "learning_rate": 1.3145056470474525e-05, "loss": 0.5295, "num_input_tokens_seen": 17267592, "step": 26390 }, { "epoch": 13.833857442348009, "grad_norm": 0.2659890353679657, "learning_rate": 1.313499075630899e-05, "loss": 0.3538, "num_input_tokens_seen": 17270888, "step": 26395 }, { "epoch": 13.836477987421384, "grad_norm": 0.15824349224567413, "learning_rate": 1.312492752416802e-05, "loss": 0.446, "num_input_tokens_seen": 17274152, "step": 26400 }, { "epoch": 13.83909853249476, "grad_norm": 0.13320834934711456, "learning_rate": 1.3114866776156748e-05, "loss": 0.7126, "num_input_tokens_seen": 17276808, "step": 26405 }, { "epoch": 13.841719077568134, "grad_norm": 0.10216176509857178, "learning_rate": 1.3104808514379748e-05, "loss": 0.3576, "num_input_tokens_seen": 17281192, "step": 26410 }, { "epoch": 13.84433962264151, "grad_norm": 0.24389874935150146, "learning_rate": 1.3094752740941094e-05, "loss": 0.4016, "num_input_tokens_seen": 17284936, "step": 26415 }, { "epoch": 13.846960167714885, "grad_norm": 0.19083231687545776, "learning_rate": 1.3084699457944344e-05, "loss": 0.582, "num_input_tokens_seen": 17287624, "step": 26420 }, { "epoch": 13.84958071278826, "grad_norm": 0.3683542311191559, "learning_rate": 1.3074648667492528e-05, "loss": 0.5331, "num_input_tokens_seen": 17290600, "step": 26425 }, { "epoch": 13.852201257861635, "grad_norm": 0.1782427430152893, "learning_rate": 1.3064600371688171e-05, "loss": 0.4894, "num_input_tokens_seen": 17292904, "step": 26430 }, { "epoch": 13.85482180293501, "grad_norm": 0.14622782170772552, "learning_rate": 1.3054554572633249e-05, "loss": 0.4339, "num_input_tokens_seen": 17296232, "step": 26435 }, { "epoch": 13.857442348008385, "grad_norm": 0.12389977276325226, "learning_rate": 1.3044511272429222e-05, "loss": 0.5363, "num_input_tokens_seen": 17300360, "step": 26440 }, { "epoch": 13.86006289308176, "grad_norm": 0.16969498991966248, "learning_rate": 1.3034470473177038e-05, "loss": 0.516, "num_input_tokens_seen": 17303720, "step": 26445 }, { "epoch": 13.862683438155136, "grad_norm": 0.30411863327026367, "learning_rate": 1.3024432176977125e-05, "loss": 0.5182, "num_input_tokens_seen": 17306568, "step": 26450 }, { "epoch": 13.865303983228511, "grad_norm": 0.21317987143993378, "learning_rate": 1.3014396385929372e-05, "loss": 0.4538, "num_input_tokens_seen": 17310280, "step": 26455 }, { "epoch": 13.867924528301886, "grad_norm": 0.18976734578609467, "learning_rate": 1.3004363102133165e-05, "loss": 0.4274, "num_input_tokens_seen": 17315272, "step": 26460 }, { "epoch": 13.870545073375261, "grad_norm": 0.1549684703350067, "learning_rate": 1.2994332327687348e-05, "loss": 0.381, "num_input_tokens_seen": 17319016, "step": 26465 }, { "epoch": 13.873165618448636, "grad_norm": 0.22638289630413055, "learning_rate": 1.2984304064690228e-05, "loss": 0.4566, "num_input_tokens_seen": 17323496, "step": 26470 }, { "epoch": 13.875786163522012, "grad_norm": 0.271724134683609, "learning_rate": 1.2974278315239619e-05, "loss": 0.4674, "num_input_tokens_seen": 17325992, "step": 26475 }, { "epoch": 13.878406708595389, "grad_norm": 0.1979234516620636, "learning_rate": 1.2964255081432786e-05, "loss": 0.5467, "num_input_tokens_seen": 17328488, "step": 26480 }, { "epoch": 13.881027253668764, "grad_norm": 0.1508432775735855, "learning_rate": 1.2954234365366488e-05, "loss": 0.6096, "num_input_tokens_seen": 17331144, "step": 26485 }, { "epoch": 13.883647798742139, "grad_norm": 0.06927583366632462, "learning_rate": 1.2944216169136936e-05, "loss": 0.3655, "num_input_tokens_seen": 17335592, "step": 26490 }, { "epoch": 13.886268343815514, "grad_norm": 0.19330565631389618, "learning_rate": 1.2934200494839815e-05, "loss": 0.5204, "num_input_tokens_seen": 17338856, "step": 26495 }, { "epoch": 13.88888888888889, "grad_norm": 0.19060944020748138, "learning_rate": 1.2924187344570293e-05, "loss": 0.5211, "num_input_tokens_seen": 17342664, "step": 26500 }, { "epoch": 13.891509433962264, "grad_norm": 0.182791605591774, "learning_rate": 1.2914176720423011e-05, "loss": 0.5195, "num_input_tokens_seen": 17346120, "step": 26505 }, { "epoch": 13.89412997903564, "grad_norm": 0.34697097539901733, "learning_rate": 1.2904168624492088e-05, "loss": 0.5487, "num_input_tokens_seen": 17349864, "step": 26510 }, { "epoch": 13.896750524109015, "grad_norm": 0.20541633665561676, "learning_rate": 1.2894163058871092e-05, "loss": 0.5062, "num_input_tokens_seen": 17352808, "step": 26515 }, { "epoch": 13.89937106918239, "grad_norm": 0.3800756335258484, "learning_rate": 1.2884160025653064e-05, "loss": 0.7156, "num_input_tokens_seen": 17356232, "step": 26520 }, { "epoch": 13.901991614255765, "grad_norm": 0.1580822914838791, "learning_rate": 1.2874159526930532e-05, "loss": 0.5618, "num_input_tokens_seen": 17359272, "step": 26525 }, { "epoch": 13.90461215932914, "grad_norm": 0.11482523381710052, "learning_rate": 1.286416156479549e-05, "loss": 0.4769, "num_input_tokens_seen": 17363880, "step": 26530 }, { "epoch": 13.907232704402515, "grad_norm": 0.20109917223453522, "learning_rate": 1.2854166141339394e-05, "loss": 0.4561, "num_input_tokens_seen": 17366600, "step": 26535 }, { "epoch": 13.90985324947589, "grad_norm": 0.18899348378181458, "learning_rate": 1.2844173258653183e-05, "loss": 0.4818, "num_input_tokens_seen": 17369544, "step": 26540 }, { "epoch": 13.912473794549266, "grad_norm": 0.14528468251228333, "learning_rate": 1.2834182918827247e-05, "loss": 0.4663, "num_input_tokens_seen": 17372360, "step": 26545 }, { "epoch": 13.915094339622641, "grad_norm": 0.39758017659187317, "learning_rate": 1.282419512395144e-05, "loss": 0.5227, "num_input_tokens_seen": 17375112, "step": 26550 }, { "epoch": 13.917714884696016, "grad_norm": 0.22827495634555817, "learning_rate": 1.2814209876115102e-05, "loss": 0.3879, "num_input_tokens_seen": 17378024, "step": 26555 }, { "epoch": 13.920335429769391, "grad_norm": 0.15461087226867676, "learning_rate": 1.2804227177407038e-05, "loss": 0.3837, "num_input_tokens_seen": 17380744, "step": 26560 }, { "epoch": 13.922955974842766, "grad_norm": 0.1128971129655838, "learning_rate": 1.279424702991552e-05, "loss": 0.4827, "num_input_tokens_seen": 17383560, "step": 26565 }, { "epoch": 13.925576519916142, "grad_norm": 0.1509537696838379, "learning_rate": 1.2784269435728264e-05, "loss": 0.4175, "num_input_tokens_seen": 17386344, "step": 26570 }, { "epoch": 13.928197064989519, "grad_norm": 0.2406889796257019, "learning_rate": 1.2774294396932488e-05, "loss": 0.5762, "num_input_tokens_seen": 17389416, "step": 26575 }, { "epoch": 13.930817610062894, "grad_norm": 0.1843259632587433, "learning_rate": 1.2764321915614835e-05, "loss": 0.4058, "num_input_tokens_seen": 17392264, "step": 26580 }, { "epoch": 13.933438155136269, "grad_norm": 0.2544418275356293, "learning_rate": 1.2754351993861444e-05, "loss": 0.4487, "num_input_tokens_seen": 17395720, "step": 26585 }, { "epoch": 13.936058700209644, "grad_norm": 0.11151643842458725, "learning_rate": 1.2744384633757922e-05, "loss": 0.4175, "num_input_tokens_seen": 17398568, "step": 26590 }, { "epoch": 13.93867924528302, "grad_norm": 0.30574318766593933, "learning_rate": 1.2734419837389306e-05, "loss": 0.3936, "num_input_tokens_seen": 17401160, "step": 26595 }, { "epoch": 13.941299790356394, "grad_norm": 0.21584804356098175, "learning_rate": 1.2724457606840123e-05, "loss": 0.5376, "num_input_tokens_seen": 17404232, "step": 26600 }, { "epoch": 13.94392033542977, "grad_norm": 0.15822434425354004, "learning_rate": 1.2714497944194376e-05, "loss": 0.3375, "num_input_tokens_seen": 17407368, "step": 26605 }, { "epoch": 13.946540880503145, "grad_norm": 0.13549689948558807, "learning_rate": 1.2704540851535484e-05, "loss": 0.4658, "num_input_tokens_seen": 17410888, "step": 26610 }, { "epoch": 13.94916142557652, "grad_norm": 0.14505429565906525, "learning_rate": 1.2694586330946376e-05, "loss": 0.4918, "num_input_tokens_seen": 17413896, "step": 26615 }, { "epoch": 13.951781970649895, "grad_norm": 0.13702502846717834, "learning_rate": 1.2684634384509427e-05, "loss": 0.5866, "num_input_tokens_seen": 17417416, "step": 26620 }, { "epoch": 13.95440251572327, "grad_norm": 0.13671723008155823, "learning_rate": 1.2674685014306456e-05, "loss": 0.4025, "num_input_tokens_seen": 17420616, "step": 26625 }, { "epoch": 13.957023060796645, "grad_norm": 0.1326499879360199, "learning_rate": 1.2664738222418776e-05, "loss": 0.5057, "num_input_tokens_seen": 17424008, "step": 26630 }, { "epoch": 13.95964360587002, "grad_norm": 0.4133944511413574, "learning_rate": 1.2654794010927118e-05, "loss": 0.6228, "num_input_tokens_seen": 17427272, "step": 26635 }, { "epoch": 13.962264150943396, "grad_norm": 0.19287481904029846, "learning_rate": 1.2644852381911715e-05, "loss": 0.4708, "num_input_tokens_seen": 17430408, "step": 26640 }, { "epoch": 13.964884696016771, "grad_norm": 0.13764794170856476, "learning_rate": 1.2634913337452248e-05, "loss": 0.4851, "num_input_tokens_seen": 17433192, "step": 26645 }, { "epoch": 13.967505241090146, "grad_norm": 0.19642230868339539, "learning_rate": 1.2624976879627832e-05, "loss": 0.3726, "num_input_tokens_seen": 17436232, "step": 26650 }, { "epoch": 13.970125786163521, "grad_norm": 0.24528713524341583, "learning_rate": 1.2615043010517069e-05, "loss": 0.5598, "num_input_tokens_seen": 17439688, "step": 26655 }, { "epoch": 13.972746331236896, "grad_norm": 0.20934109389781952, "learning_rate": 1.2605111732198027e-05, "loss": 0.4874, "num_input_tokens_seen": 17442088, "step": 26660 }, { "epoch": 13.975366876310272, "grad_norm": 0.09419694542884827, "learning_rate": 1.2595183046748188e-05, "loss": 0.4907, "num_input_tokens_seen": 17445960, "step": 26665 }, { "epoch": 13.977987421383649, "grad_norm": 0.13363803923130035, "learning_rate": 1.2585256956244545e-05, "loss": 0.5168, "num_input_tokens_seen": 17449096, "step": 26670 }, { "epoch": 13.980607966457024, "grad_norm": 0.10243934392929077, "learning_rate": 1.2575333462763506e-05, "loss": 0.4822, "num_input_tokens_seen": 17452744, "step": 26675 }, { "epoch": 13.983228511530399, "grad_norm": 0.20582690834999084, "learning_rate": 1.2565412568380957e-05, "loss": 0.5376, "num_input_tokens_seen": 17456872, "step": 26680 }, { "epoch": 13.985849056603774, "grad_norm": 0.1494031399488449, "learning_rate": 1.2555494275172237e-05, "loss": 0.4453, "num_input_tokens_seen": 17460008, "step": 26685 }, { "epoch": 13.98846960167715, "grad_norm": 0.1687234789133072, "learning_rate": 1.2545578585212148e-05, "loss": 0.3921, "num_input_tokens_seen": 17463336, "step": 26690 }, { "epoch": 13.991090146750524, "grad_norm": 0.2650141716003418, "learning_rate": 1.2535665500574922e-05, "loss": 0.4853, "num_input_tokens_seen": 17465960, "step": 26695 }, { "epoch": 13.9937106918239, "grad_norm": 0.12071648985147476, "learning_rate": 1.2525755023334285e-05, "loss": 0.4808, "num_input_tokens_seen": 17469448, "step": 26700 }, { "epoch": 13.996331236897275, "grad_norm": 0.17693674564361572, "learning_rate": 1.251584715556337e-05, "loss": 0.4818, "num_input_tokens_seen": 17473128, "step": 26705 }, { "epoch": 13.99895178197065, "grad_norm": 0.213114395737648, "learning_rate": 1.2505941899334805e-05, "loss": 0.5275, "num_input_tokens_seen": 17476616, "step": 26710 }, { "epoch": 14.0, "eval_loss": 0.46142688393592834, "eval_runtime": 9.8439, "eval_samples_per_second": 86.145, "eval_steps_per_second": 21.536, "num_input_tokens_seen": 17477224, "step": 26712 }, { "epoch": 14.001572327044025, "grad_norm": 0.1707608699798584, "learning_rate": 1.2496039256720665e-05, "loss": 0.4547, "num_input_tokens_seen": 17479176, "step": 26715 }, { "epoch": 14.0041928721174, "grad_norm": 0.1355728954076767, "learning_rate": 1.2486139229792452e-05, "loss": 0.4541, "num_input_tokens_seen": 17482760, "step": 26720 }, { "epoch": 14.006813417190775, "grad_norm": 0.12752865254878998, "learning_rate": 1.2476241820621152e-05, "loss": 0.5788, "num_input_tokens_seen": 17485992, "step": 26725 }, { "epoch": 14.00943396226415, "grad_norm": 0.16670356690883636, "learning_rate": 1.2466347031277176e-05, "loss": 0.4153, "num_input_tokens_seen": 17488712, "step": 26730 }, { "epoch": 14.012054507337526, "grad_norm": 0.12567654252052307, "learning_rate": 1.2456454863830411e-05, "loss": 0.7219, "num_input_tokens_seen": 17491784, "step": 26735 }, { "epoch": 14.014675052410901, "grad_norm": 0.22588852047920227, "learning_rate": 1.2446565320350185e-05, "loss": 0.4434, "num_input_tokens_seen": 17495368, "step": 26740 }, { "epoch": 14.017295597484276, "grad_norm": 0.1385565996170044, "learning_rate": 1.2436678402905283e-05, "loss": 0.3679, "num_input_tokens_seen": 17499272, "step": 26745 }, { "epoch": 14.019916142557651, "grad_norm": 0.18053771555423737, "learning_rate": 1.2426794113563933e-05, "loss": 0.575, "num_input_tokens_seen": 17502728, "step": 26750 }, { "epoch": 14.022536687631026, "grad_norm": 0.23042170703411102, "learning_rate": 1.24169124543938e-05, "loss": 0.6026, "num_input_tokens_seen": 17506984, "step": 26755 }, { "epoch": 14.025157232704403, "grad_norm": 0.12555494904518127, "learning_rate": 1.2407033427462026e-05, "loss": 0.6354, "num_input_tokens_seen": 17511848, "step": 26760 }, { "epoch": 14.027777777777779, "grad_norm": 0.16453826427459717, "learning_rate": 1.2397157034835192e-05, "loss": 0.5027, "num_input_tokens_seen": 17514696, "step": 26765 }, { "epoch": 14.030398322851154, "grad_norm": 0.15020757913589478, "learning_rate": 1.2387283278579325e-05, "loss": 0.4514, "num_input_tokens_seen": 17517352, "step": 26770 }, { "epoch": 14.033018867924529, "grad_norm": 0.09749677777290344, "learning_rate": 1.237741216075991e-05, "loss": 0.4814, "num_input_tokens_seen": 17520872, "step": 26775 }, { "epoch": 14.035639412997904, "grad_norm": 0.1699298620223999, "learning_rate": 1.2367543683441868e-05, "loss": 0.386, "num_input_tokens_seen": 17523880, "step": 26780 }, { "epoch": 14.03825995807128, "grad_norm": 0.08878279477357864, "learning_rate": 1.2357677848689558e-05, "loss": 0.3901, "num_input_tokens_seen": 17527016, "step": 26785 }, { "epoch": 14.040880503144654, "grad_norm": 0.38342180848121643, "learning_rate": 1.2347814658566811e-05, "loss": 0.5764, "num_input_tokens_seen": 17531688, "step": 26790 }, { "epoch": 14.04350104821803, "grad_norm": 0.1394280046224594, "learning_rate": 1.2337954115136894e-05, "loss": 0.3668, "num_input_tokens_seen": 17534760, "step": 26795 }, { "epoch": 14.046121593291405, "grad_norm": 0.18338985741138458, "learning_rate": 1.2328096220462529e-05, "loss": 0.5167, "num_input_tokens_seen": 17537672, "step": 26800 }, { "epoch": 14.04874213836478, "grad_norm": 0.22454604506492615, "learning_rate": 1.2318240976605865e-05, "loss": 0.4318, "num_input_tokens_seen": 17541480, "step": 26805 }, { "epoch": 14.051362683438155, "grad_norm": 0.1258103996515274, "learning_rate": 1.23083883856285e-05, "loss": 0.3422, "num_input_tokens_seen": 17544040, "step": 26810 }, { "epoch": 14.05398322851153, "grad_norm": 0.13242444396018982, "learning_rate": 1.229853844959149e-05, "loss": 0.3676, "num_input_tokens_seen": 17546376, "step": 26815 }, { "epoch": 14.056603773584905, "grad_norm": 0.1553943157196045, "learning_rate": 1.228869117055533e-05, "loss": 0.3975, "num_input_tokens_seen": 17549576, "step": 26820 }, { "epoch": 14.05922431865828, "grad_norm": 0.11892350018024445, "learning_rate": 1.227884655057997e-05, "loss": 0.4098, "num_input_tokens_seen": 17552776, "step": 26825 }, { "epoch": 14.061844863731656, "grad_norm": 0.1651279777288437, "learning_rate": 1.2269004591724777e-05, "loss": 0.5103, "num_input_tokens_seen": 17556328, "step": 26830 }, { "epoch": 14.064465408805031, "grad_norm": 0.2293032705783844, "learning_rate": 1.2259165296048575e-05, "loss": 0.8033, "num_input_tokens_seen": 17559144, "step": 26835 }, { "epoch": 14.067085953878406, "grad_norm": 0.22095432877540588, "learning_rate": 1.2249328665609638e-05, "loss": 0.5878, "num_input_tokens_seen": 17561960, "step": 26840 }, { "epoch": 14.069706498951781, "grad_norm": 0.15593917667865753, "learning_rate": 1.2239494702465676e-05, "loss": 0.5541, "num_input_tokens_seen": 17565480, "step": 26845 }, { "epoch": 14.072327044025156, "grad_norm": 0.21788258850574493, "learning_rate": 1.2229663408673842e-05, "loss": 0.5132, "num_input_tokens_seen": 17568456, "step": 26850 }, { "epoch": 14.074947589098532, "grad_norm": 0.21538451313972473, "learning_rate": 1.221983478629074e-05, "loss": 0.411, "num_input_tokens_seen": 17571336, "step": 26855 }, { "epoch": 14.077568134171909, "grad_norm": 0.22671759128570557, "learning_rate": 1.22100088373724e-05, "loss": 0.665, "num_input_tokens_seen": 17577576, "step": 26860 }, { "epoch": 14.080188679245284, "grad_norm": 0.13217739760875702, "learning_rate": 1.2200185563974283e-05, "loss": 0.3863, "num_input_tokens_seen": 17581352, "step": 26865 }, { "epoch": 14.082809224318659, "grad_norm": 0.12825904786586761, "learning_rate": 1.219036496815132e-05, "loss": 0.5687, "num_input_tokens_seen": 17584424, "step": 26870 }, { "epoch": 14.085429769392034, "grad_norm": 0.11032892018556595, "learning_rate": 1.2180547051957863e-05, "loss": 0.4772, "num_input_tokens_seen": 17587656, "step": 26875 }, { "epoch": 14.08805031446541, "grad_norm": 0.17165160179138184, "learning_rate": 1.2170731817447723e-05, "loss": 0.4632, "num_input_tokens_seen": 17591528, "step": 26880 }, { "epoch": 14.090670859538784, "grad_norm": 0.19324006140232086, "learning_rate": 1.216091926667412e-05, "loss": 0.4958, "num_input_tokens_seen": 17594536, "step": 26885 }, { "epoch": 14.09329140461216, "grad_norm": 0.19779342412948608, "learning_rate": 1.2151109401689723e-05, "loss": 0.5483, "num_input_tokens_seen": 17597640, "step": 26890 }, { "epoch": 14.095911949685535, "grad_norm": 0.21557998657226562, "learning_rate": 1.214130222454665e-05, "loss": 0.5359, "num_input_tokens_seen": 17600136, "step": 26895 }, { "epoch": 14.09853249475891, "grad_norm": 0.16388975083827972, "learning_rate": 1.2131497737296452e-05, "loss": 0.5264, "num_input_tokens_seen": 17602856, "step": 26900 }, { "epoch": 14.101153039832285, "grad_norm": 0.22398048639297485, "learning_rate": 1.2121695941990124e-05, "loss": 0.4542, "num_input_tokens_seen": 17606344, "step": 26905 }, { "epoch": 14.10377358490566, "grad_norm": 0.21053245663642883, "learning_rate": 1.2111896840678075e-05, "loss": 0.464, "num_input_tokens_seen": 17609064, "step": 26910 }, { "epoch": 14.106394129979035, "grad_norm": 0.10218216478824615, "learning_rate": 1.210210043541018e-05, "loss": 0.4891, "num_input_tokens_seen": 17612456, "step": 26915 }, { "epoch": 14.10901467505241, "grad_norm": 0.22537599503993988, "learning_rate": 1.2092306728235717e-05, "loss": 0.3732, "num_input_tokens_seen": 17616424, "step": 26920 }, { "epoch": 14.111635220125786, "grad_norm": 0.3096492886543274, "learning_rate": 1.2082515721203427e-05, "loss": 0.5269, "num_input_tokens_seen": 17619944, "step": 26925 }, { "epoch": 14.114255765199161, "grad_norm": 0.11984028667211533, "learning_rate": 1.2072727416361479e-05, "loss": 0.4525, "num_input_tokens_seen": 17623368, "step": 26930 }, { "epoch": 14.116876310272536, "grad_norm": 0.20468905568122864, "learning_rate": 1.2062941815757484e-05, "loss": 0.6562, "num_input_tokens_seen": 17626760, "step": 26935 }, { "epoch": 14.119496855345911, "grad_norm": 0.23934558033943176, "learning_rate": 1.2053158921438454e-05, "loss": 0.5702, "num_input_tokens_seen": 17629608, "step": 26940 }, { "epoch": 14.122117400419286, "grad_norm": 0.17061232030391693, "learning_rate": 1.2043378735450885e-05, "loss": 0.4396, "num_input_tokens_seen": 17632072, "step": 26945 }, { "epoch": 14.124737945492662, "grad_norm": 0.15111421048641205, "learning_rate": 1.2033601259840657e-05, "loss": 0.4632, "num_input_tokens_seen": 17635720, "step": 26950 }, { "epoch": 14.127358490566039, "grad_norm": 0.13722138106822968, "learning_rate": 1.2023826496653118e-05, "loss": 0.4478, "num_input_tokens_seen": 17638568, "step": 26955 }, { "epoch": 14.129979035639414, "grad_norm": 0.20486536622047424, "learning_rate": 1.2014054447933044e-05, "loss": 0.4003, "num_input_tokens_seen": 17643080, "step": 26960 }, { "epoch": 14.132599580712789, "grad_norm": 0.27084818482398987, "learning_rate": 1.200428511572462e-05, "loss": 0.4431, "num_input_tokens_seen": 17646024, "step": 26965 }, { "epoch": 14.135220125786164, "grad_norm": 0.13583549857139587, "learning_rate": 1.1994518502071484e-05, "loss": 0.4484, "num_input_tokens_seen": 17648936, "step": 26970 }, { "epoch": 14.13784067085954, "grad_norm": 0.09780305624008179, "learning_rate": 1.1984754609016716e-05, "loss": 0.4794, "num_input_tokens_seen": 17653160, "step": 26975 }, { "epoch": 14.140461215932914, "grad_norm": 0.09321776777505875, "learning_rate": 1.1974993438602785e-05, "loss": 0.5291, "num_input_tokens_seen": 17657384, "step": 26980 }, { "epoch": 14.14308176100629, "grad_norm": 0.20314575731754303, "learning_rate": 1.1965234992871641e-05, "loss": 0.493, "num_input_tokens_seen": 17660424, "step": 26985 }, { "epoch": 14.145702306079665, "grad_norm": 0.1439664661884308, "learning_rate": 1.195547927386462e-05, "loss": 0.5022, "num_input_tokens_seen": 17664424, "step": 26990 }, { "epoch": 14.14832285115304, "grad_norm": 0.2286505401134491, "learning_rate": 1.1945726283622514e-05, "loss": 0.5071, "num_input_tokens_seen": 17667368, "step": 26995 }, { "epoch": 14.150943396226415, "grad_norm": 0.1667547971010208, "learning_rate": 1.1935976024185547e-05, "loss": 0.4541, "num_input_tokens_seen": 17670184, "step": 27000 }, { "epoch": 14.15356394129979, "grad_norm": 0.16800905764102936, "learning_rate": 1.1926228497593341e-05, "loss": 0.528, "num_input_tokens_seen": 17673448, "step": 27005 }, { "epoch": 14.156184486373165, "grad_norm": 0.21516141295433044, "learning_rate": 1.1916483705884985e-05, "loss": 0.4325, "num_input_tokens_seen": 17676328, "step": 27010 }, { "epoch": 14.15880503144654, "grad_norm": 0.16835199296474457, "learning_rate": 1.1906741651098983e-05, "loss": 0.52, "num_input_tokens_seen": 17680232, "step": 27015 }, { "epoch": 14.161425576519916, "grad_norm": 0.1520242840051651, "learning_rate": 1.1897002335273239e-05, "loss": 0.4816, "num_input_tokens_seen": 17683880, "step": 27020 }, { "epoch": 14.164046121593291, "grad_norm": 0.14212770760059357, "learning_rate": 1.1887265760445121e-05, "loss": 0.6169, "num_input_tokens_seen": 17687080, "step": 27025 }, { "epoch": 14.166666666666666, "grad_norm": 0.14502422511577606, "learning_rate": 1.1877531928651417e-05, "loss": 0.4879, "num_input_tokens_seen": 17690344, "step": 27030 }, { "epoch": 14.169287211740041, "grad_norm": 0.1862606555223465, "learning_rate": 1.1867800841928313e-05, "loss": 0.5417, "num_input_tokens_seen": 17695272, "step": 27035 }, { "epoch": 14.171907756813416, "grad_norm": 0.15213662385940552, "learning_rate": 1.1858072502311465e-05, "loss": 0.344, "num_input_tokens_seen": 17698216, "step": 27040 }, { "epoch": 14.174528301886792, "grad_norm": 0.13475899398326874, "learning_rate": 1.1848346911835908e-05, "loss": 0.4213, "num_input_tokens_seen": 17700968, "step": 27045 }, { "epoch": 14.177148846960169, "grad_norm": 0.21120178699493408, "learning_rate": 1.1838624072536137e-05, "loss": 0.3769, "num_input_tokens_seen": 17703496, "step": 27050 }, { "epoch": 14.179769392033544, "grad_norm": 0.1088583841919899, "learning_rate": 1.1828903986446055e-05, "loss": 0.3968, "num_input_tokens_seen": 17706536, "step": 27055 }, { "epoch": 14.182389937106919, "grad_norm": 0.19752347469329834, "learning_rate": 1.1819186655599007e-05, "loss": 0.4191, "num_input_tokens_seen": 17710376, "step": 27060 }, { "epoch": 14.185010482180294, "grad_norm": 0.10466302186250687, "learning_rate": 1.1809472082027734e-05, "loss": 0.4104, "num_input_tokens_seen": 17713320, "step": 27065 }, { "epoch": 14.18763102725367, "grad_norm": 0.1197916641831398, "learning_rate": 1.1799760267764407e-05, "loss": 0.5269, "num_input_tokens_seen": 17716392, "step": 27070 }, { "epoch": 14.190251572327044, "grad_norm": 0.23658055067062378, "learning_rate": 1.1790051214840634e-05, "loss": 0.5458, "num_input_tokens_seen": 17720488, "step": 27075 }, { "epoch": 14.19287211740042, "grad_norm": 0.1588512659072876, "learning_rate": 1.1780344925287443e-05, "loss": 0.4792, "num_input_tokens_seen": 17724040, "step": 27080 }, { "epoch": 14.195492662473795, "grad_norm": 0.14578770101070404, "learning_rate": 1.1770641401135282e-05, "loss": 0.587, "num_input_tokens_seen": 17727208, "step": 27085 }, { "epoch": 14.19811320754717, "grad_norm": 0.1613471657037735, "learning_rate": 1.1760940644414003e-05, "loss": 0.3721, "num_input_tokens_seen": 17730504, "step": 27090 }, { "epoch": 14.200733752620545, "grad_norm": 0.18205825984477997, "learning_rate": 1.1751242657152914e-05, "loss": 0.46, "num_input_tokens_seen": 17733800, "step": 27095 }, { "epoch": 14.20335429769392, "grad_norm": 0.22970205545425415, "learning_rate": 1.1741547441380698e-05, "loss": 0.5639, "num_input_tokens_seen": 17736072, "step": 27100 }, { "epoch": 14.205974842767295, "grad_norm": 0.19459936022758484, "learning_rate": 1.17318549991255e-05, "loss": 0.4941, "num_input_tokens_seen": 17739336, "step": 27105 }, { "epoch": 14.20859538784067, "grad_norm": 0.17590253055095673, "learning_rate": 1.1722165332414863e-05, "loss": 0.507, "num_input_tokens_seen": 17743528, "step": 27110 }, { "epoch": 14.211215932914046, "grad_norm": 0.18270955979824066, "learning_rate": 1.1712478443275768e-05, "loss": 0.388, "num_input_tokens_seen": 17746024, "step": 27115 }, { "epoch": 14.213836477987421, "grad_norm": 0.183600053191185, "learning_rate": 1.170279433373459e-05, "loss": 0.4058, "num_input_tokens_seen": 17750088, "step": 27120 }, { "epoch": 14.216457023060796, "grad_norm": 0.29123207926750183, "learning_rate": 1.1693113005817125e-05, "loss": 0.6197, "num_input_tokens_seen": 17753224, "step": 27125 }, { "epoch": 14.219077568134171, "grad_norm": 0.1383047103881836, "learning_rate": 1.1683434461548606e-05, "loss": 0.5046, "num_input_tokens_seen": 17756488, "step": 27130 }, { "epoch": 14.221698113207546, "grad_norm": 0.19033434987068176, "learning_rate": 1.1673758702953675e-05, "loss": 0.5045, "num_input_tokens_seen": 17759336, "step": 27135 }, { "epoch": 14.224318658280922, "grad_norm": 0.23244550824165344, "learning_rate": 1.1664085732056398e-05, "loss": 0.6078, "num_input_tokens_seen": 17761672, "step": 27140 }, { "epoch": 14.226939203354299, "grad_norm": 0.25061342120170593, "learning_rate": 1.1654415550880243e-05, "loss": 0.5063, "num_input_tokens_seen": 17764136, "step": 27145 }, { "epoch": 14.229559748427674, "grad_norm": 0.11275365203619003, "learning_rate": 1.1644748161448088e-05, "loss": 0.4457, "num_input_tokens_seen": 17767048, "step": 27150 }, { "epoch": 14.232180293501049, "grad_norm": 0.10527676343917847, "learning_rate": 1.1635083565782257e-05, "loss": 0.5492, "num_input_tokens_seen": 17770184, "step": 27155 }, { "epoch": 14.234800838574424, "grad_norm": 0.1533377319574356, "learning_rate": 1.1625421765904465e-05, "loss": 0.4292, "num_input_tokens_seen": 17772680, "step": 27160 }, { "epoch": 14.2374213836478, "grad_norm": 0.23246973752975464, "learning_rate": 1.1615762763835858e-05, "loss": 0.5022, "num_input_tokens_seen": 17775784, "step": 27165 }, { "epoch": 14.240041928721174, "grad_norm": 0.13485659658908844, "learning_rate": 1.1606106561596993e-05, "loss": 0.3546, "num_input_tokens_seen": 17779016, "step": 27170 }, { "epoch": 14.24266247379455, "grad_norm": 0.12957903742790222, "learning_rate": 1.1596453161207829e-05, "loss": 0.405, "num_input_tokens_seen": 17782248, "step": 27175 }, { "epoch": 14.245283018867925, "grad_norm": 0.15187261998653412, "learning_rate": 1.1586802564687741e-05, "loss": 0.5261, "num_input_tokens_seen": 17785320, "step": 27180 }, { "epoch": 14.2479035639413, "grad_norm": 0.25622427463531494, "learning_rate": 1.1577154774055528e-05, "loss": 0.4241, "num_input_tokens_seen": 17788520, "step": 27185 }, { "epoch": 14.250524109014675, "grad_norm": 0.11479068547487259, "learning_rate": 1.1567509791329401e-05, "loss": 0.9196, "num_input_tokens_seen": 17792200, "step": 27190 }, { "epoch": 14.25314465408805, "grad_norm": 0.1474035680294037, "learning_rate": 1.155786761852699e-05, "loss": 0.3907, "num_input_tokens_seen": 17795816, "step": 27195 }, { "epoch": 14.255765199161425, "grad_norm": 0.15588946640491486, "learning_rate": 1.1548228257665317e-05, "loss": 0.4879, "num_input_tokens_seen": 17798568, "step": 27200 }, { "epoch": 14.2583857442348, "grad_norm": 0.17048124969005585, "learning_rate": 1.1538591710760815e-05, "loss": 0.4116, "num_input_tokens_seen": 17801064, "step": 27205 }, { "epoch": 14.261006289308176, "grad_norm": 0.16609278321266174, "learning_rate": 1.1528957979829355e-05, "loss": 0.4563, "num_input_tokens_seen": 17803400, "step": 27210 }, { "epoch": 14.26362683438155, "grad_norm": 0.2275407612323761, "learning_rate": 1.1519327066886198e-05, "loss": 0.4705, "num_input_tokens_seen": 17806696, "step": 27215 }, { "epoch": 14.266247379454926, "grad_norm": 0.13745199143886566, "learning_rate": 1.1509698973946032e-05, "loss": 0.5269, "num_input_tokens_seen": 17810216, "step": 27220 }, { "epoch": 14.268867924528301, "grad_norm": 0.22873455286026, "learning_rate": 1.1500073703022926e-05, "loss": 0.5396, "num_input_tokens_seen": 17813544, "step": 27225 }, { "epoch": 14.271488469601676, "grad_norm": 0.16438604891300201, "learning_rate": 1.1490451256130394e-05, "loss": 0.6058, "num_input_tokens_seen": 17816712, "step": 27230 }, { "epoch": 14.274109014675052, "grad_norm": 0.17360694706439972, "learning_rate": 1.1480831635281328e-05, "loss": 0.3508, "num_input_tokens_seen": 17819016, "step": 27235 }, { "epoch": 14.276729559748428, "grad_norm": 0.16219690442085266, "learning_rate": 1.1471214842488046e-05, "loss": 0.5491, "num_input_tokens_seen": 17822536, "step": 27240 }, { "epoch": 14.279350104821804, "grad_norm": 0.11963340640068054, "learning_rate": 1.1461600879762277e-05, "loss": 0.4455, "num_input_tokens_seen": 17826056, "step": 27245 }, { "epoch": 14.281970649895179, "grad_norm": 0.12442411482334137, "learning_rate": 1.1451989749115157e-05, "loss": 0.4772, "num_input_tokens_seen": 17828712, "step": 27250 }, { "epoch": 14.284591194968554, "grad_norm": 0.2326178401708603, "learning_rate": 1.1442381452557219e-05, "loss": 0.4393, "num_input_tokens_seen": 17832072, "step": 27255 }, { "epoch": 14.28721174004193, "grad_norm": 0.12392515689134598, "learning_rate": 1.1432775992098397e-05, "loss": 0.4635, "num_input_tokens_seen": 17834920, "step": 27260 }, { "epoch": 14.289832285115304, "grad_norm": 0.16938626766204834, "learning_rate": 1.1423173369748056e-05, "loss": 0.5238, "num_input_tokens_seen": 17838568, "step": 27265 }, { "epoch": 14.29245283018868, "grad_norm": 0.47811293601989746, "learning_rate": 1.1413573587514952e-05, "loss": 0.469, "num_input_tokens_seen": 17842504, "step": 27270 }, { "epoch": 14.295073375262055, "grad_norm": 0.20536109805107117, "learning_rate": 1.140397664740726e-05, "loss": 0.6367, "num_input_tokens_seen": 17846312, "step": 27275 }, { "epoch": 14.29769392033543, "grad_norm": 0.1790262907743454, "learning_rate": 1.1394382551432534e-05, "loss": 0.455, "num_input_tokens_seen": 17848424, "step": 27280 }, { "epoch": 14.300314465408805, "grad_norm": 0.2309419810771942, "learning_rate": 1.1384791301597767e-05, "loss": 0.5025, "num_input_tokens_seen": 17851176, "step": 27285 }, { "epoch": 14.30293501048218, "grad_norm": 0.18193551898002625, "learning_rate": 1.1375202899909317e-05, "loss": 0.3918, "num_input_tokens_seen": 17853800, "step": 27290 }, { "epoch": 14.305555555555555, "grad_norm": 0.19909368455410004, "learning_rate": 1.1365617348372984e-05, "loss": 0.486, "num_input_tokens_seen": 17857160, "step": 27295 }, { "epoch": 14.30817610062893, "grad_norm": 0.17856790125370026, "learning_rate": 1.135603464899396e-05, "loss": 0.5203, "num_input_tokens_seen": 17860968, "step": 27300 }, { "epoch": 14.310796645702306, "grad_norm": 0.19644632935523987, "learning_rate": 1.1346454803776821e-05, "loss": 0.5612, "num_input_tokens_seen": 17863848, "step": 27305 }, { "epoch": 14.31341719077568, "grad_norm": 0.17267242074012756, "learning_rate": 1.1336877814725566e-05, "loss": 0.3968, "num_input_tokens_seen": 17868616, "step": 27310 }, { "epoch": 14.316037735849056, "grad_norm": 0.15409652888774872, "learning_rate": 1.1327303683843607e-05, "loss": 0.4356, "num_input_tokens_seen": 17871720, "step": 27315 }, { "epoch": 14.318658280922431, "grad_norm": 0.11739696562290192, "learning_rate": 1.1317732413133723e-05, "loss": 0.4538, "num_input_tokens_seen": 17874664, "step": 27320 }, { "epoch": 14.321278825995806, "grad_norm": 0.14650790393352509, "learning_rate": 1.1308164004598118e-05, "loss": 0.5178, "num_input_tokens_seen": 17878184, "step": 27325 }, { "epoch": 14.323899371069182, "grad_norm": 0.19665125012397766, "learning_rate": 1.129859846023841e-05, "loss": 0.507, "num_input_tokens_seen": 17881416, "step": 27330 }, { "epoch": 14.326519916142558, "grad_norm": 0.14248360693454742, "learning_rate": 1.1289035782055582e-05, "loss": 0.641, "num_input_tokens_seen": 17884136, "step": 27335 }, { "epoch": 14.329140461215934, "grad_norm": 0.2181408852338791, "learning_rate": 1.1279475972050044e-05, "loss": 0.5952, "num_input_tokens_seen": 17886792, "step": 27340 }, { "epoch": 14.331761006289309, "grad_norm": 0.16725610196590424, "learning_rate": 1.1269919032221612e-05, "loss": 0.489, "num_input_tokens_seen": 17889864, "step": 27345 }, { "epoch": 14.334381551362684, "grad_norm": 0.17385287582874298, "learning_rate": 1.1260364964569466e-05, "loss": 0.455, "num_input_tokens_seen": 17893160, "step": 27350 }, { "epoch": 14.33700209643606, "grad_norm": 0.1301846206188202, "learning_rate": 1.125081377109223e-05, "loss": 0.5755, "num_input_tokens_seen": 17896360, "step": 27355 }, { "epoch": 14.339622641509434, "grad_norm": 0.1343252807855606, "learning_rate": 1.1241265453787882e-05, "loss": 0.5507, "num_input_tokens_seen": 17899208, "step": 27360 }, { "epoch": 14.34224318658281, "grad_norm": 0.1594858020544052, "learning_rate": 1.1231720014653837e-05, "loss": 0.4453, "num_input_tokens_seen": 17901864, "step": 27365 }, { "epoch": 14.344863731656185, "grad_norm": 0.13750900328159332, "learning_rate": 1.1222177455686896e-05, "loss": 0.4409, "num_input_tokens_seen": 17905288, "step": 27370 }, { "epoch": 14.34748427672956, "grad_norm": 0.141549214720726, "learning_rate": 1.1212637778883237e-05, "loss": 0.4327, "num_input_tokens_seen": 17908104, "step": 27375 }, { "epoch": 14.350104821802935, "grad_norm": 0.121793732047081, "learning_rate": 1.1203100986238468e-05, "loss": 0.4173, "num_input_tokens_seen": 17912008, "step": 27380 }, { "epoch": 14.35272536687631, "grad_norm": 0.14105145633220673, "learning_rate": 1.1193567079747561e-05, "loss": 0.4302, "num_input_tokens_seen": 17914568, "step": 27385 }, { "epoch": 14.355345911949685, "grad_norm": 0.11218443512916565, "learning_rate": 1.1184036061404907e-05, "loss": 0.4558, "num_input_tokens_seen": 17918536, "step": 27390 }, { "epoch": 14.35796645702306, "grad_norm": 0.16305117309093475, "learning_rate": 1.117450793320429e-05, "loss": 0.4451, "num_input_tokens_seen": 17921320, "step": 27395 }, { "epoch": 14.360587002096436, "grad_norm": 0.2539186477661133, "learning_rate": 1.1164982697138893e-05, "loss": 0.5103, "num_input_tokens_seen": 17924552, "step": 27400 }, { "epoch": 14.36320754716981, "grad_norm": 0.172560915350914, "learning_rate": 1.1155460355201267e-05, "loss": 0.5607, "num_input_tokens_seen": 17928072, "step": 27405 }, { "epoch": 14.365828092243186, "grad_norm": 0.23283836245536804, "learning_rate": 1.1145940909383399e-05, "loss": 0.4929, "num_input_tokens_seen": 17930408, "step": 27410 }, { "epoch": 14.368448637316561, "grad_norm": 0.1758519411087036, "learning_rate": 1.1136424361676626e-05, "loss": 0.4148, "num_input_tokens_seen": 17933416, "step": 27415 }, { "epoch": 14.371069182389936, "grad_norm": 0.18161478638648987, "learning_rate": 1.1126910714071711e-05, "loss": 0.4784, "num_input_tokens_seen": 17936232, "step": 27420 }, { "epoch": 14.373689727463312, "grad_norm": 0.12725567817687988, "learning_rate": 1.1117399968558806e-05, "loss": 0.6129, "num_input_tokens_seen": 17940648, "step": 27425 }, { "epoch": 14.376310272536688, "grad_norm": 0.20146378874778748, "learning_rate": 1.1107892127127454e-05, "loss": 0.7199, "num_input_tokens_seen": 17943304, "step": 27430 }, { "epoch": 14.378930817610064, "grad_norm": 0.20091094076633453, "learning_rate": 1.109838719176658e-05, "loss": 0.4698, "num_input_tokens_seen": 17945864, "step": 27435 }, { "epoch": 14.381551362683439, "grad_norm": 0.21149174869060516, "learning_rate": 1.1088885164464497e-05, "loss": 0.4407, "num_input_tokens_seen": 17950152, "step": 27440 }, { "epoch": 14.384171907756814, "grad_norm": 0.13996022939682007, "learning_rate": 1.1079386047208932e-05, "loss": 0.641, "num_input_tokens_seen": 17954024, "step": 27445 }, { "epoch": 14.38679245283019, "grad_norm": 0.11480192095041275, "learning_rate": 1.1069889841986991e-05, "loss": 0.3822, "num_input_tokens_seen": 17957896, "step": 27450 }, { "epoch": 14.389412997903564, "grad_norm": 0.09731482714414597, "learning_rate": 1.1060396550785182e-05, "loss": 0.5684, "num_input_tokens_seen": 17960936, "step": 27455 }, { "epoch": 14.39203354297694, "grad_norm": 0.3928844630718231, "learning_rate": 1.1050906175589382e-05, "loss": 0.4376, "num_input_tokens_seen": 17963496, "step": 27460 }, { "epoch": 14.394654088050315, "grad_norm": 0.2343730926513672, "learning_rate": 1.1041418718384863e-05, "loss": 0.5529, "num_input_tokens_seen": 17966344, "step": 27465 }, { "epoch": 14.39727463312369, "grad_norm": 0.11725296825170517, "learning_rate": 1.1031934181156298e-05, "loss": 0.7443, "num_input_tokens_seen": 17969640, "step": 27470 }, { "epoch": 14.399895178197065, "grad_norm": 0.10472790151834488, "learning_rate": 1.1022452565887747e-05, "loss": 0.4696, "num_input_tokens_seen": 17973064, "step": 27475 }, { "epoch": 14.40251572327044, "grad_norm": 0.16559943556785583, "learning_rate": 1.1012973874562657e-05, "loss": 0.4477, "num_input_tokens_seen": 17977192, "step": 27480 }, { "epoch": 14.405136268343815, "grad_norm": 0.14278757572174072, "learning_rate": 1.1003498109163865e-05, "loss": 0.4266, "num_input_tokens_seen": 17981160, "step": 27485 }, { "epoch": 14.40775681341719, "grad_norm": 0.17758002877235413, "learning_rate": 1.0994025271673592e-05, "loss": 0.4388, "num_input_tokens_seen": 17984744, "step": 27490 }, { "epoch": 14.410377358490566, "grad_norm": 0.13975732028484344, "learning_rate": 1.0984555364073431e-05, "loss": 0.5527, "num_input_tokens_seen": 17988328, "step": 27495 }, { "epoch": 14.41299790356394, "grad_norm": 0.20719169080257416, "learning_rate": 1.0975088388344393e-05, "loss": 0.4107, "num_input_tokens_seen": 17991144, "step": 27500 }, { "epoch": 14.415618448637316, "grad_norm": 0.32218191027641296, "learning_rate": 1.096562434646686e-05, "loss": 0.445, "num_input_tokens_seen": 17994344, "step": 27505 }, { "epoch": 14.418238993710691, "grad_norm": 0.09490402787923813, "learning_rate": 1.0956163240420611e-05, "loss": 0.4706, "num_input_tokens_seen": 17997704, "step": 27510 }, { "epoch": 14.420859538784066, "grad_norm": 0.15418653190135956, "learning_rate": 1.094670507218479e-05, "loss": 0.4439, "num_input_tokens_seen": 18001896, "step": 27515 }, { "epoch": 14.423480083857442, "grad_norm": 0.19186362624168396, "learning_rate": 1.093724984373793e-05, "loss": 0.5493, "num_input_tokens_seen": 18005288, "step": 27520 }, { "epoch": 14.426100628930818, "grad_norm": 0.14306332170963287, "learning_rate": 1.0927797557057968e-05, "loss": 0.5738, "num_input_tokens_seen": 18008744, "step": 27525 }, { "epoch": 14.428721174004194, "grad_norm": 0.16356761753559113, "learning_rate": 1.0918348214122209e-05, "loss": 0.5191, "num_input_tokens_seen": 18012712, "step": 27530 }, { "epoch": 14.431341719077569, "grad_norm": 0.15043224394321442, "learning_rate": 1.0908901816907366e-05, "loss": 0.491, "num_input_tokens_seen": 18015656, "step": 27535 }, { "epoch": 14.433962264150944, "grad_norm": 0.2449239045381546, "learning_rate": 1.08994583673895e-05, "loss": 0.6172, "num_input_tokens_seen": 18018120, "step": 27540 }, { "epoch": 14.43658280922432, "grad_norm": 0.13692085444927216, "learning_rate": 1.0890017867544068e-05, "loss": 0.4961, "num_input_tokens_seen": 18021736, "step": 27545 }, { "epoch": 14.439203354297694, "grad_norm": 0.2658749520778656, "learning_rate": 1.088058031934592e-05, "loss": 0.4618, "num_input_tokens_seen": 18024584, "step": 27550 }, { "epoch": 14.44182389937107, "grad_norm": 0.20665863156318665, "learning_rate": 1.087114572476929e-05, "loss": 0.466, "num_input_tokens_seen": 18027688, "step": 27555 }, { "epoch": 14.444444444444445, "grad_norm": 0.19211746752262115, "learning_rate": 1.0861714085787783e-05, "loss": 0.522, "num_input_tokens_seen": 18030568, "step": 27560 }, { "epoch": 14.44706498951782, "grad_norm": 0.19318857789039612, "learning_rate": 1.0852285404374402e-05, "loss": 0.4522, "num_input_tokens_seen": 18033128, "step": 27565 }, { "epoch": 14.449685534591195, "grad_norm": 0.2122289538383484, "learning_rate": 1.0842859682501505e-05, "loss": 0.504, "num_input_tokens_seen": 18036776, "step": 27570 }, { "epoch": 14.45230607966457, "grad_norm": 0.15325888991355896, "learning_rate": 1.0833436922140841e-05, "loss": 0.4081, "num_input_tokens_seen": 18040392, "step": 27575 }, { "epoch": 14.454926624737945, "grad_norm": 0.1445191502571106, "learning_rate": 1.0824017125263555e-05, "loss": 0.6183, "num_input_tokens_seen": 18043464, "step": 27580 }, { "epoch": 14.45754716981132, "grad_norm": 0.12114197760820389, "learning_rate": 1.0814600293840154e-05, "loss": 0.4953, "num_input_tokens_seen": 18046792, "step": 27585 }, { "epoch": 14.460167714884696, "grad_norm": 0.1502271294593811, "learning_rate": 1.080518642984055e-05, "loss": 0.3817, "num_input_tokens_seen": 18050024, "step": 27590 }, { "epoch": 14.46278825995807, "grad_norm": 0.1623488813638687, "learning_rate": 1.0795775535233988e-05, "loss": 0.5095, "num_input_tokens_seen": 18055624, "step": 27595 }, { "epoch": 14.465408805031446, "grad_norm": 0.16321757435798645, "learning_rate": 1.0786367611989145e-05, "loss": 0.323, "num_input_tokens_seen": 18060712, "step": 27600 }, { "epoch": 14.468029350104821, "grad_norm": 0.13951082527637482, "learning_rate": 1.0776962662074031e-05, "loss": 0.6067, "num_input_tokens_seen": 18063176, "step": 27605 }, { "epoch": 14.470649895178196, "grad_norm": 0.15659905970096588, "learning_rate": 1.0767560687456062e-05, "loss": 0.4726, "num_input_tokens_seen": 18066696, "step": 27610 }, { "epoch": 14.473270440251572, "grad_norm": 0.1373632848262787, "learning_rate": 1.0758161690102036e-05, "loss": 0.4631, "num_input_tokens_seen": 18069640, "step": 27615 }, { "epoch": 14.475890985324948, "grad_norm": 0.23267684876918793, "learning_rate": 1.0748765671978095e-05, "loss": 0.4235, "num_input_tokens_seen": 18072232, "step": 27620 }, { "epoch": 14.478511530398324, "grad_norm": 0.22222800552845, "learning_rate": 1.0739372635049797e-05, "loss": 0.4598, "num_input_tokens_seen": 18074760, "step": 27625 }, { "epoch": 14.481132075471699, "grad_norm": 0.08575745671987534, "learning_rate": 1.0729982581282042e-05, "loss": 0.3354, "num_input_tokens_seen": 18078056, "step": 27630 }, { "epoch": 14.483752620545074, "grad_norm": 0.18053412437438965, "learning_rate": 1.0720595512639129e-05, "loss": 0.4817, "num_input_tokens_seen": 18081672, "step": 27635 }, { "epoch": 14.48637316561845, "grad_norm": 0.16131512820720673, "learning_rate": 1.0711211431084728e-05, "loss": 0.4792, "num_input_tokens_seen": 18085736, "step": 27640 }, { "epoch": 14.488993710691824, "grad_norm": 0.16787129640579224, "learning_rate": 1.0701830338581887e-05, "loss": 0.493, "num_input_tokens_seen": 18089608, "step": 27645 }, { "epoch": 14.4916142557652, "grad_norm": 0.28201350569725037, "learning_rate": 1.069245223709301e-05, "loss": 0.7042, "num_input_tokens_seen": 18092424, "step": 27650 }, { "epoch": 14.494234800838575, "grad_norm": 0.12220361083745956, "learning_rate": 1.0683077128579904e-05, "loss": 0.5718, "num_input_tokens_seen": 18095848, "step": 27655 }, { "epoch": 14.49685534591195, "grad_norm": 0.18190898001194, "learning_rate": 1.0673705015003718e-05, "loss": 0.5019, "num_input_tokens_seen": 18099080, "step": 27660 }, { "epoch": 14.499475890985325, "grad_norm": 0.18292993307113647, "learning_rate": 1.0664335898325002e-05, "loss": 0.4428, "num_input_tokens_seen": 18101640, "step": 27665 }, { "epoch": 14.5020964360587, "grad_norm": 0.17437227070331573, "learning_rate": 1.065496978050367e-05, "loss": 0.4139, "num_input_tokens_seen": 18104840, "step": 27670 }, { "epoch": 14.504716981132075, "grad_norm": 0.2699746787548065, "learning_rate": 1.0645606663499e-05, "loss": 0.4673, "num_input_tokens_seen": 18107240, "step": 27675 }, { "epoch": 14.50733752620545, "grad_norm": 0.20652689039707184, "learning_rate": 1.0636246549269649e-05, "loss": 0.5906, "num_input_tokens_seen": 18110664, "step": 27680 }, { "epoch": 14.509958071278826, "grad_norm": 0.19455301761627197, "learning_rate": 1.062688943977366e-05, "loss": 0.5126, "num_input_tokens_seen": 18114120, "step": 27685 }, { "epoch": 14.5125786163522, "grad_norm": 0.1952134668827057, "learning_rate": 1.0617535336968413e-05, "loss": 0.5251, "num_input_tokens_seen": 18117032, "step": 27690 }, { "epoch": 14.515199161425576, "grad_norm": 0.14226070046424866, "learning_rate": 1.0608184242810698e-05, "loss": 0.4796, "num_input_tokens_seen": 18120072, "step": 27695 }, { "epoch": 14.517819706498951, "grad_norm": 0.4423946440219879, "learning_rate": 1.0598836159256642e-05, "loss": 0.6365, "num_input_tokens_seen": 18122824, "step": 27700 }, { "epoch": 14.520440251572326, "grad_norm": 0.13067330420017242, "learning_rate": 1.0589491088261763e-05, "loss": 0.4093, "num_input_tokens_seen": 18126120, "step": 27705 }, { "epoch": 14.523060796645701, "grad_norm": 0.15592293441295624, "learning_rate": 1.0580149031780945e-05, "loss": 0.4877, "num_input_tokens_seen": 18129320, "step": 27710 }, { "epoch": 14.525681341719078, "grad_norm": 0.2102329283952713, "learning_rate": 1.0570809991768447e-05, "loss": 0.4856, "num_input_tokens_seen": 18132648, "step": 27715 }, { "epoch": 14.528301886792454, "grad_norm": 0.1510329395532608, "learning_rate": 1.0561473970177873e-05, "loss": 0.4384, "num_input_tokens_seen": 18135432, "step": 27720 }, { "epoch": 14.530922431865829, "grad_norm": 0.25606104731559753, "learning_rate": 1.0552140968962232e-05, "loss": 0.5074, "num_input_tokens_seen": 18138536, "step": 27725 }, { "epoch": 14.533542976939204, "grad_norm": 0.27313828468322754, "learning_rate": 1.054281099007386e-05, "loss": 0.3619, "num_input_tokens_seen": 18142568, "step": 27730 }, { "epoch": 14.536163522012579, "grad_norm": 0.1425168812274933, "learning_rate": 1.0533484035464492e-05, "loss": 0.5133, "num_input_tokens_seen": 18145352, "step": 27735 }, { "epoch": 14.538784067085954, "grad_norm": 0.15172964334487915, "learning_rate": 1.0524160107085231e-05, "loss": 0.4944, "num_input_tokens_seen": 18152584, "step": 27740 }, { "epoch": 14.54140461215933, "grad_norm": 0.2899590730667114, "learning_rate": 1.0514839206886518e-05, "loss": 0.5008, "num_input_tokens_seen": 18155624, "step": 27745 }, { "epoch": 14.544025157232705, "grad_norm": 0.23724429309368134, "learning_rate": 1.0505521336818195e-05, "loss": 0.4352, "num_input_tokens_seen": 18158728, "step": 27750 }, { "epoch": 14.54664570230608, "grad_norm": 0.13819442689418793, "learning_rate": 1.0496206498829438e-05, "loss": 0.4173, "num_input_tokens_seen": 18161512, "step": 27755 }, { "epoch": 14.549266247379455, "grad_norm": 0.14145344495773315, "learning_rate": 1.0486894694868814e-05, "loss": 0.532, "num_input_tokens_seen": 18164616, "step": 27760 }, { "epoch": 14.55188679245283, "grad_norm": 0.2640262246131897, "learning_rate": 1.0477585926884246e-05, "loss": 0.5714, "num_input_tokens_seen": 18167080, "step": 27765 }, { "epoch": 14.554507337526205, "grad_norm": 0.09494990855455399, "learning_rate": 1.046828019682303e-05, "loss": 0.4449, "num_input_tokens_seen": 18170792, "step": 27770 }, { "epoch": 14.55712788259958, "grad_norm": 0.3952247202396393, "learning_rate": 1.0458977506631808e-05, "loss": 0.5137, "num_input_tokens_seen": 18173960, "step": 27775 }, { "epoch": 14.559748427672956, "grad_norm": 0.11331541836261749, "learning_rate": 1.0449677858256593e-05, "loss": 0.499, "num_input_tokens_seen": 18177864, "step": 27780 }, { "epoch": 14.56236897274633, "grad_norm": 0.16721795499324799, "learning_rate": 1.0440381253642773e-05, "loss": 0.5264, "num_input_tokens_seen": 18181672, "step": 27785 }, { "epoch": 14.564989517819706, "grad_norm": 0.16490767896175385, "learning_rate": 1.0431087694735087e-05, "loss": 0.4047, "num_input_tokens_seen": 18185000, "step": 27790 }, { "epoch": 14.567610062893081, "grad_norm": 0.11474040150642395, "learning_rate": 1.0421797183477644e-05, "loss": 0.4378, "num_input_tokens_seen": 18188104, "step": 27795 }, { "epoch": 14.570230607966456, "grad_norm": 0.13209177553653717, "learning_rate": 1.041250972181392e-05, "loss": 0.5388, "num_input_tokens_seen": 18192872, "step": 27800 }, { "epoch": 14.572851153039831, "grad_norm": 0.1841621845960617, "learning_rate": 1.0403225311686743e-05, "loss": 0.4962, "num_input_tokens_seen": 18196040, "step": 27805 }, { "epoch": 14.575471698113208, "grad_norm": 0.11995605379343033, "learning_rate": 1.0393943955038293e-05, "loss": 0.4323, "num_input_tokens_seen": 18199592, "step": 27810 }, { "epoch": 14.578092243186584, "grad_norm": 0.11204858869314194, "learning_rate": 1.038466565381013e-05, "loss": 0.5106, "num_input_tokens_seen": 18203016, "step": 27815 }, { "epoch": 14.580712788259959, "grad_norm": 0.1164533942937851, "learning_rate": 1.0375390409943175e-05, "loss": 0.471, "num_input_tokens_seen": 18206152, "step": 27820 }, { "epoch": 14.583333333333334, "grad_norm": 0.13831716775894165, "learning_rate": 1.0366118225377708e-05, "loss": 0.3925, "num_input_tokens_seen": 18209096, "step": 27825 }, { "epoch": 14.585953878406709, "grad_norm": 0.12095804512500763, "learning_rate": 1.0356849102053353e-05, "loss": 0.4411, "num_input_tokens_seen": 18212680, "step": 27830 }, { "epoch": 14.588574423480084, "grad_norm": 0.16068145632743835, "learning_rate": 1.0347583041909104e-05, "loss": 0.41, "num_input_tokens_seen": 18215720, "step": 27835 }, { "epoch": 14.59119496855346, "grad_norm": 0.15337540209293365, "learning_rate": 1.0338320046883315e-05, "loss": 0.4605, "num_input_tokens_seen": 18218888, "step": 27840 }, { "epoch": 14.593815513626835, "grad_norm": 0.2661544978618622, "learning_rate": 1.0329060118913705e-05, "loss": 0.4214, "num_input_tokens_seen": 18221768, "step": 27845 }, { "epoch": 14.59643605870021, "grad_norm": 0.26111260056495667, "learning_rate": 1.0319803259937353e-05, "loss": 0.443, "num_input_tokens_seen": 18225640, "step": 27850 }, { "epoch": 14.599056603773585, "grad_norm": 0.09376541525125504, "learning_rate": 1.0310549471890676e-05, "loss": 0.5359, "num_input_tokens_seen": 18229928, "step": 27855 }, { "epoch": 14.60167714884696, "grad_norm": 0.21328787505626678, "learning_rate": 1.0301298756709454e-05, "loss": 0.5904, "num_input_tokens_seen": 18232840, "step": 27860 }, { "epoch": 14.604297693920335, "grad_norm": 0.2079644650220871, "learning_rate": 1.0292051116328844e-05, "loss": 0.4727, "num_input_tokens_seen": 18236200, "step": 27865 }, { "epoch": 14.60691823899371, "grad_norm": 0.19182194769382477, "learning_rate": 1.028280655268334e-05, "loss": 0.4943, "num_input_tokens_seen": 18240072, "step": 27870 }, { "epoch": 14.609538784067086, "grad_norm": 0.15687738358974457, "learning_rate": 1.0273565067706806e-05, "loss": 0.4999, "num_input_tokens_seen": 18243464, "step": 27875 }, { "epoch": 14.61215932914046, "grad_norm": 0.12450572103261948, "learning_rate": 1.0264326663332457e-05, "loss": 0.5612, "num_input_tokens_seen": 18246088, "step": 27880 }, { "epoch": 14.614779874213836, "grad_norm": 0.2109750658273697, "learning_rate": 1.0255091341492859e-05, "loss": 0.4811, "num_input_tokens_seen": 18249512, "step": 27885 }, { "epoch": 14.617400419287211, "grad_norm": 0.13081444799900055, "learning_rate": 1.0245859104119923e-05, "loss": 0.5261, "num_input_tokens_seen": 18252328, "step": 27890 }, { "epoch": 14.620020964360586, "grad_norm": 0.19355599582195282, "learning_rate": 1.0236629953144938e-05, "loss": 0.4037, "num_input_tokens_seen": 18256168, "step": 27895 }, { "epoch": 14.622641509433961, "grad_norm": 0.20335689187049866, "learning_rate": 1.0227403890498537e-05, "loss": 0.6068, "num_input_tokens_seen": 18259208, "step": 27900 }, { "epoch": 14.625262054507338, "grad_norm": 0.21277174353599548, "learning_rate": 1.021818091811072e-05, "loss": 0.478, "num_input_tokens_seen": 18261992, "step": 27905 }, { "epoch": 14.627882599580714, "grad_norm": 0.15030857920646667, "learning_rate": 1.0208961037910813e-05, "loss": 0.5014, "num_input_tokens_seen": 18265384, "step": 27910 }, { "epoch": 14.630503144654089, "grad_norm": 0.11176320910453796, "learning_rate": 1.0199744251827503e-05, "loss": 0.5038, "num_input_tokens_seen": 18268008, "step": 27915 }, { "epoch": 14.633123689727464, "grad_norm": 0.15949422121047974, "learning_rate": 1.0190530561788845e-05, "loss": 0.6034, "num_input_tokens_seen": 18270920, "step": 27920 }, { "epoch": 14.635744234800839, "grad_norm": 0.17041969299316406, "learning_rate": 1.0181319969722236e-05, "loss": 0.3252, "num_input_tokens_seen": 18275112, "step": 27925 }, { "epoch": 14.638364779874214, "grad_norm": 0.16007989645004272, "learning_rate": 1.0172112477554441e-05, "loss": 0.3638, "num_input_tokens_seen": 18278888, "step": 27930 }, { "epoch": 14.64098532494759, "grad_norm": 0.1489357054233551, "learning_rate": 1.016290808721154e-05, "loss": 0.4148, "num_input_tokens_seen": 18281768, "step": 27935 }, { "epoch": 14.643605870020965, "grad_norm": 0.1917947232723236, "learning_rate": 1.0153706800619007e-05, "loss": 0.4895, "num_input_tokens_seen": 18284360, "step": 27940 }, { "epoch": 14.64622641509434, "grad_norm": 0.20506048202514648, "learning_rate": 1.0144508619701623e-05, "loss": 0.5314, "num_input_tokens_seen": 18287048, "step": 27945 }, { "epoch": 14.648846960167715, "grad_norm": 0.20065994560718536, "learning_rate": 1.013531354638356e-05, "loss": 0.6103, "num_input_tokens_seen": 18289672, "step": 27950 }, { "epoch": 14.65146750524109, "grad_norm": 0.18182623386383057, "learning_rate": 1.0126121582588316e-05, "loss": 0.5928, "num_input_tokens_seen": 18293096, "step": 27955 }, { "epoch": 14.654088050314465, "grad_norm": 0.0784686952829361, "learning_rate": 1.0116932730238754e-05, "loss": 0.3697, "num_input_tokens_seen": 18297608, "step": 27960 }, { "epoch": 14.65670859538784, "grad_norm": 0.1364201307296753, "learning_rate": 1.0107746991257063e-05, "loss": 0.4638, "num_input_tokens_seen": 18301352, "step": 27965 }, { "epoch": 14.659329140461216, "grad_norm": 0.17464827001094818, "learning_rate": 1.0098564367564811e-05, "loss": 0.4121, "num_input_tokens_seen": 18304552, "step": 27970 }, { "epoch": 14.66194968553459, "grad_norm": 0.09933681786060333, "learning_rate": 1.0089384861082876e-05, "loss": 0.375, "num_input_tokens_seen": 18307400, "step": 27975 }, { "epoch": 14.664570230607966, "grad_norm": 0.5487833023071289, "learning_rate": 1.0080208473731523e-05, "loss": 0.856, "num_input_tokens_seen": 18311432, "step": 27980 }, { "epoch": 14.667190775681341, "grad_norm": 0.20290376245975494, "learning_rate": 1.0071035207430352e-05, "loss": 0.4223, "num_input_tokens_seen": 18315720, "step": 27985 }, { "epoch": 14.669811320754716, "grad_norm": 0.13285116851329803, "learning_rate": 1.0061865064098285e-05, "loss": 0.4722, "num_input_tokens_seen": 18318952, "step": 27990 }, { "epoch": 14.672431865828091, "grad_norm": 0.44050323963165283, "learning_rate": 1.0052698045653625e-05, "loss": 0.5061, "num_input_tokens_seen": 18322856, "step": 27995 }, { "epoch": 14.675052410901468, "grad_norm": 0.13365893065929413, "learning_rate": 1.0043534154014015e-05, "loss": 0.4053, "num_input_tokens_seen": 18326248, "step": 28000 }, { "epoch": 14.677672955974844, "grad_norm": 0.19956758618354797, "learning_rate": 1.003437339109642e-05, "loss": 0.5568, "num_input_tokens_seen": 18329704, "step": 28005 }, { "epoch": 14.680293501048219, "grad_norm": 0.18274223804473877, "learning_rate": 1.0025215758817183e-05, "loss": 0.5303, "num_input_tokens_seen": 18333160, "step": 28010 }, { "epoch": 14.682914046121594, "grad_norm": 0.31236085295677185, "learning_rate": 1.0016061259091958e-05, "loss": 0.5846, "num_input_tokens_seen": 18336232, "step": 28015 }, { "epoch": 14.685534591194969, "grad_norm": 0.27414077520370483, "learning_rate": 1.000690989383577e-05, "loss": 0.4014, "num_input_tokens_seen": 18339304, "step": 28020 }, { "epoch": 14.688155136268344, "grad_norm": 0.14290133118629456, "learning_rate": 9.997761664962993e-06, "loss": 0.4608, "num_input_tokens_seen": 18342536, "step": 28025 }, { "epoch": 14.69077568134172, "grad_norm": 0.21430321037769318, "learning_rate": 9.988616574387314e-06, "loss": 0.4549, "num_input_tokens_seen": 18345448, "step": 28030 }, { "epoch": 14.693396226415095, "grad_norm": 0.1772220879793167, "learning_rate": 9.979474624021787e-06, "loss": 0.4129, "num_input_tokens_seen": 18348296, "step": 28035 }, { "epoch": 14.69601677148847, "grad_norm": 0.2142249047756195, "learning_rate": 9.970335815778815e-06, "loss": 0.5333, "num_input_tokens_seen": 18351656, "step": 28040 }, { "epoch": 14.698637316561845, "grad_norm": 0.18360552191734314, "learning_rate": 9.96120015157011e-06, "loss": 0.5207, "num_input_tokens_seen": 18355016, "step": 28045 }, { "epoch": 14.70125786163522, "grad_norm": 0.10873500257730484, "learning_rate": 9.952067633306767e-06, "loss": 0.5679, "num_input_tokens_seen": 18357800, "step": 28050 }, { "epoch": 14.703878406708595, "grad_norm": 0.22976654767990112, "learning_rate": 9.942938262899204e-06, "loss": 0.6195, "num_input_tokens_seen": 18360904, "step": 28055 }, { "epoch": 14.70649895178197, "grad_norm": 0.16568371653556824, "learning_rate": 9.93381204225717e-06, "loss": 0.5227, "num_input_tokens_seen": 18363528, "step": 28060 }, { "epoch": 14.709119496855346, "grad_norm": 0.10699297487735748, "learning_rate": 9.924688973289782e-06, "loss": 0.4415, "num_input_tokens_seen": 18367688, "step": 28065 }, { "epoch": 14.71174004192872, "grad_norm": 0.11054772138595581, "learning_rate": 9.915569057905463e-06, "loss": 0.5632, "num_input_tokens_seen": 18370824, "step": 28070 }, { "epoch": 14.714360587002096, "grad_norm": 0.13955634832382202, "learning_rate": 9.906452298012003e-06, "loss": 0.4627, "num_input_tokens_seen": 18373320, "step": 28075 }, { "epoch": 14.716981132075471, "grad_norm": 0.2850915491580963, "learning_rate": 9.897338695516529e-06, "loss": 0.576, "num_input_tokens_seen": 18376360, "step": 28080 }, { "epoch": 14.719601677148846, "grad_norm": 0.12340999394655228, "learning_rate": 9.888228252325507e-06, "loss": 0.5431, "num_input_tokens_seen": 18382984, "step": 28085 }, { "epoch": 14.722222222222221, "grad_norm": 0.14902114868164062, "learning_rate": 9.87912097034473e-06, "loss": 0.4371, "num_input_tokens_seen": 18387016, "step": 28090 }, { "epoch": 14.724842767295598, "grad_norm": 0.21384938061237335, "learning_rate": 9.870016851479327e-06, "loss": 0.4621, "num_input_tokens_seen": 18390120, "step": 28095 }, { "epoch": 14.727463312368974, "grad_norm": 0.1653827279806137, "learning_rate": 9.860915897633791e-06, "loss": 0.5092, "num_input_tokens_seen": 18392936, "step": 28100 }, { "epoch": 14.730083857442349, "grad_norm": 0.17303495109081268, "learning_rate": 9.85181811071193e-06, "loss": 0.5346, "num_input_tokens_seen": 18395720, "step": 28105 }, { "epoch": 14.732704402515724, "grad_norm": 0.1352793276309967, "learning_rate": 9.842723492616913e-06, "loss": 0.4137, "num_input_tokens_seen": 18398312, "step": 28110 }, { "epoch": 14.735324947589099, "grad_norm": 0.17849455773830414, "learning_rate": 9.833632045251207e-06, "loss": 0.4523, "num_input_tokens_seen": 18401608, "step": 28115 }, { "epoch": 14.737945492662474, "grad_norm": 0.20107366144657135, "learning_rate": 9.824543770516656e-06, "loss": 0.5411, "num_input_tokens_seen": 18404424, "step": 28120 }, { "epoch": 14.74056603773585, "grad_norm": 0.10359369963407516, "learning_rate": 9.815458670314412e-06, "loss": 0.4388, "num_input_tokens_seen": 18408040, "step": 28125 }, { "epoch": 14.743186582809225, "grad_norm": 0.13992464542388916, "learning_rate": 9.80637674654498e-06, "loss": 0.5111, "num_input_tokens_seen": 18410792, "step": 28130 }, { "epoch": 14.7458071278826, "grad_norm": 0.27091532945632935, "learning_rate": 9.797298001108193e-06, "loss": 0.4866, "num_input_tokens_seen": 18413320, "step": 28135 }, { "epoch": 14.748427672955975, "grad_norm": 0.2827821373939514, "learning_rate": 9.788222435903232e-06, "loss": 0.4596, "num_input_tokens_seen": 18416200, "step": 28140 }, { "epoch": 14.75104821802935, "grad_norm": 0.3584716022014618, "learning_rate": 9.779150052828596e-06, "loss": 0.652, "num_input_tokens_seen": 18418760, "step": 28145 }, { "epoch": 14.753668763102725, "grad_norm": 0.2671789824962616, "learning_rate": 9.770080853782108e-06, "loss": 0.5954, "num_input_tokens_seen": 18421832, "step": 28150 }, { "epoch": 14.7562893081761, "grad_norm": 0.1369187980890274, "learning_rate": 9.76101484066095e-06, "loss": 0.3946, "num_input_tokens_seen": 18425608, "step": 28155 }, { "epoch": 14.758909853249476, "grad_norm": 0.16934213042259216, "learning_rate": 9.751952015361637e-06, "loss": 0.466, "num_input_tokens_seen": 18429256, "step": 28160 }, { "epoch": 14.76153039832285, "grad_norm": 0.2388768196105957, "learning_rate": 9.742892379780003e-06, "loss": 0.4479, "num_input_tokens_seen": 18434600, "step": 28165 }, { "epoch": 14.764150943396226, "grad_norm": 0.30334949493408203, "learning_rate": 9.733835935811237e-06, "loss": 0.4811, "num_input_tokens_seen": 18437416, "step": 28170 }, { "epoch": 14.766771488469601, "grad_norm": 0.16502873599529266, "learning_rate": 9.724782685349809e-06, "loss": 0.4549, "num_input_tokens_seen": 18440776, "step": 28175 }, { "epoch": 14.769392033542976, "grad_norm": 0.13132865726947784, "learning_rate": 9.715732630289576e-06, "loss": 0.3663, "num_input_tokens_seen": 18443528, "step": 28180 }, { "epoch": 14.772012578616351, "grad_norm": 0.1730416715145111, "learning_rate": 9.706685772523701e-06, "loss": 0.4346, "num_input_tokens_seen": 18446504, "step": 28185 }, { "epoch": 14.774633123689728, "grad_norm": 0.1153404712677002, "learning_rate": 9.697642113944692e-06, "loss": 0.4601, "num_input_tokens_seen": 18449864, "step": 28190 }, { "epoch": 14.777253668763104, "grad_norm": 0.1300610452890396, "learning_rate": 9.688601656444381e-06, "loss": 0.4431, "num_input_tokens_seen": 18455816, "step": 28195 }, { "epoch": 14.779874213836479, "grad_norm": 0.1786344051361084, "learning_rate": 9.679564401913921e-06, "loss": 0.4104, "num_input_tokens_seen": 18459144, "step": 28200 }, { "epoch": 14.782494758909854, "grad_norm": 0.16830338537693024, "learning_rate": 9.670530352243798e-06, "loss": 0.5398, "num_input_tokens_seen": 18462056, "step": 28205 }, { "epoch": 14.785115303983229, "grad_norm": 0.15832439064979553, "learning_rate": 9.661499509323835e-06, "loss": 0.5983, "num_input_tokens_seen": 18465352, "step": 28210 }, { "epoch": 14.787735849056604, "grad_norm": 0.15460701286792755, "learning_rate": 9.652471875043182e-06, "loss": 0.5944, "num_input_tokens_seen": 18468776, "step": 28215 }, { "epoch": 14.79035639412998, "grad_norm": 0.10834750533103943, "learning_rate": 9.64344745129033e-06, "loss": 0.5409, "num_input_tokens_seen": 18472296, "step": 28220 }, { "epoch": 14.792976939203355, "grad_norm": 0.14413104951381683, "learning_rate": 9.634426239953073e-06, "loss": 0.5612, "num_input_tokens_seen": 18476424, "step": 28225 }, { "epoch": 14.79559748427673, "grad_norm": 0.14943070709705353, "learning_rate": 9.62540824291854e-06, "loss": 0.4287, "num_input_tokens_seen": 18479912, "step": 28230 }, { "epoch": 14.798218029350105, "grad_norm": 0.12894709408283234, "learning_rate": 9.616393462073197e-06, "loss": 0.3949, "num_input_tokens_seen": 18482632, "step": 28235 }, { "epoch": 14.80083857442348, "grad_norm": 0.14034225046634674, "learning_rate": 9.607381899302839e-06, "loss": 0.4289, "num_input_tokens_seen": 18485928, "step": 28240 }, { "epoch": 14.803459119496855, "grad_norm": 0.1307535320520401, "learning_rate": 9.59837355649258e-06, "loss": 0.5454, "num_input_tokens_seen": 18489768, "step": 28245 }, { "epoch": 14.80607966457023, "grad_norm": 0.29909345507621765, "learning_rate": 9.589368435526868e-06, "loss": 0.5013, "num_input_tokens_seen": 18492968, "step": 28250 }, { "epoch": 14.808700209643606, "grad_norm": 0.21395769715309143, "learning_rate": 9.580366538289468e-06, "loss": 0.5781, "num_input_tokens_seen": 18496136, "step": 28255 }, { "epoch": 14.81132075471698, "grad_norm": 0.12419603019952774, "learning_rate": 9.57136786666346e-06, "loss": 0.4489, "num_input_tokens_seen": 18499880, "step": 28260 }, { "epoch": 14.813941299790356, "grad_norm": 0.13131685554981232, "learning_rate": 9.562372422531276e-06, "loss": 0.5816, "num_input_tokens_seen": 18504488, "step": 28265 }, { "epoch": 14.816561844863731, "grad_norm": 0.1142895296216011, "learning_rate": 9.553380207774662e-06, "loss": 0.4118, "num_input_tokens_seen": 18507304, "step": 28270 }, { "epoch": 14.819182389937106, "grad_norm": 0.19730806350708008, "learning_rate": 9.54439122427469e-06, "loss": 0.4564, "num_input_tokens_seen": 18510152, "step": 28275 }, { "epoch": 14.821802935010481, "grad_norm": 0.16736039519309998, "learning_rate": 9.535405473911748e-06, "loss": 0.5273, "num_input_tokens_seen": 18513960, "step": 28280 }, { "epoch": 14.824423480083858, "grad_norm": 0.1818382292985916, "learning_rate": 9.526422958565546e-06, "loss": 0.6132, "num_input_tokens_seen": 18516520, "step": 28285 }, { "epoch": 14.827044025157234, "grad_norm": 0.19267447292804718, "learning_rate": 9.517443680115124e-06, "loss": 0.4271, "num_input_tokens_seen": 18519336, "step": 28290 }, { "epoch": 14.829664570230609, "grad_norm": 0.2571471631526947, "learning_rate": 9.50846764043885e-06, "loss": 0.5094, "num_input_tokens_seen": 18522024, "step": 28295 }, { "epoch": 14.832285115303984, "grad_norm": 0.16739094257354736, "learning_rate": 9.499494841414422e-06, "loss": 0.6277, "num_input_tokens_seen": 18526504, "step": 28300 }, { "epoch": 14.834905660377359, "grad_norm": 0.12937992811203003, "learning_rate": 9.490525284918822e-06, "loss": 0.4582, "num_input_tokens_seen": 18529736, "step": 28305 }, { "epoch": 14.837526205450734, "grad_norm": 0.17696869373321533, "learning_rate": 9.4815589728284e-06, "loss": 0.4809, "num_input_tokens_seen": 18532648, "step": 28310 }, { "epoch": 14.84014675052411, "grad_norm": 0.21461214125156403, "learning_rate": 9.472595907018788e-06, "loss": 0.5027, "num_input_tokens_seen": 18535112, "step": 28315 }, { "epoch": 14.842767295597485, "grad_norm": 0.1368129849433899, "learning_rate": 9.463636089364968e-06, "loss": 0.5048, "num_input_tokens_seen": 18538376, "step": 28320 }, { "epoch": 14.84538784067086, "grad_norm": 0.1317794770002365, "learning_rate": 9.454679521741239e-06, "loss": 0.5255, "num_input_tokens_seen": 18541864, "step": 28325 }, { "epoch": 14.848008385744235, "grad_norm": 0.14575760066509247, "learning_rate": 9.445726206021194e-06, "loss": 0.5382, "num_input_tokens_seen": 18545320, "step": 28330 }, { "epoch": 14.85062893081761, "grad_norm": 0.21895833313465118, "learning_rate": 9.436776144077772e-06, "loss": 0.5267, "num_input_tokens_seen": 18547720, "step": 28335 }, { "epoch": 14.853249475890985, "grad_norm": 0.19868077337741852, "learning_rate": 9.427829337783237e-06, "loss": 0.505, "num_input_tokens_seen": 18551880, "step": 28340 }, { "epoch": 14.85587002096436, "grad_norm": 0.1626415103673935, "learning_rate": 9.418885789009138e-06, "loss": 0.3174, "num_input_tokens_seen": 18554568, "step": 28345 }, { "epoch": 14.858490566037736, "grad_norm": 0.13165664672851562, "learning_rate": 9.409945499626374e-06, "loss": 0.5652, "num_input_tokens_seen": 18557448, "step": 28350 }, { "epoch": 14.86111111111111, "grad_norm": 0.3379449248313904, "learning_rate": 9.40100847150516e-06, "loss": 0.5334, "num_input_tokens_seen": 18559656, "step": 28355 }, { "epoch": 14.863731656184486, "grad_norm": 0.12499475479125977, "learning_rate": 9.392074706515003e-06, "loss": 0.463, "num_input_tokens_seen": 18563752, "step": 28360 }, { "epoch": 14.866352201257861, "grad_norm": 0.18233761191368103, "learning_rate": 9.383144206524753e-06, "loss": 0.5376, "num_input_tokens_seen": 18566696, "step": 28365 }, { "epoch": 14.868972746331236, "grad_norm": 0.14045356214046478, "learning_rate": 9.374216973402575e-06, "loss": 0.512, "num_input_tokens_seen": 18570280, "step": 28370 }, { "epoch": 14.871593291404611, "grad_norm": 0.13896797597408295, "learning_rate": 9.365293009015926e-06, "loss": 0.4032, "num_input_tokens_seen": 18573416, "step": 28375 }, { "epoch": 14.874213836477988, "grad_norm": 0.15480411052703857, "learning_rate": 9.356372315231623e-06, "loss": 0.3381, "num_input_tokens_seen": 18576712, "step": 28380 }, { "epoch": 14.876834381551364, "grad_norm": 0.2024257779121399, "learning_rate": 9.347454893915749e-06, "loss": 0.4965, "num_input_tokens_seen": 18579912, "step": 28385 }, { "epoch": 14.879454926624739, "grad_norm": 0.2769528329372406, "learning_rate": 9.338540746933741e-06, "loss": 0.5174, "num_input_tokens_seen": 18582664, "step": 28390 }, { "epoch": 14.882075471698114, "grad_norm": 0.18601690232753754, "learning_rate": 9.329629876150339e-06, "loss": 0.4749, "num_input_tokens_seen": 18585704, "step": 28395 }, { "epoch": 14.884696016771489, "grad_norm": 0.2186938226222992, "learning_rate": 9.320722283429581e-06, "loss": 0.3321, "num_input_tokens_seen": 18588456, "step": 28400 }, { "epoch": 14.887316561844864, "grad_norm": 0.26600518822669983, "learning_rate": 9.311817970634854e-06, "loss": 0.426, "num_input_tokens_seen": 18591848, "step": 28405 }, { "epoch": 14.88993710691824, "grad_norm": 0.27303194999694824, "learning_rate": 9.302916939628814e-06, "loss": 0.6383, "num_input_tokens_seen": 18594440, "step": 28410 }, { "epoch": 14.892557651991615, "grad_norm": 0.16487662494182587, "learning_rate": 9.29401919227347e-06, "loss": 0.5514, "num_input_tokens_seen": 18597960, "step": 28415 }, { "epoch": 14.89517819706499, "grad_norm": 0.13811026513576508, "learning_rate": 9.28512473043013e-06, "loss": 0.5982, "num_input_tokens_seen": 18600712, "step": 28420 }, { "epoch": 14.897798742138365, "grad_norm": 0.19151049852371216, "learning_rate": 9.276233555959419e-06, "loss": 0.4508, "num_input_tokens_seen": 18603944, "step": 28425 }, { "epoch": 14.90041928721174, "grad_norm": 0.16326549649238586, "learning_rate": 9.267345670721254e-06, "loss": 0.4405, "num_input_tokens_seen": 18607016, "step": 28430 }, { "epoch": 14.903039832285115, "grad_norm": 0.23110586404800415, "learning_rate": 9.258461076574895e-06, "loss": 0.5915, "num_input_tokens_seen": 18609832, "step": 28435 }, { "epoch": 14.90566037735849, "grad_norm": 0.118589848279953, "learning_rate": 9.24957977537888e-06, "loss": 0.354, "num_input_tokens_seen": 18613064, "step": 28440 }, { "epoch": 14.908280922431866, "grad_norm": 0.14714573323726654, "learning_rate": 9.240701768991086e-06, "loss": 0.4905, "num_input_tokens_seen": 18616648, "step": 28445 }, { "epoch": 14.91090146750524, "grad_norm": 0.1313396543264389, "learning_rate": 9.231827059268692e-06, "loss": 0.5441, "num_input_tokens_seen": 18619688, "step": 28450 }, { "epoch": 14.913522012578616, "grad_norm": 0.18296989798545837, "learning_rate": 9.222955648068193e-06, "loss": 0.5181, "num_input_tokens_seen": 18622568, "step": 28455 }, { "epoch": 14.916142557651991, "grad_norm": 0.16389068961143494, "learning_rate": 9.214087537245378e-06, "loss": 0.4413, "num_input_tokens_seen": 18626216, "step": 28460 }, { "epoch": 14.918763102725366, "grad_norm": 0.21400196850299835, "learning_rate": 9.205222728655344e-06, "loss": 0.5451, "num_input_tokens_seen": 18629512, "step": 28465 }, { "epoch": 14.921383647798741, "grad_norm": 0.15300673246383667, "learning_rate": 9.196361224152522e-06, "loss": 0.435, "num_input_tokens_seen": 18632744, "step": 28470 }, { "epoch": 14.924004192872118, "grad_norm": 0.2617981731891632, "learning_rate": 9.187503025590632e-06, "loss": 0.4519, "num_input_tokens_seen": 18635368, "step": 28475 }, { "epoch": 14.926624737945493, "grad_norm": 0.1815819889307022, "learning_rate": 9.17864813482272e-06, "loss": 0.529, "num_input_tokens_seen": 18638376, "step": 28480 }, { "epoch": 14.929245283018869, "grad_norm": 0.21963247656822205, "learning_rate": 9.169796553701122e-06, "loss": 0.4339, "num_input_tokens_seen": 18641736, "step": 28485 }, { "epoch": 14.931865828092244, "grad_norm": 0.25682660937309265, "learning_rate": 9.160948284077473e-06, "loss": 0.6125, "num_input_tokens_seen": 18644136, "step": 28490 }, { "epoch": 14.934486373165619, "grad_norm": 0.17415741086006165, "learning_rate": 9.152103327802738e-06, "loss": 0.4514, "num_input_tokens_seen": 18646504, "step": 28495 }, { "epoch": 14.937106918238994, "grad_norm": 0.3947198688983917, "learning_rate": 9.143261686727189e-06, "loss": 0.4366, "num_input_tokens_seen": 18650536, "step": 28500 }, { "epoch": 14.93972746331237, "grad_norm": 0.11638510227203369, "learning_rate": 9.134423362700392e-06, "loss": 0.5895, "num_input_tokens_seen": 18655944, "step": 28505 }, { "epoch": 14.942348008385745, "grad_norm": 0.15490292012691498, "learning_rate": 9.12558835757123e-06, "loss": 0.3504, "num_input_tokens_seen": 18659496, "step": 28510 }, { "epoch": 14.94496855345912, "grad_norm": 0.1851971596479416, "learning_rate": 9.116756673187878e-06, "loss": 0.6233, "num_input_tokens_seen": 18662152, "step": 28515 }, { "epoch": 14.947589098532495, "grad_norm": 0.5356983542442322, "learning_rate": 9.107928311397817e-06, "loss": 0.571, "num_input_tokens_seen": 18664648, "step": 28520 }, { "epoch": 14.95020964360587, "grad_norm": 0.1081514060497284, "learning_rate": 9.09910327404785e-06, "loss": 0.5384, "num_input_tokens_seen": 18668968, "step": 28525 }, { "epoch": 14.952830188679245, "grad_norm": 0.12446070462465286, "learning_rate": 9.090281562984068e-06, "loss": 0.3672, "num_input_tokens_seen": 18672104, "step": 28530 }, { "epoch": 14.95545073375262, "grad_norm": 0.1667233109474182, "learning_rate": 9.081463180051883e-06, "loss": 0.3858, "num_input_tokens_seen": 18674472, "step": 28535 }, { "epoch": 14.958071278825996, "grad_norm": 0.15906213223934174, "learning_rate": 9.072648127095996e-06, "loss": 0.3973, "num_input_tokens_seen": 18677576, "step": 28540 }, { "epoch": 14.96069182389937, "grad_norm": 0.13153626024723053, "learning_rate": 9.0638364059604e-06, "loss": 0.467, "num_input_tokens_seen": 18681832, "step": 28545 }, { "epoch": 14.963312368972746, "grad_norm": 0.2545868456363678, "learning_rate": 9.055028018488417e-06, "loss": 0.6449, "num_input_tokens_seen": 18684104, "step": 28550 }, { "epoch": 14.965932914046121, "grad_norm": 0.1457047462463379, "learning_rate": 9.046222966522665e-06, "loss": 0.4436, "num_input_tokens_seen": 18687208, "step": 28555 }, { "epoch": 14.968553459119496, "grad_norm": 0.20367197692394257, "learning_rate": 9.037421251905059e-06, "loss": 0.5488, "num_input_tokens_seen": 18689640, "step": 28560 }, { "epoch": 14.971174004192871, "grad_norm": 0.1808795928955078, "learning_rate": 9.028622876476833e-06, "loss": 0.4028, "num_input_tokens_seen": 18692616, "step": 28565 }, { "epoch": 14.973794549266248, "grad_norm": 0.0945696234703064, "learning_rate": 9.019827842078474e-06, "loss": 0.3036, "num_input_tokens_seen": 18696328, "step": 28570 }, { "epoch": 14.976415094339622, "grad_norm": 0.18977488577365875, "learning_rate": 9.011036150549817e-06, "loss": 0.3619, "num_input_tokens_seen": 18699240, "step": 28575 }, { "epoch": 14.979035639412999, "grad_norm": 0.26173290610313416, "learning_rate": 9.002247803729987e-06, "loss": 0.5489, "num_input_tokens_seen": 18701704, "step": 28580 }, { "epoch": 14.981656184486374, "grad_norm": 0.15383173525333405, "learning_rate": 8.993462803457404e-06, "loss": 0.419, "num_input_tokens_seen": 18704936, "step": 28585 }, { "epoch": 14.984276729559749, "grad_norm": 0.31047704815864563, "learning_rate": 8.984681151569799e-06, "loss": 0.5682, "num_input_tokens_seen": 18708808, "step": 28590 }, { "epoch": 14.986897274633124, "grad_norm": 0.20396892726421356, "learning_rate": 8.975902849904183e-06, "loss": 0.4772, "num_input_tokens_seen": 18711880, "step": 28595 }, { "epoch": 14.9895178197065, "grad_norm": 0.14910607039928436, "learning_rate": 8.967127900296874e-06, "loss": 0.4046, "num_input_tokens_seen": 18715304, "step": 28600 }, { "epoch": 14.992138364779874, "grad_norm": 0.19259649515151978, "learning_rate": 8.95835630458349e-06, "loss": 0.4217, "num_input_tokens_seen": 18718600, "step": 28605 }, { "epoch": 14.99475890985325, "grad_norm": 0.16011564433574677, "learning_rate": 8.949588064598958e-06, "loss": 0.5079, "num_input_tokens_seen": 18721256, "step": 28610 }, { "epoch": 14.997379454926625, "grad_norm": 0.16667142510414124, "learning_rate": 8.940823182177499e-06, "loss": 0.4846, "num_input_tokens_seen": 18724008, "step": 28615 }, { "epoch": 15.0, "grad_norm": 0.39940473437309265, "learning_rate": 8.932061659152607e-06, "loss": 0.4814, "num_input_tokens_seen": 18726216, "step": 28620 }, { "epoch": 15.0, "eval_loss": 0.46101847290992737, "eval_runtime": 9.8565, "eval_samples_per_second": 86.035, "eval_steps_per_second": 21.509, "num_input_tokens_seen": 18726216, "step": 28620 }, { "epoch": 15.002620545073375, "grad_norm": 0.1955687403678894, "learning_rate": 8.92330349735711e-06, "loss": 0.4309, "num_input_tokens_seen": 18729384, "step": 28625 }, { "epoch": 15.00524109014675, "grad_norm": 0.1271033138036728, "learning_rate": 8.914548698623104e-06, "loss": 0.4944, "num_input_tokens_seen": 18731944, "step": 28630 }, { "epoch": 15.007861635220126, "grad_norm": 0.22263596951961517, "learning_rate": 8.905797264781995e-06, "loss": 0.4747, "num_input_tokens_seen": 18735432, "step": 28635 }, { "epoch": 15.0104821802935, "grad_norm": 0.1008601039648056, "learning_rate": 8.89704919766449e-06, "loss": 0.4513, "num_input_tokens_seen": 18739208, "step": 28640 }, { "epoch": 15.013102725366876, "grad_norm": 0.195879727602005, "learning_rate": 8.888304499100586e-06, "loss": 0.5983, "num_input_tokens_seen": 18742472, "step": 28645 }, { "epoch": 15.015723270440251, "grad_norm": 0.09416131675243378, "learning_rate": 8.879563170919566e-06, "loss": 0.3531, "num_input_tokens_seen": 18746280, "step": 28650 }, { "epoch": 15.018343815513626, "grad_norm": 0.27263110876083374, "learning_rate": 8.870825214950016e-06, "loss": 0.3612, "num_input_tokens_seen": 18748968, "step": 28655 }, { "epoch": 15.020964360587001, "grad_norm": 0.1865784376859665, "learning_rate": 8.862090633019818e-06, "loss": 0.635, "num_input_tokens_seen": 18752008, "step": 28660 }, { "epoch": 15.023584905660377, "grad_norm": 0.08886892348527908, "learning_rate": 8.853359426956146e-06, "loss": 0.3709, "num_input_tokens_seen": 18756264, "step": 28665 }, { "epoch": 15.026205450733753, "grad_norm": 0.24814465641975403, "learning_rate": 8.844631598585482e-06, "loss": 0.566, "num_input_tokens_seen": 18759560, "step": 28670 }, { "epoch": 15.028825995807129, "grad_norm": 0.2261136919260025, "learning_rate": 8.835907149733569e-06, "loss": 0.381, "num_input_tokens_seen": 18762280, "step": 28675 }, { "epoch": 15.031446540880504, "grad_norm": 0.16098609566688538, "learning_rate": 8.827186082225477e-06, "loss": 0.3654, "num_input_tokens_seen": 18765288, "step": 28680 }, { "epoch": 15.034067085953879, "grad_norm": 0.15309053659439087, "learning_rate": 8.81846839788554e-06, "loss": 0.4713, "num_input_tokens_seen": 18769096, "step": 28685 }, { "epoch": 15.036687631027254, "grad_norm": 0.10358195006847382, "learning_rate": 8.809754098537407e-06, "loss": 0.526, "num_input_tokens_seen": 18771496, "step": 28690 }, { "epoch": 15.03930817610063, "grad_norm": 0.22546492516994476, "learning_rate": 8.801043186004016e-06, "loss": 0.433, "num_input_tokens_seen": 18774856, "step": 28695 }, { "epoch": 15.041928721174004, "grad_norm": 0.1321110725402832, "learning_rate": 8.792335662107575e-06, "loss": 0.6334, "num_input_tokens_seen": 18777576, "step": 28700 }, { "epoch": 15.04454926624738, "grad_norm": 0.20590388774871826, "learning_rate": 8.783631528669605e-06, "loss": 0.5032, "num_input_tokens_seen": 18780008, "step": 28705 }, { "epoch": 15.047169811320755, "grad_norm": 0.11738691478967667, "learning_rate": 8.774930787510924e-06, "loss": 0.5595, "num_input_tokens_seen": 18782984, "step": 28710 }, { "epoch": 15.04979035639413, "grad_norm": 0.13688595592975616, "learning_rate": 8.766233440451608e-06, "loss": 0.4619, "num_input_tokens_seen": 18785800, "step": 28715 }, { "epoch": 15.052410901467505, "grad_norm": 0.1775360107421875, "learning_rate": 8.75753948931105e-06, "loss": 0.407, "num_input_tokens_seen": 18788040, "step": 28720 }, { "epoch": 15.05503144654088, "grad_norm": 0.1853412538766861, "learning_rate": 8.748848935907936e-06, "loss": 0.5106, "num_input_tokens_seen": 18791912, "step": 28725 }, { "epoch": 15.057651991614255, "grad_norm": 0.1143873780965805, "learning_rate": 8.740161782060214e-06, "loss": 0.3634, "num_input_tokens_seen": 18795208, "step": 28730 }, { "epoch": 15.06027253668763, "grad_norm": 0.17624065279960632, "learning_rate": 8.731478029585144e-06, "loss": 0.4457, "num_input_tokens_seen": 18797960, "step": 28735 }, { "epoch": 15.062893081761006, "grad_norm": 0.21835924685001373, "learning_rate": 8.722797680299278e-06, "loss": 0.521, "num_input_tokens_seen": 18801544, "step": 28740 }, { "epoch": 15.065513626834381, "grad_norm": 0.6148537993431091, "learning_rate": 8.71412073601843e-06, "loss": 0.7667, "num_input_tokens_seen": 18804072, "step": 28745 }, { "epoch": 15.068134171907756, "grad_norm": 0.16153831779956818, "learning_rate": 8.705447198557731e-06, "loss": 0.4441, "num_input_tokens_seen": 18806696, "step": 28750 }, { "epoch": 15.070754716981131, "grad_norm": 0.10712689906358719, "learning_rate": 8.696777069731574e-06, "loss": 0.379, "num_input_tokens_seen": 18809864, "step": 28755 }, { "epoch": 15.073375262054507, "grad_norm": 0.15115222334861755, "learning_rate": 8.688110351353654e-06, "loss": 0.554, "num_input_tokens_seen": 18814440, "step": 28760 }, { "epoch": 15.075995807127883, "grad_norm": 0.22568367421627045, "learning_rate": 8.679447045236962e-06, "loss": 0.5481, "num_input_tokens_seen": 18818024, "step": 28765 }, { "epoch": 15.078616352201259, "grad_norm": 0.1470826268196106, "learning_rate": 8.670787153193746e-06, "loss": 0.541, "num_input_tokens_seen": 18822216, "step": 28770 }, { "epoch": 15.081236897274634, "grad_norm": 0.19108504056930542, "learning_rate": 8.662130677035574e-06, "loss": 0.5491, "num_input_tokens_seen": 18824616, "step": 28775 }, { "epoch": 15.083857442348009, "grad_norm": 0.19250686466693878, "learning_rate": 8.65347761857326e-06, "loss": 0.4371, "num_input_tokens_seen": 18827624, "step": 28780 }, { "epoch": 15.086477987421384, "grad_norm": 0.16418644785881042, "learning_rate": 8.64482797961694e-06, "loss": 0.5406, "num_input_tokens_seen": 18830888, "step": 28785 }, { "epoch": 15.08909853249476, "grad_norm": 0.1533607542514801, "learning_rate": 8.636181761976016e-06, "loss": 0.4588, "num_input_tokens_seen": 18833640, "step": 28790 }, { "epoch": 15.091719077568134, "grad_norm": 0.19751819968223572, "learning_rate": 8.62753896745919e-06, "loss": 0.5746, "num_input_tokens_seen": 18836424, "step": 28795 }, { "epoch": 15.09433962264151, "grad_norm": 0.1892959326505661, "learning_rate": 8.618899597874413e-06, "loss": 0.4984, "num_input_tokens_seen": 18839496, "step": 28800 }, { "epoch": 15.096960167714885, "grad_norm": 0.1909046769142151, "learning_rate": 8.610263655028964e-06, "loss": 0.4158, "num_input_tokens_seen": 18841928, "step": 28805 }, { "epoch": 15.09958071278826, "grad_norm": 0.2691115736961365, "learning_rate": 8.601631140729366e-06, "loss": 0.4038, "num_input_tokens_seen": 18844648, "step": 28810 }, { "epoch": 15.102201257861635, "grad_norm": 0.22633202373981476, "learning_rate": 8.593002056781451e-06, "loss": 0.4484, "num_input_tokens_seen": 18847656, "step": 28815 }, { "epoch": 15.10482180293501, "grad_norm": 0.13340915739536285, "learning_rate": 8.584376404990326e-06, "loss": 0.4753, "num_input_tokens_seen": 18850760, "step": 28820 }, { "epoch": 15.107442348008385, "grad_norm": 0.193513885140419, "learning_rate": 8.575754187160384e-06, "loss": 0.4061, "num_input_tokens_seen": 18853864, "step": 28825 }, { "epoch": 15.11006289308176, "grad_norm": 0.14018668234348297, "learning_rate": 8.567135405095291e-06, "loss": 0.4916, "num_input_tokens_seen": 18857992, "step": 28830 }, { "epoch": 15.112683438155136, "grad_norm": 0.18216575682163239, "learning_rate": 8.558520060597985e-06, "loss": 0.4626, "num_input_tokens_seen": 18861576, "step": 28835 }, { "epoch": 15.115303983228511, "grad_norm": 0.14992855489253998, "learning_rate": 8.54990815547071e-06, "loss": 0.4777, "num_input_tokens_seen": 18865736, "step": 28840 }, { "epoch": 15.117924528301886, "grad_norm": 0.12968213856220245, "learning_rate": 8.541299691514974e-06, "loss": 0.4066, "num_input_tokens_seen": 18869736, "step": 28845 }, { "epoch": 15.120545073375261, "grad_norm": 0.13691851496696472, "learning_rate": 8.53269467053158e-06, "loss": 0.5665, "num_input_tokens_seen": 18872520, "step": 28850 }, { "epoch": 15.123165618448636, "grad_norm": 0.15245719254016876, "learning_rate": 8.524093094320593e-06, "loss": 0.4521, "num_input_tokens_seen": 18876200, "step": 28855 }, { "epoch": 15.125786163522013, "grad_norm": 0.14029185473918915, "learning_rate": 8.515494964681357e-06, "loss": 0.4596, "num_input_tokens_seen": 18878856, "step": 28860 }, { "epoch": 15.128406708595389, "grad_norm": 0.15465693175792694, "learning_rate": 8.506900283412506e-06, "loss": 0.5751, "num_input_tokens_seen": 18883112, "step": 28865 }, { "epoch": 15.131027253668764, "grad_norm": 0.11837983876466751, "learning_rate": 8.498309052311953e-06, "loss": 0.4419, "num_input_tokens_seen": 18886344, "step": 28870 }, { "epoch": 15.133647798742139, "grad_norm": 0.12312176823616028, "learning_rate": 8.489721273176887e-06, "loss": 0.5259, "num_input_tokens_seen": 18890248, "step": 28875 }, { "epoch": 15.136268343815514, "grad_norm": 0.16486863791942596, "learning_rate": 8.481136947803777e-06, "loss": 0.4709, "num_input_tokens_seen": 18893512, "step": 28880 }, { "epoch": 15.13888888888889, "grad_norm": 0.1921038031578064, "learning_rate": 8.472556077988362e-06, "loss": 0.4795, "num_input_tokens_seen": 18896552, "step": 28885 }, { "epoch": 15.141509433962264, "grad_norm": 0.14073580503463745, "learning_rate": 8.46397866552565e-06, "loss": 0.5282, "num_input_tokens_seen": 18899912, "step": 28890 }, { "epoch": 15.14412997903564, "grad_norm": 0.22605320811271667, "learning_rate": 8.455404712209952e-06, "loss": 0.5124, "num_input_tokens_seen": 18903336, "step": 28895 }, { "epoch": 15.146750524109015, "grad_norm": 0.2004874348640442, "learning_rate": 8.446834219834836e-06, "loss": 0.4627, "num_input_tokens_seen": 18905736, "step": 28900 }, { "epoch": 15.14937106918239, "grad_norm": 0.16737978160381317, "learning_rate": 8.43826719019316e-06, "loss": 0.3077, "num_input_tokens_seen": 18909864, "step": 28905 }, { "epoch": 15.151991614255765, "grad_norm": 0.19801193475723267, "learning_rate": 8.429703625077043e-06, "loss": 0.4798, "num_input_tokens_seen": 18914280, "step": 28910 }, { "epoch": 15.15461215932914, "grad_norm": 0.14337509870529175, "learning_rate": 8.421143526277875e-06, "loss": 0.521, "num_input_tokens_seen": 18917160, "step": 28915 }, { "epoch": 15.157232704402515, "grad_norm": 0.15857329964637756, "learning_rate": 8.41258689558634e-06, "loss": 0.4804, "num_input_tokens_seen": 18920616, "step": 28920 }, { "epoch": 15.15985324947589, "grad_norm": 0.12515711784362793, "learning_rate": 8.404033734792386e-06, "loss": 0.546, "num_input_tokens_seen": 18924616, "step": 28925 }, { "epoch": 15.162473794549266, "grad_norm": 0.389415442943573, "learning_rate": 8.395484045685248e-06, "loss": 0.559, "num_input_tokens_seen": 18928488, "step": 28930 }, { "epoch": 15.165094339622641, "grad_norm": 0.1796364188194275, "learning_rate": 8.386937830053412e-06, "loss": 0.4817, "num_input_tokens_seen": 18931240, "step": 28935 }, { "epoch": 15.167714884696016, "grad_norm": 0.18772758543491364, "learning_rate": 8.37839508968464e-06, "loss": 0.7614, "num_input_tokens_seen": 18933928, "step": 28940 }, { "epoch": 15.170335429769391, "grad_norm": 0.14571993052959442, "learning_rate": 8.369855826365988e-06, "loss": 0.4397, "num_input_tokens_seen": 18936776, "step": 28945 }, { "epoch": 15.172955974842766, "grad_norm": 0.10374784469604492, "learning_rate": 8.361320041883772e-06, "loss": 0.4457, "num_input_tokens_seen": 18939912, "step": 28950 }, { "epoch": 15.175576519916143, "grad_norm": 0.1916365623474121, "learning_rate": 8.352787738023576e-06, "loss": 0.41, "num_input_tokens_seen": 18942600, "step": 28955 }, { "epoch": 15.178197064989519, "grad_norm": 0.36920785903930664, "learning_rate": 8.344258916570274e-06, "loss": 0.469, "num_input_tokens_seen": 18945640, "step": 28960 }, { "epoch": 15.180817610062894, "grad_norm": 0.1845765858888626, "learning_rate": 8.335733579307988e-06, "loss": 0.3861, "num_input_tokens_seen": 18950152, "step": 28965 }, { "epoch": 15.183438155136269, "grad_norm": 0.13940784335136414, "learning_rate": 8.327211728020113e-06, "loss": 0.4409, "num_input_tokens_seen": 18952520, "step": 28970 }, { "epoch": 15.186058700209644, "grad_norm": 0.15708142518997192, "learning_rate": 8.31869336448933e-06, "loss": 0.5839, "num_input_tokens_seen": 18955336, "step": 28975 }, { "epoch": 15.18867924528302, "grad_norm": 0.13407683372497559, "learning_rate": 8.310178490497586e-06, "loss": 0.4335, "num_input_tokens_seen": 18958216, "step": 28980 }, { "epoch": 15.191299790356394, "grad_norm": 0.119174525141716, "learning_rate": 8.301667107826103e-06, "loss": 0.6151, "num_input_tokens_seen": 18961128, "step": 28985 }, { "epoch": 15.19392033542977, "grad_norm": 0.21360717713832855, "learning_rate": 8.293159218255345e-06, "loss": 0.5189, "num_input_tokens_seen": 18964200, "step": 28990 }, { "epoch": 15.196540880503145, "grad_norm": 0.196243554353714, "learning_rate": 8.284654823565088e-06, "loss": 0.6528, "num_input_tokens_seen": 18967368, "step": 28995 }, { "epoch": 15.19916142557652, "grad_norm": 0.16435520350933075, "learning_rate": 8.276153925534333e-06, "loss": 0.4291, "num_input_tokens_seen": 18970696, "step": 29000 }, { "epoch": 15.201781970649895, "grad_norm": 0.1919059306383133, "learning_rate": 8.267656525941383e-06, "loss": 0.5275, "num_input_tokens_seen": 18974280, "step": 29005 }, { "epoch": 15.20440251572327, "grad_norm": 0.17570860683918, "learning_rate": 8.259162626563801e-06, "loss": 0.3861, "num_input_tokens_seen": 18977320, "step": 29010 }, { "epoch": 15.207023060796645, "grad_norm": 0.1088847890496254, "learning_rate": 8.250672229178402e-06, "loss": 0.4301, "num_input_tokens_seen": 18980712, "step": 29015 }, { "epoch": 15.20964360587002, "grad_norm": 0.14161625504493713, "learning_rate": 8.242185335561295e-06, "loss": 0.5265, "num_input_tokens_seen": 18984168, "step": 29020 }, { "epoch": 15.212264150943396, "grad_norm": 0.11118394136428833, "learning_rate": 8.23370194748782e-06, "loss": 0.4286, "num_input_tokens_seen": 18987528, "step": 29025 }, { "epoch": 15.214884696016771, "grad_norm": 0.16666053235530853, "learning_rate": 8.22522206673262e-06, "loss": 0.4686, "num_input_tokens_seen": 18991560, "step": 29030 }, { "epoch": 15.217505241090146, "grad_norm": 0.23979541659355164, "learning_rate": 8.216745695069589e-06, "loss": 0.5413, "num_input_tokens_seen": 18996360, "step": 29035 }, { "epoch": 15.220125786163521, "grad_norm": 0.18544259667396545, "learning_rate": 8.208272834271894e-06, "loss": 0.4729, "num_input_tokens_seen": 18999304, "step": 29040 }, { "epoch": 15.222746331236896, "grad_norm": 0.2580074667930603, "learning_rate": 8.19980348611194e-06, "loss": 0.4685, "num_input_tokens_seen": 19002888, "step": 29045 }, { "epoch": 15.225366876310272, "grad_norm": 0.15098343789577484, "learning_rate": 8.191337652361439e-06, "loss": 0.5646, "num_input_tokens_seen": 19006760, "step": 29050 }, { "epoch": 15.227987421383649, "grad_norm": 0.17708571255207062, "learning_rate": 8.182875334791332e-06, "loss": 0.514, "num_input_tokens_seen": 19009832, "step": 29055 }, { "epoch": 15.230607966457024, "grad_norm": 0.10836754739284515, "learning_rate": 8.174416535171841e-06, "loss": 0.3765, "num_input_tokens_seen": 19013000, "step": 29060 }, { "epoch": 15.233228511530399, "grad_norm": 0.21094253659248352, "learning_rate": 8.165961255272467e-06, "loss": 0.5435, "num_input_tokens_seen": 19017288, "step": 29065 }, { "epoch": 15.235849056603774, "grad_norm": 0.19841551780700684, "learning_rate": 8.157509496861935e-06, "loss": 0.3912, "num_input_tokens_seen": 19020936, "step": 29070 }, { "epoch": 15.23846960167715, "grad_norm": 0.20940864086151123, "learning_rate": 8.149061261708266e-06, "loss": 0.4671, "num_input_tokens_seen": 19024488, "step": 29075 }, { "epoch": 15.241090146750524, "grad_norm": 0.23037944734096527, "learning_rate": 8.140616551578745e-06, "loss": 0.5198, "num_input_tokens_seen": 19027752, "step": 29080 }, { "epoch": 15.2437106918239, "grad_norm": 0.17249219119548798, "learning_rate": 8.13217536823989e-06, "loss": 0.44, "num_input_tokens_seen": 19030152, "step": 29085 }, { "epoch": 15.246331236897275, "grad_norm": 0.21186964213848114, "learning_rate": 8.12373771345752e-06, "loss": 0.4651, "num_input_tokens_seen": 19033032, "step": 29090 }, { "epoch": 15.24895178197065, "grad_norm": 0.1597525179386139, "learning_rate": 8.115303588996676e-06, "loss": 0.517, "num_input_tokens_seen": 19036424, "step": 29095 }, { "epoch": 15.251572327044025, "grad_norm": 0.15559060871601105, "learning_rate": 8.10687299662169e-06, "loss": 0.3519, "num_input_tokens_seen": 19039592, "step": 29100 }, { "epoch": 15.2541928721174, "grad_norm": 0.13744673132896423, "learning_rate": 8.098445938096147e-06, "loss": 0.3601, "num_input_tokens_seen": 19041992, "step": 29105 }, { "epoch": 15.256813417190775, "grad_norm": 0.18412069976329803, "learning_rate": 8.090022415182898e-06, "loss": 0.4528, "num_input_tokens_seen": 19044968, "step": 29110 }, { "epoch": 15.25943396226415, "grad_norm": 0.19327674806118011, "learning_rate": 8.081602429644033e-06, "loss": 0.4717, "num_input_tokens_seen": 19047656, "step": 29115 }, { "epoch": 15.262054507337526, "grad_norm": 0.2238064706325531, "learning_rate": 8.073185983240932e-06, "loss": 0.6002, "num_input_tokens_seen": 19051272, "step": 29120 }, { "epoch": 15.264675052410901, "grad_norm": 0.19343328475952148, "learning_rate": 8.064773077734206e-06, "loss": 0.4878, "num_input_tokens_seen": 19054088, "step": 29125 }, { "epoch": 15.267295597484276, "grad_norm": 0.21966463327407837, "learning_rate": 8.05636371488374e-06, "loss": 0.4632, "num_input_tokens_seen": 19058248, "step": 29130 }, { "epoch": 15.269916142557651, "grad_norm": 0.1734531968832016, "learning_rate": 8.047957896448696e-06, "loss": 0.5146, "num_input_tokens_seen": 19063240, "step": 29135 }, { "epoch": 15.272536687631026, "grad_norm": 0.14592143893241882, "learning_rate": 8.039555624187451e-06, "loss": 0.4278, "num_input_tokens_seen": 19065768, "step": 29140 }, { "epoch": 15.275157232704402, "grad_norm": 0.19763709604740143, "learning_rate": 8.031156899857681e-06, "loss": 0.5729, "num_input_tokens_seen": 19069256, "step": 29145 }, { "epoch": 15.277777777777779, "grad_norm": 0.2202293872833252, "learning_rate": 8.022761725216288e-06, "loss": 0.5341, "num_input_tokens_seen": 19072072, "step": 29150 }, { "epoch": 15.280398322851154, "grad_norm": 0.17644745111465454, "learning_rate": 8.014370102019456e-06, "loss": 0.4231, "num_input_tokens_seen": 19075272, "step": 29155 }, { "epoch": 15.283018867924529, "grad_norm": 0.14498436450958252, "learning_rate": 8.005982032022616e-06, "loss": 0.5804, "num_input_tokens_seen": 19078120, "step": 29160 }, { "epoch": 15.285639412997904, "grad_norm": 0.17002961039543152, "learning_rate": 7.997597516980467e-06, "loss": 0.4576, "num_input_tokens_seen": 19080744, "step": 29165 }, { "epoch": 15.28825995807128, "grad_norm": 0.11069372296333313, "learning_rate": 7.989216558646942e-06, "loss": 0.3962, "num_input_tokens_seen": 19084200, "step": 29170 }, { "epoch": 15.290880503144654, "grad_norm": 0.17287267744541168, "learning_rate": 7.980839158775235e-06, "loss": 0.46, "num_input_tokens_seen": 19087400, "step": 29175 }, { "epoch": 15.29350104821803, "grad_norm": 0.2509255111217499, "learning_rate": 7.972465319117814e-06, "loss": 0.3741, "num_input_tokens_seen": 19090472, "step": 29180 }, { "epoch": 15.296121593291405, "grad_norm": 0.13906697928905487, "learning_rate": 7.964095041426386e-06, "loss": 0.5299, "num_input_tokens_seen": 19093896, "step": 29185 }, { "epoch": 15.29874213836478, "grad_norm": 0.13307473063468933, "learning_rate": 7.95572832745192e-06, "loss": 0.5982, "num_input_tokens_seen": 19097384, "step": 29190 }, { "epoch": 15.301362683438155, "grad_norm": 0.18641769886016846, "learning_rate": 7.947365178944643e-06, "loss": 0.4428, "num_input_tokens_seen": 19101160, "step": 29195 }, { "epoch": 15.30398322851153, "grad_norm": 0.15091195702552795, "learning_rate": 7.939005597654025e-06, "loss": 0.4115, "num_input_tokens_seen": 19105064, "step": 29200 }, { "epoch": 15.306603773584905, "grad_norm": 0.22832748293876648, "learning_rate": 7.930649585328787e-06, "loss": 0.576, "num_input_tokens_seen": 19107272, "step": 29205 }, { "epoch": 15.30922431865828, "grad_norm": 0.10766837000846863, "learning_rate": 7.922297143716918e-06, "loss": 0.5217, "num_input_tokens_seen": 19111144, "step": 29210 }, { "epoch": 15.311844863731656, "grad_norm": 0.16801930963993073, "learning_rate": 7.913948274565652e-06, "loss": 0.3867, "num_input_tokens_seen": 19114056, "step": 29215 }, { "epoch": 15.314465408805031, "grad_norm": 0.14244964718818665, "learning_rate": 7.905602979621491e-06, "loss": 0.3937, "num_input_tokens_seen": 19117416, "step": 29220 }, { "epoch": 15.317085953878406, "grad_norm": 0.132787823677063, "learning_rate": 7.897261260630159e-06, "loss": 0.4098, "num_input_tokens_seen": 19120680, "step": 29225 }, { "epoch": 15.319706498951781, "grad_norm": 0.1107519119977951, "learning_rate": 7.888923119336647e-06, "loss": 0.381, "num_input_tokens_seen": 19124136, "step": 29230 }, { "epoch": 15.322327044025156, "grad_norm": 0.2486545741558075, "learning_rate": 7.880588557485203e-06, "loss": 0.7038, "num_input_tokens_seen": 19126984, "step": 29235 }, { "epoch": 15.324947589098532, "grad_norm": 0.2420126348733902, "learning_rate": 7.872257576819325e-06, "loss": 0.4387, "num_input_tokens_seen": 19129704, "step": 29240 }, { "epoch": 15.327568134171909, "grad_norm": 0.19410420954227448, "learning_rate": 7.863930179081769e-06, "loss": 0.5903, "num_input_tokens_seen": 19132232, "step": 29245 }, { "epoch": 15.330188679245284, "grad_norm": 0.1909257024526596, "learning_rate": 7.855606366014517e-06, "loss": 0.4467, "num_input_tokens_seen": 19135016, "step": 29250 }, { "epoch": 15.332809224318659, "grad_norm": 0.12678277492523193, "learning_rate": 7.847286139358814e-06, "loss": 0.4316, "num_input_tokens_seen": 19137576, "step": 29255 }, { "epoch": 15.335429769392034, "grad_norm": 0.14953549206256866, "learning_rate": 7.838969500855162e-06, "loss": 0.5078, "num_input_tokens_seen": 19140488, "step": 29260 }, { "epoch": 15.33805031446541, "grad_norm": 0.14204955101013184, "learning_rate": 7.830656452243307e-06, "loss": 0.5046, "num_input_tokens_seen": 19144456, "step": 29265 }, { "epoch": 15.340670859538784, "grad_norm": 0.1228019967675209, "learning_rate": 7.822346995262241e-06, "loss": 0.3841, "num_input_tokens_seen": 19148200, "step": 29270 }, { "epoch": 15.34329140461216, "grad_norm": 0.15319320559501648, "learning_rate": 7.814041131650223e-06, "loss": 0.5452, "num_input_tokens_seen": 19152040, "step": 29275 }, { "epoch": 15.345911949685535, "grad_norm": 0.18105100095272064, "learning_rate": 7.805738863144731e-06, "loss": 0.5351, "num_input_tokens_seen": 19155272, "step": 29280 }, { "epoch": 15.34853249475891, "grad_norm": 0.14054851233959198, "learning_rate": 7.797440191482496e-06, "loss": 0.6428, "num_input_tokens_seen": 19158216, "step": 29285 }, { "epoch": 15.351153039832285, "grad_norm": 0.19306644797325134, "learning_rate": 7.789145118399518e-06, "loss": 0.4171, "num_input_tokens_seen": 19161800, "step": 29290 }, { "epoch": 15.35377358490566, "grad_norm": 0.25050586462020874, "learning_rate": 7.780853645631032e-06, "loss": 0.3579, "num_input_tokens_seen": 19165160, "step": 29295 }, { "epoch": 15.356394129979035, "grad_norm": 0.12395643442869186, "learning_rate": 7.772565774911522e-06, "loss": 0.5136, "num_input_tokens_seen": 19167752, "step": 29300 }, { "epoch": 15.35901467505241, "grad_norm": 0.08579085767269135, "learning_rate": 7.76428150797471e-06, "loss": 0.4159, "num_input_tokens_seen": 19171432, "step": 29305 }, { "epoch": 15.361635220125786, "grad_norm": 0.20818626880645752, "learning_rate": 7.756000846553562e-06, "loss": 0.3598, "num_input_tokens_seen": 19175720, "step": 29310 }, { "epoch": 15.364255765199161, "grad_norm": 0.11807728558778763, "learning_rate": 7.74772379238031e-06, "loss": 0.484, "num_input_tokens_seen": 19178696, "step": 29315 }, { "epoch": 15.366876310272536, "grad_norm": 0.1896473616361618, "learning_rate": 7.739450347186417e-06, "loss": 0.4865, "num_input_tokens_seen": 19181480, "step": 29320 }, { "epoch": 15.369496855345911, "grad_norm": 0.1703144609928131, "learning_rate": 7.731180512702599e-06, "loss": 0.4239, "num_input_tokens_seen": 19184136, "step": 29325 }, { "epoch": 15.372117400419286, "grad_norm": 0.16520622372627258, "learning_rate": 7.722914290658795e-06, "loss": 0.6735, "num_input_tokens_seen": 19187144, "step": 29330 }, { "epoch": 15.374737945492662, "grad_norm": 0.12903159856796265, "learning_rate": 7.714651682784224e-06, "loss": 0.5527, "num_input_tokens_seen": 19190440, "step": 29335 }, { "epoch": 15.377358490566039, "grad_norm": 0.1253461241722107, "learning_rate": 7.706392690807313e-06, "loss": 0.4622, "num_input_tokens_seen": 19193736, "step": 29340 }, { "epoch": 15.379979035639414, "grad_norm": 0.11282984167337418, "learning_rate": 7.69813731645575e-06, "loss": 0.3694, "num_input_tokens_seen": 19197416, "step": 29345 }, { "epoch": 15.382599580712789, "grad_norm": 0.11311893910169601, "learning_rate": 7.689885561456475e-06, "loss": 0.4939, "num_input_tokens_seen": 19200232, "step": 29350 }, { "epoch": 15.385220125786164, "grad_norm": 0.4518137276172638, "learning_rate": 7.681637427535663e-06, "loss": 0.3862, "num_input_tokens_seen": 19202824, "step": 29355 }, { "epoch": 15.38784067085954, "grad_norm": 0.15719106793403625, "learning_rate": 7.673392916418715e-06, "loss": 0.3689, "num_input_tokens_seen": 19205832, "step": 29360 }, { "epoch": 15.390461215932914, "grad_norm": 0.30361759662628174, "learning_rate": 7.665152029830303e-06, "loss": 0.54, "num_input_tokens_seen": 19209128, "step": 29365 }, { "epoch": 15.39308176100629, "grad_norm": 0.13409024477005005, "learning_rate": 7.656914769494314e-06, "loss": 0.4866, "num_input_tokens_seen": 19211912, "step": 29370 }, { "epoch": 15.395702306079665, "grad_norm": 0.20417708158493042, "learning_rate": 7.648681137133892e-06, "loss": 0.5373, "num_input_tokens_seen": 19215080, "step": 29375 }, { "epoch": 15.39832285115304, "grad_norm": 0.2790868878364563, "learning_rate": 7.640451134471432e-06, "loss": 0.5461, "num_input_tokens_seen": 19218056, "step": 29380 }, { "epoch": 15.400943396226415, "grad_norm": 0.19340860843658447, "learning_rate": 7.632224763228538e-06, "loss": 0.4212, "num_input_tokens_seen": 19221096, "step": 29385 }, { "epoch": 15.40356394129979, "grad_norm": 0.14680685102939606, "learning_rate": 7.6240020251260896e-06, "loss": 0.5001, "num_input_tokens_seen": 19223816, "step": 29390 }, { "epoch": 15.406184486373165, "grad_norm": 0.12308167666196823, "learning_rate": 7.615782921884174e-06, "loss": 0.4324, "num_input_tokens_seen": 19226184, "step": 29395 }, { "epoch": 15.40880503144654, "grad_norm": 0.1350962221622467, "learning_rate": 7.607567455222142e-06, "loss": 0.483, "num_input_tokens_seen": 19229032, "step": 29400 }, { "epoch": 15.411425576519916, "grad_norm": 0.14930284023284912, "learning_rate": 7.599355626858582e-06, "loss": 0.4719, "num_input_tokens_seen": 19232584, "step": 29405 }, { "epoch": 15.414046121593291, "grad_norm": 0.18502244353294373, "learning_rate": 7.591147438511298e-06, "loss": 0.4636, "num_input_tokens_seen": 19235592, "step": 29410 }, { "epoch": 15.416666666666666, "grad_norm": 0.19537898898124695, "learning_rate": 7.5829428918973625e-06, "loss": 0.5604, "num_input_tokens_seen": 19239304, "step": 29415 }, { "epoch": 15.419287211740041, "grad_norm": 0.12948189675807953, "learning_rate": 7.574741988733075e-06, "loss": 0.4226, "num_input_tokens_seen": 19241800, "step": 29420 }, { "epoch": 15.421907756813416, "grad_norm": 0.17748379707336426, "learning_rate": 7.56654473073396e-06, "loss": 0.423, "num_input_tokens_seen": 19245256, "step": 29425 }, { "epoch": 15.424528301886792, "grad_norm": 0.24394474923610687, "learning_rate": 7.558351119614796e-06, "loss": 0.6442, "num_input_tokens_seen": 19247880, "step": 29430 }, { "epoch": 15.427148846960169, "grad_norm": 0.13668760657310486, "learning_rate": 7.5501611570896e-06, "loss": 0.5215, "num_input_tokens_seen": 19251240, "step": 29435 }, { "epoch": 15.429769392033544, "grad_norm": 0.139641672372818, "learning_rate": 7.541974844871602e-06, "loss": 0.5466, "num_input_tokens_seen": 19254856, "step": 29440 }, { "epoch": 15.432389937106919, "grad_norm": 0.12751241028308868, "learning_rate": 7.5337921846732965e-06, "loss": 0.5035, "num_input_tokens_seen": 19258376, "step": 29445 }, { "epoch": 15.435010482180294, "grad_norm": 0.18388934433460236, "learning_rate": 7.525613178206409e-06, "loss": 0.5595, "num_input_tokens_seen": 19262312, "step": 29450 }, { "epoch": 15.43763102725367, "grad_norm": 0.08668304234743118, "learning_rate": 7.517437827181878e-06, "loss": 0.6289, "num_input_tokens_seen": 19266216, "step": 29455 }, { "epoch": 15.440251572327044, "grad_norm": 0.1983749270439148, "learning_rate": 7.509266133309908e-06, "loss": 0.7085, "num_input_tokens_seen": 19269256, "step": 29460 }, { "epoch": 15.44287211740042, "grad_norm": 0.16793760657310486, "learning_rate": 7.501098098299914e-06, "loss": 0.4596, "num_input_tokens_seen": 19272008, "step": 29465 }, { "epoch": 15.445492662473795, "grad_norm": 0.16005238890647888, "learning_rate": 7.492933723860557e-06, "loss": 0.5203, "num_input_tokens_seen": 19275304, "step": 29470 }, { "epoch": 15.44811320754717, "grad_norm": 0.25394102931022644, "learning_rate": 7.4847730116997334e-06, "loss": 0.4115, "num_input_tokens_seen": 19277320, "step": 29475 }, { "epoch": 15.450733752620545, "grad_norm": 0.13399234414100647, "learning_rate": 7.476615963524583e-06, "loss": 0.6252, "num_input_tokens_seen": 19280680, "step": 29480 }, { "epoch": 15.45335429769392, "grad_norm": 0.1150561273097992, "learning_rate": 7.468462581041452e-06, "loss": 0.4382, "num_input_tokens_seen": 19283560, "step": 29485 }, { "epoch": 15.455974842767295, "grad_norm": 0.18590961396694183, "learning_rate": 7.460312865955932e-06, "loss": 0.4765, "num_input_tokens_seen": 19287048, "step": 29490 }, { "epoch": 15.45859538784067, "grad_norm": 0.22140805423259735, "learning_rate": 7.4521668199728584e-06, "loss": 0.5085, "num_input_tokens_seen": 19289672, "step": 29495 }, { "epoch": 15.461215932914046, "grad_norm": 0.17914584279060364, "learning_rate": 7.4440244447962884e-06, "loss": 0.4984, "num_input_tokens_seen": 19292712, "step": 29500 }, { "epoch": 15.463836477987421, "grad_norm": 0.13300952315330505, "learning_rate": 7.435885742129523e-06, "loss": 0.5576, "num_input_tokens_seen": 19296520, "step": 29505 }, { "epoch": 15.466457023060796, "grad_norm": 0.10885097086429596, "learning_rate": 7.427750713675071e-06, "loss": 0.4916, "num_input_tokens_seen": 19299816, "step": 29510 }, { "epoch": 15.469077568134171, "grad_norm": 0.11042025685310364, "learning_rate": 7.419619361134702e-06, "loss": 0.3603, "num_input_tokens_seen": 19302472, "step": 29515 }, { "epoch": 15.471698113207546, "grad_norm": 0.19994507730007172, "learning_rate": 7.411491686209387e-06, "loss": 0.4811, "num_input_tokens_seen": 19305640, "step": 29520 }, { "epoch": 15.474318658280922, "grad_norm": 0.1306641846895218, "learning_rate": 7.4033676905993525e-06, "loss": 0.4613, "num_input_tokens_seen": 19308680, "step": 29525 }, { "epoch": 15.476939203354299, "grad_norm": 0.14563840627670288, "learning_rate": 7.3952473760040444e-06, "loss": 0.3803, "num_input_tokens_seen": 19311240, "step": 29530 }, { "epoch": 15.479559748427674, "grad_norm": 0.14877183735370636, "learning_rate": 7.387130744122148e-06, "loss": 0.439, "num_input_tokens_seen": 19314632, "step": 29535 }, { "epoch": 15.482180293501049, "grad_norm": 0.17788133025169373, "learning_rate": 7.379017796651558e-06, "loss": 0.5169, "num_input_tokens_seen": 19317736, "step": 29540 }, { "epoch": 15.484800838574424, "grad_norm": 0.3868527114391327, "learning_rate": 7.3709085352894085e-06, "loss": 0.4757, "num_input_tokens_seen": 19320936, "step": 29545 }, { "epoch": 15.4874213836478, "grad_norm": 0.2418363392353058, "learning_rate": 7.362802961732071e-06, "loss": 0.3949, "num_input_tokens_seen": 19324360, "step": 29550 }, { "epoch": 15.490041928721174, "grad_norm": 0.18152779340744019, "learning_rate": 7.354701077675136e-06, "loss": 0.4685, "num_input_tokens_seen": 19328296, "step": 29555 }, { "epoch": 15.49266247379455, "grad_norm": 0.21412143111228943, "learning_rate": 7.346602884813439e-06, "loss": 0.4722, "num_input_tokens_seen": 19331336, "step": 29560 }, { "epoch": 15.495283018867925, "grad_norm": 0.10616133362054825, "learning_rate": 7.338508384841014e-06, "loss": 0.5422, "num_input_tokens_seen": 19334760, "step": 29565 }, { "epoch": 15.4979035639413, "grad_norm": 0.19306470453739166, "learning_rate": 7.3304175794511345e-06, "loss": 0.588, "num_input_tokens_seen": 19337320, "step": 29570 }, { "epoch": 15.500524109014675, "grad_norm": 0.19359202682971954, "learning_rate": 7.3223304703363135e-06, "loss": 0.5642, "num_input_tokens_seen": 19340584, "step": 29575 }, { "epoch": 15.50314465408805, "grad_norm": 0.24208447337150574, "learning_rate": 7.314247059188281e-06, "loss": 0.5818, "num_input_tokens_seen": 19344904, "step": 29580 }, { "epoch": 15.505765199161425, "grad_norm": 0.23972438275814056, "learning_rate": 7.306167347697992e-06, "loss": 0.4587, "num_input_tokens_seen": 19348648, "step": 29585 }, { "epoch": 15.5083857442348, "grad_norm": 0.13197389245033264, "learning_rate": 7.29809133755564e-06, "loss": 0.4905, "num_input_tokens_seen": 19351816, "step": 29590 }, { "epoch": 15.511006289308176, "grad_norm": 0.17600901424884796, "learning_rate": 7.290019030450629e-06, "loss": 0.5055, "num_input_tokens_seen": 19355144, "step": 29595 }, { "epoch": 15.51362683438155, "grad_norm": 0.15798532962799072, "learning_rate": 7.2819504280715794e-06, "loss": 0.4482, "num_input_tokens_seen": 19357608, "step": 29600 }, { "epoch": 15.516247379454926, "grad_norm": 0.24449887871742249, "learning_rate": 7.273885532106364e-06, "loss": 0.5493, "num_input_tokens_seen": 19360936, "step": 29605 }, { "epoch": 15.518867924528301, "grad_norm": 0.12042616307735443, "learning_rate": 7.265824344242064e-06, "loss": 0.5509, "num_input_tokens_seen": 19364232, "step": 29610 }, { "epoch": 15.521488469601676, "grad_norm": 0.11070730537176132, "learning_rate": 7.257766866165e-06, "loss": 0.5422, "num_input_tokens_seen": 19367720, "step": 29615 }, { "epoch": 15.524109014675052, "grad_norm": 0.274036705493927, "learning_rate": 7.24971309956069e-06, "loss": 0.462, "num_input_tokens_seen": 19370408, "step": 29620 }, { "epoch": 15.526729559748428, "grad_norm": 0.1725706309080124, "learning_rate": 7.241663046113887e-06, "loss": 0.5041, "num_input_tokens_seen": 19373576, "step": 29625 }, { "epoch": 15.529350104821804, "grad_norm": 0.1338939666748047, "learning_rate": 7.233616707508576e-06, "loss": 0.4308, "num_input_tokens_seen": 19376232, "step": 29630 }, { "epoch": 15.531970649895179, "grad_norm": 0.1707034707069397, "learning_rate": 7.225574085427961e-06, "loss": 0.6398, "num_input_tokens_seen": 19379144, "step": 29635 }, { "epoch": 15.534591194968554, "grad_norm": 0.16818378865718842, "learning_rate": 7.217535181554474e-06, "loss": 0.4404, "num_input_tokens_seen": 19381992, "step": 29640 }, { "epoch": 15.53721174004193, "grad_norm": 0.27081212401390076, "learning_rate": 7.209499997569747e-06, "loss": 0.6096, "num_input_tokens_seen": 19385416, "step": 29645 }, { "epoch": 15.539832285115304, "grad_norm": 0.21208880841732025, "learning_rate": 7.201468535154663e-06, "loss": 0.4785, "num_input_tokens_seen": 19388328, "step": 29650 }, { "epoch": 15.54245283018868, "grad_norm": 0.11599812656641006, "learning_rate": 7.193440795989295e-06, "loss": 0.396, "num_input_tokens_seen": 19391624, "step": 29655 }, { "epoch": 15.545073375262055, "grad_norm": 0.2219507396221161, "learning_rate": 7.185416781752965e-06, "loss": 0.537, "num_input_tokens_seen": 19394376, "step": 29660 }, { "epoch": 15.54769392033543, "grad_norm": 0.15076620876789093, "learning_rate": 7.177396494124206e-06, "loss": 0.5424, "num_input_tokens_seen": 19397928, "step": 29665 }, { "epoch": 15.550314465408805, "grad_norm": 0.18068736791610718, "learning_rate": 7.169379934780779e-06, "loss": 0.4942, "num_input_tokens_seen": 19400936, "step": 29670 }, { "epoch": 15.55293501048218, "grad_norm": 0.15936563909053802, "learning_rate": 7.161367105399644e-06, "loss": 0.5763, "num_input_tokens_seen": 19404360, "step": 29675 }, { "epoch": 15.555555555555555, "grad_norm": 0.2727125585079193, "learning_rate": 7.153358007656991e-06, "loss": 0.5534, "num_input_tokens_seen": 19406568, "step": 29680 }, { "epoch": 15.55817610062893, "grad_norm": 0.11086954176425934, "learning_rate": 7.145352643228237e-06, "loss": 0.5066, "num_input_tokens_seen": 19410184, "step": 29685 }, { "epoch": 15.560796645702306, "grad_norm": 0.13539576530456543, "learning_rate": 7.137351013788013e-06, "loss": 0.4432, "num_input_tokens_seen": 19413480, "step": 29690 }, { "epoch": 15.56341719077568, "grad_norm": 0.25808170437812805, "learning_rate": 7.129353121010177e-06, "loss": 0.4312, "num_input_tokens_seen": 19416328, "step": 29695 }, { "epoch": 15.566037735849056, "grad_norm": 0.2025403529405594, "learning_rate": 7.121358966567779e-06, "loss": 0.5774, "num_input_tokens_seen": 19419656, "step": 29700 }, { "epoch": 15.568658280922431, "grad_norm": 0.14487552642822266, "learning_rate": 7.113368552133126e-06, "loss": 0.4381, "num_input_tokens_seen": 19424456, "step": 29705 }, { "epoch": 15.571278825995806, "grad_norm": 0.176505446434021, "learning_rate": 7.105381879377701e-06, "loss": 0.5349, "num_input_tokens_seen": 19427304, "step": 29710 }, { "epoch": 15.573899371069182, "grad_norm": 0.19671787321567535, "learning_rate": 7.097398949972234e-06, "loss": 0.3795, "num_input_tokens_seen": 19429704, "step": 29715 }, { "epoch": 15.576519916142558, "grad_norm": 0.09351888298988342, "learning_rate": 7.089419765586672e-06, "loss": 0.4072, "num_input_tokens_seen": 19433000, "step": 29720 }, { "epoch": 15.579140461215934, "grad_norm": 0.11994009464979172, "learning_rate": 7.081444327890152e-06, "loss": 0.3721, "num_input_tokens_seen": 19437096, "step": 29725 }, { "epoch": 15.581761006289309, "grad_norm": 0.17902925610542297, "learning_rate": 7.073472638551054e-06, "loss": 0.5944, "num_input_tokens_seen": 19440232, "step": 29730 }, { "epoch": 15.584381551362684, "grad_norm": 0.19164995849132538, "learning_rate": 7.065504699236969e-06, "loss": 0.4476, "num_input_tokens_seen": 19443528, "step": 29735 }, { "epoch": 15.58700209643606, "grad_norm": 0.15829266607761383, "learning_rate": 7.057540511614688e-06, "loss": 0.4179, "num_input_tokens_seen": 19448200, "step": 29740 }, { "epoch": 15.589622641509434, "grad_norm": 0.20835566520690918, "learning_rate": 7.049580077350232e-06, "loss": 0.37, "num_input_tokens_seen": 19450792, "step": 29745 }, { "epoch": 15.59224318658281, "grad_norm": 0.21400950849056244, "learning_rate": 7.041623398108843e-06, "loss": 0.3904, "num_input_tokens_seen": 19456744, "step": 29750 }, { "epoch": 15.594863731656185, "grad_norm": 0.12399124354124069, "learning_rate": 7.033670475554949e-06, "loss": 0.5155, "num_input_tokens_seen": 19459368, "step": 29755 }, { "epoch": 15.59748427672956, "grad_norm": 0.16990922391414642, "learning_rate": 7.025721311352224e-06, "loss": 0.4688, "num_input_tokens_seen": 19462952, "step": 29760 }, { "epoch": 15.600104821802935, "grad_norm": 0.12073136121034622, "learning_rate": 7.017775907163546e-06, "loss": 0.535, "num_input_tokens_seen": 19467720, "step": 29765 }, { "epoch": 15.60272536687631, "grad_norm": 0.11202865839004517, "learning_rate": 7.009834264650989e-06, "loss": 0.3586, "num_input_tokens_seen": 19470952, "step": 29770 }, { "epoch": 15.605345911949685, "grad_norm": 0.10834439843893051, "learning_rate": 7.001896385475867e-06, "loss": 0.4904, "num_input_tokens_seen": 19474280, "step": 29775 }, { "epoch": 15.60796645702306, "grad_norm": 0.20108240842819214, "learning_rate": 6.993962271298682e-06, "loss": 0.4514, "num_input_tokens_seen": 19478312, "step": 29780 }, { "epoch": 15.610587002096436, "grad_norm": 0.2685476541519165, "learning_rate": 6.986031923779166e-06, "loss": 0.4988, "num_input_tokens_seen": 19482536, "step": 29785 }, { "epoch": 15.61320754716981, "grad_norm": 0.20732930302619934, "learning_rate": 6.978105344576264e-06, "loss": 0.5676, "num_input_tokens_seen": 19486696, "step": 29790 }, { "epoch": 15.615828092243186, "grad_norm": 0.1192607507109642, "learning_rate": 6.970182535348111e-06, "loss": 0.3445, "num_input_tokens_seen": 19490792, "step": 29795 }, { "epoch": 15.618448637316561, "grad_norm": 0.1520768254995346, "learning_rate": 6.962263497752086e-06, "loss": 0.4253, "num_input_tokens_seen": 19493768, "step": 29800 }, { "epoch": 15.621069182389936, "grad_norm": 0.2010933756828308, "learning_rate": 6.9543482334447436e-06, "loss": 0.5237, "num_input_tokens_seen": 19496968, "step": 29805 }, { "epoch": 15.623689727463312, "grad_norm": 0.10721849650144577, "learning_rate": 6.946436744081875e-06, "loss": 0.4332, "num_input_tokens_seen": 19500136, "step": 29810 }, { "epoch": 15.626310272536688, "grad_norm": 0.15885291993618011, "learning_rate": 6.938529031318472e-06, "loss": 0.4432, "num_input_tokens_seen": 19503432, "step": 29815 }, { "epoch": 15.628930817610064, "grad_norm": 0.16307495534420013, "learning_rate": 6.9306250968087485e-06, "loss": 0.5022, "num_input_tokens_seen": 19506856, "step": 29820 }, { "epoch": 15.631551362683439, "grad_norm": 0.1958056539297104, "learning_rate": 6.922724942206102e-06, "loss": 0.553, "num_input_tokens_seen": 19513512, "step": 29825 }, { "epoch": 15.634171907756814, "grad_norm": 0.29314354062080383, "learning_rate": 6.914828569163167e-06, "loss": 0.3649, "num_input_tokens_seen": 19515624, "step": 29830 }, { "epoch": 15.63679245283019, "grad_norm": 0.13756176829338074, "learning_rate": 6.906935979331763e-06, "loss": 0.5475, "num_input_tokens_seen": 19518696, "step": 29835 }, { "epoch": 15.639412997903564, "grad_norm": 0.3126300871372223, "learning_rate": 6.8990471743629356e-06, "loss": 0.4514, "num_input_tokens_seen": 19521608, "step": 29840 }, { "epoch": 15.64203354297694, "grad_norm": 0.18520861864089966, "learning_rate": 6.89116215590693e-06, "loss": 0.4975, "num_input_tokens_seen": 19525416, "step": 29845 }, { "epoch": 15.644654088050315, "grad_norm": 0.1416291892528534, "learning_rate": 6.8832809256132146e-06, "loss": 0.4444, "num_input_tokens_seen": 19528584, "step": 29850 }, { "epoch": 15.64727463312369, "grad_norm": 0.11352306604385376, "learning_rate": 6.875403485130444e-06, "loss": 0.4261, "num_input_tokens_seen": 19531784, "step": 29855 }, { "epoch": 15.649895178197065, "grad_norm": 0.34962594509124756, "learning_rate": 6.867529836106479e-06, "loss": 0.4302, "num_input_tokens_seen": 19534728, "step": 29860 }, { "epoch": 15.65251572327044, "grad_norm": 0.24204102158546448, "learning_rate": 6.8596599801884045e-06, "loss": 0.5595, "num_input_tokens_seen": 19537288, "step": 29865 }, { "epoch": 15.655136268343815, "grad_norm": 0.1499422937631607, "learning_rate": 6.851793919022509e-06, "loss": 0.3797, "num_input_tokens_seen": 19540296, "step": 29870 }, { "epoch": 15.65775681341719, "grad_norm": 0.22225838899612427, "learning_rate": 6.843931654254285e-06, "loss": 0.4557, "num_input_tokens_seen": 19542504, "step": 29875 }, { "epoch": 15.660377358490566, "grad_norm": 0.1579173356294632, "learning_rate": 6.836073187528425e-06, "loss": 0.4127, "num_input_tokens_seen": 19546184, "step": 29880 }, { "epoch": 15.66299790356394, "grad_norm": 0.16838324069976807, "learning_rate": 6.828218520488821e-06, "loss": 0.4289, "num_input_tokens_seen": 19549096, "step": 29885 }, { "epoch": 15.665618448637316, "grad_norm": 0.13161645829677582, "learning_rate": 6.820367654778589e-06, "loss": 0.4592, "num_input_tokens_seen": 19552296, "step": 29890 }, { "epoch": 15.668238993710691, "grad_norm": 0.22486308217048645, "learning_rate": 6.812520592040039e-06, "loss": 0.6021, "num_input_tokens_seen": 19556520, "step": 29895 }, { "epoch": 15.670859538784066, "grad_norm": 0.20925460755825043, "learning_rate": 6.804677333914689e-06, "loss": 0.4821, "num_input_tokens_seen": 19560712, "step": 29900 }, { "epoch": 15.673480083857442, "grad_norm": 0.17425771057605743, "learning_rate": 6.796837882043261e-06, "loss": 0.5376, "num_input_tokens_seen": 19564072, "step": 29905 }, { "epoch": 15.676100628930818, "grad_norm": 0.09920870512723923, "learning_rate": 6.7890022380656785e-06, "loss": 0.572, "num_input_tokens_seen": 19567176, "step": 29910 }, { "epoch": 15.678721174004194, "grad_norm": 0.17061269283294678, "learning_rate": 6.781170403621056e-06, "loss": 0.4762, "num_input_tokens_seen": 19570248, "step": 29915 }, { "epoch": 15.681341719077569, "grad_norm": 0.12055838853120804, "learning_rate": 6.773342380347736e-06, "loss": 0.5114, "num_input_tokens_seen": 19573000, "step": 29920 }, { "epoch": 15.683962264150944, "grad_norm": 0.23524142801761627, "learning_rate": 6.765518169883248e-06, "loss": 0.536, "num_input_tokens_seen": 19575272, "step": 29925 }, { "epoch": 15.68658280922432, "grad_norm": 0.2933511435985565, "learning_rate": 6.757697773864338e-06, "loss": 0.4968, "num_input_tokens_seen": 19577544, "step": 29930 }, { "epoch": 15.689203354297694, "grad_norm": 0.11892002075910568, "learning_rate": 6.749881193926932e-06, "loss": 0.3695, "num_input_tokens_seen": 19581192, "step": 29935 }, { "epoch": 15.69182389937107, "grad_norm": 0.22144754230976105, "learning_rate": 6.742068431706167e-06, "loss": 0.4456, "num_input_tokens_seen": 19584456, "step": 29940 }, { "epoch": 15.694444444444445, "grad_norm": 0.3539850115776062, "learning_rate": 6.734259488836386e-06, "loss": 0.5352, "num_input_tokens_seen": 19587176, "step": 29945 }, { "epoch": 15.69706498951782, "grad_norm": 0.15207628905773163, "learning_rate": 6.726454366951135e-06, "loss": 0.4538, "num_input_tokens_seen": 19590664, "step": 29950 }, { "epoch": 15.699685534591195, "grad_norm": 0.23632818460464478, "learning_rate": 6.7186530676831616e-06, "loss": 0.4934, "num_input_tokens_seen": 19594792, "step": 29955 }, { "epoch": 15.70230607966457, "grad_norm": 0.1345374584197998, "learning_rate": 6.710855592664403e-06, "loss": 0.4773, "num_input_tokens_seen": 19597352, "step": 29960 }, { "epoch": 15.704926624737945, "grad_norm": 0.21391597390174866, "learning_rate": 6.703061943525993e-06, "loss": 0.4095, "num_input_tokens_seen": 19600168, "step": 29965 }, { "epoch": 15.70754716981132, "grad_norm": 0.1406823992729187, "learning_rate": 6.69527212189828e-06, "loss": 0.5476, "num_input_tokens_seen": 19603560, "step": 29970 }, { "epoch": 15.710167714884696, "grad_norm": 0.11885727196931839, "learning_rate": 6.687486129410811e-06, "loss": 0.3841, "num_input_tokens_seen": 19607240, "step": 29975 }, { "epoch": 15.71278825995807, "grad_norm": 0.18517327308654785, "learning_rate": 6.679703967692322e-06, "loss": 0.3809, "num_input_tokens_seen": 19610088, "step": 29980 }, { "epoch": 15.715408805031446, "grad_norm": 0.22268080711364746, "learning_rate": 6.67192563837076e-06, "loss": 0.5468, "num_input_tokens_seen": 19613064, "step": 29985 }, { "epoch": 15.718029350104821, "grad_norm": 0.16236932575702667, "learning_rate": 6.664151143073258e-06, "loss": 0.4189, "num_input_tokens_seen": 19616264, "step": 29990 }, { "epoch": 15.720649895178196, "grad_norm": 0.14252150058746338, "learning_rate": 6.656380483426141e-06, "loss": 0.4753, "num_input_tokens_seen": 19619560, "step": 29995 }, { "epoch": 15.723270440251572, "grad_norm": 0.20021845400333405, "learning_rate": 6.648613661054956e-06, "loss": 0.4883, "num_input_tokens_seen": 19622088, "step": 30000 }, { "epoch": 15.725890985324948, "grad_norm": 0.15497297048568726, "learning_rate": 6.6408506775844256e-06, "loss": 0.4327, "num_input_tokens_seen": 19626312, "step": 30005 }, { "epoch": 15.728511530398324, "grad_norm": 0.3171003758907318, "learning_rate": 6.633091534638492e-06, "loss": 0.4262, "num_input_tokens_seen": 19628840, "step": 30010 }, { "epoch": 15.731132075471699, "grad_norm": 0.173609659075737, "learning_rate": 6.625336233840257e-06, "loss": 0.4404, "num_input_tokens_seen": 19631720, "step": 30015 }, { "epoch": 15.733752620545074, "grad_norm": 0.15848155319690704, "learning_rate": 6.617584776812064e-06, "loss": 0.5762, "num_input_tokens_seen": 19634312, "step": 30020 }, { "epoch": 15.73637316561845, "grad_norm": 0.2961007058620453, "learning_rate": 6.6098371651754085e-06, "loss": 0.5161, "num_input_tokens_seen": 19637064, "step": 30025 }, { "epoch": 15.738993710691824, "grad_norm": 0.11766737699508667, "learning_rate": 6.602093400551012e-06, "loss": 0.4964, "num_input_tokens_seen": 19640552, "step": 30030 }, { "epoch": 15.7416142557652, "grad_norm": 0.2518046200275421, "learning_rate": 6.59435348455879e-06, "loss": 0.5112, "num_input_tokens_seen": 19643624, "step": 30035 }, { "epoch": 15.744234800838575, "grad_norm": 0.12356976419687271, "learning_rate": 6.586617418817828e-06, "loss": 0.4065, "num_input_tokens_seen": 19646280, "step": 30040 }, { "epoch": 15.74685534591195, "grad_norm": 0.18698282539844513, "learning_rate": 6.578885204946439e-06, "loss": 0.4632, "num_input_tokens_seen": 19649160, "step": 30045 }, { "epoch": 15.749475890985325, "grad_norm": 0.18344134092330933, "learning_rate": 6.571156844562098e-06, "loss": 0.5396, "num_input_tokens_seen": 19652616, "step": 30050 }, { "epoch": 15.7520964360587, "grad_norm": 0.3734492361545563, "learning_rate": 6.563432339281497e-06, "loss": 0.4013, "num_input_tokens_seen": 19655208, "step": 30055 }, { "epoch": 15.754716981132075, "grad_norm": 0.146528422832489, "learning_rate": 6.555711690720517e-06, "loss": 0.5134, "num_input_tokens_seen": 19657928, "step": 30060 }, { "epoch": 15.75733752620545, "grad_norm": 0.1440056562423706, "learning_rate": 6.547994900494234e-06, "loss": 0.4588, "num_input_tokens_seen": 19661544, "step": 30065 }, { "epoch": 15.759958071278826, "grad_norm": 0.12528114020824432, "learning_rate": 6.540281970216899e-06, "loss": 0.5005, "num_input_tokens_seen": 19664776, "step": 30070 }, { "epoch": 15.7625786163522, "grad_norm": 0.1389610767364502, "learning_rate": 6.5325729015019845e-06, "loss": 0.4652, "num_input_tokens_seen": 19668136, "step": 30075 }, { "epoch": 15.765199161425576, "grad_norm": 0.17968431115150452, "learning_rate": 6.524867695962122e-06, "loss": 0.7083, "num_input_tokens_seen": 19671592, "step": 30080 }, { "epoch": 15.767819706498951, "grad_norm": 0.17330409586429596, "learning_rate": 6.517166355209165e-06, "loss": 0.5093, "num_input_tokens_seen": 19676136, "step": 30085 }, { "epoch": 15.770440251572326, "grad_norm": 0.13179421424865723, "learning_rate": 6.50946888085415e-06, "loss": 0.4537, "num_input_tokens_seen": 19679368, "step": 30090 }, { "epoch": 15.773060796645701, "grad_norm": 0.10682916641235352, "learning_rate": 6.5017752745072855e-06, "loss": 0.4303, "num_input_tokens_seen": 19682856, "step": 30095 }, { "epoch": 15.775681341719078, "grad_norm": 0.7724379897117615, "learning_rate": 6.494085537777994e-06, "loss": 0.6486, "num_input_tokens_seen": 19685480, "step": 30100 }, { "epoch": 15.778301886792454, "grad_norm": 0.15728524327278137, "learning_rate": 6.48639967227489e-06, "loss": 0.3692, "num_input_tokens_seen": 19688104, "step": 30105 }, { "epoch": 15.780922431865829, "grad_norm": 0.24858437478542328, "learning_rate": 6.47871767960575e-06, "loss": 0.4656, "num_input_tokens_seen": 19691656, "step": 30110 }, { "epoch": 15.783542976939204, "grad_norm": 0.18469220399856567, "learning_rate": 6.471039561377581e-06, "loss": 0.4792, "num_input_tokens_seen": 19695336, "step": 30115 }, { "epoch": 15.786163522012579, "grad_norm": 0.10933714359998703, "learning_rate": 6.463365319196538e-06, "loss": 0.488, "num_input_tokens_seen": 19698088, "step": 30120 }, { "epoch": 15.788784067085954, "grad_norm": 0.15081621706485748, "learning_rate": 6.4556949546679905e-06, "loss": 0.48, "num_input_tokens_seen": 19701384, "step": 30125 }, { "epoch": 15.79140461215933, "grad_norm": 0.2449856847524643, "learning_rate": 6.448028469396497e-06, "loss": 0.4366, "num_input_tokens_seen": 19705800, "step": 30130 }, { "epoch": 15.794025157232705, "grad_norm": 0.3470495939254761, "learning_rate": 6.440365864985801e-06, "loss": 0.3743, "num_input_tokens_seen": 19709384, "step": 30135 }, { "epoch": 15.79664570230608, "grad_norm": 0.2108994424343109, "learning_rate": 6.432707143038818e-06, "loss": 0.4275, "num_input_tokens_seen": 19712072, "step": 30140 }, { "epoch": 15.799266247379455, "grad_norm": 0.1323147714138031, "learning_rate": 6.4250523051576834e-06, "loss": 0.4734, "num_input_tokens_seen": 19715528, "step": 30145 }, { "epoch": 15.80188679245283, "grad_norm": 0.11312436312437057, "learning_rate": 6.417401352943686e-06, "loss": 0.5212, "num_input_tokens_seen": 19719016, "step": 30150 }, { "epoch": 15.804507337526205, "grad_norm": 0.16912256181240082, "learning_rate": 6.409754287997322e-06, "loss": 0.3792, "num_input_tokens_seen": 19722344, "step": 30155 }, { "epoch": 15.80712788259958, "grad_norm": 0.09556393325328827, "learning_rate": 6.40211111191828e-06, "loss": 0.5527, "num_input_tokens_seen": 19725768, "step": 30160 }, { "epoch": 15.809748427672956, "grad_norm": 0.21467122435569763, "learning_rate": 6.394471826305409e-06, "loss": 0.6289, "num_input_tokens_seen": 19728200, "step": 30165 }, { "epoch": 15.81236897274633, "grad_norm": 0.1990504264831543, "learning_rate": 6.3868364327567795e-06, "loss": 0.4946, "num_input_tokens_seen": 19730920, "step": 30170 }, { "epoch": 15.814989517819706, "grad_norm": 0.17981427907943726, "learning_rate": 6.379204932869606e-06, "loss": 0.5514, "num_input_tokens_seen": 19733800, "step": 30175 }, { "epoch": 15.817610062893081, "grad_norm": 0.2610228359699249, "learning_rate": 6.371577328240327e-06, "loss": 0.4421, "num_input_tokens_seen": 19736040, "step": 30180 }, { "epoch": 15.820230607966456, "grad_norm": 0.36762818694114685, "learning_rate": 6.363953620464547e-06, "loss": 0.5176, "num_input_tokens_seen": 19739112, "step": 30185 }, { "epoch": 15.822851153039831, "grad_norm": 0.20995382964611053, "learning_rate": 6.356333811137064e-06, "loss": 0.5368, "num_input_tokens_seen": 19742984, "step": 30190 }, { "epoch": 15.825471698113208, "grad_norm": 0.14011850953102112, "learning_rate": 6.34871790185185e-06, "loss": 0.6901, "num_input_tokens_seen": 19746120, "step": 30195 }, { "epoch": 15.828092243186584, "grad_norm": 0.1193254292011261, "learning_rate": 6.341105894202057e-06, "loss": 0.4764, "num_input_tokens_seen": 19749992, "step": 30200 }, { "epoch": 15.830712788259959, "grad_norm": 0.15675170719623566, "learning_rate": 6.333497789780041e-06, "loss": 0.5874, "num_input_tokens_seen": 19753128, "step": 30205 }, { "epoch": 15.833333333333334, "grad_norm": 0.0853499248623848, "learning_rate": 6.325893590177329e-06, "loss": 0.5996, "num_input_tokens_seen": 19756328, "step": 30210 }, { "epoch": 15.835953878406709, "grad_norm": 0.1932818740606308, "learning_rate": 6.318293296984631e-06, "loss": 0.4271, "num_input_tokens_seen": 19760456, "step": 30215 }, { "epoch": 15.838574423480084, "grad_norm": 0.13906313478946686, "learning_rate": 6.3106969117918495e-06, "loss": 0.3851, "num_input_tokens_seen": 19764168, "step": 30220 }, { "epoch": 15.84119496855346, "grad_norm": 0.2115221917629242, "learning_rate": 6.303104436188057e-06, "loss": 0.6133, "num_input_tokens_seen": 19768040, "step": 30225 }, { "epoch": 15.843815513626835, "grad_norm": 0.16997791826725006, "learning_rate": 6.2955158717615036e-06, "loss": 0.4869, "num_input_tokens_seen": 19771144, "step": 30230 }, { "epoch": 15.84643605870021, "grad_norm": 0.2225014567375183, "learning_rate": 6.287931220099638e-06, "loss": 0.4879, "num_input_tokens_seen": 19774536, "step": 30235 }, { "epoch": 15.849056603773585, "grad_norm": 0.17126543819904327, "learning_rate": 6.280350482789082e-06, "loss": 0.3766, "num_input_tokens_seen": 19778024, "step": 30240 }, { "epoch": 15.85167714884696, "grad_norm": 0.14900009334087372, "learning_rate": 6.272773661415645e-06, "loss": 0.5437, "num_input_tokens_seen": 19780936, "step": 30245 }, { "epoch": 15.854297693920335, "grad_norm": 0.14022108912467957, "learning_rate": 6.26520075756431e-06, "loss": 0.3854, "num_input_tokens_seen": 19784392, "step": 30250 }, { "epoch": 15.85691823899371, "grad_norm": 0.1589786261320114, "learning_rate": 6.2576317728192304e-06, "loss": 0.4472, "num_input_tokens_seen": 19787336, "step": 30255 }, { "epoch": 15.859538784067086, "grad_norm": 0.14803864061832428, "learning_rate": 6.250066708763761e-06, "loss": 0.5156, "num_input_tokens_seen": 19790664, "step": 30260 }, { "epoch": 15.86215932914046, "grad_norm": 0.19892048835754395, "learning_rate": 6.242505566980422e-06, "loss": 0.7385, "num_input_tokens_seen": 19793320, "step": 30265 }, { "epoch": 15.864779874213836, "grad_norm": 0.14915801584720612, "learning_rate": 6.234948349050931e-06, "loss": 0.4609, "num_input_tokens_seen": 19796360, "step": 30270 }, { "epoch": 15.867400419287211, "grad_norm": 0.9240593314170837, "learning_rate": 6.227395056556162e-06, "loss": 0.5281, "num_input_tokens_seen": 19803272, "step": 30275 }, { "epoch": 15.870020964360586, "grad_norm": 0.17112582921981812, "learning_rate": 6.219845691076173e-06, "loss": 0.5258, "num_input_tokens_seen": 19812264, "step": 30280 }, { "epoch": 15.872641509433961, "grad_norm": 0.13640007376670837, "learning_rate": 6.212300254190206e-06, "loss": 0.4737, "num_input_tokens_seen": 19815144, "step": 30285 }, { "epoch": 15.875262054507338, "grad_norm": 0.13164666295051575, "learning_rate": 6.204758747476688e-06, "loss": 0.6387, "num_input_tokens_seen": 19819368, "step": 30290 }, { "epoch": 15.877882599580714, "grad_norm": 0.2952955663204193, "learning_rate": 6.1972211725132095e-06, "loss": 0.4345, "num_input_tokens_seen": 19821736, "step": 30295 }, { "epoch": 15.880503144654089, "grad_norm": 0.16300417482852936, "learning_rate": 6.189687530876559e-06, "loss": 0.7038, "num_input_tokens_seen": 19825320, "step": 30300 }, { "epoch": 15.883123689727464, "grad_norm": 0.28026092052459717, "learning_rate": 6.182157824142676e-06, "loss": 0.5316, "num_input_tokens_seen": 19829000, "step": 30305 }, { "epoch": 15.885744234800839, "grad_norm": 0.162939190864563, "learning_rate": 6.174632053886681e-06, "loss": 0.4608, "num_input_tokens_seen": 19831592, "step": 30310 }, { "epoch": 15.888364779874214, "grad_norm": 0.13944275677204132, "learning_rate": 6.167110221682893e-06, "loss": 0.3847, "num_input_tokens_seen": 19834664, "step": 30315 }, { "epoch": 15.89098532494759, "grad_norm": 0.3124698996543884, "learning_rate": 6.159592329104788e-06, "loss": 0.4978, "num_input_tokens_seen": 19837640, "step": 30320 }, { "epoch": 15.893605870020965, "grad_norm": 0.10220777243375778, "learning_rate": 6.152078377725032e-06, "loss": 0.4311, "num_input_tokens_seen": 19841192, "step": 30325 }, { "epoch": 15.89622641509434, "grad_norm": 0.11602038145065308, "learning_rate": 6.144568369115453e-06, "loss": 0.4994, "num_input_tokens_seen": 19845032, "step": 30330 }, { "epoch": 15.898846960167715, "grad_norm": 0.10910898447036743, "learning_rate": 6.137062304847046e-06, "loss": 0.3757, "num_input_tokens_seen": 19848232, "step": 30335 }, { "epoch": 15.90146750524109, "grad_norm": 0.1391504853963852, "learning_rate": 6.129560186490008e-06, "loss": 0.5422, "num_input_tokens_seen": 19851336, "step": 30340 }, { "epoch": 15.904088050314465, "grad_norm": 0.18239669501781464, "learning_rate": 6.122062015613694e-06, "loss": 0.5128, "num_input_tokens_seen": 19856040, "step": 30345 }, { "epoch": 15.90670859538784, "grad_norm": 0.14162370562553406, "learning_rate": 6.114567793786641e-06, "loss": 0.5049, "num_input_tokens_seen": 19859176, "step": 30350 }, { "epoch": 15.909329140461216, "grad_norm": 0.1896304488182068, "learning_rate": 6.107077522576543e-06, "loss": 0.3819, "num_input_tokens_seen": 19862056, "step": 30355 }, { "epoch": 15.91194968553459, "grad_norm": 0.25841397047042847, "learning_rate": 6.099591203550292e-06, "loss": 0.4411, "num_input_tokens_seen": 19864968, "step": 30360 }, { "epoch": 15.914570230607966, "grad_norm": 0.0936349406838417, "learning_rate": 6.092108838273927e-06, "loss": 0.4039, "num_input_tokens_seen": 19868456, "step": 30365 }, { "epoch": 15.917190775681341, "grad_norm": 0.16541625559329987, "learning_rate": 6.08463042831268e-06, "loss": 0.4901, "num_input_tokens_seen": 19872424, "step": 30370 }, { "epoch": 15.919811320754716, "grad_norm": 0.18587489426136017, "learning_rate": 6.0771559752309496e-06, "loss": 0.5123, "num_input_tokens_seen": 19875400, "step": 30375 }, { "epoch": 15.922431865828091, "grad_norm": 0.21119649708271027, "learning_rate": 6.069685480592313e-06, "loss": 0.3911, "num_input_tokens_seen": 19878184, "step": 30380 }, { "epoch": 15.925052410901468, "grad_norm": 0.24608221650123596, "learning_rate": 6.062218945959497e-06, "loss": 0.5642, "num_input_tokens_seen": 19880744, "step": 30385 }, { "epoch": 15.927672955974844, "grad_norm": 0.12586864829063416, "learning_rate": 6.054756372894435e-06, "loss": 0.504, "num_input_tokens_seen": 19883784, "step": 30390 }, { "epoch": 15.930293501048219, "grad_norm": 0.26836714148521423, "learning_rate": 6.04729776295819e-06, "loss": 0.5558, "num_input_tokens_seen": 19887304, "step": 30395 }, { "epoch": 15.932914046121594, "grad_norm": 0.14239759743213654, "learning_rate": 6.0398431177110306e-06, "loss": 0.537, "num_input_tokens_seen": 19890312, "step": 30400 }, { "epoch": 15.935534591194969, "grad_norm": 0.11377755552530289, "learning_rate": 6.032392438712389e-06, "loss": 0.4264, "num_input_tokens_seen": 19893704, "step": 30405 }, { "epoch": 15.938155136268344, "grad_norm": 0.29929783940315247, "learning_rate": 6.024945727520847e-06, "loss": 0.4044, "num_input_tokens_seen": 19896552, "step": 30410 }, { "epoch": 15.94077568134172, "grad_norm": 0.314374715089798, "learning_rate": 6.01750298569419e-06, "loss": 0.5551, "num_input_tokens_seen": 19899592, "step": 30415 }, { "epoch": 15.943396226415095, "grad_norm": 0.2217947393655777, "learning_rate": 6.010064214789335e-06, "loss": 0.5646, "num_input_tokens_seen": 19902952, "step": 30420 }, { "epoch": 15.94601677148847, "grad_norm": 0.1406683474779129, "learning_rate": 6.002629416362399e-06, "loss": 0.4503, "num_input_tokens_seen": 19906344, "step": 30425 }, { "epoch": 15.948637316561845, "grad_norm": 0.17375175654888153, "learning_rate": 5.995198591968662e-06, "loss": 0.4942, "num_input_tokens_seen": 19909864, "step": 30430 }, { "epoch": 15.95125786163522, "grad_norm": 0.15055330097675323, "learning_rate": 5.987771743162554e-06, "loss": 0.448, "num_input_tokens_seen": 19913160, "step": 30435 }, { "epoch": 15.953878406708595, "grad_norm": 0.1998915821313858, "learning_rate": 5.9803488714976955e-06, "loss": 0.4616, "num_input_tokens_seen": 19917160, "step": 30440 }, { "epoch": 15.95649895178197, "grad_norm": 0.10486982017755508, "learning_rate": 5.972929978526872e-06, "loss": 0.4366, "num_input_tokens_seen": 19921672, "step": 30445 }, { "epoch": 15.959119496855346, "grad_norm": 0.12995590269565582, "learning_rate": 5.965515065802019e-06, "loss": 0.4465, "num_input_tokens_seen": 19924392, "step": 30450 }, { "epoch": 15.96174004192872, "grad_norm": 0.1517358422279358, "learning_rate": 5.958104134874254e-06, "loss": 0.4515, "num_input_tokens_seen": 19927720, "step": 30455 }, { "epoch": 15.964360587002096, "grad_norm": 0.2724747955799103, "learning_rate": 5.950697187293872e-06, "loss": 0.5232, "num_input_tokens_seen": 19931112, "step": 30460 }, { "epoch": 15.966981132075471, "grad_norm": 0.13836534321308136, "learning_rate": 5.943294224610305e-06, "loss": 0.4174, "num_input_tokens_seen": 19934152, "step": 30465 }, { "epoch": 15.969601677148846, "grad_norm": 0.18578946590423584, "learning_rate": 5.935895248372175e-06, "loss": 0.5541, "num_input_tokens_seen": 19937160, "step": 30470 }, { "epoch": 15.972222222222221, "grad_norm": 0.13032633066177368, "learning_rate": 5.928500260127273e-06, "loss": 0.5593, "num_input_tokens_seen": 19940424, "step": 30475 }, { "epoch": 15.974842767295598, "grad_norm": 0.33406728506088257, "learning_rate": 5.921109261422531e-06, "loss": 0.408, "num_input_tokens_seen": 19942728, "step": 30480 }, { "epoch": 15.977463312368974, "grad_norm": 0.18584761023521423, "learning_rate": 5.913722253804071e-06, "loss": 0.5313, "num_input_tokens_seen": 19946056, "step": 30485 }, { "epoch": 15.980083857442349, "grad_norm": 0.16721118986606598, "learning_rate": 5.906339238817163e-06, "loss": 0.5712, "num_input_tokens_seen": 19950632, "step": 30490 }, { "epoch": 15.982704402515724, "grad_norm": 0.17515921592712402, "learning_rate": 5.8989602180062554e-06, "loss": 0.4225, "num_input_tokens_seen": 19954280, "step": 30495 }, { "epoch": 15.985324947589099, "grad_norm": 0.28043562173843384, "learning_rate": 5.891585192914953e-06, "loss": 0.3693, "num_input_tokens_seen": 19957768, "step": 30500 }, { "epoch": 15.987945492662474, "grad_norm": 0.22971521317958832, "learning_rate": 5.884214165086036e-06, "loss": 0.5179, "num_input_tokens_seen": 19960328, "step": 30505 }, { "epoch": 15.99056603773585, "grad_norm": 0.1479586511850357, "learning_rate": 5.876847136061428e-06, "loss": 0.4653, "num_input_tokens_seen": 19963272, "step": 30510 }, { "epoch": 15.993186582809225, "grad_norm": 0.12173083424568176, "learning_rate": 5.869484107382228e-06, "loss": 0.3744, "num_input_tokens_seen": 19966376, "step": 30515 }, { "epoch": 15.9958071278826, "grad_norm": 0.17523299157619476, "learning_rate": 5.862125080588696e-06, "loss": 0.5305, "num_input_tokens_seen": 19968968, "step": 30520 }, { "epoch": 15.998427672955975, "grad_norm": 0.16300156712532043, "learning_rate": 5.854770057220263e-06, "loss": 0.3421, "num_input_tokens_seen": 19971624, "step": 30525 }, { "epoch": 16.0, "eval_loss": 0.4605244994163513, "eval_runtime": 9.9496, "eval_samples_per_second": 85.23, "eval_steps_per_second": 21.307, "num_input_tokens_seen": 19973408, "step": 30528 }, { "epoch": 16.00104821802935, "grad_norm": 0.2071686089038849, "learning_rate": 5.8474190388155216e-06, "loss": 0.5123, "num_input_tokens_seen": 19974496, "step": 30530 }, { "epoch": 16.003668763102727, "grad_norm": 0.14605481922626495, "learning_rate": 5.840072026912205e-06, "loss": 0.4484, "num_input_tokens_seen": 19977888, "step": 30535 }, { "epoch": 16.0062893081761, "grad_norm": 0.37048307061195374, "learning_rate": 5.83272902304724e-06, "loss": 0.4481, "num_input_tokens_seen": 19980480, "step": 30540 }, { "epoch": 16.008909853249477, "grad_norm": 0.14462046325206757, "learning_rate": 5.825390028756683e-06, "loss": 0.5573, "num_input_tokens_seen": 19983200, "step": 30545 }, { "epoch": 16.01153039832285, "grad_norm": 0.17380540072917938, "learning_rate": 5.818055045575776e-06, "loss": 0.4423, "num_input_tokens_seen": 19987200, "step": 30550 }, { "epoch": 16.014150943396228, "grad_norm": 0.25728869438171387, "learning_rate": 5.810724075038917e-06, "loss": 0.5105, "num_input_tokens_seen": 19989920, "step": 30555 }, { "epoch": 16.0167714884696, "grad_norm": 0.12649425864219666, "learning_rate": 5.803397118679665e-06, "loss": 0.4292, "num_input_tokens_seen": 19993376, "step": 30560 }, { "epoch": 16.019392033542978, "grad_norm": 0.34242677688598633, "learning_rate": 5.796074178030727e-06, "loss": 0.4626, "num_input_tokens_seen": 19995264, "step": 30565 }, { "epoch": 16.02201257861635, "grad_norm": 0.17120347917079926, "learning_rate": 5.788755254623976e-06, "loss": 0.6482, "num_input_tokens_seen": 19997824, "step": 30570 }, { "epoch": 16.02463312368973, "grad_norm": 0.2814457416534424, "learning_rate": 5.7814403499904525e-06, "loss": 0.5183, "num_input_tokens_seen": 20000640, "step": 30575 }, { "epoch": 16.0272536687631, "grad_norm": 0.2558935582637787, "learning_rate": 5.774129465660352e-06, "loss": 0.4133, "num_input_tokens_seen": 20003264, "step": 30580 }, { "epoch": 16.02987421383648, "grad_norm": 0.12366840988397598, "learning_rate": 5.7668226031630246e-06, "loss": 0.4804, "num_input_tokens_seen": 20006688, "step": 30585 }, { "epoch": 16.032494758909852, "grad_norm": 0.310158908367157, "learning_rate": 5.759519764026999e-06, "loss": 0.4976, "num_input_tokens_seen": 20009280, "step": 30590 }, { "epoch": 16.03511530398323, "grad_norm": 0.22367094457149506, "learning_rate": 5.7522209497799185e-06, "loss": 0.4919, "num_input_tokens_seen": 20012960, "step": 30595 }, { "epoch": 16.037735849056602, "grad_norm": 0.29588040709495544, "learning_rate": 5.7449261619486246e-06, "loss": 0.5426, "num_input_tokens_seen": 20015616, "step": 30600 }, { "epoch": 16.04035639412998, "grad_norm": 0.1276334524154663, "learning_rate": 5.737635402059105e-06, "loss": 0.5728, "num_input_tokens_seen": 20018048, "step": 30605 }, { "epoch": 16.042976939203353, "grad_norm": 0.24486806988716125, "learning_rate": 5.730348671636501e-06, "loss": 0.4794, "num_input_tokens_seen": 20021152, "step": 30610 }, { "epoch": 16.04559748427673, "grad_norm": 0.21602536737918854, "learning_rate": 5.723065972205124e-06, "loss": 0.4241, "num_input_tokens_seen": 20024032, "step": 30615 }, { "epoch": 16.048218029350103, "grad_norm": 0.18283942341804504, "learning_rate": 5.715787305288423e-06, "loss": 0.4495, "num_input_tokens_seen": 20026944, "step": 30620 }, { "epoch": 16.05083857442348, "grad_norm": 0.20460328459739685, "learning_rate": 5.7085126724090024e-06, "loss": 0.4948, "num_input_tokens_seen": 20029536, "step": 30625 }, { "epoch": 16.053459119496857, "grad_norm": 0.19383595883846283, "learning_rate": 5.7012420750886445e-06, "loss": 0.4493, "num_input_tokens_seen": 20032768, "step": 30630 }, { "epoch": 16.05607966457023, "grad_norm": 0.14867785573005676, "learning_rate": 5.693975514848271e-06, "loss": 0.5234, "num_input_tokens_seen": 20035712, "step": 30635 }, { "epoch": 16.058700209643607, "grad_norm": 0.19187133014202118, "learning_rate": 5.686712993207971e-06, "loss": 0.5277, "num_input_tokens_seen": 20038336, "step": 30640 }, { "epoch": 16.06132075471698, "grad_norm": 0.13164806365966797, "learning_rate": 5.679454511686977e-06, "loss": 0.4759, "num_input_tokens_seen": 20042688, "step": 30645 }, { "epoch": 16.063941299790358, "grad_norm": 0.12563857436180115, "learning_rate": 5.672200071803671e-06, "loss": 0.506, "num_input_tokens_seen": 20045568, "step": 30650 }, { "epoch": 16.06656184486373, "grad_norm": 0.1615312397480011, "learning_rate": 5.664949675075607e-06, "loss": 0.5039, "num_input_tokens_seen": 20048928, "step": 30655 }, { "epoch": 16.069182389937108, "grad_norm": 0.1143355444073677, "learning_rate": 5.657703323019486e-06, "loss": 0.5001, "num_input_tokens_seen": 20051840, "step": 30660 }, { "epoch": 16.07180293501048, "grad_norm": 0.11271457374095917, "learning_rate": 5.65046101715116e-06, "loss": 0.513, "num_input_tokens_seen": 20054592, "step": 30665 }, { "epoch": 16.07442348008386, "grad_norm": 0.15148966014385223, "learning_rate": 5.643222758985644e-06, "loss": 0.6436, "num_input_tokens_seen": 20058624, "step": 30670 }, { "epoch": 16.07704402515723, "grad_norm": 0.15159045159816742, "learning_rate": 5.635988550037094e-06, "loss": 0.4098, "num_input_tokens_seen": 20061632, "step": 30675 }, { "epoch": 16.07966457023061, "grad_norm": 0.10855105519294739, "learning_rate": 5.628758391818817e-06, "loss": 0.3843, "num_input_tokens_seen": 20065472, "step": 30680 }, { "epoch": 16.082285115303982, "grad_norm": 0.16376063227653503, "learning_rate": 5.621532285843287e-06, "loss": 0.477, "num_input_tokens_seen": 20068128, "step": 30685 }, { "epoch": 16.08490566037736, "grad_norm": 0.12111061811447144, "learning_rate": 5.614310233622119e-06, "loss": 0.4874, "num_input_tokens_seen": 20071872, "step": 30690 }, { "epoch": 16.087526205450732, "grad_norm": 0.1648908257484436, "learning_rate": 5.607092236666093e-06, "loss": 0.5665, "num_input_tokens_seen": 20075232, "step": 30695 }, { "epoch": 16.09014675052411, "grad_norm": 0.1244032233953476, "learning_rate": 5.599878296485125e-06, "loss": 0.4146, "num_input_tokens_seen": 20078272, "step": 30700 }, { "epoch": 16.092767295597483, "grad_norm": 0.24332255125045776, "learning_rate": 5.592668414588284e-06, "loss": 0.5862, "num_input_tokens_seen": 20082304, "step": 30705 }, { "epoch": 16.09538784067086, "grad_norm": 0.199802428483963, "learning_rate": 5.585462592483801e-06, "loss": 0.4596, "num_input_tokens_seen": 20086112, "step": 30710 }, { "epoch": 16.098008385744233, "grad_norm": 0.15532249212265015, "learning_rate": 5.578260831679047e-06, "loss": 0.3886, "num_input_tokens_seen": 20090272, "step": 30715 }, { "epoch": 16.10062893081761, "grad_norm": 0.16247323155403137, "learning_rate": 5.571063133680562e-06, "loss": 0.5554, "num_input_tokens_seen": 20093280, "step": 30720 }, { "epoch": 16.103249475890987, "grad_norm": 0.14368897676467896, "learning_rate": 5.563869499994004e-06, "loss": 0.3829, "num_input_tokens_seen": 20096128, "step": 30725 }, { "epoch": 16.10587002096436, "grad_norm": 0.22163796424865723, "learning_rate": 5.556679932124215e-06, "loss": 0.4652, "num_input_tokens_seen": 20099136, "step": 30730 }, { "epoch": 16.108490566037737, "grad_norm": 0.08700559288263321, "learning_rate": 5.549494431575159e-06, "loss": 0.5056, "num_input_tokens_seen": 20102240, "step": 30735 }, { "epoch": 16.11111111111111, "grad_norm": 0.3983261287212372, "learning_rate": 5.542312999849961e-06, "loss": 0.6359, "num_input_tokens_seen": 20104960, "step": 30740 }, { "epoch": 16.113731656184488, "grad_norm": 0.1841597855091095, "learning_rate": 5.5351356384509e-06, "loss": 0.4671, "num_input_tokens_seen": 20108288, "step": 30745 }, { "epoch": 16.11635220125786, "grad_norm": 0.1919950693845749, "learning_rate": 5.527962348879401e-06, "loss": 0.4717, "num_input_tokens_seen": 20111072, "step": 30750 }, { "epoch": 16.118972746331238, "grad_norm": 0.17315664887428284, "learning_rate": 5.520793132636027e-06, "loss": 0.5977, "num_input_tokens_seen": 20114240, "step": 30755 }, { "epoch": 16.12159329140461, "grad_norm": 0.08593977242708206, "learning_rate": 5.513627991220502e-06, "loss": 0.426, "num_input_tokens_seen": 20117920, "step": 30760 }, { "epoch": 16.12421383647799, "grad_norm": 0.14697057008743286, "learning_rate": 5.506466926131684e-06, "loss": 0.5481, "num_input_tokens_seen": 20121056, "step": 30765 }, { "epoch": 16.12683438155136, "grad_norm": 0.16962498426437378, "learning_rate": 5.4993099388675916e-06, "loss": 0.5506, "num_input_tokens_seen": 20124160, "step": 30770 }, { "epoch": 16.12945492662474, "grad_norm": 0.147096648812294, "learning_rate": 5.49215703092539e-06, "loss": 0.4742, "num_input_tokens_seen": 20127264, "step": 30775 }, { "epoch": 16.132075471698112, "grad_norm": 0.1353737860918045, "learning_rate": 5.4850082038013755e-06, "loss": 0.5158, "num_input_tokens_seen": 20129760, "step": 30780 }, { "epoch": 16.13469601677149, "grad_norm": 0.1327427625656128, "learning_rate": 5.4778634589910115e-06, "loss": 0.4778, "num_input_tokens_seen": 20133056, "step": 30785 }, { "epoch": 16.137316561844862, "grad_norm": 0.16662472486495972, "learning_rate": 5.4707227979888834e-06, "loss": 0.7169, "num_input_tokens_seen": 20139776, "step": 30790 }, { "epoch": 16.13993710691824, "grad_norm": 0.2536458671092987, "learning_rate": 5.463586222288747e-06, "loss": 0.4379, "num_input_tokens_seen": 20143680, "step": 30795 }, { "epoch": 16.142557651991613, "grad_norm": 0.26045945286750793, "learning_rate": 5.456453733383493e-06, "loss": 0.4716, "num_input_tokens_seen": 20147776, "step": 30800 }, { "epoch": 16.14517819706499, "grad_norm": 0.10390490293502808, "learning_rate": 5.449325332765151e-06, "loss": 0.4689, "num_input_tokens_seen": 20151680, "step": 30805 }, { "epoch": 16.147798742138363, "grad_norm": 0.1605377048254013, "learning_rate": 5.4422010219249e-06, "loss": 0.3813, "num_input_tokens_seen": 20155712, "step": 30810 }, { "epoch": 16.15041928721174, "grad_norm": 0.19971998035907745, "learning_rate": 5.435080802353077e-06, "loss": 0.3779, "num_input_tokens_seen": 20158592, "step": 30815 }, { "epoch": 16.153039832285117, "grad_norm": 0.11341574788093567, "learning_rate": 5.4279646755391325e-06, "loss": 0.4174, "num_input_tokens_seen": 20162752, "step": 30820 }, { "epoch": 16.15566037735849, "grad_norm": 0.14862892031669617, "learning_rate": 5.420852642971691e-06, "loss": 0.5963, "num_input_tokens_seen": 20165440, "step": 30825 }, { "epoch": 16.158280922431867, "grad_norm": 0.13110655546188354, "learning_rate": 5.41374470613851e-06, "loss": 0.5322, "num_input_tokens_seen": 20167936, "step": 30830 }, { "epoch": 16.16090146750524, "grad_norm": 0.24828726053237915, "learning_rate": 5.40664086652648e-06, "loss": 0.2985, "num_input_tokens_seen": 20170400, "step": 30835 }, { "epoch": 16.163522012578618, "grad_norm": 0.12304413318634033, "learning_rate": 5.399541125621646e-06, "loss": 0.4451, "num_input_tokens_seen": 20173792, "step": 30840 }, { "epoch": 16.16614255765199, "grad_norm": 0.2064286768436432, "learning_rate": 5.392445484909203e-06, "loss": 0.4425, "num_input_tokens_seen": 20177088, "step": 30845 }, { "epoch": 16.168763102725368, "grad_norm": 0.18801581859588623, "learning_rate": 5.385353945873464e-06, "loss": 0.4098, "num_input_tokens_seen": 20181088, "step": 30850 }, { "epoch": 16.17138364779874, "grad_norm": 0.1563880890607834, "learning_rate": 5.378266509997909e-06, "loss": 0.38, "num_input_tokens_seen": 20183808, "step": 30855 }, { "epoch": 16.17400419287212, "grad_norm": 0.15051105618476868, "learning_rate": 5.371183178765135e-06, "loss": 0.453, "num_input_tokens_seen": 20186656, "step": 30860 }, { "epoch": 16.17662473794549, "grad_norm": 0.14185954630374908, "learning_rate": 5.364103953656904e-06, "loss": 0.3867, "num_input_tokens_seen": 20190688, "step": 30865 }, { "epoch": 16.17924528301887, "grad_norm": 0.09941218048334122, "learning_rate": 5.35702883615411e-06, "loss": 0.509, "num_input_tokens_seen": 20194560, "step": 30870 }, { "epoch": 16.181865828092242, "grad_norm": 0.2788904309272766, "learning_rate": 5.34995782773679e-06, "loss": 0.5527, "num_input_tokens_seen": 20197184, "step": 30875 }, { "epoch": 16.18448637316562, "grad_norm": 0.15502162277698517, "learning_rate": 5.3428909298841115e-06, "loss": 0.4514, "num_input_tokens_seen": 20200640, "step": 30880 }, { "epoch": 16.187106918238992, "grad_norm": 0.20243194699287415, "learning_rate": 5.335828144074384e-06, "loss": 0.6678, "num_input_tokens_seen": 20203488, "step": 30885 }, { "epoch": 16.18972746331237, "grad_norm": 0.20323975384235382, "learning_rate": 5.3287694717850695e-06, "loss": 0.3962, "num_input_tokens_seen": 20208800, "step": 30890 }, { "epoch": 16.192348008385743, "grad_norm": 0.160000741481781, "learning_rate": 5.321714914492759e-06, "loss": 0.3508, "num_input_tokens_seen": 20213024, "step": 30895 }, { "epoch": 16.19496855345912, "grad_norm": 0.2711268663406372, "learning_rate": 5.314664473673195e-06, "loss": 0.566, "num_input_tokens_seen": 20215968, "step": 30900 }, { "epoch": 16.197589098532493, "grad_norm": 0.10989008098840714, "learning_rate": 5.307618150801239e-06, "loss": 0.4502, "num_input_tokens_seen": 20219264, "step": 30905 }, { "epoch": 16.20020964360587, "grad_norm": 0.1400270164012909, "learning_rate": 5.300575947350897e-06, "loss": 0.4899, "num_input_tokens_seen": 20222208, "step": 30910 }, { "epoch": 16.202830188679247, "grad_norm": 0.1816757619380951, "learning_rate": 5.293537864795325e-06, "loss": 0.4511, "num_input_tokens_seen": 20225760, "step": 30915 }, { "epoch": 16.20545073375262, "grad_norm": 0.1358131766319275, "learning_rate": 5.2865039046068075e-06, "loss": 0.5244, "num_input_tokens_seen": 20229120, "step": 30920 }, { "epoch": 16.208071278825997, "grad_norm": 0.2097487896680832, "learning_rate": 5.279474068256767e-06, "loss": 0.6025, "num_input_tokens_seen": 20231744, "step": 30925 }, { "epoch": 16.21069182389937, "grad_norm": 0.2091621607542038, "learning_rate": 5.272448357215779e-06, "loss": 0.6081, "num_input_tokens_seen": 20235808, "step": 30930 }, { "epoch": 16.213312368972748, "grad_norm": 0.21000367403030396, "learning_rate": 5.26542677295353e-06, "loss": 0.5768, "num_input_tokens_seen": 20238784, "step": 30935 }, { "epoch": 16.21593291404612, "grad_norm": 0.15234775841236115, "learning_rate": 5.258409316938848e-06, "loss": 0.4987, "num_input_tokens_seen": 20241952, "step": 30940 }, { "epoch": 16.218553459119498, "grad_norm": 0.242877796292305, "learning_rate": 5.251395990639713e-06, "loss": 0.6373, "num_input_tokens_seen": 20244608, "step": 30945 }, { "epoch": 16.22117400419287, "grad_norm": 0.15433768928050995, "learning_rate": 5.244386795523235e-06, "loss": 0.4023, "num_input_tokens_seen": 20247424, "step": 30950 }, { "epoch": 16.22379454926625, "grad_norm": 0.14572018384933472, "learning_rate": 5.23738173305566e-06, "loss": 0.4274, "num_input_tokens_seen": 20250240, "step": 30955 }, { "epoch": 16.22641509433962, "grad_norm": 0.1316547840833664, "learning_rate": 5.230380804702362e-06, "loss": 0.36, "num_input_tokens_seen": 20253440, "step": 30960 }, { "epoch": 16.229035639413, "grad_norm": 0.1438085287809372, "learning_rate": 5.223384011927851e-06, "loss": 0.4868, "num_input_tokens_seen": 20257440, "step": 30965 }, { "epoch": 16.231656184486372, "grad_norm": 0.4745487868785858, "learning_rate": 5.216391356195779e-06, "loss": 0.683, "num_input_tokens_seen": 20260320, "step": 30970 }, { "epoch": 16.23427672955975, "grad_norm": 0.216372549533844, "learning_rate": 5.209402838968935e-06, "loss": 0.7588, "num_input_tokens_seen": 20263424, "step": 30975 }, { "epoch": 16.236897274633122, "grad_norm": 0.11222842335700989, "learning_rate": 5.202418461709233e-06, "loss": 0.3654, "num_input_tokens_seen": 20266752, "step": 30980 }, { "epoch": 16.2395178197065, "grad_norm": 0.12281301617622375, "learning_rate": 5.195438225877738e-06, "loss": 0.464, "num_input_tokens_seen": 20269792, "step": 30985 }, { "epoch": 16.242138364779873, "grad_norm": 0.3118172585964203, "learning_rate": 5.18846213293461e-06, "loss": 0.5083, "num_input_tokens_seen": 20272608, "step": 30990 }, { "epoch": 16.24475890985325, "grad_norm": 0.19394294917583466, "learning_rate": 5.181490184339185e-06, "loss": 0.3565, "num_input_tokens_seen": 20278688, "step": 30995 }, { "epoch": 16.247379454926623, "grad_norm": 0.22766350209712982, "learning_rate": 5.17452238154991e-06, "loss": 0.3453, "num_input_tokens_seen": 20280736, "step": 31000 }, { "epoch": 16.25, "grad_norm": 0.1496267467737198, "learning_rate": 5.1675587260243715e-06, "loss": 0.5449, "num_input_tokens_seen": 20284800, "step": 31005 }, { "epoch": 16.252620545073377, "grad_norm": 0.19443167746067047, "learning_rate": 5.160599219219295e-06, "loss": 0.6313, "num_input_tokens_seen": 20288608, "step": 31010 }, { "epoch": 16.25524109014675, "grad_norm": 0.2020690143108368, "learning_rate": 5.1536438625905185e-06, "loss": 0.5255, "num_input_tokens_seen": 20291712, "step": 31015 }, { "epoch": 16.257861635220127, "grad_norm": 0.23984819650650024, "learning_rate": 5.146692657593025e-06, "loss": 0.4725, "num_input_tokens_seen": 20294880, "step": 31020 }, { "epoch": 16.2604821802935, "grad_norm": 0.19222915172576904, "learning_rate": 5.139745605680926e-06, "loss": 0.3999, "num_input_tokens_seen": 20297824, "step": 31025 }, { "epoch": 16.263102725366878, "grad_norm": 0.22163954377174377, "learning_rate": 5.132802708307469e-06, "loss": 0.4885, "num_input_tokens_seen": 20300672, "step": 31030 }, { "epoch": 16.26572327044025, "grad_norm": 0.11488968133926392, "learning_rate": 5.125863966925035e-06, "loss": 0.4434, "num_input_tokens_seen": 20303232, "step": 31035 }, { "epoch": 16.268343815513628, "grad_norm": 0.14427363872528076, "learning_rate": 5.118929382985116e-06, "loss": 0.4971, "num_input_tokens_seen": 20307232, "step": 31040 }, { "epoch": 16.270964360587, "grad_norm": 0.15324236452579498, "learning_rate": 5.111998957938363e-06, "loss": 0.4941, "num_input_tokens_seen": 20310912, "step": 31045 }, { "epoch": 16.27358490566038, "grad_norm": 0.149980828166008, "learning_rate": 5.105072693234525e-06, "loss": 0.558, "num_input_tokens_seen": 20313760, "step": 31050 }, { "epoch": 16.27620545073375, "grad_norm": 0.38846710324287415, "learning_rate": 5.098150590322506e-06, "loss": 0.4452, "num_input_tokens_seen": 20316960, "step": 31055 }, { "epoch": 16.27882599580713, "grad_norm": 0.25605687499046326, "learning_rate": 5.091232650650335e-06, "loss": 0.4563, "num_input_tokens_seen": 20319552, "step": 31060 }, { "epoch": 16.281446540880502, "grad_norm": 0.17281916737556458, "learning_rate": 5.0843188756651645e-06, "loss": 0.5172, "num_input_tokens_seen": 20322656, "step": 31065 }, { "epoch": 16.28406708595388, "grad_norm": 0.19800688326358795, "learning_rate": 5.077409266813276e-06, "loss": 0.5687, "num_input_tokens_seen": 20325376, "step": 31070 }, { "epoch": 16.286687631027252, "grad_norm": 0.1343289017677307, "learning_rate": 5.0705038255400765e-06, "loss": 0.4536, "num_input_tokens_seen": 20328960, "step": 31075 }, { "epoch": 16.28930817610063, "grad_norm": 0.13042905926704407, "learning_rate": 5.063602553290106e-06, "loss": 0.3833, "num_input_tokens_seen": 20332608, "step": 31080 }, { "epoch": 16.291928721174003, "grad_norm": 0.18041086196899414, "learning_rate": 5.056705451507035e-06, "loss": 0.4361, "num_input_tokens_seen": 20335744, "step": 31085 }, { "epoch": 16.29454926624738, "grad_norm": 0.15480414032936096, "learning_rate": 5.049812521633668e-06, "loss": 0.4754, "num_input_tokens_seen": 20339520, "step": 31090 }, { "epoch": 16.297169811320753, "grad_norm": 0.23264233767986298, "learning_rate": 5.042923765111907e-06, "loss": 0.6397, "num_input_tokens_seen": 20341952, "step": 31095 }, { "epoch": 16.29979035639413, "grad_norm": 0.10560127347707748, "learning_rate": 5.036039183382821e-06, "loss": 0.3834, "num_input_tokens_seen": 20346112, "step": 31100 }, { "epoch": 16.302410901467507, "grad_norm": 0.1282140165567398, "learning_rate": 5.02915877788657e-06, "loss": 0.5122, "num_input_tokens_seen": 20348224, "step": 31105 }, { "epoch": 16.30503144654088, "grad_norm": 0.20184049010276794, "learning_rate": 5.0222825500624625e-06, "loss": 0.511, "num_input_tokens_seen": 20351040, "step": 31110 }, { "epoch": 16.307651991614257, "grad_norm": 0.09292982518672943, "learning_rate": 5.015410501348935e-06, "loss": 0.3206, "num_input_tokens_seen": 20356032, "step": 31115 }, { "epoch": 16.31027253668763, "grad_norm": 0.1551925241947174, "learning_rate": 5.008542633183527e-06, "loss": 0.4076, "num_input_tokens_seen": 20359936, "step": 31120 }, { "epoch": 16.312893081761008, "grad_norm": 0.15959839522838593, "learning_rate": 5.001678947002927e-06, "loss": 0.3698, "num_input_tokens_seen": 20362880, "step": 31125 }, { "epoch": 16.31551362683438, "grad_norm": 0.3306739330291748, "learning_rate": 4.994819444242943e-06, "loss": 0.4345, "num_input_tokens_seen": 20366784, "step": 31130 }, { "epoch": 16.318134171907758, "grad_norm": 0.12682242691516876, "learning_rate": 4.9879641263384956e-06, "loss": 0.5694, "num_input_tokens_seen": 20370496, "step": 31135 }, { "epoch": 16.32075471698113, "grad_norm": 0.2165040522813797, "learning_rate": 4.981112994723644e-06, "loss": 0.4526, "num_input_tokens_seen": 20374016, "step": 31140 }, { "epoch": 16.32337526205451, "grad_norm": 0.11575154215097427, "learning_rate": 4.974266050831572e-06, "loss": 0.7158, "num_input_tokens_seen": 20377280, "step": 31145 }, { "epoch": 16.32599580712788, "grad_norm": 0.31972721219062805, "learning_rate": 4.967423296094567e-06, "loss": 0.489, "num_input_tokens_seen": 20379872, "step": 31150 }, { "epoch": 16.32861635220126, "grad_norm": 0.21713867783546448, "learning_rate": 4.960584731944076e-06, "loss": 0.5418, "num_input_tokens_seen": 20382080, "step": 31155 }, { "epoch": 16.331236897274632, "grad_norm": 0.13542483747005463, "learning_rate": 4.953750359810627e-06, "loss": 0.4743, "num_input_tokens_seen": 20384992, "step": 31160 }, { "epoch": 16.33385744234801, "grad_norm": 0.12319041043519974, "learning_rate": 4.946920181123904e-06, "loss": 0.4111, "num_input_tokens_seen": 20389280, "step": 31165 }, { "epoch": 16.336477987421382, "grad_norm": 0.2162010371685028, "learning_rate": 4.940094197312709e-06, "loss": 0.5298, "num_input_tokens_seen": 20391808, "step": 31170 }, { "epoch": 16.33909853249476, "grad_norm": 0.18672581017017365, "learning_rate": 4.933272409804946e-06, "loss": 0.4233, "num_input_tokens_seen": 20394304, "step": 31175 }, { "epoch": 16.341719077568133, "grad_norm": 0.10825641453266144, "learning_rate": 4.926454820027663e-06, "loss": 0.3858, "num_input_tokens_seen": 20398528, "step": 31180 }, { "epoch": 16.34433962264151, "grad_norm": 0.21652600169181824, "learning_rate": 4.919641429407024e-06, "loss": 0.5874, "num_input_tokens_seen": 20400864, "step": 31185 }, { "epoch": 16.346960167714883, "grad_norm": 0.11153803020715714, "learning_rate": 4.912832239368304e-06, "loss": 0.3883, "num_input_tokens_seen": 20403904, "step": 31190 }, { "epoch": 16.34958071278826, "grad_norm": 0.21196526288986206, "learning_rate": 4.906027251335918e-06, "loss": 0.444, "num_input_tokens_seen": 20407360, "step": 31195 }, { "epoch": 16.352201257861637, "grad_norm": 0.1747768223285675, "learning_rate": 4.899226466733381e-06, "loss": 0.4132, "num_input_tokens_seen": 20409728, "step": 31200 }, { "epoch": 16.35482180293501, "grad_norm": 0.17040172219276428, "learning_rate": 4.892429886983346e-06, "loss": 0.4127, "num_input_tokens_seen": 20412800, "step": 31205 }, { "epoch": 16.357442348008387, "grad_norm": 0.2297702133655548, "learning_rate": 4.88563751350758e-06, "loss": 0.3842, "num_input_tokens_seen": 20415680, "step": 31210 }, { "epoch": 16.36006289308176, "grad_norm": 0.12443127483129501, "learning_rate": 4.878849347726977e-06, "loss": 0.4636, "num_input_tokens_seen": 20419744, "step": 31215 }, { "epoch": 16.362683438155138, "grad_norm": 0.12396614253520966, "learning_rate": 4.872065391061526e-06, "loss": 0.7291, "num_input_tokens_seen": 20422304, "step": 31220 }, { "epoch": 16.36530398322851, "grad_norm": 0.35739371180534363, "learning_rate": 4.865285644930373e-06, "loss": 0.3968, "num_input_tokens_seen": 20424832, "step": 31225 }, { "epoch": 16.367924528301888, "grad_norm": 0.1361643671989441, "learning_rate": 4.858510110751746e-06, "loss": 0.4347, "num_input_tokens_seen": 20428256, "step": 31230 }, { "epoch": 16.37054507337526, "grad_norm": 0.3204270899295807, "learning_rate": 4.851738789943019e-06, "loss": 0.4437, "num_input_tokens_seen": 20430784, "step": 31235 }, { "epoch": 16.37316561844864, "grad_norm": 0.1614740937948227, "learning_rate": 4.844971683920674e-06, "loss": 0.531, "num_input_tokens_seen": 20433760, "step": 31240 }, { "epoch": 16.37578616352201, "grad_norm": 0.19032804667949677, "learning_rate": 4.838208794100316e-06, "loss": 0.3997, "num_input_tokens_seen": 20436416, "step": 31245 }, { "epoch": 16.37840670859539, "grad_norm": 0.22828681766986847, "learning_rate": 4.831450121896661e-06, "loss": 0.3803, "num_input_tokens_seen": 20439328, "step": 31250 }, { "epoch": 16.381027253668762, "grad_norm": 0.12931720912456512, "learning_rate": 4.824695668723542e-06, "loss": 0.5906, "num_input_tokens_seen": 20442016, "step": 31255 }, { "epoch": 16.38364779874214, "grad_norm": 0.12444164603948593, "learning_rate": 4.817945435993915e-06, "loss": 0.4688, "num_input_tokens_seen": 20444800, "step": 31260 }, { "epoch": 16.386268343815512, "grad_norm": 0.29472827911376953, "learning_rate": 4.811199425119853e-06, "loss": 0.3649, "num_input_tokens_seen": 20447456, "step": 31265 }, { "epoch": 16.38888888888889, "grad_norm": 0.09966301918029785, "learning_rate": 4.804457637512552e-06, "loss": 0.4363, "num_input_tokens_seen": 20451904, "step": 31270 }, { "epoch": 16.391509433962263, "grad_norm": 0.18310080468654633, "learning_rate": 4.79772007458231e-06, "loss": 0.5519, "num_input_tokens_seen": 20455520, "step": 31275 }, { "epoch": 16.39412997903564, "grad_norm": 0.14254474639892578, "learning_rate": 4.7909867377385424e-06, "loss": 0.381, "num_input_tokens_seen": 20459232, "step": 31280 }, { "epoch": 16.396750524109013, "grad_norm": 0.26289454102516174, "learning_rate": 4.784257628389794e-06, "loss": 0.6218, "num_input_tokens_seen": 20462400, "step": 31285 }, { "epoch": 16.39937106918239, "grad_norm": 0.121615469455719, "learning_rate": 4.7775327479437166e-06, "loss": 0.6516, "num_input_tokens_seen": 20465120, "step": 31290 }, { "epoch": 16.401991614255767, "grad_norm": 0.34512680768966675, "learning_rate": 4.770812097807076e-06, "loss": 0.5408, "num_input_tokens_seen": 20467616, "step": 31295 }, { "epoch": 16.40461215932914, "grad_norm": 0.16155332326889038, "learning_rate": 4.764095679385766e-06, "loss": 0.5353, "num_input_tokens_seen": 20472192, "step": 31300 }, { "epoch": 16.407232704402517, "grad_norm": 0.13232561945915222, "learning_rate": 4.7573834940847745e-06, "loss": 0.4101, "num_input_tokens_seen": 20475776, "step": 31305 }, { "epoch": 16.40985324947589, "grad_norm": 0.08666323870420456, "learning_rate": 4.750675543308212e-06, "loss": 0.5006, "num_input_tokens_seen": 20479456, "step": 31310 }, { "epoch": 16.412473794549268, "grad_norm": 0.16039876639842987, "learning_rate": 4.743971828459307e-06, "loss": 0.525, "num_input_tokens_seen": 20481952, "step": 31315 }, { "epoch": 16.41509433962264, "grad_norm": 0.11312945932149887, "learning_rate": 4.7372723509404025e-06, "loss": 0.4856, "num_input_tokens_seen": 20486016, "step": 31320 }, { "epoch": 16.417714884696018, "grad_norm": 0.17398256063461304, "learning_rate": 4.7305771121529615e-06, "loss": 0.3855, "num_input_tokens_seen": 20493312, "step": 31325 }, { "epoch": 16.42033542976939, "grad_norm": 0.12797138094902039, "learning_rate": 4.723886113497542e-06, "loss": 0.4788, "num_input_tokens_seen": 20496224, "step": 31330 }, { "epoch": 16.42295597484277, "grad_norm": 0.18034914135932922, "learning_rate": 4.7171993563738195e-06, "loss": 0.5556, "num_input_tokens_seen": 20499776, "step": 31335 }, { "epoch": 16.42557651991614, "grad_norm": 0.2172548621892929, "learning_rate": 4.7105168421805924e-06, "loss": 0.4235, "num_input_tokens_seen": 20502816, "step": 31340 }, { "epoch": 16.42819706498952, "grad_norm": 0.1224079355597496, "learning_rate": 4.703838572315767e-06, "loss": 0.4569, "num_input_tokens_seen": 20506240, "step": 31345 }, { "epoch": 16.430817610062892, "grad_norm": 0.17390327155590057, "learning_rate": 4.697164548176367e-06, "loss": 0.5696, "num_input_tokens_seen": 20509408, "step": 31350 }, { "epoch": 16.43343815513627, "grad_norm": 0.16263629496097565, "learning_rate": 4.690494771158519e-06, "loss": 0.4199, "num_input_tokens_seen": 20511680, "step": 31355 }, { "epoch": 16.436058700209642, "grad_norm": 0.1830083429813385, "learning_rate": 4.683829242657453e-06, "loss": 0.423, "num_input_tokens_seen": 20515072, "step": 31360 }, { "epoch": 16.43867924528302, "grad_norm": 0.1531004160642624, "learning_rate": 4.677167964067533e-06, "loss": 0.5146, "num_input_tokens_seen": 20517856, "step": 31365 }, { "epoch": 16.441299790356393, "grad_norm": 0.14452391862869263, "learning_rate": 4.670510936782216e-06, "loss": 0.4162, "num_input_tokens_seen": 20521344, "step": 31370 }, { "epoch": 16.44392033542977, "grad_norm": 0.11831651628017426, "learning_rate": 4.6638581621940815e-06, "loss": 0.8596, "num_input_tokens_seen": 20524896, "step": 31375 }, { "epoch": 16.446540880503143, "grad_norm": 0.11990001797676086, "learning_rate": 4.657209641694821e-06, "loss": 0.4545, "num_input_tokens_seen": 20528576, "step": 31380 }, { "epoch": 16.44916142557652, "grad_norm": 0.1813855916261673, "learning_rate": 4.65056537667522e-06, "loss": 0.4106, "num_input_tokens_seen": 20531712, "step": 31385 }, { "epoch": 16.451781970649897, "grad_norm": 0.13841424882411957, "learning_rate": 4.643925368525176e-06, "loss": 0.5628, "num_input_tokens_seen": 20534432, "step": 31390 }, { "epoch": 16.45440251572327, "grad_norm": 0.11729581654071808, "learning_rate": 4.637289618633711e-06, "loss": 0.5625, "num_input_tokens_seen": 20538752, "step": 31395 }, { "epoch": 16.457023060796647, "grad_norm": 0.16085343062877655, "learning_rate": 4.630658128388948e-06, "loss": 0.4606, "num_input_tokens_seen": 20541376, "step": 31400 }, { "epoch": 16.45964360587002, "grad_norm": 0.1111992746591568, "learning_rate": 4.624030899178125e-06, "loss": 0.593, "num_input_tokens_seen": 20544256, "step": 31405 }, { "epoch": 16.462264150943398, "grad_norm": 0.1979360580444336, "learning_rate": 4.617407932387569e-06, "loss": 0.4538, "num_input_tokens_seen": 20548576, "step": 31410 }, { "epoch": 16.46488469601677, "grad_norm": 0.17962153255939484, "learning_rate": 4.610789229402743e-06, "loss": 0.5142, "num_input_tokens_seen": 20551968, "step": 31415 }, { "epoch": 16.467505241090148, "grad_norm": 0.14282070100307465, "learning_rate": 4.604174791608193e-06, "loss": 0.445, "num_input_tokens_seen": 20555008, "step": 31420 }, { "epoch": 16.47012578616352, "grad_norm": 0.10735706984996796, "learning_rate": 4.597564620387587e-06, "loss": 0.3882, "num_input_tokens_seen": 20558144, "step": 31425 }, { "epoch": 16.4727463312369, "grad_norm": 0.15700691938400269, "learning_rate": 4.590958717123708e-06, "loss": 0.8842, "num_input_tokens_seen": 20560704, "step": 31430 }, { "epoch": 16.47536687631027, "grad_norm": 0.14834076166152954, "learning_rate": 4.5843570831984174e-06, "loss": 0.5617, "num_input_tokens_seen": 20564608, "step": 31435 }, { "epoch": 16.47798742138365, "grad_norm": 0.15902695059776306, "learning_rate": 4.577759719992719e-06, "loss": 0.5051, "num_input_tokens_seen": 20568096, "step": 31440 }, { "epoch": 16.480607966457022, "grad_norm": 0.13090580701828003, "learning_rate": 4.571166628886692e-06, "loss": 0.45, "num_input_tokens_seen": 20571744, "step": 31445 }, { "epoch": 16.4832285115304, "grad_norm": 0.17265918850898743, "learning_rate": 4.564577811259541e-06, "loss": 0.4511, "num_input_tokens_seen": 20574368, "step": 31450 }, { "epoch": 16.485849056603772, "grad_norm": 0.16086220741271973, "learning_rate": 4.557993268489574e-06, "loss": 0.4462, "num_input_tokens_seen": 20578080, "step": 31455 }, { "epoch": 16.48846960167715, "grad_norm": 0.1437596082687378, "learning_rate": 4.551413001954205e-06, "loss": 0.5088, "num_input_tokens_seen": 20581248, "step": 31460 }, { "epoch": 16.491090146750523, "grad_norm": 0.16586549580097198, "learning_rate": 4.54483701302994e-06, "loss": 0.45, "num_input_tokens_seen": 20585664, "step": 31465 }, { "epoch": 16.4937106918239, "grad_norm": 0.11822512000799179, "learning_rate": 4.538265303092415e-06, "loss": 0.4477, "num_input_tokens_seen": 20589024, "step": 31470 }, { "epoch": 16.496331236897273, "grad_norm": 0.1043430045247078, "learning_rate": 4.531697873516338e-06, "loss": 0.3746, "num_input_tokens_seen": 20592320, "step": 31475 }, { "epoch": 16.49895178197065, "grad_norm": 0.16504648327827454, "learning_rate": 4.525134725675556e-06, "loss": 0.4232, "num_input_tokens_seen": 20594848, "step": 31480 }, { "epoch": 16.501572327044027, "grad_norm": 0.13790060579776764, "learning_rate": 4.518575860943003e-06, "loss": 0.4524, "num_input_tokens_seen": 20597504, "step": 31485 }, { "epoch": 16.5041928721174, "grad_norm": 0.14802797138690948, "learning_rate": 4.512021280690712e-06, "loss": 0.544, "num_input_tokens_seen": 20601152, "step": 31490 }, { "epoch": 16.506813417190777, "grad_norm": 0.19817696511745453, "learning_rate": 4.505470986289828e-06, "loss": 0.4604, "num_input_tokens_seen": 20603936, "step": 31495 }, { "epoch": 16.50943396226415, "grad_norm": 0.15679530799388885, "learning_rate": 4.498924979110605e-06, "loss": 0.5578, "num_input_tokens_seen": 20607872, "step": 31500 }, { "epoch": 16.512054507337528, "grad_norm": 0.12046060711145401, "learning_rate": 4.492383260522384e-06, "loss": 0.6811, "num_input_tokens_seen": 20612288, "step": 31505 }, { "epoch": 16.5146750524109, "grad_norm": 0.11766422539949417, "learning_rate": 4.485845831893626e-06, "loss": 0.4143, "num_input_tokens_seen": 20615072, "step": 31510 }, { "epoch": 16.517295597484278, "grad_norm": 0.19476161897182465, "learning_rate": 4.479312694591878e-06, "loss": 0.4868, "num_input_tokens_seen": 20620736, "step": 31515 }, { "epoch": 16.51991614255765, "grad_norm": 0.12819500267505646, "learning_rate": 4.4727838499838e-06, "loss": 0.454, "num_input_tokens_seen": 20624640, "step": 31520 }, { "epoch": 16.52253668763103, "grad_norm": 0.17466092109680176, "learning_rate": 4.4662592994351545e-06, "loss": 0.5708, "num_input_tokens_seen": 20627968, "step": 31525 }, { "epoch": 16.5251572327044, "grad_norm": 0.18634472787380219, "learning_rate": 4.4597390443108065e-06, "loss": 0.5072, "num_input_tokens_seen": 20631360, "step": 31530 }, { "epoch": 16.52777777777778, "grad_norm": 0.19423051178455353, "learning_rate": 4.453223085974712e-06, "loss": 0.6132, "num_input_tokens_seen": 20633760, "step": 31535 }, { "epoch": 16.530398322851152, "grad_norm": 0.17009980976581573, "learning_rate": 4.4467114257899425e-06, "loss": 0.4962, "num_input_tokens_seen": 20636640, "step": 31540 }, { "epoch": 16.53301886792453, "grad_norm": 0.22789537906646729, "learning_rate": 4.44020406511865e-06, "loss": 0.4248, "num_input_tokens_seen": 20639648, "step": 31545 }, { "epoch": 16.535639412997902, "grad_norm": 0.23141197860240936, "learning_rate": 4.43370100532211e-06, "loss": 0.4941, "num_input_tokens_seen": 20642464, "step": 31550 }, { "epoch": 16.53825995807128, "grad_norm": 0.17656821012496948, "learning_rate": 4.427202247760695e-06, "loss": 0.5198, "num_input_tokens_seen": 20645760, "step": 31555 }, { "epoch": 16.540880503144653, "grad_norm": 0.20128215849399567, "learning_rate": 4.420707793793854e-06, "loss": 0.4143, "num_input_tokens_seen": 20648960, "step": 31560 }, { "epoch": 16.54350104821803, "grad_norm": 0.9268805384635925, "learning_rate": 4.414217644780172e-06, "loss": 0.5479, "num_input_tokens_seen": 20652000, "step": 31565 }, { "epoch": 16.546121593291403, "grad_norm": 0.30662673711776733, "learning_rate": 4.407731802077294e-06, "loss": 0.4898, "num_input_tokens_seen": 20654592, "step": 31570 }, { "epoch": 16.54874213836478, "grad_norm": 0.1617019921541214, "learning_rate": 4.401250267041996e-06, "loss": 0.5273, "num_input_tokens_seen": 20657472, "step": 31575 }, { "epoch": 16.551362683438157, "grad_norm": 0.1353091150522232, "learning_rate": 4.394773041030137e-06, "loss": 0.519, "num_input_tokens_seen": 20660192, "step": 31580 }, { "epoch": 16.55398322851153, "grad_norm": 0.15837781131267548, "learning_rate": 4.388300125396685e-06, "loss": 0.422, "num_input_tokens_seen": 20663488, "step": 31585 }, { "epoch": 16.556603773584907, "grad_norm": 0.13256584107875824, "learning_rate": 4.381831521495699e-06, "loss": 0.5574, "num_input_tokens_seen": 20666080, "step": 31590 }, { "epoch": 16.55922431865828, "grad_norm": 0.14373555779457092, "learning_rate": 4.37536723068033e-06, "loss": 0.3984, "num_input_tokens_seen": 20669056, "step": 31595 }, { "epoch": 16.561844863731658, "grad_norm": 0.2583335041999817, "learning_rate": 4.368907254302837e-06, "loss": 0.433, "num_input_tokens_seen": 20672320, "step": 31600 }, { "epoch": 16.56446540880503, "grad_norm": 0.2427964061498642, "learning_rate": 4.3624515937145725e-06, "loss": 0.6309, "num_input_tokens_seen": 20675424, "step": 31605 }, { "epoch": 16.567085953878408, "grad_norm": 0.18308955430984497, "learning_rate": 4.3560002502659905e-06, "loss": 0.4091, "num_input_tokens_seen": 20678624, "step": 31610 }, { "epoch": 16.56970649895178, "grad_norm": 0.15705376863479614, "learning_rate": 4.349553225306644e-06, "loss": 0.5391, "num_input_tokens_seen": 20682176, "step": 31615 }, { "epoch": 16.572327044025158, "grad_norm": 0.1021399050951004, "learning_rate": 4.343110520185167e-06, "loss": 0.4825, "num_input_tokens_seen": 20686688, "step": 31620 }, { "epoch": 16.57494758909853, "grad_norm": 0.10044217109680176, "learning_rate": 4.336672136249298e-06, "loss": 0.4303, "num_input_tokens_seen": 20689376, "step": 31625 }, { "epoch": 16.57756813417191, "grad_norm": 0.1351451277732849, "learning_rate": 4.33023807484588e-06, "loss": 0.4393, "num_input_tokens_seen": 20692960, "step": 31630 }, { "epoch": 16.580188679245282, "grad_norm": 0.16254450380802155, "learning_rate": 4.323808337320842e-06, "loss": 0.4917, "num_input_tokens_seen": 20695520, "step": 31635 }, { "epoch": 16.58280922431866, "grad_norm": 0.2675369679927826, "learning_rate": 4.31738292501922e-06, "loss": 0.4412, "num_input_tokens_seen": 20699136, "step": 31640 }, { "epoch": 16.585429769392032, "grad_norm": 0.1371058076620102, "learning_rate": 4.31096183928513e-06, "loss": 0.3775, "num_input_tokens_seen": 20703104, "step": 31645 }, { "epoch": 16.58805031446541, "grad_norm": 0.12849147617816925, "learning_rate": 4.304545081461783e-06, "loss": 0.4895, "num_input_tokens_seen": 20705984, "step": 31650 }, { "epoch": 16.590670859538783, "grad_norm": 0.096686452627182, "learning_rate": 4.298132652891501e-06, "loss": 0.4435, "num_input_tokens_seen": 20712416, "step": 31655 }, { "epoch": 16.59329140461216, "grad_norm": 0.2972187101840973, "learning_rate": 4.291724554915688e-06, "loss": 0.4779, "num_input_tokens_seen": 20715456, "step": 31660 }, { "epoch": 16.595911949685533, "grad_norm": 0.18735632300376892, "learning_rate": 4.285320788874853e-06, "loss": 0.4067, "num_input_tokens_seen": 20718144, "step": 31665 }, { "epoch": 16.59853249475891, "grad_norm": 0.14908501505851746, "learning_rate": 4.278921356108581e-06, "loss": 0.448, "num_input_tokens_seen": 20720640, "step": 31670 }, { "epoch": 16.601153039832283, "grad_norm": 0.14419443905353546, "learning_rate": 4.272526257955556e-06, "loss": 0.5269, "num_input_tokens_seen": 20723424, "step": 31675 }, { "epoch": 16.60377358490566, "grad_norm": 0.16711561381816864, "learning_rate": 4.266135495753567e-06, "loss": 0.504, "num_input_tokens_seen": 20726528, "step": 31680 }, { "epoch": 16.606394129979037, "grad_norm": 0.3718298673629761, "learning_rate": 4.259749070839491e-06, "loss": 0.4629, "num_input_tokens_seen": 20729216, "step": 31685 }, { "epoch": 16.60901467505241, "grad_norm": 0.15914691984653473, "learning_rate": 4.253366984549289e-06, "loss": 0.5682, "num_input_tokens_seen": 20733344, "step": 31690 }, { "epoch": 16.611635220125788, "grad_norm": 0.19785861670970917, "learning_rate": 4.24698923821803e-06, "loss": 0.5651, "num_input_tokens_seen": 20736224, "step": 31695 }, { "epoch": 16.61425576519916, "grad_norm": 0.16724397242069244, "learning_rate": 4.240615833179861e-06, "loss": 0.4107, "num_input_tokens_seen": 20739424, "step": 31700 }, { "epoch": 16.616876310272538, "grad_norm": 0.16914808750152588, "learning_rate": 4.234246770768016e-06, "loss": 0.3935, "num_input_tokens_seen": 20741824, "step": 31705 }, { "epoch": 16.61949685534591, "grad_norm": 0.1967436820268631, "learning_rate": 4.227882052314841e-06, "loss": 0.4265, "num_input_tokens_seen": 20744896, "step": 31710 }, { "epoch": 16.622117400419288, "grad_norm": 0.1970960795879364, "learning_rate": 4.2215216791517575e-06, "loss": 0.4854, "num_input_tokens_seen": 20747776, "step": 31715 }, { "epoch": 16.62473794549266, "grad_norm": 0.1410248726606369, "learning_rate": 4.215165652609293e-06, "loss": 0.4935, "num_input_tokens_seen": 20750496, "step": 31720 }, { "epoch": 16.62735849056604, "grad_norm": 0.22723929584026337, "learning_rate": 4.208813974017048e-06, "loss": 0.5056, "num_input_tokens_seen": 20753216, "step": 31725 }, { "epoch": 16.629979035639412, "grad_norm": 0.39910727739334106, "learning_rate": 4.202466644703712e-06, "loss": 0.5995, "num_input_tokens_seen": 20756608, "step": 31730 }, { "epoch": 16.63259958071279, "grad_norm": 0.24910396337509155, "learning_rate": 4.196123665997087e-06, "loss": 0.4297, "num_input_tokens_seen": 20759264, "step": 31735 }, { "epoch": 16.635220125786162, "grad_norm": 0.14716866612434387, "learning_rate": 4.189785039224045e-06, "loss": 0.394, "num_input_tokens_seen": 20762432, "step": 31740 }, { "epoch": 16.63784067085954, "grad_norm": 0.2187880426645279, "learning_rate": 4.183450765710567e-06, "loss": 0.4495, "num_input_tokens_seen": 20765760, "step": 31745 }, { "epoch": 16.640461215932913, "grad_norm": 0.14766916632652283, "learning_rate": 4.177120846781693e-06, "loss": 0.5723, "num_input_tokens_seen": 20769376, "step": 31750 }, { "epoch": 16.64308176100629, "grad_norm": 0.14575859904289246, "learning_rate": 4.170795283761583e-06, "loss": 0.5111, "num_input_tokens_seen": 20772608, "step": 31755 }, { "epoch": 16.645702306079663, "grad_norm": 0.25022274255752563, "learning_rate": 4.164474077973462e-06, "loss": 0.4758, "num_input_tokens_seen": 20776288, "step": 31760 }, { "epoch": 16.64832285115304, "grad_norm": 0.3994915187358856, "learning_rate": 4.158157230739659e-06, "loss": 0.5464, "num_input_tokens_seen": 20780064, "step": 31765 }, { "epoch": 16.650943396226417, "grad_norm": 0.17424094676971436, "learning_rate": 4.151844743381589e-06, "loss": 0.4535, "num_input_tokens_seen": 20783040, "step": 31770 }, { "epoch": 16.65356394129979, "grad_norm": 0.19006390869617462, "learning_rate": 4.145536617219756e-06, "loss": 0.4982, "num_input_tokens_seen": 20785952, "step": 31775 }, { "epoch": 16.656184486373167, "grad_norm": 0.20243188738822937, "learning_rate": 4.139232853573735e-06, "loss": 0.4924, "num_input_tokens_seen": 20788640, "step": 31780 }, { "epoch": 16.65880503144654, "grad_norm": 0.15734495222568512, "learning_rate": 4.132933453762214e-06, "loss": 0.4649, "num_input_tokens_seen": 20792832, "step": 31785 }, { "epoch": 16.661425576519918, "grad_norm": 0.11835560947656631, "learning_rate": 4.126638419102946e-06, "loss": 0.4175, "num_input_tokens_seen": 20796512, "step": 31790 }, { "epoch": 16.66404612159329, "grad_norm": 0.1026022732257843, "learning_rate": 4.120347750912784e-06, "loss": 0.4089, "num_input_tokens_seen": 20800160, "step": 31795 }, { "epoch": 16.666666666666668, "grad_norm": 0.23337768018245697, "learning_rate": 4.1140614505076724e-06, "loss": 0.4147, "num_input_tokens_seen": 20802624, "step": 31800 }, { "epoch": 16.66928721174004, "grad_norm": 0.18782755732536316, "learning_rate": 4.10777951920262e-06, "loss": 0.3072, "num_input_tokens_seen": 20805984, "step": 31805 }, { "epoch": 16.671907756813418, "grad_norm": 0.12657707929611206, "learning_rate": 4.1015019583117455e-06, "loss": 0.4646, "num_input_tokens_seen": 20809504, "step": 31810 }, { "epoch": 16.67452830188679, "grad_norm": 0.25373575091362, "learning_rate": 4.095228769148232e-06, "loss": 0.4468, "num_input_tokens_seen": 20812736, "step": 31815 }, { "epoch": 16.67714884696017, "grad_norm": 0.1687624603509903, "learning_rate": 4.088959953024368e-06, "loss": 0.475, "num_input_tokens_seen": 20815680, "step": 31820 }, { "epoch": 16.679769392033542, "grad_norm": 0.15978007018566132, "learning_rate": 4.082695511251522e-06, "loss": 0.4067, "num_input_tokens_seen": 20818880, "step": 31825 }, { "epoch": 16.68238993710692, "grad_norm": 0.18700554966926575, "learning_rate": 4.076435445140131e-06, "loss": 0.5754, "num_input_tokens_seen": 20821632, "step": 31830 }, { "epoch": 16.685010482180292, "grad_norm": 0.11234436184167862, "learning_rate": 4.070179755999737e-06, "loss": 0.4639, "num_input_tokens_seen": 20824512, "step": 31835 }, { "epoch": 16.68763102725367, "grad_norm": 0.10337506234645844, "learning_rate": 4.063928445138962e-06, "loss": 0.618, "num_input_tokens_seen": 20828192, "step": 31840 }, { "epoch": 16.690251572327043, "grad_norm": 0.10726263374090195, "learning_rate": 4.057681513865499e-06, "loss": 0.5225, "num_input_tokens_seen": 20831072, "step": 31845 }, { "epoch": 16.69287211740042, "grad_norm": 0.1321054846048355, "learning_rate": 4.0514389634861426e-06, "loss": 0.4764, "num_input_tokens_seen": 20834880, "step": 31850 }, { "epoch": 16.695492662473793, "grad_norm": 0.1634228378534317, "learning_rate": 4.045200795306767e-06, "loss": 0.3719, "num_input_tokens_seen": 20840928, "step": 31855 }, { "epoch": 16.69811320754717, "grad_norm": 0.23846445977687836, "learning_rate": 4.038967010632314e-06, "loss": 0.464, "num_input_tokens_seen": 20843808, "step": 31860 }, { "epoch": 16.700733752620543, "grad_norm": 0.202473983168602, "learning_rate": 4.032737610766826e-06, "loss": 0.4627, "num_input_tokens_seen": 20847264, "step": 31865 }, { "epoch": 16.70335429769392, "grad_norm": 0.21513965725898743, "learning_rate": 4.026512597013429e-06, "loss": 0.4912, "num_input_tokens_seen": 20851264, "step": 31870 }, { "epoch": 16.705974842767297, "grad_norm": 0.1691327840089798, "learning_rate": 4.020291970674312e-06, "loss": 0.4507, "num_input_tokens_seen": 20854048, "step": 31875 }, { "epoch": 16.70859538784067, "grad_norm": 0.2802753150463104, "learning_rate": 4.0140757330507705e-06, "loss": 0.4113, "num_input_tokens_seen": 20857472, "step": 31880 }, { "epoch": 16.711215932914047, "grad_norm": 0.1099739596247673, "learning_rate": 4.007863885443161e-06, "loss": 0.4385, "num_input_tokens_seen": 20860352, "step": 31885 }, { "epoch": 16.71383647798742, "grad_norm": 0.11868003010749817, "learning_rate": 4.001656429150935e-06, "loss": 0.3684, "num_input_tokens_seen": 20863808, "step": 31890 }, { "epoch": 16.716457023060798, "grad_norm": 0.25282618403434753, "learning_rate": 3.9954533654726235e-06, "loss": 0.5712, "num_input_tokens_seen": 20867488, "step": 31895 }, { "epoch": 16.71907756813417, "grad_norm": 0.14886470139026642, "learning_rate": 3.989254695705841e-06, "loss": 0.5019, "num_input_tokens_seen": 20870336, "step": 31900 }, { "epoch": 16.721698113207548, "grad_norm": 0.10836051404476166, "learning_rate": 3.983060421147272e-06, "loss": 0.4525, "num_input_tokens_seen": 20874624, "step": 31905 }, { "epoch": 16.72431865828092, "grad_norm": 0.143297016620636, "learning_rate": 3.976870543092684e-06, "loss": 0.6332, "num_input_tokens_seen": 20879456, "step": 31910 }, { "epoch": 16.7269392033543, "grad_norm": 0.12724487483501434, "learning_rate": 3.970685062836932e-06, "loss": 0.6169, "num_input_tokens_seen": 20883040, "step": 31915 }, { "epoch": 16.729559748427672, "grad_norm": 0.1669701337814331, "learning_rate": 3.96450398167395e-06, "loss": 0.4111, "num_input_tokens_seen": 20886304, "step": 31920 }, { "epoch": 16.73218029350105, "grad_norm": 0.11444266885519028, "learning_rate": 3.958327300896758e-06, "loss": 0.3854, "num_input_tokens_seen": 20889888, "step": 31925 }, { "epoch": 16.734800838574422, "grad_norm": 0.21436548233032227, "learning_rate": 3.952155021797429e-06, "loss": 0.467, "num_input_tokens_seen": 20892896, "step": 31930 }, { "epoch": 16.7374213836478, "grad_norm": 0.27750512957572937, "learning_rate": 3.945987145667151e-06, "loss": 0.4863, "num_input_tokens_seen": 20896672, "step": 31935 }, { "epoch": 16.740041928721173, "grad_norm": 0.14945659041404724, "learning_rate": 3.939823673796162e-06, "loss": 0.4709, "num_input_tokens_seen": 20900128, "step": 31940 }, { "epoch": 16.74266247379455, "grad_norm": 0.15680502355098724, "learning_rate": 3.933664607473789e-06, "loss": 0.5179, "num_input_tokens_seen": 20903872, "step": 31945 }, { "epoch": 16.745283018867923, "grad_norm": 0.18060176074504852, "learning_rate": 3.927509947988445e-06, "loss": 0.5662, "num_input_tokens_seen": 20907008, "step": 31950 }, { "epoch": 16.7479035639413, "grad_norm": 0.22082586586475372, "learning_rate": 3.9213596966276164e-06, "loss": 0.5574, "num_input_tokens_seen": 20910048, "step": 31955 }, { "epoch": 16.750524109014677, "grad_norm": 0.13508179783821106, "learning_rate": 3.9152138546778625e-06, "loss": 0.5436, "num_input_tokens_seen": 20914784, "step": 31960 }, { "epoch": 16.75314465408805, "grad_norm": 0.11658589541912079, "learning_rate": 3.9090724234248175e-06, "loss": 0.4225, "num_input_tokens_seen": 20918880, "step": 31965 }, { "epoch": 16.755765199161427, "grad_norm": 0.1547943353652954, "learning_rate": 3.902935404153202e-06, "loss": 0.4437, "num_input_tokens_seen": 20921696, "step": 31970 }, { "epoch": 16.7583857442348, "grad_norm": 0.20642296969890594, "learning_rate": 3.896802798146809e-06, "loss": 0.4739, "num_input_tokens_seen": 20925440, "step": 31975 }, { "epoch": 16.761006289308177, "grad_norm": 0.15479736030101776, "learning_rate": 3.890674606688524e-06, "loss": 0.4296, "num_input_tokens_seen": 20928032, "step": 31980 }, { "epoch": 16.76362683438155, "grad_norm": 0.203607439994812, "learning_rate": 3.88455083106028e-06, "loss": 0.5521, "num_input_tokens_seen": 20930688, "step": 31985 }, { "epoch": 16.766247379454928, "grad_norm": 0.16762875020503998, "learning_rate": 3.878431472543098e-06, "loss": 0.402, "num_input_tokens_seen": 20933888, "step": 31990 }, { "epoch": 16.7688679245283, "grad_norm": 0.19109834730625153, "learning_rate": 3.872316532417086e-06, "loss": 0.3571, "num_input_tokens_seen": 20936896, "step": 31995 }, { "epoch": 16.771488469601678, "grad_norm": 0.12929199635982513, "learning_rate": 3.866206011961418e-06, "loss": 0.4983, "num_input_tokens_seen": 20940192, "step": 32000 }, { "epoch": 16.77410901467505, "grad_norm": 0.15497948229312897, "learning_rate": 3.860099912454346e-06, "loss": 0.4598, "num_input_tokens_seen": 20943168, "step": 32005 }, { "epoch": 16.77672955974843, "grad_norm": 0.15397724509239197, "learning_rate": 3.853998235173201e-06, "loss": 0.5333, "num_input_tokens_seen": 20946272, "step": 32010 }, { "epoch": 16.779350104821802, "grad_norm": 0.2272733747959137, "learning_rate": 3.847900981394379e-06, "loss": 0.5153, "num_input_tokens_seen": 20949024, "step": 32015 }, { "epoch": 16.78197064989518, "grad_norm": 0.18897034227848053, "learning_rate": 3.841808152393348e-06, "loss": 0.488, "num_input_tokens_seen": 20952160, "step": 32020 }, { "epoch": 16.784591194968552, "grad_norm": 0.24312737584114075, "learning_rate": 3.835719749444666e-06, "loss": 0.5406, "num_input_tokens_seen": 20956192, "step": 32025 }, { "epoch": 16.78721174004193, "grad_norm": 0.21379345655441284, "learning_rate": 3.8296357738219606e-06, "loss": 0.4298, "num_input_tokens_seen": 20958816, "step": 32030 }, { "epoch": 16.789832285115303, "grad_norm": 0.18314826488494873, "learning_rate": 3.823556226797931e-06, "loss": 0.4594, "num_input_tokens_seen": 20961600, "step": 32035 }, { "epoch": 16.79245283018868, "grad_norm": 0.1376228630542755, "learning_rate": 3.817481109644344e-06, "loss": 0.6119, "num_input_tokens_seen": 20964832, "step": 32040 }, { "epoch": 16.795073375262053, "grad_norm": 0.202622652053833, "learning_rate": 3.811410423632042e-06, "loss": 0.5199, "num_input_tokens_seen": 20967232, "step": 32045 }, { "epoch": 16.79769392033543, "grad_norm": 0.18136674165725708, "learning_rate": 3.8053441700309473e-06, "loss": 0.512, "num_input_tokens_seen": 20969952, "step": 32050 }, { "epoch": 16.800314465408803, "grad_norm": 0.14286549389362335, "learning_rate": 3.7992823501100507e-06, "loss": 0.4653, "num_input_tokens_seen": 20973856, "step": 32055 }, { "epoch": 16.80293501048218, "grad_norm": 0.09340503811836243, "learning_rate": 3.7932249651374207e-06, "loss": 0.5737, "num_input_tokens_seen": 20977472, "step": 32060 }, { "epoch": 16.805555555555557, "grad_norm": 0.11689142882823944, "learning_rate": 3.787172016380183e-06, "loss": 0.4454, "num_input_tokens_seen": 20981248, "step": 32065 }, { "epoch": 16.80817610062893, "grad_norm": 0.19255787134170532, "learning_rate": 3.7811235051045558e-06, "loss": 0.4772, "num_input_tokens_seen": 20985024, "step": 32070 }, { "epoch": 16.810796645702307, "grad_norm": 0.11862067878246307, "learning_rate": 3.775079432575809e-06, "loss": 0.4097, "num_input_tokens_seen": 20988128, "step": 32075 }, { "epoch": 16.81341719077568, "grad_norm": 0.19692544639110565, "learning_rate": 3.7690398000582964e-06, "loss": 0.5151, "num_input_tokens_seen": 20991552, "step": 32080 }, { "epoch": 16.816037735849058, "grad_norm": 0.27058595418930054, "learning_rate": 3.763004608815443e-06, "loss": 0.5985, "num_input_tokens_seen": 20995296, "step": 32085 }, { "epoch": 16.81865828092243, "grad_norm": 0.11552972346544266, "learning_rate": 3.756973860109747e-06, "loss": 0.3633, "num_input_tokens_seen": 20999136, "step": 32090 }, { "epoch": 16.821278825995808, "grad_norm": 0.21302530169487, "learning_rate": 3.7509475552027663e-06, "loss": 0.5063, "num_input_tokens_seen": 21002592, "step": 32095 }, { "epoch": 16.82389937106918, "grad_norm": 0.1466178596019745, "learning_rate": 3.744925695355131e-06, "loss": 0.4681, "num_input_tokens_seen": 21005440, "step": 32100 }, { "epoch": 16.82651991614256, "grad_norm": 0.15434204041957855, "learning_rate": 3.7389082818265476e-06, "loss": 0.5205, "num_input_tokens_seen": 21008832, "step": 32105 }, { "epoch": 16.829140461215932, "grad_norm": 0.2139495611190796, "learning_rate": 3.732895315875792e-06, "loss": 0.3797, "num_input_tokens_seen": 21012128, "step": 32110 }, { "epoch": 16.83176100628931, "grad_norm": 0.24048545956611633, "learning_rate": 3.726886798760715e-06, "loss": 0.4465, "num_input_tokens_seen": 21014944, "step": 32115 }, { "epoch": 16.834381551362682, "grad_norm": 0.13483139872550964, "learning_rate": 3.720882731738215e-06, "loss": 0.4376, "num_input_tokens_seen": 21018784, "step": 32120 }, { "epoch": 16.83700209643606, "grad_norm": 0.14379650354385376, "learning_rate": 3.714883116064291e-06, "loss": 0.4059, "num_input_tokens_seen": 21021728, "step": 32125 }, { "epoch": 16.839622641509433, "grad_norm": 0.1730331927537918, "learning_rate": 3.708887952993978e-06, "loss": 0.3843, "num_input_tokens_seen": 21024768, "step": 32130 }, { "epoch": 16.84224318658281, "grad_norm": 0.18401017785072327, "learning_rate": 3.7028972437814057e-06, "loss": 0.5176, "num_input_tokens_seen": 21031456, "step": 32135 }, { "epoch": 16.844863731656183, "grad_norm": 0.1182977482676506, "learning_rate": 3.696910989679764e-06, "loss": 0.4526, "num_input_tokens_seen": 21034624, "step": 32140 }, { "epoch": 16.84748427672956, "grad_norm": 0.11641670018434525, "learning_rate": 3.6909291919413003e-06, "loss": 0.4581, "num_input_tokens_seen": 21038560, "step": 32145 }, { "epoch": 16.850104821802937, "grad_norm": 0.12872493267059326, "learning_rate": 3.684951851817342e-06, "loss": 0.4634, "num_input_tokens_seen": 21041920, "step": 32150 }, { "epoch": 16.85272536687631, "grad_norm": 0.14624513685703278, "learning_rate": 3.678978970558286e-06, "loss": 0.3492, "num_input_tokens_seen": 21045344, "step": 32155 }, { "epoch": 16.855345911949687, "grad_norm": 0.1808498650789261, "learning_rate": 3.673010549413583e-06, "loss": 0.423, "num_input_tokens_seen": 21047936, "step": 32160 }, { "epoch": 16.85796645702306, "grad_norm": 0.16086545586585999, "learning_rate": 3.6670465896317592e-06, "loss": 0.5564, "num_input_tokens_seen": 21050752, "step": 32165 }, { "epoch": 16.860587002096437, "grad_norm": 0.18537652492523193, "learning_rate": 3.661087092460419e-06, "loss": 0.4082, "num_input_tokens_seen": 21054112, "step": 32170 }, { "epoch": 16.86320754716981, "grad_norm": 0.18266882002353668, "learning_rate": 3.655132059146202e-06, "loss": 0.6043, "num_input_tokens_seen": 21056800, "step": 32175 }, { "epoch": 16.865828092243188, "grad_norm": 0.19127333164215088, "learning_rate": 3.6491814909348526e-06, "loss": 0.4127, "num_input_tokens_seen": 21059712, "step": 32180 }, { "epoch": 16.86844863731656, "grad_norm": 0.23457300662994385, "learning_rate": 3.6432353890711424e-06, "loss": 0.5305, "num_input_tokens_seen": 21062304, "step": 32185 }, { "epoch": 16.871069182389938, "grad_norm": 0.18318085372447968, "learning_rate": 3.6372937547989425e-06, "loss": 0.4529, "num_input_tokens_seen": 21065408, "step": 32190 }, { "epoch": 16.87368972746331, "grad_norm": 0.19697602093219757, "learning_rate": 3.631356589361173e-06, "loss": 0.6202, "num_input_tokens_seen": 21069216, "step": 32195 }, { "epoch": 16.87631027253669, "grad_norm": 0.23066212236881256, "learning_rate": 3.625423893999813e-06, "loss": 0.4661, "num_input_tokens_seen": 21071872, "step": 32200 }, { "epoch": 16.878930817610062, "grad_norm": 0.13062722980976105, "learning_rate": 3.6194956699559217e-06, "loss": 0.4283, "num_input_tokens_seen": 21075552, "step": 32205 }, { "epoch": 16.88155136268344, "grad_norm": 0.1352423131465912, "learning_rate": 3.6135719184696188e-06, "loss": 0.3766, "num_input_tokens_seen": 21078656, "step": 32210 }, { "epoch": 16.884171907756812, "grad_norm": 0.19579461216926575, "learning_rate": 3.6076526407800766e-06, "loss": 0.4354, "num_input_tokens_seen": 21081216, "step": 32215 }, { "epoch": 16.88679245283019, "grad_norm": 0.15753531455993652, "learning_rate": 3.601737838125552e-06, "loss": 0.4495, "num_input_tokens_seen": 21084096, "step": 32220 }, { "epoch": 16.889412997903563, "grad_norm": 0.15619143843650818, "learning_rate": 3.595827511743341e-06, "loss": 0.389, "num_input_tokens_seen": 21087232, "step": 32225 }, { "epoch": 16.89203354297694, "grad_norm": 0.1438504159450531, "learning_rate": 3.5899216628698217e-06, "loss": 0.494, "num_input_tokens_seen": 21090272, "step": 32230 }, { "epoch": 16.894654088050313, "grad_norm": 0.13265283405780792, "learning_rate": 3.584020292740431e-06, "loss": 0.4706, "num_input_tokens_seen": 21093152, "step": 32235 }, { "epoch": 16.89727463312369, "grad_norm": 0.19698280096054077, "learning_rate": 3.578123402589675e-06, "loss": 0.5746, "num_input_tokens_seen": 21096224, "step": 32240 }, { "epoch": 16.899895178197063, "grad_norm": 0.1778794825077057, "learning_rate": 3.5722309936511044e-06, "loss": 0.5305, "num_input_tokens_seen": 21099072, "step": 32245 }, { "epoch": 16.90251572327044, "grad_norm": 0.13455286622047424, "learning_rate": 3.5663430671573572e-06, "loss": 0.4764, "num_input_tokens_seen": 21101888, "step": 32250 }, { "epoch": 16.905136268343817, "grad_norm": 0.11876221001148224, "learning_rate": 3.5604596243401038e-06, "loss": 0.4915, "num_input_tokens_seen": 21105152, "step": 32255 }, { "epoch": 16.90775681341719, "grad_norm": 0.1527056097984314, "learning_rate": 3.554580666430102e-06, "loss": 0.5614, "num_input_tokens_seen": 21108576, "step": 32260 }, { "epoch": 16.910377358490567, "grad_norm": 0.08870512247085571, "learning_rate": 3.5487061946571614e-06, "loss": 0.5297, "num_input_tokens_seen": 21112192, "step": 32265 }, { "epoch": 16.91299790356394, "grad_norm": 0.19619429111480713, "learning_rate": 3.542836210250161e-06, "loss": 0.4203, "num_input_tokens_seen": 21115104, "step": 32270 }, { "epoch": 16.915618448637318, "grad_norm": 0.30184680223464966, "learning_rate": 3.536970714437032e-06, "loss": 0.5001, "num_input_tokens_seen": 21118624, "step": 32275 }, { "epoch": 16.91823899371069, "grad_norm": 0.14342549443244934, "learning_rate": 3.5311097084447575e-06, "loss": 0.3561, "num_input_tokens_seen": 21122784, "step": 32280 }, { "epoch": 16.920859538784068, "grad_norm": 0.274167001247406, "learning_rate": 3.5252531934994033e-06, "loss": 0.2976, "num_input_tokens_seen": 21125920, "step": 32285 }, { "epoch": 16.92348008385744, "grad_norm": 0.13889962434768677, "learning_rate": 3.5194011708260866e-06, "loss": 0.5082, "num_input_tokens_seen": 21131488, "step": 32290 }, { "epoch": 16.92610062893082, "grad_norm": 0.16810034215450287, "learning_rate": 3.513553641648984e-06, "loss": 0.5582, "num_input_tokens_seen": 21134688, "step": 32295 }, { "epoch": 16.928721174004192, "grad_norm": 0.16956517100334167, "learning_rate": 3.5077106071913307e-06, "loss": 0.4589, "num_input_tokens_seen": 21137248, "step": 32300 }, { "epoch": 16.93134171907757, "grad_norm": 0.13563981652259827, "learning_rate": 3.5018720686754133e-06, "loss": 0.571, "num_input_tokens_seen": 21140800, "step": 32305 }, { "epoch": 16.933962264150942, "grad_norm": 0.21316753327846527, "learning_rate": 3.4960380273225985e-06, "loss": 0.5606, "num_input_tokens_seen": 21144032, "step": 32310 }, { "epoch": 16.93658280922432, "grad_norm": 0.1386411041021347, "learning_rate": 3.490208484353297e-06, "loss": 0.4162, "num_input_tokens_seen": 21146528, "step": 32315 }, { "epoch": 16.939203354297693, "grad_norm": 0.4381694197654724, "learning_rate": 3.4843834409869824e-06, "loss": 0.4983, "num_input_tokens_seen": 21149728, "step": 32320 }, { "epoch": 16.94182389937107, "grad_norm": 0.20926643908023834, "learning_rate": 3.4785628984421977e-06, "loss": 0.5333, "num_input_tokens_seen": 21153568, "step": 32325 }, { "epoch": 16.944444444444443, "grad_norm": 0.16736070811748505, "learning_rate": 3.4727468579365235e-06, "loss": 0.5089, "num_input_tokens_seen": 21158144, "step": 32330 }, { "epoch": 16.94706498951782, "grad_norm": 0.1673283576965332, "learning_rate": 3.466935320686604e-06, "loss": 0.5301, "num_input_tokens_seen": 21161344, "step": 32335 }, { "epoch": 16.949685534591197, "grad_norm": 0.10482493788003922, "learning_rate": 3.4611282879081553e-06, "loss": 0.425, "num_input_tokens_seen": 21166368, "step": 32340 }, { "epoch": 16.95230607966457, "grad_norm": 0.11176900565624237, "learning_rate": 3.455325760815939e-06, "loss": 0.6874, "num_input_tokens_seen": 21169440, "step": 32345 }, { "epoch": 16.954926624737947, "grad_norm": 0.15593187510967255, "learning_rate": 3.4495277406237843e-06, "loss": 0.4857, "num_input_tokens_seen": 21172544, "step": 32350 }, { "epoch": 16.95754716981132, "grad_norm": 0.11098587512969971, "learning_rate": 3.443734228544565e-06, "loss": 0.3817, "num_input_tokens_seen": 21176192, "step": 32355 }, { "epoch": 16.960167714884697, "grad_norm": 0.15792666375637054, "learning_rate": 3.437945225790215e-06, "loss": 0.5115, "num_input_tokens_seen": 21179424, "step": 32360 }, { "epoch": 16.96278825995807, "grad_norm": 0.14737887680530548, "learning_rate": 3.432160733571729e-06, "loss": 0.3729, "num_input_tokens_seen": 21182272, "step": 32365 }, { "epoch": 16.965408805031448, "grad_norm": 0.210773766040802, "learning_rate": 3.4263807530991595e-06, "loss": 0.5459, "num_input_tokens_seen": 21186048, "step": 32370 }, { "epoch": 16.96802935010482, "grad_norm": 0.13552281260490417, "learning_rate": 3.4206052855816155e-06, "loss": 0.4441, "num_input_tokens_seen": 21189024, "step": 32375 }, { "epoch": 16.970649895178198, "grad_norm": 0.15451224148273468, "learning_rate": 3.4148343322272567e-06, "loss": 0.4957, "num_input_tokens_seen": 21192576, "step": 32380 }, { "epoch": 16.97327044025157, "grad_norm": 0.14184223115444183, "learning_rate": 3.4090678942432923e-06, "loss": 0.5, "num_input_tokens_seen": 21195616, "step": 32385 }, { "epoch": 16.97589098532495, "grad_norm": 0.12010640650987625, "learning_rate": 3.4033059728360018e-06, "loss": 0.4372, "num_input_tokens_seen": 21199584, "step": 32390 }, { "epoch": 16.978511530398322, "grad_norm": 0.08896529674530029, "learning_rate": 3.397548569210715e-06, "loss": 0.5007, "num_input_tokens_seen": 21203008, "step": 32395 }, { "epoch": 16.9811320754717, "grad_norm": 0.2244568020105362, "learning_rate": 3.3917956845718136e-06, "loss": 0.5047, "num_input_tokens_seen": 21206080, "step": 32400 }, { "epoch": 16.983752620545072, "grad_norm": 0.2905758321285248, "learning_rate": 3.38604732012274e-06, "loss": 0.5898, "num_input_tokens_seen": 21209088, "step": 32405 }, { "epoch": 16.98637316561845, "grad_norm": 0.19351495802402496, "learning_rate": 3.3803034770659825e-06, "loss": 0.461, "num_input_tokens_seen": 21212288, "step": 32410 }, { "epoch": 16.988993710691823, "grad_norm": 0.1598917543888092, "learning_rate": 3.3745641566030826e-06, "loss": 0.5137, "num_input_tokens_seen": 21214560, "step": 32415 }, { "epoch": 16.9916142557652, "grad_norm": 0.3404638171195984, "learning_rate": 3.3688293599346442e-06, "loss": 0.8135, "num_input_tokens_seen": 21217056, "step": 32420 }, { "epoch": 16.994234800838573, "grad_norm": 0.1232319176197052, "learning_rate": 3.3630990882603213e-06, "loss": 0.4927, "num_input_tokens_seen": 21220224, "step": 32425 }, { "epoch": 16.99685534591195, "grad_norm": 0.15715467929840088, "learning_rate": 3.35737334277883e-06, "loss": 0.5563, "num_input_tokens_seen": 21223360, "step": 32430 }, { "epoch": 16.999475890985323, "grad_norm": 0.15976758301258087, "learning_rate": 3.351652124687918e-06, "loss": 0.4392, "num_input_tokens_seen": 21226496, "step": 32435 }, { "epoch": 17.0, "eval_loss": 0.46053868532180786, "eval_runtime": 10.4343, "eval_samples_per_second": 81.271, "eval_steps_per_second": 20.318, "num_input_tokens_seen": 21226656, "step": 32436 }, { "epoch": 17.0020964360587, "grad_norm": 0.15005454421043396, "learning_rate": 3.345935435184408e-06, "loss": 0.4677, "num_input_tokens_seen": 21228800, "step": 32440 }, { "epoch": 17.004716981132077, "grad_norm": 0.13471859693527222, "learning_rate": 3.3402232754641576e-06, "loss": 0.6853, "num_input_tokens_seen": 21232416, "step": 32445 }, { "epoch": 17.00733752620545, "grad_norm": 0.14313043653964996, "learning_rate": 3.334515646722089e-06, "loss": 0.4601, "num_input_tokens_seen": 21235168, "step": 32450 }, { "epoch": 17.009958071278827, "grad_norm": 0.11903907358646393, "learning_rate": 3.32881255015218e-06, "loss": 0.4133, "num_input_tokens_seen": 21237952, "step": 32455 }, { "epoch": 17.0125786163522, "grad_norm": 0.2005765289068222, "learning_rate": 3.3231139869474433e-06, "loss": 0.4547, "num_input_tokens_seen": 21241056, "step": 32460 }, { "epoch": 17.015199161425578, "grad_norm": 0.1889645904302597, "learning_rate": 3.317419958299961e-06, "loss": 0.4151, "num_input_tokens_seen": 21244960, "step": 32465 }, { "epoch": 17.01781970649895, "grad_norm": 0.15060190856456757, "learning_rate": 3.31173046540085e-06, "loss": 0.5818, "num_input_tokens_seen": 21247936, "step": 32470 }, { "epoch": 17.020440251572328, "grad_norm": 0.1907084435224533, "learning_rate": 3.306045509440295e-06, "loss": 0.4772, "num_input_tokens_seen": 21251104, "step": 32475 }, { "epoch": 17.0230607966457, "grad_norm": 0.1596502810716629, "learning_rate": 3.300365091607521e-06, "loss": 0.5258, "num_input_tokens_seen": 21255040, "step": 32480 }, { "epoch": 17.02568134171908, "grad_norm": 0.18705198168754578, "learning_rate": 3.2946892130908127e-06, "loss": 0.5363, "num_input_tokens_seen": 21258336, "step": 32485 }, { "epoch": 17.028301886792452, "grad_norm": 0.3941434621810913, "learning_rate": 3.289017875077488e-06, "loss": 0.5927, "num_input_tokens_seen": 21261760, "step": 32490 }, { "epoch": 17.03092243186583, "grad_norm": 0.1358773559331894, "learning_rate": 3.283351078753938e-06, "loss": 0.447, "num_input_tokens_seen": 21265056, "step": 32495 }, { "epoch": 17.033542976939202, "grad_norm": 0.2874052822589874, "learning_rate": 3.2776888253055777e-06, "loss": 0.5979, "num_input_tokens_seen": 21271744, "step": 32500 }, { "epoch": 17.03616352201258, "grad_norm": 0.1527155488729477, "learning_rate": 3.272031115916896e-06, "loss": 0.5157, "num_input_tokens_seen": 21274368, "step": 32505 }, { "epoch": 17.038784067085953, "grad_norm": 0.3084935247898102, "learning_rate": 3.2663779517714238e-06, "loss": 0.4424, "num_input_tokens_seen": 21277120, "step": 32510 }, { "epoch": 17.04140461215933, "grad_norm": 0.15257801115512848, "learning_rate": 3.260729334051729e-06, "loss": 0.4627, "num_input_tokens_seen": 21281632, "step": 32515 }, { "epoch": 17.044025157232703, "grad_norm": 0.0919812023639679, "learning_rate": 3.2550852639394446e-06, "loss": 0.5011, "num_input_tokens_seen": 21284160, "step": 32520 }, { "epoch": 17.04664570230608, "grad_norm": 0.18771332502365112, "learning_rate": 3.2494457426152452e-06, "loss": 0.6054, "num_input_tokens_seen": 21286752, "step": 32525 }, { "epoch": 17.049266247379453, "grad_norm": 0.15081100165843964, "learning_rate": 3.243810771258851e-06, "loss": 0.6019, "num_input_tokens_seen": 21290016, "step": 32530 }, { "epoch": 17.05188679245283, "grad_norm": 0.1586064100265503, "learning_rate": 3.2381803510490387e-06, "loss": 0.4639, "num_input_tokens_seen": 21292832, "step": 32535 }, { "epoch": 17.054507337526207, "grad_norm": 0.1410011202096939, "learning_rate": 3.2325544831636217e-06, "loss": 0.3897, "num_input_tokens_seen": 21296032, "step": 32540 }, { "epoch": 17.05712788259958, "grad_norm": 0.27423739433288574, "learning_rate": 3.22693316877947e-06, "loss": 0.5165, "num_input_tokens_seen": 21298624, "step": 32545 }, { "epoch": 17.059748427672957, "grad_norm": 0.20265479385852814, "learning_rate": 3.2213164090725057e-06, "loss": 0.3854, "num_input_tokens_seen": 21301472, "step": 32550 }, { "epoch": 17.06236897274633, "grad_norm": 0.2223329097032547, "learning_rate": 3.2157042052176763e-06, "loss": 0.6227, "num_input_tokens_seen": 21305600, "step": 32555 }, { "epoch": 17.064989517819708, "grad_norm": 0.1673414707183838, "learning_rate": 3.2100965583890003e-06, "loss": 0.4779, "num_input_tokens_seen": 21308064, "step": 32560 }, { "epoch": 17.06761006289308, "grad_norm": 0.12670476734638214, "learning_rate": 3.204493469759537e-06, "loss": 0.4711, "num_input_tokens_seen": 21311200, "step": 32565 }, { "epoch": 17.070230607966458, "grad_norm": 0.23075856268405914, "learning_rate": 3.1988949405013764e-06, "loss": 0.521, "num_input_tokens_seen": 21314528, "step": 32570 }, { "epoch": 17.07285115303983, "grad_norm": 0.213589146733284, "learning_rate": 3.193300971785676e-06, "loss": 0.4928, "num_input_tokens_seen": 21317952, "step": 32575 }, { "epoch": 17.07547169811321, "grad_norm": 0.4042034447193146, "learning_rate": 3.1877115647826335e-06, "loss": 0.4343, "num_input_tokens_seen": 21321440, "step": 32580 }, { "epoch": 17.078092243186582, "grad_norm": 0.1361500322818756, "learning_rate": 3.1821267206614797e-06, "loss": 0.5246, "num_input_tokens_seen": 21324704, "step": 32585 }, { "epoch": 17.08071278825996, "grad_norm": 0.17773623764514923, "learning_rate": 3.176546440590508e-06, "loss": 0.3757, "num_input_tokens_seen": 21327040, "step": 32590 }, { "epoch": 17.083333333333332, "grad_norm": 0.24042101204395294, "learning_rate": 3.170970725737041e-06, "loss": 0.452, "num_input_tokens_seen": 21329824, "step": 32595 }, { "epoch": 17.08595387840671, "grad_norm": 0.1102425754070282, "learning_rate": 3.165399577267461e-06, "loss": 0.4679, "num_input_tokens_seen": 21333024, "step": 32600 }, { "epoch": 17.088574423480082, "grad_norm": 0.14385244250297546, "learning_rate": 3.159832996347184e-06, "loss": 0.4549, "num_input_tokens_seen": 21336480, "step": 32605 }, { "epoch": 17.09119496855346, "grad_norm": 0.19511103630065918, "learning_rate": 3.1542709841406865e-06, "loss": 0.4478, "num_input_tokens_seen": 21340160, "step": 32610 }, { "epoch": 17.093815513626833, "grad_norm": 0.13722606003284454, "learning_rate": 3.148713541811468e-06, "loss": 0.4142, "num_input_tokens_seen": 21344096, "step": 32615 }, { "epoch": 17.09643605870021, "grad_norm": 0.12223446369171143, "learning_rate": 3.1431606705220803e-06, "loss": 0.4264, "num_input_tokens_seen": 21347712, "step": 32620 }, { "epoch": 17.099056603773583, "grad_norm": 0.13350531458854675, "learning_rate": 3.1376123714341234e-06, "loss": 0.5163, "num_input_tokens_seen": 21351168, "step": 32625 }, { "epoch": 17.10167714884696, "grad_norm": 0.29946064949035645, "learning_rate": 3.1320686457082387e-06, "loss": 0.4211, "num_input_tokens_seen": 21354016, "step": 32630 }, { "epoch": 17.104297693920337, "grad_norm": 0.21688038110733032, "learning_rate": 3.126529494504113e-06, "loss": 0.7462, "num_input_tokens_seen": 21357440, "step": 32635 }, { "epoch": 17.10691823899371, "grad_norm": 0.13136781752109528, "learning_rate": 3.1209949189804744e-06, "loss": 0.5878, "num_input_tokens_seen": 21360512, "step": 32640 }, { "epoch": 17.109538784067087, "grad_norm": 0.24760620296001434, "learning_rate": 3.115464920295089e-06, "loss": 0.439, "num_input_tokens_seen": 21363040, "step": 32645 }, { "epoch": 17.11215932914046, "grad_norm": 0.14389099180698395, "learning_rate": 3.1099394996047666e-06, "loss": 0.4673, "num_input_tokens_seen": 21365568, "step": 32650 }, { "epoch": 17.114779874213838, "grad_norm": 0.15760059654712677, "learning_rate": 3.104418658065367e-06, "loss": 0.6063, "num_input_tokens_seen": 21368960, "step": 32655 }, { "epoch": 17.11740041928721, "grad_norm": 0.23276183009147644, "learning_rate": 3.098902396831782e-06, "loss": 0.5757, "num_input_tokens_seen": 21371712, "step": 32660 }, { "epoch": 17.120020964360588, "grad_norm": 0.18330597877502441, "learning_rate": 3.0933907170579646e-06, "loss": 0.4586, "num_input_tokens_seen": 21375040, "step": 32665 }, { "epoch": 17.12264150943396, "grad_norm": 0.14896318316459656, "learning_rate": 3.0878836198968835e-06, "loss": 0.4477, "num_input_tokens_seen": 21378880, "step": 32670 }, { "epoch": 17.12526205450734, "grad_norm": 0.19532868266105652, "learning_rate": 3.0823811065005564e-06, "loss": 0.4638, "num_input_tokens_seen": 21382048, "step": 32675 }, { "epoch": 17.127882599580712, "grad_norm": 0.11111519485712051, "learning_rate": 3.0768831780200542e-06, "loss": 0.4281, "num_input_tokens_seen": 21385888, "step": 32680 }, { "epoch": 17.13050314465409, "grad_norm": 0.10421176999807358, "learning_rate": 3.071389835605476e-06, "loss": 0.4557, "num_input_tokens_seen": 21390208, "step": 32685 }, { "epoch": 17.133123689727462, "grad_norm": 0.17346252501010895, "learning_rate": 3.065901080405978e-06, "loss": 0.4001, "num_input_tokens_seen": 21393824, "step": 32690 }, { "epoch": 17.13574423480084, "grad_norm": 0.10613640397787094, "learning_rate": 3.060416913569733e-06, "loss": 0.4, "num_input_tokens_seen": 21396960, "step": 32695 }, { "epoch": 17.138364779874212, "grad_norm": 0.19716444611549377, "learning_rate": 3.054937336243968e-06, "loss": 0.597, "num_input_tokens_seen": 21400832, "step": 32700 }, { "epoch": 17.14098532494759, "grad_norm": 0.0879833847284317, "learning_rate": 3.049462349574947e-06, "loss": 0.4283, "num_input_tokens_seen": 21405120, "step": 32705 }, { "epoch": 17.143605870020963, "grad_norm": 0.32085028290748596, "learning_rate": 3.043991954707981e-06, "loss": 0.4812, "num_input_tokens_seen": 21407840, "step": 32710 }, { "epoch": 17.14622641509434, "grad_norm": 0.1237717792391777, "learning_rate": 3.0385261527874075e-06, "loss": 0.4849, "num_input_tokens_seen": 21411616, "step": 32715 }, { "epoch": 17.148846960167713, "grad_norm": 0.2308860868215561, "learning_rate": 3.0330649449566205e-06, "loss": 0.4003, "num_input_tokens_seen": 21414528, "step": 32720 }, { "epoch": 17.15146750524109, "grad_norm": 0.17544332146644592, "learning_rate": 3.027608332358034e-06, "loss": 0.4543, "num_input_tokens_seen": 21417344, "step": 32725 }, { "epoch": 17.154088050314467, "grad_norm": 0.2607289254665375, "learning_rate": 3.022156316133104e-06, "loss": 0.5306, "num_input_tokens_seen": 21420000, "step": 32730 }, { "epoch": 17.15670859538784, "grad_norm": 0.24313852190971375, "learning_rate": 3.016708897422335e-06, "loss": 0.501, "num_input_tokens_seen": 21423104, "step": 32735 }, { "epoch": 17.159329140461217, "grad_norm": 0.15403011441230774, "learning_rate": 3.0112660773652685e-06, "loss": 0.4867, "num_input_tokens_seen": 21427744, "step": 32740 }, { "epoch": 17.16194968553459, "grad_norm": 0.1027519628405571, "learning_rate": 3.005827857100482e-06, "loss": 0.4413, "num_input_tokens_seen": 21430976, "step": 32745 }, { "epoch": 17.164570230607968, "grad_norm": 0.23457472026348114, "learning_rate": 3.000394237765586e-06, "loss": 0.5417, "num_input_tokens_seen": 21433664, "step": 32750 }, { "epoch": 17.16719077568134, "grad_norm": 0.2410973161458969, "learning_rate": 2.9949652204972254e-06, "loss": 0.5715, "num_input_tokens_seen": 21436384, "step": 32755 }, { "epoch": 17.169811320754718, "grad_norm": 0.14471273124217987, "learning_rate": 2.989540806431093e-06, "loss": 0.4721, "num_input_tokens_seen": 21439520, "step": 32760 }, { "epoch": 17.17243186582809, "grad_norm": 0.12147015333175659, "learning_rate": 2.984120996701917e-06, "loss": 0.3732, "num_input_tokens_seen": 21445504, "step": 32765 }, { "epoch": 17.17505241090147, "grad_norm": 0.19216035306453705, "learning_rate": 2.9787057924434618e-06, "loss": 0.4749, "num_input_tokens_seen": 21451264, "step": 32770 }, { "epoch": 17.177672955974842, "grad_norm": 0.10168194770812988, "learning_rate": 2.9732951947885166e-06, "loss": 0.4818, "num_input_tokens_seen": 21454144, "step": 32775 }, { "epoch": 17.18029350104822, "grad_norm": 0.10847928375005722, "learning_rate": 2.967889204868929e-06, "loss": 0.4296, "num_input_tokens_seen": 21456928, "step": 32780 }, { "epoch": 17.182914046121592, "grad_norm": 0.1293691247701645, "learning_rate": 2.96248782381556e-06, "loss": 0.3855, "num_input_tokens_seen": 21460032, "step": 32785 }, { "epoch": 17.18553459119497, "grad_norm": 0.28910496830940247, "learning_rate": 2.957091052758321e-06, "loss": 0.5622, "num_input_tokens_seen": 21463040, "step": 32790 }, { "epoch": 17.188155136268342, "grad_norm": 0.1957262009382248, "learning_rate": 2.9516988928261517e-06, "loss": 0.8584, "num_input_tokens_seen": 21466336, "step": 32795 }, { "epoch": 17.19077568134172, "grad_norm": 0.19389605522155762, "learning_rate": 2.9463113451470403e-06, "loss": 0.4815, "num_input_tokens_seen": 21469440, "step": 32800 }, { "epoch": 17.193396226415093, "grad_norm": 0.1983446180820465, "learning_rate": 2.94092841084799e-06, "loss": 0.4901, "num_input_tokens_seen": 21472288, "step": 32805 }, { "epoch": 17.19601677148847, "grad_norm": 0.2660341262817383, "learning_rate": 2.9355500910550586e-06, "loss": 0.4592, "num_input_tokens_seen": 21475136, "step": 32810 }, { "epoch": 17.198637316561843, "grad_norm": 0.1619134098291397, "learning_rate": 2.9301763868933157e-06, "loss": 0.48, "num_input_tokens_seen": 21477632, "step": 32815 }, { "epoch": 17.20125786163522, "grad_norm": 0.11507324129343033, "learning_rate": 2.9248072994868896e-06, "loss": 0.4892, "num_input_tokens_seen": 21481120, "step": 32820 }, { "epoch": 17.203878406708597, "grad_norm": 0.3015812337398529, "learning_rate": 2.919442829958932e-06, "loss": 0.427, "num_input_tokens_seen": 21483872, "step": 32825 }, { "epoch": 17.20649895178197, "grad_norm": 0.11748552322387695, "learning_rate": 2.9140829794316234e-06, "loss": 0.4537, "num_input_tokens_seen": 21487840, "step": 32830 }, { "epoch": 17.209119496855347, "grad_norm": 0.1320207417011261, "learning_rate": 2.908727749026191e-06, "loss": 0.4034, "num_input_tokens_seen": 21490880, "step": 32835 }, { "epoch": 17.21174004192872, "grad_norm": 0.14179345965385437, "learning_rate": 2.9033771398628756e-06, "loss": 0.4278, "num_input_tokens_seen": 21494656, "step": 32840 }, { "epoch": 17.214360587002098, "grad_norm": 0.13321056962013245, "learning_rate": 2.898031153060973e-06, "loss": 0.4054, "num_input_tokens_seen": 21497696, "step": 32845 }, { "epoch": 17.21698113207547, "grad_norm": 0.17853136360645294, "learning_rate": 2.8926897897388033e-06, "loss": 0.5312, "num_input_tokens_seen": 21500640, "step": 32850 }, { "epoch": 17.219601677148848, "grad_norm": 0.1298704743385315, "learning_rate": 2.8873530510137136e-06, "loss": 0.5363, "num_input_tokens_seen": 21503808, "step": 32855 }, { "epoch": 17.22222222222222, "grad_norm": 0.19052338600158691, "learning_rate": 2.8820209380020892e-06, "loss": 0.6993, "num_input_tokens_seen": 21506944, "step": 32860 }, { "epoch": 17.2248427672956, "grad_norm": 0.25055673718452454, "learning_rate": 2.8766934518193566e-06, "loss": 0.5468, "num_input_tokens_seen": 21509824, "step": 32865 }, { "epoch": 17.22746331236897, "grad_norm": 0.2621743977069855, "learning_rate": 2.87137059357995e-06, "loss": 0.3959, "num_input_tokens_seen": 21512864, "step": 32870 }, { "epoch": 17.23008385744235, "grad_norm": 0.14963707327842712, "learning_rate": 2.8660523643973618e-06, "loss": 0.4217, "num_input_tokens_seen": 21516320, "step": 32875 }, { "epoch": 17.232704402515722, "grad_norm": 0.19242416322231293, "learning_rate": 2.860738765384105e-06, "loss": 0.5065, "num_input_tokens_seen": 21522240, "step": 32880 }, { "epoch": 17.2353249475891, "grad_norm": 0.22614851593971252, "learning_rate": 2.8554297976517154e-06, "loss": 0.544, "num_input_tokens_seen": 21525952, "step": 32885 }, { "epoch": 17.237945492662472, "grad_norm": 0.14457164704799652, "learning_rate": 2.8501254623107707e-06, "loss": 0.3927, "num_input_tokens_seen": 21530528, "step": 32890 }, { "epoch": 17.24056603773585, "grad_norm": 0.2872898578643799, "learning_rate": 2.84482576047089e-06, "loss": 0.7117, "num_input_tokens_seen": 21534400, "step": 32895 }, { "epoch": 17.243186582809223, "grad_norm": 0.12332871556282043, "learning_rate": 2.839530693240694e-06, "loss": 0.3907, "num_input_tokens_seen": 21537248, "step": 32900 }, { "epoch": 17.2458071278826, "grad_norm": 0.1504756212234497, "learning_rate": 2.8342402617278606e-06, "loss": 0.51, "num_input_tokens_seen": 21539808, "step": 32905 }, { "epoch": 17.248427672955973, "grad_norm": 0.29201486706733704, "learning_rate": 2.82895446703908e-06, "loss": 0.4358, "num_input_tokens_seen": 21546656, "step": 32910 }, { "epoch": 17.25104821802935, "grad_norm": 0.1445881426334381, "learning_rate": 2.823673310280084e-06, "loss": 0.5121, "num_input_tokens_seen": 21549152, "step": 32915 }, { "epoch": 17.253668763102727, "grad_norm": 0.13799995183944702, "learning_rate": 2.8183967925556305e-06, "loss": 0.5081, "num_input_tokens_seen": 21552832, "step": 32920 }, { "epoch": 17.2562893081761, "grad_norm": 0.21189330518245697, "learning_rate": 2.8131249149695115e-06, "loss": 0.4179, "num_input_tokens_seen": 21555776, "step": 32925 }, { "epoch": 17.258909853249477, "grad_norm": 0.2906430959701538, "learning_rate": 2.8078576786245397e-06, "loss": 0.4931, "num_input_tokens_seen": 21558624, "step": 32930 }, { "epoch": 17.26153039832285, "grad_norm": 0.11303403228521347, "learning_rate": 2.8025950846225557e-06, "loss": 0.3103, "num_input_tokens_seen": 21562432, "step": 32935 }, { "epoch": 17.264150943396228, "grad_norm": 0.18607117235660553, "learning_rate": 2.7973371340644373e-06, "loss": 0.4297, "num_input_tokens_seen": 21565856, "step": 32940 }, { "epoch": 17.2667714884696, "grad_norm": 0.15296177566051483, "learning_rate": 2.7920838280500907e-06, "loss": 0.3949, "num_input_tokens_seen": 21569312, "step": 32945 }, { "epoch": 17.269392033542978, "grad_norm": 0.2778351306915283, "learning_rate": 2.7868351676784504e-06, "loss": 0.5231, "num_input_tokens_seen": 21573152, "step": 32950 }, { "epoch": 17.27201257861635, "grad_norm": 0.15737558901309967, "learning_rate": 2.7815911540474665e-06, "loss": 0.5778, "num_input_tokens_seen": 21575840, "step": 32955 }, { "epoch": 17.27463312368973, "grad_norm": 0.12082887440919876, "learning_rate": 2.7763517882541364e-06, "loss": 0.4339, "num_input_tokens_seen": 21578304, "step": 32960 }, { "epoch": 17.2772536687631, "grad_norm": 0.2109118551015854, "learning_rate": 2.771117071394466e-06, "loss": 0.6253, "num_input_tokens_seen": 21580832, "step": 32965 }, { "epoch": 17.27987421383648, "grad_norm": 0.15689323842525482, "learning_rate": 2.7658870045635055e-06, "loss": 0.5441, "num_input_tokens_seen": 21583456, "step": 32970 }, { "epoch": 17.282494758909852, "grad_norm": 0.3499176502227783, "learning_rate": 2.7606615888553234e-06, "loss": 0.454, "num_input_tokens_seen": 21585856, "step": 32975 }, { "epoch": 17.28511530398323, "grad_norm": 0.13455481827259064, "learning_rate": 2.755440825363023e-06, "loss": 0.4854, "num_input_tokens_seen": 21588960, "step": 32980 }, { "epoch": 17.287735849056602, "grad_norm": 0.2010573446750641, "learning_rate": 2.75022471517872e-06, "loss": 0.4956, "num_input_tokens_seen": 21591904, "step": 32985 }, { "epoch": 17.29035639412998, "grad_norm": 0.19460171461105347, "learning_rate": 2.745013259393564e-06, "loss": 0.5202, "num_input_tokens_seen": 21594848, "step": 32990 }, { "epoch": 17.292976939203353, "grad_norm": 0.18053561449050903, "learning_rate": 2.739806459097735e-06, "loss": 0.4357, "num_input_tokens_seen": 21598112, "step": 32995 }, { "epoch": 17.29559748427673, "grad_norm": 0.21135126054286957, "learning_rate": 2.7346043153804403e-06, "loss": 0.5856, "num_input_tokens_seen": 21601312, "step": 33000 }, { "epoch": 17.298218029350103, "grad_norm": 0.20175795257091522, "learning_rate": 2.729406829329903e-06, "loss": 0.5159, "num_input_tokens_seen": 21604160, "step": 33005 }, { "epoch": 17.30083857442348, "grad_norm": 0.12421818822622299, "learning_rate": 2.72421400203339e-06, "loss": 0.4626, "num_input_tokens_seen": 21607008, "step": 33010 }, { "epoch": 17.303459119496857, "grad_norm": 0.21158504486083984, "learning_rate": 2.719025834577163e-06, "loss": 0.3746, "num_input_tokens_seen": 21610144, "step": 33015 }, { "epoch": 17.30607966457023, "grad_norm": 0.14295819401741028, "learning_rate": 2.713842328046537e-06, "loss": 0.6349, "num_input_tokens_seen": 21613312, "step": 33020 }, { "epoch": 17.308700209643607, "grad_norm": 0.2792227268218994, "learning_rate": 2.7086634835258416e-06, "loss": 0.5438, "num_input_tokens_seen": 21615936, "step": 33025 }, { "epoch": 17.31132075471698, "grad_norm": 0.16076837480068207, "learning_rate": 2.7034893020984333e-06, "loss": 0.5933, "num_input_tokens_seen": 21619200, "step": 33030 }, { "epoch": 17.313941299790358, "grad_norm": 0.15534788370132446, "learning_rate": 2.6983197848466952e-06, "loss": 0.4566, "num_input_tokens_seen": 21621792, "step": 33035 }, { "epoch": 17.31656184486373, "grad_norm": 0.13577939569950104, "learning_rate": 2.6931549328520265e-06, "loss": 0.4345, "num_input_tokens_seen": 21624448, "step": 33040 }, { "epoch": 17.319182389937108, "grad_norm": 0.19445325434207916, "learning_rate": 2.6879947471948523e-06, "loss": 0.3776, "num_input_tokens_seen": 21628576, "step": 33045 }, { "epoch": 17.32180293501048, "grad_norm": 0.13467080891132355, "learning_rate": 2.6828392289546285e-06, "loss": 0.5374, "num_input_tokens_seen": 21632128, "step": 33050 }, { "epoch": 17.32442348008386, "grad_norm": 0.1417396366596222, "learning_rate": 2.677688379209828e-06, "loss": 0.4965, "num_input_tokens_seen": 21635616, "step": 33055 }, { "epoch": 17.32704402515723, "grad_norm": 0.15724840760231018, "learning_rate": 2.6725421990379586e-06, "loss": 0.354, "num_input_tokens_seen": 21638592, "step": 33060 }, { "epoch": 17.32966457023061, "grad_norm": 0.1408902406692505, "learning_rate": 2.667400689515537e-06, "loss": 0.4448, "num_input_tokens_seen": 21641920, "step": 33065 }, { "epoch": 17.332285115303982, "grad_norm": 0.1523241251707077, "learning_rate": 2.6622638517181e-06, "loss": 0.6119, "num_input_tokens_seen": 21645024, "step": 33070 }, { "epoch": 17.33490566037736, "grad_norm": 0.20360121130943298, "learning_rate": 2.657131686720224e-06, "loss": 0.6772, "num_input_tokens_seen": 21647904, "step": 33075 }, { "epoch": 17.337526205450732, "grad_norm": 0.18193738162517548, "learning_rate": 2.652004195595498e-06, "loss": 0.6088, "num_input_tokens_seen": 21650528, "step": 33080 }, { "epoch": 17.34014675052411, "grad_norm": 0.23330719769001007, "learning_rate": 2.6468813794165356e-06, "loss": 0.5836, "num_input_tokens_seen": 21653664, "step": 33085 }, { "epoch": 17.342767295597483, "grad_norm": 0.2512466013431549, "learning_rate": 2.641763239254977e-06, "loss": 0.5939, "num_input_tokens_seen": 21656640, "step": 33090 }, { "epoch": 17.34538784067086, "grad_norm": 0.2997932732105255, "learning_rate": 2.6366497761814713e-06, "loss": 0.4418, "num_input_tokens_seen": 21659616, "step": 33095 }, { "epoch": 17.348008385744233, "grad_norm": 0.1422974318265915, "learning_rate": 2.631540991265691e-06, "loss": 0.5264, "num_input_tokens_seen": 21663040, "step": 33100 }, { "epoch": 17.35062893081761, "grad_norm": 0.16045941412448883, "learning_rate": 2.6264368855763443e-06, "loss": 0.547, "num_input_tokens_seen": 21666688, "step": 33105 }, { "epoch": 17.353249475890987, "grad_norm": 0.2339455783367157, "learning_rate": 2.6213374601811504e-06, "loss": 0.5686, "num_input_tokens_seen": 21669440, "step": 33110 }, { "epoch": 17.35587002096436, "grad_norm": 0.16447517275810242, "learning_rate": 2.616242716146855e-06, "loss": 0.4899, "num_input_tokens_seen": 21672448, "step": 33115 }, { "epoch": 17.358490566037737, "grad_norm": 0.14714890718460083, "learning_rate": 2.6111526545392145e-06, "loss": 0.5674, "num_input_tokens_seen": 21675488, "step": 33120 }, { "epoch": 17.36111111111111, "grad_norm": 0.15988098084926605, "learning_rate": 2.6060672764230106e-06, "loss": 0.4894, "num_input_tokens_seen": 21678272, "step": 33125 }, { "epoch": 17.363731656184488, "grad_norm": 0.17828631401062012, "learning_rate": 2.600986582862047e-06, "loss": 0.4151, "num_input_tokens_seen": 21681088, "step": 33130 }, { "epoch": 17.36635220125786, "grad_norm": 0.19104944169521332, "learning_rate": 2.5959105749191498e-06, "loss": 0.5348, "num_input_tokens_seen": 21683872, "step": 33135 }, { "epoch": 17.368972746331238, "grad_norm": 0.18070848286151886, "learning_rate": 2.590839253656166e-06, "loss": 0.539, "num_input_tokens_seen": 21687104, "step": 33140 }, { "epoch": 17.37159329140461, "grad_norm": 0.08777166157960892, "learning_rate": 2.5857726201339473e-06, "loss": 0.4393, "num_input_tokens_seen": 21689952, "step": 33145 }, { "epoch": 17.37421383647799, "grad_norm": 0.14102299511432648, "learning_rate": 2.5807106754123873e-06, "loss": 0.5925, "num_input_tokens_seen": 21693088, "step": 33150 }, { "epoch": 17.37683438155136, "grad_norm": 0.1793557107448578, "learning_rate": 2.575653420550378e-06, "loss": 0.5176, "num_input_tokens_seen": 21695872, "step": 33155 }, { "epoch": 17.37945492662474, "grad_norm": 0.2299502044916153, "learning_rate": 2.57060085660584e-06, "loss": 0.622, "num_input_tokens_seen": 21698272, "step": 33160 }, { "epoch": 17.382075471698112, "grad_norm": 0.173005148768425, "learning_rate": 2.565552984635716e-06, "loss": 0.4753, "num_input_tokens_seen": 21701984, "step": 33165 }, { "epoch": 17.38469601677149, "grad_norm": 0.11179093271493912, "learning_rate": 2.5605098056959652e-06, "loss": 0.4755, "num_input_tokens_seen": 21705856, "step": 33170 }, { "epoch": 17.387316561844862, "grad_norm": 0.21763181686401367, "learning_rate": 2.555471320841557e-06, "loss": 0.3855, "num_input_tokens_seen": 21708160, "step": 33175 }, { "epoch": 17.38993710691824, "grad_norm": 0.18383964896202087, "learning_rate": 2.55043753112649e-06, "loss": 0.4987, "num_input_tokens_seen": 21713760, "step": 33180 }, { "epoch": 17.392557651991613, "grad_norm": 0.17398923635482788, "learning_rate": 2.5454084376037717e-06, "loss": 0.5305, "num_input_tokens_seen": 21716768, "step": 33185 }, { "epoch": 17.39517819706499, "grad_norm": 0.15349897742271423, "learning_rate": 2.5403840413254304e-06, "loss": 0.531, "num_input_tokens_seen": 21720128, "step": 33190 }, { "epoch": 17.397798742138363, "grad_norm": 0.1614995151758194, "learning_rate": 2.5353643433425196e-06, "loss": 0.4521, "num_input_tokens_seen": 21724384, "step": 33195 }, { "epoch": 17.40041928721174, "grad_norm": 0.2642351984977722, "learning_rate": 2.5303493447050965e-06, "loss": 0.4944, "num_input_tokens_seen": 21727136, "step": 33200 }, { "epoch": 17.403039832285117, "grad_norm": 0.11556317657232285, "learning_rate": 2.525339046462244e-06, "loss": 0.4118, "num_input_tokens_seen": 21731136, "step": 33205 }, { "epoch": 17.40566037735849, "grad_norm": 0.24079196155071259, "learning_rate": 2.5203334496620544e-06, "loss": 0.4965, "num_input_tokens_seen": 21734272, "step": 33210 }, { "epoch": 17.408280922431867, "grad_norm": 0.18777979910373688, "learning_rate": 2.515332555351646e-06, "loss": 0.4613, "num_input_tokens_seen": 21738112, "step": 33215 }, { "epoch": 17.41090146750524, "grad_norm": 0.12224578112363815, "learning_rate": 2.510336364577154e-06, "loss": 0.6226, "num_input_tokens_seen": 21740896, "step": 33220 }, { "epoch": 17.413522012578618, "grad_norm": 0.21876071393489838, "learning_rate": 2.5053448783837143e-06, "loss": 0.537, "num_input_tokens_seen": 21744448, "step": 33225 }, { "epoch": 17.41614255765199, "grad_norm": 0.13733293116092682, "learning_rate": 2.5003580978154917e-06, "loss": 0.41, "num_input_tokens_seen": 21747200, "step": 33230 }, { "epoch": 17.418763102725368, "grad_norm": 0.15715420246124268, "learning_rate": 2.4953760239156716e-06, "loss": 0.5119, "num_input_tokens_seen": 21750432, "step": 33235 }, { "epoch": 17.42138364779874, "grad_norm": 0.12656746804714203, "learning_rate": 2.4903986577264366e-06, "loss": 0.5442, "num_input_tokens_seen": 21753856, "step": 33240 }, { "epoch": 17.42400419287212, "grad_norm": 0.2030569612979889, "learning_rate": 2.4854260002889986e-06, "loss": 0.4186, "num_input_tokens_seen": 21756960, "step": 33245 }, { "epoch": 17.42662473794549, "grad_norm": 0.13721628487110138, "learning_rate": 2.480458052643586e-06, "loss": 0.4341, "num_input_tokens_seen": 21759968, "step": 33250 }, { "epoch": 17.42924528301887, "grad_norm": 0.10953066498041153, "learning_rate": 2.4754948158294293e-06, "loss": 0.4542, "num_input_tokens_seen": 21762592, "step": 33255 }, { "epoch": 17.431865828092242, "grad_norm": 0.20473583042621613, "learning_rate": 2.4705362908847814e-06, "loss": 0.371, "num_input_tokens_seen": 21765440, "step": 33260 }, { "epoch": 17.43448637316562, "grad_norm": 0.22081345319747925, "learning_rate": 2.4655824788469172e-06, "loss": 0.5428, "num_input_tokens_seen": 21768736, "step": 33265 }, { "epoch": 17.437106918238992, "grad_norm": 0.18410557508468628, "learning_rate": 2.4606333807521087e-06, "loss": 0.4744, "num_input_tokens_seen": 21771392, "step": 33270 }, { "epoch": 17.43972746331237, "grad_norm": 0.19730596244335175, "learning_rate": 2.455688997635658e-06, "loss": 0.4704, "num_input_tokens_seen": 21774176, "step": 33275 }, { "epoch": 17.442348008385743, "grad_norm": 0.13374517858028412, "learning_rate": 2.4507493305318652e-06, "loss": 0.4128, "num_input_tokens_seen": 21776736, "step": 33280 }, { "epoch": 17.44496855345912, "grad_norm": 0.1694106161594391, "learning_rate": 2.445814380474057e-06, "loss": 0.4572, "num_input_tokens_seen": 21780480, "step": 33285 }, { "epoch": 17.447589098532493, "grad_norm": 0.13360704481601715, "learning_rate": 2.4408841484945684e-06, "loss": 0.4798, "num_input_tokens_seen": 21784032, "step": 33290 }, { "epoch": 17.45020964360587, "grad_norm": 0.2129502296447754, "learning_rate": 2.435958635624755e-06, "loss": 0.4436, "num_input_tokens_seen": 21787584, "step": 33295 }, { "epoch": 17.452830188679247, "grad_norm": 0.19959281384944916, "learning_rate": 2.4310378428949678e-06, "loss": 0.4351, "num_input_tokens_seen": 21790336, "step": 33300 }, { "epoch": 17.45545073375262, "grad_norm": 0.22147582471370697, "learning_rate": 2.4261217713345804e-06, "loss": 0.5933, "num_input_tokens_seen": 21793216, "step": 33305 }, { "epoch": 17.458071278825997, "grad_norm": 0.22280967235565186, "learning_rate": 2.421210421971981e-06, "loss": 0.5296, "num_input_tokens_seen": 21796160, "step": 33310 }, { "epoch": 17.46069182389937, "grad_norm": 0.18773727118968964, "learning_rate": 2.416303795834571e-06, "loss": 0.5149, "num_input_tokens_seen": 21798656, "step": 33315 }, { "epoch": 17.463312368972748, "grad_norm": 0.2523854374885559, "learning_rate": 2.4114018939487615e-06, "loss": 0.382, "num_input_tokens_seen": 21802912, "step": 33320 }, { "epoch": 17.46593291404612, "grad_norm": 0.21096044778823853, "learning_rate": 2.4065047173399662e-06, "loss": 0.4092, "num_input_tokens_seen": 21806304, "step": 33325 }, { "epoch": 17.468553459119498, "grad_norm": 0.13346201181411743, "learning_rate": 2.4016122670326295e-06, "loss": 0.5909, "num_input_tokens_seen": 21809760, "step": 33330 }, { "epoch": 17.47117400419287, "grad_norm": 0.1116323322057724, "learning_rate": 2.396724544050186e-06, "loss": 0.3804, "num_input_tokens_seen": 21812576, "step": 33335 }, { "epoch": 17.47379454926625, "grad_norm": 0.24100564420223236, "learning_rate": 2.3918415494150944e-06, "loss": 0.4345, "num_input_tokens_seen": 21815392, "step": 33340 }, { "epoch": 17.47641509433962, "grad_norm": 0.2223561704158783, "learning_rate": 2.3869632841488245e-06, "loss": 0.4552, "num_input_tokens_seen": 21817920, "step": 33345 }, { "epoch": 17.479035639413, "grad_norm": 0.24561406672000885, "learning_rate": 2.382089749271857e-06, "loss": 0.4929, "num_input_tokens_seen": 21820704, "step": 33350 }, { "epoch": 17.481656184486372, "grad_norm": 0.21661429107189178, "learning_rate": 2.3772209458036737e-06, "loss": 0.3927, "num_input_tokens_seen": 21823872, "step": 33355 }, { "epoch": 17.48427672955975, "grad_norm": 0.23347558081150055, "learning_rate": 2.372356874762771e-06, "loss": 0.4759, "num_input_tokens_seen": 21827584, "step": 33360 }, { "epoch": 17.486897274633122, "grad_norm": 0.17925989627838135, "learning_rate": 2.36749753716666e-06, "loss": 0.4241, "num_input_tokens_seen": 21830624, "step": 33365 }, { "epoch": 17.4895178197065, "grad_norm": 0.09876593202352524, "learning_rate": 2.3626429340318613e-06, "loss": 0.4511, "num_input_tokens_seen": 21834208, "step": 33370 }, { "epoch": 17.492138364779873, "grad_norm": 0.19050852954387665, "learning_rate": 2.357793066373901e-06, "loss": 0.67, "num_input_tokens_seen": 21837408, "step": 33375 }, { "epoch": 17.49475890985325, "grad_norm": 0.15785253047943115, "learning_rate": 2.352947935207317e-06, "loss": 0.4236, "num_input_tokens_seen": 21840160, "step": 33380 }, { "epoch": 17.497379454926623, "grad_norm": 0.40560388565063477, "learning_rate": 2.3481075415456495e-06, "loss": 0.4595, "num_input_tokens_seen": 21844288, "step": 33385 }, { "epoch": 17.5, "grad_norm": 0.39280131459236145, "learning_rate": 2.3432718864014576e-06, "loss": 0.6559, "num_input_tokens_seen": 21846880, "step": 33390 }, { "epoch": 17.502620545073377, "grad_norm": 0.14678841829299927, "learning_rate": 2.338440970786304e-06, "loss": 0.5205, "num_input_tokens_seen": 21850048, "step": 33395 }, { "epoch": 17.50524109014675, "grad_norm": 0.12094510346651077, "learning_rate": 2.3336147957107644e-06, "loss": 0.374, "num_input_tokens_seen": 21853088, "step": 33400 }, { "epoch": 17.507861635220127, "grad_norm": 0.2423229217529297, "learning_rate": 2.3287933621844227e-06, "loss": 0.4242, "num_input_tokens_seen": 21856000, "step": 33405 }, { "epoch": 17.5104821802935, "grad_norm": 0.18057703971862793, "learning_rate": 2.3239766712158633e-06, "loss": 0.4056, "num_input_tokens_seen": 21859392, "step": 33410 }, { "epoch": 17.513102725366878, "grad_norm": 0.11706189066171646, "learning_rate": 2.319164723812675e-06, "loss": 0.4382, "num_input_tokens_seen": 21862912, "step": 33415 }, { "epoch": 17.51572327044025, "grad_norm": 0.1415494829416275, "learning_rate": 2.314357520981472e-06, "loss": 0.5512, "num_input_tokens_seen": 21866016, "step": 33420 }, { "epoch": 17.518343815513628, "grad_norm": 0.17838752269744873, "learning_rate": 2.309555063727864e-06, "loss": 0.5469, "num_input_tokens_seen": 21869248, "step": 33425 }, { "epoch": 17.520964360587, "grad_norm": 0.2102775126695633, "learning_rate": 2.304757353056475e-06, "loss": 0.4748, "num_input_tokens_seen": 21871968, "step": 33430 }, { "epoch": 17.52358490566038, "grad_norm": 0.1604049652814865, "learning_rate": 2.2999643899709267e-06, "loss": 0.58, "num_input_tokens_seen": 21875456, "step": 33435 }, { "epoch": 17.52620545073375, "grad_norm": 0.32317182421684265, "learning_rate": 2.2951761754738456e-06, "loss": 0.3908, "num_input_tokens_seen": 21877792, "step": 33440 }, { "epoch": 17.52882599580713, "grad_norm": 0.16889970004558563, "learning_rate": 2.29039271056688e-06, "loss": 0.5125, "num_input_tokens_seen": 21881440, "step": 33445 }, { "epoch": 17.531446540880502, "grad_norm": 0.236192986369133, "learning_rate": 2.285613996250674e-06, "loss": 0.4094, "num_input_tokens_seen": 21884256, "step": 33450 }, { "epoch": 17.53406708595388, "grad_norm": 0.19628599286079407, "learning_rate": 2.2808400335248865e-06, "loss": 0.4925, "num_input_tokens_seen": 21887136, "step": 33455 }, { "epoch": 17.536687631027252, "grad_norm": 0.18942125141620636, "learning_rate": 2.2760708233881654e-06, "loss": 0.4894, "num_input_tokens_seen": 21889568, "step": 33460 }, { "epoch": 17.53930817610063, "grad_norm": 0.21586760878562927, "learning_rate": 2.271306366838183e-06, "loss": 0.6483, "num_input_tokens_seen": 21892864, "step": 33465 }, { "epoch": 17.541928721174003, "grad_norm": 0.2647119164466858, "learning_rate": 2.266546664871602e-06, "loss": 0.5524, "num_input_tokens_seen": 21895968, "step": 33470 }, { "epoch": 17.54454926624738, "grad_norm": 0.14087508618831635, "learning_rate": 2.261791718484102e-06, "loss": 0.5719, "num_input_tokens_seen": 21898912, "step": 33475 }, { "epoch": 17.547169811320753, "grad_norm": 0.12738977372646332, "learning_rate": 2.257041528670362e-06, "loss": 0.3993, "num_input_tokens_seen": 21903456, "step": 33480 }, { "epoch": 17.54979035639413, "grad_norm": 0.13313037157058716, "learning_rate": 2.2522960964240753e-06, "loss": 0.5092, "num_input_tokens_seen": 21906336, "step": 33485 }, { "epoch": 17.552410901467507, "grad_norm": 0.1719088852405548, "learning_rate": 2.247555422737924e-06, "loss": 0.3935, "num_input_tokens_seen": 21910624, "step": 33490 }, { "epoch": 17.55503144654088, "grad_norm": 0.23525424301624298, "learning_rate": 2.2428195086036024e-06, "loss": 0.4224, "num_input_tokens_seen": 21914816, "step": 33495 }, { "epoch": 17.557651991614257, "grad_norm": 0.15251488983631134, "learning_rate": 2.238088355011811e-06, "loss": 0.3705, "num_input_tokens_seen": 21919456, "step": 33500 }, { "epoch": 17.56027253668763, "grad_norm": 0.126212015748024, "learning_rate": 2.2333619629522518e-06, "loss": 0.5305, "num_input_tokens_seen": 21922880, "step": 33505 }, { "epoch": 17.562893081761008, "grad_norm": 0.1650032103061676, "learning_rate": 2.2286403334136424e-06, "loss": 0.4839, "num_input_tokens_seen": 21926240, "step": 33510 }, { "epoch": 17.56551362683438, "grad_norm": 0.4597201645374298, "learning_rate": 2.223923467383679e-06, "loss": 0.4987, "num_input_tokens_seen": 21929024, "step": 33515 }, { "epoch": 17.568134171907758, "grad_norm": 0.20210000872612, "learning_rate": 2.2192113658490886e-06, "loss": 0.395, "num_input_tokens_seen": 21931904, "step": 33520 }, { "epoch": 17.57075471698113, "grad_norm": 0.23351728916168213, "learning_rate": 2.2145040297955776e-06, "loss": 0.7432, "num_input_tokens_seen": 21934784, "step": 33525 }, { "epoch": 17.57337526205451, "grad_norm": 0.1589120328426361, "learning_rate": 2.209801460207875e-06, "loss": 0.4103, "num_input_tokens_seen": 21937376, "step": 33530 }, { "epoch": 17.57599580712788, "grad_norm": 0.12507380545139313, "learning_rate": 2.2051036580697042e-06, "loss": 0.5224, "num_input_tokens_seen": 21941568, "step": 33535 }, { "epoch": 17.57861635220126, "grad_norm": 0.1314265877008438, "learning_rate": 2.2004106243637885e-06, "loss": 0.4171, "num_input_tokens_seen": 21944800, "step": 33540 }, { "epoch": 17.581236897274632, "grad_norm": 0.22998790442943573, "learning_rate": 2.1957223600718587e-06, "loss": 0.4877, "num_input_tokens_seen": 21947680, "step": 33545 }, { "epoch": 17.58385744234801, "grad_norm": 0.1362880915403366, "learning_rate": 2.1910388661746493e-06, "loss": 0.5116, "num_input_tokens_seen": 21951392, "step": 33550 }, { "epoch": 17.586477987421382, "grad_norm": 0.18940067291259766, "learning_rate": 2.186360143651886e-06, "loss": 0.5685, "num_input_tokens_seen": 21954464, "step": 33555 }, { "epoch": 17.58909853249476, "grad_norm": 0.294528603553772, "learning_rate": 2.1816861934823103e-06, "loss": 0.4538, "num_input_tokens_seen": 21957504, "step": 33560 }, { "epoch": 17.591719077568133, "grad_norm": 0.13910183310508728, "learning_rate": 2.1770170166436595e-06, "loss": 0.4688, "num_input_tokens_seen": 21960320, "step": 33565 }, { "epoch": 17.59433962264151, "grad_norm": 0.15034641325473785, "learning_rate": 2.172352614112669e-06, "loss": 0.5168, "num_input_tokens_seen": 21963648, "step": 33570 }, { "epoch": 17.596960167714883, "grad_norm": 0.20202088356018066, "learning_rate": 2.167692986865083e-06, "loss": 0.5202, "num_input_tokens_seen": 21967264, "step": 33575 }, { "epoch": 17.59958071278826, "grad_norm": 0.133761465549469, "learning_rate": 2.1630381358756362e-06, "loss": 0.488, "num_input_tokens_seen": 21970464, "step": 33580 }, { "epoch": 17.602201257861637, "grad_norm": 0.21785087883472443, "learning_rate": 2.158388062118072e-06, "loss": 0.4805, "num_input_tokens_seen": 21973536, "step": 33585 }, { "epoch": 17.60482180293501, "grad_norm": 0.12608449161052704, "learning_rate": 2.1537427665651393e-06, "loss": 0.4334, "num_input_tokens_seen": 21976864, "step": 33590 }, { "epoch": 17.607442348008387, "grad_norm": 0.2415756732225418, "learning_rate": 2.1491022501885726e-06, "loss": 0.5072, "num_input_tokens_seen": 21980512, "step": 33595 }, { "epoch": 17.61006289308176, "grad_norm": 0.28121325373649597, "learning_rate": 2.1444665139591174e-06, "loss": 0.5651, "num_input_tokens_seen": 21983744, "step": 33600 }, { "epoch": 17.612683438155138, "grad_norm": 0.1442888230085373, "learning_rate": 2.139835558846523e-06, "loss": 0.4454, "num_input_tokens_seen": 21986336, "step": 33605 }, { "epoch": 17.61530398322851, "grad_norm": 0.12407325953245163, "learning_rate": 2.1352093858195225e-06, "loss": 0.3995, "num_input_tokens_seen": 21989280, "step": 33610 }, { "epoch": 17.617924528301888, "grad_norm": 0.1702502965927124, "learning_rate": 2.1305879958458676e-06, "loss": 0.4466, "num_input_tokens_seen": 21993312, "step": 33615 }, { "epoch": 17.62054507337526, "grad_norm": 0.13380418717861176, "learning_rate": 2.1259713898922894e-06, "loss": 0.5082, "num_input_tokens_seen": 21996160, "step": 33620 }, { "epoch": 17.62316561844864, "grad_norm": 0.11822265386581421, "learning_rate": 2.1213595689245386e-06, "loss": 0.3637, "num_input_tokens_seen": 21999264, "step": 33625 }, { "epoch": 17.62578616352201, "grad_norm": 0.1331692487001419, "learning_rate": 2.1167525339073513e-06, "loss": 0.5063, "num_input_tokens_seen": 22001984, "step": 33630 }, { "epoch": 17.62840670859539, "grad_norm": 0.1533791571855545, "learning_rate": 2.112150285804473e-06, "loss": 0.475, "num_input_tokens_seen": 22007296, "step": 33635 }, { "epoch": 17.631027253668762, "grad_norm": 0.30999311804771423, "learning_rate": 2.107552825578632e-06, "loss": 0.3718, "num_input_tokens_seen": 22010272, "step": 33640 }, { "epoch": 17.63364779874214, "grad_norm": 0.16531676054000854, "learning_rate": 2.102960154191572e-06, "loss": 0.5186, "num_input_tokens_seen": 22013984, "step": 33645 }, { "epoch": 17.636268343815512, "grad_norm": 0.16308538615703583, "learning_rate": 2.098372272604021e-06, "loss": 0.4532, "num_input_tokens_seen": 22017216, "step": 33650 }, { "epoch": 17.63888888888889, "grad_norm": 0.19376099109649658, "learning_rate": 2.0937891817757155e-06, "loss": 0.5334, "num_input_tokens_seen": 22019872, "step": 33655 }, { "epoch": 17.641509433962263, "grad_norm": 0.15739375352859497, "learning_rate": 2.089210882665385e-06, "loss": 0.4353, "num_input_tokens_seen": 22022432, "step": 33660 }, { "epoch": 17.64412997903564, "grad_norm": 0.38212257623672485, "learning_rate": 2.0846373762307635e-06, "loss": 0.5429, "num_input_tokens_seen": 22025120, "step": 33665 }, { "epoch": 17.646750524109013, "grad_norm": 0.16905729472637177, "learning_rate": 2.080068663428569e-06, "loss": 0.4844, "num_input_tokens_seen": 22028896, "step": 33670 }, { "epoch": 17.64937106918239, "grad_norm": 0.43596041202545166, "learning_rate": 2.075504745214521e-06, "loss": 0.4877, "num_input_tokens_seen": 22031168, "step": 33675 }, { "epoch": 17.651991614255767, "grad_norm": 0.18228982388973236, "learning_rate": 2.0709456225433437e-06, "loss": 0.4438, "num_input_tokens_seen": 22034400, "step": 33680 }, { "epoch": 17.65461215932914, "grad_norm": 0.12327168136835098, "learning_rate": 2.0663912963687554e-06, "loss": 0.373, "num_input_tokens_seen": 22037984, "step": 33685 }, { "epoch": 17.657232704402517, "grad_norm": 0.17015044391155243, "learning_rate": 2.0618417676434713e-06, "loss": 0.4663, "num_input_tokens_seen": 22040992, "step": 33690 }, { "epoch": 17.65985324947589, "grad_norm": 0.21610453724861145, "learning_rate": 2.0572970373191975e-06, "loss": 0.507, "num_input_tokens_seen": 22043616, "step": 33695 }, { "epoch": 17.662473794549268, "grad_norm": 0.11677396297454834, "learning_rate": 2.052757106346634e-06, "loss": 0.4046, "num_input_tokens_seen": 22045920, "step": 33700 }, { "epoch": 17.66509433962264, "grad_norm": 0.10889659821987152, "learning_rate": 2.048221975675488e-06, "loss": 0.6303, "num_input_tokens_seen": 22050144, "step": 33705 }, { "epoch": 17.667714884696018, "grad_norm": 0.14625906944274902, "learning_rate": 2.043691646254456e-06, "loss": 0.3874, "num_input_tokens_seen": 22054432, "step": 33710 }, { "epoch": 17.67033542976939, "grad_norm": 0.1660102754831314, "learning_rate": 2.039166119031233e-06, "loss": 0.5272, "num_input_tokens_seen": 22058272, "step": 33715 }, { "epoch": 17.67295597484277, "grad_norm": 0.12914764881134033, "learning_rate": 2.034645394952511e-06, "loss": 0.5109, "num_input_tokens_seen": 22061088, "step": 33720 }, { "epoch": 17.67557651991614, "grad_norm": 0.2232295274734497, "learning_rate": 2.030129474963971e-06, "loss": 0.3875, "num_input_tokens_seen": 22064160, "step": 33725 }, { "epoch": 17.67819706498952, "grad_norm": 0.2007080614566803, "learning_rate": 2.025618360010284e-06, "loss": 0.5084, "num_input_tokens_seen": 22066720, "step": 33730 }, { "epoch": 17.680817610062892, "grad_norm": 0.1032390221953392, "learning_rate": 2.0211120510351317e-06, "loss": 0.5735, "num_input_tokens_seen": 22069952, "step": 33735 }, { "epoch": 17.68343815513627, "grad_norm": 0.21528726816177368, "learning_rate": 2.0166105489811822e-06, "loss": 0.4577, "num_input_tokens_seen": 22074144, "step": 33740 }, { "epoch": 17.686058700209642, "grad_norm": 0.10639598220586777, "learning_rate": 2.0121138547901016e-06, "loss": 0.3977, "num_input_tokens_seen": 22077632, "step": 33745 }, { "epoch": 17.68867924528302, "grad_norm": 0.3734437823295593, "learning_rate": 2.0076219694025427e-06, "loss": 0.4217, "num_input_tokens_seen": 22081536, "step": 33750 }, { "epoch": 17.691299790356393, "grad_norm": 0.17245610058307648, "learning_rate": 2.003134893758152e-06, "loss": 0.5337, "num_input_tokens_seen": 22085600, "step": 33755 }, { "epoch": 17.69392033542977, "grad_norm": 0.16466742753982544, "learning_rate": 1.9986526287955805e-06, "loss": 0.5449, "num_input_tokens_seen": 22089088, "step": 33760 }, { "epoch": 17.696540880503143, "grad_norm": 0.15309946238994598, "learning_rate": 1.9941751754524644e-06, "loss": 0.4228, "num_input_tokens_seen": 22091584, "step": 33765 }, { "epoch": 17.69916142557652, "grad_norm": 0.19101876020431519, "learning_rate": 1.9897025346654412e-06, "loss": 0.3438, "num_input_tokens_seen": 22094496, "step": 33770 }, { "epoch": 17.701781970649897, "grad_norm": 0.15740656852722168, "learning_rate": 1.985234707370134e-06, "loss": 0.453, "num_input_tokens_seen": 22097728, "step": 33775 }, { "epoch": 17.70440251572327, "grad_norm": 0.14404481649398804, "learning_rate": 1.9807716945011544e-06, "loss": 0.6341, "num_input_tokens_seen": 22101696, "step": 33780 }, { "epoch": 17.707023060796647, "grad_norm": 0.31038200855255127, "learning_rate": 1.9763134969921215e-06, "loss": 0.5691, "num_input_tokens_seen": 22104256, "step": 33785 }, { "epoch": 17.70964360587002, "grad_norm": 0.3831585943698883, "learning_rate": 1.971860115775634e-06, "loss": 0.4845, "num_input_tokens_seen": 22107968, "step": 33790 }, { "epoch": 17.712264150943398, "grad_norm": 0.14229436218738556, "learning_rate": 1.967411551783294e-06, "loss": 0.5088, "num_input_tokens_seen": 22110464, "step": 33795 }, { "epoch": 17.71488469601677, "grad_norm": 0.18537180125713348, "learning_rate": 1.9629678059456895e-06, "loss": 0.4476, "num_input_tokens_seen": 22113344, "step": 33800 }, { "epoch": 17.717505241090148, "grad_norm": 0.18958726525306702, "learning_rate": 1.9585288791924004e-06, "loss": 0.372, "num_input_tokens_seen": 22116640, "step": 33805 }, { "epoch": 17.72012578616352, "grad_norm": 0.1803087592124939, "learning_rate": 1.9540947724519947e-06, "loss": 0.4163, "num_input_tokens_seen": 22119808, "step": 33810 }, { "epoch": 17.7227463312369, "grad_norm": 0.24912211298942566, "learning_rate": 1.9496654866520414e-06, "loss": 0.5975, "num_input_tokens_seen": 22123328, "step": 33815 }, { "epoch": 17.72536687631027, "grad_norm": 0.15539845824241638, "learning_rate": 1.9452410227190937e-06, "loss": 0.4608, "num_input_tokens_seen": 22127648, "step": 33820 }, { "epoch": 17.72798742138365, "grad_norm": 0.1309879869222641, "learning_rate": 1.9408213815787087e-06, "loss": 0.4594, "num_input_tokens_seen": 22130816, "step": 33825 }, { "epoch": 17.730607966457022, "grad_norm": 0.16201767325401306, "learning_rate": 1.936406564155413e-06, "loss": 0.4119, "num_input_tokens_seen": 22133568, "step": 33830 }, { "epoch": 17.7332285115304, "grad_norm": 0.15341298282146454, "learning_rate": 1.9319965713727435e-06, "loss": 0.5069, "num_input_tokens_seen": 22137888, "step": 33835 }, { "epoch": 17.735849056603772, "grad_norm": 0.15065781772136688, "learning_rate": 1.9275914041532144e-06, "loss": 0.3766, "num_input_tokens_seen": 22140448, "step": 33840 }, { "epoch": 17.73846960167715, "grad_norm": 0.18652772903442383, "learning_rate": 1.9231910634183394e-06, "loss": 0.5431, "num_input_tokens_seen": 22142400, "step": 33845 }, { "epoch": 17.741090146750523, "grad_norm": 0.2559434771537781, "learning_rate": 1.918795550088623e-06, "loss": 0.5411, "num_input_tokens_seen": 22145824, "step": 33850 }, { "epoch": 17.7437106918239, "grad_norm": 0.15123145282268524, "learning_rate": 1.9144048650835496e-06, "loss": 0.5642, "num_input_tokens_seen": 22149536, "step": 33855 }, { "epoch": 17.746331236897273, "grad_norm": 0.22332151234149933, "learning_rate": 1.9100190093216093e-06, "loss": 0.5181, "num_input_tokens_seen": 22152384, "step": 33860 }, { "epoch": 17.74895178197065, "grad_norm": 0.12252865731716156, "learning_rate": 1.9056379837202627e-06, "loss": 0.5304, "num_input_tokens_seen": 22156480, "step": 33865 }, { "epoch": 17.751572327044027, "grad_norm": 0.11042728275060654, "learning_rate": 1.9012617891959766e-06, "loss": 0.3778, "num_input_tokens_seen": 22160928, "step": 33870 }, { "epoch": 17.7541928721174, "grad_norm": 0.1479407250881195, "learning_rate": 1.8968904266641996e-06, "loss": 0.5238, "num_input_tokens_seen": 22164576, "step": 33875 }, { "epoch": 17.756813417190777, "grad_norm": 0.11081526428461075, "learning_rate": 1.892523897039372e-06, "loss": 0.427, "num_input_tokens_seen": 22168704, "step": 33880 }, { "epoch": 17.75943396226415, "grad_norm": 0.12915189564228058, "learning_rate": 1.888162201234922e-06, "loss": 0.4885, "num_input_tokens_seen": 22172896, "step": 33885 }, { "epoch": 17.762054507337528, "grad_norm": 0.1892303228378296, "learning_rate": 1.8838053401632666e-06, "loss": 0.4972, "num_input_tokens_seen": 22175552, "step": 33890 }, { "epoch": 17.7646750524109, "grad_norm": 0.15676695108413696, "learning_rate": 1.8794533147358074e-06, "loss": 0.4474, "num_input_tokens_seen": 22179200, "step": 33895 }, { "epoch": 17.767295597484278, "grad_norm": 0.0867478996515274, "learning_rate": 1.875106125862941e-06, "loss": 0.4647, "num_input_tokens_seen": 22183456, "step": 33900 }, { "epoch": 17.76991614255765, "grad_norm": 0.14811192452907562, "learning_rate": 1.870763774454057e-06, "loss": 0.5475, "num_input_tokens_seen": 22187584, "step": 33905 }, { "epoch": 17.77253668763103, "grad_norm": 0.19776295125484467, "learning_rate": 1.8664262614175115e-06, "loss": 0.4785, "num_input_tokens_seen": 22191296, "step": 33910 }, { "epoch": 17.7751572327044, "grad_norm": 0.12826399505138397, "learning_rate": 1.862093587660671e-06, "loss": 0.5424, "num_input_tokens_seen": 22195552, "step": 33915 }, { "epoch": 17.77777777777778, "grad_norm": 0.12889911234378815, "learning_rate": 1.8577657540898851e-06, "loss": 0.4409, "num_input_tokens_seen": 22199296, "step": 33920 }, { "epoch": 17.780398322851152, "grad_norm": 0.16812379658222198, "learning_rate": 1.8534427616104798e-06, "loss": 0.4877, "num_input_tokens_seen": 22202144, "step": 33925 }, { "epoch": 17.78301886792453, "grad_norm": 0.14626793563365936, "learning_rate": 1.8491246111267846e-06, "loss": 0.557, "num_input_tokens_seen": 22205440, "step": 33930 }, { "epoch": 17.785639412997902, "grad_norm": 0.16644854843616486, "learning_rate": 1.8448113035420934e-06, "loss": 0.4811, "num_input_tokens_seen": 22209120, "step": 33935 }, { "epoch": 17.78825995807128, "grad_norm": 0.11768826097249985, "learning_rate": 1.840502839758712e-06, "loss": 0.3953, "num_input_tokens_seen": 22213120, "step": 33940 }, { "epoch": 17.790880503144653, "grad_norm": 0.17855696380138397, "learning_rate": 1.8361992206779226e-06, "loss": 0.4426, "num_input_tokens_seen": 22216768, "step": 33945 }, { "epoch": 17.79350104821803, "grad_norm": 0.15213927626609802, "learning_rate": 1.8319004471999856e-06, "loss": 0.3485, "num_input_tokens_seen": 22219968, "step": 33950 }, { "epoch": 17.796121593291403, "grad_norm": 0.17934654653072357, "learning_rate": 1.8276065202241592e-06, "loss": 0.4592, "num_input_tokens_seen": 22222368, "step": 33955 }, { "epoch": 17.79874213836478, "grad_norm": 0.12043432146310806, "learning_rate": 1.8233174406486863e-06, "loss": 0.4408, "num_input_tokens_seen": 22226112, "step": 33960 }, { "epoch": 17.801362683438157, "grad_norm": 0.2204705774784088, "learning_rate": 1.8190332093707907e-06, "loss": 0.4833, "num_input_tokens_seen": 22229728, "step": 33965 }, { "epoch": 17.80398322851153, "grad_norm": 0.13415049016475677, "learning_rate": 1.8147538272866833e-06, "loss": 0.4669, "num_input_tokens_seen": 22232928, "step": 33970 }, { "epoch": 17.806603773584907, "grad_norm": 0.14373904466629028, "learning_rate": 1.8104792952915672e-06, "loss": 0.4909, "num_input_tokens_seen": 22235776, "step": 33975 }, { "epoch": 17.80922431865828, "grad_norm": 0.157755047082901, "learning_rate": 1.8062096142796192e-06, "loss": 0.4354, "num_input_tokens_seen": 22239136, "step": 33980 }, { "epoch": 17.811844863731658, "grad_norm": 0.3329833447933197, "learning_rate": 1.8019447851440163e-06, "loss": 0.415, "num_input_tokens_seen": 22243872, "step": 33985 }, { "epoch": 17.81446540880503, "grad_norm": 0.26848700642585754, "learning_rate": 1.7976848087769032e-06, "loss": 0.5447, "num_input_tokens_seen": 22247168, "step": 33990 }, { "epoch": 17.817085953878408, "grad_norm": 0.15242570638656616, "learning_rate": 1.79342968606942e-06, "loss": 0.5947, "num_input_tokens_seen": 22251648, "step": 33995 }, { "epoch": 17.81970649895178, "grad_norm": 0.1358623504638672, "learning_rate": 1.7891794179116933e-06, "loss": 0.5452, "num_input_tokens_seen": 22254240, "step": 34000 }, { "epoch": 17.822327044025158, "grad_norm": 0.13589583337306976, "learning_rate": 1.7849340051928316e-06, "loss": 0.6048, "num_input_tokens_seen": 22258176, "step": 34005 }, { "epoch": 17.82494758909853, "grad_norm": 0.14950957894325256, "learning_rate": 1.780693448800927e-06, "loss": 0.4837, "num_input_tokens_seen": 22261120, "step": 34010 }, { "epoch": 17.82756813417191, "grad_norm": 0.15030470490455627, "learning_rate": 1.7764577496230478e-06, "loss": 0.4827, "num_input_tokens_seen": 22264480, "step": 34015 }, { "epoch": 17.830188679245282, "grad_norm": 0.17151960730552673, "learning_rate": 1.7722269085452602e-06, "loss": 0.365, "num_input_tokens_seen": 22267392, "step": 34020 }, { "epoch": 17.83280922431866, "grad_norm": 0.16639253497123718, "learning_rate": 1.7680009264526038e-06, "loss": 0.5475, "num_input_tokens_seen": 22270336, "step": 34025 }, { "epoch": 17.835429769392032, "grad_norm": 0.21091869473457336, "learning_rate": 1.7637798042291126e-06, "loss": 0.5532, "num_input_tokens_seen": 22273472, "step": 34030 }, { "epoch": 17.83805031446541, "grad_norm": 0.08577179163694382, "learning_rate": 1.7595635427577944e-06, "loss": 0.336, "num_input_tokens_seen": 22276320, "step": 34035 }, { "epoch": 17.840670859538783, "grad_norm": 0.1299702227115631, "learning_rate": 1.7553521429206465e-06, "loss": 0.4216, "num_input_tokens_seen": 22280224, "step": 34040 }, { "epoch": 17.84329140461216, "grad_norm": 0.13113676011562347, "learning_rate": 1.7511456055986335e-06, "loss": 0.4773, "num_input_tokens_seen": 22283584, "step": 34045 }, { "epoch": 17.845911949685533, "grad_norm": 0.1879521906375885, "learning_rate": 1.746943931671724e-06, "loss": 0.4223, "num_input_tokens_seen": 22285952, "step": 34050 }, { "epoch": 17.84853249475891, "grad_norm": 0.21671420335769653, "learning_rate": 1.742747122018859e-06, "loss": 0.4611, "num_input_tokens_seen": 22288608, "step": 34055 }, { "epoch": 17.851153039832283, "grad_norm": 0.18629132211208344, "learning_rate": 1.738555177517967e-06, "loss": 0.4039, "num_input_tokens_seen": 22292416, "step": 34060 }, { "epoch": 17.85377358490566, "grad_norm": 0.16529591381549835, "learning_rate": 1.734368099045952e-06, "loss": 0.4328, "num_input_tokens_seen": 22295776, "step": 34065 }, { "epoch": 17.856394129979037, "grad_norm": 0.18109619617462158, "learning_rate": 1.730185887478697e-06, "loss": 0.6249, "num_input_tokens_seen": 22298272, "step": 34070 }, { "epoch": 17.85901467505241, "grad_norm": 0.21557116508483887, "learning_rate": 1.7260085436910794e-06, "loss": 0.3639, "num_input_tokens_seen": 22300960, "step": 34075 }, { "epoch": 17.861635220125788, "grad_norm": 0.2429838925600052, "learning_rate": 1.7218360685569506e-06, "loss": 0.4456, "num_input_tokens_seen": 22303904, "step": 34080 }, { "epoch": 17.86425576519916, "grad_norm": 0.16982010006904602, "learning_rate": 1.7176684629491513e-06, "loss": 0.3945, "num_input_tokens_seen": 22307584, "step": 34085 }, { "epoch": 17.866876310272538, "grad_norm": 0.14166691899299622, "learning_rate": 1.7135057277394868e-06, "loss": 0.4491, "num_input_tokens_seen": 22311008, "step": 34090 }, { "epoch": 17.86949685534591, "grad_norm": 0.10137303918600082, "learning_rate": 1.7093478637987548e-06, "loss": 0.3813, "num_input_tokens_seen": 22313888, "step": 34095 }, { "epoch": 17.872117400419288, "grad_norm": 0.1188005656003952, "learning_rate": 1.7051948719967375e-06, "loss": 0.4673, "num_input_tokens_seen": 22317376, "step": 34100 }, { "epoch": 17.87473794549266, "grad_norm": 0.14896894991397858, "learning_rate": 1.7010467532021901e-06, "loss": 0.5637, "num_input_tokens_seen": 22320320, "step": 34105 }, { "epoch": 17.87735849056604, "grad_norm": 0.1968148648738861, "learning_rate": 1.6969035082828544e-06, "loss": 0.4852, "num_input_tokens_seen": 22323136, "step": 34110 }, { "epoch": 17.879979035639412, "grad_norm": 0.1400105357170105, "learning_rate": 1.692765138105451e-06, "loss": 0.4781, "num_input_tokens_seen": 22326848, "step": 34115 }, { "epoch": 17.88259958071279, "grad_norm": 0.1170341819524765, "learning_rate": 1.6886316435356792e-06, "loss": 0.3897, "num_input_tokens_seen": 22329856, "step": 34120 }, { "epoch": 17.885220125786162, "grad_norm": 0.6191663146018982, "learning_rate": 1.6845030254382138e-06, "loss": 0.6477, "num_input_tokens_seen": 22332032, "step": 34125 }, { "epoch": 17.88784067085954, "grad_norm": 0.196343332529068, "learning_rate": 1.6803792846767197e-06, "loss": 0.5862, "num_input_tokens_seen": 22334656, "step": 34130 }, { "epoch": 17.890461215932913, "grad_norm": 0.11411429941654205, "learning_rate": 1.6762604221138316e-06, "loss": 0.4367, "num_input_tokens_seen": 22337632, "step": 34135 }, { "epoch": 17.89308176100629, "grad_norm": 0.18897497653961182, "learning_rate": 1.6721464386111774e-06, "loss": 0.5974, "num_input_tokens_seen": 22340288, "step": 34140 }, { "epoch": 17.895702306079663, "grad_norm": 0.1516292691230774, "learning_rate": 1.6680373350293516e-06, "loss": 0.4659, "num_input_tokens_seen": 22342848, "step": 34145 }, { "epoch": 17.89832285115304, "grad_norm": 0.1526830494403839, "learning_rate": 1.6639331122279223e-06, "loss": 0.5474, "num_input_tokens_seen": 22346400, "step": 34150 }, { "epoch": 17.900943396226417, "grad_norm": 0.13476763665676117, "learning_rate": 1.6598337710654583e-06, "loss": 0.4199, "num_input_tokens_seen": 22350592, "step": 34155 }, { "epoch": 17.90356394129979, "grad_norm": 0.1000182032585144, "learning_rate": 1.6557393123994874e-06, "loss": 0.403, "num_input_tokens_seen": 22354560, "step": 34160 }, { "epoch": 17.906184486373167, "grad_norm": 0.11585796624422073, "learning_rate": 1.651649737086533e-06, "loss": 0.51, "num_input_tokens_seen": 22357888, "step": 34165 }, { "epoch": 17.90880503144654, "grad_norm": 0.40289342403411865, "learning_rate": 1.6475650459820774e-06, "loss": 0.4998, "num_input_tokens_seen": 22360224, "step": 34170 }, { "epoch": 17.911425576519918, "grad_norm": 0.12006480991840363, "learning_rate": 1.643485239940598e-06, "loss": 0.4866, "num_input_tokens_seen": 22364480, "step": 34175 }, { "epoch": 17.91404612159329, "grad_norm": 0.2812703549861908, "learning_rate": 1.6394103198155375e-06, "loss": 0.4755, "num_input_tokens_seen": 22367808, "step": 34180 }, { "epoch": 17.916666666666668, "grad_norm": 0.1411048322916031, "learning_rate": 1.6353402864593276e-06, "loss": 0.4283, "num_input_tokens_seen": 22371328, "step": 34185 }, { "epoch": 17.91928721174004, "grad_norm": 0.12701334059238434, "learning_rate": 1.6312751407233706e-06, "loss": 0.5223, "num_input_tokens_seen": 22374080, "step": 34190 }, { "epoch": 17.921907756813418, "grad_norm": 0.15132685005664825, "learning_rate": 1.6272148834580559e-06, "loss": 0.523, "num_input_tokens_seen": 22377600, "step": 34195 }, { "epoch": 17.92452830188679, "grad_norm": 0.12737073004245758, "learning_rate": 1.6231595155127317e-06, "loss": 0.3935, "num_input_tokens_seen": 22380608, "step": 34200 }, { "epoch": 17.92714884696017, "grad_norm": 0.2778818905353546, "learning_rate": 1.6191090377357414e-06, "loss": 0.4936, "num_input_tokens_seen": 22383328, "step": 34205 }, { "epoch": 17.929769392033542, "grad_norm": 0.1508496254682541, "learning_rate": 1.6150634509743939e-06, "loss": 0.3784, "num_input_tokens_seen": 22387520, "step": 34210 }, { "epoch": 17.93238993710692, "grad_norm": 0.17114673554897308, "learning_rate": 1.611022756074984e-06, "loss": 0.4996, "num_input_tokens_seen": 22390368, "step": 34215 }, { "epoch": 17.935010482180292, "grad_norm": 0.16694939136505127, "learning_rate": 1.606986953882783e-06, "loss": 0.4804, "num_input_tokens_seen": 22392864, "step": 34220 }, { "epoch": 17.93763102725367, "grad_norm": 0.18523253500461578, "learning_rate": 1.6029560452420238e-06, "loss": 0.4118, "num_input_tokens_seen": 22395616, "step": 34225 }, { "epoch": 17.940251572327043, "grad_norm": 0.20336094498634338, "learning_rate": 1.5989300309959377e-06, "loss": 0.3984, "num_input_tokens_seen": 22399040, "step": 34230 }, { "epoch": 17.94287211740042, "grad_norm": 0.18180407583713531, "learning_rate": 1.5949089119867116e-06, "loss": 0.5355, "num_input_tokens_seen": 22402016, "step": 34235 }, { "epoch": 17.945492662473793, "grad_norm": 0.1449154168367386, "learning_rate": 1.5908926890555204e-06, "loss": 0.4887, "num_input_tokens_seen": 22405440, "step": 34240 }, { "epoch": 17.94811320754717, "grad_norm": 0.2911665439605713, "learning_rate": 1.5868813630425167e-06, "loss": 0.4549, "num_input_tokens_seen": 22408640, "step": 34245 }, { "epoch": 17.950733752620543, "grad_norm": 0.15249328315258026, "learning_rate": 1.5828749347868183e-06, "loss": 0.4958, "num_input_tokens_seen": 22411808, "step": 34250 }, { "epoch": 17.95335429769392, "grad_norm": 0.19467417895793915, "learning_rate": 1.5788734051265268e-06, "loss": 0.55, "num_input_tokens_seen": 22415168, "step": 34255 }, { "epoch": 17.955974842767297, "grad_norm": 0.1881045252084732, "learning_rate": 1.5748767748987231e-06, "loss": 0.4134, "num_input_tokens_seen": 22419264, "step": 34260 }, { "epoch": 17.95859538784067, "grad_norm": 0.17987947165966034, "learning_rate": 1.5708850449394436e-06, "loss": 0.4755, "num_input_tokens_seen": 22422688, "step": 34265 }, { "epoch": 17.961215932914047, "grad_norm": 0.18076317012310028, "learning_rate": 1.5668982160837176e-06, "loss": 0.4821, "num_input_tokens_seen": 22425600, "step": 34270 }, { "epoch": 17.96383647798742, "grad_norm": 0.21569330990314484, "learning_rate": 1.562916289165553e-06, "loss": 0.3927, "num_input_tokens_seen": 22430496, "step": 34275 }, { "epoch": 17.966457023060798, "grad_norm": 0.13743112981319427, "learning_rate": 1.5589392650179114e-06, "loss": 0.4514, "num_input_tokens_seen": 22433504, "step": 34280 }, { "epoch": 17.96907756813417, "grad_norm": 0.12305878102779388, "learning_rate": 1.5549671444727437e-06, "loss": 0.5766, "num_input_tokens_seen": 22436640, "step": 34285 }, { "epoch": 17.971698113207548, "grad_norm": 0.3122941553592682, "learning_rate": 1.55099992836098e-06, "loss": 0.3908, "num_input_tokens_seen": 22439648, "step": 34290 }, { "epoch": 17.97431865828092, "grad_norm": 0.13969430327415466, "learning_rate": 1.5470376175125061e-06, "loss": 0.493, "num_input_tokens_seen": 22443360, "step": 34295 }, { "epoch": 17.9769392033543, "grad_norm": 0.22011211514472961, "learning_rate": 1.5430802127562038e-06, "loss": 0.5082, "num_input_tokens_seen": 22445888, "step": 34300 }, { "epoch": 17.979559748427672, "grad_norm": 0.27814629673957825, "learning_rate": 1.5391277149199023e-06, "loss": 0.4229, "num_input_tokens_seen": 22449536, "step": 34305 }, { "epoch": 17.98218029350105, "grad_norm": 0.2950837314128876, "learning_rate": 1.5351801248304292e-06, "loss": 0.5337, "num_input_tokens_seen": 22453504, "step": 34310 }, { "epoch": 17.984800838574422, "grad_norm": 0.23942215740680695, "learning_rate": 1.5312374433135767e-06, "loss": 0.6384, "num_input_tokens_seen": 22457312, "step": 34315 }, { "epoch": 17.9874213836478, "grad_norm": 0.21826203167438507, "learning_rate": 1.5272996711941018e-06, "loss": 0.4564, "num_input_tokens_seen": 22459840, "step": 34320 }, { "epoch": 17.990041928721173, "grad_norm": 0.1341613233089447, "learning_rate": 1.5233668092957482e-06, "loss": 0.4475, "num_input_tokens_seen": 22462528, "step": 34325 }, { "epoch": 17.99266247379455, "grad_norm": 0.19809505343437195, "learning_rate": 1.5194388584412217e-06, "loss": 0.491, "num_input_tokens_seen": 22465696, "step": 34330 }, { "epoch": 17.995283018867923, "grad_norm": 0.14684215188026428, "learning_rate": 1.5155158194522013e-06, "loss": 0.588, "num_input_tokens_seen": 22468448, "step": 34335 }, { "epoch": 17.9979035639413, "grad_norm": 0.30247294902801514, "learning_rate": 1.5115976931493498e-06, "loss": 0.4993, "num_input_tokens_seen": 22471168, "step": 34340 }, { "epoch": 18.0, "eval_loss": 0.46048542857170105, "eval_runtime": 9.9687, "eval_samples_per_second": 85.067, "eval_steps_per_second": 21.267, "num_input_tokens_seen": 22472696, "step": 34344 }, { "epoch": 18.000524109014677, "grad_norm": 0.2072472721338272, "learning_rate": 1.5076844803522922e-06, "loss": 0.5703, "num_input_tokens_seen": 22473112, "step": 34345 }, { "epoch": 18.00314465408805, "grad_norm": 0.12914173305034637, "learning_rate": 1.5037761818796236e-06, "loss": 0.5742, "num_input_tokens_seen": 22475928, "step": 34350 }, { "epoch": 18.005765199161427, "grad_norm": 0.1668587625026703, "learning_rate": 1.4998727985489235e-06, "loss": 0.4662, "num_input_tokens_seen": 22478424, "step": 34355 }, { "epoch": 18.0083857442348, "grad_norm": 0.09338057041168213, "learning_rate": 1.4959743311767249e-06, "loss": 0.418, "num_input_tokens_seen": 22481496, "step": 34360 }, { "epoch": 18.011006289308177, "grad_norm": 0.1474262923002243, "learning_rate": 1.4920807805785502e-06, "loss": 0.5656, "num_input_tokens_seen": 22488472, "step": 34365 }, { "epoch": 18.01362683438155, "grad_norm": 0.22766278684139252, "learning_rate": 1.4881921475688815e-06, "loss": 0.6708, "num_input_tokens_seen": 22491384, "step": 34370 }, { "epoch": 18.016247379454928, "grad_norm": 0.11454703658819199, "learning_rate": 1.4843084329611844e-06, "loss": 0.3975, "num_input_tokens_seen": 22494104, "step": 34375 }, { "epoch": 18.0188679245283, "grad_norm": 0.11611614376306534, "learning_rate": 1.480429637567879e-06, "loss": 0.5789, "num_input_tokens_seen": 22497656, "step": 34380 }, { "epoch": 18.021488469601678, "grad_norm": 0.16449935734272003, "learning_rate": 1.4765557622003684e-06, "loss": 0.4682, "num_input_tokens_seen": 22500920, "step": 34385 }, { "epoch": 18.02410901467505, "grad_norm": 0.3218139111995697, "learning_rate": 1.4726868076690186e-06, "loss": 0.4058, "num_input_tokens_seen": 22504408, "step": 34390 }, { "epoch": 18.02672955974843, "grad_norm": 0.15064780414104462, "learning_rate": 1.4688227747831767e-06, "loss": 0.4418, "num_input_tokens_seen": 22506904, "step": 34395 }, { "epoch": 18.029350104821802, "grad_norm": 0.1838901937007904, "learning_rate": 1.4649636643511573e-06, "loss": 0.5627, "num_input_tokens_seen": 22510200, "step": 34400 }, { "epoch": 18.03197064989518, "grad_norm": 0.13326404988765717, "learning_rate": 1.4611094771802364e-06, "loss": 0.3859, "num_input_tokens_seen": 22513112, "step": 34405 }, { "epoch": 18.034591194968552, "grad_norm": 0.18153484165668488, "learning_rate": 1.457260214076664e-06, "loss": 0.474, "num_input_tokens_seen": 22516824, "step": 34410 }, { "epoch": 18.03721174004193, "grad_norm": 0.22373613715171814, "learning_rate": 1.4534158758456623e-06, "loss": 0.4386, "num_input_tokens_seen": 22520184, "step": 34415 }, { "epoch": 18.039832285115303, "grad_norm": 0.15143199265003204, "learning_rate": 1.4495764632914271e-06, "loss": 0.4256, "num_input_tokens_seen": 22523000, "step": 34420 }, { "epoch": 18.04245283018868, "grad_norm": 0.16034886240959167, "learning_rate": 1.4457419772171215e-06, "loss": 0.3795, "num_input_tokens_seen": 22525400, "step": 34425 }, { "epoch": 18.045073375262053, "grad_norm": 0.2506571412086487, "learning_rate": 1.4419124184248733e-06, "loss": 0.4892, "num_input_tokens_seen": 22528344, "step": 34430 }, { "epoch": 18.04769392033543, "grad_norm": 0.1849154531955719, "learning_rate": 1.4380877877157834e-06, "loss": 0.5077, "num_input_tokens_seen": 22530776, "step": 34435 }, { "epoch": 18.050314465408807, "grad_norm": 0.2567785084247589, "learning_rate": 1.4342680858899143e-06, "loss": 0.4478, "num_input_tokens_seen": 22534584, "step": 34440 }, { "epoch": 18.05293501048218, "grad_norm": 0.11052153259515762, "learning_rate": 1.4304533137463134e-06, "loss": 0.4641, "num_input_tokens_seen": 22539192, "step": 34445 }, { "epoch": 18.055555555555557, "grad_norm": 0.10505472123622894, "learning_rate": 1.4266434720829808e-06, "loss": 0.381, "num_input_tokens_seen": 22541624, "step": 34450 }, { "epoch": 18.05817610062893, "grad_norm": 0.12941524386405945, "learning_rate": 1.4228385616968987e-06, "loss": 0.4605, "num_input_tokens_seen": 22544888, "step": 34455 }, { "epoch": 18.060796645702307, "grad_norm": 0.18556518852710724, "learning_rate": 1.419038583384008e-06, "loss": 0.444, "num_input_tokens_seen": 22547704, "step": 34460 }, { "epoch": 18.06341719077568, "grad_norm": 0.14927975833415985, "learning_rate": 1.4152435379392148e-06, "loss": 0.4033, "num_input_tokens_seen": 22551352, "step": 34465 }, { "epoch": 18.066037735849058, "grad_norm": 0.1903902292251587, "learning_rate": 1.411453426156406e-06, "loss": 0.5284, "num_input_tokens_seen": 22554520, "step": 34470 }, { "epoch": 18.06865828092243, "grad_norm": 0.14909712970256805, "learning_rate": 1.4076682488284282e-06, "loss": 0.4672, "num_input_tokens_seen": 22557528, "step": 34475 }, { "epoch": 18.071278825995808, "grad_norm": 0.24003584682941437, "learning_rate": 1.4038880067470978e-06, "loss": 0.3988, "num_input_tokens_seen": 22560632, "step": 34480 }, { "epoch": 18.07389937106918, "grad_norm": 0.24249865114688873, "learning_rate": 1.400112700703196e-06, "loss": 0.5013, "num_input_tokens_seen": 22563896, "step": 34485 }, { "epoch": 18.07651991614256, "grad_norm": 0.17551782727241516, "learning_rate": 1.3963423314864805e-06, "loss": 0.4952, "num_input_tokens_seen": 22566136, "step": 34490 }, { "epoch": 18.079140461215932, "grad_norm": 0.1996762752532959, "learning_rate": 1.3925768998856587e-06, "loss": 0.3911, "num_input_tokens_seen": 22568824, "step": 34495 }, { "epoch": 18.08176100628931, "grad_norm": 0.15998566150665283, "learning_rate": 1.388816406688423e-06, "loss": 0.6082, "num_input_tokens_seen": 22573304, "step": 34500 }, { "epoch": 18.084381551362682, "grad_norm": 0.18063920736312866, "learning_rate": 1.3850608526814246e-06, "loss": 0.7192, "num_input_tokens_seen": 22577304, "step": 34505 }, { "epoch": 18.08700209643606, "grad_norm": 0.19433344900608063, "learning_rate": 1.3813102386502853e-06, "loss": 0.5476, "num_input_tokens_seen": 22580568, "step": 34510 }, { "epoch": 18.089622641509433, "grad_norm": 0.14805136620998383, "learning_rate": 1.3775645653795881e-06, "loss": 0.5404, "num_input_tokens_seen": 22583416, "step": 34515 }, { "epoch": 18.09224318658281, "grad_norm": 0.18727295100688934, "learning_rate": 1.373823833652882e-06, "loss": 0.3679, "num_input_tokens_seen": 22587544, "step": 34520 }, { "epoch": 18.094863731656183, "grad_norm": 0.12470871210098267, "learning_rate": 1.3700880442526876e-06, "loss": 0.5682, "num_input_tokens_seen": 22590712, "step": 34525 }, { "epoch": 18.09748427672956, "grad_norm": 0.1506742686033249, "learning_rate": 1.3663571979604884e-06, "loss": 0.4529, "num_input_tokens_seen": 22593528, "step": 34530 }, { "epoch": 18.100104821802937, "grad_norm": 0.16138553619384766, "learning_rate": 1.3626312955567406e-06, "loss": 0.5727, "num_input_tokens_seen": 22597496, "step": 34535 }, { "epoch": 18.10272536687631, "grad_norm": 0.15754441916942596, "learning_rate": 1.358910337820854e-06, "loss": 0.5279, "num_input_tokens_seen": 22599960, "step": 34540 }, { "epoch": 18.105345911949687, "grad_norm": 0.1517811417579651, "learning_rate": 1.3551943255312172e-06, "loss": 0.6697, "num_input_tokens_seen": 22602872, "step": 34545 }, { "epoch": 18.10796645702306, "grad_norm": 0.22210127115249634, "learning_rate": 1.3514832594651694e-06, "loss": 0.5304, "num_input_tokens_seen": 22605592, "step": 34550 }, { "epoch": 18.110587002096437, "grad_norm": 0.15000088512897491, "learning_rate": 1.347777140399026e-06, "loss": 0.6491, "num_input_tokens_seen": 22609112, "step": 34555 }, { "epoch": 18.11320754716981, "grad_norm": 0.1329689621925354, "learning_rate": 1.344075969108069e-06, "loss": 0.5176, "num_input_tokens_seen": 22611736, "step": 34560 }, { "epoch": 18.115828092243188, "grad_norm": 0.18162012100219727, "learning_rate": 1.3403797463665352e-06, "loss": 0.3969, "num_input_tokens_seen": 22615896, "step": 34565 }, { "epoch": 18.11844863731656, "grad_norm": 0.14835740625858307, "learning_rate": 1.3366884729476365e-06, "loss": 0.3913, "num_input_tokens_seen": 22618296, "step": 34570 }, { "epoch": 18.121069182389938, "grad_norm": 0.22938182950019836, "learning_rate": 1.3330021496235467e-06, "loss": 0.5432, "num_input_tokens_seen": 22621976, "step": 34575 }, { "epoch": 18.12368972746331, "grad_norm": 0.1634887307882309, "learning_rate": 1.3293207771653965e-06, "loss": 0.3696, "num_input_tokens_seen": 22625720, "step": 34580 }, { "epoch": 18.12631027253669, "grad_norm": 0.1257457435131073, "learning_rate": 1.3256443563432891e-06, "loss": 0.4727, "num_input_tokens_seen": 22629496, "step": 34585 }, { "epoch": 18.128930817610062, "grad_norm": 0.21787552535533905, "learning_rate": 1.3219728879262956e-06, "loss": 0.5472, "num_input_tokens_seen": 22632920, "step": 34590 }, { "epoch": 18.13155136268344, "grad_norm": 0.09897024184465408, "learning_rate": 1.3183063726824375e-06, "loss": 0.4062, "num_input_tokens_seen": 22636536, "step": 34595 }, { "epoch": 18.134171907756812, "grad_norm": 0.1625937968492508, "learning_rate": 1.3146448113787151e-06, "loss": 0.6769, "num_input_tokens_seen": 22639800, "step": 34600 }, { "epoch": 18.13679245283019, "grad_norm": 0.1644824594259262, "learning_rate": 1.3109882047810773e-06, "loss": 0.5722, "num_input_tokens_seen": 22642520, "step": 34605 }, { "epoch": 18.139412997903563, "grad_norm": 0.14785516262054443, "learning_rate": 1.307336553654448e-06, "loss": 0.4135, "num_input_tokens_seen": 22645784, "step": 34610 }, { "epoch": 18.14203354297694, "grad_norm": 0.265320748090744, "learning_rate": 1.303689858762716e-06, "loss": 0.4314, "num_input_tokens_seen": 22648664, "step": 34615 }, { "epoch": 18.144654088050313, "grad_norm": 0.1926974058151245, "learning_rate": 1.3000481208687188e-06, "loss": 0.5569, "num_input_tokens_seen": 22651832, "step": 34620 }, { "epoch": 18.14727463312369, "grad_norm": 0.10718917846679688, "learning_rate": 1.2964113407342743e-06, "loss": 0.5003, "num_input_tokens_seen": 22654808, "step": 34625 }, { "epoch": 18.149895178197063, "grad_norm": 0.12652422487735748, "learning_rate": 1.292779519120152e-06, "loss": 0.4888, "num_input_tokens_seen": 22657624, "step": 34630 }, { "epoch": 18.15251572327044, "grad_norm": 0.13877490162849426, "learning_rate": 1.2891526567860856e-06, "loss": 0.448, "num_input_tokens_seen": 22659960, "step": 34635 }, { "epoch": 18.155136268343817, "grad_norm": 0.20563609898090363, "learning_rate": 1.2855307544907825e-06, "loss": 0.5013, "num_input_tokens_seen": 22665080, "step": 34640 }, { "epoch": 18.15775681341719, "grad_norm": 0.20290158689022064, "learning_rate": 1.281913812991889e-06, "loss": 0.6503, "num_input_tokens_seen": 22667544, "step": 34645 }, { "epoch": 18.160377358490567, "grad_norm": 0.17641036212444305, "learning_rate": 1.2783018330460389e-06, "loss": 0.4207, "num_input_tokens_seen": 22671032, "step": 34650 }, { "epoch": 18.16299790356394, "grad_norm": 0.11822318285703659, "learning_rate": 1.2746948154088111e-06, "loss": 0.3905, "num_input_tokens_seen": 22675416, "step": 34655 }, { "epoch": 18.165618448637318, "grad_norm": 0.15399996936321259, "learning_rate": 1.2710927608347605e-06, "loss": 0.5044, "num_input_tokens_seen": 22678008, "step": 34660 }, { "epoch": 18.16823899371069, "grad_norm": 0.24411316215991974, "learning_rate": 1.2674956700773872e-06, "loss": 0.3958, "num_input_tokens_seen": 22680408, "step": 34665 }, { "epoch": 18.170859538784068, "grad_norm": 0.16557267308235168, "learning_rate": 1.2639035438891666e-06, "loss": 0.5063, "num_input_tokens_seen": 22685176, "step": 34670 }, { "epoch": 18.17348008385744, "grad_norm": 0.24923883378505707, "learning_rate": 1.260316383021526e-06, "loss": 0.3254, "num_input_tokens_seen": 22687960, "step": 34675 }, { "epoch": 18.17610062893082, "grad_norm": 0.14272989332675934, "learning_rate": 1.2567341882248645e-06, "loss": 0.3163, "num_input_tokens_seen": 22690840, "step": 34680 }, { "epoch": 18.178721174004192, "grad_norm": 0.11995474994182587, "learning_rate": 1.2531569602485305e-06, "loss": 0.4252, "num_input_tokens_seen": 22694424, "step": 34685 }, { "epoch": 18.18134171907757, "grad_norm": 0.13971605896949768, "learning_rate": 1.2495846998408445e-06, "loss": 0.4816, "num_input_tokens_seen": 22697976, "step": 34690 }, { "epoch": 18.183962264150942, "grad_norm": 0.2487202137708664, "learning_rate": 1.246017407749084e-06, "loss": 0.5506, "num_input_tokens_seen": 22701176, "step": 34695 }, { "epoch": 18.18658280922432, "grad_norm": 0.174480602145195, "learning_rate": 1.2424550847194739e-06, "loss": 0.4753, "num_input_tokens_seen": 22703864, "step": 34700 }, { "epoch": 18.189203354297693, "grad_norm": 0.21306580305099487, "learning_rate": 1.2388977314972238e-06, "loss": 0.3906, "num_input_tokens_seen": 22706616, "step": 34705 }, { "epoch": 18.19182389937107, "grad_norm": 0.2008492350578308, "learning_rate": 1.235345348826486e-06, "loss": 0.426, "num_input_tokens_seen": 22709272, "step": 34710 }, { "epoch": 18.194444444444443, "grad_norm": 0.15458689630031586, "learning_rate": 1.2317979374503825e-06, "loss": 0.4043, "num_input_tokens_seen": 22712568, "step": 34715 }, { "epoch": 18.19706498951782, "grad_norm": 0.20549015700817108, "learning_rate": 1.228255498110989e-06, "loss": 0.4596, "num_input_tokens_seen": 22715576, "step": 34720 }, { "epoch": 18.199685534591197, "grad_norm": 0.26814544200897217, "learning_rate": 1.224718031549338e-06, "loss": 0.5444, "num_input_tokens_seen": 22718584, "step": 34725 }, { "epoch": 18.20230607966457, "grad_norm": 0.15149268507957458, "learning_rate": 1.2211855385054344e-06, "loss": 0.421, "num_input_tokens_seen": 22722456, "step": 34730 }, { "epoch": 18.204926624737947, "grad_norm": 0.11509960144758224, "learning_rate": 1.217658019718232e-06, "loss": 0.5408, "num_input_tokens_seen": 22726392, "step": 34735 }, { "epoch": 18.20754716981132, "grad_norm": 0.14337754249572754, "learning_rate": 1.2141354759256484e-06, "loss": 0.4804, "num_input_tokens_seen": 22729560, "step": 34740 }, { "epoch": 18.210167714884697, "grad_norm": 0.14354068040847778, "learning_rate": 1.2106179078645641e-06, "loss": 0.4789, "num_input_tokens_seen": 22733496, "step": 34745 }, { "epoch": 18.21278825995807, "grad_norm": 0.2645416557788849, "learning_rate": 1.2071053162708096e-06, "loss": 0.3589, "num_input_tokens_seen": 22736120, "step": 34750 }, { "epoch": 18.215408805031448, "grad_norm": 0.2011193186044693, "learning_rate": 1.203597701879175e-06, "loss": 0.4967, "num_input_tokens_seen": 22738872, "step": 34755 }, { "epoch": 18.21802935010482, "grad_norm": 0.2982608377933502, "learning_rate": 1.2000950654234177e-06, "loss": 0.5513, "num_input_tokens_seen": 22741336, "step": 34760 }, { "epoch": 18.220649895178198, "grad_norm": 0.1451343595981598, "learning_rate": 1.1965974076362491e-06, "loss": 0.4503, "num_input_tokens_seen": 22744632, "step": 34765 }, { "epoch": 18.22327044025157, "grad_norm": 0.13702374696731567, "learning_rate": 1.193104729249342e-06, "loss": 0.5739, "num_input_tokens_seen": 22749048, "step": 34770 }, { "epoch": 18.22589098532495, "grad_norm": 0.1449889838695526, "learning_rate": 1.189617030993323e-06, "loss": 0.4162, "num_input_tokens_seen": 22753048, "step": 34775 }, { "epoch": 18.228511530398322, "grad_norm": 0.19557203352451324, "learning_rate": 1.1861343135977753e-06, "loss": 0.5088, "num_input_tokens_seen": 22756440, "step": 34780 }, { "epoch": 18.2311320754717, "grad_norm": 0.2759457528591156, "learning_rate": 1.182656577791244e-06, "loss": 0.5221, "num_input_tokens_seen": 22759096, "step": 34785 }, { "epoch": 18.233752620545072, "grad_norm": 0.1154932752251625, "learning_rate": 1.1791838243012359e-06, "loss": 0.3899, "num_input_tokens_seen": 22762424, "step": 34790 }, { "epoch": 18.23637316561845, "grad_norm": 0.17511940002441406, "learning_rate": 1.1757160538542117e-06, "loss": 0.4335, "num_input_tokens_seen": 22764600, "step": 34795 }, { "epoch": 18.238993710691823, "grad_norm": 0.14695152640342712, "learning_rate": 1.1722532671755882e-06, "loss": 0.5049, "num_input_tokens_seen": 22767672, "step": 34800 }, { "epoch": 18.2416142557652, "grad_norm": 0.16660796105861664, "learning_rate": 1.168795464989736e-06, "loss": 0.5133, "num_input_tokens_seen": 22770776, "step": 34805 }, { "epoch": 18.244234800838573, "grad_norm": 0.12618492543697357, "learning_rate": 1.165342648019993e-06, "loss": 0.3084, "num_input_tokens_seen": 22774968, "step": 34810 }, { "epoch": 18.24685534591195, "grad_norm": 0.1617441624403, "learning_rate": 1.1618948169886457e-06, "loss": 0.5022, "num_input_tokens_seen": 22778040, "step": 34815 }, { "epoch": 18.249475890985323, "grad_norm": 0.19815587997436523, "learning_rate": 1.1584519726169446e-06, "loss": 0.634, "num_input_tokens_seen": 22780760, "step": 34820 }, { "epoch": 18.2520964360587, "grad_norm": 0.17586718499660492, "learning_rate": 1.1550141156250942e-06, "loss": 0.4559, "num_input_tokens_seen": 22783480, "step": 34825 }, { "epoch": 18.254716981132077, "grad_norm": 0.14622487127780914, "learning_rate": 1.1515812467322556e-06, "loss": 0.5191, "num_input_tokens_seen": 22787704, "step": 34830 }, { "epoch": 18.25733752620545, "grad_norm": 0.12504127621650696, "learning_rate": 1.1481533666565375e-06, "loss": 0.4028, "num_input_tokens_seen": 22791384, "step": 34835 }, { "epoch": 18.259958071278827, "grad_norm": 0.19689393043518066, "learning_rate": 1.144730476115019e-06, "loss": 0.5048, "num_input_tokens_seen": 22795000, "step": 34840 }, { "epoch": 18.2625786163522, "grad_norm": 0.1815737783908844, "learning_rate": 1.1413125758237304e-06, "loss": 0.4431, "num_input_tokens_seen": 22798072, "step": 34845 }, { "epoch": 18.265199161425578, "grad_norm": 0.20057903230190277, "learning_rate": 1.1378996664976577e-06, "loss": 0.4751, "num_input_tokens_seen": 22801240, "step": 34850 }, { "epoch": 18.26781970649895, "grad_norm": 0.1434997171163559, "learning_rate": 1.134491748850741e-06, "loss": 0.4635, "num_input_tokens_seen": 22804056, "step": 34855 }, { "epoch": 18.270440251572328, "grad_norm": 0.19844795763492584, "learning_rate": 1.13108882359588e-06, "loss": 0.4787, "num_input_tokens_seen": 22807480, "step": 34860 }, { "epoch": 18.2730607966457, "grad_norm": 0.10371407121419907, "learning_rate": 1.127690891444924e-06, "loss": 0.5461, "num_input_tokens_seen": 22810648, "step": 34865 }, { "epoch": 18.27568134171908, "grad_norm": 0.18713130056858063, "learning_rate": 1.1242979531086829e-06, "loss": 0.4985, "num_input_tokens_seen": 22814520, "step": 34870 }, { "epoch": 18.278301886792452, "grad_norm": 0.21897423267364502, "learning_rate": 1.1209100092969244e-06, "loss": 0.5387, "num_input_tokens_seen": 22818264, "step": 34875 }, { "epoch": 18.28092243186583, "grad_norm": 0.19706685841083527, "learning_rate": 1.1175270607183625e-06, "loss": 0.4864, "num_input_tokens_seen": 22821208, "step": 34880 }, { "epoch": 18.283542976939202, "grad_norm": 0.11211360991001129, "learning_rate": 1.114149108080678e-06, "loss": 0.4579, "num_input_tokens_seen": 22824696, "step": 34885 }, { "epoch": 18.28616352201258, "grad_norm": 0.13142459094524384, "learning_rate": 1.1107761520904925e-06, "loss": 0.4006, "num_input_tokens_seen": 22828024, "step": 34890 }, { "epoch": 18.288784067085953, "grad_norm": 0.23197007179260254, "learning_rate": 1.1074081934533908e-06, "loss": 0.4759, "num_input_tokens_seen": 22830456, "step": 34895 }, { "epoch": 18.29140461215933, "grad_norm": 0.16867458820343018, "learning_rate": 1.1040452328739153e-06, "loss": 0.414, "num_input_tokens_seen": 22833592, "step": 34900 }, { "epoch": 18.294025157232703, "grad_norm": 0.21563293039798737, "learning_rate": 1.1006872710555615e-06, "loss": 0.5425, "num_input_tokens_seen": 22836696, "step": 34905 }, { "epoch": 18.29664570230608, "grad_norm": 0.185817688703537, "learning_rate": 1.0973343087007703e-06, "loss": 0.4145, "num_input_tokens_seen": 22839960, "step": 34910 }, { "epoch": 18.299266247379457, "grad_norm": 0.25249460339546204, "learning_rate": 1.0939863465109473e-06, "loss": 0.5019, "num_input_tokens_seen": 22843000, "step": 34915 }, { "epoch": 18.30188679245283, "grad_norm": 0.13312792778015137, "learning_rate": 1.0906433851864433e-06, "loss": 0.3914, "num_input_tokens_seen": 22847032, "step": 34920 }, { "epoch": 18.304507337526207, "grad_norm": 0.14662539958953857, "learning_rate": 1.0873054254265713e-06, "loss": 0.4471, "num_input_tokens_seen": 22849976, "step": 34925 }, { "epoch": 18.30712788259958, "grad_norm": 0.6452591419219971, "learning_rate": 1.0839724679295976e-06, "loss": 0.4616, "num_input_tokens_seen": 22852888, "step": 34930 }, { "epoch": 18.309748427672957, "grad_norm": 0.13242465257644653, "learning_rate": 1.0806445133927311e-06, "loss": 0.4377, "num_input_tokens_seen": 22855672, "step": 34935 }, { "epoch": 18.31236897274633, "grad_norm": 0.12864463031291962, "learning_rate": 1.0773215625121457e-06, "loss": 0.5032, "num_input_tokens_seen": 22859032, "step": 34940 }, { "epoch": 18.314989517819708, "grad_norm": 0.20386798679828644, "learning_rate": 1.0740036159829715e-06, "loss": 0.4008, "num_input_tokens_seen": 22863160, "step": 34945 }, { "epoch": 18.31761006289308, "grad_norm": 0.16106030344963074, "learning_rate": 1.0706906744992724e-06, "loss": 0.4707, "num_input_tokens_seen": 22866104, "step": 34950 }, { "epoch": 18.320230607966458, "grad_norm": 0.18171583116054535, "learning_rate": 1.0673827387540857e-06, "loss": 0.4917, "num_input_tokens_seen": 22869112, "step": 34955 }, { "epoch": 18.32285115303983, "grad_norm": 0.2647184431552887, "learning_rate": 1.0640798094393916e-06, "loss": 0.4129, "num_input_tokens_seen": 22872184, "step": 34960 }, { "epoch": 18.32547169811321, "grad_norm": 0.15274524688720703, "learning_rate": 1.060781887246126e-06, "loss": 0.4316, "num_input_tokens_seen": 22876376, "step": 34965 }, { "epoch": 18.328092243186582, "grad_norm": 0.1809810847043991, "learning_rate": 1.057488972864179e-06, "loss": 0.4191, "num_input_tokens_seen": 22879640, "step": 34970 }, { "epoch": 18.33071278825996, "grad_norm": 0.1939687877893448, "learning_rate": 1.0542010669823855e-06, "loss": 0.5422, "num_input_tokens_seen": 22883832, "step": 34975 }, { "epoch": 18.333333333333332, "grad_norm": 0.20737959444522858, "learning_rate": 1.0509181702885395e-06, "loss": 0.4435, "num_input_tokens_seen": 22886328, "step": 34980 }, { "epoch": 18.33595387840671, "grad_norm": 0.16872413456439972, "learning_rate": 1.047640283469392e-06, "loss": 0.5137, "num_input_tokens_seen": 22889080, "step": 34985 }, { "epoch": 18.338574423480082, "grad_norm": 0.16307006776332855, "learning_rate": 1.0443674072106308e-06, "loss": 0.4944, "num_input_tokens_seen": 22891576, "step": 34990 }, { "epoch": 18.34119496855346, "grad_norm": 0.17743471264839172, "learning_rate": 1.041099542196905e-06, "loss": 0.6822, "num_input_tokens_seen": 22894808, "step": 34995 }, { "epoch": 18.343815513626833, "grad_norm": 0.3014918267726898, "learning_rate": 1.0378366891118234e-06, "loss": 0.4714, "num_input_tokens_seen": 22897912, "step": 35000 }, { "epoch": 18.34643605870021, "grad_norm": 0.33899009227752686, "learning_rate": 1.0345788486379265e-06, "loss": 0.525, "num_input_tokens_seen": 22900824, "step": 35005 }, { "epoch": 18.349056603773583, "grad_norm": 0.14864106476306915, "learning_rate": 1.0313260214567272e-06, "loss": 0.5599, "num_input_tokens_seen": 22903864, "step": 35010 }, { "epoch": 18.35167714884696, "grad_norm": 0.24362225830554962, "learning_rate": 1.0280782082486729e-06, "loss": 0.3804, "num_input_tokens_seen": 22906744, "step": 35015 }, { "epoch": 18.354297693920337, "grad_norm": 0.12208877503871918, "learning_rate": 1.02483540969317e-06, "loss": 0.4778, "num_input_tokens_seen": 22910904, "step": 35020 }, { "epoch": 18.35691823899371, "grad_norm": 0.16530275344848633, "learning_rate": 1.0215976264685763e-06, "loss": 0.4801, "num_input_tokens_seen": 22914072, "step": 35025 }, { "epoch": 18.359538784067087, "grad_norm": 0.11282419413328171, "learning_rate": 1.0183648592522021e-06, "loss": 0.4311, "num_input_tokens_seen": 22916856, "step": 35030 }, { "epoch": 18.36215932914046, "grad_norm": 0.2768338620662689, "learning_rate": 1.0151371087203015e-06, "loss": 0.4292, "num_input_tokens_seen": 22922008, "step": 35035 }, { "epoch": 18.364779874213838, "grad_norm": 0.18862909078598022, "learning_rate": 1.0119143755480815e-06, "loss": 0.4914, "num_input_tokens_seen": 22924760, "step": 35040 }, { "epoch": 18.36740041928721, "grad_norm": 0.1601143330335617, "learning_rate": 1.0086966604097026e-06, "loss": 0.5315, "num_input_tokens_seen": 22927416, "step": 35045 }, { "epoch": 18.370020964360588, "grad_norm": 0.1382061094045639, "learning_rate": 1.0054839639782766e-06, "loss": 0.5386, "num_input_tokens_seen": 22930200, "step": 35050 }, { "epoch": 18.37264150943396, "grad_norm": 0.23389185965061188, "learning_rate": 1.00227628692586e-06, "loss": 0.523, "num_input_tokens_seen": 22933592, "step": 35055 }, { "epoch": 18.37526205450734, "grad_norm": 0.16364283859729767, "learning_rate": 9.990736299234692e-07, "loss": 0.4048, "num_input_tokens_seen": 22936312, "step": 35060 }, { "epoch": 18.377882599580712, "grad_norm": 0.1355990767478943, "learning_rate": 9.958759936410573e-07, "loss": 0.4799, "num_input_tokens_seen": 22939352, "step": 35065 }, { "epoch": 18.38050314465409, "grad_norm": 0.1837499588727951, "learning_rate": 9.926833787475304e-07, "loss": 0.5763, "num_input_tokens_seen": 22942552, "step": 35070 }, { "epoch": 18.383123689727462, "grad_norm": 0.16990086436271667, "learning_rate": 9.89495785910749e-07, "loss": 0.4246, "num_input_tokens_seen": 22945592, "step": 35075 }, { "epoch": 18.38574423480084, "grad_norm": 0.19585321843624115, "learning_rate": 9.86313215797524e-07, "loss": 0.4357, "num_input_tokens_seen": 22948280, "step": 35080 }, { "epoch": 18.388364779874212, "grad_norm": 0.1719750165939331, "learning_rate": 9.831356690736149e-07, "loss": 0.513, "num_input_tokens_seen": 22951512, "step": 35085 }, { "epoch": 18.39098532494759, "grad_norm": 0.13258694112300873, "learning_rate": 9.799631464037256e-07, "loss": 0.5619, "num_input_tokens_seen": 22954904, "step": 35090 }, { "epoch": 18.393605870020963, "grad_norm": 0.1364264339208603, "learning_rate": 9.76795648451509e-07, "loss": 0.4168, "num_input_tokens_seen": 22958136, "step": 35095 }, { "epoch": 18.39622641509434, "grad_norm": 0.31488558650016785, "learning_rate": 9.73633175879568e-07, "loss": 0.4215, "num_input_tokens_seen": 22961720, "step": 35100 }, { "epoch": 18.398846960167713, "grad_norm": 0.23805765807628632, "learning_rate": 9.704757293494627e-07, "loss": 0.4475, "num_input_tokens_seen": 22964856, "step": 35105 }, { "epoch": 18.40146750524109, "grad_norm": 0.18545715510845184, "learning_rate": 9.673233095216922e-07, "loss": 0.4311, "num_input_tokens_seen": 22967864, "step": 35110 }, { "epoch": 18.404088050314467, "grad_norm": 0.13583269715309143, "learning_rate": 9.641759170557068e-07, "loss": 0.4627, "num_input_tokens_seen": 22971704, "step": 35115 }, { "epoch": 18.40670859538784, "grad_norm": 0.1729562133550644, "learning_rate": 9.610335526099024e-07, "loss": 0.5091, "num_input_tokens_seen": 22975384, "step": 35120 }, { "epoch": 18.409329140461217, "grad_norm": 0.12775447964668274, "learning_rate": 9.578962168416278e-07, "loss": 0.4464, "num_input_tokens_seen": 22978488, "step": 35125 }, { "epoch": 18.41194968553459, "grad_norm": 0.2645840346813202, "learning_rate": 9.547639104071775e-07, "loss": 0.5807, "num_input_tokens_seen": 22982136, "step": 35130 }, { "epoch": 18.414570230607968, "grad_norm": 0.1573702096939087, "learning_rate": 9.516366339617944e-07, "loss": 0.5492, "num_input_tokens_seen": 22984440, "step": 35135 }, { "epoch": 18.41719077568134, "grad_norm": 0.11444928497076035, "learning_rate": 9.485143881596714e-07, "loss": 0.4498, "num_input_tokens_seen": 22987800, "step": 35140 }, { "epoch": 18.419811320754718, "grad_norm": 0.17804715037345886, "learning_rate": 9.453971736539447e-07, "loss": 0.5556, "num_input_tokens_seen": 22990584, "step": 35145 }, { "epoch": 18.42243186582809, "grad_norm": 0.21147149801254272, "learning_rate": 9.42284991096698e-07, "loss": 0.4473, "num_input_tokens_seen": 22994008, "step": 35150 }, { "epoch": 18.42505241090147, "grad_norm": 0.22576233744621277, "learning_rate": 9.391778411389634e-07, "loss": 0.5171, "num_input_tokens_seen": 22998136, "step": 35155 }, { "epoch": 18.427672955974842, "grad_norm": 0.2072986513376236, "learning_rate": 9.360757244307267e-07, "loss": 0.4183, "num_input_tokens_seen": 23000568, "step": 35160 }, { "epoch": 18.43029350104822, "grad_norm": 0.1658669263124466, "learning_rate": 9.329786416209157e-07, "loss": 0.4124, "num_input_tokens_seen": 23004056, "step": 35165 }, { "epoch": 18.432914046121592, "grad_norm": 0.14455977082252502, "learning_rate": 9.298865933574014e-07, "loss": 0.5877, "num_input_tokens_seen": 23007640, "step": 35170 }, { "epoch": 18.43553459119497, "grad_norm": 0.13507188856601715, "learning_rate": 9.267995802870022e-07, "loss": 0.4367, "num_input_tokens_seen": 23010488, "step": 35175 }, { "epoch": 18.438155136268342, "grad_norm": 0.13492047786712646, "learning_rate": 9.237176030554878e-07, "loss": 0.4479, "num_input_tokens_seen": 23014552, "step": 35180 }, { "epoch": 18.44077568134172, "grad_norm": 0.18807733058929443, "learning_rate": 9.206406623075758e-07, "loss": 0.5711, "num_input_tokens_seen": 23018840, "step": 35185 }, { "epoch": 18.443396226415093, "grad_norm": 0.16832603514194489, "learning_rate": 9.175687586869292e-07, "loss": 0.683, "num_input_tokens_seen": 23022520, "step": 35190 }, { "epoch": 18.44601677148847, "grad_norm": 0.1198088750243187, "learning_rate": 9.145018928361476e-07, "loss": 0.6145, "num_input_tokens_seen": 23026104, "step": 35195 }, { "epoch": 18.448637316561843, "grad_norm": 0.13303302228450775, "learning_rate": 9.114400653967903e-07, "loss": 0.6265, "num_input_tokens_seen": 23030296, "step": 35200 }, { "epoch": 18.45125786163522, "grad_norm": 0.16809892654418945, "learning_rate": 9.083832770093531e-07, "loss": 0.5068, "num_input_tokens_seen": 23033944, "step": 35205 }, { "epoch": 18.453878406708597, "grad_norm": 0.13555780053138733, "learning_rate": 9.053315283132802e-07, "loss": 0.6121, "num_input_tokens_seen": 23037912, "step": 35210 }, { "epoch": 18.45649895178197, "grad_norm": 0.2522129714488983, "learning_rate": 9.022848199469663e-07, "loss": 0.7106, "num_input_tokens_seen": 23041048, "step": 35215 }, { "epoch": 18.459119496855347, "grad_norm": 0.25247082114219666, "learning_rate": 8.992431525477518e-07, "loss": 0.4647, "num_input_tokens_seen": 23043992, "step": 35220 }, { "epoch": 18.46174004192872, "grad_norm": 0.1700076460838318, "learning_rate": 8.962065267519082e-07, "loss": 0.5636, "num_input_tokens_seen": 23047352, "step": 35225 }, { "epoch": 18.464360587002098, "grad_norm": 0.11075383424758911, "learning_rate": 8.931749431946718e-07, "loss": 0.6347, "num_input_tokens_seen": 23050488, "step": 35230 }, { "epoch": 18.46698113207547, "grad_norm": 0.18602116405963898, "learning_rate": 8.901484025102103e-07, "loss": 0.5291, "num_input_tokens_seen": 23053496, "step": 35235 }, { "epoch": 18.469601677148848, "grad_norm": 0.15570369362831116, "learning_rate": 8.871269053316422e-07, "loss": 0.4665, "num_input_tokens_seen": 23056760, "step": 35240 }, { "epoch": 18.47222222222222, "grad_norm": 0.15851382911205292, "learning_rate": 8.841104522910343e-07, "loss": 0.5167, "num_input_tokens_seen": 23060152, "step": 35245 }, { "epoch": 18.4748427672956, "grad_norm": 0.21539323031902313, "learning_rate": 8.8109904401939e-07, "loss": 0.4855, "num_input_tokens_seen": 23063448, "step": 35250 }, { "epoch": 18.47746331236897, "grad_norm": 0.13976530730724335, "learning_rate": 8.780926811466639e-07, "loss": 0.5039, "num_input_tokens_seen": 23066776, "step": 35255 }, { "epoch": 18.48008385744235, "grad_norm": 0.19901129603385925, "learning_rate": 8.7509136430175e-07, "loss": 0.5793, "num_input_tokens_seen": 23069784, "step": 35260 }, { "epoch": 18.482704402515722, "grad_norm": 0.1883413940668106, "learning_rate": 8.720950941124877e-07, "loss": 0.5226, "num_input_tokens_seen": 23072920, "step": 35265 }, { "epoch": 18.4853249475891, "grad_norm": 0.1281731128692627, "learning_rate": 8.691038712056732e-07, "loss": 0.5257, "num_input_tokens_seen": 23076408, "step": 35270 }, { "epoch": 18.487945492662472, "grad_norm": 0.2190302461385727, "learning_rate": 8.66117696207025e-07, "loss": 0.5476, "num_input_tokens_seen": 23079256, "step": 35275 }, { "epoch": 18.49056603773585, "grad_norm": 0.1680862307548523, "learning_rate": 8.631365697412186e-07, "loss": 0.4202, "num_input_tokens_seen": 23082488, "step": 35280 }, { "epoch": 18.493186582809223, "grad_norm": 0.11544285714626312, "learning_rate": 8.601604924318802e-07, "loss": 0.4436, "num_input_tokens_seen": 23089176, "step": 35285 }, { "epoch": 18.4958071278826, "grad_norm": 0.2439521700143814, "learning_rate": 8.571894649015588e-07, "loss": 0.4797, "num_input_tokens_seen": 23091928, "step": 35290 }, { "epoch": 18.498427672955973, "grad_norm": 0.10612642019987106, "learning_rate": 8.542234877717686e-07, "loss": 0.4302, "num_input_tokens_seen": 23095064, "step": 35295 }, { "epoch": 18.50104821802935, "grad_norm": 0.2246517837047577, "learning_rate": 8.512625616629549e-07, "loss": 0.4382, "num_input_tokens_seen": 23098616, "step": 35300 }, { "epoch": 18.503668763102727, "grad_norm": 0.2663184702396393, "learning_rate": 8.483066871945083e-07, "loss": 0.4852, "num_input_tokens_seen": 23101784, "step": 35305 }, { "epoch": 18.5062893081761, "grad_norm": 0.17435388267040253, "learning_rate": 8.453558649847676e-07, "loss": 0.3613, "num_input_tokens_seen": 23105816, "step": 35310 }, { "epoch": 18.508909853249477, "grad_norm": 0.1492895483970642, "learning_rate": 8.424100956510112e-07, "loss": 0.3183, "num_input_tokens_seen": 23110232, "step": 35315 }, { "epoch": 18.51153039832285, "grad_norm": 0.20011992752552032, "learning_rate": 8.394693798094544e-07, "loss": 0.3992, "num_input_tokens_seen": 23113624, "step": 35320 }, { "epoch": 18.514150943396228, "grad_norm": 0.3963407576084137, "learning_rate": 8.365337180752691e-07, "loss": 0.4478, "num_input_tokens_seen": 23117176, "step": 35325 }, { "epoch": 18.5167714884696, "grad_norm": 0.1152348518371582, "learning_rate": 8.336031110625558e-07, "loss": 0.4524, "num_input_tokens_seen": 23120472, "step": 35330 }, { "epoch": 18.519392033542978, "grad_norm": 0.17094336450099945, "learning_rate": 8.306775593843657e-07, "loss": 0.421, "num_input_tokens_seen": 23124344, "step": 35335 }, { "epoch": 18.52201257861635, "grad_norm": 0.1698130965232849, "learning_rate": 8.277570636526955e-07, "loss": 0.529, "num_input_tokens_seen": 23127608, "step": 35340 }, { "epoch": 18.52463312368973, "grad_norm": 0.12294552475214005, "learning_rate": 8.24841624478473e-07, "loss": 0.4967, "num_input_tokens_seen": 23130552, "step": 35345 }, { "epoch": 18.5272536687631, "grad_norm": 0.1852927953004837, "learning_rate": 8.219312424715825e-07, "loss": 0.5452, "num_input_tokens_seen": 23133784, "step": 35350 }, { "epoch": 18.52987421383648, "grad_norm": 0.20566239953041077, "learning_rate": 8.190259182408345e-07, "loss": 0.5205, "num_input_tokens_seen": 23138072, "step": 35355 }, { "epoch": 18.532494758909852, "grad_norm": 0.12471915036439896, "learning_rate": 8.161256523939925e-07, "loss": 0.4505, "num_input_tokens_seen": 23141080, "step": 35360 }, { "epoch": 18.53511530398323, "grad_norm": 0.245405375957489, "learning_rate": 8.132304455377576e-07, "loss": 0.6087, "num_input_tokens_seen": 23143960, "step": 35365 }, { "epoch": 18.537735849056602, "grad_norm": 0.14164862036705017, "learning_rate": 8.103402982777813e-07, "loss": 0.5396, "num_input_tokens_seen": 23147608, "step": 35370 }, { "epoch": 18.54035639412998, "grad_norm": 0.1453869789838791, "learning_rate": 8.07455211218644e-07, "loss": 0.4987, "num_input_tokens_seen": 23150872, "step": 35375 }, { "epoch": 18.542976939203353, "grad_norm": 0.14743715524673462, "learning_rate": 8.045751849638738e-07, "loss": 0.5209, "num_input_tokens_seen": 23153656, "step": 35380 }, { "epoch": 18.54559748427673, "grad_norm": 0.12697751820087433, "learning_rate": 8.017002201159363e-07, "loss": 0.3764, "num_input_tokens_seen": 23157176, "step": 35385 }, { "epoch": 18.548218029350103, "grad_norm": 0.16540050506591797, "learning_rate": 7.988303172762418e-07, "loss": 0.4226, "num_input_tokens_seen": 23160280, "step": 35390 }, { "epoch": 18.55083857442348, "grad_norm": 0.13201306760311127, "learning_rate": 7.959654770451463e-07, "loss": 0.3135, "num_input_tokens_seen": 23166552, "step": 35395 }, { "epoch": 18.553459119496857, "grad_norm": 0.15464647114276886, "learning_rate": 7.9310570002194e-07, "loss": 0.4052, "num_input_tokens_seen": 23170008, "step": 35400 }, { "epoch": 18.55607966457023, "grad_norm": 0.16742058098316193, "learning_rate": 7.902509868048552e-07, "loss": 0.4969, "num_input_tokens_seen": 23173336, "step": 35405 }, { "epoch": 18.558700209643607, "grad_norm": 0.3117983937263489, "learning_rate": 7.874013379910616e-07, "loss": 0.5654, "num_input_tokens_seen": 23177336, "step": 35410 }, { "epoch": 18.56132075471698, "grad_norm": 0.11861320585012436, "learning_rate": 7.845567541766769e-07, "loss": 0.5062, "num_input_tokens_seen": 23180632, "step": 35415 }, { "epoch": 18.563941299790358, "grad_norm": 0.19733545184135437, "learning_rate": 7.817172359567526e-07, "loss": 0.5294, "num_input_tokens_seen": 23183192, "step": 35420 }, { "epoch": 18.56656184486373, "grad_norm": 0.10538095235824585, "learning_rate": 7.788827839252888e-07, "loss": 0.4098, "num_input_tokens_seen": 23185912, "step": 35425 }, { "epoch": 18.569182389937108, "grad_norm": 0.2278834879398346, "learning_rate": 7.760533986752222e-07, "loss": 0.6971, "num_input_tokens_seen": 23189368, "step": 35430 }, { "epoch": 18.57180293501048, "grad_norm": 0.10026554763317108, "learning_rate": 7.732290807984155e-07, "loss": 0.4616, "num_input_tokens_seen": 23192728, "step": 35435 }, { "epoch": 18.57442348008386, "grad_norm": 0.10938360542058945, "learning_rate": 7.704098308856933e-07, "loss": 0.549, "num_input_tokens_seen": 23195640, "step": 35440 }, { "epoch": 18.57704402515723, "grad_norm": 0.22754833102226257, "learning_rate": 7.675956495268088e-07, "loss": 0.6627, "num_input_tokens_seen": 23198200, "step": 35445 }, { "epoch": 18.57966457023061, "grad_norm": 0.18190540373325348, "learning_rate": 7.647865373104551e-07, "loss": 0.4174, "num_input_tokens_seen": 23200760, "step": 35450 }, { "epoch": 18.582285115303982, "grad_norm": 0.13877901434898376, "learning_rate": 7.619824948242732e-07, "loss": 0.4365, "num_input_tokens_seen": 23204280, "step": 35455 }, { "epoch": 18.58490566037736, "grad_norm": 0.29161524772644043, "learning_rate": 7.5918352265483e-07, "loss": 0.6299, "num_input_tokens_seen": 23207800, "step": 35460 }, { "epoch": 18.587526205450732, "grad_norm": 0.14263026416301727, "learning_rate": 7.563896213876348e-07, "loss": 0.4446, "num_input_tokens_seen": 23210648, "step": 35465 }, { "epoch": 18.59014675052411, "grad_norm": 0.12308794260025024, "learning_rate": 7.53600791607148e-07, "loss": 0.3473, "num_input_tokens_seen": 23213752, "step": 35470 }, { "epoch": 18.592767295597483, "grad_norm": 0.10768318176269531, "learning_rate": 7.508170338967585e-07, "loss": 0.3807, "num_input_tokens_seen": 23216408, "step": 35475 }, { "epoch": 18.59538784067086, "grad_norm": 0.13814277946949005, "learning_rate": 7.480383488387976e-07, "loss": 0.4423, "num_input_tokens_seen": 23219480, "step": 35480 }, { "epoch": 18.598008385744233, "grad_norm": 0.19410169124603271, "learning_rate": 7.452647370145338e-07, "loss": 0.5725, "num_input_tokens_seen": 23222872, "step": 35485 }, { "epoch": 18.60062893081761, "grad_norm": 0.17913369834423065, "learning_rate": 7.424961990041696e-07, "loss": 0.6235, "num_input_tokens_seen": 23226584, "step": 35490 }, { "epoch": 18.603249475890987, "grad_norm": 0.13238994777202606, "learning_rate": 7.397327353868583e-07, "loss": 0.4885, "num_input_tokens_seen": 23229112, "step": 35495 }, { "epoch": 18.60587002096436, "grad_norm": 0.09059464931488037, "learning_rate": 7.369743467406848e-07, "loss": 0.3569, "num_input_tokens_seen": 23231800, "step": 35500 }, { "epoch": 18.608490566037737, "grad_norm": 0.19412827491760254, "learning_rate": 7.34221033642668e-07, "loss": 0.4356, "num_input_tokens_seen": 23235192, "step": 35505 }, { "epoch": 18.61111111111111, "grad_norm": 0.25881466269493103, "learning_rate": 7.314727966687751e-07, "loss": 0.4818, "num_input_tokens_seen": 23238008, "step": 35510 }, { "epoch": 18.613731656184488, "grad_norm": 0.18182532489299774, "learning_rate": 7.287296363939045e-07, "loss": 0.5606, "num_input_tokens_seen": 23240536, "step": 35515 }, { "epoch": 18.61635220125786, "grad_norm": 0.12551917135715485, "learning_rate": 7.259915533918888e-07, "loss": 0.6084, "num_input_tokens_seen": 23243576, "step": 35520 }, { "epoch": 18.618972746331238, "grad_norm": 0.1465214043855667, "learning_rate": 7.232585482355086e-07, "loss": 0.4726, "num_input_tokens_seen": 23247256, "step": 35525 }, { "epoch": 18.62159329140461, "grad_norm": 0.1023135855793953, "learning_rate": 7.205306214964736e-07, "loss": 0.3654, "num_input_tokens_seen": 23250456, "step": 35530 }, { "epoch": 18.62421383647799, "grad_norm": 0.08337704092264175, "learning_rate": 7.178077737454436e-07, "loss": 0.4683, "num_input_tokens_seen": 23253272, "step": 35535 }, { "epoch": 18.62683438155136, "grad_norm": 0.14153094589710236, "learning_rate": 7.150900055519993e-07, "loss": 0.5101, "num_input_tokens_seen": 23256824, "step": 35540 }, { "epoch": 18.62945492662474, "grad_norm": 0.2365569919347763, "learning_rate": 7.123773174846693e-07, "loss": 0.6234, "num_input_tokens_seen": 23261048, "step": 35545 }, { "epoch": 18.632075471698112, "grad_norm": 0.22858569025993347, "learning_rate": 7.096697101109134e-07, "loss": 0.5511, "num_input_tokens_seen": 23264216, "step": 35550 }, { "epoch": 18.63469601677149, "grad_norm": 0.15493035316467285, "learning_rate": 7.069671839971371e-07, "loss": 0.4745, "num_input_tokens_seen": 23267448, "step": 35555 }, { "epoch": 18.637316561844862, "grad_norm": 0.16786204278469086, "learning_rate": 7.042697397086795e-07, "loss": 0.4758, "num_input_tokens_seen": 23271448, "step": 35560 }, { "epoch": 18.63993710691824, "grad_norm": 0.19679318368434906, "learning_rate": 7.015773778098089e-07, "loss": 0.5495, "num_input_tokens_seen": 23274296, "step": 35565 }, { "epoch": 18.642557651991613, "grad_norm": 0.43201813101768494, "learning_rate": 6.988900988637415e-07, "loss": 0.4423, "num_input_tokens_seen": 23276664, "step": 35570 }, { "epoch": 18.64517819706499, "grad_norm": 0.1134626716375351, "learning_rate": 6.962079034326218e-07, "loss": 0.4514, "num_input_tokens_seen": 23280632, "step": 35575 }, { "epoch": 18.647798742138363, "grad_norm": 0.15684792399406433, "learning_rate": 6.935307920775346e-07, "loss": 0.3468, "num_input_tokens_seen": 23285624, "step": 35580 }, { "epoch": 18.65041928721174, "grad_norm": 0.15409739315509796, "learning_rate": 6.90858765358507e-07, "loss": 0.4262, "num_input_tokens_seen": 23288632, "step": 35585 }, { "epoch": 18.653039832285117, "grad_norm": 0.18587733805179596, "learning_rate": 6.881918238344915e-07, "loss": 0.4678, "num_input_tokens_seen": 23291800, "step": 35590 }, { "epoch": 18.65566037735849, "grad_norm": 0.15394210815429688, "learning_rate": 6.855299680633808e-07, "loss": 0.4762, "num_input_tokens_seen": 23296088, "step": 35595 }, { "epoch": 18.658280922431867, "grad_norm": 0.1454034149646759, "learning_rate": 6.828731986020104e-07, "loss": 0.4142, "num_input_tokens_seen": 23299096, "step": 35600 }, { "epoch": 18.66090146750524, "grad_norm": 0.12805604934692383, "learning_rate": 6.802215160061381e-07, "loss": 0.5308, "num_input_tokens_seen": 23301624, "step": 35605 }, { "epoch": 18.663522012578618, "grad_norm": 0.16617627441883087, "learning_rate": 6.775749208304704e-07, "loss": 0.5657, "num_input_tokens_seen": 23305240, "step": 35610 }, { "epoch": 18.66614255765199, "grad_norm": 0.18613943457603455, "learning_rate": 6.749334136286478e-07, "loss": 0.4368, "num_input_tokens_seen": 23308344, "step": 35615 }, { "epoch": 18.668763102725368, "grad_norm": 0.15785321593284607, "learning_rate": 6.722969949532365e-07, "loss": 0.5459, "num_input_tokens_seen": 23313272, "step": 35620 }, { "epoch": 18.67138364779874, "grad_norm": 0.17821985483169556, "learning_rate": 6.69665665355751e-07, "loss": 0.5975, "num_input_tokens_seen": 23316088, "step": 35625 }, { "epoch": 18.67400419287212, "grad_norm": 0.23582856357097626, "learning_rate": 6.670394253866314e-07, "loss": 0.4565, "num_input_tokens_seen": 23320312, "step": 35630 }, { "epoch": 18.67662473794549, "grad_norm": 0.41371434926986694, "learning_rate": 6.644182755952605e-07, "loss": 0.459, "num_input_tokens_seen": 23322936, "step": 35635 }, { "epoch": 18.67924528301887, "grad_norm": 0.15203233063220978, "learning_rate": 6.618022165299526e-07, "loss": 0.5622, "num_input_tokens_seen": 23326968, "step": 35640 }, { "epoch": 18.681865828092242, "grad_norm": 0.231183722615242, "learning_rate": 6.591912487379531e-07, "loss": 0.494, "num_input_tokens_seen": 23330264, "step": 35645 }, { "epoch": 18.68448637316562, "grad_norm": 0.1445288360118866, "learning_rate": 6.565853727654503e-07, "loss": 0.4155, "num_input_tokens_seen": 23333912, "step": 35650 }, { "epoch": 18.687106918238992, "grad_norm": 0.20121899247169495, "learning_rate": 6.539845891575636e-07, "loss": 0.4055, "num_input_tokens_seen": 23336920, "step": 35655 }, { "epoch": 18.68972746331237, "grad_norm": 0.168095201253891, "learning_rate": 6.513888984583466e-07, "loss": 0.5409, "num_input_tokens_seen": 23339800, "step": 35660 }, { "epoch": 18.692348008385743, "grad_norm": 0.1659780740737915, "learning_rate": 6.487983012107873e-07, "loss": 0.4348, "num_input_tokens_seen": 23343608, "step": 35665 }, { "epoch": 18.69496855345912, "grad_norm": 0.14867079257965088, "learning_rate": 6.462127979568133e-07, "loss": 0.5328, "num_input_tokens_seen": 23346456, "step": 35670 }, { "epoch": 18.697589098532493, "grad_norm": 0.11880723387002945, "learning_rate": 6.436323892372781e-07, "loss": 0.5113, "num_input_tokens_seen": 23349336, "step": 35675 }, { "epoch": 18.70020964360587, "grad_norm": 0.15839196741580963, "learning_rate": 6.410570755919721e-07, "loss": 0.3929, "num_input_tokens_seen": 23352344, "step": 35680 }, { "epoch": 18.702830188679247, "grad_norm": 0.10473884642124176, "learning_rate": 6.384868575596281e-07, "loss": 0.392, "num_input_tokens_seen": 23355064, "step": 35685 }, { "epoch": 18.70545073375262, "grad_norm": 0.1811973601579666, "learning_rate": 6.359217356779024e-07, "loss": 0.4769, "num_input_tokens_seen": 23358680, "step": 35690 }, { "epoch": 18.708071278825997, "grad_norm": 0.22983238101005554, "learning_rate": 6.333617104833878e-07, "loss": 0.4165, "num_input_tokens_seen": 23361496, "step": 35695 }, { "epoch": 18.71069182389937, "grad_norm": 0.21028994023799896, "learning_rate": 6.308067825116142e-07, "loss": 0.5628, "num_input_tokens_seen": 23366264, "step": 35700 }, { "epoch": 18.713312368972748, "grad_norm": 0.18946202099323273, "learning_rate": 6.282569522970433e-07, "loss": 0.461, "num_input_tokens_seen": 23368888, "step": 35705 }, { "epoch": 18.71593291404612, "grad_norm": 0.10366552323102951, "learning_rate": 6.257122203730731e-07, "loss": 0.4173, "num_input_tokens_seen": 23372216, "step": 35710 }, { "epoch": 18.718553459119498, "grad_norm": 0.12662650644779205, "learning_rate": 6.231725872720251e-07, "loss": 0.4868, "num_input_tokens_seen": 23376248, "step": 35715 }, { "epoch": 18.72117400419287, "grad_norm": 0.09904143959283829, "learning_rate": 6.206380535251716e-07, "loss": 0.475, "num_input_tokens_seen": 23379544, "step": 35720 }, { "epoch": 18.72379454926625, "grad_norm": 0.13829922676086426, "learning_rate": 6.181086196626995e-07, "loss": 0.5589, "num_input_tokens_seen": 23383576, "step": 35725 }, { "epoch": 18.72641509433962, "grad_norm": 0.08851338177919388, "learning_rate": 6.155842862137384e-07, "loss": 0.6033, "num_input_tokens_seen": 23387352, "step": 35730 }, { "epoch": 18.729035639413, "grad_norm": 0.1584225594997406, "learning_rate": 6.130650537063548e-07, "loss": 0.3343, "num_input_tokens_seen": 23390808, "step": 35735 }, { "epoch": 18.731656184486372, "grad_norm": 0.15340648591518402, "learning_rate": 6.105509226675438e-07, "loss": 0.4273, "num_input_tokens_seen": 23393432, "step": 35740 }, { "epoch": 18.73427672955975, "grad_norm": 0.10446885973215103, "learning_rate": 6.08041893623229e-07, "loss": 0.3518, "num_input_tokens_seen": 23396536, "step": 35745 }, { "epoch": 18.736897274633122, "grad_norm": 0.14634044468402863, "learning_rate": 6.055379670982713e-07, "loss": 0.3303, "num_input_tokens_seen": 23399416, "step": 35750 }, { "epoch": 18.7395178197065, "grad_norm": 0.1863401085138321, "learning_rate": 6.030391436164628e-07, "loss": 0.4318, "num_input_tokens_seen": 23402392, "step": 35755 }, { "epoch": 18.742138364779873, "grad_norm": 0.15784099698066711, "learning_rate": 6.005454237005298e-07, "loss": 0.4161, "num_input_tokens_seen": 23404792, "step": 35760 }, { "epoch": 18.74475890985325, "grad_norm": 0.20591451227664948, "learning_rate": 5.980568078721327e-07, "loss": 0.3883, "num_input_tokens_seen": 23408472, "step": 35765 }, { "epoch": 18.747379454926623, "grad_norm": 0.1089872494339943, "learning_rate": 5.955732966518606e-07, "loss": 0.4364, "num_input_tokens_seen": 23411768, "step": 35770 }, { "epoch": 18.75, "grad_norm": 0.2343067079782486, "learning_rate": 5.930948905592343e-07, "loss": 0.5086, "num_input_tokens_seen": 23415128, "step": 35775 }, { "epoch": 18.752620545073377, "grad_norm": 0.14823320508003235, "learning_rate": 5.906215901127054e-07, "loss": 0.4579, "num_input_tokens_seen": 23418008, "step": 35780 }, { "epoch": 18.75524109014675, "grad_norm": 0.26376286149024963, "learning_rate": 5.881533958296631e-07, "loss": 0.47, "num_input_tokens_seen": 23420888, "step": 35785 }, { "epoch": 18.757861635220127, "grad_norm": 0.1704409271478653, "learning_rate": 5.856903082264248e-07, "loss": 0.4955, "num_input_tokens_seen": 23423128, "step": 35790 }, { "epoch": 18.7604821802935, "grad_norm": 0.14757631719112396, "learning_rate": 5.832323278182422e-07, "loss": 0.5125, "num_input_tokens_seen": 23426136, "step": 35795 }, { "epoch": 18.763102725366878, "grad_norm": 0.1399320811033249, "learning_rate": 5.807794551192986e-07, "loss": 0.4892, "num_input_tokens_seen": 23429144, "step": 35800 }, { "epoch": 18.76572327044025, "grad_norm": 0.19146089255809784, "learning_rate": 5.783316906427e-07, "loss": 0.4704, "num_input_tokens_seen": 23432216, "step": 35805 }, { "epoch": 18.768343815513628, "grad_norm": 0.15021823346614838, "learning_rate": 5.758890349004925e-07, "loss": 0.4314, "num_input_tokens_seen": 23435000, "step": 35810 }, { "epoch": 18.770964360587, "grad_norm": 0.27386239171028137, "learning_rate": 5.734514884036563e-07, "loss": 0.3905, "num_input_tokens_seen": 23438008, "step": 35815 }, { "epoch": 18.77358490566038, "grad_norm": 0.17283028364181519, "learning_rate": 5.710190516620945e-07, "loss": 0.6414, "num_input_tokens_seen": 23440504, "step": 35820 }, { "epoch": 18.77620545073375, "grad_norm": 0.20703725516796112, "learning_rate": 5.685917251846501e-07, "loss": 0.5118, "num_input_tokens_seen": 23443096, "step": 35825 }, { "epoch": 18.77882599580713, "grad_norm": 0.15392731130123138, "learning_rate": 5.661695094790892e-07, "loss": 0.4417, "num_input_tokens_seen": 23446392, "step": 35830 }, { "epoch": 18.781446540880502, "grad_norm": 0.18098975718021393, "learning_rate": 5.637524050521064e-07, "loss": 0.3767, "num_input_tokens_seen": 23449208, "step": 35835 }, { "epoch": 18.78406708595388, "grad_norm": 0.19624848663806915, "learning_rate": 5.613404124093391e-07, "loss": 0.4455, "num_input_tokens_seen": 23452376, "step": 35840 }, { "epoch": 18.786687631027252, "grad_norm": 0.23464085161685944, "learning_rate": 5.589335320553446e-07, "loss": 0.3739, "num_input_tokens_seen": 23455768, "step": 35845 }, { "epoch": 18.78930817610063, "grad_norm": 0.1977182775735855, "learning_rate": 5.565317644936202e-07, "loss": 0.4515, "num_input_tokens_seen": 23458680, "step": 35850 }, { "epoch": 18.791928721174003, "grad_norm": 0.19210901856422424, "learning_rate": 5.541351102265863e-07, "loss": 0.4891, "num_input_tokens_seen": 23462456, "step": 35855 }, { "epoch": 18.79454926624738, "grad_norm": 0.15838882327079773, "learning_rate": 5.51743569755589e-07, "loss": 0.5985, "num_input_tokens_seen": 23465656, "step": 35860 }, { "epoch": 18.797169811320753, "grad_norm": 0.25797027349472046, "learning_rate": 5.493571435809197e-07, "loss": 0.5729, "num_input_tokens_seen": 23468472, "step": 35865 }, { "epoch": 18.79979035639413, "grad_norm": 0.1291886419057846, "learning_rate": 5.469758322017849e-07, "loss": 0.4344, "num_input_tokens_seen": 23472248, "step": 35870 }, { "epoch": 18.802410901467507, "grad_norm": 0.2018984854221344, "learning_rate": 5.445996361163358e-07, "loss": 0.3942, "num_input_tokens_seen": 23474680, "step": 35875 }, { "epoch": 18.80503144654088, "grad_norm": 0.1472114771604538, "learning_rate": 5.422285558216389e-07, "loss": 0.5704, "num_input_tokens_seen": 23477368, "step": 35880 }, { "epoch": 18.807651991614257, "grad_norm": 0.16011784970760345, "learning_rate": 5.398625918137001e-07, "loss": 0.5353, "num_input_tokens_seen": 23481112, "step": 35885 }, { "epoch": 18.81027253668763, "grad_norm": 0.20143374800682068, "learning_rate": 5.375017445874486e-07, "loss": 0.5075, "num_input_tokens_seen": 23484088, "step": 35890 }, { "epoch": 18.812893081761008, "grad_norm": 0.15197354555130005, "learning_rate": 5.351460146367504e-07, "loss": 0.4496, "num_input_tokens_seen": 23487288, "step": 35895 }, { "epoch": 18.81551362683438, "grad_norm": 0.18547849357128143, "learning_rate": 5.327954024543946e-07, "loss": 0.5226, "num_input_tokens_seen": 23490680, "step": 35900 }, { "epoch": 18.818134171907758, "grad_norm": 0.14616137742996216, "learning_rate": 5.304499085321074e-07, "loss": 0.502, "num_input_tokens_seen": 23493144, "step": 35905 }, { "epoch": 18.82075471698113, "grad_norm": 0.17525573074817657, "learning_rate": 5.281095333605352e-07, "loss": 0.5483, "num_input_tokens_seen": 23495608, "step": 35910 }, { "epoch": 18.82337526205451, "grad_norm": 0.15912069380283356, "learning_rate": 5.257742774292556e-07, "loss": 0.429, "num_input_tokens_seen": 23499544, "step": 35915 }, { "epoch": 18.82599580712788, "grad_norm": 0.2694525420665741, "learning_rate": 5.234441412267837e-07, "loss": 0.4402, "num_input_tokens_seen": 23502168, "step": 35920 }, { "epoch": 18.82861635220126, "grad_norm": 0.12172813713550568, "learning_rate": 5.211191252405517e-07, "loss": 0.4929, "num_input_tokens_seen": 23505816, "step": 35925 }, { "epoch": 18.831236897274632, "grad_norm": 0.1903015524148941, "learning_rate": 5.187992299569316e-07, "loss": 0.4952, "num_input_tokens_seen": 23509176, "step": 35930 }, { "epoch": 18.83385744234801, "grad_norm": 0.2338835746049881, "learning_rate": 5.164844558612131e-07, "loss": 0.514, "num_input_tokens_seen": 23512024, "step": 35935 }, { "epoch": 18.836477987421382, "grad_norm": 0.12440633773803711, "learning_rate": 5.141748034376254e-07, "loss": 0.5429, "num_input_tokens_seen": 23514552, "step": 35940 }, { "epoch": 18.83909853249476, "grad_norm": 0.20150543749332428, "learning_rate": 5.118702731693209e-07, "loss": 0.5493, "num_input_tokens_seen": 23517528, "step": 35945 }, { "epoch": 18.841719077568133, "grad_norm": 0.1117669939994812, "learning_rate": 5.095708655383751e-07, "loss": 0.4429, "num_input_tokens_seen": 23521272, "step": 35950 }, { "epoch": 18.84433962264151, "grad_norm": 0.1611711084842682, "learning_rate": 5.07276581025809e-07, "loss": 0.531, "num_input_tokens_seen": 23524632, "step": 35955 }, { "epoch": 18.846960167714883, "grad_norm": 0.11787273734807968, "learning_rate": 5.049874201115495e-07, "loss": 0.4351, "num_input_tokens_seen": 23531160, "step": 35960 }, { "epoch": 18.84958071278826, "grad_norm": 0.16735437512397766, "learning_rate": 5.027033832744693e-07, "loss": 0.5625, "num_input_tokens_seen": 23534104, "step": 35965 }, { "epoch": 18.852201257861637, "grad_norm": 0.15885809063911438, "learning_rate": 5.004244709923639e-07, "loss": 0.511, "num_input_tokens_seen": 23536856, "step": 35970 }, { "epoch": 18.85482180293501, "grad_norm": 0.13839022815227509, "learning_rate": 4.981506837419519e-07, "loss": 0.624, "num_input_tokens_seen": 23540152, "step": 35975 }, { "epoch": 18.857442348008387, "grad_norm": 0.15032942593097687, "learning_rate": 4.958820219988836e-07, "loss": 0.5412, "num_input_tokens_seen": 23543640, "step": 35980 }, { "epoch": 18.86006289308176, "grad_norm": 0.16266286373138428, "learning_rate": 4.936184862377402e-07, "loss": 0.4774, "num_input_tokens_seen": 23547064, "step": 35985 }, { "epoch": 18.862683438155138, "grad_norm": 0.19967998564243317, "learning_rate": 4.913600769320237e-07, "loss": 0.5904, "num_input_tokens_seen": 23549944, "step": 35990 }, { "epoch": 18.86530398322851, "grad_norm": 0.14975054562091827, "learning_rate": 4.891067945541728e-07, "loss": 0.5376, "num_input_tokens_seen": 23553560, "step": 35995 }, { "epoch": 18.867924528301888, "grad_norm": 0.18183879554271698, "learning_rate": 4.868586395755437e-07, "loss": 0.5196, "num_input_tokens_seen": 23556504, "step": 36000 }, { "epoch": 18.87054507337526, "grad_norm": 0.15392442047595978, "learning_rate": 4.846156124664242e-07, "loss": 0.6269, "num_input_tokens_seen": 23560376, "step": 36005 }, { "epoch": 18.87316561844864, "grad_norm": 0.2729138731956482, "learning_rate": 4.823777136960362e-07, "loss": 0.3694, "num_input_tokens_seen": 23563064, "step": 36010 }, { "epoch": 18.87578616352201, "grad_norm": 0.10096997022628784, "learning_rate": 4.801449437325162e-07, "loss": 0.428, "num_input_tokens_seen": 23566456, "step": 36015 }, { "epoch": 18.87840670859539, "grad_norm": 0.123235784471035, "learning_rate": 4.779173030429379e-07, "loss": 0.4041, "num_input_tokens_seen": 23569976, "step": 36020 }, { "epoch": 18.881027253668762, "grad_norm": 0.1490115225315094, "learning_rate": 4.7569479209329804e-07, "loss": 0.6043, "num_input_tokens_seen": 23573368, "step": 36025 }, { "epoch": 18.88364779874214, "grad_norm": 0.21732334792613983, "learning_rate": 4.734774113485163e-07, "loss": 0.3957, "num_input_tokens_seen": 23576952, "step": 36030 }, { "epoch": 18.886268343815512, "grad_norm": 0.3234836459159851, "learning_rate": 4.7126516127244945e-07, "loss": 0.467, "num_input_tokens_seen": 23580248, "step": 36035 }, { "epoch": 18.88888888888889, "grad_norm": 0.16766193509101868, "learning_rate": 4.690580423278718e-07, "loss": 0.3925, "num_input_tokens_seen": 23583128, "step": 36040 }, { "epoch": 18.891509433962263, "grad_norm": 0.20802074670791626, "learning_rate": 4.668560549764861e-07, "loss": 0.5206, "num_input_tokens_seen": 23586104, "step": 36045 }, { "epoch": 18.89412997903564, "grad_norm": 0.16332507133483887, "learning_rate": 4.646591996789268e-07, "loss": 0.5588, "num_input_tokens_seen": 23589400, "step": 36050 }, { "epoch": 18.896750524109013, "grad_norm": 0.18098598718643188, "learning_rate": 4.6246747689474847e-07, "loss": 0.5582, "num_input_tokens_seen": 23592184, "step": 36055 }, { "epoch": 18.89937106918239, "grad_norm": 0.16041427850723267, "learning_rate": 4.6028088708243434e-07, "loss": 0.5544, "num_input_tokens_seen": 23595064, "step": 36060 }, { "epoch": 18.901991614255767, "grad_norm": 0.2116982340812683, "learning_rate": 4.580994306993991e-07, "loss": 0.4218, "num_input_tokens_seen": 23597432, "step": 36065 }, { "epoch": 18.90461215932914, "grad_norm": 0.17759020626544952, "learning_rate": 4.5592310820196936e-07, "loss": 0.5158, "num_input_tokens_seen": 23600600, "step": 36070 }, { "epoch": 18.907232704402517, "grad_norm": 0.1333121657371521, "learning_rate": 4.537519200454116e-07, "loss": 0.488, "num_input_tokens_seen": 23604440, "step": 36075 }, { "epoch": 18.90985324947589, "grad_norm": 0.1395299881696701, "learning_rate": 4.515858666839151e-07, "loss": 0.628, "num_input_tokens_seen": 23607736, "step": 36080 }, { "epoch": 18.912473794549268, "grad_norm": 0.23151300847530365, "learning_rate": 4.494249485705926e-07, "loss": 0.5803, "num_input_tokens_seen": 23615096, "step": 36085 }, { "epoch": 18.91509433962264, "grad_norm": 0.15528781712055206, "learning_rate": 4.472691661574824e-07, "loss": 0.4019, "num_input_tokens_seen": 23618584, "step": 36090 }, { "epoch": 18.917714884696018, "grad_norm": 0.20657579600811005, "learning_rate": 4.4511851989555156e-07, "loss": 0.6358, "num_input_tokens_seen": 23622808, "step": 36095 }, { "epoch": 18.92033542976939, "grad_norm": 0.22150340676307678, "learning_rate": 4.429730102346874e-07, "loss": 0.5031, "num_input_tokens_seen": 23625752, "step": 36100 }, { "epoch": 18.92295597484277, "grad_norm": 0.11893639713525772, "learning_rate": 4.4083263762370873e-07, "loss": 0.5184, "num_input_tokens_seen": 23629144, "step": 36105 }, { "epoch": 18.92557651991614, "grad_norm": 0.28903505206108093, "learning_rate": 4.386974025103574e-07, "loss": 0.4126, "num_input_tokens_seen": 23632440, "step": 36110 }, { "epoch": 18.92819706498952, "grad_norm": 0.09387104213237762, "learning_rate": 4.365673053413011e-07, "loss": 0.4222, "num_input_tokens_seen": 23635448, "step": 36115 }, { "epoch": 18.930817610062892, "grad_norm": 0.13455109298229218, "learning_rate": 4.34442346562125e-07, "loss": 0.6249, "num_input_tokens_seen": 23639160, "step": 36120 }, { "epoch": 18.93343815513627, "grad_norm": 0.12629276514053345, "learning_rate": 4.323225266173542e-07, "loss": 0.3949, "num_input_tokens_seen": 23642904, "step": 36125 }, { "epoch": 18.936058700209642, "grad_norm": 0.11636258661746979, "learning_rate": 4.302078459504283e-07, "loss": 0.3271, "num_input_tokens_seen": 23646200, "step": 36130 }, { "epoch": 18.93867924528302, "grad_norm": 0.14097599685192108, "learning_rate": 4.2809830500371294e-07, "loss": 0.446, "num_input_tokens_seen": 23649304, "step": 36135 }, { "epoch": 18.941299790356393, "grad_norm": 0.13458964228630066, "learning_rate": 4.2599390421850236e-07, "loss": 0.5776, "num_input_tokens_seen": 23652024, "step": 36140 }, { "epoch": 18.94392033542977, "grad_norm": 0.2361467033624649, "learning_rate": 4.2389464403501113e-07, "loss": 0.4515, "num_input_tokens_seen": 23655608, "step": 36145 }, { "epoch": 18.946540880503143, "grad_norm": 0.16179728507995605, "learning_rate": 4.218005248923823e-07, "loss": 0.4778, "num_input_tokens_seen": 23658616, "step": 36150 }, { "epoch": 18.94916142557652, "grad_norm": 0.22855643928050995, "learning_rate": 4.1971154722867946e-07, "loss": 0.4258, "num_input_tokens_seen": 23661976, "step": 36155 }, { "epoch": 18.951781970649897, "grad_norm": 0.21579329669475555, "learning_rate": 4.176277114808946e-07, "loss": 0.457, "num_input_tokens_seen": 23665080, "step": 36160 }, { "epoch": 18.95440251572327, "grad_norm": 0.1610776036977768, "learning_rate": 4.155490180849431e-07, "loss": 0.5719, "num_input_tokens_seen": 23667736, "step": 36165 }, { "epoch": 18.957023060796647, "grad_norm": 0.11198046058416367, "learning_rate": 4.1347546747566304e-07, "loss": 0.3277, "num_input_tokens_seen": 23672824, "step": 36170 }, { "epoch": 18.95964360587002, "grad_norm": 0.1907840520143509, "learning_rate": 4.114070600868131e-07, "loss": 0.4821, "num_input_tokens_seen": 23676024, "step": 36175 }, { "epoch": 18.962264150943398, "grad_norm": 0.17284688353538513, "learning_rate": 4.093437963510888e-07, "loss": 0.5258, "num_input_tokens_seen": 23678936, "step": 36180 }, { "epoch": 18.96488469601677, "grad_norm": 0.18646858632564545, "learning_rate": 4.072856767000949e-07, "loss": 0.6003, "num_input_tokens_seen": 23683160, "step": 36185 }, { "epoch": 18.967505241090148, "grad_norm": 0.1591888815164566, "learning_rate": 4.052327015643703e-07, "loss": 0.4514, "num_input_tokens_seen": 23685912, "step": 36190 }, { "epoch": 18.97012578616352, "grad_norm": 0.1649118810892105, "learning_rate": 4.031848713733744e-07, "loss": 0.472, "num_input_tokens_seen": 23688536, "step": 36195 }, { "epoch": 18.9727463312369, "grad_norm": 0.13531093299388885, "learning_rate": 4.011421865554837e-07, "loss": 0.4442, "num_input_tokens_seen": 23691608, "step": 36200 }, { "epoch": 18.97536687631027, "grad_norm": 0.11656789481639862, "learning_rate": 3.9910464753801224e-07, "loss": 0.4475, "num_input_tokens_seen": 23694936, "step": 36205 }, { "epoch": 18.97798742138365, "grad_norm": 0.12687397003173828, "learning_rate": 3.9707225474718557e-07, "loss": 0.4886, "num_input_tokens_seen": 23698168, "step": 36210 }, { "epoch": 18.980607966457022, "grad_norm": 0.14669866859912872, "learning_rate": 3.950450086081553e-07, "loss": 0.4901, "num_input_tokens_seen": 23701080, "step": 36215 }, { "epoch": 18.9832285115304, "grad_norm": 0.1695481687784195, "learning_rate": 3.930229095450072e-07, "loss": 0.439, "num_input_tokens_seen": 23704088, "step": 36220 }, { "epoch": 18.985849056603772, "grad_norm": 0.19305379688739777, "learning_rate": 3.9100595798073345e-07, "loss": 0.5683, "num_input_tokens_seen": 23706520, "step": 36225 }, { "epoch": 18.98846960167715, "grad_norm": 0.15158917009830475, "learning_rate": 3.8899415433725493e-07, "loss": 0.475, "num_input_tokens_seen": 23709656, "step": 36230 }, { "epoch": 18.991090146750523, "grad_norm": 0.2310827374458313, "learning_rate": 3.8698749903542666e-07, "loss": 0.4465, "num_input_tokens_seen": 23712280, "step": 36235 }, { "epoch": 18.9937106918239, "grad_norm": 0.1524980366230011, "learning_rate": 3.8498599249501e-07, "loss": 0.4537, "num_input_tokens_seen": 23715224, "step": 36240 }, { "epoch": 18.996331236897273, "grad_norm": 0.18087439239025116, "learning_rate": 3.8298963513470345e-07, "loss": 0.368, "num_input_tokens_seen": 23719224, "step": 36245 }, { "epoch": 18.99895178197065, "grad_norm": 0.17890985310077667, "learning_rate": 3.8099842737211734e-07, "loss": 0.4855, "num_input_tokens_seen": 23721848, "step": 36250 }, { "epoch": 19.0, "eval_loss": 0.46043843030929565, "eval_runtime": 10.0001, "eval_samples_per_second": 84.799, "eval_steps_per_second": 21.2, "num_input_tokens_seen": 23722376, "step": 36252 }, { "epoch": 19.001572327044027, "grad_norm": 0.21338342130184174, "learning_rate": 3.7901236962379627e-07, "loss": 0.4944, "num_input_tokens_seen": 23723848, "step": 36255 }, { "epoch": 19.0041928721174, "grad_norm": 0.16699914634227753, "learning_rate": 3.7703146230519126e-07, "loss": 0.513, "num_input_tokens_seen": 23727912, "step": 36260 }, { "epoch": 19.006813417190777, "grad_norm": 0.27170777320861816, "learning_rate": 3.7505570583069306e-07, "loss": 0.4644, "num_input_tokens_seen": 23730696, "step": 36265 }, { "epoch": 19.00943396226415, "grad_norm": 0.13136684894561768, "learning_rate": 3.7308510061360435e-07, "loss": 0.418, "num_input_tokens_seen": 23733960, "step": 36270 }, { "epoch": 19.012054507337528, "grad_norm": 0.25057077407836914, "learning_rate": 3.7111964706615374e-07, "loss": 0.5433, "num_input_tokens_seen": 23736712, "step": 36275 }, { "epoch": 19.0146750524109, "grad_norm": 0.1894306093454361, "learning_rate": 3.691593455994929e-07, "loss": 0.6228, "num_input_tokens_seen": 23739624, "step": 36280 }, { "epoch": 19.017295597484278, "grad_norm": 0.1525566130876541, "learning_rate": 3.672041966236911e-07, "loss": 0.3884, "num_input_tokens_seen": 23742696, "step": 36285 }, { "epoch": 19.01991614255765, "grad_norm": 0.1684286892414093, "learning_rate": 3.6525420054774884e-07, "loss": 0.5689, "num_input_tokens_seen": 23746472, "step": 36290 }, { "epoch": 19.02253668763103, "grad_norm": 0.1571873277425766, "learning_rate": 3.6330935777957597e-07, "loss": 0.3479, "num_input_tokens_seen": 23749640, "step": 36295 }, { "epoch": 19.0251572327044, "grad_norm": 0.11285557597875595, "learning_rate": 3.613696687260193e-07, "loss": 0.5365, "num_input_tokens_seen": 23752328, "step": 36300 }, { "epoch": 19.02777777777778, "grad_norm": 0.15304675698280334, "learning_rate": 3.59435133792832e-07, "loss": 0.4426, "num_input_tokens_seen": 23756392, "step": 36305 }, { "epoch": 19.030398322851152, "grad_norm": 0.16987065970897675, "learning_rate": 3.575057533847043e-07, "loss": 0.562, "num_input_tokens_seen": 23759784, "step": 36310 }, { "epoch": 19.03301886792453, "grad_norm": 0.12068280577659607, "learning_rate": 3.555815279052327e-07, "loss": 0.4906, "num_input_tokens_seen": 23762408, "step": 36315 }, { "epoch": 19.035639412997902, "grad_norm": 0.11133819818496704, "learning_rate": 3.5366245775694805e-07, "loss": 0.5064, "num_input_tokens_seen": 23765960, "step": 36320 }, { "epoch": 19.03825995807128, "grad_norm": 0.16936160624027252, "learning_rate": 3.517485433412987e-07, "loss": 0.5425, "num_input_tokens_seen": 23769032, "step": 36325 }, { "epoch": 19.040880503144653, "grad_norm": 0.18867641687393188, "learning_rate": 3.4983978505865044e-07, "loss": 0.4153, "num_input_tokens_seen": 23771528, "step": 36330 }, { "epoch": 19.04350104821803, "grad_norm": 0.1421738862991333, "learning_rate": 3.47936183308295e-07, "loss": 0.551, "num_input_tokens_seen": 23775400, "step": 36335 }, { "epoch": 19.046121593291403, "grad_norm": 0.20928950607776642, "learning_rate": 3.460377384884472e-07, "loss": 0.4604, "num_input_tokens_seen": 23778536, "step": 36340 }, { "epoch": 19.04874213836478, "grad_norm": 0.12507405877113342, "learning_rate": 3.4414445099623384e-07, "loss": 0.4593, "num_input_tokens_seen": 23781832, "step": 36345 }, { "epoch": 19.051362683438157, "grad_norm": 0.20078034698963165, "learning_rate": 3.42256321227713e-07, "loss": 0.5351, "num_input_tokens_seen": 23789288, "step": 36350 }, { "epoch": 19.05398322851153, "grad_norm": 0.13138268887996674, "learning_rate": 3.403733495778605e-07, "loss": 0.4831, "num_input_tokens_seen": 23792296, "step": 36355 }, { "epoch": 19.056603773584907, "grad_norm": 0.11497970670461655, "learning_rate": 3.3849553644056955e-07, "loss": 0.4757, "num_input_tokens_seen": 23795464, "step": 36360 }, { "epoch": 19.05922431865828, "grad_norm": 0.1541260927915573, "learning_rate": 3.36622882208662e-07, "loss": 0.5705, "num_input_tokens_seen": 23798504, "step": 36365 }, { "epoch": 19.061844863731658, "grad_norm": 0.18062898516654968, "learning_rate": 3.347553872738718e-07, "loss": 0.5393, "num_input_tokens_seen": 23802024, "step": 36370 }, { "epoch": 19.06446540880503, "grad_norm": 0.16167014837265015, "learning_rate": 3.3289305202685583e-07, "loss": 0.3599, "num_input_tokens_seen": 23805512, "step": 36375 }, { "epoch": 19.067085953878408, "grad_norm": 0.17663192749023438, "learning_rate": 3.310358768571997e-07, "loss": 0.4211, "num_input_tokens_seen": 23808712, "step": 36380 }, { "epoch": 19.06970649895178, "grad_norm": 0.17067895829677582, "learning_rate": 3.2918386215339827e-07, "loss": 0.4264, "num_input_tokens_seen": 23814280, "step": 36385 }, { "epoch": 19.072327044025158, "grad_norm": 0.268686980009079, "learning_rate": 3.2733700830287205e-07, "loss": 0.4612, "num_input_tokens_seen": 23817448, "step": 36390 }, { "epoch": 19.07494758909853, "grad_norm": 0.1893751472234726, "learning_rate": 3.254953156919649e-07, "loss": 0.5679, "num_input_tokens_seen": 23820520, "step": 36395 }, { "epoch": 19.07756813417191, "grad_norm": 0.03376735374331474, "learning_rate": 3.2365878470593527e-07, "loss": 0.3404, "num_input_tokens_seen": 23826664, "step": 36400 }, { "epoch": 19.080188679245282, "grad_norm": 0.09550774097442627, "learning_rate": 3.2182741572896747e-07, "loss": 0.3966, "num_input_tokens_seen": 23830120, "step": 36405 }, { "epoch": 19.08280922431866, "grad_norm": 0.1367565095424652, "learning_rate": 3.2000120914415786e-07, "loss": 0.6182, "num_input_tokens_seen": 23833480, "step": 36410 }, { "epoch": 19.085429769392032, "grad_norm": 0.19932705163955688, "learning_rate": 3.181801653335315e-07, "loss": 0.5475, "num_input_tokens_seen": 23836712, "step": 36415 }, { "epoch": 19.08805031446541, "grad_norm": 0.32542675733566284, "learning_rate": 3.163642846780307e-07, "loss": 0.6484, "num_input_tokens_seen": 23840072, "step": 36420 }, { "epoch": 19.090670859538783, "grad_norm": 0.15066921710968018, "learning_rate": 3.1455356755751565e-07, "loss": 0.5286, "num_input_tokens_seen": 23843176, "step": 36425 }, { "epoch": 19.09329140461216, "grad_norm": 0.08954090625047684, "learning_rate": 3.127480143507666e-07, "loss": 0.4393, "num_input_tokens_seen": 23846344, "step": 36430 }, { "epoch": 19.095911949685533, "grad_norm": 0.14365844428539276, "learning_rate": 3.1094762543548704e-07, "loss": 0.4332, "num_input_tokens_seen": 23850216, "step": 36435 }, { "epoch": 19.09853249475891, "grad_norm": 0.41772326827049255, "learning_rate": 3.091524011882951e-07, "loss": 0.632, "num_input_tokens_seen": 23853352, "step": 36440 }, { "epoch": 19.101153039832287, "grad_norm": 0.11029034852981567, "learning_rate": 3.073623419847349e-07, "loss": 0.5352, "num_input_tokens_seen": 23855912, "step": 36445 }, { "epoch": 19.10377358490566, "grad_norm": 0.20869427919387817, "learning_rate": 3.0557744819926235e-07, "loss": 0.3299, "num_input_tokens_seen": 23858120, "step": 36450 }, { "epoch": 19.106394129979037, "grad_norm": 0.16307806968688965, "learning_rate": 3.037977202052594e-07, "loss": 0.4709, "num_input_tokens_seen": 23861736, "step": 36455 }, { "epoch": 19.10901467505241, "grad_norm": 0.15456242859363556, "learning_rate": 3.020231583750255e-07, "loss": 0.3435, "num_input_tokens_seen": 23865864, "step": 36460 }, { "epoch": 19.111635220125788, "grad_norm": 0.2676209509372711, "learning_rate": 3.002537630797747e-07, "loss": 0.466, "num_input_tokens_seen": 23869160, "step": 36465 }, { "epoch": 19.11425576519916, "grad_norm": 0.1400267779827118, "learning_rate": 2.984895346896499e-07, "loss": 0.4175, "num_input_tokens_seen": 23873064, "step": 36470 }, { "epoch": 19.116876310272538, "grad_norm": 0.10097195208072662, "learning_rate": 2.967304735737031e-07, "loss": 0.4567, "num_input_tokens_seen": 23876328, "step": 36475 }, { "epoch": 19.11949685534591, "grad_norm": 0.10729735344648361, "learning_rate": 2.9497658009991214e-07, "loss": 0.4815, "num_input_tokens_seen": 23878984, "step": 36480 }, { "epoch": 19.122117400419288, "grad_norm": 0.1428992748260498, "learning_rate": 2.932278546351724e-07, "loss": 0.3391, "num_input_tokens_seen": 23881800, "step": 36485 }, { "epoch": 19.12473794549266, "grad_norm": 0.16122370958328247, "learning_rate": 2.9148429754529137e-07, "loss": 0.4588, "num_input_tokens_seen": 23884296, "step": 36490 }, { "epoch": 19.12735849056604, "grad_norm": 0.3736553192138672, "learning_rate": 2.897459091950078e-07, "loss": 0.7303, "num_input_tokens_seen": 23887560, "step": 36495 }, { "epoch": 19.129979035639412, "grad_norm": 0.1416180282831192, "learning_rate": 2.880126899479724e-07, "loss": 0.3847, "num_input_tokens_seen": 23890408, "step": 36500 }, { "epoch": 19.13259958071279, "grad_norm": 0.18528926372528076, "learning_rate": 2.862846401667507e-07, "loss": 0.5199, "num_input_tokens_seen": 23893288, "step": 36505 }, { "epoch": 19.135220125786162, "grad_norm": 0.26803261041641235, "learning_rate": 2.8456176021283697e-07, "loss": 0.6767, "num_input_tokens_seen": 23897000, "step": 36510 }, { "epoch": 19.13784067085954, "grad_norm": 0.1635769158601761, "learning_rate": 2.8284405044662897e-07, "loss": 0.4875, "num_input_tokens_seen": 23900392, "step": 36515 }, { "epoch": 19.140461215932913, "grad_norm": 0.09965407848358154, "learning_rate": 2.8113151122745874e-07, "loss": 0.3642, "num_input_tokens_seen": 23904520, "step": 36520 }, { "epoch": 19.14308176100629, "grad_norm": 0.21155573427677155, "learning_rate": 2.7942414291357023e-07, "loss": 0.5249, "num_input_tokens_seen": 23907560, "step": 36525 }, { "epoch": 19.145702306079663, "grad_norm": 0.18746395409107208, "learning_rate": 2.777219458621222e-07, "loss": 0.5192, "num_input_tokens_seen": 23911784, "step": 36530 }, { "epoch": 19.14832285115304, "grad_norm": 0.13312575221061707, "learning_rate": 2.7602492042919923e-07, "loss": 0.4598, "num_input_tokens_seen": 23914952, "step": 36535 }, { "epoch": 19.150943396226417, "grad_norm": 0.1458963006734848, "learning_rate": 2.7433306696979523e-07, "loss": 0.5141, "num_input_tokens_seen": 23918408, "step": 36540 }, { "epoch": 19.15356394129979, "grad_norm": 0.18890805542469025, "learning_rate": 2.7264638583782973e-07, "loss": 0.5217, "num_input_tokens_seen": 23921896, "step": 36545 }, { "epoch": 19.156184486373167, "grad_norm": 0.14153039455413818, "learning_rate": 2.7096487738613177e-07, "loss": 0.459, "num_input_tokens_seen": 23925512, "step": 36550 }, { "epoch": 19.15880503144654, "grad_norm": 0.1903965175151825, "learning_rate": 2.6928854196646147e-07, "loss": 0.4737, "num_input_tokens_seen": 23928072, "step": 36555 }, { "epoch": 19.161425576519918, "grad_norm": 0.15473511815071106, "learning_rate": 2.6761737992948565e-07, "loss": 0.4539, "num_input_tokens_seen": 23930632, "step": 36560 }, { "epoch": 19.16404612159329, "grad_norm": 0.1434604972600937, "learning_rate": 2.6595139162479134e-07, "loss": 0.4207, "num_input_tokens_seen": 23934408, "step": 36565 }, { "epoch": 19.166666666666668, "grad_norm": 0.13700918853282928, "learning_rate": 2.6429057740088303e-07, "loss": 0.3157, "num_input_tokens_seen": 23937768, "step": 36570 }, { "epoch": 19.16928721174004, "grad_norm": 0.12325067073106766, "learning_rate": 2.6263493760518564e-07, "loss": 0.3811, "num_input_tokens_seen": 23941288, "step": 36575 }, { "epoch": 19.171907756813418, "grad_norm": 0.21319293975830078, "learning_rate": 2.6098447258404157e-07, "loss": 0.3928, "num_input_tokens_seen": 23945384, "step": 36580 }, { "epoch": 19.17452830188679, "grad_norm": 0.12853427231311798, "learning_rate": 2.593391826827107e-07, "loss": 0.4124, "num_input_tokens_seen": 23948680, "step": 36585 }, { "epoch": 19.17714884696017, "grad_norm": 0.23971770703792572, "learning_rate": 2.5769906824536225e-07, "loss": 0.4823, "num_input_tokens_seen": 23951912, "step": 36590 }, { "epoch": 19.179769392033542, "grad_norm": 0.19852417707443237, "learning_rate": 2.560641296150967e-07, "loss": 0.4955, "num_input_tokens_seen": 23954568, "step": 36595 }, { "epoch": 19.18238993710692, "grad_norm": 0.13259416818618774, "learning_rate": 2.544343671339211e-07, "loss": 0.65, "num_input_tokens_seen": 23958376, "step": 36600 }, { "epoch": 19.185010482180292, "grad_norm": 0.17097625136375427, "learning_rate": 2.5280978114276264e-07, "loss": 0.5459, "num_input_tokens_seen": 23961768, "step": 36605 }, { "epoch": 19.18763102725367, "grad_norm": 0.16998961567878723, "learning_rate": 2.511903719814662e-07, "loss": 0.5298, "num_input_tokens_seen": 23964936, "step": 36610 }, { "epoch": 19.190251572327043, "grad_norm": 0.15922799706459045, "learning_rate": 2.4957613998879695e-07, "loss": 0.5607, "num_input_tokens_seen": 23967976, "step": 36615 }, { "epoch": 19.19287211740042, "grad_norm": 0.18966972827911377, "learning_rate": 2.479670855024291e-07, "loss": 0.48, "num_input_tokens_seen": 23970984, "step": 36620 }, { "epoch": 19.195492662473793, "grad_norm": 0.19564729928970337, "learning_rate": 2.463632088589629e-07, "loss": 0.408, "num_input_tokens_seen": 23975400, "step": 36625 }, { "epoch": 19.19811320754717, "grad_norm": 0.2675345242023468, "learning_rate": 2.447645103939078e-07, "loss": 0.5727, "num_input_tokens_seen": 23978504, "step": 36630 }, { "epoch": 19.200733752620547, "grad_norm": 0.3485344648361206, "learning_rate": 2.431709904416962e-07, "loss": 0.4116, "num_input_tokens_seen": 23981992, "step": 36635 }, { "epoch": 19.20335429769392, "grad_norm": 0.1378069519996643, "learning_rate": 2.4158264933567264e-07, "loss": 0.4596, "num_input_tokens_seen": 23985416, "step": 36640 }, { "epoch": 19.205974842767297, "grad_norm": 0.2440204620361328, "learning_rate": 2.399994874080963e-07, "loss": 0.5142, "num_input_tokens_seen": 23988328, "step": 36645 }, { "epoch": 19.20859538784067, "grad_norm": 0.17354321479797363, "learning_rate": 2.3842150499015504e-07, "loss": 0.5701, "num_input_tokens_seen": 23991464, "step": 36650 }, { "epoch": 19.211215932914047, "grad_norm": 0.17138494551181793, "learning_rate": 2.368487024119348e-07, "loss": 0.5009, "num_input_tokens_seen": 23994536, "step": 36655 }, { "epoch": 19.21383647798742, "grad_norm": 0.2245580106973648, "learning_rate": 2.3528108000245573e-07, "loss": 0.4012, "num_input_tokens_seen": 23998472, "step": 36660 }, { "epoch": 19.216457023060798, "grad_norm": 0.15397199988365173, "learning_rate": 2.3371863808964445e-07, "loss": 0.4634, "num_input_tokens_seen": 24001384, "step": 36665 }, { "epoch": 19.21907756813417, "grad_norm": 0.15245260298252106, "learning_rate": 2.32161377000345e-07, "loss": 0.5505, "num_input_tokens_seen": 24004712, "step": 36670 }, { "epoch": 19.221698113207548, "grad_norm": 0.2605886161327362, "learning_rate": 2.306092970603163e-07, "loss": 0.467, "num_input_tokens_seen": 24007144, "step": 36675 }, { "epoch": 19.22431865828092, "grad_norm": 0.2938864827156067, "learning_rate": 2.2906239859424306e-07, "loss": 0.4605, "num_input_tokens_seen": 24010056, "step": 36680 }, { "epoch": 19.2269392033543, "grad_norm": 0.117266945540905, "learning_rate": 2.2752068192571084e-07, "loss": 0.4115, "num_input_tokens_seen": 24013352, "step": 36685 }, { "epoch": 19.229559748427672, "grad_norm": 0.28976210951805115, "learning_rate": 2.2598414737723396e-07, "loss": 0.4611, "num_input_tokens_seen": 24016488, "step": 36690 }, { "epoch": 19.23218029350105, "grad_norm": 0.20156365633010864, "learning_rate": 2.2445279527023578e-07, "loss": 0.6488, "num_input_tokens_seen": 24020008, "step": 36695 }, { "epoch": 19.234800838574422, "grad_norm": 0.23837970197200775, "learning_rate": 2.2292662592506007e-07, "loss": 0.5073, "num_input_tokens_seen": 24022472, "step": 36700 }, { "epoch": 19.2374213836478, "grad_norm": 0.12204676121473312, "learning_rate": 2.2140563966095985e-07, "loss": 0.55, "num_input_tokens_seen": 24026856, "step": 36705 }, { "epoch": 19.240041928721173, "grad_norm": 0.153603196144104, "learning_rate": 2.1988983679611386e-07, "loss": 0.4884, "num_input_tokens_seen": 24030376, "step": 36710 }, { "epoch": 19.24266247379455, "grad_norm": 0.19705234467983246, "learning_rate": 2.1837921764760462e-07, "loss": 0.4456, "num_input_tokens_seen": 24033384, "step": 36715 }, { "epoch": 19.245283018867923, "grad_norm": 0.19040322303771973, "learning_rate": 2.1687378253144043e-07, "loss": 0.3881, "num_input_tokens_seen": 24039144, "step": 36720 }, { "epoch": 19.2479035639413, "grad_norm": 0.141669362783432, "learning_rate": 2.153735317625416e-07, "loss": 0.4036, "num_input_tokens_seen": 24043112, "step": 36725 }, { "epoch": 19.250524109014677, "grad_norm": 0.15746209025382996, "learning_rate": 2.1387846565474045e-07, "loss": 0.5425, "num_input_tokens_seen": 24046824, "step": 36730 }, { "epoch": 19.25314465408805, "grad_norm": 0.17442162334918976, "learning_rate": 2.123885845207896e-07, "loss": 0.5427, "num_input_tokens_seen": 24049576, "step": 36735 }, { "epoch": 19.255765199161427, "grad_norm": 0.09752263128757477, "learning_rate": 2.1090388867235356e-07, "loss": 0.5465, "num_input_tokens_seen": 24052456, "step": 36740 }, { "epoch": 19.2583857442348, "grad_norm": 0.24528375267982483, "learning_rate": 2.0942437842001728e-07, "loss": 0.3924, "num_input_tokens_seen": 24055272, "step": 36745 }, { "epoch": 19.261006289308177, "grad_norm": 0.15038833022117615, "learning_rate": 2.0795005407327483e-07, "loss": 0.4238, "num_input_tokens_seen": 24058728, "step": 36750 }, { "epoch": 19.26362683438155, "grad_norm": 0.14850656688213348, "learning_rate": 2.0648091594053508e-07, "loss": 0.5196, "num_input_tokens_seen": 24061896, "step": 36755 }, { "epoch": 19.266247379454928, "grad_norm": 0.2814099192619324, "learning_rate": 2.0501696432913275e-07, "loss": 0.6217, "num_input_tokens_seen": 24065000, "step": 36760 }, { "epoch": 19.2688679245283, "grad_norm": 0.17552830278873444, "learning_rate": 2.035581995453034e-07, "loss": 0.4823, "num_input_tokens_seen": 24068264, "step": 36765 }, { "epoch": 19.271488469601678, "grad_norm": 0.26492705941200256, "learning_rate": 2.0210462189420577e-07, "loss": 0.4746, "num_input_tokens_seen": 24072008, "step": 36770 }, { "epoch": 19.27410901467505, "grad_norm": 0.0996968224644661, "learning_rate": 2.006562316799132e-07, "loss": 0.4881, "num_input_tokens_seen": 24075752, "step": 36775 }, { "epoch": 19.27672955974843, "grad_norm": 0.15846814215183258, "learning_rate": 1.992130292054084e-07, "loss": 0.4259, "num_input_tokens_seen": 24079528, "step": 36780 }, { "epoch": 19.279350104821802, "grad_norm": 0.11932676285505295, "learning_rate": 1.9777501477259708e-07, "loss": 0.4249, "num_input_tokens_seen": 24085544, "step": 36785 }, { "epoch": 19.28197064989518, "grad_norm": 0.25201937556266785, "learning_rate": 1.9634218868229414e-07, "loss": 0.4966, "num_input_tokens_seen": 24089480, "step": 36790 }, { "epoch": 19.284591194968552, "grad_norm": 0.17730781435966492, "learning_rate": 1.9491455123423207e-07, "loss": 0.3911, "num_input_tokens_seen": 24092968, "step": 36795 }, { "epoch": 19.28721174004193, "grad_norm": 0.11458392441272736, "learning_rate": 1.9349210272705253e-07, "loss": 0.392, "num_input_tokens_seen": 24096680, "step": 36800 }, { "epoch": 19.289832285115303, "grad_norm": 0.15314726531505585, "learning_rate": 1.920748434583175e-07, "loss": 0.4517, "num_input_tokens_seen": 24100456, "step": 36805 }, { "epoch": 19.29245283018868, "grad_norm": 0.19001375138759613, "learning_rate": 1.906627737245009e-07, "loss": 0.505, "num_input_tokens_seen": 24103816, "step": 36810 }, { "epoch": 19.295073375262053, "grad_norm": 0.16121095418930054, "learning_rate": 1.8925589382099152e-07, "loss": 0.4523, "num_input_tokens_seen": 24107752, "step": 36815 }, { "epoch": 19.29769392033543, "grad_norm": 0.1794433444738388, "learning_rate": 1.8785420404209553e-07, "loss": 0.4871, "num_input_tokens_seen": 24110856, "step": 36820 }, { "epoch": 19.300314465408803, "grad_norm": 0.2179066687822342, "learning_rate": 1.864577046810284e-07, "loss": 0.5915, "num_input_tokens_seen": 24113896, "step": 36825 }, { "epoch": 19.30293501048218, "grad_norm": 0.20105616748332977, "learning_rate": 1.8506639602992038e-07, "loss": 0.5168, "num_input_tokens_seen": 24116872, "step": 36830 }, { "epoch": 19.305555555555557, "grad_norm": 0.10366977006196976, "learning_rate": 1.836802783798164e-07, "loss": 0.4033, "num_input_tokens_seen": 24120072, "step": 36835 }, { "epoch": 19.30817610062893, "grad_norm": 0.11133383214473724, "learning_rate": 1.822993520206817e-07, "loss": 0.4005, "num_input_tokens_seen": 24124232, "step": 36840 }, { "epoch": 19.310796645702307, "grad_norm": 0.14515843987464905, "learning_rate": 1.8092361724138528e-07, "loss": 0.4226, "num_input_tokens_seen": 24127720, "step": 36845 }, { "epoch": 19.31341719077568, "grad_norm": 0.17483580112457275, "learning_rate": 1.7955307432971914e-07, "loss": 0.5214, "num_input_tokens_seen": 24131624, "step": 36850 }, { "epoch": 19.316037735849058, "grad_norm": 0.1951581984758377, "learning_rate": 1.7818772357238445e-07, "loss": 0.3922, "num_input_tokens_seen": 24134568, "step": 36855 }, { "epoch": 19.31865828092243, "grad_norm": 0.24406923353672028, "learning_rate": 1.7682756525499166e-07, "loss": 0.3928, "num_input_tokens_seen": 24137768, "step": 36860 }, { "epoch": 19.321278825995808, "grad_norm": 0.3141036927700043, "learning_rate": 1.7547259966207708e-07, "loss": 0.5052, "num_input_tokens_seen": 24142184, "step": 36865 }, { "epoch": 19.32389937106918, "grad_norm": 0.15831489861011505, "learning_rate": 1.7412282707708062e-07, "loss": 0.4631, "num_input_tokens_seen": 24145192, "step": 36870 }, { "epoch": 19.32651991614256, "grad_norm": 0.15840449929237366, "learning_rate": 1.7277824778235974e-07, "loss": 0.4785, "num_input_tokens_seen": 24148648, "step": 36875 }, { "epoch": 19.329140461215932, "grad_norm": 0.08700932562351227, "learning_rate": 1.714388620591867e-07, "loss": 0.3744, "num_input_tokens_seen": 24152616, "step": 36880 }, { "epoch": 19.33176100628931, "grad_norm": 0.15420298278331757, "learning_rate": 1.701046701877429e-07, "loss": 0.5781, "num_input_tokens_seen": 24155272, "step": 36885 }, { "epoch": 19.334381551362682, "grad_norm": 0.08712145686149597, "learning_rate": 1.6877567244713011e-07, "loss": 0.4834, "num_input_tokens_seen": 24158536, "step": 36890 }, { "epoch": 19.33700209643606, "grad_norm": 0.10123677551746368, "learning_rate": 1.6745186911535372e-07, "loss": 0.5215, "num_input_tokens_seen": 24163144, "step": 36895 }, { "epoch": 19.339622641509433, "grad_norm": 0.1794186681509018, "learning_rate": 1.6613326046934497e-07, "loss": 0.4432, "num_input_tokens_seen": 24165864, "step": 36900 }, { "epoch": 19.34224318658281, "grad_norm": 0.12692898511886597, "learning_rate": 1.64819846784936e-07, "loss": 0.47, "num_input_tokens_seen": 24169032, "step": 36905 }, { "epoch": 19.344863731656183, "grad_norm": 0.21299102902412415, "learning_rate": 1.6351162833688204e-07, "loss": 0.5498, "num_input_tokens_seen": 24171592, "step": 36910 }, { "epoch": 19.34748427672956, "grad_norm": 0.1441163718700409, "learning_rate": 1.6220860539884475e-07, "loss": 0.4982, "num_input_tokens_seen": 24174824, "step": 36915 }, { "epoch": 19.350104821802937, "grad_norm": 0.2125617265701294, "learning_rate": 1.6091077824340327e-07, "loss": 0.6142, "num_input_tokens_seen": 24177864, "step": 36920 }, { "epoch": 19.35272536687631, "grad_norm": 0.11449088901281357, "learning_rate": 1.5961814714204605e-07, "loss": 0.4501, "num_input_tokens_seen": 24181288, "step": 36925 }, { "epoch": 19.355345911949687, "grad_norm": 0.1502729207277298, "learning_rate": 1.5833071236518172e-07, "loss": 0.6221, "num_input_tokens_seen": 24184680, "step": 36930 }, { "epoch": 19.35796645702306, "grad_norm": 0.1560410112142563, "learning_rate": 1.5704847418212266e-07, "loss": 0.3714, "num_input_tokens_seen": 24188072, "step": 36935 }, { "epoch": 19.360587002096437, "grad_norm": 0.2135968655347824, "learning_rate": 1.5577143286109875e-07, "loss": 0.4497, "num_input_tokens_seen": 24190664, "step": 36940 }, { "epoch": 19.36320754716981, "grad_norm": 0.12031497806310654, "learning_rate": 1.5449958866925185e-07, "loss": 0.2737, "num_input_tokens_seen": 24193832, "step": 36945 }, { "epoch": 19.365828092243188, "grad_norm": 0.17087769508361816, "learning_rate": 1.532329418726386e-07, "loss": 0.4937, "num_input_tokens_seen": 24197256, "step": 36950 }, { "epoch": 19.36844863731656, "grad_norm": 0.1879883110523224, "learning_rate": 1.5197149273623036e-07, "loss": 0.4302, "num_input_tokens_seen": 24199816, "step": 36955 }, { "epoch": 19.371069182389938, "grad_norm": 0.1401451677083969, "learning_rate": 1.5071524152389936e-07, "loss": 0.5574, "num_input_tokens_seen": 24203272, "step": 36960 }, { "epoch": 19.37368972746331, "grad_norm": 0.06559418141841888, "learning_rate": 1.4946418849844934e-07, "loss": 0.3884, "num_input_tokens_seen": 24206728, "step": 36965 }, { "epoch": 19.37631027253669, "grad_norm": 0.2195548713207245, "learning_rate": 1.4821833392157923e-07, "loss": 0.4121, "num_input_tokens_seen": 24209416, "step": 36970 }, { "epoch": 19.378930817610062, "grad_norm": 0.1835935264825821, "learning_rate": 1.4697767805390838e-07, "loss": 0.5397, "num_input_tokens_seen": 24212456, "step": 36975 }, { "epoch": 19.38155136268344, "grad_norm": 0.19496558606624603, "learning_rate": 1.4574222115497084e-07, "loss": 0.5188, "num_input_tokens_seen": 24214856, "step": 36980 }, { "epoch": 19.384171907756812, "grad_norm": 0.15437878668308258, "learning_rate": 1.4451196348320706e-07, "loss": 0.4634, "num_input_tokens_seen": 24218024, "step": 36985 }, { "epoch": 19.38679245283019, "grad_norm": 0.14962178468704224, "learning_rate": 1.4328690529597232e-07, "loss": 0.474, "num_input_tokens_seen": 24220808, "step": 36990 }, { "epoch": 19.389412997903563, "grad_norm": 0.12157072871923447, "learning_rate": 1.4206704684953943e-07, "loss": 0.6322, "num_input_tokens_seen": 24223816, "step": 36995 }, { "epoch": 19.39203354297694, "grad_norm": 0.08949358016252518, "learning_rate": 1.4085238839908476e-07, "loss": 0.453, "num_input_tokens_seen": 24227464, "step": 37000 }, { "epoch": 19.394654088050313, "grad_norm": 0.11085688322782516, "learning_rate": 1.3964293019870235e-07, "loss": 0.5666, "num_input_tokens_seen": 24230376, "step": 37005 }, { "epoch": 19.39727463312369, "grad_norm": 0.20240609347820282, "learning_rate": 1.3843867250139807e-07, "loss": 0.4607, "num_input_tokens_seen": 24232744, "step": 37010 }, { "epoch": 19.399895178197063, "grad_norm": 0.12557530403137207, "learning_rate": 1.3723961555908715e-07, "loss": 0.4721, "num_input_tokens_seen": 24235848, "step": 37015 }, { "epoch": 19.40251572327044, "grad_norm": 0.16675075888633728, "learning_rate": 1.3604575962259947e-07, "loss": 0.4993, "num_input_tokens_seen": 24239912, "step": 37020 }, { "epoch": 19.405136268343817, "grad_norm": 0.1257612258195877, "learning_rate": 1.3485710494167692e-07, "loss": 0.6632, "num_input_tokens_seen": 24243784, "step": 37025 }, { "epoch": 19.40775681341719, "grad_norm": 0.14047859609127045, "learning_rate": 1.3367365176496783e-07, "loss": 0.4007, "num_input_tokens_seen": 24246792, "step": 37030 }, { "epoch": 19.410377358490567, "grad_norm": 0.10793323814868927, "learning_rate": 1.324954003400436e-07, "loss": 0.405, "num_input_tokens_seen": 24249704, "step": 37035 }, { "epoch": 19.41299790356394, "grad_norm": 0.16463510692119598, "learning_rate": 1.313223509133793e-07, "loss": 0.4523, "num_input_tokens_seen": 24252616, "step": 37040 }, { "epoch": 19.415618448637318, "grad_norm": 0.15308398008346558, "learning_rate": 1.301545037303592e-07, "loss": 0.4049, "num_input_tokens_seen": 24256136, "step": 37045 }, { "epoch": 19.41823899371069, "grad_norm": 0.16975037753582, "learning_rate": 1.2899185903529066e-07, "loss": 0.4737, "num_input_tokens_seen": 24259400, "step": 37050 }, { "epoch": 19.420859538784068, "grad_norm": 0.09433935582637787, "learning_rate": 1.278344170713791e-07, "loss": 0.4073, "num_input_tokens_seen": 24263880, "step": 37055 }, { "epoch": 19.42348008385744, "grad_norm": 0.1970500499010086, "learning_rate": 1.2668217808075033e-07, "loss": 0.4962, "num_input_tokens_seen": 24266888, "step": 37060 }, { "epoch": 19.42610062893082, "grad_norm": 0.2207590639591217, "learning_rate": 1.255351423044393e-07, "loss": 0.5191, "num_input_tokens_seen": 24269768, "step": 37065 }, { "epoch": 19.428721174004192, "grad_norm": 0.18199381232261658, "learning_rate": 1.2439330998239572e-07, "loss": 0.3859, "num_input_tokens_seen": 24272808, "step": 37070 }, { "epoch": 19.43134171907757, "grad_norm": 0.1733158528804779, "learning_rate": 1.2325668135347578e-07, "loss": 0.416, "num_input_tokens_seen": 24275592, "step": 37075 }, { "epoch": 19.433962264150942, "grad_norm": 0.17566703259944916, "learning_rate": 1.221252566554476e-07, "loss": 0.5001, "num_input_tokens_seen": 24278984, "step": 37080 }, { "epoch": 19.43658280922432, "grad_norm": 0.18384389579296112, "learning_rate": 1.2099903612499685e-07, "loss": 0.6115, "num_input_tokens_seen": 24282248, "step": 37085 }, { "epoch": 19.439203354297693, "grad_norm": 0.28172725439071655, "learning_rate": 1.198780199977101e-07, "loss": 0.4655, "num_input_tokens_seen": 24285960, "step": 37090 }, { "epoch": 19.44182389937107, "grad_norm": 0.146259605884552, "learning_rate": 1.1876220850809416e-07, "loss": 0.3805, "num_input_tokens_seen": 24288392, "step": 37095 }, { "epoch": 19.444444444444443, "grad_norm": 0.16033519804477692, "learning_rate": 1.1765160188956514e-07, "loss": 0.4549, "num_input_tokens_seen": 24291880, "step": 37100 }, { "epoch": 19.44706498951782, "grad_norm": 0.13042843341827393, "learning_rate": 1.1654620037445107e-07, "loss": 0.4067, "num_input_tokens_seen": 24295112, "step": 37105 }, { "epoch": 19.449685534591197, "grad_norm": 0.14764460921287537, "learning_rate": 1.1544600419398088e-07, "loss": 0.4779, "num_input_tokens_seen": 24298152, "step": 37110 }, { "epoch": 19.45230607966457, "grad_norm": 0.13975220918655396, "learning_rate": 1.1435101357831213e-07, "loss": 0.403, "num_input_tokens_seen": 24302024, "step": 37115 }, { "epoch": 19.454926624737947, "grad_norm": 0.2413650006055832, "learning_rate": 1.1326122875650047e-07, "loss": 0.5377, "num_input_tokens_seen": 24305480, "step": 37120 }, { "epoch": 19.45754716981132, "grad_norm": 0.14100484549999237, "learning_rate": 1.1217664995651633e-07, "loss": 0.4405, "num_input_tokens_seen": 24309800, "step": 37125 }, { "epoch": 19.460167714884697, "grad_norm": 0.21374161541461945, "learning_rate": 1.1109727740523935e-07, "loss": 0.5007, "num_input_tokens_seen": 24313096, "step": 37130 }, { "epoch": 19.46278825995807, "grad_norm": 0.15396039187908173, "learning_rate": 1.1002311132846944e-07, "loss": 0.5675, "num_input_tokens_seen": 24315784, "step": 37135 }, { "epoch": 19.465408805031448, "grad_norm": 0.24052539467811584, "learning_rate": 1.0895415195090186e-07, "loss": 0.534, "num_input_tokens_seen": 24318440, "step": 37140 }, { "epoch": 19.46802935010482, "grad_norm": 0.14894290268421173, "learning_rate": 1.0789039949615221e-07, "loss": 0.5429, "num_input_tokens_seen": 24321544, "step": 37145 }, { "epoch": 19.470649895178198, "grad_norm": 0.1299264281988144, "learning_rate": 1.0683185418674801e-07, "loss": 0.4693, "num_input_tokens_seen": 24325192, "step": 37150 }, { "epoch": 19.47327044025157, "grad_norm": 0.18159228563308716, "learning_rate": 1.0577851624412604e-07, "loss": 0.5168, "num_input_tokens_seen": 24328072, "step": 37155 }, { "epoch": 19.47589098532495, "grad_norm": 0.2035619616508484, "learning_rate": 1.0473038588862672e-07, "loss": 0.3668, "num_input_tokens_seen": 24331208, "step": 37160 }, { "epoch": 19.478511530398322, "grad_norm": 0.14214397966861725, "learning_rate": 1.0368746333951351e-07, "loss": 0.3963, "num_input_tokens_seen": 24334600, "step": 37165 }, { "epoch": 19.4811320754717, "grad_norm": 0.14401687681674957, "learning_rate": 1.0264974881494805e-07, "loss": 0.6235, "num_input_tokens_seen": 24338120, "step": 37170 }, { "epoch": 19.483752620545072, "grad_norm": 0.13079531490802765, "learning_rate": 1.0161724253201222e-07, "loss": 0.4807, "num_input_tokens_seen": 24341672, "step": 37175 }, { "epoch": 19.48637316561845, "grad_norm": 0.2559153139591217, "learning_rate": 1.0058994470669436e-07, "loss": 0.5723, "num_input_tokens_seen": 24345512, "step": 37180 }, { "epoch": 19.488993710691823, "grad_norm": 0.1311088502407074, "learning_rate": 9.956785555388926e-08, "loss": 0.4934, "num_input_tokens_seen": 24348616, "step": 37185 }, { "epoch": 19.4916142557652, "grad_norm": 0.13562019169330597, "learning_rate": 9.855097528741475e-08, "loss": 0.4139, "num_input_tokens_seen": 24351496, "step": 37190 }, { "epoch": 19.494234800838573, "grad_norm": 0.12686660885810852, "learning_rate": 9.753930411998402e-08, "loss": 0.4772, "num_input_tokens_seen": 24354248, "step": 37195 }, { "epoch": 19.49685534591195, "grad_norm": 0.2592089772224426, "learning_rate": 9.653284226322501e-08, "loss": 0.6193, "num_input_tokens_seen": 24357320, "step": 37200 }, { "epoch": 19.499475890985323, "grad_norm": 0.1381346583366394, "learning_rate": 9.553158992768318e-08, "loss": 0.571, "num_input_tokens_seen": 24360520, "step": 37205 }, { "epoch": 19.5020964360587, "grad_norm": 0.16734464466571808, "learning_rate": 9.453554732281044e-08, "loss": 0.4414, "num_input_tokens_seen": 24364552, "step": 37210 }, { "epoch": 19.504716981132077, "grad_norm": 0.24486996233463287, "learning_rate": 9.354471465696236e-08, "loss": 0.4461, "num_input_tokens_seen": 24367464, "step": 37215 }, { "epoch": 19.50733752620545, "grad_norm": 0.18948788940906525, "learning_rate": 9.255909213741199e-08, "loss": 0.5275, "num_input_tokens_seen": 24370536, "step": 37220 }, { "epoch": 19.509958071278827, "grad_norm": 0.11290345340967178, "learning_rate": 9.157867997034164e-08, "loss": 0.5964, "num_input_tokens_seen": 24374760, "step": 37225 }, { "epoch": 19.5125786163522, "grad_norm": 0.2513023018836975, "learning_rate": 9.060347836084004e-08, "loss": 0.522, "num_input_tokens_seen": 24377896, "step": 37230 }, { "epoch": 19.515199161425578, "grad_norm": 0.15540871024131775, "learning_rate": 8.96334875129079e-08, "loss": 0.5462, "num_input_tokens_seen": 24380392, "step": 37235 }, { "epoch": 19.51781970649895, "grad_norm": 0.1963898241519928, "learning_rate": 8.866870762946067e-08, "loss": 0.479, "num_input_tokens_seen": 24383272, "step": 37240 }, { "epoch": 19.520440251572328, "grad_norm": 0.23436833918094635, "learning_rate": 8.77091389123147e-08, "loss": 0.6258, "num_input_tokens_seen": 24386600, "step": 37245 }, { "epoch": 19.5230607966457, "grad_norm": 0.17297807335853577, "learning_rate": 8.675478156220385e-08, "loss": 0.4663, "num_input_tokens_seen": 24389832, "step": 37250 }, { "epoch": 19.52568134171908, "grad_norm": 0.1223045289516449, "learning_rate": 8.580563577876566e-08, "loss": 0.4774, "num_input_tokens_seen": 24393096, "step": 37255 }, { "epoch": 19.528301886792452, "grad_norm": 0.1387222856283188, "learning_rate": 8.486170176055242e-08, "loss": 0.4532, "num_input_tokens_seen": 24396008, "step": 37260 }, { "epoch": 19.53092243186583, "grad_norm": 0.10523883253335953, "learning_rate": 8.392297970502282e-08, "loss": 0.5395, "num_input_tokens_seen": 24398632, "step": 37265 }, { "epoch": 19.533542976939202, "grad_norm": 0.12136691808700562, "learning_rate": 8.298946980855315e-08, "loss": 0.5356, "num_input_tokens_seen": 24401992, "step": 37270 }, { "epoch": 19.53616352201258, "grad_norm": 0.15239834785461426, "learning_rate": 8.206117226641497e-08, "loss": 0.5704, "num_input_tokens_seen": 24405384, "step": 37275 }, { "epoch": 19.538784067085953, "grad_norm": 0.22317621111869812, "learning_rate": 8.113808727280015e-08, "loss": 0.561, "num_input_tokens_seen": 24408136, "step": 37280 }, { "epoch": 19.54140461215933, "grad_norm": 0.11090641468763351, "learning_rate": 8.022021502080979e-08, "loss": 0.5988, "num_input_tokens_seen": 24412040, "step": 37285 }, { "epoch": 19.544025157232703, "grad_norm": 0.1676618903875351, "learning_rate": 7.930755570244863e-08, "loss": 0.5125, "num_input_tokens_seen": 24415080, "step": 37290 }, { "epoch": 19.54664570230608, "grad_norm": 0.2664019763469696, "learning_rate": 7.840010950863619e-08, "loss": 0.4847, "num_input_tokens_seen": 24417736, "step": 37295 }, { "epoch": 19.549266247379457, "grad_norm": 0.23617224395275116, "learning_rate": 7.749787662920116e-08, "loss": 0.5832, "num_input_tokens_seen": 24421224, "step": 37300 }, { "epoch": 19.55188679245283, "grad_norm": 0.11906848847866058, "learning_rate": 7.660085725288146e-08, "loss": 0.4568, "num_input_tokens_seen": 24427304, "step": 37305 }, { "epoch": 19.554507337526207, "grad_norm": 0.1720941960811615, "learning_rate": 7.570905156731867e-08, "loss": 0.5014, "num_input_tokens_seen": 24431080, "step": 37310 }, { "epoch": 19.55712788259958, "grad_norm": 0.1454467475414276, "learning_rate": 7.482245975907188e-08, "loss": 0.5463, "num_input_tokens_seen": 24433960, "step": 37315 }, { "epoch": 19.559748427672957, "grad_norm": 0.10545894503593445, "learning_rate": 7.394108201360383e-08, "loss": 0.405, "num_input_tokens_seen": 24436616, "step": 37320 }, { "epoch": 19.56236897274633, "grad_norm": 0.3340006172657013, "learning_rate": 7.306491851529206e-08, "loss": 0.5261, "num_input_tokens_seen": 24439240, "step": 37325 }, { "epoch": 19.564989517819708, "grad_norm": 0.18657685816287994, "learning_rate": 7.219396944741774e-08, "loss": 0.5544, "num_input_tokens_seen": 24441992, "step": 37330 }, { "epoch": 19.56761006289308, "grad_norm": 0.1288353055715561, "learning_rate": 7.132823499217123e-08, "loss": 0.4461, "num_input_tokens_seen": 24444776, "step": 37335 }, { "epoch": 19.570230607966458, "grad_norm": 0.13630224764347076, "learning_rate": 7.046771533065766e-08, "loss": 0.4847, "num_input_tokens_seen": 24447720, "step": 37340 }, { "epoch": 19.57285115303983, "grad_norm": 0.10074987262487411, "learning_rate": 6.961241064288859e-08, "loss": 0.4665, "num_input_tokens_seen": 24452520, "step": 37345 }, { "epoch": 19.57547169811321, "grad_norm": 0.16902242600917816, "learning_rate": 6.876232110778202e-08, "loss": 0.4469, "num_input_tokens_seen": 24454888, "step": 37350 }, { "epoch": 19.578092243186582, "grad_norm": 0.18638712167739868, "learning_rate": 6.791744690316515e-08, "loss": 0.4302, "num_input_tokens_seen": 24458152, "step": 37355 }, { "epoch": 19.58071278825996, "grad_norm": 0.09941733628511429, "learning_rate": 6.707778820577715e-08, "loss": 0.54, "num_input_tokens_seen": 24460904, "step": 37360 }, { "epoch": 19.583333333333332, "grad_norm": 0.14433208107948303, "learning_rate": 6.624334519126919e-08, "loss": 0.4536, "num_input_tokens_seen": 24464328, "step": 37365 }, { "epoch": 19.58595387840671, "grad_norm": 0.18686780333518982, "learning_rate": 6.541411803419329e-08, "loss": 0.5945, "num_input_tokens_seen": 24468072, "step": 37370 }, { "epoch": 19.588574423480082, "grad_norm": 0.14320817589759827, "learning_rate": 6.459010690801626e-08, "loss": 0.568, "num_input_tokens_seen": 24472424, "step": 37375 }, { "epoch": 19.59119496855346, "grad_norm": 0.21376337110996246, "learning_rate": 6.377131198510855e-08, "loss": 0.3289, "num_input_tokens_seen": 24475656, "step": 37380 }, { "epoch": 19.593815513626833, "grad_norm": 0.19009177386760712, "learning_rate": 6.295773343675537e-08, "loss": 0.3937, "num_input_tokens_seen": 24479368, "step": 37385 }, { "epoch": 19.59643605870021, "grad_norm": 0.1365593671798706, "learning_rate": 6.214937143314837e-08, "loss": 0.3961, "num_input_tokens_seen": 24483048, "step": 37390 }, { "epoch": 19.599056603773583, "grad_norm": 0.139037624001503, "learning_rate": 6.134622614338559e-08, "loss": 0.4829, "num_input_tokens_seen": 24486152, "step": 37395 }, { "epoch": 19.60167714884696, "grad_norm": 0.11047127097845078, "learning_rate": 6.05482977354771e-08, "loss": 0.4359, "num_input_tokens_seen": 24489192, "step": 37400 }, { "epoch": 19.604297693920337, "grad_norm": 0.13069336116313934, "learning_rate": 5.975558637634215e-08, "loss": 0.5636, "num_input_tokens_seen": 24492392, "step": 37405 }, { "epoch": 19.60691823899371, "grad_norm": 0.2937853932380676, "learning_rate": 5.896809223180366e-08, "loss": 0.4611, "num_input_tokens_seen": 24494760, "step": 37410 }, { "epoch": 19.609538784067087, "grad_norm": 0.21550868451595306, "learning_rate": 5.8185815466596516e-08, "loss": 0.4796, "num_input_tokens_seen": 24497672, "step": 37415 }, { "epoch": 19.61215932914046, "grad_norm": 0.12538066506385803, "learning_rate": 5.7408756244370365e-08, "loss": 0.483, "num_input_tokens_seen": 24501704, "step": 37420 }, { "epoch": 19.614779874213838, "grad_norm": 0.13043677806854248, "learning_rate": 5.663691472766741e-08, "loss": 0.4112, "num_input_tokens_seen": 24505576, "step": 37425 }, { "epoch": 19.61740041928721, "grad_norm": 0.1991686224937439, "learning_rate": 5.587029107795572e-08, "loss": 0.5525, "num_input_tokens_seen": 24508616, "step": 37430 }, { "epoch": 19.620020964360588, "grad_norm": 0.13780008256435394, "learning_rate": 5.510888545559867e-08, "loss": 0.4939, "num_input_tokens_seen": 24511528, "step": 37435 }, { "epoch": 19.62264150943396, "grad_norm": 0.18331186473369598, "learning_rate": 5.435269801987997e-08, "loss": 0.5954, "num_input_tokens_seen": 24514632, "step": 37440 }, { "epoch": 19.62526205450734, "grad_norm": 0.18091078102588654, "learning_rate": 5.360172892897863e-08, "loss": 0.54, "num_input_tokens_seen": 24518088, "step": 37445 }, { "epoch": 19.627882599580712, "grad_norm": 0.17323866486549377, "learning_rate": 5.285597833999401e-08, "loss": 0.4757, "num_input_tokens_seen": 24520904, "step": 37450 }, { "epoch": 19.63050314465409, "grad_norm": 0.19055086374282837, "learning_rate": 5.2115446408926295e-08, "loss": 0.4708, "num_input_tokens_seen": 24524104, "step": 37455 }, { "epoch": 19.633123689727462, "grad_norm": 0.19902677834033966, "learning_rate": 5.138013329068769e-08, "loss": 0.5333, "num_input_tokens_seen": 24527624, "step": 37460 }, { "epoch": 19.63574423480084, "grad_norm": 0.1580783575773239, "learning_rate": 5.065003913909405e-08, "loss": 0.4866, "num_input_tokens_seen": 24529960, "step": 37465 }, { "epoch": 19.638364779874212, "grad_norm": 0.2489105761051178, "learning_rate": 4.9925164106875975e-08, "loss": 0.796, "num_input_tokens_seen": 24532616, "step": 37470 }, { "epoch": 19.64098532494759, "grad_norm": 0.29306653141975403, "learning_rate": 4.9205508345670506e-08, "loss": 0.3821, "num_input_tokens_seen": 24535880, "step": 37475 }, { "epoch": 19.643605870020963, "grad_norm": 0.16236846148967743, "learning_rate": 4.8491072006015546e-08, "loss": 0.4936, "num_input_tokens_seen": 24540296, "step": 37480 }, { "epoch": 19.64622641509434, "grad_norm": 0.09986996650695801, "learning_rate": 4.778185523736933e-08, "loss": 0.4603, "num_input_tokens_seen": 24543496, "step": 37485 }, { "epoch": 19.648846960167717, "grad_norm": 0.2567112445831299, "learning_rate": 4.7077858188085414e-08, "loss": 0.4118, "num_input_tokens_seen": 24550056, "step": 37490 }, { "epoch": 19.65146750524109, "grad_norm": 0.1335621029138565, "learning_rate": 4.6379081005437644e-08, "loss": 0.4652, "num_input_tokens_seen": 24553608, "step": 37495 }, { "epoch": 19.654088050314467, "grad_norm": 0.356979638338089, "learning_rate": 4.568552383560076e-08, "loss": 0.6286, "num_input_tokens_seen": 24557224, "step": 37500 }, { "epoch": 19.65670859538784, "grad_norm": 0.20321232080459595, "learning_rate": 4.4997186823655945e-08, "loss": 0.6754, "num_input_tokens_seen": 24560616, "step": 37505 }, { "epoch": 19.659329140461217, "grad_norm": 0.1425585150718689, "learning_rate": 4.431407011360189e-08, "loss": 0.7475, "num_input_tokens_seen": 24563656, "step": 37510 }, { "epoch": 19.66194968553459, "grad_norm": 0.15579426288604736, "learning_rate": 4.3636173848332627e-08, "loss": 0.4707, "num_input_tokens_seen": 24566312, "step": 37515 }, { "epoch": 19.664570230607968, "grad_norm": 0.16748233139514923, "learning_rate": 4.2963498169656946e-08, "loss": 0.5903, "num_input_tokens_seen": 24569640, "step": 37520 }, { "epoch": 19.66719077568134, "grad_norm": 0.15467636287212372, "learning_rate": 4.229604321829561e-08, "loss": 0.5676, "num_input_tokens_seen": 24572264, "step": 37525 }, { "epoch": 19.669811320754718, "grad_norm": 0.15053166449069977, "learning_rate": 4.16338091338675e-08, "loss": 0.455, "num_input_tokens_seen": 24575048, "step": 37530 }, { "epoch": 19.67243186582809, "grad_norm": 0.1540466696023941, "learning_rate": 4.097679605490623e-08, "loss": 0.476, "num_input_tokens_seen": 24577704, "step": 37535 }, { "epoch": 19.67505241090147, "grad_norm": 0.1897435188293457, "learning_rate": 4.032500411885465e-08, "loss": 0.4768, "num_input_tokens_seen": 24581352, "step": 37540 }, { "epoch": 19.677672955974842, "grad_norm": 0.24245432019233704, "learning_rate": 3.967843346205369e-08, "loss": 0.4964, "num_input_tokens_seen": 24584168, "step": 37545 }, { "epoch": 19.68029350104822, "grad_norm": 0.1414809376001358, "learning_rate": 3.903708421976182e-08, "loss": 0.3521, "num_input_tokens_seen": 24588200, "step": 37550 }, { "epoch": 19.682914046121592, "grad_norm": 0.20541071891784668, "learning_rate": 3.840095652614395e-08, "loss": 0.668, "num_input_tokens_seen": 24590568, "step": 37555 }, { "epoch": 19.68553459119497, "grad_norm": 0.1669038087129593, "learning_rate": 3.777005051426863e-08, "loss": 0.4867, "num_input_tokens_seen": 24593320, "step": 37560 }, { "epoch": 19.688155136268342, "grad_norm": 0.2753700017929077, "learning_rate": 3.71443663161164e-08, "loss": 0.6177, "num_input_tokens_seen": 24595848, "step": 37565 }, { "epoch": 19.69077568134172, "grad_norm": 0.13355323672294617, "learning_rate": 3.652390406256867e-08, "loss": 0.4224, "num_input_tokens_seen": 24598600, "step": 37570 }, { "epoch": 19.693396226415093, "grad_norm": 0.19485250115394592, "learning_rate": 3.5908663883421624e-08, "loss": 0.5039, "num_input_tokens_seen": 24601544, "step": 37575 }, { "epoch": 19.69601677148847, "grad_norm": 0.16565676033496857, "learning_rate": 3.529864590737786e-08, "loss": 0.4562, "num_input_tokens_seen": 24604552, "step": 37580 }, { "epoch": 19.698637316561843, "grad_norm": 0.16363346576690674, "learning_rate": 3.4693850262046415e-08, "loss": 0.5479, "num_input_tokens_seen": 24607432, "step": 37585 }, { "epoch": 19.70125786163522, "grad_norm": 0.18249231576919556, "learning_rate": 3.409427707394275e-08, "loss": 0.524, "num_input_tokens_seen": 24611464, "step": 37590 }, { "epoch": 19.703878406708597, "grad_norm": 0.25984251499176025, "learning_rate": 3.349992646848599e-08, "loss": 0.5421, "num_input_tokens_seen": 24614344, "step": 37595 }, { "epoch": 19.70649895178197, "grad_norm": 0.1455971747636795, "learning_rate": 3.291079857001556e-08, "loss": 0.538, "num_input_tokens_seen": 24617864, "step": 37600 }, { "epoch": 19.709119496855347, "grad_norm": 0.19083064794540405, "learning_rate": 3.2326893501763436e-08, "loss": 0.4257, "num_input_tokens_seen": 24620904, "step": 37605 }, { "epoch": 19.71174004192872, "grad_norm": 0.2181023806333542, "learning_rate": 3.174821138588191e-08, "loss": 0.3699, "num_input_tokens_seen": 24623560, "step": 37610 }, { "epoch": 19.714360587002098, "grad_norm": 0.1588418036699295, "learning_rate": 3.11747523434186e-08, "loss": 0.4244, "num_input_tokens_seen": 24627016, "step": 37615 }, { "epoch": 19.71698113207547, "grad_norm": 0.15258468687534332, "learning_rate": 3.060651649434143e-08, "loss": 0.4248, "num_input_tokens_seen": 24630600, "step": 37620 }, { "epoch": 19.719601677148848, "grad_norm": 0.1769072711467743, "learning_rate": 3.004350395751088e-08, "loss": 0.5755, "num_input_tokens_seen": 24634248, "step": 37625 }, { "epoch": 19.72222222222222, "grad_norm": 0.1436106115579605, "learning_rate": 2.94857148507105e-08, "loss": 0.3468, "num_input_tokens_seen": 24637352, "step": 37630 }, { "epoch": 19.7248427672956, "grad_norm": 0.19880931079387665, "learning_rate": 2.893314929061641e-08, "loss": 0.4844, "num_input_tokens_seen": 24639912, "step": 37635 }, { "epoch": 19.72746331236897, "grad_norm": 0.17629389464855194, "learning_rate": 2.8385807392825037e-08, "loss": 0.408, "num_input_tokens_seen": 24642568, "step": 37640 }, { "epoch": 19.73008385744235, "grad_norm": 0.16429302096366882, "learning_rate": 2.7843689271830897e-08, "loss": 0.3681, "num_input_tokens_seen": 24645416, "step": 37645 }, { "epoch": 19.732704402515722, "grad_norm": 0.18816062808036804, "learning_rate": 2.7306795041040502e-08, "loss": 0.4507, "num_input_tokens_seen": 24648360, "step": 37650 }, { "epoch": 19.7353249475891, "grad_norm": 0.3032641112804413, "learning_rate": 2.677512481276401e-08, "loss": 0.592, "num_input_tokens_seen": 24650728, "step": 37655 }, { "epoch": 19.737945492662472, "grad_norm": 0.4980916976928711, "learning_rate": 2.6248678698220785e-08, "loss": 0.5537, "num_input_tokens_seen": 24653928, "step": 37660 }, { "epoch": 19.74056603773585, "grad_norm": 0.15978731215000153, "learning_rate": 2.5727456807542183e-08, "loss": 0.4537, "num_input_tokens_seen": 24657608, "step": 37665 }, { "epoch": 19.743186582809223, "grad_norm": 0.4312606155872345, "learning_rate": 2.5211459249754875e-08, "loss": 0.3858, "num_input_tokens_seen": 24660744, "step": 37670 }, { "epoch": 19.7458071278826, "grad_norm": 0.21108944714069366, "learning_rate": 2.4700686132803076e-08, "loss": 0.4639, "num_input_tokens_seen": 24664136, "step": 37675 }, { "epoch": 19.748427672955973, "grad_norm": 0.13459078967571259, "learning_rate": 2.4195137563534643e-08, "loss": 0.4952, "num_input_tokens_seen": 24666920, "step": 37680 }, { "epoch": 19.75104821802935, "grad_norm": 0.2793905436992645, "learning_rate": 2.369481364770387e-08, "loss": 0.5248, "num_input_tokens_seen": 24669800, "step": 37685 }, { "epoch": 19.753668763102727, "grad_norm": 0.12815840542316437, "learning_rate": 2.3199714489974266e-08, "loss": 0.4982, "num_input_tokens_seen": 24672488, "step": 37690 }, { "epoch": 19.7562893081761, "grad_norm": 0.18409955501556396, "learning_rate": 2.2709840193912978e-08, "loss": 0.5191, "num_input_tokens_seen": 24676168, "step": 37695 }, { "epoch": 19.758909853249477, "grad_norm": 0.27444782853126526, "learning_rate": 2.2225190861999146e-08, "loss": 0.5854, "num_input_tokens_seen": 24679176, "step": 37700 }, { "epoch": 19.76153039832285, "grad_norm": 0.22802391648292542, "learning_rate": 2.1745766595612783e-08, "loss": 0.3934, "num_input_tokens_seen": 24681864, "step": 37705 }, { "epoch": 19.764150943396228, "grad_norm": 0.14902575314044952, "learning_rate": 2.1271567495045886e-08, "loss": 0.5215, "num_input_tokens_seen": 24684712, "step": 37710 }, { "epoch": 19.7667714884696, "grad_norm": 0.18634945154190063, "learning_rate": 2.0802593659491333e-08, "loss": 0.57, "num_input_tokens_seen": 24687336, "step": 37715 }, { "epoch": 19.769392033542978, "grad_norm": 0.1312103420495987, "learning_rate": 2.0338845187059528e-08, "loss": 0.3629, "num_input_tokens_seen": 24690664, "step": 37720 }, { "epoch": 19.77201257861635, "grad_norm": 0.16319003701210022, "learning_rate": 1.988032217475899e-08, "loss": 0.5045, "num_input_tokens_seen": 24694216, "step": 37725 }, { "epoch": 19.77463312368973, "grad_norm": 0.3332056701183319, "learning_rate": 1.942702471850466e-08, "loss": 0.5347, "num_input_tokens_seen": 24696936, "step": 37730 }, { "epoch": 19.7772536687631, "grad_norm": 0.15078172087669373, "learning_rate": 1.897895291312346e-08, "loss": 0.4484, "num_input_tokens_seen": 24699400, "step": 37735 }, { "epoch": 19.77987421383648, "grad_norm": 0.1798410415649414, "learning_rate": 1.8536106852348746e-08, "loss": 0.5301, "num_input_tokens_seen": 24702376, "step": 37740 }, { "epoch": 19.782494758909852, "grad_norm": 0.13771399855613708, "learning_rate": 1.8098486628814748e-08, "loss": 0.3871, "num_input_tokens_seen": 24706568, "step": 37745 }, { "epoch": 19.78511530398323, "grad_norm": 0.133550763130188, "learning_rate": 1.7666092334067684e-08, "loss": 0.5487, "num_input_tokens_seen": 24709384, "step": 37750 }, { "epoch": 19.787735849056602, "grad_norm": 0.13799448311328888, "learning_rate": 1.7238924058562977e-08, "loss": 0.4157, "num_input_tokens_seen": 24712552, "step": 37755 }, { "epoch": 19.79035639412998, "grad_norm": 0.15399646759033203, "learning_rate": 1.6816981891656924e-08, "loss": 0.4025, "num_input_tokens_seen": 24715400, "step": 37760 }, { "epoch": 19.792976939203353, "grad_norm": 0.2040373533964157, "learning_rate": 1.6400265921615032e-08, "loss": 0.4971, "num_input_tokens_seen": 24718376, "step": 37765 }, { "epoch": 19.79559748427673, "grad_norm": 0.16619619727134705, "learning_rate": 1.598877623560924e-08, "loss": 0.6371, "num_input_tokens_seen": 24721416, "step": 37770 }, { "epoch": 19.798218029350103, "grad_norm": 0.1944495141506195, "learning_rate": 1.5582512919717908e-08, "loss": 0.5273, "num_input_tokens_seen": 24724232, "step": 37775 }, { "epoch": 19.80083857442348, "grad_norm": 0.20846021175384521, "learning_rate": 1.5181476058928613e-08, "loss": 0.6416, "num_input_tokens_seen": 24731656, "step": 37780 }, { "epoch": 19.803459119496857, "grad_norm": 0.17632530629634857, "learning_rate": 1.4785665737129805e-08, "loss": 0.5544, "num_input_tokens_seen": 24734344, "step": 37785 }, { "epoch": 19.80607966457023, "grad_norm": 0.21462710201740265, "learning_rate": 1.4395082037127471e-08, "loss": 0.439, "num_input_tokens_seen": 24737416, "step": 37790 }, { "epoch": 19.808700209643607, "grad_norm": 0.10325885564088821, "learning_rate": 1.4009725040620148e-08, "loss": 0.4365, "num_input_tokens_seen": 24740648, "step": 37795 }, { "epoch": 19.81132075471698, "grad_norm": 0.0929674580693245, "learning_rate": 1.3629594828223902e-08, "loss": 0.3988, "num_input_tokens_seen": 24744712, "step": 37800 }, { "epoch": 19.813941299790358, "grad_norm": 0.18178535997867584, "learning_rate": 1.3254691479458458e-08, "loss": 0.4173, "num_input_tokens_seen": 24747592, "step": 37805 }, { "epoch": 19.81656184486373, "grad_norm": 0.16163374483585358, "learning_rate": 1.2885015072747197e-08, "loss": 0.445, "num_input_tokens_seen": 24750920, "step": 37810 }, { "epoch": 19.819182389937108, "grad_norm": 0.1531374156475067, "learning_rate": 1.2520565685419927e-08, "loss": 0.4063, "num_input_tokens_seen": 24756200, "step": 37815 }, { "epoch": 19.82180293501048, "grad_norm": 0.1784660816192627, "learning_rate": 1.2161343393721214e-08, "loss": 0.4477, "num_input_tokens_seen": 24759144, "step": 37820 }, { "epoch": 19.82442348008386, "grad_norm": 0.2445695996284485, "learning_rate": 1.1807348272790952e-08, "loss": 0.4704, "num_input_tokens_seen": 24761992, "step": 37825 }, { "epoch": 19.82704402515723, "grad_norm": 0.24436317384243011, "learning_rate": 1.1458580396683794e-08, "loss": 0.4935, "num_input_tokens_seen": 24764744, "step": 37830 }, { "epoch": 19.82966457023061, "grad_norm": 0.1752176135778427, "learning_rate": 1.1115039838358044e-08, "loss": 0.6048, "num_input_tokens_seen": 24767336, "step": 37835 }, { "epoch": 19.832285115303982, "grad_norm": 0.14647918939590454, "learning_rate": 1.0776726669678439e-08, "loss": 0.5022, "num_input_tokens_seen": 24770344, "step": 37840 }, { "epoch": 19.83490566037736, "grad_norm": 0.27024802565574646, "learning_rate": 1.044364096141337e-08, "loss": 0.5001, "num_input_tokens_seen": 24773704, "step": 37845 }, { "epoch": 19.837526205450732, "grad_norm": 0.1805339902639389, "learning_rate": 1.0115782783243211e-08, "loss": 0.4754, "num_input_tokens_seen": 24776840, "step": 37850 }, { "epoch": 19.84014675052411, "grad_norm": 0.19293014705181122, "learning_rate": 9.793152203751988e-09, "loss": 0.5162, "num_input_tokens_seen": 24780904, "step": 37855 }, { "epoch": 19.842767295597483, "grad_norm": 0.12110371887683868, "learning_rate": 9.475749290430158e-09, "loss": 0.4259, "num_input_tokens_seen": 24784104, "step": 37860 }, { "epoch": 19.84538784067086, "grad_norm": 0.15429159998893738, "learning_rate": 9.16357410967461e-09, "loss": 0.4739, "num_input_tokens_seen": 24788712, "step": 37865 }, { "epoch": 19.848008385744233, "grad_norm": 0.18414530158042908, "learning_rate": 8.856626726788664e-09, "loss": 0.4566, "num_input_tokens_seen": 24791784, "step": 37870 }, { "epoch": 19.85062893081761, "grad_norm": 0.11875008791685104, "learning_rate": 8.554907205982066e-09, "loss": 0.573, "num_input_tokens_seen": 24795016, "step": 37875 }, { "epoch": 19.853249475890987, "grad_norm": 0.1861838847398758, "learning_rate": 8.258415610373772e-09, "loss": 0.5234, "num_input_tokens_seen": 24798600, "step": 37880 }, { "epoch": 19.85587002096436, "grad_norm": 0.2427704632282257, "learning_rate": 7.967152001980838e-09, "loss": 0.5661, "num_input_tokens_seen": 24801288, "step": 37885 }, { "epoch": 19.858490566037737, "grad_norm": 0.14472423493862152, "learning_rate": 7.681116441737856e-09, "loss": 0.4219, "num_input_tokens_seen": 24803976, "step": 37890 }, { "epoch": 19.86111111111111, "grad_norm": 0.23244138062000275, "learning_rate": 7.4003089894747464e-09, "loss": 0.5735, "num_input_tokens_seen": 24806632, "step": 37895 }, { "epoch": 19.863731656184488, "grad_norm": 0.14798128604888916, "learning_rate": 7.12472970393896e-09, "loss": 0.3952, "num_input_tokens_seen": 24809160, "step": 37900 }, { "epoch": 19.86635220125786, "grad_norm": 0.13232672214508057, "learning_rate": 6.854378642776049e-09, "loss": 0.5538, "num_input_tokens_seen": 24812584, "step": 37905 }, { "epoch": 19.868972746331238, "grad_norm": 0.17481976747512817, "learning_rate": 6.589255862538002e-09, "loss": 0.574, "num_input_tokens_seen": 24816104, "step": 37910 }, { "epoch": 19.87159329140461, "grad_norm": 0.20985861122608185, "learning_rate": 6.32936141869156e-09, "loss": 0.486, "num_input_tokens_seen": 24819112, "step": 37915 }, { "epoch": 19.87421383647799, "grad_norm": 0.15954995155334473, "learning_rate": 6.074695365596017e-09, "loss": 0.3436, "num_input_tokens_seen": 24822504, "step": 37920 }, { "epoch": 19.87683438155136, "grad_norm": 0.17819340527057648, "learning_rate": 5.8252577565309775e-09, "loss": 0.4609, "num_input_tokens_seen": 24825224, "step": 37925 }, { "epoch": 19.87945492662474, "grad_norm": 0.2811594307422638, "learning_rate": 5.581048643671371e-09, "loss": 0.5745, "num_input_tokens_seen": 24828104, "step": 37930 }, { "epoch": 19.882075471698112, "grad_norm": 0.10496808588504791, "learning_rate": 5.342068078106888e-09, "loss": 0.4444, "num_input_tokens_seen": 24831592, "step": 37935 }, { "epoch": 19.88469601677149, "grad_norm": 0.13789483904838562, "learning_rate": 5.1083161098253175e-09, "loss": 0.5103, "num_input_tokens_seen": 24834824, "step": 37940 }, { "epoch": 19.887316561844862, "grad_norm": 0.1478915512561798, "learning_rate": 4.87979278772921e-09, "loss": 0.516, "num_input_tokens_seen": 24838184, "step": 37945 }, { "epoch": 19.88993710691824, "grad_norm": 0.12062834203243256, "learning_rate": 4.656498159619216e-09, "loss": 0.4348, "num_input_tokens_seen": 24841576, "step": 37950 }, { "epoch": 19.892557651991613, "grad_norm": 0.30749547481536865, "learning_rate": 4.438432272210746e-09, "loss": 0.5816, "num_input_tokens_seen": 24844232, "step": 37955 }, { "epoch": 19.89517819706499, "grad_norm": 0.16325660049915314, "learning_rate": 4.225595171114538e-09, "loss": 0.4326, "num_input_tokens_seen": 24847144, "step": 37960 }, { "epoch": 19.897798742138363, "grad_norm": 0.27309027314186096, "learning_rate": 4.017986900858861e-09, "loss": 0.4804, "num_input_tokens_seen": 24850568, "step": 37965 }, { "epoch": 19.90041928721174, "grad_norm": 0.2130437046289444, "learning_rate": 3.815607504870089e-09, "loss": 0.4041, "num_input_tokens_seen": 24852808, "step": 37970 }, { "epoch": 19.903039832285117, "grad_norm": 0.15159563720226288, "learning_rate": 3.618457025486577e-09, "loss": 0.5465, "num_input_tokens_seen": 24856744, "step": 37975 }, { "epoch": 19.90566037735849, "grad_norm": 0.17373710870742798, "learning_rate": 3.426535503947559e-09, "loss": 0.5211, "num_input_tokens_seen": 24860296, "step": 37980 }, { "epoch": 19.908280922431867, "grad_norm": 0.1368177831172943, "learning_rate": 3.2398429803986997e-09, "loss": 0.4127, "num_input_tokens_seen": 24863560, "step": 37985 }, { "epoch": 19.91090146750524, "grad_norm": 0.12757249176502228, "learning_rate": 3.058379493900421e-09, "loss": 0.4181, "num_input_tokens_seen": 24866344, "step": 37990 }, { "epoch": 19.913522012578618, "grad_norm": 0.13402430713176727, "learning_rate": 2.882145082405696e-09, "loss": 0.5193, "num_input_tokens_seen": 24869800, "step": 37995 }, { "epoch": 19.91614255765199, "grad_norm": 0.17928536236286163, "learning_rate": 2.711139782785033e-09, "loss": 0.3986, "num_input_tokens_seen": 24872264, "step": 38000 }, { "epoch": 19.918763102725368, "grad_norm": 0.2001665085554123, "learning_rate": 2.5453636308098163e-09, "loss": 0.4931, "num_input_tokens_seen": 24875816, "step": 38005 }, { "epoch": 19.92138364779874, "grad_norm": 0.19707053899765015, "learning_rate": 2.3848166611606383e-09, "loss": 0.455, "num_input_tokens_seen": 24878856, "step": 38010 }, { "epoch": 19.92400419287212, "grad_norm": 0.1968899518251419, "learning_rate": 2.2294989074161943e-09, "loss": 0.4713, "num_input_tokens_seen": 24881864, "step": 38015 }, { "epoch": 19.92662473794549, "grad_norm": 0.2687709629535675, "learning_rate": 2.0794104020727125e-09, "loss": 0.4924, "num_input_tokens_seen": 24884488, "step": 38020 }, { "epoch": 19.92924528301887, "grad_norm": 0.2479383498430252, "learning_rate": 1.9345511765272993e-09, "loss": 0.3714, "num_input_tokens_seen": 24887944, "step": 38025 }, { "epoch": 19.931865828092242, "grad_norm": 0.1267872154712677, "learning_rate": 1.794921261077942e-09, "loss": 0.4657, "num_input_tokens_seen": 24891400, "step": 38030 }, { "epoch": 19.93448637316562, "grad_norm": 0.3501501679420471, "learning_rate": 1.6605206849373833e-09, "loss": 0.4257, "num_input_tokens_seen": 24893672, "step": 38035 }, { "epoch": 19.937106918238992, "grad_norm": 0.15275442600250244, "learning_rate": 1.531349476219246e-09, "loss": 0.477, "num_input_tokens_seen": 24897544, "step": 38040 }, { "epoch": 19.93972746331237, "grad_norm": 0.1045738086104393, "learning_rate": 1.4074076619463584e-09, "loss": 0.6454, "num_input_tokens_seen": 24900904, "step": 38045 }, { "epoch": 19.942348008385743, "grad_norm": 0.10335492342710495, "learning_rate": 1.2886952680452036e-09, "loss": 0.4312, "num_input_tokens_seen": 24905064, "step": 38050 }, { "epoch": 19.94496855345912, "grad_norm": 0.16458602249622345, "learning_rate": 1.1752123193459197e-09, "loss": 0.4801, "num_input_tokens_seen": 24908392, "step": 38055 }, { "epoch": 19.947589098532493, "grad_norm": 0.11385472863912582, "learning_rate": 1.0669588395934017e-09, "loss": 0.4506, "num_input_tokens_seen": 24911336, "step": 38060 }, { "epoch": 19.95020964360587, "grad_norm": 0.1474597305059433, "learning_rate": 9.639348514278722e-10, "loss": 0.4068, "num_input_tokens_seen": 24915880, "step": 38065 }, { "epoch": 19.952830188679247, "grad_norm": 0.10823281854391098, "learning_rate": 8.661403764043119e-10, "loss": 0.6218, "num_input_tokens_seen": 24919464, "step": 38070 }, { "epoch": 19.95545073375262, "grad_norm": 0.18070589005947113, "learning_rate": 7.735754349785795e-10, "loss": 0.3182, "num_input_tokens_seen": 24922312, "step": 38075 }, { "epoch": 19.958071278825997, "grad_norm": 0.12365792691707611, "learning_rate": 6.862400465157403e-10, "loss": 0.5052, "num_input_tokens_seen": 24925928, "step": 38080 }, { "epoch": 19.96069182389937, "grad_norm": 0.2131180614233017, "learning_rate": 6.041342292817387e-10, "loss": 0.4061, "num_input_tokens_seen": 24929224, "step": 38085 }, { "epoch": 19.963312368972748, "grad_norm": 0.2116619348526001, "learning_rate": 5.272580004572758e-10, "loss": 0.4203, "num_input_tokens_seen": 24931752, "step": 38090 }, { "epoch": 19.96593291404612, "grad_norm": 0.2223587930202484, "learning_rate": 4.556113761211567e-10, "loss": 0.7313, "num_input_tokens_seen": 24934504, "step": 38095 }, { "epoch": 19.968553459119498, "grad_norm": 0.15172865986824036, "learning_rate": 3.8919437125861657e-10, "loss": 0.5883, "num_input_tokens_seen": 24937512, "step": 38100 }, { "epoch": 19.97117400419287, "grad_norm": 0.11010848730802536, "learning_rate": 3.280069997696478e-10, "loss": 0.3638, "num_input_tokens_seen": 24940520, "step": 38105 }, { "epoch": 19.97379454926625, "grad_norm": 0.15786530077457428, "learning_rate": 2.720492744467951e-10, "loss": 0.4232, "num_input_tokens_seen": 24942888, "step": 38110 }, { "epoch": 19.97641509433962, "grad_norm": 0.2397204041481018, "learning_rate": 2.2132120700291136e-10, "loss": 0.5253, "num_input_tokens_seen": 24946408, "step": 38115 }, { "epoch": 19.979035639413, "grad_norm": 0.3279951512813568, "learning_rate": 1.7582280804340213e-10, "loss": 0.6977, "num_input_tokens_seen": 24948968, "step": 38120 }, { "epoch": 19.981656184486372, "grad_norm": 0.15765748918056488, "learning_rate": 1.3555408708842977e-10, "loss": 0.4165, "num_input_tokens_seen": 24951752, "step": 38125 }, { "epoch": 19.98427672955975, "grad_norm": 0.20530462265014648, "learning_rate": 1.0051505256458704e-10, "loss": 0.493, "num_input_tokens_seen": 24954696, "step": 38130 }, { "epoch": 19.986897274633122, "grad_norm": 0.17588496208190918, "learning_rate": 7.070571179657037e-11, "loss": 0.5078, "num_input_tokens_seen": 24959048, "step": 38135 }, { "epoch": 19.9895178197065, "grad_norm": 0.11779162287712097, "learning_rate": 4.612607102383315e-11, "loss": 0.3245, "num_input_tokens_seen": 24961640, "step": 38140 }, { "epoch": 19.992138364779873, "grad_norm": 0.20211637020111084, "learning_rate": 2.677613538670798e-11, "loss": 0.4316, "num_input_tokens_seen": 24963976, "step": 38145 }, { "epoch": 19.99475890985325, "grad_norm": 0.20663121342658997, "learning_rate": 1.2655908931957782e-11, "loss": 0.4469, "num_input_tokens_seen": 24966760, "step": 38150 }, { "epoch": 19.997379454926623, "grad_norm": 0.13123875856399536, "learning_rate": 3.765394615551366e-12, "loss": 0.406, "num_input_tokens_seen": 24970216, "step": 38155 }, { "epoch": 20.0, "grad_norm": 2.0712826251983643, "learning_rate": 1.0459429711229262e-13, "loss": 0.9468, "num_input_tokens_seen": 24973864, "step": 38160 }, { "epoch": 20.0, "eval_loss": 0.4603464603424072, "eval_runtime": 9.925, "eval_samples_per_second": 85.441, "eval_steps_per_second": 21.36, "num_input_tokens_seen": 24973864, "step": 38160 }, { "epoch": 20.0, "num_input_tokens_seen": 24973864, "step": 38160, "total_flos": 1.1245912689013555e+18, "train_loss": 0.5332972578844934, "train_runtime": 7012.9299, "train_samples_per_second": 21.757, "train_steps_per_second": 5.441 } ], "logging_steps": 5, "max_steps": 38160, "num_input_tokens_seen": 24973864, "num_train_epochs": 20, "save_steps": 1908, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.1245912689013555e+18, "train_batch_size": 4, "trial_name": null, "trial_params": null }