{ "best_global_step": 39200, "best_metric": 0.4438072144985199, "best_model_checkpoint": "saves/ia3/llama-3-8b-instruct/train_record_1745950250/checkpoint-39200", "epoch": 1.2803277639075603, "eval_steps": 200, "global_step": 40000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00016004097048844505, "grad_norm": 10.631747245788574, "learning_rate": 4.999999876629946e-05, "loss": 5.3771, "num_input_tokens_seen": 6960, "step": 5 }, { "epoch": 0.0003200819409768901, "grad_norm": 5.854475021362305, "learning_rate": 4.999999375439123e-05, "loss": 5.2197, "num_input_tokens_seen": 13728, "step": 10 }, { "epoch": 0.0004801229114653351, "grad_norm": 6.329836368560791, "learning_rate": 4.9999984887169785e-05, "loss": 4.859, "num_input_tokens_seen": 20688, "step": 15 }, { "epoch": 0.0006401638819537802, "grad_norm": 9.468019485473633, "learning_rate": 4.9999972164636506e-05, "loss": 4.4766, "num_input_tokens_seen": 27616, "step": 20 }, { "epoch": 0.0008002048524422252, "grad_norm": 17.838653564453125, "learning_rate": 4.999995558679334e-05, "loss": 5.2567, "num_input_tokens_seen": 34368, "step": 25 }, { "epoch": 0.0009602458229306702, "grad_norm": 8.72453784942627, "learning_rate": 4.999993515364287e-05, "loss": 5.9932, "num_input_tokens_seen": 41664, "step": 30 }, { "epoch": 0.0011202867934191152, "grad_norm": 9.447735786437988, "learning_rate": 4.999991086518822e-05, "loss": 4.765, "num_input_tokens_seen": 48544, "step": 35 }, { "epoch": 0.0012803277639075604, "grad_norm": 11.310359954833984, "learning_rate": 4.999988272143315e-05, "loss": 5.1247, "num_input_tokens_seen": 55360, "step": 40 }, { "epoch": 0.0014403687343960054, "grad_norm": 11.689703941345215, "learning_rate": 4.999985072238199e-05, "loss": 5.1403, "num_input_tokens_seen": 62176, "step": 45 }, { "epoch": 0.0016004097048844504, "grad_norm": 7.682919979095459, "learning_rate": 4.999981486803969e-05, "loss": 5.1026, "num_input_tokens_seen": 68768, "step": 50 }, { "epoch": 0.0017604506753728954, "grad_norm": 8.247049331665039, "learning_rate": 4.999977515841176e-05, "loss": 3.4508, "num_input_tokens_seen": 75536, "step": 55 }, { "epoch": 0.0019204916458613404, "grad_norm": 9.915614128112793, "learning_rate": 4.9999731593504344e-05, "loss": 3.988, "num_input_tokens_seen": 82320, "step": 60 }, { "epoch": 0.0020805326163497854, "grad_norm": 5.647215843200684, "learning_rate": 4.999968417332415e-05, "loss": 2.8134, "num_input_tokens_seen": 89184, "step": 65 }, { "epoch": 0.0022405735868382304, "grad_norm": 5.858863830566406, "learning_rate": 4.999963289787848e-05, "loss": 2.1689, "num_input_tokens_seen": 96384, "step": 70 }, { "epoch": 0.002400614557326676, "grad_norm": 6.9189019203186035, "learning_rate": 4.999957776717526e-05, "loss": 1.975, "num_input_tokens_seen": 102960, "step": 75 }, { "epoch": 0.002560655527815121, "grad_norm": 3.134488821029663, "learning_rate": 4.9999518781222984e-05, "loss": 1.655, "num_input_tokens_seen": 109664, "step": 80 }, { "epoch": 0.002720696498303566, "grad_norm": 7.075713157653809, "learning_rate": 4.9999455940030746e-05, "loss": 2.2553, "num_input_tokens_seen": 116896, "step": 85 }, { "epoch": 0.002880737468792011, "grad_norm": 3.8718740940093994, "learning_rate": 4.999938924360824e-05, "loss": 2.4298, "num_input_tokens_seen": 123808, "step": 90 }, { "epoch": 0.003040778439280456, "grad_norm": 4.8035430908203125, "learning_rate": 4.999931869196575e-05, "loss": 1.821, "num_input_tokens_seen": 130928, "step": 95 }, { "epoch": 0.003200819409768901, "grad_norm": 3.5584030151367188, "learning_rate": 4.999924428511416e-05, "loss": 2.323, "num_input_tokens_seen": 137840, "step": 100 }, { "epoch": 0.003360860380257346, "grad_norm": 3.0133185386657715, "learning_rate": 4.999916602306494e-05, "loss": 1.3469, "num_input_tokens_seen": 144720, "step": 105 }, { "epoch": 0.003520901350745791, "grad_norm": 2.535939931869507, "learning_rate": 4.999908390583016e-05, "loss": 0.8366, "num_input_tokens_seen": 151488, "step": 110 }, { "epoch": 0.003680942321234236, "grad_norm": 8.671341896057129, "learning_rate": 4.999899793342247e-05, "loss": 1.7019, "num_input_tokens_seen": 158416, "step": 115 }, { "epoch": 0.003840983291722681, "grad_norm": 4.431541442871094, "learning_rate": 4.999890810585516e-05, "loss": 1.8669, "num_input_tokens_seen": 165024, "step": 120 }, { "epoch": 0.004001024262211126, "grad_norm": 3.6779515743255615, "learning_rate": 4.999881442314206e-05, "loss": 1.7956, "num_input_tokens_seen": 171600, "step": 125 }, { "epoch": 0.004161065232699571, "grad_norm": 4.927439212799072, "learning_rate": 4.9998716885297617e-05, "loss": 1.8931, "num_input_tokens_seen": 178304, "step": 130 }, { "epoch": 0.004321106203188016, "grad_norm": 5.270958423614502, "learning_rate": 4.999861549233688e-05, "loss": 1.4858, "num_input_tokens_seen": 185216, "step": 135 }, { "epoch": 0.004481147173676461, "grad_norm": 5.4405436515808105, "learning_rate": 4.999851024427548e-05, "loss": 1.9552, "num_input_tokens_seen": 191680, "step": 140 }, { "epoch": 0.004641188144164906, "grad_norm": 7.162008762359619, "learning_rate": 4.999840114112965e-05, "loss": 1.318, "num_input_tokens_seen": 198080, "step": 145 }, { "epoch": 0.004801229114653352, "grad_norm": 4.280998229980469, "learning_rate": 4.999828818291621e-05, "loss": 2.1878, "num_input_tokens_seen": 204752, "step": 150 }, { "epoch": 0.004961270085141797, "grad_norm": 6.074519157409668, "learning_rate": 4.999817136965259e-05, "loss": 1.4981, "num_input_tokens_seen": 211168, "step": 155 }, { "epoch": 0.005121311055630242, "grad_norm": 3.8545613288879395, "learning_rate": 4.9998050701356794e-05, "loss": 1.9851, "num_input_tokens_seen": 217552, "step": 160 }, { "epoch": 0.005281352026118687, "grad_norm": 3.6796529293060303, "learning_rate": 4.999792617804744e-05, "loss": 1.2692, "num_input_tokens_seen": 224464, "step": 165 }, { "epoch": 0.005441392996607132, "grad_norm": 4.540348529815674, "learning_rate": 4.9997797799743724e-05, "loss": 1.9699, "num_input_tokens_seen": 231232, "step": 170 }, { "epoch": 0.005601433967095577, "grad_norm": 2.4423911571502686, "learning_rate": 4.999766556646545e-05, "loss": 1.5609, "num_input_tokens_seen": 238368, "step": 175 }, { "epoch": 0.005761474937584022, "grad_norm": 5.4121413230896, "learning_rate": 4.9997529478232996e-05, "loss": 1.7259, "num_input_tokens_seen": 245248, "step": 180 }, { "epoch": 0.005921515908072467, "grad_norm": 1.7838279008865356, "learning_rate": 4.9997389535067365e-05, "loss": 1.5299, "num_input_tokens_seen": 252128, "step": 185 }, { "epoch": 0.006081556878560912, "grad_norm": 4.345981597900391, "learning_rate": 4.999724573699012e-05, "loss": 1.4578, "num_input_tokens_seen": 259232, "step": 190 }, { "epoch": 0.006241597849049357, "grad_norm": 5.287743091583252, "learning_rate": 4.9997098084023457e-05, "loss": 1.6972, "num_input_tokens_seen": 266192, "step": 195 }, { "epoch": 0.006401638819537802, "grad_norm": 3.2870564460754395, "learning_rate": 4.999694657619013e-05, "loss": 1.0376, "num_input_tokens_seen": 272992, "step": 200 }, { "epoch": 0.006401638819537802, "eval_loss": 1.5286588668823242, "eval_runtime": 499.335, "eval_samples_per_second": 27.809, "eval_steps_per_second": 13.904, "num_input_tokens_seen": 272992, "step": 200 }, { "epoch": 0.006561679790026247, "grad_norm": 7.579558849334717, "learning_rate": 4.999679121351352e-05, "loss": 1.4579, "num_input_tokens_seen": 279584, "step": 205 }, { "epoch": 0.006721720760514692, "grad_norm": 3.243178367614746, "learning_rate": 4.9996631996017565e-05, "loss": 1.2227, "num_input_tokens_seen": 286704, "step": 210 }, { "epoch": 0.006881761731003137, "grad_norm": 4.267862319946289, "learning_rate": 4.9996468923726835e-05, "loss": 1.3894, "num_input_tokens_seen": 293520, "step": 215 }, { "epoch": 0.007041802701491582, "grad_norm": 3.4980695247650146, "learning_rate": 4.999630199666647e-05, "loss": 2.1183, "num_input_tokens_seen": 300032, "step": 220 }, { "epoch": 0.007201843671980027, "grad_norm": 9.907645225524902, "learning_rate": 4.999613121486222e-05, "loss": 0.9698, "num_input_tokens_seen": 306832, "step": 225 }, { "epoch": 0.007361884642468472, "grad_norm": 6.705359935760498, "learning_rate": 4.999595657834041e-05, "loss": 1.3032, "num_input_tokens_seen": 313664, "step": 230 }, { "epoch": 0.007521925612956917, "grad_norm": 2.118607521057129, "learning_rate": 4.999577808712798e-05, "loss": 1.2461, "num_input_tokens_seen": 320384, "step": 235 }, { "epoch": 0.007681966583445362, "grad_norm": 3.999077558517456, "learning_rate": 4.999559574125244e-05, "loss": 1.4843, "num_input_tokens_seen": 326896, "step": 240 }, { "epoch": 0.007842007553933807, "grad_norm": 4.418499946594238, "learning_rate": 4.9995409540741934e-05, "loss": 1.4939, "num_input_tokens_seen": 333760, "step": 245 }, { "epoch": 0.008002048524422252, "grad_norm": 3.2386155128479004, "learning_rate": 4.999521948562516e-05, "loss": 1.358, "num_input_tokens_seen": 340288, "step": 250 }, { "epoch": 0.008162089494910697, "grad_norm": 11.279526710510254, "learning_rate": 4.999502557593143e-05, "loss": 2.0171, "num_input_tokens_seen": 347136, "step": 255 }, { "epoch": 0.008322130465399142, "grad_norm": 2.1358070373535156, "learning_rate": 4.999482781169066e-05, "loss": 1.0911, "num_input_tokens_seen": 353664, "step": 260 }, { "epoch": 0.008482171435887587, "grad_norm": 3.8295624256134033, "learning_rate": 4.9994626192933324e-05, "loss": 1.1733, "num_input_tokens_seen": 360368, "step": 265 }, { "epoch": 0.008642212406376032, "grad_norm": 2.768850088119507, "learning_rate": 4.999442071969054e-05, "loss": 1.2712, "num_input_tokens_seen": 366928, "step": 270 }, { "epoch": 0.008802253376864477, "grad_norm": 3.563295841217041, "learning_rate": 4.999421139199397e-05, "loss": 1.2139, "num_input_tokens_seen": 373904, "step": 275 }, { "epoch": 0.008962294347352922, "grad_norm": 3.714517593383789, "learning_rate": 4.999399820987592e-05, "loss": 1.2626, "num_input_tokens_seen": 380432, "step": 280 }, { "epoch": 0.009122335317841367, "grad_norm": 3.0251293182373047, "learning_rate": 4.999378117336924e-05, "loss": 1.5545, "num_input_tokens_seen": 387312, "step": 285 }, { "epoch": 0.009282376288329812, "grad_norm": 2.7002358436584473, "learning_rate": 4.9993560282507415e-05, "loss": 1.1327, "num_input_tokens_seen": 393904, "step": 290 }, { "epoch": 0.009442417258818258, "grad_norm": 7.081448078155518, "learning_rate": 4.9993335537324495e-05, "loss": 1.3385, "num_input_tokens_seen": 400688, "step": 295 }, { "epoch": 0.009602458229306703, "grad_norm": 2.914933681488037, "learning_rate": 4.999310693785516e-05, "loss": 0.8559, "num_input_tokens_seen": 407248, "step": 300 }, { "epoch": 0.009762499199795148, "grad_norm": 3.25386381149292, "learning_rate": 4.9992874484134653e-05, "loss": 1.1548, "num_input_tokens_seen": 414096, "step": 305 }, { "epoch": 0.009922540170283593, "grad_norm": 3.85601806640625, "learning_rate": 4.999263817619882e-05, "loss": 1.4724, "num_input_tokens_seen": 420832, "step": 310 }, { "epoch": 0.010082581140772038, "grad_norm": 3.9993858337402344, "learning_rate": 4.9992398014084105e-05, "loss": 1.3027, "num_input_tokens_seen": 427312, "step": 315 }, { "epoch": 0.010242622111260483, "grad_norm": 3.829296112060547, "learning_rate": 4.999215399782754e-05, "loss": 1.2334, "num_input_tokens_seen": 433680, "step": 320 }, { "epoch": 0.010402663081748928, "grad_norm": 2.6288225650787354, "learning_rate": 4.999190612746675e-05, "loss": 1.1577, "num_input_tokens_seen": 440368, "step": 325 }, { "epoch": 0.010562704052237373, "grad_norm": 1.4602971076965332, "learning_rate": 4.999165440303998e-05, "loss": 0.9965, "num_input_tokens_seen": 446736, "step": 330 }, { "epoch": 0.010722745022725818, "grad_norm": 2.96651291847229, "learning_rate": 4.999139882458603e-05, "loss": 1.3196, "num_input_tokens_seen": 453120, "step": 335 }, { "epoch": 0.010882785993214263, "grad_norm": 6.280483722686768, "learning_rate": 4.9991139392144314e-05, "loss": 1.4354, "num_input_tokens_seen": 459840, "step": 340 }, { "epoch": 0.011042826963702708, "grad_norm": 1.8988151550292969, "learning_rate": 4.999087610575485e-05, "loss": 1.283, "num_input_tokens_seen": 466384, "step": 345 }, { "epoch": 0.011202867934191153, "grad_norm": 3.0920567512512207, "learning_rate": 4.999060896545824e-05, "loss": 1.0106, "num_input_tokens_seen": 473328, "step": 350 }, { "epoch": 0.011362908904679598, "grad_norm": 2.444110870361328, "learning_rate": 4.999033797129568e-05, "loss": 1.1432, "num_input_tokens_seen": 480448, "step": 355 }, { "epoch": 0.011522949875168043, "grad_norm": 3.7875943183898926, "learning_rate": 4.999006312330894e-05, "loss": 1.5731, "num_input_tokens_seen": 487408, "step": 360 }, { "epoch": 0.011682990845656488, "grad_norm": 3.788261651992798, "learning_rate": 4.998978442154043e-05, "loss": 1.5409, "num_input_tokens_seen": 494160, "step": 365 }, { "epoch": 0.011843031816144933, "grad_norm": 3.536120891571045, "learning_rate": 4.9989501866033125e-05, "loss": 1.3932, "num_input_tokens_seen": 500480, "step": 370 }, { "epoch": 0.012003072786633378, "grad_norm": 7.952110290527344, "learning_rate": 4.998921545683059e-05, "loss": 1.5224, "num_input_tokens_seen": 507424, "step": 375 }, { "epoch": 0.012163113757121823, "grad_norm": 0.9300363063812256, "learning_rate": 4.9988925193976996e-05, "loss": 0.9871, "num_input_tokens_seen": 514064, "step": 380 }, { "epoch": 0.012323154727610268, "grad_norm": 3.809927225112915, "learning_rate": 4.998863107751711e-05, "loss": 1.0829, "num_input_tokens_seen": 520816, "step": 385 }, { "epoch": 0.012483195698098713, "grad_norm": 1.7734637260437012, "learning_rate": 4.998833310749629e-05, "loss": 0.8983, "num_input_tokens_seen": 527328, "step": 390 }, { "epoch": 0.012643236668587158, "grad_norm": 2.0998594760894775, "learning_rate": 4.998803128396047e-05, "loss": 1.0647, "num_input_tokens_seen": 534656, "step": 395 }, { "epoch": 0.012803277639075603, "grad_norm": 1.7723344564437866, "learning_rate": 4.9987725606956215e-05, "loss": 1.1312, "num_input_tokens_seen": 541536, "step": 400 }, { "epoch": 0.012803277639075603, "eval_loss": 1.1516139507293701, "eval_runtime": 499.3705, "eval_samples_per_second": 27.807, "eval_steps_per_second": 13.904, "num_input_tokens_seen": 541536, "step": 400 }, { "epoch": 0.012963318609564048, "grad_norm": 4.00723934173584, "learning_rate": 4.998741607653066e-05, "loss": 1.4514, "num_input_tokens_seen": 548160, "step": 405 }, { "epoch": 0.013123359580052493, "grad_norm": 2.893052101135254, "learning_rate": 4.9987102692731523e-05, "loss": 0.8361, "num_input_tokens_seen": 555024, "step": 410 }, { "epoch": 0.013283400550540938, "grad_norm": 4.802688121795654, "learning_rate": 4.9986785455607157e-05, "loss": 0.9404, "num_input_tokens_seen": 561792, "step": 415 }, { "epoch": 0.013443441521029383, "grad_norm": 6.887642860412598, "learning_rate": 4.9986464365206456e-05, "loss": 1.1203, "num_input_tokens_seen": 568464, "step": 420 }, { "epoch": 0.013603482491517828, "grad_norm": 4.728724479675293, "learning_rate": 4.9986139421578956e-05, "loss": 1.155, "num_input_tokens_seen": 574928, "step": 425 }, { "epoch": 0.013763523462006273, "grad_norm": 2.083258628845215, "learning_rate": 4.998581062477477e-05, "loss": 0.689, "num_input_tokens_seen": 581792, "step": 430 }, { "epoch": 0.013923564432494718, "grad_norm": 5.526749610900879, "learning_rate": 4.998547797484458e-05, "loss": 1.2622, "num_input_tokens_seen": 588512, "step": 435 }, { "epoch": 0.014083605402983163, "grad_norm": 3.6472902297973633, "learning_rate": 4.9985141471839706e-05, "loss": 1.2678, "num_input_tokens_seen": 595168, "step": 440 }, { "epoch": 0.014243646373471608, "grad_norm": 4.441622257232666, "learning_rate": 4.998480111581203e-05, "loss": 0.8233, "num_input_tokens_seen": 601968, "step": 445 }, { "epoch": 0.014403687343960053, "grad_norm": 4.394436359405518, "learning_rate": 4.998445690681405e-05, "loss": 1.064, "num_input_tokens_seen": 608432, "step": 450 }, { "epoch": 0.014563728314448498, "grad_norm": 2.7628531455993652, "learning_rate": 4.9984108844898834e-05, "loss": 1.56, "num_input_tokens_seen": 615248, "step": 455 }, { "epoch": 0.014723769284936943, "grad_norm": 3.4770851135253906, "learning_rate": 4.9983756930120076e-05, "loss": 1.1974, "num_input_tokens_seen": 622432, "step": 460 }, { "epoch": 0.014883810255425388, "grad_norm": 2.444336414337158, "learning_rate": 4.9983401162532025e-05, "loss": 0.79, "num_input_tokens_seen": 629600, "step": 465 }, { "epoch": 0.015043851225913833, "grad_norm": 2.335761785507202, "learning_rate": 4.998304154218955e-05, "loss": 0.6561, "num_input_tokens_seen": 636640, "step": 470 }, { "epoch": 0.015203892196402278, "grad_norm": 6.057685852050781, "learning_rate": 4.998267806914812e-05, "loss": 0.811, "num_input_tokens_seen": 644128, "step": 475 }, { "epoch": 0.015363933166890723, "grad_norm": 3.036181926727295, "learning_rate": 4.998231074346378e-05, "loss": 0.9155, "num_input_tokens_seen": 650416, "step": 480 }, { "epoch": 0.015523974137379168, "grad_norm": 1.9813753366470337, "learning_rate": 4.998193956519317e-05, "loss": 0.7974, "num_input_tokens_seen": 656896, "step": 485 }, { "epoch": 0.015684015107867613, "grad_norm": 1.6644023656845093, "learning_rate": 4.9981564534393545e-05, "loss": 0.8807, "num_input_tokens_seen": 663440, "step": 490 }, { "epoch": 0.01584405607835606, "grad_norm": 2.3722410202026367, "learning_rate": 4.998118565112272e-05, "loss": 0.964, "num_input_tokens_seen": 670128, "step": 495 }, { "epoch": 0.016004097048844503, "grad_norm": 2.670053720474243, "learning_rate": 4.998080291543914e-05, "loss": 1.2029, "num_input_tokens_seen": 676656, "step": 500 }, { "epoch": 0.01616413801933295, "grad_norm": 1.7139312028884888, "learning_rate": 4.9980416327401826e-05, "loss": 0.8691, "num_input_tokens_seen": 683392, "step": 505 }, { "epoch": 0.016324178989821393, "grad_norm": 1.8376898765563965, "learning_rate": 4.998002588707038e-05, "loss": 0.7164, "num_input_tokens_seen": 690496, "step": 510 }, { "epoch": 0.01648421996030984, "grad_norm": 1.8574295043945312, "learning_rate": 4.997963159450503e-05, "loss": 0.5865, "num_input_tokens_seen": 697632, "step": 515 }, { "epoch": 0.016644260930798283, "grad_norm": 4.908980846405029, "learning_rate": 4.9979233449766575e-05, "loss": 0.8008, "num_input_tokens_seen": 705136, "step": 520 }, { "epoch": 0.01680430190128673, "grad_norm": 2.0985333919525146, "learning_rate": 4.997883145291641e-05, "loss": 1.2376, "num_input_tokens_seen": 711712, "step": 525 }, { "epoch": 0.016964342871775173, "grad_norm": 2.659440755844116, "learning_rate": 4.9978425604016536e-05, "loss": 0.9338, "num_input_tokens_seen": 718688, "step": 530 }, { "epoch": 0.01712438384226362, "grad_norm": 3.3857309818267822, "learning_rate": 4.9978015903129536e-05, "loss": 1.0381, "num_input_tokens_seen": 725664, "step": 535 }, { "epoch": 0.017284424812752063, "grad_norm": 6.8295578956604, "learning_rate": 4.997760235031859e-05, "loss": 1.9325, "num_input_tokens_seen": 732336, "step": 540 }, { "epoch": 0.01744446578324051, "grad_norm": 3.6256051063537598, "learning_rate": 4.9977184945647473e-05, "loss": 0.9061, "num_input_tokens_seen": 739024, "step": 545 }, { "epoch": 0.017604506753728953, "grad_norm": 4.776090621948242, "learning_rate": 4.997676368918055e-05, "loss": 0.9882, "num_input_tokens_seen": 745952, "step": 550 }, { "epoch": 0.0177645477242174, "grad_norm": 2.826582670211792, "learning_rate": 4.9976338580982794e-05, "loss": 1.1425, "num_input_tokens_seen": 752336, "step": 555 }, { "epoch": 0.017924588694705843, "grad_norm": 2.3464062213897705, "learning_rate": 4.9975909621119755e-05, "loss": 1.2225, "num_input_tokens_seen": 759616, "step": 560 }, { "epoch": 0.01808462966519429, "grad_norm": 0.9105109572410583, "learning_rate": 4.997547680965758e-05, "loss": 0.9557, "num_input_tokens_seen": 766768, "step": 565 }, { "epoch": 0.018244670635682733, "grad_norm": 2.127803087234497, "learning_rate": 4.997504014666302e-05, "loss": 0.6445, "num_input_tokens_seen": 773200, "step": 570 }, { "epoch": 0.01840471160617118, "grad_norm": 2.10912823677063, "learning_rate": 4.997459963220342e-05, "loss": 1.0027, "num_input_tokens_seen": 779920, "step": 575 }, { "epoch": 0.018564752576659623, "grad_norm": 2.0040628910064697, "learning_rate": 4.997415526634671e-05, "loss": 0.6426, "num_input_tokens_seen": 786688, "step": 580 }, { "epoch": 0.01872479354714807, "grad_norm": 0.9174013733863831, "learning_rate": 4.99737070491614e-05, "loss": 1.0805, "num_input_tokens_seen": 793840, "step": 585 }, { "epoch": 0.018884834517636517, "grad_norm": 2.972369432449341, "learning_rate": 4.997325498071663e-05, "loss": 0.9798, "num_input_tokens_seen": 800448, "step": 590 }, { "epoch": 0.01904487548812496, "grad_norm": 3.182211399078369, "learning_rate": 4.997279906108211e-05, "loss": 1.2379, "num_input_tokens_seen": 807456, "step": 595 }, { "epoch": 0.019204916458613407, "grad_norm": 3.074709892272949, "learning_rate": 4.9972339290328155e-05, "loss": 1.0905, "num_input_tokens_seen": 813648, "step": 600 }, { "epoch": 0.019204916458613407, "eval_loss": 0.9674872756004333, "eval_runtime": 500.2033, "eval_samples_per_second": 27.761, "eval_steps_per_second": 13.88, "num_input_tokens_seen": 813648, "step": 600 }, { "epoch": 0.01936495742910185, "grad_norm": 2.732778787612915, "learning_rate": 4.9971875668525646e-05, "loss": 0.9379, "num_input_tokens_seen": 820368, "step": 605 }, { "epoch": 0.019524998399590297, "grad_norm": 2.8296258449554443, "learning_rate": 4.997140819574609e-05, "loss": 1.1092, "num_input_tokens_seen": 827792, "step": 610 }, { "epoch": 0.01968503937007874, "grad_norm": 2.662261962890625, "learning_rate": 4.997093687206159e-05, "loss": 0.9656, "num_input_tokens_seen": 834256, "step": 615 }, { "epoch": 0.019845080340567187, "grad_norm": 2.740631341934204, "learning_rate": 4.997046169754482e-05, "loss": 0.6785, "num_input_tokens_seen": 840960, "step": 620 }, { "epoch": 0.02000512131105563, "grad_norm": 3.378535032272339, "learning_rate": 4.996998267226905e-05, "loss": 1.0644, "num_input_tokens_seen": 847616, "step": 625 }, { "epoch": 0.020165162281544077, "grad_norm": 4.176841735839844, "learning_rate": 4.996949979630817e-05, "loss": 0.7746, "num_input_tokens_seen": 854320, "step": 630 }, { "epoch": 0.02032520325203252, "grad_norm": 4.468045234680176, "learning_rate": 4.996901306973663e-05, "loss": 1.0154, "num_input_tokens_seen": 860752, "step": 635 }, { "epoch": 0.020485244222520967, "grad_norm": 2.9589319229125977, "learning_rate": 4.996852249262949e-05, "loss": 0.9261, "num_input_tokens_seen": 867792, "step": 640 }, { "epoch": 0.02064528519300941, "grad_norm": 4.132810115814209, "learning_rate": 4.996802806506241e-05, "loss": 1.1863, "num_input_tokens_seen": 874768, "step": 645 }, { "epoch": 0.020805326163497857, "grad_norm": 1.3483903408050537, "learning_rate": 4.996752978711164e-05, "loss": 0.9018, "num_input_tokens_seen": 881712, "step": 650 }, { "epoch": 0.0209653671339863, "grad_norm": 5.841209888458252, "learning_rate": 4.996702765885401e-05, "loss": 0.6772, "num_input_tokens_seen": 888848, "step": 655 }, { "epoch": 0.021125408104474747, "grad_norm": 3.240293025970459, "learning_rate": 4.9966521680366964e-05, "loss": 0.6774, "num_input_tokens_seen": 895168, "step": 660 }, { "epoch": 0.02128544907496319, "grad_norm": 1.513742446899414, "learning_rate": 4.9966011851728524e-05, "loss": 0.8691, "num_input_tokens_seen": 902224, "step": 665 }, { "epoch": 0.021445490045451637, "grad_norm": 3.5859735012054443, "learning_rate": 4.996549817301731e-05, "loss": 0.9052, "num_input_tokens_seen": 909136, "step": 670 }, { "epoch": 0.02160553101594008, "grad_norm": 2.586045026779175, "learning_rate": 4.9964980644312544e-05, "loss": 0.6873, "num_input_tokens_seen": 915776, "step": 675 }, { "epoch": 0.021765571986428527, "grad_norm": 1.9876937866210938, "learning_rate": 4.996445926569403e-05, "loss": 0.9151, "num_input_tokens_seen": 922544, "step": 680 }, { "epoch": 0.02192561295691697, "grad_norm": 3.87119197845459, "learning_rate": 4.996393403724218e-05, "loss": 0.8919, "num_input_tokens_seen": 929120, "step": 685 }, { "epoch": 0.022085653927405417, "grad_norm": 3.2684218883514404, "learning_rate": 4.9963404959037985e-05, "loss": 1.1815, "num_input_tokens_seen": 936528, "step": 690 }, { "epoch": 0.02224569489789386, "grad_norm": 4.776121616363525, "learning_rate": 4.996287203116303e-05, "loss": 0.8994, "num_input_tokens_seen": 943296, "step": 695 }, { "epoch": 0.022405735868382307, "grad_norm": 3.8519155979156494, "learning_rate": 4.996233525369951e-05, "loss": 1.1514, "num_input_tokens_seen": 949616, "step": 700 }, { "epoch": 0.02256577683887075, "grad_norm": 3.357659339904785, "learning_rate": 4.99617946267302e-05, "loss": 0.8955, "num_input_tokens_seen": 956048, "step": 705 }, { "epoch": 0.022725817809359197, "grad_norm": 1.5590137243270874, "learning_rate": 4.996125015033846e-05, "loss": 0.6783, "num_input_tokens_seen": 962736, "step": 710 }, { "epoch": 0.02288585877984764, "grad_norm": 2.5808844566345215, "learning_rate": 4.996070182460827e-05, "loss": 0.7479, "num_input_tokens_seen": 969856, "step": 715 }, { "epoch": 0.023045899750336087, "grad_norm": 3.72711181640625, "learning_rate": 4.996014964962418e-05, "loss": 0.6812, "num_input_tokens_seen": 976672, "step": 720 }, { "epoch": 0.02320594072082453, "grad_norm": 1.7476166486740112, "learning_rate": 4.9959593625471344e-05, "loss": 0.8789, "num_input_tokens_seen": 983696, "step": 725 }, { "epoch": 0.023365981691312977, "grad_norm": 2.3559341430664062, "learning_rate": 4.995903375223552e-05, "loss": 1.1109, "num_input_tokens_seen": 990512, "step": 730 }, { "epoch": 0.02352602266180142, "grad_norm": 1.7309989929199219, "learning_rate": 4.995847003000302e-05, "loss": 0.6937, "num_input_tokens_seen": 997312, "step": 735 }, { "epoch": 0.023686063632289867, "grad_norm": 2.2178313732147217, "learning_rate": 4.9957902458860804e-05, "loss": 0.8254, "num_input_tokens_seen": 1003712, "step": 740 }, { "epoch": 0.02384610460277831, "grad_norm": 1.5787416696548462, "learning_rate": 4.995733103889639e-05, "loss": 0.9664, "num_input_tokens_seen": 1010368, "step": 745 }, { "epoch": 0.024006145573266757, "grad_norm": 1.4440639019012451, "learning_rate": 4.99567557701979e-05, "loss": 0.9381, "num_input_tokens_seen": 1017152, "step": 750 }, { "epoch": 0.0241661865437552, "grad_norm": 0.9079903960227966, "learning_rate": 4.995617665285403e-05, "loss": 1.074, "num_input_tokens_seen": 1023808, "step": 755 }, { "epoch": 0.024326227514243647, "grad_norm": 1.9543590545654297, "learning_rate": 4.99555936869541e-05, "loss": 0.7141, "num_input_tokens_seen": 1030656, "step": 760 }, { "epoch": 0.02448626848473209, "grad_norm": 2.0576348304748535, "learning_rate": 4.995500687258803e-05, "loss": 1.1623, "num_input_tokens_seen": 1037696, "step": 765 }, { "epoch": 0.024646309455220537, "grad_norm": 1.4402270317077637, "learning_rate": 4.995441620984628e-05, "loss": 0.6925, "num_input_tokens_seen": 1044240, "step": 770 }, { "epoch": 0.02480635042570898, "grad_norm": 2.8482413291931152, "learning_rate": 4.995382169881996e-05, "loss": 1.1764, "num_input_tokens_seen": 1050528, "step": 775 }, { "epoch": 0.024966391396197427, "grad_norm": 4.411271572113037, "learning_rate": 4.9953223339600755e-05, "loss": 0.8092, "num_input_tokens_seen": 1057408, "step": 780 }, { "epoch": 0.025126432366685873, "grad_norm": 2.0938594341278076, "learning_rate": 4.995262113228091e-05, "loss": 0.8183, "num_input_tokens_seen": 1063984, "step": 785 }, { "epoch": 0.025286473337174317, "grad_norm": 3.648866891860962, "learning_rate": 4.995201507695332e-05, "loss": 0.7361, "num_input_tokens_seen": 1070640, "step": 790 }, { "epoch": 0.025446514307662763, "grad_norm": 3.21466064453125, "learning_rate": 4.995140517371144e-05, "loss": 0.7999, "num_input_tokens_seen": 1077456, "step": 795 }, { "epoch": 0.025606555278151207, "grad_norm": 1.7965236902236938, "learning_rate": 4.995079142264932e-05, "loss": 0.6506, "num_input_tokens_seen": 1084496, "step": 800 }, { "epoch": 0.025606555278151207, "eval_loss": 0.852758526802063, "eval_runtime": 499.9099, "eval_samples_per_second": 27.777, "eval_steps_per_second": 13.889, "num_input_tokens_seen": 1084496, "step": 800 }, { "epoch": 0.025766596248639653, "grad_norm": 2.5905487537384033, "learning_rate": 4.995017382386162e-05, "loss": 1.1995, "num_input_tokens_seen": 1091504, "step": 805 }, { "epoch": 0.025926637219128097, "grad_norm": 2.5093581676483154, "learning_rate": 4.994955237744356e-05, "loss": 0.7933, "num_input_tokens_seen": 1098496, "step": 810 }, { "epoch": 0.026086678189616543, "grad_norm": 1.3717204332351685, "learning_rate": 4.994892708349101e-05, "loss": 0.832, "num_input_tokens_seen": 1105024, "step": 815 }, { "epoch": 0.026246719160104987, "grad_norm": 3.032470464706421, "learning_rate": 4.994829794210035e-05, "loss": 0.8264, "num_input_tokens_seen": 1111408, "step": 820 }, { "epoch": 0.026406760130593433, "grad_norm": 1.2625950574874878, "learning_rate": 4.994766495336864e-05, "loss": 0.533, "num_input_tokens_seen": 1118496, "step": 825 }, { "epoch": 0.026566801101081877, "grad_norm": 1.5072897672653198, "learning_rate": 4.994702811739348e-05, "loss": 0.6874, "num_input_tokens_seen": 1125504, "step": 830 }, { "epoch": 0.026726842071570323, "grad_norm": 1.216132402420044, "learning_rate": 4.994638743427308e-05, "loss": 0.5937, "num_input_tokens_seen": 1132544, "step": 835 }, { "epoch": 0.026886883042058767, "grad_norm": 1.4909512996673584, "learning_rate": 4.994574290410624e-05, "loss": 0.8412, "num_input_tokens_seen": 1139296, "step": 840 }, { "epoch": 0.027046924012547213, "grad_norm": 2.3725953102111816, "learning_rate": 4.9945094526992364e-05, "loss": 0.8614, "num_input_tokens_seen": 1145696, "step": 845 }, { "epoch": 0.027206964983035656, "grad_norm": 2.6574463844299316, "learning_rate": 4.994444230303142e-05, "loss": 0.8758, "num_input_tokens_seen": 1152416, "step": 850 }, { "epoch": 0.027367005953524103, "grad_norm": 2.2465133666992188, "learning_rate": 4.994378623232402e-05, "loss": 0.6822, "num_input_tokens_seen": 1159072, "step": 855 }, { "epoch": 0.027527046924012546, "grad_norm": 1.344522476196289, "learning_rate": 4.99431263149713e-05, "loss": 0.9155, "num_input_tokens_seen": 1165696, "step": 860 }, { "epoch": 0.027687087894500993, "grad_norm": 2.560678482055664, "learning_rate": 4.9942462551075056e-05, "loss": 1.1908, "num_input_tokens_seen": 1172416, "step": 865 }, { "epoch": 0.027847128864989436, "grad_norm": 2.1276960372924805, "learning_rate": 4.994179494073764e-05, "loss": 0.9527, "num_input_tokens_seen": 1179136, "step": 870 }, { "epoch": 0.028007169835477883, "grad_norm": 2.931952714920044, "learning_rate": 4.9941123484062e-05, "loss": 0.9433, "num_input_tokens_seen": 1186144, "step": 875 }, { "epoch": 0.028167210805966326, "grad_norm": 1.9368730783462524, "learning_rate": 4.99404481811517e-05, "loss": 0.8234, "num_input_tokens_seen": 1192624, "step": 880 }, { "epoch": 0.028327251776454773, "grad_norm": 2.4757330417633057, "learning_rate": 4.9939769032110864e-05, "loss": 0.8315, "num_input_tokens_seen": 1199088, "step": 885 }, { "epoch": 0.028487292746943216, "grad_norm": 1.3337767124176025, "learning_rate": 4.993908603704423e-05, "loss": 0.7634, "num_input_tokens_seen": 1205520, "step": 890 }, { "epoch": 0.028647333717431663, "grad_norm": 2.3001906871795654, "learning_rate": 4.9938399196057126e-05, "loss": 0.7627, "num_input_tokens_seen": 1212288, "step": 895 }, { "epoch": 0.028807374687920106, "grad_norm": 1.755315899848938, "learning_rate": 4.993770850925547e-05, "loss": 0.9215, "num_input_tokens_seen": 1219104, "step": 900 }, { "epoch": 0.028967415658408553, "grad_norm": 1.757630705833435, "learning_rate": 4.993701397674577e-05, "loss": 0.9201, "num_input_tokens_seen": 1225728, "step": 905 }, { "epoch": 0.029127456628896996, "grad_norm": 2.298445224761963, "learning_rate": 4.993631559863515e-05, "loss": 0.6608, "num_input_tokens_seen": 1232592, "step": 910 }, { "epoch": 0.029287497599385443, "grad_norm": 0.9833984971046448, "learning_rate": 4.9935613375031283e-05, "loss": 0.6748, "num_input_tokens_seen": 1239056, "step": 915 }, { "epoch": 0.029447538569873886, "grad_norm": 2.1960432529449463, "learning_rate": 4.993490730604248e-05, "loss": 0.9236, "num_input_tokens_seen": 1245776, "step": 920 }, { "epoch": 0.029607579540362333, "grad_norm": 2.133148431777954, "learning_rate": 4.993419739177761e-05, "loss": 0.9215, "num_input_tokens_seen": 1252896, "step": 925 }, { "epoch": 0.029767620510850776, "grad_norm": 1.5579707622528076, "learning_rate": 4.9933483632346164e-05, "loss": 0.7278, "num_input_tokens_seen": 1259840, "step": 930 }, { "epoch": 0.029927661481339223, "grad_norm": 1.6195976734161377, "learning_rate": 4.993276602785821e-05, "loss": 0.7485, "num_input_tokens_seen": 1266336, "step": 935 }, { "epoch": 0.030087702451827666, "grad_norm": 1.801526665687561, "learning_rate": 4.993204457842441e-05, "loss": 0.79, "num_input_tokens_seen": 1272720, "step": 940 }, { "epoch": 0.030247743422316113, "grad_norm": 1.8278778791427612, "learning_rate": 4.993131928415602e-05, "loss": 1.2293, "num_input_tokens_seen": 1279072, "step": 945 }, { "epoch": 0.030407784392804556, "grad_norm": 1.6149331331253052, "learning_rate": 4.993059014516489e-05, "loss": 0.8734, "num_input_tokens_seen": 1285712, "step": 950 }, { "epoch": 0.030567825363293003, "grad_norm": 3.6033174991607666, "learning_rate": 4.9929857161563464e-05, "loss": 0.9133, "num_input_tokens_seen": 1293488, "step": 955 }, { "epoch": 0.030727866333781446, "grad_norm": 1.8702874183654785, "learning_rate": 4.992912033346477e-05, "loss": 0.6814, "num_input_tokens_seen": 1300432, "step": 960 }, { "epoch": 0.030887907304269893, "grad_norm": 2.0555481910705566, "learning_rate": 4.992837966098245e-05, "loss": 0.7365, "num_input_tokens_seen": 1307568, "step": 965 }, { "epoch": 0.031047948274758336, "grad_norm": 1.7290964126586914, "learning_rate": 4.992763514423071e-05, "loss": 1.1769, "num_input_tokens_seen": 1314464, "step": 970 }, { "epoch": 0.031207989245246783, "grad_norm": 1.3037787675857544, "learning_rate": 4.992688678332437e-05, "loss": 0.6527, "num_input_tokens_seen": 1321120, "step": 975 }, { "epoch": 0.031368030215735226, "grad_norm": 1.952317714691162, "learning_rate": 4.992613457837884e-05, "loss": 0.8675, "num_input_tokens_seen": 1327680, "step": 980 }, { "epoch": 0.03152807118622367, "grad_norm": 2.548884868621826, "learning_rate": 4.992537852951011e-05, "loss": 0.5239, "num_input_tokens_seen": 1334576, "step": 985 }, { "epoch": 0.03168811215671212, "grad_norm": 1.9931682348251343, "learning_rate": 4.9924618636834785e-05, "loss": 0.5883, "num_input_tokens_seen": 1341824, "step": 990 }, { "epoch": 0.03184815312720057, "grad_norm": 2.3997437953948975, "learning_rate": 4.9923854900470046e-05, "loss": 0.9502, "num_input_tokens_seen": 1348512, "step": 995 }, { "epoch": 0.032008194097689006, "grad_norm": 2.4233994483947754, "learning_rate": 4.992308732053367e-05, "loss": 0.5621, "num_input_tokens_seen": 1355472, "step": 1000 }, { "epoch": 0.032008194097689006, "eval_loss": 0.783753514289856, "eval_runtime": 500.3135, "eval_samples_per_second": 27.755, "eval_steps_per_second": 13.877, "num_input_tokens_seen": 1355472, "step": 1000 }, { "epoch": 0.03216823506817745, "grad_norm": 1.9165464639663696, "learning_rate": 4.992231589714402e-05, "loss": 0.929, "num_input_tokens_seen": 1362192, "step": 1005 }, { "epoch": 0.0323282760386659, "grad_norm": 1.5531820058822632, "learning_rate": 4.992154063042007e-05, "loss": 0.7933, "num_input_tokens_seen": 1369040, "step": 1010 }, { "epoch": 0.03248831700915435, "grad_norm": 1.153718113899231, "learning_rate": 4.992076152048136e-05, "loss": 0.6429, "num_input_tokens_seen": 1375616, "step": 1015 }, { "epoch": 0.032648357979642786, "grad_norm": 1.4072681665420532, "learning_rate": 4.991997856744807e-05, "loss": 1.1251, "num_input_tokens_seen": 1382240, "step": 1020 }, { "epoch": 0.03280839895013123, "grad_norm": 2.2868101596832275, "learning_rate": 4.9919191771440905e-05, "loss": 0.8096, "num_input_tokens_seen": 1388752, "step": 1025 }, { "epoch": 0.03296843992061968, "grad_norm": 2.354259967803955, "learning_rate": 4.991840113258122e-05, "loss": 1.0128, "num_input_tokens_seen": 1395648, "step": 1030 }, { "epoch": 0.03312848089110813, "grad_norm": 1.6359614133834839, "learning_rate": 4.9917606650990933e-05, "loss": 0.4854, "num_input_tokens_seen": 1402480, "step": 1035 }, { "epoch": 0.033288521861596566, "grad_norm": 2.762009859085083, "learning_rate": 4.9916808326792566e-05, "loss": 0.9316, "num_input_tokens_seen": 1409552, "step": 1040 }, { "epoch": 0.03344856283208501, "grad_norm": 2.3339738845825195, "learning_rate": 4.9916006160109235e-05, "loss": 0.6684, "num_input_tokens_seen": 1416160, "step": 1045 }, { "epoch": 0.03360860380257346, "grad_norm": 1.7853730916976929, "learning_rate": 4.991520015106464e-05, "loss": 0.8729, "num_input_tokens_seen": 1423072, "step": 1050 }, { "epoch": 0.03376864477306191, "grad_norm": 1.4317148923873901, "learning_rate": 4.991439029978308e-05, "loss": 0.7487, "num_input_tokens_seen": 1429680, "step": 1055 }, { "epoch": 0.033928685743550346, "grad_norm": 2.0221827030181885, "learning_rate": 4.9913576606389434e-05, "loss": 0.6754, "num_input_tokens_seen": 1436096, "step": 1060 }, { "epoch": 0.03408872671403879, "grad_norm": 1.2689259052276611, "learning_rate": 4.991275907100919e-05, "loss": 0.6677, "num_input_tokens_seen": 1442704, "step": 1065 }, { "epoch": 0.03424876768452724, "grad_norm": 2.246431589126587, "learning_rate": 4.9911937693768434e-05, "loss": 0.7536, "num_input_tokens_seen": 1449472, "step": 1070 }, { "epoch": 0.03440880865501569, "grad_norm": 0.9613668918609619, "learning_rate": 4.991111247479382e-05, "loss": 0.71, "num_input_tokens_seen": 1456560, "step": 1075 }, { "epoch": 0.034568849625504126, "grad_norm": 2.3319127559661865, "learning_rate": 4.9910283414212605e-05, "loss": 0.6824, "num_input_tokens_seen": 1463232, "step": 1080 }, { "epoch": 0.03472889059599257, "grad_norm": 1.9184538125991821, "learning_rate": 4.990945051215265e-05, "loss": 0.9386, "num_input_tokens_seen": 1469904, "step": 1085 }, { "epoch": 0.03488893156648102, "grad_norm": 1.5281503200531006, "learning_rate": 4.99086137687424e-05, "loss": 0.6121, "num_input_tokens_seen": 1476464, "step": 1090 }, { "epoch": 0.03504897253696947, "grad_norm": 1.058401346206665, "learning_rate": 4.9907773184110874e-05, "loss": 0.7115, "num_input_tokens_seen": 1483248, "step": 1095 }, { "epoch": 0.035209013507457906, "grad_norm": 1.4014394283294678, "learning_rate": 4.9906928758387715e-05, "loss": 0.5775, "num_input_tokens_seen": 1490464, "step": 1100 }, { "epoch": 0.03536905447794635, "grad_norm": 1.2033225297927856, "learning_rate": 4.9906080491703146e-05, "loss": 0.7893, "num_input_tokens_seen": 1497328, "step": 1105 }, { "epoch": 0.0355290954484348, "grad_norm": 2.4747986793518066, "learning_rate": 4.990522838418797e-05, "loss": 0.992, "num_input_tokens_seen": 1503808, "step": 1110 }, { "epoch": 0.03568913641892325, "grad_norm": 2.2361409664154053, "learning_rate": 4.9904372435973604e-05, "loss": 0.8122, "num_input_tokens_seen": 1510336, "step": 1115 }, { "epoch": 0.035849177389411686, "grad_norm": 0.8653162717819214, "learning_rate": 4.990351264719203e-05, "loss": 0.6984, "num_input_tokens_seen": 1517392, "step": 1120 }, { "epoch": 0.03600921835990013, "grad_norm": 1.1116927862167358, "learning_rate": 4.990264901797586e-05, "loss": 0.5205, "num_input_tokens_seen": 1524032, "step": 1125 }, { "epoch": 0.03616925933038858, "grad_norm": 2.284291982650757, "learning_rate": 4.990178154845826e-05, "loss": 0.6283, "num_input_tokens_seen": 1530944, "step": 1130 }, { "epoch": 0.03632930030087703, "grad_norm": 2.9557135105133057, "learning_rate": 4.9900910238773014e-05, "loss": 0.9726, "num_input_tokens_seen": 1537328, "step": 1135 }, { "epoch": 0.036489341271365466, "grad_norm": 1.510709524154663, "learning_rate": 4.990003508905448e-05, "loss": 0.7199, "num_input_tokens_seen": 1544112, "step": 1140 }, { "epoch": 0.03664938224185391, "grad_norm": 1.3763887882232666, "learning_rate": 4.989915609943763e-05, "loss": 0.7857, "num_input_tokens_seen": 1550640, "step": 1145 }, { "epoch": 0.03680942321234236, "grad_norm": 1.4859611988067627, "learning_rate": 4.9898273270058e-05, "loss": 0.619, "num_input_tokens_seen": 1557440, "step": 1150 }, { "epoch": 0.03696946418283081, "grad_norm": 1.565308928489685, "learning_rate": 4.989738660105174e-05, "loss": 0.9413, "num_input_tokens_seen": 1563920, "step": 1155 }, { "epoch": 0.037129505153319246, "grad_norm": 3.783872127532959, "learning_rate": 4.989649609255559e-05, "loss": 0.9164, "num_input_tokens_seen": 1570800, "step": 1160 }, { "epoch": 0.03728954612380769, "grad_norm": 1.3093242645263672, "learning_rate": 4.989560174470687e-05, "loss": 0.5309, "num_input_tokens_seen": 1577408, "step": 1165 }, { "epoch": 0.03744958709429614, "grad_norm": 2.2422001361846924, "learning_rate": 4.989470355764351e-05, "loss": 0.7887, "num_input_tokens_seen": 1583792, "step": 1170 }, { "epoch": 0.03760962806478459, "grad_norm": 0.9357190132141113, "learning_rate": 4.9893801531504e-05, "loss": 0.6372, "num_input_tokens_seen": 1590432, "step": 1175 }, { "epoch": 0.03776966903527303, "grad_norm": 1.1340535879135132, "learning_rate": 4.9892895666427475e-05, "loss": 0.6118, "num_input_tokens_seen": 1597248, "step": 1180 }, { "epoch": 0.03792971000576147, "grad_norm": 3.3373711109161377, "learning_rate": 4.9891985962553606e-05, "loss": 0.8148, "num_input_tokens_seen": 1603952, "step": 1185 }, { "epoch": 0.03808975097624992, "grad_norm": 1.501222848892212, "learning_rate": 4.989107242002269e-05, "loss": 0.4074, "num_input_tokens_seen": 1610672, "step": 1190 }, { "epoch": 0.038249791946738367, "grad_norm": 1.551308512687683, "learning_rate": 4.989015503897561e-05, "loss": 0.6384, "num_input_tokens_seen": 1617520, "step": 1195 }, { "epoch": 0.03840983291722681, "grad_norm": 1.3420650959014893, "learning_rate": 4.988923381955383e-05, "loss": 0.7799, "num_input_tokens_seen": 1624048, "step": 1200 }, { "epoch": 0.03840983291722681, "eval_loss": 0.7391271591186523, "eval_runtime": 500.0394, "eval_samples_per_second": 27.77, "eval_steps_per_second": 13.885, "num_input_tokens_seen": 1624048, "step": 1200 }, { "epoch": 0.03856987388771525, "grad_norm": 1.4659066200256348, "learning_rate": 4.988830876189942e-05, "loss": 0.6625, "num_input_tokens_seen": 1630848, "step": 1205 }, { "epoch": 0.0387299148582037, "grad_norm": 1.1772994995117188, "learning_rate": 4.988737986615503e-05, "loss": 0.8427, "num_input_tokens_seen": 1637504, "step": 1210 }, { "epoch": 0.038889955828692147, "grad_norm": 1.2651716470718384, "learning_rate": 4.988644713246391e-05, "loss": 0.5674, "num_input_tokens_seen": 1644272, "step": 1215 }, { "epoch": 0.03904999679918059, "grad_norm": 1.331751823425293, "learning_rate": 4.988551056096991e-05, "loss": 0.7974, "num_input_tokens_seen": 1650992, "step": 1220 }, { "epoch": 0.03921003776966903, "grad_norm": 0.6829292178153992, "learning_rate": 4.988457015181743e-05, "loss": 0.5599, "num_input_tokens_seen": 1657840, "step": 1225 }, { "epoch": 0.03937007874015748, "grad_norm": 1.5594487190246582, "learning_rate": 4.988362590515153e-05, "loss": 0.5258, "num_input_tokens_seen": 1664608, "step": 1230 }, { "epoch": 0.039530119710645927, "grad_norm": 1.945160150527954, "learning_rate": 4.9882677821117805e-05, "loss": 0.5502, "num_input_tokens_seen": 1671568, "step": 1235 }, { "epoch": 0.03969016068113437, "grad_norm": 2.059187889099121, "learning_rate": 4.988172589986246e-05, "loss": 0.6496, "num_input_tokens_seen": 1678048, "step": 1240 }, { "epoch": 0.03985020165162281, "grad_norm": 2.0227110385894775, "learning_rate": 4.9880770141532304e-05, "loss": 0.5633, "num_input_tokens_seen": 1684688, "step": 1245 }, { "epoch": 0.04001024262211126, "grad_norm": 4.27987813949585, "learning_rate": 4.987981054627472e-05, "loss": 1.075, "num_input_tokens_seen": 1691360, "step": 1250 }, { "epoch": 0.040170283592599706, "grad_norm": 1.3451513051986694, "learning_rate": 4.987884711423769e-05, "loss": 0.7348, "num_input_tokens_seen": 1698064, "step": 1255 }, { "epoch": 0.04033032456308815, "grad_norm": 1.5368647575378418, "learning_rate": 4.9877879845569784e-05, "loss": 0.6079, "num_input_tokens_seen": 1705040, "step": 1260 }, { "epoch": 0.04049036553357659, "grad_norm": 0.9459667801856995, "learning_rate": 4.9876908740420175e-05, "loss": 0.5858, "num_input_tokens_seen": 1711952, "step": 1265 }, { "epoch": 0.04065040650406504, "grad_norm": 1.0551687479019165, "learning_rate": 4.987593379893861e-05, "loss": 0.6067, "num_input_tokens_seen": 1718480, "step": 1270 }, { "epoch": 0.040810447474553486, "grad_norm": 2.313131809234619, "learning_rate": 4.987495502127545e-05, "loss": 0.6632, "num_input_tokens_seen": 1725264, "step": 1275 }, { "epoch": 0.04097048844504193, "grad_norm": 1.7927626371383667, "learning_rate": 4.987397240758162e-05, "loss": 0.8062, "num_input_tokens_seen": 1731792, "step": 1280 }, { "epoch": 0.04113052941553037, "grad_norm": 0.9816111922264099, "learning_rate": 4.9872985958008664e-05, "loss": 0.5557, "num_input_tokens_seen": 1738880, "step": 1285 }, { "epoch": 0.04129057038601882, "grad_norm": 1.6864323616027832, "learning_rate": 4.987199567270871e-05, "loss": 0.7636, "num_input_tokens_seen": 1745712, "step": 1290 }, { "epoch": 0.041450611356507266, "grad_norm": 1.8753619194030762, "learning_rate": 4.9871001551834444e-05, "loss": 0.506, "num_input_tokens_seen": 1752400, "step": 1295 }, { "epoch": 0.04161065232699571, "grad_norm": 3.3018600940704346, "learning_rate": 4.98700035955392e-05, "loss": 0.8308, "num_input_tokens_seen": 1758960, "step": 1300 }, { "epoch": 0.04177069329748415, "grad_norm": 2.9311769008636475, "learning_rate": 4.986900180397686e-05, "loss": 0.89, "num_input_tokens_seen": 1765840, "step": 1305 }, { "epoch": 0.0419307342679726, "grad_norm": 1.8081365823745728, "learning_rate": 4.9867996177301926e-05, "loss": 0.6188, "num_input_tokens_seen": 1772208, "step": 1310 }, { "epoch": 0.042090775238461046, "grad_norm": 3.6146719455718994, "learning_rate": 4.9866986715669464e-05, "loss": 0.8147, "num_input_tokens_seen": 1778976, "step": 1315 }, { "epoch": 0.04225081620894949, "grad_norm": 1.4383448362350464, "learning_rate": 4.9865973419235155e-05, "loss": 0.5111, "num_input_tokens_seen": 1786400, "step": 1320 }, { "epoch": 0.04241085717943793, "grad_norm": 1.979627013206482, "learning_rate": 4.986495628815526e-05, "loss": 0.8136, "num_input_tokens_seen": 1793200, "step": 1325 }, { "epoch": 0.04257089814992638, "grad_norm": 0.8688178062438965, "learning_rate": 4.986393532258663e-05, "loss": 0.4571, "num_input_tokens_seen": 1799824, "step": 1330 }, { "epoch": 0.042730939120414826, "grad_norm": 1.2468594312667847, "learning_rate": 4.986291052268671e-05, "loss": 0.6806, "num_input_tokens_seen": 1806880, "step": 1335 }, { "epoch": 0.04289098009090327, "grad_norm": 2.63332462310791, "learning_rate": 4.986188188861355e-05, "loss": 0.7428, "num_input_tokens_seen": 1813328, "step": 1340 }, { "epoch": 0.04305102106139171, "grad_norm": 2.032743215560913, "learning_rate": 4.9860849420525766e-05, "loss": 0.6476, "num_input_tokens_seen": 1819904, "step": 1345 }, { "epoch": 0.04321106203188016, "grad_norm": 1.991755485534668, "learning_rate": 4.9859813118582575e-05, "loss": 0.6405, "num_input_tokens_seen": 1826896, "step": 1350 }, { "epoch": 0.043371103002368606, "grad_norm": 1.4239627122879028, "learning_rate": 4.98587729829438e-05, "loss": 0.5557, "num_input_tokens_seen": 1833520, "step": 1355 }, { "epoch": 0.04353114397285705, "grad_norm": 2.805006504058838, "learning_rate": 4.985772901376983e-05, "loss": 0.5965, "num_input_tokens_seen": 1840032, "step": 1360 }, { "epoch": 0.0436911849433455, "grad_norm": 1.8132771253585815, "learning_rate": 4.9856681211221666e-05, "loss": 0.7542, "num_input_tokens_seen": 1846464, "step": 1365 }, { "epoch": 0.04385122591383394, "grad_norm": 5.047675609588623, "learning_rate": 4.985562957546089e-05, "loss": 0.7864, "num_input_tokens_seen": 1853296, "step": 1370 }, { "epoch": 0.044011266884322386, "grad_norm": 4.057877063751221, "learning_rate": 4.9854574106649686e-05, "loss": 0.6631, "num_input_tokens_seen": 1859904, "step": 1375 }, { "epoch": 0.04417130785481083, "grad_norm": 0.9616785645484924, "learning_rate": 4.985351480495081e-05, "loss": 0.6268, "num_input_tokens_seen": 1867024, "step": 1380 }, { "epoch": 0.04433134882529928, "grad_norm": 1.3877830505371094, "learning_rate": 4.985245167052762e-05, "loss": 0.8139, "num_input_tokens_seen": 1873664, "step": 1385 }, { "epoch": 0.04449138979578772, "grad_norm": 1.8002235889434814, "learning_rate": 4.9851384703544066e-05, "loss": 0.7169, "num_input_tokens_seen": 1880320, "step": 1390 }, { "epoch": 0.044651430766276166, "grad_norm": 1.517683744430542, "learning_rate": 4.985031390416469e-05, "loss": 0.8611, "num_input_tokens_seen": 1887168, "step": 1395 }, { "epoch": 0.04481147173676461, "grad_norm": 3.30706524848938, "learning_rate": 4.984923927255461e-05, "loss": 0.7229, "num_input_tokens_seen": 1893968, "step": 1400 }, { "epoch": 0.04481147173676461, "eval_loss": 0.7115038633346558, "eval_runtime": 500.0182, "eval_samples_per_second": 27.771, "eval_steps_per_second": 13.885, "num_input_tokens_seen": 1893968, "step": 1400 }, { "epoch": 0.04497151270725306, "grad_norm": 1.4948740005493164, "learning_rate": 4.984816080887958e-05, "loss": 0.61, "num_input_tokens_seen": 1900480, "step": 1405 }, { "epoch": 0.0451315536777415, "grad_norm": 2.4947400093078613, "learning_rate": 4.9847078513305875e-05, "loss": 0.8499, "num_input_tokens_seen": 1907296, "step": 1410 }, { "epoch": 0.045291594648229946, "grad_norm": 1.5645006895065308, "learning_rate": 4.984599238600043e-05, "loss": 0.6845, "num_input_tokens_seen": 1913744, "step": 1415 }, { "epoch": 0.04545163561871839, "grad_norm": 2.0254640579223633, "learning_rate": 4.9844902427130716e-05, "loss": 0.4733, "num_input_tokens_seen": 1920304, "step": 1420 }, { "epoch": 0.04561167658920684, "grad_norm": 1.6247129440307617, "learning_rate": 4.984380863686482e-05, "loss": 0.6108, "num_input_tokens_seen": 1926736, "step": 1425 }, { "epoch": 0.04577171755969528, "grad_norm": 1.4100557565689087, "learning_rate": 4.984271101537143e-05, "loss": 0.6489, "num_input_tokens_seen": 1933792, "step": 1430 }, { "epoch": 0.045931758530183726, "grad_norm": 1.272983431816101, "learning_rate": 4.9841609562819816e-05, "loss": 0.729, "num_input_tokens_seen": 1940176, "step": 1435 }, { "epoch": 0.04609179950067217, "grad_norm": 1.1890103816986084, "learning_rate": 4.984050427937983e-05, "loss": 0.6292, "num_input_tokens_seen": 1946784, "step": 1440 }, { "epoch": 0.04625184047116062, "grad_norm": 1.382767915725708, "learning_rate": 4.983939516522191e-05, "loss": 0.5415, "num_input_tokens_seen": 1953792, "step": 1445 }, { "epoch": 0.04641188144164906, "grad_norm": 1.3623853921890259, "learning_rate": 4.983828222051711e-05, "loss": 0.6454, "num_input_tokens_seen": 1960704, "step": 1450 }, { "epoch": 0.046571922412137506, "grad_norm": 0.8900788426399231, "learning_rate": 4.983716544543705e-05, "loss": 0.6482, "num_input_tokens_seen": 1967280, "step": 1455 }, { "epoch": 0.04673196338262595, "grad_norm": 0.8309162259101868, "learning_rate": 4.983604484015395e-05, "loss": 0.6615, "num_input_tokens_seen": 1973936, "step": 1460 }, { "epoch": 0.0468920043531144, "grad_norm": 1.5014712810516357, "learning_rate": 4.983492040484064e-05, "loss": 0.8117, "num_input_tokens_seen": 1980624, "step": 1465 }, { "epoch": 0.04705204532360284, "grad_norm": 1.6353802680969238, "learning_rate": 4.98337921396705e-05, "loss": 0.5266, "num_input_tokens_seen": 1987056, "step": 1470 }, { "epoch": 0.047212086294091286, "grad_norm": 1.8683596849441528, "learning_rate": 4.983266004481753e-05, "loss": 0.733, "num_input_tokens_seen": 1993888, "step": 1475 }, { "epoch": 0.04737212726457973, "grad_norm": 0.9962350726127625, "learning_rate": 4.9831524120456316e-05, "loss": 0.6915, "num_input_tokens_seen": 2001008, "step": 1480 }, { "epoch": 0.04753216823506818, "grad_norm": 1.7482852935791016, "learning_rate": 4.9830384366762026e-05, "loss": 0.757, "num_input_tokens_seen": 2007312, "step": 1485 }, { "epoch": 0.04769220920555662, "grad_norm": 1.5035279989242554, "learning_rate": 4.9829240783910436e-05, "loss": 0.7016, "num_input_tokens_seen": 2013904, "step": 1490 }, { "epoch": 0.047852250176045066, "grad_norm": 3.2971935272216797, "learning_rate": 4.982809337207789e-05, "loss": 0.9514, "num_input_tokens_seen": 2020384, "step": 1495 }, { "epoch": 0.04801229114653351, "grad_norm": 2.1334195137023926, "learning_rate": 4.9826942131441337e-05, "loss": 0.6444, "num_input_tokens_seen": 2027104, "step": 1500 }, { "epoch": 0.04817233211702196, "grad_norm": 1.9241082668304443, "learning_rate": 4.9825787062178315e-05, "loss": 0.6024, "num_input_tokens_seen": 2034240, "step": 1505 }, { "epoch": 0.0483323730875104, "grad_norm": 1.592378854751587, "learning_rate": 4.9824628164466945e-05, "loss": 0.5526, "num_input_tokens_seen": 2040672, "step": 1510 }, { "epoch": 0.048492414057998846, "grad_norm": 2.911402940750122, "learning_rate": 4.982346543848595e-05, "loss": 0.6589, "num_input_tokens_seen": 2047136, "step": 1515 }, { "epoch": 0.04865245502848729, "grad_norm": 2.0352001190185547, "learning_rate": 4.9822298884414626e-05, "loss": 0.5296, "num_input_tokens_seen": 2053520, "step": 1520 }, { "epoch": 0.04881249599897574, "grad_norm": 1.8432916402816772, "learning_rate": 4.982112850243288e-05, "loss": 0.7072, "num_input_tokens_seen": 2060896, "step": 1525 }, { "epoch": 0.04897253696946418, "grad_norm": 2.2160940170288086, "learning_rate": 4.98199542927212e-05, "loss": 0.9932, "num_input_tokens_seen": 2067136, "step": 1530 }, { "epoch": 0.049132577939952626, "grad_norm": 1.570143222808838, "learning_rate": 4.981877625546066e-05, "loss": 0.6402, "num_input_tokens_seen": 2074048, "step": 1535 }, { "epoch": 0.04929261891044107, "grad_norm": 2.528520107269287, "learning_rate": 4.981759439083293e-05, "loss": 0.9943, "num_input_tokens_seen": 2080416, "step": 1540 }, { "epoch": 0.04945265988092952, "grad_norm": 2.252286911010742, "learning_rate": 4.981640869902027e-05, "loss": 0.5971, "num_input_tokens_seen": 2087072, "step": 1545 }, { "epoch": 0.04961270085141796, "grad_norm": 1.7440580129623413, "learning_rate": 4.9815219180205517e-05, "loss": 0.51, "num_input_tokens_seen": 2093872, "step": 1550 }, { "epoch": 0.049772741821906406, "grad_norm": 2.132216215133667, "learning_rate": 4.9814025834572126e-05, "loss": 0.6232, "num_input_tokens_seen": 2100784, "step": 1555 }, { "epoch": 0.04993278279239485, "grad_norm": 1.5267823934555054, "learning_rate": 4.981282866230411e-05, "loss": 0.9391, "num_input_tokens_seen": 2107488, "step": 1560 }, { "epoch": 0.0500928237628833, "grad_norm": 2.1715755462646484, "learning_rate": 4.981162766358611e-05, "loss": 0.786, "num_input_tokens_seen": 2114224, "step": 1565 }, { "epoch": 0.05025286473337175, "grad_norm": 1.6525623798370361, "learning_rate": 4.9810422838603316e-05, "loss": 0.6222, "num_input_tokens_seen": 2120880, "step": 1570 }, { "epoch": 0.050412905703860186, "grad_norm": 1.9564493894577026, "learning_rate": 4.9809214187541533e-05, "loss": 0.5284, "num_input_tokens_seen": 2128112, "step": 1575 }, { "epoch": 0.05057294667434863, "grad_norm": 1.6428784132003784, "learning_rate": 4.980800171058715e-05, "loss": 0.688, "num_input_tokens_seen": 2134928, "step": 1580 }, { "epoch": 0.05073298764483708, "grad_norm": 3.21675181388855, "learning_rate": 4.980678540792715e-05, "loss": 0.7891, "num_input_tokens_seen": 2141808, "step": 1585 }, { "epoch": 0.050893028615325527, "grad_norm": 2.9934847354888916, "learning_rate": 4.980556527974909e-05, "loss": 0.7207, "num_input_tokens_seen": 2149376, "step": 1590 }, { "epoch": 0.051053069585813966, "grad_norm": 2.091613292694092, "learning_rate": 4.980434132624114e-05, "loss": 0.5459, "num_input_tokens_seen": 2156352, "step": 1595 }, { "epoch": 0.05121311055630241, "grad_norm": 0.871289849281311, "learning_rate": 4.980311354759205e-05, "loss": 0.6776, "num_input_tokens_seen": 2163024, "step": 1600 }, { "epoch": 0.05121311055630241, "eval_loss": 0.6931127905845642, "eval_runtime": 499.2961, "eval_samples_per_second": 27.811, "eval_steps_per_second": 13.906, "num_input_tokens_seen": 2163024, "step": 1600 }, { "epoch": 0.05137315152679086, "grad_norm": 2.2532589435577393, "learning_rate": 4.980188194399116e-05, "loss": 0.7528, "num_input_tokens_seen": 2169744, "step": 1605 }, { "epoch": 0.051533192497279307, "grad_norm": 1.9042249917984009, "learning_rate": 4.9800646515628384e-05, "loss": 0.7342, "num_input_tokens_seen": 2176896, "step": 1610 }, { "epoch": 0.051693233467767746, "grad_norm": 1.0084905624389648, "learning_rate": 4.979940726269426e-05, "loss": 0.6064, "num_input_tokens_seen": 2183696, "step": 1615 }, { "epoch": 0.05185327443825619, "grad_norm": 3.6197292804718018, "learning_rate": 4.979816418537988e-05, "loss": 0.8934, "num_input_tokens_seen": 2190400, "step": 1620 }, { "epoch": 0.05201331540874464, "grad_norm": 2.5878512859344482, "learning_rate": 4.979691728387696e-05, "loss": 0.6714, "num_input_tokens_seen": 2197088, "step": 1625 }, { "epoch": 0.052173356379233086, "grad_norm": 0.8807833790779114, "learning_rate": 4.979566655837776e-05, "loss": 0.5227, "num_input_tokens_seen": 2204064, "step": 1630 }, { "epoch": 0.052333397349721526, "grad_norm": 2.3137073516845703, "learning_rate": 4.9794412009075184e-05, "loss": 0.6368, "num_input_tokens_seen": 2210704, "step": 1635 }, { "epoch": 0.05249343832020997, "grad_norm": 1.538280963897705, "learning_rate": 4.979315363616269e-05, "loss": 0.5619, "num_input_tokens_seen": 2217536, "step": 1640 }, { "epoch": 0.05265347929069842, "grad_norm": 1.8922220468521118, "learning_rate": 4.979189143983434e-05, "loss": 0.5674, "num_input_tokens_seen": 2224080, "step": 1645 }, { "epoch": 0.052813520261186866, "grad_norm": 1.9439672231674194, "learning_rate": 4.979062542028478e-05, "loss": 0.688, "num_input_tokens_seen": 2230848, "step": 1650 }, { "epoch": 0.052973561231675306, "grad_norm": 0.921069860458374, "learning_rate": 4.978935557770923e-05, "loss": 0.7633, "num_input_tokens_seen": 2237888, "step": 1655 }, { "epoch": 0.05313360220216375, "grad_norm": 1.198657512664795, "learning_rate": 4.978808191230353e-05, "loss": 0.7057, "num_input_tokens_seen": 2244640, "step": 1660 }, { "epoch": 0.0532936431726522, "grad_norm": 1.4965147972106934, "learning_rate": 4.9786804424264085e-05, "loss": 0.7001, "num_input_tokens_seen": 2251760, "step": 1665 }, { "epoch": 0.053453684143140646, "grad_norm": 0.9641931056976318, "learning_rate": 4.978552311378792e-05, "loss": 0.6243, "num_input_tokens_seen": 2258208, "step": 1670 }, { "epoch": 0.053613725113629086, "grad_norm": 2.9343056678771973, "learning_rate": 4.978423798107261e-05, "loss": 0.6539, "num_input_tokens_seen": 2265088, "step": 1675 }, { "epoch": 0.05377376608411753, "grad_norm": 1.5263572931289673, "learning_rate": 4.978294902631635e-05, "loss": 0.5498, "num_input_tokens_seen": 2271888, "step": 1680 }, { "epoch": 0.05393380705460598, "grad_norm": 3.0947532653808594, "learning_rate": 4.9781656249717914e-05, "loss": 0.9118, "num_input_tokens_seen": 2278976, "step": 1685 }, { "epoch": 0.054093848025094426, "grad_norm": 1.6733607053756714, "learning_rate": 4.9780359651476645e-05, "loss": 0.4398, "num_input_tokens_seen": 2285520, "step": 1690 }, { "epoch": 0.054253888995582866, "grad_norm": 2.051729202270508, "learning_rate": 4.977905923179251e-05, "loss": 0.6719, "num_input_tokens_seen": 2292352, "step": 1695 }, { "epoch": 0.05441392996607131, "grad_norm": 0.835533857345581, "learning_rate": 4.977775499086606e-05, "loss": 0.906, "num_input_tokens_seen": 2299696, "step": 1700 }, { "epoch": 0.05457397093655976, "grad_norm": 1.9358129501342773, "learning_rate": 4.97764469288984e-05, "loss": 0.5944, "num_input_tokens_seen": 2306496, "step": 1705 }, { "epoch": 0.054734011907048206, "grad_norm": 2.2589449882507324, "learning_rate": 4.977513504609127e-05, "loss": 0.8304, "num_input_tokens_seen": 2313600, "step": 1710 }, { "epoch": 0.054894052877536646, "grad_norm": 1.4483399391174316, "learning_rate": 4.9773819342646965e-05, "loss": 0.7034, "num_input_tokens_seen": 2320480, "step": 1715 }, { "epoch": 0.05505409384802509, "grad_norm": 1.8005095720291138, "learning_rate": 4.97724998187684e-05, "loss": 0.6514, "num_input_tokens_seen": 2327904, "step": 1720 }, { "epoch": 0.05521413481851354, "grad_norm": 1.3457320928573608, "learning_rate": 4.9771176474659045e-05, "loss": 0.6289, "num_input_tokens_seen": 2334400, "step": 1725 }, { "epoch": 0.055374175789001986, "grad_norm": 1.1806566715240479, "learning_rate": 4.976984931052299e-05, "loss": 0.5103, "num_input_tokens_seen": 2341264, "step": 1730 }, { "epoch": 0.055534216759490426, "grad_norm": 1.0168970823287964, "learning_rate": 4.976851832656489e-05, "loss": 0.4621, "num_input_tokens_seen": 2348080, "step": 1735 }, { "epoch": 0.05569425772997887, "grad_norm": 1.140221118927002, "learning_rate": 4.9767183522990004e-05, "loss": 0.5396, "num_input_tokens_seen": 2354800, "step": 1740 }, { "epoch": 0.05585429870046732, "grad_norm": 1.4006198644638062, "learning_rate": 4.9765844900004176e-05, "loss": 0.8926, "num_input_tokens_seen": 2361968, "step": 1745 }, { "epoch": 0.056014339670955766, "grad_norm": 2.6563689708709717, "learning_rate": 4.9764502457813834e-05, "loss": 0.8425, "num_input_tokens_seen": 2368448, "step": 1750 }, { "epoch": 0.05617438064144421, "grad_norm": 2.6535303592681885, "learning_rate": 4.9763156196626005e-05, "loss": 0.9433, "num_input_tokens_seen": 2375248, "step": 1755 }, { "epoch": 0.05633442161193265, "grad_norm": 1.6705288887023926, "learning_rate": 4.97618061166483e-05, "loss": 0.6232, "num_input_tokens_seen": 2381920, "step": 1760 }, { "epoch": 0.0564944625824211, "grad_norm": 1.5895678997039795, "learning_rate": 4.9760452218088915e-05, "loss": 0.6712, "num_input_tokens_seen": 2388320, "step": 1765 }, { "epoch": 0.056654503552909546, "grad_norm": 2.8711256980895996, "learning_rate": 4.975909450115663e-05, "loss": 0.6878, "num_input_tokens_seen": 2395200, "step": 1770 }, { "epoch": 0.05681454452339799, "grad_norm": 2.0239272117614746, "learning_rate": 4.975773296606084e-05, "loss": 0.697, "num_input_tokens_seen": 2402032, "step": 1775 }, { "epoch": 0.05697458549388643, "grad_norm": 1.2835094928741455, "learning_rate": 4.97563676130115e-05, "loss": 0.7064, "num_input_tokens_seen": 2409296, "step": 1780 }, { "epoch": 0.05713462646437488, "grad_norm": 1.3554253578186035, "learning_rate": 4.9754998442219166e-05, "loss": 0.6886, "num_input_tokens_seen": 2416096, "step": 1785 }, { "epoch": 0.057294667434863326, "grad_norm": 1.4881749153137207, "learning_rate": 4.9753625453894984e-05, "loss": 0.5472, "num_input_tokens_seen": 2422992, "step": 1790 }, { "epoch": 0.05745470840535177, "grad_norm": 1.4302127361297607, "learning_rate": 4.975224864825068e-05, "loss": 0.8563, "num_input_tokens_seen": 2429280, "step": 1795 }, { "epoch": 0.05761474937584021, "grad_norm": 1.8669564723968506, "learning_rate": 4.9750868025498576e-05, "loss": 0.5712, "num_input_tokens_seen": 2436032, "step": 1800 }, { "epoch": 0.05761474937584021, "eval_loss": 0.6745192408561707, "eval_runtime": 499.3629, "eval_samples_per_second": 27.807, "eval_steps_per_second": 13.904, "num_input_tokens_seen": 2436032, "step": 1800 }, { "epoch": 0.05777479034632866, "grad_norm": 1.091322898864746, "learning_rate": 4.974948358585158e-05, "loss": 0.6881, "num_input_tokens_seen": 2442544, "step": 1805 }, { "epoch": 0.057934831316817106, "grad_norm": 1.2532943487167358, "learning_rate": 4.9748095329523205e-05, "loss": 0.6683, "num_input_tokens_seen": 2449312, "step": 1810 }, { "epoch": 0.05809487228730555, "grad_norm": 1.151248574256897, "learning_rate": 4.974670325672752e-05, "loss": 0.4545, "num_input_tokens_seen": 2456208, "step": 1815 }, { "epoch": 0.05825491325779399, "grad_norm": 1.506740689277649, "learning_rate": 4.974530736767921e-05, "loss": 0.4773, "num_input_tokens_seen": 2462896, "step": 1820 }, { "epoch": 0.05841495422828244, "grad_norm": 1.5447936058044434, "learning_rate": 4.9743907662593524e-05, "loss": 0.7948, "num_input_tokens_seen": 2469296, "step": 1825 }, { "epoch": 0.058574995198770886, "grad_norm": 2.0630154609680176, "learning_rate": 4.974250414168633e-05, "loss": 0.5989, "num_input_tokens_seen": 2476064, "step": 1830 }, { "epoch": 0.05873503616925933, "grad_norm": 1.338614583015442, "learning_rate": 4.974109680517407e-05, "loss": 0.6307, "num_input_tokens_seen": 2482832, "step": 1835 }, { "epoch": 0.05889507713974777, "grad_norm": 4.365056991577148, "learning_rate": 4.973968565327376e-05, "loss": 0.7501, "num_input_tokens_seen": 2489472, "step": 1840 }, { "epoch": 0.05905511811023622, "grad_norm": 3.3107845783233643, "learning_rate": 4.973827068620303e-05, "loss": 0.6158, "num_input_tokens_seen": 2496528, "step": 1845 }, { "epoch": 0.059215159080724666, "grad_norm": 1.1562244892120361, "learning_rate": 4.973685190418008e-05, "loss": 0.5355, "num_input_tokens_seen": 2503184, "step": 1850 }, { "epoch": 0.05937520005121311, "grad_norm": 3.193216323852539, "learning_rate": 4.97354293074237e-05, "loss": 0.6788, "num_input_tokens_seen": 2509504, "step": 1855 }, { "epoch": 0.05953524102170155, "grad_norm": 1.328066110610962, "learning_rate": 4.9734002896153276e-05, "loss": 0.593, "num_input_tokens_seen": 2516304, "step": 1860 }, { "epoch": 0.05969528199219, "grad_norm": 1.4483569860458374, "learning_rate": 4.973257267058877e-05, "loss": 0.6365, "num_input_tokens_seen": 2522736, "step": 1865 }, { "epoch": 0.059855322962678446, "grad_norm": 1.228033185005188, "learning_rate": 4.973113863095076e-05, "loss": 0.5365, "num_input_tokens_seen": 2530128, "step": 1870 }, { "epoch": 0.06001536393316689, "grad_norm": 3.7626900672912598, "learning_rate": 4.9729700777460384e-05, "loss": 0.7036, "num_input_tokens_seen": 2536752, "step": 1875 }, { "epoch": 0.06017540490365533, "grad_norm": 1.3567233085632324, "learning_rate": 4.972825911033937e-05, "loss": 0.5232, "num_input_tokens_seen": 2543920, "step": 1880 }, { "epoch": 0.06033544587414378, "grad_norm": 0.6721556782722473, "learning_rate": 4.9726813629810056e-05, "loss": 0.6354, "num_input_tokens_seen": 2550528, "step": 1885 }, { "epoch": 0.060495486844632226, "grad_norm": 2.088487148284912, "learning_rate": 4.9725364336095326e-05, "loss": 0.8389, "num_input_tokens_seen": 2557184, "step": 1890 }, { "epoch": 0.06065552781512067, "grad_norm": 1.6961166858673096, "learning_rate": 4.972391122941871e-05, "loss": 0.4486, "num_input_tokens_seen": 2563536, "step": 1895 }, { "epoch": 0.06081556878560911, "grad_norm": 2.0360026359558105, "learning_rate": 4.972245431000428e-05, "loss": 0.8587, "num_input_tokens_seen": 2570656, "step": 1900 }, { "epoch": 0.06097560975609756, "grad_norm": 1.3231332302093506, "learning_rate": 4.972099357807671e-05, "loss": 0.495, "num_input_tokens_seen": 2577248, "step": 1905 }, { "epoch": 0.061135650726586006, "grad_norm": 1.2886990308761597, "learning_rate": 4.971952903386127e-05, "loss": 0.4624, "num_input_tokens_seen": 2583904, "step": 1910 }, { "epoch": 0.06129569169707445, "grad_norm": 1.259263277053833, "learning_rate": 4.971806067758381e-05, "loss": 0.6078, "num_input_tokens_seen": 2590848, "step": 1915 }, { "epoch": 0.06145573266756289, "grad_norm": 1.7064348459243774, "learning_rate": 4.971658850947076e-05, "loss": 0.7977, "num_input_tokens_seen": 2597392, "step": 1920 }, { "epoch": 0.06161577363805134, "grad_norm": 1.8530136346817017, "learning_rate": 4.9715112529749165e-05, "loss": 0.7385, "num_input_tokens_seen": 2603712, "step": 1925 }, { "epoch": 0.061775814608539786, "grad_norm": 1.7915751934051514, "learning_rate": 4.9713632738646624e-05, "loss": 0.9181, "num_input_tokens_seen": 2610528, "step": 1930 }, { "epoch": 0.06193585557902823, "grad_norm": 1.110027551651001, "learning_rate": 4.971214913639134e-05, "loss": 0.7158, "num_input_tokens_seen": 2617456, "step": 1935 }, { "epoch": 0.06209589654951667, "grad_norm": 0.930117666721344, "learning_rate": 4.9710661723212104e-05, "loss": 0.5835, "num_input_tokens_seen": 2624384, "step": 1940 }, { "epoch": 0.06225593752000512, "grad_norm": 2.112504005432129, "learning_rate": 4.9709170499338295e-05, "loss": 0.75, "num_input_tokens_seen": 2630992, "step": 1945 }, { "epoch": 0.062415978490493566, "grad_norm": 0.965133547782898, "learning_rate": 4.9707675464999895e-05, "loss": 0.4221, "num_input_tokens_seen": 2637808, "step": 1950 }, { "epoch": 0.06257601946098201, "grad_norm": 1.5320302248001099, "learning_rate": 4.970617662042743e-05, "loss": 0.4666, "num_input_tokens_seen": 2644688, "step": 1955 }, { "epoch": 0.06273606043147045, "grad_norm": 0.8927896618843079, "learning_rate": 4.970467396585206e-05, "loss": 0.6125, "num_input_tokens_seen": 2651328, "step": 1960 }, { "epoch": 0.0628961014019589, "grad_norm": 1.1503963470458984, "learning_rate": 4.97031675015055e-05, "loss": 0.6697, "num_input_tokens_seen": 2657760, "step": 1965 }, { "epoch": 0.06305614237244735, "grad_norm": 1.1730012893676758, "learning_rate": 4.9701657227620075e-05, "loss": 0.418, "num_input_tokens_seen": 2666016, "step": 1970 }, { "epoch": 0.06321618334293579, "grad_norm": 1.1966904401779175, "learning_rate": 4.9700143144428685e-05, "loss": 0.8318, "num_input_tokens_seen": 2673008, "step": 1975 }, { "epoch": 0.06337622431342424, "grad_norm": 0.6530946493148804, "learning_rate": 4.969862525216482e-05, "loss": 0.5845, "num_input_tokens_seen": 2679808, "step": 1980 }, { "epoch": 0.06353626528391268, "grad_norm": 1.2357233762741089, "learning_rate": 4.9697103551062556e-05, "loss": 0.4321, "num_input_tokens_seen": 2686640, "step": 1985 }, { "epoch": 0.06369630625440113, "grad_norm": 2.3963091373443604, "learning_rate": 4.9695578041356565e-05, "loss": 0.6138, "num_input_tokens_seen": 2693696, "step": 1990 }, { "epoch": 0.06385634722488957, "grad_norm": 1.1640299558639526, "learning_rate": 4.969404872328209e-05, "loss": 0.6721, "num_input_tokens_seen": 2700240, "step": 1995 }, { "epoch": 0.06401638819537801, "grad_norm": 2.592754364013672, "learning_rate": 4.969251559707498e-05, "loss": 0.6222, "num_input_tokens_seen": 2706960, "step": 2000 }, { "epoch": 0.06401638819537801, "eval_loss": 0.6617449522018433, "eval_runtime": 499.2928, "eval_samples_per_second": 27.811, "eval_steps_per_second": 13.906, "num_input_tokens_seen": 2706960, "step": 2000 }, { "epoch": 0.06417642916586647, "grad_norm": 1.374665379524231, "learning_rate": 4.9690978662971674e-05, "loss": 0.6842, "num_input_tokens_seen": 2714000, "step": 2005 }, { "epoch": 0.0643364701363549, "grad_norm": 1.9962787628173828, "learning_rate": 4.968943792120916e-05, "loss": 0.8885, "num_input_tokens_seen": 2720784, "step": 2010 }, { "epoch": 0.06449651110684335, "grad_norm": 0.5948417782783508, "learning_rate": 4.9687893372025046e-05, "loss": 0.6187, "num_input_tokens_seen": 2727136, "step": 2015 }, { "epoch": 0.0646565520773318, "grad_norm": 1.8634579181671143, "learning_rate": 4.9686345015657535e-05, "loss": 0.6843, "num_input_tokens_seen": 2733776, "step": 2020 }, { "epoch": 0.06481659304782024, "grad_norm": 2.119891881942749, "learning_rate": 4.968479285234538e-05, "loss": 0.5493, "num_input_tokens_seen": 2740512, "step": 2025 }, { "epoch": 0.0649766340183087, "grad_norm": 1.1546019315719604, "learning_rate": 4.9683236882327974e-05, "loss": 0.5436, "num_input_tokens_seen": 2747200, "step": 2030 }, { "epoch": 0.06513667498879713, "grad_norm": 1.9396260976791382, "learning_rate": 4.968167710584526e-05, "loss": 0.5914, "num_input_tokens_seen": 2753824, "step": 2035 }, { "epoch": 0.06529671595928557, "grad_norm": 1.895516276359558, "learning_rate": 4.968011352313775e-05, "loss": 0.8842, "num_input_tokens_seen": 2760384, "step": 2040 }, { "epoch": 0.06545675692977403, "grad_norm": 0.8529548645019531, "learning_rate": 4.967854613444659e-05, "loss": 0.4609, "num_input_tokens_seen": 2767312, "step": 2045 }, { "epoch": 0.06561679790026247, "grad_norm": 1.5243620872497559, "learning_rate": 4.967697494001349e-05, "loss": 0.5065, "num_input_tokens_seen": 2774544, "step": 2050 }, { "epoch": 0.0657768388707509, "grad_norm": 1.2484145164489746, "learning_rate": 4.9675399940080736e-05, "loss": 0.703, "num_input_tokens_seen": 2781280, "step": 2055 }, { "epoch": 0.06593687984123936, "grad_norm": 1.8233182430267334, "learning_rate": 4.9673821134891226e-05, "loss": 0.6007, "num_input_tokens_seen": 2788416, "step": 2060 }, { "epoch": 0.0660969208117278, "grad_norm": 0.7996780276298523, "learning_rate": 4.967223852468842e-05, "loss": 0.7593, "num_input_tokens_seen": 2794752, "step": 2065 }, { "epoch": 0.06625696178221625, "grad_norm": 1.0501976013183594, "learning_rate": 4.967065210971639e-05, "loss": 0.4795, "num_input_tokens_seen": 2801328, "step": 2070 }, { "epoch": 0.06641700275270469, "grad_norm": 1.3323936462402344, "learning_rate": 4.966906189021977e-05, "loss": 0.4547, "num_input_tokens_seen": 2808224, "step": 2075 }, { "epoch": 0.06657704372319313, "grad_norm": 1.1079918146133423, "learning_rate": 4.966746786644379e-05, "loss": 0.7037, "num_input_tokens_seen": 2815184, "step": 2080 }, { "epoch": 0.06673708469368159, "grad_norm": 1.8151075839996338, "learning_rate": 4.966587003863429e-05, "loss": 0.4159, "num_input_tokens_seen": 2822032, "step": 2085 }, { "epoch": 0.06689712566417003, "grad_norm": 2.469982862472534, "learning_rate": 4.966426840703765e-05, "loss": 0.7159, "num_input_tokens_seen": 2829248, "step": 2090 }, { "epoch": 0.06705716663465847, "grad_norm": 1.3917090892791748, "learning_rate": 4.9662662971900875e-05, "loss": 0.8349, "num_input_tokens_seen": 2835568, "step": 2095 }, { "epoch": 0.06721720760514692, "grad_norm": 1.9768903255462646, "learning_rate": 4.9661053733471534e-05, "loss": 0.6037, "num_input_tokens_seen": 2842224, "step": 2100 }, { "epoch": 0.06737724857563536, "grad_norm": 2.4103965759277344, "learning_rate": 4.965944069199781e-05, "loss": 0.7625, "num_input_tokens_seen": 2848880, "step": 2105 }, { "epoch": 0.06753728954612381, "grad_norm": 1.7424540519714355, "learning_rate": 4.965782384772842e-05, "loss": 0.5874, "num_input_tokens_seen": 2855664, "step": 2110 }, { "epoch": 0.06769733051661225, "grad_norm": 1.468516230583191, "learning_rate": 4.9656203200912734e-05, "loss": 0.7707, "num_input_tokens_seen": 2862720, "step": 2115 }, { "epoch": 0.06785737148710069, "grad_norm": 1.1950609683990479, "learning_rate": 4.965457875180067e-05, "loss": 0.6026, "num_input_tokens_seen": 2869376, "step": 2120 }, { "epoch": 0.06801741245758915, "grad_norm": 2.28385591506958, "learning_rate": 4.9652950500642724e-05, "loss": 0.6625, "num_input_tokens_seen": 2876272, "step": 2125 }, { "epoch": 0.06817745342807759, "grad_norm": 1.4865267276763916, "learning_rate": 4.965131844769001e-05, "loss": 0.6258, "num_input_tokens_seen": 2882848, "step": 2130 }, { "epoch": 0.06833749439856603, "grad_norm": 1.8171504735946655, "learning_rate": 4.96496825931942e-05, "loss": 0.604, "num_input_tokens_seen": 2889856, "step": 2135 }, { "epoch": 0.06849753536905448, "grad_norm": 1.1843619346618652, "learning_rate": 4.9648042937407566e-05, "loss": 0.4754, "num_input_tokens_seen": 2896768, "step": 2140 }, { "epoch": 0.06865757633954292, "grad_norm": 0.9499813914299011, "learning_rate": 4.964639948058297e-05, "loss": 0.5944, "num_input_tokens_seen": 2903344, "step": 2145 }, { "epoch": 0.06881761731003137, "grad_norm": 1.2397598028182983, "learning_rate": 4.9644752222973846e-05, "loss": 0.6487, "num_input_tokens_seen": 2910080, "step": 2150 }, { "epoch": 0.06897765828051981, "grad_norm": 2.0509896278381348, "learning_rate": 4.964310116483422e-05, "loss": 0.8045, "num_input_tokens_seen": 2916688, "step": 2155 }, { "epoch": 0.06913769925100825, "grad_norm": 1.5339853763580322, "learning_rate": 4.964144630641872e-05, "loss": 0.4716, "num_input_tokens_seen": 2923168, "step": 2160 }, { "epoch": 0.0692977402214967, "grad_norm": 1.093266248703003, "learning_rate": 4.9639787647982525e-05, "loss": 0.6708, "num_input_tokens_seen": 2930032, "step": 2165 }, { "epoch": 0.06945778119198515, "grad_norm": 1.2658212184906006, "learning_rate": 4.963812518978143e-05, "loss": 0.6151, "num_input_tokens_seen": 2936640, "step": 2170 }, { "epoch": 0.0696178221624736, "grad_norm": 0.7749289870262146, "learning_rate": 4.963645893207182e-05, "loss": 0.65, "num_input_tokens_seen": 2943536, "step": 2175 }, { "epoch": 0.06977786313296204, "grad_norm": 0.8754503130912781, "learning_rate": 4.963478887511063e-05, "loss": 0.9442, "num_input_tokens_seen": 2950016, "step": 2180 }, { "epoch": 0.06993790410345048, "grad_norm": 2.1379129886627197, "learning_rate": 4.963311501915542e-05, "loss": 0.6401, "num_input_tokens_seen": 2956528, "step": 2185 }, { "epoch": 0.07009794507393893, "grad_norm": 3.12528920173645, "learning_rate": 4.963143736446432e-05, "loss": 0.7149, "num_input_tokens_seen": 2962848, "step": 2190 }, { "epoch": 0.07025798604442737, "grad_norm": 1.7292726039886475, "learning_rate": 4.962975591129603e-05, "loss": 0.5431, "num_input_tokens_seen": 2969248, "step": 2195 }, { "epoch": 0.07041802701491581, "grad_norm": 2.1271955966949463, "learning_rate": 4.962807065990986e-05, "loss": 1.1155, "num_input_tokens_seen": 2976144, "step": 2200 }, { "epoch": 0.07041802701491581, "eval_loss": 0.6501045227050781, "eval_runtime": 499.554, "eval_samples_per_second": 27.797, "eval_steps_per_second": 13.898, "num_input_tokens_seen": 2976144, "step": 2200 }, { "epoch": 0.07057806798540427, "grad_norm": 1.6237902641296387, "learning_rate": 4.9626381610565714e-05, "loss": 0.5489, "num_input_tokens_seen": 2982992, "step": 2205 }, { "epoch": 0.0707381089558927, "grad_norm": 2.0779547691345215, "learning_rate": 4.9624688763524043e-05, "loss": 0.4984, "num_input_tokens_seen": 2989568, "step": 2210 }, { "epoch": 0.07089814992638116, "grad_norm": 0.6326608061790466, "learning_rate": 4.962299211904591e-05, "loss": 0.6041, "num_input_tokens_seen": 2996336, "step": 2215 }, { "epoch": 0.0710581908968696, "grad_norm": 3.3485047817230225, "learning_rate": 4.962129167739296e-05, "loss": 0.7132, "num_input_tokens_seen": 3003232, "step": 2220 }, { "epoch": 0.07121823186735804, "grad_norm": 1.7626702785491943, "learning_rate": 4.961958743882742e-05, "loss": 0.5412, "num_input_tokens_seen": 3010144, "step": 2225 }, { "epoch": 0.0713782728378465, "grad_norm": 0.6314999461174011, "learning_rate": 4.961787940361211e-05, "loss": 0.474, "num_input_tokens_seen": 3017264, "step": 2230 }, { "epoch": 0.07153831380833493, "grad_norm": 1.4913572072982788, "learning_rate": 4.961616757201043e-05, "loss": 0.67, "num_input_tokens_seen": 3023968, "step": 2235 }, { "epoch": 0.07169835477882337, "grad_norm": 2.5230863094329834, "learning_rate": 4.961445194428637e-05, "loss": 0.5908, "num_input_tokens_seen": 3030480, "step": 2240 }, { "epoch": 0.07185839574931183, "grad_norm": 1.7315547466278076, "learning_rate": 4.9612732520704486e-05, "loss": 0.5452, "num_input_tokens_seen": 3037520, "step": 2245 }, { "epoch": 0.07201843671980027, "grad_norm": 1.2889426946640015, "learning_rate": 4.961100930152994e-05, "loss": 0.6719, "num_input_tokens_seen": 3043808, "step": 2250 }, { "epoch": 0.07217847769028872, "grad_norm": 1.7510902881622314, "learning_rate": 4.960928228702849e-05, "loss": 0.7738, "num_input_tokens_seen": 3050544, "step": 2255 }, { "epoch": 0.07233851866077716, "grad_norm": 0.931219756603241, "learning_rate": 4.960755147746645e-05, "loss": 0.5525, "num_input_tokens_seen": 3057408, "step": 2260 }, { "epoch": 0.0724985596312656, "grad_norm": 1.269982933998108, "learning_rate": 4.9605816873110736e-05, "loss": 0.5133, "num_input_tokens_seen": 3063792, "step": 2265 }, { "epoch": 0.07265860060175405, "grad_norm": 2.2028913497924805, "learning_rate": 4.960407847422883e-05, "loss": 0.4982, "num_input_tokens_seen": 3070672, "step": 2270 }, { "epoch": 0.07281864157224249, "grad_norm": 1.6192666292190552, "learning_rate": 4.960233628108885e-05, "loss": 0.778, "num_input_tokens_seen": 3077200, "step": 2275 }, { "epoch": 0.07297868254273093, "grad_norm": 1.6341867446899414, "learning_rate": 4.960059029395942e-05, "loss": 0.7416, "num_input_tokens_seen": 3084784, "step": 2280 }, { "epoch": 0.07313872351321939, "grad_norm": 1.5017443895339966, "learning_rate": 4.959884051310983e-05, "loss": 0.8073, "num_input_tokens_seen": 3091712, "step": 2285 }, { "epoch": 0.07329876448370783, "grad_norm": 1.7320019006729126, "learning_rate": 4.959708693880991e-05, "loss": 0.7511, "num_input_tokens_seen": 3098512, "step": 2290 }, { "epoch": 0.07345880545419628, "grad_norm": 1.4258651733398438, "learning_rate": 4.9595329571330074e-05, "loss": 0.8224, "num_input_tokens_seen": 3105648, "step": 2295 }, { "epoch": 0.07361884642468472, "grad_norm": 1.334587574005127, "learning_rate": 4.9593568410941326e-05, "loss": 0.6375, "num_input_tokens_seen": 3112608, "step": 2300 }, { "epoch": 0.07377888739517316, "grad_norm": 2.194685697555542, "learning_rate": 4.959180345791528e-05, "loss": 0.6769, "num_input_tokens_seen": 3119088, "step": 2305 }, { "epoch": 0.07393892836566161, "grad_norm": 1.1999809741973877, "learning_rate": 4.9590034712524086e-05, "loss": 0.6504, "num_input_tokens_seen": 3126336, "step": 2310 }, { "epoch": 0.07409896933615005, "grad_norm": 1.5584723949432373, "learning_rate": 4.958826217504053e-05, "loss": 0.4108, "num_input_tokens_seen": 3133200, "step": 2315 }, { "epoch": 0.07425901030663849, "grad_norm": 0.6625922918319702, "learning_rate": 4.958648584573795e-05, "loss": 0.4439, "num_input_tokens_seen": 3140096, "step": 2320 }, { "epoch": 0.07441905127712695, "grad_norm": 2.1022861003875732, "learning_rate": 4.958470572489028e-05, "loss": 0.6881, "num_input_tokens_seen": 3146880, "step": 2325 }, { "epoch": 0.07457909224761539, "grad_norm": 1.0459328889846802, "learning_rate": 4.958292181277203e-05, "loss": 0.6454, "num_input_tokens_seen": 3153808, "step": 2330 }, { "epoch": 0.07473913321810384, "grad_norm": 0.9681134819984436, "learning_rate": 4.958113410965832e-05, "loss": 0.574, "num_input_tokens_seen": 3160288, "step": 2335 }, { "epoch": 0.07489917418859228, "grad_norm": 1.5802910327911377, "learning_rate": 4.957934261582481e-05, "loss": 0.6172, "num_input_tokens_seen": 3166624, "step": 2340 }, { "epoch": 0.07505921515908072, "grad_norm": 2.09704852104187, "learning_rate": 4.95775473315478e-05, "loss": 0.7953, "num_input_tokens_seen": 3173232, "step": 2345 }, { "epoch": 0.07521925612956917, "grad_norm": 2.0834052562713623, "learning_rate": 4.9575748257104124e-05, "loss": 0.5897, "num_input_tokens_seen": 3180000, "step": 2350 }, { "epoch": 0.07537929710005761, "grad_norm": 1.2619205713272095, "learning_rate": 4.9573945392771224e-05, "loss": 0.7537, "num_input_tokens_seen": 3186544, "step": 2355 }, { "epoch": 0.07553933807054607, "grad_norm": 0.9917206168174744, "learning_rate": 4.9572138738827134e-05, "loss": 0.4415, "num_input_tokens_seen": 3193376, "step": 2360 }, { "epoch": 0.0756993790410345, "grad_norm": 1.9260914325714111, "learning_rate": 4.957032829555046e-05, "loss": 0.7732, "num_input_tokens_seen": 3200000, "step": 2365 }, { "epoch": 0.07585942001152295, "grad_norm": 0.9392750859260559, "learning_rate": 4.956851406322039e-05, "loss": 0.5793, "num_input_tokens_seen": 3206720, "step": 2370 }, { "epoch": 0.0760194609820114, "grad_norm": 1.4734145402908325, "learning_rate": 4.9566696042116704e-05, "loss": 0.591, "num_input_tokens_seen": 3213248, "step": 2375 }, { "epoch": 0.07617950195249984, "grad_norm": 1.9602208137512207, "learning_rate": 4.9564874232519766e-05, "loss": 0.6095, "num_input_tokens_seen": 3220704, "step": 2380 }, { "epoch": 0.07633954292298828, "grad_norm": 1.0957311391830444, "learning_rate": 4.9563048634710516e-05, "loss": 0.5802, "num_input_tokens_seen": 3228048, "step": 2385 }, { "epoch": 0.07649958389347673, "grad_norm": 1.2085758447647095, "learning_rate": 4.956121924897049e-05, "loss": 0.5301, "num_input_tokens_seen": 3234832, "step": 2390 }, { "epoch": 0.07665962486396517, "grad_norm": 1.9824390411376953, "learning_rate": 4.955938607558181e-05, "loss": 0.7403, "num_input_tokens_seen": 3241952, "step": 2395 }, { "epoch": 0.07681966583445363, "grad_norm": 3.1434226036071777, "learning_rate": 4.955754911482715e-05, "loss": 0.6136, "num_input_tokens_seen": 3248384, "step": 2400 }, { "epoch": 0.07681966583445363, "eval_loss": 0.6399428248405457, "eval_runtime": 499.228, "eval_samples_per_second": 27.815, "eval_steps_per_second": 13.907, "num_input_tokens_seen": 3248384, "step": 2400 }, { "epoch": 0.07697970680494207, "grad_norm": 1.1707093715667725, "learning_rate": 4.9555708366989804e-05, "loss": 0.4704, "num_input_tokens_seen": 3254784, "step": 2405 }, { "epoch": 0.0771397477754305, "grad_norm": 1.4581555128097534, "learning_rate": 4.9553863832353655e-05, "loss": 0.5433, "num_input_tokens_seen": 3261504, "step": 2410 }, { "epoch": 0.07729978874591896, "grad_norm": 1.3650652170181274, "learning_rate": 4.955201551120313e-05, "loss": 0.4189, "num_input_tokens_seen": 3268160, "step": 2415 }, { "epoch": 0.0774598297164074, "grad_norm": 1.6664940118789673, "learning_rate": 4.955016340382328e-05, "loss": 0.3923, "num_input_tokens_seen": 3275184, "step": 2420 }, { "epoch": 0.07761987068689584, "grad_norm": 1.777537226676941, "learning_rate": 4.954830751049972e-05, "loss": 0.4485, "num_input_tokens_seen": 3282064, "step": 2425 }, { "epoch": 0.07777991165738429, "grad_norm": 2.0689289569854736, "learning_rate": 4.954644783151864e-05, "loss": 0.6904, "num_input_tokens_seen": 3288704, "step": 2430 }, { "epoch": 0.07793995262787273, "grad_norm": 1.7038086652755737, "learning_rate": 4.954458436716684e-05, "loss": 0.5796, "num_input_tokens_seen": 3295984, "step": 2435 }, { "epoch": 0.07809999359836119, "grad_norm": 1.5774867534637451, "learning_rate": 4.954271711773168e-05, "loss": 0.5701, "num_input_tokens_seen": 3302592, "step": 2440 }, { "epoch": 0.07826003456884963, "grad_norm": 0.8975289463996887, "learning_rate": 4.9540846083501115e-05, "loss": 0.4854, "num_input_tokens_seen": 3309424, "step": 2445 }, { "epoch": 0.07842007553933807, "grad_norm": 2.3344757556915283, "learning_rate": 4.953897126476369e-05, "loss": 0.6613, "num_input_tokens_seen": 3315728, "step": 2450 }, { "epoch": 0.07858011650982652, "grad_norm": 1.7461193799972534, "learning_rate": 4.9537092661808514e-05, "loss": 0.5223, "num_input_tokens_seen": 3322448, "step": 2455 }, { "epoch": 0.07874015748031496, "grad_norm": 2.2397263050079346, "learning_rate": 4.9535210274925306e-05, "loss": 0.671, "num_input_tokens_seen": 3329456, "step": 2460 }, { "epoch": 0.0789001984508034, "grad_norm": 1.1154276132583618, "learning_rate": 4.953332410440435e-05, "loss": 0.4853, "num_input_tokens_seen": 3336368, "step": 2465 }, { "epoch": 0.07906023942129185, "grad_norm": 2.6213080883026123, "learning_rate": 4.9531434150536496e-05, "loss": 0.8673, "num_input_tokens_seen": 3343280, "step": 2470 }, { "epoch": 0.07922028039178029, "grad_norm": 2.552342176437378, "learning_rate": 4.952954041361322e-05, "loss": 0.6916, "num_input_tokens_seen": 3350112, "step": 2475 }, { "epoch": 0.07938032136226875, "grad_norm": 0.8118849396705627, "learning_rate": 4.952764289392655e-05, "loss": 0.5546, "num_input_tokens_seen": 3356528, "step": 2480 }, { "epoch": 0.07954036233275719, "grad_norm": 2.0779640674591064, "learning_rate": 4.952574159176912e-05, "loss": 0.5556, "num_input_tokens_seen": 3363024, "step": 2485 }, { "epoch": 0.07970040330324563, "grad_norm": 1.3807529211044312, "learning_rate": 4.952383650743413e-05, "loss": 0.975, "num_input_tokens_seen": 3370016, "step": 2490 }, { "epoch": 0.07986044427373408, "grad_norm": 2.175570487976074, "learning_rate": 4.952192764121536e-05, "loss": 0.5457, "num_input_tokens_seen": 3377120, "step": 2495 }, { "epoch": 0.08002048524422252, "grad_norm": 1.7279391288757324, "learning_rate": 4.9520014993407185e-05, "loss": 0.7895, "num_input_tokens_seen": 3383760, "step": 2500 }, { "epoch": 0.08018052621471096, "grad_norm": 1.347938895225525, "learning_rate": 4.951809856430456e-05, "loss": 0.517, "num_input_tokens_seen": 3390400, "step": 2505 }, { "epoch": 0.08034056718519941, "grad_norm": 1.3388969898223877, "learning_rate": 4.951617835420303e-05, "loss": 0.5439, "num_input_tokens_seen": 3397168, "step": 2510 }, { "epoch": 0.08050060815568785, "grad_norm": 1.57168447971344, "learning_rate": 4.951425436339869e-05, "loss": 0.488, "num_input_tokens_seen": 3403824, "step": 2515 }, { "epoch": 0.0806606491261763, "grad_norm": 1.9764736890792847, "learning_rate": 4.9512326592188274e-05, "loss": 0.5392, "num_input_tokens_seen": 3410288, "step": 2520 }, { "epoch": 0.08082069009666475, "grad_norm": 0.8925207853317261, "learning_rate": 4.9510395040869054e-05, "loss": 0.5752, "num_input_tokens_seen": 3416864, "step": 2525 }, { "epoch": 0.08098073106715319, "grad_norm": 1.015109658241272, "learning_rate": 4.9508459709738905e-05, "loss": 0.5456, "num_input_tokens_seen": 3423568, "step": 2530 }, { "epoch": 0.08114077203764164, "grad_norm": 3.265580892562866, "learning_rate": 4.950652059909627e-05, "loss": 0.7342, "num_input_tokens_seen": 3431104, "step": 2535 }, { "epoch": 0.08130081300813008, "grad_norm": 1.7511273622512817, "learning_rate": 4.95045777092402e-05, "loss": 0.655, "num_input_tokens_seen": 3437680, "step": 2540 }, { "epoch": 0.08146085397861853, "grad_norm": 3.641244411468506, "learning_rate": 4.950263104047031e-05, "loss": 0.5956, "num_input_tokens_seen": 3444208, "step": 2545 }, { "epoch": 0.08162089494910697, "grad_norm": 1.6874326467514038, "learning_rate": 4.9500680593086775e-05, "loss": 0.6194, "num_input_tokens_seen": 3450864, "step": 2550 }, { "epoch": 0.08178093591959541, "grad_norm": 1.1440551280975342, "learning_rate": 4.94987263673904e-05, "loss": 0.5732, "num_input_tokens_seen": 3457312, "step": 2555 }, { "epoch": 0.08194097689008387, "grad_norm": 0.6923086643218994, "learning_rate": 4.949676836368256e-05, "loss": 0.7021, "num_input_tokens_seen": 3464272, "step": 2560 }, { "epoch": 0.0821010178605723, "grad_norm": 1.2431640625, "learning_rate": 4.949480658226518e-05, "loss": 0.6252, "num_input_tokens_seen": 3470720, "step": 2565 }, { "epoch": 0.08226105883106075, "grad_norm": 1.9398747682571411, "learning_rate": 4.949284102344082e-05, "loss": 0.6505, "num_input_tokens_seen": 3477744, "step": 2570 }, { "epoch": 0.0824210998015492, "grad_norm": 1.5418306589126587, "learning_rate": 4.9490871687512565e-05, "loss": 0.7103, "num_input_tokens_seen": 3484688, "step": 2575 }, { "epoch": 0.08258114077203764, "grad_norm": 2.1064612865448, "learning_rate": 4.948889857478413e-05, "loss": 0.6465, "num_input_tokens_seen": 3492096, "step": 2580 }, { "epoch": 0.0827411817425261, "grad_norm": 1.9225594997406006, "learning_rate": 4.948692168555978e-05, "loss": 0.5402, "num_input_tokens_seen": 3498784, "step": 2585 }, { "epoch": 0.08290122271301453, "grad_norm": 1.828643560409546, "learning_rate": 4.94849410201444e-05, "loss": 0.6499, "num_input_tokens_seen": 3505648, "step": 2590 }, { "epoch": 0.08306126368350297, "grad_norm": 2.1692001819610596, "learning_rate": 4.948295657884341e-05, "loss": 0.8383, "num_input_tokens_seen": 3512272, "step": 2595 }, { "epoch": 0.08322130465399143, "grad_norm": 1.18948233127594, "learning_rate": 4.9480968361962835e-05, "loss": 0.6635, "num_input_tokens_seen": 3519088, "step": 2600 }, { "epoch": 0.08322130465399143, "eval_loss": 0.6310076117515564, "eval_runtime": 500.0402, "eval_samples_per_second": 27.77, "eval_steps_per_second": 13.885, "num_input_tokens_seen": 3519088, "step": 2600 }, { "epoch": 0.08338134562447987, "grad_norm": 1.7686326503753662, "learning_rate": 4.9478976369809305e-05, "loss": 0.5434, "num_input_tokens_seen": 3525696, "step": 2605 }, { "epoch": 0.0835413865949683, "grad_norm": 1.2000911235809326, "learning_rate": 4.947698060268999e-05, "loss": 0.7545, "num_input_tokens_seen": 3532640, "step": 2610 }, { "epoch": 0.08370142756545676, "grad_norm": 1.2088048458099365, "learning_rate": 4.9474981060912665e-05, "loss": 0.5275, "num_input_tokens_seen": 3539840, "step": 2615 }, { "epoch": 0.0838614685359452, "grad_norm": 1.0895233154296875, "learning_rate": 4.94729777447857e-05, "loss": 0.5643, "num_input_tokens_seen": 3546544, "step": 2620 }, { "epoch": 0.08402150950643365, "grad_norm": 1.6665369272232056, "learning_rate": 4.947097065461801e-05, "loss": 0.5235, "num_input_tokens_seen": 3553584, "step": 2625 }, { "epoch": 0.08418155047692209, "grad_norm": 1.97194242477417, "learning_rate": 4.9468959790719125e-05, "loss": 0.5245, "num_input_tokens_seen": 3560368, "step": 2630 }, { "epoch": 0.08434159144741053, "grad_norm": 1.8798918724060059, "learning_rate": 4.9466945153399146e-05, "loss": 0.8656, "num_input_tokens_seen": 3566768, "step": 2635 }, { "epoch": 0.08450163241789899, "grad_norm": 3.343107223510742, "learning_rate": 4.9464926742968755e-05, "loss": 0.6683, "num_input_tokens_seen": 3574112, "step": 2640 }, { "epoch": 0.08466167338838743, "grad_norm": 1.794844150543213, "learning_rate": 4.946290455973921e-05, "loss": 0.7433, "num_input_tokens_seen": 3581024, "step": 2645 }, { "epoch": 0.08482171435887587, "grad_norm": 1.077354907989502, "learning_rate": 4.9460878604022365e-05, "loss": 0.6334, "num_input_tokens_seen": 3588016, "step": 2650 }, { "epoch": 0.08498175532936432, "grad_norm": 1.6723650693893433, "learning_rate": 4.945884887613065e-05, "loss": 0.7053, "num_input_tokens_seen": 3594416, "step": 2655 }, { "epoch": 0.08514179629985276, "grad_norm": 0.8486961126327515, "learning_rate": 4.9456815376377055e-05, "loss": 0.6574, "num_input_tokens_seen": 3601072, "step": 2660 }, { "epoch": 0.08530183727034121, "grad_norm": 1.6983387470245361, "learning_rate": 4.9454778105075195e-05, "loss": 0.7645, "num_input_tokens_seen": 3607456, "step": 2665 }, { "epoch": 0.08546187824082965, "grad_norm": 1.5200512409210205, "learning_rate": 4.945273706253924e-05, "loss": 0.6736, "num_input_tokens_seen": 3614352, "step": 2670 }, { "epoch": 0.08562191921131809, "grad_norm": 1.4292455911636353, "learning_rate": 4.9450692249083925e-05, "loss": 0.3911, "num_input_tokens_seen": 3621376, "step": 2675 }, { "epoch": 0.08578196018180655, "grad_norm": 2.47149658203125, "learning_rate": 4.9448643665024605e-05, "loss": 0.6209, "num_input_tokens_seen": 3628176, "step": 2680 }, { "epoch": 0.08594200115229499, "grad_norm": 2.5674448013305664, "learning_rate": 4.944659131067719e-05, "loss": 0.429, "num_input_tokens_seen": 3634832, "step": 2685 }, { "epoch": 0.08610204212278343, "grad_norm": 0.7140019536018372, "learning_rate": 4.944453518635818e-05, "loss": 0.4569, "num_input_tokens_seen": 3641360, "step": 2690 }, { "epoch": 0.08626208309327188, "grad_norm": 1.1509854793548584, "learning_rate": 4.944247529238465e-05, "loss": 0.7817, "num_input_tokens_seen": 3648560, "step": 2695 }, { "epoch": 0.08642212406376032, "grad_norm": 2.0408506393432617, "learning_rate": 4.944041162907427e-05, "loss": 0.5234, "num_input_tokens_seen": 3655088, "step": 2700 }, { "epoch": 0.08658216503424877, "grad_norm": 1.9578735828399658, "learning_rate": 4.943834419674529e-05, "loss": 0.6752, "num_input_tokens_seen": 3661744, "step": 2705 }, { "epoch": 0.08674220600473721, "grad_norm": 1.8262184858322144, "learning_rate": 4.9436272995716506e-05, "loss": 0.4593, "num_input_tokens_seen": 3668448, "step": 2710 }, { "epoch": 0.08690224697522565, "grad_norm": 0.912365198135376, "learning_rate": 4.943419802630735e-05, "loss": 0.3987, "num_input_tokens_seen": 3674944, "step": 2715 }, { "epoch": 0.0870622879457141, "grad_norm": 0.7842931747436523, "learning_rate": 4.94321192888378e-05, "loss": 0.5999, "num_input_tokens_seen": 3682128, "step": 2720 }, { "epoch": 0.08722232891620255, "grad_norm": 1.3744995594024658, "learning_rate": 4.943003678362842e-05, "loss": 0.5094, "num_input_tokens_seen": 3689152, "step": 2725 }, { "epoch": 0.087382369886691, "grad_norm": 1.5239976644515991, "learning_rate": 4.942795051100036e-05, "loss": 0.6902, "num_input_tokens_seen": 3696144, "step": 2730 }, { "epoch": 0.08754241085717944, "grad_norm": 1.654106855392456, "learning_rate": 4.942586047127536e-05, "loss": 0.6109, "num_input_tokens_seen": 3702512, "step": 2735 }, { "epoch": 0.08770245182766788, "grad_norm": 2.1009886264801025, "learning_rate": 4.942376666477571e-05, "loss": 0.6941, "num_input_tokens_seen": 3709744, "step": 2740 }, { "epoch": 0.08786249279815633, "grad_norm": 1.5852429866790771, "learning_rate": 4.9421669091824304e-05, "loss": 0.5927, "num_input_tokens_seen": 3716176, "step": 2745 }, { "epoch": 0.08802253376864477, "grad_norm": 0.9219412207603455, "learning_rate": 4.9419567752744634e-05, "loss": 0.4155, "num_input_tokens_seen": 3723168, "step": 2750 }, { "epoch": 0.08818257473913321, "grad_norm": 0.9451608657836914, "learning_rate": 4.941746264786074e-05, "loss": 0.4928, "num_input_tokens_seen": 3729584, "step": 2755 }, { "epoch": 0.08834261570962167, "grad_norm": 2.871488094329834, "learning_rate": 4.9415353777497254e-05, "loss": 0.4671, "num_input_tokens_seen": 3736304, "step": 2760 }, { "epoch": 0.0885026566801101, "grad_norm": 1.467054843902588, "learning_rate": 4.9413241141979394e-05, "loss": 0.5255, "num_input_tokens_seen": 3743088, "step": 2765 }, { "epoch": 0.08866269765059856, "grad_norm": 1.8138264417648315, "learning_rate": 4.9411124741632956e-05, "loss": 0.5804, "num_input_tokens_seen": 3749776, "step": 2770 }, { "epoch": 0.088822738621087, "grad_norm": 0.9919851422309875, "learning_rate": 4.940900457678431e-05, "loss": 0.4527, "num_input_tokens_seen": 3756464, "step": 2775 }, { "epoch": 0.08898277959157544, "grad_norm": 1.3550885915756226, "learning_rate": 4.9406880647760425e-05, "loss": 0.4833, "num_input_tokens_seen": 3763136, "step": 2780 }, { "epoch": 0.08914282056206389, "grad_norm": 1.9220818281173706, "learning_rate": 4.9404752954888824e-05, "loss": 0.6568, "num_input_tokens_seen": 3769680, "step": 2785 }, { "epoch": 0.08930286153255233, "grad_norm": 1.0657620429992676, "learning_rate": 4.940262149849762e-05, "loss": 0.6259, "num_input_tokens_seen": 3776512, "step": 2790 }, { "epoch": 0.08946290250304077, "grad_norm": 0.8283984065055847, "learning_rate": 4.9400486278915526e-05, "loss": 0.5829, "num_input_tokens_seen": 3783424, "step": 2795 }, { "epoch": 0.08962294347352923, "grad_norm": 1.0895153284072876, "learning_rate": 4.939834729647181e-05, "loss": 0.6066, "num_input_tokens_seen": 3790208, "step": 2800 }, { "epoch": 0.08962294347352923, "eval_loss": 0.623992919921875, "eval_runtime": 499.4277, "eval_samples_per_second": 27.804, "eval_steps_per_second": 13.902, "num_input_tokens_seen": 3790208, "step": 2800 }, { "epoch": 0.08978298444401767, "grad_norm": 1.968063473701477, "learning_rate": 4.9396204551496326e-05, "loss": 0.5499, "num_input_tokens_seen": 3796864, "step": 2805 }, { "epoch": 0.08994302541450612, "grad_norm": 1.4292185306549072, "learning_rate": 4.939405804431952e-05, "loss": 0.6475, "num_input_tokens_seen": 3803632, "step": 2810 }, { "epoch": 0.09010306638499456, "grad_norm": 0.8056857585906982, "learning_rate": 4.9391907775272414e-05, "loss": 0.4767, "num_input_tokens_seen": 3810224, "step": 2815 }, { "epoch": 0.090263107355483, "grad_norm": 2.9722235202789307, "learning_rate": 4.9389753744686604e-05, "loss": 0.4354, "num_input_tokens_seen": 3817040, "step": 2820 }, { "epoch": 0.09042314832597145, "grad_norm": 0.9839230179786682, "learning_rate": 4.938759595289426e-05, "loss": 0.5273, "num_input_tokens_seen": 3823760, "step": 2825 }, { "epoch": 0.09058318929645989, "grad_norm": 1.8901482820510864, "learning_rate": 4.938543440022815e-05, "loss": 0.7782, "num_input_tokens_seen": 3830864, "step": 2830 }, { "epoch": 0.09074323026694833, "grad_norm": 1.7058407068252563, "learning_rate": 4.938326908702161e-05, "loss": 0.7286, "num_input_tokens_seen": 3837616, "step": 2835 }, { "epoch": 0.09090327123743679, "grad_norm": 2.004265785217285, "learning_rate": 4.9381100013608554e-05, "loss": 0.6459, "num_input_tokens_seen": 3844080, "step": 2840 }, { "epoch": 0.09106331220792523, "grad_norm": 1.1668764352798462, "learning_rate": 4.9378927180323485e-05, "loss": 0.4228, "num_input_tokens_seen": 3850848, "step": 2845 }, { "epoch": 0.09122335317841368, "grad_norm": 1.0055654048919678, "learning_rate": 4.937675058750148e-05, "loss": 0.4478, "num_input_tokens_seen": 3857344, "step": 2850 }, { "epoch": 0.09138339414890212, "grad_norm": 0.943962037563324, "learning_rate": 4.937457023547819e-05, "loss": 0.5014, "num_input_tokens_seen": 3864144, "step": 2855 }, { "epoch": 0.09154343511939056, "grad_norm": 1.4931411743164062, "learning_rate": 4.9372386124589876e-05, "loss": 0.5928, "num_input_tokens_seen": 3870672, "step": 2860 }, { "epoch": 0.09170347608987901, "grad_norm": 1.5373198986053467, "learning_rate": 4.937019825517333e-05, "loss": 0.5703, "num_input_tokens_seen": 3877664, "step": 2865 }, { "epoch": 0.09186351706036745, "grad_norm": 0.9837549924850464, "learning_rate": 4.9368006627565954e-05, "loss": 0.968, "num_input_tokens_seen": 3884336, "step": 2870 }, { "epoch": 0.09202355803085589, "grad_norm": 1.41847562789917, "learning_rate": 4.936581124210573e-05, "loss": 0.6505, "num_input_tokens_seen": 3891280, "step": 2875 }, { "epoch": 0.09218359900134435, "grad_norm": 1.7370367050170898, "learning_rate": 4.9363612099131216e-05, "loss": 0.4591, "num_input_tokens_seen": 3897536, "step": 2880 }, { "epoch": 0.09234363997183279, "grad_norm": 2.1180579662323, "learning_rate": 4.936140919898155e-05, "loss": 0.9306, "num_input_tokens_seen": 3904624, "step": 2885 }, { "epoch": 0.09250368094232124, "grad_norm": 1.3696763515472412, "learning_rate": 4.9359202541996426e-05, "loss": 0.5751, "num_input_tokens_seen": 3911040, "step": 2890 }, { "epoch": 0.09266372191280968, "grad_norm": 1.4400138854980469, "learning_rate": 4.935699212851616e-05, "loss": 0.5351, "num_input_tokens_seen": 3917680, "step": 2895 }, { "epoch": 0.09282376288329812, "grad_norm": 0.9346661567687988, "learning_rate": 4.935477795888162e-05, "loss": 0.4732, "num_input_tokens_seen": 3924112, "step": 2900 }, { "epoch": 0.09298380385378657, "grad_norm": 1.2357497215270996, "learning_rate": 4.935256003343426e-05, "loss": 0.948, "num_input_tokens_seen": 3930816, "step": 2905 }, { "epoch": 0.09314384482427501, "grad_norm": 1.2597625255584717, "learning_rate": 4.93503383525161e-05, "loss": 0.5312, "num_input_tokens_seen": 3937744, "step": 2910 }, { "epoch": 0.09330388579476345, "grad_norm": 1.6343860626220703, "learning_rate": 4.934811291646977e-05, "loss": 0.626, "num_input_tokens_seen": 3944272, "step": 2915 }, { "epoch": 0.0934639267652519, "grad_norm": 2.5125808715820312, "learning_rate": 4.934588372563845e-05, "loss": 0.6088, "num_input_tokens_seen": 3951024, "step": 2920 }, { "epoch": 0.09362396773574035, "grad_norm": 2.2290401458740234, "learning_rate": 4.93436507803659e-05, "loss": 0.6609, "num_input_tokens_seen": 3957456, "step": 2925 }, { "epoch": 0.0937840087062288, "grad_norm": 4.417358875274658, "learning_rate": 4.934141408099649e-05, "loss": 0.5556, "num_input_tokens_seen": 3963856, "step": 2930 }, { "epoch": 0.09394404967671724, "grad_norm": 1.4276096820831299, "learning_rate": 4.9339173627875135e-05, "loss": 0.5169, "num_input_tokens_seen": 3970528, "step": 2935 }, { "epoch": 0.09410409064720568, "grad_norm": 1.7926644086837769, "learning_rate": 4.9336929421347335e-05, "loss": 0.4528, "num_input_tokens_seen": 3977056, "step": 2940 }, { "epoch": 0.09426413161769413, "grad_norm": 1.5520756244659424, "learning_rate": 4.933468146175918e-05, "loss": 0.6137, "num_input_tokens_seen": 3983952, "step": 2945 }, { "epoch": 0.09442417258818257, "grad_norm": 1.3950293064117432, "learning_rate": 4.933242974945734e-05, "loss": 0.4974, "num_input_tokens_seen": 3990752, "step": 2950 }, { "epoch": 0.09458421355867103, "grad_norm": 1.6298500299453735, "learning_rate": 4.933017428478906e-05, "loss": 0.7629, "num_input_tokens_seen": 3997344, "step": 2955 }, { "epoch": 0.09474425452915947, "grad_norm": 2.2025249004364014, "learning_rate": 4.932791506810214e-05, "loss": 0.5353, "num_input_tokens_seen": 4004016, "step": 2960 }, { "epoch": 0.0949042954996479, "grad_norm": 0.8745484352111816, "learning_rate": 4.932565209974499e-05, "loss": 0.7507, "num_input_tokens_seen": 4010912, "step": 2965 }, { "epoch": 0.09506433647013636, "grad_norm": 2.1232616901397705, "learning_rate": 4.93233853800666e-05, "loss": 0.6023, "num_input_tokens_seen": 4017536, "step": 2970 }, { "epoch": 0.0952243774406248, "grad_norm": 1.6794153451919556, "learning_rate": 4.932111490941651e-05, "loss": 0.4502, "num_input_tokens_seen": 4024624, "step": 2975 }, { "epoch": 0.09538441841111324, "grad_norm": 2.2388198375701904, "learning_rate": 4.9318840688144876e-05, "loss": 0.6019, "num_input_tokens_seen": 4031776, "step": 2980 }, { "epoch": 0.09554445938160169, "grad_norm": 0.7292364239692688, "learning_rate": 4.9316562716602387e-05, "loss": 0.5269, "num_input_tokens_seen": 4039104, "step": 2985 }, { "epoch": 0.09570450035209013, "grad_norm": 0.48280322551727295, "learning_rate": 4.9314280995140346e-05, "loss": 0.841, "num_input_tokens_seen": 4046048, "step": 2990 }, { "epoch": 0.09586454132257859, "grad_norm": 1.0223358869552612, "learning_rate": 4.931199552411063e-05, "loss": 0.7113, "num_input_tokens_seen": 4052512, "step": 2995 }, { "epoch": 0.09602458229306703, "grad_norm": 2.5484964847564697, "learning_rate": 4.930970630386568e-05, "loss": 0.5585, "num_input_tokens_seen": 4059472, "step": 3000 }, { "epoch": 0.09602458229306703, "eval_loss": 0.6168442368507385, "eval_runtime": 499.1958, "eval_samples_per_second": 27.817, "eval_steps_per_second": 13.908, "num_input_tokens_seen": 4059472, "step": 3000 }, { "epoch": 0.09618462326355547, "grad_norm": 2.1462059020996094, "learning_rate": 4.9307413334758524e-05, "loss": 0.6953, "num_input_tokens_seen": 4066288, "step": 3005 }, { "epoch": 0.09634466423404392, "grad_norm": 0.7908008694648743, "learning_rate": 4.930511661714276e-05, "loss": 0.6264, "num_input_tokens_seen": 4072912, "step": 3010 }, { "epoch": 0.09650470520453236, "grad_norm": 0.824087917804718, "learning_rate": 4.9302816151372576e-05, "loss": 0.6543, "num_input_tokens_seen": 4080160, "step": 3015 }, { "epoch": 0.0966647461750208, "grad_norm": 0.7950488924980164, "learning_rate": 4.930051193780274e-05, "loss": 0.5782, "num_input_tokens_seen": 4086848, "step": 3020 }, { "epoch": 0.09682478714550925, "grad_norm": 1.4592303037643433, "learning_rate": 4.929820397678858e-05, "loss": 0.8233, "num_input_tokens_seen": 4093392, "step": 3025 }, { "epoch": 0.09698482811599769, "grad_norm": 2.1402671337127686, "learning_rate": 4.9295892268686015e-05, "loss": 0.5662, "num_input_tokens_seen": 4100160, "step": 3030 }, { "epoch": 0.09714486908648615, "grad_norm": 0.9423943161964417, "learning_rate": 4.9293576813851536e-05, "loss": 0.5044, "num_input_tokens_seen": 4107232, "step": 3035 }, { "epoch": 0.09730491005697459, "grad_norm": 1.5452136993408203, "learning_rate": 4.929125761264223e-05, "loss": 0.5858, "num_input_tokens_seen": 4113856, "step": 3040 }, { "epoch": 0.09746495102746303, "grad_norm": 1.6369479894638062, "learning_rate": 4.928893466541573e-05, "loss": 0.4162, "num_input_tokens_seen": 4120528, "step": 3045 }, { "epoch": 0.09762499199795148, "grad_norm": 2.951826572418213, "learning_rate": 4.928660797253027e-05, "loss": 0.9052, "num_input_tokens_seen": 4127088, "step": 3050 }, { "epoch": 0.09778503296843992, "grad_norm": 1.4742246866226196, "learning_rate": 4.928427753434467e-05, "loss": 0.5265, "num_input_tokens_seen": 4133392, "step": 3055 }, { "epoch": 0.09794507393892836, "grad_norm": 1.288465976715088, "learning_rate": 4.9281943351218286e-05, "loss": 0.6314, "num_input_tokens_seen": 4140320, "step": 3060 }, { "epoch": 0.09810511490941681, "grad_norm": 1.088657259941101, "learning_rate": 4.9279605423511095e-05, "loss": 0.5428, "num_input_tokens_seen": 4147104, "step": 3065 }, { "epoch": 0.09826515587990525, "grad_norm": 2.386012077331543, "learning_rate": 4.927726375158363e-05, "loss": 0.6005, "num_input_tokens_seen": 4153360, "step": 3070 }, { "epoch": 0.0984251968503937, "grad_norm": 1.4134021997451782, "learning_rate": 4.9274918335797004e-05, "loss": 0.5438, "num_input_tokens_seen": 4160000, "step": 3075 }, { "epoch": 0.09858523782088215, "grad_norm": 0.6076513528823853, "learning_rate": 4.927256917651292e-05, "loss": 0.4156, "num_input_tokens_seen": 4166880, "step": 3080 }, { "epoch": 0.09874527879137059, "grad_norm": 1.138389229774475, "learning_rate": 4.927021627409364e-05, "loss": 0.5187, "num_input_tokens_seen": 4173568, "step": 3085 }, { "epoch": 0.09890531976185904, "grad_norm": 1.4184844493865967, "learning_rate": 4.9267859628902005e-05, "loss": 0.8398, "num_input_tokens_seen": 4180000, "step": 3090 }, { "epoch": 0.09906536073234748, "grad_norm": 1.171850562095642, "learning_rate": 4.9265499241301454e-05, "loss": 0.6241, "num_input_tokens_seen": 4186736, "step": 3095 }, { "epoch": 0.09922540170283592, "grad_norm": 1.5810189247131348, "learning_rate": 4.926313511165598e-05, "loss": 0.5478, "num_input_tokens_seen": 4193344, "step": 3100 }, { "epoch": 0.09938544267332437, "grad_norm": 1.0761613845825195, "learning_rate": 4.926076724033016e-05, "loss": 0.5219, "num_input_tokens_seen": 4200512, "step": 3105 }, { "epoch": 0.09954548364381281, "grad_norm": 1.3708460330963135, "learning_rate": 4.9258395627689146e-05, "loss": 0.7004, "num_input_tokens_seen": 4207104, "step": 3110 }, { "epoch": 0.09970552461430127, "grad_norm": 1.2405073642730713, "learning_rate": 4.925602027409868e-05, "loss": 0.6047, "num_input_tokens_seen": 4214240, "step": 3115 }, { "epoch": 0.0998655655847897, "grad_norm": 0.9437921643257141, "learning_rate": 4.925364117992507e-05, "loss": 0.5504, "num_input_tokens_seen": 4220784, "step": 3120 }, { "epoch": 0.10002560655527815, "grad_norm": 1.2724955081939697, "learning_rate": 4.92512583455352e-05, "loss": 0.5586, "num_input_tokens_seen": 4227728, "step": 3125 }, { "epoch": 0.1001856475257666, "grad_norm": 1.2205309867858887, "learning_rate": 4.9248871771296536e-05, "loss": 0.7943, "num_input_tokens_seen": 4234544, "step": 3130 }, { "epoch": 0.10034568849625504, "grad_norm": 1.8284504413604736, "learning_rate": 4.924648145757711e-05, "loss": 0.6962, "num_input_tokens_seen": 4241024, "step": 3135 }, { "epoch": 0.1005057294667435, "grad_norm": 1.5644618272781372, "learning_rate": 4.924408740474554e-05, "loss": 0.6277, "num_input_tokens_seen": 4248288, "step": 3140 }, { "epoch": 0.10066577043723193, "grad_norm": 1.7254225015640259, "learning_rate": 4.924168961317103e-05, "loss": 0.8255, "num_input_tokens_seen": 4255104, "step": 3145 }, { "epoch": 0.10082581140772037, "grad_norm": 1.3672162294387817, "learning_rate": 4.9239288083223334e-05, "loss": 0.5204, "num_input_tokens_seen": 4262400, "step": 3150 }, { "epoch": 0.10098585237820883, "grad_norm": 1.3189842700958252, "learning_rate": 4.9236882815272803e-05, "loss": 0.4296, "num_input_tokens_seen": 4269504, "step": 3155 }, { "epoch": 0.10114589334869727, "grad_norm": 0.7810506224632263, "learning_rate": 4.9234473809690365e-05, "loss": 0.56, "num_input_tokens_seen": 4276320, "step": 3160 }, { "epoch": 0.1013059343191857, "grad_norm": 0.873445451259613, "learning_rate": 4.923206106684752e-05, "loss": 0.5938, "num_input_tokens_seen": 4283392, "step": 3165 }, { "epoch": 0.10146597528967416, "grad_norm": 3.192701578140259, "learning_rate": 4.922964458711634e-05, "loss": 0.533, "num_input_tokens_seen": 4290080, "step": 3170 }, { "epoch": 0.1016260162601626, "grad_norm": 0.5550103783607483, "learning_rate": 4.9227224370869474e-05, "loss": 0.7387, "num_input_tokens_seen": 4296688, "step": 3175 }, { "epoch": 0.10178605723065105, "grad_norm": 2.903463363647461, "learning_rate": 4.9224800418480155e-05, "loss": 0.6804, "num_input_tokens_seen": 4303792, "step": 3180 }, { "epoch": 0.10194609820113949, "grad_norm": 1.322322964668274, "learning_rate": 4.9222372730322176e-05, "loss": 0.7822, "num_input_tokens_seen": 4310784, "step": 3185 }, { "epoch": 0.10210613917162793, "grad_norm": 1.6664186716079712, "learning_rate": 4.921994130676993e-05, "loss": 0.6533, "num_input_tokens_seen": 4317680, "step": 3190 }, { "epoch": 0.10226618014211639, "grad_norm": 1.433256983757019, "learning_rate": 4.9217506148198366e-05, "loss": 0.4398, "num_input_tokens_seen": 4324368, "step": 3195 }, { "epoch": 0.10242622111260483, "grad_norm": 1.2094494104385376, "learning_rate": 4.921506725498302e-05, "loss": 0.5138, "num_input_tokens_seen": 4331088, "step": 3200 }, { "epoch": 0.10242622111260483, "eval_loss": 0.6087194681167603, "eval_runtime": 499.916, "eval_samples_per_second": 27.777, "eval_steps_per_second": 13.888, "num_input_tokens_seen": 4331088, "step": 3200 }, { "epoch": 0.10258626208309327, "grad_norm": 0.9863497614860535, "learning_rate": 4.9212624627499994e-05, "loss": 0.458, "num_input_tokens_seen": 4337776, "step": 3205 }, { "epoch": 0.10274630305358172, "grad_norm": 1.752831220626831, "learning_rate": 4.921017826612597e-05, "loss": 0.5019, "num_input_tokens_seen": 4344624, "step": 3210 }, { "epoch": 0.10290634402407016, "grad_norm": 0.5727003216743469, "learning_rate": 4.9207728171238223e-05, "loss": 0.4176, "num_input_tokens_seen": 4351664, "step": 3215 }, { "epoch": 0.10306638499455861, "grad_norm": 1.165632724761963, "learning_rate": 4.920527434321458e-05, "loss": 0.5831, "num_input_tokens_seen": 4358576, "step": 3220 }, { "epoch": 0.10322642596504705, "grad_norm": 0.7876461744308472, "learning_rate": 4.920281678243345e-05, "loss": 0.6354, "num_input_tokens_seen": 4365216, "step": 3225 }, { "epoch": 0.10338646693553549, "grad_norm": 1.295302152633667, "learning_rate": 4.920035548927381e-05, "loss": 0.6102, "num_input_tokens_seen": 4371936, "step": 3230 }, { "epoch": 0.10354650790602395, "grad_norm": 1.3349084854125977, "learning_rate": 4.919789046411525e-05, "loss": 0.7085, "num_input_tokens_seen": 4378640, "step": 3235 }, { "epoch": 0.10370654887651239, "grad_norm": 1.9360296726226807, "learning_rate": 4.919542170733787e-05, "loss": 0.6896, "num_input_tokens_seen": 4385248, "step": 3240 }, { "epoch": 0.10386658984700083, "grad_norm": 2.81720232963562, "learning_rate": 4.919294921932242e-05, "loss": 0.5992, "num_input_tokens_seen": 4392080, "step": 3245 }, { "epoch": 0.10402663081748928, "grad_norm": 1.7356699705123901, "learning_rate": 4.919047300045016e-05, "loss": 0.6387, "num_input_tokens_seen": 4398784, "step": 3250 }, { "epoch": 0.10418667178797772, "grad_norm": 2.520958423614502, "learning_rate": 4.918799305110299e-05, "loss": 0.762, "num_input_tokens_seen": 4405376, "step": 3255 }, { "epoch": 0.10434671275846617, "grad_norm": 1.6158876419067383, "learning_rate": 4.918550937166331e-05, "loss": 0.5665, "num_input_tokens_seen": 4412112, "step": 3260 }, { "epoch": 0.10450675372895461, "grad_norm": 1.0740933418273926, "learning_rate": 4.918302196251415e-05, "loss": 0.4932, "num_input_tokens_seen": 4418816, "step": 3265 }, { "epoch": 0.10466679469944305, "grad_norm": 1.6066100597381592, "learning_rate": 4.91805308240391e-05, "loss": 0.5378, "num_input_tokens_seen": 4425472, "step": 3270 }, { "epoch": 0.1048268356699315, "grad_norm": 1.7822129726409912, "learning_rate": 4.9178035956622326e-05, "loss": 0.495, "num_input_tokens_seen": 4432176, "step": 3275 }, { "epoch": 0.10498687664041995, "grad_norm": 1.2913322448730469, "learning_rate": 4.917553736064857e-05, "loss": 0.6111, "num_input_tokens_seen": 4438928, "step": 3280 }, { "epoch": 0.10514691761090839, "grad_norm": 0.9547336101531982, "learning_rate": 4.917303503650314e-05, "loss": 0.5144, "num_input_tokens_seen": 4445216, "step": 3285 }, { "epoch": 0.10530695858139684, "grad_norm": 0.8696664571762085, "learning_rate": 4.917052898457194e-05, "loss": 0.6825, "num_input_tokens_seen": 4451760, "step": 3290 }, { "epoch": 0.10546699955188528, "grad_norm": 1.0713074207305908, "learning_rate": 4.916801920524141e-05, "loss": 0.5426, "num_input_tokens_seen": 4458752, "step": 3295 }, { "epoch": 0.10562704052237373, "grad_norm": 1.621087670326233, "learning_rate": 4.916550569889862e-05, "loss": 0.4746, "num_input_tokens_seen": 4465632, "step": 3300 }, { "epoch": 0.10578708149286217, "grad_norm": 1.7436963319778442, "learning_rate": 4.916298846593116e-05, "loss": 0.775, "num_input_tokens_seen": 4472224, "step": 3305 }, { "epoch": 0.10594712246335061, "grad_norm": 1.3318462371826172, "learning_rate": 4.916046750672722e-05, "loss": 0.5553, "num_input_tokens_seen": 4478624, "step": 3310 }, { "epoch": 0.10610716343383907, "grad_norm": 2.170288562774658, "learning_rate": 4.915794282167559e-05, "loss": 0.6452, "num_input_tokens_seen": 4485488, "step": 3315 }, { "epoch": 0.1062672044043275, "grad_norm": 1.47194242477417, "learning_rate": 4.915541441116558e-05, "loss": 0.5559, "num_input_tokens_seen": 4492736, "step": 3320 }, { "epoch": 0.10642724537481596, "grad_norm": 0.8458205461502075, "learning_rate": 4.915288227558711e-05, "loss": 0.4861, "num_input_tokens_seen": 4499504, "step": 3325 }, { "epoch": 0.1065872863453044, "grad_norm": 1.7209618091583252, "learning_rate": 4.915034641533066e-05, "loss": 0.663, "num_input_tokens_seen": 4506192, "step": 3330 }, { "epoch": 0.10674732731579284, "grad_norm": 0.965593695640564, "learning_rate": 4.914780683078731e-05, "loss": 0.5089, "num_input_tokens_seen": 4512880, "step": 3335 }, { "epoch": 0.10690736828628129, "grad_norm": 2.297039270401001, "learning_rate": 4.9145263522348695e-05, "loss": 0.6091, "num_input_tokens_seen": 4519824, "step": 3340 }, { "epoch": 0.10706740925676973, "grad_norm": 0.736360490322113, "learning_rate": 4.9142716490407e-05, "loss": 0.4587, "num_input_tokens_seen": 4526880, "step": 3345 }, { "epoch": 0.10722745022725817, "grad_norm": 1.5027530193328857, "learning_rate": 4.914016573535504e-05, "loss": 0.668, "num_input_tokens_seen": 4534304, "step": 3350 }, { "epoch": 0.10738749119774663, "grad_norm": 1.294787883758545, "learning_rate": 4.9137611257586154e-05, "loss": 0.5795, "num_input_tokens_seen": 4540880, "step": 3355 }, { "epoch": 0.10754753216823507, "grad_norm": 1.094363808631897, "learning_rate": 4.9135053057494274e-05, "loss": 0.6117, "num_input_tokens_seen": 4547520, "step": 3360 }, { "epoch": 0.10770757313872352, "grad_norm": 1.3163937330245972, "learning_rate": 4.913249113547392e-05, "loss": 0.6674, "num_input_tokens_seen": 4554240, "step": 3365 }, { "epoch": 0.10786761410921196, "grad_norm": 2.4132399559020996, "learning_rate": 4.912992549192016e-05, "loss": 0.5668, "num_input_tokens_seen": 4561040, "step": 3370 }, { "epoch": 0.1080276550797004, "grad_norm": 1.6645458936691284, "learning_rate": 4.9127356127228665e-05, "loss": 0.855, "num_input_tokens_seen": 4567920, "step": 3375 }, { "epoch": 0.10818769605018885, "grad_norm": 2.0332345962524414, "learning_rate": 4.912478304179564e-05, "loss": 0.5803, "num_input_tokens_seen": 4574592, "step": 3380 }, { "epoch": 0.10834773702067729, "grad_norm": 1.445198655128479, "learning_rate": 4.9122206236017896e-05, "loss": 0.5524, "num_input_tokens_seen": 4581184, "step": 3385 }, { "epoch": 0.10850777799116573, "grad_norm": 2.482248306274414, "learning_rate": 4.911962571029282e-05, "loss": 0.635, "num_input_tokens_seen": 4587968, "step": 3390 }, { "epoch": 0.10866781896165419, "grad_norm": 1.3639321327209473, "learning_rate": 4.9117041465018353e-05, "loss": 0.7157, "num_input_tokens_seen": 4594368, "step": 3395 }, { "epoch": 0.10882785993214263, "grad_norm": 1.0654444694519043, "learning_rate": 4.911445350059302e-05, "loss": 0.539, "num_input_tokens_seen": 4601728, "step": 3400 }, { "epoch": 0.10882785993214263, "eval_loss": 0.6047332882881165, "eval_runtime": 500.1544, "eval_samples_per_second": 27.763, "eval_steps_per_second": 13.882, "num_input_tokens_seen": 4601728, "step": 3400 }, { "epoch": 0.10898790090263108, "grad_norm": 0.9278212189674377, "learning_rate": 4.9111861817415905e-05, "loss": 0.5952, "num_input_tokens_seen": 4608336, "step": 3405 }, { "epoch": 0.10914794187311952, "grad_norm": 1.4870777130126953, "learning_rate": 4.91092664158867e-05, "loss": 0.6876, "num_input_tokens_seen": 4614848, "step": 3410 }, { "epoch": 0.10930798284360796, "grad_norm": 2.9939186573028564, "learning_rate": 4.910666729640563e-05, "loss": 0.9156, "num_input_tokens_seen": 4622000, "step": 3415 }, { "epoch": 0.10946802381409641, "grad_norm": 1.3264721632003784, "learning_rate": 4.910406445937353e-05, "loss": 0.5498, "num_input_tokens_seen": 4628848, "step": 3420 }, { "epoch": 0.10962806478458485, "grad_norm": 3.2926154136657715, "learning_rate": 4.9101457905191774e-05, "loss": 0.6382, "num_input_tokens_seen": 4635952, "step": 3425 }, { "epoch": 0.10978810575507329, "grad_norm": 1.2220007181167603, "learning_rate": 4.909884763426233e-05, "loss": 0.5569, "num_input_tokens_seen": 4642480, "step": 3430 }, { "epoch": 0.10994814672556175, "grad_norm": 1.5739187002182007, "learning_rate": 4.9096233646987736e-05, "loss": 0.6527, "num_input_tokens_seen": 4650032, "step": 3435 }, { "epoch": 0.11010818769605019, "grad_norm": 2.064406394958496, "learning_rate": 4.9093615943771104e-05, "loss": 0.8282, "num_input_tokens_seen": 4657488, "step": 3440 }, { "epoch": 0.11026822866653864, "grad_norm": 1.7302321195602417, "learning_rate": 4.909099452501611e-05, "loss": 0.5838, "num_input_tokens_seen": 4663776, "step": 3445 }, { "epoch": 0.11042826963702708, "grad_norm": 1.8613730669021606, "learning_rate": 4.908836939112702e-05, "loss": 0.561, "num_input_tokens_seen": 4670608, "step": 3450 }, { "epoch": 0.11058831060751552, "grad_norm": 0.4942485988140106, "learning_rate": 4.908574054250865e-05, "loss": 0.3502, "num_input_tokens_seen": 4677520, "step": 3455 }, { "epoch": 0.11074835157800397, "grad_norm": 0.6757672429084778, "learning_rate": 4.9083107979566414e-05, "loss": 0.448, "num_input_tokens_seen": 4684320, "step": 3460 }, { "epoch": 0.11090839254849241, "grad_norm": 1.5762038230895996, "learning_rate": 4.908047170270628e-05, "loss": 0.7793, "num_input_tokens_seen": 4691280, "step": 3465 }, { "epoch": 0.11106843351898085, "grad_norm": 0.9814431071281433, "learning_rate": 4.9077831712334784e-05, "loss": 0.5794, "num_input_tokens_seen": 4697648, "step": 3470 }, { "epoch": 0.1112284744894693, "grad_norm": 1.7980353832244873, "learning_rate": 4.907518800885907e-05, "loss": 0.4712, "num_input_tokens_seen": 4704224, "step": 3475 }, { "epoch": 0.11138851545995775, "grad_norm": 0.7281551957130432, "learning_rate": 4.907254059268681e-05, "loss": 0.7249, "num_input_tokens_seen": 4711216, "step": 3480 }, { "epoch": 0.1115485564304462, "grad_norm": 0.678212583065033, "learning_rate": 4.906988946422628e-05, "loss": 0.6752, "num_input_tokens_seen": 4717984, "step": 3485 }, { "epoch": 0.11170859740093464, "grad_norm": 1.8170686960220337, "learning_rate": 4.9067234623886315e-05, "loss": 0.5803, "num_input_tokens_seen": 4724816, "step": 3490 }, { "epoch": 0.11186863837142308, "grad_norm": 0.4114993214607239, "learning_rate": 4.9064576072076316e-05, "loss": 0.4237, "num_input_tokens_seen": 4731584, "step": 3495 }, { "epoch": 0.11202867934191153, "grad_norm": 0.6318671703338623, "learning_rate": 4.906191380920628e-05, "loss": 0.5491, "num_input_tokens_seen": 4738640, "step": 3500 }, { "epoch": 0.11218872031239997, "grad_norm": 0.8237159848213196, "learning_rate": 4.905924783568675e-05, "loss": 0.545, "num_input_tokens_seen": 4745872, "step": 3505 }, { "epoch": 0.11234876128288843, "grad_norm": 0.8686559796333313, "learning_rate": 4.905657815192886e-05, "loss": 0.5698, "num_input_tokens_seen": 4752672, "step": 3510 }, { "epoch": 0.11250880225337687, "grad_norm": 1.2189011573791504, "learning_rate": 4.90539047583443e-05, "loss": 0.6172, "num_input_tokens_seen": 4759488, "step": 3515 }, { "epoch": 0.1126688432238653, "grad_norm": 1.3878720998764038, "learning_rate": 4.905122765534534e-05, "loss": 0.5464, "num_input_tokens_seen": 4765904, "step": 3520 }, { "epoch": 0.11282888419435376, "grad_norm": 1.6992074251174927, "learning_rate": 4.9048546843344846e-05, "loss": 0.6488, "num_input_tokens_seen": 4773488, "step": 3525 }, { "epoch": 0.1129889251648422, "grad_norm": 1.528090476989746, "learning_rate": 4.9045862322756206e-05, "loss": 0.584, "num_input_tokens_seen": 4780400, "step": 3530 }, { "epoch": 0.11314896613533064, "grad_norm": 2.425684928894043, "learning_rate": 4.904317409399342e-05, "loss": 0.6558, "num_input_tokens_seen": 4787152, "step": 3535 }, { "epoch": 0.11330900710581909, "grad_norm": 0.786251425743103, "learning_rate": 4.904048215747104e-05, "loss": 0.5965, "num_input_tokens_seen": 4793664, "step": 3540 }, { "epoch": 0.11346904807630753, "grad_norm": 1.0858500003814697, "learning_rate": 4.90377865136042e-05, "loss": 0.7947, "num_input_tokens_seen": 4803888, "step": 3545 }, { "epoch": 0.11362908904679599, "grad_norm": 0.8607036471366882, "learning_rate": 4.90350871628086e-05, "loss": 0.5181, "num_input_tokens_seen": 4810480, "step": 3550 }, { "epoch": 0.11378913001728443, "grad_norm": 2.661574125289917, "learning_rate": 4.903238410550052e-05, "loss": 0.6238, "num_input_tokens_seen": 4817424, "step": 3555 }, { "epoch": 0.11394917098777287, "grad_norm": 0.806450605392456, "learning_rate": 4.90296773420968e-05, "loss": 0.6148, "num_input_tokens_seen": 4824032, "step": 3560 }, { "epoch": 0.11410921195826132, "grad_norm": 1.2536786794662476, "learning_rate": 4.902696687301486e-05, "loss": 0.5673, "num_input_tokens_seen": 4830848, "step": 3565 }, { "epoch": 0.11426925292874976, "grad_norm": 4.59646463394165, "learning_rate": 4.902425269867268e-05, "loss": 0.7192, "num_input_tokens_seen": 4837424, "step": 3570 }, { "epoch": 0.1144292938992382, "grad_norm": 0.8728713989257812, "learning_rate": 4.902153481948883e-05, "loss": 0.5187, "num_input_tokens_seen": 4843952, "step": 3575 }, { "epoch": 0.11458933486972665, "grad_norm": 1.4345433712005615, "learning_rate": 4.901881323588244e-05, "loss": 0.5475, "num_input_tokens_seen": 4850512, "step": 3580 }, { "epoch": 0.11474937584021509, "grad_norm": 1.1448978185653687, "learning_rate": 4.90160879482732e-05, "loss": 0.4858, "num_input_tokens_seen": 4857264, "step": 3585 }, { "epoch": 0.11490941681070355, "grad_norm": 0.9748579859733582, "learning_rate": 4.9013358957081405e-05, "loss": 0.5171, "num_input_tokens_seen": 4864064, "step": 3590 }, { "epoch": 0.11506945778119199, "grad_norm": 0.9702478051185608, "learning_rate": 4.901062626272789e-05, "loss": 0.6992, "num_input_tokens_seen": 4870512, "step": 3595 }, { "epoch": 0.11522949875168043, "grad_norm": 1.6208127737045288, "learning_rate": 4.900788986563406e-05, "loss": 0.4948, "num_input_tokens_seen": 4877104, "step": 3600 }, { "epoch": 0.11522949875168043, "eval_loss": 0.5977566838264465, "eval_runtime": 499.1816, "eval_samples_per_second": 27.818, "eval_steps_per_second": 13.909, "num_input_tokens_seen": 4877104, "step": 3600 }, { "epoch": 0.11538953972216888, "grad_norm": 0.7896119952201843, "learning_rate": 4.9005149766221915e-05, "loss": 0.6309, "num_input_tokens_seen": 4884000, "step": 3605 }, { "epoch": 0.11554958069265732, "grad_norm": 0.9273166656494141, "learning_rate": 4.9002405964914e-05, "loss": 0.6315, "num_input_tokens_seen": 4891184, "step": 3610 }, { "epoch": 0.11570962166314576, "grad_norm": 0.7066910862922668, "learning_rate": 4.899965846213346e-05, "loss": 0.2915, "num_input_tokens_seen": 4897792, "step": 3615 }, { "epoch": 0.11586966263363421, "grad_norm": 2.0816190242767334, "learning_rate": 4.899690725830399e-05, "loss": 0.6477, "num_input_tokens_seen": 4904864, "step": 3620 }, { "epoch": 0.11602970360412265, "grad_norm": 1.0076359510421753, "learning_rate": 4.899415235384985e-05, "loss": 0.6514, "num_input_tokens_seen": 4911936, "step": 3625 }, { "epoch": 0.1161897445746111, "grad_norm": 1.1959031820297241, "learning_rate": 4.899139374919589e-05, "loss": 0.4931, "num_input_tokens_seen": 4918368, "step": 3630 }, { "epoch": 0.11634978554509955, "grad_norm": 1.2262952327728271, "learning_rate": 4.898863144476752e-05, "loss": 0.5479, "num_input_tokens_seen": 4925616, "step": 3635 }, { "epoch": 0.11650982651558799, "grad_norm": 1.275153398513794, "learning_rate": 4.898586544099072e-05, "loss": 0.494, "num_input_tokens_seen": 4932416, "step": 3640 }, { "epoch": 0.11666986748607644, "grad_norm": 1.19944167137146, "learning_rate": 4.898309573829204e-05, "loss": 0.4645, "num_input_tokens_seen": 4938976, "step": 3645 }, { "epoch": 0.11682990845656488, "grad_norm": 1.425705909729004, "learning_rate": 4.898032233709862e-05, "loss": 0.5476, "num_input_tokens_seen": 4946112, "step": 3650 }, { "epoch": 0.11698994942705332, "grad_norm": 2.7030248641967773, "learning_rate": 4.8977545237838123e-05, "loss": 0.8684, "num_input_tokens_seen": 4952512, "step": 3655 }, { "epoch": 0.11714999039754177, "grad_norm": 1.3626596927642822, "learning_rate": 4.8974764440938836e-05, "loss": 0.4848, "num_input_tokens_seen": 4959008, "step": 3660 }, { "epoch": 0.11731003136803021, "grad_norm": 0.9310556054115295, "learning_rate": 4.897197994682959e-05, "loss": 0.593, "num_input_tokens_seen": 4965872, "step": 3665 }, { "epoch": 0.11747007233851867, "grad_norm": 1.493070363998413, "learning_rate": 4.8969191755939786e-05, "loss": 0.4739, "num_input_tokens_seen": 4972976, "step": 3670 }, { "epoch": 0.1176301133090071, "grad_norm": 1.3138806819915771, "learning_rate": 4.8966399868699396e-05, "loss": 0.7222, "num_input_tokens_seen": 4979984, "step": 3675 }, { "epoch": 0.11779015427949555, "grad_norm": 0.8768885731697083, "learning_rate": 4.8963604285538965e-05, "loss": 0.4182, "num_input_tokens_seen": 4987152, "step": 3680 }, { "epoch": 0.117950195249984, "grad_norm": 1.0247873067855835, "learning_rate": 4.8960805006889604e-05, "loss": 0.4449, "num_input_tokens_seen": 4994912, "step": 3685 }, { "epoch": 0.11811023622047244, "grad_norm": 0.7942581176757812, "learning_rate": 4.8958002033183004e-05, "loss": 0.543, "num_input_tokens_seen": 5001824, "step": 3690 }, { "epoch": 0.11827027719096088, "grad_norm": 2.5215110778808594, "learning_rate": 4.8955195364851414e-05, "loss": 0.6459, "num_input_tokens_seen": 5008992, "step": 3695 }, { "epoch": 0.11843031816144933, "grad_norm": 0.880209743976593, "learning_rate": 4.895238500232766e-05, "loss": 0.6238, "num_input_tokens_seen": 5015616, "step": 3700 }, { "epoch": 0.11859035913193777, "grad_norm": 1.030182123184204, "learning_rate": 4.8949570946045143e-05, "loss": 0.5825, "num_input_tokens_seen": 5022208, "step": 3705 }, { "epoch": 0.11875040010242623, "grad_norm": 2.061307191848755, "learning_rate": 4.89467531964378e-05, "loss": 0.6843, "num_input_tokens_seen": 5028480, "step": 3710 }, { "epoch": 0.11891044107291467, "grad_norm": 1.5313334465026855, "learning_rate": 4.894393175394019e-05, "loss": 0.5451, "num_input_tokens_seen": 5035360, "step": 3715 }, { "epoch": 0.1190704820434031, "grad_norm": 1.0037893056869507, "learning_rate": 4.8941106618987406e-05, "loss": 0.7125, "num_input_tokens_seen": 5042112, "step": 3720 }, { "epoch": 0.11923052301389156, "grad_norm": 1.263504981994629, "learning_rate": 4.893827779201512e-05, "loss": 0.4572, "num_input_tokens_seen": 5049072, "step": 3725 }, { "epoch": 0.11939056398438, "grad_norm": 2.518761157989502, "learning_rate": 4.893544527345957e-05, "loss": 0.5256, "num_input_tokens_seen": 5055440, "step": 3730 }, { "epoch": 0.11955060495486845, "grad_norm": 0.7844130396842957, "learning_rate": 4.8932609063757563e-05, "loss": 0.3911, "num_input_tokens_seen": 5062432, "step": 3735 }, { "epoch": 0.11971064592535689, "grad_norm": 1.7208037376403809, "learning_rate": 4.8929769163346484e-05, "loss": 0.6083, "num_input_tokens_seen": 5069280, "step": 3740 }, { "epoch": 0.11987068689584533, "grad_norm": 2.1308913230895996, "learning_rate": 4.892692557266429e-05, "loss": 0.56, "num_input_tokens_seen": 5076016, "step": 3745 }, { "epoch": 0.12003072786633379, "grad_norm": 0.6621193289756775, "learning_rate": 4.8924078292149464e-05, "loss": 0.5993, "num_input_tokens_seen": 5083024, "step": 3750 }, { "epoch": 0.12019076883682223, "grad_norm": 1.4653432369232178, "learning_rate": 4.892122732224114e-05, "loss": 0.624, "num_input_tokens_seen": 5089840, "step": 3755 }, { "epoch": 0.12035080980731067, "grad_norm": 0.8312159180641174, "learning_rate": 4.8918372663378944e-05, "loss": 0.5846, "num_input_tokens_seen": 5096768, "step": 3760 }, { "epoch": 0.12051085077779912, "grad_norm": 3.0056469440460205, "learning_rate": 4.89155143160031e-05, "loss": 0.9993, "num_input_tokens_seen": 5104032, "step": 3765 }, { "epoch": 0.12067089174828756, "grad_norm": 1.0460724830627441, "learning_rate": 4.891265228055441e-05, "loss": 0.6609, "num_input_tokens_seen": 5110544, "step": 3770 }, { "epoch": 0.12083093271877601, "grad_norm": 1.1402467489242554, "learning_rate": 4.890978655747424e-05, "loss": 0.6923, "num_input_tokens_seen": 5117344, "step": 3775 }, { "epoch": 0.12099097368926445, "grad_norm": 1.102414608001709, "learning_rate": 4.89069171472045e-05, "loss": 0.5187, "num_input_tokens_seen": 5124064, "step": 3780 }, { "epoch": 0.12115101465975289, "grad_norm": 1.7425309419631958, "learning_rate": 4.890404405018772e-05, "loss": 0.5173, "num_input_tokens_seen": 5130560, "step": 3785 }, { "epoch": 0.12131105563024135, "grad_norm": 1.1629689931869507, "learning_rate": 4.8901167266866934e-05, "loss": 0.6379, "num_input_tokens_seen": 5137088, "step": 3790 }, { "epoch": 0.12147109660072979, "grad_norm": 0.9707894325256348, "learning_rate": 4.88982867976858e-05, "loss": 0.5825, "num_input_tokens_seen": 5143792, "step": 3795 }, { "epoch": 0.12163113757121823, "grad_norm": 0.8016886115074158, "learning_rate": 4.889540264308852e-05, "loss": 0.5902, "num_input_tokens_seen": 5150656, "step": 3800 }, { "epoch": 0.12163113757121823, "eval_loss": 0.5923164486885071, "eval_runtime": 500.1183, "eval_samples_per_second": 27.765, "eval_steps_per_second": 13.883, "num_input_tokens_seen": 5150656, "step": 3800 }, { "epoch": 0.12179117854170668, "grad_norm": 1.9624301195144653, "learning_rate": 4.889251480351986e-05, "loss": 0.6462, "num_input_tokens_seen": 5157456, "step": 3805 }, { "epoch": 0.12195121951219512, "grad_norm": 1.8644828796386719, "learning_rate": 4.888962327942517e-05, "loss": 0.5127, "num_input_tokens_seen": 5164240, "step": 3810 }, { "epoch": 0.12211126048268357, "grad_norm": 1.0446077585220337, "learning_rate": 4.8886728071250356e-05, "loss": 0.6613, "num_input_tokens_seen": 5171040, "step": 3815 }, { "epoch": 0.12227130145317201, "grad_norm": 1.7576881647109985, "learning_rate": 4.8883829179441884e-05, "loss": 0.4506, "num_input_tokens_seen": 5178016, "step": 3820 }, { "epoch": 0.12243134242366045, "grad_norm": 0.864372730255127, "learning_rate": 4.888092660444682e-05, "loss": 0.6069, "num_input_tokens_seen": 5184528, "step": 3825 }, { "epoch": 0.1225913833941489, "grad_norm": 2.090334177017212, "learning_rate": 4.887802034671276e-05, "loss": 0.3581, "num_input_tokens_seen": 5191328, "step": 3830 }, { "epoch": 0.12275142436463735, "grad_norm": 0.8666548132896423, "learning_rate": 4.88751104066879e-05, "loss": 0.5864, "num_input_tokens_seen": 5198960, "step": 3835 }, { "epoch": 0.12291146533512579, "grad_norm": 1.533862590789795, "learning_rate": 4.887219678482098e-05, "loss": 0.5835, "num_input_tokens_seen": 5205536, "step": 3840 }, { "epoch": 0.12307150630561424, "grad_norm": 2.5892200469970703, "learning_rate": 4.8869279481561316e-05, "loss": 0.5615, "num_input_tokens_seen": 5212368, "step": 3845 }, { "epoch": 0.12323154727610268, "grad_norm": 0.9856249094009399, "learning_rate": 4.88663584973588e-05, "loss": 0.4593, "num_input_tokens_seen": 5218912, "step": 3850 }, { "epoch": 0.12339158824659113, "grad_norm": 1.5026112794876099, "learning_rate": 4.8863433832663874e-05, "loss": 0.4123, "num_input_tokens_seen": 5225408, "step": 3855 }, { "epoch": 0.12355162921707957, "grad_norm": 2.2003509998321533, "learning_rate": 4.886050548792757e-05, "loss": 0.5806, "num_input_tokens_seen": 5232016, "step": 3860 }, { "epoch": 0.12371167018756801, "grad_norm": 2.483973741531372, "learning_rate": 4.8857573463601465e-05, "loss": 0.5625, "num_input_tokens_seen": 5238560, "step": 3865 }, { "epoch": 0.12387171115805647, "grad_norm": 1.400128960609436, "learning_rate": 4.885463776013772e-05, "loss": 0.6201, "num_input_tokens_seen": 5245664, "step": 3870 }, { "epoch": 0.1240317521285449, "grad_norm": 1.5218695402145386, "learning_rate": 4.8851698377989056e-05, "loss": 0.6092, "num_input_tokens_seen": 5252288, "step": 3875 }, { "epoch": 0.12419179309903335, "grad_norm": 1.3885204792022705, "learning_rate": 4.884875531760876e-05, "loss": 0.524, "num_input_tokens_seen": 5258528, "step": 3880 }, { "epoch": 0.1243518340695218, "grad_norm": 1.799195408821106, "learning_rate": 4.88458085794507e-05, "loss": 0.4901, "num_input_tokens_seen": 5265264, "step": 3885 }, { "epoch": 0.12451187504001024, "grad_norm": 0.9448595643043518, "learning_rate": 4.884285816396929e-05, "loss": 0.5407, "num_input_tokens_seen": 5272032, "step": 3890 }, { "epoch": 0.12467191601049869, "grad_norm": 0.6818589568138123, "learning_rate": 4.8839904071619526e-05, "loss": 0.4409, "num_input_tokens_seen": 5278896, "step": 3895 }, { "epoch": 0.12483195698098713, "grad_norm": 1.7469438314437866, "learning_rate": 4.8836946302856955e-05, "loss": 0.6019, "num_input_tokens_seen": 5285808, "step": 3900 }, { "epoch": 0.12499199795147557, "grad_norm": 1.2480159997940063, "learning_rate": 4.8833984858137715e-05, "loss": 0.6821, "num_input_tokens_seen": 5292624, "step": 3905 }, { "epoch": 0.12515203892196403, "grad_norm": 1.6796382665634155, "learning_rate": 4.8831019737918494e-05, "loss": 0.6045, "num_input_tokens_seen": 5300064, "step": 3910 }, { "epoch": 0.12531207989245247, "grad_norm": 1.8034220933914185, "learning_rate": 4.882805094265655e-05, "loss": 0.5802, "num_input_tokens_seen": 5307056, "step": 3915 }, { "epoch": 0.1254721208629409, "grad_norm": 1.2517703771591187, "learning_rate": 4.8825078472809706e-05, "loss": 0.6019, "num_input_tokens_seen": 5313680, "step": 3920 }, { "epoch": 0.12563216183342935, "grad_norm": 1.1493606567382812, "learning_rate": 4.882210232883635e-05, "loss": 0.5881, "num_input_tokens_seen": 5320368, "step": 3925 }, { "epoch": 0.1257922028039178, "grad_norm": 1.3094041347503662, "learning_rate": 4.881912251119546e-05, "loss": 0.4429, "num_input_tokens_seen": 5326912, "step": 3930 }, { "epoch": 0.12595224377440625, "grad_norm": 1.0288978815078735, "learning_rate": 4.881613902034654e-05, "loss": 0.6191, "num_input_tokens_seen": 5333856, "step": 3935 }, { "epoch": 0.1261122847448947, "grad_norm": 1.093440055847168, "learning_rate": 4.88131518567497e-05, "loss": 0.5173, "num_input_tokens_seen": 5340880, "step": 3940 }, { "epoch": 0.12627232571538313, "grad_norm": 0.9543217420578003, "learning_rate": 4.881016102086558e-05, "loss": 0.7039, "num_input_tokens_seen": 5347424, "step": 3945 }, { "epoch": 0.12643236668587157, "grad_norm": 0.8789275288581848, "learning_rate": 4.8807166513155425e-05, "loss": 0.4394, "num_input_tokens_seen": 5354736, "step": 3950 }, { "epoch": 0.12659240765636004, "grad_norm": 1.7663863897323608, "learning_rate": 4.8804168334081004e-05, "loss": 0.4662, "num_input_tokens_seen": 5361584, "step": 3955 }, { "epoch": 0.12675244862684848, "grad_norm": 1.572444200515747, "learning_rate": 4.880116648410468e-05, "loss": 0.5748, "num_input_tokens_seen": 5368368, "step": 3960 }, { "epoch": 0.12691248959733692, "grad_norm": 0.7662741541862488, "learning_rate": 4.879816096368939e-05, "loss": 0.4419, "num_input_tokens_seen": 5375136, "step": 3965 }, { "epoch": 0.12707253056782536, "grad_norm": 1.2366459369659424, "learning_rate": 4.879515177329861e-05, "loss": 0.5371, "num_input_tokens_seen": 5382048, "step": 3970 }, { "epoch": 0.1272325715383138, "grad_norm": 1.0474512577056885, "learning_rate": 4.8792138913396394e-05, "loss": 0.6707, "num_input_tokens_seen": 5389056, "step": 3975 }, { "epoch": 0.12739261250880227, "grad_norm": 0.586124837398529, "learning_rate": 4.8789122384447374e-05, "loss": 0.3879, "num_input_tokens_seen": 5395904, "step": 3980 }, { "epoch": 0.1275526534792907, "grad_norm": 1.8838211297988892, "learning_rate": 4.878610218691673e-05, "loss": 0.6477, "num_input_tokens_seen": 5402688, "step": 3985 }, { "epoch": 0.12771269444977915, "grad_norm": 2.274089813232422, "learning_rate": 4.87830783212702e-05, "loss": 0.7873, "num_input_tokens_seen": 5409600, "step": 3990 }, { "epoch": 0.12787273542026759, "grad_norm": 1.040404200553894, "learning_rate": 4.878005078797413e-05, "loss": 0.873, "num_input_tokens_seen": 5416384, "step": 3995 }, { "epoch": 0.12803277639075603, "grad_norm": 1.06839919090271, "learning_rate": 4.877701958749539e-05, "loss": 0.512, "num_input_tokens_seen": 5422944, "step": 4000 }, { "epoch": 0.12803277639075603, "eval_loss": 0.5887617468833923, "eval_runtime": 500.2429, "eval_samples_per_second": 27.759, "eval_steps_per_second": 13.879, "num_input_tokens_seen": 5422944, "step": 4000 }, { "epoch": 0.12819281736124447, "grad_norm": 0.6450153589248657, "learning_rate": 4.877398472030142e-05, "loss": 0.4096, "num_input_tokens_seen": 5429664, "step": 4005 }, { "epoch": 0.12835285833173293, "grad_norm": 1.3732224702835083, "learning_rate": 4.877094618686024e-05, "loss": 0.5455, "num_input_tokens_seen": 5436208, "step": 4010 }, { "epoch": 0.12851289930222137, "grad_norm": 1.7121384143829346, "learning_rate": 4.876790398764045e-05, "loss": 0.7999, "num_input_tokens_seen": 5442704, "step": 4015 }, { "epoch": 0.1286729402727098, "grad_norm": 1.5706578493118286, "learning_rate": 4.8764858123111167e-05, "loss": 0.5683, "num_input_tokens_seen": 5449808, "step": 4020 }, { "epoch": 0.12883298124319825, "grad_norm": 1.7975893020629883, "learning_rate": 4.876180859374212e-05, "loss": 0.5063, "num_input_tokens_seen": 5456288, "step": 4025 }, { "epoch": 0.1289930222136867, "grad_norm": 1.3366271257400513, "learning_rate": 4.875875540000357e-05, "loss": 0.5918, "num_input_tokens_seen": 5463120, "step": 4030 }, { "epoch": 0.12915306318417516, "grad_norm": 1.005702018737793, "learning_rate": 4.8755698542366376e-05, "loss": 0.5364, "num_input_tokens_seen": 5469552, "step": 4035 }, { "epoch": 0.1293131041546636, "grad_norm": 2.743002414703369, "learning_rate": 4.875263802130193e-05, "loss": 0.9455, "num_input_tokens_seen": 5475840, "step": 4040 }, { "epoch": 0.12947314512515204, "grad_norm": 1.3250582218170166, "learning_rate": 4.8749573837282207e-05, "loss": 0.4777, "num_input_tokens_seen": 5482384, "step": 4045 }, { "epoch": 0.12963318609564048, "grad_norm": 1.215370774269104, "learning_rate": 4.874650599077974e-05, "loss": 0.5173, "num_input_tokens_seen": 5488544, "step": 4050 }, { "epoch": 0.12979322706612892, "grad_norm": 0.5121477842330933, "learning_rate": 4.874343448226764e-05, "loss": 0.4328, "num_input_tokens_seen": 5495344, "step": 4055 }, { "epoch": 0.1299532680366174, "grad_norm": 2.6686174869537354, "learning_rate": 4.874035931221955e-05, "loss": 0.6213, "num_input_tokens_seen": 5501824, "step": 4060 }, { "epoch": 0.13011330900710583, "grad_norm": 1.4531075954437256, "learning_rate": 4.8737280481109724e-05, "loss": 0.4798, "num_input_tokens_seen": 5508928, "step": 4065 }, { "epoch": 0.13027334997759427, "grad_norm": 1.2262979745864868, "learning_rate": 4.873419798941294e-05, "loss": 0.5813, "num_input_tokens_seen": 5515440, "step": 4070 }, { "epoch": 0.1304333909480827, "grad_norm": 1.0218592882156372, "learning_rate": 4.873111183760458e-05, "loss": 0.5656, "num_input_tokens_seen": 5522448, "step": 4075 }, { "epoch": 0.13059343191857115, "grad_norm": 1.3307251930236816, "learning_rate": 4.8728022026160537e-05, "loss": 0.537, "num_input_tokens_seen": 5528992, "step": 4080 }, { "epoch": 0.13075347288905959, "grad_norm": 0.8362858891487122, "learning_rate": 4.872492855555732e-05, "loss": 0.5562, "num_input_tokens_seen": 5535520, "step": 4085 }, { "epoch": 0.13091351385954805, "grad_norm": 1.4478092193603516, "learning_rate": 4.8721831426271956e-05, "loss": 0.4542, "num_input_tokens_seen": 5542224, "step": 4090 }, { "epoch": 0.1310735548300365, "grad_norm": 1.5214439630508423, "learning_rate": 4.87187306387821e-05, "loss": 0.4595, "num_input_tokens_seen": 5549024, "step": 4095 }, { "epoch": 0.13123359580052493, "grad_norm": 1.1281710863113403, "learning_rate": 4.87156261935659e-05, "loss": 0.6717, "num_input_tokens_seen": 5556032, "step": 4100 }, { "epoch": 0.13139363677101337, "grad_norm": 1.196312427520752, "learning_rate": 4.871251809110211e-05, "loss": 0.4739, "num_input_tokens_seen": 5562416, "step": 4105 }, { "epoch": 0.1315536777415018, "grad_norm": 1.1967214345932007, "learning_rate": 4.8709406331870044e-05, "loss": 0.6586, "num_input_tokens_seen": 5569488, "step": 4110 }, { "epoch": 0.13171371871199028, "grad_norm": 2.2152271270751953, "learning_rate": 4.8706290916349574e-05, "loss": 0.7484, "num_input_tokens_seen": 5575936, "step": 4115 }, { "epoch": 0.13187375968247872, "grad_norm": 1.0064342021942139, "learning_rate": 4.8703171845021134e-05, "loss": 0.569, "num_input_tokens_seen": 5582496, "step": 4120 }, { "epoch": 0.13203380065296716, "grad_norm": 0.9726685285568237, "learning_rate": 4.870004911836572e-05, "loss": 0.6867, "num_input_tokens_seen": 5589536, "step": 4125 }, { "epoch": 0.1321938416234556, "grad_norm": 0.6890978217124939, "learning_rate": 4.869692273686489e-05, "loss": 0.5863, "num_input_tokens_seen": 5596304, "step": 4130 }, { "epoch": 0.13235388259394404, "grad_norm": 0.8063042163848877, "learning_rate": 4.869379270100079e-05, "loss": 0.5097, "num_input_tokens_seen": 5603488, "step": 4135 }, { "epoch": 0.1325139235644325, "grad_norm": 1.173121452331543, "learning_rate": 4.86906590112561e-05, "loss": 0.5907, "num_input_tokens_seen": 5610304, "step": 4140 }, { "epoch": 0.13267396453492095, "grad_norm": 3.2397730350494385, "learning_rate": 4.8687521668114064e-05, "loss": 0.7374, "num_input_tokens_seen": 5617136, "step": 4145 }, { "epoch": 0.13283400550540939, "grad_norm": 0.5447896718978882, "learning_rate": 4.868438067205853e-05, "loss": 0.6597, "num_input_tokens_seen": 5623808, "step": 4150 }, { "epoch": 0.13299404647589783, "grad_norm": 1.8226162195205688, "learning_rate": 4.8681236023573844e-05, "loss": 0.5831, "num_input_tokens_seen": 5630672, "step": 4155 }, { "epoch": 0.13315408744638627, "grad_norm": 1.6576893329620361, "learning_rate": 4.867808772314497e-05, "loss": 0.5095, "num_input_tokens_seen": 5637152, "step": 4160 }, { "epoch": 0.13331412841687473, "grad_norm": 0.7029037475585938, "learning_rate": 4.867493577125741e-05, "loss": 0.4433, "num_input_tokens_seen": 5643552, "step": 4165 }, { "epoch": 0.13347416938736317, "grad_norm": 1.0722373723983765, "learning_rate": 4.867178016839725e-05, "loss": 0.4454, "num_input_tokens_seen": 5650512, "step": 4170 }, { "epoch": 0.1336342103578516, "grad_norm": 1.0535943508148193, "learning_rate": 4.8668620915051094e-05, "loss": 0.3421, "num_input_tokens_seen": 5657040, "step": 4175 }, { "epoch": 0.13379425132834005, "grad_norm": 0.9176779985427856, "learning_rate": 4.866545801170616e-05, "loss": 0.5705, "num_input_tokens_seen": 5663744, "step": 4180 }, { "epoch": 0.1339542922988285, "grad_norm": 1.1852340698242188, "learning_rate": 4.86622914588502e-05, "loss": 0.5924, "num_input_tokens_seen": 5671200, "step": 4185 }, { "epoch": 0.13411433326931693, "grad_norm": 1.9435229301452637, "learning_rate": 4.865912125697154e-05, "loss": 0.6114, "num_input_tokens_seen": 5678112, "step": 4190 }, { "epoch": 0.1342743742398054, "grad_norm": 1.9229742288589478, "learning_rate": 4.865594740655907e-05, "loss": 0.5706, "num_input_tokens_seen": 5685136, "step": 4195 }, { "epoch": 0.13443441521029384, "grad_norm": 1.5711063146591187, "learning_rate": 4.865276990810222e-05, "loss": 0.6774, "num_input_tokens_seen": 5692368, "step": 4200 }, { "epoch": 0.13443441521029384, "eval_loss": 0.5858265161514282, "eval_runtime": 500.0858, "eval_samples_per_second": 27.767, "eval_steps_per_second": 13.884, "num_input_tokens_seen": 5692368, "step": 4200 }, { "epoch": 0.13459445618078228, "grad_norm": 1.3223601579666138, "learning_rate": 4.8649588762091016e-05, "loss": 0.5255, "num_input_tokens_seen": 5698944, "step": 4205 }, { "epoch": 0.13475449715127072, "grad_norm": 1.6618058681488037, "learning_rate": 4.8646403969016016e-05, "loss": 0.6925, "num_input_tokens_seen": 5705952, "step": 4210 }, { "epoch": 0.13491453812175916, "grad_norm": 2.1845693588256836, "learning_rate": 4.864321552936838e-05, "loss": 0.5628, "num_input_tokens_seen": 5712464, "step": 4215 }, { "epoch": 0.13507457909224763, "grad_norm": 1.2678258419036865, "learning_rate": 4.864002344363978e-05, "loss": 0.6696, "num_input_tokens_seen": 5719392, "step": 4220 }, { "epoch": 0.13523462006273607, "grad_norm": 0.8848970532417297, "learning_rate": 4.863682771232248e-05, "loss": 0.5679, "num_input_tokens_seen": 5726304, "step": 4225 }, { "epoch": 0.1353946610332245, "grad_norm": 1.5904029607772827, "learning_rate": 4.8633628335909324e-05, "loss": 0.4903, "num_input_tokens_seen": 5733360, "step": 4230 }, { "epoch": 0.13555470200371295, "grad_norm": 1.3585723638534546, "learning_rate": 4.8630425314893676e-05, "loss": 0.5243, "num_input_tokens_seen": 5740432, "step": 4235 }, { "epoch": 0.13571474297420139, "grad_norm": 0.901386022567749, "learning_rate": 4.862721864976948e-05, "loss": 0.5175, "num_input_tokens_seen": 5747280, "step": 4240 }, { "epoch": 0.13587478394468985, "grad_norm": 1.069374680519104, "learning_rate": 4.862400834103125e-05, "loss": 0.4901, "num_input_tokens_seen": 5754176, "step": 4245 }, { "epoch": 0.1360348249151783, "grad_norm": 1.6738476753234863, "learning_rate": 4.862079438917406e-05, "loss": 0.5653, "num_input_tokens_seen": 5761008, "step": 4250 }, { "epoch": 0.13619486588566673, "grad_norm": 0.6785828471183777, "learning_rate": 4.8617576794693536e-05, "loss": 0.5622, "num_input_tokens_seen": 5767584, "step": 4255 }, { "epoch": 0.13635490685615517, "grad_norm": 1.3693795204162598, "learning_rate": 4.8614355558085875e-05, "loss": 0.4522, "num_input_tokens_seen": 5775072, "step": 4260 }, { "epoch": 0.1365149478266436, "grad_norm": 1.05167555809021, "learning_rate": 4.861113067984783e-05, "loss": 0.375, "num_input_tokens_seen": 5781792, "step": 4265 }, { "epoch": 0.13667498879713205, "grad_norm": 0.7470888495445251, "learning_rate": 4.860790216047671e-05, "loss": 0.477, "num_input_tokens_seen": 5788176, "step": 4270 }, { "epoch": 0.13683502976762052, "grad_norm": 1.415254831314087, "learning_rate": 4.860467000047041e-05, "loss": 0.5343, "num_input_tokens_seen": 5794608, "step": 4275 }, { "epoch": 0.13699507073810896, "grad_norm": 1.3446568250656128, "learning_rate": 4.860143420032737e-05, "loss": 0.4255, "num_input_tokens_seen": 5801584, "step": 4280 }, { "epoch": 0.1371551117085974, "grad_norm": 1.9262135028839111, "learning_rate": 4.859819476054657e-05, "loss": 0.511, "num_input_tokens_seen": 5808144, "step": 4285 }, { "epoch": 0.13731515267908584, "grad_norm": 1.0160304307937622, "learning_rate": 4.859495168162758e-05, "loss": 0.427, "num_input_tokens_seen": 5815072, "step": 4290 }, { "epoch": 0.13747519364957428, "grad_norm": 1.1060086488723755, "learning_rate": 4.859170496407054e-05, "loss": 0.8422, "num_input_tokens_seen": 5822000, "step": 4295 }, { "epoch": 0.13763523462006275, "grad_norm": 1.6212292909622192, "learning_rate": 4.8588454608376114e-05, "loss": 0.5857, "num_input_tokens_seen": 5828464, "step": 4300 }, { "epoch": 0.1377952755905512, "grad_norm": 2.202256202697754, "learning_rate": 4.8585200615045555e-05, "loss": 0.6604, "num_input_tokens_seen": 5835376, "step": 4305 }, { "epoch": 0.13795531656103963, "grad_norm": 1.297555923461914, "learning_rate": 4.8581942984580674e-05, "loss": 0.4613, "num_input_tokens_seen": 5842320, "step": 4310 }, { "epoch": 0.13811535753152807, "grad_norm": 1.5790575742721558, "learning_rate": 4.857868171748384e-05, "loss": 0.6223, "num_input_tokens_seen": 5849168, "step": 4315 }, { "epoch": 0.1382753985020165, "grad_norm": 1.2954707145690918, "learning_rate": 4.8575416814257976e-05, "loss": 0.5866, "num_input_tokens_seen": 5856112, "step": 4320 }, { "epoch": 0.13843543947250497, "grad_norm": 1.3154687881469727, "learning_rate": 4.857214827540657e-05, "loss": 0.5809, "num_input_tokens_seen": 5863424, "step": 4325 }, { "epoch": 0.1385954804429934, "grad_norm": 1.8123202323913574, "learning_rate": 4.856887610143367e-05, "loss": 0.7442, "num_input_tokens_seen": 5870592, "step": 4330 }, { "epoch": 0.13875552141348185, "grad_norm": 1.3981682062149048, "learning_rate": 4.8565600292843896e-05, "loss": 0.4744, "num_input_tokens_seen": 5877088, "step": 4335 }, { "epoch": 0.1389155623839703, "grad_norm": 1.3098477125167847, "learning_rate": 4.856232085014241e-05, "loss": 0.6793, "num_input_tokens_seen": 5884112, "step": 4340 }, { "epoch": 0.13907560335445873, "grad_norm": 2.1962366104125977, "learning_rate": 4.855903777383495e-05, "loss": 0.54, "num_input_tokens_seen": 5890640, "step": 4345 }, { "epoch": 0.1392356443249472, "grad_norm": 1.1837531328201294, "learning_rate": 4.85557510644278e-05, "loss": 0.4998, "num_input_tokens_seen": 5897312, "step": 4350 }, { "epoch": 0.13939568529543564, "grad_norm": 1.5166264772415161, "learning_rate": 4.855246072242782e-05, "loss": 0.7438, "num_input_tokens_seen": 5904320, "step": 4355 }, { "epoch": 0.13955572626592408, "grad_norm": 1.35394287109375, "learning_rate": 4.8549166748342414e-05, "loss": 0.446, "num_input_tokens_seen": 5911072, "step": 4360 }, { "epoch": 0.13971576723641252, "grad_norm": 0.9766973257064819, "learning_rate": 4.8545869142679556e-05, "loss": 0.4362, "num_input_tokens_seen": 5917984, "step": 4365 }, { "epoch": 0.13987580820690096, "grad_norm": 0.8325091600418091, "learning_rate": 4.8542567905947776e-05, "loss": 0.5902, "num_input_tokens_seen": 5925872, "step": 4370 }, { "epoch": 0.1400358491773894, "grad_norm": 1.8909403085708618, "learning_rate": 4.853926303865618e-05, "loss": 0.5493, "num_input_tokens_seen": 5932544, "step": 4375 }, { "epoch": 0.14019589014787787, "grad_norm": 2.80780291557312, "learning_rate": 4.853595454131441e-05, "loss": 0.6738, "num_input_tokens_seen": 5939008, "step": 4380 }, { "epoch": 0.1403559311183663, "grad_norm": 1.4202829599380493, "learning_rate": 4.8532642414432674e-05, "loss": 0.6584, "num_input_tokens_seen": 5945824, "step": 4385 }, { "epoch": 0.14051597208885475, "grad_norm": 1.6724512577056885, "learning_rate": 4.8529326658521754e-05, "loss": 0.5328, "num_input_tokens_seen": 5952384, "step": 4390 }, { "epoch": 0.14067601305934319, "grad_norm": 0.9704434275627136, "learning_rate": 4.8526007274092965e-05, "loss": 0.4638, "num_input_tokens_seen": 5958864, "step": 4395 }, { "epoch": 0.14083605402983163, "grad_norm": 0.8374595046043396, "learning_rate": 4.852268426165822e-05, "loss": 0.6641, "num_input_tokens_seen": 5965440, "step": 4400 }, { "epoch": 0.14083605402983163, "eval_loss": 0.5806149840354919, "eval_runtime": 499.6315, "eval_samples_per_second": 27.792, "eval_steps_per_second": 13.896, "num_input_tokens_seen": 5965440, "step": 4400 }, { "epoch": 0.1409960950003201, "grad_norm": 1.2809666395187378, "learning_rate": 4.851935762172995e-05, "loss": 0.3489, "num_input_tokens_seen": 5972464, "step": 4405 }, { "epoch": 0.14115613597080853, "grad_norm": 1.0373201370239258, "learning_rate": 4.8516027354821175e-05, "loss": 0.5907, "num_input_tokens_seen": 5979344, "step": 4410 }, { "epoch": 0.14131617694129697, "grad_norm": 1.9336143732070923, "learning_rate": 4.851269346144546e-05, "loss": 0.6362, "num_input_tokens_seen": 5986384, "step": 4415 }, { "epoch": 0.1414762179117854, "grad_norm": 1.3165044784545898, "learning_rate": 4.850935594211693e-05, "loss": 0.6064, "num_input_tokens_seen": 5993840, "step": 4420 }, { "epoch": 0.14163625888227385, "grad_norm": 1.0207160711288452, "learning_rate": 4.850601479735029e-05, "loss": 0.5169, "num_input_tokens_seen": 6000672, "step": 4425 }, { "epoch": 0.14179629985276232, "grad_norm": 1.2229961156845093, "learning_rate": 4.850267002766076e-05, "loss": 0.5579, "num_input_tokens_seen": 6007408, "step": 4430 }, { "epoch": 0.14195634082325076, "grad_norm": 2.0587055683135986, "learning_rate": 4.849932163356417e-05, "loss": 0.502, "num_input_tokens_seen": 6014096, "step": 4435 }, { "epoch": 0.1421163817937392, "grad_norm": 1.2133523225784302, "learning_rate": 4.8495969615576864e-05, "loss": 0.5396, "num_input_tokens_seen": 6020592, "step": 4440 }, { "epoch": 0.14227642276422764, "grad_norm": 1.950038194656372, "learning_rate": 4.849261397421577e-05, "loss": 0.6976, "num_input_tokens_seen": 6026976, "step": 4445 }, { "epoch": 0.14243646373471608, "grad_norm": 1.7990703582763672, "learning_rate": 4.848925470999839e-05, "loss": 0.5986, "num_input_tokens_seen": 6033648, "step": 4450 }, { "epoch": 0.14259650470520452, "grad_norm": 1.5461764335632324, "learning_rate": 4.848589182344273e-05, "loss": 0.5518, "num_input_tokens_seen": 6040384, "step": 4455 }, { "epoch": 0.142756545675693, "grad_norm": 1.1024677753448486, "learning_rate": 4.848252531506742e-05, "loss": 0.7002, "num_input_tokens_seen": 6047440, "step": 4460 }, { "epoch": 0.14291658664618143, "grad_norm": 1.3455286026000977, "learning_rate": 4.847915518539161e-05, "loss": 0.6151, "num_input_tokens_seen": 6054400, "step": 4465 }, { "epoch": 0.14307662761666987, "grad_norm": 1.5695325136184692, "learning_rate": 4.847578143493501e-05, "loss": 0.4855, "num_input_tokens_seen": 6061264, "step": 4470 }, { "epoch": 0.1432366685871583, "grad_norm": 1.4452133178710938, "learning_rate": 4.847240406421789e-05, "loss": 0.3849, "num_input_tokens_seen": 6067888, "step": 4475 }, { "epoch": 0.14339670955764675, "grad_norm": 2.252708911895752, "learning_rate": 4.84690230737611e-05, "loss": 0.6303, "num_input_tokens_seen": 6074640, "step": 4480 }, { "epoch": 0.1435567505281352, "grad_norm": 1.2751214504241943, "learning_rate": 4.846563846408602e-05, "loss": 0.4647, "num_input_tokens_seen": 6080992, "step": 4485 }, { "epoch": 0.14371679149862365, "grad_norm": 0.8906923532485962, "learning_rate": 4.84622502357146e-05, "loss": 0.7688, "num_input_tokens_seen": 6088160, "step": 4490 }, { "epoch": 0.1438768324691121, "grad_norm": 2.9467241764068604, "learning_rate": 4.8458858389169345e-05, "loss": 0.708, "num_input_tokens_seen": 6095008, "step": 4495 }, { "epoch": 0.14403687343960053, "grad_norm": 1.3613200187683105, "learning_rate": 4.8455462924973334e-05, "loss": 0.5046, "num_input_tokens_seen": 6102608, "step": 4500 }, { "epoch": 0.14419691441008897, "grad_norm": 1.8200634717941284, "learning_rate": 4.845206384365018e-05, "loss": 0.528, "num_input_tokens_seen": 6109216, "step": 4505 }, { "epoch": 0.14435695538057744, "grad_norm": 1.6324577331542969, "learning_rate": 4.844866114572405e-05, "loss": 0.6463, "num_input_tokens_seen": 6115760, "step": 4510 }, { "epoch": 0.14451699635106588, "grad_norm": 1.8181596994400024, "learning_rate": 4.8445254831719706e-05, "loss": 0.5243, "num_input_tokens_seen": 6122352, "step": 4515 }, { "epoch": 0.14467703732155432, "grad_norm": 1.1740621328353882, "learning_rate": 4.8441844902162434e-05, "loss": 0.5829, "num_input_tokens_seen": 6129024, "step": 4520 }, { "epoch": 0.14483707829204276, "grad_norm": 1.0331931114196777, "learning_rate": 4.843843135757809e-05, "loss": 0.5113, "num_input_tokens_seen": 6135952, "step": 4525 }, { "epoch": 0.1449971192625312, "grad_norm": 0.8117127418518066, "learning_rate": 4.843501419849308e-05, "loss": 0.5846, "num_input_tokens_seen": 6142656, "step": 4530 }, { "epoch": 0.14515716023301967, "grad_norm": 1.7335504293441772, "learning_rate": 4.8431593425434386e-05, "loss": 0.4893, "num_input_tokens_seen": 6149904, "step": 4535 }, { "epoch": 0.1453172012035081, "grad_norm": 1.6174001693725586, "learning_rate": 4.8428169038929526e-05, "loss": 0.5391, "num_input_tokens_seen": 6156672, "step": 4540 }, { "epoch": 0.14547724217399655, "grad_norm": 0.7672015428543091, "learning_rate": 4.8424741039506575e-05, "loss": 0.4272, "num_input_tokens_seen": 6163296, "step": 4545 }, { "epoch": 0.14563728314448499, "grad_norm": 0.9390732645988464, "learning_rate": 4.842130942769419e-05, "loss": 0.6152, "num_input_tokens_seen": 6169952, "step": 4550 }, { "epoch": 0.14579732411497343, "grad_norm": 1.1791462898254395, "learning_rate": 4.841787420402156e-05, "loss": 0.635, "num_input_tokens_seen": 6176512, "step": 4555 }, { "epoch": 0.14595736508546187, "grad_norm": 0.49255234003067017, "learning_rate": 4.841443536901844e-05, "loss": 0.6506, "num_input_tokens_seen": 6183248, "step": 4560 }, { "epoch": 0.14611740605595033, "grad_norm": 0.9091460108757019, "learning_rate": 4.841099292321514e-05, "loss": 0.4478, "num_input_tokens_seen": 6189872, "step": 4565 }, { "epoch": 0.14627744702643877, "grad_norm": 0.7146796584129333, "learning_rate": 4.8407546867142525e-05, "loss": 0.5189, "num_input_tokens_seen": 6196512, "step": 4570 }, { "epoch": 0.1464374879969272, "grad_norm": 1.3036879301071167, "learning_rate": 4.840409720133203e-05, "loss": 0.5569, "num_input_tokens_seen": 6203424, "step": 4575 }, { "epoch": 0.14659752896741565, "grad_norm": 1.2865960597991943, "learning_rate": 4.8400643926315634e-05, "loss": 0.4729, "num_input_tokens_seen": 6210320, "step": 4580 }, { "epoch": 0.1467575699379041, "grad_norm": 1.0012855529785156, "learning_rate": 4.839718704262587e-05, "loss": 0.4482, "num_input_tokens_seen": 6217216, "step": 4585 }, { "epoch": 0.14691761090839256, "grad_norm": 1.4223068952560425, "learning_rate": 4.839372655079585e-05, "loss": 0.4229, "num_input_tokens_seen": 6223952, "step": 4590 }, { "epoch": 0.147077651878881, "grad_norm": 1.4190385341644287, "learning_rate": 4.83902624513592e-05, "loss": 0.5288, "num_input_tokens_seen": 6230544, "step": 4595 }, { "epoch": 0.14723769284936944, "grad_norm": 0.8154120445251465, "learning_rate": 4.838679474485014e-05, "loss": 0.3172, "num_input_tokens_seen": 6237632, "step": 4600 }, { "epoch": 0.14723769284936944, "eval_loss": 0.5787048935890198, "eval_runtime": 499.8125, "eval_samples_per_second": 27.782, "eval_steps_per_second": 13.891, "num_input_tokens_seen": 6237632, "step": 4600 }, { "epoch": 0.14739773381985788, "grad_norm": 1.051804542541504, "learning_rate": 4.838332343180343e-05, "loss": 0.568, "num_input_tokens_seen": 6244192, "step": 4605 }, { "epoch": 0.14755777479034632, "grad_norm": 0.9206752777099609, "learning_rate": 4.83798485127544e-05, "loss": 0.5504, "num_input_tokens_seen": 6251280, "step": 4610 }, { "epoch": 0.1477178157608348, "grad_norm": 1.381803274154663, "learning_rate": 4.837636998823892e-05, "loss": 0.6341, "num_input_tokens_seen": 6257680, "step": 4615 }, { "epoch": 0.14787785673132323, "grad_norm": 1.1477891206741333, "learning_rate": 4.8372887858793414e-05, "loss": 0.4944, "num_input_tokens_seen": 6264480, "step": 4620 }, { "epoch": 0.14803789770181167, "grad_norm": 0.9135470986366272, "learning_rate": 4.836940212495489e-05, "loss": 0.4456, "num_input_tokens_seen": 6271296, "step": 4625 }, { "epoch": 0.1481979386723001, "grad_norm": 0.9701776504516602, "learning_rate": 4.836591278726087e-05, "loss": 0.5319, "num_input_tokens_seen": 6278528, "step": 4630 }, { "epoch": 0.14835797964278855, "grad_norm": 1.3073416948318481, "learning_rate": 4.836241984624947e-05, "loss": 0.8599, "num_input_tokens_seen": 6284928, "step": 4635 }, { "epoch": 0.14851802061327699, "grad_norm": 2.2973814010620117, "learning_rate": 4.8358923302459336e-05, "loss": 0.4392, "num_input_tokens_seen": 6291488, "step": 4640 }, { "epoch": 0.14867806158376545, "grad_norm": 1.5526819229125977, "learning_rate": 4.835542315642968e-05, "loss": 0.6214, "num_input_tokens_seen": 6298368, "step": 4645 }, { "epoch": 0.1488381025542539, "grad_norm": 1.4363521337509155, "learning_rate": 4.8351919408700274e-05, "loss": 0.4833, "num_input_tokens_seen": 6304976, "step": 4650 }, { "epoch": 0.14899814352474233, "grad_norm": 0.8393040299415588, "learning_rate": 4.834841205981144e-05, "loss": 0.5489, "num_input_tokens_seen": 6311376, "step": 4655 }, { "epoch": 0.14915818449523077, "grad_norm": 1.2044909000396729, "learning_rate": 4.8344901110304054e-05, "loss": 0.5621, "num_input_tokens_seen": 6318608, "step": 4660 }, { "epoch": 0.1493182254657192, "grad_norm": 0.9125856757164001, "learning_rate": 4.8341386560719534e-05, "loss": 0.5156, "num_input_tokens_seen": 6325008, "step": 4665 }, { "epoch": 0.14947826643620768, "grad_norm": 1.4286174774169922, "learning_rate": 4.833786841159989e-05, "loss": 0.7111, "num_input_tokens_seen": 6331568, "step": 4670 }, { "epoch": 0.14963830740669612, "grad_norm": 0.6347394585609436, "learning_rate": 4.833434666348765e-05, "loss": 0.4587, "num_input_tokens_seen": 6338448, "step": 4675 }, { "epoch": 0.14979834837718456, "grad_norm": 1.2453171014785767, "learning_rate": 4.833082131692592e-05, "loss": 0.5692, "num_input_tokens_seen": 6344976, "step": 4680 }, { "epoch": 0.149958389347673, "grad_norm": 0.9095887541770935, "learning_rate": 4.832729237245835e-05, "loss": 0.7272, "num_input_tokens_seen": 6351360, "step": 4685 }, { "epoch": 0.15011843031816144, "grad_norm": 1.564306616783142, "learning_rate": 4.8323759830629145e-05, "loss": 0.5679, "num_input_tokens_seen": 6357952, "step": 4690 }, { "epoch": 0.1502784712886499, "grad_norm": 1.2256649732589722, "learning_rate": 4.8320223691983066e-05, "loss": 0.5427, "num_input_tokens_seen": 6365088, "step": 4695 }, { "epoch": 0.15043851225913835, "grad_norm": 1.5054031610488892, "learning_rate": 4.831668395706544e-05, "loss": 0.6153, "num_input_tokens_seen": 6371312, "step": 4700 }, { "epoch": 0.15059855322962679, "grad_norm": 1.9634891748428345, "learning_rate": 4.8313140626422125e-05, "loss": 0.6009, "num_input_tokens_seen": 6377840, "step": 4705 }, { "epoch": 0.15075859420011523, "grad_norm": 1.5624769926071167, "learning_rate": 4.830959370059956e-05, "loss": 0.5312, "num_input_tokens_seen": 6384496, "step": 4710 }, { "epoch": 0.15091863517060367, "grad_norm": 1.794328212738037, "learning_rate": 4.830604318014472e-05, "loss": 0.6657, "num_input_tokens_seen": 6391120, "step": 4715 }, { "epoch": 0.15107867614109213, "grad_norm": 1.5560240745544434, "learning_rate": 4.830248906560514e-05, "loss": 0.5294, "num_input_tokens_seen": 6398256, "step": 4720 }, { "epoch": 0.15123871711158057, "grad_norm": 1.2361658811569214, "learning_rate": 4.829893135752891e-05, "loss": 0.4896, "num_input_tokens_seen": 6405552, "step": 4725 }, { "epoch": 0.151398758082069, "grad_norm": 1.2431578636169434, "learning_rate": 4.829537005646466e-05, "loss": 0.481, "num_input_tokens_seen": 6412336, "step": 4730 }, { "epoch": 0.15155879905255745, "grad_norm": 1.1346452236175537, "learning_rate": 4.8291805162961615e-05, "loss": 0.4703, "num_input_tokens_seen": 6419248, "step": 4735 }, { "epoch": 0.1517188400230459, "grad_norm": 1.1485849618911743, "learning_rate": 4.82882366775695e-05, "loss": 0.4048, "num_input_tokens_seen": 6425808, "step": 4740 }, { "epoch": 0.15187888099353433, "grad_norm": 1.2473962306976318, "learning_rate": 4.828466460083864e-05, "loss": 0.5355, "num_input_tokens_seen": 6432432, "step": 4745 }, { "epoch": 0.1520389219640228, "grad_norm": 1.6496169567108154, "learning_rate": 4.8281088933319877e-05, "loss": 0.5882, "num_input_tokens_seen": 6438960, "step": 4750 }, { "epoch": 0.15219896293451124, "grad_norm": 2.09450101852417, "learning_rate": 4.827750967556464e-05, "loss": 0.6934, "num_input_tokens_seen": 6445568, "step": 4755 }, { "epoch": 0.15235900390499968, "grad_norm": 1.563759207725525, "learning_rate": 4.827392682812488e-05, "loss": 0.8473, "num_input_tokens_seen": 6452096, "step": 4760 }, { "epoch": 0.15251904487548812, "grad_norm": 1.1767644882202148, "learning_rate": 4.827034039155312e-05, "loss": 0.4774, "num_input_tokens_seen": 6458848, "step": 4765 }, { "epoch": 0.15267908584597656, "grad_norm": 2.278261661529541, "learning_rate": 4.8266750366402445e-05, "loss": 0.5379, "num_input_tokens_seen": 6465440, "step": 4770 }, { "epoch": 0.15283912681646503, "grad_norm": 1.8104804754257202, "learning_rate": 4.8263156753226476e-05, "loss": 0.661, "num_input_tokens_seen": 6471984, "step": 4775 }, { "epoch": 0.15299916778695347, "grad_norm": 0.7289422750473022, "learning_rate": 4.8259559552579394e-05, "loss": 0.4626, "num_input_tokens_seen": 6478720, "step": 4780 }, { "epoch": 0.1531592087574419, "grad_norm": 0.9393396973609924, "learning_rate": 4.825595876501593e-05, "loss": 0.4447, "num_input_tokens_seen": 6485728, "step": 4785 }, { "epoch": 0.15331924972793035, "grad_norm": 1.0660970211029053, "learning_rate": 4.825235439109137e-05, "loss": 0.4066, "num_input_tokens_seen": 6492528, "step": 4790 }, { "epoch": 0.15347929069841879, "grad_norm": 0.9549815058708191, "learning_rate": 4.824874643136156e-05, "loss": 0.5345, "num_input_tokens_seen": 6499648, "step": 4795 }, { "epoch": 0.15363933166890725, "grad_norm": 0.8878547549247742, "learning_rate": 4.824513488638288e-05, "loss": 0.4636, "num_input_tokens_seen": 6506256, "step": 4800 }, { "epoch": 0.15363933166890725, "eval_loss": 0.5753360986709595, "eval_runtime": 499.2501, "eval_samples_per_second": 27.814, "eval_steps_per_second": 13.907, "num_input_tokens_seen": 6506256, "step": 4800 }, { "epoch": 0.1537993726393957, "grad_norm": 1.6956210136413574, "learning_rate": 4.8241519756712293e-05, "loss": 0.6226, "num_input_tokens_seen": 6513024, "step": 4805 }, { "epoch": 0.15395941360988413, "grad_norm": 1.581808090209961, "learning_rate": 4.8237901042907285e-05, "loss": 0.5779, "num_input_tokens_seen": 6520368, "step": 4810 }, { "epoch": 0.15411945458037257, "grad_norm": 0.7928503155708313, "learning_rate": 4.823427874552591e-05, "loss": 0.4666, "num_input_tokens_seen": 6527136, "step": 4815 }, { "epoch": 0.154279495550861, "grad_norm": 2.195369243621826, "learning_rate": 4.823065286512677e-05, "loss": 0.5998, "num_input_tokens_seen": 6533760, "step": 4820 }, { "epoch": 0.15443953652134945, "grad_norm": 2.1125290393829346, "learning_rate": 4.8227023402269025e-05, "loss": 0.5324, "num_input_tokens_seen": 6540624, "step": 4825 }, { "epoch": 0.15459957749183792, "grad_norm": 1.0186901092529297, "learning_rate": 4.822339035751239e-05, "loss": 0.5599, "num_input_tokens_seen": 6547392, "step": 4830 }, { "epoch": 0.15475961846232636, "grad_norm": 1.9037517309188843, "learning_rate": 4.8219753731417104e-05, "loss": 0.5202, "num_input_tokens_seen": 6554192, "step": 4835 }, { "epoch": 0.1549196594328148, "grad_norm": 1.1341607570648193, "learning_rate": 4.821611352454401e-05, "loss": 0.6495, "num_input_tokens_seen": 6561024, "step": 4840 }, { "epoch": 0.15507970040330324, "grad_norm": 1.1982109546661377, "learning_rate": 4.8212469737454444e-05, "loss": 0.6223, "num_input_tokens_seen": 6567792, "step": 4845 }, { "epoch": 0.15523974137379168, "grad_norm": 1.6425983905792236, "learning_rate": 4.820882237071035e-05, "loss": 0.5986, "num_input_tokens_seen": 6574400, "step": 4850 }, { "epoch": 0.15539978234428015, "grad_norm": 2.5139899253845215, "learning_rate": 4.820517142487417e-05, "loss": 0.6854, "num_input_tokens_seen": 6581408, "step": 4855 }, { "epoch": 0.15555982331476859, "grad_norm": 2.2904186248779297, "learning_rate": 4.8201516900508956e-05, "loss": 0.5513, "num_input_tokens_seen": 6587904, "step": 4860 }, { "epoch": 0.15571986428525703, "grad_norm": 0.6361176371574402, "learning_rate": 4.819785879817827e-05, "loss": 0.5891, "num_input_tokens_seen": 6594560, "step": 4865 }, { "epoch": 0.15587990525574547, "grad_norm": 1.344751238822937, "learning_rate": 4.8194197118446226e-05, "loss": 0.4869, "num_input_tokens_seen": 6601312, "step": 4870 }, { "epoch": 0.1560399462262339, "grad_norm": 1.4495054483413696, "learning_rate": 4.819053186187752e-05, "loss": 0.5998, "num_input_tokens_seen": 6607872, "step": 4875 }, { "epoch": 0.15619998719672237, "grad_norm": 1.1042215824127197, "learning_rate": 4.818686302903736e-05, "loss": 0.6517, "num_input_tokens_seen": 6614352, "step": 4880 }, { "epoch": 0.1563600281672108, "grad_norm": 1.1449968814849854, "learning_rate": 4.818319062049154e-05, "loss": 0.3844, "num_input_tokens_seen": 6621424, "step": 4885 }, { "epoch": 0.15652006913769925, "grad_norm": 1.4311776161193848, "learning_rate": 4.817951463680639e-05, "loss": 0.6746, "num_input_tokens_seen": 6629040, "step": 4890 }, { "epoch": 0.1566801101081877, "grad_norm": 1.7795534133911133, "learning_rate": 4.817583507854879e-05, "loss": 0.4226, "num_input_tokens_seen": 6635776, "step": 4895 }, { "epoch": 0.15684015107867613, "grad_norm": 1.152775764465332, "learning_rate": 4.817215194628617e-05, "loss": 0.649, "num_input_tokens_seen": 6643008, "step": 4900 }, { "epoch": 0.1570001920491646, "grad_norm": 5.207508563995361, "learning_rate": 4.816846524058653e-05, "loss": 0.695, "num_input_tokens_seen": 6649872, "step": 4905 }, { "epoch": 0.15716023301965304, "grad_norm": 3.137828826904297, "learning_rate": 4.816477496201839e-05, "loss": 0.6081, "num_input_tokens_seen": 6656640, "step": 4910 }, { "epoch": 0.15732027399014148, "grad_norm": 1.7264360189437866, "learning_rate": 4.8161081111150845e-05, "loss": 0.9495, "num_input_tokens_seen": 6663264, "step": 4915 }, { "epoch": 0.15748031496062992, "grad_norm": 2.003566265106201, "learning_rate": 4.815738368855354e-05, "loss": 0.657, "num_input_tokens_seen": 6669888, "step": 4920 }, { "epoch": 0.15764035593111836, "grad_norm": 1.0114836692810059, "learning_rate": 4.815368269479664e-05, "loss": 0.4297, "num_input_tokens_seen": 6676672, "step": 4925 }, { "epoch": 0.1578003969016068, "grad_norm": 1.4116028547286987, "learning_rate": 4.814997813045092e-05, "loss": 0.3643, "num_input_tokens_seen": 6683744, "step": 4930 }, { "epoch": 0.15796043787209527, "grad_norm": 0.6076855659484863, "learning_rate": 4.814626999608764e-05, "loss": 0.6089, "num_input_tokens_seen": 6690480, "step": 4935 }, { "epoch": 0.1581204788425837, "grad_norm": 1.3282337188720703, "learning_rate": 4.814255829227865e-05, "loss": 0.6735, "num_input_tokens_seen": 6696880, "step": 4940 }, { "epoch": 0.15828051981307215, "grad_norm": 1.0866166353225708, "learning_rate": 4.813884301959635e-05, "loss": 0.6007, "num_input_tokens_seen": 6704016, "step": 4945 }, { "epoch": 0.15844056078356059, "grad_norm": 1.8759983777999878, "learning_rate": 4.813512417861368e-05, "loss": 0.7381, "num_input_tokens_seen": 6711008, "step": 4950 }, { "epoch": 0.15860060175404903, "grad_norm": 1.8705642223358154, "learning_rate": 4.813140176990411e-05, "loss": 0.4462, "num_input_tokens_seen": 6717776, "step": 4955 }, { "epoch": 0.1587606427245375, "grad_norm": 1.9628822803497314, "learning_rate": 4.8127675794041714e-05, "loss": 0.6473, "num_input_tokens_seen": 6724304, "step": 4960 }, { "epoch": 0.15892068369502593, "grad_norm": 1.8416460752487183, "learning_rate": 4.812394625160107e-05, "loss": 0.5986, "num_input_tokens_seen": 6731424, "step": 4965 }, { "epoch": 0.15908072466551437, "grad_norm": 1.0453847646713257, "learning_rate": 4.812021314315732e-05, "loss": 0.4477, "num_input_tokens_seen": 6738032, "step": 4970 }, { "epoch": 0.1592407656360028, "grad_norm": 0.97403883934021, "learning_rate": 4.811647646928616e-05, "loss": 0.5531, "num_input_tokens_seen": 6745280, "step": 4975 }, { "epoch": 0.15940080660649125, "grad_norm": 0.8723864555358887, "learning_rate": 4.8112736230563814e-05, "loss": 0.4847, "num_input_tokens_seen": 6752272, "step": 4980 }, { "epoch": 0.15956084757697972, "grad_norm": 1.2588845491409302, "learning_rate": 4.81089924275671e-05, "loss": 0.6534, "num_input_tokens_seen": 6758912, "step": 4985 }, { "epoch": 0.15972088854746816, "grad_norm": 1.9051278829574585, "learning_rate": 4.810524506087335e-05, "loss": 0.5072, "num_input_tokens_seen": 6765808, "step": 4990 }, { "epoch": 0.1598809295179566, "grad_norm": 0.9740105271339417, "learning_rate": 4.810149413106044e-05, "loss": 0.2787, "num_input_tokens_seen": 6772880, "step": 4995 }, { "epoch": 0.16004097048844504, "grad_norm": 1.5361086130142212, "learning_rate": 4.809773963870684e-05, "loss": 0.6575, "num_input_tokens_seen": 6779376, "step": 5000 }, { "epoch": 0.16004097048844504, "eval_loss": 0.5703968405723572, "eval_runtime": 499.6056, "eval_samples_per_second": 27.794, "eval_steps_per_second": 13.897, "num_input_tokens_seen": 6779376, "step": 5000 }, { "epoch": 0.16020101145893348, "grad_norm": 1.710780143737793, "learning_rate": 4.809398158439151e-05, "loss": 0.5206, "num_input_tokens_seen": 6786432, "step": 5005 }, { "epoch": 0.16036105242942192, "grad_norm": 1.3448823690414429, "learning_rate": 4.8090219968694005e-05, "loss": 0.5511, "num_input_tokens_seen": 6793456, "step": 5010 }, { "epoch": 0.1605210933999104, "grad_norm": 0.9224622845649719, "learning_rate": 4.808645479219442e-05, "loss": 0.4439, "num_input_tokens_seen": 6800064, "step": 5015 }, { "epoch": 0.16068113437039883, "grad_norm": 0.6730850338935852, "learning_rate": 4.8082686055473375e-05, "loss": 0.4703, "num_input_tokens_seen": 6807728, "step": 5020 }, { "epoch": 0.16084117534088727, "grad_norm": 1.165832281112671, "learning_rate": 4.8078913759112066e-05, "loss": 0.5113, "num_input_tokens_seen": 6814304, "step": 5025 }, { "epoch": 0.1610012163113757, "grad_norm": 1.1174284219741821, "learning_rate": 4.807513790369223e-05, "loss": 0.4716, "num_input_tokens_seen": 6821232, "step": 5030 }, { "epoch": 0.16116125728186415, "grad_norm": 1.222005844116211, "learning_rate": 4.8071358489796145e-05, "loss": 0.4974, "num_input_tokens_seen": 6828048, "step": 5035 }, { "epoch": 0.1613212982523526, "grad_norm": 1.2502039670944214, "learning_rate": 4.806757551800665e-05, "loss": 0.6721, "num_input_tokens_seen": 6834832, "step": 5040 }, { "epoch": 0.16148133922284105, "grad_norm": 1.8408608436584473, "learning_rate": 4.806378898890713e-05, "loss": 0.9216, "num_input_tokens_seen": 6841216, "step": 5045 }, { "epoch": 0.1616413801933295, "grad_norm": 2.225895404815674, "learning_rate": 4.80599989030815e-05, "loss": 0.8894, "num_input_tokens_seen": 6848176, "step": 5050 }, { "epoch": 0.16180142116381793, "grad_norm": 0.7284032702445984, "learning_rate": 4.805620526111426e-05, "loss": 0.6963, "num_input_tokens_seen": 6854944, "step": 5055 }, { "epoch": 0.16196146213430637, "grad_norm": 0.814484179019928, "learning_rate": 4.805240806359042e-05, "loss": 0.5104, "num_input_tokens_seen": 6861808, "step": 5060 }, { "epoch": 0.16212150310479484, "grad_norm": 1.387211799621582, "learning_rate": 4.804860731109557e-05, "loss": 0.5289, "num_input_tokens_seen": 6868416, "step": 5065 }, { "epoch": 0.16228154407528328, "grad_norm": 0.9599493741989136, "learning_rate": 4.804480300421581e-05, "loss": 0.3501, "num_input_tokens_seen": 6875248, "step": 5070 }, { "epoch": 0.16244158504577172, "grad_norm": 1.5098975896835327, "learning_rate": 4.804099514353784e-05, "loss": 0.6765, "num_input_tokens_seen": 6882176, "step": 5075 }, { "epoch": 0.16260162601626016, "grad_norm": 1.1014306545257568, "learning_rate": 4.8037183729648867e-05, "loss": 0.5384, "num_input_tokens_seen": 6889056, "step": 5080 }, { "epoch": 0.1627616669867486, "grad_norm": 1.2958494424819946, "learning_rate": 4.803336876313666e-05, "loss": 0.629, "num_input_tokens_seen": 6895856, "step": 5085 }, { "epoch": 0.16292170795723707, "grad_norm": 0.7723139524459839, "learning_rate": 4.802955024458953e-05, "loss": 0.4158, "num_input_tokens_seen": 6902752, "step": 5090 }, { "epoch": 0.1630817489277255, "grad_norm": 1.625628113746643, "learning_rate": 4.802572817459634e-05, "loss": 0.5027, "num_input_tokens_seen": 6909104, "step": 5095 }, { "epoch": 0.16324178989821395, "grad_norm": 2.289360284805298, "learning_rate": 4.802190255374651e-05, "loss": 0.5053, "num_input_tokens_seen": 6916144, "step": 5100 }, { "epoch": 0.16340183086870239, "grad_norm": 1.2605319023132324, "learning_rate": 4.801807338263e-05, "loss": 0.4151, "num_input_tokens_seen": 6922608, "step": 5105 }, { "epoch": 0.16356187183919083, "grad_norm": 1.0404231548309326, "learning_rate": 4.8014240661837306e-05, "loss": 0.4816, "num_input_tokens_seen": 6929680, "step": 5110 }, { "epoch": 0.16372191280967927, "grad_norm": 1.0379739999771118, "learning_rate": 4.80104043919595e-05, "loss": 0.4457, "num_input_tokens_seen": 6936352, "step": 5115 }, { "epoch": 0.16388195378016773, "grad_norm": 0.7350260615348816, "learning_rate": 4.800656457358815e-05, "loss": 0.5001, "num_input_tokens_seen": 6943072, "step": 5120 }, { "epoch": 0.16404199475065617, "grad_norm": 1.0276367664337158, "learning_rate": 4.800272120731544e-05, "loss": 0.553, "num_input_tokens_seen": 6949504, "step": 5125 }, { "epoch": 0.1642020357211446, "grad_norm": 1.5413873195648193, "learning_rate": 4.799887429373404e-05, "loss": 0.5799, "num_input_tokens_seen": 6956544, "step": 5130 }, { "epoch": 0.16436207669163305, "grad_norm": 0.624157190322876, "learning_rate": 4.79950238334372e-05, "loss": 0.3409, "num_input_tokens_seen": 6963632, "step": 5135 }, { "epoch": 0.1645221176621215, "grad_norm": 1.0230700969696045, "learning_rate": 4.799116982701872e-05, "loss": 0.6162, "num_input_tokens_seen": 6970320, "step": 5140 }, { "epoch": 0.16468215863260996, "grad_norm": 1.690079689025879, "learning_rate": 4.7987312275072926e-05, "loss": 0.8125, "num_input_tokens_seen": 6977040, "step": 5145 }, { "epoch": 0.1648421996030984, "grad_norm": 0.5994982123374939, "learning_rate": 4.79834511781947e-05, "loss": 0.3637, "num_input_tokens_seen": 6984096, "step": 5150 }, { "epoch": 0.16500224057358684, "grad_norm": 1.0959999561309814, "learning_rate": 4.797958653697947e-05, "loss": 0.4064, "num_input_tokens_seen": 6991088, "step": 5155 }, { "epoch": 0.16516228154407528, "grad_norm": 0.7132736444473267, "learning_rate": 4.7975718352023225e-05, "loss": 0.5068, "num_input_tokens_seen": 6997840, "step": 5160 }, { "epoch": 0.16532232251456372, "grad_norm": 0.31892508268356323, "learning_rate": 4.7971846623922476e-05, "loss": 0.4993, "num_input_tokens_seen": 7004880, "step": 5165 }, { "epoch": 0.1654823634850522, "grad_norm": 0.7030327916145325, "learning_rate": 4.7967971353274294e-05, "loss": 0.3934, "num_input_tokens_seen": 7011616, "step": 5170 }, { "epoch": 0.16564240445554063, "grad_norm": 1.165489912033081, "learning_rate": 4.79640925406763e-05, "loss": 0.4393, "num_input_tokens_seen": 7018112, "step": 5175 }, { "epoch": 0.16580244542602907, "grad_norm": 1.8878840208053589, "learning_rate": 4.796021018672664e-05, "loss": 0.6294, "num_input_tokens_seen": 7024528, "step": 5180 }, { "epoch": 0.1659624863965175, "grad_norm": 1.0397015810012817, "learning_rate": 4.795632429202405e-05, "loss": 0.5814, "num_input_tokens_seen": 7031232, "step": 5185 }, { "epoch": 0.16612252736700595, "grad_norm": 1.5864473581314087, "learning_rate": 4.795243485716775e-05, "loss": 0.384, "num_input_tokens_seen": 7037728, "step": 5190 }, { "epoch": 0.16628256833749439, "grad_norm": 2.8212218284606934, "learning_rate": 4.794854188275757e-05, "loss": 0.5831, "num_input_tokens_seen": 7044608, "step": 5195 }, { "epoch": 0.16644260930798285, "grad_norm": 2.1875150203704834, "learning_rate": 4.794464536939384e-05, "loss": 0.5575, "num_input_tokens_seen": 7051504, "step": 5200 }, { "epoch": 0.16644260930798285, "eval_loss": 0.5675193071365356, "eval_runtime": 499.4256, "eval_samples_per_second": 27.804, "eval_steps_per_second": 13.902, "num_input_tokens_seen": 7051504, "step": 5200 }, { "epoch": 0.1666026502784713, "grad_norm": 0.9476208686828613, "learning_rate": 4.794074531767745e-05, "loss": 0.5785, "num_input_tokens_seen": 7058224, "step": 5205 }, { "epoch": 0.16676269124895973, "grad_norm": 0.9203051328659058, "learning_rate": 4.7936841728209834e-05, "loss": 0.6797, "num_input_tokens_seen": 7065088, "step": 5210 }, { "epoch": 0.16692273221944817, "grad_norm": 1.0352853536605835, "learning_rate": 4.7932934601593e-05, "loss": 0.5344, "num_input_tokens_seen": 7071632, "step": 5215 }, { "epoch": 0.1670827731899366, "grad_norm": 1.6758694648742676, "learning_rate": 4.792902393842943e-05, "loss": 0.403, "num_input_tokens_seen": 7078416, "step": 5220 }, { "epoch": 0.16724281416042508, "grad_norm": 1.2555921077728271, "learning_rate": 4.792510973932225e-05, "loss": 0.4754, "num_input_tokens_seen": 7085248, "step": 5225 }, { "epoch": 0.16740285513091352, "grad_norm": 1.7796260118484497, "learning_rate": 4.7921192004875036e-05, "loss": 0.5688, "num_input_tokens_seen": 7092464, "step": 5230 }, { "epoch": 0.16756289610140196, "grad_norm": 0.8062920570373535, "learning_rate": 4.791727073569198e-05, "loss": 0.4025, "num_input_tokens_seen": 7098832, "step": 5235 }, { "epoch": 0.1677229370718904, "grad_norm": 4.188822269439697, "learning_rate": 4.7913345932377775e-05, "loss": 0.6336, "num_input_tokens_seen": 7105904, "step": 5240 }, { "epoch": 0.16788297804237884, "grad_norm": 0.5266505479812622, "learning_rate": 4.790941759553769e-05, "loss": 0.4701, "num_input_tokens_seen": 7112496, "step": 5245 }, { "epoch": 0.1680430190128673, "grad_norm": 0.990487277507782, "learning_rate": 4.79054857257775e-05, "loss": 0.5602, "num_input_tokens_seen": 7118880, "step": 5250 }, { "epoch": 0.16820305998335575, "grad_norm": 2.2860472202301025, "learning_rate": 4.790155032370357e-05, "loss": 0.7342, "num_input_tokens_seen": 7125344, "step": 5255 }, { "epoch": 0.16836310095384419, "grad_norm": 0.9360238909721375, "learning_rate": 4.789761138992278e-05, "loss": 0.4447, "num_input_tokens_seen": 7132128, "step": 5260 }, { "epoch": 0.16852314192433263, "grad_norm": 1.4162949323654175, "learning_rate": 4.7893668925042565e-05, "loss": 0.6214, "num_input_tokens_seen": 7138880, "step": 5265 }, { "epoch": 0.16868318289482107, "grad_norm": 1.0670174360275269, "learning_rate": 4.78897229296709e-05, "loss": 0.8318, "num_input_tokens_seen": 7145632, "step": 5270 }, { "epoch": 0.16884322386530953, "grad_norm": 1.0668786764144897, "learning_rate": 4.7885773404416315e-05, "loss": 0.4241, "num_input_tokens_seen": 7152896, "step": 5275 }, { "epoch": 0.16900326483579797, "grad_norm": 1.6461520195007324, "learning_rate": 4.788182034988786e-05, "loss": 0.6285, "num_input_tokens_seen": 7159920, "step": 5280 }, { "epoch": 0.1691633058062864, "grad_norm": 1.2619601488113403, "learning_rate": 4.787786376669516e-05, "loss": 0.4618, "num_input_tokens_seen": 7166512, "step": 5285 }, { "epoch": 0.16932334677677485, "grad_norm": 1.6620131731033325, "learning_rate": 4.787390365544837e-05, "loss": 0.5635, "num_input_tokens_seen": 7173296, "step": 5290 }, { "epoch": 0.1694833877472633, "grad_norm": 0.9641928672790527, "learning_rate": 4.786994001675818e-05, "loss": 0.5258, "num_input_tokens_seen": 7179824, "step": 5295 }, { "epoch": 0.16964342871775173, "grad_norm": 0.925886332988739, "learning_rate": 4.786597285123584e-05, "loss": 0.5909, "num_input_tokens_seen": 7186720, "step": 5300 }, { "epoch": 0.1698034696882402, "grad_norm": 0.9247031807899475, "learning_rate": 4.7862002159493135e-05, "loss": 0.5248, "num_input_tokens_seen": 7193216, "step": 5305 }, { "epoch": 0.16996351065872864, "grad_norm": 0.9691357016563416, "learning_rate": 4.785802794214239e-05, "loss": 0.4921, "num_input_tokens_seen": 7200240, "step": 5310 }, { "epoch": 0.17012355162921708, "grad_norm": 2.0752017498016357, "learning_rate": 4.7854050199796495e-05, "loss": 0.6066, "num_input_tokens_seen": 7207264, "step": 5315 }, { "epoch": 0.17028359259970552, "grad_norm": 0.9385653138160706, "learning_rate": 4.7850068933068845e-05, "loss": 0.4603, "num_input_tokens_seen": 7213984, "step": 5320 }, { "epoch": 0.17044363357019396, "grad_norm": 1.6507858037948608, "learning_rate": 4.7846084142573425e-05, "loss": 0.6547, "num_input_tokens_seen": 7220592, "step": 5325 }, { "epoch": 0.17060367454068243, "grad_norm": 1.220496654510498, "learning_rate": 4.7842095828924725e-05, "loss": 0.5967, "num_input_tokens_seen": 7227232, "step": 5330 }, { "epoch": 0.17076371551117087, "grad_norm": 1.1794770956039429, "learning_rate": 4.783810399273779e-05, "loss": 0.694, "num_input_tokens_seen": 7234128, "step": 5335 }, { "epoch": 0.1709237564816593, "grad_norm": 1.8519383668899536, "learning_rate": 4.7834108634628226e-05, "loss": 0.6098, "num_input_tokens_seen": 7240656, "step": 5340 }, { "epoch": 0.17108379745214775, "grad_norm": 1.9080554246902466, "learning_rate": 4.783010975521216e-05, "loss": 0.5952, "num_input_tokens_seen": 7247440, "step": 5345 }, { "epoch": 0.17124383842263619, "grad_norm": 1.7409663200378418, "learning_rate": 4.782610735510626e-05, "loss": 0.7948, "num_input_tokens_seen": 7254224, "step": 5350 }, { "epoch": 0.17140387939312465, "grad_norm": 2.5259487628936768, "learning_rate": 4.782210143492776e-05, "loss": 0.5667, "num_input_tokens_seen": 7261136, "step": 5355 }, { "epoch": 0.1715639203636131, "grad_norm": 1.5389524698257446, "learning_rate": 4.781809199529442e-05, "loss": 0.6631, "num_input_tokens_seen": 7267504, "step": 5360 }, { "epoch": 0.17172396133410153, "grad_norm": 1.6900672912597656, "learning_rate": 4.781407903682454e-05, "loss": 0.5468, "num_input_tokens_seen": 7274352, "step": 5365 }, { "epoch": 0.17188400230458997, "grad_norm": 1.446083664894104, "learning_rate": 4.781006256013698e-05, "loss": 0.4725, "num_input_tokens_seen": 7281136, "step": 5370 }, { "epoch": 0.1720440432750784, "grad_norm": 1.4743455648422241, "learning_rate": 4.7806042565851115e-05, "loss": 0.523, "num_input_tokens_seen": 7287680, "step": 5375 }, { "epoch": 0.17220408424556685, "grad_norm": 0.7905374765396118, "learning_rate": 4.7802019054586895e-05, "loss": 0.5473, "num_input_tokens_seen": 7294224, "step": 5380 }, { "epoch": 0.17236412521605532, "grad_norm": 1.5238337516784668, "learning_rate": 4.779799202696479e-05, "loss": 0.3652, "num_input_tokens_seen": 7300848, "step": 5385 }, { "epoch": 0.17252416618654376, "grad_norm": 0.944736659526825, "learning_rate": 4.779396148360581e-05, "loss": 0.6433, "num_input_tokens_seen": 7308080, "step": 5390 }, { "epoch": 0.1726842071570322, "grad_norm": 1.01700758934021, "learning_rate": 4.7789927425131517e-05, "loss": 0.572, "num_input_tokens_seen": 7315152, "step": 5395 }, { "epoch": 0.17284424812752064, "grad_norm": 1.4659349918365479, "learning_rate": 4.778588985216403e-05, "loss": 0.5267, "num_input_tokens_seen": 7321552, "step": 5400 }, { "epoch": 0.17284424812752064, "eval_loss": 0.5632306933403015, "eval_runtime": 499.31, "eval_samples_per_second": 27.81, "eval_steps_per_second": 13.905, "num_input_tokens_seen": 7321552, "step": 5400 }, { "epoch": 0.17300428909800908, "grad_norm": 1.584977388381958, "learning_rate": 4.778184876532598e-05, "loss": 0.5718, "num_input_tokens_seen": 7328288, "step": 5405 }, { "epoch": 0.17316433006849755, "grad_norm": 1.2497507333755493, "learning_rate": 4.7777804165240556e-05, "loss": 0.6289, "num_input_tokens_seen": 7334784, "step": 5410 }, { "epoch": 0.17332437103898599, "grad_norm": 1.027827501296997, "learning_rate": 4.7773756052531485e-05, "loss": 0.6377, "num_input_tokens_seen": 7341904, "step": 5415 }, { "epoch": 0.17348441200947443, "grad_norm": 1.560953974723816, "learning_rate": 4.7769704427823035e-05, "loss": 0.4525, "num_input_tokens_seen": 7348368, "step": 5420 }, { "epoch": 0.17364445297996287, "grad_norm": 0.5367477536201477, "learning_rate": 4.776564929174003e-05, "loss": 0.5656, "num_input_tokens_seen": 7355040, "step": 5425 }, { "epoch": 0.1738044939504513, "grad_norm": 1.7808163166046143, "learning_rate": 4.7761590644907806e-05, "loss": 0.4886, "num_input_tokens_seen": 7361792, "step": 5430 }, { "epoch": 0.17396453492093977, "grad_norm": 1.6531871557235718, "learning_rate": 4.7757528487952263e-05, "loss": 0.5557, "num_input_tokens_seen": 7368720, "step": 5435 }, { "epoch": 0.1741245758914282, "grad_norm": 0.970160186290741, "learning_rate": 4.7753462821499836e-05, "loss": 0.4392, "num_input_tokens_seen": 7375408, "step": 5440 }, { "epoch": 0.17428461686191665, "grad_norm": 0.8460079431533813, "learning_rate": 4.774939364617751e-05, "loss": 0.5897, "num_input_tokens_seen": 7381952, "step": 5445 }, { "epoch": 0.1744446578324051, "grad_norm": 1.5912821292877197, "learning_rate": 4.7745320962612795e-05, "loss": 0.6325, "num_input_tokens_seen": 7388688, "step": 5450 }, { "epoch": 0.17460469880289353, "grad_norm": 1.0543689727783203, "learning_rate": 4.7741244771433756e-05, "loss": 0.5702, "num_input_tokens_seen": 7395392, "step": 5455 }, { "epoch": 0.174764739773382, "grad_norm": 0.9620958566665649, "learning_rate": 4.7737165073268985e-05, "loss": 0.5536, "num_input_tokens_seen": 7402320, "step": 5460 }, { "epoch": 0.17492478074387044, "grad_norm": 2.7592227458953857, "learning_rate": 4.7733081868747626e-05, "loss": 0.3199, "num_input_tokens_seen": 7409040, "step": 5465 }, { "epoch": 0.17508482171435888, "grad_norm": 1.1757193803787231, "learning_rate": 4.772899515849936e-05, "loss": 0.5234, "num_input_tokens_seen": 7415904, "step": 5470 }, { "epoch": 0.17524486268484732, "grad_norm": 1.3724862337112427, "learning_rate": 4.7724904943154414e-05, "loss": 0.5724, "num_input_tokens_seen": 7422768, "step": 5475 }, { "epoch": 0.17540490365533576, "grad_norm": 1.802110195159912, "learning_rate": 4.772081122334354e-05, "loss": 0.5828, "num_input_tokens_seen": 7429712, "step": 5480 }, { "epoch": 0.1755649446258242, "grad_norm": 0.9235439300537109, "learning_rate": 4.771671399969806e-05, "loss": 0.6549, "num_input_tokens_seen": 7436480, "step": 5485 }, { "epoch": 0.17572498559631267, "grad_norm": 0.7320806384086609, "learning_rate": 4.7712613272849794e-05, "loss": 0.4173, "num_input_tokens_seen": 7443648, "step": 5490 }, { "epoch": 0.1758850265668011, "grad_norm": 1.4556978940963745, "learning_rate": 4.770850904343114e-05, "loss": 0.5875, "num_input_tokens_seen": 7450816, "step": 5495 }, { "epoch": 0.17604506753728955, "grad_norm": 1.5207078456878662, "learning_rate": 4.770440131207502e-05, "loss": 0.6092, "num_input_tokens_seen": 7457536, "step": 5500 }, { "epoch": 0.17620510850777799, "grad_norm": 1.1611863374710083, "learning_rate": 4.7700290079414896e-05, "loss": 0.5977, "num_input_tokens_seen": 7464336, "step": 5505 }, { "epoch": 0.17636514947826643, "grad_norm": 1.5573790073394775, "learning_rate": 4.769617534608477e-05, "loss": 0.3927, "num_input_tokens_seen": 7471136, "step": 5510 }, { "epoch": 0.1765251904487549, "grad_norm": 1.4893620014190674, "learning_rate": 4.7692057112719193e-05, "loss": 0.5441, "num_input_tokens_seen": 7478192, "step": 5515 }, { "epoch": 0.17668523141924333, "grad_norm": 1.3339462280273438, "learning_rate": 4.7687935379953234e-05, "loss": 0.4791, "num_input_tokens_seen": 7484752, "step": 5520 }, { "epoch": 0.17684527238973177, "grad_norm": 2.1734931468963623, "learning_rate": 4.7683810148422534e-05, "loss": 0.7634, "num_input_tokens_seen": 7491648, "step": 5525 }, { "epoch": 0.1770053133602202, "grad_norm": 4.4601898193359375, "learning_rate": 4.767968141876324e-05, "loss": 0.6744, "num_input_tokens_seen": 7498496, "step": 5530 }, { "epoch": 0.17716535433070865, "grad_norm": 1.5628461837768555, "learning_rate": 4.767554919161207e-05, "loss": 0.4994, "num_input_tokens_seen": 7504992, "step": 5535 }, { "epoch": 0.17732539530119712, "grad_norm": 1.6502095460891724, "learning_rate": 4.767141346760624e-05, "loss": 0.3609, "num_input_tokens_seen": 7511520, "step": 5540 }, { "epoch": 0.17748543627168556, "grad_norm": 1.621883749961853, "learning_rate": 4.766727424738356e-05, "loss": 0.4273, "num_input_tokens_seen": 7518464, "step": 5545 }, { "epoch": 0.177645477242174, "grad_norm": 1.4665484428405762, "learning_rate": 4.7663131531582325e-05, "loss": 0.5542, "num_input_tokens_seen": 7524896, "step": 5550 }, { "epoch": 0.17780551821266244, "grad_norm": 1.3871748447418213, "learning_rate": 4.765898532084142e-05, "loss": 0.484, "num_input_tokens_seen": 7531568, "step": 5555 }, { "epoch": 0.17796555918315088, "grad_norm": 1.144497036933899, "learning_rate": 4.765483561580022e-05, "loss": 0.5585, "num_input_tokens_seen": 7538352, "step": 5560 }, { "epoch": 0.17812560015363932, "grad_norm": 1.5267374515533447, "learning_rate": 4.7650682417098666e-05, "loss": 0.6369, "num_input_tokens_seen": 7545440, "step": 5565 }, { "epoch": 0.17828564112412779, "grad_norm": 0.9913976192474365, "learning_rate": 4.7646525725377244e-05, "loss": 0.5221, "num_input_tokens_seen": 7551808, "step": 5570 }, { "epoch": 0.17844568209461623, "grad_norm": 0.7842673063278198, "learning_rate": 4.764236554127696e-05, "loss": 0.474, "num_input_tokens_seen": 7558848, "step": 5575 }, { "epoch": 0.17860572306510467, "grad_norm": 0.9168071746826172, "learning_rate": 4.7638201865439356e-05, "loss": 0.511, "num_input_tokens_seen": 7565440, "step": 5580 }, { "epoch": 0.1787657640355931, "grad_norm": 2.5766913890838623, "learning_rate": 4.7634034698506545e-05, "loss": 0.5987, "num_input_tokens_seen": 7572320, "step": 5585 }, { "epoch": 0.17892580500608155, "grad_norm": 0.9749208688735962, "learning_rate": 4.762986404112115e-05, "loss": 0.5743, "num_input_tokens_seen": 7578880, "step": 5590 }, { "epoch": 0.17908584597657, "grad_norm": 1.48108971118927, "learning_rate": 4.762568989392633e-05, "loss": 0.3523, "num_input_tokens_seen": 7585728, "step": 5595 }, { "epoch": 0.17924588694705845, "grad_norm": 1.3042562007904053, "learning_rate": 4.76215122575658e-05, "loss": 0.5367, "num_input_tokens_seen": 7592304, "step": 5600 }, { "epoch": 0.17924588694705845, "eval_loss": 0.559414803981781, "eval_runtime": 499.5323, "eval_samples_per_second": 27.798, "eval_steps_per_second": 13.899, "num_input_tokens_seen": 7592304, "step": 5600 }, { "epoch": 0.1794059279175469, "grad_norm": 1.626962661743164, "learning_rate": 4.7617331132683795e-05, "loss": 0.6019, "num_input_tokens_seen": 7598784, "step": 5605 }, { "epoch": 0.17956596888803533, "grad_norm": 1.0079940557479858, "learning_rate": 4.7613146519925105e-05, "loss": 0.5365, "num_input_tokens_seen": 7605536, "step": 5610 }, { "epoch": 0.17972600985852377, "grad_norm": 0.9455792903900146, "learning_rate": 4.7608958419935045e-05, "loss": 0.3923, "num_input_tokens_seen": 7612576, "step": 5615 }, { "epoch": 0.17988605082901224, "grad_norm": 1.7698194980621338, "learning_rate": 4.760476683335948e-05, "loss": 0.5591, "num_input_tokens_seen": 7619104, "step": 5620 }, { "epoch": 0.18004609179950068, "grad_norm": 1.4199872016906738, "learning_rate": 4.760057176084479e-05, "loss": 0.5188, "num_input_tokens_seen": 7625584, "step": 5625 }, { "epoch": 0.18020613276998912, "grad_norm": 0.8572186827659607, "learning_rate": 4.759637320303793e-05, "loss": 0.3524, "num_input_tokens_seen": 7632128, "step": 5630 }, { "epoch": 0.18036617374047756, "grad_norm": 1.4517016410827637, "learning_rate": 4.759217116058635e-05, "loss": 0.4911, "num_input_tokens_seen": 7638720, "step": 5635 }, { "epoch": 0.180526214710966, "grad_norm": 0.8337253332138062, "learning_rate": 4.758796563413807e-05, "loss": 0.4542, "num_input_tokens_seen": 7645856, "step": 5640 }, { "epoch": 0.18068625568145444, "grad_norm": 0.9037920236587524, "learning_rate": 4.758375662434163e-05, "loss": 0.4053, "num_input_tokens_seen": 7652544, "step": 5645 }, { "epoch": 0.1808462966519429, "grad_norm": 2.350764751434326, "learning_rate": 4.7579544131846114e-05, "loss": 0.5879, "num_input_tokens_seen": 7659568, "step": 5650 }, { "epoch": 0.18100633762243135, "grad_norm": 1.1671278476715088, "learning_rate": 4.757532815730114e-05, "loss": 0.4725, "num_input_tokens_seen": 7666384, "step": 5655 }, { "epoch": 0.18116637859291979, "grad_norm": 1.096500039100647, "learning_rate": 4.7571108701356865e-05, "loss": 0.4861, "num_input_tokens_seen": 7672944, "step": 5660 }, { "epoch": 0.18132641956340823, "grad_norm": 1.692191243171692, "learning_rate": 4.756688576466398e-05, "loss": 0.595, "num_input_tokens_seen": 7679792, "step": 5665 }, { "epoch": 0.18148646053389667, "grad_norm": 1.5199904441833496, "learning_rate": 4.756265934787372e-05, "loss": 0.5037, "num_input_tokens_seen": 7686944, "step": 5670 }, { "epoch": 0.18164650150438513, "grad_norm": 0.8691008687019348, "learning_rate": 4.755842945163785e-05, "loss": 0.4885, "num_input_tokens_seen": 7694416, "step": 5675 }, { "epoch": 0.18180654247487357, "grad_norm": 1.6944910287857056, "learning_rate": 4.755419607660867e-05, "loss": 0.6085, "num_input_tokens_seen": 7701504, "step": 5680 }, { "epoch": 0.181966583445362, "grad_norm": 1.7315553426742554, "learning_rate": 4.7549959223439016e-05, "loss": 0.5693, "num_input_tokens_seen": 7708080, "step": 5685 }, { "epoch": 0.18212662441585045, "grad_norm": 1.5747355222702026, "learning_rate": 4.754571889278228e-05, "loss": 0.5514, "num_input_tokens_seen": 7714896, "step": 5690 }, { "epoch": 0.1822866653863389, "grad_norm": 0.8164492249488831, "learning_rate": 4.754147508529235e-05, "loss": 0.5177, "num_input_tokens_seen": 7721376, "step": 5695 }, { "epoch": 0.18244670635682736, "grad_norm": 0.6455702185630798, "learning_rate": 4.75372278016237e-05, "loss": 0.5091, "num_input_tokens_seen": 7728160, "step": 5700 }, { "epoch": 0.1826067473273158, "grad_norm": 0.8040542602539062, "learning_rate": 4.753297704243129e-05, "loss": 0.4321, "num_input_tokens_seen": 7734672, "step": 5705 }, { "epoch": 0.18276678829780424, "grad_norm": 1.0208930969238281, "learning_rate": 4.752872280837066e-05, "loss": 0.549, "num_input_tokens_seen": 7741760, "step": 5710 }, { "epoch": 0.18292682926829268, "grad_norm": 2.3513236045837402, "learning_rate": 4.752446510009786e-05, "loss": 0.6238, "num_input_tokens_seen": 7748992, "step": 5715 }, { "epoch": 0.18308687023878112, "grad_norm": 1.2859725952148438, "learning_rate": 4.7520203918269476e-05, "loss": 0.6692, "num_input_tokens_seen": 7756016, "step": 5720 }, { "epoch": 0.1832469112092696, "grad_norm": 1.5006886720657349, "learning_rate": 4.751593926354265e-05, "loss": 0.4178, "num_input_tokens_seen": 7762800, "step": 5725 }, { "epoch": 0.18340695217975803, "grad_norm": 0.9897240996360779, "learning_rate": 4.751167113657503e-05, "loss": 0.5478, "num_input_tokens_seen": 7769616, "step": 5730 }, { "epoch": 0.18356699315024647, "grad_norm": 0.6185619831085205, "learning_rate": 4.7507399538024834e-05, "loss": 0.3582, "num_input_tokens_seen": 7776208, "step": 5735 }, { "epoch": 0.1837270341207349, "grad_norm": 0.8893991112709045, "learning_rate": 4.750312446855077e-05, "loss": 0.6644, "num_input_tokens_seen": 7783200, "step": 5740 }, { "epoch": 0.18388707509122335, "grad_norm": 1.2159312963485718, "learning_rate": 4.749884592881212e-05, "loss": 0.6159, "num_input_tokens_seen": 7789776, "step": 5745 }, { "epoch": 0.18404711606171179, "grad_norm": 1.8486347198486328, "learning_rate": 4.74945639194687e-05, "loss": 0.4321, "num_input_tokens_seen": 7796544, "step": 5750 }, { "epoch": 0.18420715703220025, "grad_norm": 1.607284426689148, "learning_rate": 4.749027844118083e-05, "loss": 0.5153, "num_input_tokens_seen": 7803344, "step": 5755 }, { "epoch": 0.1843671980026887, "grad_norm": 1.2704803943634033, "learning_rate": 4.7485989494609395e-05, "loss": 0.6109, "num_input_tokens_seen": 7810272, "step": 5760 }, { "epoch": 0.18452723897317713, "grad_norm": 1.178690791130066, "learning_rate": 4.748169708041581e-05, "loss": 0.5312, "num_input_tokens_seen": 7817072, "step": 5765 }, { "epoch": 0.18468727994366557, "grad_norm": 0.9169661998748779, "learning_rate": 4.7477401199262004e-05, "loss": 0.4995, "num_input_tokens_seen": 7823984, "step": 5770 }, { "epoch": 0.184847320914154, "grad_norm": 1.8737175464630127, "learning_rate": 4.747310185181048e-05, "loss": 0.4906, "num_input_tokens_seen": 7831088, "step": 5775 }, { "epoch": 0.18500736188464248, "grad_norm": 1.975856900215149, "learning_rate": 4.746879903872422e-05, "loss": 0.5667, "num_input_tokens_seen": 7838000, "step": 5780 }, { "epoch": 0.18516740285513092, "grad_norm": 1.4703028202056885, "learning_rate": 4.746449276066679e-05, "loss": 0.4441, "num_input_tokens_seen": 7844912, "step": 5785 }, { "epoch": 0.18532744382561936, "grad_norm": 0.6547195315361023, "learning_rate": 4.746018301830227e-05, "loss": 0.4816, "num_input_tokens_seen": 7851488, "step": 5790 }, { "epoch": 0.1854874847961078, "grad_norm": 0.9997721314430237, "learning_rate": 4.7455869812295275e-05, "loss": 0.6007, "num_input_tokens_seen": 7858736, "step": 5795 }, { "epoch": 0.18564752576659624, "grad_norm": 1.329163670539856, "learning_rate": 4.7451553143310964e-05, "loss": 0.5104, "num_input_tokens_seen": 7865632, "step": 5800 }, { "epoch": 0.18564752576659624, "eval_loss": 0.5570062398910522, "eval_runtime": 499.4279, "eval_samples_per_second": 27.804, "eval_steps_per_second": 13.902, "num_input_tokens_seen": 7865632, "step": 5800 }, { "epoch": 0.1858075667370847, "grad_norm": 0.9851922392845154, "learning_rate": 4.744723301201501e-05, "loss": 0.4707, "num_input_tokens_seen": 7872240, "step": 5805 }, { "epoch": 0.18596760770757315, "grad_norm": 1.1137782335281372, "learning_rate": 4.744290941907364e-05, "loss": 0.4957, "num_input_tokens_seen": 7879648, "step": 5810 }, { "epoch": 0.18612764867806159, "grad_norm": 1.4623578786849976, "learning_rate": 4.7438582365153594e-05, "loss": 0.5664, "num_input_tokens_seen": 7886656, "step": 5815 }, { "epoch": 0.18628768964855003, "grad_norm": 1.7241991758346558, "learning_rate": 4.743425185092217e-05, "loss": 0.5368, "num_input_tokens_seen": 7893280, "step": 5820 }, { "epoch": 0.18644773061903847, "grad_norm": 0.893438458442688, "learning_rate": 4.742991787704719e-05, "loss": 0.522, "num_input_tokens_seen": 7900176, "step": 5825 }, { "epoch": 0.1866077715895269, "grad_norm": 1.1876647472381592, "learning_rate": 4.7425580444196994e-05, "loss": 0.3548, "num_input_tokens_seen": 7906784, "step": 5830 }, { "epoch": 0.18676781256001537, "grad_norm": 1.2325985431671143, "learning_rate": 4.742123955304048e-05, "loss": 0.5269, "num_input_tokens_seen": 7914016, "step": 5835 }, { "epoch": 0.1869278535305038, "grad_norm": 1.5432966947555542, "learning_rate": 4.741689520424706e-05, "loss": 0.5088, "num_input_tokens_seen": 7920528, "step": 5840 }, { "epoch": 0.18708789450099225, "grad_norm": 1.4208271503448486, "learning_rate": 4.741254739848669e-05, "loss": 0.6057, "num_input_tokens_seen": 7927504, "step": 5845 }, { "epoch": 0.1872479354714807, "grad_norm": 1.2025295495986938, "learning_rate": 4.740819613642987e-05, "loss": 0.4873, "num_input_tokens_seen": 7934688, "step": 5850 }, { "epoch": 0.18740797644196913, "grad_norm": 1.2804224491119385, "learning_rate": 4.74038414187476e-05, "loss": 0.4441, "num_input_tokens_seen": 7941264, "step": 5855 }, { "epoch": 0.1875680174124576, "grad_norm": 2.1793251037597656, "learning_rate": 4.739948324611144e-05, "loss": 0.587, "num_input_tokens_seen": 7948112, "step": 5860 }, { "epoch": 0.18772805838294604, "grad_norm": 0.9177089333534241, "learning_rate": 4.7395121619193465e-05, "loss": 0.4688, "num_input_tokens_seen": 7954672, "step": 5865 }, { "epoch": 0.18788809935343448, "grad_norm": 2.1292526721954346, "learning_rate": 4.7390756538666313e-05, "loss": 0.7277, "num_input_tokens_seen": 7961360, "step": 5870 }, { "epoch": 0.18804814032392292, "grad_norm": 0.8389447927474976, "learning_rate": 4.738638800520311e-05, "loss": 0.448, "num_input_tokens_seen": 7967984, "step": 5875 }, { "epoch": 0.18820818129441136, "grad_norm": 1.0759825706481934, "learning_rate": 4.738201601947757e-05, "loss": 0.6854, "num_input_tokens_seen": 7974736, "step": 5880 }, { "epoch": 0.18836822226489983, "grad_norm": 2.716204881668091, "learning_rate": 4.7377640582163876e-05, "loss": 0.4592, "num_input_tokens_seen": 7981280, "step": 5885 }, { "epoch": 0.18852826323538827, "grad_norm": 0.8038913607597351, "learning_rate": 4.7373261693936786e-05, "loss": 0.614, "num_input_tokens_seen": 7987728, "step": 5890 }, { "epoch": 0.1886883042058767, "grad_norm": 1.7021831274032593, "learning_rate": 4.7368879355471595e-05, "loss": 0.6414, "num_input_tokens_seen": 7994368, "step": 5895 }, { "epoch": 0.18884834517636515, "grad_norm": 1.2488754987716675, "learning_rate": 4.736449356744409e-05, "loss": 0.4162, "num_input_tokens_seen": 8001216, "step": 5900 }, { "epoch": 0.18900838614685359, "grad_norm": 1.1302567720413208, "learning_rate": 4.736010433053064e-05, "loss": 0.5406, "num_input_tokens_seen": 8008048, "step": 5905 }, { "epoch": 0.18916842711734205, "grad_norm": 1.4172189235687256, "learning_rate": 4.73557116454081e-05, "loss": 0.5266, "num_input_tokens_seen": 8015040, "step": 5910 }, { "epoch": 0.1893284680878305, "grad_norm": 0.9638739824295044, "learning_rate": 4.735131551275389e-05, "loss": 0.479, "num_input_tokens_seen": 8021600, "step": 5915 }, { "epoch": 0.18948850905831893, "grad_norm": 1.0422859191894531, "learning_rate": 4.734691593324594e-05, "loss": 0.537, "num_input_tokens_seen": 8027856, "step": 5920 }, { "epoch": 0.18964855002880737, "grad_norm": 1.6121307611465454, "learning_rate": 4.734251290756272e-05, "loss": 0.5886, "num_input_tokens_seen": 8034560, "step": 5925 }, { "epoch": 0.1898085909992958, "grad_norm": 1.055753469467163, "learning_rate": 4.7338106436383246e-05, "loss": 0.5272, "num_input_tokens_seen": 8041088, "step": 5930 }, { "epoch": 0.18996863196978425, "grad_norm": 2.2536258697509766, "learning_rate": 4.733369652038703e-05, "loss": 0.632, "num_input_tokens_seen": 8047792, "step": 5935 }, { "epoch": 0.19012867294027272, "grad_norm": 2.1544320583343506, "learning_rate": 4.7329283160254156e-05, "loss": 0.5896, "num_input_tokens_seen": 8054736, "step": 5940 }, { "epoch": 0.19028871391076116, "grad_norm": 1.4799691438674927, "learning_rate": 4.732486635666521e-05, "loss": 0.5298, "num_input_tokens_seen": 8061872, "step": 5945 }, { "epoch": 0.1904487548812496, "grad_norm": 1.061737298965454, "learning_rate": 4.732044611030132e-05, "loss": 0.6161, "num_input_tokens_seen": 8068384, "step": 5950 }, { "epoch": 0.19060879585173804, "grad_norm": 0.9693939089775085, "learning_rate": 4.731602242184414e-05, "loss": 0.5508, "num_input_tokens_seen": 8075312, "step": 5955 }, { "epoch": 0.19076883682222648, "grad_norm": 1.5295010805130005, "learning_rate": 4.7311595291975864e-05, "loss": 0.4428, "num_input_tokens_seen": 8081808, "step": 5960 }, { "epoch": 0.19092887779271495, "grad_norm": 1.9300684928894043, "learning_rate": 4.7307164721379216e-05, "loss": 0.6498, "num_input_tokens_seen": 8088496, "step": 5965 }, { "epoch": 0.19108891876320339, "grad_norm": 1.8397740125656128, "learning_rate": 4.730273071073743e-05, "loss": 0.4559, "num_input_tokens_seen": 8095104, "step": 5970 }, { "epoch": 0.19124895973369183, "grad_norm": 0.8847364783287048, "learning_rate": 4.729829326073429e-05, "loss": 0.3025, "num_input_tokens_seen": 8102144, "step": 5975 }, { "epoch": 0.19140900070418027, "grad_norm": 0.6775109767913818, "learning_rate": 4.7293852372054126e-05, "loss": 0.577, "num_input_tokens_seen": 8109088, "step": 5980 }, { "epoch": 0.1915690416746687, "grad_norm": 0.8566697239875793, "learning_rate": 4.728940804538176e-05, "loss": 0.524, "num_input_tokens_seen": 8115792, "step": 5985 }, { "epoch": 0.19172908264515717, "grad_norm": 0.9764266014099121, "learning_rate": 4.7284960281402556e-05, "loss": 0.5337, "num_input_tokens_seen": 8122208, "step": 5990 }, { "epoch": 0.1918891236156456, "grad_norm": 1.516564130783081, "learning_rate": 4.728050908080244e-05, "loss": 0.462, "num_input_tokens_seen": 8128912, "step": 5995 }, { "epoch": 0.19204916458613405, "grad_norm": 1.509053349494934, "learning_rate": 4.727605444426782e-05, "loss": 0.4609, "num_input_tokens_seen": 8135936, "step": 6000 }, { "epoch": 0.19204916458613405, "eval_loss": 0.555640697479248, "eval_runtime": 499.6831, "eval_samples_per_second": 27.79, "eval_steps_per_second": 13.895, "num_input_tokens_seen": 8135936, "step": 6000 }, { "epoch": 0.1922092055566225, "grad_norm": 1.252049446105957, "learning_rate": 4.727159637248567e-05, "loss": 0.3929, "num_input_tokens_seen": 8142464, "step": 6005 }, { "epoch": 0.19236924652711093, "grad_norm": 1.1601756811141968, "learning_rate": 4.7267134866143474e-05, "loss": 0.6568, "num_input_tokens_seen": 8149456, "step": 6010 }, { "epoch": 0.19252928749759937, "grad_norm": 0.6003838181495667, "learning_rate": 4.726266992592926e-05, "loss": 0.5114, "num_input_tokens_seen": 8156192, "step": 6015 }, { "epoch": 0.19268932846808784, "grad_norm": 0.6521717309951782, "learning_rate": 4.725820155253157e-05, "loss": 0.5148, "num_input_tokens_seen": 8162992, "step": 6020 }, { "epoch": 0.19284936943857628, "grad_norm": 1.2710778713226318, "learning_rate": 4.725372974663948e-05, "loss": 0.4094, "num_input_tokens_seen": 8169984, "step": 6025 }, { "epoch": 0.19300941040906472, "grad_norm": 1.5693533420562744, "learning_rate": 4.724925450894262e-05, "loss": 0.4249, "num_input_tokens_seen": 8177008, "step": 6030 }, { "epoch": 0.19316945137955316, "grad_norm": 1.179375410079956, "learning_rate": 4.72447758401311e-05, "loss": 0.4743, "num_input_tokens_seen": 8183408, "step": 6035 }, { "epoch": 0.1933294923500416, "grad_norm": 1.4688984155654907, "learning_rate": 4.7240293740895616e-05, "loss": 0.8249, "num_input_tokens_seen": 8190368, "step": 6040 }, { "epoch": 0.19348953332053007, "grad_norm": 1.0370347499847412, "learning_rate": 4.723580821192733e-05, "loss": 0.6142, "num_input_tokens_seen": 8197152, "step": 6045 }, { "epoch": 0.1936495742910185, "grad_norm": 1.3216522932052612, "learning_rate": 4.7231319253917996e-05, "loss": 0.5621, "num_input_tokens_seen": 8203920, "step": 6050 }, { "epoch": 0.19380961526150695, "grad_norm": 0.7342177033424377, "learning_rate": 4.722682686755986e-05, "loss": 0.4366, "num_input_tokens_seen": 8210656, "step": 6055 }, { "epoch": 0.19396965623199539, "grad_norm": 0.9559981822967529, "learning_rate": 4.722233105354569e-05, "loss": 0.6158, "num_input_tokens_seen": 8217568, "step": 6060 }, { "epoch": 0.19412969720248383, "grad_norm": 1.2518150806427002, "learning_rate": 4.7217831812568815e-05, "loss": 0.4461, "num_input_tokens_seen": 8224704, "step": 6065 }, { "epoch": 0.1942897381729723, "grad_norm": 1.8399450778961182, "learning_rate": 4.721332914532307e-05, "loss": 0.9037, "num_input_tokens_seen": 8231600, "step": 6070 }, { "epoch": 0.19444977914346073, "grad_norm": 2.281182289123535, "learning_rate": 4.720882305250281e-05, "loss": 0.6976, "num_input_tokens_seen": 8238032, "step": 6075 }, { "epoch": 0.19460982011394917, "grad_norm": 1.441968321800232, "learning_rate": 4.720431353480295e-05, "loss": 0.5078, "num_input_tokens_seen": 8244848, "step": 6080 }, { "epoch": 0.1947698610844376, "grad_norm": 2.847297191619873, "learning_rate": 4.719980059291891e-05, "loss": 0.5123, "num_input_tokens_seen": 8251376, "step": 6085 }, { "epoch": 0.19492990205492605, "grad_norm": 1.6190491914749146, "learning_rate": 4.7195284227546634e-05, "loss": 0.4998, "num_input_tokens_seen": 8258032, "step": 6090 }, { "epoch": 0.19508994302541452, "grad_norm": 0.8559789061546326, "learning_rate": 4.7190764439382604e-05, "loss": 0.4653, "num_input_tokens_seen": 8264784, "step": 6095 }, { "epoch": 0.19524998399590296, "grad_norm": 1.1290305852890015, "learning_rate": 4.7186241229123826e-05, "loss": 0.6361, "num_input_tokens_seen": 8271792, "step": 6100 }, { "epoch": 0.1954100249663914, "grad_norm": 0.7173714637756348, "learning_rate": 4.718171459746785e-05, "loss": 0.5752, "num_input_tokens_seen": 8279008, "step": 6105 }, { "epoch": 0.19557006593687984, "grad_norm": 1.056814193725586, "learning_rate": 4.717718454511273e-05, "loss": 0.3823, "num_input_tokens_seen": 8285760, "step": 6110 }, { "epoch": 0.19573010690736828, "grad_norm": 1.2984260320663452, "learning_rate": 4.7172651072757056e-05, "loss": 0.5566, "num_input_tokens_seen": 8292768, "step": 6115 }, { "epoch": 0.19589014787785672, "grad_norm": 1.6415011882781982, "learning_rate": 4.7168114181099945e-05, "loss": 0.6641, "num_input_tokens_seen": 8299440, "step": 6120 }, { "epoch": 0.19605018884834519, "grad_norm": 0.7111352682113647, "learning_rate": 4.716357387084105e-05, "loss": 0.3272, "num_input_tokens_seen": 8306288, "step": 6125 }, { "epoch": 0.19621022981883363, "grad_norm": 0.8466628193855286, "learning_rate": 4.715903014268054e-05, "loss": 0.5837, "num_input_tokens_seen": 8313376, "step": 6130 }, { "epoch": 0.19637027078932207, "grad_norm": 1.9052448272705078, "learning_rate": 4.715448299731911e-05, "loss": 0.575, "num_input_tokens_seen": 8319648, "step": 6135 }, { "epoch": 0.1965303117598105, "grad_norm": 1.961034893989563, "learning_rate": 4.7149932435457986e-05, "loss": 0.6112, "num_input_tokens_seen": 8326592, "step": 6140 }, { "epoch": 0.19669035273029895, "grad_norm": 0.7574659585952759, "learning_rate": 4.714537845779894e-05, "loss": 0.4138, "num_input_tokens_seen": 8333920, "step": 6145 }, { "epoch": 0.1968503937007874, "grad_norm": 1.5863101482391357, "learning_rate": 4.714082106504423e-05, "loss": 0.4067, "num_input_tokens_seen": 8340896, "step": 6150 }, { "epoch": 0.19701043467127585, "grad_norm": 1.2942237854003906, "learning_rate": 4.713626025789667e-05, "loss": 0.5345, "num_input_tokens_seen": 8347488, "step": 6155 }, { "epoch": 0.1971704756417643, "grad_norm": 0.9243707656860352, "learning_rate": 4.7131696037059606e-05, "loss": 0.4767, "num_input_tokens_seen": 8354528, "step": 6160 }, { "epoch": 0.19733051661225273, "grad_norm": 1.235894799232483, "learning_rate": 4.712712840323689e-05, "loss": 0.5945, "num_input_tokens_seen": 8361376, "step": 6165 }, { "epoch": 0.19749055758274117, "grad_norm": 1.1548211574554443, "learning_rate": 4.71225573571329e-05, "loss": 0.5783, "num_input_tokens_seen": 8368176, "step": 6170 }, { "epoch": 0.19765059855322964, "grad_norm": 0.9415286779403687, "learning_rate": 4.711798289945256e-05, "loss": 0.4856, "num_input_tokens_seen": 8375024, "step": 6175 }, { "epoch": 0.19781063952371808, "grad_norm": 0.6809241771697998, "learning_rate": 4.71134050309013e-05, "loss": 0.4446, "num_input_tokens_seen": 8381504, "step": 6180 }, { "epoch": 0.19797068049420652, "grad_norm": 1.7175668478012085, "learning_rate": 4.710882375218509e-05, "loss": 0.5376, "num_input_tokens_seen": 8388160, "step": 6185 }, { "epoch": 0.19813072146469496, "grad_norm": 0.8382947444915771, "learning_rate": 4.7104239064010424e-05, "loss": 0.3339, "num_input_tokens_seen": 8394864, "step": 6190 }, { "epoch": 0.1982907624351834, "grad_norm": 1.6291899681091309, "learning_rate": 4.709965096708432e-05, "loss": 0.5423, "num_input_tokens_seen": 8402048, "step": 6195 }, { "epoch": 0.19845080340567184, "grad_norm": 1.2450478076934814, "learning_rate": 4.709505946211431e-05, "loss": 0.601, "num_input_tokens_seen": 8408624, "step": 6200 }, { "epoch": 0.19845080340567184, "eval_loss": 0.5515230894088745, "eval_runtime": 499.9143, "eval_samples_per_second": 27.777, "eval_steps_per_second": 13.888, "num_input_tokens_seen": 8408624, "step": 6200 }, { "epoch": 0.1986108443761603, "grad_norm": 2.168785333633423, "learning_rate": 4.709046454980846e-05, "loss": 0.5229, "num_input_tokens_seen": 8415440, "step": 6205 }, { "epoch": 0.19877088534664875, "grad_norm": 0.6756417155265808, "learning_rate": 4.708586623087538e-05, "loss": 0.6175, "num_input_tokens_seen": 8421888, "step": 6210 }, { "epoch": 0.19893092631713719, "grad_norm": 0.6810401082038879, "learning_rate": 4.708126450602418e-05, "loss": 0.4403, "num_input_tokens_seen": 8428560, "step": 6215 }, { "epoch": 0.19909096728762563, "grad_norm": 2.587329149246216, "learning_rate": 4.7076659375964495e-05, "loss": 0.5501, "num_input_tokens_seen": 8435520, "step": 6220 }, { "epoch": 0.19925100825811407, "grad_norm": 0.9964320659637451, "learning_rate": 4.707205084140651e-05, "loss": 0.4739, "num_input_tokens_seen": 8442336, "step": 6225 }, { "epoch": 0.19941104922860253, "grad_norm": 0.594363808631897, "learning_rate": 4.7067438903060904e-05, "loss": 0.4918, "num_input_tokens_seen": 8449056, "step": 6230 }, { "epoch": 0.19957109019909097, "grad_norm": 1.5069773197174072, "learning_rate": 4.70628235616389e-05, "loss": 0.7452, "num_input_tokens_seen": 8455632, "step": 6235 }, { "epoch": 0.1997311311695794, "grad_norm": 1.243535041809082, "learning_rate": 4.7058204817852256e-05, "loss": 0.4857, "num_input_tokens_seen": 8462864, "step": 6240 }, { "epoch": 0.19989117214006785, "grad_norm": 1.8732831478118896, "learning_rate": 4.705358267241322e-05, "loss": 0.7403, "num_input_tokens_seen": 8469520, "step": 6245 }, { "epoch": 0.2000512131105563, "grad_norm": 1.3340702056884766, "learning_rate": 4.704895712603459e-05, "loss": 0.5214, "num_input_tokens_seen": 8475936, "step": 6250 }, { "epoch": 0.20021125408104476, "grad_norm": 0.7943743467330933, "learning_rate": 4.704432817942969e-05, "loss": 0.5506, "num_input_tokens_seen": 8482928, "step": 6255 }, { "epoch": 0.2003712950515332, "grad_norm": 1.6983168125152588, "learning_rate": 4.703969583331236e-05, "loss": 0.4379, "num_input_tokens_seen": 8489536, "step": 6260 }, { "epoch": 0.20053133602202164, "grad_norm": 0.8265430331230164, "learning_rate": 4.7035060088396965e-05, "loss": 0.4804, "num_input_tokens_seen": 8496400, "step": 6265 }, { "epoch": 0.20069137699251008, "grad_norm": 0.8851269483566284, "learning_rate": 4.703042094539839e-05, "loss": 0.4189, "num_input_tokens_seen": 8502784, "step": 6270 }, { "epoch": 0.20085141796299852, "grad_norm": 1.252740502357483, "learning_rate": 4.702577840503206e-05, "loss": 0.4962, "num_input_tokens_seen": 8509232, "step": 6275 }, { "epoch": 0.201011458933487, "grad_norm": 1.5545530319213867, "learning_rate": 4.70211324680139e-05, "loss": 0.3685, "num_input_tokens_seen": 8515904, "step": 6280 }, { "epoch": 0.20117149990397543, "grad_norm": 1.546854853630066, "learning_rate": 4.7016483135060386e-05, "loss": 0.4031, "num_input_tokens_seen": 8522736, "step": 6285 }, { "epoch": 0.20133154087446387, "grad_norm": 1.1003090143203735, "learning_rate": 4.701183040688849e-05, "loss": 0.7305, "num_input_tokens_seen": 8529232, "step": 6290 }, { "epoch": 0.2014915818449523, "grad_norm": 0.65513676404953, "learning_rate": 4.700717428421573e-05, "loss": 0.5648, "num_input_tokens_seen": 8536080, "step": 6295 }, { "epoch": 0.20165162281544075, "grad_norm": 2.120959758758545, "learning_rate": 4.700251476776014e-05, "loss": 0.4089, "num_input_tokens_seen": 8542464, "step": 6300 }, { "epoch": 0.20181166378592919, "grad_norm": 1.607968807220459, "learning_rate": 4.699785185824026e-05, "loss": 0.3945, "num_input_tokens_seen": 8549408, "step": 6305 }, { "epoch": 0.20197170475641765, "grad_norm": 1.4245680570602417, "learning_rate": 4.699318555637519e-05, "loss": 0.567, "num_input_tokens_seen": 8556208, "step": 6310 }, { "epoch": 0.2021317457269061, "grad_norm": 1.4906729459762573, "learning_rate": 4.6988515862884525e-05, "loss": 0.5268, "num_input_tokens_seen": 8563040, "step": 6315 }, { "epoch": 0.20229178669739453, "grad_norm": 1.015112042427063, "learning_rate": 4.698384277848838e-05, "loss": 0.5025, "num_input_tokens_seen": 8569712, "step": 6320 }, { "epoch": 0.20245182766788297, "grad_norm": 1.2687880992889404, "learning_rate": 4.6979166303907425e-05, "loss": 0.4516, "num_input_tokens_seen": 8576752, "step": 6325 }, { "epoch": 0.2026118686383714, "grad_norm": 1.1787205934524536, "learning_rate": 4.697448643986281e-05, "loss": 0.3825, "num_input_tokens_seen": 8583632, "step": 6330 }, { "epoch": 0.20277190960885988, "grad_norm": 1.1719363927841187, "learning_rate": 4.696980318707624e-05, "loss": 0.6491, "num_input_tokens_seen": 8590320, "step": 6335 }, { "epoch": 0.20293195057934832, "grad_norm": 1.9873487949371338, "learning_rate": 4.6965116546269924e-05, "loss": 0.6169, "num_input_tokens_seen": 8596992, "step": 6340 }, { "epoch": 0.20309199154983676, "grad_norm": 1.1269888877868652, "learning_rate": 4.6960426518166615e-05, "loss": 0.8171, "num_input_tokens_seen": 8603760, "step": 6345 }, { "epoch": 0.2032520325203252, "grad_norm": 1.2940865755081177, "learning_rate": 4.6955733103489556e-05, "loss": 0.5111, "num_input_tokens_seen": 8610448, "step": 6350 }, { "epoch": 0.20341207349081364, "grad_norm": 1.9645777940750122, "learning_rate": 4.695103630296255e-05, "loss": 0.6953, "num_input_tokens_seen": 8617488, "step": 6355 }, { "epoch": 0.2035721144613021, "grad_norm": 1.7625967264175415, "learning_rate": 4.694633611730988e-05, "loss": 0.4688, "num_input_tokens_seen": 8624096, "step": 6360 }, { "epoch": 0.20373215543179055, "grad_norm": 1.1603354215621948, "learning_rate": 4.694163254725639e-05, "loss": 0.5271, "num_input_tokens_seen": 8630688, "step": 6365 }, { "epoch": 0.20389219640227899, "grad_norm": 1.0903339385986328, "learning_rate": 4.693692559352743e-05, "loss": 0.62, "num_input_tokens_seen": 8637216, "step": 6370 }, { "epoch": 0.20405223737276743, "grad_norm": 1.2826030254364014, "learning_rate": 4.693221525684886e-05, "loss": 0.458, "num_input_tokens_seen": 8644224, "step": 6375 }, { "epoch": 0.20421227834325587, "grad_norm": 0.9085673689842224, "learning_rate": 4.6927501537947084e-05, "loss": 0.5148, "num_input_tokens_seen": 8651088, "step": 6380 }, { "epoch": 0.2043723193137443, "grad_norm": 2.4471540451049805, "learning_rate": 4.692278443754901e-05, "loss": 0.8046, "num_input_tokens_seen": 8657680, "step": 6385 }, { "epoch": 0.20453236028423277, "grad_norm": 1.8873424530029297, "learning_rate": 4.691806395638208e-05, "loss": 0.7242, "num_input_tokens_seen": 8664560, "step": 6390 }, { "epoch": 0.2046924012547212, "grad_norm": 1.7071795463562012, "learning_rate": 4.6913340095174255e-05, "loss": 0.3594, "num_input_tokens_seen": 8671136, "step": 6395 }, { "epoch": 0.20485244222520965, "grad_norm": 1.1694531440734863, "learning_rate": 4.690861285465399e-05, "loss": 0.5025, "num_input_tokens_seen": 8677888, "step": 6400 }, { "epoch": 0.20485244222520965, "eval_loss": 0.5480764508247375, "eval_runtime": 499.9558, "eval_samples_per_second": 27.774, "eval_steps_per_second": 13.887, "num_input_tokens_seen": 8677888, "step": 6400 }, { "epoch": 0.2050124831956981, "grad_norm": 3.1378273963928223, "learning_rate": 4.690388223555031e-05, "loss": 0.5362, "num_input_tokens_seen": 8684240, "step": 6405 }, { "epoch": 0.20517252416618653, "grad_norm": 0.9686762690544128, "learning_rate": 4.689914823859273e-05, "loss": 0.5475, "num_input_tokens_seen": 8691184, "step": 6410 }, { "epoch": 0.205332565136675, "grad_norm": 1.6599533557891846, "learning_rate": 4.689441086451129e-05, "loss": 0.4522, "num_input_tokens_seen": 8698096, "step": 6415 }, { "epoch": 0.20549260610716344, "grad_norm": 1.274120807647705, "learning_rate": 4.688967011403655e-05, "loss": 0.3087, "num_input_tokens_seen": 8705024, "step": 6420 }, { "epoch": 0.20565264707765188, "grad_norm": 0.859351634979248, "learning_rate": 4.68849259878996e-05, "loss": 0.4379, "num_input_tokens_seen": 8711520, "step": 6425 }, { "epoch": 0.20581268804814032, "grad_norm": 0.8091830611228943, "learning_rate": 4.6880178486832036e-05, "loss": 0.5082, "num_input_tokens_seen": 8718208, "step": 6430 }, { "epoch": 0.20597272901862876, "grad_norm": 1.514302372932434, "learning_rate": 4.687542761156598e-05, "loss": 0.5222, "num_input_tokens_seen": 8724592, "step": 6435 }, { "epoch": 0.20613276998911723, "grad_norm": 1.2558414936065674, "learning_rate": 4.6870673362834096e-05, "loss": 0.5143, "num_input_tokens_seen": 8731760, "step": 6440 }, { "epoch": 0.20629281095960567, "grad_norm": 0.9201895594596863, "learning_rate": 4.6865915741369526e-05, "loss": 0.3508, "num_input_tokens_seen": 8738352, "step": 6445 }, { "epoch": 0.2064528519300941, "grad_norm": 0.6903417110443115, "learning_rate": 4.686115474790597e-05, "loss": 0.4929, "num_input_tokens_seen": 8745376, "step": 6450 }, { "epoch": 0.20661289290058255, "grad_norm": 0.7482292056083679, "learning_rate": 4.685639038317762e-05, "loss": 0.6186, "num_input_tokens_seen": 8752304, "step": 6455 }, { "epoch": 0.20677293387107099, "grad_norm": 1.237593650817871, "learning_rate": 4.685162264791921e-05, "loss": 0.4686, "num_input_tokens_seen": 8758784, "step": 6460 }, { "epoch": 0.20693297484155945, "grad_norm": 1.2855192422866821, "learning_rate": 4.684685154286599e-05, "loss": 0.4983, "num_input_tokens_seen": 8765728, "step": 6465 }, { "epoch": 0.2070930158120479, "grad_norm": 0.8556150794029236, "learning_rate": 4.684207706875371e-05, "loss": 0.5, "num_input_tokens_seen": 8772224, "step": 6470 }, { "epoch": 0.20725305678253633, "grad_norm": 1.3031196594238281, "learning_rate": 4.683729922631866e-05, "loss": 0.6993, "num_input_tokens_seen": 8778688, "step": 6475 }, { "epoch": 0.20741309775302477, "grad_norm": 0.9673957228660583, "learning_rate": 4.683251801629765e-05, "loss": 0.5918, "num_input_tokens_seen": 8785312, "step": 6480 }, { "epoch": 0.2075731387235132, "grad_norm": 0.7012938857078552, "learning_rate": 4.6827733439428e-05, "loss": 0.4858, "num_input_tokens_seen": 8791968, "step": 6485 }, { "epoch": 0.20773317969400165, "grad_norm": 1.3180755376815796, "learning_rate": 4.682294549644754e-05, "loss": 0.8712, "num_input_tokens_seen": 8798784, "step": 6490 }, { "epoch": 0.20789322066449012, "grad_norm": 1.0156959295272827, "learning_rate": 4.681815418809464e-05, "loss": 0.6233, "num_input_tokens_seen": 8805728, "step": 6495 }, { "epoch": 0.20805326163497856, "grad_norm": 0.6821162700653076, "learning_rate": 4.681335951510819e-05, "loss": 0.3385, "num_input_tokens_seen": 8812416, "step": 6500 }, { "epoch": 0.208213302605467, "grad_norm": 1.4075595140457153, "learning_rate": 4.6808561478227576e-05, "loss": 0.4869, "num_input_tokens_seen": 8819312, "step": 6505 }, { "epoch": 0.20837334357595544, "grad_norm": 1.5202465057373047, "learning_rate": 4.680376007819271e-05, "loss": 0.5889, "num_input_tokens_seen": 8825776, "step": 6510 }, { "epoch": 0.20853338454644388, "grad_norm": 0.9014536738395691, "learning_rate": 4.679895531574405e-05, "loss": 0.4665, "num_input_tokens_seen": 8832384, "step": 6515 }, { "epoch": 0.20869342551693235, "grad_norm": 1.7033271789550781, "learning_rate": 4.679414719162253e-05, "loss": 0.5455, "num_input_tokens_seen": 8839040, "step": 6520 }, { "epoch": 0.20885346648742079, "grad_norm": 0.6427426934242249, "learning_rate": 4.6789335706569635e-05, "loss": 0.3874, "num_input_tokens_seen": 8845696, "step": 6525 }, { "epoch": 0.20901350745790923, "grad_norm": 1.0254576206207275, "learning_rate": 4.678452086132734e-05, "loss": 0.5121, "num_input_tokens_seen": 8852352, "step": 6530 }, { "epoch": 0.20917354842839767, "grad_norm": 1.2750768661499023, "learning_rate": 4.677970265663818e-05, "loss": 0.7057, "num_input_tokens_seen": 8858944, "step": 6535 }, { "epoch": 0.2093335893988861, "grad_norm": 0.9340868592262268, "learning_rate": 4.677488109324517e-05, "loss": 0.5021, "num_input_tokens_seen": 8865808, "step": 6540 }, { "epoch": 0.20949363036937457, "grad_norm": 0.9170828461647034, "learning_rate": 4.6770056171891846e-05, "loss": 0.5485, "num_input_tokens_seen": 8873168, "step": 6545 }, { "epoch": 0.209653671339863, "grad_norm": 1.3377536535263062, "learning_rate": 4.6765227893322286e-05, "loss": 0.4889, "num_input_tokens_seen": 8879984, "step": 6550 }, { "epoch": 0.20981371231035145, "grad_norm": 1.5099526643753052, "learning_rate": 4.676039625828107e-05, "loss": 0.5782, "num_input_tokens_seen": 8886768, "step": 6555 }, { "epoch": 0.2099737532808399, "grad_norm": 1.300022840499878, "learning_rate": 4.675556126751328e-05, "loss": 0.7687, "num_input_tokens_seen": 8893664, "step": 6560 }, { "epoch": 0.21013379425132833, "grad_norm": 0.7946189641952515, "learning_rate": 4.6750722921764556e-05, "loss": 0.5044, "num_input_tokens_seen": 8900592, "step": 6565 }, { "epoch": 0.21029383522181677, "grad_norm": 1.6128039360046387, "learning_rate": 4.674588122178102e-05, "loss": 0.4341, "num_input_tokens_seen": 8907104, "step": 6570 }, { "epoch": 0.21045387619230524, "grad_norm": 0.8158290386199951, "learning_rate": 4.674103616830931e-05, "loss": 0.4597, "num_input_tokens_seen": 8913856, "step": 6575 }, { "epoch": 0.21061391716279368, "grad_norm": 0.6871660351753235, "learning_rate": 4.673618776209663e-05, "loss": 0.5175, "num_input_tokens_seen": 8920384, "step": 6580 }, { "epoch": 0.21077395813328212, "grad_norm": 1.3151086568832397, "learning_rate": 4.673133600389063e-05, "loss": 0.6104, "num_input_tokens_seen": 8926816, "step": 6585 }, { "epoch": 0.21093399910377056, "grad_norm": 2.948497772216797, "learning_rate": 4.672648089443953e-05, "loss": 0.5974, "num_input_tokens_seen": 8933424, "step": 6590 }, { "epoch": 0.211094040074259, "grad_norm": 0.985835075378418, "learning_rate": 4.672162243449204e-05, "loss": 0.4302, "num_input_tokens_seen": 8940096, "step": 6595 }, { "epoch": 0.21125408104474747, "grad_norm": 1.562153935432434, "learning_rate": 4.67167606247974e-05, "loss": 0.5747, "num_input_tokens_seen": 8947120, "step": 6600 }, { "epoch": 0.21125408104474747, "eval_loss": 0.5447772741317749, "eval_runtime": 499.3152, "eval_samples_per_second": 27.81, "eval_steps_per_second": 13.905, "num_input_tokens_seen": 8947120, "step": 6600 }, { "epoch": 0.2114141220152359, "grad_norm": 1.8915024995803833, "learning_rate": 4.671189546610536e-05, "loss": 0.587, "num_input_tokens_seen": 8954176, "step": 6605 }, { "epoch": 0.21157416298572435, "grad_norm": 1.189724087715149, "learning_rate": 4.67070269591662e-05, "loss": 0.6891, "num_input_tokens_seen": 8960528, "step": 6610 }, { "epoch": 0.21173420395621279, "grad_norm": 1.257282018661499, "learning_rate": 4.670215510473068e-05, "loss": 0.6651, "num_input_tokens_seen": 8967152, "step": 6615 }, { "epoch": 0.21189424492670123, "grad_norm": 2.2665834426879883, "learning_rate": 4.669727990355013e-05, "loss": 0.5792, "num_input_tokens_seen": 8973808, "step": 6620 }, { "epoch": 0.2120542858971897, "grad_norm": 0.8101881742477417, "learning_rate": 4.669240135637635e-05, "loss": 0.4686, "num_input_tokens_seen": 8980352, "step": 6625 }, { "epoch": 0.21221432686767813, "grad_norm": 1.8815263509750366, "learning_rate": 4.6687519463961675e-05, "loss": 0.7223, "num_input_tokens_seen": 8986928, "step": 6630 }, { "epoch": 0.21237436783816657, "grad_norm": 1.1393914222717285, "learning_rate": 4.668263422705896e-05, "loss": 0.4158, "num_input_tokens_seen": 8993472, "step": 6635 }, { "epoch": 0.212534408808655, "grad_norm": 1.0607550144195557, "learning_rate": 4.667774564642156e-05, "loss": 0.5293, "num_input_tokens_seen": 9000320, "step": 6640 }, { "epoch": 0.21269444977914345, "grad_norm": 1.4364094734191895, "learning_rate": 4.6672853722803365e-05, "loss": 0.7872, "num_input_tokens_seen": 9007040, "step": 6645 }, { "epoch": 0.21285449074963192, "grad_norm": 0.8655857443809509, "learning_rate": 4.666795845695877e-05, "loss": 0.5921, "num_input_tokens_seen": 9013664, "step": 6650 }, { "epoch": 0.21301453172012036, "grad_norm": 0.4469898045063019, "learning_rate": 4.666305984964269e-05, "loss": 0.618, "num_input_tokens_seen": 9020624, "step": 6655 }, { "epoch": 0.2131745726906088, "grad_norm": 1.056881308555603, "learning_rate": 4.6658157901610535e-05, "loss": 0.67, "num_input_tokens_seen": 9027376, "step": 6660 }, { "epoch": 0.21333461366109724, "grad_norm": 1.0240585803985596, "learning_rate": 4.665325261361826e-05, "loss": 0.429, "num_input_tokens_seen": 9033696, "step": 6665 }, { "epoch": 0.21349465463158568, "grad_norm": 1.2754712104797363, "learning_rate": 4.664834398642232e-05, "loss": 0.6283, "num_input_tokens_seen": 9039936, "step": 6670 }, { "epoch": 0.21365469560207412, "grad_norm": 1.433945894241333, "learning_rate": 4.6643432020779686e-05, "loss": 0.517, "num_input_tokens_seen": 9047056, "step": 6675 }, { "epoch": 0.21381473657256259, "grad_norm": 1.4668545722961426, "learning_rate": 4.663851671744786e-05, "loss": 0.5535, "num_input_tokens_seen": 9054464, "step": 6680 }, { "epoch": 0.21397477754305103, "grad_norm": 1.1077494621276855, "learning_rate": 4.6633598077184815e-05, "loss": 0.498, "num_input_tokens_seen": 9061040, "step": 6685 }, { "epoch": 0.21413481851353947, "grad_norm": 1.4751697778701782, "learning_rate": 4.662867610074908e-05, "loss": 0.5422, "num_input_tokens_seen": 9067776, "step": 6690 }, { "epoch": 0.2142948594840279, "grad_norm": 1.295549750328064, "learning_rate": 4.6623750788899696e-05, "loss": 0.5766, "num_input_tokens_seen": 9074576, "step": 6695 }, { "epoch": 0.21445490045451635, "grad_norm": 1.391587734222412, "learning_rate": 4.6618822142396195e-05, "loss": 0.5506, "num_input_tokens_seen": 9081312, "step": 6700 }, { "epoch": 0.2146149414250048, "grad_norm": 0.7861068844795227, "learning_rate": 4.661389016199864e-05, "loss": 0.4586, "num_input_tokens_seen": 9088144, "step": 6705 }, { "epoch": 0.21477498239549325, "grad_norm": 1.5441235303878784, "learning_rate": 4.660895484846761e-05, "loss": 0.5176, "num_input_tokens_seen": 9095104, "step": 6710 }, { "epoch": 0.2149350233659817, "grad_norm": 1.0374798774719238, "learning_rate": 4.660401620256418e-05, "loss": 0.5771, "num_input_tokens_seen": 9101808, "step": 6715 }, { "epoch": 0.21509506433647013, "grad_norm": 0.7345461249351501, "learning_rate": 4.659907422504997e-05, "loss": 0.5702, "num_input_tokens_seen": 9109088, "step": 6720 }, { "epoch": 0.21525510530695857, "grad_norm": 1.13515305519104, "learning_rate": 4.6594128916687074e-05, "loss": 0.4629, "num_input_tokens_seen": 9115504, "step": 6725 }, { "epoch": 0.21541514627744704, "grad_norm": 0.9308738708496094, "learning_rate": 4.658918027823813e-05, "loss": 0.3559, "num_input_tokens_seen": 9122144, "step": 6730 }, { "epoch": 0.21557518724793548, "grad_norm": 1.2689313888549805, "learning_rate": 4.658422831046628e-05, "loss": 0.6092, "num_input_tokens_seen": 9129008, "step": 6735 }, { "epoch": 0.21573522821842392, "grad_norm": 1.0939037799835205, "learning_rate": 4.657927301413518e-05, "loss": 0.5611, "num_input_tokens_seen": 9135520, "step": 6740 }, { "epoch": 0.21589526918891236, "grad_norm": 0.9709073305130005, "learning_rate": 4.657431439000901e-05, "loss": 0.3939, "num_input_tokens_seen": 9142336, "step": 6745 }, { "epoch": 0.2160553101594008, "grad_norm": 1.2582662105560303, "learning_rate": 4.656935243885243e-05, "loss": 0.5394, "num_input_tokens_seen": 9148576, "step": 6750 }, { "epoch": 0.21621535112988924, "grad_norm": 0.7235316634178162, "learning_rate": 4.656438716143066e-05, "loss": 0.5185, "num_input_tokens_seen": 9155488, "step": 6755 }, { "epoch": 0.2163753921003777, "grad_norm": 0.43029719591140747, "learning_rate": 4.6559418558509384e-05, "loss": 0.6061, "num_input_tokens_seen": 9162208, "step": 6760 }, { "epoch": 0.21653543307086615, "grad_norm": 1.6924638748168945, "learning_rate": 4.6554446630854833e-05, "loss": 0.3873, "num_input_tokens_seen": 9168912, "step": 6765 }, { "epoch": 0.21669547404135459, "grad_norm": 1.7330669164657593, "learning_rate": 4.654947137923374e-05, "loss": 0.5355, "num_input_tokens_seen": 9175328, "step": 6770 }, { "epoch": 0.21685551501184303, "grad_norm": 1.2427656650543213, "learning_rate": 4.654449280441335e-05, "loss": 0.54, "num_input_tokens_seen": 9182464, "step": 6775 }, { "epoch": 0.21701555598233147, "grad_norm": 0.9708102345466614, "learning_rate": 4.653951090716143e-05, "loss": 0.323, "num_input_tokens_seen": 9189456, "step": 6780 }, { "epoch": 0.21717559695281993, "grad_norm": 0.8329290747642517, "learning_rate": 4.653452568824625e-05, "loss": 0.3093, "num_input_tokens_seen": 9196080, "step": 6785 }, { "epoch": 0.21733563792330837, "grad_norm": 1.0236690044403076, "learning_rate": 4.6529537148436585e-05, "loss": 0.4791, "num_input_tokens_seen": 9202736, "step": 6790 }, { "epoch": 0.2174956788937968, "grad_norm": 1.8117016553878784, "learning_rate": 4.6524545288501734e-05, "loss": 0.5017, "num_input_tokens_seen": 9209472, "step": 6795 }, { "epoch": 0.21765571986428525, "grad_norm": 1.1594256162643433, "learning_rate": 4.6519550109211506e-05, "loss": 0.4186, "num_input_tokens_seen": 9216336, "step": 6800 }, { "epoch": 0.21765571986428525, "eval_loss": 0.5422279834747314, "eval_runtime": 499.3881, "eval_samples_per_second": 27.806, "eval_steps_per_second": 13.903, "num_input_tokens_seen": 9216336, "step": 6800 }, { "epoch": 0.2178157608347737, "grad_norm": 0.8992421627044678, "learning_rate": 4.651455161133622e-05, "loss": 0.5606, "num_input_tokens_seen": 9222688, "step": 6805 }, { "epoch": 0.21797580180526216, "grad_norm": 1.6440964937210083, "learning_rate": 4.6509549795646704e-05, "loss": 0.4962, "num_input_tokens_seen": 9229472, "step": 6810 }, { "epoch": 0.2181358427757506, "grad_norm": 1.2564865350723267, "learning_rate": 4.6504544662914306e-05, "loss": 0.5591, "num_input_tokens_seen": 9236128, "step": 6815 }, { "epoch": 0.21829588374623904, "grad_norm": 1.3990319967269897, "learning_rate": 4.6499536213910876e-05, "loss": 0.5775, "num_input_tokens_seen": 9242832, "step": 6820 }, { "epoch": 0.21845592471672748, "grad_norm": 0.5736234188079834, "learning_rate": 4.6494524449408786e-05, "loss": 0.4909, "num_input_tokens_seen": 9249408, "step": 6825 }, { "epoch": 0.21861596568721592, "grad_norm": 1.1834681034088135, "learning_rate": 4.6489509370180903e-05, "loss": 0.4597, "num_input_tokens_seen": 9255968, "step": 6830 }, { "epoch": 0.21877600665770439, "grad_norm": 2.239495038986206, "learning_rate": 4.648449097700063e-05, "loss": 0.528, "num_input_tokens_seen": 9263136, "step": 6835 }, { "epoch": 0.21893604762819283, "grad_norm": 2.5173709392547607, "learning_rate": 4.647946927064185e-05, "loss": 0.6597, "num_input_tokens_seen": 9269744, "step": 6840 }, { "epoch": 0.21909608859868127, "grad_norm": 1.7258572578430176, "learning_rate": 4.647444425187898e-05, "loss": 0.581, "num_input_tokens_seen": 9276560, "step": 6845 }, { "epoch": 0.2192561295691697, "grad_norm": 1.9782707691192627, "learning_rate": 4.646941592148695e-05, "loss": 0.5515, "num_input_tokens_seen": 9283408, "step": 6850 }, { "epoch": 0.21941617053965815, "grad_norm": 1.2292060852050781, "learning_rate": 4.646438428024117e-05, "loss": 0.5748, "num_input_tokens_seen": 9290560, "step": 6855 }, { "epoch": 0.21957621151014659, "grad_norm": 1.1703166961669922, "learning_rate": 4.64593493289176e-05, "loss": 0.5204, "num_input_tokens_seen": 9297168, "step": 6860 }, { "epoch": 0.21973625248063505, "grad_norm": 1.8392573595046997, "learning_rate": 4.64543110682927e-05, "loss": 0.3508, "num_input_tokens_seen": 9303504, "step": 6865 }, { "epoch": 0.2198962934511235, "grad_norm": 2.1359987258911133, "learning_rate": 4.644926949914341e-05, "loss": 0.7515, "num_input_tokens_seen": 9310400, "step": 6870 }, { "epoch": 0.22005633442161193, "grad_norm": 0.7053820490837097, "learning_rate": 4.644422462224722e-05, "loss": 0.5475, "num_input_tokens_seen": 9317168, "step": 6875 }, { "epoch": 0.22021637539210037, "grad_norm": 1.4546183347702026, "learning_rate": 4.643917643838211e-05, "loss": 0.6937, "num_input_tokens_seen": 9324176, "step": 6880 }, { "epoch": 0.2203764163625888, "grad_norm": 1.820238709449768, "learning_rate": 4.6434124948326564e-05, "loss": 0.5235, "num_input_tokens_seen": 9331040, "step": 6885 }, { "epoch": 0.22053645733307728, "grad_norm": 1.243900179862976, "learning_rate": 4.6429070152859594e-05, "loss": 0.3675, "num_input_tokens_seen": 9337712, "step": 6890 }, { "epoch": 0.22069649830356572, "grad_norm": 1.9495902061462402, "learning_rate": 4.6424012052760714e-05, "loss": 0.5805, "num_input_tokens_seen": 9344624, "step": 6895 }, { "epoch": 0.22085653927405416, "grad_norm": 1.1992663145065308, "learning_rate": 4.6418950648809945e-05, "loss": 0.5749, "num_input_tokens_seen": 9351104, "step": 6900 }, { "epoch": 0.2210165802445426, "grad_norm": 1.488749623298645, "learning_rate": 4.641388594178782e-05, "loss": 0.4793, "num_input_tokens_seen": 9358160, "step": 6905 }, { "epoch": 0.22117662121503104, "grad_norm": 0.8287264108657837, "learning_rate": 4.640881793247538e-05, "loss": 0.5417, "num_input_tokens_seen": 9364592, "step": 6910 }, { "epoch": 0.2213366621855195, "grad_norm": 1.2906569242477417, "learning_rate": 4.6403746621654173e-05, "loss": 0.5377, "num_input_tokens_seen": 9371312, "step": 6915 }, { "epoch": 0.22149670315600795, "grad_norm": 0.8427451252937317, "learning_rate": 4.639867201010626e-05, "loss": 0.4224, "num_input_tokens_seen": 9377792, "step": 6920 }, { "epoch": 0.22165674412649639, "grad_norm": 1.166329026222229, "learning_rate": 4.6393594098614204e-05, "loss": 0.6615, "num_input_tokens_seen": 9384592, "step": 6925 }, { "epoch": 0.22181678509698483, "grad_norm": 0.9563876390457153, "learning_rate": 4.63885128879611e-05, "loss": 0.6422, "num_input_tokens_seen": 9391424, "step": 6930 }, { "epoch": 0.22197682606747327, "grad_norm": 0.7072016000747681, "learning_rate": 4.638342837893052e-05, "loss": 0.4913, "num_input_tokens_seen": 9398464, "step": 6935 }, { "epoch": 0.2221368670379617, "grad_norm": 1.4750087261199951, "learning_rate": 4.6378340572306565e-05, "loss": 0.4897, "num_input_tokens_seen": 9405488, "step": 6940 }, { "epoch": 0.22229690800845017, "grad_norm": 1.6060733795166016, "learning_rate": 4.6373249468873833e-05, "loss": 0.5161, "num_input_tokens_seen": 9412304, "step": 6945 }, { "epoch": 0.2224569489789386, "grad_norm": 2.148770332336426, "learning_rate": 4.636815506941744e-05, "loss": 0.4375, "num_input_tokens_seen": 9418976, "step": 6950 }, { "epoch": 0.22261698994942705, "grad_norm": 1.342275619506836, "learning_rate": 4.6363057374723004e-05, "loss": 0.5099, "num_input_tokens_seen": 9425472, "step": 6955 }, { "epoch": 0.2227770309199155, "grad_norm": 0.772084653377533, "learning_rate": 4.635795638557666e-05, "loss": 0.6495, "num_input_tokens_seen": 9431920, "step": 6960 }, { "epoch": 0.22293707189040393, "grad_norm": 0.7171277403831482, "learning_rate": 4.635285210276504e-05, "loss": 0.671, "num_input_tokens_seen": 9438928, "step": 6965 }, { "epoch": 0.2230971128608924, "grad_norm": 1.158316731452942, "learning_rate": 4.6347744527075295e-05, "loss": 0.4987, "num_input_tokens_seen": 9445680, "step": 6970 }, { "epoch": 0.22325715383138084, "grad_norm": 1.6410486698150635, "learning_rate": 4.634263365929506e-05, "loss": 0.4903, "num_input_tokens_seen": 9452272, "step": 6975 }, { "epoch": 0.22341719480186928, "grad_norm": 1.1982887983322144, "learning_rate": 4.6337519500212515e-05, "loss": 0.4613, "num_input_tokens_seen": 9458672, "step": 6980 }, { "epoch": 0.22357723577235772, "grad_norm": 1.8238122463226318, "learning_rate": 4.633240205061632e-05, "loss": 0.555, "num_input_tokens_seen": 9465168, "step": 6985 }, { "epoch": 0.22373727674284616, "grad_norm": 0.9847112894058228, "learning_rate": 4.632728131129565e-05, "loss": 0.436, "num_input_tokens_seen": 9471872, "step": 6990 }, { "epoch": 0.22389731771333463, "grad_norm": 3.1304681301116943, "learning_rate": 4.632215728304018e-05, "loss": 0.7363, "num_input_tokens_seen": 9478736, "step": 6995 }, { "epoch": 0.22405735868382307, "grad_norm": 1.1022239923477173, "learning_rate": 4.63170299666401e-05, "loss": 0.4501, "num_input_tokens_seen": 9485568, "step": 7000 }, { "epoch": 0.22405735868382307, "eval_loss": 0.5398515462875366, "eval_runtime": 499.2745, "eval_samples_per_second": 27.812, "eval_steps_per_second": 13.906, "num_input_tokens_seen": 9485568, "step": 7000 }, { "epoch": 0.2242173996543115, "grad_norm": 2.0127410888671875, "learning_rate": 4.631189936288612e-05, "loss": 0.54, "num_input_tokens_seen": 9492112, "step": 7005 }, { "epoch": 0.22437744062479995, "grad_norm": 1.7292139530181885, "learning_rate": 4.630676547256944e-05, "loss": 0.4729, "num_input_tokens_seen": 9498816, "step": 7010 }, { "epoch": 0.22453748159528839, "grad_norm": 0.6568143963813782, "learning_rate": 4.630162829648176e-05, "loss": 0.4697, "num_input_tokens_seen": 9506112, "step": 7015 }, { "epoch": 0.22469752256577685, "grad_norm": 0.8667856454849243, "learning_rate": 4.629648783541531e-05, "loss": 0.736, "num_input_tokens_seen": 9512864, "step": 7020 }, { "epoch": 0.2248575635362653, "grad_norm": 2.4077236652374268, "learning_rate": 4.6291344090162804e-05, "loss": 0.542, "num_input_tokens_seen": 9520224, "step": 7025 }, { "epoch": 0.22501760450675373, "grad_norm": 0.8394725322723389, "learning_rate": 4.628619706151748e-05, "loss": 0.5645, "num_input_tokens_seen": 9526848, "step": 7030 }, { "epoch": 0.22517764547724217, "grad_norm": 1.0622202157974243, "learning_rate": 4.628104675027306e-05, "loss": 0.3751, "num_input_tokens_seen": 9533520, "step": 7035 }, { "epoch": 0.2253376864477306, "grad_norm": 1.6154053211212158, "learning_rate": 4.6275893157223805e-05, "loss": 0.3742, "num_input_tokens_seen": 9540624, "step": 7040 }, { "epoch": 0.22549772741821905, "grad_norm": 0.9070065021514893, "learning_rate": 4.627073628316445e-05, "loss": 0.6123, "num_input_tokens_seen": 9547712, "step": 7045 }, { "epoch": 0.22565776838870752, "grad_norm": 0.6116254329681396, "learning_rate": 4.626557612889026e-05, "loss": 0.3851, "num_input_tokens_seen": 9554160, "step": 7050 }, { "epoch": 0.22581780935919596, "grad_norm": 1.4316489696502686, "learning_rate": 4.626041269519699e-05, "loss": 0.5012, "num_input_tokens_seen": 9560848, "step": 7055 }, { "epoch": 0.2259778503296844, "grad_norm": 1.148542046546936, "learning_rate": 4.6255245982880905e-05, "loss": 0.5848, "num_input_tokens_seen": 9567808, "step": 7060 }, { "epoch": 0.22613789130017284, "grad_norm": 1.4129564762115479, "learning_rate": 4.625007599273879e-05, "loss": 0.491, "num_input_tokens_seen": 9574864, "step": 7065 }, { "epoch": 0.22629793227066128, "grad_norm": 0.8636768460273743, "learning_rate": 4.6244902725567895e-05, "loss": 0.4797, "num_input_tokens_seen": 9581776, "step": 7070 }, { "epoch": 0.22645797324114975, "grad_norm": 1.2321527004241943, "learning_rate": 4.6239726182166024e-05, "loss": 0.523, "num_input_tokens_seen": 9588208, "step": 7075 }, { "epoch": 0.22661801421163819, "grad_norm": 1.4085458517074585, "learning_rate": 4.623454636333147e-05, "loss": 0.4483, "num_input_tokens_seen": 9595024, "step": 7080 }, { "epoch": 0.22677805518212663, "grad_norm": 0.7179081439971924, "learning_rate": 4.622936326986301e-05, "loss": 0.6067, "num_input_tokens_seen": 9601680, "step": 7085 }, { "epoch": 0.22693809615261507, "grad_norm": 1.4834625720977783, "learning_rate": 4.6224176902559946e-05, "loss": 0.408, "num_input_tokens_seen": 9608560, "step": 7090 }, { "epoch": 0.2270981371231035, "grad_norm": 1.0459877252578735, "learning_rate": 4.621898726222209e-05, "loss": 0.4121, "num_input_tokens_seen": 9615136, "step": 7095 }, { "epoch": 0.22725817809359197, "grad_norm": 1.0778549909591675, "learning_rate": 4.6213794349649744e-05, "loss": 0.4116, "num_input_tokens_seen": 9622112, "step": 7100 }, { "epoch": 0.2274182190640804, "grad_norm": 0.7998549938201904, "learning_rate": 4.6208598165643715e-05, "loss": 0.5428, "num_input_tokens_seen": 9628576, "step": 7105 }, { "epoch": 0.22757826003456885, "grad_norm": 1.2425010204315186, "learning_rate": 4.620339871100533e-05, "loss": 0.4895, "num_input_tokens_seen": 9635392, "step": 7110 }, { "epoch": 0.2277383010050573, "grad_norm": 1.5886095762252808, "learning_rate": 4.6198195986536394e-05, "loss": 0.775, "num_input_tokens_seen": 9641936, "step": 7115 }, { "epoch": 0.22789834197554573, "grad_norm": 1.9605600833892822, "learning_rate": 4.619298999303926e-05, "loss": 0.4415, "num_input_tokens_seen": 9648672, "step": 7120 }, { "epoch": 0.22805838294603417, "grad_norm": 0.8239843249320984, "learning_rate": 4.618778073131673e-05, "loss": 0.5701, "num_input_tokens_seen": 9655568, "step": 7125 }, { "epoch": 0.22821842391652264, "grad_norm": 1.3281946182250977, "learning_rate": 4.618256820217215e-05, "loss": 0.512, "num_input_tokens_seen": 9662016, "step": 7130 }, { "epoch": 0.22837846488701108, "grad_norm": 1.0523624420166016, "learning_rate": 4.617735240640936e-05, "loss": 0.4615, "num_input_tokens_seen": 9668592, "step": 7135 }, { "epoch": 0.22853850585749952, "grad_norm": 1.692785620689392, "learning_rate": 4.6172133344832705e-05, "loss": 0.6871, "num_input_tokens_seen": 9675504, "step": 7140 }, { "epoch": 0.22869854682798796, "grad_norm": 1.652413010597229, "learning_rate": 4.6166911018247004e-05, "loss": 0.6198, "num_input_tokens_seen": 9682224, "step": 7145 }, { "epoch": 0.2288585877984764, "grad_norm": 1.7675650119781494, "learning_rate": 4.616168542745764e-05, "loss": 0.676, "num_input_tokens_seen": 9688864, "step": 7150 }, { "epoch": 0.22901862876896487, "grad_norm": 1.7590374946594238, "learning_rate": 4.6156456573270446e-05, "loss": 0.514, "num_input_tokens_seen": 9696368, "step": 7155 }, { "epoch": 0.2291786697394533, "grad_norm": 2.371413230895996, "learning_rate": 4.615122445649177e-05, "loss": 0.6633, "num_input_tokens_seen": 9703136, "step": 7160 }, { "epoch": 0.22933871070994175, "grad_norm": 0.8155451416969299, "learning_rate": 4.6145989077928486e-05, "loss": 0.4947, "num_input_tokens_seen": 9709600, "step": 7165 }, { "epoch": 0.22949875168043019, "grad_norm": 2.451124429702759, "learning_rate": 4.6140750438387953e-05, "loss": 0.4719, "num_input_tokens_seen": 9716336, "step": 7170 }, { "epoch": 0.22965879265091863, "grad_norm": 0.972311794757843, "learning_rate": 4.613550853867803e-05, "loss": 0.3969, "num_input_tokens_seen": 9723232, "step": 7175 }, { "epoch": 0.2298188336214071, "grad_norm": 1.6020489931106567, "learning_rate": 4.613026337960708e-05, "loss": 0.4646, "num_input_tokens_seen": 9730128, "step": 7180 }, { "epoch": 0.22997887459189553, "grad_norm": 1.9207465648651123, "learning_rate": 4.612501496198398e-05, "loss": 0.5447, "num_input_tokens_seen": 9737104, "step": 7185 }, { "epoch": 0.23013891556238397, "grad_norm": 0.6608176827430725, "learning_rate": 4.61197632866181e-05, "loss": 0.3156, "num_input_tokens_seen": 9743872, "step": 7190 }, { "epoch": 0.2302989565328724, "grad_norm": 1.491390347480774, "learning_rate": 4.611450835431931e-05, "loss": 0.5355, "num_input_tokens_seen": 9750784, "step": 7195 }, { "epoch": 0.23045899750336085, "grad_norm": 1.3720523118972778, "learning_rate": 4.6109250165898e-05, "loss": 0.4739, "num_input_tokens_seen": 9758160, "step": 7200 }, { "epoch": 0.23045899750336085, "eval_loss": 0.5383065342903137, "eval_runtime": 499.2991, "eval_samples_per_second": 27.811, "eval_steps_per_second": 13.905, "num_input_tokens_seen": 9758160, "step": 7200 }, { "epoch": 0.2306190384738493, "grad_norm": 1.013185739517212, "learning_rate": 4.610398872216503e-05, "loss": 0.5536, "num_input_tokens_seen": 9765360, "step": 7205 }, { "epoch": 0.23077907944433776, "grad_norm": 1.5094572305679321, "learning_rate": 4.6098724023931796e-05, "loss": 0.3831, "num_input_tokens_seen": 9772176, "step": 7210 }, { "epoch": 0.2309391204148262, "grad_norm": 1.4985663890838623, "learning_rate": 4.609345607201017e-05, "loss": 0.55, "num_input_tokens_seen": 9778928, "step": 7215 }, { "epoch": 0.23109916138531464, "grad_norm": 1.4270066022872925, "learning_rate": 4.608818486721254e-05, "loss": 0.4506, "num_input_tokens_seen": 9785424, "step": 7220 }, { "epoch": 0.23125920235580308, "grad_norm": 1.9932390451431274, "learning_rate": 4.608291041035179e-05, "loss": 0.4873, "num_input_tokens_seen": 9792096, "step": 7225 }, { "epoch": 0.23141924332629152, "grad_norm": 1.1999074220657349, "learning_rate": 4.607763270224132e-05, "loss": 0.4803, "num_input_tokens_seen": 9798944, "step": 7230 }, { "epoch": 0.23157928429677999, "grad_norm": 0.6567263007164001, "learning_rate": 4.6072351743695e-05, "loss": 0.4606, "num_input_tokens_seen": 9805552, "step": 7235 }, { "epoch": 0.23173932526726843, "grad_norm": 1.1438443660736084, "learning_rate": 4.606706753552723e-05, "loss": 0.4338, "num_input_tokens_seen": 9812592, "step": 7240 }, { "epoch": 0.23189936623775687, "grad_norm": 0.5389841198921204, "learning_rate": 4.6061780078552906e-05, "loss": 0.3985, "num_input_tokens_seen": 9819184, "step": 7245 }, { "epoch": 0.2320594072082453, "grad_norm": 1.2328115701675415, "learning_rate": 4.605648937358742e-05, "loss": 0.6145, "num_input_tokens_seen": 9825984, "step": 7250 }, { "epoch": 0.23221944817873375, "grad_norm": 0.9432075619697571, "learning_rate": 4.605119542144665e-05, "loss": 0.4825, "num_input_tokens_seen": 9832736, "step": 7255 }, { "epoch": 0.2323794891492222, "grad_norm": 1.1249159574508667, "learning_rate": 4.604589822294701e-05, "loss": 0.3831, "num_input_tokens_seen": 9839392, "step": 7260 }, { "epoch": 0.23253953011971065, "grad_norm": 1.0060701370239258, "learning_rate": 4.604059777890537e-05, "loss": 0.4717, "num_input_tokens_seen": 9846512, "step": 7265 }, { "epoch": 0.2326995710901991, "grad_norm": 1.279415488243103, "learning_rate": 4.6035294090139145e-05, "loss": 0.4007, "num_input_tokens_seen": 9853680, "step": 7270 }, { "epoch": 0.23285961206068753, "grad_norm": 0.8974608778953552, "learning_rate": 4.6029987157466226e-05, "loss": 0.4323, "num_input_tokens_seen": 9860448, "step": 7275 }, { "epoch": 0.23301965303117597, "grad_norm": 0.8495081067085266, "learning_rate": 4.602467698170502e-05, "loss": 0.6205, "num_input_tokens_seen": 9867712, "step": 7280 }, { "epoch": 0.23317969400166444, "grad_norm": 1.1730116605758667, "learning_rate": 4.601936356367439e-05, "loss": 0.5843, "num_input_tokens_seen": 9874528, "step": 7285 }, { "epoch": 0.23333973497215288, "grad_norm": 1.6725258827209473, "learning_rate": 4.601404690419377e-05, "loss": 0.6499, "num_input_tokens_seen": 9881392, "step": 7290 }, { "epoch": 0.23349977594264132, "grad_norm": 0.9711195826530457, "learning_rate": 4.600872700408303e-05, "loss": 0.6355, "num_input_tokens_seen": 9888144, "step": 7295 }, { "epoch": 0.23365981691312976, "grad_norm": 2.289893627166748, "learning_rate": 4.600340386416258e-05, "loss": 0.6653, "num_input_tokens_seen": 9894576, "step": 7300 }, { "epoch": 0.2338198578836182, "grad_norm": 0.7499582767486572, "learning_rate": 4.5998077485253296e-05, "loss": 0.3612, "num_input_tokens_seen": 9901408, "step": 7305 }, { "epoch": 0.23397989885410664, "grad_norm": 1.3398569822311401, "learning_rate": 4.59927478681766e-05, "loss": 0.4633, "num_input_tokens_seen": 9908096, "step": 7310 }, { "epoch": 0.2341399398245951, "grad_norm": 2.016604423522949, "learning_rate": 4.5987415013754366e-05, "loss": 0.7239, "num_input_tokens_seen": 9915376, "step": 7315 }, { "epoch": 0.23429998079508355, "grad_norm": 1.1508654356002808, "learning_rate": 4.598207892280899e-05, "loss": 0.5398, "num_input_tokens_seen": 9922000, "step": 7320 }, { "epoch": 0.23446002176557199, "grad_norm": 0.9655758738517761, "learning_rate": 4.597673959616337e-05, "loss": 0.6603, "num_input_tokens_seen": 9928528, "step": 7325 }, { "epoch": 0.23462006273606043, "grad_norm": 1.2707858085632324, "learning_rate": 4.597139703464089e-05, "loss": 0.5089, "num_input_tokens_seen": 9935360, "step": 7330 }, { "epoch": 0.23478010370654886, "grad_norm": 0.6787548661231995, "learning_rate": 4.596605123906545e-05, "loss": 0.3491, "num_input_tokens_seen": 9941936, "step": 7335 }, { "epoch": 0.23494014467703733, "grad_norm": 1.4716780185699463, "learning_rate": 4.596070221026143e-05, "loss": 0.4648, "num_input_tokens_seen": 9948288, "step": 7340 }, { "epoch": 0.23510018564752577, "grad_norm": 1.7128762006759644, "learning_rate": 4.595534994905372e-05, "loss": 0.45, "num_input_tokens_seen": 9954992, "step": 7345 }, { "epoch": 0.2352602266180142, "grad_norm": 1.9289724826812744, "learning_rate": 4.594999445626771e-05, "loss": 0.4558, "num_input_tokens_seen": 9961616, "step": 7350 }, { "epoch": 0.23542026758850265, "grad_norm": 1.7530616521835327, "learning_rate": 4.5944635732729276e-05, "loss": 0.5923, "num_input_tokens_seen": 9968064, "step": 7355 }, { "epoch": 0.2355803085589911, "grad_norm": 1.435866355895996, "learning_rate": 4.5939273779264804e-05, "loss": 0.6288, "num_input_tokens_seen": 9974832, "step": 7360 }, { "epoch": 0.23574034952947956, "grad_norm": 1.7785533666610718, "learning_rate": 4.593390859670118e-05, "loss": 0.6262, "num_input_tokens_seen": 9981536, "step": 7365 }, { "epoch": 0.235900390499968, "grad_norm": 1.560059905052185, "learning_rate": 4.5928540185865776e-05, "loss": 0.4803, "num_input_tokens_seen": 9988128, "step": 7370 }, { "epoch": 0.23606043147045644, "grad_norm": 1.7590683698654175, "learning_rate": 4.592316854758648e-05, "loss": 0.6752, "num_input_tokens_seen": 9994704, "step": 7375 }, { "epoch": 0.23622047244094488, "grad_norm": 1.1824499368667603, "learning_rate": 4.5917793682691646e-05, "loss": 0.6521, "num_input_tokens_seen": 10001280, "step": 7380 }, { "epoch": 0.23638051341143332, "grad_norm": 0.879286527633667, "learning_rate": 4.5912415592010164e-05, "loss": 0.4696, "num_input_tokens_seen": 10007760, "step": 7385 }, { "epoch": 0.23654055438192176, "grad_norm": 1.3359671831130981, "learning_rate": 4.5907034276371386e-05, "loss": 0.5371, "num_input_tokens_seen": 10014032, "step": 7390 }, { "epoch": 0.23670059535241023, "grad_norm": 0.9076265096664429, "learning_rate": 4.5901649736605196e-05, "loss": 0.2424, "num_input_tokens_seen": 10021232, "step": 7395 }, { "epoch": 0.23686063632289867, "grad_norm": 1.1304335594177246, "learning_rate": 4.589626197354195e-05, "loss": 0.6005, "num_input_tokens_seen": 10028256, "step": 7400 }, { "epoch": 0.23686063632289867, "eval_loss": 0.536597490310669, "eval_runtime": 500.1145, "eval_samples_per_second": 27.766, "eval_steps_per_second": 13.883, "num_input_tokens_seen": 10028256, "step": 7400 }, { "epoch": 0.2370206772933871, "grad_norm": 0.9677615761756897, "learning_rate": 4.5890870988012504e-05, "loss": 0.4489, "num_input_tokens_seen": 10034672, "step": 7405 }, { "epoch": 0.23718071826387555, "grad_norm": 1.2914496660232544, "learning_rate": 4.5885476780848226e-05, "loss": 0.6068, "num_input_tokens_seen": 10041936, "step": 7410 }, { "epoch": 0.23734075923436398, "grad_norm": 0.8231095671653748, "learning_rate": 4.5880079352880964e-05, "loss": 0.3875, "num_input_tokens_seen": 10048704, "step": 7415 }, { "epoch": 0.23750080020485245, "grad_norm": 1.398574709892273, "learning_rate": 4.5874678704943065e-05, "loss": 0.4949, "num_input_tokens_seen": 10055696, "step": 7420 }, { "epoch": 0.2376608411753409, "grad_norm": 1.2567541599273682, "learning_rate": 4.5869274837867394e-05, "loss": 0.5904, "num_input_tokens_seen": 10062272, "step": 7425 }, { "epoch": 0.23782088214582933, "grad_norm": 3.0228826999664307, "learning_rate": 4.5863867752487275e-05, "loss": 0.4529, "num_input_tokens_seen": 10068816, "step": 7430 }, { "epoch": 0.23798092311631777, "grad_norm": 0.9411541223526001, "learning_rate": 4.5858457449636554e-05, "loss": 0.4202, "num_input_tokens_seen": 10075936, "step": 7435 }, { "epoch": 0.2381409640868062, "grad_norm": 2.2166619300842285, "learning_rate": 4.5853043930149574e-05, "loss": 0.5704, "num_input_tokens_seen": 10082720, "step": 7440 }, { "epoch": 0.23830100505729468, "grad_norm": 2.509620428085327, "learning_rate": 4.584762719486117e-05, "loss": 0.5122, "num_input_tokens_seen": 10089264, "step": 7445 }, { "epoch": 0.23846104602778312, "grad_norm": 1.9582653045654297, "learning_rate": 4.584220724460665e-05, "loss": 0.6151, "num_input_tokens_seen": 10095744, "step": 7450 }, { "epoch": 0.23862108699827156, "grad_norm": 1.6258671283721924, "learning_rate": 4.5836784080221865e-05, "loss": 0.5124, "num_input_tokens_seen": 10102432, "step": 7455 }, { "epoch": 0.23878112796876, "grad_norm": 0.9754686951637268, "learning_rate": 4.583135770254312e-05, "loss": 0.616, "num_input_tokens_seen": 10109648, "step": 7460 }, { "epoch": 0.23894116893924844, "grad_norm": 0.6965294480323792, "learning_rate": 4.5825928112407236e-05, "loss": 0.4906, "num_input_tokens_seen": 10116384, "step": 7465 }, { "epoch": 0.2391012099097369, "grad_norm": 0.47072485089302063, "learning_rate": 4.582049531065152e-05, "loss": 0.3954, "num_input_tokens_seen": 10123424, "step": 7470 }, { "epoch": 0.23926125088022535, "grad_norm": 2.281127452850342, "learning_rate": 4.5815059298113783e-05, "loss": 0.4258, "num_input_tokens_seen": 10129968, "step": 7475 }, { "epoch": 0.23942129185071379, "grad_norm": 1.568251132965088, "learning_rate": 4.580962007563232e-05, "loss": 0.4399, "num_input_tokens_seen": 10136784, "step": 7480 }, { "epoch": 0.23958133282120223, "grad_norm": 0.9418609738349915, "learning_rate": 4.5804177644045935e-05, "loss": 0.4283, "num_input_tokens_seen": 10143600, "step": 7485 }, { "epoch": 0.23974137379169067, "grad_norm": 0.5623632073402405, "learning_rate": 4.579873200419391e-05, "loss": 0.4136, "num_input_tokens_seen": 10150448, "step": 7490 }, { "epoch": 0.2399014147621791, "grad_norm": 1.655792236328125, "learning_rate": 4.5793283156916046e-05, "loss": 0.5708, "num_input_tokens_seen": 10157024, "step": 7495 }, { "epoch": 0.24006145573266757, "grad_norm": 1.4846251010894775, "learning_rate": 4.578783110305261e-05, "loss": 0.4602, "num_input_tokens_seen": 10164208, "step": 7500 }, { "epoch": 0.240221496703156, "grad_norm": 1.274402379989624, "learning_rate": 4.578237584344438e-05, "loss": 0.504, "num_input_tokens_seen": 10170976, "step": 7505 }, { "epoch": 0.24038153767364445, "grad_norm": 1.1090466976165771, "learning_rate": 4.577691737893263e-05, "loss": 0.5425, "num_input_tokens_seen": 10177840, "step": 7510 }, { "epoch": 0.2405415786441329, "grad_norm": 0.9569831490516663, "learning_rate": 4.577145571035912e-05, "loss": 0.5803, "num_input_tokens_seen": 10184560, "step": 7515 }, { "epoch": 0.24070161961462133, "grad_norm": 0.6900236010551453, "learning_rate": 4.576599083856611e-05, "loss": 0.5616, "num_input_tokens_seen": 10191456, "step": 7520 }, { "epoch": 0.2408616605851098, "grad_norm": 0.5075759291648865, "learning_rate": 4.576052276439635e-05, "loss": 0.5678, "num_input_tokens_seen": 10198000, "step": 7525 }, { "epoch": 0.24102170155559824, "grad_norm": 1.0180484056472778, "learning_rate": 4.575505148869308e-05, "loss": 0.512, "num_input_tokens_seen": 10204800, "step": 7530 }, { "epoch": 0.24118174252608668, "grad_norm": 0.7440587878227234, "learning_rate": 4.574957701230006e-05, "loss": 0.4405, "num_input_tokens_seen": 10211808, "step": 7535 }, { "epoch": 0.24134178349657512, "grad_norm": 1.7569961547851562, "learning_rate": 4.57440993360615e-05, "loss": 0.4168, "num_input_tokens_seen": 10219008, "step": 7540 }, { "epoch": 0.24150182446706356, "grad_norm": 0.584306001663208, "learning_rate": 4.5738618460822134e-05, "loss": 0.5166, "num_input_tokens_seen": 10225872, "step": 7545 }, { "epoch": 0.24166186543755203, "grad_norm": 0.7648707032203674, "learning_rate": 4.573313438742719e-05, "loss": 0.43, "num_input_tokens_seen": 10232704, "step": 7550 }, { "epoch": 0.24182190640804047, "grad_norm": 1.5131372213363647, "learning_rate": 4.5727647116722374e-05, "loss": 0.5033, "num_input_tokens_seen": 10239744, "step": 7555 }, { "epoch": 0.2419819473785289, "grad_norm": 0.7131675481796265, "learning_rate": 4.5722156649553884e-05, "loss": 0.5647, "num_input_tokens_seen": 10246704, "step": 7560 }, { "epoch": 0.24214198834901735, "grad_norm": 1.31417715549469, "learning_rate": 4.571666298676843e-05, "loss": 0.5011, "num_input_tokens_seen": 10253264, "step": 7565 }, { "epoch": 0.24230202931950579, "grad_norm": 2.2475597858428955, "learning_rate": 4.571116612921321e-05, "loss": 0.7317, "num_input_tokens_seen": 10259712, "step": 7570 }, { "epoch": 0.24246207028999422, "grad_norm": 1.6092478036880493, "learning_rate": 4.57056660777359e-05, "loss": 0.5465, "num_input_tokens_seen": 10266400, "step": 7575 }, { "epoch": 0.2426221112604827, "grad_norm": 1.6207998991012573, "learning_rate": 4.5700162833184666e-05, "loss": 0.4891, "num_input_tokens_seen": 10273568, "step": 7580 }, { "epoch": 0.24278215223097113, "grad_norm": 0.9088239669799805, "learning_rate": 4.5694656396408195e-05, "loss": 0.4659, "num_input_tokens_seen": 10280432, "step": 7585 }, { "epoch": 0.24294219320145957, "grad_norm": 0.4658275842666626, "learning_rate": 4.5689146768255646e-05, "loss": 0.5774, "num_input_tokens_seen": 10286880, "step": 7590 }, { "epoch": 0.243102234171948, "grad_norm": 1.8107444047927856, "learning_rate": 4.568363394957667e-05, "loss": 0.6508, "num_input_tokens_seen": 10293984, "step": 7595 }, { "epoch": 0.24326227514243645, "grad_norm": 1.5462727546691895, "learning_rate": 4.567811794122141e-05, "loss": 0.5946, "num_input_tokens_seen": 10300544, "step": 7600 }, { "epoch": 0.24326227514243645, "eval_loss": 0.5359199047088623, "eval_runtime": 499.2246, "eval_samples_per_second": 27.815, "eval_steps_per_second": 13.908, "num_input_tokens_seen": 10300544, "step": 7600 }, { "epoch": 0.24342231611292492, "grad_norm": 1.6730244159698486, "learning_rate": 4.56725987440405e-05, "loss": 0.5504, "num_input_tokens_seen": 10307504, "step": 7605 }, { "epoch": 0.24358235708341336, "grad_norm": 0.6525645852088928, "learning_rate": 4.566707635888508e-05, "loss": 0.5038, "num_input_tokens_seen": 10314432, "step": 7610 }, { "epoch": 0.2437423980539018, "grad_norm": 1.4655321836471558, "learning_rate": 4.566155078660677e-05, "loss": 0.4752, "num_input_tokens_seen": 10321472, "step": 7615 }, { "epoch": 0.24390243902439024, "grad_norm": 2.6231024265289307, "learning_rate": 4.565602202805768e-05, "loss": 0.6004, "num_input_tokens_seen": 10328160, "step": 7620 }, { "epoch": 0.24406247999487868, "grad_norm": 1.1495932340621948, "learning_rate": 4.56504900840904e-05, "loss": 0.4972, "num_input_tokens_seen": 10335360, "step": 7625 }, { "epoch": 0.24422252096536715, "grad_norm": 1.4101595878601074, "learning_rate": 4.564495495555805e-05, "loss": 0.4719, "num_input_tokens_seen": 10341888, "step": 7630 }, { "epoch": 0.24438256193585559, "grad_norm": 3.005585193634033, "learning_rate": 4.5639416643314204e-05, "loss": 0.5916, "num_input_tokens_seen": 10348528, "step": 7635 }, { "epoch": 0.24454260290634403, "grad_norm": 1.42691969871521, "learning_rate": 4.5633875148212946e-05, "loss": 0.6682, "num_input_tokens_seen": 10355440, "step": 7640 }, { "epoch": 0.24470264387683247, "grad_norm": 1.0711387395858765, "learning_rate": 4.562833047110883e-05, "loss": 0.609, "num_input_tokens_seen": 10362592, "step": 7645 }, { "epoch": 0.2448626848473209, "grad_norm": 1.102066159248352, "learning_rate": 4.5622782612856923e-05, "loss": 0.5236, "num_input_tokens_seen": 10369392, "step": 7650 }, { "epoch": 0.24502272581780937, "grad_norm": 1.147270917892456, "learning_rate": 4.561723157431278e-05, "loss": 0.5102, "num_input_tokens_seen": 10376144, "step": 7655 }, { "epoch": 0.2451827667882978, "grad_norm": 0.7662567496299744, "learning_rate": 4.5611677356332435e-05, "loss": 0.431, "num_input_tokens_seen": 10383136, "step": 7660 }, { "epoch": 0.24534280775878625, "grad_norm": 0.7765305638313293, "learning_rate": 4.560611995977242e-05, "loss": 0.3424, "num_input_tokens_seen": 10389968, "step": 7665 }, { "epoch": 0.2455028487292747, "grad_norm": 0.6752556562423706, "learning_rate": 4.560055938548975e-05, "loss": 0.4653, "num_input_tokens_seen": 10396528, "step": 7670 }, { "epoch": 0.24566288969976313, "grad_norm": 1.2438746690750122, "learning_rate": 4.5594995634341944e-05, "loss": 0.4608, "num_input_tokens_seen": 10403664, "step": 7675 }, { "epoch": 0.24582293067025157, "grad_norm": 1.3232313394546509, "learning_rate": 4.5589428707187e-05, "loss": 0.4689, "num_input_tokens_seen": 10410256, "step": 7680 }, { "epoch": 0.24598297164074004, "grad_norm": 1.214390516281128, "learning_rate": 4.55838586048834e-05, "loss": 0.43, "num_input_tokens_seen": 10416560, "step": 7685 }, { "epoch": 0.24614301261122848, "grad_norm": 1.5305570363998413, "learning_rate": 4.557828532829013e-05, "loss": 0.608, "num_input_tokens_seen": 10423296, "step": 7690 }, { "epoch": 0.24630305358171692, "grad_norm": 1.6248533725738525, "learning_rate": 4.557270887826667e-05, "loss": 0.4233, "num_input_tokens_seen": 10430192, "step": 7695 }, { "epoch": 0.24646309455220536, "grad_norm": 0.7047428488731384, "learning_rate": 4.556712925567296e-05, "loss": 0.4184, "num_input_tokens_seen": 10437136, "step": 7700 }, { "epoch": 0.2466231355226938, "grad_norm": 1.01638925075531, "learning_rate": 4.5561546461369454e-05, "loss": 0.451, "num_input_tokens_seen": 10444176, "step": 7705 }, { "epoch": 0.24678317649318227, "grad_norm": 1.2871638536453247, "learning_rate": 4.55559604962171e-05, "loss": 0.4777, "num_input_tokens_seen": 10451504, "step": 7710 }, { "epoch": 0.2469432174636707, "grad_norm": 0.7652977108955383, "learning_rate": 4.55503713610773e-05, "loss": 0.3716, "num_input_tokens_seen": 10458288, "step": 7715 }, { "epoch": 0.24710325843415915, "grad_norm": 1.2268036603927612, "learning_rate": 4.5544779056812e-05, "loss": 0.3783, "num_input_tokens_seen": 10464992, "step": 7720 }, { "epoch": 0.24726329940464759, "grad_norm": 0.7304170727729797, "learning_rate": 4.553918358428358e-05, "loss": 0.5186, "num_input_tokens_seen": 10471520, "step": 7725 }, { "epoch": 0.24742334037513602, "grad_norm": 1.4225326776504517, "learning_rate": 4.553358494435494e-05, "loss": 0.4168, "num_input_tokens_seen": 10478032, "step": 7730 }, { "epoch": 0.2475833813456245, "grad_norm": 1.0462071895599365, "learning_rate": 4.5527983137889464e-05, "loss": 0.663, "num_input_tokens_seen": 10484864, "step": 7735 }, { "epoch": 0.24774342231611293, "grad_norm": 1.2546560764312744, "learning_rate": 4.5522378165751015e-05, "loss": 0.4679, "num_input_tokens_seen": 10492272, "step": 7740 }, { "epoch": 0.24790346328660137, "grad_norm": 0.6911768317222595, "learning_rate": 4.5516770028803954e-05, "loss": 0.4494, "num_input_tokens_seen": 10499104, "step": 7745 }, { "epoch": 0.2480635042570898, "grad_norm": 2.726475954055786, "learning_rate": 4.5511158727913116e-05, "loss": 0.7343, "num_input_tokens_seen": 10505808, "step": 7750 }, { "epoch": 0.24822354522757825, "grad_norm": 1.2297518253326416, "learning_rate": 4.5505544263943856e-05, "loss": 0.4694, "num_input_tokens_seen": 10513200, "step": 7755 }, { "epoch": 0.2483835861980667, "grad_norm": 0.6116795539855957, "learning_rate": 4.549992663776197e-05, "loss": 0.4177, "num_input_tokens_seen": 10520000, "step": 7760 }, { "epoch": 0.24854362716855516, "grad_norm": 3.8296895027160645, "learning_rate": 4.5494305850233786e-05, "loss": 0.8581, "num_input_tokens_seen": 10526320, "step": 7765 }, { "epoch": 0.2487036681390436, "grad_norm": 1.0428571701049805, "learning_rate": 4.5488681902226094e-05, "loss": 0.5415, "num_input_tokens_seen": 10533280, "step": 7770 }, { "epoch": 0.24886370910953204, "grad_norm": 0.7656095623970032, "learning_rate": 4.5483054794606174e-05, "loss": 0.5279, "num_input_tokens_seen": 10540736, "step": 7775 }, { "epoch": 0.24902375008002048, "grad_norm": 1.0327565670013428, "learning_rate": 4.547742452824179e-05, "loss": 0.453, "num_input_tokens_seen": 10547264, "step": 7780 }, { "epoch": 0.24918379105050892, "grad_norm": 0.6005085706710815, "learning_rate": 4.5471791104001215e-05, "loss": 0.4001, "num_input_tokens_seen": 10554048, "step": 7785 }, { "epoch": 0.24934383202099739, "grad_norm": 1.3737940788269043, "learning_rate": 4.546615452275319e-05, "loss": 0.5412, "num_input_tokens_seen": 10560928, "step": 7790 }, { "epoch": 0.24950387299148583, "grad_norm": 1.4989463090896606, "learning_rate": 4.5460514785366944e-05, "loss": 0.7869, "num_input_tokens_seen": 10567408, "step": 7795 }, { "epoch": 0.24966391396197427, "grad_norm": 0.9101086854934692, "learning_rate": 4.545487189271219e-05, "loss": 0.4071, "num_input_tokens_seen": 10574192, "step": 7800 }, { "epoch": 0.24966391396197427, "eval_loss": 0.5330924391746521, "eval_runtime": 499.3538, "eval_samples_per_second": 27.808, "eval_steps_per_second": 13.904, "num_input_tokens_seen": 10574192, "step": 7800 }, { "epoch": 0.2498239549324627, "grad_norm": 3.2908220291137695, "learning_rate": 4.544922584565914e-05, "loss": 0.4914, "num_input_tokens_seen": 10580624, "step": 7805 }, { "epoch": 0.24998399590295114, "grad_norm": 1.148220181465149, "learning_rate": 4.544357664507848e-05, "loss": 0.4599, "num_input_tokens_seen": 10587536, "step": 7810 }, { "epoch": 0.2501440368734396, "grad_norm": 1.0043233633041382, "learning_rate": 4.54379242918414e-05, "loss": 0.4266, "num_input_tokens_seen": 10594592, "step": 7815 }, { "epoch": 0.25030407784392805, "grad_norm": 1.0675941705703735, "learning_rate": 4.543226878681955e-05, "loss": 0.4313, "num_input_tokens_seen": 10601232, "step": 7820 }, { "epoch": 0.2504641188144165, "grad_norm": 1.404863715171814, "learning_rate": 4.5426610130885087e-05, "loss": 0.4774, "num_input_tokens_seen": 10607616, "step": 7825 }, { "epoch": 0.25062415978490493, "grad_norm": 1.0892804861068726, "learning_rate": 4.542094832491064e-05, "loss": 0.6123, "num_input_tokens_seen": 10614400, "step": 7830 }, { "epoch": 0.25078420075539337, "grad_norm": 1.0933963060379028, "learning_rate": 4.541528336976934e-05, "loss": 0.4935, "num_input_tokens_seen": 10620992, "step": 7835 }, { "epoch": 0.2509442417258818, "grad_norm": 0.7845276594161987, "learning_rate": 4.540961526633479e-05, "loss": 0.5535, "num_input_tokens_seen": 10627792, "step": 7840 }, { "epoch": 0.25110428269637025, "grad_norm": 1.0899966955184937, "learning_rate": 4.540394401548108e-05, "loss": 0.6601, "num_input_tokens_seen": 10634656, "step": 7845 }, { "epoch": 0.2512643236668587, "grad_norm": 1.6143686771392822, "learning_rate": 4.539826961808279e-05, "loss": 0.666, "num_input_tokens_seen": 10641616, "step": 7850 }, { "epoch": 0.2514243646373472, "grad_norm": 1.690708875656128, "learning_rate": 4.5392592075014994e-05, "loss": 0.5339, "num_input_tokens_seen": 10648832, "step": 7855 }, { "epoch": 0.2515844056078356, "grad_norm": 1.3806477785110474, "learning_rate": 4.538691138715322e-05, "loss": 0.5932, "num_input_tokens_seen": 10655616, "step": 7860 }, { "epoch": 0.25174444657832407, "grad_norm": 0.9032357931137085, "learning_rate": 4.5381227555373516e-05, "loss": 0.3895, "num_input_tokens_seen": 10662256, "step": 7865 }, { "epoch": 0.2519044875488125, "grad_norm": 0.6362234354019165, "learning_rate": 4.537554058055239e-05, "loss": 0.5102, "num_input_tokens_seen": 10669136, "step": 7870 }, { "epoch": 0.25206452851930095, "grad_norm": 1.0120145082473755, "learning_rate": 4.5369850463566865e-05, "loss": 0.5041, "num_input_tokens_seen": 10675744, "step": 7875 }, { "epoch": 0.2522245694897894, "grad_norm": 5.081171989440918, "learning_rate": 4.5364157205294404e-05, "loss": 0.5866, "num_input_tokens_seen": 10682464, "step": 7880 }, { "epoch": 0.2523846104602778, "grad_norm": 1.4173463582992554, "learning_rate": 4.5358460806612996e-05, "loss": 0.6151, "num_input_tokens_seen": 10689696, "step": 7885 }, { "epoch": 0.25254465143076626, "grad_norm": 0.8250836133956909, "learning_rate": 4.535276126840109e-05, "loss": 0.5635, "num_input_tokens_seen": 10696608, "step": 7890 }, { "epoch": 0.2527046924012547, "grad_norm": 1.0813637971878052, "learning_rate": 4.5347058591537626e-05, "loss": 0.623, "num_input_tokens_seen": 10703424, "step": 7895 }, { "epoch": 0.25286473337174314, "grad_norm": 1.6483180522918701, "learning_rate": 4.534135277690203e-05, "loss": 0.5044, "num_input_tokens_seen": 10710800, "step": 7900 }, { "epoch": 0.2530247743422316, "grad_norm": 0.6125934720039368, "learning_rate": 4.533564382537421e-05, "loss": 0.4036, "num_input_tokens_seen": 10717120, "step": 7905 }, { "epoch": 0.2531848153127201, "grad_norm": 1.0303373336791992, "learning_rate": 4.532993173783456e-05, "loss": 0.3432, "num_input_tokens_seen": 10723776, "step": 7910 }, { "epoch": 0.2533448562832085, "grad_norm": 2.184128522872925, "learning_rate": 4.5324216515163954e-05, "loss": 0.6039, "num_input_tokens_seen": 10730512, "step": 7915 }, { "epoch": 0.25350489725369696, "grad_norm": 0.8100058436393738, "learning_rate": 4.531849815824375e-05, "loss": 0.5165, "num_input_tokens_seen": 10737472, "step": 7920 }, { "epoch": 0.2536649382241854, "grad_norm": 1.296100378036499, "learning_rate": 4.5312776667955795e-05, "loss": 0.3968, "num_input_tokens_seen": 10744304, "step": 7925 }, { "epoch": 0.25382497919467384, "grad_norm": 2.1709518432617188, "learning_rate": 4.5307052045182405e-05, "loss": 0.7637, "num_input_tokens_seen": 10750848, "step": 7930 }, { "epoch": 0.2539850201651623, "grad_norm": 1.4150619506835938, "learning_rate": 4.53013242908064e-05, "loss": 0.428, "num_input_tokens_seen": 10757456, "step": 7935 }, { "epoch": 0.2541450611356507, "grad_norm": 1.5176609754562378, "learning_rate": 4.529559340571107e-05, "loss": 0.5704, "num_input_tokens_seen": 10763808, "step": 7940 }, { "epoch": 0.25430510210613916, "grad_norm": 1.7204252481460571, "learning_rate": 4.528985939078018e-05, "loss": 0.7689, "num_input_tokens_seen": 10770112, "step": 7945 }, { "epoch": 0.2544651430766276, "grad_norm": 1.4091955423355103, "learning_rate": 4.5284122246898e-05, "loss": 0.542, "num_input_tokens_seen": 10777408, "step": 7950 }, { "epoch": 0.25462518404711604, "grad_norm": 2.574108362197876, "learning_rate": 4.527838197494926e-05, "loss": 0.3858, "num_input_tokens_seen": 10784272, "step": 7955 }, { "epoch": 0.25478522501760453, "grad_norm": 1.4695645570755005, "learning_rate": 4.527263857581918e-05, "loss": 0.5871, "num_input_tokens_seen": 10791056, "step": 7960 }, { "epoch": 0.254945265988093, "grad_norm": 1.0423882007598877, "learning_rate": 4.526689205039347e-05, "loss": 0.5368, "num_input_tokens_seen": 10797664, "step": 7965 }, { "epoch": 0.2551053069585814, "grad_norm": 1.1264718770980835, "learning_rate": 4.5261142399558324e-05, "loss": 0.3409, "num_input_tokens_seen": 10804272, "step": 7970 }, { "epoch": 0.25526534792906985, "grad_norm": 1.0163960456848145, "learning_rate": 4.525538962420041e-05, "loss": 0.4252, "num_input_tokens_seen": 10811168, "step": 7975 }, { "epoch": 0.2554253888995583, "grad_norm": 1.2508965730667114, "learning_rate": 4.524963372520685e-05, "loss": 0.5271, "num_input_tokens_seen": 10818016, "step": 7980 }, { "epoch": 0.25558542987004673, "grad_norm": 1.033935785293579, "learning_rate": 4.524387470346531e-05, "loss": 0.5107, "num_input_tokens_seen": 10824704, "step": 7985 }, { "epoch": 0.25574547084053517, "grad_norm": 0.7293351888656616, "learning_rate": 4.5238112559863885e-05, "loss": 0.6397, "num_input_tokens_seen": 10831552, "step": 7990 }, { "epoch": 0.2559055118110236, "grad_norm": 1.0253721475601196, "learning_rate": 4.5232347295291175e-05, "loss": 0.4171, "num_input_tokens_seen": 10838192, "step": 7995 }, { "epoch": 0.25606555278151205, "grad_norm": 0.825881838798523, "learning_rate": 4.522657891063626e-05, "loss": 0.7391, "num_input_tokens_seen": 10844928, "step": 8000 }, { "epoch": 0.25606555278151205, "eval_loss": 0.5298962593078613, "eval_runtime": 499.4224, "eval_samples_per_second": 27.804, "eval_steps_per_second": 13.902, "num_input_tokens_seen": 10844928, "step": 8000 }, { "epoch": 0.2562255937520005, "grad_norm": 0.6639760136604309, "learning_rate": 4.52208074067887e-05, "loss": 0.4318, "num_input_tokens_seen": 10851856, "step": 8005 }, { "epoch": 0.25638563472248893, "grad_norm": 1.0416303873062134, "learning_rate": 4.5215032784638516e-05, "loss": 0.4046, "num_input_tokens_seen": 10858496, "step": 8010 }, { "epoch": 0.2565456756929774, "grad_norm": 1.3042923212051392, "learning_rate": 4.5209255045076245e-05, "loss": 0.5291, "num_input_tokens_seen": 10865152, "step": 8015 }, { "epoch": 0.25670571666346587, "grad_norm": 0.8995494246482849, "learning_rate": 4.5203474188992875e-05, "loss": 0.5263, "num_input_tokens_seen": 10872016, "step": 8020 }, { "epoch": 0.2568657576339543, "grad_norm": 0.9807570576667786, "learning_rate": 4.51976902172799e-05, "loss": 0.3317, "num_input_tokens_seen": 10878624, "step": 8025 }, { "epoch": 0.25702579860444275, "grad_norm": 1.6277841329574585, "learning_rate": 4.519190313082927e-05, "loss": 0.738, "num_input_tokens_seen": 10885312, "step": 8030 }, { "epoch": 0.2571858395749312, "grad_norm": 1.0400348901748657, "learning_rate": 4.518611293053343e-05, "loss": 0.4034, "num_input_tokens_seen": 10892416, "step": 8035 }, { "epoch": 0.2573458805454196, "grad_norm": 1.983249306678772, "learning_rate": 4.51803196172853e-05, "loss": 0.5949, "num_input_tokens_seen": 10899136, "step": 8040 }, { "epoch": 0.25750592151590807, "grad_norm": 0.8724808692932129, "learning_rate": 4.517452319197828e-05, "loss": 0.5003, "num_input_tokens_seen": 10906048, "step": 8045 }, { "epoch": 0.2576659624863965, "grad_norm": 1.5508265495300293, "learning_rate": 4.5168723655506265e-05, "loss": 0.6987, "num_input_tokens_seen": 10912800, "step": 8050 }, { "epoch": 0.25782600345688494, "grad_norm": 0.8645637035369873, "learning_rate": 4.51629210087636e-05, "loss": 0.4428, "num_input_tokens_seen": 10919584, "step": 8055 }, { "epoch": 0.2579860444273734, "grad_norm": 1.130049467086792, "learning_rate": 4.515711525264513e-05, "loss": 0.2758, "num_input_tokens_seen": 10926576, "step": 8060 }, { "epoch": 0.2581460853978619, "grad_norm": 0.9833599925041199, "learning_rate": 4.5151306388046175e-05, "loss": 0.6158, "num_input_tokens_seen": 10933376, "step": 8065 }, { "epoch": 0.2583061263683503, "grad_norm": 0.83016437292099, "learning_rate": 4.514549441586255e-05, "loss": 0.4701, "num_input_tokens_seen": 10940448, "step": 8070 }, { "epoch": 0.25846616733883876, "grad_norm": 0.5228014588356018, "learning_rate": 4.513967933699051e-05, "loss": 0.5396, "num_input_tokens_seen": 10947216, "step": 8075 }, { "epoch": 0.2586262083093272, "grad_norm": 2.6453230381011963, "learning_rate": 4.513386115232684e-05, "loss": 0.6799, "num_input_tokens_seen": 10954080, "step": 8080 }, { "epoch": 0.25878624927981564, "grad_norm": 0.7906056046485901, "learning_rate": 4.5128039862768745e-05, "loss": 0.5585, "num_input_tokens_seen": 10960624, "step": 8085 }, { "epoch": 0.2589462902503041, "grad_norm": 0.7489363551139832, "learning_rate": 4.512221546921397e-05, "loss": 0.4772, "num_input_tokens_seen": 10967504, "step": 8090 }, { "epoch": 0.2591063312207925, "grad_norm": 1.9018465280532837, "learning_rate": 4.5116387972560694e-05, "loss": 0.4474, "num_input_tokens_seen": 10974144, "step": 8095 }, { "epoch": 0.25926637219128096, "grad_norm": 0.8658369779586792, "learning_rate": 4.511055737370759e-05, "loss": 0.3435, "num_input_tokens_seen": 10980592, "step": 8100 }, { "epoch": 0.2594264131617694, "grad_norm": 0.5584046244621277, "learning_rate": 4.510472367355383e-05, "loss": 0.51, "num_input_tokens_seen": 10987344, "step": 8105 }, { "epoch": 0.25958645413225784, "grad_norm": 2.0763611793518066, "learning_rate": 4.509888687299901e-05, "loss": 0.5049, "num_input_tokens_seen": 10993872, "step": 8110 }, { "epoch": 0.2597464951027463, "grad_norm": 0.9486345052719116, "learning_rate": 4.5093046972943266e-05, "loss": 0.376, "num_input_tokens_seen": 11001152, "step": 8115 }, { "epoch": 0.2599065360732348, "grad_norm": 1.558605432510376, "learning_rate": 4.508720397428717e-05, "loss": 0.4246, "num_input_tokens_seen": 11007904, "step": 8120 }, { "epoch": 0.2600665770437232, "grad_norm": 1.2356528043746948, "learning_rate": 4.508135787793178e-05, "loss": 0.4907, "num_input_tokens_seen": 11014624, "step": 8125 }, { "epoch": 0.26022661801421165, "grad_norm": 0.7581706643104553, "learning_rate": 4.5075508684778664e-05, "loss": 0.5818, "num_input_tokens_seen": 11021632, "step": 8130 }, { "epoch": 0.2603866589847001, "grad_norm": 2.099635601043701, "learning_rate": 4.506965639572982e-05, "loss": 0.5763, "num_input_tokens_seen": 11028688, "step": 8135 }, { "epoch": 0.26054669995518853, "grad_norm": 1.7318363189697266, "learning_rate": 4.506380101168774e-05, "loss": 0.5325, "num_input_tokens_seen": 11035120, "step": 8140 }, { "epoch": 0.26070674092567697, "grad_norm": 0.6898256540298462, "learning_rate": 4.505794253355542e-05, "loss": 0.3405, "num_input_tokens_seen": 11042208, "step": 8145 }, { "epoch": 0.2608667818961654, "grad_norm": 1.4710197448730469, "learning_rate": 4.5052080962236286e-05, "loss": 0.6399, "num_input_tokens_seen": 11048976, "step": 8150 }, { "epoch": 0.26102682286665385, "grad_norm": 1.5666197538375854, "learning_rate": 4.504621629863428e-05, "loss": 0.6177, "num_input_tokens_seen": 11055376, "step": 8155 }, { "epoch": 0.2611868638371423, "grad_norm": 0.8078778386116028, "learning_rate": 4.504034854365381e-05, "loss": 0.5029, "num_input_tokens_seen": 11061952, "step": 8160 }, { "epoch": 0.26134690480763073, "grad_norm": 0.9171946048736572, "learning_rate": 4.503447769819974e-05, "loss": 0.4539, "num_input_tokens_seen": 11068304, "step": 8165 }, { "epoch": 0.26150694577811917, "grad_norm": 0.7005523443222046, "learning_rate": 4.502860376317745e-05, "loss": 0.4647, "num_input_tokens_seen": 11075168, "step": 8170 }, { "epoch": 0.26166698674860767, "grad_norm": 0.9333193302154541, "learning_rate": 4.502272673949276e-05, "loss": 0.304, "num_input_tokens_seen": 11081728, "step": 8175 }, { "epoch": 0.2618270277190961, "grad_norm": 1.6200072765350342, "learning_rate": 4.501684662805199e-05, "loss": 0.5831, "num_input_tokens_seen": 11088016, "step": 8180 }, { "epoch": 0.26198706868958455, "grad_norm": 1.3477858304977417, "learning_rate": 4.5010963429761924e-05, "loss": 0.4971, "num_input_tokens_seen": 11094816, "step": 8185 }, { "epoch": 0.262147109660073, "grad_norm": 0.9684727787971497, "learning_rate": 4.500507714552982e-05, "loss": 0.5005, "num_input_tokens_seen": 11101824, "step": 8190 }, { "epoch": 0.2623071506305614, "grad_norm": 0.5936802625656128, "learning_rate": 4.499918777626342e-05, "loss": 0.3984, "num_input_tokens_seen": 11108192, "step": 8195 }, { "epoch": 0.26246719160104987, "grad_norm": 1.0756895542144775, "learning_rate": 4.499329532287093e-05, "loss": 0.4257, "num_input_tokens_seen": 11114800, "step": 8200 }, { "epoch": 0.26246719160104987, "eval_loss": 0.5277846455574036, "eval_runtime": 499.2023, "eval_samples_per_second": 27.816, "eval_steps_per_second": 13.908, "num_input_tokens_seen": 11114800, "step": 8200 }, { "epoch": 0.2626272325715383, "grad_norm": 0.9314586520195007, "learning_rate": 4.4987399786261064e-05, "loss": 0.5688, "num_input_tokens_seen": 11121440, "step": 8205 }, { "epoch": 0.26278727354202674, "grad_norm": 0.9255468249320984, "learning_rate": 4.498150116734297e-05, "loss": 0.5025, "num_input_tokens_seen": 11127584, "step": 8210 }, { "epoch": 0.2629473145125152, "grad_norm": 1.6308468580245972, "learning_rate": 4.4975599467026294e-05, "loss": 0.6871, "num_input_tokens_seen": 11134496, "step": 8215 }, { "epoch": 0.2631073554830036, "grad_norm": 0.6297937035560608, "learning_rate": 4.496969468622114e-05, "loss": 0.3693, "num_input_tokens_seen": 11140928, "step": 8220 }, { "epoch": 0.2632673964534921, "grad_norm": 0.8554880619049072, "learning_rate": 4.496378682583813e-05, "loss": 0.471, "num_input_tokens_seen": 11147568, "step": 8225 }, { "epoch": 0.26342743742398056, "grad_norm": 1.4572534561157227, "learning_rate": 4.495787588678829e-05, "loss": 0.4875, "num_input_tokens_seen": 11154400, "step": 8230 }, { "epoch": 0.263587478394469, "grad_norm": 1.1934386491775513, "learning_rate": 4.4951961869983196e-05, "loss": 0.5494, "num_input_tokens_seen": 11161024, "step": 8235 }, { "epoch": 0.26374751936495744, "grad_norm": 1.520181655883789, "learning_rate": 4.494604477633485e-05, "loss": 0.6205, "num_input_tokens_seen": 11167536, "step": 8240 }, { "epoch": 0.2639075603354459, "grad_norm": 1.8711615800857544, "learning_rate": 4.4940124606755734e-05, "loss": 0.5391, "num_input_tokens_seen": 11174608, "step": 8245 }, { "epoch": 0.2640676013059343, "grad_norm": 1.0323874950408936, "learning_rate": 4.493420136215882e-05, "loss": 0.4809, "num_input_tokens_seen": 11181360, "step": 8250 }, { "epoch": 0.26422764227642276, "grad_norm": 1.3629484176635742, "learning_rate": 4.492827504345756e-05, "loss": 0.523, "num_input_tokens_seen": 11188576, "step": 8255 }, { "epoch": 0.2643876832469112, "grad_norm": 0.7696519494056702, "learning_rate": 4.492234565156584e-05, "loss": 0.3414, "num_input_tokens_seen": 11194976, "step": 8260 }, { "epoch": 0.26454772421739964, "grad_norm": 0.883080005645752, "learning_rate": 4.491641318739807e-05, "loss": 0.4533, "num_input_tokens_seen": 11201712, "step": 8265 }, { "epoch": 0.2647077651878881, "grad_norm": 1.1559021472930908, "learning_rate": 4.4910477651869096e-05, "loss": 0.5229, "num_input_tokens_seen": 11208096, "step": 8270 }, { "epoch": 0.2648678061583765, "grad_norm": 1.4430172443389893, "learning_rate": 4.4904539045894254e-05, "loss": 0.5883, "num_input_tokens_seen": 11214752, "step": 8275 }, { "epoch": 0.265027847128865, "grad_norm": 1.0525788068771362, "learning_rate": 4.4898597370389364e-05, "loss": 0.3688, "num_input_tokens_seen": 11221840, "step": 8280 }, { "epoch": 0.26518788809935345, "grad_norm": 1.2254334688186646, "learning_rate": 4.489265262627069e-05, "loss": 0.3899, "num_input_tokens_seen": 11228256, "step": 8285 }, { "epoch": 0.2653479290698419, "grad_norm": 0.5568065643310547, "learning_rate": 4.488670481445499e-05, "loss": 0.4591, "num_input_tokens_seen": 11234896, "step": 8290 }, { "epoch": 0.26550797004033033, "grad_norm": 1.41274094581604, "learning_rate": 4.488075393585951e-05, "loss": 0.6047, "num_input_tokens_seen": 11241648, "step": 8295 }, { "epoch": 0.26566801101081877, "grad_norm": 1.1253918409347534, "learning_rate": 4.487479999140193e-05, "loss": 0.5627, "num_input_tokens_seen": 11248800, "step": 8300 }, { "epoch": 0.2658280519813072, "grad_norm": 1.3859901428222656, "learning_rate": 4.4868842982000425e-05, "loss": 0.5125, "num_input_tokens_seen": 11255120, "step": 8305 }, { "epoch": 0.26598809295179565, "grad_norm": 1.82327401638031, "learning_rate": 4.486288290857365e-05, "loss": 0.5873, "num_input_tokens_seen": 11261760, "step": 8310 }, { "epoch": 0.2661481339222841, "grad_norm": 0.9681459069252014, "learning_rate": 4.4856919772040715e-05, "loss": 0.2686, "num_input_tokens_seen": 11268384, "step": 8315 }, { "epoch": 0.26630817489277253, "grad_norm": 0.7566037178039551, "learning_rate": 4.485095357332122e-05, "loss": 0.4759, "num_input_tokens_seen": 11274880, "step": 8320 }, { "epoch": 0.26646821586326097, "grad_norm": 1.106957197189331, "learning_rate": 4.484498431333521e-05, "loss": 0.5133, "num_input_tokens_seen": 11281696, "step": 8325 }, { "epoch": 0.26662825683374947, "grad_norm": 2.4027864933013916, "learning_rate": 4.4839011993003245e-05, "loss": 0.5475, "num_input_tokens_seen": 11288640, "step": 8330 }, { "epoch": 0.2667882978042379, "grad_norm": 1.3712412118911743, "learning_rate": 4.4833036613246305e-05, "loss": 0.4224, "num_input_tokens_seen": 11295648, "step": 8335 }, { "epoch": 0.26694833877472635, "grad_norm": 0.6962305307388306, "learning_rate": 4.482705817498589e-05, "loss": 0.4511, "num_input_tokens_seen": 11302176, "step": 8340 }, { "epoch": 0.2671083797452148, "grad_norm": 0.9766198992729187, "learning_rate": 4.4821076679143934e-05, "loss": 0.3943, "num_input_tokens_seen": 11308816, "step": 8345 }, { "epoch": 0.2672684207157032, "grad_norm": 0.9331104159355164, "learning_rate": 4.481509212664288e-05, "loss": 0.4472, "num_input_tokens_seen": 11316736, "step": 8350 }, { "epoch": 0.26742846168619167, "grad_norm": 1.544039011001587, "learning_rate": 4.480910451840559e-05, "loss": 0.7057, "num_input_tokens_seen": 11323408, "step": 8355 }, { "epoch": 0.2675885026566801, "grad_norm": 1.6054325103759766, "learning_rate": 4.480311385535546e-05, "loss": 0.6958, "num_input_tokens_seen": 11330256, "step": 8360 }, { "epoch": 0.26774854362716854, "grad_norm": 1.2121975421905518, "learning_rate": 4.47971201384163e-05, "loss": 0.4637, "num_input_tokens_seen": 11336768, "step": 8365 }, { "epoch": 0.267908584597657, "grad_norm": 1.381933569908142, "learning_rate": 4.4791123368512446e-05, "loss": 0.4254, "num_input_tokens_seen": 11343296, "step": 8370 }, { "epoch": 0.2680686255681454, "grad_norm": 1.0109390020370483, "learning_rate": 4.478512354656864e-05, "loss": 0.528, "num_input_tokens_seen": 11349728, "step": 8375 }, { "epoch": 0.26822866653863386, "grad_norm": 0.9433351755142212, "learning_rate": 4.477912067351016e-05, "loss": 0.4855, "num_input_tokens_seen": 11356624, "step": 8380 }, { "epoch": 0.26838870750912236, "grad_norm": 1.1433069705963135, "learning_rate": 4.477311475026271e-05, "loss": 0.4942, "num_input_tokens_seen": 11363136, "step": 8385 }, { "epoch": 0.2685487484796108, "grad_norm": 1.310153603553772, "learning_rate": 4.476710577775248e-05, "loss": 0.3387, "num_input_tokens_seen": 11370112, "step": 8390 }, { "epoch": 0.26870878945009924, "grad_norm": 1.805811882019043, "learning_rate": 4.476109375690612e-05, "loss": 0.6054, "num_input_tokens_seen": 11376688, "step": 8395 }, { "epoch": 0.2688688304205877, "grad_norm": 2.5247600078582764, "learning_rate": 4.4755078688650784e-05, "loss": 0.5219, "num_input_tokens_seen": 11383280, "step": 8400 }, { "epoch": 0.2688688304205877, "eval_loss": 0.5253506898880005, "eval_runtime": 499.7196, "eval_samples_per_second": 27.788, "eval_steps_per_second": 13.894, "num_input_tokens_seen": 11383280, "step": 8400 }, { "epoch": 0.2690288713910761, "grad_norm": 1.3715037107467651, "learning_rate": 4.474906057391406e-05, "loss": 0.417, "num_input_tokens_seen": 11390208, "step": 8405 }, { "epoch": 0.26918891236156456, "grad_norm": 0.9021508097648621, "learning_rate": 4.4743039413624e-05, "loss": 0.4494, "num_input_tokens_seen": 11397072, "step": 8410 }, { "epoch": 0.269348953332053, "grad_norm": 1.7077730894088745, "learning_rate": 4.473701520870916e-05, "loss": 0.4755, "num_input_tokens_seen": 11403584, "step": 8415 }, { "epoch": 0.26950899430254144, "grad_norm": 1.1275111436843872, "learning_rate": 4.4730987960098544e-05, "loss": 0.4138, "num_input_tokens_seen": 11409904, "step": 8420 }, { "epoch": 0.2696690352730299, "grad_norm": 1.2779653072357178, "learning_rate": 4.4724957668721635e-05, "loss": 0.7193, "num_input_tokens_seen": 11416336, "step": 8425 }, { "epoch": 0.2698290762435183, "grad_norm": 2.2274463176727295, "learning_rate": 4.471892433550836e-05, "loss": 0.5974, "num_input_tokens_seen": 11423184, "step": 8430 }, { "epoch": 0.2699891172140068, "grad_norm": 1.2414251565933228, "learning_rate": 4.471288796138916e-05, "loss": 0.5192, "num_input_tokens_seen": 11429968, "step": 8435 }, { "epoch": 0.27014915818449525, "grad_norm": 0.9367305040359497, "learning_rate": 4.470684854729491e-05, "loss": 0.4032, "num_input_tokens_seen": 11436816, "step": 8440 }, { "epoch": 0.2703091991549837, "grad_norm": 1.0045889616012573, "learning_rate": 4.4700806094156955e-05, "loss": 0.3875, "num_input_tokens_seen": 11444192, "step": 8445 }, { "epoch": 0.27046924012547213, "grad_norm": 1.0842821598052979, "learning_rate": 4.469476060290713e-05, "loss": 0.4704, "num_input_tokens_seen": 11450976, "step": 8450 }, { "epoch": 0.27062928109596057, "grad_norm": 1.2253859043121338, "learning_rate": 4.468871207447772e-05, "loss": 0.3441, "num_input_tokens_seen": 11457616, "step": 8455 }, { "epoch": 0.270789322066449, "grad_norm": 1.7265057563781738, "learning_rate": 4.4682660509801486e-05, "loss": 0.512, "num_input_tokens_seen": 11464096, "step": 8460 }, { "epoch": 0.27094936303693745, "grad_norm": 1.0592697858810425, "learning_rate": 4.467660590981165e-05, "loss": 0.4134, "num_input_tokens_seen": 11470800, "step": 8465 }, { "epoch": 0.2711094040074259, "grad_norm": 1.526940941810608, "learning_rate": 4.467054827544191e-05, "loss": 0.5704, "num_input_tokens_seen": 11477184, "step": 8470 }, { "epoch": 0.27126944497791433, "grad_norm": 1.1034660339355469, "learning_rate": 4.4664487607626434e-05, "loss": 0.6477, "num_input_tokens_seen": 11484096, "step": 8475 }, { "epoch": 0.27142948594840277, "grad_norm": 1.898706316947937, "learning_rate": 4.4658423907299845e-05, "loss": 0.5559, "num_input_tokens_seen": 11490864, "step": 8480 }, { "epoch": 0.2715895269188912, "grad_norm": 1.6473475694656372, "learning_rate": 4.465235717539725e-05, "loss": 0.5319, "num_input_tokens_seen": 11497456, "step": 8485 }, { "epoch": 0.2717495678893797, "grad_norm": 0.7651184797286987, "learning_rate": 4.464628741285421e-05, "loss": 0.6369, "num_input_tokens_seen": 11503968, "step": 8490 }, { "epoch": 0.27190960885986815, "grad_norm": 1.3632402420043945, "learning_rate": 4.4640214620606754e-05, "loss": 0.5735, "num_input_tokens_seen": 11510336, "step": 8495 }, { "epoch": 0.2720696498303566, "grad_norm": 0.5683907866477966, "learning_rate": 4.46341387995914e-05, "loss": 0.3338, "num_input_tokens_seen": 11517136, "step": 8500 }, { "epoch": 0.272229690800845, "grad_norm": 0.9059939980506897, "learning_rate": 4.4628059950745106e-05, "loss": 0.7273, "num_input_tokens_seen": 11524240, "step": 8505 }, { "epoch": 0.27238973177133347, "grad_norm": 0.7958516478538513, "learning_rate": 4.4621978075005297e-05, "loss": 0.3359, "num_input_tokens_seen": 11531456, "step": 8510 }, { "epoch": 0.2725497727418219, "grad_norm": 0.7560676336288452, "learning_rate": 4.461589317330989e-05, "loss": 0.456, "num_input_tokens_seen": 11538816, "step": 8515 }, { "epoch": 0.27270981371231034, "grad_norm": 0.8036216497421265, "learning_rate": 4.460980524659724e-05, "loss": 0.4006, "num_input_tokens_seen": 11545520, "step": 8520 }, { "epoch": 0.2728698546827988, "grad_norm": 1.8713688850402832, "learning_rate": 4.46037142958062e-05, "loss": 0.6654, "num_input_tokens_seen": 11552256, "step": 8525 }, { "epoch": 0.2730298956532872, "grad_norm": 2.241523265838623, "learning_rate": 4.4597620321876046e-05, "loss": 0.6072, "num_input_tokens_seen": 11559040, "step": 8530 }, { "epoch": 0.27318993662377566, "grad_norm": 1.7243746519088745, "learning_rate": 4.459152332574656e-05, "loss": 0.4868, "num_input_tokens_seen": 11565616, "step": 8535 }, { "epoch": 0.2733499775942641, "grad_norm": 2.3676555156707764, "learning_rate": 4.4585423308357985e-05, "loss": 0.54, "num_input_tokens_seen": 11572256, "step": 8540 }, { "epoch": 0.2735100185647526, "grad_norm": 0.8019701242446899, "learning_rate": 4.457932027065102e-05, "loss": 0.5159, "num_input_tokens_seen": 11579136, "step": 8545 }, { "epoch": 0.27367005953524104, "grad_norm": 0.9422193169593811, "learning_rate": 4.45732142135668e-05, "loss": 0.4781, "num_input_tokens_seen": 11586080, "step": 8550 }, { "epoch": 0.2738301005057295, "grad_norm": 0.9451255798339844, "learning_rate": 4.4567105138046986e-05, "loss": 0.39, "num_input_tokens_seen": 11592800, "step": 8555 }, { "epoch": 0.2739901414762179, "grad_norm": 1.27446448802948, "learning_rate": 4.456099304503365e-05, "loss": 0.4248, "num_input_tokens_seen": 11599680, "step": 8560 }, { "epoch": 0.27415018244670636, "grad_norm": 0.9363679885864258, "learning_rate": 4.455487793546939e-05, "loss": 0.5359, "num_input_tokens_seen": 11606464, "step": 8565 }, { "epoch": 0.2743102234171948, "grad_norm": 1.213097095489502, "learning_rate": 4.454875981029719e-05, "loss": 0.4884, "num_input_tokens_seen": 11613216, "step": 8570 }, { "epoch": 0.27447026438768324, "grad_norm": 1.6290624141693115, "learning_rate": 4.454263867046057e-05, "loss": 0.433, "num_input_tokens_seen": 11619632, "step": 8575 }, { "epoch": 0.2746303053581717, "grad_norm": 1.5222400426864624, "learning_rate": 4.4536514516903484e-05, "loss": 0.5716, "num_input_tokens_seen": 11625952, "step": 8580 }, { "epoch": 0.2747903463286601, "grad_norm": 0.9805195927619934, "learning_rate": 4.453038735057034e-05, "loss": 0.4944, "num_input_tokens_seen": 11632528, "step": 8585 }, { "epoch": 0.27495038729914856, "grad_norm": 1.6169747114181519, "learning_rate": 4.4524257172406034e-05, "loss": 0.443, "num_input_tokens_seen": 11639200, "step": 8590 }, { "epoch": 0.27511042826963705, "grad_norm": 0.9627138376235962, "learning_rate": 4.451812398335592e-05, "loss": 0.3356, "num_input_tokens_seen": 11645696, "step": 8595 }, { "epoch": 0.2752704692401255, "grad_norm": 1.5043470859527588, "learning_rate": 4.4511987784365805e-05, "loss": 0.6746, "num_input_tokens_seen": 11652336, "step": 8600 }, { "epoch": 0.2752704692401255, "eval_loss": 0.5234595537185669, "eval_runtime": 499.6722, "eval_samples_per_second": 27.79, "eval_steps_per_second": 13.895, "num_input_tokens_seen": 11652336, "step": 8600 }, { "epoch": 0.27543051021061393, "grad_norm": 1.198299527168274, "learning_rate": 4.450584857638197e-05, "loss": 0.6854, "num_input_tokens_seen": 11658688, "step": 8605 }, { "epoch": 0.2755905511811024, "grad_norm": 0.9054769277572632, "learning_rate": 4.449970636035116e-05, "loss": 0.3386, "num_input_tokens_seen": 11665488, "step": 8610 }, { "epoch": 0.2757505921515908, "grad_norm": 1.3837451934814453, "learning_rate": 4.4493561137220574e-05, "loss": 0.4327, "num_input_tokens_seen": 11672528, "step": 8615 }, { "epoch": 0.27591063312207925, "grad_norm": 0.9844025373458862, "learning_rate": 4.44874129079379e-05, "loss": 0.4657, "num_input_tokens_seen": 11679248, "step": 8620 }, { "epoch": 0.2760706740925677, "grad_norm": 1.972878098487854, "learning_rate": 4.4481261673451255e-05, "loss": 0.4815, "num_input_tokens_seen": 11685968, "step": 8625 }, { "epoch": 0.27623071506305613, "grad_norm": 0.7521002292633057, "learning_rate": 4.4475107434709245e-05, "loss": 0.485, "num_input_tokens_seen": 11692688, "step": 8630 }, { "epoch": 0.27639075603354457, "grad_norm": 1.5680036544799805, "learning_rate": 4.446895019266093e-05, "loss": 0.4543, "num_input_tokens_seen": 11699376, "step": 8635 }, { "epoch": 0.276550797004033, "grad_norm": 0.6951817870140076, "learning_rate": 4.446278994825583e-05, "loss": 0.7993, "num_input_tokens_seen": 11705776, "step": 8640 }, { "epoch": 0.27671083797452145, "grad_norm": 1.608963966369629, "learning_rate": 4.445662670244394e-05, "loss": 0.4372, "num_input_tokens_seen": 11712480, "step": 8645 }, { "epoch": 0.27687087894500995, "grad_norm": 2.473538637161255, "learning_rate": 4.44504604561757e-05, "loss": 0.4993, "num_input_tokens_seen": 11719376, "step": 8650 }, { "epoch": 0.2770309199154984, "grad_norm": 0.5289309620857239, "learning_rate": 4.4444291210402035e-05, "loss": 0.5887, "num_input_tokens_seen": 11725776, "step": 8655 }, { "epoch": 0.2771909608859868, "grad_norm": 0.8701899647712708, "learning_rate": 4.443811896607431e-05, "loss": 0.5115, "num_input_tokens_seen": 11732880, "step": 8660 }, { "epoch": 0.27735100185647527, "grad_norm": 1.3158544301986694, "learning_rate": 4.443194372414436e-05, "loss": 0.4376, "num_input_tokens_seen": 11739536, "step": 8665 }, { "epoch": 0.2775110428269637, "grad_norm": 1.1290161609649658, "learning_rate": 4.442576548556449e-05, "loss": 0.5235, "num_input_tokens_seen": 11746544, "step": 8670 }, { "epoch": 0.27767108379745215, "grad_norm": 1.6832879781723022, "learning_rate": 4.441958425128747e-05, "loss": 0.7763, "num_input_tokens_seen": 11753472, "step": 8675 }, { "epoch": 0.2778311247679406, "grad_norm": 1.3742324113845825, "learning_rate": 4.4413400022266515e-05, "loss": 0.4435, "num_input_tokens_seen": 11760400, "step": 8680 }, { "epoch": 0.277991165738429, "grad_norm": 1.6787407398223877, "learning_rate": 4.4407212799455313e-05, "loss": 0.6279, "num_input_tokens_seen": 11767440, "step": 8685 }, { "epoch": 0.27815120670891746, "grad_norm": 0.6921550035476685, "learning_rate": 4.4401022583808003e-05, "loss": 0.4704, "num_input_tokens_seen": 11774192, "step": 8690 }, { "epoch": 0.2783112476794059, "grad_norm": 1.4841036796569824, "learning_rate": 4.439482937627921e-05, "loss": 0.5604, "num_input_tokens_seen": 11780960, "step": 8695 }, { "epoch": 0.2784712886498944, "grad_norm": 1.7233000993728638, "learning_rate": 4.4388633177824004e-05, "loss": 0.5384, "num_input_tokens_seen": 11787312, "step": 8700 }, { "epoch": 0.27863132962038284, "grad_norm": 1.4677599668502808, "learning_rate": 4.4382433989397895e-05, "loss": 0.6289, "num_input_tokens_seen": 11793920, "step": 8705 }, { "epoch": 0.2787913705908713, "grad_norm": 1.4541090726852417, "learning_rate": 4.4376231811956895e-05, "loss": 0.5877, "num_input_tokens_seen": 11800848, "step": 8710 }, { "epoch": 0.2789514115613597, "grad_norm": 1.2148993015289307, "learning_rate": 4.437002664645745e-05, "loss": 0.5638, "num_input_tokens_seen": 11807888, "step": 8715 }, { "epoch": 0.27911145253184816, "grad_norm": 2.558711290359497, "learning_rate": 4.436381849385649e-05, "loss": 0.5789, "num_input_tokens_seen": 11814688, "step": 8720 }, { "epoch": 0.2792714935023366, "grad_norm": 1.4763002395629883, "learning_rate": 4.435760735511136e-05, "loss": 0.5911, "num_input_tokens_seen": 11821696, "step": 8725 }, { "epoch": 0.27943153447282504, "grad_norm": 0.724153995513916, "learning_rate": 4.435139323117992e-05, "loss": 0.5228, "num_input_tokens_seen": 11828368, "step": 8730 }, { "epoch": 0.2795915754433135, "grad_norm": 1.4200477600097656, "learning_rate": 4.434517612302046e-05, "loss": 0.5532, "num_input_tokens_seen": 11834896, "step": 8735 }, { "epoch": 0.2797516164138019, "grad_norm": 0.8560039401054382, "learning_rate": 4.433895603159174e-05, "loss": 0.4656, "num_input_tokens_seen": 11841584, "step": 8740 }, { "epoch": 0.27991165738429036, "grad_norm": 0.863024115562439, "learning_rate": 4.433273295785296e-05, "loss": 0.3992, "num_input_tokens_seen": 11848304, "step": 8745 }, { "epoch": 0.2800716983547788, "grad_norm": 1.207786202430725, "learning_rate": 4.432650690276382e-05, "loss": 0.3804, "num_input_tokens_seen": 11855200, "step": 8750 }, { "epoch": 0.2802317393252673, "grad_norm": 1.2124871015548706, "learning_rate": 4.4320277867284435e-05, "loss": 0.4248, "num_input_tokens_seen": 11862400, "step": 8755 }, { "epoch": 0.28039178029575573, "grad_norm": 3.116804361343384, "learning_rate": 4.431404585237541e-05, "loss": 0.6388, "num_input_tokens_seen": 11869232, "step": 8760 }, { "epoch": 0.2805518212662442, "grad_norm": 0.5838012099266052, "learning_rate": 4.43078108589978e-05, "loss": 0.4379, "num_input_tokens_seen": 11875760, "step": 8765 }, { "epoch": 0.2807118622367326, "grad_norm": 1.1304106712341309, "learning_rate": 4.4301572888113116e-05, "loss": 0.5083, "num_input_tokens_seen": 11882480, "step": 8770 }, { "epoch": 0.28087190320722105, "grad_norm": 0.6193222999572754, "learning_rate": 4.4295331940683337e-05, "loss": 0.4272, "num_input_tokens_seen": 11889472, "step": 8775 }, { "epoch": 0.2810319441777095, "grad_norm": 1.0016785860061646, "learning_rate": 4.428908801767089e-05, "loss": 0.4223, "num_input_tokens_seen": 11896464, "step": 8780 }, { "epoch": 0.28119198514819793, "grad_norm": 1.4528626203536987, "learning_rate": 4.428284112003868e-05, "loss": 0.5397, "num_input_tokens_seen": 11902912, "step": 8785 }, { "epoch": 0.28135202611868637, "grad_norm": 0.9973322749137878, "learning_rate": 4.4276591248750033e-05, "loss": 0.6219, "num_input_tokens_seen": 11910192, "step": 8790 }, { "epoch": 0.2815120670891748, "grad_norm": 0.922696053981781, "learning_rate": 4.4270338404768774e-05, "loss": 0.5659, "num_input_tokens_seen": 11917424, "step": 8795 }, { "epoch": 0.28167210805966325, "grad_norm": 1.6435879468917847, "learning_rate": 4.426408258905917e-05, "loss": 0.5384, "num_input_tokens_seen": 11924224, "step": 8800 }, { "epoch": 0.28167210805966325, "eval_loss": 0.5215270519256592, "eval_runtime": 499.2322, "eval_samples_per_second": 27.815, "eval_steps_per_second": 13.907, "num_input_tokens_seen": 11924224, "step": 8800 }, { "epoch": 0.28183214903015175, "grad_norm": 0.8104385137557983, "learning_rate": 4.425782380258594e-05, "loss": 0.576, "num_input_tokens_seen": 11930928, "step": 8805 }, { "epoch": 0.2819921900006402, "grad_norm": 1.1225333213806152, "learning_rate": 4.425156204631427e-05, "loss": 0.5657, "num_input_tokens_seen": 11937632, "step": 8810 }, { "epoch": 0.2821522309711286, "grad_norm": 0.7239481806755066, "learning_rate": 4.424529732120981e-05, "loss": 0.5294, "num_input_tokens_seen": 11944208, "step": 8815 }, { "epoch": 0.28231227194161707, "grad_norm": 2.0991196632385254, "learning_rate": 4.423902962823864e-05, "loss": 0.8009, "num_input_tokens_seen": 11951056, "step": 8820 }, { "epoch": 0.2824723129121055, "grad_norm": 0.6740694642066956, "learning_rate": 4.423275896836733e-05, "loss": 0.5013, "num_input_tokens_seen": 11958192, "step": 8825 }, { "epoch": 0.28263235388259395, "grad_norm": 2.047917366027832, "learning_rate": 4.42264853425629e-05, "loss": 0.6301, "num_input_tokens_seen": 11965024, "step": 8830 }, { "epoch": 0.2827923948530824, "grad_norm": 1.0677800178527832, "learning_rate": 4.4220208751792816e-05, "loss": 0.3906, "num_input_tokens_seen": 11971968, "step": 8835 }, { "epoch": 0.2829524358235708, "grad_norm": 0.7410857081413269, "learning_rate": 4.421392919702499e-05, "loss": 0.444, "num_input_tokens_seen": 11978560, "step": 8840 }, { "epoch": 0.28311247679405926, "grad_norm": 0.9561506509780884, "learning_rate": 4.4207646679227846e-05, "loss": 0.554, "num_input_tokens_seen": 11985296, "step": 8845 }, { "epoch": 0.2832725177645477, "grad_norm": 1.019690990447998, "learning_rate": 4.42013611993702e-05, "loss": 0.4775, "num_input_tokens_seen": 11991888, "step": 8850 }, { "epoch": 0.28343255873503614, "grad_norm": 0.39964666962623596, "learning_rate": 4.419507275842135e-05, "loss": 0.5395, "num_input_tokens_seen": 11998512, "step": 8855 }, { "epoch": 0.28359259970552464, "grad_norm": 0.908686637878418, "learning_rate": 4.418878135735106e-05, "loss": 0.5275, "num_input_tokens_seen": 12005696, "step": 8860 }, { "epoch": 0.2837526406760131, "grad_norm": 1.1452174186706543, "learning_rate": 4.418248699712955e-05, "loss": 0.7084, "num_input_tokens_seen": 12012128, "step": 8865 }, { "epoch": 0.2839126816465015, "grad_norm": 0.9133548736572266, "learning_rate": 4.417618967872748e-05, "loss": 0.5272, "num_input_tokens_seen": 12018704, "step": 8870 }, { "epoch": 0.28407272261698996, "grad_norm": 2.087153196334839, "learning_rate": 4.4169889403115985e-05, "loss": 0.6458, "num_input_tokens_seen": 12025616, "step": 8875 }, { "epoch": 0.2842327635874784, "grad_norm": 1.1419086456298828, "learning_rate": 4.4163586171266627e-05, "loss": 0.5309, "num_input_tokens_seen": 12032304, "step": 8880 }, { "epoch": 0.28439280455796684, "grad_norm": 0.8752790689468384, "learning_rate": 4.415727998415147e-05, "loss": 0.4249, "num_input_tokens_seen": 12039344, "step": 8885 }, { "epoch": 0.2845528455284553, "grad_norm": 0.7574382424354553, "learning_rate": 4.4150970842742985e-05, "loss": 0.3611, "num_input_tokens_seen": 12046064, "step": 8890 }, { "epoch": 0.2847128864989437, "grad_norm": 1.7689530849456787, "learning_rate": 4.4144658748014134e-05, "loss": 0.6292, "num_input_tokens_seen": 12052832, "step": 8895 }, { "epoch": 0.28487292746943216, "grad_norm": 2.0875461101531982, "learning_rate": 4.413834370093831e-05, "loss": 0.6299, "num_input_tokens_seen": 12059632, "step": 8900 }, { "epoch": 0.2850329684399206, "grad_norm": 1.540197491645813, "learning_rate": 4.413202570248939e-05, "loss": 0.7435, "num_input_tokens_seen": 12066880, "step": 8905 }, { "epoch": 0.28519300941040904, "grad_norm": 1.7674944400787354, "learning_rate": 4.412570475364167e-05, "loss": 0.4849, "num_input_tokens_seen": 12073600, "step": 8910 }, { "epoch": 0.28535305038089753, "grad_norm": 1.2549128532409668, "learning_rate": 4.411938085536994e-05, "loss": 0.5256, "num_input_tokens_seen": 12080368, "step": 8915 }, { "epoch": 0.285513091351386, "grad_norm": 2.145019769668579, "learning_rate": 4.41130540086494e-05, "loss": 0.6025, "num_input_tokens_seen": 12087120, "step": 8920 }, { "epoch": 0.2856731323218744, "grad_norm": 1.4076712131500244, "learning_rate": 4.4106724214455754e-05, "loss": 0.688, "num_input_tokens_seen": 12093776, "step": 8925 }, { "epoch": 0.28583317329236285, "grad_norm": 1.2410351037979126, "learning_rate": 4.4100391473765115e-05, "loss": 0.4576, "num_input_tokens_seen": 12100208, "step": 8930 }, { "epoch": 0.2859932142628513, "grad_norm": 1.9160709381103516, "learning_rate": 4.409405578755408e-05, "loss": 0.4151, "num_input_tokens_seen": 12107008, "step": 8935 }, { "epoch": 0.28615325523333973, "grad_norm": 1.2917217016220093, "learning_rate": 4.4087717156799705e-05, "loss": 0.4634, "num_input_tokens_seen": 12113520, "step": 8940 }, { "epoch": 0.28631329620382817, "grad_norm": 1.021485447883606, "learning_rate": 4.408137558247946e-05, "loss": 0.4819, "num_input_tokens_seen": 12120192, "step": 8945 }, { "epoch": 0.2864733371743166, "grad_norm": 1.1919537782669067, "learning_rate": 4.4075031065571306e-05, "loss": 0.4158, "num_input_tokens_seen": 12127328, "step": 8950 }, { "epoch": 0.28663337814480505, "grad_norm": 1.7275784015655518, "learning_rate": 4.406868360705366e-05, "loss": 0.4173, "num_input_tokens_seen": 12134000, "step": 8955 }, { "epoch": 0.2867934191152935, "grad_norm": 0.9077951908111572, "learning_rate": 4.406233320790536e-05, "loss": 0.2374, "num_input_tokens_seen": 12140736, "step": 8960 }, { "epoch": 0.286953460085782, "grad_norm": 0.7635043859481812, "learning_rate": 4.4055979869105734e-05, "loss": 0.4502, "num_input_tokens_seen": 12147408, "step": 8965 }, { "epoch": 0.2871135010562704, "grad_norm": 1.1353572607040405, "learning_rate": 4.404962359163454e-05, "loss": 0.4697, "num_input_tokens_seen": 12154464, "step": 8970 }, { "epoch": 0.28727354202675887, "grad_norm": 1.1400611400604248, "learning_rate": 4.404326437647199e-05, "loss": 0.377, "num_input_tokens_seen": 12161136, "step": 8975 }, { "epoch": 0.2874335829972473, "grad_norm": 1.1957370042800903, "learning_rate": 4.403690222459877e-05, "loss": 0.4945, "num_input_tokens_seen": 12168000, "step": 8980 }, { "epoch": 0.28759362396773575, "grad_norm": 0.9448075890541077, "learning_rate": 4.4030537136995984e-05, "loss": 0.5296, "num_input_tokens_seen": 12174480, "step": 8985 }, { "epoch": 0.2877536649382242, "grad_norm": 1.0116946697235107, "learning_rate": 4.402416911464523e-05, "loss": 0.7003, "num_input_tokens_seen": 12181296, "step": 8990 }, { "epoch": 0.2879137059087126, "grad_norm": 1.409566044807434, "learning_rate": 4.4017798158528516e-05, "loss": 0.4788, "num_input_tokens_seen": 12188192, "step": 8995 }, { "epoch": 0.28807374687920106, "grad_norm": 1.1772701740264893, "learning_rate": 4.401142426962834e-05, "loss": 0.4862, "num_input_tokens_seen": 12194800, "step": 9000 }, { "epoch": 0.28807374687920106, "eval_loss": 0.5198710560798645, "eval_runtime": 499.5973, "eval_samples_per_second": 27.794, "eval_steps_per_second": 13.897, "num_input_tokens_seen": 12194800, "step": 9000 }, { "epoch": 0.2882337878496895, "grad_norm": 1.2711893320083618, "learning_rate": 4.400504744892763e-05, "loss": 0.5788, "num_input_tokens_seen": 12201840, "step": 9005 }, { "epoch": 0.28839382882017794, "grad_norm": 0.7466726303100586, "learning_rate": 4.399866769740975e-05, "loss": 0.4407, "num_input_tokens_seen": 12208672, "step": 9010 }, { "epoch": 0.2885538697906664, "grad_norm": 0.7525951266288757, "learning_rate": 4.399228501605859e-05, "loss": 0.4788, "num_input_tokens_seen": 12215936, "step": 9015 }, { "epoch": 0.2887139107611549, "grad_norm": 0.56183260679245, "learning_rate": 4.398589940585839e-05, "loss": 0.5937, "num_input_tokens_seen": 12222640, "step": 9020 }, { "epoch": 0.2888739517316433, "grad_norm": 0.40617090463638306, "learning_rate": 4.3979510867793917e-05, "loss": 0.4785, "num_input_tokens_seen": 12229872, "step": 9025 }, { "epoch": 0.28903399270213176, "grad_norm": 1.5806193351745605, "learning_rate": 4.3973119402850346e-05, "loss": 0.599, "num_input_tokens_seen": 12236480, "step": 9030 }, { "epoch": 0.2891940336726202, "grad_norm": 0.7764523029327393, "learning_rate": 4.396672501201334e-05, "loss": 0.2713, "num_input_tokens_seen": 12243328, "step": 9035 }, { "epoch": 0.28935407464310864, "grad_norm": 0.9065009951591492, "learning_rate": 4.396032769626899e-05, "loss": 0.3919, "num_input_tokens_seen": 12250128, "step": 9040 }, { "epoch": 0.2895141156135971, "grad_norm": 2.4001550674438477, "learning_rate": 4.395392745660384e-05, "loss": 0.512, "num_input_tokens_seen": 12256864, "step": 9045 }, { "epoch": 0.2896741565840855, "grad_norm": 0.9992057681083679, "learning_rate": 4.394752429400488e-05, "loss": 0.3866, "num_input_tokens_seen": 12263568, "step": 9050 }, { "epoch": 0.28983419755457396, "grad_norm": 0.8980765342712402, "learning_rate": 4.394111820945957e-05, "loss": 0.4651, "num_input_tokens_seen": 12270624, "step": 9055 }, { "epoch": 0.2899942385250624, "grad_norm": 0.6319069862365723, "learning_rate": 4.393470920395579e-05, "loss": 0.5422, "num_input_tokens_seen": 12277088, "step": 9060 }, { "epoch": 0.29015427949555084, "grad_norm": 1.5611071586608887, "learning_rate": 4.392829727848192e-05, "loss": 0.5744, "num_input_tokens_seen": 12283792, "step": 9065 }, { "epoch": 0.29031432046603933, "grad_norm": 1.3051036596298218, "learning_rate": 4.392188243402673e-05, "loss": 0.5981, "num_input_tokens_seen": 12290768, "step": 9070 }, { "epoch": 0.2904743614365278, "grad_norm": 1.0296107530593872, "learning_rate": 4.391546467157949e-05, "loss": 0.4438, "num_input_tokens_seen": 12297600, "step": 9075 }, { "epoch": 0.2906344024070162, "grad_norm": 0.8152165412902832, "learning_rate": 4.390904399212988e-05, "loss": 0.4031, "num_input_tokens_seen": 12304112, "step": 9080 }, { "epoch": 0.29079444337750465, "grad_norm": 1.1001421213150024, "learning_rate": 4.390262039666807e-05, "loss": 0.4932, "num_input_tokens_seen": 12310960, "step": 9085 }, { "epoch": 0.2909544843479931, "grad_norm": 0.7355651259422302, "learning_rate": 4.389619388618464e-05, "loss": 0.5569, "num_input_tokens_seen": 12317776, "step": 9090 }, { "epoch": 0.29111452531848153, "grad_norm": 1.4298479557037354, "learning_rate": 4.3889764461670655e-05, "loss": 0.3776, "num_input_tokens_seen": 12324288, "step": 9095 }, { "epoch": 0.29127456628896997, "grad_norm": 1.085601806640625, "learning_rate": 4.38833321241176e-05, "loss": 0.6225, "num_input_tokens_seen": 12330848, "step": 9100 }, { "epoch": 0.2914346072594584, "grad_norm": 1.2805695533752441, "learning_rate": 4.3876896874517434e-05, "loss": 0.4002, "num_input_tokens_seen": 12337696, "step": 9105 }, { "epoch": 0.29159464822994685, "grad_norm": 1.7919728755950928, "learning_rate": 4.3870458713862554e-05, "loss": 0.4728, "num_input_tokens_seen": 12344688, "step": 9110 }, { "epoch": 0.2917546892004353, "grad_norm": 1.1666122674942017, "learning_rate": 4.386401764314579e-05, "loss": 0.6008, "num_input_tokens_seen": 12351776, "step": 9115 }, { "epoch": 0.29191473017092373, "grad_norm": 0.9255778193473816, "learning_rate": 4.385757366336045e-05, "loss": 0.3956, "num_input_tokens_seen": 12358384, "step": 9120 }, { "epoch": 0.2920747711414122, "grad_norm": 0.842799723148346, "learning_rate": 4.385112677550027e-05, "loss": 0.5503, "num_input_tokens_seen": 12365056, "step": 9125 }, { "epoch": 0.29223481211190067, "grad_norm": 1.3720324039459229, "learning_rate": 4.384467698055945e-05, "loss": 0.5095, "num_input_tokens_seen": 12371904, "step": 9130 }, { "epoch": 0.2923948530823891, "grad_norm": 0.9420685768127441, "learning_rate": 4.383822427953261e-05, "loss": 0.644, "num_input_tokens_seen": 12378752, "step": 9135 }, { "epoch": 0.29255489405287755, "grad_norm": 1.2418285608291626, "learning_rate": 4.3831768673414864e-05, "loss": 0.5995, "num_input_tokens_seen": 12385680, "step": 9140 }, { "epoch": 0.292714935023366, "grad_norm": 1.2591557502746582, "learning_rate": 4.382531016320173e-05, "loss": 0.5792, "num_input_tokens_seen": 12392224, "step": 9145 }, { "epoch": 0.2928749759938544, "grad_norm": 1.0159894227981567, "learning_rate": 4.3818848749889184e-05, "loss": 0.5399, "num_input_tokens_seen": 12399008, "step": 9150 }, { "epoch": 0.29303501696434286, "grad_norm": 1.423160195350647, "learning_rate": 4.381238443447368e-05, "loss": 0.6158, "num_input_tokens_seen": 12405920, "step": 9155 }, { "epoch": 0.2931950579348313, "grad_norm": 1.972316861152649, "learning_rate": 4.380591721795208e-05, "loss": 0.4424, "num_input_tokens_seen": 12412800, "step": 9160 }, { "epoch": 0.29335509890531974, "grad_norm": 0.49511873722076416, "learning_rate": 4.3799447101321723e-05, "loss": 0.5361, "num_input_tokens_seen": 12419552, "step": 9165 }, { "epoch": 0.2935151398758082, "grad_norm": 1.197746992111206, "learning_rate": 4.379297408558036e-05, "loss": 0.6222, "num_input_tokens_seen": 12426304, "step": 9170 }, { "epoch": 0.2936751808462967, "grad_norm": 1.6103413105010986, "learning_rate": 4.378649817172624e-05, "loss": 0.4101, "num_input_tokens_seen": 12432832, "step": 9175 }, { "epoch": 0.2938352218167851, "grad_norm": 0.8368043899536133, "learning_rate": 4.378001936075801e-05, "loss": 0.5132, "num_input_tokens_seen": 12439616, "step": 9180 }, { "epoch": 0.29399526278727356, "grad_norm": 1.0705041885375977, "learning_rate": 4.377353765367479e-05, "loss": 0.437, "num_input_tokens_seen": 12446048, "step": 9185 }, { "epoch": 0.294155303757762, "grad_norm": 1.4478522539138794, "learning_rate": 4.376705305147614e-05, "loss": 0.5499, "num_input_tokens_seen": 12452496, "step": 9190 }, { "epoch": 0.29431534472825044, "grad_norm": 1.1363686323165894, "learning_rate": 4.376056555516206e-05, "loss": 0.438, "num_input_tokens_seen": 12459296, "step": 9195 }, { "epoch": 0.2944753856987389, "grad_norm": 3.2893404960632324, "learning_rate": 4.375407516573302e-05, "loss": 0.4256, "num_input_tokens_seen": 12466288, "step": 9200 }, { "epoch": 0.2944753856987389, "eval_loss": 0.5186631679534912, "eval_runtime": 499.6755, "eval_samples_per_second": 27.79, "eval_steps_per_second": 13.895, "num_input_tokens_seen": 12466288, "step": 9200 }, { "epoch": 0.2946354266692273, "grad_norm": 1.0090843439102173, "learning_rate": 4.3747581884189913e-05, "loss": 0.4523, "num_input_tokens_seen": 12472896, "step": 9205 }, { "epoch": 0.29479546763971576, "grad_norm": 1.1058765649795532, "learning_rate": 4.374108571153408e-05, "loss": 0.4248, "num_input_tokens_seen": 12479584, "step": 9210 }, { "epoch": 0.2949555086102042, "grad_norm": 1.1168686151504517, "learning_rate": 4.3734586648767316e-05, "loss": 0.5395, "num_input_tokens_seen": 12486304, "step": 9215 }, { "epoch": 0.29511554958069264, "grad_norm": 1.600426435470581, "learning_rate": 4.372808469689186e-05, "loss": 0.6375, "num_input_tokens_seen": 12493136, "step": 9220 }, { "epoch": 0.2952755905511811, "grad_norm": 0.9625144004821777, "learning_rate": 4.372157985691039e-05, "loss": 0.5628, "num_input_tokens_seen": 12499808, "step": 9225 }, { "epoch": 0.2954356315216696, "grad_norm": 1.771409273147583, "learning_rate": 4.371507212982603e-05, "loss": 0.5692, "num_input_tokens_seen": 12506432, "step": 9230 }, { "epoch": 0.295595672492158, "grad_norm": 0.9242608547210693, "learning_rate": 4.370856151664236e-05, "loss": 0.4194, "num_input_tokens_seen": 12512848, "step": 9235 }, { "epoch": 0.29575571346264645, "grad_norm": 0.9328120946884155, "learning_rate": 4.3702048018363404e-05, "loss": 0.5132, "num_input_tokens_seen": 12519856, "step": 9240 }, { "epoch": 0.2959157544331349, "grad_norm": 2.8572134971618652, "learning_rate": 4.369553163599362e-05, "loss": 0.7582, "num_input_tokens_seen": 12526688, "step": 9245 }, { "epoch": 0.29607579540362333, "grad_norm": 0.8801164031028748, "learning_rate": 4.3689012370537904e-05, "loss": 0.3626, "num_input_tokens_seen": 12533696, "step": 9250 }, { "epoch": 0.29623583637411177, "grad_norm": 1.0710172653198242, "learning_rate": 4.368249022300164e-05, "loss": 0.5556, "num_input_tokens_seen": 12539952, "step": 9255 }, { "epoch": 0.2963958773446002, "grad_norm": 0.8910101056098938, "learning_rate": 4.367596519439059e-05, "loss": 0.4843, "num_input_tokens_seen": 12546976, "step": 9260 }, { "epoch": 0.29655591831508865, "grad_norm": 0.7682412266731262, "learning_rate": 4.366943728571101e-05, "loss": 0.3109, "num_input_tokens_seen": 12553264, "step": 9265 }, { "epoch": 0.2967159592855771, "grad_norm": 1.2353624105453491, "learning_rate": 4.366290649796959e-05, "loss": 0.4701, "num_input_tokens_seen": 12560208, "step": 9270 }, { "epoch": 0.29687600025606553, "grad_norm": 1.8953580856323242, "learning_rate": 4.3656372832173456e-05, "loss": 0.4754, "num_input_tokens_seen": 12566832, "step": 9275 }, { "epoch": 0.29703604122655397, "grad_norm": 0.737454891204834, "learning_rate": 4.364983628933017e-05, "loss": 0.5306, "num_input_tokens_seen": 12573264, "step": 9280 }, { "epoch": 0.29719608219704247, "grad_norm": 1.7987117767333984, "learning_rate": 4.364329687044777e-05, "loss": 0.5192, "num_input_tokens_seen": 12580224, "step": 9285 }, { "epoch": 0.2973561231675309, "grad_norm": 1.320178747177124, "learning_rate": 4.36367545765347e-05, "loss": 0.7117, "num_input_tokens_seen": 12586560, "step": 9290 }, { "epoch": 0.29751616413801935, "grad_norm": 1.8982415199279785, "learning_rate": 4.363020940859988e-05, "loss": 0.5129, "num_input_tokens_seen": 12592944, "step": 9295 }, { "epoch": 0.2976762051085078, "grad_norm": 0.9117767810821533, "learning_rate": 4.362366136765263e-05, "loss": 0.3934, "num_input_tokens_seen": 12599776, "step": 9300 }, { "epoch": 0.2978362460789962, "grad_norm": 1.7693703174591064, "learning_rate": 4.361711045470278e-05, "loss": 0.6122, "num_input_tokens_seen": 12606640, "step": 9305 }, { "epoch": 0.29799628704948466, "grad_norm": 0.5383793115615845, "learning_rate": 4.3610556670760524e-05, "loss": 0.3101, "num_input_tokens_seen": 12613136, "step": 9310 }, { "epoch": 0.2981563280199731, "grad_norm": 1.179479718208313, "learning_rate": 4.360400001683657e-05, "loss": 0.4776, "num_input_tokens_seen": 12620240, "step": 9315 }, { "epoch": 0.29831636899046154, "grad_norm": 0.7947261333465576, "learning_rate": 4.3597440493942e-05, "loss": 0.5403, "num_input_tokens_seen": 12626768, "step": 9320 }, { "epoch": 0.29847640996095, "grad_norm": 1.1443322896957397, "learning_rate": 4.3590878103088405e-05, "loss": 0.4603, "num_input_tokens_seen": 12633360, "step": 9325 }, { "epoch": 0.2986364509314384, "grad_norm": 1.173248291015625, "learning_rate": 4.358431284528779e-05, "loss": 0.4451, "num_input_tokens_seen": 12640288, "step": 9330 }, { "epoch": 0.2987964919019269, "grad_norm": 0.8970076441764832, "learning_rate": 4.357774472155257e-05, "loss": 0.3976, "num_input_tokens_seen": 12647232, "step": 9335 }, { "epoch": 0.29895653287241536, "grad_norm": 1.2251631021499634, "learning_rate": 4.3571173732895664e-05, "loss": 0.6457, "num_input_tokens_seen": 12653760, "step": 9340 }, { "epoch": 0.2991165738429038, "grad_norm": 1.0419772863388062, "learning_rate": 4.356459988033039e-05, "loss": 0.7071, "num_input_tokens_seen": 12660592, "step": 9345 }, { "epoch": 0.29927661481339224, "grad_norm": 0.7106374502182007, "learning_rate": 4.355802316487051e-05, "loss": 0.4407, "num_input_tokens_seen": 12667024, "step": 9350 }, { "epoch": 0.2994366557838807, "grad_norm": 1.0809282064437866, "learning_rate": 4.355144358753025e-05, "loss": 0.4969, "num_input_tokens_seen": 12673584, "step": 9355 }, { "epoch": 0.2995966967543691, "grad_norm": 1.2163047790527344, "learning_rate": 4.354486114932425e-05, "loss": 0.4182, "num_input_tokens_seen": 12680096, "step": 9360 }, { "epoch": 0.29975673772485756, "grad_norm": 1.354811668395996, "learning_rate": 4.353827585126762e-05, "loss": 0.5378, "num_input_tokens_seen": 12686720, "step": 9365 }, { "epoch": 0.299916778695346, "grad_norm": 0.8613168597221375, "learning_rate": 4.353168769437588e-05, "loss": 0.5179, "num_input_tokens_seen": 12693840, "step": 9370 }, { "epoch": 0.30007681966583444, "grad_norm": 0.8837816715240479, "learning_rate": 4.3525096679665014e-05, "loss": 0.4619, "num_input_tokens_seen": 12700640, "step": 9375 }, { "epoch": 0.3002368606363229, "grad_norm": 0.8180656433105469, "learning_rate": 4.351850280815144e-05, "loss": 0.4897, "num_input_tokens_seen": 12707888, "step": 9380 }, { "epoch": 0.3003969016068113, "grad_norm": 0.40282681584358215, "learning_rate": 4.3511906080852014e-05, "loss": 0.4161, "num_input_tokens_seen": 12714992, "step": 9385 }, { "epoch": 0.3005569425772998, "grad_norm": 1.6249908208847046, "learning_rate": 4.350530649878404e-05, "loss": 0.6917, "num_input_tokens_seen": 12721776, "step": 9390 }, { "epoch": 0.30071698354778825, "grad_norm": 0.978312075138092, "learning_rate": 4.3498704062965246e-05, "loss": 0.3611, "num_input_tokens_seen": 12728608, "step": 9395 }, { "epoch": 0.3008770245182767, "grad_norm": 0.8576020002365112, "learning_rate": 4.3492098774413815e-05, "loss": 0.431, "num_input_tokens_seen": 12735104, "step": 9400 }, { "epoch": 0.3008770245182767, "eval_loss": 0.5154364705085754, "eval_runtime": 499.0518, "eval_samples_per_second": 27.825, "eval_steps_per_second": 13.912, "num_input_tokens_seen": 12735104, "step": 9400 }, { "epoch": 0.30103706548876513, "grad_norm": 1.019360065460205, "learning_rate": 4.3485490634148375e-05, "loss": 0.4481, "num_input_tokens_seen": 12741664, "step": 9405 }, { "epoch": 0.30119710645925357, "grad_norm": 0.9882532358169556, "learning_rate": 4.347887964318797e-05, "loss": 0.638, "num_input_tokens_seen": 12748288, "step": 9410 }, { "epoch": 0.301357147429742, "grad_norm": 1.6544272899627686, "learning_rate": 4.34722658025521e-05, "loss": 0.4631, "num_input_tokens_seen": 12755344, "step": 9415 }, { "epoch": 0.30151718840023045, "grad_norm": 0.5869632363319397, "learning_rate": 4.346564911326071e-05, "loss": 0.4359, "num_input_tokens_seen": 12762096, "step": 9420 }, { "epoch": 0.3016772293707189, "grad_norm": 1.2045897245407104, "learning_rate": 4.345902957633418e-05, "loss": 0.4526, "num_input_tokens_seen": 12768928, "step": 9425 }, { "epoch": 0.30183727034120733, "grad_norm": 1.5490912199020386, "learning_rate": 4.345240719279331e-05, "loss": 0.5109, "num_input_tokens_seen": 12775856, "step": 9430 }, { "epoch": 0.30199731131169577, "grad_norm": 0.9508474469184875, "learning_rate": 4.3445781963659374e-05, "loss": 0.4033, "num_input_tokens_seen": 12782160, "step": 9435 }, { "epoch": 0.30215735228218427, "grad_norm": 0.9111089706420898, "learning_rate": 4.3439153889954045e-05, "loss": 0.5446, "num_input_tokens_seen": 12788992, "step": 9440 }, { "epoch": 0.3023173932526727, "grad_norm": 0.6706569194793701, "learning_rate": 4.343252297269946e-05, "loss": 0.3775, "num_input_tokens_seen": 12795808, "step": 9445 }, { "epoch": 0.30247743422316115, "grad_norm": 2.8129398822784424, "learning_rate": 4.342588921291821e-05, "loss": 0.4921, "num_input_tokens_seen": 12802640, "step": 9450 }, { "epoch": 0.3026374751936496, "grad_norm": 0.7577125430107117, "learning_rate": 4.341925261163328e-05, "loss": 0.4231, "num_input_tokens_seen": 12809552, "step": 9455 }, { "epoch": 0.302797516164138, "grad_norm": 0.8871411681175232, "learning_rate": 4.341261316986813e-05, "loss": 0.4723, "num_input_tokens_seen": 12816048, "step": 9460 }, { "epoch": 0.30295755713462647, "grad_norm": 0.944135844707489, "learning_rate": 4.340597088864664e-05, "loss": 0.5584, "num_input_tokens_seen": 12823024, "step": 9465 }, { "epoch": 0.3031175981051149, "grad_norm": 1.0453360080718994, "learning_rate": 4.339932576899313e-05, "loss": 0.4672, "num_input_tokens_seen": 12829904, "step": 9470 }, { "epoch": 0.30327763907560334, "grad_norm": 1.3465038537979126, "learning_rate": 4.3392677811932375e-05, "loss": 0.5777, "num_input_tokens_seen": 12836592, "step": 9475 }, { "epoch": 0.3034376800460918, "grad_norm": 0.7241483926773071, "learning_rate": 4.338602701848956e-05, "loss": 0.4065, "num_input_tokens_seen": 12843280, "step": 9480 }, { "epoch": 0.3035977210165802, "grad_norm": 1.180215835571289, "learning_rate": 4.337937338969033e-05, "loss": 0.5399, "num_input_tokens_seen": 12849840, "step": 9485 }, { "epoch": 0.30375776198706866, "grad_norm": 0.9173693060874939, "learning_rate": 4.337271692656075e-05, "loss": 0.3614, "num_input_tokens_seen": 12856896, "step": 9490 }, { "epoch": 0.30391780295755716, "grad_norm": 1.485828161239624, "learning_rate": 4.336605763012733e-05, "loss": 0.4446, "num_input_tokens_seen": 12864176, "step": 9495 }, { "epoch": 0.3040778439280456, "grad_norm": 1.1311774253845215, "learning_rate": 4.3359395501417026e-05, "loss": 0.4434, "num_input_tokens_seen": 12870704, "step": 9500 }, { "epoch": 0.30423788489853404, "grad_norm": 1.460583209991455, "learning_rate": 4.335273054145722e-05, "loss": 0.7562, "num_input_tokens_seen": 12877088, "step": 9505 }, { "epoch": 0.3043979258690225, "grad_norm": 1.315495491027832, "learning_rate": 4.334606275127572e-05, "loss": 0.6144, "num_input_tokens_seen": 12883440, "step": 9510 }, { "epoch": 0.3045579668395109, "grad_norm": 0.7104380130767822, "learning_rate": 4.33393921319008e-05, "loss": 0.4276, "num_input_tokens_seen": 12889776, "step": 9515 }, { "epoch": 0.30471800780999936, "grad_norm": 0.818681538105011, "learning_rate": 4.3332718684361146e-05, "loss": 0.4875, "num_input_tokens_seen": 12896448, "step": 9520 }, { "epoch": 0.3048780487804878, "grad_norm": 1.1894735097885132, "learning_rate": 4.332604240968588e-05, "loss": 0.4467, "num_input_tokens_seen": 12903280, "step": 9525 }, { "epoch": 0.30503808975097624, "grad_norm": 0.5754191279411316, "learning_rate": 4.331936330890459e-05, "loss": 0.4108, "num_input_tokens_seen": 12910208, "step": 9530 }, { "epoch": 0.3051981307214647, "grad_norm": 1.4703660011291504, "learning_rate": 4.331268138304725e-05, "loss": 0.5291, "num_input_tokens_seen": 12916384, "step": 9535 }, { "epoch": 0.3053581716919531, "grad_norm": 3.9957985877990723, "learning_rate": 4.330599663314431e-05, "loss": 0.4787, "num_input_tokens_seen": 12922992, "step": 9540 }, { "epoch": 0.3055182126624416, "grad_norm": 1.64626944065094, "learning_rate": 4.329930906022665e-05, "loss": 0.4741, "num_input_tokens_seen": 12929664, "step": 9545 }, { "epoch": 0.30567825363293005, "grad_norm": 0.9111537933349609, "learning_rate": 4.3292618665325564e-05, "loss": 0.3269, "num_input_tokens_seen": 12936368, "step": 9550 }, { "epoch": 0.3058382946034185, "grad_norm": 2.5575315952301025, "learning_rate": 4.3285925449472796e-05, "loss": 0.5778, "num_input_tokens_seen": 12943072, "step": 9555 }, { "epoch": 0.30599833557390693, "grad_norm": 0.8210638761520386, "learning_rate": 4.327922941370054e-05, "loss": 0.5244, "num_input_tokens_seen": 12950016, "step": 9560 }, { "epoch": 0.30615837654439537, "grad_norm": 1.4683457612991333, "learning_rate": 4.3272530559041384e-05, "loss": 0.4502, "num_input_tokens_seen": 12956976, "step": 9565 }, { "epoch": 0.3063184175148838, "grad_norm": 1.162642240524292, "learning_rate": 4.32658288865284e-05, "loss": 0.4234, "num_input_tokens_seen": 12963648, "step": 9570 }, { "epoch": 0.30647845848537225, "grad_norm": 1.098061442375183, "learning_rate": 4.325912439719505e-05, "loss": 0.5064, "num_input_tokens_seen": 12969936, "step": 9575 }, { "epoch": 0.3066384994558607, "grad_norm": 1.6249048709869385, "learning_rate": 4.3252417092075266e-05, "loss": 0.6023, "num_input_tokens_seen": 12976208, "step": 9580 }, { "epoch": 0.30679854042634913, "grad_norm": 1.3157069683074951, "learning_rate": 4.3245706972203385e-05, "loss": 0.441, "num_input_tokens_seen": 12982896, "step": 9585 }, { "epoch": 0.30695858139683757, "grad_norm": 0.83592289686203, "learning_rate": 4.323899403861421e-05, "loss": 0.4308, "num_input_tokens_seen": 12989520, "step": 9590 }, { "epoch": 0.307118622367326, "grad_norm": 1.038456678390503, "learning_rate": 4.3232278292342935e-05, "loss": 0.6239, "num_input_tokens_seen": 12996544, "step": 9595 }, { "epoch": 0.3072786633378145, "grad_norm": 0.9990360140800476, "learning_rate": 4.322555973442524e-05, "loss": 0.4597, "num_input_tokens_seen": 13003216, "step": 9600 }, { "epoch": 0.3072786633378145, "eval_loss": 0.5135153532028198, "eval_runtime": 499.4741, "eval_samples_per_second": 27.801, "eval_steps_per_second": 13.901, "num_input_tokens_seen": 13003216, "step": 9600 }, { "epoch": 0.30743870430830295, "grad_norm": 0.913765549659729, "learning_rate": 4.3218838365897184e-05, "loss": 0.3893, "num_input_tokens_seen": 13009808, "step": 9605 }, { "epoch": 0.3075987452787914, "grad_norm": 0.7195990681648254, "learning_rate": 4.3212114187795306e-05, "loss": 0.4274, "num_input_tokens_seen": 13016816, "step": 9610 }, { "epoch": 0.3077587862492798, "grad_norm": 1.223717451095581, "learning_rate": 4.320538720115656e-05, "loss": 0.4105, "num_input_tokens_seen": 13023472, "step": 9615 }, { "epoch": 0.30791882721976827, "grad_norm": 0.7213975787162781, "learning_rate": 4.319865740701831e-05, "loss": 0.6827, "num_input_tokens_seen": 13030192, "step": 9620 }, { "epoch": 0.3080788681902567, "grad_norm": 0.988945484161377, "learning_rate": 4.3191924806418396e-05, "loss": 0.4043, "num_input_tokens_seen": 13036896, "step": 9625 }, { "epoch": 0.30823890916074514, "grad_norm": 1.3011093139648438, "learning_rate": 4.318518940039507e-05, "loss": 0.6324, "num_input_tokens_seen": 13043280, "step": 9630 }, { "epoch": 0.3083989501312336, "grad_norm": 2.7505524158477783, "learning_rate": 4.3178451189987e-05, "loss": 0.2878, "num_input_tokens_seen": 13050336, "step": 9635 }, { "epoch": 0.308558991101722, "grad_norm": 1.194014072418213, "learning_rate": 4.3171710176233315e-05, "loss": 0.3919, "num_input_tokens_seen": 13057296, "step": 9640 }, { "epoch": 0.30871903207221046, "grad_norm": 0.7643795609474182, "learning_rate": 4.316496636017355e-05, "loss": 0.4853, "num_input_tokens_seen": 13063984, "step": 9645 }, { "epoch": 0.3088790730426989, "grad_norm": 1.3463685512542725, "learning_rate": 4.315821974284771e-05, "loss": 0.531, "num_input_tokens_seen": 13070896, "step": 9650 }, { "epoch": 0.3090391140131874, "grad_norm": 0.7531354427337646, "learning_rate": 4.315147032529619e-05, "loss": 0.3344, "num_input_tokens_seen": 13077856, "step": 9655 }, { "epoch": 0.30919915498367584, "grad_norm": 1.099165916442871, "learning_rate": 4.3144718108559845e-05, "loss": 0.4084, "num_input_tokens_seen": 13084800, "step": 9660 }, { "epoch": 0.3093591959541643, "grad_norm": 0.7598006725311279, "learning_rate": 4.3137963093679945e-05, "loss": 0.5124, "num_input_tokens_seen": 13092048, "step": 9665 }, { "epoch": 0.3095192369246527, "grad_norm": 0.8009821772575378, "learning_rate": 4.31312052816982e-05, "loss": 0.4576, "num_input_tokens_seen": 13098960, "step": 9670 }, { "epoch": 0.30967927789514116, "grad_norm": 1.6942332983016968, "learning_rate": 4.312444467365675e-05, "loss": 0.4634, "num_input_tokens_seen": 13105968, "step": 9675 }, { "epoch": 0.3098393188656296, "grad_norm": 0.8738687634468079, "learning_rate": 4.311768127059816e-05, "loss": 0.3837, "num_input_tokens_seen": 13112496, "step": 9680 }, { "epoch": 0.30999935983611804, "grad_norm": 0.8284592628479004, "learning_rate": 4.3110915073565444e-05, "loss": 0.395, "num_input_tokens_seen": 13119056, "step": 9685 }, { "epoch": 0.3101594008066065, "grad_norm": 1.9888790845870972, "learning_rate": 4.310414608360203e-05, "loss": 0.6066, "num_input_tokens_seen": 13125792, "step": 9690 }, { "epoch": 0.3103194417770949, "grad_norm": 1.0048553943634033, "learning_rate": 4.309737430175177e-05, "loss": 0.5072, "num_input_tokens_seen": 13132560, "step": 9695 }, { "epoch": 0.31047948274758336, "grad_norm": 1.277685523033142, "learning_rate": 4.309059972905897e-05, "loss": 0.4311, "num_input_tokens_seen": 13139248, "step": 9700 }, { "epoch": 0.31063952371807185, "grad_norm": 1.6521636247634888, "learning_rate": 4.308382236656836e-05, "loss": 0.5059, "num_input_tokens_seen": 13146080, "step": 9705 }, { "epoch": 0.3107995646885603, "grad_norm": 1.1301528215408325, "learning_rate": 4.307704221532507e-05, "loss": 0.3796, "num_input_tokens_seen": 13152640, "step": 9710 }, { "epoch": 0.31095960565904873, "grad_norm": 1.2536603212356567, "learning_rate": 4.307025927637471e-05, "loss": 0.4801, "num_input_tokens_seen": 13159552, "step": 9715 }, { "epoch": 0.31111964662953717, "grad_norm": 1.888142704963684, "learning_rate": 4.306347355076328e-05, "loss": 0.4028, "num_input_tokens_seen": 13166128, "step": 9720 }, { "epoch": 0.3112796876000256, "grad_norm": 1.0957109928131104, "learning_rate": 4.305668503953724e-05, "loss": 0.5745, "num_input_tokens_seen": 13172768, "step": 9725 }, { "epoch": 0.31143972857051405, "grad_norm": 0.7988583445549011, "learning_rate": 4.3049893743743436e-05, "loss": 0.625, "num_input_tokens_seen": 13179888, "step": 9730 }, { "epoch": 0.3115997695410025, "grad_norm": 4.223735332489014, "learning_rate": 4.304309966442919e-05, "loss": 0.6443, "num_input_tokens_seen": 13186672, "step": 9735 }, { "epoch": 0.31175981051149093, "grad_norm": 1.372458815574646, "learning_rate": 4.303630280264224e-05, "loss": 0.5019, "num_input_tokens_seen": 13193568, "step": 9740 }, { "epoch": 0.31191985148197937, "grad_norm": 0.8214612007141113, "learning_rate": 4.302950315943074e-05, "loss": 0.7225, "num_input_tokens_seen": 13200256, "step": 9745 }, { "epoch": 0.3120798924524678, "grad_norm": 0.5425807237625122, "learning_rate": 4.3022700735843275e-05, "loss": 0.3425, "num_input_tokens_seen": 13206672, "step": 9750 }, { "epoch": 0.31223993342295625, "grad_norm": 1.081162691116333, "learning_rate": 4.301589553292887e-05, "loss": 0.559, "num_input_tokens_seen": 13213168, "step": 9755 }, { "epoch": 0.31239997439344475, "grad_norm": 0.4362720549106598, "learning_rate": 4.300908755173697e-05, "loss": 0.4584, "num_input_tokens_seen": 13220032, "step": 9760 }, { "epoch": 0.3125600153639332, "grad_norm": 1.1715987920761108, "learning_rate": 4.300227679331745e-05, "loss": 0.4574, "num_input_tokens_seen": 13226896, "step": 9765 }, { "epoch": 0.3127200563344216, "grad_norm": 0.7407302856445312, "learning_rate": 4.299546325872063e-05, "loss": 0.3981, "num_input_tokens_seen": 13233568, "step": 9770 }, { "epoch": 0.31288009730491007, "grad_norm": 0.6686212420463562, "learning_rate": 4.2988646948997225e-05, "loss": 0.4222, "num_input_tokens_seen": 13240496, "step": 9775 }, { "epoch": 0.3130401382753985, "grad_norm": 1.425742506980896, "learning_rate": 4.29818278651984e-05, "loss": 0.5703, "num_input_tokens_seen": 13247120, "step": 9780 }, { "epoch": 0.31320017924588694, "grad_norm": 2.430969476699829, "learning_rate": 4.297500600837574e-05, "loss": 0.646, "num_input_tokens_seen": 13253952, "step": 9785 }, { "epoch": 0.3133602202163754, "grad_norm": 1.8540688753128052, "learning_rate": 4.2968181379581276e-05, "loss": 0.6041, "num_input_tokens_seen": 13260736, "step": 9790 }, { "epoch": 0.3135202611868638, "grad_norm": 1.378282904624939, "learning_rate": 4.296135397986743e-05, "loss": 0.5549, "num_input_tokens_seen": 13267120, "step": 9795 }, { "epoch": 0.31368030215735226, "grad_norm": 1.2499737739562988, "learning_rate": 4.295452381028709e-05, "loss": 0.4979, "num_input_tokens_seen": 13273680, "step": 9800 }, { "epoch": 0.31368030215735226, "eval_loss": 0.5123282670974731, "eval_runtime": 499.937, "eval_samples_per_second": 27.776, "eval_steps_per_second": 13.888, "num_input_tokens_seen": 13273680, "step": 9800 }, { "epoch": 0.3138403431278407, "grad_norm": 1.1926053762435913, "learning_rate": 4.294769087189354e-05, "loss": 0.445, "num_input_tokens_seen": 13280800, "step": 9805 }, { "epoch": 0.3140003840983292, "grad_norm": 1.7200326919555664, "learning_rate": 4.294085516574052e-05, "loss": 0.4267, "num_input_tokens_seen": 13287824, "step": 9810 }, { "epoch": 0.31416042506881764, "grad_norm": 1.2852084636688232, "learning_rate": 4.2934016692882176e-05, "loss": 0.5815, "num_input_tokens_seen": 13294480, "step": 9815 }, { "epoch": 0.3143204660393061, "grad_norm": 1.0496866703033447, "learning_rate": 4.292717545437308e-05, "loss": 0.3685, "num_input_tokens_seen": 13301024, "step": 9820 }, { "epoch": 0.3144805070097945, "grad_norm": 0.8488487601280212, "learning_rate": 4.292033145126825e-05, "loss": 0.4131, "num_input_tokens_seen": 13307424, "step": 9825 }, { "epoch": 0.31464054798028296, "grad_norm": 3.3237011432647705, "learning_rate": 4.29134846846231e-05, "loss": 0.6169, "num_input_tokens_seen": 13313872, "step": 9830 }, { "epoch": 0.3148005889507714, "grad_norm": 1.0091743469238281, "learning_rate": 4.29066351554935e-05, "loss": 0.3414, "num_input_tokens_seen": 13320448, "step": 9835 }, { "epoch": 0.31496062992125984, "grad_norm": 2.222684621810913, "learning_rate": 4.289978286493574e-05, "loss": 0.4125, "num_input_tokens_seen": 13327040, "step": 9840 }, { "epoch": 0.3151206708917483, "grad_norm": 2.2756364345550537, "learning_rate": 4.28929278140065e-05, "loss": 0.5966, "num_input_tokens_seen": 13333808, "step": 9845 }, { "epoch": 0.3152807118622367, "grad_norm": 1.1344319581985474, "learning_rate": 4.288607000376295e-05, "loss": 0.4149, "num_input_tokens_seen": 13340224, "step": 9850 }, { "epoch": 0.31544075283272516, "grad_norm": 1.502275824546814, "learning_rate": 4.2879209435262624e-05, "loss": 0.5622, "num_input_tokens_seen": 13347552, "step": 9855 }, { "epoch": 0.3156007938032136, "grad_norm": 0.685745894908905, "learning_rate": 4.287234610956353e-05, "loss": 0.4485, "num_input_tokens_seen": 13354560, "step": 9860 }, { "epoch": 0.3157608347737021, "grad_norm": 1.315229892730713, "learning_rate": 4.2865480027724056e-05, "loss": 0.4633, "num_input_tokens_seen": 13361296, "step": 9865 }, { "epoch": 0.31592087574419053, "grad_norm": 0.5259175896644592, "learning_rate": 4.285861119080306e-05, "loss": 0.5912, "num_input_tokens_seen": 13368288, "step": 9870 }, { "epoch": 0.31608091671467897, "grad_norm": 1.7481180429458618, "learning_rate": 4.2851739599859784e-05, "loss": 0.4506, "num_input_tokens_seen": 13374960, "step": 9875 }, { "epoch": 0.3162409576851674, "grad_norm": 0.6122747659683228, "learning_rate": 4.2844865255953934e-05, "loss": 0.3984, "num_input_tokens_seen": 13381888, "step": 9880 }, { "epoch": 0.31640099865565585, "grad_norm": 0.8488221764564514, "learning_rate": 4.2837988160145605e-05, "loss": 0.8733, "num_input_tokens_seen": 13388688, "step": 9885 }, { "epoch": 0.3165610396261443, "grad_norm": 1.7119896411895752, "learning_rate": 4.2831108313495336e-05, "loss": 0.4401, "num_input_tokens_seen": 13395888, "step": 9890 }, { "epoch": 0.31672108059663273, "grad_norm": 0.7904788851737976, "learning_rate": 4.282422571706408e-05, "loss": 0.2795, "num_input_tokens_seen": 13403296, "step": 9895 }, { "epoch": 0.31688112156712117, "grad_norm": 1.4088704586029053, "learning_rate": 4.281734037191323e-05, "loss": 0.4177, "num_input_tokens_seen": 13409696, "step": 9900 }, { "epoch": 0.3170411625376096, "grad_norm": 1.083701729774475, "learning_rate": 4.281045227910459e-05, "loss": 0.362, "num_input_tokens_seen": 13416416, "step": 9905 }, { "epoch": 0.31720120350809805, "grad_norm": 1.0887023210525513, "learning_rate": 4.280356143970038e-05, "loss": 0.5564, "num_input_tokens_seen": 13423152, "step": 9910 }, { "epoch": 0.3173612444785865, "grad_norm": 1.0324517488479614, "learning_rate": 4.279666785476327e-05, "loss": 0.561, "num_input_tokens_seen": 13430368, "step": 9915 }, { "epoch": 0.317521285449075, "grad_norm": 1.1440773010253906, "learning_rate": 4.2789771525356325e-05, "loss": 0.5798, "num_input_tokens_seen": 13437840, "step": 9920 }, { "epoch": 0.3176813264195634, "grad_norm": 0.9306548237800598, "learning_rate": 4.2782872452543056e-05, "loss": 0.4918, "num_input_tokens_seen": 13444528, "step": 9925 }, { "epoch": 0.31784136739005187, "grad_norm": 1.148480772972107, "learning_rate": 4.2775970637387376e-05, "loss": 0.5315, "num_input_tokens_seen": 13451376, "step": 9930 }, { "epoch": 0.3180014083605403, "grad_norm": 0.6381218433380127, "learning_rate": 4.276906608095363e-05, "loss": 0.2745, "num_input_tokens_seen": 13458304, "step": 9935 }, { "epoch": 0.31816144933102875, "grad_norm": 1.4769713878631592, "learning_rate": 4.276215878430661e-05, "loss": 0.4818, "num_input_tokens_seen": 13464752, "step": 9940 }, { "epoch": 0.3183214903015172, "grad_norm": 0.785839319229126, "learning_rate": 4.275524874851149e-05, "loss": 0.4226, "num_input_tokens_seen": 13471216, "step": 9945 }, { "epoch": 0.3184815312720056, "grad_norm": 1.2755985260009766, "learning_rate": 4.274833597463388e-05, "loss": 0.4837, "num_input_tokens_seen": 13478080, "step": 9950 }, { "epoch": 0.31864157224249406, "grad_norm": 1.1048901081085205, "learning_rate": 4.2741420463739824e-05, "loss": 0.5062, "num_input_tokens_seen": 13485040, "step": 9955 }, { "epoch": 0.3188016132129825, "grad_norm": 2.258420705795288, "learning_rate": 4.273450221689578e-05, "loss": 0.5196, "num_input_tokens_seen": 13491808, "step": 9960 }, { "epoch": 0.31896165418347094, "grad_norm": 3.0660934448242188, "learning_rate": 4.272758123516863e-05, "loss": 0.6012, "num_input_tokens_seen": 13498592, "step": 9965 }, { "epoch": 0.31912169515395944, "grad_norm": 1.3236833810806274, "learning_rate": 4.272065751962567e-05, "loss": 0.5218, "num_input_tokens_seen": 13505632, "step": 9970 }, { "epoch": 0.3192817361244479, "grad_norm": 1.1446890830993652, "learning_rate": 4.271373107133464e-05, "loss": 0.6394, "num_input_tokens_seen": 13512288, "step": 9975 }, { "epoch": 0.3194417770949363, "grad_norm": 0.7269842028617859, "learning_rate": 4.270680189136366e-05, "loss": 0.5036, "num_input_tokens_seen": 13519232, "step": 9980 }, { "epoch": 0.31960181806542476, "grad_norm": 1.056738018989563, "learning_rate": 4.269986998078132e-05, "loss": 0.46, "num_input_tokens_seen": 13525584, "step": 9985 }, { "epoch": 0.3197618590359132, "grad_norm": 1.6615192890167236, "learning_rate": 4.2692935340656595e-05, "loss": 0.6701, "num_input_tokens_seen": 13532592, "step": 9990 }, { "epoch": 0.31992190000640164, "grad_norm": 0.6908645629882812, "learning_rate": 4.26859979720589e-05, "loss": 0.416, "num_input_tokens_seen": 13539472, "step": 9995 }, { "epoch": 0.3200819409768901, "grad_norm": 1.0011719465255737, "learning_rate": 4.267905787605806e-05, "loss": 0.6121, "num_input_tokens_seen": 13545840, "step": 10000 }, { "epoch": 0.3200819409768901, "eval_loss": 0.510772705078125, "eval_runtime": 499.4352, "eval_samples_per_second": 27.803, "eval_steps_per_second": 13.902, "num_input_tokens_seen": 13545840, "step": 10000 }, { "epoch": 0.3202419819473785, "grad_norm": 1.0586612224578857, "learning_rate": 4.267211505372433e-05, "loss": 0.6113, "num_input_tokens_seen": 13552768, "step": 10005 }, { "epoch": 0.32040202291786696, "grad_norm": 0.7963998317718506, "learning_rate": 4.266516950612837e-05, "loss": 0.3726, "num_input_tokens_seen": 13559424, "step": 10010 }, { "epoch": 0.3205620638883554, "grad_norm": 2.0225722789764404, "learning_rate": 4.265822123434128e-05, "loss": 0.4522, "num_input_tokens_seen": 13566384, "step": 10015 }, { "epoch": 0.32072210485884384, "grad_norm": 0.9040815234184265, "learning_rate": 4.265127023943457e-05, "loss": 0.3948, "num_input_tokens_seen": 13573328, "step": 10020 }, { "epoch": 0.32088214582933233, "grad_norm": 1.088191270828247, "learning_rate": 4.2644316522480176e-05, "loss": 0.3433, "num_input_tokens_seen": 13580288, "step": 10025 }, { "epoch": 0.3210421867998208, "grad_norm": 1.8973618745803833, "learning_rate": 4.263736008455044e-05, "loss": 0.5714, "num_input_tokens_seen": 13587008, "step": 10030 }, { "epoch": 0.3212022277703092, "grad_norm": 1.1275900602340698, "learning_rate": 4.2630400926718125e-05, "loss": 0.5916, "num_input_tokens_seen": 13594608, "step": 10035 }, { "epoch": 0.32136226874079765, "grad_norm": 0.8555486798286438, "learning_rate": 4.262343905005644e-05, "loss": 0.5085, "num_input_tokens_seen": 13601232, "step": 10040 }, { "epoch": 0.3215223097112861, "grad_norm": 1.0065593719482422, "learning_rate": 4.261647445563897e-05, "loss": 0.4515, "num_input_tokens_seen": 13608608, "step": 10045 }, { "epoch": 0.32168235068177453, "grad_norm": 2.556973457336426, "learning_rate": 4.260950714453976e-05, "loss": 0.4477, "num_input_tokens_seen": 13615120, "step": 10050 }, { "epoch": 0.32184239165226297, "grad_norm": 1.8654223680496216, "learning_rate": 4.2602537117833266e-05, "loss": 0.4944, "num_input_tokens_seen": 13621472, "step": 10055 }, { "epoch": 0.3220024326227514, "grad_norm": 0.6920918822288513, "learning_rate": 4.259556437659433e-05, "loss": 0.38, "num_input_tokens_seen": 13628096, "step": 10060 }, { "epoch": 0.32216247359323985, "grad_norm": 1.7909040451049805, "learning_rate": 4.258858892189825e-05, "loss": 0.664, "num_input_tokens_seen": 13634784, "step": 10065 }, { "epoch": 0.3223225145637283, "grad_norm": 1.2453696727752686, "learning_rate": 4.2581610754820725e-05, "loss": 0.4498, "num_input_tokens_seen": 13641728, "step": 10070 }, { "epoch": 0.3224825555342168, "grad_norm": 0.6363314390182495, "learning_rate": 4.2574629876437876e-05, "loss": 0.6551, "num_input_tokens_seen": 13648224, "step": 10075 }, { "epoch": 0.3226425965047052, "grad_norm": 1.0672516822814941, "learning_rate": 4.256764628782625e-05, "loss": 0.5318, "num_input_tokens_seen": 13654944, "step": 10080 }, { "epoch": 0.32280263747519367, "grad_norm": 1.2357068061828613, "learning_rate": 4.256065999006279e-05, "loss": 0.624, "num_input_tokens_seen": 13661392, "step": 10085 }, { "epoch": 0.3229626784456821, "grad_norm": 1.0741037130355835, "learning_rate": 4.2553670984224885e-05, "loss": 0.4555, "num_input_tokens_seen": 13667920, "step": 10090 }, { "epoch": 0.32312271941617055, "grad_norm": 0.9258628487586975, "learning_rate": 4.254667927139032e-05, "loss": 0.4431, "num_input_tokens_seen": 13674528, "step": 10095 }, { "epoch": 0.323282760386659, "grad_norm": 1.0763276815414429, "learning_rate": 4.2539684852637295e-05, "loss": 0.6253, "num_input_tokens_seen": 13681568, "step": 10100 }, { "epoch": 0.3234428013571474, "grad_norm": 1.1624611616134644, "learning_rate": 4.253268772904446e-05, "loss": 0.4552, "num_input_tokens_seen": 13688112, "step": 10105 }, { "epoch": 0.32360284232763586, "grad_norm": 1.0338085889816284, "learning_rate": 4.252568790169085e-05, "loss": 0.4514, "num_input_tokens_seen": 13695296, "step": 10110 }, { "epoch": 0.3237628832981243, "grad_norm": 1.2672921419143677, "learning_rate": 4.251868537165592e-05, "loss": 0.3601, "num_input_tokens_seen": 13702224, "step": 10115 }, { "epoch": 0.32392292426861274, "grad_norm": 1.2852483987808228, "learning_rate": 4.251168014001955e-05, "loss": 0.5355, "num_input_tokens_seen": 13708752, "step": 10120 }, { "epoch": 0.3240829652391012, "grad_norm": 1.0522832870483398, "learning_rate": 4.250467220786204e-05, "loss": 0.3662, "num_input_tokens_seen": 13715536, "step": 10125 }, { "epoch": 0.3242430062095897, "grad_norm": 1.7802231311798096, "learning_rate": 4.249766157626409e-05, "loss": 0.656, "num_input_tokens_seen": 13722544, "step": 10130 }, { "epoch": 0.3244030471800781, "grad_norm": 1.707129716873169, "learning_rate": 4.249064824630684e-05, "loss": 0.4099, "num_input_tokens_seen": 13729056, "step": 10135 }, { "epoch": 0.32456308815056656, "grad_norm": 1.1074413061141968, "learning_rate": 4.248363221907183e-05, "loss": 0.5082, "num_input_tokens_seen": 13735936, "step": 10140 }, { "epoch": 0.324723129121055, "grad_norm": 2.144171714782715, "learning_rate": 4.2476613495641026e-05, "loss": 0.6235, "num_input_tokens_seen": 13742624, "step": 10145 }, { "epoch": 0.32488317009154344, "grad_norm": 0.6461657285690308, "learning_rate": 4.246959207709679e-05, "loss": 0.3895, "num_input_tokens_seen": 13749200, "step": 10150 }, { "epoch": 0.3250432110620319, "grad_norm": 0.5876768827438354, "learning_rate": 4.246256796452192e-05, "loss": 0.2308, "num_input_tokens_seen": 13755760, "step": 10155 }, { "epoch": 0.3252032520325203, "grad_norm": 1.9805841445922852, "learning_rate": 4.245554115899962e-05, "loss": 0.576, "num_input_tokens_seen": 13762304, "step": 10160 }, { "epoch": 0.32536329300300876, "grad_norm": 1.2357723712921143, "learning_rate": 4.2448511661613514e-05, "loss": 0.5856, "num_input_tokens_seen": 13769120, "step": 10165 }, { "epoch": 0.3255233339734972, "grad_norm": 2.2740256786346436, "learning_rate": 4.2441479473447635e-05, "loss": 0.5526, "num_input_tokens_seen": 13776080, "step": 10170 }, { "epoch": 0.32568337494398564, "grad_norm": 1.5322740077972412, "learning_rate": 4.243444459558644e-05, "loss": 0.4784, "num_input_tokens_seen": 13783280, "step": 10175 }, { "epoch": 0.32584341591447413, "grad_norm": 1.3432756662368774, "learning_rate": 4.24274070291148e-05, "loss": 0.4427, "num_input_tokens_seen": 13790000, "step": 10180 }, { "epoch": 0.3260034568849626, "grad_norm": 0.8831014633178711, "learning_rate": 4.242036677511798e-05, "loss": 0.5146, "num_input_tokens_seen": 13796560, "step": 10185 }, { "epoch": 0.326163497855451, "grad_norm": 0.5200150609016418, "learning_rate": 4.241332383468169e-05, "loss": 0.4904, "num_input_tokens_seen": 13803568, "step": 10190 }, { "epoch": 0.32632353882593945, "grad_norm": 0.9536044597625732, "learning_rate": 4.2406278208892034e-05, "loss": 0.3635, "num_input_tokens_seen": 13810368, "step": 10195 }, { "epoch": 0.3264835797964279, "grad_norm": 0.7046099901199341, "learning_rate": 4.2399229898835536e-05, "loss": 0.5801, "num_input_tokens_seen": 13817104, "step": 10200 }, { "epoch": 0.3264835797964279, "eval_loss": 0.5097899436950684, "eval_runtime": 499.2555, "eval_samples_per_second": 27.813, "eval_steps_per_second": 13.907, "num_input_tokens_seen": 13817104, "step": 10200 }, { "epoch": 0.32664362076691633, "grad_norm": 1.3654016256332397, "learning_rate": 4.239217890559914e-05, "loss": 0.5078, "num_input_tokens_seen": 13823952, "step": 10205 }, { "epoch": 0.32680366173740477, "grad_norm": 2.5637359619140625, "learning_rate": 4.238512523027019e-05, "loss": 0.478, "num_input_tokens_seen": 13830816, "step": 10210 }, { "epoch": 0.3269637027078932, "grad_norm": 0.5371450185775757, "learning_rate": 4.237806887393645e-05, "loss": 0.3333, "num_input_tokens_seen": 13837664, "step": 10215 }, { "epoch": 0.32712374367838165, "grad_norm": 0.9858543276786804, "learning_rate": 4.237100983768611e-05, "loss": 0.5584, "num_input_tokens_seen": 13844752, "step": 10220 }, { "epoch": 0.3272837846488701, "grad_norm": 1.4177255630493164, "learning_rate": 4.2363948122607756e-05, "loss": 0.4014, "num_input_tokens_seen": 13851376, "step": 10225 }, { "epoch": 0.32744382561935853, "grad_norm": 2.182512044906616, "learning_rate": 4.235688372979039e-05, "loss": 0.5414, "num_input_tokens_seen": 13857696, "step": 10230 }, { "epoch": 0.327603866589847, "grad_norm": 1.157058596611023, "learning_rate": 4.234981666032343e-05, "loss": 0.5189, "num_input_tokens_seen": 13864448, "step": 10235 }, { "epoch": 0.32776390756033547, "grad_norm": 0.9449080228805542, "learning_rate": 4.2342746915296704e-05, "loss": 0.5137, "num_input_tokens_seen": 13871584, "step": 10240 }, { "epoch": 0.3279239485308239, "grad_norm": 1.1002483367919922, "learning_rate": 4.233567449580047e-05, "loss": 0.4036, "num_input_tokens_seen": 13878256, "step": 10245 }, { "epoch": 0.32808398950131235, "grad_norm": 0.8430812358856201, "learning_rate": 4.232859940292537e-05, "loss": 0.416, "num_input_tokens_seen": 13884624, "step": 10250 }, { "epoch": 0.3282440304718008, "grad_norm": 0.9132580757141113, "learning_rate": 4.232152163776248e-05, "loss": 0.5301, "num_input_tokens_seen": 13891792, "step": 10255 }, { "epoch": 0.3284040714422892, "grad_norm": 1.2965800762176514, "learning_rate": 4.231444120140328e-05, "loss": 0.4572, "num_input_tokens_seen": 13898464, "step": 10260 }, { "epoch": 0.32856411241277766, "grad_norm": 0.996343731880188, "learning_rate": 4.230735809493967e-05, "loss": 0.3923, "num_input_tokens_seen": 13905216, "step": 10265 }, { "epoch": 0.3287241533832661, "grad_norm": 1.5490655899047852, "learning_rate": 4.2300272319463926e-05, "loss": 0.5104, "num_input_tokens_seen": 13912112, "step": 10270 }, { "epoch": 0.32888419435375454, "grad_norm": 1.033564567565918, "learning_rate": 4.2293183876068786e-05, "loss": 0.43, "num_input_tokens_seen": 13918416, "step": 10275 }, { "epoch": 0.329044235324243, "grad_norm": 1.1530870199203491, "learning_rate": 4.228609276584737e-05, "loss": 0.4697, "num_input_tokens_seen": 13925440, "step": 10280 }, { "epoch": 0.3292042762947314, "grad_norm": 0.8389992117881775, "learning_rate": 4.227899898989323e-05, "loss": 0.4595, "num_input_tokens_seen": 13932304, "step": 10285 }, { "epoch": 0.3293643172652199, "grad_norm": 1.3569247722625732, "learning_rate": 4.2271902549300293e-05, "loss": 0.5853, "num_input_tokens_seen": 13939328, "step": 10290 }, { "epoch": 0.32952435823570836, "grad_norm": 1.518251895904541, "learning_rate": 4.226480344516294e-05, "loss": 0.5939, "num_input_tokens_seen": 13946032, "step": 10295 }, { "epoch": 0.3296843992061968, "grad_norm": 1.2084466218948364, "learning_rate": 4.2257701678575925e-05, "loss": 0.5329, "num_input_tokens_seen": 13953072, "step": 10300 }, { "epoch": 0.32984444017668524, "grad_norm": 3.3548359870910645, "learning_rate": 4.225059725063444e-05, "loss": 0.6243, "num_input_tokens_seen": 13959824, "step": 10305 }, { "epoch": 0.3300044811471737, "grad_norm": 0.9076951742172241, "learning_rate": 4.2243490162434074e-05, "loss": 0.436, "num_input_tokens_seen": 13966496, "step": 10310 }, { "epoch": 0.3301645221176621, "grad_norm": 1.3291454315185547, "learning_rate": 4.223638041507083e-05, "loss": 0.4009, "num_input_tokens_seen": 13973472, "step": 10315 }, { "epoch": 0.33032456308815056, "grad_norm": 0.8479189276695251, "learning_rate": 4.2229268009641124e-05, "loss": 0.3838, "num_input_tokens_seen": 13979968, "step": 10320 }, { "epoch": 0.330484604058639, "grad_norm": 0.5785309672355652, "learning_rate": 4.222215294724177e-05, "loss": 0.4541, "num_input_tokens_seen": 13986512, "step": 10325 }, { "epoch": 0.33064464502912744, "grad_norm": 0.911967933177948, "learning_rate": 4.2215035228970005e-05, "loss": 0.3698, "num_input_tokens_seen": 13993760, "step": 10330 }, { "epoch": 0.3308046859996159, "grad_norm": 1.6852694749832153, "learning_rate": 4.2207914855923464e-05, "loss": 0.5084, "num_input_tokens_seen": 14000400, "step": 10335 }, { "epoch": 0.3309647269701044, "grad_norm": 1.1965476274490356, "learning_rate": 4.220079182920021e-05, "loss": 0.359, "num_input_tokens_seen": 14006848, "step": 10340 }, { "epoch": 0.3311247679405928, "grad_norm": 0.9386085271835327, "learning_rate": 4.2193666149898705e-05, "loss": 0.5386, "num_input_tokens_seen": 14013360, "step": 10345 }, { "epoch": 0.33128480891108125, "grad_norm": 1.3520170450210571, "learning_rate": 4.21865378191178e-05, "loss": 0.3762, "num_input_tokens_seen": 14020480, "step": 10350 }, { "epoch": 0.3314448498815697, "grad_norm": 1.9329190254211426, "learning_rate": 4.217940683795678e-05, "loss": 0.4961, "num_input_tokens_seen": 14027216, "step": 10355 }, { "epoch": 0.33160489085205813, "grad_norm": 0.9994560480117798, "learning_rate": 4.217227320751534e-05, "loss": 0.5026, "num_input_tokens_seen": 14034032, "step": 10360 }, { "epoch": 0.33176493182254657, "grad_norm": 0.9188523888587952, "learning_rate": 4.216513692889358e-05, "loss": 0.4821, "num_input_tokens_seen": 14040544, "step": 10365 }, { "epoch": 0.331924972793035, "grad_norm": 1.5020651817321777, "learning_rate": 4.215799800319199e-05, "loss": 0.4263, "num_input_tokens_seen": 14047488, "step": 10370 }, { "epoch": 0.33208501376352345, "grad_norm": 2.0331966876983643, "learning_rate": 4.2150856431511485e-05, "loss": 0.6279, "num_input_tokens_seen": 14054128, "step": 10375 }, { "epoch": 0.3322450547340119, "grad_norm": 1.4499249458312988, "learning_rate": 4.214371221495339e-05, "loss": 0.555, "num_input_tokens_seen": 14060576, "step": 10380 }, { "epoch": 0.33240509570450033, "grad_norm": 2.0847291946411133, "learning_rate": 4.213656535461942e-05, "loss": 0.4885, "num_input_tokens_seen": 14067392, "step": 10385 }, { "epoch": 0.33256513667498877, "grad_norm": 0.7537692189216614, "learning_rate": 4.2129415851611734e-05, "loss": 0.5923, "num_input_tokens_seen": 14074064, "step": 10390 }, { "epoch": 0.33272517764547727, "grad_norm": 1.5796473026275635, "learning_rate": 4.2122263707032855e-05, "loss": 0.4257, "num_input_tokens_seen": 14080896, "step": 10395 }, { "epoch": 0.3328852186159657, "grad_norm": 1.7728558778762817, "learning_rate": 4.211510892198574e-05, "loss": 0.4489, "num_input_tokens_seen": 14088032, "step": 10400 }, { "epoch": 0.3328852186159657, "eval_loss": 0.50827956199646, "eval_runtime": 499.2638, "eval_samples_per_second": 27.813, "eval_steps_per_second": 13.906, "num_input_tokens_seen": 14088032, "step": 10400 }, { "epoch": 0.33304525958645415, "grad_norm": 0.6319695711135864, "learning_rate": 4.210795149757375e-05, "loss": 0.4187, "num_input_tokens_seen": 14094608, "step": 10405 }, { "epoch": 0.3332053005569426, "grad_norm": 1.2269965410232544, "learning_rate": 4.210079143490065e-05, "loss": 0.4047, "num_input_tokens_seen": 14101504, "step": 10410 }, { "epoch": 0.333365341527431, "grad_norm": 1.1124961376190186, "learning_rate": 4.2093628735070604e-05, "loss": 0.4663, "num_input_tokens_seen": 14108496, "step": 10415 }, { "epoch": 0.33352538249791946, "grad_norm": 0.8874116539955139, "learning_rate": 4.208646339918819e-05, "loss": 0.3384, "num_input_tokens_seen": 14115248, "step": 10420 }, { "epoch": 0.3336854234684079, "grad_norm": 1.2242389917373657, "learning_rate": 4.2079295428358414e-05, "loss": 0.4928, "num_input_tokens_seen": 14121952, "step": 10425 }, { "epoch": 0.33384546443889634, "grad_norm": 1.5004092454910278, "learning_rate": 4.207212482368664e-05, "loss": 0.6544, "num_input_tokens_seen": 14128848, "step": 10430 }, { "epoch": 0.3340055054093848, "grad_norm": 1.0549813508987427, "learning_rate": 4.206495158627867e-05, "loss": 0.432, "num_input_tokens_seen": 14135104, "step": 10435 }, { "epoch": 0.3341655463798732, "grad_norm": 1.268010139465332, "learning_rate": 4.205777571724073e-05, "loss": 0.5255, "num_input_tokens_seen": 14142032, "step": 10440 }, { "epoch": 0.3343255873503617, "grad_norm": 1.466845154762268, "learning_rate": 4.20505972176794e-05, "loss": 0.5669, "num_input_tokens_seen": 14148816, "step": 10445 }, { "epoch": 0.33448562832085016, "grad_norm": 1.2457060813903809, "learning_rate": 4.204341608870171e-05, "loss": 0.4172, "num_input_tokens_seen": 14155488, "step": 10450 }, { "epoch": 0.3346456692913386, "grad_norm": 0.927788257598877, "learning_rate": 4.203623233141508e-05, "loss": 0.4682, "num_input_tokens_seen": 14162240, "step": 10455 }, { "epoch": 0.33480571026182704, "grad_norm": 1.4316604137420654, "learning_rate": 4.2029045946927334e-05, "loss": 0.4673, "num_input_tokens_seen": 14168832, "step": 10460 }, { "epoch": 0.3349657512323155, "grad_norm": 0.9315875768661499, "learning_rate": 4.20218569363467e-05, "loss": 0.5147, "num_input_tokens_seen": 14175856, "step": 10465 }, { "epoch": 0.3351257922028039, "grad_norm": 1.1403148174285889, "learning_rate": 4.2014665300781834e-05, "loss": 0.4349, "num_input_tokens_seen": 14183088, "step": 10470 }, { "epoch": 0.33528583317329236, "grad_norm": 1.3778798580169678, "learning_rate": 4.200747104134174e-05, "loss": 0.3954, "num_input_tokens_seen": 14189952, "step": 10475 }, { "epoch": 0.3354458741437808, "grad_norm": 0.7836365103721619, "learning_rate": 4.200027415913588e-05, "loss": 0.5503, "num_input_tokens_seen": 14196672, "step": 10480 }, { "epoch": 0.33560591511426924, "grad_norm": 1.5909184217453003, "learning_rate": 4.1993074655274126e-05, "loss": 0.5203, "num_input_tokens_seen": 14203968, "step": 10485 }, { "epoch": 0.3357659560847577, "grad_norm": 0.8196151256561279, "learning_rate": 4.198587253086669e-05, "loss": 0.4294, "num_input_tokens_seen": 14211168, "step": 10490 }, { "epoch": 0.3359259970552461, "grad_norm": 1.2099329233169556, "learning_rate": 4.197866778702426e-05, "loss": 0.4952, "num_input_tokens_seen": 14217920, "step": 10495 }, { "epoch": 0.3360860380257346, "grad_norm": 0.67756587266922, "learning_rate": 4.197146042485789e-05, "loss": 0.372, "num_input_tokens_seen": 14224608, "step": 10500 }, { "epoch": 0.33624607899622305, "grad_norm": 1.004448413848877, "learning_rate": 4.1964250445479046e-05, "loss": 0.4551, "num_input_tokens_seen": 14231488, "step": 10505 }, { "epoch": 0.3364061199667115, "grad_norm": 2.813499689102173, "learning_rate": 4.19570378499996e-05, "loss": 0.5913, "num_input_tokens_seen": 14238784, "step": 10510 }, { "epoch": 0.33656616093719993, "grad_norm": 1.1935044527053833, "learning_rate": 4.194982263953182e-05, "loss": 0.6067, "num_input_tokens_seen": 14245360, "step": 10515 }, { "epoch": 0.33672620190768837, "grad_norm": 0.684749186038971, "learning_rate": 4.194260481518838e-05, "loss": 0.5985, "num_input_tokens_seen": 14251808, "step": 10520 }, { "epoch": 0.3368862428781768, "grad_norm": 0.8829809427261353, "learning_rate": 4.1935384378082366e-05, "loss": 0.4349, "num_input_tokens_seen": 14258624, "step": 10525 }, { "epoch": 0.33704628384866525, "grad_norm": 1.0038429498672485, "learning_rate": 4.1928161329327267e-05, "loss": 0.4045, "num_input_tokens_seen": 14265952, "step": 10530 }, { "epoch": 0.3372063248191537, "grad_norm": 0.8171869516372681, "learning_rate": 4.1920935670036945e-05, "loss": 0.4398, "num_input_tokens_seen": 14273264, "step": 10535 }, { "epoch": 0.33736636578964213, "grad_norm": 0.6512106657028198, "learning_rate": 4.1913707401325705e-05, "loss": 0.4346, "num_input_tokens_seen": 14279968, "step": 10540 }, { "epoch": 0.33752640676013057, "grad_norm": 1.2039273977279663, "learning_rate": 4.1906476524308235e-05, "loss": 0.4247, "num_input_tokens_seen": 14286624, "step": 10545 }, { "epoch": 0.33768644773061907, "grad_norm": 1.3127390146255493, "learning_rate": 4.189924304009962e-05, "loss": 0.6233, "num_input_tokens_seen": 14293600, "step": 10550 }, { "epoch": 0.3378464887011075, "grad_norm": 0.5015636682510376, "learning_rate": 4.189200694981537e-05, "loss": 0.393, "num_input_tokens_seen": 14300272, "step": 10555 }, { "epoch": 0.33800652967159595, "grad_norm": 1.23370361328125, "learning_rate": 4.188476825457136e-05, "loss": 0.4744, "num_input_tokens_seen": 14307280, "step": 10560 }, { "epoch": 0.3381665706420844, "grad_norm": 2.357438802719116, "learning_rate": 4.18775269554839e-05, "loss": 0.5057, "num_input_tokens_seen": 14313920, "step": 10565 }, { "epoch": 0.3383266116125728, "grad_norm": 1.6273813247680664, "learning_rate": 4.187028305366969e-05, "loss": 0.4913, "num_input_tokens_seen": 14321040, "step": 10570 }, { "epoch": 0.33848665258306126, "grad_norm": 0.6701481938362122, "learning_rate": 4.1863036550245824e-05, "loss": 0.4556, "num_input_tokens_seen": 14327952, "step": 10575 }, { "epoch": 0.3386466935535497, "grad_norm": 1.2676129341125488, "learning_rate": 4.1855787446329806e-05, "loss": 0.4145, "num_input_tokens_seen": 14334976, "step": 10580 }, { "epoch": 0.33880673452403814, "grad_norm": 0.8033298850059509, "learning_rate": 4.184853574303955e-05, "loss": 0.4526, "num_input_tokens_seen": 14341456, "step": 10585 }, { "epoch": 0.3389667754945266, "grad_norm": 1.6404563188552856, "learning_rate": 4.184128144149334e-05, "loss": 0.5729, "num_input_tokens_seen": 14348384, "step": 10590 }, { "epoch": 0.339126816465015, "grad_norm": 0.6764892339706421, "learning_rate": 4.1834024542809896e-05, "loss": 0.4767, "num_input_tokens_seen": 14354912, "step": 10595 }, { "epoch": 0.33928685743550346, "grad_norm": 1.2032915353775024, "learning_rate": 4.1826765048108315e-05, "loss": 0.5318, "num_input_tokens_seen": 14361280, "step": 10600 }, { "epoch": 0.33928685743550346, "eval_loss": 0.506580114364624, "eval_runtime": 499.0992, "eval_samples_per_second": 27.822, "eval_steps_per_second": 13.911, "num_input_tokens_seen": 14361280, "step": 10600 }, { "epoch": 0.33944689840599196, "grad_norm": 1.803606629371643, "learning_rate": 4.181950295850811e-05, "loss": 0.418, "num_input_tokens_seen": 14367760, "step": 10605 }, { "epoch": 0.3396069393764804, "grad_norm": 1.254783034324646, "learning_rate": 4.181223827512918e-05, "loss": 0.4447, "num_input_tokens_seen": 14374416, "step": 10610 }, { "epoch": 0.33976698034696884, "grad_norm": 1.339561104774475, "learning_rate": 4.180497099909183e-05, "loss": 0.4505, "num_input_tokens_seen": 14381184, "step": 10615 }, { "epoch": 0.3399270213174573, "grad_norm": 1.9094033241271973, "learning_rate": 4.179770113151677e-05, "loss": 0.6265, "num_input_tokens_seen": 14387776, "step": 10620 }, { "epoch": 0.3400870622879457, "grad_norm": 1.3686723709106445, "learning_rate": 4.179042867352511e-05, "loss": 0.4167, "num_input_tokens_seen": 14394672, "step": 10625 }, { "epoch": 0.34024710325843416, "grad_norm": 1.599878191947937, "learning_rate": 4.1783153626238334e-05, "loss": 0.4794, "num_input_tokens_seen": 14401264, "step": 10630 }, { "epoch": 0.3404071442289226, "grad_norm": 1.0200510025024414, "learning_rate": 4.177587599077836e-05, "loss": 0.7566, "num_input_tokens_seen": 14408064, "step": 10635 }, { "epoch": 0.34056718519941104, "grad_norm": 1.361063838005066, "learning_rate": 4.1768595768267494e-05, "loss": 0.6268, "num_input_tokens_seen": 14415200, "step": 10640 }, { "epoch": 0.3407272261698995, "grad_norm": 1.0505410432815552, "learning_rate": 4.176131295982843e-05, "loss": 0.3761, "num_input_tokens_seen": 14421600, "step": 10645 }, { "epoch": 0.3408872671403879, "grad_norm": 1.5557825565338135, "learning_rate": 4.1754027566584276e-05, "loss": 0.6265, "num_input_tokens_seen": 14428352, "step": 10650 }, { "epoch": 0.34104730811087636, "grad_norm": 1.268153190612793, "learning_rate": 4.174673958965852e-05, "loss": 0.447, "num_input_tokens_seen": 14435536, "step": 10655 }, { "epoch": 0.34120734908136485, "grad_norm": 2.185276508331299, "learning_rate": 4.173944903017507e-05, "loss": 0.5341, "num_input_tokens_seen": 14442240, "step": 10660 }, { "epoch": 0.3413673900518533, "grad_norm": 4.1220598220825195, "learning_rate": 4.173215588925822e-05, "loss": 0.6461, "num_input_tokens_seen": 14448848, "step": 10665 }, { "epoch": 0.34152743102234173, "grad_norm": 1.267124891281128, "learning_rate": 4.172486016803266e-05, "loss": 0.4318, "num_input_tokens_seen": 14455696, "step": 10670 }, { "epoch": 0.34168747199283017, "grad_norm": 1.5405570268630981, "learning_rate": 4.171756186762349e-05, "loss": 0.6461, "num_input_tokens_seen": 14462448, "step": 10675 }, { "epoch": 0.3418475129633186, "grad_norm": 0.6091380715370178, "learning_rate": 4.171026098915619e-05, "loss": 0.4866, "num_input_tokens_seen": 14469056, "step": 10680 }, { "epoch": 0.34200755393380705, "grad_norm": 1.0471056699752808, "learning_rate": 4.170295753375665e-05, "loss": 0.3735, "num_input_tokens_seen": 14475840, "step": 10685 }, { "epoch": 0.3421675949042955, "grad_norm": 1.4750642776489258, "learning_rate": 4.169565150255117e-05, "loss": 0.3975, "num_input_tokens_seen": 14482912, "step": 10690 }, { "epoch": 0.34232763587478393, "grad_norm": 1.3290529251098633, "learning_rate": 4.16883428966664e-05, "loss": 0.3502, "num_input_tokens_seen": 14489600, "step": 10695 }, { "epoch": 0.34248767684527237, "grad_norm": 0.9327175617218018, "learning_rate": 4.168103171722944e-05, "loss": 0.5398, "num_input_tokens_seen": 14496256, "step": 10700 }, { "epoch": 0.3426477178157608, "grad_norm": 2.6017444133758545, "learning_rate": 4.167371796536777e-05, "loss": 0.4505, "num_input_tokens_seen": 14503344, "step": 10705 }, { "epoch": 0.3428077587862493, "grad_norm": 1.9632556438446045, "learning_rate": 4.166640164220924e-05, "loss": 0.4809, "num_input_tokens_seen": 14510528, "step": 10710 }, { "epoch": 0.34296779975673775, "grad_norm": 0.9727330803871155, "learning_rate": 4.1659082748882144e-05, "loss": 0.4796, "num_input_tokens_seen": 14517248, "step": 10715 }, { "epoch": 0.3431278407272262, "grad_norm": 1.3004717826843262, "learning_rate": 4.1651761286515135e-05, "loss": 0.924, "num_input_tokens_seen": 14523984, "step": 10720 }, { "epoch": 0.3432878816977146, "grad_norm": 0.9196271300315857, "learning_rate": 4.164443725623728e-05, "loss": 0.5202, "num_input_tokens_seen": 14530448, "step": 10725 }, { "epoch": 0.34344792266820307, "grad_norm": 0.8963585495948792, "learning_rate": 4.163711065917802e-05, "loss": 0.4168, "num_input_tokens_seen": 14537280, "step": 10730 }, { "epoch": 0.3436079636386915, "grad_norm": 1.3868473768234253, "learning_rate": 4.1629781496467234e-05, "loss": 0.4478, "num_input_tokens_seen": 14544400, "step": 10735 }, { "epoch": 0.34376800460917994, "grad_norm": 0.8625908493995667, "learning_rate": 4.1622449769235164e-05, "loss": 0.5757, "num_input_tokens_seen": 14551264, "step": 10740 }, { "epoch": 0.3439280455796684, "grad_norm": 1.4137210845947266, "learning_rate": 4.161511547861243e-05, "loss": 0.5416, "num_input_tokens_seen": 14558176, "step": 10745 }, { "epoch": 0.3440880865501568, "grad_norm": 0.5587725639343262, "learning_rate": 4.1607778625730104e-05, "loss": 0.3709, "num_input_tokens_seen": 14564896, "step": 10750 }, { "epoch": 0.34424812752064526, "grad_norm": 0.6897183656692505, "learning_rate": 4.160043921171961e-05, "loss": 0.3354, "num_input_tokens_seen": 14571904, "step": 10755 }, { "epoch": 0.3444081684911337, "grad_norm": 1.3198459148406982, "learning_rate": 4.159309723771276e-05, "loss": 0.3384, "num_input_tokens_seen": 14578640, "step": 10760 }, { "epoch": 0.3445682094616222, "grad_norm": 1.260388731956482, "learning_rate": 4.158575270484181e-05, "loss": 0.5821, "num_input_tokens_seen": 14584944, "step": 10765 }, { "epoch": 0.34472825043211064, "grad_norm": 0.8619983792304993, "learning_rate": 4.157840561423936e-05, "loss": 0.6953, "num_input_tokens_seen": 14591328, "step": 10770 }, { "epoch": 0.3448882914025991, "grad_norm": 0.7821494340896606, "learning_rate": 4.1571055967038416e-05, "loss": 0.5409, "num_input_tokens_seen": 14597744, "step": 10775 }, { "epoch": 0.3450483323730875, "grad_norm": 0.994651198387146, "learning_rate": 4.156370376437241e-05, "loss": 0.3827, "num_input_tokens_seen": 14604368, "step": 10780 }, { "epoch": 0.34520837334357596, "grad_norm": 1.8565911054611206, "learning_rate": 4.155634900737513e-05, "loss": 0.551, "num_input_tokens_seen": 14611040, "step": 10785 }, { "epoch": 0.3453684143140644, "grad_norm": 0.9824589490890503, "learning_rate": 4.1548991697180764e-05, "loss": 0.4278, "num_input_tokens_seen": 14617328, "step": 10790 }, { "epoch": 0.34552845528455284, "grad_norm": 0.7743909955024719, "learning_rate": 4.1541631834923914e-05, "loss": 0.557, "num_input_tokens_seen": 14623888, "step": 10795 }, { "epoch": 0.3456884962550413, "grad_norm": 0.9285649657249451, "learning_rate": 4.153426942173956e-05, "loss": 0.4673, "num_input_tokens_seen": 14631040, "step": 10800 }, { "epoch": 0.3456884962550413, "eval_loss": 0.5039613246917725, "eval_runtime": 499.7223, "eval_samples_per_second": 27.787, "eval_steps_per_second": 13.894, "num_input_tokens_seen": 14631040, "step": 10800 }, { "epoch": 0.3458485372255297, "grad_norm": 1.816778302192688, "learning_rate": 4.152690445876308e-05, "loss": 0.5482, "num_input_tokens_seen": 14638160, "step": 10805 }, { "epoch": 0.34600857819601816, "grad_norm": 1.4833041429519653, "learning_rate": 4.1519536947130245e-05, "loss": 0.4575, "num_input_tokens_seen": 14645136, "step": 10810 }, { "epoch": 0.34616861916650665, "grad_norm": 1.000950574874878, "learning_rate": 4.151216688797722e-05, "loss": 0.4703, "num_input_tokens_seen": 14651728, "step": 10815 }, { "epoch": 0.3463286601369951, "grad_norm": 0.5843793153762817, "learning_rate": 4.150479428244054e-05, "loss": 0.443, "num_input_tokens_seen": 14658192, "step": 10820 }, { "epoch": 0.34648870110748353, "grad_norm": 1.1761219501495361, "learning_rate": 4.1497419131657176e-05, "loss": 0.3442, "num_input_tokens_seen": 14664672, "step": 10825 }, { "epoch": 0.34664874207797197, "grad_norm": 1.6813493967056274, "learning_rate": 4.149004143676447e-05, "loss": 0.4709, "num_input_tokens_seen": 14671280, "step": 10830 }, { "epoch": 0.3468087830484604, "grad_norm": 1.5688389539718628, "learning_rate": 4.148266119890015e-05, "loss": 0.6075, "num_input_tokens_seen": 14678368, "step": 10835 }, { "epoch": 0.34696882401894885, "grad_norm": 2.5942232608795166, "learning_rate": 4.1475278419202324e-05, "loss": 0.5591, "num_input_tokens_seen": 14685216, "step": 10840 }, { "epoch": 0.3471288649894373, "grad_norm": 0.6468730568885803, "learning_rate": 4.146789309880953e-05, "loss": 0.4792, "num_input_tokens_seen": 14691952, "step": 10845 }, { "epoch": 0.34728890595992573, "grad_norm": 1.2713536024093628, "learning_rate": 4.146050523886068e-05, "loss": 0.7174, "num_input_tokens_seen": 14698864, "step": 10850 }, { "epoch": 0.34744894693041417, "grad_norm": 0.8733694553375244, "learning_rate": 4.1453114840495055e-05, "loss": 0.4286, "num_input_tokens_seen": 14705408, "step": 10855 }, { "epoch": 0.3476089879009026, "grad_norm": 0.6345691680908203, "learning_rate": 4.1445721904852364e-05, "loss": 0.3567, "num_input_tokens_seen": 14712304, "step": 10860 }, { "epoch": 0.34776902887139105, "grad_norm": 1.1635444164276123, "learning_rate": 4.143832643307269e-05, "loss": 0.5641, "num_input_tokens_seen": 14718624, "step": 10865 }, { "epoch": 0.34792906984187955, "grad_norm": 1.1272826194763184, "learning_rate": 4.1430928426296503e-05, "loss": 0.3972, "num_input_tokens_seen": 14725920, "step": 10870 }, { "epoch": 0.348089110812368, "grad_norm": 1.6799989938735962, "learning_rate": 4.142352788566466e-05, "loss": 0.6033, "num_input_tokens_seen": 14733008, "step": 10875 }, { "epoch": 0.3482491517828564, "grad_norm": 1.5143465995788574, "learning_rate": 4.1416124812318424e-05, "loss": 0.6364, "num_input_tokens_seen": 14739456, "step": 10880 }, { "epoch": 0.34840919275334487, "grad_norm": 1.7614117860794067, "learning_rate": 4.1408719207399453e-05, "loss": 0.3633, "num_input_tokens_seen": 14746320, "step": 10885 }, { "epoch": 0.3485692337238333, "grad_norm": 0.9197250604629517, "learning_rate": 4.140131107204978e-05, "loss": 0.4786, "num_input_tokens_seen": 14753392, "step": 10890 }, { "epoch": 0.34872927469432174, "grad_norm": 2.6097095012664795, "learning_rate": 4.139390040741182e-05, "loss": 0.5425, "num_input_tokens_seen": 14760064, "step": 10895 }, { "epoch": 0.3488893156648102, "grad_norm": 0.5363460779190063, "learning_rate": 4.1386487214628396e-05, "loss": 0.3095, "num_input_tokens_seen": 14766592, "step": 10900 }, { "epoch": 0.3490493566352986, "grad_norm": 1.5613389015197754, "learning_rate": 4.137907149484272e-05, "loss": 0.578, "num_input_tokens_seen": 14773504, "step": 10905 }, { "epoch": 0.34920939760578706, "grad_norm": 0.7888514399528503, "learning_rate": 4.137165324919839e-05, "loss": 0.3945, "num_input_tokens_seen": 14780080, "step": 10910 }, { "epoch": 0.3493694385762755, "grad_norm": 1.8622318506240845, "learning_rate": 4.136423247883939e-05, "loss": 0.4385, "num_input_tokens_seen": 14787008, "step": 10915 }, { "epoch": 0.349529479546764, "grad_norm": 1.143871545791626, "learning_rate": 4.135680918491009e-05, "loss": 0.3292, "num_input_tokens_seen": 14793824, "step": 10920 }, { "epoch": 0.34968952051725244, "grad_norm": 2.2671000957489014, "learning_rate": 4.1349383368555265e-05, "loss": 0.5254, "num_input_tokens_seen": 14800672, "step": 10925 }, { "epoch": 0.3498495614877409, "grad_norm": 1.2484685182571411, "learning_rate": 4.1341955030920065e-05, "loss": 0.4427, "num_input_tokens_seen": 14807088, "step": 10930 }, { "epoch": 0.3500096024582293, "grad_norm": 0.8098333477973938, "learning_rate": 4.1334524173150036e-05, "loss": 0.452, "num_input_tokens_seen": 14814224, "step": 10935 }, { "epoch": 0.35016964342871776, "grad_norm": 0.6180601716041565, "learning_rate": 4.13270907963911e-05, "loss": 0.4905, "num_input_tokens_seen": 14820896, "step": 10940 }, { "epoch": 0.3503296843992062, "grad_norm": 1.6008198261260986, "learning_rate": 4.131965490178959e-05, "loss": 0.4628, "num_input_tokens_seen": 14827728, "step": 10945 }, { "epoch": 0.35048972536969464, "grad_norm": 1.2560547590255737, "learning_rate": 4.131221649049222e-05, "loss": 0.6127, "num_input_tokens_seen": 14834432, "step": 10950 }, { "epoch": 0.3506497663401831, "grad_norm": 0.6045575737953186, "learning_rate": 4.130477556364606e-05, "loss": 0.441, "num_input_tokens_seen": 14841040, "step": 10955 }, { "epoch": 0.3508098073106715, "grad_norm": 2.934203624725342, "learning_rate": 4.129733212239861e-05, "loss": 0.4645, "num_input_tokens_seen": 14847744, "step": 10960 }, { "epoch": 0.35096984828115996, "grad_norm": 0.7151792645454407, "learning_rate": 4.128988616789774e-05, "loss": 0.5152, "num_input_tokens_seen": 14854768, "step": 10965 }, { "epoch": 0.3511298892516484, "grad_norm": 0.768779456615448, "learning_rate": 4.1282437701291724e-05, "loss": 0.4538, "num_input_tokens_seen": 14861712, "step": 10970 }, { "epoch": 0.3512899302221369, "grad_norm": 1.142235279083252, "learning_rate": 4.1274986723729184e-05, "loss": 0.4185, "num_input_tokens_seen": 14868640, "step": 10975 }, { "epoch": 0.35144997119262533, "grad_norm": 2.121398687362671, "learning_rate": 4.126753323635917e-05, "loss": 0.5053, "num_input_tokens_seen": 14875040, "step": 10980 }, { "epoch": 0.35161001216311377, "grad_norm": 1.7165642976760864, "learning_rate": 4.12600772403311e-05, "loss": 0.6432, "num_input_tokens_seen": 14881728, "step": 10985 }, { "epoch": 0.3517700531336022, "grad_norm": 2.2280051708221436, "learning_rate": 4.125261873679479e-05, "loss": 0.6201, "num_input_tokens_seen": 14888304, "step": 10990 }, { "epoch": 0.35193009410409065, "grad_norm": 4.267910003662109, "learning_rate": 4.124515772690042e-05, "loss": 0.5896, "num_input_tokens_seen": 14895120, "step": 10995 }, { "epoch": 0.3520901350745791, "grad_norm": 0.9202848076820374, "learning_rate": 4.123769421179858e-05, "loss": 0.443, "num_input_tokens_seen": 14901648, "step": 11000 }, { "epoch": 0.3520901350745791, "eval_loss": 0.5028252601623535, "eval_runtime": 499.3046, "eval_samples_per_second": 27.811, "eval_steps_per_second": 13.905, "num_input_tokens_seen": 14901648, "step": 11000 }, { "epoch": 0.35225017604506753, "grad_norm": 1.5529565811157227, "learning_rate": 4.1230228192640236e-05, "loss": 0.4736, "num_input_tokens_seen": 14908368, "step": 11005 }, { "epoch": 0.35241021701555597, "grad_norm": 0.7895974516868591, "learning_rate": 4.122275967057675e-05, "loss": 0.4001, "num_input_tokens_seen": 14915040, "step": 11010 }, { "epoch": 0.3525702579860444, "grad_norm": 1.7748141288757324, "learning_rate": 4.1215288646759846e-05, "loss": 0.4564, "num_input_tokens_seen": 14921872, "step": 11015 }, { "epoch": 0.35273029895653285, "grad_norm": 0.8626769185066223, "learning_rate": 4.120781512234166e-05, "loss": 0.6432, "num_input_tokens_seen": 14928752, "step": 11020 }, { "epoch": 0.3528903399270213, "grad_norm": 1.9801942110061646, "learning_rate": 4.120033909847471e-05, "loss": 0.4656, "num_input_tokens_seen": 14935424, "step": 11025 }, { "epoch": 0.3530503808975098, "grad_norm": 1.4488427639007568, "learning_rate": 4.119286057631187e-05, "loss": 0.6787, "num_input_tokens_seen": 14942528, "step": 11030 }, { "epoch": 0.3532104218679982, "grad_norm": 1.909358263015747, "learning_rate": 4.118537955700646e-05, "loss": 0.5097, "num_input_tokens_seen": 14949312, "step": 11035 }, { "epoch": 0.35337046283848667, "grad_norm": 1.5327423810958862, "learning_rate": 4.11778960417121e-05, "loss": 0.5236, "num_input_tokens_seen": 14955808, "step": 11040 }, { "epoch": 0.3535305038089751, "grad_norm": 1.3806483745574951, "learning_rate": 4.117041003158288e-05, "loss": 0.3807, "num_input_tokens_seen": 14962448, "step": 11045 }, { "epoch": 0.35369054477946354, "grad_norm": 4.2006516456604, "learning_rate": 4.1162921527773215e-05, "loss": 0.6487, "num_input_tokens_seen": 14969040, "step": 11050 }, { "epoch": 0.353850585749952, "grad_norm": 1.1339268684387207, "learning_rate": 4.115543053143794e-05, "loss": 0.6278, "num_input_tokens_seen": 14975856, "step": 11055 }, { "epoch": 0.3540106267204404, "grad_norm": 2.385099411010742, "learning_rate": 4.114793704373226e-05, "loss": 0.4506, "num_input_tokens_seen": 14982496, "step": 11060 }, { "epoch": 0.35417066769092886, "grad_norm": 1.1297718286514282, "learning_rate": 4.114044106581175e-05, "loss": 0.5976, "num_input_tokens_seen": 14989376, "step": 11065 }, { "epoch": 0.3543307086614173, "grad_norm": 1.308853030204773, "learning_rate": 4.11329425988324e-05, "loss": 0.3705, "num_input_tokens_seen": 14995936, "step": 11070 }, { "epoch": 0.35449074963190574, "grad_norm": 1.3750596046447754, "learning_rate": 4.112544164395056e-05, "loss": 0.442, "num_input_tokens_seen": 15002160, "step": 11075 }, { "epoch": 0.35465079060239424, "grad_norm": 0.8018393516540527, "learning_rate": 4.111793820232297e-05, "loss": 0.4749, "num_input_tokens_seen": 15009376, "step": 11080 }, { "epoch": 0.3548108315728827, "grad_norm": 0.5046307444572449, "learning_rate": 4.1110432275106767e-05, "loss": 0.347, "num_input_tokens_seen": 15016032, "step": 11085 }, { "epoch": 0.3549708725433711, "grad_norm": 0.9191176295280457, "learning_rate": 4.110292386345944e-05, "loss": 0.5521, "num_input_tokens_seen": 15022816, "step": 11090 }, { "epoch": 0.35513091351385956, "grad_norm": 0.8376020789146423, "learning_rate": 4.109541296853891e-05, "loss": 0.3475, "num_input_tokens_seen": 15029232, "step": 11095 }, { "epoch": 0.355290954484348, "grad_norm": 1.4961484670639038, "learning_rate": 4.108789959150341e-05, "loss": 0.4361, "num_input_tokens_seen": 15035984, "step": 11100 }, { "epoch": 0.35545099545483644, "grad_norm": 0.9828759431838989, "learning_rate": 4.108038373351163e-05, "loss": 0.3901, "num_input_tokens_seen": 15042608, "step": 11105 }, { "epoch": 0.3556110364253249, "grad_norm": 0.835957407951355, "learning_rate": 4.10728653957226e-05, "loss": 0.4343, "num_input_tokens_seen": 15049472, "step": 11110 }, { "epoch": 0.3557710773958133, "grad_norm": 0.4980391561985016, "learning_rate": 4.106534457929575e-05, "loss": 0.4925, "num_input_tokens_seen": 15056800, "step": 11115 }, { "epoch": 0.35593111836630176, "grad_norm": 0.8027119040489197, "learning_rate": 4.105782128539086e-05, "loss": 0.4133, "num_input_tokens_seen": 15063856, "step": 11120 }, { "epoch": 0.3560911593367902, "grad_norm": 1.820878505706787, "learning_rate": 4.1050295515168144e-05, "loss": 0.4875, "num_input_tokens_seen": 15070336, "step": 11125 }, { "epoch": 0.35625120030727864, "grad_norm": 0.5430300831794739, "learning_rate": 4.1042767269788155e-05, "loss": 0.5606, "num_input_tokens_seen": 15076992, "step": 11130 }, { "epoch": 0.35641124127776713, "grad_norm": 0.8998264074325562, "learning_rate": 4.103523655041185e-05, "loss": 0.5388, "num_input_tokens_seen": 15083408, "step": 11135 }, { "epoch": 0.35657128224825557, "grad_norm": 1.0872142314910889, "learning_rate": 4.102770335820055e-05, "loss": 0.3993, "num_input_tokens_seen": 15090448, "step": 11140 }, { "epoch": 0.356731323218744, "grad_norm": 1.2763464450836182, "learning_rate": 4.1020167694315984e-05, "loss": 0.5403, "num_input_tokens_seen": 15097488, "step": 11145 }, { "epoch": 0.35689136418923245, "grad_norm": 0.7884325981140137, "learning_rate": 4.101262955992023e-05, "loss": 0.5364, "num_input_tokens_seen": 15104272, "step": 11150 }, { "epoch": 0.3570514051597209, "grad_norm": 0.5054988265037537, "learning_rate": 4.100508895617578e-05, "loss": 0.5199, "num_input_tokens_seen": 15110928, "step": 11155 }, { "epoch": 0.35721144613020933, "grad_norm": 1.3052643537521362, "learning_rate": 4.099754588424547e-05, "loss": 0.4313, "num_input_tokens_seen": 15117536, "step": 11160 }, { "epoch": 0.35737148710069777, "grad_norm": 0.7754626870155334, "learning_rate": 4.0990000345292546e-05, "loss": 0.4872, "num_input_tokens_seen": 15124416, "step": 11165 }, { "epoch": 0.3575315280711862, "grad_norm": 1.6019110679626465, "learning_rate": 4.098245234048064e-05, "loss": 0.5197, "num_input_tokens_seen": 15130800, "step": 11170 }, { "epoch": 0.35769156904167465, "grad_norm": 1.610197901725769, "learning_rate": 4.0974901870973726e-05, "loss": 0.4751, "num_input_tokens_seen": 15137408, "step": 11175 }, { "epoch": 0.3578516100121631, "grad_norm": 2.442423105239868, "learning_rate": 4.096734893793619e-05, "loss": 0.5068, "num_input_tokens_seen": 15144240, "step": 11180 }, { "epoch": 0.3580116509826516, "grad_norm": 0.6917253732681274, "learning_rate": 4.095979354253279e-05, "loss": 0.3397, "num_input_tokens_seen": 15150928, "step": 11185 }, { "epoch": 0.35817169195314, "grad_norm": 1.183289647102356, "learning_rate": 4.0952235685928656e-05, "loss": 0.5148, "num_input_tokens_seen": 15157552, "step": 11190 }, { "epoch": 0.35833173292362847, "grad_norm": 1.4474695920944214, "learning_rate": 4.094467536928932e-05, "loss": 0.5189, "num_input_tokens_seen": 15164352, "step": 11195 }, { "epoch": 0.3584917738941169, "grad_norm": 0.8005366325378418, "learning_rate": 4.093711259378067e-05, "loss": 0.4586, "num_input_tokens_seen": 15170800, "step": 11200 }, { "epoch": 0.3584917738941169, "eval_loss": 0.5014805793762207, "eval_runtime": 499.1971, "eval_samples_per_second": 27.817, "eval_steps_per_second": 13.908, "num_input_tokens_seen": 15170800, "step": 11200 }, { "epoch": 0.35865181486460534, "grad_norm": 1.6778837442398071, "learning_rate": 4.092954736056897e-05, "loss": 0.5368, "num_input_tokens_seen": 15177744, "step": 11205 }, { "epoch": 0.3588118558350938, "grad_norm": 1.0022774934768677, "learning_rate": 4.09219796708209e-05, "loss": 0.3363, "num_input_tokens_seen": 15184368, "step": 11210 }, { "epoch": 0.3589718968055822, "grad_norm": 1.3268153667449951, "learning_rate": 4.0914409525703464e-05, "loss": 0.5016, "num_input_tokens_seen": 15191296, "step": 11215 }, { "epoch": 0.35913193777607066, "grad_norm": 1.554551362991333, "learning_rate": 4.090683692638408e-05, "loss": 0.609, "num_input_tokens_seen": 15197776, "step": 11220 }, { "epoch": 0.3592919787465591, "grad_norm": 1.1985417604446411, "learning_rate": 4.089926187403056e-05, "loss": 0.4207, "num_input_tokens_seen": 15204512, "step": 11225 }, { "epoch": 0.35945201971704754, "grad_norm": 0.7179904580116272, "learning_rate": 4.0891684369811044e-05, "loss": 0.4276, "num_input_tokens_seen": 15210768, "step": 11230 }, { "epoch": 0.359612060687536, "grad_norm": 0.6407595276832581, "learning_rate": 4.0884104414894107e-05, "loss": 0.3445, "num_input_tokens_seen": 15217936, "step": 11235 }, { "epoch": 0.3597721016580245, "grad_norm": 0.9772320985794067, "learning_rate": 4.087652201044864e-05, "loss": 0.5884, "num_input_tokens_seen": 15224528, "step": 11240 }, { "epoch": 0.3599321426285129, "grad_norm": 1.1278475522994995, "learning_rate": 4.086893715764397e-05, "loss": 0.5648, "num_input_tokens_seen": 15231440, "step": 11245 }, { "epoch": 0.36009218359900136, "grad_norm": 0.8351871371269226, "learning_rate": 4.086134985764977e-05, "loss": 0.4525, "num_input_tokens_seen": 15238096, "step": 11250 }, { "epoch": 0.3602522245694898, "grad_norm": 1.3419387340545654, "learning_rate": 4.0853760111636085e-05, "loss": 0.5542, "num_input_tokens_seen": 15244640, "step": 11255 }, { "epoch": 0.36041226553997824, "grad_norm": 0.8896730542182922, "learning_rate": 4.084616792077337e-05, "loss": 0.4048, "num_input_tokens_seen": 15251216, "step": 11260 }, { "epoch": 0.3605723065104667, "grad_norm": 1.3549461364746094, "learning_rate": 4.083857328623243e-05, "loss": 0.3574, "num_input_tokens_seen": 15258432, "step": 11265 }, { "epoch": 0.3607323474809551, "grad_norm": 0.803250253200531, "learning_rate": 4.083097620918444e-05, "loss": 0.5598, "num_input_tokens_seen": 15265200, "step": 11270 }, { "epoch": 0.36089238845144356, "grad_norm": 0.8141766786575317, "learning_rate": 4.082337669080097e-05, "loss": 0.5922, "num_input_tokens_seen": 15271680, "step": 11275 }, { "epoch": 0.361052429421932, "grad_norm": 0.8890913724899292, "learning_rate": 4.081577473225398e-05, "loss": 0.3821, "num_input_tokens_seen": 15278384, "step": 11280 }, { "epoch": 0.36121247039242044, "grad_norm": 0.622784435749054, "learning_rate": 4.080817033471577e-05, "loss": 0.4248, "num_input_tokens_seen": 15285248, "step": 11285 }, { "epoch": 0.3613725113629089, "grad_norm": 0.9804825186729431, "learning_rate": 4.080056349935903e-05, "loss": 0.3133, "num_input_tokens_seen": 15292256, "step": 11290 }, { "epoch": 0.3615325523333974, "grad_norm": 0.9105521440505981, "learning_rate": 4.079295422735684e-05, "loss": 0.4801, "num_input_tokens_seen": 15298800, "step": 11295 }, { "epoch": 0.3616925933038858, "grad_norm": 1.2963299751281738, "learning_rate": 4.078534251988264e-05, "loss": 0.6071, "num_input_tokens_seen": 15305920, "step": 11300 }, { "epoch": 0.36185263427437425, "grad_norm": 1.2125080823898315, "learning_rate": 4.077772837811025e-05, "loss": 0.5019, "num_input_tokens_seen": 15312544, "step": 11305 }, { "epoch": 0.3620126752448627, "grad_norm": 0.7684194445610046, "learning_rate": 4.0770111803213874e-05, "loss": 0.3495, "num_input_tokens_seen": 15319344, "step": 11310 }, { "epoch": 0.36217271621535113, "grad_norm": 1.2789503335952759, "learning_rate": 4.076249279636807e-05, "loss": 0.5352, "num_input_tokens_seen": 15326192, "step": 11315 }, { "epoch": 0.36233275718583957, "grad_norm": 0.9402282238006592, "learning_rate": 4.075487135874781e-05, "loss": 0.7706, "num_input_tokens_seen": 15332480, "step": 11320 }, { "epoch": 0.362492798156328, "grad_norm": 1.4170706272125244, "learning_rate": 4.074724749152837e-05, "loss": 0.4411, "num_input_tokens_seen": 15339024, "step": 11325 }, { "epoch": 0.36265283912681645, "grad_norm": 1.483804702758789, "learning_rate": 4.07396211958855e-05, "loss": 0.4681, "num_input_tokens_seen": 15345584, "step": 11330 }, { "epoch": 0.3628128800973049, "grad_norm": 1.4068814516067505, "learning_rate": 4.073199247299523e-05, "loss": 0.4199, "num_input_tokens_seen": 15352176, "step": 11335 }, { "epoch": 0.36297292106779333, "grad_norm": 2.038173198699951, "learning_rate": 4.072436132403403e-05, "loss": 0.3975, "num_input_tokens_seen": 15359008, "step": 11340 }, { "epoch": 0.3631329620382818, "grad_norm": 1.24567449092865, "learning_rate": 4.0716727750178704e-05, "loss": 0.4606, "num_input_tokens_seen": 15366096, "step": 11345 }, { "epoch": 0.36329300300877027, "grad_norm": 0.8500298261642456, "learning_rate": 4.0709091752606455e-05, "loss": 0.3455, "num_input_tokens_seen": 15373312, "step": 11350 }, { "epoch": 0.3634530439792587, "grad_norm": 1.4270206689834595, "learning_rate": 4.070145333249484e-05, "loss": 0.5653, "num_input_tokens_seen": 15379856, "step": 11355 }, { "epoch": 0.36361308494974715, "grad_norm": 1.2811425924301147, "learning_rate": 4.069381249102181e-05, "loss": 0.4195, "num_input_tokens_seen": 15386880, "step": 11360 }, { "epoch": 0.3637731259202356, "grad_norm": 0.8748136162757874, "learning_rate": 4.0686169229365665e-05, "loss": 0.4579, "num_input_tokens_seen": 15393392, "step": 11365 }, { "epoch": 0.363933166890724, "grad_norm": 0.9128889441490173, "learning_rate": 4.067852354870511e-05, "loss": 0.653, "num_input_tokens_seen": 15400176, "step": 11370 }, { "epoch": 0.36409320786121246, "grad_norm": 0.9973955154418945, "learning_rate": 4.067087545021919e-05, "loss": 0.3545, "num_input_tokens_seen": 15406656, "step": 11375 }, { "epoch": 0.3642532488317009, "grad_norm": 0.9467155337333679, "learning_rate": 4.066322493508734e-05, "loss": 0.5042, "num_input_tokens_seen": 15413296, "step": 11380 }, { "epoch": 0.36441328980218934, "grad_norm": 0.8848451972007751, "learning_rate": 4.065557200448937e-05, "loss": 0.3922, "num_input_tokens_seen": 15419888, "step": 11385 }, { "epoch": 0.3645733307726778, "grad_norm": 0.8245273232460022, "learning_rate": 4.064791665960546e-05, "loss": 0.7276, "num_input_tokens_seen": 15427536, "step": 11390 }, { "epoch": 0.3647333717431662, "grad_norm": 0.7726662158966064, "learning_rate": 4.064025890161615e-05, "loss": 0.4647, "num_input_tokens_seen": 15433920, "step": 11395 }, { "epoch": 0.3648934127136547, "grad_norm": 1.805343747138977, "learning_rate": 4.0632598731702373e-05, "loss": 0.6426, "num_input_tokens_seen": 15440592, "step": 11400 }, { "epoch": 0.3648934127136547, "eval_loss": 0.5002552270889282, "eval_runtime": 499.4967, "eval_samples_per_second": 27.8, "eval_steps_per_second": 13.9, "num_input_tokens_seen": 15440592, "step": 11400 }, { "epoch": 0.36505345368414316, "grad_norm": 1.3154895305633545, "learning_rate": 4.0624936151045426e-05, "loss": 0.5349, "num_input_tokens_seen": 15447136, "step": 11405 }, { "epoch": 0.3652134946546316, "grad_norm": 1.8564798831939697, "learning_rate": 4.061727116082696e-05, "loss": 0.4134, "num_input_tokens_seen": 15453952, "step": 11410 }, { "epoch": 0.36537353562512004, "grad_norm": 4.419527053833008, "learning_rate": 4.060960376222903e-05, "loss": 0.613, "num_input_tokens_seen": 15461168, "step": 11415 }, { "epoch": 0.3655335765956085, "grad_norm": 1.550950288772583, "learning_rate": 4.0601933956434034e-05, "loss": 0.4453, "num_input_tokens_seen": 15468464, "step": 11420 }, { "epoch": 0.3656936175660969, "grad_norm": 1.098772406578064, "learning_rate": 4.059426174462476e-05, "loss": 0.4578, "num_input_tokens_seen": 15474896, "step": 11425 }, { "epoch": 0.36585365853658536, "grad_norm": 1.6283856630325317, "learning_rate": 4.058658712798435e-05, "loss": 0.6294, "num_input_tokens_seen": 15481616, "step": 11430 }, { "epoch": 0.3660136995070738, "grad_norm": 0.9326215982437134, "learning_rate": 4.0578910107696336e-05, "loss": 0.3859, "num_input_tokens_seen": 15488624, "step": 11435 }, { "epoch": 0.36617374047756224, "grad_norm": 0.8015322089195251, "learning_rate": 4.05712306849446e-05, "loss": 0.4553, "num_input_tokens_seen": 15494960, "step": 11440 }, { "epoch": 0.3663337814480507, "grad_norm": 2.7774174213409424, "learning_rate": 4.0563548860913415e-05, "loss": 0.5841, "num_input_tokens_seen": 15501568, "step": 11445 }, { "epoch": 0.3664938224185392, "grad_norm": 0.5405027866363525, "learning_rate": 4.0555864636787414e-05, "loss": 0.3106, "num_input_tokens_seen": 15508416, "step": 11450 }, { "epoch": 0.3666538633890276, "grad_norm": 1.6626800298690796, "learning_rate": 4.054817801375159e-05, "loss": 0.4581, "num_input_tokens_seen": 15515424, "step": 11455 }, { "epoch": 0.36681390435951605, "grad_norm": 0.7660883069038391, "learning_rate": 4.054048899299134e-05, "loss": 0.3939, "num_input_tokens_seen": 15522192, "step": 11460 }, { "epoch": 0.3669739453300045, "grad_norm": 1.8248203992843628, "learning_rate": 4.0532797575692385e-05, "loss": 0.5845, "num_input_tokens_seen": 15529008, "step": 11465 }, { "epoch": 0.36713398630049293, "grad_norm": 1.5612927675247192, "learning_rate": 4.052510376304085e-05, "loss": 0.465, "num_input_tokens_seen": 15536112, "step": 11470 }, { "epoch": 0.36729402727098137, "grad_norm": 1.0311906337738037, "learning_rate": 4.051740755622321e-05, "loss": 0.5477, "num_input_tokens_seen": 15542544, "step": 11475 }, { "epoch": 0.3674540682414698, "grad_norm": 3.0460426807403564, "learning_rate": 4.050970895642632e-05, "loss": 0.5769, "num_input_tokens_seen": 15549248, "step": 11480 }, { "epoch": 0.36761410921195825, "grad_norm": 2.31939959526062, "learning_rate": 4.050200796483741e-05, "loss": 0.4617, "num_input_tokens_seen": 15556512, "step": 11485 }, { "epoch": 0.3677741501824467, "grad_norm": 0.7550028562545776, "learning_rate": 4.049430458264405e-05, "loss": 0.6474, "num_input_tokens_seen": 15563296, "step": 11490 }, { "epoch": 0.36793419115293513, "grad_norm": 1.5004864931106567, "learning_rate": 4.048659881103422e-05, "loss": 0.3763, "num_input_tokens_seen": 15569952, "step": 11495 }, { "epoch": 0.36809423212342357, "grad_norm": 1.1577764749526978, "learning_rate": 4.0478890651196235e-05, "loss": 0.5792, "num_input_tokens_seen": 15576656, "step": 11500 }, { "epoch": 0.36825427309391207, "grad_norm": 1.1105594635009766, "learning_rate": 4.047118010431879e-05, "loss": 0.3473, "num_input_tokens_seen": 15583104, "step": 11505 }, { "epoch": 0.3684143140644005, "grad_norm": 1.1361488103866577, "learning_rate": 4.046346717159094e-05, "loss": 0.3787, "num_input_tokens_seen": 15590064, "step": 11510 }, { "epoch": 0.36857435503488895, "grad_norm": 3.0357894897460938, "learning_rate": 4.045575185420214e-05, "loss": 0.8406, "num_input_tokens_seen": 15596816, "step": 11515 }, { "epoch": 0.3687343960053774, "grad_norm": 1.9264060258865356, "learning_rate": 4.0448034153342165e-05, "loss": 0.6829, "num_input_tokens_seen": 15602992, "step": 11520 }, { "epoch": 0.3688944369758658, "grad_norm": 0.6027604341506958, "learning_rate": 4.0440314070201194e-05, "loss": 0.4268, "num_input_tokens_seen": 15609344, "step": 11525 }, { "epoch": 0.36905447794635426, "grad_norm": 1.1515944004058838, "learning_rate": 4.043259160596976e-05, "loss": 0.3562, "num_input_tokens_seen": 15615504, "step": 11530 }, { "epoch": 0.3692145189168427, "grad_norm": 0.8568633794784546, "learning_rate": 4.0424866761838767e-05, "loss": 0.7429, "num_input_tokens_seen": 15621968, "step": 11535 }, { "epoch": 0.36937455988733114, "grad_norm": 0.4297129511833191, "learning_rate": 4.041713953899948e-05, "loss": 0.3484, "num_input_tokens_seen": 15628832, "step": 11540 }, { "epoch": 0.3695346008578196, "grad_norm": 0.9059651494026184, "learning_rate": 4.0409409938643515e-05, "loss": 0.4598, "num_input_tokens_seen": 15635552, "step": 11545 }, { "epoch": 0.369694641828308, "grad_norm": 1.1946102380752563, "learning_rate": 4.0401677961962904e-05, "loss": 0.5467, "num_input_tokens_seen": 15642368, "step": 11550 }, { "epoch": 0.3698546827987965, "grad_norm": 0.778847336769104, "learning_rate": 4.039394361015001e-05, "loss": 0.4506, "num_input_tokens_seen": 15648928, "step": 11555 }, { "epoch": 0.37001472376928496, "grad_norm": 1.7900466918945312, "learning_rate": 4.038620688439755e-05, "loss": 0.4784, "num_input_tokens_seen": 15655632, "step": 11560 }, { "epoch": 0.3701747647397734, "grad_norm": 2.0230016708374023, "learning_rate": 4.037846778589862e-05, "loss": 0.4962, "num_input_tokens_seen": 15662496, "step": 11565 }, { "epoch": 0.37033480571026184, "grad_norm": 0.6292299032211304, "learning_rate": 4.0370726315846715e-05, "loss": 0.4033, "num_input_tokens_seen": 15669248, "step": 11570 }, { "epoch": 0.3704948466807503, "grad_norm": 1.0858404636383057, "learning_rate": 4.036298247543565e-05, "loss": 0.4628, "num_input_tokens_seen": 15675776, "step": 11575 }, { "epoch": 0.3706548876512387, "grad_norm": 0.9768267273902893, "learning_rate": 4.035523626585962e-05, "loss": 0.5745, "num_input_tokens_seen": 15682912, "step": 11580 }, { "epoch": 0.37081492862172716, "grad_norm": 0.8052564263343811, "learning_rate": 4.0347487688313194e-05, "loss": 0.6458, "num_input_tokens_seen": 15689520, "step": 11585 }, { "epoch": 0.3709749695922156, "grad_norm": 2.068467378616333, "learning_rate": 4.0339736743991296e-05, "loss": 0.4277, "num_input_tokens_seen": 15696224, "step": 11590 }, { "epoch": 0.37113501056270404, "grad_norm": 0.9117299318313599, "learning_rate": 4.0331983434089227e-05, "loss": 0.39, "num_input_tokens_seen": 15703376, "step": 11595 }, { "epoch": 0.3712950515331925, "grad_norm": 5.004610061645508, "learning_rate": 4.032422775980264e-05, "loss": 0.5997, "num_input_tokens_seen": 15710608, "step": 11600 }, { "epoch": 0.3712950515331925, "eval_loss": 0.49791860580444336, "eval_runtime": 499.4943, "eval_samples_per_second": 27.8, "eval_steps_per_second": 13.9, "num_input_tokens_seen": 15710608, "step": 11600 }, { "epoch": 0.3714550925036809, "grad_norm": 0.8349342942237854, "learning_rate": 4.031646972232754e-05, "loss": 0.4965, "num_input_tokens_seen": 15717184, "step": 11605 }, { "epoch": 0.3716151334741694, "grad_norm": 0.7856435775756836, "learning_rate": 4.0308709322860344e-05, "loss": 0.4733, "num_input_tokens_seen": 15723936, "step": 11610 }, { "epoch": 0.37177517444465785, "grad_norm": 0.49639251828193665, "learning_rate": 4.0300946562597784e-05, "loss": 0.502, "num_input_tokens_seen": 15730912, "step": 11615 }, { "epoch": 0.3719352154151463, "grad_norm": 0.7474204301834106, "learning_rate": 4.029318144273698e-05, "loss": 0.4026, "num_input_tokens_seen": 15737376, "step": 11620 }, { "epoch": 0.37209525638563473, "grad_norm": 1.4799505472183228, "learning_rate": 4.0285413964475415e-05, "loss": 0.3886, "num_input_tokens_seen": 15744080, "step": 11625 }, { "epoch": 0.37225529735612317, "grad_norm": 1.216090440750122, "learning_rate": 4.0277644129010927e-05, "loss": 0.57, "num_input_tokens_seen": 15750784, "step": 11630 }, { "epoch": 0.3724153383266116, "grad_norm": 0.7647700309753418, "learning_rate": 4.0269871937541724e-05, "loss": 0.4625, "num_input_tokens_seen": 15757520, "step": 11635 }, { "epoch": 0.37257537929710005, "grad_norm": 0.7623739242553711, "learning_rate": 4.026209739126637e-05, "loss": 0.4909, "num_input_tokens_seen": 15764128, "step": 11640 }, { "epoch": 0.3727354202675885, "grad_norm": 1.5581421852111816, "learning_rate": 4.025432049138381e-05, "loss": 0.4655, "num_input_tokens_seen": 15770944, "step": 11645 }, { "epoch": 0.37289546123807693, "grad_norm": 0.8139201402664185, "learning_rate": 4.0246541239093325e-05, "loss": 0.4694, "num_input_tokens_seen": 15777776, "step": 11650 }, { "epoch": 0.37305550220856537, "grad_norm": 1.06538724899292, "learning_rate": 4.023875963559459e-05, "loss": 0.4503, "num_input_tokens_seen": 15784608, "step": 11655 }, { "epoch": 0.3732155431790538, "grad_norm": 1.0887013673782349, "learning_rate": 4.023097568208761e-05, "loss": 0.5415, "num_input_tokens_seen": 15791328, "step": 11660 }, { "epoch": 0.3733755841495423, "grad_norm": 0.6487378478050232, "learning_rate": 4.022318937977277e-05, "loss": 0.4823, "num_input_tokens_seen": 15798224, "step": 11665 }, { "epoch": 0.37353562512003075, "grad_norm": 0.5733945369720459, "learning_rate": 4.021540072985084e-05, "loss": 0.4246, "num_input_tokens_seen": 15804672, "step": 11670 }, { "epoch": 0.3736956660905192, "grad_norm": 0.6273130178451538, "learning_rate": 4.020760973352289e-05, "loss": 0.552, "num_input_tokens_seen": 15811760, "step": 11675 }, { "epoch": 0.3738557070610076, "grad_norm": 1.838482141494751, "learning_rate": 4.019981639199042e-05, "loss": 0.3889, "num_input_tokens_seen": 15818000, "step": 11680 }, { "epoch": 0.37401574803149606, "grad_norm": 1.6964653730392456, "learning_rate": 4.0192020706455245e-05, "loss": 0.5816, "num_input_tokens_seen": 15824784, "step": 11685 }, { "epoch": 0.3741757890019845, "grad_norm": 1.957187533378601, "learning_rate": 4.018422267811956e-05, "loss": 0.6184, "num_input_tokens_seen": 15831136, "step": 11690 }, { "epoch": 0.37433582997247294, "grad_norm": 1.3746083974838257, "learning_rate": 4.017642230818592e-05, "loss": 0.4788, "num_input_tokens_seen": 15838160, "step": 11695 }, { "epoch": 0.3744958709429614, "grad_norm": 1.0554546117782593, "learning_rate": 4.0168619597857246e-05, "loss": 0.3948, "num_input_tokens_seen": 15844992, "step": 11700 }, { "epoch": 0.3746559119134498, "grad_norm": 1.4627346992492676, "learning_rate": 4.016081454833681e-05, "loss": 0.4347, "num_input_tokens_seen": 15851536, "step": 11705 }, { "epoch": 0.37481595288393826, "grad_norm": 0.8973015546798706, "learning_rate": 4.0153007160828245e-05, "loss": 0.5713, "num_input_tokens_seen": 15858432, "step": 11710 }, { "epoch": 0.37497599385442676, "grad_norm": 1.3361269235610962, "learning_rate": 4.0145197436535555e-05, "loss": 0.4541, "num_input_tokens_seen": 15865104, "step": 11715 }, { "epoch": 0.3751360348249152, "grad_norm": 1.2732428312301636, "learning_rate": 4.0137385376663095e-05, "loss": 0.6032, "num_input_tokens_seen": 15871856, "step": 11720 }, { "epoch": 0.37529607579540364, "grad_norm": 0.7262862920761108, "learning_rate": 4.012957098241558e-05, "loss": 0.3911, "num_input_tokens_seen": 15878976, "step": 11725 }, { "epoch": 0.3754561167658921, "grad_norm": 1.1386181116104126, "learning_rate": 4.0121754254998076e-05, "loss": 0.4135, "num_input_tokens_seen": 15885584, "step": 11730 }, { "epoch": 0.3756161577363805, "grad_norm": 0.9386233687400818, "learning_rate": 4.011393519561606e-05, "loss": 0.4357, "num_input_tokens_seen": 15892336, "step": 11735 }, { "epoch": 0.37577619870686896, "grad_norm": 3.071956157684326, "learning_rate": 4.010611380547529e-05, "loss": 0.5807, "num_input_tokens_seen": 15899104, "step": 11740 }, { "epoch": 0.3759362396773574, "grad_norm": 1.7414143085479736, "learning_rate": 4.009829008578192e-05, "loss": 0.3205, "num_input_tokens_seen": 15905760, "step": 11745 }, { "epoch": 0.37609628064784584, "grad_norm": 0.9766247868537903, "learning_rate": 4.00904640377425e-05, "loss": 0.4352, "num_input_tokens_seen": 15912496, "step": 11750 }, { "epoch": 0.3762563216183343, "grad_norm": 0.9409669637680054, "learning_rate": 4.0082635662563886e-05, "loss": 0.5649, "num_input_tokens_seen": 15919568, "step": 11755 }, { "epoch": 0.3764163625888227, "grad_norm": 1.9068489074707031, "learning_rate": 4.007480496145331e-05, "loss": 0.582, "num_input_tokens_seen": 15926720, "step": 11760 }, { "epoch": 0.37657640355931116, "grad_norm": 0.8418974280357361, "learning_rate": 4.006697193561837e-05, "loss": 0.5119, "num_input_tokens_seen": 15933264, "step": 11765 }, { "epoch": 0.37673644452979965, "grad_norm": 1.0560352802276611, "learning_rate": 4.005913658626701e-05, "loss": 0.565, "num_input_tokens_seen": 15939808, "step": 11770 }, { "epoch": 0.3768964855002881, "grad_norm": 1.1095263957977295, "learning_rate": 4.005129891460754e-05, "loss": 0.5282, "num_input_tokens_seen": 15946352, "step": 11775 }, { "epoch": 0.37705652647077653, "grad_norm": 0.5470088720321655, "learning_rate": 4.004345892184864e-05, "loss": 0.2667, "num_input_tokens_seen": 15953312, "step": 11780 }, { "epoch": 0.37721656744126497, "grad_norm": 0.7332589626312256, "learning_rate": 4.003561660919932e-05, "loss": 0.4009, "num_input_tokens_seen": 15960112, "step": 11785 }, { "epoch": 0.3773766084117534, "grad_norm": 1.0698373317718506, "learning_rate": 4.002777197786897e-05, "loss": 0.3878, "num_input_tokens_seen": 15966784, "step": 11790 }, { "epoch": 0.37753664938224185, "grad_norm": 1.1442567110061646, "learning_rate": 4.0019925029067326e-05, "loss": 0.4444, "num_input_tokens_seen": 15973488, "step": 11795 }, { "epoch": 0.3776966903527303, "grad_norm": 0.7286717891693115, "learning_rate": 4.0012075764004495e-05, "loss": 0.3011, "num_input_tokens_seen": 15980176, "step": 11800 }, { "epoch": 0.3776966903527303, "eval_loss": 0.4966561198234558, "eval_runtime": 499.5644, "eval_samples_per_second": 27.796, "eval_steps_per_second": 13.898, "num_input_tokens_seen": 15980176, "step": 11800 }, { "epoch": 0.37785673132321873, "grad_norm": 0.9240331649780273, "learning_rate": 4.000422418389094e-05, "loss": 0.2796, "num_input_tokens_seen": 15986800, "step": 11805 }, { "epoch": 0.37801677229370717, "grad_norm": 1.714596152305603, "learning_rate": 3.999637028993744e-05, "loss": 0.6575, "num_input_tokens_seen": 15993840, "step": 11810 }, { "epoch": 0.3781768132641956, "grad_norm": 1.5203198194503784, "learning_rate": 3.99885140833552e-05, "loss": 0.3962, "num_input_tokens_seen": 16000224, "step": 11815 }, { "epoch": 0.3783368542346841, "grad_norm": 1.0760658979415894, "learning_rate": 3.998065556535572e-05, "loss": 0.5474, "num_input_tokens_seen": 16006752, "step": 11820 }, { "epoch": 0.37849689520517255, "grad_norm": 0.7373936176300049, "learning_rate": 3.9972794737150895e-05, "loss": 0.5155, "num_input_tokens_seen": 16013552, "step": 11825 }, { "epoch": 0.378656936175661, "grad_norm": 1.543158769607544, "learning_rate": 3.996493159995297e-05, "loss": 0.4318, "num_input_tokens_seen": 16020240, "step": 11830 }, { "epoch": 0.3788169771461494, "grad_norm": 1.4298326969146729, "learning_rate": 3.995706615497453e-05, "loss": 0.479, "num_input_tokens_seen": 16026896, "step": 11835 }, { "epoch": 0.37897701811663786, "grad_norm": 1.5980489253997803, "learning_rate": 3.994919840342852e-05, "loss": 0.5355, "num_input_tokens_seen": 16033840, "step": 11840 }, { "epoch": 0.3791370590871263, "grad_norm": 1.0330426692962646, "learning_rate": 3.994132834652825e-05, "loss": 0.407, "num_input_tokens_seen": 16040784, "step": 11845 }, { "epoch": 0.37929710005761474, "grad_norm": 2.455202341079712, "learning_rate": 3.99334559854874e-05, "loss": 0.6668, "num_input_tokens_seen": 16047648, "step": 11850 }, { "epoch": 0.3794571410281032, "grad_norm": 1.9064117670059204, "learning_rate": 3.9925581321519955e-05, "loss": 0.5039, "num_input_tokens_seen": 16054272, "step": 11855 }, { "epoch": 0.3796171819985916, "grad_norm": 1.2616887092590332, "learning_rate": 3.991770435584031e-05, "loss": 0.4666, "num_input_tokens_seen": 16060896, "step": 11860 }, { "epoch": 0.37977722296908006, "grad_norm": 0.7131656408309937, "learning_rate": 3.990982508966319e-05, "loss": 0.3373, "num_input_tokens_seen": 16067536, "step": 11865 }, { "epoch": 0.3799372639395685, "grad_norm": 2.687293291091919, "learning_rate": 3.990194352420367e-05, "loss": 0.5493, "num_input_tokens_seen": 16074288, "step": 11870 }, { "epoch": 0.380097304910057, "grad_norm": 1.173909068107605, "learning_rate": 3.9894059660677184e-05, "loss": 0.5003, "num_input_tokens_seen": 16080640, "step": 11875 }, { "epoch": 0.38025734588054544, "grad_norm": 0.8888762593269348, "learning_rate": 3.9886173500299526e-05, "loss": 0.4794, "num_input_tokens_seen": 16087472, "step": 11880 }, { "epoch": 0.3804173868510339, "grad_norm": 1.2150977849960327, "learning_rate": 3.987828504428685e-05, "loss": 0.3801, "num_input_tokens_seen": 16094048, "step": 11885 }, { "epoch": 0.3805774278215223, "grad_norm": 1.1483957767486572, "learning_rate": 3.987039429385565e-05, "loss": 0.6324, "num_input_tokens_seen": 16100784, "step": 11890 }, { "epoch": 0.38073746879201076, "grad_norm": 2.220524787902832, "learning_rate": 3.986250125022277e-05, "loss": 0.3828, "num_input_tokens_seen": 16107696, "step": 11895 }, { "epoch": 0.3808975097624992, "grad_norm": 2.192621946334839, "learning_rate": 3.985460591460544e-05, "loss": 0.4734, "num_input_tokens_seen": 16113920, "step": 11900 }, { "epoch": 0.38105755073298764, "grad_norm": 1.2085812091827393, "learning_rate": 3.984670828822118e-05, "loss": 0.3874, "num_input_tokens_seen": 16120944, "step": 11905 }, { "epoch": 0.3812175917034761, "grad_norm": 1.0206410884857178, "learning_rate": 3.983880837228794e-05, "loss": 0.4485, "num_input_tokens_seen": 16128176, "step": 11910 }, { "epoch": 0.3813776326739645, "grad_norm": 1.9847652912139893, "learning_rate": 3.983090616802396e-05, "loss": 0.5484, "num_input_tokens_seen": 16135216, "step": 11915 }, { "epoch": 0.38153767364445296, "grad_norm": 0.8532048463821411, "learning_rate": 3.982300167664788e-05, "loss": 0.2762, "num_input_tokens_seen": 16141712, "step": 11920 }, { "epoch": 0.38169771461494145, "grad_norm": 1.5080589056015015, "learning_rate": 3.981509489937868e-05, "loss": 0.4747, "num_input_tokens_seen": 16148208, "step": 11925 }, { "epoch": 0.3818577555854299, "grad_norm": 1.3091744184494019, "learning_rate": 3.9807185837435643e-05, "loss": 0.5125, "num_input_tokens_seen": 16154992, "step": 11930 }, { "epoch": 0.38201779655591833, "grad_norm": 1.167306661605835, "learning_rate": 3.9799274492038484e-05, "loss": 0.4328, "num_input_tokens_seen": 16161328, "step": 11935 }, { "epoch": 0.38217783752640677, "grad_norm": 0.9650282859802246, "learning_rate": 3.979136086440722e-05, "loss": 0.4646, "num_input_tokens_seen": 16168272, "step": 11940 }, { "epoch": 0.3823378784968952, "grad_norm": 1.4778625965118408, "learning_rate": 3.9783444955762226e-05, "loss": 0.457, "num_input_tokens_seen": 16175552, "step": 11945 }, { "epoch": 0.38249791946738365, "grad_norm": 1.2486610412597656, "learning_rate": 3.977552676732424e-05, "loss": 0.3836, "num_input_tokens_seen": 16182496, "step": 11950 }, { "epoch": 0.3826579604378721, "grad_norm": 0.990148663520813, "learning_rate": 3.976760630031435e-05, "loss": 0.2956, "num_input_tokens_seen": 16189040, "step": 11955 }, { "epoch": 0.38281800140836053, "grad_norm": 1.0497684478759766, "learning_rate": 3.975968355595398e-05, "loss": 0.5725, "num_input_tokens_seen": 16195664, "step": 11960 }, { "epoch": 0.38297804237884897, "grad_norm": 1.1223945617675781, "learning_rate": 3.9751758535464935e-05, "loss": 0.3538, "num_input_tokens_seen": 16202192, "step": 11965 }, { "epoch": 0.3831380833493374, "grad_norm": 1.1488978862762451, "learning_rate": 3.9743831240069326e-05, "loss": 0.4512, "num_input_tokens_seen": 16209104, "step": 11970 }, { "epoch": 0.38329812431982585, "grad_norm": 1.3212956190109253, "learning_rate": 3.9735901670989675e-05, "loss": 0.3342, "num_input_tokens_seen": 16215504, "step": 11975 }, { "epoch": 0.38345816529031435, "grad_norm": 1.2407859563827515, "learning_rate": 3.97279698294488e-05, "loss": 0.4529, "num_input_tokens_seen": 16222240, "step": 11980 }, { "epoch": 0.3836182062608028, "grad_norm": 1.8083115816116333, "learning_rate": 3.9720035716669876e-05, "loss": 0.516, "num_input_tokens_seen": 16229040, "step": 11985 }, { "epoch": 0.3837782472312912, "grad_norm": 1.0966668128967285, "learning_rate": 3.9712099333876474e-05, "loss": 0.4062, "num_input_tokens_seen": 16235728, "step": 11990 }, { "epoch": 0.38393828820177966, "grad_norm": 0.9549819827079773, "learning_rate": 3.9704160682292475e-05, "loss": 0.4623, "num_input_tokens_seen": 16242272, "step": 11995 }, { "epoch": 0.3840983291722681, "grad_norm": 1.6221727132797241, "learning_rate": 3.9696219763142106e-05, "loss": 0.676, "num_input_tokens_seen": 16249072, "step": 12000 }, { "epoch": 0.3840983291722681, "eval_loss": 0.4954199492931366, "eval_runtime": 499.2959, "eval_samples_per_second": 27.811, "eval_steps_per_second": 13.906, "num_input_tokens_seen": 16249072, "step": 12000 }, { "epoch": 0.38425837014275654, "grad_norm": 0.8860795497894287, "learning_rate": 3.968827657764997e-05, "loss": 0.5461, "num_input_tokens_seen": 16255728, "step": 12005 }, { "epoch": 0.384418411113245, "grad_norm": 0.9262051582336426, "learning_rate": 3.9680331127041e-05, "loss": 0.4839, "num_input_tokens_seen": 16262432, "step": 12010 }, { "epoch": 0.3845784520837334, "grad_norm": 1.580402135848999, "learning_rate": 3.9672383412540495e-05, "loss": 0.6743, "num_input_tokens_seen": 16268880, "step": 12015 }, { "epoch": 0.38473849305422186, "grad_norm": 0.8950178027153015, "learning_rate": 3.966443343537407e-05, "loss": 0.4369, "num_input_tokens_seen": 16275824, "step": 12020 }, { "epoch": 0.3848985340247103, "grad_norm": 1.4160100221633911, "learning_rate": 3.965648119676772e-05, "loss": 0.3155, "num_input_tokens_seen": 16283024, "step": 12025 }, { "epoch": 0.38505857499519874, "grad_norm": 1.6358193159103394, "learning_rate": 3.96485266979478e-05, "loss": 0.5731, "num_input_tokens_seen": 16289616, "step": 12030 }, { "epoch": 0.38521861596568724, "grad_norm": 1.9567488431930542, "learning_rate": 3.9640569940140974e-05, "loss": 0.3791, "num_input_tokens_seen": 16296144, "step": 12035 }, { "epoch": 0.3853786569361757, "grad_norm": 0.7074329257011414, "learning_rate": 3.963261092457428e-05, "loss": 0.4335, "num_input_tokens_seen": 16303200, "step": 12040 }, { "epoch": 0.3855386979066641, "grad_norm": 3.7683188915252686, "learning_rate": 3.962464965247509e-05, "loss": 0.6192, "num_input_tokens_seen": 16310080, "step": 12045 }, { "epoch": 0.38569873887715256, "grad_norm": 1.2144947052001953, "learning_rate": 3.9616686125071135e-05, "loss": 0.5153, "num_input_tokens_seen": 16316944, "step": 12050 }, { "epoch": 0.385858779847641, "grad_norm": 0.6956453919410706, "learning_rate": 3.9608720343590506e-05, "loss": 0.2074, "num_input_tokens_seen": 16323856, "step": 12055 }, { "epoch": 0.38601882081812944, "grad_norm": 1.4459606409072876, "learning_rate": 3.960075230926161e-05, "loss": 0.4471, "num_input_tokens_seen": 16330336, "step": 12060 }, { "epoch": 0.3861788617886179, "grad_norm": 0.49433279037475586, "learning_rate": 3.959278202331322e-05, "loss": 0.471, "num_input_tokens_seen": 16337280, "step": 12065 }, { "epoch": 0.3863389027591063, "grad_norm": 0.7183104753494263, "learning_rate": 3.958480948697446e-05, "loss": 0.4879, "num_input_tokens_seen": 16344464, "step": 12070 }, { "epoch": 0.38649894372959476, "grad_norm": 0.7969166040420532, "learning_rate": 3.95768347014748e-05, "loss": 0.5207, "num_input_tokens_seen": 16350848, "step": 12075 }, { "epoch": 0.3866589847000832, "grad_norm": 1.5091416835784912, "learning_rate": 3.956885766804404e-05, "loss": 0.523, "num_input_tokens_seen": 16357600, "step": 12080 }, { "epoch": 0.3868190256705717, "grad_norm": 2.164944887161255, "learning_rate": 3.956087838791235e-05, "loss": 0.5126, "num_input_tokens_seen": 16364192, "step": 12085 }, { "epoch": 0.38697906664106013, "grad_norm": 1.459604263305664, "learning_rate": 3.955289686231022e-05, "loss": 0.4111, "num_input_tokens_seen": 16370944, "step": 12090 }, { "epoch": 0.38713910761154857, "grad_norm": 0.8548316955566406, "learning_rate": 3.9544913092468504e-05, "loss": 0.4848, "num_input_tokens_seen": 16377280, "step": 12095 }, { "epoch": 0.387299148582037, "grad_norm": 2.080235719680786, "learning_rate": 3.9536927079618425e-05, "loss": 0.3393, "num_input_tokens_seen": 16383680, "step": 12100 }, { "epoch": 0.38745918955252545, "grad_norm": 3.16471004486084, "learning_rate": 3.9528938824991494e-05, "loss": 0.7004, "num_input_tokens_seen": 16390464, "step": 12105 }, { "epoch": 0.3876192305230139, "grad_norm": 1.9220075607299805, "learning_rate": 3.952094832981962e-05, "loss": 0.4134, "num_input_tokens_seen": 16397120, "step": 12110 }, { "epoch": 0.38777927149350233, "grad_norm": 1.9563905000686646, "learning_rate": 3.951295559533503e-05, "loss": 0.463, "num_input_tokens_seen": 16403600, "step": 12115 }, { "epoch": 0.38793931246399077, "grad_norm": 1.169453740119934, "learning_rate": 3.95049606227703e-05, "loss": 0.4886, "num_input_tokens_seen": 16410144, "step": 12120 }, { "epoch": 0.3880993534344792, "grad_norm": 1.0109748840332031, "learning_rate": 3.949696341335838e-05, "loss": 0.4612, "num_input_tokens_seen": 16417184, "step": 12125 }, { "epoch": 0.38825939440496765, "grad_norm": 1.9129148721694946, "learning_rate": 3.9488963968332503e-05, "loss": 0.4972, "num_input_tokens_seen": 16423520, "step": 12130 }, { "epoch": 0.3884194353754561, "grad_norm": 0.9784203767776489, "learning_rate": 3.948096228892631e-05, "loss": 0.5437, "num_input_tokens_seen": 16430416, "step": 12135 }, { "epoch": 0.3885794763459446, "grad_norm": 2.8666532039642334, "learning_rate": 3.947295837637375e-05, "loss": 0.5158, "num_input_tokens_seen": 16437040, "step": 12140 }, { "epoch": 0.388739517316433, "grad_norm": 3.319727659225464, "learning_rate": 3.9464952231909135e-05, "loss": 0.6595, "num_input_tokens_seen": 16443648, "step": 12145 }, { "epoch": 0.38889955828692147, "grad_norm": 1.064665675163269, "learning_rate": 3.945694385676711e-05, "loss": 0.6088, "num_input_tokens_seen": 16450624, "step": 12150 }, { "epoch": 0.3890595992574099, "grad_norm": 1.2467575073242188, "learning_rate": 3.944893325218265e-05, "loss": 0.5488, "num_input_tokens_seen": 16457968, "step": 12155 }, { "epoch": 0.38921964022789834, "grad_norm": 1.5814895629882812, "learning_rate": 3.944092041939112e-05, "loss": 0.5732, "num_input_tokens_seen": 16465120, "step": 12160 }, { "epoch": 0.3893796811983868, "grad_norm": 0.8629751205444336, "learning_rate": 3.943290535962818e-05, "loss": 0.5912, "num_input_tokens_seen": 16471840, "step": 12165 }, { "epoch": 0.3895397221688752, "grad_norm": 1.7958718538284302, "learning_rate": 3.942488807412985e-05, "loss": 0.6118, "num_input_tokens_seen": 16478480, "step": 12170 }, { "epoch": 0.38969976313936366, "grad_norm": 1.0110682249069214, "learning_rate": 3.941686856413251e-05, "loss": 0.3701, "num_input_tokens_seen": 16485280, "step": 12175 }, { "epoch": 0.3898598041098521, "grad_norm": 0.36539652943611145, "learning_rate": 3.9408846830872874e-05, "loss": 0.3807, "num_input_tokens_seen": 16491968, "step": 12180 }, { "epoch": 0.39001984508034054, "grad_norm": 0.862739622592926, "learning_rate": 3.940082287558798e-05, "loss": 0.3826, "num_input_tokens_seen": 16498848, "step": 12185 }, { "epoch": 0.39017988605082904, "grad_norm": 2.9061155319213867, "learning_rate": 3.939279669951522e-05, "loss": 0.8514, "num_input_tokens_seen": 16509136, "step": 12190 }, { "epoch": 0.3903399270213175, "grad_norm": 1.8322062492370605, "learning_rate": 3.938476830389234e-05, "loss": 0.4537, "num_input_tokens_seen": 16515952, "step": 12195 }, { "epoch": 0.3904999679918059, "grad_norm": 1.141493797302246, "learning_rate": 3.937673768995742e-05, "loss": 0.3801, "num_input_tokens_seen": 16522704, "step": 12200 }, { "epoch": 0.3904999679918059, "eval_loss": 0.49414393305778503, "eval_runtime": 499.3643, "eval_samples_per_second": 27.807, "eval_steps_per_second": 13.904, "num_input_tokens_seen": 16522704, "step": 12200 }, { "epoch": 0.39066000896229436, "grad_norm": 1.9492233991622925, "learning_rate": 3.936870485894888e-05, "loss": 0.4962, "num_input_tokens_seen": 16529184, "step": 12205 }, { "epoch": 0.3908200499327828, "grad_norm": 0.9528722763061523, "learning_rate": 3.9360669812105475e-05, "loss": 0.4943, "num_input_tokens_seen": 16535920, "step": 12210 }, { "epoch": 0.39098009090327124, "grad_norm": 0.779717743396759, "learning_rate": 3.9352632550666325e-05, "loss": 0.5197, "num_input_tokens_seen": 16542736, "step": 12215 }, { "epoch": 0.3911401318737597, "grad_norm": 1.0160809755325317, "learning_rate": 3.9344593075870866e-05, "loss": 0.3429, "num_input_tokens_seen": 16549552, "step": 12220 }, { "epoch": 0.3913001728442481, "grad_norm": 1.06925368309021, "learning_rate": 3.933655138895889e-05, "loss": 0.4544, "num_input_tokens_seen": 16556400, "step": 12225 }, { "epoch": 0.39146021381473656, "grad_norm": 0.6929334998130798, "learning_rate": 3.932850749117053e-05, "loss": 0.6122, "num_input_tokens_seen": 16563264, "step": 12230 }, { "epoch": 0.391620254785225, "grad_norm": 1.0197222232818604, "learning_rate": 3.932046138374624e-05, "loss": 0.5161, "num_input_tokens_seen": 16569952, "step": 12235 }, { "epoch": 0.39178029575571344, "grad_norm": 0.36828282475471497, "learning_rate": 3.9312413067926854e-05, "loss": 0.3939, "num_input_tokens_seen": 16576480, "step": 12240 }, { "epoch": 0.39194033672620193, "grad_norm": 2.2814652919769287, "learning_rate": 3.9304362544953506e-05, "loss": 0.6236, "num_input_tokens_seen": 16583248, "step": 12245 }, { "epoch": 0.39210037769669037, "grad_norm": 0.9984094500541687, "learning_rate": 3.929630981606769e-05, "loss": 0.5789, "num_input_tokens_seen": 16589824, "step": 12250 }, { "epoch": 0.3922604186671788, "grad_norm": 1.152068018913269, "learning_rate": 3.928825488251124e-05, "loss": 0.4374, "num_input_tokens_seen": 16596800, "step": 12255 }, { "epoch": 0.39242045963766725, "grad_norm": 1.4343278408050537, "learning_rate": 3.9280197745526344e-05, "loss": 0.378, "num_input_tokens_seen": 16603520, "step": 12260 }, { "epoch": 0.3925805006081557, "grad_norm": 1.1242254972457886, "learning_rate": 3.9272138406355495e-05, "loss": 0.5094, "num_input_tokens_seen": 16610336, "step": 12265 }, { "epoch": 0.39274054157864413, "grad_norm": 1.0600792169570923, "learning_rate": 3.926407686624154e-05, "loss": 0.7011, "num_input_tokens_seen": 16617136, "step": 12270 }, { "epoch": 0.39290058254913257, "grad_norm": 1.027989387512207, "learning_rate": 3.9256013126427684e-05, "loss": 0.5014, "num_input_tokens_seen": 16623520, "step": 12275 }, { "epoch": 0.393060623519621, "grad_norm": 0.9437049031257629, "learning_rate": 3.9247947188157455e-05, "loss": 0.3536, "num_input_tokens_seen": 16630832, "step": 12280 }, { "epoch": 0.39322066449010945, "grad_norm": 2.170215606689453, "learning_rate": 3.9239879052674715e-05, "loss": 0.6582, "num_input_tokens_seen": 16637728, "step": 12285 }, { "epoch": 0.3933807054605979, "grad_norm": 1.3477551937103271, "learning_rate": 3.9231808721223673e-05, "loss": 0.5516, "num_input_tokens_seen": 16644608, "step": 12290 }, { "epoch": 0.3935407464310864, "grad_norm": 1.0114003419876099, "learning_rate": 3.9223736195048886e-05, "loss": 0.3797, "num_input_tokens_seen": 16651536, "step": 12295 }, { "epoch": 0.3937007874015748, "grad_norm": 1.423549771308899, "learning_rate": 3.921566147539523e-05, "loss": 0.3695, "num_input_tokens_seen": 16658224, "step": 12300 }, { "epoch": 0.39386082837206327, "grad_norm": 0.4400905668735504, "learning_rate": 3.920758456350792e-05, "loss": 0.3864, "num_input_tokens_seen": 16665056, "step": 12305 }, { "epoch": 0.3940208693425517, "grad_norm": 2.123849630355835, "learning_rate": 3.919950546063253e-05, "loss": 0.6635, "num_input_tokens_seen": 16671984, "step": 12310 }, { "epoch": 0.39418091031304014, "grad_norm": 0.8671271800994873, "learning_rate": 3.919142416801496e-05, "loss": 0.4042, "num_input_tokens_seen": 16678624, "step": 12315 }, { "epoch": 0.3943409512835286, "grad_norm": 1.4973191022872925, "learning_rate": 3.918334068690144e-05, "loss": 0.5393, "num_input_tokens_seen": 16685712, "step": 12320 }, { "epoch": 0.394500992254017, "grad_norm": 1.6401687860488892, "learning_rate": 3.917525501853855e-05, "loss": 0.4555, "num_input_tokens_seen": 16692464, "step": 12325 }, { "epoch": 0.39466103322450546, "grad_norm": 1.2333238124847412, "learning_rate": 3.916716716417319e-05, "loss": 0.4482, "num_input_tokens_seen": 16698992, "step": 12330 }, { "epoch": 0.3948210741949939, "grad_norm": 0.909326434135437, "learning_rate": 3.915907712505263e-05, "loss": 0.5026, "num_input_tokens_seen": 16705456, "step": 12335 }, { "epoch": 0.39498111516548234, "grad_norm": 0.9906455874443054, "learning_rate": 3.915098490242444e-05, "loss": 0.4858, "num_input_tokens_seen": 16711936, "step": 12340 }, { "epoch": 0.3951411561359708, "grad_norm": 1.5221924781799316, "learning_rate": 3.914289049753654e-05, "loss": 0.5814, "num_input_tokens_seen": 16718704, "step": 12345 }, { "epoch": 0.3953011971064593, "grad_norm": 0.8214284181594849, "learning_rate": 3.913479391163719e-05, "loss": 0.4103, "num_input_tokens_seen": 16725952, "step": 12350 }, { "epoch": 0.3954612380769477, "grad_norm": 0.8111627697944641, "learning_rate": 3.9126695145975e-05, "loss": 0.402, "num_input_tokens_seen": 16732960, "step": 12355 }, { "epoch": 0.39562127904743616, "grad_norm": 1.0625360012054443, "learning_rate": 3.911859420179889e-05, "loss": 0.6901, "num_input_tokens_seen": 16739472, "step": 12360 }, { "epoch": 0.3957813200179246, "grad_norm": 0.5441197752952576, "learning_rate": 3.911049108035813e-05, "loss": 0.3323, "num_input_tokens_seen": 16745984, "step": 12365 }, { "epoch": 0.39594136098841304, "grad_norm": 0.7806872725486755, "learning_rate": 3.910238578290232e-05, "loss": 0.3756, "num_input_tokens_seen": 16752448, "step": 12370 }, { "epoch": 0.3961014019589015, "grad_norm": 1.2112842798233032, "learning_rate": 3.90942783106814e-05, "loss": 0.3987, "num_input_tokens_seen": 16759216, "step": 12375 }, { "epoch": 0.3962614429293899, "grad_norm": 1.1144860982894897, "learning_rate": 3.908616866494564e-05, "loss": 0.5454, "num_input_tokens_seen": 16766160, "step": 12380 }, { "epoch": 0.39642148389987836, "grad_norm": 1.5429900884628296, "learning_rate": 3.907805684694566e-05, "loss": 0.6741, "num_input_tokens_seen": 16773520, "step": 12385 }, { "epoch": 0.3965815248703668, "grad_norm": 1.7234848737716675, "learning_rate": 3.90699428579324e-05, "loss": 0.5559, "num_input_tokens_seen": 16780480, "step": 12390 }, { "epoch": 0.39674156584085524, "grad_norm": 0.9100705981254578, "learning_rate": 3.906182669915713e-05, "loss": 0.6751, "num_input_tokens_seen": 16787568, "step": 12395 }, { "epoch": 0.3969016068113437, "grad_norm": 2.0288968086242676, "learning_rate": 3.9053708371871476e-05, "loss": 0.7394, "num_input_tokens_seen": 16794064, "step": 12400 }, { "epoch": 0.3969016068113437, "eval_loss": 0.49245819449424744, "eval_runtime": 499.2784, "eval_samples_per_second": 27.812, "eval_steps_per_second": 13.906, "num_input_tokens_seen": 16794064, "step": 12400 }, { "epoch": 0.39706164778183217, "grad_norm": 0.7770348191261292, "learning_rate": 3.904558787732738e-05, "loss": 0.4187, "num_input_tokens_seen": 16801264, "step": 12405 }, { "epoch": 0.3972216887523206, "grad_norm": 1.2351347208023071, "learning_rate": 3.9037465216777135e-05, "loss": 0.3855, "num_input_tokens_seen": 16807936, "step": 12410 }, { "epoch": 0.39738172972280905, "grad_norm": 0.934526264667511, "learning_rate": 3.902934039147334e-05, "loss": 0.5178, "num_input_tokens_seen": 16814384, "step": 12415 }, { "epoch": 0.3975417706932975, "grad_norm": 0.9298099279403687, "learning_rate": 3.902121340266894e-05, "loss": 0.3731, "num_input_tokens_seen": 16821088, "step": 12420 }, { "epoch": 0.39770181166378593, "grad_norm": 0.8087689876556396, "learning_rate": 3.9013084251617246e-05, "loss": 0.4444, "num_input_tokens_seen": 16827552, "step": 12425 }, { "epoch": 0.39786185263427437, "grad_norm": 1.1575212478637695, "learning_rate": 3.9004952939571865e-05, "loss": 0.4495, "num_input_tokens_seen": 16834304, "step": 12430 }, { "epoch": 0.3980218936047628, "grad_norm": 2.291872262954712, "learning_rate": 3.899681946778673e-05, "loss": 0.4987, "num_input_tokens_seen": 16841008, "step": 12435 }, { "epoch": 0.39818193457525125, "grad_norm": 0.8366824984550476, "learning_rate": 3.898868383751615e-05, "loss": 0.377, "num_input_tokens_seen": 16847872, "step": 12440 }, { "epoch": 0.3983419755457397, "grad_norm": 1.534074068069458, "learning_rate": 3.8980546050014724e-05, "loss": 0.5335, "num_input_tokens_seen": 16854960, "step": 12445 }, { "epoch": 0.39850201651622813, "grad_norm": 0.829552173614502, "learning_rate": 3.897240610653741e-05, "loss": 0.5792, "num_input_tokens_seen": 16861152, "step": 12450 }, { "epoch": 0.3986620574867166, "grad_norm": 0.9508932828903198, "learning_rate": 3.896426400833948e-05, "loss": 0.2804, "num_input_tokens_seen": 16867648, "step": 12455 }, { "epoch": 0.39882209845720507, "grad_norm": 1.861354947090149, "learning_rate": 3.895611975667656e-05, "loss": 0.5964, "num_input_tokens_seen": 16874496, "step": 12460 }, { "epoch": 0.3989821394276935, "grad_norm": 1.9429627656936646, "learning_rate": 3.8947973352804584e-05, "loss": 0.5381, "num_input_tokens_seen": 16881632, "step": 12465 }, { "epoch": 0.39914218039818194, "grad_norm": 0.7233342528343201, "learning_rate": 3.893982479797984e-05, "loss": 0.4604, "num_input_tokens_seen": 16888656, "step": 12470 }, { "epoch": 0.3993022213686704, "grad_norm": 0.8550416827201843, "learning_rate": 3.8931674093458926e-05, "loss": 0.3302, "num_input_tokens_seen": 16895248, "step": 12475 }, { "epoch": 0.3994622623391588, "grad_norm": 1.1210764646530151, "learning_rate": 3.89235212404988e-05, "loss": 0.3727, "num_input_tokens_seen": 16901744, "step": 12480 }, { "epoch": 0.39962230330964726, "grad_norm": 1.0930149555206299, "learning_rate": 3.891536624035672e-05, "loss": 0.3465, "num_input_tokens_seen": 16908640, "step": 12485 }, { "epoch": 0.3997823442801357, "grad_norm": 0.7611419558525085, "learning_rate": 3.8907209094290295e-05, "loss": 0.4313, "num_input_tokens_seen": 16916208, "step": 12490 }, { "epoch": 0.39994238525062414, "grad_norm": 1.2703920602798462, "learning_rate": 3.8899049803557466e-05, "loss": 0.4707, "num_input_tokens_seen": 16923136, "step": 12495 }, { "epoch": 0.4001024262211126, "grad_norm": 0.8650281429290771, "learning_rate": 3.889088836941648e-05, "loss": 0.41, "num_input_tokens_seen": 16929552, "step": 12500 }, { "epoch": 0.400262467191601, "grad_norm": 1.2626574039459229, "learning_rate": 3.8882724793125946e-05, "loss": 0.2957, "num_input_tokens_seen": 16936240, "step": 12505 }, { "epoch": 0.4004225081620895, "grad_norm": 1.4133200645446777, "learning_rate": 3.8874559075944794e-05, "loss": 0.4741, "num_input_tokens_seen": 16942928, "step": 12510 }, { "epoch": 0.40058254913257796, "grad_norm": 2.1422319412231445, "learning_rate": 3.886639121913227e-05, "loss": 0.4995, "num_input_tokens_seen": 16949104, "step": 12515 }, { "epoch": 0.4007425901030664, "grad_norm": 1.7684494256973267, "learning_rate": 3.885822122394797e-05, "loss": 0.5797, "num_input_tokens_seen": 16955936, "step": 12520 }, { "epoch": 0.40090263107355484, "grad_norm": 0.7588819861412048, "learning_rate": 3.8850049091651794e-05, "loss": 0.4784, "num_input_tokens_seen": 16962848, "step": 12525 }, { "epoch": 0.4010626720440433, "grad_norm": 1.8123620748519897, "learning_rate": 3.8841874823504e-05, "loss": 0.3933, "num_input_tokens_seen": 16969440, "step": 12530 }, { "epoch": 0.4012227130145317, "grad_norm": 0.7609497904777527, "learning_rate": 3.8833698420765157e-05, "loss": 0.3562, "num_input_tokens_seen": 16976080, "step": 12535 }, { "epoch": 0.40138275398502016, "grad_norm": 1.7243574857711792, "learning_rate": 3.882551988469618e-05, "loss": 0.5668, "num_input_tokens_seen": 16982528, "step": 12540 }, { "epoch": 0.4015427949555086, "grad_norm": 0.8111590147018433, "learning_rate": 3.881733921655829e-05, "loss": 0.384, "num_input_tokens_seen": 16988864, "step": 12545 }, { "epoch": 0.40170283592599704, "grad_norm": 0.844003438949585, "learning_rate": 3.8809156417613054e-05, "loss": 0.4454, "num_input_tokens_seen": 16995952, "step": 12550 }, { "epoch": 0.4018628768964855, "grad_norm": 0.7457853555679321, "learning_rate": 3.8800971489122364e-05, "loss": 0.4169, "num_input_tokens_seen": 17002816, "step": 12555 }, { "epoch": 0.402022917866974, "grad_norm": 0.5478670001029968, "learning_rate": 3.8792784432348434e-05, "loss": 0.2746, "num_input_tokens_seen": 17009312, "step": 12560 }, { "epoch": 0.4021829588374624, "grad_norm": 1.4546078443527222, "learning_rate": 3.878459524855381e-05, "loss": 0.3987, "num_input_tokens_seen": 17016000, "step": 12565 }, { "epoch": 0.40234299980795085, "grad_norm": 1.2123771905899048, "learning_rate": 3.8776403939001384e-05, "loss": 0.3798, "num_input_tokens_seen": 17022736, "step": 12570 }, { "epoch": 0.4025030407784393, "grad_norm": 1.5222357511520386, "learning_rate": 3.876821050495433e-05, "loss": 0.6153, "num_input_tokens_seen": 17029104, "step": 12575 }, { "epoch": 0.40266308174892773, "grad_norm": 0.5009467005729675, "learning_rate": 3.87600149476762e-05, "loss": 0.3284, "num_input_tokens_seen": 17035696, "step": 12580 }, { "epoch": 0.40282312271941617, "grad_norm": 1.276177167892456, "learning_rate": 3.8751817268430843e-05, "loss": 0.2973, "num_input_tokens_seen": 17042272, "step": 12585 }, { "epoch": 0.4029831636899046, "grad_norm": 0.6938021779060364, "learning_rate": 3.8743617468482464e-05, "loss": 0.4525, "num_input_tokens_seen": 17048848, "step": 12590 }, { "epoch": 0.40314320466039305, "grad_norm": 1.0311808586120605, "learning_rate": 3.8735415549095535e-05, "loss": 0.3115, "num_input_tokens_seen": 17055744, "step": 12595 }, { "epoch": 0.4033032456308815, "grad_norm": 0.9665417075157166, "learning_rate": 3.8727211511534934e-05, "loss": 0.4433, "num_input_tokens_seen": 17062288, "step": 12600 }, { "epoch": 0.4033032456308815, "eval_loss": 0.4915654957294464, "eval_runtime": 499.1444, "eval_samples_per_second": 27.82, "eval_steps_per_second": 13.91, "num_input_tokens_seen": 17062288, "step": 12600 }, { "epoch": 0.40346328660136993, "grad_norm": 0.9853880405426025, "learning_rate": 3.8719005357065804e-05, "loss": 0.512, "num_input_tokens_seen": 17068912, "step": 12605 }, { "epoch": 0.40362332757185837, "grad_norm": 0.9677982330322266, "learning_rate": 3.8710797086953645e-05, "loss": 0.5117, "num_input_tokens_seen": 17075600, "step": 12610 }, { "epoch": 0.40378336854234687, "grad_norm": 1.3169025182724, "learning_rate": 3.870258670246427e-05, "loss": 0.3873, "num_input_tokens_seen": 17082128, "step": 12615 }, { "epoch": 0.4039434095128353, "grad_norm": 0.6890947222709656, "learning_rate": 3.869437420486384e-05, "loss": 0.4757, "num_input_tokens_seen": 17088640, "step": 12620 }, { "epoch": 0.40410345048332375, "grad_norm": 1.6340458393096924, "learning_rate": 3.8686159595418805e-05, "loss": 0.3909, "num_input_tokens_seen": 17095648, "step": 12625 }, { "epoch": 0.4042634914538122, "grad_norm": 1.7872098684310913, "learning_rate": 3.867794287539597e-05, "loss": 0.4557, "num_input_tokens_seen": 17102768, "step": 12630 }, { "epoch": 0.4044235324243006, "grad_norm": 1.3784291744232178, "learning_rate": 3.866972404606245e-05, "loss": 0.4837, "num_input_tokens_seen": 17109584, "step": 12635 }, { "epoch": 0.40458357339478906, "grad_norm": 0.948330819606781, "learning_rate": 3.866150310868571e-05, "loss": 0.405, "num_input_tokens_seen": 17115968, "step": 12640 }, { "epoch": 0.4047436143652775, "grad_norm": 1.5563360452651978, "learning_rate": 3.8653280064533506e-05, "loss": 0.4131, "num_input_tokens_seen": 17122672, "step": 12645 }, { "epoch": 0.40490365533576594, "grad_norm": 0.9172854423522949, "learning_rate": 3.864505491487394e-05, "loss": 0.3832, "num_input_tokens_seen": 17129440, "step": 12650 }, { "epoch": 0.4050636963062544, "grad_norm": 0.5972158908843994, "learning_rate": 3.8636827660975414e-05, "loss": 0.4086, "num_input_tokens_seen": 17136064, "step": 12655 }, { "epoch": 0.4052237372767428, "grad_norm": 1.778615951538086, "learning_rate": 3.862859830410671e-05, "loss": 0.3884, "num_input_tokens_seen": 17142544, "step": 12660 }, { "epoch": 0.4053837782472313, "grad_norm": 0.624601423740387, "learning_rate": 3.862036684553688e-05, "loss": 0.4327, "num_input_tokens_seen": 17149120, "step": 12665 }, { "epoch": 0.40554381921771976, "grad_norm": 1.099181056022644, "learning_rate": 3.8612133286535314e-05, "loss": 0.4563, "num_input_tokens_seen": 17155680, "step": 12670 }, { "epoch": 0.4057038601882082, "grad_norm": 3.216230630874634, "learning_rate": 3.860389762837173e-05, "loss": 0.743, "num_input_tokens_seen": 17162432, "step": 12675 }, { "epoch": 0.40586390115869664, "grad_norm": 0.8137784004211426, "learning_rate": 3.859565987231618e-05, "loss": 0.3982, "num_input_tokens_seen": 17169376, "step": 12680 }, { "epoch": 0.4060239421291851, "grad_norm": 0.7532979249954224, "learning_rate": 3.858742001963902e-05, "loss": 0.4308, "num_input_tokens_seen": 17176096, "step": 12685 }, { "epoch": 0.4061839830996735, "grad_norm": 1.7739810943603516, "learning_rate": 3.857917807161094e-05, "loss": 0.4097, "num_input_tokens_seen": 17182480, "step": 12690 }, { "epoch": 0.40634402407016196, "grad_norm": 2.46500301361084, "learning_rate": 3.857093402950296e-05, "loss": 0.4333, "num_input_tokens_seen": 17189440, "step": 12695 }, { "epoch": 0.4065040650406504, "grad_norm": 0.7688290476799011, "learning_rate": 3.8562687894586414e-05, "loss": 0.3854, "num_input_tokens_seen": 17196640, "step": 12700 }, { "epoch": 0.40666410601113884, "grad_norm": 0.8830394148826599, "learning_rate": 3.8554439668132946e-05, "loss": 0.3166, "num_input_tokens_seen": 17203440, "step": 12705 }, { "epoch": 0.4068241469816273, "grad_norm": 0.7157168388366699, "learning_rate": 3.854618935141455e-05, "loss": 0.6015, "num_input_tokens_seen": 17210032, "step": 12710 }, { "epoch": 0.4069841879521157, "grad_norm": 2.52504301071167, "learning_rate": 3.8537936945703525e-05, "loss": 0.5876, "num_input_tokens_seen": 17216880, "step": 12715 }, { "epoch": 0.4071442289226042, "grad_norm": 1.448081612586975, "learning_rate": 3.852968245227249e-05, "loss": 0.7596, "num_input_tokens_seen": 17223440, "step": 12720 }, { "epoch": 0.40730426989309265, "grad_norm": 1.71774160861969, "learning_rate": 3.85214258723944e-05, "loss": 0.4459, "num_input_tokens_seen": 17230432, "step": 12725 }, { "epoch": 0.4074643108635811, "grad_norm": 0.9752745628356934, "learning_rate": 3.8513167207342524e-05, "loss": 0.4018, "num_input_tokens_seen": 17236880, "step": 12730 }, { "epoch": 0.40762435183406953, "grad_norm": 0.6221851706504822, "learning_rate": 3.850490645839044e-05, "loss": 0.4631, "num_input_tokens_seen": 17243680, "step": 12735 }, { "epoch": 0.40778439280455797, "grad_norm": 1.075121283531189, "learning_rate": 3.849664362681207e-05, "loss": 0.49, "num_input_tokens_seen": 17250464, "step": 12740 }, { "epoch": 0.4079444337750464, "grad_norm": 0.9508699178695679, "learning_rate": 3.848837871388165e-05, "loss": 0.5381, "num_input_tokens_seen": 17257168, "step": 12745 }, { "epoch": 0.40810447474553485, "grad_norm": 1.7270439863204956, "learning_rate": 3.848011172087371e-05, "loss": 0.5988, "num_input_tokens_seen": 17264096, "step": 12750 }, { "epoch": 0.4082645157160233, "grad_norm": 0.9508028626441956, "learning_rate": 3.847184264906315e-05, "loss": 0.409, "num_input_tokens_seen": 17271088, "step": 12755 }, { "epoch": 0.40842455668651173, "grad_norm": 0.9474120736122131, "learning_rate": 3.846357149972516e-05, "loss": 0.3509, "num_input_tokens_seen": 17277680, "step": 12760 }, { "epoch": 0.40858459765700017, "grad_norm": 0.8399868607521057, "learning_rate": 3.8455298274135246e-05, "loss": 0.4666, "num_input_tokens_seen": 17284272, "step": 12765 }, { "epoch": 0.4087446386274886, "grad_norm": 0.9663114547729492, "learning_rate": 3.8447022973569254e-05, "loss": 0.4874, "num_input_tokens_seen": 17291120, "step": 12770 }, { "epoch": 0.4089046795979771, "grad_norm": 1.794861912727356, "learning_rate": 3.843874559930332e-05, "loss": 0.5747, "num_input_tokens_seen": 17297472, "step": 12775 }, { "epoch": 0.40906472056846555, "grad_norm": 0.8246877193450928, "learning_rate": 3.843046615261394e-05, "loss": 0.6242, "num_input_tokens_seen": 17304224, "step": 12780 }, { "epoch": 0.409224761538954, "grad_norm": 1.2510700225830078, "learning_rate": 3.842218463477791e-05, "loss": 0.6847, "num_input_tokens_seen": 17311216, "step": 12785 }, { "epoch": 0.4093848025094424, "grad_norm": 2.8272182941436768, "learning_rate": 3.841390104707233e-05, "loss": 0.5018, "num_input_tokens_seen": 17317936, "step": 12790 }, { "epoch": 0.40954484347993086, "grad_norm": 1.4561794996261597, "learning_rate": 3.8405615390774643e-05, "loss": 0.3846, "num_input_tokens_seen": 17324512, "step": 12795 }, { "epoch": 0.4097048844504193, "grad_norm": 0.7042632102966309, "learning_rate": 3.839732766716259e-05, "loss": 0.5369, "num_input_tokens_seen": 17331072, "step": 12800 }, { "epoch": 0.4097048844504193, "eval_loss": 0.48986726999282837, "eval_runtime": 499.3187, "eval_samples_per_second": 27.81, "eval_steps_per_second": 13.905, "num_input_tokens_seen": 17331072, "step": 12800 }, { "epoch": 0.40986492542090774, "grad_norm": 2.163691759109497, "learning_rate": 3.838903787751425e-05, "loss": 0.5353, "num_input_tokens_seen": 17338080, "step": 12805 }, { "epoch": 0.4100249663913962, "grad_norm": 0.6763746738433838, "learning_rate": 3.838074602310802e-05, "loss": 0.4613, "num_input_tokens_seen": 17344368, "step": 12810 }, { "epoch": 0.4101850073618846, "grad_norm": 0.7158538699150085, "learning_rate": 3.837245210522258e-05, "loss": 0.5167, "num_input_tokens_seen": 17351152, "step": 12815 }, { "epoch": 0.41034504833237306, "grad_norm": 1.207715392112732, "learning_rate": 3.8364156125136996e-05, "loss": 0.4164, "num_input_tokens_seen": 17357616, "step": 12820 }, { "epoch": 0.41050508930286156, "grad_norm": 0.9788351655006409, "learning_rate": 3.835585808413059e-05, "loss": 0.3538, "num_input_tokens_seen": 17364352, "step": 12825 }, { "epoch": 0.41066513027335, "grad_norm": 1.4715487957000732, "learning_rate": 3.8347557983483024e-05, "loss": 0.5245, "num_input_tokens_seen": 17371056, "step": 12830 }, { "epoch": 0.41082517124383844, "grad_norm": 0.8583643436431885, "learning_rate": 3.833925582447428e-05, "loss": 0.6057, "num_input_tokens_seen": 17378032, "step": 12835 }, { "epoch": 0.4109852122143269, "grad_norm": 0.961536169052124, "learning_rate": 3.8330951608384656e-05, "loss": 0.459, "num_input_tokens_seen": 17385104, "step": 12840 }, { "epoch": 0.4111452531848153, "grad_norm": 1.2229011058807373, "learning_rate": 3.832264533649477e-05, "loss": 0.4652, "num_input_tokens_seen": 17391616, "step": 12845 }, { "epoch": 0.41130529415530376, "grad_norm": 0.9692617058753967, "learning_rate": 3.8314337010085555e-05, "loss": 0.3713, "num_input_tokens_seen": 17398336, "step": 12850 }, { "epoch": 0.4114653351257922, "grad_norm": 0.7631275057792664, "learning_rate": 3.830602663043824e-05, "loss": 0.3393, "num_input_tokens_seen": 17404848, "step": 12855 }, { "epoch": 0.41162537609628064, "grad_norm": 0.5206641554832458, "learning_rate": 3.8297714198834414e-05, "loss": 0.4592, "num_input_tokens_seen": 17411312, "step": 12860 }, { "epoch": 0.4117854170667691, "grad_norm": 0.6280103921890259, "learning_rate": 3.828939971655595e-05, "loss": 0.5236, "num_input_tokens_seen": 17418128, "step": 12865 }, { "epoch": 0.4119454580372575, "grad_norm": 1.3228216171264648, "learning_rate": 3.828108318488505e-05, "loss": 0.4122, "num_input_tokens_seen": 17424672, "step": 12870 }, { "epoch": 0.41210549900774596, "grad_norm": 1.0528162717819214, "learning_rate": 3.8272764605104216e-05, "loss": 0.3302, "num_input_tokens_seen": 17431472, "step": 12875 }, { "epoch": 0.41226553997823445, "grad_norm": 0.9918424487113953, "learning_rate": 3.826444397849628e-05, "loss": 0.3469, "num_input_tokens_seen": 17438480, "step": 12880 }, { "epoch": 0.4124255809487229, "grad_norm": 1.9508111476898193, "learning_rate": 3.825612130634439e-05, "loss": 0.4596, "num_input_tokens_seen": 17444800, "step": 12885 }, { "epoch": 0.41258562191921133, "grad_norm": 0.6042405962944031, "learning_rate": 3.824779658993202e-05, "loss": 0.3379, "num_input_tokens_seen": 17451568, "step": 12890 }, { "epoch": 0.41274566288969977, "grad_norm": 0.42383307218551636, "learning_rate": 3.823946983054292e-05, "loss": 0.319, "num_input_tokens_seen": 17458144, "step": 12895 }, { "epoch": 0.4129057038601882, "grad_norm": 1.107054352760315, "learning_rate": 3.82311410294612e-05, "loss": 0.4861, "num_input_tokens_seen": 17464736, "step": 12900 }, { "epoch": 0.41306574483067665, "grad_norm": 0.7433157563209534, "learning_rate": 3.822281018797127e-05, "loss": 0.3646, "num_input_tokens_seen": 17471760, "step": 12905 }, { "epoch": 0.4132257858011651, "grad_norm": 0.7336546778678894, "learning_rate": 3.821447730735783e-05, "loss": 0.5202, "num_input_tokens_seen": 17478176, "step": 12910 }, { "epoch": 0.41338582677165353, "grad_norm": 1.4123060703277588, "learning_rate": 3.820614238890592e-05, "loss": 0.3965, "num_input_tokens_seen": 17484400, "step": 12915 }, { "epoch": 0.41354586774214197, "grad_norm": 1.7735894918441772, "learning_rate": 3.819780543390091e-05, "loss": 0.4407, "num_input_tokens_seen": 17491488, "step": 12920 }, { "epoch": 0.4137059087126304, "grad_norm": 0.8377105593681335, "learning_rate": 3.818946644362844e-05, "loss": 0.4846, "num_input_tokens_seen": 17497776, "step": 12925 }, { "epoch": 0.4138659496831189, "grad_norm": 4.492810249328613, "learning_rate": 3.81811254193745e-05, "loss": 0.3111, "num_input_tokens_seen": 17504960, "step": 12930 }, { "epoch": 0.41402599065360735, "grad_norm": 0.5231596231460571, "learning_rate": 3.8172782362425366e-05, "loss": 0.4206, "num_input_tokens_seen": 17511824, "step": 12935 }, { "epoch": 0.4141860316240958, "grad_norm": 1.053873896598816, "learning_rate": 3.816443727406765e-05, "loss": 0.4254, "num_input_tokens_seen": 17518544, "step": 12940 }, { "epoch": 0.4143460725945842, "grad_norm": 1.4670652151107788, "learning_rate": 3.815609015558829e-05, "loss": 0.4957, "num_input_tokens_seen": 17525440, "step": 12945 }, { "epoch": 0.41450611356507266, "grad_norm": 1.7582231760025024, "learning_rate": 3.814774100827448e-05, "loss": 0.3226, "num_input_tokens_seen": 17532128, "step": 12950 }, { "epoch": 0.4146661545355611, "grad_norm": 1.3294622898101807, "learning_rate": 3.813938983341379e-05, "loss": 0.3809, "num_input_tokens_seen": 17538976, "step": 12955 }, { "epoch": 0.41482619550604954, "grad_norm": 2.4223992824554443, "learning_rate": 3.813103663229407e-05, "loss": 0.5015, "num_input_tokens_seen": 17545936, "step": 12960 }, { "epoch": 0.414986236476538, "grad_norm": 0.6114048361778259, "learning_rate": 3.812268140620349e-05, "loss": 0.4807, "num_input_tokens_seen": 17552816, "step": 12965 }, { "epoch": 0.4151462774470264, "grad_norm": 2.3744072914123535, "learning_rate": 3.811432415643051e-05, "loss": 0.5467, "num_input_tokens_seen": 17559616, "step": 12970 }, { "epoch": 0.41530631841751486, "grad_norm": 0.9999582767486572, "learning_rate": 3.8105964884263954e-05, "loss": 0.5645, "num_input_tokens_seen": 17566400, "step": 12975 }, { "epoch": 0.4154663593880033, "grad_norm": 1.1104124784469604, "learning_rate": 3.809760359099291e-05, "loss": 0.4351, "num_input_tokens_seen": 17573024, "step": 12980 }, { "epoch": 0.4156264003584918, "grad_norm": 0.9254279732704163, "learning_rate": 3.8089240277906804e-05, "loss": 0.4061, "num_input_tokens_seen": 17579616, "step": 12985 }, { "epoch": 0.41578644132898024, "grad_norm": 1.6548757553100586, "learning_rate": 3.808087494629535e-05, "loss": 0.5011, "num_input_tokens_seen": 17586624, "step": 12990 }, { "epoch": 0.4159464822994687, "grad_norm": 0.837333083152771, "learning_rate": 3.8072507597448595e-05, "loss": 0.4548, "num_input_tokens_seen": 17592944, "step": 12995 }, { "epoch": 0.4161065232699571, "grad_norm": 2.8793528079986572, "learning_rate": 3.806413823265689e-05, "loss": 0.3812, "num_input_tokens_seen": 17599616, "step": 13000 }, { "epoch": 0.4161065232699571, "eval_loss": 0.48960432410240173, "eval_runtime": 499.6136, "eval_samples_per_second": 27.793, "eval_steps_per_second": 13.897, "num_input_tokens_seen": 17599616, "step": 13000 }, { "epoch": 0.41626656424044556, "grad_norm": 1.72840416431427, "learning_rate": 3.805576685321089e-05, "loss": 0.5548, "num_input_tokens_seen": 17606688, "step": 13005 }, { "epoch": 0.416426605210934, "grad_norm": 0.633876621723175, "learning_rate": 3.804739346040158e-05, "loss": 0.468, "num_input_tokens_seen": 17613568, "step": 13010 }, { "epoch": 0.41658664618142244, "grad_norm": 1.488675832748413, "learning_rate": 3.8039018055520234e-05, "loss": 0.401, "num_input_tokens_seen": 17620896, "step": 13015 }, { "epoch": 0.4167466871519109, "grad_norm": 0.87236088514328, "learning_rate": 3.803064063985844e-05, "loss": 0.3382, "num_input_tokens_seen": 17627536, "step": 13020 }, { "epoch": 0.4169067281223993, "grad_norm": 0.7323044538497925, "learning_rate": 3.802226121470811e-05, "loss": 0.322, "num_input_tokens_seen": 17634000, "step": 13025 }, { "epoch": 0.41706676909288776, "grad_norm": 1.657210111618042, "learning_rate": 3.801387978136145e-05, "loss": 0.4769, "num_input_tokens_seen": 17640688, "step": 13030 }, { "epoch": 0.4172268100633762, "grad_norm": 1.5200703144073486, "learning_rate": 3.800549634111099e-05, "loss": 0.5835, "num_input_tokens_seen": 17647648, "step": 13035 }, { "epoch": 0.4173868510338647, "grad_norm": 3.0174331665039062, "learning_rate": 3.799711089524955e-05, "loss": 0.5692, "num_input_tokens_seen": 17654752, "step": 13040 }, { "epoch": 0.41754689200435313, "grad_norm": 0.838486909866333, "learning_rate": 3.7988723445070285e-05, "loss": 0.4858, "num_input_tokens_seen": 17661568, "step": 13045 }, { "epoch": 0.41770693297484157, "grad_norm": 0.6753174662590027, "learning_rate": 3.798033399186663e-05, "loss": 0.4296, "num_input_tokens_seen": 17668224, "step": 13050 }, { "epoch": 0.41786697394533, "grad_norm": 0.7888809442520142, "learning_rate": 3.797194253693237e-05, "loss": 0.5079, "num_input_tokens_seen": 17675392, "step": 13055 }, { "epoch": 0.41802701491581845, "grad_norm": 1.2384167909622192, "learning_rate": 3.796354908156153e-05, "loss": 0.497, "num_input_tokens_seen": 17682688, "step": 13060 }, { "epoch": 0.4181870558863069, "grad_norm": 2.165670156478882, "learning_rate": 3.795515362704853e-05, "loss": 0.4707, "num_input_tokens_seen": 17689568, "step": 13065 }, { "epoch": 0.41834709685679533, "grad_norm": 1.8334547281265259, "learning_rate": 3.794675617468803e-05, "loss": 0.5588, "num_input_tokens_seen": 17696032, "step": 13070 }, { "epoch": 0.41850713782728377, "grad_norm": 1.20870840549469, "learning_rate": 3.793835672577503e-05, "loss": 0.435, "num_input_tokens_seen": 17702512, "step": 13075 }, { "epoch": 0.4186671787977722, "grad_norm": 0.8258212208747864, "learning_rate": 3.7929955281604826e-05, "loss": 0.4874, "num_input_tokens_seen": 17709168, "step": 13080 }, { "epoch": 0.41882721976826065, "grad_norm": 0.7629877924919128, "learning_rate": 3.7921551843473036e-05, "loss": 0.5215, "num_input_tokens_seen": 17715776, "step": 13085 }, { "epoch": 0.41898726073874915, "grad_norm": 1.0885834693908691, "learning_rate": 3.791314641267557e-05, "loss": 0.4361, "num_input_tokens_seen": 17722432, "step": 13090 }, { "epoch": 0.4191473017092376, "grad_norm": 0.9115943312644958, "learning_rate": 3.790473899050864e-05, "loss": 0.5125, "num_input_tokens_seen": 17728752, "step": 13095 }, { "epoch": 0.419307342679726, "grad_norm": 0.8615649938583374, "learning_rate": 3.7896329578268794e-05, "loss": 0.4786, "num_input_tokens_seen": 17735808, "step": 13100 }, { "epoch": 0.41946738365021446, "grad_norm": 1.8561675548553467, "learning_rate": 3.7887918177252855e-05, "loss": 0.4916, "num_input_tokens_seen": 17742512, "step": 13105 }, { "epoch": 0.4196274246207029, "grad_norm": 0.8011605143547058, "learning_rate": 3.787950478875798e-05, "loss": 0.4375, "num_input_tokens_seen": 17749408, "step": 13110 }, { "epoch": 0.41978746559119134, "grad_norm": 0.9086322784423828, "learning_rate": 3.787108941408162e-05, "loss": 0.616, "num_input_tokens_seen": 17756224, "step": 13115 }, { "epoch": 0.4199475065616798, "grad_norm": 1.1114732027053833, "learning_rate": 3.786267205452151e-05, "loss": 0.3543, "num_input_tokens_seen": 17763152, "step": 13120 }, { "epoch": 0.4201075475321682, "grad_norm": 2.8018598556518555, "learning_rate": 3.785425271137573e-05, "loss": 0.6109, "num_input_tokens_seen": 17769840, "step": 13125 }, { "epoch": 0.42026758850265666, "grad_norm": 1.6687711477279663, "learning_rate": 3.7845831385942655e-05, "loss": 0.4883, "num_input_tokens_seen": 17776416, "step": 13130 }, { "epoch": 0.4204276294731451, "grad_norm": 1.7674247026443481, "learning_rate": 3.7837408079520944e-05, "loss": 0.6949, "num_input_tokens_seen": 17782992, "step": 13135 }, { "epoch": 0.42058767044363354, "grad_norm": 1.3479863405227661, "learning_rate": 3.782898279340957e-05, "loss": 0.6505, "num_input_tokens_seen": 17789360, "step": 13140 }, { "epoch": 0.42074771141412204, "grad_norm": 1.1119484901428223, "learning_rate": 3.782055552890784e-05, "loss": 0.4775, "num_input_tokens_seen": 17795760, "step": 13145 }, { "epoch": 0.4209077523846105, "grad_norm": 0.7871937155723572, "learning_rate": 3.781212628731534e-05, "loss": 0.396, "num_input_tokens_seen": 17802112, "step": 13150 }, { "epoch": 0.4210677933550989, "grad_norm": 1.1705217361450195, "learning_rate": 3.7803695069931946e-05, "loss": 0.4959, "num_input_tokens_seen": 17808880, "step": 13155 }, { "epoch": 0.42122783432558736, "grad_norm": 0.6338385939598083, "learning_rate": 3.779526187805789e-05, "loss": 0.5203, "num_input_tokens_seen": 17815696, "step": 13160 }, { "epoch": 0.4213878752960758, "grad_norm": 0.6527264714241028, "learning_rate": 3.778682671299364e-05, "loss": 0.4297, "num_input_tokens_seen": 17822528, "step": 13165 }, { "epoch": 0.42154791626656424, "grad_norm": 1.2767224311828613, "learning_rate": 3.777838957604003e-05, "loss": 0.6928, "num_input_tokens_seen": 17829328, "step": 13170 }, { "epoch": 0.4217079572370527, "grad_norm": 0.5964704155921936, "learning_rate": 3.776995046849816e-05, "loss": 0.5713, "num_input_tokens_seen": 17836000, "step": 13175 }, { "epoch": 0.4218679982075411, "grad_norm": 1.1955498456954956, "learning_rate": 3.776150939166945e-05, "loss": 0.475, "num_input_tokens_seen": 17842800, "step": 13180 }, { "epoch": 0.42202803917802956, "grad_norm": 0.8636594414710999, "learning_rate": 3.775306634685562e-05, "loss": 0.4314, "num_input_tokens_seen": 17849680, "step": 13185 }, { "epoch": 0.422188080148518, "grad_norm": 1.0173832178115845, "learning_rate": 3.7744621335358696e-05, "loss": 0.4403, "num_input_tokens_seen": 17856400, "step": 13190 }, { "epoch": 0.4223481211190065, "grad_norm": 1.1984474658966064, "learning_rate": 3.7736174358481e-05, "loss": 0.3801, "num_input_tokens_seen": 17862784, "step": 13195 }, { "epoch": 0.42250816208949493, "grad_norm": 1.751861333847046, "learning_rate": 3.7727725417525175e-05, "loss": 0.2993, "num_input_tokens_seen": 17869424, "step": 13200 }, { "epoch": 0.42250816208949493, "eval_loss": 0.4891446530818939, "eval_runtime": 499.5536, "eval_samples_per_second": 27.797, "eval_steps_per_second": 13.898, "num_input_tokens_seen": 17869424, "step": 13200 }, { "epoch": 0.42266820305998337, "grad_norm": 1.2818089723587036, "learning_rate": 3.771927451379414e-05, "loss": 0.4848, "num_input_tokens_seen": 17876800, "step": 13205 }, { "epoch": 0.4228282440304718, "grad_norm": 1.3831088542938232, "learning_rate": 3.7710821648591135e-05, "loss": 0.5297, "num_input_tokens_seen": 17883200, "step": 13210 }, { "epoch": 0.42298828500096025, "grad_norm": 0.8105955719947815, "learning_rate": 3.7702366823219694e-05, "loss": 0.2854, "num_input_tokens_seen": 17890144, "step": 13215 }, { "epoch": 0.4231483259714487, "grad_norm": 1.1777011156082153, "learning_rate": 3.769391003898366e-05, "loss": 0.4316, "num_input_tokens_seen": 17896928, "step": 13220 }, { "epoch": 0.42330836694193713, "grad_norm": 1.6935023069381714, "learning_rate": 3.768545129718718e-05, "loss": 0.535, "num_input_tokens_seen": 17903728, "step": 13225 }, { "epoch": 0.42346840791242557, "grad_norm": 1.2679020166397095, "learning_rate": 3.7676990599134686e-05, "loss": 0.5624, "num_input_tokens_seen": 17910816, "step": 13230 }, { "epoch": 0.423628448882914, "grad_norm": 1.9389687776565552, "learning_rate": 3.766852794613095e-05, "loss": 0.4804, "num_input_tokens_seen": 17917504, "step": 13235 }, { "epoch": 0.42378848985340245, "grad_norm": 2.412212371826172, "learning_rate": 3.766006333948099e-05, "loss": 0.5117, "num_input_tokens_seen": 17924544, "step": 13240 }, { "epoch": 0.4239485308238909, "grad_norm": 0.8496106266975403, "learning_rate": 3.765159678049017e-05, "loss": 0.5626, "num_input_tokens_seen": 17931136, "step": 13245 }, { "epoch": 0.4241085717943794, "grad_norm": 1.7047284841537476, "learning_rate": 3.7643128270464134e-05, "loss": 0.6097, "num_input_tokens_seen": 17938416, "step": 13250 }, { "epoch": 0.4242686127648678, "grad_norm": 1.3039520978927612, "learning_rate": 3.763465781070884e-05, "loss": 0.4207, "num_input_tokens_seen": 17945632, "step": 13255 }, { "epoch": 0.42442865373535626, "grad_norm": 3.612677812576294, "learning_rate": 3.762618540253052e-05, "loss": 0.5411, "num_input_tokens_seen": 17953104, "step": 13260 }, { "epoch": 0.4245886947058447, "grad_norm": 0.8205096125602722, "learning_rate": 3.761771104723576e-05, "loss": 0.5224, "num_input_tokens_seen": 17959760, "step": 13265 }, { "epoch": 0.42474873567633314, "grad_norm": 1.6651432514190674, "learning_rate": 3.7609234746131386e-05, "loss": 0.381, "num_input_tokens_seen": 17966576, "step": 13270 }, { "epoch": 0.4249087766468216, "grad_norm": 0.8508557677268982, "learning_rate": 3.7600756500524556e-05, "loss": 0.454, "num_input_tokens_seen": 17973264, "step": 13275 }, { "epoch": 0.42506881761731, "grad_norm": 0.9222484230995178, "learning_rate": 3.759227631172271e-05, "loss": 0.5615, "num_input_tokens_seen": 17979824, "step": 13280 }, { "epoch": 0.42522885858779846, "grad_norm": 0.9261692762374878, "learning_rate": 3.758379418103363e-05, "loss": 0.3279, "num_input_tokens_seen": 17986592, "step": 13285 }, { "epoch": 0.4253888995582869, "grad_norm": 1.0967872142791748, "learning_rate": 3.757531010976534e-05, "loss": 0.3051, "num_input_tokens_seen": 17993648, "step": 13290 }, { "epoch": 0.42554894052877534, "grad_norm": 1.0332833528518677, "learning_rate": 3.75668240992262e-05, "loss": 0.4217, "num_input_tokens_seen": 18000528, "step": 13295 }, { "epoch": 0.42570898149926384, "grad_norm": 1.6419490575790405, "learning_rate": 3.7558336150724865e-05, "loss": 0.5473, "num_input_tokens_seen": 18007792, "step": 13300 }, { "epoch": 0.4258690224697523, "grad_norm": 0.7376577854156494, "learning_rate": 3.754984626557028e-05, "loss": 0.2872, "num_input_tokens_seen": 18014448, "step": 13305 }, { "epoch": 0.4260290634402407, "grad_norm": 2.356626510620117, "learning_rate": 3.754135444507168e-05, "loss": 0.7707, "num_input_tokens_seen": 18021120, "step": 13310 }, { "epoch": 0.42618910441072916, "grad_norm": 1.165686011314392, "learning_rate": 3.753286069053863e-05, "loss": 0.5909, "num_input_tokens_seen": 18027536, "step": 13315 }, { "epoch": 0.4263491453812176, "grad_norm": 1.492568850517273, "learning_rate": 3.7524365003280945e-05, "loss": 0.573, "num_input_tokens_seen": 18034576, "step": 13320 }, { "epoch": 0.42650918635170604, "grad_norm": 1.4644469022750854, "learning_rate": 3.75158673846088e-05, "loss": 0.4018, "num_input_tokens_seen": 18040896, "step": 13325 }, { "epoch": 0.4266692273221945, "grad_norm": 4.0237884521484375, "learning_rate": 3.750736783583262e-05, "loss": 0.3794, "num_input_tokens_seen": 18047776, "step": 13330 }, { "epoch": 0.4268292682926829, "grad_norm": 1.5948702096939087, "learning_rate": 3.7498866358263144e-05, "loss": 0.5498, "num_input_tokens_seen": 18054544, "step": 13335 }, { "epoch": 0.42698930926317136, "grad_norm": 2.652315616607666, "learning_rate": 3.74903629532114e-05, "loss": 0.4289, "num_input_tokens_seen": 18060992, "step": 13340 }, { "epoch": 0.4271493502336598, "grad_norm": 1.7528314590454102, "learning_rate": 3.748185762198873e-05, "loss": 0.4289, "num_input_tokens_seen": 18067424, "step": 13345 }, { "epoch": 0.42730939120414824, "grad_norm": 0.6120824217796326, "learning_rate": 3.747335036590676e-05, "loss": 0.348, "num_input_tokens_seen": 18074032, "step": 13350 }, { "epoch": 0.42746943217463673, "grad_norm": 1.1253732442855835, "learning_rate": 3.7464841186277405e-05, "loss": 0.558, "num_input_tokens_seen": 18080544, "step": 13355 }, { "epoch": 0.42762947314512517, "grad_norm": 0.9544677734375, "learning_rate": 3.7456330084412896e-05, "loss": 0.3908, "num_input_tokens_seen": 18087376, "step": 13360 }, { "epoch": 0.4277895141156136, "grad_norm": 0.9456689357757568, "learning_rate": 3.744781706162576e-05, "loss": 0.3222, "num_input_tokens_seen": 18094208, "step": 13365 }, { "epoch": 0.42794955508610205, "grad_norm": 1.5868470668792725, "learning_rate": 3.743930211922879e-05, "loss": 0.5778, "num_input_tokens_seen": 18100752, "step": 13370 }, { "epoch": 0.4281095960565905, "grad_norm": 1.736526370048523, "learning_rate": 3.743078525853513e-05, "loss": 0.4636, "num_input_tokens_seen": 18107248, "step": 13375 }, { "epoch": 0.42826963702707893, "grad_norm": 1.3311762809753418, "learning_rate": 3.7422266480858154e-05, "loss": 0.3307, "num_input_tokens_seen": 18113856, "step": 13380 }, { "epoch": 0.42842967799756737, "grad_norm": 1.7233184576034546, "learning_rate": 3.741374578751158e-05, "loss": 0.6089, "num_input_tokens_seen": 18120384, "step": 13385 }, { "epoch": 0.4285897189680558, "grad_norm": 1.232845425605774, "learning_rate": 3.740522317980941e-05, "loss": 0.4974, "num_input_tokens_seen": 18127472, "step": 13390 }, { "epoch": 0.42874975993854425, "grad_norm": 1.5095579624176025, "learning_rate": 3.739669865906593e-05, "loss": 0.4199, "num_input_tokens_seen": 18134224, "step": 13395 }, { "epoch": 0.4289098009090327, "grad_norm": 1.2462756633758545, "learning_rate": 3.738817222659573e-05, "loss": 0.4166, "num_input_tokens_seen": 18141136, "step": 13400 }, { "epoch": 0.4289098009090327, "eval_loss": 0.48740243911743164, "eval_runtime": 499.6927, "eval_samples_per_second": 27.789, "eval_steps_per_second": 13.895, "num_input_tokens_seen": 18141136, "step": 13400 }, { "epoch": 0.42906984187952113, "grad_norm": 1.3661267757415771, "learning_rate": 3.73796438837137e-05, "loss": 0.3601, "num_input_tokens_seen": 18148160, "step": 13405 }, { "epoch": 0.4292298828500096, "grad_norm": 0.7704752683639526, "learning_rate": 3.7371113631735e-05, "loss": 0.4013, "num_input_tokens_seen": 18154768, "step": 13410 }, { "epoch": 0.42938992382049807, "grad_norm": 2.0250418186187744, "learning_rate": 3.736258147197512e-05, "loss": 0.4554, "num_input_tokens_seen": 18161632, "step": 13415 }, { "epoch": 0.4295499647909865, "grad_norm": 0.6641039848327637, "learning_rate": 3.735404740574981e-05, "loss": 0.4541, "num_input_tokens_seen": 18168688, "step": 13420 }, { "epoch": 0.42971000576147494, "grad_norm": 1.0427111387252808, "learning_rate": 3.7345511434375145e-05, "loss": 0.3665, "num_input_tokens_seen": 18175296, "step": 13425 }, { "epoch": 0.4298700467319634, "grad_norm": 2.5654361248016357, "learning_rate": 3.733697355916748e-05, "loss": 0.5708, "num_input_tokens_seen": 18182544, "step": 13430 }, { "epoch": 0.4300300877024518, "grad_norm": 1.0560684204101562, "learning_rate": 3.732843378144345e-05, "loss": 0.5159, "num_input_tokens_seen": 18189584, "step": 13435 }, { "epoch": 0.43019012867294026, "grad_norm": 1.4290529489517212, "learning_rate": 3.7319892102519995e-05, "loss": 0.3931, "num_input_tokens_seen": 18196736, "step": 13440 }, { "epoch": 0.4303501696434287, "grad_norm": 1.4684476852416992, "learning_rate": 3.731134852371436e-05, "loss": 0.4525, "num_input_tokens_seen": 18203424, "step": 13445 }, { "epoch": 0.43051021061391714, "grad_norm": 0.9879925847053528, "learning_rate": 3.730280304634408e-05, "loss": 0.4169, "num_input_tokens_seen": 18210256, "step": 13450 }, { "epoch": 0.4306702515844056, "grad_norm": 1.2991739511489868, "learning_rate": 3.729425567172696e-05, "loss": 0.4823, "num_input_tokens_seen": 18217008, "step": 13455 }, { "epoch": 0.4308302925548941, "grad_norm": 1.29166841506958, "learning_rate": 3.728570640118111e-05, "loss": 0.6127, "num_input_tokens_seen": 18223712, "step": 13460 }, { "epoch": 0.4309903335253825, "grad_norm": 0.883462131023407, "learning_rate": 3.727715523602494e-05, "loss": 0.2866, "num_input_tokens_seen": 18230192, "step": 13465 }, { "epoch": 0.43115037449587096, "grad_norm": 1.6438575983047485, "learning_rate": 3.726860217757715e-05, "loss": 0.462, "num_input_tokens_seen": 18236912, "step": 13470 }, { "epoch": 0.4313104154663594, "grad_norm": 0.7591443657875061, "learning_rate": 3.726004722715673e-05, "loss": 0.6893, "num_input_tokens_seen": 18244176, "step": 13475 }, { "epoch": 0.43147045643684784, "grad_norm": 0.6087927222251892, "learning_rate": 3.725149038608296e-05, "loss": 0.3165, "num_input_tokens_seen": 18251184, "step": 13480 }, { "epoch": 0.4316304974073363, "grad_norm": 0.996601939201355, "learning_rate": 3.7242931655675404e-05, "loss": 0.4031, "num_input_tokens_seen": 18258496, "step": 13485 }, { "epoch": 0.4317905383778247, "grad_norm": 1.742184042930603, "learning_rate": 3.7234371037253937e-05, "loss": 0.4583, "num_input_tokens_seen": 18265328, "step": 13490 }, { "epoch": 0.43195057934831316, "grad_norm": 1.6735329627990723, "learning_rate": 3.7225808532138705e-05, "loss": 0.508, "num_input_tokens_seen": 18272480, "step": 13495 }, { "epoch": 0.4321106203188016, "grad_norm": 0.7237532138824463, "learning_rate": 3.721724414165016e-05, "loss": 0.3518, "num_input_tokens_seen": 18278976, "step": 13500 }, { "epoch": 0.43227066128929004, "grad_norm": 1.9745450019836426, "learning_rate": 3.720867786710904e-05, "loss": 0.4282, "num_input_tokens_seen": 18285680, "step": 13505 }, { "epoch": 0.4324307022597785, "grad_norm": 0.8879016041755676, "learning_rate": 3.7200109709836366e-05, "loss": 0.3646, "num_input_tokens_seen": 18292192, "step": 13510 }, { "epoch": 0.43259074323026697, "grad_norm": 0.7496688961982727, "learning_rate": 3.7191539671153465e-05, "loss": 0.321, "num_input_tokens_seen": 18298960, "step": 13515 }, { "epoch": 0.4327507842007554, "grad_norm": 1.2021459341049194, "learning_rate": 3.718296775238193e-05, "loss": 0.3399, "num_input_tokens_seen": 18305904, "step": 13520 }, { "epoch": 0.43291082517124385, "grad_norm": 0.90215003490448, "learning_rate": 3.7174393954843675e-05, "loss": 0.4159, "num_input_tokens_seen": 18312720, "step": 13525 }, { "epoch": 0.4330708661417323, "grad_norm": 1.1435595750808716, "learning_rate": 3.716581827986087e-05, "loss": 0.5466, "num_input_tokens_seen": 18319568, "step": 13530 }, { "epoch": 0.43323090711222073, "grad_norm": 1.5336519479751587, "learning_rate": 3.7157240728756004e-05, "loss": 0.4613, "num_input_tokens_seen": 18326144, "step": 13535 }, { "epoch": 0.43339094808270917, "grad_norm": 1.7514586448669434, "learning_rate": 3.714866130285184e-05, "loss": 0.4505, "num_input_tokens_seen": 18333216, "step": 13540 }, { "epoch": 0.4335509890531976, "grad_norm": 1.0832946300506592, "learning_rate": 3.714008000347143e-05, "loss": 0.5126, "num_input_tokens_seen": 18340368, "step": 13545 }, { "epoch": 0.43371103002368605, "grad_norm": 1.4967936277389526, "learning_rate": 3.7131496831938126e-05, "loss": 0.6438, "num_input_tokens_seen": 18347264, "step": 13550 }, { "epoch": 0.4338710709941745, "grad_norm": 0.6123723983764648, "learning_rate": 3.7122911789575565e-05, "loss": 0.6208, "num_input_tokens_seen": 18354304, "step": 13555 }, { "epoch": 0.43403111196466293, "grad_norm": 1.0551024675369263, "learning_rate": 3.711432487770765e-05, "loss": 0.54, "num_input_tokens_seen": 18360992, "step": 13560 }, { "epoch": 0.4341911529351514, "grad_norm": 2.540195941925049, "learning_rate": 3.710573609765861e-05, "loss": 0.4332, "num_input_tokens_seen": 18367312, "step": 13565 }, { "epoch": 0.43435119390563987, "grad_norm": 1.0287153720855713, "learning_rate": 3.709714545075292e-05, "loss": 0.4304, "num_input_tokens_seen": 18373984, "step": 13570 }, { "epoch": 0.4345112348761283, "grad_norm": 0.9913532137870789, "learning_rate": 3.708855293831538e-05, "loss": 0.3479, "num_input_tokens_seen": 18380720, "step": 13575 }, { "epoch": 0.43467127584661674, "grad_norm": 1.994874358177185, "learning_rate": 3.707995856167107e-05, "loss": 0.4858, "num_input_tokens_seen": 18387552, "step": 13580 }, { "epoch": 0.4348313168171052, "grad_norm": 0.6902313828468323, "learning_rate": 3.707136232214534e-05, "loss": 0.513, "num_input_tokens_seen": 18394272, "step": 13585 }, { "epoch": 0.4349913577875936, "grad_norm": 0.30044224858283997, "learning_rate": 3.7062764221063844e-05, "loss": 0.5044, "num_input_tokens_seen": 18400928, "step": 13590 }, { "epoch": 0.43515139875808206, "grad_norm": 1.6446484327316284, "learning_rate": 3.705416425975252e-05, "loss": 0.4957, "num_input_tokens_seen": 18407472, "step": 13595 }, { "epoch": 0.4353114397285705, "grad_norm": 0.6772439479827881, "learning_rate": 3.704556243953758e-05, "loss": 0.494, "num_input_tokens_seen": 18414272, "step": 13600 }, { "epoch": 0.4353114397285705, "eval_loss": 0.48551803827285767, "eval_runtime": 500.3168, "eval_samples_per_second": 27.754, "eval_steps_per_second": 13.877, "num_input_tokens_seen": 18414272, "step": 13600 }, { "epoch": 0.43547148069905894, "grad_norm": 1.591130018234253, "learning_rate": 3.7036958761745535e-05, "loss": 0.6066, "num_input_tokens_seen": 18420864, "step": 13605 }, { "epoch": 0.4356315216695474, "grad_norm": 1.5195238590240479, "learning_rate": 3.702835322770318e-05, "loss": 0.3645, "num_input_tokens_seen": 18427136, "step": 13610 }, { "epoch": 0.4357915626400358, "grad_norm": 0.9020497798919678, "learning_rate": 3.701974583873761e-05, "loss": 0.5903, "num_input_tokens_seen": 18433920, "step": 13615 }, { "epoch": 0.4359516036105243, "grad_norm": 1.1312472820281982, "learning_rate": 3.701113659617618e-05, "loss": 0.3963, "num_input_tokens_seen": 18440528, "step": 13620 }, { "epoch": 0.43611164458101276, "grad_norm": 1.1973997354507446, "learning_rate": 3.7002525501346535e-05, "loss": 0.6411, "num_input_tokens_seen": 18447088, "step": 13625 }, { "epoch": 0.4362716855515012, "grad_norm": 1.540264368057251, "learning_rate": 3.699391255557664e-05, "loss": 0.497, "num_input_tokens_seen": 18453488, "step": 13630 }, { "epoch": 0.43643172652198964, "grad_norm": 0.5908782482147217, "learning_rate": 3.69852977601947e-05, "loss": 0.4134, "num_input_tokens_seen": 18460608, "step": 13635 }, { "epoch": 0.4365917674924781, "grad_norm": 0.891940712928772, "learning_rate": 3.697668111652922e-05, "loss": 0.4831, "num_input_tokens_seen": 18467536, "step": 13640 }, { "epoch": 0.4367518084629665, "grad_norm": 0.4332504868507385, "learning_rate": 3.6968062625909005e-05, "loss": 0.6633, "num_input_tokens_seen": 18474880, "step": 13645 }, { "epoch": 0.43691184943345496, "grad_norm": 0.9427095055580139, "learning_rate": 3.6959442289663135e-05, "loss": 0.6956, "num_input_tokens_seen": 18482048, "step": 13650 }, { "epoch": 0.4370718904039434, "grad_norm": 1.19693124294281, "learning_rate": 3.695082010912098e-05, "loss": 0.4878, "num_input_tokens_seen": 18488752, "step": 13655 }, { "epoch": 0.43723193137443184, "grad_norm": 2.9506022930145264, "learning_rate": 3.694219608561217e-05, "loss": 0.393, "num_input_tokens_seen": 18495360, "step": 13660 }, { "epoch": 0.4373919723449203, "grad_norm": 1.5186288356781006, "learning_rate": 3.693357022046665e-05, "loss": 0.437, "num_input_tokens_seen": 18502208, "step": 13665 }, { "epoch": 0.43755201331540877, "grad_norm": 0.8250058889389038, "learning_rate": 3.6924942515014644e-05, "loss": 0.4819, "num_input_tokens_seen": 18509232, "step": 13670 }, { "epoch": 0.4377120542858972, "grad_norm": 1.624508023262024, "learning_rate": 3.691631297058664e-05, "loss": 0.5334, "num_input_tokens_seen": 18515888, "step": 13675 }, { "epoch": 0.43787209525638565, "grad_norm": 0.639493465423584, "learning_rate": 3.6907681588513424e-05, "loss": 0.5049, "num_input_tokens_seen": 18522784, "step": 13680 }, { "epoch": 0.4380321362268741, "grad_norm": 0.9850865602493286, "learning_rate": 3.689904837012606e-05, "loss": 0.3295, "num_input_tokens_seen": 18529552, "step": 13685 }, { "epoch": 0.43819217719736253, "grad_norm": 1.089218020439148, "learning_rate": 3.689041331675591e-05, "loss": 0.5867, "num_input_tokens_seen": 18536512, "step": 13690 }, { "epoch": 0.43835221816785097, "grad_norm": 0.41318678855895996, "learning_rate": 3.688177642973461e-05, "loss": 0.3716, "num_input_tokens_seen": 18543312, "step": 13695 }, { "epoch": 0.4385122591383394, "grad_norm": 1.1943000555038452, "learning_rate": 3.687313771039406e-05, "loss": 0.5608, "num_input_tokens_seen": 18550048, "step": 13700 }, { "epoch": 0.43867230010882785, "grad_norm": 1.458820104598999, "learning_rate": 3.686449716006647e-05, "loss": 0.6182, "num_input_tokens_seen": 18556864, "step": 13705 }, { "epoch": 0.4388323410793163, "grad_norm": 1.0045361518859863, "learning_rate": 3.685585478008432e-05, "loss": 0.537, "num_input_tokens_seen": 18563936, "step": 13710 }, { "epoch": 0.43899238204980473, "grad_norm": 1.3685108423233032, "learning_rate": 3.6847210571780364e-05, "loss": 0.5755, "num_input_tokens_seen": 18570688, "step": 13715 }, { "epoch": 0.43915242302029317, "grad_norm": 1.3044136762619019, "learning_rate": 3.683856453648767e-05, "loss": 0.4958, "num_input_tokens_seen": 18577520, "step": 13720 }, { "epoch": 0.43931246399078167, "grad_norm": 0.6089051961898804, "learning_rate": 3.682991667553954e-05, "loss": 0.4587, "num_input_tokens_seen": 18584192, "step": 13725 }, { "epoch": 0.4394725049612701, "grad_norm": 0.7801868319511414, "learning_rate": 3.6821266990269606e-05, "loss": 0.4304, "num_input_tokens_seen": 18591104, "step": 13730 }, { "epoch": 0.43963254593175854, "grad_norm": 1.2598552703857422, "learning_rate": 3.681261548201174e-05, "loss": 0.4579, "num_input_tokens_seen": 18597792, "step": 13735 }, { "epoch": 0.439792586902247, "grad_norm": 1.044095516204834, "learning_rate": 3.6803962152100125e-05, "loss": 0.6473, "num_input_tokens_seen": 18604704, "step": 13740 }, { "epoch": 0.4399526278727354, "grad_norm": 1.2951431274414062, "learning_rate": 3.67953070018692e-05, "loss": 0.522, "num_input_tokens_seen": 18611440, "step": 13745 }, { "epoch": 0.44011266884322386, "grad_norm": 2.19628643989563, "learning_rate": 3.678665003265371e-05, "loss": 0.5707, "num_input_tokens_seen": 18617840, "step": 13750 }, { "epoch": 0.4402727098137123, "grad_norm": 1.145975112915039, "learning_rate": 3.677799124578867e-05, "loss": 0.4969, "num_input_tokens_seen": 18624304, "step": 13755 }, { "epoch": 0.44043275078420074, "grad_norm": 1.3256946802139282, "learning_rate": 3.676933064260937e-05, "loss": 0.3884, "num_input_tokens_seen": 18631328, "step": 13760 }, { "epoch": 0.4405927917546892, "grad_norm": 0.9727473855018616, "learning_rate": 3.6760668224451365e-05, "loss": 0.4613, "num_input_tokens_seen": 18638192, "step": 13765 }, { "epoch": 0.4407528327251776, "grad_norm": 0.9552671909332275, "learning_rate": 3.675200399265054e-05, "loss": 0.4622, "num_input_tokens_seen": 18645120, "step": 13770 }, { "epoch": 0.44091287369566606, "grad_norm": 1.1420150995254517, "learning_rate": 3.6743337948543014e-05, "loss": 0.5251, "num_input_tokens_seen": 18651776, "step": 13775 }, { "epoch": 0.44107291466615456, "grad_norm": 0.8266005516052246, "learning_rate": 3.6734670093465204e-05, "loss": 0.3598, "num_input_tokens_seen": 18658432, "step": 13780 }, { "epoch": 0.441232955636643, "grad_norm": 0.7505054473876953, "learning_rate": 3.672600042875379e-05, "loss": 0.5144, "num_input_tokens_seen": 18665200, "step": 13785 }, { "epoch": 0.44139299660713144, "grad_norm": 0.7537105679512024, "learning_rate": 3.671732895574575e-05, "loss": 0.4086, "num_input_tokens_seen": 18672016, "step": 13790 }, { "epoch": 0.4415530375776199, "grad_norm": 1.2295483350753784, "learning_rate": 3.670865567577834e-05, "loss": 0.5198, "num_input_tokens_seen": 18678608, "step": 13795 }, { "epoch": 0.4417130785481083, "grad_norm": 0.8564761877059937, "learning_rate": 3.669998059018909e-05, "loss": 0.4413, "num_input_tokens_seen": 18685264, "step": 13800 }, { "epoch": 0.4417130785481083, "eval_loss": 0.4849293529987335, "eval_runtime": 500.8048, "eval_samples_per_second": 27.727, "eval_steps_per_second": 13.864, "num_input_tokens_seen": 18685264, "step": 13800 }, { "epoch": 0.44187311951859676, "grad_norm": 1.790171504020691, "learning_rate": 3.6691303700315796e-05, "loss": 0.622, "num_input_tokens_seen": 18692656, "step": 13805 }, { "epoch": 0.4420331604890852, "grad_norm": 1.5247178077697754, "learning_rate": 3.668262500749655e-05, "loss": 0.557, "num_input_tokens_seen": 18699376, "step": 13810 }, { "epoch": 0.44219320145957364, "grad_norm": 1.0094013214111328, "learning_rate": 3.667394451306971e-05, "loss": 0.3714, "num_input_tokens_seen": 18706224, "step": 13815 }, { "epoch": 0.4423532424300621, "grad_norm": 0.6003481149673462, "learning_rate": 3.666526221837393e-05, "loss": 0.3493, "num_input_tokens_seen": 18713232, "step": 13820 }, { "epoch": 0.4425132834005505, "grad_norm": 1.8146451711654663, "learning_rate": 3.665657812474812e-05, "loss": 0.4668, "num_input_tokens_seen": 18719680, "step": 13825 }, { "epoch": 0.442673324371039, "grad_norm": 1.2503876686096191, "learning_rate": 3.664789223353147e-05, "loss": 0.4195, "num_input_tokens_seen": 18726496, "step": 13830 }, { "epoch": 0.44283336534152745, "grad_norm": 1.4267961978912354, "learning_rate": 3.663920454606347e-05, "loss": 0.5417, "num_input_tokens_seen": 18733216, "step": 13835 }, { "epoch": 0.4429934063120159, "grad_norm": 1.2013121843338013, "learning_rate": 3.6630515063683856e-05, "loss": 0.4488, "num_input_tokens_seen": 18740144, "step": 13840 }, { "epoch": 0.44315344728250433, "grad_norm": 0.6096800565719604, "learning_rate": 3.662182378773267e-05, "loss": 0.3136, "num_input_tokens_seen": 18746864, "step": 13845 }, { "epoch": 0.44331348825299277, "grad_norm": 1.3305836915969849, "learning_rate": 3.66131307195502e-05, "loss": 0.4181, "num_input_tokens_seen": 18753744, "step": 13850 }, { "epoch": 0.4434735292234812, "grad_norm": 1.5617079734802246, "learning_rate": 3.6604435860477034e-05, "loss": 0.4, "num_input_tokens_seen": 18760624, "step": 13855 }, { "epoch": 0.44363357019396965, "grad_norm": 1.5536127090454102, "learning_rate": 3.6595739211854025e-05, "loss": 0.4204, "num_input_tokens_seen": 18767264, "step": 13860 }, { "epoch": 0.4437936111644581, "grad_norm": 1.0311758518218994, "learning_rate": 3.658704077502231e-05, "loss": 0.5528, "num_input_tokens_seen": 18773952, "step": 13865 }, { "epoch": 0.44395365213494653, "grad_norm": 0.9659668207168579, "learning_rate": 3.65783405513233e-05, "loss": 0.4529, "num_input_tokens_seen": 18780336, "step": 13870 }, { "epoch": 0.44411369310543497, "grad_norm": 0.8181794285774231, "learning_rate": 3.656963854209867e-05, "loss": 0.7336, "num_input_tokens_seen": 18786848, "step": 13875 }, { "epoch": 0.4442737340759234, "grad_norm": 1.883866786956787, "learning_rate": 3.656093474869038e-05, "loss": 0.6249, "num_input_tokens_seen": 18793632, "step": 13880 }, { "epoch": 0.4444337750464119, "grad_norm": 0.8231064081192017, "learning_rate": 3.655222917244068e-05, "loss": 0.4449, "num_input_tokens_seen": 18800512, "step": 13885 }, { "epoch": 0.44459381601690035, "grad_norm": 0.984035313129425, "learning_rate": 3.6543521814692054e-05, "loss": 0.5281, "num_input_tokens_seen": 18807200, "step": 13890 }, { "epoch": 0.4447538569873888, "grad_norm": 0.5275024771690369, "learning_rate": 3.653481267678731e-05, "loss": 0.3911, "num_input_tokens_seen": 18813856, "step": 13895 }, { "epoch": 0.4449138979578772, "grad_norm": 1.2783823013305664, "learning_rate": 3.652610176006949e-05, "loss": 0.3809, "num_input_tokens_seen": 18820576, "step": 13900 }, { "epoch": 0.44507393892836566, "grad_norm": 1.004033088684082, "learning_rate": 3.6517389065881925e-05, "loss": 0.3824, "num_input_tokens_seen": 18827184, "step": 13905 }, { "epoch": 0.4452339798988541, "grad_norm": 0.9164702892303467, "learning_rate": 3.650867459556824e-05, "loss": 0.5215, "num_input_tokens_seen": 18833856, "step": 13910 }, { "epoch": 0.44539402086934254, "grad_norm": 0.7845243215560913, "learning_rate": 3.64999583504723e-05, "loss": 0.328, "num_input_tokens_seen": 18840656, "step": 13915 }, { "epoch": 0.445554061839831, "grad_norm": 0.7811794877052307, "learning_rate": 3.649124033193827e-05, "loss": 0.4212, "num_input_tokens_seen": 18847552, "step": 13920 }, { "epoch": 0.4457141028103194, "grad_norm": 0.994693398475647, "learning_rate": 3.648252054131057e-05, "loss": 0.5874, "num_input_tokens_seen": 18854464, "step": 13925 }, { "epoch": 0.44587414378080786, "grad_norm": 1.1100469827651978, "learning_rate": 3.647379897993391e-05, "loss": 0.5913, "num_input_tokens_seen": 18861568, "step": 13930 }, { "epoch": 0.44603418475129636, "grad_norm": 1.4418449401855469, "learning_rate": 3.646507564915325e-05, "loss": 0.3858, "num_input_tokens_seen": 18868240, "step": 13935 }, { "epoch": 0.4461942257217848, "grad_norm": 1.1928914785385132, "learning_rate": 3.645635055031385e-05, "loss": 0.5682, "num_input_tokens_seen": 18875376, "step": 13940 }, { "epoch": 0.44635426669227324, "grad_norm": 0.9794670939445496, "learning_rate": 3.6447623684761224e-05, "loss": 0.4799, "num_input_tokens_seen": 18881904, "step": 13945 }, { "epoch": 0.4465143076627617, "grad_norm": 0.8230293989181519, "learning_rate": 3.643889505384117e-05, "loss": 0.4019, "num_input_tokens_seen": 18888608, "step": 13950 }, { "epoch": 0.4466743486332501, "grad_norm": 0.514246940612793, "learning_rate": 3.6430164658899744e-05, "loss": 0.3934, "num_input_tokens_seen": 18895744, "step": 13955 }, { "epoch": 0.44683438960373856, "grad_norm": 2.474999189376831, "learning_rate": 3.642143250128329e-05, "loss": 0.7182, "num_input_tokens_seen": 18902688, "step": 13960 }, { "epoch": 0.446994430574227, "grad_norm": 1.5628974437713623, "learning_rate": 3.641269858233841e-05, "loss": 0.4255, "num_input_tokens_seen": 18909280, "step": 13965 }, { "epoch": 0.44715447154471544, "grad_norm": 1.342161774635315, "learning_rate": 3.640396290341199e-05, "loss": 0.4698, "num_input_tokens_seen": 18916208, "step": 13970 }, { "epoch": 0.4473145125152039, "grad_norm": 3.0874881744384766, "learning_rate": 3.639522546585118e-05, "loss": 0.6004, "num_input_tokens_seen": 18923120, "step": 13975 }, { "epoch": 0.4474745534856923, "grad_norm": 0.797806441783905, "learning_rate": 3.6386486271003404e-05, "loss": 0.4054, "num_input_tokens_seen": 18930256, "step": 13980 }, { "epoch": 0.44763459445618076, "grad_norm": 1.7286043167114258, "learning_rate": 3.6377745320216346e-05, "loss": 0.4971, "num_input_tokens_seen": 18936880, "step": 13985 }, { "epoch": 0.44779463542666925, "grad_norm": 1.3670763969421387, "learning_rate": 3.636900261483798e-05, "loss": 0.5646, "num_input_tokens_seen": 18943552, "step": 13990 }, { "epoch": 0.4479546763971577, "grad_norm": 0.7805337309837341, "learning_rate": 3.636025815621654e-05, "loss": 0.4073, "num_input_tokens_seen": 18950176, "step": 13995 }, { "epoch": 0.44811471736764613, "grad_norm": 1.943440318107605, "learning_rate": 3.635151194570054e-05, "loss": 0.4482, "num_input_tokens_seen": 18957072, "step": 14000 }, { "epoch": 0.44811471736764613, "eval_loss": 0.4830978810787201, "eval_runtime": 500.0905, "eval_samples_per_second": 27.767, "eval_steps_per_second": 13.883, "num_input_tokens_seen": 18957072, "step": 14000 }, { "epoch": 0.44827475833813457, "grad_norm": 0.9233578443527222, "learning_rate": 3.634276398463873e-05, "loss": 0.4338, "num_input_tokens_seen": 18963792, "step": 14005 }, { "epoch": 0.448434799308623, "grad_norm": 0.8580632209777832, "learning_rate": 3.633401427438018e-05, "loss": 0.5619, "num_input_tokens_seen": 18970752, "step": 14010 }, { "epoch": 0.44859484027911145, "grad_norm": 1.0136831998825073, "learning_rate": 3.63252628162742e-05, "loss": 0.3805, "num_input_tokens_seen": 18977424, "step": 14015 }, { "epoch": 0.4487548812495999, "grad_norm": 0.5699650645256042, "learning_rate": 3.6316509611670364e-05, "loss": 0.4804, "num_input_tokens_seen": 18984224, "step": 14020 }, { "epoch": 0.44891492222008833, "grad_norm": 1.63930344581604, "learning_rate": 3.630775466191854e-05, "loss": 0.3867, "num_input_tokens_seen": 18990784, "step": 14025 }, { "epoch": 0.44907496319057677, "grad_norm": 1.3964060544967651, "learning_rate": 3.629899796836884e-05, "loss": 0.5135, "num_input_tokens_seen": 18997488, "step": 14030 }, { "epoch": 0.4492350041610652, "grad_norm": 1.2912498712539673, "learning_rate": 3.6290239532371666e-05, "loss": 0.3993, "num_input_tokens_seen": 19004416, "step": 14035 }, { "epoch": 0.4493950451315537, "grad_norm": 1.9138729572296143, "learning_rate": 3.628147935527767e-05, "loss": 0.5477, "num_input_tokens_seen": 19011600, "step": 14040 }, { "epoch": 0.44955508610204215, "grad_norm": 0.7207756042480469, "learning_rate": 3.627271743843779e-05, "loss": 0.2632, "num_input_tokens_seen": 19018304, "step": 14045 }, { "epoch": 0.4497151270725306, "grad_norm": 0.9704263210296631, "learning_rate": 3.626395378320321e-05, "loss": 0.6186, "num_input_tokens_seen": 19024992, "step": 14050 }, { "epoch": 0.449875168043019, "grad_norm": 0.7231895923614502, "learning_rate": 3.625518839092541e-05, "loss": 0.4393, "num_input_tokens_seen": 19031856, "step": 14055 }, { "epoch": 0.45003520901350746, "grad_norm": 0.6289212107658386, "learning_rate": 3.624642126295612e-05, "loss": 0.4229, "num_input_tokens_seen": 19038432, "step": 14060 }, { "epoch": 0.4501952499839959, "grad_norm": 0.5614485740661621, "learning_rate": 3.6237652400647345e-05, "loss": 0.5197, "num_input_tokens_seen": 19045264, "step": 14065 }, { "epoch": 0.45035529095448434, "grad_norm": 1.1746249198913574, "learning_rate": 3.622888180535134e-05, "loss": 0.3648, "num_input_tokens_seen": 19052496, "step": 14070 }, { "epoch": 0.4505153319249728, "grad_norm": 0.43515390157699585, "learning_rate": 3.6220109478420655e-05, "loss": 0.4078, "num_input_tokens_seen": 19059232, "step": 14075 }, { "epoch": 0.4506753728954612, "grad_norm": 0.7689614295959473, "learning_rate": 3.6211335421208084e-05, "loss": 0.4537, "num_input_tokens_seen": 19065952, "step": 14080 }, { "epoch": 0.45083541386594966, "grad_norm": 1.5187575817108154, "learning_rate": 3.62025596350667e-05, "loss": 0.3745, "num_input_tokens_seen": 19072384, "step": 14085 }, { "epoch": 0.4509954548364381, "grad_norm": 0.7014159560203552, "learning_rate": 3.619378212134984e-05, "loss": 0.4243, "num_input_tokens_seen": 19079440, "step": 14090 }, { "epoch": 0.4511554958069266, "grad_norm": 0.34386080503463745, "learning_rate": 3.618500288141111e-05, "loss": 0.3363, "num_input_tokens_seen": 19086112, "step": 14095 }, { "epoch": 0.45131553677741504, "grad_norm": 1.8484607934951782, "learning_rate": 3.617622191660438e-05, "loss": 0.8706, "num_input_tokens_seen": 19093264, "step": 14100 }, { "epoch": 0.4514755777479035, "grad_norm": 1.488945722579956, "learning_rate": 3.616743922828377e-05, "loss": 0.4109, "num_input_tokens_seen": 19100560, "step": 14105 }, { "epoch": 0.4516356187183919, "grad_norm": 1.6456570625305176, "learning_rate": 3.615865481780371e-05, "loss": 0.4122, "num_input_tokens_seen": 19107344, "step": 14110 }, { "epoch": 0.45179565968888036, "grad_norm": 1.0008739233016968, "learning_rate": 3.614986868651883e-05, "loss": 0.54, "num_input_tokens_seen": 19114064, "step": 14115 }, { "epoch": 0.4519557006593688, "grad_norm": 2.83911395072937, "learning_rate": 3.614108083578409e-05, "loss": 0.4444, "num_input_tokens_seen": 19120512, "step": 14120 }, { "epoch": 0.45211574162985724, "grad_norm": 2.1459691524505615, "learning_rate": 3.613229126695467e-05, "loss": 0.5451, "num_input_tokens_seen": 19127584, "step": 14125 }, { "epoch": 0.4522757826003457, "grad_norm": 0.6380545496940613, "learning_rate": 3.612349998138605e-05, "loss": 0.4203, "num_input_tokens_seen": 19134320, "step": 14130 }, { "epoch": 0.4524358235708341, "grad_norm": 2.43410587310791, "learning_rate": 3.6114706980433946e-05, "loss": 0.6257, "num_input_tokens_seen": 19141056, "step": 14135 }, { "epoch": 0.45259586454132256, "grad_norm": 2.4855759143829346, "learning_rate": 3.610591226545435e-05, "loss": 0.5306, "num_input_tokens_seen": 19148144, "step": 14140 }, { "epoch": 0.452755905511811, "grad_norm": 1.116119623184204, "learning_rate": 3.6097115837803505e-05, "loss": 0.5139, "num_input_tokens_seen": 19154912, "step": 14145 }, { "epoch": 0.4529159464822995, "grad_norm": 1.4451366662979126, "learning_rate": 3.608831769883795e-05, "loss": 0.7167, "num_input_tokens_seen": 19162224, "step": 14150 }, { "epoch": 0.45307598745278793, "grad_norm": 1.1747711896896362, "learning_rate": 3.607951784991446e-05, "loss": 0.5168, "num_input_tokens_seen": 19169152, "step": 14155 }, { "epoch": 0.45323602842327637, "grad_norm": 0.3952060341835022, "learning_rate": 3.6070716292390085e-05, "loss": 0.301, "num_input_tokens_seen": 19175648, "step": 14160 }, { "epoch": 0.4533960693937648, "grad_norm": 2.642686605453491, "learning_rate": 3.606191302762213e-05, "loss": 0.6645, "num_input_tokens_seen": 19182048, "step": 14165 }, { "epoch": 0.45355611036425325, "grad_norm": 1.1455645561218262, "learning_rate": 3.605310805696818e-05, "loss": 0.4075, "num_input_tokens_seen": 19188704, "step": 14170 }, { "epoch": 0.4537161513347417, "grad_norm": 0.809980571269989, "learning_rate": 3.6044301381786067e-05, "loss": 0.3379, "num_input_tokens_seen": 19195408, "step": 14175 }, { "epoch": 0.45387619230523013, "grad_norm": 0.6401693820953369, "learning_rate": 3.6035493003433883e-05, "loss": 0.4419, "num_input_tokens_seen": 19202256, "step": 14180 }, { "epoch": 0.45403623327571857, "grad_norm": 1.2380050420761108, "learning_rate": 3.6026682923269994e-05, "loss": 0.4738, "num_input_tokens_seen": 19208992, "step": 14185 }, { "epoch": 0.454196274246207, "grad_norm": 1.4398342370986938, "learning_rate": 3.6017871142653034e-05, "loss": 0.5648, "num_input_tokens_seen": 19216208, "step": 14190 }, { "epoch": 0.45435631521669545, "grad_norm": 1.0439597368240356, "learning_rate": 3.600905766294189e-05, "loss": 0.3466, "num_input_tokens_seen": 19223328, "step": 14195 }, { "epoch": 0.45451635618718395, "grad_norm": 0.6344534158706665, "learning_rate": 3.60002424854957e-05, "loss": 0.3391, "num_input_tokens_seen": 19230480, "step": 14200 }, { "epoch": 0.45451635618718395, "eval_loss": 0.48136696219444275, "eval_runtime": 500.8146, "eval_samples_per_second": 27.727, "eval_steps_per_second": 13.863, "num_input_tokens_seen": 19230480, "step": 14200 }, { "epoch": 0.4546763971576724, "grad_norm": 0.6400049328804016, "learning_rate": 3.5991425611673876e-05, "loss": 0.4883, "num_input_tokens_seen": 19236976, "step": 14205 }, { "epoch": 0.4548364381281608, "grad_norm": 1.1569499969482422, "learning_rate": 3.5982607042836105e-05, "loss": 0.4088, "num_input_tokens_seen": 19243616, "step": 14210 }, { "epoch": 0.45499647909864926, "grad_norm": 1.0892066955566406, "learning_rate": 3.597378678034231e-05, "loss": 0.3759, "num_input_tokens_seen": 19250352, "step": 14215 }, { "epoch": 0.4551565200691377, "grad_norm": 1.4516080617904663, "learning_rate": 3.596496482555269e-05, "loss": 0.4102, "num_input_tokens_seen": 19257328, "step": 14220 }, { "epoch": 0.45531656103962614, "grad_norm": 1.2145699262619019, "learning_rate": 3.595614117982769e-05, "loss": 0.4401, "num_input_tokens_seen": 19264160, "step": 14225 }, { "epoch": 0.4554766020101146, "grad_norm": 0.6741130948066711, "learning_rate": 3.594731584452805e-05, "loss": 0.444, "num_input_tokens_seen": 19271216, "step": 14230 }, { "epoch": 0.455636642980603, "grad_norm": 1.5875670909881592, "learning_rate": 3.593848882101472e-05, "loss": 0.4575, "num_input_tokens_seen": 19277984, "step": 14235 }, { "epoch": 0.45579668395109146, "grad_norm": 1.2097394466400146, "learning_rate": 3.592966011064896e-05, "loss": 0.5488, "num_input_tokens_seen": 19285040, "step": 14240 }, { "epoch": 0.4559567249215799, "grad_norm": 1.670649528503418, "learning_rate": 3.592082971479226e-05, "loss": 0.2709, "num_input_tokens_seen": 19292352, "step": 14245 }, { "epoch": 0.45611676589206834, "grad_norm": 1.1232287883758545, "learning_rate": 3.5911997634806385e-05, "loss": 0.3748, "num_input_tokens_seen": 19299104, "step": 14250 }, { "epoch": 0.45627680686255684, "grad_norm": 1.2791017293930054, "learning_rate": 3.5903163872053336e-05, "loss": 0.4816, "num_input_tokens_seen": 19305936, "step": 14255 }, { "epoch": 0.4564368478330453, "grad_norm": 0.4760448932647705, "learning_rate": 3.58943284278954e-05, "loss": 0.5738, "num_input_tokens_seen": 19312208, "step": 14260 }, { "epoch": 0.4565968888035337, "grad_norm": 1.4405056238174438, "learning_rate": 3.588549130369512e-05, "loss": 0.5111, "num_input_tokens_seen": 19319232, "step": 14265 }, { "epoch": 0.45675692977402216, "grad_norm": 2.2243738174438477, "learning_rate": 3.5876652500815274e-05, "loss": 0.3716, "num_input_tokens_seen": 19326144, "step": 14270 }, { "epoch": 0.4569169707445106, "grad_norm": 0.936850905418396, "learning_rate": 3.586781202061894e-05, "loss": 0.2987, "num_input_tokens_seen": 19332928, "step": 14275 }, { "epoch": 0.45707701171499904, "grad_norm": 0.8314521312713623, "learning_rate": 3.585896986446942e-05, "loss": 0.4308, "num_input_tokens_seen": 19339472, "step": 14280 }, { "epoch": 0.4572370526854875, "grad_norm": 0.8478997349739075, "learning_rate": 3.585012603373028e-05, "loss": 0.5589, "num_input_tokens_seen": 19346144, "step": 14285 }, { "epoch": 0.4573970936559759, "grad_norm": 0.9310707449913025, "learning_rate": 3.584128052976535e-05, "loss": 0.4794, "num_input_tokens_seen": 19352704, "step": 14290 }, { "epoch": 0.45755713462646436, "grad_norm": 0.7885555624961853, "learning_rate": 3.5832433353938724e-05, "loss": 0.3951, "num_input_tokens_seen": 19359440, "step": 14295 }, { "epoch": 0.4577171755969528, "grad_norm": 2.4220614433288574, "learning_rate": 3.5823584507614746e-05, "loss": 0.5257, "num_input_tokens_seen": 19366320, "step": 14300 }, { "epoch": 0.4578772165674413, "grad_norm": 1.7753539085388184, "learning_rate": 3.581473399215802e-05, "loss": 0.3262, "num_input_tokens_seen": 19373264, "step": 14305 }, { "epoch": 0.45803725753792973, "grad_norm": 0.6998465657234192, "learning_rate": 3.580588180893341e-05, "loss": 0.5454, "num_input_tokens_seen": 19380272, "step": 14310 }, { "epoch": 0.45819729850841817, "grad_norm": 0.8707395195960999, "learning_rate": 3.579702795930602e-05, "loss": 0.5246, "num_input_tokens_seen": 19387360, "step": 14315 }, { "epoch": 0.4583573394789066, "grad_norm": 1.4228497743606567, "learning_rate": 3.578817244464125e-05, "loss": 0.5386, "num_input_tokens_seen": 19394272, "step": 14320 }, { "epoch": 0.45851738044939505, "grad_norm": 1.0615442991256714, "learning_rate": 3.577931526630471e-05, "loss": 0.4059, "num_input_tokens_seen": 19401664, "step": 14325 }, { "epoch": 0.4586774214198835, "grad_norm": 1.385966181755066, "learning_rate": 3.577045642566229e-05, "loss": 0.515, "num_input_tokens_seen": 19408144, "step": 14330 }, { "epoch": 0.45883746239037193, "grad_norm": 0.7431609034538269, "learning_rate": 3.576159592408014e-05, "loss": 0.6382, "num_input_tokens_seen": 19414864, "step": 14335 }, { "epoch": 0.45899750336086037, "grad_norm": 0.7443194389343262, "learning_rate": 3.575273376292466e-05, "loss": 0.3778, "num_input_tokens_seen": 19421104, "step": 14340 }, { "epoch": 0.4591575443313488, "grad_norm": 0.6374250650405884, "learning_rate": 3.574386994356251e-05, "loss": 0.5434, "num_input_tokens_seen": 19427888, "step": 14345 }, { "epoch": 0.45931758530183725, "grad_norm": 1.4108318090438843, "learning_rate": 3.573500446736059e-05, "loss": 0.563, "num_input_tokens_seen": 19434288, "step": 14350 }, { "epoch": 0.4594776262723257, "grad_norm": 1.2196769714355469, "learning_rate": 3.5726137335686094e-05, "loss": 0.4038, "num_input_tokens_seen": 19441536, "step": 14355 }, { "epoch": 0.4596376672428142, "grad_norm": 0.7833574414253235, "learning_rate": 3.571726854990642e-05, "loss": 0.5032, "num_input_tokens_seen": 19448624, "step": 14360 }, { "epoch": 0.4597977082133026, "grad_norm": 1.2358342409133911, "learning_rate": 3.570839811138925e-05, "loss": 0.597, "num_input_tokens_seen": 19455296, "step": 14365 }, { "epoch": 0.45995774918379106, "grad_norm": 1.8241093158721924, "learning_rate": 3.569952602150252e-05, "loss": 0.7032, "num_input_tokens_seen": 19462112, "step": 14370 }, { "epoch": 0.4601177901542795, "grad_norm": 0.8857606649398804, "learning_rate": 3.569065228161442e-05, "loss": 0.5486, "num_input_tokens_seen": 19468816, "step": 14375 }, { "epoch": 0.46027783112476794, "grad_norm": 0.6612582802772522, "learning_rate": 3.5681776893093395e-05, "loss": 0.5195, "num_input_tokens_seen": 19475776, "step": 14380 }, { "epoch": 0.4604378720952564, "grad_norm": 1.893301010131836, "learning_rate": 3.5672899857308134e-05, "loss": 0.5992, "num_input_tokens_seen": 19482224, "step": 14385 }, { "epoch": 0.4605979130657448, "grad_norm": 0.8282082080841064, "learning_rate": 3.566402117562759e-05, "loss": 0.3575, "num_input_tokens_seen": 19489312, "step": 14390 }, { "epoch": 0.46075795403623326, "grad_norm": 0.7572600841522217, "learning_rate": 3.565514084942097e-05, "loss": 0.481, "num_input_tokens_seen": 19496576, "step": 14395 }, { "epoch": 0.4609179950067217, "grad_norm": 1.014831304550171, "learning_rate": 3.564625888005773e-05, "loss": 0.3414, "num_input_tokens_seen": 19503472, "step": 14400 }, { "epoch": 0.4609179950067217, "eval_loss": 0.48162662982940674, "eval_runtime": 499.8836, "eval_samples_per_second": 27.778, "eval_steps_per_second": 13.889, "num_input_tokens_seen": 19503472, "step": 14400 }, { "epoch": 0.46107803597721014, "grad_norm": 1.3166347742080688, "learning_rate": 3.563737526890759e-05, "loss": 0.4782, "num_input_tokens_seen": 19510016, "step": 14405 }, { "epoch": 0.4612380769476986, "grad_norm": 1.4899733066558838, "learning_rate": 3.562849001734049e-05, "loss": 0.5936, "num_input_tokens_seen": 19516832, "step": 14410 }, { "epoch": 0.4613981179181871, "grad_norm": 2.0248970985412598, "learning_rate": 3.561960312672667e-05, "loss": 0.4671, "num_input_tokens_seen": 19523152, "step": 14415 }, { "epoch": 0.4615581588886755, "grad_norm": 1.194138526916504, "learning_rate": 3.5610714598436596e-05, "loss": 0.6793, "num_input_tokens_seen": 19529904, "step": 14420 }, { "epoch": 0.46171819985916396, "grad_norm": 1.0233031511306763, "learning_rate": 3.5601824433840986e-05, "loss": 0.3157, "num_input_tokens_seen": 19536592, "step": 14425 }, { "epoch": 0.4618782408296524, "grad_norm": 1.0056926012039185, "learning_rate": 3.559293263431082e-05, "loss": 0.5308, "num_input_tokens_seen": 19543104, "step": 14430 }, { "epoch": 0.46203828180014084, "grad_norm": 0.6064141392707825, "learning_rate": 3.558403920121732e-05, "loss": 0.4432, "num_input_tokens_seen": 19549824, "step": 14435 }, { "epoch": 0.4621983227706293, "grad_norm": 4.109951496124268, "learning_rate": 3.557514413593197e-05, "loss": 0.4885, "num_input_tokens_seen": 19556560, "step": 14440 }, { "epoch": 0.4623583637411177, "grad_norm": 0.5797832012176514, "learning_rate": 3.55662474398265e-05, "loss": 0.4107, "num_input_tokens_seen": 19563600, "step": 14445 }, { "epoch": 0.46251840471160616, "grad_norm": 0.8193063139915466, "learning_rate": 3.555734911427288e-05, "loss": 0.4517, "num_input_tokens_seen": 19570272, "step": 14450 }, { "epoch": 0.4626784456820946, "grad_norm": 1.920416235923767, "learning_rate": 3.5548449160643363e-05, "loss": 0.499, "num_input_tokens_seen": 19577136, "step": 14455 }, { "epoch": 0.46283848665258304, "grad_norm": 2.342120409011841, "learning_rate": 3.553954758031043e-05, "loss": 0.3614, "num_input_tokens_seen": 19584016, "step": 14460 }, { "epoch": 0.46299852762307153, "grad_norm": 1.5728763341903687, "learning_rate": 3.5530644374646815e-05, "loss": 0.4905, "num_input_tokens_seen": 19590640, "step": 14465 }, { "epoch": 0.46315856859355997, "grad_norm": 0.8008591532707214, "learning_rate": 3.552173954502549e-05, "loss": 0.3062, "num_input_tokens_seen": 19597296, "step": 14470 }, { "epoch": 0.4633186095640484, "grad_norm": 1.1356137990951538, "learning_rate": 3.55128330928197e-05, "loss": 0.5462, "num_input_tokens_seen": 19603776, "step": 14475 }, { "epoch": 0.46347865053453685, "grad_norm": 1.8308320045471191, "learning_rate": 3.550392501940294e-05, "loss": 0.4908, "num_input_tokens_seen": 19610320, "step": 14480 }, { "epoch": 0.4636386915050253, "grad_norm": 1.3581712245941162, "learning_rate": 3.5495015326148945e-05, "loss": 0.4164, "num_input_tokens_seen": 19617168, "step": 14485 }, { "epoch": 0.46379873247551373, "grad_norm": 0.7251900434494019, "learning_rate": 3.548610401443169e-05, "loss": 0.5036, "num_input_tokens_seen": 19623936, "step": 14490 }, { "epoch": 0.46395877344600217, "grad_norm": 0.8181205987930298, "learning_rate": 3.547719108562543e-05, "loss": 0.3521, "num_input_tokens_seen": 19630512, "step": 14495 }, { "epoch": 0.4641188144164906, "grad_norm": 1.9359350204467773, "learning_rate": 3.546827654110464e-05, "loss": 0.5703, "num_input_tokens_seen": 19637008, "step": 14500 }, { "epoch": 0.46427885538697905, "grad_norm": 0.6775594353675842, "learning_rate": 3.545936038224405e-05, "loss": 0.3825, "num_input_tokens_seen": 19643712, "step": 14505 }, { "epoch": 0.4644388963574675, "grad_norm": 1.7063446044921875, "learning_rate": 3.545044261041864e-05, "loss": 0.3211, "num_input_tokens_seen": 19650496, "step": 14510 }, { "epoch": 0.46459893732795593, "grad_norm": 0.5602341294288635, "learning_rate": 3.5441523227003657e-05, "loss": 0.5802, "num_input_tokens_seen": 19657840, "step": 14515 }, { "epoch": 0.4647589782984444, "grad_norm": 0.9830994606018066, "learning_rate": 3.543260223337459e-05, "loss": 0.3726, "num_input_tokens_seen": 19664528, "step": 14520 }, { "epoch": 0.46491901926893286, "grad_norm": 1.8685451745986938, "learning_rate": 3.542367963090714e-05, "loss": 0.4691, "num_input_tokens_seen": 19671488, "step": 14525 }, { "epoch": 0.4650790602394213, "grad_norm": 1.8598010540008545, "learning_rate": 3.5414755420977295e-05, "loss": 0.4652, "num_input_tokens_seen": 19678384, "step": 14530 }, { "epoch": 0.46523910120990974, "grad_norm": 0.9169935584068298, "learning_rate": 3.54058296049613e-05, "loss": 0.5105, "num_input_tokens_seen": 19685504, "step": 14535 }, { "epoch": 0.4653991421803982, "grad_norm": 1.147271752357483, "learning_rate": 3.53969021842356e-05, "loss": 0.421, "num_input_tokens_seen": 19692192, "step": 14540 }, { "epoch": 0.4655591831508866, "grad_norm": 1.8532894849777222, "learning_rate": 3.5387973160176926e-05, "loss": 0.4644, "num_input_tokens_seen": 19698880, "step": 14545 }, { "epoch": 0.46571922412137506, "grad_norm": 0.8858861327171326, "learning_rate": 3.537904253416224e-05, "loss": 0.3919, "num_input_tokens_seen": 19705776, "step": 14550 }, { "epoch": 0.4658792650918635, "grad_norm": 2.019740581512451, "learning_rate": 3.537011030756878e-05, "loss": 1.042, "num_input_tokens_seen": 19715952, "step": 14555 }, { "epoch": 0.46603930606235194, "grad_norm": 1.095686674118042, "learning_rate": 3.536117648177399e-05, "loss": 0.5661, "num_input_tokens_seen": 19722896, "step": 14560 }, { "epoch": 0.4661993470328404, "grad_norm": 1.9411228895187378, "learning_rate": 3.535224105815558e-05, "loss": 0.4116, "num_input_tokens_seen": 19729648, "step": 14565 }, { "epoch": 0.4663593880033289, "grad_norm": 1.102890968322754, "learning_rate": 3.5343304038091494e-05, "loss": 0.4959, "num_input_tokens_seen": 19736368, "step": 14570 }, { "epoch": 0.4665194289738173, "grad_norm": 1.0962048768997192, "learning_rate": 3.5334365422959955e-05, "loss": 0.5552, "num_input_tokens_seen": 19743152, "step": 14575 }, { "epoch": 0.46667946994430576, "grad_norm": 0.6894339323043823, "learning_rate": 3.5325425214139396e-05, "loss": 0.4066, "num_input_tokens_seen": 19749872, "step": 14580 }, { "epoch": 0.4668395109147942, "grad_norm": 0.8786467909812927, "learning_rate": 3.531648341300851e-05, "loss": 0.7374, "num_input_tokens_seen": 19756864, "step": 14585 }, { "epoch": 0.46699955188528264, "grad_norm": 0.818324625492096, "learning_rate": 3.530754002094623e-05, "loss": 0.4483, "num_input_tokens_seen": 19763648, "step": 14590 }, { "epoch": 0.4671595928557711, "grad_norm": 1.3910943269729614, "learning_rate": 3.529859503933175e-05, "loss": 0.3959, "num_input_tokens_seen": 19770544, "step": 14595 }, { "epoch": 0.4673196338262595, "grad_norm": 0.9683017730712891, "learning_rate": 3.52896484695445e-05, "loss": 0.5706, "num_input_tokens_seen": 19777344, "step": 14600 }, { "epoch": 0.4673196338262595, "eval_loss": 0.4806956946849823, "eval_runtime": 500.4399, "eval_samples_per_second": 27.748, "eval_steps_per_second": 13.874, "num_input_tokens_seen": 19777344, "step": 14600 }, { "epoch": 0.46747967479674796, "grad_norm": 1.0134902000427246, "learning_rate": 3.528070031296414e-05, "loss": 0.5682, "num_input_tokens_seen": 19784496, "step": 14605 }, { "epoch": 0.4676397157672364, "grad_norm": 0.6001584529876709, "learning_rate": 3.5271750570970605e-05, "loss": 0.4665, "num_input_tokens_seen": 19791520, "step": 14610 }, { "epoch": 0.46779975673772484, "grad_norm": 1.4064861536026, "learning_rate": 3.526279924494405e-05, "loss": 0.6092, "num_input_tokens_seen": 19798368, "step": 14615 }, { "epoch": 0.4679597977082133, "grad_norm": 0.6329464316368103, "learning_rate": 3.5253846336264874e-05, "loss": 0.2766, "num_input_tokens_seen": 19805152, "step": 14620 }, { "epoch": 0.46811983867870177, "grad_norm": 1.3291484117507935, "learning_rate": 3.5244891846313736e-05, "loss": 0.4191, "num_input_tokens_seen": 19811792, "step": 14625 }, { "epoch": 0.4682798796491902, "grad_norm": 1.1188850402832031, "learning_rate": 3.5235935776471527e-05, "loss": 0.4731, "num_input_tokens_seen": 19818384, "step": 14630 }, { "epoch": 0.46843992061967865, "grad_norm": 0.7876870036125183, "learning_rate": 3.522697812811939e-05, "loss": 0.4354, "num_input_tokens_seen": 19825360, "step": 14635 }, { "epoch": 0.4685999615901671, "grad_norm": 1.0386309623718262, "learning_rate": 3.521801890263871e-05, "loss": 0.6507, "num_input_tokens_seen": 19831968, "step": 14640 }, { "epoch": 0.46876000256065553, "grad_norm": 0.8016396760940552, "learning_rate": 3.5209058101411114e-05, "loss": 0.406, "num_input_tokens_seen": 19838928, "step": 14645 }, { "epoch": 0.46892004353114397, "grad_norm": 0.660746157169342, "learning_rate": 3.520009572581845e-05, "loss": 0.353, "num_input_tokens_seen": 19845696, "step": 14650 }, { "epoch": 0.4690800845016324, "grad_norm": 0.7438501119613647, "learning_rate": 3.519113177724285e-05, "loss": 0.302, "num_input_tokens_seen": 19852704, "step": 14655 }, { "epoch": 0.46924012547212085, "grad_norm": 3.317383050918579, "learning_rate": 3.5182166257066656e-05, "loss": 0.58, "num_input_tokens_seen": 19859872, "step": 14660 }, { "epoch": 0.4694001664426093, "grad_norm": 2.250864028930664, "learning_rate": 3.517319916667247e-05, "loss": 0.5101, "num_input_tokens_seen": 19866464, "step": 14665 }, { "epoch": 0.46956020741309773, "grad_norm": 0.6705631613731384, "learning_rate": 3.516423050744313e-05, "loss": 0.4259, "num_input_tokens_seen": 19873328, "step": 14670 }, { "epoch": 0.4697202483835862, "grad_norm": 1.025203824043274, "learning_rate": 3.5155260280761704e-05, "loss": 0.479, "num_input_tokens_seen": 19879936, "step": 14675 }, { "epoch": 0.46988028935407467, "grad_norm": 1.1361336708068848, "learning_rate": 3.514628848801154e-05, "loss": 0.4069, "num_input_tokens_seen": 19886640, "step": 14680 }, { "epoch": 0.4700403303245631, "grad_norm": 2.194176197052002, "learning_rate": 3.5137315130576174e-05, "loss": 0.5511, "num_input_tokens_seen": 19893696, "step": 14685 }, { "epoch": 0.47020037129505154, "grad_norm": 2.268200397491455, "learning_rate": 3.512834020983942e-05, "loss": 0.5918, "num_input_tokens_seen": 19900224, "step": 14690 }, { "epoch": 0.47036041226554, "grad_norm": 1.2893956899642944, "learning_rate": 3.5119363727185334e-05, "loss": 0.4598, "num_input_tokens_seen": 19906688, "step": 14695 }, { "epoch": 0.4705204532360284, "grad_norm": 0.8260759711265564, "learning_rate": 3.511038568399819e-05, "loss": 0.4384, "num_input_tokens_seen": 19913728, "step": 14700 }, { "epoch": 0.47068049420651686, "grad_norm": 1.3019566535949707, "learning_rate": 3.510140608166251e-05, "loss": 0.5849, "num_input_tokens_seen": 19920896, "step": 14705 }, { "epoch": 0.4708405351770053, "grad_norm": 0.6874845027923584, "learning_rate": 3.509242492156308e-05, "loss": 0.5619, "num_input_tokens_seen": 19927584, "step": 14710 }, { "epoch": 0.47100057614749374, "grad_norm": 0.9883540272712708, "learning_rate": 3.5083442205084896e-05, "loss": 0.4563, "num_input_tokens_seen": 19933904, "step": 14715 }, { "epoch": 0.4711606171179822, "grad_norm": 0.7534438967704773, "learning_rate": 3.507445793361321e-05, "loss": 0.3552, "num_input_tokens_seen": 19940752, "step": 14720 }, { "epoch": 0.4713206580884706, "grad_norm": 1.024632453918457, "learning_rate": 3.5065472108533505e-05, "loss": 0.4218, "num_input_tokens_seen": 19947408, "step": 14725 }, { "epoch": 0.4714806990589591, "grad_norm": 1.196765422821045, "learning_rate": 3.5056484731231504e-05, "loss": 0.2957, "num_input_tokens_seen": 19954304, "step": 14730 }, { "epoch": 0.47164074002944756, "grad_norm": 1.3347100019454956, "learning_rate": 3.504749580309319e-05, "loss": 0.5814, "num_input_tokens_seen": 19961440, "step": 14735 }, { "epoch": 0.471800780999936, "grad_norm": 1.1721893548965454, "learning_rate": 3.5038505325504753e-05, "loss": 0.4031, "num_input_tokens_seen": 19968208, "step": 14740 }, { "epoch": 0.47196082197042444, "grad_norm": 1.0686076879501343, "learning_rate": 3.502951329985264e-05, "loss": 0.4749, "num_input_tokens_seen": 19974960, "step": 14745 }, { "epoch": 0.4721208629409129, "grad_norm": 1.5502748489379883, "learning_rate": 3.502051972752354e-05, "loss": 0.5183, "num_input_tokens_seen": 19981760, "step": 14750 }, { "epoch": 0.4722809039114013, "grad_norm": 1.1720256805419922, "learning_rate": 3.5011524609904374e-05, "loss": 0.3832, "num_input_tokens_seen": 19988736, "step": 14755 }, { "epoch": 0.47244094488188976, "grad_norm": 0.7488219738006592, "learning_rate": 3.50025279483823e-05, "loss": 0.4527, "num_input_tokens_seen": 19995728, "step": 14760 }, { "epoch": 0.4726009858523782, "grad_norm": 0.7865149974822998, "learning_rate": 3.499352974434472e-05, "loss": 0.3503, "num_input_tokens_seen": 20002672, "step": 14765 }, { "epoch": 0.47276102682286664, "grad_norm": 1.8324015140533447, "learning_rate": 3.498452999917926e-05, "loss": 0.5752, "num_input_tokens_seen": 20009408, "step": 14770 }, { "epoch": 0.4729210677933551, "grad_norm": 2.4782018661499023, "learning_rate": 3.4975528714273795e-05, "loss": 0.63, "num_input_tokens_seen": 20016080, "step": 14775 }, { "epoch": 0.4730811087638435, "grad_norm": 1.249840497970581, "learning_rate": 3.4966525891016454e-05, "loss": 0.6087, "num_input_tokens_seen": 20022816, "step": 14780 }, { "epoch": 0.473241149734332, "grad_norm": 0.7342067956924438, "learning_rate": 3.495752153079557e-05, "loss": 0.3229, "num_input_tokens_seen": 20029136, "step": 14785 }, { "epoch": 0.47340119070482045, "grad_norm": 1.1227582693099976, "learning_rate": 3.494851563499974e-05, "loss": 0.4586, "num_input_tokens_seen": 20035648, "step": 14790 }, { "epoch": 0.4735612316753089, "grad_norm": 0.8856181502342224, "learning_rate": 3.493950820501777e-05, "loss": 0.6574, "num_input_tokens_seen": 20042240, "step": 14795 }, { "epoch": 0.47372127264579733, "grad_norm": 0.5393078327178955, "learning_rate": 3.493049924223872e-05, "loss": 0.3137, "num_input_tokens_seen": 20049328, "step": 14800 }, { "epoch": 0.47372127264579733, "eval_loss": 0.4796735644340515, "eval_runtime": 500.7301, "eval_samples_per_second": 27.732, "eval_steps_per_second": 13.866, "num_input_tokens_seen": 20049328, "step": 14800 }, { "epoch": 0.47388131361628577, "grad_norm": 0.9289237260818481, "learning_rate": 3.49214887480519e-05, "loss": 0.3758, "num_input_tokens_seen": 20056064, "step": 14805 }, { "epoch": 0.4740413545867742, "grad_norm": 0.7639195322990417, "learning_rate": 3.4912476723846834e-05, "loss": 0.3118, "num_input_tokens_seen": 20063104, "step": 14810 }, { "epoch": 0.47420139555726265, "grad_norm": 0.5412048697471619, "learning_rate": 3.490346317101328e-05, "loss": 0.307, "num_input_tokens_seen": 20069680, "step": 14815 }, { "epoch": 0.4743614365277511, "grad_norm": 1.0012516975402832, "learning_rate": 3.4894448090941266e-05, "loss": 0.3864, "num_input_tokens_seen": 20076512, "step": 14820 }, { "epoch": 0.47452147749823953, "grad_norm": 1.0609753131866455, "learning_rate": 3.488543148502101e-05, "loss": 0.535, "num_input_tokens_seen": 20082992, "step": 14825 }, { "epoch": 0.47468151846872797, "grad_norm": 1.6707249879837036, "learning_rate": 3.487641335464299e-05, "loss": 0.3184, "num_input_tokens_seen": 20089536, "step": 14830 }, { "epoch": 0.47484155943921647, "grad_norm": 1.0930161476135254, "learning_rate": 3.4867393701197914e-05, "loss": 0.3631, "num_input_tokens_seen": 20096288, "step": 14835 }, { "epoch": 0.4750016004097049, "grad_norm": 1.9397534132003784, "learning_rate": 3.485837252607673e-05, "loss": 0.332, "num_input_tokens_seen": 20102992, "step": 14840 }, { "epoch": 0.47516164138019334, "grad_norm": 1.3005367517471313, "learning_rate": 3.4849349830670615e-05, "loss": 0.5492, "num_input_tokens_seen": 20109904, "step": 14845 }, { "epoch": 0.4753216823506818, "grad_norm": 1.063464641571045, "learning_rate": 3.4840325616370976e-05, "loss": 0.4577, "num_input_tokens_seen": 20116608, "step": 14850 }, { "epoch": 0.4754817233211702, "grad_norm": 0.7499387860298157, "learning_rate": 3.483129988456947e-05, "loss": 0.3501, "num_input_tokens_seen": 20123552, "step": 14855 }, { "epoch": 0.47564176429165866, "grad_norm": 1.4741413593292236, "learning_rate": 3.482227263665797e-05, "loss": 0.4418, "num_input_tokens_seen": 20130560, "step": 14860 }, { "epoch": 0.4758018052621471, "grad_norm": 0.9751808047294617, "learning_rate": 3.48132438740286e-05, "loss": 0.4281, "num_input_tokens_seen": 20137248, "step": 14865 }, { "epoch": 0.47596184623263554, "grad_norm": 0.6100972890853882, "learning_rate": 3.48042135980737e-05, "loss": 0.2328, "num_input_tokens_seen": 20143760, "step": 14870 }, { "epoch": 0.476121887203124, "grad_norm": 0.848004162311554, "learning_rate": 3.479518181018586e-05, "loss": 0.5479, "num_input_tokens_seen": 20150512, "step": 14875 }, { "epoch": 0.4762819281736124, "grad_norm": 2.12319278717041, "learning_rate": 3.4786148511757886e-05, "loss": 0.5769, "num_input_tokens_seen": 20157248, "step": 14880 }, { "epoch": 0.47644196914410086, "grad_norm": 1.6553499698638916, "learning_rate": 3.477711370418284e-05, "loss": 0.4642, "num_input_tokens_seen": 20164192, "step": 14885 }, { "epoch": 0.47660201011458936, "grad_norm": 0.9628608226776123, "learning_rate": 3.476807738885399e-05, "loss": 0.4045, "num_input_tokens_seen": 20170976, "step": 14890 }, { "epoch": 0.4767620510850778, "grad_norm": 0.6365246176719666, "learning_rate": 3.475903956716485e-05, "loss": 0.3027, "num_input_tokens_seen": 20177488, "step": 14895 }, { "epoch": 0.47692209205556624, "grad_norm": 0.7141900658607483, "learning_rate": 3.475000024050917e-05, "loss": 0.4325, "num_input_tokens_seen": 20183904, "step": 14900 }, { "epoch": 0.4770821330260547, "grad_norm": 1.4656904935836792, "learning_rate": 3.4740959410280926e-05, "loss": 0.4052, "num_input_tokens_seen": 20190528, "step": 14905 }, { "epoch": 0.4772421739965431, "grad_norm": 0.6531578302383423, "learning_rate": 3.4731917077874324e-05, "loss": 0.3324, "num_input_tokens_seen": 20197216, "step": 14910 }, { "epoch": 0.47740221496703156, "grad_norm": 0.9292129278182983, "learning_rate": 3.4722873244683816e-05, "loss": 0.3769, "num_input_tokens_seen": 20203792, "step": 14915 }, { "epoch": 0.47756225593752, "grad_norm": 3.051236152648926, "learning_rate": 3.4713827912104065e-05, "loss": 0.6224, "num_input_tokens_seen": 20210608, "step": 14920 }, { "epoch": 0.47772229690800844, "grad_norm": 0.9970913529396057, "learning_rate": 3.470478108152998e-05, "loss": 0.3984, "num_input_tokens_seen": 20217456, "step": 14925 }, { "epoch": 0.4778823378784969, "grad_norm": 1.0929675102233887, "learning_rate": 3.4695732754356695e-05, "loss": 0.482, "num_input_tokens_seen": 20224384, "step": 14930 }, { "epoch": 0.4780423788489853, "grad_norm": 0.9224982857704163, "learning_rate": 3.4686682931979576e-05, "loss": 0.4251, "num_input_tokens_seen": 20230848, "step": 14935 }, { "epoch": 0.4782024198194738, "grad_norm": 0.7020156979560852, "learning_rate": 3.467763161579422e-05, "loss": 0.3889, "num_input_tokens_seen": 20237856, "step": 14940 }, { "epoch": 0.47836246078996225, "grad_norm": 1.6286154985427856, "learning_rate": 3.466857880719645e-05, "loss": 0.4813, "num_input_tokens_seen": 20245040, "step": 14945 }, { "epoch": 0.4785225017604507, "grad_norm": 1.9407763481140137, "learning_rate": 3.465952450758233e-05, "loss": 0.4233, "num_input_tokens_seen": 20251504, "step": 14950 }, { "epoch": 0.47868254273093913, "grad_norm": 0.6835934519767761, "learning_rate": 3.4650468718348126e-05, "loss": 0.5465, "num_input_tokens_seen": 20258528, "step": 14955 }, { "epoch": 0.47884258370142757, "grad_norm": 1.1470600366592407, "learning_rate": 3.464141144089038e-05, "loss": 0.6462, "num_input_tokens_seen": 20265200, "step": 14960 }, { "epoch": 0.479002624671916, "grad_norm": 1.0198142528533936, "learning_rate": 3.463235267660583e-05, "loss": 0.4242, "num_input_tokens_seen": 20271840, "step": 14965 }, { "epoch": 0.47916266564240445, "grad_norm": 0.6723988652229309, "learning_rate": 3.462329242689145e-05, "loss": 0.3349, "num_input_tokens_seen": 20278592, "step": 14970 }, { "epoch": 0.4793227066128929, "grad_norm": 1.0750370025634766, "learning_rate": 3.461423069314444e-05, "loss": 0.3839, "num_input_tokens_seen": 20285280, "step": 14975 }, { "epoch": 0.47948274758338133, "grad_norm": 2.289430856704712, "learning_rate": 3.460516747676224e-05, "loss": 0.3101, "num_input_tokens_seen": 20292032, "step": 14980 }, { "epoch": 0.47964278855386977, "grad_norm": 1.5249427556991577, "learning_rate": 3.459610277914251e-05, "loss": 0.4098, "num_input_tokens_seen": 20298768, "step": 14985 }, { "epoch": 0.4798028295243582, "grad_norm": 0.7204676866531372, "learning_rate": 3.458703660168314e-05, "loss": 0.5449, "num_input_tokens_seen": 20305728, "step": 14990 }, { "epoch": 0.4799628704948467, "grad_norm": 0.9393253326416016, "learning_rate": 3.457796894578224e-05, "loss": 0.3749, "num_input_tokens_seen": 20312432, "step": 14995 }, { "epoch": 0.48012291146533514, "grad_norm": 1.1559858322143555, "learning_rate": 3.456889981283817e-05, "loss": 0.4268, "num_input_tokens_seen": 20319488, "step": 15000 }, { "epoch": 0.48012291146533514, "eval_loss": 0.47980818152427673, "eval_runtime": 500.2956, "eval_samples_per_second": 27.756, "eval_steps_per_second": 13.878, "num_input_tokens_seen": 20319488, "step": 15000 }, { "epoch": 0.4802829524358236, "grad_norm": 0.6509031057357788, "learning_rate": 3.45598292042495e-05, "loss": 0.6192, "num_input_tokens_seen": 20326416, "step": 15005 }, { "epoch": 0.480442993406312, "grad_norm": 0.8763227462768555, "learning_rate": 3.4550757121415035e-05, "loss": 0.3337, "num_input_tokens_seen": 20332816, "step": 15010 }, { "epoch": 0.48060303437680046, "grad_norm": 2.9906227588653564, "learning_rate": 3.454168356573378e-05, "loss": 0.5938, "num_input_tokens_seen": 20339072, "step": 15015 }, { "epoch": 0.4807630753472889, "grad_norm": 0.8945491909980774, "learning_rate": 3.453260853860503e-05, "loss": 0.3839, "num_input_tokens_seen": 20346176, "step": 15020 }, { "epoch": 0.48092311631777734, "grad_norm": 1.3223730325698853, "learning_rate": 3.452353204142824e-05, "loss": 0.5074, "num_input_tokens_seen": 20352800, "step": 15025 }, { "epoch": 0.4810831572882658, "grad_norm": 1.9683173894882202, "learning_rate": 3.4514454075603136e-05, "loss": 0.4793, "num_input_tokens_seen": 20359520, "step": 15030 }, { "epoch": 0.4812431982587542, "grad_norm": 1.3677237033843994, "learning_rate": 3.450537464252964e-05, "loss": 0.5296, "num_input_tokens_seen": 20366720, "step": 15035 }, { "epoch": 0.48140323922924266, "grad_norm": 1.9386136531829834, "learning_rate": 3.4496293743607925e-05, "loss": 0.6383, "num_input_tokens_seen": 20373216, "step": 15040 }, { "epoch": 0.48156328019973116, "grad_norm": 0.9601774215698242, "learning_rate": 3.448721138023838e-05, "loss": 0.2314, "num_input_tokens_seen": 20380016, "step": 15045 }, { "epoch": 0.4817233211702196, "grad_norm": 1.086519479751587, "learning_rate": 3.447812755382162e-05, "loss": 0.6105, "num_input_tokens_seen": 20386624, "step": 15050 }, { "epoch": 0.48188336214070804, "grad_norm": 1.0412492752075195, "learning_rate": 3.446904226575847e-05, "loss": 0.4355, "num_input_tokens_seen": 20392960, "step": 15055 }, { "epoch": 0.4820434031111965, "grad_norm": 2.1380691528320312, "learning_rate": 3.445995551745002e-05, "loss": 0.6434, "num_input_tokens_seen": 20400032, "step": 15060 }, { "epoch": 0.4822034440816849, "grad_norm": 1.0173437595367432, "learning_rate": 3.445086731029753e-05, "loss": 0.5613, "num_input_tokens_seen": 20406800, "step": 15065 }, { "epoch": 0.48236348505217336, "grad_norm": 1.8643763065338135, "learning_rate": 3.444177764570255e-05, "loss": 0.4482, "num_input_tokens_seen": 20413536, "step": 15070 }, { "epoch": 0.4825235260226618, "grad_norm": 1.5961236953735352, "learning_rate": 3.44326865250668e-05, "loss": 0.5458, "num_input_tokens_seen": 20420144, "step": 15075 }, { "epoch": 0.48268356699315024, "grad_norm": 0.561506986618042, "learning_rate": 3.442359394979225e-05, "loss": 0.5464, "num_input_tokens_seen": 20426784, "step": 15080 }, { "epoch": 0.4828436079636387, "grad_norm": 0.4927544593811035, "learning_rate": 3.441449992128108e-05, "loss": 0.4803, "num_input_tokens_seen": 20433824, "step": 15085 }, { "epoch": 0.4830036489341271, "grad_norm": 1.4877617359161377, "learning_rate": 3.440540444093573e-05, "loss": 0.4406, "num_input_tokens_seen": 20440464, "step": 15090 }, { "epoch": 0.48316368990461556, "grad_norm": 1.319882869720459, "learning_rate": 3.43963075101588e-05, "loss": 0.52, "num_input_tokens_seen": 20447376, "step": 15095 }, { "epoch": 0.48332373087510405, "grad_norm": 1.094825267791748, "learning_rate": 3.438720913035318e-05, "loss": 0.3381, "num_input_tokens_seen": 20454096, "step": 15100 }, { "epoch": 0.4834837718455925, "grad_norm": 1.0192081928253174, "learning_rate": 3.437810930292195e-05, "loss": 0.3931, "num_input_tokens_seen": 20461104, "step": 15105 }, { "epoch": 0.48364381281608093, "grad_norm": 1.2102344036102295, "learning_rate": 3.43690080292684e-05, "loss": 0.3711, "num_input_tokens_seen": 20467728, "step": 15110 }, { "epoch": 0.48380385378656937, "grad_norm": 1.1361358165740967, "learning_rate": 3.435990531079608e-05, "loss": 0.3243, "num_input_tokens_seen": 20474160, "step": 15115 }, { "epoch": 0.4839638947570578, "grad_norm": 0.2306067794561386, "learning_rate": 3.435080114890874e-05, "loss": 0.326, "num_input_tokens_seen": 20480896, "step": 15120 }, { "epoch": 0.48412393572754625, "grad_norm": 0.5004125833511353, "learning_rate": 3.434169554501035e-05, "loss": 0.4025, "num_input_tokens_seen": 20487664, "step": 15125 }, { "epoch": 0.4842839766980347, "grad_norm": 1.2813096046447754, "learning_rate": 3.433258850050511e-05, "loss": 0.4976, "num_input_tokens_seen": 20494624, "step": 15130 }, { "epoch": 0.48444401766852313, "grad_norm": 1.6965992450714111, "learning_rate": 3.4323480016797446e-05, "loss": 0.2668, "num_input_tokens_seen": 20501520, "step": 15135 }, { "epoch": 0.48460405863901157, "grad_norm": 2.2886290550231934, "learning_rate": 3.4314370095291995e-05, "loss": 0.5533, "num_input_tokens_seen": 20508560, "step": 15140 }, { "epoch": 0.4847640996095, "grad_norm": 0.9750733971595764, "learning_rate": 3.430525873739363e-05, "loss": 0.6197, "num_input_tokens_seen": 20515504, "step": 15145 }, { "epoch": 0.48492414057998845, "grad_norm": 1.571469783782959, "learning_rate": 3.429614594450743e-05, "loss": 0.5144, "num_input_tokens_seen": 20522640, "step": 15150 }, { "epoch": 0.48508418155047694, "grad_norm": 2.29028058052063, "learning_rate": 3.428703171803869e-05, "loss": 0.6368, "num_input_tokens_seen": 20529648, "step": 15155 }, { "epoch": 0.4852442225209654, "grad_norm": 0.5672218799591064, "learning_rate": 3.4277916059392964e-05, "loss": 0.4015, "num_input_tokens_seen": 20536496, "step": 15160 }, { "epoch": 0.4854042634914538, "grad_norm": 0.758071780204773, "learning_rate": 3.426879896997598e-05, "loss": 0.4194, "num_input_tokens_seen": 20543552, "step": 15165 }, { "epoch": 0.48556430446194226, "grad_norm": 1.0433034896850586, "learning_rate": 3.425968045119372e-05, "loss": 0.3569, "num_input_tokens_seen": 20550176, "step": 15170 }, { "epoch": 0.4857243454324307, "grad_norm": 0.5815495848655701, "learning_rate": 3.425056050445237e-05, "loss": 0.3915, "num_input_tokens_seen": 20556960, "step": 15175 }, { "epoch": 0.48588438640291914, "grad_norm": 1.8381257057189941, "learning_rate": 3.4241439131158336e-05, "loss": 0.3793, "num_input_tokens_seen": 20563808, "step": 15180 }, { "epoch": 0.4860444273734076, "grad_norm": 1.231325626373291, "learning_rate": 3.423231633271825e-05, "loss": 0.4285, "num_input_tokens_seen": 20570304, "step": 15185 }, { "epoch": 0.486204468343896, "grad_norm": 1.1594280004501343, "learning_rate": 3.4223192110538985e-05, "loss": 0.3118, "num_input_tokens_seen": 20576656, "step": 15190 }, { "epoch": 0.48636450931438446, "grad_norm": 1.3347910642623901, "learning_rate": 3.4214066466027575e-05, "loss": 0.4489, "num_input_tokens_seen": 20583184, "step": 15195 }, { "epoch": 0.4865245502848729, "grad_norm": 0.4990743398666382, "learning_rate": 3.4204939400591325e-05, "loss": 0.5074, "num_input_tokens_seen": 20589760, "step": 15200 }, { "epoch": 0.4865245502848729, "eval_loss": 0.4789141118526459, "eval_runtime": 500.2092, "eval_samples_per_second": 27.76, "eval_steps_per_second": 13.88, "num_input_tokens_seen": 20589760, "step": 15200 }, { "epoch": 0.4866845912553614, "grad_norm": 1.2303533554077148, "learning_rate": 3.419581091563775e-05, "loss": 0.4711, "num_input_tokens_seen": 20596368, "step": 15205 }, { "epoch": 0.48684463222584984, "grad_norm": 1.144346833229065, "learning_rate": 3.418668101257456e-05, "loss": 0.4035, "num_input_tokens_seen": 20603328, "step": 15210 }, { "epoch": 0.4870046731963383, "grad_norm": 0.6788691282272339, "learning_rate": 3.417754969280971e-05, "loss": 0.5099, "num_input_tokens_seen": 20610368, "step": 15215 }, { "epoch": 0.4871647141668267, "grad_norm": 2.837827444076538, "learning_rate": 3.416841695775137e-05, "loss": 0.633, "num_input_tokens_seen": 20616992, "step": 15220 }, { "epoch": 0.48732475513731516, "grad_norm": 1.9259226322174072, "learning_rate": 3.415928280880792e-05, "loss": 0.4628, "num_input_tokens_seen": 20623696, "step": 15225 }, { "epoch": 0.4874847961078036, "grad_norm": 1.5810127258300781, "learning_rate": 3.4150147247387965e-05, "loss": 0.5294, "num_input_tokens_seen": 20630224, "step": 15230 }, { "epoch": 0.48764483707829204, "grad_norm": 0.520296573638916, "learning_rate": 3.4141010274900306e-05, "loss": 0.3832, "num_input_tokens_seen": 20636944, "step": 15235 }, { "epoch": 0.4878048780487805, "grad_norm": 0.9293533563613892, "learning_rate": 3.413187189275399e-05, "loss": 0.4098, "num_input_tokens_seen": 20643632, "step": 15240 }, { "epoch": 0.4879649190192689, "grad_norm": 1.4843974113464355, "learning_rate": 3.4122732102358265e-05, "loss": 0.386, "num_input_tokens_seen": 20650352, "step": 15245 }, { "epoch": 0.48812495998975736, "grad_norm": 1.0251734256744385, "learning_rate": 3.411359090512261e-05, "loss": 0.6274, "num_input_tokens_seen": 20656864, "step": 15250 }, { "epoch": 0.4882850009602458, "grad_norm": 0.4214751124382019, "learning_rate": 3.410444830245672e-05, "loss": 0.4093, "num_input_tokens_seen": 20663584, "step": 15255 }, { "epoch": 0.4884450419307343, "grad_norm": 0.9342290759086609, "learning_rate": 3.409530429577048e-05, "loss": 0.5233, "num_input_tokens_seen": 20670256, "step": 15260 }, { "epoch": 0.48860508290122273, "grad_norm": 0.8321547508239746, "learning_rate": 3.408615888647402e-05, "loss": 0.4242, "num_input_tokens_seen": 20676800, "step": 15265 }, { "epoch": 0.48876512387171117, "grad_norm": 0.7290105819702148, "learning_rate": 3.4077012075977675e-05, "loss": 0.4975, "num_input_tokens_seen": 20683632, "step": 15270 }, { "epoch": 0.4889251648421996, "grad_norm": 1.5755035877227783, "learning_rate": 3.4067863865692e-05, "loss": 0.67, "num_input_tokens_seen": 20690448, "step": 15275 }, { "epoch": 0.48908520581268805, "grad_norm": 1.3513765335083008, "learning_rate": 3.4058714257027755e-05, "loss": 0.5606, "num_input_tokens_seen": 20697456, "step": 15280 }, { "epoch": 0.4892452467831765, "grad_norm": 0.7169976830482483, "learning_rate": 3.404956325139594e-05, "loss": 0.5064, "num_input_tokens_seen": 20704112, "step": 15285 }, { "epoch": 0.48940528775366493, "grad_norm": 1.1273927688598633, "learning_rate": 3.404041085020775e-05, "loss": 0.3852, "num_input_tokens_seen": 20710928, "step": 15290 }, { "epoch": 0.48956532872415337, "grad_norm": 1.3820854425430298, "learning_rate": 3.403125705487459e-05, "loss": 0.5817, "num_input_tokens_seen": 20717504, "step": 15295 }, { "epoch": 0.4897253696946418, "grad_norm": 1.4693703651428223, "learning_rate": 3.402210186680811e-05, "loss": 0.4199, "num_input_tokens_seen": 20724640, "step": 15300 }, { "epoch": 0.48988541066513025, "grad_norm": 1.215417504310608, "learning_rate": 3.4012945287420137e-05, "loss": 0.4123, "num_input_tokens_seen": 20731472, "step": 15305 }, { "epoch": 0.49004545163561875, "grad_norm": 0.7326523661613464, "learning_rate": 3.400378731812274e-05, "loss": 0.4783, "num_input_tokens_seen": 20739040, "step": 15310 }, { "epoch": 0.4902054926061072, "grad_norm": 1.3662270307540894, "learning_rate": 3.399462796032817e-05, "loss": 0.5591, "num_input_tokens_seen": 20745344, "step": 15315 }, { "epoch": 0.4903655335765956, "grad_norm": 0.8520097136497498, "learning_rate": 3.3985467215448954e-05, "loss": 0.535, "num_input_tokens_seen": 20751904, "step": 15320 }, { "epoch": 0.49052557454708406, "grad_norm": 1.2992100715637207, "learning_rate": 3.3976305084897776e-05, "loss": 0.3714, "num_input_tokens_seen": 20758256, "step": 15325 }, { "epoch": 0.4906856155175725, "grad_norm": 0.7989327907562256, "learning_rate": 3.3967141570087544e-05, "loss": 0.3314, "num_input_tokens_seen": 20765040, "step": 15330 }, { "epoch": 0.49084565648806094, "grad_norm": 1.2473801374435425, "learning_rate": 3.39579766724314e-05, "loss": 0.4684, "num_input_tokens_seen": 20771600, "step": 15335 }, { "epoch": 0.4910056974585494, "grad_norm": 1.8319666385650635, "learning_rate": 3.3948810393342677e-05, "loss": 0.55, "num_input_tokens_seen": 20778128, "step": 15340 }, { "epoch": 0.4911657384290378, "grad_norm": 1.097423791885376, "learning_rate": 3.3939642734234936e-05, "loss": 0.5254, "num_input_tokens_seen": 20785376, "step": 15345 }, { "epoch": 0.49132577939952626, "grad_norm": 0.4848121106624603, "learning_rate": 3.393047369652194e-05, "loss": 0.311, "num_input_tokens_seen": 20792496, "step": 15350 }, { "epoch": 0.4914858203700147, "grad_norm": 0.7616662979125977, "learning_rate": 3.3921303281617664e-05, "loss": 0.3701, "num_input_tokens_seen": 20799248, "step": 15355 }, { "epoch": 0.49164586134050314, "grad_norm": 1.7491310834884644, "learning_rate": 3.391213149093632e-05, "loss": 0.5065, "num_input_tokens_seen": 20806304, "step": 15360 }, { "epoch": 0.49180590231099164, "grad_norm": 1.690969467163086, "learning_rate": 3.3902958325892303e-05, "loss": 0.5502, "num_input_tokens_seen": 20813104, "step": 15365 }, { "epoch": 0.4919659432814801, "grad_norm": 2.5741679668426514, "learning_rate": 3.389378378790023e-05, "loss": 0.5572, "num_input_tokens_seen": 20820368, "step": 15370 }, { "epoch": 0.4921259842519685, "grad_norm": 1.0268610715866089, "learning_rate": 3.388460787837493e-05, "loss": 0.584, "num_input_tokens_seen": 20826944, "step": 15375 }, { "epoch": 0.49228602522245696, "grad_norm": 0.9988959431648254, "learning_rate": 3.387543059873145e-05, "loss": 0.4043, "num_input_tokens_seen": 20834000, "step": 15380 }, { "epoch": 0.4924460661929454, "grad_norm": 1.031823992729187, "learning_rate": 3.386625195038503e-05, "loss": 0.4389, "num_input_tokens_seen": 20841248, "step": 15385 }, { "epoch": 0.49260610716343384, "grad_norm": 1.6973538398742676, "learning_rate": 3.3857071934751136e-05, "loss": 0.4506, "num_input_tokens_seen": 20847744, "step": 15390 }, { "epoch": 0.4927661481339223, "grad_norm": 0.7339852452278137, "learning_rate": 3.384789055324544e-05, "loss": 0.4298, "num_input_tokens_seen": 20854192, "step": 15395 }, { "epoch": 0.4929261891044107, "grad_norm": 0.9939761757850647, "learning_rate": 3.3838707807283843e-05, "loss": 0.5278, "num_input_tokens_seen": 20860624, "step": 15400 }, { "epoch": 0.4929261891044107, "eval_loss": 0.4764499068260193, "eval_runtime": 500.0695, "eval_samples_per_second": 27.768, "eval_steps_per_second": 13.884, "num_input_tokens_seen": 20860624, "step": 15400 }, { "epoch": 0.49308623007489916, "grad_norm": 1.1926863193511963, "learning_rate": 3.382952369828243e-05, "loss": 0.3653, "num_input_tokens_seen": 20867536, "step": 15405 }, { "epoch": 0.4932462710453876, "grad_norm": 0.8735642433166504, "learning_rate": 3.38203382276575e-05, "loss": 0.4172, "num_input_tokens_seen": 20874000, "step": 15410 }, { "epoch": 0.4934063120158761, "grad_norm": 2.1028807163238525, "learning_rate": 3.381115139682557e-05, "loss": 0.3525, "num_input_tokens_seen": 20881280, "step": 15415 }, { "epoch": 0.49356635298636453, "grad_norm": 0.796198844909668, "learning_rate": 3.3801963207203366e-05, "loss": 0.5343, "num_input_tokens_seen": 20887936, "step": 15420 }, { "epoch": 0.49372639395685297, "grad_norm": 1.1151204109191895, "learning_rate": 3.379277366020782e-05, "loss": 0.7077, "num_input_tokens_seen": 20895312, "step": 15425 }, { "epoch": 0.4938864349273414, "grad_norm": 1.0404897928237915, "learning_rate": 3.3783582757256085e-05, "loss": 0.5406, "num_input_tokens_seen": 20902048, "step": 15430 }, { "epoch": 0.49404647589782985, "grad_norm": 1.2840877771377563, "learning_rate": 3.3774390499765504e-05, "loss": 0.4937, "num_input_tokens_seen": 20909072, "step": 15435 }, { "epoch": 0.4942065168683183, "grad_norm": 1.6717321872711182, "learning_rate": 3.376519688915364e-05, "loss": 0.4396, "num_input_tokens_seen": 20915520, "step": 15440 }, { "epoch": 0.49436655783880673, "grad_norm": 2.3797261714935303, "learning_rate": 3.3756001926838273e-05, "loss": 0.5184, "num_input_tokens_seen": 20922208, "step": 15445 }, { "epoch": 0.49452659880929517, "grad_norm": 1.0830192565917969, "learning_rate": 3.374680561423737e-05, "loss": 0.4492, "num_input_tokens_seen": 20928752, "step": 15450 }, { "epoch": 0.4946866397797836, "grad_norm": 1.1284582614898682, "learning_rate": 3.373760795276912e-05, "loss": 0.4435, "num_input_tokens_seen": 20935312, "step": 15455 }, { "epoch": 0.49484668075027205, "grad_norm": 1.1844452619552612, "learning_rate": 3.372840894385192e-05, "loss": 0.4315, "num_input_tokens_seen": 20942208, "step": 15460 }, { "epoch": 0.4950067217207605, "grad_norm": 1.0947697162628174, "learning_rate": 3.3719208588904375e-05, "loss": 0.6606, "num_input_tokens_seen": 20948592, "step": 15465 }, { "epoch": 0.495166762691249, "grad_norm": 0.6747429966926575, "learning_rate": 3.371000688934529e-05, "loss": 0.4432, "num_input_tokens_seen": 20955440, "step": 15470 }, { "epoch": 0.4953268036617374, "grad_norm": 1.0333305597305298, "learning_rate": 3.370080384659369e-05, "loss": 0.4518, "num_input_tokens_seen": 20962352, "step": 15475 }, { "epoch": 0.49548684463222586, "grad_norm": 1.8212271928787231, "learning_rate": 3.36915994620688e-05, "loss": 0.5677, "num_input_tokens_seen": 20969312, "step": 15480 }, { "epoch": 0.4956468856027143, "grad_norm": 0.5433996915817261, "learning_rate": 3.3682393737190035e-05, "loss": 0.4493, "num_input_tokens_seen": 20976128, "step": 15485 }, { "epoch": 0.49580692657320274, "grad_norm": 1.9545003175735474, "learning_rate": 3.3673186673377054e-05, "loss": 0.5027, "num_input_tokens_seen": 20982752, "step": 15490 }, { "epoch": 0.4959669675436912, "grad_norm": 0.9600472450256348, "learning_rate": 3.366397827204969e-05, "loss": 0.4994, "num_input_tokens_seen": 20989872, "step": 15495 }, { "epoch": 0.4961270085141796, "grad_norm": 0.7893664240837097, "learning_rate": 3.3654768534628e-05, "loss": 0.3924, "num_input_tokens_seen": 20996256, "step": 15500 }, { "epoch": 0.49628704948466806, "grad_norm": 0.7298755645751953, "learning_rate": 3.3645557462532245e-05, "loss": 0.3191, "num_input_tokens_seen": 21002560, "step": 15505 }, { "epoch": 0.4964470904551565, "grad_norm": 1.5158474445343018, "learning_rate": 3.363634505718288e-05, "loss": 0.398, "num_input_tokens_seen": 21009248, "step": 15510 }, { "epoch": 0.49660713142564494, "grad_norm": 1.1701042652130127, "learning_rate": 3.362713132000057e-05, "loss": 0.3781, "num_input_tokens_seen": 21016048, "step": 15515 }, { "epoch": 0.4967671723961334, "grad_norm": 0.6304591298103333, "learning_rate": 3.36179162524062e-05, "loss": 0.4931, "num_input_tokens_seen": 21023136, "step": 15520 }, { "epoch": 0.4969272133666219, "grad_norm": 1.438050627708435, "learning_rate": 3.3608699855820846e-05, "loss": 0.4018, "num_input_tokens_seen": 21030192, "step": 15525 }, { "epoch": 0.4970872543371103, "grad_norm": 0.7720056772232056, "learning_rate": 3.359948213166578e-05, "loss": 0.5177, "num_input_tokens_seen": 21037456, "step": 15530 }, { "epoch": 0.49724729530759876, "grad_norm": 2.3318703174591064, "learning_rate": 3.359026308136252e-05, "loss": 0.5802, "num_input_tokens_seen": 21044304, "step": 15535 }, { "epoch": 0.4974073362780872, "grad_norm": 1.3434076309204102, "learning_rate": 3.358104270633272e-05, "loss": 0.4511, "num_input_tokens_seen": 21051232, "step": 15540 }, { "epoch": 0.49756737724857564, "grad_norm": 1.3285702466964722, "learning_rate": 3.357182100799831e-05, "loss": 0.3936, "num_input_tokens_seen": 21057968, "step": 15545 }, { "epoch": 0.4977274182190641, "grad_norm": 0.44185590744018555, "learning_rate": 3.3562597987781384e-05, "loss": 0.4252, "num_input_tokens_seen": 21064624, "step": 15550 }, { "epoch": 0.4978874591895525, "grad_norm": 0.8283523321151733, "learning_rate": 3.355337364710424e-05, "loss": 0.4021, "num_input_tokens_seen": 21071680, "step": 15555 }, { "epoch": 0.49804750016004096, "grad_norm": 1.0108243227005005, "learning_rate": 3.354414798738939e-05, "loss": 0.4406, "num_input_tokens_seen": 21078672, "step": 15560 }, { "epoch": 0.4982075411305294, "grad_norm": 1.3912503719329834, "learning_rate": 3.353492101005955e-05, "loss": 0.4423, "num_input_tokens_seen": 21085328, "step": 15565 }, { "epoch": 0.49836758210101784, "grad_norm": 0.5895864367485046, "learning_rate": 3.352569271653763e-05, "loss": 0.4174, "num_input_tokens_seen": 21092272, "step": 15570 }, { "epoch": 0.49852762307150633, "grad_norm": 1.116746187210083, "learning_rate": 3.351646310824675e-05, "loss": 0.529, "num_input_tokens_seen": 21099008, "step": 15575 }, { "epoch": 0.49868766404199477, "grad_norm": 0.9733281135559082, "learning_rate": 3.350723218661023e-05, "loss": 0.4239, "num_input_tokens_seen": 21106080, "step": 15580 }, { "epoch": 0.4988477050124832, "grad_norm": 1.1268880367279053, "learning_rate": 3.349799995305162e-05, "loss": 0.5949, "num_input_tokens_seen": 21112928, "step": 15585 }, { "epoch": 0.49900774598297165, "grad_norm": 1.122602105140686, "learning_rate": 3.348876640899461e-05, "loss": 0.5514, "num_input_tokens_seen": 21119664, "step": 15590 }, { "epoch": 0.4991677869534601, "grad_norm": 0.8493016362190247, "learning_rate": 3.3479531555863144e-05, "loss": 0.3957, "num_input_tokens_seen": 21126576, "step": 15595 }, { "epoch": 0.49932782792394853, "grad_norm": 1.9113789796829224, "learning_rate": 3.3470295395081344e-05, "loss": 0.4903, "num_input_tokens_seen": 21133104, "step": 15600 }, { "epoch": 0.49932782792394853, "eval_loss": 0.47540467977523804, "eval_runtime": 499.9231, "eval_samples_per_second": 27.776, "eval_steps_per_second": 13.888, "num_input_tokens_seen": 21133104, "step": 15600 }, { "epoch": 0.49948786889443697, "grad_norm": 1.0215049982070923, "learning_rate": 3.3461057928073556e-05, "loss": 0.4848, "num_input_tokens_seen": 21140128, "step": 15605 }, { "epoch": 0.4996479098649254, "grad_norm": 1.4341892004013062, "learning_rate": 3.345181915626431e-05, "loss": 0.4466, "num_input_tokens_seen": 21147040, "step": 15610 }, { "epoch": 0.49980795083541385, "grad_norm": 1.8776317834854126, "learning_rate": 3.344257908107834e-05, "loss": 0.4691, "num_input_tokens_seen": 21153696, "step": 15615 }, { "epoch": 0.4999679918059023, "grad_norm": 0.5883113741874695, "learning_rate": 3.343333770394058e-05, "loss": 0.3828, "num_input_tokens_seen": 21160176, "step": 15620 }, { "epoch": 0.5001280327763907, "grad_norm": 0.9838226437568665, "learning_rate": 3.342409502627616e-05, "loss": 0.4181, "num_input_tokens_seen": 21166672, "step": 15625 }, { "epoch": 0.5002880737468792, "grad_norm": 1.5867677927017212, "learning_rate": 3.341485104951043e-05, "loss": 0.5117, "num_input_tokens_seen": 21173712, "step": 15630 }, { "epoch": 0.5004481147173676, "grad_norm": 0.6405961513519287, "learning_rate": 3.340560577506892e-05, "loss": 0.368, "num_input_tokens_seen": 21179984, "step": 15635 }, { "epoch": 0.5006081556878561, "grad_norm": 2.103783369064331, "learning_rate": 3.339635920437735e-05, "loss": 0.4851, "num_input_tokens_seen": 21186528, "step": 15640 }, { "epoch": 0.5007681966583445, "grad_norm": 1.7581442594528198, "learning_rate": 3.338711133886169e-05, "loss": 0.5222, "num_input_tokens_seen": 21193248, "step": 15645 }, { "epoch": 0.500928237628833, "grad_norm": 1.8324600458145142, "learning_rate": 3.3377862179948064e-05, "loss": 0.4651, "num_input_tokens_seen": 21199904, "step": 15650 }, { "epoch": 0.5010882785993215, "grad_norm": 0.9375972151756287, "learning_rate": 3.336861172906281e-05, "loss": 0.4296, "num_input_tokens_seen": 21206528, "step": 15655 }, { "epoch": 0.5012483195698099, "grad_norm": 0.5844959020614624, "learning_rate": 3.335935998763245e-05, "loss": 0.3342, "num_input_tokens_seen": 21213520, "step": 15660 }, { "epoch": 0.5014083605402984, "grad_norm": 0.5143135786056519, "learning_rate": 3.3350106957083744e-05, "loss": 0.4753, "num_input_tokens_seen": 21220384, "step": 15665 }, { "epoch": 0.5015684015107867, "grad_norm": 1.1347848176956177, "learning_rate": 3.33408526388436e-05, "loss": 0.5671, "num_input_tokens_seen": 21227136, "step": 15670 }, { "epoch": 0.5017284424812752, "grad_norm": 0.5970630049705505, "learning_rate": 3.3331597034339166e-05, "loss": 0.4478, "num_input_tokens_seen": 21233968, "step": 15675 }, { "epoch": 0.5018884834517636, "grad_norm": 1.3823813199996948, "learning_rate": 3.3322340144997764e-05, "loss": 0.4281, "num_input_tokens_seen": 21240560, "step": 15680 }, { "epoch": 0.5020485244222521, "grad_norm": 0.7659940123558044, "learning_rate": 3.331308197224693e-05, "loss": 0.4521, "num_input_tokens_seen": 21247888, "step": 15685 }, { "epoch": 0.5022085653927405, "grad_norm": 2.543308973312378, "learning_rate": 3.330382251751438e-05, "loss": 0.6436, "num_input_tokens_seen": 21254624, "step": 15690 }, { "epoch": 0.502368606363229, "grad_norm": 1.3097037076950073, "learning_rate": 3.3294561782228054e-05, "loss": 0.4203, "num_input_tokens_seen": 21261344, "step": 15695 }, { "epoch": 0.5025286473337174, "grad_norm": 1.204756736755371, "learning_rate": 3.328529976781607e-05, "loss": 0.3951, "num_input_tokens_seen": 21268560, "step": 15700 }, { "epoch": 0.5026886883042059, "grad_norm": 0.9107356667518616, "learning_rate": 3.327603647570673e-05, "loss": 0.4878, "num_input_tokens_seen": 21275408, "step": 15705 }, { "epoch": 0.5028487292746944, "grad_norm": 0.5673076510429382, "learning_rate": 3.326677190732857e-05, "loss": 0.3575, "num_input_tokens_seen": 21282192, "step": 15710 }, { "epoch": 0.5030087702451828, "grad_norm": 0.6507924795150757, "learning_rate": 3.325750606411029e-05, "loss": 0.4553, "num_input_tokens_seen": 21289360, "step": 15715 }, { "epoch": 0.5031688112156713, "grad_norm": 0.8991279602050781, "learning_rate": 3.3248238947480804e-05, "loss": 0.4329, "num_input_tokens_seen": 21296096, "step": 15720 }, { "epoch": 0.5033288521861596, "grad_norm": 1.13148033618927, "learning_rate": 3.323897055886922e-05, "loss": 0.4714, "num_input_tokens_seen": 21302944, "step": 15725 }, { "epoch": 0.5034888931566481, "grad_norm": 1.3351290225982666, "learning_rate": 3.322970089970484e-05, "loss": 0.5001, "num_input_tokens_seen": 21309584, "step": 15730 }, { "epoch": 0.5036489341271365, "grad_norm": 0.8026302456855774, "learning_rate": 3.3220429971417165e-05, "loss": 0.6295, "num_input_tokens_seen": 21316288, "step": 15735 }, { "epoch": 0.503808975097625, "grad_norm": 1.1859347820281982, "learning_rate": 3.321115777543588e-05, "loss": 0.3213, "num_input_tokens_seen": 21322736, "step": 15740 }, { "epoch": 0.5039690160681134, "grad_norm": 3.1086437702178955, "learning_rate": 3.320188431319088e-05, "loss": 0.6735, "num_input_tokens_seen": 21329936, "step": 15745 }, { "epoch": 0.5041290570386019, "grad_norm": 1.0176512002944946, "learning_rate": 3.319260958611224e-05, "loss": 0.4966, "num_input_tokens_seen": 21336288, "step": 15750 }, { "epoch": 0.5042890980090903, "grad_norm": 1.0969423055648804, "learning_rate": 3.3183333595630256e-05, "loss": 0.4015, "num_input_tokens_seen": 21342832, "step": 15755 }, { "epoch": 0.5044491389795788, "grad_norm": 1.8815343379974365, "learning_rate": 3.317405634317538e-05, "loss": 0.454, "num_input_tokens_seen": 21349632, "step": 15760 }, { "epoch": 0.5046091799500673, "grad_norm": 0.7637792825698853, "learning_rate": 3.3164777830178315e-05, "loss": 0.3793, "num_input_tokens_seen": 21355904, "step": 15765 }, { "epoch": 0.5047692209205557, "grad_norm": 0.8535454273223877, "learning_rate": 3.315549805806989e-05, "loss": 0.6212, "num_input_tokens_seen": 21362384, "step": 15770 }, { "epoch": 0.5049292618910441, "grad_norm": 1.0510270595550537, "learning_rate": 3.314621702828118e-05, "loss": 0.4019, "num_input_tokens_seen": 21369200, "step": 15775 }, { "epoch": 0.5050893028615325, "grad_norm": 1.4131830930709839, "learning_rate": 3.313693474224342e-05, "loss": 0.4598, "num_input_tokens_seen": 21376176, "step": 15780 }, { "epoch": 0.505249343832021, "grad_norm": 2.9776673316955566, "learning_rate": 3.312765120138809e-05, "loss": 0.4053, "num_input_tokens_seen": 21382976, "step": 15785 }, { "epoch": 0.5054093848025094, "grad_norm": 1.076229453086853, "learning_rate": 3.311836640714679e-05, "loss": 0.6086, "num_input_tokens_seen": 21389472, "step": 15790 }, { "epoch": 0.5055694257729979, "grad_norm": 0.9390943050384521, "learning_rate": 3.310908036095137e-05, "loss": 0.3382, "num_input_tokens_seen": 21396192, "step": 15795 }, { "epoch": 0.5057294667434863, "grad_norm": 1.975279450416565, "learning_rate": 3.309979306423386e-05, "loss": 0.5416, "num_input_tokens_seen": 21403072, "step": 15800 }, { "epoch": 0.5057294667434863, "eval_loss": 0.47564899921417236, "eval_runtime": 500.4254, "eval_samples_per_second": 27.748, "eval_steps_per_second": 13.874, "num_input_tokens_seen": 21403072, "step": 15800 }, { "epoch": 0.5058895077139748, "grad_norm": 0.5142216682434082, "learning_rate": 3.309050451842647e-05, "loss": 0.3937, "num_input_tokens_seen": 21410096, "step": 15805 }, { "epoch": 0.5060495486844632, "grad_norm": 1.3892488479614258, "learning_rate": 3.3081214724961604e-05, "loss": 0.4181, "num_input_tokens_seen": 21416768, "step": 15810 }, { "epoch": 0.5062095896549517, "grad_norm": 0.29830697178840637, "learning_rate": 3.307192368527188e-05, "loss": 0.2957, "num_input_tokens_seen": 21423856, "step": 15815 }, { "epoch": 0.5063696306254402, "grad_norm": 1.51951003074646, "learning_rate": 3.306263140079008e-05, "loss": 0.3007, "num_input_tokens_seen": 21430464, "step": 15820 }, { "epoch": 0.5065296715959285, "grad_norm": 0.9022490978240967, "learning_rate": 3.30533378729492e-05, "loss": 0.4227, "num_input_tokens_seen": 21437568, "step": 15825 }, { "epoch": 0.506689712566417, "grad_norm": 1.0275042057037354, "learning_rate": 3.304404310318242e-05, "loss": 0.4645, "num_input_tokens_seen": 21443888, "step": 15830 }, { "epoch": 0.5068497535369054, "grad_norm": 0.33434075117111206, "learning_rate": 3.3034747092923105e-05, "loss": 0.2814, "num_input_tokens_seen": 21450896, "step": 15835 }, { "epoch": 0.5070097945073939, "grad_norm": 1.271256685256958, "learning_rate": 3.3025449843604806e-05, "loss": 0.4932, "num_input_tokens_seen": 21457600, "step": 15840 }, { "epoch": 0.5071698354778823, "grad_norm": 1.6157004833221436, "learning_rate": 3.30161513566613e-05, "loss": 0.5561, "num_input_tokens_seen": 21464880, "step": 15845 }, { "epoch": 0.5073298764483708, "grad_norm": 1.4120070934295654, "learning_rate": 3.3006851633526506e-05, "loss": 0.4861, "num_input_tokens_seen": 21471744, "step": 15850 }, { "epoch": 0.5074899174188592, "grad_norm": 0.8811519145965576, "learning_rate": 3.2997550675634584e-05, "loss": 0.4755, "num_input_tokens_seen": 21478368, "step": 15855 }, { "epoch": 0.5076499583893477, "grad_norm": 0.9546794891357422, "learning_rate": 3.2988248484419825e-05, "loss": 0.3869, "num_input_tokens_seen": 21485136, "step": 15860 }, { "epoch": 0.5078099993598362, "grad_norm": 0.8140272498130798, "learning_rate": 3.2978945061316776e-05, "loss": 0.4948, "num_input_tokens_seen": 21491456, "step": 15865 }, { "epoch": 0.5079700403303246, "grad_norm": 0.9598685503005981, "learning_rate": 3.296964040776013e-05, "loss": 0.4233, "num_input_tokens_seen": 21497760, "step": 15870 }, { "epoch": 0.508130081300813, "grad_norm": 0.5828906893730164, "learning_rate": 3.296033452518478e-05, "loss": 0.4409, "num_input_tokens_seen": 21504240, "step": 15875 }, { "epoch": 0.5082901222713014, "grad_norm": 1.467200756072998, "learning_rate": 3.2951027415025806e-05, "loss": 0.4545, "num_input_tokens_seen": 21511024, "step": 15880 }, { "epoch": 0.5084501632417899, "grad_norm": 1.5770796537399292, "learning_rate": 3.294171907871849e-05, "loss": 0.39, "num_input_tokens_seen": 21517760, "step": 15885 }, { "epoch": 0.5086102042122783, "grad_norm": 0.8234081268310547, "learning_rate": 3.293240951769828e-05, "loss": 0.386, "num_input_tokens_seen": 21524576, "step": 15890 }, { "epoch": 0.5087702451827668, "grad_norm": 1.425205111503601, "learning_rate": 3.2923098733400846e-05, "loss": 0.4623, "num_input_tokens_seen": 21531984, "step": 15895 }, { "epoch": 0.5089302861532552, "grad_norm": 0.7212979793548584, "learning_rate": 3.291378672726202e-05, "loss": 0.3439, "num_input_tokens_seen": 21538576, "step": 15900 }, { "epoch": 0.5090903271237437, "grad_norm": 0.6673359274864197, "learning_rate": 3.2904473500717824e-05, "loss": 0.4843, "num_input_tokens_seen": 21545728, "step": 15905 }, { "epoch": 0.5092503680942321, "grad_norm": 1.3027461767196655, "learning_rate": 3.289515905520449e-05, "loss": 0.4242, "num_input_tokens_seen": 21552432, "step": 15910 }, { "epoch": 0.5094104090647206, "grad_norm": 3.3588876724243164, "learning_rate": 3.288584339215841e-05, "loss": 0.4744, "num_input_tokens_seen": 21559040, "step": 15915 }, { "epoch": 0.5095704500352091, "grad_norm": 0.7356337308883667, "learning_rate": 3.287652651301617e-05, "loss": 0.3557, "num_input_tokens_seen": 21565792, "step": 15920 }, { "epoch": 0.5097304910056975, "grad_norm": 1.135225772857666, "learning_rate": 3.286720841921457e-05, "loss": 0.4577, "num_input_tokens_seen": 21572400, "step": 15925 }, { "epoch": 0.509890531976186, "grad_norm": 1.3884891271591187, "learning_rate": 3.285788911219056e-05, "loss": 0.5278, "num_input_tokens_seen": 21579488, "step": 15930 }, { "epoch": 0.5100505729466743, "grad_norm": 1.3977618217468262, "learning_rate": 3.284856859338131e-05, "loss": 0.5206, "num_input_tokens_seen": 21586560, "step": 15935 }, { "epoch": 0.5102106139171628, "grad_norm": 0.884822428226471, "learning_rate": 3.283924686422414e-05, "loss": 0.4047, "num_input_tokens_seen": 21593376, "step": 15940 }, { "epoch": 0.5103706548876512, "grad_norm": 1.4627041816711426, "learning_rate": 3.282992392615659e-05, "loss": 0.4075, "num_input_tokens_seen": 21600352, "step": 15945 }, { "epoch": 0.5105306958581397, "grad_norm": 1.8483883142471313, "learning_rate": 3.282059978061638e-05, "loss": 0.549, "num_input_tokens_seen": 21607136, "step": 15950 }, { "epoch": 0.5106907368286281, "grad_norm": 0.6347278356552124, "learning_rate": 3.28112744290414e-05, "loss": 0.4338, "num_input_tokens_seen": 21614656, "step": 15955 }, { "epoch": 0.5108507777991166, "grad_norm": 1.480844497680664, "learning_rate": 3.280194787286974e-05, "loss": 0.4465, "num_input_tokens_seen": 21621744, "step": 15960 }, { "epoch": 0.511010818769605, "grad_norm": 1.3625954389572144, "learning_rate": 3.2792620113539674e-05, "loss": 0.4525, "num_input_tokens_seen": 21628224, "step": 15965 }, { "epoch": 0.5111708597400935, "grad_norm": 0.980945885181427, "learning_rate": 3.278329115248966e-05, "loss": 0.4302, "num_input_tokens_seen": 21635152, "step": 15970 }, { "epoch": 0.511330900710582, "grad_norm": 0.9938319325447083, "learning_rate": 3.277396099115834e-05, "loss": 0.4685, "num_input_tokens_seen": 21641952, "step": 15975 }, { "epoch": 0.5114909416810703, "grad_norm": 0.46764448285102844, "learning_rate": 3.276462963098454e-05, "loss": 0.4194, "num_input_tokens_seen": 21648736, "step": 15980 }, { "epoch": 0.5116509826515588, "grad_norm": 1.161483645439148, "learning_rate": 3.275529707340728e-05, "loss": 0.3513, "num_input_tokens_seen": 21655296, "step": 15985 }, { "epoch": 0.5118110236220472, "grad_norm": 1.0730863809585571, "learning_rate": 3.274596331986574e-05, "loss": 0.3438, "num_input_tokens_seen": 21661648, "step": 15990 }, { "epoch": 0.5119710645925357, "grad_norm": 1.7645937204360962, "learning_rate": 3.273662837179932e-05, "loss": 0.301, "num_input_tokens_seen": 21668688, "step": 15995 }, { "epoch": 0.5121311055630241, "grad_norm": 0.6950116157531738, "learning_rate": 3.272729223064758e-05, "loss": 0.3926, "num_input_tokens_seen": 21675712, "step": 16000 }, { "epoch": 0.5121311055630241, "eval_loss": 0.4760884642601013, "eval_runtime": 500.5067, "eval_samples_per_second": 27.744, "eval_steps_per_second": 13.872, "num_input_tokens_seen": 21675712, "step": 16000 }, { "epoch": 0.5122911465335126, "grad_norm": 1.6403142213821411, "learning_rate": 3.2717954897850264e-05, "loss": 0.6023, "num_input_tokens_seen": 21682272, "step": 16005 }, { "epoch": 0.512451187504001, "grad_norm": 0.8814744353294373, "learning_rate": 3.270861637484733e-05, "loss": 0.3557, "num_input_tokens_seen": 21689088, "step": 16010 }, { "epoch": 0.5126112284744895, "grad_norm": 0.5780954360961914, "learning_rate": 3.2699276663078867e-05, "loss": 0.188, "num_input_tokens_seen": 21696672, "step": 16015 }, { "epoch": 0.5127712694449779, "grad_norm": 0.788901686668396, "learning_rate": 3.268993576398519e-05, "loss": 0.5075, "num_input_tokens_seen": 21703504, "step": 16020 }, { "epoch": 0.5129313104154664, "grad_norm": 0.4769628942012787, "learning_rate": 3.268059367900678e-05, "loss": 0.4547, "num_input_tokens_seen": 21710048, "step": 16025 }, { "epoch": 0.5130913513859549, "grad_norm": 1.3757686614990234, "learning_rate": 3.26712504095843e-05, "loss": 0.4524, "num_input_tokens_seen": 21716432, "step": 16030 }, { "epoch": 0.5132513923564432, "grad_norm": 1.963362455368042, "learning_rate": 3.2661905957158615e-05, "loss": 0.525, "num_input_tokens_seen": 21722864, "step": 16035 }, { "epoch": 0.5134114333269317, "grad_norm": 4.193247318267822, "learning_rate": 3.2652560323170734e-05, "loss": 0.5856, "num_input_tokens_seen": 21729776, "step": 16040 }, { "epoch": 0.5135714742974201, "grad_norm": 0.6771774888038635, "learning_rate": 3.264321350906189e-05, "loss": 0.3756, "num_input_tokens_seen": 21736608, "step": 16045 }, { "epoch": 0.5137315152679086, "grad_norm": 0.9702341556549072, "learning_rate": 3.263386551627346e-05, "loss": 0.3721, "num_input_tokens_seen": 21743024, "step": 16050 }, { "epoch": 0.513891556238397, "grad_norm": 1.6442067623138428, "learning_rate": 3.2624516346247055e-05, "loss": 0.6721, "num_input_tokens_seen": 21749712, "step": 16055 }, { "epoch": 0.5140515972088855, "grad_norm": 1.1422524452209473, "learning_rate": 3.2615166000424404e-05, "loss": 0.3483, "num_input_tokens_seen": 21756848, "step": 16060 }, { "epoch": 0.5142116381793739, "grad_norm": 1.4027554988861084, "learning_rate": 3.260581448024745e-05, "loss": 0.497, "num_input_tokens_seen": 21763008, "step": 16065 }, { "epoch": 0.5143716791498624, "grad_norm": 0.533864438533783, "learning_rate": 3.2596461787158335e-05, "loss": 0.4587, "num_input_tokens_seen": 21769504, "step": 16070 }, { "epoch": 0.5145317201203508, "grad_norm": 1.086184024810791, "learning_rate": 3.258710792259934e-05, "loss": 0.3998, "num_input_tokens_seen": 21776624, "step": 16075 }, { "epoch": 0.5146917610908393, "grad_norm": 0.9015721082687378, "learning_rate": 3.257775288801296e-05, "loss": 0.485, "num_input_tokens_seen": 21783680, "step": 16080 }, { "epoch": 0.5148518020613277, "grad_norm": 0.7007731795310974, "learning_rate": 3.256839668484186e-05, "loss": 0.4288, "num_input_tokens_seen": 21790048, "step": 16085 }, { "epoch": 0.5150118430318161, "grad_norm": 1.0637321472167969, "learning_rate": 3.255903931452888e-05, "loss": 0.484, "num_input_tokens_seen": 21796624, "step": 16090 }, { "epoch": 0.5151718840023046, "grad_norm": 1.2318757772445679, "learning_rate": 3.2549680778517045e-05, "loss": 0.4106, "num_input_tokens_seen": 21803568, "step": 16095 }, { "epoch": 0.515331924972793, "grad_norm": 0.9341947436332703, "learning_rate": 3.2540321078249556e-05, "loss": 0.4273, "num_input_tokens_seen": 21810000, "step": 16100 }, { "epoch": 0.5154919659432815, "grad_norm": 0.8414938449859619, "learning_rate": 3.2530960215169795e-05, "loss": 0.4349, "num_input_tokens_seen": 21816560, "step": 16105 }, { "epoch": 0.5156520069137699, "grad_norm": 1.3913719654083252, "learning_rate": 3.2521598190721345e-05, "loss": 0.5263, "num_input_tokens_seen": 21823904, "step": 16110 }, { "epoch": 0.5158120478842584, "grad_norm": 0.9573763608932495, "learning_rate": 3.251223500634792e-05, "loss": 0.3844, "num_input_tokens_seen": 21830944, "step": 16115 }, { "epoch": 0.5159720888547468, "grad_norm": 1.1484527587890625, "learning_rate": 3.2502870663493445e-05, "loss": 0.4175, "num_input_tokens_seen": 21837792, "step": 16120 }, { "epoch": 0.5161321298252353, "grad_norm": 1.453602910041809, "learning_rate": 3.249350516360203e-05, "loss": 0.4147, "num_input_tokens_seen": 21844224, "step": 16125 }, { "epoch": 0.5162921707957238, "grad_norm": 0.4189009964466095, "learning_rate": 3.248413850811797e-05, "loss": 0.487, "num_input_tokens_seen": 21851024, "step": 16130 }, { "epoch": 0.5164522117662121, "grad_norm": 1.6050121784210205, "learning_rate": 3.2474770698485677e-05, "loss": 0.7138, "num_input_tokens_seen": 21857920, "step": 16135 }, { "epoch": 0.5166122527367006, "grad_norm": 0.9524714350700378, "learning_rate": 3.246540173614983e-05, "loss": 0.6209, "num_input_tokens_seen": 21864432, "step": 16140 }, { "epoch": 0.516772293707189, "grad_norm": 0.7223261594772339, "learning_rate": 3.2456031622555197e-05, "loss": 0.4569, "num_input_tokens_seen": 21871072, "step": 16145 }, { "epoch": 0.5169323346776775, "grad_norm": 1.5369328260421753, "learning_rate": 3.2446660359146794e-05, "loss": 0.4399, "num_input_tokens_seen": 21877680, "step": 16150 }, { "epoch": 0.5170923756481659, "grad_norm": 0.6187332272529602, "learning_rate": 3.2437287947369786e-05, "loss": 0.289, "num_input_tokens_seen": 21884304, "step": 16155 }, { "epoch": 0.5172524166186544, "grad_norm": 1.177997350692749, "learning_rate": 3.2427914388669525e-05, "loss": 0.3789, "num_input_tokens_seen": 21891440, "step": 16160 }, { "epoch": 0.5174124575891428, "grad_norm": 1.6176203489303589, "learning_rate": 3.241853968449151e-05, "loss": 0.5091, "num_input_tokens_seen": 21899056, "step": 16165 }, { "epoch": 0.5175724985596313, "grad_norm": 1.1696354150772095, "learning_rate": 3.240916383628144e-05, "loss": 0.2798, "num_input_tokens_seen": 21906192, "step": 16170 }, { "epoch": 0.5177325395301197, "grad_norm": 1.1266751289367676, "learning_rate": 3.239978684548521e-05, "loss": 0.41, "num_input_tokens_seen": 21912768, "step": 16175 }, { "epoch": 0.5178925805006082, "grad_norm": 1.338550090789795, "learning_rate": 3.239040871354885e-05, "loss": 0.5436, "num_input_tokens_seen": 21919088, "step": 16180 }, { "epoch": 0.5180526214710967, "grad_norm": 0.8327107429504395, "learning_rate": 3.2381029441918596e-05, "loss": 0.5736, "num_input_tokens_seen": 21926416, "step": 16185 }, { "epoch": 0.518212662441585, "grad_norm": 1.0950024127960205, "learning_rate": 3.2371649032040845e-05, "loss": 0.4558, "num_input_tokens_seen": 21933008, "step": 16190 }, { "epoch": 0.5183727034120735, "grad_norm": 0.8167585730552673, "learning_rate": 3.2362267485362174e-05, "loss": 0.4245, "num_input_tokens_seen": 21940000, "step": 16195 }, { "epoch": 0.5185327443825619, "grad_norm": 0.9349043965339661, "learning_rate": 3.235288480332934e-05, "loss": 0.4037, "num_input_tokens_seen": 21946528, "step": 16200 }, { "epoch": 0.5185327443825619, "eval_loss": 0.4749816656112671, "eval_runtime": 500.5045, "eval_samples_per_second": 27.744, "eval_steps_per_second": 13.872, "num_input_tokens_seen": 21946528, "step": 16200 }, { "epoch": 0.5186927853530504, "grad_norm": 1.213371992111206, "learning_rate": 3.234350098738927e-05, "loss": 0.4319, "num_input_tokens_seen": 21953568, "step": 16205 }, { "epoch": 0.5188528263235388, "grad_norm": 1.1117613315582275, "learning_rate": 3.233411603898906e-05, "loss": 0.2941, "num_input_tokens_seen": 21960160, "step": 16210 }, { "epoch": 0.5190128672940273, "grad_norm": 2.151010751724243, "learning_rate": 3.232472995957599e-05, "loss": 0.498, "num_input_tokens_seen": 21967104, "step": 16215 }, { "epoch": 0.5191729082645157, "grad_norm": 0.7887100577354431, "learning_rate": 3.231534275059751e-05, "loss": 0.3894, "num_input_tokens_seen": 21973968, "step": 16220 }, { "epoch": 0.5193329492350042, "grad_norm": 1.3512219190597534, "learning_rate": 3.230595441350125e-05, "loss": 0.451, "num_input_tokens_seen": 21980848, "step": 16225 }, { "epoch": 0.5194929902054926, "grad_norm": 0.8565742373466492, "learning_rate": 3.2296564949735e-05, "loss": 0.3726, "num_input_tokens_seen": 21987392, "step": 16230 }, { "epoch": 0.519653031175981, "grad_norm": 3.4894561767578125, "learning_rate": 3.228717436074675e-05, "loss": 0.5646, "num_input_tokens_seen": 21994528, "step": 16235 }, { "epoch": 0.5198130721464695, "grad_norm": 1.1757254600524902, "learning_rate": 3.227778264798463e-05, "loss": 0.5243, "num_input_tokens_seen": 22001280, "step": 16240 }, { "epoch": 0.5199731131169579, "grad_norm": 1.7620736360549927, "learning_rate": 3.226838981289698e-05, "loss": 0.4371, "num_input_tokens_seen": 22008128, "step": 16245 }, { "epoch": 0.5201331540874464, "grad_norm": 1.0812488794326782, "learning_rate": 3.225899585693227e-05, "loss": 0.5096, "num_input_tokens_seen": 22014656, "step": 16250 }, { "epoch": 0.5202931950579348, "grad_norm": 1.2202714681625366, "learning_rate": 3.224960078153918e-05, "loss": 0.5504, "num_input_tokens_seen": 22021440, "step": 16255 }, { "epoch": 0.5204532360284233, "grad_norm": 1.2173504829406738, "learning_rate": 3.224020458816655e-05, "loss": 0.4175, "num_input_tokens_seen": 22028176, "step": 16260 }, { "epoch": 0.5206132769989117, "grad_norm": 1.6090906858444214, "learning_rate": 3.223080727826337e-05, "loss": 0.4442, "num_input_tokens_seen": 22034560, "step": 16265 }, { "epoch": 0.5207733179694002, "grad_norm": 0.7172626256942749, "learning_rate": 3.222140885327885e-05, "loss": 0.5056, "num_input_tokens_seen": 22041424, "step": 16270 }, { "epoch": 0.5209333589398886, "grad_norm": 0.3413030803203583, "learning_rate": 3.221200931466234e-05, "loss": 0.5028, "num_input_tokens_seen": 22048528, "step": 16275 }, { "epoch": 0.5210933999103771, "grad_norm": 1.3838393688201904, "learning_rate": 3.220260866386336e-05, "loss": 0.4759, "num_input_tokens_seen": 22055344, "step": 16280 }, { "epoch": 0.5212534408808654, "grad_norm": 1.9181067943572998, "learning_rate": 3.21932069023316e-05, "loss": 0.4581, "num_input_tokens_seen": 22061968, "step": 16285 }, { "epoch": 0.5214134818513539, "grad_norm": 0.6750077605247498, "learning_rate": 3.218380403151695e-05, "loss": 0.3568, "num_input_tokens_seen": 22068496, "step": 16290 }, { "epoch": 0.5215735228218424, "grad_norm": 1.8963160514831543, "learning_rate": 3.217440005286943e-05, "loss": 0.4408, "num_input_tokens_seen": 22075536, "step": 16295 }, { "epoch": 0.5217335637923308, "grad_norm": 0.5985034108161926, "learning_rate": 3.216499496783928e-05, "loss": 0.4823, "num_input_tokens_seen": 22082592, "step": 16300 }, { "epoch": 0.5218936047628193, "grad_norm": 0.9656236171722412, "learning_rate": 3.2155588777876856e-05, "loss": 0.3612, "num_input_tokens_seen": 22089248, "step": 16305 }, { "epoch": 0.5220536457333077, "grad_norm": 1.0423214435577393, "learning_rate": 3.214618148443273e-05, "loss": 0.6836, "num_input_tokens_seen": 22095856, "step": 16310 }, { "epoch": 0.5222136867037962, "grad_norm": 3.30823016166687, "learning_rate": 3.2136773088957595e-05, "loss": 0.4896, "num_input_tokens_seen": 22102416, "step": 16315 }, { "epoch": 0.5223737276742846, "grad_norm": 1.3011152744293213, "learning_rate": 3.2127363592902374e-05, "loss": 0.5854, "num_input_tokens_seen": 22109280, "step": 16320 }, { "epoch": 0.5225337686447731, "grad_norm": 0.8331037759780884, "learning_rate": 3.211795299771812e-05, "loss": 0.5363, "num_input_tokens_seen": 22116784, "step": 16325 }, { "epoch": 0.5226938096152615, "grad_norm": 1.7971923351287842, "learning_rate": 3.210854130485605e-05, "loss": 0.4043, "num_input_tokens_seen": 22123232, "step": 16330 }, { "epoch": 0.52285385058575, "grad_norm": 1.7514352798461914, "learning_rate": 3.209912851576759e-05, "loss": 0.4497, "num_input_tokens_seen": 22130016, "step": 16335 }, { "epoch": 0.5230138915562383, "grad_norm": 1.2777384519577026, "learning_rate": 3.208971463190431e-05, "loss": 0.5033, "num_input_tokens_seen": 22136768, "step": 16340 }, { "epoch": 0.5231739325267268, "grad_norm": 1.7466914653778076, "learning_rate": 3.208029965471793e-05, "loss": 0.5169, "num_input_tokens_seen": 22142992, "step": 16345 }, { "epoch": 0.5233339734972153, "grad_norm": 2.766667127609253, "learning_rate": 3.2070883585660364e-05, "loss": 0.6949, "num_input_tokens_seen": 22149440, "step": 16350 }, { "epoch": 0.5234940144677037, "grad_norm": 0.5944870114326477, "learning_rate": 3.20614664261837e-05, "loss": 0.513, "num_input_tokens_seen": 22155872, "step": 16355 }, { "epoch": 0.5236540554381922, "grad_norm": 0.8773806691169739, "learning_rate": 3.205204817774016e-05, "loss": 0.5232, "num_input_tokens_seen": 22162752, "step": 16360 }, { "epoch": 0.5238140964086806, "grad_norm": 1.5803898572921753, "learning_rate": 3.204262884178218e-05, "loss": 0.3902, "num_input_tokens_seen": 22169184, "step": 16365 }, { "epoch": 0.5239741373791691, "grad_norm": 0.6795974373817444, "learning_rate": 3.2033208419762314e-05, "loss": 0.4893, "num_input_tokens_seen": 22175872, "step": 16370 }, { "epoch": 0.5241341783496575, "grad_norm": 1.0615888833999634, "learning_rate": 3.2023786913133344e-05, "loss": 0.5253, "num_input_tokens_seen": 22183216, "step": 16375 }, { "epoch": 0.524294219320146, "grad_norm": 1.3055377006530762, "learning_rate": 3.201436432334816e-05, "loss": 0.4582, "num_input_tokens_seen": 22190112, "step": 16380 }, { "epoch": 0.5244542602906344, "grad_norm": 1.339911699295044, "learning_rate": 3.2004940651859844e-05, "loss": 0.5462, "num_input_tokens_seen": 22197440, "step": 16385 }, { "epoch": 0.5246143012611229, "grad_norm": 0.9269368648529053, "learning_rate": 3.1995515900121655e-05, "loss": 0.3575, "num_input_tokens_seen": 22204160, "step": 16390 }, { "epoch": 0.5247743422316113, "grad_norm": 1.693435549736023, "learning_rate": 3.1986090069587e-05, "loss": 0.6502, "num_input_tokens_seen": 22210992, "step": 16395 }, { "epoch": 0.5249343832020997, "grad_norm": 2.004215717315674, "learning_rate": 3.1976663161709466e-05, "loss": 0.4913, "num_input_tokens_seen": 22217936, "step": 16400 }, { "epoch": 0.5249343832020997, "eval_loss": 0.4724019765853882, "eval_runtime": 499.8853, "eval_samples_per_second": 27.778, "eval_steps_per_second": 13.889, "num_input_tokens_seen": 22217936, "step": 16400 }, { "epoch": 0.5250944241725882, "grad_norm": 2.0785062313079834, "learning_rate": 3.196723517794279e-05, "loss": 0.5618, "num_input_tokens_seen": 22224864, "step": 16405 }, { "epoch": 0.5252544651430766, "grad_norm": 1.140337586402893, "learning_rate": 3.19578061197409e-05, "loss": 0.4, "num_input_tokens_seen": 22231200, "step": 16410 }, { "epoch": 0.5254145061135651, "grad_norm": 0.48438557982444763, "learning_rate": 3.194837598855787e-05, "loss": 0.4872, "num_input_tokens_seen": 22238064, "step": 16415 }, { "epoch": 0.5255745470840535, "grad_norm": 1.4118822813034058, "learning_rate": 3.193894478584794e-05, "loss": 0.4633, "num_input_tokens_seen": 22244768, "step": 16420 }, { "epoch": 0.525734588054542, "grad_norm": 0.9405772686004639, "learning_rate": 3.192951251306553e-05, "loss": 0.4111, "num_input_tokens_seen": 22251824, "step": 16425 }, { "epoch": 0.5258946290250304, "grad_norm": 1.2684836387634277, "learning_rate": 3.192007917166521e-05, "loss": 0.5535, "num_input_tokens_seen": 22258752, "step": 16430 }, { "epoch": 0.5260546699955189, "grad_norm": 1.2612377405166626, "learning_rate": 3.191064476310171e-05, "loss": 0.6647, "num_input_tokens_seen": 22265472, "step": 16435 }, { "epoch": 0.5262147109660072, "grad_norm": 0.8869642615318298, "learning_rate": 3.1901209288829944e-05, "loss": 0.2682, "num_input_tokens_seen": 22272192, "step": 16440 }, { "epoch": 0.5263747519364957, "grad_norm": 0.7128597497940063, "learning_rate": 3.1891772750304985e-05, "loss": 0.4605, "num_input_tokens_seen": 22278832, "step": 16445 }, { "epoch": 0.5265347929069842, "grad_norm": 1.1636236906051636, "learning_rate": 3.188233514898206e-05, "loss": 0.4181, "num_input_tokens_seen": 22286208, "step": 16450 }, { "epoch": 0.5266948338774726, "grad_norm": 1.2325164079666138, "learning_rate": 3.187289648631657e-05, "loss": 0.4491, "num_input_tokens_seen": 22292848, "step": 16455 }, { "epoch": 0.5268548748479611, "grad_norm": 1.9284791946411133, "learning_rate": 3.186345676376406e-05, "loss": 0.4141, "num_input_tokens_seen": 22299760, "step": 16460 }, { "epoch": 0.5270149158184495, "grad_norm": 0.7550022602081299, "learning_rate": 3.1854015982780275e-05, "loss": 0.4501, "num_input_tokens_seen": 22306464, "step": 16465 }, { "epoch": 0.527174956788938, "grad_norm": 1.5276771783828735, "learning_rate": 3.1844574144821084e-05, "loss": 0.4807, "num_input_tokens_seen": 22313328, "step": 16470 }, { "epoch": 0.5273349977594264, "grad_norm": 1.3695076704025269, "learning_rate": 3.1835131251342554e-05, "loss": 0.4303, "num_input_tokens_seen": 22320016, "step": 16475 }, { "epoch": 0.5274950387299149, "grad_norm": 0.642967164516449, "learning_rate": 3.182568730380089e-05, "loss": 0.4192, "num_input_tokens_seen": 22327088, "step": 16480 }, { "epoch": 0.5276550797004033, "grad_norm": 0.49298614263534546, "learning_rate": 3.181624230365245e-05, "loss": 0.4321, "num_input_tokens_seen": 22334224, "step": 16485 }, { "epoch": 0.5278151206708918, "grad_norm": 1.10055673122406, "learning_rate": 3.180679625235381e-05, "loss": 0.4747, "num_input_tokens_seen": 22340816, "step": 16490 }, { "epoch": 0.5279751616413801, "grad_norm": 0.7807582020759583, "learning_rate": 3.1797349151361646e-05, "loss": 0.4544, "num_input_tokens_seen": 22347440, "step": 16495 }, { "epoch": 0.5281352026118686, "grad_norm": 1.4509652853012085, "learning_rate": 3.178790100213281e-05, "loss": 0.4939, "num_input_tokens_seen": 22354064, "step": 16500 }, { "epoch": 0.5282952435823571, "grad_norm": 0.8174197673797607, "learning_rate": 3.1778451806124346e-05, "loss": 0.3931, "num_input_tokens_seen": 22360928, "step": 16505 }, { "epoch": 0.5284552845528455, "grad_norm": 1.2090401649475098, "learning_rate": 3.176900156479342e-05, "loss": 0.4027, "num_input_tokens_seen": 22367824, "step": 16510 }, { "epoch": 0.528615325523334, "grad_norm": 0.6965067982673645, "learning_rate": 3.17595502795974e-05, "loss": 0.3389, "num_input_tokens_seen": 22374896, "step": 16515 }, { "epoch": 0.5287753664938224, "grad_norm": 1.18043851852417, "learning_rate": 3.175009795199377e-05, "loss": 0.5661, "num_input_tokens_seen": 22381824, "step": 16520 }, { "epoch": 0.5289354074643109, "grad_norm": 0.7305789589881897, "learning_rate": 3.1740644583440224e-05, "loss": 0.4392, "num_input_tokens_seen": 22388272, "step": 16525 }, { "epoch": 0.5290954484347993, "grad_norm": 1.1816078424453735, "learning_rate": 3.173119017539457e-05, "loss": 0.5704, "num_input_tokens_seen": 22394448, "step": 16530 }, { "epoch": 0.5292554894052878, "grad_norm": 0.8863591551780701, "learning_rate": 3.172173472931479e-05, "loss": 0.5848, "num_input_tokens_seen": 22401424, "step": 16535 }, { "epoch": 0.5294155303757762, "grad_norm": 2.573823928833008, "learning_rate": 3.1712278246659055e-05, "loss": 0.5801, "num_input_tokens_seen": 22408528, "step": 16540 }, { "epoch": 0.5295755713462647, "grad_norm": 1.187854528427124, "learning_rate": 3.170282072888566e-05, "loss": 0.436, "num_input_tokens_seen": 22415536, "step": 16545 }, { "epoch": 0.529735612316753, "grad_norm": 1.911020278930664, "learning_rate": 3.169336217745307e-05, "loss": 0.4891, "num_input_tokens_seen": 22422240, "step": 16550 }, { "epoch": 0.5298956532872415, "grad_norm": 0.7812690734863281, "learning_rate": 3.1683902593819924e-05, "loss": 0.3263, "num_input_tokens_seen": 22429152, "step": 16555 }, { "epoch": 0.53005569425773, "grad_norm": 0.6993247866630554, "learning_rate": 3.1674441979445e-05, "loss": 0.3431, "num_input_tokens_seen": 22435904, "step": 16560 }, { "epoch": 0.5302157352282184, "grad_norm": 0.8987290263175964, "learning_rate": 3.166498033578725e-05, "loss": 0.4145, "num_input_tokens_seen": 22442976, "step": 16565 }, { "epoch": 0.5303757761987069, "grad_norm": 2.0029456615448, "learning_rate": 3.165551766430578e-05, "loss": 0.5746, "num_input_tokens_seen": 22449360, "step": 16570 }, { "epoch": 0.5305358171691953, "grad_norm": 0.46734410524368286, "learning_rate": 3.164605396645984e-05, "loss": 0.4054, "num_input_tokens_seen": 22455984, "step": 16575 }, { "epoch": 0.5306958581396838, "grad_norm": 1.169528603553772, "learning_rate": 3.163658924370886e-05, "loss": 0.5177, "num_input_tokens_seen": 22462304, "step": 16580 }, { "epoch": 0.5308558991101722, "grad_norm": 1.0726202726364136, "learning_rate": 3.1627123497512415e-05, "loss": 0.4555, "num_input_tokens_seen": 22469312, "step": 16585 }, { "epoch": 0.5310159400806607, "grad_norm": 2.151975154876709, "learning_rate": 3.1617656729330245e-05, "loss": 0.337, "num_input_tokens_seen": 22475776, "step": 16590 }, { "epoch": 0.531175981051149, "grad_norm": 0.8338291049003601, "learning_rate": 3.1608188940622255e-05, "loss": 0.6165, "num_input_tokens_seen": 22482512, "step": 16595 }, { "epoch": 0.5313360220216375, "grad_norm": 0.8200504183769226, "learning_rate": 3.159872013284847e-05, "loss": 0.4442, "num_input_tokens_seen": 22489168, "step": 16600 }, { "epoch": 0.5313360220216375, "eval_loss": 0.472185879945755, "eval_runtime": 503.3653, "eval_samples_per_second": 27.586, "eval_steps_per_second": 13.793, "num_input_tokens_seen": 22489168, "step": 16600 }, { "epoch": 0.531496062992126, "grad_norm": 1.738380789756775, "learning_rate": 3.1589250307469134e-05, "loss": 0.5828, "num_input_tokens_seen": 22495904, "step": 16605 }, { "epoch": 0.5316561039626144, "grad_norm": 0.828385591506958, "learning_rate": 3.1579779465944586e-05, "loss": 0.4734, "num_input_tokens_seen": 22502608, "step": 16610 }, { "epoch": 0.5318161449331029, "grad_norm": 1.526009440422058, "learning_rate": 3.1570307609735363e-05, "loss": 0.5316, "num_input_tokens_seen": 22509968, "step": 16615 }, { "epoch": 0.5319761859035913, "grad_norm": 0.7534289360046387, "learning_rate": 3.156083474030213e-05, "loss": 0.447, "num_input_tokens_seen": 22516624, "step": 16620 }, { "epoch": 0.5321362268740798, "grad_norm": 1.1236211061477661, "learning_rate": 3.155136085910573e-05, "loss": 0.7023, "num_input_tokens_seen": 22523328, "step": 16625 }, { "epoch": 0.5322962678445682, "grad_norm": 1.200156569480896, "learning_rate": 3.154188596760717e-05, "loss": 0.3926, "num_input_tokens_seen": 22530160, "step": 16630 }, { "epoch": 0.5324563088150567, "grad_norm": 0.6264258027076721, "learning_rate": 3.153241006726757e-05, "loss": 0.3458, "num_input_tokens_seen": 22536912, "step": 16635 }, { "epoch": 0.5326163497855451, "grad_norm": 0.7725962400436401, "learning_rate": 3.152293315954825e-05, "loss": 0.3274, "num_input_tokens_seen": 22543744, "step": 16640 }, { "epoch": 0.5327763907560336, "grad_norm": 1.9488739967346191, "learning_rate": 3.1513455245910666e-05, "loss": 0.6, "num_input_tokens_seen": 22550656, "step": 16645 }, { "epoch": 0.5329364317265219, "grad_norm": 1.2958955764770508, "learning_rate": 3.150397632781643e-05, "loss": 0.4217, "num_input_tokens_seen": 22557376, "step": 16650 }, { "epoch": 0.5330964726970104, "grad_norm": 3.3930060863494873, "learning_rate": 3.149449640672731e-05, "loss": 0.5347, "num_input_tokens_seen": 22564272, "step": 16655 }, { "epoch": 0.5332565136674989, "grad_norm": 1.4003616571426392, "learning_rate": 3.148501548410523e-05, "loss": 0.5288, "num_input_tokens_seen": 22570896, "step": 16660 }, { "epoch": 0.5334165546379873, "grad_norm": 1.243383765220642, "learning_rate": 3.1475533561412256e-05, "loss": 0.5353, "num_input_tokens_seen": 22577520, "step": 16665 }, { "epoch": 0.5335765956084758, "grad_norm": 1.190394401550293, "learning_rate": 3.146605064011065e-05, "loss": 0.2702, "num_input_tokens_seen": 22583952, "step": 16670 }, { "epoch": 0.5337366365789642, "grad_norm": 0.9655085206031799, "learning_rate": 3.145656672166277e-05, "loss": 0.3534, "num_input_tokens_seen": 22590576, "step": 16675 }, { "epoch": 0.5338966775494527, "grad_norm": 1.1032023429870605, "learning_rate": 3.144708180753116e-05, "loss": 0.5589, "num_input_tokens_seen": 22597440, "step": 16680 }, { "epoch": 0.5340567185199411, "grad_norm": 1.6474472284317017, "learning_rate": 3.143759589917851e-05, "loss": 0.5648, "num_input_tokens_seen": 22604096, "step": 16685 }, { "epoch": 0.5342167594904296, "grad_norm": 1.455150842666626, "learning_rate": 3.142810899806768e-05, "loss": 0.6602, "num_input_tokens_seen": 22610912, "step": 16690 }, { "epoch": 0.534376800460918, "grad_norm": 1.3691067695617676, "learning_rate": 3.141862110566166e-05, "loss": 0.5489, "num_input_tokens_seen": 22618032, "step": 16695 }, { "epoch": 0.5345368414314065, "grad_norm": 0.7650402784347534, "learning_rate": 3.1409132223423606e-05, "loss": 0.3112, "num_input_tokens_seen": 22624816, "step": 16700 }, { "epoch": 0.5346968824018948, "grad_norm": 0.9162204265594482, "learning_rate": 3.139964235281682e-05, "loss": 0.3438, "num_input_tokens_seen": 22631680, "step": 16705 }, { "epoch": 0.5348569233723833, "grad_norm": 0.676585853099823, "learning_rate": 3.139015149530476e-05, "loss": 0.4173, "num_input_tokens_seen": 22638208, "step": 16710 }, { "epoch": 0.5350169643428718, "grad_norm": 1.6216754913330078, "learning_rate": 3.1380659652351034e-05, "loss": 0.5132, "num_input_tokens_seen": 22644704, "step": 16715 }, { "epoch": 0.5351770053133602, "grad_norm": 1.477292537689209, "learning_rate": 3.137116682541941e-05, "loss": 0.4633, "num_input_tokens_seen": 22651568, "step": 16720 }, { "epoch": 0.5353370462838487, "grad_norm": 1.5562282800674438, "learning_rate": 3.136167301597379e-05, "loss": 0.3662, "num_input_tokens_seen": 22658576, "step": 16725 }, { "epoch": 0.5354970872543371, "grad_norm": 1.3426560163497925, "learning_rate": 3.1352178225478254e-05, "loss": 0.456, "num_input_tokens_seen": 22665152, "step": 16730 }, { "epoch": 0.5356571282248256, "grad_norm": 2.0576634407043457, "learning_rate": 3.1342682455396996e-05, "loss": 0.6192, "num_input_tokens_seen": 22671632, "step": 16735 }, { "epoch": 0.535817169195314, "grad_norm": 2.0354292392730713, "learning_rate": 3.133318570719441e-05, "loss": 0.4888, "num_input_tokens_seen": 22678720, "step": 16740 }, { "epoch": 0.5359772101658025, "grad_norm": 1.4772762060165405, "learning_rate": 3.132368798233499e-05, "loss": 0.332, "num_input_tokens_seen": 22684944, "step": 16745 }, { "epoch": 0.5361372511362908, "grad_norm": 0.43709391355514526, "learning_rate": 3.131418928228342e-05, "loss": 0.6186, "num_input_tokens_seen": 22691392, "step": 16750 }, { "epoch": 0.5362972921067793, "grad_norm": 0.518375813961029, "learning_rate": 3.1304689608504514e-05, "loss": 0.2477, "num_input_tokens_seen": 22697984, "step": 16755 }, { "epoch": 0.5364573330772677, "grad_norm": 1.8707096576690674, "learning_rate": 3.129518896246324e-05, "loss": 0.5286, "num_input_tokens_seen": 22704736, "step": 16760 }, { "epoch": 0.5366173740477562, "grad_norm": 0.4369646906852722, "learning_rate": 3.128568734562472e-05, "loss": 0.4827, "num_input_tokens_seen": 22711552, "step": 16765 }, { "epoch": 0.5367774150182447, "grad_norm": 1.232566475868225, "learning_rate": 3.127618475945421e-05, "loss": 0.5227, "num_input_tokens_seen": 22718240, "step": 16770 }, { "epoch": 0.5369374559887331, "grad_norm": 0.7940579652786255, "learning_rate": 3.126668120541715e-05, "loss": 0.3253, "num_input_tokens_seen": 22724944, "step": 16775 }, { "epoch": 0.5370974969592216, "grad_norm": 1.6275674104690552, "learning_rate": 3.1257176684979096e-05, "loss": 0.5272, "num_input_tokens_seen": 22732096, "step": 16780 }, { "epoch": 0.53725753792971, "grad_norm": 1.058823823928833, "learning_rate": 3.124767119960576e-05, "loss": 0.529, "num_input_tokens_seen": 22739040, "step": 16785 }, { "epoch": 0.5374175789001985, "grad_norm": 1.1432007551193237, "learning_rate": 3.123816475076301e-05, "loss": 0.4418, "num_input_tokens_seen": 22745744, "step": 16790 }, { "epoch": 0.5375776198706869, "grad_norm": 3.155411958694458, "learning_rate": 3.122865733991687e-05, "loss": 0.3456, "num_input_tokens_seen": 22752592, "step": 16795 }, { "epoch": 0.5377376608411754, "grad_norm": 1.5163155794143677, "learning_rate": 3.1219148968533486e-05, "loss": 0.3534, "num_input_tokens_seen": 22759200, "step": 16800 }, { "epoch": 0.5377376608411754, "eval_loss": 0.4719852805137634, "eval_runtime": 500.6695, "eval_samples_per_second": 27.735, "eval_steps_per_second": 13.867, "num_input_tokens_seen": 22759200, "step": 16800 }, { "epoch": 0.5378977018116637, "grad_norm": 1.324681043624878, "learning_rate": 3.120963963807918e-05, "loss": 0.4814, "num_input_tokens_seen": 22765728, "step": 16805 }, { "epoch": 0.5380577427821522, "grad_norm": 1.0173190832138062, "learning_rate": 3.12001293500204e-05, "loss": 0.3377, "num_input_tokens_seen": 22772240, "step": 16810 }, { "epoch": 0.5382177837526406, "grad_norm": 1.2956092357635498, "learning_rate": 3.1190618105823765e-05, "loss": 0.4123, "num_input_tokens_seen": 22778800, "step": 16815 }, { "epoch": 0.5383778247231291, "grad_norm": 1.1378813982009888, "learning_rate": 3.118110590695603e-05, "loss": 0.3925, "num_input_tokens_seen": 22785408, "step": 16820 }, { "epoch": 0.5385378656936176, "grad_norm": 0.9359673261642456, "learning_rate": 3.117159275488407e-05, "loss": 0.3843, "num_input_tokens_seen": 22792032, "step": 16825 }, { "epoch": 0.538697906664106, "grad_norm": 0.8550867438316345, "learning_rate": 3.1162078651074956e-05, "loss": 0.3207, "num_input_tokens_seen": 22798992, "step": 16830 }, { "epoch": 0.5388579476345945, "grad_norm": 1.0343866348266602, "learning_rate": 3.1152563596995885e-05, "loss": 0.5478, "num_input_tokens_seen": 22805904, "step": 16835 }, { "epoch": 0.5390179886050829, "grad_norm": 0.8515803813934326, "learning_rate": 3.1143047594114186e-05, "loss": 0.4511, "num_input_tokens_seen": 22812592, "step": 16840 }, { "epoch": 0.5391780295755714, "grad_norm": 2.5803561210632324, "learning_rate": 3.113353064389734e-05, "loss": 0.4394, "num_input_tokens_seen": 22819312, "step": 16845 }, { "epoch": 0.5393380705460598, "grad_norm": 2.057608127593994, "learning_rate": 3.1124012747812993e-05, "loss": 0.4433, "num_input_tokens_seen": 22825568, "step": 16850 }, { "epoch": 0.5394981115165483, "grad_norm": 0.6199699640274048, "learning_rate": 3.1114493907328936e-05, "loss": 0.4336, "num_input_tokens_seen": 22832480, "step": 16855 }, { "epoch": 0.5396581524870366, "grad_norm": 0.9614711403846741, "learning_rate": 3.110497412391306e-05, "loss": 0.5372, "num_input_tokens_seen": 22839616, "step": 16860 }, { "epoch": 0.5398181934575251, "grad_norm": 4.225128650665283, "learning_rate": 3.1095453399033466e-05, "loss": 0.6432, "num_input_tokens_seen": 22846096, "step": 16865 }, { "epoch": 0.5399782344280136, "grad_norm": 1.454642653465271, "learning_rate": 3.108593173415835e-05, "loss": 0.423, "num_input_tokens_seen": 22852912, "step": 16870 }, { "epoch": 0.540138275398502, "grad_norm": 1.3055955171585083, "learning_rate": 3.107640913075609e-05, "loss": 0.405, "num_input_tokens_seen": 22860416, "step": 16875 }, { "epoch": 0.5402983163689905, "grad_norm": 1.6328028440475464, "learning_rate": 3.106688559029517e-05, "loss": 0.3501, "num_input_tokens_seen": 22867392, "step": 16880 }, { "epoch": 0.5404583573394789, "grad_norm": 1.6892728805541992, "learning_rate": 3.105736111424425e-05, "loss": 0.6091, "num_input_tokens_seen": 22873872, "step": 16885 }, { "epoch": 0.5406183983099674, "grad_norm": 1.1549303531646729, "learning_rate": 3.1047835704072136e-05, "loss": 0.4639, "num_input_tokens_seen": 22881152, "step": 16890 }, { "epoch": 0.5407784392804558, "grad_norm": 0.9525120854377747, "learning_rate": 3.103830936124775e-05, "loss": 0.3555, "num_input_tokens_seen": 22887872, "step": 16895 }, { "epoch": 0.5409384802509443, "grad_norm": 1.4862827062606812, "learning_rate": 3.102878208724018e-05, "loss": 0.4851, "num_input_tokens_seen": 22894560, "step": 16900 }, { "epoch": 0.5410985212214326, "grad_norm": 0.7146314978599548, "learning_rate": 3.101925388351865e-05, "loss": 0.4039, "num_input_tokens_seen": 22900992, "step": 16905 }, { "epoch": 0.5412585621919211, "grad_norm": 0.732488751411438, "learning_rate": 3.1009724751552515e-05, "loss": 0.3708, "num_input_tokens_seen": 22907536, "step": 16910 }, { "epoch": 0.5414186031624095, "grad_norm": 0.7567753791809082, "learning_rate": 3.100019469281131e-05, "loss": 0.516, "num_input_tokens_seen": 22914288, "step": 16915 }, { "epoch": 0.541578644132898, "grad_norm": 1.2377433776855469, "learning_rate": 3.0990663708764685e-05, "loss": 0.4991, "num_input_tokens_seen": 22920928, "step": 16920 }, { "epoch": 0.5417386851033865, "grad_norm": 0.8736951947212219, "learning_rate": 3.098113180088243e-05, "loss": 0.3602, "num_input_tokens_seen": 22927456, "step": 16925 }, { "epoch": 0.5418987260738749, "grad_norm": 1.0059163570404053, "learning_rate": 3.097159897063448e-05, "loss": 0.3501, "num_input_tokens_seen": 22934096, "step": 16930 }, { "epoch": 0.5420587670443634, "grad_norm": 2.2788174152374268, "learning_rate": 3.096206521949094e-05, "loss": 0.6341, "num_input_tokens_seen": 22940704, "step": 16935 }, { "epoch": 0.5422188080148518, "grad_norm": 0.6172890067100525, "learning_rate": 3.0952530548922006e-05, "loss": 0.4358, "num_input_tokens_seen": 22947584, "step": 16940 }, { "epoch": 0.5423788489853403, "grad_norm": 1.3158622980117798, "learning_rate": 3.0942994960398064e-05, "loss": 0.299, "num_input_tokens_seen": 22954272, "step": 16945 }, { "epoch": 0.5425388899558287, "grad_norm": 0.9406221508979797, "learning_rate": 3.093345845538961e-05, "loss": 0.3942, "num_input_tokens_seen": 22961152, "step": 16950 }, { "epoch": 0.5426989309263172, "grad_norm": 0.6390989422798157, "learning_rate": 3.09239210353673e-05, "loss": 0.4494, "num_input_tokens_seen": 22967936, "step": 16955 }, { "epoch": 0.5428589718968055, "grad_norm": 1.6861305236816406, "learning_rate": 3.0914382701801926e-05, "loss": 0.3681, "num_input_tokens_seen": 22974848, "step": 16960 }, { "epoch": 0.543019012867294, "grad_norm": 1.0386093854904175, "learning_rate": 3.090484345616441e-05, "loss": 0.5534, "num_input_tokens_seen": 22981408, "step": 16965 }, { "epoch": 0.5431790538377824, "grad_norm": 0.6641054153442383, "learning_rate": 3.0895303299925825e-05, "loss": 0.3815, "num_input_tokens_seen": 22987968, "step": 16970 }, { "epoch": 0.5433390948082709, "grad_norm": 0.6045334935188293, "learning_rate": 3.0885762234557393e-05, "loss": 0.5095, "num_input_tokens_seen": 22994736, "step": 16975 }, { "epoch": 0.5434991357787594, "grad_norm": 1.397790551185608, "learning_rate": 3.087622026153045e-05, "loss": 0.4868, "num_input_tokens_seen": 23001664, "step": 16980 }, { "epoch": 0.5436591767492478, "grad_norm": 1.7094486951828003, "learning_rate": 3.086667738231651e-05, "loss": 0.6814, "num_input_tokens_seen": 23008480, "step": 16985 }, { "epoch": 0.5438192177197363, "grad_norm": 0.6398731470108032, "learning_rate": 3.085713359838718e-05, "loss": 0.3757, "num_input_tokens_seen": 23015168, "step": 16990 }, { "epoch": 0.5439792586902247, "grad_norm": 0.8815991282463074, "learning_rate": 3.084758891121425e-05, "loss": 0.3414, "num_input_tokens_seen": 23021552, "step": 16995 }, { "epoch": 0.5441392996607132, "grad_norm": 0.5231814384460449, "learning_rate": 3.083804332226963e-05, "loss": 0.4472, "num_input_tokens_seen": 23028128, "step": 17000 }, { "epoch": 0.5441392996607132, "eval_loss": 0.4709340035915375, "eval_runtime": 500.6312, "eval_samples_per_second": 27.737, "eval_steps_per_second": 13.868, "num_input_tokens_seen": 23028128, "step": 17000 }, { "epoch": 0.5442993406312016, "grad_norm": 1.8761488199234009, "learning_rate": 3.082849683302536e-05, "loss": 0.5179, "num_input_tokens_seen": 23034880, "step": 17005 }, { "epoch": 0.54445938160169, "grad_norm": 1.7704322338104248, "learning_rate": 3.081894944495363e-05, "loss": 0.5032, "num_input_tokens_seen": 23041856, "step": 17010 }, { "epoch": 0.5446194225721784, "grad_norm": 1.2778639793395996, "learning_rate": 3.080940115952677e-05, "loss": 0.5905, "num_input_tokens_seen": 23048848, "step": 17015 }, { "epoch": 0.5447794635426669, "grad_norm": 0.8917871713638306, "learning_rate": 3.0799851978217245e-05, "loss": 0.5813, "num_input_tokens_seen": 23055776, "step": 17020 }, { "epoch": 0.5449395045131553, "grad_norm": 0.7696419358253479, "learning_rate": 3.0790301902497666e-05, "loss": 0.4097, "num_input_tokens_seen": 23062976, "step": 17025 }, { "epoch": 0.5450995454836438, "grad_norm": 0.7319458723068237, "learning_rate": 3.078075093384076e-05, "loss": 0.5387, "num_input_tokens_seen": 23069920, "step": 17030 }, { "epoch": 0.5452595864541323, "grad_norm": 0.5631580948829651, "learning_rate": 3.077119907371942e-05, "loss": 0.3669, "num_input_tokens_seen": 23076640, "step": 17035 }, { "epoch": 0.5454196274246207, "grad_norm": 0.9407795667648315, "learning_rate": 3.076164632360666e-05, "loss": 0.5914, "num_input_tokens_seen": 23083600, "step": 17040 }, { "epoch": 0.5455796683951092, "grad_norm": 2.8735592365264893, "learning_rate": 3.075209268497563e-05, "loss": 0.335, "num_input_tokens_seen": 23090592, "step": 17045 }, { "epoch": 0.5457397093655976, "grad_norm": 1.5130388736724854, "learning_rate": 3.074253815929961e-05, "loss": 0.4364, "num_input_tokens_seen": 23097408, "step": 17050 }, { "epoch": 0.5458997503360861, "grad_norm": 1.1230281591415405, "learning_rate": 3.0732982748052054e-05, "loss": 0.6476, "num_input_tokens_seen": 23104176, "step": 17055 }, { "epoch": 0.5460597913065744, "grad_norm": 1.5298831462860107, "learning_rate": 3.072342645270651e-05, "loss": 0.7034, "num_input_tokens_seen": 23110880, "step": 17060 }, { "epoch": 0.5462198322770629, "grad_norm": 1.1416162252426147, "learning_rate": 3.071386927473668e-05, "loss": 0.3964, "num_input_tokens_seen": 23117728, "step": 17065 }, { "epoch": 0.5463798732475513, "grad_norm": 1.3927971124649048, "learning_rate": 3.0704311215616404e-05, "loss": 0.4809, "num_input_tokens_seen": 23124240, "step": 17070 }, { "epoch": 0.5465399142180398, "grad_norm": 0.9725024700164795, "learning_rate": 3.0694752276819656e-05, "loss": 0.3959, "num_input_tokens_seen": 23130560, "step": 17075 }, { "epoch": 0.5466999551885282, "grad_norm": 2.549879550933838, "learning_rate": 3.068519245982054e-05, "loss": 0.5515, "num_input_tokens_seen": 23137104, "step": 17080 }, { "epoch": 0.5468599961590167, "grad_norm": 1.0024152994155884, "learning_rate": 3.0675631766093304e-05, "loss": 0.4558, "num_input_tokens_seen": 23144144, "step": 17085 }, { "epoch": 0.5470200371295052, "grad_norm": 1.2265243530273438, "learning_rate": 3.066607019711232e-05, "loss": 0.389, "num_input_tokens_seen": 23150704, "step": 17090 }, { "epoch": 0.5471800780999936, "grad_norm": 2.261796712875366, "learning_rate": 3.065650775435211e-05, "loss": 0.3969, "num_input_tokens_seen": 23157312, "step": 17095 }, { "epoch": 0.5473401190704821, "grad_norm": 1.0674378871917725, "learning_rate": 3.0646944439287326e-05, "loss": 0.6106, "num_input_tokens_seen": 23163920, "step": 17100 }, { "epoch": 0.5475001600409705, "grad_norm": 1.1703532934188843, "learning_rate": 3.0637380253392736e-05, "loss": 0.3356, "num_input_tokens_seen": 23170480, "step": 17105 }, { "epoch": 0.547660201011459, "grad_norm": 0.8088896870613098, "learning_rate": 3.062781519814327e-05, "loss": 0.2779, "num_input_tokens_seen": 23176720, "step": 17110 }, { "epoch": 0.5478202419819473, "grad_norm": 1.0365177392959595, "learning_rate": 3.0618249275013985e-05, "loss": 0.371, "num_input_tokens_seen": 23183280, "step": 17115 }, { "epoch": 0.5479802829524358, "grad_norm": 1.1039057970046997, "learning_rate": 3.060868248548005e-05, "loss": 0.519, "num_input_tokens_seen": 23190128, "step": 17120 }, { "epoch": 0.5481403239229242, "grad_norm": 0.8647306561470032, "learning_rate": 3.0599114831016796e-05, "loss": 0.3942, "num_input_tokens_seen": 23196944, "step": 17125 }, { "epoch": 0.5483003648934127, "grad_norm": 1.8292896747589111, "learning_rate": 3.0589546313099666e-05, "loss": 0.587, "num_input_tokens_seen": 23203936, "step": 17130 }, { "epoch": 0.5484604058639012, "grad_norm": 0.7763909101486206, "learning_rate": 3.0579976933204255e-05, "loss": 0.4361, "num_input_tokens_seen": 23210400, "step": 17135 }, { "epoch": 0.5486204468343896, "grad_norm": 0.8977211117744446, "learning_rate": 3.0570406692806284e-05, "loss": 0.446, "num_input_tokens_seen": 23217568, "step": 17140 }, { "epoch": 0.5487804878048781, "grad_norm": 0.8155872225761414, "learning_rate": 3.05608355933816e-05, "loss": 0.4511, "num_input_tokens_seen": 23224592, "step": 17145 }, { "epoch": 0.5489405287753665, "grad_norm": 0.768828809261322, "learning_rate": 3.055126363640618e-05, "loss": 0.4602, "num_input_tokens_seen": 23231264, "step": 17150 }, { "epoch": 0.549100569745855, "grad_norm": 1.857652187347412, "learning_rate": 3.0541690823356146e-05, "loss": 0.443, "num_input_tokens_seen": 23238320, "step": 17155 }, { "epoch": 0.5492606107163434, "grad_norm": 0.9448212385177612, "learning_rate": 3.053211715570775e-05, "loss": 0.2568, "num_input_tokens_seen": 23245472, "step": 17160 }, { "epoch": 0.5494206516868319, "grad_norm": 0.8667546510696411, "learning_rate": 3.052254263493736e-05, "loss": 0.4496, "num_input_tokens_seen": 23252736, "step": 17165 }, { "epoch": 0.5495806926573202, "grad_norm": 1.3398067951202393, "learning_rate": 3.0512967262521498e-05, "loss": 0.3922, "num_input_tokens_seen": 23260016, "step": 17170 }, { "epoch": 0.5497407336278087, "grad_norm": 1.3408631086349487, "learning_rate": 3.0503391039936803e-05, "loss": 0.516, "num_input_tokens_seen": 23266592, "step": 17175 }, { "epoch": 0.5499007745982971, "grad_norm": 1.3815749883651733, "learning_rate": 3.0493813968660056e-05, "loss": 0.3543, "num_input_tokens_seen": 23273504, "step": 17180 }, { "epoch": 0.5500608155687856, "grad_norm": 1.5220706462860107, "learning_rate": 3.0484236050168153e-05, "loss": 0.6365, "num_input_tokens_seen": 23280480, "step": 17185 }, { "epoch": 0.5502208565392741, "grad_norm": 0.88559490442276, "learning_rate": 3.0474657285938123e-05, "loss": 0.5248, "num_input_tokens_seen": 23287328, "step": 17190 }, { "epoch": 0.5503808975097625, "grad_norm": 1.4071999788284302, "learning_rate": 3.046507767744715e-05, "loss": 0.4955, "num_input_tokens_seen": 23293888, "step": 17195 }, { "epoch": 0.550540938480251, "grad_norm": 1.3665951490402222, "learning_rate": 3.045549722617252e-05, "loss": 0.3981, "num_input_tokens_seen": 23300528, "step": 17200 }, { "epoch": 0.550540938480251, "eval_loss": 0.4705389738082886, "eval_runtime": 500.4303, "eval_samples_per_second": 27.748, "eval_steps_per_second": 13.874, "num_input_tokens_seen": 23300528, "step": 17200 }, { "epoch": 0.5507009794507394, "grad_norm": 1.4155055284500122, "learning_rate": 3.0445915933591658e-05, "loss": 0.4618, "num_input_tokens_seen": 23307360, "step": 17205 }, { "epoch": 0.5508610204212279, "grad_norm": 1.2358405590057373, "learning_rate": 3.0436333801182114e-05, "loss": 0.4228, "num_input_tokens_seen": 23314272, "step": 17210 }, { "epoch": 0.5510210613917162, "grad_norm": 2.386981725692749, "learning_rate": 3.0426750830421596e-05, "loss": 0.4132, "num_input_tokens_seen": 23320848, "step": 17215 }, { "epoch": 0.5511811023622047, "grad_norm": 0.8809069991111755, "learning_rate": 3.0417167022787897e-05, "loss": 0.3215, "num_input_tokens_seen": 23327744, "step": 17220 }, { "epoch": 0.5513411433326931, "grad_norm": 0.8140941858291626, "learning_rate": 3.0407582379758966e-05, "loss": 0.3801, "num_input_tokens_seen": 23334320, "step": 17225 }, { "epoch": 0.5515011843031816, "grad_norm": 1.9330437183380127, "learning_rate": 3.039799690281287e-05, "loss": 0.5427, "num_input_tokens_seen": 23340720, "step": 17230 }, { "epoch": 0.55166122527367, "grad_norm": 1.172855019569397, "learning_rate": 3.0388410593427823e-05, "loss": 0.8807, "num_input_tokens_seen": 23347360, "step": 17235 }, { "epoch": 0.5518212662441585, "grad_norm": 0.7078018188476562, "learning_rate": 3.0378823453082146e-05, "loss": 0.5395, "num_input_tokens_seen": 23353952, "step": 17240 }, { "epoch": 0.551981307214647, "grad_norm": 0.9479028582572937, "learning_rate": 3.03692354832543e-05, "loss": 0.3061, "num_input_tokens_seen": 23360480, "step": 17245 }, { "epoch": 0.5521413481851354, "grad_norm": 2.4552407264709473, "learning_rate": 3.0359646685422865e-05, "loss": 0.4499, "num_input_tokens_seen": 23367120, "step": 17250 }, { "epoch": 0.5523013891556239, "grad_norm": 0.6049360036849976, "learning_rate": 3.035005706106656e-05, "loss": 0.4312, "num_input_tokens_seen": 23373856, "step": 17255 }, { "epoch": 0.5524614301261123, "grad_norm": 0.6374049782752991, "learning_rate": 3.034046661166422e-05, "loss": 0.6513, "num_input_tokens_seen": 23380224, "step": 17260 }, { "epoch": 0.5526214710966008, "grad_norm": 0.46854016184806824, "learning_rate": 3.033087533869482e-05, "loss": 0.2643, "num_input_tokens_seen": 23386848, "step": 17265 }, { "epoch": 0.5527815120670891, "grad_norm": 1.316092848777771, "learning_rate": 3.0321283243637444e-05, "loss": 0.4115, "num_input_tokens_seen": 23393680, "step": 17270 }, { "epoch": 0.5529415530375776, "grad_norm": 2.454890489578247, "learning_rate": 3.0311690327971326e-05, "loss": 0.5854, "num_input_tokens_seen": 23400448, "step": 17275 }, { "epoch": 0.553101594008066, "grad_norm": 1.5543898344039917, "learning_rate": 3.030209659317581e-05, "loss": 0.4688, "num_input_tokens_seen": 23407248, "step": 17280 }, { "epoch": 0.5532616349785545, "grad_norm": 1.2922133207321167, "learning_rate": 3.0292502040730362e-05, "loss": 0.5112, "num_input_tokens_seen": 23413824, "step": 17285 }, { "epoch": 0.5534216759490429, "grad_norm": 0.6186715960502625, "learning_rate": 3.0282906672114597e-05, "loss": 0.4213, "num_input_tokens_seen": 23420624, "step": 17290 }, { "epoch": 0.5535817169195314, "grad_norm": 1.0705941915512085, "learning_rate": 3.027331048880823e-05, "loss": 0.5055, "num_input_tokens_seen": 23427440, "step": 17295 }, { "epoch": 0.5537417578900199, "grad_norm": 0.9691600203514099, "learning_rate": 3.0263713492291123e-05, "loss": 0.3278, "num_input_tokens_seen": 23434672, "step": 17300 }, { "epoch": 0.5539017988605083, "grad_norm": 1.6861310005187988, "learning_rate": 3.0254115684043242e-05, "loss": 0.5223, "num_input_tokens_seen": 23441472, "step": 17305 }, { "epoch": 0.5540618398309968, "grad_norm": 1.5157634019851685, "learning_rate": 3.024451706554469e-05, "loss": 0.5177, "num_input_tokens_seen": 23448352, "step": 17310 }, { "epoch": 0.5542218808014852, "grad_norm": 1.3144575357437134, "learning_rate": 3.0234917638275705e-05, "loss": 0.517, "num_input_tokens_seen": 23454992, "step": 17315 }, { "epoch": 0.5543819217719737, "grad_norm": 0.9761263728141785, "learning_rate": 3.0225317403716635e-05, "loss": 0.4705, "num_input_tokens_seen": 23461632, "step": 17320 }, { "epoch": 0.554541962742462, "grad_norm": 1.0696152448654175, "learning_rate": 3.0215716363347956e-05, "loss": 0.5136, "num_input_tokens_seen": 23468272, "step": 17325 }, { "epoch": 0.5547020037129505, "grad_norm": 2.4588987827301025, "learning_rate": 3.0206114518650275e-05, "loss": 0.4311, "num_input_tokens_seen": 23474832, "step": 17330 }, { "epoch": 0.5548620446834389, "grad_norm": 0.7631997466087341, "learning_rate": 3.0196511871104304e-05, "loss": 0.6125, "num_input_tokens_seen": 23481648, "step": 17335 }, { "epoch": 0.5550220856539274, "grad_norm": 0.7313428521156311, "learning_rate": 3.01869084221909e-05, "loss": 0.6288, "num_input_tokens_seen": 23488160, "step": 17340 }, { "epoch": 0.5551821266244159, "grad_norm": 0.5594280362129211, "learning_rate": 3.0177304173391037e-05, "loss": 0.4549, "num_input_tokens_seen": 23495280, "step": 17345 }, { "epoch": 0.5553421675949043, "grad_norm": 2.658188581466675, "learning_rate": 3.01676991261858e-05, "loss": 0.5328, "num_input_tokens_seen": 23501776, "step": 17350 }, { "epoch": 0.5555022085653928, "grad_norm": 1.6003661155700684, "learning_rate": 3.015809328205642e-05, "loss": 0.4455, "num_input_tokens_seen": 23508400, "step": 17355 }, { "epoch": 0.5556622495358812, "grad_norm": 0.6276563405990601, "learning_rate": 3.0148486642484248e-05, "loss": 0.4097, "num_input_tokens_seen": 23515312, "step": 17360 }, { "epoch": 0.5558222905063697, "grad_norm": 0.47599780559539795, "learning_rate": 3.0138879208950722e-05, "loss": 0.2676, "num_input_tokens_seen": 23521712, "step": 17365 }, { "epoch": 0.555982331476858, "grad_norm": 0.8649271130561829, "learning_rate": 3.012927098293744e-05, "loss": 0.5594, "num_input_tokens_seen": 23528592, "step": 17370 }, { "epoch": 0.5561423724473465, "grad_norm": 1.3225241899490356, "learning_rate": 3.0119661965926123e-05, "loss": 0.7301, "num_input_tokens_seen": 23535744, "step": 17375 }, { "epoch": 0.5563024134178349, "grad_norm": 0.7326554656028748, "learning_rate": 3.0110052159398587e-05, "loss": 0.4676, "num_input_tokens_seen": 23542400, "step": 17380 }, { "epoch": 0.5564624543883234, "grad_norm": 1.3685307502746582, "learning_rate": 3.0100441564836802e-05, "loss": 0.5136, "num_input_tokens_seen": 23548912, "step": 17385 }, { "epoch": 0.5566224953588118, "grad_norm": 0.8527954816818237, "learning_rate": 3.0090830183722817e-05, "loss": 0.3883, "num_input_tokens_seen": 23555888, "step": 17390 }, { "epoch": 0.5567825363293003, "grad_norm": 0.9303795695304871, "learning_rate": 3.0081218017538852e-05, "loss": 0.3771, "num_input_tokens_seen": 23562736, "step": 17395 }, { "epoch": 0.5569425772997888, "grad_norm": 1.4532116651535034, "learning_rate": 3.0071605067767212e-05, "loss": 0.4462, "num_input_tokens_seen": 23569728, "step": 17400 }, { "epoch": 0.5569425772997888, "eval_loss": 0.4684276282787323, "eval_runtime": 500.8448, "eval_samples_per_second": 27.725, "eval_steps_per_second": 13.863, "num_input_tokens_seen": 23569728, "step": 17400 }, { "epoch": 0.5571026182702772, "grad_norm": 1.0846753120422363, "learning_rate": 3.006199133589034e-05, "loss": 0.3683, "num_input_tokens_seen": 23576656, "step": 17405 }, { "epoch": 0.5572626592407657, "grad_norm": 1.8030296564102173, "learning_rate": 3.005237682339079e-05, "loss": 0.5697, "num_input_tokens_seen": 23583280, "step": 17410 }, { "epoch": 0.5574227002112541, "grad_norm": 2.4795384407043457, "learning_rate": 3.0042761531751228e-05, "loss": 0.4224, "num_input_tokens_seen": 23589952, "step": 17415 }, { "epoch": 0.5575827411817426, "grad_norm": 2.1323983669281006, "learning_rate": 3.0033145462454482e-05, "loss": 0.4653, "num_input_tokens_seen": 23596608, "step": 17420 }, { "epoch": 0.5577427821522309, "grad_norm": 0.9479411840438843, "learning_rate": 3.002352861698345e-05, "loss": 0.4648, "num_input_tokens_seen": 23603120, "step": 17425 }, { "epoch": 0.5579028231227194, "grad_norm": 1.1161293983459473, "learning_rate": 3.0013910996821178e-05, "loss": 0.416, "num_input_tokens_seen": 23609712, "step": 17430 }, { "epoch": 0.5580628640932078, "grad_norm": 1.6247797012329102, "learning_rate": 3.0004292603450817e-05, "loss": 0.367, "num_input_tokens_seen": 23616352, "step": 17435 }, { "epoch": 0.5582229050636963, "grad_norm": 1.0802717208862305, "learning_rate": 2.9994673438355653e-05, "loss": 0.5287, "num_input_tokens_seen": 23623200, "step": 17440 }, { "epoch": 0.5583829460341847, "grad_norm": 0.966475784778595, "learning_rate": 2.9985053503019078e-05, "loss": 0.6893, "num_input_tokens_seen": 23629568, "step": 17445 }, { "epoch": 0.5585429870046732, "grad_norm": 0.8473954796791077, "learning_rate": 2.99754327989246e-05, "loss": 0.4554, "num_input_tokens_seen": 23636400, "step": 17450 }, { "epoch": 0.5587030279751617, "grad_norm": 0.999119222164154, "learning_rate": 2.9965811327555864e-05, "loss": 0.3484, "num_input_tokens_seen": 23642928, "step": 17455 }, { "epoch": 0.5588630689456501, "grad_norm": 1.8298819065093994, "learning_rate": 2.995618909039662e-05, "loss": 0.4814, "num_input_tokens_seen": 23649744, "step": 17460 }, { "epoch": 0.5590231099161386, "grad_norm": 1.347137451171875, "learning_rate": 2.9946566088930727e-05, "loss": 0.4026, "num_input_tokens_seen": 23656416, "step": 17465 }, { "epoch": 0.559183150886627, "grad_norm": 0.6441223621368408, "learning_rate": 2.9936942324642192e-05, "loss": 0.4295, "num_input_tokens_seen": 23662736, "step": 17470 }, { "epoch": 0.5593431918571155, "grad_norm": 1.7260538339614868, "learning_rate": 2.9927317799015097e-05, "loss": 0.4088, "num_input_tokens_seen": 23669600, "step": 17475 }, { "epoch": 0.5595032328276038, "grad_norm": 1.647323727607727, "learning_rate": 2.9917692513533685e-05, "loss": 0.5666, "num_input_tokens_seen": 23676400, "step": 17480 }, { "epoch": 0.5596632737980923, "grad_norm": 1.9676440954208374, "learning_rate": 2.990806646968229e-05, "loss": 0.7412, "num_input_tokens_seen": 23683168, "step": 17485 }, { "epoch": 0.5598233147685807, "grad_norm": 1.2849279642105103, "learning_rate": 2.989843966894536e-05, "loss": 0.6332, "num_input_tokens_seen": 23689984, "step": 17490 }, { "epoch": 0.5599833557390692, "grad_norm": 0.7257857918739319, "learning_rate": 2.9888812112807472e-05, "loss": 0.4097, "num_input_tokens_seen": 23697200, "step": 17495 }, { "epoch": 0.5601433967095576, "grad_norm": 1.5271313190460205, "learning_rate": 2.987918380275333e-05, "loss": 0.5535, "num_input_tokens_seen": 23703952, "step": 17500 }, { "epoch": 0.5603034376800461, "grad_norm": 1.1042759418487549, "learning_rate": 2.9869554740267724e-05, "loss": 0.5916, "num_input_tokens_seen": 23711136, "step": 17505 }, { "epoch": 0.5604634786505346, "grad_norm": 1.993161678314209, "learning_rate": 2.9859924926835585e-05, "loss": 0.3529, "num_input_tokens_seen": 23718304, "step": 17510 }, { "epoch": 0.560623519621023, "grad_norm": 2.2873806953430176, "learning_rate": 2.9850294363941944e-05, "loss": 0.4577, "num_input_tokens_seen": 23725216, "step": 17515 }, { "epoch": 0.5607835605915115, "grad_norm": 1.2414973974227905, "learning_rate": 2.9840663053071967e-05, "loss": 0.4908, "num_input_tokens_seen": 23731696, "step": 17520 }, { "epoch": 0.5609436015619998, "grad_norm": 1.1095921993255615, "learning_rate": 2.983103099571091e-05, "loss": 0.3489, "num_input_tokens_seen": 23738544, "step": 17525 }, { "epoch": 0.5611036425324883, "grad_norm": 1.923628568649292, "learning_rate": 2.9821398193344164e-05, "loss": 0.4259, "num_input_tokens_seen": 23745840, "step": 17530 }, { "epoch": 0.5612636835029767, "grad_norm": 1.6104085445404053, "learning_rate": 2.9811764647457226e-05, "loss": 0.6361, "num_input_tokens_seen": 23752400, "step": 17535 }, { "epoch": 0.5614237244734652, "grad_norm": 2.6314609050750732, "learning_rate": 2.9802130359535714e-05, "loss": 0.7839, "num_input_tokens_seen": 23759040, "step": 17540 }, { "epoch": 0.5615837654439536, "grad_norm": 0.44926148653030396, "learning_rate": 2.979249533106535e-05, "loss": 0.2667, "num_input_tokens_seen": 23765872, "step": 17545 }, { "epoch": 0.5617438064144421, "grad_norm": 1.7611095905303955, "learning_rate": 2.9782859563531986e-05, "loss": 0.365, "num_input_tokens_seen": 23772336, "step": 17550 }, { "epoch": 0.5619038473849305, "grad_norm": 0.5593433380126953, "learning_rate": 2.977322305842156e-05, "loss": 0.3774, "num_input_tokens_seen": 23778864, "step": 17555 }, { "epoch": 0.562063888355419, "grad_norm": 1.9680862426757812, "learning_rate": 2.9763585817220162e-05, "loss": 0.6151, "num_input_tokens_seen": 23785136, "step": 17560 }, { "epoch": 0.5622239293259075, "grad_norm": 0.7440273761749268, "learning_rate": 2.975394784141397e-05, "loss": 0.3196, "num_input_tokens_seen": 23791664, "step": 17565 }, { "epoch": 0.5623839702963959, "grad_norm": 1.0687668323516846, "learning_rate": 2.974430913248928e-05, "loss": 0.4315, "num_input_tokens_seen": 23797904, "step": 17570 }, { "epoch": 0.5625440112668844, "grad_norm": 1.3317886590957642, "learning_rate": 2.9734669691932497e-05, "loss": 0.5026, "num_input_tokens_seen": 23804560, "step": 17575 }, { "epoch": 0.5627040522373727, "grad_norm": 1.3399075269699097, "learning_rate": 2.9725029521230147e-05, "loss": 0.6855, "num_input_tokens_seen": 23811456, "step": 17580 }, { "epoch": 0.5628640932078612, "grad_norm": 1.4389522075653076, "learning_rate": 2.9715388621868873e-05, "loss": 0.5781, "num_input_tokens_seen": 23818256, "step": 17585 }, { "epoch": 0.5630241341783496, "grad_norm": 1.6458529233932495, "learning_rate": 2.970574699533541e-05, "loss": 0.3832, "num_input_tokens_seen": 23825120, "step": 17590 }, { "epoch": 0.5631841751488381, "grad_norm": 0.653683602809906, "learning_rate": 2.969610464311662e-05, "loss": 0.3141, "num_input_tokens_seen": 23831984, "step": 17595 }, { "epoch": 0.5633442161193265, "grad_norm": 1.6110225915908813, "learning_rate": 2.9686461566699487e-05, "loss": 0.4062, "num_input_tokens_seen": 23838464, "step": 17600 }, { "epoch": 0.5633442161193265, "eval_loss": 0.4680066406726837, "eval_runtime": 501.0002, "eval_samples_per_second": 27.717, "eval_steps_per_second": 13.858, "num_input_tokens_seen": 23838464, "step": 17600 }, { "epoch": 0.563504257089815, "grad_norm": 0.6246126890182495, "learning_rate": 2.9676817767571086e-05, "loss": 0.5041, "num_input_tokens_seen": 23845088, "step": 17605 }, { "epoch": 0.5636642980603035, "grad_norm": 1.0653685331344604, "learning_rate": 2.966717324721861e-05, "loss": 0.6133, "num_input_tokens_seen": 23851952, "step": 17610 }, { "epoch": 0.5638243390307919, "grad_norm": 1.2612569332122803, "learning_rate": 2.9657528007129366e-05, "loss": 0.4559, "num_input_tokens_seen": 23858576, "step": 17615 }, { "epoch": 0.5639843800012804, "grad_norm": 0.6612778306007385, "learning_rate": 2.9647882048790777e-05, "loss": 0.3988, "num_input_tokens_seen": 23865664, "step": 17620 }, { "epoch": 0.5641444209717688, "grad_norm": 1.3265182971954346, "learning_rate": 2.963823537369037e-05, "loss": 0.3317, "num_input_tokens_seen": 23872240, "step": 17625 }, { "epoch": 0.5643044619422573, "grad_norm": 1.068508505821228, "learning_rate": 2.9628587983315775e-05, "loss": 0.5955, "num_input_tokens_seen": 23878928, "step": 17630 }, { "epoch": 0.5644645029127456, "grad_norm": 1.4244142770767212, "learning_rate": 2.9618939879154746e-05, "loss": 0.3891, "num_input_tokens_seen": 23885504, "step": 17635 }, { "epoch": 0.5646245438832341, "grad_norm": 0.8167967796325684, "learning_rate": 2.9609291062695143e-05, "loss": 0.4761, "num_input_tokens_seen": 23892240, "step": 17640 }, { "epoch": 0.5647845848537225, "grad_norm": 0.6771546006202698, "learning_rate": 2.9599641535424938e-05, "loss": 0.3608, "num_input_tokens_seen": 23898912, "step": 17645 }, { "epoch": 0.564944625824211, "grad_norm": 1.5181591510772705, "learning_rate": 2.9589991298832202e-05, "loss": 0.4114, "num_input_tokens_seen": 23905872, "step": 17650 }, { "epoch": 0.5651046667946994, "grad_norm": 0.5113324522972107, "learning_rate": 2.958034035440513e-05, "loss": 0.3689, "num_input_tokens_seen": 23912816, "step": 17655 }, { "epoch": 0.5652647077651879, "grad_norm": 1.2330983877182007, "learning_rate": 2.957068870363201e-05, "loss": 0.5075, "num_input_tokens_seen": 23919264, "step": 17660 }, { "epoch": 0.5654247487356764, "grad_norm": 1.8853645324707031, "learning_rate": 2.956103634800126e-05, "loss": 0.489, "num_input_tokens_seen": 23925936, "step": 17665 }, { "epoch": 0.5655847897061648, "grad_norm": 0.9224727153778076, "learning_rate": 2.9551383289001384e-05, "loss": 0.4783, "num_input_tokens_seen": 23932688, "step": 17670 }, { "epoch": 0.5657448306766533, "grad_norm": 0.9164918065071106, "learning_rate": 2.9541729528121005e-05, "loss": 0.4817, "num_input_tokens_seen": 23940032, "step": 17675 }, { "epoch": 0.5659048716471416, "grad_norm": 0.7354864478111267, "learning_rate": 2.9532075066848856e-05, "loss": 0.5949, "num_input_tokens_seen": 23946768, "step": 17680 }, { "epoch": 0.5660649126176301, "grad_norm": 1.1124248504638672, "learning_rate": 2.9522419906673786e-05, "loss": 0.5318, "num_input_tokens_seen": 23953728, "step": 17685 }, { "epoch": 0.5662249535881185, "grad_norm": 1.0801992416381836, "learning_rate": 2.951276404908474e-05, "loss": 0.432, "num_input_tokens_seen": 23960928, "step": 17690 }, { "epoch": 0.566384994558607, "grad_norm": 0.7199653387069702, "learning_rate": 2.9503107495570752e-05, "loss": 0.459, "num_input_tokens_seen": 23967440, "step": 17695 }, { "epoch": 0.5665450355290954, "grad_norm": 0.8425434827804565, "learning_rate": 2.9493450247621003e-05, "loss": 0.4352, "num_input_tokens_seen": 23974160, "step": 17700 }, { "epoch": 0.5667050764995839, "grad_norm": 1.0763251781463623, "learning_rate": 2.948379230672476e-05, "loss": 0.441, "num_input_tokens_seen": 23981248, "step": 17705 }, { "epoch": 0.5668651174700723, "grad_norm": 0.9062819480895996, "learning_rate": 2.9474133674371396e-05, "loss": 0.4277, "num_input_tokens_seen": 23988064, "step": 17710 }, { "epoch": 0.5670251584405608, "grad_norm": 1.345561146736145, "learning_rate": 2.9464474352050387e-05, "loss": 0.4445, "num_input_tokens_seen": 23994896, "step": 17715 }, { "epoch": 0.5671851994110493, "grad_norm": 0.9352979063987732, "learning_rate": 2.9454814341251336e-05, "loss": 0.5386, "num_input_tokens_seen": 24001584, "step": 17720 }, { "epoch": 0.5673452403815377, "grad_norm": 1.195769190788269, "learning_rate": 2.9445153643463942e-05, "loss": 0.4788, "num_input_tokens_seen": 24008544, "step": 17725 }, { "epoch": 0.5675052813520262, "grad_norm": 2.03393816947937, "learning_rate": 2.943549226017798e-05, "loss": 0.6793, "num_input_tokens_seen": 24015232, "step": 17730 }, { "epoch": 0.5676653223225145, "grad_norm": 1.2939295768737793, "learning_rate": 2.942583019288337e-05, "loss": 0.4385, "num_input_tokens_seen": 24022048, "step": 17735 }, { "epoch": 0.567825363293003, "grad_norm": 1.6739765405654907, "learning_rate": 2.9416167443070132e-05, "loss": 0.5372, "num_input_tokens_seen": 24028512, "step": 17740 }, { "epoch": 0.5679854042634914, "grad_norm": 0.8093753457069397, "learning_rate": 2.9406504012228375e-05, "loss": 0.5759, "num_input_tokens_seen": 24035168, "step": 17745 }, { "epoch": 0.5681454452339799, "grad_norm": 1.8613135814666748, "learning_rate": 2.939683990184832e-05, "loss": 0.5454, "num_input_tokens_seen": 24041664, "step": 17750 }, { "epoch": 0.5683054862044683, "grad_norm": 0.5815603137016296, "learning_rate": 2.93871751134203e-05, "loss": 0.42, "num_input_tokens_seen": 24048528, "step": 17755 }, { "epoch": 0.5684655271749568, "grad_norm": 1.1281766891479492, "learning_rate": 2.9377509648434752e-05, "loss": 0.4875, "num_input_tokens_seen": 24054864, "step": 17760 }, { "epoch": 0.5686255681454452, "grad_norm": 1.379487156867981, "learning_rate": 2.9367843508382203e-05, "loss": 0.331, "num_input_tokens_seen": 24061536, "step": 17765 }, { "epoch": 0.5687856091159337, "grad_norm": 2.929609775543213, "learning_rate": 2.9358176694753293e-05, "loss": 0.4633, "num_input_tokens_seen": 24068848, "step": 17770 }, { "epoch": 0.5689456500864222, "grad_norm": 1.7682311534881592, "learning_rate": 2.9348509209038766e-05, "loss": 0.3297, "num_input_tokens_seen": 24075680, "step": 17775 }, { "epoch": 0.5691056910569106, "grad_norm": 0.9147989153862, "learning_rate": 2.933884105272947e-05, "loss": 0.5235, "num_input_tokens_seen": 24082912, "step": 17780 }, { "epoch": 0.569265732027399, "grad_norm": 1.2142060995101929, "learning_rate": 2.9329172227316366e-05, "loss": 0.479, "num_input_tokens_seen": 24089648, "step": 17785 }, { "epoch": 0.5694257729978874, "grad_norm": 0.8319917917251587, "learning_rate": 2.93195027342905e-05, "loss": 0.3644, "num_input_tokens_seen": 24096160, "step": 17790 }, { "epoch": 0.5695858139683759, "grad_norm": 1.2882407903671265, "learning_rate": 2.9309832575143024e-05, "loss": 0.6144, "num_input_tokens_seen": 24102832, "step": 17795 }, { "epoch": 0.5697458549388643, "grad_norm": 1.1356488466262817, "learning_rate": 2.930016175136521e-05, "loss": 0.3186, "num_input_tokens_seen": 24109808, "step": 17800 }, { "epoch": 0.5697458549388643, "eval_loss": 0.46679630875587463, "eval_runtime": 500.9409, "eval_samples_per_second": 27.72, "eval_steps_per_second": 13.86, "num_input_tokens_seen": 24109808, "step": 17800 }, { "epoch": 0.5699058959093528, "grad_norm": 0.8049488067626953, "learning_rate": 2.9290490264448412e-05, "loss": 0.6824, "num_input_tokens_seen": 24116448, "step": 17805 }, { "epoch": 0.5700659368798412, "grad_norm": 0.8990801572799683, "learning_rate": 2.9280818115884094e-05, "loss": 0.3222, "num_input_tokens_seen": 24123440, "step": 17810 }, { "epoch": 0.5702259778503297, "grad_norm": 1.0932972431182861, "learning_rate": 2.9271145307163828e-05, "loss": 0.3709, "num_input_tokens_seen": 24130112, "step": 17815 }, { "epoch": 0.5703860188208181, "grad_norm": 1.2762202024459839, "learning_rate": 2.9261471839779287e-05, "loss": 0.3883, "num_input_tokens_seen": 24136944, "step": 17820 }, { "epoch": 0.5705460597913066, "grad_norm": 0.9216403961181641, "learning_rate": 2.925179771522223e-05, "loss": 0.36, "num_input_tokens_seen": 24143072, "step": 17825 }, { "epoch": 0.5707061007617951, "grad_norm": 2.917677879333496, "learning_rate": 2.9242122934984535e-05, "loss": 0.4198, "num_input_tokens_seen": 24149664, "step": 17830 }, { "epoch": 0.5708661417322834, "grad_norm": 0.839994490146637, "learning_rate": 2.9232447500558176e-05, "loss": 0.4835, "num_input_tokens_seen": 24156896, "step": 17835 }, { "epoch": 0.571026182702772, "grad_norm": 2.2371482849121094, "learning_rate": 2.9222771413435225e-05, "loss": 0.4641, "num_input_tokens_seen": 24163648, "step": 17840 }, { "epoch": 0.5711862236732603, "grad_norm": 1.5915287733078003, "learning_rate": 2.9213094675107848e-05, "loss": 0.5887, "num_input_tokens_seen": 24170464, "step": 17845 }, { "epoch": 0.5713462646437488, "grad_norm": 1.2887705564498901, "learning_rate": 2.9203417287068335e-05, "loss": 0.4987, "num_input_tokens_seen": 24177136, "step": 17850 }, { "epoch": 0.5715063056142372, "grad_norm": 0.5165514945983887, "learning_rate": 2.9193739250809042e-05, "loss": 0.2639, "num_input_tokens_seen": 24183744, "step": 17855 }, { "epoch": 0.5716663465847257, "grad_norm": 0.4397483468055725, "learning_rate": 2.9184060567822463e-05, "loss": 0.4139, "num_input_tokens_seen": 24190336, "step": 17860 }, { "epoch": 0.5718263875552141, "grad_norm": 1.5979816913604736, "learning_rate": 2.9174381239601166e-05, "loss": 0.4479, "num_input_tokens_seen": 24197472, "step": 17865 }, { "epoch": 0.5719864285257026, "grad_norm": 1.0404245853424072, "learning_rate": 2.916470126763783e-05, "loss": 0.3031, "num_input_tokens_seen": 24204080, "step": 17870 }, { "epoch": 0.5721464694961911, "grad_norm": 1.869951605796814, "learning_rate": 2.9155020653425203e-05, "loss": 0.4929, "num_input_tokens_seen": 24210784, "step": 17875 }, { "epoch": 0.5723065104666795, "grad_norm": 0.6884189248085022, "learning_rate": 2.9145339398456184e-05, "loss": 0.5563, "num_input_tokens_seen": 24217472, "step": 17880 }, { "epoch": 0.572466551437168, "grad_norm": 0.7710629105567932, "learning_rate": 2.913565750422374e-05, "loss": 0.2994, "num_input_tokens_seen": 24224032, "step": 17885 }, { "epoch": 0.5726265924076563, "grad_norm": 1.3434978723526, "learning_rate": 2.9125974972220938e-05, "loss": 0.438, "num_input_tokens_seen": 24230624, "step": 17890 }, { "epoch": 0.5727866333781448, "grad_norm": 0.9452151656150818, "learning_rate": 2.9116291803940932e-05, "loss": 0.4975, "num_input_tokens_seen": 24237584, "step": 17895 }, { "epoch": 0.5729466743486332, "grad_norm": 0.8884077072143555, "learning_rate": 2.910660800087701e-05, "loss": 0.3105, "num_input_tokens_seen": 24244048, "step": 17900 }, { "epoch": 0.5731067153191217, "grad_norm": 1.4022823572158813, "learning_rate": 2.909692356452254e-05, "loss": 0.4152, "num_input_tokens_seen": 24250960, "step": 17905 }, { "epoch": 0.5732667562896101, "grad_norm": 0.7346158623695374, "learning_rate": 2.9087238496370962e-05, "loss": 0.3303, "num_input_tokens_seen": 24257856, "step": 17910 }, { "epoch": 0.5734267972600986, "grad_norm": 0.8874785304069519, "learning_rate": 2.907755279791583e-05, "loss": 0.4695, "num_input_tokens_seen": 24264768, "step": 17915 }, { "epoch": 0.573586838230587, "grad_norm": 1.9887045621871948, "learning_rate": 2.906786647065083e-05, "loss": 0.5726, "num_input_tokens_seen": 24271728, "step": 17920 }, { "epoch": 0.5737468792010755, "grad_norm": 1.5918570756912231, "learning_rate": 2.9058179516069695e-05, "loss": 0.4247, "num_input_tokens_seen": 24278480, "step": 17925 }, { "epoch": 0.573906920171564, "grad_norm": 0.6525139212608337, "learning_rate": 2.9048491935666282e-05, "loss": 0.5345, "num_input_tokens_seen": 24285360, "step": 17930 }, { "epoch": 0.5740669611420524, "grad_norm": 0.6509631276130676, "learning_rate": 2.9038803730934534e-05, "loss": 0.3611, "num_input_tokens_seen": 24292192, "step": 17935 }, { "epoch": 0.5742270021125409, "grad_norm": 1.6881515979766846, "learning_rate": 2.9029114903368503e-05, "loss": 0.3936, "num_input_tokens_seen": 24299136, "step": 17940 }, { "epoch": 0.5743870430830292, "grad_norm": 1.0273772478103638, "learning_rate": 2.9019425454462318e-05, "loss": 0.4297, "num_input_tokens_seen": 24305680, "step": 17945 }, { "epoch": 0.5745470840535177, "grad_norm": 0.8799623847007751, "learning_rate": 2.9009735385710212e-05, "loss": 0.4542, "num_input_tokens_seen": 24313248, "step": 17950 }, { "epoch": 0.5747071250240061, "grad_norm": 1.936453104019165, "learning_rate": 2.900004469860652e-05, "loss": 0.5665, "num_input_tokens_seen": 24319968, "step": 17955 }, { "epoch": 0.5748671659944946, "grad_norm": 2.749157190322876, "learning_rate": 2.8990353394645668e-05, "loss": 0.4866, "num_input_tokens_seen": 24326912, "step": 17960 }, { "epoch": 0.575027206964983, "grad_norm": 1.5146172046661377, "learning_rate": 2.8980661475322186e-05, "loss": 0.4855, "num_input_tokens_seen": 24333200, "step": 17965 }, { "epoch": 0.5751872479354715, "grad_norm": 1.7150492668151855, "learning_rate": 2.897096894213067e-05, "loss": 0.5048, "num_input_tokens_seen": 24339872, "step": 17970 }, { "epoch": 0.5753472889059599, "grad_norm": 0.7181293964385986, "learning_rate": 2.8961275796565845e-05, "loss": 0.3345, "num_input_tokens_seen": 24346448, "step": 17975 }, { "epoch": 0.5755073298764484, "grad_norm": 1.167405366897583, "learning_rate": 2.8951582040122517e-05, "loss": 0.4668, "num_input_tokens_seen": 24353040, "step": 17980 }, { "epoch": 0.5756673708469369, "grad_norm": 1.0826358795166016, "learning_rate": 2.894188767429557e-05, "loss": 0.4485, "num_input_tokens_seen": 24359600, "step": 17985 }, { "epoch": 0.5758274118174252, "grad_norm": 1.5191144943237305, "learning_rate": 2.8932192700580014e-05, "loss": 0.468, "num_input_tokens_seen": 24366576, "step": 17990 }, { "epoch": 0.5759874527879137, "grad_norm": 0.8508705496788025, "learning_rate": 2.8922497120470916e-05, "loss": 0.4172, "num_input_tokens_seen": 24373584, "step": 17995 }, { "epoch": 0.5761474937584021, "grad_norm": 2.245709180831909, "learning_rate": 2.891280093546348e-05, "loss": 0.4891, "num_input_tokens_seen": 24380336, "step": 18000 }, { "epoch": 0.5761474937584021, "eval_loss": 0.46631523966789246, "eval_runtime": 501.1105, "eval_samples_per_second": 27.71, "eval_steps_per_second": 13.855, "num_input_tokens_seen": 24380336, "step": 18000 }, { "epoch": 0.5763075347288906, "grad_norm": 1.023675799369812, "learning_rate": 2.890310414705297e-05, "loss": 0.4075, "num_input_tokens_seen": 24387856, "step": 18005 }, { "epoch": 0.576467575699379, "grad_norm": 0.7621611952781677, "learning_rate": 2.8893406756734742e-05, "loss": 0.426, "num_input_tokens_seen": 24394976, "step": 18010 }, { "epoch": 0.5766276166698675, "grad_norm": 1.430893898010254, "learning_rate": 2.888370876600427e-05, "loss": 0.4815, "num_input_tokens_seen": 24402256, "step": 18015 }, { "epoch": 0.5767876576403559, "grad_norm": 0.8639985918998718, "learning_rate": 2.8874010176357104e-05, "loss": 0.6245, "num_input_tokens_seen": 24409536, "step": 18020 }, { "epoch": 0.5769476986108444, "grad_norm": 0.8131205439567566, "learning_rate": 2.886431098928888e-05, "loss": 0.4129, "num_input_tokens_seen": 24416016, "step": 18025 }, { "epoch": 0.5771077395813328, "grad_norm": 0.3935171365737915, "learning_rate": 2.885461120629534e-05, "loss": 0.3494, "num_input_tokens_seen": 24422656, "step": 18030 }, { "epoch": 0.5772677805518213, "grad_norm": 0.5913761258125305, "learning_rate": 2.8844910828872317e-05, "loss": 0.5551, "num_input_tokens_seen": 24430112, "step": 18035 }, { "epoch": 0.5774278215223098, "grad_norm": 0.7747605443000793, "learning_rate": 2.8835209858515715e-05, "loss": 0.3944, "num_input_tokens_seen": 24437280, "step": 18040 }, { "epoch": 0.5775878624927981, "grad_norm": 1.6908458471298218, "learning_rate": 2.8825508296721566e-05, "loss": 0.4726, "num_input_tokens_seen": 24444096, "step": 18045 }, { "epoch": 0.5777479034632866, "grad_norm": 0.45634621381759644, "learning_rate": 2.881580614498596e-05, "loss": 0.4431, "num_input_tokens_seen": 24450704, "step": 18050 }, { "epoch": 0.577907944433775, "grad_norm": 0.8778092265129089, "learning_rate": 2.8806103404805103e-05, "loss": 0.3933, "num_input_tokens_seen": 24457168, "step": 18055 }, { "epoch": 0.5780679854042635, "grad_norm": 1.039109706878662, "learning_rate": 2.8796400077675257e-05, "loss": 0.462, "num_input_tokens_seen": 24463968, "step": 18060 }, { "epoch": 0.5782280263747519, "grad_norm": 0.5666877031326294, "learning_rate": 2.8786696165092812e-05, "loss": 0.5062, "num_input_tokens_seen": 24470576, "step": 18065 }, { "epoch": 0.5783880673452404, "grad_norm": 1.378974437713623, "learning_rate": 2.8776991668554236e-05, "loss": 0.5572, "num_input_tokens_seen": 24477440, "step": 18070 }, { "epoch": 0.5785481083157288, "grad_norm": 1.8320152759552002, "learning_rate": 2.876728658955608e-05, "loss": 0.4141, "num_input_tokens_seen": 24484400, "step": 18075 }, { "epoch": 0.5787081492862173, "grad_norm": 0.733855664730072, "learning_rate": 2.8757580929594986e-05, "loss": 0.4557, "num_input_tokens_seen": 24490992, "step": 18080 }, { "epoch": 0.5788681902567057, "grad_norm": 1.4276132583618164, "learning_rate": 2.87478746901677e-05, "loss": 0.4421, "num_input_tokens_seen": 24497952, "step": 18085 }, { "epoch": 0.5790282312271942, "grad_norm": 1.4138861894607544, "learning_rate": 2.873816787277103e-05, "loss": 0.5624, "num_input_tokens_seen": 24504112, "step": 18090 }, { "epoch": 0.5791882721976827, "grad_norm": 0.8598321676254272, "learning_rate": 2.8728460478901903e-05, "loss": 0.4845, "num_input_tokens_seen": 24510704, "step": 18095 }, { "epoch": 0.579348313168171, "grad_norm": 0.9675015807151794, "learning_rate": 2.8718752510057307e-05, "loss": 0.3076, "num_input_tokens_seen": 24517376, "step": 18100 }, { "epoch": 0.5795083541386595, "grad_norm": 0.6562041640281677, "learning_rate": 2.870904396773435e-05, "loss": 0.4649, "num_input_tokens_seen": 24524304, "step": 18105 }, { "epoch": 0.5796683951091479, "grad_norm": 1.142846941947937, "learning_rate": 2.86993348534302e-05, "loss": 0.5545, "num_input_tokens_seen": 24530624, "step": 18110 }, { "epoch": 0.5798284360796364, "grad_norm": 0.7166200876235962, "learning_rate": 2.868962516864212e-05, "loss": 0.3285, "num_input_tokens_seen": 24537120, "step": 18115 }, { "epoch": 0.5799884770501248, "grad_norm": 1.2290209531784058, "learning_rate": 2.8679914914867477e-05, "loss": 0.4941, "num_input_tokens_seen": 24543904, "step": 18120 }, { "epoch": 0.5801485180206133, "grad_norm": 2.1227951049804688, "learning_rate": 2.8670204093603713e-05, "loss": 0.4729, "num_input_tokens_seen": 24550800, "step": 18125 }, { "epoch": 0.5803085589911017, "grad_norm": 0.8657891154289246, "learning_rate": 2.8660492706348357e-05, "loss": 0.3637, "num_input_tokens_seen": 24557552, "step": 18130 }, { "epoch": 0.5804685999615902, "grad_norm": 0.8357455134391785, "learning_rate": 2.8650780754599022e-05, "loss": 0.3717, "num_input_tokens_seen": 24564192, "step": 18135 }, { "epoch": 0.5806286409320787, "grad_norm": 1.0062628984451294, "learning_rate": 2.8641068239853407e-05, "loss": 0.3787, "num_input_tokens_seen": 24570752, "step": 18140 }, { "epoch": 0.580788681902567, "grad_norm": 0.7108971476554871, "learning_rate": 2.863135516360932e-05, "loss": 0.3967, "num_input_tokens_seen": 24577392, "step": 18145 }, { "epoch": 0.5809487228730555, "grad_norm": 1.2039738893508911, "learning_rate": 2.8621641527364633e-05, "loss": 0.5998, "num_input_tokens_seen": 24584016, "step": 18150 }, { "epoch": 0.5811087638435439, "grad_norm": 0.9864029884338379, "learning_rate": 2.8611927332617313e-05, "loss": 0.5534, "num_input_tokens_seen": 24590960, "step": 18155 }, { "epoch": 0.5812688048140324, "grad_norm": 2.221626043319702, "learning_rate": 2.8602212580865405e-05, "loss": 0.4773, "num_input_tokens_seen": 24598384, "step": 18160 }, { "epoch": 0.5814288457845208, "grad_norm": 0.7146998047828674, "learning_rate": 2.859249727360705e-05, "loss": 0.6558, "num_input_tokens_seen": 24605568, "step": 18165 }, { "epoch": 0.5815888867550093, "grad_norm": 1.454753041267395, "learning_rate": 2.8582781412340465e-05, "loss": 0.4492, "num_input_tokens_seen": 24611968, "step": 18170 }, { "epoch": 0.5817489277254977, "grad_norm": 1.123781442642212, "learning_rate": 2.857306499856397e-05, "loss": 0.4406, "num_input_tokens_seen": 24619184, "step": 18175 }, { "epoch": 0.5819089686959862, "grad_norm": 3.26322603225708, "learning_rate": 2.856334803377594e-05, "loss": 0.4408, "num_input_tokens_seen": 24626400, "step": 18180 }, { "epoch": 0.5820690096664746, "grad_norm": 0.2836594879627228, "learning_rate": 2.8553630519474867e-05, "loss": 0.3332, "num_input_tokens_seen": 24632720, "step": 18185 }, { "epoch": 0.5822290506369631, "grad_norm": 1.2833739519119263, "learning_rate": 2.8543912457159317e-05, "loss": 0.3813, "num_input_tokens_seen": 24639536, "step": 18190 }, { "epoch": 0.5823890916074516, "grad_norm": 1.8142666816711426, "learning_rate": 2.853419384832792e-05, "loss": 0.4058, "num_input_tokens_seen": 24646304, "step": 18195 }, { "epoch": 0.5825491325779399, "grad_norm": 0.5139685869216919, "learning_rate": 2.8524474694479423e-05, "loss": 0.458, "num_input_tokens_seen": 24653072, "step": 18200 }, { "epoch": 0.5825491325779399, "eval_loss": 0.46614232659339905, "eval_runtime": 500.706, "eval_samples_per_second": 27.733, "eval_steps_per_second": 13.866, "num_input_tokens_seen": 24653072, "step": 18200 }, { "epoch": 0.5827091735484284, "grad_norm": 0.9849366545677185, "learning_rate": 2.851475499711264e-05, "loss": 0.5571, "num_input_tokens_seen": 24660000, "step": 18205 }, { "epoch": 0.5828692145189168, "grad_norm": 1.2649853229522705, "learning_rate": 2.8505034757726468e-05, "loss": 0.5878, "num_input_tokens_seen": 24666544, "step": 18210 }, { "epoch": 0.5830292554894053, "grad_norm": 0.4858602285385132, "learning_rate": 2.8495313977819886e-05, "loss": 0.6133, "num_input_tokens_seen": 24673168, "step": 18215 }, { "epoch": 0.5831892964598937, "grad_norm": 1.136054515838623, "learning_rate": 2.8485592658891956e-05, "loss": 0.4707, "num_input_tokens_seen": 24680144, "step": 18220 }, { "epoch": 0.5833493374303822, "grad_norm": 1.8750320672988892, "learning_rate": 2.8475870802441844e-05, "loss": 0.5173, "num_input_tokens_seen": 24687648, "step": 18225 }, { "epoch": 0.5835093784008706, "grad_norm": 0.7645443081855774, "learning_rate": 2.8466148409968774e-05, "loss": 0.5322, "num_input_tokens_seen": 24694336, "step": 18230 }, { "epoch": 0.5836694193713591, "grad_norm": 0.7903144359588623, "learning_rate": 2.8456425482972067e-05, "loss": 0.3323, "num_input_tokens_seen": 24701008, "step": 18235 }, { "epoch": 0.5838294603418475, "grad_norm": 0.8228718638420105, "learning_rate": 2.84467020229511e-05, "loss": 0.434, "num_input_tokens_seen": 24707872, "step": 18240 }, { "epoch": 0.583989501312336, "grad_norm": 2.22483491897583, "learning_rate": 2.8436978031405375e-05, "loss": 0.3679, "num_input_tokens_seen": 24714944, "step": 18245 }, { "epoch": 0.5841495422828245, "grad_norm": 1.4832876920700073, "learning_rate": 2.842725350983445e-05, "loss": 0.4102, "num_input_tokens_seen": 24721616, "step": 18250 }, { "epoch": 0.5843095832533128, "grad_norm": 2.62211012840271, "learning_rate": 2.8417528459737957e-05, "loss": 0.4089, "num_input_tokens_seen": 24728048, "step": 18255 }, { "epoch": 0.5844696242238013, "grad_norm": 1.8954955339431763, "learning_rate": 2.8407802882615624e-05, "loss": 0.4622, "num_input_tokens_seen": 24734480, "step": 18260 }, { "epoch": 0.5846296651942897, "grad_norm": 0.5865403413772583, "learning_rate": 2.8398076779967277e-05, "loss": 0.4288, "num_input_tokens_seen": 24741296, "step": 18265 }, { "epoch": 0.5847897061647782, "grad_norm": 0.5985564589500427, "learning_rate": 2.8388350153292774e-05, "loss": 0.5329, "num_input_tokens_seen": 24747632, "step": 18270 }, { "epoch": 0.5849497471352666, "grad_norm": 0.9115424752235413, "learning_rate": 2.8378623004092103e-05, "loss": 0.6879, "num_input_tokens_seen": 24754416, "step": 18275 }, { "epoch": 0.5851097881057551, "grad_norm": 1.5386015176773071, "learning_rate": 2.8368895333865302e-05, "loss": 0.5281, "num_input_tokens_seen": 24760976, "step": 18280 }, { "epoch": 0.5852698290762435, "grad_norm": 0.5414811372756958, "learning_rate": 2.835916714411251e-05, "loss": 0.2378, "num_input_tokens_seen": 24767648, "step": 18285 }, { "epoch": 0.585429870046732, "grad_norm": 0.5024917125701904, "learning_rate": 2.8349438436333926e-05, "loss": 0.6092, "num_input_tokens_seen": 24774112, "step": 18290 }, { "epoch": 0.5855899110172204, "grad_norm": 0.83122718334198, "learning_rate": 2.833970921202984e-05, "loss": 0.2767, "num_input_tokens_seen": 24780752, "step": 18295 }, { "epoch": 0.5857499519877089, "grad_norm": 1.3910300731658936, "learning_rate": 2.8329979472700628e-05, "loss": 0.4636, "num_input_tokens_seen": 24787504, "step": 18300 }, { "epoch": 0.5859099929581973, "grad_norm": 1.0969585180282593, "learning_rate": 2.832024921984674e-05, "loss": 0.4242, "num_input_tokens_seen": 24794480, "step": 18305 }, { "epoch": 0.5860700339286857, "grad_norm": 0.9538201093673706, "learning_rate": 2.8310518454968693e-05, "loss": 0.415, "num_input_tokens_seen": 24801280, "step": 18310 }, { "epoch": 0.5862300748991742, "grad_norm": 3.73749041557312, "learning_rate": 2.8300787179567095e-05, "loss": 0.4625, "num_input_tokens_seen": 24808080, "step": 18315 }, { "epoch": 0.5863901158696626, "grad_norm": 1.072460651397705, "learning_rate": 2.8291055395142636e-05, "loss": 0.3387, "num_input_tokens_seen": 24815088, "step": 18320 }, { "epoch": 0.5865501568401511, "grad_norm": 0.9983315467834473, "learning_rate": 2.8281323103196073e-05, "loss": 0.3955, "num_input_tokens_seen": 24821520, "step": 18325 }, { "epoch": 0.5867101978106395, "grad_norm": 0.5777068138122559, "learning_rate": 2.8271590305228256e-05, "loss": 0.5996, "num_input_tokens_seen": 24828592, "step": 18330 }, { "epoch": 0.586870238781128, "grad_norm": 1.299630880355835, "learning_rate": 2.82618570027401e-05, "loss": 0.4693, "num_input_tokens_seen": 24835424, "step": 18335 }, { "epoch": 0.5870302797516164, "grad_norm": 1.5122543573379517, "learning_rate": 2.8252123197232604e-05, "loss": 0.3993, "num_input_tokens_seen": 24842448, "step": 18340 }, { "epoch": 0.5871903207221049, "grad_norm": 1.0888378620147705, "learning_rate": 2.8242388890206843e-05, "loss": 0.4928, "num_input_tokens_seen": 24849408, "step": 18345 }, { "epoch": 0.5873503616925934, "grad_norm": 1.0421876907348633, "learning_rate": 2.8232654083163967e-05, "loss": 0.4755, "num_input_tokens_seen": 24856416, "step": 18350 }, { "epoch": 0.5875104026630817, "grad_norm": 1.1809478998184204, "learning_rate": 2.822291877760521e-05, "loss": 0.4144, "num_input_tokens_seen": 24863136, "step": 18355 }, { "epoch": 0.5876704436335702, "grad_norm": 0.6731178164482117, "learning_rate": 2.8213182975031864e-05, "loss": 0.305, "num_input_tokens_seen": 24869904, "step": 18360 }, { "epoch": 0.5878304846040586, "grad_norm": 0.7517751455307007, "learning_rate": 2.8203446676945337e-05, "loss": 0.5245, "num_input_tokens_seen": 24877248, "step": 18365 }, { "epoch": 0.5879905255745471, "grad_norm": 1.1457079648971558, "learning_rate": 2.8193709884847075e-05, "loss": 0.5355, "num_input_tokens_seen": 24883472, "step": 18370 }, { "epoch": 0.5881505665450355, "grad_norm": 0.6922146081924438, "learning_rate": 2.8183972600238605e-05, "loss": 0.4581, "num_input_tokens_seen": 24890432, "step": 18375 }, { "epoch": 0.588310607515524, "grad_norm": 0.8709003329277039, "learning_rate": 2.817423482462156e-05, "loss": 0.4502, "num_input_tokens_seen": 24896944, "step": 18380 }, { "epoch": 0.5884706484860124, "grad_norm": 0.801264226436615, "learning_rate": 2.8164496559497605e-05, "loss": 0.4955, "num_input_tokens_seen": 24904016, "step": 18385 }, { "epoch": 0.5886306894565009, "grad_norm": 1.6901888847351074, "learning_rate": 2.815475780636852e-05, "loss": 0.3852, "num_input_tokens_seen": 24911008, "step": 18390 }, { "epoch": 0.5887907304269893, "grad_norm": 0.6560359597206116, "learning_rate": 2.814501856673613e-05, "loss": 0.3611, "num_input_tokens_seen": 24918032, "step": 18395 }, { "epoch": 0.5889507713974778, "grad_norm": 1.246032476425171, "learning_rate": 2.8135278842102353e-05, "loss": 0.4189, "num_input_tokens_seen": 24924912, "step": 18400 }, { "epoch": 0.5889507713974778, "eval_loss": 0.4656590521335602, "eval_runtime": 501.7783, "eval_samples_per_second": 27.674, "eval_steps_per_second": 13.837, "num_input_tokens_seen": 24924912, "step": 18400 }, { "epoch": 0.5891108123679663, "grad_norm": 0.7574769854545593, "learning_rate": 2.8125538633969183e-05, "loss": 0.3894, "num_input_tokens_seen": 24931712, "step": 18405 }, { "epoch": 0.5892708533384546, "grad_norm": 1.513626217842102, "learning_rate": 2.8115797943838677e-05, "loss": 0.3323, "num_input_tokens_seen": 24938288, "step": 18410 }, { "epoch": 0.5894308943089431, "grad_norm": 1.8038785457611084, "learning_rate": 2.810605677321298e-05, "loss": 0.4652, "num_input_tokens_seen": 24945264, "step": 18415 }, { "epoch": 0.5895909352794315, "grad_norm": 0.8533377647399902, "learning_rate": 2.809631512359428e-05, "loss": 0.413, "num_input_tokens_seen": 24951920, "step": 18420 }, { "epoch": 0.58975097624992, "grad_norm": 0.9171306490898132, "learning_rate": 2.8086572996484884e-05, "loss": 0.3317, "num_input_tokens_seen": 24958752, "step": 18425 }, { "epoch": 0.5899110172204084, "grad_norm": 0.8547083735466003, "learning_rate": 2.8076830393387143e-05, "loss": 0.4227, "num_input_tokens_seen": 24965408, "step": 18430 }, { "epoch": 0.5900710581908969, "grad_norm": 0.7318635582923889, "learning_rate": 2.8067087315803497e-05, "loss": 0.4334, "num_input_tokens_seen": 24972208, "step": 18435 }, { "epoch": 0.5902310991613853, "grad_norm": 1.1962274312973022, "learning_rate": 2.8057343765236433e-05, "loss": 0.3694, "num_input_tokens_seen": 24978896, "step": 18440 }, { "epoch": 0.5903911401318738, "grad_norm": 2.0307424068450928, "learning_rate": 2.804759974318854e-05, "loss": 0.4662, "num_input_tokens_seen": 24986144, "step": 18445 }, { "epoch": 0.5905511811023622, "grad_norm": 0.8388469815254211, "learning_rate": 2.8037855251162482e-05, "loss": 0.4027, "num_input_tokens_seen": 24992896, "step": 18450 }, { "epoch": 0.5907112220728507, "grad_norm": 1.095177173614502, "learning_rate": 2.802811029066096e-05, "loss": 0.3984, "num_input_tokens_seen": 24999552, "step": 18455 }, { "epoch": 0.5908712630433391, "grad_norm": 1.1936161518096924, "learning_rate": 2.8018364863186764e-05, "loss": 0.3792, "num_input_tokens_seen": 25006752, "step": 18460 }, { "epoch": 0.5910313040138275, "grad_norm": 0.927971601486206, "learning_rate": 2.800861897024279e-05, "loss": 0.4591, "num_input_tokens_seen": 25013472, "step": 18465 }, { "epoch": 0.591191344984316, "grad_norm": 2.121753454208374, "learning_rate": 2.799887261333196e-05, "loss": 0.4869, "num_input_tokens_seen": 25020176, "step": 18470 }, { "epoch": 0.5913513859548044, "grad_norm": 0.7883803248405457, "learning_rate": 2.798912579395728e-05, "loss": 0.3514, "num_input_tokens_seen": 25026992, "step": 18475 }, { "epoch": 0.5915114269252929, "grad_norm": 1.1712121963500977, "learning_rate": 2.797937851362185e-05, "loss": 0.3134, "num_input_tokens_seen": 25033920, "step": 18480 }, { "epoch": 0.5916714678957813, "grad_norm": 0.3858097493648529, "learning_rate": 2.7969630773828802e-05, "loss": 0.4074, "num_input_tokens_seen": 25040800, "step": 18485 }, { "epoch": 0.5918315088662698, "grad_norm": 2.573045015335083, "learning_rate": 2.7959882576081382e-05, "loss": 0.3424, "num_input_tokens_seen": 25047248, "step": 18490 }, { "epoch": 0.5919915498367582, "grad_norm": 1.1976600885391235, "learning_rate": 2.795013392188286e-05, "loss": 0.3615, "num_input_tokens_seen": 25053792, "step": 18495 }, { "epoch": 0.5921515908072467, "grad_norm": 1.0723965167999268, "learning_rate": 2.7940384812736614e-05, "loss": 0.4025, "num_input_tokens_seen": 25060688, "step": 18500 }, { "epoch": 0.592311631777735, "grad_norm": 1.5031468868255615, "learning_rate": 2.7930635250146087e-05, "loss": 0.3954, "num_input_tokens_seen": 25067376, "step": 18505 }, { "epoch": 0.5924716727482235, "grad_norm": 1.172600269317627, "learning_rate": 2.792088523561477e-05, "loss": 0.3562, "num_input_tokens_seen": 25074064, "step": 18510 }, { "epoch": 0.592631713718712, "grad_norm": 0.8321752548217773, "learning_rate": 2.7911134770646246e-05, "loss": 0.5114, "num_input_tokens_seen": 25081344, "step": 18515 }, { "epoch": 0.5927917546892004, "grad_norm": 1.23635995388031, "learning_rate": 2.7901383856744157e-05, "loss": 0.443, "num_input_tokens_seen": 25088208, "step": 18520 }, { "epoch": 0.5929517956596889, "grad_norm": 0.8324930667877197, "learning_rate": 2.7891632495412217e-05, "loss": 0.4517, "num_input_tokens_seen": 25095008, "step": 18525 }, { "epoch": 0.5931118366301773, "grad_norm": 1.218210220336914, "learning_rate": 2.7881880688154205e-05, "loss": 0.4003, "num_input_tokens_seen": 25101632, "step": 18530 }, { "epoch": 0.5932718776006658, "grad_norm": 1.0514323711395264, "learning_rate": 2.7872128436473977e-05, "loss": 0.4496, "num_input_tokens_seen": 25108432, "step": 18535 }, { "epoch": 0.5934319185711542, "grad_norm": 1.2559621334075928, "learning_rate": 2.7862375741875448e-05, "loss": 0.2511, "num_input_tokens_seen": 25114928, "step": 18540 }, { "epoch": 0.5935919595416427, "grad_norm": 1.504115104675293, "learning_rate": 2.785262260586261e-05, "loss": 0.5673, "num_input_tokens_seen": 25121920, "step": 18545 }, { "epoch": 0.5937520005121311, "grad_norm": 1.2239793539047241, "learning_rate": 2.7842869029939517e-05, "loss": 0.5896, "num_input_tokens_seen": 25128304, "step": 18550 }, { "epoch": 0.5939120414826196, "grad_norm": 1.163636565208435, "learning_rate": 2.7833115015610296e-05, "loss": 0.4775, "num_input_tokens_seen": 25135168, "step": 18555 }, { "epoch": 0.5940720824531079, "grad_norm": 0.8722857236862183, "learning_rate": 2.7823360564379136e-05, "loss": 0.4517, "num_input_tokens_seen": 25141632, "step": 18560 }, { "epoch": 0.5942321234235964, "grad_norm": 0.8473311066627502, "learning_rate": 2.7813605677750297e-05, "loss": 0.4167, "num_input_tokens_seen": 25148352, "step": 18565 }, { "epoch": 0.5943921643940849, "grad_norm": 0.6133707761764526, "learning_rate": 2.7803850357228102e-05, "loss": 0.3814, "num_input_tokens_seen": 25154960, "step": 18570 }, { "epoch": 0.5945522053645733, "grad_norm": 1.033425211906433, "learning_rate": 2.779409460431695e-05, "loss": 0.5586, "num_input_tokens_seen": 25161616, "step": 18575 }, { "epoch": 0.5947122463350618, "grad_norm": 0.896207332611084, "learning_rate": 2.778433842052129e-05, "loss": 0.3564, "num_input_tokens_seen": 25168656, "step": 18580 }, { "epoch": 0.5948722873055502, "grad_norm": 1.8070207834243774, "learning_rate": 2.7774581807345664e-05, "loss": 0.5562, "num_input_tokens_seen": 25175984, "step": 18585 }, { "epoch": 0.5950323282760387, "grad_norm": 0.4439444839954376, "learning_rate": 2.776482476629465e-05, "loss": 0.4229, "num_input_tokens_seen": 25182848, "step": 18590 }, { "epoch": 0.5951923692465271, "grad_norm": 0.5393457412719727, "learning_rate": 2.7755067298872924e-05, "loss": 0.4761, "num_input_tokens_seen": 25189824, "step": 18595 }, { "epoch": 0.5953524102170156, "grad_norm": 0.9157475233078003, "learning_rate": 2.774530940658518e-05, "loss": 0.4599, "num_input_tokens_seen": 25196400, "step": 18600 }, { "epoch": 0.5953524102170156, "eval_loss": 0.46608665585517883, "eval_runtime": 502.3547, "eval_samples_per_second": 27.642, "eval_steps_per_second": 13.821, "num_input_tokens_seen": 25196400, "step": 18600 }, { "epoch": 0.595512451187504, "grad_norm": 1.0350232124328613, "learning_rate": 2.7735551090936236e-05, "loss": 0.4856, "num_input_tokens_seen": 25203536, "step": 18605 }, { "epoch": 0.5956724921579925, "grad_norm": 0.9314146637916565, "learning_rate": 2.7725792353430934e-05, "loss": 0.4661, "num_input_tokens_seen": 25210224, "step": 18610 }, { "epoch": 0.595832533128481, "grad_norm": 1.0977489948272705, "learning_rate": 2.77160331955742e-05, "loss": 0.2986, "num_input_tokens_seen": 25217296, "step": 18615 }, { "epoch": 0.5959925740989693, "grad_norm": 0.7888285517692566, "learning_rate": 2.7706273618871008e-05, "loss": 0.4085, "num_input_tokens_seen": 25224096, "step": 18620 }, { "epoch": 0.5961526150694578, "grad_norm": 1.7822703123092651, "learning_rate": 2.769651362482642e-05, "loss": 0.3375, "num_input_tokens_seen": 25230896, "step": 18625 }, { "epoch": 0.5963126560399462, "grad_norm": 1.4718092679977417, "learning_rate": 2.768675321494555e-05, "loss": 0.556, "num_input_tokens_seen": 25238080, "step": 18630 }, { "epoch": 0.5964726970104347, "grad_norm": 2.147491693496704, "learning_rate": 2.7676992390733565e-05, "loss": 0.5953, "num_input_tokens_seen": 25244720, "step": 18635 }, { "epoch": 0.5966327379809231, "grad_norm": 1.0054174661636353, "learning_rate": 2.766723115369571e-05, "loss": 0.3384, "num_input_tokens_seen": 25251104, "step": 18640 }, { "epoch": 0.5967927789514116, "grad_norm": 0.7019771337509155, "learning_rate": 2.765746950533729e-05, "loss": 0.4887, "num_input_tokens_seen": 25257664, "step": 18645 }, { "epoch": 0.5969528199219, "grad_norm": 0.8868275880813599, "learning_rate": 2.7647707447163684e-05, "loss": 0.3164, "num_input_tokens_seen": 25264128, "step": 18650 }, { "epoch": 0.5971128608923885, "grad_norm": 1.1252919435501099, "learning_rate": 2.7637944980680315e-05, "loss": 0.437, "num_input_tokens_seen": 25270704, "step": 18655 }, { "epoch": 0.5972729018628768, "grad_norm": 2.471052646636963, "learning_rate": 2.762818210739268e-05, "loss": 0.5079, "num_input_tokens_seen": 25277248, "step": 18660 }, { "epoch": 0.5974329428333653, "grad_norm": 1.4431928396224976, "learning_rate": 2.7618418828806332e-05, "loss": 0.4164, "num_input_tokens_seen": 25283936, "step": 18665 }, { "epoch": 0.5975929838038538, "grad_norm": 2.0910451412200928, "learning_rate": 2.76086551464269e-05, "loss": 0.5276, "num_input_tokens_seen": 25290400, "step": 18670 }, { "epoch": 0.5977530247743422, "grad_norm": 0.9017327427864075, "learning_rate": 2.759889106176006e-05, "loss": 0.6397, "num_input_tokens_seen": 25297264, "step": 18675 }, { "epoch": 0.5979130657448307, "grad_norm": 0.856281042098999, "learning_rate": 2.758912657631156e-05, "loss": 0.3997, "num_input_tokens_seen": 25304176, "step": 18680 }, { "epoch": 0.5980731067153191, "grad_norm": 0.6228924989700317, "learning_rate": 2.7579361691587198e-05, "loss": 0.4043, "num_input_tokens_seen": 25310928, "step": 18685 }, { "epoch": 0.5982331476858076, "grad_norm": 1.4014313220977783, "learning_rate": 2.756959640909285e-05, "loss": 0.5803, "num_input_tokens_seen": 25317664, "step": 18690 }, { "epoch": 0.598393188656296, "grad_norm": 1.2718595266342163, "learning_rate": 2.7559830730334452e-05, "loss": 0.415, "num_input_tokens_seen": 25324512, "step": 18695 }, { "epoch": 0.5985532296267845, "grad_norm": 1.183663010597229, "learning_rate": 2.7550064656817988e-05, "loss": 0.4983, "num_input_tokens_seen": 25331552, "step": 18700 }, { "epoch": 0.5987132705972729, "grad_norm": 0.8490984439849854, "learning_rate": 2.7540298190049503e-05, "loss": 0.415, "num_input_tokens_seen": 25338048, "step": 18705 }, { "epoch": 0.5988733115677614, "grad_norm": 1.617436170578003, "learning_rate": 2.7530531331535107e-05, "loss": 0.5138, "num_input_tokens_seen": 25344736, "step": 18710 }, { "epoch": 0.5990333525382497, "grad_norm": 1.61823308467865, "learning_rate": 2.752076408278099e-05, "loss": 0.4951, "num_input_tokens_seen": 25352400, "step": 18715 }, { "epoch": 0.5991933935087382, "grad_norm": 1.1962251663208008, "learning_rate": 2.751099644529337e-05, "loss": 0.4645, "num_input_tokens_seen": 25359328, "step": 18720 }, { "epoch": 0.5993534344792267, "grad_norm": 1.5761051177978516, "learning_rate": 2.7501228420578533e-05, "loss": 0.5616, "num_input_tokens_seen": 25366224, "step": 18725 }, { "epoch": 0.5995134754497151, "grad_norm": 1.5993478298187256, "learning_rate": 2.7491460010142857e-05, "loss": 0.4904, "num_input_tokens_seen": 25372816, "step": 18730 }, { "epoch": 0.5996735164202036, "grad_norm": 2.441901683807373, "learning_rate": 2.7481691215492727e-05, "loss": 0.6252, "num_input_tokens_seen": 25379808, "step": 18735 }, { "epoch": 0.599833557390692, "grad_norm": 0.7238442301750183, "learning_rate": 2.747192203813463e-05, "loss": 0.5726, "num_input_tokens_seen": 25386560, "step": 18740 }, { "epoch": 0.5999935983611805, "grad_norm": 4.354560375213623, "learning_rate": 2.7462152479575087e-05, "loss": 0.5831, "num_input_tokens_seen": 25393024, "step": 18745 }, { "epoch": 0.6001536393316689, "grad_norm": 0.5313559770584106, "learning_rate": 2.7452382541320697e-05, "loss": 0.3961, "num_input_tokens_seen": 25399888, "step": 18750 }, { "epoch": 0.6003136803021574, "grad_norm": 0.6854838132858276, "learning_rate": 2.7442612224878096e-05, "loss": 0.4098, "num_input_tokens_seen": 25406544, "step": 18755 }, { "epoch": 0.6004737212726458, "grad_norm": 0.8668001890182495, "learning_rate": 2.7432841531753994e-05, "loss": 0.327, "num_input_tokens_seen": 25413520, "step": 18760 }, { "epoch": 0.6006337622431343, "grad_norm": 0.41424766182899475, "learning_rate": 2.7423070463455147e-05, "loss": 0.5134, "num_input_tokens_seen": 25420432, "step": 18765 }, { "epoch": 0.6007938032136226, "grad_norm": 0.47481605410575867, "learning_rate": 2.7413299021488397e-05, "loss": 0.5975, "num_input_tokens_seen": 25427024, "step": 18770 }, { "epoch": 0.6009538441841111, "grad_norm": 1.4907073974609375, "learning_rate": 2.7403527207360615e-05, "loss": 0.4844, "num_input_tokens_seen": 25433968, "step": 18775 }, { "epoch": 0.6011138851545996, "grad_norm": 0.6130492687225342, "learning_rate": 2.7393755022578722e-05, "loss": 0.6208, "num_input_tokens_seen": 25441072, "step": 18780 }, { "epoch": 0.601273926125088, "grad_norm": 1.6054235696792603, "learning_rate": 2.7383982468649714e-05, "loss": 0.2996, "num_input_tokens_seen": 25448320, "step": 18785 }, { "epoch": 0.6014339670955765, "grad_norm": 1.6459356546401978, "learning_rate": 2.7374209547080665e-05, "loss": 0.5081, "num_input_tokens_seen": 25455744, "step": 18790 }, { "epoch": 0.6015940080660649, "grad_norm": 0.6801779270172119, "learning_rate": 2.7364436259378663e-05, "loss": 0.3918, "num_input_tokens_seen": 25462080, "step": 18795 }, { "epoch": 0.6017540490365534, "grad_norm": 1.1158188581466675, "learning_rate": 2.735466260705088e-05, "loss": 0.3823, "num_input_tokens_seen": 25468816, "step": 18800 }, { "epoch": 0.6017540490365534, "eval_loss": 0.4649721086025238, "eval_runtime": 502.6061, "eval_samples_per_second": 27.628, "eval_steps_per_second": 13.814, "num_input_tokens_seen": 25468816, "step": 18800 }, { "epoch": 0.6019140900070418, "grad_norm": 2.236921548843384, "learning_rate": 2.7344888591604524e-05, "loss": 0.6107, "num_input_tokens_seen": 25475648, "step": 18805 }, { "epoch": 0.6020741309775303, "grad_norm": 1.695690393447876, "learning_rate": 2.7335114214546893e-05, "loss": 0.543, "num_input_tokens_seen": 25482336, "step": 18810 }, { "epoch": 0.6022341719480186, "grad_norm": 0.860981822013855, "learning_rate": 2.7325339477385293e-05, "loss": 0.2885, "num_input_tokens_seen": 25488960, "step": 18815 }, { "epoch": 0.6023942129185071, "grad_norm": 1.0557172298431396, "learning_rate": 2.7315564381627128e-05, "loss": 0.4149, "num_input_tokens_seen": 25495776, "step": 18820 }, { "epoch": 0.6025542538889955, "grad_norm": 2.07210636138916, "learning_rate": 2.7305788928779835e-05, "loss": 0.3957, "num_input_tokens_seen": 25502480, "step": 18825 }, { "epoch": 0.602714294859484, "grad_norm": 0.9985405802726746, "learning_rate": 2.729601312035091e-05, "loss": 0.3989, "num_input_tokens_seen": 25509184, "step": 18830 }, { "epoch": 0.6028743358299725, "grad_norm": 1.103640079498291, "learning_rate": 2.7286236957847915e-05, "loss": 0.6182, "num_input_tokens_seen": 25516032, "step": 18835 }, { "epoch": 0.6030343768004609, "grad_norm": 0.6023880839347839, "learning_rate": 2.7276460442778446e-05, "loss": 0.5341, "num_input_tokens_seen": 25523104, "step": 18840 }, { "epoch": 0.6031944177709494, "grad_norm": 1.6575411558151245, "learning_rate": 2.726668357665017e-05, "loss": 0.4045, "num_input_tokens_seen": 25530272, "step": 18845 }, { "epoch": 0.6033544587414378, "grad_norm": 0.816554605960846, "learning_rate": 2.7256906360970808e-05, "loss": 0.4856, "num_input_tokens_seen": 25536704, "step": 18850 }, { "epoch": 0.6035144997119263, "grad_norm": 0.6883811354637146, "learning_rate": 2.7247128797248117e-05, "loss": 0.4854, "num_input_tokens_seen": 25543504, "step": 18855 }, { "epoch": 0.6036745406824147, "grad_norm": 1.933488130569458, "learning_rate": 2.7237350886989925e-05, "loss": 0.6238, "num_input_tokens_seen": 25550752, "step": 18860 }, { "epoch": 0.6038345816529032, "grad_norm": 1.1757093667984009, "learning_rate": 2.7227572631704107e-05, "loss": 0.3769, "num_input_tokens_seen": 25557504, "step": 18865 }, { "epoch": 0.6039946226233915, "grad_norm": 1.4518988132476807, "learning_rate": 2.7217794032898596e-05, "loss": 0.5781, "num_input_tokens_seen": 25564272, "step": 18870 }, { "epoch": 0.60415466359388, "grad_norm": 1.0864216089248657, "learning_rate": 2.7208015092081384e-05, "loss": 0.4324, "num_input_tokens_seen": 25570704, "step": 18875 }, { "epoch": 0.6043147045643685, "grad_norm": 4.184746265411377, "learning_rate": 2.719823581076049e-05, "loss": 0.6718, "num_input_tokens_seen": 25577424, "step": 18880 }, { "epoch": 0.6044747455348569, "grad_norm": 0.5868710279464722, "learning_rate": 2.718845619044401e-05, "loss": 0.2685, "num_input_tokens_seen": 25584224, "step": 18885 }, { "epoch": 0.6046347865053454, "grad_norm": 2.4495418071746826, "learning_rate": 2.7178676232640088e-05, "loss": 0.4755, "num_input_tokens_seen": 25591120, "step": 18890 }, { "epoch": 0.6047948274758338, "grad_norm": 1.3036906719207764, "learning_rate": 2.716889593885691e-05, "loss": 0.3514, "num_input_tokens_seen": 25598032, "step": 18895 }, { "epoch": 0.6049548684463223, "grad_norm": 0.5910568833351135, "learning_rate": 2.7159115310602716e-05, "loss": 0.4264, "num_input_tokens_seen": 25605360, "step": 18900 }, { "epoch": 0.6051149094168107, "grad_norm": 1.0373882055282593, "learning_rate": 2.7149334349385814e-05, "loss": 0.3937, "num_input_tokens_seen": 25612128, "step": 18905 }, { "epoch": 0.6052749503872992, "grad_norm": 1.0516316890716553, "learning_rate": 2.713955305671454e-05, "loss": 0.443, "num_input_tokens_seen": 25618592, "step": 18910 }, { "epoch": 0.6054349913577876, "grad_norm": 2.181600332260132, "learning_rate": 2.71297714340973e-05, "loss": 0.6152, "num_input_tokens_seen": 25625280, "step": 18915 }, { "epoch": 0.605595032328276, "grad_norm": 0.33862748742103577, "learning_rate": 2.7119989483042545e-05, "loss": 0.529, "num_input_tokens_seen": 25631920, "step": 18920 }, { "epoch": 0.6057550732987644, "grad_norm": 1.2909302711486816, "learning_rate": 2.7110207205058768e-05, "loss": 0.6358, "num_input_tokens_seen": 25638784, "step": 18925 }, { "epoch": 0.6059151142692529, "grad_norm": 1.1537878513336182, "learning_rate": 2.7100424601654517e-05, "loss": 0.3396, "num_input_tokens_seen": 25645536, "step": 18930 }, { "epoch": 0.6060751552397414, "grad_norm": 0.7206286191940308, "learning_rate": 2.7090641674338403e-05, "loss": 0.4265, "num_input_tokens_seen": 25652400, "step": 18935 }, { "epoch": 0.6062351962102298, "grad_norm": 1.1685107946395874, "learning_rate": 2.7080858424619072e-05, "loss": 0.3993, "num_input_tokens_seen": 25659248, "step": 18940 }, { "epoch": 0.6063952371807183, "grad_norm": 1.0122935771942139, "learning_rate": 2.707107485400521e-05, "loss": 0.5596, "num_input_tokens_seen": 25666128, "step": 18945 }, { "epoch": 0.6065552781512067, "grad_norm": 0.825553297996521, "learning_rate": 2.7061290964005586e-05, "loss": 0.3652, "num_input_tokens_seen": 25672976, "step": 18950 }, { "epoch": 0.6067153191216952, "grad_norm": 0.9771496057510376, "learning_rate": 2.7051506756129e-05, "loss": 0.5431, "num_input_tokens_seen": 25679552, "step": 18955 }, { "epoch": 0.6068753600921836, "grad_norm": 0.8399747610092163, "learning_rate": 2.704172223188428e-05, "loss": 0.4555, "num_input_tokens_seen": 25686864, "step": 18960 }, { "epoch": 0.6070354010626721, "grad_norm": 1.6616089344024658, "learning_rate": 2.7031937392780334e-05, "loss": 0.4267, "num_input_tokens_seen": 25693472, "step": 18965 }, { "epoch": 0.6071954420331604, "grad_norm": 1.2497731447219849, "learning_rate": 2.702215224032611e-05, "loss": 0.6346, "num_input_tokens_seen": 25700976, "step": 18970 }, { "epoch": 0.6073554830036489, "grad_norm": 1.4080023765563965, "learning_rate": 2.70123667760306e-05, "loss": 0.3668, "num_input_tokens_seen": 25707792, "step": 18975 }, { "epoch": 0.6075155239741373, "grad_norm": 0.9106693863868713, "learning_rate": 2.7002581001402845e-05, "loss": 0.4366, "num_input_tokens_seen": 25714768, "step": 18980 }, { "epoch": 0.6076755649446258, "grad_norm": 0.7297866940498352, "learning_rate": 2.6992794917951923e-05, "loss": 0.5023, "num_input_tokens_seen": 25721248, "step": 18985 }, { "epoch": 0.6078356059151143, "grad_norm": 1.3915965557098389, "learning_rate": 2.6983008527187e-05, "loss": 0.4866, "num_input_tokens_seen": 25728320, "step": 18990 }, { "epoch": 0.6079956468856027, "grad_norm": 0.8520409464836121, "learning_rate": 2.697322183061723e-05, "loss": 0.4989, "num_input_tokens_seen": 25734704, "step": 18995 }, { "epoch": 0.6081556878560912, "grad_norm": 1.165097951889038, "learning_rate": 2.696343482975186e-05, "loss": 0.3631, "num_input_tokens_seen": 25741776, "step": 19000 }, { "epoch": 0.6081556878560912, "eval_loss": 0.46407023072242737, "eval_runtime": 502.4443, "eval_samples_per_second": 27.637, "eval_steps_per_second": 13.818, "num_input_tokens_seen": 25741776, "step": 19000 }, { "epoch": 0.6083157288265796, "grad_norm": 0.5579731464385986, "learning_rate": 2.695364752610016e-05, "loss": 0.3492, "num_input_tokens_seen": 25748336, "step": 19005 }, { "epoch": 0.6084757697970681, "grad_norm": 1.3559132814407349, "learning_rate": 2.6943859921171467e-05, "loss": 0.5479, "num_input_tokens_seen": 25755024, "step": 19010 }, { "epoch": 0.6086358107675565, "grad_norm": 1.5926709175109863, "learning_rate": 2.6934072016475143e-05, "loss": 0.6446, "num_input_tokens_seen": 25762512, "step": 19015 }, { "epoch": 0.608795851738045, "grad_norm": 0.9719430208206177, "learning_rate": 2.6924283813520606e-05, "loss": 0.4845, "num_input_tokens_seen": 25769808, "step": 19020 }, { "epoch": 0.6089558927085333, "grad_norm": 1.7029434442520142, "learning_rate": 2.691449531381733e-05, "loss": 0.4135, "num_input_tokens_seen": 25776656, "step": 19025 }, { "epoch": 0.6091159336790218, "grad_norm": 1.1391280889511108, "learning_rate": 2.6904706518874816e-05, "loss": 0.5085, "num_input_tokens_seen": 25783248, "step": 19030 }, { "epoch": 0.6092759746495102, "grad_norm": 1.7234108448028564, "learning_rate": 2.6894917430202615e-05, "loss": 0.4129, "num_input_tokens_seen": 25789984, "step": 19035 }, { "epoch": 0.6094360156199987, "grad_norm": 0.6335171461105347, "learning_rate": 2.6885128049310343e-05, "loss": 0.5242, "num_input_tokens_seen": 25796704, "step": 19040 }, { "epoch": 0.6095960565904872, "grad_norm": 0.47420674562454224, "learning_rate": 2.687533837770762e-05, "loss": 0.4335, "num_input_tokens_seen": 25803472, "step": 19045 }, { "epoch": 0.6097560975609756, "grad_norm": 1.12116539478302, "learning_rate": 2.6865548416904162e-05, "loss": 0.3852, "num_input_tokens_seen": 25810384, "step": 19050 }, { "epoch": 0.6099161385314641, "grad_norm": 4.678715229034424, "learning_rate": 2.68557581684097e-05, "loss": 0.3195, "num_input_tokens_seen": 25817248, "step": 19055 }, { "epoch": 0.6100761795019525, "grad_norm": 1.431725025177002, "learning_rate": 2.6845967633733998e-05, "loss": 0.3169, "num_input_tokens_seen": 25824000, "step": 19060 }, { "epoch": 0.610236220472441, "grad_norm": 0.685271680355072, "learning_rate": 2.683617681438689e-05, "loss": 0.3224, "num_input_tokens_seen": 25830624, "step": 19065 }, { "epoch": 0.6103962614429294, "grad_norm": 0.6334336996078491, "learning_rate": 2.682638571187825e-05, "loss": 0.3871, "num_input_tokens_seen": 25837488, "step": 19070 }, { "epoch": 0.6105563024134179, "grad_norm": 1.0490221977233887, "learning_rate": 2.6816594327717976e-05, "loss": 0.3574, "num_input_tokens_seen": 25843984, "step": 19075 }, { "epoch": 0.6107163433839062, "grad_norm": 2.7271947860717773, "learning_rate": 2.680680266341603e-05, "loss": 0.4036, "num_input_tokens_seen": 25850752, "step": 19080 }, { "epoch": 0.6108763843543947, "grad_norm": 1.4213489294052124, "learning_rate": 2.67970107204824e-05, "loss": 0.3514, "num_input_tokens_seen": 25857296, "step": 19085 }, { "epoch": 0.6110364253248832, "grad_norm": 2.3915696144104004, "learning_rate": 2.6787218500427142e-05, "loss": 0.5541, "num_input_tokens_seen": 25864368, "step": 19090 }, { "epoch": 0.6111964662953716, "grad_norm": 0.727970540523529, "learning_rate": 2.6777426004760332e-05, "loss": 0.4125, "num_input_tokens_seen": 25871744, "step": 19095 }, { "epoch": 0.6113565072658601, "grad_norm": 1.420711636543274, "learning_rate": 2.6767633234992094e-05, "loss": 0.5123, "num_input_tokens_seen": 25878992, "step": 19100 }, { "epoch": 0.6115165482363485, "grad_norm": 1.1944373846054077, "learning_rate": 2.6757840192632598e-05, "loss": 0.3895, "num_input_tokens_seen": 25885872, "step": 19105 }, { "epoch": 0.611676589206837, "grad_norm": 0.8834598064422607, "learning_rate": 2.6748046879192052e-05, "loss": 0.3454, "num_input_tokens_seen": 25892608, "step": 19110 }, { "epoch": 0.6118366301773254, "grad_norm": 1.0825512409210205, "learning_rate": 2.673825329618071e-05, "loss": 0.5004, "num_input_tokens_seen": 25899344, "step": 19115 }, { "epoch": 0.6119966711478139, "grad_norm": 1.6629574298858643, "learning_rate": 2.6728459445108866e-05, "loss": 0.4601, "num_input_tokens_seen": 25906400, "step": 19120 }, { "epoch": 0.6121567121183022, "grad_norm": 1.187004804611206, "learning_rate": 2.6718665327486854e-05, "loss": 0.4691, "num_input_tokens_seen": 25913472, "step": 19125 }, { "epoch": 0.6123167530887907, "grad_norm": 2.069305896759033, "learning_rate": 2.6708870944825048e-05, "loss": 0.4864, "num_input_tokens_seen": 25920448, "step": 19130 }, { "epoch": 0.6124767940592791, "grad_norm": 1.2762442827224731, "learning_rate": 2.6699076298633874e-05, "loss": 0.5042, "num_input_tokens_seen": 25927408, "step": 19135 }, { "epoch": 0.6126368350297676, "grad_norm": 1.0459587574005127, "learning_rate": 2.6689281390423788e-05, "loss": 0.3803, "num_input_tokens_seen": 25934480, "step": 19140 }, { "epoch": 0.6127968760002561, "grad_norm": 1.4500843286514282, "learning_rate": 2.667948622170527e-05, "loss": 0.4451, "num_input_tokens_seen": 25941424, "step": 19145 }, { "epoch": 0.6129569169707445, "grad_norm": 0.8355640769004822, "learning_rate": 2.6669690793988873e-05, "loss": 0.3915, "num_input_tokens_seen": 25947968, "step": 19150 }, { "epoch": 0.613116957941233, "grad_norm": 1.897025465965271, "learning_rate": 2.665989510878518e-05, "loss": 0.5341, "num_input_tokens_seen": 25954816, "step": 19155 }, { "epoch": 0.6132769989117214, "grad_norm": 1.3352255821228027, "learning_rate": 2.6650099167604793e-05, "loss": 0.5068, "num_input_tokens_seen": 25962128, "step": 19160 }, { "epoch": 0.6134370398822099, "grad_norm": 1.3687583208084106, "learning_rate": 2.6640302971958376e-05, "loss": 0.4565, "num_input_tokens_seen": 25968352, "step": 19165 }, { "epoch": 0.6135970808526983, "grad_norm": 1.6482281684875488, "learning_rate": 2.6630506523356635e-05, "loss": 0.4849, "num_input_tokens_seen": 25975104, "step": 19170 }, { "epoch": 0.6137571218231868, "grad_norm": 1.5710700750350952, "learning_rate": 2.6620709823310297e-05, "loss": 0.5724, "num_input_tokens_seen": 25982816, "step": 19175 }, { "epoch": 0.6139171627936751, "grad_norm": 1.5019547939300537, "learning_rate": 2.661091287333014e-05, "loss": 0.3017, "num_input_tokens_seen": 25989520, "step": 19180 }, { "epoch": 0.6140772037641636, "grad_norm": 0.7536382675170898, "learning_rate": 2.660111567492696e-05, "loss": 0.5653, "num_input_tokens_seen": 25996560, "step": 19185 }, { "epoch": 0.614237244734652, "grad_norm": 0.97004634141922, "learning_rate": 2.6591318229611635e-05, "loss": 0.2861, "num_input_tokens_seen": 26003152, "step": 19190 }, { "epoch": 0.6143972857051405, "grad_norm": 0.4902622401714325, "learning_rate": 2.6581520538895037e-05, "loss": 0.5669, "num_input_tokens_seen": 26010080, "step": 19195 }, { "epoch": 0.614557326675629, "grad_norm": 2.792673110961914, "learning_rate": 2.6571722604288102e-05, "loss": 0.6834, "num_input_tokens_seen": 26017088, "step": 19200 }, { "epoch": 0.614557326675629, "eval_loss": 0.4636697471141815, "eval_runtime": 502.4823, "eval_samples_per_second": 27.635, "eval_steps_per_second": 13.817, "num_input_tokens_seen": 26017088, "step": 19200 }, { "epoch": 0.6147173676461174, "grad_norm": 1.2518984079360962, "learning_rate": 2.656192442730179e-05, "loss": 0.5607, "num_input_tokens_seen": 26023952, "step": 19205 }, { "epoch": 0.6148774086166059, "grad_norm": 2.411432981491089, "learning_rate": 2.6552126009447098e-05, "loss": 0.4832, "num_input_tokens_seen": 26030848, "step": 19210 }, { "epoch": 0.6150374495870943, "grad_norm": 1.9856972694396973, "learning_rate": 2.654232735223507e-05, "loss": 0.5099, "num_input_tokens_seen": 26037760, "step": 19215 }, { "epoch": 0.6151974905575828, "grad_norm": 1.287429928779602, "learning_rate": 2.6532528457176787e-05, "loss": 0.4467, "num_input_tokens_seen": 26044608, "step": 19220 }, { "epoch": 0.6153575315280712, "grad_norm": 2.185647964477539, "learning_rate": 2.6522729325783348e-05, "loss": 0.5376, "num_input_tokens_seen": 26051440, "step": 19225 }, { "epoch": 0.6155175724985597, "grad_norm": 0.9052348732948303, "learning_rate": 2.6512929959565914e-05, "loss": 0.431, "num_input_tokens_seen": 26057776, "step": 19230 }, { "epoch": 0.615677613469048, "grad_norm": 1.1728533506393433, "learning_rate": 2.6503130360035673e-05, "loss": 0.6323, "num_input_tokens_seen": 26064368, "step": 19235 }, { "epoch": 0.6158376544395365, "grad_norm": 2.6535251140594482, "learning_rate": 2.6493330528703835e-05, "loss": 0.4102, "num_input_tokens_seen": 26070800, "step": 19240 }, { "epoch": 0.6159976954100249, "grad_norm": 1.4450148344039917, "learning_rate": 2.648353046708167e-05, "loss": 0.495, "num_input_tokens_seen": 26077520, "step": 19245 }, { "epoch": 0.6161577363805134, "grad_norm": 0.9935912489891052, "learning_rate": 2.647373017668046e-05, "loss": 0.627, "num_input_tokens_seen": 26084000, "step": 19250 }, { "epoch": 0.6163177773510019, "grad_norm": 1.6232712268829346, "learning_rate": 2.6463929659011537e-05, "loss": 0.5563, "num_input_tokens_seen": 26090832, "step": 19255 }, { "epoch": 0.6164778183214903, "grad_norm": 0.9919853806495667, "learning_rate": 2.6454128915586262e-05, "loss": 0.3765, "num_input_tokens_seen": 26097440, "step": 19260 }, { "epoch": 0.6166378592919788, "grad_norm": 2.2183609008789062, "learning_rate": 2.6444327947916036e-05, "loss": 0.4472, "num_input_tokens_seen": 26104288, "step": 19265 }, { "epoch": 0.6167979002624672, "grad_norm": 1.178234338760376, "learning_rate": 2.6434526757512292e-05, "loss": 0.4363, "num_input_tokens_seen": 26110864, "step": 19270 }, { "epoch": 0.6169579412329557, "grad_norm": 0.9137439131736755, "learning_rate": 2.6424725345886486e-05, "loss": 0.4338, "num_input_tokens_seen": 26117536, "step": 19275 }, { "epoch": 0.617117982203444, "grad_norm": 1.304365873336792, "learning_rate": 2.641492371455014e-05, "loss": 0.5206, "num_input_tokens_seen": 26124592, "step": 19280 }, { "epoch": 0.6172780231739325, "grad_norm": 0.9110331535339355, "learning_rate": 2.640512186501477e-05, "loss": 0.4275, "num_input_tokens_seen": 26131008, "step": 19285 }, { "epoch": 0.6174380641444209, "grad_norm": 0.8513411283493042, "learning_rate": 2.639531979879195e-05, "loss": 0.2999, "num_input_tokens_seen": 26137600, "step": 19290 }, { "epoch": 0.6175981051149094, "grad_norm": 1.1622071266174316, "learning_rate": 2.638551751739328e-05, "loss": 0.5218, "num_input_tokens_seen": 26144064, "step": 19295 }, { "epoch": 0.6177581460853978, "grad_norm": 0.5864351391792297, "learning_rate": 2.6375715022330404e-05, "loss": 0.5669, "num_input_tokens_seen": 26150528, "step": 19300 }, { "epoch": 0.6179181870558863, "grad_norm": 1.4701262712478638, "learning_rate": 2.6365912315114976e-05, "loss": 0.3411, "num_input_tokens_seen": 26157680, "step": 19305 }, { "epoch": 0.6180782280263748, "grad_norm": 2.584657669067383, "learning_rate": 2.6356109397258704e-05, "loss": 0.5501, "num_input_tokens_seen": 26164816, "step": 19310 }, { "epoch": 0.6182382689968632, "grad_norm": 0.9865561723709106, "learning_rate": 2.6346306270273325e-05, "loss": 0.5607, "num_input_tokens_seen": 26172224, "step": 19315 }, { "epoch": 0.6183983099673517, "grad_norm": 0.8413369059562683, "learning_rate": 2.6336502935670608e-05, "loss": 0.5161, "num_input_tokens_seen": 26179168, "step": 19320 }, { "epoch": 0.6185583509378401, "grad_norm": 2.01426362991333, "learning_rate": 2.6326699394962333e-05, "loss": 0.6361, "num_input_tokens_seen": 26186336, "step": 19325 }, { "epoch": 0.6187183919083286, "grad_norm": 1.47652006149292, "learning_rate": 2.6316895649660334e-05, "loss": 0.4602, "num_input_tokens_seen": 26193120, "step": 19330 }, { "epoch": 0.6188784328788169, "grad_norm": 1.337424397468567, "learning_rate": 2.6307091701276486e-05, "loss": 0.4197, "num_input_tokens_seen": 26200368, "step": 19335 }, { "epoch": 0.6190384738493054, "grad_norm": 0.8963878154754639, "learning_rate": 2.629728755132267e-05, "loss": 0.3624, "num_input_tokens_seen": 26207152, "step": 19340 }, { "epoch": 0.6191985148197938, "grad_norm": 1.0921396017074585, "learning_rate": 2.628748320131081e-05, "loss": 0.5027, "num_input_tokens_seen": 26214272, "step": 19345 }, { "epoch": 0.6193585557902823, "grad_norm": 1.6084487438201904, "learning_rate": 2.6277678652752856e-05, "loss": 0.6621, "num_input_tokens_seen": 26221312, "step": 19350 }, { "epoch": 0.6195185967607708, "grad_norm": 1.7490733861923218, "learning_rate": 2.6267873907160807e-05, "loss": 0.4281, "num_input_tokens_seen": 26227920, "step": 19355 }, { "epoch": 0.6196786377312592, "grad_norm": 0.8270451426506042, "learning_rate": 2.6258068966046668e-05, "loss": 0.4495, "num_input_tokens_seen": 26234432, "step": 19360 }, { "epoch": 0.6198386787017477, "grad_norm": 1.475791096687317, "learning_rate": 2.6248263830922475e-05, "loss": 0.7498, "num_input_tokens_seen": 26240752, "step": 19365 }, { "epoch": 0.6199987196722361, "grad_norm": 0.8762374520301819, "learning_rate": 2.6238458503300318e-05, "loss": 0.4029, "num_input_tokens_seen": 26247056, "step": 19370 }, { "epoch": 0.6201587606427246, "grad_norm": 1.2544633150100708, "learning_rate": 2.6228652984692292e-05, "loss": 0.4257, "num_input_tokens_seen": 26253568, "step": 19375 }, { "epoch": 0.620318801613213, "grad_norm": 1.3706837892532349, "learning_rate": 2.621884727661054e-05, "loss": 0.3961, "num_input_tokens_seen": 26260112, "step": 19380 }, { "epoch": 0.6204788425837015, "grad_norm": 1.3625080585479736, "learning_rate": 2.6209041380567222e-05, "loss": 0.6498, "num_input_tokens_seen": 26266656, "step": 19385 }, { "epoch": 0.6206388835541898, "grad_norm": 0.6594556570053101, "learning_rate": 2.6199235298074527e-05, "loss": 0.4649, "num_input_tokens_seen": 26273280, "step": 19390 }, { "epoch": 0.6207989245246783, "grad_norm": 0.9726179242134094, "learning_rate": 2.618942903064468e-05, "loss": 0.2809, "num_input_tokens_seen": 26279808, "step": 19395 }, { "epoch": 0.6209589654951667, "grad_norm": 0.3949177861213684, "learning_rate": 2.6179622579789932e-05, "loss": 0.3855, "num_input_tokens_seen": 26286480, "step": 19400 }, { "epoch": 0.6209589654951667, "eval_loss": 0.46274009346961975, "eval_runtime": 502.6682, "eval_samples_per_second": 27.625, "eval_steps_per_second": 13.812, "num_input_tokens_seen": 26286480, "step": 19400 }, { "epoch": 0.6211190064656552, "grad_norm": 1.276833415031433, "learning_rate": 2.6169815947022553e-05, "loss": 0.5894, "num_input_tokens_seen": 26293328, "step": 19405 }, { "epoch": 0.6212790474361437, "grad_norm": 1.377231240272522, "learning_rate": 2.6160009133854853e-05, "loss": 0.3274, "num_input_tokens_seen": 26300496, "step": 19410 }, { "epoch": 0.6214390884066321, "grad_norm": 0.7963985204696655, "learning_rate": 2.6150202141799168e-05, "loss": 0.3055, "num_input_tokens_seen": 26307200, "step": 19415 }, { "epoch": 0.6215991293771206, "grad_norm": 1.8815661668777466, "learning_rate": 2.614039497236786e-05, "loss": 0.5324, "num_input_tokens_seen": 26314000, "step": 19420 }, { "epoch": 0.621759170347609, "grad_norm": 0.7305392622947693, "learning_rate": 2.6130587627073315e-05, "loss": 0.4626, "num_input_tokens_seen": 26320800, "step": 19425 }, { "epoch": 0.6219192113180975, "grad_norm": 0.7659434080123901, "learning_rate": 2.6120780107427956e-05, "loss": 0.3076, "num_input_tokens_seen": 26327712, "step": 19430 }, { "epoch": 0.6220792522885858, "grad_norm": 5.0824408531188965, "learning_rate": 2.6110972414944214e-05, "loss": 0.4436, "num_input_tokens_seen": 26334432, "step": 19435 }, { "epoch": 0.6222392932590743, "grad_norm": 0.5677497982978821, "learning_rate": 2.6101164551134565e-05, "loss": 0.4205, "num_input_tokens_seen": 26341056, "step": 19440 }, { "epoch": 0.6223993342295627, "grad_norm": 1.188108205795288, "learning_rate": 2.6091356517511505e-05, "loss": 0.4186, "num_input_tokens_seen": 26347408, "step": 19445 }, { "epoch": 0.6225593752000512, "grad_norm": 0.7368975281715393, "learning_rate": 2.608154831558755e-05, "loss": 0.4284, "num_input_tokens_seen": 26354256, "step": 19450 }, { "epoch": 0.6227194161705396, "grad_norm": 1.115251064300537, "learning_rate": 2.607173994687526e-05, "loss": 0.6317, "num_input_tokens_seen": 26361024, "step": 19455 }, { "epoch": 0.6228794571410281, "grad_norm": 0.6924639344215393, "learning_rate": 2.6061931412887196e-05, "loss": 0.4868, "num_input_tokens_seen": 26367488, "step": 19460 }, { "epoch": 0.6230394981115166, "grad_norm": 0.7413017153739929, "learning_rate": 2.6052122715135973e-05, "loss": 0.3092, "num_input_tokens_seen": 26374416, "step": 19465 }, { "epoch": 0.623199539082005, "grad_norm": 2.9270517826080322, "learning_rate": 2.60423138551342e-05, "loss": 0.5593, "num_input_tokens_seen": 26381264, "step": 19470 }, { "epoch": 0.6233595800524935, "grad_norm": 0.6721493601799011, "learning_rate": 2.6032504834394527e-05, "loss": 0.4023, "num_input_tokens_seen": 26388096, "step": 19475 }, { "epoch": 0.6235196210229819, "grad_norm": 1.423473596572876, "learning_rate": 2.602269565442964e-05, "loss": 0.4372, "num_input_tokens_seen": 26394864, "step": 19480 }, { "epoch": 0.6236796619934704, "grad_norm": 0.8739303350448608, "learning_rate": 2.6012886316752227e-05, "loss": 0.5022, "num_input_tokens_seen": 26401456, "step": 19485 }, { "epoch": 0.6238397029639587, "grad_norm": 0.6765075922012329, "learning_rate": 2.6003076822875018e-05, "loss": 0.447, "num_input_tokens_seen": 26408160, "step": 19490 }, { "epoch": 0.6239997439344472, "grad_norm": 1.64859139919281, "learning_rate": 2.5993267174310755e-05, "loss": 0.4663, "num_input_tokens_seen": 26414608, "step": 19495 }, { "epoch": 0.6241597849049356, "grad_norm": 1.0251209735870361, "learning_rate": 2.5983457372572218e-05, "loss": 0.4449, "num_input_tokens_seen": 26421248, "step": 19500 }, { "epoch": 0.6243198258754241, "grad_norm": 0.9135609865188599, "learning_rate": 2.597364741917219e-05, "loss": 0.4458, "num_input_tokens_seen": 26428192, "step": 19505 }, { "epoch": 0.6244798668459125, "grad_norm": 0.899238109588623, "learning_rate": 2.5963837315623492e-05, "loss": 0.6427, "num_input_tokens_seen": 26434384, "step": 19510 }, { "epoch": 0.624639907816401, "grad_norm": 0.7622813582420349, "learning_rate": 2.595402706343897e-05, "loss": 0.289, "num_input_tokens_seen": 26441360, "step": 19515 }, { "epoch": 0.6247999487868895, "grad_norm": 1.6332199573516846, "learning_rate": 2.594421666413148e-05, "loss": 0.6122, "num_input_tokens_seen": 26447664, "step": 19520 }, { "epoch": 0.6249599897573779, "grad_norm": 0.8093807101249695, "learning_rate": 2.5934406119213928e-05, "loss": 0.34, "num_input_tokens_seen": 26454992, "step": 19525 }, { "epoch": 0.6251200307278664, "grad_norm": 1.1147756576538086, "learning_rate": 2.5924595430199193e-05, "loss": 0.4493, "num_input_tokens_seen": 26461920, "step": 19530 }, { "epoch": 0.6252800716983548, "grad_norm": 0.6115877628326416, "learning_rate": 2.5914784598600238e-05, "loss": 0.3814, "num_input_tokens_seen": 26468608, "step": 19535 }, { "epoch": 0.6254401126688433, "grad_norm": 3.769653797149658, "learning_rate": 2.5904973625930002e-05, "loss": 0.4816, "num_input_tokens_seen": 26476064, "step": 19540 }, { "epoch": 0.6256001536393316, "grad_norm": 1.2651067972183228, "learning_rate": 2.5895162513701456e-05, "loss": 0.3828, "num_input_tokens_seen": 26482560, "step": 19545 }, { "epoch": 0.6257601946098201, "grad_norm": 0.7219253778457642, "learning_rate": 2.5885351263427593e-05, "loss": 0.2634, "num_input_tokens_seen": 26489584, "step": 19550 }, { "epoch": 0.6259202355803085, "grad_norm": 0.4908851981163025, "learning_rate": 2.5875539876621448e-05, "loss": 0.4076, "num_input_tokens_seen": 26496480, "step": 19555 }, { "epoch": 0.626080276550797, "grad_norm": 0.7463430762290955, "learning_rate": 2.586572835479605e-05, "loss": 0.3236, "num_input_tokens_seen": 26502944, "step": 19560 }, { "epoch": 0.6262403175212854, "grad_norm": 0.9161614179611206, "learning_rate": 2.585591669946446e-05, "loss": 0.3309, "num_input_tokens_seen": 26509776, "step": 19565 }, { "epoch": 0.6264003584917739, "grad_norm": 0.6861354112625122, "learning_rate": 2.5846104912139756e-05, "loss": 0.4305, "num_input_tokens_seen": 26516736, "step": 19570 }, { "epoch": 0.6265603994622624, "grad_norm": 0.6225859522819519, "learning_rate": 2.583629299433505e-05, "loss": 0.3059, "num_input_tokens_seen": 26523408, "step": 19575 }, { "epoch": 0.6267204404327508, "grad_norm": 0.9053742289543152, "learning_rate": 2.582648094756345e-05, "loss": 0.4608, "num_input_tokens_seen": 26529776, "step": 19580 }, { "epoch": 0.6268804814032393, "grad_norm": 0.8995998501777649, "learning_rate": 2.5816668773338098e-05, "loss": 0.464, "num_input_tokens_seen": 26536528, "step": 19585 }, { "epoch": 0.6270405223737276, "grad_norm": 1.2825632095336914, "learning_rate": 2.580685647317216e-05, "loss": 0.4982, "num_input_tokens_seen": 26543520, "step": 19590 }, { "epoch": 0.6272005633442161, "grad_norm": 0.6317079663276672, "learning_rate": 2.5797044048578818e-05, "loss": 0.2293, "num_input_tokens_seen": 26550352, "step": 19595 }, { "epoch": 0.6273606043147045, "grad_norm": 2.3113651275634766, "learning_rate": 2.5787231501071262e-05, "loss": 0.4292, "num_input_tokens_seen": 26557200, "step": 19600 }, { "epoch": 0.6273606043147045, "eval_loss": 0.46376025676727295, "eval_runtime": 502.4386, "eval_samples_per_second": 27.637, "eval_steps_per_second": 13.819, "num_input_tokens_seen": 26557200, "step": 19600 }, { "epoch": 0.627520645285193, "grad_norm": 1.1186703443527222, "learning_rate": 2.577741883216272e-05, "loss": 0.5463, "num_input_tokens_seen": 26563824, "step": 19605 }, { "epoch": 0.6276806862556814, "grad_norm": 1.855559229850769, "learning_rate": 2.576760604336642e-05, "loss": 0.5383, "num_input_tokens_seen": 26570384, "step": 19610 }, { "epoch": 0.6278407272261699, "grad_norm": 1.0081326961517334, "learning_rate": 2.575779313619563e-05, "loss": 0.5492, "num_input_tokens_seen": 26576816, "step": 19615 }, { "epoch": 0.6280007681966584, "grad_norm": 1.0228947401046753, "learning_rate": 2.5747980112163605e-05, "loss": 0.4141, "num_input_tokens_seen": 26583360, "step": 19620 }, { "epoch": 0.6281608091671468, "grad_norm": 1.924477219581604, "learning_rate": 2.5738166972783656e-05, "loss": 0.3939, "num_input_tokens_seen": 26590160, "step": 19625 }, { "epoch": 0.6283208501376353, "grad_norm": 0.9741834402084351, "learning_rate": 2.5728353719569075e-05, "loss": 0.56, "num_input_tokens_seen": 26597168, "step": 19630 }, { "epoch": 0.6284808911081237, "grad_norm": 0.515294075012207, "learning_rate": 2.57185403540332e-05, "loss": 0.4659, "num_input_tokens_seen": 26603856, "step": 19635 }, { "epoch": 0.6286409320786122, "grad_norm": 0.49492838978767395, "learning_rate": 2.5708726877689375e-05, "loss": 0.619, "num_input_tokens_seen": 26610368, "step": 19640 }, { "epoch": 0.6288009730491005, "grad_norm": 1.0940935611724854, "learning_rate": 2.5698913292050964e-05, "loss": 0.4168, "num_input_tokens_seen": 26616928, "step": 19645 }, { "epoch": 0.628961014019589, "grad_norm": 2.253431558609009, "learning_rate": 2.568909959863133e-05, "loss": 0.4573, "num_input_tokens_seen": 26623600, "step": 19650 }, { "epoch": 0.6291210549900774, "grad_norm": 0.8563117980957031, "learning_rate": 2.5679285798943887e-05, "loss": 0.3577, "num_input_tokens_seen": 26630560, "step": 19655 }, { "epoch": 0.6292810959605659, "grad_norm": 1.863661289215088, "learning_rate": 2.5669471894502035e-05, "loss": 0.4992, "num_input_tokens_seen": 26636704, "step": 19660 }, { "epoch": 0.6294411369310543, "grad_norm": 0.9890412092208862, "learning_rate": 2.56596578868192e-05, "loss": 0.4061, "num_input_tokens_seen": 26643344, "step": 19665 }, { "epoch": 0.6296011779015428, "grad_norm": 1.6476123332977295, "learning_rate": 2.564984377740883e-05, "loss": 0.6522, "num_input_tokens_seen": 26650400, "step": 19670 }, { "epoch": 0.6297612188720313, "grad_norm": 1.3420590162277222, "learning_rate": 2.564002956778438e-05, "loss": 0.3342, "num_input_tokens_seen": 26657104, "step": 19675 }, { "epoch": 0.6299212598425197, "grad_norm": 0.8274680376052856, "learning_rate": 2.563021525945934e-05, "loss": 0.4712, "num_input_tokens_seen": 26663552, "step": 19680 }, { "epoch": 0.6300813008130082, "grad_norm": 1.456097960472107, "learning_rate": 2.562040085394718e-05, "loss": 0.4247, "num_input_tokens_seen": 26670048, "step": 19685 }, { "epoch": 0.6302413417834966, "grad_norm": 1.187318205833435, "learning_rate": 2.56105863527614e-05, "loss": 0.3792, "num_input_tokens_seen": 26676848, "step": 19690 }, { "epoch": 0.630401382753985, "grad_norm": 0.9778875708580017, "learning_rate": 2.5600771757415548e-05, "loss": 0.5101, "num_input_tokens_seen": 26683568, "step": 19695 }, { "epoch": 0.6305614237244734, "grad_norm": 1.0063872337341309, "learning_rate": 2.5590957069423134e-05, "loss": 0.357, "num_input_tokens_seen": 26690160, "step": 19700 }, { "epoch": 0.6307214646949619, "grad_norm": 0.9051551818847656, "learning_rate": 2.5581142290297716e-05, "loss": 0.2447, "num_input_tokens_seen": 26696896, "step": 19705 }, { "epoch": 0.6308815056654503, "grad_norm": 0.5703232884407043, "learning_rate": 2.557132742155285e-05, "loss": 0.4173, "num_input_tokens_seen": 26703520, "step": 19710 }, { "epoch": 0.6310415466359388, "grad_norm": 1.3363605737686157, "learning_rate": 2.556151246470212e-05, "loss": 0.4245, "num_input_tokens_seen": 26710640, "step": 19715 }, { "epoch": 0.6312015876064272, "grad_norm": 1.871976375579834, "learning_rate": 2.5551697421259114e-05, "loss": 0.4987, "num_input_tokens_seen": 26717536, "step": 19720 }, { "epoch": 0.6313616285769157, "grad_norm": 0.6450322866439819, "learning_rate": 2.554188229273743e-05, "loss": 0.4996, "num_input_tokens_seen": 26724368, "step": 19725 }, { "epoch": 0.6315216695474042, "grad_norm": 0.977324903011322, "learning_rate": 2.5532067080650678e-05, "loss": 0.4441, "num_input_tokens_seen": 26731168, "step": 19730 }, { "epoch": 0.6316817105178926, "grad_norm": 2.0576136112213135, "learning_rate": 2.55222517865125e-05, "loss": 0.4209, "num_input_tokens_seen": 26738160, "step": 19735 }, { "epoch": 0.6318417514883811, "grad_norm": 1.5598218441009521, "learning_rate": 2.5512436411836538e-05, "loss": 0.3102, "num_input_tokens_seen": 26745840, "step": 19740 }, { "epoch": 0.6320017924588694, "grad_norm": 0.9182868003845215, "learning_rate": 2.5502620958136443e-05, "loss": 0.4418, "num_input_tokens_seen": 26752448, "step": 19745 }, { "epoch": 0.6321618334293579, "grad_norm": 1.3303121328353882, "learning_rate": 2.5492805426925874e-05, "loss": 0.4124, "num_input_tokens_seen": 26759088, "step": 19750 }, { "epoch": 0.6323218743998463, "grad_norm": 0.9293755888938904, "learning_rate": 2.5482989819718523e-05, "loss": 0.4293, "num_input_tokens_seen": 26766288, "step": 19755 }, { "epoch": 0.6324819153703348, "grad_norm": 2.638044834136963, "learning_rate": 2.5473174138028065e-05, "loss": 0.3864, "num_input_tokens_seen": 26773024, "step": 19760 }, { "epoch": 0.6326419563408232, "grad_norm": 0.6385533213615417, "learning_rate": 2.5463358383368212e-05, "loss": 0.4117, "num_input_tokens_seen": 26779552, "step": 19765 }, { "epoch": 0.6328019973113117, "grad_norm": 0.40871334075927734, "learning_rate": 2.545354255725267e-05, "loss": 0.4919, "num_input_tokens_seen": 26786272, "step": 19770 }, { "epoch": 0.6329620382818001, "grad_norm": 1.4573328495025635, "learning_rate": 2.5443726661195165e-05, "loss": 0.5657, "num_input_tokens_seen": 26792992, "step": 19775 }, { "epoch": 0.6331220792522886, "grad_norm": 0.37001729011535645, "learning_rate": 2.543391069670944e-05, "loss": 0.3628, "num_input_tokens_seen": 26799920, "step": 19780 }, { "epoch": 0.6332821202227771, "grad_norm": 1.0735337734222412, "learning_rate": 2.5424094665309228e-05, "loss": 0.4219, "num_input_tokens_seen": 26806800, "step": 19785 }, { "epoch": 0.6334421611932655, "grad_norm": 0.5547364354133606, "learning_rate": 2.5414278568508292e-05, "loss": 0.4743, "num_input_tokens_seen": 26813968, "step": 19790 }, { "epoch": 0.633602202163754, "grad_norm": 0.9064767360687256, "learning_rate": 2.540446240782039e-05, "loss": 0.3509, "num_input_tokens_seen": 26820928, "step": 19795 }, { "epoch": 0.6337622431342423, "grad_norm": 1.8917241096496582, "learning_rate": 2.5394646184759307e-05, "loss": 0.437, "num_input_tokens_seen": 26827696, "step": 19800 }, { "epoch": 0.6337622431342423, "eval_loss": 0.46269965171813965, "eval_runtime": 502.5819, "eval_samples_per_second": 27.629, "eval_steps_per_second": 13.815, "num_input_tokens_seen": 26827696, "step": 19800 }, { "epoch": 0.6339222841047308, "grad_norm": 1.5854829549789429, "learning_rate": 2.538482990083882e-05, "loss": 0.5039, "num_input_tokens_seen": 26834288, "step": 19805 }, { "epoch": 0.6340823250752192, "grad_norm": 0.8148331046104431, "learning_rate": 2.5375013557572725e-05, "loss": 0.2801, "num_input_tokens_seen": 26840544, "step": 19810 }, { "epoch": 0.6342423660457077, "grad_norm": 1.157702922821045, "learning_rate": 2.536519715647483e-05, "loss": 0.345, "num_input_tokens_seen": 26847456, "step": 19815 }, { "epoch": 0.6344024070161961, "grad_norm": 0.9639371633529663, "learning_rate": 2.535538069905894e-05, "loss": 0.3513, "num_input_tokens_seen": 26854080, "step": 19820 }, { "epoch": 0.6345624479866846, "grad_norm": 2.459845781326294, "learning_rate": 2.534556418683888e-05, "loss": 0.5644, "num_input_tokens_seen": 26860480, "step": 19825 }, { "epoch": 0.634722488957173, "grad_norm": 1.3581452369689941, "learning_rate": 2.5335747621328486e-05, "loss": 0.3143, "num_input_tokens_seen": 26867584, "step": 19830 }, { "epoch": 0.6348825299276615, "grad_norm": 1.7225196361541748, "learning_rate": 2.5325931004041586e-05, "loss": 0.399, "num_input_tokens_seen": 26874144, "step": 19835 }, { "epoch": 0.63504257089815, "grad_norm": 1.9507102966308594, "learning_rate": 2.5316114336492032e-05, "loss": 0.5316, "num_input_tokens_seen": 26880528, "step": 19840 }, { "epoch": 0.6352026118686384, "grad_norm": 1.582119107246399, "learning_rate": 2.530629762019367e-05, "loss": 0.4818, "num_input_tokens_seen": 26887584, "step": 19845 }, { "epoch": 0.6353626528391269, "grad_norm": 1.0967015027999878, "learning_rate": 2.5296480856660364e-05, "loss": 0.5052, "num_input_tokens_seen": 26894144, "step": 19850 }, { "epoch": 0.6355226938096152, "grad_norm": 2.0392980575561523, "learning_rate": 2.528666404740599e-05, "loss": 0.4507, "num_input_tokens_seen": 26900944, "step": 19855 }, { "epoch": 0.6356827347801037, "grad_norm": 1.6139755249023438, "learning_rate": 2.527684719394442e-05, "loss": 0.347, "num_input_tokens_seen": 26907456, "step": 19860 }, { "epoch": 0.6358427757505921, "grad_norm": 1.357455849647522, "learning_rate": 2.526703029778953e-05, "loss": 0.4389, "num_input_tokens_seen": 26914976, "step": 19865 }, { "epoch": 0.6360028167210806, "grad_norm": 0.8803106546401978, "learning_rate": 2.5257213360455208e-05, "loss": 0.3926, "num_input_tokens_seen": 26921392, "step": 19870 }, { "epoch": 0.636162857691569, "grad_norm": 1.1058928966522217, "learning_rate": 2.5247396383455353e-05, "loss": 0.6788, "num_input_tokens_seen": 26927760, "step": 19875 }, { "epoch": 0.6363228986620575, "grad_norm": 2.170840263366699, "learning_rate": 2.523757936830387e-05, "loss": 0.6234, "num_input_tokens_seen": 26934560, "step": 19880 }, { "epoch": 0.636482939632546, "grad_norm": 0.9731006622314453, "learning_rate": 2.5227762316514662e-05, "loss": 0.4796, "num_input_tokens_seen": 26941136, "step": 19885 }, { "epoch": 0.6366429806030344, "grad_norm": 0.6055054068565369, "learning_rate": 2.5217945229601648e-05, "loss": 0.5186, "num_input_tokens_seen": 26947808, "step": 19890 }, { "epoch": 0.6368030215735229, "grad_norm": 0.811434805393219, "learning_rate": 2.5208128109078738e-05, "loss": 0.405, "num_input_tokens_seen": 26954608, "step": 19895 }, { "epoch": 0.6369630625440112, "grad_norm": 0.8712993860244751, "learning_rate": 2.5198310956459853e-05, "loss": 0.4668, "num_input_tokens_seen": 26961280, "step": 19900 }, { "epoch": 0.6371231035144997, "grad_norm": 1.402013897895813, "learning_rate": 2.518849377325893e-05, "loss": 0.5439, "num_input_tokens_seen": 26968192, "step": 19905 }, { "epoch": 0.6372831444849881, "grad_norm": 1.0787690877914429, "learning_rate": 2.51786765609899e-05, "loss": 0.656, "num_input_tokens_seen": 26975024, "step": 19910 }, { "epoch": 0.6374431854554766, "grad_norm": 0.9335280656814575, "learning_rate": 2.5168859321166694e-05, "loss": 0.2939, "num_input_tokens_seen": 26981616, "step": 19915 }, { "epoch": 0.637603226425965, "grad_norm": 1.052337884902954, "learning_rate": 2.515904205530326e-05, "loss": 0.43, "num_input_tokens_seen": 26988592, "step": 19920 }, { "epoch": 0.6377632673964535, "grad_norm": 1.113929271697998, "learning_rate": 2.514922476491355e-05, "loss": 0.3532, "num_input_tokens_seen": 26995264, "step": 19925 }, { "epoch": 0.6379233083669419, "grad_norm": 0.7300124764442444, "learning_rate": 2.51394074515115e-05, "loss": 0.2938, "num_input_tokens_seen": 27002128, "step": 19930 }, { "epoch": 0.6380833493374304, "grad_norm": 0.8457174301147461, "learning_rate": 2.5129590116611067e-05, "loss": 0.5584, "num_input_tokens_seen": 27009120, "step": 19935 }, { "epoch": 0.6382433903079189, "grad_norm": 1.1566333770751953, "learning_rate": 2.5119772761726212e-05, "loss": 0.4201, "num_input_tokens_seen": 27015824, "step": 19940 }, { "epoch": 0.6384034312784073, "grad_norm": 0.8028162121772766, "learning_rate": 2.5109955388370893e-05, "loss": 0.2957, "num_input_tokens_seen": 27022992, "step": 19945 }, { "epoch": 0.6385634722488958, "grad_norm": 0.6784633994102478, "learning_rate": 2.510013799805907e-05, "loss": 0.2575, "num_input_tokens_seen": 27029712, "step": 19950 }, { "epoch": 0.6387235132193841, "grad_norm": 1.067744493484497, "learning_rate": 2.5090320592304706e-05, "loss": 0.4175, "num_input_tokens_seen": 27036784, "step": 19955 }, { "epoch": 0.6388835541898726, "grad_norm": 1.2903486490249634, "learning_rate": 2.5080503172621777e-05, "loss": 0.3904, "num_input_tokens_seen": 27043392, "step": 19960 }, { "epoch": 0.639043595160361, "grad_norm": 1.042386531829834, "learning_rate": 2.5070685740524246e-05, "loss": 0.3571, "num_input_tokens_seen": 27050112, "step": 19965 }, { "epoch": 0.6392036361308495, "grad_norm": 2.138936996459961, "learning_rate": 2.5060868297526084e-05, "loss": 0.4852, "num_input_tokens_seen": 27056752, "step": 19970 }, { "epoch": 0.6393636771013379, "grad_norm": 0.9103673696517944, "learning_rate": 2.5051050845141267e-05, "loss": 0.3734, "num_input_tokens_seen": 27064176, "step": 19975 }, { "epoch": 0.6395237180718264, "grad_norm": 0.601779580116272, "learning_rate": 2.5041233384883765e-05, "loss": 0.3839, "num_input_tokens_seen": 27071008, "step": 19980 }, { "epoch": 0.6396837590423148, "grad_norm": 1.138095498085022, "learning_rate": 2.5031415918267564e-05, "loss": 0.4334, "num_input_tokens_seen": 27077744, "step": 19985 }, { "epoch": 0.6398438000128033, "grad_norm": 1.2888764142990112, "learning_rate": 2.5021598446806626e-05, "loss": 0.4312, "num_input_tokens_seen": 27084304, "step": 19990 }, { "epoch": 0.6400038409832918, "grad_norm": 5.164597511291504, "learning_rate": 2.5011780972014937e-05, "loss": 0.4618, "num_input_tokens_seen": 27091248, "step": 19995 }, { "epoch": 0.6401638819537802, "grad_norm": 0.4004306197166443, "learning_rate": 2.5001963495406478e-05, "loss": 0.3012, "num_input_tokens_seen": 27098112, "step": 20000 }, { "epoch": 0.6401638819537802, "eval_loss": 0.46179184317588806, "eval_runtime": 502.6653, "eval_samples_per_second": 27.625, "eval_steps_per_second": 13.812, "num_input_tokens_seen": 27098112, "step": 20000 }, { "epoch": 0.6403239229242687, "grad_norm": 0.49682721495628357, "learning_rate": 2.499214601849522e-05, "loss": 0.2906, "num_input_tokens_seen": 27104848, "step": 20005 }, { "epoch": 0.640483963894757, "grad_norm": 0.7872929573059082, "learning_rate": 2.4982328542795148e-05, "loss": 0.2969, "num_input_tokens_seen": 27111344, "step": 20010 }, { "epoch": 0.6406440048652455, "grad_norm": 0.7133567929267883, "learning_rate": 2.497251106982024e-05, "loss": 0.4397, "num_input_tokens_seen": 27118112, "step": 20015 }, { "epoch": 0.6408040458357339, "grad_norm": 3.3602049350738525, "learning_rate": 2.4962693601084458e-05, "loss": 0.3603, "num_input_tokens_seen": 27124624, "step": 20020 }, { "epoch": 0.6409640868062224, "grad_norm": 1.0881205797195435, "learning_rate": 2.4952876138101794e-05, "loss": 0.6503, "num_input_tokens_seen": 27131312, "step": 20025 }, { "epoch": 0.6411241277767108, "grad_norm": 2.378541946411133, "learning_rate": 2.4943058682386233e-05, "loss": 0.4787, "num_input_tokens_seen": 27138160, "step": 20030 }, { "epoch": 0.6412841687471993, "grad_norm": 2.013639450073242, "learning_rate": 2.493324123545173e-05, "loss": 0.5075, "num_input_tokens_seen": 27144784, "step": 20035 }, { "epoch": 0.6414442097176877, "grad_norm": 1.164308786392212, "learning_rate": 2.4923423798812272e-05, "loss": 0.588, "num_input_tokens_seen": 27151264, "step": 20040 }, { "epoch": 0.6416042506881762, "grad_norm": 0.7506194114685059, "learning_rate": 2.4913606373981825e-05, "loss": 0.4389, "num_input_tokens_seen": 27158112, "step": 20045 }, { "epoch": 0.6417642916586647, "grad_norm": 2.846315622329712, "learning_rate": 2.4903788962474357e-05, "loss": 0.6274, "num_input_tokens_seen": 27164752, "step": 20050 }, { "epoch": 0.641924332629153, "grad_norm": 0.8597906231880188, "learning_rate": 2.489397156580385e-05, "loss": 0.3699, "num_input_tokens_seen": 27171424, "step": 20055 }, { "epoch": 0.6420843735996415, "grad_norm": 1.6260597705841064, "learning_rate": 2.4884154185484246e-05, "loss": 0.4679, "num_input_tokens_seen": 27177920, "step": 20060 }, { "epoch": 0.6422444145701299, "grad_norm": 0.7856186628341675, "learning_rate": 2.4874336823029526e-05, "loss": 0.5428, "num_input_tokens_seen": 27185088, "step": 20065 }, { "epoch": 0.6424044555406184, "grad_norm": 2.9259605407714844, "learning_rate": 2.4864519479953656e-05, "loss": 0.5924, "num_input_tokens_seen": 27191824, "step": 20070 }, { "epoch": 0.6425644965111068, "grad_norm": 1.5606067180633545, "learning_rate": 2.485470215777058e-05, "loss": 0.5319, "num_input_tokens_seen": 27198384, "step": 20075 }, { "epoch": 0.6427245374815953, "grad_norm": 1.0898425579071045, "learning_rate": 2.4844884857994258e-05, "loss": 0.4232, "num_input_tokens_seen": 27205216, "step": 20080 }, { "epoch": 0.6428845784520837, "grad_norm": 0.9408832788467407, "learning_rate": 2.4835067582138638e-05, "loss": 0.5257, "num_input_tokens_seen": 27212240, "step": 20085 }, { "epoch": 0.6430446194225722, "grad_norm": 1.6821718215942383, "learning_rate": 2.4825250331717666e-05, "loss": 0.5842, "num_input_tokens_seen": 27219184, "step": 20090 }, { "epoch": 0.6432046603930607, "grad_norm": 1.3382997512817383, "learning_rate": 2.4815433108245298e-05, "loss": 0.586, "num_input_tokens_seen": 27225952, "step": 20095 }, { "epoch": 0.6433647013635491, "grad_norm": 0.6200948357582092, "learning_rate": 2.4805615913235456e-05, "loss": 0.3344, "num_input_tokens_seen": 27232720, "step": 20100 }, { "epoch": 0.6435247423340376, "grad_norm": 0.9662649035453796, "learning_rate": 2.479579874820208e-05, "loss": 0.5485, "num_input_tokens_seen": 27239424, "step": 20105 }, { "epoch": 0.6436847833045259, "grad_norm": 2.423508405685425, "learning_rate": 2.4785981614659115e-05, "loss": 0.5622, "num_input_tokens_seen": 27245904, "step": 20110 }, { "epoch": 0.6438448242750144, "grad_norm": 1.803612470626831, "learning_rate": 2.477616451412047e-05, "loss": 0.7522, "num_input_tokens_seen": 27253504, "step": 20115 }, { "epoch": 0.6440048652455028, "grad_norm": 1.524285078048706, "learning_rate": 2.476634744810007e-05, "loss": 0.4432, "num_input_tokens_seen": 27260192, "step": 20120 }, { "epoch": 0.6441649062159913, "grad_norm": 1.4581034183502197, "learning_rate": 2.475653041811183e-05, "loss": 0.5701, "num_input_tokens_seen": 27266864, "step": 20125 }, { "epoch": 0.6443249471864797, "grad_norm": 1.5394526720046997, "learning_rate": 2.4746713425669652e-05, "loss": 0.4176, "num_input_tokens_seen": 27274096, "step": 20130 }, { "epoch": 0.6444849881569682, "grad_norm": 0.9581401348114014, "learning_rate": 2.4736896472287458e-05, "loss": 0.3187, "num_input_tokens_seen": 27280784, "step": 20135 }, { "epoch": 0.6446450291274566, "grad_norm": 1.923144817352295, "learning_rate": 2.4727079559479124e-05, "loss": 0.4181, "num_input_tokens_seen": 27288208, "step": 20140 }, { "epoch": 0.6448050700979451, "grad_norm": 1.9955828189849854, "learning_rate": 2.4717262688758557e-05, "loss": 0.6018, "num_input_tokens_seen": 27295200, "step": 20145 }, { "epoch": 0.6449651110684336, "grad_norm": 1.4766621589660645, "learning_rate": 2.4707445861639637e-05, "loss": 0.4852, "num_input_tokens_seen": 27301968, "step": 20150 }, { "epoch": 0.645125152038922, "grad_norm": 1.2269874811172485, "learning_rate": 2.4697629079636244e-05, "loss": 0.4722, "num_input_tokens_seen": 27308544, "step": 20155 }, { "epoch": 0.6452851930094105, "grad_norm": 1.3312008380889893, "learning_rate": 2.4687812344262244e-05, "loss": 0.5763, "num_input_tokens_seen": 27315792, "step": 20160 }, { "epoch": 0.6454452339798988, "grad_norm": 1.1780202388763428, "learning_rate": 2.46779956570315e-05, "loss": 0.3481, "num_input_tokens_seen": 27322848, "step": 20165 }, { "epoch": 0.6456052749503873, "grad_norm": 0.5938460230827332, "learning_rate": 2.466817901945787e-05, "loss": 0.3704, "num_input_tokens_seen": 27329296, "step": 20170 }, { "epoch": 0.6457653159208757, "grad_norm": 1.1095637083053589, "learning_rate": 2.4658362433055217e-05, "loss": 0.4126, "num_input_tokens_seen": 27335712, "step": 20175 }, { "epoch": 0.6459253568913642, "grad_norm": 0.7867429852485657, "learning_rate": 2.4648545899337356e-05, "loss": 0.4363, "num_input_tokens_seen": 27342288, "step": 20180 }, { "epoch": 0.6460853978618526, "grad_norm": 0.9986233711242676, "learning_rate": 2.4638729419818143e-05, "loss": 0.3569, "num_input_tokens_seen": 27348848, "step": 20185 }, { "epoch": 0.6462454388323411, "grad_norm": 1.1104828119277954, "learning_rate": 2.46289129960114e-05, "loss": 0.4426, "num_input_tokens_seen": 27355696, "step": 20190 }, { "epoch": 0.6464054798028295, "grad_norm": 0.9568548202514648, "learning_rate": 2.4619096629430924e-05, "loss": 0.6245, "num_input_tokens_seen": 27363056, "step": 20195 }, { "epoch": 0.646565520773318, "grad_norm": 0.7925690412521362, "learning_rate": 2.4609280321590543e-05, "loss": 0.3044, "num_input_tokens_seen": 27369984, "step": 20200 }, { "epoch": 0.646565520773318, "eval_loss": 0.46089276671409607, "eval_runtime": 502.6082, "eval_samples_per_second": 27.628, "eval_steps_per_second": 13.814, "num_input_tokens_seen": 27369984, "step": 20200 }, { "epoch": 0.6467255617438065, "grad_norm": 4.099344730377197, "learning_rate": 2.4599464074004037e-05, "loss": 0.5831, "num_input_tokens_seen": 27376736, "step": 20205 }, { "epoch": 0.6468856027142948, "grad_norm": 1.4935015439987183, "learning_rate": 2.4589647888185204e-05, "loss": 0.6189, "num_input_tokens_seen": 27383952, "step": 20210 }, { "epoch": 0.6470456436847833, "grad_norm": 0.9650381803512573, "learning_rate": 2.4579831765647836e-05, "loss": 0.4631, "num_input_tokens_seen": 27390496, "step": 20215 }, { "epoch": 0.6472056846552717, "grad_norm": 0.8739038705825806, "learning_rate": 2.4570015707905676e-05, "loss": 0.3897, "num_input_tokens_seen": 27397456, "step": 20220 }, { "epoch": 0.6473657256257602, "grad_norm": 0.801494300365448, "learning_rate": 2.4560199716472508e-05, "loss": 0.3798, "num_input_tokens_seen": 27404144, "step": 20225 }, { "epoch": 0.6475257665962486, "grad_norm": 0.8734854459762573, "learning_rate": 2.455038379286207e-05, "loss": 0.6545, "num_input_tokens_seen": 27411024, "step": 20230 }, { "epoch": 0.6476858075667371, "grad_norm": 1.5059138536453247, "learning_rate": 2.4540567938588095e-05, "loss": 0.4278, "num_input_tokens_seen": 27417824, "step": 20235 }, { "epoch": 0.6478458485372255, "grad_norm": 1.0476032495498657, "learning_rate": 2.4530752155164328e-05, "loss": 0.6298, "num_input_tokens_seen": 27424896, "step": 20240 }, { "epoch": 0.648005889507714, "grad_norm": 1.5662373304367065, "learning_rate": 2.4520936444104463e-05, "loss": 0.4295, "num_input_tokens_seen": 27431360, "step": 20245 }, { "epoch": 0.6481659304782024, "grad_norm": 1.0478121042251587, "learning_rate": 2.4511120806922218e-05, "loss": 0.3783, "num_input_tokens_seen": 27438416, "step": 20250 }, { "epoch": 0.6483259714486909, "grad_norm": 0.6824548840522766, "learning_rate": 2.45013052451313e-05, "loss": 0.4073, "num_input_tokens_seen": 27445248, "step": 20255 }, { "epoch": 0.6484860124191794, "grad_norm": 0.5337820053100586, "learning_rate": 2.4491489760245376e-05, "loss": 0.3391, "num_input_tokens_seen": 27452128, "step": 20260 }, { "epoch": 0.6486460533896677, "grad_norm": 1.1193565130233765, "learning_rate": 2.4481674353778115e-05, "loss": 0.4941, "num_input_tokens_seen": 27458880, "step": 20265 }, { "epoch": 0.6488060943601562, "grad_norm": 1.7811610698699951, "learning_rate": 2.447185902724319e-05, "loss": 0.3784, "num_input_tokens_seen": 27465808, "step": 20270 }, { "epoch": 0.6489661353306446, "grad_norm": 1.4347734451293945, "learning_rate": 2.4462043782154233e-05, "loss": 0.4417, "num_input_tokens_seen": 27472256, "step": 20275 }, { "epoch": 0.6491261763011331, "grad_norm": 0.6616083979606628, "learning_rate": 2.4452228620024895e-05, "loss": 0.3352, "num_input_tokens_seen": 27478736, "step": 20280 }, { "epoch": 0.6492862172716215, "grad_norm": 1.2172554731369019, "learning_rate": 2.4442413542368776e-05, "loss": 0.4968, "num_input_tokens_seen": 27485344, "step": 20285 }, { "epoch": 0.64944625824211, "grad_norm": 1.0974398851394653, "learning_rate": 2.4432598550699502e-05, "loss": 0.5538, "num_input_tokens_seen": 27492336, "step": 20290 }, { "epoch": 0.6496062992125984, "grad_norm": 2.7792866230010986, "learning_rate": 2.4422783646530663e-05, "loss": 0.5248, "num_input_tokens_seen": 27499072, "step": 20295 }, { "epoch": 0.6497663401830869, "grad_norm": 1.6997166872024536, "learning_rate": 2.441296883137584e-05, "loss": 0.5526, "num_input_tokens_seen": 27505728, "step": 20300 }, { "epoch": 0.6499263811535753, "grad_norm": 1.16422700881958, "learning_rate": 2.4403154106748592e-05, "loss": 0.6423, "num_input_tokens_seen": 27512432, "step": 20305 }, { "epoch": 0.6500864221240638, "grad_norm": 0.9466081857681274, "learning_rate": 2.4393339474162494e-05, "loss": 0.4448, "num_input_tokens_seen": 27519328, "step": 20310 }, { "epoch": 0.6502464630945523, "grad_norm": 2.31803560256958, "learning_rate": 2.4383524935131062e-05, "loss": 0.4944, "num_input_tokens_seen": 27526448, "step": 20315 }, { "epoch": 0.6504065040650406, "grad_norm": 1.1614247560501099, "learning_rate": 2.437371049116784e-05, "loss": 0.453, "num_input_tokens_seen": 27533376, "step": 20320 }, { "epoch": 0.6505665450355291, "grad_norm": 0.771190345287323, "learning_rate": 2.436389614378632e-05, "loss": 0.4754, "num_input_tokens_seen": 27539968, "step": 20325 }, { "epoch": 0.6507265860060175, "grad_norm": 0.7757164835929871, "learning_rate": 2.435408189450002e-05, "loss": 0.355, "num_input_tokens_seen": 27546624, "step": 20330 }, { "epoch": 0.650886626976506, "grad_norm": 1.0329428911209106, "learning_rate": 2.4344267744822406e-05, "loss": 0.4102, "num_input_tokens_seen": 27553408, "step": 20335 }, { "epoch": 0.6510466679469944, "grad_norm": 0.45103663206100464, "learning_rate": 2.4334453696266944e-05, "loss": 0.4378, "num_input_tokens_seen": 27559888, "step": 20340 }, { "epoch": 0.6512067089174829, "grad_norm": 1.5316184759140015, "learning_rate": 2.432463975034708e-05, "loss": 0.4503, "num_input_tokens_seen": 27566368, "step": 20345 }, { "epoch": 0.6513667498879713, "grad_norm": 1.0930469036102295, "learning_rate": 2.4314825908576265e-05, "loss": 0.4083, "num_input_tokens_seen": 27572624, "step": 20350 }, { "epoch": 0.6515267908584598, "grad_norm": 1.2167447805404663, "learning_rate": 2.4305012172467897e-05, "loss": 0.5991, "num_input_tokens_seen": 27579328, "step": 20355 }, { "epoch": 0.6516868318289483, "grad_norm": 1.0172560214996338, "learning_rate": 2.4295198543535393e-05, "loss": 0.3734, "num_input_tokens_seen": 27586064, "step": 20360 }, { "epoch": 0.6518468727994366, "grad_norm": 0.9198071956634521, "learning_rate": 2.4285385023292124e-05, "loss": 0.2769, "num_input_tokens_seen": 27592832, "step": 20365 }, { "epoch": 0.6520069137699251, "grad_norm": 2.0872631072998047, "learning_rate": 2.427557161325147e-05, "loss": 0.5394, "num_input_tokens_seen": 27599728, "step": 20370 }, { "epoch": 0.6521669547404135, "grad_norm": 0.9336918592453003, "learning_rate": 2.4265758314926778e-05, "loss": 0.4082, "num_input_tokens_seen": 27606640, "step": 20375 }, { "epoch": 0.652326995710902, "grad_norm": 1.7495778799057007, "learning_rate": 2.4255945129831373e-05, "loss": 0.5761, "num_input_tokens_seen": 27613360, "step": 20380 }, { "epoch": 0.6524870366813904, "grad_norm": 0.7791070342063904, "learning_rate": 2.4246132059478578e-05, "loss": 0.506, "num_input_tokens_seen": 27620048, "step": 20385 }, { "epoch": 0.6526470776518789, "grad_norm": 1.6850075721740723, "learning_rate": 2.4236319105381706e-05, "loss": 0.4375, "num_input_tokens_seen": 27627680, "step": 20390 }, { "epoch": 0.6528071186223673, "grad_norm": 0.4529598653316498, "learning_rate": 2.422650626905401e-05, "loss": 0.3226, "num_input_tokens_seen": 27634560, "step": 20395 }, { "epoch": 0.6529671595928558, "grad_norm": 1.3418370485305786, "learning_rate": 2.4216693552008785e-05, "loss": 0.5599, "num_input_tokens_seen": 27640768, "step": 20400 }, { "epoch": 0.6529671595928558, "eval_loss": 0.45975616574287415, "eval_runtime": 502.4431, "eval_samples_per_second": 27.637, "eval_steps_per_second": 13.818, "num_input_tokens_seen": 27640768, "step": 20400 }, { "epoch": 0.6531272005633442, "grad_norm": 1.2320592403411865, "learning_rate": 2.4206880955759247e-05, "loss": 0.6945, "num_input_tokens_seen": 27647424, "step": 20405 }, { "epoch": 0.6532872415338327, "grad_norm": 0.9389510154724121, "learning_rate": 2.419706848181863e-05, "loss": 0.3641, "num_input_tokens_seen": 27654304, "step": 20410 }, { "epoch": 0.6534472825043212, "grad_norm": 0.8555676341056824, "learning_rate": 2.4187256131700153e-05, "loss": 0.4468, "num_input_tokens_seen": 27661328, "step": 20415 }, { "epoch": 0.6536073234748095, "grad_norm": 2.2362289428710938, "learning_rate": 2.4177443906916985e-05, "loss": 0.5344, "num_input_tokens_seen": 27669104, "step": 20420 }, { "epoch": 0.653767364445298, "grad_norm": 0.8022750616073608, "learning_rate": 2.4167631808982303e-05, "loss": 0.3824, "num_input_tokens_seen": 27676496, "step": 20425 }, { "epoch": 0.6539274054157864, "grad_norm": 1.28188955783844, "learning_rate": 2.4157819839409264e-05, "loss": 0.5289, "num_input_tokens_seen": 27683200, "step": 20430 }, { "epoch": 0.6540874463862749, "grad_norm": 1.0495871305465698, "learning_rate": 2.414800799971098e-05, "loss": 0.5209, "num_input_tokens_seen": 27690272, "step": 20435 }, { "epoch": 0.6542474873567633, "grad_norm": 1.6311619281768799, "learning_rate": 2.4138196291400582e-05, "loss": 0.5765, "num_input_tokens_seen": 27696816, "step": 20440 }, { "epoch": 0.6544075283272518, "grad_norm": 1.4258079528808594, "learning_rate": 2.412838471599114e-05, "loss": 0.4239, "num_input_tokens_seen": 27703856, "step": 20445 }, { "epoch": 0.6545675692977402, "grad_norm": 1.005157709121704, "learning_rate": 2.411857327499572e-05, "loss": 0.5192, "num_input_tokens_seen": 27710528, "step": 20450 }, { "epoch": 0.6547276102682287, "grad_norm": 0.8010101914405823, "learning_rate": 2.410876196992739e-05, "loss": 0.4267, "num_input_tokens_seen": 27717056, "step": 20455 }, { "epoch": 0.6548876512387171, "grad_norm": 2.108834743499756, "learning_rate": 2.4098950802299156e-05, "loss": 0.5824, "num_input_tokens_seen": 27723856, "step": 20460 }, { "epoch": 0.6550476922092056, "grad_norm": 1.6484503746032715, "learning_rate": 2.4089139773624027e-05, "loss": 0.4889, "num_input_tokens_seen": 27730416, "step": 20465 }, { "epoch": 0.655207733179694, "grad_norm": 0.8266956806182861, "learning_rate": 2.4079328885415007e-05, "loss": 0.608, "num_input_tokens_seen": 27737328, "step": 20470 }, { "epoch": 0.6553677741501824, "grad_norm": 0.7103060483932495, "learning_rate": 2.4069518139185036e-05, "loss": 0.3644, "num_input_tokens_seen": 27743792, "step": 20475 }, { "epoch": 0.6555278151206709, "grad_norm": 0.3555283844470978, "learning_rate": 2.405970753644706e-05, "loss": 0.2506, "num_input_tokens_seen": 27750400, "step": 20480 }, { "epoch": 0.6556878560911593, "grad_norm": 0.5952970385551453, "learning_rate": 2.4049897078714e-05, "loss": 0.344, "num_input_tokens_seen": 27757360, "step": 20485 }, { "epoch": 0.6558478970616478, "grad_norm": 1.7720859050750732, "learning_rate": 2.404008676749874e-05, "loss": 0.3813, "num_input_tokens_seen": 27763840, "step": 20490 }, { "epoch": 0.6560079380321362, "grad_norm": 0.9890759587287903, "learning_rate": 2.403027660431418e-05, "loss": 0.3325, "num_input_tokens_seen": 27770336, "step": 20495 }, { "epoch": 0.6561679790026247, "grad_norm": 1.1558117866516113, "learning_rate": 2.402046659067314e-05, "loss": 0.4957, "num_input_tokens_seen": 27776960, "step": 20500 }, { "epoch": 0.6563280199731131, "grad_norm": 0.8957875967025757, "learning_rate": 2.401065672808847e-05, "loss": 0.4214, "num_input_tokens_seen": 27783472, "step": 20505 }, { "epoch": 0.6564880609436016, "grad_norm": 0.5670983791351318, "learning_rate": 2.400084701807296e-05, "loss": 0.3925, "num_input_tokens_seen": 27789984, "step": 20510 }, { "epoch": 0.65664810191409, "grad_norm": 1.719618558883667, "learning_rate": 2.39910374621394e-05, "loss": 0.4415, "num_input_tokens_seen": 27796576, "step": 20515 }, { "epoch": 0.6568081428845784, "grad_norm": 0.8680922985076904, "learning_rate": 2.3981228061800544e-05, "loss": 0.5038, "num_input_tokens_seen": 27802960, "step": 20520 }, { "epoch": 0.656968183855067, "grad_norm": 0.9573259949684143, "learning_rate": 2.3971418818569115e-05, "loss": 0.5813, "num_input_tokens_seen": 27809680, "step": 20525 }, { "epoch": 0.6571282248255553, "grad_norm": 1.120613694190979, "learning_rate": 2.3961609733957832e-05, "loss": 0.4777, "num_input_tokens_seen": 27816272, "step": 20530 }, { "epoch": 0.6572882657960438, "grad_norm": 0.8722102046012878, "learning_rate": 2.395180080947939e-05, "loss": 0.3865, "num_input_tokens_seen": 27822992, "step": 20535 }, { "epoch": 0.6574483067665322, "grad_norm": 0.8517833948135376, "learning_rate": 2.394199204664642e-05, "loss": 0.3912, "num_input_tokens_seen": 27829568, "step": 20540 }, { "epoch": 0.6576083477370207, "grad_norm": 0.8262898921966553, "learning_rate": 2.3932183446971583e-05, "loss": 0.5531, "num_input_tokens_seen": 27836448, "step": 20545 }, { "epoch": 0.6577683887075091, "grad_norm": 1.9160122871398926, "learning_rate": 2.3922375011967473e-05, "loss": 0.4538, "num_input_tokens_seen": 27843504, "step": 20550 }, { "epoch": 0.6579284296779976, "grad_norm": 0.8452538251876831, "learning_rate": 2.3912566743146676e-05, "loss": 0.4703, "num_input_tokens_seen": 27850240, "step": 20555 }, { "epoch": 0.658088470648486, "grad_norm": 0.864054262638092, "learning_rate": 2.390275864202176e-05, "loss": 0.331, "num_input_tokens_seen": 27856912, "step": 20560 }, { "epoch": 0.6582485116189745, "grad_norm": 0.6120114326477051, "learning_rate": 2.3892950710105243e-05, "loss": 0.3676, "num_input_tokens_seen": 27863504, "step": 20565 }, { "epoch": 0.6584085525894628, "grad_norm": 1.5379517078399658, "learning_rate": 2.3883142948909635e-05, "loss": 0.4177, "num_input_tokens_seen": 27870096, "step": 20570 }, { "epoch": 0.6585685935599513, "grad_norm": 0.5363937020301819, "learning_rate": 2.3873335359947433e-05, "loss": 0.359, "num_input_tokens_seen": 27876656, "step": 20575 }, { "epoch": 0.6587286345304398, "grad_norm": 0.49278077483177185, "learning_rate": 2.3863527944731066e-05, "loss": 0.3108, "num_input_tokens_seen": 27883392, "step": 20580 }, { "epoch": 0.6588886755009282, "grad_norm": 1.2295503616333008, "learning_rate": 2.385372070477298e-05, "loss": 0.4425, "num_input_tokens_seen": 27890032, "step": 20585 }, { "epoch": 0.6590487164714167, "grad_norm": 1.9223800897598267, "learning_rate": 2.384391364158556e-05, "loss": 0.5333, "num_input_tokens_seen": 27896976, "step": 20590 }, { "epoch": 0.6592087574419051, "grad_norm": 1.0595468282699585, "learning_rate": 2.3834106756681185e-05, "loss": 0.5395, "num_input_tokens_seen": 27903600, "step": 20595 }, { "epoch": 0.6593687984123936, "grad_norm": 2.05108642578125, "learning_rate": 2.3824300051572206e-05, "loss": 0.3936, "num_input_tokens_seen": 27910480, "step": 20600 }, { "epoch": 0.6593687984123936, "eval_loss": 0.45920807123184204, "eval_runtime": 502.4841, "eval_samples_per_second": 27.635, "eval_steps_per_second": 13.817, "num_input_tokens_seen": 27910480, "step": 20600 }, { "epoch": 0.659528839382882, "grad_norm": 1.4748965501785278, "learning_rate": 2.3814493527770923e-05, "loss": 0.482, "num_input_tokens_seen": 27917056, "step": 20605 }, { "epoch": 0.6596888803533705, "grad_norm": 0.9572867751121521, "learning_rate": 2.3804687186789637e-05, "loss": 0.3717, "num_input_tokens_seen": 27924048, "step": 20610 }, { "epoch": 0.6598489213238589, "grad_norm": 0.7072188854217529, "learning_rate": 2.379488103014062e-05, "loss": 0.4657, "num_input_tokens_seen": 27930736, "step": 20615 }, { "epoch": 0.6600089622943474, "grad_norm": 0.8445866703987122, "learning_rate": 2.3785075059336086e-05, "loss": 0.4068, "num_input_tokens_seen": 27937856, "step": 20620 }, { "epoch": 0.6601690032648359, "grad_norm": 0.7934162616729736, "learning_rate": 2.3775269275888248e-05, "loss": 0.4547, "num_input_tokens_seen": 27944816, "step": 20625 }, { "epoch": 0.6603290442353242, "grad_norm": 1.6215077638626099, "learning_rate": 2.3765463681309274e-05, "loss": 0.5383, "num_input_tokens_seen": 27951504, "step": 20630 }, { "epoch": 0.6604890852058127, "grad_norm": 1.701661229133606, "learning_rate": 2.3755658277111313e-05, "loss": 0.4588, "num_input_tokens_seen": 27958160, "step": 20635 }, { "epoch": 0.6606491261763011, "grad_norm": 2.6753990650177, "learning_rate": 2.374585306480649e-05, "loss": 0.4825, "num_input_tokens_seen": 27964880, "step": 20640 }, { "epoch": 0.6608091671467896, "grad_norm": 0.8581603169441223, "learning_rate": 2.3736048045906877e-05, "loss": 0.446, "num_input_tokens_seen": 27971424, "step": 20645 }, { "epoch": 0.660969208117278, "grad_norm": 0.9199719429016113, "learning_rate": 2.372624322192454e-05, "loss": 0.4594, "num_input_tokens_seen": 27977632, "step": 20650 }, { "epoch": 0.6611292490877665, "grad_norm": 1.3351387977600098, "learning_rate": 2.3716438594371516e-05, "loss": 0.4881, "num_input_tokens_seen": 27984512, "step": 20655 }, { "epoch": 0.6612892900582549, "grad_norm": 0.9169932007789612, "learning_rate": 2.3706634164759784e-05, "loss": 0.3617, "num_input_tokens_seen": 27991184, "step": 20660 }, { "epoch": 0.6614493310287434, "grad_norm": 1.4637168645858765, "learning_rate": 2.3696829934601323e-05, "loss": 0.4492, "num_input_tokens_seen": 27997840, "step": 20665 }, { "epoch": 0.6616093719992318, "grad_norm": 1.0879899263381958, "learning_rate": 2.3687025905408053e-05, "loss": 0.4625, "num_input_tokens_seen": 28004832, "step": 20670 }, { "epoch": 0.6617694129697202, "grad_norm": 1.8825902938842773, "learning_rate": 2.3677222078691886e-05, "loss": 0.5841, "num_input_tokens_seen": 28011344, "step": 20675 }, { "epoch": 0.6619294539402087, "grad_norm": 1.0498782396316528, "learning_rate": 2.366741845596471e-05, "loss": 0.3071, "num_input_tokens_seen": 28018112, "step": 20680 }, { "epoch": 0.6620894949106971, "grad_norm": 1.76450777053833, "learning_rate": 2.3657615038738343e-05, "loss": 0.4788, "num_input_tokens_seen": 28024976, "step": 20685 }, { "epoch": 0.6622495358811856, "grad_norm": 1.1589082479476929, "learning_rate": 2.3647811828524614e-05, "loss": 0.4599, "num_input_tokens_seen": 28031632, "step": 20690 }, { "epoch": 0.662409576851674, "grad_norm": 0.8871309757232666, "learning_rate": 2.363800882683529e-05, "loss": 0.3894, "num_input_tokens_seen": 28038464, "step": 20695 }, { "epoch": 0.6625696178221625, "grad_norm": 1.6359187364578247, "learning_rate": 2.3628206035182125e-05, "loss": 0.4511, "num_input_tokens_seen": 28044928, "step": 20700 }, { "epoch": 0.6627296587926509, "grad_norm": 1.0421271324157715, "learning_rate": 2.361840345507683e-05, "loss": 0.4578, "num_input_tokens_seen": 28051552, "step": 20705 }, { "epoch": 0.6628896997631394, "grad_norm": 1.201812505722046, "learning_rate": 2.3608601088031073e-05, "loss": 0.5432, "num_input_tokens_seen": 28058560, "step": 20710 }, { "epoch": 0.6630497407336278, "grad_norm": 1.15079665184021, "learning_rate": 2.3598798935556516e-05, "loss": 0.6892, "num_input_tokens_seen": 28065536, "step": 20715 }, { "epoch": 0.6632097817041163, "grad_norm": 0.828709602355957, "learning_rate": 2.3588996999164784e-05, "loss": 0.3677, "num_input_tokens_seen": 28072512, "step": 20720 }, { "epoch": 0.6633698226746046, "grad_norm": 0.6142131090164185, "learning_rate": 2.3579195280367434e-05, "loss": 0.2756, "num_input_tokens_seen": 28079360, "step": 20725 }, { "epoch": 0.6635298636450931, "grad_norm": 0.7956287860870361, "learning_rate": 2.356939378067603e-05, "loss": 0.5295, "num_input_tokens_seen": 28086304, "step": 20730 }, { "epoch": 0.6636899046155816, "grad_norm": 0.6568320989608765, "learning_rate": 2.3559592501602092e-05, "loss": 0.469, "num_input_tokens_seen": 28092832, "step": 20735 }, { "epoch": 0.66384994558607, "grad_norm": 1.1758859157562256, "learning_rate": 2.3549791444657076e-05, "loss": 0.3951, "num_input_tokens_seen": 28099328, "step": 20740 }, { "epoch": 0.6640099865565585, "grad_norm": 0.9873849749565125, "learning_rate": 2.353999061135246e-05, "loss": 0.2756, "num_input_tokens_seen": 28106192, "step": 20745 }, { "epoch": 0.6641700275270469, "grad_norm": 0.7410650849342346, "learning_rate": 2.3530190003199626e-05, "loss": 0.593, "num_input_tokens_seen": 28112928, "step": 20750 }, { "epoch": 0.6643300684975354, "grad_norm": 0.8740013241767883, "learning_rate": 2.3520389621709965e-05, "loss": 0.4246, "num_input_tokens_seen": 28119968, "step": 20755 }, { "epoch": 0.6644901094680238, "grad_norm": 1.5406877994537354, "learning_rate": 2.351058946839483e-05, "loss": 0.4175, "num_input_tokens_seen": 28126800, "step": 20760 }, { "epoch": 0.6646501504385123, "grad_norm": 1.659627914428711, "learning_rate": 2.350078954476551e-05, "loss": 0.3457, "num_input_tokens_seen": 28133040, "step": 20765 }, { "epoch": 0.6648101914090007, "grad_norm": 0.7541285157203674, "learning_rate": 2.3490989852333272e-05, "loss": 0.3172, "num_input_tokens_seen": 28140016, "step": 20770 }, { "epoch": 0.6649702323794892, "grad_norm": 0.4567452073097229, "learning_rate": 2.3481190392609377e-05, "loss": 0.3883, "num_input_tokens_seen": 28146512, "step": 20775 }, { "epoch": 0.6651302733499775, "grad_norm": 0.4751623868942261, "learning_rate": 2.3471391167105e-05, "loss": 0.2944, "num_input_tokens_seen": 28153648, "step": 20780 }, { "epoch": 0.665290314320466, "grad_norm": 1.8126635551452637, "learning_rate": 2.3461592177331325e-05, "loss": 0.4376, "num_input_tokens_seen": 28160304, "step": 20785 }, { "epoch": 0.6654503552909545, "grad_norm": 0.7794167995452881, "learning_rate": 2.345179342479946e-05, "loss": 0.3213, "num_input_tokens_seen": 28167008, "step": 20790 }, { "epoch": 0.6656103962614429, "grad_norm": 1.0302480459213257, "learning_rate": 2.3441994911020503e-05, "loss": 0.3571, "num_input_tokens_seen": 28173936, "step": 20795 }, { "epoch": 0.6657704372319314, "grad_norm": 0.8023236393928528, "learning_rate": 2.3432196637505522e-05, "loss": 0.4015, "num_input_tokens_seen": 28180240, "step": 20800 }, { "epoch": 0.6657704372319314, "eval_loss": 0.45871827006340027, "eval_runtime": 502.4228, "eval_samples_per_second": 27.638, "eval_steps_per_second": 13.819, "num_input_tokens_seen": 28180240, "step": 20800 }, { "epoch": 0.6659304782024198, "grad_norm": 1.0687702894210815, "learning_rate": 2.3422398605765515e-05, "loss": 0.4672, "num_input_tokens_seen": 28186832, "step": 20805 }, { "epoch": 0.6660905191729083, "grad_norm": 1.5032145977020264, "learning_rate": 2.3412600817311462e-05, "loss": 0.5729, "num_input_tokens_seen": 28194368, "step": 20810 }, { "epoch": 0.6662505601433967, "grad_norm": 1.44588303565979, "learning_rate": 2.3402803273654326e-05, "loss": 0.4316, "num_input_tokens_seen": 28200976, "step": 20815 }, { "epoch": 0.6664106011138852, "grad_norm": 1.5080820322036743, "learning_rate": 2.3393005976304983e-05, "loss": 0.3802, "num_input_tokens_seen": 28207984, "step": 20820 }, { "epoch": 0.6665706420843736, "grad_norm": 1.0068719387054443, "learning_rate": 2.338320892677432e-05, "loss": 0.5169, "num_input_tokens_seen": 28214720, "step": 20825 }, { "epoch": 0.666730683054862, "grad_norm": 0.9536910653114319, "learning_rate": 2.3373412126573155e-05, "loss": 0.5289, "num_input_tokens_seen": 28221312, "step": 20830 }, { "epoch": 0.6668907240253504, "grad_norm": 1.1679376363754272, "learning_rate": 2.3363615577212285e-05, "loss": 0.4844, "num_input_tokens_seen": 28228176, "step": 20835 }, { "epoch": 0.6670507649958389, "grad_norm": 1.4089418649673462, "learning_rate": 2.3353819280202455e-05, "loss": 0.4631, "num_input_tokens_seen": 28235056, "step": 20840 }, { "epoch": 0.6672108059663274, "grad_norm": 0.9316611886024475, "learning_rate": 2.334402323705438e-05, "loss": 0.49, "num_input_tokens_seen": 28241840, "step": 20845 }, { "epoch": 0.6673708469368158, "grad_norm": 1.0765846967697144, "learning_rate": 2.3334227449278725e-05, "loss": 0.3164, "num_input_tokens_seen": 28248560, "step": 20850 }, { "epoch": 0.6675308879073043, "grad_norm": 1.9219536781311035, "learning_rate": 2.3324431918386143e-05, "loss": 0.4364, "num_input_tokens_seen": 28255440, "step": 20855 }, { "epoch": 0.6676909288777927, "grad_norm": 1.085113763809204, "learning_rate": 2.3314636645887207e-05, "loss": 0.3527, "num_input_tokens_seen": 28262336, "step": 20860 }, { "epoch": 0.6678509698482812, "grad_norm": 0.5674912929534912, "learning_rate": 2.3304841633292487e-05, "loss": 0.3407, "num_input_tokens_seen": 28268768, "step": 20865 }, { "epoch": 0.6680110108187696, "grad_norm": 1.9897706508636475, "learning_rate": 2.329504688211248e-05, "loss": 0.4742, "num_input_tokens_seen": 28275344, "step": 20870 }, { "epoch": 0.6681710517892581, "grad_norm": 1.6326402425765991, "learning_rate": 2.3285252393857677e-05, "loss": 0.4157, "num_input_tokens_seen": 28282192, "step": 20875 }, { "epoch": 0.6683310927597464, "grad_norm": 1.1234625577926636, "learning_rate": 2.327545817003851e-05, "loss": 0.5245, "num_input_tokens_seen": 28288720, "step": 20880 }, { "epoch": 0.6684911337302349, "grad_norm": 1.2769502401351929, "learning_rate": 2.326566421216535e-05, "loss": 0.4493, "num_input_tokens_seen": 28295232, "step": 20885 }, { "epoch": 0.6686511747007234, "grad_norm": 0.7133677005767822, "learning_rate": 2.3255870521748565e-05, "loss": 0.5111, "num_input_tokens_seen": 28301872, "step": 20890 }, { "epoch": 0.6688112156712118, "grad_norm": 0.6759841442108154, "learning_rate": 2.3246077100298474e-05, "loss": 0.3738, "num_input_tokens_seen": 28308576, "step": 20895 }, { "epoch": 0.6689712566417003, "grad_norm": 0.7668102383613586, "learning_rate": 2.3236283949325328e-05, "loss": 0.3737, "num_input_tokens_seen": 28314880, "step": 20900 }, { "epoch": 0.6691312976121887, "grad_norm": 0.5300349593162537, "learning_rate": 2.3226491070339368e-05, "loss": 0.4153, "num_input_tokens_seen": 28321872, "step": 20905 }, { "epoch": 0.6692913385826772, "grad_norm": 1.007910132408142, "learning_rate": 2.3216698464850762e-05, "loss": 0.4197, "num_input_tokens_seen": 28328848, "step": 20910 }, { "epoch": 0.6694513795531656, "grad_norm": 1.144014835357666, "learning_rate": 2.320690613436967e-05, "loss": 0.3999, "num_input_tokens_seen": 28335536, "step": 20915 }, { "epoch": 0.6696114205236541, "grad_norm": 0.9240982532501221, "learning_rate": 2.3197114080406192e-05, "loss": 0.3774, "num_input_tokens_seen": 28342192, "step": 20920 }, { "epoch": 0.6697714614941425, "grad_norm": 3.191511631011963, "learning_rate": 2.3187322304470365e-05, "loss": 0.6107, "num_input_tokens_seen": 28349200, "step": 20925 }, { "epoch": 0.669931502464631, "grad_norm": 0.7959356307983398, "learning_rate": 2.3177530808072222e-05, "loss": 0.4594, "num_input_tokens_seen": 28356064, "step": 20930 }, { "epoch": 0.6700915434351193, "grad_norm": 0.8077986240386963, "learning_rate": 2.316773959272174e-05, "loss": 0.5367, "num_input_tokens_seen": 28362944, "step": 20935 }, { "epoch": 0.6702515844056078, "grad_norm": 0.914588212966919, "learning_rate": 2.3157948659928823e-05, "loss": 0.4954, "num_input_tokens_seen": 28369696, "step": 20940 }, { "epoch": 0.6704116253760963, "grad_norm": 0.5605587363243103, "learning_rate": 2.3148158011203388e-05, "loss": 0.3184, "num_input_tokens_seen": 28376384, "step": 20945 }, { "epoch": 0.6705716663465847, "grad_norm": 1.0825809240341187, "learning_rate": 2.3138367648055253e-05, "loss": 0.567, "num_input_tokens_seen": 28383200, "step": 20950 }, { "epoch": 0.6707317073170732, "grad_norm": 0.6083789467811584, "learning_rate": 2.312857757199422e-05, "loss": 0.538, "num_input_tokens_seen": 28390112, "step": 20955 }, { "epoch": 0.6708917482875616, "grad_norm": 1.0265264511108398, "learning_rate": 2.3118787784530048e-05, "loss": 0.468, "num_input_tokens_seen": 28397072, "step": 20960 }, { "epoch": 0.6710517892580501, "grad_norm": 1.9771690368652344, "learning_rate": 2.310899828717243e-05, "loss": 0.5543, "num_input_tokens_seen": 28404048, "step": 20965 }, { "epoch": 0.6712118302285385, "grad_norm": 0.9903253316879272, "learning_rate": 2.309920908143104e-05, "loss": 0.3811, "num_input_tokens_seen": 28411072, "step": 20970 }, { "epoch": 0.671371871199027, "grad_norm": 0.7776872515678406, "learning_rate": 2.308942016881551e-05, "loss": 0.3229, "num_input_tokens_seen": 28418016, "step": 20975 }, { "epoch": 0.6715319121695154, "grad_norm": 1.251903772354126, "learning_rate": 2.307963155083539e-05, "loss": 0.5435, "num_input_tokens_seen": 28424480, "step": 20980 }, { "epoch": 0.6716919531400039, "grad_norm": 0.7784940004348755, "learning_rate": 2.306984322900022e-05, "loss": 0.4169, "num_input_tokens_seen": 28431584, "step": 20985 }, { "epoch": 0.6718519941104922, "grad_norm": 1.1364219188690186, "learning_rate": 2.3060055204819482e-05, "loss": 0.4446, "num_input_tokens_seen": 28438336, "step": 20990 }, { "epoch": 0.6720120350809807, "grad_norm": 1.2341430187225342, "learning_rate": 2.3050267479802604e-05, "loss": 0.4114, "num_input_tokens_seen": 28445312, "step": 20995 }, { "epoch": 0.6721720760514692, "grad_norm": 1.4752107858657837, "learning_rate": 2.304048005545899e-05, "loss": 0.5022, "num_input_tokens_seen": 28451984, "step": 21000 }, { "epoch": 0.6721720760514692, "eval_loss": 0.4579222798347473, "eval_runtime": 502.4867, "eval_samples_per_second": 27.635, "eval_steps_per_second": 13.817, "num_input_tokens_seen": 28451984, "step": 21000 }, { "epoch": 0.6723321170219576, "grad_norm": 2.2729740142822266, "learning_rate": 2.3030692933297972e-05, "loss": 0.397, "num_input_tokens_seen": 28458928, "step": 21005 }, { "epoch": 0.6724921579924461, "grad_norm": 0.44885945320129395, "learning_rate": 2.3020906114828843e-05, "loss": 0.5769, "num_input_tokens_seen": 28465648, "step": 21010 }, { "epoch": 0.6726521989629345, "grad_norm": 3.6690850257873535, "learning_rate": 2.301111960156088e-05, "loss": 0.3001, "num_input_tokens_seen": 28473008, "step": 21015 }, { "epoch": 0.672812239933423, "grad_norm": 1.1513967514038086, "learning_rate": 2.300133339500326e-05, "loss": 0.5243, "num_input_tokens_seen": 28479952, "step": 21020 }, { "epoch": 0.6729722809039114, "grad_norm": 0.699674665927887, "learning_rate": 2.2991547496665148e-05, "loss": 0.3321, "num_input_tokens_seen": 28486288, "step": 21025 }, { "epoch": 0.6731323218743999, "grad_norm": 0.8281735181808472, "learning_rate": 2.298176190805565e-05, "loss": 0.6391, "num_input_tokens_seen": 28492912, "step": 21030 }, { "epoch": 0.6732923628448882, "grad_norm": 1.1726322174072266, "learning_rate": 2.2971976630683826e-05, "loss": 0.5765, "num_input_tokens_seen": 28499872, "step": 21035 }, { "epoch": 0.6734524038153767, "grad_norm": 2.0504302978515625, "learning_rate": 2.29621916660587e-05, "loss": 0.5045, "num_input_tokens_seen": 28506560, "step": 21040 }, { "epoch": 0.6736124447858651, "grad_norm": 1.872726321220398, "learning_rate": 2.295240701568922e-05, "loss": 0.4237, "num_input_tokens_seen": 28513360, "step": 21045 }, { "epoch": 0.6737724857563536, "grad_norm": 1.4914218187332153, "learning_rate": 2.2942622681084312e-05, "loss": 0.4301, "num_input_tokens_seen": 28520320, "step": 21050 }, { "epoch": 0.6739325267268421, "grad_norm": 1.0911279916763306, "learning_rate": 2.293283866375284e-05, "loss": 0.4651, "num_input_tokens_seen": 28526800, "step": 21055 }, { "epoch": 0.6740925676973305, "grad_norm": 1.2936582565307617, "learning_rate": 2.2923054965203627e-05, "loss": 0.6638, "num_input_tokens_seen": 28534080, "step": 21060 }, { "epoch": 0.674252608667819, "grad_norm": 0.4043818712234497, "learning_rate": 2.2913271586945443e-05, "loss": 0.4323, "num_input_tokens_seen": 28540544, "step": 21065 }, { "epoch": 0.6744126496383074, "grad_norm": 1.3424934148788452, "learning_rate": 2.290348853048699e-05, "loss": 0.3895, "num_input_tokens_seen": 28547568, "step": 21070 }, { "epoch": 0.6745726906087959, "grad_norm": 0.8908501267433167, "learning_rate": 2.2893705797336956e-05, "loss": 0.461, "num_input_tokens_seen": 28554288, "step": 21075 }, { "epoch": 0.6747327315792843, "grad_norm": 0.7045111656188965, "learning_rate": 2.288392338900397e-05, "loss": 0.4494, "num_input_tokens_seen": 28561216, "step": 21080 }, { "epoch": 0.6748927725497728, "grad_norm": 1.4020875692367554, "learning_rate": 2.2874141306996576e-05, "loss": 0.4402, "num_input_tokens_seen": 28568304, "step": 21085 }, { "epoch": 0.6750528135202611, "grad_norm": 1.8430441617965698, "learning_rate": 2.2864359552823312e-05, "loss": 0.4025, "num_input_tokens_seen": 28575024, "step": 21090 }, { "epoch": 0.6752128544907496, "grad_norm": 1.368208408355713, "learning_rate": 2.2854578127992648e-05, "loss": 0.3728, "num_input_tokens_seen": 28582096, "step": 21095 }, { "epoch": 0.6753728954612381, "grad_norm": 1.3552517890930176, "learning_rate": 2.2844797034012988e-05, "loss": 0.3613, "num_input_tokens_seen": 28588704, "step": 21100 }, { "epoch": 0.6755329364317265, "grad_norm": 1.2592076063156128, "learning_rate": 2.2835016272392722e-05, "loss": 0.4976, "num_input_tokens_seen": 28595904, "step": 21105 }, { "epoch": 0.675692977402215, "grad_norm": 1.0053962469100952, "learning_rate": 2.2825235844640142e-05, "loss": 0.3857, "num_input_tokens_seen": 28602304, "step": 21110 }, { "epoch": 0.6758530183727034, "grad_norm": 0.9266933798789978, "learning_rate": 2.2815455752263522e-05, "loss": 0.5805, "num_input_tokens_seen": 28609344, "step": 21115 }, { "epoch": 0.6760130593431919, "grad_norm": 1.0970829725265503, "learning_rate": 2.2805675996771092e-05, "loss": 0.4276, "num_input_tokens_seen": 28616192, "step": 21120 }, { "epoch": 0.6761731003136803, "grad_norm": 1.1949089765548706, "learning_rate": 2.2795896579670987e-05, "loss": 0.4146, "num_input_tokens_seen": 28623104, "step": 21125 }, { "epoch": 0.6763331412841688, "grad_norm": 0.4506973922252655, "learning_rate": 2.2786117502471337e-05, "loss": 0.4297, "num_input_tokens_seen": 28630144, "step": 21130 }, { "epoch": 0.6764931822546572, "grad_norm": 1.1950397491455078, "learning_rate": 2.2776338766680185e-05, "loss": 0.4678, "num_input_tokens_seen": 28636704, "step": 21135 }, { "epoch": 0.6766532232251457, "grad_norm": 1.419776439666748, "learning_rate": 2.2766560373805533e-05, "loss": 0.6612, "num_input_tokens_seen": 28643408, "step": 21140 }, { "epoch": 0.676813264195634, "grad_norm": 0.7642685770988464, "learning_rate": 2.2756782325355353e-05, "loss": 0.412, "num_input_tokens_seen": 28650624, "step": 21145 }, { "epoch": 0.6769733051661225, "grad_norm": 1.0203287601470947, "learning_rate": 2.2747004622837514e-05, "loss": 0.4009, "num_input_tokens_seen": 28657072, "step": 21150 }, { "epoch": 0.677133346136611, "grad_norm": 0.6041578054428101, "learning_rate": 2.2737227267759878e-05, "loss": 0.5849, "num_input_tokens_seen": 28663536, "step": 21155 }, { "epoch": 0.6772933871070994, "grad_norm": 0.7313211560249329, "learning_rate": 2.272745026163024e-05, "loss": 0.6129, "num_input_tokens_seen": 28670192, "step": 21160 }, { "epoch": 0.6774534280775879, "grad_norm": 1.795999526977539, "learning_rate": 2.271767360595633e-05, "loss": 0.5543, "num_input_tokens_seen": 28676640, "step": 21165 }, { "epoch": 0.6776134690480763, "grad_norm": 1.9468907117843628, "learning_rate": 2.270789730224583e-05, "loss": 0.3347, "num_input_tokens_seen": 28683264, "step": 21170 }, { "epoch": 0.6777735100185648, "grad_norm": 1.0401729345321655, "learning_rate": 2.2698121352006367e-05, "loss": 0.3373, "num_input_tokens_seen": 28689696, "step": 21175 }, { "epoch": 0.6779335509890532, "grad_norm": 1.5371499061584473, "learning_rate": 2.2688345756745517e-05, "loss": 0.5545, "num_input_tokens_seen": 28696032, "step": 21180 }, { "epoch": 0.6780935919595417, "grad_norm": 1.5167714357376099, "learning_rate": 2.267857051797081e-05, "loss": 0.2887, "num_input_tokens_seen": 28703088, "step": 21185 }, { "epoch": 0.67825363293003, "grad_norm": 1.0241421461105347, "learning_rate": 2.2668795637189695e-05, "loss": 0.3308, "num_input_tokens_seen": 28710304, "step": 21190 }, { "epoch": 0.6784136739005185, "grad_norm": 2.4584438800811768, "learning_rate": 2.2659021115909586e-05, "loss": 0.5353, "num_input_tokens_seen": 28717136, "step": 21195 }, { "epoch": 0.6785737148710069, "grad_norm": 0.28509923815727234, "learning_rate": 2.2649246955637847e-05, "loss": 0.3381, "num_input_tokens_seen": 28723904, "step": 21200 }, { "epoch": 0.6785737148710069, "eval_loss": 0.457673043012619, "eval_runtime": 502.6962, "eval_samples_per_second": 27.623, "eval_steps_per_second": 13.812, "num_input_tokens_seen": 28723904, "step": 21200 }, { "epoch": 0.6787337558414954, "grad_norm": 0.8449125289916992, "learning_rate": 2.2639473157881766e-05, "loss": 0.3606, "num_input_tokens_seen": 28730512, "step": 21205 }, { "epoch": 0.6788937968119839, "grad_norm": 1.1974719762802124, "learning_rate": 2.2629699724148594e-05, "loss": 0.6921, "num_input_tokens_seen": 28737232, "step": 21210 }, { "epoch": 0.6790538377824723, "grad_norm": 0.995990514755249, "learning_rate": 2.26199266559455e-05, "loss": 0.4656, "num_input_tokens_seen": 28744608, "step": 21215 }, { "epoch": 0.6792138787529608, "grad_norm": 3.0710527896881104, "learning_rate": 2.2610153954779625e-05, "loss": 0.3454, "num_input_tokens_seen": 28751776, "step": 21220 }, { "epoch": 0.6793739197234492, "grad_norm": 2.570422649383545, "learning_rate": 2.2600381622158056e-05, "loss": 0.6749, "num_input_tokens_seen": 28758672, "step": 21225 }, { "epoch": 0.6795339606939377, "grad_norm": 1.0665867328643799, "learning_rate": 2.2590609659587783e-05, "loss": 0.5, "num_input_tokens_seen": 28765440, "step": 21230 }, { "epoch": 0.6796940016644261, "grad_norm": 0.9047220945358276, "learning_rate": 2.2580838068575787e-05, "loss": 0.4123, "num_input_tokens_seen": 28772240, "step": 21235 }, { "epoch": 0.6798540426349146, "grad_norm": 1.2000715732574463, "learning_rate": 2.257106685062896e-05, "loss": 0.5945, "num_input_tokens_seen": 28779072, "step": 21240 }, { "epoch": 0.6800140836054029, "grad_norm": 0.7811800241470337, "learning_rate": 2.256129600725415e-05, "loss": 0.4416, "num_input_tokens_seen": 28785904, "step": 21245 }, { "epoch": 0.6801741245758914, "grad_norm": 1.2652956247329712, "learning_rate": 2.2551525539958145e-05, "loss": 0.4579, "num_input_tokens_seen": 28792528, "step": 21250 }, { "epoch": 0.6803341655463798, "grad_norm": 0.8846961855888367, "learning_rate": 2.2541755450247663e-05, "loss": 0.5068, "num_input_tokens_seen": 28799344, "step": 21255 }, { "epoch": 0.6804942065168683, "grad_norm": 1.6842089891433716, "learning_rate": 2.2531985739629382e-05, "loss": 0.5164, "num_input_tokens_seen": 28806016, "step": 21260 }, { "epoch": 0.6806542474873568, "grad_norm": 1.1834909915924072, "learning_rate": 2.2522216409609924e-05, "loss": 0.4271, "num_input_tokens_seen": 28813056, "step": 21265 }, { "epoch": 0.6808142884578452, "grad_norm": 3.060763359069824, "learning_rate": 2.2512447461695826e-05, "loss": 0.634, "num_input_tokens_seen": 28819904, "step": 21270 }, { "epoch": 0.6809743294283337, "grad_norm": 2.1217668056488037, "learning_rate": 2.2502678897393593e-05, "loss": 0.4831, "num_input_tokens_seen": 28826656, "step": 21275 }, { "epoch": 0.6811343703988221, "grad_norm": 0.8291046023368835, "learning_rate": 2.2492910718209665e-05, "loss": 0.5563, "num_input_tokens_seen": 28833392, "step": 21280 }, { "epoch": 0.6812944113693106, "grad_norm": 0.7641519904136658, "learning_rate": 2.2483142925650398e-05, "loss": 0.4292, "num_input_tokens_seen": 28840016, "step": 21285 }, { "epoch": 0.681454452339799, "grad_norm": 0.6318883895874023, "learning_rate": 2.247337552122213e-05, "loss": 0.5432, "num_input_tokens_seen": 28846304, "step": 21290 }, { "epoch": 0.6816144933102875, "grad_norm": 0.8163531422615051, "learning_rate": 2.24636085064311e-05, "loss": 0.5886, "num_input_tokens_seen": 28853040, "step": 21295 }, { "epoch": 0.6817745342807758, "grad_norm": 1.0244183540344238, "learning_rate": 2.245384188278351e-05, "loss": 0.3975, "num_input_tokens_seen": 28859584, "step": 21300 }, { "epoch": 0.6819345752512643, "grad_norm": 0.9370933771133423, "learning_rate": 2.2444075651785513e-05, "loss": 0.4419, "num_input_tokens_seen": 28865952, "step": 21305 }, { "epoch": 0.6820946162217527, "grad_norm": 0.9829869270324707, "learning_rate": 2.243430981494316e-05, "loss": 0.4167, "num_input_tokens_seen": 28872640, "step": 21310 }, { "epoch": 0.6822546571922412, "grad_norm": 0.7440448999404907, "learning_rate": 2.2424544373762475e-05, "loss": 0.4339, "num_input_tokens_seen": 28879360, "step": 21315 }, { "epoch": 0.6824146981627297, "grad_norm": 1.3838633298873901, "learning_rate": 2.2414779329749418e-05, "loss": 0.4227, "num_input_tokens_seen": 28885952, "step": 21320 }, { "epoch": 0.6825747391332181, "grad_norm": 0.959227442741394, "learning_rate": 2.2405014684409873e-05, "loss": 0.4992, "num_input_tokens_seen": 28892896, "step": 21325 }, { "epoch": 0.6827347801037066, "grad_norm": 1.30539071559906, "learning_rate": 2.239525043924968e-05, "loss": 0.4274, "num_input_tokens_seen": 28899488, "step": 21330 }, { "epoch": 0.682894821074195, "grad_norm": 1.717134952545166, "learning_rate": 2.2385486595774592e-05, "loss": 0.4037, "num_input_tokens_seen": 28906320, "step": 21335 }, { "epoch": 0.6830548620446835, "grad_norm": 1.2643247842788696, "learning_rate": 2.237572315549033e-05, "loss": 0.5775, "num_input_tokens_seen": 28912912, "step": 21340 }, { "epoch": 0.6832149030151718, "grad_norm": 0.9388083815574646, "learning_rate": 2.2365960119902545e-05, "loss": 0.4143, "num_input_tokens_seen": 28919328, "step": 21345 }, { "epoch": 0.6833749439856603, "grad_norm": 1.456993818283081, "learning_rate": 2.2356197490516806e-05, "loss": 0.5424, "num_input_tokens_seen": 28926016, "step": 21350 }, { "epoch": 0.6835349849561487, "grad_norm": 0.8241636753082275, "learning_rate": 2.234643526883863e-05, "loss": 0.3558, "num_input_tokens_seen": 28932784, "step": 21355 }, { "epoch": 0.6836950259266372, "grad_norm": 0.9469195604324341, "learning_rate": 2.2336673456373497e-05, "loss": 0.4007, "num_input_tokens_seen": 28939776, "step": 21360 }, { "epoch": 0.6838550668971257, "grad_norm": 2.4584529399871826, "learning_rate": 2.2326912054626772e-05, "loss": 0.4634, "num_input_tokens_seen": 28946512, "step": 21365 }, { "epoch": 0.6840151078676141, "grad_norm": 0.8919292688369751, "learning_rate": 2.2317151065103813e-05, "loss": 0.3991, "num_input_tokens_seen": 28953408, "step": 21370 }, { "epoch": 0.6841751488381026, "grad_norm": 0.7790409326553345, "learning_rate": 2.2307390489309865e-05, "loss": 0.2998, "num_input_tokens_seen": 28960032, "step": 21375 }, { "epoch": 0.684335189808591, "grad_norm": 0.7634439468383789, "learning_rate": 2.2297630328750146e-05, "loss": 0.4561, "num_input_tokens_seen": 28966992, "step": 21380 }, { "epoch": 0.6844952307790795, "grad_norm": 0.9198728799819946, "learning_rate": 2.228787058492979e-05, "loss": 0.3393, "num_input_tokens_seen": 28973680, "step": 21385 }, { "epoch": 0.6846552717495679, "grad_norm": 1.184288501739502, "learning_rate": 2.2278111259353875e-05, "loss": 0.5542, "num_input_tokens_seen": 28980624, "step": 21390 }, { "epoch": 0.6848153127200564, "grad_norm": 0.7562806010246277, "learning_rate": 2.2268352353527395e-05, "loss": 0.3422, "num_input_tokens_seen": 28987408, "step": 21395 }, { "epoch": 0.6849753536905447, "grad_norm": 1.0730257034301758, "learning_rate": 2.225859386895533e-05, "loss": 0.6385, "num_input_tokens_seen": 28994096, "step": 21400 }, { "epoch": 0.6849753536905447, "eval_loss": 0.45760253071784973, "eval_runtime": 502.6832, "eval_samples_per_second": 27.624, "eval_steps_per_second": 13.812, "num_input_tokens_seen": 28994096, "step": 21400 }, { "epoch": 0.6851353946610332, "grad_norm": 1.176235318183899, "learning_rate": 2.2248835807142525e-05, "loss": 0.4126, "num_input_tokens_seen": 29001456, "step": 21405 }, { "epoch": 0.6852954356315216, "grad_norm": 2.318406105041504, "learning_rate": 2.2239078169593826e-05, "loss": 0.528, "num_input_tokens_seen": 29008448, "step": 21410 }, { "epoch": 0.6854554766020101, "grad_norm": 1.4872275590896606, "learning_rate": 2.222932095781396e-05, "loss": 0.2944, "num_input_tokens_seen": 29016000, "step": 21415 }, { "epoch": 0.6856155175724986, "grad_norm": 0.9234164953231812, "learning_rate": 2.221956417330762e-05, "loss": 0.2117, "num_input_tokens_seen": 29022784, "step": 21420 }, { "epoch": 0.685775558542987, "grad_norm": 0.9660922884941101, "learning_rate": 2.2209807817579438e-05, "loss": 0.3825, "num_input_tokens_seen": 29030080, "step": 21425 }, { "epoch": 0.6859355995134755, "grad_norm": 1.4086791276931763, "learning_rate": 2.220005189213394e-05, "loss": 0.4671, "num_input_tokens_seen": 29036752, "step": 21430 }, { "epoch": 0.6860956404839639, "grad_norm": 1.9710803031921387, "learning_rate": 2.2190296398475624e-05, "loss": 0.302, "num_input_tokens_seen": 29043616, "step": 21435 }, { "epoch": 0.6862556814544524, "grad_norm": 0.9519757032394409, "learning_rate": 2.2180541338108926e-05, "loss": 0.4063, "num_input_tokens_seen": 29050288, "step": 21440 }, { "epoch": 0.6864157224249408, "grad_norm": 0.9607702493667603, "learning_rate": 2.2170786712538176e-05, "loss": 0.5087, "num_input_tokens_seen": 29056960, "step": 21445 }, { "epoch": 0.6865757633954293, "grad_norm": 0.9382809996604919, "learning_rate": 2.216103252326768e-05, "loss": 0.3756, "num_input_tokens_seen": 29063872, "step": 21450 }, { "epoch": 0.6867358043659176, "grad_norm": 1.3597066402435303, "learning_rate": 2.2151278771801635e-05, "loss": 0.3576, "num_input_tokens_seen": 29070400, "step": 21455 }, { "epoch": 0.6868958453364061, "grad_norm": 0.9553912878036499, "learning_rate": 2.21415254596442e-05, "loss": 0.4234, "num_input_tokens_seen": 29077344, "step": 21460 }, { "epoch": 0.6870558863068945, "grad_norm": 0.6661427021026611, "learning_rate": 2.213177258829947e-05, "loss": 0.3505, "num_input_tokens_seen": 29084048, "step": 21465 }, { "epoch": 0.687215927277383, "grad_norm": 0.7686975002288818, "learning_rate": 2.2122020159271445e-05, "loss": 0.538, "num_input_tokens_seen": 29091008, "step": 21470 }, { "epoch": 0.6873759682478715, "grad_norm": 2.3698136806488037, "learning_rate": 2.2112268174064075e-05, "loss": 0.4397, "num_input_tokens_seen": 29097792, "step": 21475 }, { "epoch": 0.6875360092183599, "grad_norm": 2.151636838912964, "learning_rate": 2.2102516634181253e-05, "loss": 0.4422, "num_input_tokens_seen": 29104640, "step": 21480 }, { "epoch": 0.6876960501888484, "grad_norm": 0.8113951683044434, "learning_rate": 2.209276554112677e-05, "loss": 0.5004, "num_input_tokens_seen": 29111568, "step": 21485 }, { "epoch": 0.6878560911593368, "grad_norm": 1.2703937292099, "learning_rate": 2.2083014896404384e-05, "loss": 0.6009, "num_input_tokens_seen": 29118432, "step": 21490 }, { "epoch": 0.6880161321298253, "grad_norm": 0.7546266913414001, "learning_rate": 2.207326470151775e-05, "loss": 0.3802, "num_input_tokens_seen": 29124976, "step": 21495 }, { "epoch": 0.6881761731003136, "grad_norm": 1.2853213548660278, "learning_rate": 2.2063514957970477e-05, "loss": 0.5387, "num_input_tokens_seen": 29131488, "step": 21500 }, { "epoch": 0.6883362140708021, "grad_norm": 1.3504544496536255, "learning_rate": 2.205376566726611e-05, "loss": 0.4608, "num_input_tokens_seen": 29138672, "step": 21505 }, { "epoch": 0.6884962550412905, "grad_norm": 2.489807367324829, "learning_rate": 2.204401683090809e-05, "loss": 0.4243, "num_input_tokens_seen": 29145440, "step": 21510 }, { "epoch": 0.688656296011779, "grad_norm": 2.184542417526245, "learning_rate": 2.203426845039982e-05, "loss": 0.5283, "num_input_tokens_seen": 29152192, "step": 21515 }, { "epoch": 0.6888163369822674, "grad_norm": 1.2406383752822876, "learning_rate": 2.202452052724464e-05, "loss": 0.4108, "num_input_tokens_seen": 29159088, "step": 21520 }, { "epoch": 0.6889763779527559, "grad_norm": 1.0956746339797974, "learning_rate": 2.2014773062945777e-05, "loss": 0.5834, "num_input_tokens_seen": 29165632, "step": 21525 }, { "epoch": 0.6891364189232444, "grad_norm": 1.7474130392074585, "learning_rate": 2.2005026059006427e-05, "loss": 0.6546, "num_input_tokens_seen": 29172432, "step": 21530 }, { "epoch": 0.6892964598937328, "grad_norm": 0.9825288653373718, "learning_rate": 2.1995279516929695e-05, "loss": 0.4055, "num_input_tokens_seen": 29178992, "step": 21535 }, { "epoch": 0.6894565008642213, "grad_norm": 0.8008619546890259, "learning_rate": 2.1985533438218613e-05, "loss": 0.4266, "num_input_tokens_seen": 29185872, "step": 21540 }, { "epoch": 0.6896165418347097, "grad_norm": 0.989151120185852, "learning_rate": 2.197578782437617e-05, "loss": 0.341, "num_input_tokens_seen": 29192544, "step": 21545 }, { "epoch": 0.6897765828051982, "grad_norm": 1.0652811527252197, "learning_rate": 2.196604267690524e-05, "loss": 0.3611, "num_input_tokens_seen": 29199040, "step": 21550 }, { "epoch": 0.6899366237756865, "grad_norm": 3.609128713607788, "learning_rate": 2.195629799730865e-05, "loss": 0.6273, "num_input_tokens_seen": 29205840, "step": 21555 }, { "epoch": 0.690096664746175, "grad_norm": 1.2655434608459473, "learning_rate": 2.1946553787089173e-05, "loss": 0.6441, "num_input_tokens_seen": 29212336, "step": 21560 }, { "epoch": 0.6902567057166634, "grad_norm": 1.2770289182662964, "learning_rate": 2.193681004774947e-05, "loss": 0.3324, "num_input_tokens_seen": 29219136, "step": 21565 }, { "epoch": 0.6904167466871519, "grad_norm": 2.1873297691345215, "learning_rate": 2.1927066780792154e-05, "loss": 0.4182, "num_input_tokens_seen": 29226480, "step": 21570 }, { "epoch": 0.6905767876576403, "grad_norm": 1.6228234767913818, "learning_rate": 2.191732398771975e-05, "loss": 0.4133, "num_input_tokens_seen": 29233376, "step": 21575 }, { "epoch": 0.6907368286281288, "grad_norm": 1.366385817527771, "learning_rate": 2.1907581670034725e-05, "loss": 0.4681, "num_input_tokens_seen": 29240464, "step": 21580 }, { "epoch": 0.6908968695986173, "grad_norm": 0.9923104047775269, "learning_rate": 2.189783982923948e-05, "loss": 0.3455, "num_input_tokens_seen": 29247360, "step": 21585 }, { "epoch": 0.6910569105691057, "grad_norm": 0.2734001576900482, "learning_rate": 2.1888098466836303e-05, "loss": 0.3641, "num_input_tokens_seen": 29254112, "step": 21590 }, { "epoch": 0.6912169515395942, "grad_norm": 2.227954387664795, "learning_rate": 2.1878357584327457e-05, "loss": 0.4945, "num_input_tokens_seen": 29260960, "step": 21595 }, { "epoch": 0.6913769925100826, "grad_norm": 2.145073652267456, "learning_rate": 2.1868617183215103e-05, "loss": 0.5204, "num_input_tokens_seen": 29267904, "step": 21600 }, { "epoch": 0.6913769925100826, "eval_loss": 0.4570278823375702, "eval_runtime": 502.5083, "eval_samples_per_second": 27.633, "eval_steps_per_second": 13.817, "num_input_tokens_seen": 29267904, "step": 21600 }, { "epoch": 0.691537033480571, "grad_norm": 1.7431622743606567, "learning_rate": 2.1858877265001327e-05, "loss": 0.3946, "num_input_tokens_seen": 29274496, "step": 21605 }, { "epoch": 0.6916970744510594, "grad_norm": 0.7392350435256958, "learning_rate": 2.184913783118816e-05, "loss": 0.2856, "num_input_tokens_seen": 29281120, "step": 21610 }, { "epoch": 0.6918571154215479, "grad_norm": 0.7555866837501526, "learning_rate": 2.1839398883277522e-05, "loss": 0.4543, "num_input_tokens_seen": 29287808, "step": 21615 }, { "epoch": 0.6920171563920363, "grad_norm": 1.1804871559143066, "learning_rate": 2.182966042277129e-05, "loss": 0.4741, "num_input_tokens_seen": 29294640, "step": 21620 }, { "epoch": 0.6921771973625248, "grad_norm": 1.0550107955932617, "learning_rate": 2.181992245117128e-05, "loss": 0.4501, "num_input_tokens_seen": 29301536, "step": 21625 }, { "epoch": 0.6923372383330133, "grad_norm": 0.6082996726036072, "learning_rate": 2.181018496997918e-05, "loss": 0.3596, "num_input_tokens_seen": 29308432, "step": 21630 }, { "epoch": 0.6924972793035017, "grad_norm": 1.1398391723632812, "learning_rate": 2.1800447980696648e-05, "loss": 0.3484, "num_input_tokens_seen": 29315152, "step": 21635 }, { "epoch": 0.6926573202739902, "grad_norm": 1.0742075443267822, "learning_rate": 2.1790711484825248e-05, "loss": 0.3885, "num_input_tokens_seen": 29322032, "step": 21640 }, { "epoch": 0.6928173612444786, "grad_norm": 1.3343851566314697, "learning_rate": 2.178097548386646e-05, "loss": 0.4578, "num_input_tokens_seen": 29328848, "step": 21645 }, { "epoch": 0.6929774022149671, "grad_norm": 0.5552470088005066, "learning_rate": 2.1771239979321712e-05, "loss": 0.4059, "num_input_tokens_seen": 29335504, "step": 21650 }, { "epoch": 0.6931374431854554, "grad_norm": 1.1492058038711548, "learning_rate": 2.1761504972692327e-05, "loss": 0.3904, "num_input_tokens_seen": 29342368, "step": 21655 }, { "epoch": 0.6932974841559439, "grad_norm": 0.5526182651519775, "learning_rate": 2.1751770465479572e-05, "loss": 0.331, "num_input_tokens_seen": 29349296, "step": 21660 }, { "epoch": 0.6934575251264323, "grad_norm": 0.9947152137756348, "learning_rate": 2.174203645918464e-05, "loss": 0.429, "num_input_tokens_seen": 29355904, "step": 21665 }, { "epoch": 0.6936175660969208, "grad_norm": 0.9824190735816956, "learning_rate": 2.1732302955308624e-05, "loss": 0.3796, "num_input_tokens_seen": 29362624, "step": 21670 }, { "epoch": 0.6937776070674092, "grad_norm": 1.0688307285308838, "learning_rate": 2.172256995535255e-05, "loss": 0.4494, "num_input_tokens_seen": 29369680, "step": 21675 }, { "epoch": 0.6939376480378977, "grad_norm": 0.4111238420009613, "learning_rate": 2.171283746081739e-05, "loss": 0.2147, "num_input_tokens_seen": 29376480, "step": 21680 }, { "epoch": 0.6940976890083862, "grad_norm": 1.1220698356628418, "learning_rate": 2.1703105473203988e-05, "loss": 0.5384, "num_input_tokens_seen": 29383056, "step": 21685 }, { "epoch": 0.6942577299788746, "grad_norm": 1.0830365419387817, "learning_rate": 2.1693373994013168e-05, "loss": 0.4923, "num_input_tokens_seen": 29390032, "step": 21690 }, { "epoch": 0.6944177709493631, "grad_norm": 0.9912424087524414, "learning_rate": 2.168364302474562e-05, "loss": 0.6117, "num_input_tokens_seen": 29396688, "step": 21695 }, { "epoch": 0.6945778119198515, "grad_norm": 1.376664161682129, "learning_rate": 2.167391256690199e-05, "loss": 0.5066, "num_input_tokens_seen": 29403408, "step": 21700 }, { "epoch": 0.69473785289034, "grad_norm": 1.1614609956741333, "learning_rate": 2.1664182621982855e-05, "loss": 0.5257, "num_input_tokens_seen": 29410048, "step": 21705 }, { "epoch": 0.6948978938608283, "grad_norm": 1.3544764518737793, "learning_rate": 2.1654453191488673e-05, "loss": 0.354, "num_input_tokens_seen": 29416576, "step": 21710 }, { "epoch": 0.6950579348313168, "grad_norm": 0.3963971734046936, "learning_rate": 2.1644724276919846e-05, "loss": 0.3838, "num_input_tokens_seen": 29423120, "step": 21715 }, { "epoch": 0.6952179758018052, "grad_norm": 0.606066107749939, "learning_rate": 2.1634995879776715e-05, "loss": 0.4122, "num_input_tokens_seen": 29429568, "step": 21720 }, { "epoch": 0.6953780167722937, "grad_norm": 1.1019282341003418, "learning_rate": 2.162526800155949e-05, "loss": 0.4057, "num_input_tokens_seen": 29436176, "step": 21725 }, { "epoch": 0.6955380577427821, "grad_norm": 0.863886833190918, "learning_rate": 2.1615540643768363e-05, "loss": 0.4522, "num_input_tokens_seen": 29443872, "step": 21730 }, { "epoch": 0.6956980987132706, "grad_norm": 1.2731515169143677, "learning_rate": 2.160581380790339e-05, "loss": 0.4953, "num_input_tokens_seen": 29450944, "step": 21735 }, { "epoch": 0.6958581396837591, "grad_norm": 0.7245847582817078, "learning_rate": 2.1596087495464586e-05, "loss": 0.4781, "num_input_tokens_seen": 29458240, "step": 21740 }, { "epoch": 0.6960181806542475, "grad_norm": 1.287158489227295, "learning_rate": 2.1586361707951866e-05, "loss": 0.3613, "num_input_tokens_seen": 29465376, "step": 21745 }, { "epoch": 0.696178221624736, "grad_norm": 1.6101305484771729, "learning_rate": 2.157663644686507e-05, "loss": 0.4239, "num_input_tokens_seen": 29472080, "step": 21750 }, { "epoch": 0.6963382625952244, "grad_norm": 0.8269065022468567, "learning_rate": 2.156691171370396e-05, "loss": 0.3249, "num_input_tokens_seen": 29479168, "step": 21755 }, { "epoch": 0.6964983035657129, "grad_norm": 1.505050778388977, "learning_rate": 2.1557187509968195e-05, "loss": 0.5778, "num_input_tokens_seen": 29485984, "step": 21760 }, { "epoch": 0.6966583445362012, "grad_norm": 3.5188021659851074, "learning_rate": 2.1547463837157382e-05, "loss": 0.7205, "num_input_tokens_seen": 29492624, "step": 21765 }, { "epoch": 0.6968183855066897, "grad_norm": 1.2641348838806152, "learning_rate": 2.1537740696771045e-05, "loss": 0.3594, "num_input_tokens_seen": 29499504, "step": 21770 }, { "epoch": 0.6969784264771781, "grad_norm": 0.9232179522514343, "learning_rate": 2.1528018090308587e-05, "loss": 0.4255, "num_input_tokens_seen": 29506672, "step": 21775 }, { "epoch": 0.6971384674476666, "grad_norm": 0.8403621912002563, "learning_rate": 2.151829601926938e-05, "loss": 0.4089, "num_input_tokens_seen": 29513488, "step": 21780 }, { "epoch": 0.697298508418155, "grad_norm": 0.9264079332351685, "learning_rate": 2.1508574485152684e-05, "loss": 0.3548, "num_input_tokens_seen": 29520112, "step": 21785 }, { "epoch": 0.6974585493886435, "grad_norm": 0.8086733818054199, "learning_rate": 2.1498853489457667e-05, "loss": 0.4886, "num_input_tokens_seen": 29526768, "step": 21790 }, { "epoch": 0.697618590359132, "grad_norm": 0.8741891980171204, "learning_rate": 2.1489133033683455e-05, "loss": 0.3511, "num_input_tokens_seen": 29533664, "step": 21795 }, { "epoch": 0.6977786313296204, "grad_norm": 1.0420881509780884, "learning_rate": 2.1479413119329038e-05, "loss": 0.3454, "num_input_tokens_seen": 29540768, "step": 21800 }, { "epoch": 0.6977786313296204, "eval_loss": 0.45701009035110474, "eval_runtime": 502.6036, "eval_samples_per_second": 27.628, "eval_steps_per_second": 13.814, "num_input_tokens_seen": 29540768, "step": 21800 }, { "epoch": 0.6979386723001089, "grad_norm": 0.7651846408843994, "learning_rate": 2.1469693747893355e-05, "loss": 0.5083, "num_input_tokens_seen": 29547536, "step": 21805 }, { "epoch": 0.6980987132705972, "grad_norm": 1.1068943738937378, "learning_rate": 2.1459974920875274e-05, "loss": 0.5288, "num_input_tokens_seen": 29554128, "step": 21810 }, { "epoch": 0.6982587542410857, "grad_norm": 0.8256592154502869, "learning_rate": 2.145025663977354e-05, "loss": 0.5534, "num_input_tokens_seen": 29560864, "step": 21815 }, { "epoch": 0.6984187952115741, "grad_norm": 1.2968958616256714, "learning_rate": 2.1440538906086844e-05, "loss": 0.5398, "num_input_tokens_seen": 29567088, "step": 21820 }, { "epoch": 0.6985788361820626, "grad_norm": 0.281297892332077, "learning_rate": 2.1430821721313782e-05, "loss": 0.2695, "num_input_tokens_seen": 29573984, "step": 21825 }, { "epoch": 0.698738877152551, "grad_norm": 1.2859023809432983, "learning_rate": 2.142110508695286e-05, "loss": 0.3721, "num_input_tokens_seen": 29581104, "step": 21830 }, { "epoch": 0.6988989181230395, "grad_norm": 0.826005220413208, "learning_rate": 2.1411389004502515e-05, "loss": 0.407, "num_input_tokens_seen": 29587968, "step": 21835 }, { "epoch": 0.699058959093528, "grad_norm": 0.871752917766571, "learning_rate": 2.140167347546107e-05, "loss": 0.4777, "num_input_tokens_seen": 29594720, "step": 21840 }, { "epoch": 0.6992190000640164, "grad_norm": 1.1230984926223755, "learning_rate": 2.1391958501326793e-05, "loss": 0.506, "num_input_tokens_seen": 29601648, "step": 21845 }, { "epoch": 0.6993790410345049, "grad_norm": 1.473122000694275, "learning_rate": 2.1382244083597873e-05, "loss": 0.3501, "num_input_tokens_seen": 29608032, "step": 21850 }, { "epoch": 0.6995390820049933, "grad_norm": 1.3639519214630127, "learning_rate": 2.137253022377237e-05, "loss": 0.3713, "num_input_tokens_seen": 29614336, "step": 21855 }, { "epoch": 0.6996991229754818, "grad_norm": 1.6808404922485352, "learning_rate": 2.136281692334829e-05, "loss": 0.4813, "num_input_tokens_seen": 29621808, "step": 21860 }, { "epoch": 0.6998591639459701, "grad_norm": 0.8794233798980713, "learning_rate": 2.135310418382356e-05, "loss": 0.3213, "num_input_tokens_seen": 29628624, "step": 21865 }, { "epoch": 0.7000192049164586, "grad_norm": 1.3569539785385132, "learning_rate": 2.134339200669598e-05, "loss": 0.5131, "num_input_tokens_seen": 29635376, "step": 21870 }, { "epoch": 0.700179245886947, "grad_norm": 0.825372040271759, "learning_rate": 2.133368039346331e-05, "loss": 0.5141, "num_input_tokens_seen": 29642368, "step": 21875 }, { "epoch": 0.7003392868574355, "grad_norm": 1.1431909799575806, "learning_rate": 2.1323969345623195e-05, "loss": 0.4999, "num_input_tokens_seen": 29649264, "step": 21880 }, { "epoch": 0.7004993278279239, "grad_norm": 0.942758321762085, "learning_rate": 2.1314258864673207e-05, "loss": 0.4558, "num_input_tokens_seen": 29655664, "step": 21885 }, { "epoch": 0.7006593687984124, "grad_norm": 2.8122167587280273, "learning_rate": 2.130454895211082e-05, "loss": 0.4603, "num_input_tokens_seen": 29662336, "step": 21890 }, { "epoch": 0.7008194097689009, "grad_norm": 0.8863010406494141, "learning_rate": 2.129483960943342e-05, "loss": 0.4068, "num_input_tokens_seen": 29669072, "step": 21895 }, { "epoch": 0.7009794507393893, "grad_norm": 0.8029206395149231, "learning_rate": 2.128513083813831e-05, "loss": 0.3543, "num_input_tokens_seen": 29676336, "step": 21900 }, { "epoch": 0.7011394917098778, "grad_norm": 1.0152686834335327, "learning_rate": 2.1275422639722724e-05, "loss": 0.4699, "num_input_tokens_seen": 29683104, "step": 21905 }, { "epoch": 0.7012995326803662, "grad_norm": 0.5898716449737549, "learning_rate": 2.126571501568376e-05, "loss": 0.432, "num_input_tokens_seen": 29690256, "step": 21910 }, { "epoch": 0.7014595736508547, "grad_norm": 1.5407851934432983, "learning_rate": 2.1256007967518478e-05, "loss": 0.4161, "num_input_tokens_seen": 29697040, "step": 21915 }, { "epoch": 0.701619614621343, "grad_norm": 0.916541576385498, "learning_rate": 2.124630149672381e-05, "loss": 0.4894, "num_input_tokens_seen": 29703984, "step": 21920 }, { "epoch": 0.7017796555918315, "grad_norm": 0.7216461300849915, "learning_rate": 2.1236595604796624e-05, "loss": 0.3611, "num_input_tokens_seen": 29710736, "step": 21925 }, { "epoch": 0.7019396965623199, "grad_norm": 0.7638271450996399, "learning_rate": 2.1226890293233693e-05, "loss": 0.3927, "num_input_tokens_seen": 29717040, "step": 21930 }, { "epoch": 0.7020997375328084, "grad_norm": 3.2947487831115723, "learning_rate": 2.1217185563531694e-05, "loss": 0.6172, "num_input_tokens_seen": 29723152, "step": 21935 }, { "epoch": 0.7022597785032968, "grad_norm": 1.7101643085479736, "learning_rate": 2.120748141718721e-05, "loss": 0.4034, "num_input_tokens_seen": 29729936, "step": 21940 }, { "epoch": 0.7024198194737853, "grad_norm": 1.856994867324829, "learning_rate": 2.1197777855696765e-05, "loss": 0.5477, "num_input_tokens_seen": 29736736, "step": 21945 }, { "epoch": 0.7025798604442738, "grad_norm": 3.1223533153533936, "learning_rate": 2.1188074880556746e-05, "loss": 0.5024, "num_input_tokens_seen": 29743392, "step": 21950 }, { "epoch": 0.7027399014147622, "grad_norm": 0.8338295817375183, "learning_rate": 2.1178372493263495e-05, "loss": 0.4889, "num_input_tokens_seen": 29750096, "step": 21955 }, { "epoch": 0.7028999423852507, "grad_norm": 1.1259950399398804, "learning_rate": 2.116867069531322e-05, "loss": 0.385, "num_input_tokens_seen": 29757184, "step": 21960 }, { "epoch": 0.703059983355739, "grad_norm": 0.6765305995941162, "learning_rate": 2.1158969488202073e-05, "loss": 0.3568, "num_input_tokens_seen": 29763600, "step": 21965 }, { "epoch": 0.7032200243262275, "grad_norm": 1.3507922887802124, "learning_rate": 2.114926887342611e-05, "loss": 0.3448, "num_input_tokens_seen": 29770304, "step": 21970 }, { "epoch": 0.7033800652967159, "grad_norm": 1.489423394203186, "learning_rate": 2.113956885248127e-05, "loss": 0.4565, "num_input_tokens_seen": 29777424, "step": 21975 }, { "epoch": 0.7035401062672044, "grad_norm": 2.435607433319092, "learning_rate": 2.112986942686342e-05, "loss": 0.5437, "num_input_tokens_seen": 29784912, "step": 21980 }, { "epoch": 0.7037001472376928, "grad_norm": 1.620453953742981, "learning_rate": 2.112017059806835e-05, "loss": 0.411, "num_input_tokens_seen": 29791616, "step": 21985 }, { "epoch": 0.7038601882081813, "grad_norm": 0.8055664896965027, "learning_rate": 2.1110472367591724e-05, "loss": 0.6303, "num_input_tokens_seen": 29798432, "step": 21990 }, { "epoch": 0.7040202291786697, "grad_norm": 1.0610816478729248, "learning_rate": 2.1100774736929145e-05, "loss": 0.4416, "num_input_tokens_seen": 29805168, "step": 21995 }, { "epoch": 0.7041802701491582, "grad_norm": 0.6009805798530579, "learning_rate": 2.10910777075761e-05, "loss": 0.4744, "num_input_tokens_seen": 29812480, "step": 22000 }, { "epoch": 0.7041802701491582, "eval_loss": 0.4565279185771942, "eval_runtime": 502.8166, "eval_samples_per_second": 27.616, "eval_steps_per_second": 13.808, "num_input_tokens_seen": 29812480, "step": 22000 }, { "epoch": 0.7043403111196467, "grad_norm": 0.9534848928451538, "learning_rate": 2.108138128102799e-05, "loss": 0.5268, "num_input_tokens_seen": 29819264, "step": 22005 }, { "epoch": 0.7045003520901351, "grad_norm": 2.0086257457733154, "learning_rate": 2.107168545878014e-05, "loss": 0.6507, "num_input_tokens_seen": 29825808, "step": 22010 }, { "epoch": 0.7046603930606236, "grad_norm": 1.0710643529891968, "learning_rate": 2.106199024232775e-05, "loss": 0.5272, "num_input_tokens_seen": 29832640, "step": 22015 }, { "epoch": 0.7048204340311119, "grad_norm": 1.6261117458343506, "learning_rate": 2.105229563316595e-05, "loss": 0.4714, "num_input_tokens_seen": 29839904, "step": 22020 }, { "epoch": 0.7049804750016004, "grad_norm": 1.004955768585205, "learning_rate": 2.1042601632789784e-05, "loss": 0.3825, "num_input_tokens_seen": 29847248, "step": 22025 }, { "epoch": 0.7051405159720888, "grad_norm": 0.7845077514648438, "learning_rate": 2.103290824269417e-05, "loss": 0.3418, "num_input_tokens_seen": 29853840, "step": 22030 }, { "epoch": 0.7053005569425773, "grad_norm": 2.0189619064331055, "learning_rate": 2.1023215464373965e-05, "loss": 0.5284, "num_input_tokens_seen": 29860352, "step": 22035 }, { "epoch": 0.7054605979130657, "grad_norm": 1.4230380058288574, "learning_rate": 2.1013523299323908e-05, "loss": 0.4767, "num_input_tokens_seen": 29866512, "step": 22040 }, { "epoch": 0.7056206388835542, "grad_norm": 1.3062210083007812, "learning_rate": 2.1003831749038654e-05, "loss": 0.363, "num_input_tokens_seen": 29873424, "step": 22045 }, { "epoch": 0.7057806798540426, "grad_norm": 1.0758014917373657, "learning_rate": 2.099414081501277e-05, "loss": 0.4718, "num_input_tokens_seen": 29879856, "step": 22050 }, { "epoch": 0.7059407208245311, "grad_norm": 2.6005048751831055, "learning_rate": 2.09844504987407e-05, "loss": 0.3478, "num_input_tokens_seen": 29886816, "step": 22055 }, { "epoch": 0.7061007617950196, "grad_norm": 0.9969660043716431, "learning_rate": 2.097476080171683e-05, "loss": 0.4725, "num_input_tokens_seen": 29893728, "step": 22060 }, { "epoch": 0.706260802765508, "grad_norm": 0.9778667688369751, "learning_rate": 2.0965071725435436e-05, "loss": 0.4107, "num_input_tokens_seen": 29900096, "step": 22065 }, { "epoch": 0.7064208437359965, "grad_norm": 0.9715601801872253, "learning_rate": 2.0955383271390684e-05, "loss": 0.4204, "num_input_tokens_seen": 29907280, "step": 22070 }, { "epoch": 0.7065808847064848, "grad_norm": 0.5698919296264648, "learning_rate": 2.094569544107666e-05, "loss": 0.3525, "num_input_tokens_seen": 29913904, "step": 22075 }, { "epoch": 0.7067409256769733, "grad_norm": 2.006958484649658, "learning_rate": 2.093600823598735e-05, "loss": 0.3309, "num_input_tokens_seen": 29920608, "step": 22080 }, { "epoch": 0.7069009666474617, "grad_norm": 2.129295825958252, "learning_rate": 2.092632165761663e-05, "loss": 0.3685, "num_input_tokens_seen": 29927424, "step": 22085 }, { "epoch": 0.7070610076179502, "grad_norm": 0.7301648855209351, "learning_rate": 2.091663570745832e-05, "loss": 0.4918, "num_input_tokens_seen": 29934944, "step": 22090 }, { "epoch": 0.7072210485884386, "grad_norm": 1.7757843732833862, "learning_rate": 2.0906950387006086e-05, "loss": 0.389, "num_input_tokens_seen": 29941744, "step": 22095 }, { "epoch": 0.7073810895589271, "grad_norm": 0.7441300749778748, "learning_rate": 2.0897265697753543e-05, "loss": 0.31, "num_input_tokens_seen": 29948560, "step": 22100 }, { "epoch": 0.7075411305294156, "grad_norm": 1.0899310111999512, "learning_rate": 2.088758164119419e-05, "loss": 0.3309, "num_input_tokens_seen": 29955088, "step": 22105 }, { "epoch": 0.707701171499904, "grad_norm": 0.8804386258125305, "learning_rate": 2.0877898218821428e-05, "loss": 0.3963, "num_input_tokens_seen": 29961824, "step": 22110 }, { "epoch": 0.7078612124703925, "grad_norm": 0.6076482534408569, "learning_rate": 2.0868215432128565e-05, "loss": 0.4852, "num_input_tokens_seen": 29968240, "step": 22115 }, { "epoch": 0.7080212534408808, "grad_norm": 0.8119211792945862, "learning_rate": 2.0858533282608796e-05, "loss": 0.3015, "num_input_tokens_seen": 29974784, "step": 22120 }, { "epoch": 0.7081812944113693, "grad_norm": 1.0860369205474854, "learning_rate": 2.084885177175524e-05, "loss": 0.5204, "num_input_tokens_seen": 29981328, "step": 22125 }, { "epoch": 0.7083413353818577, "grad_norm": 1.8282768726348877, "learning_rate": 2.0839170901060917e-05, "loss": 0.507, "num_input_tokens_seen": 29988272, "step": 22130 }, { "epoch": 0.7085013763523462, "grad_norm": 0.7347087264060974, "learning_rate": 2.082949067201872e-05, "loss": 0.4845, "num_input_tokens_seen": 29994992, "step": 22135 }, { "epoch": 0.7086614173228346, "grad_norm": 1.155158519744873, "learning_rate": 2.0819811086121475e-05, "loss": 0.2958, "num_input_tokens_seen": 30001616, "step": 22140 }, { "epoch": 0.7088214582933231, "grad_norm": 1.800554871559143, "learning_rate": 2.08101321448619e-05, "loss": 0.361, "num_input_tokens_seen": 30008112, "step": 22145 }, { "epoch": 0.7089814992638115, "grad_norm": 1.2978497743606567, "learning_rate": 2.080045384973259e-05, "loss": 0.6266, "num_input_tokens_seen": 30014592, "step": 22150 }, { "epoch": 0.7091415402343, "grad_norm": 0.6640035510063171, "learning_rate": 2.0790776202226082e-05, "loss": 0.475, "num_input_tokens_seen": 30021440, "step": 22155 }, { "epoch": 0.7093015812047885, "grad_norm": 0.9764406681060791, "learning_rate": 2.078109920383477e-05, "loss": 0.364, "num_input_tokens_seen": 30027984, "step": 22160 }, { "epoch": 0.7094616221752769, "grad_norm": 1.1410423517227173, "learning_rate": 2.0771422856050978e-05, "loss": 0.499, "num_input_tokens_seen": 30035008, "step": 22165 }, { "epoch": 0.7096216631457654, "grad_norm": 1.6418325901031494, "learning_rate": 2.076174716036693e-05, "loss": 0.4139, "num_input_tokens_seen": 30041488, "step": 22170 }, { "epoch": 0.7097817041162537, "grad_norm": 1.6015722751617432, "learning_rate": 2.075207211827472e-05, "loss": 0.5127, "num_input_tokens_seen": 30048064, "step": 22175 }, { "epoch": 0.7099417450867422, "grad_norm": 0.46480536460876465, "learning_rate": 2.074239773126638e-05, "loss": 0.2017, "num_input_tokens_seen": 30054928, "step": 22180 }, { "epoch": 0.7101017860572306, "grad_norm": 1.1882027387619019, "learning_rate": 2.073272400083382e-05, "loss": 0.4818, "num_input_tokens_seen": 30061424, "step": 22185 }, { "epoch": 0.7102618270277191, "grad_norm": 0.9924299716949463, "learning_rate": 2.072305092846883e-05, "loss": 0.4074, "num_input_tokens_seen": 30067776, "step": 22190 }, { "epoch": 0.7104218679982075, "grad_norm": 0.9732227325439453, "learning_rate": 2.0713378515663152e-05, "loss": 0.4771, "num_input_tokens_seen": 30074304, "step": 22195 }, { "epoch": 0.710581908968696, "grad_norm": 0.6458679437637329, "learning_rate": 2.070370676390836e-05, "loss": 0.3103, "num_input_tokens_seen": 30080624, "step": 22200 }, { "epoch": 0.710581908968696, "eval_loss": 0.45582857728004456, "eval_runtime": 502.473, "eval_samples_per_second": 27.635, "eval_steps_per_second": 13.818, "num_input_tokens_seen": 30080624, "step": 22200 }, { "epoch": 0.7107419499391844, "grad_norm": 2.4121015071868896, "learning_rate": 2.0694035674695974e-05, "loss": 0.5287, "num_input_tokens_seen": 30087520, "step": 22205 }, { "epoch": 0.7109019909096729, "grad_norm": 2.210980176925659, "learning_rate": 2.0684365249517416e-05, "loss": 0.3972, "num_input_tokens_seen": 30093904, "step": 22210 }, { "epoch": 0.7110620318801614, "grad_norm": 0.3556768596172333, "learning_rate": 2.067469548986396e-05, "loss": 0.5842, "num_input_tokens_seen": 30100576, "step": 22215 }, { "epoch": 0.7112220728506498, "grad_norm": 0.7957329154014587, "learning_rate": 2.066502639722681e-05, "loss": 0.4538, "num_input_tokens_seen": 30107312, "step": 22220 }, { "epoch": 0.7113821138211383, "grad_norm": 1.3435372114181519, "learning_rate": 2.065535797309708e-05, "loss": 0.5324, "num_input_tokens_seen": 30114304, "step": 22225 }, { "epoch": 0.7115421547916266, "grad_norm": 1.5869349241256714, "learning_rate": 2.0645690218965736e-05, "loss": 0.549, "num_input_tokens_seen": 30121120, "step": 22230 }, { "epoch": 0.7117021957621151, "grad_norm": 1.018259048461914, "learning_rate": 2.063602313632369e-05, "loss": 0.3901, "num_input_tokens_seen": 30127664, "step": 22235 }, { "epoch": 0.7118622367326035, "grad_norm": 0.7801620364189148, "learning_rate": 2.0626356726661704e-05, "loss": 0.3973, "num_input_tokens_seen": 30135008, "step": 22240 }, { "epoch": 0.712022277703092, "grad_norm": 0.6564340591430664, "learning_rate": 2.0616690991470477e-05, "loss": 0.3417, "num_input_tokens_seen": 30141504, "step": 22245 }, { "epoch": 0.7121823186735804, "grad_norm": 2.423217535018921, "learning_rate": 2.0607025932240595e-05, "loss": 0.4318, "num_input_tokens_seen": 30147792, "step": 22250 }, { "epoch": 0.7123423596440689, "grad_norm": 0.9277958869934082, "learning_rate": 2.059736155046251e-05, "loss": 0.5195, "num_input_tokens_seen": 30154672, "step": 22255 }, { "epoch": 0.7125024006145573, "grad_norm": 0.5719887614250183, "learning_rate": 2.0587697847626603e-05, "loss": 0.4342, "num_input_tokens_seen": 30161216, "step": 22260 }, { "epoch": 0.7126624415850458, "grad_norm": 0.6912057399749756, "learning_rate": 2.057803482522314e-05, "loss": 0.3028, "num_input_tokens_seen": 30168016, "step": 22265 }, { "epoch": 0.7128224825555343, "grad_norm": 0.4421541392803192, "learning_rate": 2.056837248474227e-05, "loss": 0.4095, "num_input_tokens_seen": 30175312, "step": 22270 }, { "epoch": 0.7129825235260226, "grad_norm": 0.9962394833564758, "learning_rate": 2.0558710827674064e-05, "loss": 0.3263, "num_input_tokens_seen": 30181904, "step": 22275 }, { "epoch": 0.7131425644965111, "grad_norm": 0.8381558060646057, "learning_rate": 2.054904985550845e-05, "loss": 0.4665, "num_input_tokens_seen": 30188544, "step": 22280 }, { "epoch": 0.7133026054669995, "grad_norm": 0.5967469215393066, "learning_rate": 2.0539389569735287e-05, "loss": 0.4749, "num_input_tokens_seen": 30195232, "step": 22285 }, { "epoch": 0.713462646437488, "grad_norm": 1.500099778175354, "learning_rate": 2.052972997184431e-05, "loss": 0.6707, "num_input_tokens_seen": 30201680, "step": 22290 }, { "epoch": 0.7136226874079764, "grad_norm": 1.6594992876052856, "learning_rate": 2.0520071063325146e-05, "loss": 0.4164, "num_input_tokens_seen": 30208336, "step": 22295 }, { "epoch": 0.7137827283784649, "grad_norm": 0.9387924671173096, "learning_rate": 2.051041284566732e-05, "loss": 0.4005, "num_input_tokens_seen": 30215008, "step": 22300 }, { "epoch": 0.7139427693489533, "grad_norm": 1.9467695951461792, "learning_rate": 2.050075532036026e-05, "loss": 0.5366, "num_input_tokens_seen": 30221600, "step": 22305 }, { "epoch": 0.7141028103194418, "grad_norm": 0.7960267066955566, "learning_rate": 2.0491098488893264e-05, "loss": 0.3751, "num_input_tokens_seen": 30228768, "step": 22310 }, { "epoch": 0.7142628512899302, "grad_norm": 0.7942819595336914, "learning_rate": 2.0481442352755546e-05, "loss": 0.427, "num_input_tokens_seen": 30235424, "step": 22315 }, { "epoch": 0.7144228922604187, "grad_norm": 0.7265687584877014, "learning_rate": 2.0471786913436198e-05, "loss": 0.4418, "num_input_tokens_seen": 30242160, "step": 22320 }, { "epoch": 0.7145829332309072, "grad_norm": 1.1366403102874756, "learning_rate": 2.0462132172424218e-05, "loss": 0.4679, "num_input_tokens_seen": 30248928, "step": 22325 }, { "epoch": 0.7147429742013955, "grad_norm": 0.8196418881416321, "learning_rate": 2.0452478131208484e-05, "loss": 0.4323, "num_input_tokens_seen": 30255536, "step": 22330 }, { "epoch": 0.714903015171884, "grad_norm": 1.2295631170272827, "learning_rate": 2.0442824791277765e-05, "loss": 0.3591, "num_input_tokens_seen": 30262016, "step": 22335 }, { "epoch": 0.7150630561423724, "grad_norm": 0.9659268856048584, "learning_rate": 2.0433172154120727e-05, "loss": 0.3717, "num_input_tokens_seen": 30269200, "step": 22340 }, { "epoch": 0.7152230971128609, "grad_norm": 1.7553415298461914, "learning_rate": 2.0423520221225947e-05, "loss": 0.3371, "num_input_tokens_seen": 30275792, "step": 22345 }, { "epoch": 0.7153831380833493, "grad_norm": 2.3269670009613037, "learning_rate": 2.0413868994081848e-05, "loss": 0.5166, "num_input_tokens_seen": 30282720, "step": 22350 }, { "epoch": 0.7155431790538378, "grad_norm": 0.5317374467849731, "learning_rate": 2.0404218474176795e-05, "loss": 0.4067, "num_input_tokens_seen": 30289936, "step": 22355 }, { "epoch": 0.7157032200243262, "grad_norm": 0.43755269050598145, "learning_rate": 2.0394568662999002e-05, "loss": 0.4437, "num_input_tokens_seen": 30296368, "step": 22360 }, { "epoch": 0.7158632609948147, "grad_norm": 0.8474901914596558, "learning_rate": 2.0384919562036593e-05, "loss": 0.4262, "num_input_tokens_seen": 30303488, "step": 22365 }, { "epoch": 0.7160233019653032, "grad_norm": 3.1599459648132324, "learning_rate": 2.0375271172777593e-05, "loss": 0.3356, "num_input_tokens_seen": 30310816, "step": 22370 }, { "epoch": 0.7161833429357916, "grad_norm": 1.1983476877212524, "learning_rate": 2.0365623496709885e-05, "loss": 0.4149, "num_input_tokens_seen": 30317440, "step": 22375 }, { "epoch": 0.71634338390628, "grad_norm": 1.7862974405288696, "learning_rate": 2.0355976535321283e-05, "loss": 0.5453, "num_input_tokens_seen": 30324512, "step": 22380 }, { "epoch": 0.7165034248767684, "grad_norm": 1.3024381399154663, "learning_rate": 2.034633029009945e-05, "loss": 0.326, "num_input_tokens_seen": 30331136, "step": 22385 }, { "epoch": 0.7166634658472569, "grad_norm": 0.8247941732406616, "learning_rate": 2.0336684762531972e-05, "loss": 0.4501, "num_input_tokens_seen": 30338480, "step": 22390 }, { "epoch": 0.7168235068177453, "grad_norm": 0.6794570684432983, "learning_rate": 2.032703995410631e-05, "loss": 0.3769, "num_input_tokens_seen": 30345424, "step": 22395 }, { "epoch": 0.7169835477882338, "grad_norm": 2.6238913536071777, "learning_rate": 2.031739586630981e-05, "loss": 0.5805, "num_input_tokens_seen": 30352256, "step": 22400 }, { "epoch": 0.7169835477882338, "eval_loss": 0.4556000232696533, "eval_runtime": 502.5869, "eval_samples_per_second": 27.629, "eval_steps_per_second": 13.815, "num_input_tokens_seen": 30352256, "step": 22400 }, { "epoch": 0.7171435887587222, "grad_norm": 1.0908492803573608, "learning_rate": 2.0307752500629707e-05, "loss": 0.5615, "num_input_tokens_seen": 30359232, "step": 22405 }, { "epoch": 0.7173036297292107, "grad_norm": 0.8965224623680115, "learning_rate": 2.0298109858553144e-05, "loss": 0.4225, "num_input_tokens_seen": 30366128, "step": 22410 }, { "epoch": 0.7174636706996991, "grad_norm": 1.0642634630203247, "learning_rate": 2.028846794156712e-05, "loss": 0.5369, "num_input_tokens_seen": 30373488, "step": 22415 }, { "epoch": 0.7176237116701876, "grad_norm": 1.019187331199646, "learning_rate": 2.027882675115856e-05, "loss": 0.3142, "num_input_tokens_seen": 30380560, "step": 22420 }, { "epoch": 0.7177837526406761, "grad_norm": 1.5257593393325806, "learning_rate": 2.026918628881423e-05, "loss": 0.5218, "num_input_tokens_seen": 30387504, "step": 22425 }, { "epoch": 0.7179437936111644, "grad_norm": 0.5841609239578247, "learning_rate": 2.0259546556020833e-05, "loss": 0.294, "num_input_tokens_seen": 30394016, "step": 22430 }, { "epoch": 0.7181038345816529, "grad_norm": 1.0577337741851807, "learning_rate": 2.024990755426493e-05, "loss": 0.3288, "num_input_tokens_seen": 30400912, "step": 22435 }, { "epoch": 0.7182638755521413, "grad_norm": 0.7847697734832764, "learning_rate": 2.0240269285032975e-05, "loss": 0.4341, "num_input_tokens_seen": 30407552, "step": 22440 }, { "epoch": 0.7184239165226298, "grad_norm": 1.285758137702942, "learning_rate": 2.0230631749811306e-05, "loss": 0.3831, "num_input_tokens_seen": 30414528, "step": 22445 }, { "epoch": 0.7185839574931182, "grad_norm": 2.479592800140381, "learning_rate": 2.0220994950086162e-05, "loss": 0.5392, "num_input_tokens_seen": 30421264, "step": 22450 }, { "epoch": 0.7187439984636067, "grad_norm": 1.2658581733703613, "learning_rate": 2.021135888734365e-05, "loss": 0.5252, "num_input_tokens_seen": 30428160, "step": 22455 }, { "epoch": 0.7189040394340951, "grad_norm": 2.556316614151001, "learning_rate": 2.0201723563069783e-05, "loss": 0.4871, "num_input_tokens_seen": 30434784, "step": 22460 }, { "epoch": 0.7190640804045836, "grad_norm": 1.3655381202697754, "learning_rate": 2.0192088978750433e-05, "loss": 0.4513, "num_input_tokens_seen": 30441184, "step": 22465 }, { "epoch": 0.719224121375072, "grad_norm": 1.1636736392974854, "learning_rate": 2.0182455135871385e-05, "loss": 0.4194, "num_input_tokens_seen": 30447760, "step": 22470 }, { "epoch": 0.7193841623455605, "grad_norm": 3.1340160369873047, "learning_rate": 2.0172822035918305e-05, "loss": 0.3682, "num_input_tokens_seen": 30454544, "step": 22475 }, { "epoch": 0.719544203316049, "grad_norm": 1.3514162302017212, "learning_rate": 2.016318968037671e-05, "loss": 0.3585, "num_input_tokens_seen": 30461136, "step": 22480 }, { "epoch": 0.7197042442865373, "grad_norm": 1.6323970556259155, "learning_rate": 2.015355807073206e-05, "loss": 0.5229, "num_input_tokens_seen": 30468096, "step": 22485 }, { "epoch": 0.7198642852570258, "grad_norm": 1.2552788257598877, "learning_rate": 2.0143927208469664e-05, "loss": 0.312, "num_input_tokens_seen": 30474512, "step": 22490 }, { "epoch": 0.7200243262275142, "grad_norm": 1.7172200679779053, "learning_rate": 2.0134297095074708e-05, "loss": 0.6434, "num_input_tokens_seen": 30481376, "step": 22495 }, { "epoch": 0.7201843671980027, "grad_norm": 0.8497475981712341, "learning_rate": 2.0124667732032297e-05, "loss": 0.749, "num_input_tokens_seen": 30488368, "step": 22500 }, { "epoch": 0.7203444081684911, "grad_norm": 0.6441019773483276, "learning_rate": 2.011503912082738e-05, "loss": 0.4561, "num_input_tokens_seen": 30494960, "step": 22505 }, { "epoch": 0.7205044491389796, "grad_norm": 0.6979666352272034, "learning_rate": 2.0105411262944823e-05, "loss": 0.3429, "num_input_tokens_seen": 30501648, "step": 22510 }, { "epoch": 0.720664490109468, "grad_norm": 0.7955120801925659, "learning_rate": 2.0095784159869366e-05, "loss": 0.3019, "num_input_tokens_seen": 30508064, "step": 22515 }, { "epoch": 0.7208245310799565, "grad_norm": 1.1841650009155273, "learning_rate": 2.0086157813085608e-05, "loss": 0.3487, "num_input_tokens_seen": 30514880, "step": 22520 }, { "epoch": 0.7209845720504449, "grad_norm": 0.40560823678970337, "learning_rate": 2.0076532224078068e-05, "loss": 0.3021, "num_input_tokens_seen": 30521520, "step": 22525 }, { "epoch": 0.7211446130209334, "grad_norm": 0.6920286417007446, "learning_rate": 2.0066907394331142e-05, "loss": 0.4447, "num_input_tokens_seen": 30528272, "step": 22530 }, { "epoch": 0.7213046539914219, "grad_norm": 0.7283238768577576, "learning_rate": 2.0057283325329077e-05, "loss": 0.4776, "num_input_tokens_seen": 30534768, "step": 22535 }, { "epoch": 0.7214646949619102, "grad_norm": 0.7139344215393066, "learning_rate": 2.0047660018556047e-05, "loss": 0.3378, "num_input_tokens_seen": 30541152, "step": 22540 }, { "epoch": 0.7216247359323987, "grad_norm": 1.7861437797546387, "learning_rate": 2.0038037475496075e-05, "loss": 0.6109, "num_input_tokens_seen": 30547936, "step": 22545 }, { "epoch": 0.7217847769028871, "grad_norm": 0.9268988966941833, "learning_rate": 2.0028415697633073e-05, "loss": 0.3152, "num_input_tokens_seen": 30554208, "step": 22550 }, { "epoch": 0.7219448178733756, "grad_norm": 0.9078839421272278, "learning_rate": 2.0018794686450858e-05, "loss": 0.4157, "num_input_tokens_seen": 30560480, "step": 22555 }, { "epoch": 0.722104858843864, "grad_norm": 1.247725248336792, "learning_rate": 2.0009174443433088e-05, "loss": 0.3878, "num_input_tokens_seen": 30567232, "step": 22560 }, { "epoch": 0.7222648998143525, "grad_norm": 0.8305630087852478, "learning_rate": 1.999955497006334e-05, "loss": 0.3975, "num_input_tokens_seen": 30574144, "step": 22565 }, { "epoch": 0.7224249407848409, "grad_norm": 1.7479262351989746, "learning_rate": 1.9989936267825067e-05, "loss": 0.4899, "num_input_tokens_seen": 30580896, "step": 22570 }, { "epoch": 0.7225849817553294, "grad_norm": 1.466342806816101, "learning_rate": 1.9980318338201572e-05, "loss": 0.5126, "num_input_tokens_seen": 30588096, "step": 22575 }, { "epoch": 0.7227450227258178, "grad_norm": 1.1129177808761597, "learning_rate": 1.997070118267607e-05, "loss": 0.5168, "num_input_tokens_seen": 30594800, "step": 22580 }, { "epoch": 0.7229050636963062, "grad_norm": 1.5406255722045898, "learning_rate": 1.9961084802731654e-05, "loss": 0.4662, "num_input_tokens_seen": 30601632, "step": 22585 }, { "epoch": 0.7230651046667947, "grad_norm": 0.5147783160209656, "learning_rate": 1.9951469199851273e-05, "loss": 0.4161, "num_input_tokens_seen": 30608096, "step": 22590 }, { "epoch": 0.7232251456372831, "grad_norm": 1.681506872177124, "learning_rate": 1.99418543755178e-05, "loss": 0.464, "num_input_tokens_seen": 30615088, "step": 22595 }, { "epoch": 0.7233851866077716, "grad_norm": 0.9727017283439636, "learning_rate": 1.9932240331213936e-05, "loss": 0.4824, "num_input_tokens_seen": 30622032, "step": 22600 }, { "epoch": 0.7233851866077716, "eval_loss": 0.45522257685661316, "eval_runtime": 502.4457, "eval_samples_per_second": 27.637, "eval_steps_per_second": 13.818, "num_input_tokens_seen": 30622032, "step": 22600 }, { "epoch": 0.72354522757826, "grad_norm": 1.3577693700790405, "learning_rate": 1.9922627068422297e-05, "loss": 0.3879, "num_input_tokens_seen": 30628816, "step": 22605 }, { "epoch": 0.7237052685487485, "grad_norm": 1.044047474861145, "learning_rate": 1.991301458862538e-05, "loss": 0.4228, "num_input_tokens_seen": 30635984, "step": 22610 }, { "epoch": 0.7238653095192369, "grad_norm": 1.1776493787765503, "learning_rate": 1.9903402893305536e-05, "loss": 0.5416, "num_input_tokens_seen": 30642512, "step": 22615 }, { "epoch": 0.7240253504897254, "grad_norm": 0.8102881908416748, "learning_rate": 1.9893791983945016e-05, "loss": 0.5352, "num_input_tokens_seen": 30649408, "step": 22620 }, { "epoch": 0.7241853914602138, "grad_norm": 2.2643628120422363, "learning_rate": 1.988418186202594e-05, "loss": 0.5639, "num_input_tokens_seen": 30656096, "step": 22625 }, { "epoch": 0.7243454324307023, "grad_norm": 1.7027387619018555, "learning_rate": 1.98745725290303e-05, "loss": 0.548, "num_input_tokens_seen": 30662832, "step": 22630 }, { "epoch": 0.7245054734011908, "grad_norm": 1.0369569063186646, "learning_rate": 1.986496398644e-05, "loss": 0.421, "num_input_tokens_seen": 30669744, "step": 22635 }, { "epoch": 0.7246655143716791, "grad_norm": 1.1164624691009521, "learning_rate": 1.9855356235736777e-05, "loss": 0.5109, "num_input_tokens_seen": 30676752, "step": 22640 }, { "epoch": 0.7248255553421676, "grad_norm": 1.013738751411438, "learning_rate": 1.9845749278402277e-05, "loss": 0.4444, "num_input_tokens_seen": 30683136, "step": 22645 }, { "epoch": 0.724985596312656, "grad_norm": 0.582649290561676, "learning_rate": 1.9836143115918006e-05, "loss": 0.4408, "num_input_tokens_seen": 30690112, "step": 22650 }, { "epoch": 0.7251456372831445, "grad_norm": 0.9423133134841919, "learning_rate": 1.9826537749765367e-05, "loss": 0.6525, "num_input_tokens_seen": 30696896, "step": 22655 }, { "epoch": 0.7253056782536329, "grad_norm": 1.2231156826019287, "learning_rate": 1.9816933181425625e-05, "loss": 0.4382, "num_input_tokens_seen": 30703488, "step": 22660 }, { "epoch": 0.7254657192241214, "grad_norm": 0.5015915632247925, "learning_rate": 1.9807329412379903e-05, "loss": 0.5443, "num_input_tokens_seen": 30710144, "step": 22665 }, { "epoch": 0.7256257601946098, "grad_norm": 1.1551538705825806, "learning_rate": 1.9797726444109247e-05, "loss": 0.3969, "num_input_tokens_seen": 30716832, "step": 22670 }, { "epoch": 0.7257858011650983, "grad_norm": 1.2470252513885498, "learning_rate": 1.9788124278094557e-05, "loss": 0.4493, "num_input_tokens_seen": 30723488, "step": 22675 }, { "epoch": 0.7259458421355867, "grad_norm": 0.7177825570106506, "learning_rate": 1.9778522915816594e-05, "loss": 0.3411, "num_input_tokens_seen": 30730160, "step": 22680 }, { "epoch": 0.7261058831060752, "grad_norm": 1.7886077165603638, "learning_rate": 1.9768922358756014e-05, "loss": 0.4833, "num_input_tokens_seen": 30736672, "step": 22685 }, { "epoch": 0.7262659240765637, "grad_norm": 1.0217829942703247, "learning_rate": 1.9759322608393353e-05, "loss": 0.4838, "num_input_tokens_seen": 30743504, "step": 22690 }, { "epoch": 0.726425965047052, "grad_norm": 0.7662832736968994, "learning_rate": 1.9749723666208992e-05, "loss": 0.5029, "num_input_tokens_seen": 30750256, "step": 22695 }, { "epoch": 0.7265860060175405, "grad_norm": 2.948239803314209, "learning_rate": 1.9740125533683235e-05, "loss": 0.4533, "num_input_tokens_seen": 30757056, "step": 22700 }, { "epoch": 0.7267460469880289, "grad_norm": 0.9676111340522766, "learning_rate": 1.9730528212296208e-05, "loss": 0.4759, "num_input_tokens_seen": 30763744, "step": 22705 }, { "epoch": 0.7269060879585174, "grad_norm": 1.3450411558151245, "learning_rate": 1.9720931703527945e-05, "loss": 0.4393, "num_input_tokens_seen": 30770432, "step": 22710 }, { "epoch": 0.7270661289290058, "grad_norm": 0.8603992462158203, "learning_rate": 1.9711336008858373e-05, "loss": 0.493, "num_input_tokens_seen": 30777328, "step": 22715 }, { "epoch": 0.7272261698994943, "grad_norm": 1.848406434059143, "learning_rate": 1.9701741129767233e-05, "loss": 0.6424, "num_input_tokens_seen": 30784048, "step": 22720 }, { "epoch": 0.7273862108699827, "grad_norm": 0.7987347841262817, "learning_rate": 1.9692147067734202e-05, "loss": 0.3623, "num_input_tokens_seen": 30790704, "step": 22725 }, { "epoch": 0.7275462518404712, "grad_norm": 0.9247506856918335, "learning_rate": 1.96825538242388e-05, "loss": 0.3142, "num_input_tokens_seen": 30798016, "step": 22730 }, { "epoch": 0.7277062928109596, "grad_norm": 0.9469716548919678, "learning_rate": 1.967296140076041e-05, "loss": 0.3967, "num_input_tokens_seen": 30805088, "step": 22735 }, { "epoch": 0.727866333781448, "grad_norm": 1.7942882776260376, "learning_rate": 1.966336979877833e-05, "loss": 0.5685, "num_input_tokens_seen": 30811744, "step": 22740 }, { "epoch": 0.7280263747519365, "grad_norm": 2.618889570236206, "learning_rate": 1.9653779019771678e-05, "loss": 0.3278, "num_input_tokens_seen": 30818496, "step": 22745 }, { "epoch": 0.7281864157224249, "grad_norm": 1.2838298082351685, "learning_rate": 1.9644189065219488e-05, "loss": 0.5124, "num_input_tokens_seen": 30825408, "step": 22750 }, { "epoch": 0.7283464566929134, "grad_norm": 1.8478446006774902, "learning_rate": 1.9634599936600655e-05, "loss": 0.529, "num_input_tokens_seen": 30831920, "step": 22755 }, { "epoch": 0.7285064976634018, "grad_norm": 1.2058359384536743, "learning_rate": 1.9625011635393935e-05, "loss": 0.4291, "num_input_tokens_seen": 30838896, "step": 22760 }, { "epoch": 0.7286665386338903, "grad_norm": 0.6871253252029419, "learning_rate": 1.9615424163077963e-05, "loss": 0.4101, "num_input_tokens_seen": 30845392, "step": 22765 }, { "epoch": 0.7288265796043787, "grad_norm": 1.53274667263031, "learning_rate": 1.9605837521131263e-05, "loss": 0.3979, "num_input_tokens_seen": 30852528, "step": 22770 }, { "epoch": 0.7289866205748672, "grad_norm": 1.9656320810317993, "learning_rate": 1.9596251711032192e-05, "loss": 0.5502, "num_input_tokens_seen": 30859232, "step": 22775 }, { "epoch": 0.7291466615453556, "grad_norm": 1.0744328498840332, "learning_rate": 1.958666673425903e-05, "loss": 0.4432, "num_input_tokens_seen": 30866592, "step": 22780 }, { "epoch": 0.7293067025158441, "grad_norm": 0.7345167398452759, "learning_rate": 1.957708259228987e-05, "loss": 0.4244, "num_input_tokens_seen": 30873152, "step": 22785 }, { "epoch": 0.7294667434863324, "grad_norm": 1.1824913024902344, "learning_rate": 1.956749928660273e-05, "loss": 0.4705, "num_input_tokens_seen": 30880384, "step": 22790 }, { "epoch": 0.7296267844568209, "grad_norm": 0.5317119359970093, "learning_rate": 1.955791681867547e-05, "loss": 0.2681, "num_input_tokens_seen": 30887168, "step": 22795 }, { "epoch": 0.7297868254273094, "grad_norm": 0.9119118452072144, "learning_rate": 1.9548335189985824e-05, "loss": 0.3745, "num_input_tokens_seen": 30894016, "step": 22800 }, { "epoch": 0.7297868254273094, "eval_loss": 0.45485708117485046, "eval_runtime": 502.4473, "eval_samples_per_second": 27.637, "eval_steps_per_second": 13.818, "num_input_tokens_seen": 30894016, "step": 22800 }, { "epoch": 0.7299468663977978, "grad_norm": 1.111621379852295, "learning_rate": 1.9538754402011396e-05, "loss": 0.519, "num_input_tokens_seen": 30900768, "step": 22805 }, { "epoch": 0.7301069073682863, "grad_norm": 1.9493647813796997, "learning_rate": 1.952917445622968e-05, "loss": 0.4317, "num_input_tokens_seen": 30907744, "step": 22810 }, { "epoch": 0.7302669483387747, "grad_norm": 0.8506498336791992, "learning_rate": 1.9519595354118005e-05, "loss": 0.3828, "num_input_tokens_seen": 30913904, "step": 22815 }, { "epoch": 0.7304269893092632, "grad_norm": 0.6675489544868469, "learning_rate": 1.951001709715361e-05, "loss": 0.3863, "num_input_tokens_seen": 30920256, "step": 22820 }, { "epoch": 0.7305870302797516, "grad_norm": 0.5247955918312073, "learning_rate": 1.9500439686813556e-05, "loss": 0.3518, "num_input_tokens_seen": 30926688, "step": 22825 }, { "epoch": 0.7307470712502401, "grad_norm": 0.7715054154396057, "learning_rate": 1.949086312457482e-05, "loss": 0.4452, "num_input_tokens_seen": 30933392, "step": 22830 }, { "epoch": 0.7309071122207285, "grad_norm": 1.2842952013015747, "learning_rate": 1.9481287411914223e-05, "loss": 0.4433, "num_input_tokens_seen": 30940032, "step": 22835 }, { "epoch": 0.731067153191217, "grad_norm": 1.378877878189087, "learning_rate": 1.9471712550308457e-05, "loss": 0.601, "num_input_tokens_seen": 30947008, "step": 22840 }, { "epoch": 0.7312271941617055, "grad_norm": 0.8747061491012573, "learning_rate": 1.946213854123409e-05, "loss": 0.3344, "num_input_tokens_seen": 30953968, "step": 22845 }, { "epoch": 0.7313872351321938, "grad_norm": 1.2039531469345093, "learning_rate": 1.9452565386167554e-05, "loss": 0.476, "num_input_tokens_seen": 30960656, "step": 22850 }, { "epoch": 0.7315472761026823, "grad_norm": 0.7641540169715881, "learning_rate": 1.9442993086585142e-05, "loss": 0.4001, "num_input_tokens_seen": 30967168, "step": 22855 }, { "epoch": 0.7317073170731707, "grad_norm": 1.258208155632019, "learning_rate": 1.9433421643963043e-05, "loss": 0.3218, "num_input_tokens_seen": 30974128, "step": 22860 }, { "epoch": 0.7318673580436592, "grad_norm": 1.4244115352630615, "learning_rate": 1.942385105977727e-05, "loss": 0.375, "num_input_tokens_seen": 30980656, "step": 22865 }, { "epoch": 0.7320273990141476, "grad_norm": 1.0311172008514404, "learning_rate": 1.9414281335503743e-05, "loss": 0.4228, "num_input_tokens_seen": 30987360, "step": 22870 }, { "epoch": 0.7321874399846361, "grad_norm": 1.0923854112625122, "learning_rate": 1.9404712472618232e-05, "loss": 0.3898, "num_input_tokens_seen": 30994224, "step": 22875 }, { "epoch": 0.7323474809551245, "grad_norm": 1.6335012912750244, "learning_rate": 1.939514447259636e-05, "loss": 0.4809, "num_input_tokens_seen": 31001024, "step": 22880 }, { "epoch": 0.732507521925613, "grad_norm": 1.3943500518798828, "learning_rate": 1.938557733691365e-05, "loss": 0.5514, "num_input_tokens_seen": 31007552, "step": 22885 }, { "epoch": 0.7326675628961014, "grad_norm": 2.7723562717437744, "learning_rate": 1.9376011067045476e-05, "loss": 0.475, "num_input_tokens_seen": 31014304, "step": 22890 }, { "epoch": 0.7328276038665898, "grad_norm": 0.8531527519226074, "learning_rate": 1.9366445664467065e-05, "loss": 0.4789, "num_input_tokens_seen": 31021104, "step": 22895 }, { "epoch": 0.7329876448370783, "grad_norm": 1.6238770484924316, "learning_rate": 1.9356881130653533e-05, "loss": 0.5275, "num_input_tokens_seen": 31028032, "step": 22900 }, { "epoch": 0.7331476858075667, "grad_norm": 2.574993133544922, "learning_rate": 1.9347317467079846e-05, "loss": 0.4538, "num_input_tokens_seen": 31034496, "step": 22905 }, { "epoch": 0.7333077267780552, "grad_norm": 1.356863021850586, "learning_rate": 1.9337754675220836e-05, "loss": 0.3596, "num_input_tokens_seen": 31041152, "step": 22910 }, { "epoch": 0.7334677677485436, "grad_norm": 0.7431120276451111, "learning_rate": 1.9328192756551218e-05, "loss": 0.3554, "num_input_tokens_seen": 31047760, "step": 22915 }, { "epoch": 0.7336278087190321, "grad_norm": 1.5123956203460693, "learning_rate": 1.931863171254555e-05, "loss": 0.4846, "num_input_tokens_seen": 31054768, "step": 22920 }, { "epoch": 0.7337878496895205, "grad_norm": 1.8042949438095093, "learning_rate": 1.930907154467826e-05, "loss": 0.4162, "num_input_tokens_seen": 31061344, "step": 22925 }, { "epoch": 0.733947890660009, "grad_norm": 1.555599331855774, "learning_rate": 1.9299512254423673e-05, "loss": 0.661, "num_input_tokens_seen": 31068080, "step": 22930 }, { "epoch": 0.7341079316304974, "grad_norm": 0.6235243082046509, "learning_rate": 1.9289953843255914e-05, "loss": 0.3164, "num_input_tokens_seen": 31074880, "step": 22935 }, { "epoch": 0.7342679726009859, "grad_norm": 5.71909761428833, "learning_rate": 1.9280396312649048e-05, "loss": 0.4351, "num_input_tokens_seen": 31081792, "step": 22940 }, { "epoch": 0.7344280135714742, "grad_norm": 0.5520308017730713, "learning_rate": 1.9270839664076936e-05, "loss": 0.2563, "num_input_tokens_seen": 31088544, "step": 22945 }, { "epoch": 0.7345880545419627, "grad_norm": 1.202600121498108, "learning_rate": 1.9261283899013345e-05, "loss": 0.4792, "num_input_tokens_seen": 31095408, "step": 22950 }, { "epoch": 0.7347480955124512, "grad_norm": 0.7364420890808105, "learning_rate": 1.92517290189319e-05, "loss": 0.4629, "num_input_tokens_seen": 31102272, "step": 22955 }, { "epoch": 0.7349081364829396, "grad_norm": 1.5127619504928589, "learning_rate": 1.924217502530607e-05, "loss": 0.5923, "num_input_tokens_seen": 31108640, "step": 22960 }, { "epoch": 0.7350681774534281, "grad_norm": 1.3547604084014893, "learning_rate": 1.9232621919609207e-05, "loss": 0.4781, "num_input_tokens_seen": 31115168, "step": 22965 }, { "epoch": 0.7352282184239165, "grad_norm": 1.2195615768432617, "learning_rate": 1.9223069703314534e-05, "loss": 0.5156, "num_input_tokens_seen": 31121744, "step": 22970 }, { "epoch": 0.735388259394405, "grad_norm": 1.6630210876464844, "learning_rate": 1.92135183778951e-05, "loss": 0.4001, "num_input_tokens_seen": 31128560, "step": 22975 }, { "epoch": 0.7355483003648934, "grad_norm": 0.8096179366111755, "learning_rate": 1.9203967944823857e-05, "loss": 0.5094, "num_input_tokens_seen": 31135408, "step": 22980 }, { "epoch": 0.7357083413353819, "grad_norm": 1.4212677478790283, "learning_rate": 1.9194418405573588e-05, "loss": 0.4278, "num_input_tokens_seen": 31142112, "step": 22985 }, { "epoch": 0.7358683823058703, "grad_norm": 0.6473661065101624, "learning_rate": 1.9184869761616954e-05, "loss": 0.3375, "num_input_tokens_seen": 31149232, "step": 22990 }, { "epoch": 0.7360284232763588, "grad_norm": 0.6513486504554749, "learning_rate": 1.9175322014426495e-05, "loss": 0.3635, "num_input_tokens_seen": 31156368, "step": 22995 }, { "epoch": 0.7361884642468471, "grad_norm": 1.703829288482666, "learning_rate": 1.9165775165474565e-05, "loss": 0.5018, "num_input_tokens_seen": 31162736, "step": 23000 }, { "epoch": 0.7361884642468471, "eval_loss": 0.4544597864151001, "eval_runtime": 502.5894, "eval_samples_per_second": 27.629, "eval_steps_per_second": 13.814, "num_input_tokens_seen": 31162736, "step": 23000 }, { "epoch": 0.7363485052173356, "grad_norm": 1.4073764085769653, "learning_rate": 1.9156229216233434e-05, "loss": 0.4597, "num_input_tokens_seen": 31169120, "step": 23005 }, { "epoch": 0.7365085461878241, "grad_norm": 1.1401244401931763, "learning_rate": 1.9146684168175184e-05, "loss": 0.44, "num_input_tokens_seen": 31175760, "step": 23010 }, { "epoch": 0.7366685871583125, "grad_norm": 0.7074934840202332, "learning_rate": 1.9137140022771796e-05, "loss": 0.5074, "num_input_tokens_seen": 31182720, "step": 23015 }, { "epoch": 0.736828628128801, "grad_norm": 1.1211318969726562, "learning_rate": 1.9127596781495103e-05, "loss": 0.5222, "num_input_tokens_seen": 31189808, "step": 23020 }, { "epoch": 0.7369886690992894, "grad_norm": 1.3536326885223389, "learning_rate": 1.9118054445816767e-05, "loss": 0.4415, "num_input_tokens_seen": 31197008, "step": 23025 }, { "epoch": 0.7371487100697779, "grad_norm": 0.66172194480896, "learning_rate": 1.9108513017208356e-05, "loss": 0.3915, "num_input_tokens_seen": 31203552, "step": 23030 }, { "epoch": 0.7373087510402663, "grad_norm": 2.2516138553619385, "learning_rate": 1.9098972497141287e-05, "loss": 0.4648, "num_input_tokens_seen": 31211408, "step": 23035 }, { "epoch": 0.7374687920107548, "grad_norm": 1.0167611837387085, "learning_rate": 1.9089432887086806e-05, "loss": 0.4095, "num_input_tokens_seen": 31218016, "step": 23040 }, { "epoch": 0.7376288329812432, "grad_norm": 1.3325809240341187, "learning_rate": 1.9079894188516056e-05, "loss": 0.4235, "num_input_tokens_seen": 31225120, "step": 23045 }, { "epoch": 0.7377888739517316, "grad_norm": 0.9894768595695496, "learning_rate": 1.907035640290002e-05, "loss": 0.4411, "num_input_tokens_seen": 31231744, "step": 23050 }, { "epoch": 0.73794891492222, "grad_norm": 0.9603154063224792, "learning_rate": 1.9060819531709534e-05, "loss": 0.6553, "num_input_tokens_seen": 31238176, "step": 23055 }, { "epoch": 0.7381089558927085, "grad_norm": 1.124348759651184, "learning_rate": 1.9051283576415325e-05, "loss": 0.39, "num_input_tokens_seen": 31244720, "step": 23060 }, { "epoch": 0.738268996863197, "grad_norm": 2.408407211303711, "learning_rate": 1.904174853848793e-05, "loss": 0.5597, "num_input_tokens_seen": 31251568, "step": 23065 }, { "epoch": 0.7384290378336854, "grad_norm": 0.7452348470687866, "learning_rate": 1.903221441939779e-05, "loss": 0.315, "num_input_tokens_seen": 31258096, "step": 23070 }, { "epoch": 0.7385890788041739, "grad_norm": 0.9460397958755493, "learning_rate": 1.9022681220615194e-05, "loss": 0.2624, "num_input_tokens_seen": 31264992, "step": 23075 }, { "epoch": 0.7387491197746623, "grad_norm": 0.9274387359619141, "learning_rate": 1.9013148943610255e-05, "loss": 0.4204, "num_input_tokens_seen": 31271616, "step": 23080 }, { "epoch": 0.7389091607451508, "grad_norm": 0.9782979488372803, "learning_rate": 1.9003617589852998e-05, "loss": 0.4344, "num_input_tokens_seen": 31278272, "step": 23085 }, { "epoch": 0.7390692017156392, "grad_norm": 1.3536471128463745, "learning_rate": 1.899408716081326e-05, "loss": 0.6773, "num_input_tokens_seen": 31285040, "step": 23090 }, { "epoch": 0.7392292426861277, "grad_norm": 0.5304598212242126, "learning_rate": 1.898455765796075e-05, "loss": 0.3551, "num_input_tokens_seen": 31292160, "step": 23095 }, { "epoch": 0.739389283656616, "grad_norm": 0.6278796195983887, "learning_rate": 1.8975029082765053e-05, "loss": 0.581, "num_input_tokens_seen": 31298384, "step": 23100 }, { "epoch": 0.7395493246271045, "grad_norm": 1.1274698972702026, "learning_rate": 1.8965501436695577e-05, "loss": 0.5124, "num_input_tokens_seen": 31304944, "step": 23105 }, { "epoch": 0.739709365597593, "grad_norm": 0.3840019702911377, "learning_rate": 1.895597472122161e-05, "loss": 0.27, "num_input_tokens_seen": 31311520, "step": 23110 }, { "epoch": 0.7398694065680814, "grad_norm": 1.3979649543762207, "learning_rate": 1.894644893781231e-05, "loss": 0.7375, "num_input_tokens_seen": 31318128, "step": 23115 }, { "epoch": 0.7400294475385699, "grad_norm": 0.9921591877937317, "learning_rate": 1.893692408793665e-05, "loss": 0.514, "num_input_tokens_seen": 31324544, "step": 23120 }, { "epoch": 0.7401894885090583, "grad_norm": 1.3536372184753418, "learning_rate": 1.8927400173063493e-05, "loss": 0.3143, "num_input_tokens_seen": 31331328, "step": 23125 }, { "epoch": 0.7403495294795468, "grad_norm": 0.7523252964019775, "learning_rate": 1.891787719466154e-05, "loss": 0.3859, "num_input_tokens_seen": 31337936, "step": 23130 }, { "epoch": 0.7405095704500352, "grad_norm": 1.1278263330459595, "learning_rate": 1.8908355154199346e-05, "loss": 0.4664, "num_input_tokens_seen": 31344736, "step": 23135 }, { "epoch": 0.7406696114205237, "grad_norm": 1.160677194595337, "learning_rate": 1.8898834053145357e-05, "loss": 0.6576, "num_input_tokens_seen": 31351712, "step": 23140 }, { "epoch": 0.7408296523910121, "grad_norm": 0.8018305897712708, "learning_rate": 1.8889313892967813e-05, "loss": 0.3633, "num_input_tokens_seen": 31358528, "step": 23145 }, { "epoch": 0.7409896933615006, "grad_norm": 1.1917294263839722, "learning_rate": 1.8879794675134863e-05, "loss": 0.2762, "num_input_tokens_seen": 31365456, "step": 23150 }, { "epoch": 0.7411497343319889, "grad_norm": 1.4879956245422363, "learning_rate": 1.8870276401114494e-05, "loss": 0.4798, "num_input_tokens_seen": 31372080, "step": 23155 }, { "epoch": 0.7413097753024774, "grad_norm": 0.8045569062232971, "learning_rate": 1.886075907237453e-05, "loss": 0.3991, "num_input_tokens_seen": 31378416, "step": 23160 }, { "epoch": 0.7414698162729659, "grad_norm": 1.2325587272644043, "learning_rate": 1.8851242690382672e-05, "loss": 0.4263, "num_input_tokens_seen": 31385168, "step": 23165 }, { "epoch": 0.7416298572434543, "grad_norm": 1.4541890621185303, "learning_rate": 1.884172725660645e-05, "loss": 0.4796, "num_input_tokens_seen": 31392144, "step": 23170 }, { "epoch": 0.7417898982139428, "grad_norm": 1.4589476585388184, "learning_rate": 1.8832212772513277e-05, "loss": 0.3179, "num_input_tokens_seen": 31398816, "step": 23175 }, { "epoch": 0.7419499391844312, "grad_norm": 1.2576509714126587, "learning_rate": 1.8822699239570414e-05, "loss": 0.4025, "num_input_tokens_seen": 31405840, "step": 23180 }, { "epoch": 0.7421099801549197, "grad_norm": 1.2103338241577148, "learning_rate": 1.8813186659244943e-05, "loss": 0.4387, "num_input_tokens_seen": 31412720, "step": 23185 }, { "epoch": 0.7422700211254081, "grad_norm": 1.0530376434326172, "learning_rate": 1.880367503300385e-05, "loss": 0.3698, "num_input_tokens_seen": 31419440, "step": 23190 }, { "epoch": 0.7424300620958966, "grad_norm": 0.7831695675849915, "learning_rate": 1.8794164362313927e-05, "loss": 0.385, "num_input_tokens_seen": 31426560, "step": 23195 }, { "epoch": 0.742590103066385, "grad_norm": 0.9720648527145386, "learning_rate": 1.878465464864185e-05, "loss": 0.4904, "num_input_tokens_seen": 31433344, "step": 23200 }, { "epoch": 0.742590103066385, "eval_loss": 0.45405229926109314, "eval_runtime": 503.382, "eval_samples_per_second": 27.585, "eval_steps_per_second": 13.793, "num_input_tokens_seen": 31433344, "step": 23200 }, { "epoch": 0.7427501440368734, "grad_norm": 0.5792034864425659, "learning_rate": 1.877514589345414e-05, "loss": 0.5416, "num_input_tokens_seen": 31440400, "step": 23205 }, { "epoch": 0.7429101850073618, "grad_norm": 1.2271339893341064, "learning_rate": 1.876563809821715e-05, "loss": 0.4801, "num_input_tokens_seen": 31447136, "step": 23210 }, { "epoch": 0.7430702259778503, "grad_norm": 1.5053329467773438, "learning_rate": 1.8756131264397106e-05, "loss": 0.4152, "num_input_tokens_seen": 31453904, "step": 23215 }, { "epoch": 0.7432302669483388, "grad_norm": 1.171112060546875, "learning_rate": 1.87466253934601e-05, "loss": 0.4512, "num_input_tokens_seen": 31460976, "step": 23220 }, { "epoch": 0.7433903079188272, "grad_norm": 0.735288143157959, "learning_rate": 1.8737120486872033e-05, "loss": 0.442, "num_input_tokens_seen": 31468000, "step": 23225 }, { "epoch": 0.7435503488893157, "grad_norm": 0.822211503982544, "learning_rate": 1.8727616546098696e-05, "loss": 0.5153, "num_input_tokens_seen": 31474576, "step": 23230 }, { "epoch": 0.7437103898598041, "grad_norm": 1.167911171913147, "learning_rate": 1.8718113572605716e-05, "loss": 0.5828, "num_input_tokens_seen": 31481456, "step": 23235 }, { "epoch": 0.7438704308302926, "grad_norm": 1.1750048398971558, "learning_rate": 1.8708611567858554e-05, "loss": 0.3826, "num_input_tokens_seen": 31488208, "step": 23240 }, { "epoch": 0.744030471800781, "grad_norm": 1.5413395166397095, "learning_rate": 1.8699110533322565e-05, "loss": 0.5304, "num_input_tokens_seen": 31495712, "step": 23245 }, { "epoch": 0.7441905127712695, "grad_norm": 1.0608340501785278, "learning_rate": 1.8689610470462897e-05, "loss": 0.4866, "num_input_tokens_seen": 31502576, "step": 23250 }, { "epoch": 0.7443505537417578, "grad_norm": 1.2556896209716797, "learning_rate": 1.8680111380744604e-05, "loss": 0.3185, "num_input_tokens_seen": 31509216, "step": 23255 }, { "epoch": 0.7445105947122463, "grad_norm": 1.126851201057434, "learning_rate": 1.8670613265632564e-05, "loss": 0.4662, "num_input_tokens_seen": 31515424, "step": 23260 }, { "epoch": 0.7446706356827347, "grad_norm": 1.0156989097595215, "learning_rate": 1.866111612659149e-05, "loss": 0.3935, "num_input_tokens_seen": 31522016, "step": 23265 }, { "epoch": 0.7448306766532232, "grad_norm": 1.0945996046066284, "learning_rate": 1.8651619965085967e-05, "loss": 0.3695, "num_input_tokens_seen": 31528848, "step": 23270 }, { "epoch": 0.7449907176237117, "grad_norm": 1.4260053634643555, "learning_rate": 1.8642124782580433e-05, "loss": 0.3842, "num_input_tokens_seen": 31535936, "step": 23275 }, { "epoch": 0.7451507585942001, "grad_norm": 0.8352589011192322, "learning_rate": 1.8632630580539144e-05, "loss": 0.305, "num_input_tokens_seen": 31543008, "step": 23280 }, { "epoch": 0.7453107995646886, "grad_norm": 0.46649911999702454, "learning_rate": 1.862313736042625e-05, "loss": 0.4836, "num_input_tokens_seen": 31550208, "step": 23285 }, { "epoch": 0.745470840535177, "grad_norm": 0.9204493165016174, "learning_rate": 1.8613645123705703e-05, "loss": 0.4084, "num_input_tokens_seen": 31556752, "step": 23290 }, { "epoch": 0.7456308815056655, "grad_norm": 1.1889985799789429, "learning_rate": 1.8604153871841328e-05, "loss": 0.6777, "num_input_tokens_seen": 31563680, "step": 23295 }, { "epoch": 0.7457909224761539, "grad_norm": 1.5045557022094727, "learning_rate": 1.859466360629682e-05, "loss": 0.6396, "num_input_tokens_seen": 31570384, "step": 23300 }, { "epoch": 0.7459509634466424, "grad_norm": 0.9895270466804504, "learning_rate": 1.8585174328535666e-05, "loss": 0.4892, "num_input_tokens_seen": 31577440, "step": 23305 }, { "epoch": 0.7461110044171307, "grad_norm": 1.2441744804382324, "learning_rate": 1.857568604002124e-05, "loss": 0.5501, "num_input_tokens_seen": 31584512, "step": 23310 }, { "epoch": 0.7462710453876192, "grad_norm": 0.6995877027511597, "learning_rate": 1.8566198742216774e-05, "loss": 0.2893, "num_input_tokens_seen": 31591344, "step": 23315 }, { "epoch": 0.7464310863581076, "grad_norm": 1.1675786972045898, "learning_rate": 1.85567124365853e-05, "loss": 0.3185, "num_input_tokens_seen": 31598160, "step": 23320 }, { "epoch": 0.7465911273285961, "grad_norm": 0.7830284237861633, "learning_rate": 1.854722712458975e-05, "loss": 0.4217, "num_input_tokens_seen": 31605072, "step": 23325 }, { "epoch": 0.7467511682990846, "grad_norm": 0.809174120426178, "learning_rate": 1.853774280769286e-05, "loss": 0.467, "num_input_tokens_seen": 31611760, "step": 23330 }, { "epoch": 0.746911209269573, "grad_norm": 1.9357867240905762, "learning_rate": 1.852825948735724e-05, "loss": 0.4372, "num_input_tokens_seen": 31619264, "step": 23335 }, { "epoch": 0.7470712502400615, "grad_norm": 1.758955955505371, "learning_rate": 1.851877716504534e-05, "loss": 0.4155, "num_input_tokens_seen": 31626384, "step": 23340 }, { "epoch": 0.7472312912105499, "grad_norm": 1.4103389978408813, "learning_rate": 1.8509295842219448e-05, "loss": 0.4357, "num_input_tokens_seen": 31632624, "step": 23345 }, { "epoch": 0.7473913321810384, "grad_norm": 0.6663206219673157, "learning_rate": 1.8499815520341697e-05, "loss": 0.3377, "num_input_tokens_seen": 31639584, "step": 23350 }, { "epoch": 0.7475513731515268, "grad_norm": 1.0164093971252441, "learning_rate": 1.8490336200874094e-05, "loss": 0.2993, "num_input_tokens_seen": 31646176, "step": 23355 }, { "epoch": 0.7477114141220152, "grad_norm": 0.7201242446899414, "learning_rate": 1.848085788527844e-05, "loss": 0.3602, "num_input_tokens_seen": 31652896, "step": 23360 }, { "epoch": 0.7478714550925036, "grad_norm": 0.4895721673965454, "learning_rate": 1.847138057501644e-05, "loss": 0.4536, "num_input_tokens_seen": 31659472, "step": 23365 }, { "epoch": 0.7480314960629921, "grad_norm": 0.6944164633750916, "learning_rate": 1.8461904271549582e-05, "loss": 0.4141, "num_input_tokens_seen": 31666672, "step": 23370 }, { "epoch": 0.7481915370334806, "grad_norm": 1.4860292673110962, "learning_rate": 1.845242897633926e-05, "loss": 0.4346, "num_input_tokens_seen": 31673472, "step": 23375 }, { "epoch": 0.748351578003969, "grad_norm": 0.9831204414367676, "learning_rate": 1.844295469084667e-05, "loss": 0.5228, "num_input_tokens_seen": 31680464, "step": 23380 }, { "epoch": 0.7485116189744575, "grad_norm": 2.4687540531158447, "learning_rate": 1.843348141653286e-05, "loss": 0.4668, "num_input_tokens_seen": 31687264, "step": 23385 }, { "epoch": 0.7486716599449459, "grad_norm": 1.1187057495117188, "learning_rate": 1.842400915485874e-05, "loss": 0.3369, "num_input_tokens_seen": 31694128, "step": 23390 }, { "epoch": 0.7488317009154344, "grad_norm": 0.5582628846168518, "learning_rate": 1.8414537907285053e-05, "loss": 0.3823, "num_input_tokens_seen": 31701168, "step": 23395 }, { "epoch": 0.7489917418859228, "grad_norm": 1.1492524147033691, "learning_rate": 1.840506767527237e-05, "loss": 0.5793, "num_input_tokens_seen": 31708288, "step": 23400 }, { "epoch": 0.7489917418859228, "eval_loss": 0.453313410282135, "eval_runtime": 502.3971, "eval_samples_per_second": 27.639, "eval_steps_per_second": 13.82, "num_input_tokens_seen": 31708288, "step": 23400 }, { "epoch": 0.7491517828564113, "grad_norm": 2.9157965183258057, "learning_rate": 1.8395598460281137e-05, "loss": 0.6211, "num_input_tokens_seen": 31715040, "step": 23405 }, { "epoch": 0.7493118238268996, "grad_norm": 0.8476764559745789, "learning_rate": 1.838613026377161e-05, "loss": 0.2823, "num_input_tokens_seen": 31721920, "step": 23410 }, { "epoch": 0.7494718647973881, "grad_norm": 2.1062803268432617, "learning_rate": 1.8376663087203917e-05, "loss": 0.5875, "num_input_tokens_seen": 31728352, "step": 23415 }, { "epoch": 0.7496319057678765, "grad_norm": 1.736702561378479, "learning_rate": 1.8367196932038014e-05, "loss": 0.5849, "num_input_tokens_seen": 31735488, "step": 23420 }, { "epoch": 0.749791946738365, "grad_norm": 1.6178343296051025, "learning_rate": 1.8357731799733686e-05, "loss": 0.4687, "num_input_tokens_seen": 31742336, "step": 23425 }, { "epoch": 0.7499519877088535, "grad_norm": 1.018661379814148, "learning_rate": 1.8348267691750586e-05, "loss": 0.5314, "num_input_tokens_seen": 31748752, "step": 23430 }, { "epoch": 0.7501120286793419, "grad_norm": 0.8531097769737244, "learning_rate": 1.833880460954821e-05, "loss": 0.4373, "num_input_tokens_seen": 31755376, "step": 23435 }, { "epoch": 0.7502720696498304, "grad_norm": 1.2604491710662842, "learning_rate": 1.8329342554585866e-05, "loss": 0.4241, "num_input_tokens_seen": 31762176, "step": 23440 }, { "epoch": 0.7504321106203188, "grad_norm": 1.137133002281189, "learning_rate": 1.8319881528322735e-05, "loss": 0.4709, "num_input_tokens_seen": 31768896, "step": 23445 }, { "epoch": 0.7505921515908073, "grad_norm": 0.7314408421516418, "learning_rate": 1.8310421532217815e-05, "loss": 0.3796, "num_input_tokens_seen": 31775648, "step": 23450 }, { "epoch": 0.7507521925612957, "grad_norm": 1.174717664718628, "learning_rate": 1.8300962567729958e-05, "loss": 0.4699, "num_input_tokens_seen": 31782880, "step": 23455 }, { "epoch": 0.7509122335317842, "grad_norm": 1.6129670143127441, "learning_rate": 1.8291504636317866e-05, "loss": 0.4944, "num_input_tokens_seen": 31789120, "step": 23460 }, { "epoch": 0.7510722745022725, "grad_norm": 1.097809910774231, "learning_rate": 1.8282047739440055e-05, "loss": 0.3811, "num_input_tokens_seen": 31796192, "step": 23465 }, { "epoch": 0.751232315472761, "grad_norm": 0.4840260148048401, "learning_rate": 1.8272591878554903e-05, "loss": 0.2282, "num_input_tokens_seen": 31803184, "step": 23470 }, { "epoch": 0.7513923564432494, "grad_norm": 1.5210437774658203, "learning_rate": 1.8263137055120638e-05, "loss": 0.439, "num_input_tokens_seen": 31810176, "step": 23475 }, { "epoch": 0.7515523974137379, "grad_norm": 1.882192850112915, "learning_rate": 1.8253683270595295e-05, "loss": 0.5614, "num_input_tokens_seen": 31817088, "step": 23480 }, { "epoch": 0.7517124383842264, "grad_norm": 1.0754852294921875, "learning_rate": 1.824423052643677e-05, "loss": 0.6261, "num_input_tokens_seen": 31823728, "step": 23485 }, { "epoch": 0.7518724793547148, "grad_norm": 1.1207233667373657, "learning_rate": 1.82347788241028e-05, "loss": 0.6461, "num_input_tokens_seen": 31830704, "step": 23490 }, { "epoch": 0.7520325203252033, "grad_norm": 1.1066449880599976, "learning_rate": 1.8225328165050942e-05, "loss": 0.3633, "num_input_tokens_seen": 31837936, "step": 23495 }, { "epoch": 0.7521925612956917, "grad_norm": 0.7996084690093994, "learning_rate": 1.821587855073863e-05, "loss": 0.4509, "num_input_tokens_seen": 31845488, "step": 23500 }, { "epoch": 0.7523526022661802, "grad_norm": 1.1953394412994385, "learning_rate": 1.8206429982623086e-05, "loss": 0.4613, "num_input_tokens_seen": 31852128, "step": 23505 }, { "epoch": 0.7525126432366686, "grad_norm": 1.2282423973083496, "learning_rate": 1.8196982462161416e-05, "loss": 0.4449, "num_input_tokens_seen": 31858624, "step": 23510 }, { "epoch": 0.752672684207157, "grad_norm": 1.1248033046722412, "learning_rate": 1.818753599081055e-05, "loss": 0.651, "num_input_tokens_seen": 31865360, "step": 23515 }, { "epoch": 0.7528327251776454, "grad_norm": 1.4817156791687012, "learning_rate": 1.817809057002724e-05, "loss": 0.5433, "num_input_tokens_seen": 31871712, "step": 23520 }, { "epoch": 0.7529927661481339, "grad_norm": 0.7126525640487671, "learning_rate": 1.8168646201268096e-05, "loss": 0.3136, "num_input_tokens_seen": 31878688, "step": 23525 }, { "epoch": 0.7531528071186223, "grad_norm": 0.7782636880874634, "learning_rate": 1.8159202885989557e-05, "loss": 0.3386, "num_input_tokens_seen": 31885536, "step": 23530 }, { "epoch": 0.7533128480891108, "grad_norm": 1.262356162071228, "learning_rate": 1.814976062564789e-05, "loss": 0.3921, "num_input_tokens_seen": 31892912, "step": 23535 }, { "epoch": 0.7534728890595993, "grad_norm": 0.7488334774971008, "learning_rate": 1.8140319421699234e-05, "loss": 0.3078, "num_input_tokens_seen": 31899984, "step": 23540 }, { "epoch": 0.7536329300300877, "grad_norm": 0.994171142578125, "learning_rate": 1.8130879275599515e-05, "loss": 0.349, "num_input_tokens_seen": 31906784, "step": 23545 }, { "epoch": 0.7537929710005762, "grad_norm": 1.8683176040649414, "learning_rate": 1.8121440188804544e-05, "loss": 0.4167, "num_input_tokens_seen": 31913488, "step": 23550 }, { "epoch": 0.7539530119710646, "grad_norm": 0.8736310005187988, "learning_rate": 1.811200216276993e-05, "loss": 0.5569, "num_input_tokens_seen": 31920672, "step": 23555 }, { "epoch": 0.7541130529415531, "grad_norm": 0.8198381662368774, "learning_rate": 1.810256519895115e-05, "loss": 0.4275, "num_input_tokens_seen": 31927424, "step": 23560 }, { "epoch": 0.7542730939120414, "grad_norm": 0.6132899522781372, "learning_rate": 1.8093129298803494e-05, "loss": 0.3935, "num_input_tokens_seen": 31934496, "step": 23565 }, { "epoch": 0.7544331348825299, "grad_norm": 1.2358503341674805, "learning_rate": 1.808369446378209e-05, "loss": 0.3826, "num_input_tokens_seen": 31941392, "step": 23570 }, { "epoch": 0.7545931758530183, "grad_norm": 1.3520578145980835, "learning_rate": 1.8074260695341914e-05, "loss": 0.4227, "num_input_tokens_seen": 31948144, "step": 23575 }, { "epoch": 0.7547532168235068, "grad_norm": 2.542442798614502, "learning_rate": 1.8064827994937782e-05, "loss": 0.6459, "num_input_tokens_seen": 31954976, "step": 23580 }, { "epoch": 0.7549132577939953, "grad_norm": 2.049315929412842, "learning_rate": 1.8055396364024317e-05, "loss": 0.5675, "num_input_tokens_seen": 31961632, "step": 23585 }, { "epoch": 0.7550732987644837, "grad_norm": 0.7700329422950745, "learning_rate": 1.804596580405601e-05, "loss": 0.4032, "num_input_tokens_seen": 31968656, "step": 23590 }, { "epoch": 0.7552333397349722, "grad_norm": 2.04097580909729, "learning_rate": 1.8036536316487174e-05, "loss": 0.4842, "num_input_tokens_seen": 31975184, "step": 23595 }, { "epoch": 0.7553933807054606, "grad_norm": 0.8322122693061829, "learning_rate": 1.802710790277193e-05, "loss": 0.5206, "num_input_tokens_seen": 31982128, "step": 23600 }, { "epoch": 0.7553933807054606, "eval_loss": 0.45335549116134644, "eval_runtime": 502.7972, "eval_samples_per_second": 27.617, "eval_steps_per_second": 13.809, "num_input_tokens_seen": 31982128, "step": 23600 }, { "epoch": 0.7555534216759491, "grad_norm": 0.5826030373573303, "learning_rate": 1.801768056436429e-05, "loss": 0.4985, "num_input_tokens_seen": 31988560, "step": 23605 }, { "epoch": 0.7557134626464375, "grad_norm": 0.8320585489273071, "learning_rate": 1.8008254302718035e-05, "loss": 0.4803, "num_input_tokens_seen": 31995552, "step": 23610 }, { "epoch": 0.755873503616926, "grad_norm": 0.7206472754478455, "learning_rate": 1.7998829119286837e-05, "loss": 0.5336, "num_input_tokens_seen": 32002496, "step": 23615 }, { "epoch": 0.7560335445874143, "grad_norm": 0.9988227486610413, "learning_rate": 1.798940501552418e-05, "loss": 0.3719, "num_input_tokens_seen": 32009216, "step": 23620 }, { "epoch": 0.7561935855579028, "grad_norm": 1.2997978925704956, "learning_rate": 1.797998199288336e-05, "loss": 0.3365, "num_input_tokens_seen": 32015456, "step": 23625 }, { "epoch": 0.7563536265283912, "grad_norm": 0.5526777505874634, "learning_rate": 1.7970560052817543e-05, "loss": 0.3841, "num_input_tokens_seen": 32022832, "step": 23630 }, { "epoch": 0.7565136674988797, "grad_norm": 0.45208755135536194, "learning_rate": 1.7961139196779702e-05, "loss": 0.3705, "num_input_tokens_seen": 32029328, "step": 23635 }, { "epoch": 0.7566737084693682, "grad_norm": 0.9762411713600159, "learning_rate": 1.7951719426222647e-05, "loss": 0.4707, "num_input_tokens_seen": 32036016, "step": 23640 }, { "epoch": 0.7568337494398566, "grad_norm": 1.1968648433685303, "learning_rate": 1.794230074259904e-05, "loss": 0.3963, "num_input_tokens_seen": 32042608, "step": 23645 }, { "epoch": 0.7569937904103451, "grad_norm": 0.786974310874939, "learning_rate": 1.7932883147361336e-05, "loss": 0.4307, "num_input_tokens_seen": 32049392, "step": 23650 }, { "epoch": 0.7571538313808335, "grad_norm": 1.6454559564590454, "learning_rate": 1.7923466641961865e-05, "loss": 0.464, "num_input_tokens_seen": 32056176, "step": 23655 }, { "epoch": 0.757313872351322, "grad_norm": 1.0527316331863403, "learning_rate": 1.791405122785278e-05, "loss": 0.4157, "num_input_tokens_seen": 32062992, "step": 23660 }, { "epoch": 0.7574739133218104, "grad_norm": 1.2597399950027466, "learning_rate": 1.7904636906486037e-05, "loss": 0.5365, "num_input_tokens_seen": 32069552, "step": 23665 }, { "epoch": 0.7576339542922989, "grad_norm": 0.996519148349762, "learning_rate": 1.7895223679313448e-05, "loss": 0.3728, "num_input_tokens_seen": 32076304, "step": 23670 }, { "epoch": 0.7577939952627872, "grad_norm": 1.3071792125701904, "learning_rate": 1.7885811547786653e-05, "loss": 0.4169, "num_input_tokens_seen": 32082928, "step": 23675 }, { "epoch": 0.7579540362332757, "grad_norm": 1.1237061023712158, "learning_rate": 1.7876400513357115e-05, "loss": 0.4267, "num_input_tokens_seen": 32089888, "step": 23680 }, { "epoch": 0.7581140772037641, "grad_norm": 1.0029144287109375, "learning_rate": 1.7866990577476146e-05, "loss": 0.3119, "num_input_tokens_seen": 32096432, "step": 23685 }, { "epoch": 0.7582741181742526, "grad_norm": 1.598319172859192, "learning_rate": 1.7857581741594863e-05, "loss": 0.435, "num_input_tokens_seen": 32103088, "step": 23690 }, { "epoch": 0.7584341591447411, "grad_norm": 1.077052116394043, "learning_rate": 1.7848174007164237e-05, "loss": 0.362, "num_input_tokens_seen": 32109744, "step": 23695 }, { "epoch": 0.7585942001152295, "grad_norm": 1.3883322477340698, "learning_rate": 1.7838767375635052e-05, "loss": 0.5006, "num_input_tokens_seen": 32116512, "step": 23700 }, { "epoch": 0.758754241085718, "grad_norm": 0.8895444273948669, "learning_rate": 1.782936184845793e-05, "loss": 0.4384, "num_input_tokens_seen": 32123120, "step": 23705 }, { "epoch": 0.7589142820562064, "grad_norm": 1.308905839920044, "learning_rate": 1.7819957427083334e-05, "loss": 0.4803, "num_input_tokens_seen": 32129840, "step": 23710 }, { "epoch": 0.7590743230266949, "grad_norm": 1.0074717998504639, "learning_rate": 1.7810554112961516e-05, "loss": 0.4082, "num_input_tokens_seen": 32136544, "step": 23715 }, { "epoch": 0.7592343639971832, "grad_norm": 0.6311304569244385, "learning_rate": 1.7801151907542607e-05, "loss": 0.3669, "num_input_tokens_seen": 32143072, "step": 23720 }, { "epoch": 0.7593944049676717, "grad_norm": 0.9234102964401245, "learning_rate": 1.7791750812276547e-05, "loss": 0.5653, "num_input_tokens_seen": 32149888, "step": 23725 }, { "epoch": 0.7595544459381601, "grad_norm": 0.7204322814941406, "learning_rate": 1.778235082861309e-05, "loss": 0.4796, "num_input_tokens_seen": 32157088, "step": 23730 }, { "epoch": 0.7597144869086486, "grad_norm": 0.7160316109657288, "learning_rate": 1.777295195800184e-05, "loss": 0.4471, "num_input_tokens_seen": 32163424, "step": 23735 }, { "epoch": 0.759874527879137, "grad_norm": 1.2926750183105469, "learning_rate": 1.7763554201892215e-05, "loss": 0.5115, "num_input_tokens_seen": 32170352, "step": 23740 }, { "epoch": 0.7600345688496255, "grad_norm": 1.4761971235275269, "learning_rate": 1.7754157561733476e-05, "loss": 0.4436, "num_input_tokens_seen": 32177312, "step": 23745 }, { "epoch": 0.760194609820114, "grad_norm": 0.9608497619628906, "learning_rate": 1.7744762038974702e-05, "loss": 0.3501, "num_input_tokens_seen": 32184336, "step": 23750 }, { "epoch": 0.7603546507906024, "grad_norm": 1.361330509185791, "learning_rate": 1.7735367635064788e-05, "loss": 0.3784, "num_input_tokens_seen": 32191200, "step": 23755 }, { "epoch": 0.7605146917610909, "grad_norm": 1.3501476049423218, "learning_rate": 1.7725974351452474e-05, "loss": 0.5104, "num_input_tokens_seen": 32197904, "step": 23760 }, { "epoch": 0.7606747327315793, "grad_norm": 1.724169373512268, "learning_rate": 1.771658218958634e-05, "loss": 0.7806, "num_input_tokens_seen": 32204768, "step": 23765 }, { "epoch": 0.7608347737020678, "grad_norm": 0.7665554881095886, "learning_rate": 1.770719115091475e-05, "loss": 0.5152, "num_input_tokens_seen": 32211728, "step": 23770 }, { "epoch": 0.7609948146725561, "grad_norm": 0.8097151517868042, "learning_rate": 1.7697801236885935e-05, "loss": 0.4612, "num_input_tokens_seen": 32218784, "step": 23775 }, { "epoch": 0.7611548556430446, "grad_norm": 0.5712356567382812, "learning_rate": 1.7688412448947944e-05, "loss": 0.3525, "num_input_tokens_seen": 32225488, "step": 23780 }, { "epoch": 0.761314896613533, "grad_norm": 1.040644645690918, "learning_rate": 1.767902478854862e-05, "loss": 0.4661, "num_input_tokens_seen": 32232944, "step": 23785 }, { "epoch": 0.7614749375840215, "grad_norm": 1.061659812927246, "learning_rate": 1.766963825713569e-05, "loss": 0.4173, "num_input_tokens_seen": 32239504, "step": 23790 }, { "epoch": 0.7616349785545099, "grad_norm": 0.9516823887825012, "learning_rate": 1.766025285615665e-05, "loss": 0.4794, "num_input_tokens_seen": 32246304, "step": 23795 }, { "epoch": 0.7617950195249984, "grad_norm": 0.7223002910614014, "learning_rate": 1.7650868587058854e-05, "loss": 0.4382, "num_input_tokens_seen": 32253040, "step": 23800 }, { "epoch": 0.7617950195249984, "eval_loss": 0.4533097743988037, "eval_runtime": 502.6846, "eval_samples_per_second": 27.624, "eval_steps_per_second": 13.812, "num_input_tokens_seen": 32253040, "step": 23800 }, { "epoch": 0.7619550604954869, "grad_norm": 1.40898859500885, "learning_rate": 1.7641485451289484e-05, "loss": 0.5016, "num_input_tokens_seen": 32259824, "step": 23805 }, { "epoch": 0.7621151014659753, "grad_norm": 1.11431086063385, "learning_rate": 1.7632103450295534e-05, "loss": 0.5214, "num_input_tokens_seen": 32266896, "step": 23810 }, { "epoch": 0.7622751424364638, "grad_norm": 0.9113686680793762, "learning_rate": 1.762272258552381e-05, "loss": 0.3729, "num_input_tokens_seen": 32273920, "step": 23815 }, { "epoch": 0.7624351834069522, "grad_norm": 1.4125170707702637, "learning_rate": 1.7613342858420988e-05, "loss": 0.444, "num_input_tokens_seen": 32280560, "step": 23820 }, { "epoch": 0.7625952243774407, "grad_norm": 2.528960704803467, "learning_rate": 1.760396427043351e-05, "loss": 0.5361, "num_input_tokens_seen": 32287248, "step": 23825 }, { "epoch": 0.762755265347929, "grad_norm": 1.2621294260025024, "learning_rate": 1.7594586823007696e-05, "loss": 0.4141, "num_input_tokens_seen": 32293616, "step": 23830 }, { "epoch": 0.7629153063184175, "grad_norm": 1.1746522188186646, "learning_rate": 1.7585210517589646e-05, "loss": 0.3904, "num_input_tokens_seen": 32300704, "step": 23835 }, { "epoch": 0.7630753472889059, "grad_norm": 0.4140104353427887, "learning_rate": 1.7575835355625314e-05, "loss": 0.3227, "num_input_tokens_seen": 32307840, "step": 23840 }, { "epoch": 0.7632353882593944, "grad_norm": 1.1973003149032593, "learning_rate": 1.756646133856048e-05, "loss": 0.4106, "num_input_tokens_seen": 32314800, "step": 23845 }, { "epoch": 0.7633954292298829, "grad_norm": 0.39789363741874695, "learning_rate": 1.7557088467840714e-05, "loss": 0.2879, "num_input_tokens_seen": 32321952, "step": 23850 }, { "epoch": 0.7635554702003713, "grad_norm": 0.9275329113006592, "learning_rate": 1.7547716744911438e-05, "loss": 0.431, "num_input_tokens_seen": 32328704, "step": 23855 }, { "epoch": 0.7637155111708598, "grad_norm": 1.0254762172698975, "learning_rate": 1.7538346171217902e-05, "loss": 0.532, "num_input_tokens_seen": 32335536, "step": 23860 }, { "epoch": 0.7638755521413482, "grad_norm": 1.0997556447982788, "learning_rate": 1.7528976748205146e-05, "loss": 0.2838, "num_input_tokens_seen": 32342432, "step": 23865 }, { "epoch": 0.7640355931118367, "grad_norm": 0.8651045560836792, "learning_rate": 1.751960847731807e-05, "loss": 0.3679, "num_input_tokens_seen": 32349024, "step": 23870 }, { "epoch": 0.764195634082325, "grad_norm": 2.0053510665893555, "learning_rate": 1.7510241360001362e-05, "loss": 0.437, "num_input_tokens_seen": 32355520, "step": 23875 }, { "epoch": 0.7643556750528135, "grad_norm": 0.9419711232185364, "learning_rate": 1.7500875397699562e-05, "loss": 0.4501, "num_input_tokens_seen": 32362352, "step": 23880 }, { "epoch": 0.7645157160233019, "grad_norm": 0.8395537734031677, "learning_rate": 1.7491510591857015e-05, "loss": 0.4109, "num_input_tokens_seen": 32369392, "step": 23885 }, { "epoch": 0.7646757569937904, "grad_norm": 1.2820402383804321, "learning_rate": 1.7482146943917896e-05, "loss": 0.4031, "num_input_tokens_seen": 32376672, "step": 23890 }, { "epoch": 0.7648357979642788, "grad_norm": 0.7714034914970398, "learning_rate": 1.7472784455326185e-05, "loss": 0.5898, "num_input_tokens_seen": 32383312, "step": 23895 }, { "epoch": 0.7649958389347673, "grad_norm": 1.4772957563400269, "learning_rate": 1.746342312752572e-05, "loss": 0.4845, "num_input_tokens_seen": 32390208, "step": 23900 }, { "epoch": 0.7651558799052558, "grad_norm": 1.2001577615737915, "learning_rate": 1.74540629619601e-05, "loss": 0.4367, "num_input_tokens_seen": 32397040, "step": 23905 }, { "epoch": 0.7653159208757442, "grad_norm": 0.9709440469741821, "learning_rate": 1.7444703960072815e-05, "loss": 0.4111, "num_input_tokens_seen": 32403584, "step": 23910 }, { "epoch": 0.7654759618462327, "grad_norm": 1.0610339641571045, "learning_rate": 1.7435346123307118e-05, "loss": 0.663, "num_input_tokens_seen": 32410320, "step": 23915 }, { "epoch": 0.7656360028167211, "grad_norm": 0.8542266488075256, "learning_rate": 1.742598945310611e-05, "loss": 0.4438, "num_input_tokens_seen": 32417168, "step": 23920 }, { "epoch": 0.7657960437872096, "grad_norm": 0.8770877122879028, "learning_rate": 1.741663395091272e-05, "loss": 0.4713, "num_input_tokens_seen": 32423536, "step": 23925 }, { "epoch": 0.7659560847576979, "grad_norm": 1.3699569702148438, "learning_rate": 1.7407279618169657e-05, "loss": 0.5638, "num_input_tokens_seen": 32430080, "step": 23930 }, { "epoch": 0.7661161257281864, "grad_norm": 1.6465157270431519, "learning_rate": 1.73979264563195e-05, "loss": 0.5039, "num_input_tokens_seen": 32436608, "step": 23935 }, { "epoch": 0.7662761666986748, "grad_norm": 1.365645408630371, "learning_rate": 1.7388574466804625e-05, "loss": 0.5305, "num_input_tokens_seen": 32443424, "step": 23940 }, { "epoch": 0.7664362076691633, "grad_norm": 0.41198745369911194, "learning_rate": 1.7379223651067207e-05, "loss": 0.3571, "num_input_tokens_seen": 32450096, "step": 23945 }, { "epoch": 0.7665962486396517, "grad_norm": 1.6836379766464233, "learning_rate": 1.736987401054928e-05, "loss": 0.5137, "num_input_tokens_seen": 32456960, "step": 23950 }, { "epoch": 0.7667562896101402, "grad_norm": 0.6789073944091797, "learning_rate": 1.736052554669266e-05, "loss": 0.3279, "num_input_tokens_seen": 32463552, "step": 23955 }, { "epoch": 0.7669163305806287, "grad_norm": 1.1893285512924194, "learning_rate": 1.7351178260939007e-05, "loss": 0.5574, "num_input_tokens_seen": 32470288, "step": 23960 }, { "epoch": 0.7670763715511171, "grad_norm": 1.7354059219360352, "learning_rate": 1.7341832154729794e-05, "loss": 0.4005, "num_input_tokens_seen": 32477088, "step": 23965 }, { "epoch": 0.7672364125216056, "grad_norm": 1.3459430932998657, "learning_rate": 1.7332487229506286e-05, "loss": 0.4443, "num_input_tokens_seen": 32484240, "step": 23970 }, { "epoch": 0.767396453492094, "grad_norm": 0.9039115905761719, "learning_rate": 1.732314348670961e-05, "loss": 0.253, "num_input_tokens_seen": 32491152, "step": 23975 }, { "epoch": 0.7675564944625825, "grad_norm": 1.7129099369049072, "learning_rate": 1.7313800927780686e-05, "loss": 0.5178, "num_input_tokens_seen": 32497680, "step": 23980 }, { "epoch": 0.7677165354330708, "grad_norm": 1.9325884580612183, "learning_rate": 1.7304459554160245e-05, "loss": 0.5974, "num_input_tokens_seen": 32504576, "step": 23985 }, { "epoch": 0.7678765764035593, "grad_norm": 1.0054341554641724, "learning_rate": 1.7295119367288853e-05, "loss": 0.3967, "num_input_tokens_seen": 32511136, "step": 23990 }, { "epoch": 0.7680366173740477, "grad_norm": 2.0906951427459717, "learning_rate": 1.728578036860688e-05, "loss": 0.4555, "num_input_tokens_seen": 32517792, "step": 23995 }, { "epoch": 0.7681966583445362, "grad_norm": 2.437523365020752, "learning_rate": 1.7276442559554513e-05, "loss": 0.464, "num_input_tokens_seen": 32524464, "step": 24000 }, { "epoch": 0.7681966583445362, "eval_loss": 0.4531278610229492, "eval_runtime": 502.7046, "eval_samples_per_second": 27.623, "eval_steps_per_second": 13.811, "num_input_tokens_seen": 32524464, "step": 24000 }, { "epoch": 0.7683566993150246, "grad_norm": 0.7984288334846497, "learning_rate": 1.726710594157177e-05, "loss": 0.3242, "num_input_tokens_seen": 32531040, "step": 24005 }, { "epoch": 0.7685167402855131, "grad_norm": 0.7580831050872803, "learning_rate": 1.725777051609846e-05, "loss": 0.3357, "num_input_tokens_seen": 32537664, "step": 24010 }, { "epoch": 0.7686767812560016, "grad_norm": 0.8964778184890747, "learning_rate": 1.7248436284574228e-05, "loss": 0.5268, "num_input_tokens_seen": 32544336, "step": 24015 }, { "epoch": 0.76883682222649, "grad_norm": 1.4862940311431885, "learning_rate": 1.723910324843855e-05, "loss": 0.3517, "num_input_tokens_seen": 32551280, "step": 24020 }, { "epoch": 0.7689968631969785, "grad_norm": 1.1065462827682495, "learning_rate": 1.722977140913067e-05, "loss": 0.3108, "num_input_tokens_seen": 32558304, "step": 24025 }, { "epoch": 0.7691569041674668, "grad_norm": 1.0971518754959106, "learning_rate": 1.7220440768089688e-05, "loss": 0.5821, "num_input_tokens_seen": 32564720, "step": 24030 }, { "epoch": 0.7693169451379553, "grad_norm": 1.3702201843261719, "learning_rate": 1.7211111326754505e-05, "loss": 0.4672, "num_input_tokens_seen": 32571760, "step": 24035 }, { "epoch": 0.7694769861084437, "grad_norm": 0.48422813415527344, "learning_rate": 1.720178308656383e-05, "loss": 0.5295, "num_input_tokens_seen": 32578160, "step": 24040 }, { "epoch": 0.7696370270789322, "grad_norm": 0.828641414642334, "learning_rate": 1.719245604895621e-05, "loss": 0.3506, "num_input_tokens_seen": 32584832, "step": 24045 }, { "epoch": 0.7697970680494206, "grad_norm": 3.651768922805786, "learning_rate": 1.7183130215369972e-05, "loss": 0.5058, "num_input_tokens_seen": 32591296, "step": 24050 }, { "epoch": 0.7699571090199091, "grad_norm": 2.3396122455596924, "learning_rate": 1.7173805587243292e-05, "loss": 0.464, "num_input_tokens_seen": 32597920, "step": 24055 }, { "epoch": 0.7701171499903975, "grad_norm": 1.6659170389175415, "learning_rate": 1.7164482166014147e-05, "loss": 0.5095, "num_input_tokens_seen": 32604704, "step": 24060 }, { "epoch": 0.770277190960886, "grad_norm": 0.6066564917564392, "learning_rate": 1.7155159953120313e-05, "loss": 0.4671, "num_input_tokens_seen": 32611456, "step": 24065 }, { "epoch": 0.7704372319313745, "grad_norm": 1.249688744544983, "learning_rate": 1.714583894999941e-05, "loss": 0.4507, "num_input_tokens_seen": 32618064, "step": 24070 }, { "epoch": 0.7705972729018629, "grad_norm": 0.9270922541618347, "learning_rate": 1.7136519158088826e-05, "loss": 0.4014, "num_input_tokens_seen": 32624672, "step": 24075 }, { "epoch": 0.7707573138723514, "grad_norm": 0.6318855881690979, "learning_rate": 1.712720057882581e-05, "loss": 0.3437, "num_input_tokens_seen": 32631248, "step": 24080 }, { "epoch": 0.7709173548428397, "grad_norm": 0.4902842938899994, "learning_rate": 1.7117883213647413e-05, "loss": 0.4187, "num_input_tokens_seen": 32637856, "step": 24085 }, { "epoch": 0.7710773958133282, "grad_norm": 0.7766571640968323, "learning_rate": 1.710856706399046e-05, "loss": 0.3587, "num_input_tokens_seen": 32644768, "step": 24090 }, { "epoch": 0.7712374367838166, "grad_norm": 1.6034799814224243, "learning_rate": 1.7099252131291648e-05, "loss": 0.4248, "num_input_tokens_seen": 32651744, "step": 24095 }, { "epoch": 0.7713974777543051, "grad_norm": 1.6370586156845093, "learning_rate": 1.708993841698744e-05, "loss": 0.4439, "num_input_tokens_seen": 32658176, "step": 24100 }, { "epoch": 0.7715575187247935, "grad_norm": 1.4095202684402466, "learning_rate": 1.7080625922514132e-05, "loss": 0.6032, "num_input_tokens_seen": 32664944, "step": 24105 }, { "epoch": 0.771717559695282, "grad_norm": 0.5341885685920715, "learning_rate": 1.7071314649307836e-05, "loss": 0.3181, "num_input_tokens_seen": 32671456, "step": 24110 }, { "epoch": 0.7718776006657705, "grad_norm": 0.6938349604606628, "learning_rate": 1.7062004598804448e-05, "loss": 0.3726, "num_input_tokens_seen": 32678368, "step": 24115 }, { "epoch": 0.7720376416362589, "grad_norm": 0.8587434887886047, "learning_rate": 1.7052695772439702e-05, "loss": 0.342, "num_input_tokens_seen": 32685248, "step": 24120 }, { "epoch": 0.7721976826067474, "grad_norm": 1.166662335395813, "learning_rate": 1.7043388171649154e-05, "loss": 0.4627, "num_input_tokens_seen": 32691952, "step": 24125 }, { "epoch": 0.7723577235772358, "grad_norm": 0.790028989315033, "learning_rate": 1.7034081797868127e-05, "loss": 0.3095, "num_input_tokens_seen": 32699600, "step": 24130 }, { "epoch": 0.7725177645477243, "grad_norm": 0.8514086604118347, "learning_rate": 1.70247766525318e-05, "loss": 0.5009, "num_input_tokens_seen": 32706576, "step": 24135 }, { "epoch": 0.7726778055182126, "grad_norm": 1.6523756980895996, "learning_rate": 1.701547273707514e-05, "loss": 0.6293, "num_input_tokens_seen": 32713488, "step": 24140 }, { "epoch": 0.7728378464887011, "grad_norm": 0.9783941507339478, "learning_rate": 1.7006170052932916e-05, "loss": 0.468, "num_input_tokens_seen": 32720496, "step": 24145 }, { "epoch": 0.7729978874591895, "grad_norm": 0.8477450609207153, "learning_rate": 1.6996868601539735e-05, "loss": 0.5465, "num_input_tokens_seen": 32726896, "step": 24150 }, { "epoch": 0.773157928429678, "grad_norm": 1.4983817338943481, "learning_rate": 1.6987568384329977e-05, "loss": 0.3969, "num_input_tokens_seen": 32733840, "step": 24155 }, { "epoch": 0.7733179694001664, "grad_norm": 1.0196316242218018, "learning_rate": 1.6978269402737866e-05, "loss": 0.409, "num_input_tokens_seen": 32740336, "step": 24160 }, { "epoch": 0.7734780103706549, "grad_norm": 3.3653182983398438, "learning_rate": 1.696897165819743e-05, "loss": 0.5524, "num_input_tokens_seen": 32747232, "step": 24165 }, { "epoch": 0.7736380513411434, "grad_norm": 1.6687778234481812, "learning_rate": 1.6959675152142487e-05, "loss": 0.4677, "num_input_tokens_seen": 32754000, "step": 24170 }, { "epoch": 0.7737980923116318, "grad_norm": 0.880802571773529, "learning_rate": 1.6950379886006667e-05, "loss": 0.3824, "num_input_tokens_seen": 32760448, "step": 24175 }, { "epoch": 0.7739581332821203, "grad_norm": 0.6581676006317139, "learning_rate": 1.6941085861223438e-05, "loss": 0.3687, "num_input_tokens_seen": 32767296, "step": 24180 }, { "epoch": 0.7741181742526086, "grad_norm": 0.42516231536865234, "learning_rate": 1.6931793079226034e-05, "loss": 0.3848, "num_input_tokens_seen": 32774672, "step": 24185 }, { "epoch": 0.7742782152230971, "grad_norm": 0.7712870836257935, "learning_rate": 1.692250154144754e-05, "loss": 0.298, "num_input_tokens_seen": 32781072, "step": 24190 }, { "epoch": 0.7744382561935855, "grad_norm": 0.711495578289032, "learning_rate": 1.6913211249320807e-05, "loss": 0.3858, "num_input_tokens_seen": 32787936, "step": 24195 }, { "epoch": 0.774598297164074, "grad_norm": 0.9523311853408813, "learning_rate": 1.6903922204278522e-05, "loss": 0.4827, "num_input_tokens_seen": 32794928, "step": 24200 }, { "epoch": 0.774598297164074, "eval_loss": 0.45266446471214294, "eval_runtime": 502.5258, "eval_samples_per_second": 27.632, "eval_steps_per_second": 13.816, "num_input_tokens_seen": 32794928, "step": 24200 }, { "epoch": 0.7747583381345624, "grad_norm": 1.631130337715149, "learning_rate": 1.6894634407753186e-05, "loss": 0.4256, "num_input_tokens_seen": 32801632, "step": 24205 }, { "epoch": 0.7749183791050509, "grad_norm": 0.844215452671051, "learning_rate": 1.6885347861177077e-05, "loss": 0.4951, "num_input_tokens_seen": 32808704, "step": 24210 }, { "epoch": 0.7750784200755393, "grad_norm": 1.0881154537200928, "learning_rate": 1.6876062565982298e-05, "loss": 0.3088, "num_input_tokens_seen": 32815184, "step": 24215 }, { "epoch": 0.7752384610460278, "grad_norm": 0.9074248671531677, "learning_rate": 1.6866778523600774e-05, "loss": 0.4799, "num_input_tokens_seen": 32821888, "step": 24220 }, { "epoch": 0.7753985020165163, "grad_norm": 0.6240242719650269, "learning_rate": 1.6857495735464195e-05, "loss": 0.3616, "num_input_tokens_seen": 32828480, "step": 24225 }, { "epoch": 0.7755585429870047, "grad_norm": 0.928105890750885, "learning_rate": 1.6848214203004115e-05, "loss": 0.5101, "num_input_tokens_seen": 32835056, "step": 24230 }, { "epoch": 0.7757185839574932, "grad_norm": 1.0496116876602173, "learning_rate": 1.6838933927651835e-05, "loss": 0.3444, "num_input_tokens_seen": 32842080, "step": 24235 }, { "epoch": 0.7758786249279815, "grad_norm": 0.6388212442398071, "learning_rate": 1.6829654910838506e-05, "loss": 0.5458, "num_input_tokens_seen": 32848960, "step": 24240 }, { "epoch": 0.77603866589847, "grad_norm": 1.1474863290786743, "learning_rate": 1.6820377153995065e-05, "loss": 0.427, "num_input_tokens_seen": 32855728, "step": 24245 }, { "epoch": 0.7761987068689584, "grad_norm": 0.5774698853492737, "learning_rate": 1.681110065855226e-05, "loss": 0.3882, "num_input_tokens_seen": 32862304, "step": 24250 }, { "epoch": 0.7763587478394469, "grad_norm": 0.8264679908752441, "learning_rate": 1.6801825425940642e-05, "loss": 0.2969, "num_input_tokens_seen": 32868848, "step": 24255 }, { "epoch": 0.7765187888099353, "grad_norm": 0.5351314544677734, "learning_rate": 1.679255145759056e-05, "loss": 0.2577, "num_input_tokens_seen": 32876096, "step": 24260 }, { "epoch": 0.7766788297804238, "grad_norm": 1.1994549036026, "learning_rate": 1.6783278754932187e-05, "loss": 0.468, "num_input_tokens_seen": 32883072, "step": 24265 }, { "epoch": 0.7768388707509122, "grad_norm": 0.4832893908023834, "learning_rate": 1.6774007319395496e-05, "loss": 0.4362, "num_input_tokens_seen": 32890256, "step": 24270 }, { "epoch": 0.7769989117214007, "grad_norm": 0.8279063701629639, "learning_rate": 1.6764737152410243e-05, "loss": 0.4883, "num_input_tokens_seen": 32897200, "step": 24275 }, { "epoch": 0.7771589526918892, "grad_norm": 0.7618823647499084, "learning_rate": 1.6755468255406016e-05, "loss": 0.3939, "num_input_tokens_seen": 32903792, "step": 24280 }, { "epoch": 0.7773189936623776, "grad_norm": 0.6774752736091614, "learning_rate": 1.674620062981219e-05, "loss": 0.2108, "num_input_tokens_seen": 32910592, "step": 24285 }, { "epoch": 0.777479034632866, "grad_norm": 0.7138208746910095, "learning_rate": 1.6736934277057947e-05, "loss": 0.5017, "num_input_tokens_seen": 32917200, "step": 24290 }, { "epoch": 0.7776390756033544, "grad_norm": 0.7166140079498291, "learning_rate": 1.6727669198572286e-05, "loss": 0.3291, "num_input_tokens_seen": 32924016, "step": 24295 }, { "epoch": 0.7777991165738429, "grad_norm": 1.2484456300735474, "learning_rate": 1.6718405395783984e-05, "loss": 0.4991, "num_input_tokens_seen": 32930880, "step": 24300 }, { "epoch": 0.7779591575443313, "grad_norm": 1.2347561120986938, "learning_rate": 1.6709142870121643e-05, "loss": 0.4397, "num_input_tokens_seen": 32937584, "step": 24305 }, { "epoch": 0.7781191985148198, "grad_norm": 1.2694231271743774, "learning_rate": 1.669988162301367e-05, "loss": 0.5292, "num_input_tokens_seen": 32944384, "step": 24310 }, { "epoch": 0.7782792394853082, "grad_norm": 0.9410802125930786, "learning_rate": 1.6690621655888243e-05, "loss": 0.4265, "num_input_tokens_seen": 32951584, "step": 24315 }, { "epoch": 0.7784392804557967, "grad_norm": 1.9944524765014648, "learning_rate": 1.6681362970173386e-05, "loss": 0.4883, "num_input_tokens_seen": 32958320, "step": 24320 }, { "epoch": 0.7785993214262851, "grad_norm": 0.9488465785980225, "learning_rate": 1.6672105567296904e-05, "loss": 0.2921, "num_input_tokens_seen": 32965360, "step": 24325 }, { "epoch": 0.7787593623967736, "grad_norm": 2.3085532188415527, "learning_rate": 1.666284944868639e-05, "loss": 0.5748, "num_input_tokens_seen": 32972176, "step": 24330 }, { "epoch": 0.7789194033672621, "grad_norm": 1.5444865226745605, "learning_rate": 1.665359461576927e-05, "loss": 0.3722, "num_input_tokens_seen": 32978672, "step": 24335 }, { "epoch": 0.7790794443377504, "grad_norm": 1.338379144668579, "learning_rate": 1.6644341069972736e-05, "loss": 0.4628, "num_input_tokens_seen": 32986080, "step": 24340 }, { "epoch": 0.7792394853082389, "grad_norm": 1.3110549449920654, "learning_rate": 1.6635088812723813e-05, "loss": 0.4511, "num_input_tokens_seen": 32992816, "step": 24345 }, { "epoch": 0.7793995262787273, "grad_norm": 0.5982165932655334, "learning_rate": 1.6625837845449328e-05, "loss": 0.4953, "num_input_tokens_seen": 32999504, "step": 24350 }, { "epoch": 0.7795595672492158, "grad_norm": 0.5660097599029541, "learning_rate": 1.6616588169575874e-05, "loss": 0.4168, "num_input_tokens_seen": 33006192, "step": 24355 }, { "epoch": 0.7797196082197042, "grad_norm": 0.9575988054275513, "learning_rate": 1.6607339786529878e-05, "loss": 0.4886, "num_input_tokens_seen": 33012656, "step": 24360 }, { "epoch": 0.7798796491901927, "grad_norm": 2.5156137943267822, "learning_rate": 1.659809269773756e-05, "loss": 0.5881, "num_input_tokens_seen": 33020080, "step": 24365 }, { "epoch": 0.7800396901606811, "grad_norm": 1.5882970094680786, "learning_rate": 1.658884690462493e-05, "loss": 0.4505, "num_input_tokens_seen": 33026736, "step": 24370 }, { "epoch": 0.7801997311311696, "grad_norm": 1.227412223815918, "learning_rate": 1.6579602408617813e-05, "loss": 0.5972, "num_input_tokens_seen": 33033312, "step": 24375 }, { "epoch": 0.7803597721016581, "grad_norm": 1.004668116569519, "learning_rate": 1.657035921114181e-05, "loss": 0.486, "num_input_tokens_seen": 33040656, "step": 24380 }, { "epoch": 0.7805198130721465, "grad_norm": 0.6040318608283997, "learning_rate": 1.656111731362236e-05, "loss": 0.4978, "num_input_tokens_seen": 33047600, "step": 24385 }, { "epoch": 0.780679854042635, "grad_norm": 1.8481765985488892, "learning_rate": 1.6551876717484666e-05, "loss": 0.5421, "num_input_tokens_seen": 33054064, "step": 24390 }, { "epoch": 0.7808398950131233, "grad_norm": 0.6983646154403687, "learning_rate": 1.6542637424153752e-05, "loss": 0.2981, "num_input_tokens_seen": 33060992, "step": 24395 }, { "epoch": 0.7809999359836118, "grad_norm": 1.3139327764511108, "learning_rate": 1.6533399435054418e-05, "loss": 0.5373, "num_input_tokens_seen": 33067904, "step": 24400 }, { "epoch": 0.7809999359836118, "eval_loss": 0.45206522941589355, "eval_runtime": 502.5642, "eval_samples_per_second": 27.63, "eval_steps_per_second": 13.815, "num_input_tokens_seen": 33067904, "step": 24400 }, { "epoch": 0.7811599769541002, "grad_norm": 0.4810771942138672, "learning_rate": 1.6524162751611304e-05, "loss": 0.4101, "num_input_tokens_seen": 33074064, "step": 24405 }, { "epoch": 0.7813200179245887, "grad_norm": 0.6321762800216675, "learning_rate": 1.6514927375248796e-05, "loss": 0.3269, "num_input_tokens_seen": 33080640, "step": 24410 }, { "epoch": 0.7814800588950771, "grad_norm": 1.1481627225875854, "learning_rate": 1.6505693307391127e-05, "loss": 0.4138, "num_input_tokens_seen": 33087264, "step": 24415 }, { "epoch": 0.7816400998655656, "grad_norm": 1.070896863937378, "learning_rate": 1.6496460549462288e-05, "loss": 0.4121, "num_input_tokens_seen": 33093872, "step": 24420 }, { "epoch": 0.781800140836054, "grad_norm": 1.2583262920379639, "learning_rate": 1.6487229102886097e-05, "loss": 0.6045, "num_input_tokens_seen": 33100752, "step": 24425 }, { "epoch": 0.7819601818065425, "grad_norm": 0.6985873579978943, "learning_rate": 1.6477998969086155e-05, "loss": 0.357, "num_input_tokens_seen": 33107616, "step": 24430 }, { "epoch": 0.782120222777031, "grad_norm": 0.6438146829605103, "learning_rate": 1.646877014948587e-05, "loss": 0.3034, "num_input_tokens_seen": 33114736, "step": 24435 }, { "epoch": 0.7822802637475194, "grad_norm": 2.084343433380127, "learning_rate": 1.6459542645508433e-05, "loss": 0.3904, "num_input_tokens_seen": 33121504, "step": 24440 }, { "epoch": 0.7824403047180079, "grad_norm": 1.8416566848754883, "learning_rate": 1.6450316458576852e-05, "loss": 0.4953, "num_input_tokens_seen": 33127936, "step": 24445 }, { "epoch": 0.7826003456884962, "grad_norm": 0.7675683498382568, "learning_rate": 1.6441091590113912e-05, "loss": 0.2797, "num_input_tokens_seen": 33134688, "step": 24450 }, { "epoch": 0.7827603866589847, "grad_norm": 1.7859748601913452, "learning_rate": 1.6431868041542213e-05, "loss": 0.5712, "num_input_tokens_seen": 33141520, "step": 24455 }, { "epoch": 0.7829204276294731, "grad_norm": 0.7565663456916809, "learning_rate": 1.6422645814284123e-05, "loss": 0.4333, "num_input_tokens_seen": 33148512, "step": 24460 }, { "epoch": 0.7830804685999616, "grad_norm": 2.3434181213378906, "learning_rate": 1.6413424909761846e-05, "loss": 0.3887, "num_input_tokens_seen": 33155232, "step": 24465 }, { "epoch": 0.78324050957045, "grad_norm": 0.6894811987876892, "learning_rate": 1.640420532939736e-05, "loss": 0.3908, "num_input_tokens_seen": 33162032, "step": 24470 }, { "epoch": 0.7834005505409385, "grad_norm": 0.793387234210968, "learning_rate": 1.639498707461242e-05, "loss": 0.3328, "num_input_tokens_seen": 33168800, "step": 24475 }, { "epoch": 0.7835605915114269, "grad_norm": 0.7592688798904419, "learning_rate": 1.6385770146828614e-05, "loss": 0.2111, "num_input_tokens_seen": 33175632, "step": 24480 }, { "epoch": 0.7837206324819154, "grad_norm": 1.4252291917800903, "learning_rate": 1.637655454746731e-05, "loss": 0.541, "num_input_tokens_seen": 33183024, "step": 24485 }, { "epoch": 0.7838806734524039, "grad_norm": 1.7297406196594238, "learning_rate": 1.6367340277949658e-05, "loss": 0.4578, "num_input_tokens_seen": 33190032, "step": 24490 }, { "epoch": 0.7840407144228922, "grad_norm": 0.5748306512832642, "learning_rate": 1.635812733969663e-05, "loss": 0.2865, "num_input_tokens_seen": 33196992, "step": 24495 }, { "epoch": 0.7842007553933807, "grad_norm": 1.7022373676300049, "learning_rate": 1.634891573412896e-05, "loss": 0.4227, "num_input_tokens_seen": 33203584, "step": 24500 }, { "epoch": 0.7843607963638691, "grad_norm": 0.7675620317459106, "learning_rate": 1.6339705462667196e-05, "loss": 0.3346, "num_input_tokens_seen": 33210096, "step": 24505 }, { "epoch": 0.7845208373343576, "grad_norm": 1.4294102191925049, "learning_rate": 1.633049652673169e-05, "loss": 0.622, "num_input_tokens_seen": 33216912, "step": 24510 }, { "epoch": 0.784680878304846, "grad_norm": 0.6076758503913879, "learning_rate": 1.632128892774256e-05, "loss": 0.3186, "num_input_tokens_seen": 33223568, "step": 24515 }, { "epoch": 0.7848409192753345, "grad_norm": 2.3925793170928955, "learning_rate": 1.6312082667119737e-05, "loss": 0.419, "num_input_tokens_seen": 33230272, "step": 24520 }, { "epoch": 0.7850009602458229, "grad_norm": 1.1440761089324951, "learning_rate": 1.630287774628296e-05, "loss": 0.4315, "num_input_tokens_seen": 33236848, "step": 24525 }, { "epoch": 0.7851610012163114, "grad_norm": 0.663750171661377, "learning_rate": 1.6293674166651718e-05, "loss": 0.46, "num_input_tokens_seen": 33243952, "step": 24530 }, { "epoch": 0.7853210421867998, "grad_norm": 0.757514238357544, "learning_rate": 1.6284471929645338e-05, "loss": 0.375, "num_input_tokens_seen": 33250864, "step": 24535 }, { "epoch": 0.7854810831572883, "grad_norm": 0.771578848361969, "learning_rate": 1.627527103668291e-05, "loss": 0.3943, "num_input_tokens_seen": 33257632, "step": 24540 }, { "epoch": 0.7856411241277768, "grad_norm": 1.0102533102035522, "learning_rate": 1.6266071489183327e-05, "loss": 0.355, "num_input_tokens_seen": 33264000, "step": 24545 }, { "epoch": 0.7858011650982651, "grad_norm": 1.1234298944473267, "learning_rate": 1.6256873288565283e-05, "loss": 0.5918, "num_input_tokens_seen": 33270560, "step": 24550 }, { "epoch": 0.7859612060687536, "grad_norm": 0.43409407138824463, "learning_rate": 1.6247676436247245e-05, "loss": 0.2135, "num_input_tokens_seen": 33277104, "step": 24555 }, { "epoch": 0.786121247039242, "grad_norm": 1.6151421070098877, "learning_rate": 1.6238480933647486e-05, "loss": 0.4553, "num_input_tokens_seen": 33283648, "step": 24560 }, { "epoch": 0.7862812880097305, "grad_norm": 1.059715986251831, "learning_rate": 1.6229286782184083e-05, "loss": 0.4547, "num_input_tokens_seen": 33290096, "step": 24565 }, { "epoch": 0.7864413289802189, "grad_norm": 1.3508695363998413, "learning_rate": 1.622009398327487e-05, "loss": 0.4762, "num_input_tokens_seen": 33296528, "step": 24570 }, { "epoch": 0.7866013699507074, "grad_norm": 1.293613314628601, "learning_rate": 1.6210902538337502e-05, "loss": 0.4654, "num_input_tokens_seen": 33303808, "step": 24575 }, { "epoch": 0.7867614109211958, "grad_norm": 0.43149349093437195, "learning_rate": 1.6201712448789413e-05, "loss": 0.5973, "num_input_tokens_seen": 33310224, "step": 24580 }, { "epoch": 0.7869214518916843, "grad_norm": 2.234614610671997, "learning_rate": 1.6192523716047827e-05, "loss": 0.4573, "num_input_tokens_seen": 33316576, "step": 24585 }, { "epoch": 0.7870814928621728, "grad_norm": 1.3993088006973267, "learning_rate": 1.6183336341529776e-05, "loss": 0.3597, "num_input_tokens_seen": 33323152, "step": 24590 }, { "epoch": 0.7872415338326612, "grad_norm": 1.3162837028503418, "learning_rate": 1.6174150326652047e-05, "loss": 0.383, "num_input_tokens_seen": 33329920, "step": 24595 }, { "epoch": 0.7874015748031497, "grad_norm": 0.5054768323898315, "learning_rate": 1.6164965672831256e-05, "loss": 0.3557, "num_input_tokens_seen": 33336480, "step": 24600 }, { "epoch": 0.7874015748031497, "eval_loss": 0.45190489292144775, "eval_runtime": 502.5412, "eval_samples_per_second": 27.632, "eval_steps_per_second": 13.816, "num_input_tokens_seen": 33336480, "step": 24600 }, { "epoch": 0.787561615773638, "grad_norm": 0.6907312273979187, "learning_rate": 1.6155782381483784e-05, "loss": 0.4162, "num_input_tokens_seen": 33343088, "step": 24605 }, { "epoch": 0.7877216567441265, "grad_norm": 1.582440972328186, "learning_rate": 1.6146600454025813e-05, "loss": 0.5562, "num_input_tokens_seen": 33350272, "step": 24610 }, { "epoch": 0.7878816977146149, "grad_norm": 0.9724825024604797, "learning_rate": 1.6137419891873317e-05, "loss": 0.3218, "num_input_tokens_seen": 33357312, "step": 24615 }, { "epoch": 0.7880417386851034, "grad_norm": 1.4050183296203613, "learning_rate": 1.6128240696442038e-05, "loss": 0.3908, "num_input_tokens_seen": 33363760, "step": 24620 }, { "epoch": 0.7882017796555918, "grad_norm": 1.4195295572280884, "learning_rate": 1.611906286914753e-05, "loss": 0.5848, "num_input_tokens_seen": 33370784, "step": 24625 }, { "epoch": 0.7883618206260803, "grad_norm": 1.0658535957336426, "learning_rate": 1.6109886411405144e-05, "loss": 0.2439, "num_input_tokens_seen": 33377728, "step": 24630 }, { "epoch": 0.7885218615965687, "grad_norm": 1.3852101564407349, "learning_rate": 1.6100711324629985e-05, "loss": 0.3595, "num_input_tokens_seen": 33384560, "step": 24635 }, { "epoch": 0.7886819025670572, "grad_norm": 0.841738760471344, "learning_rate": 1.609153761023698e-05, "loss": 0.4167, "num_input_tokens_seen": 33391264, "step": 24640 }, { "epoch": 0.7888419435375457, "grad_norm": 0.6824131011962891, "learning_rate": 1.608236526964083e-05, "loss": 0.4474, "num_input_tokens_seen": 33397856, "step": 24645 }, { "epoch": 0.789001984508034, "grad_norm": 0.9834127426147461, "learning_rate": 1.607319430425601e-05, "loss": 0.4866, "num_input_tokens_seen": 33404864, "step": 24650 }, { "epoch": 0.7891620254785225, "grad_norm": 0.6939612627029419, "learning_rate": 1.606402471549682e-05, "loss": 0.3421, "num_input_tokens_seen": 33411712, "step": 24655 }, { "epoch": 0.7893220664490109, "grad_norm": 2.1135621070861816, "learning_rate": 1.6054856504777312e-05, "loss": 0.5326, "num_input_tokens_seen": 33418320, "step": 24660 }, { "epoch": 0.7894821074194994, "grad_norm": 1.0963897705078125, "learning_rate": 1.6045689673511334e-05, "loss": 0.3985, "num_input_tokens_seen": 33424752, "step": 24665 }, { "epoch": 0.7896421483899878, "grad_norm": 1.2745198011398315, "learning_rate": 1.6036524223112548e-05, "loss": 0.6072, "num_input_tokens_seen": 33431392, "step": 24670 }, { "epoch": 0.7898021893604763, "grad_norm": 0.40869495272636414, "learning_rate": 1.602736015499436e-05, "loss": 0.3875, "num_input_tokens_seen": 33438096, "step": 24675 }, { "epoch": 0.7899622303309647, "grad_norm": 0.4046911299228668, "learning_rate": 1.601819747057e-05, "loss": 0.2919, "num_input_tokens_seen": 33444608, "step": 24680 }, { "epoch": 0.7901222713014532, "grad_norm": 2.670774459838867, "learning_rate": 1.6009036171252465e-05, "loss": 0.3333, "num_input_tokens_seen": 33451168, "step": 24685 }, { "epoch": 0.7902823122719416, "grad_norm": 0.7028703093528748, "learning_rate": 1.599987625845453e-05, "loss": 0.2893, "num_input_tokens_seen": 33457696, "step": 24690 }, { "epoch": 0.7904423532424301, "grad_norm": 1.3665441274642944, "learning_rate": 1.599071773358879e-05, "loss": 0.5799, "num_input_tokens_seen": 33464768, "step": 24695 }, { "epoch": 0.7906023942129186, "grad_norm": 0.445985347032547, "learning_rate": 1.598156059806758e-05, "loss": 0.2972, "num_input_tokens_seen": 33471856, "step": 24700 }, { "epoch": 0.7907624351834069, "grad_norm": 0.8753498792648315, "learning_rate": 1.5972404853303062e-05, "loss": 0.3777, "num_input_tokens_seen": 33478512, "step": 24705 }, { "epoch": 0.7909224761538954, "grad_norm": 1.004380226135254, "learning_rate": 1.5963250500707172e-05, "loss": 0.462, "num_input_tokens_seen": 33485120, "step": 24710 }, { "epoch": 0.7910825171243838, "grad_norm": 0.6628986597061157, "learning_rate": 1.5954097541691612e-05, "loss": 0.4435, "num_input_tokens_seen": 33491856, "step": 24715 }, { "epoch": 0.7912425580948723, "grad_norm": 1.261520504951477, "learning_rate": 1.5944945977667884e-05, "loss": 0.4947, "num_input_tokens_seen": 33498416, "step": 24720 }, { "epoch": 0.7914025990653607, "grad_norm": 0.47043874859809875, "learning_rate": 1.593579581004729e-05, "loss": 0.4326, "num_input_tokens_seen": 33505360, "step": 24725 }, { "epoch": 0.7915626400358492, "grad_norm": 1.978331208229065, "learning_rate": 1.592664704024088e-05, "loss": 0.457, "num_input_tokens_seen": 33511680, "step": 24730 }, { "epoch": 0.7917226810063376, "grad_norm": 0.4698430895805359, "learning_rate": 1.591749966965953e-05, "loss": 0.5029, "num_input_tokens_seen": 33518624, "step": 24735 }, { "epoch": 0.7918827219768261, "grad_norm": 3.588357925415039, "learning_rate": 1.5908353699713856e-05, "loss": 0.5554, "num_input_tokens_seen": 33525536, "step": 24740 }, { "epoch": 0.7920427629473145, "grad_norm": 1.280158281326294, "learning_rate": 1.5899209131814298e-05, "loss": 0.517, "num_input_tokens_seen": 33532272, "step": 24745 }, { "epoch": 0.792202803917803, "grad_norm": 1.026283860206604, "learning_rate": 1.5890065967371067e-05, "loss": 0.3654, "num_input_tokens_seen": 33539152, "step": 24750 }, { "epoch": 0.7923628448882915, "grad_norm": 1.7357436418533325, "learning_rate": 1.5880924207794144e-05, "loss": 0.3984, "num_input_tokens_seen": 33546048, "step": 24755 }, { "epoch": 0.7925228858587798, "grad_norm": 0.8547391295433044, "learning_rate": 1.5871783854493298e-05, "loss": 0.3725, "num_input_tokens_seen": 33552624, "step": 24760 }, { "epoch": 0.7926829268292683, "grad_norm": 0.7671740055084229, "learning_rate": 1.5862644908878106e-05, "loss": 0.4132, "num_input_tokens_seen": 33559472, "step": 24765 }, { "epoch": 0.7928429677997567, "grad_norm": 1.758953332901001, "learning_rate": 1.5853507372357885e-05, "loss": 0.5557, "num_input_tokens_seen": 33566448, "step": 24770 }, { "epoch": 0.7930030087702452, "grad_norm": 1.2154065370559692, "learning_rate": 1.5844371246341776e-05, "loss": 0.5125, "num_input_tokens_seen": 33572816, "step": 24775 }, { "epoch": 0.7931630497407336, "grad_norm": 0.6301549077033997, "learning_rate": 1.5835236532238674e-05, "loss": 0.4166, "num_input_tokens_seen": 33579344, "step": 24780 }, { "epoch": 0.7933230907112221, "grad_norm": 0.676328182220459, "learning_rate": 1.582610323145727e-05, "loss": 0.3409, "num_input_tokens_seen": 33585840, "step": 24785 }, { "epoch": 0.7934831316817105, "grad_norm": 0.8341811299324036, "learning_rate": 1.5816971345406035e-05, "loss": 0.382, "num_input_tokens_seen": 33592528, "step": 24790 }, { "epoch": 0.793643172652199, "grad_norm": 0.6919166445732117, "learning_rate": 1.5807840875493225e-05, "loss": 0.4481, "num_input_tokens_seen": 33599488, "step": 24795 }, { "epoch": 0.7938032136226874, "grad_norm": 0.9276781678199768, "learning_rate": 1.5798711823126854e-05, "loss": 0.4961, "num_input_tokens_seen": 33606096, "step": 24800 }, { "epoch": 0.7938032136226874, "eval_loss": 0.4517000913619995, "eval_runtime": 502.869, "eval_samples_per_second": 27.614, "eval_steps_per_second": 13.807, "num_input_tokens_seen": 33606096, "step": 24800 }, { "epoch": 0.7939632545931758, "grad_norm": 0.8533514738082886, "learning_rate": 1.578958418971477e-05, "loss": 0.3807, "num_input_tokens_seen": 33612928, "step": 24805 }, { "epoch": 0.7941232955636643, "grad_norm": 0.9207338690757751, "learning_rate": 1.578045797666453e-05, "loss": 0.3269, "num_input_tokens_seen": 33619296, "step": 24810 }, { "epoch": 0.7942833365341527, "grad_norm": 1.561726689338684, "learning_rate": 1.5771333185383548e-05, "loss": 0.5968, "num_input_tokens_seen": 33625792, "step": 24815 }, { "epoch": 0.7944433775046412, "grad_norm": 0.6948306560516357, "learning_rate": 1.576220981727895e-05, "loss": 0.4285, "num_input_tokens_seen": 33632784, "step": 24820 }, { "epoch": 0.7946034184751296, "grad_norm": 1.2157717943191528, "learning_rate": 1.575308787375769e-05, "loss": 0.5617, "num_input_tokens_seen": 33639520, "step": 24825 }, { "epoch": 0.7947634594456181, "grad_norm": 1.3724883794784546, "learning_rate": 1.5743967356226492e-05, "loss": 0.606, "num_input_tokens_seen": 33646640, "step": 24830 }, { "epoch": 0.7949235004161065, "grad_norm": 1.5864043235778809, "learning_rate": 1.5734848266091835e-05, "loss": 0.6472, "num_input_tokens_seen": 33653440, "step": 24835 }, { "epoch": 0.795083541386595, "grad_norm": 1.030842661857605, "learning_rate": 1.572573060476001e-05, "loss": 0.2648, "num_input_tokens_seen": 33660240, "step": 24840 }, { "epoch": 0.7952435823570834, "grad_norm": 0.5789699554443359, "learning_rate": 1.5716614373637085e-05, "loss": 0.4093, "num_input_tokens_seen": 33667168, "step": 24845 }, { "epoch": 0.7954036233275719, "grad_norm": 0.5810915231704712, "learning_rate": 1.570749957412887e-05, "loss": 0.4423, "num_input_tokens_seen": 33674064, "step": 24850 }, { "epoch": 0.7955636642980604, "grad_norm": 1.8907854557037354, "learning_rate": 1.5698386207641013e-05, "loss": 0.3915, "num_input_tokens_seen": 33680544, "step": 24855 }, { "epoch": 0.7957237052685487, "grad_norm": 2.0503482818603516, "learning_rate": 1.5689274275578884e-05, "loss": 0.3849, "num_input_tokens_seen": 33687616, "step": 24860 }, { "epoch": 0.7958837462390372, "grad_norm": 1.2882179021835327, "learning_rate": 1.5680163779347667e-05, "loss": 0.3726, "num_input_tokens_seen": 33694608, "step": 24865 }, { "epoch": 0.7960437872095256, "grad_norm": 2.1665263175964355, "learning_rate": 1.5671054720352327e-05, "loss": 0.5146, "num_input_tokens_seen": 33701536, "step": 24870 }, { "epoch": 0.7962038281800141, "grad_norm": 2.176478147506714, "learning_rate": 1.566194709999757e-05, "loss": 0.3971, "num_input_tokens_seen": 33708064, "step": 24875 }, { "epoch": 0.7963638691505025, "grad_norm": 0.8492569923400879, "learning_rate": 1.5652840919687933e-05, "loss": 0.4196, "num_input_tokens_seen": 33714736, "step": 24880 }, { "epoch": 0.796523910120991, "grad_norm": 1.171231746673584, "learning_rate": 1.5643736180827676e-05, "loss": 0.4137, "num_input_tokens_seen": 33721536, "step": 24885 }, { "epoch": 0.7966839510914794, "grad_norm": 0.7169893980026245, "learning_rate": 1.5634632884820878e-05, "loss": 0.4901, "num_input_tokens_seen": 33728576, "step": 24890 }, { "epoch": 0.7968439920619679, "grad_norm": 1.2162046432495117, "learning_rate": 1.5625531033071395e-05, "loss": 0.3577, "num_input_tokens_seen": 33735248, "step": 24895 }, { "epoch": 0.7970040330324563, "grad_norm": 0.6884238123893738, "learning_rate": 1.5616430626982828e-05, "loss": 0.6337, "num_input_tokens_seen": 33741808, "step": 24900 }, { "epoch": 0.7971640740029448, "grad_norm": 1.5811727046966553, "learning_rate": 1.5607331667958575e-05, "loss": 0.421, "num_input_tokens_seen": 33748880, "step": 24905 }, { "epoch": 0.7973241149734333, "grad_norm": 0.5580316185951233, "learning_rate": 1.5598234157401824e-05, "loss": 0.3448, "num_input_tokens_seen": 33755216, "step": 24910 }, { "epoch": 0.7974841559439216, "grad_norm": 1.6347506046295166, "learning_rate": 1.5589138096715503e-05, "loss": 0.5195, "num_input_tokens_seen": 33762048, "step": 24915 }, { "epoch": 0.7976441969144101, "grad_norm": 0.8157460689544678, "learning_rate": 1.5580043487302365e-05, "loss": 0.3454, "num_input_tokens_seen": 33768896, "step": 24920 }, { "epoch": 0.7978042378848985, "grad_norm": 1.295173168182373, "learning_rate": 1.5570950330564888e-05, "loss": 0.55, "num_input_tokens_seen": 33775760, "step": 24925 }, { "epoch": 0.797964278855387, "grad_norm": 1.4716787338256836, "learning_rate": 1.5561858627905367e-05, "loss": 0.4255, "num_input_tokens_seen": 33782880, "step": 24930 }, { "epoch": 0.7981243198258754, "grad_norm": 0.924547553062439, "learning_rate": 1.5552768380725857e-05, "loss": 0.377, "num_input_tokens_seen": 33789904, "step": 24935 }, { "epoch": 0.7982843607963639, "grad_norm": 0.8663755655288696, "learning_rate": 1.5543679590428183e-05, "loss": 0.4161, "num_input_tokens_seen": 33797664, "step": 24940 }, { "epoch": 0.7984444017668523, "grad_norm": 1.3619253635406494, "learning_rate": 1.5534592258413943e-05, "loss": 0.4477, "num_input_tokens_seen": 33804240, "step": 24945 }, { "epoch": 0.7986044427373408, "grad_norm": 0.6744622588157654, "learning_rate": 1.5525506386084538e-05, "loss": 0.267, "num_input_tokens_seen": 33810976, "step": 24950 }, { "epoch": 0.7987644837078292, "grad_norm": 1.226279616355896, "learning_rate": 1.55164219748411e-05, "loss": 0.437, "num_input_tokens_seen": 33817680, "step": 24955 }, { "epoch": 0.7989245246783176, "grad_norm": 0.6927685141563416, "learning_rate": 1.550733902608459e-05, "loss": 0.3513, "num_input_tokens_seen": 33824432, "step": 24960 }, { "epoch": 0.7990845656488061, "grad_norm": 1.0618040561676025, "learning_rate": 1.549825754121568e-05, "loss": 0.2817, "num_input_tokens_seen": 33831296, "step": 24965 }, { "epoch": 0.7992446066192945, "grad_norm": 0.5821548700332642, "learning_rate": 1.5489177521634864e-05, "loss": 0.3894, "num_input_tokens_seen": 33838096, "step": 24970 }, { "epoch": 0.799404647589783, "grad_norm": 0.6881456971168518, "learning_rate": 1.5480098968742402e-05, "loss": 0.3845, "num_input_tokens_seen": 33845056, "step": 24975 }, { "epoch": 0.7995646885602714, "grad_norm": 1.065528154373169, "learning_rate": 1.5471021883938304e-05, "loss": 0.347, "num_input_tokens_seen": 33852064, "step": 24980 }, { "epoch": 0.7997247295307599, "grad_norm": 0.6455118060112, "learning_rate": 1.546194626862238e-05, "loss": 0.4439, "num_input_tokens_seen": 33858608, "step": 24985 }, { "epoch": 0.7998847705012483, "grad_norm": 0.9154494404792786, "learning_rate": 1.5452872124194216e-05, "loss": 0.4891, "num_input_tokens_seen": 33865376, "step": 24990 }, { "epoch": 0.8000448114717368, "grad_norm": 1.035416603088379, "learning_rate": 1.5443799452053136e-05, "loss": 0.469, "num_input_tokens_seen": 33872064, "step": 24995 }, { "epoch": 0.8002048524422252, "grad_norm": 1.1223764419555664, "learning_rate": 1.543472825359828e-05, "loss": 0.6283, "num_input_tokens_seen": 33878720, "step": 25000 }, { "epoch": 0.8002048524422252, "eval_loss": 0.4515012204647064, "eval_runtime": 502.8843, "eval_samples_per_second": 27.613, "eval_steps_per_second": 13.806, "num_input_tokens_seen": 33878720, "step": 25000 }, { "epoch": 0.8003648934127137, "grad_norm": 1.0212681293487549, "learning_rate": 1.5425658530228522e-05, "loss": 0.4375, "num_input_tokens_seen": 33885824, "step": 25005 }, { "epoch": 0.800524934383202, "grad_norm": 0.8597103357315063, "learning_rate": 1.5416590283342546e-05, "loss": 0.3299, "num_input_tokens_seen": 33892576, "step": 25010 }, { "epoch": 0.8006849753536905, "grad_norm": 0.7688319683074951, "learning_rate": 1.5407523514338783e-05, "loss": 0.3506, "num_input_tokens_seen": 33899344, "step": 25015 }, { "epoch": 0.800845016324179, "grad_norm": 0.6656736135482788, "learning_rate": 1.539845822461543e-05, "loss": 0.2972, "num_input_tokens_seen": 33906144, "step": 25020 }, { "epoch": 0.8010050572946674, "grad_norm": 0.8438025116920471, "learning_rate": 1.538939441557048e-05, "loss": 0.6197, "num_input_tokens_seen": 33912768, "step": 25025 }, { "epoch": 0.8011650982651559, "grad_norm": 1.1721946001052856, "learning_rate": 1.5380332088601696e-05, "loss": 0.3326, "num_input_tokens_seen": 33919936, "step": 25030 }, { "epoch": 0.8013251392356443, "grad_norm": 1.2966264486312866, "learning_rate": 1.537127124510658e-05, "loss": 0.3966, "num_input_tokens_seen": 33926672, "step": 25035 }, { "epoch": 0.8014851802061328, "grad_norm": 0.8619105815887451, "learning_rate": 1.5362211886482457e-05, "loss": 0.3309, "num_input_tokens_seen": 33933488, "step": 25040 }, { "epoch": 0.8016452211766212, "grad_norm": 1.2793126106262207, "learning_rate": 1.5353154014126363e-05, "loss": 0.3328, "num_input_tokens_seen": 33940256, "step": 25045 }, { "epoch": 0.8018052621471097, "grad_norm": 1.221885085105896, "learning_rate": 1.534409762943515e-05, "loss": 0.3963, "num_input_tokens_seen": 33946656, "step": 25050 }, { "epoch": 0.8019653031175981, "grad_norm": 0.9724823832511902, "learning_rate": 1.5335042733805438e-05, "loss": 0.4119, "num_input_tokens_seen": 33953120, "step": 25055 }, { "epoch": 0.8021253440880866, "grad_norm": 1.1590347290039062, "learning_rate": 1.532598932863358e-05, "loss": 0.3856, "num_input_tokens_seen": 33959968, "step": 25060 }, { "epoch": 0.8022853850585749, "grad_norm": 1.824790358543396, "learning_rate": 1.531693741531574e-05, "loss": 0.4712, "num_input_tokens_seen": 33967248, "step": 25065 }, { "epoch": 0.8024454260290634, "grad_norm": 1.2772191762924194, "learning_rate": 1.5307886995247844e-05, "loss": 0.3231, "num_input_tokens_seen": 33973664, "step": 25070 }, { "epoch": 0.8026054669995519, "grad_norm": 0.46105825901031494, "learning_rate": 1.529883806982557e-05, "loss": 0.3558, "num_input_tokens_seen": 33980032, "step": 25075 }, { "epoch": 0.8027655079700403, "grad_norm": 0.9439039826393127, "learning_rate": 1.5289790640444376e-05, "loss": 0.345, "num_input_tokens_seen": 33986736, "step": 25080 }, { "epoch": 0.8029255489405288, "grad_norm": 0.3847595751285553, "learning_rate": 1.5280744708499494e-05, "loss": 0.454, "num_input_tokens_seen": 33993584, "step": 25085 }, { "epoch": 0.8030855899110172, "grad_norm": 0.534122109413147, "learning_rate": 1.527170027538591e-05, "loss": 0.4548, "num_input_tokens_seen": 34000368, "step": 25090 }, { "epoch": 0.8032456308815057, "grad_norm": 0.7970086336135864, "learning_rate": 1.5262657342498407e-05, "loss": 0.4297, "num_input_tokens_seen": 34006752, "step": 25095 }, { "epoch": 0.8034056718519941, "grad_norm": 0.8312856554985046, "learning_rate": 1.52536159112315e-05, "loss": 0.4005, "num_input_tokens_seen": 34013648, "step": 25100 }, { "epoch": 0.8035657128224826, "grad_norm": 1.7777254581451416, "learning_rate": 1.5244575982979497e-05, "loss": 0.5754, "num_input_tokens_seen": 34020080, "step": 25105 }, { "epoch": 0.803725753792971, "grad_norm": 1.5141980648040771, "learning_rate": 1.5235537559136487e-05, "loss": 0.5675, "num_input_tokens_seen": 34026992, "step": 25110 }, { "epoch": 0.8038857947634594, "grad_norm": 0.3669368624687195, "learning_rate": 1.5226500641096286e-05, "loss": 0.3774, "num_input_tokens_seen": 34033552, "step": 25115 }, { "epoch": 0.804045835733948, "grad_norm": 1.8644657135009766, "learning_rate": 1.5217465230252509e-05, "loss": 0.5331, "num_input_tokens_seen": 34040096, "step": 25120 }, { "epoch": 0.8042058767044363, "grad_norm": 1.1668742895126343, "learning_rate": 1.5208431327998523e-05, "loss": 0.5158, "num_input_tokens_seen": 34046896, "step": 25125 }, { "epoch": 0.8043659176749248, "grad_norm": 2.909109592437744, "learning_rate": 1.5199398935727477e-05, "loss": 0.4948, "num_input_tokens_seen": 34053728, "step": 25130 }, { "epoch": 0.8045259586454132, "grad_norm": 1.3403364419937134, "learning_rate": 1.5190368054832282e-05, "loss": 0.4312, "num_input_tokens_seen": 34060144, "step": 25135 }, { "epoch": 0.8046859996159017, "grad_norm": 1.800445556640625, "learning_rate": 1.5181338686705601e-05, "loss": 0.4248, "num_input_tokens_seen": 34067120, "step": 25140 }, { "epoch": 0.8048460405863901, "grad_norm": 0.511084258556366, "learning_rate": 1.5172310832739889e-05, "loss": 0.4356, "num_input_tokens_seen": 34073920, "step": 25145 }, { "epoch": 0.8050060815568786, "grad_norm": 1.9993101358413696, "learning_rate": 1.5163284494327346e-05, "loss": 0.397, "num_input_tokens_seen": 34080576, "step": 25150 }, { "epoch": 0.805166122527367, "grad_norm": 2.0841872692108154, "learning_rate": 1.5154259672859952e-05, "loss": 0.6908, "num_input_tokens_seen": 34087792, "step": 25155 }, { "epoch": 0.8053261634978555, "grad_norm": 1.0753036737442017, "learning_rate": 1.5145236369729452e-05, "loss": 0.4131, "num_input_tokens_seen": 34094512, "step": 25160 }, { "epoch": 0.8054862044683438, "grad_norm": 0.6777392029762268, "learning_rate": 1.5136214586327335e-05, "loss": 0.3037, "num_input_tokens_seen": 34101344, "step": 25165 }, { "epoch": 0.8056462454388323, "grad_norm": 1.126760482788086, "learning_rate": 1.5127194324044885e-05, "loss": 0.4037, "num_input_tokens_seen": 34108304, "step": 25170 }, { "epoch": 0.8058062864093208, "grad_norm": 1.9543229341506958, "learning_rate": 1.5118175584273148e-05, "loss": 0.4753, "num_input_tokens_seen": 34115136, "step": 25175 }, { "epoch": 0.8059663273798092, "grad_norm": 1.569501280784607, "learning_rate": 1.5109158368402909e-05, "loss": 0.4004, "num_input_tokens_seen": 34121744, "step": 25180 }, { "epoch": 0.8061263683502977, "grad_norm": 1.06857430934906, "learning_rate": 1.5100142677824753e-05, "loss": 0.4461, "num_input_tokens_seen": 34128400, "step": 25185 }, { "epoch": 0.8062864093207861, "grad_norm": 1.6837189197540283, "learning_rate": 1.509112851392901e-05, "loss": 0.3861, "num_input_tokens_seen": 34134880, "step": 25190 }, { "epoch": 0.8064464502912746, "grad_norm": 0.48088565468788147, "learning_rate": 1.5082115878105763e-05, "loss": 0.3039, "num_input_tokens_seen": 34141504, "step": 25195 }, { "epoch": 0.806606491261763, "grad_norm": 1.0038210153579712, "learning_rate": 1.5073104771744892e-05, "loss": 0.3892, "num_input_tokens_seen": 34148496, "step": 25200 }, { "epoch": 0.806606491261763, "eval_loss": 0.4512828588485718, "eval_runtime": 503.817, "eval_samples_per_second": 27.562, "eval_steps_per_second": 13.781, "num_input_tokens_seen": 34148496, "step": 25200 }, { "epoch": 0.8067665322322515, "grad_norm": 0.7792969942092896, "learning_rate": 1.5064095196236006e-05, "loss": 0.3022, "num_input_tokens_seen": 34155136, "step": 25205 }, { "epoch": 0.8069265732027399, "grad_norm": 0.8890697956085205, "learning_rate": 1.50550871529685e-05, "loss": 0.3188, "num_input_tokens_seen": 34162016, "step": 25210 }, { "epoch": 0.8070866141732284, "grad_norm": 0.9759238958358765, "learning_rate": 1.5046080643331546e-05, "loss": 0.4488, "num_input_tokens_seen": 34169072, "step": 25215 }, { "epoch": 0.8072466551437167, "grad_norm": 1.314168930053711, "learning_rate": 1.5037075668714028e-05, "loss": 0.463, "num_input_tokens_seen": 34176176, "step": 25220 }, { "epoch": 0.8074066961142052, "grad_norm": 1.2123035192489624, "learning_rate": 1.5028072230504656e-05, "loss": 0.4848, "num_input_tokens_seen": 34182736, "step": 25225 }, { "epoch": 0.8075667370846937, "grad_norm": 0.973718523979187, "learning_rate": 1.5019070330091861e-05, "loss": 0.3398, "num_input_tokens_seen": 34189536, "step": 25230 }, { "epoch": 0.8077267780551821, "grad_norm": 0.9337305426597595, "learning_rate": 1.5010069968863843e-05, "loss": 0.5278, "num_input_tokens_seen": 34196400, "step": 25235 }, { "epoch": 0.8078868190256706, "grad_norm": 0.9303128123283386, "learning_rate": 1.5001071148208584e-05, "loss": 0.3175, "num_input_tokens_seen": 34203056, "step": 25240 }, { "epoch": 0.808046859996159, "grad_norm": 2.2519474029541016, "learning_rate": 1.49920738695138e-05, "loss": 0.415, "num_input_tokens_seen": 34210112, "step": 25245 }, { "epoch": 0.8082069009666475, "grad_norm": 1.220367193222046, "learning_rate": 1.4983078134166995e-05, "loss": 0.3754, "num_input_tokens_seen": 34216704, "step": 25250 }, { "epoch": 0.8083669419371359, "grad_norm": 0.9004956483840942, "learning_rate": 1.4974083943555428e-05, "loss": 0.3544, "num_input_tokens_seen": 34223376, "step": 25255 }, { "epoch": 0.8085269829076244, "grad_norm": 1.078702449798584, "learning_rate": 1.496509129906611e-05, "loss": 0.5219, "num_input_tokens_seen": 34230416, "step": 25260 }, { "epoch": 0.8086870238781128, "grad_norm": 1.5573173761367798, "learning_rate": 1.4956100202085809e-05, "loss": 0.4841, "num_input_tokens_seen": 34237072, "step": 25265 }, { "epoch": 0.8088470648486012, "grad_norm": 0.9188011884689331, "learning_rate": 1.4947110654001093e-05, "loss": 0.4378, "num_input_tokens_seen": 34243696, "step": 25270 }, { "epoch": 0.8090071058190896, "grad_norm": 1.7737702131271362, "learning_rate": 1.4938122656198234e-05, "loss": 0.4114, "num_input_tokens_seen": 34251296, "step": 25275 }, { "epoch": 0.8091671467895781, "grad_norm": 1.1937108039855957, "learning_rate": 1.4929136210063316e-05, "loss": 0.4229, "num_input_tokens_seen": 34258320, "step": 25280 }, { "epoch": 0.8093271877600666, "grad_norm": 0.6855153441429138, "learning_rate": 1.4920151316982146e-05, "loss": 0.2922, "num_input_tokens_seen": 34265152, "step": 25285 }, { "epoch": 0.809487228730555, "grad_norm": 1.4897502660751343, "learning_rate": 1.4911167978340312e-05, "loss": 0.5041, "num_input_tokens_seen": 34272096, "step": 25290 }, { "epoch": 0.8096472697010435, "grad_norm": 1.0495408773422241, "learning_rate": 1.4902186195523166e-05, "loss": 0.4783, "num_input_tokens_seen": 34278464, "step": 25295 }, { "epoch": 0.8098073106715319, "grad_norm": 1.397287130355835, "learning_rate": 1.4893205969915805e-05, "loss": 0.3246, "num_input_tokens_seen": 34284976, "step": 25300 }, { "epoch": 0.8099673516420204, "grad_norm": 1.6649020910263062, "learning_rate": 1.4884227302903086e-05, "loss": 0.385, "num_input_tokens_seen": 34291440, "step": 25305 }, { "epoch": 0.8101273926125088, "grad_norm": 1.1406660079956055, "learning_rate": 1.4875250195869653e-05, "loss": 0.4663, "num_input_tokens_seen": 34298192, "step": 25310 }, { "epoch": 0.8102874335829973, "grad_norm": 1.444870948791504, "learning_rate": 1.4866274650199862e-05, "loss": 0.3829, "num_input_tokens_seen": 34304800, "step": 25315 }, { "epoch": 0.8104474745534856, "grad_norm": 1.0164146423339844, "learning_rate": 1.485730066727788e-05, "loss": 0.5124, "num_input_tokens_seen": 34311424, "step": 25320 }, { "epoch": 0.8106075155239741, "grad_norm": 1.2866088151931763, "learning_rate": 1.4848328248487586e-05, "loss": 0.6626, "num_input_tokens_seen": 34317984, "step": 25325 }, { "epoch": 0.8107675564944626, "grad_norm": 1.1262321472167969, "learning_rate": 1.4839357395212656e-05, "loss": 0.5936, "num_input_tokens_seen": 34324896, "step": 25330 }, { "epoch": 0.810927597464951, "grad_norm": 1.425831913948059, "learning_rate": 1.4830388108836502e-05, "loss": 0.518, "num_input_tokens_seen": 34331888, "step": 25335 }, { "epoch": 0.8110876384354395, "grad_norm": 0.9509392380714417, "learning_rate": 1.4821420390742299e-05, "loss": 0.635, "num_input_tokens_seen": 34338832, "step": 25340 }, { "epoch": 0.8112476794059279, "grad_norm": 1.0634727478027344, "learning_rate": 1.4812454242312979e-05, "loss": 0.408, "num_input_tokens_seen": 34345472, "step": 25345 }, { "epoch": 0.8114077203764164, "grad_norm": 1.05191969871521, "learning_rate": 1.4803489664931253e-05, "loss": 0.3974, "num_input_tokens_seen": 34352192, "step": 25350 }, { "epoch": 0.8115677613469048, "grad_norm": 1.2794851064682007, "learning_rate": 1.4794526659979544e-05, "loss": 0.3834, "num_input_tokens_seen": 34359136, "step": 25355 }, { "epoch": 0.8117278023173933, "grad_norm": 0.9629953503608704, "learning_rate": 1.4785565228840086e-05, "loss": 0.3407, "num_input_tokens_seen": 34366192, "step": 25360 }, { "epoch": 0.8118878432878817, "grad_norm": 0.764295220375061, "learning_rate": 1.4776605372894819e-05, "loss": 0.468, "num_input_tokens_seen": 34373280, "step": 25365 }, { "epoch": 0.8120478842583702, "grad_norm": 3.195910930633545, "learning_rate": 1.4767647093525488e-05, "loss": 0.3637, "num_input_tokens_seen": 34379824, "step": 25370 }, { "epoch": 0.8122079252288585, "grad_norm": 1.7660589218139648, "learning_rate": 1.4758690392113566e-05, "loss": 0.3391, "num_input_tokens_seen": 34387104, "step": 25375 }, { "epoch": 0.812367966199347, "grad_norm": 0.6352086067199707, "learning_rate": 1.4749735270040276e-05, "loss": 0.4961, "num_input_tokens_seen": 34393872, "step": 25380 }, { "epoch": 0.8125280071698355, "grad_norm": 0.922492265701294, "learning_rate": 1.4740781728686623e-05, "loss": 0.309, "num_input_tokens_seen": 34400384, "step": 25385 }, { "epoch": 0.8126880481403239, "grad_norm": 0.6925721764564514, "learning_rate": 1.4731829769433358e-05, "loss": 0.3653, "num_input_tokens_seen": 34407392, "step": 25390 }, { "epoch": 0.8128480891108124, "grad_norm": 0.8918465375900269, "learning_rate": 1.4722879393660976e-05, "loss": 0.6019, "num_input_tokens_seen": 34414240, "step": 25395 }, { "epoch": 0.8130081300813008, "grad_norm": 1.094306230545044, "learning_rate": 1.4713930602749748e-05, "loss": 0.4803, "num_input_tokens_seen": 34421392, "step": 25400 }, { "epoch": 0.8130081300813008, "eval_loss": 0.45179006457328796, "eval_runtime": 503.1245, "eval_samples_per_second": 27.6, "eval_steps_per_second": 13.8, "num_input_tokens_seen": 34421392, "step": 25400 }, { "epoch": 0.8131681710517893, "grad_norm": 1.0160390138626099, "learning_rate": 1.470498339807968e-05, "loss": 0.5386, "num_input_tokens_seen": 34428288, "step": 25405 }, { "epoch": 0.8133282120222777, "grad_norm": 1.2024755477905273, "learning_rate": 1.4696037781030542e-05, "loss": 0.4886, "num_input_tokens_seen": 34435104, "step": 25410 }, { "epoch": 0.8134882529927662, "grad_norm": 1.9684914350509644, "learning_rate": 1.4687093752981876e-05, "loss": 0.5097, "num_input_tokens_seen": 34441968, "step": 25415 }, { "epoch": 0.8136482939632546, "grad_norm": 0.8266091346740723, "learning_rate": 1.4678151315312943e-05, "loss": 0.3582, "num_input_tokens_seen": 34448768, "step": 25420 }, { "epoch": 0.813808334933743, "grad_norm": 1.6699708700180054, "learning_rate": 1.4669210469402789e-05, "loss": 0.5764, "num_input_tokens_seen": 34455584, "step": 25425 }, { "epoch": 0.8139683759042314, "grad_norm": 2.199354887008667, "learning_rate": 1.4660271216630218e-05, "loss": 0.4797, "num_input_tokens_seen": 34462880, "step": 25430 }, { "epoch": 0.8141284168747199, "grad_norm": 1.1717474460601807, "learning_rate": 1.4651333558373748e-05, "loss": 0.5035, "num_input_tokens_seen": 34469744, "step": 25435 }, { "epoch": 0.8142884578452084, "grad_norm": 0.5555474162101746, "learning_rate": 1.4642397496011707e-05, "loss": 0.4417, "num_input_tokens_seen": 34476592, "step": 25440 }, { "epoch": 0.8144484988156968, "grad_norm": 1.585731029510498, "learning_rate": 1.4633463030922129e-05, "loss": 0.5954, "num_input_tokens_seen": 34483232, "step": 25445 }, { "epoch": 0.8146085397861853, "grad_norm": 0.8087971806526184, "learning_rate": 1.462453016448282e-05, "loss": 0.3891, "num_input_tokens_seen": 34490176, "step": 25450 }, { "epoch": 0.8147685807566737, "grad_norm": 0.6044691205024719, "learning_rate": 1.4615598898071354e-05, "loss": 0.5227, "num_input_tokens_seen": 34496960, "step": 25455 }, { "epoch": 0.8149286217271622, "grad_norm": 0.6104463934898376, "learning_rate": 1.4606669233065026e-05, "loss": 0.3531, "num_input_tokens_seen": 34503824, "step": 25460 }, { "epoch": 0.8150886626976506, "grad_norm": 1.2554875612258911, "learning_rate": 1.4597741170840914e-05, "loss": 0.3547, "num_input_tokens_seen": 34510480, "step": 25465 }, { "epoch": 0.8152487036681391, "grad_norm": 1.154374361038208, "learning_rate": 1.4588814712775853e-05, "loss": 0.5036, "num_input_tokens_seen": 34517344, "step": 25470 }, { "epoch": 0.8154087446386274, "grad_norm": 1.2083964347839355, "learning_rate": 1.4579889860246382e-05, "loss": 0.401, "num_input_tokens_seen": 34524112, "step": 25475 }, { "epoch": 0.8155687856091159, "grad_norm": 1.8208110332489014, "learning_rate": 1.457096661462885e-05, "loss": 0.403, "num_input_tokens_seen": 34530480, "step": 25480 }, { "epoch": 0.8157288265796043, "grad_norm": 0.4479103982448578, "learning_rate": 1.4562044977299322e-05, "loss": 0.6954, "num_input_tokens_seen": 34537152, "step": 25485 }, { "epoch": 0.8158888675500928, "grad_norm": 0.7658055424690247, "learning_rate": 1.4553124949633623e-05, "loss": 0.5174, "num_input_tokens_seen": 34544016, "step": 25490 }, { "epoch": 0.8160489085205813, "grad_norm": 1.2255648374557495, "learning_rate": 1.4544206533007354e-05, "loss": 0.589, "num_input_tokens_seen": 34550608, "step": 25495 }, { "epoch": 0.8162089494910697, "grad_norm": 1.552894949913025, "learning_rate": 1.4535289728795821e-05, "loss": 0.3609, "num_input_tokens_seen": 34557760, "step": 25500 }, { "epoch": 0.8163689904615582, "grad_norm": 1.0303122997283936, "learning_rate": 1.4526374538374132e-05, "loss": 0.5493, "num_input_tokens_seen": 34564256, "step": 25505 }, { "epoch": 0.8165290314320466, "grad_norm": 2.4676780700683594, "learning_rate": 1.4517460963117097e-05, "loss": 0.6753, "num_input_tokens_seen": 34571296, "step": 25510 }, { "epoch": 0.8166890724025351, "grad_norm": 0.5882137417793274, "learning_rate": 1.4508549004399314e-05, "loss": 0.3743, "num_input_tokens_seen": 34577968, "step": 25515 }, { "epoch": 0.8168491133730235, "grad_norm": 1.6472806930541992, "learning_rate": 1.449963866359513e-05, "loss": 0.5548, "num_input_tokens_seen": 34584816, "step": 25520 }, { "epoch": 0.817009154343512, "grad_norm": 2.2328202724456787, "learning_rate": 1.4490729942078607e-05, "loss": 0.3918, "num_input_tokens_seen": 34591344, "step": 25525 }, { "epoch": 0.8171691953140003, "grad_norm": 0.7861356139183044, "learning_rate": 1.4481822841223608e-05, "loss": 0.4147, "num_input_tokens_seen": 34597776, "step": 25530 }, { "epoch": 0.8173292362844888, "grad_norm": 0.9414953589439392, "learning_rate": 1.4472917362403704e-05, "loss": 0.3091, "num_input_tokens_seen": 34604384, "step": 25535 }, { "epoch": 0.8174892772549772, "grad_norm": 1.0501925945281982, "learning_rate": 1.4464013506992224e-05, "loss": 0.4685, "num_input_tokens_seen": 34611328, "step": 25540 }, { "epoch": 0.8176493182254657, "grad_norm": 1.4395208358764648, "learning_rate": 1.4455111276362277e-05, "loss": 0.4884, "num_input_tokens_seen": 34618448, "step": 25545 }, { "epoch": 0.8178093591959542, "grad_norm": 0.9753334522247314, "learning_rate": 1.4446210671886676e-05, "loss": 0.2889, "num_input_tokens_seen": 34625328, "step": 25550 }, { "epoch": 0.8179694001664426, "grad_norm": 0.2837475836277008, "learning_rate": 1.4437311694938015e-05, "loss": 0.4024, "num_input_tokens_seen": 34631920, "step": 25555 }, { "epoch": 0.8181294411369311, "grad_norm": 1.021338701248169, "learning_rate": 1.442841434688864e-05, "loss": 0.2511, "num_input_tokens_seen": 34638480, "step": 25560 }, { "epoch": 0.8182894821074195, "grad_norm": 0.93056720495224, "learning_rate": 1.4419518629110615e-05, "loss": 0.4626, "num_input_tokens_seen": 34645072, "step": 25565 }, { "epoch": 0.818449523077908, "grad_norm": 0.4449019432067871, "learning_rate": 1.4410624542975778e-05, "loss": 0.4289, "num_input_tokens_seen": 34652528, "step": 25570 }, { "epoch": 0.8186095640483964, "grad_norm": 0.8920873403549194, "learning_rate": 1.4401732089855724e-05, "loss": 0.5804, "num_input_tokens_seen": 34658976, "step": 25575 }, { "epoch": 0.8187696050188848, "grad_norm": 1.3865141868591309, "learning_rate": 1.4392841271121754e-05, "loss": 0.4469, "num_input_tokens_seen": 34665488, "step": 25580 }, { "epoch": 0.8189296459893732, "grad_norm": 1.7837616205215454, "learning_rate": 1.438395208814497e-05, "loss": 0.4669, "num_input_tokens_seen": 34672624, "step": 25585 }, { "epoch": 0.8190896869598617, "grad_norm": 0.7877084612846375, "learning_rate": 1.4375064542296174e-05, "loss": 0.4342, "num_input_tokens_seen": 34679152, "step": 25590 }, { "epoch": 0.8192497279303502, "grad_norm": 1.0518970489501953, "learning_rate": 1.4366178634945946e-05, "loss": 0.433, "num_input_tokens_seen": 34685952, "step": 25595 }, { "epoch": 0.8194097689008386, "grad_norm": 1.176080346107483, "learning_rate": 1.4357294367464616e-05, "loss": 0.4706, "num_input_tokens_seen": 34692880, "step": 25600 }, { "epoch": 0.8194097689008386, "eval_loss": 0.45109474658966064, "eval_runtime": 503.0629, "eval_samples_per_second": 27.603, "eval_steps_per_second": 13.801, "num_input_tokens_seen": 34692880, "step": 25600 }, { "epoch": 0.8195698098713271, "grad_norm": 1.0059654712677002, "learning_rate": 1.434841174122224e-05, "loss": 0.4895, "num_input_tokens_seen": 34699712, "step": 25605 }, { "epoch": 0.8197298508418155, "grad_norm": 0.8708590269088745, "learning_rate": 1.4339530757588615e-05, "loss": 0.5323, "num_input_tokens_seen": 34706976, "step": 25610 }, { "epoch": 0.819889891812304, "grad_norm": 2.027834892272949, "learning_rate": 1.433065141793333e-05, "loss": 0.5521, "num_input_tokens_seen": 34713904, "step": 25615 }, { "epoch": 0.8200499327827924, "grad_norm": 1.0243840217590332, "learning_rate": 1.4321773723625665e-05, "loss": 0.3741, "num_input_tokens_seen": 34720160, "step": 25620 }, { "epoch": 0.8202099737532809, "grad_norm": 0.8289214372634888, "learning_rate": 1.4312897676034693e-05, "loss": 0.3242, "num_input_tokens_seen": 34726816, "step": 25625 }, { "epoch": 0.8203700147237692, "grad_norm": 0.8140386939048767, "learning_rate": 1.4304023276529188e-05, "loss": 0.3551, "num_input_tokens_seen": 34733584, "step": 25630 }, { "epoch": 0.8205300556942577, "grad_norm": 1.4009830951690674, "learning_rate": 1.4295150526477712e-05, "loss": 0.4519, "num_input_tokens_seen": 34740544, "step": 25635 }, { "epoch": 0.8206900966647461, "grad_norm": 1.0317585468292236, "learning_rate": 1.4286279427248562e-05, "loss": 0.5734, "num_input_tokens_seen": 34747504, "step": 25640 }, { "epoch": 0.8208501376352346, "grad_norm": 2.186417579650879, "learning_rate": 1.4277409980209747e-05, "loss": 0.4767, "num_input_tokens_seen": 34754624, "step": 25645 }, { "epoch": 0.8210101786057231, "grad_norm": 0.5648555755615234, "learning_rate": 1.4268542186729061e-05, "loss": 0.2958, "num_input_tokens_seen": 34761376, "step": 25650 }, { "epoch": 0.8211702195762115, "grad_norm": 0.8802680969238281, "learning_rate": 1.4259676048174043e-05, "loss": 0.5753, "num_input_tokens_seen": 34767888, "step": 25655 }, { "epoch": 0.8213302605467, "grad_norm": 0.8089010715484619, "learning_rate": 1.4250811565911937e-05, "loss": 0.418, "num_input_tokens_seen": 34774432, "step": 25660 }, { "epoch": 0.8214903015171884, "grad_norm": 1.3740124702453613, "learning_rate": 1.4241948741309782e-05, "loss": 0.3933, "num_input_tokens_seen": 34781152, "step": 25665 }, { "epoch": 0.8216503424876769, "grad_norm": 1.168876051902771, "learning_rate": 1.4233087575734317e-05, "loss": 0.4023, "num_input_tokens_seen": 34787584, "step": 25670 }, { "epoch": 0.8218103834581653, "grad_norm": 0.45532920956611633, "learning_rate": 1.422422807055206e-05, "loss": 0.3371, "num_input_tokens_seen": 34794624, "step": 25675 }, { "epoch": 0.8219704244286538, "grad_norm": 1.896723985671997, "learning_rate": 1.4215370227129243e-05, "loss": 0.4933, "num_input_tokens_seen": 34801392, "step": 25680 }, { "epoch": 0.8221304653991421, "grad_norm": 1.5246368646621704, "learning_rate": 1.4206514046831876e-05, "loss": 0.4957, "num_input_tokens_seen": 34808208, "step": 25685 }, { "epoch": 0.8222905063696306, "grad_norm": 0.829714834690094, "learning_rate": 1.419765953102567e-05, "loss": 0.4111, "num_input_tokens_seen": 34814864, "step": 25690 }, { "epoch": 0.822450547340119, "grad_norm": 0.8949618339538574, "learning_rate": 1.4188806681076125e-05, "loss": 0.4673, "num_input_tokens_seen": 34821472, "step": 25695 }, { "epoch": 0.8226105883106075, "grad_norm": 1.3834075927734375, "learning_rate": 1.4179955498348443e-05, "loss": 0.4578, "num_input_tokens_seen": 34828064, "step": 25700 }, { "epoch": 0.822770629281096, "grad_norm": 3.6078603267669678, "learning_rate": 1.4171105984207605e-05, "loss": 0.6362, "num_input_tokens_seen": 34834848, "step": 25705 }, { "epoch": 0.8229306702515844, "grad_norm": 0.8326038718223572, "learning_rate": 1.4162258140018304e-05, "loss": 0.4036, "num_input_tokens_seen": 34841712, "step": 25710 }, { "epoch": 0.8230907112220729, "grad_norm": 0.9624081254005432, "learning_rate": 1.4153411967144986e-05, "loss": 0.4123, "num_input_tokens_seen": 34848800, "step": 25715 }, { "epoch": 0.8232507521925613, "grad_norm": 1.8614740371704102, "learning_rate": 1.4144567466951864e-05, "loss": 0.5433, "num_input_tokens_seen": 34855184, "step": 25720 }, { "epoch": 0.8234107931630498, "grad_norm": 0.6166229844093323, "learning_rate": 1.4135724640802844e-05, "loss": 0.4643, "num_input_tokens_seen": 34862160, "step": 25725 }, { "epoch": 0.8235708341335382, "grad_norm": 0.9241877198219299, "learning_rate": 1.4126883490061615e-05, "loss": 0.4738, "num_input_tokens_seen": 34868768, "step": 25730 }, { "epoch": 0.8237308751040266, "grad_norm": 1.7183406352996826, "learning_rate": 1.4118044016091603e-05, "loss": 0.4452, "num_input_tokens_seen": 34875296, "step": 25735 }, { "epoch": 0.823890916074515, "grad_norm": 1.0577869415283203, "learning_rate": 1.410920622025594e-05, "loss": 0.4796, "num_input_tokens_seen": 34882064, "step": 25740 }, { "epoch": 0.8240509570450035, "grad_norm": 1.6547385454177856, "learning_rate": 1.4100370103917554e-05, "loss": 0.481, "num_input_tokens_seen": 34889008, "step": 25745 }, { "epoch": 0.8242109980154919, "grad_norm": 1.2681812047958374, "learning_rate": 1.409153566843907e-05, "loss": 0.3754, "num_input_tokens_seen": 34895616, "step": 25750 }, { "epoch": 0.8243710389859804, "grad_norm": 2.03560733795166, "learning_rate": 1.408270291518286e-05, "loss": 0.4506, "num_input_tokens_seen": 34902624, "step": 25755 }, { "epoch": 0.8245310799564689, "grad_norm": 1.6217045783996582, "learning_rate": 1.407387184551107e-05, "loss": 0.5481, "num_input_tokens_seen": 34909312, "step": 25760 }, { "epoch": 0.8246911209269573, "grad_norm": 0.7230198383331299, "learning_rate": 1.4065042460785532e-05, "loss": 0.4365, "num_input_tokens_seen": 34916176, "step": 25765 }, { "epoch": 0.8248511618974458, "grad_norm": 1.2759935855865479, "learning_rate": 1.405621476236787e-05, "loss": 0.4092, "num_input_tokens_seen": 34922992, "step": 25770 }, { "epoch": 0.8250112028679342, "grad_norm": 0.9458349347114563, "learning_rate": 1.4047388751619423e-05, "loss": 0.4696, "num_input_tokens_seen": 34929440, "step": 25775 }, { "epoch": 0.8251712438384227, "grad_norm": 0.9093321561813354, "learning_rate": 1.4038564429901264e-05, "loss": 0.2381, "num_input_tokens_seen": 34936096, "step": 25780 }, { "epoch": 0.825331284808911, "grad_norm": 0.9671236872673035, "learning_rate": 1.4029741798574227e-05, "loss": 0.3643, "num_input_tokens_seen": 34943056, "step": 25785 }, { "epoch": 0.8254913257793995, "grad_norm": 0.6398288607597351, "learning_rate": 1.402092085899886e-05, "loss": 0.2275, "num_input_tokens_seen": 34949904, "step": 25790 }, { "epoch": 0.8256513667498879, "grad_norm": 0.7064740657806396, "learning_rate": 1.4012101612535464e-05, "loss": 0.4456, "num_input_tokens_seen": 34956992, "step": 25795 }, { "epoch": 0.8258114077203764, "grad_norm": 1.447035312652588, "learning_rate": 1.4003284060544092e-05, "loss": 0.421, "num_input_tokens_seen": 34964656, "step": 25800 }, { "epoch": 0.8258114077203764, "eval_loss": 0.4504276514053345, "eval_runtime": 503.7701, "eval_samples_per_second": 27.564, "eval_steps_per_second": 13.782, "num_input_tokens_seen": 34964656, "step": 25800 }, { "epoch": 0.8259714486908648, "grad_norm": 1.2241755723953247, "learning_rate": 1.3994468204384504e-05, "loss": 0.3941, "num_input_tokens_seen": 34971184, "step": 25805 }, { "epoch": 0.8261314896613533, "grad_norm": 1.1004713773727417, "learning_rate": 1.398565404541622e-05, "loss": 0.4526, "num_input_tokens_seen": 34977824, "step": 25810 }, { "epoch": 0.8262915306318418, "grad_norm": 1.207359790802002, "learning_rate": 1.3976841584998513e-05, "loss": 0.5937, "num_input_tokens_seen": 34984336, "step": 25815 }, { "epoch": 0.8264515716023302, "grad_norm": 1.0834406614303589, "learning_rate": 1.3968030824490352e-05, "loss": 0.2498, "num_input_tokens_seen": 34990944, "step": 25820 }, { "epoch": 0.8266116125728187, "grad_norm": 1.394444465637207, "learning_rate": 1.3959221765250469e-05, "loss": 0.2955, "num_input_tokens_seen": 34997600, "step": 25825 }, { "epoch": 0.8267716535433071, "grad_norm": 1.0799819231033325, "learning_rate": 1.3950414408637343e-05, "loss": 0.4901, "num_input_tokens_seen": 35004464, "step": 25830 }, { "epoch": 0.8269316945137956, "grad_norm": 1.5854614973068237, "learning_rate": 1.3941608756009166e-05, "loss": 0.4562, "num_input_tokens_seen": 35010736, "step": 25835 }, { "epoch": 0.8270917354842839, "grad_norm": 1.007348895072937, "learning_rate": 1.3932804808723898e-05, "loss": 0.3162, "num_input_tokens_seen": 35017232, "step": 25840 }, { "epoch": 0.8272517764547724, "grad_norm": 1.516115427017212, "learning_rate": 1.3924002568139194e-05, "loss": 0.6671, "num_input_tokens_seen": 35023472, "step": 25845 }, { "epoch": 0.8274118174252608, "grad_norm": 1.3623595237731934, "learning_rate": 1.3915202035612485e-05, "loss": 0.5661, "num_input_tokens_seen": 35030096, "step": 25850 }, { "epoch": 0.8275718583957493, "grad_norm": 0.7556722164154053, "learning_rate": 1.3906403212500935e-05, "loss": 0.5322, "num_input_tokens_seen": 35036784, "step": 25855 }, { "epoch": 0.8277318993662378, "grad_norm": 0.8384426832199097, "learning_rate": 1.3897606100161409e-05, "loss": 0.3947, "num_input_tokens_seen": 35043632, "step": 25860 }, { "epoch": 0.8278919403367262, "grad_norm": 0.8374789953231812, "learning_rate": 1.388881069995055e-05, "loss": 0.4493, "num_input_tokens_seen": 35050544, "step": 25865 }, { "epoch": 0.8280519813072147, "grad_norm": 0.9689523577690125, "learning_rate": 1.3880017013224708e-05, "loss": 0.5037, "num_input_tokens_seen": 35058208, "step": 25870 }, { "epoch": 0.8282120222777031, "grad_norm": 1.337121605873108, "learning_rate": 1.3871225041339984e-05, "loss": 0.552, "num_input_tokens_seen": 35065216, "step": 25875 }, { "epoch": 0.8283720632481916, "grad_norm": 1.2044637203216553, "learning_rate": 1.386243478565222e-05, "loss": 0.5908, "num_input_tokens_seen": 35071168, "step": 25880 }, { "epoch": 0.82853210421868, "grad_norm": 1.4428751468658447, "learning_rate": 1.3853646247516966e-05, "loss": 0.3541, "num_input_tokens_seen": 35078224, "step": 25885 }, { "epoch": 0.8286921451891684, "grad_norm": 1.1697030067443848, "learning_rate": 1.3844859428289545e-05, "loss": 0.394, "num_input_tokens_seen": 35085248, "step": 25890 }, { "epoch": 0.8288521861596568, "grad_norm": 1.074967622756958, "learning_rate": 1.3836074329324984e-05, "loss": 0.3528, "num_input_tokens_seen": 35091520, "step": 25895 }, { "epoch": 0.8290122271301453, "grad_norm": 0.6935981512069702, "learning_rate": 1.3827290951978044e-05, "loss": 0.3788, "num_input_tokens_seen": 35098208, "step": 25900 }, { "epoch": 0.8291722681006337, "grad_norm": 0.34131795167922974, "learning_rate": 1.381850929760326e-05, "loss": 0.5324, "num_input_tokens_seen": 35105360, "step": 25905 }, { "epoch": 0.8293323090711222, "grad_norm": 0.5398966670036316, "learning_rate": 1.3809729367554842e-05, "loss": 0.3312, "num_input_tokens_seen": 35112496, "step": 25910 }, { "epoch": 0.8294923500416107, "grad_norm": 0.5615841746330261, "learning_rate": 1.3800951163186784e-05, "loss": 0.5101, "num_input_tokens_seen": 35119248, "step": 25915 }, { "epoch": 0.8296523910120991, "grad_norm": 0.6493860483169556, "learning_rate": 1.3792174685852801e-05, "loss": 0.2611, "num_input_tokens_seen": 35126000, "step": 25920 }, { "epoch": 0.8298124319825876, "grad_norm": 1.3403602838516235, "learning_rate": 1.378339993690632e-05, "loss": 0.5081, "num_input_tokens_seen": 35132688, "step": 25925 }, { "epoch": 0.829972472953076, "grad_norm": 1.8076492547988892, "learning_rate": 1.3774626917700523e-05, "loss": 0.5136, "num_input_tokens_seen": 35139488, "step": 25930 }, { "epoch": 0.8301325139235645, "grad_norm": 0.37496134638786316, "learning_rate": 1.3765855629588334e-05, "loss": 0.3484, "num_input_tokens_seen": 35146016, "step": 25935 }, { "epoch": 0.8302925548940528, "grad_norm": 0.4138462245464325, "learning_rate": 1.3757086073922374e-05, "loss": 0.3972, "num_input_tokens_seen": 35152832, "step": 25940 }, { "epoch": 0.8304525958645413, "grad_norm": 0.9186856746673584, "learning_rate": 1.3748318252055038e-05, "loss": 0.4417, "num_input_tokens_seen": 35160272, "step": 25945 }, { "epoch": 0.8306126368350297, "grad_norm": 1.287805199623108, "learning_rate": 1.3739552165338416e-05, "loss": 0.3352, "num_input_tokens_seen": 35166672, "step": 25950 }, { "epoch": 0.8307726778055182, "grad_norm": 0.8282560706138611, "learning_rate": 1.3730787815124354e-05, "loss": 0.6978, "num_input_tokens_seen": 35173360, "step": 25955 }, { "epoch": 0.8309327187760066, "grad_norm": 0.7043783664703369, "learning_rate": 1.3722025202764443e-05, "loss": 0.4495, "num_input_tokens_seen": 35180640, "step": 25960 }, { "epoch": 0.8310927597464951, "grad_norm": 3.073270320892334, "learning_rate": 1.371326432960997e-05, "loss": 0.6886, "num_input_tokens_seen": 35187248, "step": 25965 }, { "epoch": 0.8312528007169836, "grad_norm": 1.062348484992981, "learning_rate": 1.3704505197011969e-05, "loss": 0.6176, "num_input_tokens_seen": 35194192, "step": 25970 }, { "epoch": 0.831412841687472, "grad_norm": 0.5976991057395935, "learning_rate": 1.3695747806321224e-05, "loss": 0.4913, "num_input_tokens_seen": 35201008, "step": 25975 }, { "epoch": 0.8315728826579605, "grad_norm": 0.9134689569473267, "learning_rate": 1.3686992158888212e-05, "loss": 0.4228, "num_input_tokens_seen": 35207840, "step": 25980 }, { "epoch": 0.8317329236284489, "grad_norm": 0.9449376463890076, "learning_rate": 1.367823825606319e-05, "loss": 0.7988, "num_input_tokens_seen": 35214544, "step": 25985 }, { "epoch": 0.8318929645989374, "grad_norm": 0.7036371827125549, "learning_rate": 1.36694860991961e-05, "loss": 0.3432, "num_input_tokens_seen": 35221344, "step": 25990 }, { "epoch": 0.8320530055694257, "grad_norm": 0.9958868622779846, "learning_rate": 1.3660735689636636e-05, "loss": 0.3643, "num_input_tokens_seen": 35228000, "step": 25995 }, { "epoch": 0.8322130465399142, "grad_norm": 1.2328925132751465, "learning_rate": 1.365198702873424e-05, "loss": 0.5967, "num_input_tokens_seen": 35234256, "step": 26000 }, { "epoch": 0.8322130465399142, "eval_loss": 0.4503823220729828, "eval_runtime": 503.0508, "eval_samples_per_second": 27.604, "eval_steps_per_second": 13.802, "num_input_tokens_seen": 35234256, "step": 26000 }, { "epoch": 0.8323730875104026, "grad_norm": 0.6350032687187195, "learning_rate": 1.364324011783804e-05, "loss": 0.4256, "num_input_tokens_seen": 35240416, "step": 26005 }, { "epoch": 0.8325331284808911, "grad_norm": 1.804351806640625, "learning_rate": 1.3634494958296934e-05, "loss": 0.5285, "num_input_tokens_seen": 35246864, "step": 26010 }, { "epoch": 0.8326931694513795, "grad_norm": 0.5879729986190796, "learning_rate": 1.3625751551459542e-05, "loss": 0.3983, "num_input_tokens_seen": 35253616, "step": 26015 }, { "epoch": 0.832853210421868, "grad_norm": 1.319486141204834, "learning_rate": 1.3617009898674188e-05, "loss": 0.5352, "num_input_tokens_seen": 35260304, "step": 26020 }, { "epoch": 0.8330132513923565, "grad_norm": 1.314268946647644, "learning_rate": 1.3608270001288967e-05, "loss": 0.3969, "num_input_tokens_seen": 35267280, "step": 26025 }, { "epoch": 0.8331732923628449, "grad_norm": 1.756417155265808, "learning_rate": 1.359953186065166e-05, "loss": 0.4683, "num_input_tokens_seen": 35274288, "step": 26030 }, { "epoch": 0.8333333333333334, "grad_norm": 1.6140371561050415, "learning_rate": 1.3590795478109814e-05, "loss": 0.5428, "num_input_tokens_seen": 35280832, "step": 26035 }, { "epoch": 0.8334933743038218, "grad_norm": 0.8924579620361328, "learning_rate": 1.3582060855010675e-05, "loss": 0.4517, "num_input_tokens_seen": 35287808, "step": 26040 }, { "epoch": 0.8336534152743102, "grad_norm": 0.8689470887184143, "learning_rate": 1.3573327992701245e-05, "loss": 0.3779, "num_input_tokens_seen": 35294160, "step": 26045 }, { "epoch": 0.8338134562447986, "grad_norm": 1.0036723613739014, "learning_rate": 1.356459689252823e-05, "loss": 0.303, "num_input_tokens_seen": 35300944, "step": 26050 }, { "epoch": 0.8339734972152871, "grad_norm": 0.8757918477058411, "learning_rate": 1.3555867555838087e-05, "loss": 0.3819, "num_input_tokens_seen": 35307680, "step": 26055 }, { "epoch": 0.8341335381857755, "grad_norm": 1.0149078369140625, "learning_rate": 1.3547139983976975e-05, "loss": 0.4219, "num_input_tokens_seen": 35314704, "step": 26060 }, { "epoch": 0.834293579156264, "grad_norm": 0.7193334102630615, "learning_rate": 1.3538414178290815e-05, "loss": 0.383, "num_input_tokens_seen": 35321744, "step": 26065 }, { "epoch": 0.8344536201267524, "grad_norm": 0.628097414970398, "learning_rate": 1.3529690140125209e-05, "loss": 0.3749, "num_input_tokens_seen": 35328384, "step": 26070 }, { "epoch": 0.8346136610972409, "grad_norm": 1.1393030881881714, "learning_rate": 1.352096787082553e-05, "loss": 0.4165, "num_input_tokens_seen": 35335424, "step": 26075 }, { "epoch": 0.8347737020677294, "grad_norm": 2.3954906463623047, "learning_rate": 1.3512247371736871e-05, "loss": 0.6209, "num_input_tokens_seen": 35342464, "step": 26080 }, { "epoch": 0.8349337430382178, "grad_norm": 1.4629451036453247, "learning_rate": 1.3503528644204022e-05, "loss": 0.4854, "num_input_tokens_seen": 35348976, "step": 26085 }, { "epoch": 0.8350937840087063, "grad_norm": 2.3944478034973145, "learning_rate": 1.349481168957153e-05, "loss": 0.4709, "num_input_tokens_seen": 35355952, "step": 26090 }, { "epoch": 0.8352538249791946, "grad_norm": 0.9004864692687988, "learning_rate": 1.3486096509183665e-05, "loss": 0.5068, "num_input_tokens_seen": 35362320, "step": 26095 }, { "epoch": 0.8354138659496831, "grad_norm": 1.6042723655700684, "learning_rate": 1.3477383104384406e-05, "loss": 0.4591, "num_input_tokens_seen": 35369568, "step": 26100 }, { "epoch": 0.8355739069201715, "grad_norm": 0.4576559364795685, "learning_rate": 1.3468671476517481e-05, "loss": 0.2859, "num_input_tokens_seen": 35376576, "step": 26105 }, { "epoch": 0.83573394789066, "grad_norm": 0.9797880053520203, "learning_rate": 1.3459961626926326e-05, "loss": 0.374, "num_input_tokens_seen": 35383248, "step": 26110 }, { "epoch": 0.8358939888611484, "grad_norm": 0.5582695603370667, "learning_rate": 1.3451253556954101e-05, "loss": 0.4561, "num_input_tokens_seen": 35390640, "step": 26115 }, { "epoch": 0.8360540298316369, "grad_norm": 0.5881622433662415, "learning_rate": 1.3442547267943717e-05, "loss": 0.3316, "num_input_tokens_seen": 35397376, "step": 26120 }, { "epoch": 0.8362140708021254, "grad_norm": 0.8296788334846497, "learning_rate": 1.3433842761237774e-05, "loss": 0.5, "num_input_tokens_seen": 35404256, "step": 26125 }, { "epoch": 0.8363741117726138, "grad_norm": 0.7105377316474915, "learning_rate": 1.3425140038178639e-05, "loss": 0.327, "num_input_tokens_seen": 35410800, "step": 26130 }, { "epoch": 0.8365341527431023, "grad_norm": 1.4422366619110107, "learning_rate": 1.3416439100108358e-05, "loss": 0.4517, "num_input_tokens_seen": 35417328, "step": 26135 }, { "epoch": 0.8366941937135907, "grad_norm": 1.9510430097579956, "learning_rate": 1.3407739948368734e-05, "loss": 0.4563, "num_input_tokens_seen": 35424144, "step": 26140 }, { "epoch": 0.8368542346840792, "grad_norm": 0.588683009147644, "learning_rate": 1.3399042584301298e-05, "loss": 0.333, "num_input_tokens_seen": 35431008, "step": 26145 }, { "epoch": 0.8370142756545675, "grad_norm": 1.1061758995056152, "learning_rate": 1.3390347009247272e-05, "loss": 0.4087, "num_input_tokens_seen": 35437456, "step": 26150 }, { "epoch": 0.837174316625056, "grad_norm": 0.763550877571106, "learning_rate": 1.3381653224547635e-05, "loss": 0.4301, "num_input_tokens_seen": 35445104, "step": 26155 }, { "epoch": 0.8373343575955444, "grad_norm": 0.5590509176254272, "learning_rate": 1.3372961231543086e-05, "loss": 0.3208, "num_input_tokens_seen": 35451472, "step": 26160 }, { "epoch": 0.8374943985660329, "grad_norm": 1.564724087715149, "learning_rate": 1.3364271031574016e-05, "loss": 0.3687, "num_input_tokens_seen": 35458176, "step": 26165 }, { "epoch": 0.8376544395365213, "grad_norm": 0.7072741389274597, "learning_rate": 1.335558262598059e-05, "loss": 0.42, "num_input_tokens_seen": 35464896, "step": 26170 }, { "epoch": 0.8378144805070098, "grad_norm": 0.7719861268997192, "learning_rate": 1.3346896016102645e-05, "loss": 0.5134, "num_input_tokens_seen": 35471712, "step": 26175 }, { "epoch": 0.8379745214774983, "grad_norm": 0.7853825688362122, "learning_rate": 1.3338211203279788e-05, "loss": 0.4561, "num_input_tokens_seen": 35478096, "step": 26180 }, { "epoch": 0.8381345624479867, "grad_norm": 0.6827985048294067, "learning_rate": 1.3329528188851303e-05, "loss": 0.3128, "num_input_tokens_seen": 35484832, "step": 26185 }, { "epoch": 0.8382946034184752, "grad_norm": 0.9926403164863586, "learning_rate": 1.3320846974156242e-05, "loss": 0.4282, "num_input_tokens_seen": 35491344, "step": 26190 }, { "epoch": 0.8384546443889636, "grad_norm": 0.7206461429595947, "learning_rate": 1.3312167560533337e-05, "loss": 0.4897, "num_input_tokens_seen": 35498048, "step": 26195 }, { "epoch": 0.838614685359452, "grad_norm": 0.9023417234420776, "learning_rate": 1.3303489949321082e-05, "loss": 0.4502, "num_input_tokens_seen": 35504864, "step": 26200 }, { "epoch": 0.838614685359452, "eval_loss": 0.4503050148487091, "eval_runtime": 502.8303, "eval_samples_per_second": 27.616, "eval_steps_per_second": 13.808, "num_input_tokens_seen": 35504864, "step": 26200 }, { "epoch": 0.8387747263299404, "grad_norm": 0.9697619080543518, "learning_rate": 1.3294814141857653e-05, "loss": 0.3654, "num_input_tokens_seen": 35511760, "step": 26205 }, { "epoch": 0.8389347673004289, "grad_norm": 0.7081337571144104, "learning_rate": 1.3286140139480992e-05, "loss": 0.3851, "num_input_tokens_seen": 35518736, "step": 26210 }, { "epoch": 0.8390948082709173, "grad_norm": 0.8838334679603577, "learning_rate": 1.3277467943528719e-05, "loss": 0.352, "num_input_tokens_seen": 35525360, "step": 26215 }, { "epoch": 0.8392548492414058, "grad_norm": 0.7081719040870667, "learning_rate": 1.3268797555338203e-05, "loss": 0.5153, "num_input_tokens_seen": 35532368, "step": 26220 }, { "epoch": 0.8394148902118942, "grad_norm": 1.2438362836837769, "learning_rate": 1.3260128976246533e-05, "loss": 0.2481, "num_input_tokens_seen": 35538976, "step": 26225 }, { "epoch": 0.8395749311823827, "grad_norm": 1.3202707767486572, "learning_rate": 1.32514622075905e-05, "loss": 0.4593, "num_input_tokens_seen": 35546080, "step": 26230 }, { "epoch": 0.8397349721528712, "grad_norm": 1.0216838121414185, "learning_rate": 1.3242797250706638e-05, "loss": 0.4386, "num_input_tokens_seen": 35552832, "step": 26235 }, { "epoch": 0.8398950131233596, "grad_norm": 1.1654163599014282, "learning_rate": 1.3234134106931195e-05, "loss": 0.3533, "num_input_tokens_seen": 35559392, "step": 26240 }, { "epoch": 0.8400550540938481, "grad_norm": 1.0375697612762451, "learning_rate": 1.322547277760013e-05, "loss": 0.5551, "num_input_tokens_seen": 35566512, "step": 26245 }, { "epoch": 0.8402150950643364, "grad_norm": 0.7770195007324219, "learning_rate": 1.3216813264049132e-05, "loss": 0.3353, "num_input_tokens_seen": 35573152, "step": 26250 }, { "epoch": 0.8403751360348249, "grad_norm": 6.112483024597168, "learning_rate": 1.32081555676136e-05, "loss": 0.4747, "num_input_tokens_seen": 35579952, "step": 26255 }, { "epoch": 0.8405351770053133, "grad_norm": 1.2336256504058838, "learning_rate": 1.3199499689628674e-05, "loss": 0.4315, "num_input_tokens_seen": 35586752, "step": 26260 }, { "epoch": 0.8406952179758018, "grad_norm": 1.2815037965774536, "learning_rate": 1.3190845631429192e-05, "loss": 0.4601, "num_input_tokens_seen": 35594064, "step": 26265 }, { "epoch": 0.8408552589462902, "grad_norm": 0.4700641334056854, "learning_rate": 1.3182193394349704e-05, "loss": 0.2921, "num_input_tokens_seen": 35600624, "step": 26270 }, { "epoch": 0.8410152999167787, "grad_norm": 0.7726446390151978, "learning_rate": 1.3173542979724507e-05, "loss": 0.424, "num_input_tokens_seen": 35607472, "step": 26275 }, { "epoch": 0.8411753408872671, "grad_norm": 0.8494935631752014, "learning_rate": 1.3164894388887617e-05, "loss": 0.3579, "num_input_tokens_seen": 35614256, "step": 26280 }, { "epoch": 0.8413353818577556, "grad_norm": 1.1257555484771729, "learning_rate": 1.3156247623172727e-05, "loss": 0.4189, "num_input_tokens_seen": 35620592, "step": 26285 }, { "epoch": 0.8414954228282441, "grad_norm": 0.870383620262146, "learning_rate": 1.3147602683913302e-05, "loss": 0.5325, "num_input_tokens_seen": 35627088, "step": 26290 }, { "epoch": 0.8416554637987325, "grad_norm": 1.4688589572906494, "learning_rate": 1.3138959572442481e-05, "loss": 0.5111, "num_input_tokens_seen": 35633680, "step": 26295 }, { "epoch": 0.841815504769221, "grad_norm": 0.8316260576248169, "learning_rate": 1.3130318290093146e-05, "loss": 0.4614, "num_input_tokens_seen": 35640464, "step": 26300 }, { "epoch": 0.8419755457397093, "grad_norm": 0.6681721806526184, "learning_rate": 1.3121678838197909e-05, "loss": 0.5264, "num_input_tokens_seen": 35646976, "step": 26305 }, { "epoch": 0.8421355867101978, "grad_norm": 0.6301558017730713, "learning_rate": 1.3113041218089056e-05, "loss": 0.3484, "num_input_tokens_seen": 35653984, "step": 26310 }, { "epoch": 0.8422956276806862, "grad_norm": 1.312416911125183, "learning_rate": 1.3104405431098626e-05, "loss": 0.4348, "num_input_tokens_seen": 35660688, "step": 26315 }, { "epoch": 0.8424556686511747, "grad_norm": 1.702574372291565, "learning_rate": 1.3095771478558377e-05, "loss": 0.5868, "num_input_tokens_seen": 35667280, "step": 26320 }, { "epoch": 0.8426157096216631, "grad_norm": 1.3538000583648682, "learning_rate": 1.3087139361799766e-05, "loss": 0.3423, "num_input_tokens_seen": 35674064, "step": 26325 }, { "epoch": 0.8427757505921516, "grad_norm": 1.3969719409942627, "learning_rate": 1.3078509082153964e-05, "loss": 0.3174, "num_input_tokens_seen": 35681040, "step": 26330 }, { "epoch": 0.8429357915626401, "grad_norm": 0.5000377297401428, "learning_rate": 1.3069880640951885e-05, "loss": 0.3467, "num_input_tokens_seen": 35687712, "step": 26335 }, { "epoch": 0.8430958325331285, "grad_norm": 1.1883724927902222, "learning_rate": 1.3061254039524123e-05, "loss": 0.399, "num_input_tokens_seen": 35694208, "step": 26340 }, { "epoch": 0.843255873503617, "grad_norm": 1.319361925125122, "learning_rate": 1.3052629279201028e-05, "loss": 0.3266, "num_input_tokens_seen": 35701184, "step": 26345 }, { "epoch": 0.8434159144741054, "grad_norm": 2.033771276473999, "learning_rate": 1.3044006361312633e-05, "loss": 0.393, "num_input_tokens_seen": 35708000, "step": 26350 }, { "epoch": 0.8435759554445939, "grad_norm": 1.6975018978118896, "learning_rate": 1.30353852871887e-05, "loss": 0.572, "num_input_tokens_seen": 35714816, "step": 26355 }, { "epoch": 0.8437359964150822, "grad_norm": 1.1084200143814087, "learning_rate": 1.302676605815873e-05, "loss": 0.4625, "num_input_tokens_seen": 35721680, "step": 26360 }, { "epoch": 0.8438960373855707, "grad_norm": 1.2515171766281128, "learning_rate": 1.3018148675551884e-05, "loss": 0.5109, "num_input_tokens_seen": 35729344, "step": 26365 }, { "epoch": 0.8440560783560591, "grad_norm": 2.4058730602264404, "learning_rate": 1.3009533140697094e-05, "loss": 0.6041, "num_input_tokens_seen": 35736256, "step": 26370 }, { "epoch": 0.8442161193265476, "grad_norm": 1.2949960231781006, "learning_rate": 1.3000919454922966e-05, "loss": 0.3534, "num_input_tokens_seen": 35743040, "step": 26375 }, { "epoch": 0.844376160297036, "grad_norm": 0.8214130401611328, "learning_rate": 1.299230761955785e-05, "loss": 0.4683, "num_input_tokens_seen": 35750320, "step": 26380 }, { "epoch": 0.8445362012675245, "grad_norm": 0.7390081286430359, "learning_rate": 1.2983697635929807e-05, "loss": 0.5145, "num_input_tokens_seen": 35757200, "step": 26385 }, { "epoch": 0.844696242238013, "grad_norm": 0.6877449154853821, "learning_rate": 1.2975089505366584e-05, "loss": 0.5347, "num_input_tokens_seen": 35764112, "step": 26390 }, { "epoch": 0.8448562832085014, "grad_norm": 2.1901679039001465, "learning_rate": 1.2966483229195683e-05, "loss": 0.5076, "num_input_tokens_seen": 35770832, "step": 26395 }, { "epoch": 0.8450163241789899, "grad_norm": 1.7150601148605347, "learning_rate": 1.2957878808744283e-05, "loss": 0.5948, "num_input_tokens_seen": 35777296, "step": 26400 }, { "epoch": 0.8450163241789899, "eval_loss": 0.45004814863204956, "eval_runtime": 503.4656, "eval_samples_per_second": 27.581, "eval_steps_per_second": 13.79, "num_input_tokens_seen": 35777296, "step": 26400 }, { "epoch": 0.8451763651494782, "grad_norm": 1.3758656978607178, "learning_rate": 1.294927624533931e-05, "loss": 0.3911, "num_input_tokens_seen": 35783840, "step": 26405 }, { "epoch": 0.8453364061199667, "grad_norm": 0.981448769569397, "learning_rate": 1.2940675540307378e-05, "loss": 0.4465, "num_input_tokens_seen": 35790400, "step": 26410 }, { "epoch": 0.8454964470904551, "grad_norm": 1.0470880270004272, "learning_rate": 1.2932076694974814e-05, "loss": 0.3589, "num_input_tokens_seen": 35797456, "step": 26415 }, { "epoch": 0.8456564880609436, "grad_norm": 1.280579924583435, "learning_rate": 1.2923479710667682e-05, "loss": 0.4093, "num_input_tokens_seen": 35804160, "step": 26420 }, { "epoch": 0.845816529031432, "grad_norm": 1.2628257274627686, "learning_rate": 1.2914884588711751e-05, "loss": 0.3294, "num_input_tokens_seen": 35810864, "step": 26425 }, { "epoch": 0.8459765700019205, "grad_norm": 1.7116619348526, "learning_rate": 1.2906291330432475e-05, "loss": 0.5817, "num_input_tokens_seen": 35817088, "step": 26430 }, { "epoch": 0.8461366109724089, "grad_norm": 4.988977432250977, "learning_rate": 1.2897699937155055e-05, "loss": 0.4418, "num_input_tokens_seen": 35823616, "step": 26435 }, { "epoch": 0.8462966519428974, "grad_norm": 0.3531813323497772, "learning_rate": 1.2889110410204403e-05, "loss": 0.4354, "num_input_tokens_seen": 35830384, "step": 26440 }, { "epoch": 0.8464566929133859, "grad_norm": 1.2506219148635864, "learning_rate": 1.2880522750905111e-05, "loss": 0.526, "num_input_tokens_seen": 35837216, "step": 26445 }, { "epoch": 0.8466167338838743, "grad_norm": 0.6922656893730164, "learning_rate": 1.2871936960581523e-05, "loss": 0.4738, "num_input_tokens_seen": 35843760, "step": 26450 }, { "epoch": 0.8467767748543628, "grad_norm": 2.331763505935669, "learning_rate": 1.2863353040557658e-05, "loss": 0.4268, "num_input_tokens_seen": 35850080, "step": 26455 }, { "epoch": 0.8469368158248511, "grad_norm": 1.9792859554290771, "learning_rate": 1.2854770992157273e-05, "loss": 0.5008, "num_input_tokens_seen": 35856720, "step": 26460 }, { "epoch": 0.8470968567953396, "grad_norm": 1.8973491191864014, "learning_rate": 1.2846190816703835e-05, "loss": 0.4362, "num_input_tokens_seen": 35863056, "step": 26465 }, { "epoch": 0.847256897765828, "grad_norm": 0.6732066869735718, "learning_rate": 1.2837612515520498e-05, "loss": 0.329, "num_input_tokens_seen": 35869632, "step": 26470 }, { "epoch": 0.8474169387363165, "grad_norm": 1.6017045974731445, "learning_rate": 1.2829036089930163e-05, "loss": 0.5226, "num_input_tokens_seen": 35877024, "step": 26475 }, { "epoch": 0.8475769797068049, "grad_norm": 1.2184171676635742, "learning_rate": 1.2820461541255412e-05, "loss": 0.4244, "num_input_tokens_seen": 35883776, "step": 26480 }, { "epoch": 0.8477370206772934, "grad_norm": 1.5944713354110718, "learning_rate": 1.2811888870818543e-05, "loss": 0.4551, "num_input_tokens_seen": 35890496, "step": 26485 }, { "epoch": 0.8478970616477818, "grad_norm": 0.8106061220169067, "learning_rate": 1.2803318079941581e-05, "loss": 0.407, "num_input_tokens_seen": 35897136, "step": 26490 }, { "epoch": 0.8480571026182703, "grad_norm": 0.7104806900024414, "learning_rate": 1.2794749169946235e-05, "loss": 0.3203, "num_input_tokens_seen": 35903776, "step": 26495 }, { "epoch": 0.8482171435887588, "grad_norm": 1.6074472665786743, "learning_rate": 1.2786182142153952e-05, "loss": 0.4448, "num_input_tokens_seen": 35910272, "step": 26500 }, { "epoch": 0.8483771845592472, "grad_norm": 1.0076357126235962, "learning_rate": 1.2777616997885878e-05, "loss": 0.3777, "num_input_tokens_seen": 35917104, "step": 26505 }, { "epoch": 0.8485372255297357, "grad_norm": 1.023148536682129, "learning_rate": 1.2769053738462847e-05, "loss": 0.385, "num_input_tokens_seen": 35924160, "step": 26510 }, { "epoch": 0.848697266500224, "grad_norm": 0.5624628663063049, "learning_rate": 1.2760492365205434e-05, "loss": 0.4256, "num_input_tokens_seen": 35930896, "step": 26515 }, { "epoch": 0.8488573074707125, "grad_norm": 1.5528383255004883, "learning_rate": 1.2751932879433919e-05, "loss": 0.5602, "num_input_tokens_seen": 35937584, "step": 26520 }, { "epoch": 0.8490173484412009, "grad_norm": 1.0362168550491333, "learning_rate": 1.2743375282468267e-05, "loss": 0.3676, "num_input_tokens_seen": 35944384, "step": 26525 }, { "epoch": 0.8491773894116894, "grad_norm": 1.102369785308838, "learning_rate": 1.2734819575628182e-05, "loss": 0.4179, "num_input_tokens_seen": 35950944, "step": 26530 }, { "epoch": 0.8493374303821778, "grad_norm": 1.5860427618026733, "learning_rate": 1.2726265760233039e-05, "loss": 0.4765, "num_input_tokens_seen": 35957616, "step": 26535 }, { "epoch": 0.8494974713526663, "grad_norm": 0.9812548756599426, "learning_rate": 1.271771383760197e-05, "loss": 0.3934, "num_input_tokens_seen": 35964464, "step": 26540 }, { "epoch": 0.8496575123231547, "grad_norm": 1.4465925693511963, "learning_rate": 1.2709163809053764e-05, "loss": 0.4288, "num_input_tokens_seen": 35971216, "step": 26545 }, { "epoch": 0.8498175532936432, "grad_norm": 0.40670883655548096, "learning_rate": 1.2700615675906963e-05, "loss": 0.33, "num_input_tokens_seen": 35978272, "step": 26550 }, { "epoch": 0.8499775942641317, "grad_norm": 0.8171785473823547, "learning_rate": 1.269206943947978e-05, "loss": 0.4432, "num_input_tokens_seen": 35984992, "step": 26555 }, { "epoch": 0.85013763523462, "grad_norm": 1.208061695098877, "learning_rate": 1.2683525101090177e-05, "loss": 0.5004, "num_input_tokens_seen": 35991520, "step": 26560 }, { "epoch": 0.8502976762051085, "grad_norm": 1.0238304138183594, "learning_rate": 1.2674982662055765e-05, "loss": 0.4719, "num_input_tokens_seen": 35998560, "step": 26565 }, { "epoch": 0.8504577171755969, "grad_norm": 0.8499165177345276, "learning_rate": 1.2666442123693922e-05, "loss": 0.3729, "num_input_tokens_seen": 36005200, "step": 26570 }, { "epoch": 0.8506177581460854, "grad_norm": 2.877845048904419, "learning_rate": 1.265790348732169e-05, "loss": 0.4393, "num_input_tokens_seen": 36012112, "step": 26575 }, { "epoch": 0.8507777991165738, "grad_norm": 1.7891558408737183, "learning_rate": 1.264936675425584e-05, "loss": 0.406, "num_input_tokens_seen": 36018512, "step": 26580 }, { "epoch": 0.8509378400870623, "grad_norm": 2.2154998779296875, "learning_rate": 1.2640831925812852e-05, "loss": 0.5829, "num_input_tokens_seen": 36025120, "step": 26585 }, { "epoch": 0.8510978810575507, "grad_norm": 1.2540571689605713, "learning_rate": 1.263229900330889e-05, "loss": 0.5312, "num_input_tokens_seen": 36031936, "step": 26590 }, { "epoch": 0.8512579220280392, "grad_norm": 1.2480758428573608, "learning_rate": 1.2623767988059843e-05, "loss": 0.2978, "num_input_tokens_seen": 36038560, "step": 26595 }, { "epoch": 0.8514179629985277, "grad_norm": 0.6364144086837769, "learning_rate": 1.2615238881381309e-05, "loss": 0.3845, "num_input_tokens_seen": 36045376, "step": 26600 }, { "epoch": 0.8514179629985277, "eval_loss": 0.4493965804576874, "eval_runtime": 503.49, "eval_samples_per_second": 27.579, "eval_steps_per_second": 13.79, "num_input_tokens_seen": 36045376, "step": 26600 }, { "epoch": 0.8515780039690161, "grad_norm": 0.5651815533638, "learning_rate": 1.2606711684588568e-05, "loss": 0.6678, "num_input_tokens_seen": 36052272, "step": 26605 }, { "epoch": 0.8517380449395046, "grad_norm": 1.083119511604309, "learning_rate": 1.2598186398996636e-05, "loss": 0.3556, "num_input_tokens_seen": 36058848, "step": 26610 }, { "epoch": 0.8518980859099929, "grad_norm": 0.7432171702384949, "learning_rate": 1.2589663025920207e-05, "loss": 0.4544, "num_input_tokens_seen": 36065376, "step": 26615 }, { "epoch": 0.8520581268804814, "grad_norm": 1.2714964151382446, "learning_rate": 1.2581141566673705e-05, "loss": 0.5968, "num_input_tokens_seen": 36072048, "step": 26620 }, { "epoch": 0.8522181678509698, "grad_norm": 0.7764636278152466, "learning_rate": 1.257262202257124e-05, "loss": 0.4761, "num_input_tokens_seen": 36078816, "step": 26625 }, { "epoch": 0.8523782088214583, "grad_norm": 3.1453585624694824, "learning_rate": 1.2564104394926618e-05, "loss": 0.6576, "num_input_tokens_seen": 36085968, "step": 26630 }, { "epoch": 0.8525382497919467, "grad_norm": 1.1808048486709595, "learning_rate": 1.2555588685053383e-05, "loss": 0.5585, "num_input_tokens_seen": 36092640, "step": 26635 }, { "epoch": 0.8526982907624352, "grad_norm": 0.989133358001709, "learning_rate": 1.2547074894264762e-05, "loss": 0.5412, "num_input_tokens_seen": 36099920, "step": 26640 }, { "epoch": 0.8528583317329236, "grad_norm": 0.7757540345191956, "learning_rate": 1.2538563023873679e-05, "loss": 0.45, "num_input_tokens_seen": 36106624, "step": 26645 }, { "epoch": 0.8530183727034121, "grad_norm": 0.6240809559822083, "learning_rate": 1.2530053075192789e-05, "loss": 0.3127, "num_input_tokens_seen": 36113392, "step": 26650 }, { "epoch": 0.8531784136739006, "grad_norm": 0.7750654220581055, "learning_rate": 1.252154504953441e-05, "loss": 0.4722, "num_input_tokens_seen": 36120176, "step": 26655 }, { "epoch": 0.853338454644389, "grad_norm": 0.5008593201637268, "learning_rate": 1.25130389482106e-05, "loss": 0.3717, "num_input_tokens_seen": 36127008, "step": 26660 }, { "epoch": 0.8534984956148775, "grad_norm": 0.8817144632339478, "learning_rate": 1.2504534772533116e-05, "loss": 0.4114, "num_input_tokens_seen": 36133712, "step": 26665 }, { "epoch": 0.8536585365853658, "grad_norm": 0.6729145646095276, "learning_rate": 1.2496032523813387e-05, "loss": 0.4648, "num_input_tokens_seen": 36140352, "step": 26670 }, { "epoch": 0.8538185775558543, "grad_norm": 0.7878817319869995, "learning_rate": 1.2487532203362576e-05, "loss": 0.7405, "num_input_tokens_seen": 36146992, "step": 26675 }, { "epoch": 0.8539786185263427, "grad_norm": 1.8124516010284424, "learning_rate": 1.247903381249155e-05, "loss": 0.576, "num_input_tokens_seen": 36153936, "step": 26680 }, { "epoch": 0.8541386594968312, "grad_norm": 1.2497133016586304, "learning_rate": 1.2470537352510853e-05, "loss": 0.5304, "num_input_tokens_seen": 36160992, "step": 26685 }, { "epoch": 0.8542987004673196, "grad_norm": 0.9209546446800232, "learning_rate": 1.2462042824730758e-05, "loss": 0.2886, "num_input_tokens_seen": 36167712, "step": 26690 }, { "epoch": 0.8544587414378081, "grad_norm": 1.308232307434082, "learning_rate": 1.245355023046122e-05, "loss": 0.4473, "num_input_tokens_seen": 36174528, "step": 26695 }, { "epoch": 0.8546187824082965, "grad_norm": 1.215441346168518, "learning_rate": 1.2445059571011896e-05, "loss": 0.3286, "num_input_tokens_seen": 36181264, "step": 26700 }, { "epoch": 0.854778823378785, "grad_norm": 0.9542867541313171, "learning_rate": 1.2436570847692173e-05, "loss": 0.4173, "num_input_tokens_seen": 36187872, "step": 26705 }, { "epoch": 0.8549388643492735, "grad_norm": 1.7727339267730713, "learning_rate": 1.2428084061811096e-05, "loss": 0.4897, "num_input_tokens_seen": 36194352, "step": 26710 }, { "epoch": 0.8550989053197618, "grad_norm": 0.6187965273857117, "learning_rate": 1.2419599214677447e-05, "loss": 0.3505, "num_input_tokens_seen": 36201168, "step": 26715 }, { "epoch": 0.8552589462902503, "grad_norm": 2.9853663444519043, "learning_rate": 1.2411116307599702e-05, "loss": 0.6473, "num_input_tokens_seen": 36207712, "step": 26720 }, { "epoch": 0.8554189872607387, "grad_norm": 1.4117766618728638, "learning_rate": 1.2402635341886016e-05, "loss": 0.3763, "num_input_tokens_seen": 36214688, "step": 26725 }, { "epoch": 0.8555790282312272, "grad_norm": 0.923589825630188, "learning_rate": 1.2394156318844278e-05, "loss": 0.3865, "num_input_tokens_seen": 36221424, "step": 26730 }, { "epoch": 0.8557390692017156, "grad_norm": 1.0391427278518677, "learning_rate": 1.2385679239782039e-05, "loss": 0.3844, "num_input_tokens_seen": 36227728, "step": 26735 }, { "epoch": 0.8558991101722041, "grad_norm": 1.4685474634170532, "learning_rate": 1.2377204106006585e-05, "loss": 0.4596, "num_input_tokens_seen": 36234624, "step": 26740 }, { "epoch": 0.8560591511426925, "grad_norm": 1.0669506788253784, "learning_rate": 1.2368730918824891e-05, "loss": 0.298, "num_input_tokens_seen": 36241360, "step": 26745 }, { "epoch": 0.856219192113181, "grad_norm": 0.9823153018951416, "learning_rate": 1.236025967954362e-05, "loss": 0.4829, "num_input_tokens_seen": 36248608, "step": 26750 }, { "epoch": 0.8563792330836694, "grad_norm": 1.6871843338012695, "learning_rate": 1.2351790389469153e-05, "loss": 0.5617, "num_input_tokens_seen": 36255056, "step": 26755 }, { "epoch": 0.8565392740541579, "grad_norm": 1.2095954418182373, "learning_rate": 1.234332304990755e-05, "loss": 0.4143, "num_input_tokens_seen": 36261328, "step": 26760 }, { "epoch": 0.8566993150246464, "grad_norm": 0.6884724497795105, "learning_rate": 1.2334857662164593e-05, "loss": 0.5273, "num_input_tokens_seen": 36268224, "step": 26765 }, { "epoch": 0.8568593559951347, "grad_norm": 1.816605806350708, "learning_rate": 1.2326394227545743e-05, "loss": 0.5663, "num_input_tokens_seen": 36274880, "step": 26770 }, { "epoch": 0.8570193969656232, "grad_norm": 1.5426048040390015, "learning_rate": 1.2317932747356162e-05, "loss": 0.4625, "num_input_tokens_seen": 36281824, "step": 26775 }, { "epoch": 0.8571794379361116, "grad_norm": 1.174087405204773, "learning_rate": 1.2309473222900726e-05, "loss": 0.4891, "num_input_tokens_seen": 36288560, "step": 26780 }, { "epoch": 0.8573394789066001, "grad_norm": 0.9136653542518616, "learning_rate": 1.2301015655484006e-05, "loss": 0.4944, "num_input_tokens_seen": 36295232, "step": 26785 }, { "epoch": 0.8574995198770885, "grad_norm": 0.7457116842269897, "learning_rate": 1.2292560046410245e-05, "loss": 0.2864, "num_input_tokens_seen": 36302416, "step": 26790 }, { "epoch": 0.857659560847577, "grad_norm": 0.5541020631790161, "learning_rate": 1.228410639698343e-05, "loss": 0.3342, "num_input_tokens_seen": 36309216, "step": 26795 }, { "epoch": 0.8578196018180654, "grad_norm": 1.3721375465393066, "learning_rate": 1.2275654708507195e-05, "loss": 0.5572, "num_input_tokens_seen": 36315872, "step": 26800 }, { "epoch": 0.8578196018180654, "eval_loss": 0.449122816324234, "eval_runtime": 503.5168, "eval_samples_per_second": 27.578, "eval_steps_per_second": 13.789, "num_input_tokens_seen": 36315872, "step": 26800 }, { "epoch": 0.8579796427885539, "grad_norm": 1.2162848711013794, "learning_rate": 1.2267204982284908e-05, "loss": 0.5856, "num_input_tokens_seen": 36322464, "step": 26805 }, { "epoch": 0.8581396837590423, "grad_norm": 1.8109462261199951, "learning_rate": 1.2258757219619635e-05, "loss": 0.4822, "num_input_tokens_seen": 36329168, "step": 26810 }, { "epoch": 0.8582997247295308, "grad_norm": 1.0816696882247925, "learning_rate": 1.2250311421814104e-05, "loss": 0.3691, "num_input_tokens_seen": 36336352, "step": 26815 }, { "epoch": 0.8584597657000193, "grad_norm": 1.3267526626586914, "learning_rate": 1.2241867590170772e-05, "loss": 0.4219, "num_input_tokens_seen": 36343232, "step": 26820 }, { "epoch": 0.8586198066705076, "grad_norm": 1.0834532976150513, "learning_rate": 1.2233425725991799e-05, "loss": 0.7723, "num_input_tokens_seen": 36350240, "step": 26825 }, { "epoch": 0.8587798476409961, "grad_norm": 1.9266515970230103, "learning_rate": 1.2224985830579003e-05, "loss": 0.463, "num_input_tokens_seen": 36356992, "step": 26830 }, { "epoch": 0.8589398886114845, "grad_norm": 2.344548463821411, "learning_rate": 1.2216547905233944e-05, "loss": 0.5909, "num_input_tokens_seen": 36363488, "step": 26835 }, { "epoch": 0.859099929581973, "grad_norm": 0.2828824818134308, "learning_rate": 1.2208111951257842e-05, "loss": 0.3661, "num_input_tokens_seen": 36370576, "step": 26840 }, { "epoch": 0.8592599705524614, "grad_norm": 2.8714048862457275, "learning_rate": 1.2199677969951622e-05, "loss": 0.5749, "num_input_tokens_seen": 36377840, "step": 26845 }, { "epoch": 0.8594200115229499, "grad_norm": 0.8391273617744446, "learning_rate": 1.2191245962615927e-05, "loss": 0.3429, "num_input_tokens_seen": 36384624, "step": 26850 }, { "epoch": 0.8595800524934383, "grad_norm": 0.7368427515029907, "learning_rate": 1.218281593055106e-05, "loss": 0.2864, "num_input_tokens_seen": 36391504, "step": 26855 }, { "epoch": 0.8597400934639268, "grad_norm": 0.9111384749412537, "learning_rate": 1.217438787505705e-05, "loss": 0.407, "num_input_tokens_seen": 36398032, "step": 26860 }, { "epoch": 0.8599001344344153, "grad_norm": 1.2656930685043335, "learning_rate": 1.2165961797433615e-05, "loss": 0.4518, "num_input_tokens_seen": 36404880, "step": 26865 }, { "epoch": 0.8600601754049036, "grad_norm": 0.6806905269622803, "learning_rate": 1.215753769898014e-05, "loss": 0.4727, "num_input_tokens_seen": 36411360, "step": 26870 }, { "epoch": 0.8602202163753921, "grad_norm": 0.6741381287574768, "learning_rate": 1.2149115580995755e-05, "loss": 0.4275, "num_input_tokens_seen": 36418000, "step": 26875 }, { "epoch": 0.8603802573458805, "grad_norm": 0.8296957612037659, "learning_rate": 1.2140695444779227e-05, "loss": 0.2772, "num_input_tokens_seen": 36424960, "step": 26880 }, { "epoch": 0.860540298316369, "grad_norm": 1.8219162225723267, "learning_rate": 1.2132277291629066e-05, "loss": 0.415, "num_input_tokens_seen": 36432096, "step": 26885 }, { "epoch": 0.8607003392868574, "grad_norm": 1.1920270919799805, "learning_rate": 1.2123861122843458e-05, "loss": 0.5508, "num_input_tokens_seen": 36439024, "step": 26890 }, { "epoch": 0.8608603802573459, "grad_norm": 0.914595901966095, "learning_rate": 1.2115446939720271e-05, "loss": 0.3449, "num_input_tokens_seen": 36446176, "step": 26895 }, { "epoch": 0.8610204212278343, "grad_norm": 0.9425821900367737, "learning_rate": 1.210703474355708e-05, "loss": 0.37, "num_input_tokens_seen": 36452896, "step": 26900 }, { "epoch": 0.8611804621983228, "grad_norm": 0.6695361733436584, "learning_rate": 1.2098624535651164e-05, "loss": 0.4469, "num_input_tokens_seen": 36459456, "step": 26905 }, { "epoch": 0.8613405031688112, "grad_norm": 1.0786678791046143, "learning_rate": 1.2090216317299477e-05, "loss": 0.3873, "num_input_tokens_seen": 36466240, "step": 26910 }, { "epoch": 0.8615005441392997, "grad_norm": 1.3967851400375366, "learning_rate": 1.2081810089798668e-05, "loss": 0.5912, "num_input_tokens_seen": 36472848, "step": 26915 }, { "epoch": 0.8616605851097882, "grad_norm": 2.3500194549560547, "learning_rate": 1.2073405854445072e-05, "loss": 0.4146, "num_input_tokens_seen": 36479792, "step": 26920 }, { "epoch": 0.8618206260802765, "grad_norm": 0.5053597688674927, "learning_rate": 1.206500361253474e-05, "loss": 0.3589, "num_input_tokens_seen": 36486544, "step": 26925 }, { "epoch": 0.861980667050765, "grad_norm": 0.5712835788726807, "learning_rate": 1.2056603365363409e-05, "loss": 0.3266, "num_input_tokens_seen": 36493152, "step": 26930 }, { "epoch": 0.8621407080212534, "grad_norm": 0.70162433385849, "learning_rate": 1.2048205114226487e-05, "loss": 0.434, "num_input_tokens_seen": 36500256, "step": 26935 }, { "epoch": 0.8623007489917419, "grad_norm": 1.254035234451294, "learning_rate": 1.2039808860419102e-05, "loss": 0.3054, "num_input_tokens_seen": 36507536, "step": 26940 }, { "epoch": 0.8624607899622303, "grad_norm": 0.42864277958869934, "learning_rate": 1.2031414605236066e-05, "loss": 0.3419, "num_input_tokens_seen": 36514624, "step": 26945 }, { "epoch": 0.8626208309327188, "grad_norm": 1.0280373096466064, "learning_rate": 1.2023022349971862e-05, "loss": 0.5221, "num_input_tokens_seen": 36521504, "step": 26950 }, { "epoch": 0.8627808719032072, "grad_norm": 1.512273907661438, "learning_rate": 1.20146320959207e-05, "loss": 0.4019, "num_input_tokens_seen": 36528384, "step": 26955 }, { "epoch": 0.8629409128736957, "grad_norm": 0.6891092658042908, "learning_rate": 1.2006243844376445e-05, "loss": 0.5036, "num_input_tokens_seen": 36534864, "step": 26960 }, { "epoch": 0.8631009538441841, "grad_norm": 0.6790077686309814, "learning_rate": 1.1997857596632678e-05, "loss": 0.5023, "num_input_tokens_seen": 36541792, "step": 26965 }, { "epoch": 0.8632609948146726, "grad_norm": 0.6911222338676453, "learning_rate": 1.1989473353982672e-05, "loss": 0.4079, "num_input_tokens_seen": 36548464, "step": 26970 }, { "epoch": 0.863421035785161, "grad_norm": 0.8421101570129395, "learning_rate": 1.198109111771937e-05, "loss": 0.4885, "num_input_tokens_seen": 36555360, "step": 26975 }, { "epoch": 0.8635810767556494, "grad_norm": 0.7379137277603149, "learning_rate": 1.197271088913543e-05, "loss": 0.5936, "num_input_tokens_seen": 36562256, "step": 26980 }, { "epoch": 0.8637411177261379, "grad_norm": 0.9360712766647339, "learning_rate": 1.1964332669523182e-05, "loss": 0.4427, "num_input_tokens_seen": 36569600, "step": 26985 }, { "epoch": 0.8639011586966263, "grad_norm": 1.7769428491592407, "learning_rate": 1.1955956460174645e-05, "loss": 0.4319, "num_input_tokens_seen": 36576400, "step": 26990 }, { "epoch": 0.8640611996671148, "grad_norm": 1.0358775854110718, "learning_rate": 1.1947582262381552e-05, "loss": 0.4336, "num_input_tokens_seen": 36583248, "step": 26995 }, { "epoch": 0.8642212406376032, "grad_norm": 0.8951472640037537, "learning_rate": 1.1939210077435293e-05, "loss": 0.5925, "num_input_tokens_seen": 36590336, "step": 27000 }, { "epoch": 0.8642212406376032, "eval_loss": 0.4487477242946625, "eval_runtime": 502.8649, "eval_samples_per_second": 27.614, "eval_steps_per_second": 13.807, "num_input_tokens_seen": 36590336, "step": 27000 }, { "epoch": 0.8643812816080917, "grad_norm": 1.5739692449569702, "learning_rate": 1.193083990662697e-05, "loss": 0.5265, "num_input_tokens_seen": 36596720, "step": 27005 }, { "epoch": 0.8645413225785801, "grad_norm": 2.2499876022338867, "learning_rate": 1.192247175124738e-05, "loss": 0.5342, "num_input_tokens_seen": 36603088, "step": 27010 }, { "epoch": 0.8647013635490686, "grad_norm": 1.6708104610443115, "learning_rate": 1.191410561258698e-05, "loss": 0.4483, "num_input_tokens_seen": 36609504, "step": 27015 }, { "epoch": 0.864861404519557, "grad_norm": 0.4231661260128021, "learning_rate": 1.1905741491935944e-05, "loss": 0.4631, "num_input_tokens_seen": 36616336, "step": 27020 }, { "epoch": 0.8650214454900454, "grad_norm": 1.0590406656265259, "learning_rate": 1.1897379390584129e-05, "loss": 0.3914, "num_input_tokens_seen": 36623216, "step": 27025 }, { "epoch": 0.8651814864605339, "grad_norm": 0.9172616600990295, "learning_rate": 1.1889019309821062e-05, "loss": 0.4284, "num_input_tokens_seen": 36630032, "step": 27030 }, { "epoch": 0.8653415274310223, "grad_norm": 0.7870034575462341, "learning_rate": 1.188066125093599e-05, "loss": 0.3071, "num_input_tokens_seen": 36636880, "step": 27035 }, { "epoch": 0.8655015684015108, "grad_norm": 1.2334054708480835, "learning_rate": 1.1872305215217811e-05, "loss": 0.4244, "num_input_tokens_seen": 36643312, "step": 27040 }, { "epoch": 0.8656616093719992, "grad_norm": 2.453392267227173, "learning_rate": 1.186395120395514e-05, "loss": 0.3975, "num_input_tokens_seen": 36649952, "step": 27045 }, { "epoch": 0.8658216503424877, "grad_norm": 1.7396773099899292, "learning_rate": 1.1855599218436283e-05, "loss": 0.4912, "num_input_tokens_seen": 36657024, "step": 27050 }, { "epoch": 0.8659816913129761, "grad_norm": 0.5438675880432129, "learning_rate": 1.1847249259949209e-05, "loss": 0.263, "num_input_tokens_seen": 36663664, "step": 27055 }, { "epoch": 0.8661417322834646, "grad_norm": 1.5180739164352417, "learning_rate": 1.1838901329781574e-05, "loss": 0.5217, "num_input_tokens_seen": 36670384, "step": 27060 }, { "epoch": 0.866301773253953, "grad_norm": 0.8423869013786316, "learning_rate": 1.1830555429220758e-05, "loss": 0.6859, "num_input_tokens_seen": 36676768, "step": 27065 }, { "epoch": 0.8664618142244415, "grad_norm": 1.245092749595642, "learning_rate": 1.1822211559553784e-05, "loss": 0.4915, "num_input_tokens_seen": 36683328, "step": 27070 }, { "epoch": 0.86662185519493, "grad_norm": 1.1192084550857544, "learning_rate": 1.18138697220674e-05, "loss": 0.431, "num_input_tokens_seen": 36689984, "step": 27075 }, { "epoch": 0.8667818961654183, "grad_norm": 1.1449657678604126, "learning_rate": 1.1805529918048e-05, "loss": 0.4213, "num_input_tokens_seen": 36696544, "step": 27080 }, { "epoch": 0.8669419371359068, "grad_norm": 2.5009231567382812, "learning_rate": 1.1797192148781702e-05, "loss": 0.4765, "num_input_tokens_seen": 36703072, "step": 27085 }, { "epoch": 0.8671019781063952, "grad_norm": 1.3071098327636719, "learning_rate": 1.1788856415554297e-05, "loss": 0.5674, "num_input_tokens_seen": 36709632, "step": 27090 }, { "epoch": 0.8672620190768837, "grad_norm": 0.846396803855896, "learning_rate": 1.1780522719651249e-05, "loss": 0.3194, "num_input_tokens_seen": 36716960, "step": 27095 }, { "epoch": 0.8674220600473721, "grad_norm": 1.6878865957260132, "learning_rate": 1.1772191062357721e-05, "loss": 0.534, "num_input_tokens_seen": 36723584, "step": 27100 }, { "epoch": 0.8675821010178606, "grad_norm": 0.8932645320892334, "learning_rate": 1.1763861444958573e-05, "loss": 0.455, "num_input_tokens_seen": 36729968, "step": 27105 }, { "epoch": 0.867742141988349, "grad_norm": 2.0219357013702393, "learning_rate": 1.1755533868738317e-05, "loss": 0.3826, "num_input_tokens_seen": 36736944, "step": 27110 }, { "epoch": 0.8679021829588375, "grad_norm": 1.0444616079330444, "learning_rate": 1.1747208334981185e-05, "loss": 0.3599, "num_input_tokens_seen": 36743792, "step": 27115 }, { "epoch": 0.8680622239293259, "grad_norm": 0.7170119285583496, "learning_rate": 1.1738884844971067e-05, "loss": 0.5191, "num_input_tokens_seen": 36750432, "step": 27120 }, { "epoch": 0.8682222648998144, "grad_norm": 1.078748345375061, "learning_rate": 1.1730563399991563e-05, "loss": 0.3506, "num_input_tokens_seen": 36757216, "step": 27125 }, { "epoch": 0.8683823058703029, "grad_norm": 1.7223236560821533, "learning_rate": 1.1722244001325938e-05, "loss": 0.6554, "num_input_tokens_seen": 36764480, "step": 27130 }, { "epoch": 0.8685423468407912, "grad_norm": 1.3745828866958618, "learning_rate": 1.1713926650257137e-05, "loss": 0.3133, "num_input_tokens_seen": 36770688, "step": 27135 }, { "epoch": 0.8687023878112797, "grad_norm": 0.7210537195205688, "learning_rate": 1.170561134806781e-05, "loss": 0.403, "num_input_tokens_seen": 36777760, "step": 27140 }, { "epoch": 0.8688624287817681, "grad_norm": 1.6497156620025635, "learning_rate": 1.1697298096040287e-05, "loss": 0.5784, "num_input_tokens_seen": 36784464, "step": 27145 }, { "epoch": 0.8690224697522566, "grad_norm": 0.7046680450439453, "learning_rate": 1.1688986895456567e-05, "loss": 0.4555, "num_input_tokens_seen": 36791216, "step": 27150 }, { "epoch": 0.869182510722745, "grad_norm": 0.97605299949646, "learning_rate": 1.1680677747598349e-05, "loss": 0.5043, "num_input_tokens_seen": 36797824, "step": 27155 }, { "epoch": 0.8693425516932335, "grad_norm": 0.9169461727142334, "learning_rate": 1.1672370653746995e-05, "loss": 0.555, "num_input_tokens_seen": 36804768, "step": 27160 }, { "epoch": 0.8695025926637219, "grad_norm": 1.7800122499465942, "learning_rate": 1.166406561518357e-05, "loss": 0.638, "num_input_tokens_seen": 36811120, "step": 27165 }, { "epoch": 0.8696626336342104, "grad_norm": 1.1462585926055908, "learning_rate": 1.1655762633188826e-05, "loss": 0.565, "num_input_tokens_seen": 36817712, "step": 27170 }, { "epoch": 0.8698226746046988, "grad_norm": 0.9153957366943359, "learning_rate": 1.1647461709043172e-05, "loss": 0.3794, "num_input_tokens_seen": 36824880, "step": 27175 }, { "epoch": 0.8699827155751872, "grad_norm": 0.8228662014007568, "learning_rate": 1.1639162844026722e-05, "loss": 0.3273, "num_input_tokens_seen": 36831728, "step": 27180 }, { "epoch": 0.8701427565456757, "grad_norm": 1.849541425704956, "learning_rate": 1.163086603941927e-05, "loss": 0.4695, "num_input_tokens_seen": 36838192, "step": 27185 }, { "epoch": 0.8703027975161641, "grad_norm": 0.7762446403503418, "learning_rate": 1.1622571296500273e-05, "loss": 0.3572, "num_input_tokens_seen": 36844848, "step": 27190 }, { "epoch": 0.8704628384866526, "grad_norm": 1.2121782302856445, "learning_rate": 1.1614278616548904e-05, "loss": 0.3954, "num_input_tokens_seen": 36851392, "step": 27195 }, { "epoch": 0.870622879457141, "grad_norm": 0.7130492329597473, "learning_rate": 1.1605988000843986e-05, "loss": 0.3107, "num_input_tokens_seen": 36858080, "step": 27200 }, { "epoch": 0.870622879457141, "eval_loss": 0.44856885075569153, "eval_runtime": 502.7278, "eval_samples_per_second": 27.621, "eval_steps_per_second": 13.811, "num_input_tokens_seen": 36858080, "step": 27200 }, { "epoch": 0.8707829204276295, "grad_norm": 0.9871045351028442, "learning_rate": 1.1597699450664028e-05, "loss": 0.4023, "num_input_tokens_seen": 36864640, "step": 27205 }, { "epoch": 0.8709429613981179, "grad_norm": 1.1157474517822266, "learning_rate": 1.1589412967287252e-05, "loss": 0.5627, "num_input_tokens_seen": 36871392, "step": 27210 }, { "epoch": 0.8711030023686064, "grad_norm": 0.9548100233078003, "learning_rate": 1.1581128551991514e-05, "loss": 0.3866, "num_input_tokens_seen": 36878176, "step": 27215 }, { "epoch": 0.8712630433390948, "grad_norm": 0.9648100733757019, "learning_rate": 1.1572846206054383e-05, "loss": 0.4799, "num_input_tokens_seen": 36885152, "step": 27220 }, { "epoch": 0.8714230843095833, "grad_norm": 1.3638347387313843, "learning_rate": 1.1564565930753113e-05, "loss": 0.4055, "num_input_tokens_seen": 36891616, "step": 27225 }, { "epoch": 0.8715831252800716, "grad_norm": 1.538414716720581, "learning_rate": 1.1556287727364606e-05, "loss": 0.4809, "num_input_tokens_seen": 36898672, "step": 27230 }, { "epoch": 0.8717431662505601, "grad_norm": 2.074052095413208, "learning_rate": 1.1548011597165489e-05, "loss": 0.5264, "num_input_tokens_seen": 36905840, "step": 27235 }, { "epoch": 0.8719032072210486, "grad_norm": 0.750031590461731, "learning_rate": 1.1539737541432019e-05, "loss": 0.3399, "num_input_tokens_seen": 36912672, "step": 27240 }, { "epoch": 0.872063248191537, "grad_norm": 0.9083012342453003, "learning_rate": 1.1531465561440174e-05, "loss": 0.5642, "num_input_tokens_seen": 36919264, "step": 27245 }, { "epoch": 0.8722232891620255, "grad_norm": 0.9028396606445312, "learning_rate": 1.1523195658465605e-05, "loss": 0.2877, "num_input_tokens_seen": 36925760, "step": 27250 }, { "epoch": 0.8723833301325139, "grad_norm": 1.2721507549285889, "learning_rate": 1.1514927833783618e-05, "loss": 0.4464, "num_input_tokens_seen": 36932224, "step": 27255 }, { "epoch": 0.8725433711030024, "grad_norm": 1.448289394378662, "learning_rate": 1.150666208866922e-05, "loss": 0.4968, "num_input_tokens_seen": 36938880, "step": 27260 }, { "epoch": 0.8727034120734908, "grad_norm": 0.9839437007904053, "learning_rate": 1.1498398424397106e-05, "loss": 0.3199, "num_input_tokens_seen": 36945392, "step": 27265 }, { "epoch": 0.8728634530439793, "grad_norm": 2.1645939350128174, "learning_rate": 1.1490136842241628e-05, "loss": 0.4604, "num_input_tokens_seen": 36951936, "step": 27270 }, { "epoch": 0.8730234940144677, "grad_norm": 1.6482384204864502, "learning_rate": 1.1481877343476813e-05, "loss": 0.5721, "num_input_tokens_seen": 36958768, "step": 27275 }, { "epoch": 0.8731835349849562, "grad_norm": 0.5316797494888306, "learning_rate": 1.14736199293764e-05, "loss": 0.5007, "num_input_tokens_seen": 36965552, "step": 27280 }, { "epoch": 0.8733435759554445, "grad_norm": 0.5267472863197327, "learning_rate": 1.1465364601213771e-05, "loss": 0.4004, "num_input_tokens_seen": 36972336, "step": 27285 }, { "epoch": 0.873503616925933, "grad_norm": 0.8713369965553284, "learning_rate": 1.1457111360262012e-05, "loss": 0.4665, "num_input_tokens_seen": 36978832, "step": 27290 }, { "epoch": 0.8736636578964215, "grad_norm": 1.2984884977340698, "learning_rate": 1.1448860207793869e-05, "loss": 0.5222, "num_input_tokens_seen": 36985424, "step": 27295 }, { "epoch": 0.8738236988669099, "grad_norm": 0.7935431003570557, "learning_rate": 1.144061114508177e-05, "loss": 0.354, "num_input_tokens_seen": 36992240, "step": 27300 }, { "epoch": 0.8739837398373984, "grad_norm": 0.9421049356460571, "learning_rate": 1.1432364173397842e-05, "loss": 0.4682, "num_input_tokens_seen": 36998976, "step": 27305 }, { "epoch": 0.8741437808078868, "grad_norm": 0.9458938241004944, "learning_rate": 1.1424119294013852e-05, "loss": 0.4954, "num_input_tokens_seen": 37005584, "step": 27310 }, { "epoch": 0.8743038217783753, "grad_norm": 0.7435742616653442, "learning_rate": 1.1415876508201279e-05, "loss": 0.3584, "num_input_tokens_seen": 37012336, "step": 27315 }, { "epoch": 0.8744638627488637, "grad_norm": 1.5317599773406982, "learning_rate": 1.140763581723125e-05, "loss": 0.4902, "num_input_tokens_seen": 37019184, "step": 27320 }, { "epoch": 0.8746239037193522, "grad_norm": 1.7208186388015747, "learning_rate": 1.1399397222374588e-05, "loss": 0.3791, "num_input_tokens_seen": 37026288, "step": 27325 }, { "epoch": 0.8747839446898406, "grad_norm": 0.5259827375411987, "learning_rate": 1.1391160724901804e-05, "loss": 0.3115, "num_input_tokens_seen": 37032752, "step": 27330 }, { "epoch": 0.874943985660329, "grad_norm": 0.8727607727050781, "learning_rate": 1.138292632608304e-05, "loss": 0.421, "num_input_tokens_seen": 37039216, "step": 27335 }, { "epoch": 0.8751040266308175, "grad_norm": 0.3841390907764435, "learning_rate": 1.1374694027188174e-05, "loss": 0.3353, "num_input_tokens_seen": 37046064, "step": 27340 }, { "epoch": 0.8752640676013059, "grad_norm": 1.7420356273651123, "learning_rate": 1.1366463829486711e-05, "loss": 0.3517, "num_input_tokens_seen": 37052576, "step": 27345 }, { "epoch": 0.8754241085717944, "grad_norm": 0.5459234714508057, "learning_rate": 1.1358235734247849e-05, "loss": 0.2929, "num_input_tokens_seen": 37059200, "step": 27350 }, { "epoch": 0.8755841495422828, "grad_norm": 1.4723687171936035, "learning_rate": 1.1350009742740478e-05, "loss": 0.481, "num_input_tokens_seen": 37065824, "step": 27355 }, { "epoch": 0.8757441905127713, "grad_norm": 1.0642058849334717, "learning_rate": 1.134178585623313e-05, "loss": 0.3383, "num_input_tokens_seen": 37072336, "step": 27360 }, { "epoch": 0.8759042314832597, "grad_norm": 1.4033042192459106, "learning_rate": 1.1333564075994047e-05, "loss": 0.3267, "num_input_tokens_seen": 37078816, "step": 27365 }, { "epoch": 0.8760642724537482, "grad_norm": 1.1637858152389526, "learning_rate": 1.1325344403291133e-05, "loss": 0.5743, "num_input_tokens_seen": 37085504, "step": 27370 }, { "epoch": 0.8762243134242366, "grad_norm": 1.5868821144104004, "learning_rate": 1.1317126839391951e-05, "loss": 0.3125, "num_input_tokens_seen": 37091824, "step": 27375 }, { "epoch": 0.8763843543947251, "grad_norm": 0.9351597428321838, "learning_rate": 1.1308911385563766e-05, "loss": 0.4006, "num_input_tokens_seen": 37098624, "step": 27380 }, { "epoch": 0.8765443953652134, "grad_norm": 0.6468963623046875, "learning_rate": 1.1300698043073494e-05, "loss": 0.465, "num_input_tokens_seen": 37105296, "step": 27385 }, { "epoch": 0.8767044363357019, "grad_norm": 0.8130002617835999, "learning_rate": 1.1292486813187736e-05, "loss": 0.463, "num_input_tokens_seen": 37111888, "step": 27390 }, { "epoch": 0.8768644773061904, "grad_norm": 1.9455331563949585, "learning_rate": 1.1284277697172782e-05, "loss": 0.472, "num_input_tokens_seen": 37118448, "step": 27395 }, { "epoch": 0.8770245182766788, "grad_norm": 1.0061167478561401, "learning_rate": 1.127607069629456e-05, "loss": 0.441, "num_input_tokens_seen": 37125216, "step": 27400 }, { "epoch": 0.8770245182766788, "eval_loss": 0.4482533633708954, "eval_runtime": 502.8135, "eval_samples_per_second": 27.617, "eval_steps_per_second": 13.808, "num_input_tokens_seen": 37125216, "step": 27400 }, { "epoch": 0.8771845592471673, "grad_norm": 0.7468124628067017, "learning_rate": 1.1267865811818701e-05, "loss": 0.608, "num_input_tokens_seen": 37132288, "step": 27405 }, { "epoch": 0.8773446002176557, "grad_norm": 0.9082098603248596, "learning_rate": 1.1259663045010513e-05, "loss": 0.653, "num_input_tokens_seen": 37139056, "step": 27410 }, { "epoch": 0.8775046411881442, "grad_norm": 1.1705461740493774, "learning_rate": 1.1251462397134957e-05, "loss": 0.3241, "num_input_tokens_seen": 37145840, "step": 27415 }, { "epoch": 0.8776646821586326, "grad_norm": 2.2039904594421387, "learning_rate": 1.1243263869456664e-05, "loss": 0.4457, "num_input_tokens_seen": 37152432, "step": 27420 }, { "epoch": 0.8778247231291211, "grad_norm": 1.740042805671692, "learning_rate": 1.1235067463239967e-05, "loss": 0.3629, "num_input_tokens_seen": 37159280, "step": 27425 }, { "epoch": 0.8779847640996095, "grad_norm": 1.3924862146377563, "learning_rate": 1.122687317974884e-05, "loss": 0.6707, "num_input_tokens_seen": 37166160, "step": 27430 }, { "epoch": 0.878144805070098, "grad_norm": 1.9281939268112183, "learning_rate": 1.1218681020246963e-05, "loss": 0.4888, "num_input_tokens_seen": 37173184, "step": 27435 }, { "epoch": 0.8783048460405863, "grad_norm": 0.8583048582077026, "learning_rate": 1.1210490985997652e-05, "loss": 0.5073, "num_input_tokens_seen": 37180272, "step": 27440 }, { "epoch": 0.8784648870110748, "grad_norm": 1.4619991779327393, "learning_rate": 1.1202303078263917e-05, "loss": 0.5512, "num_input_tokens_seen": 37187104, "step": 27445 }, { "epoch": 0.8786249279815633, "grad_norm": 0.6075193881988525, "learning_rate": 1.1194117298308451e-05, "loss": 0.3509, "num_input_tokens_seen": 37193776, "step": 27450 }, { "epoch": 0.8787849689520517, "grad_norm": 1.1808745861053467, "learning_rate": 1.1185933647393585e-05, "loss": 0.4671, "num_input_tokens_seen": 37200672, "step": 27455 }, { "epoch": 0.8789450099225402, "grad_norm": 0.6358294486999512, "learning_rate": 1.1177752126781354e-05, "loss": 0.5121, "num_input_tokens_seen": 37207360, "step": 27460 }, { "epoch": 0.8791050508930286, "grad_norm": 1.5624879598617554, "learning_rate": 1.1169572737733441e-05, "loss": 0.5273, "num_input_tokens_seen": 37214304, "step": 27465 }, { "epoch": 0.8792650918635171, "grad_norm": 0.915926992893219, "learning_rate": 1.1161395481511216e-05, "loss": 0.3709, "num_input_tokens_seen": 37220928, "step": 27470 }, { "epoch": 0.8794251328340055, "grad_norm": 1.4066784381866455, "learning_rate": 1.1153220359375722e-05, "loss": 0.7509, "num_input_tokens_seen": 37227920, "step": 27475 }, { "epoch": 0.879585173804494, "grad_norm": 0.9820419549942017, "learning_rate": 1.114504737258765e-05, "loss": 0.5403, "num_input_tokens_seen": 37234864, "step": 27480 }, { "epoch": 0.8797452147749824, "grad_norm": 1.566272258758545, "learning_rate": 1.1136876522407393e-05, "loss": 0.5399, "num_input_tokens_seen": 37241536, "step": 27485 }, { "epoch": 0.8799052557454708, "grad_norm": 0.70066899061203, "learning_rate": 1.1128707810094985e-05, "loss": 0.4004, "num_input_tokens_seen": 37247968, "step": 27490 }, { "epoch": 0.8800652967159592, "grad_norm": 1.6954975128173828, "learning_rate": 1.1120541236910157e-05, "loss": 0.5081, "num_input_tokens_seen": 37254384, "step": 27495 }, { "epoch": 0.8802253376864477, "grad_norm": 1.084061861038208, "learning_rate": 1.111237680411229e-05, "loss": 0.4052, "num_input_tokens_seen": 37260832, "step": 27500 }, { "epoch": 0.8803853786569362, "grad_norm": 1.214590311050415, "learning_rate": 1.1104214512960433e-05, "loss": 0.4357, "num_input_tokens_seen": 37267952, "step": 27505 }, { "epoch": 0.8805454196274246, "grad_norm": 0.5086069703102112, "learning_rate": 1.1096054364713327e-05, "loss": 0.3567, "num_input_tokens_seen": 37274672, "step": 27510 }, { "epoch": 0.8807054605979131, "grad_norm": 0.8737255930900574, "learning_rate": 1.1087896360629371e-05, "loss": 0.512, "num_input_tokens_seen": 37281584, "step": 27515 }, { "epoch": 0.8808655015684015, "grad_norm": 0.7574036121368408, "learning_rate": 1.107974050196662e-05, "loss": 0.3041, "num_input_tokens_seen": 37288208, "step": 27520 }, { "epoch": 0.88102554253889, "grad_norm": 1.3396276235580444, "learning_rate": 1.1071586789982816e-05, "loss": 0.4873, "num_input_tokens_seen": 37294944, "step": 27525 }, { "epoch": 0.8811855835093784, "grad_norm": 1.0439802408218384, "learning_rate": 1.1063435225935373e-05, "loss": 0.383, "num_input_tokens_seen": 37301808, "step": 27530 }, { "epoch": 0.8813456244798669, "grad_norm": 0.9531407356262207, "learning_rate": 1.1055285811081348e-05, "loss": 0.4575, "num_input_tokens_seen": 37308528, "step": 27535 }, { "epoch": 0.8815056654503552, "grad_norm": 1.3769832849502563, "learning_rate": 1.1047138546677499e-05, "loss": 0.5177, "num_input_tokens_seen": 37316336, "step": 27540 }, { "epoch": 0.8816657064208437, "grad_norm": 0.8803538084030151, "learning_rate": 1.1038993433980219e-05, "loss": 0.5425, "num_input_tokens_seen": 37323088, "step": 27545 }, { "epoch": 0.8818257473913321, "grad_norm": 1.2710989713668823, "learning_rate": 1.1030850474245597e-05, "loss": 0.5406, "num_input_tokens_seen": 37329504, "step": 27550 }, { "epoch": 0.8819857883618206, "grad_norm": 0.8267504572868347, "learning_rate": 1.102270966872939e-05, "loss": 0.4421, "num_input_tokens_seen": 37336432, "step": 27555 }, { "epoch": 0.8821458293323091, "grad_norm": 1.4133812189102173, "learning_rate": 1.1014571018687e-05, "loss": 0.4059, "num_input_tokens_seen": 37343184, "step": 27560 }, { "epoch": 0.8823058703027975, "grad_norm": 1.0904086828231812, "learning_rate": 1.1006434525373502e-05, "loss": 0.3325, "num_input_tokens_seen": 37350208, "step": 27565 }, { "epoch": 0.882465911273286, "grad_norm": 2.4467289447784424, "learning_rate": 1.0998300190043664e-05, "loss": 0.4845, "num_input_tokens_seen": 37357040, "step": 27570 }, { "epoch": 0.8826259522437744, "grad_norm": 0.8374495506286621, "learning_rate": 1.0990168013951882e-05, "loss": 0.3468, "num_input_tokens_seen": 37363776, "step": 27575 }, { "epoch": 0.8827859932142629, "grad_norm": 1.7369000911712646, "learning_rate": 1.0982037998352263e-05, "loss": 0.4384, "num_input_tokens_seen": 37370256, "step": 27580 }, { "epoch": 0.8829460341847513, "grad_norm": 2.320688009262085, "learning_rate": 1.0973910144498534e-05, "loss": 0.5675, "num_input_tokens_seen": 37376944, "step": 27585 }, { "epoch": 0.8831060751552398, "grad_norm": 1.3236581087112427, "learning_rate": 1.0965784453644123e-05, "loss": 0.5435, "num_input_tokens_seen": 37383952, "step": 27590 }, { "epoch": 0.8832661161257281, "grad_norm": 1.1440809965133667, "learning_rate": 1.0957660927042127e-05, "loss": 0.4228, "num_input_tokens_seen": 37390864, "step": 27595 }, { "epoch": 0.8834261570962166, "grad_norm": 1.1134992837905884, "learning_rate": 1.094953956594527e-05, "loss": 0.5021, "num_input_tokens_seen": 37397648, "step": 27600 }, { "epoch": 0.8834261570962166, "eval_loss": 0.4478852152824402, "eval_runtime": 503.0343, "eval_samples_per_second": 27.604, "eval_steps_per_second": 13.802, "num_input_tokens_seen": 37397648, "step": 27600 }, { "epoch": 0.8835861980667051, "grad_norm": 1.1003061532974243, "learning_rate": 1.0941420371605981e-05, "loss": 0.3939, "num_input_tokens_seen": 37404528, "step": 27605 }, { "epoch": 0.8837462390371935, "grad_norm": 2.2770581245422363, "learning_rate": 1.0933303345276354e-05, "loss": 0.4427, "num_input_tokens_seen": 37411296, "step": 27610 }, { "epoch": 0.883906280007682, "grad_norm": 1.5954548120498657, "learning_rate": 1.0925188488208112e-05, "loss": 0.5325, "num_input_tokens_seen": 37418064, "step": 27615 }, { "epoch": 0.8840663209781704, "grad_norm": 2.8964197635650635, "learning_rate": 1.0917075801652694e-05, "loss": 0.384, "num_input_tokens_seen": 37425008, "step": 27620 }, { "epoch": 0.8842263619486589, "grad_norm": 0.8397064805030823, "learning_rate": 1.0908965286861151e-05, "loss": 0.4445, "num_input_tokens_seen": 37431264, "step": 27625 }, { "epoch": 0.8843864029191473, "grad_norm": 1.1368988752365112, "learning_rate": 1.090085694508425e-05, "loss": 0.3255, "num_input_tokens_seen": 37438000, "step": 27630 }, { "epoch": 0.8845464438896358, "grad_norm": 0.8813546895980835, "learning_rate": 1.089275077757238e-05, "loss": 0.4181, "num_input_tokens_seen": 37444144, "step": 27635 }, { "epoch": 0.8847064848601242, "grad_norm": 1.7681704759597778, "learning_rate": 1.0884646785575633e-05, "loss": 0.5228, "num_input_tokens_seen": 37450672, "step": 27640 }, { "epoch": 0.8848665258306126, "grad_norm": 0.8534215688705444, "learning_rate": 1.0876544970343728e-05, "loss": 0.2725, "num_input_tokens_seen": 37457712, "step": 27645 }, { "epoch": 0.885026566801101, "grad_norm": 0.6031243205070496, "learning_rate": 1.0868445333126082e-05, "loss": 0.3674, "num_input_tokens_seen": 37464368, "step": 27650 }, { "epoch": 0.8851866077715895, "grad_norm": 1.4411413669586182, "learning_rate": 1.0860347875171745e-05, "loss": 0.3992, "num_input_tokens_seen": 37471152, "step": 27655 }, { "epoch": 0.885346648742078, "grad_norm": 0.6085433959960938, "learning_rate": 1.0852252597729465e-05, "loss": 0.4033, "num_input_tokens_seen": 37477744, "step": 27660 }, { "epoch": 0.8855066897125664, "grad_norm": 0.5418941974639893, "learning_rate": 1.0844159502047615e-05, "loss": 0.4644, "num_input_tokens_seen": 37484320, "step": 27665 }, { "epoch": 0.8856667306830549, "grad_norm": 1.0127078294754028, "learning_rate": 1.0836068589374265e-05, "loss": 0.5436, "num_input_tokens_seen": 37491520, "step": 27670 }, { "epoch": 0.8858267716535433, "grad_norm": 0.8287951350212097, "learning_rate": 1.0827979860957144e-05, "loss": 0.5344, "num_input_tokens_seen": 37498560, "step": 27675 }, { "epoch": 0.8859868126240318, "grad_norm": 1.4978973865509033, "learning_rate": 1.0819893318043615e-05, "loss": 0.5814, "num_input_tokens_seen": 37505392, "step": 27680 }, { "epoch": 0.8861468535945202, "grad_norm": 0.8566482067108154, "learning_rate": 1.0811808961880734e-05, "loss": 0.3909, "num_input_tokens_seen": 37512432, "step": 27685 }, { "epoch": 0.8863068945650087, "grad_norm": 1.8248610496520996, "learning_rate": 1.080372679371522e-05, "loss": 0.5058, "num_input_tokens_seen": 37519248, "step": 27690 }, { "epoch": 0.886466935535497, "grad_norm": 2.5786635875701904, "learning_rate": 1.0795646814793428e-05, "loss": 0.3651, "num_input_tokens_seen": 37525744, "step": 27695 }, { "epoch": 0.8866269765059855, "grad_norm": 1.2512470483779907, "learning_rate": 1.078756902636141e-05, "loss": 0.5888, "num_input_tokens_seen": 37532848, "step": 27700 }, { "epoch": 0.8867870174764739, "grad_norm": 1.7457586526870728, "learning_rate": 1.077949342966485e-05, "loss": 0.5273, "num_input_tokens_seen": 37539728, "step": 27705 }, { "epoch": 0.8869470584469624, "grad_norm": 2.0418860912323, "learning_rate": 1.0771420025949103e-05, "loss": 0.5254, "num_input_tokens_seen": 37547056, "step": 27710 }, { "epoch": 0.8871070994174509, "grad_norm": 0.6624867916107178, "learning_rate": 1.0763348816459204e-05, "loss": 0.6002, "num_input_tokens_seen": 37553744, "step": 27715 }, { "epoch": 0.8872671403879393, "grad_norm": 1.043891191482544, "learning_rate": 1.0755279802439816e-05, "loss": 0.587, "num_input_tokens_seen": 37560320, "step": 27720 }, { "epoch": 0.8874271813584278, "grad_norm": 1.023565411567688, "learning_rate": 1.0747212985135293e-05, "loss": 0.373, "num_input_tokens_seen": 37566944, "step": 27725 }, { "epoch": 0.8875872223289162, "grad_norm": 0.7699140906333923, "learning_rate": 1.073914836578965e-05, "loss": 0.3116, "num_input_tokens_seen": 37573712, "step": 27730 }, { "epoch": 0.8877472632994047, "grad_norm": 1.2139508724212646, "learning_rate": 1.0731085945646529e-05, "loss": 0.4839, "num_input_tokens_seen": 37580560, "step": 27735 }, { "epoch": 0.8879073042698931, "grad_norm": 0.5243580937385559, "learning_rate": 1.0723025725949285e-05, "loss": 0.2941, "num_input_tokens_seen": 37587568, "step": 27740 }, { "epoch": 0.8880673452403816, "grad_norm": 0.574816882610321, "learning_rate": 1.0714967707940875e-05, "loss": 0.3143, "num_input_tokens_seen": 37594608, "step": 27745 }, { "epoch": 0.8882273862108699, "grad_norm": 0.5788937211036682, "learning_rate": 1.0706911892863963e-05, "loss": 0.3424, "num_input_tokens_seen": 37601296, "step": 27750 }, { "epoch": 0.8883874271813584, "grad_norm": 1.5099307298660278, "learning_rate": 1.0698858281960866e-05, "loss": 0.3824, "num_input_tokens_seen": 37608000, "step": 27755 }, { "epoch": 0.8885474681518468, "grad_norm": 1.4133718013763428, "learning_rate": 1.069080687647353e-05, "loss": 0.3724, "num_input_tokens_seen": 37614384, "step": 27760 }, { "epoch": 0.8887075091223353, "grad_norm": 1.0262134075164795, "learning_rate": 1.0682757677643596e-05, "loss": 0.4607, "num_input_tokens_seen": 37620816, "step": 27765 }, { "epoch": 0.8888675500928238, "grad_norm": 1.6637871265411377, "learning_rate": 1.0674710686712359e-05, "loss": 0.3945, "num_input_tokens_seen": 37627664, "step": 27770 }, { "epoch": 0.8890275910633122, "grad_norm": 0.861086905002594, "learning_rate": 1.0666665904920756e-05, "loss": 0.3482, "num_input_tokens_seen": 37633920, "step": 27775 }, { "epoch": 0.8891876320338007, "grad_norm": 1.16426682472229, "learning_rate": 1.0658623333509385e-05, "loss": 0.4121, "num_input_tokens_seen": 37640384, "step": 27780 }, { "epoch": 0.8893476730042891, "grad_norm": 1.4560109376907349, "learning_rate": 1.0650582973718532e-05, "loss": 0.3481, "num_input_tokens_seen": 37646896, "step": 27785 }, { "epoch": 0.8895077139747776, "grad_norm": 0.6430191993713379, "learning_rate": 1.0642544826788098e-05, "loss": 0.5854, "num_input_tokens_seen": 37654064, "step": 27790 }, { "epoch": 0.889667754945266, "grad_norm": 1.3758777379989624, "learning_rate": 1.063450889395769e-05, "loss": 0.4747, "num_input_tokens_seen": 37660864, "step": 27795 }, { "epoch": 0.8898277959157544, "grad_norm": 1.4694424867630005, "learning_rate": 1.062647517646653e-05, "loss": 0.4081, "num_input_tokens_seen": 37667456, "step": 27800 }, { "epoch": 0.8898277959157544, "eval_loss": 0.4479265809059143, "eval_runtime": 503.5699, "eval_samples_per_second": 27.575, "eval_steps_per_second": 13.788, "num_input_tokens_seen": 37667456, "step": 27800 }, { "epoch": 0.8899878368862428, "grad_norm": 1.0701826810836792, "learning_rate": 1.0618443675553527e-05, "loss": 0.3332, "num_input_tokens_seen": 37673872, "step": 27805 }, { "epoch": 0.8901478778567313, "grad_norm": 0.7766188383102417, "learning_rate": 1.0610414392457247e-05, "loss": 0.502, "num_input_tokens_seen": 37680288, "step": 27810 }, { "epoch": 0.8903079188272197, "grad_norm": 2.10648512840271, "learning_rate": 1.0602387328415888e-05, "loss": 0.5232, "num_input_tokens_seen": 37687296, "step": 27815 }, { "epoch": 0.8904679597977082, "grad_norm": 1.2605239152908325, "learning_rate": 1.0594362484667347e-05, "loss": 0.647, "num_input_tokens_seen": 37694016, "step": 27820 }, { "epoch": 0.8906280007681967, "grad_norm": 0.8705422282218933, "learning_rate": 1.0586339862449132e-05, "loss": 0.4551, "num_input_tokens_seen": 37700752, "step": 27825 }, { "epoch": 0.8907880417386851, "grad_norm": 0.634976327419281, "learning_rate": 1.0578319462998445e-05, "loss": 0.37, "num_input_tokens_seen": 37707392, "step": 27830 }, { "epoch": 0.8909480827091736, "grad_norm": 1.7400603294372559, "learning_rate": 1.057030128755214e-05, "loss": 0.4696, "num_input_tokens_seen": 37714064, "step": 27835 }, { "epoch": 0.891108123679662, "grad_norm": 2.444538116455078, "learning_rate": 1.0562285337346703e-05, "loss": 0.6507, "num_input_tokens_seen": 37720624, "step": 27840 }, { "epoch": 0.8912681646501505, "grad_norm": 0.6649273037910461, "learning_rate": 1.0554271613618308e-05, "loss": 0.4478, "num_input_tokens_seen": 37727520, "step": 27845 }, { "epoch": 0.8914282056206388, "grad_norm": 0.9790852069854736, "learning_rate": 1.054626011760276e-05, "loss": 0.5446, "num_input_tokens_seen": 37734784, "step": 27850 }, { "epoch": 0.8915882465911273, "grad_norm": 0.6829090714454651, "learning_rate": 1.0538250850535549e-05, "loss": 0.5413, "num_input_tokens_seen": 37741056, "step": 27855 }, { "epoch": 0.8917482875616157, "grad_norm": 0.5898301601409912, "learning_rate": 1.0530243813651794e-05, "loss": 0.4368, "num_input_tokens_seen": 37747728, "step": 27860 }, { "epoch": 0.8919083285321042, "grad_norm": 1.7837529182434082, "learning_rate": 1.0522239008186271e-05, "loss": 0.4475, "num_input_tokens_seen": 37754496, "step": 27865 }, { "epoch": 0.8920683695025927, "grad_norm": 0.488763689994812, "learning_rate": 1.0514236435373434e-05, "loss": 0.404, "num_input_tokens_seen": 37760800, "step": 27870 }, { "epoch": 0.8922284104730811, "grad_norm": 0.8125307559967041, "learning_rate": 1.0506236096447386e-05, "loss": 0.3835, "num_input_tokens_seen": 37767376, "step": 27875 }, { "epoch": 0.8923884514435696, "grad_norm": 0.8249881267547607, "learning_rate": 1.049823799264186e-05, "loss": 0.3602, "num_input_tokens_seen": 37774240, "step": 27880 }, { "epoch": 0.892548492414058, "grad_norm": 0.6085098385810852, "learning_rate": 1.049024212519028e-05, "loss": 0.4641, "num_input_tokens_seen": 37780912, "step": 27885 }, { "epoch": 0.8927085333845465, "grad_norm": 1.5371012687683105, "learning_rate": 1.0482248495325713e-05, "loss": 0.5357, "num_input_tokens_seen": 37787792, "step": 27890 }, { "epoch": 0.8928685743550349, "grad_norm": 2.415452003479004, "learning_rate": 1.047425710428086e-05, "loss": 0.497, "num_input_tokens_seen": 37794432, "step": 27895 }, { "epoch": 0.8930286153255234, "grad_norm": 0.9846587181091309, "learning_rate": 1.0466267953288114e-05, "loss": 0.5266, "num_input_tokens_seen": 37801232, "step": 27900 }, { "epoch": 0.8931886562960117, "grad_norm": 0.7846252918243408, "learning_rate": 1.0458281043579482e-05, "loss": 0.4576, "num_input_tokens_seen": 37808144, "step": 27905 }, { "epoch": 0.8933486972665002, "grad_norm": 0.9878780841827393, "learning_rate": 1.0450296376386657e-05, "loss": 0.5762, "num_input_tokens_seen": 37814864, "step": 27910 }, { "epoch": 0.8935087382369886, "grad_norm": 0.6839149594306946, "learning_rate": 1.044231395294098e-05, "loss": 0.4746, "num_input_tokens_seen": 37821616, "step": 27915 }, { "epoch": 0.8936687792074771, "grad_norm": 1.4309921264648438, "learning_rate": 1.0434333774473435e-05, "loss": 0.4044, "num_input_tokens_seen": 37828752, "step": 27920 }, { "epoch": 0.8938288201779656, "grad_norm": 0.6506509184837341, "learning_rate": 1.0426355842214657e-05, "loss": 0.3989, "num_input_tokens_seen": 37835056, "step": 27925 }, { "epoch": 0.893988861148454, "grad_norm": 1.5658860206604004, "learning_rate": 1.0418380157394963e-05, "loss": 0.3968, "num_input_tokens_seen": 37842160, "step": 27930 }, { "epoch": 0.8941489021189425, "grad_norm": 0.8953266739845276, "learning_rate": 1.0410406721244281e-05, "loss": 0.506, "num_input_tokens_seen": 37848976, "step": 27935 }, { "epoch": 0.8943089430894309, "grad_norm": 1.210089087486267, "learning_rate": 1.0402435534992238e-05, "loss": 0.4254, "num_input_tokens_seen": 37855824, "step": 27940 }, { "epoch": 0.8944689840599194, "grad_norm": 1.0293519496917725, "learning_rate": 1.0394466599868071e-05, "loss": 0.5136, "num_input_tokens_seen": 37862624, "step": 27945 }, { "epoch": 0.8946290250304078, "grad_norm": 3.1042428016662598, "learning_rate": 1.0386499917100697e-05, "loss": 0.6174, "num_input_tokens_seen": 37869408, "step": 27950 }, { "epoch": 0.8947890660008962, "grad_norm": 1.3758273124694824, "learning_rate": 1.0378535487918692e-05, "loss": 0.4866, "num_input_tokens_seen": 37875952, "step": 27955 }, { "epoch": 0.8949491069713846, "grad_norm": 0.7571635842323303, "learning_rate": 1.037057331355025e-05, "loss": 0.3181, "num_input_tokens_seen": 37882704, "step": 27960 }, { "epoch": 0.8951091479418731, "grad_norm": 1.577998399734497, "learning_rate": 1.0362613395223247e-05, "loss": 0.6891, "num_input_tokens_seen": 37889360, "step": 27965 }, { "epoch": 0.8952691889123615, "grad_norm": 1.1432147026062012, "learning_rate": 1.0354655734165212e-05, "loss": 0.4684, "num_input_tokens_seen": 37896368, "step": 27970 }, { "epoch": 0.89542922988285, "grad_norm": 2.3780741691589355, "learning_rate": 1.03467003316033e-05, "loss": 0.4007, "num_input_tokens_seen": 37902576, "step": 27975 }, { "epoch": 0.8955892708533385, "grad_norm": 0.7362944483757019, "learning_rate": 1.033874718876435e-05, "loss": 0.5464, "num_input_tokens_seen": 37909216, "step": 27980 }, { "epoch": 0.8957493118238269, "grad_norm": 1.4075714349746704, "learning_rate": 1.0330796306874818e-05, "loss": 0.4291, "num_input_tokens_seen": 37915760, "step": 27985 }, { "epoch": 0.8959093527943154, "grad_norm": 1.3392956256866455, "learning_rate": 1.032284768716085e-05, "loss": 0.5428, "num_input_tokens_seen": 37922736, "step": 27990 }, { "epoch": 0.8960693937648038, "grad_norm": 1.2315138578414917, "learning_rate": 1.0314901330848206e-05, "loss": 0.2909, "num_input_tokens_seen": 37929312, "step": 27995 }, { "epoch": 0.8962294347352923, "grad_norm": 0.7239326238632202, "learning_rate": 1.030695723916233e-05, "loss": 0.4245, "num_input_tokens_seen": 37935760, "step": 28000 }, { "epoch": 0.8962294347352923, "eval_loss": 0.4473896026611328, "eval_runtime": 502.8151, "eval_samples_per_second": 27.617, "eval_steps_per_second": 13.808, "num_input_tokens_seen": 37935760, "step": 28000 }, { "epoch": 0.8963894757057806, "grad_norm": 0.962806224822998, "learning_rate": 1.0299015413328289e-05, "loss": 0.5645, "num_input_tokens_seen": 37942464, "step": 28005 }, { "epoch": 0.8965495166762691, "grad_norm": 0.7911249399185181, "learning_rate": 1.0291075854570809e-05, "loss": 0.5595, "num_input_tokens_seen": 37949264, "step": 28010 }, { "epoch": 0.8967095576467575, "grad_norm": 1.3906611204147339, "learning_rate": 1.0283138564114275e-05, "loss": 0.4318, "num_input_tokens_seen": 37955888, "step": 28015 }, { "epoch": 0.896869598617246, "grad_norm": 0.8552301526069641, "learning_rate": 1.027520354318273e-05, "loss": 0.4143, "num_input_tokens_seen": 37962688, "step": 28020 }, { "epoch": 0.8970296395877344, "grad_norm": 0.5581190586090088, "learning_rate": 1.0267270792999828e-05, "loss": 0.3489, "num_input_tokens_seen": 37969296, "step": 28025 }, { "epoch": 0.8971896805582229, "grad_norm": 0.704081118106842, "learning_rate": 1.0259340314788919e-05, "loss": 0.3538, "num_input_tokens_seen": 37975616, "step": 28030 }, { "epoch": 0.8973497215287114, "grad_norm": 1.0467689037322998, "learning_rate": 1.0251412109772979e-05, "loss": 0.4222, "num_input_tokens_seen": 37982432, "step": 28035 }, { "epoch": 0.8975097624991998, "grad_norm": 1.9408351182937622, "learning_rate": 1.0243486179174627e-05, "loss": 0.4921, "num_input_tokens_seen": 37989008, "step": 28040 }, { "epoch": 0.8976698034696883, "grad_norm": 1.2899929285049438, "learning_rate": 1.0235562524216158e-05, "loss": 0.5928, "num_input_tokens_seen": 37996336, "step": 28045 }, { "epoch": 0.8978298444401767, "grad_norm": 1.9680625200271606, "learning_rate": 1.022764114611948e-05, "loss": 0.7363, "num_input_tokens_seen": 38003104, "step": 28050 }, { "epoch": 0.8979898854106652, "grad_norm": 1.3342915773391724, "learning_rate": 1.0219722046106178e-05, "loss": 0.4736, "num_input_tokens_seen": 38009424, "step": 28055 }, { "epoch": 0.8981499263811535, "grad_norm": 0.6893227696418762, "learning_rate": 1.0211805225397486e-05, "loss": 0.324, "num_input_tokens_seen": 38016208, "step": 28060 }, { "epoch": 0.898309967351642, "grad_norm": 2.0163629055023193, "learning_rate": 1.020389068521426e-05, "loss": 0.6699, "num_input_tokens_seen": 38022544, "step": 28065 }, { "epoch": 0.8984700083221304, "grad_norm": 0.517765462398529, "learning_rate": 1.0195978426777039e-05, "loss": 0.4772, "num_input_tokens_seen": 38029888, "step": 28070 }, { "epoch": 0.8986300492926189, "grad_norm": 1.322505235671997, "learning_rate": 1.0188068451305982e-05, "loss": 0.4515, "num_input_tokens_seen": 38036608, "step": 28075 }, { "epoch": 0.8987900902631074, "grad_norm": 0.7886798977851868, "learning_rate": 1.0180160760020902e-05, "loss": 0.3855, "num_input_tokens_seen": 38043216, "step": 28080 }, { "epoch": 0.8989501312335958, "grad_norm": 1.3048319816589355, "learning_rate": 1.0172255354141278e-05, "loss": 0.5376, "num_input_tokens_seen": 38049840, "step": 28085 }, { "epoch": 0.8991101722040843, "grad_norm": 1.5632708072662354, "learning_rate": 1.0164352234886205e-05, "loss": 0.4001, "num_input_tokens_seen": 38056640, "step": 28090 }, { "epoch": 0.8992702131745727, "grad_norm": 0.7354735136032104, "learning_rate": 1.0156451403474454e-05, "loss": 0.341, "num_input_tokens_seen": 38063424, "step": 28095 }, { "epoch": 0.8994302541450612, "grad_norm": 1.3639843463897705, "learning_rate": 1.0148552861124443e-05, "loss": 0.4831, "num_input_tokens_seen": 38070608, "step": 28100 }, { "epoch": 0.8995902951155496, "grad_norm": 2.040687322616577, "learning_rate": 1.0140656609054205e-05, "loss": 0.3657, "num_input_tokens_seen": 38077616, "step": 28105 }, { "epoch": 0.899750336086038, "grad_norm": 0.8861901164054871, "learning_rate": 1.0132762648481455e-05, "loss": 0.487, "num_input_tokens_seen": 38084560, "step": 28110 }, { "epoch": 0.8999103770565264, "grad_norm": 1.7804811000823975, "learning_rate": 1.0124870980623543e-05, "loss": 0.5321, "num_input_tokens_seen": 38091408, "step": 28115 }, { "epoch": 0.9000704180270149, "grad_norm": 1.0581941604614258, "learning_rate": 1.0116981606697453e-05, "loss": 0.4999, "num_input_tokens_seen": 38097952, "step": 28120 }, { "epoch": 0.9002304589975033, "grad_norm": 2.202836036682129, "learning_rate": 1.0109094527919838e-05, "loss": 0.5598, "num_input_tokens_seen": 38104544, "step": 28125 }, { "epoch": 0.9003904999679918, "grad_norm": 1.0996066331863403, "learning_rate": 1.010120974550697e-05, "loss": 0.4053, "num_input_tokens_seen": 38111280, "step": 28130 }, { "epoch": 0.9005505409384803, "grad_norm": 1.119792103767395, "learning_rate": 1.0093327260674795e-05, "loss": 0.4541, "num_input_tokens_seen": 38118160, "step": 28135 }, { "epoch": 0.9007105819089687, "grad_norm": 1.1695590019226074, "learning_rate": 1.0085447074638878e-05, "loss": 0.3679, "num_input_tokens_seen": 38124640, "step": 28140 }, { "epoch": 0.9008706228794572, "grad_norm": 0.6080266833305359, "learning_rate": 1.0077569188614461e-05, "loss": 0.3076, "num_input_tokens_seen": 38131168, "step": 28145 }, { "epoch": 0.9010306638499456, "grad_norm": 2.519484043121338, "learning_rate": 1.0069693603816393e-05, "loss": 0.5523, "num_input_tokens_seen": 38137728, "step": 28150 }, { "epoch": 0.9011907048204341, "grad_norm": 0.5659096240997314, "learning_rate": 1.0061820321459204e-05, "loss": 0.2899, "num_input_tokens_seen": 38144464, "step": 28155 }, { "epoch": 0.9013507457909224, "grad_norm": 0.6995646953582764, "learning_rate": 1.0053949342757038e-05, "loss": 0.4218, "num_input_tokens_seen": 38150928, "step": 28160 }, { "epoch": 0.9015107867614109, "grad_norm": 1.9614917039871216, "learning_rate": 1.0046080668923717e-05, "loss": 0.6125, "num_input_tokens_seen": 38157872, "step": 28165 }, { "epoch": 0.9016708277318993, "grad_norm": 2.252455472946167, "learning_rate": 1.003821430117267e-05, "loss": 0.3651, "num_input_tokens_seen": 38164896, "step": 28170 }, { "epoch": 0.9018308687023878, "grad_norm": 1.1107290983200073, "learning_rate": 1.0030350240716999e-05, "loss": 0.3723, "num_input_tokens_seen": 38171520, "step": 28175 }, { "epoch": 0.9019909096728762, "grad_norm": 0.8814994096755981, "learning_rate": 1.0022488488769449e-05, "loss": 0.4752, "num_input_tokens_seen": 38178000, "step": 28180 }, { "epoch": 0.9021509506433647, "grad_norm": 0.6124024391174316, "learning_rate": 1.0014629046542387e-05, "loss": 0.4375, "num_input_tokens_seen": 38184912, "step": 28185 }, { "epoch": 0.9023109916138532, "grad_norm": 0.9732754230499268, "learning_rate": 1.0006771915247842e-05, "loss": 0.4724, "num_input_tokens_seen": 38191552, "step": 28190 }, { "epoch": 0.9024710325843416, "grad_norm": 1.7876957654953003, "learning_rate": 9.998917096097495e-06, "loss": 0.255, "num_input_tokens_seen": 38198224, "step": 28195 }, { "epoch": 0.9026310735548301, "grad_norm": 0.7991209626197815, "learning_rate": 9.991064590302638e-06, "loss": 0.4141, "num_input_tokens_seen": 38204832, "step": 28200 }, { "epoch": 0.9026310735548301, "eval_loss": 0.4473317265510559, "eval_runtime": 502.8457, "eval_samples_per_second": 27.615, "eval_steps_per_second": 13.807, "num_input_tokens_seen": 38204832, "step": 28200 }, { "epoch": 0.9027911145253185, "grad_norm": 0.4048801064491272, "learning_rate": 9.983214399074241e-06, "loss": 0.3668, "num_input_tokens_seen": 38211344, "step": 28205 }, { "epoch": 0.902951155495807, "grad_norm": 1.0169267654418945, "learning_rate": 9.975366523622893e-06, "loss": 0.4932, "num_input_tokens_seen": 38218096, "step": 28210 }, { "epoch": 0.9031111964662953, "grad_norm": 2.718863010406494, "learning_rate": 9.967520965158841e-06, "loss": 0.6828, "num_input_tokens_seen": 38225648, "step": 28215 }, { "epoch": 0.9032712374367838, "grad_norm": 1.394716501235962, "learning_rate": 9.95967772489197e-06, "loss": 0.4154, "num_input_tokens_seen": 38232512, "step": 28220 }, { "epoch": 0.9034312784072722, "grad_norm": 1.614880919456482, "learning_rate": 9.951836804031794e-06, "loss": 0.4075, "num_input_tokens_seen": 38238944, "step": 28225 }, { "epoch": 0.9035913193777607, "grad_norm": 0.5394642353057861, "learning_rate": 9.943998203787489e-06, "loss": 0.332, "num_input_tokens_seen": 38245712, "step": 28230 }, { "epoch": 0.9037513603482491, "grad_norm": 0.8718390464782715, "learning_rate": 9.936161925367874e-06, "loss": 0.4975, "num_input_tokens_seen": 38252272, "step": 28235 }, { "epoch": 0.9039114013187376, "grad_norm": 0.4716379642486572, "learning_rate": 9.928327969981386e-06, "loss": 0.3813, "num_input_tokens_seen": 38259040, "step": 28240 }, { "epoch": 0.9040714422892261, "grad_norm": 0.7430824637413025, "learning_rate": 9.920496338836135e-06, "loss": 0.2746, "num_input_tokens_seen": 38265664, "step": 28245 }, { "epoch": 0.9042314832597145, "grad_norm": 1.042675256729126, "learning_rate": 9.912667033139844e-06, "loss": 0.3923, "num_input_tokens_seen": 38272432, "step": 28250 }, { "epoch": 0.904391524230203, "grad_norm": 0.5394790768623352, "learning_rate": 9.904840054099893e-06, "loss": 0.4215, "num_input_tokens_seen": 38279984, "step": 28255 }, { "epoch": 0.9045515652006914, "grad_norm": 1.8303722143173218, "learning_rate": 9.897015402923312e-06, "loss": 0.3217, "num_input_tokens_seen": 38286432, "step": 28260 }, { "epoch": 0.9047116061711798, "grad_norm": 0.8991348743438721, "learning_rate": 9.889193080816744e-06, "loss": 0.5484, "num_input_tokens_seen": 38293728, "step": 28265 }, { "epoch": 0.9048716471416682, "grad_norm": 1.0607234239578247, "learning_rate": 9.881373088986498e-06, "loss": 0.3315, "num_input_tokens_seen": 38300736, "step": 28270 }, { "epoch": 0.9050316881121567, "grad_norm": 0.9181934595108032, "learning_rate": 9.873555428638523e-06, "loss": 0.457, "num_input_tokens_seen": 38307456, "step": 28275 }, { "epoch": 0.9051917290826451, "grad_norm": 1.6627824306488037, "learning_rate": 9.865740100978383e-06, "loss": 0.5078, "num_input_tokens_seen": 38314128, "step": 28280 }, { "epoch": 0.9053517700531336, "grad_norm": 1.1497366428375244, "learning_rate": 9.857927107211315e-06, "loss": 0.4522, "num_input_tokens_seen": 38320832, "step": 28285 }, { "epoch": 0.905511811023622, "grad_norm": 0.9037163257598877, "learning_rate": 9.850116448542177e-06, "loss": 0.5129, "num_input_tokens_seen": 38327296, "step": 28290 }, { "epoch": 0.9056718519941105, "grad_norm": 1.5598374605178833, "learning_rate": 9.842308126175457e-06, "loss": 0.57, "num_input_tokens_seen": 38334480, "step": 28295 }, { "epoch": 0.905831892964599, "grad_norm": 1.1839035749435425, "learning_rate": 9.834502141315315e-06, "loss": 0.3657, "num_input_tokens_seen": 38341504, "step": 28300 }, { "epoch": 0.9059919339350874, "grad_norm": 0.6624149680137634, "learning_rate": 9.82669849516552e-06, "loss": 0.3967, "num_input_tokens_seen": 38348112, "step": 28305 }, { "epoch": 0.9061519749055759, "grad_norm": 2.293757915496826, "learning_rate": 9.818897188929493e-06, "loss": 0.457, "num_input_tokens_seen": 38354848, "step": 28310 }, { "epoch": 0.9063120158760642, "grad_norm": 1.1332803964614868, "learning_rate": 9.811098223810309e-06, "loss": 0.4548, "num_input_tokens_seen": 38361872, "step": 28315 }, { "epoch": 0.9064720568465527, "grad_norm": 1.2649662494659424, "learning_rate": 9.803301601010641e-06, "loss": 0.5423, "num_input_tokens_seen": 38368432, "step": 28320 }, { "epoch": 0.9066320978170411, "grad_norm": 0.7153981924057007, "learning_rate": 9.795507321732853e-06, "loss": 0.3251, "num_input_tokens_seen": 38375024, "step": 28325 }, { "epoch": 0.9067921387875296, "grad_norm": 1.1392743587493896, "learning_rate": 9.787715387178898e-06, "loss": 0.511, "num_input_tokens_seen": 38381632, "step": 28330 }, { "epoch": 0.906952179758018, "grad_norm": 2.221482992172241, "learning_rate": 9.779925798550399e-06, "loss": 0.4922, "num_input_tokens_seen": 38388352, "step": 28335 }, { "epoch": 0.9071122207285065, "grad_norm": 1.1309136152267456, "learning_rate": 9.772138557048619e-06, "loss": 0.3603, "num_input_tokens_seen": 38395024, "step": 28340 }, { "epoch": 0.907272261698995, "grad_norm": 0.638218879699707, "learning_rate": 9.764353663874426e-06, "loss": 0.2374, "num_input_tokens_seen": 38401856, "step": 28345 }, { "epoch": 0.9074323026694834, "grad_norm": 1.4256640672683716, "learning_rate": 9.756571120228375e-06, "loss": 0.5168, "num_input_tokens_seen": 38408848, "step": 28350 }, { "epoch": 0.9075923436399719, "grad_norm": 1.4236916303634644, "learning_rate": 9.748790927310605e-06, "loss": 0.3629, "num_input_tokens_seen": 38415440, "step": 28355 }, { "epoch": 0.9077523846104603, "grad_norm": 2.112926721572876, "learning_rate": 9.741013086320946e-06, "loss": 0.4008, "num_input_tokens_seen": 38422256, "step": 28360 }, { "epoch": 0.9079124255809488, "grad_norm": 1.302713394165039, "learning_rate": 9.733237598458821e-06, "loss": 0.4085, "num_input_tokens_seen": 38429200, "step": 28365 }, { "epoch": 0.9080724665514371, "grad_norm": 1.5834625959396362, "learning_rate": 9.725464464923308e-06, "loss": 0.5002, "num_input_tokens_seen": 38435808, "step": 28370 }, { "epoch": 0.9082325075219256, "grad_norm": 0.694423496723175, "learning_rate": 9.717693686913123e-06, "loss": 0.4662, "num_input_tokens_seen": 38442416, "step": 28375 }, { "epoch": 0.908392548492414, "grad_norm": 1.2975995540618896, "learning_rate": 9.709925265626632e-06, "loss": 0.5092, "num_input_tokens_seen": 38449072, "step": 28380 }, { "epoch": 0.9085525894629025, "grad_norm": 0.4555673599243164, "learning_rate": 9.702159202261801e-06, "loss": 0.3649, "num_input_tokens_seen": 38455568, "step": 28385 }, { "epoch": 0.9087126304333909, "grad_norm": 0.9193395972251892, "learning_rate": 9.694395498016268e-06, "loss": 0.3059, "num_input_tokens_seen": 38462320, "step": 28390 }, { "epoch": 0.9088726714038794, "grad_norm": 1.0653893947601318, "learning_rate": 9.686634154087298e-06, "loss": 0.3711, "num_input_tokens_seen": 38468784, "step": 28395 }, { "epoch": 0.9090327123743679, "grad_norm": 1.4043737649917603, "learning_rate": 9.678875171671776e-06, "loss": 0.4092, "num_input_tokens_seen": 38475552, "step": 28400 }, { "epoch": 0.9090327123743679, "eval_loss": 0.44691312313079834, "eval_runtime": 502.1467, "eval_samples_per_second": 27.653, "eval_steps_per_second": 13.827, "num_input_tokens_seen": 38475552, "step": 28400 }, { "epoch": 0.9091927533448563, "grad_norm": 2.9908607006073, "learning_rate": 9.671118551966246e-06, "loss": 0.4548, "num_input_tokens_seen": 38482112, "step": 28405 }, { "epoch": 0.9093527943153448, "grad_norm": 0.8936663866043091, "learning_rate": 9.66336429616686e-06, "loss": 0.4389, "num_input_tokens_seen": 38488592, "step": 28410 }, { "epoch": 0.9095128352858332, "grad_norm": 1.3224040269851685, "learning_rate": 9.655612405469436e-06, "loss": 0.3608, "num_input_tokens_seen": 38495456, "step": 28415 }, { "epoch": 0.9096728762563216, "grad_norm": 1.0727665424346924, "learning_rate": 9.647862881069413e-06, "loss": 0.357, "num_input_tokens_seen": 38502352, "step": 28420 }, { "epoch": 0.90983291722681, "grad_norm": 1.4226837158203125, "learning_rate": 9.640115724161855e-06, "loss": 0.5128, "num_input_tokens_seen": 38509760, "step": 28425 }, { "epoch": 0.9099929581972985, "grad_norm": 0.9809828400611877, "learning_rate": 9.632370935941483e-06, "loss": 0.3724, "num_input_tokens_seen": 38516912, "step": 28430 }, { "epoch": 0.9101529991677869, "grad_norm": 1.5203149318695068, "learning_rate": 9.624628517602634e-06, "loss": 0.4364, "num_input_tokens_seen": 38523792, "step": 28435 }, { "epoch": 0.9103130401382754, "grad_norm": 1.527077317237854, "learning_rate": 9.61688847033928e-06, "loss": 0.4388, "num_input_tokens_seen": 38530432, "step": 28440 }, { "epoch": 0.9104730811087638, "grad_norm": 1.2591172456741333, "learning_rate": 9.609150795345051e-06, "loss": 0.4291, "num_input_tokens_seen": 38537280, "step": 28445 }, { "epoch": 0.9106331220792523, "grad_norm": 0.7165976166725159, "learning_rate": 9.601415493813171e-06, "loss": 0.3075, "num_input_tokens_seen": 38544032, "step": 28450 }, { "epoch": 0.9107931630497408, "grad_norm": 0.9677860736846924, "learning_rate": 9.593682566936533e-06, "loss": 0.3474, "num_input_tokens_seen": 38550768, "step": 28455 }, { "epoch": 0.9109532040202292, "grad_norm": 1.2147001028060913, "learning_rate": 9.58595201590766e-06, "loss": 0.3593, "num_input_tokens_seen": 38557904, "step": 28460 }, { "epoch": 0.9111132449907177, "grad_norm": 1.1886470317840576, "learning_rate": 9.578223841918681e-06, "loss": 0.5595, "num_input_tokens_seen": 38564192, "step": 28465 }, { "epoch": 0.911273285961206, "grad_norm": 0.9289337992668152, "learning_rate": 9.570498046161389e-06, "loss": 0.3442, "num_input_tokens_seen": 38570592, "step": 28470 }, { "epoch": 0.9114333269316945, "grad_norm": 1.4837377071380615, "learning_rate": 9.562774629827206e-06, "loss": 0.3848, "num_input_tokens_seen": 38577392, "step": 28475 }, { "epoch": 0.9115933679021829, "grad_norm": 0.8995751142501831, "learning_rate": 9.555053594107163e-06, "loss": 0.4026, "num_input_tokens_seen": 38584160, "step": 28480 }, { "epoch": 0.9117534088726714, "grad_norm": 2.1170456409454346, "learning_rate": 9.547334940191957e-06, "loss": 0.5442, "num_input_tokens_seen": 38590544, "step": 28485 }, { "epoch": 0.9119134498431598, "grad_norm": 0.531827449798584, "learning_rate": 9.539618669271886e-06, "loss": 0.4722, "num_input_tokens_seen": 38597584, "step": 28490 }, { "epoch": 0.9120734908136483, "grad_norm": 0.6818651556968689, "learning_rate": 9.531904782536904e-06, "loss": 0.5023, "num_input_tokens_seen": 38604176, "step": 28495 }, { "epoch": 0.9122335317841367, "grad_norm": 1.444675087928772, "learning_rate": 9.524193281176597e-06, "loss": 0.3883, "num_input_tokens_seen": 38611104, "step": 28500 }, { "epoch": 0.9123935727546252, "grad_norm": 0.47742292284965515, "learning_rate": 9.516484166380165e-06, "loss": 0.3283, "num_input_tokens_seen": 38617728, "step": 28505 }, { "epoch": 0.9125536137251137, "grad_norm": 1.9919450283050537, "learning_rate": 9.508777439336447e-06, "loss": 0.4862, "num_input_tokens_seen": 38624448, "step": 28510 }, { "epoch": 0.9127136546956021, "grad_norm": 0.30114802718162537, "learning_rate": 9.50107310123393e-06, "loss": 0.4274, "num_input_tokens_seen": 38631280, "step": 28515 }, { "epoch": 0.9128736956660906, "grad_norm": 1.513085961341858, "learning_rate": 9.493371153260702e-06, "loss": 0.4762, "num_input_tokens_seen": 38637696, "step": 28520 }, { "epoch": 0.9130337366365789, "grad_norm": 1.9561827182769775, "learning_rate": 9.485671596604523e-06, "loss": 0.2888, "num_input_tokens_seen": 38644896, "step": 28525 }, { "epoch": 0.9131937776070674, "grad_norm": 1.5585153102874756, "learning_rate": 9.477974432452738e-06, "loss": 0.4663, "num_input_tokens_seen": 38651584, "step": 28530 }, { "epoch": 0.9133538185775558, "grad_norm": 0.8265997171401978, "learning_rate": 9.470279661992356e-06, "loss": 0.4312, "num_input_tokens_seen": 38658592, "step": 28535 }, { "epoch": 0.9135138595480443, "grad_norm": 1.135668396949768, "learning_rate": 9.462587286410021e-06, "loss": 0.3932, "num_input_tokens_seen": 38665312, "step": 28540 }, { "epoch": 0.9136739005185327, "grad_norm": 1.2269251346588135, "learning_rate": 9.454897306891972e-06, "loss": 0.4446, "num_input_tokens_seen": 38672176, "step": 28545 }, { "epoch": 0.9138339414890212, "grad_norm": 1.348482608795166, "learning_rate": 9.44720972462411e-06, "loss": 0.3362, "num_input_tokens_seen": 38678624, "step": 28550 }, { "epoch": 0.9139939824595096, "grad_norm": 0.9339847564697266, "learning_rate": 9.439524540791964e-06, "loss": 0.437, "num_input_tokens_seen": 38685248, "step": 28555 }, { "epoch": 0.9141540234299981, "grad_norm": 1.4007827043533325, "learning_rate": 9.431841756580673e-06, "loss": 0.3291, "num_input_tokens_seen": 38692240, "step": 28560 }, { "epoch": 0.9143140644004866, "grad_norm": 1.285663366317749, "learning_rate": 9.42416137317503e-06, "loss": 0.4265, "num_input_tokens_seen": 38699664, "step": 28565 }, { "epoch": 0.914474105370975, "grad_norm": 1.5064975023269653, "learning_rate": 9.416483391759437e-06, "loss": 0.4933, "num_input_tokens_seen": 38706496, "step": 28570 }, { "epoch": 0.9146341463414634, "grad_norm": 1.6790990829467773, "learning_rate": 9.408807813517945e-06, "loss": 0.4755, "num_input_tokens_seen": 38713008, "step": 28575 }, { "epoch": 0.9147941873119518, "grad_norm": 1.211506962776184, "learning_rate": 9.401134639634221e-06, "loss": 0.5038, "num_input_tokens_seen": 38719648, "step": 28580 }, { "epoch": 0.9149542282824403, "grad_norm": 1.2137051820755005, "learning_rate": 9.393463871291555e-06, "loss": 0.384, "num_input_tokens_seen": 38726512, "step": 28585 }, { "epoch": 0.9151142692529287, "grad_norm": 1.3469098806381226, "learning_rate": 9.385795509672881e-06, "loss": 0.4065, "num_input_tokens_seen": 38733104, "step": 28590 }, { "epoch": 0.9152743102234172, "grad_norm": 2.10937237739563, "learning_rate": 9.378129555960771e-06, "loss": 0.6413, "num_input_tokens_seen": 38739600, "step": 28595 }, { "epoch": 0.9154343511939056, "grad_norm": 1.094238042831421, "learning_rate": 9.370466011337392e-06, "loss": 0.3323, "num_input_tokens_seen": 38746560, "step": 28600 }, { "epoch": 0.9154343511939056, "eval_loss": 0.44689232110977173, "eval_runtime": 501.9574, "eval_samples_per_second": 27.664, "eval_steps_per_second": 13.832, "num_input_tokens_seen": 38746560, "step": 28600 }, { "epoch": 0.9155943921643941, "grad_norm": 0.5626973509788513, "learning_rate": 9.362804876984573e-06, "loss": 0.3484, "num_input_tokens_seen": 38753344, "step": 28605 }, { "epoch": 0.9157544331348826, "grad_norm": 2.401491403579712, "learning_rate": 9.355146154083747e-06, "loss": 0.519, "num_input_tokens_seen": 38759728, "step": 28610 }, { "epoch": 0.915914474105371, "grad_norm": 0.575781524181366, "learning_rate": 9.347489843815987e-06, "loss": 0.3786, "num_input_tokens_seen": 38766800, "step": 28615 }, { "epoch": 0.9160745150758595, "grad_norm": 1.7480665445327759, "learning_rate": 9.339835947362002e-06, "loss": 0.4204, "num_input_tokens_seen": 38773632, "step": 28620 }, { "epoch": 0.9162345560463478, "grad_norm": 0.9935767650604248, "learning_rate": 9.332184465902105e-06, "loss": 0.3237, "num_input_tokens_seen": 38780192, "step": 28625 }, { "epoch": 0.9163945970168363, "grad_norm": 1.435620665550232, "learning_rate": 9.324535400616266e-06, "loss": 0.528, "num_input_tokens_seen": 38787200, "step": 28630 }, { "epoch": 0.9165546379873247, "grad_norm": 0.5525312423706055, "learning_rate": 9.31688875268405e-06, "loss": 0.5577, "num_input_tokens_seen": 38794176, "step": 28635 }, { "epoch": 0.9167146789578132, "grad_norm": 1.3160815238952637, "learning_rate": 9.309244523284674e-06, "loss": 0.405, "num_input_tokens_seen": 38801504, "step": 28640 }, { "epoch": 0.9168747199283016, "grad_norm": 1.3762916326522827, "learning_rate": 9.301602713596982e-06, "loss": 0.5405, "num_input_tokens_seen": 38808736, "step": 28645 }, { "epoch": 0.9170347608987901, "grad_norm": 2.027595281600952, "learning_rate": 9.293963324799432e-06, "loss": 0.4309, "num_input_tokens_seen": 38815376, "step": 28650 }, { "epoch": 0.9171948018692785, "grad_norm": 1.530226707458496, "learning_rate": 9.286326358070104e-06, "loss": 0.4687, "num_input_tokens_seen": 38821904, "step": 28655 }, { "epoch": 0.917354842839767, "grad_norm": 0.6656329035758972, "learning_rate": 9.278691814586729e-06, "loss": 0.4677, "num_input_tokens_seen": 38828672, "step": 28660 }, { "epoch": 0.9175148838102555, "grad_norm": 0.8194789886474609, "learning_rate": 9.271059695526635e-06, "loss": 0.2462, "num_input_tokens_seen": 38834944, "step": 28665 }, { "epoch": 0.9176749247807439, "grad_norm": 0.47458261251449585, "learning_rate": 9.263430002066805e-06, "loss": 0.3665, "num_input_tokens_seen": 38841440, "step": 28670 }, { "epoch": 0.9178349657512324, "grad_norm": 1.249041199684143, "learning_rate": 9.25580273538382e-06, "loss": 0.3474, "num_input_tokens_seen": 38848112, "step": 28675 }, { "epoch": 0.9179950067217207, "grad_norm": 2.862274646759033, "learning_rate": 9.248177896653907e-06, "loss": 0.5496, "num_input_tokens_seen": 38854704, "step": 28680 }, { "epoch": 0.9181550476922092, "grad_norm": 2.116982936859131, "learning_rate": 9.240555487052918e-06, "loss": 0.494, "num_input_tokens_seen": 38861856, "step": 28685 }, { "epoch": 0.9183150886626976, "grad_norm": 0.6126220226287842, "learning_rate": 9.232935507756313e-06, "loss": 0.3678, "num_input_tokens_seen": 38868176, "step": 28690 }, { "epoch": 0.9184751296331861, "grad_norm": 0.7455644607543945, "learning_rate": 9.225317959939193e-06, "loss": 0.4201, "num_input_tokens_seen": 38874928, "step": 28695 }, { "epoch": 0.9186351706036745, "grad_norm": 1.001046895980835, "learning_rate": 9.217702844776287e-06, "loss": 0.3302, "num_input_tokens_seen": 38881456, "step": 28700 }, { "epoch": 0.918795211574163, "grad_norm": 0.849030613899231, "learning_rate": 9.210090163441929e-06, "loss": 0.3958, "num_input_tokens_seen": 38888048, "step": 28705 }, { "epoch": 0.9189552525446514, "grad_norm": 1.6393040418624878, "learning_rate": 9.202479917110105e-06, "loss": 0.4566, "num_input_tokens_seen": 38894704, "step": 28710 }, { "epoch": 0.9191152935151399, "grad_norm": 1.9871826171875, "learning_rate": 9.194872106954392e-06, "loss": 0.6585, "num_input_tokens_seen": 38901728, "step": 28715 }, { "epoch": 0.9192753344856284, "grad_norm": 1.708677053451538, "learning_rate": 9.187266734148029e-06, "loss": 0.6035, "num_input_tokens_seen": 38908528, "step": 28720 }, { "epoch": 0.9194353754561168, "grad_norm": 1.7734719514846802, "learning_rate": 9.179663799863849e-06, "loss": 0.3465, "num_input_tokens_seen": 38915552, "step": 28725 }, { "epoch": 0.9195954164266052, "grad_norm": 0.6046491265296936, "learning_rate": 9.172063305274317e-06, "loss": 0.3499, "num_input_tokens_seen": 38922032, "step": 28730 }, { "epoch": 0.9197554573970936, "grad_norm": 0.9695170521736145, "learning_rate": 9.164465251551527e-06, "loss": 0.3175, "num_input_tokens_seen": 38928336, "step": 28735 }, { "epoch": 0.9199154983675821, "grad_norm": 2.200470209121704, "learning_rate": 9.156869639867205e-06, "loss": 0.5456, "num_input_tokens_seen": 38934720, "step": 28740 }, { "epoch": 0.9200755393380705, "grad_norm": 1.5349398851394653, "learning_rate": 9.149276471392677e-06, "loss": 0.4548, "num_input_tokens_seen": 38941312, "step": 28745 }, { "epoch": 0.920235580308559, "grad_norm": 1.7893506288528442, "learning_rate": 9.141685747298914e-06, "loss": 0.4693, "num_input_tokens_seen": 38948064, "step": 28750 }, { "epoch": 0.9203956212790474, "grad_norm": 0.9098327159881592, "learning_rate": 9.13409746875649e-06, "loss": 0.5827, "num_input_tokens_seen": 38954224, "step": 28755 }, { "epoch": 0.9205556622495359, "grad_norm": 1.3711967468261719, "learning_rate": 9.12651163693562e-06, "loss": 0.5419, "num_input_tokens_seen": 38961120, "step": 28760 }, { "epoch": 0.9207157032200243, "grad_norm": 1.4098546504974365, "learning_rate": 9.11892825300614e-06, "loss": 0.4346, "num_input_tokens_seen": 38967712, "step": 28765 }, { "epoch": 0.9208757441905128, "grad_norm": 0.4862283766269684, "learning_rate": 9.111347318137491e-06, "loss": 0.4384, "num_input_tokens_seen": 38974448, "step": 28770 }, { "epoch": 0.9210357851610013, "grad_norm": 1.2541356086730957, "learning_rate": 9.103768833498755e-06, "loss": 0.5356, "num_input_tokens_seen": 38981488, "step": 28775 }, { "epoch": 0.9211958261314896, "grad_norm": 1.828629732131958, "learning_rate": 9.096192800258639e-06, "loss": 0.5109, "num_input_tokens_seen": 38988752, "step": 28780 }, { "epoch": 0.9213558671019781, "grad_norm": 0.5587706565856934, "learning_rate": 9.088619219585443e-06, "loss": 0.4436, "num_input_tokens_seen": 38995472, "step": 28785 }, { "epoch": 0.9215159080724665, "grad_norm": 0.8300322890281677, "learning_rate": 9.081048092647127e-06, "loss": 0.4551, "num_input_tokens_seen": 39002528, "step": 28790 }, { "epoch": 0.921675949042955, "grad_norm": 0.9258273839950562, "learning_rate": 9.073479420611245e-06, "loss": 0.4487, "num_input_tokens_seen": 39009600, "step": 28795 }, { "epoch": 0.9218359900134434, "grad_norm": 1.3483490943908691, "learning_rate": 9.065913204644974e-06, "loss": 0.5035, "num_input_tokens_seen": 39016288, "step": 28800 }, { "epoch": 0.9218359900134434, "eval_loss": 0.4467929005622864, "eval_runtime": 502.9832, "eval_samples_per_second": 27.607, "eval_steps_per_second": 13.804, "num_input_tokens_seen": 39016288, "step": 28800 }, { "epoch": 0.9219960309839319, "grad_norm": 1.2587915658950806, "learning_rate": 9.058349445915135e-06, "loss": 0.4584, "num_input_tokens_seen": 39023328, "step": 28805 }, { "epoch": 0.9221560719544203, "grad_norm": 0.6415718197822571, "learning_rate": 9.050788145588138e-06, "loss": 0.3908, "num_input_tokens_seen": 39030160, "step": 28810 }, { "epoch": 0.9223161129249088, "grad_norm": 2.011758804321289, "learning_rate": 9.043229304830039e-06, "loss": 0.5028, "num_input_tokens_seen": 39036992, "step": 28815 }, { "epoch": 0.9224761538953972, "grad_norm": 1.0946365594863892, "learning_rate": 9.035672924806515e-06, "loss": 0.435, "num_input_tokens_seen": 39043616, "step": 28820 }, { "epoch": 0.9226361948658857, "grad_norm": 1.2542861700057983, "learning_rate": 9.028119006682839e-06, "loss": 0.6475, "num_input_tokens_seen": 39050656, "step": 28825 }, { "epoch": 0.9227962358363742, "grad_norm": 1.6148035526275635, "learning_rate": 9.020567551623935e-06, "loss": 0.3799, "num_input_tokens_seen": 39057456, "step": 28830 }, { "epoch": 0.9229562768068625, "grad_norm": 1.0831598043441772, "learning_rate": 9.013018560794318e-06, "loss": 0.3973, "num_input_tokens_seen": 39064288, "step": 28835 }, { "epoch": 0.923116317777351, "grad_norm": 1.568123698234558, "learning_rate": 9.005472035358139e-06, "loss": 0.3455, "num_input_tokens_seen": 39071328, "step": 28840 }, { "epoch": 0.9232763587478394, "grad_norm": 3.450308322906494, "learning_rate": 8.997927976479185e-06, "loss": 0.4151, "num_input_tokens_seen": 39077904, "step": 28845 }, { "epoch": 0.9234363997183279, "grad_norm": 1.4101487398147583, "learning_rate": 8.99038638532082e-06, "loss": 0.3461, "num_input_tokens_seen": 39084656, "step": 28850 }, { "epoch": 0.9235964406888163, "grad_norm": 1.986095905303955, "learning_rate": 8.982847263046065e-06, "loss": 0.449, "num_input_tokens_seen": 39091408, "step": 28855 }, { "epoch": 0.9237564816593048, "grad_norm": 0.5200111865997314, "learning_rate": 8.975310610817555e-06, "loss": 0.4489, "num_input_tokens_seen": 39097872, "step": 28860 }, { "epoch": 0.9239165226297932, "grad_norm": 1.6863083839416504, "learning_rate": 8.967776429797528e-06, "loss": 0.4629, "num_input_tokens_seen": 39105232, "step": 28865 }, { "epoch": 0.9240765636002817, "grad_norm": 0.7022715210914612, "learning_rate": 8.960244721147842e-06, "loss": 0.2379, "num_input_tokens_seen": 39111792, "step": 28870 }, { "epoch": 0.9242366045707702, "grad_norm": 0.9722920656204224, "learning_rate": 8.952715486029995e-06, "loss": 0.3567, "num_input_tokens_seen": 39118736, "step": 28875 }, { "epoch": 0.9243966455412586, "grad_norm": 0.7161692380905151, "learning_rate": 8.945188725605075e-06, "loss": 0.2541, "num_input_tokens_seen": 39125216, "step": 28880 }, { "epoch": 0.924556686511747, "grad_norm": 0.9279947876930237, "learning_rate": 8.937664441033817e-06, "loss": 0.4162, "num_input_tokens_seen": 39131952, "step": 28885 }, { "epoch": 0.9247167274822354, "grad_norm": 0.8987071514129639, "learning_rate": 8.930142633476549e-06, "loss": 0.4086, "num_input_tokens_seen": 39138832, "step": 28890 }, { "epoch": 0.9248767684527239, "grad_norm": 1.2405922412872314, "learning_rate": 8.92262330409323e-06, "loss": 0.509, "num_input_tokens_seen": 39145792, "step": 28895 }, { "epoch": 0.9250368094232123, "grad_norm": 2.0430877208709717, "learning_rate": 8.915106454043448e-06, "loss": 0.5038, "num_input_tokens_seen": 39152880, "step": 28900 }, { "epoch": 0.9251968503937008, "grad_norm": 0.9004833102226257, "learning_rate": 8.90759208448638e-06, "loss": 0.3726, "num_input_tokens_seen": 39159760, "step": 28905 }, { "epoch": 0.9253568913641892, "grad_norm": 1.027367115020752, "learning_rate": 8.900080196580848e-06, "loss": 0.5423, "num_input_tokens_seen": 39166752, "step": 28910 }, { "epoch": 0.9255169323346777, "grad_norm": 1.1744440793991089, "learning_rate": 8.892570791485267e-06, "loss": 0.4516, "num_input_tokens_seen": 39173760, "step": 28915 }, { "epoch": 0.9256769733051661, "grad_norm": 1.3665746450424194, "learning_rate": 8.885063870357688e-06, "loss": 0.5046, "num_input_tokens_seen": 39180448, "step": 28920 }, { "epoch": 0.9258370142756546, "grad_norm": 1.3737531900405884, "learning_rate": 8.87755943435578e-06, "loss": 0.4249, "num_input_tokens_seen": 39187392, "step": 28925 }, { "epoch": 0.9259970552461431, "grad_norm": 0.34172847867012024, "learning_rate": 8.87005748463681e-06, "loss": 0.4143, "num_input_tokens_seen": 39194208, "step": 28930 }, { "epoch": 0.9261570962166314, "grad_norm": 1.4998482465744019, "learning_rate": 8.862558022357681e-06, "loss": 0.3755, "num_input_tokens_seen": 39200720, "step": 28935 }, { "epoch": 0.9263171371871199, "grad_norm": 0.9449548125267029, "learning_rate": 8.855061048674903e-06, "loss": 0.4013, "num_input_tokens_seen": 39207552, "step": 28940 }, { "epoch": 0.9264771781576083, "grad_norm": 1.1614364385604858, "learning_rate": 8.847566564744595e-06, "loss": 0.4181, "num_input_tokens_seen": 39214272, "step": 28945 }, { "epoch": 0.9266372191280968, "grad_norm": 0.7235198616981506, "learning_rate": 8.840074571722512e-06, "loss": 0.4633, "num_input_tokens_seen": 39221056, "step": 28950 }, { "epoch": 0.9267972600985852, "grad_norm": 1.1377004384994507, "learning_rate": 8.832585070764002e-06, "loss": 0.4393, "num_input_tokens_seen": 39227568, "step": 28955 }, { "epoch": 0.9269573010690737, "grad_norm": 0.3911716043949127, "learning_rate": 8.825098063024045e-06, "loss": 0.2928, "num_input_tokens_seen": 39234256, "step": 28960 }, { "epoch": 0.9271173420395621, "grad_norm": 1.3337799310684204, "learning_rate": 8.817613549657244e-06, "loss": 0.3989, "num_input_tokens_seen": 39240864, "step": 28965 }, { "epoch": 0.9272773830100506, "grad_norm": 0.8897550702095032, "learning_rate": 8.810131531817783e-06, "loss": 0.5069, "num_input_tokens_seen": 39247696, "step": 28970 }, { "epoch": 0.927437423980539, "grad_norm": 1.4986854791641235, "learning_rate": 8.802652010659496e-06, "loss": 0.3331, "num_input_tokens_seen": 39254032, "step": 28975 }, { "epoch": 0.9275974649510275, "grad_norm": 0.914304792881012, "learning_rate": 8.795174987335827e-06, "loss": 0.3473, "num_input_tokens_seen": 39260656, "step": 28980 }, { "epoch": 0.927757505921516, "grad_norm": 0.5059024691581726, "learning_rate": 8.787700462999807e-06, "loss": 0.3639, "num_input_tokens_seen": 39267296, "step": 28985 }, { "epoch": 0.9279175468920043, "grad_norm": 2.0512619018554688, "learning_rate": 8.780228438804122e-06, "loss": 0.4436, "num_input_tokens_seen": 39273840, "step": 28990 }, { "epoch": 0.9280775878624928, "grad_norm": 1.2629001140594482, "learning_rate": 8.772758915901032e-06, "loss": 0.473, "num_input_tokens_seen": 39281136, "step": 28995 }, { "epoch": 0.9282376288329812, "grad_norm": 1.1432416439056396, "learning_rate": 8.765291895442443e-06, "loss": 0.3608, "num_input_tokens_seen": 39287360, "step": 29000 }, { "epoch": 0.9282376288329812, "eval_loss": 0.446988970041275, "eval_runtime": 501.8546, "eval_samples_per_second": 27.669, "eval_steps_per_second": 13.835, "num_input_tokens_seen": 39287360, "step": 29000 }, { "epoch": 0.9283976698034697, "grad_norm": 0.6134523153305054, "learning_rate": 8.75782737857987e-06, "loss": 0.4408, "num_input_tokens_seen": 39293776, "step": 29005 }, { "epoch": 0.9285577107739581, "grad_norm": 1.3912686109542847, "learning_rate": 8.750365366464425e-06, "loss": 0.3697, "num_input_tokens_seen": 39300688, "step": 29010 }, { "epoch": 0.9287177517444466, "grad_norm": 1.206839680671692, "learning_rate": 8.742905860246838e-06, "loss": 0.3214, "num_input_tokens_seen": 39307216, "step": 29015 }, { "epoch": 0.928877792714935, "grad_norm": 0.9101176857948303, "learning_rate": 8.735448861077478e-06, "loss": 0.4626, "num_input_tokens_seen": 39314064, "step": 29020 }, { "epoch": 0.9290378336854235, "grad_norm": 1.3230373859405518, "learning_rate": 8.727994370106288e-06, "loss": 0.4242, "num_input_tokens_seen": 39321072, "step": 29025 }, { "epoch": 0.9291978746559119, "grad_norm": 0.9018651843070984, "learning_rate": 8.720542388482861e-06, "loss": 0.4937, "num_input_tokens_seen": 39327696, "step": 29030 }, { "epoch": 0.9293579156264004, "grad_norm": 0.6811857223510742, "learning_rate": 8.71309291735637e-06, "loss": 0.396, "num_input_tokens_seen": 39335008, "step": 29035 }, { "epoch": 0.9295179565968889, "grad_norm": 0.7794992923736572, "learning_rate": 8.705645957875621e-06, "loss": 0.5036, "num_input_tokens_seen": 39341776, "step": 29040 }, { "epoch": 0.9296779975673772, "grad_norm": 1.6521902084350586, "learning_rate": 8.698201511189048e-06, "loss": 0.4921, "num_input_tokens_seen": 39348320, "step": 29045 }, { "epoch": 0.9298380385378657, "grad_norm": 1.5447715520858765, "learning_rate": 8.690759578444649e-06, "loss": 0.4605, "num_input_tokens_seen": 39355232, "step": 29050 }, { "epoch": 0.9299980795083541, "grad_norm": 1.864150881767273, "learning_rate": 8.68332016079008e-06, "loss": 0.4626, "num_input_tokens_seen": 39361680, "step": 29055 }, { "epoch": 0.9301581204788426, "grad_norm": 1.1327736377716064, "learning_rate": 8.6758832593726e-06, "loss": 0.6201, "num_input_tokens_seen": 39368048, "step": 29060 }, { "epoch": 0.930318161449331, "grad_norm": 3.2385659217834473, "learning_rate": 8.668448875339053e-06, "loss": 0.2667, "num_input_tokens_seen": 39374576, "step": 29065 }, { "epoch": 0.9304782024198195, "grad_norm": 0.5153788924217224, "learning_rate": 8.661017009835933e-06, "loss": 0.4206, "num_input_tokens_seen": 39381264, "step": 29070 }, { "epoch": 0.9306382433903079, "grad_norm": 0.952303946018219, "learning_rate": 8.653587664009311e-06, "loss": 0.3714, "num_input_tokens_seen": 39388368, "step": 29075 }, { "epoch": 0.9307982843607964, "grad_norm": 0.759958803653717, "learning_rate": 8.646160839004902e-06, "loss": 0.4262, "num_input_tokens_seen": 39395088, "step": 29080 }, { "epoch": 0.9309583253312849, "grad_norm": 0.9036935567855835, "learning_rate": 8.638736535967998e-06, "loss": 0.4014, "num_input_tokens_seen": 39401792, "step": 29085 }, { "epoch": 0.9311183663017732, "grad_norm": 0.9158445596694946, "learning_rate": 8.631314756043535e-06, "loss": 0.4578, "num_input_tokens_seen": 39408240, "step": 29090 }, { "epoch": 0.9312784072722617, "grad_norm": 1.1139249801635742, "learning_rate": 8.62389550037603e-06, "loss": 0.3006, "num_input_tokens_seen": 39415376, "step": 29095 }, { "epoch": 0.9314384482427501, "grad_norm": 1.4153125286102295, "learning_rate": 8.616478770109646e-06, "loss": 0.3827, "num_input_tokens_seen": 39421744, "step": 29100 }, { "epoch": 0.9315984892132386, "grad_norm": 1.32168447971344, "learning_rate": 8.609064566388111e-06, "loss": 0.4426, "num_input_tokens_seen": 39428304, "step": 29105 }, { "epoch": 0.931758530183727, "grad_norm": 0.8111050724983215, "learning_rate": 8.601652890354815e-06, "loss": 0.2657, "num_input_tokens_seen": 39434896, "step": 29110 }, { "epoch": 0.9319185711542155, "grad_norm": 1.190636157989502, "learning_rate": 8.594243743152705e-06, "loss": 0.3247, "num_input_tokens_seen": 39441552, "step": 29115 }, { "epoch": 0.9320786121247039, "grad_norm": 0.9310837388038635, "learning_rate": 8.58683712592438e-06, "loss": 0.5612, "num_input_tokens_seen": 39448496, "step": 29120 }, { "epoch": 0.9322386530951924, "grad_norm": 0.6398417353630066, "learning_rate": 8.579433039812037e-06, "loss": 0.4354, "num_input_tokens_seen": 39455472, "step": 29125 }, { "epoch": 0.9323986940656808, "grad_norm": 1.40254807472229, "learning_rate": 8.572031485957466e-06, "loss": 0.4635, "num_input_tokens_seen": 39461920, "step": 29130 }, { "epoch": 0.9325587350361693, "grad_norm": 1.3724803924560547, "learning_rate": 8.564632465502084e-06, "loss": 0.513, "num_input_tokens_seen": 39468656, "step": 29135 }, { "epoch": 0.9327187760066578, "grad_norm": 0.6110796332359314, "learning_rate": 8.557235979586928e-06, "loss": 0.4666, "num_input_tokens_seen": 39475456, "step": 29140 }, { "epoch": 0.9328788169771461, "grad_norm": 1.2413134574890137, "learning_rate": 8.549842029352606e-06, "loss": 0.5764, "num_input_tokens_seen": 39482160, "step": 29145 }, { "epoch": 0.9330388579476346, "grad_norm": 0.987693190574646, "learning_rate": 8.542450615939376e-06, "loss": 0.3398, "num_input_tokens_seen": 39489296, "step": 29150 }, { "epoch": 0.933198898918123, "grad_norm": 1.0791261196136475, "learning_rate": 8.535061740487082e-06, "loss": 0.436, "num_input_tokens_seen": 39495856, "step": 29155 }, { "epoch": 0.9333589398886115, "grad_norm": 1.0150728225708008, "learning_rate": 8.527675404135168e-06, "loss": 0.3964, "num_input_tokens_seen": 39502608, "step": 29160 }, { "epoch": 0.9335189808590999, "grad_norm": 0.6819791793823242, "learning_rate": 8.520291608022724e-06, "loss": 0.5532, "num_input_tokens_seen": 39509680, "step": 29165 }, { "epoch": 0.9336790218295884, "grad_norm": 1.2592693567276, "learning_rate": 8.512910353288398e-06, "loss": 0.5237, "num_input_tokens_seen": 39516624, "step": 29170 }, { "epoch": 0.9338390628000768, "grad_norm": 0.9921311736106873, "learning_rate": 8.505531641070486e-06, "loss": 0.3396, "num_input_tokens_seen": 39523472, "step": 29175 }, { "epoch": 0.9339991037705653, "grad_norm": 0.6248499751091003, "learning_rate": 8.498155472506885e-06, "loss": 0.2642, "num_input_tokens_seen": 39530832, "step": 29180 }, { "epoch": 0.9341591447410537, "grad_norm": 1.2825303077697754, "learning_rate": 8.49078184873508e-06, "loss": 0.501, "num_input_tokens_seen": 39537376, "step": 29185 }, { "epoch": 0.9343191857115422, "grad_norm": 1.403912901878357, "learning_rate": 8.483410770892188e-06, "loss": 0.3604, "num_input_tokens_seen": 39544176, "step": 29190 }, { "epoch": 0.9344792266820307, "grad_norm": 1.2790606021881104, "learning_rate": 8.476042240114909e-06, "loss": 0.6119, "num_input_tokens_seen": 39550960, "step": 29195 }, { "epoch": 0.934639267652519, "grad_norm": 0.8067705631256104, "learning_rate": 8.468676257539568e-06, "loss": 0.4684, "num_input_tokens_seen": 39557440, "step": 29200 }, { "epoch": 0.934639267652519, "eval_loss": 0.4466246962547302, "eval_runtime": 502.0716, "eval_samples_per_second": 27.657, "eval_steps_per_second": 13.829, "num_input_tokens_seen": 39557440, "step": 29200 }, { "epoch": 0.9347993086230075, "grad_norm": 1.8372219800949097, "learning_rate": 8.4613128243021e-06, "loss": 0.3034, "num_input_tokens_seen": 39564512, "step": 29205 }, { "epoch": 0.9349593495934959, "grad_norm": 0.5446352958679199, "learning_rate": 8.453951941538028e-06, "loss": 0.5045, "num_input_tokens_seen": 39571456, "step": 29210 }, { "epoch": 0.9351193905639844, "grad_norm": 0.8312909603118896, "learning_rate": 8.446593610382495e-06, "loss": 0.6862, "num_input_tokens_seen": 39578624, "step": 29215 }, { "epoch": 0.9352794315344728, "grad_norm": 3.0480711460113525, "learning_rate": 8.439237831970259e-06, "loss": 0.4825, "num_input_tokens_seen": 39585712, "step": 29220 }, { "epoch": 0.9354394725049613, "grad_norm": 1.0838195085525513, "learning_rate": 8.431884607435667e-06, "loss": 0.2966, "num_input_tokens_seen": 39591904, "step": 29225 }, { "epoch": 0.9355995134754497, "grad_norm": 0.9340204000473022, "learning_rate": 8.424533937912665e-06, "loss": 0.4521, "num_input_tokens_seen": 39598800, "step": 29230 }, { "epoch": 0.9357595544459382, "grad_norm": 1.0092605352401733, "learning_rate": 8.41718582453484e-06, "loss": 0.5158, "num_input_tokens_seen": 39605680, "step": 29235 }, { "epoch": 0.9359195954164266, "grad_norm": 2.206325054168701, "learning_rate": 8.409840268435346e-06, "loss": 0.4284, "num_input_tokens_seen": 39612528, "step": 29240 }, { "epoch": 0.936079636386915, "grad_norm": 1.4139004945755005, "learning_rate": 8.402497270746976e-06, "loss": 0.45, "num_input_tokens_seen": 39619680, "step": 29245 }, { "epoch": 0.9362396773574035, "grad_norm": 2.4772329330444336, "learning_rate": 8.395156832602095e-06, "loss": 0.5174, "num_input_tokens_seen": 39626096, "step": 29250 }, { "epoch": 0.9363997183278919, "grad_norm": 0.8652684092521667, "learning_rate": 8.387818955132707e-06, "loss": 0.3816, "num_input_tokens_seen": 39632656, "step": 29255 }, { "epoch": 0.9365597592983804, "grad_norm": 0.9761320352554321, "learning_rate": 8.38048363947039e-06, "loss": 0.5765, "num_input_tokens_seen": 39639600, "step": 29260 }, { "epoch": 0.9367198002688688, "grad_norm": 1.5928584337234497, "learning_rate": 8.373150886746351e-06, "loss": 0.4485, "num_input_tokens_seen": 39646320, "step": 29265 }, { "epoch": 0.9368798412393573, "grad_norm": 1.6521097421646118, "learning_rate": 8.365820698091397e-06, "loss": 0.4904, "num_input_tokens_seen": 39653504, "step": 29270 }, { "epoch": 0.9370398822098457, "grad_norm": 2.573869466781616, "learning_rate": 8.358493074635922e-06, "loss": 0.4742, "num_input_tokens_seen": 39660032, "step": 29275 }, { "epoch": 0.9371999231803342, "grad_norm": 1.3574162721633911, "learning_rate": 8.351168017509948e-06, "loss": 0.2929, "num_input_tokens_seen": 39666752, "step": 29280 }, { "epoch": 0.9373599641508226, "grad_norm": 0.9829820394515991, "learning_rate": 8.343845527843094e-06, "loss": 0.3395, "num_input_tokens_seen": 39673344, "step": 29285 }, { "epoch": 0.9375200051213111, "grad_norm": 1.1947203874588013, "learning_rate": 8.336525606764566e-06, "loss": 0.4463, "num_input_tokens_seen": 39679808, "step": 29290 }, { "epoch": 0.9376800460917994, "grad_norm": 0.6517603993415833, "learning_rate": 8.329208255403204e-06, "loss": 0.3368, "num_input_tokens_seen": 39686416, "step": 29295 }, { "epoch": 0.9378400870622879, "grad_norm": 1.6426318883895874, "learning_rate": 8.321893474887426e-06, "loss": 0.4802, "num_input_tokens_seen": 39693264, "step": 29300 }, { "epoch": 0.9380001280327764, "grad_norm": 0.7532843947410583, "learning_rate": 8.31458126634526e-06, "loss": 0.2546, "num_input_tokens_seen": 39700112, "step": 29305 }, { "epoch": 0.9381601690032648, "grad_norm": 1.4365205764770508, "learning_rate": 8.30727163090435e-06, "loss": 0.4622, "num_input_tokens_seen": 39706848, "step": 29310 }, { "epoch": 0.9383202099737533, "grad_norm": 1.600210189819336, "learning_rate": 8.29996456969192e-06, "loss": 0.5137, "num_input_tokens_seen": 39713472, "step": 29315 }, { "epoch": 0.9384802509442417, "grad_norm": 0.9494938254356384, "learning_rate": 8.292660083834818e-06, "loss": 0.3734, "num_input_tokens_seen": 39720208, "step": 29320 }, { "epoch": 0.9386402919147302, "grad_norm": 1.322889804840088, "learning_rate": 8.2853581744595e-06, "loss": 0.4775, "num_input_tokens_seen": 39727200, "step": 29325 }, { "epoch": 0.9388003328852186, "grad_norm": 0.67103111743927, "learning_rate": 8.278058842691991e-06, "loss": 0.3962, "num_input_tokens_seen": 39734112, "step": 29330 }, { "epoch": 0.9389603738557071, "grad_norm": 2.258406400680542, "learning_rate": 8.27076208965796e-06, "loss": 0.5514, "num_input_tokens_seen": 39741328, "step": 29335 }, { "epoch": 0.9391204148261955, "grad_norm": 1.4450788497924805, "learning_rate": 8.263467916482637e-06, "loss": 0.4115, "num_input_tokens_seen": 39748144, "step": 29340 }, { "epoch": 0.939280455796684, "grad_norm": 1.406304955482483, "learning_rate": 8.256176324290885e-06, "loss": 0.4742, "num_input_tokens_seen": 39754544, "step": 29345 }, { "epoch": 0.9394404967671725, "grad_norm": 1.4462047815322876, "learning_rate": 8.248887314207168e-06, "loss": 0.4047, "num_input_tokens_seen": 39762496, "step": 29350 }, { "epoch": 0.9396005377376608, "grad_norm": 0.8478049635887146, "learning_rate": 8.24160088735553e-06, "loss": 0.3477, "num_input_tokens_seen": 39769536, "step": 29355 }, { "epoch": 0.9397605787081493, "grad_norm": 0.9706061482429504, "learning_rate": 8.234317044859629e-06, "loss": 0.4006, "num_input_tokens_seen": 39776848, "step": 29360 }, { "epoch": 0.9399206196786377, "grad_norm": 1.9254733324050903, "learning_rate": 8.227035787842744e-06, "loss": 0.5045, "num_input_tokens_seen": 39783440, "step": 29365 }, { "epoch": 0.9400806606491262, "grad_norm": 1.5120891332626343, "learning_rate": 8.219757117427721e-06, "loss": 0.7038, "num_input_tokens_seen": 39790512, "step": 29370 }, { "epoch": 0.9402407016196146, "grad_norm": 0.35816293954849243, "learning_rate": 8.212481034737014e-06, "loss": 0.3168, "num_input_tokens_seen": 39797488, "step": 29375 }, { "epoch": 0.9404007425901031, "grad_norm": 2.053136110305786, "learning_rate": 8.205207540892707e-06, "loss": 0.4448, "num_input_tokens_seen": 39804064, "step": 29380 }, { "epoch": 0.9405607835605915, "grad_norm": 1.4115327596664429, "learning_rate": 8.197936637016442e-06, "loss": 0.5872, "num_input_tokens_seen": 39810576, "step": 29385 }, { "epoch": 0.94072082453108, "grad_norm": 0.9614371061325073, "learning_rate": 8.190668324229508e-06, "loss": 0.5211, "num_input_tokens_seen": 39817152, "step": 29390 }, { "epoch": 0.9408808655015684, "grad_norm": 0.5774345397949219, "learning_rate": 8.183402603652749e-06, "loss": 0.4335, "num_input_tokens_seen": 39823792, "step": 29395 }, { "epoch": 0.9410409064720568, "grad_norm": 1.1318367719650269, "learning_rate": 8.176139476406635e-06, "loss": 0.2902, "num_input_tokens_seen": 39830256, "step": 29400 }, { "epoch": 0.9410409064720568, "eval_loss": 0.4463297724723816, "eval_runtime": 501.5944, "eval_samples_per_second": 27.684, "eval_steps_per_second": 13.842, "num_input_tokens_seen": 39830256, "step": 29400 }, { "epoch": 0.9412009474425453, "grad_norm": 0.594544529914856, "learning_rate": 8.16887894361125e-06, "loss": 0.3126, "num_input_tokens_seen": 39837200, "step": 29405 }, { "epoch": 0.9413609884130337, "grad_norm": 0.8185716271400452, "learning_rate": 8.161621006386233e-06, "loss": 0.3577, "num_input_tokens_seen": 39844304, "step": 29410 }, { "epoch": 0.9415210293835222, "grad_norm": 2.002852439880371, "learning_rate": 8.154365665850869e-06, "loss": 0.5155, "num_input_tokens_seen": 39851024, "step": 29415 }, { "epoch": 0.9416810703540106, "grad_norm": 0.8106648325920105, "learning_rate": 8.147112923124005e-06, "loss": 0.4724, "num_input_tokens_seen": 39857648, "step": 29420 }, { "epoch": 0.9418411113244991, "grad_norm": 1.7479777336120605, "learning_rate": 8.13986277932412e-06, "loss": 0.4544, "num_input_tokens_seen": 39864352, "step": 29425 }, { "epoch": 0.9420011522949875, "grad_norm": 0.6317320466041565, "learning_rate": 8.132615235569277e-06, "loss": 0.3586, "num_input_tokens_seen": 39871072, "step": 29430 }, { "epoch": 0.942161193265476, "grad_norm": 0.8768548369407654, "learning_rate": 8.125370292977124e-06, "loss": 0.2883, "num_input_tokens_seen": 39877968, "step": 29435 }, { "epoch": 0.9423212342359644, "grad_norm": 0.5289057493209839, "learning_rate": 8.118127952664944e-06, "loss": 0.3204, "num_input_tokens_seen": 39884592, "step": 29440 }, { "epoch": 0.9424812752064529, "grad_norm": 1.0740056037902832, "learning_rate": 8.110888215749574e-06, "loss": 0.5523, "num_input_tokens_seen": 39891408, "step": 29445 }, { "epoch": 0.9426413161769412, "grad_norm": 0.8389222621917725, "learning_rate": 8.10365108334749e-06, "loss": 0.3921, "num_input_tokens_seen": 39898224, "step": 29450 }, { "epoch": 0.9428013571474297, "grad_norm": 2.0079894065856934, "learning_rate": 8.096416556574743e-06, "loss": 0.3738, "num_input_tokens_seen": 39905120, "step": 29455 }, { "epoch": 0.9429613981179182, "grad_norm": 0.9042963981628418, "learning_rate": 8.08918463654698e-06, "loss": 0.2828, "num_input_tokens_seen": 39911712, "step": 29460 }, { "epoch": 0.9431214390884066, "grad_norm": 0.898002028465271, "learning_rate": 8.081955324379458e-06, "loss": 0.4716, "num_input_tokens_seen": 39918544, "step": 29465 }, { "epoch": 0.9432814800588951, "grad_norm": 0.5860347151756287, "learning_rate": 8.074728621187039e-06, "loss": 0.3079, "num_input_tokens_seen": 39925312, "step": 29470 }, { "epoch": 0.9434415210293835, "grad_norm": 1.4151105880737305, "learning_rate": 8.067504528084158e-06, "loss": 0.5424, "num_input_tokens_seen": 39932256, "step": 29475 }, { "epoch": 0.943601561999872, "grad_norm": 1.104325532913208, "learning_rate": 8.060283046184861e-06, "loss": 0.5113, "num_input_tokens_seen": 39939552, "step": 29480 }, { "epoch": 0.9437616029703604, "grad_norm": 1.0580766201019287, "learning_rate": 8.053064176602806e-06, "loss": 0.2541, "num_input_tokens_seen": 39946352, "step": 29485 }, { "epoch": 0.9439216439408489, "grad_norm": 0.6876772046089172, "learning_rate": 8.045847920451216e-06, "loss": 0.4489, "num_input_tokens_seen": 39953344, "step": 29490 }, { "epoch": 0.9440816849113373, "grad_norm": 1.4056748151779175, "learning_rate": 8.038634278842944e-06, "loss": 0.617, "num_input_tokens_seen": 39960384, "step": 29495 }, { "epoch": 0.9442417258818258, "grad_norm": 0.9337983131408691, "learning_rate": 8.031423252890408e-06, "loss": 0.4878, "num_input_tokens_seen": 39967632, "step": 29500 }, { "epoch": 0.9444017668523141, "grad_norm": 1.7216322422027588, "learning_rate": 8.024214843705646e-06, "loss": 0.4572, "num_input_tokens_seen": 39974064, "step": 29505 }, { "epoch": 0.9445618078228026, "grad_norm": 1.7219592332839966, "learning_rate": 8.017009052400295e-06, "loss": 0.4161, "num_input_tokens_seen": 39981104, "step": 29510 }, { "epoch": 0.9447218487932911, "grad_norm": 0.9445180296897888, "learning_rate": 8.00980588008557e-06, "loss": 0.3832, "num_input_tokens_seen": 39987616, "step": 29515 }, { "epoch": 0.9448818897637795, "grad_norm": 1.1591988801956177, "learning_rate": 8.002605327872282e-06, "loss": 0.3839, "num_input_tokens_seen": 39994416, "step": 29520 }, { "epoch": 0.945041930734268, "grad_norm": 0.8680149912834167, "learning_rate": 7.995407396870862e-06, "loss": 0.4844, "num_input_tokens_seen": 40001136, "step": 29525 }, { "epoch": 0.9452019717047564, "grad_norm": 0.6328607201576233, "learning_rate": 7.988212088191307e-06, "loss": 0.3848, "num_input_tokens_seen": 40008496, "step": 29530 }, { "epoch": 0.9453620126752449, "grad_norm": 1.4071489572525024, "learning_rate": 7.98101940294324e-06, "loss": 0.4315, "num_input_tokens_seen": 40015136, "step": 29535 }, { "epoch": 0.9455220536457333, "grad_norm": 1.3675003051757812, "learning_rate": 7.973829342235847e-06, "loss": 0.3798, "num_input_tokens_seen": 40021760, "step": 29540 }, { "epoch": 0.9456820946162218, "grad_norm": 1.6150625944137573, "learning_rate": 7.966641907177936e-06, "loss": 0.3836, "num_input_tokens_seen": 40028272, "step": 29545 }, { "epoch": 0.9458421355867102, "grad_norm": 3.1703732013702393, "learning_rate": 7.959457098877901e-06, "loss": 0.6326, "num_input_tokens_seen": 40035152, "step": 29550 }, { "epoch": 0.9460021765571986, "grad_norm": 1.24990713596344, "learning_rate": 7.952274918443719e-06, "loss": 0.4808, "num_input_tokens_seen": 40041936, "step": 29555 }, { "epoch": 0.946162217527687, "grad_norm": 1.0010720491409302, "learning_rate": 7.945095366982983e-06, "loss": 0.4971, "num_input_tokens_seen": 40049328, "step": 29560 }, { "epoch": 0.9463222584981755, "grad_norm": 0.9347095489501953, "learning_rate": 7.937918445602871e-06, "loss": 0.421, "num_input_tokens_seen": 40056096, "step": 29565 }, { "epoch": 0.946482299468664, "grad_norm": 0.7142376899719238, "learning_rate": 7.930744155410145e-06, "loss": 0.3421, "num_input_tokens_seen": 40062880, "step": 29570 }, { "epoch": 0.9466423404391524, "grad_norm": 1.3780429363250732, "learning_rate": 7.923572497511181e-06, "loss": 0.4426, "num_input_tokens_seen": 40069696, "step": 29575 }, { "epoch": 0.9468023814096409, "grad_norm": 1.3671735525131226, "learning_rate": 7.916403473011927e-06, "loss": 0.41, "num_input_tokens_seen": 40076112, "step": 29580 }, { "epoch": 0.9469624223801293, "grad_norm": 0.56328946352005, "learning_rate": 7.909237083017953e-06, "loss": 0.3389, "num_input_tokens_seen": 40082544, "step": 29585 }, { "epoch": 0.9471224633506178, "grad_norm": 0.8534747958183289, "learning_rate": 7.902073328634389e-06, "loss": 0.3684, "num_input_tokens_seen": 40089328, "step": 29590 }, { "epoch": 0.9472825043211062, "grad_norm": 2.11899471282959, "learning_rate": 7.894912210965987e-06, "loss": 0.3343, "num_input_tokens_seen": 40095936, "step": 29595 }, { "epoch": 0.9474425452915947, "grad_norm": 1.764528751373291, "learning_rate": 7.887753731117075e-06, "loss": 0.4933, "num_input_tokens_seen": 40102464, "step": 29600 }, { "epoch": 0.9474425452915947, "eval_loss": 0.4463340640068054, "eval_runtime": 502.0938, "eval_samples_per_second": 27.656, "eval_steps_per_second": 13.828, "num_input_tokens_seen": 40102464, "step": 29600 }, { "epoch": 0.947602586262083, "grad_norm": 1.176598072052002, "learning_rate": 7.880597890191587e-06, "loss": 0.4211, "num_input_tokens_seen": 40108848, "step": 29605 }, { "epoch": 0.9477626272325715, "grad_norm": 0.6428348422050476, "learning_rate": 7.873444689293036e-06, "loss": 0.5552, "num_input_tokens_seen": 40115600, "step": 29610 }, { "epoch": 0.94792266820306, "grad_norm": 0.8888952732086182, "learning_rate": 7.866294129524548e-06, "loss": 0.4723, "num_input_tokens_seen": 40122784, "step": 29615 }, { "epoch": 0.9480827091735484, "grad_norm": 0.9078834056854248, "learning_rate": 7.859146211988811e-06, "loss": 0.5216, "num_input_tokens_seen": 40129376, "step": 29620 }, { "epoch": 0.9482427501440369, "grad_norm": 0.6113062500953674, "learning_rate": 7.852000937788134e-06, "loss": 0.3395, "num_input_tokens_seen": 40135984, "step": 29625 }, { "epoch": 0.9484027911145253, "grad_norm": 1.1219913959503174, "learning_rate": 7.844858308024416e-06, "loss": 0.6493, "num_input_tokens_seen": 40142880, "step": 29630 }, { "epoch": 0.9485628320850138, "grad_norm": 1.2802636623382568, "learning_rate": 7.837718323799122e-06, "loss": 0.5264, "num_input_tokens_seen": 40149424, "step": 29635 }, { "epoch": 0.9487228730555022, "grad_norm": 1.1616028547286987, "learning_rate": 7.83058098621334e-06, "loss": 0.5136, "num_input_tokens_seen": 40155888, "step": 29640 }, { "epoch": 0.9488829140259907, "grad_norm": 0.7440544962882996, "learning_rate": 7.823446296367739e-06, "loss": 0.4999, "num_input_tokens_seen": 40162544, "step": 29645 }, { "epoch": 0.9490429549964791, "grad_norm": 0.6742283701896667, "learning_rate": 7.81631425536257e-06, "loss": 0.4618, "num_input_tokens_seen": 40169328, "step": 29650 }, { "epoch": 0.9492029959669676, "grad_norm": 1.9738585948944092, "learning_rate": 7.809184864297689e-06, "loss": 0.368, "num_input_tokens_seen": 40175808, "step": 29655 }, { "epoch": 0.9493630369374559, "grad_norm": 1.416695237159729, "learning_rate": 7.802058124272532e-06, "loss": 0.4389, "num_input_tokens_seen": 40182640, "step": 29660 }, { "epoch": 0.9495230779079444, "grad_norm": 1.512817621231079, "learning_rate": 7.79493403638614e-06, "loss": 0.4941, "num_input_tokens_seen": 40189264, "step": 29665 }, { "epoch": 0.9496831188784329, "grad_norm": 1.060327410697937, "learning_rate": 7.787812601737132e-06, "loss": 0.4912, "num_input_tokens_seen": 40196224, "step": 29670 }, { "epoch": 0.9498431598489213, "grad_norm": 0.8688853979110718, "learning_rate": 7.780693821423715e-06, "loss": 0.4351, "num_input_tokens_seen": 40203088, "step": 29675 }, { "epoch": 0.9500032008194098, "grad_norm": 0.5394354462623596, "learning_rate": 7.773577696543705e-06, "loss": 0.272, "num_input_tokens_seen": 40210096, "step": 29680 }, { "epoch": 0.9501632417898982, "grad_norm": 1.657495379447937, "learning_rate": 7.7664642281945e-06, "loss": 0.5109, "num_input_tokens_seen": 40217072, "step": 29685 }, { "epoch": 0.9503232827603867, "grad_norm": 0.7388549447059631, "learning_rate": 7.759353417473072e-06, "loss": 0.3737, "num_input_tokens_seen": 40223760, "step": 29690 }, { "epoch": 0.9504833237308751, "grad_norm": 1.16424560546875, "learning_rate": 7.752245265476016e-06, "loss": 0.414, "num_input_tokens_seen": 40230288, "step": 29695 }, { "epoch": 0.9506433647013636, "grad_norm": 1.0180622339248657, "learning_rate": 7.745139773299481e-06, "loss": 0.5074, "num_input_tokens_seen": 40237152, "step": 29700 }, { "epoch": 0.950803405671852, "grad_norm": 2.1823770999908447, "learning_rate": 7.738036942039232e-06, "loss": 0.5906, "num_input_tokens_seen": 40243712, "step": 29705 }, { "epoch": 0.9509634466423404, "grad_norm": 1.0746866464614868, "learning_rate": 7.73093677279062e-06, "loss": 0.5863, "num_input_tokens_seen": 40250528, "step": 29710 }, { "epoch": 0.9511234876128288, "grad_norm": 1.2242847681045532, "learning_rate": 7.72383926664857e-06, "loss": 0.4928, "num_input_tokens_seen": 40257168, "step": 29715 }, { "epoch": 0.9512835285833173, "grad_norm": 0.8382877707481384, "learning_rate": 7.716744424707606e-06, "loss": 0.5106, "num_input_tokens_seen": 40264064, "step": 29720 }, { "epoch": 0.9514435695538058, "grad_norm": 1.1583232879638672, "learning_rate": 7.709652248061858e-06, "loss": 0.5831, "num_input_tokens_seen": 40270720, "step": 29725 }, { "epoch": 0.9516036105242942, "grad_norm": 1.2832956314086914, "learning_rate": 7.702562737805017e-06, "loss": 0.4386, "num_input_tokens_seen": 40277312, "step": 29730 }, { "epoch": 0.9517636514947827, "grad_norm": 1.1347779035568237, "learning_rate": 7.695475895030365e-06, "loss": 0.5705, "num_input_tokens_seen": 40283904, "step": 29735 }, { "epoch": 0.9519236924652711, "grad_norm": 1.2380725145339966, "learning_rate": 7.6883917208308e-06, "loss": 0.4975, "num_input_tokens_seen": 40290656, "step": 29740 }, { "epoch": 0.9520837334357596, "grad_norm": 0.6505553126335144, "learning_rate": 7.681310216298778e-06, "loss": 0.3218, "num_input_tokens_seen": 40297216, "step": 29745 }, { "epoch": 0.952243774406248, "grad_norm": 1.1128994226455688, "learning_rate": 7.674231382526367e-06, "loss": 0.3721, "num_input_tokens_seen": 40303632, "step": 29750 }, { "epoch": 0.9524038153767365, "grad_norm": 3.0256621837615967, "learning_rate": 7.667155220605198e-06, "loss": 0.4346, "num_input_tokens_seen": 40310416, "step": 29755 }, { "epoch": 0.9525638563472248, "grad_norm": 0.8765048384666443, "learning_rate": 7.660081731626515e-06, "loss": 0.4758, "num_input_tokens_seen": 40316896, "step": 29760 }, { "epoch": 0.9527238973177133, "grad_norm": 1.3428226709365845, "learning_rate": 7.653010916681141e-06, "loss": 0.3291, "num_input_tokens_seen": 40323344, "step": 29765 }, { "epoch": 0.9528839382882017, "grad_norm": 1.1866421699523926, "learning_rate": 7.645942776859472e-06, "loss": 0.4779, "num_input_tokens_seen": 40330176, "step": 29770 }, { "epoch": 0.9530439792586902, "grad_norm": 0.7131648063659668, "learning_rate": 7.63887731325152e-06, "loss": 0.4037, "num_input_tokens_seen": 40337344, "step": 29775 }, { "epoch": 0.9532040202291787, "grad_norm": 1.0078396797180176, "learning_rate": 7.63181452694685e-06, "loss": 0.3387, "num_input_tokens_seen": 40344064, "step": 29780 }, { "epoch": 0.9533640611996671, "grad_norm": 2.1597092151641846, "learning_rate": 7.624754419034644e-06, "loss": 0.482, "num_input_tokens_seen": 40351136, "step": 29785 }, { "epoch": 0.9535241021701556, "grad_norm": 1.0582417249679565, "learning_rate": 7.6176969906036645e-06, "loss": 0.4475, "num_input_tokens_seen": 40358272, "step": 29790 }, { "epoch": 0.953684143140644, "grad_norm": 1.0018419027328491, "learning_rate": 7.610642242742242e-06, "loss": 0.3822, "num_input_tokens_seen": 40365216, "step": 29795 }, { "epoch": 0.9538441841111325, "grad_norm": 0.7854601740837097, "learning_rate": 7.603590176538322e-06, "loss": 0.2688, "num_input_tokens_seen": 40371968, "step": 29800 }, { "epoch": 0.9538441841111325, "eval_loss": 0.44614383578300476, "eval_runtime": 501.8545, "eval_samples_per_second": 27.669, "eval_steps_per_second": 13.835, "num_input_tokens_seen": 40371968, "step": 29800 }, { "epoch": 0.9540042250816209, "grad_norm": 1.0010384321212769, "learning_rate": 7.596540793079404e-06, "loss": 0.4504, "num_input_tokens_seen": 40379104, "step": 29805 }, { "epoch": 0.9541642660521094, "grad_norm": 0.9398353695869446, "learning_rate": 7.5894940934526125e-06, "loss": 0.4042, "num_input_tokens_seen": 40385904, "step": 29810 }, { "epoch": 0.9543243070225977, "grad_norm": 1.5266778469085693, "learning_rate": 7.582450078744621e-06, "loss": 0.3982, "num_input_tokens_seen": 40392480, "step": 29815 }, { "epoch": 0.9544843479930862, "grad_norm": 2.7431395053863525, "learning_rate": 7.575408750041707e-06, "loss": 0.6872, "num_input_tokens_seen": 40398976, "step": 29820 }, { "epoch": 0.9546443889635747, "grad_norm": 1.2148798704147339, "learning_rate": 7.568370108429732e-06, "loss": 0.4655, "num_input_tokens_seen": 40405504, "step": 29825 }, { "epoch": 0.9548044299340631, "grad_norm": 1.2787766456604004, "learning_rate": 7.561334154994154e-06, "loss": 0.4026, "num_input_tokens_seen": 40412192, "step": 29830 }, { "epoch": 0.9549644709045516, "grad_norm": 0.3727843761444092, "learning_rate": 7.55430089081999e-06, "loss": 0.4079, "num_input_tokens_seen": 40419040, "step": 29835 }, { "epoch": 0.95512451187504, "grad_norm": 1.056492567062378, "learning_rate": 7.547270316991864e-06, "loss": 0.5088, "num_input_tokens_seen": 40425968, "step": 29840 }, { "epoch": 0.9552845528455285, "grad_norm": 1.6068685054779053, "learning_rate": 7.5402424345939884e-06, "loss": 0.4619, "num_input_tokens_seen": 40432688, "step": 29845 }, { "epoch": 0.9554445938160169, "grad_norm": 0.9926839470863342, "learning_rate": 7.533217244710133e-06, "loss": 0.4035, "num_input_tokens_seen": 40439152, "step": 29850 }, { "epoch": 0.9556046347865054, "grad_norm": 0.9034361839294434, "learning_rate": 7.52619474842369e-06, "loss": 0.3393, "num_input_tokens_seen": 40445776, "step": 29855 }, { "epoch": 0.9557646757569938, "grad_norm": 0.7694793343544006, "learning_rate": 7.519174946817597e-06, "loss": 0.382, "num_input_tokens_seen": 40452624, "step": 29860 }, { "epoch": 0.9559247167274822, "grad_norm": 1.7837719917297363, "learning_rate": 7.512157840974407e-06, "loss": 0.5946, "num_input_tokens_seen": 40459184, "step": 29865 }, { "epoch": 0.9560847576979706, "grad_norm": 1.055553674697876, "learning_rate": 7.5051434319762496e-06, "loss": 0.3815, "num_input_tokens_seen": 40465904, "step": 29870 }, { "epoch": 0.9562447986684591, "grad_norm": 1.3711150884628296, "learning_rate": 7.498131720904822e-06, "loss": 0.4469, "num_input_tokens_seen": 40472432, "step": 29875 }, { "epoch": 0.9564048396389476, "grad_norm": 0.41669389605522156, "learning_rate": 7.491122708841433e-06, "loss": 0.4138, "num_input_tokens_seen": 40479824, "step": 29880 }, { "epoch": 0.956564880609436, "grad_norm": 0.9328253269195557, "learning_rate": 7.4841163968669524e-06, "loss": 0.3113, "num_input_tokens_seen": 40486448, "step": 29885 }, { "epoch": 0.9567249215799245, "grad_norm": 0.7677809596061707, "learning_rate": 7.4771127860618355e-06, "loss": 0.4169, "num_input_tokens_seen": 40493088, "step": 29890 }, { "epoch": 0.9568849625504129, "grad_norm": 0.6468613147735596, "learning_rate": 7.470111877506139e-06, "loss": 0.3247, "num_input_tokens_seen": 40500176, "step": 29895 }, { "epoch": 0.9570450035209014, "grad_norm": 1.1668704748153687, "learning_rate": 7.463113672279479e-06, "loss": 0.4272, "num_input_tokens_seen": 40507104, "step": 29900 }, { "epoch": 0.9572050444913898, "grad_norm": 2.5045435428619385, "learning_rate": 7.456118171461071e-06, "loss": 0.4058, "num_input_tokens_seen": 40514112, "step": 29905 }, { "epoch": 0.9573650854618783, "grad_norm": 0.9293186068534851, "learning_rate": 7.449125376129721e-06, "loss": 0.4796, "num_input_tokens_seen": 40521088, "step": 29910 }, { "epoch": 0.9575251264323666, "grad_norm": 1.645723581314087, "learning_rate": 7.442135287363788e-06, "loss": 0.5559, "num_input_tokens_seen": 40528048, "step": 29915 }, { "epoch": 0.9576851674028551, "grad_norm": 0.6115571856498718, "learning_rate": 7.435147906241247e-06, "loss": 0.3006, "num_input_tokens_seen": 40534736, "step": 29920 }, { "epoch": 0.9578452083733435, "grad_norm": 1.063376545906067, "learning_rate": 7.428163233839624e-06, "loss": 0.4931, "num_input_tokens_seen": 40541232, "step": 29925 }, { "epoch": 0.958005249343832, "grad_norm": 0.9095811247825623, "learning_rate": 7.4211812712360525e-06, "loss": 0.3227, "num_input_tokens_seen": 40548480, "step": 29930 }, { "epoch": 0.9581652903143205, "grad_norm": 0.7489847540855408, "learning_rate": 7.4142020195072464e-06, "loss": 0.4385, "num_input_tokens_seen": 40555200, "step": 29935 }, { "epoch": 0.9583253312848089, "grad_norm": 2.0515849590301514, "learning_rate": 7.407225479729479e-06, "loss": 0.3518, "num_input_tokens_seen": 40562192, "step": 29940 }, { "epoch": 0.9584853722552974, "grad_norm": 1.1447302103042603, "learning_rate": 7.400251652978632e-06, "loss": 0.3792, "num_input_tokens_seen": 40569408, "step": 29945 }, { "epoch": 0.9586454132257858, "grad_norm": 1.0466312170028687, "learning_rate": 7.393280540330147e-06, "loss": 0.504, "num_input_tokens_seen": 40576272, "step": 29950 }, { "epoch": 0.9588054541962743, "grad_norm": 1.1922645568847656, "learning_rate": 7.386312142859069e-06, "loss": 0.4737, "num_input_tokens_seen": 40582976, "step": 29955 }, { "epoch": 0.9589654951667627, "grad_norm": 1.5757163763046265, "learning_rate": 7.379346461640008e-06, "loss": 0.5668, "num_input_tokens_seen": 40589696, "step": 29960 }, { "epoch": 0.9591255361372512, "grad_norm": 1.9215655326843262, "learning_rate": 7.372383497747149e-06, "loss": 0.5664, "num_input_tokens_seen": 40596560, "step": 29965 }, { "epoch": 0.9592855771077395, "grad_norm": 1.274596095085144, "learning_rate": 7.3654232522542775e-06, "loss": 0.5512, "num_input_tokens_seen": 40603504, "step": 29970 }, { "epoch": 0.959445618078228, "grad_norm": 1.9080455303192139, "learning_rate": 7.358465726234756e-06, "loss": 0.4412, "num_input_tokens_seen": 40610176, "step": 29975 }, { "epoch": 0.9596056590487164, "grad_norm": 1.5043662786483765, "learning_rate": 7.351510920761512e-06, "loss": 0.4043, "num_input_tokens_seen": 40617040, "step": 29980 }, { "epoch": 0.9597657000192049, "grad_norm": 0.5037497282028198, "learning_rate": 7.344558836907067e-06, "loss": 0.3062, "num_input_tokens_seen": 40623888, "step": 29985 }, { "epoch": 0.9599257409896934, "grad_norm": 0.778590202331543, "learning_rate": 7.3376094757435285e-06, "loss": 0.323, "num_input_tokens_seen": 40630352, "step": 29990 }, { "epoch": 0.9600857819601818, "grad_norm": 1.6041080951690674, "learning_rate": 7.330662838342561e-06, "loss": 0.4343, "num_input_tokens_seen": 40637120, "step": 29995 }, { "epoch": 0.9602458229306703, "grad_norm": 1.2248390913009644, "learning_rate": 7.323718925775438e-06, "loss": 0.3888, "num_input_tokens_seen": 40643632, "step": 30000 }, { "epoch": 0.9602458229306703, "eval_loss": 0.446245938539505, "eval_runtime": 501.5243, "eval_samples_per_second": 27.688, "eval_steps_per_second": 13.844, "num_input_tokens_seen": 40643632, "step": 30000 }, { "epoch": 0.9604058639011587, "grad_norm": 1.2774256467819214, "learning_rate": 7.316777739112985e-06, "loss": 0.4311, "num_input_tokens_seen": 40650960, "step": 30005 }, { "epoch": 0.9605659048716472, "grad_norm": 0.8181086182594299, "learning_rate": 7.309839279425626e-06, "loss": 0.4508, "num_input_tokens_seen": 40657856, "step": 30010 }, { "epoch": 0.9607259458421356, "grad_norm": 0.7233179211616516, "learning_rate": 7.302903547783366e-06, "loss": 0.3872, "num_input_tokens_seen": 40664800, "step": 30015 }, { "epoch": 0.960885986812624, "grad_norm": 1.0370615720748901, "learning_rate": 7.2959705452557644e-06, "loss": 0.3799, "num_input_tokens_seen": 40671408, "step": 30020 }, { "epoch": 0.9610460277831124, "grad_norm": 0.8653162717819214, "learning_rate": 7.289040272911996e-06, "loss": 0.4457, "num_input_tokens_seen": 40678048, "step": 30025 }, { "epoch": 0.9612060687536009, "grad_norm": 1.4548200368881226, "learning_rate": 7.282112731820789e-06, "loss": 0.5028, "num_input_tokens_seen": 40684672, "step": 30030 }, { "epoch": 0.9613661097240893, "grad_norm": 1.3645825386047363, "learning_rate": 7.275187923050447e-06, "loss": 0.4124, "num_input_tokens_seen": 40691264, "step": 30035 }, { "epoch": 0.9615261506945778, "grad_norm": 0.9164348840713501, "learning_rate": 7.268265847668879e-06, "loss": 0.3804, "num_input_tokens_seen": 40698160, "step": 30040 }, { "epoch": 0.9616861916650663, "grad_norm": 1.4317917823791504, "learning_rate": 7.261346506743538e-06, "loss": 0.4036, "num_input_tokens_seen": 40704816, "step": 30045 }, { "epoch": 0.9618462326355547, "grad_norm": 1.0706809759140015, "learning_rate": 7.254429901341486e-06, "loss": 0.5289, "num_input_tokens_seen": 40711552, "step": 30050 }, { "epoch": 0.9620062736060432, "grad_norm": 1.021320104598999, "learning_rate": 7.247516032529356e-06, "loss": 0.427, "num_input_tokens_seen": 40717952, "step": 30055 }, { "epoch": 0.9621663145765316, "grad_norm": 0.9345084428787231, "learning_rate": 7.240604901373338e-06, "loss": 0.29, "num_input_tokens_seen": 40725024, "step": 30060 }, { "epoch": 0.9623263555470201, "grad_norm": 0.4407006800174713, "learning_rate": 7.233696508939223e-06, "loss": 0.3873, "num_input_tokens_seen": 40731744, "step": 30065 }, { "epoch": 0.9624863965175084, "grad_norm": 1.2444400787353516, "learning_rate": 7.226790856292376e-06, "loss": 0.4671, "num_input_tokens_seen": 40738208, "step": 30070 }, { "epoch": 0.9626464374879969, "grad_norm": 0.6253748536109924, "learning_rate": 7.219887944497727e-06, "loss": 0.3062, "num_input_tokens_seen": 40745168, "step": 30075 }, { "epoch": 0.9628064784584853, "grad_norm": 1.2710375785827637, "learning_rate": 7.2129877746198e-06, "loss": 0.3954, "num_input_tokens_seen": 40752064, "step": 30080 }, { "epoch": 0.9629665194289738, "grad_norm": 1.3633201122283936, "learning_rate": 7.20609034772268e-06, "loss": 0.4918, "num_input_tokens_seen": 40758736, "step": 30085 }, { "epoch": 0.9631265603994623, "grad_norm": 1.0774425268173218, "learning_rate": 7.19919566487004e-06, "loss": 0.4288, "num_input_tokens_seen": 40765392, "step": 30090 }, { "epoch": 0.9632866013699507, "grad_norm": 0.9419002532958984, "learning_rate": 7.192303727125132e-06, "loss": 0.4973, "num_input_tokens_seen": 40772688, "step": 30095 }, { "epoch": 0.9634466423404392, "grad_norm": 1.3350034952163696, "learning_rate": 7.185414535550777e-06, "loss": 0.43, "num_input_tokens_seen": 40779568, "step": 30100 }, { "epoch": 0.9636066833109276, "grad_norm": 5.463657855987549, "learning_rate": 7.178528091209363e-06, "loss": 0.4269, "num_input_tokens_seen": 40786240, "step": 30105 }, { "epoch": 0.9637667242814161, "grad_norm": 0.9549700617790222, "learning_rate": 7.171644395162888e-06, "loss": 0.4018, "num_input_tokens_seen": 40793648, "step": 30110 }, { "epoch": 0.9639267652519045, "grad_norm": 0.8380830883979797, "learning_rate": 7.164763448472881e-06, "loss": 0.3671, "num_input_tokens_seen": 40800304, "step": 30115 }, { "epoch": 0.964086806222393, "grad_norm": 0.7509099841117859, "learning_rate": 7.157885252200491e-06, "loss": 0.3095, "num_input_tokens_seen": 40806544, "step": 30120 }, { "epoch": 0.9642468471928813, "grad_norm": 1.0742989778518677, "learning_rate": 7.151009807406403e-06, "loss": 0.4341, "num_input_tokens_seen": 40813152, "step": 30125 }, { "epoch": 0.9644068881633698, "grad_norm": 1.4010231494903564, "learning_rate": 7.144137115150909e-06, "loss": 0.3842, "num_input_tokens_seen": 40820032, "step": 30130 }, { "epoch": 0.9645669291338582, "grad_norm": 2.1622979640960693, "learning_rate": 7.1372671764938725e-06, "loss": 0.5491, "num_input_tokens_seen": 40827168, "step": 30135 }, { "epoch": 0.9647269701043467, "grad_norm": 2.4561188220977783, "learning_rate": 7.130399992494705e-06, "loss": 0.4085, "num_input_tokens_seen": 40833360, "step": 30140 }, { "epoch": 0.9648870110748352, "grad_norm": 0.6576330661773682, "learning_rate": 7.123535564212419e-06, "loss": 0.3403, "num_input_tokens_seen": 40840032, "step": 30145 }, { "epoch": 0.9650470520453236, "grad_norm": 1.1841516494750977, "learning_rate": 7.116673892705611e-06, "loss": 0.4412, "num_input_tokens_seen": 40846368, "step": 30150 }, { "epoch": 0.9652070930158121, "grad_norm": 1.1548019647598267, "learning_rate": 7.109814979032415e-06, "loss": 0.6378, "num_input_tokens_seen": 40853296, "step": 30155 }, { "epoch": 0.9653671339863005, "grad_norm": 0.3775942325592041, "learning_rate": 7.102958824250577e-06, "loss": 0.3348, "num_input_tokens_seen": 40859936, "step": 30160 }, { "epoch": 0.965527174956789, "grad_norm": 1.0163546800613403, "learning_rate": 7.096105429417393e-06, "loss": 0.419, "num_input_tokens_seen": 40866672, "step": 30165 }, { "epoch": 0.9656872159272774, "grad_norm": 0.8947678208351135, "learning_rate": 7.0892547955897506e-06, "loss": 0.5218, "num_input_tokens_seen": 40873584, "step": 30170 }, { "epoch": 0.9658472568977658, "grad_norm": 1.5928438901901245, "learning_rate": 7.0824069238241e-06, "loss": 0.3516, "num_input_tokens_seen": 40879808, "step": 30175 }, { "epoch": 0.9660072978682542, "grad_norm": 1.499416470527649, "learning_rate": 7.075561815176462e-06, "loss": 0.5253, "num_input_tokens_seen": 40886656, "step": 30180 }, { "epoch": 0.9661673388387427, "grad_norm": 0.6686384081840515, "learning_rate": 7.068719470702445e-06, "loss": 0.3891, "num_input_tokens_seen": 40893648, "step": 30185 }, { "epoch": 0.9663273798092311, "grad_norm": 0.7004643678665161, "learning_rate": 7.061879891457229e-06, "loss": 0.3183, "num_input_tokens_seen": 40900400, "step": 30190 }, { "epoch": 0.9664874207797196, "grad_norm": 2.608196496963501, "learning_rate": 7.0550430784955515e-06, "loss": 0.615, "num_input_tokens_seen": 40907328, "step": 30195 }, { "epoch": 0.9666474617502081, "grad_norm": 1.2183170318603516, "learning_rate": 7.048209032871752e-06, "loss": 0.4356, "num_input_tokens_seen": 40914064, "step": 30200 }, { "epoch": 0.9666474617502081, "eval_loss": 0.4461461007595062, "eval_runtime": 500.9459, "eval_samples_per_second": 27.72, "eval_steps_per_second": 13.86, "num_input_tokens_seen": 40914064, "step": 30200 }, { "epoch": 0.9668075027206965, "grad_norm": 0.9397175312042236, "learning_rate": 7.0413777556397055e-06, "loss": 0.4047, "num_input_tokens_seen": 40920576, "step": 30205 }, { "epoch": 0.966967543691185, "grad_norm": 0.5613161325454712, "learning_rate": 7.0345492478528925e-06, "loss": 0.4712, "num_input_tokens_seen": 40927200, "step": 30210 }, { "epoch": 0.9671275846616734, "grad_norm": 0.6913599371910095, "learning_rate": 7.02772351056436e-06, "loss": 0.477, "num_input_tokens_seen": 40933456, "step": 30215 }, { "epoch": 0.9672876256321619, "grad_norm": 1.2289369106292725, "learning_rate": 7.020900544826709e-06, "loss": 0.5839, "num_input_tokens_seen": 40940096, "step": 30220 }, { "epoch": 0.9674476666026502, "grad_norm": 0.643688440322876, "learning_rate": 7.014080351692134e-06, "loss": 0.3154, "num_input_tokens_seen": 40946720, "step": 30225 }, { "epoch": 0.9676077075731387, "grad_norm": 1.8247597217559814, "learning_rate": 7.0072629322124024e-06, "loss": 0.3928, "num_input_tokens_seen": 40953600, "step": 30230 }, { "epoch": 0.9677677485436271, "grad_norm": 1.930310606956482, "learning_rate": 7.000448287438827e-06, "loss": 0.5323, "num_input_tokens_seen": 40960208, "step": 30235 }, { "epoch": 0.9679277895141156, "grad_norm": 0.8993620276451111, "learning_rate": 6.993636418422331e-06, "loss": 0.5037, "num_input_tokens_seen": 40966848, "step": 30240 }, { "epoch": 0.968087830484604, "grad_norm": 0.8171431422233582, "learning_rate": 6.986827326213383e-06, "loss": 0.3747, "num_input_tokens_seen": 40973600, "step": 30245 }, { "epoch": 0.9682478714550925, "grad_norm": 1.1214388608932495, "learning_rate": 6.9800210118620205e-06, "loss": 0.3328, "num_input_tokens_seen": 40981120, "step": 30250 }, { "epoch": 0.968407912425581, "grad_norm": 1.1642892360687256, "learning_rate": 6.973217476417876e-06, "loss": 0.6028, "num_input_tokens_seen": 40988096, "step": 30255 }, { "epoch": 0.9685679533960694, "grad_norm": 2.084151268005371, "learning_rate": 6.96641672093013e-06, "loss": 0.4868, "num_input_tokens_seen": 40994704, "step": 30260 }, { "epoch": 0.9687279943665579, "grad_norm": 1.0834803581237793, "learning_rate": 6.95961874644755e-06, "loss": 0.4804, "num_input_tokens_seen": 41001024, "step": 30265 }, { "epoch": 0.9688880353370463, "grad_norm": 0.948494553565979, "learning_rate": 6.952823554018476e-06, "loss": 0.3969, "num_input_tokens_seen": 41007712, "step": 30270 }, { "epoch": 0.9690480763075348, "grad_norm": 1.2346675395965576, "learning_rate": 6.946031144690798e-06, "loss": 0.4072, "num_input_tokens_seen": 41014496, "step": 30275 }, { "epoch": 0.9692081172780231, "grad_norm": 1.4045625925064087, "learning_rate": 6.939241519512005e-06, "loss": 0.4434, "num_input_tokens_seen": 41021184, "step": 30280 }, { "epoch": 0.9693681582485116, "grad_norm": 1.4893313646316528, "learning_rate": 6.932454679529129e-06, "loss": 0.4144, "num_input_tokens_seen": 41027648, "step": 30285 }, { "epoch": 0.969528199219, "grad_norm": 2.399909496307373, "learning_rate": 6.925670625788791e-06, "loss": 0.4077, "num_input_tokens_seen": 41034512, "step": 30290 }, { "epoch": 0.9696882401894885, "grad_norm": 0.30964288115501404, "learning_rate": 6.918889359337186e-06, "loss": 0.3518, "num_input_tokens_seen": 41041440, "step": 30295 }, { "epoch": 0.9698482811599769, "grad_norm": 1.3744370937347412, "learning_rate": 6.912110881220058e-06, "loss": 0.4372, "num_input_tokens_seen": 41048016, "step": 30300 }, { "epoch": 0.9700083221304654, "grad_norm": 0.820833683013916, "learning_rate": 6.905335192482735e-06, "loss": 0.2899, "num_input_tokens_seen": 41054912, "step": 30305 }, { "epoch": 0.9701683631009539, "grad_norm": 0.931501030921936, "learning_rate": 6.8985622941701275e-06, "loss": 0.4329, "num_input_tokens_seen": 41061792, "step": 30310 }, { "epoch": 0.9703284040714423, "grad_norm": 1.0155941247940063, "learning_rate": 6.89179218732669e-06, "loss": 0.423, "num_input_tokens_seen": 41068144, "step": 30315 }, { "epoch": 0.9704884450419308, "grad_norm": 0.4437260925769806, "learning_rate": 6.8850248729964595e-06, "loss": 0.3753, "num_input_tokens_seen": 41074784, "step": 30320 }, { "epoch": 0.9706484860124192, "grad_norm": 1.0803383588790894, "learning_rate": 6.8782603522230314e-06, "loss": 0.3965, "num_input_tokens_seen": 41081280, "step": 30325 }, { "epoch": 0.9708085269829076, "grad_norm": 1.3110935688018799, "learning_rate": 6.871498626049591e-06, "loss": 0.2581, "num_input_tokens_seen": 41088000, "step": 30330 }, { "epoch": 0.970968567953396, "grad_norm": 1.087380051612854, "learning_rate": 6.8647396955188875e-06, "loss": 0.5718, "num_input_tokens_seen": 41094256, "step": 30335 }, { "epoch": 0.9711286089238845, "grad_norm": 1.0877859592437744, "learning_rate": 6.857983561673218e-06, "loss": 0.3535, "num_input_tokens_seen": 41100848, "step": 30340 }, { "epoch": 0.9712886498943729, "grad_norm": 0.8706737756729126, "learning_rate": 6.851230225554467e-06, "loss": 0.4763, "num_input_tokens_seen": 41107568, "step": 30345 }, { "epoch": 0.9714486908648614, "grad_norm": 0.8548260927200317, "learning_rate": 6.8444796882040946e-06, "loss": 0.6091, "num_input_tokens_seen": 41114096, "step": 30350 }, { "epoch": 0.9716087318353499, "grad_norm": 0.5772595405578613, "learning_rate": 6.837731950663106e-06, "loss": 0.238, "num_input_tokens_seen": 41120736, "step": 30355 }, { "epoch": 0.9717687728058383, "grad_norm": 0.9033230543136597, "learning_rate": 6.830987013972098e-06, "loss": 0.3334, "num_input_tokens_seen": 41127488, "step": 30360 }, { "epoch": 0.9719288137763268, "grad_norm": 2.0272414684295654, "learning_rate": 6.82424487917121e-06, "loss": 0.6341, "num_input_tokens_seen": 41133696, "step": 30365 }, { "epoch": 0.9720888547468152, "grad_norm": 1.043502926826477, "learning_rate": 6.8175055473001735e-06, "loss": 0.4276, "num_input_tokens_seen": 41140832, "step": 30370 }, { "epoch": 0.9722488957173037, "grad_norm": 0.9907441735267639, "learning_rate": 6.8107690193982855e-06, "loss": 0.3666, "num_input_tokens_seen": 41147664, "step": 30375 }, { "epoch": 0.972408936687792, "grad_norm": 1.3301494121551514, "learning_rate": 6.804035296504385e-06, "loss": 0.3015, "num_input_tokens_seen": 41154464, "step": 30380 }, { "epoch": 0.9725689776582805, "grad_norm": 0.9800562858581543, "learning_rate": 6.797304379656916e-06, "loss": 0.3868, "num_input_tokens_seen": 41161024, "step": 30385 }, { "epoch": 0.9727290186287689, "grad_norm": 0.6560123562812805, "learning_rate": 6.790576269893861e-06, "loss": 0.4862, "num_input_tokens_seen": 41168000, "step": 30390 }, { "epoch": 0.9728890595992574, "grad_norm": 0.8217708468437195, "learning_rate": 6.783850968252772e-06, "loss": 0.2903, "num_input_tokens_seen": 41174976, "step": 30395 }, { "epoch": 0.9730491005697458, "grad_norm": 1.0641233921051025, "learning_rate": 6.777128475770789e-06, "loss": 0.419, "num_input_tokens_seen": 41182128, "step": 30400 }, { "epoch": 0.9730491005697458, "eval_loss": 0.4460110664367676, "eval_runtime": 501.9106, "eval_samples_per_second": 27.666, "eval_steps_per_second": 13.833, "num_input_tokens_seen": 41182128, "step": 30400 }, { "epoch": 0.9732091415402343, "grad_norm": 1.885340690612793, "learning_rate": 6.77040879348459e-06, "loss": 0.5261, "num_input_tokens_seen": 41188800, "step": 30405 }, { "epoch": 0.9733691825107228, "grad_norm": 1.1981825828552246, "learning_rate": 6.763691922430443e-06, "loss": 0.4552, "num_input_tokens_seen": 41195904, "step": 30410 }, { "epoch": 0.9735292234812112, "grad_norm": 0.8644784092903137, "learning_rate": 6.756977863644178e-06, "loss": 0.4955, "num_input_tokens_seen": 41202912, "step": 30415 }, { "epoch": 0.9736892644516997, "grad_norm": 1.3282780647277832, "learning_rate": 6.7502666181611804e-06, "loss": 0.3177, "num_input_tokens_seen": 41209424, "step": 30420 }, { "epoch": 0.9738493054221881, "grad_norm": 1.315554141998291, "learning_rate": 6.743558187016405e-06, "loss": 0.5786, "num_input_tokens_seen": 41216720, "step": 30425 }, { "epoch": 0.9740093463926766, "grad_norm": 0.9152923822402954, "learning_rate": 6.7368525712443925e-06, "loss": 0.4483, "num_input_tokens_seen": 41223280, "step": 30430 }, { "epoch": 0.9741693873631649, "grad_norm": 1.23293137550354, "learning_rate": 6.7301497718792155e-06, "loss": 0.4788, "num_input_tokens_seen": 41230016, "step": 30435 }, { "epoch": 0.9743294283336534, "grad_norm": 2.222691774368286, "learning_rate": 6.723449789954544e-06, "loss": 0.4367, "num_input_tokens_seen": 41237120, "step": 30440 }, { "epoch": 0.9744894693041418, "grad_norm": 1.4505932331085205, "learning_rate": 6.716752626503586e-06, "loss": 0.5213, "num_input_tokens_seen": 41243776, "step": 30445 }, { "epoch": 0.9746495102746303, "grad_norm": 0.8001918792724609, "learning_rate": 6.710058282559131e-06, "loss": 0.398, "num_input_tokens_seen": 41250592, "step": 30450 }, { "epoch": 0.9748095512451187, "grad_norm": 1.4430410861968994, "learning_rate": 6.703366759153545e-06, "loss": 0.4404, "num_input_tokens_seen": 41257168, "step": 30455 }, { "epoch": 0.9749695922156072, "grad_norm": 1.7899460792541504, "learning_rate": 6.6966780573187335e-06, "loss": 0.4615, "num_input_tokens_seen": 41264128, "step": 30460 }, { "epoch": 0.9751296331860957, "grad_norm": 0.6676501631736755, "learning_rate": 6.689992178086174e-06, "loss": 0.3436, "num_input_tokens_seen": 41270752, "step": 30465 }, { "epoch": 0.9752896741565841, "grad_norm": 0.867149829864502, "learning_rate": 6.683309122486925e-06, "loss": 0.3319, "num_input_tokens_seen": 41277392, "step": 30470 }, { "epoch": 0.9754497151270726, "grad_norm": 1.6216449737548828, "learning_rate": 6.676628891551584e-06, "loss": 0.4337, "num_input_tokens_seen": 41284000, "step": 30475 }, { "epoch": 0.975609756097561, "grad_norm": 0.9839417338371277, "learning_rate": 6.6699514863103385e-06, "loss": 0.3503, "num_input_tokens_seen": 41290816, "step": 30480 }, { "epoch": 0.9757697970680494, "grad_norm": 1.3410111665725708, "learning_rate": 6.663276907792921e-06, "loss": 0.4948, "num_input_tokens_seen": 41297504, "step": 30485 }, { "epoch": 0.9759298380385378, "grad_norm": 1.607400894165039, "learning_rate": 6.656605157028634e-06, "loss": 0.4586, "num_input_tokens_seen": 41303984, "step": 30490 }, { "epoch": 0.9760898790090263, "grad_norm": 1.118848443031311, "learning_rate": 6.649936235046358e-06, "loss": 0.4587, "num_input_tokens_seen": 41310432, "step": 30495 }, { "epoch": 0.9762499199795147, "grad_norm": 0.8951090574264526, "learning_rate": 6.643270142874508e-06, "loss": 0.3088, "num_input_tokens_seen": 41317120, "step": 30500 }, { "epoch": 0.9764099609500032, "grad_norm": 0.7267104983329773, "learning_rate": 6.636606881541094e-06, "loss": 0.6528, "num_input_tokens_seen": 41323600, "step": 30505 }, { "epoch": 0.9765700019204916, "grad_norm": 1.0456053018569946, "learning_rate": 6.629946452073662e-06, "loss": 0.6792, "num_input_tokens_seen": 41330256, "step": 30510 }, { "epoch": 0.9767300428909801, "grad_norm": 0.45813265442848206, "learning_rate": 6.6232888554993375e-06, "loss": 0.3829, "num_input_tokens_seen": 41336688, "step": 30515 }, { "epoch": 0.9768900838614686, "grad_norm": 0.5766186714172363, "learning_rate": 6.616634092844817e-06, "loss": 0.3891, "num_input_tokens_seen": 41343504, "step": 30520 }, { "epoch": 0.977050124831957, "grad_norm": 1.7966957092285156, "learning_rate": 6.609982165136331e-06, "loss": 0.4165, "num_input_tokens_seen": 41349920, "step": 30525 }, { "epoch": 0.9772101658024455, "grad_norm": 1.2053016424179077, "learning_rate": 6.603333073399706e-06, "loss": 0.5934, "num_input_tokens_seen": 41356816, "step": 30530 }, { "epoch": 0.9773702067729338, "grad_norm": 0.5871291160583496, "learning_rate": 6.596686818660308e-06, "loss": 0.4064, "num_input_tokens_seen": 41363808, "step": 30535 }, { "epoch": 0.9775302477434223, "grad_norm": 1.7941380739212036, "learning_rate": 6.590043401943066e-06, "loss": 0.4165, "num_input_tokens_seen": 41370928, "step": 30540 }, { "epoch": 0.9776902887139107, "grad_norm": 0.8436145186424255, "learning_rate": 6.583402824272494e-06, "loss": 0.5081, "num_input_tokens_seen": 41377936, "step": 30545 }, { "epoch": 0.9778503296843992, "grad_norm": 1.9722564220428467, "learning_rate": 6.576765086672634e-06, "loss": 0.4357, "num_input_tokens_seen": 41384704, "step": 30550 }, { "epoch": 0.9780103706548876, "grad_norm": 1.5073752403259277, "learning_rate": 6.57013019016712e-06, "loss": 0.5292, "num_input_tokens_seen": 41391440, "step": 30555 }, { "epoch": 0.9781704116253761, "grad_norm": 2.378268241882324, "learning_rate": 6.563498135779142e-06, "loss": 0.392, "num_input_tokens_seen": 41399136, "step": 30560 }, { "epoch": 0.9783304525958645, "grad_norm": 1.0848886966705322, "learning_rate": 6.556868924531431e-06, "loss": 0.3366, "num_input_tokens_seen": 41405760, "step": 30565 }, { "epoch": 0.978490493566353, "grad_norm": 1.0350288152694702, "learning_rate": 6.550242557446304e-06, "loss": 0.3566, "num_input_tokens_seen": 41412560, "step": 30570 }, { "epoch": 0.9786505345368415, "grad_norm": 1.3603718280792236, "learning_rate": 6.543619035545634e-06, "loss": 0.358, "num_input_tokens_seen": 41419040, "step": 30575 }, { "epoch": 0.9788105755073299, "grad_norm": 1.2406671047210693, "learning_rate": 6.53699835985084e-06, "loss": 0.4672, "num_input_tokens_seen": 41425760, "step": 30580 }, { "epoch": 0.9789706164778184, "grad_norm": 1.6960318088531494, "learning_rate": 6.530380531382927e-06, "loss": 0.4421, "num_input_tokens_seen": 41432432, "step": 30585 }, { "epoch": 0.9791306574483067, "grad_norm": 1.6929625272750854, "learning_rate": 6.523765551162433e-06, "loss": 0.3639, "num_input_tokens_seen": 41439104, "step": 30590 }, { "epoch": 0.9792906984187952, "grad_norm": 2.78501558303833, "learning_rate": 6.517153420209476e-06, "loss": 0.4575, "num_input_tokens_seen": 41445952, "step": 30595 }, { "epoch": 0.9794507393892836, "grad_norm": 1.171384572982788, "learning_rate": 6.510544139543739e-06, "loss": 0.3897, "num_input_tokens_seen": 41452688, "step": 30600 }, { "epoch": 0.9794507393892836, "eval_loss": 0.445690393447876, "eval_runtime": 501.7664, "eval_samples_per_second": 27.674, "eval_steps_per_second": 13.837, "num_input_tokens_seen": 41452688, "step": 30600 }, { "epoch": 0.9796107803597721, "grad_norm": 1.5316319465637207, "learning_rate": 6.503937710184452e-06, "loss": 0.4896, "num_input_tokens_seen": 41459408, "step": 30605 }, { "epoch": 0.9797708213302605, "grad_norm": 1.2698698043823242, "learning_rate": 6.4973341331503954e-06, "loss": 0.5197, "num_input_tokens_seen": 41466032, "step": 30610 }, { "epoch": 0.979930862300749, "grad_norm": 1.02549147605896, "learning_rate": 6.490733409459942e-06, "loss": 0.4123, "num_input_tokens_seen": 41472816, "step": 30615 }, { "epoch": 0.9800909032712375, "grad_norm": 1.3487932682037354, "learning_rate": 6.484135540130995e-06, "loss": 0.3573, "num_input_tokens_seen": 41479888, "step": 30620 }, { "epoch": 0.9802509442417259, "grad_norm": 1.0293954610824585, "learning_rate": 6.4775405261810364e-06, "loss": 0.3956, "num_input_tokens_seen": 41486800, "step": 30625 }, { "epoch": 0.9804109852122144, "grad_norm": 0.9150331616401672, "learning_rate": 6.470948368627092e-06, "loss": 0.2848, "num_input_tokens_seen": 41493648, "step": 30630 }, { "epoch": 0.9805710261827028, "grad_norm": 1.0535428524017334, "learning_rate": 6.464359068485756e-06, "loss": 0.3225, "num_input_tokens_seen": 41500080, "step": 30635 }, { "epoch": 0.9807310671531912, "grad_norm": 1.1888868808746338, "learning_rate": 6.457772626773195e-06, "loss": 0.5442, "num_input_tokens_seen": 41506912, "step": 30640 }, { "epoch": 0.9808911081236796, "grad_norm": 1.1894805431365967, "learning_rate": 6.451189044505104e-06, "loss": 0.5356, "num_input_tokens_seen": 41513392, "step": 30645 }, { "epoch": 0.9810511490941681, "grad_norm": 6.961277961730957, "learning_rate": 6.44460832269676e-06, "loss": 0.4564, "num_input_tokens_seen": 41520192, "step": 30650 }, { "epoch": 0.9812111900646565, "grad_norm": 0.8098112344741821, "learning_rate": 6.438030462363001e-06, "loss": 0.5944, "num_input_tokens_seen": 41526592, "step": 30655 }, { "epoch": 0.981371231035145, "grad_norm": 0.9655308127403259, "learning_rate": 6.431455464518205e-06, "loss": 0.3544, "num_input_tokens_seen": 41533408, "step": 30660 }, { "epoch": 0.9815312720056334, "grad_norm": 0.8079832196235657, "learning_rate": 6.424883330176326e-06, "loss": 0.2588, "num_input_tokens_seen": 41540128, "step": 30665 }, { "epoch": 0.9816913129761219, "grad_norm": 1.130310297012329, "learning_rate": 6.418314060350864e-06, "loss": 0.6384, "num_input_tokens_seen": 41546784, "step": 30670 }, { "epoch": 0.9818513539466104, "grad_norm": 1.1721150875091553, "learning_rate": 6.4117476560548895e-06, "loss": 0.4485, "num_input_tokens_seen": 41553808, "step": 30675 }, { "epoch": 0.9820113949170988, "grad_norm": 1.0286532640457153, "learning_rate": 6.405184118301016e-06, "loss": 0.531, "num_input_tokens_seen": 41560432, "step": 30680 }, { "epoch": 0.9821714358875873, "grad_norm": 1.0516339540481567, "learning_rate": 6.398623448101434e-06, "loss": 0.4551, "num_input_tokens_seen": 41567232, "step": 30685 }, { "epoch": 0.9823314768580756, "grad_norm": 0.3837610185146332, "learning_rate": 6.392065646467871e-06, "loss": 0.2549, "num_input_tokens_seen": 41573568, "step": 30690 }, { "epoch": 0.9824915178285641, "grad_norm": 1.2185453176498413, "learning_rate": 6.385510714411632e-06, "loss": 0.4389, "num_input_tokens_seen": 41580192, "step": 30695 }, { "epoch": 0.9826515587990525, "grad_norm": 0.702276885509491, "learning_rate": 6.378958652943559e-06, "loss": 0.32, "num_input_tokens_seen": 41586768, "step": 30700 }, { "epoch": 0.982811599769541, "grad_norm": 2.2059733867645264, "learning_rate": 6.3724094630740776e-06, "loss": 0.4326, "num_input_tokens_seen": 41593648, "step": 30705 }, { "epoch": 0.9829716407400294, "grad_norm": 0.7845094799995422, "learning_rate": 6.365863145813136e-06, "loss": 0.4309, "num_input_tokens_seen": 41600368, "step": 30710 }, { "epoch": 0.9831316817105179, "grad_norm": 1.1968613862991333, "learning_rate": 6.359319702170269e-06, "loss": 0.4359, "num_input_tokens_seen": 41607056, "step": 30715 }, { "epoch": 0.9832917226810063, "grad_norm": 1.3280103206634521, "learning_rate": 6.352779133154566e-06, "loss": 0.6453, "num_input_tokens_seen": 41613776, "step": 30720 }, { "epoch": 0.9834517636514948, "grad_norm": 1.4683239459991455, "learning_rate": 6.346241439774648e-06, "loss": 0.2543, "num_input_tokens_seen": 41620288, "step": 30725 }, { "epoch": 0.9836118046219833, "grad_norm": 1.374292254447937, "learning_rate": 6.339706623038716e-06, "loss": 0.36, "num_input_tokens_seen": 41627088, "step": 30730 }, { "epoch": 0.9837718455924717, "grad_norm": 1.0977721214294434, "learning_rate": 6.333174683954532e-06, "loss": 0.4385, "num_input_tokens_seen": 41633776, "step": 30735 }, { "epoch": 0.9839318865629602, "grad_norm": 0.8704602122306824, "learning_rate": 6.326645623529387e-06, "loss": 0.4285, "num_input_tokens_seen": 41640688, "step": 30740 }, { "epoch": 0.9840919275334485, "grad_norm": 0.8487563729286194, "learning_rate": 6.320119442770156e-06, "loss": 0.2685, "num_input_tokens_seen": 41647232, "step": 30745 }, { "epoch": 0.984251968503937, "grad_norm": 1.017513394355774, "learning_rate": 6.313596142683254e-06, "loss": 0.2695, "num_input_tokens_seen": 41654064, "step": 30750 }, { "epoch": 0.9844120094744254, "grad_norm": 0.5065831542015076, "learning_rate": 6.307075724274647e-06, "loss": 0.3654, "num_input_tokens_seen": 41660800, "step": 30755 }, { "epoch": 0.9845720504449139, "grad_norm": 1.3692450523376465, "learning_rate": 6.300558188549882e-06, "loss": 0.7546, "num_input_tokens_seen": 41667456, "step": 30760 }, { "epoch": 0.9847320914154023, "grad_norm": 2.6476709842681885, "learning_rate": 6.29404353651403e-06, "loss": 0.3885, "num_input_tokens_seen": 41674048, "step": 30765 }, { "epoch": 0.9848921323858908, "grad_norm": 1.0338984727859497, "learning_rate": 6.287531769171737e-06, "loss": 0.5605, "num_input_tokens_seen": 41680608, "step": 30770 }, { "epoch": 0.9850521733563792, "grad_norm": 1.3871488571166992, "learning_rate": 6.2810228875272045e-06, "loss": 0.5276, "num_input_tokens_seen": 41687056, "step": 30775 }, { "epoch": 0.9852122143268677, "grad_norm": 1.356458306312561, "learning_rate": 6.274516892584179e-06, "loss": 0.3082, "num_input_tokens_seen": 41693760, "step": 30780 }, { "epoch": 0.9853722552973562, "grad_norm": 1.9188599586486816, "learning_rate": 6.268013785345969e-06, "loss": 0.5283, "num_input_tokens_seen": 41700672, "step": 30785 }, { "epoch": 0.9855322962678446, "grad_norm": 1.516463041305542, "learning_rate": 6.26151356681543e-06, "loss": 0.2795, "num_input_tokens_seen": 41707312, "step": 30790 }, { "epoch": 0.985692337238333, "grad_norm": 0.8800028562545776, "learning_rate": 6.255016237994981e-06, "loss": 0.4712, "num_input_tokens_seen": 41714064, "step": 30795 }, { "epoch": 0.9858523782088214, "grad_norm": 0.9363225698471069, "learning_rate": 6.248521799886603e-06, "loss": 0.4594, "num_input_tokens_seen": 41721056, "step": 30800 }, { "epoch": 0.9858523782088214, "eval_loss": 0.44577673077583313, "eval_runtime": 501.0033, "eval_samples_per_second": 27.716, "eval_steps_per_second": 13.858, "num_input_tokens_seen": 41721056, "step": 30800 }, { "epoch": 0.9860124191793099, "grad_norm": 1.5999748706817627, "learning_rate": 6.242030253491798e-06, "loss": 0.5044, "num_input_tokens_seen": 41728128, "step": 30805 }, { "epoch": 0.9861724601497983, "grad_norm": 0.832691490650177, "learning_rate": 6.235541599811656e-06, "loss": 0.4367, "num_input_tokens_seen": 41734864, "step": 30810 }, { "epoch": 0.9863325011202868, "grad_norm": 1.2278368473052979, "learning_rate": 6.229055839846814e-06, "loss": 0.5277, "num_input_tokens_seen": 41741472, "step": 30815 }, { "epoch": 0.9864925420907752, "grad_norm": 0.9893569946289062, "learning_rate": 6.222572974597455e-06, "loss": 0.3193, "num_input_tokens_seen": 41747936, "step": 30820 }, { "epoch": 0.9866525830612637, "grad_norm": 0.7110177278518677, "learning_rate": 6.216093005063306e-06, "loss": 0.4713, "num_input_tokens_seen": 41754800, "step": 30825 }, { "epoch": 0.9868126240317522, "grad_norm": 1.095641851425171, "learning_rate": 6.209615932243678e-06, "loss": 0.3931, "num_input_tokens_seen": 41761600, "step": 30830 }, { "epoch": 0.9869726650022406, "grad_norm": 1.4423309564590454, "learning_rate": 6.203141757137399e-06, "loss": 0.3564, "num_input_tokens_seen": 41768496, "step": 30835 }, { "epoch": 0.9871327059727291, "grad_norm": 0.6323695182800293, "learning_rate": 6.196670480742886e-06, "loss": 0.3187, "num_input_tokens_seen": 41775264, "step": 30840 }, { "epoch": 0.9872927469432174, "grad_norm": 1.188193678855896, "learning_rate": 6.190202104058074e-06, "loss": 0.4253, "num_input_tokens_seen": 41782048, "step": 30845 }, { "epoch": 0.9874527879137059, "grad_norm": 0.9230363368988037, "learning_rate": 6.183736628080475e-06, "loss": 0.3388, "num_input_tokens_seen": 41789168, "step": 30850 }, { "epoch": 0.9876128288841943, "grad_norm": 1.3017799854278564, "learning_rate": 6.177274053807155e-06, "loss": 0.3989, "num_input_tokens_seen": 41795680, "step": 30855 }, { "epoch": 0.9877728698546828, "grad_norm": 1.1091880798339844, "learning_rate": 6.170814382234713e-06, "loss": 0.5003, "num_input_tokens_seen": 41802432, "step": 30860 }, { "epoch": 0.9879329108251712, "grad_norm": 2.274216651916504, "learning_rate": 6.16435761435932e-06, "loss": 0.7587, "num_input_tokens_seen": 41809280, "step": 30865 }, { "epoch": 0.9880929517956597, "grad_norm": 0.6413001418113708, "learning_rate": 6.157903751176681e-06, "loss": 0.3575, "num_input_tokens_seen": 41815968, "step": 30870 }, { "epoch": 0.9882529927661481, "grad_norm": 0.6014311909675598, "learning_rate": 6.151452793682066e-06, "loss": 0.4644, "num_input_tokens_seen": 41822672, "step": 30875 }, { "epoch": 0.9884130337366366, "grad_norm": 1.0056841373443604, "learning_rate": 6.145004742870305e-06, "loss": 0.5461, "num_input_tokens_seen": 41829888, "step": 30880 }, { "epoch": 0.9885730747071251, "grad_norm": 1.7159849405288696, "learning_rate": 6.138559599735752e-06, "loss": 0.5832, "num_input_tokens_seen": 41836464, "step": 30885 }, { "epoch": 0.9887331156776135, "grad_norm": 1.4584059715270996, "learning_rate": 6.132117365272344e-06, "loss": 0.3024, "num_input_tokens_seen": 41843312, "step": 30890 }, { "epoch": 0.988893156648102, "grad_norm": 1.236695408821106, "learning_rate": 6.125678040473545e-06, "loss": 0.4374, "num_input_tokens_seen": 41850320, "step": 30895 }, { "epoch": 0.9890531976185903, "grad_norm": 0.9269846081733704, "learning_rate": 6.1192416263323755e-06, "loss": 0.4446, "num_input_tokens_seen": 41857344, "step": 30900 }, { "epoch": 0.9892132385890788, "grad_norm": 0.8143202662467957, "learning_rate": 6.112808123841424e-06, "loss": 0.3693, "num_input_tokens_seen": 41864096, "step": 30905 }, { "epoch": 0.9893732795595672, "grad_norm": 1.2989996671676636, "learning_rate": 6.106377533992805e-06, "loss": 0.5269, "num_input_tokens_seen": 41870848, "step": 30910 }, { "epoch": 0.9895333205300557, "grad_norm": 1.1592174768447876, "learning_rate": 6.099949857778204e-06, "loss": 0.3322, "num_input_tokens_seen": 41877696, "step": 30915 }, { "epoch": 0.9896933615005441, "grad_norm": 0.6294031143188477, "learning_rate": 6.093525096188852e-06, "loss": 0.3541, "num_input_tokens_seen": 41884384, "step": 30920 }, { "epoch": 0.9898534024710326, "grad_norm": 2.4088809490203857, "learning_rate": 6.087103250215518e-06, "loss": 0.5901, "num_input_tokens_seen": 41890976, "step": 30925 }, { "epoch": 0.990013443441521, "grad_norm": 0.8307891488075256, "learning_rate": 6.080684320848537e-06, "loss": 0.4281, "num_input_tokens_seen": 41897760, "step": 30930 }, { "epoch": 0.9901734844120095, "grad_norm": 0.876263439655304, "learning_rate": 6.074268309077794e-06, "loss": 0.523, "num_input_tokens_seen": 41904752, "step": 30935 }, { "epoch": 0.990333525382498, "grad_norm": 1.2479394674301147, "learning_rate": 6.067855215892709e-06, "loss": 0.411, "num_input_tokens_seen": 41911552, "step": 30940 }, { "epoch": 0.9904935663529864, "grad_norm": 0.8556258678436279, "learning_rate": 6.061445042282271e-06, "loss": 0.4395, "num_input_tokens_seen": 41918384, "step": 30945 }, { "epoch": 0.9906536073234748, "grad_norm": 1.446900486946106, "learning_rate": 6.055037789234999e-06, "loss": 0.5215, "num_input_tokens_seen": 41925568, "step": 30950 }, { "epoch": 0.9908136482939632, "grad_norm": 1.2658387422561646, "learning_rate": 6.048633457738975e-06, "loss": 0.5399, "num_input_tokens_seen": 41932592, "step": 30955 }, { "epoch": 0.9909736892644517, "grad_norm": 1.646807074546814, "learning_rate": 6.042232048781837e-06, "loss": 0.5048, "num_input_tokens_seen": 41939264, "step": 30960 }, { "epoch": 0.9911337302349401, "grad_norm": 1.2716395854949951, "learning_rate": 6.035833563350757e-06, "loss": 0.4622, "num_input_tokens_seen": 41946080, "step": 30965 }, { "epoch": 0.9912937712054286, "grad_norm": 1.8633925914764404, "learning_rate": 6.0294380024324525e-06, "loss": 0.4596, "num_input_tokens_seen": 41952944, "step": 30970 }, { "epoch": 0.991453812175917, "grad_norm": 0.9348977208137512, "learning_rate": 6.023045367013213e-06, "loss": 0.346, "num_input_tokens_seen": 41959504, "step": 30975 }, { "epoch": 0.9916138531464055, "grad_norm": 0.4229811131954193, "learning_rate": 6.016655658078851e-06, "loss": 0.4563, "num_input_tokens_seen": 41966000, "step": 30980 }, { "epoch": 0.9917738941168939, "grad_norm": 1.536251187324524, "learning_rate": 6.010268876614753e-06, "loss": 0.396, "num_input_tokens_seen": 41972768, "step": 30985 }, { "epoch": 0.9919339350873824, "grad_norm": 1.0870556831359863, "learning_rate": 6.0038850236058266e-06, "loss": 0.3904, "num_input_tokens_seen": 41979584, "step": 30990 }, { "epoch": 0.9920939760578709, "grad_norm": 0.6724305152893066, "learning_rate": 5.997504100036549e-06, "loss": 0.4379, "num_input_tokens_seen": 41986784, "step": 30995 }, { "epoch": 0.9922540170283592, "grad_norm": 0.43206632137298584, "learning_rate": 5.991126106890949e-06, "loss": 0.3648, "num_input_tokens_seen": 41993584, "step": 31000 }, { "epoch": 0.9922540170283592, "eval_loss": 0.4453795850276947, "eval_runtime": 501.1576, "eval_samples_per_second": 27.708, "eval_steps_per_second": 13.854, "num_input_tokens_seen": 41993584, "step": 31000 }, { "epoch": 0.9924140579988477, "grad_norm": 1.3589123487472534, "learning_rate": 5.984751045152576e-06, "loss": 0.5037, "num_input_tokens_seen": 42000864, "step": 31005 }, { "epoch": 0.9925740989693361, "grad_norm": 0.4816036820411682, "learning_rate": 5.978378915804553e-06, "loss": 0.3428, "num_input_tokens_seen": 42008128, "step": 31010 }, { "epoch": 0.9927341399398246, "grad_norm": 0.4345457851886749, "learning_rate": 5.972009719829547e-06, "loss": 0.3017, "num_input_tokens_seen": 42014944, "step": 31015 }, { "epoch": 0.992894180910313, "grad_norm": 0.940072238445282, "learning_rate": 5.965643458209755e-06, "loss": 0.4426, "num_input_tokens_seen": 42021936, "step": 31020 }, { "epoch": 0.9930542218808015, "grad_norm": 1.9944168329238892, "learning_rate": 5.95928013192695e-06, "loss": 0.3504, "num_input_tokens_seen": 42028592, "step": 31025 }, { "epoch": 0.9932142628512899, "grad_norm": 0.7064417600631714, "learning_rate": 5.952919741962423e-06, "loss": 0.4232, "num_input_tokens_seen": 42035216, "step": 31030 }, { "epoch": 0.9933743038217784, "grad_norm": 1.4450619220733643, "learning_rate": 5.946562289297042e-06, "loss": 0.3372, "num_input_tokens_seen": 42041936, "step": 31035 }, { "epoch": 0.9935343447922668, "grad_norm": 1.7914341688156128, "learning_rate": 5.9402077749111855e-06, "loss": 0.5809, "num_input_tokens_seen": 42048832, "step": 31040 }, { "epoch": 0.9936943857627553, "grad_norm": 0.8997992873191833, "learning_rate": 5.933856199784821e-06, "loss": 0.2665, "num_input_tokens_seen": 42055840, "step": 31045 }, { "epoch": 0.9938544267332438, "grad_norm": 0.6693968772888184, "learning_rate": 5.927507564897419e-06, "loss": 0.3378, "num_input_tokens_seen": 42062960, "step": 31050 }, { "epoch": 0.9940144677037321, "grad_norm": 0.5640976428985596, "learning_rate": 5.9211618712280395e-06, "loss": 0.465, "num_input_tokens_seen": 42069456, "step": 31055 }, { "epoch": 0.9941745086742206, "grad_norm": 2.4292712211608887, "learning_rate": 5.914819119755255e-06, "loss": 0.4176, "num_input_tokens_seen": 42075952, "step": 31060 }, { "epoch": 0.994334549644709, "grad_norm": 0.9103301167488098, "learning_rate": 5.908479311457205e-06, "loss": 0.3345, "num_input_tokens_seen": 42083136, "step": 31065 }, { "epoch": 0.9944945906151975, "grad_norm": 1.1921007633209229, "learning_rate": 5.902142447311559e-06, "loss": 0.6017, "num_input_tokens_seen": 42090048, "step": 31070 }, { "epoch": 0.9946546315856859, "grad_norm": 0.7005053758621216, "learning_rate": 5.895808528295546e-06, "loss": 0.3402, "num_input_tokens_seen": 42096624, "step": 31075 }, { "epoch": 0.9948146725561744, "grad_norm": 0.5902649164199829, "learning_rate": 5.889477555385941e-06, "loss": 0.3827, "num_input_tokens_seen": 42103184, "step": 31080 }, { "epoch": 0.9949747135266628, "grad_norm": 0.9239402413368225, "learning_rate": 5.883149529559051e-06, "loss": 0.4823, "num_input_tokens_seen": 42110272, "step": 31085 }, { "epoch": 0.9951347544971513, "grad_norm": 1.280263900756836, "learning_rate": 5.876824451790738e-06, "loss": 0.5015, "num_input_tokens_seen": 42117216, "step": 31090 }, { "epoch": 0.9952947954676398, "grad_norm": 0.28580427169799805, "learning_rate": 5.87050232305642e-06, "loss": 0.2503, "num_input_tokens_seen": 42123856, "step": 31095 }, { "epoch": 0.9954548364381282, "grad_norm": 0.618948757648468, "learning_rate": 5.864183144331034e-06, "loss": 0.4168, "num_input_tokens_seen": 42130800, "step": 31100 }, { "epoch": 0.9956148774086166, "grad_norm": 0.7555756568908691, "learning_rate": 5.857866916589089e-06, "loss": 0.3826, "num_input_tokens_seen": 42138240, "step": 31105 }, { "epoch": 0.995774918379105, "grad_norm": 1.2948495149612427, "learning_rate": 5.8515536408046216e-06, "loss": 0.6599, "num_input_tokens_seen": 42144576, "step": 31110 }, { "epoch": 0.9959349593495935, "grad_norm": 1.2020784616470337, "learning_rate": 5.845243317951208e-06, "loss": 0.3507, "num_input_tokens_seen": 42151728, "step": 31115 }, { "epoch": 0.9960950003200819, "grad_norm": 0.79340660572052, "learning_rate": 5.838935949001997e-06, "loss": 0.2863, "num_input_tokens_seen": 42158864, "step": 31120 }, { "epoch": 0.9962550412905704, "grad_norm": 1.2789652347564697, "learning_rate": 5.8326315349296476e-06, "loss": 0.3701, "num_input_tokens_seen": 42165648, "step": 31125 }, { "epoch": 0.9964150822610588, "grad_norm": 0.6513067483901978, "learning_rate": 5.826330076706396e-06, "loss": 0.4749, "num_input_tokens_seen": 42172384, "step": 31130 }, { "epoch": 0.9965751232315473, "grad_norm": 1.565772294998169, "learning_rate": 5.820031575303988e-06, "loss": 0.4675, "num_input_tokens_seen": 42179200, "step": 31135 }, { "epoch": 0.9967351642020357, "grad_norm": 0.9397836923599243, "learning_rate": 5.813736031693745e-06, "loss": 0.4402, "num_input_tokens_seen": 42185904, "step": 31140 }, { "epoch": 0.9968952051725242, "grad_norm": 1.1123874187469482, "learning_rate": 5.807443446846522e-06, "loss": 0.5664, "num_input_tokens_seen": 42192592, "step": 31145 }, { "epoch": 0.9970552461430127, "grad_norm": 0.5799927711486816, "learning_rate": 5.801153821732699e-06, "loss": 0.3513, "num_input_tokens_seen": 42199024, "step": 31150 }, { "epoch": 0.997215287113501, "grad_norm": 0.9659283757209778, "learning_rate": 5.794867157322229e-06, "loss": 0.4947, "num_input_tokens_seen": 42206496, "step": 31155 }, { "epoch": 0.9973753280839895, "grad_norm": 1.8757792711257935, "learning_rate": 5.788583454584593e-06, "loss": 0.4789, "num_input_tokens_seen": 42213152, "step": 31160 }, { "epoch": 0.9975353690544779, "grad_norm": 0.7584568858146667, "learning_rate": 5.7823027144888075e-06, "loss": 0.4237, "num_input_tokens_seen": 42219744, "step": 31165 }, { "epoch": 0.9976954100249664, "grad_norm": 1.2403180599212646, "learning_rate": 5.776024938003455e-06, "loss": 0.5146, "num_input_tokens_seen": 42226304, "step": 31170 }, { "epoch": 0.9978554509954548, "grad_norm": 0.734687089920044, "learning_rate": 5.7697501260966345e-06, "loss": 0.3873, "num_input_tokens_seen": 42232960, "step": 31175 }, { "epoch": 0.9980154919659433, "grad_norm": 0.8196642398834229, "learning_rate": 5.7634782797360145e-06, "loss": 0.4058, "num_input_tokens_seen": 42239568, "step": 31180 }, { "epoch": 0.9981755329364317, "grad_norm": 3.3226754665374756, "learning_rate": 5.757209399888777e-06, "loss": 0.4057, "num_input_tokens_seen": 42246624, "step": 31185 }, { "epoch": 0.9983355739069202, "grad_norm": 1.0242668390274048, "learning_rate": 5.750943487521679e-06, "loss": 0.4521, "num_input_tokens_seen": 42253376, "step": 31190 }, { "epoch": 0.9984956148774086, "grad_norm": 0.9957196712493896, "learning_rate": 5.744680543600986e-06, "loss": 0.4553, "num_input_tokens_seen": 42259968, "step": 31195 }, { "epoch": 0.9986556558478971, "grad_norm": 1.5246458053588867, "learning_rate": 5.738420569092537e-06, "loss": 0.4395, "num_input_tokens_seen": 42266304, "step": 31200 }, { "epoch": 0.9986556558478971, "eval_loss": 0.4456407427787781, "eval_runtime": 500.9109, "eval_samples_per_second": 27.721, "eval_steps_per_second": 13.861, "num_input_tokens_seen": 42266304, "step": 31200 }, { "epoch": 0.9988156968183856, "grad_norm": 0.9095137715339661, "learning_rate": 5.732163564961684e-06, "loss": 0.4584, "num_input_tokens_seen": 42273360, "step": 31205 }, { "epoch": 0.9989757377888739, "grad_norm": 1.4562259912490845, "learning_rate": 5.725909532173354e-06, "loss": 0.3872, "num_input_tokens_seen": 42279920, "step": 31210 }, { "epoch": 0.9991357787593624, "grad_norm": 0.8921005725860596, "learning_rate": 5.719658471691977e-06, "loss": 0.3951, "num_input_tokens_seen": 42287072, "step": 31215 }, { "epoch": 0.9992958197298508, "grad_norm": 1.4483238458633423, "learning_rate": 5.71341038448156e-06, "loss": 0.4169, "num_input_tokens_seen": 42294064, "step": 31220 }, { "epoch": 0.9994558607003393, "grad_norm": 0.6083152294158936, "learning_rate": 5.707165271505635e-06, "loss": 0.4373, "num_input_tokens_seen": 42300832, "step": 31225 }, { "epoch": 0.9996159016708277, "grad_norm": 0.6133499145507812, "learning_rate": 5.700923133727271e-06, "loss": 0.4107, "num_input_tokens_seen": 42308272, "step": 31230 }, { "epoch": 0.9997759426413162, "grad_norm": 0.7265134453773499, "learning_rate": 5.694683972109083e-06, "loss": 0.2927, "num_input_tokens_seen": 42314800, "step": 31235 }, { "epoch": 0.9999359836118046, "grad_norm": 0.9158039689064026, "learning_rate": 5.688447787613241e-06, "loss": 0.431, "num_input_tokens_seen": 42321728, "step": 31240 }, { "epoch": 1.000096024582293, "grad_norm": 0.6086492538452148, "learning_rate": 5.6822145812014285e-06, "loss": 0.252, "num_input_tokens_seen": 42328064, "step": 31245 }, { "epoch": 1.0002560655527815, "grad_norm": 1.7223200798034668, "learning_rate": 5.675984353834896e-06, "loss": 0.51, "num_input_tokens_seen": 42334896, "step": 31250 }, { "epoch": 1.00041610652327, "grad_norm": 0.9262574911117554, "learning_rate": 5.66975710647441e-06, "loss": 0.4586, "num_input_tokens_seen": 42341760, "step": 31255 }, { "epoch": 1.0005761474937584, "grad_norm": 0.4695618450641632, "learning_rate": 5.663532840080304e-06, "loss": 0.4558, "num_input_tokens_seen": 42348448, "step": 31260 }, { "epoch": 1.000736188464247, "grad_norm": 0.8580643534660339, "learning_rate": 5.6573115556124325e-06, "loss": 0.4473, "num_input_tokens_seen": 42355344, "step": 31265 }, { "epoch": 1.0008962294347352, "grad_norm": 1.2768162488937378, "learning_rate": 5.651093254030185e-06, "loss": 0.3925, "num_input_tokens_seen": 42361968, "step": 31270 }, { "epoch": 1.0010562704052237, "grad_norm": 2.914611577987671, "learning_rate": 5.644877936292514e-06, "loss": 0.4467, "num_input_tokens_seen": 42368944, "step": 31275 }, { "epoch": 1.0012163113757122, "grad_norm": 1.1882870197296143, "learning_rate": 5.638665603357901e-06, "loss": 0.4162, "num_input_tokens_seen": 42375600, "step": 31280 }, { "epoch": 1.0013763523462007, "grad_norm": 0.8105599880218506, "learning_rate": 5.632456256184357e-06, "loss": 0.4312, "num_input_tokens_seen": 42382784, "step": 31285 }, { "epoch": 1.001536393316689, "grad_norm": 1.2732245922088623, "learning_rate": 5.626249895729452e-06, "loss": 0.5703, "num_input_tokens_seen": 42389856, "step": 31290 }, { "epoch": 1.0016964342871775, "grad_norm": 0.6595498919487, "learning_rate": 5.620046522950273e-06, "loss": 0.3589, "num_input_tokens_seen": 42396384, "step": 31295 }, { "epoch": 1.001856475257666, "grad_norm": 1.2382923364639282, "learning_rate": 5.613846138803464e-06, "loss": 0.3582, "num_input_tokens_seen": 42403072, "step": 31300 }, { "epoch": 1.0020165162281545, "grad_norm": 5.605134010314941, "learning_rate": 5.607648744245206e-06, "loss": 0.4214, "num_input_tokens_seen": 42409648, "step": 31305 }, { "epoch": 1.002176557198643, "grad_norm": 0.8961493968963623, "learning_rate": 5.601454340231207e-06, "loss": 0.4135, "num_input_tokens_seen": 42416176, "step": 31310 }, { "epoch": 1.0023365981691312, "grad_norm": 1.2817256450653076, "learning_rate": 5.595262927716724e-06, "loss": 0.3849, "num_input_tokens_seen": 42422976, "step": 31315 }, { "epoch": 1.0024966391396197, "grad_norm": 1.199858546257019, "learning_rate": 5.589074507656561e-06, "loss": 0.5103, "num_input_tokens_seen": 42429744, "step": 31320 }, { "epoch": 1.0026566801101082, "grad_norm": 1.4189008474349976, "learning_rate": 5.582889081005044e-06, "loss": 0.4375, "num_input_tokens_seen": 42436432, "step": 31325 }, { "epoch": 1.0028167210805967, "grad_norm": 0.7300707697868347, "learning_rate": 5.5767066487160316e-06, "loss": 0.2888, "num_input_tokens_seen": 42442816, "step": 31330 }, { "epoch": 1.002976762051085, "grad_norm": 0.6395755410194397, "learning_rate": 5.570527211742949e-06, "loss": 0.3641, "num_input_tokens_seen": 42449376, "step": 31335 }, { "epoch": 1.0031368030215735, "grad_norm": 1.5808132886886597, "learning_rate": 5.564350771038731e-06, "loss": 0.4671, "num_input_tokens_seen": 42456240, "step": 31340 }, { "epoch": 1.003296843992062, "grad_norm": 1.1805752515792847, "learning_rate": 5.558177327555875e-06, "loss": 0.5154, "num_input_tokens_seen": 42462656, "step": 31345 }, { "epoch": 1.0034568849625505, "grad_norm": 1.0817821025848389, "learning_rate": 5.552006882246388e-06, "loss": 0.3731, "num_input_tokens_seen": 42469152, "step": 31350 }, { "epoch": 1.0036169259330388, "grad_norm": 0.5460583567619324, "learning_rate": 5.545839436061839e-06, "loss": 0.4271, "num_input_tokens_seen": 42476048, "step": 31355 }, { "epoch": 1.0037769669035272, "grad_norm": 1.0268394947052002, "learning_rate": 5.539674989953331e-06, "loss": 0.4931, "num_input_tokens_seen": 42482848, "step": 31360 }, { "epoch": 1.0039370078740157, "grad_norm": 0.6136949062347412, "learning_rate": 5.533513544871488e-06, "loss": 0.4077, "num_input_tokens_seen": 42489888, "step": 31365 }, { "epoch": 1.0040970488445042, "grad_norm": 0.5411801338195801, "learning_rate": 5.527355101766493e-06, "loss": 0.289, "num_input_tokens_seen": 42496464, "step": 31370 }, { "epoch": 1.0042570898149927, "grad_norm": 1.7788398265838623, "learning_rate": 5.521199661588044e-06, "loss": 0.6087, "num_input_tokens_seen": 42502912, "step": 31375 }, { "epoch": 1.004417130785481, "grad_norm": 0.7336798906326294, "learning_rate": 5.5150472252853944e-06, "loss": 0.2767, "num_input_tokens_seen": 42509520, "step": 31380 }, { "epoch": 1.0045771717559695, "grad_norm": 2.083256959915161, "learning_rate": 5.50889779380733e-06, "loss": 0.3259, "num_input_tokens_seen": 42515872, "step": 31385 }, { "epoch": 1.004737212726458, "grad_norm": 1.3187085390090942, "learning_rate": 5.5027513681021605e-06, "loss": 0.5031, "num_input_tokens_seen": 42522352, "step": 31390 }, { "epoch": 1.0048972536969465, "grad_norm": 0.8585872650146484, "learning_rate": 5.4966079491177545e-06, "loss": 0.4077, "num_input_tokens_seen": 42529104, "step": 31395 }, { "epoch": 1.0050572946674348, "grad_norm": 1.6713510751724243, "learning_rate": 5.490467537801491e-06, "loss": 0.3717, "num_input_tokens_seen": 42536720, "step": 31400 }, { "epoch": 1.0050572946674348, "eval_loss": 0.44537195563316345, "eval_runtime": 501.5697, "eval_samples_per_second": 27.685, "eval_steps_per_second": 13.843, "num_input_tokens_seen": 42536720, "step": 31400 }, { "epoch": 1.0052173356379233, "grad_norm": 1.578993558883667, "learning_rate": 5.484330135100313e-06, "loss": 0.6355, "num_input_tokens_seen": 42544096, "step": 31405 }, { "epoch": 1.0053773766084118, "grad_norm": 1.2393903732299805, "learning_rate": 5.4781957419606785e-06, "loss": 0.3438, "num_input_tokens_seen": 42550832, "step": 31410 }, { "epoch": 1.0055374175789002, "grad_norm": 1.4198237657546997, "learning_rate": 5.472064359328577e-06, "loss": 0.4344, "num_input_tokens_seen": 42558336, "step": 31415 }, { "epoch": 1.0056974585493887, "grad_norm": 2.053567409515381, "learning_rate": 5.4659359881495565e-06, "loss": 0.4241, "num_input_tokens_seen": 42565536, "step": 31420 }, { "epoch": 1.005857499519877, "grad_norm": 0.9955751895904541, "learning_rate": 5.4598106293686916e-06, "loss": 0.4262, "num_input_tokens_seen": 42572208, "step": 31425 }, { "epoch": 1.0060175404903655, "grad_norm": 2.1548502445220947, "learning_rate": 5.45368828393058e-06, "loss": 0.4201, "num_input_tokens_seen": 42578816, "step": 31430 }, { "epoch": 1.006177581460854, "grad_norm": 1.3106170892715454, "learning_rate": 5.44756895277937e-06, "loss": 0.4828, "num_input_tokens_seen": 42585520, "step": 31435 }, { "epoch": 1.0063376224313425, "grad_norm": 0.6091107726097107, "learning_rate": 5.441452636858746e-06, "loss": 0.3758, "num_input_tokens_seen": 42592192, "step": 31440 }, { "epoch": 1.0064976634018308, "grad_norm": 0.9442325830459595, "learning_rate": 5.435339337111905e-06, "loss": 0.4698, "num_input_tokens_seen": 42599040, "step": 31445 }, { "epoch": 1.0066577043723193, "grad_norm": 0.8319540619850159, "learning_rate": 5.42922905448161e-06, "loss": 0.2771, "num_input_tokens_seen": 42605856, "step": 31450 }, { "epoch": 1.0068177453428078, "grad_norm": 2.5325522422790527, "learning_rate": 5.423121789910129e-06, "loss": 0.5098, "num_input_tokens_seen": 42612752, "step": 31455 }, { "epoch": 1.0069777863132963, "grad_norm": 0.7632439732551575, "learning_rate": 5.417017544339287e-06, "loss": 0.5871, "num_input_tokens_seen": 42620032, "step": 31460 }, { "epoch": 1.0071378272837848, "grad_norm": 0.8105471134185791, "learning_rate": 5.410916318710443e-06, "loss": 0.6031, "num_input_tokens_seen": 42626992, "step": 31465 }, { "epoch": 1.007297868254273, "grad_norm": 1.1565766334533691, "learning_rate": 5.404818113964466e-06, "loss": 0.3203, "num_input_tokens_seen": 42634016, "step": 31470 }, { "epoch": 1.0074579092247615, "grad_norm": 1.1089942455291748, "learning_rate": 5.398722931041792e-06, "loss": 0.3904, "num_input_tokens_seen": 42640784, "step": 31475 }, { "epoch": 1.00761795019525, "grad_norm": 0.8894555568695068, "learning_rate": 5.392630770882367e-06, "loss": 0.3615, "num_input_tokens_seen": 42648080, "step": 31480 }, { "epoch": 1.0077779911657385, "grad_norm": 1.343505620956421, "learning_rate": 5.3865416344256705e-06, "loss": 0.3939, "num_input_tokens_seen": 42655072, "step": 31485 }, { "epoch": 1.0079380321362268, "grad_norm": 1.0969749689102173, "learning_rate": 5.380455522610742e-06, "loss": 0.4353, "num_input_tokens_seen": 42661728, "step": 31490 }, { "epoch": 1.0080980731067153, "grad_norm": 1.0579112768173218, "learning_rate": 5.374372436376116e-06, "loss": 0.3421, "num_input_tokens_seen": 42668160, "step": 31495 }, { "epoch": 1.0082581140772038, "grad_norm": 1.1856166124343872, "learning_rate": 5.368292376659895e-06, "loss": 0.4957, "num_input_tokens_seen": 42674736, "step": 31500 }, { "epoch": 1.0084181550476923, "grad_norm": 0.9646955728530884, "learning_rate": 5.362215344399701e-06, "loss": 0.4112, "num_input_tokens_seen": 42681648, "step": 31505 }, { "epoch": 1.0085781960181806, "grad_norm": 1.032920002937317, "learning_rate": 5.356141340532678e-06, "loss": 0.3646, "num_input_tokens_seen": 42688176, "step": 31510 }, { "epoch": 1.008738236988669, "grad_norm": 1.1826317310333252, "learning_rate": 5.350070365995522e-06, "loss": 0.321, "num_input_tokens_seen": 42695360, "step": 31515 }, { "epoch": 1.0088982779591575, "grad_norm": 1.2019803524017334, "learning_rate": 5.344002421724459e-06, "loss": 0.6263, "num_input_tokens_seen": 42701952, "step": 31520 }, { "epoch": 1.009058318929646, "grad_norm": 0.5078164339065552, "learning_rate": 5.337937508655228e-06, "loss": 0.3615, "num_input_tokens_seen": 42708752, "step": 31525 }, { "epoch": 1.0092183599001345, "grad_norm": 0.6504508256912231, "learning_rate": 5.331875627723126e-06, "loss": 0.3912, "num_input_tokens_seen": 42715488, "step": 31530 }, { "epoch": 1.0093784008706228, "grad_norm": 2.3341870307922363, "learning_rate": 5.325816779862963e-06, "loss": 0.4309, "num_input_tokens_seen": 42722144, "step": 31535 }, { "epoch": 1.0095384418411113, "grad_norm": 0.6671712398529053, "learning_rate": 5.319760966009102e-06, "loss": 0.2715, "num_input_tokens_seen": 42729024, "step": 31540 }, { "epoch": 1.0096984828115998, "grad_norm": 0.9101867079734802, "learning_rate": 5.3137081870954096e-06, "loss": 0.3067, "num_input_tokens_seen": 42735712, "step": 31545 }, { "epoch": 1.0098585237820883, "grad_norm": 1.5090556144714355, "learning_rate": 5.307658444055313e-06, "loss": 0.5501, "num_input_tokens_seen": 42742448, "step": 31550 }, { "epoch": 1.0100185647525766, "grad_norm": 1.1343148946762085, "learning_rate": 5.301611737821749e-06, "loss": 0.4654, "num_input_tokens_seen": 42749456, "step": 31555 }, { "epoch": 1.010178605723065, "grad_norm": 0.6558846235275269, "learning_rate": 5.295568069327206e-06, "loss": 0.2702, "num_input_tokens_seen": 42756544, "step": 31560 }, { "epoch": 1.0103386466935536, "grad_norm": 1.0771875381469727, "learning_rate": 5.289527439503683e-06, "loss": 0.475, "num_input_tokens_seen": 42763216, "step": 31565 }, { "epoch": 1.010498687664042, "grad_norm": 1.3546149730682373, "learning_rate": 5.28348984928273e-06, "loss": 0.3886, "num_input_tokens_seen": 42769968, "step": 31570 }, { "epoch": 1.0106587286345305, "grad_norm": 0.9112178683280945, "learning_rate": 5.27745529959541e-06, "loss": 0.2989, "num_input_tokens_seen": 42776416, "step": 31575 }, { "epoch": 1.0108187696050188, "grad_norm": 0.6591923236846924, "learning_rate": 5.271423791372335e-06, "loss": 0.492, "num_input_tokens_seen": 42783392, "step": 31580 }, { "epoch": 1.0109788105755073, "grad_norm": 1.280197262763977, "learning_rate": 5.26539532554364e-06, "loss": 0.4713, "num_input_tokens_seen": 42790128, "step": 31585 }, { "epoch": 1.0111388515459958, "grad_norm": 1.195160984992981, "learning_rate": 5.25936990303898e-06, "loss": 0.482, "num_input_tokens_seen": 42797344, "step": 31590 }, { "epoch": 1.0112988925164843, "grad_norm": 1.4191516637802124, "learning_rate": 5.253347524787555e-06, "loss": 0.3477, "num_input_tokens_seen": 42804080, "step": 31595 }, { "epoch": 1.0114589334869726, "grad_norm": 0.6198561787605286, "learning_rate": 5.2473281917181035e-06, "loss": 0.3335, "num_input_tokens_seen": 42810528, "step": 31600 }, { "epoch": 1.0114589334869726, "eval_loss": 0.44565606117248535, "eval_runtime": 501.3378, "eval_samples_per_second": 27.698, "eval_steps_per_second": 13.849, "num_input_tokens_seen": 42810528, "step": 31600 }, { "epoch": 1.011618974457461, "grad_norm": 0.8523576259613037, "learning_rate": 5.241311904758864e-06, "loss": 0.3268, "num_input_tokens_seen": 42817248, "step": 31605 }, { "epoch": 1.0117790154279496, "grad_norm": 0.48575425148010254, "learning_rate": 5.23529866483764e-06, "loss": 0.3775, "num_input_tokens_seen": 42823936, "step": 31610 }, { "epoch": 1.011939056398438, "grad_norm": 0.8616943955421448, "learning_rate": 5.229288472881732e-06, "loss": 0.3675, "num_input_tokens_seen": 42831072, "step": 31615 }, { "epoch": 1.0120990973689263, "grad_norm": 0.5186178684234619, "learning_rate": 5.2232813298180025e-06, "loss": 0.3196, "num_input_tokens_seen": 42837840, "step": 31620 }, { "epoch": 1.0122591383394148, "grad_norm": 5.314910888671875, "learning_rate": 5.217277236572824e-06, "loss": 0.5975, "num_input_tokens_seen": 42844736, "step": 31625 }, { "epoch": 1.0124191793099033, "grad_norm": 0.7313738465309143, "learning_rate": 5.211276194072093e-06, "loss": 0.3222, "num_input_tokens_seen": 42851536, "step": 31630 }, { "epoch": 1.0125792202803918, "grad_norm": 1.2444413900375366, "learning_rate": 5.205278203241254e-06, "loss": 0.334, "num_input_tokens_seen": 42858144, "step": 31635 }, { "epoch": 1.0127392612508803, "grad_norm": 1.2944804430007935, "learning_rate": 5.199283265005278e-06, "loss": 0.4797, "num_input_tokens_seen": 42864576, "step": 31640 }, { "epoch": 1.0128993022213686, "grad_norm": 0.4896613657474518, "learning_rate": 5.193291380288648e-06, "loss": 0.2986, "num_input_tokens_seen": 42871280, "step": 31645 }, { "epoch": 1.013059343191857, "grad_norm": 0.9621142745018005, "learning_rate": 5.1873025500153995e-06, "loss": 0.4521, "num_input_tokens_seen": 42878272, "step": 31650 }, { "epoch": 1.0132193841623456, "grad_norm": 0.6306197643280029, "learning_rate": 5.181316775109071e-06, "loss": 0.3721, "num_input_tokens_seen": 42884976, "step": 31655 }, { "epoch": 1.013379425132834, "grad_norm": 0.5738018751144409, "learning_rate": 5.1753340564927564e-06, "loss": 0.3279, "num_input_tokens_seen": 42891888, "step": 31660 }, { "epoch": 1.0135394661033224, "grad_norm": 0.7671627402305603, "learning_rate": 5.169354395089068e-06, "loss": 0.4334, "num_input_tokens_seen": 42898352, "step": 31665 }, { "epoch": 1.0136995070738108, "grad_norm": 0.6781479120254517, "learning_rate": 5.1633777918201346e-06, "loss": 0.412, "num_input_tokens_seen": 42905296, "step": 31670 }, { "epoch": 1.0138595480442993, "grad_norm": 0.50380939245224, "learning_rate": 5.157404247607625e-06, "loss": 0.3536, "num_input_tokens_seen": 42911648, "step": 31675 }, { "epoch": 1.0140195890147878, "grad_norm": 0.8476006984710693, "learning_rate": 5.1514337633727454e-06, "loss": 0.3491, "num_input_tokens_seen": 42918256, "step": 31680 }, { "epoch": 1.0141796299852763, "grad_norm": 0.8537139892578125, "learning_rate": 5.145466340036206e-06, "loss": 0.4698, "num_input_tokens_seen": 42924880, "step": 31685 }, { "epoch": 1.0143396709557646, "grad_norm": 1.3440275192260742, "learning_rate": 5.139501978518274e-06, "loss": 0.618, "num_input_tokens_seen": 42931328, "step": 31690 }, { "epoch": 1.014499711926253, "grad_norm": 0.7397693395614624, "learning_rate": 5.133540679738716e-06, "loss": 0.4953, "num_input_tokens_seen": 42938368, "step": 31695 }, { "epoch": 1.0146597528967416, "grad_norm": 1.035059928894043, "learning_rate": 5.127582444616838e-06, "loss": 0.4732, "num_input_tokens_seen": 42944848, "step": 31700 }, { "epoch": 1.01481979386723, "grad_norm": 1.304296612739563, "learning_rate": 5.121627274071486e-06, "loss": 0.5893, "num_input_tokens_seen": 42951440, "step": 31705 }, { "epoch": 1.0149798348377184, "grad_norm": 2.0853540897369385, "learning_rate": 5.115675169021009e-06, "loss": 0.6651, "num_input_tokens_seen": 42958336, "step": 31710 }, { "epoch": 1.0151398758082069, "grad_norm": 1.8042218685150146, "learning_rate": 5.1097261303832994e-06, "loss": 0.6331, "num_input_tokens_seen": 42965312, "step": 31715 }, { "epoch": 1.0152999167786954, "grad_norm": 1.4842718839645386, "learning_rate": 5.103780159075788e-06, "loss": 0.3594, "num_input_tokens_seen": 42971968, "step": 31720 }, { "epoch": 1.0154599577491839, "grad_norm": 1.2732831239700317, "learning_rate": 5.0978372560154e-06, "loss": 0.5963, "num_input_tokens_seen": 42979040, "step": 31725 }, { "epoch": 1.0156199987196723, "grad_norm": 3.1601033210754395, "learning_rate": 5.091897422118619e-06, "loss": 0.4271, "num_input_tokens_seen": 42986048, "step": 31730 }, { "epoch": 1.0157800396901606, "grad_norm": 0.6781535148620605, "learning_rate": 5.0859606583014305e-06, "loss": 0.2152, "num_input_tokens_seen": 42992880, "step": 31735 }, { "epoch": 1.0159400806606491, "grad_norm": 0.9836328625679016, "learning_rate": 5.080026965479365e-06, "loss": 0.4408, "num_input_tokens_seen": 42999392, "step": 31740 }, { "epoch": 1.0161001216311376, "grad_norm": 0.7362932562828064, "learning_rate": 5.074096344567475e-06, "loss": 0.5594, "num_input_tokens_seen": 43006272, "step": 31745 }, { "epoch": 1.016260162601626, "grad_norm": 0.67617267370224, "learning_rate": 5.0681687964803294e-06, "loss": 0.3762, "num_input_tokens_seen": 43013328, "step": 31750 }, { "epoch": 1.0164202035721144, "grad_norm": 1.4360413551330566, "learning_rate": 5.06224432213204e-06, "loss": 0.3567, "num_input_tokens_seen": 43019856, "step": 31755 }, { "epoch": 1.0165802445426029, "grad_norm": 0.9780874252319336, "learning_rate": 5.056322922436224e-06, "loss": 0.4487, "num_input_tokens_seen": 43026512, "step": 31760 }, { "epoch": 1.0167402855130914, "grad_norm": 1.8378485441207886, "learning_rate": 5.0504045983060465e-06, "loss": 0.4833, "num_input_tokens_seen": 43033248, "step": 31765 }, { "epoch": 1.0169003264835799, "grad_norm": 0.7816437482833862, "learning_rate": 5.044489350654183e-06, "loss": 0.3908, "num_input_tokens_seen": 43039952, "step": 31770 }, { "epoch": 1.0170603674540681, "grad_norm": 0.9591552019119263, "learning_rate": 5.038577180392831e-06, "loss": 0.5338, "num_input_tokens_seen": 43047024, "step": 31775 }, { "epoch": 1.0172204084245566, "grad_norm": 1.1571815013885498, "learning_rate": 5.032668088433729e-06, "loss": 0.5025, "num_input_tokens_seen": 43053456, "step": 31780 }, { "epoch": 1.0173804493950451, "grad_norm": 2.49697208404541, "learning_rate": 5.02676207568814e-06, "loss": 0.6822, "num_input_tokens_seen": 43060416, "step": 31785 }, { "epoch": 1.0175404903655336, "grad_norm": 1.4488673210144043, "learning_rate": 5.02085914306683e-06, "loss": 0.4252, "num_input_tokens_seen": 43066944, "step": 31790 }, { "epoch": 1.0177005313360221, "grad_norm": 0.6712659001350403, "learning_rate": 5.014959291480123e-06, "loss": 0.4214, "num_input_tokens_seen": 43074432, "step": 31795 }, { "epoch": 1.0178605723065104, "grad_norm": 1.8140188455581665, "learning_rate": 5.009062521837835e-06, "loss": 0.603, "num_input_tokens_seen": 43081488, "step": 31800 }, { "epoch": 1.0178605723065104, "eval_loss": 0.4452488422393799, "eval_runtime": 501.0128, "eval_samples_per_second": 27.716, "eval_steps_per_second": 13.858, "num_input_tokens_seen": 43081488, "step": 31800 }, { "epoch": 1.0180206132769989, "grad_norm": 0.9014360308647156, "learning_rate": 5.003168835049324e-06, "loss": 0.304, "num_input_tokens_seen": 43088224, "step": 31805 }, { "epoch": 1.0181806542474874, "grad_norm": 1.1074211597442627, "learning_rate": 4.997278232023483e-06, "loss": 0.377, "num_input_tokens_seen": 43094912, "step": 31810 }, { "epoch": 1.0183406952179759, "grad_norm": 1.3256993293762207, "learning_rate": 4.9913907136687036e-06, "loss": 0.3386, "num_input_tokens_seen": 43102064, "step": 31815 }, { "epoch": 1.0185007361884642, "grad_norm": 1.3634294271469116, "learning_rate": 4.985506280892918e-06, "loss": 0.4387, "num_input_tokens_seen": 43109120, "step": 31820 }, { "epoch": 1.0186607771589526, "grad_norm": 0.8035393357276917, "learning_rate": 4.979624934603589e-06, "loss": 0.4904, "num_input_tokens_seen": 43115696, "step": 31825 }, { "epoch": 1.0188208181294411, "grad_norm": 0.7266831398010254, "learning_rate": 4.97374667570768e-06, "loss": 0.4346, "num_input_tokens_seen": 43122272, "step": 31830 }, { "epoch": 1.0189808590999296, "grad_norm": 0.6810875535011292, "learning_rate": 4.967871505111704e-06, "loss": 0.5032, "num_input_tokens_seen": 43129168, "step": 31835 }, { "epoch": 1.0191409000704181, "grad_norm": 1.5542895793914795, "learning_rate": 4.961999423721686e-06, "loss": 0.4305, "num_input_tokens_seen": 43135888, "step": 31840 }, { "epoch": 1.0193009410409064, "grad_norm": 0.6554122567176819, "learning_rate": 4.956130432443159e-06, "loss": 0.3692, "num_input_tokens_seen": 43142352, "step": 31845 }, { "epoch": 1.019460982011395, "grad_norm": 1.2876396179199219, "learning_rate": 4.950264532181215e-06, "loss": 0.3514, "num_input_tokens_seen": 43148912, "step": 31850 }, { "epoch": 1.0196210229818834, "grad_norm": 1.3972300291061401, "learning_rate": 4.944401723840433e-06, "loss": 0.4097, "num_input_tokens_seen": 43155392, "step": 31855 }, { "epoch": 1.019781063952372, "grad_norm": 1.5580452680587769, "learning_rate": 4.938542008324942e-06, "loss": 0.6433, "num_input_tokens_seen": 43162288, "step": 31860 }, { "epoch": 1.0199411049228602, "grad_norm": 1.1446906328201294, "learning_rate": 4.9326853865383855e-06, "loss": 0.3945, "num_input_tokens_seen": 43169280, "step": 31865 }, { "epoch": 1.0201011458933487, "grad_norm": 0.8684237599372864, "learning_rate": 4.926831859383918e-06, "loss": 0.3735, "num_input_tokens_seen": 43175872, "step": 31870 }, { "epoch": 1.0202611868638372, "grad_norm": 0.6597183346748352, "learning_rate": 4.92098142776424e-06, "loss": 0.3799, "num_input_tokens_seen": 43182544, "step": 31875 }, { "epoch": 1.0204212278343257, "grad_norm": 1.615783452987671, "learning_rate": 4.91513409258155e-06, "loss": 0.6886, "num_input_tokens_seen": 43189200, "step": 31880 }, { "epoch": 1.020581268804814, "grad_norm": 3.7855992317199707, "learning_rate": 4.909289854737581e-06, "loss": 0.3925, "num_input_tokens_seen": 43196064, "step": 31885 }, { "epoch": 1.0207413097753024, "grad_norm": 0.7037519216537476, "learning_rate": 4.903448715133602e-06, "loss": 0.4802, "num_input_tokens_seen": 43203024, "step": 31890 }, { "epoch": 1.020901350745791, "grad_norm": 1.2873591184616089, "learning_rate": 4.897610674670372e-06, "loss": 0.5226, "num_input_tokens_seen": 43209600, "step": 31895 }, { "epoch": 1.0210613917162794, "grad_norm": 0.625558078289032, "learning_rate": 4.8917757342482e-06, "loss": 0.3882, "num_input_tokens_seen": 43216416, "step": 31900 }, { "epoch": 1.021221432686768, "grad_norm": 0.7951403260231018, "learning_rate": 4.885943894766909e-06, "loss": 0.3468, "num_input_tokens_seen": 43223936, "step": 31905 }, { "epoch": 1.0213814736572562, "grad_norm": 0.9892215728759766, "learning_rate": 4.880115157125842e-06, "loss": 0.4841, "num_input_tokens_seen": 43230592, "step": 31910 }, { "epoch": 1.0215415146277447, "grad_norm": 1.540888786315918, "learning_rate": 4.874289522223857e-06, "loss": 0.4817, "num_input_tokens_seen": 43237568, "step": 31915 }, { "epoch": 1.0217015555982332, "grad_norm": 2.2415738105773926, "learning_rate": 4.868466990959339e-06, "loss": 0.3672, "num_input_tokens_seen": 43245024, "step": 31920 }, { "epoch": 1.0218615965687217, "grad_norm": 0.7552374601364136, "learning_rate": 4.8626475642301964e-06, "loss": 0.2906, "num_input_tokens_seen": 43251328, "step": 31925 }, { "epoch": 1.02202163753921, "grad_norm": 1.1923805475234985, "learning_rate": 4.856831242933871e-06, "loss": 0.5864, "num_input_tokens_seen": 43258288, "step": 31930 }, { "epoch": 1.0221816785096984, "grad_norm": 1.2733575105667114, "learning_rate": 4.851018027967294e-06, "loss": 0.407, "num_input_tokens_seen": 43265040, "step": 31935 }, { "epoch": 1.022341719480187, "grad_norm": 0.9557252526283264, "learning_rate": 4.845207920226946e-06, "loss": 0.4201, "num_input_tokens_seen": 43271520, "step": 31940 }, { "epoch": 1.0225017604506754, "grad_norm": 2.3483407497406006, "learning_rate": 4.839400920608825e-06, "loss": 0.5208, "num_input_tokens_seen": 43278384, "step": 31945 }, { "epoch": 1.022661801421164, "grad_norm": 0.8227044343948364, "learning_rate": 4.83359703000843e-06, "loss": 0.3125, "num_input_tokens_seen": 43285024, "step": 31950 }, { "epoch": 1.0228218423916522, "grad_norm": 0.6290159821510315, "learning_rate": 4.827796249320804e-06, "loss": 0.3423, "num_input_tokens_seen": 43291184, "step": 31955 }, { "epoch": 1.0229818833621407, "grad_norm": 1.171950101852417, "learning_rate": 4.82199857944049e-06, "loss": 0.3338, "num_input_tokens_seen": 43297616, "step": 31960 }, { "epoch": 1.0231419243326292, "grad_norm": 2.9772069454193115, "learning_rate": 4.8162040212615695e-06, "loss": 0.4917, "num_input_tokens_seen": 43304592, "step": 31965 }, { "epoch": 1.0233019653031177, "grad_norm": 0.6651450395584106, "learning_rate": 4.810412575677639e-06, "loss": 0.3679, "num_input_tokens_seen": 43311664, "step": 31970 }, { "epoch": 1.023462006273606, "grad_norm": 1.6412664651870728, "learning_rate": 4.804624243581801e-06, "loss": 0.2598, "num_input_tokens_seen": 43318320, "step": 31975 }, { "epoch": 1.0236220472440944, "grad_norm": 1.501949429512024, "learning_rate": 4.798839025866703e-06, "loss": 0.4569, "num_input_tokens_seen": 43325552, "step": 31980 }, { "epoch": 1.023782088214583, "grad_norm": 0.4933648705482483, "learning_rate": 4.793056923424491e-06, "loss": 0.2442, "num_input_tokens_seen": 43331920, "step": 31985 }, { "epoch": 1.0239421291850714, "grad_norm": 0.6322157979011536, "learning_rate": 4.78727793714683e-06, "loss": 0.4, "num_input_tokens_seen": 43338400, "step": 31990 }, { "epoch": 1.02410217015556, "grad_norm": 1.0661009550094604, "learning_rate": 4.7815020679249285e-06, "loss": 0.3588, "num_input_tokens_seen": 43345296, "step": 31995 }, { "epoch": 1.0242622111260482, "grad_norm": 0.8326596617698669, "learning_rate": 4.775729316649483e-06, "loss": 0.3702, "num_input_tokens_seen": 43351904, "step": 32000 }, { "epoch": 1.0242622111260482, "eval_loss": 0.4453503489494324, "eval_runtime": 500.8982, "eval_samples_per_second": 27.722, "eval_steps_per_second": 13.861, "num_input_tokens_seen": 43351904, "step": 32000 }, { "epoch": 1.0244222520965367, "grad_norm": 0.7269607782363892, "learning_rate": 4.769959684210728e-06, "loss": 0.4638, "num_input_tokens_seen": 43358480, "step": 32005 }, { "epoch": 1.0245822930670252, "grad_norm": 1.4143383502960205, "learning_rate": 4.764193171498426e-06, "loss": 0.4762, "num_input_tokens_seen": 43365072, "step": 32010 }, { "epoch": 1.0247423340375137, "grad_norm": 1.5402613878250122, "learning_rate": 4.75842977940183e-06, "loss": 0.3175, "num_input_tokens_seen": 43372032, "step": 32015 }, { "epoch": 1.024902375008002, "grad_norm": 0.619432806968689, "learning_rate": 4.752669508809729e-06, "loss": 0.2665, "num_input_tokens_seen": 43378816, "step": 32020 }, { "epoch": 1.0250624159784905, "grad_norm": 1.4663982391357422, "learning_rate": 4.746912360610445e-06, "loss": 0.4143, "num_input_tokens_seen": 43385456, "step": 32025 }, { "epoch": 1.025222456948979, "grad_norm": 1.2308330535888672, "learning_rate": 4.741158335691781e-06, "loss": 0.4558, "num_input_tokens_seen": 43392096, "step": 32030 }, { "epoch": 1.0253824979194675, "grad_norm": 1.6806074380874634, "learning_rate": 4.7354074349410994e-06, "loss": 0.4349, "num_input_tokens_seen": 43398848, "step": 32035 }, { "epoch": 1.0255425388899557, "grad_norm": 0.9349518418312073, "learning_rate": 4.729659659245245e-06, "loss": 0.3545, "num_input_tokens_seen": 43405376, "step": 32040 }, { "epoch": 1.0257025798604442, "grad_norm": 0.9432089328765869, "learning_rate": 4.723915009490601e-06, "loss": 0.4484, "num_input_tokens_seen": 43412176, "step": 32045 }, { "epoch": 1.0258626208309327, "grad_norm": 1.8469637632369995, "learning_rate": 4.718173486563077e-06, "loss": 0.4824, "num_input_tokens_seen": 43418992, "step": 32050 }, { "epoch": 1.0260226618014212, "grad_norm": 2.0519585609436035, "learning_rate": 4.71243509134808e-06, "loss": 0.444, "num_input_tokens_seen": 43425328, "step": 32055 }, { "epoch": 1.0261827027719097, "grad_norm": 0.9538591504096985, "learning_rate": 4.706699824730532e-06, "loss": 0.392, "num_input_tokens_seen": 43432368, "step": 32060 }, { "epoch": 1.026342743742398, "grad_norm": 0.5527302026748657, "learning_rate": 4.700967687594901e-06, "loss": 0.5938, "num_input_tokens_seen": 43439248, "step": 32065 }, { "epoch": 1.0265027847128865, "grad_norm": 0.971819281578064, "learning_rate": 4.69523868082514e-06, "loss": 0.4263, "num_input_tokens_seen": 43445824, "step": 32070 }, { "epoch": 1.026662825683375, "grad_norm": 2.191654920578003, "learning_rate": 4.689512805304747e-06, "loss": 0.4307, "num_input_tokens_seen": 43452448, "step": 32075 }, { "epoch": 1.0268228666538635, "grad_norm": 0.7807109355926514, "learning_rate": 4.683790061916707e-06, "loss": 0.339, "num_input_tokens_seen": 43459600, "step": 32080 }, { "epoch": 1.0269829076243517, "grad_norm": 1.2068679332733154, "learning_rate": 4.678070451543551e-06, "loss": 0.3591, "num_input_tokens_seen": 43466144, "step": 32085 }, { "epoch": 1.0271429485948402, "grad_norm": 0.9853156805038452, "learning_rate": 4.6723539750673204e-06, "loss": 0.3762, "num_input_tokens_seen": 43472688, "step": 32090 }, { "epoch": 1.0273029895653287, "grad_norm": 2.7648465633392334, "learning_rate": 4.666640633369551e-06, "loss": 0.4496, "num_input_tokens_seen": 43479248, "step": 32095 }, { "epoch": 1.0274630305358172, "grad_norm": 0.9868075847625732, "learning_rate": 4.660930427331323e-06, "loss": 0.3091, "num_input_tokens_seen": 43485984, "step": 32100 }, { "epoch": 1.0276230715063057, "grad_norm": 2.975717544555664, "learning_rate": 4.6552233578332244e-06, "loss": 0.3989, "num_input_tokens_seen": 43492848, "step": 32105 }, { "epoch": 1.027783112476794, "grad_norm": 0.9340004920959473, "learning_rate": 4.649519425755347e-06, "loss": 0.5577, "num_input_tokens_seen": 43499312, "step": 32110 }, { "epoch": 1.0279431534472825, "grad_norm": 1.026879072189331, "learning_rate": 4.64381863197732e-06, "loss": 0.4956, "num_input_tokens_seen": 43506320, "step": 32115 }, { "epoch": 1.028103194417771, "grad_norm": 3.0130035877227783, "learning_rate": 4.638120977378269e-06, "loss": 0.5504, "num_input_tokens_seen": 43512864, "step": 32120 }, { "epoch": 1.0282632353882595, "grad_norm": 1.2810207605361938, "learning_rate": 4.632426462836848e-06, "loss": 0.6605, "num_input_tokens_seen": 43520064, "step": 32125 }, { "epoch": 1.0284232763587478, "grad_norm": 1.6067932844161987, "learning_rate": 4.626735089231224e-06, "loss": 0.5103, "num_input_tokens_seen": 43526528, "step": 32130 }, { "epoch": 1.0285833173292362, "grad_norm": 1.1490448713302612, "learning_rate": 4.621046857439068e-06, "loss": 0.4153, "num_input_tokens_seen": 43533216, "step": 32135 }, { "epoch": 1.0287433582997247, "grad_norm": 1.71611487865448, "learning_rate": 4.615361768337587e-06, "loss": 0.5666, "num_input_tokens_seen": 43540976, "step": 32140 }, { "epoch": 1.0289033992702132, "grad_norm": 1.067458987236023, "learning_rate": 4.6096798228034946e-06, "loss": 0.5159, "num_input_tokens_seen": 43547680, "step": 32145 }, { "epoch": 1.0290634402407015, "grad_norm": 1.1113018989562988, "learning_rate": 4.604001021713008e-06, "loss": 0.4469, "num_input_tokens_seen": 43554960, "step": 32150 }, { "epoch": 1.02922348121119, "grad_norm": 1.787441611289978, "learning_rate": 4.598325365941883e-06, "loss": 0.4542, "num_input_tokens_seen": 43561664, "step": 32155 }, { "epoch": 1.0293835221816785, "grad_norm": 0.9360641837120056, "learning_rate": 4.5926528563653645e-06, "loss": 0.3032, "num_input_tokens_seen": 43568320, "step": 32160 }, { "epoch": 1.029543563152167, "grad_norm": 1.2393953800201416, "learning_rate": 4.5869834938582295e-06, "loss": 0.4802, "num_input_tokens_seen": 43575024, "step": 32165 }, { "epoch": 1.0297036041226555, "grad_norm": 0.7838313579559326, "learning_rate": 4.581317279294772e-06, "loss": 0.326, "num_input_tokens_seen": 43582032, "step": 32170 }, { "epoch": 1.0298636450931438, "grad_norm": 0.6424607038497925, "learning_rate": 4.57565421354878e-06, "loss": 0.4613, "num_input_tokens_seen": 43588752, "step": 32175 }, { "epoch": 1.0300236860636323, "grad_norm": 1.0028183460235596, "learning_rate": 4.569994297493579e-06, "loss": 0.3961, "num_input_tokens_seen": 43595680, "step": 32180 }, { "epoch": 1.0301837270341208, "grad_norm": 0.6193357110023499, "learning_rate": 4.564337532002002e-06, "loss": 0.3847, "num_input_tokens_seen": 43602416, "step": 32185 }, { "epoch": 1.0303437680046093, "grad_norm": 1.0647668838500977, "learning_rate": 4.55868391794638e-06, "loss": 0.4349, "num_input_tokens_seen": 43609440, "step": 32190 }, { "epoch": 1.0305038089750975, "grad_norm": 0.8914259672164917, "learning_rate": 4.553033456198588e-06, "loss": 0.5232, "num_input_tokens_seen": 43615920, "step": 32195 }, { "epoch": 1.030663849945586, "grad_norm": 4.439581871032715, "learning_rate": 4.54738614762999e-06, "loss": 0.67, "num_input_tokens_seen": 43622640, "step": 32200 }, { "epoch": 1.030663849945586, "eval_loss": 0.445011168718338, "eval_runtime": 501.6304, "eval_samples_per_second": 27.682, "eval_steps_per_second": 13.841, "num_input_tokens_seen": 43622640, "step": 32200 }, { "epoch": 1.0308238909160745, "grad_norm": 0.6873922944068909, "learning_rate": 4.541741993111465e-06, "loss": 0.3224, "num_input_tokens_seen": 43629408, "step": 32205 }, { "epoch": 1.030983931886563, "grad_norm": 0.7988283038139343, "learning_rate": 4.536100993513423e-06, "loss": 0.4373, "num_input_tokens_seen": 43636064, "step": 32210 }, { "epoch": 1.0311439728570515, "grad_norm": 1.0114686489105225, "learning_rate": 4.530463149705768e-06, "loss": 0.5777, "num_input_tokens_seen": 43642672, "step": 32215 }, { "epoch": 1.0313040138275398, "grad_norm": 1.2610316276550293, "learning_rate": 4.524828462557934e-06, "loss": 0.4203, "num_input_tokens_seen": 43649344, "step": 32220 }, { "epoch": 1.0314640547980283, "grad_norm": 0.6575474739074707, "learning_rate": 4.5191969329388625e-06, "loss": 0.5666, "num_input_tokens_seen": 43655648, "step": 32225 }, { "epoch": 1.0316240957685168, "grad_norm": 0.6484273672103882, "learning_rate": 4.5135685617169965e-06, "loss": 0.4217, "num_input_tokens_seen": 43662720, "step": 32230 }, { "epoch": 1.0317841367390053, "grad_norm": 0.7471108436584473, "learning_rate": 4.507943349760313e-06, "loss": 0.2756, "num_input_tokens_seen": 43669136, "step": 32235 }, { "epoch": 1.0319441777094935, "grad_norm": 1.0032742023468018, "learning_rate": 4.502321297936277e-06, "loss": 0.4304, "num_input_tokens_seen": 43676224, "step": 32240 }, { "epoch": 1.032104218679982, "grad_norm": 0.41298648715019226, "learning_rate": 4.496702407111888e-06, "loss": 0.4108, "num_input_tokens_seen": 43683728, "step": 32245 }, { "epoch": 1.0322642596504705, "grad_norm": 0.8688769936561584, "learning_rate": 4.491086678153653e-06, "loss": 0.3564, "num_input_tokens_seen": 43690416, "step": 32250 }, { "epoch": 1.032424300620959, "grad_norm": 1.1121184825897217, "learning_rate": 4.485474111927579e-06, "loss": 0.4013, "num_input_tokens_seen": 43696880, "step": 32255 }, { "epoch": 1.0325843415914475, "grad_norm": 0.7271167635917664, "learning_rate": 4.479864709299197e-06, "loss": 0.2763, "num_input_tokens_seen": 43703520, "step": 32260 }, { "epoch": 1.0327443825619358, "grad_norm": 1.771864891052246, "learning_rate": 4.474258471133555e-06, "loss": 0.3497, "num_input_tokens_seen": 43710800, "step": 32265 }, { "epoch": 1.0329044235324243, "grad_norm": 1.9539477825164795, "learning_rate": 4.4686553982952014e-06, "loss": 0.5926, "num_input_tokens_seen": 43717296, "step": 32270 }, { "epoch": 1.0330644645029128, "grad_norm": 1.6434307098388672, "learning_rate": 4.463055491648191e-06, "loss": 0.3984, "num_input_tokens_seen": 43724032, "step": 32275 }, { "epoch": 1.0332245054734013, "grad_norm": 1.2653944492340088, "learning_rate": 4.457458752056112e-06, "loss": 0.3347, "num_input_tokens_seen": 43730976, "step": 32280 }, { "epoch": 1.0333845464438896, "grad_norm": 2.285806179046631, "learning_rate": 4.451865180382042e-06, "loss": 0.6487, "num_input_tokens_seen": 43737744, "step": 32285 }, { "epoch": 1.033544587414378, "grad_norm": 1.3868675231933594, "learning_rate": 4.4462747774885936e-06, "loss": 0.4191, "num_input_tokens_seen": 43744528, "step": 32290 }, { "epoch": 1.0337046283848665, "grad_norm": 0.7096356153488159, "learning_rate": 4.440687544237859e-06, "loss": 0.6325, "num_input_tokens_seen": 43751376, "step": 32295 }, { "epoch": 1.033864669355355, "grad_norm": 0.6493476629257202, "learning_rate": 4.435103481491471e-06, "loss": 0.3772, "num_input_tokens_seen": 43758192, "step": 32300 }, { "epoch": 1.0340247103258433, "grad_norm": 1.486671805381775, "learning_rate": 4.429522590110569e-06, "loss": 0.5249, "num_input_tokens_seen": 43765184, "step": 32305 }, { "epoch": 1.0341847512963318, "grad_norm": 1.2244304418563843, "learning_rate": 4.423944870955779e-06, "loss": 0.4128, "num_input_tokens_seen": 43772176, "step": 32310 }, { "epoch": 1.0343447922668203, "grad_norm": 1.5699673891067505, "learning_rate": 4.418370324887272e-06, "loss": 0.6267, "num_input_tokens_seen": 43778864, "step": 32315 }, { "epoch": 1.0345048332373088, "grad_norm": 0.5639461278915405, "learning_rate": 4.412798952764699e-06, "loss": 0.3685, "num_input_tokens_seen": 43785408, "step": 32320 }, { "epoch": 1.0346648742077973, "grad_norm": 1.0257762670516968, "learning_rate": 4.407230755447245e-06, "loss": 0.4401, "num_input_tokens_seen": 43792544, "step": 32325 }, { "epoch": 1.0348249151782856, "grad_norm": 0.9900389909744263, "learning_rate": 4.401665733793598e-06, "loss": 0.305, "num_input_tokens_seen": 43799104, "step": 32330 }, { "epoch": 1.034984956148774, "grad_norm": 1.367155909538269, "learning_rate": 4.3961038886619425e-06, "loss": 0.4767, "num_input_tokens_seen": 43806336, "step": 32335 }, { "epoch": 1.0351449971192626, "grad_norm": 1.2191370725631714, "learning_rate": 4.39054522091e-06, "loss": 0.6634, "num_input_tokens_seen": 43812912, "step": 32340 }, { "epoch": 1.035305038089751, "grad_norm": 1.7201319932937622, "learning_rate": 4.384989731394979e-06, "loss": 0.4339, "num_input_tokens_seen": 43819760, "step": 32345 }, { "epoch": 1.0354650790602393, "grad_norm": 1.651322603225708, "learning_rate": 4.379437420973598e-06, "loss": 0.4042, "num_input_tokens_seen": 43826736, "step": 32350 }, { "epoch": 1.0356251200307278, "grad_norm": 0.6659154295921326, "learning_rate": 4.373888290502107e-06, "loss": 0.3395, "num_input_tokens_seen": 43833600, "step": 32355 }, { "epoch": 1.0357851610012163, "grad_norm": 0.7232133746147156, "learning_rate": 4.36834234083624e-06, "loss": 0.5918, "num_input_tokens_seen": 43839920, "step": 32360 }, { "epoch": 1.0359452019717048, "grad_norm": 1.493891716003418, "learning_rate": 4.362799572831258e-06, "loss": 0.6487, "num_input_tokens_seen": 43846112, "step": 32365 }, { "epoch": 1.0361052429421933, "grad_norm": 1.1899746656417847, "learning_rate": 4.35725998734193e-06, "loss": 0.4905, "num_input_tokens_seen": 43852896, "step": 32370 }, { "epoch": 1.0362652839126816, "grad_norm": 0.6722431778907776, "learning_rate": 4.3517235852225195e-06, "loss": 0.4051, "num_input_tokens_seen": 43859648, "step": 32375 }, { "epoch": 1.03642532488317, "grad_norm": 0.6855292916297913, "learning_rate": 4.346190367326822e-06, "loss": 0.3747, "num_input_tokens_seen": 43866576, "step": 32380 }, { "epoch": 1.0365853658536586, "grad_norm": 1.6139882802963257, "learning_rate": 4.340660334508115e-06, "loss": 0.5761, "num_input_tokens_seen": 43873232, "step": 32385 }, { "epoch": 1.036745406824147, "grad_norm": 0.6306357383728027, "learning_rate": 4.335133487619206e-06, "loss": 0.5078, "num_input_tokens_seen": 43879872, "step": 32390 }, { "epoch": 1.0369054477946353, "grad_norm": 0.6948071122169495, "learning_rate": 4.329609827512409e-06, "loss": 0.3224, "num_input_tokens_seen": 43887472, "step": 32395 }, { "epoch": 1.0370654887651238, "grad_norm": 0.6053580045700073, "learning_rate": 4.324089355039531e-06, "loss": 0.4796, "num_input_tokens_seen": 43893856, "step": 32400 }, { "epoch": 1.0370654887651238, "eval_loss": 0.4449980556964874, "eval_runtime": 501.5633, "eval_samples_per_second": 27.685, "eval_steps_per_second": 13.843, "num_input_tokens_seen": 43893856, "step": 32400 }, { "epoch": 1.0372255297356123, "grad_norm": 3.581075668334961, "learning_rate": 4.3185720710519075e-06, "loss": 0.367, "num_input_tokens_seen": 43900704, "step": 32405 }, { "epoch": 1.0373855707061008, "grad_norm": 1.3087162971496582, "learning_rate": 4.3130579764003724e-06, "loss": 0.5391, "num_input_tokens_seen": 43907376, "step": 32410 }, { "epoch": 1.037545611676589, "grad_norm": 1.1812323331832886, "learning_rate": 4.307547071935267e-06, "loss": 0.3744, "num_input_tokens_seen": 43913760, "step": 32415 }, { "epoch": 1.0377056526470776, "grad_norm": 0.8852864503860474, "learning_rate": 4.302039358506435e-06, "loss": 0.4179, "num_input_tokens_seen": 43920496, "step": 32420 }, { "epoch": 1.037865693617566, "grad_norm": 0.6563858389854431, "learning_rate": 4.296534836963245e-06, "loss": 0.2215, "num_input_tokens_seen": 43927360, "step": 32425 }, { "epoch": 1.0380257345880546, "grad_norm": 0.42414748668670654, "learning_rate": 4.291033508154555e-06, "loss": 0.3924, "num_input_tokens_seen": 43934496, "step": 32430 }, { "epoch": 1.038185775558543, "grad_norm": 1.5701488256454468, "learning_rate": 4.285535372928748e-06, "loss": 0.3477, "num_input_tokens_seen": 43941280, "step": 32435 }, { "epoch": 1.0383458165290314, "grad_norm": 1.3131167888641357, "learning_rate": 4.280040432133695e-06, "loss": 0.3527, "num_input_tokens_seen": 43947776, "step": 32440 }, { "epoch": 1.0385058574995198, "grad_norm": 0.8473873734474182, "learning_rate": 4.274548686616789e-06, "loss": 0.3449, "num_input_tokens_seen": 43954800, "step": 32445 }, { "epoch": 1.0386658984700083, "grad_norm": 0.8690330982208252, "learning_rate": 4.2690601372249364e-06, "loss": 0.5228, "num_input_tokens_seen": 43962128, "step": 32450 }, { "epoch": 1.0388259394404968, "grad_norm": 1.063979148864746, "learning_rate": 4.263574784804525e-06, "loss": 0.4093, "num_input_tokens_seen": 43968864, "step": 32455 }, { "epoch": 1.038985980410985, "grad_norm": 1.0473884344100952, "learning_rate": 4.258092630201479e-06, "loss": 0.3486, "num_input_tokens_seen": 43975456, "step": 32460 }, { "epoch": 1.0391460213814736, "grad_norm": 2.382899045944214, "learning_rate": 4.252613674261202e-06, "loss": 0.6023, "num_input_tokens_seen": 43982592, "step": 32465 }, { "epoch": 1.039306062351962, "grad_norm": 0.773908793926239, "learning_rate": 4.2471379178286224e-06, "loss": 0.4184, "num_input_tokens_seen": 43989440, "step": 32470 }, { "epoch": 1.0394661033224506, "grad_norm": 0.8778852820396423, "learning_rate": 4.241665361748181e-06, "loss": 0.4231, "num_input_tokens_seen": 43996432, "step": 32475 }, { "epoch": 1.039626144292939, "grad_norm": 0.9651718139648438, "learning_rate": 4.2361960068637994e-06, "loss": 0.5074, "num_input_tokens_seen": 44003264, "step": 32480 }, { "epoch": 1.0397861852634274, "grad_norm": 0.7764536738395691, "learning_rate": 4.230729854018933e-06, "loss": 0.2942, "num_input_tokens_seen": 44010080, "step": 32485 }, { "epoch": 1.0399462262339159, "grad_norm": 0.5729247331619263, "learning_rate": 4.225266904056521e-06, "loss": 0.3426, "num_input_tokens_seen": 44016896, "step": 32490 }, { "epoch": 1.0401062672044044, "grad_norm": 0.8599652647972107, "learning_rate": 4.21980715781903e-06, "loss": 0.4717, "num_input_tokens_seen": 44023488, "step": 32495 }, { "epoch": 1.0402663081748929, "grad_norm": 2.2672665119171143, "learning_rate": 4.214350616148416e-06, "loss": 0.3469, "num_input_tokens_seen": 44029824, "step": 32500 }, { "epoch": 1.0404263491453811, "grad_norm": 1.341922640800476, "learning_rate": 4.20889727988614e-06, "loss": 0.4662, "num_input_tokens_seen": 44036800, "step": 32505 }, { "epoch": 1.0405863901158696, "grad_norm": 2.7308454513549805, "learning_rate": 4.20344714987318e-06, "loss": 0.5751, "num_input_tokens_seen": 44043360, "step": 32510 }, { "epoch": 1.0407464310863581, "grad_norm": 0.7205327749252319, "learning_rate": 4.198000226950022e-06, "loss": 0.3341, "num_input_tokens_seen": 44049856, "step": 32515 }, { "epoch": 1.0409064720568466, "grad_norm": 1.5440162420272827, "learning_rate": 4.192556511956635e-06, "loss": 0.3925, "num_input_tokens_seen": 44056432, "step": 32520 }, { "epoch": 1.041066513027335, "grad_norm": 1.39932119846344, "learning_rate": 4.18711600573252e-06, "loss": 0.6353, "num_input_tokens_seen": 44063296, "step": 32525 }, { "epoch": 1.0412265539978234, "grad_norm": 0.7267877459526062, "learning_rate": 4.181678709116671e-06, "loss": 0.3047, "num_input_tokens_seen": 44070688, "step": 32530 }, { "epoch": 1.0413865949683119, "grad_norm": 1.0105676651000977, "learning_rate": 4.1762446229475785e-06, "loss": 0.4027, "num_input_tokens_seen": 44077472, "step": 32535 }, { "epoch": 1.0415466359388004, "grad_norm": 1.1077840328216553, "learning_rate": 4.17081374806326e-06, "loss": 0.4727, "num_input_tokens_seen": 44083904, "step": 32540 }, { "epoch": 1.0417066769092889, "grad_norm": 0.9234069585800171, "learning_rate": 4.165386085301212e-06, "loss": 0.3015, "num_input_tokens_seen": 44090560, "step": 32545 }, { "epoch": 1.0418667178797771, "grad_norm": 1.571810007095337, "learning_rate": 4.1599616354984525e-06, "loss": 0.419, "num_input_tokens_seen": 44097056, "step": 32550 }, { "epoch": 1.0420267588502656, "grad_norm": 1.444084644317627, "learning_rate": 4.154540399491508e-06, "loss": 0.482, "num_input_tokens_seen": 44103376, "step": 32555 }, { "epoch": 1.0421867998207541, "grad_norm": 1.2806100845336914, "learning_rate": 4.149122378116394e-06, "loss": 0.3998, "num_input_tokens_seen": 44110336, "step": 32560 }, { "epoch": 1.0423468407912426, "grad_norm": 1.6786314249038696, "learning_rate": 4.14370757220863e-06, "loss": 0.654, "num_input_tokens_seen": 44117056, "step": 32565 }, { "epoch": 1.042506881761731, "grad_norm": 0.7664747834205627, "learning_rate": 4.138295982603263e-06, "loss": 0.3442, "num_input_tokens_seen": 44123920, "step": 32570 }, { "epoch": 1.0426669227322194, "grad_norm": 1.775693416595459, "learning_rate": 4.132887610134814e-06, "loss": 0.6638, "num_input_tokens_seen": 44130528, "step": 32575 }, { "epoch": 1.0428269637027079, "grad_norm": 0.6315363049507141, "learning_rate": 4.127482455637335e-06, "loss": 0.6524, "num_input_tokens_seen": 44137360, "step": 32580 }, { "epoch": 1.0429870046731964, "grad_norm": 0.9203594923019409, "learning_rate": 4.1220805199443545e-06, "loss": 0.5457, "num_input_tokens_seen": 44143792, "step": 32585 }, { "epoch": 1.0431470456436849, "grad_norm": 1.4310870170593262, "learning_rate": 4.116681803888925e-06, "loss": 0.4465, "num_input_tokens_seen": 44150800, "step": 32590 }, { "epoch": 1.0433070866141732, "grad_norm": 1.277604341506958, "learning_rate": 4.111286308303605e-06, "loss": 0.4625, "num_input_tokens_seen": 44157728, "step": 32595 }, { "epoch": 1.0434671275846616, "grad_norm": 0.4481387436389923, "learning_rate": 4.105894034020433e-06, "loss": 0.2008, "num_input_tokens_seen": 44164592, "step": 32600 }, { "epoch": 1.0434671275846616, "eval_loss": 0.4448828101158142, "eval_runtime": 501.5489, "eval_samples_per_second": 27.686, "eval_steps_per_second": 13.843, "num_input_tokens_seen": 44164592, "step": 32600 }, { "epoch": 1.0436271685551501, "grad_norm": 2.1572265625, "learning_rate": 4.100504981870975e-06, "loss": 0.5282, "num_input_tokens_seen": 44171600, "step": 32605 }, { "epoch": 1.0437872095256386, "grad_norm": 1.2954511642456055, "learning_rate": 4.0951191526862915e-06, "loss": 0.4162, "num_input_tokens_seen": 44178016, "step": 32610 }, { "epoch": 1.043947250496127, "grad_norm": 3.4958465099334717, "learning_rate": 4.089736547296938e-06, "loss": 0.495, "num_input_tokens_seen": 44184752, "step": 32615 }, { "epoch": 1.0441072914666154, "grad_norm": 3.1809401512145996, "learning_rate": 4.08435716653299e-06, "loss": 0.5338, "num_input_tokens_seen": 44191104, "step": 32620 }, { "epoch": 1.044267332437104, "grad_norm": 1.3581409454345703, "learning_rate": 4.0789810112240005e-06, "loss": 0.4495, "num_input_tokens_seen": 44197520, "step": 32625 }, { "epoch": 1.0444273734075924, "grad_norm": 0.9781064987182617, "learning_rate": 4.073608082199057e-06, "loss": 0.3918, "num_input_tokens_seen": 44204080, "step": 32630 }, { "epoch": 1.044587414378081, "grad_norm": 0.8079739212989807, "learning_rate": 4.068238380286718e-06, "loss": 0.5751, "num_input_tokens_seen": 44210560, "step": 32635 }, { "epoch": 1.0447474553485692, "grad_norm": 1.8671835660934448, "learning_rate": 4.062871906315072e-06, "loss": 0.4768, "num_input_tokens_seen": 44217024, "step": 32640 }, { "epoch": 1.0449074963190577, "grad_norm": 0.8129122257232666, "learning_rate": 4.057508661111686e-06, "loss": 0.2535, "num_input_tokens_seen": 44223936, "step": 32645 }, { "epoch": 1.0450675372895462, "grad_norm": 0.46589338779449463, "learning_rate": 4.052148645503648e-06, "loss": 0.4411, "num_input_tokens_seen": 44231264, "step": 32650 }, { "epoch": 1.0452275782600347, "grad_norm": 1.10280179977417, "learning_rate": 4.046791860317531e-06, "loss": 0.5213, "num_input_tokens_seen": 44238464, "step": 32655 }, { "epoch": 1.045387619230523, "grad_norm": 4.040499687194824, "learning_rate": 4.041438306379431e-06, "loss": 0.3783, "num_input_tokens_seen": 44245488, "step": 32660 }, { "epoch": 1.0455476602010114, "grad_norm": 0.8582992553710938, "learning_rate": 4.036087984514916e-06, "loss": 0.3432, "num_input_tokens_seen": 44252608, "step": 32665 }, { "epoch": 1.0457077011715, "grad_norm": 0.6505030393600464, "learning_rate": 4.030740895549084e-06, "loss": 0.416, "num_input_tokens_seen": 44259344, "step": 32670 }, { "epoch": 1.0458677421419884, "grad_norm": 2.7729616165161133, "learning_rate": 4.025397040306531e-06, "loss": 0.3266, "num_input_tokens_seen": 44266000, "step": 32675 }, { "epoch": 1.0460277831124767, "grad_norm": 0.6535462141036987, "learning_rate": 4.0200564196113285e-06, "loss": 0.3913, "num_input_tokens_seen": 44272912, "step": 32680 }, { "epoch": 1.0461878240829652, "grad_norm": 1.1820893287658691, "learning_rate": 4.014719034287079e-06, "loss": 0.4776, "num_input_tokens_seen": 44280144, "step": 32685 }, { "epoch": 1.0463478650534537, "grad_norm": 0.8751479983329773, "learning_rate": 4.0093848851568775e-06, "loss": 0.3758, "num_input_tokens_seen": 44287072, "step": 32690 }, { "epoch": 1.0465079060239422, "grad_norm": 1.1110424995422363, "learning_rate": 4.004053973043304e-06, "loss": 0.5023, "num_input_tokens_seen": 44293696, "step": 32695 }, { "epoch": 1.0466679469944307, "grad_norm": 1.0001972913742065, "learning_rate": 3.998726298768465e-06, "loss": 0.3578, "num_input_tokens_seen": 44300464, "step": 32700 }, { "epoch": 1.046827987964919, "grad_norm": 1.005786418914795, "learning_rate": 3.99340186315395e-06, "loss": 0.4847, "num_input_tokens_seen": 44306944, "step": 32705 }, { "epoch": 1.0469880289354074, "grad_norm": 0.6503677368164062, "learning_rate": 3.988080667020849e-06, "loss": 0.4058, "num_input_tokens_seen": 44313984, "step": 32710 }, { "epoch": 1.047148069905896, "grad_norm": 0.9233301877975464, "learning_rate": 3.982762711189766e-06, "loss": 0.4892, "num_input_tokens_seen": 44320912, "step": 32715 }, { "epoch": 1.0473081108763844, "grad_norm": 1.526137351989746, "learning_rate": 3.977447996480785e-06, "loss": 0.4628, "num_input_tokens_seen": 44327968, "step": 32720 }, { "epoch": 1.0474681518468727, "grad_norm": 0.8516893982887268, "learning_rate": 3.97213652371351e-06, "loss": 0.4779, "num_input_tokens_seen": 44335248, "step": 32725 }, { "epoch": 1.0476281928173612, "grad_norm": 1.3801323175430298, "learning_rate": 3.966828293707042e-06, "loss": 0.3429, "num_input_tokens_seen": 44342304, "step": 32730 }, { "epoch": 1.0477882337878497, "grad_norm": 1.380100965499878, "learning_rate": 3.961523307279963e-06, "loss": 0.4263, "num_input_tokens_seen": 44349088, "step": 32735 }, { "epoch": 1.0479482747583382, "grad_norm": 0.76335209608078, "learning_rate": 3.956221565250382e-06, "loss": 0.4734, "num_input_tokens_seen": 44356272, "step": 32740 }, { "epoch": 1.0481083157288267, "grad_norm": 0.6781879663467407, "learning_rate": 3.950923068435883e-06, "loss": 0.3679, "num_input_tokens_seen": 44363632, "step": 32745 }, { "epoch": 1.048268356699315, "grad_norm": 1.2759313583374023, "learning_rate": 3.945627817653566e-06, "loss": 0.533, "num_input_tokens_seen": 44370288, "step": 32750 }, { "epoch": 1.0484283976698034, "grad_norm": 0.9691600203514099, "learning_rate": 3.9403358137200335e-06, "loss": 0.2176, "num_input_tokens_seen": 44377520, "step": 32755 }, { "epoch": 1.048588438640292, "grad_norm": 2.4607980251312256, "learning_rate": 3.9350470574513605e-06, "loss": 0.5679, "num_input_tokens_seen": 44384544, "step": 32760 }, { "epoch": 1.0487484796107804, "grad_norm": 1.353050708770752, "learning_rate": 3.9297615496631525e-06, "loss": 0.3892, "num_input_tokens_seen": 44391536, "step": 32765 }, { "epoch": 1.0489085205812687, "grad_norm": 1.5255568027496338, "learning_rate": 3.924479291170505e-06, "loss": 0.3972, "num_input_tokens_seen": 44398368, "step": 32770 }, { "epoch": 1.0490685615517572, "grad_norm": 1.4711014032363892, "learning_rate": 3.919200282788002e-06, "loss": 0.3305, "num_input_tokens_seen": 44405536, "step": 32775 }, { "epoch": 1.0492286025222457, "grad_norm": 2.1548638343811035, "learning_rate": 3.913924525329726e-06, "loss": 0.5089, "num_input_tokens_seen": 44412176, "step": 32780 }, { "epoch": 1.0493886434927342, "grad_norm": 1.9540053606033325, "learning_rate": 3.908652019609279e-06, "loss": 0.6532, "num_input_tokens_seen": 44418880, "step": 32785 }, { "epoch": 1.0495486844632227, "grad_norm": 1.2323369979858398, "learning_rate": 3.9033827664397364e-06, "loss": 0.4579, "num_input_tokens_seen": 44425648, "step": 32790 }, { "epoch": 1.049708725433711, "grad_norm": 1.5515201091766357, "learning_rate": 3.898116766633694e-06, "loss": 0.373, "num_input_tokens_seen": 44431952, "step": 32795 }, { "epoch": 1.0498687664041995, "grad_norm": 1.4485222101211548, "learning_rate": 3.8928540210032225e-06, "loss": 0.4444, "num_input_tokens_seen": 44438640, "step": 32800 }, { "epoch": 1.0498687664041995, "eval_loss": 0.44471845030784607, "eval_runtime": 500.9796, "eval_samples_per_second": 27.718, "eval_steps_per_second": 13.859, "num_input_tokens_seen": 44438640, "step": 32800 }, { "epoch": 1.050028807374688, "grad_norm": 1.0671170949935913, "learning_rate": 3.887594530359909e-06, "loss": 0.3199, "num_input_tokens_seen": 44445600, "step": 32805 }, { "epoch": 1.0501888483451765, "grad_norm": 1.0413471460342407, "learning_rate": 3.88233829551484e-06, "loss": 0.5105, "num_input_tokens_seen": 44452560, "step": 32810 }, { "epoch": 1.0503488893156647, "grad_norm": 0.9948884844779968, "learning_rate": 3.877085317278581e-06, "loss": 0.4064, "num_input_tokens_seen": 44459424, "step": 32815 }, { "epoch": 1.0505089302861532, "grad_norm": 0.881770670413971, "learning_rate": 3.87183559646122e-06, "loss": 0.4339, "num_input_tokens_seen": 44466192, "step": 32820 }, { "epoch": 1.0506689712566417, "grad_norm": 1.1565032005310059, "learning_rate": 3.866589133872317e-06, "loss": 0.4713, "num_input_tokens_seen": 44473360, "step": 32825 }, { "epoch": 1.0508290122271302, "grad_norm": 1.6381442546844482, "learning_rate": 3.861345930320948e-06, "loss": 0.4253, "num_input_tokens_seen": 44479872, "step": 32830 }, { "epoch": 1.0509890531976185, "grad_norm": 1.7960660457611084, "learning_rate": 3.856105986615688e-06, "loss": 0.5715, "num_input_tokens_seen": 44486912, "step": 32835 }, { "epoch": 1.051149094168107, "grad_norm": 1.1886063814163208, "learning_rate": 3.850869303564589e-06, "loss": 0.3507, "num_input_tokens_seen": 44493984, "step": 32840 }, { "epoch": 1.0513091351385955, "grad_norm": 1.6625556945800781, "learning_rate": 3.845635881975226e-06, "loss": 0.7029, "num_input_tokens_seen": 44501280, "step": 32845 }, { "epoch": 1.051469176109084, "grad_norm": 0.7180764079093933, "learning_rate": 3.840405722654647e-06, "loss": 0.535, "num_input_tokens_seen": 44508544, "step": 32850 }, { "epoch": 1.0516292170795725, "grad_norm": 1.3185179233551025, "learning_rate": 3.835178826409419e-06, "loss": 0.4153, "num_input_tokens_seen": 44515248, "step": 32855 }, { "epoch": 1.0517892580500607, "grad_norm": 1.7634938955307007, "learning_rate": 3.8299551940455895e-06, "loss": 0.4553, "num_input_tokens_seen": 44522016, "step": 32860 }, { "epoch": 1.0519492990205492, "grad_norm": 1.1067172288894653, "learning_rate": 3.824734826368703e-06, "loss": 0.3181, "num_input_tokens_seen": 44528928, "step": 32865 }, { "epoch": 1.0521093399910377, "grad_norm": 0.9451293349266052, "learning_rate": 3.819517724183813e-06, "loss": 0.3129, "num_input_tokens_seen": 44535936, "step": 32870 }, { "epoch": 1.0522693809615262, "grad_norm": 1.839853048324585, "learning_rate": 3.8143038882954648e-06, "loss": 0.4726, "num_input_tokens_seen": 44542544, "step": 32875 }, { "epoch": 1.0524294219320145, "grad_norm": 0.8956995606422424, "learning_rate": 3.8090933195076867e-06, "loss": 0.4901, "num_input_tokens_seen": 44549472, "step": 32880 }, { "epoch": 1.052589462902503, "grad_norm": 0.41987112164497375, "learning_rate": 3.8038860186240198e-06, "loss": 0.3607, "num_input_tokens_seen": 44556768, "step": 32885 }, { "epoch": 1.0527495038729915, "grad_norm": 1.3187240362167358, "learning_rate": 3.7986819864475026e-06, "loss": 0.4837, "num_input_tokens_seen": 44562928, "step": 32890 }, { "epoch": 1.05290954484348, "grad_norm": 0.6023634076118469, "learning_rate": 3.793481223780651e-06, "loss": 0.4127, "num_input_tokens_seen": 44569440, "step": 32895 }, { "epoch": 1.0530695858139685, "grad_norm": 0.9861136674880981, "learning_rate": 3.788283731425496e-06, "loss": 0.3735, "num_input_tokens_seen": 44576304, "step": 32900 }, { "epoch": 1.0532296267844568, "grad_norm": 0.7822355628013611, "learning_rate": 3.7830895101835488e-06, "loss": 0.3214, "num_input_tokens_seen": 44583296, "step": 32905 }, { "epoch": 1.0533896677549452, "grad_norm": 1.4548354148864746, "learning_rate": 3.7778985608558274e-06, "loss": 0.4165, "num_input_tokens_seen": 44590128, "step": 32910 }, { "epoch": 1.0535497087254337, "grad_norm": 1.6952074766159058, "learning_rate": 3.7727108842428443e-06, "loss": 0.5095, "num_input_tokens_seen": 44596624, "step": 32915 }, { "epoch": 1.0537097496959222, "grad_norm": 2.388482093811035, "learning_rate": 3.7675264811446065e-06, "loss": 0.5903, "num_input_tokens_seen": 44603456, "step": 32920 }, { "epoch": 1.0538697906664105, "grad_norm": 0.6622839570045471, "learning_rate": 3.7623453523605994e-06, "loss": 0.4344, "num_input_tokens_seen": 44610064, "step": 32925 }, { "epoch": 1.054029831636899, "grad_norm": 0.45028820633888245, "learning_rate": 3.757167498689834e-06, "loss": 0.3168, "num_input_tokens_seen": 44616832, "step": 32930 }, { "epoch": 1.0541898726073875, "grad_norm": 0.7560675144195557, "learning_rate": 3.7519929209307914e-06, "loss": 0.3702, "num_input_tokens_seen": 44623360, "step": 32935 }, { "epoch": 1.054349913577876, "grad_norm": 0.5275657773017883, "learning_rate": 3.746821619881463e-06, "loss": 0.3787, "num_input_tokens_seen": 44630480, "step": 32940 }, { "epoch": 1.0545099545483643, "grad_norm": 1.707765817642212, "learning_rate": 3.74165359633932e-06, "loss": 0.4584, "num_input_tokens_seen": 44637184, "step": 32945 }, { "epoch": 1.0546699955188528, "grad_norm": 0.6088129878044128, "learning_rate": 3.736488851101341e-06, "loss": 0.6279, "num_input_tokens_seen": 44644096, "step": 32950 }, { "epoch": 1.0548300364893413, "grad_norm": 1.7298036813735962, "learning_rate": 3.7313273849640035e-06, "loss": 0.621, "num_input_tokens_seen": 44651120, "step": 32955 }, { "epoch": 1.0549900774598298, "grad_norm": 1.9383600950241089, "learning_rate": 3.7261691987232533e-06, "loss": 0.5169, "num_input_tokens_seen": 44658208, "step": 32960 }, { "epoch": 1.0551501184303183, "grad_norm": 1.0484882593154907, "learning_rate": 3.7210142931745575e-06, "loss": 0.4613, "num_input_tokens_seen": 44664448, "step": 32965 }, { "epoch": 1.0553101594008065, "grad_norm": 0.9316795468330383, "learning_rate": 3.7158626691128712e-06, "loss": 0.4388, "num_input_tokens_seen": 44671584, "step": 32970 }, { "epoch": 1.055470200371295, "grad_norm": 0.8065432906150818, "learning_rate": 3.710714327332629e-06, "loss": 0.4214, "num_input_tokens_seen": 44678080, "step": 32975 }, { "epoch": 1.0556302413417835, "grad_norm": 0.8156113624572754, "learning_rate": 3.7055692686277815e-06, "loss": 0.3593, "num_input_tokens_seen": 44684736, "step": 32980 }, { "epoch": 1.055790282312272, "grad_norm": 0.6411776542663574, "learning_rate": 3.70042749379175e-06, "loss": 0.5295, "num_input_tokens_seen": 44692656, "step": 32985 }, { "epoch": 1.0559503232827603, "grad_norm": 0.9408909678459167, "learning_rate": 3.6952890036174693e-06, "loss": 0.4505, "num_input_tokens_seen": 44699328, "step": 32990 }, { "epoch": 1.0561103642532488, "grad_norm": 1.130380630493164, "learning_rate": 3.690153798897353e-06, "loss": 0.4051, "num_input_tokens_seen": 44706000, "step": 32995 }, { "epoch": 1.0562704052237373, "grad_norm": 0.865004301071167, "learning_rate": 3.6850218804233225e-06, "loss": 0.467, "num_input_tokens_seen": 44712640, "step": 33000 }, { "epoch": 1.0562704052237373, "eval_loss": 0.4448479115962982, "eval_runtime": 501.7071, "eval_samples_per_second": 27.678, "eval_steps_per_second": 13.839, "num_input_tokens_seen": 44712640, "step": 33000 }, { "epoch": 1.0564304461942258, "grad_norm": 1.1134823560714722, "learning_rate": 3.679893248986779e-06, "loss": 0.4677, "num_input_tokens_seen": 44719136, "step": 33005 }, { "epoch": 1.0565904871647143, "grad_norm": 1.364106297492981, "learning_rate": 3.6747679053786147e-06, "loss": 0.3084, "num_input_tokens_seen": 44725824, "step": 33010 }, { "epoch": 1.0567505281352025, "grad_norm": 1.5568822622299194, "learning_rate": 3.669645850389228e-06, "loss": 0.4602, "num_input_tokens_seen": 44732816, "step": 33015 }, { "epoch": 1.056910569105691, "grad_norm": 1.2088191509246826, "learning_rate": 3.664527084808514e-06, "loss": 0.5392, "num_input_tokens_seen": 44739520, "step": 33020 }, { "epoch": 1.0570706100761795, "grad_norm": 1.2872676849365234, "learning_rate": 3.6594116094258337e-06, "loss": 0.3794, "num_input_tokens_seen": 44746768, "step": 33025 }, { "epoch": 1.057230651046668, "grad_norm": 1.611397624015808, "learning_rate": 3.6542994250300665e-06, "loss": 0.5922, "num_input_tokens_seen": 44753184, "step": 33030 }, { "epoch": 1.0573906920171563, "grad_norm": 1.5864357948303223, "learning_rate": 3.6491905324095825e-06, "loss": 0.4487, "num_input_tokens_seen": 44759776, "step": 33035 }, { "epoch": 1.0575507329876448, "grad_norm": 1.6682631969451904, "learning_rate": 3.644084932352221e-06, "loss": 0.4496, "num_input_tokens_seen": 44766400, "step": 33040 }, { "epoch": 1.0577107739581333, "grad_norm": 1.4696780443191528, "learning_rate": 3.6389826256453457e-06, "loss": 0.4428, "num_input_tokens_seen": 44772896, "step": 33045 }, { "epoch": 1.0578708149286218, "grad_norm": 0.49091777205467224, "learning_rate": 3.633883613075781e-06, "loss": 0.3785, "num_input_tokens_seen": 44779456, "step": 33050 }, { "epoch": 1.0580308558991103, "grad_norm": 0.7616659998893738, "learning_rate": 3.6287878954298693e-06, "loss": 0.5668, "num_input_tokens_seen": 44786064, "step": 33055 }, { "epoch": 1.0581908968695986, "grad_norm": 0.8297250866889954, "learning_rate": 3.6236954734934354e-06, "loss": 0.3488, "num_input_tokens_seen": 44792976, "step": 33060 }, { "epoch": 1.058350937840087, "grad_norm": 1.281243085861206, "learning_rate": 3.618606348051784e-06, "loss": 0.5411, "num_input_tokens_seen": 44799648, "step": 33065 }, { "epoch": 1.0585109788105755, "grad_norm": 1.5614736080169678, "learning_rate": 3.6135205198897376e-06, "loss": 0.4598, "num_input_tokens_seen": 44806080, "step": 33070 }, { "epoch": 1.058671019781064, "grad_norm": 0.7378279566764832, "learning_rate": 3.6084379897915854e-06, "loss": 0.524, "num_input_tokens_seen": 44812592, "step": 33075 }, { "epoch": 1.0588310607515523, "grad_norm": 0.48224276304244995, "learning_rate": 3.6033587585411115e-06, "loss": 0.4588, "num_input_tokens_seen": 44819232, "step": 33080 }, { "epoch": 1.0589911017220408, "grad_norm": 1.0379754304885864, "learning_rate": 3.5982828269216117e-06, "loss": 0.3563, "num_input_tokens_seen": 44825696, "step": 33085 }, { "epoch": 1.0591511426925293, "grad_norm": 1.9733483791351318, "learning_rate": 3.593210195715843e-06, "loss": 0.3934, "num_input_tokens_seen": 44832400, "step": 33090 }, { "epoch": 1.0593111836630178, "grad_norm": 0.33704787492752075, "learning_rate": 3.5881408657060773e-06, "loss": 0.3056, "num_input_tokens_seen": 44839152, "step": 33095 }, { "epoch": 1.059471224633506, "grad_norm": 1.128968596458435, "learning_rate": 3.583074837674075e-06, "loss": 0.4448, "num_input_tokens_seen": 44845952, "step": 33100 }, { "epoch": 1.0596312656039946, "grad_norm": 0.9223785400390625, "learning_rate": 3.578012112401069e-06, "loss": 0.43, "num_input_tokens_seen": 44852384, "step": 33105 }, { "epoch": 1.059791306574483, "grad_norm": 1.4886101484298706, "learning_rate": 3.5729526906677996e-06, "loss": 0.5083, "num_input_tokens_seen": 44859392, "step": 33110 }, { "epoch": 1.0599513475449716, "grad_norm": 0.9559779167175293, "learning_rate": 3.5678965732545007e-06, "loss": 0.4536, "num_input_tokens_seen": 44866752, "step": 33115 }, { "epoch": 1.06011138851546, "grad_norm": 1.589030385017395, "learning_rate": 3.562843760940876e-06, "loss": 0.3268, "num_input_tokens_seen": 44873456, "step": 33120 }, { "epoch": 1.0602714294859483, "grad_norm": 1.6658720970153809, "learning_rate": 3.5577942545061473e-06, "loss": 0.5389, "num_input_tokens_seen": 44880112, "step": 33125 }, { "epoch": 1.0604314704564368, "grad_norm": 0.7882829904556274, "learning_rate": 3.5527480547289967e-06, "loss": 0.3515, "num_input_tokens_seen": 44886512, "step": 33130 }, { "epoch": 1.0605915114269253, "grad_norm": 1.8917040824890137, "learning_rate": 3.547705162387624e-06, "loss": 0.419, "num_input_tokens_seen": 44893136, "step": 33135 }, { "epoch": 1.0607515523974138, "grad_norm": 1.2291369438171387, "learning_rate": 3.542665578259699e-06, "loss": 0.503, "num_input_tokens_seen": 44900944, "step": 33140 }, { "epoch": 1.060911593367902, "grad_norm": 1.2912412881851196, "learning_rate": 3.5376293031223945e-06, "loss": 0.6098, "num_input_tokens_seen": 44907552, "step": 33145 }, { "epoch": 1.0610716343383906, "grad_norm": 1.4064124822616577, "learning_rate": 3.5325963377523614e-06, "loss": 0.5835, "num_input_tokens_seen": 44913888, "step": 33150 }, { "epoch": 1.061231675308879, "grad_norm": 1.6598020792007446, "learning_rate": 3.5275666829257536e-06, "loss": 0.5039, "num_input_tokens_seen": 44920848, "step": 33155 }, { "epoch": 1.0613917162793676, "grad_norm": 1.1705710887908936, "learning_rate": 3.5225403394181955e-06, "loss": 0.4614, "num_input_tokens_seen": 44927664, "step": 33160 }, { "epoch": 1.061551757249856, "grad_norm": 0.5210960507392883, "learning_rate": 3.517517308004828e-06, "loss": 0.4074, "num_input_tokens_seen": 44934400, "step": 33165 }, { "epoch": 1.0617117982203443, "grad_norm": 1.272634744644165, "learning_rate": 3.512497589460251e-06, "loss": 0.4503, "num_input_tokens_seen": 44941360, "step": 33170 }, { "epoch": 1.0618718391908328, "grad_norm": 0.6314456462860107, "learning_rate": 3.5074811845585727e-06, "loss": 0.2887, "num_input_tokens_seen": 44948160, "step": 33175 }, { "epoch": 1.0620318801613213, "grad_norm": 1.6934120655059814, "learning_rate": 3.5024680940733937e-06, "loss": 0.4398, "num_input_tokens_seen": 44954912, "step": 33180 }, { "epoch": 1.0621919211318098, "grad_norm": 1.166081190109253, "learning_rate": 3.4974583187777852e-06, "loss": 0.4285, "num_input_tokens_seen": 44961488, "step": 33185 }, { "epoch": 1.062351962102298, "grad_norm": 1.1894233226776123, "learning_rate": 3.4924518594443204e-06, "loss": 0.4971, "num_input_tokens_seen": 44967984, "step": 33190 }, { "epoch": 1.0625120030727866, "grad_norm": 1.067564606666565, "learning_rate": 3.4874487168450682e-06, "loss": 0.5106, "num_input_tokens_seen": 44974624, "step": 33195 }, { "epoch": 1.062672044043275, "grad_norm": 0.6987452507019043, "learning_rate": 3.482448891751558e-06, "loss": 0.3529, "num_input_tokens_seen": 44980912, "step": 33200 }, { "epoch": 1.062672044043275, "eval_loss": 0.44460782408714294, "eval_runtime": 502.7706, "eval_samples_per_second": 27.619, "eval_steps_per_second": 13.809, "num_input_tokens_seen": 44980912, "step": 33200 }, { "epoch": 1.0628320850137636, "grad_norm": 1.4670153856277466, "learning_rate": 3.477452384934843e-06, "loss": 0.5372, "num_input_tokens_seen": 44987664, "step": 33205 }, { "epoch": 1.0629921259842519, "grad_norm": 1.9598435163497925, "learning_rate": 3.472459197165434e-06, "loss": 0.6133, "num_input_tokens_seen": 44994384, "step": 33210 }, { "epoch": 1.0631521669547404, "grad_norm": 0.525270938873291, "learning_rate": 3.4674693292133518e-06, "loss": 0.3369, "num_input_tokens_seen": 45000768, "step": 33215 }, { "epoch": 1.0633122079252288, "grad_norm": 0.998742938041687, "learning_rate": 3.4624827818480977e-06, "loss": 0.3436, "num_input_tokens_seen": 45007568, "step": 33220 }, { "epoch": 1.0634722488957173, "grad_norm": 0.7194823026657104, "learning_rate": 3.4574995558386474e-06, "loss": 0.4219, "num_input_tokens_seen": 45014336, "step": 33225 }, { "epoch": 1.0636322898662058, "grad_norm": 1.1040691137313843, "learning_rate": 3.452519651953487e-06, "loss": 0.31, "num_input_tokens_seen": 45021168, "step": 33230 }, { "epoch": 1.0637923308366941, "grad_norm": 0.9371485710144043, "learning_rate": 3.447543070960585e-06, "loss": 0.5123, "num_input_tokens_seen": 45027776, "step": 33235 }, { "epoch": 1.0639523718071826, "grad_norm": 1.2656536102294922, "learning_rate": 3.4425698136273778e-06, "loss": 0.3128, "num_input_tokens_seen": 45034176, "step": 33240 }, { "epoch": 1.064112412777671, "grad_norm": 0.6413699984550476, "learning_rate": 3.437599880720821e-06, "loss": 0.3547, "num_input_tokens_seen": 45040992, "step": 33245 }, { "epoch": 1.0642724537481596, "grad_norm": 0.8408285975456238, "learning_rate": 3.4326332730073267e-06, "loss": 0.4099, "num_input_tokens_seen": 45047808, "step": 33250 }, { "epoch": 1.0644324947186479, "grad_norm": 1.036442756652832, "learning_rate": 3.427669991252813e-06, "loss": 0.365, "num_input_tokens_seen": 45054384, "step": 33255 }, { "epoch": 1.0645925356891364, "grad_norm": 2.6669797897338867, "learning_rate": 3.42271003622269e-06, "loss": 0.4815, "num_input_tokens_seen": 45061456, "step": 33260 }, { "epoch": 1.0647525766596249, "grad_norm": 0.6784858107566833, "learning_rate": 3.4177534086818286e-06, "loss": 0.3922, "num_input_tokens_seen": 45068144, "step": 33265 }, { "epoch": 1.0649126176301134, "grad_norm": 1.1835790872573853, "learning_rate": 3.412800109394612e-06, "loss": 0.3187, "num_input_tokens_seen": 45074944, "step": 33270 }, { "epoch": 1.0650726586006019, "grad_norm": 0.9168053865432739, "learning_rate": 3.4078501391249044e-06, "loss": 0.5397, "num_input_tokens_seen": 45081888, "step": 33275 }, { "epoch": 1.0652326995710901, "grad_norm": 0.7989587783813477, "learning_rate": 3.4029034986360453e-06, "loss": 0.4462, "num_input_tokens_seen": 45088912, "step": 33280 }, { "epoch": 1.0653927405415786, "grad_norm": 0.9610612392425537, "learning_rate": 3.397960188690877e-06, "loss": 0.3805, "num_input_tokens_seen": 45095728, "step": 33285 }, { "epoch": 1.0655527815120671, "grad_norm": 1.190757393836975, "learning_rate": 3.393020210051717e-06, "loss": 0.4079, "num_input_tokens_seen": 45102448, "step": 33290 }, { "epoch": 1.0657128224825556, "grad_norm": 1.2726505994796753, "learning_rate": 3.3880835634803655e-06, "loss": 0.3442, "num_input_tokens_seen": 45109600, "step": 33295 }, { "epoch": 1.0658728634530439, "grad_norm": 3.0072684288024902, "learning_rate": 3.383150249738126e-06, "loss": 0.454, "num_input_tokens_seen": 45116240, "step": 33300 }, { "epoch": 1.0660329044235324, "grad_norm": 0.6104142665863037, "learning_rate": 3.3782202695857663e-06, "loss": 0.2995, "num_input_tokens_seen": 45122784, "step": 33305 }, { "epoch": 1.0661929453940209, "grad_norm": 0.8755791783332825, "learning_rate": 3.373293623783558e-06, "loss": 0.3425, "num_input_tokens_seen": 45129136, "step": 33310 }, { "epoch": 1.0663529863645094, "grad_norm": 0.8118401169776917, "learning_rate": 3.368370313091257e-06, "loss": 0.2962, "num_input_tokens_seen": 45135856, "step": 33315 }, { "epoch": 1.0665130273349979, "grad_norm": 1.2099074125289917, "learning_rate": 3.363450338268087e-06, "loss": 0.4313, "num_input_tokens_seen": 45142992, "step": 33320 }, { "epoch": 1.0666730683054861, "grad_norm": 1.990649700164795, "learning_rate": 3.358533700072783e-06, "loss": 0.514, "num_input_tokens_seen": 45149872, "step": 33325 }, { "epoch": 1.0668331092759746, "grad_norm": 0.5385273694992065, "learning_rate": 3.3536203992635377e-06, "loss": 0.4242, "num_input_tokens_seen": 45156512, "step": 33330 }, { "epoch": 1.0669931502464631, "grad_norm": 0.572413444519043, "learning_rate": 3.348710436598057e-06, "loss": 0.5374, "num_input_tokens_seen": 45163440, "step": 33335 }, { "epoch": 1.0671531912169516, "grad_norm": 0.8856343626976013, "learning_rate": 3.3438038128335155e-06, "loss": 0.3291, "num_input_tokens_seen": 45170384, "step": 33340 }, { "epoch": 1.06731323218744, "grad_norm": 0.6202824711799622, "learning_rate": 3.338900528726571e-06, "loss": 0.3837, "num_input_tokens_seen": 45177072, "step": 33345 }, { "epoch": 1.0674732731579284, "grad_norm": 0.9020460247993469, "learning_rate": 3.3340005850333812e-06, "loss": 0.4205, "num_input_tokens_seen": 45183936, "step": 33350 }, { "epoch": 1.0676333141284169, "grad_norm": 1.0570766925811768, "learning_rate": 3.329103982509568e-06, "loss": 0.4571, "num_input_tokens_seen": 45190736, "step": 33355 }, { "epoch": 1.0677933550989054, "grad_norm": 2.056213140487671, "learning_rate": 3.324210721910259e-06, "loss": 0.3783, "num_input_tokens_seen": 45197616, "step": 33360 }, { "epoch": 1.0679533960693939, "grad_norm": 0.6890876293182373, "learning_rate": 3.319320803990053e-06, "loss": 0.3513, "num_input_tokens_seen": 45204576, "step": 33365 }, { "epoch": 1.0681134370398822, "grad_norm": 0.5926291346549988, "learning_rate": 3.3144342295030274e-06, "loss": 0.2615, "num_input_tokens_seen": 45210960, "step": 33370 }, { "epoch": 1.0682734780103706, "grad_norm": 0.9363026022911072, "learning_rate": 3.309550999202765e-06, "loss": 0.4182, "num_input_tokens_seen": 45217824, "step": 33375 }, { "epoch": 1.0684335189808591, "grad_norm": 0.4235239624977112, "learning_rate": 3.3046711138423197e-06, "loss": 0.2419, "num_input_tokens_seen": 45224528, "step": 33380 }, { "epoch": 1.0685935599513476, "grad_norm": 1.3822154998779297, "learning_rate": 3.2997945741742255e-06, "loss": 0.4452, "num_input_tokens_seen": 45231168, "step": 33385 }, { "epoch": 1.068753600921836, "grad_norm": 2.742744207382202, "learning_rate": 3.2949213809505082e-06, "loss": 0.3318, "num_input_tokens_seen": 45237600, "step": 33390 }, { "epoch": 1.0689136418923244, "grad_norm": 1.5927644968032837, "learning_rate": 3.2900515349226834e-06, "loss": 0.3441, "num_input_tokens_seen": 45244512, "step": 33395 }, { "epoch": 1.069073682862813, "grad_norm": 3.3958373069763184, "learning_rate": 3.285185036841731e-06, "loss": 0.5797, "num_input_tokens_seen": 45251328, "step": 33400 }, { "epoch": 1.069073682862813, "eval_loss": 0.44465845823287964, "eval_runtime": 502.7919, "eval_samples_per_second": 27.618, "eval_steps_per_second": 13.809, "num_input_tokens_seen": 45251328, "step": 33400 }, { "epoch": 1.0692337238333014, "grad_norm": 2.0126802921295166, "learning_rate": 3.2803218874581377e-06, "loss": 0.5091, "num_input_tokens_seen": 45257632, "step": 33405 }, { "epoch": 1.0693937648037897, "grad_norm": 1.04922354221344, "learning_rate": 3.2754620875218494e-06, "loss": 0.4439, "num_input_tokens_seen": 45264496, "step": 33410 }, { "epoch": 1.0695538057742782, "grad_norm": 1.098311424255371, "learning_rate": 3.2706056377823146e-06, "loss": 0.5387, "num_input_tokens_seen": 45271440, "step": 33415 }, { "epoch": 1.0697138467447667, "grad_norm": 1.6668246984481812, "learning_rate": 3.2657525389884647e-06, "loss": 0.5267, "num_input_tokens_seen": 45278336, "step": 33420 }, { "epoch": 1.0698738877152552, "grad_norm": 1.4827327728271484, "learning_rate": 3.260902791888698e-06, "loss": 0.3685, "num_input_tokens_seen": 45285168, "step": 33425 }, { "epoch": 1.0700339286857437, "grad_norm": 0.8115019798278809, "learning_rate": 3.2560563972309166e-06, "loss": 0.3339, "num_input_tokens_seen": 45292736, "step": 33430 }, { "epoch": 1.070193969656232, "grad_norm": 2.347435474395752, "learning_rate": 3.251213355762489e-06, "loss": 0.4697, "num_input_tokens_seen": 45299552, "step": 33435 }, { "epoch": 1.0703540106267204, "grad_norm": 1.3242868185043335, "learning_rate": 3.2463736682302707e-06, "loss": 0.4432, "num_input_tokens_seen": 45306704, "step": 33440 }, { "epoch": 1.070514051597209, "grad_norm": 1.0009342432022095, "learning_rate": 3.2415373353806124e-06, "loss": 0.4141, "num_input_tokens_seen": 45313808, "step": 33445 }, { "epoch": 1.0706740925676974, "grad_norm": 1.0403082370758057, "learning_rate": 3.236704357959322e-06, "loss": 0.4158, "num_input_tokens_seen": 45320544, "step": 33450 }, { "epoch": 1.0708341335381857, "grad_norm": 1.4042457342147827, "learning_rate": 3.2318747367117154e-06, "loss": 0.5127, "num_input_tokens_seen": 45327248, "step": 33455 }, { "epoch": 1.0709941745086742, "grad_norm": 1.302717685699463, "learning_rate": 3.227048472382585e-06, "loss": 0.419, "num_input_tokens_seen": 45334336, "step": 33460 }, { "epoch": 1.0711542154791627, "grad_norm": 0.6671117544174194, "learning_rate": 3.2222255657161915e-06, "loss": 0.5972, "num_input_tokens_seen": 45340688, "step": 33465 }, { "epoch": 1.0713142564496512, "grad_norm": 0.5375148057937622, "learning_rate": 3.2174060174562924e-06, "loss": 0.4013, "num_input_tokens_seen": 45347952, "step": 33470 }, { "epoch": 1.0714742974201394, "grad_norm": 0.6392238736152649, "learning_rate": 3.2125898283461298e-06, "loss": 0.3946, "num_input_tokens_seen": 45354576, "step": 33475 }, { "epoch": 1.071634338390628, "grad_norm": 1.5154513120651245, "learning_rate": 3.207776999128406e-06, "loss": 0.4605, "num_input_tokens_seen": 45361472, "step": 33480 }, { "epoch": 1.0717943793611164, "grad_norm": 1.001235842704773, "learning_rate": 3.202967530545331e-06, "loss": 0.4677, "num_input_tokens_seen": 45368464, "step": 33485 }, { "epoch": 1.071954420331605, "grad_norm": 1.7098166942596436, "learning_rate": 3.1981614233385778e-06, "loss": 0.4367, "num_input_tokens_seen": 45375872, "step": 33490 }, { "epoch": 1.0721144613020934, "grad_norm": 0.625732421875, "learning_rate": 3.1933586782493115e-06, "loss": 0.3259, "num_input_tokens_seen": 45382640, "step": 33495 }, { "epoch": 1.0722745022725817, "grad_norm": 1.8184020519256592, "learning_rate": 3.188559296018184e-06, "loss": 0.581, "num_input_tokens_seen": 45389344, "step": 33500 }, { "epoch": 1.0724345432430702, "grad_norm": 0.9905248880386353, "learning_rate": 3.1837632773853098e-06, "loss": 0.453, "num_input_tokens_seen": 45395680, "step": 33505 }, { "epoch": 1.0725945842135587, "grad_norm": 1.4244022369384766, "learning_rate": 3.178970623090294e-06, "loss": 0.359, "num_input_tokens_seen": 45402384, "step": 33510 }, { "epoch": 1.0727546251840472, "grad_norm": 1.0605289936065674, "learning_rate": 3.174181333872234e-06, "loss": 0.4588, "num_input_tokens_seen": 45408752, "step": 33515 }, { "epoch": 1.0729146661545355, "grad_norm": 0.8161154985427856, "learning_rate": 3.169395410469686e-06, "loss": 0.4094, "num_input_tokens_seen": 45415328, "step": 33520 }, { "epoch": 1.073074707125024, "grad_norm": 0.40605428814888, "learning_rate": 3.164612853620713e-06, "loss": 0.2653, "num_input_tokens_seen": 45422000, "step": 33525 }, { "epoch": 1.0732347480955124, "grad_norm": 1.9523229598999023, "learning_rate": 3.1598336640628333e-06, "loss": 0.5566, "num_input_tokens_seen": 45429040, "step": 33530 }, { "epoch": 1.073394789066001, "grad_norm": 0.8111315965652466, "learning_rate": 3.155057842533063e-06, "loss": 0.3517, "num_input_tokens_seen": 45435360, "step": 33535 }, { "epoch": 1.0735548300364894, "grad_norm": 0.9385542869567871, "learning_rate": 3.1502853897678984e-06, "loss": 0.5728, "num_input_tokens_seen": 45441920, "step": 33540 }, { "epoch": 1.0737148710069777, "grad_norm": 2.2268357276916504, "learning_rate": 3.1455163065033017e-06, "loss": 0.5184, "num_input_tokens_seen": 45448864, "step": 33545 }, { "epoch": 1.0738749119774662, "grad_norm": 0.6700196266174316, "learning_rate": 3.140750593474734e-06, "loss": 0.2729, "num_input_tokens_seen": 45456176, "step": 33550 }, { "epoch": 1.0740349529479547, "grad_norm": 0.44418150186538696, "learning_rate": 3.1359882514171294e-06, "loss": 0.3924, "num_input_tokens_seen": 45462992, "step": 33555 }, { "epoch": 1.0741949939184432, "grad_norm": 1.5925755500793457, "learning_rate": 3.1312292810648903e-06, "loss": 0.3629, "num_input_tokens_seen": 45469728, "step": 33560 }, { "epoch": 1.0743550348889315, "grad_norm": 0.7240942716598511, "learning_rate": 3.1264736831519204e-06, "loss": 0.5825, "num_input_tokens_seen": 45476576, "step": 33565 }, { "epoch": 1.07451507585942, "grad_norm": 0.6587312817573547, "learning_rate": 3.1217214584115863e-06, "loss": 0.3295, "num_input_tokens_seen": 45483552, "step": 33570 }, { "epoch": 1.0746751168299085, "grad_norm": 1.2045358419418335, "learning_rate": 3.116972607576746e-06, "loss": 0.4699, "num_input_tokens_seen": 45490192, "step": 33575 }, { "epoch": 1.074835157800397, "grad_norm": 3.051600217819214, "learning_rate": 3.1122271313797303e-06, "loss": 0.4647, "num_input_tokens_seen": 45496800, "step": 33580 }, { "epoch": 1.0749951987708855, "grad_norm": 1.1766403913497925, "learning_rate": 3.107485030552343e-06, "loss": 0.353, "num_input_tokens_seen": 45503328, "step": 33585 }, { "epoch": 1.0751552397413737, "grad_norm": 1.1011971235275269, "learning_rate": 3.1027463058258848e-06, "loss": 0.2703, "num_input_tokens_seen": 45510112, "step": 33590 }, { "epoch": 1.0753152807118622, "grad_norm": 1.3815720081329346, "learning_rate": 3.0980109579311273e-06, "loss": 0.4306, "num_input_tokens_seen": 45516752, "step": 33595 }, { "epoch": 1.0754753216823507, "grad_norm": 0.8650616407394409, "learning_rate": 3.093278987598314e-06, "loss": 0.5556, "num_input_tokens_seen": 45523792, "step": 33600 }, { "epoch": 1.0754753216823507, "eval_loss": 0.4446703791618347, "eval_runtime": 503.0175, "eval_samples_per_second": 27.605, "eval_steps_per_second": 13.803, "num_input_tokens_seen": 45523792, "step": 33600 }, { "epoch": 1.0756353626528392, "grad_norm": 1.271663784980774, "learning_rate": 3.0885503955571826e-06, "loss": 0.6192, "num_input_tokens_seen": 45530608, "step": 33605 }, { "epoch": 1.0757954036233275, "grad_norm": 0.9228882193565369, "learning_rate": 3.0838251825369313e-06, "loss": 0.2944, "num_input_tokens_seen": 45537568, "step": 33610 }, { "epoch": 1.075955444593816, "grad_norm": 2.2598724365234375, "learning_rate": 3.0791033492662517e-06, "loss": 0.5739, "num_input_tokens_seen": 45544592, "step": 33615 }, { "epoch": 1.0761154855643045, "grad_norm": 0.8038756251335144, "learning_rate": 3.0743848964733203e-06, "loss": 0.383, "num_input_tokens_seen": 45551440, "step": 33620 }, { "epoch": 1.076275526534793, "grad_norm": 0.9066963791847229, "learning_rate": 3.0696698248857625e-06, "loss": 0.4581, "num_input_tokens_seen": 45558096, "step": 33625 }, { "epoch": 1.0764355675052815, "grad_norm": 1.1844284534454346, "learning_rate": 3.0649581352307192e-06, "loss": 0.4189, "num_input_tokens_seen": 45564736, "step": 33630 }, { "epoch": 1.0765956084757697, "grad_norm": 1.4909249544143677, "learning_rate": 3.060249828234776e-06, "loss": 0.4458, "num_input_tokens_seen": 45571504, "step": 33635 }, { "epoch": 1.0767556494462582, "grad_norm": 0.8021509051322937, "learning_rate": 3.055544904624025e-06, "loss": 0.2728, "num_input_tokens_seen": 45578848, "step": 33640 }, { "epoch": 1.0769156904167467, "grad_norm": 0.7465742826461792, "learning_rate": 3.050843365124026e-06, "loss": 0.4963, "num_input_tokens_seen": 45585472, "step": 33645 }, { "epoch": 1.0770757313872352, "grad_norm": 1.9507709741592407, "learning_rate": 3.0461452104598083e-06, "loss": 0.4693, "num_input_tokens_seen": 45592560, "step": 33650 }, { "epoch": 1.0772357723577235, "grad_norm": 1.426255226135254, "learning_rate": 3.0414504413558836e-06, "loss": 0.4209, "num_input_tokens_seen": 45599792, "step": 33655 }, { "epoch": 1.077395813328212, "grad_norm": 1.1476352214813232, "learning_rate": 3.0367590585362564e-06, "loss": 0.2931, "num_input_tokens_seen": 45606608, "step": 33660 }, { "epoch": 1.0775558542987005, "grad_norm": 1.3245009183883667, "learning_rate": 3.0320710627243813e-06, "loss": 0.3608, "num_input_tokens_seen": 45614192, "step": 33665 }, { "epoch": 1.077715895269189, "grad_norm": 0.915649950504303, "learning_rate": 3.027386454643222e-06, "loss": 0.4813, "num_input_tokens_seen": 45620720, "step": 33670 }, { "epoch": 1.0778759362396773, "grad_norm": 0.35003072023391724, "learning_rate": 3.0227052350151914e-06, "loss": 0.3921, "num_input_tokens_seen": 45627392, "step": 33675 }, { "epoch": 1.0780359772101658, "grad_norm": 0.676565945148468, "learning_rate": 3.0180274045621957e-06, "loss": 0.2742, "num_input_tokens_seen": 45633776, "step": 33680 }, { "epoch": 1.0781960181806542, "grad_norm": 0.9660544395446777, "learning_rate": 3.013352964005625e-06, "loss": 0.3435, "num_input_tokens_seen": 45640912, "step": 33685 }, { "epoch": 1.0783560591511427, "grad_norm": 0.4874199628829956, "learning_rate": 3.0086819140663218e-06, "loss": 0.358, "num_input_tokens_seen": 45647456, "step": 33690 }, { "epoch": 1.0785161001216312, "grad_norm": 1.3395614624023438, "learning_rate": 3.0040142554646265e-06, "loss": 0.5967, "num_input_tokens_seen": 45654048, "step": 33695 }, { "epoch": 1.0786761410921195, "grad_norm": 2.6767826080322266, "learning_rate": 2.999349988920361e-06, "loss": 0.4866, "num_input_tokens_seen": 45661216, "step": 33700 }, { "epoch": 1.078836182062608, "grad_norm": 0.6378637552261353, "learning_rate": 2.994689115152796e-06, "loss": 0.3162, "num_input_tokens_seen": 45667968, "step": 33705 }, { "epoch": 1.0789962230330965, "grad_norm": 0.700838029384613, "learning_rate": 2.9900316348807105e-06, "loss": 0.3376, "num_input_tokens_seen": 45674592, "step": 33710 }, { "epoch": 1.079156264003585, "grad_norm": 1.2221850156784058, "learning_rate": 2.985377548822338e-06, "loss": 0.4141, "num_input_tokens_seen": 45681632, "step": 33715 }, { "epoch": 1.0793163049740733, "grad_norm": 0.836270809173584, "learning_rate": 2.980726857695404e-06, "loss": 0.3543, "num_input_tokens_seen": 45687984, "step": 33720 }, { "epoch": 1.0794763459445618, "grad_norm": 0.4565238356590271, "learning_rate": 2.9760795622171017e-06, "loss": 0.4208, "num_input_tokens_seen": 45694224, "step": 33725 }, { "epoch": 1.0796363869150503, "grad_norm": 1.718531847000122, "learning_rate": 2.971435663104094e-06, "loss": 0.629, "num_input_tokens_seen": 45701056, "step": 33730 }, { "epoch": 1.0797964278855388, "grad_norm": 1.1408711671829224, "learning_rate": 2.9667951610725385e-06, "loss": 0.5274, "num_input_tokens_seen": 45707456, "step": 33735 }, { "epoch": 1.079956468856027, "grad_norm": 2.47945237159729, "learning_rate": 2.9621580568380575e-06, "loss": 0.4746, "num_input_tokens_seen": 45714752, "step": 33740 }, { "epoch": 1.0801165098265155, "grad_norm": 0.33378246426582336, "learning_rate": 2.9575243511157453e-06, "loss": 0.2578, "num_input_tokens_seen": 45721392, "step": 33745 }, { "epoch": 1.080276550797004, "grad_norm": 0.6492066383361816, "learning_rate": 2.952894044620186e-06, "loss": 0.4164, "num_input_tokens_seen": 45727664, "step": 33750 }, { "epoch": 1.0804365917674925, "grad_norm": 0.8198605179786682, "learning_rate": 2.948267138065419e-06, "loss": 0.3875, "num_input_tokens_seen": 45734976, "step": 33755 }, { "epoch": 1.080596632737981, "grad_norm": 1.400511384010315, "learning_rate": 2.943643632164983e-06, "loss": 0.5418, "num_input_tokens_seen": 45741936, "step": 33760 }, { "epoch": 1.0807566737084693, "grad_norm": 1.9108517169952393, "learning_rate": 2.939023527631879e-06, "loss": 0.397, "num_input_tokens_seen": 45748784, "step": 33765 }, { "epoch": 1.0809167146789578, "grad_norm": 1.047094702720642, "learning_rate": 2.934406825178576e-06, "loss": 0.4874, "num_input_tokens_seen": 45755616, "step": 33770 }, { "epoch": 1.0810767556494463, "grad_norm": 0.7765690088272095, "learning_rate": 2.9297935255170357e-06, "loss": 0.365, "num_input_tokens_seen": 45762352, "step": 33775 }, { "epoch": 1.0812367966199348, "grad_norm": 1.0933986902236938, "learning_rate": 2.925183629358691e-06, "loss": 0.4243, "num_input_tokens_seen": 45768816, "step": 33780 }, { "epoch": 1.081396837590423, "grad_norm": 1.1923935413360596, "learning_rate": 2.9205771374144346e-06, "loss": 0.356, "num_input_tokens_seen": 45775776, "step": 33785 }, { "epoch": 1.0815568785609115, "grad_norm": 1.3781518936157227, "learning_rate": 2.915974050394657e-06, "loss": 0.5655, "num_input_tokens_seen": 45782736, "step": 33790 }, { "epoch": 1.0817169195314, "grad_norm": 1.0250869989395142, "learning_rate": 2.9113743690092067e-06, "loss": 0.5282, "num_input_tokens_seen": 45789968, "step": 33795 }, { "epoch": 1.0818769605018885, "grad_norm": 1.137986660003662, "learning_rate": 2.906778093967402e-06, "loss": 0.3862, "num_input_tokens_seen": 45796960, "step": 33800 }, { "epoch": 1.0818769605018885, "eval_loss": 0.4445202648639679, "eval_runtime": 502.817, "eval_samples_per_second": 27.616, "eval_steps_per_second": 13.808, "num_input_tokens_seen": 45796960, "step": 33800 }, { "epoch": 1.082037001472377, "grad_norm": 1.2508784532546997, "learning_rate": 2.9021852259780656e-06, "loss": 0.6183, "num_input_tokens_seen": 45803328, "step": 33805 }, { "epoch": 1.0821970424428653, "grad_norm": 1.2071690559387207, "learning_rate": 2.8975957657494583e-06, "loss": 0.5732, "num_input_tokens_seen": 45810352, "step": 33810 }, { "epoch": 1.0823570834133538, "grad_norm": 0.706307590007782, "learning_rate": 2.8930097139893417e-06, "loss": 0.3819, "num_input_tokens_seen": 45817344, "step": 33815 }, { "epoch": 1.0825171243838423, "grad_norm": 0.4975731074810028, "learning_rate": 2.888427071404945e-06, "loss": 0.2931, "num_input_tokens_seen": 45824032, "step": 33820 }, { "epoch": 1.0826771653543308, "grad_norm": 1.4898641109466553, "learning_rate": 2.8838478387029606e-06, "loss": 0.7345, "num_input_tokens_seen": 45830944, "step": 33825 }, { "epoch": 1.082837206324819, "grad_norm": 0.973395586013794, "learning_rate": 2.8792720165895737e-06, "loss": 0.3451, "num_input_tokens_seen": 45837632, "step": 33830 }, { "epoch": 1.0829972472953076, "grad_norm": 0.555379331111908, "learning_rate": 2.874699605770423e-06, "loss": 0.3608, "num_input_tokens_seen": 45844800, "step": 33835 }, { "epoch": 1.083157288265796, "grad_norm": 0.6658756136894226, "learning_rate": 2.8701306069506383e-06, "loss": 0.378, "num_input_tokens_seen": 45851456, "step": 33840 }, { "epoch": 1.0833173292362845, "grad_norm": 0.7843957543373108, "learning_rate": 2.8655650208348178e-06, "loss": 0.4319, "num_input_tokens_seen": 45858176, "step": 33845 }, { "epoch": 1.083477370206773, "grad_norm": 0.7227602005004883, "learning_rate": 2.8610028481270257e-06, "loss": 0.3737, "num_input_tokens_seen": 45864864, "step": 33850 }, { "epoch": 1.0836374111772613, "grad_norm": 1.616110920906067, "learning_rate": 2.856444089530813e-06, "loss": 0.5974, "num_input_tokens_seen": 45871856, "step": 33855 }, { "epoch": 1.0837974521477498, "grad_norm": 1.5797860622406006, "learning_rate": 2.8518887457491955e-06, "loss": 0.3978, "num_input_tokens_seen": 45878368, "step": 33860 }, { "epoch": 1.0839574931182383, "grad_norm": 0.6262904405593872, "learning_rate": 2.8473368174846666e-06, "loss": 0.4576, "num_input_tokens_seen": 45885328, "step": 33865 }, { "epoch": 1.0841175340887268, "grad_norm": 0.6495195031166077, "learning_rate": 2.842788305439184e-06, "loss": 0.4003, "num_input_tokens_seen": 45892304, "step": 33870 }, { "epoch": 1.084277575059215, "grad_norm": 1.1416215896606445, "learning_rate": 2.8382432103141925e-06, "loss": 0.4986, "num_input_tokens_seen": 45899120, "step": 33875 }, { "epoch": 1.0844376160297036, "grad_norm": 1.9488329887390137, "learning_rate": 2.833701532810598e-06, "loss": 0.573, "num_input_tokens_seen": 45905696, "step": 33880 }, { "epoch": 1.084597657000192, "grad_norm": 1.7601947784423828, "learning_rate": 2.8291632736287877e-06, "loss": 0.5237, "num_input_tokens_seen": 45912496, "step": 33885 }, { "epoch": 1.0847576979706806, "grad_norm": 0.5052145719528198, "learning_rate": 2.824628433468615e-06, "loss": 0.3315, "num_input_tokens_seen": 45919600, "step": 33890 }, { "epoch": 1.084917738941169, "grad_norm": 1.678670048713684, "learning_rate": 2.8200970130294073e-06, "loss": 0.4962, "num_input_tokens_seen": 45926864, "step": 33895 }, { "epoch": 1.0850777799116573, "grad_norm": 1.1642227172851562, "learning_rate": 2.8155690130099775e-06, "loss": 0.642, "num_input_tokens_seen": 45933536, "step": 33900 }, { "epoch": 1.0852378208821458, "grad_norm": 1.3740437030792236, "learning_rate": 2.8110444341085895e-06, "loss": 0.3578, "num_input_tokens_seen": 45940304, "step": 33905 }, { "epoch": 1.0853978618526343, "grad_norm": 0.6542921662330627, "learning_rate": 2.806523277022996e-06, "loss": 0.363, "num_input_tokens_seen": 45946768, "step": 33910 }, { "epoch": 1.0855579028231228, "grad_norm": 1.1505308151245117, "learning_rate": 2.802005542450409e-06, "loss": 0.4747, "num_input_tokens_seen": 45953360, "step": 33915 }, { "epoch": 1.085717943793611, "grad_norm": 2.3430116176605225, "learning_rate": 2.797491231087526e-06, "loss": 0.5884, "num_input_tokens_seen": 45959904, "step": 33920 }, { "epoch": 1.0858779847640996, "grad_norm": 0.7858442068099976, "learning_rate": 2.7929803436305137e-06, "loss": 0.4347, "num_input_tokens_seen": 45966512, "step": 33925 }, { "epoch": 1.086038025734588, "grad_norm": 1.433327078819275, "learning_rate": 2.788472880774998e-06, "loss": 0.467, "num_input_tokens_seen": 45973264, "step": 33930 }, { "epoch": 1.0861980667050766, "grad_norm": 1.848134994506836, "learning_rate": 2.7839688432160977e-06, "loss": 0.54, "num_input_tokens_seen": 45980416, "step": 33935 }, { "epoch": 1.0863581076755648, "grad_norm": 0.577246367931366, "learning_rate": 2.779468231648383e-06, "loss": 0.3398, "num_input_tokens_seen": 45987008, "step": 33940 }, { "epoch": 1.0865181486460533, "grad_norm": 1.0914698839187622, "learning_rate": 2.774971046765906e-06, "loss": 0.4831, "num_input_tokens_seen": 45993984, "step": 33945 }, { "epoch": 1.0866781896165418, "grad_norm": 1.2421008348464966, "learning_rate": 2.770477289262194e-06, "loss": 0.3324, "num_input_tokens_seen": 46000464, "step": 33950 }, { "epoch": 1.0868382305870303, "grad_norm": 0.5743545293807983, "learning_rate": 2.765986959830233e-06, "loss": 0.609, "num_input_tokens_seen": 46007584, "step": 33955 }, { "epoch": 1.0869982715575188, "grad_norm": 1.1957156658172607, "learning_rate": 2.761500059162492e-06, "loss": 0.2785, "num_input_tokens_seen": 46014160, "step": 33960 }, { "epoch": 1.087158312528007, "grad_norm": 0.707548201084137, "learning_rate": 2.757016587950914e-06, "loss": 0.3598, "num_input_tokens_seen": 46020912, "step": 33965 }, { "epoch": 1.0873183534984956, "grad_norm": 1.028656005859375, "learning_rate": 2.752536546886897e-06, "loss": 0.3913, "num_input_tokens_seen": 46027616, "step": 33970 }, { "epoch": 1.087478394468984, "grad_norm": 2.778656244277954, "learning_rate": 2.7480599366613234e-06, "loss": 0.5099, "num_input_tokens_seen": 46034208, "step": 33975 }, { "epoch": 1.0876384354394726, "grad_norm": 0.7964799404144287, "learning_rate": 2.7435867579645473e-06, "loss": 0.3831, "num_input_tokens_seen": 46041296, "step": 33980 }, { "epoch": 1.0877984764099609, "grad_norm": 1.149257779121399, "learning_rate": 2.739117011486378e-06, "loss": 0.4035, "num_input_tokens_seen": 46048032, "step": 33985 }, { "epoch": 1.0879585173804494, "grad_norm": 1.3496973514556885, "learning_rate": 2.7346506979161216e-06, "loss": 0.4028, "num_input_tokens_seen": 46055056, "step": 33990 }, { "epoch": 1.0881185583509378, "grad_norm": 0.7763625979423523, "learning_rate": 2.7301878179425227e-06, "loss": 0.3238, "num_input_tokens_seen": 46061632, "step": 33995 }, { "epoch": 1.0882785993214263, "grad_norm": 0.9190540313720703, "learning_rate": 2.7257283722538244e-06, "loss": 0.389, "num_input_tokens_seen": 46067712, "step": 34000 }, { "epoch": 1.0882785993214263, "eval_loss": 0.4443865120410919, "eval_runtime": 502.8959, "eval_samples_per_second": 27.612, "eval_steps_per_second": 13.806, "num_input_tokens_seen": 46067712, "step": 34000 }, { "epoch": 1.0884386402919146, "grad_norm": 0.8316935896873474, "learning_rate": 2.7212723615377326e-06, "loss": 0.4751, "num_input_tokens_seen": 46074592, "step": 34005 }, { "epoch": 1.0885986812624031, "grad_norm": 0.6808618307113647, "learning_rate": 2.7168197864814145e-06, "loss": 0.4925, "num_input_tokens_seen": 46081232, "step": 34010 }, { "epoch": 1.0887587222328916, "grad_norm": 1.0696309804916382, "learning_rate": 2.712370647771509e-06, "loss": 0.3255, "num_input_tokens_seen": 46087744, "step": 34015 }, { "epoch": 1.08891876320338, "grad_norm": 0.7853757739067078, "learning_rate": 2.707924946094137e-06, "loss": 0.3904, "num_input_tokens_seen": 46094208, "step": 34020 }, { "epoch": 1.0890788041738686, "grad_norm": 0.6183369755744934, "learning_rate": 2.7034826821348723e-06, "loss": 0.3031, "num_input_tokens_seen": 46101104, "step": 34025 }, { "epoch": 1.0892388451443569, "grad_norm": 1.3745207786560059, "learning_rate": 2.6990438565787786e-06, "loss": 0.4478, "num_input_tokens_seen": 46107920, "step": 34030 }, { "epoch": 1.0893988861148454, "grad_norm": 1.2598720788955688, "learning_rate": 2.6946084701103714e-06, "loss": 0.4803, "num_input_tokens_seen": 46114752, "step": 34035 }, { "epoch": 1.0895589270853339, "grad_norm": 1.1498141288757324, "learning_rate": 2.6901765234136428e-06, "loss": 0.3992, "num_input_tokens_seen": 46122320, "step": 34040 }, { "epoch": 1.0897189680558224, "grad_norm": 1.1681876182556152, "learning_rate": 2.685748017172063e-06, "loss": 0.4791, "num_input_tokens_seen": 46128752, "step": 34045 }, { "epoch": 1.0898790090263106, "grad_norm": 1.2102574110031128, "learning_rate": 2.681322952068549e-06, "loss": 0.4536, "num_input_tokens_seen": 46135792, "step": 34050 }, { "epoch": 1.0900390499967991, "grad_norm": 1.2761175632476807, "learning_rate": 2.6769013287855137e-06, "loss": 0.4033, "num_input_tokens_seen": 46142240, "step": 34055 }, { "epoch": 1.0901990909672876, "grad_norm": 0.6228249669075012, "learning_rate": 2.6724831480048286e-06, "loss": 0.4302, "num_input_tokens_seen": 46149120, "step": 34060 }, { "epoch": 1.0903591319377761, "grad_norm": 1.179463267326355, "learning_rate": 2.66806841040782e-06, "loss": 0.3688, "num_input_tokens_seen": 46155904, "step": 34065 }, { "epoch": 1.0905191729082646, "grad_norm": 2.1063709259033203, "learning_rate": 2.6636571166753083e-06, "loss": 0.5055, "num_input_tokens_seen": 46162784, "step": 34070 }, { "epoch": 1.0906792138787529, "grad_norm": 0.9148886799812317, "learning_rate": 2.6592492674875598e-06, "loss": 0.3701, "num_input_tokens_seen": 46169424, "step": 34075 }, { "epoch": 1.0908392548492414, "grad_norm": 0.905065655708313, "learning_rate": 2.6548448635243305e-06, "loss": 0.4615, "num_input_tokens_seen": 46176144, "step": 34080 }, { "epoch": 1.0909992958197299, "grad_norm": 1.7496651411056519, "learning_rate": 2.650443905464828e-06, "loss": 0.5063, "num_input_tokens_seen": 46182976, "step": 34085 }, { "epoch": 1.0911593367902184, "grad_norm": 0.9222688674926758, "learning_rate": 2.646046393987739e-06, "loss": 0.4074, "num_input_tokens_seen": 46189264, "step": 34090 }, { "epoch": 1.0913193777607066, "grad_norm": 1.8100777864456177, "learning_rate": 2.64165232977121e-06, "loss": 0.322, "num_input_tokens_seen": 46195856, "step": 34095 }, { "epoch": 1.0914794187311951, "grad_norm": 0.8745064735412598, "learning_rate": 2.6372617134928695e-06, "loss": 0.3881, "num_input_tokens_seen": 46202208, "step": 34100 }, { "epoch": 1.0916394597016836, "grad_norm": 0.9327558875083923, "learning_rate": 2.6328745458297943e-06, "loss": 0.4248, "num_input_tokens_seen": 46208800, "step": 34105 }, { "epoch": 1.0917995006721721, "grad_norm": 0.8306865096092224, "learning_rate": 2.6284908274585546e-06, "loss": 0.5121, "num_input_tokens_seen": 46215520, "step": 34110 }, { "epoch": 1.0919595416426606, "grad_norm": 0.8784945011138916, "learning_rate": 2.6241105590551595e-06, "loss": 0.3502, "num_input_tokens_seen": 46222208, "step": 34115 }, { "epoch": 1.092119582613149, "grad_norm": 1.4159274101257324, "learning_rate": 2.6197337412951105e-06, "loss": 0.5449, "num_input_tokens_seen": 46229008, "step": 34120 }, { "epoch": 1.0922796235836374, "grad_norm": 0.8490428924560547, "learning_rate": 2.6153603748533705e-06, "loss": 0.5539, "num_input_tokens_seen": 46235680, "step": 34125 }, { "epoch": 1.0924396645541259, "grad_norm": 0.9915359616279602, "learning_rate": 2.6109904604043585e-06, "loss": 0.4471, "num_input_tokens_seen": 46242416, "step": 34130 }, { "epoch": 1.0925997055246144, "grad_norm": 0.9093831181526184, "learning_rate": 2.6066239986219765e-06, "loss": 0.3782, "num_input_tokens_seen": 46248928, "step": 34135 }, { "epoch": 1.0927597464951027, "grad_norm": 1.0937585830688477, "learning_rate": 2.602260990179592e-06, "loss": 0.4669, "num_input_tokens_seen": 46255984, "step": 34140 }, { "epoch": 1.0929197874655912, "grad_norm": 0.9037122130393982, "learning_rate": 2.5979014357500248e-06, "loss": 0.5269, "num_input_tokens_seen": 46262480, "step": 34145 }, { "epoch": 1.0930798284360796, "grad_norm": 1.5674335956573486, "learning_rate": 2.5935453360055844e-06, "loss": 0.5747, "num_input_tokens_seen": 46269168, "step": 34150 }, { "epoch": 1.0932398694065681, "grad_norm": 1.2715978622436523, "learning_rate": 2.5891926916180283e-06, "loss": 0.5548, "num_input_tokens_seen": 46275536, "step": 34155 }, { "epoch": 1.0933999103770566, "grad_norm": 0.619799017906189, "learning_rate": 2.5848435032585883e-06, "loss": 0.3273, "num_input_tokens_seen": 46282656, "step": 34160 }, { "epoch": 1.093559951347545, "grad_norm": 0.7707210779190063, "learning_rate": 2.58049777159797e-06, "loss": 0.299, "num_input_tokens_seen": 46289200, "step": 34165 }, { "epoch": 1.0937199923180334, "grad_norm": 0.46836432814598083, "learning_rate": 2.576155497306332e-06, "loss": 0.3884, "num_input_tokens_seen": 46296576, "step": 34170 }, { "epoch": 1.093880033288522, "grad_norm": 0.9437069296836853, "learning_rate": 2.57181668105331e-06, "loss": 0.2224, "num_input_tokens_seen": 46303728, "step": 34175 }, { "epoch": 1.0940400742590104, "grad_norm": 1.0982567071914673, "learning_rate": 2.567481323508014e-06, "loss": 0.492, "num_input_tokens_seen": 46310352, "step": 34180 }, { "epoch": 1.0942001152294987, "grad_norm": 0.3943024277687073, "learning_rate": 2.5631494253389954e-06, "loss": 0.4123, "num_input_tokens_seen": 46316960, "step": 34185 }, { "epoch": 1.0943601561999872, "grad_norm": 2.556516647338867, "learning_rate": 2.5588209872142997e-06, "loss": 0.55, "num_input_tokens_seen": 46324288, "step": 34190 }, { "epoch": 1.0945201971704757, "grad_norm": 1.2429263591766357, "learning_rate": 2.5544960098014186e-06, "loss": 0.3449, "num_input_tokens_seen": 46330880, "step": 34195 }, { "epoch": 1.0946802381409642, "grad_norm": 0.8967440724372864, "learning_rate": 2.550174493767318e-06, "loss": 0.3956, "num_input_tokens_seen": 46337408, "step": 34200 }, { "epoch": 1.0946802381409642, "eval_loss": 0.44434598088264465, "eval_runtime": 503.6563, "eval_samples_per_second": 27.57, "eval_steps_per_second": 13.785, "num_input_tokens_seen": 46337408, "step": 34200 }, { "epoch": 1.0948402791114524, "grad_norm": 1.2407855987548828, "learning_rate": 2.545856439778438e-06, "loss": 0.4129, "num_input_tokens_seen": 46344112, "step": 34205 }, { "epoch": 1.095000320081941, "grad_norm": 0.5687593221664429, "learning_rate": 2.541541848500667e-06, "loss": 0.4377, "num_input_tokens_seen": 46350784, "step": 34210 }, { "epoch": 1.0951603610524294, "grad_norm": 1.1773306131362915, "learning_rate": 2.5372307205993733e-06, "loss": 0.2452, "num_input_tokens_seen": 46357904, "step": 34215 }, { "epoch": 1.095320402022918, "grad_norm": 0.7462024092674255, "learning_rate": 2.5329230567393917e-06, "loss": 0.4057, "num_input_tokens_seen": 46364912, "step": 34220 }, { "epoch": 1.0954804429934064, "grad_norm": 0.62778639793396, "learning_rate": 2.5286188575850164e-06, "loss": 0.4315, "num_input_tokens_seen": 46371824, "step": 34225 }, { "epoch": 1.0956404839638947, "grad_norm": 0.5797130465507507, "learning_rate": 2.5243181237999984e-06, "loss": 0.3997, "num_input_tokens_seen": 46378800, "step": 34230 }, { "epoch": 1.0958005249343832, "grad_norm": 1.6674059629440308, "learning_rate": 2.520020856047578e-06, "loss": 0.5355, "num_input_tokens_seen": 46385872, "step": 34235 }, { "epoch": 1.0959605659048717, "grad_norm": 0.7018206119537354, "learning_rate": 2.515727054990438e-06, "loss": 0.4523, "num_input_tokens_seen": 46392704, "step": 34240 }, { "epoch": 1.0961206068753602, "grad_norm": 1.8269706964492798, "learning_rate": 2.511436721290747e-06, "loss": 0.4514, "num_input_tokens_seen": 46400144, "step": 34245 }, { "epoch": 1.0962806478458484, "grad_norm": 1.578245997428894, "learning_rate": 2.5071498556101164e-06, "loss": 0.6053, "num_input_tokens_seen": 46406880, "step": 34250 }, { "epoch": 1.096440688816337, "grad_norm": 1.6239943504333496, "learning_rate": 2.5028664586096485e-06, "loss": 0.4186, "num_input_tokens_seen": 46413488, "step": 34255 }, { "epoch": 1.0966007297868254, "grad_norm": 0.5684275031089783, "learning_rate": 2.498586530949881e-06, "loss": 0.4118, "num_input_tokens_seen": 46420320, "step": 34260 }, { "epoch": 1.096760770757314, "grad_norm": 1.1810650825500488, "learning_rate": 2.4943100732908427e-06, "loss": 0.7715, "num_input_tokens_seen": 46427120, "step": 34265 }, { "epoch": 1.0969208117278022, "grad_norm": 0.7537782192230225, "learning_rate": 2.4900370862920188e-06, "loss": 0.2598, "num_input_tokens_seen": 46434272, "step": 34270 }, { "epoch": 1.0970808526982907, "grad_norm": 1.179762601852417, "learning_rate": 2.4857675706123518e-06, "loss": 0.3117, "num_input_tokens_seen": 46441104, "step": 34275 }, { "epoch": 1.0972408936687792, "grad_norm": 1.07290518283844, "learning_rate": 2.4815015269102543e-06, "loss": 0.3729, "num_input_tokens_seen": 46448048, "step": 34280 }, { "epoch": 1.0974009346392677, "grad_norm": 0.9391834139823914, "learning_rate": 2.477238955843611e-06, "loss": 0.4342, "num_input_tokens_seen": 46455136, "step": 34285 }, { "epoch": 1.0975609756097562, "grad_norm": 1.510071039199829, "learning_rate": 2.4729798580697573e-06, "loss": 0.5644, "num_input_tokens_seen": 46461872, "step": 34290 }, { "epoch": 1.0977210165802445, "grad_norm": 1.4620819091796875, "learning_rate": 2.4687242342455034e-06, "loss": 0.5041, "num_input_tokens_seen": 46468752, "step": 34295 }, { "epoch": 1.097881057550733, "grad_norm": 0.6041153073310852, "learning_rate": 2.4644720850271196e-06, "loss": 0.3851, "num_input_tokens_seen": 46475536, "step": 34300 }, { "epoch": 1.0980410985212214, "grad_norm": 0.5011281967163086, "learning_rate": 2.4602234110703364e-06, "loss": 0.4048, "num_input_tokens_seen": 46482208, "step": 34305 }, { "epoch": 1.09820113949171, "grad_norm": 0.7118455171585083, "learning_rate": 2.4559782130303576e-06, "loss": 0.3628, "num_input_tokens_seen": 46489280, "step": 34310 }, { "epoch": 1.0983611804621982, "grad_norm": 0.8173003792762756, "learning_rate": 2.451736491561843e-06, "loss": 0.4662, "num_input_tokens_seen": 46495984, "step": 34315 }, { "epoch": 1.0985212214326867, "grad_norm": 1.2417197227478027, "learning_rate": 2.4474982473189163e-06, "loss": 0.335, "num_input_tokens_seen": 46502400, "step": 34320 }, { "epoch": 1.0986812624031752, "grad_norm": 2.060359001159668, "learning_rate": 2.4432634809551796e-06, "loss": 0.5307, "num_input_tokens_seen": 46509216, "step": 34325 }, { "epoch": 1.0988413033736637, "grad_norm": 0.9202799201011658, "learning_rate": 2.439032193123675e-06, "loss": 0.4875, "num_input_tokens_seen": 46516144, "step": 34330 }, { "epoch": 1.0990013443441522, "grad_norm": 1.3665153980255127, "learning_rate": 2.4348043844769297e-06, "loss": 0.5312, "num_input_tokens_seen": 46523024, "step": 34335 }, { "epoch": 1.0991613853146405, "grad_norm": 0.6525409817695618, "learning_rate": 2.4305800556669146e-06, "loss": 0.396, "num_input_tokens_seen": 46529728, "step": 34340 }, { "epoch": 1.099321426285129, "grad_norm": 1.605680227279663, "learning_rate": 2.426359207345083e-06, "loss": 0.3903, "num_input_tokens_seen": 46536560, "step": 34345 }, { "epoch": 1.0994814672556175, "grad_norm": 0.6360828876495361, "learning_rate": 2.4221418401623396e-06, "loss": 0.3898, "num_input_tokens_seen": 46544128, "step": 34350 }, { "epoch": 1.099641508226106, "grad_norm": 2.1078362464904785, "learning_rate": 2.4179279547690557e-06, "loss": 0.4402, "num_input_tokens_seen": 46550880, "step": 34355 }, { "epoch": 1.0998015491965942, "grad_norm": 0.6433507800102234, "learning_rate": 2.413717551815062e-06, "loss": 0.2695, "num_input_tokens_seen": 46557408, "step": 34360 }, { "epoch": 1.0999615901670827, "grad_norm": 0.5911909937858582, "learning_rate": 2.409510631949666e-06, "loss": 0.2816, "num_input_tokens_seen": 46563840, "step": 34365 }, { "epoch": 1.1001216311375712, "grad_norm": 0.9246445298194885, "learning_rate": 2.405307195821618e-06, "loss": 0.5647, "num_input_tokens_seen": 46570384, "step": 34370 }, { "epoch": 1.1002816721080597, "grad_norm": 1.1587462425231934, "learning_rate": 2.4011072440791372e-06, "loss": 0.4636, "num_input_tokens_seen": 46577168, "step": 34375 }, { "epoch": 1.1004417130785482, "grad_norm": 0.8357776403427124, "learning_rate": 2.3969107773699233e-06, "loss": 0.36, "num_input_tokens_seen": 46583936, "step": 34380 }, { "epoch": 1.1006017540490365, "grad_norm": 0.5907230973243713, "learning_rate": 2.3927177963411096e-06, "loss": 0.3162, "num_input_tokens_seen": 46590736, "step": 34385 }, { "epoch": 1.100761795019525, "grad_norm": 0.8806478977203369, "learning_rate": 2.3885283016393144e-06, "loss": 0.4488, "num_input_tokens_seen": 46597824, "step": 34390 }, { "epoch": 1.1009218359900135, "grad_norm": 1.1280182600021362, "learning_rate": 2.3843422939106076e-06, "loss": 0.3603, "num_input_tokens_seen": 46604656, "step": 34395 }, { "epoch": 1.101081876960502, "grad_norm": 1.2924327850341797, "learning_rate": 2.380159773800525e-06, "loss": 0.5701, "num_input_tokens_seen": 46611232, "step": 34400 }, { "epoch": 1.101081876960502, "eval_loss": 0.44431573152542114, "eval_runtime": 503.3707, "eval_samples_per_second": 27.586, "eval_steps_per_second": 13.793, "num_input_tokens_seen": 46611232, "step": 34400 }, { "epoch": 1.1012419179309902, "grad_norm": 0.49073857069015503, "learning_rate": 2.3759807419540675e-06, "loss": 0.3589, "num_input_tokens_seen": 46618000, "step": 34405 }, { "epoch": 1.1014019589014787, "grad_norm": 1.0741201639175415, "learning_rate": 2.3718051990156835e-06, "loss": 0.5397, "num_input_tokens_seen": 46624464, "step": 34410 }, { "epoch": 1.1015619998719672, "grad_norm": 0.7202144861221313, "learning_rate": 2.367633145629311e-06, "loss": 0.4337, "num_input_tokens_seen": 46631088, "step": 34415 }, { "epoch": 1.1017220408424557, "grad_norm": 1.1068446636199951, "learning_rate": 2.363464582438316e-06, "loss": 0.5715, "num_input_tokens_seen": 46637664, "step": 34420 }, { "epoch": 1.1018820818129442, "grad_norm": 1.5588197708129883, "learning_rate": 2.3592995100855526e-06, "loss": 0.4104, "num_input_tokens_seen": 46644608, "step": 34425 }, { "epoch": 1.1020421227834325, "grad_norm": 1.1119931936264038, "learning_rate": 2.3551379292133273e-06, "loss": 0.3646, "num_input_tokens_seen": 46651088, "step": 34430 }, { "epoch": 1.102202163753921, "grad_norm": 1.2950819730758667, "learning_rate": 2.3509798404634047e-06, "loss": 0.4158, "num_input_tokens_seen": 46657472, "step": 34435 }, { "epoch": 1.1023622047244095, "grad_norm": 1.207358479499817, "learning_rate": 2.346825244477019e-06, "loss": 0.402, "num_input_tokens_seen": 46664016, "step": 34440 }, { "epoch": 1.102522245694898, "grad_norm": 1.3423686027526855, "learning_rate": 2.3426741418948545e-06, "loss": 0.3671, "num_input_tokens_seen": 46671008, "step": 34445 }, { "epoch": 1.1026822866653863, "grad_norm": 0.569341778755188, "learning_rate": 2.3385265333570715e-06, "loss": 0.5116, "num_input_tokens_seen": 46677632, "step": 34450 }, { "epoch": 1.1028423276358748, "grad_norm": 1.5854418277740479, "learning_rate": 2.334382419503278e-06, "loss": 0.3796, "num_input_tokens_seen": 46684288, "step": 34455 }, { "epoch": 1.1030023686063632, "grad_norm": 0.8382490277290344, "learning_rate": 2.3302418009725465e-06, "loss": 0.296, "num_input_tokens_seen": 46691264, "step": 34460 }, { "epoch": 1.1031624095768517, "grad_norm": 0.546868622303009, "learning_rate": 2.326104678403415e-06, "loss": 0.3731, "num_input_tokens_seen": 46698464, "step": 34465 }, { "epoch": 1.10332245054734, "grad_norm": 1.3369425535202026, "learning_rate": 2.321971052433883e-06, "loss": 0.4307, "num_input_tokens_seen": 46705408, "step": 34470 }, { "epoch": 1.1034824915178285, "grad_norm": 1.696777582168579, "learning_rate": 2.3178409237014004e-06, "loss": 0.81, "num_input_tokens_seen": 46711968, "step": 34475 }, { "epoch": 1.103642532488317, "grad_norm": 1.1149709224700928, "learning_rate": 2.313714292842889e-06, "loss": 0.5222, "num_input_tokens_seen": 46718928, "step": 34480 }, { "epoch": 1.1038025734588055, "grad_norm": 0.6703688502311707, "learning_rate": 2.309591160494734e-06, "loss": 0.4463, "num_input_tokens_seen": 46725264, "step": 34485 }, { "epoch": 1.103962614429294, "grad_norm": 1.197686791419983, "learning_rate": 2.305471527292763e-06, "loss": 0.3944, "num_input_tokens_seen": 46732256, "step": 34490 }, { "epoch": 1.1041226553997823, "grad_norm": 0.6857261657714844, "learning_rate": 2.3013553938722817e-06, "loss": 0.3989, "num_input_tokens_seen": 46738560, "step": 34495 }, { "epoch": 1.1042826963702708, "grad_norm": 2.1586551666259766, "learning_rate": 2.297242760868043e-06, "loss": 0.5258, "num_input_tokens_seen": 46745504, "step": 34500 }, { "epoch": 1.1044427373407593, "grad_norm": 0.9389007687568665, "learning_rate": 2.2931336289142735e-06, "loss": 0.4018, "num_input_tokens_seen": 46752128, "step": 34505 }, { "epoch": 1.1046027783112478, "grad_norm": 2.0289740562438965, "learning_rate": 2.289027998644655e-06, "loss": 0.4572, "num_input_tokens_seen": 46759136, "step": 34510 }, { "epoch": 1.104762819281736, "grad_norm": 1.0364410877227783, "learning_rate": 2.2849258706923228e-06, "loss": 0.4268, "num_input_tokens_seen": 46766272, "step": 34515 }, { "epoch": 1.1049228602522245, "grad_norm": 0.8788717985153198, "learning_rate": 2.2808272456898705e-06, "loss": 0.4341, "num_input_tokens_seen": 46773200, "step": 34520 }, { "epoch": 1.105082901222713, "grad_norm": 0.6072565317153931, "learning_rate": 2.2767321242693707e-06, "loss": 0.3962, "num_input_tokens_seen": 46779440, "step": 34525 }, { "epoch": 1.1052429421932015, "grad_norm": 0.9052245020866394, "learning_rate": 2.272640507062329e-06, "loss": 0.4597, "num_input_tokens_seen": 46786448, "step": 34530 }, { "epoch": 1.1054029831636898, "grad_norm": 1.5648554563522339, "learning_rate": 2.2685523946997382e-06, "loss": 0.4147, "num_input_tokens_seen": 46793024, "step": 34535 }, { "epoch": 1.1055630241341783, "grad_norm": 1.534589171409607, "learning_rate": 2.2644677878120245e-06, "loss": 0.5619, "num_input_tokens_seen": 46799888, "step": 34540 }, { "epoch": 1.1057230651046668, "grad_norm": 0.7190722823143005, "learning_rate": 2.2603866870290897e-06, "loss": 0.4018, "num_input_tokens_seen": 46806384, "step": 34545 }, { "epoch": 1.1058831060751553, "grad_norm": 1.3453158140182495, "learning_rate": 2.256309092980294e-06, "loss": 0.3648, "num_input_tokens_seen": 46813216, "step": 34550 }, { "epoch": 1.1060431470456438, "grad_norm": 1.8577580451965332, "learning_rate": 2.252235006294448e-06, "loss": 0.3771, "num_input_tokens_seen": 46819424, "step": 34555 }, { "epoch": 1.106203188016132, "grad_norm": 1.3479423522949219, "learning_rate": 2.2481644275998333e-06, "loss": 0.3358, "num_input_tokens_seen": 46826576, "step": 34560 }, { "epoch": 1.1063632289866205, "grad_norm": 1.334773063659668, "learning_rate": 2.2440973575241832e-06, "loss": 0.401, "num_input_tokens_seen": 46833248, "step": 34565 }, { "epoch": 1.106523269957109, "grad_norm": 0.9620363116264343, "learning_rate": 2.240033796694685e-06, "loss": 0.3314, "num_input_tokens_seen": 46839680, "step": 34570 }, { "epoch": 1.1066833109275975, "grad_norm": 1.909392237663269, "learning_rate": 2.235973745737999e-06, "loss": 0.5871, "num_input_tokens_seen": 46846304, "step": 34575 }, { "epoch": 1.1068433518980858, "grad_norm": 1.3146034479141235, "learning_rate": 2.2319172052802263e-06, "loss": 0.503, "num_input_tokens_seen": 46852976, "step": 34580 }, { "epoch": 1.1070033928685743, "grad_norm": 1.2200151681900024, "learning_rate": 2.2278641759469477e-06, "loss": 0.4285, "num_input_tokens_seen": 46859808, "step": 34585 }, { "epoch": 1.1071634338390628, "grad_norm": 0.7789304852485657, "learning_rate": 2.2238146583631825e-06, "loss": 0.2783, "num_input_tokens_seen": 46866512, "step": 34590 }, { "epoch": 1.1073234748095513, "grad_norm": 0.828667402267456, "learning_rate": 2.2197686531534256e-06, "loss": 0.4381, "num_input_tokens_seen": 46873232, "step": 34595 }, { "epoch": 1.1074835157800398, "grad_norm": 1.9009288549423218, "learning_rate": 2.2157261609416087e-06, "loss": 0.4403, "num_input_tokens_seen": 46879824, "step": 34600 }, { "epoch": 1.1074835157800398, "eval_loss": 0.44427207112312317, "eval_runtime": 503.4135, "eval_samples_per_second": 27.584, "eval_steps_per_second": 13.792, "num_input_tokens_seen": 46879824, "step": 34600 }, { "epoch": 1.107643556750528, "grad_norm": 0.7000159621238708, "learning_rate": 2.211687182351149e-06, "loss": 0.3105, "num_input_tokens_seen": 46886368, "step": 34605 }, { "epoch": 1.1078035977210166, "grad_norm": 1.1770694255828857, "learning_rate": 2.2076517180048993e-06, "loss": 0.4195, "num_input_tokens_seen": 46893424, "step": 34610 }, { "epoch": 1.107963638691505, "grad_norm": 1.1102718114852905, "learning_rate": 2.2036197685251834e-06, "loss": 0.4286, "num_input_tokens_seen": 46899936, "step": 34615 }, { "epoch": 1.1081236796619935, "grad_norm": 0.7865011096000671, "learning_rate": 2.199591334533771e-06, "loss": 0.4109, "num_input_tokens_seen": 46906512, "step": 34620 }, { "epoch": 1.1082837206324818, "grad_norm": 1.3009308576583862, "learning_rate": 2.1955664166519036e-06, "loss": 0.456, "num_input_tokens_seen": 46913456, "step": 34625 }, { "epoch": 1.1084437616029703, "grad_norm": 1.098808765411377, "learning_rate": 2.1915450155002793e-06, "loss": 0.4429, "num_input_tokens_seen": 46920016, "step": 34630 }, { "epoch": 1.1086038025734588, "grad_norm": 0.923337459564209, "learning_rate": 2.187527131699038e-06, "loss": 0.3806, "num_input_tokens_seen": 46926960, "step": 34635 }, { "epoch": 1.1087638435439473, "grad_norm": 1.0896118879318237, "learning_rate": 2.18351276586779e-06, "loss": 0.4861, "num_input_tokens_seen": 46934048, "step": 34640 }, { "epoch": 1.1089238845144358, "grad_norm": 0.6005719900131226, "learning_rate": 2.1795019186256092e-06, "loss": 0.4195, "num_input_tokens_seen": 46940928, "step": 34645 }, { "epoch": 1.109083925484924, "grad_norm": 0.7754842638969421, "learning_rate": 2.1754945905910094e-06, "loss": 0.4502, "num_input_tokens_seen": 46948032, "step": 34650 }, { "epoch": 1.1092439664554126, "grad_norm": 1.7801138162612915, "learning_rate": 2.171490782381977e-06, "loss": 0.281, "num_input_tokens_seen": 46955120, "step": 34655 }, { "epoch": 1.109404007425901, "grad_norm": 2.4017133712768555, "learning_rate": 2.1674904946159425e-06, "loss": 0.5004, "num_input_tokens_seen": 46962208, "step": 34660 }, { "epoch": 1.1095640483963896, "grad_norm": 2.1609091758728027, "learning_rate": 2.16349372790981e-06, "loss": 0.3959, "num_input_tokens_seen": 46969168, "step": 34665 }, { "epoch": 1.1097240893668778, "grad_norm": 0.880286455154419, "learning_rate": 2.159500482879928e-06, "loss": 0.3921, "num_input_tokens_seen": 46976592, "step": 34670 }, { "epoch": 1.1098841303373663, "grad_norm": 1.9662102460861206, "learning_rate": 2.155510760142096e-06, "loss": 0.3587, "num_input_tokens_seen": 46983536, "step": 34675 }, { "epoch": 1.1100441713078548, "grad_norm": 1.6522181034088135, "learning_rate": 2.151524560311588e-06, "loss": 0.4541, "num_input_tokens_seen": 46990704, "step": 34680 }, { "epoch": 1.1102042122783433, "grad_norm": 1.1401124000549316, "learning_rate": 2.147541884003129e-06, "loss": 0.4546, "num_input_tokens_seen": 46997616, "step": 34685 }, { "epoch": 1.1103642532488318, "grad_norm": 1.611288070678711, "learning_rate": 2.1435627318308895e-06, "loss": 0.3371, "num_input_tokens_seen": 47004544, "step": 34690 }, { "epoch": 1.11052429421932, "grad_norm": 0.8389620184898376, "learning_rate": 2.139587104408511e-06, "loss": 0.3335, "num_input_tokens_seen": 47011520, "step": 34695 }, { "epoch": 1.1106843351898086, "grad_norm": 1.434902548789978, "learning_rate": 2.1356150023490783e-06, "loss": 0.3982, "num_input_tokens_seen": 47018704, "step": 34700 }, { "epoch": 1.110844376160297, "grad_norm": 1.659942626953125, "learning_rate": 2.1316464262651464e-06, "loss": 0.377, "num_input_tokens_seen": 47025376, "step": 34705 }, { "epoch": 1.1110044171307856, "grad_norm": 2.2290565967559814, "learning_rate": 2.1276813767687224e-06, "loss": 0.4209, "num_input_tokens_seen": 47031872, "step": 34710 }, { "epoch": 1.1111644581012738, "grad_norm": 0.7516611814498901, "learning_rate": 2.123719854471254e-06, "loss": 0.4304, "num_input_tokens_seen": 47038464, "step": 34715 }, { "epoch": 1.1113244990717623, "grad_norm": 0.7761006355285645, "learning_rate": 2.119761859983668e-06, "loss": 0.5004, "num_input_tokens_seen": 47045296, "step": 34720 }, { "epoch": 1.1114845400422508, "grad_norm": 1.168711543083191, "learning_rate": 2.1158073939163386e-06, "loss": 0.4249, "num_input_tokens_seen": 47052368, "step": 34725 }, { "epoch": 1.1116445810127393, "grad_norm": 0.5725624561309814, "learning_rate": 2.111856456879088e-06, "loss": 0.3085, "num_input_tokens_seen": 47059104, "step": 34730 }, { "epoch": 1.1118046219832276, "grad_norm": 1.052215576171875, "learning_rate": 2.1079090494811993e-06, "loss": 0.403, "num_input_tokens_seen": 47066144, "step": 34735 }, { "epoch": 1.111964662953716, "grad_norm": 1.1027679443359375, "learning_rate": 2.103965172331418e-06, "loss": 0.3888, "num_input_tokens_seen": 47072848, "step": 34740 }, { "epoch": 1.1121247039242046, "grad_norm": 0.9674995541572571, "learning_rate": 2.100024826037933e-06, "loss": 0.3677, "num_input_tokens_seen": 47079376, "step": 34745 }, { "epoch": 1.112284744894693, "grad_norm": 0.8414745926856995, "learning_rate": 2.0960880112084027e-06, "loss": 0.4018, "num_input_tokens_seen": 47086160, "step": 34750 }, { "epoch": 1.1124447858651816, "grad_norm": 1.927930235862732, "learning_rate": 2.092154728449927e-06, "loss": 0.5169, "num_input_tokens_seen": 47092848, "step": 34755 }, { "epoch": 1.1126048268356699, "grad_norm": 0.6423677206039429, "learning_rate": 2.0882249783690687e-06, "loss": 0.3348, "num_input_tokens_seen": 47100064, "step": 34760 }, { "epoch": 1.1127648678061584, "grad_norm": 0.5691007971763611, "learning_rate": 2.084298761571851e-06, "loss": 0.2971, "num_input_tokens_seen": 47106928, "step": 34765 }, { "epoch": 1.1129249087766468, "grad_norm": 1.4348748922348022, "learning_rate": 2.080376078663737e-06, "loss": 0.4205, "num_input_tokens_seen": 47113776, "step": 34770 }, { "epoch": 1.1130849497471353, "grad_norm": 1.2111897468566895, "learning_rate": 2.0764569302496593e-06, "loss": 0.3909, "num_input_tokens_seen": 47120352, "step": 34775 }, { "epoch": 1.1132449907176236, "grad_norm": 0.5073950886726379, "learning_rate": 2.0725413169339957e-06, "loss": 0.4922, "num_input_tokens_seen": 47127280, "step": 34780 }, { "epoch": 1.1134050316881121, "grad_norm": 0.5270991921424866, "learning_rate": 2.068629239320588e-06, "loss": 0.4181, "num_input_tokens_seen": 47134160, "step": 34785 }, { "epoch": 1.1135650726586006, "grad_norm": 1.0801142454147339, "learning_rate": 2.064720698012726e-06, "loss": 0.3846, "num_input_tokens_seen": 47141184, "step": 34790 }, { "epoch": 1.113725113629089, "grad_norm": 0.8227710127830505, "learning_rate": 2.0608156936131522e-06, "loss": 0.433, "num_input_tokens_seen": 47148000, "step": 34795 }, { "epoch": 1.1138851545995774, "grad_norm": 1.8358877897262573, "learning_rate": 2.056914226724074e-06, "loss": 0.5335, "num_input_tokens_seen": 47155008, "step": 34800 }, { "epoch": 1.1138851545995774, "eval_loss": 0.44412797689437866, "eval_runtime": 503.4732, "eval_samples_per_second": 27.58, "eval_steps_per_second": 13.79, "num_input_tokens_seen": 47155008, "step": 34800 }, { "epoch": 1.1140451955700659, "grad_norm": 1.0741350650787354, "learning_rate": 2.0530162979471385e-06, "loss": 0.4323, "num_input_tokens_seen": 47161728, "step": 34805 }, { "epoch": 1.1142052365405544, "grad_norm": 1.0094916820526123, "learning_rate": 2.0491219078834667e-06, "loss": 0.2695, "num_input_tokens_seen": 47168816, "step": 34810 }, { "epoch": 1.1143652775110429, "grad_norm": 0.52582186460495, "learning_rate": 2.045231057133612e-06, "loss": 0.3744, "num_input_tokens_seen": 47175744, "step": 34815 }, { "epoch": 1.1145253184815314, "grad_norm": 0.8528621196746826, "learning_rate": 2.0413437462975944e-06, "loss": 0.5987, "num_input_tokens_seen": 47183120, "step": 34820 }, { "epoch": 1.1146853594520196, "grad_norm": 0.7121173739433289, "learning_rate": 2.0374599759748843e-06, "loss": 0.5458, "num_input_tokens_seen": 47189568, "step": 34825 }, { "epoch": 1.1148454004225081, "grad_norm": 0.5316007733345032, "learning_rate": 2.033579746764419e-06, "loss": 0.3773, "num_input_tokens_seen": 47196480, "step": 34830 }, { "epoch": 1.1150054413929966, "grad_norm": 1.6323024034500122, "learning_rate": 2.029703059264565e-06, "loss": 0.5254, "num_input_tokens_seen": 47203568, "step": 34835 }, { "epoch": 1.1151654823634851, "grad_norm": 0.40274471044540405, "learning_rate": 2.02582991407316e-06, "loss": 0.4523, "num_input_tokens_seen": 47210208, "step": 34840 }, { "epoch": 1.1153255233339734, "grad_norm": 1.0052783489227295, "learning_rate": 2.0219603117874992e-06, "loss": 0.5536, "num_input_tokens_seen": 47217008, "step": 34845 }, { "epoch": 1.1154855643044619, "grad_norm": 1.4743932485580444, "learning_rate": 2.0180942530043156e-06, "loss": 0.3934, "num_input_tokens_seen": 47224320, "step": 34850 }, { "epoch": 1.1156456052749504, "grad_norm": 1.4120233058929443, "learning_rate": 2.0142317383198107e-06, "loss": 0.5052, "num_input_tokens_seen": 47231168, "step": 34855 }, { "epoch": 1.1158056462454389, "grad_norm": 0.8287968039512634, "learning_rate": 2.0103727683296243e-06, "loss": 0.4812, "num_input_tokens_seen": 47238336, "step": 34860 }, { "epoch": 1.1159656872159274, "grad_norm": 1.3502825498580933, "learning_rate": 2.0065173436288636e-06, "loss": 0.4262, "num_input_tokens_seen": 47245056, "step": 34865 }, { "epoch": 1.1161257281864156, "grad_norm": 1.175682544708252, "learning_rate": 2.002665464812087e-06, "loss": 0.265, "num_input_tokens_seen": 47251920, "step": 34870 }, { "epoch": 1.1162857691569041, "grad_norm": 1.33686101436615, "learning_rate": 1.998817132473291e-06, "loss": 0.4841, "num_input_tokens_seen": 47258864, "step": 34875 }, { "epoch": 1.1164458101273926, "grad_norm": 1.8672327995300293, "learning_rate": 1.9949723472059507e-06, "loss": 0.4923, "num_input_tokens_seen": 47265488, "step": 34880 }, { "epoch": 1.1166058510978811, "grad_norm": 1.2391730546951294, "learning_rate": 1.9911311096029726e-06, "loss": 0.4888, "num_input_tokens_seen": 47272736, "step": 34885 }, { "epoch": 1.1167658920683694, "grad_norm": 1.256331443786621, "learning_rate": 1.9872934202567224e-06, "loss": 0.6001, "num_input_tokens_seen": 47279408, "step": 34890 }, { "epoch": 1.116925933038858, "grad_norm": 1.255893349647522, "learning_rate": 1.9834592797590257e-06, "loss": 0.3877, "num_input_tokens_seen": 47285984, "step": 34895 }, { "epoch": 1.1170859740093464, "grad_norm": 1.182491421699524, "learning_rate": 1.979628688701149e-06, "loss": 0.4752, "num_input_tokens_seen": 47293056, "step": 34900 }, { "epoch": 1.117246014979835, "grad_norm": 1.5064542293548584, "learning_rate": 1.9758016476738193e-06, "loss": 0.5071, "num_input_tokens_seen": 47299392, "step": 34905 }, { "epoch": 1.1174060559503234, "grad_norm": 1.0826629400253296, "learning_rate": 1.971978157267221e-06, "loss": 0.5191, "num_input_tokens_seen": 47305904, "step": 34910 }, { "epoch": 1.1175660969208117, "grad_norm": 0.7236371040344238, "learning_rate": 1.968158218070973e-06, "loss": 0.4889, "num_input_tokens_seen": 47312896, "step": 34915 }, { "epoch": 1.1177261378913002, "grad_norm": 0.9491059184074402, "learning_rate": 1.9643418306741682e-06, "loss": 0.4899, "num_input_tokens_seen": 47319456, "step": 34920 }, { "epoch": 1.1178861788617886, "grad_norm": 2.0076773166656494, "learning_rate": 1.9605289956653337e-06, "loss": 0.3722, "num_input_tokens_seen": 47326848, "step": 34925 }, { "epoch": 1.1180462198322771, "grad_norm": 0.6479073166847229, "learning_rate": 1.9567197136324626e-06, "loss": 0.3479, "num_input_tokens_seen": 47333520, "step": 34930 }, { "epoch": 1.1182062608027654, "grad_norm": 1.3479279279708862, "learning_rate": 1.9529139851629935e-06, "loss": 0.4185, "num_input_tokens_seen": 47340176, "step": 34935 }, { "epoch": 1.118366301773254, "grad_norm": 1.0278080701828003, "learning_rate": 1.949111810843812e-06, "loss": 0.3164, "num_input_tokens_seen": 47346448, "step": 34940 }, { "epoch": 1.1185263427437424, "grad_norm": 1.2169313430786133, "learning_rate": 1.9453131912612694e-06, "loss": 0.4929, "num_input_tokens_seen": 47353120, "step": 34945 }, { "epoch": 1.118686383714231, "grad_norm": 1.6444990634918213, "learning_rate": 1.941518127001149e-06, "loss": 0.4574, "num_input_tokens_seen": 47359664, "step": 34950 }, { "epoch": 1.1188464246847194, "grad_norm": 3.7008068561553955, "learning_rate": 1.9377266186487107e-06, "loss": 0.6098, "num_input_tokens_seen": 47366464, "step": 34955 }, { "epoch": 1.1190064656552077, "grad_norm": 1.0299123525619507, "learning_rate": 1.9339386667886483e-06, "loss": 0.3766, "num_input_tokens_seen": 47373120, "step": 34960 }, { "epoch": 1.1191665066256962, "grad_norm": 1.1237770318984985, "learning_rate": 1.9301542720051024e-06, "loss": 0.3686, "num_input_tokens_seen": 47379568, "step": 34965 }, { "epoch": 1.1193265475961847, "grad_norm": 1.2509024143218994, "learning_rate": 1.926373434881684e-06, "loss": 0.5336, "num_input_tokens_seen": 47386176, "step": 34970 }, { "epoch": 1.1194865885666732, "grad_norm": 1.3120558261871338, "learning_rate": 1.9225961560014468e-06, "loss": 0.4261, "num_input_tokens_seen": 47393168, "step": 34975 }, { "epoch": 1.1196466295371614, "grad_norm": 0.854357898235321, "learning_rate": 1.918822435946885e-06, "loss": 0.3858, "num_input_tokens_seen": 47399696, "step": 34980 }, { "epoch": 1.11980667050765, "grad_norm": 1.5461808443069458, "learning_rate": 1.915052275299961e-06, "loss": 0.3476, "num_input_tokens_seen": 47406576, "step": 34985 }, { "epoch": 1.1199667114781384, "grad_norm": 1.2374181747436523, "learning_rate": 1.9112856746420854e-06, "loss": 0.4154, "num_input_tokens_seen": 47413744, "step": 34990 }, { "epoch": 1.120126752448627, "grad_norm": 0.38258159160614014, "learning_rate": 1.907522634554104e-06, "loss": 0.3866, "num_input_tokens_seen": 47420480, "step": 34995 }, { "epoch": 1.1202867934191152, "grad_norm": 1.009314775466919, "learning_rate": 1.9037631556163337e-06, "loss": 0.339, "num_input_tokens_seen": 47426864, "step": 35000 }, { "epoch": 1.1202867934191152, "eval_loss": 0.44422513246536255, "eval_runtime": 503.9744, "eval_samples_per_second": 27.553, "eval_steps_per_second": 13.776, "num_input_tokens_seen": 47426864, "step": 35000 }, { "epoch": 1.1204468343896037, "grad_norm": 0.9076944589614868, "learning_rate": 1.9000072384085272e-06, "loss": 0.4248, "num_input_tokens_seen": 47434352, "step": 35005 }, { "epoch": 1.1206068753600922, "grad_norm": 1.3025217056274414, "learning_rate": 1.8962548835098987e-06, "loss": 0.5796, "num_input_tokens_seen": 47440976, "step": 35010 }, { "epoch": 1.1207669163305807, "grad_norm": 1.1091804504394531, "learning_rate": 1.8925060914991077e-06, "loss": 0.4193, "num_input_tokens_seen": 47447696, "step": 35015 }, { "epoch": 1.1209269573010692, "grad_norm": 1.2766536474227905, "learning_rate": 1.888760862954264e-06, "loss": 0.5568, "num_input_tokens_seen": 47454256, "step": 35020 }, { "epoch": 1.1210869982715574, "grad_norm": 0.3552403748035431, "learning_rate": 1.8850191984529309e-06, "loss": 0.3505, "num_input_tokens_seen": 47461136, "step": 35025 }, { "epoch": 1.121247039242046, "grad_norm": 1.5170117616653442, "learning_rate": 1.8812810985721186e-06, "loss": 0.5157, "num_input_tokens_seen": 47468256, "step": 35030 }, { "epoch": 1.1214070802125344, "grad_norm": 0.4583956003189087, "learning_rate": 1.8775465638882856e-06, "loss": 0.3216, "num_input_tokens_seen": 47474912, "step": 35035 }, { "epoch": 1.121567121183023, "grad_norm": 1.2665172815322876, "learning_rate": 1.8738155949773517e-06, "loss": 0.5653, "num_input_tokens_seen": 47481696, "step": 35040 }, { "epoch": 1.1217271621535112, "grad_norm": 2.0293407440185547, "learning_rate": 1.8700881924146707e-06, "loss": 0.5426, "num_input_tokens_seen": 47488448, "step": 35045 }, { "epoch": 1.1218872031239997, "grad_norm": 1.043534755706787, "learning_rate": 1.8663643567750577e-06, "loss": 0.5492, "num_input_tokens_seen": 47495360, "step": 35050 }, { "epoch": 1.1220472440944882, "grad_norm": 1.0207297801971436, "learning_rate": 1.8626440886327813e-06, "loss": 0.3965, "num_input_tokens_seen": 47502176, "step": 35055 }, { "epoch": 1.1222072850649767, "grad_norm": 0.699758768081665, "learning_rate": 1.8589273885615432e-06, "loss": 0.4692, "num_input_tokens_seen": 47509280, "step": 35060 }, { "epoch": 1.122367326035465, "grad_norm": 2.0344882011413574, "learning_rate": 1.8552142571345133e-06, "loss": 0.3912, "num_input_tokens_seen": 47515536, "step": 35065 }, { "epoch": 1.1225273670059535, "grad_norm": 1.4432473182678223, "learning_rate": 1.8515046949243025e-06, "loss": 0.5244, "num_input_tokens_seen": 47522192, "step": 35070 }, { "epoch": 1.122687407976442, "grad_norm": 1.007572054862976, "learning_rate": 1.8477987025029674e-06, "loss": 0.3879, "num_input_tokens_seen": 47528752, "step": 35075 }, { "epoch": 1.1228474489469304, "grad_norm": 0.38948744535446167, "learning_rate": 1.8440962804420232e-06, "loss": 0.4069, "num_input_tokens_seen": 47535904, "step": 35080 }, { "epoch": 1.123007489917419, "grad_norm": 1.782987356185913, "learning_rate": 1.8403974293124265e-06, "loss": 0.4488, "num_input_tokens_seen": 47543120, "step": 35085 }, { "epoch": 1.1231675308879072, "grad_norm": 0.6880477070808411, "learning_rate": 1.8367021496845854e-06, "loss": 0.661, "num_input_tokens_seen": 47549984, "step": 35090 }, { "epoch": 1.1233275718583957, "grad_norm": 1.1179394721984863, "learning_rate": 1.8330104421283662e-06, "loss": 0.3779, "num_input_tokens_seen": 47556960, "step": 35095 }, { "epoch": 1.1234876128288842, "grad_norm": 1.5800013542175293, "learning_rate": 1.8293223072130717e-06, "loss": 0.5357, "num_input_tokens_seen": 47563616, "step": 35100 }, { "epoch": 1.1236476537993727, "grad_norm": 0.8227736949920654, "learning_rate": 1.8256377455074525e-06, "loss": 0.432, "num_input_tokens_seen": 47570144, "step": 35105 }, { "epoch": 1.123807694769861, "grad_norm": 1.0085945129394531, "learning_rate": 1.8219567575797263e-06, "loss": 0.5324, "num_input_tokens_seen": 47576576, "step": 35110 }, { "epoch": 1.1239677357403495, "grad_norm": 0.9960586428642273, "learning_rate": 1.8182793439975365e-06, "loss": 0.3665, "num_input_tokens_seen": 47583584, "step": 35115 }, { "epoch": 1.124127776710838, "grad_norm": 1.481881856918335, "learning_rate": 1.8146055053279958e-06, "loss": 0.3975, "num_input_tokens_seen": 47590304, "step": 35120 }, { "epoch": 1.1242878176813265, "grad_norm": 0.917873740196228, "learning_rate": 1.8109352421376486e-06, "loss": 0.3903, "num_input_tokens_seen": 47596560, "step": 35125 }, { "epoch": 1.124447858651815, "grad_norm": 2.608555316925049, "learning_rate": 1.8072685549924972e-06, "loss": 0.4685, "num_input_tokens_seen": 47603424, "step": 35130 }, { "epoch": 1.1246078996223032, "grad_norm": 1.1373822689056396, "learning_rate": 1.8036054444579982e-06, "loss": 0.553, "num_input_tokens_seen": 47610160, "step": 35135 }, { "epoch": 1.1247679405927917, "grad_norm": 1.0006051063537598, "learning_rate": 1.7999459110990407e-06, "loss": 0.3964, "num_input_tokens_seen": 47617136, "step": 35140 }, { "epoch": 1.1249279815632802, "grad_norm": 0.7448509335517883, "learning_rate": 1.7962899554799712e-06, "loss": 0.3997, "num_input_tokens_seen": 47623536, "step": 35145 }, { "epoch": 1.1250880225337687, "grad_norm": 1.0814709663391113, "learning_rate": 1.7926375781645937e-06, "loss": 0.4816, "num_input_tokens_seen": 47630400, "step": 35150 }, { "epoch": 1.125248063504257, "grad_norm": 1.3344557285308838, "learning_rate": 1.7889887797161359e-06, "loss": 0.5574, "num_input_tokens_seen": 47637072, "step": 35155 }, { "epoch": 1.1254081044747455, "grad_norm": 1.1795766353607178, "learning_rate": 1.7853435606973028e-06, "loss": 0.6141, "num_input_tokens_seen": 47643760, "step": 35160 }, { "epoch": 1.125568145445234, "grad_norm": 0.8842289447784424, "learning_rate": 1.781701921670223e-06, "loss": 0.231, "num_input_tokens_seen": 47650768, "step": 35165 }, { "epoch": 1.1257281864157225, "grad_norm": 0.8406931757926941, "learning_rate": 1.7780638631964886e-06, "loss": 0.5353, "num_input_tokens_seen": 47657600, "step": 35170 }, { "epoch": 1.125888227386211, "grad_norm": 1.1351861953735352, "learning_rate": 1.7744293858371314e-06, "loss": 0.3524, "num_input_tokens_seen": 47664304, "step": 35175 }, { "epoch": 1.1260482683566992, "grad_norm": 1.493247628211975, "learning_rate": 1.770798490152631e-06, "loss": 0.4448, "num_input_tokens_seen": 47670992, "step": 35180 }, { "epoch": 1.1262083093271877, "grad_norm": 0.7848457098007202, "learning_rate": 1.767171176702917e-06, "loss": 0.32, "num_input_tokens_seen": 47677856, "step": 35185 }, { "epoch": 1.1263683502976762, "grad_norm": 1.168838620185852, "learning_rate": 1.7635474460473755e-06, "loss": 0.5033, "num_input_tokens_seen": 47684640, "step": 35190 }, { "epoch": 1.1265283912681647, "grad_norm": 1.454136848449707, "learning_rate": 1.7599272987448206e-06, "loss": 0.3219, "num_input_tokens_seen": 47691424, "step": 35195 }, { "epoch": 1.126688432238653, "grad_norm": 1.0712227821350098, "learning_rate": 1.7563107353535362e-06, "loss": 0.4668, "num_input_tokens_seen": 47698224, "step": 35200 }, { "epoch": 1.126688432238653, "eval_loss": 0.4440913498401642, "eval_runtime": 503.4138, "eval_samples_per_second": 27.584, "eval_steps_per_second": 13.792, "num_input_tokens_seen": 47698224, "step": 35200 }, { "epoch": 1.1268484732091415, "grad_norm": 0.7475598454475403, "learning_rate": 1.7526977564312263e-06, "loss": 0.4416, "num_input_tokens_seen": 47705296, "step": 35205 }, { "epoch": 1.12700851417963, "grad_norm": 1.3866052627563477, "learning_rate": 1.7490883625350701e-06, "loss": 0.496, "num_input_tokens_seen": 47711824, "step": 35210 }, { "epoch": 1.1271685551501185, "grad_norm": 3.6265406608581543, "learning_rate": 1.7454825542216807e-06, "loss": 0.381, "num_input_tokens_seen": 47718032, "step": 35215 }, { "epoch": 1.127328596120607, "grad_norm": 0.887493908405304, "learning_rate": 1.7418803320471105e-06, "loss": 0.4683, "num_input_tokens_seen": 47724544, "step": 35220 }, { "epoch": 1.1274886370910953, "grad_norm": 1.783775806427002, "learning_rate": 1.7382816965668737e-06, "loss": 0.5659, "num_input_tokens_seen": 47730992, "step": 35225 }, { "epoch": 1.1276486780615838, "grad_norm": 0.7765684723854065, "learning_rate": 1.7346866483359285e-06, "loss": 0.3436, "num_input_tokens_seen": 47737712, "step": 35230 }, { "epoch": 1.1278087190320722, "grad_norm": 0.9132009744644165, "learning_rate": 1.7310951879086657e-06, "loss": 0.701, "num_input_tokens_seen": 47744128, "step": 35235 }, { "epoch": 1.1279687600025607, "grad_norm": 1.1442065238952637, "learning_rate": 1.7275073158389471e-06, "loss": 0.4575, "num_input_tokens_seen": 47750768, "step": 35240 }, { "epoch": 1.128128800973049, "grad_norm": 0.6994971036911011, "learning_rate": 1.723923032680061e-06, "loss": 0.5049, "num_input_tokens_seen": 47757232, "step": 35245 }, { "epoch": 1.1282888419435375, "grad_norm": 0.8392181396484375, "learning_rate": 1.7203423389847428e-06, "loss": 0.4197, "num_input_tokens_seen": 47763696, "step": 35250 }, { "epoch": 1.128448882914026, "grad_norm": 1.5535463094711304, "learning_rate": 1.7167652353051928e-06, "loss": 0.3802, "num_input_tokens_seen": 47770944, "step": 35255 }, { "epoch": 1.1286089238845145, "grad_norm": 1.0678396224975586, "learning_rate": 1.7131917221930333e-06, "loss": 0.4545, "num_input_tokens_seen": 47777536, "step": 35260 }, { "epoch": 1.128768964855003, "grad_norm": 1.0072598457336426, "learning_rate": 1.7096218001993513e-06, "loss": 0.4191, "num_input_tokens_seen": 47784320, "step": 35265 }, { "epoch": 1.1289290058254913, "grad_norm": 0.856016993522644, "learning_rate": 1.706055469874676e-06, "loss": 0.4829, "num_input_tokens_seen": 47791488, "step": 35270 }, { "epoch": 1.1290890467959798, "grad_norm": 0.874214768409729, "learning_rate": 1.702492731768976e-06, "loss": 0.3312, "num_input_tokens_seen": 47798400, "step": 35275 }, { "epoch": 1.1292490877664683, "grad_norm": 0.5580058693885803, "learning_rate": 1.6989335864316724e-06, "loss": 0.4271, "num_input_tokens_seen": 47804848, "step": 35280 }, { "epoch": 1.1294091287369568, "grad_norm": 2.3586623668670654, "learning_rate": 1.6953780344116265e-06, "loss": 0.4941, "num_input_tokens_seen": 47811808, "step": 35285 }, { "epoch": 1.129569169707445, "grad_norm": 3.369934558868408, "learning_rate": 1.6918260762571497e-06, "loss": 0.5386, "num_input_tokens_seen": 47818416, "step": 35290 }, { "epoch": 1.1297292106779335, "grad_norm": 1.0021133422851562, "learning_rate": 1.6882777125160093e-06, "loss": 0.2658, "num_input_tokens_seen": 47825296, "step": 35295 }, { "epoch": 1.129889251648422, "grad_norm": 0.9986805319786072, "learning_rate": 1.6847329437353899e-06, "loss": 0.3376, "num_input_tokens_seen": 47832512, "step": 35300 }, { "epoch": 1.1300492926189105, "grad_norm": 0.5188847780227661, "learning_rate": 1.6811917704619511e-06, "loss": 0.342, "num_input_tokens_seen": 47838816, "step": 35305 }, { "epoch": 1.1302093335893988, "grad_norm": 0.623921275138855, "learning_rate": 1.67765419324179e-06, "loss": 0.3144, "num_input_tokens_seen": 47845440, "step": 35310 }, { "epoch": 1.1303693745598873, "grad_norm": 1.1298258304595947, "learning_rate": 1.6741202126204364e-06, "loss": 0.476, "num_input_tokens_seen": 47852144, "step": 35315 }, { "epoch": 1.1305294155303758, "grad_norm": 1.4030464887619019, "learning_rate": 1.6705898291428767e-06, "loss": 0.3869, "num_input_tokens_seen": 47859008, "step": 35320 }, { "epoch": 1.1306894565008643, "grad_norm": 1.2536754608154297, "learning_rate": 1.6670630433535395e-06, "loss": 0.5028, "num_input_tokens_seen": 47865904, "step": 35325 }, { "epoch": 1.1308494974713525, "grad_norm": 1.2922700643539429, "learning_rate": 1.6635398557962979e-06, "loss": 0.3769, "num_input_tokens_seen": 47872736, "step": 35330 }, { "epoch": 1.131009538441841, "grad_norm": 0.9555620551109314, "learning_rate": 1.660020267014481e-06, "loss": 0.291, "num_input_tokens_seen": 47879744, "step": 35335 }, { "epoch": 1.1311695794123295, "grad_norm": 1.3466657400131226, "learning_rate": 1.6565042775508438e-06, "loss": 0.5021, "num_input_tokens_seen": 47886256, "step": 35340 }, { "epoch": 1.131329620382818, "grad_norm": 0.4067636728286743, "learning_rate": 1.6529918879475997e-06, "loss": 0.4188, "num_input_tokens_seen": 47893344, "step": 35345 }, { "epoch": 1.1314896613533065, "grad_norm": 0.9124087691307068, "learning_rate": 1.6494830987464043e-06, "loss": 0.5212, "num_input_tokens_seen": 47900752, "step": 35350 }, { "epoch": 1.1316497023237948, "grad_norm": 0.937445342540741, "learning_rate": 1.6459779104883555e-06, "loss": 0.3016, "num_input_tokens_seen": 47907408, "step": 35355 }, { "epoch": 1.1318097432942833, "grad_norm": 0.7048527598381042, "learning_rate": 1.6424763237140013e-06, "loss": 0.3687, "num_input_tokens_seen": 47914432, "step": 35360 }, { "epoch": 1.1319697842647718, "grad_norm": 0.9132658839225769, "learning_rate": 1.6389783389633207e-06, "loss": 0.2875, "num_input_tokens_seen": 47921328, "step": 35365 }, { "epoch": 1.1321298252352603, "grad_norm": 1.709003210067749, "learning_rate": 1.6354839567757546e-06, "loss": 0.4081, "num_input_tokens_seen": 47928096, "step": 35370 }, { "epoch": 1.1322898662057486, "grad_norm": 0.6180753111839294, "learning_rate": 1.6319931776901831e-06, "loss": 0.2613, "num_input_tokens_seen": 47934608, "step": 35375 }, { "epoch": 1.132449907176237, "grad_norm": 1.4564663171768188, "learning_rate": 1.6285060022449229e-06, "loss": 0.4077, "num_input_tokens_seen": 47941216, "step": 35380 }, { "epoch": 1.1326099481467256, "grad_norm": 0.8443557024002075, "learning_rate": 1.6250224309777434e-06, "loss": 0.403, "num_input_tokens_seen": 47947584, "step": 35385 }, { "epoch": 1.132769989117214, "grad_norm": 1.261839509010315, "learning_rate": 1.6215424644258515e-06, "loss": 0.5556, "num_input_tokens_seen": 47954240, "step": 35390 }, { "epoch": 1.1329300300877025, "grad_norm": 0.8759411573410034, "learning_rate": 1.6180661031259036e-06, "loss": 0.3381, "num_input_tokens_seen": 47961168, "step": 35395 }, { "epoch": 1.1330900710581908, "grad_norm": 0.4766831696033478, "learning_rate": 1.614593347613999e-06, "loss": 0.5801, "num_input_tokens_seen": 47967840, "step": 35400 }, { "epoch": 1.1330900710581908, "eval_loss": 0.443901389837265, "eval_runtime": 503.4494, "eval_samples_per_second": 27.582, "eval_steps_per_second": 13.791, "num_input_tokens_seen": 47967840, "step": 35400 }, { "epoch": 1.1332501120286793, "grad_norm": 0.712995707988739, "learning_rate": 1.6111241984256758e-06, "loss": 0.2765, "num_input_tokens_seen": 47974240, "step": 35405 }, { "epoch": 1.1334101529991678, "grad_norm": 0.7422541379928589, "learning_rate": 1.6076586560959257e-06, "loss": 0.3628, "num_input_tokens_seen": 47980848, "step": 35410 }, { "epoch": 1.1335701939696563, "grad_norm": 0.7134740352630615, "learning_rate": 1.604196721159182e-06, "loss": 0.3506, "num_input_tokens_seen": 47987168, "step": 35415 }, { "epoch": 1.1337302349401446, "grad_norm": 0.4995236098766327, "learning_rate": 1.6007383941493092e-06, "loss": 0.5511, "num_input_tokens_seen": 47994256, "step": 35420 }, { "epoch": 1.133890275910633, "grad_norm": 1.6036150455474854, "learning_rate": 1.5972836755996285e-06, "loss": 0.4029, "num_input_tokens_seen": 48001232, "step": 35425 }, { "epoch": 1.1340503168811216, "grad_norm": 0.4971206784248352, "learning_rate": 1.5938325660429076e-06, "loss": 0.332, "num_input_tokens_seen": 48008080, "step": 35430 }, { "epoch": 1.13421035785161, "grad_norm": 0.8692333102226257, "learning_rate": 1.5903850660113378e-06, "loss": 0.4862, "num_input_tokens_seen": 48014608, "step": 35435 }, { "epoch": 1.1343703988220986, "grad_norm": 1.4519412517547607, "learning_rate": 1.5869411760365826e-06, "loss": 0.4978, "num_input_tokens_seen": 48021584, "step": 35440 }, { "epoch": 1.1345304397925868, "grad_norm": 1.7095757722854614, "learning_rate": 1.58350089664972e-06, "loss": 0.5343, "num_input_tokens_seen": 48028192, "step": 35445 }, { "epoch": 1.1346904807630753, "grad_norm": 1.4643077850341797, "learning_rate": 1.5800642283812865e-06, "loss": 0.4945, "num_input_tokens_seen": 48034848, "step": 35450 }, { "epoch": 1.1348505217335638, "grad_norm": 0.8499975800514221, "learning_rate": 1.5766311717612698e-06, "loss": 0.5879, "num_input_tokens_seen": 48041632, "step": 35455 }, { "epoch": 1.1350105627040523, "grad_norm": 1.1612708568572998, "learning_rate": 1.5732017273190818e-06, "loss": 0.5454, "num_input_tokens_seen": 48048160, "step": 35460 }, { "epoch": 1.1351706036745406, "grad_norm": 0.6855917572975159, "learning_rate": 1.5697758955835806e-06, "loss": 0.3856, "num_input_tokens_seen": 48054784, "step": 35465 }, { "epoch": 1.135330644645029, "grad_norm": 1.2659912109375, "learning_rate": 1.566353677083085e-06, "loss": 0.3749, "num_input_tokens_seen": 48062144, "step": 35470 }, { "epoch": 1.1354906856155176, "grad_norm": 2.3552956581115723, "learning_rate": 1.562935072345334e-06, "loss": 0.4732, "num_input_tokens_seen": 48069712, "step": 35475 }, { "epoch": 1.135650726586006, "grad_norm": 1.0054748058319092, "learning_rate": 1.5595200818975281e-06, "loss": 0.4137, "num_input_tokens_seen": 48076336, "step": 35480 }, { "epoch": 1.1358107675564946, "grad_norm": 1.0223119258880615, "learning_rate": 1.5561087062662905e-06, "loss": 0.4067, "num_input_tokens_seen": 48083056, "step": 35485 }, { "epoch": 1.1359708085269828, "grad_norm": 1.0202490091323853, "learning_rate": 1.5527009459777087e-06, "loss": 0.4526, "num_input_tokens_seen": 48089440, "step": 35490 }, { "epoch": 1.1361308494974713, "grad_norm": 0.6036620140075684, "learning_rate": 1.5492968015572984e-06, "loss": 0.4871, "num_input_tokens_seen": 48096336, "step": 35495 }, { "epoch": 1.1362908904679598, "grad_norm": 0.5704581141471863, "learning_rate": 1.5458962735300203e-06, "loss": 0.3998, "num_input_tokens_seen": 48103312, "step": 35500 }, { "epoch": 1.1364509314384483, "grad_norm": 1.2274450063705444, "learning_rate": 1.54249936242028e-06, "loss": 0.4546, "num_input_tokens_seen": 48110304, "step": 35505 }, { "epoch": 1.1366109724089366, "grad_norm": 0.5254071950912476, "learning_rate": 1.5391060687519222e-06, "loss": 0.2433, "num_input_tokens_seen": 48116864, "step": 35510 }, { "epoch": 1.136771013379425, "grad_norm": 0.8729111552238464, "learning_rate": 1.5357163930482367e-06, "loss": 0.3862, "num_input_tokens_seen": 48123632, "step": 35515 }, { "epoch": 1.1369310543499136, "grad_norm": 0.8309873342514038, "learning_rate": 1.532330335831955e-06, "loss": 0.3929, "num_input_tokens_seen": 48130848, "step": 35520 }, { "epoch": 1.137091095320402, "grad_norm": 1.3436936140060425, "learning_rate": 1.5289478976252491e-06, "loss": 0.2629, "num_input_tokens_seen": 48137296, "step": 35525 }, { "epoch": 1.1372511362908906, "grad_norm": 1.6767174005508423, "learning_rate": 1.5255690789497345e-06, "loss": 0.4875, "num_input_tokens_seen": 48143840, "step": 35530 }, { "epoch": 1.1374111772613789, "grad_norm": 0.917529284954071, "learning_rate": 1.5221938803264641e-06, "loss": 0.3372, "num_input_tokens_seen": 48150272, "step": 35535 }, { "epoch": 1.1375712182318674, "grad_norm": 0.6862383484840393, "learning_rate": 1.518822302275938e-06, "loss": 0.3031, "num_input_tokens_seen": 48157024, "step": 35540 }, { "epoch": 1.1377312592023558, "grad_norm": 1.0409221649169922, "learning_rate": 1.5154543453180958e-06, "loss": 0.3253, "num_input_tokens_seen": 48163792, "step": 35545 }, { "epoch": 1.1378913001728443, "grad_norm": 2.5842859745025635, "learning_rate": 1.5120900099723167e-06, "loss": 0.4663, "num_input_tokens_seen": 48170496, "step": 35550 }, { "epoch": 1.1380513411433326, "grad_norm": 0.38596782088279724, "learning_rate": 1.5087292967574273e-06, "loss": 0.4298, "num_input_tokens_seen": 48177488, "step": 35555 }, { "epoch": 1.1382113821138211, "grad_norm": 0.9488540887832642, "learning_rate": 1.5053722061916908e-06, "loss": 0.412, "num_input_tokens_seen": 48183968, "step": 35560 }, { "epoch": 1.1383714230843096, "grad_norm": 1.119920253753662, "learning_rate": 1.5020187387928124e-06, "loss": 0.5312, "num_input_tokens_seen": 48190608, "step": 35565 }, { "epoch": 1.138531464054798, "grad_norm": 1.8783687353134155, "learning_rate": 1.4986688950779343e-06, "loss": 0.5222, "num_input_tokens_seen": 48197456, "step": 35570 }, { "epoch": 1.1386915050252864, "grad_norm": 0.961031973361969, "learning_rate": 1.495322675563654e-06, "loss": 0.3963, "num_input_tokens_seen": 48204736, "step": 35575 }, { "epoch": 1.1388515459957749, "grad_norm": 3.337756633758545, "learning_rate": 1.4919800807659922e-06, "loss": 0.4136, "num_input_tokens_seen": 48212144, "step": 35580 }, { "epoch": 1.1390115869662634, "grad_norm": 0.9437316060066223, "learning_rate": 1.4886411112004255e-06, "loss": 0.5235, "num_input_tokens_seen": 48219072, "step": 35585 }, { "epoch": 1.1391716279367519, "grad_norm": 0.5794472694396973, "learning_rate": 1.4853057673818588e-06, "loss": 0.2784, "num_input_tokens_seen": 48225696, "step": 35590 }, { "epoch": 1.1393316689072401, "grad_norm": 0.8692311644554138, "learning_rate": 1.481974049824647e-06, "loss": 0.3294, "num_input_tokens_seen": 48232496, "step": 35595 }, { "epoch": 1.1394917098777286, "grad_norm": 1.2062910795211792, "learning_rate": 1.4786459590425849e-06, "loss": 0.4932, "num_input_tokens_seen": 48239792, "step": 35600 }, { "epoch": 1.1394917098777286, "eval_loss": 0.444240003824234, "eval_runtime": 503.3772, "eval_samples_per_second": 27.586, "eval_steps_per_second": 13.793, "num_input_tokens_seen": 48239792, "step": 35600 }, { "epoch": 1.1396517508482171, "grad_norm": 1.2892450094223022, "learning_rate": 1.4753214955489036e-06, "loss": 0.4722, "num_input_tokens_seen": 48246432, "step": 35605 }, { "epoch": 1.1398117918187056, "grad_norm": 1.63887357711792, "learning_rate": 1.4720006598562737e-06, "loss": 0.372, "num_input_tokens_seen": 48253616, "step": 35610 }, { "epoch": 1.1399718327891941, "grad_norm": 0.8537741899490356, "learning_rate": 1.4686834524768185e-06, "loss": 0.3549, "num_input_tokens_seen": 48260080, "step": 35615 }, { "epoch": 1.1401318737596824, "grad_norm": 1.361542820930481, "learning_rate": 1.4653698739220844e-06, "loss": 0.6128, "num_input_tokens_seen": 48266800, "step": 35620 }, { "epoch": 1.1402919147301709, "grad_norm": 0.6834345459938049, "learning_rate": 1.4620599247030715e-06, "loss": 0.2815, "num_input_tokens_seen": 48273760, "step": 35625 }, { "epoch": 1.1404519557006594, "grad_norm": 0.3771156966686249, "learning_rate": 1.4587536053302125e-06, "loss": 0.3409, "num_input_tokens_seen": 48280784, "step": 35630 }, { "epoch": 1.1406119966711479, "grad_norm": 1.554452657699585, "learning_rate": 1.4554509163133862e-06, "loss": 0.4619, "num_input_tokens_seen": 48287552, "step": 35635 }, { "epoch": 1.1407720376416361, "grad_norm": 1.389455795288086, "learning_rate": 1.4521518581619098e-06, "loss": 0.5391, "num_input_tokens_seen": 48295008, "step": 35640 }, { "epoch": 1.1409320786121246, "grad_norm": 0.9101411700248718, "learning_rate": 1.4488564313845348e-06, "loss": 0.3999, "num_input_tokens_seen": 48301888, "step": 35645 }, { "epoch": 1.1410921195826131, "grad_norm": 1.9503648281097412, "learning_rate": 1.4455646364894603e-06, "loss": 0.7306, "num_input_tokens_seen": 48309472, "step": 35650 }, { "epoch": 1.1412521605531016, "grad_norm": 1.9261655807495117, "learning_rate": 1.4422764739843247e-06, "loss": 0.4511, "num_input_tokens_seen": 48316592, "step": 35655 }, { "epoch": 1.1414122015235901, "grad_norm": 0.8138653039932251, "learning_rate": 1.4389919443762e-06, "loss": 0.4457, "num_input_tokens_seen": 48323456, "step": 35660 }, { "epoch": 1.1415722424940784, "grad_norm": 1.1467020511627197, "learning_rate": 1.4357110481716063e-06, "loss": 0.3778, "num_input_tokens_seen": 48330688, "step": 35665 }, { "epoch": 1.141732283464567, "grad_norm": 0.7029937505722046, "learning_rate": 1.4324337858764941e-06, "loss": 0.33, "num_input_tokens_seen": 48337536, "step": 35670 }, { "epoch": 1.1418923244350554, "grad_norm": 0.45746392011642456, "learning_rate": 1.4291601579962622e-06, "loss": 0.2896, "num_input_tokens_seen": 48344336, "step": 35675 }, { "epoch": 1.142052365405544, "grad_norm": 0.7672642469406128, "learning_rate": 1.42589016503574e-06, "loss": 0.6401, "num_input_tokens_seen": 48351104, "step": 35680 }, { "epoch": 1.1422124063760322, "grad_norm": 1.839684009552002, "learning_rate": 1.4226238074992099e-06, "loss": 0.5622, "num_input_tokens_seen": 48357904, "step": 35685 }, { "epoch": 1.1423724473465207, "grad_norm": 0.887267529964447, "learning_rate": 1.4193610858903778e-06, "loss": 0.4219, "num_input_tokens_seen": 48364752, "step": 35690 }, { "epoch": 1.1425324883170092, "grad_norm": 2.1704015731811523, "learning_rate": 1.416102000712402e-06, "loss": 0.4079, "num_input_tokens_seen": 48371568, "step": 35695 }, { "epoch": 1.1426925292874976, "grad_norm": 0.7952640652656555, "learning_rate": 1.4128465524678668e-06, "loss": 0.4024, "num_input_tokens_seen": 48378576, "step": 35700 }, { "epoch": 1.1428525702579861, "grad_norm": 0.735280454158783, "learning_rate": 1.4095947416588124e-06, "loss": 0.3251, "num_input_tokens_seen": 48384928, "step": 35705 }, { "epoch": 1.1430126112284744, "grad_norm": 2.309680223464966, "learning_rate": 1.4063465687866983e-06, "loss": 0.5394, "num_input_tokens_seen": 48392224, "step": 35710 }, { "epoch": 1.143172652198963, "grad_norm": 0.6865726709365845, "learning_rate": 1.4031020343524438e-06, "loss": 0.6077, "num_input_tokens_seen": 48398928, "step": 35715 }, { "epoch": 1.1433326931694514, "grad_norm": 1.1193814277648926, "learning_rate": 1.3998611388563926e-06, "loss": 0.293, "num_input_tokens_seen": 48405824, "step": 35720 }, { "epoch": 1.14349273413994, "grad_norm": 0.562782347202301, "learning_rate": 1.3966238827983314e-06, "loss": 0.3737, "num_input_tokens_seen": 48412912, "step": 35725 }, { "epoch": 1.1436527751104282, "grad_norm": 0.9278671145439148, "learning_rate": 1.393390266677483e-06, "loss": 0.3703, "num_input_tokens_seen": 48419472, "step": 35730 }, { "epoch": 1.1438128160809167, "grad_norm": 4.076487064361572, "learning_rate": 1.3901602909925204e-06, "loss": 0.6397, "num_input_tokens_seen": 48426112, "step": 35735 }, { "epoch": 1.1439728570514052, "grad_norm": 0.7210180163383484, "learning_rate": 1.3869339562415373e-06, "loss": 0.5135, "num_input_tokens_seen": 48432736, "step": 35740 }, { "epoch": 1.1441328980218937, "grad_norm": 0.8649200201034546, "learning_rate": 1.38371126292208e-06, "loss": 0.4799, "num_input_tokens_seen": 48439616, "step": 35745 }, { "epoch": 1.1442929389923822, "grad_norm": 1.4926196336746216, "learning_rate": 1.3804922115311286e-06, "loss": 0.534, "num_input_tokens_seen": 48446240, "step": 35750 }, { "epoch": 1.1444529799628704, "grad_norm": 1.5546622276306152, "learning_rate": 1.3772768025650945e-06, "loss": 0.3952, "num_input_tokens_seen": 48453648, "step": 35755 }, { "epoch": 1.144613020933359, "grad_norm": 0.8813854455947876, "learning_rate": 1.3740650365198448e-06, "loss": 0.324, "num_input_tokens_seen": 48460336, "step": 35760 }, { "epoch": 1.1447730619038474, "grad_norm": 1.4806321859359741, "learning_rate": 1.3708569138906612e-06, "loss": 0.4503, "num_input_tokens_seen": 48467008, "step": 35765 }, { "epoch": 1.144933102874336, "grad_norm": 0.950002133846283, "learning_rate": 1.367652435172287e-06, "loss": 0.3554, "num_input_tokens_seen": 48473664, "step": 35770 }, { "epoch": 1.1450931438448242, "grad_norm": 1.0287209749221802, "learning_rate": 1.364451600858893e-06, "loss": 0.4487, "num_input_tokens_seen": 48480656, "step": 35775 }, { "epoch": 1.1452531848153127, "grad_norm": 2.555264472961426, "learning_rate": 1.3612544114440823e-06, "loss": 0.4168, "num_input_tokens_seen": 48487536, "step": 35780 }, { "epoch": 1.1454132257858012, "grad_norm": 1.2889834642410278, "learning_rate": 1.3580608674209072e-06, "loss": 0.3697, "num_input_tokens_seen": 48494416, "step": 35785 }, { "epoch": 1.1455732667562897, "grad_norm": 0.9966676235198975, "learning_rate": 1.3548709692818434e-06, "loss": 0.5089, "num_input_tokens_seen": 48501520, "step": 35790 }, { "epoch": 1.1457333077267782, "grad_norm": 1.44319748878479, "learning_rate": 1.3516847175188223e-06, "loss": 0.3634, "num_input_tokens_seen": 48508432, "step": 35795 }, { "epoch": 1.1458933486972664, "grad_norm": 2.0764338970184326, "learning_rate": 1.348502112623204e-06, "loss": 0.4009, "num_input_tokens_seen": 48514752, "step": 35800 }, { "epoch": 1.1458933486972664, "eval_loss": 0.444074809551239, "eval_runtime": 503.4822, "eval_samples_per_second": 27.58, "eval_steps_per_second": 13.79, "num_input_tokens_seen": 48514752, "step": 35800 }, { "epoch": 1.146053389667755, "grad_norm": 0.6392542719841003, "learning_rate": 1.3453231550857787e-06, "loss": 0.2582, "num_input_tokens_seen": 48521664, "step": 35805 }, { "epoch": 1.1462134306382434, "grad_norm": 1.230003833770752, "learning_rate": 1.3421478453967878e-06, "loss": 0.6495, "num_input_tokens_seen": 48528480, "step": 35810 }, { "epoch": 1.146373471608732, "grad_norm": 0.5038407444953918, "learning_rate": 1.3389761840459065e-06, "loss": 0.3085, "num_input_tokens_seen": 48534784, "step": 35815 }, { "epoch": 1.1465335125792202, "grad_norm": 1.16440749168396, "learning_rate": 1.3358081715222376e-06, "loss": 0.4898, "num_input_tokens_seen": 48541248, "step": 35820 }, { "epoch": 1.1466935535497087, "grad_norm": 1.6115550994873047, "learning_rate": 1.3326438083143295e-06, "loss": 0.5754, "num_input_tokens_seen": 48547936, "step": 35825 }, { "epoch": 1.1468535945201972, "grad_norm": 0.9154455661773682, "learning_rate": 1.3294830949101723e-06, "loss": 0.3731, "num_input_tokens_seen": 48554576, "step": 35830 }, { "epoch": 1.1470136354906857, "grad_norm": 0.9110314249992371, "learning_rate": 1.3263260317971815e-06, "loss": 0.3788, "num_input_tokens_seen": 48561360, "step": 35835 }, { "epoch": 1.1471736764611742, "grad_norm": 0.44943535327911377, "learning_rate": 1.3231726194622208e-06, "loss": 0.3968, "num_input_tokens_seen": 48568016, "step": 35840 }, { "epoch": 1.1473337174316625, "grad_norm": 1.0988537073135376, "learning_rate": 1.3200228583915814e-06, "loss": 0.4567, "num_input_tokens_seen": 48574672, "step": 35845 }, { "epoch": 1.147493758402151, "grad_norm": 1.4603229761123657, "learning_rate": 1.3168767490709971e-06, "loss": 0.475, "num_input_tokens_seen": 48581216, "step": 35850 }, { "epoch": 1.1476537993726394, "grad_norm": 1.3993102312088013, "learning_rate": 1.3137342919856437e-06, "loss": 0.5105, "num_input_tokens_seen": 48587792, "step": 35855 }, { "epoch": 1.1478138403431277, "grad_norm": 0.8453565835952759, "learning_rate": 1.310595487620117e-06, "loss": 0.354, "num_input_tokens_seen": 48594656, "step": 35860 }, { "epoch": 1.1479738813136162, "grad_norm": 1.3362709283828735, "learning_rate": 1.3074603364584715e-06, "loss": 0.4959, "num_input_tokens_seen": 48601696, "step": 35865 }, { "epoch": 1.1481339222841047, "grad_norm": 0.7668452858924866, "learning_rate": 1.3043288389841758e-06, "loss": 0.5616, "num_input_tokens_seen": 48608320, "step": 35870 }, { "epoch": 1.1482939632545932, "grad_norm": 0.9999887943267822, "learning_rate": 1.3012009956801546e-06, "loss": 0.5111, "num_input_tokens_seen": 48615184, "step": 35875 }, { "epoch": 1.1484540042250817, "grad_norm": 0.7032452821731567, "learning_rate": 1.2980768070287586e-06, "loss": 0.6154, "num_input_tokens_seen": 48622144, "step": 35880 }, { "epoch": 1.14861404519557, "grad_norm": 1.3020038604736328, "learning_rate": 1.2949562735117716e-06, "loss": 0.2938, "num_input_tokens_seen": 48628576, "step": 35885 }, { "epoch": 1.1487740861660585, "grad_norm": 1.0228379964828491, "learning_rate": 1.291839395610428e-06, "loss": 0.4529, "num_input_tokens_seen": 48635136, "step": 35890 }, { "epoch": 1.148934127136547, "grad_norm": 0.8057577013969421, "learning_rate": 1.2887261738053852e-06, "loss": 0.4034, "num_input_tokens_seen": 48641744, "step": 35895 }, { "epoch": 1.1490941681070355, "grad_norm": 1.763364553451538, "learning_rate": 1.2856166085767396e-06, "loss": 0.4691, "num_input_tokens_seen": 48648560, "step": 35900 }, { "epoch": 1.1492542090775237, "grad_norm": 0.633115291595459, "learning_rate": 1.2825107004040272e-06, "loss": 0.2636, "num_input_tokens_seen": 48655504, "step": 35905 }, { "epoch": 1.1494142500480122, "grad_norm": 1.8808486461639404, "learning_rate": 1.2794084497662146e-06, "loss": 0.4579, "num_input_tokens_seen": 48662400, "step": 35910 }, { "epoch": 1.1495742910185007, "grad_norm": 0.7640665769577026, "learning_rate": 1.276309857141711e-06, "loss": 0.3712, "num_input_tokens_seen": 48669264, "step": 35915 }, { "epoch": 1.1497343319889892, "grad_norm": 0.9363978505134583, "learning_rate": 1.273214923008359e-06, "loss": 0.4583, "num_input_tokens_seen": 48675920, "step": 35920 }, { "epoch": 1.1498943729594777, "grad_norm": 2.142697334289551, "learning_rate": 1.2701236478434352e-06, "loss": 0.468, "num_input_tokens_seen": 48682576, "step": 35925 }, { "epoch": 1.150054413929966, "grad_norm": 0.940480649471283, "learning_rate": 1.2670360321236502e-06, "loss": 0.4184, "num_input_tokens_seen": 48689024, "step": 35930 }, { "epoch": 1.1502144549004545, "grad_norm": 0.41492268443107605, "learning_rate": 1.2639520763251617e-06, "loss": 0.3152, "num_input_tokens_seen": 48695920, "step": 35935 }, { "epoch": 1.150374495870943, "grad_norm": 0.7603142857551575, "learning_rate": 1.2608717809235448e-06, "loss": 0.278, "num_input_tokens_seen": 48702464, "step": 35940 }, { "epoch": 1.1505345368414315, "grad_norm": 1.4636454582214355, "learning_rate": 1.2577951463938282e-06, "loss": 0.5875, "num_input_tokens_seen": 48708848, "step": 35945 }, { "epoch": 1.1506945778119197, "grad_norm": 1.155890703201294, "learning_rate": 1.2547221732104569e-06, "loss": 0.4267, "num_input_tokens_seen": 48715392, "step": 35950 }, { "epoch": 1.1508546187824082, "grad_norm": 1.4045783281326294, "learning_rate": 1.25165286184733e-06, "loss": 0.4921, "num_input_tokens_seen": 48722176, "step": 35955 }, { "epoch": 1.1510146597528967, "grad_norm": 1.1316436529159546, "learning_rate": 1.248587212777777e-06, "loss": 0.3894, "num_input_tokens_seen": 48728832, "step": 35960 }, { "epoch": 1.1511747007233852, "grad_norm": 1.0797176361083984, "learning_rate": 1.2455252264745532e-06, "loss": 0.3783, "num_input_tokens_seen": 48735712, "step": 35965 }, { "epoch": 1.1513347416938737, "grad_norm": 0.4843481481075287, "learning_rate": 1.2424669034098528e-06, "loss": 0.3661, "num_input_tokens_seen": 48742464, "step": 35970 }, { "epoch": 1.151494782664362, "grad_norm": 1.5303057432174683, "learning_rate": 1.2394122440553185e-06, "loss": 0.414, "num_input_tokens_seen": 48749136, "step": 35975 }, { "epoch": 1.1516548236348505, "grad_norm": 0.9672002792358398, "learning_rate": 1.2363612488820037e-06, "loss": 0.4773, "num_input_tokens_seen": 48755840, "step": 35980 }, { "epoch": 1.151814864605339, "grad_norm": 2.9928295612335205, "learning_rate": 1.2333139183604208e-06, "loss": 0.4138, "num_input_tokens_seen": 48762592, "step": 35985 }, { "epoch": 1.1519749055758275, "grad_norm": 1.159659504890442, "learning_rate": 1.2302702529604998e-06, "loss": 0.5411, "num_input_tokens_seen": 48769584, "step": 35990 }, { "epoch": 1.1521349465463158, "grad_norm": 0.648472011089325, "learning_rate": 1.227230253151615e-06, "loss": 0.3127, "num_input_tokens_seen": 48776592, "step": 35995 }, { "epoch": 1.1522949875168043, "grad_norm": 0.5720286965370178, "learning_rate": 1.2241939194025748e-06, "loss": 0.6396, "num_input_tokens_seen": 48783136, "step": 36000 }, { "epoch": 1.1522949875168043, "eval_loss": 0.4441034495830536, "eval_runtime": 503.4342, "eval_samples_per_second": 27.583, "eval_steps_per_second": 13.791, "num_input_tokens_seen": 48783136, "step": 36000 }, { "epoch": 1.1524550284872928, "grad_norm": 0.9092857837677002, "learning_rate": 1.2211612521816156e-06, "loss": 0.6046, "num_input_tokens_seen": 48789776, "step": 36005 }, { "epoch": 1.1526150694577812, "grad_norm": 0.9541423320770264, "learning_rate": 1.2181322519564137e-06, "loss": 0.5716, "num_input_tokens_seen": 48796720, "step": 36010 }, { "epoch": 1.1527751104282697, "grad_norm": 1.1634656190872192, "learning_rate": 1.2151069191940839e-06, "loss": 0.3367, "num_input_tokens_seen": 48804208, "step": 36015 }, { "epoch": 1.152935151398758, "grad_norm": 3.0665197372436523, "learning_rate": 1.2120852543611644e-06, "loss": 0.5677, "num_input_tokens_seen": 48810848, "step": 36020 }, { "epoch": 1.1530951923692465, "grad_norm": 0.5671855807304382, "learning_rate": 1.2090672579236379e-06, "loss": 0.2976, "num_input_tokens_seen": 48817440, "step": 36025 }, { "epoch": 1.153255233339735, "grad_norm": 1.5510499477386475, "learning_rate": 1.2060529303469126e-06, "loss": 0.4191, "num_input_tokens_seen": 48824064, "step": 36030 }, { "epoch": 1.1534152743102235, "grad_norm": 0.5580193996429443, "learning_rate": 1.2030422720958445e-06, "loss": 0.3224, "num_input_tokens_seen": 48830976, "step": 36035 }, { "epoch": 1.1535753152807118, "grad_norm": 1.3799303770065308, "learning_rate": 1.200035283634704e-06, "loss": 0.4763, "num_input_tokens_seen": 48837824, "step": 36040 }, { "epoch": 1.1537353562512003, "grad_norm": 0.6422731876373291, "learning_rate": 1.1970319654272144e-06, "loss": 0.473, "num_input_tokens_seen": 48844336, "step": 36045 }, { "epoch": 1.1538953972216888, "grad_norm": 0.8464412689208984, "learning_rate": 1.1940323179365192e-06, "loss": 0.3931, "num_input_tokens_seen": 48851696, "step": 36050 }, { "epoch": 1.1540554381921773, "grad_norm": 1.1159433126449585, "learning_rate": 1.1910363416252095e-06, "loss": 0.4491, "num_input_tokens_seen": 48858128, "step": 36055 }, { "epoch": 1.1542154791626658, "grad_norm": 1.0083461999893188, "learning_rate": 1.1880440369552964e-06, "loss": 0.3734, "num_input_tokens_seen": 48864928, "step": 36060 }, { "epoch": 1.154375520133154, "grad_norm": 1.0009349584579468, "learning_rate": 1.1850554043882328e-06, "loss": 0.5981, "num_input_tokens_seen": 48871808, "step": 36065 }, { "epoch": 1.1545355611036425, "grad_norm": 2.204172134399414, "learning_rate": 1.1820704443849028e-06, "loss": 0.4065, "num_input_tokens_seen": 48878768, "step": 36070 }, { "epoch": 1.154695602074131, "grad_norm": 1.391878604888916, "learning_rate": 1.1790891574056219e-06, "loss": 0.6317, "num_input_tokens_seen": 48885232, "step": 36075 }, { "epoch": 1.1548556430446195, "grad_norm": 0.9184126853942871, "learning_rate": 1.1761115439101523e-06, "loss": 0.3908, "num_input_tokens_seen": 48891728, "step": 36080 }, { "epoch": 1.1550156840151078, "grad_norm": 1.3596080541610718, "learning_rate": 1.1731376043576659e-06, "loss": 0.5944, "num_input_tokens_seen": 48898176, "step": 36085 }, { "epoch": 1.1551757249855963, "grad_norm": 0.8022765517234802, "learning_rate": 1.1701673392067875e-06, "loss": 0.3519, "num_input_tokens_seen": 48905008, "step": 36090 }, { "epoch": 1.1553357659560848, "grad_norm": 1.0108840465545654, "learning_rate": 1.1672007489155757e-06, "loss": 0.4347, "num_input_tokens_seen": 48911808, "step": 36095 }, { "epoch": 1.1554958069265733, "grad_norm": 0.45102304220199585, "learning_rate": 1.164237833941506e-06, "loss": 0.2251, "num_input_tokens_seen": 48918432, "step": 36100 }, { "epoch": 1.1556558478970618, "grad_norm": 0.8518909811973572, "learning_rate": 1.1612785947415022e-06, "loss": 0.3162, "num_input_tokens_seen": 48924912, "step": 36105 }, { "epoch": 1.15581588886755, "grad_norm": 1.0712100267410278, "learning_rate": 1.1583230317719185e-06, "loss": 0.5085, "num_input_tokens_seen": 48931488, "step": 36110 }, { "epoch": 1.1559759298380385, "grad_norm": 0.9312910437583923, "learning_rate": 1.1553711454885318e-06, "loss": 0.3227, "num_input_tokens_seen": 48938240, "step": 36115 }, { "epoch": 1.156135970808527, "grad_norm": 0.8645557165145874, "learning_rate": 1.152422936346567e-06, "loss": 0.3987, "num_input_tokens_seen": 48944880, "step": 36120 }, { "epoch": 1.1562960117790153, "grad_norm": 0.660796046257019, "learning_rate": 1.1494784048006718e-06, "loss": 0.3516, "num_input_tokens_seen": 48951392, "step": 36125 }, { "epoch": 1.1564560527495038, "grad_norm": 0.9568411707878113, "learning_rate": 1.1465375513049326e-06, "loss": 0.5401, "num_input_tokens_seen": 48958016, "step": 36130 }, { "epoch": 1.1566160937199923, "grad_norm": 1.3991910219192505, "learning_rate": 1.1436003763128616e-06, "loss": 0.5558, "num_input_tokens_seen": 48964560, "step": 36135 }, { "epoch": 1.1567761346904808, "grad_norm": 0.6096898913383484, "learning_rate": 1.1406668802774106e-06, "loss": 0.2554, "num_input_tokens_seen": 48971136, "step": 36140 }, { "epoch": 1.1569361756609693, "grad_norm": 2.0996530055999756, "learning_rate": 1.137737063650965e-06, "loss": 0.5566, "num_input_tokens_seen": 48978336, "step": 36145 }, { "epoch": 1.1570962166314576, "grad_norm": 1.0250611305236816, "learning_rate": 1.1348109268853323e-06, "loss": 0.5168, "num_input_tokens_seen": 48985280, "step": 36150 }, { "epoch": 1.157256257601946, "grad_norm": 1.2379724979400635, "learning_rate": 1.1318884704317634e-06, "loss": 0.4655, "num_input_tokens_seen": 48991936, "step": 36155 }, { "epoch": 1.1574162985724346, "grad_norm": 2.831648588180542, "learning_rate": 1.1289696947409417e-06, "loss": 0.4004, "num_input_tokens_seen": 48998736, "step": 36160 }, { "epoch": 1.157576339542923, "grad_norm": 0.4542253911495209, "learning_rate": 1.126054600262974e-06, "loss": 0.3304, "num_input_tokens_seen": 49005568, "step": 36165 }, { "epoch": 1.1577363805134113, "grad_norm": 0.5224754214286804, "learning_rate": 1.1231431874474064e-06, "loss": 0.3507, "num_input_tokens_seen": 49012128, "step": 36170 }, { "epoch": 1.1578964214838998, "grad_norm": 0.9235455989837646, "learning_rate": 1.12023545674321e-06, "loss": 0.3507, "num_input_tokens_seen": 49018960, "step": 36175 }, { "epoch": 1.1580564624543883, "grad_norm": 0.7146857976913452, "learning_rate": 1.117331408598804e-06, "loss": 0.4671, "num_input_tokens_seen": 49025520, "step": 36180 }, { "epoch": 1.1582165034248768, "grad_norm": 0.9059643149375916, "learning_rate": 1.1144310434620191e-06, "loss": 0.4636, "num_input_tokens_seen": 49032224, "step": 36185 }, { "epoch": 1.1583765443953653, "grad_norm": 1.7457575798034668, "learning_rate": 1.1115343617801365e-06, "loss": 0.462, "num_input_tokens_seen": 49039184, "step": 36190 }, { "epoch": 1.1585365853658536, "grad_norm": 0.6822516918182373, "learning_rate": 1.1086413639998515e-06, "loss": 0.519, "num_input_tokens_seen": 49045920, "step": 36195 }, { "epoch": 1.158696626336342, "grad_norm": 1.0825860500335693, "learning_rate": 1.1057520505673103e-06, "loss": 0.4138, "num_input_tokens_seen": 49052640, "step": 36200 }, { "epoch": 1.158696626336342, "eval_loss": 0.444102019071579, "eval_runtime": 504.1465, "eval_samples_per_second": 27.544, "eval_steps_per_second": 13.772, "num_input_tokens_seen": 49052640, "step": 36200 }, { "epoch": 1.1588566673068306, "grad_norm": 1.5914177894592285, "learning_rate": 1.1028664219280727e-06, "loss": 0.5329, "num_input_tokens_seen": 49059936, "step": 36205 }, { "epoch": 1.159016708277319, "grad_norm": 1.172411561012268, "learning_rate": 1.0999844785271468e-06, "loss": 0.4894, "num_input_tokens_seen": 49066624, "step": 36210 }, { "epoch": 1.1591767492478073, "grad_norm": 0.9520345330238342, "learning_rate": 1.097106220808955e-06, "loss": 0.3968, "num_input_tokens_seen": 49073024, "step": 36215 }, { "epoch": 1.1593367902182958, "grad_norm": 0.6435553431510925, "learning_rate": 1.0942316492173698e-06, "loss": 0.3616, "num_input_tokens_seen": 49079824, "step": 36220 }, { "epoch": 1.1594968311887843, "grad_norm": 1.277783751487732, "learning_rate": 1.0913607641956841e-06, "loss": 0.2819, "num_input_tokens_seen": 49086736, "step": 36225 }, { "epoch": 1.1596568721592728, "grad_norm": 0.841407835483551, "learning_rate": 1.0884935661866213e-06, "loss": 0.3744, "num_input_tokens_seen": 49093248, "step": 36230 }, { "epoch": 1.1598169131297613, "grad_norm": 1.4742896556854248, "learning_rate": 1.0856300556323418e-06, "loss": 0.4715, "num_input_tokens_seen": 49099696, "step": 36235 }, { "epoch": 1.1599769541002496, "grad_norm": 1.1748387813568115, "learning_rate": 1.0827702329744365e-06, "loss": 0.4387, "num_input_tokens_seen": 49106320, "step": 36240 }, { "epoch": 1.160136995070738, "grad_norm": 0.6009706854820251, "learning_rate": 1.0799140986539197e-06, "loss": 0.5368, "num_input_tokens_seen": 49113344, "step": 36245 }, { "epoch": 1.1602970360412266, "grad_norm": 1.2898857593536377, "learning_rate": 1.0770616531112526e-06, "loss": 0.4131, "num_input_tokens_seen": 49119952, "step": 36250 }, { "epoch": 1.160457077011715, "grad_norm": 0.6168726086616516, "learning_rate": 1.0742128967863085e-06, "loss": 0.2968, "num_input_tokens_seen": 49126848, "step": 36255 }, { "epoch": 1.1606171179822034, "grad_norm": 0.6745858788490295, "learning_rate": 1.071367830118411e-06, "loss": 0.3645, "num_input_tokens_seen": 49134112, "step": 36260 }, { "epoch": 1.1607771589526918, "grad_norm": 2.0719473361968994, "learning_rate": 1.068526453546298e-06, "loss": 0.3936, "num_input_tokens_seen": 49140864, "step": 36265 }, { "epoch": 1.1609371999231803, "grad_norm": 0.8258347511291504, "learning_rate": 1.0656887675081467e-06, "loss": 0.2814, "num_input_tokens_seen": 49148032, "step": 36270 }, { "epoch": 1.1610972408936688, "grad_norm": 0.9227496981620789, "learning_rate": 1.0628547724415628e-06, "loss": 0.5018, "num_input_tokens_seen": 49154592, "step": 36275 }, { "epoch": 1.1612572818641573, "grad_norm": 1.1122627258300781, "learning_rate": 1.0600244687835881e-06, "loss": 0.3578, "num_input_tokens_seen": 49161264, "step": 36280 }, { "epoch": 1.1614173228346456, "grad_norm": 1.0791635513305664, "learning_rate": 1.0571978569706876e-06, "loss": 0.506, "num_input_tokens_seen": 49168416, "step": 36285 }, { "epoch": 1.161577363805134, "grad_norm": 0.8785150647163391, "learning_rate": 1.0543749374387652e-06, "loss": 0.3772, "num_input_tokens_seen": 49175120, "step": 36290 }, { "epoch": 1.1617374047756226, "grad_norm": 1.742706537246704, "learning_rate": 1.051555710623142e-06, "loss": 0.5947, "num_input_tokens_seen": 49182240, "step": 36295 }, { "epoch": 1.161897445746111, "grad_norm": 0.7612987160682678, "learning_rate": 1.0487401769585847e-06, "loss": 0.3264, "num_input_tokens_seen": 49189168, "step": 36300 }, { "epoch": 1.1620574867165994, "grad_norm": 1.7901531457901, "learning_rate": 1.0459283368792845e-06, "loss": 0.4259, "num_input_tokens_seen": 49195744, "step": 36305 }, { "epoch": 1.1622175276870879, "grad_norm": 0.7883951663970947, "learning_rate": 1.043120190818858e-06, "loss": 0.3947, "num_input_tokens_seen": 49202432, "step": 36310 }, { "epoch": 1.1623775686575764, "grad_norm": 1.3827332258224487, "learning_rate": 1.0403157392103596e-06, "loss": 0.4125, "num_input_tokens_seen": 49209104, "step": 36315 }, { "epoch": 1.1625376096280648, "grad_norm": 0.783401370048523, "learning_rate": 1.0375149824862735e-06, "loss": 0.4301, "num_input_tokens_seen": 49215952, "step": 36320 }, { "epoch": 1.1626976505985533, "grad_norm": 1.2110490798950195, "learning_rate": 1.034717921078507e-06, "loss": 0.3567, "num_input_tokens_seen": 49222784, "step": 36325 }, { "epoch": 1.1628576915690416, "grad_norm": 0.3311435878276825, "learning_rate": 1.0319245554184009e-06, "loss": 0.4745, "num_input_tokens_seen": 49229472, "step": 36330 }, { "epoch": 1.1630177325395301, "grad_norm": 1.6821398735046387, "learning_rate": 1.0291348859367361e-06, "loss": 0.5733, "num_input_tokens_seen": 49235904, "step": 36335 }, { "epoch": 1.1631777735100186, "grad_norm": 0.6547881364822388, "learning_rate": 1.0263489130637016e-06, "loss": 0.5513, "num_input_tokens_seen": 49242464, "step": 36340 }, { "epoch": 1.163337814480507, "grad_norm": 1.5540788173675537, "learning_rate": 1.0235666372289427e-06, "loss": 0.4665, "num_input_tokens_seen": 49249104, "step": 36345 }, { "epoch": 1.1634978554509954, "grad_norm": 1.0587564706802368, "learning_rate": 1.0207880588615076e-06, "loss": 0.3637, "num_input_tokens_seen": 49255856, "step": 36350 }, { "epoch": 1.1636578964214839, "grad_norm": 0.5424942970275879, "learning_rate": 1.0180131783898984e-06, "loss": 0.4605, "num_input_tokens_seen": 49262800, "step": 36355 }, { "epoch": 1.1638179373919724, "grad_norm": 0.8357629179954529, "learning_rate": 1.0152419962420362e-06, "loss": 0.338, "num_input_tokens_seen": 49269216, "step": 36360 }, { "epoch": 1.1639779783624609, "grad_norm": 1.3977317810058594, "learning_rate": 1.0124745128452685e-06, "loss": 0.4602, "num_input_tokens_seen": 49275840, "step": 36365 }, { "epoch": 1.1641380193329494, "grad_norm": 0.9389896392822266, "learning_rate": 1.0097107286263758e-06, "loss": 0.4192, "num_input_tokens_seen": 49282048, "step": 36370 }, { "epoch": 1.1642980603034376, "grad_norm": 0.5398714542388916, "learning_rate": 1.00695064401157e-06, "loss": 0.3734, "num_input_tokens_seen": 49288544, "step": 36375 }, { "epoch": 1.1644581012739261, "grad_norm": 1.504989743232727, "learning_rate": 1.0041942594264886e-06, "loss": 0.3654, "num_input_tokens_seen": 49295408, "step": 36380 }, { "epoch": 1.1646181422444146, "grad_norm": 1.1057283878326416, "learning_rate": 1.001441575296208e-06, "loss": 0.3693, "num_input_tokens_seen": 49302160, "step": 36385 }, { "epoch": 1.164778183214903, "grad_norm": 0.7058910131454468, "learning_rate": 9.986925920452139e-07, "loss": 0.3883, "num_input_tokens_seen": 49308864, "step": 36390 }, { "epoch": 1.1649382241853914, "grad_norm": 0.9983810186386108, "learning_rate": 9.959473100974475e-07, "loss": 0.3138, "num_input_tokens_seen": 49315248, "step": 36395 }, { "epoch": 1.1650982651558799, "grad_norm": 1.2592873573303223, "learning_rate": 9.932057298762564e-07, "loss": 0.341, "num_input_tokens_seen": 49321648, "step": 36400 }, { "epoch": 1.1650982651558799, "eval_loss": 0.44387248158454895, "eval_runtime": 503.3728, "eval_samples_per_second": 27.586, "eval_steps_per_second": 13.793, "num_input_tokens_seen": 49321648, "step": 36400 }, { "epoch": 1.1652583061263684, "grad_norm": 1.8742083311080933, "learning_rate": 9.90467851804433e-07, "loss": 0.5283, "num_input_tokens_seen": 49328544, "step": 36405 }, { "epoch": 1.1654183470968569, "grad_norm": 1.0732091665267944, "learning_rate": 9.877336763041895e-07, "loss": 0.4174, "num_input_tokens_seen": 49335264, "step": 36410 }, { "epoch": 1.1655783880673452, "grad_norm": 0.4396938979625702, "learning_rate": 9.850032037971662e-07, "loss": 0.2483, "num_input_tokens_seen": 49341888, "step": 36415 }, { "epoch": 1.1657384290378336, "grad_norm": 0.5493690371513367, "learning_rate": 9.822764347044406e-07, "loss": 0.4735, "num_input_tokens_seen": 49348512, "step": 36420 }, { "epoch": 1.1658984700083221, "grad_norm": 0.9712446928024292, "learning_rate": 9.795533694465175e-07, "loss": 0.3711, "num_input_tokens_seen": 49355648, "step": 36425 }, { "epoch": 1.1660585109788106, "grad_norm": 0.9208157062530518, "learning_rate": 9.768340084433197e-07, "loss": 0.4625, "num_input_tokens_seen": 49362320, "step": 36430 }, { "epoch": 1.166218551949299, "grad_norm": 0.94537752866745, "learning_rate": 9.741183521142143e-07, "loss": 0.3801, "num_input_tokens_seen": 49369088, "step": 36435 }, { "epoch": 1.1663785929197874, "grad_norm": 1.239746332168579, "learning_rate": 9.714064008779889e-07, "loss": 0.3703, "num_input_tokens_seen": 49376192, "step": 36440 }, { "epoch": 1.166538633890276, "grad_norm": 1.345841646194458, "learning_rate": 9.686981551528584e-07, "loss": 0.5501, "num_input_tokens_seen": 49382848, "step": 36445 }, { "epoch": 1.1666986748607644, "grad_norm": 1.4530377388000488, "learning_rate": 9.65993615356467e-07, "loss": 0.3734, "num_input_tokens_seen": 49389840, "step": 36450 }, { "epoch": 1.166858715831253, "grad_norm": 2.1263091564178467, "learning_rate": 9.632927819058917e-07, "loss": 0.3613, "num_input_tokens_seen": 49396592, "step": 36455 }, { "epoch": 1.1670187568017412, "grad_norm": 0.5847842693328857, "learning_rate": 9.605956552176305e-07, "loss": 0.4579, "num_input_tokens_seen": 49403120, "step": 36460 }, { "epoch": 1.1671787977722297, "grad_norm": 1.335284948348999, "learning_rate": 9.579022357076223e-07, "loss": 0.509, "num_input_tokens_seen": 49409872, "step": 36465 }, { "epoch": 1.1673388387427182, "grad_norm": 0.4681970477104187, "learning_rate": 9.552125237912158e-07, "loss": 0.5331, "num_input_tokens_seen": 49416768, "step": 36470 }, { "epoch": 1.1674988797132066, "grad_norm": 1.0926570892333984, "learning_rate": 9.525265198832096e-07, "loss": 0.4925, "num_input_tokens_seen": 49423616, "step": 36475 }, { "epoch": 1.167658920683695, "grad_norm": 0.5259520411491394, "learning_rate": 9.498442243978112e-07, "loss": 0.2616, "num_input_tokens_seen": 49430384, "step": 36480 }, { "epoch": 1.1678189616541834, "grad_norm": 0.8219240307807922, "learning_rate": 9.471656377486649e-07, "loss": 0.4026, "num_input_tokens_seen": 49436752, "step": 36485 }, { "epoch": 1.167979002624672, "grad_norm": 3.143277645111084, "learning_rate": 9.444907603488456e-07, "loss": 0.3997, "num_input_tokens_seen": 49443392, "step": 36490 }, { "epoch": 1.1681390435951604, "grad_norm": 0.6060807704925537, "learning_rate": 9.418195926108514e-07, "loss": 0.2958, "num_input_tokens_seen": 49449792, "step": 36495 }, { "epoch": 1.168299084565649, "grad_norm": 2.14406418800354, "learning_rate": 9.391521349466053e-07, "loss": 0.3971, "num_input_tokens_seen": 49456800, "step": 36500 }, { "epoch": 1.1684591255361372, "grad_norm": 0.5320016145706177, "learning_rate": 9.364883877674758e-07, "loss": 0.3236, "num_input_tokens_seen": 49463840, "step": 36505 }, { "epoch": 1.1686191665066257, "grad_norm": 0.8356160521507263, "learning_rate": 9.33828351484231e-07, "loss": 0.3764, "num_input_tokens_seen": 49470944, "step": 36510 }, { "epoch": 1.1687792074771142, "grad_norm": 2.0902702808380127, "learning_rate": 9.311720265070906e-07, "loss": 0.5048, "num_input_tokens_seen": 49477904, "step": 36515 }, { "epoch": 1.1689392484476027, "grad_norm": 0.9656543731689453, "learning_rate": 9.285194132456931e-07, "loss": 0.4996, "num_input_tokens_seen": 49484352, "step": 36520 }, { "epoch": 1.169099289418091, "grad_norm": 1.7522814273834229, "learning_rate": 9.258705121091032e-07, "loss": 0.409, "num_input_tokens_seen": 49490912, "step": 36525 }, { "epoch": 1.1692593303885794, "grad_norm": 0.6940569877624512, "learning_rate": 9.232253235058136e-07, "loss": 0.4117, "num_input_tokens_seen": 49497456, "step": 36530 }, { "epoch": 1.169419371359068, "grad_norm": 1.790684461593628, "learning_rate": 9.205838478437478e-07, "loss": 0.4564, "num_input_tokens_seen": 49504400, "step": 36535 }, { "epoch": 1.1695794123295564, "grad_norm": 1.2331286668777466, "learning_rate": 9.179460855302524e-07, "loss": 0.3907, "num_input_tokens_seen": 49511104, "step": 36540 }, { "epoch": 1.169739453300045, "grad_norm": 2.889995813369751, "learning_rate": 9.153120369721046e-07, "loss": 0.4788, "num_input_tokens_seen": 49518016, "step": 36545 }, { "epoch": 1.1698994942705332, "grad_norm": 0.8544628024101257, "learning_rate": 9.126817025755103e-07, "loss": 0.4751, "num_input_tokens_seen": 49524448, "step": 36550 }, { "epoch": 1.1700595352410217, "grad_norm": 1.0654551982879639, "learning_rate": 9.100550827460947e-07, "loss": 0.3712, "num_input_tokens_seen": 49531088, "step": 36555 }, { "epoch": 1.1702195762115102, "grad_norm": 0.9000945091247559, "learning_rate": 9.0743217788892e-07, "loss": 0.4867, "num_input_tokens_seen": 49537920, "step": 36560 }, { "epoch": 1.1703796171819987, "grad_norm": 0.483447790145874, "learning_rate": 9.048129884084683e-07, "loss": 0.5061, "num_input_tokens_seen": 49544560, "step": 36565 }, { "epoch": 1.170539658152487, "grad_norm": 1.039818286895752, "learning_rate": 9.021975147086553e-07, "loss": 0.3407, "num_input_tokens_seen": 49551424, "step": 36570 }, { "epoch": 1.1706996991229754, "grad_norm": 1.855265736579895, "learning_rate": 8.995857571928141e-07, "loss": 0.4742, "num_input_tokens_seen": 49558448, "step": 36575 }, { "epoch": 1.170859740093464, "grad_norm": 1.228890061378479, "learning_rate": 8.969777162637139e-07, "loss": 0.4955, "num_input_tokens_seen": 49564912, "step": 36580 }, { "epoch": 1.1710197810639524, "grad_norm": 1.2036731243133545, "learning_rate": 8.943733923235525e-07, "loss": 0.3902, "num_input_tokens_seen": 49571984, "step": 36585 }, { "epoch": 1.171179822034441, "grad_norm": 1.7391010522842407, "learning_rate": 8.917727857739394e-07, "loss": 0.4707, "num_input_tokens_seen": 49578688, "step": 36590 }, { "epoch": 1.1713398630049292, "grad_norm": 1.0010055303573608, "learning_rate": 8.891758970159258e-07, "loss": 0.4086, "num_input_tokens_seen": 49585344, "step": 36595 }, { "epoch": 1.1714999039754177, "grad_norm": 0.46762293577194214, "learning_rate": 8.86582726449986e-07, "loss": 0.3171, "num_input_tokens_seen": 49592352, "step": 36600 }, { "epoch": 1.1714999039754177, "eval_loss": 0.4440258741378784, "eval_runtime": 501.8943, "eval_samples_per_second": 27.667, "eval_steps_per_second": 13.834, "num_input_tokens_seen": 49592352, "step": 36600 }, { "epoch": 1.1716599449459062, "grad_norm": 0.8707814812660217, "learning_rate": 8.839932744760165e-07, "loss": 0.3384, "num_input_tokens_seen": 49599200, "step": 36605 }, { "epoch": 1.1718199859163947, "grad_norm": 0.7646002769470215, "learning_rate": 8.814075414933482e-07, "loss": 0.4758, "num_input_tokens_seen": 49605952, "step": 36610 }, { "epoch": 1.171980026886883, "grad_norm": 0.8905220627784729, "learning_rate": 8.788255279007257e-07, "loss": 0.3258, "num_input_tokens_seen": 49612608, "step": 36615 }, { "epoch": 1.1721400678573715, "grad_norm": 0.9166685938835144, "learning_rate": 8.762472340963362e-07, "loss": 0.4339, "num_input_tokens_seen": 49619152, "step": 36620 }, { "epoch": 1.17230010882786, "grad_norm": 0.7794966101646423, "learning_rate": 8.736726604777811e-07, "loss": 0.3454, "num_input_tokens_seen": 49625568, "step": 36625 }, { "epoch": 1.1724601497983484, "grad_norm": 1.5526890754699707, "learning_rate": 8.711018074420901e-07, "loss": 0.4094, "num_input_tokens_seen": 49632128, "step": 36630 }, { "epoch": 1.172620190768837, "grad_norm": 0.988559365272522, "learning_rate": 8.685346753857209e-07, "loss": 0.3307, "num_input_tokens_seen": 49639040, "step": 36635 }, { "epoch": 1.1727802317393252, "grad_norm": 1.2706806659698486, "learning_rate": 8.659712647045654e-07, "loss": 0.3155, "num_input_tokens_seen": 49645632, "step": 36640 }, { "epoch": 1.1729402727098137, "grad_norm": 1.376503825187683, "learning_rate": 8.634115757939209e-07, "loss": 0.3959, "num_input_tokens_seen": 49652448, "step": 36645 }, { "epoch": 1.1731003136803022, "grad_norm": 1.5392420291900635, "learning_rate": 8.608556090485387e-07, "loss": 0.7178, "num_input_tokens_seen": 49659136, "step": 36650 }, { "epoch": 1.1732603546507905, "grad_norm": 0.6861594915390015, "learning_rate": 8.583033648625671e-07, "loss": 0.475, "num_input_tokens_seen": 49666048, "step": 36655 }, { "epoch": 1.173420395621279, "grad_norm": 3.7219674587249756, "learning_rate": 8.557548436295998e-07, "loss": 0.4441, "num_input_tokens_seen": 49672544, "step": 36660 }, { "epoch": 1.1735804365917675, "grad_norm": 0.6608366370201111, "learning_rate": 8.532100457426556e-07, "loss": 0.3924, "num_input_tokens_seen": 49679488, "step": 36665 }, { "epoch": 1.173740477562256, "grad_norm": 0.4069277346134186, "learning_rate": 8.506689715941679e-07, "loss": 0.468, "num_input_tokens_seen": 49686224, "step": 36670 }, { "epoch": 1.1739005185327445, "grad_norm": 1.1858769655227661, "learning_rate": 8.481316215760011e-07, "loss": 0.4563, "num_input_tokens_seen": 49692736, "step": 36675 }, { "epoch": 1.1740605595032327, "grad_norm": 0.9413641095161438, "learning_rate": 8.455979960794558e-07, "loss": 0.5245, "num_input_tokens_seen": 49699392, "step": 36680 }, { "epoch": 1.1742206004737212, "grad_norm": 0.5766263604164124, "learning_rate": 8.430680954952364e-07, "loss": 0.4271, "num_input_tokens_seen": 49705920, "step": 36685 }, { "epoch": 1.1743806414442097, "grad_norm": 1.4731781482696533, "learning_rate": 8.405419202134974e-07, "loss": 0.3237, "num_input_tokens_seen": 49712960, "step": 36690 }, { "epoch": 1.1745406824146982, "grad_norm": 1.5540144443511963, "learning_rate": 8.380194706237993e-07, "loss": 0.5806, "num_input_tokens_seen": 49719440, "step": 36695 }, { "epoch": 1.1747007233851865, "grad_norm": 0.8832173347473145, "learning_rate": 8.355007471151366e-07, "loss": 0.4918, "num_input_tokens_seen": 49726032, "step": 36700 }, { "epoch": 1.174860764355675, "grad_norm": 1.5215566158294678, "learning_rate": 8.329857500759292e-07, "loss": 0.518, "num_input_tokens_seen": 49732432, "step": 36705 }, { "epoch": 1.1750208053261635, "grad_norm": 1.8384060859680176, "learning_rate": 8.304744798940194e-07, "loss": 0.3556, "num_input_tokens_seen": 49739536, "step": 36710 }, { "epoch": 1.175180846296652, "grad_norm": 0.8988075852394104, "learning_rate": 8.279669369566756e-07, "loss": 0.3174, "num_input_tokens_seen": 49746224, "step": 36715 }, { "epoch": 1.1753408872671405, "grad_norm": 0.7896994352340698, "learning_rate": 8.254631216505993e-07, "loss": 0.5026, "num_input_tokens_seen": 49753344, "step": 36720 }, { "epoch": 1.1755009282376288, "grad_norm": 1.1593869924545288, "learning_rate": 8.229630343619038e-07, "loss": 0.4905, "num_input_tokens_seen": 49760096, "step": 36725 }, { "epoch": 1.1756609692081172, "grad_norm": 0.9519355893135071, "learning_rate": 8.204666754761392e-07, "loss": 0.416, "num_input_tokens_seen": 49767216, "step": 36730 }, { "epoch": 1.1758210101786057, "grad_norm": 1.1209782361984253, "learning_rate": 8.179740453782669e-07, "loss": 0.4542, "num_input_tokens_seen": 49774112, "step": 36735 }, { "epoch": 1.1759810511490942, "grad_norm": 0.5400444269180298, "learning_rate": 8.154851444526907e-07, "loss": 0.3543, "num_input_tokens_seen": 49780928, "step": 36740 }, { "epoch": 1.1761410921195825, "grad_norm": 0.7466829419136047, "learning_rate": 8.129999730832283e-07, "loss": 0.26, "num_input_tokens_seen": 49788128, "step": 36745 }, { "epoch": 1.176301133090071, "grad_norm": 2.699786424636841, "learning_rate": 8.105185316531178e-07, "loss": 0.3323, "num_input_tokens_seen": 49794864, "step": 36750 }, { "epoch": 1.1764611740605595, "grad_norm": 1.3989770412445068, "learning_rate": 8.08040820545039e-07, "loss": 0.4857, "num_input_tokens_seen": 49801648, "step": 36755 }, { "epoch": 1.176621215031048, "grad_norm": 0.5767696499824524, "learning_rate": 8.055668401410782e-07, "loss": 0.5183, "num_input_tokens_seen": 49808288, "step": 36760 }, { "epoch": 1.1767812560015365, "grad_norm": 1.375177025794983, "learning_rate": 8.030965908227578e-07, "loss": 0.4454, "num_input_tokens_seen": 49815296, "step": 36765 }, { "epoch": 1.1769412969720248, "grad_norm": 0.950416088104248, "learning_rate": 8.006300729710203e-07, "loss": 0.3894, "num_input_tokens_seen": 49822544, "step": 36770 }, { "epoch": 1.1771013379425133, "grad_norm": 1.3383256196975708, "learning_rate": 7.981672869662337e-07, "loss": 0.3427, "num_input_tokens_seen": 49829504, "step": 36775 }, { "epoch": 1.1772613789130018, "grad_norm": 0.7001045346260071, "learning_rate": 7.957082331881888e-07, "loss": 0.3572, "num_input_tokens_seen": 49836288, "step": 36780 }, { "epoch": 1.1774214198834903, "grad_norm": 1.7110857963562012, "learning_rate": 7.932529120161069e-07, "loss": 0.5551, "num_input_tokens_seen": 49842800, "step": 36785 }, { "epoch": 1.1775814608539785, "grad_norm": 1.245592713356018, "learning_rate": 7.908013238286243e-07, "loss": 0.5205, "num_input_tokens_seen": 49849312, "step": 36790 }, { "epoch": 1.177741501824467, "grad_norm": 1.6061147451400757, "learning_rate": 7.883534690038136e-07, "loss": 0.5487, "num_input_tokens_seen": 49856240, "step": 36795 }, { "epoch": 1.1779015427949555, "grad_norm": 0.7764145731925964, "learning_rate": 7.859093479191559e-07, "loss": 0.404, "num_input_tokens_seen": 49863184, "step": 36800 }, { "epoch": 1.1779015427949555, "eval_loss": 0.44402313232421875, "eval_runtime": 501.6689, "eval_samples_per_second": 27.68, "eval_steps_per_second": 13.84, "num_input_tokens_seen": 49863184, "step": 36800 }, { "epoch": 1.178061583765444, "grad_norm": 0.9129008650779724, "learning_rate": 7.834689609515722e-07, "loss": 0.4739, "num_input_tokens_seen": 49869568, "step": 36805 }, { "epoch": 1.1782216247359325, "grad_norm": 0.8168427348136902, "learning_rate": 7.810323084774002e-07, "loss": 0.2875, "num_input_tokens_seen": 49876288, "step": 36810 }, { "epoch": 1.1783816657064208, "grad_norm": 1.8976153135299683, "learning_rate": 7.785993908723976e-07, "loss": 0.5107, "num_input_tokens_seen": 49883168, "step": 36815 }, { "epoch": 1.1785417066769093, "grad_norm": 0.5822235941886902, "learning_rate": 7.761702085117534e-07, "loss": 0.4626, "num_input_tokens_seen": 49890416, "step": 36820 }, { "epoch": 1.1787017476473978, "grad_norm": 0.8265109062194824, "learning_rate": 7.737447617700844e-07, "loss": 0.2653, "num_input_tokens_seen": 49896544, "step": 36825 }, { "epoch": 1.1788617886178863, "grad_norm": 1.4136147499084473, "learning_rate": 7.713230510214136e-07, "loss": 0.3501, "num_input_tokens_seen": 49903232, "step": 36830 }, { "epoch": 1.1790218295883745, "grad_norm": 1.3927165269851685, "learning_rate": 7.689050766392092e-07, "loss": 0.3783, "num_input_tokens_seen": 49909744, "step": 36835 }, { "epoch": 1.179181870558863, "grad_norm": 2.8538124561309814, "learning_rate": 7.664908389963477e-07, "loss": 0.423, "num_input_tokens_seen": 49916288, "step": 36840 }, { "epoch": 1.1793419115293515, "grad_norm": 1.277608871459961, "learning_rate": 7.64080338465134e-07, "loss": 0.4946, "num_input_tokens_seen": 49923584, "step": 36845 }, { "epoch": 1.17950195249984, "grad_norm": 2.710480213165283, "learning_rate": 7.616735754173043e-07, "loss": 0.4929, "num_input_tokens_seen": 49930800, "step": 36850 }, { "epoch": 1.1796619934703285, "grad_norm": 0.7678565979003906, "learning_rate": 7.592705502240005e-07, "loss": 0.307, "num_input_tokens_seen": 49937744, "step": 36855 }, { "epoch": 1.1798220344408168, "grad_norm": 1.1186929941177368, "learning_rate": 7.568712632558095e-07, "loss": 0.47, "num_input_tokens_seen": 49944752, "step": 36860 }, { "epoch": 1.1799820754113053, "grad_norm": 0.607414960861206, "learning_rate": 7.544757148827297e-07, "loss": 0.3741, "num_input_tokens_seen": 49951680, "step": 36865 }, { "epoch": 1.1801421163817938, "grad_norm": 0.602993369102478, "learning_rate": 7.520839054741797e-07, "loss": 0.5382, "num_input_tokens_seen": 49958608, "step": 36870 }, { "epoch": 1.1803021573522823, "grad_norm": 0.8157167434692383, "learning_rate": 7.496958353990113e-07, "loss": 0.3284, "num_input_tokens_seen": 49965184, "step": 36875 }, { "epoch": 1.1804621983227706, "grad_norm": 0.860498309135437, "learning_rate": 7.473115050254941e-07, "loss": 0.3589, "num_input_tokens_seen": 49971936, "step": 36880 }, { "epoch": 1.180622239293259, "grad_norm": 1.1139076948165894, "learning_rate": 7.449309147213173e-07, "loss": 0.3868, "num_input_tokens_seen": 49978880, "step": 36885 }, { "epoch": 1.1807822802637475, "grad_norm": 0.9805417656898499, "learning_rate": 7.425540648536067e-07, "loss": 0.5983, "num_input_tokens_seen": 49985504, "step": 36890 }, { "epoch": 1.180942321234236, "grad_norm": 0.8720136284828186, "learning_rate": 7.40180955788894e-07, "loss": 0.3905, "num_input_tokens_seen": 49992112, "step": 36895 }, { "epoch": 1.1811023622047245, "grad_norm": 0.9071004986763, "learning_rate": 7.378115878931474e-07, "loss": 0.4974, "num_input_tokens_seen": 49998768, "step": 36900 }, { "epoch": 1.1812624031752128, "grad_norm": 1.0323779582977295, "learning_rate": 7.354459615317527e-07, "loss": 0.4029, "num_input_tokens_seen": 50005072, "step": 36905 }, { "epoch": 1.1814224441457013, "grad_norm": 1.5500725507736206, "learning_rate": 7.33084077069518e-07, "loss": 0.5592, "num_input_tokens_seen": 50011984, "step": 36910 }, { "epoch": 1.1815824851161898, "grad_norm": 2.231128692626953, "learning_rate": 7.307259348706768e-07, "loss": 0.4883, "num_input_tokens_seen": 50019232, "step": 36915 }, { "epoch": 1.181742526086678, "grad_norm": 0.730543315410614, "learning_rate": 7.283715352988801e-07, "loss": 0.4837, "num_input_tokens_seen": 50026736, "step": 36920 }, { "epoch": 1.1819025670571666, "grad_norm": 0.4576643407344818, "learning_rate": 7.260208787172068e-07, "loss": 0.4396, "num_input_tokens_seen": 50033552, "step": 36925 }, { "epoch": 1.182062608027655, "grad_norm": 0.6970105171203613, "learning_rate": 7.23673965488167e-07, "loss": 0.4508, "num_input_tokens_seen": 50040448, "step": 36930 }, { "epoch": 1.1822226489981436, "grad_norm": 1.5384695529937744, "learning_rate": 7.213307959736709e-07, "loss": 0.5628, "num_input_tokens_seen": 50046768, "step": 36935 }, { "epoch": 1.182382689968632, "grad_norm": 1.378930926322937, "learning_rate": 7.189913705350715e-07, "loss": 0.4583, "num_input_tokens_seen": 50053568, "step": 36940 }, { "epoch": 1.1825427309391203, "grad_norm": 0.5337319374084473, "learning_rate": 7.166556895331411e-07, "loss": 0.5123, "num_input_tokens_seen": 50060448, "step": 36945 }, { "epoch": 1.1827027719096088, "grad_norm": 1.69148850440979, "learning_rate": 7.143237533280639e-07, "loss": 0.569, "num_input_tokens_seen": 50067184, "step": 36950 }, { "epoch": 1.1828628128800973, "grad_norm": 1.2071020603179932, "learning_rate": 7.119955622794578e-07, "loss": 0.4432, "num_input_tokens_seen": 50073648, "step": 36955 }, { "epoch": 1.1830228538505858, "grad_norm": 0.5362337231636047, "learning_rate": 7.096711167463577e-07, "loss": 0.3395, "num_input_tokens_seen": 50080480, "step": 36960 }, { "epoch": 1.183182894821074, "grad_norm": 1.0517176389694214, "learning_rate": 7.073504170872213e-07, "loss": 0.4596, "num_input_tokens_seen": 50087392, "step": 36965 }, { "epoch": 1.1833429357915626, "grad_norm": 0.622214674949646, "learning_rate": 7.05033463659932e-07, "loss": 0.513, "num_input_tokens_seen": 50094048, "step": 36970 }, { "epoch": 1.183502976762051, "grad_norm": 1.5356903076171875, "learning_rate": 7.027202568217928e-07, "loss": 0.3422, "num_input_tokens_seen": 50100800, "step": 36975 }, { "epoch": 1.1836630177325396, "grad_norm": 2.0742595195770264, "learning_rate": 7.004107969295293e-07, "loss": 0.3781, "num_input_tokens_seen": 50107648, "step": 36980 }, { "epoch": 1.183823058703028, "grad_norm": 0.7300456166267395, "learning_rate": 6.9810508433929e-07, "loss": 0.3164, "num_input_tokens_seen": 50114800, "step": 36985 }, { "epoch": 1.1839830996735163, "grad_norm": 1.6346189975738525, "learning_rate": 6.958031194066406e-07, "loss": 0.3616, "num_input_tokens_seen": 50121776, "step": 36990 }, { "epoch": 1.1841431406440048, "grad_norm": 2.6734392642974854, "learning_rate": 6.935049024865776e-07, "loss": 0.5284, "num_input_tokens_seen": 50128400, "step": 36995 }, { "epoch": 1.1843031816144933, "grad_norm": 1.3631548881530762, "learning_rate": 6.912104339335118e-07, "loss": 0.4588, "num_input_tokens_seen": 50135184, "step": 37000 }, { "epoch": 1.1843031816144933, "eval_loss": 0.4440250098705292, "eval_runtime": 500.7723, "eval_samples_per_second": 27.729, "eval_steps_per_second": 13.865, "num_input_tokens_seen": 50135184, "step": 37000 }, { "epoch": 1.1844632225849818, "grad_norm": 0.8845051527023315, "learning_rate": 6.889197141012799e-07, "loss": 0.3657, "num_input_tokens_seen": 50142496, "step": 37005 }, { "epoch": 1.18462326355547, "grad_norm": 1.6321866512298584, "learning_rate": 6.866327433431435e-07, "loss": 0.5919, "num_input_tokens_seen": 50149744, "step": 37010 }, { "epoch": 1.1847833045259586, "grad_norm": 0.9604186415672302, "learning_rate": 6.843495220117735e-07, "loss": 0.4391, "num_input_tokens_seen": 50156592, "step": 37015 }, { "epoch": 1.184943345496447, "grad_norm": 1.2028603553771973, "learning_rate": 6.820700504592798e-07, "loss": 0.4735, "num_input_tokens_seen": 50163232, "step": 37020 }, { "epoch": 1.1851033864669356, "grad_norm": 4.514913082122803, "learning_rate": 6.797943290371839e-07, "loss": 0.2603, "num_input_tokens_seen": 50169872, "step": 37025 }, { "epoch": 1.185263427437424, "grad_norm": 1.0658690929412842, "learning_rate": 6.775223580964274e-07, "loss": 0.6204, "num_input_tokens_seen": 50176880, "step": 37030 }, { "epoch": 1.1854234684079124, "grad_norm": 0.8530901670455933, "learning_rate": 6.7525413798738e-07, "loss": 0.2372, "num_input_tokens_seen": 50183840, "step": 37035 }, { "epoch": 1.1855835093784008, "grad_norm": 0.7082592844963074, "learning_rate": 6.729896690598259e-07, "loss": 0.3518, "num_input_tokens_seen": 50190528, "step": 37040 }, { "epoch": 1.1857435503488893, "grad_norm": 1.52073335647583, "learning_rate": 6.707289516629772e-07, "loss": 0.3547, "num_input_tokens_seen": 50196992, "step": 37045 }, { "epoch": 1.1859035913193778, "grad_norm": 0.719871997833252, "learning_rate": 6.684719861454692e-07, "loss": 0.3858, "num_input_tokens_seen": 50203520, "step": 37050 }, { "epoch": 1.186063632289866, "grad_norm": 1.695510745048523, "learning_rate": 6.662187728553481e-07, "loss": 0.4419, "num_input_tokens_seen": 50210240, "step": 37055 }, { "epoch": 1.1862236732603546, "grad_norm": 1.3553118705749512, "learning_rate": 6.639693121400892e-07, "loss": 0.5943, "num_input_tokens_seen": 50216816, "step": 37060 }, { "epoch": 1.186383714230843, "grad_norm": 0.8530353307723999, "learning_rate": 6.617236043465868e-07, "loss": 0.3331, "num_input_tokens_seen": 50223648, "step": 37065 }, { "epoch": 1.1865437552013316, "grad_norm": 1.0704673528671265, "learning_rate": 6.594816498211587e-07, "loss": 0.5136, "num_input_tokens_seen": 50230160, "step": 37070 }, { "epoch": 1.18670379617182, "grad_norm": 0.9553940296173096, "learning_rate": 6.572434489095447e-07, "loss": 0.4714, "num_input_tokens_seen": 50236832, "step": 37075 }, { "epoch": 1.1868638371423084, "grad_norm": 2.267127752304077, "learning_rate": 6.550090019568994e-07, "loss": 0.481, "num_input_tokens_seen": 50243616, "step": 37080 }, { "epoch": 1.1870238781127969, "grad_norm": 2.2641398906707764, "learning_rate": 6.527783093078027e-07, "loss": 0.5103, "num_input_tokens_seen": 50251216, "step": 37085 }, { "epoch": 1.1871839190832854, "grad_norm": 0.8285337686538696, "learning_rate": 6.5055137130626e-07, "loss": 0.3173, "num_input_tokens_seen": 50257952, "step": 37090 }, { "epoch": 1.1873439600537739, "grad_norm": 1.762891173362732, "learning_rate": 6.483281882956854e-07, "loss": 0.5206, "num_input_tokens_seen": 50264544, "step": 37095 }, { "epoch": 1.1875040010242621, "grad_norm": 1.3743870258331299, "learning_rate": 6.461087606189298e-07, "loss": 0.3716, "num_input_tokens_seen": 50271696, "step": 37100 }, { "epoch": 1.1876640419947506, "grad_norm": 2.0238606929779053, "learning_rate": 6.438930886182554e-07, "loss": 0.5195, "num_input_tokens_seen": 50278544, "step": 37105 }, { "epoch": 1.1878240829652391, "grad_norm": 1.08851957321167, "learning_rate": 6.416811726353417e-07, "loss": 0.3233, "num_input_tokens_seen": 50285056, "step": 37110 }, { "epoch": 1.1879841239357276, "grad_norm": 0.8284158706665039, "learning_rate": 6.394730130112991e-07, "loss": 0.5761, "num_input_tokens_seen": 50291840, "step": 37115 }, { "epoch": 1.188144164906216, "grad_norm": 3.995879888534546, "learning_rate": 6.372686100866471e-07, "loss": 0.4091, "num_input_tokens_seen": 50298384, "step": 37120 }, { "epoch": 1.1883042058767044, "grad_norm": 1.3197343349456787, "learning_rate": 6.350679642013413e-07, "loss": 0.4062, "num_input_tokens_seen": 50305008, "step": 37125 }, { "epoch": 1.1884642468471929, "grad_norm": 0.9802370667457581, "learning_rate": 6.328710756947437e-07, "loss": 0.3963, "num_input_tokens_seen": 50311680, "step": 37130 }, { "epoch": 1.1886242878176814, "grad_norm": 1.3915656805038452, "learning_rate": 6.306779449056416e-07, "loss": 0.5373, "num_input_tokens_seen": 50319072, "step": 37135 }, { "epoch": 1.1887843287881699, "grad_norm": 0.7021545767784119, "learning_rate": 6.284885721722422e-07, "loss": 0.3601, "num_input_tokens_seen": 50325968, "step": 37140 }, { "epoch": 1.1889443697586581, "grad_norm": 1.0484217405319214, "learning_rate": 6.26302957832181e-07, "loss": 0.3774, "num_input_tokens_seen": 50332576, "step": 37145 }, { "epoch": 1.1891044107291466, "grad_norm": 1.3966586589813232, "learning_rate": 6.241211022224997e-07, "loss": 0.5133, "num_input_tokens_seen": 50339024, "step": 37150 }, { "epoch": 1.1892644516996351, "grad_norm": 1.2407824993133545, "learning_rate": 6.219430056796732e-07, "loss": 0.5365, "num_input_tokens_seen": 50345472, "step": 37155 }, { "epoch": 1.1894244926701236, "grad_norm": 2.0518648624420166, "learning_rate": 6.19768668539586e-07, "loss": 0.4132, "num_input_tokens_seen": 50352320, "step": 37160 }, { "epoch": 1.1895845336406121, "grad_norm": 0.5263341665267944, "learning_rate": 6.175980911375528e-07, "loss": 0.2253, "num_input_tokens_seen": 50359152, "step": 37165 }, { "epoch": 1.1897445746111004, "grad_norm": 1.2283700704574585, "learning_rate": 6.154312738083034e-07, "loss": 0.3184, "num_input_tokens_seen": 50365856, "step": 37170 }, { "epoch": 1.1899046155815889, "grad_norm": 0.7419281601905823, "learning_rate": 6.132682168859843e-07, "loss": 0.4254, "num_input_tokens_seen": 50372768, "step": 37175 }, { "epoch": 1.1900646565520774, "grad_norm": 1.2488420009613037, "learning_rate": 6.111089207041704e-07, "loss": 0.3438, "num_input_tokens_seen": 50379712, "step": 37180 }, { "epoch": 1.1902246975225657, "grad_norm": 0.6985135674476624, "learning_rate": 6.089533855958507e-07, "loss": 0.5408, "num_input_tokens_seen": 50386768, "step": 37185 }, { "epoch": 1.1903847384930542, "grad_norm": 1.2625058889389038, "learning_rate": 6.068016118934372e-07, "loss": 0.3599, "num_input_tokens_seen": 50393680, "step": 37190 }, { "epoch": 1.1905447794635426, "grad_norm": 0.5557836890220642, "learning_rate": 6.04653599928759e-07, "loss": 0.3967, "num_input_tokens_seen": 50400672, "step": 37195 }, { "epoch": 1.1907048204340311, "grad_norm": 0.6746461391448975, "learning_rate": 6.025093500330675e-07, "loss": 0.4547, "num_input_tokens_seen": 50407568, "step": 37200 }, { "epoch": 1.1907048204340311, "eval_loss": 0.44408589601516724, "eval_runtime": 500.6453, "eval_samples_per_second": 27.736, "eval_steps_per_second": 13.868, "num_input_tokens_seen": 50407568, "step": 37200 }, { "epoch": 1.1908648614045196, "grad_norm": 0.9541804194450378, "learning_rate": 6.003688625370291e-07, "loss": 0.4165, "num_input_tokens_seen": 50414576, "step": 37205 }, { "epoch": 1.191024902375008, "grad_norm": 0.6802651882171631, "learning_rate": 5.982321377707406e-07, "loss": 0.3303, "num_input_tokens_seen": 50420864, "step": 37210 }, { "epoch": 1.1911849433454964, "grad_norm": 1.4793685674667358, "learning_rate": 5.96099176063708e-07, "loss": 0.4703, "num_input_tokens_seen": 50427632, "step": 37215 }, { "epoch": 1.191344984315985, "grad_norm": 1.4632902145385742, "learning_rate": 5.93969977744857e-07, "loss": 0.452, "num_input_tokens_seen": 50434208, "step": 37220 }, { "epoch": 1.1915050252864734, "grad_norm": 0.7247810959815979, "learning_rate": 5.918445431425445e-07, "loss": 0.4291, "num_input_tokens_seen": 50441264, "step": 37225 }, { "epoch": 1.1916650662569617, "grad_norm": 1.3748146295547485, "learning_rate": 5.897228725845333e-07, "loss": 0.3361, "num_input_tokens_seen": 50447792, "step": 37230 }, { "epoch": 1.1918251072274502, "grad_norm": 1.187942624092102, "learning_rate": 5.876049663980171e-07, "loss": 0.4114, "num_input_tokens_seen": 50454560, "step": 37235 }, { "epoch": 1.1919851481979387, "grad_norm": 1.110108733177185, "learning_rate": 5.854908249095959e-07, "loss": 0.47, "num_input_tokens_seen": 50461088, "step": 37240 }, { "epoch": 1.1921451891684272, "grad_norm": 0.6498258709907532, "learning_rate": 5.833804484453031e-07, "loss": 0.4067, "num_input_tokens_seen": 50468064, "step": 37245 }, { "epoch": 1.1923052301389157, "grad_norm": 1.2495806217193604, "learning_rate": 5.81273837330587e-07, "loss": 0.5149, "num_input_tokens_seen": 50474864, "step": 37250 }, { "epoch": 1.192465271109404, "grad_norm": 0.9515522718429565, "learning_rate": 5.791709918903071e-07, "loss": 0.3222, "num_input_tokens_seen": 50481808, "step": 37255 }, { "epoch": 1.1926253120798924, "grad_norm": 0.743956983089447, "learning_rate": 5.770719124487483e-07, "loss": 0.5961, "num_input_tokens_seen": 50488400, "step": 37260 }, { "epoch": 1.192785353050381, "grad_norm": 1.223838210105896, "learning_rate": 5.749765993296241e-07, "loss": 0.4086, "num_input_tokens_seen": 50494896, "step": 37265 }, { "epoch": 1.1929453940208694, "grad_norm": 1.2915805578231812, "learning_rate": 5.728850528560509e-07, "loss": 0.4802, "num_input_tokens_seen": 50501264, "step": 37270 }, { "epoch": 1.1931054349913577, "grad_norm": 1.4513447284698486, "learning_rate": 5.707972733505707e-07, "loss": 0.3387, "num_input_tokens_seen": 50507792, "step": 37275 }, { "epoch": 1.1932654759618462, "grad_norm": 2.422985315322876, "learning_rate": 5.687132611351509e-07, "loss": 0.433, "num_input_tokens_seen": 50514880, "step": 37280 }, { "epoch": 1.1934255169323347, "grad_norm": 2.0840017795562744, "learning_rate": 5.666330165311651e-07, "loss": 0.5066, "num_input_tokens_seen": 50521472, "step": 37285 }, { "epoch": 1.1935855579028232, "grad_norm": 0.7797309756278992, "learning_rate": 5.645565398594204e-07, "loss": 0.2976, "num_input_tokens_seen": 50528384, "step": 37290 }, { "epoch": 1.1937455988733117, "grad_norm": 0.5514177680015564, "learning_rate": 5.624838314401304e-07, "loss": 0.3663, "num_input_tokens_seen": 50535184, "step": 37295 }, { "epoch": 1.1939056398438, "grad_norm": 0.9423919916152954, "learning_rate": 5.604148915929336e-07, "loss": 0.4696, "num_input_tokens_seen": 50542448, "step": 37300 }, { "epoch": 1.1940656808142884, "grad_norm": 1.841941475868225, "learning_rate": 5.583497206368887e-07, "loss": 0.3733, "num_input_tokens_seen": 50549024, "step": 37305 }, { "epoch": 1.194225721784777, "grad_norm": 0.7542312741279602, "learning_rate": 5.562883188904688e-07, "loss": 0.322, "num_input_tokens_seen": 50555744, "step": 37310 }, { "epoch": 1.1943857627552654, "grad_norm": 0.6190003752708435, "learning_rate": 5.542306866715724e-07, "loss": 0.4922, "num_input_tokens_seen": 50562512, "step": 37315 }, { "epoch": 1.1945458037257537, "grad_norm": 1.2349132299423218, "learning_rate": 5.52176824297504e-07, "loss": 0.285, "num_input_tokens_seen": 50569648, "step": 37320 }, { "epoch": 1.1947058446962422, "grad_norm": 0.7758252620697021, "learning_rate": 5.501267320850018e-07, "loss": 0.503, "num_input_tokens_seen": 50576768, "step": 37325 }, { "epoch": 1.1948658856667307, "grad_norm": 0.7591339349746704, "learning_rate": 5.480804103502157e-07, "loss": 0.3656, "num_input_tokens_seen": 50583344, "step": 37330 }, { "epoch": 1.1950259266372192, "grad_norm": 1.8123496770858765, "learning_rate": 5.460378594087101e-07, "loss": 0.3686, "num_input_tokens_seen": 50590208, "step": 37335 }, { "epoch": 1.1951859676077077, "grad_norm": 0.7524552345275879, "learning_rate": 5.439990795754773e-07, "loss": 0.4882, "num_input_tokens_seen": 50597024, "step": 37340 }, { "epoch": 1.195346008578196, "grad_norm": 0.6158735752105713, "learning_rate": 5.419640711649188e-07, "loss": 0.5674, "num_input_tokens_seen": 50604064, "step": 37345 }, { "epoch": 1.1955060495486844, "grad_norm": 0.9076447486877441, "learning_rate": 5.399328344908583e-07, "loss": 0.3565, "num_input_tokens_seen": 50611104, "step": 37350 }, { "epoch": 1.195666090519173, "grad_norm": 2.8327596187591553, "learning_rate": 5.379053698665399e-07, "loss": 0.4685, "num_input_tokens_seen": 50617664, "step": 37355 }, { "epoch": 1.1958261314896614, "grad_norm": 0.4810369610786438, "learning_rate": 5.358816776046216e-07, "loss": 0.351, "num_input_tokens_seen": 50624240, "step": 37360 }, { "epoch": 1.1959861724601497, "grad_norm": 0.8461878895759583, "learning_rate": 5.338617580171817e-07, "loss": 0.3661, "num_input_tokens_seen": 50631360, "step": 37365 }, { "epoch": 1.1961462134306382, "grad_norm": 1.0479086637496948, "learning_rate": 5.318456114157239e-07, "loss": 0.3854, "num_input_tokens_seen": 50638384, "step": 37370 }, { "epoch": 1.1963062544011267, "grad_norm": 0.346531480550766, "learning_rate": 5.298332381111576e-07, "loss": 0.3129, "num_input_tokens_seen": 50645040, "step": 37375 }, { "epoch": 1.1964662953716152, "grad_norm": 0.6682314872741699, "learning_rate": 5.27824638413818e-07, "loss": 0.5871, "num_input_tokens_seen": 50651904, "step": 37380 }, { "epoch": 1.1966263363421037, "grad_norm": 1.1566376686096191, "learning_rate": 5.258198126334546e-07, "loss": 0.3996, "num_input_tokens_seen": 50658576, "step": 37385 }, { "epoch": 1.196786377312592, "grad_norm": 1.574832558631897, "learning_rate": 5.238187610792367e-07, "loss": 0.3619, "num_input_tokens_seen": 50665200, "step": 37390 }, { "epoch": 1.1969464182830805, "grad_norm": 2.523665428161621, "learning_rate": 5.218214840597563e-07, "loss": 0.2947, "num_input_tokens_seen": 50671872, "step": 37395 }, { "epoch": 1.197106459253569, "grad_norm": 1.2187894582748413, "learning_rate": 5.198279818830115e-07, "loss": 0.7914, "num_input_tokens_seen": 50678192, "step": 37400 }, { "epoch": 1.197106459253569, "eval_loss": 0.44395920634269714, "eval_runtime": 500.528, "eval_samples_per_second": 27.743, "eval_steps_per_second": 13.871, "num_input_tokens_seen": 50678192, "step": 37400 }, { "epoch": 1.1972665002240575, "grad_norm": 1.1204999685287476, "learning_rate": 5.178382548564287e-07, "loss": 0.4126, "num_input_tokens_seen": 50684880, "step": 37405 }, { "epoch": 1.1974265411945457, "grad_norm": 0.9511638283729553, "learning_rate": 5.15852303286854e-07, "loss": 0.4427, "num_input_tokens_seen": 50691520, "step": 37410 }, { "epoch": 1.1975865821650342, "grad_norm": 1.5724575519561768, "learning_rate": 5.138701274805396e-07, "loss": 0.3674, "num_input_tokens_seen": 50698272, "step": 37415 }, { "epoch": 1.1977466231355227, "grad_norm": 1.4300423860549927, "learning_rate": 5.118917277431606e-07, "loss": 0.4107, "num_input_tokens_seen": 50704976, "step": 37420 }, { "epoch": 1.1979066641060112, "grad_norm": 0.5687354803085327, "learning_rate": 5.099171043798145e-07, "loss": 0.2871, "num_input_tokens_seen": 50711536, "step": 37425 }, { "epoch": 1.1980667050764997, "grad_norm": 1.8423147201538086, "learning_rate": 5.079462576950133e-07, "loss": 0.5228, "num_input_tokens_seen": 50718928, "step": 37430 }, { "epoch": 1.198226746046988, "grad_norm": 1.3073731660842896, "learning_rate": 5.059791879926862e-07, "loss": 0.3699, "num_input_tokens_seen": 50725264, "step": 37435 }, { "epoch": 1.1983867870174765, "grad_norm": 0.5688896179199219, "learning_rate": 5.040158955761793e-07, "loss": 0.4509, "num_input_tokens_seen": 50731664, "step": 37440 }, { "epoch": 1.198546827987965, "grad_norm": 0.699887216091156, "learning_rate": 5.020563807482559e-07, "loss": 0.3459, "num_input_tokens_seen": 50738704, "step": 37445 }, { "epoch": 1.1987068689584532, "grad_norm": 0.8993141651153564, "learning_rate": 5.001006438110995e-07, "loss": 0.5182, "num_input_tokens_seen": 50745136, "step": 37450 }, { "epoch": 1.1988669099289417, "grad_norm": 1.04466712474823, "learning_rate": 4.981486850663075e-07, "loss": 0.3231, "num_input_tokens_seen": 50752016, "step": 37455 }, { "epoch": 1.1990269508994302, "grad_norm": 0.4923645853996277, "learning_rate": 4.962005048149005e-07, "loss": 0.4286, "num_input_tokens_seen": 50758800, "step": 37460 }, { "epoch": 1.1991869918699187, "grad_norm": 1.7785422801971436, "learning_rate": 4.942561033573073e-07, "loss": 0.4143, "num_input_tokens_seen": 50765552, "step": 37465 }, { "epoch": 1.1993470328404072, "grad_norm": 1.631069540977478, "learning_rate": 4.923154809933827e-07, "loss": 0.5005, "num_input_tokens_seen": 50772464, "step": 37470 }, { "epoch": 1.1995070738108955, "grad_norm": 1.5711305141448975, "learning_rate": 4.903786380223957e-07, "loss": 0.5308, "num_input_tokens_seen": 50779392, "step": 37475 }, { "epoch": 1.199667114781384, "grad_norm": 0.8336905837059021, "learning_rate": 4.884455747430266e-07, "loss": 0.4848, "num_input_tokens_seen": 50786448, "step": 37480 }, { "epoch": 1.1998271557518725, "grad_norm": 0.8443978428840637, "learning_rate": 4.865162914533816e-07, "loss": 0.3596, "num_input_tokens_seen": 50793008, "step": 37485 }, { "epoch": 1.199987196722361, "grad_norm": 1.8886981010437012, "learning_rate": 4.845907884509809e-07, "loss": 0.409, "num_input_tokens_seen": 50799984, "step": 37490 }, { "epoch": 1.2001472376928493, "grad_norm": 1.4963387250900269, "learning_rate": 4.82669066032762e-07, "loss": 0.4032, "num_input_tokens_seen": 50806400, "step": 37495 }, { "epoch": 1.2003072786633378, "grad_norm": 1.0143917798995972, "learning_rate": 4.807511244950768e-07, "loss": 0.3714, "num_input_tokens_seen": 50813152, "step": 37500 }, { "epoch": 1.2004673196338262, "grad_norm": 0.9065979719161987, "learning_rate": 4.788369641336943e-07, "loss": 0.3079, "num_input_tokens_seen": 50820288, "step": 37505 }, { "epoch": 1.2006273606043147, "grad_norm": 0.996900737285614, "learning_rate": 4.769265852438032e-07, "loss": 0.5166, "num_input_tokens_seen": 50826928, "step": 37510 }, { "epoch": 1.2007874015748032, "grad_norm": 0.6357383131980896, "learning_rate": 4.750199881200124e-07, "loss": 0.4797, "num_input_tokens_seen": 50833600, "step": 37515 }, { "epoch": 1.2009474425452915, "grad_norm": 0.8319483995437622, "learning_rate": 4.7311717305633664e-07, "loss": 0.4079, "num_input_tokens_seen": 50840224, "step": 37520 }, { "epoch": 1.20110748351578, "grad_norm": 0.9680730104446411, "learning_rate": 4.7121814034621623e-07, "loss": 0.4194, "num_input_tokens_seen": 50847136, "step": 37525 }, { "epoch": 1.2012675244862685, "grad_norm": 0.398812860250473, "learning_rate": 4.693228902825114e-07, "loss": 0.4486, "num_input_tokens_seen": 50853648, "step": 37530 }, { "epoch": 1.201427565456757, "grad_norm": 1.1012845039367676, "learning_rate": 4.6743142315748277e-07, "loss": 0.3669, "num_input_tokens_seen": 50860224, "step": 37535 }, { "epoch": 1.2015876064272453, "grad_norm": 1.172308325767517, "learning_rate": 4.655437392628276e-07, "loss": 0.5346, "num_input_tokens_seen": 50866752, "step": 37540 }, { "epoch": 1.2017476473977338, "grad_norm": 0.7720732688903809, "learning_rate": 4.636598388896463e-07, "loss": 0.5503, "num_input_tokens_seen": 50873536, "step": 37545 }, { "epoch": 1.2019076883682223, "grad_norm": 1.2398746013641357, "learning_rate": 4.6177972232845925e-07, "loss": 0.3777, "num_input_tokens_seen": 50880256, "step": 37550 }, { "epoch": 1.2020677293387108, "grad_norm": 0.696431577205658, "learning_rate": 4.5990338986920953e-07, "loss": 0.4632, "num_input_tokens_seen": 50887744, "step": 37555 }, { "epoch": 1.2022277703091993, "grad_norm": 1.120730996131897, "learning_rate": 4.5803084180124633e-07, "loss": 0.3431, "num_input_tokens_seen": 50894784, "step": 37560 }, { "epoch": 1.2023878112796875, "grad_norm": 0.8054856657981873, "learning_rate": 4.561620784133386e-07, "loss": 0.3137, "num_input_tokens_seen": 50901472, "step": 37565 }, { "epoch": 1.202547852250176, "grad_norm": 3.0119409561157227, "learning_rate": 4.5429709999367796e-07, "loss": 0.3859, "num_input_tokens_seen": 50908112, "step": 37570 }, { "epoch": 1.2027078932206645, "grad_norm": 1.0139349699020386, "learning_rate": 4.5243590682986223e-07, "loss": 0.2558, "num_input_tokens_seen": 50914960, "step": 37575 }, { "epoch": 1.202867934191153, "grad_norm": 0.8065664172172546, "learning_rate": 4.5057849920891735e-07, "loss": 0.3852, "num_input_tokens_seen": 50921920, "step": 37580 }, { "epoch": 1.2030279751616413, "grad_norm": 1.3336756229400635, "learning_rate": 4.487248774172698e-07, "loss": 0.4947, "num_input_tokens_seen": 50928880, "step": 37585 }, { "epoch": 1.2031880161321298, "grad_norm": 1.141396403312683, "learning_rate": 4.4687504174077965e-07, "loss": 0.7791, "num_input_tokens_seen": 50939504, "step": 37590 }, { "epoch": 1.2033480571026183, "grad_norm": 2.9561450481414795, "learning_rate": 4.450289924647133e-07, "loss": 0.3643, "num_input_tokens_seen": 50946352, "step": 37595 }, { "epoch": 1.2035080980731068, "grad_norm": 1.0435106754302979, "learning_rate": 4.431867298737513e-07, "loss": 0.4438, "num_input_tokens_seen": 50953312, "step": 37600 }, { "epoch": 1.2035080980731068, "eval_loss": 0.4441055953502655, "eval_runtime": 500.3577, "eval_samples_per_second": 27.752, "eval_steps_per_second": 13.876, "num_input_tokens_seen": 50953312, "step": 37600 }, { "epoch": 1.2036681390435953, "grad_norm": 1.143262267112732, "learning_rate": 4.41348254251997e-07, "loss": 0.3893, "num_input_tokens_seen": 50959648, "step": 37605 }, { "epoch": 1.2038281800140835, "grad_norm": 0.9325176477432251, "learning_rate": 4.395135658829652e-07, "loss": 0.2885, "num_input_tokens_seen": 50966240, "step": 37610 }, { "epoch": 1.203988220984572, "grad_norm": 1.1619865894317627, "learning_rate": 4.376826650495852e-07, "loss": 0.6069, "num_input_tokens_seen": 50972944, "step": 37615 }, { "epoch": 1.2041482619550605, "grad_norm": 6.292259693145752, "learning_rate": 4.358555520342117e-07, "loss": 0.493, "num_input_tokens_seen": 50979328, "step": 37620 }, { "epoch": 1.204308302925549, "grad_norm": 1.0324420928955078, "learning_rate": 4.3403222711860257e-07, "loss": 0.4725, "num_input_tokens_seen": 50985792, "step": 37625 }, { "epoch": 1.2044683438960373, "grad_norm": 1.2937625646591187, "learning_rate": 4.3221269058394133e-07, "loss": 0.5606, "num_input_tokens_seen": 50993024, "step": 37630 }, { "epoch": 1.2046283848665258, "grad_norm": 0.7386814951896667, "learning_rate": 4.303969427108173e-07, "loss": 0.2755, "num_input_tokens_seen": 51000256, "step": 37635 }, { "epoch": 1.2047884258370143, "grad_norm": 1.4790308475494385, "learning_rate": 4.2858498377924825e-07, "loss": 0.3929, "num_input_tokens_seen": 51007008, "step": 37640 }, { "epoch": 1.2049484668075028, "grad_norm": 0.49496325850486755, "learning_rate": 4.267768140686579e-07, "loss": 0.3716, "num_input_tokens_seen": 51013760, "step": 37645 }, { "epoch": 1.2051085077779913, "grad_norm": 0.8378781080245972, "learning_rate": 4.2497243385788975e-07, "loss": 0.3599, "num_input_tokens_seen": 51021104, "step": 37650 }, { "epoch": 1.2052685487484796, "grad_norm": 0.9440827369689941, "learning_rate": 4.231718434251991e-07, "loss": 0.4389, "num_input_tokens_seen": 51027712, "step": 37655 }, { "epoch": 1.205428589718968, "grad_norm": 1.8318172693252563, "learning_rate": 4.213750430482666e-07, "loss": 0.4747, "num_input_tokens_seen": 51034416, "step": 37660 }, { "epoch": 1.2055886306894565, "grad_norm": 1.0668859481811523, "learning_rate": 4.1958203300417054e-07, "loss": 0.5059, "num_input_tokens_seen": 51040912, "step": 37665 }, { "epoch": 1.205748671659945, "grad_norm": 0.999131441116333, "learning_rate": 4.177928135694259e-07, "loss": 0.5207, "num_input_tokens_seen": 51047616, "step": 37670 }, { "epoch": 1.2059087126304333, "grad_norm": 1.7374309301376343, "learning_rate": 4.1600738501994807e-07, "loss": 0.3758, "num_input_tokens_seen": 51054400, "step": 37675 }, { "epoch": 1.2060687536009218, "grad_norm": 0.7783908247947693, "learning_rate": 4.1422574763107237e-07, "loss": 0.4717, "num_input_tokens_seen": 51061952, "step": 37680 }, { "epoch": 1.2062287945714103, "grad_norm": 1.550418496131897, "learning_rate": 4.124479016775512e-07, "loss": 0.3948, "num_input_tokens_seen": 51068560, "step": 37685 }, { "epoch": 1.2063888355418988, "grad_norm": 0.8850752711296082, "learning_rate": 4.106738474335514e-07, "loss": 0.4234, "num_input_tokens_seen": 51075344, "step": 37690 }, { "epoch": 1.2065488765123873, "grad_norm": 1.3585383892059326, "learning_rate": 4.089035851726486e-07, "loss": 0.4404, "num_input_tokens_seen": 51082416, "step": 37695 }, { "epoch": 1.2067089174828756, "grad_norm": 1.811026930809021, "learning_rate": 4.0713711516784937e-07, "loss": 0.4909, "num_input_tokens_seen": 51089200, "step": 37700 }, { "epoch": 1.206868958453364, "grad_norm": 1.2121673822402954, "learning_rate": 4.05374437691558e-07, "loss": 0.3548, "num_input_tokens_seen": 51095840, "step": 37705 }, { "epoch": 1.2070289994238526, "grad_norm": 0.9321725964546204, "learning_rate": 4.036155530156044e-07, "loss": 0.4475, "num_input_tokens_seen": 51102608, "step": 37710 }, { "epoch": 1.2071890403943408, "grad_norm": 0.8922818899154663, "learning_rate": 4.018604614112298e-07, "loss": 0.4336, "num_input_tokens_seen": 51108928, "step": 37715 }, { "epoch": 1.2073490813648293, "grad_norm": 0.6625494360923767, "learning_rate": 4.0010916314908996e-07, "loss": 0.4266, "num_input_tokens_seen": 51116048, "step": 37720 }, { "epoch": 1.2075091223353178, "grad_norm": 0.8631109595298767, "learning_rate": 3.983616584992578e-07, "loss": 0.4356, "num_input_tokens_seen": 51122816, "step": 37725 }, { "epoch": 1.2076691633058063, "grad_norm": 0.9588841795921326, "learning_rate": 3.9661794773122595e-07, "loss": 0.5066, "num_input_tokens_seen": 51129280, "step": 37730 }, { "epoch": 1.2078292042762948, "grad_norm": 1.3789485692977905, "learning_rate": 3.9487803111388777e-07, "loss": 0.5132, "num_input_tokens_seen": 51136208, "step": 37735 }, { "epoch": 1.207989245246783, "grad_norm": 1.4121496677398682, "learning_rate": 3.9314190891556747e-07, "loss": 0.485, "num_input_tokens_seen": 51142928, "step": 37740 }, { "epoch": 1.2081492862172716, "grad_norm": 0.9785292744636536, "learning_rate": 3.914095814039925e-07, "loss": 0.3916, "num_input_tokens_seen": 51149584, "step": 37745 }, { "epoch": 1.20830932718776, "grad_norm": 0.9364411234855652, "learning_rate": 3.896810488463104e-07, "loss": 0.3016, "num_input_tokens_seen": 51155984, "step": 37750 }, { "epoch": 1.2084693681582486, "grad_norm": 1.4379525184631348, "learning_rate": 3.8795631150908565e-07, "loss": 0.4131, "num_input_tokens_seen": 51162864, "step": 37755 }, { "epoch": 1.2086294091287368, "grad_norm": 1.518570065498352, "learning_rate": 3.862353696582888e-07, "loss": 0.4339, "num_input_tokens_seen": 51169984, "step": 37760 }, { "epoch": 1.2087894500992253, "grad_norm": 0.9311004877090454, "learning_rate": 3.8451822355931313e-07, "loss": 0.5597, "num_input_tokens_seen": 51176592, "step": 37765 }, { "epoch": 1.2089494910697138, "grad_norm": 1.0994932651519775, "learning_rate": 3.82804873476969e-07, "loss": 0.3767, "num_input_tokens_seen": 51182992, "step": 37770 }, { "epoch": 1.2091095320402023, "grad_norm": 1.1749756336212158, "learning_rate": 3.810953196754702e-07, "loss": 0.3594, "num_input_tokens_seen": 51189760, "step": 37775 }, { "epoch": 1.2092695730106908, "grad_norm": 1.1870882511138916, "learning_rate": 3.793895624184529e-07, "loss": 0.5888, "num_input_tokens_seen": 51196304, "step": 37780 }, { "epoch": 1.209429613981179, "grad_norm": 1.2090892791748047, "learning_rate": 3.776876019689679e-07, "loss": 0.5186, "num_input_tokens_seen": 51202656, "step": 37785 }, { "epoch": 1.2095896549516676, "grad_norm": 1.4478284120559692, "learning_rate": 3.7598943858947743e-07, "loss": 0.5253, "num_input_tokens_seen": 51209440, "step": 37790 }, { "epoch": 1.209749695922156, "grad_norm": 2.2720282077789307, "learning_rate": 3.742950725418637e-07, "loss": 0.4528, "num_input_tokens_seen": 51216320, "step": 37795 }, { "epoch": 1.2099097368926446, "grad_norm": 1.9021953344345093, "learning_rate": 3.726045040874093e-07, "loss": 0.4227, "num_input_tokens_seen": 51223392, "step": 37800 }, { "epoch": 1.2099097368926446, "eval_loss": 0.4440028667449951, "eval_runtime": 500.7513, "eval_samples_per_second": 27.73, "eval_steps_per_second": 13.865, "num_input_tokens_seen": 51223392, "step": 37800 }, { "epoch": 1.2100697778631329, "grad_norm": 0.9783435463905334, "learning_rate": 3.709177334868308e-07, "loss": 0.4331, "num_input_tokens_seen": 51230128, "step": 37805 }, { "epoch": 1.2102298188336214, "grad_norm": 1.1082277297973633, "learning_rate": 3.692347610002478e-07, "loss": 0.4324, "num_input_tokens_seen": 51236544, "step": 37810 }, { "epoch": 1.2103898598041098, "grad_norm": 0.5341923236846924, "learning_rate": 3.675555868871916e-07, "loss": 0.574, "num_input_tokens_seen": 51243264, "step": 37815 }, { "epoch": 1.2105499007745983, "grad_norm": 0.8164230585098267, "learning_rate": 3.658802114066162e-07, "loss": 0.5041, "num_input_tokens_seen": 51250048, "step": 37820 }, { "epoch": 1.2107099417450868, "grad_norm": 0.6786366701126099, "learning_rate": 3.6420863481688437e-07, "loss": 0.3668, "num_input_tokens_seen": 51256576, "step": 37825 }, { "epoch": 1.210869982715575, "grad_norm": 0.9046198725700378, "learning_rate": 3.625408573757705e-07, "loss": 0.3208, "num_input_tokens_seen": 51262960, "step": 37830 }, { "epoch": 1.2110300236860636, "grad_norm": 0.9415422677993774, "learning_rate": 3.608768793404743e-07, "loss": 0.4215, "num_input_tokens_seen": 51270176, "step": 37835 }, { "epoch": 1.211190064656552, "grad_norm": 0.5807902812957764, "learning_rate": 3.592167009675934e-07, "loss": 0.3914, "num_input_tokens_seen": 51276752, "step": 37840 }, { "epoch": 1.2113501056270406, "grad_norm": 0.9723078012466431, "learning_rate": 3.575603225131563e-07, "loss": 0.4724, "num_input_tokens_seen": 51283712, "step": 37845 }, { "epoch": 1.2115101465975289, "grad_norm": 1.2015422582626343, "learning_rate": 3.55907744232592e-07, "loss": 0.308, "num_input_tokens_seen": 51290112, "step": 37850 }, { "epoch": 1.2116701875680174, "grad_norm": 1.3836745023727417, "learning_rate": 3.5425896638075217e-07, "loss": 0.3964, "num_input_tokens_seen": 51296672, "step": 37855 }, { "epoch": 1.2118302285385059, "grad_norm": 0.7816332578659058, "learning_rate": 3.5261398921189736e-07, "loss": 0.359, "num_input_tokens_seen": 51303168, "step": 37860 }, { "epoch": 1.2119902695089944, "grad_norm": 0.6637858748435974, "learning_rate": 3.509728129797024e-07, "loss": 0.4814, "num_input_tokens_seen": 51309552, "step": 37865 }, { "epoch": 1.2121503104794829, "grad_norm": 1.0057563781738281, "learning_rate": 3.4933543793725656e-07, "loss": 0.4213, "num_input_tokens_seen": 51316112, "step": 37870 }, { "epoch": 1.2123103514499711, "grad_norm": 3.2992875576019287, "learning_rate": 3.4770186433707163e-07, "loss": 0.4348, "num_input_tokens_seen": 51322704, "step": 37875 }, { "epoch": 1.2124703924204596, "grad_norm": 0.5702718496322632, "learning_rate": 3.4607209243105453e-07, "loss": 0.5471, "num_input_tokens_seen": 51329696, "step": 37880 }, { "epoch": 1.2126304333909481, "grad_norm": 1.155175805091858, "learning_rate": 3.444461224705431e-07, "loss": 0.3893, "num_input_tokens_seen": 51336400, "step": 37885 }, { "epoch": 1.2127904743614366, "grad_norm": 0.7814359068870544, "learning_rate": 3.4282395470628116e-07, "loss": 0.3279, "num_input_tokens_seen": 51343184, "step": 37890 }, { "epoch": 1.2129505153319249, "grad_norm": 1.3964393138885498, "learning_rate": 3.4120558938842417e-07, "loss": 0.5342, "num_input_tokens_seen": 51350096, "step": 37895 }, { "epoch": 1.2131105563024134, "grad_norm": 1.0473166704177856, "learning_rate": 3.395910267665503e-07, "loss": 0.3085, "num_input_tokens_seen": 51356960, "step": 37900 }, { "epoch": 1.2132705972729019, "grad_norm": 0.6967160701751709, "learning_rate": 3.3798026708964094e-07, "loss": 0.475, "num_input_tokens_seen": 51363360, "step": 37905 }, { "epoch": 1.2134306382433904, "grad_norm": 0.6319978833198547, "learning_rate": 3.3637331060609456e-07, "loss": 0.3565, "num_input_tokens_seen": 51370144, "step": 37910 }, { "epoch": 1.2135906792138789, "grad_norm": 1.0844247341156006, "learning_rate": 3.3477015756372966e-07, "loss": 0.3921, "num_input_tokens_seen": 51377216, "step": 37915 }, { "epoch": 1.2137507201843671, "grad_norm": 1.0439118146896362, "learning_rate": 3.3317080820976785e-07, "loss": 0.4986, "num_input_tokens_seen": 51383712, "step": 37920 }, { "epoch": 1.2139107611548556, "grad_norm": 0.8350580334663391, "learning_rate": 3.315752627908508e-07, "loss": 0.4615, "num_input_tokens_seen": 51390016, "step": 37925 }, { "epoch": 1.2140708021253441, "grad_norm": 1.249146819114685, "learning_rate": 3.299835215530317e-07, "loss": 0.5097, "num_input_tokens_seen": 51396464, "step": 37930 }, { "epoch": 1.2142308430958326, "grad_norm": 0.682003378868103, "learning_rate": 3.2839558474177245e-07, "loss": 0.3911, "num_input_tokens_seen": 51403200, "step": 37935 }, { "epoch": 1.214390884066321, "grad_norm": 0.583314836025238, "learning_rate": 3.2681145260196056e-07, "loss": 0.298, "num_input_tokens_seen": 51409904, "step": 37940 }, { "epoch": 1.2145509250368094, "grad_norm": 1.2373440265655518, "learning_rate": 3.252311253778839e-07, "loss": 0.4167, "num_input_tokens_seen": 51417088, "step": 37945 }, { "epoch": 1.2147109660072979, "grad_norm": 1.5308172702789307, "learning_rate": 3.2365460331325034e-07, "loss": 0.5152, "num_input_tokens_seen": 51423776, "step": 37950 }, { "epoch": 1.2148710069777864, "grad_norm": 0.5970078110694885, "learning_rate": 3.2208188665117934e-07, "loss": 0.3142, "num_input_tokens_seen": 51430928, "step": 37955 }, { "epoch": 1.2150310479482749, "grad_norm": 1.0624003410339355, "learning_rate": 3.205129756342018e-07, "loss": 0.385, "num_input_tokens_seen": 51437744, "step": 37960 }, { "epoch": 1.2151910889187632, "grad_norm": 1.3825891017913818, "learning_rate": 3.189478705042659e-07, "loss": 0.2753, "num_input_tokens_seen": 51444496, "step": 37965 }, { "epoch": 1.2153511298892516, "grad_norm": 1.032371997833252, "learning_rate": 3.173865715027341e-07, "loss": 0.4271, "num_input_tokens_seen": 51451136, "step": 37970 }, { "epoch": 1.2155111708597401, "grad_norm": 1.41609525680542, "learning_rate": 3.158290788703694e-07, "loss": 0.4322, "num_input_tokens_seen": 51457680, "step": 37975 }, { "epoch": 1.2156712118302284, "grad_norm": 1.4135264158248901, "learning_rate": 3.1427539284736297e-07, "loss": 0.3056, "num_input_tokens_seen": 51464368, "step": 37980 }, { "epoch": 1.215831252800717, "grad_norm": 1.3977185487747192, "learning_rate": 3.127255136733093e-07, "loss": 0.4559, "num_input_tokens_seen": 51470880, "step": 37985 }, { "epoch": 1.2159912937712054, "grad_norm": 1.2942991256713867, "learning_rate": 3.1117944158722544e-07, "loss": 0.4047, "num_input_tokens_seen": 51477952, "step": 37990 }, { "epoch": 1.216151334741694, "grad_norm": 1.2412934303283691, "learning_rate": 3.0963717682752635e-07, "loss": 0.5419, "num_input_tokens_seen": 51485184, "step": 37995 }, { "epoch": 1.2163113757121824, "grad_norm": 0.8060369491577148, "learning_rate": 3.080987196320578e-07, "loss": 0.287, "num_input_tokens_seen": 51491824, "step": 38000 }, { "epoch": 1.2163113757121824, "eval_loss": 0.4440440237522125, "eval_runtime": 500.6292, "eval_samples_per_second": 27.737, "eval_steps_per_second": 13.869, "num_input_tokens_seen": 51491824, "step": 38000 }, { "epoch": 1.2164714166826707, "grad_norm": 0.9745396971702576, "learning_rate": 3.065640702380607e-07, "loss": 0.5369, "num_input_tokens_seen": 51498448, "step": 38005 }, { "epoch": 1.2166314576531592, "grad_norm": 0.6379018425941467, "learning_rate": 3.050332288822011e-07, "loss": 0.3475, "num_input_tokens_seen": 51505056, "step": 38010 }, { "epoch": 1.2167914986236477, "grad_norm": 0.866948127746582, "learning_rate": 3.035061958005542e-07, "loss": 0.4072, "num_input_tokens_seen": 51512032, "step": 38015 }, { "epoch": 1.2169515395941362, "grad_norm": 0.9984107613563538, "learning_rate": 3.019829712286093e-07, "loss": 0.3783, "num_input_tokens_seen": 51518816, "step": 38020 }, { "epoch": 1.2171115805646244, "grad_norm": 1.7684638500213623, "learning_rate": 3.004635554012647e-07, "loss": 0.5093, "num_input_tokens_seen": 51525248, "step": 38025 }, { "epoch": 1.217271621535113, "grad_norm": 0.8747856020927429, "learning_rate": 2.9894794855283017e-07, "loss": 0.3657, "num_input_tokens_seen": 51532144, "step": 38030 }, { "epoch": 1.2174316625056014, "grad_norm": 2.724395990371704, "learning_rate": 2.9743615091703816e-07, "loss": 0.4631, "num_input_tokens_seen": 51538912, "step": 38035 }, { "epoch": 1.21759170347609, "grad_norm": 0.7500375509262085, "learning_rate": 2.959281627270216e-07, "loss": 0.363, "num_input_tokens_seen": 51545824, "step": 38040 }, { "epoch": 1.2177517444465784, "grad_norm": 1.1358914375305176, "learning_rate": 2.944239842153362e-07, "loss": 0.3454, "num_input_tokens_seen": 51552304, "step": 38045 }, { "epoch": 1.2179117854170667, "grad_norm": 3.0498642921447754, "learning_rate": 2.929236156139381e-07, "loss": 0.5664, "num_input_tokens_seen": 51559056, "step": 38050 }, { "epoch": 1.2180718263875552, "grad_norm": 0.7869550585746765, "learning_rate": 2.9142705715420883e-07, "loss": 0.4685, "num_input_tokens_seen": 51565808, "step": 38055 }, { "epoch": 1.2182318673580437, "grad_norm": 2.5671944618225098, "learning_rate": 2.8993430906693595e-07, "loss": 0.4422, "num_input_tokens_seen": 51572448, "step": 38060 }, { "epoch": 1.2183919083285322, "grad_norm": 0.9268649816513062, "learning_rate": 2.88445371582316e-07, "loss": 0.5604, "num_input_tokens_seen": 51579232, "step": 38065 }, { "epoch": 1.2185519492990204, "grad_norm": 1.8310304880142212, "learning_rate": 2.8696024492996796e-07, "loss": 0.7252, "num_input_tokens_seen": 51585664, "step": 38070 }, { "epoch": 1.218711990269509, "grad_norm": 1.4336520433425903, "learning_rate": 2.854789293389115e-07, "loss": 0.5199, "num_input_tokens_seen": 51592464, "step": 38075 }, { "epoch": 1.2188720312399974, "grad_norm": 1.1146351099014282, "learning_rate": 2.8400142503758606e-07, "loss": 0.4785, "num_input_tokens_seen": 51599184, "step": 38080 }, { "epoch": 1.219032072210486, "grad_norm": 2.323026657104492, "learning_rate": 2.8252773225384276e-07, "loss": 0.441, "num_input_tokens_seen": 51605904, "step": 38085 }, { "epoch": 1.2191921131809744, "grad_norm": 0.5733765363693237, "learning_rate": 2.8105785121494143e-07, "loss": 0.3623, "num_input_tokens_seen": 51612800, "step": 38090 }, { "epoch": 1.2193521541514627, "grad_norm": 0.4306499660015106, "learning_rate": 2.795917821475563e-07, "loss": 0.422, "num_input_tokens_seen": 51619456, "step": 38095 }, { "epoch": 1.2195121951219512, "grad_norm": 0.9606510996818542, "learning_rate": 2.78129525277776e-07, "loss": 0.5458, "num_input_tokens_seen": 51626464, "step": 38100 }, { "epoch": 1.2196722360924397, "grad_norm": 0.9410779476165771, "learning_rate": 2.766710808310952e-07, "loss": 0.5435, "num_input_tokens_seen": 51633616, "step": 38105 }, { "epoch": 1.2198322770629282, "grad_norm": 1.237830400466919, "learning_rate": 2.7521644903242827e-07, "loss": 0.2834, "num_input_tokens_seen": 51640160, "step": 38110 }, { "epoch": 1.2199923180334165, "grad_norm": 1.508835792541504, "learning_rate": 2.7376563010609593e-07, "loss": 0.4333, "num_input_tokens_seen": 51646864, "step": 38115 }, { "epoch": 1.220152359003905, "grad_norm": 1.5274226665496826, "learning_rate": 2.72318624275833e-07, "loss": 0.6509, "num_input_tokens_seen": 51653520, "step": 38120 }, { "epoch": 1.2203123999743934, "grad_norm": 0.8049749135971069, "learning_rate": 2.7087543176478324e-07, "loss": 0.3567, "num_input_tokens_seen": 51660560, "step": 38125 }, { "epoch": 1.220472440944882, "grad_norm": 0.9415270686149597, "learning_rate": 2.694360527955103e-07, "loss": 0.2658, "num_input_tokens_seen": 51667392, "step": 38130 }, { "epoch": 1.2206324819153704, "grad_norm": 0.4672722816467285, "learning_rate": 2.680004875899811e-07, "loss": 0.3201, "num_input_tokens_seen": 51674640, "step": 38135 }, { "epoch": 1.2207925228858587, "grad_norm": 1.0316722393035889, "learning_rate": 2.665687363695768e-07, "loss": 0.4422, "num_input_tokens_seen": 51681232, "step": 38140 }, { "epoch": 1.2209525638563472, "grad_norm": 1.3662519454956055, "learning_rate": 2.6514079935509584e-07, "loss": 0.3033, "num_input_tokens_seen": 51689152, "step": 38145 }, { "epoch": 1.2211126048268357, "grad_norm": 1.2080628871917725, "learning_rate": 2.6371667676673983e-07, "loss": 0.4258, "num_input_tokens_seen": 51696208, "step": 38150 }, { "epoch": 1.2212726457973242, "grad_norm": 0.6394874453544617, "learning_rate": 2.6229636882412755e-07, "loss": 0.4751, "num_input_tokens_seen": 51702928, "step": 38155 }, { "epoch": 1.2214326867678125, "grad_norm": 0.7173629999160767, "learning_rate": 2.6087987574628935e-07, "loss": 0.3689, "num_input_tokens_seen": 51709808, "step": 38160 }, { "epoch": 1.221592727738301, "grad_norm": 1.2623305320739746, "learning_rate": 2.5946719775166437e-07, "loss": 0.3174, "num_input_tokens_seen": 51716256, "step": 38165 }, { "epoch": 1.2217527687087895, "grad_norm": 0.9159582853317261, "learning_rate": 2.5805833505810616e-07, "loss": 0.4075, "num_input_tokens_seen": 51723136, "step": 38170 }, { "epoch": 1.221912809679278, "grad_norm": 1.8974031209945679, "learning_rate": 2.566532878828798e-07, "loss": 0.4975, "num_input_tokens_seen": 51729776, "step": 38175 }, { "epoch": 1.2220728506497665, "grad_norm": 1.0245479345321655, "learning_rate": 2.552520564426619e-07, "loss": 0.4287, "num_input_tokens_seen": 51736480, "step": 38180 }, { "epoch": 1.2222328916202547, "grad_norm": 1.2523109912872314, "learning_rate": 2.5385464095353803e-07, "loss": 0.2993, "num_input_tokens_seen": 51743008, "step": 38185 }, { "epoch": 1.2223929325907432, "grad_norm": 1.165134310722351, "learning_rate": 2.5246104163100804e-07, "loss": 0.4589, "num_input_tokens_seen": 51749376, "step": 38190 }, { "epoch": 1.2225529735612317, "grad_norm": 0.7976319193840027, "learning_rate": 2.510712586899833e-07, "loss": 0.3791, "num_input_tokens_seen": 51755936, "step": 38195 }, { "epoch": 1.2227130145317202, "grad_norm": 0.7184064984321594, "learning_rate": 2.4968529234478124e-07, "loss": 0.3409, "num_input_tokens_seen": 51763040, "step": 38200 }, { "epoch": 1.2227130145317202, "eval_loss": 0.4439009726047516, "eval_runtime": 500.2173, "eval_samples_per_second": 27.76, "eval_steps_per_second": 13.88, "num_input_tokens_seen": 51763040, "step": 38200 }, { "epoch": 1.2228730555022085, "grad_norm": 2.87097430229187, "learning_rate": 2.483031428091448e-07, "loss": 0.5183, "num_input_tokens_seen": 51769456, "step": 38205 }, { "epoch": 1.223033096472697, "grad_norm": 0.658376157283783, "learning_rate": 2.469248102962091e-07, "loss": 0.2987, "num_input_tokens_seen": 51776864, "step": 38210 }, { "epoch": 1.2231931374431855, "grad_norm": 0.9630885124206543, "learning_rate": 2.4555029501853455e-07, "loss": 0.3511, "num_input_tokens_seen": 51784096, "step": 38215 }, { "epoch": 1.223353178413674, "grad_norm": 0.785548985004425, "learning_rate": 2.441795971880906e-07, "loss": 0.3566, "num_input_tokens_seen": 51790624, "step": 38220 }, { "epoch": 1.2235132193841625, "grad_norm": 0.815841555595398, "learning_rate": 2.4281271701625255e-07, "loss": 0.3292, "num_input_tokens_seen": 51797808, "step": 38225 }, { "epoch": 1.2236732603546507, "grad_norm": 0.7717330455780029, "learning_rate": 2.4144965471381007e-07, "loss": 0.5569, "num_input_tokens_seen": 51804592, "step": 38230 }, { "epoch": 1.2238333013251392, "grad_norm": 1.353119969367981, "learning_rate": 2.400904104909674e-07, "loss": 0.4352, "num_input_tokens_seen": 51811488, "step": 38235 }, { "epoch": 1.2239933422956277, "grad_norm": 1.4099441766738892, "learning_rate": 2.3873498455733725e-07, "loss": 0.5059, "num_input_tokens_seen": 51818336, "step": 38240 }, { "epoch": 1.224153383266116, "grad_norm": 1.0963051319122314, "learning_rate": 2.3738337712194137e-07, "loss": 0.5334, "num_input_tokens_seen": 51825120, "step": 38245 }, { "epoch": 1.2243134242366045, "grad_norm": 0.8449654579162598, "learning_rate": 2.3603558839321305e-07, "loss": 0.3566, "num_input_tokens_seen": 51832128, "step": 38250 }, { "epoch": 1.224473465207093, "grad_norm": 0.7448559999465942, "learning_rate": 2.3469161857900267e-07, "loss": 0.3116, "num_input_tokens_seen": 51838816, "step": 38255 }, { "epoch": 1.2246335061775815, "grad_norm": 0.6732180118560791, "learning_rate": 2.3335146788656393e-07, "loss": 0.5009, "num_input_tokens_seen": 51845248, "step": 38260 }, { "epoch": 1.22479354714807, "grad_norm": 1.2991482019424438, "learning_rate": 2.3201513652256757e-07, "loss": 0.6109, "num_input_tokens_seen": 51851744, "step": 38265 }, { "epoch": 1.2249535881185583, "grad_norm": 0.3209378123283386, "learning_rate": 2.3068262469308766e-07, "loss": 0.2905, "num_input_tokens_seen": 51858640, "step": 38270 }, { "epoch": 1.2251136290890468, "grad_norm": 1.0569955110549927, "learning_rate": 2.2935393260362093e-07, "loss": 0.5397, "num_input_tokens_seen": 51865072, "step": 38275 }, { "epoch": 1.2252736700595352, "grad_norm": 0.9124591946601868, "learning_rate": 2.2802906045906458e-07, "loss": 0.5417, "num_input_tokens_seen": 51872080, "step": 38280 }, { "epoch": 1.2254337110300237, "grad_norm": 0.6062467694282532, "learning_rate": 2.2670800846373018e-07, "loss": 0.3359, "num_input_tokens_seen": 51878784, "step": 38285 }, { "epoch": 1.225593752000512, "grad_norm": 0.8531752824783325, "learning_rate": 2.2539077682134367e-07, "loss": 0.4048, "num_input_tokens_seen": 51885360, "step": 38290 }, { "epoch": 1.2257537929710005, "grad_norm": 2.030630111694336, "learning_rate": 2.2407736573503423e-07, "loss": 0.3387, "num_input_tokens_seen": 51892848, "step": 38295 }, { "epoch": 1.225913833941489, "grad_norm": 4.621054172515869, "learning_rate": 2.2276777540735093e-07, "loss": 0.3757, "num_input_tokens_seen": 51899408, "step": 38300 }, { "epoch": 1.2260738749119775, "grad_norm": 1.2756314277648926, "learning_rate": 2.2146200604024613e-07, "loss": 0.4878, "num_input_tokens_seen": 51905920, "step": 38305 }, { "epoch": 1.226233915882466, "grad_norm": 0.6952580213546753, "learning_rate": 2.2016005783508375e-07, "loss": 0.2881, "num_input_tokens_seen": 51912560, "step": 38310 }, { "epoch": 1.2263939568529543, "grad_norm": 1.5365418195724487, "learning_rate": 2.1886193099264763e-07, "loss": 0.5288, "num_input_tokens_seen": 51919376, "step": 38315 }, { "epoch": 1.2265539978234428, "grad_norm": 1.154452919960022, "learning_rate": 2.175676257131165e-07, "loss": 0.4494, "num_input_tokens_seen": 51926432, "step": 38320 }, { "epoch": 1.2267140387939313, "grad_norm": 1.0308789014816284, "learning_rate": 2.162771421960974e-07, "loss": 0.4996, "num_input_tokens_seen": 51933232, "step": 38325 }, { "epoch": 1.2268740797644198, "grad_norm": 0.7239701747894287, "learning_rate": 2.1499048064059224e-07, "loss": 0.3354, "num_input_tokens_seen": 51940208, "step": 38330 }, { "epoch": 1.227034120734908, "grad_norm": 0.8928205966949463, "learning_rate": 2.1370764124502285e-07, "loss": 0.4491, "num_input_tokens_seen": 51947184, "step": 38335 }, { "epoch": 1.2271941617053965, "grad_norm": 0.8550989627838135, "learning_rate": 2.1242862420721988e-07, "loss": 0.3309, "num_input_tokens_seen": 51953760, "step": 38340 }, { "epoch": 1.227354202675885, "grad_norm": 1.371314287185669, "learning_rate": 2.1115342972442276e-07, "loss": 0.3895, "num_input_tokens_seen": 51960400, "step": 38345 }, { "epoch": 1.2275142436463735, "grad_norm": 0.6845566630363464, "learning_rate": 2.0988205799328252e-07, "loss": 0.4638, "num_input_tokens_seen": 51967024, "step": 38350 }, { "epoch": 1.227674284616862, "grad_norm": 1.3057547807693481, "learning_rate": 2.0861450920986182e-07, "loss": 0.4856, "num_input_tokens_seen": 51974144, "step": 38355 }, { "epoch": 1.2278343255873503, "grad_norm": 1.1157763004302979, "learning_rate": 2.07350783569632e-07, "loss": 0.3569, "num_input_tokens_seen": 51980896, "step": 38360 }, { "epoch": 1.2279943665578388, "grad_norm": 1.6174453496932983, "learning_rate": 2.060908812674761e-07, "loss": 0.4622, "num_input_tokens_seen": 51987568, "step": 38365 }, { "epoch": 1.2281544075283273, "grad_norm": 0.7758387327194214, "learning_rate": 2.0483480249768317e-07, "loss": 0.3769, "num_input_tokens_seen": 51994288, "step": 38370 }, { "epoch": 1.2283144484988158, "grad_norm": 1.0519198179244995, "learning_rate": 2.035825474539621e-07, "loss": 0.5887, "num_input_tokens_seen": 52001184, "step": 38375 }, { "epoch": 1.228474489469304, "grad_norm": 0.9843283891677856, "learning_rate": 2.0233411632942235e-07, "loss": 0.2935, "num_input_tokens_seen": 52007584, "step": 38380 }, { "epoch": 1.2286345304397925, "grad_norm": 1.13589346408844, "learning_rate": 2.0108950931658764e-07, "loss": 0.4196, "num_input_tokens_seen": 52013888, "step": 38385 }, { "epoch": 1.228794571410281, "grad_norm": 0.8834012746810913, "learning_rate": 1.998487266073934e-07, "loss": 0.3728, "num_input_tokens_seen": 52020128, "step": 38390 }, { "epoch": 1.2289546123807695, "grad_norm": 0.8443254232406616, "learning_rate": 1.986117683931865e-07, "loss": 0.4306, "num_input_tokens_seen": 52026592, "step": 38395 }, { "epoch": 1.229114653351258, "grad_norm": 0.9348466992378235, "learning_rate": 1.9737863486471442e-07, "loss": 0.5219, "num_input_tokens_seen": 52033392, "step": 38400 }, { "epoch": 1.229114653351258, "eval_loss": 0.4439164400100708, "eval_runtime": 500.5529, "eval_samples_per_second": 27.741, "eval_steps_per_second": 13.871, "num_input_tokens_seen": 52033392, "step": 38400 }, { "epoch": 1.2292746943217463, "grad_norm": 1.090994119644165, "learning_rate": 1.9614932621215e-07, "loss": 0.348, "num_input_tokens_seen": 52040176, "step": 38405 }, { "epoch": 1.2294347352922348, "grad_norm": 0.8228490352630615, "learning_rate": 1.9492384262506102e-07, "loss": 0.3248, "num_input_tokens_seen": 52046832, "step": 38410 }, { "epoch": 1.2295947762627233, "grad_norm": 2.0764236450195312, "learning_rate": 1.9370218429243524e-07, "loss": 0.4495, "num_input_tokens_seen": 52053520, "step": 38415 }, { "epoch": 1.2297548172332118, "grad_norm": 0.961789071559906, "learning_rate": 1.9248435140267197e-07, "loss": 0.5313, "num_input_tokens_seen": 52060336, "step": 38420 }, { "epoch": 1.2299148582037, "grad_norm": 1.1100610494613647, "learning_rate": 1.9127034414356814e-07, "loss": 0.4612, "num_input_tokens_seen": 52066880, "step": 38425 }, { "epoch": 1.2300748991741886, "grad_norm": 0.7514784932136536, "learning_rate": 1.9006016270234627e-07, "loss": 0.4447, "num_input_tokens_seen": 52073344, "step": 38430 }, { "epoch": 1.230234940144677, "grad_norm": 1.0143394470214844, "learning_rate": 1.888538072656293e-07, "loss": 0.4972, "num_input_tokens_seen": 52080112, "step": 38435 }, { "epoch": 1.2303949811151655, "grad_norm": 0.8382437229156494, "learning_rate": 1.8765127801944893e-07, "loss": 0.4493, "num_input_tokens_seen": 52086896, "step": 38440 }, { "epoch": 1.230555022085654, "grad_norm": 0.6388680338859558, "learning_rate": 1.8645257514925406e-07, "loss": 0.4621, "num_input_tokens_seen": 52093568, "step": 38445 }, { "epoch": 1.2307150630561423, "grad_norm": 1.6535797119140625, "learning_rate": 1.8525769883989685e-07, "loss": 0.4801, "num_input_tokens_seen": 52100496, "step": 38450 }, { "epoch": 1.2308751040266308, "grad_norm": 1.7080819606781006, "learning_rate": 1.8406664927564654e-07, "loss": 0.4562, "num_input_tokens_seen": 52107392, "step": 38455 }, { "epoch": 1.2310351449971193, "grad_norm": 1.0564548969268799, "learning_rate": 1.8287942664017566e-07, "loss": 0.3704, "num_input_tokens_seen": 52114064, "step": 38460 }, { "epoch": 1.2311951859676078, "grad_norm": 1.3552117347717285, "learning_rate": 1.8169603111656552e-07, "loss": 0.5502, "num_input_tokens_seen": 52120992, "step": 38465 }, { "epoch": 1.231355226938096, "grad_norm": 1.165908694267273, "learning_rate": 1.805164628873146e-07, "loss": 0.5823, "num_input_tokens_seen": 52128400, "step": 38470 }, { "epoch": 1.2315152679085846, "grad_norm": 3.732422113418579, "learning_rate": 1.793407221343274e-07, "loss": 0.6843, "num_input_tokens_seen": 52135360, "step": 38475 }, { "epoch": 1.231675308879073, "grad_norm": 1.1323659420013428, "learning_rate": 1.781688090389172e-07, "loss": 0.3409, "num_input_tokens_seen": 52142272, "step": 38480 }, { "epoch": 1.2318353498495616, "grad_norm": 0.8298894166946411, "learning_rate": 1.770007237818061e-07, "loss": 0.4258, "num_input_tokens_seen": 52149056, "step": 38485 }, { "epoch": 1.23199539082005, "grad_norm": 1.1682196855545044, "learning_rate": 1.7583646654313059e-07, "loss": 0.4731, "num_input_tokens_seen": 52155472, "step": 38490 }, { "epoch": 1.2321554317905383, "grad_norm": 0.8653672337532043, "learning_rate": 1.7467603750242757e-07, "loss": 0.3335, "num_input_tokens_seen": 52162176, "step": 38495 }, { "epoch": 1.2323154727610268, "grad_norm": 2.038736343383789, "learning_rate": 1.7351943683865944e-07, "loss": 0.4327, "num_input_tokens_seen": 52168352, "step": 38500 }, { "epoch": 1.2324755137315153, "grad_norm": 0.9485922455787659, "learning_rate": 1.723666647301808e-07, "loss": 0.3499, "num_input_tokens_seen": 52174944, "step": 38505 }, { "epoch": 1.2326355547020036, "grad_norm": 0.7217786908149719, "learning_rate": 1.712177213547661e-07, "loss": 0.452, "num_input_tokens_seen": 52181648, "step": 38510 }, { "epoch": 1.232795595672492, "grad_norm": 0.35707759857177734, "learning_rate": 1.7007260688959581e-07, "loss": 0.4915, "num_input_tokens_seen": 52188496, "step": 38515 }, { "epoch": 1.2329556366429806, "grad_norm": 0.9858397245407104, "learning_rate": 1.68931321511262e-07, "loss": 0.4276, "num_input_tokens_seen": 52195184, "step": 38520 }, { "epoch": 1.233115677613469, "grad_norm": 0.841559648513794, "learning_rate": 1.6779386539576835e-07, "loss": 0.4209, "num_input_tokens_seen": 52202368, "step": 38525 }, { "epoch": 1.2332757185839576, "grad_norm": 1.2594478130340576, "learning_rate": 1.666602387185162e-07, "loss": 0.4517, "num_input_tokens_seen": 52209024, "step": 38530 }, { "epoch": 1.2334357595544458, "grad_norm": 1.1916587352752686, "learning_rate": 1.655304416543352e-07, "loss": 0.4516, "num_input_tokens_seen": 52215904, "step": 38535 }, { "epoch": 1.2335958005249343, "grad_norm": 0.7658333778381348, "learning_rate": 1.6440447437744698e-07, "loss": 0.4499, "num_input_tokens_seen": 52222544, "step": 38540 }, { "epoch": 1.2337558414954228, "grad_norm": 1.7377610206604004, "learning_rate": 1.6328233706149332e-07, "loss": 0.4118, "num_input_tokens_seen": 52229440, "step": 38545 }, { "epoch": 1.2339158824659113, "grad_norm": 1.759990930557251, "learning_rate": 1.6216402987951906e-07, "loss": 0.5939, "num_input_tokens_seen": 52235680, "step": 38550 }, { "epoch": 1.2340759234363996, "grad_norm": 1.062950611114502, "learning_rate": 1.6104955300398627e-07, "loss": 0.3616, "num_input_tokens_seen": 52242464, "step": 38555 }, { "epoch": 1.234235964406888, "grad_norm": 0.5599203109741211, "learning_rate": 1.5993890660675748e-07, "loss": 0.3425, "num_input_tokens_seen": 52249296, "step": 38560 }, { "epoch": 1.2343960053773766, "grad_norm": 1.050881266593933, "learning_rate": 1.5883209085910678e-07, "loss": 0.2783, "num_input_tokens_seen": 52256480, "step": 38565 }, { "epoch": 1.234556046347865, "grad_norm": 1.0092021226882935, "learning_rate": 1.5772910593172264e-07, "loss": 0.4364, "num_input_tokens_seen": 52263424, "step": 38570 }, { "epoch": 1.2347160873183536, "grad_norm": 1.1096017360687256, "learning_rate": 1.5662995199469954e-07, "loss": 0.4549, "num_input_tokens_seen": 52270512, "step": 38575 }, { "epoch": 1.2348761282888419, "grad_norm": 1.9369268417358398, "learning_rate": 1.5553462921753802e-07, "loss": 0.511, "num_input_tokens_seen": 52277552, "step": 38580 }, { "epoch": 1.2350361692593304, "grad_norm": 0.6800172924995422, "learning_rate": 1.544431377691502e-07, "loss": 0.2879, "num_input_tokens_seen": 52284256, "step": 38585 }, { "epoch": 1.2351962102298188, "grad_norm": 1.6357531547546387, "learning_rate": 1.5335547781785975e-07, "loss": 0.3803, "num_input_tokens_seen": 52290960, "step": 38590 }, { "epoch": 1.2353562512003073, "grad_norm": 1.4135903120040894, "learning_rate": 1.5227164953139917e-07, "loss": 0.3807, "num_input_tokens_seen": 52297792, "step": 38595 }, { "epoch": 1.2355162921707956, "grad_norm": 1.4843392372131348, "learning_rate": 1.511916530769042e-07, "loss": 0.4185, "num_input_tokens_seen": 52304608, "step": 38600 }, { "epoch": 1.2355162921707956, "eval_loss": 0.4438559412956238, "eval_runtime": 500.164, "eval_samples_per_second": 27.763, "eval_steps_per_second": 13.881, "num_input_tokens_seen": 52304608, "step": 38600 }, { "epoch": 1.2356763331412841, "grad_norm": 1.0309220552444458, "learning_rate": 1.5011548862092773e-07, "loss": 0.4597, "num_input_tokens_seen": 52311568, "step": 38605 }, { "epoch": 1.2358363741117726, "grad_norm": 0.6245331168174744, "learning_rate": 1.490431563294231e-07, "loss": 0.326, "num_input_tokens_seen": 52318512, "step": 38610 }, { "epoch": 1.235996415082261, "grad_norm": 1.0547982454299927, "learning_rate": 1.4797465636776365e-07, "loss": 0.5145, "num_input_tokens_seen": 52325792, "step": 38615 }, { "epoch": 1.2361564560527496, "grad_norm": 1.2115707397460938, "learning_rate": 1.4690998890072027e-07, "loss": 0.3675, "num_input_tokens_seen": 52332016, "step": 38620 }, { "epoch": 1.2363164970232379, "grad_norm": 1.3243423700332642, "learning_rate": 1.4584915409248112e-07, "loss": 0.2883, "num_input_tokens_seen": 52338656, "step": 38625 }, { "epoch": 1.2364765379937264, "grad_norm": 0.4047757387161255, "learning_rate": 1.4479215210663754e-07, "loss": 0.6597, "num_input_tokens_seen": 52345712, "step": 38630 }, { "epoch": 1.2366365789642149, "grad_norm": 1.2268413305282593, "learning_rate": 1.4373898310619528e-07, "loss": 0.429, "num_input_tokens_seen": 52352416, "step": 38635 }, { "epoch": 1.2367966199347034, "grad_norm": 1.489470362663269, "learning_rate": 1.4268964725356604e-07, "loss": 0.4894, "num_input_tokens_seen": 52359024, "step": 38640 }, { "epoch": 1.2369566609051916, "grad_norm": 0.8272039294242859, "learning_rate": 1.4164414471056764e-07, "loss": 0.5024, "num_input_tokens_seen": 52366144, "step": 38645 }, { "epoch": 1.2371167018756801, "grad_norm": 1.3957339525222778, "learning_rate": 1.4060247563843497e-07, "loss": 0.6849, "num_input_tokens_seen": 52372992, "step": 38650 }, { "epoch": 1.2372767428461686, "grad_norm": 0.8703235387802124, "learning_rate": 1.3956464019780068e-07, "loss": 0.3661, "num_input_tokens_seen": 52379632, "step": 38655 }, { "epoch": 1.2374367838166571, "grad_norm": 1.2998178005218506, "learning_rate": 1.385306385487145e-07, "loss": 0.5281, "num_input_tokens_seen": 52386432, "step": 38660 }, { "epoch": 1.2375968247871456, "grad_norm": 1.4364482164382935, "learning_rate": 1.3750047085063222e-07, "loss": 0.3949, "num_input_tokens_seen": 52393024, "step": 38665 }, { "epoch": 1.2377568657576339, "grad_norm": 1.1510978937149048, "learning_rate": 1.3647413726242119e-07, "loss": 0.3544, "num_input_tokens_seen": 52399616, "step": 38670 }, { "epoch": 1.2379169067281224, "grad_norm": 0.6250775456428528, "learning_rate": 1.3545163794235205e-07, "loss": 0.4945, "num_input_tokens_seen": 52406368, "step": 38675 }, { "epoch": 1.2380769476986109, "grad_norm": 0.7739946842193604, "learning_rate": 1.3443297304810698e-07, "loss": 0.4157, "num_input_tokens_seen": 52412752, "step": 38680 }, { "epoch": 1.2382369886690994, "grad_norm": 1.145361065864563, "learning_rate": 1.3341814273677977e-07, "loss": 0.5961, "num_input_tokens_seen": 52419248, "step": 38685 }, { "epoch": 1.2383970296395876, "grad_norm": 1.2552411556243896, "learning_rate": 1.324071471648647e-07, "loss": 0.6476, "num_input_tokens_seen": 52426048, "step": 38690 }, { "epoch": 1.2385570706100761, "grad_norm": 0.5500972867012024, "learning_rate": 1.3139998648827312e-07, "loss": 0.4001, "num_input_tokens_seen": 52432752, "step": 38695 }, { "epoch": 1.2387171115805646, "grad_norm": 0.890416145324707, "learning_rate": 1.3039666086232526e-07, "loss": 0.4414, "num_input_tokens_seen": 52439376, "step": 38700 }, { "epoch": 1.2388771525510531, "grad_norm": 1.0097471475601196, "learning_rate": 1.2939717044174183e-07, "loss": 0.3023, "num_input_tokens_seen": 52446192, "step": 38705 }, { "epoch": 1.2390371935215416, "grad_norm": 0.9343335032463074, "learning_rate": 1.284015153806578e-07, "loss": 0.3562, "num_input_tokens_seen": 52452736, "step": 38710 }, { "epoch": 1.23919723449203, "grad_norm": 0.8743049502372742, "learning_rate": 1.274096958326171e-07, "loss": 0.4769, "num_input_tokens_seen": 52459968, "step": 38715 }, { "epoch": 1.2393572754625184, "grad_norm": 1.6490799188613892, "learning_rate": 1.2642171195056952e-07, "loss": 0.51, "num_input_tokens_seen": 52466976, "step": 38720 }, { "epoch": 1.2395173164330069, "grad_norm": 1.144740343093872, "learning_rate": 1.2543756388687377e-07, "loss": 0.3428, "num_input_tokens_seen": 52473920, "step": 38725 }, { "epoch": 1.2396773574034954, "grad_norm": 1.495513677597046, "learning_rate": 1.2445725179330014e-07, "loss": 0.3431, "num_input_tokens_seen": 52480640, "step": 38730 }, { "epoch": 1.2398373983739837, "grad_norm": 1.6469844579696655, "learning_rate": 1.2348077582102212e-07, "loss": 0.3903, "num_input_tokens_seen": 52487136, "step": 38735 }, { "epoch": 1.2399974393444722, "grad_norm": 1.6348471641540527, "learning_rate": 1.2250813612062762e-07, "loss": 0.4669, "num_input_tokens_seen": 52493824, "step": 38740 }, { "epoch": 1.2401574803149606, "grad_norm": 0.7057589888572693, "learning_rate": 1.215393328421105e-07, "loss": 0.447, "num_input_tokens_seen": 52500192, "step": 38745 }, { "epoch": 1.2403175212854491, "grad_norm": 1.047764539718628, "learning_rate": 1.2057436613486796e-07, "loss": 0.2676, "num_input_tokens_seen": 52506928, "step": 38750 }, { "epoch": 1.2404775622559376, "grad_norm": 1.6411212682724, "learning_rate": 1.1961323614771424e-07, "loss": 0.4149, "num_input_tokens_seen": 52513680, "step": 38755 }, { "epoch": 1.240637603226426, "grad_norm": 0.4434163272380829, "learning_rate": 1.1865594302886418e-07, "loss": 0.3135, "num_input_tokens_seen": 52520400, "step": 38760 }, { "epoch": 1.2407976441969144, "grad_norm": 0.5759666562080383, "learning_rate": 1.1770248692594687e-07, "loss": 0.6441, "num_input_tokens_seen": 52527168, "step": 38765 }, { "epoch": 1.240957685167403, "grad_norm": 1.488877534866333, "learning_rate": 1.167528679859975e-07, "loss": 0.5371, "num_input_tokens_seen": 52534112, "step": 38770 }, { "epoch": 1.2411177261378914, "grad_norm": 1.593049168586731, "learning_rate": 1.1580708635545446e-07, "loss": 0.4323, "num_input_tokens_seen": 52541104, "step": 38775 }, { "epoch": 1.2412777671083797, "grad_norm": 1.3168283700942993, "learning_rate": 1.1486514218017885e-07, "loss": 0.3265, "num_input_tokens_seen": 52547696, "step": 38780 }, { "epoch": 1.2414378080788682, "grad_norm": 0.8194010257720947, "learning_rate": 1.1392703560542117e-07, "loss": 0.401, "num_input_tokens_seen": 52554464, "step": 38785 }, { "epoch": 1.2415978490493567, "grad_norm": 0.855043351650238, "learning_rate": 1.129927667758518e-07, "loss": 0.5283, "num_input_tokens_seen": 52560752, "step": 38790 }, { "epoch": 1.2417578900198452, "grad_norm": 1.997436761856079, "learning_rate": 1.1206233583554992e-07, "loss": 0.4655, "num_input_tokens_seen": 52567584, "step": 38795 }, { "epoch": 1.2419179309903334, "grad_norm": 3.1251423358917236, "learning_rate": 1.1113574292799523e-07, "loss": 0.6725, "num_input_tokens_seen": 52574352, "step": 38800 }, { "epoch": 1.2419179309903334, "eval_loss": 0.44413456320762634, "eval_runtime": 500.5435, "eval_samples_per_second": 27.742, "eval_steps_per_second": 13.871, "num_input_tokens_seen": 52574352, "step": 38800 }, { "epoch": 1.242077971960822, "grad_norm": 1.0175679922103882, "learning_rate": 1.1021298819608449e-07, "loss": 0.3028, "num_input_tokens_seen": 52581264, "step": 38805 }, { "epoch": 1.2422380129313104, "grad_norm": 0.9055303335189819, "learning_rate": 1.0929407178211226e-07, "loss": 0.4562, "num_input_tokens_seen": 52587696, "step": 38810 }, { "epoch": 1.242398053901799, "grad_norm": 0.7160012125968933, "learning_rate": 1.0837899382779293e-07, "loss": 0.7485, "num_input_tokens_seen": 52594320, "step": 38815 }, { "epoch": 1.2425580948722872, "grad_norm": 0.6283361315727234, "learning_rate": 1.0746775447423862e-07, "loss": 0.339, "num_input_tokens_seen": 52601040, "step": 38820 }, { "epoch": 1.2427181358427757, "grad_norm": 0.7520672678947449, "learning_rate": 1.0656035386197583e-07, "loss": 0.4974, "num_input_tokens_seen": 52607840, "step": 38825 }, { "epoch": 1.2428781768132642, "grad_norm": 0.9113684892654419, "learning_rate": 1.0565679213093982e-07, "loss": 0.514, "num_input_tokens_seen": 52614704, "step": 38830 }, { "epoch": 1.2430382177837527, "grad_norm": 1.2904407978057861, "learning_rate": 1.0475706942046638e-07, "loss": 0.3496, "num_input_tokens_seen": 52621120, "step": 38835 }, { "epoch": 1.2431982587542412, "grad_norm": 1.223744511604309, "learning_rate": 1.0386118586930282e-07, "loss": 0.4204, "num_input_tokens_seen": 52627840, "step": 38840 }, { "epoch": 1.2433582997247294, "grad_norm": 0.6978201866149902, "learning_rate": 1.0296914161561367e-07, "loss": 0.4757, "num_input_tokens_seen": 52634336, "step": 38845 }, { "epoch": 1.243518340695218, "grad_norm": 0.7785102725028992, "learning_rate": 1.0208093679695552e-07, "loss": 0.4061, "num_input_tokens_seen": 52641008, "step": 38850 }, { "epoch": 1.2436783816657064, "grad_norm": 1.1427931785583496, "learning_rate": 1.0119657155030493e-07, "loss": 0.4668, "num_input_tokens_seen": 52647440, "step": 38855 }, { "epoch": 1.243838422636195, "grad_norm": 0.6727978587150574, "learning_rate": 1.003160460120417e-07, "loss": 0.396, "num_input_tokens_seen": 52654160, "step": 38860 }, { "epoch": 1.2439984636066832, "grad_norm": 1.1725298166275024, "learning_rate": 9.943936031795165e-08, "loss": 0.3957, "num_input_tokens_seen": 52661024, "step": 38865 }, { "epoch": 1.2441585045771717, "grad_norm": 1.1378592252731323, "learning_rate": 9.856651460323219e-08, "loss": 0.4449, "num_input_tokens_seen": 52667808, "step": 38870 }, { "epoch": 1.2443185455476602, "grad_norm": 1.0947962999343872, "learning_rate": 9.769750900248953e-08, "loss": 0.4856, "num_input_tokens_seen": 52674640, "step": 38875 }, { "epoch": 1.2444785865181487, "grad_norm": 1.1317095756530762, "learning_rate": 9.683234364973038e-08, "loss": 0.4061, "num_input_tokens_seen": 52681568, "step": 38880 }, { "epoch": 1.2446386274886372, "grad_norm": 0.5757045745849609, "learning_rate": 9.597101867837854e-08, "loss": 0.3705, "num_input_tokens_seen": 52688032, "step": 38885 }, { "epoch": 1.2447986684591255, "grad_norm": 1.125521183013916, "learning_rate": 9.511353422125835e-08, "loss": 0.4716, "num_input_tokens_seen": 52695136, "step": 38890 }, { "epoch": 1.244958709429614, "grad_norm": 0.7759419083595276, "learning_rate": 9.42598904106029e-08, "loss": 0.4351, "num_input_tokens_seen": 52702464, "step": 38895 }, { "epoch": 1.2451187504001024, "grad_norm": 1.633363127708435, "learning_rate": 9.341008737806245e-08, "loss": 0.3474, "num_input_tokens_seen": 52709232, "step": 38900 }, { "epoch": 1.245278791370591, "grad_norm": 1.554616928100586, "learning_rate": 9.256412525467661e-08, "loss": 0.5656, "num_input_tokens_seen": 52715904, "step": 38905 }, { "epoch": 1.2454388323410792, "grad_norm": 0.7316455245018005, "learning_rate": 9.172200417091326e-08, "loss": 0.4095, "num_input_tokens_seen": 52723024, "step": 38910 }, { "epoch": 1.2455988733115677, "grad_norm": 0.3365159332752228, "learning_rate": 9.088372425663239e-08, "loss": 0.4022, "num_input_tokens_seen": 52729616, "step": 38915 }, { "epoch": 1.2457589142820562, "grad_norm": 1.246146321296692, "learning_rate": 9.004928564110837e-08, "loss": 0.4818, "num_input_tokens_seen": 52736400, "step": 38920 }, { "epoch": 1.2459189552525447, "grad_norm": 1.5901607275009155, "learning_rate": 8.92186884530244e-08, "loss": 0.4327, "num_input_tokens_seen": 52743184, "step": 38925 }, { "epoch": 1.2460789962230332, "grad_norm": 1.0582091808319092, "learning_rate": 8.83919328204641e-08, "loss": 0.3763, "num_input_tokens_seen": 52750112, "step": 38930 }, { "epoch": 1.2462390371935215, "grad_norm": 1.507655382156372, "learning_rate": 8.756901887093105e-08, "loss": 0.4579, "num_input_tokens_seen": 52756992, "step": 38935 }, { "epoch": 1.24639907816401, "grad_norm": 0.8741564750671387, "learning_rate": 8.674994673132098e-08, "loss": 0.377, "num_input_tokens_seen": 52763600, "step": 38940 }, { "epoch": 1.2465591191344985, "grad_norm": 1.1830896139144897, "learning_rate": 8.593471652794949e-08, "loss": 0.4663, "num_input_tokens_seen": 52770336, "step": 38945 }, { "epoch": 1.246719160104987, "grad_norm": 0.4921850562095642, "learning_rate": 8.512332838653548e-08, "loss": 0.3791, "num_input_tokens_seen": 52777552, "step": 38950 }, { "epoch": 1.2468792010754752, "grad_norm": 1.9133402109146118, "learning_rate": 8.431578243220106e-08, "loss": 0.6263, "num_input_tokens_seen": 52784064, "step": 38955 }, { "epoch": 1.2470392420459637, "grad_norm": 0.9755556583404541, "learning_rate": 8.351207878948552e-08, "loss": 0.4519, "num_input_tokens_seen": 52791008, "step": 38960 }, { "epoch": 1.2471992830164522, "grad_norm": 0.9098991751670837, "learning_rate": 8.271221758232583e-08, "loss": 0.3222, "num_input_tokens_seen": 52798048, "step": 38965 }, { "epoch": 1.2473593239869407, "grad_norm": 0.5419732332229614, "learning_rate": 8.191619893407332e-08, "loss": 0.555, "num_input_tokens_seen": 52805008, "step": 38970 }, { "epoch": 1.2475193649574292, "grad_norm": 1.1954149007797241, "learning_rate": 8.112402296748534e-08, "loss": 0.4485, "num_input_tokens_seen": 52811840, "step": 38975 }, { "epoch": 1.2476794059279175, "grad_norm": 0.5339744687080383, "learning_rate": 8.033568980471973e-08, "loss": 0.4389, "num_input_tokens_seen": 52818800, "step": 38980 }, { "epoch": 1.247839446898406, "grad_norm": 0.6595876812934875, "learning_rate": 7.955119956735146e-08, "loss": 0.3154, "num_input_tokens_seen": 52825584, "step": 38985 }, { "epoch": 1.2479994878688945, "grad_norm": 0.8824321031570435, "learning_rate": 7.877055237636155e-08, "loss": 0.374, "num_input_tokens_seen": 52832656, "step": 38990 }, { "epoch": 1.248159528839383, "grad_norm": 2.109677314758301, "learning_rate": 7.79937483521287e-08, "loss": 0.5019, "num_input_tokens_seen": 52839104, "step": 38995 }, { "epoch": 1.2483195698098712, "grad_norm": 0.8497366905212402, "learning_rate": 7.722078761444873e-08, "loss": 0.3928, "num_input_tokens_seen": 52846048, "step": 39000 }, { "epoch": 1.2483195698098712, "eval_loss": 0.4441685378551483, "eval_runtime": 500.1341, "eval_samples_per_second": 27.765, "eval_steps_per_second": 13.882, "num_input_tokens_seen": 52846048, "step": 39000 }, { "epoch": 1.2484796107803597, "grad_norm": 0.7498650550842285, "learning_rate": 7.645167028252631e-08, "loss": 0.5506, "num_input_tokens_seen": 52853200, "step": 39005 }, { "epoch": 1.2486396517508482, "grad_norm": 1.9842500686645508, "learning_rate": 7.568639647496379e-08, "loss": 0.3007, "num_input_tokens_seen": 52860000, "step": 39010 }, { "epoch": 1.2487996927213367, "grad_norm": 0.8693788647651672, "learning_rate": 7.492496630977508e-08, "loss": 0.3687, "num_input_tokens_seen": 52866976, "step": 39015 }, { "epoch": 1.2489597336918252, "grad_norm": 1.569952130317688, "learning_rate": 7.416737990438571e-08, "loss": 0.3984, "num_input_tokens_seen": 52873968, "step": 39020 }, { "epoch": 1.2491197746623135, "grad_norm": 0.8939169645309448, "learning_rate": 7.341363737562445e-08, "loss": 0.3103, "num_input_tokens_seen": 52881184, "step": 39025 }, { "epoch": 1.249279815632802, "grad_norm": 0.8396551012992859, "learning_rate": 7.266373883972887e-08, "loss": 0.5512, "num_input_tokens_seen": 52888304, "step": 39030 }, { "epoch": 1.2494398566032905, "grad_norm": 0.960391104221344, "learning_rate": 7.191768441233981e-08, "loss": 0.4034, "num_input_tokens_seen": 52895504, "step": 39035 }, { "epoch": 1.249599897573779, "grad_norm": 1.0697718858718872, "learning_rate": 7.11754742085069e-08, "loss": 0.4861, "num_input_tokens_seen": 52902192, "step": 39040 }, { "epoch": 1.2497599385442673, "grad_norm": 1.2449636459350586, "learning_rate": 7.043710834269413e-08, "loss": 0.6369, "num_input_tokens_seen": 52908720, "step": 39045 }, { "epoch": 1.2499199795147558, "grad_norm": 0.7844606637954712, "learning_rate": 6.970258692876319e-08, "loss": 0.3587, "num_input_tokens_seen": 52915744, "step": 39050 }, { "epoch": 1.2500800204852442, "grad_norm": 0.4760741889476776, "learning_rate": 6.897191007998738e-08, "loss": 0.3945, "num_input_tokens_seen": 52922432, "step": 39055 }, { "epoch": 1.2502400614557327, "grad_norm": 0.6398620009422302, "learning_rate": 6.824507790904599e-08, "loss": 0.4065, "num_input_tokens_seen": 52929584, "step": 39060 }, { "epoch": 1.2504001024262212, "grad_norm": 0.6476759910583496, "learning_rate": 6.752209052802439e-08, "loss": 0.3781, "num_input_tokens_seen": 52936480, "step": 39065 }, { "epoch": 1.2505601433967095, "grad_norm": 0.8040982484817505, "learning_rate": 6.680294804841946e-08, "loss": 0.4603, "num_input_tokens_seen": 52943600, "step": 39070 }, { "epoch": 1.250720184367198, "grad_norm": 1.0096595287322998, "learning_rate": 6.608765058112865e-08, "loss": 0.4824, "num_input_tokens_seen": 52950208, "step": 39075 }, { "epoch": 1.2508802253376865, "grad_norm": 1.273358702659607, "learning_rate": 6.537619823646368e-08, "loss": 0.3553, "num_input_tokens_seen": 52957376, "step": 39080 }, { "epoch": 1.2510402663081748, "grad_norm": 1.2498657703399658, "learning_rate": 6.466859112413404e-08, "loss": 0.3171, "num_input_tokens_seen": 52964384, "step": 39085 }, { "epoch": 1.2512003072786633, "grad_norm": 0.8715188503265381, "learning_rate": 6.39648293532663e-08, "loss": 0.397, "num_input_tokens_seen": 52970736, "step": 39090 }, { "epoch": 1.2513603482491518, "grad_norm": 0.900442361831665, "learning_rate": 6.32649130323848e-08, "loss": 0.4285, "num_input_tokens_seen": 52977696, "step": 39095 }, { "epoch": 1.2515203892196403, "grad_norm": 0.7191373109817505, "learning_rate": 6.256884226943094e-08, "loss": 0.4071, "num_input_tokens_seen": 52984080, "step": 39100 }, { "epoch": 1.2516804301901288, "grad_norm": 1.1867945194244385, "learning_rate": 6.187661717174386e-08, "loss": 0.4333, "num_input_tokens_seen": 52990864, "step": 39105 }, { "epoch": 1.2518404711606173, "grad_norm": 1.0562407970428467, "learning_rate": 6.118823784607708e-08, "loss": 0.4364, "num_input_tokens_seen": 52997616, "step": 39110 }, { "epoch": 1.2520005121311055, "grad_norm": 1.3757878541946411, "learning_rate": 6.050370439858178e-08, "loss": 0.3531, "num_input_tokens_seen": 53004672, "step": 39115 }, { "epoch": 1.252160553101594, "grad_norm": 1.1992474794387817, "learning_rate": 5.98230169348235e-08, "loss": 0.4299, "num_input_tokens_seen": 53011328, "step": 39120 }, { "epoch": 1.2523205940720825, "grad_norm": 0.6891060471534729, "learning_rate": 5.914617555977664e-08, "loss": 0.3581, "num_input_tokens_seen": 53018096, "step": 39125 }, { "epoch": 1.2524806350425708, "grad_norm": 1.4662410020828247, "learning_rate": 5.8473180377816017e-08, "loss": 0.5629, "num_input_tokens_seen": 53024816, "step": 39130 }, { "epoch": 1.2526406760130593, "grad_norm": 2.17543363571167, "learning_rate": 5.780403149272251e-08, "loss": 0.4313, "num_input_tokens_seen": 53031488, "step": 39135 }, { "epoch": 1.2528007169835478, "grad_norm": 0.8031476140022278, "learning_rate": 5.7138729007694126e-08, "loss": 0.3001, "num_input_tokens_seen": 53038448, "step": 39140 }, { "epoch": 1.2529607579540363, "grad_norm": 0.7441738247871399, "learning_rate": 5.64772730253238e-08, "loss": 0.4814, "num_input_tokens_seen": 53045424, "step": 39145 }, { "epoch": 1.2531207989245248, "grad_norm": 0.9070810675621033, "learning_rate": 5.5819663647618814e-08, "loss": 0.4172, "num_input_tokens_seen": 53052160, "step": 39150 }, { "epoch": 1.253280839895013, "grad_norm": 0.9767752289772034, "learning_rate": 5.5165900975989723e-08, "loss": 0.7405, "num_input_tokens_seen": 53059312, "step": 39155 }, { "epoch": 1.2534408808655015, "grad_norm": 0.4705246090888977, "learning_rate": 5.451598511125311e-08, "loss": 0.3734, "num_input_tokens_seen": 53065872, "step": 39160 }, { "epoch": 1.25360092183599, "grad_norm": 0.8974220156669617, "learning_rate": 5.3869916153637124e-08, "loss": 0.3303, "num_input_tokens_seen": 53072672, "step": 39165 }, { "epoch": 1.2537609628064785, "grad_norm": 1.0597563982009888, "learning_rate": 5.322769420277318e-08, "loss": 0.6176, "num_input_tokens_seen": 53079632, "step": 39170 }, { "epoch": 1.2539210037769668, "grad_norm": 2.479215383529663, "learning_rate": 5.258931935769873e-08, "loss": 0.5341, "num_input_tokens_seen": 53086112, "step": 39175 }, { "epoch": 1.2540810447474553, "grad_norm": 2.411797046661377, "learning_rate": 5.19547917168628e-08, "loss": 0.4076, "num_input_tokens_seen": 53092544, "step": 39180 }, { "epoch": 1.2542410857179438, "grad_norm": 1.0917999744415283, "learning_rate": 5.13241113781121e-08, "loss": 0.3589, "num_input_tokens_seen": 53098768, "step": 39185 }, { "epoch": 1.2544011266884323, "grad_norm": 1.1424071788787842, "learning_rate": 5.0697278438707755e-08, "loss": 0.5889, "num_input_tokens_seen": 53105520, "step": 39190 }, { "epoch": 1.2545611676589208, "grad_norm": 0.5353949666023254, "learning_rate": 5.0074292995316854e-08, "loss": 0.4138, "num_input_tokens_seen": 53112032, "step": 39195 }, { "epoch": 1.254721208629409, "grad_norm": 0.9072209000587463, "learning_rate": 4.945515514400978e-08, "loss": 0.3627, "num_input_tokens_seen": 53118576, "step": 39200 }, { "epoch": 1.254721208629409, "eval_loss": 0.4438072144985199, "eval_runtime": 500.4902, "eval_samples_per_second": 27.745, "eval_steps_per_second": 13.872, "num_input_tokens_seen": 53118576, "step": 39200 }, { "epoch": 1.2548812495998976, "grad_norm": 0.8233124613761902, "learning_rate": 4.883986498026571e-08, "loss": 0.3014, "num_input_tokens_seen": 53125136, "step": 39205 }, { "epoch": 1.255041290570386, "grad_norm": 0.7670081257820129, "learning_rate": 4.822842259896987e-08, "loss": 0.4815, "num_input_tokens_seen": 53131696, "step": 39210 }, { "epoch": 1.2552013315408745, "grad_norm": 0.6229771375656128, "learning_rate": 4.762082809441626e-08, "loss": 0.2979, "num_input_tokens_seen": 53138496, "step": 39215 }, { "epoch": 1.2553613725113628, "grad_norm": 1.7914557456970215, "learning_rate": 4.7017081560302156e-08, "loss": 0.899, "num_input_tokens_seen": 53145024, "step": 39220 }, { "epoch": 1.2555214134818513, "grad_norm": 0.7887352705001831, "learning_rate": 4.6417183089730866e-08, "loss": 0.3741, "num_input_tokens_seen": 53151632, "step": 39225 }, { "epoch": 1.2556814544523398, "grad_norm": 1.3685722351074219, "learning_rate": 4.5821132775217265e-08, "loss": 0.4303, "num_input_tokens_seen": 53158320, "step": 39230 }, { "epoch": 1.2558414954228283, "grad_norm": 0.985630989074707, "learning_rate": 4.5228930708679504e-08, "loss": 0.4764, "num_input_tokens_seen": 53165088, "step": 39235 }, { "epoch": 1.2560015363933168, "grad_norm": 1.259887456893921, "learning_rate": 4.464057698144175e-08, "loss": 0.4505, "num_input_tokens_seen": 53171360, "step": 39240 }, { "epoch": 1.256161577363805, "grad_norm": 1.2572695016860962, "learning_rate": 4.4056071684236974e-08, "loss": 0.2891, "num_input_tokens_seen": 53178304, "step": 39245 }, { "epoch": 1.2563216183342936, "grad_norm": 3.218492269515991, "learning_rate": 4.347541490719864e-08, "loss": 0.5381, "num_input_tokens_seen": 53184912, "step": 39250 }, { "epoch": 1.256481659304782, "grad_norm": 0.7701823115348816, "learning_rate": 4.2898606739877336e-08, "loss": 0.4289, "num_input_tokens_seen": 53191952, "step": 39255 }, { "epoch": 1.2566417002752703, "grad_norm": 2.5681841373443604, "learning_rate": 4.232564727122135e-08, "loss": 0.5552, "num_input_tokens_seen": 53198848, "step": 39260 }, { "epoch": 1.2568017412457588, "grad_norm": 2.744537115097046, "learning_rate": 4.1756536589585004e-08, "loss": 0.4071, "num_input_tokens_seen": 53205728, "step": 39265 }, { "epoch": 1.2569617822162473, "grad_norm": 0.7768771648406982, "learning_rate": 4.119127478273976e-08, "loss": 0.4479, "num_input_tokens_seen": 53212544, "step": 39270 }, { "epoch": 1.2571218231867358, "grad_norm": 0.6203510761260986, "learning_rate": 4.062986193784923e-08, "loss": 0.3933, "num_input_tokens_seen": 53219296, "step": 39275 }, { "epoch": 1.2572818641572243, "grad_norm": 1.7983226776123047, "learning_rate": 4.007229814149416e-08, "loss": 0.5335, "num_input_tokens_seen": 53226016, "step": 39280 }, { "epoch": 1.2574419051277128, "grad_norm": 2.1766135692596436, "learning_rate": 3.951858347965576e-08, "loss": 0.3658, "num_input_tokens_seen": 53232848, "step": 39285 }, { "epoch": 1.257601946098201, "grad_norm": 0.8024305701255798, "learning_rate": 3.896871803772684e-08, "loss": 0.3843, "num_input_tokens_seen": 53239472, "step": 39290 }, { "epoch": 1.2577619870686896, "grad_norm": 1.2075036764144897, "learning_rate": 3.842270190050068e-08, "loss": 0.3835, "num_input_tokens_seen": 53246592, "step": 39295 }, { "epoch": 1.257922028039178, "grad_norm": 1.5712597370147705, "learning_rate": 3.7880535152179376e-08, "loss": 0.4344, "num_input_tokens_seen": 53253376, "step": 39300 }, { "epoch": 1.2580820690096663, "grad_norm": 1.483134150505066, "learning_rate": 3.734221787637382e-08, "loss": 0.2871, "num_input_tokens_seen": 53260000, "step": 39305 }, { "epoch": 1.2582421099801548, "grad_norm": 1.3794232606887817, "learning_rate": 3.680775015609817e-08, "loss": 0.5824, "num_input_tokens_seen": 53266752, "step": 39310 }, { "epoch": 1.2584021509506433, "grad_norm": 1.3618422746658325, "learning_rate": 3.627713207377537e-08, "loss": 0.326, "num_input_tokens_seen": 53273712, "step": 39315 }, { "epoch": 1.2585621919211318, "grad_norm": 1.1072149276733398, "learning_rate": 3.575036371123164e-08, "loss": 0.5157, "num_input_tokens_seen": 53280208, "step": 39320 }, { "epoch": 1.2587222328916203, "grad_norm": 1.6871726512908936, "learning_rate": 3.5227445149704776e-08, "loss": 0.5525, "num_input_tokens_seen": 53287520, "step": 39325 }, { "epoch": 1.2588822738621088, "grad_norm": 1.0898672342300415, "learning_rate": 3.470837646983027e-08, "loss": 0.4889, "num_input_tokens_seen": 53294160, "step": 39330 }, { "epoch": 1.259042314832597, "grad_norm": 1.2672381401062012, "learning_rate": 3.419315775165799e-08, "loss": 0.3594, "num_input_tokens_seen": 53301008, "step": 39335 }, { "epoch": 1.2592023558030856, "grad_norm": 1.293689250946045, "learning_rate": 3.368178907464103e-08, "loss": 0.4695, "num_input_tokens_seen": 53307664, "step": 39340 }, { "epoch": 1.259362396773574, "grad_norm": 0.9984648823738098, "learning_rate": 3.317427051763855e-08, "loss": 0.4146, "num_input_tokens_seen": 53314512, "step": 39345 }, { "epoch": 1.2595224377440624, "grad_norm": 0.6736137270927429, "learning_rate": 3.267060215891571e-08, "loss": 0.3637, "num_input_tokens_seen": 53321008, "step": 39350 }, { "epoch": 1.2596824787145509, "grad_norm": 0.5148435831069946, "learning_rate": 3.217078407614649e-08, "loss": 0.2501, "num_input_tokens_seen": 53327520, "step": 39355 }, { "epoch": 1.2598425196850394, "grad_norm": 1.233624815940857, "learning_rate": 3.1674816346405345e-08, "loss": 0.621, "num_input_tokens_seen": 53334272, "step": 39360 }, { "epoch": 1.2600025606555278, "grad_norm": 2.017483949661255, "learning_rate": 3.11826990461811e-08, "loss": 0.4648, "num_input_tokens_seen": 53340800, "step": 39365 }, { "epoch": 1.2601626016260163, "grad_norm": 0.9081746935844421, "learning_rate": 3.069443225136304e-08, "loss": 0.4076, "num_input_tokens_seen": 53347328, "step": 39370 }, { "epoch": 1.2603226425965048, "grad_norm": 0.7418732643127441, "learning_rate": 3.021001603724372e-08, "loss": 0.4073, "num_input_tokens_seen": 53353920, "step": 39375 }, { "epoch": 1.2604826835669931, "grad_norm": 0.6189103722572327, "learning_rate": 2.9729450478532818e-08, "loss": 0.3116, "num_input_tokens_seen": 53360848, "step": 39380 }, { "epoch": 1.2606427245374816, "grad_norm": 1.3061091899871826, "learning_rate": 2.9252735649337726e-08, "loss": 0.4711, "num_input_tokens_seen": 53367328, "step": 39385 }, { "epoch": 1.26080276550797, "grad_norm": 1.213159203529358, "learning_rate": 2.8779871623171863e-08, "loss": 0.5122, "num_input_tokens_seen": 53374256, "step": 39390 }, { "epoch": 1.2609628064784584, "grad_norm": 4.187745094299316, "learning_rate": 2.8310858472957448e-08, "loss": 0.4183, "num_input_tokens_seen": 53381072, "step": 39395 }, { "epoch": 1.2611228474489469, "grad_norm": 0.6127101182937622, "learning_rate": 2.784569627101996e-08, "loss": 0.3919, "num_input_tokens_seen": 53387872, "step": 39400 }, { "epoch": 1.2611228474489469, "eval_loss": 0.44392508268356323, "eval_runtime": 500.1397, "eval_samples_per_second": 27.764, "eval_steps_per_second": 13.882, "num_input_tokens_seen": 53387872, "step": 39400 }, { "epoch": 1.2612828884194354, "grad_norm": 0.7582268118858337, "learning_rate": 2.738438508909924e-08, "loss": 0.372, "num_input_tokens_seen": 53394400, "step": 39405 }, { "epoch": 1.2614429293899239, "grad_norm": 1.4506268501281738, "learning_rate": 2.692692499833005e-08, "loss": 0.4307, "num_input_tokens_seen": 53401888, "step": 39410 }, { "epoch": 1.2616029703604124, "grad_norm": 0.7452934980392456, "learning_rate": 2.647331606926151e-08, "loss": 0.3947, "num_input_tokens_seen": 53408848, "step": 39415 }, { "epoch": 1.2617630113309006, "grad_norm": 1.1457887887954712, "learning_rate": 2.6023558371843225e-08, "loss": 0.4412, "num_input_tokens_seen": 53415632, "step": 39420 }, { "epoch": 1.2619230523013891, "grad_norm": 1.2555745840072632, "learning_rate": 2.557765197543638e-08, "loss": 0.5458, "num_input_tokens_seen": 53422368, "step": 39425 }, { "epoch": 1.2620830932718776, "grad_norm": 0.6592350006103516, "learning_rate": 2.513559694880263e-08, "loss": 0.2849, "num_input_tokens_seen": 53428848, "step": 39430 }, { "epoch": 1.2622431342423661, "grad_norm": 1.4751551151275635, "learning_rate": 2.469739336011523e-08, "loss": 0.397, "num_input_tokens_seen": 53435552, "step": 39435 }, { "epoch": 1.2624031752128544, "grad_norm": 1.38583505153656, "learning_rate": 2.4263041276947894e-08, "loss": 0.3613, "num_input_tokens_seen": 53442032, "step": 39440 }, { "epoch": 1.2625632161833429, "grad_norm": 2.3212316036224365, "learning_rate": 2.3832540766283164e-08, "loss": 0.5267, "num_input_tokens_seen": 53448720, "step": 39445 }, { "epoch": 1.2627232571538314, "grad_norm": 0.6013729572296143, "learning_rate": 2.3405891894512366e-08, "loss": 0.323, "num_input_tokens_seen": 53455408, "step": 39450 }, { "epoch": 1.2628832981243199, "grad_norm": 1.546548843383789, "learning_rate": 2.29830947274301e-08, "loss": 0.3478, "num_input_tokens_seen": 53462528, "step": 39455 }, { "epoch": 1.2630433390948084, "grad_norm": 1.0295053720474243, "learning_rate": 2.2564149330231432e-08, "loss": 0.409, "num_input_tokens_seen": 53469184, "step": 39460 }, { "epoch": 1.2632033800652966, "grad_norm": 1.332435131072998, "learning_rate": 2.2149055767528572e-08, "loss": 0.3823, "num_input_tokens_seen": 53475568, "step": 39465 }, { "epoch": 1.2633634210357851, "grad_norm": 1.6076613664627075, "learning_rate": 2.1737814103334197e-08, "loss": 0.4367, "num_input_tokens_seen": 53482416, "step": 39470 }, { "epoch": 1.2635234620062736, "grad_norm": 1.4394923448562622, "learning_rate": 2.1330424401064253e-08, "loss": 0.2733, "num_input_tokens_seen": 53489088, "step": 39475 }, { "epoch": 1.2636835029767621, "grad_norm": 1.6312692165374756, "learning_rate": 2.092688672354348e-08, "loss": 0.5005, "num_input_tokens_seen": 53495760, "step": 39480 }, { "epoch": 1.2638435439472504, "grad_norm": 1.0419225692749023, "learning_rate": 2.0527201133005435e-08, "loss": 0.3515, "num_input_tokens_seen": 53502752, "step": 39485 }, { "epoch": 1.264003584917739, "grad_norm": 1.0388504266738892, "learning_rate": 2.0131367691084148e-08, "loss": 0.3649, "num_input_tokens_seen": 53509584, "step": 39490 }, { "epoch": 1.2641636258882274, "grad_norm": 1.8944343328475952, "learning_rate": 1.9739386458819675e-08, "loss": 0.3214, "num_input_tokens_seen": 53516544, "step": 39495 }, { "epoch": 1.2643236668587159, "grad_norm": 1.0084905624389648, "learning_rate": 1.9351257496666442e-08, "loss": 0.4028, "num_input_tokens_seen": 53524000, "step": 39500 }, { "epoch": 1.2644837078292044, "grad_norm": 1.3123408555984497, "learning_rate": 1.896698086447657e-08, "loss": 0.3803, "num_input_tokens_seen": 53530976, "step": 39505 }, { "epoch": 1.2646437487996927, "grad_norm": 0.5362628698348999, "learning_rate": 1.8586556621505436e-08, "loss": 0.4316, "num_input_tokens_seen": 53537600, "step": 39510 }, { "epoch": 1.2648037897701812, "grad_norm": 1.228289246559143, "learning_rate": 1.820998482642833e-08, "loss": 0.408, "num_input_tokens_seen": 53544032, "step": 39515 }, { "epoch": 1.2649638307406696, "grad_norm": 2.5737791061401367, "learning_rate": 1.7837265537309912e-08, "loss": 0.3739, "num_input_tokens_seen": 53551104, "step": 39520 }, { "epoch": 1.265123871711158, "grad_norm": 1.1731131076812744, "learning_rate": 1.7468398811629206e-08, "loss": 0.4523, "num_input_tokens_seen": 53557840, "step": 39525 }, { "epoch": 1.2652839126816464, "grad_norm": 0.7275082468986511, "learning_rate": 1.710338470627404e-08, "loss": 0.3521, "num_input_tokens_seen": 53564448, "step": 39530 }, { "epoch": 1.265443953652135, "grad_norm": 1.2026879787445068, "learning_rate": 1.6742223277529945e-08, "loss": 0.6582, "num_input_tokens_seen": 53570736, "step": 39535 }, { "epoch": 1.2656039946226234, "grad_norm": 0.7903927564620972, "learning_rate": 1.6384914581094036e-08, "loss": 0.3762, "num_input_tokens_seen": 53577792, "step": 39540 }, { "epoch": 1.265764035593112, "grad_norm": 0.7059358954429626, "learning_rate": 1.6031458672069455e-08, "loss": 0.5288, "num_input_tokens_seen": 53584688, "step": 39545 }, { "epoch": 1.2659240765636004, "grad_norm": 1.6993399858474731, "learning_rate": 1.5681855604962602e-08, "loss": 0.4358, "num_input_tokens_seen": 53591264, "step": 39550 }, { "epoch": 1.2660841175340887, "grad_norm": 1.2218399047851562, "learning_rate": 1.5336105433683135e-08, "loss": 0.4241, "num_input_tokens_seen": 53598288, "step": 39555 }, { "epoch": 1.2662441585045772, "grad_norm": 1.4761195182800293, "learning_rate": 1.499420821155506e-08, "loss": 0.3463, "num_input_tokens_seen": 53605312, "step": 39560 }, { "epoch": 1.2664041994750657, "grad_norm": 0.7499814033508301, "learning_rate": 1.4656163991302874e-08, "loss": 0.3497, "num_input_tokens_seen": 53612032, "step": 39565 }, { "epoch": 1.266564240445554, "grad_norm": 0.8623110055923462, "learning_rate": 1.4321972825051544e-08, "loss": 0.393, "num_input_tokens_seen": 53619216, "step": 39570 }, { "epoch": 1.2667242814160424, "grad_norm": 0.7121394276618958, "learning_rate": 1.3991634764345951e-08, "loss": 0.3037, "num_input_tokens_seen": 53625888, "step": 39575 }, { "epoch": 1.266884322386531, "grad_norm": 1.2580054998397827, "learning_rate": 1.3665149860120352e-08, "loss": 0.4731, "num_input_tokens_seen": 53633136, "step": 39580 }, { "epoch": 1.2670443633570194, "grad_norm": 0.7608388066291809, "learning_rate": 1.3342518162728912e-08, "loss": 0.2831, "num_input_tokens_seen": 53639776, "step": 39585 }, { "epoch": 1.267204404327508, "grad_norm": 0.6347780227661133, "learning_rate": 1.30237397219235e-08, "loss": 0.354, "num_input_tokens_seen": 53646528, "step": 39590 }, { "epoch": 1.2673644452979964, "grad_norm": 2.0987579822540283, "learning_rate": 1.2708814586862016e-08, "loss": 0.6377, "num_input_tokens_seen": 53653008, "step": 39595 }, { "epoch": 1.2675244862684847, "grad_norm": 0.8773940801620483, "learning_rate": 1.2397742806111168e-08, "loss": 0.3773, "num_input_tokens_seen": 53659856, "step": 39600 }, { "epoch": 1.2675244862684847, "eval_loss": 0.44395196437835693, "eval_runtime": 500.3578, "eval_samples_per_second": 27.752, "eval_steps_per_second": 13.876, "num_input_tokens_seen": 53659856, "step": 39600 }, { "epoch": 1.2676845272389732, "grad_norm": 0.714451014995575, "learning_rate": 1.209052442764369e-08, "loss": 0.4317, "num_input_tokens_seen": 53666864, "step": 39605 }, { "epoch": 1.2678445682094617, "grad_norm": 0.7274598479270935, "learning_rate": 1.17871594988328e-08, "loss": 0.4057, "num_input_tokens_seen": 53673408, "step": 39610 }, { "epoch": 1.26800460917995, "grad_norm": 0.5104867815971375, "learning_rate": 1.1487648066466072e-08, "loss": 0.2796, "num_input_tokens_seen": 53679936, "step": 39615 }, { "epoch": 1.2681646501504384, "grad_norm": 2.349175453186035, "learning_rate": 1.1191990176728784e-08, "loss": 0.562, "num_input_tokens_seen": 53686512, "step": 39620 }, { "epoch": 1.268324691120927, "grad_norm": 0.6163950562477112, "learning_rate": 1.0900185875215018e-08, "loss": 0.4299, "num_input_tokens_seen": 53693072, "step": 39625 }, { "epoch": 1.2684847320914154, "grad_norm": 1.3201817274093628, "learning_rate": 1.0612235206924891e-08, "loss": 0.3107, "num_input_tokens_seen": 53699664, "step": 39630 }, { "epoch": 1.268644773061904, "grad_norm": 0.7133184671401978, "learning_rate": 1.0328138216264549e-08, "loss": 0.3401, "num_input_tokens_seen": 53706320, "step": 39635 }, { "epoch": 1.2688048140323924, "grad_norm": 1.0473203659057617, "learning_rate": 1.004789494704339e-08, "loss": 0.4997, "num_input_tokens_seen": 53713072, "step": 39640 }, { "epoch": 1.2689648550028807, "grad_norm": 1.193054437637329, "learning_rate": 9.771505442482397e-09, "loss": 0.6228, "num_input_tokens_seen": 53719504, "step": 39645 }, { "epoch": 1.2691248959733692, "grad_norm": 1.123302698135376, "learning_rate": 9.498969745200259e-09, "loss": 0.4641, "num_input_tokens_seen": 53726752, "step": 39650 }, { "epoch": 1.2692849369438577, "grad_norm": 1.0045021772384644, "learning_rate": 9.230287897230017e-09, "loss": 0.3963, "num_input_tokens_seen": 53733392, "step": 39655 }, { "epoch": 1.269444977914346, "grad_norm": 1.4181526899337769, "learning_rate": 8.965459940002419e-09, "loss": 0.3789, "num_input_tokens_seen": 53739984, "step": 39660 }, { "epoch": 1.2696050188848345, "grad_norm": 0.8458113074302673, "learning_rate": 8.704485914357019e-09, "loss": 0.3192, "num_input_tokens_seen": 53746656, "step": 39665 }, { "epoch": 1.269765059855323, "grad_norm": 1.1516481637954712, "learning_rate": 8.447365860539402e-09, "loss": 0.3686, "num_input_tokens_seen": 53753648, "step": 39670 }, { "epoch": 1.2699251008258114, "grad_norm": 1.0328857898712158, "learning_rate": 8.194099818201184e-09, "loss": 0.4597, "num_input_tokens_seen": 53760592, "step": 39675 }, { "epoch": 1.2700851417963, "grad_norm": 0.9485328793525696, "learning_rate": 7.944687826400011e-09, "loss": 0.556, "num_input_tokens_seen": 53767264, "step": 39680 }, { "epoch": 1.2702451827667882, "grad_norm": 0.8928106427192688, "learning_rate": 7.699129923599557e-09, "loss": 0.5674, "num_input_tokens_seen": 53774224, "step": 39685 }, { "epoch": 1.2704052237372767, "grad_norm": 0.7957941293716431, "learning_rate": 7.457426147663982e-09, "loss": 0.3998, "num_input_tokens_seen": 53781088, "step": 39690 }, { "epoch": 1.2705652647077652, "grad_norm": 0.7082434892654419, "learning_rate": 7.219576535871797e-09, "loss": 0.6619, "num_input_tokens_seen": 53787520, "step": 39695 }, { "epoch": 1.2707253056782537, "grad_norm": 0.9046097993850708, "learning_rate": 6.985581124896445e-09, "loss": 0.2866, "num_input_tokens_seen": 53793856, "step": 39700 }, { "epoch": 1.270885346648742, "grad_norm": 1.1227258443832397, "learning_rate": 6.755439950828501e-09, "loss": 0.371, "num_input_tokens_seen": 53800896, "step": 39705 }, { "epoch": 1.2710453876192305, "grad_norm": 0.8597086668014526, "learning_rate": 6.5291530491562444e-09, "loss": 0.4249, "num_input_tokens_seen": 53807840, "step": 39710 }, { "epoch": 1.271205428589719, "grad_norm": 1.7431070804595947, "learning_rate": 6.3067204547739845e-09, "loss": 0.534, "num_input_tokens_seen": 53814128, "step": 39715 }, { "epoch": 1.2713654695602075, "grad_norm": 1.1498152017593384, "learning_rate": 6.088142201987612e-09, "loss": 0.5943, "num_input_tokens_seen": 53821072, "step": 39720 }, { "epoch": 1.271525510530696, "grad_norm": 2.2878224849700928, "learning_rate": 5.873418324503499e-09, "loss": 0.5854, "num_input_tokens_seen": 53827408, "step": 39725 }, { "epoch": 1.2716855515011842, "grad_norm": 2.421004295349121, "learning_rate": 5.6625488554340465e-09, "loss": 0.4573, "num_input_tokens_seen": 53833664, "step": 39730 }, { "epoch": 1.2718455924716727, "grad_norm": 0.9544460773468018, "learning_rate": 5.455533827297688e-09, "loss": 0.3534, "num_input_tokens_seen": 53840672, "step": 39735 }, { "epoch": 1.2720056334421612, "grad_norm": 1.4172979593276978, "learning_rate": 5.252373272018885e-09, "loss": 0.4862, "num_input_tokens_seen": 53847472, "step": 39740 }, { "epoch": 1.2721656744126497, "grad_norm": 0.8227820992469788, "learning_rate": 5.053067220925356e-09, "loss": 0.4688, "num_input_tokens_seen": 53854208, "step": 39745 }, { "epoch": 1.272325715383138, "grad_norm": 2.084866762161255, "learning_rate": 4.857615704759177e-09, "loss": 0.4183, "num_input_tokens_seen": 53860848, "step": 39750 }, { "epoch": 1.2724857563536265, "grad_norm": 0.7065160274505615, "learning_rate": 4.666018753654577e-09, "loss": 0.4242, "num_input_tokens_seen": 53867712, "step": 39755 }, { "epoch": 1.272645797324115, "grad_norm": 1.9244959354400635, "learning_rate": 4.478276397162917e-09, "loss": 0.3408, "num_input_tokens_seen": 53874368, "step": 39760 }, { "epoch": 1.2728058382946035, "grad_norm": 1.7378677129745483, "learning_rate": 4.294388664233262e-09, "loss": 0.4848, "num_input_tokens_seen": 53881024, "step": 39765 }, { "epoch": 1.272965879265092, "grad_norm": 1.2606316804885864, "learning_rate": 4.114355583223484e-09, "loss": 0.3549, "num_input_tokens_seen": 53887184, "step": 39770 }, { "epoch": 1.2731259202355802, "grad_norm": 1.2296738624572754, "learning_rate": 3.9381771818974845e-09, "loss": 0.42, "num_input_tokens_seen": 53893648, "step": 39775 }, { "epoch": 1.2732859612060687, "grad_norm": 1.0672131776809692, "learning_rate": 3.765853487427973e-09, "loss": 0.4227, "num_input_tokens_seen": 53900624, "step": 39780 }, { "epoch": 1.2734460021765572, "grad_norm": 1.0422000885009766, "learning_rate": 3.5973845263825857e-09, "loss": 0.3754, "num_input_tokens_seen": 53907744, "step": 39785 }, { "epoch": 1.2736060431470455, "grad_norm": 0.8983277082443237, "learning_rate": 3.4327703247488684e-09, "loss": 0.3358, "num_input_tokens_seen": 53914800, "step": 39790 }, { "epoch": 1.273766084117534, "grad_norm": 0.6600583791732788, "learning_rate": 3.2720109079037443e-09, "loss": 0.4566, "num_input_tokens_seen": 53921712, "step": 39795 }, { "epoch": 1.2739261250880225, "grad_norm": 3.235652446746826, "learning_rate": 3.1151063006468193e-09, "loss": 0.4209, "num_input_tokens_seen": 53928784, "step": 39800 }, { "epoch": 1.2739261250880225, "eval_loss": 0.44395196437835693, "eval_runtime": 500.2604, "eval_samples_per_second": 27.758, "eval_steps_per_second": 13.879, "num_input_tokens_seen": 53928784, "step": 39800 }, { "epoch": 1.274086166058511, "grad_norm": 0.9637734889984131, "learning_rate": 2.962056527169854e-09, "loss": 0.5255, "num_input_tokens_seen": 53935552, "step": 39805 }, { "epoch": 1.2742462070289995, "grad_norm": 0.9216920733451843, "learning_rate": 2.8128616110761898e-09, "loss": 0.3353, "num_input_tokens_seen": 53942256, "step": 39810 }, { "epoch": 1.274406247999488, "grad_norm": 0.8598169684410095, "learning_rate": 2.6675215753724223e-09, "loss": 0.3467, "num_input_tokens_seen": 53948512, "step": 39815 }, { "epoch": 1.2745662889699763, "grad_norm": 0.7026599645614624, "learning_rate": 2.5260364424739557e-09, "loss": 0.2884, "num_input_tokens_seen": 53955520, "step": 39820 }, { "epoch": 1.2747263299404648, "grad_norm": 1.436268925666809, "learning_rate": 2.3884062341994475e-09, "loss": 0.4072, "num_input_tokens_seen": 53962352, "step": 39825 }, { "epoch": 1.2748863709109532, "grad_norm": 1.4573687314987183, "learning_rate": 2.25463097177081e-09, "loss": 0.4422, "num_input_tokens_seen": 53969264, "step": 39830 }, { "epoch": 1.2750464118814415, "grad_norm": 1.978093147277832, "learning_rate": 2.1247106758215397e-09, "loss": 0.3812, "num_input_tokens_seen": 53976032, "step": 39835 }, { "epoch": 1.27520645285193, "grad_norm": 1.037670373916626, "learning_rate": 1.998645366382834e-09, "loss": 0.2977, "num_input_tokens_seen": 53982976, "step": 39840 }, { "epoch": 1.2753664938224185, "grad_norm": 0.7508617043495178, "learning_rate": 1.876435062897475e-09, "loss": 0.4371, "num_input_tokens_seen": 53990112, "step": 39845 }, { "epoch": 1.275526534792907, "grad_norm": 0.7194172739982605, "learning_rate": 1.758079784211497e-09, "loss": 0.4302, "num_input_tokens_seen": 53996960, "step": 39850 }, { "epoch": 1.2756865757633955, "grad_norm": 2.9052681922912598, "learning_rate": 1.6435795485797434e-09, "loss": 0.5152, "num_input_tokens_seen": 54003616, "step": 39855 }, { "epoch": 1.275846616733884, "grad_norm": 1.250667929649353, "learning_rate": 1.5329343736547596e-09, "loss": 0.6194, "num_input_tokens_seen": 54010320, "step": 39860 }, { "epoch": 1.2760066577043723, "grad_norm": 1.7671226263046265, "learning_rate": 1.4261442765006739e-09, "loss": 0.3193, "num_input_tokens_seen": 54016752, "step": 39865 }, { "epoch": 1.2761666986748608, "grad_norm": 1.0482640266418457, "learning_rate": 1.3232092735876445e-09, "loss": 0.5354, "num_input_tokens_seen": 54023792, "step": 39870 }, { "epoch": 1.2763267396453493, "grad_norm": 0.48150429129600525, "learning_rate": 1.2241293807918607e-09, "loss": 0.3291, "num_input_tokens_seen": 54030800, "step": 39875 }, { "epoch": 1.2764867806158375, "grad_norm": 0.7945966720581055, "learning_rate": 1.128904613387216e-09, "loss": 0.4874, "num_input_tokens_seen": 54038272, "step": 39880 }, { "epoch": 1.276646821586326, "grad_norm": 1.0335273742675781, "learning_rate": 1.0375349860591853e-09, "loss": 0.6306, "num_input_tokens_seen": 54044944, "step": 39885 }, { "epoch": 1.2768068625568145, "grad_norm": 0.7753233313560486, "learning_rate": 9.5002051290205e-10, "loss": 0.2892, "num_input_tokens_seen": 54051744, "step": 39890 }, { "epoch": 1.276966903527303, "grad_norm": 0.37868359684944153, "learning_rate": 8.663612074077954e-10, "loss": 0.3713, "num_input_tokens_seen": 54058176, "step": 39895 }, { "epoch": 1.2771269444977915, "grad_norm": 0.9414956569671631, "learning_rate": 7.865570824799884e-10, "loss": 0.5137, "num_input_tokens_seen": 54065248, "step": 39900 }, { "epoch": 1.27728698546828, "grad_norm": 1.0168520212173462, "learning_rate": 7.106081504254514e-10, "loss": 0.4038, "num_input_tokens_seen": 54071648, "step": 39905 }, { "epoch": 1.2774470264387683, "grad_norm": 0.6593319773674011, "learning_rate": 6.385144229570372e-10, "loss": 0.316, "num_input_tokens_seen": 54078176, "step": 39910 }, { "epoch": 1.2776070674092568, "grad_norm": 0.7522731423377991, "learning_rate": 5.70275911190854e-10, "loss": 0.4233, "num_input_tokens_seen": 54084912, "step": 39915 }, { "epoch": 1.2777671083797453, "grad_norm": 1.1136348247528076, "learning_rate": 5.058926256490403e-10, "loss": 0.4325, "num_input_tokens_seen": 54091424, "step": 39920 }, { "epoch": 1.2779271493502335, "grad_norm": 0.8797469139099121, "learning_rate": 4.4536457626254134e-10, "loss": 0.4069, "num_input_tokens_seen": 54098000, "step": 39925 }, { "epoch": 1.278087190320722, "grad_norm": 0.5482557415962219, "learning_rate": 3.88691772365557e-10, "loss": 0.4892, "num_input_tokens_seen": 54104816, "step": 39930 }, { "epoch": 1.2782472312912105, "grad_norm": 1.0223385095596313, "learning_rate": 3.358742226955425e-10, "loss": 0.4008, "num_input_tokens_seen": 54111488, "step": 39935 }, { "epoch": 1.278407272261699, "grad_norm": 0.5215373635292053, "learning_rate": 2.8691193539875925e-10, "loss": 0.5212, "num_input_tokens_seen": 54118352, "step": 39940 }, { "epoch": 1.2785673132321875, "grad_norm": 1.4805142879486084, "learning_rate": 2.418049180274995e-10, "loss": 0.4056, "num_input_tokens_seen": 54125232, "step": 39945 }, { "epoch": 1.2787273542026758, "grad_norm": 1.8502439260482788, "learning_rate": 2.005531775373104e-10, "loss": 0.56, "num_input_tokens_seen": 54131792, "step": 39950 }, { "epoch": 1.2788873951731643, "grad_norm": 1.786207675933838, "learning_rate": 1.6315672028699435e-10, "loss": 0.4518, "num_input_tokens_seen": 54138736, "step": 39955 }, { "epoch": 1.2790474361436528, "grad_norm": 1.0367251634597778, "learning_rate": 1.2961555204693555e-10, "loss": 0.3408, "num_input_tokens_seen": 54145120, "step": 39960 }, { "epoch": 1.2792074771141413, "grad_norm": 1.041618824005127, "learning_rate": 9.992967798799768e-11, "loss": 0.4988, "num_input_tokens_seen": 54151872, "step": 39965 }, { "epoch": 1.2793675180846296, "grad_norm": 0.6577768921852112, "learning_rate": 7.409910268707521e-11, "loss": 0.3944, "num_input_tokens_seen": 54158800, "step": 39970 }, { "epoch": 1.279527559055118, "grad_norm": 0.473045289516449, "learning_rate": 5.212383012986877e-11, "loss": 0.5413, "num_input_tokens_seen": 54165648, "step": 39975 }, { "epoch": 1.2796876000256066, "grad_norm": 1.4739238023757935, "learning_rate": 3.400386370533415e-11, "loss": 0.4645, "num_input_tokens_seen": 54172192, "step": 39980 }, { "epoch": 1.279847640996095, "grad_norm": 1.0387855768203735, "learning_rate": 1.9739206205682258e-11, "loss": 0.3328, "num_input_tokens_seen": 54178880, "step": 39985 }, { "epoch": 1.2800076819665835, "grad_norm": 1.2906767129898071, "learning_rate": 9.329859829154685e-12, "loss": 0.4789, "num_input_tokens_seen": 54185392, "step": 39990 }, { "epoch": 1.2801677229370718, "grad_norm": 0.8961595892906189, "learning_rate": 2.7758261855748148e-12, "loss": 0.3449, "num_input_tokens_seen": 54192000, "step": 39995 }, { "epoch": 1.2803277639075603, "grad_norm": 1.0028530359268188, "learning_rate": 7.710628524559838e-14, "loss": 0.3773, "num_input_tokens_seen": 54198768, "step": 40000 }, { "epoch": 1.2803277639075603, "eval_loss": 0.44395196437835693, "eval_runtime": 499.9291, "eval_samples_per_second": 27.776, "eval_steps_per_second": 13.888, "num_input_tokens_seen": 54198768, "step": 40000 }, { "epoch": 1.2803277639075603, "num_input_tokens_seen": 54198768, "step": 40000, "total_flos": 2.440609962399498e+18, "train_loss": 0.4951011966511607, "train_runtime": 118588.9714, "train_samples_per_second": 1.349, "train_steps_per_second": 0.337 } ], "logging_steps": 5, "max_steps": 40000, "num_input_tokens_seen": 54198768, "num_train_epochs": 2, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 2.440609962399498e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }