{ "best_global_step": 10600, "best_metric": 0.12110266089439392, "best_model_checkpoint": "saves/lora/gemma-3-1b-it/train_multirc_1745950261/checkpoint-10600", "epoch": 6.525328330206379, "eval_steps": 200, "global_step": 40000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0008157272208173587, "grad_norm": 36.68157196044922, "learning_rate": 4.999999876629946e-05, "loss": 2.4052, "num_input_tokens_seen": 7936, "step": 5 }, { "epoch": 0.0016314544416347174, "grad_norm": 16.811506271362305, "learning_rate": 4.999999375439123e-05, "loss": 1.3686, "num_input_tokens_seen": 17088, "step": 10 }, { "epoch": 0.002447181662452076, "grad_norm": 35.64654541015625, "learning_rate": 4.9999984887169785e-05, "loss": 0.7674, "num_input_tokens_seen": 26976, "step": 15 }, { "epoch": 0.0032629088832694347, "grad_norm": 13.873591423034668, "learning_rate": 4.9999972164636506e-05, "loss": 0.5275, "num_input_tokens_seen": 37536, "step": 20 }, { "epoch": 0.004078636104086793, "grad_norm": 6.594821453094482, "learning_rate": 4.999995558679334e-05, "loss": 0.5193, "num_input_tokens_seen": 46224, "step": 25 }, { "epoch": 0.004894363324904152, "grad_norm": 11.566936492919922, "learning_rate": 4.999993515364287e-05, "loss": 0.4229, "num_input_tokens_seen": 55984, "step": 30 }, { "epoch": 0.005710090545721511, "grad_norm": 6.5197906494140625, "learning_rate": 4.999991086518822e-05, "loss": 0.2686, "num_input_tokens_seen": 65952, "step": 35 }, { "epoch": 0.0065258177665388694, "grad_norm": 6.62571907043457, "learning_rate": 4.999988272143315e-05, "loss": 0.3117, "num_input_tokens_seen": 76256, "step": 40 }, { "epoch": 0.007341544987356228, "grad_norm": 6.953311443328857, "learning_rate": 4.999985072238199e-05, "loss": 0.2426, "num_input_tokens_seen": 86080, "step": 45 }, { "epoch": 0.008157272208173586, "grad_norm": 2.746248245239258, "learning_rate": 4.999981486803969e-05, "loss": 0.2255, "num_input_tokens_seen": 96528, "step": 50 }, { "epoch": 0.008972999428990946, "grad_norm": 3.3465235233306885, "learning_rate": 4.999977515841176e-05, "loss": 0.1462, "num_input_tokens_seen": 106144, "step": 55 }, { "epoch": 0.009788726649808304, "grad_norm": 3.739912509918213, "learning_rate": 4.9999731593504344e-05, "loss": 0.2708, "num_input_tokens_seen": 115184, "step": 60 }, { "epoch": 0.010604453870625663, "grad_norm": 4.4542999267578125, "learning_rate": 4.999968417332415e-05, "loss": 0.1493, "num_input_tokens_seen": 124096, "step": 65 }, { "epoch": 0.011420181091443021, "grad_norm": 9.93581771850586, "learning_rate": 4.999963289787848e-05, "loss": 0.161, "num_input_tokens_seen": 133280, "step": 70 }, { "epoch": 0.012235908312260381, "grad_norm": 10.192604064941406, "learning_rate": 4.999957776717526e-05, "loss": 0.2371, "num_input_tokens_seen": 143072, "step": 75 }, { "epoch": 0.013051635533077739, "grad_norm": 4.58124303817749, "learning_rate": 4.9999518781222984e-05, "loss": 0.4047, "num_input_tokens_seen": 151808, "step": 80 }, { "epoch": 0.013867362753895097, "grad_norm": 1.018736720085144, "learning_rate": 4.9999455940030746e-05, "loss": 0.1763, "num_input_tokens_seen": 161184, "step": 85 }, { "epoch": 0.014683089974712456, "grad_norm": 5.487241744995117, "learning_rate": 4.999938924360824e-05, "loss": 0.1932, "num_input_tokens_seen": 171328, "step": 90 }, { "epoch": 0.015498817195529814, "grad_norm": 5.338013172149658, "learning_rate": 4.999931869196575e-05, "loss": 0.3198, "num_input_tokens_seen": 180304, "step": 95 }, { "epoch": 0.016314544416347172, "grad_norm": 5.621579647064209, "learning_rate": 4.999924428511416e-05, "loss": 0.2359, "num_input_tokens_seen": 191024, "step": 100 }, { "epoch": 0.017130271637164532, "grad_norm": 2.5466842651367188, "learning_rate": 4.999916602306494e-05, "loss": 0.2293, "num_input_tokens_seen": 201056, "step": 105 }, { "epoch": 0.01794599885798189, "grad_norm": 10.298262596130371, "learning_rate": 4.999908390583016e-05, "loss": 0.2658, "num_input_tokens_seen": 211808, "step": 110 }, { "epoch": 0.01876172607879925, "grad_norm": 6.8406877517700195, "learning_rate": 4.999899793342247e-05, "loss": 0.1653, "num_input_tokens_seen": 220128, "step": 115 }, { "epoch": 0.019577453299616607, "grad_norm": 5.468479156494141, "learning_rate": 4.999890810585516e-05, "loss": 0.2473, "num_input_tokens_seen": 229264, "step": 120 }, { "epoch": 0.020393180520433967, "grad_norm": 4.9427947998046875, "learning_rate": 4.999881442314206e-05, "loss": 0.2832, "num_input_tokens_seen": 239120, "step": 125 }, { "epoch": 0.021208907741251327, "grad_norm": 3.189035415649414, "learning_rate": 4.9998716885297617e-05, "loss": 0.1388, "num_input_tokens_seen": 248656, "step": 130 }, { "epoch": 0.022024634962068683, "grad_norm": 6.873549938201904, "learning_rate": 4.999861549233688e-05, "loss": 0.1793, "num_input_tokens_seen": 258848, "step": 135 }, { "epoch": 0.022840362182886043, "grad_norm": 8.813206672668457, "learning_rate": 4.999851024427548e-05, "loss": 0.1571, "num_input_tokens_seen": 267936, "step": 140 }, { "epoch": 0.023656089403703402, "grad_norm": 5.31882905960083, "learning_rate": 4.999840114112965e-05, "loss": 0.2039, "num_input_tokens_seen": 277456, "step": 145 }, { "epoch": 0.024471816624520762, "grad_norm": 6.896450519561768, "learning_rate": 4.999828818291621e-05, "loss": 0.2658, "num_input_tokens_seen": 286544, "step": 150 }, { "epoch": 0.025287543845338118, "grad_norm": 2.790698289871216, "learning_rate": 4.999817136965259e-05, "loss": 0.2373, "num_input_tokens_seen": 297216, "step": 155 }, { "epoch": 0.026103271066155478, "grad_norm": 3.1215009689331055, "learning_rate": 4.9998050701356794e-05, "loss": 0.2183, "num_input_tokens_seen": 307472, "step": 160 }, { "epoch": 0.026918998286972837, "grad_norm": 0.9581290483474731, "learning_rate": 4.999792617804744e-05, "loss": 0.1058, "num_input_tokens_seen": 317984, "step": 165 }, { "epoch": 0.027734725507790194, "grad_norm": 6.643110275268555, "learning_rate": 4.9997797799743724e-05, "loss": 0.3465, "num_input_tokens_seen": 328240, "step": 170 }, { "epoch": 0.028550452728607553, "grad_norm": 7.382333278656006, "learning_rate": 4.999766556646545e-05, "loss": 0.1936, "num_input_tokens_seen": 336848, "step": 175 }, { "epoch": 0.029366179949424913, "grad_norm": 8.981882095336914, "learning_rate": 4.9997529478232996e-05, "loss": 0.143, "num_input_tokens_seen": 346304, "step": 180 }, { "epoch": 0.030181907170242273, "grad_norm": 4.760029315948486, "learning_rate": 4.9997389535067365e-05, "loss": 0.2509, "num_input_tokens_seen": 356160, "step": 185 }, { "epoch": 0.03099763439105963, "grad_norm": 7.401561737060547, "learning_rate": 4.999724573699012e-05, "loss": 0.2983, "num_input_tokens_seen": 365488, "step": 190 }, { "epoch": 0.031813361611876985, "grad_norm": 2.8404643535614014, "learning_rate": 4.9997098084023457e-05, "loss": 0.0909, "num_input_tokens_seen": 375600, "step": 195 }, { "epoch": 0.032629088832694345, "grad_norm": 3.345306396484375, "learning_rate": 4.999694657619013e-05, "loss": 0.08, "num_input_tokens_seen": 385088, "step": 200 }, { "epoch": 0.032629088832694345, "eval_loss": 0.2234242558479309, "eval_runtime": 80.4354, "eval_samples_per_second": 33.878, "eval_steps_per_second": 16.945, "num_input_tokens_seen": 385088, "step": 200 }, { "epoch": 0.033444816053511704, "grad_norm": 6.926886081695557, "learning_rate": 4.999679121351352e-05, "loss": 0.2614, "num_input_tokens_seen": 395168, "step": 205 }, { "epoch": 0.034260543274329064, "grad_norm": 5.638162612915039, "learning_rate": 4.9996631996017565e-05, "loss": 0.1616, "num_input_tokens_seen": 406016, "step": 210 }, { "epoch": 0.035076270495146424, "grad_norm": 5.596492290496826, "learning_rate": 4.9996468923726835e-05, "loss": 0.1697, "num_input_tokens_seen": 415600, "step": 215 }, { "epoch": 0.03589199771596378, "grad_norm": 4.136259078979492, "learning_rate": 4.999630199666647e-05, "loss": 0.3501, "num_input_tokens_seen": 426560, "step": 220 }, { "epoch": 0.03670772493678114, "grad_norm": 10.956855773925781, "learning_rate": 4.999613121486222e-05, "loss": 0.3154, "num_input_tokens_seen": 436640, "step": 225 }, { "epoch": 0.0375234521575985, "grad_norm": 10.116739273071289, "learning_rate": 4.999595657834041e-05, "loss": 0.2006, "num_input_tokens_seen": 445712, "step": 230 }, { "epoch": 0.038339179378415855, "grad_norm": 10.550344467163086, "learning_rate": 4.999577808712798e-05, "loss": 0.2806, "num_input_tokens_seen": 454960, "step": 235 }, { "epoch": 0.039154906599233215, "grad_norm": 3.0466341972351074, "learning_rate": 4.999559574125244e-05, "loss": 0.3143, "num_input_tokens_seen": 465104, "step": 240 }, { "epoch": 0.039970633820050575, "grad_norm": 1.1682441234588623, "learning_rate": 4.9995409540741934e-05, "loss": 0.1285, "num_input_tokens_seen": 474032, "step": 245 }, { "epoch": 0.040786361040867934, "grad_norm": 5.114032745361328, "learning_rate": 4.999521948562516e-05, "loss": 0.2862, "num_input_tokens_seen": 484032, "step": 250 }, { "epoch": 0.041602088261685294, "grad_norm": 8.975369453430176, "learning_rate": 4.999502557593143e-05, "loss": 0.1644, "num_input_tokens_seen": 492704, "step": 255 }, { "epoch": 0.042417815482502653, "grad_norm": 2.955695390701294, "learning_rate": 4.999482781169066e-05, "loss": 0.154, "num_input_tokens_seen": 501584, "step": 260 }, { "epoch": 0.04323354270332001, "grad_norm": 7.104404449462891, "learning_rate": 4.9994626192933324e-05, "loss": 0.2047, "num_input_tokens_seen": 512480, "step": 265 }, { "epoch": 0.044049269924137366, "grad_norm": 4.615366458892822, "learning_rate": 4.999442071969054e-05, "loss": 0.1446, "num_input_tokens_seen": 521520, "step": 270 }, { "epoch": 0.044864997144954726, "grad_norm": 10.349061965942383, "learning_rate": 4.999421139199397e-05, "loss": 0.2746, "num_input_tokens_seen": 531568, "step": 275 }, { "epoch": 0.045680724365772085, "grad_norm": 5.679474353790283, "learning_rate": 4.999399820987592e-05, "loss": 0.2376, "num_input_tokens_seen": 541408, "step": 280 }, { "epoch": 0.046496451586589445, "grad_norm": 4.750317096710205, "learning_rate": 4.999378117336924e-05, "loss": 0.204, "num_input_tokens_seen": 551008, "step": 285 }, { "epoch": 0.047312178807406804, "grad_norm": 6.489104747772217, "learning_rate": 4.9993560282507415e-05, "loss": 0.2562, "num_input_tokens_seen": 559984, "step": 290 }, { "epoch": 0.048127906028224164, "grad_norm": 9.624720573425293, "learning_rate": 4.9993335537324495e-05, "loss": 0.3027, "num_input_tokens_seen": 569760, "step": 295 }, { "epoch": 0.048943633249041524, "grad_norm": 5.390135765075684, "learning_rate": 4.999310693785516e-05, "loss": 0.1503, "num_input_tokens_seen": 579504, "step": 300 }, { "epoch": 0.049759360469858877, "grad_norm": 2.047227144241333, "learning_rate": 4.9992874484134653e-05, "loss": 0.1521, "num_input_tokens_seen": 590624, "step": 305 }, { "epoch": 0.050575087690676236, "grad_norm": 4.443001747131348, "learning_rate": 4.999263817619882e-05, "loss": 0.1643, "num_input_tokens_seen": 599520, "step": 310 }, { "epoch": 0.051390814911493596, "grad_norm": 2.327409029006958, "learning_rate": 4.9992398014084105e-05, "loss": 0.1998, "num_input_tokens_seen": 609808, "step": 315 }, { "epoch": 0.052206542132310955, "grad_norm": 1.611852765083313, "learning_rate": 4.999215399782754e-05, "loss": 0.118, "num_input_tokens_seen": 619856, "step": 320 }, { "epoch": 0.053022269353128315, "grad_norm": 6.84437370300293, "learning_rate": 4.999190612746675e-05, "loss": 0.1491, "num_input_tokens_seen": 629152, "step": 325 }, { "epoch": 0.053837996573945675, "grad_norm": 5.721476078033447, "learning_rate": 4.999165440303998e-05, "loss": 0.1131, "num_input_tokens_seen": 637904, "step": 330 }, { "epoch": 0.054653723794763034, "grad_norm": 7.3426971435546875, "learning_rate": 4.999139882458603e-05, "loss": 0.2898, "num_input_tokens_seen": 647856, "step": 335 }, { "epoch": 0.05546945101558039, "grad_norm": 3.2493183612823486, "learning_rate": 4.9991139392144314e-05, "loss": 0.1406, "num_input_tokens_seen": 657456, "step": 340 }, { "epoch": 0.05628517823639775, "grad_norm": 8.567588806152344, "learning_rate": 4.999087610575485e-05, "loss": 0.3754, "num_input_tokens_seen": 665968, "step": 345 }, { "epoch": 0.057100905457215106, "grad_norm": 5.963621139526367, "learning_rate": 4.999060896545824e-05, "loss": 0.1411, "num_input_tokens_seen": 675184, "step": 350 }, { "epoch": 0.057916632678032466, "grad_norm": 2.8033437728881836, "learning_rate": 4.999033797129568e-05, "loss": 0.1409, "num_input_tokens_seen": 685952, "step": 355 }, { "epoch": 0.058732359898849826, "grad_norm": 8.553593635559082, "learning_rate": 4.999006312330894e-05, "loss": 0.3955, "num_input_tokens_seen": 694960, "step": 360 }, { "epoch": 0.059548087119667185, "grad_norm": 1.1024402379989624, "learning_rate": 4.998978442154043e-05, "loss": 0.1654, "num_input_tokens_seen": 705840, "step": 365 }, { "epoch": 0.060363814340484545, "grad_norm": 5.333893299102783, "learning_rate": 4.9989501866033125e-05, "loss": 0.1621, "num_input_tokens_seen": 715648, "step": 370 }, { "epoch": 0.0611795415613019, "grad_norm": 10.682958602905273, "learning_rate": 4.998921545683059e-05, "loss": 0.1704, "num_input_tokens_seen": 725472, "step": 375 }, { "epoch": 0.06199526878211926, "grad_norm": 2.851257085800171, "learning_rate": 4.9988925193976996e-05, "loss": 0.1311, "num_input_tokens_seen": 734848, "step": 380 }, { "epoch": 0.06281099600293662, "grad_norm": 6.416214942932129, "learning_rate": 4.998863107751711e-05, "loss": 0.2051, "num_input_tokens_seen": 743776, "step": 385 }, { "epoch": 0.06362672322375397, "grad_norm": 4.887044429779053, "learning_rate": 4.998833310749629e-05, "loss": 0.354, "num_input_tokens_seen": 752160, "step": 390 }, { "epoch": 0.06444245044457134, "grad_norm": 7.312385559082031, "learning_rate": 4.998803128396047e-05, "loss": 0.2271, "num_input_tokens_seen": 760944, "step": 395 }, { "epoch": 0.06525817766538869, "grad_norm": 3.450441360473633, "learning_rate": 4.9987725606956215e-05, "loss": 0.1261, "num_input_tokens_seen": 770352, "step": 400 }, { "epoch": 0.06525817766538869, "eval_loss": 0.19892781972885132, "eval_runtime": 80.5196, "eval_samples_per_second": 33.843, "eval_steps_per_second": 16.928, "num_input_tokens_seen": 770352, "step": 400 }, { "epoch": 0.06607390488620606, "grad_norm": 6.1866936683654785, "learning_rate": 4.998741607653066e-05, "loss": 0.1671, "num_input_tokens_seen": 779504, "step": 405 }, { "epoch": 0.06688963210702341, "grad_norm": 0.6360821723937988, "learning_rate": 4.9987102692731523e-05, "loss": 0.1209, "num_input_tokens_seen": 789264, "step": 410 }, { "epoch": 0.06770535932784078, "grad_norm": 4.589506149291992, "learning_rate": 4.9986785455607157e-05, "loss": 0.4041, "num_input_tokens_seen": 799248, "step": 415 }, { "epoch": 0.06852108654865813, "grad_norm": 4.507879257202148, "learning_rate": 4.9986464365206456e-05, "loss": 0.0997, "num_input_tokens_seen": 809456, "step": 420 }, { "epoch": 0.06933681376947548, "grad_norm": 2.0600619316101074, "learning_rate": 4.9986139421578956e-05, "loss": 0.1324, "num_input_tokens_seen": 818848, "step": 425 }, { "epoch": 0.07015254099029285, "grad_norm": 6.2990217208862305, "learning_rate": 4.998581062477477e-05, "loss": 0.1638, "num_input_tokens_seen": 828656, "step": 430 }, { "epoch": 0.0709682682111102, "grad_norm": 0.21380572021007538, "learning_rate": 4.998547797484458e-05, "loss": 0.1425, "num_input_tokens_seen": 838576, "step": 435 }, { "epoch": 0.07178399543192757, "grad_norm": 3.0877318382263184, "learning_rate": 4.9985141471839706e-05, "loss": 0.3307, "num_input_tokens_seen": 847760, "step": 440 }, { "epoch": 0.07259972265274492, "grad_norm": 11.062613487243652, "learning_rate": 4.998480111581203e-05, "loss": 0.2363, "num_input_tokens_seen": 857152, "step": 445 }, { "epoch": 0.07341544987356229, "grad_norm": 4.363058090209961, "learning_rate": 4.998445690681405e-05, "loss": 0.2565, "num_input_tokens_seen": 866784, "step": 450 }, { "epoch": 0.07423117709437964, "grad_norm": 3.6255016326904297, "learning_rate": 4.9984108844898834e-05, "loss": 0.1983, "num_input_tokens_seen": 876960, "step": 455 }, { "epoch": 0.075046904315197, "grad_norm": 5.972115993499756, "learning_rate": 4.9983756930120076e-05, "loss": 0.1215, "num_input_tokens_seen": 886144, "step": 460 }, { "epoch": 0.07586263153601436, "grad_norm": 2.5401713848114014, "learning_rate": 4.9983401162532025e-05, "loss": 0.2265, "num_input_tokens_seen": 896528, "step": 465 }, { "epoch": 0.07667835875683171, "grad_norm": 4.733911037445068, "learning_rate": 4.998304154218955e-05, "loss": 0.137, "num_input_tokens_seen": 907168, "step": 470 }, { "epoch": 0.07749408597764908, "grad_norm": 7.0521135330200195, "learning_rate": 4.998267806914812e-05, "loss": 0.3018, "num_input_tokens_seen": 916656, "step": 475 }, { "epoch": 0.07830981319846643, "grad_norm": 3.367258310317993, "learning_rate": 4.998231074346378e-05, "loss": 0.1544, "num_input_tokens_seen": 925648, "step": 480 }, { "epoch": 0.0791255404192838, "grad_norm": 3.9131665229797363, "learning_rate": 4.998193956519317e-05, "loss": 0.2134, "num_input_tokens_seen": 936096, "step": 485 }, { "epoch": 0.07994126764010115, "grad_norm": 4.922147274017334, "learning_rate": 4.9981564534393545e-05, "loss": 0.257, "num_input_tokens_seen": 946464, "step": 490 }, { "epoch": 0.08075699486091852, "grad_norm": 4.784729480743408, "learning_rate": 4.998118565112272e-05, "loss": 0.2332, "num_input_tokens_seen": 954448, "step": 495 }, { "epoch": 0.08157272208173587, "grad_norm": 2.8884408473968506, "learning_rate": 4.998080291543914e-05, "loss": 0.1084, "num_input_tokens_seen": 964720, "step": 500 }, { "epoch": 0.08238844930255322, "grad_norm": 5.298649787902832, "learning_rate": 4.9980416327401826e-05, "loss": 0.1389, "num_input_tokens_seen": 974240, "step": 505 }, { "epoch": 0.08320417652337059, "grad_norm": 4.631773471832275, "learning_rate": 4.998002588707038e-05, "loss": 0.1564, "num_input_tokens_seen": 983152, "step": 510 }, { "epoch": 0.08401990374418794, "grad_norm": 2.560546875, "learning_rate": 4.997963159450503e-05, "loss": 0.1981, "num_input_tokens_seen": 993216, "step": 515 }, { "epoch": 0.08483563096500531, "grad_norm": 2.507143497467041, "learning_rate": 4.9979233449766575e-05, "loss": 0.2121, "num_input_tokens_seen": 1002480, "step": 520 }, { "epoch": 0.08565135818582266, "grad_norm": 1.597779631614685, "learning_rate": 4.997883145291641e-05, "loss": 0.0921, "num_input_tokens_seen": 1012192, "step": 525 }, { "epoch": 0.08646708540664003, "grad_norm": 3.792372465133667, "learning_rate": 4.9978425604016536e-05, "loss": 0.2231, "num_input_tokens_seen": 1021824, "step": 530 }, { "epoch": 0.08728281262745738, "grad_norm": 6.455567359924316, "learning_rate": 4.9978015903129536e-05, "loss": 0.2767, "num_input_tokens_seen": 1032384, "step": 535 }, { "epoch": 0.08809853984827473, "grad_norm": 7.154489994049072, "learning_rate": 4.997760235031859e-05, "loss": 0.227, "num_input_tokens_seen": 1041888, "step": 540 }, { "epoch": 0.0889142670690921, "grad_norm": 3.439760446548462, "learning_rate": 4.9977184945647473e-05, "loss": 0.1751, "num_input_tokens_seen": 1052032, "step": 545 }, { "epoch": 0.08972999428990945, "grad_norm": 3.6011199951171875, "learning_rate": 4.997676368918055e-05, "loss": 0.2559, "num_input_tokens_seen": 1063104, "step": 550 }, { "epoch": 0.09054572151072682, "grad_norm": 5.210806369781494, "learning_rate": 4.9976338580982794e-05, "loss": 0.1594, "num_input_tokens_seen": 1074016, "step": 555 }, { "epoch": 0.09136144873154417, "grad_norm": 1.724695086479187, "learning_rate": 4.9975909621119755e-05, "loss": 0.1242, "num_input_tokens_seen": 1083776, "step": 560 }, { "epoch": 0.09217717595236154, "grad_norm": 7.516482830047607, "learning_rate": 4.997547680965758e-05, "loss": 0.2617, "num_input_tokens_seen": 1092896, "step": 565 }, { "epoch": 0.09299290317317889, "grad_norm": 7.74547004699707, "learning_rate": 4.997504014666302e-05, "loss": 0.2124, "num_input_tokens_seen": 1101920, "step": 570 }, { "epoch": 0.09380863039399624, "grad_norm": 5.989716529846191, "learning_rate": 4.997459963220342e-05, "loss": 0.2258, "num_input_tokens_seen": 1111904, "step": 575 }, { "epoch": 0.09462435761481361, "grad_norm": 3.4793660640716553, "learning_rate": 4.997415526634671e-05, "loss": 0.145, "num_input_tokens_seen": 1121744, "step": 580 }, { "epoch": 0.09544008483563096, "grad_norm": 1.3984707593917847, "learning_rate": 4.99737070491614e-05, "loss": 0.2283, "num_input_tokens_seen": 1130400, "step": 585 }, { "epoch": 0.09625581205644833, "grad_norm": 1.713973879814148, "learning_rate": 4.997325498071663e-05, "loss": 0.1617, "num_input_tokens_seen": 1140848, "step": 590 }, { "epoch": 0.09707153927726568, "grad_norm": 1.838117003440857, "learning_rate": 4.997279906108211e-05, "loss": 0.2155, "num_input_tokens_seen": 1151008, "step": 595 }, { "epoch": 0.09788726649808305, "grad_norm": 3.036022663116455, "learning_rate": 4.9972339290328155e-05, "loss": 0.1522, "num_input_tokens_seen": 1160480, "step": 600 }, { "epoch": 0.09788726649808305, "eval_loss": 0.1722775250673294, "eval_runtime": 80.4479, "eval_samples_per_second": 33.873, "eval_steps_per_second": 16.943, "num_input_tokens_seen": 1160480, "step": 600 }, { "epoch": 0.0987029937189004, "grad_norm": 4.956845283508301, "learning_rate": 4.9971875668525646e-05, "loss": 0.1845, "num_input_tokens_seen": 1170144, "step": 605 }, { "epoch": 0.09951872093971775, "grad_norm": 6.79923677444458, "learning_rate": 4.997140819574609e-05, "loss": 0.1687, "num_input_tokens_seen": 1180064, "step": 610 }, { "epoch": 0.10033444816053512, "grad_norm": 4.365787506103516, "learning_rate": 4.997093687206159e-05, "loss": 0.2332, "num_input_tokens_seen": 1190224, "step": 615 }, { "epoch": 0.10115017538135247, "grad_norm": 3.6207454204559326, "learning_rate": 4.997046169754482e-05, "loss": 0.158, "num_input_tokens_seen": 1199504, "step": 620 }, { "epoch": 0.10196590260216984, "grad_norm": 0.9195646047592163, "learning_rate": 4.996998267226905e-05, "loss": 0.095, "num_input_tokens_seen": 1209120, "step": 625 }, { "epoch": 0.10278162982298719, "grad_norm": 2.343541383743286, "learning_rate": 4.996949979630817e-05, "loss": 0.1523, "num_input_tokens_seen": 1218432, "step": 630 }, { "epoch": 0.10359735704380456, "grad_norm": 2.1696057319641113, "learning_rate": 4.996901306973663e-05, "loss": 0.2523, "num_input_tokens_seen": 1227376, "step": 635 }, { "epoch": 0.10441308426462191, "grad_norm": 0.4341140389442444, "learning_rate": 4.996852249262949e-05, "loss": 0.1068, "num_input_tokens_seen": 1235536, "step": 640 }, { "epoch": 0.10522881148543926, "grad_norm": 6.386353015899658, "learning_rate": 4.996802806506241e-05, "loss": 0.2402, "num_input_tokens_seen": 1245152, "step": 645 }, { "epoch": 0.10604453870625663, "grad_norm": 3.0384490489959717, "learning_rate": 4.996752978711164e-05, "loss": 0.143, "num_input_tokens_seen": 1254768, "step": 650 }, { "epoch": 0.10686026592707398, "grad_norm": 5.93949556350708, "learning_rate": 4.996702765885401e-05, "loss": 0.2261, "num_input_tokens_seen": 1264864, "step": 655 }, { "epoch": 0.10767599314789135, "grad_norm": 4.585086822509766, "learning_rate": 4.9966521680366964e-05, "loss": 0.2615, "num_input_tokens_seen": 1273696, "step": 660 }, { "epoch": 0.1084917203687087, "grad_norm": 8.751458168029785, "learning_rate": 4.9966011851728524e-05, "loss": 0.2185, "num_input_tokens_seen": 1281904, "step": 665 }, { "epoch": 0.10930744758952607, "grad_norm": 3.7141060829162598, "learning_rate": 4.996549817301731e-05, "loss": 0.257, "num_input_tokens_seen": 1292144, "step": 670 }, { "epoch": 0.11012317481034342, "grad_norm": 1.6052815914154053, "learning_rate": 4.9964980644312544e-05, "loss": 0.1283, "num_input_tokens_seen": 1302000, "step": 675 }, { "epoch": 0.11093890203116077, "grad_norm": 4.7805256843566895, "learning_rate": 4.996445926569403e-05, "loss": 0.2018, "num_input_tokens_seen": 1312016, "step": 680 }, { "epoch": 0.11175462925197814, "grad_norm": 1.548027515411377, "learning_rate": 4.996393403724218e-05, "loss": 0.1954, "num_input_tokens_seen": 1320928, "step": 685 }, { "epoch": 0.1125703564727955, "grad_norm": 3.0099217891693115, "learning_rate": 4.9963404959037985e-05, "loss": 0.285, "num_input_tokens_seen": 1330464, "step": 690 }, { "epoch": 0.11338608369361286, "grad_norm": 7.591726303100586, "learning_rate": 4.996287203116303e-05, "loss": 0.2239, "num_input_tokens_seen": 1340896, "step": 695 }, { "epoch": 0.11420181091443021, "grad_norm": 1.0580211877822876, "learning_rate": 4.996233525369951e-05, "loss": 0.1688, "num_input_tokens_seen": 1350128, "step": 700 }, { "epoch": 0.11501753813524758, "grad_norm": 3.0852243900299072, "learning_rate": 4.99617946267302e-05, "loss": 0.1853, "num_input_tokens_seen": 1359552, "step": 705 }, { "epoch": 0.11583326535606493, "grad_norm": 5.4888458251953125, "learning_rate": 4.996125015033846e-05, "loss": 0.1948, "num_input_tokens_seen": 1370448, "step": 710 }, { "epoch": 0.11664899257688228, "grad_norm": 1.9162993431091309, "learning_rate": 4.996070182460827e-05, "loss": 0.1294, "num_input_tokens_seen": 1380528, "step": 715 }, { "epoch": 0.11746471979769965, "grad_norm": 0.8164911270141602, "learning_rate": 4.996014964962418e-05, "loss": 0.2038, "num_input_tokens_seen": 1389328, "step": 720 }, { "epoch": 0.118280447018517, "grad_norm": 0.6596381068229675, "learning_rate": 4.9959593625471344e-05, "loss": 0.1001, "num_input_tokens_seen": 1398944, "step": 725 }, { "epoch": 0.11909617423933437, "grad_norm": 5.215538024902344, "learning_rate": 4.995903375223552e-05, "loss": 0.4022, "num_input_tokens_seen": 1407728, "step": 730 }, { "epoch": 0.11991190146015172, "grad_norm": 4.466455459594727, "learning_rate": 4.995847003000302e-05, "loss": 0.2167, "num_input_tokens_seen": 1417744, "step": 735 }, { "epoch": 0.12072762868096909, "grad_norm": 2.4763951301574707, "learning_rate": 4.9957902458860804e-05, "loss": 0.2059, "num_input_tokens_seen": 1428016, "step": 740 }, { "epoch": 0.12154335590178644, "grad_norm": 5.920817852020264, "learning_rate": 4.995733103889639e-05, "loss": 0.2557, "num_input_tokens_seen": 1438784, "step": 745 }, { "epoch": 0.1223590831226038, "grad_norm": 4.216243743896484, "learning_rate": 4.99567557701979e-05, "loss": 0.2066, "num_input_tokens_seen": 1449120, "step": 750 }, { "epoch": 0.12317481034342116, "grad_norm": 1.9662518501281738, "learning_rate": 4.995617665285403e-05, "loss": 0.1827, "num_input_tokens_seen": 1459312, "step": 755 }, { "epoch": 0.12399053756423851, "grad_norm": 3.6196577548980713, "learning_rate": 4.99555936869541e-05, "loss": 0.2396, "num_input_tokens_seen": 1469744, "step": 760 }, { "epoch": 0.12480626478505588, "grad_norm": 3.8358724117279053, "learning_rate": 4.995500687258803e-05, "loss": 0.1619, "num_input_tokens_seen": 1479376, "step": 765 }, { "epoch": 0.12562199200587323, "grad_norm": 3.0689830780029297, "learning_rate": 4.995441620984628e-05, "loss": 0.152, "num_input_tokens_seen": 1488816, "step": 770 }, { "epoch": 0.1264377192266906, "grad_norm": 5.028448104858398, "learning_rate": 4.995382169881996e-05, "loss": 0.3288, "num_input_tokens_seen": 1497056, "step": 775 }, { "epoch": 0.12725344644750794, "grad_norm": 2.5537209510803223, "learning_rate": 4.9953223339600755e-05, "loss": 0.1297, "num_input_tokens_seen": 1506304, "step": 780 }, { "epoch": 0.12806917366832532, "grad_norm": 2.013659715652466, "learning_rate": 4.995262113228091e-05, "loss": 0.2263, "num_input_tokens_seen": 1515328, "step": 785 }, { "epoch": 0.12888490088914267, "grad_norm": 2.302478790283203, "learning_rate": 4.995201507695332e-05, "loss": 0.222, "num_input_tokens_seen": 1525440, "step": 790 }, { "epoch": 0.12970062810996003, "grad_norm": 2.5594420433044434, "learning_rate": 4.995140517371144e-05, "loss": 0.1505, "num_input_tokens_seen": 1533792, "step": 795 }, { "epoch": 0.13051635533077738, "grad_norm": 2.6681861877441406, "learning_rate": 4.995079142264932e-05, "loss": 0.1221, "num_input_tokens_seen": 1543296, "step": 800 }, { "epoch": 0.13051635533077738, "eval_loss": 0.17393815517425537, "eval_runtime": 80.6323, "eval_samples_per_second": 33.795, "eval_steps_per_second": 16.904, "num_input_tokens_seen": 1543296, "step": 800 }, { "epoch": 0.13133208255159476, "grad_norm": 2.3631277084350586, "learning_rate": 4.995017382386162e-05, "loss": 0.1947, "num_input_tokens_seen": 1553120, "step": 805 }, { "epoch": 0.1321478097724121, "grad_norm": 5.90090799331665, "learning_rate": 4.994955237744356e-05, "loss": 0.2138, "num_input_tokens_seen": 1562320, "step": 810 }, { "epoch": 0.13296353699322946, "grad_norm": 3.2624714374542236, "learning_rate": 4.994892708349101e-05, "loss": 0.1636, "num_input_tokens_seen": 1572624, "step": 815 }, { "epoch": 0.13377926421404682, "grad_norm": 2.842169761657715, "learning_rate": 4.994829794210035e-05, "loss": 0.1175, "num_input_tokens_seen": 1582400, "step": 820 }, { "epoch": 0.13459499143486417, "grad_norm": 1.1115463972091675, "learning_rate": 4.994766495336864e-05, "loss": 0.0782, "num_input_tokens_seen": 1591136, "step": 825 }, { "epoch": 0.13541071865568155, "grad_norm": 2.9647607803344727, "learning_rate": 4.994702811739348e-05, "loss": 0.103, "num_input_tokens_seen": 1600544, "step": 830 }, { "epoch": 0.1362264458764989, "grad_norm": 4.31200647354126, "learning_rate": 4.994638743427308e-05, "loss": 0.2403, "num_input_tokens_seen": 1609968, "step": 835 }, { "epoch": 0.13704217309731626, "grad_norm": 2.8722476959228516, "learning_rate": 4.994574290410624e-05, "loss": 0.1422, "num_input_tokens_seen": 1619232, "step": 840 }, { "epoch": 0.1378579003181336, "grad_norm": 2.6737687587738037, "learning_rate": 4.9945094526992364e-05, "loss": 0.268, "num_input_tokens_seen": 1628448, "step": 845 }, { "epoch": 0.13867362753895096, "grad_norm": 2.7302610874176025, "learning_rate": 4.994444230303142e-05, "loss": 0.1572, "num_input_tokens_seen": 1639536, "step": 850 }, { "epoch": 0.13948935475976834, "grad_norm": 2.8625166416168213, "learning_rate": 4.994378623232402e-05, "loss": 0.1568, "num_input_tokens_seen": 1649840, "step": 855 }, { "epoch": 0.1403050819805857, "grad_norm": 0.8174279928207397, "learning_rate": 4.99431263149713e-05, "loss": 0.1522, "num_input_tokens_seen": 1659552, "step": 860 }, { "epoch": 0.14112080920140305, "grad_norm": 4.568795680999756, "learning_rate": 4.9942462551075056e-05, "loss": 0.1579, "num_input_tokens_seen": 1668592, "step": 865 }, { "epoch": 0.1419365364222204, "grad_norm": 2.652386426925659, "learning_rate": 4.994179494073764e-05, "loss": 0.2029, "num_input_tokens_seen": 1679968, "step": 870 }, { "epoch": 0.14275226364303778, "grad_norm": 3.8700335025787354, "learning_rate": 4.9941123484062e-05, "loss": 0.1563, "num_input_tokens_seen": 1691664, "step": 875 }, { "epoch": 0.14356799086385513, "grad_norm": 1.116204023361206, "learning_rate": 4.99404481811517e-05, "loss": 0.1187, "num_input_tokens_seen": 1701088, "step": 880 }, { "epoch": 0.14438371808467249, "grad_norm": 2.5078821182250977, "learning_rate": 4.9939769032110864e-05, "loss": 0.1769, "num_input_tokens_seen": 1709904, "step": 885 }, { "epoch": 0.14519944530548984, "grad_norm": 10.270109176635742, "learning_rate": 4.993908603704423e-05, "loss": 0.2754, "num_input_tokens_seen": 1719072, "step": 890 }, { "epoch": 0.1460151725263072, "grad_norm": 0.5942384600639343, "learning_rate": 4.9938399196057126e-05, "loss": 0.091, "num_input_tokens_seen": 1727680, "step": 895 }, { "epoch": 0.14683089974712457, "grad_norm": 5.059907913208008, "learning_rate": 4.993770850925547e-05, "loss": 0.2754, "num_input_tokens_seen": 1736832, "step": 900 }, { "epoch": 0.14764662696794192, "grad_norm": 0.9482457637786865, "learning_rate": 4.993701397674577e-05, "loss": 0.1225, "num_input_tokens_seen": 1747056, "step": 905 }, { "epoch": 0.14846235418875928, "grad_norm": 0.23672747611999512, "learning_rate": 4.993631559863515e-05, "loss": 0.1842, "num_input_tokens_seen": 1758064, "step": 910 }, { "epoch": 0.14927808140957663, "grad_norm": 3.2651476860046387, "learning_rate": 4.9935613375031283e-05, "loss": 0.2701, "num_input_tokens_seen": 1767616, "step": 915 }, { "epoch": 0.150093808630394, "grad_norm": 0.6000333428382874, "learning_rate": 4.993490730604248e-05, "loss": 0.1472, "num_input_tokens_seen": 1776752, "step": 920 }, { "epoch": 0.15090953585121136, "grad_norm": 5.554628849029541, "learning_rate": 4.993419739177761e-05, "loss": 0.1796, "num_input_tokens_seen": 1786800, "step": 925 }, { "epoch": 0.15172526307202872, "grad_norm": 2.004424571990967, "learning_rate": 4.9933483632346164e-05, "loss": 0.187, "num_input_tokens_seen": 1796816, "step": 930 }, { "epoch": 0.15254099029284607, "grad_norm": 12.424083709716797, "learning_rate": 4.993276602785821e-05, "loss": 0.2038, "num_input_tokens_seen": 1807472, "step": 935 }, { "epoch": 0.15335671751366342, "grad_norm": 2.7192342281341553, "learning_rate": 4.993204457842441e-05, "loss": 0.1414, "num_input_tokens_seen": 1816944, "step": 940 }, { "epoch": 0.1541724447344808, "grad_norm": 3.58933162689209, "learning_rate": 4.993131928415602e-05, "loss": 0.1308, "num_input_tokens_seen": 1826208, "step": 945 }, { "epoch": 0.15498817195529815, "grad_norm": 4.046415328979492, "learning_rate": 4.993059014516489e-05, "loss": 0.3003, "num_input_tokens_seen": 1836544, "step": 950 }, { "epoch": 0.1558038991761155, "grad_norm": 2.476475715637207, "learning_rate": 4.9929857161563464e-05, "loss": 0.1475, "num_input_tokens_seen": 1845696, "step": 955 }, { "epoch": 0.15661962639693286, "grad_norm": 5.26346492767334, "learning_rate": 4.992912033346477e-05, "loss": 0.1477, "num_input_tokens_seen": 1856832, "step": 960 }, { "epoch": 0.1574353536177502, "grad_norm": 0.6886953115463257, "learning_rate": 4.992837966098245e-05, "loss": 0.0967, "num_input_tokens_seen": 1865776, "step": 965 }, { "epoch": 0.1582510808385676, "grad_norm": 3.675825595855713, "learning_rate": 4.992763514423071e-05, "loss": 0.1087, "num_input_tokens_seen": 1874704, "step": 970 }, { "epoch": 0.15906680805938495, "grad_norm": 4.94442081451416, "learning_rate": 4.992688678332437e-05, "loss": 0.2848, "num_input_tokens_seen": 1884384, "step": 975 }, { "epoch": 0.1598825352802023, "grad_norm": 1.190104365348816, "learning_rate": 4.992613457837884e-05, "loss": 0.0531, "num_input_tokens_seen": 1894160, "step": 980 }, { "epoch": 0.16069826250101965, "grad_norm": 5.538182258605957, "learning_rate": 4.992537852951011e-05, "loss": 0.1993, "num_input_tokens_seen": 1903488, "step": 985 }, { "epoch": 0.16151398972183703, "grad_norm": 2.3608195781707764, "learning_rate": 4.9924618636834785e-05, "loss": 0.3059, "num_input_tokens_seen": 1913728, "step": 990 }, { "epoch": 0.16232971694265438, "grad_norm": 3.2152085304260254, "learning_rate": 4.9923854900470046e-05, "loss": 0.1894, "num_input_tokens_seen": 1922976, "step": 995 }, { "epoch": 0.16314544416347174, "grad_norm": 6.828896999359131, "learning_rate": 4.992308732053367e-05, "loss": 0.1766, "num_input_tokens_seen": 1931808, "step": 1000 }, { "epoch": 0.16314544416347174, "eval_loss": 0.17772261798381805, "eval_runtime": 80.5829, "eval_samples_per_second": 33.816, "eval_steps_per_second": 16.914, "num_input_tokens_seen": 1931808, "step": 1000 }, { "epoch": 0.1639611713842891, "grad_norm": 2.164947271347046, "learning_rate": 4.992231589714402e-05, "loss": 0.1628, "num_input_tokens_seen": 1942560, "step": 1005 }, { "epoch": 0.16477689860510644, "grad_norm": 4.1967058181762695, "learning_rate": 4.992154063042007e-05, "loss": 0.1643, "num_input_tokens_seen": 1950912, "step": 1010 }, { "epoch": 0.16559262582592382, "grad_norm": 0.5201284289360046, "learning_rate": 4.992076152048136e-05, "loss": 0.1076, "num_input_tokens_seen": 1961696, "step": 1015 }, { "epoch": 0.16640835304674118, "grad_norm": 6.313735008239746, "learning_rate": 4.991997856744807e-05, "loss": 0.2818, "num_input_tokens_seen": 1971280, "step": 1020 }, { "epoch": 0.16722408026755853, "grad_norm": 3.7770400047302246, "learning_rate": 4.9919191771440905e-05, "loss": 0.1464, "num_input_tokens_seen": 1981200, "step": 1025 }, { "epoch": 0.16803980748837588, "grad_norm": 3.841866970062256, "learning_rate": 4.991840113258122e-05, "loss": 0.1527, "num_input_tokens_seen": 1990400, "step": 1030 }, { "epoch": 0.16885553470919323, "grad_norm": 2.817147970199585, "learning_rate": 4.9917606650990933e-05, "loss": 0.1042, "num_input_tokens_seen": 2000016, "step": 1035 }, { "epoch": 0.16967126193001061, "grad_norm": 4.806766033172607, "learning_rate": 4.9916808326792566e-05, "loss": 0.1441, "num_input_tokens_seen": 2009600, "step": 1040 }, { "epoch": 0.17048698915082797, "grad_norm": 2.89398193359375, "learning_rate": 4.9916006160109235e-05, "loss": 0.1776, "num_input_tokens_seen": 2019696, "step": 1045 }, { "epoch": 0.17130271637164532, "grad_norm": 7.507143497467041, "learning_rate": 4.991520015106464e-05, "loss": 0.2092, "num_input_tokens_seen": 2029776, "step": 1050 }, { "epoch": 0.17211844359246267, "grad_norm": 4.893310070037842, "learning_rate": 4.991439029978308e-05, "loss": 0.1671, "num_input_tokens_seen": 2037728, "step": 1055 }, { "epoch": 0.17293417081328005, "grad_norm": 2.774866819381714, "learning_rate": 4.9913576606389434e-05, "loss": 0.2225, "num_input_tokens_seen": 2047616, "step": 1060 }, { "epoch": 0.1737498980340974, "grad_norm": 2.9137349128723145, "learning_rate": 4.991275907100919e-05, "loss": 0.2792, "num_input_tokens_seen": 2057440, "step": 1065 }, { "epoch": 0.17456562525491476, "grad_norm": 6.442518711090088, "learning_rate": 4.9911937693768434e-05, "loss": 0.2697, "num_input_tokens_seen": 2066528, "step": 1070 }, { "epoch": 0.1753813524757321, "grad_norm": 5.197111129760742, "learning_rate": 4.991111247479382e-05, "loss": 0.1894, "num_input_tokens_seen": 2076288, "step": 1075 }, { "epoch": 0.17619707969654946, "grad_norm": 5.366116523742676, "learning_rate": 4.9910283414212605e-05, "loss": 0.2011, "num_input_tokens_seen": 2085920, "step": 1080 }, { "epoch": 0.17701280691736684, "grad_norm": 4.207390785217285, "learning_rate": 4.990945051215265e-05, "loss": 0.1608, "num_input_tokens_seen": 2095344, "step": 1085 }, { "epoch": 0.1778285341381842, "grad_norm": 1.4392685890197754, "learning_rate": 4.99086137687424e-05, "loss": 0.1176, "num_input_tokens_seen": 2104928, "step": 1090 }, { "epoch": 0.17864426135900155, "grad_norm": 5.905887126922607, "learning_rate": 4.9907773184110874e-05, "loss": 0.2544, "num_input_tokens_seen": 2117072, "step": 1095 }, { "epoch": 0.1794599885798189, "grad_norm": 3.040151357650757, "learning_rate": 4.9906928758387715e-05, "loss": 0.3657, "num_input_tokens_seen": 2126640, "step": 1100 }, { "epoch": 0.18027571580063625, "grad_norm": 2.9338362216949463, "learning_rate": 4.9906080491703146e-05, "loss": 0.1765, "num_input_tokens_seen": 2136112, "step": 1105 }, { "epoch": 0.18109144302145364, "grad_norm": 3.7915303707122803, "learning_rate": 4.990522838418797e-05, "loss": 0.2221, "num_input_tokens_seen": 2145904, "step": 1110 }, { "epoch": 0.181907170242271, "grad_norm": 1.4048444032669067, "learning_rate": 4.9904372435973604e-05, "loss": 0.2205, "num_input_tokens_seen": 2157072, "step": 1115 }, { "epoch": 0.18272289746308834, "grad_norm": 5.755616664886475, "learning_rate": 4.990351264719203e-05, "loss": 0.1354, "num_input_tokens_seen": 2166624, "step": 1120 }, { "epoch": 0.1835386246839057, "grad_norm": 2.1121668815612793, "learning_rate": 4.990264901797586e-05, "loss": 0.1545, "num_input_tokens_seen": 2176224, "step": 1125 }, { "epoch": 0.18435435190472307, "grad_norm": 1.374175786972046, "learning_rate": 4.990178154845826e-05, "loss": 0.1358, "num_input_tokens_seen": 2185856, "step": 1130 }, { "epoch": 0.18517007912554043, "grad_norm": 2.858830213546753, "learning_rate": 4.9900910238773014e-05, "loss": 0.1975, "num_input_tokens_seen": 2194608, "step": 1135 }, { "epoch": 0.18598580634635778, "grad_norm": 3.4270877838134766, "learning_rate": 4.990003508905448e-05, "loss": 0.1583, "num_input_tokens_seen": 2203312, "step": 1140 }, { "epoch": 0.18680153356717513, "grad_norm": 1.1583950519561768, "learning_rate": 4.989915609943763e-05, "loss": 0.1189, "num_input_tokens_seen": 2214048, "step": 1145 }, { "epoch": 0.18761726078799248, "grad_norm": 3.0837957859039307, "learning_rate": 4.9898273270058e-05, "loss": 0.158, "num_input_tokens_seen": 2222032, "step": 1150 }, { "epoch": 0.18843298800880987, "grad_norm": 1.2459049224853516, "learning_rate": 4.989738660105174e-05, "loss": 0.1914, "num_input_tokens_seen": 2231232, "step": 1155 }, { "epoch": 0.18924871522962722, "grad_norm": 4.662827491760254, "learning_rate": 4.989649609255559e-05, "loss": 0.255, "num_input_tokens_seen": 2240128, "step": 1160 }, { "epoch": 0.19006444245044457, "grad_norm": 3.4149911403656006, "learning_rate": 4.989560174470687e-05, "loss": 0.1206, "num_input_tokens_seen": 2250016, "step": 1165 }, { "epoch": 0.19088016967126192, "grad_norm": 7.738561630249023, "learning_rate": 4.989470355764351e-05, "loss": 0.1787, "num_input_tokens_seen": 2260192, "step": 1170 }, { "epoch": 0.19169589689207928, "grad_norm": 3.8226420879364014, "learning_rate": 4.9893801531504e-05, "loss": 0.1302, "num_input_tokens_seen": 2270288, "step": 1175 }, { "epoch": 0.19251162411289666, "grad_norm": 0.977420449256897, "learning_rate": 4.9892895666427475e-05, "loss": 0.1692, "num_input_tokens_seen": 2280112, "step": 1180 }, { "epoch": 0.193327351333714, "grad_norm": 3.7066075801849365, "learning_rate": 4.9891985962553606e-05, "loss": 0.2129, "num_input_tokens_seen": 2290160, "step": 1185 }, { "epoch": 0.19414307855453136, "grad_norm": 3.8382036685943604, "learning_rate": 4.989107242002269e-05, "loss": 0.2249, "num_input_tokens_seen": 2298560, "step": 1190 }, { "epoch": 0.19495880577534871, "grad_norm": 4.517809867858887, "learning_rate": 4.989015503897561e-05, "loss": 0.0919, "num_input_tokens_seen": 2307312, "step": 1195 }, { "epoch": 0.1957745329961661, "grad_norm": 5.698427677154541, "learning_rate": 4.988923381955383e-05, "loss": 0.163, "num_input_tokens_seen": 2315744, "step": 1200 }, { "epoch": 0.1957745329961661, "eval_loss": 0.1670651137828827, "eval_runtime": 80.6495, "eval_samples_per_second": 33.788, "eval_steps_per_second": 16.9, "num_input_tokens_seen": 2315744, "step": 1200 }, { "epoch": 0.19659026021698345, "grad_norm": 4.003925800323486, "learning_rate": 4.988830876189942e-05, "loss": 0.2159, "num_input_tokens_seen": 2325456, "step": 1205 }, { "epoch": 0.1974059874378008, "grad_norm": 2.0437228679656982, "learning_rate": 4.988737986615503e-05, "loss": 0.1526, "num_input_tokens_seen": 2334768, "step": 1210 }, { "epoch": 0.19822171465861815, "grad_norm": 5.080355644226074, "learning_rate": 4.988644713246391e-05, "loss": 0.1153, "num_input_tokens_seen": 2345232, "step": 1215 }, { "epoch": 0.1990374418794355, "grad_norm": 6.327542781829834, "learning_rate": 4.988551056096991e-05, "loss": 0.1224, "num_input_tokens_seen": 2355424, "step": 1220 }, { "epoch": 0.1998531691002529, "grad_norm": 11.857685089111328, "learning_rate": 4.988457015181743e-05, "loss": 0.2857, "num_input_tokens_seen": 2364576, "step": 1225 }, { "epoch": 0.20066889632107024, "grad_norm": 4.305271625518799, "learning_rate": 4.988362590515153e-05, "loss": 0.2914, "num_input_tokens_seen": 2375488, "step": 1230 }, { "epoch": 0.2014846235418876, "grad_norm": 4.674610137939453, "learning_rate": 4.9882677821117805e-05, "loss": 0.3731, "num_input_tokens_seen": 2386272, "step": 1235 }, { "epoch": 0.20230035076270494, "grad_norm": 4.1312761306762695, "learning_rate": 4.988172589986246e-05, "loss": 0.1389, "num_input_tokens_seen": 2395952, "step": 1240 }, { "epoch": 0.2031160779835223, "grad_norm": 1.2351837158203125, "learning_rate": 4.9880770141532304e-05, "loss": 0.1366, "num_input_tokens_seen": 2404944, "step": 1245 }, { "epoch": 0.20393180520433968, "grad_norm": 1.403663992881775, "learning_rate": 4.987981054627472e-05, "loss": 0.1825, "num_input_tokens_seen": 2414656, "step": 1250 }, { "epoch": 0.20474753242515703, "grad_norm": 7.332441806793213, "learning_rate": 4.987884711423769e-05, "loss": 0.3315, "num_input_tokens_seen": 2423456, "step": 1255 }, { "epoch": 0.20556325964597438, "grad_norm": 2.7848246097564697, "learning_rate": 4.9877879845569784e-05, "loss": 0.1648, "num_input_tokens_seen": 2434368, "step": 1260 }, { "epoch": 0.20637898686679174, "grad_norm": 2.0276377201080322, "learning_rate": 4.9876908740420175e-05, "loss": 0.2243, "num_input_tokens_seen": 2444240, "step": 1265 }, { "epoch": 0.20719471408760912, "grad_norm": 2.32285737991333, "learning_rate": 4.987593379893861e-05, "loss": 0.1922, "num_input_tokens_seen": 2453456, "step": 1270 }, { "epoch": 0.20801044130842647, "grad_norm": 1.6850742101669312, "learning_rate": 4.987495502127545e-05, "loss": 0.2116, "num_input_tokens_seen": 2463504, "step": 1275 }, { "epoch": 0.20882616852924382, "grad_norm": 2.782235860824585, "learning_rate": 4.987397240758162e-05, "loss": 0.1721, "num_input_tokens_seen": 2473408, "step": 1280 }, { "epoch": 0.20964189575006117, "grad_norm": 3.5455141067504883, "learning_rate": 4.9872985958008664e-05, "loss": 0.2366, "num_input_tokens_seen": 2482832, "step": 1285 }, { "epoch": 0.21045762297087853, "grad_norm": 0.36759334802627563, "learning_rate": 4.987199567270871e-05, "loss": 0.1015, "num_input_tokens_seen": 2492720, "step": 1290 }, { "epoch": 0.2112733501916959, "grad_norm": 2.5595929622650146, "learning_rate": 4.9871001551834444e-05, "loss": 0.1498, "num_input_tokens_seen": 2503136, "step": 1295 }, { "epoch": 0.21208907741251326, "grad_norm": 4.621960163116455, "learning_rate": 4.98700035955392e-05, "loss": 0.1634, "num_input_tokens_seen": 2511568, "step": 1300 }, { "epoch": 0.2129048046333306, "grad_norm": 3.75927734375, "learning_rate": 4.986900180397686e-05, "loss": 0.1908, "num_input_tokens_seen": 2522256, "step": 1305 }, { "epoch": 0.21372053185414797, "grad_norm": 3.564652919769287, "learning_rate": 4.9867996177301926e-05, "loss": 0.3022, "num_input_tokens_seen": 2532016, "step": 1310 }, { "epoch": 0.21453625907496532, "grad_norm": 1.410338282585144, "learning_rate": 4.9866986715669464e-05, "loss": 0.1208, "num_input_tokens_seen": 2540800, "step": 1315 }, { "epoch": 0.2153519862957827, "grad_norm": 0.7918124794960022, "learning_rate": 4.9865973419235155e-05, "loss": 0.1023, "num_input_tokens_seen": 2551200, "step": 1320 }, { "epoch": 0.21616771351660005, "grad_norm": 1.4017318487167358, "learning_rate": 4.986495628815526e-05, "loss": 0.1223, "num_input_tokens_seen": 2561312, "step": 1325 }, { "epoch": 0.2169834407374174, "grad_norm": 4.7769999504089355, "learning_rate": 4.986393532258663e-05, "loss": 0.2117, "num_input_tokens_seen": 2570784, "step": 1330 }, { "epoch": 0.21779916795823476, "grad_norm": 0.5604217052459717, "learning_rate": 4.986291052268671e-05, "loss": 0.0897, "num_input_tokens_seen": 2581168, "step": 1335 }, { "epoch": 0.21861489517905214, "grad_norm": 2.5824508666992188, "learning_rate": 4.986188188861355e-05, "loss": 0.1549, "num_input_tokens_seen": 2590416, "step": 1340 }, { "epoch": 0.2194306223998695, "grad_norm": 2.8567981719970703, "learning_rate": 4.9860849420525766e-05, "loss": 0.0621, "num_input_tokens_seen": 2601056, "step": 1345 }, { "epoch": 0.22024634962068684, "grad_norm": 6.999955654144287, "learning_rate": 4.9859813118582575e-05, "loss": 0.3058, "num_input_tokens_seen": 2611152, "step": 1350 }, { "epoch": 0.2210620768415042, "grad_norm": 1.4559204578399658, "learning_rate": 4.98587729829438e-05, "loss": 0.0967, "num_input_tokens_seen": 2620240, "step": 1355 }, { "epoch": 0.22187780406232155, "grad_norm": 3.9770984649658203, "learning_rate": 4.985772901376983e-05, "loss": 0.1434, "num_input_tokens_seen": 2630912, "step": 1360 }, { "epoch": 0.22269353128313893, "grad_norm": 4.471310615539551, "learning_rate": 4.9856681211221666e-05, "loss": 0.2058, "num_input_tokens_seen": 2640560, "step": 1365 }, { "epoch": 0.22350925850395628, "grad_norm": 0.567314088344574, "learning_rate": 4.985562957546089e-05, "loss": 0.2436, "num_input_tokens_seen": 2650864, "step": 1370 }, { "epoch": 0.22432498572477363, "grad_norm": 2.344604253768921, "learning_rate": 4.9854574106649686e-05, "loss": 0.1747, "num_input_tokens_seen": 2661920, "step": 1375 }, { "epoch": 0.225140712945591, "grad_norm": 2.8438103199005127, "learning_rate": 4.985351480495081e-05, "loss": 0.142, "num_input_tokens_seen": 2671648, "step": 1380 }, { "epoch": 0.22595644016640834, "grad_norm": 1.129413366317749, "learning_rate": 4.985245167052762e-05, "loss": 0.1122, "num_input_tokens_seen": 2681520, "step": 1385 }, { "epoch": 0.22677216738722572, "grad_norm": 0.4005439281463623, "learning_rate": 4.9851384703544066e-05, "loss": 0.1388, "num_input_tokens_seen": 2690720, "step": 1390 }, { "epoch": 0.22758789460804307, "grad_norm": 3.284651279449463, "learning_rate": 4.985031390416469e-05, "loss": 0.1684, "num_input_tokens_seen": 2701280, "step": 1395 }, { "epoch": 0.22840362182886043, "grad_norm": 4.33876371383667, "learning_rate": 4.984923927255461e-05, "loss": 0.1673, "num_input_tokens_seen": 2710208, "step": 1400 }, { "epoch": 0.22840362182886043, "eval_loss": 0.15960896015167236, "eval_runtime": 80.5959, "eval_samples_per_second": 33.811, "eval_steps_per_second": 16.912, "num_input_tokens_seen": 2710208, "step": 1400 }, { "epoch": 0.22921934904967778, "grad_norm": 0.659085750579834, "learning_rate": 4.984816080887958e-05, "loss": 0.1826, "num_input_tokens_seen": 2721632, "step": 1405 }, { "epoch": 0.23003507627049516, "grad_norm": 3.2569568157196045, "learning_rate": 4.9847078513305875e-05, "loss": 0.1716, "num_input_tokens_seen": 2731008, "step": 1410 }, { "epoch": 0.2308508034913125, "grad_norm": 3.4976158142089844, "learning_rate": 4.984599238600043e-05, "loss": 0.1045, "num_input_tokens_seen": 2740128, "step": 1415 }, { "epoch": 0.23166653071212986, "grad_norm": 4.432700157165527, "learning_rate": 4.9844902427130716e-05, "loss": 0.2998, "num_input_tokens_seen": 2749328, "step": 1420 }, { "epoch": 0.23248225793294722, "grad_norm": 3.683445453643799, "learning_rate": 4.984380863686482e-05, "loss": 0.3173, "num_input_tokens_seen": 2757456, "step": 1425 }, { "epoch": 0.23329798515376457, "grad_norm": 2.9508402347564697, "learning_rate": 4.984271101537143e-05, "loss": 0.2083, "num_input_tokens_seen": 2765200, "step": 1430 }, { "epoch": 0.23411371237458195, "grad_norm": 3.7187294960021973, "learning_rate": 4.9841609562819816e-05, "loss": 0.2005, "num_input_tokens_seen": 2775072, "step": 1435 }, { "epoch": 0.2349294395953993, "grad_norm": 2.854233503341675, "learning_rate": 4.984050427937983e-05, "loss": 0.1568, "num_input_tokens_seen": 2784896, "step": 1440 }, { "epoch": 0.23574516681621666, "grad_norm": 3.1973180770874023, "learning_rate": 4.983939516522191e-05, "loss": 0.1551, "num_input_tokens_seen": 2794272, "step": 1445 }, { "epoch": 0.236560894037034, "grad_norm": 0.6271368861198425, "learning_rate": 4.983828222051711e-05, "loss": 0.1768, "num_input_tokens_seen": 2803072, "step": 1450 }, { "epoch": 0.23737662125785136, "grad_norm": 0.9552211165428162, "learning_rate": 4.983716544543705e-05, "loss": 0.1594, "num_input_tokens_seen": 2812528, "step": 1455 }, { "epoch": 0.23819234847866874, "grad_norm": 5.364340782165527, "learning_rate": 4.983604484015395e-05, "loss": 0.1251, "num_input_tokens_seen": 2822464, "step": 1460 }, { "epoch": 0.2390080756994861, "grad_norm": 3.4730679988861084, "learning_rate": 4.983492040484064e-05, "loss": 0.2847, "num_input_tokens_seen": 2832832, "step": 1465 }, { "epoch": 0.23982380292030345, "grad_norm": 0.7594687938690186, "learning_rate": 4.98337921396705e-05, "loss": 0.1608, "num_input_tokens_seen": 2844336, "step": 1470 }, { "epoch": 0.2406395301411208, "grad_norm": 4.097828388214111, "learning_rate": 4.983266004481753e-05, "loss": 0.2354, "num_input_tokens_seen": 2854176, "step": 1475 }, { "epoch": 0.24145525736193818, "grad_norm": 4.082433700561523, "learning_rate": 4.9831524120456316e-05, "loss": 0.2694, "num_input_tokens_seen": 2863888, "step": 1480 }, { "epoch": 0.24227098458275553, "grad_norm": 3.5972423553466797, "learning_rate": 4.9830384366762026e-05, "loss": 0.1234, "num_input_tokens_seen": 2872640, "step": 1485 }, { "epoch": 0.24308671180357289, "grad_norm": 6.808229923248291, "learning_rate": 4.9829240783910436e-05, "loss": 0.2559, "num_input_tokens_seen": 2881488, "step": 1490 }, { "epoch": 0.24390243902439024, "grad_norm": 3.6657040119171143, "learning_rate": 4.982809337207789e-05, "loss": 0.2511, "num_input_tokens_seen": 2892336, "step": 1495 }, { "epoch": 0.2447181662452076, "grad_norm": 5.312696933746338, "learning_rate": 4.9826942131441337e-05, "loss": 0.2359, "num_input_tokens_seen": 2901600, "step": 1500 }, { "epoch": 0.24553389346602497, "grad_norm": 1.884000301361084, "learning_rate": 4.9825787062178315e-05, "loss": 0.1698, "num_input_tokens_seen": 2911664, "step": 1505 }, { "epoch": 0.24634962068684232, "grad_norm": 3.191110372543335, "learning_rate": 4.9824628164466945e-05, "loss": 0.1691, "num_input_tokens_seen": 2920976, "step": 1510 }, { "epoch": 0.24716534790765968, "grad_norm": 3.2921102046966553, "learning_rate": 4.982346543848595e-05, "loss": 0.2693, "num_input_tokens_seen": 2930864, "step": 1515 }, { "epoch": 0.24798107512847703, "grad_norm": 2.8637707233428955, "learning_rate": 4.9822298884414626e-05, "loss": 0.1767, "num_input_tokens_seen": 2940944, "step": 1520 }, { "epoch": 0.24879680234929438, "grad_norm": 1.6360349655151367, "learning_rate": 4.982112850243288e-05, "loss": 0.155, "num_input_tokens_seen": 2950768, "step": 1525 }, { "epoch": 0.24961252957011176, "grad_norm": 3.8389902114868164, "learning_rate": 4.98199542927212e-05, "loss": 0.2191, "num_input_tokens_seen": 2959600, "step": 1530 }, { "epoch": 0.2504282567909291, "grad_norm": 1.2851710319519043, "learning_rate": 4.981877625546066e-05, "loss": 0.2092, "num_input_tokens_seen": 2970000, "step": 1535 }, { "epoch": 0.25124398401174647, "grad_norm": 2.024242877960205, "learning_rate": 4.981759439083293e-05, "loss": 0.1298, "num_input_tokens_seen": 2980640, "step": 1540 }, { "epoch": 0.2520597112325638, "grad_norm": 1.419155478477478, "learning_rate": 4.981640869902027e-05, "loss": 0.1309, "num_input_tokens_seen": 2990448, "step": 1545 }, { "epoch": 0.2528754384533812, "grad_norm": 1.652171015739441, "learning_rate": 4.9815219180205517e-05, "loss": 0.1576, "num_input_tokens_seen": 2999568, "step": 1550 }, { "epoch": 0.2536911656741985, "grad_norm": 1.9787511825561523, "learning_rate": 4.9814025834572126e-05, "loss": 0.1545, "num_input_tokens_seen": 3009136, "step": 1555 }, { "epoch": 0.2545068928950159, "grad_norm": 5.866007328033447, "learning_rate": 4.981282866230411e-05, "loss": 0.0864, "num_input_tokens_seen": 3018272, "step": 1560 }, { "epoch": 0.2553226201158333, "grad_norm": 3.1259052753448486, "learning_rate": 4.981162766358611e-05, "loss": 0.3156, "num_input_tokens_seen": 3027856, "step": 1565 }, { "epoch": 0.25613834733665064, "grad_norm": 2.057162046432495, "learning_rate": 4.9810422838603316e-05, "loss": 0.1742, "num_input_tokens_seen": 3037584, "step": 1570 }, { "epoch": 0.256954074557468, "grad_norm": 0.5758703351020813, "learning_rate": 4.9809214187541533e-05, "loss": 0.0834, "num_input_tokens_seen": 3045856, "step": 1575 }, { "epoch": 0.25776980177828535, "grad_norm": 0.2862725555896759, "learning_rate": 4.980800171058715e-05, "loss": 0.066, "num_input_tokens_seen": 3053872, "step": 1580 }, { "epoch": 0.2585855289991027, "grad_norm": 6.652406692504883, "learning_rate": 4.980678540792715e-05, "loss": 0.1909, "num_input_tokens_seen": 3064560, "step": 1585 }, { "epoch": 0.25940125621992005, "grad_norm": 4.4123663902282715, "learning_rate": 4.980556527974909e-05, "loss": 0.1509, "num_input_tokens_seen": 3075408, "step": 1590 }, { "epoch": 0.2602169834407374, "grad_norm": 2.053060531616211, "learning_rate": 4.980434132624114e-05, "loss": 0.2524, "num_input_tokens_seen": 3086240, "step": 1595 }, { "epoch": 0.26103271066155476, "grad_norm": 2.4497063159942627, "learning_rate": 4.980311354759205e-05, "loss": 0.1716, "num_input_tokens_seen": 3095216, "step": 1600 }, { "epoch": 0.26103271066155476, "eval_loss": 0.16884565353393555, "eval_runtime": 80.4753, "eval_samples_per_second": 33.861, "eval_steps_per_second": 16.937, "num_input_tokens_seen": 3095216, "step": 1600 }, { "epoch": 0.2618484378823721, "grad_norm": 0.5283342003822327, "learning_rate": 4.980188194399116e-05, "loss": 0.0563, "num_input_tokens_seen": 3105056, "step": 1605 }, { "epoch": 0.2626641651031895, "grad_norm": 3.6346142292022705, "learning_rate": 4.9800646515628384e-05, "loss": 0.2176, "num_input_tokens_seen": 3116064, "step": 1610 }, { "epoch": 0.26347989232400687, "grad_norm": 3.2141287326812744, "learning_rate": 4.979940726269426e-05, "loss": 0.1024, "num_input_tokens_seen": 3126272, "step": 1615 }, { "epoch": 0.2642956195448242, "grad_norm": 10.714072227478027, "learning_rate": 4.979816418537988e-05, "loss": 0.3802, "num_input_tokens_seen": 3135520, "step": 1620 }, { "epoch": 0.2651113467656416, "grad_norm": 3.5528223514556885, "learning_rate": 4.979691728387696e-05, "loss": 0.2414, "num_input_tokens_seen": 3145680, "step": 1625 }, { "epoch": 0.26592707398645893, "grad_norm": 3.8710150718688965, "learning_rate": 4.979566655837776e-05, "loss": 0.2088, "num_input_tokens_seen": 3156032, "step": 1630 }, { "epoch": 0.2667428012072763, "grad_norm": 0.43514299392700195, "learning_rate": 4.9794412009075184e-05, "loss": 0.1595, "num_input_tokens_seen": 3164912, "step": 1635 }, { "epoch": 0.26755852842809363, "grad_norm": 2.0786993503570557, "learning_rate": 4.979315363616269e-05, "loss": 0.1067, "num_input_tokens_seen": 3175056, "step": 1640 }, { "epoch": 0.268374255648911, "grad_norm": 0.9220787286758423, "learning_rate": 4.979189143983434e-05, "loss": 0.1238, "num_input_tokens_seen": 3184976, "step": 1645 }, { "epoch": 0.26918998286972834, "grad_norm": 2.041755199432373, "learning_rate": 4.979062542028478e-05, "loss": 0.1339, "num_input_tokens_seen": 3195552, "step": 1650 }, { "epoch": 0.27000571009054575, "grad_norm": 2.150991916656494, "learning_rate": 4.978935557770923e-05, "loss": 0.1481, "num_input_tokens_seen": 3205312, "step": 1655 }, { "epoch": 0.2708214373113631, "grad_norm": 3.522237539291382, "learning_rate": 4.978808191230353e-05, "loss": 0.1049, "num_input_tokens_seen": 3215216, "step": 1660 }, { "epoch": 0.27163716453218045, "grad_norm": 4.184380531311035, "learning_rate": 4.9786804424264085e-05, "loss": 0.2217, "num_input_tokens_seen": 3225024, "step": 1665 }, { "epoch": 0.2724528917529978, "grad_norm": 5.816573619842529, "learning_rate": 4.978552311378792e-05, "loss": 0.3031, "num_input_tokens_seen": 3234032, "step": 1670 }, { "epoch": 0.27326861897381516, "grad_norm": 4.096407413482666, "learning_rate": 4.978423798107261e-05, "loss": 0.2426, "num_input_tokens_seen": 3244160, "step": 1675 }, { "epoch": 0.2740843461946325, "grad_norm": 3.425082206726074, "learning_rate": 4.978294902631635e-05, "loss": 0.2013, "num_input_tokens_seen": 3254144, "step": 1680 }, { "epoch": 0.27490007341544986, "grad_norm": 2.917534589767456, "learning_rate": 4.9781656249717914e-05, "loss": 0.2164, "num_input_tokens_seen": 3264400, "step": 1685 }, { "epoch": 0.2757158006362672, "grad_norm": 5.522337913513184, "learning_rate": 4.9780359651476645e-05, "loss": 0.2439, "num_input_tokens_seen": 3273456, "step": 1690 }, { "epoch": 0.27653152785708457, "grad_norm": 2.3305258750915527, "learning_rate": 4.977905923179251e-05, "loss": 0.1545, "num_input_tokens_seen": 3283680, "step": 1695 }, { "epoch": 0.2773472550779019, "grad_norm": 1.6220788955688477, "learning_rate": 4.977775499086606e-05, "loss": 0.1596, "num_input_tokens_seen": 3292960, "step": 1700 }, { "epoch": 0.27816298229871933, "grad_norm": 0.8839308619499207, "learning_rate": 4.97764469288984e-05, "loss": 0.1284, "num_input_tokens_seen": 3302160, "step": 1705 }, { "epoch": 0.2789787095195367, "grad_norm": 2.515286445617676, "learning_rate": 4.977513504609127e-05, "loss": 0.0943, "num_input_tokens_seen": 3312400, "step": 1710 }, { "epoch": 0.27979443674035404, "grad_norm": 0.24816642701625824, "learning_rate": 4.9773819342646965e-05, "loss": 0.0516, "num_input_tokens_seen": 3322352, "step": 1715 }, { "epoch": 0.2806101639611714, "grad_norm": 1.3077576160430908, "learning_rate": 4.97724998187684e-05, "loss": 0.256, "num_input_tokens_seen": 3331536, "step": 1720 }, { "epoch": 0.28142589118198874, "grad_norm": 3.6691477298736572, "learning_rate": 4.9771176474659045e-05, "loss": 0.1276, "num_input_tokens_seen": 3341632, "step": 1725 }, { "epoch": 0.2822416184028061, "grad_norm": 1.113187551498413, "learning_rate": 4.976984931052299e-05, "loss": 0.1156, "num_input_tokens_seen": 3351200, "step": 1730 }, { "epoch": 0.28305734562362345, "grad_norm": 3.5667006969451904, "learning_rate": 4.976851832656489e-05, "loss": 0.1533, "num_input_tokens_seen": 3360480, "step": 1735 }, { "epoch": 0.2838730728444408, "grad_norm": 3.796882390975952, "learning_rate": 4.9767183522990004e-05, "loss": 0.2537, "num_input_tokens_seen": 3368848, "step": 1740 }, { "epoch": 0.28468880006525815, "grad_norm": 6.122213363647461, "learning_rate": 4.9765844900004176e-05, "loss": 0.3346, "num_input_tokens_seen": 3377680, "step": 1745 }, { "epoch": 0.28550452728607556, "grad_norm": 4.5856242179870605, "learning_rate": 4.9764502457813834e-05, "loss": 0.1168, "num_input_tokens_seen": 3388528, "step": 1750 }, { "epoch": 0.2863202545068929, "grad_norm": 0.7014274001121521, "learning_rate": 4.9763156196626005e-05, "loss": 0.099, "num_input_tokens_seen": 3398448, "step": 1755 }, { "epoch": 0.28713598172771027, "grad_norm": 3.833523988723755, "learning_rate": 4.97618061166483e-05, "loss": 0.2439, "num_input_tokens_seen": 3407248, "step": 1760 }, { "epoch": 0.2879517089485276, "grad_norm": 3.1522367000579834, "learning_rate": 4.9760452218088915e-05, "loss": 0.1194, "num_input_tokens_seen": 3416640, "step": 1765 }, { "epoch": 0.28876743616934497, "grad_norm": 5.101263046264648, "learning_rate": 4.975909450115663e-05, "loss": 0.3219, "num_input_tokens_seen": 3426016, "step": 1770 }, { "epoch": 0.2895831633901623, "grad_norm": 1.8819241523742676, "learning_rate": 4.975773296606084e-05, "loss": 0.1398, "num_input_tokens_seen": 3436160, "step": 1775 }, { "epoch": 0.2903988906109797, "grad_norm": 4.468438625335693, "learning_rate": 4.97563676130115e-05, "loss": 0.1764, "num_input_tokens_seen": 3446384, "step": 1780 }, { "epoch": 0.29121461783179703, "grad_norm": 7.367520809173584, "learning_rate": 4.9754998442219166e-05, "loss": 0.1299, "num_input_tokens_seen": 3455392, "step": 1785 }, { "epoch": 0.2920303450526144, "grad_norm": 0.6520694494247437, "learning_rate": 4.9753625453894984e-05, "loss": 0.1675, "num_input_tokens_seen": 3464144, "step": 1790 }, { "epoch": 0.2928460722734318, "grad_norm": 3.273996353149414, "learning_rate": 4.975224864825068e-05, "loss": 0.182, "num_input_tokens_seen": 3473504, "step": 1795 }, { "epoch": 0.29366179949424914, "grad_norm": 2.3622043132781982, "learning_rate": 4.9750868025498576e-05, "loss": 0.1506, "num_input_tokens_seen": 3483504, "step": 1800 }, { "epoch": 0.29366179949424914, "eval_loss": 0.16898152232170105, "eval_runtime": 80.4731, "eval_samples_per_second": 33.862, "eval_steps_per_second": 16.937, "num_input_tokens_seen": 3483504, "step": 1800 }, { "epoch": 0.2944775267150665, "grad_norm": 1.079675316810608, "learning_rate": 4.974948358585158e-05, "loss": 0.1361, "num_input_tokens_seen": 3493856, "step": 1805 }, { "epoch": 0.29529325393588385, "grad_norm": 3.4213874340057373, "learning_rate": 4.9748095329523205e-05, "loss": 0.1946, "num_input_tokens_seen": 3502848, "step": 1810 }, { "epoch": 0.2961089811567012, "grad_norm": 5.383677005767822, "learning_rate": 4.974670325672752e-05, "loss": 0.1452, "num_input_tokens_seen": 3513424, "step": 1815 }, { "epoch": 0.29692470837751855, "grad_norm": 9.146373748779297, "learning_rate": 4.974530736767921e-05, "loss": 0.2157, "num_input_tokens_seen": 3520816, "step": 1820 }, { "epoch": 0.2977404355983359, "grad_norm": 5.925894737243652, "learning_rate": 4.9743907662593524e-05, "loss": 0.2954, "num_input_tokens_seen": 3531376, "step": 1825 }, { "epoch": 0.29855616281915326, "grad_norm": 0.3918537199497223, "learning_rate": 4.974250414168633e-05, "loss": 0.1157, "num_input_tokens_seen": 3542144, "step": 1830 }, { "epoch": 0.2993718900399706, "grad_norm": 1.6769565343856812, "learning_rate": 4.974109680517407e-05, "loss": 0.1904, "num_input_tokens_seen": 3551120, "step": 1835 }, { "epoch": 0.300187617260788, "grad_norm": 3.3778395652770996, "learning_rate": 4.973968565327376e-05, "loss": 0.1413, "num_input_tokens_seen": 3560800, "step": 1840 }, { "epoch": 0.3010033444816054, "grad_norm": 2.4970593452453613, "learning_rate": 4.973827068620303e-05, "loss": 0.0998, "num_input_tokens_seen": 3569280, "step": 1845 }, { "epoch": 0.3018190717024227, "grad_norm": 3.813615322113037, "learning_rate": 4.973685190418008e-05, "loss": 0.2406, "num_input_tokens_seen": 3578704, "step": 1850 }, { "epoch": 0.3026347989232401, "grad_norm": 3.413303852081299, "learning_rate": 4.97354293074237e-05, "loss": 0.1875, "num_input_tokens_seen": 3587760, "step": 1855 }, { "epoch": 0.30345052614405743, "grad_norm": 2.712235450744629, "learning_rate": 4.9734002896153276e-05, "loss": 0.1487, "num_input_tokens_seen": 3596496, "step": 1860 }, { "epoch": 0.3042662533648748, "grad_norm": 4.288201808929443, "learning_rate": 4.973257267058877e-05, "loss": 0.1756, "num_input_tokens_seen": 3605744, "step": 1865 }, { "epoch": 0.30508198058569214, "grad_norm": 2.768932342529297, "learning_rate": 4.973113863095076e-05, "loss": 0.2611, "num_input_tokens_seen": 3615648, "step": 1870 }, { "epoch": 0.3058977078065095, "grad_norm": 2.4355952739715576, "learning_rate": 4.9729700777460384e-05, "loss": 0.1913, "num_input_tokens_seen": 3625456, "step": 1875 }, { "epoch": 0.30671343502732684, "grad_norm": 0.9855892062187195, "learning_rate": 4.972825911033937e-05, "loss": 0.14, "num_input_tokens_seen": 3635424, "step": 1880 }, { "epoch": 0.3075291622481442, "grad_norm": 1.750441312789917, "learning_rate": 4.9726813629810056e-05, "loss": 0.1275, "num_input_tokens_seen": 3644816, "step": 1885 }, { "epoch": 0.3083448894689616, "grad_norm": 3.805288791656494, "learning_rate": 4.9725364336095326e-05, "loss": 0.1496, "num_input_tokens_seen": 3654592, "step": 1890 }, { "epoch": 0.30916061668977896, "grad_norm": 2.6993308067321777, "learning_rate": 4.972391122941871e-05, "loss": 0.1442, "num_input_tokens_seen": 3663568, "step": 1895 }, { "epoch": 0.3099763439105963, "grad_norm": 2.0891520977020264, "learning_rate": 4.972245431000428e-05, "loss": 0.2052, "num_input_tokens_seen": 3673520, "step": 1900 }, { "epoch": 0.31079207113141366, "grad_norm": 3.0131924152374268, "learning_rate": 4.972099357807671e-05, "loss": 0.1248, "num_input_tokens_seen": 3684416, "step": 1905 }, { "epoch": 0.311607798352231, "grad_norm": 2.8031163215637207, "learning_rate": 4.971952903386127e-05, "loss": 0.1743, "num_input_tokens_seen": 3694736, "step": 1910 }, { "epoch": 0.31242352557304837, "grad_norm": 0.8483976125717163, "learning_rate": 4.971806067758381e-05, "loss": 0.1198, "num_input_tokens_seen": 3704400, "step": 1915 }, { "epoch": 0.3132392527938657, "grad_norm": 3.996195077896118, "learning_rate": 4.971658850947076e-05, "loss": 0.2946, "num_input_tokens_seen": 3714160, "step": 1920 }, { "epoch": 0.31405498001468307, "grad_norm": 3.621753692626953, "learning_rate": 4.9715112529749165e-05, "loss": 0.1879, "num_input_tokens_seen": 3724864, "step": 1925 }, { "epoch": 0.3148707072355004, "grad_norm": 4.1725754737854, "learning_rate": 4.9713632738646624e-05, "loss": 0.1212, "num_input_tokens_seen": 3735088, "step": 1930 }, { "epoch": 0.31568643445631783, "grad_norm": 3.8697903156280518, "learning_rate": 4.971214913639134e-05, "loss": 0.1506, "num_input_tokens_seen": 3745360, "step": 1935 }, { "epoch": 0.3165021616771352, "grad_norm": 0.5149944424629211, "learning_rate": 4.9710661723212104e-05, "loss": 0.1813, "num_input_tokens_seen": 3755968, "step": 1940 }, { "epoch": 0.31731788889795254, "grad_norm": 3.4427783489227295, "learning_rate": 4.9709170499338295e-05, "loss": 0.1484, "num_input_tokens_seen": 3764672, "step": 1945 }, { "epoch": 0.3181336161187699, "grad_norm": 6.513813495635986, "learning_rate": 4.9707675464999895e-05, "loss": 0.2445, "num_input_tokens_seen": 3774208, "step": 1950 }, { "epoch": 0.31894934333958724, "grad_norm": 0.7264953255653381, "learning_rate": 4.970617662042743e-05, "loss": 0.1172, "num_input_tokens_seen": 3784528, "step": 1955 }, { "epoch": 0.3197650705604046, "grad_norm": 3.2111825942993164, "learning_rate": 4.970467396585206e-05, "loss": 0.1819, "num_input_tokens_seen": 3793696, "step": 1960 }, { "epoch": 0.32058079778122195, "grad_norm": 4.300517559051514, "learning_rate": 4.97031675015055e-05, "loss": 0.2811, "num_input_tokens_seen": 3802848, "step": 1965 }, { "epoch": 0.3213965250020393, "grad_norm": 5.739434719085693, "learning_rate": 4.9701657227620075e-05, "loss": 0.2013, "num_input_tokens_seen": 3812592, "step": 1970 }, { "epoch": 0.32221225222285665, "grad_norm": 0.7602491974830627, "learning_rate": 4.9700143144428685e-05, "loss": 0.162, "num_input_tokens_seen": 3822224, "step": 1975 }, { "epoch": 0.32302797944367406, "grad_norm": 2.067610740661621, "learning_rate": 4.969862525216482e-05, "loss": 0.0946, "num_input_tokens_seen": 3831824, "step": 1980 }, { "epoch": 0.3238437066644914, "grad_norm": 3.486706256866455, "learning_rate": 4.9697103551062556e-05, "loss": 0.0949, "num_input_tokens_seen": 3841392, "step": 1985 }, { "epoch": 0.32465943388530877, "grad_norm": 3.2342031002044678, "learning_rate": 4.9695578041356565e-05, "loss": 0.2679, "num_input_tokens_seen": 3851200, "step": 1990 }, { "epoch": 0.3254751611061261, "grad_norm": 2.0003609657287598, "learning_rate": 4.969404872328209e-05, "loss": 0.1717, "num_input_tokens_seen": 3861664, "step": 1995 }, { "epoch": 0.3262908883269435, "grad_norm": 2.425981044769287, "learning_rate": 4.969251559707498e-05, "loss": 0.2079, "num_input_tokens_seen": 3872976, "step": 2000 }, { "epoch": 0.3262908883269435, "eval_loss": 0.1561587154865265, "eval_runtime": 80.5445, "eval_samples_per_second": 33.832, "eval_steps_per_second": 16.922, "num_input_tokens_seen": 3872976, "step": 2000 }, { "epoch": 0.3271066155477608, "grad_norm": 1.5189738273620605, "learning_rate": 4.9690978662971674e-05, "loss": 0.1744, "num_input_tokens_seen": 3882896, "step": 2005 }, { "epoch": 0.3279223427685782, "grad_norm": 2.6177382469177246, "learning_rate": 4.968943792120916e-05, "loss": 0.2405, "num_input_tokens_seen": 3892848, "step": 2010 }, { "epoch": 0.32873806998939553, "grad_norm": 6.364871978759766, "learning_rate": 4.9687893372025046e-05, "loss": 0.2457, "num_input_tokens_seen": 3903072, "step": 2015 }, { "epoch": 0.3295537972102129, "grad_norm": 2.7170560359954834, "learning_rate": 4.9686345015657535e-05, "loss": 0.1865, "num_input_tokens_seen": 3913776, "step": 2020 }, { "epoch": 0.33036952443103024, "grad_norm": 3.0262880325317383, "learning_rate": 4.968479285234538e-05, "loss": 0.1612, "num_input_tokens_seen": 3923568, "step": 2025 }, { "epoch": 0.33118525165184765, "grad_norm": 3.5740468502044678, "learning_rate": 4.9683236882327974e-05, "loss": 0.1829, "num_input_tokens_seen": 3933280, "step": 2030 }, { "epoch": 0.332000978872665, "grad_norm": 3.1455540657043457, "learning_rate": 4.968167710584526e-05, "loss": 0.1736, "num_input_tokens_seen": 3942032, "step": 2035 }, { "epoch": 0.33281670609348235, "grad_norm": 5.252224922180176, "learning_rate": 4.968011352313775e-05, "loss": 0.142, "num_input_tokens_seen": 3951728, "step": 2040 }, { "epoch": 0.3336324333142997, "grad_norm": 2.394721746444702, "learning_rate": 4.967854613444659e-05, "loss": 0.1236, "num_input_tokens_seen": 3961536, "step": 2045 }, { "epoch": 0.33444816053511706, "grad_norm": 5.4086127281188965, "learning_rate": 4.967697494001349e-05, "loss": 0.2773, "num_input_tokens_seen": 3970800, "step": 2050 }, { "epoch": 0.3352638877559344, "grad_norm": 3.0963058471679688, "learning_rate": 4.9675399940080736e-05, "loss": 0.1891, "num_input_tokens_seen": 3980640, "step": 2055 }, { "epoch": 0.33607961497675176, "grad_norm": 6.214781761169434, "learning_rate": 4.9673821134891226e-05, "loss": 0.0988, "num_input_tokens_seen": 3989552, "step": 2060 }, { "epoch": 0.3368953421975691, "grad_norm": 3.0821030139923096, "learning_rate": 4.967223852468842e-05, "loss": 0.1487, "num_input_tokens_seen": 3999152, "step": 2065 }, { "epoch": 0.33771106941838647, "grad_norm": 2.782444477081299, "learning_rate": 4.967065210971639e-05, "loss": 0.075, "num_input_tokens_seen": 4009472, "step": 2070 }, { "epoch": 0.3385267966392039, "grad_norm": 5.829905986785889, "learning_rate": 4.966906189021977e-05, "loss": 0.1887, "num_input_tokens_seen": 4019552, "step": 2075 }, { "epoch": 0.33934252386002123, "grad_norm": 4.534142017364502, "learning_rate": 4.966746786644379e-05, "loss": 0.2441, "num_input_tokens_seen": 4027968, "step": 2080 }, { "epoch": 0.3401582510808386, "grad_norm": 1.1295180320739746, "learning_rate": 4.966587003863429e-05, "loss": 0.1545, "num_input_tokens_seen": 4037184, "step": 2085 }, { "epoch": 0.34097397830165593, "grad_norm": 2.0278964042663574, "learning_rate": 4.966426840703765e-05, "loss": 0.1523, "num_input_tokens_seen": 4047760, "step": 2090 }, { "epoch": 0.3417897055224733, "grad_norm": 3.5291926860809326, "learning_rate": 4.9662662971900875e-05, "loss": 0.1655, "num_input_tokens_seen": 4057152, "step": 2095 }, { "epoch": 0.34260543274329064, "grad_norm": 3.5582194328308105, "learning_rate": 4.9661053733471534e-05, "loss": 0.1785, "num_input_tokens_seen": 4066320, "step": 2100 }, { "epoch": 0.343421159964108, "grad_norm": 2.403644323348999, "learning_rate": 4.965944069199781e-05, "loss": 0.2461, "num_input_tokens_seen": 4076464, "step": 2105 }, { "epoch": 0.34423688718492534, "grad_norm": 3.1906845569610596, "learning_rate": 4.965782384772842e-05, "loss": 0.1525, "num_input_tokens_seen": 4085456, "step": 2110 }, { "epoch": 0.3450526144057427, "grad_norm": 1.5384975671768188, "learning_rate": 4.9656203200912734e-05, "loss": 0.2106, "num_input_tokens_seen": 4095104, "step": 2115 }, { "epoch": 0.3458683416265601, "grad_norm": 3.7206132411956787, "learning_rate": 4.965457875180067e-05, "loss": 0.1593, "num_input_tokens_seen": 4105312, "step": 2120 }, { "epoch": 0.34668406884737746, "grad_norm": 2.372290849685669, "learning_rate": 4.9652950500642724e-05, "loss": 0.1009, "num_input_tokens_seen": 4115248, "step": 2125 }, { "epoch": 0.3474997960681948, "grad_norm": 1.743883728981018, "learning_rate": 4.965131844769001e-05, "loss": 0.1532, "num_input_tokens_seen": 4124560, "step": 2130 }, { "epoch": 0.34831552328901216, "grad_norm": 5.934199333190918, "learning_rate": 4.96496825931942e-05, "loss": 0.2949, "num_input_tokens_seen": 4133776, "step": 2135 }, { "epoch": 0.3491312505098295, "grad_norm": 2.4048616886138916, "learning_rate": 4.9648042937407566e-05, "loss": 0.1752, "num_input_tokens_seen": 4142480, "step": 2140 }, { "epoch": 0.34994697773064687, "grad_norm": 4.917928695678711, "learning_rate": 4.964639948058297e-05, "loss": 0.1064, "num_input_tokens_seen": 4151184, "step": 2145 }, { "epoch": 0.3507627049514642, "grad_norm": 4.004197597503662, "learning_rate": 4.9644752222973846e-05, "loss": 0.1985, "num_input_tokens_seen": 4160656, "step": 2150 }, { "epoch": 0.3515784321722816, "grad_norm": 6.1970601081848145, "learning_rate": 4.964310116483422e-05, "loss": 0.1613, "num_input_tokens_seen": 4171104, "step": 2155 }, { "epoch": 0.3523941593930989, "grad_norm": 3.543532371520996, "learning_rate": 4.964144630641872e-05, "loss": 0.2142, "num_input_tokens_seen": 4180128, "step": 2160 }, { "epoch": 0.3532098866139163, "grad_norm": 4.401368141174316, "learning_rate": 4.9639787647982525e-05, "loss": 0.1648, "num_input_tokens_seen": 4188944, "step": 2165 }, { "epoch": 0.3540256138347337, "grad_norm": 2.754697561264038, "learning_rate": 4.963812518978143e-05, "loss": 0.1627, "num_input_tokens_seen": 4199472, "step": 2170 }, { "epoch": 0.35484134105555104, "grad_norm": 4.694288730621338, "learning_rate": 4.963645893207182e-05, "loss": 0.1648, "num_input_tokens_seen": 4208624, "step": 2175 }, { "epoch": 0.3556570682763684, "grad_norm": 0.41743844747543335, "learning_rate": 4.963478887511063e-05, "loss": 0.1379, "num_input_tokens_seen": 4217488, "step": 2180 }, { "epoch": 0.35647279549718575, "grad_norm": 2.421638250350952, "learning_rate": 4.963311501915542e-05, "loss": 0.0572, "num_input_tokens_seen": 4226336, "step": 2185 }, { "epoch": 0.3572885227180031, "grad_norm": 4.4105072021484375, "learning_rate": 4.963143736446432e-05, "loss": 0.2935, "num_input_tokens_seen": 4235168, "step": 2190 }, { "epoch": 0.35810424993882045, "grad_norm": 3.577188730239868, "learning_rate": 4.962975591129603e-05, "loss": 0.175, "num_input_tokens_seen": 4244912, "step": 2195 }, { "epoch": 0.3589199771596378, "grad_norm": 6.645815372467041, "learning_rate": 4.962807065990986e-05, "loss": 0.1947, "num_input_tokens_seen": 4254272, "step": 2200 }, { "epoch": 0.3589199771596378, "eval_loss": 0.1468677520751953, "eval_runtime": 80.5547, "eval_samples_per_second": 33.828, "eval_steps_per_second": 16.92, "num_input_tokens_seen": 4254272, "step": 2200 }, { "epoch": 0.35973570438045516, "grad_norm": 2.039992094039917, "learning_rate": 4.9626381610565714e-05, "loss": 0.1642, "num_input_tokens_seen": 4263248, "step": 2205 }, { "epoch": 0.3605514316012725, "grad_norm": 3.3931894302368164, "learning_rate": 4.9624688763524043e-05, "loss": 0.0926, "num_input_tokens_seen": 4273664, "step": 2210 }, { "epoch": 0.3613671588220899, "grad_norm": 3.776719808578491, "learning_rate": 4.962299211904591e-05, "loss": 0.1995, "num_input_tokens_seen": 4283360, "step": 2215 }, { "epoch": 0.36218288604290727, "grad_norm": 5.454777240753174, "learning_rate": 4.962129167739296e-05, "loss": 0.1665, "num_input_tokens_seen": 4293648, "step": 2220 }, { "epoch": 0.3629986132637246, "grad_norm": 0.8218948245048523, "learning_rate": 4.961958743882742e-05, "loss": 0.1306, "num_input_tokens_seen": 4303264, "step": 2225 }, { "epoch": 0.363814340484542, "grad_norm": 3.092089891433716, "learning_rate": 4.961787940361211e-05, "loss": 0.1562, "num_input_tokens_seen": 4313136, "step": 2230 }, { "epoch": 0.36463006770535933, "grad_norm": 3.934082508087158, "learning_rate": 4.961616757201043e-05, "loss": 0.2057, "num_input_tokens_seen": 4323424, "step": 2235 }, { "epoch": 0.3654457949261767, "grad_norm": 1.51753568649292, "learning_rate": 4.961445194428637e-05, "loss": 0.1268, "num_input_tokens_seen": 4332848, "step": 2240 }, { "epoch": 0.36626152214699403, "grad_norm": 5.6349334716796875, "learning_rate": 4.9612732520704486e-05, "loss": 0.1491, "num_input_tokens_seen": 4343344, "step": 2245 }, { "epoch": 0.3670772493678114, "grad_norm": 1.75339937210083, "learning_rate": 4.961100930152994e-05, "loss": 0.2679, "num_input_tokens_seen": 4352656, "step": 2250 }, { "epoch": 0.36789297658862874, "grad_norm": 2.293443441390991, "learning_rate": 4.960928228702849e-05, "loss": 0.1233, "num_input_tokens_seen": 4362736, "step": 2255 }, { "epoch": 0.36870870380944615, "grad_norm": 4.779496669769287, "learning_rate": 4.960755147746645e-05, "loss": 0.181, "num_input_tokens_seen": 4369776, "step": 2260 }, { "epoch": 0.3695244310302635, "grad_norm": 2.051400899887085, "learning_rate": 4.9605816873110736e-05, "loss": 0.1741, "num_input_tokens_seen": 4379200, "step": 2265 }, { "epoch": 0.37034015825108085, "grad_norm": 2.470956802368164, "learning_rate": 4.960407847422883e-05, "loss": 0.1529, "num_input_tokens_seen": 4388080, "step": 2270 }, { "epoch": 0.3711558854718982, "grad_norm": 2.723106861114502, "learning_rate": 4.960233628108885e-05, "loss": 0.1643, "num_input_tokens_seen": 4398368, "step": 2275 }, { "epoch": 0.37197161269271556, "grad_norm": 0.9187365174293518, "learning_rate": 4.960059029395942e-05, "loss": 0.1048, "num_input_tokens_seen": 4407936, "step": 2280 }, { "epoch": 0.3727873399135329, "grad_norm": 3.897923707962036, "learning_rate": 4.959884051310983e-05, "loss": 0.128, "num_input_tokens_seen": 4418464, "step": 2285 }, { "epoch": 0.37360306713435026, "grad_norm": 4.4176926612854, "learning_rate": 4.959708693880991e-05, "loss": 0.1702, "num_input_tokens_seen": 4427232, "step": 2290 }, { "epoch": 0.3744187943551676, "grad_norm": 0.6433035135269165, "learning_rate": 4.9595329571330074e-05, "loss": 0.0603, "num_input_tokens_seen": 4435920, "step": 2295 }, { "epoch": 0.37523452157598497, "grad_norm": 2.658217191696167, "learning_rate": 4.9593568410941326e-05, "loss": 0.1879, "num_input_tokens_seen": 4445232, "step": 2300 }, { "epoch": 0.3760502487968023, "grad_norm": 0.9176677465438843, "learning_rate": 4.959180345791528e-05, "loss": 0.1598, "num_input_tokens_seen": 4454464, "step": 2305 }, { "epoch": 0.37686597601761973, "grad_norm": 0.21786712110042572, "learning_rate": 4.9590034712524086e-05, "loss": 0.1941, "num_input_tokens_seen": 4464432, "step": 2310 }, { "epoch": 0.3776817032384371, "grad_norm": 2.9176032543182373, "learning_rate": 4.958826217504053e-05, "loss": 0.1927, "num_input_tokens_seen": 4474368, "step": 2315 }, { "epoch": 0.37849743045925444, "grad_norm": 5.222769737243652, "learning_rate": 4.958648584573795e-05, "loss": 0.3066, "num_input_tokens_seen": 4483856, "step": 2320 }, { "epoch": 0.3793131576800718, "grad_norm": 1.9787544012069702, "learning_rate": 4.958470572489028e-05, "loss": 0.1374, "num_input_tokens_seen": 4493104, "step": 2325 }, { "epoch": 0.38012888490088914, "grad_norm": 0.4447499215602875, "learning_rate": 4.958292181277203e-05, "loss": 0.1034, "num_input_tokens_seen": 4502848, "step": 2330 }, { "epoch": 0.3809446121217065, "grad_norm": 1.5550415515899658, "learning_rate": 4.958113410965832e-05, "loss": 0.1067, "num_input_tokens_seen": 4512384, "step": 2335 }, { "epoch": 0.38176033934252385, "grad_norm": 0.35908281803131104, "learning_rate": 4.957934261582481e-05, "loss": 0.1629, "num_input_tokens_seen": 4523760, "step": 2340 }, { "epoch": 0.3825760665633412, "grad_norm": 2.8292624950408936, "learning_rate": 4.95775473315478e-05, "loss": 0.2076, "num_input_tokens_seen": 4533472, "step": 2345 }, { "epoch": 0.38339179378415855, "grad_norm": 2.1734533309936523, "learning_rate": 4.9575748257104124e-05, "loss": 0.1381, "num_input_tokens_seen": 4542896, "step": 2350 }, { "epoch": 0.38420752100497596, "grad_norm": 3.0647635459899902, "learning_rate": 4.9573945392771224e-05, "loss": 0.1867, "num_input_tokens_seen": 4552032, "step": 2355 }, { "epoch": 0.3850232482257933, "grad_norm": 3.1652636528015137, "learning_rate": 4.9572138738827134e-05, "loss": 0.1266, "num_input_tokens_seen": 4560928, "step": 2360 }, { "epoch": 0.38583897544661067, "grad_norm": 3.6657984256744385, "learning_rate": 4.957032829555046e-05, "loss": 0.1309, "num_input_tokens_seen": 4569568, "step": 2365 }, { "epoch": 0.386654702667428, "grad_norm": 2.205678939819336, "learning_rate": 4.956851406322039e-05, "loss": 0.1899, "num_input_tokens_seen": 4578896, "step": 2370 }, { "epoch": 0.38747042988824537, "grad_norm": 0.62244713306427, "learning_rate": 4.9566696042116704e-05, "loss": 0.1112, "num_input_tokens_seen": 4588624, "step": 2375 }, { "epoch": 0.3882861571090627, "grad_norm": 2.274362802505493, "learning_rate": 4.9564874232519766e-05, "loss": 0.0771, "num_input_tokens_seen": 4598864, "step": 2380 }, { "epoch": 0.3891018843298801, "grad_norm": 3.607962131500244, "learning_rate": 4.9563048634710516e-05, "loss": 0.1854, "num_input_tokens_seen": 4608672, "step": 2385 }, { "epoch": 0.38991761155069743, "grad_norm": 1.4993677139282227, "learning_rate": 4.956121924897049e-05, "loss": 0.1545, "num_input_tokens_seen": 4618288, "step": 2390 }, { "epoch": 0.3907333387715148, "grad_norm": 1.6663634777069092, "learning_rate": 4.955938607558181e-05, "loss": 0.1407, "num_input_tokens_seen": 4627904, "step": 2395 }, { "epoch": 0.3915490659923322, "grad_norm": 2.80906081199646, "learning_rate": 4.955754911482715e-05, "loss": 0.1659, "num_input_tokens_seen": 4637376, "step": 2400 }, { "epoch": 0.3915490659923322, "eval_loss": 0.15362855792045593, "eval_runtime": 80.5933, "eval_samples_per_second": 33.812, "eval_steps_per_second": 16.912, "num_input_tokens_seen": 4637376, "step": 2400 }, { "epoch": 0.39236479321314954, "grad_norm": 3.4053850173950195, "learning_rate": 4.9555708366989804e-05, "loss": 0.1795, "num_input_tokens_seen": 4646928, "step": 2405 }, { "epoch": 0.3931805204339669, "grad_norm": 6.350583076477051, "learning_rate": 4.9553863832353655e-05, "loss": 0.1467, "num_input_tokens_seen": 4656080, "step": 2410 }, { "epoch": 0.39399624765478425, "grad_norm": 6.189652919769287, "learning_rate": 4.955201551120313e-05, "loss": 0.2426, "num_input_tokens_seen": 4666240, "step": 2415 }, { "epoch": 0.3948119748756016, "grad_norm": 1.9365154504776, "learning_rate": 4.955016340382328e-05, "loss": 0.1847, "num_input_tokens_seen": 4676704, "step": 2420 }, { "epoch": 0.39562770209641895, "grad_norm": 4.060844898223877, "learning_rate": 4.954830751049972e-05, "loss": 0.1705, "num_input_tokens_seen": 4686096, "step": 2425 }, { "epoch": 0.3964434293172363, "grad_norm": 4.018820762634277, "learning_rate": 4.954644783151864e-05, "loss": 0.1568, "num_input_tokens_seen": 4696048, "step": 2430 }, { "epoch": 0.39725915653805366, "grad_norm": 4.8212432861328125, "learning_rate": 4.954458436716684e-05, "loss": 0.1104, "num_input_tokens_seen": 4705040, "step": 2435 }, { "epoch": 0.398074883758871, "grad_norm": 0.5027077198028564, "learning_rate": 4.954271711773168e-05, "loss": 0.2213, "num_input_tokens_seen": 4715152, "step": 2440 }, { "epoch": 0.39889061097968836, "grad_norm": 2.5174286365509033, "learning_rate": 4.9540846083501115e-05, "loss": 0.105, "num_input_tokens_seen": 4725456, "step": 2445 }, { "epoch": 0.3997063382005058, "grad_norm": 3.966956377029419, "learning_rate": 4.953897126476369e-05, "loss": 0.0618, "num_input_tokens_seen": 4735552, "step": 2450 }, { "epoch": 0.4005220654213231, "grad_norm": 0.4652875065803528, "learning_rate": 4.9537092661808514e-05, "loss": 0.1822, "num_input_tokens_seen": 4745552, "step": 2455 }, { "epoch": 0.4013377926421405, "grad_norm": 7.436320781707764, "learning_rate": 4.9535210274925306e-05, "loss": 0.2691, "num_input_tokens_seen": 4755920, "step": 2460 }, { "epoch": 0.40215351986295783, "grad_norm": 0.35505548119544983, "learning_rate": 4.953332410440435e-05, "loss": 0.0931, "num_input_tokens_seen": 4766544, "step": 2465 }, { "epoch": 0.4029692470837752, "grad_norm": 0.1869570016860962, "learning_rate": 4.9531434150536496e-05, "loss": 0.0746, "num_input_tokens_seen": 4776032, "step": 2470 }, { "epoch": 0.40378497430459254, "grad_norm": 1.7337158918380737, "learning_rate": 4.952954041361322e-05, "loss": 0.1612, "num_input_tokens_seen": 4786848, "step": 2475 }, { "epoch": 0.4046007015254099, "grad_norm": 2.693042516708374, "learning_rate": 4.952764289392655e-05, "loss": 0.0812, "num_input_tokens_seen": 4796960, "step": 2480 }, { "epoch": 0.40541642874622724, "grad_norm": 2.563663959503174, "learning_rate": 4.952574159176912e-05, "loss": 0.063, "num_input_tokens_seen": 4804992, "step": 2485 }, { "epoch": 0.4062321559670446, "grad_norm": 0.21734054386615753, "learning_rate": 4.952383650743413e-05, "loss": 0.0304, "num_input_tokens_seen": 4813328, "step": 2490 }, { "epoch": 0.407047883187862, "grad_norm": 3.099566698074341, "learning_rate": 4.952192764121536e-05, "loss": 0.3103, "num_input_tokens_seen": 4823056, "step": 2495 }, { "epoch": 0.40786361040867936, "grad_norm": 10.5172758102417, "learning_rate": 4.9520014993407185e-05, "loss": 0.2057, "num_input_tokens_seen": 4831984, "step": 2500 }, { "epoch": 0.4086793376294967, "grad_norm": 1.330620527267456, "learning_rate": 4.951809856430456e-05, "loss": 0.0714, "num_input_tokens_seen": 4841920, "step": 2505 }, { "epoch": 0.40949506485031406, "grad_norm": 3.4734115600585938, "learning_rate": 4.951617835420303e-05, "loss": 0.1553, "num_input_tokens_seen": 4851104, "step": 2510 }, { "epoch": 0.4103107920711314, "grad_norm": 5.063526630401611, "learning_rate": 4.951425436339869e-05, "loss": 0.347, "num_input_tokens_seen": 4859456, "step": 2515 }, { "epoch": 0.41112651929194877, "grad_norm": 6.238959789276123, "learning_rate": 4.9512326592188274e-05, "loss": 0.2516, "num_input_tokens_seen": 4869040, "step": 2520 }, { "epoch": 0.4119422465127661, "grad_norm": 1.3521301746368408, "learning_rate": 4.9510395040869054e-05, "loss": 0.097, "num_input_tokens_seen": 4878288, "step": 2525 }, { "epoch": 0.41275797373358347, "grad_norm": 1.1200133562088013, "learning_rate": 4.9508459709738905e-05, "loss": 0.1485, "num_input_tokens_seen": 4888368, "step": 2530 }, { "epoch": 0.4135737009544008, "grad_norm": 5.664483547210693, "learning_rate": 4.950652059909627e-05, "loss": 0.2793, "num_input_tokens_seen": 4895840, "step": 2535 }, { "epoch": 0.41438942817521823, "grad_norm": 3.356574058532715, "learning_rate": 4.95045777092402e-05, "loss": 0.1705, "num_input_tokens_seen": 4904768, "step": 2540 }, { "epoch": 0.4152051553960356, "grad_norm": 3.5758256912231445, "learning_rate": 4.950263104047031e-05, "loss": 0.226, "num_input_tokens_seen": 4914272, "step": 2545 }, { "epoch": 0.41602088261685294, "grad_norm": 1.9487732648849487, "learning_rate": 4.9500680593086775e-05, "loss": 0.0879, "num_input_tokens_seen": 4924048, "step": 2550 }, { "epoch": 0.4168366098376703, "grad_norm": 1.706809401512146, "learning_rate": 4.94987263673904e-05, "loss": 0.1636, "num_input_tokens_seen": 4933552, "step": 2555 }, { "epoch": 0.41765233705848764, "grad_norm": 0.680821418762207, "learning_rate": 4.949676836368256e-05, "loss": 0.1304, "num_input_tokens_seen": 4943408, "step": 2560 }, { "epoch": 0.418468064279305, "grad_norm": 3.19102144241333, "learning_rate": 4.949480658226518e-05, "loss": 0.1307, "num_input_tokens_seen": 4954176, "step": 2565 }, { "epoch": 0.41928379150012235, "grad_norm": 1.0861793756484985, "learning_rate": 4.949284102344082e-05, "loss": 0.2472, "num_input_tokens_seen": 4962800, "step": 2570 }, { "epoch": 0.4200995187209397, "grad_norm": 1.3217942714691162, "learning_rate": 4.9490871687512565e-05, "loss": 0.1436, "num_input_tokens_seen": 4973200, "step": 2575 }, { "epoch": 0.42091524594175705, "grad_norm": 2.9040634632110596, "learning_rate": 4.948889857478413e-05, "loss": 0.0815, "num_input_tokens_seen": 4982160, "step": 2580 }, { "epoch": 0.4217309731625744, "grad_norm": 3.3095381259918213, "learning_rate": 4.948692168555978e-05, "loss": 0.0751, "num_input_tokens_seen": 4991168, "step": 2585 }, { "epoch": 0.4225467003833918, "grad_norm": 3.617753744125366, "learning_rate": 4.94849410201444e-05, "loss": 0.1395, "num_input_tokens_seen": 4999760, "step": 2590 }, { "epoch": 0.42336242760420917, "grad_norm": 6.960190296173096, "learning_rate": 4.948295657884341e-05, "loss": 0.3019, "num_input_tokens_seen": 5009984, "step": 2595 }, { "epoch": 0.4241781548250265, "grad_norm": 3.4379990100860596, "learning_rate": 4.9480968361962835e-05, "loss": 0.2789, "num_input_tokens_seen": 5019664, "step": 2600 }, { "epoch": 0.4241781548250265, "eval_loss": 0.1615014225244522, "eval_runtime": 80.6298, "eval_samples_per_second": 33.796, "eval_steps_per_second": 16.904, "num_input_tokens_seen": 5019664, "step": 2600 }, { "epoch": 0.4249938820458439, "grad_norm": 2.6258890628814697, "learning_rate": 4.9478976369809305e-05, "loss": 0.2514, "num_input_tokens_seen": 5029696, "step": 2605 }, { "epoch": 0.4258096092666612, "grad_norm": 3.477609872817993, "learning_rate": 4.947698060268999e-05, "loss": 0.1603, "num_input_tokens_seen": 5038816, "step": 2610 }, { "epoch": 0.4266253364874786, "grad_norm": 4.777493953704834, "learning_rate": 4.9474981060912665e-05, "loss": 0.1575, "num_input_tokens_seen": 5047792, "step": 2615 }, { "epoch": 0.42744106370829593, "grad_norm": 3.608173370361328, "learning_rate": 4.94729777447857e-05, "loss": 0.1816, "num_input_tokens_seen": 5056960, "step": 2620 }, { "epoch": 0.4282567909291133, "grad_norm": 2.1989176273345947, "learning_rate": 4.947097065461801e-05, "loss": 0.1882, "num_input_tokens_seen": 5066912, "step": 2625 }, { "epoch": 0.42907251814993064, "grad_norm": 1.1693536043167114, "learning_rate": 4.9468959790719125e-05, "loss": 0.0784, "num_input_tokens_seen": 5076800, "step": 2630 }, { "epoch": 0.42988824537074805, "grad_norm": 1.0613338947296143, "learning_rate": 4.9466945153399146e-05, "loss": 0.14, "num_input_tokens_seen": 5085456, "step": 2635 }, { "epoch": 0.4307039725915654, "grad_norm": 0.9011128544807434, "learning_rate": 4.9464926742968755e-05, "loss": 0.1354, "num_input_tokens_seen": 5095712, "step": 2640 }, { "epoch": 0.43151969981238275, "grad_norm": 3.597947359085083, "learning_rate": 4.946290455973921e-05, "loss": 0.1214, "num_input_tokens_seen": 5106272, "step": 2645 }, { "epoch": 0.4323354270332001, "grad_norm": 3.367994785308838, "learning_rate": 4.9460878604022365e-05, "loss": 0.3491, "num_input_tokens_seen": 5115936, "step": 2650 }, { "epoch": 0.43315115425401746, "grad_norm": 2.5142171382904053, "learning_rate": 4.945884887613065e-05, "loss": 0.1031, "num_input_tokens_seen": 5125248, "step": 2655 }, { "epoch": 0.4339668814748348, "grad_norm": 2.132192850112915, "learning_rate": 4.9456815376377055e-05, "loss": 0.1837, "num_input_tokens_seen": 5134832, "step": 2660 }, { "epoch": 0.43478260869565216, "grad_norm": 2.168283462524414, "learning_rate": 4.9454778105075195e-05, "loss": 0.0851, "num_input_tokens_seen": 5144400, "step": 2665 }, { "epoch": 0.4355983359164695, "grad_norm": 2.80080246925354, "learning_rate": 4.945273706253924e-05, "loss": 0.0609, "num_input_tokens_seen": 5154448, "step": 2670 }, { "epoch": 0.43641406313728687, "grad_norm": 3.1868531703948975, "learning_rate": 4.9450692249083925e-05, "loss": 0.2015, "num_input_tokens_seen": 5163760, "step": 2675 }, { "epoch": 0.4372297903581043, "grad_norm": 4.800107955932617, "learning_rate": 4.9448643665024605e-05, "loss": 0.1535, "num_input_tokens_seen": 5173168, "step": 2680 }, { "epoch": 0.43804551757892163, "grad_norm": 0.12405548244714737, "learning_rate": 4.944659131067719e-05, "loss": 0.0547, "num_input_tokens_seen": 5183648, "step": 2685 }, { "epoch": 0.438861244799739, "grad_norm": 2.8939900398254395, "learning_rate": 4.944453518635818e-05, "loss": 0.0734, "num_input_tokens_seen": 5192800, "step": 2690 }, { "epoch": 0.43967697202055633, "grad_norm": 5.2967753410339355, "learning_rate": 4.944247529238465e-05, "loss": 0.2466, "num_input_tokens_seen": 5201536, "step": 2695 }, { "epoch": 0.4404926992413737, "grad_norm": 0.4772220849990845, "learning_rate": 4.944041162907427e-05, "loss": 0.1495, "num_input_tokens_seen": 5209584, "step": 2700 }, { "epoch": 0.44130842646219104, "grad_norm": 2.9886202812194824, "learning_rate": 4.943834419674529e-05, "loss": 0.3011, "num_input_tokens_seen": 5219792, "step": 2705 }, { "epoch": 0.4421241536830084, "grad_norm": 5.126646518707275, "learning_rate": 4.9436272995716506e-05, "loss": 0.1767, "num_input_tokens_seen": 5228448, "step": 2710 }, { "epoch": 0.44293988090382574, "grad_norm": 0.27330824732780457, "learning_rate": 4.943419802630735e-05, "loss": 0.0787, "num_input_tokens_seen": 5237888, "step": 2715 }, { "epoch": 0.4437556081246431, "grad_norm": 4.239050388336182, "learning_rate": 4.94321192888378e-05, "loss": 0.2152, "num_input_tokens_seen": 5248512, "step": 2720 }, { "epoch": 0.4445713353454605, "grad_norm": 1.7313318252563477, "learning_rate": 4.943003678362842e-05, "loss": 0.146, "num_input_tokens_seen": 5258448, "step": 2725 }, { "epoch": 0.44538706256627786, "grad_norm": 3.4095561504364014, "learning_rate": 4.942795051100036e-05, "loss": 0.0695, "num_input_tokens_seen": 5267312, "step": 2730 }, { "epoch": 0.4462027897870952, "grad_norm": 2.3634440898895264, "learning_rate": 4.942586047127536e-05, "loss": 0.154, "num_input_tokens_seen": 5277616, "step": 2735 }, { "epoch": 0.44701851700791256, "grad_norm": 4.441982269287109, "learning_rate": 4.942376666477571e-05, "loss": 0.2451, "num_input_tokens_seen": 5287280, "step": 2740 }, { "epoch": 0.4478342442287299, "grad_norm": 5.204430103302002, "learning_rate": 4.9421669091824304e-05, "loss": 0.1477, "num_input_tokens_seen": 5295936, "step": 2745 }, { "epoch": 0.44864997144954727, "grad_norm": 1.7589601278305054, "learning_rate": 4.9419567752744634e-05, "loss": 0.2415, "num_input_tokens_seen": 5306896, "step": 2750 }, { "epoch": 0.4494656986703646, "grad_norm": 2.179410934448242, "learning_rate": 4.941746264786074e-05, "loss": 0.1849, "num_input_tokens_seen": 5316992, "step": 2755 }, { "epoch": 0.450281425891182, "grad_norm": 1.7760694026947021, "learning_rate": 4.9415353777497254e-05, "loss": 0.2242, "num_input_tokens_seen": 5327408, "step": 2760 }, { "epoch": 0.4510971531119993, "grad_norm": 3.401319742202759, "learning_rate": 4.9413241141979394e-05, "loss": 0.178, "num_input_tokens_seen": 5337712, "step": 2765 }, { "epoch": 0.4519128803328167, "grad_norm": 4.5007734298706055, "learning_rate": 4.9411124741632956e-05, "loss": 0.1256, "num_input_tokens_seen": 5347328, "step": 2770 }, { "epoch": 0.4527286075536341, "grad_norm": 0.8746948838233948, "learning_rate": 4.940900457678431e-05, "loss": 0.1199, "num_input_tokens_seen": 5357728, "step": 2775 }, { "epoch": 0.45354433477445144, "grad_norm": 1.344089150428772, "learning_rate": 4.9406880647760425e-05, "loss": 0.1955, "num_input_tokens_seen": 5367968, "step": 2780 }, { "epoch": 0.4543600619952688, "grad_norm": 3.0119545459747314, "learning_rate": 4.9404752954888824e-05, "loss": 0.1102, "num_input_tokens_seen": 5378128, "step": 2785 }, { "epoch": 0.45517578921608615, "grad_norm": 0.3304344415664673, "learning_rate": 4.940262149849762e-05, "loss": 0.0701, "num_input_tokens_seen": 5386736, "step": 2790 }, { "epoch": 0.4559915164369035, "grad_norm": 2.369185209274292, "learning_rate": 4.9400486278915526e-05, "loss": 0.1971, "num_input_tokens_seen": 5397744, "step": 2795 }, { "epoch": 0.45680724365772085, "grad_norm": 5.4565839767456055, "learning_rate": 4.939834729647181e-05, "loss": 0.3232, "num_input_tokens_seen": 5406912, "step": 2800 }, { "epoch": 0.45680724365772085, "eval_loss": 0.1564914733171463, "eval_runtime": 80.5291, "eval_samples_per_second": 33.839, "eval_steps_per_second": 16.926, "num_input_tokens_seen": 5406912, "step": 2800 }, { "epoch": 0.4576229708785382, "grad_norm": 1.9392626285552979, "learning_rate": 4.9396204551496326e-05, "loss": 0.097, "num_input_tokens_seen": 5415840, "step": 2805 }, { "epoch": 0.45843869809935556, "grad_norm": 5.553318500518799, "learning_rate": 4.939405804431952e-05, "loss": 0.1122, "num_input_tokens_seen": 5424816, "step": 2810 }, { "epoch": 0.4592544253201729, "grad_norm": 1.5567224025726318, "learning_rate": 4.9391907775272414e-05, "loss": 0.2002, "num_input_tokens_seen": 5433600, "step": 2815 }, { "epoch": 0.4600701525409903, "grad_norm": 3.800903081893921, "learning_rate": 4.9389753744686604e-05, "loss": 0.0899, "num_input_tokens_seen": 5444016, "step": 2820 }, { "epoch": 0.46088587976180767, "grad_norm": 2.8995983600616455, "learning_rate": 4.938759595289426e-05, "loss": 0.148, "num_input_tokens_seen": 5453776, "step": 2825 }, { "epoch": 0.461701606982625, "grad_norm": 3.154177188873291, "learning_rate": 4.938543440022815e-05, "loss": 0.2506, "num_input_tokens_seen": 5463456, "step": 2830 }, { "epoch": 0.4625173342034424, "grad_norm": 0.26995185017585754, "learning_rate": 4.938326908702161e-05, "loss": 0.0621, "num_input_tokens_seen": 5472496, "step": 2835 }, { "epoch": 0.46333306142425973, "grad_norm": 2.3481571674346924, "learning_rate": 4.9381100013608554e-05, "loss": 0.1483, "num_input_tokens_seen": 5482560, "step": 2840 }, { "epoch": 0.4641487886450771, "grad_norm": 2.3804385662078857, "learning_rate": 4.9378927180323485e-05, "loss": 0.074, "num_input_tokens_seen": 5492368, "step": 2845 }, { "epoch": 0.46496451586589443, "grad_norm": 3.443176031112671, "learning_rate": 4.937675058750148e-05, "loss": 0.2351, "num_input_tokens_seen": 5502048, "step": 2850 }, { "epoch": 0.4657802430867118, "grad_norm": 4.628260135650635, "learning_rate": 4.937457023547819e-05, "loss": 0.1906, "num_input_tokens_seen": 5511888, "step": 2855 }, { "epoch": 0.46659597030752914, "grad_norm": 3.4357028007507324, "learning_rate": 4.9372386124589876e-05, "loss": 0.1908, "num_input_tokens_seen": 5521680, "step": 2860 }, { "epoch": 0.46741169752834655, "grad_norm": 0.6387907862663269, "learning_rate": 4.937019825517333e-05, "loss": 0.0806, "num_input_tokens_seen": 5530688, "step": 2865 }, { "epoch": 0.4682274247491639, "grad_norm": 2.377319097518921, "learning_rate": 4.9368006627565954e-05, "loss": 0.1613, "num_input_tokens_seen": 5539872, "step": 2870 }, { "epoch": 0.46904315196998125, "grad_norm": 1.13958740234375, "learning_rate": 4.936581124210573e-05, "loss": 0.2407, "num_input_tokens_seen": 5549248, "step": 2875 }, { "epoch": 0.4698588791907986, "grad_norm": 4.096313953399658, "learning_rate": 4.9363612099131216e-05, "loss": 0.2723, "num_input_tokens_seen": 5558736, "step": 2880 }, { "epoch": 0.47067460641161596, "grad_norm": 5.320545673370361, "learning_rate": 4.936140919898155e-05, "loss": 0.1938, "num_input_tokens_seen": 5567936, "step": 2885 }, { "epoch": 0.4714903336324333, "grad_norm": 0.3549118936061859, "learning_rate": 4.9359202541996426e-05, "loss": 0.1371, "num_input_tokens_seen": 5576576, "step": 2890 }, { "epoch": 0.47230606085325066, "grad_norm": 0.6741284728050232, "learning_rate": 4.935699212851616e-05, "loss": 0.1857, "num_input_tokens_seen": 5586144, "step": 2895 }, { "epoch": 0.473121788074068, "grad_norm": 1.2061666250228882, "learning_rate": 4.935477795888162e-05, "loss": 0.122, "num_input_tokens_seen": 5596112, "step": 2900 }, { "epoch": 0.47393751529488537, "grad_norm": 0.6635583639144897, "learning_rate": 4.935256003343426e-05, "loss": 0.1331, "num_input_tokens_seen": 5604912, "step": 2905 }, { "epoch": 0.4747532425157027, "grad_norm": 3.5390546321868896, "learning_rate": 4.93503383525161e-05, "loss": 0.1951, "num_input_tokens_seen": 5614848, "step": 2910 }, { "epoch": 0.47556896973652013, "grad_norm": 0.1810443103313446, "learning_rate": 4.934811291646977e-05, "loss": 0.0296, "num_input_tokens_seen": 5624016, "step": 2915 }, { "epoch": 0.4763846969573375, "grad_norm": 2.5612597465515137, "learning_rate": 4.934588372563845e-05, "loss": 0.1144, "num_input_tokens_seen": 5633392, "step": 2920 }, { "epoch": 0.47720042417815484, "grad_norm": 7.0020880699157715, "learning_rate": 4.93436507803659e-05, "loss": 0.2443, "num_input_tokens_seen": 5642928, "step": 2925 }, { "epoch": 0.4780161513989722, "grad_norm": 1.639376163482666, "learning_rate": 4.934141408099649e-05, "loss": 0.2059, "num_input_tokens_seen": 5651776, "step": 2930 }, { "epoch": 0.47883187861978954, "grad_norm": 2.1886119842529297, "learning_rate": 4.9339173627875135e-05, "loss": 0.1083, "num_input_tokens_seen": 5661888, "step": 2935 }, { "epoch": 0.4796476058406069, "grad_norm": 5.976256847381592, "learning_rate": 4.9336929421347335e-05, "loss": 0.3268, "num_input_tokens_seen": 5670592, "step": 2940 }, { "epoch": 0.48046333306142425, "grad_norm": 1.8350181579589844, "learning_rate": 4.933468146175918e-05, "loss": 0.0487, "num_input_tokens_seen": 5678720, "step": 2945 }, { "epoch": 0.4812790602822416, "grad_norm": 2.3004918098449707, "learning_rate": 4.933242974945734e-05, "loss": 0.2619, "num_input_tokens_seen": 5688288, "step": 2950 }, { "epoch": 0.48209478750305895, "grad_norm": 0.5780215263366699, "learning_rate": 4.933017428478906e-05, "loss": 0.0601, "num_input_tokens_seen": 5698160, "step": 2955 }, { "epoch": 0.48291051472387636, "grad_norm": 2.3306679725646973, "learning_rate": 4.932791506810214e-05, "loss": 0.1416, "num_input_tokens_seen": 5707632, "step": 2960 }, { "epoch": 0.4837262419446937, "grad_norm": 3.0666050910949707, "learning_rate": 4.932565209974499e-05, "loss": 0.0438, "num_input_tokens_seen": 5718240, "step": 2965 }, { "epoch": 0.48454196916551107, "grad_norm": 2.329211711883545, "learning_rate": 4.93233853800666e-05, "loss": 0.2323, "num_input_tokens_seen": 5727552, "step": 2970 }, { "epoch": 0.4853576963863284, "grad_norm": 2.399857759475708, "learning_rate": 4.932111490941651e-05, "loss": 0.2556, "num_input_tokens_seen": 5737040, "step": 2975 }, { "epoch": 0.48617342360714577, "grad_norm": 1.7209950685501099, "learning_rate": 4.9318840688144876e-05, "loss": 0.1233, "num_input_tokens_seen": 5747696, "step": 2980 }, { "epoch": 0.4869891508279631, "grad_norm": 1.906476616859436, "learning_rate": 4.9316562716602387e-05, "loss": 0.197, "num_input_tokens_seen": 5757488, "step": 2985 }, { "epoch": 0.4878048780487805, "grad_norm": 3.324341297149658, "learning_rate": 4.9314280995140346e-05, "loss": 0.2009, "num_input_tokens_seen": 5766368, "step": 2990 }, { "epoch": 0.48862060526959783, "grad_norm": 4.0968217849731445, "learning_rate": 4.931199552411063e-05, "loss": 0.1005, "num_input_tokens_seen": 5776288, "step": 2995 }, { "epoch": 0.4894363324904152, "grad_norm": 1.614572525024414, "learning_rate": 4.930970630386568e-05, "loss": 0.2185, "num_input_tokens_seen": 5786080, "step": 3000 }, { "epoch": 0.4894363324904152, "eval_loss": 0.14820662140846252, "eval_runtime": 80.5851, "eval_samples_per_second": 33.815, "eval_steps_per_second": 16.914, "num_input_tokens_seen": 5786080, "step": 3000 }, { "epoch": 0.4902520597112326, "grad_norm": 3.551241874694824, "learning_rate": 4.9307413334758524e-05, "loss": 0.102, "num_input_tokens_seen": 5796080, "step": 3005 }, { "epoch": 0.49106778693204994, "grad_norm": 1.0762125253677368, "learning_rate": 4.930511661714276e-05, "loss": 0.2125, "num_input_tokens_seen": 5805696, "step": 3010 }, { "epoch": 0.4918835141528673, "grad_norm": 2.372960090637207, "learning_rate": 4.9302816151372576e-05, "loss": 0.1831, "num_input_tokens_seen": 5814576, "step": 3015 }, { "epoch": 0.49269924137368465, "grad_norm": 2.5572612285614014, "learning_rate": 4.930051193780274e-05, "loss": 0.177, "num_input_tokens_seen": 5823520, "step": 3020 }, { "epoch": 0.493514968594502, "grad_norm": 3.4741194248199463, "learning_rate": 4.929820397678858e-05, "loss": 0.1023, "num_input_tokens_seen": 5832624, "step": 3025 }, { "epoch": 0.49433069581531935, "grad_norm": 4.418854236602783, "learning_rate": 4.9295892268686015e-05, "loss": 0.1383, "num_input_tokens_seen": 5841920, "step": 3030 }, { "epoch": 0.4951464230361367, "grad_norm": 0.8202112317085266, "learning_rate": 4.9293576813851536e-05, "loss": 0.1035, "num_input_tokens_seen": 5851552, "step": 3035 }, { "epoch": 0.49596215025695406, "grad_norm": 0.5917057394981384, "learning_rate": 4.929125761264223e-05, "loss": 0.1064, "num_input_tokens_seen": 5861632, "step": 3040 }, { "epoch": 0.4967778774777714, "grad_norm": 4.721303462982178, "learning_rate": 4.928893466541573e-05, "loss": 0.1501, "num_input_tokens_seen": 5872000, "step": 3045 }, { "epoch": 0.49759360469858877, "grad_norm": 2.8229689598083496, "learning_rate": 4.928660797253027e-05, "loss": 0.107, "num_input_tokens_seen": 5882336, "step": 3050 }, { "epoch": 0.4984093319194062, "grad_norm": 6.890613079071045, "learning_rate": 4.928427753434467e-05, "loss": 0.1558, "num_input_tokens_seen": 5892304, "step": 3055 }, { "epoch": 0.4992250591402235, "grad_norm": 4.540615558624268, "learning_rate": 4.9281943351218286e-05, "loss": 0.1724, "num_input_tokens_seen": 5901216, "step": 3060 }, { "epoch": 0.5000407863610409, "grad_norm": 4.315707206726074, "learning_rate": 4.9279605423511095e-05, "loss": 0.13, "num_input_tokens_seen": 5910496, "step": 3065 }, { "epoch": 0.5008565135818582, "grad_norm": 6.459292888641357, "learning_rate": 4.927726375158363e-05, "loss": 0.2389, "num_input_tokens_seen": 5920112, "step": 3070 }, { "epoch": 0.5016722408026756, "grad_norm": 0.41955384612083435, "learning_rate": 4.9274918335797004e-05, "loss": 0.1095, "num_input_tokens_seen": 5929024, "step": 3075 }, { "epoch": 0.5024879680234929, "grad_norm": 4.416693687438965, "learning_rate": 4.927256917651292e-05, "loss": 0.2664, "num_input_tokens_seen": 5938448, "step": 3080 }, { "epoch": 0.5033036952443103, "grad_norm": 2.2961530685424805, "learning_rate": 4.927021627409364e-05, "loss": 0.1262, "num_input_tokens_seen": 5947760, "step": 3085 }, { "epoch": 0.5041194224651276, "grad_norm": 6.18982458114624, "learning_rate": 4.9267859628902005e-05, "loss": 0.1965, "num_input_tokens_seen": 5956624, "step": 3090 }, { "epoch": 0.504935149685945, "grad_norm": 3.466203212738037, "learning_rate": 4.9265499241301454e-05, "loss": 0.2725, "num_input_tokens_seen": 5966304, "step": 3095 }, { "epoch": 0.5057508769067623, "grad_norm": 2.9157326221466064, "learning_rate": 4.926313511165598e-05, "loss": 0.215, "num_input_tokens_seen": 5975632, "step": 3100 }, { "epoch": 0.5065666041275797, "grad_norm": 0.6411234140396118, "learning_rate": 4.926076724033016e-05, "loss": 0.1453, "num_input_tokens_seen": 5985840, "step": 3105 }, { "epoch": 0.507382331348397, "grad_norm": 3.000697612762451, "learning_rate": 4.9258395627689146e-05, "loss": 0.1398, "num_input_tokens_seen": 5996224, "step": 3110 }, { "epoch": 0.5081980585692144, "grad_norm": 1.9581067562103271, "learning_rate": 4.925602027409868e-05, "loss": 0.123, "num_input_tokens_seen": 6005808, "step": 3115 }, { "epoch": 0.5090137857900318, "grad_norm": 2.8008856773376465, "learning_rate": 4.925364117992507e-05, "loss": 0.1747, "num_input_tokens_seen": 6015600, "step": 3120 }, { "epoch": 0.5098295130108492, "grad_norm": 3.199464797973633, "learning_rate": 4.92512583455352e-05, "loss": 0.1692, "num_input_tokens_seen": 6025440, "step": 3125 }, { "epoch": 0.5106452402316666, "grad_norm": 2.41405987739563, "learning_rate": 4.9248871771296536e-05, "loss": 0.2213, "num_input_tokens_seen": 6034560, "step": 3130 }, { "epoch": 0.5114609674524839, "grad_norm": 2.346818447113037, "learning_rate": 4.924648145757711e-05, "loss": 0.1446, "num_input_tokens_seen": 6043312, "step": 3135 }, { "epoch": 0.5122766946733013, "grad_norm": 0.9452957510948181, "learning_rate": 4.924408740474554e-05, "loss": 0.1229, "num_input_tokens_seen": 6052912, "step": 3140 }, { "epoch": 0.5130924218941186, "grad_norm": 3.122060537338257, "learning_rate": 4.924168961317103e-05, "loss": 0.1056, "num_input_tokens_seen": 6062032, "step": 3145 }, { "epoch": 0.513908149114936, "grad_norm": 0.9986603856086731, "learning_rate": 4.9239288083223334e-05, "loss": 0.1185, "num_input_tokens_seen": 6070496, "step": 3150 }, { "epoch": 0.5147238763357533, "grad_norm": 2.29278826713562, "learning_rate": 4.9236882815272803e-05, "loss": 0.1973, "num_input_tokens_seen": 6078736, "step": 3155 }, { "epoch": 0.5155396035565707, "grad_norm": 1.1455597877502441, "learning_rate": 4.9234473809690365e-05, "loss": 0.0805, "num_input_tokens_seen": 6087472, "step": 3160 }, { "epoch": 0.516355330777388, "grad_norm": 0.31629613041877747, "learning_rate": 4.923206106684752e-05, "loss": 0.1437, "num_input_tokens_seen": 6097760, "step": 3165 }, { "epoch": 0.5171710579982054, "grad_norm": 4.097001075744629, "learning_rate": 4.922964458711634e-05, "loss": 0.1506, "num_input_tokens_seen": 6108160, "step": 3170 }, { "epoch": 0.5179867852190227, "grad_norm": 3.7709035873413086, "learning_rate": 4.9227224370869474e-05, "loss": 0.2375, "num_input_tokens_seen": 6117440, "step": 3175 }, { "epoch": 0.5188025124398401, "grad_norm": 0.9901310801506042, "learning_rate": 4.9224800418480155e-05, "loss": 0.1582, "num_input_tokens_seen": 6126880, "step": 3180 }, { "epoch": 0.5196182396606575, "grad_norm": 5.35906982421875, "learning_rate": 4.9222372730322176e-05, "loss": 0.2288, "num_input_tokens_seen": 6137136, "step": 3185 }, { "epoch": 0.5204339668814748, "grad_norm": 4.591630458831787, "learning_rate": 4.921994130676993e-05, "loss": 0.1426, "num_input_tokens_seen": 6148080, "step": 3190 }, { "epoch": 0.5212496941022922, "grad_norm": 1.5423104763031006, "learning_rate": 4.9217506148198366e-05, "loss": 0.1291, "num_input_tokens_seen": 6157840, "step": 3195 }, { "epoch": 0.5220654213231095, "grad_norm": 4.438913822174072, "learning_rate": 4.921506725498302e-05, "loss": 0.2453, "num_input_tokens_seen": 6167600, "step": 3200 }, { "epoch": 0.5220654213231095, "eval_loss": 0.1551085114479065, "eval_runtime": 80.4784, "eval_samples_per_second": 33.86, "eval_steps_per_second": 16.936, "num_input_tokens_seen": 6167600, "step": 3200 }, { "epoch": 0.5228811485439269, "grad_norm": 3.6703076362609863, "learning_rate": 4.9212624627499994e-05, "loss": 0.1741, "num_input_tokens_seen": 6177536, "step": 3205 }, { "epoch": 0.5236968757647442, "grad_norm": 2.683462381362915, "learning_rate": 4.921017826612597e-05, "loss": 0.1413, "num_input_tokens_seen": 6187008, "step": 3210 }, { "epoch": 0.5245126029855617, "grad_norm": 3.870368719100952, "learning_rate": 4.9207728171238223e-05, "loss": 0.1998, "num_input_tokens_seen": 6196736, "step": 3215 }, { "epoch": 0.525328330206379, "grad_norm": 1.1318142414093018, "learning_rate": 4.920527434321458e-05, "loss": 0.2115, "num_input_tokens_seen": 6205408, "step": 3220 }, { "epoch": 0.5261440574271964, "grad_norm": 3.3105297088623047, "learning_rate": 4.920281678243345e-05, "loss": 0.1685, "num_input_tokens_seen": 6215248, "step": 3225 }, { "epoch": 0.5269597846480137, "grad_norm": 0.5964555144309998, "learning_rate": 4.920035548927381e-05, "loss": 0.1173, "num_input_tokens_seen": 6224512, "step": 3230 }, { "epoch": 0.5277755118688311, "grad_norm": 1.1205974817276, "learning_rate": 4.919789046411525e-05, "loss": 0.0745, "num_input_tokens_seen": 6233536, "step": 3235 }, { "epoch": 0.5285912390896484, "grad_norm": 5.821063041687012, "learning_rate": 4.919542170733787e-05, "loss": 0.2178, "num_input_tokens_seen": 6243584, "step": 3240 }, { "epoch": 0.5294069663104658, "grad_norm": 2.9373950958251953, "learning_rate": 4.919294921932242e-05, "loss": 0.1672, "num_input_tokens_seen": 6252672, "step": 3245 }, { "epoch": 0.5302226935312832, "grad_norm": 0.8087368011474609, "learning_rate": 4.919047300045016e-05, "loss": 0.1899, "num_input_tokens_seen": 6262320, "step": 3250 }, { "epoch": 0.5310384207521005, "grad_norm": 2.8918981552124023, "learning_rate": 4.918799305110299e-05, "loss": 0.2278, "num_input_tokens_seen": 6272448, "step": 3255 }, { "epoch": 0.5318541479729179, "grad_norm": 4.511650562286377, "learning_rate": 4.918550937166331e-05, "loss": 0.2702, "num_input_tokens_seen": 6281744, "step": 3260 }, { "epoch": 0.5326698751937352, "grad_norm": 1.4461735486984253, "learning_rate": 4.918302196251415e-05, "loss": 0.1351, "num_input_tokens_seen": 6291568, "step": 3265 }, { "epoch": 0.5334856024145526, "grad_norm": 2.1107871532440186, "learning_rate": 4.91805308240391e-05, "loss": 0.2162, "num_input_tokens_seen": 6301168, "step": 3270 }, { "epoch": 0.5343013296353699, "grad_norm": 5.4295973777771, "learning_rate": 4.9178035956622326e-05, "loss": 0.2648, "num_input_tokens_seen": 6311296, "step": 3275 }, { "epoch": 0.5351170568561873, "grad_norm": 0.9894177913665771, "learning_rate": 4.917553736064857e-05, "loss": 0.1211, "num_input_tokens_seen": 6322032, "step": 3280 }, { "epoch": 0.5359327840770046, "grad_norm": 3.3221254348754883, "learning_rate": 4.917303503650314e-05, "loss": 0.1923, "num_input_tokens_seen": 6332208, "step": 3285 }, { "epoch": 0.536748511297822, "grad_norm": 2.1342341899871826, "learning_rate": 4.917052898457194e-05, "loss": 0.1144, "num_input_tokens_seen": 6341408, "step": 3290 }, { "epoch": 0.5375642385186393, "grad_norm": 3.11775541305542, "learning_rate": 4.916801920524141e-05, "loss": 0.1454, "num_input_tokens_seen": 6350848, "step": 3295 }, { "epoch": 0.5383799657394567, "grad_norm": 0.7882725596427917, "learning_rate": 4.916550569889862e-05, "loss": 0.1029, "num_input_tokens_seen": 6360560, "step": 3300 }, { "epoch": 0.539195692960274, "grad_norm": 1.3684604167938232, "learning_rate": 4.916298846593116e-05, "loss": 0.0888, "num_input_tokens_seen": 6370880, "step": 3305 }, { "epoch": 0.5400114201810915, "grad_norm": 4.003333568572998, "learning_rate": 4.916046750672722e-05, "loss": 0.1111, "num_input_tokens_seen": 6380832, "step": 3310 }, { "epoch": 0.5408271474019088, "grad_norm": 0.11373095214366913, "learning_rate": 4.915794282167559e-05, "loss": 0.1099, "num_input_tokens_seen": 6390848, "step": 3315 }, { "epoch": 0.5416428746227262, "grad_norm": 8.744851112365723, "learning_rate": 4.915541441116558e-05, "loss": 0.2081, "num_input_tokens_seen": 6400128, "step": 3320 }, { "epoch": 0.5424586018435436, "grad_norm": 6.503535747528076, "learning_rate": 4.915288227558711e-05, "loss": 0.2119, "num_input_tokens_seen": 6409824, "step": 3325 }, { "epoch": 0.5432743290643609, "grad_norm": 2.8347294330596924, "learning_rate": 4.915034641533066e-05, "loss": 0.1812, "num_input_tokens_seen": 6418976, "step": 3330 }, { "epoch": 0.5440900562851783, "grad_norm": 2.438267946243286, "learning_rate": 4.914780683078731e-05, "loss": 0.1213, "num_input_tokens_seen": 6428384, "step": 3335 }, { "epoch": 0.5449057835059956, "grad_norm": 3.6933469772338867, "learning_rate": 4.9145263522348695e-05, "loss": 0.1227, "num_input_tokens_seen": 6437888, "step": 3340 }, { "epoch": 0.545721510726813, "grad_norm": 0.4935023784637451, "learning_rate": 4.9142716490407e-05, "loss": 0.1498, "num_input_tokens_seen": 6447712, "step": 3345 }, { "epoch": 0.5465372379476303, "grad_norm": 3.6719777584075928, "learning_rate": 4.914016573535504e-05, "loss": 0.136, "num_input_tokens_seen": 6456864, "step": 3350 }, { "epoch": 0.5473529651684477, "grad_norm": 2.2959506511688232, "learning_rate": 4.9137611257586154e-05, "loss": 0.1683, "num_input_tokens_seen": 6465824, "step": 3355 }, { "epoch": 0.548168692389265, "grad_norm": 3.3483498096466064, "learning_rate": 4.9135053057494274e-05, "loss": 0.1031, "num_input_tokens_seen": 6475584, "step": 3360 }, { "epoch": 0.5489844196100824, "grad_norm": 4.142162799835205, "learning_rate": 4.913249113547392e-05, "loss": 0.0959, "num_input_tokens_seen": 6485264, "step": 3365 }, { "epoch": 0.5498001468308997, "grad_norm": 4.0762553215026855, "learning_rate": 4.912992549192016e-05, "loss": 0.1577, "num_input_tokens_seen": 6494736, "step": 3370 }, { "epoch": 0.5506158740517171, "grad_norm": 1.166689157485962, "learning_rate": 4.9127356127228665e-05, "loss": 0.171, "num_input_tokens_seen": 6504512, "step": 3375 }, { "epoch": 0.5514316012725344, "grad_norm": 4.085173606872559, "learning_rate": 4.912478304179564e-05, "loss": 0.1174, "num_input_tokens_seen": 6514544, "step": 3380 }, { "epoch": 0.5522473284933518, "grad_norm": 3.1466543674468994, "learning_rate": 4.9122206236017896e-05, "loss": 0.165, "num_input_tokens_seen": 6525504, "step": 3385 }, { "epoch": 0.5530630557141691, "grad_norm": 4.53146505355835, "learning_rate": 4.911962571029282e-05, "loss": 0.1595, "num_input_tokens_seen": 6535584, "step": 3390 }, { "epoch": 0.5538787829349865, "grad_norm": 2.1106245517730713, "learning_rate": 4.9117041465018353e-05, "loss": 0.2362, "num_input_tokens_seen": 6545104, "step": 3395 }, { "epoch": 0.5546945101558038, "grad_norm": 1.685410976409912, "learning_rate": 4.911445350059302e-05, "loss": 0.1106, "num_input_tokens_seen": 6553904, "step": 3400 }, { "epoch": 0.5546945101558038, "eval_loss": 0.1407547891139984, "eval_runtime": 80.6066, "eval_samples_per_second": 33.806, "eval_steps_per_second": 16.909, "num_input_tokens_seen": 6553904, "step": 3400 }, { "epoch": 0.5555102373766213, "grad_norm": 1.6961088180541992, "learning_rate": 4.9111861817415905e-05, "loss": 0.1278, "num_input_tokens_seen": 6563056, "step": 3405 }, { "epoch": 0.5563259645974387, "grad_norm": 1.0433721542358398, "learning_rate": 4.91092664158867e-05, "loss": 0.0628, "num_input_tokens_seen": 6571568, "step": 3410 }, { "epoch": 0.557141691818256, "grad_norm": 2.7166194915771484, "learning_rate": 4.910666729640563e-05, "loss": 0.1078, "num_input_tokens_seen": 6581072, "step": 3415 }, { "epoch": 0.5579574190390734, "grad_norm": 3.790647506713867, "learning_rate": 4.910406445937353e-05, "loss": 0.2131, "num_input_tokens_seen": 6591136, "step": 3420 }, { "epoch": 0.5587731462598907, "grad_norm": 1.1787406206130981, "learning_rate": 4.9101457905191774e-05, "loss": 0.1129, "num_input_tokens_seen": 6600528, "step": 3425 }, { "epoch": 0.5595888734807081, "grad_norm": 6.369274139404297, "learning_rate": 4.909884763426233e-05, "loss": 0.1903, "num_input_tokens_seen": 6609536, "step": 3430 }, { "epoch": 0.5604046007015254, "grad_norm": 7.302731990814209, "learning_rate": 4.9096233646987736e-05, "loss": 0.2874, "num_input_tokens_seen": 6619600, "step": 3435 }, { "epoch": 0.5612203279223428, "grad_norm": 3.3084843158721924, "learning_rate": 4.9093615943771104e-05, "loss": 0.1775, "num_input_tokens_seen": 6628928, "step": 3440 }, { "epoch": 0.5620360551431601, "grad_norm": 3.4115612506866455, "learning_rate": 4.909099452501611e-05, "loss": 0.1355, "num_input_tokens_seen": 6637008, "step": 3445 }, { "epoch": 0.5628517823639775, "grad_norm": 0.23730197548866272, "learning_rate": 4.908836939112702e-05, "loss": 0.1131, "num_input_tokens_seen": 6646512, "step": 3450 }, { "epoch": 0.5636675095847948, "grad_norm": 0.4997255504131317, "learning_rate": 4.908574054250865e-05, "loss": 0.178, "num_input_tokens_seen": 6657168, "step": 3455 }, { "epoch": 0.5644832368056122, "grad_norm": 0.5094509720802307, "learning_rate": 4.9083107979566414e-05, "loss": 0.2248, "num_input_tokens_seen": 6666944, "step": 3460 }, { "epoch": 0.5652989640264295, "grad_norm": 1.9671709537506104, "learning_rate": 4.908047170270628e-05, "loss": 0.1607, "num_input_tokens_seen": 6677072, "step": 3465 }, { "epoch": 0.5661146912472469, "grad_norm": 4.724954605102539, "learning_rate": 4.9077831712334784e-05, "loss": 0.1523, "num_input_tokens_seen": 6686528, "step": 3470 }, { "epoch": 0.5669304184680642, "grad_norm": 0.31526219844818115, "learning_rate": 4.907518800885907e-05, "loss": 0.1534, "num_input_tokens_seen": 6696624, "step": 3475 }, { "epoch": 0.5677461456888816, "grad_norm": 5.740746974945068, "learning_rate": 4.907254059268681e-05, "loss": 0.267, "num_input_tokens_seen": 6705376, "step": 3480 }, { "epoch": 0.568561872909699, "grad_norm": 3.421851396560669, "learning_rate": 4.906988946422628e-05, "loss": 0.1526, "num_input_tokens_seen": 6715168, "step": 3485 }, { "epoch": 0.5693776001305163, "grad_norm": 3.7238056659698486, "learning_rate": 4.9067234623886315e-05, "loss": 0.2012, "num_input_tokens_seen": 6725008, "step": 3490 }, { "epoch": 0.5701933273513338, "grad_norm": 0.502248227596283, "learning_rate": 4.9064576072076316e-05, "loss": 0.127, "num_input_tokens_seen": 6735520, "step": 3495 }, { "epoch": 0.5710090545721511, "grad_norm": 4.325214862823486, "learning_rate": 4.906191380920628e-05, "loss": 0.1378, "num_input_tokens_seen": 6744992, "step": 3500 }, { "epoch": 0.5718247817929685, "grad_norm": 1.3171849250793457, "learning_rate": 4.905924783568675e-05, "loss": 0.1029, "num_input_tokens_seen": 6753312, "step": 3505 }, { "epoch": 0.5726405090137858, "grad_norm": 5.915157794952393, "learning_rate": 4.905657815192886e-05, "loss": 0.1901, "num_input_tokens_seen": 6763264, "step": 3510 }, { "epoch": 0.5734562362346032, "grad_norm": 9.157356262207031, "learning_rate": 4.90539047583443e-05, "loss": 0.2079, "num_input_tokens_seen": 6773568, "step": 3515 }, { "epoch": 0.5742719634554205, "grad_norm": 3.2176194190979004, "learning_rate": 4.905122765534534e-05, "loss": 0.137, "num_input_tokens_seen": 6782672, "step": 3520 }, { "epoch": 0.5750876906762379, "grad_norm": 0.793321967124939, "learning_rate": 4.9048546843344846e-05, "loss": 0.1146, "num_input_tokens_seen": 6791728, "step": 3525 }, { "epoch": 0.5759034178970552, "grad_norm": 4.470217704772949, "learning_rate": 4.9045862322756206e-05, "loss": 0.0682, "num_input_tokens_seen": 6801856, "step": 3530 }, { "epoch": 0.5767191451178726, "grad_norm": 0.4832174777984619, "learning_rate": 4.904317409399342e-05, "loss": 0.1976, "num_input_tokens_seen": 6812272, "step": 3535 }, { "epoch": 0.5775348723386899, "grad_norm": 1.7956924438476562, "learning_rate": 4.904048215747104e-05, "loss": 0.1855, "num_input_tokens_seen": 6821552, "step": 3540 }, { "epoch": 0.5783505995595073, "grad_norm": 0.7552207112312317, "learning_rate": 4.90377865136042e-05, "loss": 0.1511, "num_input_tokens_seen": 6831040, "step": 3545 }, { "epoch": 0.5791663267803246, "grad_norm": 5.582232475280762, "learning_rate": 4.90350871628086e-05, "loss": 0.2709, "num_input_tokens_seen": 6840832, "step": 3550 }, { "epoch": 0.579982054001142, "grad_norm": 3.9449994564056396, "learning_rate": 4.903238410550052e-05, "loss": 0.1328, "num_input_tokens_seen": 6850832, "step": 3555 }, { "epoch": 0.5807977812219594, "grad_norm": 3.3539955615997314, "learning_rate": 4.90296773420968e-05, "loss": 0.1243, "num_input_tokens_seen": 6861296, "step": 3560 }, { "epoch": 0.5816135084427767, "grad_norm": 4.2276291847229, "learning_rate": 4.902696687301486e-05, "loss": 0.1325, "num_input_tokens_seen": 6870480, "step": 3565 }, { "epoch": 0.5824292356635941, "grad_norm": 0.8770161867141724, "learning_rate": 4.902425269867268e-05, "loss": 0.1645, "num_input_tokens_seen": 6879648, "step": 3570 }, { "epoch": 0.5832449628844114, "grad_norm": 3.911221504211426, "learning_rate": 4.902153481948883e-05, "loss": 0.1243, "num_input_tokens_seen": 6889728, "step": 3575 }, { "epoch": 0.5840606901052288, "grad_norm": 0.7289730906486511, "learning_rate": 4.901881323588244e-05, "loss": 0.1267, "num_input_tokens_seen": 6898416, "step": 3580 }, { "epoch": 0.5848764173260461, "grad_norm": 0.2815453112125397, "learning_rate": 4.90160879482732e-05, "loss": 0.0674, "num_input_tokens_seen": 6907072, "step": 3585 }, { "epoch": 0.5856921445468636, "grad_norm": 0.9623736143112183, "learning_rate": 4.9013358957081405e-05, "loss": 0.1122, "num_input_tokens_seen": 6916880, "step": 3590 }, { "epoch": 0.5865078717676809, "grad_norm": 0.3805393576622009, "learning_rate": 4.901062626272789e-05, "loss": 0.0949, "num_input_tokens_seen": 6927504, "step": 3595 }, { "epoch": 0.5873235989884983, "grad_norm": 0.3686760365962982, "learning_rate": 4.900788986563406e-05, "loss": 0.0803, "num_input_tokens_seen": 6936656, "step": 3600 }, { "epoch": 0.5873235989884983, "eval_loss": 0.1551746428012848, "eval_runtime": 80.6369, "eval_samples_per_second": 33.793, "eval_steps_per_second": 16.903, "num_input_tokens_seen": 6936656, "step": 3600 }, { "epoch": 0.5881393262093156, "grad_norm": 3.8675129413604736, "learning_rate": 4.9005149766221915e-05, "loss": 0.1958, "num_input_tokens_seen": 6946256, "step": 3605 }, { "epoch": 0.588955053430133, "grad_norm": 0.8626136183738708, "learning_rate": 4.9002405964914e-05, "loss": 0.096, "num_input_tokens_seen": 6955936, "step": 3610 }, { "epoch": 0.5897707806509503, "grad_norm": 3.679800033569336, "learning_rate": 4.899965846213346e-05, "loss": 0.3361, "num_input_tokens_seen": 6965504, "step": 3615 }, { "epoch": 0.5905865078717677, "grad_norm": 0.2893342077732086, "learning_rate": 4.899690725830399e-05, "loss": 0.1923, "num_input_tokens_seen": 6974768, "step": 3620 }, { "epoch": 0.591402235092585, "grad_norm": 2.6497230529785156, "learning_rate": 4.899415235384985e-05, "loss": 0.1875, "num_input_tokens_seen": 6985872, "step": 3625 }, { "epoch": 0.5922179623134024, "grad_norm": 2.3696537017822266, "learning_rate": 4.899139374919589e-05, "loss": 0.3257, "num_input_tokens_seen": 6995248, "step": 3630 }, { "epoch": 0.5930336895342198, "grad_norm": 4.514228820800781, "learning_rate": 4.898863144476752e-05, "loss": 0.1735, "num_input_tokens_seen": 7004928, "step": 3635 }, { "epoch": 0.5938494167550371, "grad_norm": 5.458553314208984, "learning_rate": 4.898586544099072e-05, "loss": 0.1151, "num_input_tokens_seen": 7013552, "step": 3640 }, { "epoch": 0.5946651439758545, "grad_norm": 0.7432141900062561, "learning_rate": 4.898309573829204e-05, "loss": 0.1319, "num_input_tokens_seen": 7022720, "step": 3645 }, { "epoch": 0.5954808711966718, "grad_norm": 3.1244962215423584, "learning_rate": 4.898032233709862e-05, "loss": 0.1666, "num_input_tokens_seen": 7032624, "step": 3650 }, { "epoch": 0.5962965984174892, "grad_norm": 1.9265172481536865, "learning_rate": 4.8977545237838123e-05, "loss": 0.231, "num_input_tokens_seen": 7043456, "step": 3655 }, { "epoch": 0.5971123256383065, "grad_norm": 1.9855777025222778, "learning_rate": 4.8974764440938836e-05, "loss": 0.1095, "num_input_tokens_seen": 7054048, "step": 3660 }, { "epoch": 0.5979280528591239, "grad_norm": 1.4470611810684204, "learning_rate": 4.897197994682959e-05, "loss": 0.2064, "num_input_tokens_seen": 7063280, "step": 3665 }, { "epoch": 0.5987437800799412, "grad_norm": 4.288484573364258, "learning_rate": 4.8969191755939786e-05, "loss": 0.1693, "num_input_tokens_seen": 7072192, "step": 3670 }, { "epoch": 0.5995595073007586, "grad_norm": 2.750138759613037, "learning_rate": 4.8966399868699396e-05, "loss": 0.1391, "num_input_tokens_seen": 7082288, "step": 3675 }, { "epoch": 0.600375234521576, "grad_norm": 4.737759590148926, "learning_rate": 4.8963604285538965e-05, "loss": 0.115, "num_input_tokens_seen": 7093952, "step": 3680 }, { "epoch": 0.6011909617423934, "grad_norm": 2.9757328033447266, "learning_rate": 4.8960805006889604e-05, "loss": 0.104, "num_input_tokens_seen": 7101680, "step": 3685 }, { "epoch": 0.6020066889632107, "grad_norm": 3.3790760040283203, "learning_rate": 4.8958002033183004e-05, "loss": 0.1921, "num_input_tokens_seen": 7111984, "step": 3690 }, { "epoch": 0.6028224161840281, "grad_norm": 0.38780757784843445, "learning_rate": 4.8955195364851414e-05, "loss": 0.0837, "num_input_tokens_seen": 7120272, "step": 3695 }, { "epoch": 0.6036381434048455, "grad_norm": 2.269780158996582, "learning_rate": 4.895238500232766e-05, "loss": 0.0966, "num_input_tokens_seen": 7129456, "step": 3700 }, { "epoch": 0.6044538706256628, "grad_norm": 5.299923896789551, "learning_rate": 4.8949570946045143e-05, "loss": 0.1672, "num_input_tokens_seen": 7139824, "step": 3705 }, { "epoch": 0.6052695978464802, "grad_norm": 6.501714706420898, "learning_rate": 4.89467531964378e-05, "loss": 0.19, "num_input_tokens_seen": 7148640, "step": 3710 }, { "epoch": 0.6060853250672975, "grad_norm": 5.929072380065918, "learning_rate": 4.894393175394019e-05, "loss": 0.1846, "num_input_tokens_seen": 7157920, "step": 3715 }, { "epoch": 0.6069010522881149, "grad_norm": 2.0287060737609863, "learning_rate": 4.8941106618987406e-05, "loss": 0.1132, "num_input_tokens_seen": 7166928, "step": 3720 }, { "epoch": 0.6077167795089322, "grad_norm": 3.6768293380737305, "learning_rate": 4.893827779201512e-05, "loss": 0.2157, "num_input_tokens_seen": 7176688, "step": 3725 }, { "epoch": 0.6085325067297496, "grad_norm": 3.2829365730285645, "learning_rate": 4.893544527345957e-05, "loss": 0.169, "num_input_tokens_seen": 7185824, "step": 3730 }, { "epoch": 0.6093482339505669, "grad_norm": 2.597349166870117, "learning_rate": 4.8932609063757563e-05, "loss": 0.211, "num_input_tokens_seen": 7197040, "step": 3735 }, { "epoch": 0.6101639611713843, "grad_norm": 2.3274824619293213, "learning_rate": 4.8929769163346484e-05, "loss": 0.2348, "num_input_tokens_seen": 7207136, "step": 3740 }, { "epoch": 0.6109796883922016, "grad_norm": 6.169339656829834, "learning_rate": 4.892692557266429e-05, "loss": 0.1403, "num_input_tokens_seen": 7216624, "step": 3745 }, { "epoch": 0.611795415613019, "grad_norm": 6.8795671463012695, "learning_rate": 4.8924078292149464e-05, "loss": 0.1544, "num_input_tokens_seen": 7226320, "step": 3750 }, { "epoch": 0.6126111428338363, "grad_norm": 4.703738212585449, "learning_rate": 4.892122732224114e-05, "loss": 0.1351, "num_input_tokens_seen": 7235728, "step": 3755 }, { "epoch": 0.6134268700546537, "grad_norm": 1.0521414279937744, "learning_rate": 4.8918372663378944e-05, "loss": 0.1098, "num_input_tokens_seen": 7244656, "step": 3760 }, { "epoch": 0.614242597275471, "grad_norm": 0.5880059599876404, "learning_rate": 4.89155143160031e-05, "loss": 0.0645, "num_input_tokens_seen": 7254624, "step": 3765 }, { "epoch": 0.6150583244962884, "grad_norm": 4.690751552581787, "learning_rate": 4.891265228055441e-05, "loss": 0.0989, "num_input_tokens_seen": 7265168, "step": 3770 }, { "epoch": 0.6158740517171059, "grad_norm": 5.46951150894165, "learning_rate": 4.890978655747424e-05, "loss": 0.2385, "num_input_tokens_seen": 7275504, "step": 3775 }, { "epoch": 0.6166897789379232, "grad_norm": 0.0853017270565033, "learning_rate": 4.89069171472045e-05, "loss": 0.2812, "num_input_tokens_seen": 7284864, "step": 3780 }, { "epoch": 0.6175055061587406, "grad_norm": 8.25926685333252, "learning_rate": 4.890404405018772e-05, "loss": 0.2746, "num_input_tokens_seen": 7293616, "step": 3785 }, { "epoch": 0.6183212333795579, "grad_norm": 4.170375823974609, "learning_rate": 4.8901167266866934e-05, "loss": 0.1741, "num_input_tokens_seen": 7303152, "step": 3790 }, { "epoch": 0.6191369606003753, "grad_norm": 4.519219398498535, "learning_rate": 4.88982867976858e-05, "loss": 0.2382, "num_input_tokens_seen": 7312672, "step": 3795 }, { "epoch": 0.6199526878211926, "grad_norm": 0.6724659204483032, "learning_rate": 4.889540264308852e-05, "loss": 0.1782, "num_input_tokens_seen": 7321136, "step": 3800 }, { "epoch": 0.6199526878211926, "eval_loss": 0.1527053713798523, "eval_runtime": 80.8219, "eval_samples_per_second": 33.716, "eval_steps_per_second": 16.864, "num_input_tokens_seen": 7321136, "step": 3800 }, { "epoch": 0.62076841504201, "grad_norm": 2.5497395992279053, "learning_rate": 4.889251480351986e-05, "loss": 0.1917, "num_input_tokens_seen": 7330464, "step": 3805 }, { "epoch": 0.6215841422628273, "grad_norm": 3.6651859283447266, "learning_rate": 4.888962327942517e-05, "loss": 0.1391, "num_input_tokens_seen": 7339760, "step": 3810 }, { "epoch": 0.6223998694836447, "grad_norm": 0.6441165804862976, "learning_rate": 4.8886728071250356e-05, "loss": 0.1618, "num_input_tokens_seen": 7351008, "step": 3815 }, { "epoch": 0.623215596704462, "grad_norm": 2.7631776332855225, "learning_rate": 4.8883829179441884e-05, "loss": 0.1693, "num_input_tokens_seen": 7360928, "step": 3820 }, { "epoch": 0.6240313239252794, "grad_norm": 3.169785976409912, "learning_rate": 4.888092660444682e-05, "loss": 0.1884, "num_input_tokens_seen": 7370672, "step": 3825 }, { "epoch": 0.6248470511460967, "grad_norm": 0.22548553347587585, "learning_rate": 4.887802034671276e-05, "loss": 0.161, "num_input_tokens_seen": 7380448, "step": 3830 }, { "epoch": 0.6256627783669141, "grad_norm": 0.44410935044288635, "learning_rate": 4.88751104066879e-05, "loss": 0.1719, "num_input_tokens_seen": 7389392, "step": 3835 }, { "epoch": 0.6264785055877314, "grad_norm": 3.1614513397216797, "learning_rate": 4.887219678482098e-05, "loss": 0.1157, "num_input_tokens_seen": 7399984, "step": 3840 }, { "epoch": 0.6272942328085488, "grad_norm": 0.5027387738227844, "learning_rate": 4.8869279481561316e-05, "loss": 0.0636, "num_input_tokens_seen": 7408432, "step": 3845 }, { "epoch": 0.6281099600293661, "grad_norm": 4.561949253082275, "learning_rate": 4.88663584973588e-05, "loss": 0.1475, "num_input_tokens_seen": 7418864, "step": 3850 }, { "epoch": 0.6289256872501835, "grad_norm": 0.3950614035129547, "learning_rate": 4.8863433832663874e-05, "loss": 0.1435, "num_input_tokens_seen": 7426688, "step": 3855 }, { "epoch": 0.6297414144710008, "grad_norm": 0.3233862519264221, "learning_rate": 4.886050548792757e-05, "loss": 0.143, "num_input_tokens_seen": 7436592, "step": 3860 }, { "epoch": 0.6305571416918182, "grad_norm": 0.3350463807582855, "learning_rate": 4.8857573463601465e-05, "loss": 0.0639, "num_input_tokens_seen": 7446512, "step": 3865 }, { "epoch": 0.6313728689126357, "grad_norm": 0.4833509922027588, "learning_rate": 4.885463776013772e-05, "loss": 0.2142, "num_input_tokens_seen": 7456896, "step": 3870 }, { "epoch": 0.632188596133453, "grad_norm": 3.6676464080810547, "learning_rate": 4.8851698377989056e-05, "loss": 0.1849, "num_input_tokens_seen": 7466832, "step": 3875 }, { "epoch": 0.6330043233542704, "grad_norm": 3.7590878009796143, "learning_rate": 4.884875531760876e-05, "loss": 0.1749, "num_input_tokens_seen": 7476528, "step": 3880 }, { "epoch": 0.6338200505750877, "grad_norm": 0.9205212593078613, "learning_rate": 4.88458085794507e-05, "loss": 0.0847, "num_input_tokens_seen": 7485280, "step": 3885 }, { "epoch": 0.6346357777959051, "grad_norm": 2.0356521606445312, "learning_rate": 4.884285816396929e-05, "loss": 0.2444, "num_input_tokens_seen": 7494336, "step": 3890 }, { "epoch": 0.6354515050167224, "grad_norm": 3.3976945877075195, "learning_rate": 4.8839904071619526e-05, "loss": 0.1411, "num_input_tokens_seen": 7503792, "step": 3895 }, { "epoch": 0.6362672322375398, "grad_norm": 2.557546377182007, "learning_rate": 4.8836946302856955e-05, "loss": 0.1612, "num_input_tokens_seen": 7514912, "step": 3900 }, { "epoch": 0.6370829594583571, "grad_norm": 3.6499030590057373, "learning_rate": 4.8833984858137715e-05, "loss": 0.1139, "num_input_tokens_seen": 7524112, "step": 3905 }, { "epoch": 0.6378986866791745, "grad_norm": 2.0728607177734375, "learning_rate": 4.8831019737918494e-05, "loss": 0.0936, "num_input_tokens_seen": 7534912, "step": 3910 }, { "epoch": 0.6387144138999918, "grad_norm": 0.16674146056175232, "learning_rate": 4.882805094265655e-05, "loss": 0.0391, "num_input_tokens_seen": 7544112, "step": 3915 }, { "epoch": 0.6395301411208092, "grad_norm": 2.358222723007202, "learning_rate": 4.8825078472809706e-05, "loss": 0.1545, "num_input_tokens_seen": 7553424, "step": 3920 }, { "epoch": 0.6403458683416265, "grad_norm": 0.5397843718528748, "learning_rate": 4.882210232883635e-05, "loss": 0.1768, "num_input_tokens_seen": 7563616, "step": 3925 }, { "epoch": 0.6411615955624439, "grad_norm": 1.6982215642929077, "learning_rate": 4.881912251119546e-05, "loss": 0.1399, "num_input_tokens_seen": 7573824, "step": 3930 }, { "epoch": 0.6419773227832613, "grad_norm": 6.231251239776611, "learning_rate": 4.881613902034654e-05, "loss": 0.1239, "num_input_tokens_seen": 7583456, "step": 3935 }, { "epoch": 0.6427930500040786, "grad_norm": 1.3060141801834106, "learning_rate": 4.88131518567497e-05, "loss": 0.3056, "num_input_tokens_seen": 7593280, "step": 3940 }, { "epoch": 0.643608777224896, "grad_norm": 3.620422124862671, "learning_rate": 4.881016102086558e-05, "loss": 0.1046, "num_input_tokens_seen": 7602976, "step": 3945 }, { "epoch": 0.6444245044457133, "grad_norm": 2.013489007949829, "learning_rate": 4.8807166513155425e-05, "loss": 0.2312, "num_input_tokens_seen": 7613328, "step": 3950 }, { "epoch": 0.6452402316665307, "grad_norm": 3.4957027435302734, "learning_rate": 4.8804168334081004e-05, "loss": 0.2819, "num_input_tokens_seen": 7623664, "step": 3955 }, { "epoch": 0.6460559588873481, "grad_norm": 0.5384356379508972, "learning_rate": 4.880116648410468e-05, "loss": 0.1569, "num_input_tokens_seen": 7633552, "step": 3960 }, { "epoch": 0.6468716861081655, "grad_norm": 3.2785487174987793, "learning_rate": 4.879816096368939e-05, "loss": 0.1205, "num_input_tokens_seen": 7642912, "step": 3965 }, { "epoch": 0.6476874133289828, "grad_norm": 4.629924297332764, "learning_rate": 4.879515177329861e-05, "loss": 0.1557, "num_input_tokens_seen": 7651792, "step": 3970 }, { "epoch": 0.6485031405498002, "grad_norm": 2.818220853805542, "learning_rate": 4.8792138913396394e-05, "loss": 0.1107, "num_input_tokens_seen": 7662016, "step": 3975 }, { "epoch": 0.6493188677706175, "grad_norm": 1.6979237794876099, "learning_rate": 4.8789122384447374e-05, "loss": 0.1878, "num_input_tokens_seen": 7671552, "step": 3980 }, { "epoch": 0.6501345949914349, "grad_norm": 1.2491258382797241, "learning_rate": 4.878610218691673e-05, "loss": 0.2282, "num_input_tokens_seen": 7681152, "step": 3985 }, { "epoch": 0.6509503222122522, "grad_norm": 2.703573703765869, "learning_rate": 4.87830783212702e-05, "loss": 0.2628, "num_input_tokens_seen": 7690960, "step": 3990 }, { "epoch": 0.6517660494330696, "grad_norm": 2.694197654724121, "learning_rate": 4.878005078797413e-05, "loss": 0.1283, "num_input_tokens_seen": 7700032, "step": 3995 }, { "epoch": 0.652581776653887, "grad_norm": 3.0511672496795654, "learning_rate": 4.877701958749539e-05, "loss": 0.1269, "num_input_tokens_seen": 7709856, "step": 4000 }, { "epoch": 0.652581776653887, "eval_loss": 0.1362190544605255, "eval_runtime": 80.5871, "eval_samples_per_second": 33.814, "eval_steps_per_second": 16.913, "num_input_tokens_seen": 7709856, "step": 4000 }, { "epoch": 0.6533975038747043, "grad_norm": 1.4650704860687256, "learning_rate": 4.877398472030142e-05, "loss": 0.1159, "num_input_tokens_seen": 7719216, "step": 4005 }, { "epoch": 0.6542132310955217, "grad_norm": 1.092560052871704, "learning_rate": 4.877094618686024e-05, "loss": 0.1344, "num_input_tokens_seen": 7729584, "step": 4010 }, { "epoch": 0.655028958316339, "grad_norm": 3.0615835189819336, "learning_rate": 4.876790398764045e-05, "loss": 0.1539, "num_input_tokens_seen": 7739328, "step": 4015 }, { "epoch": 0.6558446855371564, "grad_norm": 2.4090230464935303, "learning_rate": 4.8764858123111167e-05, "loss": 0.1764, "num_input_tokens_seen": 7748544, "step": 4020 }, { "epoch": 0.6566604127579737, "grad_norm": 2.8026821613311768, "learning_rate": 4.876180859374212e-05, "loss": 0.0866, "num_input_tokens_seen": 7758400, "step": 4025 }, { "epoch": 0.6574761399787911, "grad_norm": 4.526634693145752, "learning_rate": 4.875875540000357e-05, "loss": 0.0635, "num_input_tokens_seen": 7769088, "step": 4030 }, { "epoch": 0.6582918671996084, "grad_norm": 11.626662254333496, "learning_rate": 4.8755698542366376e-05, "loss": 0.225, "num_input_tokens_seen": 7779616, "step": 4035 }, { "epoch": 0.6591075944204258, "grad_norm": 3.7482352256774902, "learning_rate": 4.875263802130193e-05, "loss": 0.2078, "num_input_tokens_seen": 7790928, "step": 4040 }, { "epoch": 0.6599233216412431, "grad_norm": 8.145423889160156, "learning_rate": 4.8749573837282207e-05, "loss": 0.4037, "num_input_tokens_seen": 7800032, "step": 4045 }, { "epoch": 0.6607390488620605, "grad_norm": 0.5119229555130005, "learning_rate": 4.874650599077974e-05, "loss": 0.125, "num_input_tokens_seen": 7809184, "step": 4050 }, { "epoch": 0.6615547760828779, "grad_norm": 2.29842472076416, "learning_rate": 4.874343448226764e-05, "loss": 0.1635, "num_input_tokens_seen": 7819232, "step": 4055 }, { "epoch": 0.6623705033036953, "grad_norm": 3.1747753620147705, "learning_rate": 4.874035931221955e-05, "loss": 0.152, "num_input_tokens_seen": 7829392, "step": 4060 }, { "epoch": 0.6631862305245126, "grad_norm": 5.154107570648193, "learning_rate": 4.8737280481109724e-05, "loss": 0.1831, "num_input_tokens_seen": 7838640, "step": 4065 }, { "epoch": 0.66400195774533, "grad_norm": 2.6247146129608154, "learning_rate": 4.873419798941294e-05, "loss": 0.0747, "num_input_tokens_seen": 7848912, "step": 4070 }, { "epoch": 0.6648176849661473, "grad_norm": 4.7680439949035645, "learning_rate": 4.873111183760458e-05, "loss": 0.0855, "num_input_tokens_seen": 7860272, "step": 4075 }, { "epoch": 0.6656334121869647, "grad_norm": 6.751440525054932, "learning_rate": 4.8728022026160537e-05, "loss": 0.1194, "num_input_tokens_seen": 7868640, "step": 4080 }, { "epoch": 0.666449139407782, "grad_norm": 1.1776163578033447, "learning_rate": 4.872492855555732e-05, "loss": 0.1715, "num_input_tokens_seen": 7878256, "step": 4085 }, { "epoch": 0.6672648666285994, "grad_norm": 0.6247519850730896, "learning_rate": 4.8721831426271956e-05, "loss": 0.1791, "num_input_tokens_seen": 7886576, "step": 4090 }, { "epoch": 0.6680805938494168, "grad_norm": 2.5957367420196533, "learning_rate": 4.87187306387821e-05, "loss": 0.1414, "num_input_tokens_seen": 7896944, "step": 4095 }, { "epoch": 0.6688963210702341, "grad_norm": 4.266890048980713, "learning_rate": 4.87156261935659e-05, "loss": 0.1119, "num_input_tokens_seen": 7907472, "step": 4100 }, { "epoch": 0.6697120482910515, "grad_norm": 0.5319278836250305, "learning_rate": 4.871251809110211e-05, "loss": 0.1623, "num_input_tokens_seen": 7917328, "step": 4105 }, { "epoch": 0.6705277755118688, "grad_norm": 2.35076904296875, "learning_rate": 4.8709406331870044e-05, "loss": 0.2686, "num_input_tokens_seen": 7927344, "step": 4110 }, { "epoch": 0.6713435027326862, "grad_norm": 4.342655181884766, "learning_rate": 4.8706290916349574e-05, "loss": 0.1247, "num_input_tokens_seen": 7935840, "step": 4115 }, { "epoch": 0.6721592299535035, "grad_norm": 1.0053589344024658, "learning_rate": 4.8703171845021134e-05, "loss": 0.1523, "num_input_tokens_seen": 7945696, "step": 4120 }, { "epoch": 0.6729749571743209, "grad_norm": 2.1029160022735596, "learning_rate": 4.870004911836572e-05, "loss": 0.2719, "num_input_tokens_seen": 7955536, "step": 4125 }, { "epoch": 0.6737906843951382, "grad_norm": 1.8603630065917969, "learning_rate": 4.869692273686489e-05, "loss": 0.1802, "num_input_tokens_seen": 7965200, "step": 4130 }, { "epoch": 0.6746064116159556, "grad_norm": 0.17985346913337708, "learning_rate": 4.869379270100079e-05, "loss": 0.0676, "num_input_tokens_seen": 7975072, "step": 4135 }, { "epoch": 0.6754221388367729, "grad_norm": 0.5448839664459229, "learning_rate": 4.86906590112561e-05, "loss": 0.1093, "num_input_tokens_seen": 7984080, "step": 4140 }, { "epoch": 0.6762378660575903, "grad_norm": 3.4063217639923096, "learning_rate": 4.8687521668114064e-05, "loss": 0.1339, "num_input_tokens_seen": 7993088, "step": 4145 }, { "epoch": 0.6770535932784078, "grad_norm": 4.136303901672363, "learning_rate": 4.868438067205853e-05, "loss": 0.1487, "num_input_tokens_seen": 8002320, "step": 4150 }, { "epoch": 0.6778693204992251, "grad_norm": 6.266186237335205, "learning_rate": 4.8681236023573844e-05, "loss": 0.247, "num_input_tokens_seen": 8012096, "step": 4155 }, { "epoch": 0.6786850477200425, "grad_norm": 1.9196983575820923, "learning_rate": 4.867808772314497e-05, "loss": 0.1219, "num_input_tokens_seen": 8022160, "step": 4160 }, { "epoch": 0.6795007749408598, "grad_norm": 4.915822982788086, "learning_rate": 4.867493577125741e-05, "loss": 0.1279, "num_input_tokens_seen": 8031552, "step": 4165 }, { "epoch": 0.6803165021616772, "grad_norm": 0.5322265625, "learning_rate": 4.867178016839725e-05, "loss": 0.2319, "num_input_tokens_seen": 8041504, "step": 4170 }, { "epoch": 0.6811322293824945, "grad_norm": 2.928309202194214, "learning_rate": 4.8668620915051094e-05, "loss": 0.1193, "num_input_tokens_seen": 8051168, "step": 4175 }, { "epoch": 0.6819479566033119, "grad_norm": 1.39497709274292, "learning_rate": 4.866545801170616e-05, "loss": 0.1761, "num_input_tokens_seen": 8061936, "step": 4180 }, { "epoch": 0.6827636838241292, "grad_norm": 2.3256664276123047, "learning_rate": 4.86622914588502e-05, "loss": 0.2202, "num_input_tokens_seen": 8071936, "step": 4185 }, { "epoch": 0.6835794110449466, "grad_norm": 0.6909777522087097, "learning_rate": 4.865912125697154e-05, "loss": 0.1861, "num_input_tokens_seen": 8081568, "step": 4190 }, { "epoch": 0.6843951382657639, "grad_norm": 3.0979361534118652, "learning_rate": 4.865594740655907e-05, "loss": 0.3101, "num_input_tokens_seen": 8090592, "step": 4195 }, { "epoch": 0.6852108654865813, "grad_norm": 0.45151495933532715, "learning_rate": 4.865276990810222e-05, "loss": 0.0605, "num_input_tokens_seen": 8100560, "step": 4200 }, { "epoch": 0.6852108654865813, "eval_loss": 0.13741271197795868, "eval_runtime": 80.6201, "eval_samples_per_second": 33.801, "eval_steps_per_second": 16.906, "num_input_tokens_seen": 8100560, "step": 4200 }, { "epoch": 0.6860265927073986, "grad_norm": 1.4192712306976318, "learning_rate": 4.8649588762091016e-05, "loss": 0.1465, "num_input_tokens_seen": 8110688, "step": 4205 }, { "epoch": 0.686842319928216, "grad_norm": 3.1347901821136475, "learning_rate": 4.8646403969016016e-05, "loss": 0.1408, "num_input_tokens_seen": 8120448, "step": 4210 }, { "epoch": 0.6876580471490333, "grad_norm": 1.5444177389144897, "learning_rate": 4.864321552936838e-05, "loss": 0.0862, "num_input_tokens_seen": 8130608, "step": 4215 }, { "epoch": 0.6884737743698507, "grad_norm": 0.19252422451972961, "learning_rate": 4.864002344363978e-05, "loss": 0.0827, "num_input_tokens_seen": 8140592, "step": 4220 }, { "epoch": 0.689289501590668, "grad_norm": 2.3816380500793457, "learning_rate": 4.863682771232248e-05, "loss": 0.1154, "num_input_tokens_seen": 8150176, "step": 4225 }, { "epoch": 0.6901052288114854, "grad_norm": 2.4156367778778076, "learning_rate": 4.8633628335909324e-05, "loss": 0.1176, "num_input_tokens_seen": 8161568, "step": 4230 }, { "epoch": 0.6909209560323027, "grad_norm": 5.494925498962402, "learning_rate": 4.8630425314893676e-05, "loss": 0.1603, "num_input_tokens_seen": 8170896, "step": 4235 }, { "epoch": 0.6917366832531202, "grad_norm": 3.7069551944732666, "learning_rate": 4.862721864976948e-05, "loss": 0.1523, "num_input_tokens_seen": 8179616, "step": 4240 }, { "epoch": 0.6925524104739376, "grad_norm": 4.5818400382995605, "learning_rate": 4.862400834103125e-05, "loss": 0.1084, "num_input_tokens_seen": 8188608, "step": 4245 }, { "epoch": 0.6933681376947549, "grad_norm": 1.9326019287109375, "learning_rate": 4.862079438917406e-05, "loss": 0.1255, "num_input_tokens_seen": 8197952, "step": 4250 }, { "epoch": 0.6941838649155723, "grad_norm": 0.22906509041786194, "learning_rate": 4.8617576794693536e-05, "loss": 0.1131, "num_input_tokens_seen": 8206256, "step": 4255 }, { "epoch": 0.6949995921363896, "grad_norm": 0.5365654230117798, "learning_rate": 4.8614355558085875e-05, "loss": 0.0596, "num_input_tokens_seen": 8215408, "step": 4260 }, { "epoch": 0.695815319357207, "grad_norm": 2.553936004638672, "learning_rate": 4.861113067984783e-05, "loss": 0.069, "num_input_tokens_seen": 8225376, "step": 4265 }, { "epoch": 0.6966310465780243, "grad_norm": 0.682935893535614, "learning_rate": 4.860790216047671e-05, "loss": 0.1086, "num_input_tokens_seen": 8234512, "step": 4270 }, { "epoch": 0.6974467737988417, "grad_norm": 3.3546807765960693, "learning_rate": 4.860467000047041e-05, "loss": 0.0756, "num_input_tokens_seen": 8243152, "step": 4275 }, { "epoch": 0.698262501019659, "grad_norm": 1.6133331060409546, "learning_rate": 4.860143420032737e-05, "loss": 0.1886, "num_input_tokens_seen": 8253056, "step": 4280 }, { "epoch": 0.6990782282404764, "grad_norm": 0.23937354981899261, "learning_rate": 4.859819476054657e-05, "loss": 0.2421, "num_input_tokens_seen": 8262416, "step": 4285 }, { "epoch": 0.6998939554612937, "grad_norm": 8.026215553283691, "learning_rate": 4.859495168162758e-05, "loss": 0.1539, "num_input_tokens_seen": 8271856, "step": 4290 }, { "epoch": 0.7007096826821111, "grad_norm": 3.369351387023926, "learning_rate": 4.859170496407054e-05, "loss": 0.2073, "num_input_tokens_seen": 8280048, "step": 4295 }, { "epoch": 0.7015254099029284, "grad_norm": 2.9132862091064453, "learning_rate": 4.8588454608376114e-05, "loss": 0.0703, "num_input_tokens_seen": 8288464, "step": 4300 }, { "epoch": 0.7023411371237458, "grad_norm": 3.724048376083374, "learning_rate": 4.8585200615045555e-05, "loss": 0.2675, "num_input_tokens_seen": 8297616, "step": 4305 }, { "epoch": 0.7031568643445631, "grad_norm": 0.8881903290748596, "learning_rate": 4.8581942984580674e-05, "loss": 0.1532, "num_input_tokens_seen": 8307248, "step": 4310 }, { "epoch": 0.7039725915653805, "grad_norm": 7.1685285568237305, "learning_rate": 4.857868171748384e-05, "loss": 0.1854, "num_input_tokens_seen": 8316384, "step": 4315 }, { "epoch": 0.7047883187861979, "grad_norm": 4.948118209838867, "learning_rate": 4.8575416814257976e-05, "loss": 0.2107, "num_input_tokens_seen": 8326416, "step": 4320 }, { "epoch": 0.7056040460070152, "grad_norm": 0.9349321722984314, "learning_rate": 4.857214827540657e-05, "loss": 0.1006, "num_input_tokens_seen": 8336544, "step": 4325 }, { "epoch": 0.7064197732278326, "grad_norm": 2.6882545948028564, "learning_rate": 4.856887610143367e-05, "loss": 0.1363, "num_input_tokens_seen": 8347376, "step": 4330 }, { "epoch": 0.70723550044865, "grad_norm": 4.172438144683838, "learning_rate": 4.8565600292843896e-05, "loss": 0.1621, "num_input_tokens_seen": 8356672, "step": 4335 }, { "epoch": 0.7080512276694674, "grad_norm": 4.936478137969971, "learning_rate": 4.856232085014241e-05, "loss": 0.1857, "num_input_tokens_seen": 8367408, "step": 4340 }, { "epoch": 0.7088669548902847, "grad_norm": 0.399890273809433, "learning_rate": 4.855903777383495e-05, "loss": 0.1403, "num_input_tokens_seen": 8378032, "step": 4345 }, { "epoch": 0.7096826821111021, "grad_norm": 0.5126187205314636, "learning_rate": 4.85557510644278e-05, "loss": 0.198, "num_input_tokens_seen": 8388080, "step": 4350 }, { "epoch": 0.7104984093319194, "grad_norm": 0.9699304699897766, "learning_rate": 4.855246072242782e-05, "loss": 0.0903, "num_input_tokens_seen": 8398160, "step": 4355 }, { "epoch": 0.7113141365527368, "grad_norm": 1.20470130443573, "learning_rate": 4.8549166748342414e-05, "loss": 0.1252, "num_input_tokens_seen": 8406944, "step": 4360 }, { "epoch": 0.7121298637735541, "grad_norm": 1.1030476093292236, "learning_rate": 4.8545869142679556e-05, "loss": 0.1134, "num_input_tokens_seen": 8415424, "step": 4365 }, { "epoch": 0.7129455909943715, "grad_norm": 4.2523674964904785, "learning_rate": 4.8542567905947776e-05, "loss": 0.2409, "num_input_tokens_seen": 8425488, "step": 4370 }, { "epoch": 0.7137613182151888, "grad_norm": 1.5076823234558105, "learning_rate": 4.853926303865618e-05, "loss": 0.2119, "num_input_tokens_seen": 8434704, "step": 4375 }, { "epoch": 0.7145770454360062, "grad_norm": 0.9500270485877991, "learning_rate": 4.853595454131441e-05, "loss": 0.127, "num_input_tokens_seen": 8443488, "step": 4380 }, { "epoch": 0.7153927726568236, "grad_norm": 2.6300976276397705, "learning_rate": 4.8532642414432674e-05, "loss": 0.1135, "num_input_tokens_seen": 8453120, "step": 4385 }, { "epoch": 0.7162084998776409, "grad_norm": 6.135566711425781, "learning_rate": 4.8529326658521754e-05, "loss": 0.1758, "num_input_tokens_seen": 8463088, "step": 4390 }, { "epoch": 0.7170242270984583, "grad_norm": 4.097250461578369, "learning_rate": 4.8526007274092965e-05, "loss": 0.1626, "num_input_tokens_seen": 8472368, "step": 4395 }, { "epoch": 0.7178399543192756, "grad_norm": 1.7617202997207642, "learning_rate": 4.852268426165822e-05, "loss": 0.1342, "num_input_tokens_seen": 8482208, "step": 4400 }, { "epoch": 0.7178399543192756, "eval_loss": 0.13959182798862457, "eval_runtime": 80.6425, "eval_samples_per_second": 33.791, "eval_steps_per_second": 16.902, "num_input_tokens_seen": 8482208, "step": 4400 }, { "epoch": 0.718655681540093, "grad_norm": 5.042740821838379, "learning_rate": 4.851935762172995e-05, "loss": 0.1165, "num_input_tokens_seen": 8491040, "step": 4405 }, { "epoch": 0.7194714087609103, "grad_norm": 4.390962600708008, "learning_rate": 4.8516027354821175e-05, "loss": 0.197, "num_input_tokens_seen": 8501152, "step": 4410 }, { "epoch": 0.7202871359817277, "grad_norm": 9.274046897888184, "learning_rate": 4.851269346144546e-05, "loss": 0.2401, "num_input_tokens_seen": 8509984, "step": 4415 }, { "epoch": 0.721102863202545, "grad_norm": 5.710137367248535, "learning_rate": 4.850935594211693e-05, "loss": 0.2153, "num_input_tokens_seen": 8519200, "step": 4420 }, { "epoch": 0.7219185904233625, "grad_norm": 0.295536607503891, "learning_rate": 4.850601479735029e-05, "loss": 0.1124, "num_input_tokens_seen": 8528624, "step": 4425 }, { "epoch": 0.7227343176441798, "grad_norm": 1.8885139226913452, "learning_rate": 4.850267002766076e-05, "loss": 0.1351, "num_input_tokens_seen": 8537408, "step": 4430 }, { "epoch": 0.7235500448649972, "grad_norm": 1.5898939371109009, "learning_rate": 4.849932163356417e-05, "loss": 0.1582, "num_input_tokens_seen": 8547408, "step": 4435 }, { "epoch": 0.7243657720858145, "grad_norm": 3.127882719039917, "learning_rate": 4.8495969615576864e-05, "loss": 0.1044, "num_input_tokens_seen": 8557360, "step": 4440 }, { "epoch": 0.7251814993066319, "grad_norm": 1.1207481622695923, "learning_rate": 4.849261397421577e-05, "loss": 0.1262, "num_input_tokens_seen": 8567088, "step": 4445 }, { "epoch": 0.7259972265274492, "grad_norm": 2.5280630588531494, "learning_rate": 4.848925470999839e-05, "loss": 0.1707, "num_input_tokens_seen": 8576320, "step": 4450 }, { "epoch": 0.7268129537482666, "grad_norm": 3.248206615447998, "learning_rate": 4.848589182344273e-05, "loss": 0.2246, "num_input_tokens_seen": 8585392, "step": 4455 }, { "epoch": 0.727628680969084, "grad_norm": 2.7375388145446777, "learning_rate": 4.848252531506742e-05, "loss": 0.1463, "num_input_tokens_seen": 8594656, "step": 4460 }, { "epoch": 0.7284444081899013, "grad_norm": 4.474180698394775, "learning_rate": 4.847915518539161e-05, "loss": 0.1622, "num_input_tokens_seen": 8605664, "step": 4465 }, { "epoch": 0.7292601354107187, "grad_norm": 5.724634170532227, "learning_rate": 4.847578143493501e-05, "loss": 0.1826, "num_input_tokens_seen": 8616608, "step": 4470 }, { "epoch": 0.730075862631536, "grad_norm": 6.1944684982299805, "learning_rate": 4.847240406421789e-05, "loss": 0.3089, "num_input_tokens_seen": 8626272, "step": 4475 }, { "epoch": 0.7308915898523534, "grad_norm": 2.3474655151367188, "learning_rate": 4.84690230737611e-05, "loss": 0.2247, "num_input_tokens_seen": 8636544, "step": 4480 }, { "epoch": 0.7317073170731707, "grad_norm": 0.7543244957923889, "learning_rate": 4.846563846408602e-05, "loss": 0.1593, "num_input_tokens_seen": 8646176, "step": 4485 }, { "epoch": 0.7325230442939881, "grad_norm": 2.308561325073242, "learning_rate": 4.84622502357146e-05, "loss": 0.197, "num_input_tokens_seen": 8656016, "step": 4490 }, { "epoch": 0.7333387715148054, "grad_norm": 2.716529369354248, "learning_rate": 4.8458858389169345e-05, "loss": 0.1635, "num_input_tokens_seen": 8665584, "step": 4495 }, { "epoch": 0.7341544987356228, "grad_norm": 2.734009027481079, "learning_rate": 4.8455462924973334e-05, "loss": 0.1778, "num_input_tokens_seen": 8674192, "step": 4500 }, { "epoch": 0.7349702259564401, "grad_norm": 3.23504376411438, "learning_rate": 4.845206384365018e-05, "loss": 0.1939, "num_input_tokens_seen": 8683696, "step": 4505 }, { "epoch": 0.7357859531772575, "grad_norm": 2.556208848953247, "learning_rate": 4.844866114572405e-05, "loss": 0.1162, "num_input_tokens_seen": 8692272, "step": 4510 }, { "epoch": 0.7366016803980748, "grad_norm": 1.1252615451812744, "learning_rate": 4.8445254831719706e-05, "loss": 0.1621, "num_input_tokens_seen": 8701792, "step": 4515 }, { "epoch": 0.7374174076188923, "grad_norm": 0.9145962595939636, "learning_rate": 4.8441844902162434e-05, "loss": 0.1006, "num_input_tokens_seen": 8710160, "step": 4520 }, { "epoch": 0.7382331348397096, "grad_norm": 1.665910243988037, "learning_rate": 4.843843135757809e-05, "loss": 0.1278, "num_input_tokens_seen": 8720960, "step": 4525 }, { "epoch": 0.739048862060527, "grad_norm": 2.717635154724121, "learning_rate": 4.843501419849308e-05, "loss": 0.1189, "num_input_tokens_seen": 8731152, "step": 4530 }, { "epoch": 0.7398645892813444, "grad_norm": 1.8891475200653076, "learning_rate": 4.8431593425434386e-05, "loss": 0.0827, "num_input_tokens_seen": 8741888, "step": 4535 }, { "epoch": 0.7406803165021617, "grad_norm": 2.427234172821045, "learning_rate": 4.8428169038929526e-05, "loss": 0.1529, "num_input_tokens_seen": 8751728, "step": 4540 }, { "epoch": 0.7414960437229791, "grad_norm": 6.330031871795654, "learning_rate": 4.8424741039506575e-05, "loss": 0.1211, "num_input_tokens_seen": 8760544, "step": 4545 }, { "epoch": 0.7423117709437964, "grad_norm": 2.4435484409332275, "learning_rate": 4.842130942769419e-05, "loss": 0.0864, "num_input_tokens_seen": 8770928, "step": 4550 }, { "epoch": 0.7431274981646138, "grad_norm": 4.529120922088623, "learning_rate": 4.841787420402156e-05, "loss": 0.181, "num_input_tokens_seen": 8780032, "step": 4555 }, { "epoch": 0.7439432253854311, "grad_norm": 0.6497309803962708, "learning_rate": 4.841443536901844e-05, "loss": 0.1343, "num_input_tokens_seen": 8790320, "step": 4560 }, { "epoch": 0.7447589526062485, "grad_norm": 3.723041534423828, "learning_rate": 4.841099292321514e-05, "loss": 0.0821, "num_input_tokens_seen": 8799072, "step": 4565 }, { "epoch": 0.7455746798270658, "grad_norm": 4.3906965255737305, "learning_rate": 4.8407546867142525e-05, "loss": 0.2545, "num_input_tokens_seen": 8807968, "step": 4570 }, { "epoch": 0.7463904070478832, "grad_norm": 0.23931767046451569, "learning_rate": 4.840409720133203e-05, "loss": 0.1927, "num_input_tokens_seen": 8817824, "step": 4575 }, { "epoch": 0.7472061342687005, "grad_norm": 1.8283854722976685, "learning_rate": 4.8400643926315634e-05, "loss": 0.1574, "num_input_tokens_seen": 8827392, "step": 4580 }, { "epoch": 0.7480218614895179, "grad_norm": 1.8128526210784912, "learning_rate": 4.839718704262587e-05, "loss": 0.1894, "num_input_tokens_seen": 8836992, "step": 4585 }, { "epoch": 0.7488375887103352, "grad_norm": 4.573443412780762, "learning_rate": 4.839372655079585e-05, "loss": 0.0774, "num_input_tokens_seen": 8846176, "step": 4590 }, { "epoch": 0.7496533159311526, "grad_norm": 4.988640785217285, "learning_rate": 4.83902624513592e-05, "loss": 0.1649, "num_input_tokens_seen": 8857664, "step": 4595 }, { "epoch": 0.7504690431519699, "grad_norm": 3.9652764797210693, "learning_rate": 4.838679474485014e-05, "loss": 0.1543, "num_input_tokens_seen": 8868016, "step": 4600 }, { "epoch": 0.7504690431519699, "eval_loss": 0.13483543694019318, "eval_runtime": 80.7235, "eval_samples_per_second": 33.757, "eval_steps_per_second": 16.885, "num_input_tokens_seen": 8868016, "step": 4600 }, { "epoch": 0.7512847703727873, "grad_norm": 0.5873069167137146, "learning_rate": 4.838332343180343e-05, "loss": 0.0408, "num_input_tokens_seen": 8877904, "step": 4605 }, { "epoch": 0.7521004975936046, "grad_norm": 2.6304619312286377, "learning_rate": 4.83798485127544e-05, "loss": 0.1372, "num_input_tokens_seen": 8888288, "step": 4610 }, { "epoch": 0.7529162248144221, "grad_norm": 4.439068794250488, "learning_rate": 4.837636998823892e-05, "loss": 0.0769, "num_input_tokens_seen": 8897376, "step": 4615 }, { "epoch": 0.7537319520352395, "grad_norm": 0.2544367015361786, "learning_rate": 4.8372887858793414e-05, "loss": 0.146, "num_input_tokens_seen": 8906240, "step": 4620 }, { "epoch": 0.7545476792560568, "grad_norm": 3.952855348587036, "learning_rate": 4.836940212495489e-05, "loss": 0.2003, "num_input_tokens_seen": 8915712, "step": 4625 }, { "epoch": 0.7553634064768742, "grad_norm": 3.742826223373413, "learning_rate": 4.836591278726087e-05, "loss": 0.1391, "num_input_tokens_seen": 8924240, "step": 4630 }, { "epoch": 0.7561791336976915, "grad_norm": 2.5761172771453857, "learning_rate": 4.836241984624947e-05, "loss": 0.1241, "num_input_tokens_seen": 8933488, "step": 4635 }, { "epoch": 0.7569948609185089, "grad_norm": 3.988523006439209, "learning_rate": 4.8358923302459336e-05, "loss": 0.1306, "num_input_tokens_seen": 8943264, "step": 4640 }, { "epoch": 0.7578105881393262, "grad_norm": 3.8204593658447266, "learning_rate": 4.835542315642968e-05, "loss": 0.0862, "num_input_tokens_seen": 8952912, "step": 4645 }, { "epoch": 0.7586263153601436, "grad_norm": 3.798645257949829, "learning_rate": 4.8351919408700274e-05, "loss": 0.1318, "num_input_tokens_seen": 8963248, "step": 4650 }, { "epoch": 0.7594420425809609, "grad_norm": 3.5681300163269043, "learning_rate": 4.834841205981144e-05, "loss": 0.2171, "num_input_tokens_seen": 8972608, "step": 4655 }, { "epoch": 0.7602577698017783, "grad_norm": 4.404212951660156, "learning_rate": 4.8344901110304054e-05, "loss": 0.2287, "num_input_tokens_seen": 8982064, "step": 4660 }, { "epoch": 0.7610734970225956, "grad_norm": 0.6038201451301575, "learning_rate": 4.8341386560719534e-05, "loss": 0.0803, "num_input_tokens_seen": 8992240, "step": 4665 }, { "epoch": 0.761889224243413, "grad_norm": 0.9359808564186096, "learning_rate": 4.833786841159989e-05, "loss": 0.1018, "num_input_tokens_seen": 9002784, "step": 4670 }, { "epoch": 0.7627049514642303, "grad_norm": 1.4026508331298828, "learning_rate": 4.833434666348765e-05, "loss": 0.2399, "num_input_tokens_seen": 9012240, "step": 4675 }, { "epoch": 0.7635206786850477, "grad_norm": 1.6320940256118774, "learning_rate": 4.833082131692592e-05, "loss": 0.1561, "num_input_tokens_seen": 9021520, "step": 4680 }, { "epoch": 0.764336405905865, "grad_norm": 1.521120309829712, "learning_rate": 4.832729237245835e-05, "loss": 0.2433, "num_input_tokens_seen": 9030752, "step": 4685 }, { "epoch": 0.7651521331266824, "grad_norm": 2.048569679260254, "learning_rate": 4.8323759830629145e-05, "loss": 0.1505, "num_input_tokens_seen": 9039376, "step": 4690 }, { "epoch": 0.7659678603474998, "grad_norm": 0.8252496719360352, "learning_rate": 4.8320223691983066e-05, "loss": 0.1024, "num_input_tokens_seen": 9049424, "step": 4695 }, { "epoch": 0.7667835875683171, "grad_norm": 2.605323076248169, "learning_rate": 4.831668395706544e-05, "loss": 0.1181, "num_input_tokens_seen": 9060480, "step": 4700 }, { "epoch": 0.7675993147891346, "grad_norm": 0.5781183838844299, "learning_rate": 4.8313140626422125e-05, "loss": 0.0458, "num_input_tokens_seen": 9070384, "step": 4705 }, { "epoch": 0.7684150420099519, "grad_norm": 1.0541704893112183, "learning_rate": 4.830959370059956e-05, "loss": 0.0732, "num_input_tokens_seen": 9080016, "step": 4710 }, { "epoch": 0.7692307692307693, "grad_norm": 1.009090542793274, "learning_rate": 4.830604318014472e-05, "loss": 0.0832, "num_input_tokens_seen": 9089440, "step": 4715 }, { "epoch": 0.7700464964515866, "grad_norm": 0.7465659379959106, "learning_rate": 4.830248906560514e-05, "loss": 0.0472, "num_input_tokens_seen": 9098416, "step": 4720 }, { "epoch": 0.770862223672404, "grad_norm": 1.034303903579712, "learning_rate": 4.829893135752891e-05, "loss": 0.1285, "num_input_tokens_seen": 9108000, "step": 4725 }, { "epoch": 0.7716779508932213, "grad_norm": 0.4256904721260071, "learning_rate": 4.829537005646466e-05, "loss": 0.1751, "num_input_tokens_seen": 9117920, "step": 4730 }, { "epoch": 0.7724936781140387, "grad_norm": 1.2676197290420532, "learning_rate": 4.8291805162961615e-05, "loss": 0.0722, "num_input_tokens_seen": 9128704, "step": 4735 }, { "epoch": 0.773309405334856, "grad_norm": 3.3601551055908203, "learning_rate": 4.82882366775695e-05, "loss": 0.1551, "num_input_tokens_seen": 9137024, "step": 4740 }, { "epoch": 0.7741251325556734, "grad_norm": 1.0484751462936401, "learning_rate": 4.828466460083864e-05, "loss": 0.0723, "num_input_tokens_seen": 9146768, "step": 4745 }, { "epoch": 0.7749408597764907, "grad_norm": 0.0913279727101326, "learning_rate": 4.8281088933319877e-05, "loss": 0.0496, "num_input_tokens_seen": 9155968, "step": 4750 }, { "epoch": 0.7757565869973081, "grad_norm": 0.9332654476165771, "learning_rate": 4.827750967556464e-05, "loss": 0.1361, "num_input_tokens_seen": 9166176, "step": 4755 }, { "epoch": 0.7765723142181254, "grad_norm": 0.17551936209201813, "learning_rate": 4.827392682812488e-05, "loss": 0.2683, "num_input_tokens_seen": 9175632, "step": 4760 }, { "epoch": 0.7773880414389428, "grad_norm": 3.053764581680298, "learning_rate": 4.827034039155312e-05, "loss": 0.1048, "num_input_tokens_seen": 9186160, "step": 4765 }, { "epoch": 0.7782037686597602, "grad_norm": 4.091749668121338, "learning_rate": 4.8266750366402445e-05, "loss": 0.0663, "num_input_tokens_seen": 9196096, "step": 4770 }, { "epoch": 0.7790194958805775, "grad_norm": 5.57669734954834, "learning_rate": 4.8263156753226476e-05, "loss": 0.1113, "num_input_tokens_seen": 9204448, "step": 4775 }, { "epoch": 0.7798352231013949, "grad_norm": 1.6489096879959106, "learning_rate": 4.8259559552579394e-05, "loss": 0.2188, "num_input_tokens_seen": 9214240, "step": 4780 }, { "epoch": 0.7806509503222122, "grad_norm": 5.952878475189209, "learning_rate": 4.825595876501593e-05, "loss": 0.1456, "num_input_tokens_seen": 9224528, "step": 4785 }, { "epoch": 0.7814666775430296, "grad_norm": 4.060801029205322, "learning_rate": 4.825235439109137e-05, "loss": 0.3098, "num_input_tokens_seen": 9234496, "step": 4790 }, { "epoch": 0.7822824047638469, "grad_norm": 2.3125877380371094, "learning_rate": 4.824874643136156e-05, "loss": 0.1448, "num_input_tokens_seen": 9244288, "step": 4795 }, { "epoch": 0.7830981319846644, "grad_norm": 4.625110626220703, "learning_rate": 4.824513488638288e-05, "loss": 0.1647, "num_input_tokens_seen": 9254560, "step": 4800 }, { "epoch": 0.7830981319846644, "eval_loss": 0.1545816957950592, "eval_runtime": 80.6241, "eval_samples_per_second": 33.799, "eval_steps_per_second": 16.906, "num_input_tokens_seen": 9254560, "step": 4800 }, { "epoch": 0.7839138592054817, "grad_norm": 1.840641736984253, "learning_rate": 4.8241519756712293e-05, "loss": 0.2541, "num_input_tokens_seen": 9263248, "step": 4805 }, { "epoch": 0.7847295864262991, "grad_norm": 2.911818265914917, "learning_rate": 4.8237901042907285e-05, "loss": 0.2545, "num_input_tokens_seen": 9273568, "step": 4810 }, { "epoch": 0.7855453136471164, "grad_norm": 1.525213599205017, "learning_rate": 4.823427874552591e-05, "loss": 0.0649, "num_input_tokens_seen": 9282384, "step": 4815 }, { "epoch": 0.7863610408679338, "grad_norm": 3.0976057052612305, "learning_rate": 4.823065286512677e-05, "loss": 0.0783, "num_input_tokens_seen": 9291408, "step": 4820 }, { "epoch": 0.7871767680887511, "grad_norm": 2.3862435817718506, "learning_rate": 4.8227023402269025e-05, "loss": 0.173, "num_input_tokens_seen": 9301408, "step": 4825 }, { "epoch": 0.7879924953095685, "grad_norm": 2.7953343391418457, "learning_rate": 4.822339035751239e-05, "loss": 0.1791, "num_input_tokens_seen": 9310272, "step": 4830 }, { "epoch": 0.7888082225303858, "grad_norm": 0.47344309091567993, "learning_rate": 4.8219753731417104e-05, "loss": 0.0962, "num_input_tokens_seen": 9320112, "step": 4835 }, { "epoch": 0.7896239497512032, "grad_norm": 1.6930607557296753, "learning_rate": 4.821611352454401e-05, "loss": 0.1086, "num_input_tokens_seen": 9329584, "step": 4840 }, { "epoch": 0.7904396769720206, "grad_norm": 5.080583572387695, "learning_rate": 4.8212469737454444e-05, "loss": 0.0763, "num_input_tokens_seen": 9338416, "step": 4845 }, { "epoch": 0.7912554041928379, "grad_norm": 4.219272136688232, "learning_rate": 4.820882237071035e-05, "loss": 0.2224, "num_input_tokens_seen": 9348624, "step": 4850 }, { "epoch": 0.7920711314136553, "grad_norm": 6.10535192489624, "learning_rate": 4.820517142487417e-05, "loss": 0.1126, "num_input_tokens_seen": 9356432, "step": 4855 }, { "epoch": 0.7928868586344726, "grad_norm": 0.2164386361837387, "learning_rate": 4.8201516900508956e-05, "loss": 0.1716, "num_input_tokens_seen": 9364816, "step": 4860 }, { "epoch": 0.79370258585529, "grad_norm": 4.652078151702881, "learning_rate": 4.819785879817827e-05, "loss": 0.2729, "num_input_tokens_seen": 9374880, "step": 4865 }, { "epoch": 0.7945183130761073, "grad_norm": 4.6241865158081055, "learning_rate": 4.8194197118446226e-05, "loss": 0.1745, "num_input_tokens_seen": 9383904, "step": 4870 }, { "epoch": 0.7953340402969247, "grad_norm": 2.290480852127075, "learning_rate": 4.819053186187752e-05, "loss": 0.162, "num_input_tokens_seen": 9393504, "step": 4875 }, { "epoch": 0.796149767517742, "grad_norm": 0.6085495352745056, "learning_rate": 4.818686302903736e-05, "loss": 0.1511, "num_input_tokens_seen": 9402576, "step": 4880 }, { "epoch": 0.7969654947385594, "grad_norm": 1.4558889865875244, "learning_rate": 4.818319062049154e-05, "loss": 0.0949, "num_input_tokens_seen": 9412592, "step": 4885 }, { "epoch": 0.7977812219593767, "grad_norm": 3.490464448928833, "learning_rate": 4.817951463680639e-05, "loss": 0.2249, "num_input_tokens_seen": 9422368, "step": 4890 }, { "epoch": 0.7985969491801942, "grad_norm": 2.179405450820923, "learning_rate": 4.817583507854879e-05, "loss": 0.1209, "num_input_tokens_seen": 9432352, "step": 4895 }, { "epoch": 0.7994126764010115, "grad_norm": 5.833091735839844, "learning_rate": 4.817215194628617e-05, "loss": 0.1032, "num_input_tokens_seen": 9441056, "step": 4900 }, { "epoch": 0.8002284036218289, "grad_norm": 2.4020493030548096, "learning_rate": 4.816846524058653e-05, "loss": 0.1052, "num_input_tokens_seen": 9451456, "step": 4905 }, { "epoch": 0.8010441308426463, "grad_norm": 4.465664863586426, "learning_rate": 4.816477496201839e-05, "loss": 0.0877, "num_input_tokens_seen": 9460064, "step": 4910 }, { "epoch": 0.8018598580634636, "grad_norm": 2.61080265045166, "learning_rate": 4.8161081111150845e-05, "loss": 0.295, "num_input_tokens_seen": 9469824, "step": 4915 }, { "epoch": 0.802675585284281, "grad_norm": 1.478885531425476, "learning_rate": 4.815738368855354e-05, "loss": 0.0876, "num_input_tokens_seen": 9477680, "step": 4920 }, { "epoch": 0.8034913125050983, "grad_norm": 4.458764553070068, "learning_rate": 4.815368269479664e-05, "loss": 0.1174, "num_input_tokens_seen": 9486720, "step": 4925 }, { "epoch": 0.8043070397259157, "grad_norm": 3.3011839389801025, "learning_rate": 4.814997813045092e-05, "loss": 0.1597, "num_input_tokens_seen": 9496352, "step": 4930 }, { "epoch": 0.805122766946733, "grad_norm": 2.0186870098114014, "learning_rate": 4.814626999608764e-05, "loss": 0.1015, "num_input_tokens_seen": 9505488, "step": 4935 }, { "epoch": 0.8059384941675504, "grad_norm": 2.355733871459961, "learning_rate": 4.814255829227865e-05, "loss": 0.1497, "num_input_tokens_seen": 9515264, "step": 4940 }, { "epoch": 0.8067542213883677, "grad_norm": 2.8522887229919434, "learning_rate": 4.813884301959635e-05, "loss": 0.1042, "num_input_tokens_seen": 9525760, "step": 4945 }, { "epoch": 0.8075699486091851, "grad_norm": 4.386489391326904, "learning_rate": 4.813512417861368e-05, "loss": 0.2226, "num_input_tokens_seen": 9536320, "step": 4950 }, { "epoch": 0.8083856758300024, "grad_norm": 5.399264812469482, "learning_rate": 4.813140176990411e-05, "loss": 0.0854, "num_input_tokens_seen": 9546672, "step": 4955 }, { "epoch": 0.8092014030508198, "grad_norm": 3.0540225505828857, "learning_rate": 4.8127675794041714e-05, "loss": 0.1436, "num_input_tokens_seen": 9554544, "step": 4960 }, { "epoch": 0.8100171302716371, "grad_norm": 7.095914840698242, "learning_rate": 4.812394625160107e-05, "loss": 0.1546, "num_input_tokens_seen": 9564416, "step": 4965 }, { "epoch": 0.8108328574924545, "grad_norm": 0.5479191541671753, "learning_rate": 4.812021314315732e-05, "loss": 0.1424, "num_input_tokens_seen": 9574768, "step": 4970 }, { "epoch": 0.8116485847132718, "grad_norm": 4.17957067489624, "learning_rate": 4.811647646928616e-05, "loss": 0.1564, "num_input_tokens_seen": 9585856, "step": 4975 }, { "epoch": 0.8124643119340892, "grad_norm": 2.6341397762298584, "learning_rate": 4.8112736230563814e-05, "loss": 0.0644, "num_input_tokens_seen": 9595440, "step": 4980 }, { "epoch": 0.8132800391549067, "grad_norm": 2.2055516242980957, "learning_rate": 4.81089924275671e-05, "loss": 0.0887, "num_input_tokens_seen": 9605424, "step": 4985 }, { "epoch": 0.814095766375724, "grad_norm": 0.44006484746932983, "learning_rate": 4.810524506087335e-05, "loss": 0.1094, "num_input_tokens_seen": 9615808, "step": 4990 }, { "epoch": 0.8149114935965414, "grad_norm": 4.546008110046387, "learning_rate": 4.810149413106044e-05, "loss": 0.2214, "num_input_tokens_seen": 9624432, "step": 4995 }, { "epoch": 0.8157272208173587, "grad_norm": 0.31072524189949036, "learning_rate": 4.809773963870684e-05, "loss": 0.0188, "num_input_tokens_seen": 9634544, "step": 5000 }, { "epoch": 0.8157272208173587, "eval_loss": 0.1498986780643463, "eval_runtime": 80.4601, "eval_samples_per_second": 33.868, "eval_steps_per_second": 16.94, "num_input_tokens_seen": 9634544, "step": 5000 }, { "epoch": 0.8165429480381761, "grad_norm": 0.10580665618181229, "learning_rate": 4.809398158439151e-05, "loss": 0.1071, "num_input_tokens_seen": 9644320, "step": 5005 }, { "epoch": 0.8173586752589934, "grad_norm": 7.004683971405029, "learning_rate": 4.8090219968694005e-05, "loss": 0.1366, "num_input_tokens_seen": 9654656, "step": 5010 }, { "epoch": 0.8181744024798108, "grad_norm": 5.847246170043945, "learning_rate": 4.808645479219442e-05, "loss": 0.1681, "num_input_tokens_seen": 9663280, "step": 5015 }, { "epoch": 0.8189901297006281, "grad_norm": 0.6261177659034729, "learning_rate": 4.8082686055473375e-05, "loss": 0.1812, "num_input_tokens_seen": 9673904, "step": 5020 }, { "epoch": 0.8198058569214455, "grad_norm": 6.074930191040039, "learning_rate": 4.8078913759112066e-05, "loss": 0.1099, "num_input_tokens_seen": 9683392, "step": 5025 }, { "epoch": 0.8206215841422628, "grad_norm": 1.7997792959213257, "learning_rate": 4.807513790369223e-05, "loss": 0.1634, "num_input_tokens_seen": 9693008, "step": 5030 }, { "epoch": 0.8214373113630802, "grad_norm": 6.689655780792236, "learning_rate": 4.8071358489796145e-05, "loss": 0.2687, "num_input_tokens_seen": 9703376, "step": 5035 }, { "epoch": 0.8222530385838975, "grad_norm": 6.271470546722412, "learning_rate": 4.806757551800665e-05, "loss": 0.2111, "num_input_tokens_seen": 9712624, "step": 5040 }, { "epoch": 0.8230687658047149, "grad_norm": 5.511608600616455, "learning_rate": 4.806378898890713e-05, "loss": 0.0469, "num_input_tokens_seen": 9721680, "step": 5045 }, { "epoch": 0.8238844930255322, "grad_norm": 1.6162246465682983, "learning_rate": 4.80599989030815e-05, "loss": 0.1547, "num_input_tokens_seen": 9731312, "step": 5050 }, { "epoch": 0.8247002202463496, "grad_norm": 4.731105327606201, "learning_rate": 4.805620526111426e-05, "loss": 0.2831, "num_input_tokens_seen": 9741312, "step": 5055 }, { "epoch": 0.8255159474671669, "grad_norm": 4.140326499938965, "learning_rate": 4.805240806359042e-05, "loss": 0.3001, "num_input_tokens_seen": 9750560, "step": 5060 }, { "epoch": 0.8263316746879843, "grad_norm": 0.9011407494544983, "learning_rate": 4.804860731109557e-05, "loss": 0.0903, "num_input_tokens_seen": 9758800, "step": 5065 }, { "epoch": 0.8271474019088016, "grad_norm": 0.3067605197429657, "learning_rate": 4.804480300421581e-05, "loss": 0.0894, "num_input_tokens_seen": 9767472, "step": 5070 }, { "epoch": 0.827963129129619, "grad_norm": 0.5962313413619995, "learning_rate": 4.804099514353784e-05, "loss": 0.0994, "num_input_tokens_seen": 9777056, "step": 5075 }, { "epoch": 0.8287788563504365, "grad_norm": 0.4712335467338562, "learning_rate": 4.8037183729648867e-05, "loss": 0.0679, "num_input_tokens_seen": 9785888, "step": 5080 }, { "epoch": 0.8295945835712538, "grad_norm": 3.1823198795318604, "learning_rate": 4.803336876313666e-05, "loss": 0.0839, "num_input_tokens_seen": 9796080, "step": 5085 }, { "epoch": 0.8304103107920712, "grad_norm": 3.8663296699523926, "learning_rate": 4.802955024458953e-05, "loss": 0.1754, "num_input_tokens_seen": 9804512, "step": 5090 }, { "epoch": 0.8312260380128885, "grad_norm": 0.47383931279182434, "learning_rate": 4.802572817459634e-05, "loss": 0.059, "num_input_tokens_seen": 9813984, "step": 5095 }, { "epoch": 0.8320417652337059, "grad_norm": 2.2694787979125977, "learning_rate": 4.802190255374651e-05, "loss": 0.16, "num_input_tokens_seen": 9823504, "step": 5100 }, { "epoch": 0.8328574924545232, "grad_norm": 6.950901985168457, "learning_rate": 4.801807338263e-05, "loss": 0.1818, "num_input_tokens_seen": 9833296, "step": 5105 }, { "epoch": 0.8336732196753406, "grad_norm": 0.3967662751674652, "learning_rate": 4.8014240661837306e-05, "loss": 0.0335, "num_input_tokens_seen": 9843360, "step": 5110 }, { "epoch": 0.8344889468961579, "grad_norm": 4.315240859985352, "learning_rate": 4.80104043919595e-05, "loss": 0.1962, "num_input_tokens_seen": 9853264, "step": 5115 }, { "epoch": 0.8353046741169753, "grad_norm": 4.0286664962768555, "learning_rate": 4.800656457358815e-05, "loss": 0.114, "num_input_tokens_seen": 9861952, "step": 5120 }, { "epoch": 0.8361204013377926, "grad_norm": 3.1694421768188477, "learning_rate": 4.800272120731544e-05, "loss": 0.067, "num_input_tokens_seen": 9871424, "step": 5125 }, { "epoch": 0.83693612855861, "grad_norm": 3.774060010910034, "learning_rate": 4.799887429373404e-05, "loss": 0.1886, "num_input_tokens_seen": 9881344, "step": 5130 }, { "epoch": 0.8377518557794273, "grad_norm": 1.4915474653244019, "learning_rate": 4.79950238334372e-05, "loss": 0.094, "num_input_tokens_seen": 9891104, "step": 5135 }, { "epoch": 0.8385675830002447, "grad_norm": 1.2744650840759277, "learning_rate": 4.799116982701872e-05, "loss": 0.1696, "num_input_tokens_seen": 9900656, "step": 5140 }, { "epoch": 0.839383310221062, "grad_norm": 3.9799931049346924, "learning_rate": 4.7987312275072926e-05, "loss": 0.0525, "num_input_tokens_seen": 9908944, "step": 5145 }, { "epoch": 0.8401990374418794, "grad_norm": 0.35056570172309875, "learning_rate": 4.79834511781947e-05, "loss": 0.1015, "num_input_tokens_seen": 9917344, "step": 5150 }, { "epoch": 0.8410147646626968, "grad_norm": 0.16002899408340454, "learning_rate": 4.797958653697947e-05, "loss": 0.1004, "num_input_tokens_seen": 9925616, "step": 5155 }, { "epoch": 0.8418304918835141, "grad_norm": 10.594069480895996, "learning_rate": 4.7975718352023225e-05, "loss": 0.1359, "num_input_tokens_seen": 9936288, "step": 5160 }, { "epoch": 0.8426462191043315, "grad_norm": 6.265426158905029, "learning_rate": 4.7971846623922476e-05, "loss": 0.2317, "num_input_tokens_seen": 9946224, "step": 5165 }, { "epoch": 0.8434619463251488, "grad_norm": 2.085150718688965, "learning_rate": 4.7967971353274294e-05, "loss": 0.117, "num_input_tokens_seen": 9954896, "step": 5170 }, { "epoch": 0.8442776735459663, "grad_norm": 0.2845820486545563, "learning_rate": 4.79640925406763e-05, "loss": 0.2031, "num_input_tokens_seen": 9964960, "step": 5175 }, { "epoch": 0.8450934007667836, "grad_norm": 4.702597141265869, "learning_rate": 4.796021018672664e-05, "loss": 0.2108, "num_input_tokens_seen": 9975344, "step": 5180 }, { "epoch": 0.845909127987601, "grad_norm": 0.6973244547843933, "learning_rate": 4.795632429202405e-05, "loss": 0.2057, "num_input_tokens_seen": 9984624, "step": 5185 }, { "epoch": 0.8467248552084183, "grad_norm": 3.856330394744873, "learning_rate": 4.795243485716775e-05, "loss": 0.2346, "num_input_tokens_seen": 9995008, "step": 5190 }, { "epoch": 0.8475405824292357, "grad_norm": 3.7143280506134033, "learning_rate": 4.794854188275757e-05, "loss": 0.2088, "num_input_tokens_seen": 10005200, "step": 5195 }, { "epoch": 0.848356309650053, "grad_norm": 1.7199541330337524, "learning_rate": 4.794464536939384e-05, "loss": 0.1237, "num_input_tokens_seen": 10013984, "step": 5200 }, { "epoch": 0.848356309650053, "eval_loss": 0.13520273566246033, "eval_runtime": 80.5014, "eval_samples_per_second": 33.85, "eval_steps_per_second": 16.931, "num_input_tokens_seen": 10013984, "step": 5200 }, { "epoch": 0.8491720368708704, "grad_norm": 3.0086209774017334, "learning_rate": 4.794074531767745e-05, "loss": 0.0808, "num_input_tokens_seen": 10023648, "step": 5205 }, { "epoch": 0.8499877640916877, "grad_norm": 3.2653331756591797, "learning_rate": 4.7936841728209834e-05, "loss": 0.2055, "num_input_tokens_seen": 10031968, "step": 5210 }, { "epoch": 0.8508034913125051, "grad_norm": 2.625401258468628, "learning_rate": 4.7932934601593e-05, "loss": 0.1435, "num_input_tokens_seen": 10041040, "step": 5215 }, { "epoch": 0.8516192185333225, "grad_norm": 5.378385543823242, "learning_rate": 4.792902393842943e-05, "loss": 0.2076, "num_input_tokens_seen": 10051664, "step": 5220 }, { "epoch": 0.8524349457541398, "grad_norm": 2.1942522525787354, "learning_rate": 4.792510973932225e-05, "loss": 0.2355, "num_input_tokens_seen": 10060432, "step": 5225 }, { "epoch": 0.8532506729749572, "grad_norm": 0.9434137940406799, "learning_rate": 4.7921192004875036e-05, "loss": 0.1219, "num_input_tokens_seen": 10070288, "step": 5230 }, { "epoch": 0.8540664001957745, "grad_norm": 2.270223379135132, "learning_rate": 4.791727073569198e-05, "loss": 0.1552, "num_input_tokens_seen": 10079392, "step": 5235 }, { "epoch": 0.8548821274165919, "grad_norm": 0.5697221159934998, "learning_rate": 4.7913345932377775e-05, "loss": 0.0523, "num_input_tokens_seen": 10088320, "step": 5240 }, { "epoch": 0.8556978546374092, "grad_norm": 0.19636160135269165, "learning_rate": 4.790941759553769e-05, "loss": 0.0688, "num_input_tokens_seen": 10097456, "step": 5245 }, { "epoch": 0.8565135818582266, "grad_norm": 2.3194801807403564, "learning_rate": 4.79054857257775e-05, "loss": 0.1684, "num_input_tokens_seen": 10108048, "step": 5250 }, { "epoch": 0.8573293090790439, "grad_norm": 0.21476587653160095, "learning_rate": 4.790155032370357e-05, "loss": 0.104, "num_input_tokens_seen": 10116608, "step": 5255 }, { "epoch": 0.8581450362998613, "grad_norm": 8.721016883850098, "learning_rate": 4.789761138992278e-05, "loss": 0.2, "num_input_tokens_seen": 10125936, "step": 5260 }, { "epoch": 0.8589607635206787, "grad_norm": 0.49516695737838745, "learning_rate": 4.7893668925042565e-05, "loss": 0.0673, "num_input_tokens_seen": 10135856, "step": 5265 }, { "epoch": 0.8597764907414961, "grad_norm": 3.8834545612335205, "learning_rate": 4.78897229296709e-05, "loss": 0.1028, "num_input_tokens_seen": 10145696, "step": 5270 }, { "epoch": 0.8605922179623134, "grad_norm": 0.11218233406543732, "learning_rate": 4.7885773404416315e-05, "loss": 0.1991, "num_input_tokens_seen": 10153936, "step": 5275 }, { "epoch": 0.8614079451831308, "grad_norm": 6.863733768463135, "learning_rate": 4.788182034988786e-05, "loss": 0.1008, "num_input_tokens_seen": 10164192, "step": 5280 }, { "epoch": 0.8622236724039481, "grad_norm": 3.0058929920196533, "learning_rate": 4.787786376669516e-05, "loss": 0.1441, "num_input_tokens_seen": 10173968, "step": 5285 }, { "epoch": 0.8630393996247655, "grad_norm": 3.5830395221710205, "learning_rate": 4.787390365544837e-05, "loss": 0.1142, "num_input_tokens_seen": 10184160, "step": 5290 }, { "epoch": 0.8638551268455829, "grad_norm": 0.7656301259994507, "learning_rate": 4.786994001675818e-05, "loss": 0.1671, "num_input_tokens_seen": 10194352, "step": 5295 }, { "epoch": 0.8646708540664002, "grad_norm": 5.132259368896484, "learning_rate": 4.786597285123584e-05, "loss": 0.142, "num_input_tokens_seen": 10202464, "step": 5300 }, { "epoch": 0.8654865812872176, "grad_norm": 1.9765105247497559, "learning_rate": 4.7862002159493135e-05, "loss": 0.3672, "num_input_tokens_seen": 10211696, "step": 5305 }, { "epoch": 0.8663023085080349, "grad_norm": 4.249225616455078, "learning_rate": 4.785802794214239e-05, "loss": 0.1728, "num_input_tokens_seen": 10221264, "step": 5310 }, { "epoch": 0.8671180357288523, "grad_norm": 4.551701068878174, "learning_rate": 4.7854050199796495e-05, "loss": 0.2327, "num_input_tokens_seen": 10232208, "step": 5315 }, { "epoch": 0.8679337629496696, "grad_norm": 3.1114048957824707, "learning_rate": 4.7850068933068845e-05, "loss": 0.1353, "num_input_tokens_seen": 10242272, "step": 5320 }, { "epoch": 0.868749490170487, "grad_norm": 0.47481420636177063, "learning_rate": 4.7846084142573425e-05, "loss": 0.0422, "num_input_tokens_seen": 10252912, "step": 5325 }, { "epoch": 0.8695652173913043, "grad_norm": 0.5661172866821289, "learning_rate": 4.7842095828924725e-05, "loss": 0.1516, "num_input_tokens_seen": 10261584, "step": 5330 }, { "epoch": 0.8703809446121217, "grad_norm": 3.394564628601074, "learning_rate": 4.783810399273779e-05, "loss": 0.1492, "num_input_tokens_seen": 10271456, "step": 5335 }, { "epoch": 0.871196671832939, "grad_norm": 7.724820137023926, "learning_rate": 4.7834108634628226e-05, "loss": 0.1705, "num_input_tokens_seen": 10281296, "step": 5340 }, { "epoch": 0.8720123990537564, "grad_norm": 1.8087080717086792, "learning_rate": 4.783010975521216e-05, "loss": 0.223, "num_input_tokens_seen": 10291408, "step": 5345 }, { "epoch": 0.8728281262745737, "grad_norm": 0.4670826494693756, "learning_rate": 4.782610735510626e-05, "loss": 0.1514, "num_input_tokens_seen": 10301232, "step": 5350 }, { "epoch": 0.8736438534953911, "grad_norm": 3.6589407920837402, "learning_rate": 4.782210143492776e-05, "loss": 0.2838, "num_input_tokens_seen": 10310864, "step": 5355 }, { "epoch": 0.8744595807162086, "grad_norm": 1.8293226957321167, "learning_rate": 4.781809199529442e-05, "loss": 0.136, "num_input_tokens_seen": 10321840, "step": 5360 }, { "epoch": 0.8752753079370259, "grad_norm": 2.6135945320129395, "learning_rate": 4.781407903682454e-05, "loss": 0.1353, "num_input_tokens_seen": 10332208, "step": 5365 }, { "epoch": 0.8760910351578433, "grad_norm": 0.7979068160057068, "learning_rate": 4.781006256013698e-05, "loss": 0.0971, "num_input_tokens_seen": 10342848, "step": 5370 }, { "epoch": 0.8769067623786606, "grad_norm": 2.805459499359131, "learning_rate": 4.7806042565851115e-05, "loss": 0.0922, "num_input_tokens_seen": 10352304, "step": 5375 }, { "epoch": 0.877722489599478, "grad_norm": 3.7861948013305664, "learning_rate": 4.7802019054586895e-05, "loss": 0.211, "num_input_tokens_seen": 10361536, "step": 5380 }, { "epoch": 0.8785382168202953, "grad_norm": 7.0739359855651855, "learning_rate": 4.779799202696479e-05, "loss": 0.1129, "num_input_tokens_seen": 10370320, "step": 5385 }, { "epoch": 0.8793539440411127, "grad_norm": 2.320420503616333, "learning_rate": 4.779396148360581e-05, "loss": 0.1539, "num_input_tokens_seen": 10380288, "step": 5390 }, { "epoch": 0.88016967126193, "grad_norm": 0.7168295383453369, "learning_rate": 4.7789927425131517e-05, "loss": 0.0676, "num_input_tokens_seen": 10389024, "step": 5395 }, { "epoch": 0.8809853984827474, "grad_norm": 1.2502236366271973, "learning_rate": 4.778588985216403e-05, "loss": 0.1465, "num_input_tokens_seen": 10397792, "step": 5400 }, { "epoch": 0.8809853984827474, "eval_loss": 0.15891703963279724, "eval_runtime": 80.4562, "eval_samples_per_second": 33.869, "eval_steps_per_second": 16.941, "num_input_tokens_seen": 10397792, "step": 5400 }, { "epoch": 0.8818011257035647, "grad_norm": 1.8683559894561768, "learning_rate": 4.778184876532598e-05, "loss": 0.1919, "num_input_tokens_seen": 10407952, "step": 5405 }, { "epoch": 0.8826168529243821, "grad_norm": 2.2263288497924805, "learning_rate": 4.7777804165240556e-05, "loss": 0.1037, "num_input_tokens_seen": 10416096, "step": 5410 }, { "epoch": 0.8834325801451994, "grad_norm": 3.2507574558258057, "learning_rate": 4.7773756052531485e-05, "loss": 0.0628, "num_input_tokens_seen": 10426176, "step": 5415 }, { "epoch": 0.8842483073660168, "grad_norm": 6.396968841552734, "learning_rate": 4.7769704427823035e-05, "loss": 0.0948, "num_input_tokens_seen": 10435264, "step": 5420 }, { "epoch": 0.8850640345868341, "grad_norm": 0.22850437462329865, "learning_rate": 4.776564929174003e-05, "loss": 0.132, "num_input_tokens_seen": 10444448, "step": 5425 }, { "epoch": 0.8858797618076515, "grad_norm": 0.7065241932868958, "learning_rate": 4.7761590644907806e-05, "loss": 0.1048, "num_input_tokens_seen": 10455152, "step": 5430 }, { "epoch": 0.8866954890284688, "grad_norm": 4.800972938537598, "learning_rate": 4.7757528487952263e-05, "loss": 0.274, "num_input_tokens_seen": 10464848, "step": 5435 }, { "epoch": 0.8875112162492862, "grad_norm": 2.3675737380981445, "learning_rate": 4.7753462821499836e-05, "loss": 0.213, "num_input_tokens_seen": 10474704, "step": 5440 }, { "epoch": 0.8883269434701035, "grad_norm": 0.2686684727668762, "learning_rate": 4.774939364617751e-05, "loss": 0.1102, "num_input_tokens_seen": 10484400, "step": 5445 }, { "epoch": 0.889142670690921, "grad_norm": 3.0900137424468994, "learning_rate": 4.7745320962612795e-05, "loss": 0.1642, "num_input_tokens_seen": 10494480, "step": 5450 }, { "epoch": 0.8899583979117384, "grad_norm": 3.8391315937042236, "learning_rate": 4.7741244771433756e-05, "loss": 0.2939, "num_input_tokens_seen": 10503328, "step": 5455 }, { "epoch": 0.8907741251325557, "grad_norm": 1.6832275390625, "learning_rate": 4.7737165073268985e-05, "loss": 0.1316, "num_input_tokens_seen": 10513008, "step": 5460 }, { "epoch": 0.8915898523533731, "grad_norm": 1.4710336923599243, "learning_rate": 4.7733081868747626e-05, "loss": 0.1072, "num_input_tokens_seen": 10521344, "step": 5465 }, { "epoch": 0.8924055795741904, "grad_norm": 0.4739426374435425, "learning_rate": 4.772899515849936e-05, "loss": 0.0732, "num_input_tokens_seen": 10530736, "step": 5470 }, { "epoch": 0.8932213067950078, "grad_norm": 0.5044422149658203, "learning_rate": 4.7724904943154414e-05, "loss": 0.1055, "num_input_tokens_seen": 10540992, "step": 5475 }, { "epoch": 0.8940370340158251, "grad_norm": 0.2806858420372009, "learning_rate": 4.772081122334354e-05, "loss": 0.1469, "num_input_tokens_seen": 10551120, "step": 5480 }, { "epoch": 0.8948527612366425, "grad_norm": 4.677550315856934, "learning_rate": 4.771671399969806e-05, "loss": 0.1268, "num_input_tokens_seen": 10560752, "step": 5485 }, { "epoch": 0.8956684884574598, "grad_norm": 3.851579189300537, "learning_rate": 4.7712613272849794e-05, "loss": 0.2256, "num_input_tokens_seen": 10570176, "step": 5490 }, { "epoch": 0.8964842156782772, "grad_norm": 0.13420923054218292, "learning_rate": 4.770850904343114e-05, "loss": 0.1186, "num_input_tokens_seen": 10580992, "step": 5495 }, { "epoch": 0.8972999428990945, "grad_norm": 0.9728893637657166, "learning_rate": 4.770440131207502e-05, "loss": 0.0435, "num_input_tokens_seen": 10591296, "step": 5500 }, { "epoch": 0.8981156701199119, "grad_norm": 4.392834186553955, "learning_rate": 4.7700290079414896e-05, "loss": 0.3146, "num_input_tokens_seen": 10600960, "step": 5505 }, { "epoch": 0.8989313973407292, "grad_norm": 4.846619606018066, "learning_rate": 4.769617534608477e-05, "loss": 0.1306, "num_input_tokens_seen": 10610704, "step": 5510 }, { "epoch": 0.8997471245615466, "grad_norm": 0.736326277256012, "learning_rate": 4.7692057112719193e-05, "loss": 0.1226, "num_input_tokens_seen": 10620224, "step": 5515 }, { "epoch": 0.900562851782364, "grad_norm": 1.690291166305542, "learning_rate": 4.7687935379953234e-05, "loss": 0.1824, "num_input_tokens_seen": 10629248, "step": 5520 }, { "epoch": 0.9013785790031813, "grad_norm": 4.207988262176514, "learning_rate": 4.7683810148422534e-05, "loss": 0.1551, "num_input_tokens_seen": 10638400, "step": 5525 }, { "epoch": 0.9021943062239987, "grad_norm": 4.191145420074463, "learning_rate": 4.767968141876324e-05, "loss": 0.2012, "num_input_tokens_seen": 10648176, "step": 5530 }, { "epoch": 0.903010033444816, "grad_norm": 2.995443344116211, "learning_rate": 4.767554919161207e-05, "loss": 0.1248, "num_input_tokens_seen": 10656848, "step": 5535 }, { "epoch": 0.9038257606656334, "grad_norm": 2.8024709224700928, "learning_rate": 4.767141346760624e-05, "loss": 0.0933, "num_input_tokens_seen": 10666944, "step": 5540 }, { "epoch": 0.9046414878864508, "grad_norm": 1.8596307039260864, "learning_rate": 4.766727424738356e-05, "loss": 0.1366, "num_input_tokens_seen": 10676720, "step": 5545 }, { "epoch": 0.9054572151072682, "grad_norm": 3.6329288482666016, "learning_rate": 4.7663131531582325e-05, "loss": 0.2569, "num_input_tokens_seen": 10686896, "step": 5550 }, { "epoch": 0.9062729423280855, "grad_norm": 4.417144775390625, "learning_rate": 4.765898532084142e-05, "loss": 0.0968, "num_input_tokens_seen": 10696560, "step": 5555 }, { "epoch": 0.9070886695489029, "grad_norm": 4.522747993469238, "learning_rate": 4.765483561580022e-05, "loss": 0.2482, "num_input_tokens_seen": 10707296, "step": 5560 }, { "epoch": 0.9079043967697202, "grad_norm": 2.047534942626953, "learning_rate": 4.7650682417098666e-05, "loss": 0.0847, "num_input_tokens_seen": 10715664, "step": 5565 }, { "epoch": 0.9087201239905376, "grad_norm": 1.5404976606369019, "learning_rate": 4.7646525725377244e-05, "loss": 0.139, "num_input_tokens_seen": 10725648, "step": 5570 }, { "epoch": 0.9095358512113549, "grad_norm": 1.1270887851715088, "learning_rate": 4.764236554127696e-05, "loss": 0.1292, "num_input_tokens_seen": 10736192, "step": 5575 }, { "epoch": 0.9103515784321723, "grad_norm": 4.0132060050964355, "learning_rate": 4.7638201865439356e-05, "loss": 0.1649, "num_input_tokens_seen": 10746176, "step": 5580 }, { "epoch": 0.9111673056529896, "grad_norm": 2.2463841438293457, "learning_rate": 4.7634034698506545e-05, "loss": 0.2183, "num_input_tokens_seen": 10755200, "step": 5585 }, { "epoch": 0.911983032873807, "grad_norm": 2.4663426876068115, "learning_rate": 4.762986404112115e-05, "loss": 0.0778, "num_input_tokens_seen": 10764816, "step": 5590 }, { "epoch": 0.9127987600946244, "grad_norm": 1.4591420888900757, "learning_rate": 4.762568989392633e-05, "loss": 0.2251, "num_input_tokens_seen": 10774272, "step": 5595 }, { "epoch": 0.9136144873154417, "grad_norm": 2.291351795196533, "learning_rate": 4.76215122575658e-05, "loss": 0.1414, "num_input_tokens_seen": 10784512, "step": 5600 }, { "epoch": 0.9136144873154417, "eval_loss": 0.1298934519290924, "eval_runtime": 80.5647, "eval_samples_per_second": 33.824, "eval_steps_per_second": 16.918, "num_input_tokens_seen": 10784512, "step": 5600 }, { "epoch": 0.9144302145362591, "grad_norm": 2.4497451782226562, "learning_rate": 4.7617331132683795e-05, "loss": 0.1728, "num_input_tokens_seen": 10794256, "step": 5605 }, { "epoch": 0.9152459417570764, "grad_norm": 2.0188148021698, "learning_rate": 4.7613146519925105e-05, "loss": 0.1124, "num_input_tokens_seen": 10804112, "step": 5610 }, { "epoch": 0.9160616689778938, "grad_norm": 2.0468132495880127, "learning_rate": 4.7608958419935045e-05, "loss": 0.1142, "num_input_tokens_seen": 10812752, "step": 5615 }, { "epoch": 0.9168773961987111, "grad_norm": 0.6019260883331299, "learning_rate": 4.760476683335948e-05, "loss": 0.0951, "num_input_tokens_seen": 10821552, "step": 5620 }, { "epoch": 0.9176931234195285, "grad_norm": 3.437530994415283, "learning_rate": 4.760057176084479e-05, "loss": 0.0831, "num_input_tokens_seen": 10830704, "step": 5625 }, { "epoch": 0.9185088506403458, "grad_norm": 4.703093528747559, "learning_rate": 4.759637320303793e-05, "loss": 0.143, "num_input_tokens_seen": 10840816, "step": 5630 }, { "epoch": 0.9193245778611632, "grad_norm": 2.235828161239624, "learning_rate": 4.759217116058635e-05, "loss": 0.1064, "num_input_tokens_seen": 10849472, "step": 5635 }, { "epoch": 0.9201403050819806, "grad_norm": 1.0159187316894531, "learning_rate": 4.758796563413807e-05, "loss": 0.1775, "num_input_tokens_seen": 10859296, "step": 5640 }, { "epoch": 0.920956032302798, "grad_norm": 5.161415100097656, "learning_rate": 4.758375662434163e-05, "loss": 0.1593, "num_input_tokens_seen": 10868448, "step": 5645 }, { "epoch": 0.9217717595236153, "grad_norm": 1.013604998588562, "learning_rate": 4.7579544131846114e-05, "loss": 0.064, "num_input_tokens_seen": 10877312, "step": 5650 }, { "epoch": 0.9225874867444327, "grad_norm": 0.5827361345291138, "learning_rate": 4.757532815730114e-05, "loss": 0.1064, "num_input_tokens_seen": 10886912, "step": 5655 }, { "epoch": 0.92340321396525, "grad_norm": 3.5018632411956787, "learning_rate": 4.7571108701356865e-05, "loss": 0.1643, "num_input_tokens_seen": 10896032, "step": 5660 }, { "epoch": 0.9242189411860674, "grad_norm": 2.754272699356079, "learning_rate": 4.756688576466398e-05, "loss": 0.1705, "num_input_tokens_seen": 10906400, "step": 5665 }, { "epoch": 0.9250346684068848, "grad_norm": 1.4439191818237305, "learning_rate": 4.756265934787372e-05, "loss": 0.2473, "num_input_tokens_seen": 10915552, "step": 5670 }, { "epoch": 0.9258503956277021, "grad_norm": 3.359255313873291, "learning_rate": 4.755842945163785e-05, "loss": 0.1277, "num_input_tokens_seen": 10924720, "step": 5675 }, { "epoch": 0.9266661228485195, "grad_norm": 3.790789842605591, "learning_rate": 4.755419607660867e-05, "loss": 0.0964, "num_input_tokens_seen": 10934768, "step": 5680 }, { "epoch": 0.9274818500693368, "grad_norm": 1.6309388875961304, "learning_rate": 4.7549959223439016e-05, "loss": 0.1746, "num_input_tokens_seen": 10944336, "step": 5685 }, { "epoch": 0.9282975772901542, "grad_norm": 6.307379245758057, "learning_rate": 4.754571889278228e-05, "loss": 0.3014, "num_input_tokens_seen": 10954480, "step": 5690 }, { "epoch": 0.9291133045109715, "grad_norm": 0.42783334851264954, "learning_rate": 4.754147508529235e-05, "loss": 0.1432, "num_input_tokens_seen": 10964432, "step": 5695 }, { "epoch": 0.9299290317317889, "grad_norm": 1.5752546787261963, "learning_rate": 4.75372278016237e-05, "loss": 0.2148, "num_input_tokens_seen": 10973728, "step": 5700 }, { "epoch": 0.9307447589526062, "grad_norm": 0.16247229278087616, "learning_rate": 4.753297704243129e-05, "loss": 0.1726, "num_input_tokens_seen": 10983376, "step": 5705 }, { "epoch": 0.9315604861734236, "grad_norm": 0.39484113454818726, "learning_rate": 4.752872280837066e-05, "loss": 0.1102, "num_input_tokens_seen": 10993472, "step": 5710 }, { "epoch": 0.9323762133942409, "grad_norm": 2.9760091304779053, "learning_rate": 4.752446510009786e-05, "loss": 0.1326, "num_input_tokens_seen": 11004240, "step": 5715 }, { "epoch": 0.9331919406150583, "grad_norm": 2.9162209033966064, "learning_rate": 4.7520203918269476e-05, "loss": 0.1637, "num_input_tokens_seen": 11013168, "step": 5720 }, { "epoch": 0.9340076678358756, "grad_norm": 4.753259181976318, "learning_rate": 4.751593926354265e-05, "loss": 0.2623, "num_input_tokens_seen": 11022832, "step": 5725 }, { "epoch": 0.9348233950566931, "grad_norm": 2.479592800140381, "learning_rate": 4.751167113657503e-05, "loss": 0.1549, "num_input_tokens_seen": 11031376, "step": 5730 }, { "epoch": 0.9356391222775104, "grad_norm": 1.651119589805603, "learning_rate": 4.7507399538024834e-05, "loss": 0.121, "num_input_tokens_seen": 11041152, "step": 5735 }, { "epoch": 0.9364548494983278, "grad_norm": 2.2114720344543457, "learning_rate": 4.750312446855077e-05, "loss": 0.1488, "num_input_tokens_seen": 11050688, "step": 5740 }, { "epoch": 0.9372705767191452, "grad_norm": 2.6062026023864746, "learning_rate": 4.749884592881212e-05, "loss": 0.2582, "num_input_tokens_seen": 11060400, "step": 5745 }, { "epoch": 0.9380863039399625, "grad_norm": 2.4557442665100098, "learning_rate": 4.74945639194687e-05, "loss": 0.1524, "num_input_tokens_seen": 11070480, "step": 5750 }, { "epoch": 0.9389020311607799, "grad_norm": 0.6860836148262024, "learning_rate": 4.749027844118083e-05, "loss": 0.1472, "num_input_tokens_seen": 11079824, "step": 5755 }, { "epoch": 0.9397177583815972, "grad_norm": 0.6759036779403687, "learning_rate": 4.7485989494609395e-05, "loss": 0.0511, "num_input_tokens_seen": 11090752, "step": 5760 }, { "epoch": 0.9405334856024146, "grad_norm": 0.7650130987167358, "learning_rate": 4.748169708041581e-05, "loss": 0.1013, "num_input_tokens_seen": 11100608, "step": 5765 }, { "epoch": 0.9413492128232319, "grad_norm": 1.1126277446746826, "learning_rate": 4.7477401199262004e-05, "loss": 0.2075, "num_input_tokens_seen": 11110416, "step": 5770 }, { "epoch": 0.9421649400440493, "grad_norm": 0.2707446813583374, "learning_rate": 4.747310185181048e-05, "loss": 0.1825, "num_input_tokens_seen": 11120256, "step": 5775 }, { "epoch": 0.9429806672648666, "grad_norm": 2.8019909858703613, "learning_rate": 4.746879903872422e-05, "loss": 0.1363, "num_input_tokens_seen": 11128672, "step": 5780 }, { "epoch": 0.943796394485684, "grad_norm": 1.2708525657653809, "learning_rate": 4.746449276066679e-05, "loss": 0.1547, "num_input_tokens_seen": 11137248, "step": 5785 }, { "epoch": 0.9446121217065013, "grad_norm": 0.19789566099643707, "learning_rate": 4.746018301830227e-05, "loss": 0.1011, "num_input_tokens_seen": 11146640, "step": 5790 }, { "epoch": 0.9454278489273187, "grad_norm": 2.7243425846099854, "learning_rate": 4.7455869812295275e-05, "loss": 0.259, "num_input_tokens_seen": 11155664, "step": 5795 }, { "epoch": 0.946243576148136, "grad_norm": 0.4580719470977783, "learning_rate": 4.7451553143310964e-05, "loss": 0.1428, "num_input_tokens_seen": 11165168, "step": 5800 }, { "epoch": 0.946243576148136, "eval_loss": 0.12852467596530914, "eval_runtime": 80.6396, "eval_samples_per_second": 33.792, "eval_steps_per_second": 16.902, "num_input_tokens_seen": 11165168, "step": 5800 }, { "epoch": 0.9470593033689534, "grad_norm": 0.5369745492935181, "learning_rate": 4.744723301201501e-05, "loss": 0.1137, "num_input_tokens_seen": 11175088, "step": 5805 }, { "epoch": 0.9478750305897707, "grad_norm": 0.26575082540512085, "learning_rate": 4.744290941907364e-05, "loss": 0.0769, "num_input_tokens_seen": 11184752, "step": 5810 }, { "epoch": 0.9486907578105881, "grad_norm": 4.374102592468262, "learning_rate": 4.7438582365153594e-05, "loss": 0.2378, "num_input_tokens_seen": 11194624, "step": 5815 }, { "epoch": 0.9495064850314054, "grad_norm": 3.1562659740448, "learning_rate": 4.743425185092217e-05, "loss": 0.2069, "num_input_tokens_seen": 11204992, "step": 5820 }, { "epoch": 0.9503222122522229, "grad_norm": 0.6647024750709534, "learning_rate": 4.742991787704719e-05, "loss": 0.081, "num_input_tokens_seen": 11215728, "step": 5825 }, { "epoch": 0.9511379394730403, "grad_norm": 1.2246392965316772, "learning_rate": 4.7425580444196994e-05, "loss": 0.0414, "num_input_tokens_seen": 11226064, "step": 5830 }, { "epoch": 0.9519536666938576, "grad_norm": 2.0717482566833496, "learning_rate": 4.742123955304048e-05, "loss": 0.1472, "num_input_tokens_seen": 11235504, "step": 5835 }, { "epoch": 0.952769393914675, "grad_norm": 1.50772225856781, "learning_rate": 4.741689520424706e-05, "loss": 0.0781, "num_input_tokens_seen": 11244624, "step": 5840 }, { "epoch": 0.9535851211354923, "grad_norm": 0.32258620858192444, "learning_rate": 4.741254739848669e-05, "loss": 0.0142, "num_input_tokens_seen": 11254416, "step": 5845 }, { "epoch": 0.9544008483563097, "grad_norm": 1.9959949254989624, "learning_rate": 4.740819613642987e-05, "loss": 0.1869, "num_input_tokens_seen": 11263408, "step": 5850 }, { "epoch": 0.955216575577127, "grad_norm": 1.760833740234375, "learning_rate": 4.74038414187476e-05, "loss": 0.0345, "num_input_tokens_seen": 11272560, "step": 5855 }, { "epoch": 0.9560323027979444, "grad_norm": 2.8897526264190674, "learning_rate": 4.739948324611144e-05, "loss": 0.181, "num_input_tokens_seen": 11282704, "step": 5860 }, { "epoch": 0.9568480300187617, "grad_norm": 5.540792465209961, "learning_rate": 4.7395121619193465e-05, "loss": 0.2121, "num_input_tokens_seen": 11291712, "step": 5865 }, { "epoch": 0.9576637572395791, "grad_norm": 0.38117343187332153, "learning_rate": 4.7390756538666313e-05, "loss": 0.1015, "num_input_tokens_seen": 11301424, "step": 5870 }, { "epoch": 0.9584794844603964, "grad_norm": 0.2286301851272583, "learning_rate": 4.738638800520311e-05, "loss": 0.1538, "num_input_tokens_seen": 11310368, "step": 5875 }, { "epoch": 0.9592952116812138, "grad_norm": 0.6991433501243591, "learning_rate": 4.738201601947757e-05, "loss": 0.1241, "num_input_tokens_seen": 11320064, "step": 5880 }, { "epoch": 0.9601109389020311, "grad_norm": 5.350255966186523, "learning_rate": 4.7377640582163876e-05, "loss": 0.1884, "num_input_tokens_seen": 11329904, "step": 5885 }, { "epoch": 0.9609266661228485, "grad_norm": 0.9993571639060974, "learning_rate": 4.7373261693936786e-05, "loss": 0.1483, "num_input_tokens_seen": 11338848, "step": 5890 }, { "epoch": 0.9617423933436658, "grad_norm": 4.93531608581543, "learning_rate": 4.7368879355471595e-05, "loss": 0.0807, "num_input_tokens_seen": 11348112, "step": 5895 }, { "epoch": 0.9625581205644832, "grad_norm": 3.3866868019104004, "learning_rate": 4.736449356744409e-05, "loss": 0.1847, "num_input_tokens_seen": 11356880, "step": 5900 }, { "epoch": 0.9633738477853006, "grad_norm": 1.1954598426818848, "learning_rate": 4.736010433053064e-05, "loss": 0.1268, "num_input_tokens_seen": 11366512, "step": 5905 }, { "epoch": 0.9641895750061179, "grad_norm": 4.293285369873047, "learning_rate": 4.73557116454081e-05, "loss": 0.1507, "num_input_tokens_seen": 11375856, "step": 5910 }, { "epoch": 0.9650053022269353, "grad_norm": 0.45887061953544617, "learning_rate": 4.735131551275389e-05, "loss": 0.1391, "num_input_tokens_seen": 11386208, "step": 5915 }, { "epoch": 0.9658210294477527, "grad_norm": 0.15395455062389374, "learning_rate": 4.734691593324594e-05, "loss": 0.0699, "num_input_tokens_seen": 11396224, "step": 5920 }, { "epoch": 0.9666367566685701, "grad_norm": 4.018805503845215, "learning_rate": 4.734251290756272e-05, "loss": 0.1256, "num_input_tokens_seen": 11405904, "step": 5925 }, { "epoch": 0.9674524838893874, "grad_norm": 7.336617946624756, "learning_rate": 4.7338106436383246e-05, "loss": 0.1421, "num_input_tokens_seen": 11415120, "step": 5930 }, { "epoch": 0.9682682111102048, "grad_norm": 1.6519732475280762, "learning_rate": 4.733369652038703e-05, "loss": 0.0929, "num_input_tokens_seen": 11425648, "step": 5935 }, { "epoch": 0.9690839383310221, "grad_norm": 0.16236504912376404, "learning_rate": 4.7329283160254156e-05, "loss": 0.1358, "num_input_tokens_seen": 11436576, "step": 5940 }, { "epoch": 0.9698996655518395, "grad_norm": 0.4637192487716675, "learning_rate": 4.732486635666521e-05, "loss": 0.1473, "num_input_tokens_seen": 11446672, "step": 5945 }, { "epoch": 0.9707153927726568, "grad_norm": 5.539768218994141, "learning_rate": 4.732044611030132e-05, "loss": 0.2031, "num_input_tokens_seen": 11456480, "step": 5950 }, { "epoch": 0.9715311199934742, "grad_norm": 5.84586763381958, "learning_rate": 4.731602242184414e-05, "loss": 0.1701, "num_input_tokens_seen": 11465872, "step": 5955 }, { "epoch": 0.9723468472142915, "grad_norm": 5.760857105255127, "learning_rate": 4.7311595291975864e-05, "loss": 0.2442, "num_input_tokens_seen": 11476112, "step": 5960 }, { "epoch": 0.9731625744351089, "grad_norm": 0.5871009826660156, "learning_rate": 4.7307164721379216e-05, "loss": 0.0738, "num_input_tokens_seen": 11486816, "step": 5965 }, { "epoch": 0.9739783016559262, "grad_norm": 0.3269072473049164, "learning_rate": 4.730273071073743e-05, "loss": 0.0989, "num_input_tokens_seen": 11496400, "step": 5970 }, { "epoch": 0.9747940288767436, "grad_norm": 5.06233549118042, "learning_rate": 4.729829326073429e-05, "loss": 0.2929, "num_input_tokens_seen": 11505376, "step": 5975 }, { "epoch": 0.975609756097561, "grad_norm": 4.4989094734191895, "learning_rate": 4.7293852372054126e-05, "loss": 0.2074, "num_input_tokens_seen": 11515472, "step": 5980 }, { "epoch": 0.9764254833183783, "grad_norm": 3.4501335620880127, "learning_rate": 4.728940804538176e-05, "loss": 0.1321, "num_input_tokens_seen": 11524208, "step": 5985 }, { "epoch": 0.9772412105391957, "grad_norm": 0.48771893978118896, "learning_rate": 4.7284960281402556e-05, "loss": 0.1077, "num_input_tokens_seen": 11534176, "step": 5990 }, { "epoch": 0.978056937760013, "grad_norm": 3.455949068069458, "learning_rate": 4.728050908080244e-05, "loss": 0.2447, "num_input_tokens_seen": 11544592, "step": 5995 }, { "epoch": 0.9788726649808304, "grad_norm": 4.439403057098389, "learning_rate": 4.727605444426782e-05, "loss": 0.1544, "num_input_tokens_seen": 11553056, "step": 6000 }, { "epoch": 0.9788726649808304, "eval_loss": 0.1307803988456726, "eval_runtime": 80.6654, "eval_samples_per_second": 33.782, "eval_steps_per_second": 16.897, "num_input_tokens_seen": 11553056, "step": 6000 }, { "epoch": 0.9796883922016477, "grad_norm": 3.7535886764526367, "learning_rate": 4.727159637248567e-05, "loss": 0.1424, "num_input_tokens_seen": 11562896, "step": 6005 }, { "epoch": 0.9805041194224652, "grad_norm": 1.8671404123306274, "learning_rate": 4.7267134866143474e-05, "loss": 0.2438, "num_input_tokens_seen": 11572016, "step": 6010 }, { "epoch": 0.9813198466432825, "grad_norm": 0.4553922712802887, "learning_rate": 4.726266992592926e-05, "loss": 0.0601, "num_input_tokens_seen": 11580560, "step": 6015 }, { "epoch": 0.9821355738640999, "grad_norm": 1.0979448556900024, "learning_rate": 4.725820155253157e-05, "loss": 0.1571, "num_input_tokens_seen": 11590672, "step": 6020 }, { "epoch": 0.9829513010849172, "grad_norm": 1.6835674047470093, "learning_rate": 4.725372974663948e-05, "loss": 0.0914, "num_input_tokens_seen": 11600848, "step": 6025 }, { "epoch": 0.9837670283057346, "grad_norm": 2.824171543121338, "learning_rate": 4.724925450894262e-05, "loss": 0.1391, "num_input_tokens_seen": 11611232, "step": 6030 }, { "epoch": 0.984582755526552, "grad_norm": 0.6735730171203613, "learning_rate": 4.72447758401311e-05, "loss": 0.1895, "num_input_tokens_seen": 11620832, "step": 6035 }, { "epoch": 0.9853984827473693, "grad_norm": 2.489548921585083, "learning_rate": 4.7240293740895616e-05, "loss": 0.178, "num_input_tokens_seen": 11630464, "step": 6040 }, { "epoch": 0.9862142099681867, "grad_norm": 0.1599188596010208, "learning_rate": 4.723580821192733e-05, "loss": 0.0501, "num_input_tokens_seen": 11639632, "step": 6045 }, { "epoch": 0.987029937189004, "grad_norm": 6.345241546630859, "learning_rate": 4.7231319253917996e-05, "loss": 0.1879, "num_input_tokens_seen": 11648144, "step": 6050 }, { "epoch": 0.9878456644098214, "grad_norm": 1.1078344583511353, "learning_rate": 4.722682686755986e-05, "loss": 0.0763, "num_input_tokens_seen": 11658688, "step": 6055 }, { "epoch": 0.9886613916306387, "grad_norm": 3.304954767227173, "learning_rate": 4.722233105354569e-05, "loss": 0.2592, "num_input_tokens_seen": 11668480, "step": 6060 }, { "epoch": 0.9894771188514561, "grad_norm": 2.136212110519409, "learning_rate": 4.7217831812568815e-05, "loss": 0.1873, "num_input_tokens_seen": 11678576, "step": 6065 }, { "epoch": 0.9902928460722734, "grad_norm": 0.4164341986179352, "learning_rate": 4.721332914532307e-05, "loss": 0.0882, "num_input_tokens_seen": 11688480, "step": 6070 }, { "epoch": 0.9911085732930908, "grad_norm": 0.27655455470085144, "learning_rate": 4.720882305250281e-05, "loss": 0.2036, "num_input_tokens_seen": 11698128, "step": 6075 }, { "epoch": 0.9919243005139081, "grad_norm": 1.2263612747192383, "learning_rate": 4.720431353480295e-05, "loss": 0.1796, "num_input_tokens_seen": 11708000, "step": 6080 }, { "epoch": 0.9927400277347255, "grad_norm": 2.2550110816955566, "learning_rate": 4.719980059291891e-05, "loss": 0.0977, "num_input_tokens_seen": 11716352, "step": 6085 }, { "epoch": 0.9935557549555428, "grad_norm": 1.1590876579284668, "learning_rate": 4.7195284227546634e-05, "loss": 0.1113, "num_input_tokens_seen": 11725632, "step": 6090 }, { "epoch": 0.9943714821763602, "grad_norm": 5.560396194458008, "learning_rate": 4.7190764439382604e-05, "loss": 0.1146, "num_input_tokens_seen": 11736448, "step": 6095 }, { "epoch": 0.9951872093971775, "grad_norm": 0.11833511292934418, "learning_rate": 4.7186241229123826e-05, "loss": 0.1016, "num_input_tokens_seen": 11745712, "step": 6100 }, { "epoch": 0.996002936617995, "grad_norm": 0.08502493798732758, "learning_rate": 4.718171459746785e-05, "loss": 0.1288, "num_input_tokens_seen": 11754512, "step": 6105 }, { "epoch": 0.9968186638388123, "grad_norm": 0.31896206736564636, "learning_rate": 4.717718454511273e-05, "loss": 0.0559, "num_input_tokens_seen": 11763808, "step": 6110 }, { "epoch": 0.9976343910596297, "grad_norm": 0.49657514691352844, "learning_rate": 4.7172651072757056e-05, "loss": 0.1291, "num_input_tokens_seen": 11773632, "step": 6115 }, { "epoch": 0.998450118280447, "grad_norm": 2.1673614978790283, "learning_rate": 4.7168114181099945e-05, "loss": 0.0955, "num_input_tokens_seen": 11783840, "step": 6120 }, { "epoch": 0.9992658455012644, "grad_norm": 5.683025360107422, "learning_rate": 4.716357387084105e-05, "loss": 0.2449, "num_input_tokens_seen": 11794208, "step": 6125 }, { "epoch": 1.0, "grad_norm": 3.105323076248169, "learning_rate": 4.715903014268054e-05, "loss": 0.0648, "num_input_tokens_seen": 11803328, "step": 6130 }, { "epoch": 1.0008157272208174, "grad_norm": 2.432279109954834, "learning_rate": 4.715448299731911e-05, "loss": 0.0785, "num_input_tokens_seen": 11813488, "step": 6135 }, { "epoch": 1.0016314544416347, "grad_norm": 1.5147275924682617, "learning_rate": 4.7149932435457986e-05, "loss": 0.0709, "num_input_tokens_seen": 11823680, "step": 6140 }, { "epoch": 1.002447181662452, "grad_norm": 3.2791101932525635, "learning_rate": 4.714537845779894e-05, "loss": 0.2011, "num_input_tokens_seen": 11833184, "step": 6145 }, { "epoch": 1.0032629088832694, "grad_norm": 0.6574708819389343, "learning_rate": 4.714082106504423e-05, "loss": 0.0872, "num_input_tokens_seen": 11843792, "step": 6150 }, { "epoch": 1.0040786361040868, "grad_norm": 2.6987409591674805, "learning_rate": 4.713626025789667e-05, "loss": 0.1119, "num_input_tokens_seen": 11853344, "step": 6155 }, { "epoch": 1.0048943633249041, "grad_norm": 0.17500954866409302, "learning_rate": 4.7131696037059606e-05, "loss": 0.1906, "num_input_tokens_seen": 11862416, "step": 6160 }, { "epoch": 1.0057100905457215, "grad_norm": 1.1480648517608643, "learning_rate": 4.712712840323689e-05, "loss": 0.1038, "num_input_tokens_seen": 11872544, "step": 6165 }, { "epoch": 1.0065258177665388, "grad_norm": 1.874083399772644, "learning_rate": 4.71225573571329e-05, "loss": 0.1875, "num_input_tokens_seen": 11882896, "step": 6170 }, { "epoch": 1.0073415449873562, "grad_norm": 0.248834490776062, "learning_rate": 4.711798289945256e-05, "loss": 0.1409, "num_input_tokens_seen": 11892512, "step": 6175 }, { "epoch": 1.0081572722081735, "grad_norm": 2.1607401371002197, "learning_rate": 4.71134050309013e-05, "loss": 0.1118, "num_input_tokens_seen": 11902592, "step": 6180 }, { "epoch": 1.0089729994289909, "grad_norm": 4.3786234855651855, "learning_rate": 4.710882375218509e-05, "loss": 0.1241, "num_input_tokens_seen": 11912752, "step": 6185 }, { "epoch": 1.0097887266498082, "grad_norm": 0.4070148169994354, "learning_rate": 4.7104239064010424e-05, "loss": 0.1164, "num_input_tokens_seen": 11922368, "step": 6190 }, { "epoch": 1.0106044538706256, "grad_norm": 0.3100547194480896, "learning_rate": 4.709965096708432e-05, "loss": 0.0859, "num_input_tokens_seen": 11930176, "step": 6195 }, { "epoch": 1.011420181091443, "grad_norm": 0.3534668982028961, "learning_rate": 4.709505946211431e-05, "loss": 0.0942, "num_input_tokens_seen": 11940352, "step": 6200 }, { "epoch": 1.011420181091443, "eval_loss": 0.14942443370819092, "eval_runtime": 80.558, "eval_samples_per_second": 33.827, "eval_steps_per_second": 16.919, "num_input_tokens_seen": 11940352, "step": 6200 }, { "epoch": 1.0122359083122603, "grad_norm": 1.079227328300476, "learning_rate": 4.709046454980846e-05, "loss": 0.0941, "num_input_tokens_seen": 11951168, "step": 6205 }, { "epoch": 1.0130516355330776, "grad_norm": 6.435112476348877, "learning_rate": 4.708586623087538e-05, "loss": 0.1031, "num_input_tokens_seen": 11960720, "step": 6210 }, { "epoch": 1.013867362753895, "grad_norm": 0.21830955147743225, "learning_rate": 4.708126450602418e-05, "loss": 0.1016, "num_input_tokens_seen": 11971488, "step": 6215 }, { "epoch": 1.0146830899747123, "grad_norm": 10.455132484436035, "learning_rate": 4.7076659375964495e-05, "loss": 0.0317, "num_input_tokens_seen": 11981648, "step": 6220 }, { "epoch": 1.01549881719553, "grad_norm": 2.627993583679199, "learning_rate": 4.707205084140651e-05, "loss": 0.0911, "num_input_tokens_seen": 11989584, "step": 6225 }, { "epoch": 1.0163145444163473, "grad_norm": 2.9207210540771484, "learning_rate": 4.7067438903060904e-05, "loss": 0.108, "num_input_tokens_seen": 11999600, "step": 6230 }, { "epoch": 1.0171302716371646, "grad_norm": 5.3224616050720215, "learning_rate": 4.70628235616389e-05, "loss": 0.1346, "num_input_tokens_seen": 12008752, "step": 6235 }, { "epoch": 1.017945998857982, "grad_norm": 10.404356956481934, "learning_rate": 4.7058204817852256e-05, "loss": 0.0991, "num_input_tokens_seen": 12018128, "step": 6240 }, { "epoch": 1.0187617260787993, "grad_norm": 1.2825971841812134, "learning_rate": 4.705358267241322e-05, "loss": 0.1736, "num_input_tokens_seen": 12028368, "step": 6245 }, { "epoch": 1.0195774532996167, "grad_norm": 2.927537679672241, "learning_rate": 4.704895712603459e-05, "loss": 0.1495, "num_input_tokens_seen": 12038336, "step": 6250 }, { "epoch": 1.020393180520434, "grad_norm": 7.289804458618164, "learning_rate": 4.704432817942969e-05, "loss": 0.1679, "num_input_tokens_seen": 12047808, "step": 6255 }, { "epoch": 1.0212089077412514, "grad_norm": 3.9436042308807373, "learning_rate": 4.703969583331236e-05, "loss": 0.322, "num_input_tokens_seen": 12057680, "step": 6260 }, { "epoch": 1.0220246349620687, "grad_norm": 1.1020015478134155, "learning_rate": 4.7035060088396965e-05, "loss": 0.0689, "num_input_tokens_seen": 12067088, "step": 6265 }, { "epoch": 1.022840362182886, "grad_norm": 4.092426300048828, "learning_rate": 4.703042094539839e-05, "loss": 0.1128, "num_input_tokens_seen": 12075392, "step": 6270 }, { "epoch": 1.0236560894037035, "grad_norm": 1.1230463981628418, "learning_rate": 4.702577840503206e-05, "loss": 0.1725, "num_input_tokens_seen": 12084752, "step": 6275 }, { "epoch": 1.0244718166245208, "grad_norm": 0.22472038865089417, "learning_rate": 4.70211324680139e-05, "loss": 0.1351, "num_input_tokens_seen": 12094448, "step": 6280 }, { "epoch": 1.0252875438453382, "grad_norm": 2.4995243549346924, "learning_rate": 4.7016483135060386e-05, "loss": 0.0954, "num_input_tokens_seen": 12105504, "step": 6285 }, { "epoch": 1.0261032710661555, "grad_norm": 2.2113635540008545, "learning_rate": 4.701183040688849e-05, "loss": 0.0846, "num_input_tokens_seen": 12114992, "step": 6290 }, { "epoch": 1.0269189982869729, "grad_norm": 1.9180456399917603, "learning_rate": 4.700717428421573e-05, "loss": 0.0323, "num_input_tokens_seen": 12125424, "step": 6295 }, { "epoch": 1.0277347255077902, "grad_norm": 0.5761340856552124, "learning_rate": 4.700251476776014e-05, "loss": 0.194, "num_input_tokens_seen": 12134704, "step": 6300 }, { "epoch": 1.0285504527286076, "grad_norm": 11.763626098632812, "learning_rate": 4.699785185824026e-05, "loss": 0.1032, "num_input_tokens_seen": 12144320, "step": 6305 }, { "epoch": 1.029366179949425, "grad_norm": 7.28200626373291, "learning_rate": 4.699318555637519e-05, "loss": 0.0818, "num_input_tokens_seen": 12153680, "step": 6310 }, { "epoch": 1.0301819071702423, "grad_norm": 0.42700058221817017, "learning_rate": 4.6988515862884525e-05, "loss": 0.1318, "num_input_tokens_seen": 12164656, "step": 6315 }, { "epoch": 1.0309976343910596, "grad_norm": 1.439990520477295, "learning_rate": 4.698384277848838e-05, "loss": 0.1842, "num_input_tokens_seen": 12174416, "step": 6320 }, { "epoch": 1.031813361611877, "grad_norm": 0.12069860100746155, "learning_rate": 4.6979166303907425e-05, "loss": 0.2469, "num_input_tokens_seen": 12184384, "step": 6325 }, { "epoch": 1.0326290888326943, "grad_norm": 7.40447473526001, "learning_rate": 4.697448643986281e-05, "loss": 0.0965, "num_input_tokens_seen": 12194688, "step": 6330 }, { "epoch": 1.0334448160535117, "grad_norm": 0.557852029800415, "learning_rate": 4.696980318707624e-05, "loss": 0.1518, "num_input_tokens_seen": 12204032, "step": 6335 }, { "epoch": 1.034260543274329, "grad_norm": 1.7957943677902222, "learning_rate": 4.6965116546269924e-05, "loss": 0.1178, "num_input_tokens_seen": 12213728, "step": 6340 }, { "epoch": 1.0350762704951464, "grad_norm": 4.007907867431641, "learning_rate": 4.6960426518166615e-05, "loss": 0.1189, "num_input_tokens_seen": 12223888, "step": 6345 }, { "epoch": 1.0358919977159637, "grad_norm": 5.036341667175293, "learning_rate": 4.6955733103489556e-05, "loss": 0.0513, "num_input_tokens_seen": 12234544, "step": 6350 }, { "epoch": 1.036707724936781, "grad_norm": 0.19433759152889252, "learning_rate": 4.695103630296255e-05, "loss": 0.1402, "num_input_tokens_seen": 12244800, "step": 6355 }, { "epoch": 1.0375234521575984, "grad_norm": 7.055074691772461, "learning_rate": 4.694633611730988e-05, "loss": 0.164, "num_input_tokens_seen": 12254240, "step": 6360 }, { "epoch": 1.0383391793784158, "grad_norm": 2.105761766433716, "learning_rate": 4.694163254725639e-05, "loss": 0.1212, "num_input_tokens_seen": 12264800, "step": 6365 }, { "epoch": 1.0391549065992332, "grad_norm": 5.389227390289307, "learning_rate": 4.693692559352743e-05, "loss": 0.1735, "num_input_tokens_seen": 12274816, "step": 6370 }, { "epoch": 1.0399706338200505, "grad_norm": 3.2645490169525146, "learning_rate": 4.693221525684886e-05, "loss": 0.0292, "num_input_tokens_seen": 12283904, "step": 6375 }, { "epoch": 1.0407863610408679, "grad_norm": 1.4031956195831299, "learning_rate": 4.6927501537947084e-05, "loss": 0.1231, "num_input_tokens_seen": 12293088, "step": 6380 }, { "epoch": 1.0416020882616852, "grad_norm": 0.29801538586616516, "learning_rate": 4.692278443754901e-05, "loss": 0.1449, "num_input_tokens_seen": 12302448, "step": 6385 }, { "epoch": 1.0424178154825026, "grad_norm": 3.5880112648010254, "learning_rate": 4.691806395638208e-05, "loss": 0.1482, "num_input_tokens_seen": 12311984, "step": 6390 }, { "epoch": 1.04323354270332, "grad_norm": 2.959843158721924, "learning_rate": 4.6913340095174255e-05, "loss": 0.114, "num_input_tokens_seen": 12323232, "step": 6395 }, { "epoch": 1.0440492699241373, "grad_norm": 3.3240721225738525, "learning_rate": 4.690861285465399e-05, "loss": 0.1933, "num_input_tokens_seen": 12331920, "step": 6400 }, { "epoch": 1.0440492699241373, "eval_loss": 0.14100046455860138, "eval_runtime": 80.6256, "eval_samples_per_second": 33.798, "eval_steps_per_second": 16.905, "num_input_tokens_seen": 12331920, "step": 6400 }, { "epoch": 1.0448649971449546, "grad_norm": 5.61729621887207, "learning_rate": 4.690388223555031e-05, "loss": 0.1121, "num_input_tokens_seen": 12341824, "step": 6405 }, { "epoch": 1.0456807243657722, "grad_norm": 3.153292417526245, "learning_rate": 4.689914823859273e-05, "loss": 0.0415, "num_input_tokens_seen": 12352224, "step": 6410 }, { "epoch": 1.0464964515865895, "grad_norm": 0.22295893728733063, "learning_rate": 4.689441086451129e-05, "loss": 0.0252, "num_input_tokens_seen": 12360544, "step": 6415 }, { "epoch": 1.047312178807407, "grad_norm": 0.5936170816421509, "learning_rate": 4.688967011403655e-05, "loss": 0.1551, "num_input_tokens_seen": 12370592, "step": 6420 }, { "epoch": 1.0481279060282243, "grad_norm": 1.3052730560302734, "learning_rate": 4.68849259878996e-05, "loss": 0.0399, "num_input_tokens_seen": 12380912, "step": 6425 }, { "epoch": 1.0489436332490416, "grad_norm": 7.119204044342041, "learning_rate": 4.6880178486832036e-05, "loss": 0.1527, "num_input_tokens_seen": 12390432, "step": 6430 }, { "epoch": 1.049759360469859, "grad_norm": 12.710539817810059, "learning_rate": 4.687542761156598e-05, "loss": 0.1558, "num_input_tokens_seen": 12400240, "step": 6435 }, { "epoch": 1.0505750876906763, "grad_norm": 0.499844491481781, "learning_rate": 4.6870673362834096e-05, "loss": 0.0321, "num_input_tokens_seen": 12409024, "step": 6440 }, { "epoch": 1.0513908149114937, "grad_norm": 2.986776351928711, "learning_rate": 4.6865915741369526e-05, "loss": 0.0978, "num_input_tokens_seen": 12419792, "step": 6445 }, { "epoch": 1.052206542132311, "grad_norm": 4.514587879180908, "learning_rate": 4.686115474790597e-05, "loss": 0.1755, "num_input_tokens_seen": 12429600, "step": 6450 }, { "epoch": 1.0530222693531284, "grad_norm": 3.26855731010437, "learning_rate": 4.685639038317762e-05, "loss": 0.1368, "num_input_tokens_seen": 12438064, "step": 6455 }, { "epoch": 1.0538379965739457, "grad_norm": 0.04682502895593643, "learning_rate": 4.685162264791921e-05, "loss": 0.0139, "num_input_tokens_seen": 12449232, "step": 6460 }, { "epoch": 1.054653723794763, "grad_norm": 5.880825042724609, "learning_rate": 4.684685154286599e-05, "loss": 0.1463, "num_input_tokens_seen": 12458448, "step": 6465 }, { "epoch": 1.0554694510155804, "grad_norm": 3.8950655460357666, "learning_rate": 4.684207706875371e-05, "loss": 0.0855, "num_input_tokens_seen": 12468144, "step": 6470 }, { "epoch": 1.0562851782363978, "grad_norm": 0.2789478600025177, "learning_rate": 4.683729922631866e-05, "loss": 0.0915, "num_input_tokens_seen": 12477936, "step": 6475 }, { "epoch": 1.0571009054572151, "grad_norm": 5.7919840812683105, "learning_rate": 4.683251801629765e-05, "loss": 0.1252, "num_input_tokens_seen": 12487888, "step": 6480 }, { "epoch": 1.0579166326780325, "grad_norm": 0.07130250334739685, "learning_rate": 4.6827733439428e-05, "loss": 0.1003, "num_input_tokens_seen": 12498176, "step": 6485 }, { "epoch": 1.0587323598988498, "grad_norm": 0.19239214062690735, "learning_rate": 4.682294549644754e-05, "loss": 0.1228, "num_input_tokens_seen": 12508560, "step": 6490 }, { "epoch": 1.0595480871196672, "grad_norm": 3.171140670776367, "learning_rate": 4.681815418809464e-05, "loss": 0.0883, "num_input_tokens_seen": 12517632, "step": 6495 }, { "epoch": 1.0603638143404845, "grad_norm": 3.271909713745117, "learning_rate": 4.681335951510819e-05, "loss": 0.0623, "num_input_tokens_seen": 12526112, "step": 6500 }, { "epoch": 1.061179541561302, "grad_norm": 3.097656011581421, "learning_rate": 4.6808561478227576e-05, "loss": 0.1835, "num_input_tokens_seen": 12536544, "step": 6505 }, { "epoch": 1.0619952687821193, "grad_norm": 2.187546968460083, "learning_rate": 4.680376007819271e-05, "loss": 0.0805, "num_input_tokens_seen": 12546784, "step": 6510 }, { "epoch": 1.0628109960029366, "grad_norm": 3.9908971786499023, "learning_rate": 4.679895531574405e-05, "loss": 0.1405, "num_input_tokens_seen": 12557072, "step": 6515 }, { "epoch": 1.063626723223754, "grad_norm": 4.448425769805908, "learning_rate": 4.679414719162253e-05, "loss": 0.2206, "num_input_tokens_seen": 12566176, "step": 6520 }, { "epoch": 1.0644424504445713, "grad_norm": 5.317002296447754, "learning_rate": 4.6789335706569635e-05, "loss": 0.0839, "num_input_tokens_seen": 12575632, "step": 6525 }, { "epoch": 1.0652581776653887, "grad_norm": 3.3642897605895996, "learning_rate": 4.678452086132734e-05, "loss": 0.0983, "num_input_tokens_seen": 12586000, "step": 6530 }, { "epoch": 1.066073904886206, "grad_norm": 4.7375569343566895, "learning_rate": 4.677970265663818e-05, "loss": 0.1503, "num_input_tokens_seen": 12596704, "step": 6535 }, { "epoch": 1.0668896321070234, "grad_norm": 0.898496687412262, "learning_rate": 4.677488109324517e-05, "loss": 0.0941, "num_input_tokens_seen": 12605264, "step": 6540 }, { "epoch": 1.0677053593278407, "grad_norm": 5.1676154136657715, "learning_rate": 4.6770056171891846e-05, "loss": 0.1031, "num_input_tokens_seen": 12616192, "step": 6545 }, { "epoch": 1.068521086548658, "grad_norm": 2.5144147872924805, "learning_rate": 4.6765227893322286e-05, "loss": 0.045, "num_input_tokens_seen": 12626096, "step": 6550 }, { "epoch": 1.0693368137694754, "grad_norm": 3.1739723682403564, "learning_rate": 4.676039625828107e-05, "loss": 0.1586, "num_input_tokens_seen": 12636208, "step": 6555 }, { "epoch": 1.0701525409902928, "grad_norm": 2.8332507610321045, "learning_rate": 4.675556126751328e-05, "loss": 0.1111, "num_input_tokens_seen": 12646144, "step": 6560 }, { "epoch": 1.0709682682111101, "grad_norm": 0.07937043905258179, "learning_rate": 4.6750722921764556e-05, "loss": 0.0256, "num_input_tokens_seen": 12656880, "step": 6565 }, { "epoch": 1.0717839954319275, "grad_norm": 0.11062305420637131, "learning_rate": 4.674588122178102e-05, "loss": 0.0763, "num_input_tokens_seen": 12666640, "step": 6570 }, { "epoch": 1.0725997226527448, "grad_norm": 3.3560116291046143, "learning_rate": 4.674103616830931e-05, "loss": 0.0691, "num_input_tokens_seen": 12677264, "step": 6575 }, { "epoch": 1.0734154498735622, "grad_norm": 3.1575372219085693, "learning_rate": 4.673618776209663e-05, "loss": 0.2272, "num_input_tokens_seen": 12687808, "step": 6580 }, { "epoch": 1.0742311770943795, "grad_norm": 0.2855183184146881, "learning_rate": 4.673133600389063e-05, "loss": 0.0615, "num_input_tokens_seen": 12697024, "step": 6585 }, { "epoch": 1.075046904315197, "grad_norm": 0.0894378125667572, "learning_rate": 4.672648089443953e-05, "loss": 0.0827, "num_input_tokens_seen": 12706304, "step": 6590 }, { "epoch": 1.0758626315360145, "grad_norm": 0.7645026445388794, "learning_rate": 4.672162243449204e-05, "loss": 0.2231, "num_input_tokens_seen": 12716768, "step": 6595 }, { "epoch": 1.0766783587568316, "grad_norm": 0.8109590411186218, "learning_rate": 4.67167606247974e-05, "loss": 0.0982, "num_input_tokens_seen": 12726352, "step": 6600 }, { "epoch": 1.0766783587568316, "eval_loss": 0.14559750258922577, "eval_runtime": 80.557, "eval_samples_per_second": 33.827, "eval_steps_per_second": 16.92, "num_input_tokens_seen": 12726352, "step": 6600 }, { "epoch": 1.0774940859776492, "grad_norm": 0.5200181603431702, "learning_rate": 4.671189546610536e-05, "loss": 0.0293, "num_input_tokens_seen": 12736480, "step": 6605 }, { "epoch": 1.0783098131984665, "grad_norm": 0.7032972574234009, "learning_rate": 4.67070269591662e-05, "loss": 0.0638, "num_input_tokens_seen": 12745984, "step": 6610 }, { "epoch": 1.0791255404192839, "grad_norm": 3.9539859294891357, "learning_rate": 4.670215510473068e-05, "loss": 0.0554, "num_input_tokens_seen": 12755920, "step": 6615 }, { "epoch": 1.0799412676401012, "grad_norm": 5.935372352600098, "learning_rate": 4.669727990355013e-05, "loss": 0.0915, "num_input_tokens_seen": 12765376, "step": 6620 }, { "epoch": 1.0807569948609186, "grad_norm": 6.215117454528809, "learning_rate": 4.669240135637635e-05, "loss": 0.1013, "num_input_tokens_seen": 12773776, "step": 6625 }, { "epoch": 1.081572722081736, "grad_norm": 5.286499977111816, "learning_rate": 4.6687519463961675e-05, "loss": 0.1489, "num_input_tokens_seen": 12784288, "step": 6630 }, { "epoch": 1.0823884493025533, "grad_norm": 2.1881895065307617, "learning_rate": 4.668263422705896e-05, "loss": 0.0835, "num_input_tokens_seen": 12793264, "step": 6635 }, { "epoch": 1.0832041765233706, "grad_norm": 5.180602073669434, "learning_rate": 4.667774564642156e-05, "loss": 0.1338, "num_input_tokens_seen": 12801728, "step": 6640 }, { "epoch": 1.084019903744188, "grad_norm": 0.12274446338415146, "learning_rate": 4.6672853722803365e-05, "loss": 0.1104, "num_input_tokens_seen": 12811952, "step": 6645 }, { "epoch": 1.0848356309650053, "grad_norm": 0.23717054724693298, "learning_rate": 4.666795845695877e-05, "loss": 0.0771, "num_input_tokens_seen": 12821632, "step": 6650 }, { "epoch": 1.0856513581858227, "grad_norm": 1.8630797863006592, "learning_rate": 4.666305984964269e-05, "loss": 0.202, "num_input_tokens_seen": 12830832, "step": 6655 }, { "epoch": 1.08646708540664, "grad_norm": 5.630118370056152, "learning_rate": 4.6658157901610535e-05, "loss": 0.3061, "num_input_tokens_seen": 12841520, "step": 6660 }, { "epoch": 1.0872828126274574, "grad_norm": 9.373684883117676, "learning_rate": 4.665325261361826e-05, "loss": 0.1183, "num_input_tokens_seen": 12852560, "step": 6665 }, { "epoch": 1.0880985398482748, "grad_norm": 2.9588541984558105, "learning_rate": 4.664834398642232e-05, "loss": 0.1435, "num_input_tokens_seen": 12861952, "step": 6670 }, { "epoch": 1.0889142670690921, "grad_norm": 0.5051087737083435, "learning_rate": 4.6643432020779686e-05, "loss": 0.1147, "num_input_tokens_seen": 12871600, "step": 6675 }, { "epoch": 1.0897299942899095, "grad_norm": 0.6051807403564453, "learning_rate": 4.663851671744786e-05, "loss": 0.1058, "num_input_tokens_seen": 12881520, "step": 6680 }, { "epoch": 1.0905457215107268, "grad_norm": 1.0254082679748535, "learning_rate": 4.6633598077184815e-05, "loss": 0.0975, "num_input_tokens_seen": 12890368, "step": 6685 }, { "epoch": 1.0913614487315442, "grad_norm": 2.262422561645508, "learning_rate": 4.662867610074908e-05, "loss": 0.1462, "num_input_tokens_seen": 12900384, "step": 6690 }, { "epoch": 1.0921771759523615, "grad_norm": 0.711699903011322, "learning_rate": 4.6623750788899696e-05, "loss": 0.0568, "num_input_tokens_seen": 12909056, "step": 6695 }, { "epoch": 1.0929929031731789, "grad_norm": 5.2430925369262695, "learning_rate": 4.6618822142396195e-05, "loss": 0.151, "num_input_tokens_seen": 12919088, "step": 6700 }, { "epoch": 1.0938086303939962, "grad_norm": 2.798293113708496, "learning_rate": 4.661389016199864e-05, "loss": 0.1348, "num_input_tokens_seen": 12929504, "step": 6705 }, { "epoch": 1.0946243576148136, "grad_norm": 3.313573122024536, "learning_rate": 4.660895484846761e-05, "loss": 0.0942, "num_input_tokens_seen": 12939520, "step": 6710 }, { "epoch": 1.095440084835631, "grad_norm": 7.987139701843262, "learning_rate": 4.660401620256418e-05, "loss": 0.0526, "num_input_tokens_seen": 12948832, "step": 6715 }, { "epoch": 1.0962558120564483, "grad_norm": 1.2409541606903076, "learning_rate": 4.659907422504997e-05, "loss": 0.0351, "num_input_tokens_seen": 12958176, "step": 6720 }, { "epoch": 1.0970715392772656, "grad_norm": 3.3917908668518066, "learning_rate": 4.6594128916687074e-05, "loss": 0.1264, "num_input_tokens_seen": 12967168, "step": 6725 }, { "epoch": 1.097887266498083, "grad_norm": 2.8892648220062256, "learning_rate": 4.658918027823813e-05, "loss": 0.0867, "num_input_tokens_seen": 12976096, "step": 6730 }, { "epoch": 1.0987029937189003, "grad_norm": 0.04671201482415199, "learning_rate": 4.658422831046628e-05, "loss": 0.0217, "num_input_tokens_seen": 12984880, "step": 6735 }, { "epoch": 1.0995187209397177, "grad_norm": 0.0538431853055954, "learning_rate": 4.657927301413518e-05, "loss": 0.0783, "num_input_tokens_seen": 12995232, "step": 6740 }, { "epoch": 1.100334448160535, "grad_norm": 0.14195217192173004, "learning_rate": 4.657431439000901e-05, "loss": 0.107, "num_input_tokens_seen": 13002576, "step": 6745 }, { "epoch": 1.1011501753813524, "grad_norm": 7.4815239906311035, "learning_rate": 4.656935243885243e-05, "loss": 0.1428, "num_input_tokens_seen": 13011584, "step": 6750 }, { "epoch": 1.1019659026021698, "grad_norm": 1.131117820739746, "learning_rate": 4.656438716143066e-05, "loss": 0.0194, "num_input_tokens_seen": 13022288, "step": 6755 }, { "epoch": 1.102781629822987, "grad_norm": 7.3729777336120605, "learning_rate": 4.6559418558509384e-05, "loss": 0.119, "num_input_tokens_seen": 13032000, "step": 6760 }, { "epoch": 1.1035973570438045, "grad_norm": 3.9622273445129395, "learning_rate": 4.6554446630854833e-05, "loss": 0.1655, "num_input_tokens_seen": 13040048, "step": 6765 }, { "epoch": 1.1044130842646218, "grad_norm": 11.169100761413574, "learning_rate": 4.654947137923374e-05, "loss": 0.1255, "num_input_tokens_seen": 13049248, "step": 6770 }, { "epoch": 1.1052288114854392, "grad_norm": 6.0267815589904785, "learning_rate": 4.654449280441335e-05, "loss": 0.1434, "num_input_tokens_seen": 13057616, "step": 6775 }, { "epoch": 1.1060445387062567, "grad_norm": 1.0917819738388062, "learning_rate": 4.653951090716143e-05, "loss": 0.0835, "num_input_tokens_seen": 13066320, "step": 6780 }, { "epoch": 1.1068602659270739, "grad_norm": 4.001070499420166, "learning_rate": 4.653452568824625e-05, "loss": 0.2111, "num_input_tokens_seen": 13075696, "step": 6785 }, { "epoch": 1.1076759931478914, "grad_norm": 0.24964171648025513, "learning_rate": 4.6529537148436585e-05, "loss": 0.0807, "num_input_tokens_seen": 13085872, "step": 6790 }, { "epoch": 1.1084917203687088, "grad_norm": 6.971999645233154, "learning_rate": 4.6524545288501734e-05, "loss": 0.1779, "num_input_tokens_seen": 13095248, "step": 6795 }, { "epoch": 1.1093074475895262, "grad_norm": 1.4290037155151367, "learning_rate": 4.6519550109211506e-05, "loss": 0.2651, "num_input_tokens_seen": 13105200, "step": 6800 }, { "epoch": 1.1093074475895262, "eval_loss": 0.1489081233739853, "eval_runtime": 80.5055, "eval_samples_per_second": 33.849, "eval_steps_per_second": 16.931, "num_input_tokens_seen": 13105200, "step": 6800 }, { "epoch": 1.1101231748103435, "grad_norm": 1.9280354976654053, "learning_rate": 4.651455161133622e-05, "loss": 0.1436, "num_input_tokens_seen": 13113760, "step": 6805 }, { "epoch": 1.1109389020311609, "grad_norm": 0.2798200249671936, "learning_rate": 4.6509549795646704e-05, "loss": 0.0835, "num_input_tokens_seen": 13124768, "step": 6810 }, { "epoch": 1.1117546292519782, "grad_norm": 2.3192732334136963, "learning_rate": 4.6504544662914306e-05, "loss": 0.1966, "num_input_tokens_seen": 13134608, "step": 6815 }, { "epoch": 1.1125703564727956, "grad_norm": 2.8248159885406494, "learning_rate": 4.6499536213910876e-05, "loss": 0.0714, "num_input_tokens_seen": 13144000, "step": 6820 }, { "epoch": 1.113386083693613, "grad_norm": 4.8287353515625, "learning_rate": 4.6494524449408786e-05, "loss": 0.1044, "num_input_tokens_seen": 13152928, "step": 6825 }, { "epoch": 1.1142018109144303, "grad_norm": 0.1656479388475418, "learning_rate": 4.6489509370180903e-05, "loss": 0.0539, "num_input_tokens_seen": 13162288, "step": 6830 }, { "epoch": 1.1150175381352476, "grad_norm": 2.4762351512908936, "learning_rate": 4.648449097700063e-05, "loss": 0.0782, "num_input_tokens_seen": 13170192, "step": 6835 }, { "epoch": 1.115833265356065, "grad_norm": 0.8100693225860596, "learning_rate": 4.647946927064185e-05, "loss": 0.042, "num_input_tokens_seen": 13179584, "step": 6840 }, { "epoch": 1.1166489925768823, "grad_norm": 2.3920793533325195, "learning_rate": 4.647444425187898e-05, "loss": 0.0439, "num_input_tokens_seen": 13188624, "step": 6845 }, { "epoch": 1.1174647197976997, "grad_norm": 5.076511383056641, "learning_rate": 4.646941592148695e-05, "loss": 0.1835, "num_input_tokens_seen": 13198944, "step": 6850 }, { "epoch": 1.118280447018517, "grad_norm": 0.8614566922187805, "learning_rate": 4.646438428024117e-05, "loss": 0.1292, "num_input_tokens_seen": 13208176, "step": 6855 }, { "epoch": 1.1190961742393344, "grad_norm": 3.874589443206787, "learning_rate": 4.64593493289176e-05, "loss": 0.2033, "num_input_tokens_seen": 13218000, "step": 6860 }, { "epoch": 1.1199119014601517, "grad_norm": 0.2249961644411087, "learning_rate": 4.64543110682927e-05, "loss": 0.0712, "num_input_tokens_seen": 13227952, "step": 6865 }, { "epoch": 1.120727628680969, "grad_norm": 5.264800548553467, "learning_rate": 4.644926949914341e-05, "loss": 0.1364, "num_input_tokens_seen": 13238480, "step": 6870 }, { "epoch": 1.1215433559017864, "grad_norm": 5.794719696044922, "learning_rate": 4.644422462224722e-05, "loss": 0.1567, "num_input_tokens_seen": 13247488, "step": 6875 }, { "epoch": 1.1223590831226038, "grad_norm": 1.0872552394866943, "learning_rate": 4.643917643838211e-05, "loss": 0.0279, "num_input_tokens_seen": 13256000, "step": 6880 }, { "epoch": 1.1231748103434211, "grad_norm": 0.09562472254037857, "learning_rate": 4.6434124948326564e-05, "loss": 0.0921, "num_input_tokens_seen": 13265488, "step": 6885 }, { "epoch": 1.1239905375642385, "grad_norm": 6.509157657623291, "learning_rate": 4.6429070152859594e-05, "loss": 0.2564, "num_input_tokens_seen": 13274128, "step": 6890 }, { "epoch": 1.1248062647850559, "grad_norm": 2.2184109687805176, "learning_rate": 4.6424012052760714e-05, "loss": 0.1562, "num_input_tokens_seen": 13284384, "step": 6895 }, { "epoch": 1.1256219920058732, "grad_norm": 1.015695571899414, "learning_rate": 4.6418950648809945e-05, "loss": 0.091, "num_input_tokens_seen": 13294560, "step": 6900 }, { "epoch": 1.1264377192266906, "grad_norm": 5.068578243255615, "learning_rate": 4.641388594178782e-05, "loss": 0.1662, "num_input_tokens_seen": 13304656, "step": 6905 }, { "epoch": 1.127253446447508, "grad_norm": 3.3149309158325195, "learning_rate": 4.640881793247538e-05, "loss": 0.106, "num_input_tokens_seen": 13315296, "step": 6910 }, { "epoch": 1.1280691736683253, "grad_norm": 3.050391435623169, "learning_rate": 4.6403746621654173e-05, "loss": 0.0562, "num_input_tokens_seen": 13325152, "step": 6915 }, { "epoch": 1.1288849008891426, "grad_norm": 0.5820143818855286, "learning_rate": 4.639867201010626e-05, "loss": 0.2027, "num_input_tokens_seen": 13333728, "step": 6920 }, { "epoch": 1.12970062810996, "grad_norm": 3.4149270057678223, "learning_rate": 4.6393594098614204e-05, "loss": 0.1546, "num_input_tokens_seen": 13342880, "step": 6925 }, { "epoch": 1.1305163553307773, "grad_norm": 0.4158870577812195, "learning_rate": 4.63885128879611e-05, "loss": 0.1402, "num_input_tokens_seen": 13351616, "step": 6930 }, { "epoch": 1.1313320825515947, "grad_norm": 0.5381664633750916, "learning_rate": 4.638342837893052e-05, "loss": 0.2242, "num_input_tokens_seen": 13361872, "step": 6935 }, { "epoch": 1.132147809772412, "grad_norm": 4.160445690155029, "learning_rate": 4.6378340572306565e-05, "loss": 0.2235, "num_input_tokens_seen": 13370336, "step": 6940 }, { "epoch": 1.1329635369932294, "grad_norm": 1.1728819608688354, "learning_rate": 4.6373249468873833e-05, "loss": 0.0861, "num_input_tokens_seen": 13378720, "step": 6945 }, { "epoch": 1.1337792642140467, "grad_norm": 1.7499843835830688, "learning_rate": 4.636815506941744e-05, "loss": 0.1003, "num_input_tokens_seen": 13388720, "step": 6950 }, { "epoch": 1.134594991434864, "grad_norm": 2.484630823135376, "learning_rate": 4.6363057374723004e-05, "loss": 0.0975, "num_input_tokens_seen": 13398304, "step": 6955 }, { "epoch": 1.1354107186556814, "grad_norm": 5.819054126739502, "learning_rate": 4.635795638557666e-05, "loss": 0.0622, "num_input_tokens_seen": 13408368, "step": 6960 }, { "epoch": 1.136226445876499, "grad_norm": 1.465413212776184, "learning_rate": 4.635285210276504e-05, "loss": 0.1307, "num_input_tokens_seen": 13418512, "step": 6965 }, { "epoch": 1.1370421730973161, "grad_norm": 5.592293739318848, "learning_rate": 4.6347744527075295e-05, "loss": 0.0888, "num_input_tokens_seen": 13427296, "step": 6970 }, { "epoch": 1.1378579003181337, "grad_norm": 3.404665470123291, "learning_rate": 4.634263365929506e-05, "loss": 0.0698, "num_input_tokens_seen": 13435264, "step": 6975 }, { "epoch": 1.1386736275389508, "grad_norm": 6.279942512512207, "learning_rate": 4.6337519500212515e-05, "loss": 0.1668, "num_input_tokens_seen": 13444608, "step": 6980 }, { "epoch": 1.1394893547597684, "grad_norm": 0.18506431579589844, "learning_rate": 4.633240205061632e-05, "loss": 0.0603, "num_input_tokens_seen": 13454528, "step": 6985 }, { "epoch": 1.1403050819805858, "grad_norm": 5.586252689361572, "learning_rate": 4.632728131129565e-05, "loss": 0.092, "num_input_tokens_seen": 13464016, "step": 6990 }, { "epoch": 1.1411208092014031, "grad_norm": 8.9928617477417, "learning_rate": 4.632215728304018e-05, "loss": 0.1711, "num_input_tokens_seen": 13473488, "step": 6995 }, { "epoch": 1.1419365364222205, "grad_norm": 0.38201066851615906, "learning_rate": 4.63170299666401e-05, "loss": 0.106, "num_input_tokens_seen": 13483648, "step": 7000 }, { "epoch": 1.1419365364222205, "eval_loss": 0.14278674125671387, "eval_runtime": 80.5153, "eval_samples_per_second": 33.844, "eval_steps_per_second": 16.928, "num_input_tokens_seen": 13483648, "step": 7000 }, { "epoch": 1.1427522636430378, "grad_norm": 4.863647937774658, "learning_rate": 4.631189936288612e-05, "loss": 0.1229, "num_input_tokens_seen": 13492992, "step": 7005 }, { "epoch": 1.1435679908638552, "grad_norm": 0.17430612444877625, "learning_rate": 4.630676547256944e-05, "loss": 0.1081, "num_input_tokens_seen": 13502208, "step": 7010 }, { "epoch": 1.1443837180846725, "grad_norm": 3.265918016433716, "learning_rate": 4.630162829648176e-05, "loss": 0.0326, "num_input_tokens_seen": 13511600, "step": 7015 }, { "epoch": 1.14519944530549, "grad_norm": 2.7660720348358154, "learning_rate": 4.629648783541531e-05, "loss": 0.1669, "num_input_tokens_seen": 13521760, "step": 7020 }, { "epoch": 1.1460151725263072, "grad_norm": 2.559823751449585, "learning_rate": 4.6291344090162804e-05, "loss": 0.1906, "num_input_tokens_seen": 13530752, "step": 7025 }, { "epoch": 1.1468308997471246, "grad_norm": 10.027872085571289, "learning_rate": 4.628619706151748e-05, "loss": 0.163, "num_input_tokens_seen": 13540256, "step": 7030 }, { "epoch": 1.147646626967942, "grad_norm": 2.6856164932250977, "learning_rate": 4.628104675027306e-05, "loss": 0.0843, "num_input_tokens_seen": 13551168, "step": 7035 }, { "epoch": 1.1484623541887593, "grad_norm": 0.9430164694786072, "learning_rate": 4.6275893157223805e-05, "loss": 0.1483, "num_input_tokens_seen": 13560608, "step": 7040 }, { "epoch": 1.1492780814095767, "grad_norm": 2.9642155170440674, "learning_rate": 4.627073628316445e-05, "loss": 0.1621, "num_input_tokens_seen": 13569424, "step": 7045 }, { "epoch": 1.150093808630394, "grad_norm": 6.7635674476623535, "learning_rate": 4.626557612889026e-05, "loss": 0.0491, "num_input_tokens_seen": 13578272, "step": 7050 }, { "epoch": 1.1509095358512114, "grad_norm": 2.852968692779541, "learning_rate": 4.626041269519699e-05, "loss": 0.1151, "num_input_tokens_seen": 13588192, "step": 7055 }, { "epoch": 1.1517252630720287, "grad_norm": 3.020711898803711, "learning_rate": 4.6255245982880905e-05, "loss": 0.1487, "num_input_tokens_seen": 13596976, "step": 7060 }, { "epoch": 1.152540990292846, "grad_norm": 0.9684382081031799, "learning_rate": 4.625007599273879e-05, "loss": 0.0772, "num_input_tokens_seen": 13608016, "step": 7065 }, { "epoch": 1.1533567175136634, "grad_norm": 0.2033701390028, "learning_rate": 4.6244902725567895e-05, "loss": 0.1646, "num_input_tokens_seen": 13617040, "step": 7070 }, { "epoch": 1.1541724447344808, "grad_norm": 0.38435736298561096, "learning_rate": 4.6239726182166024e-05, "loss": 0.08, "num_input_tokens_seen": 13625856, "step": 7075 }, { "epoch": 1.1549881719552981, "grad_norm": 3.7058603763580322, "learning_rate": 4.623454636333147e-05, "loss": 0.1165, "num_input_tokens_seen": 13634928, "step": 7080 }, { "epoch": 1.1558038991761155, "grad_norm": 1.8812377452850342, "learning_rate": 4.622936326986301e-05, "loss": 0.1281, "num_input_tokens_seen": 13644736, "step": 7085 }, { "epoch": 1.1566196263969328, "grad_norm": 0.17432352900505066, "learning_rate": 4.6224176902559946e-05, "loss": 0.0678, "num_input_tokens_seen": 13655136, "step": 7090 }, { "epoch": 1.1574353536177502, "grad_norm": 1.5097625255584717, "learning_rate": 4.621898726222209e-05, "loss": 0.1737, "num_input_tokens_seen": 13664656, "step": 7095 }, { "epoch": 1.1582510808385675, "grad_norm": 0.46114033460617065, "learning_rate": 4.6213794349649744e-05, "loss": 0.1494, "num_input_tokens_seen": 13672880, "step": 7100 }, { "epoch": 1.159066808059385, "grad_norm": 0.22114896774291992, "learning_rate": 4.6208598165643715e-05, "loss": 0.0834, "num_input_tokens_seen": 13682240, "step": 7105 }, { "epoch": 1.1598825352802022, "grad_norm": 0.18082383275032043, "learning_rate": 4.620339871100533e-05, "loss": 0.0493, "num_input_tokens_seen": 13690720, "step": 7110 }, { "epoch": 1.1606982625010196, "grad_norm": 0.29677510261535645, "learning_rate": 4.6198195986536394e-05, "loss": 0.0234, "num_input_tokens_seen": 13699888, "step": 7115 }, { "epoch": 1.161513989721837, "grad_norm": 0.8579052686691284, "learning_rate": 4.619298999303926e-05, "loss": 0.0307, "num_input_tokens_seen": 13709360, "step": 7120 }, { "epoch": 1.1623297169426543, "grad_norm": 0.22020839154720306, "learning_rate": 4.618778073131673e-05, "loss": 0.1643, "num_input_tokens_seen": 13719040, "step": 7125 }, { "epoch": 1.1631454441634717, "grad_norm": 7.603555202484131, "learning_rate": 4.618256820217215e-05, "loss": 0.2368, "num_input_tokens_seen": 13729328, "step": 7130 }, { "epoch": 1.163961171384289, "grad_norm": 0.30058780312538147, "learning_rate": 4.617735240640936e-05, "loss": 0.1419, "num_input_tokens_seen": 13739792, "step": 7135 }, { "epoch": 1.1647768986051064, "grad_norm": 3.178114175796509, "learning_rate": 4.6172133344832705e-05, "loss": 0.1923, "num_input_tokens_seen": 13748912, "step": 7140 }, { "epoch": 1.1655926258259237, "grad_norm": 3.937889575958252, "learning_rate": 4.6166911018247004e-05, "loss": 0.1105, "num_input_tokens_seen": 13758656, "step": 7145 }, { "epoch": 1.1664083530467413, "grad_norm": 0.1439642757177353, "learning_rate": 4.616168542745764e-05, "loss": 0.1382, "num_input_tokens_seen": 13768944, "step": 7150 }, { "epoch": 1.1672240802675584, "grad_norm": 3.297572374343872, "learning_rate": 4.6156456573270446e-05, "loss": 0.1767, "num_input_tokens_seen": 13776880, "step": 7155 }, { "epoch": 1.168039807488376, "grad_norm": 2.02905535697937, "learning_rate": 4.615122445649177e-05, "loss": 0.0638, "num_input_tokens_seen": 13785488, "step": 7160 }, { "epoch": 1.1688555347091931, "grad_norm": 4.342958927154541, "learning_rate": 4.6145989077928486e-05, "loss": 0.0321, "num_input_tokens_seen": 13795136, "step": 7165 }, { "epoch": 1.1696712619300107, "grad_norm": 4.525722980499268, "learning_rate": 4.6140750438387953e-05, "loss": 0.114, "num_input_tokens_seen": 13804720, "step": 7170 }, { "epoch": 1.170486989150828, "grad_norm": 9.688820838928223, "learning_rate": 4.613550853867803e-05, "loss": 0.1322, "num_input_tokens_seen": 13814272, "step": 7175 }, { "epoch": 1.1713027163716454, "grad_norm": 1.4124161005020142, "learning_rate": 4.613026337960708e-05, "loss": 0.1005, "num_input_tokens_seen": 13824624, "step": 7180 }, { "epoch": 1.1721184435924628, "grad_norm": 0.573037326335907, "learning_rate": 4.612501496198398e-05, "loss": 0.0889, "num_input_tokens_seen": 13833328, "step": 7185 }, { "epoch": 1.17293417081328, "grad_norm": 3.3043930530548096, "learning_rate": 4.61197632866181e-05, "loss": 0.0859, "num_input_tokens_seen": 13843520, "step": 7190 }, { "epoch": 1.1737498980340975, "grad_norm": 0.8584202527999878, "learning_rate": 4.611450835431931e-05, "loss": 0.0763, "num_input_tokens_seen": 13852736, "step": 7195 }, { "epoch": 1.1745656252549148, "grad_norm": 0.15480048954486847, "learning_rate": 4.6109250165898e-05, "loss": 0.0809, "num_input_tokens_seen": 13862816, "step": 7200 }, { "epoch": 1.1745656252549148, "eval_loss": 0.14510487020015717, "eval_runtime": 80.59, "eval_samples_per_second": 33.813, "eval_steps_per_second": 16.913, "num_input_tokens_seen": 13862816, "step": 7200 }, { "epoch": 1.1753813524757322, "grad_norm": 0.38698479533195496, "learning_rate": 4.610398872216503e-05, "loss": 0.1054, "num_input_tokens_seen": 13872784, "step": 7205 }, { "epoch": 1.1761970796965495, "grad_norm": 0.7491023540496826, "learning_rate": 4.6098724023931796e-05, "loss": 0.2095, "num_input_tokens_seen": 13883040, "step": 7210 }, { "epoch": 1.1770128069173669, "grad_norm": 0.1377689689397812, "learning_rate": 4.609345607201017e-05, "loss": 0.0433, "num_input_tokens_seen": 13891632, "step": 7215 }, { "epoch": 1.1778285341381842, "grad_norm": 4.985355854034424, "learning_rate": 4.608818486721254e-05, "loss": 0.0949, "num_input_tokens_seen": 13900928, "step": 7220 }, { "epoch": 1.1786442613590016, "grad_norm": 3.6047027111053467, "learning_rate": 4.608291041035179e-05, "loss": 0.2016, "num_input_tokens_seen": 13910304, "step": 7225 }, { "epoch": 1.179459988579819, "grad_norm": 2.1672701835632324, "learning_rate": 4.607763270224132e-05, "loss": 0.0451, "num_input_tokens_seen": 13919824, "step": 7230 }, { "epoch": 1.1802757158006363, "grad_norm": 8.675230026245117, "learning_rate": 4.6072351743695e-05, "loss": 0.2784, "num_input_tokens_seen": 13929184, "step": 7235 }, { "epoch": 1.1810914430214536, "grad_norm": 0.4864087998867035, "learning_rate": 4.606706753552723e-05, "loss": 0.0823, "num_input_tokens_seen": 13939744, "step": 7240 }, { "epoch": 1.181907170242271, "grad_norm": 2.802443742752075, "learning_rate": 4.6061780078552906e-05, "loss": 0.1021, "num_input_tokens_seen": 13949632, "step": 7245 }, { "epoch": 1.1827228974630883, "grad_norm": 0.21235381066799164, "learning_rate": 4.605648937358742e-05, "loss": 0.0476, "num_input_tokens_seen": 13959872, "step": 7250 }, { "epoch": 1.1835386246839057, "grad_norm": 9.524452209472656, "learning_rate": 4.605119542144665e-05, "loss": 0.1647, "num_input_tokens_seen": 13969520, "step": 7255 }, { "epoch": 1.184354351904723, "grad_norm": 8.258328437805176, "learning_rate": 4.604589822294701e-05, "loss": 0.1017, "num_input_tokens_seen": 13977680, "step": 7260 }, { "epoch": 1.1851700791255404, "grad_norm": 7.56138801574707, "learning_rate": 4.604059777890537e-05, "loss": 0.2035, "num_input_tokens_seen": 13987440, "step": 7265 }, { "epoch": 1.1859858063463578, "grad_norm": 5.597310543060303, "learning_rate": 4.6035294090139145e-05, "loss": 0.2121, "num_input_tokens_seen": 13996976, "step": 7270 }, { "epoch": 1.186801533567175, "grad_norm": 5.320793151855469, "learning_rate": 4.6029987157466226e-05, "loss": 0.1219, "num_input_tokens_seen": 14007424, "step": 7275 }, { "epoch": 1.1876172607879925, "grad_norm": 1.7267528772354126, "learning_rate": 4.602467698170502e-05, "loss": 0.1346, "num_input_tokens_seen": 14019120, "step": 7280 }, { "epoch": 1.1884329880088098, "grad_norm": 0.4745176136493683, "learning_rate": 4.601936356367439e-05, "loss": 0.2016, "num_input_tokens_seen": 14028976, "step": 7285 }, { "epoch": 1.1892487152296272, "grad_norm": 1.250327467918396, "learning_rate": 4.601404690419377e-05, "loss": 0.1076, "num_input_tokens_seen": 14039200, "step": 7290 }, { "epoch": 1.1900644424504445, "grad_norm": 3.091090679168701, "learning_rate": 4.600872700408303e-05, "loss": 0.1731, "num_input_tokens_seen": 14047984, "step": 7295 }, { "epoch": 1.1908801696712619, "grad_norm": 2.6107940673828125, "learning_rate": 4.600340386416258e-05, "loss": 0.2404, "num_input_tokens_seen": 14057792, "step": 7300 }, { "epoch": 1.1916958968920792, "grad_norm": 2.6314947605133057, "learning_rate": 4.5998077485253296e-05, "loss": 0.1837, "num_input_tokens_seen": 14066048, "step": 7305 }, { "epoch": 1.1925116241128966, "grad_norm": 1.3942351341247559, "learning_rate": 4.59927478681766e-05, "loss": 0.1399, "num_input_tokens_seen": 14075744, "step": 7310 }, { "epoch": 1.193327351333714, "grad_norm": 0.734959065914154, "learning_rate": 4.5987415013754366e-05, "loss": 0.1053, "num_input_tokens_seen": 14086224, "step": 7315 }, { "epoch": 1.1941430785545313, "grad_norm": 2.435173273086548, "learning_rate": 4.598207892280899e-05, "loss": 0.0842, "num_input_tokens_seen": 14097744, "step": 7320 }, { "epoch": 1.1949588057753486, "grad_norm": 1.0310214757919312, "learning_rate": 4.597673959616337e-05, "loss": 0.0911, "num_input_tokens_seen": 14106752, "step": 7325 }, { "epoch": 1.195774532996166, "grad_norm": 4.333062171936035, "learning_rate": 4.597139703464089e-05, "loss": 0.1206, "num_input_tokens_seen": 14116256, "step": 7330 }, { "epoch": 1.1965902602169836, "grad_norm": 4.3255534172058105, "learning_rate": 4.596605123906545e-05, "loss": 0.0822, "num_input_tokens_seen": 14126608, "step": 7335 }, { "epoch": 1.1974059874378007, "grad_norm": 0.4104560613632202, "learning_rate": 4.596070221026143e-05, "loss": 0.1049, "num_input_tokens_seen": 14136064, "step": 7340 }, { "epoch": 1.1982217146586183, "grad_norm": 0.3880508542060852, "learning_rate": 4.595534994905372e-05, "loss": 0.0729, "num_input_tokens_seen": 14146160, "step": 7345 }, { "epoch": 1.1990374418794354, "grad_norm": 10.567431449890137, "learning_rate": 4.594999445626771e-05, "loss": 0.0456, "num_input_tokens_seen": 14155888, "step": 7350 }, { "epoch": 1.199853169100253, "grad_norm": 4.409867286682129, "learning_rate": 4.5944635732729276e-05, "loss": 0.1298, "num_input_tokens_seen": 14164912, "step": 7355 }, { "epoch": 1.2006688963210703, "grad_norm": 8.081650733947754, "learning_rate": 4.5939273779264804e-05, "loss": 0.1426, "num_input_tokens_seen": 14174096, "step": 7360 }, { "epoch": 1.2014846235418877, "grad_norm": 6.273383617401123, "learning_rate": 4.593390859670118e-05, "loss": 0.0606, "num_input_tokens_seen": 14183760, "step": 7365 }, { "epoch": 1.202300350762705, "grad_norm": 14.794708251953125, "learning_rate": 4.5928540185865776e-05, "loss": 0.2541, "num_input_tokens_seen": 14192704, "step": 7370 }, { "epoch": 1.2031160779835224, "grad_norm": 4.932717800140381, "learning_rate": 4.592316854758648e-05, "loss": 0.1973, "num_input_tokens_seen": 14202800, "step": 7375 }, { "epoch": 1.2039318052043397, "grad_norm": 5.262067794799805, "learning_rate": 4.5917793682691646e-05, "loss": 0.0176, "num_input_tokens_seen": 14212480, "step": 7380 }, { "epoch": 1.204747532425157, "grad_norm": 12.247564315795898, "learning_rate": 4.5912415592010164e-05, "loss": 0.1886, "num_input_tokens_seen": 14221648, "step": 7385 }, { "epoch": 1.2055632596459744, "grad_norm": 3.8080806732177734, "learning_rate": 4.5907034276371386e-05, "loss": 0.1906, "num_input_tokens_seen": 14231680, "step": 7390 }, { "epoch": 1.2063789868667918, "grad_norm": 1.3288699388504028, "learning_rate": 4.5901649736605196e-05, "loss": 0.0406, "num_input_tokens_seen": 14241824, "step": 7395 }, { "epoch": 1.2071947140876091, "grad_norm": 2.42669939994812, "learning_rate": 4.589626197354195e-05, "loss": 0.0991, "num_input_tokens_seen": 14252288, "step": 7400 }, { "epoch": 1.2071947140876091, "eval_loss": 0.13805000483989716, "eval_runtime": 80.6534, "eval_samples_per_second": 33.787, "eval_steps_per_second": 16.899, "num_input_tokens_seen": 14252288, "step": 7400 }, { "epoch": 1.2080104413084265, "grad_norm": 3.4057037830352783, "learning_rate": 4.5890870988012504e-05, "loss": 0.1313, "num_input_tokens_seen": 14261088, "step": 7405 }, { "epoch": 1.2088261685292438, "grad_norm": 4.414740085601807, "learning_rate": 4.5885476780848226e-05, "loss": 0.1136, "num_input_tokens_seen": 14271248, "step": 7410 }, { "epoch": 1.2096418957500612, "grad_norm": 5.0028300285339355, "learning_rate": 4.5880079352880964e-05, "loss": 0.1231, "num_input_tokens_seen": 14281216, "step": 7415 }, { "epoch": 1.2104576229708786, "grad_norm": 0.11986750364303589, "learning_rate": 4.5874678704943065e-05, "loss": 0.0818, "num_input_tokens_seen": 14291184, "step": 7420 }, { "epoch": 1.211273350191696, "grad_norm": 0.659478485584259, "learning_rate": 4.5869274837867394e-05, "loss": 0.0609, "num_input_tokens_seen": 14301472, "step": 7425 }, { "epoch": 1.2120890774125133, "grad_norm": 8.84797477722168, "learning_rate": 4.5863867752487275e-05, "loss": 0.1907, "num_input_tokens_seen": 14310880, "step": 7430 }, { "epoch": 1.2129048046333306, "grad_norm": 12.011825561523438, "learning_rate": 4.5858457449636554e-05, "loss": 0.0369, "num_input_tokens_seen": 14321408, "step": 7435 }, { "epoch": 1.213720531854148, "grad_norm": 3.5923550128936768, "learning_rate": 4.5853043930149574e-05, "loss": 0.0315, "num_input_tokens_seen": 14330608, "step": 7440 }, { "epoch": 1.2145362590749653, "grad_norm": 0.11774462461471558, "learning_rate": 4.584762719486117e-05, "loss": 0.1587, "num_input_tokens_seen": 14341072, "step": 7445 }, { "epoch": 1.2153519862957827, "grad_norm": 9.749540328979492, "learning_rate": 4.584220724460665e-05, "loss": 0.1933, "num_input_tokens_seen": 14351360, "step": 7450 }, { "epoch": 1.2161677135166, "grad_norm": 7.010673999786377, "learning_rate": 4.5836784080221865e-05, "loss": 0.0368, "num_input_tokens_seen": 14360880, "step": 7455 }, { "epoch": 1.2169834407374174, "grad_norm": 13.14829158782959, "learning_rate": 4.583135770254312e-05, "loss": 0.1702, "num_input_tokens_seen": 14371120, "step": 7460 }, { "epoch": 1.2177991679582347, "grad_norm": 7.564314842224121, "learning_rate": 4.5825928112407236e-05, "loss": 0.1485, "num_input_tokens_seen": 14380544, "step": 7465 }, { "epoch": 1.218614895179052, "grad_norm": 5.833128929138184, "learning_rate": 4.582049531065152e-05, "loss": 0.1973, "num_input_tokens_seen": 14389824, "step": 7470 }, { "epoch": 1.2194306223998694, "grad_norm": 0.2599206864833832, "learning_rate": 4.5815059298113783e-05, "loss": 0.2094, "num_input_tokens_seen": 14398816, "step": 7475 }, { "epoch": 1.2202463496206868, "grad_norm": 1.580451488494873, "learning_rate": 4.580962007563232e-05, "loss": 0.0879, "num_input_tokens_seen": 14408096, "step": 7480 }, { "epoch": 1.2210620768415041, "grad_norm": 3.6115918159484863, "learning_rate": 4.5804177644045935e-05, "loss": 0.1202, "num_input_tokens_seen": 14417552, "step": 7485 }, { "epoch": 1.2218778040623215, "grad_norm": 6.7855963706970215, "learning_rate": 4.579873200419391e-05, "loss": 0.1542, "num_input_tokens_seen": 14426816, "step": 7490 }, { "epoch": 1.2226935312831388, "grad_norm": 1.8887115716934204, "learning_rate": 4.5793283156916046e-05, "loss": 0.2297, "num_input_tokens_seen": 14435472, "step": 7495 }, { "epoch": 1.2235092585039562, "grad_norm": 2.497213840484619, "learning_rate": 4.578783110305261e-05, "loss": 0.1396, "num_input_tokens_seen": 14445344, "step": 7500 }, { "epoch": 1.2243249857247736, "grad_norm": 0.6618476510047913, "learning_rate": 4.578237584344438e-05, "loss": 0.088, "num_input_tokens_seen": 14456080, "step": 7505 }, { "epoch": 1.225140712945591, "grad_norm": 0.8020130395889282, "learning_rate": 4.577691737893263e-05, "loss": 0.1176, "num_input_tokens_seen": 14465536, "step": 7510 }, { "epoch": 1.2259564401664083, "grad_norm": 0.15240824222564697, "learning_rate": 4.577145571035912e-05, "loss": 0.1099, "num_input_tokens_seen": 14475152, "step": 7515 }, { "epoch": 1.2267721673872258, "grad_norm": 0.1636204570531845, "learning_rate": 4.576599083856611e-05, "loss": 0.086, "num_input_tokens_seen": 14484720, "step": 7520 }, { "epoch": 1.227587894608043, "grad_norm": 6.443145751953125, "learning_rate": 4.576052276439635e-05, "loss": 0.1184, "num_input_tokens_seen": 14493744, "step": 7525 }, { "epoch": 1.2284036218288605, "grad_norm": 0.42518267035484314, "learning_rate": 4.575505148869308e-05, "loss": 0.2109, "num_input_tokens_seen": 14503440, "step": 7530 }, { "epoch": 1.2292193490496777, "grad_norm": 4.848672866821289, "learning_rate": 4.574957701230006e-05, "loss": 0.1181, "num_input_tokens_seen": 14512432, "step": 7535 }, { "epoch": 1.2300350762704952, "grad_norm": 4.051324367523193, "learning_rate": 4.57440993360615e-05, "loss": 0.1243, "num_input_tokens_seen": 14522176, "step": 7540 }, { "epoch": 1.2308508034913126, "grad_norm": 3.219104051589966, "learning_rate": 4.5738618460822134e-05, "loss": 0.1319, "num_input_tokens_seen": 14531728, "step": 7545 }, { "epoch": 1.23166653071213, "grad_norm": 6.3264970779418945, "learning_rate": 4.573313438742719e-05, "loss": 0.0991, "num_input_tokens_seen": 14541184, "step": 7550 }, { "epoch": 1.2324822579329473, "grad_norm": 6.224959373474121, "learning_rate": 4.5727647116722374e-05, "loss": 0.0838, "num_input_tokens_seen": 14551232, "step": 7555 }, { "epoch": 1.2332979851537647, "grad_norm": 2.573307514190674, "learning_rate": 4.5722156649553884e-05, "loss": 0.2042, "num_input_tokens_seen": 14561008, "step": 7560 }, { "epoch": 1.234113712374582, "grad_norm": 6.027124404907227, "learning_rate": 4.571666298676843e-05, "loss": 0.173, "num_input_tokens_seen": 14571264, "step": 7565 }, { "epoch": 1.2349294395953994, "grad_norm": 7.128942012786865, "learning_rate": 4.571116612921321e-05, "loss": 0.0677, "num_input_tokens_seen": 14580240, "step": 7570 }, { "epoch": 1.2357451668162167, "grad_norm": 0.8201281428337097, "learning_rate": 4.57056660777359e-05, "loss": 0.0167, "num_input_tokens_seen": 14591040, "step": 7575 }, { "epoch": 1.236560894037034, "grad_norm": 0.5220300555229187, "learning_rate": 4.5700162833184666e-05, "loss": 0.1035, "num_input_tokens_seen": 14600368, "step": 7580 }, { "epoch": 1.2373766212578514, "grad_norm": 0.4787207841873169, "learning_rate": 4.5694656396408195e-05, "loss": 0.0428, "num_input_tokens_seen": 14609712, "step": 7585 }, { "epoch": 1.2381923484786688, "grad_norm": 8.41710090637207, "learning_rate": 4.5689146768255646e-05, "loss": 0.1936, "num_input_tokens_seen": 14620128, "step": 7590 }, { "epoch": 1.2390080756994861, "grad_norm": 3.7160089015960693, "learning_rate": 4.568363394957667e-05, "loss": 0.1791, "num_input_tokens_seen": 14629488, "step": 7595 }, { "epoch": 1.2398238029203035, "grad_norm": 9.52303695678711, "learning_rate": 4.567811794122141e-05, "loss": 0.0932, "num_input_tokens_seen": 14638816, "step": 7600 }, { "epoch": 1.2398238029203035, "eval_loss": 0.14474880695343018, "eval_runtime": 80.6295, "eval_samples_per_second": 33.797, "eval_steps_per_second": 16.904, "num_input_tokens_seen": 14638816, "step": 7600 }, { "epoch": 1.2406395301411208, "grad_norm": 2.16888689994812, "learning_rate": 4.56725987440405e-05, "loss": 0.1821, "num_input_tokens_seen": 14648240, "step": 7605 }, { "epoch": 1.2414552573619382, "grad_norm": 7.614789009094238, "learning_rate": 4.566707635888508e-05, "loss": 0.2218, "num_input_tokens_seen": 14657520, "step": 7610 }, { "epoch": 1.2422709845827555, "grad_norm": 5.429116249084473, "learning_rate": 4.566155078660677e-05, "loss": 0.1998, "num_input_tokens_seen": 14665696, "step": 7615 }, { "epoch": 1.2430867118035729, "grad_norm": 0.18809430301189423, "learning_rate": 4.565602202805768e-05, "loss": 0.1043, "num_input_tokens_seen": 14675392, "step": 7620 }, { "epoch": 1.2439024390243902, "grad_norm": 4.276175498962402, "learning_rate": 4.56504900840904e-05, "loss": 0.1047, "num_input_tokens_seen": 14684496, "step": 7625 }, { "epoch": 1.2447181662452076, "grad_norm": 3.7656161785125732, "learning_rate": 4.564495495555805e-05, "loss": 0.0512, "num_input_tokens_seen": 14693408, "step": 7630 }, { "epoch": 1.245533893466025, "grad_norm": 3.3400533199310303, "learning_rate": 4.5639416643314204e-05, "loss": 0.0422, "num_input_tokens_seen": 14703328, "step": 7635 }, { "epoch": 1.2463496206868423, "grad_norm": 1.8792784214019775, "learning_rate": 4.5633875148212946e-05, "loss": 0.0699, "num_input_tokens_seen": 14713536, "step": 7640 }, { "epoch": 1.2471653479076596, "grad_norm": 7.94925594329834, "learning_rate": 4.562833047110883e-05, "loss": 0.1024, "num_input_tokens_seen": 14723856, "step": 7645 }, { "epoch": 1.247981075128477, "grad_norm": 6.68981409072876, "learning_rate": 4.5622782612856923e-05, "loss": 0.1666, "num_input_tokens_seen": 14734352, "step": 7650 }, { "epoch": 1.2487968023492944, "grad_norm": 0.8546966910362244, "learning_rate": 4.561723157431278e-05, "loss": 0.119, "num_input_tokens_seen": 14743168, "step": 7655 }, { "epoch": 1.2496125295701117, "grad_norm": 10.474210739135742, "learning_rate": 4.5611677356332435e-05, "loss": 0.203, "num_input_tokens_seen": 14752480, "step": 7660 }, { "epoch": 1.250428256790929, "grad_norm": 2.726994752883911, "learning_rate": 4.560611995977242e-05, "loss": 0.206, "num_input_tokens_seen": 14762096, "step": 7665 }, { "epoch": 1.2512439840117464, "grad_norm": 2.1252636909484863, "learning_rate": 4.560055938548975e-05, "loss": 0.3419, "num_input_tokens_seen": 14773104, "step": 7670 }, { "epoch": 1.2520597112325638, "grad_norm": 2.3569560050964355, "learning_rate": 4.5594995634341944e-05, "loss": 0.1547, "num_input_tokens_seen": 14781968, "step": 7675 }, { "epoch": 1.2528754384533811, "grad_norm": 1.1456671953201294, "learning_rate": 4.5589428707187e-05, "loss": 0.0295, "num_input_tokens_seen": 14792016, "step": 7680 }, { "epoch": 1.2536911656741985, "grad_norm": 3.7267637252807617, "learning_rate": 4.55838586048834e-05, "loss": 0.0661, "num_input_tokens_seen": 14801440, "step": 7685 }, { "epoch": 1.2545068928950158, "grad_norm": 2.6881964206695557, "learning_rate": 4.557828532829013e-05, "loss": 0.1324, "num_input_tokens_seen": 14812496, "step": 7690 }, { "epoch": 1.2553226201158334, "grad_norm": 0.9688513278961182, "learning_rate": 4.557270887826667e-05, "loss": 0.1086, "num_input_tokens_seen": 14822144, "step": 7695 }, { "epoch": 1.2561383473366505, "grad_norm": 6.903630256652832, "learning_rate": 4.556712925567296e-05, "loss": 0.2371, "num_input_tokens_seen": 14832000, "step": 7700 }, { "epoch": 1.256954074557468, "grad_norm": 0.17563799023628235, "learning_rate": 4.5561546461369454e-05, "loss": 0.0804, "num_input_tokens_seen": 14842064, "step": 7705 }, { "epoch": 1.2577698017782852, "grad_norm": 0.08517669141292572, "learning_rate": 4.55559604962171e-05, "loss": 0.0882, "num_input_tokens_seen": 14850880, "step": 7710 }, { "epoch": 1.2585855289991028, "grad_norm": 0.25196418166160583, "learning_rate": 4.55503713610773e-05, "loss": 0.0503, "num_input_tokens_seen": 14859456, "step": 7715 }, { "epoch": 1.25940125621992, "grad_norm": 2.826772928237915, "learning_rate": 4.5544779056812e-05, "loss": 0.1248, "num_input_tokens_seen": 14870112, "step": 7720 }, { "epoch": 1.2602169834407375, "grad_norm": 5.233460426330566, "learning_rate": 4.553918358428358e-05, "loss": 0.1246, "num_input_tokens_seen": 14879088, "step": 7725 }, { "epoch": 1.2610327106615546, "grad_norm": 0.1777113974094391, "learning_rate": 4.553358494435494e-05, "loss": 0.1855, "num_input_tokens_seen": 14888560, "step": 7730 }, { "epoch": 1.2618484378823722, "grad_norm": 0.44819384813308716, "learning_rate": 4.5527983137889464e-05, "loss": 0.0434, "num_input_tokens_seen": 14898720, "step": 7735 }, { "epoch": 1.2626641651031896, "grad_norm": 0.6356499791145325, "learning_rate": 4.5522378165751015e-05, "loss": 0.015, "num_input_tokens_seen": 14908176, "step": 7740 }, { "epoch": 1.263479892324007, "grad_norm": 0.08788683265447617, "learning_rate": 4.5516770028803954e-05, "loss": 0.0923, "num_input_tokens_seen": 14918736, "step": 7745 }, { "epoch": 1.2642956195448243, "grad_norm": 0.5418336987495422, "learning_rate": 4.5511158727913116e-05, "loss": 0.1621, "num_input_tokens_seen": 14926784, "step": 7750 }, { "epoch": 1.2651113467656416, "grad_norm": 1.426465630531311, "learning_rate": 4.5505544263943856e-05, "loss": 0.2148, "num_input_tokens_seen": 14936688, "step": 7755 }, { "epoch": 1.265927073986459, "grad_norm": 0.28603413701057434, "learning_rate": 4.549992663776197e-05, "loss": 0.0116, "num_input_tokens_seen": 14946112, "step": 7760 }, { "epoch": 1.2667428012072763, "grad_norm": 3.732353687286377, "learning_rate": 4.5494305850233786e-05, "loss": 0.1449, "num_input_tokens_seen": 14955904, "step": 7765 }, { "epoch": 1.2675585284280937, "grad_norm": 4.775760650634766, "learning_rate": 4.5488681902226094e-05, "loss": 0.0676, "num_input_tokens_seen": 14965600, "step": 7770 }, { "epoch": 1.268374255648911, "grad_norm": 1.376336693763733, "learning_rate": 4.5483054794606174e-05, "loss": 0.1418, "num_input_tokens_seen": 14977296, "step": 7775 }, { "epoch": 1.2691899828697284, "grad_norm": 0.2971770167350769, "learning_rate": 4.547742452824179e-05, "loss": 0.095, "num_input_tokens_seen": 14986496, "step": 7780 }, { "epoch": 1.2700057100905457, "grad_norm": 0.40695950388908386, "learning_rate": 4.5471791104001215e-05, "loss": 0.1466, "num_input_tokens_seen": 14996256, "step": 7785 }, { "epoch": 1.270821437311363, "grad_norm": 0.23649172484874725, "learning_rate": 4.546615452275319e-05, "loss": 0.0201, "num_input_tokens_seen": 15006032, "step": 7790 }, { "epoch": 1.2716371645321805, "grad_norm": 1.8381125926971436, "learning_rate": 4.5460514785366944e-05, "loss": 0.126, "num_input_tokens_seen": 15015216, "step": 7795 }, { "epoch": 1.2724528917529978, "grad_norm": 0.1474628448486328, "learning_rate": 4.545487189271219e-05, "loss": 0.1439, "num_input_tokens_seen": 15024560, "step": 7800 }, { "epoch": 1.2724528917529978, "eval_loss": 0.14069685339927673, "eval_runtime": 80.4426, "eval_samples_per_second": 33.875, "eval_steps_per_second": 16.944, "num_input_tokens_seen": 15024560, "step": 7800 }, { "epoch": 1.2732686189738152, "grad_norm": 4.243251800537109, "learning_rate": 4.544922584565914e-05, "loss": 0.2049, "num_input_tokens_seen": 15034032, "step": 7805 }, { "epoch": 1.2740843461946325, "grad_norm": 1.4447987079620361, "learning_rate": 4.544357664507848e-05, "loss": 0.0728, "num_input_tokens_seen": 15043456, "step": 7810 }, { "epoch": 1.2749000734154499, "grad_norm": 5.8437957763671875, "learning_rate": 4.54379242918414e-05, "loss": 0.0776, "num_input_tokens_seen": 15052496, "step": 7815 }, { "epoch": 1.2757158006362672, "grad_norm": 1.2743998765945435, "learning_rate": 4.543226878681955e-05, "loss": 0.0838, "num_input_tokens_seen": 15063120, "step": 7820 }, { "epoch": 1.2765315278570846, "grad_norm": 2.2658214569091797, "learning_rate": 4.5426610130885087e-05, "loss": 0.2926, "num_input_tokens_seen": 15072192, "step": 7825 }, { "epoch": 1.277347255077902, "grad_norm": 4.095370769500732, "learning_rate": 4.542094832491064e-05, "loss": 0.1326, "num_input_tokens_seen": 15082784, "step": 7830 }, { "epoch": 1.2781629822987193, "grad_norm": 6.295642375946045, "learning_rate": 4.541528336976934e-05, "loss": 0.1335, "num_input_tokens_seen": 15093520, "step": 7835 }, { "epoch": 1.2789787095195366, "grad_norm": 2.89682674407959, "learning_rate": 4.540961526633479e-05, "loss": 0.1346, "num_input_tokens_seen": 15103696, "step": 7840 }, { "epoch": 1.279794436740354, "grad_norm": 3.9352378845214844, "learning_rate": 4.540394401548108e-05, "loss": 0.285, "num_input_tokens_seen": 15113072, "step": 7845 }, { "epoch": 1.2806101639611713, "grad_norm": 0.8565982580184937, "learning_rate": 4.539826961808279e-05, "loss": 0.1843, "num_input_tokens_seen": 15122272, "step": 7850 }, { "epoch": 1.2814258911819887, "grad_norm": 7.764593601226807, "learning_rate": 4.5392592075014994e-05, "loss": 0.0874, "num_input_tokens_seen": 15131232, "step": 7855 }, { "epoch": 1.282241618402806, "grad_norm": 4.285320281982422, "learning_rate": 4.538691138715322e-05, "loss": 0.137, "num_input_tokens_seen": 15140384, "step": 7860 }, { "epoch": 1.2830573456236234, "grad_norm": 0.3120591938495636, "learning_rate": 4.5381227555373516e-05, "loss": 0.0987, "num_input_tokens_seen": 15150336, "step": 7865 }, { "epoch": 1.2838730728444407, "grad_norm": 0.5004242658615112, "learning_rate": 4.537554058055239e-05, "loss": 0.0984, "num_input_tokens_seen": 15159408, "step": 7870 }, { "epoch": 1.284688800065258, "grad_norm": 2.1900999546051025, "learning_rate": 4.5369850463566865e-05, "loss": 0.1531, "num_input_tokens_seen": 15168288, "step": 7875 }, { "epoch": 1.2855045272860757, "grad_norm": 0.6691837310791016, "learning_rate": 4.5364157205294404e-05, "loss": 0.151, "num_input_tokens_seen": 15178576, "step": 7880 }, { "epoch": 1.2863202545068928, "grad_norm": 0.578788161277771, "learning_rate": 4.5358460806612996e-05, "loss": 0.1559, "num_input_tokens_seen": 15188240, "step": 7885 }, { "epoch": 1.2871359817277104, "grad_norm": 0.4566475450992584, "learning_rate": 4.535276126840109e-05, "loss": 0.0568, "num_input_tokens_seen": 15198448, "step": 7890 }, { "epoch": 1.2879517089485275, "grad_norm": 1.4952336549758911, "learning_rate": 4.5347058591537626e-05, "loss": 0.1636, "num_input_tokens_seen": 15208368, "step": 7895 }, { "epoch": 1.288767436169345, "grad_norm": 8.476726531982422, "learning_rate": 4.534135277690203e-05, "loss": 0.1178, "num_input_tokens_seen": 15217840, "step": 7900 }, { "epoch": 1.2895831633901622, "grad_norm": 0.29233303666114807, "learning_rate": 4.533564382537421e-05, "loss": 0.1225, "num_input_tokens_seen": 15227184, "step": 7905 }, { "epoch": 1.2903988906109798, "grad_norm": 3.672699213027954, "learning_rate": 4.532993173783456e-05, "loss": 0.2969, "num_input_tokens_seen": 15237136, "step": 7910 }, { "epoch": 1.291214617831797, "grad_norm": 3.1047465801239014, "learning_rate": 4.5324216515163954e-05, "loss": 0.1411, "num_input_tokens_seen": 15247600, "step": 7915 }, { "epoch": 1.2920303450526145, "grad_norm": 0.5802624821662903, "learning_rate": 4.531849815824375e-05, "loss": 0.1071, "num_input_tokens_seen": 15256976, "step": 7920 }, { "epoch": 1.2928460722734318, "grad_norm": 0.20674435794353485, "learning_rate": 4.5312776667955795e-05, "loss": 0.0314, "num_input_tokens_seen": 15266992, "step": 7925 }, { "epoch": 1.2936617994942492, "grad_norm": 2.4930789470672607, "learning_rate": 4.5307052045182405e-05, "loss": 0.0896, "num_input_tokens_seen": 15276416, "step": 7930 }, { "epoch": 1.2944775267150666, "grad_norm": 2.631802797317505, "learning_rate": 4.53013242908064e-05, "loss": 0.102, "num_input_tokens_seen": 15286352, "step": 7935 }, { "epoch": 1.295293253935884, "grad_norm": 0.19311267137527466, "learning_rate": 4.529559340571107e-05, "loss": 0.0741, "num_input_tokens_seen": 15295616, "step": 7940 }, { "epoch": 1.2961089811567013, "grad_norm": 5.840487003326416, "learning_rate": 4.528985939078018e-05, "loss": 0.068, "num_input_tokens_seen": 15304704, "step": 7945 }, { "epoch": 1.2969247083775186, "grad_norm": 4.033778667449951, "learning_rate": 4.5284122246898e-05, "loss": 0.152, "num_input_tokens_seen": 15315168, "step": 7950 }, { "epoch": 1.297740435598336, "grad_norm": 4.287422180175781, "learning_rate": 4.527838197494926e-05, "loss": 0.1617, "num_input_tokens_seen": 15324912, "step": 7955 }, { "epoch": 1.2985561628191533, "grad_norm": 3.7947020530700684, "learning_rate": 4.527263857581918e-05, "loss": 0.1508, "num_input_tokens_seen": 15335152, "step": 7960 }, { "epoch": 1.2993718900399707, "grad_norm": 6.960402965545654, "learning_rate": 4.526689205039347e-05, "loss": 0.1641, "num_input_tokens_seen": 15345936, "step": 7965 }, { "epoch": 1.300187617260788, "grad_norm": 3.6413371562957764, "learning_rate": 4.5261142399558324e-05, "loss": 0.1083, "num_input_tokens_seen": 15355376, "step": 7970 }, { "epoch": 1.3010033444816054, "grad_norm": 0.4401666224002838, "learning_rate": 4.525538962420041e-05, "loss": 0.0244, "num_input_tokens_seen": 15364304, "step": 7975 }, { "epoch": 1.3018190717024227, "grad_norm": 1.3188586235046387, "learning_rate": 4.524963372520685e-05, "loss": 0.1249, "num_input_tokens_seen": 15374336, "step": 7980 }, { "epoch": 1.30263479892324, "grad_norm": 7.42828893661499, "learning_rate": 4.524387470346531e-05, "loss": 0.0695, "num_input_tokens_seen": 15384128, "step": 7985 }, { "epoch": 1.3034505261440574, "grad_norm": 0.41795971989631653, "learning_rate": 4.5238112559863885e-05, "loss": 0.0213, "num_input_tokens_seen": 15394496, "step": 7990 }, { "epoch": 1.3042662533648748, "grad_norm": 0.2175310105085373, "learning_rate": 4.5232347295291175e-05, "loss": 0.0942, "num_input_tokens_seen": 15402992, "step": 7995 }, { "epoch": 1.3050819805856921, "grad_norm": 3.0505483150482178, "learning_rate": 4.522657891063626e-05, "loss": 0.0544, "num_input_tokens_seen": 15412000, "step": 8000 }, { "epoch": 1.3050819805856921, "eval_loss": 0.1473751664161682, "eval_runtime": 80.5651, "eval_samples_per_second": 33.824, "eval_steps_per_second": 16.918, "num_input_tokens_seen": 15412000, "step": 8000 }, { "epoch": 1.3058977078065095, "grad_norm": 0.10646350681781769, "learning_rate": 4.52208074067887e-05, "loss": 0.0832, "num_input_tokens_seen": 15421504, "step": 8005 }, { "epoch": 1.3067134350273268, "grad_norm": 7.038905143737793, "learning_rate": 4.5215032784638516e-05, "loss": 0.217, "num_input_tokens_seen": 15429856, "step": 8010 }, { "epoch": 1.3075291622481442, "grad_norm": 2.6406450271606445, "learning_rate": 4.5209255045076245e-05, "loss": 0.2323, "num_input_tokens_seen": 15439712, "step": 8015 }, { "epoch": 1.3083448894689615, "grad_norm": 0.3574487566947937, "learning_rate": 4.5203474188992875e-05, "loss": 0.0835, "num_input_tokens_seen": 15449888, "step": 8020 }, { "epoch": 1.309160616689779, "grad_norm": 4.508933067321777, "learning_rate": 4.51976902172799e-05, "loss": 0.0959, "num_input_tokens_seen": 15459440, "step": 8025 }, { "epoch": 1.3099763439105963, "grad_norm": 0.35206571221351624, "learning_rate": 4.519190313082927e-05, "loss": 0.1673, "num_input_tokens_seen": 15468704, "step": 8030 }, { "epoch": 1.3107920711314136, "grad_norm": 0.8002821207046509, "learning_rate": 4.518611293053343e-05, "loss": 0.1359, "num_input_tokens_seen": 15477984, "step": 8035 }, { "epoch": 1.311607798352231, "grad_norm": 3.5747299194335938, "learning_rate": 4.51803196172853e-05, "loss": 0.0475, "num_input_tokens_seen": 15485520, "step": 8040 }, { "epoch": 1.3124235255730483, "grad_norm": 3.691861867904663, "learning_rate": 4.517452319197828e-05, "loss": 0.2286, "num_input_tokens_seen": 15494592, "step": 8045 }, { "epoch": 1.3132392527938657, "grad_norm": 10.118099212646484, "learning_rate": 4.5168723655506265e-05, "loss": 0.0558, "num_input_tokens_seen": 15503136, "step": 8050 }, { "epoch": 1.314054980014683, "grad_norm": 0.9821534156799316, "learning_rate": 4.51629210087636e-05, "loss": 0.0764, "num_input_tokens_seen": 15511856, "step": 8055 }, { "epoch": 1.3148707072355004, "grad_norm": 5.31689453125, "learning_rate": 4.515711525264513e-05, "loss": 0.1837, "num_input_tokens_seen": 15520880, "step": 8060 }, { "epoch": 1.315686434456318, "grad_norm": 2.1364428997039795, "learning_rate": 4.5151306388046175e-05, "loss": 0.0516, "num_input_tokens_seen": 15531472, "step": 8065 }, { "epoch": 1.316502161677135, "grad_norm": 6.293585777282715, "learning_rate": 4.514549441586255e-05, "loss": 0.1462, "num_input_tokens_seen": 15540640, "step": 8070 }, { "epoch": 1.3173178888979526, "grad_norm": 3.7469077110290527, "learning_rate": 4.513967933699051e-05, "loss": 0.061, "num_input_tokens_seen": 15550496, "step": 8075 }, { "epoch": 1.3181336161187698, "grad_norm": 2.499439239501953, "learning_rate": 4.513386115232684e-05, "loss": 0.2481, "num_input_tokens_seen": 15559856, "step": 8080 }, { "epoch": 1.3189493433395874, "grad_norm": 1.5286515951156616, "learning_rate": 4.5128039862768745e-05, "loss": 0.0852, "num_input_tokens_seen": 15568880, "step": 8085 }, { "epoch": 1.3197650705604045, "grad_norm": 18.236595153808594, "learning_rate": 4.512221546921397e-05, "loss": 0.1412, "num_input_tokens_seen": 15578944, "step": 8090 }, { "epoch": 1.320580797781222, "grad_norm": 0.388948529958725, "learning_rate": 4.5116387972560694e-05, "loss": 0.0936, "num_input_tokens_seen": 15587600, "step": 8095 }, { "epoch": 1.3213965250020392, "grad_norm": 0.4639320969581604, "learning_rate": 4.511055737370759e-05, "loss": 0.0903, "num_input_tokens_seen": 15597264, "step": 8100 }, { "epoch": 1.3222122522228568, "grad_norm": 0.5632197856903076, "learning_rate": 4.510472367355383e-05, "loss": 0.0874, "num_input_tokens_seen": 15607392, "step": 8105 }, { "epoch": 1.3230279794436741, "grad_norm": 0.2058260440826416, "learning_rate": 4.509888687299901e-05, "loss": 0.0639, "num_input_tokens_seen": 15616048, "step": 8110 }, { "epoch": 1.3238437066644915, "grad_norm": 2.28214693069458, "learning_rate": 4.5093046972943266e-05, "loss": 0.2483, "num_input_tokens_seen": 15625488, "step": 8115 }, { "epoch": 1.3246594338853088, "grad_norm": 0.08630665391683578, "learning_rate": 4.508720397428717e-05, "loss": 0.2053, "num_input_tokens_seen": 15634000, "step": 8120 }, { "epoch": 1.3254751611061262, "grad_norm": 0.22236613929271698, "learning_rate": 4.508135787793178e-05, "loss": 0.064, "num_input_tokens_seen": 15644464, "step": 8125 }, { "epoch": 1.3262908883269435, "grad_norm": 1.869123935699463, "learning_rate": 4.5075508684778664e-05, "loss": 0.2062, "num_input_tokens_seen": 15653488, "step": 8130 }, { "epoch": 1.3271066155477609, "grad_norm": 3.5261621475219727, "learning_rate": 4.506965639572982e-05, "loss": 0.0928, "num_input_tokens_seen": 15664512, "step": 8135 }, { "epoch": 1.3279223427685782, "grad_norm": 5.115334987640381, "learning_rate": 4.506380101168774e-05, "loss": 0.1041, "num_input_tokens_seen": 15674400, "step": 8140 }, { "epoch": 1.3287380699893956, "grad_norm": 5.7173752784729, "learning_rate": 4.505794253355542e-05, "loss": 0.0861, "num_input_tokens_seen": 15683392, "step": 8145 }, { "epoch": 1.329553797210213, "grad_norm": 7.50900936126709, "learning_rate": 4.5052080962236286e-05, "loss": 0.2513, "num_input_tokens_seen": 15692848, "step": 8150 }, { "epoch": 1.3303695244310303, "grad_norm": 7.074011325836182, "learning_rate": 4.504621629863428e-05, "loss": 0.174, "num_input_tokens_seen": 15703088, "step": 8155 }, { "epoch": 1.3311852516518476, "grad_norm": 4.914999961853027, "learning_rate": 4.504034854365381e-05, "loss": 0.067, "num_input_tokens_seen": 15712800, "step": 8160 }, { "epoch": 1.332000978872665, "grad_norm": 3.463705062866211, "learning_rate": 4.503447769819974e-05, "loss": 0.092, "num_input_tokens_seen": 15721664, "step": 8165 }, { "epoch": 1.3328167060934824, "grad_norm": 0.19997531175613403, "learning_rate": 4.502860376317745e-05, "loss": 0.1982, "num_input_tokens_seen": 15731328, "step": 8170 }, { "epoch": 1.3336324333142997, "grad_norm": 3.976003885269165, "learning_rate": 4.502272673949276e-05, "loss": 0.1585, "num_input_tokens_seen": 15740448, "step": 8175 }, { "epoch": 1.334448160535117, "grad_norm": 1.0777143239974976, "learning_rate": 4.501684662805199e-05, "loss": 0.0523, "num_input_tokens_seen": 15750208, "step": 8180 }, { "epoch": 1.3352638877559344, "grad_norm": 0.41291847825050354, "learning_rate": 4.5010963429761924e-05, "loss": 0.0889, "num_input_tokens_seen": 15760384, "step": 8185 }, { "epoch": 1.3360796149767518, "grad_norm": 0.12458813190460205, "learning_rate": 4.500507714552982e-05, "loss": 0.054, "num_input_tokens_seen": 15770320, "step": 8190 }, { "epoch": 1.3368953421975691, "grad_norm": 0.22758689522743225, "learning_rate": 4.499918777626342e-05, "loss": 0.0978, "num_input_tokens_seen": 15779632, "step": 8195 }, { "epoch": 1.3377110694183865, "grad_norm": 4.605654239654541, "learning_rate": 4.499329532287093e-05, "loss": 0.1348, "num_input_tokens_seen": 15789456, "step": 8200 }, { "epoch": 1.3377110694183865, "eval_loss": 0.13672591745853424, "eval_runtime": 80.6129, "eval_samples_per_second": 33.804, "eval_steps_per_second": 16.908, "num_input_tokens_seen": 15789456, "step": 8200 }, { "epoch": 1.3385267966392038, "grad_norm": 0.5996509790420532, "learning_rate": 4.4987399786261064e-05, "loss": 0.0306, "num_input_tokens_seen": 15798384, "step": 8205 }, { "epoch": 1.3393425238600212, "grad_norm": 4.302902698516846, "learning_rate": 4.498150116734297e-05, "loss": 0.1579, "num_input_tokens_seen": 15807008, "step": 8210 }, { "epoch": 1.3401582510808385, "grad_norm": 14.511566162109375, "learning_rate": 4.4975599467026294e-05, "loss": 0.2549, "num_input_tokens_seen": 15816640, "step": 8215 }, { "epoch": 1.3409739783016559, "grad_norm": 3.23539662361145, "learning_rate": 4.496969468622114e-05, "loss": 0.0787, "num_input_tokens_seen": 15828048, "step": 8220 }, { "epoch": 1.3417897055224732, "grad_norm": 0.15357667207717896, "learning_rate": 4.496378682583813e-05, "loss": 0.0622, "num_input_tokens_seen": 15836976, "step": 8225 }, { "epoch": 1.3426054327432906, "grad_norm": 0.16332300007343292, "learning_rate": 4.495787588678829e-05, "loss": 0.0986, "num_input_tokens_seen": 15846368, "step": 8230 }, { "epoch": 1.343421159964108, "grad_norm": 5.190822601318359, "learning_rate": 4.4951961869983196e-05, "loss": 0.0316, "num_input_tokens_seen": 15856496, "step": 8235 }, { "epoch": 1.3442368871849253, "grad_norm": 13.671947479248047, "learning_rate": 4.494604477633485e-05, "loss": 0.0396, "num_input_tokens_seen": 15866528, "step": 8240 }, { "epoch": 1.3450526144057426, "grad_norm": 4.4034423828125, "learning_rate": 4.4940124606755734e-05, "loss": 0.0405, "num_input_tokens_seen": 15875936, "step": 8245 }, { "epoch": 1.3458683416265602, "grad_norm": 0.3740399479866028, "learning_rate": 4.493420136215882e-05, "loss": 0.0107, "num_input_tokens_seen": 15884560, "step": 8250 }, { "epoch": 1.3466840688473773, "grad_norm": 5.342568874359131, "learning_rate": 4.492827504345756e-05, "loss": 0.1721, "num_input_tokens_seen": 15893968, "step": 8255 }, { "epoch": 1.347499796068195, "grad_norm": 0.07206309586763382, "learning_rate": 4.492234565156584e-05, "loss": 0.1041, "num_input_tokens_seen": 15903776, "step": 8260 }, { "epoch": 1.348315523289012, "grad_norm": 2.910031795501709, "learning_rate": 4.491641318739807e-05, "loss": 0.1535, "num_input_tokens_seen": 15912208, "step": 8265 }, { "epoch": 1.3491312505098296, "grad_norm": 2.3412628173828125, "learning_rate": 4.4910477651869096e-05, "loss": 0.0795, "num_input_tokens_seen": 15922832, "step": 8270 }, { "epoch": 1.3499469777306468, "grad_norm": 0.19453145563602448, "learning_rate": 4.4904539045894254e-05, "loss": 0.0811, "num_input_tokens_seen": 15933520, "step": 8275 }, { "epoch": 1.3507627049514643, "grad_norm": 0.19833339750766754, "learning_rate": 4.4898597370389364e-05, "loss": 0.1257, "num_input_tokens_seen": 15943296, "step": 8280 }, { "epoch": 1.3515784321722815, "grad_norm": 0.5475854873657227, "learning_rate": 4.489265262627069e-05, "loss": 0.1683, "num_input_tokens_seen": 15952336, "step": 8285 }, { "epoch": 1.352394159393099, "grad_norm": 0.9318704605102539, "learning_rate": 4.488670481445499e-05, "loss": 0.116, "num_input_tokens_seen": 15961648, "step": 8290 }, { "epoch": 1.3532098866139162, "grad_norm": 2.43410325050354, "learning_rate": 4.488075393585951e-05, "loss": 0.0432, "num_input_tokens_seen": 15971072, "step": 8295 }, { "epoch": 1.3540256138347337, "grad_norm": 0.06614109873771667, "learning_rate": 4.487479999140193e-05, "loss": 0.0842, "num_input_tokens_seen": 15980128, "step": 8300 }, { "epoch": 1.354841341055551, "grad_norm": 0.08831585198640823, "learning_rate": 4.4868842982000425e-05, "loss": 0.0934, "num_input_tokens_seen": 15989584, "step": 8305 }, { "epoch": 1.3556570682763684, "grad_norm": 0.1588425487279892, "learning_rate": 4.486288290857365e-05, "loss": 0.1476, "num_input_tokens_seen": 15998880, "step": 8310 }, { "epoch": 1.3564727954971858, "grad_norm": 7.6119771003723145, "learning_rate": 4.4856919772040715e-05, "loss": 0.1723, "num_input_tokens_seen": 16009104, "step": 8315 }, { "epoch": 1.3572885227180032, "grad_norm": 0.191700741648674, "learning_rate": 4.485095357332122e-05, "loss": 0.0871, "num_input_tokens_seen": 16018640, "step": 8320 }, { "epoch": 1.3581042499388205, "grad_norm": 0.313199520111084, "learning_rate": 4.484498431333521e-05, "loss": 0.0216, "num_input_tokens_seen": 16028688, "step": 8325 }, { "epoch": 1.3589199771596379, "grad_norm": 1.2567901611328125, "learning_rate": 4.4839011993003245e-05, "loss": 0.1484, "num_input_tokens_seen": 16039232, "step": 8330 }, { "epoch": 1.3597357043804552, "grad_norm": 0.4149797856807709, "learning_rate": 4.4833036613246305e-05, "loss": 0.0862, "num_input_tokens_seen": 16049536, "step": 8335 }, { "epoch": 1.3605514316012726, "grad_norm": 1.2337677478790283, "learning_rate": 4.482705817498589e-05, "loss": 0.1136, "num_input_tokens_seen": 16060032, "step": 8340 }, { "epoch": 1.36136715882209, "grad_norm": 2.787531614303589, "learning_rate": 4.4821076679143934e-05, "loss": 0.1947, "num_input_tokens_seen": 16070064, "step": 8345 }, { "epoch": 1.3621828860429073, "grad_norm": 0.236080601811409, "learning_rate": 4.481509212664288e-05, "loss": 0.0126, "num_input_tokens_seen": 16079408, "step": 8350 }, { "epoch": 1.3629986132637246, "grad_norm": 5.98255729675293, "learning_rate": 4.480910451840559e-05, "loss": 0.1652, "num_input_tokens_seen": 16087440, "step": 8355 }, { "epoch": 1.363814340484542, "grad_norm": 0.1754792183637619, "learning_rate": 4.480311385535546e-05, "loss": 0.1168, "num_input_tokens_seen": 16097584, "step": 8360 }, { "epoch": 1.3646300677053593, "grad_norm": 4.373373985290527, "learning_rate": 4.47971201384163e-05, "loss": 0.095, "num_input_tokens_seen": 16106864, "step": 8365 }, { "epoch": 1.3654457949261767, "grad_norm": 1.9650225639343262, "learning_rate": 4.4791123368512446e-05, "loss": 0.1482, "num_input_tokens_seen": 16116656, "step": 8370 }, { "epoch": 1.366261522146994, "grad_norm": 1.332412838935852, "learning_rate": 4.478512354656864e-05, "loss": 0.1005, "num_input_tokens_seen": 16125776, "step": 8375 }, { "epoch": 1.3670772493678114, "grad_norm": 0.13586121797561646, "learning_rate": 4.477912067351016e-05, "loss": 0.1137, "num_input_tokens_seen": 16135216, "step": 8380 }, { "epoch": 1.3678929765886287, "grad_norm": 0.16403527557849884, "learning_rate": 4.477311475026271e-05, "loss": 0.0675, "num_input_tokens_seen": 16144080, "step": 8385 }, { "epoch": 1.368708703809446, "grad_norm": 4.038161754608154, "learning_rate": 4.476710577775248e-05, "loss": 0.1608, "num_input_tokens_seen": 16153392, "step": 8390 }, { "epoch": 1.3695244310302634, "grad_norm": 0.6109093427658081, "learning_rate": 4.476109375690612e-05, "loss": 0.0606, "num_input_tokens_seen": 16162784, "step": 8395 }, { "epoch": 1.3703401582510808, "grad_norm": 0.06936314702033997, "learning_rate": 4.4755078688650784e-05, "loss": 0.0182, "num_input_tokens_seen": 16173616, "step": 8400 }, { "epoch": 1.3703401582510808, "eval_loss": 0.1445779800415039, "eval_runtime": 80.5165, "eval_samples_per_second": 33.844, "eval_steps_per_second": 16.928, "num_input_tokens_seen": 16173616, "step": 8400 }, { "epoch": 1.3711558854718982, "grad_norm": 0.10116879642009735, "learning_rate": 4.474906057391406e-05, "loss": 0.0505, "num_input_tokens_seen": 16183520, "step": 8405 }, { "epoch": 1.3719716126927155, "grad_norm": 2.3805887699127197, "learning_rate": 4.4743039413624e-05, "loss": 0.3477, "num_input_tokens_seen": 16192192, "step": 8410 }, { "epoch": 1.3727873399135329, "grad_norm": 0.31667837500572205, "learning_rate": 4.473701520870916e-05, "loss": 0.086, "num_input_tokens_seen": 16201888, "step": 8415 }, { "epoch": 1.3736030671343502, "grad_norm": 3.6076037883758545, "learning_rate": 4.4730987960098544e-05, "loss": 0.2345, "num_input_tokens_seen": 16211760, "step": 8420 }, { "epoch": 1.3744187943551676, "grad_norm": 3.9711668491363525, "learning_rate": 4.4724957668721635e-05, "loss": 0.1655, "num_input_tokens_seen": 16222256, "step": 8425 }, { "epoch": 1.375234521575985, "grad_norm": 1.1858608722686768, "learning_rate": 4.471892433550836e-05, "loss": 0.0298, "num_input_tokens_seen": 16231856, "step": 8430 }, { "epoch": 1.3760502487968023, "grad_norm": 5.863003253936768, "learning_rate": 4.471288796138916e-05, "loss": 0.1449, "num_input_tokens_seen": 16240752, "step": 8435 }, { "epoch": 1.3768659760176196, "grad_norm": 5.009912967681885, "learning_rate": 4.470684854729491e-05, "loss": 0.1315, "num_input_tokens_seen": 16250960, "step": 8440 }, { "epoch": 1.3776817032384372, "grad_norm": 4.250934600830078, "learning_rate": 4.4700806094156955e-05, "loss": 0.1344, "num_input_tokens_seen": 16260016, "step": 8445 }, { "epoch": 1.3784974304592543, "grad_norm": 2.893958330154419, "learning_rate": 4.469476060290713e-05, "loss": 0.2106, "num_input_tokens_seen": 16270304, "step": 8450 }, { "epoch": 1.379313157680072, "grad_norm": 0.23388828337192535, "learning_rate": 4.468871207447772e-05, "loss": 0.0486, "num_input_tokens_seen": 16279664, "step": 8455 }, { "epoch": 1.380128884900889, "grad_norm": 1.7357455492019653, "learning_rate": 4.4682660509801486e-05, "loss": 0.0752, "num_input_tokens_seen": 16289184, "step": 8460 }, { "epoch": 1.3809446121217066, "grad_norm": 5.859502792358398, "learning_rate": 4.467660590981165e-05, "loss": 0.2016, "num_input_tokens_seen": 16300224, "step": 8465 }, { "epoch": 1.3817603393425237, "grad_norm": 6.581857204437256, "learning_rate": 4.467054827544191e-05, "loss": 0.091, "num_input_tokens_seen": 16309536, "step": 8470 }, { "epoch": 1.3825760665633413, "grad_norm": 0.5576367378234863, "learning_rate": 4.4664487607626434e-05, "loss": 0.0902, "num_input_tokens_seen": 16317600, "step": 8475 }, { "epoch": 1.3833917937841584, "grad_norm": 4.272623062133789, "learning_rate": 4.4658423907299845e-05, "loss": 0.246, "num_input_tokens_seen": 16327792, "step": 8480 }, { "epoch": 1.384207521004976, "grad_norm": 0.21360667049884796, "learning_rate": 4.465235717539725e-05, "loss": 0.1123, "num_input_tokens_seen": 16338384, "step": 8485 }, { "epoch": 1.3850232482257934, "grad_norm": 0.13778169453144073, "learning_rate": 4.464628741285421e-05, "loss": 0.0136, "num_input_tokens_seen": 16347792, "step": 8490 }, { "epoch": 1.3858389754466107, "grad_norm": 2.1035468578338623, "learning_rate": 4.4640214620606754e-05, "loss": 0.0689, "num_input_tokens_seen": 16358672, "step": 8495 }, { "epoch": 1.386654702667428, "grad_norm": 0.17581866681575775, "learning_rate": 4.46341387995914e-05, "loss": 0.0668, "num_input_tokens_seen": 16368016, "step": 8500 }, { "epoch": 1.3874704298882454, "grad_norm": 1.1194713115692139, "learning_rate": 4.4628059950745106e-05, "loss": 0.08, "num_input_tokens_seen": 16377504, "step": 8505 }, { "epoch": 1.3882861571090628, "grad_norm": 5.12571907043457, "learning_rate": 4.4621978075005297e-05, "loss": 0.3058, "num_input_tokens_seen": 16386528, "step": 8510 }, { "epoch": 1.3891018843298801, "grad_norm": 6.598723888397217, "learning_rate": 4.461589317330989e-05, "loss": 0.1223, "num_input_tokens_seen": 16397376, "step": 8515 }, { "epoch": 1.3899176115506975, "grad_norm": 3.5887181758880615, "learning_rate": 4.460980524659724e-05, "loss": 0.1195, "num_input_tokens_seen": 16406304, "step": 8520 }, { "epoch": 1.3907333387715148, "grad_norm": 1.2626854181289673, "learning_rate": 4.46037142958062e-05, "loss": 0.1507, "num_input_tokens_seen": 16415424, "step": 8525 }, { "epoch": 1.3915490659923322, "grad_norm": 3.8043477535247803, "learning_rate": 4.4597620321876046e-05, "loss": 0.072, "num_input_tokens_seen": 16425536, "step": 8530 }, { "epoch": 1.3923647932131495, "grad_norm": 7.035579681396484, "learning_rate": 4.459152332574656e-05, "loss": 0.0631, "num_input_tokens_seen": 16433248, "step": 8535 }, { "epoch": 1.393180520433967, "grad_norm": 3.2972230911254883, "learning_rate": 4.4585423308357985e-05, "loss": 0.1855, "num_input_tokens_seen": 16442416, "step": 8540 }, { "epoch": 1.3939962476547842, "grad_norm": 3.979217290878296, "learning_rate": 4.457932027065102e-05, "loss": 0.1919, "num_input_tokens_seen": 16452416, "step": 8545 }, { "epoch": 1.3948119748756016, "grad_norm": 11.088677406311035, "learning_rate": 4.45732142135668e-05, "loss": 0.1467, "num_input_tokens_seen": 16461952, "step": 8550 }, { "epoch": 1.395627702096419, "grad_norm": 4.881683826446533, "learning_rate": 4.4567105138046986e-05, "loss": 0.1708, "num_input_tokens_seen": 16472432, "step": 8555 }, { "epoch": 1.3964434293172363, "grad_norm": 4.6320977210998535, "learning_rate": 4.456099304503365e-05, "loss": 0.2164, "num_input_tokens_seen": 16482288, "step": 8560 }, { "epoch": 1.3972591565380537, "grad_norm": 3.1480393409729004, "learning_rate": 4.455487793546939e-05, "loss": 0.2086, "num_input_tokens_seen": 16492544, "step": 8565 }, { "epoch": 1.398074883758871, "grad_norm": 0.26666608452796936, "learning_rate": 4.454875981029719e-05, "loss": 0.139, "num_input_tokens_seen": 16502048, "step": 8570 }, { "epoch": 1.3988906109796884, "grad_norm": 0.37394261360168457, "learning_rate": 4.454263867046057e-05, "loss": 0.0278, "num_input_tokens_seen": 16511008, "step": 8575 }, { "epoch": 1.3997063382005057, "grad_norm": 2.394345760345459, "learning_rate": 4.4536514516903484e-05, "loss": 0.1287, "num_input_tokens_seen": 16520656, "step": 8580 }, { "epoch": 1.400522065421323, "grad_norm": 1.2217938899993896, "learning_rate": 4.453038735057034e-05, "loss": 0.117, "num_input_tokens_seen": 16529552, "step": 8585 }, { "epoch": 1.4013377926421404, "grad_norm": 0.13570037484169006, "learning_rate": 4.4524257172406034e-05, "loss": 0.0916, "num_input_tokens_seen": 16539392, "step": 8590 }, { "epoch": 1.4021535198629578, "grad_norm": 3.46962571144104, "learning_rate": 4.451812398335592e-05, "loss": 0.2925, "num_input_tokens_seen": 16548800, "step": 8595 }, { "epoch": 1.4029692470837751, "grad_norm": 0.636460542678833, "learning_rate": 4.4511987784365805e-05, "loss": 0.0708, "num_input_tokens_seen": 16558464, "step": 8600 }, { "epoch": 1.4029692470837751, "eval_loss": 0.13184860348701477, "eval_runtime": 80.5156, "eval_samples_per_second": 33.844, "eval_steps_per_second": 16.928, "num_input_tokens_seen": 16558464, "step": 8600 }, { "epoch": 1.4037849743045925, "grad_norm": 0.34172818064689636, "learning_rate": 4.450584857638197e-05, "loss": 0.2357, "num_input_tokens_seen": 16566832, "step": 8605 }, { "epoch": 1.4046007015254098, "grad_norm": 3.679274320602417, "learning_rate": 4.449970636035116e-05, "loss": 0.078, "num_input_tokens_seen": 16576912, "step": 8610 }, { "epoch": 1.4054164287462272, "grad_norm": 1.6555323600769043, "learning_rate": 4.4493561137220574e-05, "loss": 0.0755, "num_input_tokens_seen": 16586544, "step": 8615 }, { "epoch": 1.4062321559670445, "grad_norm": 5.926528453826904, "learning_rate": 4.44874129079379e-05, "loss": 0.2003, "num_input_tokens_seen": 16595808, "step": 8620 }, { "epoch": 1.407047883187862, "grad_norm": 0.13392134010791779, "learning_rate": 4.4481261673451255e-05, "loss": 0.0693, "num_input_tokens_seen": 16604848, "step": 8625 }, { "epoch": 1.4078636104086795, "grad_norm": 0.22195963561534882, "learning_rate": 4.4475107434709245e-05, "loss": 0.0642, "num_input_tokens_seen": 16614480, "step": 8630 }, { "epoch": 1.4086793376294966, "grad_norm": 0.28545960783958435, "learning_rate": 4.446895019266093e-05, "loss": 0.0296, "num_input_tokens_seen": 16623520, "step": 8635 }, { "epoch": 1.4094950648503142, "grad_norm": 2.0419342517852783, "learning_rate": 4.446278994825583e-05, "loss": 0.0805, "num_input_tokens_seen": 16633216, "step": 8640 }, { "epoch": 1.4103107920711313, "grad_norm": 0.1604013741016388, "learning_rate": 4.445662670244394e-05, "loss": 0.2129, "num_input_tokens_seen": 16642768, "step": 8645 }, { "epoch": 1.4111265192919489, "grad_norm": 2.2206602096557617, "learning_rate": 4.44504604561757e-05, "loss": 0.2205, "num_input_tokens_seen": 16652032, "step": 8650 }, { "epoch": 1.411942246512766, "grad_norm": 3.603525400161743, "learning_rate": 4.4444291210402035e-05, "loss": 0.0905, "num_input_tokens_seen": 16660976, "step": 8655 }, { "epoch": 1.4127579737335836, "grad_norm": 0.2829318642616272, "learning_rate": 4.443811896607431e-05, "loss": 0.018, "num_input_tokens_seen": 16670496, "step": 8660 }, { "epoch": 1.4135737009544007, "grad_norm": 3.4557573795318604, "learning_rate": 4.443194372414436e-05, "loss": 0.0921, "num_input_tokens_seen": 16681232, "step": 8665 }, { "epoch": 1.4143894281752183, "grad_norm": 2.2992560863494873, "learning_rate": 4.442576548556449e-05, "loss": 0.1016, "num_input_tokens_seen": 16690576, "step": 8670 }, { "epoch": 1.4152051553960356, "grad_norm": 0.3306518793106079, "learning_rate": 4.441958425128747e-05, "loss": 0.0945, "num_input_tokens_seen": 16700352, "step": 8675 }, { "epoch": 1.416020882616853, "grad_norm": 0.3908638060092926, "learning_rate": 4.4413400022266515e-05, "loss": 0.0114, "num_input_tokens_seen": 16711264, "step": 8680 }, { "epoch": 1.4168366098376703, "grad_norm": 0.240587055683136, "learning_rate": 4.4407212799455313e-05, "loss": 0.0631, "num_input_tokens_seen": 16721872, "step": 8685 }, { "epoch": 1.4176523370584877, "grad_norm": 4.559015274047852, "learning_rate": 4.4401022583808003e-05, "loss": 0.217, "num_input_tokens_seen": 16730912, "step": 8690 }, { "epoch": 1.418468064279305, "grad_norm": 0.8984541893005371, "learning_rate": 4.439482937627921e-05, "loss": 0.1469, "num_input_tokens_seen": 16741632, "step": 8695 }, { "epoch": 1.4192837915001224, "grad_norm": 0.499795138835907, "learning_rate": 4.4388633177824004e-05, "loss": 0.1716, "num_input_tokens_seen": 16750080, "step": 8700 }, { "epoch": 1.4200995187209398, "grad_norm": 4.27479887008667, "learning_rate": 4.4382433989397895e-05, "loss": 0.1829, "num_input_tokens_seen": 16758240, "step": 8705 }, { "epoch": 1.420915245941757, "grad_norm": 13.089518547058105, "learning_rate": 4.4376231811956895e-05, "loss": 0.2039, "num_input_tokens_seen": 16768224, "step": 8710 }, { "epoch": 1.4217309731625745, "grad_norm": 2.9451465606689453, "learning_rate": 4.437002664645745e-05, "loss": 0.1878, "num_input_tokens_seen": 16778688, "step": 8715 }, { "epoch": 1.4225467003833918, "grad_norm": 0.36847200989723206, "learning_rate": 4.436381849385649e-05, "loss": 0.155, "num_input_tokens_seen": 16788400, "step": 8720 }, { "epoch": 1.4233624276042092, "grad_norm": 4.059162616729736, "learning_rate": 4.435760735511136e-05, "loss": 0.1273, "num_input_tokens_seen": 16798368, "step": 8725 }, { "epoch": 1.4241781548250265, "grad_norm": 0.13560108840465546, "learning_rate": 4.435139323117992e-05, "loss": 0.1833, "num_input_tokens_seen": 16807072, "step": 8730 }, { "epoch": 1.4249938820458439, "grad_norm": 0.2017868608236313, "learning_rate": 4.434517612302046e-05, "loss": 0.1023, "num_input_tokens_seen": 16818112, "step": 8735 }, { "epoch": 1.4258096092666612, "grad_norm": 0.5734509229660034, "learning_rate": 4.433895603159174e-05, "loss": 0.1334, "num_input_tokens_seen": 16829440, "step": 8740 }, { "epoch": 1.4266253364874786, "grad_norm": 0.2788032591342926, "learning_rate": 4.433273295785296e-05, "loss": 0.1156, "num_input_tokens_seen": 16839712, "step": 8745 }, { "epoch": 1.427441063708296, "grad_norm": 6.170844078063965, "learning_rate": 4.432650690276382e-05, "loss": 0.1234, "num_input_tokens_seen": 16848608, "step": 8750 }, { "epoch": 1.4282567909291133, "grad_norm": 0.2763994038105011, "learning_rate": 4.4320277867284435e-05, "loss": 0.0971, "num_input_tokens_seen": 16859696, "step": 8755 }, { "epoch": 1.4290725181499306, "grad_norm": 7.282841205596924, "learning_rate": 4.431404585237541e-05, "loss": 0.1164, "num_input_tokens_seen": 16868832, "step": 8760 }, { "epoch": 1.429888245370748, "grad_norm": 10.706565856933594, "learning_rate": 4.43078108589978e-05, "loss": 0.0671, "num_input_tokens_seen": 16878288, "step": 8765 }, { "epoch": 1.4307039725915653, "grad_norm": 1.9361652135849, "learning_rate": 4.4301572888113116e-05, "loss": 0.1301, "num_input_tokens_seen": 16888624, "step": 8770 }, { "epoch": 1.4315196998123827, "grad_norm": 1.7338529825210571, "learning_rate": 4.4295331940683337e-05, "loss": 0.0434, "num_input_tokens_seen": 16897744, "step": 8775 }, { "epoch": 1.4323354270332, "grad_norm": 1.4866844415664673, "learning_rate": 4.428908801767089e-05, "loss": 0.0574, "num_input_tokens_seen": 16907632, "step": 8780 }, { "epoch": 1.4331511542540174, "grad_norm": 0.38377562165260315, "learning_rate": 4.428284112003868e-05, "loss": 0.075, "num_input_tokens_seen": 16916752, "step": 8785 }, { "epoch": 1.4339668814748348, "grad_norm": 0.12664037942886353, "learning_rate": 4.4276591248750033e-05, "loss": 0.1503, "num_input_tokens_seen": 16926176, "step": 8790 }, { "epoch": 1.434782608695652, "grad_norm": 0.90788733959198, "learning_rate": 4.4270338404768774e-05, "loss": 0.0425, "num_input_tokens_seen": 16936336, "step": 8795 }, { "epoch": 1.4355983359164695, "grad_norm": 0.4151371717453003, "learning_rate": 4.426408258905917e-05, "loss": 0.0205, "num_input_tokens_seen": 16945488, "step": 8800 }, { "epoch": 1.4355983359164695, "eval_loss": 0.14636728167533875, "eval_runtime": 80.7279, "eval_samples_per_second": 33.755, "eval_steps_per_second": 16.884, "num_input_tokens_seen": 16945488, "step": 8800 }, { "epoch": 1.4364140631372868, "grad_norm": 5.6986894607543945, "learning_rate": 4.425782380258594e-05, "loss": 0.2288, "num_input_tokens_seen": 16955408, "step": 8805 }, { "epoch": 1.4372297903581042, "grad_norm": 3.822000503540039, "learning_rate": 4.425156204631427e-05, "loss": 0.0959, "num_input_tokens_seen": 16965680, "step": 8810 }, { "epoch": 1.4380455175789217, "grad_norm": 4.078396320343018, "learning_rate": 4.424529732120981e-05, "loss": 0.0963, "num_input_tokens_seen": 16975632, "step": 8815 }, { "epoch": 1.4388612447997389, "grad_norm": 0.05874129384756088, "learning_rate": 4.423902962823864e-05, "loss": 0.044, "num_input_tokens_seen": 16985536, "step": 8820 }, { "epoch": 1.4396769720205564, "grad_norm": 0.3935348391532898, "learning_rate": 4.423275896836733e-05, "loss": 0.11, "num_input_tokens_seen": 16994352, "step": 8825 }, { "epoch": 1.4404926992413736, "grad_norm": 7.379292011260986, "learning_rate": 4.42264853425629e-05, "loss": 0.1348, "num_input_tokens_seen": 17004960, "step": 8830 }, { "epoch": 1.4413084264621912, "grad_norm": 0.18844330310821533, "learning_rate": 4.4220208751792816e-05, "loss": 0.1486, "num_input_tokens_seen": 17015408, "step": 8835 }, { "epoch": 1.4421241536830083, "grad_norm": 2.3379509449005127, "learning_rate": 4.421392919702499e-05, "loss": 0.2395, "num_input_tokens_seen": 17024656, "step": 8840 }, { "epoch": 1.4429398809038259, "grad_norm": 7.300658702850342, "learning_rate": 4.4207646679227846e-05, "loss": 0.1692, "num_input_tokens_seen": 17033056, "step": 8845 }, { "epoch": 1.443755608124643, "grad_norm": 4.629790306091309, "learning_rate": 4.42013611993702e-05, "loss": 0.1047, "num_input_tokens_seen": 17042384, "step": 8850 }, { "epoch": 1.4445713353454606, "grad_norm": 0.7215453386306763, "learning_rate": 4.419507275842135e-05, "loss": 0.111, "num_input_tokens_seen": 17052384, "step": 8855 }, { "epoch": 1.445387062566278, "grad_norm": 0.37270814180374146, "learning_rate": 4.418878135735106e-05, "loss": 0.1329, "num_input_tokens_seen": 17059168, "step": 8860 }, { "epoch": 1.4462027897870953, "grad_norm": 2.588278293609619, "learning_rate": 4.418248699712955e-05, "loss": 0.1371, "num_input_tokens_seen": 17069408, "step": 8865 }, { "epoch": 1.4470185170079126, "grad_norm": 0.27338147163391113, "learning_rate": 4.417618967872748e-05, "loss": 0.0271, "num_input_tokens_seen": 17080016, "step": 8870 }, { "epoch": 1.44783424422873, "grad_norm": 1.6089565753936768, "learning_rate": 4.4169889403115985e-05, "loss": 0.0979, "num_input_tokens_seen": 17089968, "step": 8875 }, { "epoch": 1.4486499714495473, "grad_norm": 0.48560529947280884, "learning_rate": 4.4163586171266627e-05, "loss": 0.1198, "num_input_tokens_seen": 17099504, "step": 8880 }, { "epoch": 1.4494656986703647, "grad_norm": 2.2669918537139893, "learning_rate": 4.415727998415147e-05, "loss": 0.2152, "num_input_tokens_seen": 17109840, "step": 8885 }, { "epoch": 1.450281425891182, "grad_norm": 5.799392223358154, "learning_rate": 4.4150970842742985e-05, "loss": 0.0588, "num_input_tokens_seen": 17118640, "step": 8890 }, { "epoch": 1.4510971531119994, "grad_norm": 0.6157407164573669, "learning_rate": 4.4144658748014134e-05, "loss": 0.051, "num_input_tokens_seen": 17128608, "step": 8895 }, { "epoch": 1.4519128803328167, "grad_norm": 10.504014015197754, "learning_rate": 4.413834370093831e-05, "loss": 0.1571, "num_input_tokens_seen": 17138384, "step": 8900 }, { "epoch": 1.452728607553634, "grad_norm": 2.533989906311035, "learning_rate": 4.413202570248939e-05, "loss": 0.2274, "num_input_tokens_seen": 17148560, "step": 8905 }, { "epoch": 1.4535443347744514, "grad_norm": 7.887779235839844, "learning_rate": 4.412570475364167e-05, "loss": 0.2292, "num_input_tokens_seen": 17159232, "step": 8910 }, { "epoch": 1.4543600619952688, "grad_norm": 0.2704218029975891, "learning_rate": 4.411938085536994e-05, "loss": 0.1101, "num_input_tokens_seen": 17168640, "step": 8915 }, { "epoch": 1.4551757892160861, "grad_norm": 2.3745076656341553, "learning_rate": 4.41130540086494e-05, "loss": 0.128, "num_input_tokens_seen": 17178000, "step": 8920 }, { "epoch": 1.4559915164369035, "grad_norm": 4.223077297210693, "learning_rate": 4.4106724214455754e-05, "loss": 0.0545, "num_input_tokens_seen": 17188272, "step": 8925 }, { "epoch": 1.4568072436577209, "grad_norm": 0.37215620279312134, "learning_rate": 4.4100391473765115e-05, "loss": 0.0785, "num_input_tokens_seen": 17199008, "step": 8930 }, { "epoch": 1.4576229708785382, "grad_norm": 2.362814426422119, "learning_rate": 4.409405578755408e-05, "loss": 0.1046, "num_input_tokens_seen": 17209008, "step": 8935 }, { "epoch": 1.4584386980993556, "grad_norm": 2.4667088985443115, "learning_rate": 4.4087717156799705e-05, "loss": 0.0272, "num_input_tokens_seen": 17219760, "step": 8940 }, { "epoch": 1.459254425320173, "grad_norm": 0.22792421281337738, "learning_rate": 4.408137558247946e-05, "loss": 0.1466, "num_input_tokens_seen": 17229728, "step": 8945 }, { "epoch": 1.4600701525409903, "grad_norm": 7.806302070617676, "learning_rate": 4.4075031065571306e-05, "loss": 0.2147, "num_input_tokens_seen": 17239888, "step": 8950 }, { "epoch": 1.4608858797618076, "grad_norm": 2.4476189613342285, "learning_rate": 4.406868360705366e-05, "loss": 0.2188, "num_input_tokens_seen": 17249376, "step": 8955 }, { "epoch": 1.461701606982625, "grad_norm": 0.35633182525634766, "learning_rate": 4.406233320790536e-05, "loss": 0.2185, "num_input_tokens_seen": 17259072, "step": 8960 }, { "epoch": 1.4625173342034423, "grad_norm": 0.4987070858478546, "learning_rate": 4.4055979869105734e-05, "loss": 0.0555, "num_input_tokens_seen": 17269680, "step": 8965 }, { "epoch": 1.4633330614242597, "grad_norm": 0.14802850782871246, "learning_rate": 4.404962359163454e-05, "loss": 0.0918, "num_input_tokens_seen": 17278944, "step": 8970 }, { "epoch": 1.464148788645077, "grad_norm": 0.04853261262178421, "learning_rate": 4.404326437647199e-05, "loss": 0.0348, "num_input_tokens_seen": 17288384, "step": 8975 }, { "epoch": 1.4649645158658944, "grad_norm": 0.2809288799762726, "learning_rate": 4.403690222459877e-05, "loss": 0.0263, "num_input_tokens_seen": 17298176, "step": 8980 }, { "epoch": 1.4657802430867117, "grad_norm": 1.2732962369918823, "learning_rate": 4.4030537136995984e-05, "loss": 0.121, "num_input_tokens_seen": 17308416, "step": 8985 }, { "epoch": 1.466595970307529, "grad_norm": 2.252814769744873, "learning_rate": 4.402416911464523e-05, "loss": 0.2046, "num_input_tokens_seen": 17317664, "step": 8990 }, { "epoch": 1.4674116975283464, "grad_norm": 3.2629783153533936, "learning_rate": 4.4017798158528516e-05, "loss": 0.0431, "num_input_tokens_seen": 17328256, "step": 8995 }, { "epoch": 1.468227424749164, "grad_norm": 2.2272188663482666, "learning_rate": 4.401142426962834e-05, "loss": 0.1505, "num_input_tokens_seen": 17338800, "step": 9000 }, { "epoch": 1.468227424749164, "eval_loss": 0.1334640234708786, "eval_runtime": 80.4849, "eval_samples_per_second": 33.857, "eval_steps_per_second": 16.935, "num_input_tokens_seen": 17338800, "step": 9000 }, { "epoch": 1.4690431519699811, "grad_norm": 4.9417524337768555, "learning_rate": 4.400504744892763e-05, "loss": 0.1247, "num_input_tokens_seen": 17348048, "step": 9005 }, { "epoch": 1.4698588791907987, "grad_norm": 0.8917207717895508, "learning_rate": 4.399866769740975e-05, "loss": 0.0817, "num_input_tokens_seen": 17356416, "step": 9010 }, { "epoch": 1.4706746064116158, "grad_norm": 4.084373950958252, "learning_rate": 4.399228501605859e-05, "loss": 0.1488, "num_input_tokens_seen": 17364048, "step": 9015 }, { "epoch": 1.4714903336324334, "grad_norm": 0.36519163846969604, "learning_rate": 4.398589940585839e-05, "loss": 0.1536, "num_input_tokens_seen": 17374272, "step": 9020 }, { "epoch": 1.4723060608532506, "grad_norm": 2.931654214859009, "learning_rate": 4.3979510867793917e-05, "loss": 0.1052, "num_input_tokens_seen": 17383152, "step": 9025 }, { "epoch": 1.4731217880740681, "grad_norm": 5.805938720703125, "learning_rate": 4.3973119402850346e-05, "loss": 0.2641, "num_input_tokens_seen": 17391632, "step": 9030 }, { "epoch": 1.4739375152948853, "grad_norm": 4.3164238929748535, "learning_rate": 4.396672501201334e-05, "loss": 0.1711, "num_input_tokens_seen": 17400960, "step": 9035 }, { "epoch": 1.4747532425157028, "grad_norm": 5.159472942352295, "learning_rate": 4.396032769626899e-05, "loss": 0.1812, "num_input_tokens_seen": 17409584, "step": 9040 }, { "epoch": 1.4755689697365202, "grad_norm": 1.525762915611267, "learning_rate": 4.395392745660384e-05, "loss": 0.0723, "num_input_tokens_seen": 17419616, "step": 9045 }, { "epoch": 1.4763846969573375, "grad_norm": 2.122459650039673, "learning_rate": 4.394752429400488e-05, "loss": 0.1415, "num_input_tokens_seen": 17429856, "step": 9050 }, { "epoch": 1.477200424178155, "grad_norm": 4.296762466430664, "learning_rate": 4.394111820945957e-05, "loss": 0.0985, "num_input_tokens_seen": 17441168, "step": 9055 }, { "epoch": 1.4780161513989722, "grad_norm": 0.3781457245349884, "learning_rate": 4.393470920395579e-05, "loss": 0.1539, "num_input_tokens_seen": 17449712, "step": 9060 }, { "epoch": 1.4788318786197896, "grad_norm": 1.192503809928894, "learning_rate": 4.392829727848192e-05, "loss": 0.1548, "num_input_tokens_seen": 17459232, "step": 9065 }, { "epoch": 1.479647605840607, "grad_norm": 4.294162273406982, "learning_rate": 4.392188243402673e-05, "loss": 0.0875, "num_input_tokens_seen": 17469120, "step": 9070 }, { "epoch": 1.4804633330614243, "grad_norm": 0.8956590294837952, "learning_rate": 4.391546467157949e-05, "loss": 0.0684, "num_input_tokens_seen": 17479472, "step": 9075 }, { "epoch": 1.4812790602822417, "grad_norm": 0.5488312840461731, "learning_rate": 4.390904399212988e-05, "loss": 0.0544, "num_input_tokens_seen": 17490384, "step": 9080 }, { "epoch": 1.482094787503059, "grad_norm": 0.9046301245689392, "learning_rate": 4.390262039666807e-05, "loss": 0.0463, "num_input_tokens_seen": 17501168, "step": 9085 }, { "epoch": 1.4829105147238764, "grad_norm": 6.886214733123779, "learning_rate": 4.389619388618464e-05, "loss": 0.2169, "num_input_tokens_seen": 17510112, "step": 9090 }, { "epoch": 1.4837262419446937, "grad_norm": 5.432490825653076, "learning_rate": 4.3889764461670655e-05, "loss": 0.1051, "num_input_tokens_seen": 17520288, "step": 9095 }, { "epoch": 1.484541969165511, "grad_norm": 0.7910898923873901, "learning_rate": 4.38833321241176e-05, "loss": 0.0903, "num_input_tokens_seen": 17530704, "step": 9100 }, { "epoch": 1.4853576963863284, "grad_norm": 3.709172248840332, "learning_rate": 4.3876896874517434e-05, "loss": 0.155, "num_input_tokens_seen": 17540320, "step": 9105 }, { "epoch": 1.4861734236071458, "grad_norm": 0.34371986985206604, "learning_rate": 4.3870458713862554e-05, "loss": 0.1527, "num_input_tokens_seen": 17550240, "step": 9110 }, { "epoch": 1.4869891508279631, "grad_norm": 0.7240642309188843, "learning_rate": 4.386401764314579e-05, "loss": 0.0659, "num_input_tokens_seen": 17559200, "step": 9115 }, { "epoch": 1.4878048780487805, "grad_norm": 7.529331684112549, "learning_rate": 4.385757366336045e-05, "loss": 0.127, "num_input_tokens_seen": 17569824, "step": 9120 }, { "epoch": 1.4886206052695978, "grad_norm": 8.859451293945312, "learning_rate": 4.385112677550027e-05, "loss": 0.3851, "num_input_tokens_seen": 17580400, "step": 9125 }, { "epoch": 1.4894363324904152, "grad_norm": 2.3119072914123535, "learning_rate": 4.384467698055945e-05, "loss": 0.1062, "num_input_tokens_seen": 17590400, "step": 9130 }, { "epoch": 1.4902520597112325, "grad_norm": 2.63891339302063, "learning_rate": 4.383822427953261e-05, "loss": 0.087, "num_input_tokens_seen": 17600768, "step": 9135 }, { "epoch": 1.4910677869320499, "grad_norm": 2.0410423278808594, "learning_rate": 4.3831768673414864e-05, "loss": 0.1478, "num_input_tokens_seen": 17611568, "step": 9140 }, { "epoch": 1.4918835141528672, "grad_norm": 0.22274629771709442, "learning_rate": 4.382531016320173e-05, "loss": 0.0483, "num_input_tokens_seen": 17622560, "step": 9145 }, { "epoch": 1.4926992413736846, "grad_norm": 0.3118540048599243, "learning_rate": 4.3818848749889184e-05, "loss": 0.063, "num_input_tokens_seen": 17634256, "step": 9150 }, { "epoch": 1.493514968594502, "grad_norm": 0.418718159198761, "learning_rate": 4.381238443447368e-05, "loss": 0.1469, "num_input_tokens_seen": 17643168, "step": 9155 }, { "epoch": 1.4943306958153193, "grad_norm": 3.2628157138824463, "learning_rate": 4.380591721795208e-05, "loss": 0.0537, "num_input_tokens_seen": 17652976, "step": 9160 }, { "epoch": 1.4951464230361367, "grad_norm": 0.05468630790710449, "learning_rate": 4.3799447101321723e-05, "loss": 0.1742, "num_input_tokens_seen": 17662256, "step": 9165 }, { "epoch": 1.495962150256954, "grad_norm": 2.8541154861450195, "learning_rate": 4.379297408558036e-05, "loss": 0.1801, "num_input_tokens_seen": 17670848, "step": 9170 }, { "epoch": 1.4967778774777714, "grad_norm": 1.189784049987793, "learning_rate": 4.378649817172624e-05, "loss": 0.1709, "num_input_tokens_seen": 17680240, "step": 9175 }, { "epoch": 1.4975936046985887, "grad_norm": 3.2905900478363037, "learning_rate": 4.378001936075801e-05, "loss": 0.0532, "num_input_tokens_seen": 17691360, "step": 9180 }, { "epoch": 1.4984093319194063, "grad_norm": 0.3179306387901306, "learning_rate": 4.377353765367479e-05, "loss": 0.0828, "num_input_tokens_seen": 17701120, "step": 9185 }, { "epoch": 1.4992250591402234, "grad_norm": 6.387816905975342, "learning_rate": 4.376705305147614e-05, "loss": 0.177, "num_input_tokens_seen": 17710144, "step": 9190 }, { "epoch": 1.500040786361041, "grad_norm": 0.858077883720398, "learning_rate": 4.376056555516206e-05, "loss": 0.1184, "num_input_tokens_seen": 17719104, "step": 9195 }, { "epoch": 1.5008565135818581, "grad_norm": 0.49382835626602173, "learning_rate": 4.375407516573302e-05, "loss": 0.1536, "num_input_tokens_seen": 17729104, "step": 9200 }, { "epoch": 1.5008565135818581, "eval_loss": 0.13459455966949463, "eval_runtime": 80.5467, "eval_samples_per_second": 33.831, "eval_steps_per_second": 16.922, "num_input_tokens_seen": 17729104, "step": 9200 }, { "epoch": 1.5016722408026757, "grad_norm": 0.8150772452354431, "learning_rate": 4.3747581884189913e-05, "loss": 0.1654, "num_input_tokens_seen": 17738768, "step": 9205 }, { "epoch": 1.5024879680234928, "grad_norm": 4.798342704772949, "learning_rate": 4.374108571153408e-05, "loss": 0.0586, "num_input_tokens_seen": 17748864, "step": 9210 }, { "epoch": 1.5033036952443104, "grad_norm": 0.9389445185661316, "learning_rate": 4.3734586648767316e-05, "loss": 0.046, "num_input_tokens_seen": 17757872, "step": 9215 }, { "epoch": 1.5041194224651275, "grad_norm": 0.1893216222524643, "learning_rate": 4.372808469689186e-05, "loss": 0.065, "num_input_tokens_seen": 17768320, "step": 9220 }, { "epoch": 1.504935149685945, "grad_norm": 0.13591302931308746, "learning_rate": 4.372157985691039e-05, "loss": 0.0431, "num_input_tokens_seen": 17776992, "step": 9225 }, { "epoch": 1.5057508769067622, "grad_norm": 0.3763396143913269, "learning_rate": 4.371507212982603e-05, "loss": 0.0876, "num_input_tokens_seen": 17786864, "step": 9230 }, { "epoch": 1.5065666041275798, "grad_norm": 4.654919147491455, "learning_rate": 4.370856151664236e-05, "loss": 0.1892, "num_input_tokens_seen": 17797248, "step": 9235 }, { "epoch": 1.507382331348397, "grad_norm": 10.770853996276855, "learning_rate": 4.3702048018363404e-05, "loss": 0.0699, "num_input_tokens_seen": 17806480, "step": 9240 }, { "epoch": 1.5081980585692145, "grad_norm": 1.0427067279815674, "learning_rate": 4.369553163599362e-05, "loss": 0.1438, "num_input_tokens_seen": 17817376, "step": 9245 }, { "epoch": 1.5090137857900316, "grad_norm": 1.4841383695602417, "learning_rate": 4.3689012370537904e-05, "loss": 0.0859, "num_input_tokens_seen": 17826224, "step": 9250 }, { "epoch": 1.5098295130108492, "grad_norm": 7.068836212158203, "learning_rate": 4.368249022300164e-05, "loss": 0.0739, "num_input_tokens_seen": 17836528, "step": 9255 }, { "epoch": 1.5106452402316666, "grad_norm": 20.660354614257812, "learning_rate": 4.367596519439059e-05, "loss": 0.2231, "num_input_tokens_seen": 17845312, "step": 9260 }, { "epoch": 1.511460967452484, "grad_norm": 1.0413957834243774, "learning_rate": 4.366943728571101e-05, "loss": 0.1273, "num_input_tokens_seen": 17853360, "step": 9265 }, { "epoch": 1.5122766946733013, "grad_norm": 0.17163482308387756, "learning_rate": 4.366290649796959e-05, "loss": 0.1099, "num_input_tokens_seen": 17862336, "step": 9270 }, { "epoch": 1.5130924218941186, "grad_norm": 5.1213483810424805, "learning_rate": 4.3656372832173456e-05, "loss": 0.3006, "num_input_tokens_seen": 17872224, "step": 9275 }, { "epoch": 1.513908149114936, "grad_norm": 0.6682316064834595, "learning_rate": 4.364983628933017e-05, "loss": 0.0719, "num_input_tokens_seen": 17881872, "step": 9280 }, { "epoch": 1.5147238763357533, "grad_norm": 1.9261053800582886, "learning_rate": 4.364329687044777e-05, "loss": 0.2299, "num_input_tokens_seen": 17890944, "step": 9285 }, { "epoch": 1.5155396035565707, "grad_norm": 3.786494731903076, "learning_rate": 4.36367545765347e-05, "loss": 0.0993, "num_input_tokens_seen": 17899392, "step": 9290 }, { "epoch": 1.516355330777388, "grad_norm": 2.0166497230529785, "learning_rate": 4.363020940859988e-05, "loss": 0.3937, "num_input_tokens_seen": 17908432, "step": 9295 }, { "epoch": 1.5171710579982054, "grad_norm": 2.429173469543457, "learning_rate": 4.362366136765263e-05, "loss": 0.1153, "num_input_tokens_seen": 17918080, "step": 9300 }, { "epoch": 1.5179867852190227, "grad_norm": 1.2514550685882568, "learning_rate": 4.361711045470278e-05, "loss": 0.138, "num_input_tokens_seen": 17927280, "step": 9305 }, { "epoch": 1.51880251243984, "grad_norm": 0.4912535548210144, "learning_rate": 4.3610556670760524e-05, "loss": 0.1204, "num_input_tokens_seen": 17937456, "step": 9310 }, { "epoch": 1.5196182396606575, "grad_norm": 4.629091262817383, "learning_rate": 4.360400001683657e-05, "loss": 0.128, "num_input_tokens_seen": 17947440, "step": 9315 }, { "epoch": 1.5204339668814748, "grad_norm": 1.9218175411224365, "learning_rate": 4.3597440493942e-05, "loss": 0.0802, "num_input_tokens_seen": 17957056, "step": 9320 }, { "epoch": 1.5212496941022922, "grad_norm": 0.25697243213653564, "learning_rate": 4.3590878103088405e-05, "loss": 0.1087, "num_input_tokens_seen": 17966752, "step": 9325 }, { "epoch": 1.5220654213231095, "grad_norm": 0.3835064470767975, "learning_rate": 4.358431284528779e-05, "loss": 0.0604, "num_input_tokens_seen": 17975984, "step": 9330 }, { "epoch": 1.5228811485439269, "grad_norm": 3.293909788131714, "learning_rate": 4.357774472155257e-05, "loss": 0.1157, "num_input_tokens_seen": 17985296, "step": 9335 }, { "epoch": 1.5236968757647442, "grad_norm": 0.1880607306957245, "learning_rate": 4.3571173732895664e-05, "loss": 0.1343, "num_input_tokens_seen": 17994304, "step": 9340 }, { "epoch": 1.5245126029855616, "grad_norm": 0.5091652274131775, "learning_rate": 4.356459988033039e-05, "loss": 0.0533, "num_input_tokens_seen": 18002768, "step": 9345 }, { "epoch": 1.5253283302063791, "grad_norm": 0.3562512993812561, "learning_rate": 4.355802316487051e-05, "loss": 0.166, "num_input_tokens_seen": 18010816, "step": 9350 }, { "epoch": 1.5261440574271963, "grad_norm": 7.136851787567139, "learning_rate": 4.355144358753025e-05, "loss": 0.2233, "num_input_tokens_seen": 18020640, "step": 9355 }, { "epoch": 1.5269597846480139, "grad_norm": 2.835458517074585, "learning_rate": 4.354486114932425e-05, "loss": 0.1569, "num_input_tokens_seen": 18030720, "step": 9360 }, { "epoch": 1.527775511868831, "grad_norm": 0.3129119873046875, "learning_rate": 4.353827585126762e-05, "loss": 0.0395, "num_input_tokens_seen": 18040576, "step": 9365 }, { "epoch": 1.5285912390896486, "grad_norm": 3.3152015209198, "learning_rate": 4.353168769437588e-05, "loss": 0.1186, "num_input_tokens_seen": 18049552, "step": 9370 }, { "epoch": 1.5294069663104657, "grad_norm": 1.4803169965744019, "learning_rate": 4.3525096679665014e-05, "loss": 0.0945, "num_input_tokens_seen": 18059104, "step": 9375 }, { "epoch": 1.5302226935312833, "grad_norm": 0.7477015256881714, "learning_rate": 4.351850280815144e-05, "loss": 0.1083, "num_input_tokens_seen": 18068736, "step": 9380 }, { "epoch": 1.5310384207521004, "grad_norm": 1.3731838464736938, "learning_rate": 4.3511906080852014e-05, "loss": 0.0942, "num_input_tokens_seen": 18077904, "step": 9385 }, { "epoch": 1.531854147972918, "grad_norm": 6.087541103363037, "learning_rate": 4.350530649878404e-05, "loss": 0.1248, "num_input_tokens_seen": 18088144, "step": 9390 }, { "epoch": 1.532669875193735, "grad_norm": 0.6714817881584167, "learning_rate": 4.3498704062965246e-05, "loss": 0.1571, "num_input_tokens_seen": 18097536, "step": 9395 }, { "epoch": 1.5334856024145527, "grad_norm": 1.3414976596832275, "learning_rate": 4.3492098774413815e-05, "loss": 0.1391, "num_input_tokens_seen": 18107328, "step": 9400 }, { "epoch": 1.5334856024145527, "eval_loss": 0.15398284792900085, "eval_runtime": 80.5794, "eval_samples_per_second": 33.818, "eval_steps_per_second": 16.915, "num_input_tokens_seen": 18107328, "step": 9400 }, { "epoch": 1.5343013296353698, "grad_norm": 2.7538387775421143, "learning_rate": 4.3485490634148375e-05, "loss": 0.1986, "num_input_tokens_seen": 18117552, "step": 9405 }, { "epoch": 1.5351170568561874, "grad_norm": 0.2651204764842987, "learning_rate": 4.347887964318797e-05, "loss": 0.0903, "num_input_tokens_seen": 18126496, "step": 9410 }, { "epoch": 1.5359327840770045, "grad_norm": 2.1179120540618896, "learning_rate": 4.34722658025521e-05, "loss": 0.0631, "num_input_tokens_seen": 18136480, "step": 9415 }, { "epoch": 1.536748511297822, "grad_norm": 0.20079469680786133, "learning_rate": 4.346564911326071e-05, "loss": 0.0333, "num_input_tokens_seen": 18146976, "step": 9420 }, { "epoch": 1.5375642385186392, "grad_norm": 0.5086643695831299, "learning_rate": 4.345902957633418e-05, "loss": 0.0401, "num_input_tokens_seen": 18157120, "step": 9425 }, { "epoch": 1.5383799657394568, "grad_norm": 5.925291538238525, "learning_rate": 4.345240719279331e-05, "loss": 0.0654, "num_input_tokens_seen": 18166000, "step": 9430 }, { "epoch": 1.539195692960274, "grad_norm": 0.043157320469617844, "learning_rate": 4.3445781963659374e-05, "loss": 0.0794, "num_input_tokens_seen": 18176128, "step": 9435 }, { "epoch": 1.5400114201810915, "grad_norm": 3.7750179767608643, "learning_rate": 4.3439153889954045e-05, "loss": 0.1045, "num_input_tokens_seen": 18186416, "step": 9440 }, { "epoch": 1.5408271474019088, "grad_norm": 9.288150787353516, "learning_rate": 4.343252297269946e-05, "loss": 0.2015, "num_input_tokens_seen": 18196144, "step": 9445 }, { "epoch": 1.5416428746227262, "grad_norm": 6.133874893188477, "learning_rate": 4.342588921291821e-05, "loss": 0.2613, "num_input_tokens_seen": 18206720, "step": 9450 }, { "epoch": 1.5424586018435436, "grad_norm": 8.732186317443848, "learning_rate": 4.341925261163328e-05, "loss": 0.1021, "num_input_tokens_seen": 18215760, "step": 9455 }, { "epoch": 1.543274329064361, "grad_norm": 0.056656159460544586, "learning_rate": 4.341261316986813e-05, "loss": 0.0735, "num_input_tokens_seen": 18225664, "step": 9460 }, { "epoch": 1.5440900562851783, "grad_norm": 0.1354185789823532, "learning_rate": 4.340597088864664e-05, "loss": 0.0811, "num_input_tokens_seen": 18236224, "step": 9465 }, { "epoch": 1.5449057835059956, "grad_norm": 2.198082208633423, "learning_rate": 4.339932576899313e-05, "loss": 0.0938, "num_input_tokens_seen": 18246192, "step": 9470 }, { "epoch": 1.545721510726813, "grad_norm": 2.7004401683807373, "learning_rate": 4.3392677811932375e-05, "loss": 0.139, "num_input_tokens_seen": 18256368, "step": 9475 }, { "epoch": 1.5465372379476303, "grad_norm": 0.042576078325510025, "learning_rate": 4.338602701848956e-05, "loss": 0.129, "num_input_tokens_seen": 18265440, "step": 9480 }, { "epoch": 1.5473529651684477, "grad_norm": 4.4360456466674805, "learning_rate": 4.337937338969033e-05, "loss": 0.1468, "num_input_tokens_seen": 18274320, "step": 9485 }, { "epoch": 1.548168692389265, "grad_norm": 3.5660579204559326, "learning_rate": 4.337271692656075e-05, "loss": 0.258, "num_input_tokens_seen": 18285520, "step": 9490 }, { "epoch": 1.5489844196100824, "grad_norm": 2.613595962524414, "learning_rate": 4.336605763012733e-05, "loss": 0.225, "num_input_tokens_seen": 18295280, "step": 9495 }, { "epoch": 1.5498001468308997, "grad_norm": 7.081236839294434, "learning_rate": 4.3359395501417026e-05, "loss": 0.2211, "num_input_tokens_seen": 18306192, "step": 9500 }, { "epoch": 1.550615874051717, "grad_norm": 1.8533295392990112, "learning_rate": 4.335273054145722e-05, "loss": 0.1906, "num_input_tokens_seen": 18316944, "step": 9505 }, { "epoch": 1.5514316012725344, "grad_norm": 0.6695801019668579, "learning_rate": 4.334606275127572e-05, "loss": 0.0569, "num_input_tokens_seen": 18326992, "step": 9510 }, { "epoch": 1.5522473284933518, "grad_norm": 2.92034912109375, "learning_rate": 4.33393921319008e-05, "loss": 0.1361, "num_input_tokens_seen": 18336256, "step": 9515 }, { "epoch": 1.5530630557141691, "grad_norm": 1.509328007698059, "learning_rate": 4.3332718684361146e-05, "loss": 0.1583, "num_input_tokens_seen": 18346112, "step": 9520 }, { "epoch": 1.5538787829349865, "grad_norm": 3.9353067874908447, "learning_rate": 4.332604240968588e-05, "loss": 0.0598, "num_input_tokens_seen": 18355024, "step": 9525 }, { "epoch": 1.5546945101558038, "grad_norm": 3.4375388622283936, "learning_rate": 4.331936330890459e-05, "loss": 0.1038, "num_input_tokens_seen": 18364272, "step": 9530 }, { "epoch": 1.5555102373766214, "grad_norm": 3.2472546100616455, "learning_rate": 4.331268138304725e-05, "loss": 0.1005, "num_input_tokens_seen": 18374544, "step": 9535 }, { "epoch": 1.5563259645974385, "grad_norm": 6.143444538116455, "learning_rate": 4.330599663314431e-05, "loss": 0.0695, "num_input_tokens_seen": 18384768, "step": 9540 }, { "epoch": 1.5571416918182561, "grad_norm": 5.093058109283447, "learning_rate": 4.329930906022665e-05, "loss": 0.1019, "num_input_tokens_seen": 18394176, "step": 9545 }, { "epoch": 1.5579574190390733, "grad_norm": 0.16181442141532898, "learning_rate": 4.3292618665325564e-05, "loss": 0.0392, "num_input_tokens_seen": 18402000, "step": 9550 }, { "epoch": 1.5587731462598908, "grad_norm": 0.43624189496040344, "learning_rate": 4.3285925449472796e-05, "loss": 0.1078, "num_input_tokens_seen": 18412096, "step": 9555 }, { "epoch": 1.559588873480708, "grad_norm": 7.548092842102051, "learning_rate": 4.327922941370054e-05, "loss": 0.1525, "num_input_tokens_seen": 18421472, "step": 9560 }, { "epoch": 1.5604046007015255, "grad_norm": 0.4828577935695648, "learning_rate": 4.3272530559041384e-05, "loss": 0.1022, "num_input_tokens_seen": 18431680, "step": 9565 }, { "epoch": 1.5612203279223427, "grad_norm": 0.6210483908653259, "learning_rate": 4.32658288865284e-05, "loss": 0.0391, "num_input_tokens_seen": 18441616, "step": 9570 }, { "epoch": 1.5620360551431602, "grad_norm": 5.474872589111328, "learning_rate": 4.325912439719505e-05, "loss": 0.2426, "num_input_tokens_seen": 18451840, "step": 9575 }, { "epoch": 1.5628517823639774, "grad_norm": 0.1774660050868988, "learning_rate": 4.3252417092075266e-05, "loss": 0.2161, "num_input_tokens_seen": 18461296, "step": 9580 }, { "epoch": 1.563667509584795, "grad_norm": 4.475576400756836, "learning_rate": 4.3245706972203385e-05, "loss": 0.1656, "num_input_tokens_seen": 18471856, "step": 9585 }, { "epoch": 1.564483236805612, "grad_norm": 3.7996292114257812, "learning_rate": 4.323899403861421e-05, "loss": 0.0963, "num_input_tokens_seen": 18479456, "step": 9590 }, { "epoch": 1.5652989640264297, "grad_norm": 0.23759880661964417, "learning_rate": 4.3232278292342935e-05, "loss": 0.1367, "num_input_tokens_seen": 18488592, "step": 9595 }, { "epoch": 1.5661146912472468, "grad_norm": 9.551024436950684, "learning_rate": 4.322555973442524e-05, "loss": 0.044, "num_input_tokens_seen": 18497776, "step": 9600 }, { "epoch": 1.5661146912472468, "eval_loss": 0.14135172963142395, "eval_runtime": 80.7075, "eval_samples_per_second": 33.764, "eval_steps_per_second": 16.888, "num_input_tokens_seen": 18497776, "step": 9600 }, { "epoch": 1.5669304184680644, "grad_norm": 0.2285165786743164, "learning_rate": 4.3218838365897184e-05, "loss": 0.0505, "num_input_tokens_seen": 18506272, "step": 9605 }, { "epoch": 1.5677461456888815, "grad_norm": 3.7989368438720703, "learning_rate": 4.3212114187795306e-05, "loss": 0.165, "num_input_tokens_seen": 18516304, "step": 9610 }, { "epoch": 1.568561872909699, "grad_norm": 0.6597906351089478, "learning_rate": 4.320538720115656e-05, "loss": 0.1697, "num_input_tokens_seen": 18525168, "step": 9615 }, { "epoch": 1.5693776001305162, "grad_norm": 3.214118480682373, "learning_rate": 4.319865740701831e-05, "loss": 0.0149, "num_input_tokens_seen": 18534400, "step": 9620 }, { "epoch": 1.5701933273513338, "grad_norm": 5.970759391784668, "learning_rate": 4.3191924806418396e-05, "loss": 0.173, "num_input_tokens_seen": 18545248, "step": 9625 }, { "epoch": 1.5710090545721511, "grad_norm": 4.782864570617676, "learning_rate": 4.318518940039507e-05, "loss": 0.1162, "num_input_tokens_seen": 18554128, "step": 9630 }, { "epoch": 1.5718247817929685, "grad_norm": 2.0661871433258057, "learning_rate": 4.3178451189987e-05, "loss": 0.1923, "num_input_tokens_seen": 18564288, "step": 9635 }, { "epoch": 1.5726405090137858, "grad_norm": 2.168940782546997, "learning_rate": 4.3171710176233315e-05, "loss": 0.0712, "num_input_tokens_seen": 18573840, "step": 9640 }, { "epoch": 1.5734562362346032, "grad_norm": 5.297238826751709, "learning_rate": 4.316496636017355e-05, "loss": 0.1223, "num_input_tokens_seen": 18583488, "step": 9645 }, { "epoch": 1.5742719634554205, "grad_norm": 5.983246803283691, "learning_rate": 4.315821974284771e-05, "loss": 0.1297, "num_input_tokens_seen": 18593904, "step": 9650 }, { "epoch": 1.5750876906762379, "grad_norm": 11.608475685119629, "learning_rate": 4.315147032529619e-05, "loss": 0.13, "num_input_tokens_seen": 18602320, "step": 9655 }, { "epoch": 1.5759034178970552, "grad_norm": 3.6334517002105713, "learning_rate": 4.3144718108559845e-05, "loss": 0.0916, "num_input_tokens_seen": 18610432, "step": 9660 }, { "epoch": 1.5767191451178726, "grad_norm": 2.285623550415039, "learning_rate": 4.3137963093679945e-05, "loss": 0.272, "num_input_tokens_seen": 18620896, "step": 9665 }, { "epoch": 1.57753487233869, "grad_norm": 3.8401331901550293, "learning_rate": 4.31312052816982e-05, "loss": 0.1654, "num_input_tokens_seen": 18629376, "step": 9670 }, { "epoch": 1.5783505995595073, "grad_norm": 0.1881566047668457, "learning_rate": 4.312444467365675e-05, "loss": 0.097, "num_input_tokens_seen": 18638624, "step": 9675 }, { "epoch": 1.5791663267803246, "grad_norm": 0.16377289593219757, "learning_rate": 4.311768127059816e-05, "loss": 0.0413, "num_input_tokens_seen": 18648224, "step": 9680 }, { "epoch": 1.579982054001142, "grad_norm": 0.27914854884147644, "learning_rate": 4.3110915073565444e-05, "loss": 0.0333, "num_input_tokens_seen": 18658944, "step": 9685 }, { "epoch": 1.5807977812219594, "grad_norm": 8.026263236999512, "learning_rate": 4.310414608360203e-05, "loss": 0.0917, "num_input_tokens_seen": 18668528, "step": 9690 }, { "epoch": 1.5816135084427767, "grad_norm": 0.7181974649429321, "learning_rate": 4.309737430175177e-05, "loss": 0.1164, "num_input_tokens_seen": 18678512, "step": 9695 }, { "epoch": 1.582429235663594, "grad_norm": 8.551152229309082, "learning_rate": 4.309059972905897e-05, "loss": 0.0754, "num_input_tokens_seen": 18688960, "step": 9700 }, { "epoch": 1.5832449628844114, "grad_norm": 0.445828378200531, "learning_rate": 4.308382236656836e-05, "loss": 0.076, "num_input_tokens_seen": 18698992, "step": 9705 }, { "epoch": 1.5840606901052288, "grad_norm": 0.1675088107585907, "learning_rate": 4.307704221532507e-05, "loss": 0.2093, "num_input_tokens_seen": 18708528, "step": 9710 }, { "epoch": 1.5848764173260461, "grad_norm": 0.7673028111457825, "learning_rate": 4.307025927637471e-05, "loss": 0.1353, "num_input_tokens_seen": 18717680, "step": 9715 }, { "epoch": 1.5856921445468637, "grad_norm": 0.10994542390108109, "learning_rate": 4.306347355076328e-05, "loss": 0.0388, "num_input_tokens_seen": 18727760, "step": 9720 }, { "epoch": 1.5865078717676808, "grad_norm": 12.926371574401855, "learning_rate": 4.305668503953724e-05, "loss": 0.0977, "num_input_tokens_seen": 18736416, "step": 9725 }, { "epoch": 1.5873235989884984, "grad_norm": 3.17745041847229, "learning_rate": 4.3049893743743436e-05, "loss": 0.1085, "num_input_tokens_seen": 18745184, "step": 9730 }, { "epoch": 1.5881393262093155, "grad_norm": 0.1052270233631134, "learning_rate": 4.304309966442919e-05, "loss": 0.0484, "num_input_tokens_seen": 18754224, "step": 9735 }, { "epoch": 1.588955053430133, "grad_norm": 7.47006368637085, "learning_rate": 4.303630280264224e-05, "loss": 0.1015, "num_input_tokens_seen": 18765392, "step": 9740 }, { "epoch": 1.5897707806509502, "grad_norm": 5.632560729980469, "learning_rate": 4.302950315943074e-05, "loss": 0.1609, "num_input_tokens_seen": 18775792, "step": 9745 }, { "epoch": 1.5905865078717678, "grad_norm": 0.24667659401893616, "learning_rate": 4.3022700735843275e-05, "loss": 0.1693, "num_input_tokens_seen": 18784544, "step": 9750 }, { "epoch": 1.591402235092585, "grad_norm": 0.21593967080116272, "learning_rate": 4.301589553292887e-05, "loss": 0.1358, "num_input_tokens_seen": 18794000, "step": 9755 }, { "epoch": 1.5922179623134025, "grad_norm": 0.5875682234764099, "learning_rate": 4.300908755173697e-05, "loss": 0.1035, "num_input_tokens_seen": 18804528, "step": 9760 }, { "epoch": 1.5930336895342196, "grad_norm": 12.273174285888672, "learning_rate": 4.300227679331745e-05, "loss": 0.1824, "num_input_tokens_seen": 18813664, "step": 9765 }, { "epoch": 1.5938494167550372, "grad_norm": 4.171061038970947, "learning_rate": 4.299546325872063e-05, "loss": 0.1357, "num_input_tokens_seen": 18823104, "step": 9770 }, { "epoch": 1.5946651439758543, "grad_norm": 1.5945556163787842, "learning_rate": 4.2988646948997225e-05, "loss": 0.083, "num_input_tokens_seen": 18832592, "step": 9775 }, { "epoch": 1.595480871196672, "grad_norm": 2.004289388656616, "learning_rate": 4.29818278651984e-05, "loss": 0.1366, "num_input_tokens_seen": 18842640, "step": 9780 }, { "epoch": 1.596296598417489, "grad_norm": 1.8276574611663818, "learning_rate": 4.297500600837574e-05, "loss": 0.151, "num_input_tokens_seen": 18852016, "step": 9785 }, { "epoch": 1.5971123256383066, "grad_norm": 9.60038948059082, "learning_rate": 4.2968181379581276e-05, "loss": 0.061, "num_input_tokens_seen": 18862880, "step": 9790 }, { "epoch": 1.5979280528591238, "grad_norm": 0.9884555339813232, "learning_rate": 4.296135397986743e-05, "loss": 0.0275, "num_input_tokens_seen": 18871312, "step": 9795 }, { "epoch": 1.5987437800799413, "grad_norm": 3.9602010250091553, "learning_rate": 4.295452381028709e-05, "loss": 0.1508, "num_input_tokens_seen": 18881008, "step": 9800 }, { "epoch": 1.5987437800799413, "eval_loss": 0.14867442846298218, "eval_runtime": 80.434, "eval_samples_per_second": 33.879, "eval_steps_per_second": 16.946, "num_input_tokens_seen": 18881008, "step": 9800 }, { "epoch": 1.5995595073007585, "grad_norm": 4.365893363952637, "learning_rate": 4.294769087189354e-05, "loss": 0.1359, "num_input_tokens_seen": 18891936, "step": 9805 }, { "epoch": 1.600375234521576, "grad_norm": 0.10508125275373459, "learning_rate": 4.294085516574052e-05, "loss": 0.0957, "num_input_tokens_seen": 18901440, "step": 9810 }, { "epoch": 1.6011909617423934, "grad_norm": 0.17188389599323273, "learning_rate": 4.2934016692882176e-05, "loss": 0.0888, "num_input_tokens_seen": 18911696, "step": 9815 }, { "epoch": 1.6020066889632107, "grad_norm": 0.13540537655353546, "learning_rate": 4.292717545437308e-05, "loss": 0.1687, "num_input_tokens_seen": 18920416, "step": 9820 }, { "epoch": 1.602822416184028, "grad_norm": 3.0576422214508057, "learning_rate": 4.292033145126825e-05, "loss": 0.3199, "num_input_tokens_seen": 18930256, "step": 9825 }, { "epoch": 1.6036381434048455, "grad_norm": 7.0821709632873535, "learning_rate": 4.29134846846231e-05, "loss": 0.2082, "num_input_tokens_seen": 18940400, "step": 9830 }, { "epoch": 1.6044538706256628, "grad_norm": 5.004809379577637, "learning_rate": 4.29066351554935e-05, "loss": 0.0378, "num_input_tokens_seen": 18950032, "step": 9835 }, { "epoch": 1.6052695978464802, "grad_norm": 0.21340826153755188, "learning_rate": 4.289978286493574e-05, "loss": 0.0691, "num_input_tokens_seen": 18959056, "step": 9840 }, { "epoch": 1.6060853250672975, "grad_norm": 0.6314195394515991, "learning_rate": 4.28929278140065e-05, "loss": 0.0617, "num_input_tokens_seen": 18969296, "step": 9845 }, { "epoch": 1.6069010522881149, "grad_norm": 0.12582612037658691, "learning_rate": 4.288607000376295e-05, "loss": 0.208, "num_input_tokens_seen": 18979920, "step": 9850 }, { "epoch": 1.6077167795089322, "grad_norm": 3.9218995571136475, "learning_rate": 4.2879209435262624e-05, "loss": 0.1415, "num_input_tokens_seen": 18989424, "step": 9855 }, { "epoch": 1.6085325067297496, "grad_norm": 4.354196071624756, "learning_rate": 4.287234610956353e-05, "loss": 0.067, "num_input_tokens_seen": 18998976, "step": 9860 }, { "epoch": 1.609348233950567, "grad_norm": 3.6133666038513184, "learning_rate": 4.2865480027724056e-05, "loss": 0.0235, "num_input_tokens_seen": 19007632, "step": 9865 }, { "epoch": 1.6101639611713843, "grad_norm": 2.559189796447754, "learning_rate": 4.285861119080306e-05, "loss": 0.2162, "num_input_tokens_seen": 19017568, "step": 9870 }, { "epoch": 1.6109796883922016, "grad_norm": 0.858832836151123, "learning_rate": 4.2851739599859784e-05, "loss": 0.1254, "num_input_tokens_seen": 19027520, "step": 9875 }, { "epoch": 1.611795415613019, "grad_norm": 0.5379680395126343, "learning_rate": 4.2844865255953934e-05, "loss": 0.0284, "num_input_tokens_seen": 19035904, "step": 9880 }, { "epoch": 1.6126111428338363, "grad_norm": 11.899154663085938, "learning_rate": 4.2837988160145605e-05, "loss": 0.6, "num_input_tokens_seen": 19045360, "step": 9885 }, { "epoch": 1.6134268700546537, "grad_norm": 3.9541616439819336, "learning_rate": 4.2831108313495336e-05, "loss": 0.1317, "num_input_tokens_seen": 19056608, "step": 9890 }, { "epoch": 1.614242597275471, "grad_norm": 4.540041923522949, "learning_rate": 4.282422571706408e-05, "loss": 0.2066, "num_input_tokens_seen": 19066464, "step": 9895 }, { "epoch": 1.6150583244962884, "grad_norm": 1.589934229850769, "learning_rate": 4.281734037191323e-05, "loss": 0.1314, "num_input_tokens_seen": 19076016, "step": 9900 }, { "epoch": 1.615874051717106, "grad_norm": 0.6914893984794617, "learning_rate": 4.281045227910459e-05, "loss": 0.0358, "num_input_tokens_seen": 19084992, "step": 9905 }, { "epoch": 1.616689778937923, "grad_norm": 0.5330763459205627, "learning_rate": 4.280356143970038e-05, "loss": 0.0528, "num_input_tokens_seen": 19093712, "step": 9910 }, { "epoch": 1.6175055061587407, "grad_norm": 0.5381929278373718, "learning_rate": 4.279666785476327e-05, "loss": 0.1651, "num_input_tokens_seen": 19103888, "step": 9915 }, { "epoch": 1.6183212333795578, "grad_norm": 1.6208304166793823, "learning_rate": 4.2789771525356325e-05, "loss": 0.1018, "num_input_tokens_seen": 19112640, "step": 9920 }, { "epoch": 1.6191369606003754, "grad_norm": 0.6785579919815063, "learning_rate": 4.2782872452543056e-05, "loss": 0.1508, "num_input_tokens_seen": 19122304, "step": 9925 }, { "epoch": 1.6199526878211925, "grad_norm": 4.610354900360107, "learning_rate": 4.2775970637387376e-05, "loss": 0.1898, "num_input_tokens_seen": 19132032, "step": 9930 }, { "epoch": 1.62076841504201, "grad_norm": 4.133893013000488, "learning_rate": 4.276906608095363e-05, "loss": 0.128, "num_input_tokens_seen": 19142432, "step": 9935 }, { "epoch": 1.6215841422628272, "grad_norm": 0.058292001485824585, "learning_rate": 4.276215878430661e-05, "loss": 0.1022, "num_input_tokens_seen": 19151840, "step": 9940 }, { "epoch": 1.6223998694836448, "grad_norm": 0.31450793147087097, "learning_rate": 4.275524874851149e-05, "loss": 0.0459, "num_input_tokens_seen": 19161680, "step": 9945 }, { "epoch": 1.623215596704462, "grad_norm": 0.565985381603241, "learning_rate": 4.274833597463388e-05, "loss": 0.0853, "num_input_tokens_seen": 19170832, "step": 9950 }, { "epoch": 1.6240313239252795, "grad_norm": 7.825998783111572, "learning_rate": 4.2741420463739824e-05, "loss": 0.1639, "num_input_tokens_seen": 19180176, "step": 9955 }, { "epoch": 1.6248470511460966, "grad_norm": 0.32194119691848755, "learning_rate": 4.273450221689578e-05, "loss": 0.0611, "num_input_tokens_seen": 19188800, "step": 9960 }, { "epoch": 1.6256627783669142, "grad_norm": 7.364459991455078, "learning_rate": 4.272758123516863e-05, "loss": 0.2871, "num_input_tokens_seen": 19197552, "step": 9965 }, { "epoch": 1.6264785055877313, "grad_norm": 10.09674072265625, "learning_rate": 4.272065751962567e-05, "loss": 0.1938, "num_input_tokens_seen": 19207536, "step": 9970 }, { "epoch": 1.627294232808549, "grad_norm": 0.892166018486023, "learning_rate": 4.271373107133464e-05, "loss": 0.0329, "num_input_tokens_seen": 19217680, "step": 9975 }, { "epoch": 1.628109960029366, "grad_norm": 4.251893043518066, "learning_rate": 4.270680189136366e-05, "loss": 0.1904, "num_input_tokens_seen": 19227328, "step": 9980 }, { "epoch": 1.6289256872501836, "grad_norm": 0.18747548758983612, "learning_rate": 4.269986998078132e-05, "loss": 0.073, "num_input_tokens_seen": 19237088, "step": 9985 }, { "epoch": 1.6297414144710007, "grad_norm": 0.2329052835702896, "learning_rate": 4.2692935340656595e-05, "loss": 0.1845, "num_input_tokens_seen": 19246416, "step": 9990 }, { "epoch": 1.6305571416918183, "grad_norm": 3.580002546310425, "learning_rate": 4.26859979720589e-05, "loss": 0.1517, "num_input_tokens_seen": 19257344, "step": 9995 }, { "epoch": 1.6313728689126357, "grad_norm": 1.9567829370498657, "learning_rate": 4.267905787605806e-05, "loss": 0.1213, "num_input_tokens_seen": 19266960, "step": 10000 }, { "epoch": 1.6313728689126357, "eval_loss": 0.13758963346481323, "eval_runtime": 80.6495, "eval_samples_per_second": 33.788, "eval_steps_per_second": 16.9, "num_input_tokens_seen": 19266960, "step": 10000 }, { "epoch": 1.632188596133453, "grad_norm": 0.11498472839593887, "learning_rate": 4.267211505372433e-05, "loss": 0.0177, "num_input_tokens_seen": 19276304, "step": 10005 }, { "epoch": 1.6330043233542704, "grad_norm": 0.28367552161216736, "learning_rate": 4.266516950612837e-05, "loss": 0.1933, "num_input_tokens_seen": 19286672, "step": 10010 }, { "epoch": 1.6338200505750877, "grad_norm": 10.796890258789062, "learning_rate": 4.265822123434128e-05, "loss": 0.1222, "num_input_tokens_seen": 19296944, "step": 10015 }, { "epoch": 1.634635777795905, "grad_norm": 4.58135461807251, "learning_rate": 4.265127023943457e-05, "loss": 0.1615, "num_input_tokens_seen": 19307120, "step": 10020 }, { "epoch": 1.6354515050167224, "grad_norm": 0.34715795516967773, "learning_rate": 4.2644316522480176e-05, "loss": 0.0926, "num_input_tokens_seen": 19316384, "step": 10025 }, { "epoch": 1.6362672322375398, "grad_norm": 0.296067476272583, "learning_rate": 4.263736008455044e-05, "loss": 0.0379, "num_input_tokens_seen": 19325792, "step": 10030 }, { "epoch": 1.6370829594583571, "grad_norm": 0.11610766500234604, "learning_rate": 4.2630400926718125e-05, "loss": 0.0999, "num_input_tokens_seen": 19336448, "step": 10035 }, { "epoch": 1.6378986866791745, "grad_norm": 7.111971855163574, "learning_rate": 4.262343905005644e-05, "loss": 0.0885, "num_input_tokens_seen": 19345280, "step": 10040 }, { "epoch": 1.6387144138999918, "grad_norm": 4.998166084289551, "learning_rate": 4.261647445563897e-05, "loss": 0.0918, "num_input_tokens_seen": 19355248, "step": 10045 }, { "epoch": 1.6395301411208092, "grad_norm": 0.38634857535362244, "learning_rate": 4.260950714453976e-05, "loss": 0.0791, "num_input_tokens_seen": 19364480, "step": 10050 }, { "epoch": 1.6403458683416265, "grad_norm": 1.6005346775054932, "learning_rate": 4.2602537117833266e-05, "loss": 0.1351, "num_input_tokens_seen": 19374688, "step": 10055 }, { "epoch": 1.641161595562444, "grad_norm": 18.269020080566406, "learning_rate": 4.259556437659433e-05, "loss": 0.1469, "num_input_tokens_seen": 19384320, "step": 10060 }, { "epoch": 1.6419773227832613, "grad_norm": 3.5548739433288574, "learning_rate": 4.258858892189825e-05, "loss": 0.0661, "num_input_tokens_seen": 19393488, "step": 10065 }, { "epoch": 1.6427930500040786, "grad_norm": 0.46315205097198486, "learning_rate": 4.2581610754820725e-05, "loss": 0.15, "num_input_tokens_seen": 19402128, "step": 10070 }, { "epoch": 1.643608777224896, "grad_norm": 9.43237018585205, "learning_rate": 4.2574629876437876e-05, "loss": 0.2613, "num_input_tokens_seen": 19410896, "step": 10075 }, { "epoch": 1.6444245044457133, "grad_norm": 6.352436065673828, "learning_rate": 4.256764628782625e-05, "loss": 0.0605, "num_input_tokens_seen": 19421360, "step": 10080 }, { "epoch": 1.6452402316665307, "grad_norm": 0.37292197346687317, "learning_rate": 4.256065999006279e-05, "loss": 0.0506, "num_input_tokens_seen": 19431200, "step": 10085 }, { "epoch": 1.6460559588873482, "grad_norm": 0.2685343325138092, "learning_rate": 4.2553670984224885e-05, "loss": 0.1717, "num_input_tokens_seen": 19441264, "step": 10090 }, { "epoch": 1.6468716861081654, "grad_norm": 0.06452862173318863, "learning_rate": 4.254667927139032e-05, "loss": 0.0786, "num_input_tokens_seen": 19451040, "step": 10095 }, { "epoch": 1.647687413328983, "grad_norm": 0.0496949777007103, "learning_rate": 4.2539684852637295e-05, "loss": 0.0365, "num_input_tokens_seen": 19461232, "step": 10100 }, { "epoch": 1.6485031405498, "grad_norm": 0.12387213110923767, "learning_rate": 4.253268772904446e-05, "loss": 0.1762, "num_input_tokens_seen": 19472000, "step": 10105 }, { "epoch": 1.6493188677706176, "grad_norm": 6.672485828399658, "learning_rate": 4.252568790169085e-05, "loss": 0.2638, "num_input_tokens_seen": 19481776, "step": 10110 }, { "epoch": 1.6501345949914348, "grad_norm": 6.468101501464844, "learning_rate": 4.251868537165592e-05, "loss": 0.1475, "num_input_tokens_seen": 19492560, "step": 10115 }, { "epoch": 1.6509503222122524, "grad_norm": 1.8678858280181885, "learning_rate": 4.251168014001955e-05, "loss": 0.2499, "num_input_tokens_seen": 19501344, "step": 10120 }, { "epoch": 1.6517660494330695, "grad_norm": 8.891501426696777, "learning_rate": 4.250467220786204e-05, "loss": 0.1686, "num_input_tokens_seen": 19512048, "step": 10125 }, { "epoch": 1.652581776653887, "grad_norm": 0.3381754755973816, "learning_rate": 4.249766157626409e-05, "loss": 0.1739, "num_input_tokens_seen": 19521248, "step": 10130 }, { "epoch": 1.6533975038747042, "grad_norm": 0.38108542561531067, "learning_rate": 4.249064824630684e-05, "loss": 0.1336, "num_input_tokens_seen": 19531136, "step": 10135 }, { "epoch": 1.6542132310955218, "grad_norm": 4.138002395629883, "learning_rate": 4.248363221907183e-05, "loss": 0.0846, "num_input_tokens_seen": 19540816, "step": 10140 }, { "epoch": 1.655028958316339, "grad_norm": 0.8310005068778992, "learning_rate": 4.2476613495641026e-05, "loss": 0.077, "num_input_tokens_seen": 19550912, "step": 10145 }, { "epoch": 1.6558446855371565, "grad_norm": 7.556066513061523, "learning_rate": 4.246959207709679e-05, "loss": 0.0861, "num_input_tokens_seen": 19560352, "step": 10150 }, { "epoch": 1.6566604127579736, "grad_norm": 12.099373817443848, "learning_rate": 4.246256796452192e-05, "loss": 0.0711, "num_input_tokens_seen": 19569488, "step": 10155 }, { "epoch": 1.6574761399787912, "grad_norm": 1.4370285272598267, "learning_rate": 4.245554115899962e-05, "loss": 0.1543, "num_input_tokens_seen": 19579024, "step": 10160 }, { "epoch": 1.6582918671996083, "grad_norm": 0.6663340926170349, "learning_rate": 4.2448511661613514e-05, "loss": 0.1271, "num_input_tokens_seen": 19587984, "step": 10165 }, { "epoch": 1.6591075944204259, "grad_norm": 4.935194969177246, "learning_rate": 4.2441479473447635e-05, "loss": 0.2113, "num_input_tokens_seen": 19596608, "step": 10170 }, { "epoch": 1.659923321641243, "grad_norm": 2.7597479820251465, "learning_rate": 4.243444459558644e-05, "loss": 0.1666, "num_input_tokens_seen": 19606400, "step": 10175 }, { "epoch": 1.6607390488620606, "grad_norm": 0.19362221658229828, "learning_rate": 4.24274070291148e-05, "loss": 0.1727, "num_input_tokens_seen": 19614928, "step": 10180 }, { "epoch": 1.661554776082878, "grad_norm": 0.34819871187210083, "learning_rate": 4.242036677511798e-05, "loss": 0.105, "num_input_tokens_seen": 19623488, "step": 10185 }, { "epoch": 1.6623705033036953, "grad_norm": 0.1182885393500328, "learning_rate": 4.241332383468169e-05, "loss": 0.1634, "num_input_tokens_seen": 19631584, "step": 10190 }, { "epoch": 1.6631862305245126, "grad_norm": 0.45154157280921936, "learning_rate": 4.2406278208892034e-05, "loss": 0.0512, "num_input_tokens_seen": 19641056, "step": 10195 }, { "epoch": 1.66400195774533, "grad_norm": 0.19053253531455994, "learning_rate": 4.2399229898835536e-05, "loss": 0.1004, "num_input_tokens_seen": 19650480, "step": 10200 }, { "epoch": 1.66400195774533, "eval_loss": 0.1352052241563797, "eval_runtime": 80.5749, "eval_samples_per_second": 33.819, "eval_steps_per_second": 16.916, "num_input_tokens_seen": 19650480, "step": 10200 }, { "epoch": 1.6648176849661473, "grad_norm": 0.10079460591077805, "learning_rate": 4.239217890559914e-05, "loss": 0.0203, "num_input_tokens_seen": 19660592, "step": 10205 }, { "epoch": 1.6656334121869647, "grad_norm": 1.1563682556152344, "learning_rate": 4.238512523027019e-05, "loss": 0.1303, "num_input_tokens_seen": 19670240, "step": 10210 }, { "epoch": 1.666449139407782, "grad_norm": 0.06009712815284729, "learning_rate": 4.237806887393645e-05, "loss": 0.024, "num_input_tokens_seen": 19679808, "step": 10215 }, { "epoch": 1.6672648666285994, "grad_norm": 0.17409536242485046, "learning_rate": 4.237100983768611e-05, "loss": 0.1115, "num_input_tokens_seen": 19690432, "step": 10220 }, { "epoch": 1.6680805938494168, "grad_norm": 0.09282304346561432, "learning_rate": 4.2363948122607756e-05, "loss": 0.0155, "num_input_tokens_seen": 19699824, "step": 10225 }, { "epoch": 1.6688963210702341, "grad_norm": 4.095532417297363, "learning_rate": 4.235688372979039e-05, "loss": 0.1771, "num_input_tokens_seen": 19709808, "step": 10230 }, { "epoch": 1.6697120482910515, "grad_norm": 1.7622507810592651, "learning_rate": 4.234981666032343e-05, "loss": 0.1415, "num_input_tokens_seen": 19719232, "step": 10235 }, { "epoch": 1.6705277755118688, "grad_norm": 0.06213398650288582, "learning_rate": 4.2342746915296704e-05, "loss": 0.0699, "num_input_tokens_seen": 19729360, "step": 10240 }, { "epoch": 1.6713435027326862, "grad_norm": 0.06153102591633797, "learning_rate": 4.233567449580047e-05, "loss": 0.0802, "num_input_tokens_seen": 19740160, "step": 10245 }, { "epoch": 1.6721592299535035, "grad_norm": 4.004356384277344, "learning_rate": 4.232859940292537e-05, "loss": 0.2589, "num_input_tokens_seen": 19748304, "step": 10250 }, { "epoch": 1.6729749571743209, "grad_norm": 0.8101032376289368, "learning_rate": 4.232152163776248e-05, "loss": 0.1382, "num_input_tokens_seen": 19758464, "step": 10255 }, { "epoch": 1.6737906843951382, "grad_norm": 1.0505543947219849, "learning_rate": 4.231444120140328e-05, "loss": 0.1003, "num_input_tokens_seen": 19768992, "step": 10260 }, { "epoch": 1.6746064116159556, "grad_norm": 9.404332160949707, "learning_rate": 4.230735809493967e-05, "loss": 0.1583, "num_input_tokens_seen": 19778768, "step": 10265 }, { "epoch": 1.675422138836773, "grad_norm": 0.965438961982727, "learning_rate": 4.2300272319463926e-05, "loss": 0.0904, "num_input_tokens_seen": 19788768, "step": 10270 }, { "epoch": 1.6762378660575903, "grad_norm": 0.15216219425201416, "learning_rate": 4.2293183876068786e-05, "loss": 0.1042, "num_input_tokens_seen": 19798416, "step": 10275 }, { "epoch": 1.6770535932784076, "grad_norm": 6.283332824707031, "learning_rate": 4.228609276584737e-05, "loss": 0.2312, "num_input_tokens_seen": 19809136, "step": 10280 }, { "epoch": 1.6778693204992252, "grad_norm": 1.0849623680114746, "learning_rate": 4.227899898989323e-05, "loss": 0.0985, "num_input_tokens_seen": 19818304, "step": 10285 }, { "epoch": 1.6786850477200423, "grad_norm": 2.8403544425964355, "learning_rate": 4.2271902549300293e-05, "loss": 0.1598, "num_input_tokens_seen": 19827120, "step": 10290 }, { "epoch": 1.67950077494086, "grad_norm": 0.469628244638443, "learning_rate": 4.226480344516294e-05, "loss": 0.214, "num_input_tokens_seen": 19837520, "step": 10295 }, { "epoch": 1.680316502161677, "grad_norm": 0.16934889554977417, "learning_rate": 4.2257701678575925e-05, "loss": 0.1083, "num_input_tokens_seen": 19846960, "step": 10300 }, { "epoch": 1.6811322293824946, "grad_norm": 0.34704264998435974, "learning_rate": 4.225059725063444e-05, "loss": 0.1145, "num_input_tokens_seen": 19856432, "step": 10305 }, { "epoch": 1.6819479566033118, "grad_norm": 0.2562759220600128, "learning_rate": 4.2243490162434074e-05, "loss": 0.1203, "num_input_tokens_seen": 19866128, "step": 10310 }, { "epoch": 1.6827636838241293, "grad_norm": 0.9014117121696472, "learning_rate": 4.223638041507083e-05, "loss": 0.0678, "num_input_tokens_seen": 19875568, "step": 10315 }, { "epoch": 1.6835794110449465, "grad_norm": 0.16773146390914917, "learning_rate": 4.2229268009641124e-05, "loss": 0.1473, "num_input_tokens_seen": 19885600, "step": 10320 }, { "epoch": 1.684395138265764, "grad_norm": 9.00635814666748, "learning_rate": 4.222215294724177e-05, "loss": 0.1485, "num_input_tokens_seen": 19895008, "step": 10325 }, { "epoch": 1.6852108654865812, "grad_norm": 3.6319644451141357, "learning_rate": 4.2215035228970005e-05, "loss": 0.1029, "num_input_tokens_seen": 19903424, "step": 10330 }, { "epoch": 1.6860265927073987, "grad_norm": 1.8751260042190552, "learning_rate": 4.2207914855923464e-05, "loss": 0.1156, "num_input_tokens_seen": 19913504, "step": 10335 }, { "epoch": 1.6868423199282159, "grad_norm": 0.3598117530345917, "learning_rate": 4.220079182920021e-05, "loss": 0.1079, "num_input_tokens_seen": 19921632, "step": 10340 }, { "epoch": 1.6876580471490334, "grad_norm": 3.388775110244751, "learning_rate": 4.2193666149898705e-05, "loss": 0.1782, "num_input_tokens_seen": 19932256, "step": 10345 }, { "epoch": 1.6884737743698506, "grad_norm": 4.825318336486816, "learning_rate": 4.21865378191178e-05, "loss": 0.1358, "num_input_tokens_seen": 19942064, "step": 10350 }, { "epoch": 1.6892895015906682, "grad_norm": 3.8079583644866943, "learning_rate": 4.217940683795678e-05, "loss": 0.0912, "num_input_tokens_seen": 19952144, "step": 10355 }, { "epoch": 1.6901052288114853, "grad_norm": 2.7265563011169434, "learning_rate": 4.217227320751534e-05, "loss": 0.0899, "num_input_tokens_seen": 19962592, "step": 10360 }, { "epoch": 1.6909209560323029, "grad_norm": 7.264439582824707, "learning_rate": 4.216513692889358e-05, "loss": 0.1568, "num_input_tokens_seen": 19971520, "step": 10365 }, { "epoch": 1.6917366832531202, "grad_norm": 2.4248204231262207, "learning_rate": 4.215799800319199e-05, "loss": 0.1481, "num_input_tokens_seen": 19980784, "step": 10370 }, { "epoch": 1.6925524104739376, "grad_norm": 2.085984706878662, "learning_rate": 4.2150856431511485e-05, "loss": 0.1331, "num_input_tokens_seen": 19990528, "step": 10375 }, { "epoch": 1.693368137694755, "grad_norm": 2.0948433876037598, "learning_rate": 4.214371221495339e-05, "loss": 0.0629, "num_input_tokens_seen": 20000000, "step": 10380 }, { "epoch": 1.6941838649155723, "grad_norm": 0.4134425222873688, "learning_rate": 4.213656535461942e-05, "loss": 0.1277, "num_input_tokens_seen": 20010544, "step": 10385 }, { "epoch": 1.6949995921363896, "grad_norm": 2.7527053356170654, "learning_rate": 4.2129415851611734e-05, "loss": 0.1353, "num_input_tokens_seen": 20020704, "step": 10390 }, { "epoch": 1.695815319357207, "grad_norm": 0.11205022037029266, "learning_rate": 4.2122263707032855e-05, "loss": 0.1937, "num_input_tokens_seen": 20030672, "step": 10395 }, { "epoch": 1.6966310465780243, "grad_norm": 1.1434587240219116, "learning_rate": 4.211510892198574e-05, "loss": 0.2218, "num_input_tokens_seen": 20041120, "step": 10400 }, { "epoch": 1.6966310465780243, "eval_loss": 0.12628936767578125, "eval_runtime": 80.6972, "eval_samples_per_second": 33.768, "eval_steps_per_second": 16.89, "num_input_tokens_seen": 20041120, "step": 10400 }, { "epoch": 1.6974467737988417, "grad_norm": 1.8085997104644775, "learning_rate": 4.210795149757375e-05, "loss": 0.0488, "num_input_tokens_seen": 20049536, "step": 10405 }, { "epoch": 1.698262501019659, "grad_norm": 5.455844402313232, "learning_rate": 4.210079143490065e-05, "loss": 0.1024, "num_input_tokens_seen": 20058064, "step": 10410 }, { "epoch": 1.6990782282404764, "grad_norm": 0.1660221964120865, "learning_rate": 4.2093628735070604e-05, "loss": 0.0873, "num_input_tokens_seen": 20066880, "step": 10415 }, { "epoch": 1.6998939554612937, "grad_norm": 0.20517145097255707, "learning_rate": 4.208646339918819e-05, "loss": 0.052, "num_input_tokens_seen": 20076432, "step": 10420 }, { "epoch": 1.700709682682111, "grad_norm": 3.9389781951904297, "learning_rate": 4.2079295428358414e-05, "loss": 0.1591, "num_input_tokens_seen": 20085856, "step": 10425 }, { "epoch": 1.7015254099029284, "grad_norm": 0.17011268436908722, "learning_rate": 4.207212482368664e-05, "loss": 0.0559, "num_input_tokens_seen": 20096624, "step": 10430 }, { "epoch": 1.7023411371237458, "grad_norm": 2.03718900680542, "learning_rate": 4.206495158627867e-05, "loss": 0.205, "num_input_tokens_seen": 20106976, "step": 10435 }, { "epoch": 1.7031568643445631, "grad_norm": 10.46908187866211, "learning_rate": 4.205777571724073e-05, "loss": 0.0369, "num_input_tokens_seen": 20117088, "step": 10440 }, { "epoch": 1.7039725915653805, "grad_norm": 0.7341259121894836, "learning_rate": 4.20505972176794e-05, "loss": 0.0974, "num_input_tokens_seen": 20126688, "step": 10445 }, { "epoch": 1.7047883187861979, "grad_norm": 8.72164535522461, "learning_rate": 4.204341608870171e-05, "loss": 0.0757, "num_input_tokens_seen": 20137424, "step": 10450 }, { "epoch": 1.7056040460070152, "grad_norm": 0.653978168964386, "learning_rate": 4.203623233141508e-05, "loss": 0.0219, "num_input_tokens_seen": 20147392, "step": 10455 }, { "epoch": 1.7064197732278326, "grad_norm": 9.612174034118652, "learning_rate": 4.2029045946927334e-05, "loss": 0.2624, "num_input_tokens_seen": 20156176, "step": 10460 }, { "epoch": 1.70723550044865, "grad_norm": 0.20520594716072083, "learning_rate": 4.20218569363467e-05, "loss": 0.0088, "num_input_tokens_seen": 20164464, "step": 10465 }, { "epoch": 1.7080512276694675, "grad_norm": 8.213727951049805, "learning_rate": 4.2014665300781834e-05, "loss": 0.038, "num_input_tokens_seen": 20174128, "step": 10470 }, { "epoch": 1.7088669548902846, "grad_norm": 4.4290032386779785, "learning_rate": 4.200747104134174e-05, "loss": 0.1197, "num_input_tokens_seen": 20182496, "step": 10475 }, { "epoch": 1.7096826821111022, "grad_norm": 0.5042591094970703, "learning_rate": 4.200027415913588e-05, "loss": 0.3324, "num_input_tokens_seen": 20191440, "step": 10480 }, { "epoch": 1.7104984093319193, "grad_norm": 0.6732475757598877, "learning_rate": 4.1993074655274126e-05, "loss": 0.2245, "num_input_tokens_seen": 20200896, "step": 10485 }, { "epoch": 1.711314136552737, "grad_norm": 6.240912914276123, "learning_rate": 4.198587253086669e-05, "loss": 0.193, "num_input_tokens_seen": 20210688, "step": 10490 }, { "epoch": 1.712129863773554, "grad_norm": 0.30935218930244446, "learning_rate": 4.197866778702426e-05, "loss": 0.0665, "num_input_tokens_seen": 20220848, "step": 10495 }, { "epoch": 1.7129455909943716, "grad_norm": 0.4225887656211853, "learning_rate": 4.197146042485789e-05, "loss": 0.2039, "num_input_tokens_seen": 20230320, "step": 10500 }, { "epoch": 1.7137613182151887, "grad_norm": 2.7910754680633545, "learning_rate": 4.1964250445479046e-05, "loss": 0.1735, "num_input_tokens_seen": 20239264, "step": 10505 }, { "epoch": 1.7145770454360063, "grad_norm": 0.9507277607917786, "learning_rate": 4.19570378499996e-05, "loss": 0.0428, "num_input_tokens_seen": 20248768, "step": 10510 }, { "epoch": 1.7153927726568234, "grad_norm": 2.233187675476074, "learning_rate": 4.194982263953182e-05, "loss": 0.2205, "num_input_tokens_seen": 20258528, "step": 10515 }, { "epoch": 1.716208499877641, "grad_norm": 6.193792819976807, "learning_rate": 4.194260481518838e-05, "loss": 0.3621, "num_input_tokens_seen": 20267536, "step": 10520 }, { "epoch": 1.7170242270984581, "grad_norm": 2.8748865127563477, "learning_rate": 4.1935384378082366e-05, "loss": 0.1037, "num_input_tokens_seen": 20277248, "step": 10525 }, { "epoch": 1.7178399543192757, "grad_norm": 1.6761696338653564, "learning_rate": 4.1928161329327267e-05, "loss": 0.0676, "num_input_tokens_seen": 20287664, "step": 10530 }, { "epoch": 1.7186556815400928, "grad_norm": 0.3354313373565674, "learning_rate": 4.1920935670036945e-05, "loss": 0.1387, "num_input_tokens_seen": 20298432, "step": 10535 }, { "epoch": 1.7194714087609104, "grad_norm": 4.631555557250977, "learning_rate": 4.1913707401325705e-05, "loss": 0.2113, "num_input_tokens_seen": 20308320, "step": 10540 }, { "epoch": 1.7202871359817276, "grad_norm": 0.1589125692844391, "learning_rate": 4.1906476524308235e-05, "loss": 0.1124, "num_input_tokens_seen": 20318976, "step": 10545 }, { "epoch": 1.7211028632025451, "grad_norm": 0.7694568634033203, "learning_rate": 4.189924304009962e-05, "loss": 0.2259, "num_input_tokens_seen": 20328544, "step": 10550 }, { "epoch": 1.7219185904233625, "grad_norm": 6.472229957580566, "learning_rate": 4.189200694981537e-05, "loss": 0.0925, "num_input_tokens_seen": 20337552, "step": 10555 }, { "epoch": 1.7227343176441798, "grad_norm": 1.3003891706466675, "learning_rate": 4.188476825457136e-05, "loss": 0.0618, "num_input_tokens_seen": 20347520, "step": 10560 }, { "epoch": 1.7235500448649972, "grad_norm": 0.21543312072753906, "learning_rate": 4.18775269554839e-05, "loss": 0.0403, "num_input_tokens_seen": 20356368, "step": 10565 }, { "epoch": 1.7243657720858145, "grad_norm": 2.751504421234131, "learning_rate": 4.187028305366969e-05, "loss": 0.1487, "num_input_tokens_seen": 20363888, "step": 10570 }, { "epoch": 1.725181499306632, "grad_norm": 3.40366268157959, "learning_rate": 4.1863036550245824e-05, "loss": 0.1468, "num_input_tokens_seen": 20373456, "step": 10575 }, { "epoch": 1.7259972265274492, "grad_norm": 2.348135471343994, "learning_rate": 4.1855787446329806e-05, "loss": 0.2004, "num_input_tokens_seen": 20383600, "step": 10580 }, { "epoch": 1.7268129537482666, "grad_norm": 3.7838361263275146, "learning_rate": 4.184853574303955e-05, "loss": 0.1831, "num_input_tokens_seen": 20393424, "step": 10585 }, { "epoch": 1.727628680969084, "grad_norm": 1.207087755203247, "learning_rate": 4.184128144149334e-05, "loss": 0.0884, "num_input_tokens_seen": 20402624, "step": 10590 }, { "epoch": 1.7284444081899013, "grad_norm": 4.648318767547607, "learning_rate": 4.1834024542809896e-05, "loss": 0.1559, "num_input_tokens_seen": 20411760, "step": 10595 }, { "epoch": 1.7292601354107187, "grad_norm": 1.89371919631958, "learning_rate": 4.1826765048108315e-05, "loss": 0.1434, "num_input_tokens_seen": 20421120, "step": 10600 }, { "epoch": 1.7292601354107187, "eval_loss": 0.12110266089439392, "eval_runtime": 80.584, "eval_samples_per_second": 33.816, "eval_steps_per_second": 16.914, "num_input_tokens_seen": 20421120, "step": 10600 }, { "epoch": 1.730075862631536, "grad_norm": 0.2747395932674408, "learning_rate": 4.181950295850811e-05, "loss": 0.1667, "num_input_tokens_seen": 20430096, "step": 10605 }, { "epoch": 1.7308915898523534, "grad_norm": 2.9165024757385254, "learning_rate": 4.181223827512918e-05, "loss": 0.0684, "num_input_tokens_seen": 20440304, "step": 10610 }, { "epoch": 1.7317073170731707, "grad_norm": 0.586040735244751, "learning_rate": 4.180497099909183e-05, "loss": 0.0521, "num_input_tokens_seen": 20448496, "step": 10615 }, { "epoch": 1.732523044293988, "grad_norm": 5.325362682342529, "learning_rate": 4.179770113151677e-05, "loss": 0.1519, "num_input_tokens_seen": 20457104, "step": 10620 }, { "epoch": 1.7333387715148054, "grad_norm": 2.49556827545166, "learning_rate": 4.179042867352511e-05, "loss": 0.1341, "num_input_tokens_seen": 20467424, "step": 10625 }, { "epoch": 1.7341544987356228, "grad_norm": 0.7466742992401123, "learning_rate": 4.1783153626238334e-05, "loss": 0.1137, "num_input_tokens_seen": 20479136, "step": 10630 }, { "epoch": 1.7349702259564401, "grad_norm": 2.7693915367126465, "learning_rate": 4.177587599077836e-05, "loss": 0.0821, "num_input_tokens_seen": 20489888, "step": 10635 }, { "epoch": 1.7357859531772575, "grad_norm": 0.21983395516872406, "learning_rate": 4.1768595768267494e-05, "loss": 0.1308, "num_input_tokens_seen": 20499392, "step": 10640 }, { "epoch": 1.7366016803980748, "grad_norm": 6.947964668273926, "learning_rate": 4.176131295982843e-05, "loss": 0.0406, "num_input_tokens_seen": 20509232, "step": 10645 }, { "epoch": 1.7374174076188922, "grad_norm": 2.0944623947143555, "learning_rate": 4.1754027566584276e-05, "loss": 0.189, "num_input_tokens_seen": 20518112, "step": 10650 }, { "epoch": 1.7382331348397098, "grad_norm": 2.164567708969116, "learning_rate": 4.174673958965852e-05, "loss": 0.2266, "num_input_tokens_seen": 20527792, "step": 10655 }, { "epoch": 1.739048862060527, "grad_norm": 1.1035722494125366, "learning_rate": 4.173944903017507e-05, "loss": 0.0791, "num_input_tokens_seen": 20537824, "step": 10660 }, { "epoch": 1.7398645892813445, "grad_norm": 6.9140143394470215, "learning_rate": 4.173215588925822e-05, "loss": 0.1338, "num_input_tokens_seen": 20546432, "step": 10665 }, { "epoch": 1.7406803165021616, "grad_norm": 3.363210678100586, "learning_rate": 4.172486016803266e-05, "loss": 0.1178, "num_input_tokens_seen": 20555632, "step": 10670 }, { "epoch": 1.7414960437229792, "grad_norm": 2.135885238647461, "learning_rate": 4.171756186762349e-05, "loss": 0.1688, "num_input_tokens_seen": 20564864, "step": 10675 }, { "epoch": 1.7423117709437963, "grad_norm": 0.20662111043930054, "learning_rate": 4.171026098915619e-05, "loss": 0.2128, "num_input_tokens_seen": 20574448, "step": 10680 }, { "epoch": 1.7431274981646139, "grad_norm": 4.189009666442871, "learning_rate": 4.170295753375665e-05, "loss": 0.2137, "num_input_tokens_seen": 20583856, "step": 10685 }, { "epoch": 1.743943225385431, "grad_norm": 0.6670907735824585, "learning_rate": 4.169565150255117e-05, "loss": 0.1043, "num_input_tokens_seen": 20592784, "step": 10690 }, { "epoch": 1.7447589526062486, "grad_norm": 1.9768602848052979, "learning_rate": 4.16883428966664e-05, "loss": 0.1193, "num_input_tokens_seen": 20602400, "step": 10695 }, { "epoch": 1.7455746798270657, "grad_norm": 5.4526753425598145, "learning_rate": 4.168103171722944e-05, "loss": 0.1002, "num_input_tokens_seen": 20612784, "step": 10700 }, { "epoch": 1.7463904070478833, "grad_norm": 1.8590067625045776, "learning_rate": 4.167371796536777e-05, "loss": 0.0813, "num_input_tokens_seen": 20623280, "step": 10705 }, { "epoch": 1.7472061342687004, "grad_norm": 0.4057891070842743, "learning_rate": 4.166640164220924e-05, "loss": 0.1092, "num_input_tokens_seen": 20632592, "step": 10710 }, { "epoch": 1.748021861489518, "grad_norm": 2.5234391689300537, "learning_rate": 4.1659082748882144e-05, "loss": 0.0709, "num_input_tokens_seen": 20641616, "step": 10715 }, { "epoch": 1.7488375887103351, "grad_norm": 4.325383186340332, "learning_rate": 4.1651761286515135e-05, "loss": 0.1036, "num_input_tokens_seen": 20652048, "step": 10720 }, { "epoch": 1.7496533159311527, "grad_norm": 0.17282424867153168, "learning_rate": 4.164443725623728e-05, "loss": 0.12, "num_input_tokens_seen": 20660720, "step": 10725 }, { "epoch": 1.7504690431519698, "grad_norm": 2.6777408123016357, "learning_rate": 4.163711065917802e-05, "loss": 0.1616, "num_input_tokens_seen": 20669680, "step": 10730 }, { "epoch": 1.7512847703727874, "grad_norm": 0.491107702255249, "learning_rate": 4.1629781496467234e-05, "loss": 0.1492, "num_input_tokens_seen": 20678800, "step": 10735 }, { "epoch": 1.7521004975936045, "grad_norm": 0.4647221565246582, "learning_rate": 4.1622449769235164e-05, "loss": 0.0979, "num_input_tokens_seen": 20688688, "step": 10740 }, { "epoch": 1.752916224814422, "grad_norm": 0.22601905465126038, "learning_rate": 4.161511547861243e-05, "loss": 0.1057, "num_input_tokens_seen": 20698912, "step": 10745 }, { "epoch": 1.7537319520352395, "grad_norm": 2.7841575145721436, "learning_rate": 4.1607778625730104e-05, "loss": 0.0607, "num_input_tokens_seen": 20708416, "step": 10750 }, { "epoch": 1.7545476792560568, "grad_norm": 4.05216646194458, "learning_rate": 4.160043921171961e-05, "loss": 0.1189, "num_input_tokens_seen": 20717328, "step": 10755 }, { "epoch": 1.7553634064768742, "grad_norm": 8.22956657409668, "learning_rate": 4.159309723771276e-05, "loss": 0.1239, "num_input_tokens_seen": 20728272, "step": 10760 }, { "epoch": 1.7561791336976915, "grad_norm": 0.21885302662849426, "learning_rate": 4.158575270484181e-05, "loss": 0.0355, "num_input_tokens_seen": 20738928, "step": 10765 }, { "epoch": 1.7569948609185089, "grad_norm": 4.154351711273193, "learning_rate": 4.157840561423936e-05, "loss": 0.0784, "num_input_tokens_seen": 20750176, "step": 10770 }, { "epoch": 1.7578105881393262, "grad_norm": 1.768066644668579, "learning_rate": 4.1571055967038416e-05, "loss": 0.0167, "num_input_tokens_seen": 20759584, "step": 10775 }, { "epoch": 1.7586263153601436, "grad_norm": 0.08059516549110413, "learning_rate": 4.156370376437241e-05, "loss": 0.0281, "num_input_tokens_seen": 20768896, "step": 10780 }, { "epoch": 1.759442042580961, "grad_norm": 1.7599976062774658, "learning_rate": 4.155634900737513e-05, "loss": 0.1451, "num_input_tokens_seen": 20779232, "step": 10785 }, { "epoch": 1.7602577698017783, "grad_norm": 0.30177006125450134, "learning_rate": 4.1548991697180764e-05, "loss": 0.1469, "num_input_tokens_seen": 20788848, "step": 10790 }, { "epoch": 1.7610734970225956, "grad_norm": 4.018011569976807, "learning_rate": 4.1541631834923914e-05, "loss": 0.2624, "num_input_tokens_seen": 20799216, "step": 10795 }, { "epoch": 1.761889224243413, "grad_norm": 4.556617736816406, "learning_rate": 4.153426942173956e-05, "loss": 0.0432, "num_input_tokens_seen": 20808496, "step": 10800 }, { "epoch": 1.761889224243413, "eval_loss": 0.13979317247867584, "eval_runtime": 80.5862, "eval_samples_per_second": 33.815, "eval_steps_per_second": 16.914, "num_input_tokens_seen": 20808496, "step": 10800 }, { "epoch": 1.7627049514642303, "grad_norm": 5.850243091583252, "learning_rate": 4.152690445876308e-05, "loss": 0.1832, "num_input_tokens_seen": 20817600, "step": 10805 }, { "epoch": 1.7635206786850477, "grad_norm": 2.3955512046813965, "learning_rate": 4.1519536947130245e-05, "loss": 0.1104, "num_input_tokens_seen": 20826496, "step": 10810 }, { "epoch": 1.764336405905865, "grad_norm": 0.1696726679801941, "learning_rate": 4.151216688797722e-05, "loss": 0.1044, "num_input_tokens_seen": 20835952, "step": 10815 }, { "epoch": 1.7651521331266824, "grad_norm": 3.0345447063446045, "learning_rate": 4.150479428244054e-05, "loss": 0.2129, "num_input_tokens_seen": 20846800, "step": 10820 }, { "epoch": 1.7659678603474998, "grad_norm": 6.0092339515686035, "learning_rate": 4.1497419131657176e-05, "loss": 0.036, "num_input_tokens_seen": 20855968, "step": 10825 }, { "epoch": 1.766783587568317, "grad_norm": 0.30964693427085876, "learning_rate": 4.149004143676447e-05, "loss": 0.0398, "num_input_tokens_seen": 20867360, "step": 10830 }, { "epoch": 1.7675993147891345, "grad_norm": 0.11161283403635025, "learning_rate": 4.148266119890015e-05, "loss": 0.0559, "num_input_tokens_seen": 20875616, "step": 10835 }, { "epoch": 1.768415042009952, "grad_norm": 0.22224774956703186, "learning_rate": 4.1475278419202324e-05, "loss": 0.1667, "num_input_tokens_seen": 20883792, "step": 10840 }, { "epoch": 1.7692307692307692, "grad_norm": 0.12360375374555588, "learning_rate": 4.146789309880953e-05, "loss": 0.194, "num_input_tokens_seen": 20893920, "step": 10845 }, { "epoch": 1.7700464964515867, "grad_norm": 9.187466621398926, "learning_rate": 4.146050523886068e-05, "loss": 0.2724, "num_input_tokens_seen": 20902704, "step": 10850 }, { "epoch": 1.7708622236724039, "grad_norm": 0.42369645833969116, "learning_rate": 4.1453114840495055e-05, "loss": 0.1627, "num_input_tokens_seen": 20913008, "step": 10855 }, { "epoch": 1.7716779508932214, "grad_norm": 3.065220355987549, "learning_rate": 4.1445721904852364e-05, "loss": 0.0697, "num_input_tokens_seen": 20922896, "step": 10860 }, { "epoch": 1.7724936781140386, "grad_norm": 0.1057853251695633, "learning_rate": 4.143832643307269e-05, "loss": 0.0971, "num_input_tokens_seen": 20931664, "step": 10865 }, { "epoch": 1.7733094053348561, "grad_norm": 0.5570223927497864, "learning_rate": 4.1430928426296503e-05, "loss": 0.1222, "num_input_tokens_seen": 20941632, "step": 10870 }, { "epoch": 1.7741251325556733, "grad_norm": 13.21889591217041, "learning_rate": 4.142352788566466e-05, "loss": 0.1587, "num_input_tokens_seen": 20950928, "step": 10875 }, { "epoch": 1.7749408597764909, "grad_norm": 1.1949715614318848, "learning_rate": 4.1416124812318424e-05, "loss": 0.0641, "num_input_tokens_seen": 20962144, "step": 10880 }, { "epoch": 1.775756586997308, "grad_norm": 0.20544469356536865, "learning_rate": 4.1408719207399453e-05, "loss": 0.1877, "num_input_tokens_seen": 20972512, "step": 10885 }, { "epoch": 1.7765723142181256, "grad_norm": 0.6022437214851379, "learning_rate": 4.140131107204978e-05, "loss": 0.0911, "num_input_tokens_seen": 20981824, "step": 10890 }, { "epoch": 1.7773880414389427, "grad_norm": 6.690586566925049, "learning_rate": 4.139390040741182e-05, "loss": 0.0609, "num_input_tokens_seen": 20991472, "step": 10895 }, { "epoch": 1.7782037686597603, "grad_norm": 5.506676197052002, "learning_rate": 4.1386487214628396e-05, "loss": 0.052, "num_input_tokens_seen": 21001664, "step": 10900 }, { "epoch": 1.7790194958805774, "grad_norm": 0.7637980580329895, "learning_rate": 4.137907149484272e-05, "loss": 0.0821, "num_input_tokens_seen": 21010864, "step": 10905 }, { "epoch": 1.779835223101395, "grad_norm": 2.8917295932769775, "learning_rate": 4.137165324919839e-05, "loss": 0.1419, "num_input_tokens_seen": 21020480, "step": 10910 }, { "epoch": 1.780650950322212, "grad_norm": 3.9098589420318604, "learning_rate": 4.136423247883939e-05, "loss": 0.0859, "num_input_tokens_seen": 21029568, "step": 10915 }, { "epoch": 1.7814666775430297, "grad_norm": 0.18496155738830566, "learning_rate": 4.135680918491009e-05, "loss": 0.1389, "num_input_tokens_seen": 21039136, "step": 10920 }, { "epoch": 1.7822824047638468, "grad_norm": 9.867603302001953, "learning_rate": 4.1349383368555265e-05, "loss": 0.2551, "num_input_tokens_seen": 21048752, "step": 10925 }, { "epoch": 1.7830981319846644, "grad_norm": 0.20130310952663422, "learning_rate": 4.1341955030920065e-05, "loss": 0.0695, "num_input_tokens_seen": 21057872, "step": 10930 }, { "epoch": 1.7839138592054817, "grad_norm": 0.3983127474784851, "learning_rate": 4.1334524173150036e-05, "loss": 0.023, "num_input_tokens_seen": 21067264, "step": 10935 }, { "epoch": 1.784729586426299, "grad_norm": 8.680363655090332, "learning_rate": 4.13270907963911e-05, "loss": 0.0794, "num_input_tokens_seen": 21076304, "step": 10940 }, { "epoch": 1.7855453136471164, "grad_norm": 0.21083983778953552, "learning_rate": 4.131965490178959e-05, "loss": 0.109, "num_input_tokens_seen": 21085696, "step": 10945 }, { "epoch": 1.7863610408679338, "grad_norm": 8.364938735961914, "learning_rate": 4.131221649049222e-05, "loss": 0.1606, "num_input_tokens_seen": 21095952, "step": 10950 }, { "epoch": 1.7871767680887511, "grad_norm": 0.2619577944278717, "learning_rate": 4.130477556364606e-05, "loss": 0.065, "num_input_tokens_seen": 21105056, "step": 10955 }, { "epoch": 1.7879924953095685, "grad_norm": 0.26402056217193604, "learning_rate": 4.129733212239861e-05, "loss": 0.0147, "num_input_tokens_seen": 21113904, "step": 10960 }, { "epoch": 1.7888082225303858, "grad_norm": 7.386570453643799, "learning_rate": 4.128988616789774e-05, "loss": 0.1257, "num_input_tokens_seen": 21124224, "step": 10965 }, { "epoch": 1.7896239497512032, "grad_norm": 1.9153218269348145, "learning_rate": 4.1282437701291724e-05, "loss": 0.1306, "num_input_tokens_seen": 21134704, "step": 10970 }, { "epoch": 1.7904396769720206, "grad_norm": 0.38396134972572327, "learning_rate": 4.1274986723729184e-05, "loss": 0.2052, "num_input_tokens_seen": 21144864, "step": 10975 }, { "epoch": 1.791255404192838, "grad_norm": 17.33257484436035, "learning_rate": 4.126753323635917e-05, "loss": 0.1383, "num_input_tokens_seen": 21155840, "step": 10980 }, { "epoch": 1.7920711314136553, "grad_norm": 0.14817292988300323, "learning_rate": 4.12600772403311e-05, "loss": 0.107, "num_input_tokens_seen": 21166272, "step": 10985 }, { "epoch": 1.7928868586344726, "grad_norm": 0.08761875331401825, "learning_rate": 4.125261873679479e-05, "loss": 0.0466, "num_input_tokens_seen": 21176416, "step": 10990 }, { "epoch": 1.79370258585529, "grad_norm": 0.24825842678546906, "learning_rate": 4.124515772690042e-05, "loss": 0.1305, "num_input_tokens_seen": 21185680, "step": 10995 }, { "epoch": 1.7945183130761073, "grad_norm": 0.16904424130916595, "learning_rate": 4.123769421179858e-05, "loss": 0.2285, "num_input_tokens_seen": 21195024, "step": 11000 }, { "epoch": 1.7945183130761073, "eval_loss": 0.14586400985717773, "eval_runtime": 80.5718, "eval_samples_per_second": 33.821, "eval_steps_per_second": 16.917, "num_input_tokens_seen": 21195024, "step": 11000 }, { "epoch": 1.7953340402969247, "grad_norm": 0.23372121155261993, "learning_rate": 4.1230228192640236e-05, "loss": 0.1051, "num_input_tokens_seen": 21203376, "step": 11005 }, { "epoch": 1.796149767517742, "grad_norm": 0.6445044279098511, "learning_rate": 4.122275967057675e-05, "loss": 0.1287, "num_input_tokens_seen": 21212672, "step": 11010 }, { "epoch": 1.7969654947385594, "grad_norm": 1.8688048124313354, "learning_rate": 4.1215288646759846e-05, "loss": 0.1066, "num_input_tokens_seen": 21222656, "step": 11015 }, { "epoch": 1.7977812219593767, "grad_norm": 0.8660359978675842, "learning_rate": 4.120781512234166e-05, "loss": 0.0961, "num_input_tokens_seen": 21231456, "step": 11020 }, { "epoch": 1.7985969491801943, "grad_norm": 2.0001707077026367, "learning_rate": 4.120033909847471e-05, "loss": 0.1074, "num_input_tokens_seen": 21241568, "step": 11025 }, { "epoch": 1.7994126764010114, "grad_norm": 3.889380693435669, "learning_rate": 4.119286057631187e-05, "loss": 0.1007, "num_input_tokens_seen": 21251136, "step": 11030 }, { "epoch": 1.800228403621829, "grad_norm": 2.2264392375946045, "learning_rate": 4.118537955700646e-05, "loss": 0.0291, "num_input_tokens_seen": 21262320, "step": 11035 }, { "epoch": 1.8010441308426461, "grad_norm": 0.4033173620700836, "learning_rate": 4.11778960417121e-05, "loss": 0.0285, "num_input_tokens_seen": 21272256, "step": 11040 }, { "epoch": 1.8018598580634637, "grad_norm": 0.4700261354446411, "learning_rate": 4.117041003158288e-05, "loss": 0.0194, "num_input_tokens_seen": 21281600, "step": 11045 }, { "epoch": 1.8026755852842808, "grad_norm": 7.367134094238281, "learning_rate": 4.1162921527773215e-05, "loss": 0.0414, "num_input_tokens_seen": 21291072, "step": 11050 }, { "epoch": 1.8034913125050984, "grad_norm": 3.4688363075256348, "learning_rate": 4.115543053143794e-05, "loss": 0.1555, "num_input_tokens_seen": 21300624, "step": 11055 }, { "epoch": 1.8043070397259156, "grad_norm": 7.36353063583374, "learning_rate": 4.114793704373226e-05, "loss": 0.0485, "num_input_tokens_seen": 21310656, "step": 11060 }, { "epoch": 1.8051227669467331, "grad_norm": 0.05091405287384987, "learning_rate": 4.114044106581175e-05, "loss": 0.1177, "num_input_tokens_seen": 21319200, "step": 11065 }, { "epoch": 1.8059384941675503, "grad_norm": 0.07151032984256744, "learning_rate": 4.11329425988324e-05, "loss": 0.0286, "num_input_tokens_seen": 21328576, "step": 11070 }, { "epoch": 1.8067542213883678, "grad_norm": 6.751795768737793, "learning_rate": 4.112544164395056e-05, "loss": 0.2685, "num_input_tokens_seen": 21338176, "step": 11075 }, { "epoch": 1.807569948609185, "grad_norm": 0.25132042169570923, "learning_rate": 4.111793820232297e-05, "loss": 0.0934, "num_input_tokens_seen": 21347936, "step": 11080 }, { "epoch": 1.8083856758300025, "grad_norm": 0.31702470779418945, "learning_rate": 4.1110432275106767e-05, "loss": 0.1413, "num_input_tokens_seen": 21356672, "step": 11085 }, { "epoch": 1.8092014030508197, "grad_norm": 0.06302165985107422, "learning_rate": 4.110292386345944e-05, "loss": 0.0822, "num_input_tokens_seen": 21366288, "step": 11090 }, { "epoch": 1.8100171302716372, "grad_norm": 5.978841781616211, "learning_rate": 4.109541296853891e-05, "loss": 0.2335, "num_input_tokens_seen": 21375024, "step": 11095 }, { "epoch": 1.8108328574924544, "grad_norm": 0.6735524535179138, "learning_rate": 4.108789959150341e-05, "loss": 0.0993, "num_input_tokens_seen": 21385136, "step": 11100 }, { "epoch": 1.811648584713272, "grad_norm": 1.9631353616714478, "learning_rate": 4.108038373351163e-05, "loss": 0.1442, "num_input_tokens_seen": 21393024, "step": 11105 }, { "epoch": 1.812464311934089, "grad_norm": 0.3213370740413666, "learning_rate": 4.10728653957226e-05, "loss": 0.0902, "num_input_tokens_seen": 21401520, "step": 11110 }, { "epoch": 1.8132800391549067, "grad_norm": 0.17569850385189056, "learning_rate": 4.106534457929575e-05, "loss": 0.088, "num_input_tokens_seen": 21410544, "step": 11115 }, { "epoch": 1.814095766375724, "grad_norm": 0.41773009300231934, "learning_rate": 4.105782128539086e-05, "loss": 0.095, "num_input_tokens_seen": 21418656, "step": 11120 }, { "epoch": 1.8149114935965414, "grad_norm": 15.266284942626953, "learning_rate": 4.1050295515168144e-05, "loss": 0.0507, "num_input_tokens_seen": 21427584, "step": 11125 }, { "epoch": 1.8157272208173587, "grad_norm": 0.3473612368106842, "learning_rate": 4.1042767269788155e-05, "loss": 0.1394, "num_input_tokens_seen": 21436416, "step": 11130 }, { "epoch": 1.816542948038176, "grad_norm": 0.17828114330768585, "learning_rate": 4.103523655041185e-05, "loss": 0.1322, "num_input_tokens_seen": 21444896, "step": 11135 }, { "epoch": 1.8173586752589934, "grad_norm": 0.2808997333049774, "learning_rate": 4.102770335820055e-05, "loss": 0.0651, "num_input_tokens_seen": 21453712, "step": 11140 }, { "epoch": 1.8181744024798108, "grad_norm": 0.1166992336511612, "learning_rate": 4.1020167694315984e-05, "loss": 0.0251, "num_input_tokens_seen": 21462704, "step": 11145 }, { "epoch": 1.8189901297006281, "grad_norm": 1.3549641370773315, "learning_rate": 4.101262955992023e-05, "loss": 0.0842, "num_input_tokens_seen": 21471024, "step": 11150 }, { "epoch": 1.8198058569214455, "grad_norm": 7.269641399383545, "learning_rate": 4.100508895617578e-05, "loss": 0.0555, "num_input_tokens_seen": 21480352, "step": 11155 }, { "epoch": 1.8206215841422628, "grad_norm": 9.566405296325684, "learning_rate": 4.099754588424547e-05, "loss": 0.155, "num_input_tokens_seen": 21490352, "step": 11160 }, { "epoch": 1.8214373113630802, "grad_norm": 0.057025421410799026, "learning_rate": 4.0990000345292546e-05, "loss": 0.14, "num_input_tokens_seen": 21501216, "step": 11165 }, { "epoch": 1.8222530385838975, "grad_norm": 0.1841605305671692, "learning_rate": 4.098245234048064e-05, "loss": 0.3381, "num_input_tokens_seen": 21510704, "step": 11170 }, { "epoch": 1.8230687658047149, "grad_norm": 5.706592559814453, "learning_rate": 4.0974901870973726e-05, "loss": 0.2451, "num_input_tokens_seen": 21520704, "step": 11175 }, { "epoch": 1.8238844930255322, "grad_norm": 4.299702167510986, "learning_rate": 4.096734893793619e-05, "loss": 0.3588, "num_input_tokens_seen": 21530016, "step": 11180 }, { "epoch": 1.8247002202463496, "grad_norm": 3.3626580238342285, "learning_rate": 4.095979354253279e-05, "loss": 0.128, "num_input_tokens_seen": 21540032, "step": 11185 }, { "epoch": 1.825515947467167, "grad_norm": 0.378899484872818, "learning_rate": 4.0952235685928656e-05, "loss": 0.0568, "num_input_tokens_seen": 21550208, "step": 11190 }, { "epoch": 1.8263316746879843, "grad_norm": 0.9802643656730652, "learning_rate": 4.094467536928932e-05, "loss": 0.1102, "num_input_tokens_seen": 21560688, "step": 11195 }, { "epoch": 1.8271474019088016, "grad_norm": 2.1961312294006348, "learning_rate": 4.093711259378067e-05, "loss": 0.1986, "num_input_tokens_seen": 21570368, "step": 11200 }, { "epoch": 1.8271474019088016, "eval_loss": 0.130798801779747, "eval_runtime": 80.4603, "eval_samples_per_second": 33.868, "eval_steps_per_second": 16.94, "num_input_tokens_seen": 21570368, "step": 11200 }, { "epoch": 1.827963129129619, "grad_norm": 1.9079866409301758, "learning_rate": 4.092954736056897e-05, "loss": 0.0715, "num_input_tokens_seen": 21580208, "step": 11205 }, { "epoch": 1.8287788563504366, "grad_norm": 3.709385395050049, "learning_rate": 4.09219796708209e-05, "loss": 0.2056, "num_input_tokens_seen": 21589920, "step": 11210 }, { "epoch": 1.8295945835712537, "grad_norm": 0.1477595865726471, "learning_rate": 4.0914409525703464e-05, "loss": 0.0765, "num_input_tokens_seen": 21598256, "step": 11215 }, { "epoch": 1.8304103107920713, "grad_norm": 0.06902894377708435, "learning_rate": 4.090683692638408e-05, "loss": 0.0475, "num_input_tokens_seen": 21607392, "step": 11220 }, { "epoch": 1.8312260380128884, "grad_norm": 2.8300037384033203, "learning_rate": 4.089926187403056e-05, "loss": 0.0989, "num_input_tokens_seen": 21617104, "step": 11225 }, { "epoch": 1.832041765233706, "grad_norm": 0.24401098489761353, "learning_rate": 4.0891684369811044e-05, "loss": 0.0637, "num_input_tokens_seen": 21625824, "step": 11230 }, { "epoch": 1.8328574924545231, "grad_norm": 1.005397081375122, "learning_rate": 4.0884104414894107e-05, "loss": 0.134, "num_input_tokens_seen": 21636224, "step": 11235 }, { "epoch": 1.8336732196753407, "grad_norm": 8.105875015258789, "learning_rate": 4.087652201044864e-05, "loss": 0.0927, "num_input_tokens_seen": 21646432, "step": 11240 }, { "epoch": 1.8344889468961578, "grad_norm": 0.1575143188238144, "learning_rate": 4.086893715764397e-05, "loss": 0.056, "num_input_tokens_seen": 21656048, "step": 11245 }, { "epoch": 1.8353046741169754, "grad_norm": 11.04617977142334, "learning_rate": 4.086134985764977e-05, "loss": 0.0736, "num_input_tokens_seen": 21665872, "step": 11250 }, { "epoch": 1.8361204013377925, "grad_norm": 3.3746514320373535, "learning_rate": 4.0853760111636085e-05, "loss": 0.0925, "num_input_tokens_seen": 21675424, "step": 11255 }, { "epoch": 1.83693612855861, "grad_norm": 0.4221266806125641, "learning_rate": 4.084616792077337e-05, "loss": 0.1874, "num_input_tokens_seen": 21685264, "step": 11260 }, { "epoch": 1.8377518557794272, "grad_norm": 0.2831117808818817, "learning_rate": 4.083857328623243e-05, "loss": 0.0847, "num_input_tokens_seen": 21694384, "step": 11265 }, { "epoch": 1.8385675830002448, "grad_norm": 0.03337699919939041, "learning_rate": 4.083097620918444e-05, "loss": 0.1277, "num_input_tokens_seen": 21704224, "step": 11270 }, { "epoch": 1.839383310221062, "grad_norm": 0.20055800676345825, "learning_rate": 4.082337669080097e-05, "loss": 0.1118, "num_input_tokens_seen": 21713440, "step": 11275 }, { "epoch": 1.8401990374418795, "grad_norm": 0.06908325850963593, "learning_rate": 4.081577473225398e-05, "loss": 0.0841, "num_input_tokens_seen": 21723952, "step": 11280 }, { "epoch": 1.8410147646626966, "grad_norm": 4.216429710388184, "learning_rate": 4.080817033471577e-05, "loss": 0.1091, "num_input_tokens_seen": 21732128, "step": 11285 }, { "epoch": 1.8418304918835142, "grad_norm": 3.329230308532715, "learning_rate": 4.080056349935903e-05, "loss": 0.1538, "num_input_tokens_seen": 21741808, "step": 11290 }, { "epoch": 1.8426462191043314, "grad_norm": 0.11614969372749329, "learning_rate": 4.079295422735684e-05, "loss": 0.0616, "num_input_tokens_seen": 21750448, "step": 11295 }, { "epoch": 1.843461946325149, "grad_norm": 8.009305953979492, "learning_rate": 4.078534251988264e-05, "loss": 0.1477, "num_input_tokens_seen": 21760048, "step": 11300 }, { "epoch": 1.8442776735459663, "grad_norm": 0.0580173097550869, "learning_rate": 4.077772837811025e-05, "loss": 0.0349, "num_input_tokens_seen": 21768592, "step": 11305 }, { "epoch": 1.8450934007667836, "grad_norm": 2.251955509185791, "learning_rate": 4.0770111803213874e-05, "loss": 0.203, "num_input_tokens_seen": 21779328, "step": 11310 }, { "epoch": 1.845909127987601, "grad_norm": 0.10223402082920074, "learning_rate": 4.076249279636807e-05, "loss": 0.2463, "num_input_tokens_seen": 21788464, "step": 11315 }, { "epoch": 1.8467248552084183, "grad_norm": 5.613015174865723, "learning_rate": 4.075487135874781e-05, "loss": 0.1112, "num_input_tokens_seen": 21798416, "step": 11320 }, { "epoch": 1.8475405824292357, "grad_norm": 3.846006393432617, "learning_rate": 4.074724749152837e-05, "loss": 0.0442, "num_input_tokens_seen": 21808288, "step": 11325 }, { "epoch": 1.848356309650053, "grad_norm": 0.18212294578552246, "learning_rate": 4.07396211958855e-05, "loss": 0.1133, "num_input_tokens_seen": 21818080, "step": 11330 }, { "epoch": 1.8491720368708704, "grad_norm": 1.1422609090805054, "learning_rate": 4.073199247299523e-05, "loss": 0.1918, "num_input_tokens_seen": 21826880, "step": 11335 }, { "epoch": 1.8499877640916877, "grad_norm": 3.6077635288238525, "learning_rate": 4.072436132403403e-05, "loss": 0.1034, "num_input_tokens_seen": 21836688, "step": 11340 }, { "epoch": 1.850803491312505, "grad_norm": 5.3536810874938965, "learning_rate": 4.0716727750178704e-05, "loss": 0.089, "num_input_tokens_seen": 21845696, "step": 11345 }, { "epoch": 1.8516192185333225, "grad_norm": 3.630195379257202, "learning_rate": 4.0709091752606455e-05, "loss": 0.2052, "num_input_tokens_seen": 21855648, "step": 11350 }, { "epoch": 1.8524349457541398, "grad_norm": 0.13481667637825012, "learning_rate": 4.070145333249484e-05, "loss": 0.0926, "num_input_tokens_seen": 21865744, "step": 11355 }, { "epoch": 1.8532506729749572, "grad_norm": 0.21452920138835907, "learning_rate": 4.069381249102181e-05, "loss": 0.072, "num_input_tokens_seen": 21876096, "step": 11360 }, { "epoch": 1.8540664001957745, "grad_norm": 2.698383331298828, "learning_rate": 4.0686169229365665e-05, "loss": 0.0346, "num_input_tokens_seen": 21886144, "step": 11365 }, { "epoch": 1.8548821274165919, "grad_norm": 0.162701815366745, "learning_rate": 4.067852354870511e-05, "loss": 0.2254, "num_input_tokens_seen": 21895344, "step": 11370 }, { "epoch": 1.8556978546374092, "grad_norm": 0.14867816865444183, "learning_rate": 4.067087545021919e-05, "loss": 0.1279, "num_input_tokens_seen": 21904192, "step": 11375 }, { "epoch": 1.8565135818582266, "grad_norm": 5.593995094299316, "learning_rate": 4.066322493508734e-05, "loss": 0.1695, "num_input_tokens_seen": 21914080, "step": 11380 }, { "epoch": 1.857329309079044, "grad_norm": 2.2306013107299805, "learning_rate": 4.065557200448937e-05, "loss": 0.1299, "num_input_tokens_seen": 21922128, "step": 11385 }, { "epoch": 1.8581450362998613, "grad_norm": 0.26800885796546936, "learning_rate": 4.064791665960546e-05, "loss": 0.2197, "num_input_tokens_seen": 21932208, "step": 11390 }, { "epoch": 1.8589607635206788, "grad_norm": 1.6480181217193604, "learning_rate": 4.064025890161615e-05, "loss": 0.0516, "num_input_tokens_seen": 21940960, "step": 11395 }, { "epoch": 1.859776490741496, "grad_norm": 2.8191721439361572, "learning_rate": 4.0632598731702373e-05, "loss": 0.172, "num_input_tokens_seen": 21950896, "step": 11400 }, { "epoch": 1.859776490741496, "eval_loss": 0.12488897144794464, "eval_runtime": 80.5599, "eval_samples_per_second": 33.826, "eval_steps_per_second": 16.919, "num_input_tokens_seen": 21950896, "step": 11400 }, { "epoch": 1.8605922179623136, "grad_norm": 3.6550400257110596, "learning_rate": 4.0624936151045426e-05, "loss": 0.163, "num_input_tokens_seen": 21960224, "step": 11405 }, { "epoch": 1.8614079451831307, "grad_norm": 2.7587954998016357, "learning_rate": 4.061727116082696e-05, "loss": 0.1985, "num_input_tokens_seen": 21969552, "step": 11410 }, { "epoch": 1.8622236724039483, "grad_norm": 1.1887667179107666, "learning_rate": 4.060960376222903e-05, "loss": 0.0814, "num_input_tokens_seen": 21977712, "step": 11415 }, { "epoch": 1.8630393996247654, "grad_norm": 2.9697184562683105, "learning_rate": 4.0601933956434034e-05, "loss": 0.1588, "num_input_tokens_seen": 21985648, "step": 11420 }, { "epoch": 1.863855126845583, "grad_norm": 0.6164626479148865, "learning_rate": 4.059426174462476e-05, "loss": 0.0636, "num_input_tokens_seen": 21995696, "step": 11425 }, { "epoch": 1.8646708540664, "grad_norm": 3.8329453468322754, "learning_rate": 4.058658712798435e-05, "loss": 0.103, "num_input_tokens_seen": 22006240, "step": 11430 }, { "epoch": 1.8654865812872177, "grad_norm": 2.8336901664733887, "learning_rate": 4.0578910107696336e-05, "loss": 0.0726, "num_input_tokens_seen": 22015520, "step": 11435 }, { "epoch": 1.8663023085080348, "grad_norm": 0.22289137542247772, "learning_rate": 4.05712306849446e-05, "loss": 0.0477, "num_input_tokens_seen": 22024528, "step": 11440 }, { "epoch": 1.8671180357288524, "grad_norm": 4.887592792510986, "learning_rate": 4.0563548860913415e-05, "loss": 0.244, "num_input_tokens_seen": 22033696, "step": 11445 }, { "epoch": 1.8679337629496695, "grad_norm": 1.1708412170410156, "learning_rate": 4.0555864636787414e-05, "loss": 0.0617, "num_input_tokens_seen": 22043600, "step": 11450 }, { "epoch": 1.868749490170487, "grad_norm": 5.352920055389404, "learning_rate": 4.054817801375159e-05, "loss": 0.0956, "num_input_tokens_seen": 22053072, "step": 11455 }, { "epoch": 1.8695652173913042, "grad_norm": 0.027494946494698524, "learning_rate": 4.054048899299134e-05, "loss": 0.1962, "num_input_tokens_seen": 22062160, "step": 11460 }, { "epoch": 1.8703809446121218, "grad_norm": 0.14919185638427734, "learning_rate": 4.0532797575692385e-05, "loss": 0.1343, "num_input_tokens_seen": 22072000, "step": 11465 }, { "epoch": 1.871196671832939, "grad_norm": 2.6734848022460938, "learning_rate": 4.052510376304085e-05, "loss": 0.1983, "num_input_tokens_seen": 22081488, "step": 11470 }, { "epoch": 1.8720123990537565, "grad_norm": 0.5846447348594666, "learning_rate": 4.051740755622321e-05, "loss": 0.085, "num_input_tokens_seen": 22090576, "step": 11475 }, { "epoch": 1.8728281262745736, "grad_norm": 1.9874963760375977, "learning_rate": 4.050970895642632e-05, "loss": 0.1897, "num_input_tokens_seen": 22098624, "step": 11480 }, { "epoch": 1.8736438534953912, "grad_norm": 1.6476434469223022, "learning_rate": 4.050200796483741e-05, "loss": 0.0211, "num_input_tokens_seen": 22107776, "step": 11485 }, { "epoch": 1.8744595807162086, "grad_norm": 5.72823429107666, "learning_rate": 4.049430458264405e-05, "loss": 0.125, "num_input_tokens_seen": 22116208, "step": 11490 }, { "epoch": 1.875275307937026, "grad_norm": 0.09581055492162704, "learning_rate": 4.048659881103422e-05, "loss": 0.0643, "num_input_tokens_seen": 22126960, "step": 11495 }, { "epoch": 1.8760910351578433, "grad_norm": 2.7020301818847656, "learning_rate": 4.0478890651196235e-05, "loss": 0.167, "num_input_tokens_seen": 22137792, "step": 11500 }, { "epoch": 1.8769067623786606, "grad_norm": 3.919119358062744, "learning_rate": 4.047118010431879e-05, "loss": 0.1081, "num_input_tokens_seen": 22147648, "step": 11505 }, { "epoch": 1.877722489599478, "grad_norm": 4.190240859985352, "learning_rate": 4.046346717159094e-05, "loss": 0.0918, "num_input_tokens_seen": 22157792, "step": 11510 }, { "epoch": 1.8785382168202953, "grad_norm": 0.32639339566230774, "learning_rate": 4.045575185420214e-05, "loss": 0.0152, "num_input_tokens_seen": 22167424, "step": 11515 }, { "epoch": 1.8793539440411127, "grad_norm": 0.14199550449848175, "learning_rate": 4.0448034153342165e-05, "loss": 0.0752, "num_input_tokens_seen": 22176368, "step": 11520 }, { "epoch": 1.88016967126193, "grad_norm": 1.8425705432891846, "learning_rate": 4.0440314070201194e-05, "loss": 0.2685, "num_input_tokens_seen": 22185744, "step": 11525 }, { "epoch": 1.8809853984827474, "grad_norm": 0.19575892388820648, "learning_rate": 4.043259160596976e-05, "loss": 0.0701, "num_input_tokens_seen": 22194704, "step": 11530 }, { "epoch": 1.8818011257035647, "grad_norm": 3.5306479930877686, "learning_rate": 4.0424866761838767e-05, "loss": 0.18, "num_input_tokens_seen": 22205808, "step": 11535 }, { "epoch": 1.882616852924382, "grad_norm": 0.2799314260482788, "learning_rate": 4.041713953899948e-05, "loss": 0.1358, "num_input_tokens_seen": 22214880, "step": 11540 }, { "epoch": 1.8834325801451994, "grad_norm": 0.508095920085907, "learning_rate": 4.0409409938643515e-05, "loss": 0.1095, "num_input_tokens_seen": 22222976, "step": 11545 }, { "epoch": 1.8842483073660168, "grad_norm": 0.2156675010919571, "learning_rate": 4.0401677961962904e-05, "loss": 0.0309, "num_input_tokens_seen": 22232096, "step": 11550 }, { "epoch": 1.8850640345868341, "grad_norm": 1.4998451471328735, "learning_rate": 4.039394361015001e-05, "loss": 0.0822, "num_input_tokens_seen": 22243168, "step": 11555 }, { "epoch": 1.8858797618076515, "grad_norm": 2.5061652660369873, "learning_rate": 4.038620688439755e-05, "loss": 0.047, "num_input_tokens_seen": 22251648, "step": 11560 }, { "epoch": 1.8866954890284688, "grad_norm": 2.324549436569214, "learning_rate": 4.037846778589862e-05, "loss": 0.0674, "num_input_tokens_seen": 22262208, "step": 11565 }, { "epoch": 1.8875112162492862, "grad_norm": 4.182657241821289, "learning_rate": 4.0370726315846715e-05, "loss": 0.3466, "num_input_tokens_seen": 22273248, "step": 11570 }, { "epoch": 1.8883269434701035, "grad_norm": 0.17911554872989655, "learning_rate": 4.036298247543565e-05, "loss": 0.0865, "num_input_tokens_seen": 22283392, "step": 11575 }, { "epoch": 1.8891426706909211, "grad_norm": 4.515605449676514, "learning_rate": 4.035523626585962e-05, "loss": 0.2373, "num_input_tokens_seen": 22294096, "step": 11580 }, { "epoch": 1.8899583979117383, "grad_norm": 3.796912670135498, "learning_rate": 4.0347487688313194e-05, "loss": 0.1683, "num_input_tokens_seen": 22303424, "step": 11585 }, { "epoch": 1.8907741251325558, "grad_norm": 3.3204028606414795, "learning_rate": 4.0339736743991296e-05, "loss": 0.0386, "num_input_tokens_seen": 22313456, "step": 11590 }, { "epoch": 1.891589852353373, "grad_norm": 0.3914947509765625, "learning_rate": 4.0331983434089227e-05, "loss": 0.1035, "num_input_tokens_seen": 22322784, "step": 11595 }, { "epoch": 1.8924055795741905, "grad_norm": 7.3957839012146, "learning_rate": 4.032422775980264e-05, "loss": 0.1569, "num_input_tokens_seen": 22333376, "step": 11600 }, { "epoch": 1.8924055795741905, "eval_loss": 0.13688918948173523, "eval_runtime": 80.6054, "eval_samples_per_second": 33.807, "eval_steps_per_second": 16.91, "num_input_tokens_seen": 22333376, "step": 11600 }, { "epoch": 1.8932213067950077, "grad_norm": 0.10853725671768188, "learning_rate": 4.031646972232754e-05, "loss": 0.1454, "num_input_tokens_seen": 22342944, "step": 11605 }, { "epoch": 1.8940370340158252, "grad_norm": 0.27401086688041687, "learning_rate": 4.0308709322860344e-05, "loss": 0.1027, "num_input_tokens_seen": 22352128, "step": 11610 }, { "epoch": 1.8948527612366424, "grad_norm": 2.555706024169922, "learning_rate": 4.0300946562597784e-05, "loss": 0.1561, "num_input_tokens_seen": 22362688, "step": 11615 }, { "epoch": 1.89566848845746, "grad_norm": 0.2040976583957672, "learning_rate": 4.029318144273698e-05, "loss": 0.1281, "num_input_tokens_seen": 22370528, "step": 11620 }, { "epoch": 1.896484215678277, "grad_norm": 0.19735084474086761, "learning_rate": 4.0285413964475415e-05, "loss": 0.1501, "num_input_tokens_seen": 22380784, "step": 11625 }, { "epoch": 1.8972999428990946, "grad_norm": 0.29281073808670044, "learning_rate": 4.0277644129010927e-05, "loss": 0.1648, "num_input_tokens_seen": 22391200, "step": 11630 }, { "epoch": 1.8981156701199118, "grad_norm": 3.232017993927002, "learning_rate": 4.0269871937541724e-05, "loss": 0.169, "num_input_tokens_seen": 22400288, "step": 11635 }, { "epoch": 1.8989313973407294, "grad_norm": 2.4696648120880127, "learning_rate": 4.026209739126637e-05, "loss": 0.1305, "num_input_tokens_seen": 22409152, "step": 11640 }, { "epoch": 1.8997471245615465, "grad_norm": 0.3631843030452728, "learning_rate": 4.025432049138381e-05, "loss": 0.1486, "num_input_tokens_seen": 22418880, "step": 11645 }, { "epoch": 1.900562851782364, "grad_norm": 2.643916130065918, "learning_rate": 4.0246541239093325e-05, "loss": 0.1397, "num_input_tokens_seen": 22428048, "step": 11650 }, { "epoch": 1.9013785790031812, "grad_norm": 2.6205484867095947, "learning_rate": 4.023875963559459e-05, "loss": 0.1035, "num_input_tokens_seen": 22436944, "step": 11655 }, { "epoch": 1.9021943062239988, "grad_norm": 3.6807544231414795, "learning_rate": 4.023097568208761e-05, "loss": 0.1148, "num_input_tokens_seen": 22446960, "step": 11660 }, { "epoch": 1.903010033444816, "grad_norm": 1.4314723014831543, "learning_rate": 4.022318937977277e-05, "loss": 0.0934, "num_input_tokens_seen": 22457728, "step": 11665 }, { "epoch": 1.9038257606656335, "grad_norm": 0.3851330578327179, "learning_rate": 4.021540072985084e-05, "loss": 0.0977, "num_input_tokens_seen": 22467712, "step": 11670 }, { "epoch": 1.9046414878864508, "grad_norm": 0.2633568346500397, "learning_rate": 4.020760973352289e-05, "loss": 0.1934, "num_input_tokens_seen": 22477824, "step": 11675 }, { "epoch": 1.9054572151072682, "grad_norm": 0.1058354303240776, "learning_rate": 4.019981639199042e-05, "loss": 0.0492, "num_input_tokens_seen": 22487408, "step": 11680 }, { "epoch": 1.9062729423280855, "grad_norm": 0.08781848102807999, "learning_rate": 4.0192020706455245e-05, "loss": 0.0995, "num_input_tokens_seen": 22496560, "step": 11685 }, { "epoch": 1.9070886695489029, "grad_norm": 1.0071496963500977, "learning_rate": 4.018422267811956e-05, "loss": 0.0873, "num_input_tokens_seen": 22506608, "step": 11690 }, { "epoch": 1.9079043967697202, "grad_norm": 6.537502288818359, "learning_rate": 4.017642230818592e-05, "loss": 0.1287, "num_input_tokens_seen": 22515808, "step": 11695 }, { "epoch": 1.9087201239905376, "grad_norm": 0.3074709475040436, "learning_rate": 4.0168619597857246e-05, "loss": 0.1053, "num_input_tokens_seen": 22524544, "step": 11700 }, { "epoch": 1.909535851211355, "grad_norm": 1.6290664672851562, "learning_rate": 4.016081454833681e-05, "loss": 0.0744, "num_input_tokens_seen": 22534080, "step": 11705 }, { "epoch": 1.9103515784321723, "grad_norm": 1.588680624961853, "learning_rate": 4.0153007160828245e-05, "loss": 0.099, "num_input_tokens_seen": 22542800, "step": 11710 }, { "epoch": 1.9111673056529896, "grad_norm": 0.1031455397605896, "learning_rate": 4.0145197436535555e-05, "loss": 0.1392, "num_input_tokens_seen": 22553392, "step": 11715 }, { "epoch": 1.911983032873807, "grad_norm": 2.433926820755005, "learning_rate": 4.0137385376663095e-05, "loss": 0.075, "num_input_tokens_seen": 22561936, "step": 11720 }, { "epoch": 1.9127987600946244, "grad_norm": 5.85143518447876, "learning_rate": 4.012957098241558e-05, "loss": 0.1363, "num_input_tokens_seen": 22571312, "step": 11725 }, { "epoch": 1.9136144873154417, "grad_norm": 2.9286928176879883, "learning_rate": 4.0121754254998076e-05, "loss": 0.0631, "num_input_tokens_seen": 22580976, "step": 11730 }, { "epoch": 1.914430214536259, "grad_norm": 0.4974684417247772, "learning_rate": 4.011393519561606e-05, "loss": 0.1084, "num_input_tokens_seen": 22589152, "step": 11735 }, { "epoch": 1.9152459417570764, "grad_norm": 1.6110217571258545, "learning_rate": 4.010611380547529e-05, "loss": 0.1011, "num_input_tokens_seen": 22597968, "step": 11740 }, { "epoch": 1.9160616689778938, "grad_norm": 4.168816089630127, "learning_rate": 4.009829008578192e-05, "loss": 0.1118, "num_input_tokens_seen": 22607760, "step": 11745 }, { "epoch": 1.9168773961987111, "grad_norm": 0.09786936640739441, "learning_rate": 4.00904640377425e-05, "loss": 0.0663, "num_input_tokens_seen": 22616816, "step": 11750 }, { "epoch": 1.9176931234195285, "grad_norm": 0.20905940234661102, "learning_rate": 4.0082635662563886e-05, "loss": 0.0533, "num_input_tokens_seen": 22626192, "step": 11755 }, { "epoch": 1.9185088506403458, "grad_norm": 3.878082513809204, "learning_rate": 4.007480496145331e-05, "loss": 0.322, "num_input_tokens_seen": 22636832, "step": 11760 }, { "epoch": 1.9193245778611632, "grad_norm": 18.13751792907715, "learning_rate": 4.006697193561837e-05, "loss": 0.071, "num_input_tokens_seen": 22647776, "step": 11765 }, { "epoch": 1.9201403050819805, "grad_norm": 4.547831058502197, "learning_rate": 4.005913658626701e-05, "loss": 0.1343, "num_input_tokens_seen": 22656256, "step": 11770 }, { "epoch": 1.920956032302798, "grad_norm": 4.68699312210083, "learning_rate": 4.005129891460754e-05, "loss": 0.1678, "num_input_tokens_seen": 22666992, "step": 11775 }, { "epoch": 1.9217717595236152, "grad_norm": 3.3224291801452637, "learning_rate": 4.004345892184864e-05, "loss": 0.1674, "num_input_tokens_seen": 22674992, "step": 11780 }, { "epoch": 1.9225874867444328, "grad_norm": 3.8731939792633057, "learning_rate": 4.003561660919932e-05, "loss": 0.2303, "num_input_tokens_seen": 22685648, "step": 11785 }, { "epoch": 1.92340321396525, "grad_norm": 2.0127811431884766, "learning_rate": 4.002777197786897e-05, "loss": 0.0613, "num_input_tokens_seen": 22694304, "step": 11790 }, { "epoch": 1.9242189411860675, "grad_norm": 7.452116012573242, "learning_rate": 4.0019925029067326e-05, "loss": 0.1392, "num_input_tokens_seen": 22703808, "step": 11795 }, { "epoch": 1.9250346684068846, "grad_norm": 4.734189510345459, "learning_rate": 4.0012075764004495e-05, "loss": 0.0447, "num_input_tokens_seen": 22714512, "step": 11800 }, { "epoch": 1.9250346684068846, "eval_loss": 0.12658576667308807, "eval_runtime": 80.4729, "eval_samples_per_second": 33.862, "eval_steps_per_second": 16.937, "num_input_tokens_seen": 22714512, "step": 11800 }, { "epoch": 1.9258503956277022, "grad_norm": 11.834879875183105, "learning_rate": 4.000422418389094e-05, "loss": 0.1459, "num_input_tokens_seen": 22724432, "step": 11805 }, { "epoch": 1.9266661228485193, "grad_norm": 4.526362895965576, "learning_rate": 3.999637028993744e-05, "loss": 0.1346, "num_input_tokens_seen": 22734736, "step": 11810 }, { "epoch": 1.927481850069337, "grad_norm": 6.557865142822266, "learning_rate": 3.99885140833552e-05, "loss": 0.2249, "num_input_tokens_seen": 22745440, "step": 11815 }, { "epoch": 1.928297577290154, "grad_norm": 6.923935890197754, "learning_rate": 3.998065556535572e-05, "loss": 0.0285, "num_input_tokens_seen": 22753568, "step": 11820 }, { "epoch": 1.9291133045109716, "grad_norm": 0.4173142910003662, "learning_rate": 3.9972794737150895e-05, "loss": 0.1078, "num_input_tokens_seen": 22763680, "step": 11825 }, { "epoch": 1.9299290317317888, "grad_norm": 3.6448886394500732, "learning_rate": 3.996493159995297e-05, "loss": 0.199, "num_input_tokens_seen": 22773264, "step": 11830 }, { "epoch": 1.9307447589526063, "grad_norm": 5.8817925453186035, "learning_rate": 3.995706615497453e-05, "loss": 0.0878, "num_input_tokens_seen": 22782784, "step": 11835 }, { "epoch": 1.9315604861734235, "grad_norm": 4.0286173820495605, "learning_rate": 3.994919840342852e-05, "loss": 0.1731, "num_input_tokens_seen": 22792064, "step": 11840 }, { "epoch": 1.932376213394241, "grad_norm": 5.825239181518555, "learning_rate": 3.994132834652825e-05, "loss": 0.1423, "num_input_tokens_seen": 22800560, "step": 11845 }, { "epoch": 1.9331919406150582, "grad_norm": 0.41830480098724365, "learning_rate": 3.99334559854874e-05, "loss": 0.0517, "num_input_tokens_seen": 22811728, "step": 11850 }, { "epoch": 1.9340076678358757, "grad_norm": 3.221958875656128, "learning_rate": 3.9925581321519955e-05, "loss": 0.1561, "num_input_tokens_seen": 22819904, "step": 11855 }, { "epoch": 1.934823395056693, "grad_norm": 1.754294514656067, "learning_rate": 3.991770435584031e-05, "loss": 0.1673, "num_input_tokens_seen": 22828592, "step": 11860 }, { "epoch": 1.9356391222775104, "grad_norm": 4.989778518676758, "learning_rate": 3.990982508966319e-05, "loss": 0.1037, "num_input_tokens_seen": 22838464, "step": 11865 }, { "epoch": 1.9364548494983278, "grad_norm": 3.6758410930633545, "learning_rate": 3.990194352420367e-05, "loss": 0.0987, "num_input_tokens_seen": 22847488, "step": 11870 }, { "epoch": 1.9372705767191452, "grad_norm": 3.6896798610687256, "learning_rate": 3.9894059660677184e-05, "loss": 0.1024, "num_input_tokens_seen": 22856384, "step": 11875 }, { "epoch": 1.9380863039399625, "grad_norm": 4.410912990570068, "learning_rate": 3.9886173500299526e-05, "loss": 0.1456, "num_input_tokens_seen": 22866224, "step": 11880 }, { "epoch": 1.9389020311607799, "grad_norm": 0.3526060879230499, "learning_rate": 3.987828504428685e-05, "loss": 0.1909, "num_input_tokens_seen": 22876512, "step": 11885 }, { "epoch": 1.9397177583815972, "grad_norm": 1.6333200931549072, "learning_rate": 3.987039429385565e-05, "loss": 0.1144, "num_input_tokens_seen": 22886800, "step": 11890 }, { "epoch": 1.9405334856024146, "grad_norm": 2.9603564739227295, "learning_rate": 3.986250125022277e-05, "loss": 0.1038, "num_input_tokens_seen": 22895728, "step": 11895 }, { "epoch": 1.941349212823232, "grad_norm": 3.5871901512145996, "learning_rate": 3.985460591460544e-05, "loss": 0.1028, "num_input_tokens_seen": 22905424, "step": 11900 }, { "epoch": 1.9421649400440493, "grad_norm": 0.09535194933414459, "learning_rate": 3.984670828822118e-05, "loss": 0.0809, "num_input_tokens_seen": 22914016, "step": 11905 }, { "epoch": 1.9429806672648666, "grad_norm": 2.660440444946289, "learning_rate": 3.983880837228794e-05, "loss": 0.1635, "num_input_tokens_seen": 22924448, "step": 11910 }, { "epoch": 1.943796394485684, "grad_norm": 0.17508381605148315, "learning_rate": 3.983090616802396e-05, "loss": 0.0779, "num_input_tokens_seen": 22933152, "step": 11915 }, { "epoch": 1.9446121217065013, "grad_norm": 3.589540719985962, "learning_rate": 3.982300167664788e-05, "loss": 0.2566, "num_input_tokens_seen": 22942112, "step": 11920 }, { "epoch": 1.9454278489273187, "grad_norm": 0.15849487483501434, "learning_rate": 3.981509489937868e-05, "loss": 0.1974, "num_input_tokens_seen": 22951360, "step": 11925 }, { "epoch": 1.946243576148136, "grad_norm": 0.4031371772289276, "learning_rate": 3.9807185837435643e-05, "loss": 0.1519, "num_input_tokens_seen": 22961824, "step": 11930 }, { "epoch": 1.9470593033689534, "grad_norm": 1.7907766103744507, "learning_rate": 3.9799274492038484e-05, "loss": 0.131, "num_input_tokens_seen": 22970784, "step": 11935 }, { "epoch": 1.9478750305897707, "grad_norm": 0.4455960690975189, "learning_rate": 3.979136086440722e-05, "loss": 0.1022, "num_input_tokens_seen": 22980960, "step": 11940 }, { "epoch": 1.948690757810588, "grad_norm": 1.3312139511108398, "learning_rate": 3.9783444955762226e-05, "loss": 0.1027, "num_input_tokens_seen": 22991136, "step": 11945 }, { "epoch": 1.9495064850314054, "grad_norm": 0.22573238611221313, "learning_rate": 3.977552676732424e-05, "loss": 0.0607, "num_input_tokens_seen": 23000896, "step": 11950 }, { "epoch": 1.9503222122522228, "grad_norm": 4.315555572509766, "learning_rate": 3.976760630031435e-05, "loss": 0.1324, "num_input_tokens_seen": 23010528, "step": 11955 }, { "epoch": 1.9511379394730404, "grad_norm": 1.4965163469314575, "learning_rate": 3.975968355595398e-05, "loss": 0.1389, "num_input_tokens_seen": 23018800, "step": 11960 }, { "epoch": 1.9519536666938575, "grad_norm": 2.0514066219329834, "learning_rate": 3.9751758535464935e-05, "loss": 0.0626, "num_input_tokens_seen": 23029392, "step": 11965 }, { "epoch": 1.952769393914675, "grad_norm": 4.3735857009887695, "learning_rate": 3.9743831240069326e-05, "loss": 0.0931, "num_input_tokens_seen": 23040336, "step": 11970 }, { "epoch": 1.9535851211354922, "grad_norm": 0.09482330828905106, "learning_rate": 3.9735901670989675e-05, "loss": 0.1788, "num_input_tokens_seen": 23050672, "step": 11975 }, { "epoch": 1.9544008483563098, "grad_norm": 3.966949224472046, "learning_rate": 3.97279698294488e-05, "loss": 0.2206, "num_input_tokens_seen": 23059984, "step": 11980 }, { "epoch": 1.955216575577127, "grad_norm": 0.5008415579795837, "learning_rate": 3.9720035716669876e-05, "loss": 0.0651, "num_input_tokens_seen": 23069296, "step": 11985 }, { "epoch": 1.9560323027979445, "grad_norm": 0.46342211961746216, "learning_rate": 3.9712099333876474e-05, "loss": 0.0489, "num_input_tokens_seen": 23079568, "step": 11990 }, { "epoch": 1.9568480300187616, "grad_norm": 1.6921188831329346, "learning_rate": 3.9704160682292475e-05, "loss": 0.0948, "num_input_tokens_seen": 23089536, "step": 11995 }, { "epoch": 1.9576637572395792, "grad_norm": 4.163602352142334, "learning_rate": 3.9696219763142106e-05, "loss": 0.081, "num_input_tokens_seen": 23099888, "step": 12000 }, { "epoch": 1.9576637572395792, "eval_loss": 0.13295525312423706, "eval_runtime": 80.7224, "eval_samples_per_second": 33.758, "eval_steps_per_second": 16.885, "num_input_tokens_seen": 23099888, "step": 12000 }, { "epoch": 1.9584794844603963, "grad_norm": 5.406636714935303, "learning_rate": 3.968827657764997e-05, "loss": 0.0909, "num_input_tokens_seen": 23108880, "step": 12005 }, { "epoch": 1.959295211681214, "grad_norm": 3.175865888595581, "learning_rate": 3.9680331127041e-05, "loss": 0.1165, "num_input_tokens_seen": 23117376, "step": 12010 }, { "epoch": 1.960110938902031, "grad_norm": 0.530909538269043, "learning_rate": 3.9672383412540495e-05, "loss": 0.0923, "num_input_tokens_seen": 23126624, "step": 12015 }, { "epoch": 1.9609266661228486, "grad_norm": 0.3424642086029053, "learning_rate": 3.966443343537407e-05, "loss": 0.2495, "num_input_tokens_seen": 23135600, "step": 12020 }, { "epoch": 1.9617423933436657, "grad_norm": 6.260643482208252, "learning_rate": 3.965648119676772e-05, "loss": 0.1055, "num_input_tokens_seen": 23145824, "step": 12025 }, { "epoch": 1.9625581205644833, "grad_norm": 4.928952693939209, "learning_rate": 3.96485266979478e-05, "loss": 0.1431, "num_input_tokens_seen": 23155808, "step": 12030 }, { "epoch": 1.9633738477853004, "grad_norm": 4.205617427825928, "learning_rate": 3.9640569940140974e-05, "loss": 0.1876, "num_input_tokens_seen": 23166720, "step": 12035 }, { "epoch": 1.964189575006118, "grad_norm": 1.6521018743515015, "learning_rate": 3.963261092457428e-05, "loss": 0.1271, "num_input_tokens_seen": 23176672, "step": 12040 }, { "epoch": 1.9650053022269351, "grad_norm": 1.8620775938034058, "learning_rate": 3.962464965247509e-05, "loss": 0.1418, "num_input_tokens_seen": 23187248, "step": 12045 }, { "epoch": 1.9658210294477527, "grad_norm": 0.4816955327987671, "learning_rate": 3.9616686125071135e-05, "loss": 0.0295, "num_input_tokens_seen": 23196416, "step": 12050 }, { "epoch": 1.96663675666857, "grad_norm": 1.6753302812576294, "learning_rate": 3.9608720343590506e-05, "loss": 0.1338, "num_input_tokens_seen": 23204736, "step": 12055 }, { "epoch": 1.9674524838893874, "grad_norm": 1.2446664571762085, "learning_rate": 3.960075230926161e-05, "loss": 0.0553, "num_input_tokens_seen": 23215360, "step": 12060 }, { "epoch": 1.9682682111102048, "grad_norm": 0.16133242845535278, "learning_rate": 3.959278202331322e-05, "loss": 0.0282, "num_input_tokens_seen": 23224848, "step": 12065 }, { "epoch": 1.9690839383310221, "grad_norm": 2.764160394668579, "learning_rate": 3.958480948697446e-05, "loss": 0.1068, "num_input_tokens_seen": 23234848, "step": 12070 }, { "epoch": 1.9698996655518395, "grad_norm": 0.12456540763378143, "learning_rate": 3.95768347014748e-05, "loss": 0.1102, "num_input_tokens_seen": 23244800, "step": 12075 }, { "epoch": 1.9707153927726568, "grad_norm": 0.25526049733161926, "learning_rate": 3.956885766804404e-05, "loss": 0.0127, "num_input_tokens_seen": 23254336, "step": 12080 }, { "epoch": 1.9715311199934742, "grad_norm": 8.183381080627441, "learning_rate": 3.956087838791235e-05, "loss": 0.0432, "num_input_tokens_seen": 23264000, "step": 12085 }, { "epoch": 1.9723468472142915, "grad_norm": 3.235450029373169, "learning_rate": 3.955289686231022e-05, "loss": 0.1326, "num_input_tokens_seen": 23273456, "step": 12090 }, { "epoch": 1.973162574435109, "grad_norm": 6.53942346572876, "learning_rate": 3.9544913092468504e-05, "loss": 0.1302, "num_input_tokens_seen": 23282624, "step": 12095 }, { "epoch": 1.9739783016559262, "grad_norm": 0.1640028953552246, "learning_rate": 3.9536927079618425e-05, "loss": 0.0454, "num_input_tokens_seen": 23292640, "step": 12100 }, { "epoch": 1.9747940288767436, "grad_norm": 3.620941638946533, "learning_rate": 3.9528938824991494e-05, "loss": 0.1884, "num_input_tokens_seen": 23301168, "step": 12105 }, { "epoch": 1.975609756097561, "grad_norm": 0.7770626544952393, "learning_rate": 3.952094832981962e-05, "loss": 0.064, "num_input_tokens_seen": 23311648, "step": 12110 }, { "epoch": 1.9764254833183783, "grad_norm": 0.3896181583404541, "learning_rate": 3.951295559533503e-05, "loss": 0.1217, "num_input_tokens_seen": 23321408, "step": 12115 }, { "epoch": 1.9772412105391957, "grad_norm": 3.109130382537842, "learning_rate": 3.95049606227703e-05, "loss": 0.2666, "num_input_tokens_seen": 23330320, "step": 12120 }, { "epoch": 1.978056937760013, "grad_norm": 6.355167865753174, "learning_rate": 3.949696341335838e-05, "loss": 0.2908, "num_input_tokens_seen": 23340160, "step": 12125 }, { "epoch": 1.9788726649808304, "grad_norm": 1.2438938617706299, "learning_rate": 3.9488963968332503e-05, "loss": 0.0664, "num_input_tokens_seen": 23349776, "step": 12130 }, { "epoch": 1.9796883922016477, "grad_norm": 0.1928200125694275, "learning_rate": 3.948096228892631e-05, "loss": 0.1394, "num_input_tokens_seen": 23360144, "step": 12135 }, { "epoch": 1.980504119422465, "grad_norm": 0.3145080804824829, "learning_rate": 3.947295837637375e-05, "loss": 0.1058, "num_input_tokens_seen": 23370128, "step": 12140 }, { "epoch": 1.9813198466432826, "grad_norm": 0.6509871482849121, "learning_rate": 3.9464952231909135e-05, "loss": 0.0586, "num_input_tokens_seen": 23379584, "step": 12145 }, { "epoch": 1.9821355738640998, "grad_norm": 0.7111116647720337, "learning_rate": 3.945694385676711e-05, "loss": 0.0716, "num_input_tokens_seen": 23388032, "step": 12150 }, { "epoch": 1.9829513010849174, "grad_norm": 4.863717079162598, "learning_rate": 3.944893325218265e-05, "loss": 0.0747, "num_input_tokens_seen": 23396960, "step": 12155 }, { "epoch": 1.9837670283057345, "grad_norm": 0.633622407913208, "learning_rate": 3.944092041939112e-05, "loss": 0.1213, "num_input_tokens_seen": 23406272, "step": 12160 }, { "epoch": 1.984582755526552, "grad_norm": 0.6268385052680969, "learning_rate": 3.943290535962818e-05, "loss": 0.1572, "num_input_tokens_seen": 23416512, "step": 12165 }, { "epoch": 1.9853984827473692, "grad_norm": 3.999596118927002, "learning_rate": 3.942488807412985e-05, "loss": 0.1441, "num_input_tokens_seen": 23425632, "step": 12170 }, { "epoch": 1.9862142099681868, "grad_norm": 6.288873672485352, "learning_rate": 3.941686856413251e-05, "loss": 0.0955, "num_input_tokens_seen": 23435504, "step": 12175 }, { "epoch": 1.987029937189004, "grad_norm": 4.092971324920654, "learning_rate": 3.9408846830872874e-05, "loss": 0.1278, "num_input_tokens_seen": 23444336, "step": 12180 }, { "epoch": 1.9878456644098215, "grad_norm": 8.371832847595215, "learning_rate": 3.940082287558798e-05, "loss": 0.0861, "num_input_tokens_seen": 23454256, "step": 12185 }, { "epoch": 1.9886613916306386, "grad_norm": 0.9795680642127991, "learning_rate": 3.939279669951522e-05, "loss": 0.0658, "num_input_tokens_seen": 23463072, "step": 12190 }, { "epoch": 1.9894771188514562, "grad_norm": 1.3817287683486938, "learning_rate": 3.938476830389234e-05, "loss": 0.1574, "num_input_tokens_seen": 23472816, "step": 12195 }, { "epoch": 1.9902928460722733, "grad_norm": 4.171404838562012, "learning_rate": 3.937673768995742e-05, "loss": 0.0422, "num_input_tokens_seen": 23482400, "step": 12200 }, { "epoch": 1.9902928460722733, "eval_loss": 0.1475783884525299, "eval_runtime": 80.5381, "eval_samples_per_second": 33.835, "eval_steps_per_second": 16.924, "num_input_tokens_seen": 23482400, "step": 12200 }, { "epoch": 1.9911085732930909, "grad_norm": 6.5513529777526855, "learning_rate": 3.936870485894888e-05, "loss": 0.1418, "num_input_tokens_seen": 23491392, "step": 12205 }, { "epoch": 1.991924300513908, "grad_norm": 0.7898595333099365, "learning_rate": 3.9360669812105475e-05, "loss": 0.0384, "num_input_tokens_seen": 23501344, "step": 12210 }, { "epoch": 1.9927400277347256, "grad_norm": 7.992228984832764, "learning_rate": 3.9352632550666325e-05, "loss": 0.112, "num_input_tokens_seen": 23510304, "step": 12215 }, { "epoch": 1.9935557549555427, "grad_norm": 9.311297416687012, "learning_rate": 3.9344593075870866e-05, "loss": 0.1603, "num_input_tokens_seen": 23520224, "step": 12220 }, { "epoch": 1.9943714821763603, "grad_norm": 3.0627493858337402, "learning_rate": 3.933655138895889e-05, "loss": 0.1286, "num_input_tokens_seen": 23528816, "step": 12225 }, { "epoch": 1.9951872093971774, "grad_norm": 6.728572368621826, "learning_rate": 3.932850749117053e-05, "loss": 0.0376, "num_input_tokens_seen": 23538176, "step": 12230 }, { "epoch": 1.996002936617995, "grad_norm": 3.116227865219116, "learning_rate": 3.932046138374624e-05, "loss": 0.1369, "num_input_tokens_seen": 23546608, "step": 12235 }, { "epoch": 1.9968186638388123, "grad_norm": 0.17552058398723602, "learning_rate": 3.9312413067926854e-05, "loss": 0.0934, "num_input_tokens_seen": 23555664, "step": 12240 }, { "epoch": 1.9976343910596297, "grad_norm": 3.5341601371765137, "learning_rate": 3.9304362544953506e-05, "loss": 0.0359, "num_input_tokens_seen": 23564912, "step": 12245 }, { "epoch": 1.998450118280447, "grad_norm": 0.3753301501274109, "learning_rate": 3.929630981606769e-05, "loss": 0.045, "num_input_tokens_seen": 23574976, "step": 12250 }, { "epoch": 1.9992658455012644, "grad_norm": 6.234209060668945, "learning_rate": 3.928825488251124e-05, "loss": 0.0779, "num_input_tokens_seen": 23585520, "step": 12255 }, { "epoch": 2.0, "grad_norm": 0.8983178734779358, "learning_rate": 3.9280197745526344e-05, "loss": 0.1414, "num_input_tokens_seen": 23594288, "step": 12260 }, { "epoch": 2.0008157272208176, "grad_norm": 9.617816925048828, "learning_rate": 3.9272138406355495e-05, "loss": 0.1358, "num_input_tokens_seen": 23603248, "step": 12265 }, { "epoch": 2.0016314544416347, "grad_norm": 2.6751084327697754, "learning_rate": 3.926407686624154e-05, "loss": 0.2254, "num_input_tokens_seen": 23611920, "step": 12270 }, { "epoch": 2.0024471816624523, "grad_norm": 1.218498945236206, "learning_rate": 3.9256013126427684e-05, "loss": 0.0869, "num_input_tokens_seen": 23622272, "step": 12275 }, { "epoch": 2.0032629088832694, "grad_norm": 10.891067504882812, "learning_rate": 3.9247947188157455e-05, "loss": 0.0782, "num_input_tokens_seen": 23631104, "step": 12280 }, { "epoch": 2.004078636104087, "grad_norm": 0.5316359400749207, "learning_rate": 3.9239879052674715e-05, "loss": 0.0145, "num_input_tokens_seen": 23639328, "step": 12285 }, { "epoch": 2.004894363324904, "grad_norm": 0.2262042909860611, "learning_rate": 3.9231808721223673e-05, "loss": 0.0988, "num_input_tokens_seen": 23648416, "step": 12290 }, { "epoch": 2.0057100905457217, "grad_norm": 0.22950005531311035, "learning_rate": 3.9223736195048886e-05, "loss": 0.1461, "num_input_tokens_seen": 23657184, "step": 12295 }, { "epoch": 2.006525817766539, "grad_norm": 9.42253303527832, "learning_rate": 3.921566147539523e-05, "loss": 0.1292, "num_input_tokens_seen": 23667312, "step": 12300 }, { "epoch": 2.0073415449873564, "grad_norm": 0.28284570574760437, "learning_rate": 3.920758456350792e-05, "loss": 0.0744, "num_input_tokens_seen": 23677344, "step": 12305 }, { "epoch": 2.0081572722081735, "grad_norm": 0.20386220514774323, "learning_rate": 3.919950546063253e-05, "loss": 0.0706, "num_input_tokens_seen": 23687856, "step": 12310 }, { "epoch": 2.008972999428991, "grad_norm": 0.15744929015636444, "learning_rate": 3.919142416801496e-05, "loss": 0.0147, "num_input_tokens_seen": 23696432, "step": 12315 }, { "epoch": 2.0097887266498082, "grad_norm": 0.14806832373142242, "learning_rate": 3.918334068690144e-05, "loss": 0.0479, "num_input_tokens_seen": 23706560, "step": 12320 }, { "epoch": 2.010604453870626, "grad_norm": 0.4919969439506531, "learning_rate": 3.917525501853855e-05, "loss": 0.0771, "num_input_tokens_seen": 23716144, "step": 12325 }, { "epoch": 2.011420181091443, "grad_norm": 0.05087290331721306, "learning_rate": 3.916716716417319e-05, "loss": 0.0398, "num_input_tokens_seen": 23725856, "step": 12330 }, { "epoch": 2.0122359083122605, "grad_norm": 0.7999910116195679, "learning_rate": 3.915907712505263e-05, "loss": 0.0166, "num_input_tokens_seen": 23734240, "step": 12335 }, { "epoch": 2.0130516355330776, "grad_norm": 0.08667811006307602, "learning_rate": 3.915098490242444e-05, "loss": 0.2305, "num_input_tokens_seen": 23744208, "step": 12340 }, { "epoch": 2.013867362753895, "grad_norm": 0.6572323441505432, "learning_rate": 3.914289049753654e-05, "loss": 0.0099, "num_input_tokens_seen": 23754432, "step": 12345 }, { "epoch": 2.0146830899747123, "grad_norm": 4.490767002105713, "learning_rate": 3.913479391163719e-05, "loss": 0.1615, "num_input_tokens_seen": 23764064, "step": 12350 }, { "epoch": 2.01549881719553, "grad_norm": 0.2821427285671234, "learning_rate": 3.9126695145975e-05, "loss": 0.1182, "num_input_tokens_seen": 23774144, "step": 12355 }, { "epoch": 2.016314544416347, "grad_norm": 0.574830949306488, "learning_rate": 3.911859420179889e-05, "loss": 0.1337, "num_input_tokens_seen": 23782624, "step": 12360 }, { "epoch": 2.0171302716371646, "grad_norm": 0.08536189049482346, "learning_rate": 3.911049108035813e-05, "loss": 0.0262, "num_input_tokens_seen": 23792896, "step": 12365 }, { "epoch": 2.0179459988579818, "grad_norm": 0.12134217470884323, "learning_rate": 3.910238578290232e-05, "loss": 0.0695, "num_input_tokens_seen": 23802864, "step": 12370 }, { "epoch": 2.0187617260787993, "grad_norm": 0.054739225655794144, "learning_rate": 3.90942783106814e-05, "loss": 0.0427, "num_input_tokens_seen": 23810784, "step": 12375 }, { "epoch": 2.0195774532996165, "grad_norm": 0.055542539805173874, "learning_rate": 3.908616866494564e-05, "loss": 0.0128, "num_input_tokens_seen": 23820096, "step": 12380 }, { "epoch": 2.020393180520434, "grad_norm": 0.2973631024360657, "learning_rate": 3.907805684694566e-05, "loss": 0.1098, "num_input_tokens_seen": 23830272, "step": 12385 }, { "epoch": 2.021208907741251, "grad_norm": 0.20365010201931, "learning_rate": 3.90699428579324e-05, "loss": 0.0673, "num_input_tokens_seen": 23839536, "step": 12390 }, { "epoch": 2.0220246349620687, "grad_norm": 3.1795175075531006, "learning_rate": 3.906182669915713e-05, "loss": 0.0686, "num_input_tokens_seen": 23849392, "step": 12395 }, { "epoch": 2.022840362182886, "grad_norm": 0.07312045991420746, "learning_rate": 3.9053708371871476e-05, "loss": 0.0054, "num_input_tokens_seen": 23860160, "step": 12400 }, { "epoch": 2.022840362182886, "eval_loss": 0.16377586126327515, "eval_runtime": 80.6469, "eval_samples_per_second": 33.789, "eval_steps_per_second": 16.901, "num_input_tokens_seen": 23860160, "step": 12400 }, { "epoch": 2.0236560894037035, "grad_norm": 0.44478288292884827, "learning_rate": 3.904558787732738e-05, "loss": 0.0705, "num_input_tokens_seen": 23869568, "step": 12405 }, { "epoch": 2.0244718166245206, "grad_norm": 0.11413396149873734, "learning_rate": 3.9037465216777135e-05, "loss": 0.1036, "num_input_tokens_seen": 23880016, "step": 12410 }, { "epoch": 2.025287543845338, "grad_norm": 0.030995188280940056, "learning_rate": 3.902934039147334e-05, "loss": 0.0733, "num_input_tokens_seen": 23889312, "step": 12415 }, { "epoch": 2.0261032710661553, "grad_norm": 19.322168350219727, "learning_rate": 3.902121340266894e-05, "loss": 0.0574, "num_input_tokens_seen": 23899536, "step": 12420 }, { "epoch": 2.026918998286973, "grad_norm": 8.862095832824707, "learning_rate": 3.9013084251617246e-05, "loss": 0.1242, "num_input_tokens_seen": 23908304, "step": 12425 }, { "epoch": 2.02773472550779, "grad_norm": 3.6933021545410156, "learning_rate": 3.9004952939571865e-05, "loss": 0.0382, "num_input_tokens_seen": 23918992, "step": 12430 }, { "epoch": 2.0285504527286076, "grad_norm": 0.1444721519947052, "learning_rate": 3.899681946778673e-05, "loss": 0.0729, "num_input_tokens_seen": 23927136, "step": 12435 }, { "epoch": 2.0293661799494247, "grad_norm": 0.15110427141189575, "learning_rate": 3.898868383751615e-05, "loss": 0.0241, "num_input_tokens_seen": 23937616, "step": 12440 }, { "epoch": 2.0301819071702423, "grad_norm": 5.250365734100342, "learning_rate": 3.8980546050014724e-05, "loss": 0.0767, "num_input_tokens_seen": 23947584, "step": 12445 }, { "epoch": 2.03099763439106, "grad_norm": 10.829490661621094, "learning_rate": 3.897240610653741e-05, "loss": 0.075, "num_input_tokens_seen": 23957632, "step": 12450 }, { "epoch": 2.031813361611877, "grad_norm": 0.2280314713716507, "learning_rate": 3.896426400833948e-05, "loss": 0.1369, "num_input_tokens_seen": 23967440, "step": 12455 }, { "epoch": 2.0326290888326946, "grad_norm": 0.16863512992858887, "learning_rate": 3.895611975667656e-05, "loss": 0.0745, "num_input_tokens_seen": 23976928, "step": 12460 }, { "epoch": 2.0334448160535117, "grad_norm": 0.2620634436607361, "learning_rate": 3.8947973352804584e-05, "loss": 0.0903, "num_input_tokens_seen": 23987728, "step": 12465 }, { "epoch": 2.0342605432743293, "grad_norm": 0.025972997769713402, "learning_rate": 3.893982479797984e-05, "loss": 0.1446, "num_input_tokens_seen": 23997568, "step": 12470 }, { "epoch": 2.0350762704951464, "grad_norm": 0.20556983351707458, "learning_rate": 3.8931674093458926e-05, "loss": 0.0257, "num_input_tokens_seen": 24007568, "step": 12475 }, { "epoch": 2.035891997715964, "grad_norm": 2.2535831928253174, "learning_rate": 3.89235212404988e-05, "loss": 0.0376, "num_input_tokens_seen": 24018032, "step": 12480 }, { "epoch": 2.036707724936781, "grad_norm": 0.030922455713152885, "learning_rate": 3.891536624035672e-05, "loss": 0.0933, "num_input_tokens_seen": 24027168, "step": 12485 }, { "epoch": 2.0375234521575987, "grad_norm": 1.0176295042037964, "learning_rate": 3.8907209094290295e-05, "loss": 0.0045, "num_input_tokens_seen": 24037296, "step": 12490 }, { "epoch": 2.038339179378416, "grad_norm": 0.12859266996383667, "learning_rate": 3.8899049803557466e-05, "loss": 0.0211, "num_input_tokens_seen": 24046112, "step": 12495 }, { "epoch": 2.0391549065992334, "grad_norm": 0.010691659525036812, "learning_rate": 3.889088836941648e-05, "loss": 0.0791, "num_input_tokens_seen": 24054768, "step": 12500 }, { "epoch": 2.0399706338200505, "grad_norm": 0.17051485180854797, "learning_rate": 3.8882724793125946e-05, "loss": 0.0026, "num_input_tokens_seen": 24065520, "step": 12505 }, { "epoch": 2.040786361040868, "grad_norm": 1.3364516496658325, "learning_rate": 3.8874559075944794e-05, "loss": 0.0083, "num_input_tokens_seen": 24076112, "step": 12510 }, { "epoch": 2.041602088261685, "grad_norm": 0.06549646705389023, "learning_rate": 3.886639121913227e-05, "loss": 0.0916, "num_input_tokens_seen": 24085824, "step": 12515 }, { "epoch": 2.042417815482503, "grad_norm": 0.5002557039260864, "learning_rate": 3.885822122394797e-05, "loss": 0.0455, "num_input_tokens_seen": 24096304, "step": 12520 }, { "epoch": 2.04323354270332, "grad_norm": 0.23360249400138855, "learning_rate": 3.8850049091651794e-05, "loss": 0.0587, "num_input_tokens_seen": 24106192, "step": 12525 }, { "epoch": 2.0440492699241375, "grad_norm": 0.19034446775913239, "learning_rate": 3.8841874823504e-05, "loss": 0.0099, "num_input_tokens_seen": 24116160, "step": 12530 }, { "epoch": 2.0448649971449546, "grad_norm": 0.050646718591451645, "learning_rate": 3.8833698420765157e-05, "loss": 0.0339, "num_input_tokens_seen": 24125408, "step": 12535 }, { "epoch": 2.045680724365772, "grad_norm": 0.13957597315311432, "learning_rate": 3.882551988469618e-05, "loss": 0.0572, "num_input_tokens_seen": 24135216, "step": 12540 }, { "epoch": 2.0464964515865893, "grad_norm": 6.799407482147217, "learning_rate": 3.881733921655829e-05, "loss": 0.0738, "num_input_tokens_seen": 24144640, "step": 12545 }, { "epoch": 2.047312178807407, "grad_norm": 0.1368280053138733, "learning_rate": 3.8809156417613054e-05, "loss": 0.0435, "num_input_tokens_seen": 24153792, "step": 12550 }, { "epoch": 2.048127906028224, "grad_norm": 2.3649189472198486, "learning_rate": 3.8800971489122364e-05, "loss": 0.3246, "num_input_tokens_seen": 24162624, "step": 12555 }, { "epoch": 2.0489436332490416, "grad_norm": 0.035921160131692886, "learning_rate": 3.8792784432348434e-05, "loss": 0.0013, "num_input_tokens_seen": 24171856, "step": 12560 }, { "epoch": 2.0497593604698587, "grad_norm": 0.032157544046640396, "learning_rate": 3.878459524855381e-05, "loss": 0.1352, "num_input_tokens_seen": 24181888, "step": 12565 }, { "epoch": 2.0505750876906763, "grad_norm": 8.70738410949707, "learning_rate": 3.8776403939001384e-05, "loss": 0.0932, "num_input_tokens_seen": 24191600, "step": 12570 }, { "epoch": 2.0513908149114934, "grad_norm": 0.01815725490450859, "learning_rate": 3.876821050495433e-05, "loss": 0.0151, "num_input_tokens_seen": 24200816, "step": 12575 }, { "epoch": 2.052206542132311, "grad_norm": 0.0051775285974144936, "learning_rate": 3.87600149476762e-05, "loss": 0.1306, "num_input_tokens_seen": 24211344, "step": 12580 }, { "epoch": 2.053022269353128, "grad_norm": 0.13302087783813477, "learning_rate": 3.8751817268430843e-05, "loss": 0.0063, "num_input_tokens_seen": 24221136, "step": 12585 }, { "epoch": 2.0538379965739457, "grad_norm": 0.3567580580711365, "learning_rate": 3.8743617468482464e-05, "loss": 0.0471, "num_input_tokens_seen": 24229616, "step": 12590 }, { "epoch": 2.054653723794763, "grad_norm": 0.9962260723114014, "learning_rate": 3.8735415549095535e-05, "loss": 0.0062, "num_input_tokens_seen": 24239232, "step": 12595 }, { "epoch": 2.0554694510155804, "grad_norm": 9.212347030639648, "learning_rate": 3.8727211511534934e-05, "loss": 0.1379, "num_input_tokens_seen": 24249008, "step": 12600 }, { "epoch": 2.0554694510155804, "eval_loss": 0.2060728222131729, "eval_runtime": 80.6616, "eval_samples_per_second": 33.783, "eval_steps_per_second": 16.898, "num_input_tokens_seen": 24249008, "step": 12600 }, { "epoch": 2.0562851782363976, "grad_norm": 0.02070227824151516, "learning_rate": 3.8719005357065804e-05, "loss": 0.085, "num_input_tokens_seen": 24258432, "step": 12605 }, { "epoch": 2.057100905457215, "grad_norm": 0.3615046441555023, "learning_rate": 3.8710797086953645e-05, "loss": 0.1083, "num_input_tokens_seen": 24268128, "step": 12610 }, { "epoch": 2.0579166326780323, "grad_norm": 0.06797393411397934, "learning_rate": 3.870258670246427e-05, "loss": 0.0447, "num_input_tokens_seen": 24278496, "step": 12615 }, { "epoch": 2.05873235989885, "grad_norm": 4.142745494842529, "learning_rate": 3.869437420486384e-05, "loss": 0.1042, "num_input_tokens_seen": 24287792, "step": 12620 }, { "epoch": 2.059548087119667, "grad_norm": 0.17283061146736145, "learning_rate": 3.8686159595418805e-05, "loss": 0.048, "num_input_tokens_seen": 24295856, "step": 12625 }, { "epoch": 2.0603638143404845, "grad_norm": 0.6650189757347107, "learning_rate": 3.867794287539597e-05, "loss": 0.0075, "num_input_tokens_seen": 24306528, "step": 12630 }, { "epoch": 2.0611795415613017, "grad_norm": 0.06833010911941528, "learning_rate": 3.866972404606245e-05, "loss": 0.0828, "num_input_tokens_seen": 24315760, "step": 12635 }, { "epoch": 2.0619952687821193, "grad_norm": 0.41577452421188354, "learning_rate": 3.866150310868571e-05, "loss": 0.0035, "num_input_tokens_seen": 24326480, "step": 12640 }, { "epoch": 2.062810996002937, "grad_norm": 0.04087607562541962, "learning_rate": 3.8653280064533506e-05, "loss": 0.0056, "num_input_tokens_seen": 24336864, "step": 12645 }, { "epoch": 2.063626723223754, "grad_norm": 3.935596466064453, "learning_rate": 3.864505491487394e-05, "loss": 0.1849, "num_input_tokens_seen": 24346736, "step": 12650 }, { "epoch": 2.0644424504445715, "grad_norm": 0.030004840344190598, "learning_rate": 3.8636827660975414e-05, "loss": 0.0123, "num_input_tokens_seen": 24356000, "step": 12655 }, { "epoch": 2.0652581776653887, "grad_norm": 0.26327913999557495, "learning_rate": 3.862859830410671e-05, "loss": 0.0028, "num_input_tokens_seen": 24366384, "step": 12660 }, { "epoch": 2.0660739048862062, "grad_norm": 7.651117324829102, "learning_rate": 3.862036684553688e-05, "loss": 0.1215, "num_input_tokens_seen": 24377440, "step": 12665 }, { "epoch": 2.0668896321070234, "grad_norm": 0.012587029486894608, "learning_rate": 3.8612133286535314e-05, "loss": 0.101, "num_input_tokens_seen": 24386832, "step": 12670 }, { "epoch": 2.067705359327841, "grad_norm": 0.0510450154542923, "learning_rate": 3.860389762837173e-05, "loss": 0.028, "num_input_tokens_seen": 24395328, "step": 12675 }, { "epoch": 2.068521086548658, "grad_norm": 0.23934341967105865, "learning_rate": 3.859565987231618e-05, "loss": 0.0032, "num_input_tokens_seen": 24403824, "step": 12680 }, { "epoch": 2.0693368137694756, "grad_norm": 6.722578525543213, "learning_rate": 3.858742001963902e-05, "loss": 0.1846, "num_input_tokens_seen": 24414448, "step": 12685 }, { "epoch": 2.0701525409902928, "grad_norm": 0.7437567114830017, "learning_rate": 3.857917807161094e-05, "loss": 0.0669, "num_input_tokens_seen": 24424128, "step": 12690 }, { "epoch": 2.0709682682111104, "grad_norm": 0.43422362208366394, "learning_rate": 3.857093402950296e-05, "loss": 0.0461, "num_input_tokens_seen": 24434096, "step": 12695 }, { "epoch": 2.0717839954319275, "grad_norm": 0.008242120034992695, "learning_rate": 3.8562687894586414e-05, "loss": 0.0925, "num_input_tokens_seen": 24444256, "step": 12700 }, { "epoch": 2.072599722652745, "grad_norm": 2.5358548164367676, "learning_rate": 3.8554439668132946e-05, "loss": 0.1565, "num_input_tokens_seen": 24454496, "step": 12705 }, { "epoch": 2.073415449873562, "grad_norm": 0.3626706600189209, "learning_rate": 3.854618935141455e-05, "loss": 0.1677, "num_input_tokens_seen": 24464432, "step": 12710 }, { "epoch": 2.0742311770943798, "grad_norm": 1.4680817127227783, "learning_rate": 3.8537936945703525e-05, "loss": 0.064, "num_input_tokens_seen": 24474816, "step": 12715 }, { "epoch": 2.075046904315197, "grad_norm": 3.0767295360565186, "learning_rate": 3.852968245227249e-05, "loss": 0.0372, "num_input_tokens_seen": 24484432, "step": 12720 }, { "epoch": 2.0758626315360145, "grad_norm": 7.714775562286377, "learning_rate": 3.85214258723944e-05, "loss": 0.0843, "num_input_tokens_seen": 24493488, "step": 12725 }, { "epoch": 2.0766783587568316, "grad_norm": 5.937942028045654, "learning_rate": 3.8513167207342524e-05, "loss": 0.0688, "num_input_tokens_seen": 24504208, "step": 12730 }, { "epoch": 2.077494085977649, "grad_norm": 0.6842219233512878, "learning_rate": 3.850490645839044e-05, "loss": 0.0798, "num_input_tokens_seen": 24513280, "step": 12735 }, { "epoch": 2.0783098131984663, "grad_norm": 0.643696665763855, "learning_rate": 3.849664362681207e-05, "loss": 0.0447, "num_input_tokens_seen": 24521920, "step": 12740 }, { "epoch": 2.079125540419284, "grad_norm": 10.593219757080078, "learning_rate": 3.848837871388165e-05, "loss": 0.1161, "num_input_tokens_seen": 24532272, "step": 12745 }, { "epoch": 2.079941267640101, "grad_norm": 0.08498917520046234, "learning_rate": 3.848011172087371e-05, "loss": 0.0533, "num_input_tokens_seen": 24543312, "step": 12750 }, { "epoch": 2.0807569948609186, "grad_norm": 1.419345736503601, "learning_rate": 3.847184264906315e-05, "loss": 0.0081, "num_input_tokens_seen": 24552816, "step": 12755 }, { "epoch": 2.0815727220817357, "grad_norm": 0.1112709566950798, "learning_rate": 3.846357149972516e-05, "loss": 0.0422, "num_input_tokens_seen": 24562800, "step": 12760 }, { "epoch": 2.0823884493025533, "grad_norm": 7.47514009475708, "learning_rate": 3.8455298274135246e-05, "loss": 0.185, "num_input_tokens_seen": 24572208, "step": 12765 }, { "epoch": 2.0832041765233704, "grad_norm": 0.770677924156189, "learning_rate": 3.8447022973569254e-05, "loss": 0.2123, "num_input_tokens_seen": 24581536, "step": 12770 }, { "epoch": 2.084019903744188, "grad_norm": 0.038420263677835464, "learning_rate": 3.843874559930332e-05, "loss": 0.1157, "num_input_tokens_seen": 24591232, "step": 12775 }, { "epoch": 2.084835630965005, "grad_norm": 2.832380771636963, "learning_rate": 3.843046615261394e-05, "loss": 0.0619, "num_input_tokens_seen": 24601152, "step": 12780 }, { "epoch": 2.0856513581858227, "grad_norm": 0.057414811104536057, "learning_rate": 3.842218463477791e-05, "loss": 0.0517, "num_input_tokens_seen": 24610944, "step": 12785 }, { "epoch": 2.08646708540664, "grad_norm": 0.08521892875432968, "learning_rate": 3.841390104707233e-05, "loss": 0.2182, "num_input_tokens_seen": 24620992, "step": 12790 }, { "epoch": 2.0872828126274574, "grad_norm": 0.059838734567165375, "learning_rate": 3.8405615390774643e-05, "loss": 0.1488, "num_input_tokens_seen": 24630864, "step": 12795 }, { "epoch": 2.0880985398482745, "grad_norm": 13.6072416305542, "learning_rate": 3.839732766716259e-05, "loss": 0.1184, "num_input_tokens_seen": 24639552, "step": 12800 }, { "epoch": 2.0880985398482745, "eval_loss": 0.1538952738046646, "eval_runtime": 80.5221, "eval_samples_per_second": 33.842, "eval_steps_per_second": 16.927, "num_input_tokens_seen": 24639552, "step": 12800 }, { "epoch": 2.088914267069092, "grad_norm": 0.05724475160241127, "learning_rate": 3.838903787751425e-05, "loss": 0.1097, "num_input_tokens_seen": 24648400, "step": 12805 }, { "epoch": 2.0897299942899092, "grad_norm": 24.370065689086914, "learning_rate": 3.838074602310802e-05, "loss": 0.0921, "num_input_tokens_seen": 24658336, "step": 12810 }, { "epoch": 2.090545721510727, "grad_norm": 0.06297491490840912, "learning_rate": 3.837245210522258e-05, "loss": 0.003, "num_input_tokens_seen": 24667792, "step": 12815 }, { "epoch": 2.0913614487315444, "grad_norm": 10.661189079284668, "learning_rate": 3.8364156125136996e-05, "loss": 0.1458, "num_input_tokens_seen": 24676208, "step": 12820 }, { "epoch": 2.0921771759523615, "grad_norm": 5.47276496887207, "learning_rate": 3.835585808413059e-05, "loss": 0.0944, "num_input_tokens_seen": 24685728, "step": 12825 }, { "epoch": 2.092992903173179, "grad_norm": 0.19460371136665344, "learning_rate": 3.8347557983483024e-05, "loss": 0.0696, "num_input_tokens_seen": 24696528, "step": 12830 }, { "epoch": 2.0938086303939962, "grad_norm": 0.9754253625869751, "learning_rate": 3.833925582447428e-05, "loss": 0.1106, "num_input_tokens_seen": 24705808, "step": 12835 }, { "epoch": 2.094624357614814, "grad_norm": 0.03532562032341957, "learning_rate": 3.8330951608384656e-05, "loss": 0.007, "num_input_tokens_seen": 24715280, "step": 12840 }, { "epoch": 2.095440084835631, "grad_norm": 12.058591842651367, "learning_rate": 3.832264533649477e-05, "loss": 0.0259, "num_input_tokens_seen": 24723360, "step": 12845 }, { "epoch": 2.0962558120564485, "grad_norm": 1.7317264080047607, "learning_rate": 3.8314337010085555e-05, "loss": 0.0527, "num_input_tokens_seen": 24733072, "step": 12850 }, { "epoch": 2.0970715392772656, "grad_norm": 0.3876380920410156, "learning_rate": 3.830602663043824e-05, "loss": 0.0053, "num_input_tokens_seen": 24742896, "step": 12855 }, { "epoch": 2.097887266498083, "grad_norm": 0.24994590878486633, "learning_rate": 3.8297714198834414e-05, "loss": 0.0145, "num_input_tokens_seen": 24752864, "step": 12860 }, { "epoch": 2.0987029937189003, "grad_norm": 0.619368314743042, "learning_rate": 3.828939971655595e-05, "loss": 0.0142, "num_input_tokens_seen": 24761392, "step": 12865 }, { "epoch": 2.099518720939718, "grad_norm": 1.91392982006073, "learning_rate": 3.828108318488505e-05, "loss": 0.111, "num_input_tokens_seen": 24772000, "step": 12870 }, { "epoch": 2.100334448160535, "grad_norm": 0.07988817244768143, "learning_rate": 3.8272764605104216e-05, "loss": 0.02, "num_input_tokens_seen": 24782064, "step": 12875 }, { "epoch": 2.1011501753813526, "grad_norm": 2.309826612472534, "learning_rate": 3.826444397849628e-05, "loss": 0.0732, "num_input_tokens_seen": 24791968, "step": 12880 }, { "epoch": 2.1019659026021698, "grad_norm": 9.357927322387695, "learning_rate": 3.825612130634439e-05, "loss": 0.2518, "num_input_tokens_seen": 24802576, "step": 12885 }, { "epoch": 2.1027816298229873, "grad_norm": 0.12266439199447632, "learning_rate": 3.824779658993202e-05, "loss": 0.0585, "num_input_tokens_seen": 24812336, "step": 12890 }, { "epoch": 2.1035973570438045, "grad_norm": 0.05362628772854805, "learning_rate": 3.823946983054292e-05, "loss": 0.0026, "num_input_tokens_seen": 24822304, "step": 12895 }, { "epoch": 2.104413084264622, "grad_norm": 0.035339873284101486, "learning_rate": 3.82311410294612e-05, "loss": 0.0156, "num_input_tokens_seen": 24832048, "step": 12900 }, { "epoch": 2.105228811485439, "grad_norm": 1.5902180671691895, "learning_rate": 3.822281018797127e-05, "loss": 0.0503, "num_input_tokens_seen": 24842272, "step": 12905 }, { "epoch": 2.1060445387062567, "grad_norm": 2.8650569915771484, "learning_rate": 3.821447730735783e-05, "loss": 0.1389, "num_input_tokens_seen": 24851408, "step": 12910 }, { "epoch": 2.106860265927074, "grad_norm": 13.985198974609375, "learning_rate": 3.820614238890592e-05, "loss": 0.3545, "num_input_tokens_seen": 24862192, "step": 12915 }, { "epoch": 2.1076759931478914, "grad_norm": 0.028613068163394928, "learning_rate": 3.819780543390091e-05, "loss": 0.0018, "num_input_tokens_seen": 24872704, "step": 12920 }, { "epoch": 2.1084917203687086, "grad_norm": 0.3170166313648224, "learning_rate": 3.818946644362844e-05, "loss": 0.0495, "num_input_tokens_seen": 24881232, "step": 12925 }, { "epoch": 2.109307447589526, "grad_norm": 0.08557803183794022, "learning_rate": 3.81811254193745e-05, "loss": 0.0114, "num_input_tokens_seen": 24891872, "step": 12930 }, { "epoch": 2.1101231748103433, "grad_norm": 0.43818581104278564, "learning_rate": 3.8172782362425366e-05, "loss": 0.1051, "num_input_tokens_seen": 24901840, "step": 12935 }, { "epoch": 2.110938902031161, "grad_norm": 3.9735848903656006, "learning_rate": 3.816443727406765e-05, "loss": 0.0074, "num_input_tokens_seen": 24911120, "step": 12940 }, { "epoch": 2.111754629251978, "grad_norm": 1.3955984115600586, "learning_rate": 3.815609015558829e-05, "loss": 0.0605, "num_input_tokens_seen": 24920784, "step": 12945 }, { "epoch": 2.1125703564727956, "grad_norm": 8.311373710632324, "learning_rate": 3.814774100827448e-05, "loss": 0.1674, "num_input_tokens_seen": 24930208, "step": 12950 }, { "epoch": 2.1133860836936127, "grad_norm": 0.1945127546787262, "learning_rate": 3.813938983341379e-05, "loss": 0.086, "num_input_tokens_seen": 24939600, "step": 12955 }, { "epoch": 2.1142018109144303, "grad_norm": 0.21422414481639862, "learning_rate": 3.813103663229407e-05, "loss": 0.0215, "num_input_tokens_seen": 24947840, "step": 12960 }, { "epoch": 2.1150175381352474, "grad_norm": 2.4101901054382324, "learning_rate": 3.812268140620349e-05, "loss": 0.2012, "num_input_tokens_seen": 24958032, "step": 12965 }, { "epoch": 2.115833265356065, "grad_norm": 3.3942089080810547, "learning_rate": 3.811432415643051e-05, "loss": 0.0617, "num_input_tokens_seen": 24966976, "step": 12970 }, { "epoch": 2.116648992576882, "grad_norm": 2.94681715965271, "learning_rate": 3.8105964884263954e-05, "loss": 0.1097, "num_input_tokens_seen": 24978288, "step": 12975 }, { "epoch": 2.1174647197976997, "grad_norm": 0.06756850332021713, "learning_rate": 3.809760359099291e-05, "loss": 0.1215, "num_input_tokens_seen": 24987616, "step": 12980 }, { "epoch": 2.118280447018517, "grad_norm": 0.01590188406407833, "learning_rate": 3.8089240277906804e-05, "loss": 0.0063, "num_input_tokens_seen": 24997824, "step": 12985 }, { "epoch": 2.1190961742393344, "grad_norm": 0.059940338134765625, "learning_rate": 3.808087494629535e-05, "loss": 0.1094, "num_input_tokens_seen": 25007712, "step": 12990 }, { "epoch": 2.1199119014601515, "grad_norm": 2.777463436126709, "learning_rate": 3.8072507597448595e-05, "loss": 0.109, "num_input_tokens_seen": 25017024, "step": 12995 }, { "epoch": 2.120727628680969, "grad_norm": 0.23852042853832245, "learning_rate": 3.806413823265689e-05, "loss": 0.0284, "num_input_tokens_seen": 25026880, "step": 13000 }, { "epoch": 2.120727628680969, "eval_loss": 0.164071187376976, "eval_runtime": 80.557, "eval_samples_per_second": 33.827, "eval_steps_per_second": 16.92, "num_input_tokens_seen": 25026880, "step": 13000 }, { "epoch": 2.121543355901786, "grad_norm": 13.082240104675293, "learning_rate": 3.805576685321089e-05, "loss": 0.0984, "num_input_tokens_seen": 25037408, "step": 13005 }, { "epoch": 2.122359083122604, "grad_norm": 8.84022331237793, "learning_rate": 3.804739346040158e-05, "loss": 0.0437, "num_input_tokens_seen": 25045760, "step": 13010 }, { "epoch": 2.1231748103434214, "grad_norm": 0.03682362660765648, "learning_rate": 3.8039018055520234e-05, "loss": 0.097, "num_input_tokens_seen": 25054896, "step": 13015 }, { "epoch": 2.1239905375642385, "grad_norm": 0.0953642949461937, "learning_rate": 3.803064063985844e-05, "loss": 0.0706, "num_input_tokens_seen": 25064768, "step": 13020 }, { "epoch": 2.124806264785056, "grad_norm": 13.43785285949707, "learning_rate": 3.802226121470811e-05, "loss": 0.1015, "num_input_tokens_seen": 25075152, "step": 13025 }, { "epoch": 2.125621992005873, "grad_norm": 3.1398744583129883, "learning_rate": 3.801387978136145e-05, "loss": 0.1342, "num_input_tokens_seen": 25085120, "step": 13030 }, { "epoch": 2.126437719226691, "grad_norm": 4.965421676635742, "learning_rate": 3.800549634111099e-05, "loss": 0.0827, "num_input_tokens_seen": 25095312, "step": 13035 }, { "epoch": 2.127253446447508, "grad_norm": 1.5016950368881226, "learning_rate": 3.799711089524955e-05, "loss": 0.1107, "num_input_tokens_seen": 25104160, "step": 13040 }, { "epoch": 2.1280691736683255, "grad_norm": 14.924964904785156, "learning_rate": 3.7988723445070285e-05, "loss": 0.0327, "num_input_tokens_seen": 25113104, "step": 13045 }, { "epoch": 2.1288849008891426, "grad_norm": 2.2680728435516357, "learning_rate": 3.798033399186663e-05, "loss": 0.2605, "num_input_tokens_seen": 25123760, "step": 13050 }, { "epoch": 2.12970062810996, "grad_norm": 0.3753352165222168, "learning_rate": 3.797194253693237e-05, "loss": 0.0085, "num_input_tokens_seen": 25133168, "step": 13055 }, { "epoch": 2.1305163553307773, "grad_norm": 0.18092961609363556, "learning_rate": 3.796354908156153e-05, "loss": 0.0593, "num_input_tokens_seen": 25142208, "step": 13060 }, { "epoch": 2.131332082551595, "grad_norm": 2.5093252658843994, "learning_rate": 3.795515362704853e-05, "loss": 0.1558, "num_input_tokens_seen": 25151616, "step": 13065 }, { "epoch": 2.132147809772412, "grad_norm": 0.7501621246337891, "learning_rate": 3.794675617468803e-05, "loss": 0.1023, "num_input_tokens_seen": 25161536, "step": 13070 }, { "epoch": 2.1329635369932296, "grad_norm": 4.109676837921143, "learning_rate": 3.793835672577503e-05, "loss": 0.1906, "num_input_tokens_seen": 25172032, "step": 13075 }, { "epoch": 2.1337792642140467, "grad_norm": 0.183066263794899, "learning_rate": 3.7929955281604826e-05, "loss": 0.0283, "num_input_tokens_seen": 25180400, "step": 13080 }, { "epoch": 2.1345949914348643, "grad_norm": 3.018068790435791, "learning_rate": 3.7921551843473036e-05, "loss": 0.1406, "num_input_tokens_seen": 25189232, "step": 13085 }, { "epoch": 2.1354107186556814, "grad_norm": 15.609221458435059, "learning_rate": 3.791314641267557e-05, "loss": 0.0881, "num_input_tokens_seen": 25199584, "step": 13090 }, { "epoch": 2.136226445876499, "grad_norm": 0.7709688544273376, "learning_rate": 3.790473899050864e-05, "loss": 0.0578, "num_input_tokens_seen": 25208288, "step": 13095 }, { "epoch": 2.137042173097316, "grad_norm": 0.8310428857803345, "learning_rate": 3.7896329578268794e-05, "loss": 0.0455, "num_input_tokens_seen": 25218576, "step": 13100 }, { "epoch": 2.1378579003181337, "grad_norm": 8.525418281555176, "learning_rate": 3.7887918177252855e-05, "loss": 0.1775, "num_input_tokens_seen": 25228032, "step": 13105 }, { "epoch": 2.138673627538951, "grad_norm": 0.8127164840698242, "learning_rate": 3.787950478875798e-05, "loss": 0.0565, "num_input_tokens_seen": 25239392, "step": 13110 }, { "epoch": 2.1394893547597684, "grad_norm": 11.037280082702637, "learning_rate": 3.787108941408162e-05, "loss": 0.0986, "num_input_tokens_seen": 25248656, "step": 13115 }, { "epoch": 2.1403050819805856, "grad_norm": 0.07471893727779388, "learning_rate": 3.786267205452151e-05, "loss": 0.003, "num_input_tokens_seen": 25257360, "step": 13120 }, { "epoch": 2.141120809201403, "grad_norm": 1.1793055534362793, "learning_rate": 3.785425271137573e-05, "loss": 0.0485, "num_input_tokens_seen": 25266576, "step": 13125 }, { "epoch": 2.1419365364222203, "grad_norm": 18.25712776184082, "learning_rate": 3.7845831385942655e-05, "loss": 0.1017, "num_input_tokens_seen": 25277456, "step": 13130 }, { "epoch": 2.142752263643038, "grad_norm": 0.049476154148578644, "learning_rate": 3.7837408079520944e-05, "loss": 0.0832, "num_input_tokens_seen": 25286160, "step": 13135 }, { "epoch": 2.143567990863855, "grad_norm": 5.074484825134277, "learning_rate": 3.782898279340957e-05, "loss": 0.122, "num_input_tokens_seen": 25295040, "step": 13140 }, { "epoch": 2.1443837180846725, "grad_norm": 0.03451668098568916, "learning_rate": 3.782055552890784e-05, "loss": 0.0362, "num_input_tokens_seen": 25304656, "step": 13145 }, { "epoch": 2.1451994453054897, "grad_norm": 11.66770076751709, "learning_rate": 3.781212628731534e-05, "loss": 0.0493, "num_input_tokens_seen": 25313616, "step": 13150 }, { "epoch": 2.1460151725263072, "grad_norm": 3.6386101245880127, "learning_rate": 3.7803695069931946e-05, "loss": 0.1374, "num_input_tokens_seen": 25323664, "step": 13155 }, { "epoch": 2.1468308997471244, "grad_norm": 0.25280144810676575, "learning_rate": 3.779526187805789e-05, "loss": 0.1427, "num_input_tokens_seen": 25332864, "step": 13160 }, { "epoch": 2.147646626967942, "grad_norm": 3.1510727405548096, "learning_rate": 3.778682671299364e-05, "loss": 0.1506, "num_input_tokens_seen": 25342512, "step": 13165 }, { "epoch": 2.148462354188759, "grad_norm": 1.0462629795074463, "learning_rate": 3.777838957604003e-05, "loss": 0.0361, "num_input_tokens_seen": 25351296, "step": 13170 }, { "epoch": 2.1492780814095767, "grad_norm": 3.379539728164673, "learning_rate": 3.776995046849816e-05, "loss": 0.124, "num_input_tokens_seen": 25360592, "step": 13175 }, { "epoch": 2.150093808630394, "grad_norm": 0.29367539286613464, "learning_rate": 3.776150939166945e-05, "loss": 0.1514, "num_input_tokens_seen": 25370096, "step": 13180 }, { "epoch": 2.1509095358512114, "grad_norm": 0.1490248292684555, "learning_rate": 3.775306634685562e-05, "loss": 0.0285, "num_input_tokens_seen": 25381392, "step": 13185 }, { "epoch": 2.151725263072029, "grad_norm": 12.442108154296875, "learning_rate": 3.7744621335358696e-05, "loss": 0.2144, "num_input_tokens_seen": 25390992, "step": 13190 }, { "epoch": 2.152540990292846, "grad_norm": 7.58488655090332, "learning_rate": 3.7736174358481e-05, "loss": 0.093, "num_input_tokens_seen": 25401136, "step": 13195 }, { "epoch": 2.153356717513663, "grad_norm": 0.07765518873929977, "learning_rate": 3.7727725417525175e-05, "loss": 0.0089, "num_input_tokens_seen": 25410448, "step": 13200 }, { "epoch": 2.153356717513663, "eval_loss": 0.15041252970695496, "eval_runtime": 80.5437, "eval_samples_per_second": 33.833, "eval_steps_per_second": 16.922, "num_input_tokens_seen": 25410448, "step": 13200 }, { "epoch": 2.1541724447344808, "grad_norm": 0.22303394973278046, "learning_rate": 3.771927451379414e-05, "loss": 0.0481, "num_input_tokens_seen": 25421264, "step": 13205 }, { "epoch": 2.1549881719552983, "grad_norm": 0.06130184605717659, "learning_rate": 3.7710821648591135e-05, "loss": 0.0091, "num_input_tokens_seen": 25429792, "step": 13210 }, { "epoch": 2.1558038991761155, "grad_norm": 3.639263153076172, "learning_rate": 3.7702366823219694e-05, "loss": 0.125, "num_input_tokens_seen": 25438736, "step": 13215 }, { "epoch": 2.156619626396933, "grad_norm": 0.0890580415725708, "learning_rate": 3.769391003898366e-05, "loss": 0.1289, "num_input_tokens_seen": 25447760, "step": 13220 }, { "epoch": 2.15743535361775, "grad_norm": 0.16205084323883057, "learning_rate": 3.768545129718718e-05, "loss": 0.1362, "num_input_tokens_seen": 25458848, "step": 13225 }, { "epoch": 2.1582510808385678, "grad_norm": 0.15712398290634155, "learning_rate": 3.7676990599134686e-05, "loss": 0.0039, "num_input_tokens_seen": 25468688, "step": 13230 }, { "epoch": 2.159066808059385, "grad_norm": 13.965954780578613, "learning_rate": 3.766852794613095e-05, "loss": 0.174, "num_input_tokens_seen": 25478032, "step": 13235 }, { "epoch": 2.1598825352802025, "grad_norm": 0.09934819489717484, "learning_rate": 3.766006333948099e-05, "loss": 0.1918, "num_input_tokens_seen": 25486976, "step": 13240 }, { "epoch": 2.1606982625010196, "grad_norm": 3.8237054347991943, "learning_rate": 3.765159678049017e-05, "loss": 0.0494, "num_input_tokens_seen": 25497392, "step": 13245 }, { "epoch": 2.161513989721837, "grad_norm": 0.09137393534183502, "learning_rate": 3.7643128270464134e-05, "loss": 0.063, "num_input_tokens_seen": 25507872, "step": 13250 }, { "epoch": 2.1623297169426543, "grad_norm": 4.510310649871826, "learning_rate": 3.763465781070884e-05, "loss": 0.0455, "num_input_tokens_seen": 25517040, "step": 13255 }, { "epoch": 2.163145444163472, "grad_norm": 0.3208509385585785, "learning_rate": 3.762618540253052e-05, "loss": 0.086, "num_input_tokens_seen": 25526816, "step": 13260 }, { "epoch": 2.163961171384289, "grad_norm": 0.06116088852286339, "learning_rate": 3.761771104723576e-05, "loss": 0.0132, "num_input_tokens_seen": 25535904, "step": 13265 }, { "epoch": 2.1647768986051066, "grad_norm": 0.11833060532808304, "learning_rate": 3.7609234746131386e-05, "loss": 0.0674, "num_input_tokens_seen": 25545552, "step": 13270 }, { "epoch": 2.1655926258259237, "grad_norm": 0.05778205767273903, "learning_rate": 3.7600756500524556e-05, "loss": 0.0613, "num_input_tokens_seen": 25554656, "step": 13275 }, { "epoch": 2.1664083530467413, "grad_norm": 0.10652953386306763, "learning_rate": 3.759227631172271e-05, "loss": 0.0367, "num_input_tokens_seen": 25563280, "step": 13280 }, { "epoch": 2.1672240802675584, "grad_norm": 4.46542501449585, "learning_rate": 3.758379418103363e-05, "loss": 0.0336, "num_input_tokens_seen": 25572688, "step": 13285 }, { "epoch": 2.168039807488376, "grad_norm": 8.399996757507324, "learning_rate": 3.757531010976534e-05, "loss": 0.1203, "num_input_tokens_seen": 25581808, "step": 13290 }, { "epoch": 2.168855534709193, "grad_norm": 4.148972511291504, "learning_rate": 3.75668240992262e-05, "loss": 0.1513, "num_input_tokens_seen": 25591856, "step": 13295 }, { "epoch": 2.1696712619300107, "grad_norm": 0.030086670070886612, "learning_rate": 3.7558336150724865e-05, "loss": 0.0278, "num_input_tokens_seen": 25601568, "step": 13300 }, { "epoch": 2.170486989150828, "grad_norm": 2.531381845474243, "learning_rate": 3.754984626557028e-05, "loss": 0.0969, "num_input_tokens_seen": 25611120, "step": 13305 }, { "epoch": 2.1713027163716454, "grad_norm": 0.244443878531456, "learning_rate": 3.754135444507168e-05, "loss": 0.1026, "num_input_tokens_seen": 25621488, "step": 13310 }, { "epoch": 2.1721184435924625, "grad_norm": 4.091286659240723, "learning_rate": 3.753286069053863e-05, "loss": 0.0705, "num_input_tokens_seen": 25631168, "step": 13315 }, { "epoch": 2.17293417081328, "grad_norm": 8.720475196838379, "learning_rate": 3.7524365003280945e-05, "loss": 0.0482, "num_input_tokens_seen": 25640400, "step": 13320 }, { "epoch": 2.1737498980340972, "grad_norm": 0.04770376905798912, "learning_rate": 3.75158673846088e-05, "loss": 0.0582, "num_input_tokens_seen": 25647808, "step": 13325 }, { "epoch": 2.174565625254915, "grad_norm": 13.262475967407227, "learning_rate": 3.750736783583262e-05, "loss": 0.0963, "num_input_tokens_seen": 25656224, "step": 13330 }, { "epoch": 2.175381352475732, "grad_norm": 0.16525927186012268, "learning_rate": 3.7498866358263144e-05, "loss": 0.1554, "num_input_tokens_seen": 25666368, "step": 13335 }, { "epoch": 2.1761970796965495, "grad_norm": 2.3303208351135254, "learning_rate": 3.74903629532114e-05, "loss": 0.0259, "num_input_tokens_seen": 25675968, "step": 13340 }, { "epoch": 2.1770128069173666, "grad_norm": 0.292331725358963, "learning_rate": 3.748185762198873e-05, "loss": 0.1647, "num_input_tokens_seen": 25684800, "step": 13345 }, { "epoch": 2.1778285341381842, "grad_norm": 0.08469189703464508, "learning_rate": 3.747335036590676e-05, "loss": 0.0032, "num_input_tokens_seen": 25693712, "step": 13350 }, { "epoch": 2.1786442613590014, "grad_norm": 0.1123327910900116, "learning_rate": 3.7464841186277405e-05, "loss": 0.0325, "num_input_tokens_seen": 25702624, "step": 13355 }, { "epoch": 2.179459988579819, "grad_norm": 0.15994194149971008, "learning_rate": 3.7456330084412896e-05, "loss": 0.0511, "num_input_tokens_seen": 25711744, "step": 13360 }, { "epoch": 2.180275715800636, "grad_norm": 4.283573627471924, "learning_rate": 3.744781706162576e-05, "loss": 0.1668, "num_input_tokens_seen": 25720064, "step": 13365 }, { "epoch": 2.1810914430214536, "grad_norm": 2.561093330383301, "learning_rate": 3.743930211922879e-05, "loss": 0.1609, "num_input_tokens_seen": 25730672, "step": 13370 }, { "epoch": 2.1819071702422708, "grad_norm": 0.3772830069065094, "learning_rate": 3.743078525853513e-05, "loss": 0.1635, "num_input_tokens_seen": 25740928, "step": 13375 }, { "epoch": 2.1827228974630883, "grad_norm": 9.789597511291504, "learning_rate": 3.7422266480858154e-05, "loss": 0.0258, "num_input_tokens_seen": 25749600, "step": 13380 }, { "epoch": 2.183538624683906, "grad_norm": 0.20130957663059235, "learning_rate": 3.741374578751158e-05, "loss": 0.1035, "num_input_tokens_seen": 25759056, "step": 13385 }, { "epoch": 2.184354351904723, "grad_norm": 32.2354850769043, "learning_rate": 3.740522317980941e-05, "loss": 0.0921, "num_input_tokens_seen": 25766576, "step": 13390 }, { "epoch": 2.1851700791255406, "grad_norm": 4.503053188323975, "learning_rate": 3.739669865906593e-05, "loss": 0.1326, "num_input_tokens_seen": 25776640, "step": 13395 }, { "epoch": 2.1859858063463578, "grad_norm": 0.08148358762264252, "learning_rate": 3.738817222659573e-05, "loss": 0.0075, "num_input_tokens_seen": 25785744, "step": 13400 }, { "epoch": 2.1859858063463578, "eval_loss": 0.1619856208562851, "eval_runtime": 80.6126, "eval_samples_per_second": 33.804, "eval_steps_per_second": 16.908, "num_input_tokens_seen": 25785744, "step": 13400 }, { "epoch": 2.1868015335671753, "grad_norm": 6.5838494300842285, "learning_rate": 3.73796438837137e-05, "loss": 0.2266, "num_input_tokens_seen": 25795504, "step": 13405 }, { "epoch": 2.1876172607879925, "grad_norm": 0.04629312828183174, "learning_rate": 3.7371113631735e-05, "loss": 0.1119, "num_input_tokens_seen": 25805776, "step": 13410 }, { "epoch": 2.18843298800881, "grad_norm": 8.007908821105957, "learning_rate": 3.736258147197512e-05, "loss": 0.029, "num_input_tokens_seen": 25815840, "step": 13415 }, { "epoch": 2.189248715229627, "grad_norm": 0.15080362558364868, "learning_rate": 3.735404740574981e-05, "loss": 0.0392, "num_input_tokens_seen": 25824112, "step": 13420 }, { "epoch": 2.1900644424504447, "grad_norm": 0.32515934109687805, "learning_rate": 3.7345511434375145e-05, "loss": 0.1795, "num_input_tokens_seen": 25833664, "step": 13425 }, { "epoch": 2.190880169671262, "grad_norm": 0.07551378756761551, "learning_rate": 3.733697355916748e-05, "loss": 0.0559, "num_input_tokens_seen": 25842656, "step": 13430 }, { "epoch": 2.1916958968920794, "grad_norm": 0.1435483992099762, "learning_rate": 3.732843378144345e-05, "loss": 0.0954, "num_input_tokens_seen": 25851584, "step": 13435 }, { "epoch": 2.1925116241128966, "grad_norm": 4.821315288543701, "learning_rate": 3.7319892102519995e-05, "loss": 0.101, "num_input_tokens_seen": 25860528, "step": 13440 }, { "epoch": 2.193327351333714, "grad_norm": 0.804589569568634, "learning_rate": 3.731134852371436e-05, "loss": 0.0699, "num_input_tokens_seen": 25870448, "step": 13445 }, { "epoch": 2.1941430785545313, "grad_norm": 2.8145649433135986, "learning_rate": 3.730280304634408e-05, "loss": 0.0869, "num_input_tokens_seen": 25880496, "step": 13450 }, { "epoch": 2.194958805775349, "grad_norm": 0.15799278020858765, "learning_rate": 3.729425567172696e-05, "loss": 0.0499, "num_input_tokens_seen": 25891504, "step": 13455 }, { "epoch": 2.195774532996166, "grad_norm": 7.647582530975342, "learning_rate": 3.728570640118111e-05, "loss": 0.0783, "num_input_tokens_seen": 25901488, "step": 13460 }, { "epoch": 2.1965902602169836, "grad_norm": 3.575045108795166, "learning_rate": 3.727715523602494e-05, "loss": 0.0775, "num_input_tokens_seen": 25910608, "step": 13465 }, { "epoch": 2.1974059874378007, "grad_norm": 0.119844950735569, "learning_rate": 3.726860217757715e-05, "loss": 0.0514, "num_input_tokens_seen": 25920864, "step": 13470 }, { "epoch": 2.1982217146586183, "grad_norm": 2.9803574085235596, "learning_rate": 3.726004722715673e-05, "loss": 0.1181, "num_input_tokens_seen": 25929040, "step": 13475 }, { "epoch": 2.1990374418794354, "grad_norm": 0.16828687489032745, "learning_rate": 3.725149038608296e-05, "loss": 0.0035, "num_input_tokens_seen": 25939008, "step": 13480 }, { "epoch": 2.199853169100253, "grad_norm": 0.1576552391052246, "learning_rate": 3.7242931655675404e-05, "loss": 0.0043, "num_input_tokens_seen": 25948720, "step": 13485 }, { "epoch": 2.20066889632107, "grad_norm": 0.3171720802783966, "learning_rate": 3.7234371037253937e-05, "loss": 0.0199, "num_input_tokens_seen": 25957360, "step": 13490 }, { "epoch": 2.2014846235418877, "grad_norm": 8.404420852661133, "learning_rate": 3.7225808532138705e-05, "loss": 0.0839, "num_input_tokens_seen": 25966048, "step": 13495 }, { "epoch": 2.202300350762705, "grad_norm": 4.111469745635986, "learning_rate": 3.721724414165016e-05, "loss": 0.1735, "num_input_tokens_seen": 25976080, "step": 13500 }, { "epoch": 2.2031160779835224, "grad_norm": 0.07462161034345627, "learning_rate": 3.720867786710904e-05, "loss": 0.0835, "num_input_tokens_seen": 25985232, "step": 13505 }, { "epoch": 2.2039318052043395, "grad_norm": 0.04356435313820839, "learning_rate": 3.7200109709836366e-05, "loss": 0.1519, "num_input_tokens_seen": 25994080, "step": 13510 }, { "epoch": 2.204747532425157, "grad_norm": 0.11485077440738678, "learning_rate": 3.7191539671153465e-05, "loss": 0.0094, "num_input_tokens_seen": 26004464, "step": 13515 }, { "epoch": 2.205563259645974, "grad_norm": 5.914989471435547, "learning_rate": 3.718296775238193e-05, "loss": 0.0291, "num_input_tokens_seen": 26013504, "step": 13520 }, { "epoch": 2.206378986866792, "grad_norm": 0.2737886905670166, "learning_rate": 3.7174393954843675e-05, "loss": 0.0077, "num_input_tokens_seen": 26022496, "step": 13525 }, { "epoch": 2.207194714087609, "grad_norm": 0.517493724822998, "learning_rate": 3.716581827986087e-05, "loss": 0.113, "num_input_tokens_seen": 26031984, "step": 13530 }, { "epoch": 2.2080104413084265, "grad_norm": 0.3777415156364441, "learning_rate": 3.7157240728756004e-05, "loss": 0.0826, "num_input_tokens_seen": 26040912, "step": 13535 }, { "epoch": 2.2088261685292436, "grad_norm": 0.13363708555698395, "learning_rate": 3.714866130285184e-05, "loss": 0.1165, "num_input_tokens_seen": 26050544, "step": 13540 }, { "epoch": 2.209641895750061, "grad_norm": 13.289350509643555, "learning_rate": 3.714008000347143e-05, "loss": 0.1248, "num_input_tokens_seen": 26059488, "step": 13545 }, { "epoch": 2.2104576229708783, "grad_norm": 1.2117440700531006, "learning_rate": 3.7131496831938126e-05, "loss": 0.1524, "num_input_tokens_seen": 26069264, "step": 13550 }, { "epoch": 2.211273350191696, "grad_norm": 0.12210609763860703, "learning_rate": 3.7122911789575565e-05, "loss": 0.0185, "num_input_tokens_seen": 26077680, "step": 13555 }, { "epoch": 2.2120890774125135, "grad_norm": 11.305815696716309, "learning_rate": 3.711432487770765e-05, "loss": 0.0619, "num_input_tokens_seen": 26087632, "step": 13560 }, { "epoch": 2.2129048046333306, "grad_norm": 0.06385408341884613, "learning_rate": 3.710573609765861e-05, "loss": 0.1784, "num_input_tokens_seen": 26097328, "step": 13565 }, { "epoch": 2.2137205318541477, "grad_norm": 9.714229583740234, "learning_rate": 3.709714545075292e-05, "loss": 0.2022, "num_input_tokens_seen": 26106288, "step": 13570 }, { "epoch": 2.2145362590749653, "grad_norm": 3.8804495334625244, "learning_rate": 3.708855293831538e-05, "loss": 0.0146, "num_input_tokens_seen": 26115792, "step": 13575 }, { "epoch": 2.215351986295783, "grad_norm": 3.206080436706543, "learning_rate": 3.707995856167107e-05, "loss": 0.2115, "num_input_tokens_seen": 26124512, "step": 13580 }, { "epoch": 2.2161677135166, "grad_norm": 0.07431373000144958, "learning_rate": 3.707136232214534e-05, "loss": 0.0389, "num_input_tokens_seen": 26134224, "step": 13585 }, { "epoch": 2.2169834407374176, "grad_norm": 8.447474479675293, "learning_rate": 3.7062764221063844e-05, "loss": 0.1845, "num_input_tokens_seen": 26143584, "step": 13590 }, { "epoch": 2.2177991679582347, "grad_norm": 0.10574942082166672, "learning_rate": 3.705416425975252e-05, "loss": 0.0061, "num_input_tokens_seen": 26153408, "step": 13595 }, { "epoch": 2.2186148951790523, "grad_norm": 0.10322944074869156, "learning_rate": 3.704556243953758e-05, "loss": 0.0404, "num_input_tokens_seen": 26163104, "step": 13600 }, { "epoch": 2.2186148951790523, "eval_loss": 0.17065684497356415, "eval_runtime": 80.5261, "eval_samples_per_second": 33.84, "eval_steps_per_second": 16.926, "num_input_tokens_seen": 26163104, "step": 13600 }, { "epoch": 2.2194306223998694, "grad_norm": 2.336543083190918, "learning_rate": 3.7036958761745535e-05, "loss": 0.1406, "num_input_tokens_seen": 26171696, "step": 13605 }, { "epoch": 2.220246349620687, "grad_norm": 7.292792320251465, "learning_rate": 3.702835322770318e-05, "loss": 0.1511, "num_input_tokens_seen": 26180912, "step": 13610 }, { "epoch": 2.221062076841504, "grad_norm": 0.7093823552131653, "learning_rate": 3.701974583873761e-05, "loss": 0.0171, "num_input_tokens_seen": 26189680, "step": 13615 }, { "epoch": 2.2218778040623217, "grad_norm": 5.899660587310791, "learning_rate": 3.701113659617618e-05, "loss": 0.095, "num_input_tokens_seen": 26200608, "step": 13620 }, { "epoch": 2.222693531283139, "grad_norm": 1.1200942993164062, "learning_rate": 3.7002525501346535e-05, "loss": 0.0573, "num_input_tokens_seen": 26209824, "step": 13625 }, { "epoch": 2.2235092585039564, "grad_norm": 9.90450668334961, "learning_rate": 3.699391255557664e-05, "loss": 0.0266, "num_input_tokens_seen": 26219680, "step": 13630 }, { "epoch": 2.2243249857247736, "grad_norm": 0.14820624887943268, "learning_rate": 3.69852977601947e-05, "loss": 0.0671, "num_input_tokens_seen": 26229136, "step": 13635 }, { "epoch": 2.225140712945591, "grad_norm": 8.976523399353027, "learning_rate": 3.697668111652922e-05, "loss": 0.0478, "num_input_tokens_seen": 26238208, "step": 13640 }, { "epoch": 2.2259564401664083, "grad_norm": 0.30997446179389954, "learning_rate": 3.6968062625909005e-05, "loss": 0.1194, "num_input_tokens_seen": 26247904, "step": 13645 }, { "epoch": 2.226772167387226, "grad_norm": 4.93719482421875, "learning_rate": 3.6959442289663135e-05, "loss": 0.1431, "num_input_tokens_seen": 26257776, "step": 13650 }, { "epoch": 2.227587894608043, "grad_norm": 0.5334243774414062, "learning_rate": 3.695082010912098e-05, "loss": 0.0667, "num_input_tokens_seen": 26267936, "step": 13655 }, { "epoch": 2.2284036218288605, "grad_norm": 0.23789095878601074, "learning_rate": 3.694219608561217e-05, "loss": 0.0379, "num_input_tokens_seen": 26277840, "step": 13660 }, { "epoch": 2.2292193490496777, "grad_norm": 0.152626171708107, "learning_rate": 3.693357022046665e-05, "loss": 0.0659, "num_input_tokens_seen": 26287296, "step": 13665 }, { "epoch": 2.2300350762704952, "grad_norm": 0.6952475905418396, "learning_rate": 3.6924942515014644e-05, "loss": 0.1387, "num_input_tokens_seen": 26298224, "step": 13670 }, { "epoch": 2.2308508034913124, "grad_norm": 1.6431065797805786, "learning_rate": 3.691631297058664e-05, "loss": 0.0416, "num_input_tokens_seen": 26307744, "step": 13675 }, { "epoch": 2.23166653071213, "grad_norm": 0.06271620094776154, "learning_rate": 3.6907681588513424e-05, "loss": 0.0389, "num_input_tokens_seen": 26316992, "step": 13680 }, { "epoch": 2.232482257932947, "grad_norm": 0.0703810378909111, "learning_rate": 3.689904837012606e-05, "loss": 0.1392, "num_input_tokens_seen": 26327424, "step": 13685 }, { "epoch": 2.2332979851537647, "grad_norm": 23.72652244567871, "learning_rate": 3.689041331675591e-05, "loss": 0.0406, "num_input_tokens_seen": 26338000, "step": 13690 }, { "epoch": 2.234113712374582, "grad_norm": 0.9204171299934387, "learning_rate": 3.688177642973461e-05, "loss": 0.3213, "num_input_tokens_seen": 26347952, "step": 13695 }, { "epoch": 2.2349294395953994, "grad_norm": 8.693523406982422, "learning_rate": 3.687313771039406e-05, "loss": 0.1075, "num_input_tokens_seen": 26358000, "step": 13700 }, { "epoch": 2.2357451668162165, "grad_norm": 0.5701942443847656, "learning_rate": 3.686449716006647e-05, "loss": 0.0841, "num_input_tokens_seen": 26367104, "step": 13705 }, { "epoch": 2.236560894037034, "grad_norm": 0.08740047365427017, "learning_rate": 3.685585478008432e-05, "loss": 0.0602, "num_input_tokens_seen": 26376848, "step": 13710 }, { "epoch": 2.237376621257851, "grad_norm": 0.06965368986129761, "learning_rate": 3.6847210571780364e-05, "loss": 0.0063, "num_input_tokens_seen": 26386128, "step": 13715 }, { "epoch": 2.2381923484786688, "grad_norm": 0.21464189887046814, "learning_rate": 3.683856453648767e-05, "loss": 0.0105, "num_input_tokens_seen": 26394688, "step": 13720 }, { "epoch": 2.239008075699486, "grad_norm": 18.57854652404785, "learning_rate": 3.682991667553954e-05, "loss": 0.0911, "num_input_tokens_seen": 26404608, "step": 13725 }, { "epoch": 2.2398238029203035, "grad_norm": 3.3385727405548096, "learning_rate": 3.6821266990269606e-05, "loss": 0.225, "num_input_tokens_seen": 26413408, "step": 13730 }, { "epoch": 2.2406395301411206, "grad_norm": 0.24209219217300415, "learning_rate": 3.681261548201174e-05, "loss": 0.0768, "num_input_tokens_seen": 26423840, "step": 13735 }, { "epoch": 2.241455257361938, "grad_norm": 1.7724837064743042, "learning_rate": 3.6803962152100125e-05, "loss": 0.0777, "num_input_tokens_seen": 26432240, "step": 13740 }, { "epoch": 2.2422709845827553, "grad_norm": 12.673678398132324, "learning_rate": 3.67953070018692e-05, "loss": 0.1224, "num_input_tokens_seen": 26443088, "step": 13745 }, { "epoch": 2.243086711803573, "grad_norm": 0.06871150434017181, "learning_rate": 3.678665003265371e-05, "loss": 0.12, "num_input_tokens_seen": 26453024, "step": 13750 }, { "epoch": 2.2439024390243905, "grad_norm": 6.4378981590271, "learning_rate": 3.677799124578867e-05, "loss": 0.196, "num_input_tokens_seen": 26462704, "step": 13755 }, { "epoch": 2.2447181662452076, "grad_norm": 20.861318588256836, "learning_rate": 3.676933064260937e-05, "loss": 0.0259, "num_input_tokens_seen": 26473200, "step": 13760 }, { "epoch": 2.245533893466025, "grad_norm": 10.523355484008789, "learning_rate": 3.6760668224451365e-05, "loss": 0.0437, "num_input_tokens_seen": 26482048, "step": 13765 }, { "epoch": 2.2463496206868423, "grad_norm": 0.2767745554447174, "learning_rate": 3.675200399265054e-05, "loss": 0.0072, "num_input_tokens_seen": 26490816, "step": 13770 }, { "epoch": 2.24716534790766, "grad_norm": 0.2473295032978058, "learning_rate": 3.6743337948543014e-05, "loss": 0.0528, "num_input_tokens_seen": 26500928, "step": 13775 }, { "epoch": 2.247981075128477, "grad_norm": 0.09700512140989304, "learning_rate": 3.6734670093465204e-05, "loss": 0.0434, "num_input_tokens_seen": 26510688, "step": 13780 }, { "epoch": 2.2487968023492946, "grad_norm": 0.1859816610813141, "learning_rate": 3.672600042875379e-05, "loss": 0.0622, "num_input_tokens_seen": 26518704, "step": 13785 }, { "epoch": 2.2496125295701117, "grad_norm": 13.333586692810059, "learning_rate": 3.671732895574575e-05, "loss": 0.0962, "num_input_tokens_seen": 26527376, "step": 13790 }, { "epoch": 2.2504282567909293, "grad_norm": 4.798047065734863, "learning_rate": 3.670865567577834e-05, "loss": 0.2155, "num_input_tokens_seen": 26536080, "step": 13795 }, { "epoch": 2.2512439840117464, "grad_norm": 0.2895274758338928, "learning_rate": 3.669998059018909e-05, "loss": 0.0478, "num_input_tokens_seen": 26546240, "step": 13800 }, { "epoch": 2.2512439840117464, "eval_loss": 0.17794707417488098, "eval_runtime": 80.5604, "eval_samples_per_second": 33.826, "eval_steps_per_second": 16.919, "num_input_tokens_seen": 26546240, "step": 13800 }, { "epoch": 2.252059711232564, "grad_norm": 5.545392990112305, "learning_rate": 3.6691303700315796e-05, "loss": 0.0871, "num_input_tokens_seen": 26554800, "step": 13805 }, { "epoch": 2.252875438453381, "grad_norm": 0.03368283063173294, "learning_rate": 3.668262500749655e-05, "loss": 0.0659, "num_input_tokens_seen": 26563408, "step": 13810 }, { "epoch": 2.2536911656741987, "grad_norm": 0.1969253271818161, "learning_rate": 3.667394451306971e-05, "loss": 0.1567, "num_input_tokens_seen": 26572720, "step": 13815 }, { "epoch": 2.254506892895016, "grad_norm": 0.3471747934818268, "learning_rate": 3.666526221837393e-05, "loss": 0.0248, "num_input_tokens_seen": 26582432, "step": 13820 }, { "epoch": 2.2553226201158334, "grad_norm": 0.8622422218322754, "learning_rate": 3.665657812474812e-05, "loss": 0.0767, "num_input_tokens_seen": 26592336, "step": 13825 }, { "epoch": 2.2561383473366505, "grad_norm": 0.08418353646993637, "learning_rate": 3.664789223353147e-05, "loss": 0.0073, "num_input_tokens_seen": 26601136, "step": 13830 }, { "epoch": 2.256954074557468, "grad_norm": 0.22948284447193146, "learning_rate": 3.663920454606347e-05, "loss": 0.1118, "num_input_tokens_seen": 26611904, "step": 13835 }, { "epoch": 2.2577698017782852, "grad_norm": 5.509349346160889, "learning_rate": 3.6630515063683856e-05, "loss": 0.1105, "num_input_tokens_seen": 26619888, "step": 13840 }, { "epoch": 2.258585528999103, "grad_norm": 0.11843826621770859, "learning_rate": 3.662182378773267e-05, "loss": 0.0056, "num_input_tokens_seen": 26629936, "step": 13845 }, { "epoch": 2.25940125621992, "grad_norm": 1.2441051006317139, "learning_rate": 3.66131307195502e-05, "loss": 0.1042, "num_input_tokens_seen": 26639424, "step": 13850 }, { "epoch": 2.2602169834407375, "grad_norm": 17.8143253326416, "learning_rate": 3.6604435860477034e-05, "loss": 0.0862, "num_input_tokens_seen": 26648288, "step": 13855 }, { "epoch": 2.2610327106615546, "grad_norm": 2.684112310409546, "learning_rate": 3.6595739211854025e-05, "loss": 0.0859, "num_input_tokens_seen": 26659328, "step": 13860 }, { "epoch": 2.261848437882372, "grad_norm": 3.7406485080718994, "learning_rate": 3.658704077502231e-05, "loss": 0.181, "num_input_tokens_seen": 26667632, "step": 13865 }, { "epoch": 2.2626641651031894, "grad_norm": 0.024368004873394966, "learning_rate": 3.65783405513233e-05, "loss": 0.017, "num_input_tokens_seen": 26676640, "step": 13870 }, { "epoch": 2.263479892324007, "grad_norm": 0.41513964533805847, "learning_rate": 3.656963854209867e-05, "loss": 0.0968, "num_input_tokens_seen": 26685904, "step": 13875 }, { "epoch": 2.264295619544824, "grad_norm": 1.3457083702087402, "learning_rate": 3.656093474869038e-05, "loss": 0.1043, "num_input_tokens_seen": 26695632, "step": 13880 }, { "epoch": 2.2651113467656416, "grad_norm": 6.291446208953857, "learning_rate": 3.655222917244068e-05, "loss": 0.1514, "num_input_tokens_seen": 26704880, "step": 13885 }, { "epoch": 2.2659270739864588, "grad_norm": 2.2184150218963623, "learning_rate": 3.6543521814692054e-05, "loss": 0.0998, "num_input_tokens_seen": 26713872, "step": 13890 }, { "epoch": 2.2667428012072763, "grad_norm": 8.803646087646484, "learning_rate": 3.653481267678731e-05, "loss": 0.1238, "num_input_tokens_seen": 26723056, "step": 13895 }, { "epoch": 2.2675585284280935, "grad_norm": 0.2353961169719696, "learning_rate": 3.652610176006949e-05, "loss": 0.1063, "num_input_tokens_seen": 26732208, "step": 13900 }, { "epoch": 2.268374255648911, "grad_norm": 4.111473560333252, "learning_rate": 3.6517389065881925e-05, "loss": 0.0769, "num_input_tokens_seen": 26741344, "step": 13905 }, { "epoch": 2.269189982869728, "grad_norm": 0.29587432742118835, "learning_rate": 3.650867459556824e-05, "loss": 0.0152, "num_input_tokens_seen": 26750592, "step": 13910 }, { "epoch": 2.2700057100905457, "grad_norm": 10.206216812133789, "learning_rate": 3.64999583504723e-05, "loss": 0.1839, "num_input_tokens_seen": 26758928, "step": 13915 }, { "epoch": 2.270821437311363, "grad_norm": 6.0826592445373535, "learning_rate": 3.649124033193827e-05, "loss": 0.1524, "num_input_tokens_seen": 26768400, "step": 13920 }, { "epoch": 2.2716371645321805, "grad_norm": 1.1036072969436646, "learning_rate": 3.648252054131057e-05, "loss": 0.064, "num_input_tokens_seen": 26776832, "step": 13925 }, { "epoch": 2.272452891752998, "grad_norm": 0.0670141652226448, "learning_rate": 3.647379897993391e-05, "loss": 0.0473, "num_input_tokens_seen": 26786336, "step": 13930 }, { "epoch": 2.273268618973815, "grad_norm": 0.146382674574852, "learning_rate": 3.646507564915325e-05, "loss": 0.1703, "num_input_tokens_seen": 26794992, "step": 13935 }, { "epoch": 2.2740843461946323, "grad_norm": 3.360471248626709, "learning_rate": 3.645635055031385e-05, "loss": 0.3162, "num_input_tokens_seen": 26805328, "step": 13940 }, { "epoch": 2.27490007341545, "grad_norm": 0.2040025144815445, "learning_rate": 3.6447623684761224e-05, "loss": 0.0189, "num_input_tokens_seen": 26814480, "step": 13945 }, { "epoch": 2.2757158006362674, "grad_norm": 4.616547107696533, "learning_rate": 3.643889505384117e-05, "loss": 0.0753, "num_input_tokens_seen": 26824832, "step": 13950 }, { "epoch": 2.2765315278570846, "grad_norm": 0.22874362766742706, "learning_rate": 3.6430164658899744e-05, "loss": 0.0211, "num_input_tokens_seen": 26835104, "step": 13955 }, { "epoch": 2.2773472550779017, "grad_norm": 0.1104869470000267, "learning_rate": 3.642143250128329e-05, "loss": 0.0085, "num_input_tokens_seen": 26845248, "step": 13960 }, { "epoch": 2.2781629822987193, "grad_norm": 0.15071013569831848, "learning_rate": 3.641269858233841e-05, "loss": 0.0486, "num_input_tokens_seen": 26855456, "step": 13965 }, { "epoch": 2.278978709519537, "grad_norm": 0.08995061367750168, "learning_rate": 3.640396290341199e-05, "loss": 0.1744, "num_input_tokens_seen": 26865168, "step": 13970 }, { "epoch": 2.279794436740354, "grad_norm": 7.559288501739502, "learning_rate": 3.639522546585118e-05, "loss": 0.1156, "num_input_tokens_seen": 26874464, "step": 13975 }, { "epoch": 2.2806101639611716, "grad_norm": 2.855253219604492, "learning_rate": 3.6386486271003404e-05, "loss": 0.0471, "num_input_tokens_seen": 26883664, "step": 13980 }, { "epoch": 2.2814258911819887, "grad_norm": 2.5901381969451904, "learning_rate": 3.6377745320216346e-05, "loss": 0.1646, "num_input_tokens_seen": 26892976, "step": 13985 }, { "epoch": 2.2822416184028063, "grad_norm": 12.906902313232422, "learning_rate": 3.636900261483798e-05, "loss": 0.0302, "num_input_tokens_seen": 26902896, "step": 13990 }, { "epoch": 2.2830573456236234, "grad_norm": 0.047966811805963516, "learning_rate": 3.636025815621654e-05, "loss": 0.0452, "num_input_tokens_seen": 26913536, "step": 13995 }, { "epoch": 2.283873072844441, "grad_norm": 0.05011178180575371, "learning_rate": 3.635151194570054e-05, "loss": 0.0021, "num_input_tokens_seen": 26923408, "step": 14000 }, { "epoch": 2.283873072844441, "eval_loss": 0.16863511502742767, "eval_runtime": 80.5907, "eval_samples_per_second": 33.813, "eval_steps_per_second": 16.913, "num_input_tokens_seen": 26923408, "step": 14000 }, { "epoch": 2.284688800065258, "grad_norm": 15.645014762878418, "learning_rate": 3.634276398463873e-05, "loss": 0.157, "num_input_tokens_seen": 26933664, "step": 14005 }, { "epoch": 2.2855045272860757, "grad_norm": 0.09257246553897858, "learning_rate": 3.633401427438018e-05, "loss": 0.078, "num_input_tokens_seen": 26943312, "step": 14010 }, { "epoch": 2.286320254506893, "grad_norm": 0.0842473953962326, "learning_rate": 3.63252628162742e-05, "loss": 0.0272, "num_input_tokens_seen": 26952288, "step": 14015 }, { "epoch": 2.2871359817277104, "grad_norm": 0.06199311837553978, "learning_rate": 3.6316509611670364e-05, "loss": 0.0299, "num_input_tokens_seen": 26962448, "step": 14020 }, { "epoch": 2.2879517089485275, "grad_norm": 0.060907430946826935, "learning_rate": 3.630775466191854e-05, "loss": 0.0477, "num_input_tokens_seen": 26971168, "step": 14025 }, { "epoch": 2.288767436169345, "grad_norm": 0.14465093612670898, "learning_rate": 3.629899796836884e-05, "loss": 0.0482, "num_input_tokens_seen": 26980336, "step": 14030 }, { "epoch": 2.289583163390162, "grad_norm": 0.4020133912563324, "learning_rate": 3.6290239532371666e-05, "loss": 0.0712, "num_input_tokens_seen": 26990160, "step": 14035 }, { "epoch": 2.29039889061098, "grad_norm": 21.011812210083008, "learning_rate": 3.628147935527767e-05, "loss": 0.0795, "num_input_tokens_seen": 26999392, "step": 14040 }, { "epoch": 2.291214617831797, "grad_norm": 0.5432901978492737, "learning_rate": 3.627271743843779e-05, "loss": 0.0673, "num_input_tokens_seen": 27007968, "step": 14045 }, { "epoch": 2.2920303450526145, "grad_norm": 0.3035838305950165, "learning_rate": 3.626395378320321e-05, "loss": 0.0047, "num_input_tokens_seen": 27017952, "step": 14050 }, { "epoch": 2.2928460722734316, "grad_norm": 0.1971852332353592, "learning_rate": 3.625518839092541e-05, "loss": 0.1122, "num_input_tokens_seen": 27027264, "step": 14055 }, { "epoch": 2.293661799494249, "grad_norm": 22.038002014160156, "learning_rate": 3.624642126295612e-05, "loss": 0.1019, "num_input_tokens_seen": 27037648, "step": 14060 }, { "epoch": 2.2944775267150663, "grad_norm": 0.07759211957454681, "learning_rate": 3.6237652400647345e-05, "loss": 0.0534, "num_input_tokens_seen": 27047056, "step": 14065 }, { "epoch": 2.295293253935884, "grad_norm": 0.24205002188682556, "learning_rate": 3.622888180535134e-05, "loss": 0.0073, "num_input_tokens_seen": 27056240, "step": 14070 }, { "epoch": 2.296108981156701, "grad_norm": 0.2805420458316803, "learning_rate": 3.6220109478420655e-05, "loss": 0.1271, "num_input_tokens_seen": 27066064, "step": 14075 }, { "epoch": 2.2969247083775186, "grad_norm": 2.8370540142059326, "learning_rate": 3.6211335421208084e-05, "loss": 0.0743, "num_input_tokens_seen": 27076064, "step": 14080 }, { "epoch": 2.2977404355983357, "grad_norm": 11.434503555297852, "learning_rate": 3.62025596350667e-05, "loss": 0.1122, "num_input_tokens_seen": 27085744, "step": 14085 }, { "epoch": 2.2985561628191533, "grad_norm": 3.028992176055908, "learning_rate": 3.619378212134984e-05, "loss": 0.1525, "num_input_tokens_seen": 27095760, "step": 14090 }, { "epoch": 2.2993718900399704, "grad_norm": 4.766038417816162, "learning_rate": 3.618500288141111e-05, "loss": 0.0305, "num_input_tokens_seen": 27105952, "step": 14095 }, { "epoch": 2.300187617260788, "grad_norm": 8.264909744262695, "learning_rate": 3.617622191660438e-05, "loss": 0.0854, "num_input_tokens_seen": 27115184, "step": 14100 }, { "epoch": 2.3010033444816056, "grad_norm": 22.20020866394043, "learning_rate": 3.616743922828377e-05, "loss": 0.0686, "num_input_tokens_seen": 27125120, "step": 14105 }, { "epoch": 2.3018190717024227, "grad_norm": 0.05316278338432312, "learning_rate": 3.615865481780371e-05, "loss": 0.0251, "num_input_tokens_seen": 27134128, "step": 14110 }, { "epoch": 2.30263479892324, "grad_norm": 0.1368291676044464, "learning_rate": 3.614986868651883e-05, "loss": 0.0072, "num_input_tokens_seen": 27143904, "step": 14115 }, { "epoch": 2.3034505261440574, "grad_norm": 2.859710454940796, "learning_rate": 3.614108083578409e-05, "loss": 0.0543, "num_input_tokens_seen": 27154112, "step": 14120 }, { "epoch": 2.304266253364875, "grad_norm": 0.12053476274013519, "learning_rate": 3.613229126695467e-05, "loss": 0.1827, "num_input_tokens_seen": 27162736, "step": 14125 }, { "epoch": 2.305081980585692, "grad_norm": 6.37370491027832, "learning_rate": 3.612349998138605e-05, "loss": 0.2436, "num_input_tokens_seen": 27172016, "step": 14130 }, { "epoch": 2.3058977078065093, "grad_norm": 0.7090364694595337, "learning_rate": 3.6114706980433946e-05, "loss": 0.0559, "num_input_tokens_seen": 27181696, "step": 14135 }, { "epoch": 2.306713435027327, "grad_norm": 0.4082042872905731, "learning_rate": 3.610591226545435e-05, "loss": 0.0105, "num_input_tokens_seen": 27190848, "step": 14140 }, { "epoch": 2.3075291622481444, "grad_norm": 5.918069362640381, "learning_rate": 3.6097115837803505e-05, "loss": 0.0425, "num_input_tokens_seen": 27200608, "step": 14145 }, { "epoch": 2.3083448894689615, "grad_norm": 0.06597770005464554, "learning_rate": 3.608831769883795e-05, "loss": 0.0292, "num_input_tokens_seen": 27211392, "step": 14150 }, { "epoch": 2.309160616689779, "grad_norm": 23.22001075744629, "learning_rate": 3.607951784991446e-05, "loss": 0.1042, "num_input_tokens_seen": 27220432, "step": 14155 }, { "epoch": 2.3099763439105963, "grad_norm": 19.264368057250977, "learning_rate": 3.6070716292390085e-05, "loss": 0.1631, "num_input_tokens_seen": 27229120, "step": 14160 }, { "epoch": 2.310792071131414, "grad_norm": 0.0609859824180603, "learning_rate": 3.606191302762213e-05, "loss": 0.0073, "num_input_tokens_seen": 27239312, "step": 14165 }, { "epoch": 2.311607798352231, "grad_norm": 8.507192611694336, "learning_rate": 3.605310805696818e-05, "loss": 0.1752, "num_input_tokens_seen": 27248464, "step": 14170 }, { "epoch": 2.3124235255730485, "grad_norm": 0.24376064538955688, "learning_rate": 3.6044301381786067e-05, "loss": 0.005, "num_input_tokens_seen": 27258096, "step": 14175 }, { "epoch": 2.3132392527938657, "grad_norm": 3.6274607181549072, "learning_rate": 3.6035493003433883e-05, "loss": 0.1998, "num_input_tokens_seen": 27269424, "step": 14180 }, { "epoch": 2.3140549800146832, "grad_norm": 16.671241760253906, "learning_rate": 3.6026682923269994e-05, "loss": 0.0447, "num_input_tokens_seen": 27278592, "step": 14185 }, { "epoch": 2.3148707072355004, "grad_norm": 4.5282368659973145, "learning_rate": 3.6017871142653034e-05, "loss": 0.0412, "num_input_tokens_seen": 27288880, "step": 14190 }, { "epoch": 2.315686434456318, "grad_norm": 12.028268814086914, "learning_rate": 3.600905766294189e-05, "loss": 0.0347, "num_input_tokens_seen": 27299728, "step": 14195 }, { "epoch": 2.316502161677135, "grad_norm": 1.0615499019622803, "learning_rate": 3.60002424854957e-05, "loss": 0.0934, "num_input_tokens_seen": 27309344, "step": 14200 }, { "epoch": 2.316502161677135, "eval_loss": 0.17132551968097687, "eval_runtime": 80.5692, "eval_samples_per_second": 33.822, "eval_steps_per_second": 16.917, "num_input_tokens_seen": 27309344, "step": 14200 }, { "epoch": 2.3173178888979526, "grad_norm": 0.0646098181605339, "learning_rate": 3.5991425611673876e-05, "loss": 0.073, "num_input_tokens_seen": 27319936, "step": 14205 }, { "epoch": 2.31813361611877, "grad_norm": 3.932451009750366, "learning_rate": 3.5982607042836105e-05, "loss": 0.2091, "num_input_tokens_seen": 27329680, "step": 14210 }, { "epoch": 2.3189493433395874, "grad_norm": 1.910987377166748, "learning_rate": 3.597378678034231e-05, "loss": 0.1155, "num_input_tokens_seen": 27340576, "step": 14215 }, { "epoch": 2.3197650705604045, "grad_norm": 0.09607714414596558, "learning_rate": 3.596496482555269e-05, "loss": 0.0919, "num_input_tokens_seen": 27351056, "step": 14220 }, { "epoch": 2.320580797781222, "grad_norm": 0.17084263265132904, "learning_rate": 3.595614117982769e-05, "loss": 0.0787, "num_input_tokens_seen": 27361232, "step": 14225 }, { "epoch": 2.321396525002039, "grad_norm": 2.9486751556396484, "learning_rate": 3.594731584452805e-05, "loss": 0.0187, "num_input_tokens_seen": 27370096, "step": 14230 }, { "epoch": 2.3222122522228568, "grad_norm": 0.4807453453540802, "learning_rate": 3.593848882101472e-05, "loss": 0.0075, "num_input_tokens_seen": 27379840, "step": 14235 }, { "epoch": 2.323027979443674, "grad_norm": 4.876959800720215, "learning_rate": 3.592966011064896e-05, "loss": 0.0677, "num_input_tokens_seen": 27388928, "step": 14240 }, { "epoch": 2.3238437066644915, "grad_norm": 16.629993438720703, "learning_rate": 3.592082971479226e-05, "loss": 0.1232, "num_input_tokens_seen": 27398288, "step": 14245 }, { "epoch": 2.3246594338853086, "grad_norm": 15.432682037353516, "learning_rate": 3.5911997634806385e-05, "loss": 0.0844, "num_input_tokens_seen": 27408960, "step": 14250 }, { "epoch": 2.325475161106126, "grad_norm": 3.673947334289551, "learning_rate": 3.5903163872053336e-05, "loss": 0.2123, "num_input_tokens_seen": 27419296, "step": 14255 }, { "epoch": 2.3262908883269433, "grad_norm": 0.5892764329910278, "learning_rate": 3.58943284278954e-05, "loss": 0.0776, "num_input_tokens_seen": 27428912, "step": 14260 }, { "epoch": 2.327106615547761, "grad_norm": 16.040721893310547, "learning_rate": 3.588549130369512e-05, "loss": 0.079, "num_input_tokens_seen": 27436144, "step": 14265 }, { "epoch": 2.327922342768578, "grad_norm": 0.11304785311222076, "learning_rate": 3.5876652500815274e-05, "loss": 0.0192, "num_input_tokens_seen": 27446272, "step": 14270 }, { "epoch": 2.3287380699893956, "grad_norm": 0.08446386456489563, "learning_rate": 3.586781202061894e-05, "loss": 0.0749, "num_input_tokens_seen": 27455536, "step": 14275 }, { "epoch": 2.3295537972102127, "grad_norm": 0.17964202165603638, "learning_rate": 3.585896986446942e-05, "loss": 0.0599, "num_input_tokens_seen": 27465584, "step": 14280 }, { "epoch": 2.3303695244310303, "grad_norm": 1.0614326000213623, "learning_rate": 3.585012603373028e-05, "loss": 0.0567, "num_input_tokens_seen": 27475280, "step": 14285 }, { "epoch": 2.3311852516518474, "grad_norm": 1.99945068359375, "learning_rate": 3.584128052976535e-05, "loss": 0.0143, "num_input_tokens_seen": 27485888, "step": 14290 }, { "epoch": 2.332000978872665, "grad_norm": 0.04088766872882843, "learning_rate": 3.5832433353938724e-05, "loss": 0.0044, "num_input_tokens_seen": 27496976, "step": 14295 }, { "epoch": 2.3328167060934826, "grad_norm": 1.0146706104278564, "learning_rate": 3.5823584507614746e-05, "loss": 0.0238, "num_input_tokens_seen": 27505424, "step": 14300 }, { "epoch": 2.3336324333142997, "grad_norm": 0.16131781041622162, "learning_rate": 3.581473399215802e-05, "loss": 0.0921, "num_input_tokens_seen": 27514336, "step": 14305 }, { "epoch": 2.334448160535117, "grad_norm": 0.025673864409327507, "learning_rate": 3.580588180893341e-05, "loss": 0.1117, "num_input_tokens_seen": 27523456, "step": 14310 }, { "epoch": 2.3352638877559344, "grad_norm": 0.2151380032300949, "learning_rate": 3.579702795930602e-05, "loss": 0.0087, "num_input_tokens_seen": 27531904, "step": 14315 }, { "epoch": 2.336079614976752, "grad_norm": 0.10722669214010239, "learning_rate": 3.578817244464125e-05, "loss": 0.0591, "num_input_tokens_seen": 27541776, "step": 14320 }, { "epoch": 2.336895342197569, "grad_norm": 6.533113956451416, "learning_rate": 3.577931526630471e-05, "loss": 0.2779, "num_input_tokens_seen": 27551648, "step": 14325 }, { "epoch": 2.3377110694183862, "grad_norm": 0.15942804515361786, "learning_rate": 3.577045642566229e-05, "loss": 0.033, "num_input_tokens_seen": 27560480, "step": 14330 }, { "epoch": 2.338526796639204, "grad_norm": 0.11167322844266891, "learning_rate": 3.576159592408014e-05, "loss": 0.0463, "num_input_tokens_seen": 27570624, "step": 14335 }, { "epoch": 2.3393425238600214, "grad_norm": 29.898887634277344, "learning_rate": 3.575273376292466e-05, "loss": 0.0771, "num_input_tokens_seen": 27580128, "step": 14340 }, { "epoch": 2.3401582510808385, "grad_norm": 0.18845194578170776, "learning_rate": 3.574386994356251e-05, "loss": 0.0309, "num_input_tokens_seen": 27589648, "step": 14345 }, { "epoch": 2.340973978301656, "grad_norm": 0.019807366654276848, "learning_rate": 3.573500446736059e-05, "loss": 0.0028, "num_input_tokens_seen": 27598752, "step": 14350 }, { "epoch": 2.3417897055224732, "grad_norm": 0.5222815275192261, "learning_rate": 3.5726137335686094e-05, "loss": 0.0276, "num_input_tokens_seen": 27607808, "step": 14355 }, { "epoch": 2.342605432743291, "grad_norm": 4.09986686706543, "learning_rate": 3.571726854990642e-05, "loss": 0.2376, "num_input_tokens_seen": 27618560, "step": 14360 }, { "epoch": 2.343421159964108, "grad_norm": 0.018806923180818558, "learning_rate": 3.570839811138925e-05, "loss": 0.0752, "num_input_tokens_seen": 27629888, "step": 14365 }, { "epoch": 2.3442368871849255, "grad_norm": 0.4475741684436798, "learning_rate": 3.569952602150252e-05, "loss": 0.163, "num_input_tokens_seen": 27639568, "step": 14370 }, { "epoch": 2.3450526144057426, "grad_norm": 0.3179117441177368, "learning_rate": 3.569065228161442e-05, "loss": 0.0041, "num_input_tokens_seen": 27649760, "step": 14375 }, { "epoch": 2.34586834162656, "grad_norm": 0.21944041550159454, "learning_rate": 3.5681776893093395e-05, "loss": 0.0773, "num_input_tokens_seen": 27660640, "step": 14380 }, { "epoch": 2.3466840688473773, "grad_norm": 0.2104196548461914, "learning_rate": 3.5672899857308134e-05, "loss": 0.0518, "num_input_tokens_seen": 27670304, "step": 14385 }, { "epoch": 2.347499796068195, "grad_norm": 0.09074784070253372, "learning_rate": 3.566402117562759e-05, "loss": 0.2015, "num_input_tokens_seen": 27680880, "step": 14390 }, { "epoch": 2.348315523289012, "grad_norm": 10.401070594787598, "learning_rate": 3.565514084942097e-05, "loss": 0.238, "num_input_tokens_seen": 27690128, "step": 14395 }, { "epoch": 2.3491312505098296, "grad_norm": 4.894956588745117, "learning_rate": 3.564625888005773e-05, "loss": 0.1286, "num_input_tokens_seen": 27698752, "step": 14400 }, { "epoch": 2.3491312505098296, "eval_loss": 0.1712968945503235, "eval_runtime": 80.5626, "eval_samples_per_second": 33.825, "eval_steps_per_second": 16.919, "num_input_tokens_seen": 27698752, "step": 14400 }, { "epoch": 2.3499469777306468, "grad_norm": 0.18049581348896027, "learning_rate": 3.563737526890759e-05, "loss": 0.0054, "num_input_tokens_seen": 27708048, "step": 14405 }, { "epoch": 2.3507627049514643, "grad_norm": 5.464495658874512, "learning_rate": 3.562849001734049e-05, "loss": 0.2001, "num_input_tokens_seen": 27716912, "step": 14410 }, { "epoch": 2.3515784321722815, "grad_norm": 0.06043294817209244, "learning_rate": 3.561960312672667e-05, "loss": 0.0382, "num_input_tokens_seen": 27726640, "step": 14415 }, { "epoch": 2.352394159393099, "grad_norm": 9.546494483947754, "learning_rate": 3.5610714598436596e-05, "loss": 0.0511, "num_input_tokens_seen": 27736976, "step": 14420 }, { "epoch": 2.353209886613916, "grad_norm": 0.04781101271510124, "learning_rate": 3.5601824433840986e-05, "loss": 0.0901, "num_input_tokens_seen": 27746128, "step": 14425 }, { "epoch": 2.3540256138347337, "grad_norm": 3.6936869621276855, "learning_rate": 3.559293263431082e-05, "loss": 0.2108, "num_input_tokens_seen": 27756576, "step": 14430 }, { "epoch": 2.354841341055551, "grad_norm": 0.18821083009243011, "learning_rate": 3.558403920121732e-05, "loss": 0.1485, "num_input_tokens_seen": 27765344, "step": 14435 }, { "epoch": 2.3556570682763684, "grad_norm": 0.13639993965625763, "learning_rate": 3.557514413593197e-05, "loss": 0.0356, "num_input_tokens_seen": 27775824, "step": 14440 }, { "epoch": 2.3564727954971856, "grad_norm": 11.83562183380127, "learning_rate": 3.55662474398265e-05, "loss": 0.035, "num_input_tokens_seen": 27785184, "step": 14445 }, { "epoch": 2.357288522718003, "grad_norm": 3.40201735496521, "learning_rate": 3.555734911427288e-05, "loss": 0.1014, "num_input_tokens_seen": 27793456, "step": 14450 }, { "epoch": 2.3581042499388203, "grad_norm": 0.2605929970741272, "learning_rate": 3.5548449160643363e-05, "loss": 0.0154, "num_input_tokens_seen": 27804288, "step": 14455 }, { "epoch": 2.358919977159638, "grad_norm": 0.6635530591011047, "learning_rate": 3.553954758031043e-05, "loss": 0.0554, "num_input_tokens_seen": 27815200, "step": 14460 }, { "epoch": 2.359735704380455, "grad_norm": 0.09983836114406586, "learning_rate": 3.5530644374646815e-05, "loss": 0.0953, "num_input_tokens_seen": 27825072, "step": 14465 }, { "epoch": 2.3605514316012726, "grad_norm": 3.9144043922424316, "learning_rate": 3.552173954502549e-05, "loss": 0.1008, "num_input_tokens_seen": 27834512, "step": 14470 }, { "epoch": 2.36136715882209, "grad_norm": 6.7046918869018555, "learning_rate": 3.55128330928197e-05, "loss": 0.0654, "num_input_tokens_seen": 27843232, "step": 14475 }, { "epoch": 2.3621828860429073, "grad_norm": 0.028566531836986542, "learning_rate": 3.550392501940294e-05, "loss": 0.1815, "num_input_tokens_seen": 27853456, "step": 14480 }, { "epoch": 2.3629986132637244, "grad_norm": 16.64316749572754, "learning_rate": 3.5495015326148945e-05, "loss": 0.1092, "num_input_tokens_seen": 27863680, "step": 14485 }, { "epoch": 2.363814340484542, "grad_norm": 0.17268559336662292, "learning_rate": 3.548610401443169e-05, "loss": 0.0046, "num_input_tokens_seen": 27872016, "step": 14490 }, { "epoch": 2.3646300677053596, "grad_norm": 2.382155656814575, "learning_rate": 3.547719108562543e-05, "loss": 0.0242, "num_input_tokens_seen": 27882048, "step": 14495 }, { "epoch": 2.3654457949261767, "grad_norm": 2.9288747310638428, "learning_rate": 3.546827654110464e-05, "loss": 0.1447, "num_input_tokens_seen": 27891040, "step": 14500 }, { "epoch": 2.366261522146994, "grad_norm": 1.1980451345443726, "learning_rate": 3.545936038224405e-05, "loss": 0.119, "num_input_tokens_seen": 27899872, "step": 14505 }, { "epoch": 2.3670772493678114, "grad_norm": 0.9690814018249512, "learning_rate": 3.545044261041864e-05, "loss": 0.0121, "num_input_tokens_seen": 27908848, "step": 14510 }, { "epoch": 2.367892976588629, "grad_norm": 0.07769708335399628, "learning_rate": 3.5441523227003657e-05, "loss": 0.022, "num_input_tokens_seen": 27918800, "step": 14515 }, { "epoch": 2.368708703809446, "grad_norm": 11.314078330993652, "learning_rate": 3.543260223337459e-05, "loss": 0.0864, "num_input_tokens_seen": 27928688, "step": 14520 }, { "epoch": 2.3695244310302637, "grad_norm": 2.828523635864258, "learning_rate": 3.542367963090714e-05, "loss": 0.0767, "num_input_tokens_seen": 27937536, "step": 14525 }, { "epoch": 2.370340158251081, "grad_norm": 0.31727510690689087, "learning_rate": 3.5414755420977295e-05, "loss": 0.0095, "num_input_tokens_seen": 27946192, "step": 14530 }, { "epoch": 2.3711558854718984, "grad_norm": 0.23679079115390778, "learning_rate": 3.54058296049613e-05, "loss": 0.0162, "num_input_tokens_seen": 27954336, "step": 14535 }, { "epoch": 2.3719716126927155, "grad_norm": 0.1319563388824463, "learning_rate": 3.53969021842356e-05, "loss": 0.0822, "num_input_tokens_seen": 27963840, "step": 14540 }, { "epoch": 2.372787339913533, "grad_norm": 22.525436401367188, "learning_rate": 3.5387973160176926e-05, "loss": 0.0935, "num_input_tokens_seen": 27973344, "step": 14545 }, { "epoch": 2.37360306713435, "grad_norm": 11.56708812713623, "learning_rate": 3.537904253416224e-05, "loss": 0.0913, "num_input_tokens_seen": 27983200, "step": 14550 }, { "epoch": 2.374418794355168, "grad_norm": 12.298150062561035, "learning_rate": 3.537011030756878e-05, "loss": 0.0871, "num_input_tokens_seen": 27992576, "step": 14555 }, { "epoch": 2.375234521575985, "grad_norm": 2.3362905979156494, "learning_rate": 3.536117648177399e-05, "loss": 0.0719, "num_input_tokens_seen": 28002128, "step": 14560 }, { "epoch": 2.3760502487968025, "grad_norm": 0.19336026906967163, "learning_rate": 3.535224105815558e-05, "loss": 0.0729, "num_input_tokens_seen": 28012160, "step": 14565 }, { "epoch": 2.3768659760176196, "grad_norm": 3.8241515159606934, "learning_rate": 3.5343304038091494e-05, "loss": 0.1153, "num_input_tokens_seen": 28021264, "step": 14570 }, { "epoch": 2.377681703238437, "grad_norm": 4.054490089416504, "learning_rate": 3.5334365422959955e-05, "loss": 0.0856, "num_input_tokens_seen": 28031472, "step": 14575 }, { "epoch": 2.3784974304592543, "grad_norm": 0.12356970459222794, "learning_rate": 3.5325425214139396e-05, "loss": 0.0628, "num_input_tokens_seen": 28041232, "step": 14580 }, { "epoch": 2.379313157680072, "grad_norm": 4.584226608276367, "learning_rate": 3.531648341300851e-05, "loss": 0.0806, "num_input_tokens_seen": 28051600, "step": 14585 }, { "epoch": 2.380128884900889, "grad_norm": 0.28714680671691895, "learning_rate": 3.530754002094623e-05, "loss": 0.0804, "num_input_tokens_seen": 28061888, "step": 14590 }, { "epoch": 2.3809446121217066, "grad_norm": 8.824941635131836, "learning_rate": 3.529859503933175e-05, "loss": 0.1169, "num_input_tokens_seen": 28072160, "step": 14595 }, { "epoch": 2.3817603393425237, "grad_norm": 8.955395698547363, "learning_rate": 3.52896484695445e-05, "loss": 0.0139, "num_input_tokens_seen": 28082208, "step": 14600 }, { "epoch": 2.3817603393425237, "eval_loss": 0.16269946098327637, "eval_runtime": 80.5452, "eval_samples_per_second": 33.832, "eval_steps_per_second": 16.922, "num_input_tokens_seen": 28082208, "step": 14600 }, { "epoch": 2.3825760665633413, "grad_norm": 0.04703907668590546, "learning_rate": 3.528070031296414e-05, "loss": 0.0048, "num_input_tokens_seen": 28091728, "step": 14605 }, { "epoch": 2.3833917937841584, "grad_norm": 0.04841151088476181, "learning_rate": 3.5271750570970605e-05, "loss": 0.0057, "num_input_tokens_seen": 28100656, "step": 14610 }, { "epoch": 2.384207521004976, "grad_norm": 0.2338748574256897, "learning_rate": 3.526279924494405e-05, "loss": 0.0045, "num_input_tokens_seen": 28109216, "step": 14615 }, { "epoch": 2.385023248225793, "grad_norm": 0.23440362513065338, "learning_rate": 3.5253846336264874e-05, "loss": 0.0042, "num_input_tokens_seen": 28118752, "step": 14620 }, { "epoch": 2.3858389754466107, "grad_norm": 2.4554007053375244, "learning_rate": 3.5244891846313736e-05, "loss": 0.0484, "num_input_tokens_seen": 28128528, "step": 14625 }, { "epoch": 2.386654702667428, "grad_norm": 34.71001052856445, "learning_rate": 3.5235935776471527e-05, "loss": 0.0392, "num_input_tokens_seen": 28139280, "step": 14630 }, { "epoch": 2.3874704298882454, "grad_norm": 0.2856120765209198, "learning_rate": 3.522697812811939e-05, "loss": 0.1148, "num_input_tokens_seen": 28148384, "step": 14635 }, { "epoch": 2.3882861571090626, "grad_norm": 4.341506004333496, "learning_rate": 3.521801890263871e-05, "loss": 0.2484, "num_input_tokens_seen": 28157968, "step": 14640 }, { "epoch": 2.38910188432988, "grad_norm": 0.017109747976064682, "learning_rate": 3.5209058101411114e-05, "loss": 0.0639, "num_input_tokens_seen": 28167712, "step": 14645 }, { "epoch": 2.3899176115506973, "grad_norm": 0.2393866777420044, "learning_rate": 3.520009572581845e-05, "loss": 0.0976, "num_input_tokens_seen": 28177680, "step": 14650 }, { "epoch": 2.390733338771515, "grad_norm": 0.055381182581186295, "learning_rate": 3.519113177724285e-05, "loss": 0.0814, "num_input_tokens_seen": 28187664, "step": 14655 }, { "epoch": 2.391549065992332, "grad_norm": 0.045955684036016464, "learning_rate": 3.5182166257066656e-05, "loss": 0.1131, "num_input_tokens_seen": 28197968, "step": 14660 }, { "epoch": 2.3923647932131495, "grad_norm": 61.29153823852539, "learning_rate": 3.517319916667247e-05, "loss": 0.1239, "num_input_tokens_seen": 28207216, "step": 14665 }, { "epoch": 2.393180520433967, "grad_norm": 4.0390424728393555, "learning_rate": 3.516423050744313e-05, "loss": 0.1441, "num_input_tokens_seen": 28218320, "step": 14670 }, { "epoch": 2.3939962476547842, "grad_norm": 3.3544230461120605, "learning_rate": 3.5155260280761704e-05, "loss": 0.0672, "num_input_tokens_seen": 28228512, "step": 14675 }, { "epoch": 2.3948119748756014, "grad_norm": 0.1413489282131195, "learning_rate": 3.514628848801154e-05, "loss": 0.0984, "num_input_tokens_seen": 28237408, "step": 14680 }, { "epoch": 2.395627702096419, "grad_norm": 0.1232454776763916, "learning_rate": 3.5137315130576174e-05, "loss": 0.0494, "num_input_tokens_seen": 28248288, "step": 14685 }, { "epoch": 2.3964434293172365, "grad_norm": 0.027249954640865326, "learning_rate": 3.512834020983942e-05, "loss": 0.225, "num_input_tokens_seen": 28257600, "step": 14690 }, { "epoch": 2.3972591565380537, "grad_norm": 1.3823583126068115, "learning_rate": 3.5119363727185334e-05, "loss": 0.1128, "num_input_tokens_seen": 28267840, "step": 14695 }, { "epoch": 2.398074883758871, "grad_norm": 2.7858450412750244, "learning_rate": 3.511038568399819e-05, "loss": 0.1977, "num_input_tokens_seen": 28276880, "step": 14700 }, { "epoch": 2.3988906109796884, "grad_norm": 0.503020703792572, "learning_rate": 3.510140608166251e-05, "loss": 0.037, "num_input_tokens_seen": 28285776, "step": 14705 }, { "epoch": 2.399706338200506, "grad_norm": 3.08534836769104, "learning_rate": 3.509242492156308e-05, "loss": 0.0771, "num_input_tokens_seen": 28296048, "step": 14710 }, { "epoch": 2.400522065421323, "grad_norm": 9.831576347351074, "learning_rate": 3.5083442205084896e-05, "loss": 0.0565, "num_input_tokens_seen": 28305616, "step": 14715 }, { "epoch": 2.4013377926421406, "grad_norm": 4.85479211807251, "learning_rate": 3.507445793361321e-05, "loss": 0.1427, "num_input_tokens_seen": 28315056, "step": 14720 }, { "epoch": 2.4021535198629578, "grad_norm": 0.4432201683521271, "learning_rate": 3.5065472108533505e-05, "loss": 0.0096, "num_input_tokens_seen": 28325104, "step": 14725 }, { "epoch": 2.4029692470837754, "grad_norm": 0.0927424356341362, "learning_rate": 3.5056484731231504e-05, "loss": 0.1154, "num_input_tokens_seen": 28335040, "step": 14730 }, { "epoch": 2.4037849743045925, "grad_norm": 0.8610493540763855, "learning_rate": 3.504749580309319e-05, "loss": 0.0476, "num_input_tokens_seen": 28344352, "step": 14735 }, { "epoch": 2.40460070152541, "grad_norm": 0.1331104338169098, "learning_rate": 3.5038505325504753e-05, "loss": 0.114, "num_input_tokens_seen": 28354064, "step": 14740 }, { "epoch": 2.405416428746227, "grad_norm": 0.16152074933052063, "learning_rate": 3.502951329985264e-05, "loss": 0.0622, "num_input_tokens_seen": 28363168, "step": 14745 }, { "epoch": 2.4062321559670448, "grad_norm": 0.07123187184333801, "learning_rate": 3.502051972752354e-05, "loss": 0.0288, "num_input_tokens_seen": 28372544, "step": 14750 }, { "epoch": 2.407047883187862, "grad_norm": 0.07864515483379364, "learning_rate": 3.5011524609904374e-05, "loss": 0.087, "num_input_tokens_seen": 28383152, "step": 14755 }, { "epoch": 2.4078636104086795, "grad_norm": 7.181223392486572, "learning_rate": 3.50025279483823e-05, "loss": 0.0874, "num_input_tokens_seen": 28391712, "step": 14760 }, { "epoch": 2.4086793376294966, "grad_norm": 0.32262009382247925, "learning_rate": 3.499352974434472e-05, "loss": 0.1544, "num_input_tokens_seen": 28401168, "step": 14765 }, { "epoch": 2.409495064850314, "grad_norm": 4.087881088256836, "learning_rate": 3.498452999917926e-05, "loss": 0.1815, "num_input_tokens_seen": 28410304, "step": 14770 }, { "epoch": 2.4103107920711313, "grad_norm": 9.209386825561523, "learning_rate": 3.4975528714273795e-05, "loss": 0.047, "num_input_tokens_seen": 28420192, "step": 14775 }, { "epoch": 2.411126519291949, "grad_norm": 0.34848275780677795, "learning_rate": 3.4966525891016454e-05, "loss": 0.0187, "num_input_tokens_seen": 28430896, "step": 14780 }, { "epoch": 2.411942246512766, "grad_norm": 2.3965518474578857, "learning_rate": 3.495752153079557e-05, "loss": 0.0259, "num_input_tokens_seen": 28441104, "step": 14785 }, { "epoch": 2.4127579737335836, "grad_norm": 0.477152556180954, "learning_rate": 3.494851563499974e-05, "loss": 0.0677, "num_input_tokens_seen": 28450000, "step": 14790 }, { "epoch": 2.4135737009544007, "grad_norm": 0.29894500970840454, "learning_rate": 3.493950820501777e-05, "loss": 0.0062, "num_input_tokens_seen": 28458672, "step": 14795 }, { "epoch": 2.4143894281752183, "grad_norm": 4.1251935958862305, "learning_rate": 3.493049924223872e-05, "loss": 0.163, "num_input_tokens_seen": 28468576, "step": 14800 }, { "epoch": 2.4143894281752183, "eval_loss": 0.15688742697238922, "eval_runtime": 80.6377, "eval_samples_per_second": 33.793, "eval_steps_per_second": 16.903, "num_input_tokens_seen": 28468576, "step": 14800 }, { "epoch": 2.4152051553960354, "grad_norm": 0.7337701916694641, "learning_rate": 3.49214887480519e-05, "loss": 0.1416, "num_input_tokens_seen": 28477920, "step": 14805 }, { "epoch": 2.416020882616853, "grad_norm": 5.623737335205078, "learning_rate": 3.4912476723846834e-05, "loss": 0.1079, "num_input_tokens_seen": 28486848, "step": 14810 }, { "epoch": 2.41683660983767, "grad_norm": 0.3262381851673126, "learning_rate": 3.490346317101328e-05, "loss": 0.0695, "num_input_tokens_seen": 28496624, "step": 14815 }, { "epoch": 2.4176523370584877, "grad_norm": 7.1893181800842285, "learning_rate": 3.4894448090941266e-05, "loss": 0.1091, "num_input_tokens_seen": 28506032, "step": 14820 }, { "epoch": 2.418468064279305, "grad_norm": 0.7399225831031799, "learning_rate": 3.488543148502101e-05, "loss": 0.0578, "num_input_tokens_seen": 28515936, "step": 14825 }, { "epoch": 2.4192837915001224, "grad_norm": 12.514142036437988, "learning_rate": 3.487641335464299e-05, "loss": 0.0413, "num_input_tokens_seen": 28523696, "step": 14830 }, { "epoch": 2.4200995187209395, "grad_norm": 0.10344403237104416, "learning_rate": 3.4867393701197914e-05, "loss": 0.1145, "num_input_tokens_seen": 28534384, "step": 14835 }, { "epoch": 2.420915245941757, "grad_norm": 0.24388010799884796, "learning_rate": 3.485837252607673e-05, "loss": 0.0277, "num_input_tokens_seen": 28544432, "step": 14840 }, { "epoch": 2.4217309731625742, "grad_norm": 2.4944679737091064, "learning_rate": 3.4849349830670615e-05, "loss": 0.0924, "num_input_tokens_seen": 28553872, "step": 14845 }, { "epoch": 2.422546700383392, "grad_norm": 0.7733018398284912, "learning_rate": 3.4840325616370976e-05, "loss": 0.0279, "num_input_tokens_seen": 28562848, "step": 14850 }, { "epoch": 2.423362427604209, "grad_norm": 8.521966934204102, "learning_rate": 3.483129988456947e-05, "loss": 0.05, "num_input_tokens_seen": 28571840, "step": 14855 }, { "epoch": 2.4241781548250265, "grad_norm": 0.08600195497274399, "learning_rate": 3.482227263665797e-05, "loss": 0.0305, "num_input_tokens_seen": 28580336, "step": 14860 }, { "epoch": 2.424993882045844, "grad_norm": 7.698124885559082, "learning_rate": 3.48132438740286e-05, "loss": 0.0531, "num_input_tokens_seen": 28591280, "step": 14865 }, { "epoch": 2.4258096092666612, "grad_norm": 7.147644519805908, "learning_rate": 3.48042135980737e-05, "loss": 0.0713, "num_input_tokens_seen": 28600768, "step": 14870 }, { "epoch": 2.4266253364874784, "grad_norm": 5.404852390289307, "learning_rate": 3.479518181018586e-05, "loss": 0.1477, "num_input_tokens_seen": 28610544, "step": 14875 }, { "epoch": 2.427441063708296, "grad_norm": 13.449983596801758, "learning_rate": 3.4786148511757886e-05, "loss": 0.0488, "num_input_tokens_seen": 28621136, "step": 14880 }, { "epoch": 2.4282567909291135, "grad_norm": 0.034942325204610825, "learning_rate": 3.477711370418284e-05, "loss": 0.0135, "num_input_tokens_seen": 28632544, "step": 14885 }, { "epoch": 2.4290725181499306, "grad_norm": 10.008339881896973, "learning_rate": 3.476807738885399e-05, "loss": 0.0779, "num_input_tokens_seen": 28642384, "step": 14890 }, { "epoch": 2.429888245370748, "grad_norm": 5.903133869171143, "learning_rate": 3.475903956716485e-05, "loss": 0.1253, "num_input_tokens_seen": 28651632, "step": 14895 }, { "epoch": 2.4307039725915653, "grad_norm": 0.025541475042700768, "learning_rate": 3.475000024050917e-05, "loss": 0.1792, "num_input_tokens_seen": 28660800, "step": 14900 }, { "epoch": 2.431519699812383, "grad_norm": 0.5507267713546753, "learning_rate": 3.4740959410280926e-05, "loss": 0.0847, "num_input_tokens_seen": 28669904, "step": 14905 }, { "epoch": 2.4323354270332, "grad_norm": 0.07428539544343948, "learning_rate": 3.4731917077874324e-05, "loss": 0.1835, "num_input_tokens_seen": 28681392, "step": 14910 }, { "epoch": 2.4331511542540176, "grad_norm": 0.08048471063375473, "learning_rate": 3.4722873244683816e-05, "loss": 0.1095, "num_input_tokens_seen": 28691824, "step": 14915 }, { "epoch": 2.4339668814748348, "grad_norm": 1.7376888990402222, "learning_rate": 3.4713827912104065e-05, "loss": 0.0659, "num_input_tokens_seen": 28701680, "step": 14920 }, { "epoch": 2.4347826086956523, "grad_norm": 10.126543045043945, "learning_rate": 3.470478108152998e-05, "loss": 0.0605, "num_input_tokens_seen": 28710688, "step": 14925 }, { "epoch": 2.4355983359164695, "grad_norm": 0.4980336129665375, "learning_rate": 3.4695732754356695e-05, "loss": 0.1269, "num_input_tokens_seen": 28720976, "step": 14930 }, { "epoch": 2.436414063137287, "grad_norm": 0.14678587019443512, "learning_rate": 3.4686682931979576e-05, "loss": 0.0564, "num_input_tokens_seen": 28731488, "step": 14935 }, { "epoch": 2.437229790358104, "grad_norm": 0.22872231900691986, "learning_rate": 3.467763161579422e-05, "loss": 0.0059, "num_input_tokens_seen": 28740064, "step": 14940 }, { "epoch": 2.4380455175789217, "grad_norm": 0.13359034061431885, "learning_rate": 3.466857880719645e-05, "loss": 0.0203, "num_input_tokens_seen": 28748752, "step": 14945 }, { "epoch": 2.438861244799739, "grad_norm": 0.07515594363212585, "learning_rate": 3.465952450758233e-05, "loss": 0.01, "num_input_tokens_seen": 28758384, "step": 14950 }, { "epoch": 2.4396769720205564, "grad_norm": 8.137689590454102, "learning_rate": 3.4650468718348126e-05, "loss": 0.1486, "num_input_tokens_seen": 28768704, "step": 14955 }, { "epoch": 2.4404926992413736, "grad_norm": 8.29736614227295, "learning_rate": 3.464141144089038e-05, "loss": 0.0557, "num_input_tokens_seen": 28777408, "step": 14960 }, { "epoch": 2.441308426462191, "grad_norm": 0.08335006237030029, "learning_rate": 3.463235267660583e-05, "loss": 0.0029, "num_input_tokens_seen": 28787456, "step": 14965 }, { "epoch": 2.4421241536830083, "grad_norm": 6.51212215423584, "learning_rate": 3.462329242689145e-05, "loss": 0.1521, "num_input_tokens_seen": 28797504, "step": 14970 }, { "epoch": 2.442939880903826, "grad_norm": 9.102468490600586, "learning_rate": 3.461423069314444e-05, "loss": 0.258, "num_input_tokens_seen": 28807824, "step": 14975 }, { "epoch": 2.443755608124643, "grad_norm": 0.13290099799633026, "learning_rate": 3.460516747676224e-05, "loss": 0.0622, "num_input_tokens_seen": 28817456, "step": 14980 }, { "epoch": 2.4445713353454606, "grad_norm": 2.7881898880004883, "learning_rate": 3.459610277914251e-05, "loss": 0.1036, "num_input_tokens_seen": 28827680, "step": 14985 }, { "epoch": 2.4453870625662777, "grad_norm": 0.12244553118944168, "learning_rate": 3.458703660168314e-05, "loss": 0.0206, "num_input_tokens_seen": 28837456, "step": 14990 }, { "epoch": 2.4462027897870953, "grad_norm": 0.04794885963201523, "learning_rate": 3.457796894578224e-05, "loss": 0.158, "num_input_tokens_seen": 28847136, "step": 14995 }, { "epoch": 2.4470185170079124, "grad_norm": 0.04190535098314285, "learning_rate": 3.456889981283817e-05, "loss": 0.0111, "num_input_tokens_seen": 28856272, "step": 15000 }, { "epoch": 2.4470185170079124, "eval_loss": 0.15830248594284058, "eval_runtime": 80.6093, "eval_samples_per_second": 33.805, "eval_steps_per_second": 16.909, "num_input_tokens_seen": 28856272, "step": 15000 }, { "epoch": 2.44783424422873, "grad_norm": 0.1355772316455841, "learning_rate": 3.45598292042495e-05, "loss": 0.1665, "num_input_tokens_seen": 28864672, "step": 15005 }, { "epoch": 2.448649971449547, "grad_norm": 3.651782989501953, "learning_rate": 3.4550757121415035e-05, "loss": 0.1849, "num_input_tokens_seen": 28872784, "step": 15010 }, { "epoch": 2.4494656986703647, "grad_norm": 0.3550991415977478, "learning_rate": 3.454168356573378e-05, "loss": 0.0111, "num_input_tokens_seen": 28882320, "step": 15015 }, { "epoch": 2.450281425891182, "grad_norm": 0.0855480283498764, "learning_rate": 3.453260853860503e-05, "loss": 0.0763, "num_input_tokens_seen": 28891296, "step": 15020 }, { "epoch": 2.4510971531119994, "grad_norm": 0.14177384972572327, "learning_rate": 3.452353204142824e-05, "loss": 0.1252, "num_input_tokens_seen": 28900304, "step": 15025 }, { "epoch": 2.4519128803328165, "grad_norm": 5.328641414642334, "learning_rate": 3.4514454075603136e-05, "loss": 0.0785, "num_input_tokens_seen": 28909440, "step": 15030 }, { "epoch": 2.452728607553634, "grad_norm": 3.670375108718872, "learning_rate": 3.450537464252964e-05, "loss": 0.1721, "num_input_tokens_seen": 28918928, "step": 15035 }, { "epoch": 2.4535443347744517, "grad_norm": 0.9455239772796631, "learning_rate": 3.4496293743607925e-05, "loss": 0.0174, "num_input_tokens_seen": 28928480, "step": 15040 }, { "epoch": 2.454360061995269, "grad_norm": 0.5147427320480347, "learning_rate": 3.448721138023838e-05, "loss": 0.0115, "num_input_tokens_seen": 28937312, "step": 15045 }, { "epoch": 2.455175789216086, "grad_norm": 0.1200718879699707, "learning_rate": 3.447812755382162e-05, "loss": 0.2322, "num_input_tokens_seen": 28947760, "step": 15050 }, { "epoch": 2.4559915164369035, "grad_norm": 3.9885387420654297, "learning_rate": 3.446904226575847e-05, "loss": 0.199, "num_input_tokens_seen": 28957920, "step": 15055 }, { "epoch": 2.456807243657721, "grad_norm": 1.368145227432251, "learning_rate": 3.445995551745002e-05, "loss": 0.1251, "num_input_tokens_seen": 28967136, "step": 15060 }, { "epoch": 2.457622970878538, "grad_norm": 0.7329331636428833, "learning_rate": 3.445086731029753e-05, "loss": 0.0217, "num_input_tokens_seen": 28976208, "step": 15065 }, { "epoch": 2.4584386980993553, "grad_norm": 2.789210557937622, "learning_rate": 3.444177764570255e-05, "loss": 0.0201, "num_input_tokens_seen": 28986656, "step": 15070 }, { "epoch": 2.459254425320173, "grad_norm": 0.3335046172142029, "learning_rate": 3.44326865250668e-05, "loss": 0.1715, "num_input_tokens_seen": 28996928, "step": 15075 }, { "epoch": 2.4600701525409905, "grad_norm": 10.273541450500488, "learning_rate": 3.442359394979225e-05, "loss": 0.0814, "num_input_tokens_seen": 29007024, "step": 15080 }, { "epoch": 2.4608858797618076, "grad_norm": 1.8824050426483154, "learning_rate": 3.441449992128108e-05, "loss": 0.1125, "num_input_tokens_seen": 29016480, "step": 15085 }, { "epoch": 2.461701606982625, "grad_norm": 6.699951648712158, "learning_rate": 3.440540444093573e-05, "loss": 0.1089, "num_input_tokens_seen": 29024832, "step": 15090 }, { "epoch": 2.4625173342034423, "grad_norm": 0.04979947581887245, "learning_rate": 3.43963075101588e-05, "loss": 0.0256, "num_input_tokens_seen": 29034064, "step": 15095 }, { "epoch": 2.46333306142426, "grad_norm": 0.05424768850207329, "learning_rate": 3.438720913035318e-05, "loss": 0.01, "num_input_tokens_seen": 29044064, "step": 15100 }, { "epoch": 2.464148788645077, "grad_norm": 0.05452780798077583, "learning_rate": 3.437810930292195e-05, "loss": 0.0834, "num_input_tokens_seen": 29053968, "step": 15105 }, { "epoch": 2.4649645158658946, "grad_norm": 0.21766474843025208, "learning_rate": 3.43690080292684e-05, "loss": 0.0407, "num_input_tokens_seen": 29062144, "step": 15110 }, { "epoch": 2.4657802430867117, "grad_norm": 0.3232106566429138, "learning_rate": 3.435990531079608e-05, "loss": 0.078, "num_input_tokens_seen": 29070672, "step": 15115 }, { "epoch": 2.4665959703075293, "grad_norm": 0.07935960590839386, "learning_rate": 3.435080114890874e-05, "loss": 0.0288, "num_input_tokens_seen": 29081072, "step": 15120 }, { "epoch": 2.4674116975283464, "grad_norm": 0.0920410230755806, "learning_rate": 3.434169554501035e-05, "loss": 0.2502, "num_input_tokens_seen": 29090192, "step": 15125 }, { "epoch": 2.468227424749164, "grad_norm": 0.07209794223308563, "learning_rate": 3.433258850050511e-05, "loss": 0.1559, "num_input_tokens_seen": 29100416, "step": 15130 }, { "epoch": 2.469043151969981, "grad_norm": 0.7566903829574585, "learning_rate": 3.4323480016797446e-05, "loss": 0.0939, "num_input_tokens_seen": 29109328, "step": 15135 }, { "epoch": 2.4698588791907987, "grad_norm": 4.48051643371582, "learning_rate": 3.4314370095291995e-05, "loss": 0.0972, "num_input_tokens_seen": 29119552, "step": 15140 }, { "epoch": 2.470674606411616, "grad_norm": 21.176368713378906, "learning_rate": 3.430525873739363e-05, "loss": 0.1656, "num_input_tokens_seen": 29128656, "step": 15145 }, { "epoch": 2.4714903336324334, "grad_norm": 0.6717095971107483, "learning_rate": 3.429614594450743e-05, "loss": 0.0495, "num_input_tokens_seen": 29137760, "step": 15150 }, { "epoch": 2.4723060608532506, "grad_norm": 0.09295976161956787, "learning_rate": 3.428703171803869e-05, "loss": 0.0746, "num_input_tokens_seen": 29146064, "step": 15155 }, { "epoch": 2.473121788074068, "grad_norm": 0.17196324467658997, "learning_rate": 3.4277916059392964e-05, "loss": 0.1269, "num_input_tokens_seen": 29156528, "step": 15160 }, { "epoch": 2.4739375152948853, "grad_norm": 3.0182976722717285, "learning_rate": 3.426879896997598e-05, "loss": 0.0643, "num_input_tokens_seen": 29165936, "step": 15165 }, { "epoch": 2.474753242515703, "grad_norm": 0.07200907170772552, "learning_rate": 3.425968045119372e-05, "loss": 0.0931, "num_input_tokens_seen": 29175536, "step": 15170 }, { "epoch": 2.47556896973652, "grad_norm": 0.06835351884365082, "learning_rate": 3.425056050445237e-05, "loss": 0.0456, "num_input_tokens_seen": 29184928, "step": 15175 }, { "epoch": 2.4763846969573375, "grad_norm": 0.35220178961753845, "learning_rate": 3.4241439131158336e-05, "loss": 0.0559, "num_input_tokens_seen": 29195232, "step": 15180 }, { "epoch": 2.4772004241781547, "grad_norm": 8.049848556518555, "learning_rate": 3.423231633271825e-05, "loss": 0.0176, "num_input_tokens_seen": 29205248, "step": 15185 }, { "epoch": 2.4780161513989722, "grad_norm": 8.349711418151855, "learning_rate": 3.4223192110538985e-05, "loss": 0.085, "num_input_tokens_seen": 29215808, "step": 15190 }, { "epoch": 2.4788318786197894, "grad_norm": 4.706721305847168, "learning_rate": 3.4214066466027575e-05, "loss": 0.1772, "num_input_tokens_seen": 29225952, "step": 15195 }, { "epoch": 2.479647605840607, "grad_norm": 0.5686995983123779, "learning_rate": 3.4204939400591325e-05, "loss": 0.1074, "num_input_tokens_seen": 29234704, "step": 15200 }, { "epoch": 2.479647605840607, "eval_loss": 0.1568540632724762, "eval_runtime": 80.6931, "eval_samples_per_second": 33.77, "eval_steps_per_second": 16.891, "num_input_tokens_seen": 29234704, "step": 15200 }, { "epoch": 2.480463333061424, "grad_norm": 33.0731201171875, "learning_rate": 3.419581091563775e-05, "loss": 0.1366, "num_input_tokens_seen": 29244176, "step": 15205 }, { "epoch": 2.4812790602822417, "grad_norm": 33.65237808227539, "learning_rate": 3.418668101257456e-05, "loss": 0.1914, "num_input_tokens_seen": 29253280, "step": 15210 }, { "epoch": 2.482094787503059, "grad_norm": 0.5840086340904236, "learning_rate": 3.417754969280971e-05, "loss": 0.1228, "num_input_tokens_seen": 29263024, "step": 15215 }, { "epoch": 2.4829105147238764, "grad_norm": 13.186771392822266, "learning_rate": 3.416841695775137e-05, "loss": 0.1023, "num_input_tokens_seen": 29272528, "step": 15220 }, { "epoch": 2.4837262419446935, "grad_norm": 0.10213175415992737, "learning_rate": 3.415928280880792e-05, "loss": 0.0857, "num_input_tokens_seen": 29282192, "step": 15225 }, { "epoch": 2.484541969165511, "grad_norm": 0.061870723962783813, "learning_rate": 3.4150147247387965e-05, "loss": 0.006, "num_input_tokens_seen": 29291504, "step": 15230 }, { "epoch": 2.4853576963863286, "grad_norm": 2.372612476348877, "learning_rate": 3.4141010274900306e-05, "loss": 0.0772, "num_input_tokens_seen": 29299424, "step": 15235 }, { "epoch": 2.4861734236071458, "grad_norm": 0.09497179836034775, "learning_rate": 3.413187189275399e-05, "loss": 0.0464, "num_input_tokens_seen": 29309376, "step": 15240 }, { "epoch": 2.486989150827963, "grad_norm": 17.44321060180664, "learning_rate": 3.4122732102358265e-05, "loss": 0.1597, "num_input_tokens_seen": 29318544, "step": 15245 }, { "epoch": 2.4878048780487805, "grad_norm": 5.019428253173828, "learning_rate": 3.411359090512261e-05, "loss": 0.0121, "num_input_tokens_seen": 29328448, "step": 15250 }, { "epoch": 2.488620605269598, "grad_norm": 0.09571440517902374, "learning_rate": 3.410444830245672e-05, "loss": 0.004, "num_input_tokens_seen": 29337936, "step": 15255 }, { "epoch": 2.489436332490415, "grad_norm": 4.790166854858398, "learning_rate": 3.409530429577048e-05, "loss": 0.0076, "num_input_tokens_seen": 29347120, "step": 15260 }, { "epoch": 2.4902520597112328, "grad_norm": 0.4514438807964325, "learning_rate": 3.408615888647402e-05, "loss": 0.1831, "num_input_tokens_seen": 29356464, "step": 15265 }, { "epoch": 2.49106778693205, "grad_norm": 4.316021919250488, "learning_rate": 3.4077012075977675e-05, "loss": 0.0407, "num_input_tokens_seen": 29365584, "step": 15270 }, { "epoch": 2.4918835141528675, "grad_norm": 7.764566898345947, "learning_rate": 3.4067863865692e-05, "loss": 0.0497, "num_input_tokens_seen": 29375424, "step": 15275 }, { "epoch": 2.4926992413736846, "grad_norm": 0.12369349598884583, "learning_rate": 3.4058714257027755e-05, "loss": 0.0259, "num_input_tokens_seen": 29385056, "step": 15280 }, { "epoch": 2.493514968594502, "grad_norm": 13.716044425964355, "learning_rate": 3.404956325139594e-05, "loss": 0.0671, "num_input_tokens_seen": 29395760, "step": 15285 }, { "epoch": 2.4943306958153193, "grad_norm": 4.364566802978516, "learning_rate": 3.404041085020775e-05, "loss": 0.0669, "num_input_tokens_seen": 29405472, "step": 15290 }, { "epoch": 2.495146423036137, "grad_norm": 5.965693473815918, "learning_rate": 3.403125705487459e-05, "loss": 0.076, "num_input_tokens_seen": 29414656, "step": 15295 }, { "epoch": 2.495962150256954, "grad_norm": 0.22130264341831207, "learning_rate": 3.402210186680811e-05, "loss": 0.0919, "num_input_tokens_seen": 29424240, "step": 15300 }, { "epoch": 2.4967778774777716, "grad_norm": 3.1909923553466797, "learning_rate": 3.4012945287420137e-05, "loss": 0.0415, "num_input_tokens_seen": 29433440, "step": 15305 }, { "epoch": 2.4975936046985887, "grad_norm": 8.003599166870117, "learning_rate": 3.400378731812274e-05, "loss": 0.0788, "num_input_tokens_seen": 29442416, "step": 15310 }, { "epoch": 2.4984093319194063, "grad_norm": 0.1885966807603836, "learning_rate": 3.399462796032817e-05, "loss": 0.0689, "num_input_tokens_seen": 29452944, "step": 15315 }, { "epoch": 2.4992250591402234, "grad_norm": 0.13772395253181458, "learning_rate": 3.3985467215448954e-05, "loss": 0.0073, "num_input_tokens_seen": 29461920, "step": 15320 }, { "epoch": 2.500040786361041, "grad_norm": 33.893821716308594, "learning_rate": 3.3976305084897776e-05, "loss": 0.0245, "num_input_tokens_seen": 29471520, "step": 15325 }, { "epoch": 2.500856513581858, "grad_norm": 0.33131346106529236, "learning_rate": 3.3967141570087544e-05, "loss": 0.0282, "num_input_tokens_seen": 29481088, "step": 15330 }, { "epoch": 2.5016722408026757, "grad_norm": 0.08547098189592361, "learning_rate": 3.39579766724314e-05, "loss": 0.0192, "num_input_tokens_seen": 29491904, "step": 15335 }, { "epoch": 2.502487968023493, "grad_norm": 0.040994394570589066, "learning_rate": 3.3948810393342677e-05, "loss": 0.1307, "num_input_tokens_seen": 29501776, "step": 15340 }, { "epoch": 2.5033036952443104, "grad_norm": 36.8040657043457, "learning_rate": 3.3939642734234936e-05, "loss": 0.1492, "num_input_tokens_seen": 29512016, "step": 15345 }, { "epoch": 2.5041194224651275, "grad_norm": 0.07192758470773697, "learning_rate": 3.393047369652194e-05, "loss": 0.0296, "num_input_tokens_seen": 29522064, "step": 15350 }, { "epoch": 2.504935149685945, "grad_norm": 0.06405460089445114, "learning_rate": 3.3921303281617664e-05, "loss": 0.0734, "num_input_tokens_seen": 29531440, "step": 15355 }, { "epoch": 2.5057508769067622, "grad_norm": 0.042540859431028366, "learning_rate": 3.391213149093632e-05, "loss": 0.1904, "num_input_tokens_seen": 29541024, "step": 15360 }, { "epoch": 2.50656660412758, "grad_norm": 0.02309397980570793, "learning_rate": 3.3902958325892303e-05, "loss": 0.0319, "num_input_tokens_seen": 29552128, "step": 15365 }, { "epoch": 2.507382331348397, "grad_norm": 0.04197274520993233, "learning_rate": 3.389378378790023e-05, "loss": 0.0792, "num_input_tokens_seen": 29561248, "step": 15370 }, { "epoch": 2.5081980585692145, "grad_norm": 0.3177729845046997, "learning_rate": 3.388460787837493e-05, "loss": 0.0043, "num_input_tokens_seen": 29571152, "step": 15375 }, { "epoch": 2.5090137857900316, "grad_norm": 7.872017860412598, "learning_rate": 3.387543059873145e-05, "loss": 0.0271, "num_input_tokens_seen": 29581248, "step": 15380 }, { "epoch": 2.5098295130108492, "grad_norm": 0.04975946247577667, "learning_rate": 3.386625195038503e-05, "loss": 0.0781, "num_input_tokens_seen": 29590896, "step": 15385 }, { "epoch": 2.510645240231667, "grad_norm": 9.99071216583252, "learning_rate": 3.3857071934751136e-05, "loss": 0.2328, "num_input_tokens_seen": 29600064, "step": 15390 }, { "epoch": 2.511460967452484, "grad_norm": 0.046065352857112885, "learning_rate": 3.384789055324544e-05, "loss": 0.0381, "num_input_tokens_seen": 29608448, "step": 15395 }, { "epoch": 2.512276694673301, "grad_norm": 0.05906765162944794, "learning_rate": 3.3838707807283843e-05, "loss": 0.0931, "num_input_tokens_seen": 29617728, "step": 15400 }, { "epoch": 2.512276694673301, "eval_loss": 0.17570002377033234, "eval_runtime": 80.4951, "eval_samples_per_second": 33.853, "eval_steps_per_second": 16.933, "num_input_tokens_seen": 29617728, "step": 15400 }, { "epoch": 2.5130924218941186, "grad_norm": 0.4470925033092499, "learning_rate": 3.382952369828243e-05, "loss": 0.1121, "num_input_tokens_seen": 29625632, "step": 15405 }, { "epoch": 2.513908149114936, "grad_norm": 0.11295468360185623, "learning_rate": 3.38203382276575e-05, "loss": 0.104, "num_input_tokens_seen": 29635936, "step": 15410 }, { "epoch": 2.5147238763357533, "grad_norm": 0.11262326687574387, "learning_rate": 3.381115139682557e-05, "loss": 0.0101, "num_input_tokens_seen": 29645408, "step": 15415 }, { "epoch": 2.5155396035565705, "grad_norm": 0.11103654652833939, "learning_rate": 3.3801963207203366e-05, "loss": 0.0035, "num_input_tokens_seen": 29656480, "step": 15420 }, { "epoch": 2.516355330777388, "grad_norm": 1.7870334386825562, "learning_rate": 3.379277366020782e-05, "loss": 0.0843, "num_input_tokens_seen": 29667024, "step": 15425 }, { "epoch": 2.5171710579982056, "grad_norm": 0.1109570637345314, "learning_rate": 3.3783582757256085e-05, "loss": 0.1824, "num_input_tokens_seen": 29676944, "step": 15430 }, { "epoch": 2.5179867852190227, "grad_norm": 6.812257766723633, "learning_rate": 3.3774390499765504e-05, "loss": 0.0232, "num_input_tokens_seen": 29686128, "step": 15435 }, { "epoch": 2.51880251243984, "grad_norm": 4.877628326416016, "learning_rate": 3.376519688915364e-05, "loss": 0.0656, "num_input_tokens_seen": 29695536, "step": 15440 }, { "epoch": 2.5196182396606575, "grad_norm": 0.2087065577507019, "learning_rate": 3.3756001926838273e-05, "loss": 0.0037, "num_input_tokens_seen": 29704928, "step": 15445 }, { "epoch": 2.520433966881475, "grad_norm": 0.09656035155057907, "learning_rate": 3.374680561423737e-05, "loss": 0.0585, "num_input_tokens_seen": 29715296, "step": 15450 }, { "epoch": 2.521249694102292, "grad_norm": 3.743852138519287, "learning_rate": 3.373760795276912e-05, "loss": 0.2436, "num_input_tokens_seen": 29725072, "step": 15455 }, { "epoch": 2.5220654213231093, "grad_norm": 11.74156379699707, "learning_rate": 3.372840894385192e-05, "loss": 0.0691, "num_input_tokens_seen": 29734592, "step": 15460 }, { "epoch": 2.522881148543927, "grad_norm": 0.14797130227088928, "learning_rate": 3.3719208588904375e-05, "loss": 0.0546, "num_input_tokens_seen": 29742800, "step": 15465 }, { "epoch": 2.5236968757647444, "grad_norm": 0.13813985884189606, "learning_rate": 3.371000688934529e-05, "loss": 0.1029, "num_input_tokens_seen": 29752368, "step": 15470 }, { "epoch": 2.5245126029855616, "grad_norm": 0.033723775297403336, "learning_rate": 3.370080384659369e-05, "loss": 0.1161, "num_input_tokens_seen": 29762176, "step": 15475 }, { "epoch": 2.525328330206379, "grad_norm": 0.19966821372509003, "learning_rate": 3.36915994620688e-05, "loss": 0.0096, "num_input_tokens_seen": 29771920, "step": 15480 }, { "epoch": 2.5261440574271963, "grad_norm": 0.1841120719909668, "learning_rate": 3.3682393737190035e-05, "loss": 0.0698, "num_input_tokens_seen": 29781792, "step": 15485 }, { "epoch": 2.526959784648014, "grad_norm": 0.735352635383606, "learning_rate": 3.3673186673377054e-05, "loss": 0.0126, "num_input_tokens_seen": 29791328, "step": 15490 }, { "epoch": 2.527775511868831, "grad_norm": 2.5135951042175293, "learning_rate": 3.366397827204969e-05, "loss": 0.1874, "num_input_tokens_seen": 29801472, "step": 15495 }, { "epoch": 2.5285912390896486, "grad_norm": 0.19453343749046326, "learning_rate": 3.3654768534628e-05, "loss": 0.1101, "num_input_tokens_seen": 29810944, "step": 15500 }, { "epoch": 2.5294069663104657, "grad_norm": 8.129416465759277, "learning_rate": 3.3645557462532245e-05, "loss": 0.0233, "num_input_tokens_seen": 29821152, "step": 15505 }, { "epoch": 2.5302226935312833, "grad_norm": 0.14572913944721222, "learning_rate": 3.363634505718288e-05, "loss": 0.0055, "num_input_tokens_seen": 29830880, "step": 15510 }, { "epoch": 2.5310384207521004, "grad_norm": 0.3689590096473694, "learning_rate": 3.362713132000057e-05, "loss": 0.0074, "num_input_tokens_seen": 29840144, "step": 15515 }, { "epoch": 2.531854147972918, "grad_norm": 0.12765167653560638, "learning_rate": 3.36179162524062e-05, "loss": 0.0172, "num_input_tokens_seen": 29849712, "step": 15520 }, { "epoch": 2.532669875193735, "grad_norm": 9.210830688476562, "learning_rate": 3.3608699855820846e-05, "loss": 0.1422, "num_input_tokens_seen": 29859520, "step": 15525 }, { "epoch": 2.5334856024145527, "grad_norm": 0.2928789258003235, "learning_rate": 3.359948213166578e-05, "loss": 0.0266, "num_input_tokens_seen": 29870016, "step": 15530 }, { "epoch": 2.53430132963537, "grad_norm": 0.16169115900993347, "learning_rate": 3.359026308136252e-05, "loss": 0.0466, "num_input_tokens_seen": 29879136, "step": 15535 }, { "epoch": 2.5351170568561874, "grad_norm": 20.760116577148438, "learning_rate": 3.358104270633272e-05, "loss": 0.1914, "num_input_tokens_seen": 29888928, "step": 15540 }, { "epoch": 2.5359327840770045, "grad_norm": 0.027660585939884186, "learning_rate": 3.357182100799831e-05, "loss": 0.0079, "num_input_tokens_seen": 29898336, "step": 15545 }, { "epoch": 2.536748511297822, "grad_norm": 0.19372710585594177, "learning_rate": 3.3562597987781384e-05, "loss": 0.1628, "num_input_tokens_seen": 29908912, "step": 15550 }, { "epoch": 2.537564238518639, "grad_norm": 5.046256065368652, "learning_rate": 3.355337364710424e-05, "loss": 0.0347, "num_input_tokens_seen": 29917584, "step": 15555 }, { "epoch": 2.538379965739457, "grad_norm": 0.058916427195072174, "learning_rate": 3.354414798738939e-05, "loss": 0.1695, "num_input_tokens_seen": 29925920, "step": 15560 }, { "epoch": 2.539195692960274, "grad_norm": 0.6457627415657043, "learning_rate": 3.353492101005955e-05, "loss": 0.0048, "num_input_tokens_seen": 29935520, "step": 15565 }, { "epoch": 2.5400114201810915, "grad_norm": 0.6542972326278687, "learning_rate": 3.352569271653763e-05, "loss": 0.1135, "num_input_tokens_seen": 29946112, "step": 15570 }, { "epoch": 2.5408271474019086, "grad_norm": 0.07525448501110077, "learning_rate": 3.351646310824675e-05, "loss": 0.1705, "num_input_tokens_seen": 29955664, "step": 15575 }, { "epoch": 2.541642874622726, "grad_norm": 0.07309743762016296, "learning_rate": 3.350723218661023e-05, "loss": 0.0157, "num_input_tokens_seen": 29966224, "step": 15580 }, { "epoch": 2.5424586018435438, "grad_norm": 0.3335886001586914, "learning_rate": 3.349799995305162e-05, "loss": 0.0187, "num_input_tokens_seen": 29976560, "step": 15585 }, { "epoch": 2.543274329064361, "grad_norm": 0.06698064506053925, "learning_rate": 3.348876640899461e-05, "loss": 0.0019, "num_input_tokens_seen": 29985536, "step": 15590 }, { "epoch": 2.544090056285178, "grad_norm": 0.21583053469657898, "learning_rate": 3.3479531555863144e-05, "loss": 0.456, "num_input_tokens_seen": 29993808, "step": 15595 }, { "epoch": 2.5449057835059956, "grad_norm": 0.09343983232975006, "learning_rate": 3.3470295395081344e-05, "loss": 0.1366, "num_input_tokens_seen": 30004032, "step": 15600 }, { "epoch": 2.5449057835059956, "eval_loss": 0.17634761333465576, "eval_runtime": 80.7075, "eval_samples_per_second": 33.764, "eval_steps_per_second": 16.888, "num_input_tokens_seen": 30004032, "step": 15600 }, { "epoch": 2.545721510726813, "grad_norm": 0.0757494643330574, "learning_rate": 3.3461057928073556e-05, "loss": 0.0042, "num_input_tokens_seen": 30012768, "step": 15605 }, { "epoch": 2.5465372379476303, "grad_norm": 0.048544742166996, "learning_rate": 3.345181915626431e-05, "loss": 0.0824, "num_input_tokens_seen": 30023216, "step": 15610 }, { "epoch": 2.5473529651684474, "grad_norm": 0.1464398205280304, "learning_rate": 3.344257908107834e-05, "loss": 0.0097, "num_input_tokens_seen": 30033232, "step": 15615 }, { "epoch": 2.548168692389265, "grad_norm": 0.3193468749523163, "learning_rate": 3.343333770394058e-05, "loss": 0.0191, "num_input_tokens_seen": 30042224, "step": 15620 }, { "epoch": 2.5489844196100826, "grad_norm": 0.1785127818584442, "learning_rate": 3.342409502627616e-05, "loss": 0.0166, "num_input_tokens_seen": 30051232, "step": 15625 }, { "epoch": 2.5498001468308997, "grad_norm": 2.414695978164673, "learning_rate": 3.341485104951043e-05, "loss": 0.0434, "num_input_tokens_seen": 30059520, "step": 15630 }, { "epoch": 2.550615874051717, "grad_norm": 0.044403936713933945, "learning_rate": 3.340560577506892e-05, "loss": 0.1745, "num_input_tokens_seen": 30067648, "step": 15635 }, { "epoch": 2.5514316012725344, "grad_norm": 0.011859099380671978, "learning_rate": 3.339635920437735e-05, "loss": 0.1463, "num_input_tokens_seen": 30077760, "step": 15640 }, { "epoch": 2.552247328493352, "grad_norm": 9.226837158203125, "learning_rate": 3.338711133886169e-05, "loss": 0.0965, "num_input_tokens_seen": 30087328, "step": 15645 }, { "epoch": 2.553063055714169, "grad_norm": 5.109345436096191, "learning_rate": 3.3377862179948064e-05, "loss": 0.1256, "num_input_tokens_seen": 30096720, "step": 15650 }, { "epoch": 2.5538787829349863, "grad_norm": 0.033507443964481354, "learning_rate": 3.336861172906281e-05, "loss": 0.0529, "num_input_tokens_seen": 30106864, "step": 15655 }, { "epoch": 2.554694510155804, "grad_norm": 1.5782979726791382, "learning_rate": 3.335935998763245e-05, "loss": 0.0523, "num_input_tokens_seen": 30116464, "step": 15660 }, { "epoch": 2.5555102373766214, "grad_norm": 5.3647141456604, "learning_rate": 3.3350106957083744e-05, "loss": 0.0129, "num_input_tokens_seen": 30127328, "step": 15665 }, { "epoch": 2.5563259645974385, "grad_norm": 14.843099594116211, "learning_rate": 3.33408526388436e-05, "loss": 0.0937, "num_input_tokens_seen": 30136512, "step": 15670 }, { "epoch": 2.557141691818256, "grad_norm": 0.03382357209920883, "learning_rate": 3.3331597034339166e-05, "loss": 0.1991, "num_input_tokens_seen": 30145360, "step": 15675 }, { "epoch": 2.5579574190390733, "grad_norm": 16.77920150756836, "learning_rate": 3.3322340144997764e-05, "loss": 0.3063, "num_input_tokens_seen": 30155120, "step": 15680 }, { "epoch": 2.558773146259891, "grad_norm": 0.0556364469230175, "learning_rate": 3.331308197224693e-05, "loss": 0.1829, "num_input_tokens_seen": 30163712, "step": 15685 }, { "epoch": 2.559588873480708, "grad_norm": 12.713562965393066, "learning_rate": 3.330382251751438e-05, "loss": 0.0161, "num_input_tokens_seen": 30173936, "step": 15690 }, { "epoch": 2.5604046007015255, "grad_norm": 0.10952233523130417, "learning_rate": 3.3294561782228054e-05, "loss": 0.0273, "num_input_tokens_seen": 30183376, "step": 15695 }, { "epoch": 2.5612203279223427, "grad_norm": 0.13090232014656067, "learning_rate": 3.328529976781607e-05, "loss": 0.0068, "num_input_tokens_seen": 30192768, "step": 15700 }, { "epoch": 2.5620360551431602, "grad_norm": 4.27801513671875, "learning_rate": 3.327603647570673e-05, "loss": 0.2159, "num_input_tokens_seen": 30202448, "step": 15705 }, { "epoch": 2.5628517823639774, "grad_norm": 0.023850997909903526, "learning_rate": 3.326677190732857e-05, "loss": 0.0435, "num_input_tokens_seen": 30213664, "step": 15710 }, { "epoch": 2.563667509584795, "grad_norm": 0.563783586025238, "learning_rate": 3.325750606411029e-05, "loss": 0.0282, "num_input_tokens_seen": 30223712, "step": 15715 }, { "epoch": 2.564483236805612, "grad_norm": 1.476179599761963, "learning_rate": 3.3248238947480804e-05, "loss": 0.038, "num_input_tokens_seen": 30232640, "step": 15720 }, { "epoch": 2.5652989640264297, "grad_norm": 0.11567360162734985, "learning_rate": 3.323897055886922e-05, "loss": 0.042, "num_input_tokens_seen": 30242048, "step": 15725 }, { "epoch": 2.566114691247247, "grad_norm": 0.1551218032836914, "learning_rate": 3.322970089970484e-05, "loss": 0.0049, "num_input_tokens_seen": 30251008, "step": 15730 }, { "epoch": 2.5669304184680644, "grad_norm": 0.07110387086868286, "learning_rate": 3.3220429971417165e-05, "loss": 0.0023, "num_input_tokens_seen": 30260160, "step": 15735 }, { "epoch": 2.5677461456888815, "grad_norm": 1.0211411714553833, "learning_rate": 3.321115777543588e-05, "loss": 0.2055, "num_input_tokens_seen": 30268144, "step": 15740 }, { "epoch": 2.568561872909699, "grad_norm": 0.0601465106010437, "learning_rate": 3.320188431319088e-05, "loss": 0.1707, "num_input_tokens_seen": 30277392, "step": 15745 }, { "epoch": 2.569377600130516, "grad_norm": 2.551783561706543, "learning_rate": 3.319260958611224e-05, "loss": 0.1205, "num_input_tokens_seen": 30286848, "step": 15750 }, { "epoch": 2.5701933273513338, "grad_norm": 5.8261237144470215, "learning_rate": 3.3183333595630256e-05, "loss": 0.0971, "num_input_tokens_seen": 30295616, "step": 15755 }, { "epoch": 2.5710090545721513, "grad_norm": 0.11420835554599762, "learning_rate": 3.317405634317538e-05, "loss": 0.1945, "num_input_tokens_seen": 30305456, "step": 15760 }, { "epoch": 2.5718247817929685, "grad_norm": 0.36658066511154175, "learning_rate": 3.3164777830178315e-05, "loss": 0.1948, "num_input_tokens_seen": 30315504, "step": 15765 }, { "epoch": 2.5726405090137856, "grad_norm": 8.797396659851074, "learning_rate": 3.315549805806989e-05, "loss": 0.0852, "num_input_tokens_seen": 30324720, "step": 15770 }, { "epoch": 2.573456236234603, "grad_norm": 0.06926865130662918, "learning_rate": 3.314621702828118e-05, "loss": 0.0526, "num_input_tokens_seen": 30334160, "step": 15775 }, { "epoch": 2.5742719634554208, "grad_norm": 14.982314109802246, "learning_rate": 3.313693474224342e-05, "loss": 0.1675, "num_input_tokens_seen": 30344848, "step": 15780 }, { "epoch": 2.575087690676238, "grad_norm": 11.807470321655273, "learning_rate": 3.312765120138809e-05, "loss": 0.0621, "num_input_tokens_seen": 30356032, "step": 15785 }, { "epoch": 2.575903417897055, "grad_norm": 6.061267375946045, "learning_rate": 3.311836640714679e-05, "loss": 0.162, "num_input_tokens_seen": 30366736, "step": 15790 }, { "epoch": 2.5767191451178726, "grad_norm": 0.0673699751496315, "learning_rate": 3.310908036095137e-05, "loss": 0.0061, "num_input_tokens_seen": 30375952, "step": 15795 }, { "epoch": 2.57753487233869, "grad_norm": 1.4928369522094727, "learning_rate": 3.309979306423386e-05, "loss": 0.0933, "num_input_tokens_seen": 30386752, "step": 15800 }, { "epoch": 2.57753487233869, "eval_loss": 0.1606510579586029, "eval_runtime": 80.5601, "eval_samples_per_second": 33.826, "eval_steps_per_second": 16.919, "num_input_tokens_seen": 30386752, "step": 15800 }, { "epoch": 2.5783505995595073, "grad_norm": 0.09719076007604599, "learning_rate": 3.309050451842647e-05, "loss": 0.0073, "num_input_tokens_seen": 30395568, "step": 15805 }, { "epoch": 2.5791663267803244, "grad_norm": 0.0952904149889946, "learning_rate": 3.3081214724961604e-05, "loss": 0.1003, "num_input_tokens_seen": 30405344, "step": 15810 }, { "epoch": 2.579982054001142, "grad_norm": 0.07281143963336945, "learning_rate": 3.307192368527188e-05, "loss": 0.0536, "num_input_tokens_seen": 30414576, "step": 15815 }, { "epoch": 2.5807977812219596, "grad_norm": 0.13854776322841644, "learning_rate": 3.306263140079008e-05, "loss": 0.0075, "num_input_tokens_seen": 30424080, "step": 15820 }, { "epoch": 2.5816135084427767, "grad_norm": 0.0520842969417572, "learning_rate": 3.30533378729492e-05, "loss": 0.0121, "num_input_tokens_seen": 30433680, "step": 15825 }, { "epoch": 2.582429235663594, "grad_norm": 0.10346539318561554, "learning_rate": 3.304404310318242e-05, "loss": 0.0386, "num_input_tokens_seen": 30442704, "step": 15830 }, { "epoch": 2.5832449628844114, "grad_norm": 3.100130558013916, "learning_rate": 3.3034747092923105e-05, "loss": 0.1194, "num_input_tokens_seen": 30452656, "step": 15835 }, { "epoch": 2.584060690105229, "grad_norm": 1.3971301317214966, "learning_rate": 3.3025449843604806e-05, "loss": 0.088, "num_input_tokens_seen": 30462448, "step": 15840 }, { "epoch": 2.584876417326046, "grad_norm": 2.420644521713257, "learning_rate": 3.30161513566613e-05, "loss": 0.1267, "num_input_tokens_seen": 30472192, "step": 15845 }, { "epoch": 2.5856921445468637, "grad_norm": 4.670839786529541, "learning_rate": 3.3006851633526506e-05, "loss": 0.1471, "num_input_tokens_seen": 30481824, "step": 15850 }, { "epoch": 2.586507871767681, "grad_norm": 0.149531751871109, "learning_rate": 3.2997550675634584e-05, "loss": 0.0213, "num_input_tokens_seen": 30491248, "step": 15855 }, { "epoch": 2.5873235989884984, "grad_norm": 2.7019717693328857, "learning_rate": 3.2988248484419825e-05, "loss": 0.0629, "num_input_tokens_seen": 30500448, "step": 15860 }, { "epoch": 2.5881393262093155, "grad_norm": 3.3140459060668945, "learning_rate": 3.2978945061316776e-05, "loss": 0.0989, "num_input_tokens_seen": 30509936, "step": 15865 }, { "epoch": 2.588955053430133, "grad_norm": 0.8181040287017822, "learning_rate": 3.296964040776013e-05, "loss": 0.0264, "num_input_tokens_seen": 30519440, "step": 15870 }, { "epoch": 2.5897707806509502, "grad_norm": 0.04026656970381737, "learning_rate": 3.296033452518478e-05, "loss": 0.0045, "num_input_tokens_seen": 30529040, "step": 15875 }, { "epoch": 2.590586507871768, "grad_norm": 0.1271546483039856, "learning_rate": 3.2951027415025806e-05, "loss": 0.0044, "num_input_tokens_seen": 30539344, "step": 15880 }, { "epoch": 2.591402235092585, "grad_norm": 2.068253517150879, "learning_rate": 3.294171907871849e-05, "loss": 0.1009, "num_input_tokens_seen": 30547440, "step": 15885 }, { "epoch": 2.5922179623134025, "grad_norm": 0.37611302733421326, "learning_rate": 3.293240951769828e-05, "loss": 0.078, "num_input_tokens_seen": 30556288, "step": 15890 }, { "epoch": 2.5930336895342196, "grad_norm": 1.498902678489685, "learning_rate": 3.2923098733400846e-05, "loss": 0.1238, "num_input_tokens_seen": 30565840, "step": 15895 }, { "epoch": 2.593849416755037, "grad_norm": 0.03217145428061485, "learning_rate": 3.291378672726202e-05, "loss": 0.1541, "num_input_tokens_seen": 30575216, "step": 15900 }, { "epoch": 2.5946651439758543, "grad_norm": 0.1144847720861435, "learning_rate": 3.2904473500717824e-05, "loss": 0.074, "num_input_tokens_seen": 30585424, "step": 15905 }, { "epoch": 2.595480871196672, "grad_norm": 0.17311295866966248, "learning_rate": 3.289515905520449e-05, "loss": 0.1273, "num_input_tokens_seen": 30593824, "step": 15910 }, { "epoch": 2.596296598417489, "grad_norm": 12.574938774108887, "learning_rate": 3.288584339215841e-05, "loss": 0.0382, "num_input_tokens_seen": 30604176, "step": 15915 }, { "epoch": 2.5971123256383066, "grad_norm": 2.541449546813965, "learning_rate": 3.287652651301617e-05, "loss": 0.1507, "num_input_tokens_seen": 30613904, "step": 15920 }, { "epoch": 2.5979280528591238, "grad_norm": 3.156761884689331, "learning_rate": 3.286720841921457e-05, "loss": 0.063, "num_input_tokens_seen": 30624080, "step": 15925 }, { "epoch": 2.5987437800799413, "grad_norm": 0.03558104857802391, "learning_rate": 3.285788911219056e-05, "loss": 0.1152, "num_input_tokens_seen": 30633952, "step": 15930 }, { "epoch": 2.5995595073007585, "grad_norm": 0.3265205919742584, "learning_rate": 3.284856859338131e-05, "loss": 0.0586, "num_input_tokens_seen": 30643872, "step": 15935 }, { "epoch": 2.600375234521576, "grad_norm": 15.3932466506958, "learning_rate": 3.283924686422414e-05, "loss": 0.0399, "num_input_tokens_seen": 30655024, "step": 15940 }, { "epoch": 2.601190961742393, "grad_norm": 2.5373785495758057, "learning_rate": 3.282992392615659e-05, "loss": 0.0672, "num_input_tokens_seen": 30664448, "step": 15945 }, { "epoch": 2.6020066889632107, "grad_norm": 0.5296359062194824, "learning_rate": 3.282059978061638e-05, "loss": 0.0057, "num_input_tokens_seen": 30674976, "step": 15950 }, { "epoch": 2.6028224161840283, "grad_norm": 0.3822036385536194, "learning_rate": 3.28112744290414e-05, "loss": 0.0062, "num_input_tokens_seen": 30684400, "step": 15955 }, { "epoch": 2.6036381434048455, "grad_norm": 0.045961037278175354, "learning_rate": 3.280194787286974e-05, "loss": 0.0301, "num_input_tokens_seen": 30695216, "step": 15960 }, { "epoch": 2.6044538706256626, "grad_norm": 9.411898612976074, "learning_rate": 3.2792620113539674e-05, "loss": 0.1554, "num_input_tokens_seen": 30705520, "step": 15965 }, { "epoch": 2.60526959784648, "grad_norm": 0.11876645684242249, "learning_rate": 3.278329115248966e-05, "loss": 0.0049, "num_input_tokens_seen": 30716304, "step": 15970 }, { "epoch": 2.6060853250672977, "grad_norm": 1.1487029790878296, "learning_rate": 3.277396099115834e-05, "loss": 0.0662, "num_input_tokens_seen": 30726000, "step": 15975 }, { "epoch": 2.606901052288115, "grad_norm": 0.10453588515520096, "learning_rate": 3.276462963098454e-05, "loss": 0.0738, "num_input_tokens_seen": 30735712, "step": 15980 }, { "epoch": 2.607716779508932, "grad_norm": 0.09837460517883301, "learning_rate": 3.275529707340728e-05, "loss": 0.0046, "num_input_tokens_seen": 30746224, "step": 15985 }, { "epoch": 2.6085325067297496, "grad_norm": 0.016503822058439255, "learning_rate": 3.274596331986574e-05, "loss": 0.0047, "num_input_tokens_seen": 30755248, "step": 15990 }, { "epoch": 2.609348233950567, "grad_norm": 4.411293029785156, "learning_rate": 3.273662837179932e-05, "loss": 0.08, "num_input_tokens_seen": 30765472, "step": 15995 }, { "epoch": 2.6101639611713843, "grad_norm": 0.3860437572002411, "learning_rate": 3.272729223064758e-05, "loss": 0.0025, "num_input_tokens_seen": 30774224, "step": 16000 }, { "epoch": 2.6101639611713843, "eval_loss": 0.2008722573518753, "eval_runtime": 80.6704, "eval_samples_per_second": 33.779, "eval_steps_per_second": 16.896, "num_input_tokens_seen": 30774224, "step": 16000 }, { "epoch": 2.6109796883922014, "grad_norm": 17.124204635620117, "learning_rate": 3.2717954897850264e-05, "loss": 0.1182, "num_input_tokens_seen": 30783648, "step": 16005 }, { "epoch": 2.611795415613019, "grad_norm": 0.27153587341308594, "learning_rate": 3.270861637484733e-05, "loss": 0.0031, "num_input_tokens_seen": 30793296, "step": 16010 }, { "epoch": 2.6126111428338366, "grad_norm": 0.08743090182542801, "learning_rate": 3.2699276663078867e-05, "loss": 0.0435, "num_input_tokens_seen": 30803392, "step": 16015 }, { "epoch": 2.6134268700546537, "grad_norm": 0.028118815273046494, "learning_rate": 3.268993576398519e-05, "loss": 0.0607, "num_input_tokens_seen": 30814528, "step": 16020 }, { "epoch": 2.614242597275471, "grad_norm": 64.26446533203125, "learning_rate": 3.268059367900678e-05, "loss": 0.093, "num_input_tokens_seen": 30824832, "step": 16025 }, { "epoch": 2.6150583244962884, "grad_norm": 2.8458428382873535, "learning_rate": 3.26712504095843e-05, "loss": 0.0803, "num_input_tokens_seen": 30834064, "step": 16030 }, { "epoch": 2.615874051717106, "grad_norm": 6.306473731994629, "learning_rate": 3.2661905957158615e-05, "loss": 0.051, "num_input_tokens_seen": 30844048, "step": 16035 }, { "epoch": 2.616689778937923, "grad_norm": 2.547318458557129, "learning_rate": 3.2652560323170734e-05, "loss": 0.0774, "num_input_tokens_seen": 30854256, "step": 16040 }, { "epoch": 2.6175055061587407, "grad_norm": 0.05460813269019127, "learning_rate": 3.264321350906189e-05, "loss": 0.0053, "num_input_tokens_seen": 30862752, "step": 16045 }, { "epoch": 2.618321233379558, "grad_norm": 6.85199499130249, "learning_rate": 3.263386551627346e-05, "loss": 0.0061, "num_input_tokens_seen": 30871536, "step": 16050 }, { "epoch": 2.6191369606003754, "grad_norm": 0.016856204718351364, "learning_rate": 3.2624516346247055e-05, "loss": 0.0782, "num_input_tokens_seen": 30881200, "step": 16055 }, { "epoch": 2.6199526878211925, "grad_norm": 0.08311116695404053, "learning_rate": 3.2615166000424404e-05, "loss": 0.0407, "num_input_tokens_seen": 30889488, "step": 16060 }, { "epoch": 2.62076841504201, "grad_norm": 1.951041579246521, "learning_rate": 3.260581448024745e-05, "loss": 0.0744, "num_input_tokens_seen": 30898416, "step": 16065 }, { "epoch": 2.621584142262827, "grad_norm": 0.11371038854122162, "learning_rate": 3.2596461787158335e-05, "loss": 0.0395, "num_input_tokens_seen": 30908176, "step": 16070 }, { "epoch": 2.622399869483645, "grad_norm": 0.39430561661720276, "learning_rate": 3.258710792259934e-05, "loss": 0.0071, "num_input_tokens_seen": 30917920, "step": 16075 }, { "epoch": 2.623215596704462, "grad_norm": 0.1092665046453476, "learning_rate": 3.257775288801296e-05, "loss": 0.063, "num_input_tokens_seen": 30927856, "step": 16080 }, { "epoch": 2.6240313239252795, "grad_norm": 0.14323636889457703, "learning_rate": 3.256839668484186e-05, "loss": 0.0892, "num_input_tokens_seen": 30937488, "step": 16085 }, { "epoch": 2.6248470511460966, "grad_norm": 0.03746979311108589, "learning_rate": 3.255903931452888e-05, "loss": 0.0239, "num_input_tokens_seen": 30947712, "step": 16090 }, { "epoch": 2.625662778366914, "grad_norm": 2.023102045059204, "learning_rate": 3.2549680778517045e-05, "loss": 0.0063, "num_input_tokens_seen": 30957008, "step": 16095 }, { "epoch": 2.6264785055877313, "grad_norm": 2.79736065864563, "learning_rate": 3.2540321078249556e-05, "loss": 0.1893, "num_input_tokens_seen": 30967408, "step": 16100 }, { "epoch": 2.627294232808549, "grad_norm": 0.019994115456938744, "learning_rate": 3.2530960215169795e-05, "loss": 0.1187, "num_input_tokens_seen": 30976256, "step": 16105 }, { "epoch": 2.628109960029366, "grad_norm": 0.24755936861038208, "learning_rate": 3.2521598190721345e-05, "loss": 0.0571, "num_input_tokens_seen": 30984976, "step": 16110 }, { "epoch": 2.6289256872501836, "grad_norm": 16.693395614624023, "learning_rate": 3.251223500634792e-05, "loss": 0.0483, "num_input_tokens_seen": 30993232, "step": 16115 }, { "epoch": 2.6297414144710007, "grad_norm": 0.04358195513486862, "learning_rate": 3.2502870663493445e-05, "loss": 0.0085, "num_input_tokens_seen": 31003632, "step": 16120 }, { "epoch": 2.6305571416918183, "grad_norm": 12.712675094604492, "learning_rate": 3.249350516360203e-05, "loss": 0.0417, "num_input_tokens_seen": 31013664, "step": 16125 }, { "epoch": 2.631372868912636, "grad_norm": 0.04553189501166344, "learning_rate": 3.248413850811797e-05, "loss": 0.0031, "num_input_tokens_seen": 31024400, "step": 16130 }, { "epoch": 2.632188596133453, "grad_norm": 3.565410852432251, "learning_rate": 3.2474770698485677e-05, "loss": 0.0637, "num_input_tokens_seen": 31033984, "step": 16135 }, { "epoch": 2.63300432335427, "grad_norm": 0.10767301172018051, "learning_rate": 3.246540173614983e-05, "loss": 0.0541, "num_input_tokens_seen": 31042848, "step": 16140 }, { "epoch": 2.6338200505750877, "grad_norm": 5.794039249420166, "learning_rate": 3.2456031622555197e-05, "loss": 0.0072, "num_input_tokens_seen": 31054128, "step": 16145 }, { "epoch": 2.6346357777959053, "grad_norm": 8.909018516540527, "learning_rate": 3.2446660359146794e-05, "loss": 0.358, "num_input_tokens_seen": 31064128, "step": 16150 }, { "epoch": 2.6354515050167224, "grad_norm": 0.07391408085823059, "learning_rate": 3.2437287947369786e-05, "loss": 0.08, "num_input_tokens_seen": 31074240, "step": 16155 }, { "epoch": 2.6362672322375396, "grad_norm": 16.275470733642578, "learning_rate": 3.2427914388669525e-05, "loss": 0.1286, "num_input_tokens_seen": 31084000, "step": 16160 }, { "epoch": 2.637082959458357, "grad_norm": 13.271692276000977, "learning_rate": 3.241853968449151e-05, "loss": 0.1036, "num_input_tokens_seen": 31093552, "step": 16165 }, { "epoch": 2.6378986866791747, "grad_norm": 0.044913746416568756, "learning_rate": 3.240916383628144e-05, "loss": 0.0507, "num_input_tokens_seen": 31103360, "step": 16170 }, { "epoch": 2.638714413899992, "grad_norm": 4.519645690917969, "learning_rate": 3.239978684548521e-05, "loss": 0.1057, "num_input_tokens_seen": 31113136, "step": 16175 }, { "epoch": 2.639530141120809, "grad_norm": 0.08373378217220306, "learning_rate": 3.239040871354885e-05, "loss": 0.0721, "num_input_tokens_seen": 31123280, "step": 16180 }, { "epoch": 2.6403458683416265, "grad_norm": 0.5234513282775879, "learning_rate": 3.2381029441918596e-05, "loss": 0.0272, "num_input_tokens_seen": 31133056, "step": 16185 }, { "epoch": 2.641161595562444, "grad_norm": 0.9690439701080322, "learning_rate": 3.2371649032040845e-05, "loss": 0.0657, "num_input_tokens_seen": 31143168, "step": 16190 }, { "epoch": 2.6419773227832613, "grad_norm": 0.38221970200538635, "learning_rate": 3.2362267485362174e-05, "loss": 0.2079, "num_input_tokens_seen": 31153328, "step": 16195 }, { "epoch": 2.6427930500040784, "grad_norm": 0.08242252469062805, "learning_rate": 3.235288480332934e-05, "loss": 0.071, "num_input_tokens_seen": 31164304, "step": 16200 }, { "epoch": 2.6427930500040784, "eval_loss": 0.16527904570102692, "eval_runtime": 80.5451, "eval_samples_per_second": 33.832, "eval_steps_per_second": 16.922, "num_input_tokens_seen": 31164304, "step": 16200 }, { "epoch": 2.643608777224896, "grad_norm": 3.278475761413574, "learning_rate": 3.234350098738927e-05, "loss": 0.0552, "num_input_tokens_seen": 31174192, "step": 16205 }, { "epoch": 2.6444245044457135, "grad_norm": 3.770191192626953, "learning_rate": 3.233411603898906e-05, "loss": 0.2037, "num_input_tokens_seen": 31184064, "step": 16210 }, { "epoch": 2.6452402316665307, "grad_norm": 0.13457424938678741, "learning_rate": 3.232472995957599e-05, "loss": 0.0775, "num_input_tokens_seen": 31193760, "step": 16215 }, { "epoch": 2.6460559588873482, "grad_norm": 4.579916954040527, "learning_rate": 3.231534275059751e-05, "loss": 0.0334, "num_input_tokens_seen": 31203232, "step": 16220 }, { "epoch": 2.6468716861081654, "grad_norm": 11.816463470458984, "learning_rate": 3.230595441350125e-05, "loss": 0.1469, "num_input_tokens_seen": 31213088, "step": 16225 }, { "epoch": 2.647687413328983, "grad_norm": 0.33601754903793335, "learning_rate": 3.2296564949735e-05, "loss": 0.0487, "num_input_tokens_seen": 31222496, "step": 16230 }, { "epoch": 2.6485031405498, "grad_norm": 2.4631764888763428, "learning_rate": 3.228717436074675e-05, "loss": 0.12, "num_input_tokens_seen": 31231584, "step": 16235 }, { "epoch": 2.6493188677706176, "grad_norm": 0.07135502249002457, "learning_rate": 3.227778264798463e-05, "loss": 0.1274, "num_input_tokens_seen": 31240416, "step": 16240 }, { "epoch": 2.6501345949914348, "grad_norm": 2.227520704269409, "learning_rate": 3.226838981289698e-05, "loss": 0.0589, "num_input_tokens_seen": 31250608, "step": 16245 }, { "epoch": 2.6509503222122524, "grad_norm": 0.1307193487882614, "learning_rate": 3.225899585693227e-05, "loss": 0.0284, "num_input_tokens_seen": 31260016, "step": 16250 }, { "epoch": 2.6517660494330695, "grad_norm": 0.16868956387043, "learning_rate": 3.224960078153918e-05, "loss": 0.0443, "num_input_tokens_seen": 31270576, "step": 16255 }, { "epoch": 2.652581776653887, "grad_norm": 0.10840169340372086, "learning_rate": 3.224020458816655e-05, "loss": 0.0764, "num_input_tokens_seen": 31279200, "step": 16260 }, { "epoch": 2.653397503874704, "grad_norm": 8.495349884033203, "learning_rate": 3.223080727826337e-05, "loss": 0.127, "num_input_tokens_seen": 31289232, "step": 16265 }, { "epoch": 2.6542132310955218, "grad_norm": 21.47642707824707, "learning_rate": 3.222140885327885e-05, "loss": 0.12, "num_input_tokens_seen": 31299328, "step": 16270 }, { "epoch": 2.655028958316339, "grad_norm": 0.1772170215845108, "learning_rate": 3.221200931466234e-05, "loss": 0.0996, "num_input_tokens_seen": 31309280, "step": 16275 }, { "epoch": 2.6558446855371565, "grad_norm": 0.15833240747451782, "learning_rate": 3.220260866386336e-05, "loss": 0.0514, "num_input_tokens_seen": 31318784, "step": 16280 }, { "epoch": 2.6566604127579736, "grad_norm": 0.7897761464118958, "learning_rate": 3.21932069023316e-05, "loss": 0.1098, "num_input_tokens_seen": 31328608, "step": 16285 }, { "epoch": 2.657476139978791, "grad_norm": 0.03329586982727051, "learning_rate": 3.218380403151695e-05, "loss": 0.0712, "num_input_tokens_seen": 31338192, "step": 16290 }, { "epoch": 2.6582918671996083, "grad_norm": 0.07167500257492065, "learning_rate": 3.217440005286943e-05, "loss": 0.0077, "num_input_tokens_seen": 31348416, "step": 16295 }, { "epoch": 2.659107594420426, "grad_norm": 0.05347897484898567, "learning_rate": 3.216499496783928e-05, "loss": 0.0483, "num_input_tokens_seen": 31358128, "step": 16300 }, { "epoch": 2.659923321641243, "grad_norm": 0.09634024649858475, "learning_rate": 3.2155588777876856e-05, "loss": 0.0195, "num_input_tokens_seen": 31366144, "step": 16305 }, { "epoch": 2.6607390488620606, "grad_norm": 9.107094764709473, "learning_rate": 3.214618148443273e-05, "loss": 0.1207, "num_input_tokens_seen": 31375456, "step": 16310 }, { "epoch": 2.6615547760828777, "grad_norm": 0.0902787297964096, "learning_rate": 3.2136773088957595e-05, "loss": 0.0223, "num_input_tokens_seen": 31384432, "step": 16315 }, { "epoch": 2.6623705033036953, "grad_norm": 0.1022786870598793, "learning_rate": 3.2127363592902374e-05, "loss": 0.2218, "num_input_tokens_seen": 31395216, "step": 16320 }, { "epoch": 2.663186230524513, "grad_norm": 0.09048007428646088, "learning_rate": 3.211795299771812e-05, "loss": 0.0936, "num_input_tokens_seen": 31403744, "step": 16325 }, { "epoch": 2.66400195774533, "grad_norm": 15.388311386108398, "learning_rate": 3.210854130485605e-05, "loss": 0.0825, "num_input_tokens_seen": 31414032, "step": 16330 }, { "epoch": 2.664817684966147, "grad_norm": 3.9954583644866943, "learning_rate": 3.209912851576759e-05, "loss": 0.1527, "num_input_tokens_seen": 31422464, "step": 16335 }, { "epoch": 2.6656334121869647, "grad_norm": 0.07482016831636429, "learning_rate": 3.208971463190431e-05, "loss": 0.0249, "num_input_tokens_seen": 31431984, "step": 16340 }, { "epoch": 2.6664491394077823, "grad_norm": 6.80096435546875, "learning_rate": 3.208029965471793e-05, "loss": 0.067, "num_input_tokens_seen": 31442480, "step": 16345 }, { "epoch": 2.6672648666285994, "grad_norm": 1.581192135810852, "learning_rate": 3.2070883585660364e-05, "loss": 0.0691, "num_input_tokens_seen": 31453088, "step": 16350 }, { "epoch": 2.6680805938494165, "grad_norm": 0.32339850068092346, "learning_rate": 3.20614664261837e-05, "loss": 0.0069, "num_input_tokens_seen": 31461728, "step": 16355 }, { "epoch": 2.668896321070234, "grad_norm": 0.36434948444366455, "learning_rate": 3.205204817774016e-05, "loss": 0.0756, "num_input_tokens_seen": 31471312, "step": 16360 }, { "epoch": 2.6697120482910517, "grad_norm": 5.442041873931885, "learning_rate": 3.204262884178218e-05, "loss": 0.1532, "num_input_tokens_seen": 31482096, "step": 16365 }, { "epoch": 2.670527775511869, "grad_norm": 0.22235794365406036, "learning_rate": 3.2033208419762314e-05, "loss": 0.0702, "num_input_tokens_seen": 31491632, "step": 16370 }, { "epoch": 2.671343502732686, "grad_norm": 0.37996554374694824, "learning_rate": 3.2023786913133344e-05, "loss": 0.0876, "num_input_tokens_seen": 31500608, "step": 16375 }, { "epoch": 2.6721592299535035, "grad_norm": 9.016222953796387, "learning_rate": 3.201436432334816e-05, "loss": 0.0896, "num_input_tokens_seen": 31510192, "step": 16380 }, { "epoch": 2.672974957174321, "grad_norm": 0.15481285750865936, "learning_rate": 3.2004940651859844e-05, "loss": 0.0065, "num_input_tokens_seen": 31520352, "step": 16385 }, { "epoch": 2.6737906843951382, "grad_norm": 3.180870532989502, "learning_rate": 3.1995515900121655e-05, "loss": 0.1561, "num_input_tokens_seen": 31530560, "step": 16390 }, { "epoch": 2.6746064116159554, "grad_norm": 0.055673398077487946, "learning_rate": 3.1986090069587e-05, "loss": 0.1499, "num_input_tokens_seen": 31540560, "step": 16395 }, { "epoch": 2.675422138836773, "grad_norm": 8.603291511535645, "learning_rate": 3.1976663161709466e-05, "loss": 0.1059, "num_input_tokens_seen": 31548832, "step": 16400 }, { "epoch": 2.675422138836773, "eval_loss": 0.17207415401935577, "eval_runtime": 80.5435, "eval_samples_per_second": 33.833, "eval_steps_per_second": 16.923, "num_input_tokens_seen": 31548832, "step": 16400 }, { "epoch": 2.6762378660575905, "grad_norm": 1.3995842933654785, "learning_rate": 3.196723517794279e-05, "loss": 0.0054, "num_input_tokens_seen": 31559040, "step": 16405 }, { "epoch": 2.6770535932784076, "grad_norm": 10.4619722366333, "learning_rate": 3.19578061197409e-05, "loss": 0.1335, "num_input_tokens_seen": 31569376, "step": 16410 }, { "epoch": 2.677869320499225, "grad_norm": 0.5852776765823364, "learning_rate": 3.194837598855787e-05, "loss": 0.1549, "num_input_tokens_seen": 31578896, "step": 16415 }, { "epoch": 2.6786850477200423, "grad_norm": 0.12173879146575928, "learning_rate": 3.193894478584794e-05, "loss": 0.0949, "num_input_tokens_seen": 31588560, "step": 16420 }, { "epoch": 2.67950077494086, "grad_norm": 3.289397716522217, "learning_rate": 3.192951251306553e-05, "loss": 0.169, "num_input_tokens_seen": 31598592, "step": 16425 }, { "epoch": 2.680316502161677, "grad_norm": 0.23264598846435547, "learning_rate": 3.192007917166521e-05, "loss": 0.1302, "num_input_tokens_seen": 31608800, "step": 16430 }, { "epoch": 2.6811322293824946, "grad_norm": 0.9112705588340759, "learning_rate": 3.191064476310171e-05, "loss": 0.1122, "num_input_tokens_seen": 31618656, "step": 16435 }, { "epoch": 2.6819479566033118, "grad_norm": 0.30793124437332153, "learning_rate": 3.1901209288829944e-05, "loss": 0.3881, "num_input_tokens_seen": 31628944, "step": 16440 }, { "epoch": 2.6827636838241293, "grad_norm": 0.10559048503637314, "learning_rate": 3.1891772750304985e-05, "loss": 0.0911, "num_input_tokens_seen": 31638384, "step": 16445 }, { "epoch": 2.6835794110449465, "grad_norm": 1.471740961074829, "learning_rate": 3.188233514898206e-05, "loss": 0.0361, "num_input_tokens_seen": 31648016, "step": 16450 }, { "epoch": 2.684395138265764, "grad_norm": 0.28553691506385803, "learning_rate": 3.187289648631657e-05, "loss": 0.1001, "num_input_tokens_seen": 31658640, "step": 16455 }, { "epoch": 2.685210865486581, "grad_norm": 0.22659878432750702, "learning_rate": 3.186345676376406e-05, "loss": 0.1103, "num_input_tokens_seen": 31668736, "step": 16460 }, { "epoch": 2.6860265927073987, "grad_norm": 3.645519256591797, "learning_rate": 3.1854015982780275e-05, "loss": 0.0658, "num_input_tokens_seen": 31679168, "step": 16465 }, { "epoch": 2.686842319928216, "grad_norm": 0.08945980668067932, "learning_rate": 3.1844574144821084e-05, "loss": 0.0079, "num_input_tokens_seen": 31689216, "step": 16470 }, { "epoch": 2.6876580471490334, "grad_norm": 0.2757154107093811, "learning_rate": 3.1835131251342554e-05, "loss": 0.0682, "num_input_tokens_seen": 31699600, "step": 16475 }, { "epoch": 2.6884737743698506, "grad_norm": 4.49289083480835, "learning_rate": 3.182568730380089e-05, "loss": 0.0197, "num_input_tokens_seen": 31707920, "step": 16480 }, { "epoch": 2.689289501590668, "grad_norm": 0.39324110746383667, "learning_rate": 3.181624230365245e-05, "loss": 0.011, "num_input_tokens_seen": 31717648, "step": 16485 }, { "epoch": 2.6901052288114853, "grad_norm": 9.281904220581055, "learning_rate": 3.180679625235381e-05, "loss": 0.0856, "num_input_tokens_seen": 31727008, "step": 16490 }, { "epoch": 2.690920956032303, "grad_norm": 4.321330547332764, "learning_rate": 3.1797349151361646e-05, "loss": 0.0223, "num_input_tokens_seen": 31737200, "step": 16495 }, { "epoch": 2.6917366832531204, "grad_norm": 0.38028445839881897, "learning_rate": 3.178790100213281e-05, "loss": 0.1098, "num_input_tokens_seen": 31746896, "step": 16500 }, { "epoch": 2.6925524104739376, "grad_norm": 0.2764654755592346, "learning_rate": 3.1778451806124346e-05, "loss": 0.1301, "num_input_tokens_seen": 31757776, "step": 16505 }, { "epoch": 2.6933681376947547, "grad_norm": 0.2944759726524353, "learning_rate": 3.176900156479342e-05, "loss": 0.0084, "num_input_tokens_seen": 31768720, "step": 16510 }, { "epoch": 2.6941838649155723, "grad_norm": 14.710145950317383, "learning_rate": 3.17595502795974e-05, "loss": 0.0185, "num_input_tokens_seen": 31777712, "step": 16515 }, { "epoch": 2.69499959213639, "grad_norm": 0.06104408949613571, "learning_rate": 3.175009795199377e-05, "loss": 0.0089, "num_input_tokens_seen": 31786944, "step": 16520 }, { "epoch": 2.695815319357207, "grad_norm": 3.9001238346099854, "learning_rate": 3.1740644583440224e-05, "loss": 0.0476, "num_input_tokens_seen": 31796976, "step": 16525 }, { "epoch": 2.696631046578024, "grad_norm": 0.06549728661775589, "learning_rate": 3.173119017539457e-05, "loss": 0.0695, "num_input_tokens_seen": 31807088, "step": 16530 }, { "epoch": 2.6974467737988417, "grad_norm": 0.8955299258232117, "learning_rate": 3.172173472931479e-05, "loss": 0.0923, "num_input_tokens_seen": 31817264, "step": 16535 }, { "epoch": 2.6982625010196593, "grad_norm": 2.8272693157196045, "learning_rate": 3.1712278246659055e-05, "loss": 0.0275, "num_input_tokens_seen": 31826416, "step": 16540 }, { "epoch": 2.6990782282404764, "grad_norm": 18.79732322692871, "learning_rate": 3.170282072888566e-05, "loss": 0.049, "num_input_tokens_seen": 31834832, "step": 16545 }, { "epoch": 2.6998939554612935, "grad_norm": 46.900062561035156, "learning_rate": 3.169336217745307e-05, "loss": 0.0939, "num_input_tokens_seen": 31845072, "step": 16550 }, { "epoch": 2.700709682682111, "grad_norm": 0.012757394462823868, "learning_rate": 3.1683902593819924e-05, "loss": 0.0032, "num_input_tokens_seen": 31855920, "step": 16555 }, { "epoch": 2.7015254099029287, "grad_norm": 0.022984705865383148, "learning_rate": 3.1674441979445e-05, "loss": 0.0013, "num_input_tokens_seen": 31866320, "step": 16560 }, { "epoch": 2.702341137123746, "grad_norm": 2.285409450531006, "learning_rate": 3.166498033578725e-05, "loss": 0.0994, "num_input_tokens_seen": 31876384, "step": 16565 }, { "epoch": 2.703156864344563, "grad_norm": 0.041536543518304825, "learning_rate": 3.165551766430578e-05, "loss": 0.1874, "num_input_tokens_seen": 31885936, "step": 16570 }, { "epoch": 2.7039725915653805, "grad_norm": 0.10091151297092438, "learning_rate": 3.164605396645984e-05, "loss": 0.0451, "num_input_tokens_seen": 31896592, "step": 16575 }, { "epoch": 2.704788318786198, "grad_norm": 0.17758750915527344, "learning_rate": 3.163658924370886e-05, "loss": 0.0738, "num_input_tokens_seen": 31907296, "step": 16580 }, { "epoch": 2.705604046007015, "grad_norm": 0.02639589086174965, "learning_rate": 3.1627123497512415e-05, "loss": 0.0852, "num_input_tokens_seen": 31917024, "step": 16585 }, { "epoch": 2.7064197732278323, "grad_norm": 0.12438645213842392, "learning_rate": 3.1617656729330245e-05, "loss": 0.0578, "num_input_tokens_seen": 31925408, "step": 16590 }, { "epoch": 2.70723550044865, "grad_norm": 11.224960327148438, "learning_rate": 3.1608188940622255e-05, "loss": 0.0783, "num_input_tokens_seen": 31933904, "step": 16595 }, { "epoch": 2.7080512276694675, "grad_norm": 3.637103796005249, "learning_rate": 3.159872013284847e-05, "loss": 0.0747, "num_input_tokens_seen": 31943568, "step": 16600 }, { "epoch": 2.7080512276694675, "eval_loss": 0.1902150809764862, "eval_runtime": 80.6496, "eval_samples_per_second": 33.788, "eval_steps_per_second": 16.9, "num_input_tokens_seen": 31943568, "step": 16600 }, { "epoch": 2.7088669548902846, "grad_norm": 0.11509933322668076, "learning_rate": 3.1589250307469134e-05, "loss": 0.09, "num_input_tokens_seen": 31953312, "step": 16605 }, { "epoch": 2.709682682111102, "grad_norm": 0.22208444774150848, "learning_rate": 3.1579779465944586e-05, "loss": 0.0663, "num_input_tokens_seen": 31963616, "step": 16610 }, { "epoch": 2.7104984093319193, "grad_norm": 0.07440277189016342, "learning_rate": 3.1570307609735363e-05, "loss": 0.0595, "num_input_tokens_seen": 31973632, "step": 16615 }, { "epoch": 2.711314136552737, "grad_norm": 0.03986494615674019, "learning_rate": 3.156083474030213e-05, "loss": 0.1544, "num_input_tokens_seen": 31983904, "step": 16620 }, { "epoch": 2.712129863773554, "grad_norm": 0.04983624815940857, "learning_rate": 3.155136085910573e-05, "loss": 0.0984, "num_input_tokens_seen": 31994096, "step": 16625 }, { "epoch": 2.7129455909943716, "grad_norm": 0.02154393121600151, "learning_rate": 3.154188596760717e-05, "loss": 0.2009, "num_input_tokens_seen": 32003600, "step": 16630 }, { "epoch": 2.7137613182151887, "grad_norm": 30.001293182373047, "learning_rate": 3.153241006726757e-05, "loss": 0.1312, "num_input_tokens_seen": 32013168, "step": 16635 }, { "epoch": 2.7145770454360063, "grad_norm": 0.1478712111711502, "learning_rate": 3.152293315954825e-05, "loss": 0.1565, "num_input_tokens_seen": 32023008, "step": 16640 }, { "epoch": 2.7153927726568234, "grad_norm": 9.745245933532715, "learning_rate": 3.1513455245910666e-05, "loss": 0.2162, "num_input_tokens_seen": 32031840, "step": 16645 }, { "epoch": 2.716208499877641, "grad_norm": 0.09713097661733627, "learning_rate": 3.150397632781643e-05, "loss": 0.1436, "num_input_tokens_seen": 32041936, "step": 16650 }, { "epoch": 2.717024227098458, "grad_norm": 0.0521564744412899, "learning_rate": 3.149449640672731e-05, "loss": 0.0385, "num_input_tokens_seen": 32050896, "step": 16655 }, { "epoch": 2.7178399543192757, "grad_norm": 3.517650604248047, "learning_rate": 3.148501548410523e-05, "loss": 0.1869, "num_input_tokens_seen": 32060736, "step": 16660 }, { "epoch": 2.718655681540093, "grad_norm": 0.18670149147510529, "learning_rate": 3.1475533561412256e-05, "loss": 0.1175, "num_input_tokens_seen": 32069792, "step": 16665 }, { "epoch": 2.7194714087609104, "grad_norm": 20.926990509033203, "learning_rate": 3.146605064011065e-05, "loss": 0.0788, "num_input_tokens_seen": 32079344, "step": 16670 }, { "epoch": 2.7202871359817276, "grad_norm": 0.1108156144618988, "learning_rate": 3.145656672166277e-05, "loss": 0.0095, "num_input_tokens_seen": 32087984, "step": 16675 }, { "epoch": 2.721102863202545, "grad_norm": 7.3731913566589355, "learning_rate": 3.144708180753116e-05, "loss": 0.079, "num_input_tokens_seen": 32097648, "step": 16680 }, { "epoch": 2.7219185904233623, "grad_norm": 1.9543476104736328, "learning_rate": 3.143759589917851e-05, "loss": 0.034, "num_input_tokens_seen": 32107232, "step": 16685 }, { "epoch": 2.72273431764418, "grad_norm": 0.09713885933160782, "learning_rate": 3.142810899806768e-05, "loss": 0.0862, "num_input_tokens_seen": 32116496, "step": 16690 }, { "epoch": 2.7235500448649974, "grad_norm": 5.08942174911499, "learning_rate": 3.141862110566166e-05, "loss": 0.0777, "num_input_tokens_seen": 32124512, "step": 16695 }, { "epoch": 2.7243657720858145, "grad_norm": 0.09973885864019394, "learning_rate": 3.1409132223423606e-05, "loss": 0.0253, "num_input_tokens_seen": 32134704, "step": 16700 }, { "epoch": 2.7251814993066317, "grad_norm": 0.06758391112089157, "learning_rate": 3.139964235281682e-05, "loss": 0.1311, "num_input_tokens_seen": 32144496, "step": 16705 }, { "epoch": 2.7259972265274492, "grad_norm": 0.33809971809387207, "learning_rate": 3.139015149530476e-05, "loss": 0.0063, "num_input_tokens_seen": 32153664, "step": 16710 }, { "epoch": 2.726812953748267, "grad_norm": 0.16983255743980408, "learning_rate": 3.1380659652351034e-05, "loss": 0.0418, "num_input_tokens_seen": 32162400, "step": 16715 }, { "epoch": 2.727628680969084, "grad_norm": 0.47525933384895325, "learning_rate": 3.137116682541941e-05, "loss": 0.0844, "num_input_tokens_seen": 32172544, "step": 16720 }, { "epoch": 2.728444408189901, "grad_norm": 0.1320856809616089, "learning_rate": 3.136167301597379e-05, "loss": 0.005, "num_input_tokens_seen": 32181792, "step": 16725 }, { "epoch": 2.7292601354107187, "grad_norm": 14.176803588867188, "learning_rate": 3.1352178225478254e-05, "loss": 0.0739, "num_input_tokens_seen": 32191408, "step": 16730 }, { "epoch": 2.7300758626315362, "grad_norm": 0.013720488175749779, "learning_rate": 3.1342682455396996e-05, "loss": 0.0587, "num_input_tokens_seen": 32200432, "step": 16735 }, { "epoch": 2.7308915898523534, "grad_norm": 0.996284008026123, "learning_rate": 3.133318570719441e-05, "loss": 0.1135, "num_input_tokens_seen": 32211424, "step": 16740 }, { "epoch": 2.7317073170731705, "grad_norm": 5.633542060852051, "learning_rate": 3.132368798233499e-05, "loss": 0.0611, "num_input_tokens_seen": 32221408, "step": 16745 }, { "epoch": 2.732523044293988, "grad_norm": 1.312124252319336, "learning_rate": 3.131418928228342e-05, "loss": 0.0103, "num_input_tokens_seen": 32230480, "step": 16750 }, { "epoch": 2.7333387715148056, "grad_norm": 0.1812501698732376, "learning_rate": 3.1304689608504514e-05, "loss": 0.0995, "num_input_tokens_seen": 32239664, "step": 16755 }, { "epoch": 2.7341544987356228, "grad_norm": 0.2647080719470978, "learning_rate": 3.129518896246324e-05, "loss": 0.0556, "num_input_tokens_seen": 32249072, "step": 16760 }, { "epoch": 2.73497022595644, "grad_norm": 4.511878967285156, "learning_rate": 3.128568734562472e-05, "loss": 0.2106, "num_input_tokens_seen": 32260960, "step": 16765 }, { "epoch": 2.7357859531772575, "grad_norm": 0.09365584701299667, "learning_rate": 3.127618475945421e-05, "loss": 0.0562, "num_input_tokens_seen": 32271696, "step": 16770 }, { "epoch": 2.736601680398075, "grad_norm": 0.036534011363983154, "learning_rate": 3.126668120541715e-05, "loss": 0.0064, "num_input_tokens_seen": 32281328, "step": 16775 }, { "epoch": 2.737417407618892, "grad_norm": 1.912710428237915, "learning_rate": 3.1257176684979096e-05, "loss": 0.1509, "num_input_tokens_seen": 32291984, "step": 16780 }, { "epoch": 2.7382331348397098, "grad_norm": 0.040170371532440186, "learning_rate": 3.124767119960576e-05, "loss": 0.2175, "num_input_tokens_seen": 32300128, "step": 16785 }, { "epoch": 2.739048862060527, "grad_norm": 9.58408260345459, "learning_rate": 3.123816475076301e-05, "loss": 0.045, "num_input_tokens_seen": 32310656, "step": 16790 }, { "epoch": 2.7398645892813445, "grad_norm": 2.6504690647125244, "learning_rate": 3.122865733991687e-05, "loss": 0.0592, "num_input_tokens_seen": 32319248, "step": 16795 }, { "epoch": 2.7406803165021616, "grad_norm": 0.24900855123996735, "learning_rate": 3.1219148968533486e-05, "loss": 0.0031, "num_input_tokens_seen": 32327088, "step": 16800 }, { "epoch": 2.7406803165021616, "eval_loss": 0.17834709584712982, "eval_runtime": 80.6056, "eval_samples_per_second": 33.807, "eval_steps_per_second": 16.909, "num_input_tokens_seen": 32327088, "step": 16800 }, { "epoch": 2.741496043722979, "grad_norm": 4.576326370239258, "learning_rate": 3.120963963807918e-05, "loss": 0.0551, "num_input_tokens_seen": 32337040, "step": 16805 }, { "epoch": 2.7423117709437963, "grad_norm": 6.334578037261963, "learning_rate": 3.12001293500204e-05, "loss": 0.096, "num_input_tokens_seen": 32346800, "step": 16810 }, { "epoch": 2.743127498164614, "grad_norm": 3.022615432739258, "learning_rate": 3.1190618105823765e-05, "loss": 0.2332, "num_input_tokens_seen": 32355728, "step": 16815 }, { "epoch": 2.743943225385431, "grad_norm": 0.01889379695057869, "learning_rate": 3.118110590695603e-05, "loss": 0.0427, "num_input_tokens_seen": 32364528, "step": 16820 }, { "epoch": 2.7447589526062486, "grad_norm": 0.13113808631896973, "learning_rate": 3.117159275488407e-05, "loss": 0.015, "num_input_tokens_seen": 32374448, "step": 16825 }, { "epoch": 2.7455746798270657, "grad_norm": 2.4752559661865234, "learning_rate": 3.1162078651074956e-05, "loss": 0.223, "num_input_tokens_seen": 32383760, "step": 16830 }, { "epoch": 2.7463904070478833, "grad_norm": 0.026394136250019073, "learning_rate": 3.1152563596995885e-05, "loss": 0.0936, "num_input_tokens_seen": 32393632, "step": 16835 }, { "epoch": 2.7472061342687004, "grad_norm": 0.09343540668487549, "learning_rate": 3.1143047594114186e-05, "loss": 0.0978, "num_input_tokens_seen": 32403456, "step": 16840 }, { "epoch": 2.748021861489518, "grad_norm": 0.20874986052513123, "learning_rate": 3.113353064389734e-05, "loss": 0.0496, "num_input_tokens_seen": 32412656, "step": 16845 }, { "epoch": 2.748837588710335, "grad_norm": 0.05170416459441185, "learning_rate": 3.1124012747812993e-05, "loss": 0.0461, "num_input_tokens_seen": 32422992, "step": 16850 }, { "epoch": 2.7496533159311527, "grad_norm": 0.044776469469070435, "learning_rate": 3.1114493907328936e-05, "loss": 0.0732, "num_input_tokens_seen": 32432816, "step": 16855 }, { "epoch": 2.75046904315197, "grad_norm": 6.415893077850342, "learning_rate": 3.110497412391306e-05, "loss": 0.0881, "num_input_tokens_seen": 32442528, "step": 16860 }, { "epoch": 2.7512847703727874, "grad_norm": 1.5927220582962036, "learning_rate": 3.1095453399033466e-05, "loss": 0.0321, "num_input_tokens_seen": 32450272, "step": 16865 }, { "epoch": 2.7521004975936045, "grad_norm": 0.08637713640928268, "learning_rate": 3.108593173415835e-05, "loss": 0.0293, "num_input_tokens_seen": 32461152, "step": 16870 }, { "epoch": 2.752916224814422, "grad_norm": 0.07184837013483047, "learning_rate": 3.107640913075609e-05, "loss": 0.241, "num_input_tokens_seen": 32470832, "step": 16875 }, { "epoch": 2.7537319520352392, "grad_norm": 2.5349411964416504, "learning_rate": 3.106688559029517e-05, "loss": 0.2238, "num_input_tokens_seen": 32480048, "step": 16880 }, { "epoch": 2.754547679256057, "grad_norm": 0.07815702259540558, "learning_rate": 3.105736111424425e-05, "loss": 0.1362, "num_input_tokens_seen": 32488848, "step": 16885 }, { "epoch": 2.7553634064768744, "grad_norm": 3.0363848209381104, "learning_rate": 3.1047835704072136e-05, "loss": 0.146, "num_input_tokens_seen": 32497328, "step": 16890 }, { "epoch": 2.7561791336976915, "grad_norm": 0.18412187695503235, "learning_rate": 3.103830936124775e-05, "loss": 0.1356, "num_input_tokens_seen": 32506832, "step": 16895 }, { "epoch": 2.7569948609185086, "grad_norm": 2.0925562381744385, "learning_rate": 3.102878208724018e-05, "loss": 0.1053, "num_input_tokens_seen": 32517664, "step": 16900 }, { "epoch": 2.7578105881393262, "grad_norm": 21.271055221557617, "learning_rate": 3.101925388351865e-05, "loss": 0.0794, "num_input_tokens_seen": 32527504, "step": 16905 }, { "epoch": 2.758626315360144, "grad_norm": 0.13240988552570343, "learning_rate": 3.1009724751552515e-05, "loss": 0.0519, "num_input_tokens_seen": 32536928, "step": 16910 }, { "epoch": 2.759442042580961, "grad_norm": 0.14117726683616638, "learning_rate": 3.100019469281131e-05, "loss": 0.0608, "num_input_tokens_seen": 32547232, "step": 16915 }, { "epoch": 2.760257769801778, "grad_norm": 0.48430973291397095, "learning_rate": 3.0990663708764685e-05, "loss": 0.2425, "num_input_tokens_seen": 32557536, "step": 16920 }, { "epoch": 2.7610734970225956, "grad_norm": 0.21623821556568146, "learning_rate": 3.098113180088243e-05, "loss": 0.1111, "num_input_tokens_seen": 32567632, "step": 16925 }, { "epoch": 2.761889224243413, "grad_norm": 0.4968336820602417, "learning_rate": 3.097159897063448e-05, "loss": 0.0115, "num_input_tokens_seen": 32577760, "step": 16930 }, { "epoch": 2.7627049514642303, "grad_norm": 0.2959360182285309, "learning_rate": 3.096206521949094e-05, "loss": 0.0629, "num_input_tokens_seen": 32586704, "step": 16935 }, { "epoch": 2.7635206786850475, "grad_norm": 0.20296329259872437, "learning_rate": 3.0952530548922006e-05, "loss": 0.0124, "num_input_tokens_seen": 32596448, "step": 16940 }, { "epoch": 2.764336405905865, "grad_norm": 0.05288166180253029, "learning_rate": 3.0942994960398064e-05, "loss": 0.0704, "num_input_tokens_seen": 32606464, "step": 16945 }, { "epoch": 2.7651521331266826, "grad_norm": 0.18069575726985931, "learning_rate": 3.093345845538961e-05, "loss": 0.1101, "num_input_tokens_seen": 32615792, "step": 16950 }, { "epoch": 2.7659678603474998, "grad_norm": 0.7753559350967407, "learning_rate": 3.09239210353673e-05, "loss": 0.0539, "num_input_tokens_seen": 32625168, "step": 16955 }, { "epoch": 2.766783587568317, "grad_norm": 0.6407883167266846, "learning_rate": 3.0914382701801926e-05, "loss": 0.0056, "num_input_tokens_seen": 32634976, "step": 16960 }, { "epoch": 2.7675993147891345, "grad_norm": 5.613101005554199, "learning_rate": 3.090484345616441e-05, "loss": 0.0724, "num_input_tokens_seen": 32644800, "step": 16965 }, { "epoch": 2.768415042009952, "grad_norm": 4.663684368133545, "learning_rate": 3.0895303299925825e-05, "loss": 0.1398, "num_input_tokens_seen": 32653856, "step": 16970 }, { "epoch": 2.769230769230769, "grad_norm": 0.0839235782623291, "learning_rate": 3.0885762234557393e-05, "loss": 0.1515, "num_input_tokens_seen": 32664192, "step": 16975 }, { "epoch": 2.7700464964515867, "grad_norm": 0.311222642660141, "learning_rate": 3.087622026153045e-05, "loss": 0.048, "num_input_tokens_seen": 32674224, "step": 16980 }, { "epoch": 2.770862223672404, "grad_norm": 3.1258432865142822, "learning_rate": 3.086667738231651e-05, "loss": 0.1774, "num_input_tokens_seen": 32683088, "step": 16985 }, { "epoch": 2.7716779508932214, "grad_norm": 2.649247169494629, "learning_rate": 3.085713359838718e-05, "loss": 0.1428, "num_input_tokens_seen": 32692800, "step": 16990 }, { "epoch": 2.7724936781140386, "grad_norm": 0.24702635407447815, "learning_rate": 3.084758891121425e-05, "loss": 0.0316, "num_input_tokens_seen": 32703104, "step": 16995 }, { "epoch": 2.773309405334856, "grad_norm": 3.6091954708099365, "learning_rate": 3.083804332226963e-05, "loss": 0.0657, "num_input_tokens_seen": 32713728, "step": 17000 }, { "epoch": 2.773309405334856, "eval_loss": 0.1591905653476715, "eval_runtime": 81.1463, "eval_samples_per_second": 33.581, "eval_steps_per_second": 16.797, "num_input_tokens_seen": 32713728, "step": 17000 }, { "epoch": 2.7741251325556733, "grad_norm": 22.30855941772461, "learning_rate": 3.082849683302536e-05, "loss": 0.1297, "num_input_tokens_seen": 32723696, "step": 17005 }, { "epoch": 2.774940859776491, "grad_norm": 0.285813570022583, "learning_rate": 3.081894944495363e-05, "loss": 0.0719, "num_input_tokens_seen": 32733120, "step": 17010 }, { "epoch": 2.775756586997308, "grad_norm": 0.3873692750930786, "learning_rate": 3.080940115952677e-05, "loss": 0.0362, "num_input_tokens_seen": 32741472, "step": 17015 }, { "epoch": 2.7765723142181256, "grad_norm": 5.100844383239746, "learning_rate": 3.0799851978217245e-05, "loss": 0.1388, "num_input_tokens_seen": 32750832, "step": 17020 }, { "epoch": 2.7773880414389427, "grad_norm": 7.554279327392578, "learning_rate": 3.0790301902497666e-05, "loss": 0.169, "num_input_tokens_seen": 32760256, "step": 17025 }, { "epoch": 2.7782037686597603, "grad_norm": 21.638582229614258, "learning_rate": 3.078075093384076e-05, "loss": 0.0223, "num_input_tokens_seen": 32770256, "step": 17030 }, { "epoch": 2.7790194958805774, "grad_norm": 4.520464897155762, "learning_rate": 3.077119907371942e-05, "loss": 0.1892, "num_input_tokens_seen": 32781088, "step": 17035 }, { "epoch": 2.779835223101395, "grad_norm": 22.07815170288086, "learning_rate": 3.076164632360666e-05, "loss": 0.1545, "num_input_tokens_seen": 32790400, "step": 17040 }, { "epoch": 2.780650950322212, "grad_norm": 1.725791573524475, "learning_rate": 3.075209268497563e-05, "loss": 0.0759, "num_input_tokens_seen": 32800464, "step": 17045 }, { "epoch": 2.7814666775430297, "grad_norm": 2.755052328109741, "learning_rate": 3.074253815929961e-05, "loss": 0.0757, "num_input_tokens_seen": 32810272, "step": 17050 }, { "epoch": 2.782282404763847, "grad_norm": 0.20379939675331116, "learning_rate": 3.0732982748052054e-05, "loss": 0.0058, "num_input_tokens_seen": 32820768, "step": 17055 }, { "epoch": 2.7830981319846644, "grad_norm": 0.4281162619590759, "learning_rate": 3.072342645270651e-05, "loss": 0.118, "num_input_tokens_seen": 32831200, "step": 17060 }, { "epoch": 2.783913859205482, "grad_norm": 15.090806007385254, "learning_rate": 3.071386927473668e-05, "loss": 0.1283, "num_input_tokens_seen": 32839216, "step": 17065 }, { "epoch": 2.784729586426299, "grad_norm": 4.6464338302612305, "learning_rate": 3.0704311215616404e-05, "loss": 0.2465, "num_input_tokens_seen": 32848688, "step": 17070 }, { "epoch": 2.785545313647116, "grad_norm": 0.7337316870689392, "learning_rate": 3.0694752276819656e-05, "loss": 0.0574, "num_input_tokens_seen": 32857152, "step": 17075 }, { "epoch": 2.786361040867934, "grad_norm": 4.126342296600342, "learning_rate": 3.068519245982054e-05, "loss": 0.0486, "num_input_tokens_seen": 32866496, "step": 17080 }, { "epoch": 2.7871767680887514, "grad_norm": 5.395965099334717, "learning_rate": 3.0675631766093304e-05, "loss": 0.1835, "num_input_tokens_seen": 32875088, "step": 17085 }, { "epoch": 2.7879924953095685, "grad_norm": 0.135576993227005, "learning_rate": 3.066607019711232e-05, "loss": 0.1057, "num_input_tokens_seen": 32885136, "step": 17090 }, { "epoch": 2.7888082225303856, "grad_norm": 1.2117407321929932, "learning_rate": 3.065650775435211e-05, "loss": 0.1708, "num_input_tokens_seen": 32894896, "step": 17095 }, { "epoch": 2.789623949751203, "grad_norm": 0.27790945768356323, "learning_rate": 3.0646944439287326e-05, "loss": 0.0372, "num_input_tokens_seen": 32903440, "step": 17100 }, { "epoch": 2.7904396769720208, "grad_norm": 0.31325864791870117, "learning_rate": 3.0637380253392736e-05, "loss": 0.0189, "num_input_tokens_seen": 32912624, "step": 17105 }, { "epoch": 2.791255404192838, "grad_norm": 0.3175453245639801, "learning_rate": 3.062781519814327e-05, "loss": 0.0814, "num_input_tokens_seen": 32921216, "step": 17110 }, { "epoch": 2.792071131413655, "grad_norm": 9.367076873779297, "learning_rate": 3.0618249275013985e-05, "loss": 0.2312, "num_input_tokens_seen": 32931136, "step": 17115 }, { "epoch": 2.7928868586344726, "grad_norm": 0.34119245409965515, "learning_rate": 3.060868248548005e-05, "loss": 0.1592, "num_input_tokens_seen": 32941056, "step": 17120 }, { "epoch": 2.79370258585529, "grad_norm": 0.19590707123279572, "learning_rate": 3.0599114831016796e-05, "loss": 0.051, "num_input_tokens_seen": 32951008, "step": 17125 }, { "epoch": 2.7945183130761073, "grad_norm": 0.06779064238071442, "learning_rate": 3.0589546313099666e-05, "loss": 0.0336, "num_input_tokens_seen": 32958864, "step": 17130 }, { "epoch": 2.7953340402969244, "grad_norm": 0.7423411011695862, "learning_rate": 3.0579976933204255e-05, "loss": 0.0054, "num_input_tokens_seen": 32967984, "step": 17135 }, { "epoch": 2.796149767517742, "grad_norm": 4.348735332489014, "learning_rate": 3.0570406692806284e-05, "loss": 0.0651, "num_input_tokens_seen": 32978512, "step": 17140 }, { "epoch": 2.7969654947385596, "grad_norm": 0.046053335070610046, "learning_rate": 3.05608355933816e-05, "loss": 0.0103, "num_input_tokens_seen": 32987952, "step": 17145 }, { "epoch": 2.7977812219593767, "grad_norm": 0.2537609934806824, "learning_rate": 3.055126363640618e-05, "loss": 0.0043, "num_input_tokens_seen": 32997104, "step": 17150 }, { "epoch": 2.7985969491801943, "grad_norm": 15.549437522888184, "learning_rate": 3.0541690823356146e-05, "loss": 0.0384, "num_input_tokens_seen": 33007216, "step": 17155 }, { "epoch": 2.7994126764010114, "grad_norm": 0.07454269379377365, "learning_rate": 3.053211715570775e-05, "loss": 0.0779, "num_input_tokens_seen": 33017280, "step": 17160 }, { "epoch": 2.800228403621829, "grad_norm": 7.692463397979736, "learning_rate": 3.052254263493736e-05, "loss": 0.1276, "num_input_tokens_seen": 33026752, "step": 17165 }, { "epoch": 2.801044130842646, "grad_norm": 0.27779167890548706, "learning_rate": 3.0512967262521498e-05, "loss": 0.1132, "num_input_tokens_seen": 33034672, "step": 17170 }, { "epoch": 2.8018598580634637, "grad_norm": 0.14414694905281067, "learning_rate": 3.0503391039936803e-05, "loss": 0.0092, "num_input_tokens_seen": 33043616, "step": 17175 }, { "epoch": 2.802675585284281, "grad_norm": 0.13373015820980072, "learning_rate": 3.0493813968660056e-05, "loss": 0.062, "num_input_tokens_seen": 33053808, "step": 17180 }, { "epoch": 2.8034913125050984, "grad_norm": 13.765204429626465, "learning_rate": 3.0484236050168153e-05, "loss": 0.0585, "num_input_tokens_seen": 33063968, "step": 17185 }, { "epoch": 2.8043070397259156, "grad_norm": 0.02121742255985737, "learning_rate": 3.0474657285938123e-05, "loss": 0.0055, "num_input_tokens_seen": 33074368, "step": 17190 }, { "epoch": 2.805122766946733, "grad_norm": 5.712313175201416, "learning_rate": 3.046507767744715e-05, "loss": 0.0631, "num_input_tokens_seen": 33084160, "step": 17195 }, { "epoch": 2.8059384941675503, "grad_norm": 4.132228851318359, "learning_rate": 3.045549722617252e-05, "loss": 0.0995, "num_input_tokens_seen": 33093744, "step": 17200 }, { "epoch": 2.8059384941675503, "eval_loss": 0.20447540283203125, "eval_runtime": 80.7107, "eval_samples_per_second": 33.763, "eval_steps_per_second": 16.887, "num_input_tokens_seen": 33093744, "step": 17200 }, { "epoch": 2.806754221388368, "grad_norm": 4.5397162437438965, "learning_rate": 3.0445915933591658e-05, "loss": 0.2528, "num_input_tokens_seen": 33103712, "step": 17205 }, { "epoch": 2.807569948609185, "grad_norm": 0.2622455060482025, "learning_rate": 3.0436333801182114e-05, "loss": 0.1325, "num_input_tokens_seen": 33113968, "step": 17210 }, { "epoch": 2.8083856758300025, "grad_norm": 16.27513885498047, "learning_rate": 3.0426750830421596e-05, "loss": 0.1517, "num_input_tokens_seen": 33123200, "step": 17215 }, { "epoch": 2.8092014030508197, "grad_norm": 11.764395713806152, "learning_rate": 3.0417167022787897e-05, "loss": 0.0755, "num_input_tokens_seen": 33131696, "step": 17220 }, { "epoch": 2.8100171302716372, "grad_norm": 2.3048324584960938, "learning_rate": 3.0407582379758966e-05, "loss": 0.0806, "num_input_tokens_seen": 33141872, "step": 17225 }, { "epoch": 2.8108328574924544, "grad_norm": 0.4447864890098572, "learning_rate": 3.039799690281287e-05, "loss": 0.0798, "num_input_tokens_seen": 33149808, "step": 17230 }, { "epoch": 2.811648584713272, "grad_norm": 0.3187989294528961, "learning_rate": 3.0388410593427823e-05, "loss": 0.0723, "num_input_tokens_seen": 33158304, "step": 17235 }, { "epoch": 2.812464311934089, "grad_norm": 2.1020667552948, "learning_rate": 3.0378823453082146e-05, "loss": 0.0243, "num_input_tokens_seen": 33168144, "step": 17240 }, { "epoch": 2.8132800391549067, "grad_norm": 0.12439366430044174, "learning_rate": 3.03692354832543e-05, "loss": 0.0038, "num_input_tokens_seen": 33179152, "step": 17245 }, { "epoch": 2.814095766375724, "grad_norm": 3.7402799129486084, "learning_rate": 3.0359646685422865e-05, "loss": 0.0942, "num_input_tokens_seen": 33190320, "step": 17250 }, { "epoch": 2.8149114935965414, "grad_norm": 0.09342723339796066, "learning_rate": 3.035005706106656e-05, "loss": 0.0629, "num_input_tokens_seen": 33199936, "step": 17255 }, { "epoch": 2.815727220817359, "grad_norm": 0.11185844987630844, "learning_rate": 3.034046661166422e-05, "loss": 0.1619, "num_input_tokens_seen": 33210000, "step": 17260 }, { "epoch": 2.816542948038176, "grad_norm": 3.994415521621704, "learning_rate": 3.033087533869482e-05, "loss": 0.0117, "num_input_tokens_seen": 33220112, "step": 17265 }, { "epoch": 2.817358675258993, "grad_norm": 0.20762285590171814, "learning_rate": 3.0321283243637444e-05, "loss": 0.073, "num_input_tokens_seen": 33230464, "step": 17270 }, { "epoch": 2.8181744024798108, "grad_norm": 0.07369744777679443, "learning_rate": 3.0311690327971326e-05, "loss": 0.1095, "num_input_tokens_seen": 33240112, "step": 17275 }, { "epoch": 2.8189901297006283, "grad_norm": 17.023223876953125, "learning_rate": 3.030209659317581e-05, "loss": 0.2338, "num_input_tokens_seen": 33250224, "step": 17280 }, { "epoch": 2.8198058569214455, "grad_norm": 5.378016948699951, "learning_rate": 3.0292502040730362e-05, "loss": 0.1724, "num_input_tokens_seen": 33260448, "step": 17285 }, { "epoch": 2.8206215841422626, "grad_norm": 0.6308000087738037, "learning_rate": 3.0282906672114597e-05, "loss": 0.0197, "num_input_tokens_seen": 33270784, "step": 17290 }, { "epoch": 2.82143731136308, "grad_norm": 0.03317071124911308, "learning_rate": 3.027331048880823e-05, "loss": 0.0852, "num_input_tokens_seen": 33281168, "step": 17295 }, { "epoch": 2.8222530385838978, "grad_norm": 0.10334622859954834, "learning_rate": 3.0263713492291123e-05, "loss": 0.0672, "num_input_tokens_seen": 33290816, "step": 17300 }, { "epoch": 2.823068765804715, "grad_norm": 11.397963523864746, "learning_rate": 3.0254115684043242e-05, "loss": 0.0688, "num_input_tokens_seen": 33301488, "step": 17305 }, { "epoch": 2.823884493025532, "grad_norm": 0.8567450046539307, "learning_rate": 3.024451706554469e-05, "loss": 0.1213, "num_input_tokens_seen": 33310592, "step": 17310 }, { "epoch": 2.8247002202463496, "grad_norm": 4.4792633056640625, "learning_rate": 3.0234917638275705e-05, "loss": 0.1557, "num_input_tokens_seen": 33319808, "step": 17315 }, { "epoch": 2.825515947467167, "grad_norm": 0.14555229246616364, "learning_rate": 3.0225317403716635e-05, "loss": 0.0393, "num_input_tokens_seen": 33330192, "step": 17320 }, { "epoch": 2.8263316746879843, "grad_norm": 3.1202945709228516, "learning_rate": 3.0215716363347956e-05, "loss": 0.1116, "num_input_tokens_seen": 33339952, "step": 17325 }, { "epoch": 2.8271474019088014, "grad_norm": 0.21646322309970856, "learning_rate": 3.0206114518650275e-05, "loss": 0.0458, "num_input_tokens_seen": 33350752, "step": 17330 }, { "epoch": 2.827963129129619, "grad_norm": 0.37651997804641724, "learning_rate": 3.0196511871104304e-05, "loss": 0.1536, "num_input_tokens_seen": 33361392, "step": 17335 }, { "epoch": 2.8287788563504366, "grad_norm": 0.018732348456978798, "learning_rate": 3.01869084221909e-05, "loss": 0.0067, "num_input_tokens_seen": 33370400, "step": 17340 }, { "epoch": 2.8295945835712537, "grad_norm": 0.1647125482559204, "learning_rate": 3.0177304173391037e-05, "loss": 0.1822, "num_input_tokens_seen": 33379760, "step": 17345 }, { "epoch": 2.8304103107920713, "grad_norm": 8.630106925964355, "learning_rate": 3.01676991261858e-05, "loss": 0.2161, "num_input_tokens_seen": 33389856, "step": 17350 }, { "epoch": 2.8312260380128884, "grad_norm": 0.3935125172138214, "learning_rate": 3.015809328205642e-05, "loss": 0.1325, "num_input_tokens_seen": 33398416, "step": 17355 }, { "epoch": 2.832041765233706, "grad_norm": 2.1077775955200195, "learning_rate": 3.0148486642484248e-05, "loss": 0.0529, "num_input_tokens_seen": 33407632, "step": 17360 }, { "epoch": 2.832857492454523, "grad_norm": 0.18089605867862701, "learning_rate": 3.0138879208950722e-05, "loss": 0.0179, "num_input_tokens_seen": 33416672, "step": 17365 }, { "epoch": 2.8336732196753407, "grad_norm": 0.8140025734901428, "learning_rate": 3.012927098293744e-05, "loss": 0.008, "num_input_tokens_seen": 33426608, "step": 17370 }, { "epoch": 2.834488946896158, "grad_norm": 0.10046727955341339, "learning_rate": 3.0119661965926123e-05, "loss": 0.0735, "num_input_tokens_seen": 33435072, "step": 17375 }, { "epoch": 2.8353046741169754, "grad_norm": 6.749020099639893, "learning_rate": 3.0110052159398587e-05, "loss": 0.218, "num_input_tokens_seen": 33444512, "step": 17380 }, { "epoch": 2.8361204013377925, "grad_norm": 0.31404900550842285, "learning_rate": 3.0100441564836802e-05, "loss": 0.145, "num_input_tokens_seen": 33454256, "step": 17385 }, { "epoch": 2.83693612855861, "grad_norm": 0.2134295403957367, "learning_rate": 3.0090830183722817e-05, "loss": 0.0823, "num_input_tokens_seen": 33464208, "step": 17390 }, { "epoch": 2.8377518557794272, "grad_norm": 6.772312164306641, "learning_rate": 3.0081218017538852e-05, "loss": 0.0437, "num_input_tokens_seen": 33474544, "step": 17395 }, { "epoch": 2.838567583000245, "grad_norm": 6.568906307220459, "learning_rate": 3.0071605067767212e-05, "loss": 0.2777, "num_input_tokens_seen": 33484336, "step": 17400 }, { "epoch": 2.838567583000245, "eval_loss": 0.15053951740264893, "eval_runtime": 80.8507, "eval_samples_per_second": 33.704, "eval_steps_per_second": 16.858, "num_input_tokens_seen": 33484336, "step": 17400 }, { "epoch": 2.839383310221062, "grad_norm": 21.26409149169922, "learning_rate": 3.006199133589034e-05, "loss": 0.2882, "num_input_tokens_seen": 33494864, "step": 17405 }, { "epoch": 2.8401990374418795, "grad_norm": 7.1878437995910645, "learning_rate": 3.005237682339079e-05, "loss": 0.0399, "num_input_tokens_seen": 33504224, "step": 17410 }, { "epoch": 2.8410147646626966, "grad_norm": 11.930059432983398, "learning_rate": 3.0042761531751228e-05, "loss": 0.0498, "num_input_tokens_seen": 33513568, "step": 17415 }, { "epoch": 2.841830491883514, "grad_norm": 1.9472090005874634, "learning_rate": 3.0033145462454482e-05, "loss": 0.1135, "num_input_tokens_seen": 33522272, "step": 17420 }, { "epoch": 2.8426462191043314, "grad_norm": 4.920133113861084, "learning_rate": 3.002352861698345e-05, "loss": 0.2042, "num_input_tokens_seen": 33532016, "step": 17425 }, { "epoch": 2.843461946325149, "grad_norm": 0.18438994884490967, "learning_rate": 3.0013910996821178e-05, "loss": 0.016, "num_input_tokens_seen": 33541984, "step": 17430 }, { "epoch": 2.8442776735459665, "grad_norm": 4.840715408325195, "learning_rate": 3.0004292603450817e-05, "loss": 0.0251, "num_input_tokens_seen": 33552320, "step": 17435 }, { "epoch": 2.8450934007667836, "grad_norm": 1.203605055809021, "learning_rate": 2.9994673438355653e-05, "loss": 0.0224, "num_input_tokens_seen": 33562336, "step": 17440 }, { "epoch": 2.8459091279876008, "grad_norm": 0.155713751912117, "learning_rate": 2.9985053503019078e-05, "loss": 0.0107, "num_input_tokens_seen": 33571952, "step": 17445 }, { "epoch": 2.8467248552084183, "grad_norm": 5.738866329193115, "learning_rate": 2.99754327989246e-05, "loss": 0.1783, "num_input_tokens_seen": 33581872, "step": 17450 }, { "epoch": 2.847540582429236, "grad_norm": 14.164285659790039, "learning_rate": 2.9965811327555864e-05, "loss": 0.0674, "num_input_tokens_seen": 33591840, "step": 17455 }, { "epoch": 2.848356309650053, "grad_norm": 1.406355857849121, "learning_rate": 2.995618909039662e-05, "loss": 0.0081, "num_input_tokens_seen": 33601824, "step": 17460 }, { "epoch": 2.84917203687087, "grad_norm": 4.104393482208252, "learning_rate": 2.9946566088930727e-05, "loss": 0.0242, "num_input_tokens_seen": 33611584, "step": 17465 }, { "epoch": 2.8499877640916877, "grad_norm": 5.611894607543945, "learning_rate": 2.9936942324642192e-05, "loss": 0.1258, "num_input_tokens_seen": 33620448, "step": 17470 }, { "epoch": 2.8508034913125053, "grad_norm": 0.29672273993492126, "learning_rate": 2.9927317799015097e-05, "loss": 0.0034, "num_input_tokens_seen": 33629776, "step": 17475 }, { "epoch": 2.8516192185333225, "grad_norm": 3.2155652046203613, "learning_rate": 2.9917692513533685e-05, "loss": 0.1525, "num_input_tokens_seen": 33639344, "step": 17480 }, { "epoch": 2.8524349457541396, "grad_norm": 0.15495926141738892, "learning_rate": 2.990806646968229e-05, "loss": 0.0046, "num_input_tokens_seen": 33649632, "step": 17485 }, { "epoch": 2.853250672974957, "grad_norm": 0.14054393768310547, "learning_rate": 2.989843966894536e-05, "loss": 0.2605, "num_input_tokens_seen": 33659280, "step": 17490 }, { "epoch": 2.8540664001957747, "grad_norm": 2.776528835296631, "learning_rate": 2.9888812112807472e-05, "loss": 0.0615, "num_input_tokens_seen": 33668192, "step": 17495 }, { "epoch": 2.854882127416592, "grad_norm": 0.7655100226402283, "learning_rate": 2.987918380275333e-05, "loss": 0.023, "num_input_tokens_seen": 33678032, "step": 17500 }, { "epoch": 2.855697854637409, "grad_norm": 0.11163699626922607, "learning_rate": 2.9869554740267724e-05, "loss": 0.0185, "num_input_tokens_seen": 33687776, "step": 17505 }, { "epoch": 2.8565135818582266, "grad_norm": 21.375566482543945, "learning_rate": 2.9859924926835585e-05, "loss": 0.1449, "num_input_tokens_seen": 33696704, "step": 17510 }, { "epoch": 2.857329309079044, "grad_norm": 1.979358434677124, "learning_rate": 2.9850294363941944e-05, "loss": 0.1721, "num_input_tokens_seen": 33705600, "step": 17515 }, { "epoch": 2.8581450362998613, "grad_norm": 7.7528557777404785, "learning_rate": 2.9840663053071967e-05, "loss": 0.1442, "num_input_tokens_seen": 33716528, "step": 17520 }, { "epoch": 2.858960763520679, "grad_norm": 0.036281365901231766, "learning_rate": 2.983103099571091e-05, "loss": 0.0121, "num_input_tokens_seen": 33726928, "step": 17525 }, { "epoch": 2.859776490741496, "grad_norm": 0.07244466245174408, "learning_rate": 2.9821398193344164e-05, "loss": 0.0132, "num_input_tokens_seen": 33737280, "step": 17530 }, { "epoch": 2.8605922179623136, "grad_norm": 0.6939958333969116, "learning_rate": 2.9811764647457226e-05, "loss": 0.0889, "num_input_tokens_seen": 33746688, "step": 17535 }, { "epoch": 2.8614079451831307, "grad_norm": 0.516800045967102, "learning_rate": 2.9802130359535714e-05, "loss": 0.1121, "num_input_tokens_seen": 33757568, "step": 17540 }, { "epoch": 2.8622236724039483, "grad_norm": 4.0538859367370605, "learning_rate": 2.979249533106535e-05, "loss": 0.051, "num_input_tokens_seen": 33766480, "step": 17545 }, { "epoch": 2.8630393996247654, "grad_norm": 0.39921948313713074, "learning_rate": 2.9782859563531986e-05, "loss": 0.0764, "num_input_tokens_seen": 33776240, "step": 17550 }, { "epoch": 2.863855126845583, "grad_norm": 2.388857126235962, "learning_rate": 2.977322305842156e-05, "loss": 0.1191, "num_input_tokens_seen": 33786016, "step": 17555 }, { "epoch": 2.8646708540664, "grad_norm": 0.06255277991294861, "learning_rate": 2.9763585817220162e-05, "loss": 0.014, "num_input_tokens_seen": 33795088, "step": 17560 }, { "epoch": 2.8654865812872177, "grad_norm": 0.20000986754894257, "learning_rate": 2.975394784141397e-05, "loss": 0.1267, "num_input_tokens_seen": 33804704, "step": 17565 }, { "epoch": 2.866302308508035, "grad_norm": 0.24220646917819977, "learning_rate": 2.974430913248928e-05, "loss": 0.0489, "num_input_tokens_seen": 33815136, "step": 17570 }, { "epoch": 2.8671180357288524, "grad_norm": 0.09929189085960388, "learning_rate": 2.9734669691932497e-05, "loss": 0.0716, "num_input_tokens_seen": 33825088, "step": 17575 }, { "epoch": 2.8679337629496695, "grad_norm": 0.3174518048763275, "learning_rate": 2.9725029521230147e-05, "loss": 0.008, "num_input_tokens_seen": 33835120, "step": 17580 }, { "epoch": 2.868749490170487, "grad_norm": 0.12606942653656006, "learning_rate": 2.9715388621868873e-05, "loss": 0.1333, "num_input_tokens_seen": 33844704, "step": 17585 }, { "epoch": 2.869565217391304, "grad_norm": 0.045185960829257965, "learning_rate": 2.970574699533541e-05, "loss": 0.0387, "num_input_tokens_seen": 33854896, "step": 17590 }, { "epoch": 2.870380944612122, "grad_norm": 0.11977942287921906, "learning_rate": 2.969610464311662e-05, "loss": 0.1104, "num_input_tokens_seen": 33864400, "step": 17595 }, { "epoch": 2.871196671832939, "grad_norm": 0.12874145805835724, "learning_rate": 2.9686461566699487e-05, "loss": 0.0597, "num_input_tokens_seen": 33875072, "step": 17600 }, { "epoch": 2.871196671832939, "eval_loss": 0.172089084982872, "eval_runtime": 80.9168, "eval_samples_per_second": 33.677, "eval_steps_per_second": 16.844, "num_input_tokens_seen": 33875072, "step": 17600 }, { "epoch": 2.8720123990537565, "grad_norm": 2.543053150177002, "learning_rate": 2.9676817767571086e-05, "loss": 0.2606, "num_input_tokens_seen": 33885840, "step": 17605 }, { "epoch": 2.8728281262745736, "grad_norm": 0.05838209390640259, "learning_rate": 2.966717324721861e-05, "loss": 0.1321, "num_input_tokens_seen": 33895728, "step": 17610 }, { "epoch": 2.873643853495391, "grad_norm": 1.4164999723434448, "learning_rate": 2.9657528007129366e-05, "loss": 0.0623, "num_input_tokens_seen": 33905344, "step": 17615 }, { "epoch": 2.8744595807162083, "grad_norm": 0.40089836716651917, "learning_rate": 2.9647882048790777e-05, "loss": 0.0884, "num_input_tokens_seen": 33916208, "step": 17620 }, { "epoch": 2.875275307937026, "grad_norm": 6.868166923522949, "learning_rate": 2.963823537369037e-05, "loss": 0.0828, "num_input_tokens_seen": 33926096, "step": 17625 }, { "epoch": 2.8760910351578435, "grad_norm": 3.242985486984253, "learning_rate": 2.9628587983315775e-05, "loss": 0.0628, "num_input_tokens_seen": 33935088, "step": 17630 }, { "epoch": 2.8769067623786606, "grad_norm": 4.8106689453125, "learning_rate": 2.9618939879154746e-05, "loss": 0.1181, "num_input_tokens_seen": 33945120, "step": 17635 }, { "epoch": 2.8777224895994777, "grad_norm": 0.009233747608959675, "learning_rate": 2.9609291062695143e-05, "loss": 0.173, "num_input_tokens_seen": 33953952, "step": 17640 }, { "epoch": 2.8785382168202953, "grad_norm": 0.10044034570455551, "learning_rate": 2.9599641535424938e-05, "loss": 0.0827, "num_input_tokens_seen": 33963072, "step": 17645 }, { "epoch": 2.879353944041113, "grad_norm": 7.626039981842041, "learning_rate": 2.9589991298832202e-05, "loss": 0.1571, "num_input_tokens_seen": 33971104, "step": 17650 }, { "epoch": 2.88016967126193, "grad_norm": 3.005821704864502, "learning_rate": 2.958034035440513e-05, "loss": 0.0766, "num_input_tokens_seen": 33980672, "step": 17655 }, { "epoch": 2.880985398482747, "grad_norm": 6.763315200805664, "learning_rate": 2.957068870363201e-05, "loss": 0.0254, "num_input_tokens_seen": 33989696, "step": 17660 }, { "epoch": 2.8818011257035647, "grad_norm": 10.557279586791992, "learning_rate": 2.956103634800126e-05, "loss": 0.12, "num_input_tokens_seen": 33999120, "step": 17665 }, { "epoch": 2.8826168529243823, "grad_norm": 0.2298445701599121, "learning_rate": 2.9551383289001384e-05, "loss": 0.021, "num_input_tokens_seen": 34009328, "step": 17670 }, { "epoch": 2.8834325801451994, "grad_norm": 5.666200637817383, "learning_rate": 2.9541729528121005e-05, "loss": 0.1389, "num_input_tokens_seen": 34019056, "step": 17675 }, { "epoch": 2.8842483073660166, "grad_norm": 1.3409146070480347, "learning_rate": 2.9532075066848856e-05, "loss": 0.0276, "num_input_tokens_seen": 34029472, "step": 17680 }, { "epoch": 2.885064034586834, "grad_norm": 0.13344275951385498, "learning_rate": 2.9522419906673786e-05, "loss": 0.0117, "num_input_tokens_seen": 34038032, "step": 17685 }, { "epoch": 2.8858797618076517, "grad_norm": 4.424880504608154, "learning_rate": 2.951276404908474e-05, "loss": 0.0319, "num_input_tokens_seen": 34047968, "step": 17690 }, { "epoch": 2.886695489028469, "grad_norm": 0.02343510091304779, "learning_rate": 2.9503107495570752e-05, "loss": 0.0714, "num_input_tokens_seen": 34058624, "step": 17695 }, { "epoch": 2.887511216249286, "grad_norm": 3.777130365371704, "learning_rate": 2.9493450247621003e-05, "loss": 0.0798, "num_input_tokens_seen": 34067248, "step": 17700 }, { "epoch": 2.8883269434701035, "grad_norm": 7.2159528732299805, "learning_rate": 2.948379230672476e-05, "loss": 0.167, "num_input_tokens_seen": 34077568, "step": 17705 }, { "epoch": 2.889142670690921, "grad_norm": 3.433037519454956, "learning_rate": 2.9474133674371396e-05, "loss": 0.1157, "num_input_tokens_seen": 34087296, "step": 17710 }, { "epoch": 2.8899583979117383, "grad_norm": 0.04161461815237999, "learning_rate": 2.9464474352050387e-05, "loss": 0.0702, "num_input_tokens_seen": 34096576, "step": 17715 }, { "epoch": 2.890774125132556, "grad_norm": 0.07237016409635544, "learning_rate": 2.9454814341251336e-05, "loss": 0.0077, "num_input_tokens_seen": 34106160, "step": 17720 }, { "epoch": 2.891589852353373, "grad_norm": 0.0507231242954731, "learning_rate": 2.9445153643463942e-05, "loss": 0.0064, "num_input_tokens_seen": 34116608, "step": 17725 }, { "epoch": 2.8924055795741905, "grad_norm": 0.07263002544641495, "learning_rate": 2.943549226017798e-05, "loss": 0.0661, "num_input_tokens_seen": 34127808, "step": 17730 }, { "epoch": 2.8932213067950077, "grad_norm": 0.037324246019124985, "learning_rate": 2.942583019288337e-05, "loss": 0.0042, "num_input_tokens_seen": 34137280, "step": 17735 }, { "epoch": 2.8940370340158252, "grad_norm": 2.8181746006011963, "learning_rate": 2.9416167443070132e-05, "loss": 0.2629, "num_input_tokens_seen": 34146848, "step": 17740 }, { "epoch": 2.8948527612366424, "grad_norm": 0.13407394289970398, "learning_rate": 2.9406504012228375e-05, "loss": 0.1481, "num_input_tokens_seen": 34155920, "step": 17745 }, { "epoch": 2.89566848845746, "grad_norm": 0.5046784281730652, "learning_rate": 2.939683990184832e-05, "loss": 0.0285, "num_input_tokens_seen": 34166160, "step": 17750 }, { "epoch": 2.896484215678277, "grad_norm": 0.290334016084671, "learning_rate": 2.93871751134203e-05, "loss": 0.0324, "num_input_tokens_seen": 34175424, "step": 17755 }, { "epoch": 2.8972999428990946, "grad_norm": 0.07454335689544678, "learning_rate": 2.9377509648434752e-05, "loss": 0.0116, "num_input_tokens_seen": 34185904, "step": 17760 }, { "epoch": 2.898115670119912, "grad_norm": 3.042870283126831, "learning_rate": 2.9367843508382203e-05, "loss": 0.1272, "num_input_tokens_seen": 34194528, "step": 17765 }, { "epoch": 2.8989313973407294, "grad_norm": 0.04248226433992386, "learning_rate": 2.9358176694753293e-05, "loss": 0.0063, "num_input_tokens_seen": 34204048, "step": 17770 }, { "epoch": 2.8997471245615465, "grad_norm": 1.3342692852020264, "learning_rate": 2.9348509209038766e-05, "loss": 0.0759, "num_input_tokens_seen": 34214432, "step": 17775 }, { "epoch": 2.900562851782364, "grad_norm": 0.07066114246845245, "learning_rate": 2.933884105272947e-05, "loss": 0.0699, "num_input_tokens_seen": 34223936, "step": 17780 }, { "epoch": 2.901378579003181, "grad_norm": 11.264946937561035, "learning_rate": 2.9329172227316366e-05, "loss": 0.1767, "num_input_tokens_seen": 34235152, "step": 17785 }, { "epoch": 2.9021943062239988, "grad_norm": 0.39995846152305603, "learning_rate": 2.93195027342905e-05, "loss": 0.1498, "num_input_tokens_seen": 34244880, "step": 17790 }, { "epoch": 2.903010033444816, "grad_norm": 8.608067512512207, "learning_rate": 2.9309832575143024e-05, "loss": 0.064, "num_input_tokens_seen": 34255312, "step": 17795 }, { "epoch": 2.9038257606656335, "grad_norm": 0.18033288419246674, "learning_rate": 2.930016175136521e-05, "loss": 0.1232, "num_input_tokens_seen": 34264832, "step": 17800 }, { "epoch": 2.9038257606656335, "eval_loss": 0.15844771265983582, "eval_runtime": 80.7686, "eval_samples_per_second": 33.738, "eval_steps_per_second": 16.875, "num_input_tokens_seen": 34264832, "step": 17800 }, { "epoch": 2.904641487886451, "grad_norm": 5.538437366485596, "learning_rate": 2.9290490264448412e-05, "loss": 0.082, "num_input_tokens_seen": 34273728, "step": 17805 }, { "epoch": 2.905457215107268, "grad_norm": 15.639097213745117, "learning_rate": 2.9280818115884094e-05, "loss": 0.1346, "num_input_tokens_seen": 34283168, "step": 17810 }, { "epoch": 2.9062729423280853, "grad_norm": 1.1149054765701294, "learning_rate": 2.9271145307163828e-05, "loss": 0.1049, "num_input_tokens_seen": 34293200, "step": 17815 }, { "epoch": 2.907088669548903, "grad_norm": 0.22711676359176636, "learning_rate": 2.9261471839779287e-05, "loss": 0.0074, "num_input_tokens_seen": 34302832, "step": 17820 }, { "epoch": 2.9079043967697205, "grad_norm": 0.1364402025938034, "learning_rate": 2.925179771522223e-05, "loss": 0.0172, "num_input_tokens_seen": 34312304, "step": 17825 }, { "epoch": 2.9087201239905376, "grad_norm": 2.0705161094665527, "learning_rate": 2.9242122934984535e-05, "loss": 0.0527, "num_input_tokens_seen": 34322144, "step": 17830 }, { "epoch": 2.9095358512113547, "grad_norm": 3.313441514968872, "learning_rate": 2.9232447500558176e-05, "loss": 0.1672, "num_input_tokens_seen": 34333024, "step": 17835 }, { "epoch": 2.9103515784321723, "grad_norm": 0.6785831451416016, "learning_rate": 2.9222771413435225e-05, "loss": 0.1224, "num_input_tokens_seen": 34342208, "step": 17840 }, { "epoch": 2.91116730565299, "grad_norm": 0.5265159606933594, "learning_rate": 2.9213094675107848e-05, "loss": 0.0824, "num_input_tokens_seen": 34350992, "step": 17845 }, { "epoch": 2.911983032873807, "grad_norm": 3.420048952102661, "learning_rate": 2.9203417287068335e-05, "loss": 0.0562, "num_input_tokens_seen": 34360096, "step": 17850 }, { "epoch": 2.912798760094624, "grad_norm": 9.52295207977295, "learning_rate": 2.9193739250809042e-05, "loss": 0.1163, "num_input_tokens_seen": 34369312, "step": 17855 }, { "epoch": 2.9136144873154417, "grad_norm": 0.29704800248146057, "learning_rate": 2.9184060567822463e-05, "loss": 0.1147, "num_input_tokens_seen": 34380880, "step": 17860 }, { "epoch": 2.9144302145362593, "grad_norm": 0.07634585350751877, "learning_rate": 2.9174381239601166e-05, "loss": 0.156, "num_input_tokens_seen": 34389408, "step": 17865 }, { "epoch": 2.9152459417570764, "grad_norm": 0.4205307960510254, "learning_rate": 2.916470126763783e-05, "loss": 0.091, "num_input_tokens_seen": 34398192, "step": 17870 }, { "epoch": 2.9160616689778935, "grad_norm": 0.07141991704702377, "learning_rate": 2.9155020653425203e-05, "loss": 0.1442, "num_input_tokens_seen": 34406608, "step": 17875 }, { "epoch": 2.916877396198711, "grad_norm": 0.10451458394527435, "learning_rate": 2.9145339398456184e-05, "loss": 0.0296, "num_input_tokens_seen": 34415808, "step": 17880 }, { "epoch": 2.9176931234195287, "grad_norm": 5.906317710876465, "learning_rate": 2.913565750422374e-05, "loss": 0.2068, "num_input_tokens_seen": 34425088, "step": 17885 }, { "epoch": 2.918508850640346, "grad_norm": 0.6820822358131409, "learning_rate": 2.9125974972220938e-05, "loss": 0.1029, "num_input_tokens_seen": 34434656, "step": 17890 }, { "epoch": 2.919324577861163, "grad_norm": 0.2048235982656479, "learning_rate": 2.9116291803940932e-05, "loss": 0.0104, "num_input_tokens_seen": 34443808, "step": 17895 }, { "epoch": 2.9201403050819805, "grad_norm": 0.1096406951546669, "learning_rate": 2.910660800087701e-05, "loss": 0.0124, "num_input_tokens_seen": 34453296, "step": 17900 }, { "epoch": 2.920956032302798, "grad_norm": 2.6003503799438477, "learning_rate": 2.909692356452254e-05, "loss": 0.1028, "num_input_tokens_seen": 34463312, "step": 17905 }, { "epoch": 2.9217717595236152, "grad_norm": 0.2839101552963257, "learning_rate": 2.9087238496370962e-05, "loss": 0.0284, "num_input_tokens_seen": 34473600, "step": 17910 }, { "epoch": 2.922587486744433, "grad_norm": 0.07950001955032349, "learning_rate": 2.907755279791583e-05, "loss": 0.0678, "num_input_tokens_seen": 34482960, "step": 17915 }, { "epoch": 2.92340321396525, "grad_norm": 0.11613748967647552, "learning_rate": 2.906786647065083e-05, "loss": 0.1124, "num_input_tokens_seen": 34493680, "step": 17920 }, { "epoch": 2.9242189411860675, "grad_norm": 0.43287670612335205, "learning_rate": 2.9058179516069695e-05, "loss": 0.0543, "num_input_tokens_seen": 34502592, "step": 17925 }, { "epoch": 2.9250346684068846, "grad_norm": 12.356428146362305, "learning_rate": 2.9048491935666282e-05, "loss": 0.0521, "num_input_tokens_seen": 34512176, "step": 17930 }, { "epoch": 2.925850395627702, "grad_norm": 0.09848584979772568, "learning_rate": 2.9038803730934534e-05, "loss": 0.191, "num_input_tokens_seen": 34522304, "step": 17935 }, { "epoch": 2.9266661228485193, "grad_norm": 0.11608539521694183, "learning_rate": 2.9029114903368503e-05, "loss": 0.0449, "num_input_tokens_seen": 34532560, "step": 17940 }, { "epoch": 2.927481850069337, "grad_norm": 0.04954452067613602, "learning_rate": 2.9019425454462318e-05, "loss": 0.1088, "num_input_tokens_seen": 34542880, "step": 17945 }, { "epoch": 2.928297577290154, "grad_norm": 7.374344348907471, "learning_rate": 2.9009735385710212e-05, "loss": 0.048, "num_input_tokens_seen": 34551792, "step": 17950 }, { "epoch": 2.9291133045109716, "grad_norm": 6.967530250549316, "learning_rate": 2.900004469860652e-05, "loss": 0.1087, "num_input_tokens_seen": 34561872, "step": 17955 }, { "epoch": 2.9299290317317888, "grad_norm": 0.03577359393239021, "learning_rate": 2.8990353394645668e-05, "loss": 0.0097, "num_input_tokens_seen": 34572032, "step": 17960 }, { "epoch": 2.9307447589526063, "grad_norm": 0.09855331480503082, "learning_rate": 2.8980661475322186e-05, "loss": 0.0104, "num_input_tokens_seen": 34582640, "step": 17965 }, { "epoch": 2.9315604861734235, "grad_norm": 2.342485189437866, "learning_rate": 2.897096894213067e-05, "loss": 0.0648, "num_input_tokens_seen": 34592400, "step": 17970 }, { "epoch": 2.932376213394241, "grad_norm": 0.17267844080924988, "learning_rate": 2.8961275796565845e-05, "loss": 0.1195, "num_input_tokens_seen": 34601664, "step": 17975 }, { "epoch": 2.933191940615058, "grad_norm": 0.12724865972995758, "learning_rate": 2.8951582040122517e-05, "loss": 0.0712, "num_input_tokens_seen": 34612064, "step": 17980 }, { "epoch": 2.9340076678358757, "grad_norm": 0.25750815868377686, "learning_rate": 2.894188767429557e-05, "loss": 0.0051, "num_input_tokens_seen": 34621056, "step": 17985 }, { "epoch": 2.934823395056693, "grad_norm": 1.539098858833313, "learning_rate": 2.8932192700580014e-05, "loss": 0.1112, "num_input_tokens_seen": 34631168, "step": 17990 }, { "epoch": 2.9356391222775104, "grad_norm": 0.050367020070552826, "learning_rate": 2.8922497120470916e-05, "loss": 0.1187, "num_input_tokens_seen": 34641488, "step": 17995 }, { "epoch": 2.936454849498328, "grad_norm": 9.326510429382324, "learning_rate": 2.891280093546348e-05, "loss": 0.2073, "num_input_tokens_seen": 34652800, "step": 18000 }, { "epoch": 2.936454849498328, "eval_loss": 0.18465572595596313, "eval_runtime": 80.8094, "eval_samples_per_second": 33.721, "eval_steps_per_second": 16.867, "num_input_tokens_seen": 34652800, "step": 18000 }, { "epoch": 2.937270576719145, "grad_norm": 0.27437692880630493, "learning_rate": 2.890310414705297e-05, "loss": 0.1171, "num_input_tokens_seen": 34663040, "step": 18005 }, { "epoch": 2.9380863039399623, "grad_norm": 4.234347820281982, "learning_rate": 2.8893406756734742e-05, "loss": 0.0073, "num_input_tokens_seen": 34672704, "step": 18010 }, { "epoch": 2.93890203116078, "grad_norm": 0.1311168670654297, "learning_rate": 2.888370876600427e-05, "loss": 0.0079, "num_input_tokens_seen": 34682384, "step": 18015 }, { "epoch": 2.9397177583815974, "grad_norm": 0.043100859969854355, "learning_rate": 2.8874010176357104e-05, "loss": 0.0386, "num_input_tokens_seen": 34692208, "step": 18020 }, { "epoch": 2.9405334856024146, "grad_norm": 1.8206192255020142, "learning_rate": 2.886431098928888e-05, "loss": 0.0876, "num_input_tokens_seen": 34702368, "step": 18025 }, { "epoch": 2.9413492128232317, "grad_norm": 0.23541559278964996, "learning_rate": 2.885461120629534e-05, "loss": 0.0073, "num_input_tokens_seen": 34711744, "step": 18030 }, { "epoch": 2.9421649400440493, "grad_norm": 0.10875163972377777, "learning_rate": 2.8844910828872317e-05, "loss": 0.0919, "num_input_tokens_seen": 34719840, "step": 18035 }, { "epoch": 2.942980667264867, "grad_norm": 3.700996160507202, "learning_rate": 2.8835209858515715e-05, "loss": 0.2704, "num_input_tokens_seen": 34729680, "step": 18040 }, { "epoch": 2.943796394485684, "grad_norm": 0.02406853809952736, "learning_rate": 2.8825508296721566e-05, "loss": 0.0238, "num_input_tokens_seen": 34740384, "step": 18045 }, { "epoch": 2.944612121706501, "grad_norm": 0.04537033289670944, "learning_rate": 2.881580614498596e-05, "loss": 0.104, "num_input_tokens_seen": 34750256, "step": 18050 }, { "epoch": 2.9454278489273187, "grad_norm": 0.08182710409164429, "learning_rate": 2.8806103404805103e-05, "loss": 0.1082, "num_input_tokens_seen": 34760688, "step": 18055 }, { "epoch": 2.9462435761481363, "grad_norm": 0.1647128313779831, "learning_rate": 2.8796400077675257e-05, "loss": 0.0127, "num_input_tokens_seen": 34771152, "step": 18060 }, { "epoch": 2.9470593033689534, "grad_norm": 0.03909978270530701, "learning_rate": 2.8786696165092812e-05, "loss": 0.068, "num_input_tokens_seen": 34781488, "step": 18065 }, { "epoch": 2.9478750305897705, "grad_norm": 0.040784742683172226, "learning_rate": 2.8776991668554236e-05, "loss": 0.1132, "num_input_tokens_seen": 34790976, "step": 18070 }, { "epoch": 2.948690757810588, "grad_norm": 0.7137571573257446, "learning_rate": 2.876728658955608e-05, "loss": 0.0408, "num_input_tokens_seen": 34801056, "step": 18075 }, { "epoch": 2.9495064850314057, "grad_norm": 0.044492531567811966, "learning_rate": 2.8757580929594986e-05, "loss": 0.0023, "num_input_tokens_seen": 34810800, "step": 18080 }, { "epoch": 2.950322212252223, "grad_norm": 0.3894851505756378, "learning_rate": 2.87478746901677e-05, "loss": 0.0576, "num_input_tokens_seen": 34818880, "step": 18085 }, { "epoch": 2.9511379394730404, "grad_norm": 13.775740623474121, "learning_rate": 2.873816787277103e-05, "loss": 0.0321, "num_input_tokens_seen": 34828448, "step": 18090 }, { "epoch": 2.9519536666938575, "grad_norm": 0.07070664316415787, "learning_rate": 2.8728460478901903e-05, "loss": 0.2981, "num_input_tokens_seen": 34835952, "step": 18095 }, { "epoch": 2.952769393914675, "grad_norm": 0.02091820538043976, "learning_rate": 2.8718752510057307e-05, "loss": 0.0069, "num_input_tokens_seen": 34845408, "step": 18100 }, { "epoch": 2.953585121135492, "grad_norm": 0.17264322936534882, "learning_rate": 2.870904396773435e-05, "loss": 0.059, "num_input_tokens_seen": 34854768, "step": 18105 }, { "epoch": 2.95440084835631, "grad_norm": 0.09772830456495285, "learning_rate": 2.86993348534302e-05, "loss": 0.0166, "num_input_tokens_seen": 34864880, "step": 18110 }, { "epoch": 2.955216575577127, "grad_norm": 4.405411720275879, "learning_rate": 2.868962516864212e-05, "loss": 0.1803, "num_input_tokens_seen": 34874096, "step": 18115 }, { "epoch": 2.9560323027979445, "grad_norm": 0.17173804342746735, "learning_rate": 2.8679914914867477e-05, "loss": 0.0132, "num_input_tokens_seen": 34883536, "step": 18120 }, { "epoch": 2.9568480300187616, "grad_norm": 21.496549606323242, "learning_rate": 2.8670204093603713e-05, "loss": 0.1164, "num_input_tokens_seen": 34893984, "step": 18125 }, { "epoch": 2.957663757239579, "grad_norm": 6.158992767333984, "learning_rate": 2.8660492706348357e-05, "loss": 0.1251, "num_input_tokens_seen": 34902848, "step": 18130 }, { "epoch": 2.9584794844603963, "grad_norm": 0.6828044652938843, "learning_rate": 2.8650780754599022e-05, "loss": 0.1002, "num_input_tokens_seen": 34913040, "step": 18135 }, { "epoch": 2.959295211681214, "grad_norm": 7.0721306800842285, "learning_rate": 2.8641068239853407e-05, "loss": 0.078, "num_input_tokens_seen": 34921600, "step": 18140 }, { "epoch": 2.960110938902031, "grad_norm": 5.852766036987305, "learning_rate": 2.863135516360932e-05, "loss": 0.0571, "num_input_tokens_seen": 34932160, "step": 18145 }, { "epoch": 2.9609266661228486, "grad_norm": 12.003578186035156, "learning_rate": 2.8621641527364633e-05, "loss": 0.1018, "num_input_tokens_seen": 34943040, "step": 18150 }, { "epoch": 2.9617423933436657, "grad_norm": 0.19126954674720764, "learning_rate": 2.8611927332617313e-05, "loss": 0.0651, "num_input_tokens_seen": 34952864, "step": 18155 }, { "epoch": 2.9625581205644833, "grad_norm": 0.03697548434138298, "learning_rate": 2.8602212580865405e-05, "loss": 0.1216, "num_input_tokens_seen": 34961840, "step": 18160 }, { "epoch": 2.9633738477853004, "grad_norm": 5.15321159362793, "learning_rate": 2.859249727360705e-05, "loss": 0.0865, "num_input_tokens_seen": 34971952, "step": 18165 }, { "epoch": 2.964189575006118, "grad_norm": 2.947138786315918, "learning_rate": 2.8582781412340465e-05, "loss": 0.064, "num_input_tokens_seen": 34982800, "step": 18170 }, { "epoch": 2.965005302226935, "grad_norm": 0.2238212525844574, "learning_rate": 2.857306499856397e-05, "loss": 0.1243, "num_input_tokens_seen": 34992464, "step": 18175 }, { "epoch": 2.9658210294477527, "grad_norm": 0.045656174421310425, "learning_rate": 2.856334803377594e-05, "loss": 0.0786, "num_input_tokens_seen": 35001360, "step": 18180 }, { "epoch": 2.96663675666857, "grad_norm": 7.961764335632324, "learning_rate": 2.8553630519474867e-05, "loss": 0.0462, "num_input_tokens_seen": 35010880, "step": 18185 }, { "epoch": 2.9674524838893874, "grad_norm": 0.1787831038236618, "learning_rate": 2.8543912457159317e-05, "loss": 0.1297, "num_input_tokens_seen": 35019360, "step": 18190 }, { "epoch": 2.968268211110205, "grad_norm": 0.22666268050670624, "learning_rate": 2.853419384832792e-05, "loss": 0.0597, "num_input_tokens_seen": 35027104, "step": 18195 }, { "epoch": 2.969083938331022, "grad_norm": 0.38136807084083557, "learning_rate": 2.8524474694479423e-05, "loss": 0.0482, "num_input_tokens_seen": 35036144, "step": 18200 }, { "epoch": 2.969083938331022, "eval_loss": 0.1611875742673874, "eval_runtime": 80.7646, "eval_samples_per_second": 33.74, "eval_steps_per_second": 16.876, "num_input_tokens_seen": 35036144, "step": 18200 }, { "epoch": 2.9698996655518393, "grad_norm": 5.3699116706848145, "learning_rate": 2.851475499711264e-05, "loss": 0.1057, "num_input_tokens_seen": 35044944, "step": 18205 }, { "epoch": 2.970715392772657, "grad_norm": 0.20283454656600952, "learning_rate": 2.8505034757726468e-05, "loss": 0.0715, "num_input_tokens_seen": 35053936, "step": 18210 }, { "epoch": 2.9715311199934744, "grad_norm": 0.22993400692939758, "learning_rate": 2.8495313977819886e-05, "loss": 0.1235, "num_input_tokens_seen": 35064032, "step": 18215 }, { "epoch": 2.9723468472142915, "grad_norm": 0.5255441665649414, "learning_rate": 2.8485592658891956e-05, "loss": 0.1158, "num_input_tokens_seen": 35072832, "step": 18220 }, { "epoch": 2.9731625744351087, "grad_norm": 1.562090277671814, "learning_rate": 2.8475870802441844e-05, "loss": 0.0392, "num_input_tokens_seen": 35082672, "step": 18225 }, { "epoch": 2.9739783016559262, "grad_norm": 0.31168147921562195, "learning_rate": 2.8466148409968774e-05, "loss": 0.1633, "num_input_tokens_seen": 35090848, "step": 18230 }, { "epoch": 2.974794028876744, "grad_norm": 0.15821650624275208, "learning_rate": 2.8456425482972067e-05, "loss": 0.0065, "num_input_tokens_seen": 35099712, "step": 18235 }, { "epoch": 2.975609756097561, "grad_norm": 2.9990270137786865, "learning_rate": 2.84467020229511e-05, "loss": 0.1186, "num_input_tokens_seen": 35108560, "step": 18240 }, { "epoch": 2.976425483318378, "grad_norm": 0.03341003134846687, "learning_rate": 2.8436978031405375e-05, "loss": 0.0043, "num_input_tokens_seen": 35117696, "step": 18245 }, { "epoch": 2.9772412105391957, "grad_norm": 3.0316381454467773, "learning_rate": 2.842725350983445e-05, "loss": 0.0923, "num_input_tokens_seen": 35127344, "step": 18250 }, { "epoch": 2.9780569377600132, "grad_norm": 0.12257864326238632, "learning_rate": 2.8417528459737957e-05, "loss": 0.153, "num_input_tokens_seen": 35135600, "step": 18255 }, { "epoch": 2.9788726649808304, "grad_norm": 9.977020263671875, "learning_rate": 2.8407802882615624e-05, "loss": 0.0687, "num_input_tokens_seen": 35145632, "step": 18260 }, { "epoch": 2.9796883922016475, "grad_norm": 0.05344606563448906, "learning_rate": 2.8398076779967277e-05, "loss": 0.0043, "num_input_tokens_seen": 35154512, "step": 18265 }, { "epoch": 2.980504119422465, "grad_norm": 0.21449266374111176, "learning_rate": 2.8388350153292774e-05, "loss": 0.0174, "num_input_tokens_seen": 35163952, "step": 18270 }, { "epoch": 2.9813198466432826, "grad_norm": 0.15286283195018768, "learning_rate": 2.8378623004092103e-05, "loss": 0.0054, "num_input_tokens_seen": 35173440, "step": 18275 }, { "epoch": 2.9821355738640998, "grad_norm": 4.296424388885498, "learning_rate": 2.8368895333865302e-05, "loss": 0.1202, "num_input_tokens_seen": 35182272, "step": 18280 }, { "epoch": 2.9829513010849174, "grad_norm": 4.632872581481934, "learning_rate": 2.835916714411251e-05, "loss": 0.1643, "num_input_tokens_seen": 35191200, "step": 18285 }, { "epoch": 2.9837670283057345, "grad_norm": 2.3406972885131836, "learning_rate": 2.8349438436333926e-05, "loss": 0.2133, "num_input_tokens_seen": 35201024, "step": 18290 }, { "epoch": 2.984582755526552, "grad_norm": 5.946495056152344, "learning_rate": 2.833970921202984e-05, "loss": 0.0843, "num_input_tokens_seen": 35212336, "step": 18295 }, { "epoch": 2.985398482747369, "grad_norm": 3.826974391937256, "learning_rate": 2.8329979472700628e-05, "loss": 0.0573, "num_input_tokens_seen": 35221408, "step": 18300 }, { "epoch": 2.9862142099681868, "grad_norm": 3.4813876152038574, "learning_rate": 2.832024921984674e-05, "loss": 0.115, "num_input_tokens_seen": 35230096, "step": 18305 }, { "epoch": 2.987029937189004, "grad_norm": 0.1108347475528717, "learning_rate": 2.8310518454968693e-05, "loss": 0.088, "num_input_tokens_seen": 35240832, "step": 18310 }, { "epoch": 2.9878456644098215, "grad_norm": 0.0998205840587616, "learning_rate": 2.8300787179567095e-05, "loss": 0.05, "num_input_tokens_seen": 35250544, "step": 18315 }, { "epoch": 2.9886613916306386, "grad_norm": 0.07689336687326431, "learning_rate": 2.8291055395142636e-05, "loss": 0.0679, "num_input_tokens_seen": 35258048, "step": 18320 }, { "epoch": 2.989477118851456, "grad_norm": 0.05307824909687042, "learning_rate": 2.8281323103196073e-05, "loss": 0.0161, "num_input_tokens_seen": 35269008, "step": 18325 }, { "epoch": 2.9902928460722733, "grad_norm": 0.1400662064552307, "learning_rate": 2.8271590305228256e-05, "loss": 0.0997, "num_input_tokens_seen": 35279616, "step": 18330 }, { "epoch": 2.991108573293091, "grad_norm": 0.2680072486400604, "learning_rate": 2.82618570027401e-05, "loss": 0.0665, "num_input_tokens_seen": 35288976, "step": 18335 }, { "epoch": 2.991924300513908, "grad_norm": 10.400407791137695, "learning_rate": 2.8252123197232604e-05, "loss": 0.0334, "num_input_tokens_seen": 35298976, "step": 18340 }, { "epoch": 2.9927400277347256, "grad_norm": 1.7880386114120483, "learning_rate": 2.8242388890206843e-05, "loss": 0.0098, "num_input_tokens_seen": 35308832, "step": 18345 }, { "epoch": 2.9935557549555427, "grad_norm": 0.08832664042711258, "learning_rate": 2.8232654083163967e-05, "loss": 0.0278, "num_input_tokens_seen": 35318320, "step": 18350 }, { "epoch": 2.9943714821763603, "grad_norm": 1.835818886756897, "learning_rate": 2.822291877760521e-05, "loss": 0.1566, "num_input_tokens_seen": 35328256, "step": 18355 }, { "epoch": 2.9951872093971774, "grad_norm": 0.09528788179159164, "learning_rate": 2.8213182975031864e-05, "loss": 0.0594, "num_input_tokens_seen": 35337984, "step": 18360 }, { "epoch": 2.996002936617995, "grad_norm": 1.7721539735794067, "learning_rate": 2.8203446676945337e-05, "loss": 0.0719, "num_input_tokens_seen": 35347136, "step": 18365 }, { "epoch": 2.9968186638388126, "grad_norm": 0.17838728427886963, "learning_rate": 2.8193709884847075e-05, "loss": 0.0086, "num_input_tokens_seen": 35357040, "step": 18370 }, { "epoch": 2.9976343910596297, "grad_norm": 5.37764835357666, "learning_rate": 2.8183972600238605e-05, "loss": 0.0657, "num_input_tokens_seen": 35366304, "step": 18375 }, { "epoch": 2.998450118280447, "grad_norm": 1.742262601852417, "learning_rate": 2.817423482462156e-05, "loss": 0.0714, "num_input_tokens_seen": 35374736, "step": 18380 }, { "epoch": 2.9992658455012644, "grad_norm": 1.122757911682129, "learning_rate": 2.8164496559497605e-05, "loss": 0.024, "num_input_tokens_seen": 35382944, "step": 18385 }, { "epoch": 3.0, "grad_norm": 8.288668632507324, "learning_rate": 2.815475780636852e-05, "loss": 0.2098, "num_input_tokens_seen": 35390256, "step": 18390 }, { "epoch": 3.0008157272208176, "grad_norm": 0.1840137243270874, "learning_rate": 2.814501856673613e-05, "loss": 0.0066, "num_input_tokens_seen": 35399136, "step": 18395 }, { "epoch": 3.0016314544416347, "grad_norm": 0.046300191432237625, "learning_rate": 2.8135278842102353e-05, "loss": 0.0212, "num_input_tokens_seen": 35410304, "step": 18400 }, { "epoch": 3.0016314544416347, "eval_loss": 0.16577646136283875, "eval_runtime": 80.7673, "eval_samples_per_second": 33.739, "eval_steps_per_second": 16.876, "num_input_tokens_seen": 35410304, "step": 18400 }, { "epoch": 3.0024471816624523, "grad_norm": 0.1343427151441574, "learning_rate": 2.8125538633969183e-05, "loss": 0.0086, "num_input_tokens_seen": 35420608, "step": 18405 }, { "epoch": 3.0032629088832694, "grad_norm": 1.9248394966125488, "learning_rate": 2.8115797943838677e-05, "loss": 0.0231, "num_input_tokens_seen": 35429248, "step": 18410 }, { "epoch": 3.004078636104087, "grad_norm": 3.2549896240234375, "learning_rate": 2.810605677321298e-05, "loss": 0.0969, "num_input_tokens_seen": 35439968, "step": 18415 }, { "epoch": 3.004894363324904, "grad_norm": 0.13404066860675812, "learning_rate": 2.809631512359428e-05, "loss": 0.0812, "num_input_tokens_seen": 35448848, "step": 18420 }, { "epoch": 3.0057100905457217, "grad_norm": 0.20380617678165436, "learning_rate": 2.8086572996484884e-05, "loss": 0.0272, "num_input_tokens_seen": 35459168, "step": 18425 }, { "epoch": 3.006525817766539, "grad_norm": 0.40683993697166443, "learning_rate": 2.8076830393387143e-05, "loss": 0.092, "num_input_tokens_seen": 35468368, "step": 18430 }, { "epoch": 3.0073415449873564, "grad_norm": 0.5867032408714294, "learning_rate": 2.8067087315803497e-05, "loss": 0.0037, "num_input_tokens_seen": 35478416, "step": 18435 }, { "epoch": 3.0081572722081735, "grad_norm": 2.1325294971466064, "learning_rate": 2.8057343765236433e-05, "loss": 0.114, "num_input_tokens_seen": 35488432, "step": 18440 }, { "epoch": 3.008972999428991, "grad_norm": 10.228546142578125, "learning_rate": 2.804759974318854e-05, "loss": 0.035, "num_input_tokens_seen": 35498592, "step": 18445 }, { "epoch": 3.0097887266498082, "grad_norm": 0.03797043487429619, "learning_rate": 2.8037855251162482e-05, "loss": 0.1108, "num_input_tokens_seen": 35508768, "step": 18450 }, { "epoch": 3.010604453870626, "grad_norm": 0.14626318216323853, "learning_rate": 2.802811029066096e-05, "loss": 0.0031, "num_input_tokens_seen": 35519824, "step": 18455 }, { "epoch": 3.011420181091443, "grad_norm": 0.019640378654003143, "learning_rate": 2.8018364863186764e-05, "loss": 0.0065, "num_input_tokens_seen": 35530592, "step": 18460 }, { "epoch": 3.0122359083122605, "grad_norm": 19.109037399291992, "learning_rate": 2.800861897024279e-05, "loss": 0.0358, "num_input_tokens_seen": 35540416, "step": 18465 }, { "epoch": 3.0130516355330776, "grad_norm": 0.1003599688410759, "learning_rate": 2.799887261333196e-05, "loss": 0.0184, "num_input_tokens_seen": 35550224, "step": 18470 }, { "epoch": 3.013867362753895, "grad_norm": 0.27475079894065857, "learning_rate": 2.798912579395728e-05, "loss": 0.1635, "num_input_tokens_seen": 35559344, "step": 18475 }, { "epoch": 3.0146830899747123, "grad_norm": 0.006660666316747665, "learning_rate": 2.797937851362185e-05, "loss": 0.002, "num_input_tokens_seen": 35569280, "step": 18480 }, { "epoch": 3.01549881719553, "grad_norm": 0.10249733924865723, "learning_rate": 2.7969630773828802e-05, "loss": 0.0018, "num_input_tokens_seen": 35577616, "step": 18485 }, { "epoch": 3.016314544416347, "grad_norm": 0.013578456826508045, "learning_rate": 2.7959882576081382e-05, "loss": 0.0041, "num_input_tokens_seen": 35586976, "step": 18490 }, { "epoch": 3.0171302716371646, "grad_norm": 0.11479266732931137, "learning_rate": 2.795013392188286e-05, "loss": 0.0031, "num_input_tokens_seen": 35596928, "step": 18495 }, { "epoch": 3.0179459988579818, "grad_norm": 11.678627014160156, "learning_rate": 2.7940384812736614e-05, "loss": 0.0368, "num_input_tokens_seen": 35607296, "step": 18500 }, { "epoch": 3.0187617260787993, "grad_norm": 5.61922550201416, "learning_rate": 2.7930635250146087e-05, "loss": 0.0668, "num_input_tokens_seen": 35617248, "step": 18505 }, { "epoch": 3.0195774532996165, "grad_norm": 0.07897403091192245, "learning_rate": 2.792088523561477e-05, "loss": 0.003, "num_input_tokens_seen": 35628432, "step": 18510 }, { "epoch": 3.020393180520434, "grad_norm": 6.669342517852783, "learning_rate": 2.7911134770646246e-05, "loss": 0.1072, "num_input_tokens_seen": 35638784, "step": 18515 }, { "epoch": 3.021208907741251, "grad_norm": 0.1882418394088745, "learning_rate": 2.7901383856744157e-05, "loss": 0.0038, "num_input_tokens_seen": 35648496, "step": 18520 }, { "epoch": 3.0220246349620687, "grad_norm": 0.10703099519014359, "learning_rate": 2.7891632495412217e-05, "loss": 0.0186, "num_input_tokens_seen": 35658784, "step": 18525 }, { "epoch": 3.022840362182886, "grad_norm": 0.010686877183616161, "learning_rate": 2.7881880688154205e-05, "loss": 0.001, "num_input_tokens_seen": 35669328, "step": 18530 }, { "epoch": 3.0236560894037035, "grad_norm": 0.05324617028236389, "learning_rate": 2.7872128436473977e-05, "loss": 0.1083, "num_input_tokens_seen": 35679200, "step": 18535 }, { "epoch": 3.0244718166245206, "grad_norm": 0.057277861982584, "learning_rate": 2.7862375741875448e-05, "loss": 0.0901, "num_input_tokens_seen": 35689296, "step": 18540 }, { "epoch": 3.025287543845338, "grad_norm": 0.09001282602548599, "learning_rate": 2.785262260586261e-05, "loss": 0.0461, "num_input_tokens_seen": 35699776, "step": 18545 }, { "epoch": 3.0261032710661553, "grad_norm": 0.14849026501178741, "learning_rate": 2.7842869029939517e-05, "loss": 0.034, "num_input_tokens_seen": 35708800, "step": 18550 }, { "epoch": 3.026918998286973, "grad_norm": 0.06037577986717224, "learning_rate": 2.7833115015610296e-05, "loss": 0.0018, "num_input_tokens_seen": 35718688, "step": 18555 }, { "epoch": 3.02773472550779, "grad_norm": 0.042600587010383606, "learning_rate": 2.7823360564379136e-05, "loss": 0.001, "num_input_tokens_seen": 35727376, "step": 18560 }, { "epoch": 3.0285504527286076, "grad_norm": 1.6241509914398193, "learning_rate": 2.7813605677750297e-05, "loss": 0.0374, "num_input_tokens_seen": 35735664, "step": 18565 }, { "epoch": 3.0293661799494247, "grad_norm": 0.0725112333893776, "learning_rate": 2.7803850357228102e-05, "loss": 0.0024, "num_input_tokens_seen": 35746832, "step": 18570 }, { "epoch": 3.0301819071702423, "grad_norm": 7.539488315582275, "learning_rate": 2.779409460431695e-05, "loss": 0.0749, "num_input_tokens_seen": 35757200, "step": 18575 }, { "epoch": 3.03099763439106, "grad_norm": 0.12899836897850037, "learning_rate": 2.778433842052129e-05, "loss": 0.0023, "num_input_tokens_seen": 35767264, "step": 18580 }, { "epoch": 3.031813361611877, "grad_norm": 0.14373008906841278, "learning_rate": 2.7774581807345664e-05, "loss": 0.101, "num_input_tokens_seen": 35778000, "step": 18585 }, { "epoch": 3.0326290888326946, "grad_norm": 10.920551300048828, "learning_rate": 2.776482476629465e-05, "loss": 0.0161, "num_input_tokens_seen": 35788080, "step": 18590 }, { "epoch": 3.0334448160535117, "grad_norm": 24.01616668701172, "learning_rate": 2.7755067298872924e-05, "loss": 0.0462, "num_input_tokens_seen": 35797264, "step": 18595 }, { "epoch": 3.0342605432743293, "grad_norm": 0.006025353912264109, "learning_rate": 2.774530940658518e-05, "loss": 0.0058, "num_input_tokens_seen": 35808688, "step": 18600 }, { "epoch": 3.0342605432743293, "eval_loss": 0.20917591452598572, "eval_runtime": 80.9308, "eval_samples_per_second": 33.671, "eval_steps_per_second": 16.842, "num_input_tokens_seen": 35808688, "step": 18600 }, { "epoch": 3.0350762704951464, "grad_norm": 0.03501379117369652, "learning_rate": 2.7735551090936236e-05, "loss": 0.0021, "num_input_tokens_seen": 35818112, "step": 18605 }, { "epoch": 3.035891997715964, "grad_norm": 4.556329250335693, "learning_rate": 2.7725792353430934e-05, "loss": 0.0309, "num_input_tokens_seen": 35828208, "step": 18610 }, { "epoch": 3.036707724936781, "grad_norm": 0.32974809408187866, "learning_rate": 2.77160331955742e-05, "loss": 0.0734, "num_input_tokens_seen": 35837984, "step": 18615 }, { "epoch": 3.0375234521575987, "grad_norm": 0.020463936030864716, "learning_rate": 2.7706273618871008e-05, "loss": 0.0245, "num_input_tokens_seen": 35848416, "step": 18620 }, { "epoch": 3.038339179378416, "grad_norm": 1.5128107070922852, "learning_rate": 2.769651362482642e-05, "loss": 0.0766, "num_input_tokens_seen": 35857680, "step": 18625 }, { "epoch": 3.0391549065992334, "grad_norm": 0.031092820689082146, "learning_rate": 2.768675321494555e-05, "loss": 0.0775, "num_input_tokens_seen": 35867248, "step": 18630 }, { "epoch": 3.0399706338200505, "grad_norm": 0.005933571606874466, "learning_rate": 2.7676992390733565e-05, "loss": 0.026, "num_input_tokens_seen": 35877440, "step": 18635 }, { "epoch": 3.040786361040868, "grad_norm": 0.04612031206488609, "learning_rate": 2.766723115369571e-05, "loss": 0.0017, "num_input_tokens_seen": 35886640, "step": 18640 }, { "epoch": 3.041602088261685, "grad_norm": 1.4577860832214355, "learning_rate": 2.765746950533729e-05, "loss": 0.0808, "num_input_tokens_seen": 35896096, "step": 18645 }, { "epoch": 3.042417815482503, "grad_norm": 6.066333770751953, "learning_rate": 2.7647707447163684e-05, "loss": 0.0867, "num_input_tokens_seen": 35907264, "step": 18650 }, { "epoch": 3.04323354270332, "grad_norm": 0.046831462532281876, "learning_rate": 2.7637944980680315e-05, "loss": 0.0391, "num_input_tokens_seen": 35916656, "step": 18655 }, { "epoch": 3.0440492699241375, "grad_norm": 0.024889947846531868, "learning_rate": 2.762818210739268e-05, "loss": 0.0877, "num_input_tokens_seen": 35926448, "step": 18660 }, { "epoch": 3.0448649971449546, "grad_norm": 0.04462418705224991, "learning_rate": 2.7618418828806332e-05, "loss": 0.0011, "num_input_tokens_seen": 35935488, "step": 18665 }, { "epoch": 3.045680724365772, "grad_norm": 0.025036808103322983, "learning_rate": 2.76086551464269e-05, "loss": 0.0843, "num_input_tokens_seen": 35944400, "step": 18670 }, { "epoch": 3.0464964515865893, "grad_norm": 0.01695391908288002, "learning_rate": 2.759889106176006e-05, "loss": 0.0023, "num_input_tokens_seen": 35955536, "step": 18675 }, { "epoch": 3.047312178807407, "grad_norm": 5.301802158355713, "learning_rate": 2.758912657631156e-05, "loss": 0.1327, "num_input_tokens_seen": 35965744, "step": 18680 }, { "epoch": 3.048127906028224, "grad_norm": 0.23066428303718567, "learning_rate": 2.7579361691587198e-05, "loss": 0.0013, "num_input_tokens_seen": 35976560, "step": 18685 }, { "epoch": 3.0489436332490416, "grad_norm": 0.20935320854187012, "learning_rate": 2.756959640909285e-05, "loss": 0.0046, "num_input_tokens_seen": 35987312, "step": 18690 }, { "epoch": 3.0497593604698587, "grad_norm": 14.311943054199219, "learning_rate": 2.7559830730334452e-05, "loss": 0.0508, "num_input_tokens_seen": 35996928, "step": 18695 }, { "epoch": 3.0505750876906763, "grad_norm": 0.0373285748064518, "learning_rate": 2.7550064656817988e-05, "loss": 0.0011, "num_input_tokens_seen": 36005888, "step": 18700 }, { "epoch": 3.0513908149114934, "grad_norm": 12.909080505371094, "learning_rate": 2.7540298190049503e-05, "loss": 0.0181, "num_input_tokens_seen": 36016432, "step": 18705 }, { "epoch": 3.052206542132311, "grad_norm": 29.47242546081543, "learning_rate": 2.7530531331535107e-05, "loss": 0.0446, "num_input_tokens_seen": 36026160, "step": 18710 }, { "epoch": 3.053022269353128, "grad_norm": 0.01104747224599123, "learning_rate": 2.752076408278099e-05, "loss": 0.1064, "num_input_tokens_seen": 36035360, "step": 18715 }, { "epoch": 3.0538379965739457, "grad_norm": 15.851882934570312, "learning_rate": 2.751099644529337e-05, "loss": 0.0337, "num_input_tokens_seen": 36046400, "step": 18720 }, { "epoch": 3.054653723794763, "grad_norm": 0.007206023205071688, "learning_rate": 2.7501228420578533e-05, "loss": 0.0777, "num_input_tokens_seen": 36054640, "step": 18725 }, { "epoch": 3.0554694510155804, "grad_norm": 0.13392828404903412, "learning_rate": 2.7491460010142857e-05, "loss": 0.1434, "num_input_tokens_seen": 36064480, "step": 18730 }, { "epoch": 3.0562851782363976, "grad_norm": 0.042804472148418427, "learning_rate": 2.7481691215492727e-05, "loss": 0.0596, "num_input_tokens_seen": 36073936, "step": 18735 }, { "epoch": 3.057100905457215, "grad_norm": 0.033152684569358826, "learning_rate": 2.747192203813463e-05, "loss": 0.0011, "num_input_tokens_seen": 36084240, "step": 18740 }, { "epoch": 3.0579166326780323, "grad_norm": 5.5546135902404785, "learning_rate": 2.7462152479575087e-05, "loss": 0.1184, "num_input_tokens_seen": 36093728, "step": 18745 }, { "epoch": 3.05873235989885, "grad_norm": 8.176497459411621, "learning_rate": 2.7452382541320697e-05, "loss": 0.0996, "num_input_tokens_seen": 36103456, "step": 18750 }, { "epoch": 3.059548087119667, "grad_norm": 0.03724995627999306, "learning_rate": 2.7442612224878096e-05, "loss": 0.058, "num_input_tokens_seen": 36113440, "step": 18755 }, { "epoch": 3.0603638143404845, "grad_norm": 2.8443779945373535, "learning_rate": 2.7432841531753994e-05, "loss": 0.074, "num_input_tokens_seen": 36123728, "step": 18760 }, { "epoch": 3.0611795415613017, "grad_norm": 0.020347826182842255, "learning_rate": 2.7423070463455147e-05, "loss": 0.0717, "num_input_tokens_seen": 36133328, "step": 18765 }, { "epoch": 3.0619952687821193, "grad_norm": 3.8361849784851074, "learning_rate": 2.7413299021488397e-05, "loss": 0.1271, "num_input_tokens_seen": 36141952, "step": 18770 }, { "epoch": 3.062810996002937, "grad_norm": 0.07812763750553131, "learning_rate": 2.7403527207360615e-05, "loss": 0.0663, "num_input_tokens_seen": 36152192, "step": 18775 }, { "epoch": 3.063626723223754, "grad_norm": 0.05638013780117035, "learning_rate": 2.7393755022578722e-05, "loss": 0.0362, "num_input_tokens_seen": 36160992, "step": 18780 }, { "epoch": 3.0644424504445715, "grad_norm": 0.1731405258178711, "learning_rate": 2.7383982468649714e-05, "loss": 0.0063, "num_input_tokens_seen": 36171088, "step": 18785 }, { "epoch": 3.0652581776653887, "grad_norm": 0.08691735565662384, "learning_rate": 2.7374209547080665e-05, "loss": 0.0411, "num_input_tokens_seen": 36180192, "step": 18790 }, { "epoch": 3.0660739048862062, "grad_norm": 0.13987454771995544, "learning_rate": 2.7364436259378663e-05, "loss": 0.0064, "num_input_tokens_seen": 36190672, "step": 18795 }, { "epoch": 3.0668896321070234, "grad_norm": 0.0814172551035881, "learning_rate": 2.735466260705088e-05, "loss": 0.048, "num_input_tokens_seen": 36200720, "step": 18800 }, { "epoch": 3.0668896321070234, "eval_loss": 0.18425863981246948, "eval_runtime": 80.81, "eval_samples_per_second": 33.721, "eval_steps_per_second": 16.867, "num_input_tokens_seen": 36200720, "step": 18800 }, { "epoch": 3.067705359327841, "grad_norm": 0.02273900993168354, "learning_rate": 2.7344888591604524e-05, "loss": 0.0094, "num_input_tokens_seen": 36210592, "step": 18805 }, { "epoch": 3.068521086548658, "grad_norm": 0.08776378631591797, "learning_rate": 2.7335114214546893e-05, "loss": 0.0133, "num_input_tokens_seen": 36219040, "step": 18810 }, { "epoch": 3.0693368137694756, "grad_norm": 0.13827188313007355, "learning_rate": 2.7325339477385293e-05, "loss": 0.0098, "num_input_tokens_seen": 36228224, "step": 18815 }, { "epoch": 3.0701525409902928, "grad_norm": 9.942596435546875, "learning_rate": 2.7315564381627128e-05, "loss": 0.0879, "num_input_tokens_seen": 36238160, "step": 18820 }, { "epoch": 3.0709682682111104, "grad_norm": 0.13486938178539276, "learning_rate": 2.7305788928779835e-05, "loss": 0.0069, "num_input_tokens_seen": 36247760, "step": 18825 }, { "epoch": 3.0717839954319275, "grad_norm": 15.199189186096191, "learning_rate": 2.729601312035091e-05, "loss": 0.0444, "num_input_tokens_seen": 36255584, "step": 18830 }, { "epoch": 3.072599722652745, "grad_norm": 0.014051619917154312, "learning_rate": 2.7286236957847915e-05, "loss": 0.0068, "num_input_tokens_seen": 36264112, "step": 18835 }, { "epoch": 3.073415449873562, "grad_norm": 0.02902876026928425, "learning_rate": 2.7276460442778446e-05, "loss": 0.0008, "num_input_tokens_seen": 36273664, "step": 18840 }, { "epoch": 3.0742311770943798, "grad_norm": 0.11926239728927612, "learning_rate": 2.726668357665017e-05, "loss": 0.0425, "num_input_tokens_seen": 36284384, "step": 18845 }, { "epoch": 3.075046904315197, "grad_norm": 0.7233933806419373, "learning_rate": 2.7256906360970808e-05, "loss": 0.0023, "num_input_tokens_seen": 36294464, "step": 18850 }, { "epoch": 3.0758626315360145, "grad_norm": 0.07577712088823318, "learning_rate": 2.7247128797248117e-05, "loss": 0.0489, "num_input_tokens_seen": 36303424, "step": 18855 }, { "epoch": 3.0766783587568316, "grad_norm": 0.016077980399131775, "learning_rate": 2.7237350886989925e-05, "loss": 0.0042, "num_input_tokens_seen": 36312240, "step": 18860 }, { "epoch": 3.077494085977649, "grad_norm": 8.831189155578613, "learning_rate": 2.7227572631704107e-05, "loss": 0.0078, "num_input_tokens_seen": 36321456, "step": 18865 }, { "epoch": 3.0783098131984663, "grad_norm": 0.0796179324388504, "learning_rate": 2.7217794032898596e-05, "loss": 0.0011, "num_input_tokens_seen": 36331296, "step": 18870 }, { "epoch": 3.079125540419284, "grad_norm": 0.01017540879547596, "learning_rate": 2.7208015092081384e-05, "loss": 0.0725, "num_input_tokens_seen": 36341344, "step": 18875 }, { "epoch": 3.079941267640101, "grad_norm": 2.7388052940368652, "learning_rate": 2.719823581076049e-05, "loss": 0.0797, "num_input_tokens_seen": 36349552, "step": 18880 }, { "epoch": 3.0807569948609186, "grad_norm": 0.03265150636434555, "learning_rate": 2.718845619044401e-05, "loss": 0.0775, "num_input_tokens_seen": 36358880, "step": 18885 }, { "epoch": 3.0815727220817357, "grad_norm": 0.02662089839577675, "learning_rate": 2.7178676232640088e-05, "loss": 0.0006, "num_input_tokens_seen": 36368576, "step": 18890 }, { "epoch": 3.0823884493025533, "grad_norm": 0.0649031326174736, "learning_rate": 2.716889593885691e-05, "loss": 0.001, "num_input_tokens_seen": 36376832, "step": 18895 }, { "epoch": 3.0832041765233704, "grad_norm": 0.013551315292716026, "learning_rate": 2.7159115310602716e-05, "loss": 0.0325, "num_input_tokens_seen": 36387712, "step": 18900 }, { "epoch": 3.084019903744188, "grad_norm": 0.008263272233307362, "learning_rate": 2.7149334349385814e-05, "loss": 0.0472, "num_input_tokens_seen": 36397456, "step": 18905 }, { "epoch": 3.084835630965005, "grad_norm": 0.013924223370850086, "learning_rate": 2.713955305671454e-05, "loss": 0.0005, "num_input_tokens_seen": 36406944, "step": 18910 }, { "epoch": 3.0856513581858227, "grad_norm": 0.02556474320590496, "learning_rate": 2.71297714340973e-05, "loss": 0.0636, "num_input_tokens_seen": 36416864, "step": 18915 }, { "epoch": 3.08646708540664, "grad_norm": 0.02115175500512123, "learning_rate": 2.7119989483042545e-05, "loss": 0.0342, "num_input_tokens_seen": 36426880, "step": 18920 }, { "epoch": 3.0872828126274574, "grad_norm": 0.018491409718990326, "learning_rate": 2.7110207205058768e-05, "loss": 0.0004, "num_input_tokens_seen": 36435840, "step": 18925 }, { "epoch": 3.0880985398482745, "grad_norm": 0.43555450439453125, "learning_rate": 2.7100424601654517e-05, "loss": 0.0017, "num_input_tokens_seen": 36444672, "step": 18930 }, { "epoch": 3.088914267069092, "grad_norm": 0.010241419076919556, "learning_rate": 2.7090641674338403e-05, "loss": 0.0033, "num_input_tokens_seen": 36453744, "step": 18935 }, { "epoch": 3.0897299942899092, "grad_norm": 0.04943776875734329, "learning_rate": 2.7080858424619072e-05, "loss": 0.0952, "num_input_tokens_seen": 36464016, "step": 18940 }, { "epoch": 3.090545721510727, "grad_norm": 7.667379856109619, "learning_rate": 2.707107485400521e-05, "loss": 0.0496, "num_input_tokens_seen": 36472800, "step": 18945 }, { "epoch": 3.0913614487315444, "grad_norm": 0.010717482306063175, "learning_rate": 2.7061290964005586e-05, "loss": 0.0018, "num_input_tokens_seen": 36482672, "step": 18950 }, { "epoch": 3.0921771759523615, "grad_norm": 0.03946332633495331, "learning_rate": 2.7051506756129e-05, "loss": 0.1169, "num_input_tokens_seen": 36493216, "step": 18955 }, { "epoch": 3.092992903173179, "grad_norm": 6.425327777862549, "learning_rate": 2.704172223188428e-05, "loss": 0.0683, "num_input_tokens_seen": 36501840, "step": 18960 }, { "epoch": 3.0938086303939962, "grad_norm": 0.06597268581390381, "learning_rate": 2.7031937392780334e-05, "loss": 0.0035, "num_input_tokens_seen": 36511456, "step": 18965 }, { "epoch": 3.094624357614814, "grad_norm": 0.01076358463615179, "learning_rate": 2.702215224032611e-05, "loss": 0.1277, "num_input_tokens_seen": 36520640, "step": 18970 }, { "epoch": 3.095440084835631, "grad_norm": 0.07807300984859467, "learning_rate": 2.70123667760306e-05, "loss": 0.0899, "num_input_tokens_seen": 36530608, "step": 18975 }, { "epoch": 3.0962558120564485, "grad_norm": 17.103946685791016, "learning_rate": 2.7002581001402845e-05, "loss": 0.0193, "num_input_tokens_seen": 36541360, "step": 18980 }, { "epoch": 3.0970715392772656, "grad_norm": 0.06313730031251907, "learning_rate": 2.6992794917951923e-05, "loss": 0.0635, "num_input_tokens_seen": 36550992, "step": 18985 }, { "epoch": 3.097887266498083, "grad_norm": 0.020368091762065887, "learning_rate": 2.6983008527187e-05, "loss": 0.0187, "num_input_tokens_seen": 36560656, "step": 18990 }, { "epoch": 3.0987029937189003, "grad_norm": 0.11044154316186905, "learning_rate": 2.697322183061723e-05, "loss": 0.0029, "num_input_tokens_seen": 36570736, "step": 18995 }, { "epoch": 3.099518720939718, "grad_norm": 0.3026280105113983, "learning_rate": 2.696343482975186e-05, "loss": 0.1001, "num_input_tokens_seen": 36580112, "step": 19000 }, { "epoch": 3.099518720939718, "eval_loss": 0.22107887268066406, "eval_runtime": 80.779, "eval_samples_per_second": 33.734, "eval_steps_per_second": 16.873, "num_input_tokens_seen": 36580112, "step": 19000 }, { "epoch": 3.100334448160535, "grad_norm": 0.01758621074259281, "learning_rate": 2.695364752610016e-05, "loss": 0.03, "num_input_tokens_seen": 36590096, "step": 19005 }, { "epoch": 3.1011501753813526, "grad_norm": 0.021818039938807487, "learning_rate": 2.6943859921171467e-05, "loss": 0.0017, "num_input_tokens_seen": 36599936, "step": 19010 }, { "epoch": 3.1019659026021698, "grad_norm": 0.2975093424320221, "learning_rate": 2.6934072016475143e-05, "loss": 0.0024, "num_input_tokens_seen": 36609280, "step": 19015 }, { "epoch": 3.1027816298229873, "grad_norm": 0.10076608508825302, "learning_rate": 2.6924283813520606e-05, "loss": 0.0015, "num_input_tokens_seen": 36618704, "step": 19020 }, { "epoch": 3.1035973570438045, "grad_norm": 0.013476711697876453, "learning_rate": 2.691449531381733e-05, "loss": 0.0697, "num_input_tokens_seen": 36628144, "step": 19025 }, { "epoch": 3.104413084264622, "grad_norm": 0.019229654222726822, "learning_rate": 2.6904706518874816e-05, "loss": 0.0053, "num_input_tokens_seen": 36638704, "step": 19030 }, { "epoch": 3.105228811485439, "grad_norm": 2.985271692276001, "learning_rate": 2.6894917430202615e-05, "loss": 0.0051, "num_input_tokens_seen": 36647680, "step": 19035 }, { "epoch": 3.1060445387062567, "grad_norm": 19.120237350463867, "learning_rate": 2.6885128049310343e-05, "loss": 0.0598, "num_input_tokens_seen": 36659360, "step": 19040 }, { "epoch": 3.106860265927074, "grad_norm": 0.018518483266234398, "learning_rate": 2.687533837770762e-05, "loss": 0.1114, "num_input_tokens_seen": 36668752, "step": 19045 }, { "epoch": 3.1076759931478914, "grad_norm": 0.021827153861522675, "learning_rate": 2.6865548416904162e-05, "loss": 0.0041, "num_input_tokens_seen": 36678480, "step": 19050 }, { "epoch": 3.1084917203687086, "grad_norm": 0.029251275584101677, "learning_rate": 2.68557581684097e-05, "loss": 0.0025, "num_input_tokens_seen": 36686720, "step": 19055 }, { "epoch": 3.109307447589526, "grad_norm": 4.608001232147217, "learning_rate": 2.6845967633733998e-05, "loss": 0.1347, "num_input_tokens_seen": 36695920, "step": 19060 }, { "epoch": 3.1101231748103433, "grad_norm": 0.18184004724025726, "learning_rate": 2.683617681438689e-05, "loss": 0.0939, "num_input_tokens_seen": 36706064, "step": 19065 }, { "epoch": 3.110938902031161, "grad_norm": 65.92681884765625, "learning_rate": 2.682638571187825e-05, "loss": 0.0343, "num_input_tokens_seen": 36714176, "step": 19070 }, { "epoch": 3.111754629251978, "grad_norm": 0.05738804116845131, "learning_rate": 2.6816594327717976e-05, "loss": 0.0019, "num_input_tokens_seen": 36724848, "step": 19075 }, { "epoch": 3.1125703564727956, "grad_norm": 0.05137452855706215, "learning_rate": 2.680680266341603e-05, "loss": 0.0864, "num_input_tokens_seen": 36733984, "step": 19080 }, { "epoch": 3.1133860836936127, "grad_norm": 14.577560424804688, "learning_rate": 2.67970107204824e-05, "loss": 0.054, "num_input_tokens_seen": 36742496, "step": 19085 }, { "epoch": 3.1142018109144303, "grad_norm": 0.02703697234392166, "learning_rate": 2.6787218500427142e-05, "loss": 0.0006, "num_input_tokens_seen": 36752368, "step": 19090 }, { "epoch": 3.1150175381352474, "grad_norm": 18.10379981994629, "learning_rate": 2.6777426004760332e-05, "loss": 0.1711, "num_input_tokens_seen": 36763344, "step": 19095 }, { "epoch": 3.115833265356065, "grad_norm": 0.10247480124235153, "learning_rate": 2.6767633234992094e-05, "loss": 0.002, "num_input_tokens_seen": 36771520, "step": 19100 }, { "epoch": 3.116648992576882, "grad_norm": 0.06014490872621536, "learning_rate": 2.6757840192632598e-05, "loss": 0.002, "num_input_tokens_seen": 36781312, "step": 19105 }, { "epoch": 3.1174647197976997, "grad_norm": 1.6001056432724, "learning_rate": 2.6748046879192052e-05, "loss": 0.1435, "num_input_tokens_seen": 36789984, "step": 19110 }, { "epoch": 3.118280447018517, "grad_norm": 0.0268429946154356, "learning_rate": 2.673825329618071e-05, "loss": 0.001, "num_input_tokens_seen": 36799520, "step": 19115 }, { "epoch": 3.1190961742393344, "grad_norm": 0.009722678922116756, "learning_rate": 2.6728459445108866e-05, "loss": 0.0682, "num_input_tokens_seen": 36809568, "step": 19120 }, { "epoch": 3.1199119014601515, "grad_norm": 6.531730651855469, "learning_rate": 2.6718665327486854e-05, "loss": 0.0779, "num_input_tokens_seen": 36819632, "step": 19125 }, { "epoch": 3.120727628680969, "grad_norm": 0.029178796336054802, "learning_rate": 2.6708870944825048e-05, "loss": 0.0033, "num_input_tokens_seen": 36829568, "step": 19130 }, { "epoch": 3.121543355901786, "grad_norm": 0.10504882037639618, "learning_rate": 2.6699076298633874e-05, "loss": 0.0723, "num_input_tokens_seen": 36839328, "step": 19135 }, { "epoch": 3.122359083122604, "grad_norm": 0.06255598366260529, "learning_rate": 2.6689281390423788e-05, "loss": 0.1543, "num_input_tokens_seen": 36848672, "step": 19140 }, { "epoch": 3.1231748103434214, "grad_norm": 0.043260470032691956, "learning_rate": 2.667948622170527e-05, "loss": 0.0464, "num_input_tokens_seen": 36858624, "step": 19145 }, { "epoch": 3.1239905375642385, "grad_norm": 0.09127151966094971, "learning_rate": 2.6669690793988873e-05, "loss": 0.0054, "num_input_tokens_seen": 36867728, "step": 19150 }, { "epoch": 3.124806264785056, "grad_norm": 0.718860387802124, "learning_rate": 2.665989510878518e-05, "loss": 0.0056, "num_input_tokens_seen": 36876832, "step": 19155 }, { "epoch": 3.125621992005873, "grad_norm": 8.616241455078125, "learning_rate": 2.6650099167604793e-05, "loss": 0.0706, "num_input_tokens_seen": 36886944, "step": 19160 }, { "epoch": 3.126437719226691, "grad_norm": 0.06876213848590851, "learning_rate": 2.6640302971958376e-05, "loss": 0.0703, "num_input_tokens_seen": 36895472, "step": 19165 }, { "epoch": 3.127253446447508, "grad_norm": 0.0613069087266922, "learning_rate": 2.6630506523356635e-05, "loss": 0.055, "num_input_tokens_seen": 36904912, "step": 19170 }, { "epoch": 3.1280691736683255, "grad_norm": 3.453028678894043, "learning_rate": 2.6620709823310297e-05, "loss": 0.0909, "num_input_tokens_seen": 36915424, "step": 19175 }, { "epoch": 3.1288849008891426, "grad_norm": 12.669260025024414, "learning_rate": 2.661091287333014e-05, "loss": 0.0067, "num_input_tokens_seen": 36924240, "step": 19180 }, { "epoch": 3.12970062810996, "grad_norm": 0.10534396767616272, "learning_rate": 2.660111567492696e-05, "loss": 0.0031, "num_input_tokens_seen": 36934448, "step": 19185 }, { "epoch": 3.1305163553307773, "grad_norm": 0.12284586578607559, "learning_rate": 2.6591318229611635e-05, "loss": 0.0015, "num_input_tokens_seen": 36943664, "step": 19190 }, { "epoch": 3.131332082551595, "grad_norm": 0.14884500205516815, "learning_rate": 2.6581520538895037e-05, "loss": 0.0026, "num_input_tokens_seen": 36954064, "step": 19195 }, { "epoch": 3.132147809772412, "grad_norm": 7.529445648193359, "learning_rate": 2.6571722604288102e-05, "loss": 0.1011, "num_input_tokens_seen": 36961872, "step": 19200 }, { "epoch": 3.132147809772412, "eval_loss": 0.22590163350105286, "eval_runtime": 80.8155, "eval_samples_per_second": 33.719, "eval_steps_per_second": 16.866, "num_input_tokens_seen": 36961872, "step": 19200 }, { "epoch": 3.1329635369932296, "grad_norm": 0.1195962205529213, "learning_rate": 2.656192442730179e-05, "loss": 0.0019, "num_input_tokens_seen": 36971328, "step": 19205 }, { "epoch": 3.1337792642140467, "grad_norm": 0.0862114429473877, "learning_rate": 2.6552126009447098e-05, "loss": 0.0024, "num_input_tokens_seen": 36980224, "step": 19210 }, { "epoch": 3.1345949914348643, "grad_norm": 14.857142448425293, "learning_rate": 2.654232735223507e-05, "loss": 0.1223, "num_input_tokens_seen": 36989024, "step": 19215 }, { "epoch": 3.1354107186556814, "grad_norm": 0.04321800172328949, "learning_rate": 2.6532528457176787e-05, "loss": 0.0517, "num_input_tokens_seen": 36997968, "step": 19220 }, { "epoch": 3.136226445876499, "grad_norm": 0.05695275589823723, "learning_rate": 2.6522729325783348e-05, "loss": 0.08, "num_input_tokens_seen": 37007136, "step": 19225 }, { "epoch": 3.137042173097316, "grad_norm": 0.03869447112083435, "learning_rate": 2.6512929959565914e-05, "loss": 0.1886, "num_input_tokens_seen": 37017264, "step": 19230 }, { "epoch": 3.1378579003181337, "grad_norm": 0.07884635776281357, "learning_rate": 2.6503130360035673e-05, "loss": 0.1244, "num_input_tokens_seen": 37026880, "step": 19235 }, { "epoch": 3.138673627538951, "grad_norm": 14.745856285095215, "learning_rate": 2.6493330528703835e-05, "loss": 0.0289, "num_input_tokens_seen": 37035984, "step": 19240 }, { "epoch": 3.1394893547597684, "grad_norm": 9.914064407348633, "learning_rate": 2.648353046708167e-05, "loss": 0.1105, "num_input_tokens_seen": 37047760, "step": 19245 }, { "epoch": 3.1403050819805856, "grad_norm": 0.09371268004179001, "learning_rate": 2.647373017668046e-05, "loss": 0.0058, "num_input_tokens_seen": 37058144, "step": 19250 }, { "epoch": 3.141120809201403, "grad_norm": 1.3496935367584229, "learning_rate": 2.6463929659011537e-05, "loss": 0.1345, "num_input_tokens_seen": 37067760, "step": 19255 }, { "epoch": 3.1419365364222203, "grad_norm": 2.2372970581054688, "learning_rate": 2.6454128915586262e-05, "loss": 0.0031, "num_input_tokens_seen": 37076384, "step": 19260 }, { "epoch": 3.142752263643038, "grad_norm": 0.03508375957608223, "learning_rate": 2.6444327947916036e-05, "loss": 0.0012, "num_input_tokens_seen": 37085968, "step": 19265 }, { "epoch": 3.143567990863855, "grad_norm": 8.18043327331543, "learning_rate": 2.6434526757512292e-05, "loss": 0.0662, "num_input_tokens_seen": 37096128, "step": 19270 }, { "epoch": 3.1443837180846725, "grad_norm": 0.0323517769575119, "learning_rate": 2.6424725345886486e-05, "loss": 0.061, "num_input_tokens_seen": 37105664, "step": 19275 }, { "epoch": 3.1451994453054897, "grad_norm": 0.035637401044368744, "learning_rate": 2.641492371455014e-05, "loss": 0.0019, "num_input_tokens_seen": 37114688, "step": 19280 }, { "epoch": 3.1460151725263072, "grad_norm": 0.010877061635255814, "learning_rate": 2.640512186501477e-05, "loss": 0.0114, "num_input_tokens_seen": 37124416, "step": 19285 }, { "epoch": 3.1468308997471244, "grad_norm": 18.501096725463867, "learning_rate": 2.639531979879195e-05, "loss": 0.1005, "num_input_tokens_seen": 37133680, "step": 19290 }, { "epoch": 3.147646626967942, "grad_norm": 0.10082552582025528, "learning_rate": 2.638551751739328e-05, "loss": 0.0021, "num_input_tokens_seen": 37144144, "step": 19295 }, { "epoch": 3.148462354188759, "grad_norm": 0.2042008936405182, "learning_rate": 2.6375715022330404e-05, "loss": 0.0099, "num_input_tokens_seen": 37154560, "step": 19300 }, { "epoch": 3.1492780814095767, "grad_norm": 0.011605328880250454, "learning_rate": 2.6365912315114976e-05, "loss": 0.0036, "num_input_tokens_seen": 37164848, "step": 19305 }, { "epoch": 3.150093808630394, "grad_norm": 25.068893432617188, "learning_rate": 2.6356109397258704e-05, "loss": 0.0981, "num_input_tokens_seen": 37176000, "step": 19310 }, { "epoch": 3.1509095358512114, "grad_norm": 0.020509082823991776, "learning_rate": 2.6346306270273325e-05, "loss": 0.0011, "num_input_tokens_seen": 37185680, "step": 19315 }, { "epoch": 3.151725263072029, "grad_norm": 0.00443639513105154, "learning_rate": 2.6336502935670608e-05, "loss": 0.0705, "num_input_tokens_seen": 37195600, "step": 19320 }, { "epoch": 3.152540990292846, "grad_norm": 0.07766662538051605, "learning_rate": 2.6326699394962333e-05, "loss": 0.0011, "num_input_tokens_seen": 37204464, "step": 19325 }, { "epoch": 3.153356717513663, "grad_norm": 4.960204124450684, "learning_rate": 2.6316895649660334e-05, "loss": 0.1484, "num_input_tokens_seen": 37213360, "step": 19330 }, { "epoch": 3.1541724447344808, "grad_norm": 0.1271876096725464, "learning_rate": 2.6307091701276486e-05, "loss": 0.0459, "num_input_tokens_seen": 37222608, "step": 19335 }, { "epoch": 3.1549881719552983, "grad_norm": 0.06578805297613144, "learning_rate": 2.629728755132267e-05, "loss": 0.0522, "num_input_tokens_seen": 37233072, "step": 19340 }, { "epoch": 3.1558038991761155, "grad_norm": 0.5317943692207336, "learning_rate": 2.628748320131081e-05, "loss": 0.0582, "num_input_tokens_seen": 37241984, "step": 19345 }, { "epoch": 3.156619626396933, "grad_norm": 0.010319624096155167, "learning_rate": 2.6277678652752856e-05, "loss": 0.0719, "num_input_tokens_seen": 37250832, "step": 19350 }, { "epoch": 3.15743535361775, "grad_norm": 0.009173870086669922, "learning_rate": 2.6267873907160807e-05, "loss": 0.0008, "num_input_tokens_seen": 37260608, "step": 19355 }, { "epoch": 3.1582510808385678, "grad_norm": 0.04896368086338043, "learning_rate": 2.6258068966046668e-05, "loss": 0.0017, "num_input_tokens_seen": 37271056, "step": 19360 }, { "epoch": 3.159066808059385, "grad_norm": 0.01659572497010231, "learning_rate": 2.6248263830922475e-05, "loss": 0.0007, "num_input_tokens_seen": 37280864, "step": 19365 }, { "epoch": 3.1598825352802025, "grad_norm": 12.67842960357666, "learning_rate": 2.6238458503300318e-05, "loss": 0.1153, "num_input_tokens_seen": 37290592, "step": 19370 }, { "epoch": 3.1606982625010196, "grad_norm": 5.866383075714111, "learning_rate": 2.6228652984692292e-05, "loss": 0.1214, "num_input_tokens_seen": 37299136, "step": 19375 }, { "epoch": 3.161513989721837, "grad_norm": 3.3789637088775635, "learning_rate": 2.621884727661054e-05, "loss": 0.0679, "num_input_tokens_seen": 37308272, "step": 19380 }, { "epoch": 3.1623297169426543, "grad_norm": 0.12869977951049805, "learning_rate": 2.6209041380567222e-05, "loss": 0.0019, "num_input_tokens_seen": 37317696, "step": 19385 }, { "epoch": 3.163145444163472, "grad_norm": 9.313343048095703, "learning_rate": 2.6199235298074527e-05, "loss": 0.0524, "num_input_tokens_seen": 37327312, "step": 19390 }, { "epoch": 3.163961171384289, "grad_norm": 0.1601957231760025, "learning_rate": 2.618942903064468e-05, "loss": 0.0023, "num_input_tokens_seen": 37336384, "step": 19395 }, { "epoch": 3.1647768986051066, "grad_norm": 1.88916015625, "learning_rate": 2.6179622579789932e-05, "loss": 0.0699, "num_input_tokens_seen": 37345136, "step": 19400 }, { "epoch": 3.1647768986051066, "eval_loss": 0.21505676209926605, "eval_runtime": 80.9094, "eval_samples_per_second": 33.68, "eval_steps_per_second": 16.846, "num_input_tokens_seen": 37345136, "step": 19400 }, { "epoch": 3.1655926258259237, "grad_norm": 4.060099124908447, "learning_rate": 2.6169815947022553e-05, "loss": 0.4192, "num_input_tokens_seen": 37354864, "step": 19405 }, { "epoch": 3.1664083530467413, "grad_norm": 0.03469521924853325, "learning_rate": 2.6160009133854853e-05, "loss": 0.0068, "num_input_tokens_seen": 37365408, "step": 19410 }, { "epoch": 3.1672240802675584, "grad_norm": 2.9531643390655518, "learning_rate": 2.6150202141799168e-05, "loss": 0.066, "num_input_tokens_seen": 37375952, "step": 19415 }, { "epoch": 3.168039807488376, "grad_norm": 3.920391082763672, "learning_rate": 2.614039497236786e-05, "loss": 0.047, "num_input_tokens_seen": 37385808, "step": 19420 }, { "epoch": 3.168855534709193, "grad_norm": 0.09784732758998871, "learning_rate": 2.6130587627073315e-05, "loss": 0.0024, "num_input_tokens_seen": 37395808, "step": 19425 }, { "epoch": 3.1696712619300107, "grad_norm": 21.382648468017578, "learning_rate": 2.6120780107427956e-05, "loss": 0.0501, "num_input_tokens_seen": 37404464, "step": 19430 }, { "epoch": 3.170486989150828, "grad_norm": 0.5798795223236084, "learning_rate": 2.6110972414944214e-05, "loss": 0.0867, "num_input_tokens_seen": 37414128, "step": 19435 }, { "epoch": 3.1713027163716454, "grad_norm": 0.048391252756118774, "learning_rate": 2.6101164551134565e-05, "loss": 0.0029, "num_input_tokens_seen": 37424240, "step": 19440 }, { "epoch": 3.1721184435924625, "grad_norm": 0.09883076697587967, "learning_rate": 2.6091356517511505e-05, "loss": 0.0132, "num_input_tokens_seen": 37433696, "step": 19445 }, { "epoch": 3.17293417081328, "grad_norm": 0.1422962248325348, "learning_rate": 2.608154831558755e-05, "loss": 0.0025, "num_input_tokens_seen": 37443392, "step": 19450 }, { "epoch": 3.1737498980340972, "grad_norm": 4.249868869781494, "learning_rate": 2.607173994687526e-05, "loss": 0.2242, "num_input_tokens_seen": 37453968, "step": 19455 }, { "epoch": 3.174565625254915, "grad_norm": 4.984252452850342, "learning_rate": 2.6061931412887196e-05, "loss": 0.1665, "num_input_tokens_seen": 37462336, "step": 19460 }, { "epoch": 3.175381352475732, "grad_norm": 0.03280777111649513, "learning_rate": 2.6052122715135973e-05, "loss": 0.0034, "num_input_tokens_seen": 37472944, "step": 19465 }, { "epoch": 3.1761970796965495, "grad_norm": 0.045917145907878876, "learning_rate": 2.60423138551342e-05, "loss": 0.0022, "num_input_tokens_seen": 37482736, "step": 19470 }, { "epoch": 3.1770128069173666, "grad_norm": 0.035162266343832016, "learning_rate": 2.6032504834394527e-05, "loss": 0.0367, "num_input_tokens_seen": 37491872, "step": 19475 }, { "epoch": 3.1778285341381842, "grad_norm": 3.4477856159210205, "learning_rate": 2.602269565442964e-05, "loss": 0.0623, "num_input_tokens_seen": 37502096, "step": 19480 }, { "epoch": 3.1786442613590014, "grad_norm": 19.46333885192871, "learning_rate": 2.6012886316752227e-05, "loss": 0.0712, "num_input_tokens_seen": 37512032, "step": 19485 }, { "epoch": 3.179459988579819, "grad_norm": 0.8408870697021484, "learning_rate": 2.6003076822875018e-05, "loss": 0.0048, "num_input_tokens_seen": 37522656, "step": 19490 }, { "epoch": 3.180275715800636, "grad_norm": 23.934167861938477, "learning_rate": 2.5993267174310755e-05, "loss": 0.0466, "num_input_tokens_seen": 37531840, "step": 19495 }, { "epoch": 3.1810914430214536, "grad_norm": 0.12252669781446457, "learning_rate": 2.5983457372572218e-05, "loss": 0.0592, "num_input_tokens_seen": 37540736, "step": 19500 }, { "epoch": 3.1819071702422708, "grad_norm": 0.19632588326931, "learning_rate": 2.597364741917219e-05, "loss": 0.0277, "num_input_tokens_seen": 37549248, "step": 19505 }, { "epoch": 3.1827228974630883, "grad_norm": 3.8361170291900635, "learning_rate": 2.5963837315623492e-05, "loss": 0.0085, "num_input_tokens_seen": 37558640, "step": 19510 }, { "epoch": 3.183538624683906, "grad_norm": 0.08668892830610275, "learning_rate": 2.595402706343897e-05, "loss": 0.002, "num_input_tokens_seen": 37568416, "step": 19515 }, { "epoch": 3.184354351904723, "grad_norm": 0.04894661530852318, "learning_rate": 2.594421666413148e-05, "loss": 0.0211, "num_input_tokens_seen": 37577808, "step": 19520 }, { "epoch": 3.1851700791255406, "grad_norm": 0.1622164100408554, "learning_rate": 2.5934406119213928e-05, "loss": 0.0017, "num_input_tokens_seen": 37588768, "step": 19525 }, { "epoch": 3.1859858063463578, "grad_norm": 0.007634622510522604, "learning_rate": 2.5924595430199193e-05, "loss": 0.0681, "num_input_tokens_seen": 37598560, "step": 19530 }, { "epoch": 3.1868015335671753, "grad_norm": 0.1320134997367859, "learning_rate": 2.5914784598600238e-05, "loss": 0.0607, "num_input_tokens_seen": 37608992, "step": 19535 }, { "epoch": 3.1876172607879925, "grad_norm": 0.19571201503276825, "learning_rate": 2.5904973625930002e-05, "loss": 0.002, "num_input_tokens_seen": 37619648, "step": 19540 }, { "epoch": 3.18843298800881, "grad_norm": 0.036804113537073135, "learning_rate": 2.5895162513701456e-05, "loss": 0.1104, "num_input_tokens_seen": 37630352, "step": 19545 }, { "epoch": 3.189248715229627, "grad_norm": 4.569889068603516, "learning_rate": 2.5885351263427593e-05, "loss": 0.0493, "num_input_tokens_seen": 37639440, "step": 19550 }, { "epoch": 3.1900644424504447, "grad_norm": 0.07193262875080109, "learning_rate": 2.5875539876621448e-05, "loss": 0.0782, "num_input_tokens_seen": 37649872, "step": 19555 }, { "epoch": 3.190880169671262, "grad_norm": 0.02591182291507721, "learning_rate": 2.586572835479605e-05, "loss": 0.0267, "num_input_tokens_seen": 37659088, "step": 19560 }, { "epoch": 3.1916958968920794, "grad_norm": 0.08124153316020966, "learning_rate": 2.585591669946446e-05, "loss": 0.0451, "num_input_tokens_seen": 37668080, "step": 19565 }, { "epoch": 3.1925116241128966, "grad_norm": 2.9490272998809814, "learning_rate": 2.5846104912139756e-05, "loss": 0.0687, "num_input_tokens_seen": 37678032, "step": 19570 }, { "epoch": 3.193327351333714, "grad_norm": 0.28466132283210754, "learning_rate": 2.583629299433505e-05, "loss": 0.0364, "num_input_tokens_seen": 37687728, "step": 19575 }, { "epoch": 3.1941430785545313, "grad_norm": 0.06095725670456886, "learning_rate": 2.582648094756345e-05, "loss": 0.0034, "num_input_tokens_seen": 37697104, "step": 19580 }, { "epoch": 3.194958805775349, "grad_norm": 0.06537720561027527, "learning_rate": 2.5816668773338098e-05, "loss": 0.0265, "num_input_tokens_seen": 37705536, "step": 19585 }, { "epoch": 3.195774532996166, "grad_norm": 0.09333989024162292, "learning_rate": 2.580685647317216e-05, "loss": 0.0017, "num_input_tokens_seen": 37714944, "step": 19590 }, { "epoch": 3.1965902602169836, "grad_norm": 0.03880314528942108, "learning_rate": 2.5797044048578818e-05, "loss": 0.1226, "num_input_tokens_seen": 37724512, "step": 19595 }, { "epoch": 3.1974059874378007, "grad_norm": 0.030203888192772865, "learning_rate": 2.5787231501071262e-05, "loss": 0.0018, "num_input_tokens_seen": 37732992, "step": 19600 }, { "epoch": 3.1974059874378007, "eval_loss": 0.22466830909252167, "eval_runtime": 80.8247, "eval_samples_per_second": 33.715, "eval_steps_per_second": 16.864, "num_input_tokens_seen": 37732992, "step": 19600 }, { "epoch": 3.1982217146586183, "grad_norm": 0.05040663480758667, "learning_rate": 2.577741883216272e-05, "loss": 0.0015, "num_input_tokens_seen": 37743120, "step": 19605 }, { "epoch": 3.1990374418794354, "grad_norm": 0.1957186907529831, "learning_rate": 2.576760604336642e-05, "loss": 0.0022, "num_input_tokens_seen": 37750736, "step": 19610 }, { "epoch": 3.199853169100253, "grad_norm": 10.73019790649414, "learning_rate": 2.575779313619563e-05, "loss": 0.0295, "num_input_tokens_seen": 37761600, "step": 19615 }, { "epoch": 3.20066889632107, "grad_norm": 0.007978960871696472, "learning_rate": 2.5747980112163605e-05, "loss": 0.0689, "num_input_tokens_seen": 37771328, "step": 19620 }, { "epoch": 3.2014846235418877, "grad_norm": 0.01160295307636261, "learning_rate": 2.5738166972783656e-05, "loss": 0.1619, "num_input_tokens_seen": 37781104, "step": 19625 }, { "epoch": 3.202300350762705, "grad_norm": 0.21324336528778076, "learning_rate": 2.5728353719569075e-05, "loss": 0.0379, "num_input_tokens_seen": 37791424, "step": 19630 }, { "epoch": 3.2031160779835224, "grad_norm": 0.04578011482954025, "learning_rate": 2.57185403540332e-05, "loss": 0.0462, "num_input_tokens_seen": 37800160, "step": 19635 }, { "epoch": 3.2039318052043395, "grad_norm": 0.010399658232927322, "learning_rate": 2.5708726877689375e-05, "loss": 0.001, "num_input_tokens_seen": 37810064, "step": 19640 }, { "epoch": 3.204747532425157, "grad_norm": 0.00537498714402318, "learning_rate": 2.5698913292050964e-05, "loss": 0.0803, "num_input_tokens_seen": 37818368, "step": 19645 }, { "epoch": 3.205563259645974, "grad_norm": 0.0331442691385746, "learning_rate": 2.568909959863133e-05, "loss": 0.0009, "num_input_tokens_seen": 37828192, "step": 19650 }, { "epoch": 3.206378986866792, "grad_norm": 0.015244841575622559, "learning_rate": 2.5679285798943887e-05, "loss": 0.0009, "num_input_tokens_seen": 37836416, "step": 19655 }, { "epoch": 3.207194714087609, "grad_norm": 0.07933181524276733, "learning_rate": 2.5669471894502035e-05, "loss": 0.0006, "num_input_tokens_seen": 37846736, "step": 19660 }, { "epoch": 3.2080104413084265, "grad_norm": 0.019060570746660233, "learning_rate": 2.56596578868192e-05, "loss": 0.044, "num_input_tokens_seen": 37855920, "step": 19665 }, { "epoch": 3.2088261685292436, "grad_norm": 0.08487197011709213, "learning_rate": 2.564984377740883e-05, "loss": 0.0035, "num_input_tokens_seen": 37865632, "step": 19670 }, { "epoch": 3.209641895750061, "grad_norm": 0.34878596663475037, "learning_rate": 2.564002956778438e-05, "loss": 0.0023, "num_input_tokens_seen": 37874336, "step": 19675 }, { "epoch": 3.2104576229708783, "grad_norm": 0.091056227684021, "learning_rate": 2.563021525945934e-05, "loss": 0.1804, "num_input_tokens_seen": 37885264, "step": 19680 }, { "epoch": 3.211273350191696, "grad_norm": 0.12491904199123383, "learning_rate": 2.562040085394718e-05, "loss": 0.0873, "num_input_tokens_seen": 37894240, "step": 19685 }, { "epoch": 3.2120890774125135, "grad_norm": 0.03672676533460617, "learning_rate": 2.56105863527614e-05, "loss": 0.1621, "num_input_tokens_seen": 37905600, "step": 19690 }, { "epoch": 3.2129048046333306, "grad_norm": 0.16493043303489685, "learning_rate": 2.5600771757415548e-05, "loss": 0.0397, "num_input_tokens_seen": 37914800, "step": 19695 }, { "epoch": 3.2137205318541477, "grad_norm": 0.1215800866484642, "learning_rate": 2.5590957069423134e-05, "loss": 0.0011, "num_input_tokens_seen": 37923520, "step": 19700 }, { "epoch": 3.2145362590749653, "grad_norm": 4.8510661125183105, "learning_rate": 2.5581142290297716e-05, "loss": 0.1061, "num_input_tokens_seen": 37933072, "step": 19705 }, { "epoch": 3.215351986295783, "grad_norm": 0.07105845957994461, "learning_rate": 2.557132742155285e-05, "loss": 0.0522, "num_input_tokens_seen": 37943440, "step": 19710 }, { "epoch": 3.2161677135166, "grad_norm": 8.152841567993164, "learning_rate": 2.556151246470212e-05, "loss": 0.0241, "num_input_tokens_seen": 37952928, "step": 19715 }, { "epoch": 3.2169834407374176, "grad_norm": 0.18338890373706818, "learning_rate": 2.5551697421259114e-05, "loss": 0.063, "num_input_tokens_seen": 37963232, "step": 19720 }, { "epoch": 3.2177991679582347, "grad_norm": 0.05328962579369545, "learning_rate": 2.554188229273743e-05, "loss": 0.0022, "num_input_tokens_seen": 37972864, "step": 19725 }, { "epoch": 3.2186148951790523, "grad_norm": 0.038942717015743256, "learning_rate": 2.5532067080650678e-05, "loss": 0.0633, "num_input_tokens_seen": 37982896, "step": 19730 }, { "epoch": 3.2194306223998694, "grad_norm": 37.78455352783203, "learning_rate": 2.55222517865125e-05, "loss": 0.051, "num_input_tokens_seen": 37992720, "step": 19735 }, { "epoch": 3.220246349620687, "grad_norm": 0.027252180501818657, "learning_rate": 2.5512436411836538e-05, "loss": 0.0924, "num_input_tokens_seen": 38003840, "step": 19740 }, { "epoch": 3.221062076841504, "grad_norm": 0.04661443084478378, "learning_rate": 2.5502620958136443e-05, "loss": 0.0054, "num_input_tokens_seen": 38014336, "step": 19745 }, { "epoch": 3.2218778040623217, "grad_norm": 0.09895910322666168, "learning_rate": 2.5492805426925874e-05, "loss": 0.0275, "num_input_tokens_seen": 38024048, "step": 19750 }, { "epoch": 3.222693531283139, "grad_norm": 0.5572999119758606, "learning_rate": 2.5482989819718523e-05, "loss": 0.0026, "num_input_tokens_seen": 38033440, "step": 19755 }, { "epoch": 3.2235092585039564, "grad_norm": 0.708504855632782, "learning_rate": 2.5473174138028065e-05, "loss": 0.1551, "num_input_tokens_seen": 38042592, "step": 19760 }, { "epoch": 3.2243249857247736, "grad_norm": 1.7377018928527832, "learning_rate": 2.5463358383368212e-05, "loss": 0.0047, "num_input_tokens_seen": 38051200, "step": 19765 }, { "epoch": 3.225140712945591, "grad_norm": 17.776660919189453, "learning_rate": 2.545354255725267e-05, "loss": 0.0259, "num_input_tokens_seen": 38059200, "step": 19770 }, { "epoch": 3.2259564401664083, "grad_norm": 0.014570559374988079, "learning_rate": 2.5443726661195165e-05, "loss": 0.1374, "num_input_tokens_seen": 38068848, "step": 19775 }, { "epoch": 3.226772167387226, "grad_norm": 0.050948455929756165, "learning_rate": 2.543391069670944e-05, "loss": 0.0451, "num_input_tokens_seen": 38079632, "step": 19780 }, { "epoch": 3.227587894608043, "grad_norm": 0.017693478614091873, "learning_rate": 2.5424094665309228e-05, "loss": 0.0276, "num_input_tokens_seen": 38089344, "step": 19785 }, { "epoch": 3.2284036218288605, "grad_norm": 28.662925720214844, "learning_rate": 2.5414278568508292e-05, "loss": 0.1633, "num_input_tokens_seen": 38099200, "step": 19790 }, { "epoch": 3.2292193490496777, "grad_norm": 0.012523802928626537, "learning_rate": 2.540446240782039e-05, "loss": 0.0024, "num_input_tokens_seen": 38108160, "step": 19795 }, { "epoch": 3.2300350762704952, "grad_norm": 42.72515869140625, "learning_rate": 2.5394646184759307e-05, "loss": 0.0134, "num_input_tokens_seen": 38118784, "step": 19800 }, { "epoch": 3.2300350762704952, "eval_loss": 0.232109934091568, "eval_runtime": 80.7152, "eval_samples_per_second": 33.761, "eval_steps_per_second": 16.887, "num_input_tokens_seen": 38118784, "step": 19800 }, { "epoch": 3.2308508034913124, "grad_norm": 0.01731063425540924, "learning_rate": 2.538482990083882e-05, "loss": 0.1011, "num_input_tokens_seen": 38127328, "step": 19805 }, { "epoch": 3.23166653071213, "grad_norm": 0.04519623517990112, "learning_rate": 2.5375013557572725e-05, "loss": 0.0369, "num_input_tokens_seen": 38136864, "step": 19810 }, { "epoch": 3.232482257932947, "grad_norm": 5.700882911682129, "learning_rate": 2.536519715647483e-05, "loss": 0.0826, "num_input_tokens_seen": 38146304, "step": 19815 }, { "epoch": 3.2332979851537647, "grad_norm": 0.17053669691085815, "learning_rate": 2.535538069905894e-05, "loss": 0.0892, "num_input_tokens_seen": 38155520, "step": 19820 }, { "epoch": 3.234113712374582, "grad_norm": 0.2600462734699249, "learning_rate": 2.534556418683888e-05, "loss": 0.0021, "num_input_tokens_seen": 38164896, "step": 19825 }, { "epoch": 3.2349294395953994, "grad_norm": 0.05564652010798454, "learning_rate": 2.5335747621328486e-05, "loss": 0.0013, "num_input_tokens_seen": 38174736, "step": 19830 }, { "epoch": 3.2357451668162165, "grad_norm": 5.512970924377441, "learning_rate": 2.5325931004041586e-05, "loss": 0.1307, "num_input_tokens_seen": 38184240, "step": 19835 }, { "epoch": 3.236560894037034, "grad_norm": 0.0037050098180770874, "learning_rate": 2.5316114336492032e-05, "loss": 0.0134, "num_input_tokens_seen": 38194560, "step": 19840 }, { "epoch": 3.237376621257851, "grad_norm": 0.31292101740837097, "learning_rate": 2.530629762019367e-05, "loss": 0.1344, "num_input_tokens_seen": 38204752, "step": 19845 }, { "epoch": 3.2381923484786688, "grad_norm": 0.32200610637664795, "learning_rate": 2.5296480856660364e-05, "loss": 0.0016, "num_input_tokens_seen": 38214880, "step": 19850 }, { "epoch": 3.239008075699486, "grad_norm": 0.046802617609500885, "learning_rate": 2.528666404740599e-05, "loss": 0.1216, "num_input_tokens_seen": 38224352, "step": 19855 }, { "epoch": 3.2398238029203035, "grad_norm": 0.08920887112617493, "learning_rate": 2.527684719394442e-05, "loss": 0.0031, "num_input_tokens_seen": 38232208, "step": 19860 }, { "epoch": 3.2406395301411206, "grad_norm": 11.138998031616211, "learning_rate": 2.526703029778953e-05, "loss": 0.1519, "num_input_tokens_seen": 38242160, "step": 19865 }, { "epoch": 3.241455257361938, "grad_norm": 0.07322834432125092, "learning_rate": 2.5257213360455208e-05, "loss": 0.0009, "num_input_tokens_seen": 38251952, "step": 19870 }, { "epoch": 3.2422709845827553, "grad_norm": 9.820852279663086, "learning_rate": 2.5247396383455353e-05, "loss": 0.0531, "num_input_tokens_seen": 38262272, "step": 19875 }, { "epoch": 3.243086711803573, "grad_norm": 0.07307173311710358, "learning_rate": 2.523757936830387e-05, "loss": 0.0248, "num_input_tokens_seen": 38271952, "step": 19880 }, { "epoch": 3.2439024390243905, "grad_norm": 13.061152458190918, "learning_rate": 2.5227762316514662e-05, "loss": 0.1279, "num_input_tokens_seen": 38282528, "step": 19885 }, { "epoch": 3.2447181662452076, "grad_norm": 0.09940548241138458, "learning_rate": 2.5217945229601648e-05, "loss": 0.0568, "num_input_tokens_seen": 38291952, "step": 19890 }, { "epoch": 3.245533893466025, "grad_norm": 0.026713307946920395, "learning_rate": 2.5208128109078738e-05, "loss": 0.0743, "num_input_tokens_seen": 38301920, "step": 19895 }, { "epoch": 3.2463496206868423, "grad_norm": 0.045057184994220734, "learning_rate": 2.5198310956459853e-05, "loss": 0.0024, "num_input_tokens_seen": 38312128, "step": 19900 }, { "epoch": 3.24716534790766, "grad_norm": 0.09078305214643478, "learning_rate": 2.518849377325893e-05, "loss": 0.091, "num_input_tokens_seen": 38321024, "step": 19905 }, { "epoch": 3.247981075128477, "grad_norm": 4.4382710456848145, "learning_rate": 2.51786765609899e-05, "loss": 0.0951, "num_input_tokens_seen": 38331520, "step": 19910 }, { "epoch": 3.2487968023492946, "grad_norm": 0.015991900116205215, "learning_rate": 2.5168859321166694e-05, "loss": 0.0157, "num_input_tokens_seen": 38341008, "step": 19915 }, { "epoch": 3.2496125295701117, "grad_norm": 0.017502445727586746, "learning_rate": 2.515904205530326e-05, "loss": 0.0807, "num_input_tokens_seen": 38350880, "step": 19920 }, { "epoch": 3.2504282567909293, "grad_norm": 0.08908609300851822, "learning_rate": 2.514922476491355e-05, "loss": 0.0664, "num_input_tokens_seen": 38359920, "step": 19925 }, { "epoch": 3.2512439840117464, "grad_norm": 0.005584254395216703, "learning_rate": 2.51394074515115e-05, "loss": 0.0018, "num_input_tokens_seen": 38369840, "step": 19930 }, { "epoch": 3.252059711232564, "grad_norm": 0.03899528086185455, "learning_rate": 2.5129590116611067e-05, "loss": 0.0173, "num_input_tokens_seen": 38380224, "step": 19935 }, { "epoch": 3.252875438453381, "grad_norm": 0.0151018425822258, "learning_rate": 2.5119772761726212e-05, "loss": 0.0619, "num_input_tokens_seen": 38389904, "step": 19940 }, { "epoch": 3.2536911656741987, "grad_norm": 6.000811576843262, "learning_rate": 2.5109955388370893e-05, "loss": 0.0054, "num_input_tokens_seen": 38400032, "step": 19945 }, { "epoch": 3.254506892895016, "grad_norm": 4.729984283447266, "learning_rate": 2.510013799805907e-05, "loss": 0.0597, "num_input_tokens_seen": 38408544, "step": 19950 }, { "epoch": 3.2553226201158334, "grad_norm": 0.12379870563745499, "learning_rate": 2.5090320592304706e-05, "loss": 0.0013, "num_input_tokens_seen": 38417312, "step": 19955 }, { "epoch": 3.2561383473366505, "grad_norm": 0.032778285443782806, "learning_rate": 2.5080503172621777e-05, "loss": 0.007, "num_input_tokens_seen": 38426416, "step": 19960 }, { "epoch": 3.256954074557468, "grad_norm": 0.021184643730521202, "learning_rate": 2.5070685740524246e-05, "loss": 0.0689, "num_input_tokens_seen": 38437264, "step": 19965 }, { "epoch": 3.2577698017782852, "grad_norm": 0.11225549876689911, "learning_rate": 2.5060868297526084e-05, "loss": 0.0021, "num_input_tokens_seen": 38446576, "step": 19970 }, { "epoch": 3.258585528999103, "grad_norm": 0.6013143062591553, "learning_rate": 2.5051050845141267e-05, "loss": 0.1109, "num_input_tokens_seen": 38457056, "step": 19975 }, { "epoch": 3.25940125621992, "grad_norm": 0.005616642534732819, "learning_rate": 2.5041233384883765e-05, "loss": 0.1549, "num_input_tokens_seen": 38465808, "step": 19980 }, { "epoch": 3.2602169834407375, "grad_norm": 0.12498565018177032, "learning_rate": 2.5031415918267564e-05, "loss": 0.0052, "num_input_tokens_seen": 38474592, "step": 19985 }, { "epoch": 3.2610327106615546, "grad_norm": 14.495500564575195, "learning_rate": 2.5021598446806626e-05, "loss": 0.0738, "num_input_tokens_seen": 38485120, "step": 19990 }, { "epoch": 3.261848437882372, "grad_norm": 0.03399031609296799, "learning_rate": 2.5011780972014937e-05, "loss": 0.0011, "num_input_tokens_seen": 38493408, "step": 19995 }, { "epoch": 3.2626641651031894, "grad_norm": 0.1805219203233719, "learning_rate": 2.5001963495406478e-05, "loss": 0.0028, "num_input_tokens_seen": 38503392, "step": 20000 }, { "epoch": 3.2626641651031894, "eval_loss": 0.21684901416301727, "eval_runtime": 80.7268, "eval_samples_per_second": 33.756, "eval_steps_per_second": 16.884, "num_input_tokens_seen": 38503392, "step": 20000 }, { "epoch": 3.263479892324007, "grad_norm": 0.17801399528980255, "learning_rate": 2.499214601849522e-05, "loss": 0.0093, "num_input_tokens_seen": 38512656, "step": 20005 }, { "epoch": 3.264295619544824, "grad_norm": 6.2993083000183105, "learning_rate": 2.4982328542795148e-05, "loss": 0.1169, "num_input_tokens_seen": 38521952, "step": 20010 }, { "epoch": 3.2651113467656416, "grad_norm": 0.0722586140036583, "learning_rate": 2.497251106982024e-05, "loss": 0.0025, "num_input_tokens_seen": 38532384, "step": 20015 }, { "epoch": 3.2659270739864588, "grad_norm": 0.23055577278137207, "learning_rate": 2.4962693601084458e-05, "loss": 0.0041, "num_input_tokens_seen": 38540800, "step": 20020 }, { "epoch": 3.2667428012072763, "grad_norm": 0.044252194464206696, "learning_rate": 2.4952876138101794e-05, "loss": 0.1156, "num_input_tokens_seen": 38550448, "step": 20025 }, { "epoch": 3.2675585284280935, "grad_norm": 11.809980392456055, "learning_rate": 2.4943058682386233e-05, "loss": 0.1904, "num_input_tokens_seen": 38560144, "step": 20030 }, { "epoch": 3.268374255648911, "grad_norm": 4.23405647277832, "learning_rate": 2.493324123545173e-05, "loss": 0.1544, "num_input_tokens_seen": 38569728, "step": 20035 }, { "epoch": 3.269189982869728, "grad_norm": 0.3212803900241852, "learning_rate": 2.4923423798812272e-05, "loss": 0.0032, "num_input_tokens_seen": 38580544, "step": 20040 }, { "epoch": 3.2700057100905457, "grad_norm": 5.281126499176025, "learning_rate": 2.4913606373981825e-05, "loss": 0.1796, "num_input_tokens_seen": 38590880, "step": 20045 }, { "epoch": 3.270821437311363, "grad_norm": 0.05637357011437416, "learning_rate": 2.4903788962474357e-05, "loss": 0.0034, "num_input_tokens_seen": 38599328, "step": 20050 }, { "epoch": 3.2716371645321805, "grad_norm": 0.03056877665221691, "learning_rate": 2.489397156580385e-05, "loss": 0.0067, "num_input_tokens_seen": 38607440, "step": 20055 }, { "epoch": 3.272452891752998, "grad_norm": 0.10729481279850006, "learning_rate": 2.4884154185484246e-05, "loss": 0.0579, "num_input_tokens_seen": 38617024, "step": 20060 }, { "epoch": 3.273268618973815, "grad_norm": 0.14060348272323608, "learning_rate": 2.4874336823029526e-05, "loss": 0.0025, "num_input_tokens_seen": 38626576, "step": 20065 }, { "epoch": 3.2740843461946323, "grad_norm": 4.989144325256348, "learning_rate": 2.4864519479953656e-05, "loss": 0.1053, "num_input_tokens_seen": 38635760, "step": 20070 }, { "epoch": 3.27490007341545, "grad_norm": 0.03696523606777191, "learning_rate": 2.485470215777058e-05, "loss": 0.0019, "num_input_tokens_seen": 38643952, "step": 20075 }, { "epoch": 3.2757158006362674, "grad_norm": 0.059735171496868134, "learning_rate": 2.4844884857994258e-05, "loss": 0.0159, "num_input_tokens_seen": 38654272, "step": 20080 }, { "epoch": 3.2765315278570846, "grad_norm": 0.05516931787133217, "learning_rate": 2.4835067582138638e-05, "loss": 0.0643, "num_input_tokens_seen": 38664336, "step": 20085 }, { "epoch": 3.2773472550779017, "grad_norm": 0.060698382556438446, "learning_rate": 2.4825250331717666e-05, "loss": 0.1383, "num_input_tokens_seen": 38674128, "step": 20090 }, { "epoch": 3.2781629822987193, "grad_norm": 0.032705117017030716, "learning_rate": 2.4815433108245298e-05, "loss": 0.0072, "num_input_tokens_seen": 38684768, "step": 20095 }, { "epoch": 3.278978709519537, "grad_norm": 0.11144495755434036, "learning_rate": 2.4805615913235456e-05, "loss": 0.0584, "num_input_tokens_seen": 38693904, "step": 20100 }, { "epoch": 3.279794436740354, "grad_norm": 0.0488751120865345, "learning_rate": 2.479579874820208e-05, "loss": 0.0828, "num_input_tokens_seen": 38704752, "step": 20105 }, { "epoch": 3.2806101639611716, "grad_norm": 4.783389091491699, "learning_rate": 2.4785981614659115e-05, "loss": 0.0588, "num_input_tokens_seen": 38714064, "step": 20110 }, { "epoch": 3.2814258911819887, "grad_norm": 0.1475493311882019, "learning_rate": 2.477616451412047e-05, "loss": 0.0603, "num_input_tokens_seen": 38724768, "step": 20115 }, { "epoch": 3.2822416184028063, "grad_norm": 0.05011129379272461, "learning_rate": 2.476634744810007e-05, "loss": 0.0531, "num_input_tokens_seen": 38733680, "step": 20120 }, { "epoch": 3.2830573456236234, "grad_norm": 3.319481372833252, "learning_rate": 2.475653041811183e-05, "loss": 0.0785, "num_input_tokens_seen": 38743424, "step": 20125 }, { "epoch": 3.283873072844441, "grad_norm": 0.02022423781454563, "learning_rate": 2.4746713425669652e-05, "loss": 0.0005, "num_input_tokens_seen": 38753760, "step": 20130 }, { "epoch": 3.284688800065258, "grad_norm": 0.03907870873808861, "learning_rate": 2.4736896472287458e-05, "loss": 0.0034, "num_input_tokens_seen": 38762528, "step": 20135 }, { "epoch": 3.2855045272860757, "grad_norm": 0.03596453368663788, "learning_rate": 2.4727079559479124e-05, "loss": 0.0017, "num_input_tokens_seen": 38771472, "step": 20140 }, { "epoch": 3.286320254506893, "grad_norm": 0.031931035220623016, "learning_rate": 2.4717262688758557e-05, "loss": 0.1659, "num_input_tokens_seen": 38781104, "step": 20145 }, { "epoch": 3.2871359817277104, "grad_norm": 0.04566275328397751, "learning_rate": 2.4707445861639637e-05, "loss": 0.0023, "num_input_tokens_seen": 38790208, "step": 20150 }, { "epoch": 3.2879517089485275, "grad_norm": 0.15743692219257355, "learning_rate": 2.4697629079636244e-05, "loss": 0.149, "num_input_tokens_seen": 38799312, "step": 20155 }, { "epoch": 3.288767436169345, "grad_norm": 4.977086544036865, "learning_rate": 2.4687812344262244e-05, "loss": 0.0131, "num_input_tokens_seen": 38808656, "step": 20160 }, { "epoch": 3.289583163390162, "grad_norm": 8.870923042297363, "learning_rate": 2.46779956570315e-05, "loss": 0.1419, "num_input_tokens_seen": 38817584, "step": 20165 }, { "epoch": 3.29039889061098, "grad_norm": 0.6305984258651733, "learning_rate": 2.466817901945787e-05, "loss": 0.0646, "num_input_tokens_seen": 38827248, "step": 20170 }, { "epoch": 3.291214617831797, "grad_norm": 4.422989845275879, "learning_rate": 2.4658362433055217e-05, "loss": 0.0073, "num_input_tokens_seen": 38836720, "step": 20175 }, { "epoch": 3.2920303450526145, "grad_norm": 4.774616718292236, "learning_rate": 2.4648545899337356e-05, "loss": 0.1365, "num_input_tokens_seen": 38845888, "step": 20180 }, { "epoch": 3.2928460722734316, "grad_norm": 0.03751023858785629, "learning_rate": 2.4638729419818143e-05, "loss": 0.003, "num_input_tokens_seen": 38855408, "step": 20185 }, { "epoch": 3.293661799494249, "grad_norm": 0.020740805193781853, "learning_rate": 2.46289129960114e-05, "loss": 0.0039, "num_input_tokens_seen": 38865744, "step": 20190 }, { "epoch": 3.2944775267150663, "grad_norm": 0.42959362268447876, "learning_rate": 2.4619096629430924e-05, "loss": 0.0537, "num_input_tokens_seen": 38876688, "step": 20195 }, { "epoch": 3.295293253935884, "grad_norm": 0.025738338008522987, "learning_rate": 2.4609280321590543e-05, "loss": 0.2017, "num_input_tokens_seen": 38885696, "step": 20200 }, { "epoch": 3.295293253935884, "eval_loss": 0.20447805523872375, "eval_runtime": 80.8227, "eval_samples_per_second": 33.716, "eval_steps_per_second": 16.864, "num_input_tokens_seen": 38885696, "step": 20200 }, { "epoch": 3.296108981156701, "grad_norm": 0.07932816445827484, "learning_rate": 2.4599464074004037e-05, "loss": 0.0513, "num_input_tokens_seen": 38895952, "step": 20205 }, { "epoch": 3.2969247083775186, "grad_norm": 0.07365990430116653, "learning_rate": 2.4589647888185204e-05, "loss": 0.0042, "num_input_tokens_seen": 38905152, "step": 20210 }, { "epoch": 3.2977404355983357, "grad_norm": 0.13862694799900055, "learning_rate": 2.4579831765647836e-05, "loss": 0.0032, "num_input_tokens_seen": 38915424, "step": 20215 }, { "epoch": 3.2985561628191533, "grad_norm": 0.05618876591324806, "learning_rate": 2.4570015707905676e-05, "loss": 0.0027, "num_input_tokens_seen": 38923648, "step": 20220 }, { "epoch": 3.2993718900399704, "grad_norm": 0.21115726232528687, "learning_rate": 2.4560199716472508e-05, "loss": 0.069, "num_input_tokens_seen": 38933072, "step": 20225 }, { "epoch": 3.300187617260788, "grad_norm": 3.8173179626464844, "learning_rate": 2.455038379286207e-05, "loss": 0.1926, "num_input_tokens_seen": 38943952, "step": 20230 }, { "epoch": 3.3010033444816056, "grad_norm": 11.747687339782715, "learning_rate": 2.4540567938588095e-05, "loss": 0.1319, "num_input_tokens_seen": 38954800, "step": 20235 }, { "epoch": 3.3018190717024227, "grad_norm": 0.17614533007144928, "learning_rate": 2.4530752155164328e-05, "loss": 0.0041, "num_input_tokens_seen": 38964880, "step": 20240 }, { "epoch": 3.30263479892324, "grad_norm": 0.09034328162670135, "learning_rate": 2.4520936444104463e-05, "loss": 0.0016, "num_input_tokens_seen": 38974064, "step": 20245 }, { "epoch": 3.3034505261440574, "grad_norm": 0.06904762983322144, "learning_rate": 2.4511120806922218e-05, "loss": 0.002, "num_input_tokens_seen": 38984752, "step": 20250 }, { "epoch": 3.304266253364875, "grad_norm": 0.01904343068599701, "learning_rate": 2.45013052451313e-05, "loss": 0.0128, "num_input_tokens_seen": 38995456, "step": 20255 }, { "epoch": 3.305081980585692, "grad_norm": 0.03267381340265274, "learning_rate": 2.4491489760245376e-05, "loss": 0.1157, "num_input_tokens_seen": 39003808, "step": 20260 }, { "epoch": 3.3058977078065093, "grad_norm": 0.05451156198978424, "learning_rate": 2.4481674353778115e-05, "loss": 0.0314, "num_input_tokens_seen": 39012992, "step": 20265 }, { "epoch": 3.306713435027327, "grad_norm": 3.930540084838867, "learning_rate": 2.447185902724319e-05, "loss": 0.0999, "num_input_tokens_seen": 39022240, "step": 20270 }, { "epoch": 3.3075291622481444, "grad_norm": 0.06211436912417412, "learning_rate": 2.4462043782154233e-05, "loss": 0.1569, "num_input_tokens_seen": 39031920, "step": 20275 }, { "epoch": 3.3083448894689615, "grad_norm": 4.841126441955566, "learning_rate": 2.4452228620024895e-05, "loss": 0.0918, "num_input_tokens_seen": 39041632, "step": 20280 }, { "epoch": 3.309160616689779, "grad_norm": 3.5549662113189697, "learning_rate": 2.4442413542368776e-05, "loss": 0.0456, "num_input_tokens_seen": 39050848, "step": 20285 }, { "epoch": 3.3099763439105963, "grad_norm": 0.03240911290049553, "learning_rate": 2.4432598550699502e-05, "loss": 0.0028, "num_input_tokens_seen": 39061184, "step": 20290 }, { "epoch": 3.310792071131414, "grad_norm": 0.12160733342170715, "learning_rate": 2.4422783646530663e-05, "loss": 0.0303, "num_input_tokens_seen": 39070816, "step": 20295 }, { "epoch": 3.311607798352231, "grad_norm": 0.06927158683538437, "learning_rate": 2.441296883137584e-05, "loss": 0.1267, "num_input_tokens_seen": 39081312, "step": 20300 }, { "epoch": 3.3124235255730485, "grad_norm": 0.16603191196918488, "learning_rate": 2.4403154106748592e-05, "loss": 0.029, "num_input_tokens_seen": 39090192, "step": 20305 }, { "epoch": 3.3132392527938657, "grad_norm": 0.05370715633034706, "learning_rate": 2.4393339474162494e-05, "loss": 0.0041, "num_input_tokens_seen": 39099488, "step": 20310 }, { "epoch": 3.3140549800146832, "grad_norm": 7.491815090179443, "learning_rate": 2.4383524935131062e-05, "loss": 0.0566, "num_input_tokens_seen": 39109104, "step": 20315 }, { "epoch": 3.3148707072355004, "grad_norm": 0.08193619549274445, "learning_rate": 2.437371049116784e-05, "loss": 0.0094, "num_input_tokens_seen": 39119088, "step": 20320 }, { "epoch": 3.315686434456318, "grad_norm": 0.051253046840429306, "learning_rate": 2.436389614378632e-05, "loss": 0.003, "num_input_tokens_seen": 39129040, "step": 20325 }, { "epoch": 3.316502161677135, "grad_norm": 0.11274677515029907, "learning_rate": 2.435408189450002e-05, "loss": 0.0198, "num_input_tokens_seen": 39138416, "step": 20330 }, { "epoch": 3.3173178888979526, "grad_norm": 0.02017173357307911, "learning_rate": 2.4344267744822406e-05, "loss": 0.0015, "num_input_tokens_seen": 39148384, "step": 20335 }, { "epoch": 3.31813361611877, "grad_norm": 6.636771202087402, "learning_rate": 2.4334453696266944e-05, "loss": 0.0386, "num_input_tokens_seen": 39157840, "step": 20340 }, { "epoch": 3.3189493433395874, "grad_norm": 0.022440392524003983, "learning_rate": 2.432463975034708e-05, "loss": 0.1552, "num_input_tokens_seen": 39167728, "step": 20345 }, { "epoch": 3.3197650705604045, "grad_norm": 0.03196731582283974, "learning_rate": 2.4314825908576265e-05, "loss": 0.1798, "num_input_tokens_seen": 39176752, "step": 20350 }, { "epoch": 3.320580797781222, "grad_norm": 0.046285662800073624, "learning_rate": 2.4305012172467897e-05, "loss": 0.0566, "num_input_tokens_seen": 39186480, "step": 20355 }, { "epoch": 3.321396525002039, "grad_norm": 0.11770118027925491, "learning_rate": 2.4295198543535393e-05, "loss": 0.0897, "num_input_tokens_seen": 39196512, "step": 20360 }, { "epoch": 3.3222122522228568, "grad_norm": 6.460652828216553, "learning_rate": 2.4285385023292124e-05, "loss": 0.0455, "num_input_tokens_seen": 39205312, "step": 20365 }, { "epoch": 3.323027979443674, "grad_norm": 0.11565186083316803, "learning_rate": 2.427557161325147e-05, "loss": 0.0424, "num_input_tokens_seen": 39214880, "step": 20370 }, { "epoch": 3.3238437066644915, "grad_norm": 2.218986988067627, "learning_rate": 2.4265758314926778e-05, "loss": 0.0708, "num_input_tokens_seen": 39225520, "step": 20375 }, { "epoch": 3.3246594338853086, "grad_norm": 0.037873901426792145, "learning_rate": 2.4255945129831373e-05, "loss": 0.0028, "num_input_tokens_seen": 39235072, "step": 20380 }, { "epoch": 3.325475161106126, "grad_norm": 0.049255404621362686, "learning_rate": 2.4246132059478578e-05, "loss": 0.0711, "num_input_tokens_seen": 39244592, "step": 20385 }, { "epoch": 3.3262908883269433, "grad_norm": 0.057214248925447464, "learning_rate": 2.4236319105381706e-05, "loss": 0.0032, "num_input_tokens_seen": 39252432, "step": 20390 }, { "epoch": 3.327106615547761, "grad_norm": 2.068699598312378, "learning_rate": 2.422650626905401e-05, "loss": 0.1068, "num_input_tokens_seen": 39261520, "step": 20395 }, { "epoch": 3.327922342768578, "grad_norm": 0.0106091583147645, "learning_rate": 2.4216693552008785e-05, "loss": 0.0441, "num_input_tokens_seen": 39270320, "step": 20400 }, { "epoch": 3.327922342768578, "eval_loss": 0.20466327667236328, "eval_runtime": 80.8771, "eval_samples_per_second": 33.693, "eval_steps_per_second": 16.853, "num_input_tokens_seen": 39270320, "step": 20400 }, { "epoch": 3.3287380699893956, "grad_norm": 0.0658707320690155, "learning_rate": 2.4206880955759247e-05, "loss": 0.0574, "num_input_tokens_seen": 39280288, "step": 20405 }, { "epoch": 3.3295537972102127, "grad_norm": 17.281679153442383, "learning_rate": 2.419706848181863e-05, "loss": 0.073, "num_input_tokens_seen": 39290240, "step": 20410 }, { "epoch": 3.3303695244310303, "grad_norm": 0.06242569908499718, "learning_rate": 2.4187256131700153e-05, "loss": 0.002, "num_input_tokens_seen": 39299616, "step": 20415 }, { "epoch": 3.3311852516518474, "grad_norm": 0.14030005037784576, "learning_rate": 2.4177443906916985e-05, "loss": 0.0677, "num_input_tokens_seen": 39309920, "step": 20420 }, { "epoch": 3.332000978872665, "grad_norm": 0.1470736712217331, "learning_rate": 2.4167631808982303e-05, "loss": 0.0704, "num_input_tokens_seen": 39319824, "step": 20425 }, { "epoch": 3.3328167060934826, "grad_norm": 0.19230392575263977, "learning_rate": 2.4157819839409264e-05, "loss": 0.0022, "num_input_tokens_seen": 39330032, "step": 20430 }, { "epoch": 3.3336324333142997, "grad_norm": 0.13322608172893524, "learning_rate": 2.414800799971098e-05, "loss": 0.002, "num_input_tokens_seen": 39339408, "step": 20435 }, { "epoch": 3.334448160535117, "grad_norm": 0.5549549460411072, "learning_rate": 2.4138196291400582e-05, "loss": 0.0209, "num_input_tokens_seen": 39349616, "step": 20440 }, { "epoch": 3.3352638877559344, "grad_norm": 0.008777991868555546, "learning_rate": 2.412838471599114e-05, "loss": 0.0022, "num_input_tokens_seen": 39359312, "step": 20445 }, { "epoch": 3.336079614976752, "grad_norm": 0.03853563964366913, "learning_rate": 2.411857327499572e-05, "loss": 0.0021, "num_input_tokens_seen": 39368080, "step": 20450 }, { "epoch": 3.336895342197569, "grad_norm": 0.08951739221811295, "learning_rate": 2.410876196992739e-05, "loss": 0.0013, "num_input_tokens_seen": 39377952, "step": 20455 }, { "epoch": 3.3377110694183862, "grad_norm": 0.013466717675328255, "learning_rate": 2.4098950802299156e-05, "loss": 0.0223, "num_input_tokens_seen": 39388064, "step": 20460 }, { "epoch": 3.338526796639204, "grad_norm": 4.8084611892700195, "learning_rate": 2.4089139773624027e-05, "loss": 0.0875, "num_input_tokens_seen": 39398704, "step": 20465 }, { "epoch": 3.3393425238600214, "grad_norm": 0.20767280459403992, "learning_rate": 2.4079328885415007e-05, "loss": 0.0742, "num_input_tokens_seen": 39407776, "step": 20470 }, { "epoch": 3.3401582510808385, "grad_norm": 0.3430621325969696, "learning_rate": 2.4069518139185036e-05, "loss": 0.0729, "num_input_tokens_seen": 39417776, "step": 20475 }, { "epoch": 3.340973978301656, "grad_norm": 0.3614760637283325, "learning_rate": 2.405970753644706e-05, "loss": 0.0501, "num_input_tokens_seen": 39427776, "step": 20480 }, { "epoch": 3.3417897055224732, "grad_norm": 1.326314091682434, "learning_rate": 2.4049897078714e-05, "loss": 0.0385, "num_input_tokens_seen": 39437584, "step": 20485 }, { "epoch": 3.342605432743291, "grad_norm": 0.012279320508241653, "learning_rate": 2.404008676749874e-05, "loss": 0.0484, "num_input_tokens_seen": 39448064, "step": 20490 }, { "epoch": 3.343421159964108, "grad_norm": 0.014506399631500244, "learning_rate": 2.403027660431418e-05, "loss": 0.0028, "num_input_tokens_seen": 39457712, "step": 20495 }, { "epoch": 3.3442368871849255, "grad_norm": 0.017034417018294334, "learning_rate": 2.402046659067314e-05, "loss": 0.0621, "num_input_tokens_seen": 39467744, "step": 20500 }, { "epoch": 3.3450526144057426, "grad_norm": 1.2526535987854004, "learning_rate": 2.401065672808847e-05, "loss": 0.0955, "num_input_tokens_seen": 39477152, "step": 20505 }, { "epoch": 3.34586834162656, "grad_norm": 0.11787302792072296, "learning_rate": 2.400084701807296e-05, "loss": 0.0006, "num_input_tokens_seen": 39488576, "step": 20510 }, { "epoch": 3.3466840688473773, "grad_norm": 0.11705147475004196, "learning_rate": 2.39910374621394e-05, "loss": 0.1154, "num_input_tokens_seen": 39498416, "step": 20515 }, { "epoch": 3.347499796068195, "grad_norm": 48.52117156982422, "learning_rate": 2.3981228061800544e-05, "loss": 0.1114, "num_input_tokens_seen": 39507712, "step": 20520 }, { "epoch": 3.348315523289012, "grad_norm": 0.04622736945748329, "learning_rate": 2.3971418818569115e-05, "loss": 0.0013, "num_input_tokens_seen": 39517184, "step": 20525 }, { "epoch": 3.3491312505098296, "grad_norm": 0.051734503358602524, "learning_rate": 2.3961609733957832e-05, "loss": 0.0625, "num_input_tokens_seen": 39526576, "step": 20530 }, { "epoch": 3.3499469777306468, "grad_norm": 3.8803582191467285, "learning_rate": 2.395180080947939e-05, "loss": 0.0942, "num_input_tokens_seen": 39536288, "step": 20535 }, { "epoch": 3.3507627049514643, "grad_norm": 0.23206037282943726, "learning_rate": 2.394199204664642e-05, "loss": 0.0693, "num_input_tokens_seen": 39546880, "step": 20540 }, { "epoch": 3.3515784321722815, "grad_norm": 0.02141346037387848, "learning_rate": 2.3932183446971583e-05, "loss": 0.0568, "num_input_tokens_seen": 39556448, "step": 20545 }, { "epoch": 3.352394159393099, "grad_norm": 10.474163055419922, "learning_rate": 2.3922375011967473e-05, "loss": 0.1404, "num_input_tokens_seen": 39566736, "step": 20550 }, { "epoch": 3.353209886613916, "grad_norm": 0.041771359741687775, "learning_rate": 2.3912566743146676e-05, "loss": 0.0231, "num_input_tokens_seen": 39577472, "step": 20555 }, { "epoch": 3.3540256138347337, "grad_norm": 0.029089482501149178, "learning_rate": 2.390275864202176e-05, "loss": 0.0529, "num_input_tokens_seen": 39586736, "step": 20560 }, { "epoch": 3.354841341055551, "grad_norm": 0.08556690067052841, "learning_rate": 2.3892950710105243e-05, "loss": 0.1349, "num_input_tokens_seen": 39596640, "step": 20565 }, { "epoch": 3.3556570682763684, "grad_norm": 0.10795444250106812, "learning_rate": 2.3883142948909635e-05, "loss": 0.128, "num_input_tokens_seen": 39606800, "step": 20570 }, { "epoch": 3.3564727954971856, "grad_norm": 1.5950967073440552, "learning_rate": 2.3873335359947433e-05, "loss": 0.0448, "num_input_tokens_seen": 39615808, "step": 20575 }, { "epoch": 3.357288522718003, "grad_norm": 0.09208367764949799, "learning_rate": 2.3863527944731066e-05, "loss": 0.0441, "num_input_tokens_seen": 39625984, "step": 20580 }, { "epoch": 3.3581042499388203, "grad_norm": 0.09142982959747314, "learning_rate": 2.385372070477298e-05, "loss": 0.0158, "num_input_tokens_seen": 39636832, "step": 20585 }, { "epoch": 3.358919977159638, "grad_norm": 0.351926326751709, "learning_rate": 2.384391364158556e-05, "loss": 0.0521, "num_input_tokens_seen": 39646416, "step": 20590 }, { "epoch": 3.359735704380455, "grad_norm": 0.05262964591383934, "learning_rate": 2.3834106756681185e-05, "loss": 0.0065, "num_input_tokens_seen": 39656128, "step": 20595 }, { "epoch": 3.3605514316012726, "grad_norm": 0.008976149372756481, "learning_rate": 2.3824300051572206e-05, "loss": 0.0007, "num_input_tokens_seen": 39665472, "step": 20600 }, { "epoch": 3.3605514316012726, "eval_loss": 0.213675394654274, "eval_runtime": 80.7311, "eval_samples_per_second": 33.754, "eval_steps_per_second": 16.883, "num_input_tokens_seen": 39665472, "step": 20600 }, { "epoch": 3.36136715882209, "grad_norm": 0.13205792009830475, "learning_rate": 2.3814493527770923e-05, "loss": 0.0064, "num_input_tokens_seen": 39674736, "step": 20605 }, { "epoch": 3.3621828860429073, "grad_norm": 0.12561243772506714, "learning_rate": 2.3804687186789637e-05, "loss": 0.0882, "num_input_tokens_seen": 39684928, "step": 20610 }, { "epoch": 3.3629986132637244, "grad_norm": 0.05673838034272194, "learning_rate": 2.379488103014062e-05, "loss": 0.0014, "num_input_tokens_seen": 39693504, "step": 20615 }, { "epoch": 3.363814340484542, "grad_norm": 0.055429428815841675, "learning_rate": 2.3785075059336086e-05, "loss": 0.0074, "num_input_tokens_seen": 39703632, "step": 20620 }, { "epoch": 3.3646300677053596, "grad_norm": 1.288129448890686, "learning_rate": 2.3775269275888248e-05, "loss": 0.0835, "num_input_tokens_seen": 39714736, "step": 20625 }, { "epoch": 3.3654457949261767, "grad_norm": 0.07330045104026794, "learning_rate": 2.3765463681309274e-05, "loss": 0.0676, "num_input_tokens_seen": 39723840, "step": 20630 }, { "epoch": 3.366261522146994, "grad_norm": 0.047891706228256226, "learning_rate": 2.3755658277111313e-05, "loss": 0.001, "num_input_tokens_seen": 39733888, "step": 20635 }, { "epoch": 3.3670772493678114, "grad_norm": 0.0494304783642292, "learning_rate": 2.374585306480649e-05, "loss": 0.1029, "num_input_tokens_seen": 39744208, "step": 20640 }, { "epoch": 3.367892976588629, "grad_norm": 0.20313802361488342, "learning_rate": 2.3736048045906877e-05, "loss": 0.1377, "num_input_tokens_seen": 39752192, "step": 20645 }, { "epoch": 3.368708703809446, "grad_norm": 7.52268648147583, "learning_rate": 2.372624322192454e-05, "loss": 0.1548, "num_input_tokens_seen": 39762288, "step": 20650 }, { "epoch": 3.3695244310302637, "grad_norm": 18.988548278808594, "learning_rate": 2.3716438594371516e-05, "loss": 0.1041, "num_input_tokens_seen": 39772256, "step": 20655 }, { "epoch": 3.370340158251081, "grad_norm": 0.04524518921971321, "learning_rate": 2.3706634164759784e-05, "loss": 0.0174, "num_input_tokens_seen": 39782128, "step": 20660 }, { "epoch": 3.3711558854718984, "grad_norm": 0.020662281662225723, "learning_rate": 2.3696829934601323e-05, "loss": 0.0013, "num_input_tokens_seen": 39791696, "step": 20665 }, { "epoch": 3.3719716126927155, "grad_norm": 1.9855762720108032, "learning_rate": 2.3687025905408053e-05, "loss": 0.0099, "num_input_tokens_seen": 39801280, "step": 20670 }, { "epoch": 3.372787339913533, "grad_norm": 0.009811599738895893, "learning_rate": 2.3677222078691886e-05, "loss": 0.0878, "num_input_tokens_seen": 39811456, "step": 20675 }, { "epoch": 3.37360306713435, "grad_norm": 0.005101102404296398, "learning_rate": 2.366741845596471e-05, "loss": 0.0653, "num_input_tokens_seen": 39820880, "step": 20680 }, { "epoch": 3.374418794355168, "grad_norm": 7.829875946044922, "learning_rate": 2.3657615038738343e-05, "loss": 0.0562, "num_input_tokens_seen": 39831504, "step": 20685 }, { "epoch": 3.375234521575985, "grad_norm": 0.21084064245224, "learning_rate": 2.3647811828524614e-05, "loss": 0.0023, "num_input_tokens_seen": 39842032, "step": 20690 }, { "epoch": 3.3760502487968025, "grad_norm": 0.06462401896715164, "learning_rate": 2.363800882683529e-05, "loss": 0.013, "num_input_tokens_seen": 39852848, "step": 20695 }, { "epoch": 3.3768659760176196, "grad_norm": 0.01769915595650673, "learning_rate": 2.3628206035182125e-05, "loss": 0.0018, "num_input_tokens_seen": 39862288, "step": 20700 }, { "epoch": 3.377681703238437, "grad_norm": 30.059499740600586, "learning_rate": 2.361840345507683e-05, "loss": 0.0253, "num_input_tokens_seen": 39872208, "step": 20705 }, { "epoch": 3.3784974304592543, "grad_norm": 0.058874815702438354, "learning_rate": 2.3608601088031073e-05, "loss": 0.1156, "num_input_tokens_seen": 39881936, "step": 20710 }, { "epoch": 3.379313157680072, "grad_norm": 0.040764641016721725, "learning_rate": 2.3598798935556516e-05, "loss": 0.0112, "num_input_tokens_seen": 39891376, "step": 20715 }, { "epoch": 3.380128884900889, "grad_norm": 0.024655137211084366, "learning_rate": 2.3588996999164784e-05, "loss": 0.0006, "num_input_tokens_seen": 39899920, "step": 20720 }, { "epoch": 3.3809446121217066, "grad_norm": 0.21816802024841309, "learning_rate": 2.3579195280367434e-05, "loss": 0.0121, "num_input_tokens_seen": 39909840, "step": 20725 }, { "epoch": 3.3817603393425237, "grad_norm": 24.410167694091797, "learning_rate": 2.356939378067603e-05, "loss": 0.0102, "num_input_tokens_seen": 39920352, "step": 20730 }, { "epoch": 3.3825760665633413, "grad_norm": 0.008033725433051586, "learning_rate": 2.3559592501602092e-05, "loss": 0.0004, "num_input_tokens_seen": 39930032, "step": 20735 }, { "epoch": 3.3833917937841584, "grad_norm": 0.029788527637720108, "learning_rate": 2.3549791444657076e-05, "loss": 0.0517, "num_input_tokens_seen": 39939104, "step": 20740 }, { "epoch": 3.384207521004976, "grad_norm": 0.2903110980987549, "learning_rate": 2.353999061135246e-05, "loss": 0.001, "num_input_tokens_seen": 39947920, "step": 20745 }, { "epoch": 3.385023248225793, "grad_norm": 0.5034065246582031, "learning_rate": 2.3530190003199626e-05, "loss": 0.0011, "num_input_tokens_seen": 39956640, "step": 20750 }, { "epoch": 3.3858389754466107, "grad_norm": 0.05904463306069374, "learning_rate": 2.3520389621709965e-05, "loss": 0.0847, "num_input_tokens_seen": 39967072, "step": 20755 }, { "epoch": 3.386654702667428, "grad_norm": 0.15477104485034943, "learning_rate": 2.351058946839483e-05, "loss": 0.0022, "num_input_tokens_seen": 39976832, "step": 20760 }, { "epoch": 3.3874704298882454, "grad_norm": 0.10751911997795105, "learning_rate": 2.350078954476551e-05, "loss": 0.0014, "num_input_tokens_seen": 39986816, "step": 20765 }, { "epoch": 3.3882861571090626, "grad_norm": 9.305068969726562, "learning_rate": 2.3490989852333272e-05, "loss": 0.1399, "num_input_tokens_seen": 39994800, "step": 20770 }, { "epoch": 3.38910188432988, "grad_norm": 0.25531333684921265, "learning_rate": 2.3481190392609377e-05, "loss": 0.0819, "num_input_tokens_seen": 40003968, "step": 20775 }, { "epoch": 3.3899176115506973, "grad_norm": 0.023949749767780304, "learning_rate": 2.3471391167105e-05, "loss": 0.0662, "num_input_tokens_seen": 40012352, "step": 20780 }, { "epoch": 3.390733338771515, "grad_norm": 0.012926201336085796, "learning_rate": 2.3461592177331325e-05, "loss": 0.1087, "num_input_tokens_seen": 40020928, "step": 20785 }, { "epoch": 3.391549065992332, "grad_norm": 0.09482630342245102, "learning_rate": 2.345179342479946e-05, "loss": 0.1017, "num_input_tokens_seen": 40031184, "step": 20790 }, { "epoch": 3.3923647932131495, "grad_norm": 0.04010402411222458, "learning_rate": 2.3441994911020503e-05, "loss": 0.0799, "num_input_tokens_seen": 40040784, "step": 20795 }, { "epoch": 3.393180520433967, "grad_norm": 0.08062505722045898, "learning_rate": 2.3432196637505522e-05, "loss": 0.0745, "num_input_tokens_seen": 40049680, "step": 20800 }, { "epoch": 3.393180520433967, "eval_loss": 0.21692736446857452, "eval_runtime": 80.8299, "eval_samples_per_second": 33.713, "eval_steps_per_second": 16.863, "num_input_tokens_seen": 40049680, "step": 20800 }, { "epoch": 3.3939962476547842, "grad_norm": 0.11864311993122101, "learning_rate": 2.3422398605765515e-05, "loss": 0.0826, "num_input_tokens_seen": 40058896, "step": 20805 }, { "epoch": 3.3948119748756014, "grad_norm": 0.04061538353562355, "learning_rate": 2.3412600817311462e-05, "loss": 0.1913, "num_input_tokens_seen": 40068944, "step": 20810 }, { "epoch": 3.395627702096419, "grad_norm": 0.08589401841163635, "learning_rate": 2.3402803273654326e-05, "loss": 0.0026, "num_input_tokens_seen": 40078144, "step": 20815 }, { "epoch": 3.3964434293172365, "grad_norm": 0.02268855646252632, "learning_rate": 2.3393005976304983e-05, "loss": 0.1525, "num_input_tokens_seen": 40087888, "step": 20820 }, { "epoch": 3.3972591565380537, "grad_norm": 0.17885352671146393, "learning_rate": 2.338320892677432e-05, "loss": 0.0029, "num_input_tokens_seen": 40098096, "step": 20825 }, { "epoch": 3.398074883758871, "grad_norm": 0.09390019625425339, "learning_rate": 2.3373412126573155e-05, "loss": 0.129, "num_input_tokens_seen": 40108816, "step": 20830 }, { "epoch": 3.3988906109796884, "grad_norm": 0.1959863156080246, "learning_rate": 2.3363615577212285e-05, "loss": 0.0027, "num_input_tokens_seen": 40117888, "step": 20835 }, { "epoch": 3.399706338200506, "grad_norm": 0.17178362607955933, "learning_rate": 2.3353819280202455e-05, "loss": 0.0077, "num_input_tokens_seen": 40126368, "step": 20840 }, { "epoch": 3.400522065421323, "grad_norm": 0.052287276834249496, "learning_rate": 2.334402323705438e-05, "loss": 0.0015, "num_input_tokens_seen": 40135728, "step": 20845 }, { "epoch": 3.4013377926421406, "grad_norm": 0.032651349902153015, "learning_rate": 2.3334227449278725e-05, "loss": 0.0683, "num_input_tokens_seen": 40145504, "step": 20850 }, { "epoch": 3.4021535198629578, "grad_norm": 0.11049804836511612, "learning_rate": 2.3324431918386143e-05, "loss": 0.1568, "num_input_tokens_seen": 40155312, "step": 20855 }, { "epoch": 3.4029692470837754, "grad_norm": 0.2654881477355957, "learning_rate": 2.3314636645887207e-05, "loss": 0.086, "num_input_tokens_seen": 40163840, "step": 20860 }, { "epoch": 3.4037849743045925, "grad_norm": 0.09101933240890503, "learning_rate": 2.3304841633292487e-05, "loss": 0.0051, "num_input_tokens_seen": 40174128, "step": 20865 }, { "epoch": 3.40460070152541, "grad_norm": 0.05697968974709511, "learning_rate": 2.329504688211248e-05, "loss": 0.0528, "num_input_tokens_seen": 40182416, "step": 20870 }, { "epoch": 3.405416428746227, "grad_norm": 7.0572428703308105, "learning_rate": 2.3285252393857677e-05, "loss": 0.0487, "num_input_tokens_seen": 40192848, "step": 20875 }, { "epoch": 3.4062321559670448, "grad_norm": 0.2321046143770218, "learning_rate": 2.327545817003851e-05, "loss": 0.0024, "num_input_tokens_seen": 40202240, "step": 20880 }, { "epoch": 3.407047883187862, "grad_norm": 0.04804759472608566, "learning_rate": 2.326566421216535e-05, "loss": 0.1604, "num_input_tokens_seen": 40211760, "step": 20885 }, { "epoch": 3.4078636104086795, "grad_norm": 0.027640866115689278, "learning_rate": 2.3255870521748565e-05, "loss": 0.0984, "num_input_tokens_seen": 40221808, "step": 20890 }, { "epoch": 3.4086793376294966, "grad_norm": 0.028458671644330025, "learning_rate": 2.3246077100298474e-05, "loss": 0.1517, "num_input_tokens_seen": 40231152, "step": 20895 }, { "epoch": 3.409495064850314, "grad_norm": 0.016782118007540703, "learning_rate": 2.3236283949325328e-05, "loss": 0.0018, "num_input_tokens_seen": 40242512, "step": 20900 }, { "epoch": 3.4103107920711313, "grad_norm": 0.022037863731384277, "learning_rate": 2.3226491070339368e-05, "loss": 0.0054, "num_input_tokens_seen": 40252160, "step": 20905 }, { "epoch": 3.411126519291949, "grad_norm": 0.056989651173353195, "learning_rate": 2.3216698464850762e-05, "loss": 0.0011, "num_input_tokens_seen": 40261264, "step": 20910 }, { "epoch": 3.411942246512766, "grad_norm": 0.01790255680680275, "learning_rate": 2.320690613436967e-05, "loss": 0.005, "num_input_tokens_seen": 40271472, "step": 20915 }, { "epoch": 3.4127579737335836, "grad_norm": 0.12813201546669006, "learning_rate": 2.3197114080406192e-05, "loss": 0.0015, "num_input_tokens_seen": 40282208, "step": 20920 }, { "epoch": 3.4135737009544007, "grad_norm": 0.4155734181404114, "learning_rate": 2.3187322304470365e-05, "loss": 0.0036, "num_input_tokens_seen": 40291008, "step": 20925 }, { "epoch": 3.4143894281752183, "grad_norm": 0.10471813380718231, "learning_rate": 2.3177530808072222e-05, "loss": 0.002, "num_input_tokens_seen": 40301440, "step": 20930 }, { "epoch": 3.4152051553960354, "grad_norm": 0.17182223498821259, "learning_rate": 2.316773959272174e-05, "loss": 0.0248, "num_input_tokens_seen": 40311328, "step": 20935 }, { "epoch": 3.416020882616853, "grad_norm": 0.05778462439775467, "learning_rate": 2.3157948659928823e-05, "loss": 0.0255, "num_input_tokens_seen": 40320208, "step": 20940 }, { "epoch": 3.41683660983767, "grad_norm": 4.645394325256348, "learning_rate": 2.3148158011203388e-05, "loss": 0.0903, "num_input_tokens_seen": 40329808, "step": 20945 }, { "epoch": 3.4176523370584877, "grad_norm": 0.020428087562322617, "learning_rate": 2.3138367648055253e-05, "loss": 0.0011, "num_input_tokens_seen": 40338720, "step": 20950 }, { "epoch": 3.418468064279305, "grad_norm": 0.2543233036994934, "learning_rate": 2.312857757199422e-05, "loss": 0.003, "num_input_tokens_seen": 40348832, "step": 20955 }, { "epoch": 3.4192837915001224, "grad_norm": 0.13757146894931793, "learning_rate": 2.3118787784530048e-05, "loss": 0.0011, "num_input_tokens_seen": 40357872, "step": 20960 }, { "epoch": 3.4200995187209395, "grad_norm": 0.012202057987451553, "learning_rate": 2.310899828717243e-05, "loss": 0.0014, "num_input_tokens_seen": 40367584, "step": 20965 }, { "epoch": 3.420915245941757, "grad_norm": 0.07405415177345276, "learning_rate": 2.309920908143104e-05, "loss": 0.0147, "num_input_tokens_seen": 40378064, "step": 20970 }, { "epoch": 3.4217309731625742, "grad_norm": 0.068646140396595, "learning_rate": 2.308942016881551e-05, "loss": 0.1209, "num_input_tokens_seen": 40387760, "step": 20975 }, { "epoch": 3.422546700383392, "grad_norm": 15.986627578735352, "learning_rate": 2.307963155083539e-05, "loss": 0.0044, "num_input_tokens_seen": 40397824, "step": 20980 }, { "epoch": 3.423362427604209, "grad_norm": 0.2839064598083496, "learning_rate": 2.306984322900022e-05, "loss": 0.0177, "num_input_tokens_seen": 40408608, "step": 20985 }, { "epoch": 3.4241781548250265, "grad_norm": 0.04615724831819534, "learning_rate": 2.3060055204819482e-05, "loss": 0.0008, "num_input_tokens_seen": 40417936, "step": 20990 }, { "epoch": 3.424993882045844, "grad_norm": 0.07986809313297272, "learning_rate": 2.3050267479802604e-05, "loss": 0.0675, "num_input_tokens_seen": 40427600, "step": 20995 }, { "epoch": 3.4258096092666612, "grad_norm": 0.028423411771655083, "learning_rate": 2.304048005545899e-05, "loss": 0.0049, "num_input_tokens_seen": 40436560, "step": 21000 }, { "epoch": 3.4258096092666612, "eval_loss": 0.24008458852767944, "eval_runtime": 80.7969, "eval_samples_per_second": 33.727, "eval_steps_per_second": 16.869, "num_input_tokens_seen": 40436560, "step": 21000 }, { "epoch": 3.4266253364874784, "grad_norm": 0.048677053302526474, "learning_rate": 2.3030692933297972e-05, "loss": 0.0969, "num_input_tokens_seen": 40446528, "step": 21005 }, { "epoch": 3.427441063708296, "grad_norm": 128.7406768798828, "learning_rate": 2.3020906114828843e-05, "loss": 0.1491, "num_input_tokens_seen": 40455296, "step": 21010 }, { "epoch": 3.4282567909291135, "grad_norm": 0.03305704519152641, "learning_rate": 2.301111960156088e-05, "loss": 0.2425, "num_input_tokens_seen": 40465712, "step": 21015 }, { "epoch": 3.4290725181499306, "grad_norm": 0.061642710119485855, "learning_rate": 2.300133339500326e-05, "loss": 0.076, "num_input_tokens_seen": 40474880, "step": 21020 }, { "epoch": 3.429888245370748, "grad_norm": 0.08140570670366287, "learning_rate": 2.2991547496665148e-05, "loss": 0.1318, "num_input_tokens_seen": 40483392, "step": 21025 }, { "epoch": 3.4307039725915653, "grad_norm": 0.03323007747530937, "learning_rate": 2.298176190805565e-05, "loss": 0.0037, "num_input_tokens_seen": 40492400, "step": 21030 }, { "epoch": 3.431519699812383, "grad_norm": 0.18762052059173584, "learning_rate": 2.2971976630683826e-05, "loss": 0.0022, "num_input_tokens_seen": 40502704, "step": 21035 }, { "epoch": 3.4323354270332, "grad_norm": 0.02898377925157547, "learning_rate": 2.29621916660587e-05, "loss": 0.0009, "num_input_tokens_seen": 40511152, "step": 21040 }, { "epoch": 3.4331511542540176, "grad_norm": 0.10076163709163666, "learning_rate": 2.295240701568922e-05, "loss": 0.0013, "num_input_tokens_seen": 40521648, "step": 21045 }, { "epoch": 3.4339668814748348, "grad_norm": 0.019315827637910843, "learning_rate": 2.2942622681084312e-05, "loss": 0.0015, "num_input_tokens_seen": 40531360, "step": 21050 }, { "epoch": 3.4347826086956523, "grad_norm": 0.2728413939476013, "learning_rate": 2.293283866375284e-05, "loss": 0.0451, "num_input_tokens_seen": 40541200, "step": 21055 }, { "epoch": 3.4355983359164695, "grad_norm": 0.6648605465888977, "learning_rate": 2.2923054965203627e-05, "loss": 0.0023, "num_input_tokens_seen": 40551104, "step": 21060 }, { "epoch": 3.436414063137287, "grad_norm": 0.039832908660173416, "learning_rate": 2.2913271586945443e-05, "loss": 0.0017, "num_input_tokens_seen": 40560608, "step": 21065 }, { "epoch": 3.437229790358104, "grad_norm": 4.344415187835693, "learning_rate": 2.290348853048699e-05, "loss": 0.0107, "num_input_tokens_seen": 40570208, "step": 21070 }, { "epoch": 3.4380455175789217, "grad_norm": 0.057867489755153656, "learning_rate": 2.2893705797336956e-05, "loss": 0.0011, "num_input_tokens_seen": 40578160, "step": 21075 }, { "epoch": 3.438861244799739, "grad_norm": 0.2916070520877838, "learning_rate": 2.288392338900397e-05, "loss": 0.1454, "num_input_tokens_seen": 40587792, "step": 21080 }, { "epoch": 3.4396769720205564, "grad_norm": 0.014510133303701878, "learning_rate": 2.2874141306996576e-05, "loss": 0.003, "num_input_tokens_seen": 40596800, "step": 21085 }, { "epoch": 3.4404926992413736, "grad_norm": 0.03915877640247345, "learning_rate": 2.2864359552823312e-05, "loss": 0.001, "num_input_tokens_seen": 40605952, "step": 21090 }, { "epoch": 3.441308426462191, "grad_norm": 5.139720916748047, "learning_rate": 2.2854578127992648e-05, "loss": 0.0833, "num_input_tokens_seen": 40615456, "step": 21095 }, { "epoch": 3.4421241536830083, "grad_norm": 0.05561329424381256, "learning_rate": 2.2844797034012988e-05, "loss": 0.1153, "num_input_tokens_seen": 40625536, "step": 21100 }, { "epoch": 3.442939880903826, "grad_norm": 0.13770297169685364, "learning_rate": 2.2835016272392722e-05, "loss": 0.0662, "num_input_tokens_seen": 40636000, "step": 21105 }, { "epoch": 3.443755608124643, "grad_norm": 0.04581727460026741, "learning_rate": 2.2825235844640142e-05, "loss": 0.0032, "num_input_tokens_seen": 40646032, "step": 21110 }, { "epoch": 3.4445713353454606, "grad_norm": 0.601372241973877, "learning_rate": 2.2815455752263522e-05, "loss": 0.0029, "num_input_tokens_seen": 40656016, "step": 21115 }, { "epoch": 3.4453870625662777, "grad_norm": 0.2832775413990021, "learning_rate": 2.2805675996771092e-05, "loss": 0.02, "num_input_tokens_seen": 40667584, "step": 21120 }, { "epoch": 3.4462027897870953, "grad_norm": 0.009084327146410942, "learning_rate": 2.2795896579670987e-05, "loss": 0.1826, "num_input_tokens_seen": 40678176, "step": 21125 }, { "epoch": 3.4470185170079124, "grad_norm": 8.594374656677246, "learning_rate": 2.2786117502471337e-05, "loss": 0.0969, "num_input_tokens_seen": 40686480, "step": 21130 }, { "epoch": 3.44783424422873, "grad_norm": 0.0675189271569252, "learning_rate": 2.2776338766680185e-05, "loss": 0.0345, "num_input_tokens_seen": 40696816, "step": 21135 }, { "epoch": 3.448649971449547, "grad_norm": 0.017639361321926117, "learning_rate": 2.2766560373805533e-05, "loss": 0.0011, "num_input_tokens_seen": 40706736, "step": 21140 }, { "epoch": 3.4494656986703647, "grad_norm": 0.01746949926018715, "learning_rate": 2.2756782325355353e-05, "loss": 0.0919, "num_input_tokens_seen": 40716192, "step": 21145 }, { "epoch": 3.450281425891182, "grad_norm": 0.1258089542388916, "learning_rate": 2.2747004622837514e-05, "loss": 0.0015, "num_input_tokens_seen": 40725520, "step": 21150 }, { "epoch": 3.4510971531119994, "grad_norm": 0.18648461997509003, "learning_rate": 2.2737227267759878e-05, "loss": 0.09, "num_input_tokens_seen": 40734800, "step": 21155 }, { "epoch": 3.4519128803328165, "grad_norm": 0.01571417786180973, "learning_rate": 2.272745026163024e-05, "loss": 0.0676, "num_input_tokens_seen": 40744208, "step": 21160 }, { "epoch": 3.452728607553634, "grad_norm": 7.043460369110107, "learning_rate": 2.271767360595633e-05, "loss": 0.0481, "num_input_tokens_seen": 40753936, "step": 21165 }, { "epoch": 3.4535443347744517, "grad_norm": 0.006073142401874065, "learning_rate": 2.270789730224583e-05, "loss": 0.0046, "num_input_tokens_seen": 40764208, "step": 21170 }, { "epoch": 3.454360061995269, "grad_norm": 4.018921852111816, "learning_rate": 2.2698121352006367e-05, "loss": 0.3085, "num_input_tokens_seen": 40774272, "step": 21175 }, { "epoch": 3.455175789216086, "grad_norm": 15.046433448791504, "learning_rate": 2.2688345756745517e-05, "loss": 0.0096, "num_input_tokens_seen": 40783696, "step": 21180 }, { "epoch": 3.4559915164369035, "grad_norm": 0.09312200546264648, "learning_rate": 2.267857051797081e-05, "loss": 0.0034, "num_input_tokens_seen": 40793376, "step": 21185 }, { "epoch": 3.456807243657721, "grad_norm": 0.04217598959803581, "learning_rate": 2.2668795637189695e-05, "loss": 0.0019, "num_input_tokens_seen": 40801920, "step": 21190 }, { "epoch": 3.457622970878538, "grad_norm": 0.03281543776392937, "learning_rate": 2.2659021115909586e-05, "loss": 0.0908, "num_input_tokens_seen": 40810880, "step": 21195 }, { "epoch": 3.4584386980993553, "grad_norm": 0.09307906776666641, "learning_rate": 2.2649246955637847e-05, "loss": 0.0513, "num_input_tokens_seen": 40820704, "step": 21200 }, { "epoch": 3.4584386980993553, "eval_loss": 0.2144426554441452, "eval_runtime": 80.7804, "eval_samples_per_second": 33.733, "eval_steps_per_second": 16.873, "num_input_tokens_seen": 40820704, "step": 21200 }, { "epoch": 3.459254425320173, "grad_norm": 0.5533918738365173, "learning_rate": 2.2639473157881766e-05, "loss": 0.0028, "num_input_tokens_seen": 40830592, "step": 21205 }, { "epoch": 3.4600701525409905, "grad_norm": 0.1816835105419159, "learning_rate": 2.2629699724148594e-05, "loss": 0.09, "num_input_tokens_seen": 40840160, "step": 21210 }, { "epoch": 3.4608858797618076, "grad_norm": 9.642165184020996, "learning_rate": 2.26199266559455e-05, "loss": 0.0887, "num_input_tokens_seen": 40849248, "step": 21215 }, { "epoch": 3.461701606982625, "grad_norm": 0.08780097216367722, "learning_rate": 2.2610153954779625e-05, "loss": 0.0012, "num_input_tokens_seen": 40859648, "step": 21220 }, { "epoch": 3.4625173342034423, "grad_norm": 0.1120121031999588, "learning_rate": 2.2600381622158056e-05, "loss": 0.0014, "num_input_tokens_seen": 40868672, "step": 21225 }, { "epoch": 3.46333306142426, "grad_norm": 0.025707721710205078, "learning_rate": 2.2590609659587783e-05, "loss": 0.1636, "num_input_tokens_seen": 40877552, "step": 21230 }, { "epoch": 3.464148788645077, "grad_norm": 0.5730763077735901, "learning_rate": 2.2580838068575787e-05, "loss": 0.0016, "num_input_tokens_seen": 40886224, "step": 21235 }, { "epoch": 3.4649645158658946, "grad_norm": 0.07311075925827026, "learning_rate": 2.257106685062896e-05, "loss": 0.0486, "num_input_tokens_seen": 40894464, "step": 21240 }, { "epoch": 3.4657802430867117, "grad_norm": 32.24970245361328, "learning_rate": 2.256129600725415e-05, "loss": 0.191, "num_input_tokens_seen": 40904688, "step": 21245 }, { "epoch": 3.4665959703075293, "grad_norm": 0.060759562999010086, "learning_rate": 2.2551525539958145e-05, "loss": 0.093, "num_input_tokens_seen": 40914304, "step": 21250 }, { "epoch": 3.4674116975283464, "grad_norm": 0.01602395996451378, "learning_rate": 2.2541755450247663e-05, "loss": 0.0664, "num_input_tokens_seen": 40923504, "step": 21255 }, { "epoch": 3.468227424749164, "grad_norm": 6.334537029266357, "learning_rate": 2.2531985739629382e-05, "loss": 0.031, "num_input_tokens_seen": 40933536, "step": 21260 }, { "epoch": 3.469043151969981, "grad_norm": 0.1286562979221344, "learning_rate": 2.2522216409609924e-05, "loss": 0.061, "num_input_tokens_seen": 40942336, "step": 21265 }, { "epoch": 3.4698588791907987, "grad_norm": 0.04189978539943695, "learning_rate": 2.2512447461695826e-05, "loss": 0.1712, "num_input_tokens_seen": 40952096, "step": 21270 }, { "epoch": 3.470674606411616, "grad_norm": 0.12571856379508972, "learning_rate": 2.2502678897393593e-05, "loss": 0.034, "num_input_tokens_seen": 40962416, "step": 21275 }, { "epoch": 3.4714903336324334, "grad_norm": 0.05844392627477646, "learning_rate": 2.2492910718209665e-05, "loss": 0.0803, "num_input_tokens_seen": 40971680, "step": 21280 }, { "epoch": 3.4723060608532506, "grad_norm": 6.6170973777771, "learning_rate": 2.2483142925650398e-05, "loss": 0.0092, "num_input_tokens_seen": 40982656, "step": 21285 }, { "epoch": 3.473121788074068, "grad_norm": 0.0231233611702919, "learning_rate": 2.247337552122213e-05, "loss": 0.0017, "num_input_tokens_seen": 40990560, "step": 21290 }, { "epoch": 3.4739375152948853, "grad_norm": 17.421934127807617, "learning_rate": 2.24636085064311e-05, "loss": 0.0562, "num_input_tokens_seen": 41000416, "step": 21295 }, { "epoch": 3.474753242515703, "grad_norm": 0.023104188963770866, "learning_rate": 2.245384188278351e-05, "loss": 0.0076, "num_input_tokens_seen": 41010144, "step": 21300 }, { "epoch": 3.47556896973652, "grad_norm": 0.7305206060409546, "learning_rate": 2.2444075651785513e-05, "loss": 0.0026, "num_input_tokens_seen": 41020288, "step": 21305 }, { "epoch": 3.4763846969573375, "grad_norm": 0.01923373155295849, "learning_rate": 2.243430981494316e-05, "loss": 0.0077, "num_input_tokens_seen": 41029040, "step": 21310 }, { "epoch": 3.4772004241781547, "grad_norm": 0.048619821667671204, "learning_rate": 2.2424544373762475e-05, "loss": 0.0016, "num_input_tokens_seen": 41039728, "step": 21315 }, { "epoch": 3.4780161513989722, "grad_norm": 0.028527721762657166, "learning_rate": 2.2414779329749418e-05, "loss": 0.0604, "num_input_tokens_seen": 41048736, "step": 21320 }, { "epoch": 3.4788318786197894, "grad_norm": 0.09809727966785431, "learning_rate": 2.2405014684409873e-05, "loss": 0.0048, "num_input_tokens_seen": 41057824, "step": 21325 }, { "epoch": 3.479647605840607, "grad_norm": 0.31992846727371216, "learning_rate": 2.239525043924968e-05, "loss": 0.0021, "num_input_tokens_seen": 41067824, "step": 21330 }, { "epoch": 3.480463333061424, "grad_norm": 0.013210638426244259, "learning_rate": 2.2385486595774592e-05, "loss": 0.0927, "num_input_tokens_seen": 41078144, "step": 21335 }, { "epoch": 3.4812790602822417, "grad_norm": 0.06236208230257034, "learning_rate": 2.237572315549033e-05, "loss": 0.0009, "num_input_tokens_seen": 41086896, "step": 21340 }, { "epoch": 3.482094787503059, "grad_norm": 0.0182952880859375, "learning_rate": 2.2365960119902545e-05, "loss": 0.1047, "num_input_tokens_seen": 41096944, "step": 21345 }, { "epoch": 3.4829105147238764, "grad_norm": 0.041738640516996384, "learning_rate": 2.2356197490516806e-05, "loss": 0.0059, "num_input_tokens_seen": 41107904, "step": 21350 }, { "epoch": 3.4837262419446935, "grad_norm": 0.09220010042190552, "learning_rate": 2.234643526883863e-05, "loss": 0.0066, "num_input_tokens_seen": 41117920, "step": 21355 }, { "epoch": 3.484541969165511, "grad_norm": 1.2405239343643188, "learning_rate": 2.2336673456373497e-05, "loss": 0.0036, "num_input_tokens_seen": 41126784, "step": 21360 }, { "epoch": 3.4853576963863286, "grad_norm": 0.07442279905080795, "learning_rate": 2.2326912054626772e-05, "loss": 0.0012, "num_input_tokens_seen": 41135792, "step": 21365 }, { "epoch": 3.4861734236071458, "grad_norm": 0.03339843824505806, "learning_rate": 2.2317151065103813e-05, "loss": 0.0048, "num_input_tokens_seen": 41145840, "step": 21370 }, { "epoch": 3.486989150827963, "grad_norm": 0.10854662954807281, "learning_rate": 2.2307390489309865e-05, "loss": 0.1579, "num_input_tokens_seen": 41156224, "step": 21375 }, { "epoch": 3.4878048780487805, "grad_norm": 0.008725069463253021, "learning_rate": 2.2297630328750146e-05, "loss": 0.2276, "num_input_tokens_seen": 41165776, "step": 21380 }, { "epoch": 3.488620605269598, "grad_norm": 0.01032553892582655, "learning_rate": 2.228787058492979e-05, "loss": 0.0886, "num_input_tokens_seen": 41175216, "step": 21385 }, { "epoch": 3.489436332490415, "grad_norm": 0.12804967164993286, "learning_rate": 2.2278111259353875e-05, "loss": 0.0925, "num_input_tokens_seen": 41184656, "step": 21390 }, { "epoch": 3.4902520597112328, "grad_norm": 0.015665683895349503, "learning_rate": 2.2268352353527395e-05, "loss": 0.0263, "num_input_tokens_seen": 41193888, "step": 21395 }, { "epoch": 3.49106778693205, "grad_norm": 5.304450035095215, "learning_rate": 2.225859386895533e-05, "loss": 0.0909, "num_input_tokens_seen": 41202080, "step": 21400 }, { "epoch": 3.49106778693205, "eval_loss": 0.23057134449481964, "eval_runtime": 80.7501, "eval_samples_per_second": 33.746, "eval_steps_per_second": 16.879, "num_input_tokens_seen": 41202080, "step": 21400 }, { "epoch": 3.4918835141528675, "grad_norm": 0.06289172172546387, "learning_rate": 2.2248835807142525e-05, "loss": 0.0502, "num_input_tokens_seen": 41211888, "step": 21405 }, { "epoch": 3.4926992413736846, "grad_norm": 1.008634090423584, "learning_rate": 2.2239078169593826e-05, "loss": 0.0815, "num_input_tokens_seen": 41220832, "step": 21410 }, { "epoch": 3.493514968594502, "grad_norm": 0.0969105064868927, "learning_rate": 2.222932095781396e-05, "loss": 0.0011, "num_input_tokens_seen": 41229600, "step": 21415 }, { "epoch": 3.4943306958153193, "grad_norm": 0.20747077465057373, "learning_rate": 2.221956417330762e-05, "loss": 0.0021, "num_input_tokens_seen": 41238416, "step": 21420 }, { "epoch": 3.495146423036137, "grad_norm": 0.036896031349897385, "learning_rate": 2.2209807817579438e-05, "loss": 0.1323, "num_input_tokens_seen": 41248816, "step": 21425 }, { "epoch": 3.495962150256954, "grad_norm": 0.06582999974489212, "learning_rate": 2.220005189213394e-05, "loss": 0.0121, "num_input_tokens_seen": 41257872, "step": 21430 }, { "epoch": 3.4967778774777716, "grad_norm": 3.0573766231536865, "learning_rate": 2.2190296398475624e-05, "loss": 0.1264, "num_input_tokens_seen": 41267920, "step": 21435 }, { "epoch": 3.4975936046985887, "grad_norm": 0.0731719359755516, "learning_rate": 2.2180541338108926e-05, "loss": 0.018, "num_input_tokens_seen": 41278000, "step": 21440 }, { "epoch": 3.4984093319194063, "grad_norm": 0.03400823473930359, "learning_rate": 2.2170786712538176e-05, "loss": 0.056, "num_input_tokens_seen": 41287872, "step": 21445 }, { "epoch": 3.4992250591402234, "grad_norm": 0.01786184124648571, "learning_rate": 2.216103252326768e-05, "loss": 0.0044, "num_input_tokens_seen": 41297840, "step": 21450 }, { "epoch": 3.500040786361041, "grad_norm": 0.05992283672094345, "learning_rate": 2.2151278771801635e-05, "loss": 0.0205, "num_input_tokens_seen": 41307440, "step": 21455 }, { "epoch": 3.500856513581858, "grad_norm": 0.018827542662620544, "learning_rate": 2.21415254596442e-05, "loss": 0.0625, "num_input_tokens_seen": 41315840, "step": 21460 }, { "epoch": 3.5016722408026757, "grad_norm": 0.042535968124866486, "learning_rate": 2.213177258829947e-05, "loss": 0.0022, "num_input_tokens_seen": 41325216, "step": 21465 }, { "epoch": 3.502487968023493, "grad_norm": 0.02988072857260704, "learning_rate": 2.2122020159271445e-05, "loss": 0.0989, "num_input_tokens_seen": 41334768, "step": 21470 }, { "epoch": 3.5033036952443104, "grad_norm": 0.009184329770505428, "learning_rate": 2.2112268174064075e-05, "loss": 0.0018, "num_input_tokens_seen": 41344432, "step": 21475 }, { "epoch": 3.5041194224651275, "grad_norm": 11.577535629272461, "learning_rate": 2.2102516634181253e-05, "loss": 0.1293, "num_input_tokens_seen": 41353408, "step": 21480 }, { "epoch": 3.504935149685945, "grad_norm": 0.017421333119273186, "learning_rate": 2.209276554112677e-05, "loss": 0.3536, "num_input_tokens_seen": 41362272, "step": 21485 }, { "epoch": 3.5057508769067622, "grad_norm": 0.046686138957738876, "learning_rate": 2.2083014896404384e-05, "loss": 0.0022, "num_input_tokens_seen": 41372336, "step": 21490 }, { "epoch": 3.50656660412758, "grad_norm": 0.1173911839723587, "learning_rate": 2.207326470151775e-05, "loss": 0.1453, "num_input_tokens_seen": 41382368, "step": 21495 }, { "epoch": 3.507382331348397, "grad_norm": 0.05564790219068527, "learning_rate": 2.2063514957970477e-05, "loss": 0.0508, "num_input_tokens_seen": 41392672, "step": 21500 }, { "epoch": 3.5081980585692145, "grad_norm": 0.27882543206214905, "learning_rate": 2.205376566726611e-05, "loss": 0.0037, "num_input_tokens_seen": 41402992, "step": 21505 }, { "epoch": 3.5090137857900316, "grad_norm": 3.270355701446533, "learning_rate": 2.204401683090809e-05, "loss": 0.1605, "num_input_tokens_seen": 41413312, "step": 21510 }, { "epoch": 3.5098295130108492, "grad_norm": 0.1702650487422943, "learning_rate": 2.203426845039982e-05, "loss": 0.0033, "num_input_tokens_seen": 41422448, "step": 21515 }, { "epoch": 3.510645240231667, "grad_norm": 0.12540952861309052, "learning_rate": 2.202452052724464e-05, "loss": 0.0901, "num_input_tokens_seen": 41432224, "step": 21520 }, { "epoch": 3.511460967452484, "grad_norm": 0.20674540102481842, "learning_rate": 2.2014773062945777e-05, "loss": 0.0671, "num_input_tokens_seen": 41441712, "step": 21525 }, { "epoch": 3.512276694673301, "grad_norm": 6.674310684204102, "learning_rate": 2.2005026059006427e-05, "loss": 0.1109, "num_input_tokens_seen": 41451232, "step": 21530 }, { "epoch": 3.5130924218941186, "grad_norm": 11.885497093200684, "learning_rate": 2.1995279516929695e-05, "loss": 0.0607, "num_input_tokens_seen": 41460960, "step": 21535 }, { "epoch": 3.513908149114936, "grad_norm": 0.8603357076644897, "learning_rate": 2.1985533438218613e-05, "loss": 0.0036, "num_input_tokens_seen": 41471072, "step": 21540 }, { "epoch": 3.5147238763357533, "grad_norm": 0.30019471049308777, "learning_rate": 2.197578782437617e-05, "loss": 0.191, "num_input_tokens_seen": 41481680, "step": 21545 }, { "epoch": 3.5155396035565705, "grad_norm": 0.06874606013298035, "learning_rate": 2.196604267690524e-05, "loss": 0.0029, "num_input_tokens_seen": 41491216, "step": 21550 }, { "epoch": 3.516355330777388, "grad_norm": 0.11149551719427109, "learning_rate": 2.195629799730865e-05, "loss": 0.0594, "num_input_tokens_seen": 41501232, "step": 21555 }, { "epoch": 3.5171710579982056, "grad_norm": 0.06482197344303131, "learning_rate": 2.1946553787089173e-05, "loss": 0.0033, "num_input_tokens_seen": 41510656, "step": 21560 }, { "epoch": 3.5179867852190227, "grad_norm": 0.05035420134663582, "learning_rate": 2.193681004774947e-05, "loss": 0.0016, "num_input_tokens_seen": 41520864, "step": 21565 }, { "epoch": 3.51880251243984, "grad_norm": 0.3141658306121826, "learning_rate": 2.1927066780792154e-05, "loss": 0.0685, "num_input_tokens_seen": 41529264, "step": 21570 }, { "epoch": 3.5196182396606575, "grad_norm": 0.07996934652328491, "learning_rate": 2.191732398771975e-05, "loss": 0.0019, "num_input_tokens_seen": 41540144, "step": 21575 }, { "epoch": 3.520433966881475, "grad_norm": 0.051349710673093796, "learning_rate": 2.1907581670034725e-05, "loss": 0.0433, "num_input_tokens_seen": 41549808, "step": 21580 }, { "epoch": 3.521249694102292, "grad_norm": 0.05561695247888565, "learning_rate": 2.189783982923948e-05, "loss": 0.0581, "num_input_tokens_seen": 41558192, "step": 21585 }, { "epoch": 3.5220654213231093, "grad_norm": 0.11467240005731583, "learning_rate": 2.1888098466836303e-05, "loss": 0.0016, "num_input_tokens_seen": 41568624, "step": 21590 }, { "epoch": 3.522881148543927, "grad_norm": 0.01126091368496418, "learning_rate": 2.1878357584327457e-05, "loss": 0.1447, "num_input_tokens_seen": 41578848, "step": 21595 }, { "epoch": 3.5236968757647444, "grad_norm": 0.05755425617098808, "learning_rate": 2.1868617183215103e-05, "loss": 0.0033, "num_input_tokens_seen": 41588560, "step": 21600 }, { "epoch": 3.5236968757647444, "eval_loss": 0.22020629048347473, "eval_runtime": 80.7587, "eval_samples_per_second": 33.743, "eval_steps_per_second": 16.877, "num_input_tokens_seen": 41588560, "step": 21600 }, { "epoch": 3.5245126029855616, "grad_norm": 0.09692621231079102, "learning_rate": 2.1858877265001327e-05, "loss": 0.0025, "num_input_tokens_seen": 41597536, "step": 21605 }, { "epoch": 3.525328330206379, "grad_norm": 17.60674285888672, "learning_rate": 2.184913783118816e-05, "loss": 0.1039, "num_input_tokens_seen": 41606080, "step": 21610 }, { "epoch": 3.5261440574271963, "grad_norm": 0.02562488056719303, "learning_rate": 2.1839398883277522e-05, "loss": 0.0347, "num_input_tokens_seen": 41615952, "step": 21615 }, { "epoch": 3.526959784648014, "grad_norm": 1.1665050983428955, "learning_rate": 2.182966042277129e-05, "loss": 0.0026, "num_input_tokens_seen": 41624480, "step": 21620 }, { "epoch": 3.527775511868831, "grad_norm": 9.621599197387695, "learning_rate": 2.181992245117128e-05, "loss": 0.1642, "num_input_tokens_seen": 41633904, "step": 21625 }, { "epoch": 3.5285912390896486, "grad_norm": 0.035974062979221344, "learning_rate": 2.181018496997918e-05, "loss": 0.0689, "num_input_tokens_seen": 41644000, "step": 21630 }, { "epoch": 3.5294069663104657, "grad_norm": 0.15023870766162872, "learning_rate": 2.1800447980696648e-05, "loss": 0.2096, "num_input_tokens_seen": 41652720, "step": 21635 }, { "epoch": 3.5302226935312833, "grad_norm": 3.077965497970581, "learning_rate": 2.1790711484825248e-05, "loss": 0.0997, "num_input_tokens_seen": 41662144, "step": 21640 }, { "epoch": 3.5310384207521004, "grad_norm": 0.04259280860424042, "learning_rate": 2.178097548386646e-05, "loss": 0.0455, "num_input_tokens_seen": 41671984, "step": 21645 }, { "epoch": 3.531854147972918, "grad_norm": 0.0323316790163517, "learning_rate": 2.1771239979321712e-05, "loss": 0.1018, "num_input_tokens_seen": 41681312, "step": 21650 }, { "epoch": 3.532669875193735, "grad_norm": 0.024091482162475586, "learning_rate": 2.1761504972692327e-05, "loss": 0.0026, "num_input_tokens_seen": 41691056, "step": 21655 }, { "epoch": 3.5334856024145527, "grad_norm": 0.03494379669427872, "learning_rate": 2.1751770465479572e-05, "loss": 0.0139, "num_input_tokens_seen": 41701520, "step": 21660 }, { "epoch": 3.53430132963537, "grad_norm": 5.63151741027832, "learning_rate": 2.174203645918464e-05, "loss": 0.0733, "num_input_tokens_seen": 41710752, "step": 21665 }, { "epoch": 3.5351170568561874, "grad_norm": 0.016763264313340187, "learning_rate": 2.1732302955308624e-05, "loss": 0.0027, "num_input_tokens_seen": 41720752, "step": 21670 }, { "epoch": 3.5359327840770045, "grad_norm": 0.04849901795387268, "learning_rate": 2.172256995535255e-05, "loss": 0.1476, "num_input_tokens_seen": 41731504, "step": 21675 }, { "epoch": 3.536748511297822, "grad_norm": 0.1306498646736145, "learning_rate": 2.171283746081739e-05, "loss": 0.161, "num_input_tokens_seen": 41741936, "step": 21680 }, { "epoch": 3.537564238518639, "grad_norm": 26.056236267089844, "learning_rate": 2.1703105473203988e-05, "loss": 0.0691, "num_input_tokens_seen": 41751968, "step": 21685 }, { "epoch": 3.538379965739457, "grad_norm": 0.11730145663022995, "learning_rate": 2.1693373994013168e-05, "loss": 0.0451, "num_input_tokens_seen": 41761808, "step": 21690 }, { "epoch": 3.539195692960274, "grad_norm": 3.8069381713867188, "learning_rate": 2.168364302474562e-05, "loss": 0.1123, "num_input_tokens_seen": 41771504, "step": 21695 }, { "epoch": 3.5400114201810915, "grad_norm": 4.251990795135498, "learning_rate": 2.167391256690199e-05, "loss": 0.0846, "num_input_tokens_seen": 41782160, "step": 21700 }, { "epoch": 3.5408271474019086, "grad_norm": 0.13969364762306213, "learning_rate": 2.1664182621982855e-05, "loss": 0.0026, "num_input_tokens_seen": 41791664, "step": 21705 }, { "epoch": 3.541642874622726, "grad_norm": 0.05560339242219925, "learning_rate": 2.1654453191488673e-05, "loss": 0.0938, "num_input_tokens_seen": 41800896, "step": 21710 }, { "epoch": 3.5424586018435438, "grad_norm": 0.04560023173689842, "learning_rate": 2.1644724276919846e-05, "loss": 0.0045, "num_input_tokens_seen": 41811424, "step": 21715 }, { "epoch": 3.543274329064361, "grad_norm": 0.05647752434015274, "learning_rate": 2.1634995879776715e-05, "loss": 0.003, "num_input_tokens_seen": 41821440, "step": 21720 }, { "epoch": 3.544090056285178, "grad_norm": 2.477468490600586, "learning_rate": 2.162526800155949e-05, "loss": 0.0876, "num_input_tokens_seen": 41829776, "step": 21725 }, { "epoch": 3.5449057835059956, "grad_norm": 9.274667739868164, "learning_rate": 2.1615540643768363e-05, "loss": 0.067, "num_input_tokens_seen": 41840512, "step": 21730 }, { "epoch": 3.545721510726813, "grad_norm": 0.10546979308128357, "learning_rate": 2.160581380790339e-05, "loss": 0.1605, "num_input_tokens_seen": 41849952, "step": 21735 }, { "epoch": 3.5465372379476303, "grad_norm": 0.18198207020759583, "learning_rate": 2.1596087495464586e-05, "loss": 0.0107, "num_input_tokens_seen": 41858944, "step": 21740 }, { "epoch": 3.5473529651684474, "grad_norm": 0.06261551380157471, "learning_rate": 2.1586361707951866e-05, "loss": 0.0935, "num_input_tokens_seen": 41869760, "step": 21745 }, { "epoch": 3.548168692389265, "grad_norm": 0.138240784406662, "learning_rate": 2.157663644686507e-05, "loss": 0.0151, "num_input_tokens_seen": 41879072, "step": 21750 }, { "epoch": 3.5489844196100826, "grad_norm": 0.36632734537124634, "learning_rate": 2.156691171370396e-05, "loss": 0.004, "num_input_tokens_seen": 41888912, "step": 21755 }, { "epoch": 3.5498001468308997, "grad_norm": 0.12360133230686188, "learning_rate": 2.1557187509968195e-05, "loss": 0.002, "num_input_tokens_seen": 41898176, "step": 21760 }, { "epoch": 3.550615874051717, "grad_norm": 0.2835187613964081, "learning_rate": 2.1547463837157382e-05, "loss": 0.2004, "num_input_tokens_seen": 41906336, "step": 21765 }, { "epoch": 3.5514316012725344, "grad_norm": 2.17880916595459, "learning_rate": 2.1537740696771045e-05, "loss": 0.0413, "num_input_tokens_seen": 41916640, "step": 21770 }, { "epoch": 3.552247328493352, "grad_norm": 0.05654972419142723, "learning_rate": 2.1528018090308587e-05, "loss": 0.0295, "num_input_tokens_seen": 41926880, "step": 21775 }, { "epoch": 3.553063055714169, "grad_norm": 10.593392372131348, "learning_rate": 2.151829601926938e-05, "loss": 0.0351, "num_input_tokens_seen": 41936240, "step": 21780 }, { "epoch": 3.5538787829349863, "grad_norm": 5.336388111114502, "learning_rate": 2.1508574485152684e-05, "loss": 0.054, "num_input_tokens_seen": 41947136, "step": 21785 }, { "epoch": 3.554694510155804, "grad_norm": 0.038806259632110596, "learning_rate": 2.1498853489457667e-05, "loss": 0.0414, "num_input_tokens_seen": 41957440, "step": 21790 }, { "epoch": 3.5555102373766214, "grad_norm": 1.2397881746292114, "learning_rate": 2.1489133033683455e-05, "loss": 0.0031, "num_input_tokens_seen": 41968000, "step": 21795 }, { "epoch": 3.5563259645974385, "grad_norm": 20.81171989440918, "learning_rate": 2.1479413119329038e-05, "loss": 0.0445, "num_input_tokens_seen": 41977888, "step": 21800 }, { "epoch": 3.5563259645974385, "eval_loss": 0.22205375134944916, "eval_runtime": 80.8068, "eval_samples_per_second": 33.722, "eval_steps_per_second": 16.867, "num_input_tokens_seen": 41977888, "step": 21800 }, { "epoch": 3.557141691818256, "grad_norm": 0.010091330856084824, "learning_rate": 2.1469693747893355e-05, "loss": 0.0128, "num_input_tokens_seen": 41988144, "step": 21805 }, { "epoch": 3.5579574190390733, "grad_norm": 0.057582512497901917, "learning_rate": 2.1459974920875274e-05, "loss": 0.0533, "num_input_tokens_seen": 41997824, "step": 21810 }, { "epoch": 3.558773146259891, "grad_norm": 0.06249683350324631, "learning_rate": 2.145025663977354e-05, "loss": 0.0019, "num_input_tokens_seen": 42007808, "step": 21815 }, { "epoch": 3.559588873480708, "grad_norm": 0.019337501376867294, "learning_rate": 2.1440538906086844e-05, "loss": 0.0013, "num_input_tokens_seen": 42016816, "step": 21820 }, { "epoch": 3.5604046007015255, "grad_norm": 0.1012239009141922, "learning_rate": 2.1430821721313782e-05, "loss": 0.0023, "num_input_tokens_seen": 42027264, "step": 21825 }, { "epoch": 3.5612203279223427, "grad_norm": 5.797224044799805, "learning_rate": 2.142110508695286e-05, "loss": 0.1673, "num_input_tokens_seen": 42037312, "step": 21830 }, { "epoch": 3.5620360551431602, "grad_norm": 0.13557836413383484, "learning_rate": 2.1411389004502515e-05, "loss": 0.0017, "num_input_tokens_seen": 42047008, "step": 21835 }, { "epoch": 3.5628517823639774, "grad_norm": 0.01056506298482418, "learning_rate": 2.140167347546107e-05, "loss": 0.0016, "num_input_tokens_seen": 42057056, "step": 21840 }, { "epoch": 3.563667509584795, "grad_norm": 3.7790839672088623, "learning_rate": 2.1391958501326793e-05, "loss": 0.0825, "num_input_tokens_seen": 42066512, "step": 21845 }, { "epoch": 3.564483236805612, "grad_norm": 16.58628273010254, "learning_rate": 2.1382244083597873e-05, "loss": 0.1055, "num_input_tokens_seen": 42075280, "step": 21850 }, { "epoch": 3.5652989640264297, "grad_norm": 0.20439057052135468, "learning_rate": 2.137253022377237e-05, "loss": 0.0562, "num_input_tokens_seen": 42085392, "step": 21855 }, { "epoch": 3.566114691247247, "grad_norm": 0.022235127165913582, "learning_rate": 2.136281692334829e-05, "loss": 0.0014, "num_input_tokens_seen": 42095504, "step": 21860 }, { "epoch": 3.5669304184680644, "grad_norm": 0.5018846988677979, "learning_rate": 2.135310418382356e-05, "loss": 0.1531, "num_input_tokens_seen": 42104912, "step": 21865 }, { "epoch": 3.5677461456888815, "grad_norm": 0.034926094114780426, "learning_rate": 2.134339200669598e-05, "loss": 0.0008, "num_input_tokens_seen": 42114736, "step": 21870 }, { "epoch": 3.568561872909699, "grad_norm": 0.013431504368782043, "learning_rate": 2.133368039346331e-05, "loss": 0.0599, "num_input_tokens_seen": 42123760, "step": 21875 }, { "epoch": 3.569377600130516, "grad_norm": 2.4528510570526123, "learning_rate": 2.1323969345623195e-05, "loss": 0.031, "num_input_tokens_seen": 42134752, "step": 21880 }, { "epoch": 3.5701933273513338, "grad_norm": 5.794452667236328, "learning_rate": 2.1314258864673207e-05, "loss": 0.0835, "num_input_tokens_seen": 42144592, "step": 21885 }, { "epoch": 3.5710090545721513, "grad_norm": 0.1625540554523468, "learning_rate": 2.130454895211082e-05, "loss": 0.0536, "num_input_tokens_seen": 42153296, "step": 21890 }, { "epoch": 3.5718247817929685, "grad_norm": 0.5794196724891663, "learning_rate": 2.129483960943342e-05, "loss": 0.04, "num_input_tokens_seen": 42163904, "step": 21895 }, { "epoch": 3.5726405090137856, "grad_norm": 0.93848717212677, "learning_rate": 2.128513083813831e-05, "loss": 0.1574, "num_input_tokens_seen": 42173488, "step": 21900 }, { "epoch": 3.573456236234603, "grad_norm": 14.913854598999023, "learning_rate": 2.1275422639722724e-05, "loss": 0.0602, "num_input_tokens_seen": 42182976, "step": 21905 }, { "epoch": 3.5742719634554208, "grad_norm": 0.02655990980565548, "learning_rate": 2.126571501568376e-05, "loss": 0.0426, "num_input_tokens_seen": 42192240, "step": 21910 }, { "epoch": 3.575087690676238, "grad_norm": 11.369963645935059, "learning_rate": 2.1256007967518478e-05, "loss": 0.0859, "num_input_tokens_seen": 42200720, "step": 21915 }, { "epoch": 3.575903417897055, "grad_norm": 0.0035934560000896454, "learning_rate": 2.124630149672381e-05, "loss": 0.0846, "num_input_tokens_seen": 42210544, "step": 21920 }, { "epoch": 3.5767191451178726, "grad_norm": 5.360978603363037, "learning_rate": 2.1236595604796624e-05, "loss": 0.0834, "num_input_tokens_seen": 42220128, "step": 21925 }, { "epoch": 3.57753487233869, "grad_norm": 0.37049639225006104, "learning_rate": 2.1226890293233693e-05, "loss": 0.0892, "num_input_tokens_seen": 42229072, "step": 21930 }, { "epoch": 3.5783505995595073, "grad_norm": 4.307316303253174, "learning_rate": 2.1217185563531694e-05, "loss": 0.0829, "num_input_tokens_seen": 42239088, "step": 21935 }, { "epoch": 3.5791663267803244, "grad_norm": 9.324857711791992, "learning_rate": 2.120748141718721e-05, "loss": 0.0577, "num_input_tokens_seen": 42248864, "step": 21940 }, { "epoch": 3.579982054001142, "grad_norm": 12.835624694824219, "learning_rate": 2.1197777855696765e-05, "loss": 0.1177, "num_input_tokens_seen": 42256960, "step": 21945 }, { "epoch": 3.5807977812219596, "grad_norm": 0.0904562920331955, "learning_rate": 2.1188074880556746e-05, "loss": 0.0031, "num_input_tokens_seen": 42266208, "step": 21950 }, { "epoch": 3.5816135084427767, "grad_norm": 0.054017119109630585, "learning_rate": 2.1178372493263495e-05, "loss": 0.0876, "num_input_tokens_seen": 42276016, "step": 21955 }, { "epoch": 3.582429235663594, "grad_norm": 40.6777229309082, "learning_rate": 2.116867069531322e-05, "loss": 0.0246, "num_input_tokens_seen": 42286128, "step": 21960 }, { "epoch": 3.5832449628844114, "grad_norm": 0.014014148153364658, "learning_rate": 2.1158969488202073e-05, "loss": 0.0439, "num_input_tokens_seen": 42295120, "step": 21965 }, { "epoch": 3.584060690105229, "grad_norm": 0.13954344391822815, "learning_rate": 2.114926887342611e-05, "loss": 0.0942, "num_input_tokens_seen": 42305456, "step": 21970 }, { "epoch": 3.584876417326046, "grad_norm": 0.22224284708499908, "learning_rate": 2.113956885248127e-05, "loss": 0.0026, "num_input_tokens_seen": 42314688, "step": 21975 }, { "epoch": 3.5856921445468637, "grad_norm": 0.05057428404688835, "learning_rate": 2.112986942686342e-05, "loss": 0.0015, "num_input_tokens_seen": 42323936, "step": 21980 }, { "epoch": 3.586507871767681, "grad_norm": 0.08724889904260635, "learning_rate": 2.112017059806835e-05, "loss": 0.001, "num_input_tokens_seen": 42332496, "step": 21985 }, { "epoch": 3.5873235989884984, "grad_norm": 0.07848905026912689, "learning_rate": 2.1110472367591724e-05, "loss": 0.1494, "num_input_tokens_seen": 42342432, "step": 21990 }, { "epoch": 3.5881393262093155, "grad_norm": 0.12545502185821533, "learning_rate": 2.1100774736929145e-05, "loss": 0.009, "num_input_tokens_seen": 42351584, "step": 21995 }, { "epoch": 3.588955053430133, "grad_norm": 0.13646265864372253, "learning_rate": 2.10910777075761e-05, "loss": 0.0054, "num_input_tokens_seen": 42361392, "step": 22000 }, { "epoch": 3.588955053430133, "eval_loss": 0.2167072296142578, "eval_runtime": 80.8022, "eval_samples_per_second": 33.724, "eval_steps_per_second": 16.868, "num_input_tokens_seen": 42361392, "step": 22000 }, { "epoch": 3.5897707806509502, "grad_norm": 0.06067923083901405, "learning_rate": 2.108138128102799e-05, "loss": 0.0014, "num_input_tokens_seen": 42371904, "step": 22005 }, { "epoch": 3.590586507871768, "grad_norm": 0.8396866917610168, "learning_rate": 2.107168545878014e-05, "loss": 0.0065, "num_input_tokens_seen": 42380688, "step": 22010 }, { "epoch": 3.591402235092585, "grad_norm": 0.028711913153529167, "learning_rate": 2.106199024232775e-05, "loss": 0.1269, "num_input_tokens_seen": 42390864, "step": 22015 }, { "epoch": 3.5922179623134025, "grad_norm": 0.2775137722492218, "learning_rate": 2.105229563316595e-05, "loss": 0.0018, "num_input_tokens_seen": 42399696, "step": 22020 }, { "epoch": 3.5930336895342196, "grad_norm": 0.033992547541856766, "learning_rate": 2.1042601632789784e-05, "loss": 0.0014, "num_input_tokens_seen": 42409360, "step": 22025 }, { "epoch": 3.593849416755037, "grad_norm": 0.020716149359941483, "learning_rate": 2.103290824269417e-05, "loss": 0.0055, "num_input_tokens_seen": 42419168, "step": 22030 }, { "epoch": 3.5946651439758543, "grad_norm": 3.4594104290008545, "learning_rate": 2.1023215464373965e-05, "loss": 0.1788, "num_input_tokens_seen": 42429456, "step": 22035 }, { "epoch": 3.595480871196672, "grad_norm": 0.028664497658610344, "learning_rate": 2.1013523299323908e-05, "loss": 0.0042, "num_input_tokens_seen": 42440112, "step": 22040 }, { "epoch": 3.596296598417489, "grad_norm": 0.797788143157959, "learning_rate": 2.1003831749038654e-05, "loss": 0.1266, "num_input_tokens_seen": 42450576, "step": 22045 }, { "epoch": 3.5971123256383066, "grad_norm": 0.016397768631577492, "learning_rate": 2.099414081501277e-05, "loss": 0.0285, "num_input_tokens_seen": 42459488, "step": 22050 }, { "epoch": 3.5979280528591238, "grad_norm": 0.018768658861517906, "learning_rate": 2.09844504987407e-05, "loss": 0.0017, "num_input_tokens_seen": 42467280, "step": 22055 }, { "epoch": 3.5987437800799413, "grad_norm": 0.07280290871858597, "learning_rate": 2.097476080171683e-05, "loss": 0.0513, "num_input_tokens_seen": 42477104, "step": 22060 }, { "epoch": 3.5995595073007585, "grad_norm": 0.2688499093055725, "learning_rate": 2.0965071725435436e-05, "loss": 0.0022, "num_input_tokens_seen": 42486960, "step": 22065 }, { "epoch": 3.600375234521576, "grad_norm": 0.08014672994613647, "learning_rate": 2.0955383271390684e-05, "loss": 0.0255, "num_input_tokens_seen": 42495104, "step": 22070 }, { "epoch": 3.601190961742393, "grad_norm": 0.22114573419094086, "learning_rate": 2.094569544107666e-05, "loss": 0.0534, "num_input_tokens_seen": 42503664, "step": 22075 }, { "epoch": 3.6020066889632107, "grad_norm": 0.02581193670630455, "learning_rate": 2.093600823598735e-05, "loss": 0.0032, "num_input_tokens_seen": 42513584, "step": 22080 }, { "epoch": 3.6028224161840283, "grad_norm": 4.565693378448486, "learning_rate": 2.092632165761663e-05, "loss": 0.0828, "num_input_tokens_seen": 42524240, "step": 22085 }, { "epoch": 3.6036381434048455, "grad_norm": 0.31235265731811523, "learning_rate": 2.091663570745832e-05, "loss": 0.0745, "num_input_tokens_seen": 42533680, "step": 22090 }, { "epoch": 3.6044538706256626, "grad_norm": 4.324159622192383, "learning_rate": 2.0906950387006086e-05, "loss": 0.07, "num_input_tokens_seen": 42543904, "step": 22095 }, { "epoch": 3.60526959784648, "grad_norm": 6.381954669952393, "learning_rate": 2.0897265697753543e-05, "loss": 0.0352, "num_input_tokens_seen": 42553344, "step": 22100 }, { "epoch": 3.6060853250672977, "grad_norm": 0.05753001943230629, "learning_rate": 2.088758164119419e-05, "loss": 0.0015, "num_input_tokens_seen": 42562256, "step": 22105 }, { "epoch": 3.606901052288115, "grad_norm": 12.138768196105957, "learning_rate": 2.0877898218821428e-05, "loss": 0.1651, "num_input_tokens_seen": 42571504, "step": 22110 }, { "epoch": 3.607716779508932, "grad_norm": 0.04119554162025452, "learning_rate": 2.0868215432128565e-05, "loss": 0.0011, "num_input_tokens_seen": 42581712, "step": 22115 }, { "epoch": 3.6085325067297496, "grad_norm": 5.614025115966797, "learning_rate": 2.0858533282608796e-05, "loss": 0.0201, "num_input_tokens_seen": 42591440, "step": 22120 }, { "epoch": 3.609348233950567, "grad_norm": 0.5002683997154236, "learning_rate": 2.084885177175524e-05, "loss": 0.0048, "num_input_tokens_seen": 42600720, "step": 22125 }, { "epoch": 3.6101639611713843, "grad_norm": 0.04250982403755188, "learning_rate": 2.0839170901060917e-05, "loss": 0.0015, "num_input_tokens_seen": 42611248, "step": 22130 }, { "epoch": 3.6109796883922014, "grad_norm": 0.02865598350763321, "learning_rate": 2.082949067201872e-05, "loss": 0.0262, "num_input_tokens_seen": 42620304, "step": 22135 }, { "epoch": 3.611795415613019, "grad_norm": 0.3998732268810272, "learning_rate": 2.0819811086121475e-05, "loss": 0.0018, "num_input_tokens_seen": 42629936, "step": 22140 }, { "epoch": 3.6126111428338366, "grad_norm": 0.05655687674880028, "learning_rate": 2.08101321448619e-05, "loss": 0.0845, "num_input_tokens_seen": 42639968, "step": 22145 }, { "epoch": 3.6134268700546537, "grad_norm": 0.04455924406647682, "learning_rate": 2.080045384973259e-05, "loss": 0.001, "num_input_tokens_seen": 42650176, "step": 22150 }, { "epoch": 3.614242597275471, "grad_norm": 1.8896393775939941, "learning_rate": 2.0790776202226082e-05, "loss": 0.0032, "num_input_tokens_seen": 42660368, "step": 22155 }, { "epoch": 3.6150583244962884, "grad_norm": 0.011446557007730007, "learning_rate": 2.078109920383477e-05, "loss": 0.0175, "num_input_tokens_seen": 42670736, "step": 22160 }, { "epoch": 3.615874051717106, "grad_norm": 0.6180949807167053, "learning_rate": 2.0771422856050978e-05, "loss": 0.2362, "num_input_tokens_seen": 42679952, "step": 22165 }, { "epoch": 3.616689778937923, "grad_norm": 0.010625842027366161, "learning_rate": 2.076174716036693e-05, "loss": 0.0029, "num_input_tokens_seen": 42689216, "step": 22170 }, { "epoch": 3.6175055061587407, "grad_norm": 0.025493783876299858, "learning_rate": 2.075207211827472e-05, "loss": 0.001, "num_input_tokens_seen": 42699392, "step": 22175 }, { "epoch": 3.618321233379558, "grad_norm": 0.006858097389340401, "learning_rate": 2.074239773126638e-05, "loss": 0.223, "num_input_tokens_seen": 42709920, "step": 22180 }, { "epoch": 3.6191369606003754, "grad_norm": 0.17257803678512573, "learning_rate": 2.073272400083382e-05, "loss": 0.0013, "num_input_tokens_seen": 42719616, "step": 22185 }, { "epoch": 3.6199526878211925, "grad_norm": 0.15820761024951935, "learning_rate": 2.072305092846883e-05, "loss": 0.0818, "num_input_tokens_seen": 42728624, "step": 22190 }, { "epoch": 3.62076841504201, "grad_norm": 0.07830089330673218, "learning_rate": 2.0713378515663152e-05, "loss": 0.1258, "num_input_tokens_seen": 42737840, "step": 22195 }, { "epoch": 3.621584142262827, "grad_norm": 0.7776676416397095, "learning_rate": 2.070370676390836e-05, "loss": 0.0047, "num_input_tokens_seen": 42746416, "step": 22200 }, { "epoch": 3.621584142262827, "eval_loss": 0.22714954614639282, "eval_runtime": 80.8005, "eval_samples_per_second": 33.725, "eval_steps_per_second": 16.869, "num_input_tokens_seen": 42746416, "step": 22200 }, { "epoch": 3.622399869483645, "grad_norm": 0.09625179320573807, "learning_rate": 2.0694035674695974e-05, "loss": 0.0309, "num_input_tokens_seen": 42756288, "step": 22205 }, { "epoch": 3.623215596704462, "grad_norm": 17.017316818237305, "learning_rate": 2.0684365249517416e-05, "loss": 0.0294, "num_input_tokens_seen": 42766688, "step": 22210 }, { "epoch": 3.6240313239252795, "grad_norm": 7.702353000640869, "learning_rate": 2.067469548986396e-05, "loss": 0.1816, "num_input_tokens_seen": 42775520, "step": 22215 }, { "epoch": 3.6248470511460966, "grad_norm": 0.1148965060710907, "learning_rate": 2.066502639722681e-05, "loss": 0.0043, "num_input_tokens_seen": 42785360, "step": 22220 }, { "epoch": 3.625662778366914, "grad_norm": 0.14774702489376068, "learning_rate": 2.065535797309708e-05, "loss": 0.0499, "num_input_tokens_seen": 42794432, "step": 22225 }, { "epoch": 3.6264785055877313, "grad_norm": 0.5200445055961609, "learning_rate": 2.0645690218965736e-05, "loss": 0.0181, "num_input_tokens_seen": 42804272, "step": 22230 }, { "epoch": 3.627294232808549, "grad_norm": 0.1617402732372284, "learning_rate": 2.063602313632369e-05, "loss": 0.1281, "num_input_tokens_seen": 42814544, "step": 22235 }, { "epoch": 3.628109960029366, "grad_norm": 0.031996745616197586, "learning_rate": 2.0626356726661704e-05, "loss": 0.0652, "num_input_tokens_seen": 42824496, "step": 22240 }, { "epoch": 3.6289256872501836, "grad_norm": 0.38198739290237427, "learning_rate": 2.0616690991470477e-05, "loss": 0.0041, "num_input_tokens_seen": 42834032, "step": 22245 }, { "epoch": 3.6297414144710007, "grad_norm": 0.010467014275491238, "learning_rate": 2.0607025932240595e-05, "loss": 0.1532, "num_input_tokens_seen": 42841344, "step": 22250 }, { "epoch": 3.6305571416918183, "grad_norm": 0.023972608149051666, "learning_rate": 2.059736155046251e-05, "loss": 0.084, "num_input_tokens_seen": 42849776, "step": 22255 }, { "epoch": 3.631372868912636, "grad_norm": 0.11304476857185364, "learning_rate": 2.0587697847626603e-05, "loss": 0.028, "num_input_tokens_seen": 42859280, "step": 22260 }, { "epoch": 3.632188596133453, "grad_norm": 0.06290195882320404, "learning_rate": 2.057803482522314e-05, "loss": 0.0014, "num_input_tokens_seen": 42868800, "step": 22265 }, { "epoch": 3.63300432335427, "grad_norm": 5.976905345916748, "learning_rate": 2.056837248474227e-05, "loss": 0.2009, "num_input_tokens_seen": 42878848, "step": 22270 }, { "epoch": 3.6338200505750877, "grad_norm": 0.017591556534171104, "learning_rate": 2.0558710827674064e-05, "loss": 0.0455, "num_input_tokens_seen": 42887648, "step": 22275 }, { "epoch": 3.6346357777959053, "grad_norm": 0.18724150955677032, "learning_rate": 2.054904985550845e-05, "loss": 0.0038, "num_input_tokens_seen": 42897392, "step": 22280 }, { "epoch": 3.6354515050167224, "grad_norm": 0.12019192427396774, "learning_rate": 2.0539389569735287e-05, "loss": 0.0021, "num_input_tokens_seen": 42906384, "step": 22285 }, { "epoch": 3.6362672322375396, "grad_norm": 0.08440645039081573, "learning_rate": 2.052972997184431e-05, "loss": 0.0015, "num_input_tokens_seen": 42915072, "step": 22290 }, { "epoch": 3.637082959458357, "grad_norm": 0.04042300581932068, "learning_rate": 2.0520071063325146e-05, "loss": 0.0013, "num_input_tokens_seen": 42924032, "step": 22295 }, { "epoch": 3.6378986866791747, "grad_norm": 0.11198041588068008, "learning_rate": 2.051041284566732e-05, "loss": 0.1234, "num_input_tokens_seen": 42933168, "step": 22300 }, { "epoch": 3.638714413899992, "grad_norm": 0.13361385464668274, "learning_rate": 2.050075532036026e-05, "loss": 0.0032, "num_input_tokens_seen": 42942944, "step": 22305 }, { "epoch": 3.639530141120809, "grad_norm": 0.21988584101200104, "learning_rate": 2.0491098488893264e-05, "loss": 0.0576, "num_input_tokens_seen": 42952464, "step": 22310 }, { "epoch": 3.6403458683416265, "grad_norm": 12.508980751037598, "learning_rate": 2.0481442352755546e-05, "loss": 0.1142, "num_input_tokens_seen": 42962880, "step": 22315 }, { "epoch": 3.641161595562444, "grad_norm": 0.17452280223369598, "learning_rate": 2.0471786913436198e-05, "loss": 0.0106, "num_input_tokens_seen": 42972976, "step": 22320 }, { "epoch": 3.6419773227832613, "grad_norm": 0.41815271973609924, "learning_rate": 2.0462132172424218e-05, "loss": 0.0059, "num_input_tokens_seen": 42983024, "step": 22325 }, { "epoch": 3.6427930500040784, "grad_norm": 7.557932376861572, "learning_rate": 2.0452478131208484e-05, "loss": 0.1628, "num_input_tokens_seen": 42993472, "step": 22330 }, { "epoch": 3.643608777224896, "grad_norm": 41.03466033935547, "learning_rate": 2.0442824791277765e-05, "loss": 0.0712, "num_input_tokens_seen": 43003344, "step": 22335 }, { "epoch": 3.6444245044457135, "grad_norm": 0.04566115140914917, "learning_rate": 2.0433172154120727e-05, "loss": 0.0047, "num_input_tokens_seen": 43012080, "step": 22340 }, { "epoch": 3.6452402316665307, "grad_norm": 0.04182480648159981, "learning_rate": 2.0423520221225947e-05, "loss": 0.0621, "num_input_tokens_seen": 43020128, "step": 22345 }, { "epoch": 3.6460559588873482, "grad_norm": 0.15893839299678802, "learning_rate": 2.0413868994081848e-05, "loss": 0.0013, "num_input_tokens_seen": 43029712, "step": 22350 }, { "epoch": 3.6468716861081654, "grad_norm": 0.04383135959506035, "learning_rate": 2.0404218474176795e-05, "loss": 0.0053, "num_input_tokens_seen": 43039024, "step": 22355 }, { "epoch": 3.647687413328983, "grad_norm": 0.1591237187385559, "learning_rate": 2.0394568662999002e-05, "loss": 0.0105, "num_input_tokens_seen": 43048816, "step": 22360 }, { "epoch": 3.6485031405498, "grad_norm": 0.21163463592529297, "learning_rate": 2.0384919562036593e-05, "loss": 0.0187, "num_input_tokens_seen": 43059136, "step": 22365 }, { "epoch": 3.6493188677706176, "grad_norm": 0.027326136827468872, "learning_rate": 2.0375271172777593e-05, "loss": 0.0012, "num_input_tokens_seen": 43067904, "step": 22370 }, { "epoch": 3.6501345949914348, "grad_norm": 0.008531281724572182, "learning_rate": 2.0365623496709885e-05, "loss": 0.0017, "num_input_tokens_seen": 43077776, "step": 22375 }, { "epoch": 3.6509503222122524, "grad_norm": 0.5450990796089172, "learning_rate": 2.0355976535321283e-05, "loss": 0.0052, "num_input_tokens_seen": 43087824, "step": 22380 }, { "epoch": 3.6517660494330695, "grad_norm": 0.017348196357488632, "learning_rate": 2.034633029009945e-05, "loss": 0.0052, "num_input_tokens_seen": 43097760, "step": 22385 }, { "epoch": 3.652581776653887, "grad_norm": 4.416871070861816, "learning_rate": 2.0336684762531972e-05, "loss": 0.0965, "num_input_tokens_seen": 43107264, "step": 22390 }, { "epoch": 3.653397503874704, "grad_norm": 0.04651743918657303, "learning_rate": 2.032703995410631e-05, "loss": 0.0172, "num_input_tokens_seen": 43115984, "step": 22395 }, { "epoch": 3.6542132310955218, "grad_norm": 0.03740594536066055, "learning_rate": 2.031739586630981e-05, "loss": 0.0009, "num_input_tokens_seen": 43126400, "step": 22400 }, { "epoch": 3.6542132310955218, "eval_loss": 0.25593921542167664, "eval_runtime": 80.8902, "eval_samples_per_second": 33.688, "eval_steps_per_second": 16.85, "num_input_tokens_seen": 43126400, "step": 22400 }, { "epoch": 3.655028958316339, "grad_norm": 0.023683616891503334, "learning_rate": 2.0307752500629707e-05, "loss": 0.045, "num_input_tokens_seen": 43136672, "step": 22405 }, { "epoch": 3.6558446855371565, "grad_norm": 2.9350478649139404, "learning_rate": 2.0298109858553144e-05, "loss": 0.089, "num_input_tokens_seen": 43145216, "step": 22410 }, { "epoch": 3.6566604127579736, "grad_norm": 0.017051000148057938, "learning_rate": 2.028846794156712e-05, "loss": 0.0026, "num_input_tokens_seen": 43154480, "step": 22415 }, { "epoch": 3.657476139978791, "grad_norm": 0.00798528641462326, "learning_rate": 2.027882675115856e-05, "loss": 0.0238, "num_input_tokens_seen": 43164496, "step": 22420 }, { "epoch": 3.6582918671996083, "grad_norm": 0.2959352135658264, "learning_rate": 2.026918628881423e-05, "loss": 0.0253, "num_input_tokens_seen": 43174192, "step": 22425 }, { "epoch": 3.659107594420426, "grad_norm": 0.09763965755701065, "learning_rate": 2.0259546556020833e-05, "loss": 0.0619, "num_input_tokens_seen": 43184688, "step": 22430 }, { "epoch": 3.659923321641243, "grad_norm": 0.08261463046073914, "learning_rate": 2.024990755426493e-05, "loss": 0.1641, "num_input_tokens_seen": 43194448, "step": 22435 }, { "epoch": 3.6607390488620606, "grad_norm": 0.03180404007434845, "learning_rate": 2.0240269285032975e-05, "loss": 0.0006, "num_input_tokens_seen": 43205184, "step": 22440 }, { "epoch": 3.6615547760828777, "grad_norm": 0.11323186755180359, "learning_rate": 2.0230631749811306e-05, "loss": 0.0904, "num_input_tokens_seen": 43214528, "step": 22445 }, { "epoch": 3.6623705033036953, "grad_norm": 3.2831404209136963, "learning_rate": 2.0220994950086162e-05, "loss": 0.103, "num_input_tokens_seen": 43223808, "step": 22450 }, { "epoch": 3.663186230524513, "grad_norm": 0.04886673390865326, "learning_rate": 2.021135888734365e-05, "loss": 0.0417, "num_input_tokens_seen": 43233712, "step": 22455 }, { "epoch": 3.66400195774533, "grad_norm": 0.015315530821681023, "learning_rate": 2.0201723563069783e-05, "loss": 0.001, "num_input_tokens_seen": 43243008, "step": 22460 }, { "epoch": 3.664817684966147, "grad_norm": 0.036480970680713654, "learning_rate": 2.0192088978750433e-05, "loss": 0.0694, "num_input_tokens_seen": 43253344, "step": 22465 }, { "epoch": 3.6656334121869647, "grad_norm": 0.0048008812591433525, "learning_rate": 2.0182455135871385e-05, "loss": 0.0007, "num_input_tokens_seen": 43261904, "step": 22470 }, { "epoch": 3.6664491394077823, "grad_norm": 11.611413955688477, "learning_rate": 2.0172822035918305e-05, "loss": 0.0587, "num_input_tokens_seen": 43271040, "step": 22475 }, { "epoch": 3.6672648666285994, "grad_norm": 0.05632075294852257, "learning_rate": 2.016318968037671e-05, "loss": 0.0023, "num_input_tokens_seen": 43281248, "step": 22480 }, { "epoch": 3.6680805938494165, "grad_norm": 0.038459353148937225, "learning_rate": 2.015355807073206e-05, "loss": 0.01, "num_input_tokens_seen": 43289312, "step": 22485 }, { "epoch": 3.668896321070234, "grad_norm": 0.037103693932294846, "learning_rate": 2.0143927208469664e-05, "loss": 0.1259, "num_input_tokens_seen": 43300016, "step": 22490 }, { "epoch": 3.6697120482910517, "grad_norm": 0.053365226835012436, "learning_rate": 2.0134297095074708e-05, "loss": 0.0021, "num_input_tokens_seen": 43310560, "step": 22495 }, { "epoch": 3.670527775511869, "grad_norm": 0.07449490576982498, "learning_rate": 2.0124667732032297e-05, "loss": 0.0684, "num_input_tokens_seen": 43320272, "step": 22500 }, { "epoch": 3.671343502732686, "grad_norm": 0.045307457447052, "learning_rate": 2.011503912082738e-05, "loss": 0.0013, "num_input_tokens_seen": 43329264, "step": 22505 }, { "epoch": 3.6721592299535035, "grad_norm": 0.01135691162198782, "learning_rate": 2.0105411262944823e-05, "loss": 0.0796, "num_input_tokens_seen": 43339136, "step": 22510 }, { "epoch": 3.672974957174321, "grad_norm": 0.015987925231456757, "learning_rate": 2.0095784159869366e-05, "loss": 0.0328, "num_input_tokens_seen": 43348160, "step": 22515 }, { "epoch": 3.6737906843951382, "grad_norm": 0.02874588593840599, "learning_rate": 2.0086157813085608e-05, "loss": 0.001, "num_input_tokens_seen": 43356848, "step": 22520 }, { "epoch": 3.6746064116159554, "grad_norm": 0.27572062611579895, "learning_rate": 2.0076532224078068e-05, "loss": 0.0069, "num_input_tokens_seen": 43366944, "step": 22525 }, { "epoch": 3.675422138836773, "grad_norm": 0.1491815745830536, "learning_rate": 2.0066907394331142e-05, "loss": 0.1037, "num_input_tokens_seen": 43377024, "step": 22530 }, { "epoch": 3.6762378660575905, "grad_norm": 0.030195435509085655, "learning_rate": 2.0057283325329077e-05, "loss": 0.0041, "num_input_tokens_seen": 43386976, "step": 22535 }, { "epoch": 3.6770535932784076, "grad_norm": 0.012974049896001816, "learning_rate": 2.0047660018556047e-05, "loss": 0.0681, "num_input_tokens_seen": 43396496, "step": 22540 }, { "epoch": 3.677869320499225, "grad_norm": 8.084233283996582, "learning_rate": 2.0038037475496075e-05, "loss": 0.1047, "num_input_tokens_seen": 43406592, "step": 22545 }, { "epoch": 3.6786850477200423, "grad_norm": 0.019553935155272484, "learning_rate": 2.0028415697633073e-05, "loss": 0.1597, "num_input_tokens_seen": 43415984, "step": 22550 }, { "epoch": 3.67950077494086, "grad_norm": 9.648763656616211, "learning_rate": 2.0018794686450858e-05, "loss": 0.0872, "num_input_tokens_seen": 43423888, "step": 22555 }, { "epoch": 3.680316502161677, "grad_norm": 0.02670772559940815, "learning_rate": 2.0009174443433088e-05, "loss": 0.0004, "num_input_tokens_seen": 43434592, "step": 22560 }, { "epoch": 3.6811322293824946, "grad_norm": 12.911754608154297, "learning_rate": 1.999955497006334e-05, "loss": 0.0864, "num_input_tokens_seen": 43443968, "step": 22565 }, { "epoch": 3.6819479566033118, "grad_norm": 0.9865021109580994, "learning_rate": 1.9989936267825067e-05, "loss": 0.0816, "num_input_tokens_seen": 43453488, "step": 22570 }, { "epoch": 3.6827636838241293, "grad_norm": 4.7919182777404785, "learning_rate": 1.9980318338201572e-05, "loss": 0.1659, "num_input_tokens_seen": 43462304, "step": 22575 }, { "epoch": 3.6835794110449465, "grad_norm": 0.05729639157652855, "learning_rate": 1.997070118267607e-05, "loss": 0.0027, "num_input_tokens_seen": 43473376, "step": 22580 }, { "epoch": 3.684395138265764, "grad_norm": 3.837707996368408, "learning_rate": 1.9961084802731654e-05, "loss": 0.0714, "num_input_tokens_seen": 43483776, "step": 22585 }, { "epoch": 3.685210865486581, "grad_norm": 0.10268513113260269, "learning_rate": 1.9951469199851273e-05, "loss": 0.0019, "num_input_tokens_seen": 43494208, "step": 22590 }, { "epoch": 3.6860265927073987, "grad_norm": 0.05837603658437729, "learning_rate": 1.99418543755178e-05, "loss": 0.0505, "num_input_tokens_seen": 43504736, "step": 22595 }, { "epoch": 3.686842319928216, "grad_norm": 4.731760025024414, "learning_rate": 1.9932240331213936e-05, "loss": 0.1059, "num_input_tokens_seen": 43513248, "step": 22600 }, { "epoch": 3.686842319928216, "eval_loss": 0.22804485261440277, "eval_runtime": 80.8484, "eval_samples_per_second": 33.705, "eval_steps_per_second": 16.859, "num_input_tokens_seen": 43513248, "step": 22600 }, { "epoch": 3.6876580471490334, "grad_norm": 0.013526206836104393, "learning_rate": 1.9922627068422297e-05, "loss": 0.001, "num_input_tokens_seen": 43521248, "step": 22605 }, { "epoch": 3.6884737743698506, "grad_norm": 0.16036534309387207, "learning_rate": 1.991301458862538e-05, "loss": 0.0669, "num_input_tokens_seen": 43532192, "step": 22610 }, { "epoch": 3.689289501590668, "grad_norm": 0.07746357470750809, "learning_rate": 1.9903402893305536e-05, "loss": 0.0282, "num_input_tokens_seen": 43541424, "step": 22615 }, { "epoch": 3.6901052288114853, "grad_norm": 0.26691165566444397, "learning_rate": 1.9893791983945016e-05, "loss": 0.0053, "num_input_tokens_seen": 43552416, "step": 22620 }, { "epoch": 3.690920956032303, "grad_norm": 0.016410479322075844, "learning_rate": 1.988418186202594e-05, "loss": 0.0027, "num_input_tokens_seen": 43562176, "step": 22625 }, { "epoch": 3.6917366832531204, "grad_norm": 0.046454448252916336, "learning_rate": 1.98745725290303e-05, "loss": 0.0682, "num_input_tokens_seen": 43572944, "step": 22630 }, { "epoch": 3.6925524104739376, "grad_norm": 0.3450881838798523, "learning_rate": 1.986496398644e-05, "loss": 0.1211, "num_input_tokens_seen": 43581360, "step": 22635 }, { "epoch": 3.6933681376947547, "grad_norm": 0.04249950870871544, "learning_rate": 1.9855356235736777e-05, "loss": 0.0065, "num_input_tokens_seen": 43592064, "step": 22640 }, { "epoch": 3.6941838649155723, "grad_norm": 0.0431499145925045, "learning_rate": 1.9845749278402277e-05, "loss": 0.0499, "num_input_tokens_seen": 43602064, "step": 22645 }, { "epoch": 3.69499959213639, "grad_norm": 5.6172871589660645, "learning_rate": 1.9836143115918006e-05, "loss": 0.0443, "num_input_tokens_seen": 43610592, "step": 22650 }, { "epoch": 3.695815319357207, "grad_norm": 0.8257471323013306, "learning_rate": 1.9826537749765367e-05, "loss": 0.0684, "num_input_tokens_seen": 43619968, "step": 22655 }, { "epoch": 3.696631046578024, "grad_norm": 3.746844530105591, "learning_rate": 1.9816933181425625e-05, "loss": 0.0977, "num_input_tokens_seen": 43629616, "step": 22660 }, { "epoch": 3.6974467737988417, "grad_norm": 3.2795908451080322, "learning_rate": 1.9807329412379903e-05, "loss": 0.0256, "num_input_tokens_seen": 43639600, "step": 22665 }, { "epoch": 3.6982625010196593, "grad_norm": 4.6556010246276855, "learning_rate": 1.9797726444109247e-05, "loss": 0.0067, "num_input_tokens_seen": 43649568, "step": 22670 }, { "epoch": 3.6990782282404764, "grad_norm": 43.83822250366211, "learning_rate": 1.9788124278094557e-05, "loss": 0.0199, "num_input_tokens_seen": 43658368, "step": 22675 }, { "epoch": 3.6998939554612935, "grad_norm": 0.057920824736356735, "learning_rate": 1.9778522915816594e-05, "loss": 0.0012, "num_input_tokens_seen": 43667888, "step": 22680 }, { "epoch": 3.700709682682111, "grad_norm": 5.699970245361328, "learning_rate": 1.9768922358756014e-05, "loss": 0.1376, "num_input_tokens_seen": 43677776, "step": 22685 }, { "epoch": 3.7015254099029287, "grad_norm": 0.05185824632644653, "learning_rate": 1.9759322608393353e-05, "loss": 0.0925, "num_input_tokens_seen": 43686608, "step": 22690 }, { "epoch": 3.702341137123746, "grad_norm": 2.7139625549316406, "learning_rate": 1.9749723666208992e-05, "loss": 0.1578, "num_input_tokens_seen": 43696880, "step": 22695 }, { "epoch": 3.703156864344563, "grad_norm": 0.058233097195625305, "learning_rate": 1.9740125533683235e-05, "loss": 0.1091, "num_input_tokens_seen": 43707040, "step": 22700 }, { "epoch": 3.7039725915653805, "grad_norm": 36.30372619628906, "learning_rate": 1.9730528212296208e-05, "loss": 0.0966, "num_input_tokens_seen": 43717568, "step": 22705 }, { "epoch": 3.704788318786198, "grad_norm": 37.402565002441406, "learning_rate": 1.9720931703527945e-05, "loss": 0.0849, "num_input_tokens_seen": 43726800, "step": 22710 }, { "epoch": 3.705604046007015, "grad_norm": 0.24851779639720917, "learning_rate": 1.9711336008858373e-05, "loss": 0.059, "num_input_tokens_seen": 43736496, "step": 22715 }, { "epoch": 3.7064197732278323, "grad_norm": 0.020434541627764702, "learning_rate": 1.9701741129767233e-05, "loss": 0.0626, "num_input_tokens_seen": 43746816, "step": 22720 }, { "epoch": 3.70723550044865, "grad_norm": 0.033518120646476746, "learning_rate": 1.9692147067734202e-05, "loss": 0.1484, "num_input_tokens_seen": 43755696, "step": 22725 }, { "epoch": 3.7080512276694675, "grad_norm": 8.61585807800293, "learning_rate": 1.96825538242388e-05, "loss": 0.0429, "num_input_tokens_seen": 43765600, "step": 22730 }, { "epoch": 3.7088669548902846, "grad_norm": 0.16771802306175232, "learning_rate": 1.967296140076041e-05, "loss": 0.0088, "num_input_tokens_seen": 43774832, "step": 22735 }, { "epoch": 3.709682682111102, "grad_norm": 0.9769271612167358, "learning_rate": 1.966336979877833e-05, "loss": 0.0041, "num_input_tokens_seen": 43784912, "step": 22740 }, { "epoch": 3.7104984093319193, "grad_norm": 6.89987850189209, "learning_rate": 1.9653779019771678e-05, "loss": 0.0504, "num_input_tokens_seen": 43793024, "step": 22745 }, { "epoch": 3.711314136552737, "grad_norm": 0.28927120566368103, "learning_rate": 1.9644189065219488e-05, "loss": 0.0267, "num_input_tokens_seen": 43802896, "step": 22750 }, { "epoch": 3.712129863773554, "grad_norm": 0.03530287370085716, "learning_rate": 1.9634599936600655e-05, "loss": 0.0952, "num_input_tokens_seen": 43812032, "step": 22755 }, { "epoch": 3.7129455909943716, "grad_norm": 0.009491709992289543, "learning_rate": 1.9625011635393935e-05, "loss": 0.0015, "num_input_tokens_seen": 43820592, "step": 22760 }, { "epoch": 3.7137613182151887, "grad_norm": 16.93254280090332, "learning_rate": 1.9615424163077963e-05, "loss": 0.0492, "num_input_tokens_seen": 43829648, "step": 22765 }, { "epoch": 3.7145770454360063, "grad_norm": 0.08667849749326706, "learning_rate": 1.9605837521131263e-05, "loss": 0.0352, "num_input_tokens_seen": 43839184, "step": 22770 }, { "epoch": 3.7153927726568234, "grad_norm": 0.039439696818590164, "learning_rate": 1.9596251711032192e-05, "loss": 0.0028, "num_input_tokens_seen": 43849488, "step": 22775 }, { "epoch": 3.716208499877641, "grad_norm": 0.020007571205496788, "learning_rate": 1.958666673425903e-05, "loss": 0.0012, "num_input_tokens_seen": 43858128, "step": 22780 }, { "epoch": 3.717024227098458, "grad_norm": 1.390222430229187, "learning_rate": 1.957708259228987e-05, "loss": 0.0059, "num_input_tokens_seen": 43869136, "step": 22785 }, { "epoch": 3.7178399543192757, "grad_norm": 6.649351596832275, "learning_rate": 1.956749928660273e-05, "loss": 0.034, "num_input_tokens_seen": 43877680, "step": 22790 }, { "epoch": 3.718655681540093, "grad_norm": 0.12640734016895294, "learning_rate": 1.955791681867547e-05, "loss": 0.0015, "num_input_tokens_seen": 43886960, "step": 22795 }, { "epoch": 3.7194714087609104, "grad_norm": 13.25074577331543, "learning_rate": 1.9548335189985824e-05, "loss": 0.1354, "num_input_tokens_seen": 43896720, "step": 22800 }, { "epoch": 3.7194714087609104, "eval_loss": 0.22966504096984863, "eval_runtime": 80.7879, "eval_samples_per_second": 33.73, "eval_steps_per_second": 16.871, "num_input_tokens_seen": 43896720, "step": 22800 }, { "epoch": 3.7202871359817276, "grad_norm": 0.3494199812412262, "learning_rate": 1.9538754402011396e-05, "loss": 0.0888, "num_input_tokens_seen": 43906848, "step": 22805 }, { "epoch": 3.721102863202545, "grad_norm": 0.4520113468170166, "learning_rate": 1.952917445622968e-05, "loss": 0.2289, "num_input_tokens_seen": 43916304, "step": 22810 }, { "epoch": 3.7219185904233623, "grad_norm": 29.169822692871094, "learning_rate": 1.9519595354118005e-05, "loss": 0.0678, "num_input_tokens_seen": 43925808, "step": 22815 }, { "epoch": 3.72273431764418, "grad_norm": 0.0465911403298378, "learning_rate": 1.951001709715361e-05, "loss": 0.1333, "num_input_tokens_seen": 43935408, "step": 22820 }, { "epoch": 3.7235500448649974, "grad_norm": 0.036910656839609146, "learning_rate": 1.9500439686813556e-05, "loss": 0.0009, "num_input_tokens_seen": 43945712, "step": 22825 }, { "epoch": 3.7243657720858145, "grad_norm": 0.05563516169786453, "learning_rate": 1.949086312457482e-05, "loss": 0.0929, "num_input_tokens_seen": 43956832, "step": 22830 }, { "epoch": 3.7251814993066317, "grad_norm": 1.110138177871704, "learning_rate": 1.9481287411914223e-05, "loss": 0.0079, "num_input_tokens_seen": 43966512, "step": 22835 }, { "epoch": 3.7259972265274492, "grad_norm": 6.751047611236572, "learning_rate": 1.9471712550308457e-05, "loss": 0.0087, "num_input_tokens_seen": 43977152, "step": 22840 }, { "epoch": 3.726812953748267, "grad_norm": 0.00920822098851204, "learning_rate": 1.946213854123409e-05, "loss": 0.1391, "num_input_tokens_seen": 43986592, "step": 22845 }, { "epoch": 3.727628680969084, "grad_norm": 2.144606351852417, "learning_rate": 1.9452565386167554e-05, "loss": 0.0025, "num_input_tokens_seen": 43996080, "step": 22850 }, { "epoch": 3.728444408189901, "grad_norm": 0.01596212014555931, "learning_rate": 1.9442993086585142e-05, "loss": 0.0966, "num_input_tokens_seen": 44005232, "step": 22855 }, { "epoch": 3.7292601354107187, "grad_norm": 0.025731945410370827, "learning_rate": 1.9433421643963043e-05, "loss": 0.0295, "num_input_tokens_seen": 44014784, "step": 22860 }, { "epoch": 3.7300758626315362, "grad_norm": 4.265120029449463, "learning_rate": 1.942385105977727e-05, "loss": 0.0734, "num_input_tokens_seen": 44025248, "step": 22865 }, { "epoch": 3.7308915898523534, "grad_norm": 5.352970123291016, "learning_rate": 1.9414281335503743e-05, "loss": 0.2226, "num_input_tokens_seen": 44035328, "step": 22870 }, { "epoch": 3.7317073170731705, "grad_norm": 0.03910936042666435, "learning_rate": 1.9404712472618232e-05, "loss": 0.1527, "num_input_tokens_seen": 44044192, "step": 22875 }, { "epoch": 3.732523044293988, "grad_norm": 0.21162573993206024, "learning_rate": 1.939514447259636e-05, "loss": 0.1022, "num_input_tokens_seen": 44053712, "step": 22880 }, { "epoch": 3.7333387715148056, "grad_norm": 0.04750150442123413, "learning_rate": 1.938557733691365e-05, "loss": 0.0026, "num_input_tokens_seen": 44063280, "step": 22885 }, { "epoch": 3.7341544987356228, "grad_norm": 0.04778676852583885, "learning_rate": 1.9376011067045476e-05, "loss": 0.0043, "num_input_tokens_seen": 44072320, "step": 22890 }, { "epoch": 3.73497022595644, "grad_norm": 0.16358153522014618, "learning_rate": 1.9366445664467065e-05, "loss": 0.0271, "num_input_tokens_seen": 44080848, "step": 22895 }, { "epoch": 3.7357859531772575, "grad_norm": 0.09779050946235657, "learning_rate": 1.9356881130653533e-05, "loss": 0.1796, "num_input_tokens_seen": 44090080, "step": 22900 }, { "epoch": 3.736601680398075, "grad_norm": 15.088316917419434, "learning_rate": 1.9347317467079846e-05, "loss": 0.0101, "num_input_tokens_seen": 44099408, "step": 22905 }, { "epoch": 3.737417407618892, "grad_norm": 0.045450735837221146, "learning_rate": 1.9337754675220836e-05, "loss": 0.0847, "num_input_tokens_seen": 44108256, "step": 22910 }, { "epoch": 3.7382331348397098, "grad_norm": 0.04258306324481964, "learning_rate": 1.9328192756551218e-05, "loss": 0.064, "num_input_tokens_seen": 44117200, "step": 22915 }, { "epoch": 3.739048862060527, "grad_norm": 1.416707992553711, "learning_rate": 1.931863171254555e-05, "loss": 0.0035, "num_input_tokens_seen": 44127408, "step": 22920 }, { "epoch": 3.7398645892813445, "grad_norm": 3.1665923595428467, "learning_rate": 1.930907154467826e-05, "loss": 0.0068, "num_input_tokens_seen": 44137136, "step": 22925 }, { "epoch": 3.7406803165021616, "grad_norm": 8.633514404296875, "learning_rate": 1.9299512254423673e-05, "loss": 0.0796, "num_input_tokens_seen": 44144944, "step": 22930 }, { "epoch": 3.741496043722979, "grad_norm": 0.14733202755451202, "learning_rate": 1.9289953843255914e-05, "loss": 0.0022, "num_input_tokens_seen": 44154656, "step": 22935 }, { "epoch": 3.7423117709437963, "grad_norm": 0.016261309385299683, "learning_rate": 1.9280396312649048e-05, "loss": 0.0661, "num_input_tokens_seen": 44164416, "step": 22940 }, { "epoch": 3.743127498164614, "grad_norm": 0.09380444884300232, "learning_rate": 1.9270839664076936e-05, "loss": 0.0013, "num_input_tokens_seen": 44173232, "step": 22945 }, { "epoch": 3.743943225385431, "grad_norm": 0.2254842221736908, "learning_rate": 1.9261283899013345e-05, "loss": 0.0032, "num_input_tokens_seen": 44182592, "step": 22950 }, { "epoch": 3.7447589526062486, "grad_norm": 0.06933482736349106, "learning_rate": 1.92517290189319e-05, "loss": 0.0014, "num_input_tokens_seen": 44191712, "step": 22955 }, { "epoch": 3.7455746798270657, "grad_norm": 0.5235021710395813, "learning_rate": 1.924217502530607e-05, "loss": 0.0794, "num_input_tokens_seen": 44201168, "step": 22960 }, { "epoch": 3.7463904070478833, "grad_norm": 0.03614088147878647, "learning_rate": 1.9232621919609207e-05, "loss": 0.0019, "num_input_tokens_seen": 44211632, "step": 22965 }, { "epoch": 3.7472061342687004, "grad_norm": 0.09564150869846344, "learning_rate": 1.9223069703314534e-05, "loss": 0.0318, "num_input_tokens_seen": 44221024, "step": 22970 }, { "epoch": 3.748021861489518, "grad_norm": 0.05535630136728287, "learning_rate": 1.92135183778951e-05, "loss": 0.0289, "num_input_tokens_seen": 44231280, "step": 22975 }, { "epoch": 3.748837588710335, "grad_norm": 0.045300986617803574, "learning_rate": 1.9203967944823857e-05, "loss": 0.0562, "num_input_tokens_seen": 44240960, "step": 22980 }, { "epoch": 3.7496533159311527, "grad_norm": 0.13363218307495117, "learning_rate": 1.9194418405573588e-05, "loss": 0.0544, "num_input_tokens_seen": 44249936, "step": 22985 }, { "epoch": 3.75046904315197, "grad_norm": 0.06751228868961334, "learning_rate": 1.9184869761616954e-05, "loss": 0.0584, "num_input_tokens_seen": 44259296, "step": 22990 }, { "epoch": 3.7512847703727874, "grad_norm": 0.06504711508750916, "learning_rate": 1.9175322014426495e-05, "loss": 0.0978, "num_input_tokens_seen": 44268800, "step": 22995 }, { "epoch": 3.7521004975936045, "grad_norm": 5.310547828674316, "learning_rate": 1.9165775165474565e-05, "loss": 0.1154, "num_input_tokens_seen": 44278640, "step": 23000 }, { "epoch": 3.7521004975936045, "eval_loss": 0.2092617154121399, "eval_runtime": 80.7969, "eval_samples_per_second": 33.727, "eval_steps_per_second": 16.869, "num_input_tokens_seen": 44278640, "step": 23000 }, { "epoch": 3.752916224814422, "grad_norm": 0.02113150618970394, "learning_rate": 1.9156229216233434e-05, "loss": 0.0021, "num_input_tokens_seen": 44287952, "step": 23005 }, { "epoch": 3.7537319520352392, "grad_norm": 0.23687447607517242, "learning_rate": 1.9146684168175184e-05, "loss": 0.0021, "num_input_tokens_seen": 44298192, "step": 23010 }, { "epoch": 3.754547679256057, "grad_norm": 0.06389934569597244, "learning_rate": 1.9137140022771796e-05, "loss": 0.0031, "num_input_tokens_seen": 44308640, "step": 23015 }, { "epoch": 3.7553634064768744, "grad_norm": 0.04574432969093323, "learning_rate": 1.9127596781495103e-05, "loss": 0.002, "num_input_tokens_seen": 44319088, "step": 23020 }, { "epoch": 3.7561791336976915, "grad_norm": 0.1004398763179779, "learning_rate": 1.9118054445816767e-05, "loss": 0.0091, "num_input_tokens_seen": 44328336, "step": 23025 }, { "epoch": 3.7569948609185086, "grad_norm": 0.10710179060697556, "learning_rate": 1.9108513017208356e-05, "loss": 0.0566, "num_input_tokens_seen": 44338176, "step": 23030 }, { "epoch": 3.7578105881393262, "grad_norm": 0.07521402090787888, "learning_rate": 1.9098972497141287e-05, "loss": 0.1442, "num_input_tokens_seen": 44347552, "step": 23035 }, { "epoch": 3.758626315360144, "grad_norm": 0.19343921542167664, "learning_rate": 1.9089432887086806e-05, "loss": 0.1095, "num_input_tokens_seen": 44356560, "step": 23040 }, { "epoch": 3.759442042580961, "grad_norm": 3.101517677307129, "learning_rate": 1.9079894188516056e-05, "loss": 0.1818, "num_input_tokens_seen": 44366560, "step": 23045 }, { "epoch": 3.760257769801778, "grad_norm": 0.028560791164636612, "learning_rate": 1.907035640290002e-05, "loss": 0.0993, "num_input_tokens_seen": 44375264, "step": 23050 }, { "epoch": 3.7610734970225956, "grad_norm": 0.16470295190811157, "learning_rate": 1.9060819531709534e-05, "loss": 0.0922, "num_input_tokens_seen": 44385968, "step": 23055 }, { "epoch": 3.761889224243413, "grad_norm": 0.020717918872833252, "learning_rate": 1.9051283576415325e-05, "loss": 0.0033, "num_input_tokens_seen": 44394208, "step": 23060 }, { "epoch": 3.7627049514642303, "grad_norm": 0.04383789002895355, "learning_rate": 1.904174853848793e-05, "loss": 0.0124, "num_input_tokens_seen": 44403808, "step": 23065 }, { "epoch": 3.7635206786850475, "grad_norm": 0.03629613667726517, "learning_rate": 1.903221441939779e-05, "loss": 0.0032, "num_input_tokens_seen": 44413728, "step": 23070 }, { "epoch": 3.764336405905865, "grad_norm": 0.16863180696964264, "learning_rate": 1.9022681220615194e-05, "loss": 0.0487, "num_input_tokens_seen": 44424256, "step": 23075 }, { "epoch": 3.7651521331266826, "grad_norm": 0.09500498324632645, "learning_rate": 1.9013148943610255e-05, "loss": 0.0017, "num_input_tokens_seen": 44434592, "step": 23080 }, { "epoch": 3.7659678603474998, "grad_norm": 0.34787318110466003, "learning_rate": 1.9003617589852998e-05, "loss": 0.0046, "num_input_tokens_seen": 44443824, "step": 23085 }, { "epoch": 3.766783587568317, "grad_norm": 0.060417868196964264, "learning_rate": 1.899408716081326e-05, "loss": 0.0836, "num_input_tokens_seen": 44453616, "step": 23090 }, { "epoch": 3.7675993147891345, "grad_norm": 0.06861728429794312, "learning_rate": 1.898455765796075e-05, "loss": 0.0026, "num_input_tokens_seen": 44462544, "step": 23095 }, { "epoch": 3.768415042009952, "grad_norm": 0.06725306808948517, "learning_rate": 1.8975029082765053e-05, "loss": 0.0575, "num_input_tokens_seen": 44473872, "step": 23100 }, { "epoch": 3.769230769230769, "grad_norm": 0.036710649728775024, "learning_rate": 1.8965501436695577e-05, "loss": 0.0013, "num_input_tokens_seen": 44483664, "step": 23105 }, { "epoch": 3.7700464964515867, "grad_norm": 9.445686340332031, "learning_rate": 1.895597472122161e-05, "loss": 0.1022, "num_input_tokens_seen": 44493120, "step": 23110 }, { "epoch": 3.770862223672404, "grad_norm": 0.050272710621356964, "learning_rate": 1.894644893781231e-05, "loss": 0.1499, "num_input_tokens_seen": 44502048, "step": 23115 }, { "epoch": 3.7716779508932214, "grad_norm": 4.240086555480957, "learning_rate": 1.893692408793665e-05, "loss": 0.0624, "num_input_tokens_seen": 44511920, "step": 23120 }, { "epoch": 3.7724936781140386, "grad_norm": 4.232926368713379, "learning_rate": 1.8927400173063493e-05, "loss": 0.1807, "num_input_tokens_seen": 44521936, "step": 23125 }, { "epoch": 3.773309405334856, "grad_norm": 0.07446613907814026, "learning_rate": 1.891787719466154e-05, "loss": 0.0124, "num_input_tokens_seen": 44531840, "step": 23130 }, { "epoch": 3.7741251325556733, "grad_norm": 0.003991317935287952, "learning_rate": 1.8908355154199346e-05, "loss": 0.0593, "num_input_tokens_seen": 44540224, "step": 23135 }, { "epoch": 3.774940859776491, "grad_norm": 0.16536380350589752, "learning_rate": 1.8898834053145357e-05, "loss": 0.0048, "num_input_tokens_seen": 44549312, "step": 23140 }, { "epoch": 3.775756586997308, "grad_norm": 6.607301712036133, "learning_rate": 1.8889313892967813e-05, "loss": 0.0979, "num_input_tokens_seen": 44559664, "step": 23145 }, { "epoch": 3.7765723142181256, "grad_norm": 7.9340596199035645, "learning_rate": 1.8879794675134863e-05, "loss": 0.1041, "num_input_tokens_seen": 44568800, "step": 23150 }, { "epoch": 3.7773880414389427, "grad_norm": 0.028190642595291138, "learning_rate": 1.8870276401114494e-05, "loss": 0.0016, "num_input_tokens_seen": 44578896, "step": 23155 }, { "epoch": 3.7782037686597603, "grad_norm": 2.711960792541504, "learning_rate": 1.886075907237453e-05, "loss": 0.1231, "num_input_tokens_seen": 44589344, "step": 23160 }, { "epoch": 3.7790194958805774, "grad_norm": 0.02677481435239315, "learning_rate": 1.8851242690382672e-05, "loss": 0.0249, "num_input_tokens_seen": 44599744, "step": 23165 }, { "epoch": 3.779835223101395, "grad_norm": 0.05434103682637215, "learning_rate": 1.884172725660645e-05, "loss": 0.0012, "num_input_tokens_seen": 44609600, "step": 23170 }, { "epoch": 3.780650950322212, "grad_norm": 0.37132877111434937, "learning_rate": 1.8832212772513277e-05, "loss": 0.0747, "num_input_tokens_seen": 44619296, "step": 23175 }, { "epoch": 3.7814666775430297, "grad_norm": 5.9592719078063965, "learning_rate": 1.8822699239570414e-05, "loss": 0.1168, "num_input_tokens_seen": 44628928, "step": 23180 }, { "epoch": 3.782282404763847, "grad_norm": 0.26694056391716003, "learning_rate": 1.8813186659244943e-05, "loss": 0.0043, "num_input_tokens_seen": 44638016, "step": 23185 }, { "epoch": 3.7830981319846644, "grad_norm": 0.049465905874967575, "learning_rate": 1.880367503300385e-05, "loss": 0.0294, "num_input_tokens_seen": 44647520, "step": 23190 }, { "epoch": 3.783913859205482, "grad_norm": 0.5975068807601929, "learning_rate": 1.8794164362313927e-05, "loss": 0.0084, "num_input_tokens_seen": 44656784, "step": 23195 }, { "epoch": 3.784729586426299, "grad_norm": 5.607474327087402, "learning_rate": 1.878465464864185e-05, "loss": 0.0173, "num_input_tokens_seen": 44666464, "step": 23200 }, { "epoch": 3.784729586426299, "eval_loss": 0.21155737340450287, "eval_runtime": 80.8198, "eval_samples_per_second": 33.717, "eval_steps_per_second": 16.865, "num_input_tokens_seen": 44666464, "step": 23200 }, { "epoch": 3.785545313647116, "grad_norm": 0.06131801754236221, "learning_rate": 1.877514589345414e-05, "loss": 0.0478, "num_input_tokens_seen": 44676944, "step": 23205 }, { "epoch": 3.786361040867934, "grad_norm": 0.09804241359233856, "learning_rate": 1.876563809821715e-05, "loss": 0.0009, "num_input_tokens_seen": 44686912, "step": 23210 }, { "epoch": 3.7871767680887514, "grad_norm": 0.1453404575586319, "learning_rate": 1.8756131264397106e-05, "loss": 0.0022, "num_input_tokens_seen": 44696080, "step": 23215 }, { "epoch": 3.7879924953095685, "grad_norm": 0.012368873693048954, "learning_rate": 1.87466253934601e-05, "loss": 0.0013, "num_input_tokens_seen": 44705408, "step": 23220 }, { "epoch": 3.7888082225303856, "grad_norm": 15.653815269470215, "learning_rate": 1.8737120486872033e-05, "loss": 0.0786, "num_input_tokens_seen": 44715424, "step": 23225 }, { "epoch": 3.789623949751203, "grad_norm": 0.0649225115776062, "learning_rate": 1.8727616546098696e-05, "loss": 0.0824, "num_input_tokens_seen": 44725536, "step": 23230 }, { "epoch": 3.7904396769720208, "grad_norm": 0.9082272052764893, "learning_rate": 1.8718113572605716e-05, "loss": 0.0083, "num_input_tokens_seen": 44735840, "step": 23235 }, { "epoch": 3.791255404192838, "grad_norm": 0.016888197511434555, "learning_rate": 1.8708611567858554e-05, "loss": 0.0031, "num_input_tokens_seen": 44745792, "step": 23240 }, { "epoch": 3.792071131413655, "grad_norm": 7.102306365966797, "learning_rate": 1.8699110533322565e-05, "loss": 0.2173, "num_input_tokens_seen": 44755424, "step": 23245 }, { "epoch": 3.7928868586344726, "grad_norm": 0.3326719105243683, "learning_rate": 1.8689610470462897e-05, "loss": 0.0738, "num_input_tokens_seen": 44765280, "step": 23250 }, { "epoch": 3.79370258585529, "grad_norm": 0.05014032870531082, "learning_rate": 1.8680111380744604e-05, "loss": 0.0009, "num_input_tokens_seen": 44773760, "step": 23255 }, { "epoch": 3.7945183130761073, "grad_norm": 11.517789840698242, "learning_rate": 1.8670613265632564e-05, "loss": 0.0402, "num_input_tokens_seen": 44783280, "step": 23260 }, { "epoch": 3.7953340402969244, "grad_norm": 0.022273020818829536, "learning_rate": 1.866111612659149e-05, "loss": 0.0665, "num_input_tokens_seen": 44793040, "step": 23265 }, { "epoch": 3.796149767517742, "grad_norm": 0.4981158673763275, "learning_rate": 1.8651619965085967e-05, "loss": 0.0036, "num_input_tokens_seen": 44803072, "step": 23270 }, { "epoch": 3.7969654947385596, "grad_norm": 0.4318300783634186, "learning_rate": 1.8642124782580433e-05, "loss": 0.0031, "num_input_tokens_seen": 44812848, "step": 23275 }, { "epoch": 3.7977812219593767, "grad_norm": 0.06421075016260147, "learning_rate": 1.8632630580539144e-05, "loss": 0.0258, "num_input_tokens_seen": 44821664, "step": 23280 }, { "epoch": 3.7985969491801943, "grad_norm": 0.06524746119976044, "learning_rate": 1.862313736042625e-05, "loss": 0.0759, "num_input_tokens_seen": 44831312, "step": 23285 }, { "epoch": 3.7994126764010114, "grad_norm": 4.225825309753418, "learning_rate": 1.8613645123705703e-05, "loss": 0.0971, "num_input_tokens_seen": 44841072, "step": 23290 }, { "epoch": 3.800228403621829, "grad_norm": 0.02680872566998005, "learning_rate": 1.8604153871841328e-05, "loss": 0.0621, "num_input_tokens_seen": 44850896, "step": 23295 }, { "epoch": 3.801044130842646, "grad_norm": 40.73585510253906, "learning_rate": 1.859466360629682e-05, "loss": 0.0954, "num_input_tokens_seen": 44859344, "step": 23300 }, { "epoch": 3.8018598580634637, "grad_norm": 0.021868009120225906, "learning_rate": 1.8585174328535666e-05, "loss": 0.0127, "num_input_tokens_seen": 44868992, "step": 23305 }, { "epoch": 3.802675585284281, "grad_norm": 0.0999327078461647, "learning_rate": 1.857568604002124e-05, "loss": 0.0022, "num_input_tokens_seen": 44880224, "step": 23310 }, { "epoch": 3.8034913125050984, "grad_norm": 0.14319844543933868, "learning_rate": 1.8566198742216774e-05, "loss": 0.0902, "num_input_tokens_seen": 44888976, "step": 23315 }, { "epoch": 3.8043070397259156, "grad_norm": 0.044932007789611816, "learning_rate": 1.85567124365853e-05, "loss": 0.0617, "num_input_tokens_seen": 44898064, "step": 23320 }, { "epoch": 3.805122766946733, "grad_norm": 25.523941040039062, "learning_rate": 1.854722712458975e-05, "loss": 0.1426, "num_input_tokens_seen": 44908000, "step": 23325 }, { "epoch": 3.8059384941675503, "grad_norm": 0.05955538526177406, "learning_rate": 1.853774280769286e-05, "loss": 0.0018, "num_input_tokens_seen": 44916768, "step": 23330 }, { "epoch": 3.806754221388368, "grad_norm": 0.035582534968853, "learning_rate": 1.852825948735724e-05, "loss": 0.0015, "num_input_tokens_seen": 44926192, "step": 23335 }, { "epoch": 3.807569948609185, "grad_norm": 0.3003494143486023, "learning_rate": 1.851877716504534e-05, "loss": 0.003, "num_input_tokens_seen": 44935472, "step": 23340 }, { "epoch": 3.8083856758300025, "grad_norm": 3.374256134033203, "learning_rate": 1.8509295842219448e-05, "loss": 0.1638, "num_input_tokens_seen": 44944560, "step": 23345 }, { "epoch": 3.8092014030508197, "grad_norm": 0.329235315322876, "learning_rate": 1.8499815520341697e-05, "loss": 0.1135, "num_input_tokens_seen": 44952576, "step": 23350 }, { "epoch": 3.8100171302716372, "grad_norm": 0.1087404116988182, "learning_rate": 1.8490336200874094e-05, "loss": 0.0489, "num_input_tokens_seen": 44962432, "step": 23355 }, { "epoch": 3.8108328574924544, "grad_norm": 0.18872381746768951, "learning_rate": 1.848085788527844e-05, "loss": 0.0014, "num_input_tokens_seen": 44972480, "step": 23360 }, { "epoch": 3.811648584713272, "grad_norm": 1.336581826210022, "learning_rate": 1.847138057501644e-05, "loss": 0.1449, "num_input_tokens_seen": 44983040, "step": 23365 }, { "epoch": 3.812464311934089, "grad_norm": 0.12635627388954163, "learning_rate": 1.8461904271549582e-05, "loss": 0.0822, "num_input_tokens_seen": 44992304, "step": 23370 }, { "epoch": 3.8132800391549067, "grad_norm": 0.07586894929409027, "learning_rate": 1.845242897633926e-05, "loss": 0.1184, "num_input_tokens_seen": 45001552, "step": 23375 }, { "epoch": 3.814095766375724, "grad_norm": 0.2112853229045868, "learning_rate": 1.844295469084667e-05, "loss": 0.0324, "num_input_tokens_seen": 45011152, "step": 23380 }, { "epoch": 3.8149114935965414, "grad_norm": 34.936180114746094, "learning_rate": 1.843348141653286e-05, "loss": 0.0528, "num_input_tokens_seen": 45020496, "step": 23385 }, { "epoch": 3.815727220817359, "grad_norm": 0.6358641982078552, "learning_rate": 1.842400915485874e-05, "loss": 0.0489, "num_input_tokens_seen": 45029632, "step": 23390 }, { "epoch": 3.816542948038176, "grad_norm": 0.031414251774549484, "learning_rate": 1.8414537907285053e-05, "loss": 0.0679, "num_input_tokens_seen": 45038288, "step": 23395 }, { "epoch": 3.817358675258993, "grad_norm": 0.20551452040672302, "learning_rate": 1.840506767527237e-05, "loss": 0.0032, "num_input_tokens_seen": 45047360, "step": 23400 }, { "epoch": 3.817358675258993, "eval_loss": 0.20784857869148254, "eval_runtime": 80.7965, "eval_samples_per_second": 33.727, "eval_steps_per_second": 16.87, "num_input_tokens_seen": 45047360, "step": 23400 }, { "epoch": 3.8181744024798108, "grad_norm": 0.5667853355407715, "learning_rate": 1.8395598460281137e-05, "loss": 0.073, "num_input_tokens_seen": 45056672, "step": 23405 }, { "epoch": 3.8189901297006283, "grad_norm": 0.05175712704658508, "learning_rate": 1.838613026377161e-05, "loss": 0.0017, "num_input_tokens_seen": 45065712, "step": 23410 }, { "epoch": 3.8198058569214455, "grad_norm": 0.01168854534626007, "learning_rate": 1.8376663087203917e-05, "loss": 0.0013, "num_input_tokens_seen": 45074112, "step": 23415 }, { "epoch": 3.8206215841422626, "grad_norm": 0.1475382298231125, "learning_rate": 1.8367196932038014e-05, "loss": 0.0041, "num_input_tokens_seen": 45081824, "step": 23420 }, { "epoch": 3.82143731136308, "grad_norm": 0.11210223287343979, "learning_rate": 1.8357731799733686e-05, "loss": 0.0105, "num_input_tokens_seen": 45091424, "step": 23425 }, { "epoch": 3.8222530385838978, "grad_norm": 0.024502063170075417, "learning_rate": 1.8348267691750586e-05, "loss": 0.0038, "num_input_tokens_seen": 45101664, "step": 23430 }, { "epoch": 3.823068765804715, "grad_norm": 0.06320466101169586, "learning_rate": 1.833880460954821e-05, "loss": 0.0028, "num_input_tokens_seen": 45112192, "step": 23435 }, { "epoch": 3.823884493025532, "grad_norm": 34.04985809326172, "learning_rate": 1.8329342554585866e-05, "loss": 0.0566, "num_input_tokens_seen": 45123392, "step": 23440 }, { "epoch": 3.8247002202463496, "grad_norm": 0.11692275106906891, "learning_rate": 1.8319881528322735e-05, "loss": 0.0958, "num_input_tokens_seen": 45133584, "step": 23445 }, { "epoch": 3.825515947467167, "grad_norm": 0.053337763994932175, "learning_rate": 1.8310421532217815e-05, "loss": 0.019, "num_input_tokens_seen": 45143552, "step": 23450 }, { "epoch": 3.8263316746879843, "grad_norm": 54.32670211791992, "learning_rate": 1.8300962567729958e-05, "loss": 0.0221, "num_input_tokens_seen": 45152832, "step": 23455 }, { "epoch": 3.8271474019088014, "grad_norm": 0.08052916079759598, "learning_rate": 1.8291504636317866e-05, "loss": 0.0628, "num_input_tokens_seen": 45161776, "step": 23460 }, { "epoch": 3.827963129129619, "grad_norm": 0.0330791100859642, "learning_rate": 1.8282047739440055e-05, "loss": 0.0536, "num_input_tokens_seen": 45172352, "step": 23465 }, { "epoch": 3.8287788563504366, "grad_norm": 0.04474806785583496, "learning_rate": 1.8272591878554903e-05, "loss": 0.0043, "num_input_tokens_seen": 45182288, "step": 23470 }, { "epoch": 3.8295945835712537, "grad_norm": 0.01441435981541872, "learning_rate": 1.8263137055120638e-05, "loss": 0.1149, "num_input_tokens_seen": 45191664, "step": 23475 }, { "epoch": 3.8304103107920713, "grad_norm": 0.1292528361082077, "learning_rate": 1.8253683270595295e-05, "loss": 0.1386, "num_input_tokens_seen": 45200240, "step": 23480 }, { "epoch": 3.8312260380128884, "grad_norm": 0.030493363738059998, "learning_rate": 1.824423052643677e-05, "loss": 0.0038, "num_input_tokens_seen": 45209968, "step": 23485 }, { "epoch": 3.832041765233706, "grad_norm": 1.485381841659546, "learning_rate": 1.82347788241028e-05, "loss": 0.0688, "num_input_tokens_seen": 45219520, "step": 23490 }, { "epoch": 3.832857492454523, "grad_norm": 7.243000030517578, "learning_rate": 1.8225328165050942e-05, "loss": 0.1909, "num_input_tokens_seen": 45228832, "step": 23495 }, { "epoch": 3.8336732196753407, "grad_norm": 6.1635422706604, "learning_rate": 1.821587855073863e-05, "loss": 0.0799, "num_input_tokens_seen": 45238704, "step": 23500 }, { "epoch": 3.834488946896158, "grad_norm": 0.0960182249546051, "learning_rate": 1.8206429982623086e-05, "loss": 0.0119, "num_input_tokens_seen": 45248208, "step": 23505 }, { "epoch": 3.8353046741169754, "grad_norm": 9.497145652770996, "learning_rate": 1.8196982462161416e-05, "loss": 0.0698, "num_input_tokens_seen": 45257296, "step": 23510 }, { "epoch": 3.8361204013377925, "grad_norm": 0.010044160299003124, "learning_rate": 1.818753599081055e-05, "loss": 0.0097, "num_input_tokens_seen": 45266672, "step": 23515 }, { "epoch": 3.83693612855861, "grad_norm": 0.7595449090003967, "learning_rate": 1.817809057002724e-05, "loss": 0.0022, "num_input_tokens_seen": 45276528, "step": 23520 }, { "epoch": 3.8377518557794272, "grad_norm": 3.591895818710327, "learning_rate": 1.8168646201268096e-05, "loss": 0.0779, "num_input_tokens_seen": 45284384, "step": 23525 }, { "epoch": 3.838567583000245, "grad_norm": 0.007172089535742998, "learning_rate": 1.8159202885989557e-05, "loss": 0.0009, "num_input_tokens_seen": 45292896, "step": 23530 }, { "epoch": 3.839383310221062, "grad_norm": 0.13678300380706787, "learning_rate": 1.814976062564789e-05, "loss": 0.0157, "num_input_tokens_seen": 45303168, "step": 23535 }, { "epoch": 3.8401990374418795, "grad_norm": 0.007601574528962374, "learning_rate": 1.8140319421699234e-05, "loss": 0.0429, "num_input_tokens_seen": 45312992, "step": 23540 }, { "epoch": 3.8410147646626966, "grad_norm": 0.015206480398774147, "learning_rate": 1.8130879275599515e-05, "loss": 0.0007, "num_input_tokens_seen": 45323296, "step": 23545 }, { "epoch": 3.841830491883514, "grad_norm": 0.07709959894418716, "learning_rate": 1.8121440188804544e-05, "loss": 0.0025, "num_input_tokens_seen": 45331424, "step": 23550 }, { "epoch": 3.8426462191043314, "grad_norm": 0.02109519951045513, "learning_rate": 1.811200216276993e-05, "loss": 0.1873, "num_input_tokens_seen": 45339904, "step": 23555 }, { "epoch": 3.843461946325149, "grad_norm": 0.0178622305393219, "learning_rate": 1.810256519895115e-05, "loss": 0.0666, "num_input_tokens_seen": 45348704, "step": 23560 }, { "epoch": 3.8442776735459665, "grad_norm": 4.269245147705078, "learning_rate": 1.8093129298803494e-05, "loss": 0.0763, "num_input_tokens_seen": 45357840, "step": 23565 }, { "epoch": 3.8450934007667836, "grad_norm": 0.512679934501648, "learning_rate": 1.808369446378209e-05, "loss": 0.0707, "num_input_tokens_seen": 45367216, "step": 23570 }, { "epoch": 3.8459091279876008, "grad_norm": 0.06984791904687881, "learning_rate": 1.8074260695341914e-05, "loss": 0.04, "num_input_tokens_seen": 45375296, "step": 23575 }, { "epoch": 3.8467248552084183, "grad_norm": 0.027215471491217613, "learning_rate": 1.8064827994937782e-05, "loss": 0.0054, "num_input_tokens_seen": 45385632, "step": 23580 }, { "epoch": 3.847540582429236, "grad_norm": 2.8548929691314697, "learning_rate": 1.8055396364024317e-05, "loss": 0.1558, "num_input_tokens_seen": 45395536, "step": 23585 }, { "epoch": 3.848356309650053, "grad_norm": 0.010109935887157917, "learning_rate": 1.804596580405601e-05, "loss": 0.0009, "num_input_tokens_seen": 45405072, "step": 23590 }, { "epoch": 3.84917203687087, "grad_norm": 8.924600601196289, "learning_rate": 1.8036536316487174e-05, "loss": 0.0719, "num_input_tokens_seen": 45416080, "step": 23595 }, { "epoch": 3.8499877640916877, "grad_norm": 16.9992733001709, "learning_rate": 1.802710790277193e-05, "loss": 0.0808, "num_input_tokens_seen": 45426496, "step": 23600 }, { "epoch": 3.8499877640916877, "eval_loss": 0.22055284678936005, "eval_runtime": 80.8506, "eval_samples_per_second": 33.704, "eval_steps_per_second": 16.858, "num_input_tokens_seen": 45426496, "step": 23600 }, { "epoch": 3.8508034913125053, "grad_norm": 0.39164093136787415, "learning_rate": 1.801768056436429e-05, "loss": 0.0038, "num_input_tokens_seen": 45435712, "step": 23605 }, { "epoch": 3.8516192185333225, "grad_norm": 0.045130133628845215, "learning_rate": 1.8008254302718035e-05, "loss": 0.0411, "num_input_tokens_seen": 45446432, "step": 23610 }, { "epoch": 3.8524349457541396, "grad_norm": 0.005062705371528864, "learning_rate": 1.7998829119286837e-05, "loss": 0.0024, "num_input_tokens_seen": 45455760, "step": 23615 }, { "epoch": 3.853250672974957, "grad_norm": 0.10634361952543259, "learning_rate": 1.798940501552418e-05, "loss": 0.0022, "num_input_tokens_seen": 45466304, "step": 23620 }, { "epoch": 3.8540664001957747, "grad_norm": 1.2514804601669312, "learning_rate": 1.797998199288336e-05, "loss": 0.0028, "num_input_tokens_seen": 45474544, "step": 23625 }, { "epoch": 3.854882127416592, "grad_norm": 0.023979084566235542, "learning_rate": 1.7970560052817543e-05, "loss": 0.0543, "num_input_tokens_seen": 45484736, "step": 23630 }, { "epoch": 3.855697854637409, "grad_norm": 34.11445236206055, "learning_rate": 1.7961139196779702e-05, "loss": 0.015, "num_input_tokens_seen": 45494064, "step": 23635 }, { "epoch": 3.8565135818582266, "grad_norm": 0.02226126939058304, "learning_rate": 1.7951719426222647e-05, "loss": 0.082, "num_input_tokens_seen": 45504016, "step": 23640 }, { "epoch": 3.857329309079044, "grad_norm": 0.020512808114290237, "learning_rate": 1.794230074259904e-05, "loss": 0.0746, "num_input_tokens_seen": 45512720, "step": 23645 }, { "epoch": 3.8581450362998613, "grad_norm": 0.04959171265363693, "learning_rate": 1.7932883147361336e-05, "loss": 0.095, "num_input_tokens_seen": 45523056, "step": 23650 }, { "epoch": 3.858960763520679, "grad_norm": 5.517549991607666, "learning_rate": 1.7923466641961865e-05, "loss": 0.0346, "num_input_tokens_seen": 45532672, "step": 23655 }, { "epoch": 3.859776490741496, "grad_norm": 0.03867103531956673, "learning_rate": 1.791405122785278e-05, "loss": 0.1098, "num_input_tokens_seen": 45541632, "step": 23660 }, { "epoch": 3.8605922179623136, "grad_norm": 11.452033042907715, "learning_rate": 1.7904636906486037e-05, "loss": 0.1102, "num_input_tokens_seen": 45552016, "step": 23665 }, { "epoch": 3.8614079451831307, "grad_norm": 0.053729575127363205, "learning_rate": 1.7895223679313448e-05, "loss": 0.0027, "num_input_tokens_seen": 45562048, "step": 23670 }, { "epoch": 3.8622236724039483, "grad_norm": 0.14861436188220978, "learning_rate": 1.7885811547786653e-05, "loss": 0.0124, "num_input_tokens_seen": 45572544, "step": 23675 }, { "epoch": 3.8630393996247654, "grad_norm": 0.06167653203010559, "learning_rate": 1.7876400513357115e-05, "loss": 0.0011, "num_input_tokens_seen": 45582672, "step": 23680 }, { "epoch": 3.863855126845583, "grad_norm": 0.11388162523508072, "learning_rate": 1.7866990577476146e-05, "loss": 0.0016, "num_input_tokens_seen": 45592464, "step": 23685 }, { "epoch": 3.8646708540664, "grad_norm": 0.271121084690094, "learning_rate": 1.7857581741594863e-05, "loss": 0.2124, "num_input_tokens_seen": 45602160, "step": 23690 }, { "epoch": 3.8654865812872177, "grad_norm": 0.07852467894554138, "learning_rate": 1.7848174007164237e-05, "loss": 0.0014, "num_input_tokens_seen": 45611504, "step": 23695 }, { "epoch": 3.866302308508035, "grad_norm": 0.024526482447981834, "learning_rate": 1.7838767375635052e-05, "loss": 0.0762, "num_input_tokens_seen": 45621232, "step": 23700 }, { "epoch": 3.8671180357288524, "grad_norm": 0.02389206923544407, "learning_rate": 1.782936184845793e-05, "loss": 0.002, "num_input_tokens_seen": 45630576, "step": 23705 }, { "epoch": 3.8679337629496695, "grad_norm": 0.06913624703884125, "learning_rate": 1.7819957427083334e-05, "loss": 0.0019, "num_input_tokens_seen": 45640240, "step": 23710 }, { "epoch": 3.868749490170487, "grad_norm": 0.054843444377183914, "learning_rate": 1.7810554112961516e-05, "loss": 0.0209, "num_input_tokens_seen": 45649680, "step": 23715 }, { "epoch": 3.869565217391304, "grad_norm": 0.1770460605621338, "learning_rate": 1.7801151907542607e-05, "loss": 0.0034, "num_input_tokens_seen": 45658832, "step": 23720 }, { "epoch": 3.870380944612122, "grad_norm": 0.02646060846745968, "learning_rate": 1.7791750812276547e-05, "loss": 0.0805, "num_input_tokens_seen": 45667744, "step": 23725 }, { "epoch": 3.871196671832939, "grad_norm": 9.558135986328125, "learning_rate": 1.778235082861309e-05, "loss": 0.1158, "num_input_tokens_seen": 45677760, "step": 23730 }, { "epoch": 3.8720123990537565, "grad_norm": 0.08422793447971344, "learning_rate": 1.777295195800184e-05, "loss": 0.0019, "num_input_tokens_seen": 45687088, "step": 23735 }, { "epoch": 3.8728281262745736, "grad_norm": 0.04100879654288292, "learning_rate": 1.7763554201892215e-05, "loss": 0.0065, "num_input_tokens_seen": 45696816, "step": 23740 }, { "epoch": 3.873643853495391, "grad_norm": 30.358654022216797, "learning_rate": 1.7754157561733476e-05, "loss": 0.0342, "num_input_tokens_seen": 45706544, "step": 23745 }, { "epoch": 3.8744595807162083, "grad_norm": 0.21118423342704773, "learning_rate": 1.7744762038974702e-05, "loss": 0.0018, "num_input_tokens_seen": 45716320, "step": 23750 }, { "epoch": 3.875275307937026, "grad_norm": 0.5819016098976135, "learning_rate": 1.7735367635064788e-05, "loss": 0.042, "num_input_tokens_seen": 45726992, "step": 23755 }, { "epoch": 3.8760910351578435, "grad_norm": 5.458822727203369, "learning_rate": 1.7725974351452474e-05, "loss": 0.1142, "num_input_tokens_seen": 45736160, "step": 23760 }, { "epoch": 3.8769067623786606, "grad_norm": 0.026134835556149483, "learning_rate": 1.771658218958634e-05, "loss": 0.062, "num_input_tokens_seen": 45744976, "step": 23765 }, { "epoch": 3.8777224895994777, "grad_norm": 0.1456063985824585, "learning_rate": 1.770719115091475e-05, "loss": 0.0018, "num_input_tokens_seen": 45754912, "step": 23770 }, { "epoch": 3.8785382168202953, "grad_norm": 0.0069508021697402, "learning_rate": 1.7697801236885935e-05, "loss": 0.0184, "num_input_tokens_seen": 45765200, "step": 23775 }, { "epoch": 3.879353944041113, "grad_norm": 0.057759083807468414, "learning_rate": 1.7688412448947944e-05, "loss": 0.0037, "num_input_tokens_seen": 45775280, "step": 23780 }, { "epoch": 3.88016967126193, "grad_norm": 0.010064798407256603, "learning_rate": 1.767902478854862e-05, "loss": 0.0014, "num_input_tokens_seen": 45784832, "step": 23785 }, { "epoch": 3.880985398482747, "grad_norm": 0.02167762815952301, "learning_rate": 1.766963825713569e-05, "loss": 0.0011, "num_input_tokens_seen": 45794032, "step": 23790 }, { "epoch": 3.8818011257035647, "grad_norm": 0.11201658844947815, "learning_rate": 1.766025285615665e-05, "loss": 0.0714, "num_input_tokens_seen": 45803360, "step": 23795 }, { "epoch": 3.8826168529243823, "grad_norm": 22.445632934570312, "learning_rate": 1.7650868587058854e-05, "loss": 0.1363, "num_input_tokens_seen": 45813536, "step": 23800 }, { "epoch": 3.8826168529243823, "eval_loss": 0.2621123790740967, "eval_runtime": 80.75, "eval_samples_per_second": 33.746, "eval_steps_per_second": 16.879, "num_input_tokens_seen": 45813536, "step": 23800 }, { "epoch": 3.8834325801451994, "grad_norm": 0.060192324221134186, "learning_rate": 1.7641485451289484e-05, "loss": 0.2097, "num_input_tokens_seen": 45823008, "step": 23805 }, { "epoch": 3.8842483073660166, "grad_norm": 0.06190494820475578, "learning_rate": 1.7632103450295534e-05, "loss": 0.0006, "num_input_tokens_seen": 45830208, "step": 23810 }, { "epoch": 3.885064034586834, "grad_norm": 0.11105020344257355, "learning_rate": 1.762272258552381e-05, "loss": 0.2892, "num_input_tokens_seen": 45840928, "step": 23815 }, { "epoch": 3.8858797618076517, "grad_norm": 0.05356283485889435, "learning_rate": 1.7613342858420988e-05, "loss": 0.0584, "num_input_tokens_seen": 45850992, "step": 23820 }, { "epoch": 3.886695489028469, "grad_norm": 0.26302629709243774, "learning_rate": 1.760396427043351e-05, "loss": 0.0637, "num_input_tokens_seen": 45860320, "step": 23825 }, { "epoch": 3.887511216249286, "grad_norm": 0.023557215929031372, "learning_rate": 1.7594586823007696e-05, "loss": 0.0016, "num_input_tokens_seen": 45870368, "step": 23830 }, { "epoch": 3.8883269434701035, "grad_norm": 7.4486517906188965, "learning_rate": 1.7585210517589646e-05, "loss": 0.2168, "num_input_tokens_seen": 45879744, "step": 23835 }, { "epoch": 3.889142670690921, "grad_norm": 5.185449123382568, "learning_rate": 1.7575835355625314e-05, "loss": 0.1038, "num_input_tokens_seen": 45889424, "step": 23840 }, { "epoch": 3.8899583979117383, "grad_norm": 5.1167097091674805, "learning_rate": 1.756646133856048e-05, "loss": 0.0787, "num_input_tokens_seen": 45899344, "step": 23845 }, { "epoch": 3.890774125132556, "grad_norm": 0.1059434711933136, "learning_rate": 1.7557088467840714e-05, "loss": 0.0426, "num_input_tokens_seen": 45907088, "step": 23850 }, { "epoch": 3.891589852353373, "grad_norm": 3.4416229724884033, "learning_rate": 1.7547716744911438e-05, "loss": 0.0734, "num_input_tokens_seen": 45915600, "step": 23855 }, { "epoch": 3.8924055795741905, "grad_norm": 0.05724203586578369, "learning_rate": 1.7538346171217902e-05, "loss": 0.0727, "num_input_tokens_seen": 45925856, "step": 23860 }, { "epoch": 3.8932213067950077, "grad_norm": 0.11686507612466812, "learning_rate": 1.7528976748205146e-05, "loss": 0.0487, "num_input_tokens_seen": 45935328, "step": 23865 }, { "epoch": 3.8940370340158252, "grad_norm": 0.11265715211629868, "learning_rate": 1.751960847731807e-05, "loss": 0.0038, "num_input_tokens_seen": 45944832, "step": 23870 }, { "epoch": 3.8948527612366424, "grad_norm": 1.7678154706954956, "learning_rate": 1.7510241360001362e-05, "loss": 0.0551, "num_input_tokens_seen": 45955264, "step": 23875 }, { "epoch": 3.89566848845746, "grad_norm": 0.10564921796321869, "learning_rate": 1.7500875397699562e-05, "loss": 0.0329, "num_input_tokens_seen": 45963312, "step": 23880 }, { "epoch": 3.896484215678277, "grad_norm": 0.019755465909838676, "learning_rate": 1.7491510591857015e-05, "loss": 0.0059, "num_input_tokens_seen": 45973056, "step": 23885 }, { "epoch": 3.8972999428990946, "grad_norm": 0.13788466155529022, "learning_rate": 1.7482146943917896e-05, "loss": 0.1032, "num_input_tokens_seen": 45982416, "step": 23890 }, { "epoch": 3.898115670119912, "grad_norm": 3.3024728298187256, "learning_rate": 1.7472784455326185e-05, "loss": 0.1281, "num_input_tokens_seen": 45993104, "step": 23895 }, { "epoch": 3.8989313973407294, "grad_norm": 0.07480019330978394, "learning_rate": 1.746342312752572e-05, "loss": 0.0313, "num_input_tokens_seen": 46002528, "step": 23900 }, { "epoch": 3.8997471245615465, "grad_norm": 0.12794730067253113, "learning_rate": 1.74540629619601e-05, "loss": 0.027, "num_input_tokens_seen": 46011408, "step": 23905 }, { "epoch": 3.900562851782364, "grad_norm": 0.302467405796051, "learning_rate": 1.7444703960072815e-05, "loss": 0.0139, "num_input_tokens_seen": 46019280, "step": 23910 }, { "epoch": 3.901378579003181, "grad_norm": 5.127807140350342, "learning_rate": 1.7435346123307118e-05, "loss": 0.0162, "num_input_tokens_seen": 46028720, "step": 23915 }, { "epoch": 3.9021943062239988, "grad_norm": 0.21729010343551636, "learning_rate": 1.742598945310611e-05, "loss": 0.0054, "num_input_tokens_seen": 46038192, "step": 23920 }, { "epoch": 3.903010033444816, "grad_norm": 0.062309809029102325, "learning_rate": 1.741663395091272e-05, "loss": 0.0019, "num_input_tokens_seen": 46047120, "step": 23925 }, { "epoch": 3.9038257606656335, "grad_norm": 0.03170297294855118, "learning_rate": 1.7407279618169657e-05, "loss": 0.1203, "num_input_tokens_seen": 46057008, "step": 23930 }, { "epoch": 3.904641487886451, "grad_norm": 0.13014638423919678, "learning_rate": 1.73979264563195e-05, "loss": 0.0364, "num_input_tokens_seen": 46066176, "step": 23935 }, { "epoch": 3.905457215107268, "grad_norm": 0.041178636252880096, "learning_rate": 1.7388574466804625e-05, "loss": 0.0016, "num_input_tokens_seen": 46076464, "step": 23940 }, { "epoch": 3.9062729423280853, "grad_norm": 9.996639251708984, "learning_rate": 1.7379223651067207e-05, "loss": 0.1013, "num_input_tokens_seen": 46086656, "step": 23945 }, { "epoch": 3.907088669548903, "grad_norm": 0.157805934548378, "learning_rate": 1.736987401054928e-05, "loss": 0.003, "num_input_tokens_seen": 46095648, "step": 23950 }, { "epoch": 3.9079043967697205, "grad_norm": 0.06987877190113068, "learning_rate": 1.736052554669266e-05, "loss": 0.1237, "num_input_tokens_seen": 46104960, "step": 23955 }, { "epoch": 3.9087201239905376, "grad_norm": 0.05432639271020889, "learning_rate": 1.7351178260939007e-05, "loss": 0.0302, "num_input_tokens_seen": 46114992, "step": 23960 }, { "epoch": 3.9095358512113547, "grad_norm": 0.10174841433763504, "learning_rate": 1.7341832154729794e-05, "loss": 0.0077, "num_input_tokens_seen": 46123728, "step": 23965 }, { "epoch": 3.9103515784321723, "grad_norm": 0.06271301209926605, "learning_rate": 1.7332487229506286e-05, "loss": 0.062, "num_input_tokens_seen": 46134016, "step": 23970 }, { "epoch": 3.91116730565299, "grad_norm": 0.015306372195482254, "learning_rate": 1.732314348670961e-05, "loss": 0.0828, "num_input_tokens_seen": 46145088, "step": 23975 }, { "epoch": 3.911983032873807, "grad_norm": 0.2262871414422989, "learning_rate": 1.7313800927780686e-05, "loss": 0.0028, "num_input_tokens_seen": 46155728, "step": 23980 }, { "epoch": 3.912798760094624, "grad_norm": 0.987302839756012, "learning_rate": 1.7304459554160245e-05, "loss": 0.074, "num_input_tokens_seen": 46165072, "step": 23985 }, { "epoch": 3.9136144873154417, "grad_norm": 0.17659014463424683, "learning_rate": 1.7295119367288853e-05, "loss": 0.1266, "num_input_tokens_seen": 46174416, "step": 23990 }, { "epoch": 3.9144302145362593, "grad_norm": 0.049260493367910385, "learning_rate": 1.728578036860688e-05, "loss": 0.0008, "num_input_tokens_seen": 46182912, "step": 23995 }, { "epoch": 3.9152459417570764, "grad_norm": 0.06781288236379623, "learning_rate": 1.7276442559554513e-05, "loss": 0.1103, "num_input_tokens_seen": 46192656, "step": 24000 }, { "epoch": 3.9152459417570764, "eval_loss": 0.22623005509376526, "eval_runtime": 80.8747, "eval_samples_per_second": 33.694, "eval_steps_per_second": 16.853, "num_input_tokens_seen": 46192656, "step": 24000 }, { "epoch": 3.9160616689778935, "grad_norm": 0.24848079681396484, "learning_rate": 1.726710594157177e-05, "loss": 0.0017, "num_input_tokens_seen": 46201376, "step": 24005 }, { "epoch": 3.916877396198711, "grad_norm": 7.47152042388916, "learning_rate": 1.725777051609846e-05, "loss": 0.1399, "num_input_tokens_seen": 46211728, "step": 24010 }, { "epoch": 3.9176931234195287, "grad_norm": 0.02873879484832287, "learning_rate": 1.7248436284574228e-05, "loss": 0.0012, "num_input_tokens_seen": 46219136, "step": 24015 }, { "epoch": 3.918508850640346, "grad_norm": 0.1884484738111496, "learning_rate": 1.723910324843855e-05, "loss": 0.0824, "num_input_tokens_seen": 46228384, "step": 24020 }, { "epoch": 3.919324577861163, "grad_norm": 0.058901358395814896, "learning_rate": 1.722977140913067e-05, "loss": 0.1015, "num_input_tokens_seen": 46238288, "step": 24025 }, { "epoch": 3.9201403050819805, "grad_norm": 4.424781799316406, "learning_rate": 1.7220440768089688e-05, "loss": 0.0911, "num_input_tokens_seen": 46248096, "step": 24030 }, { "epoch": 3.920956032302798, "grad_norm": 0.047605425119400024, "learning_rate": 1.7211111326754505e-05, "loss": 0.0933, "num_input_tokens_seen": 46257248, "step": 24035 }, { "epoch": 3.9217717595236152, "grad_norm": 9.389705657958984, "learning_rate": 1.720178308656383e-05, "loss": 0.1284, "num_input_tokens_seen": 46267296, "step": 24040 }, { "epoch": 3.922587486744433, "grad_norm": 36.66328811645508, "learning_rate": 1.719245604895621e-05, "loss": 0.0235, "num_input_tokens_seen": 46276000, "step": 24045 }, { "epoch": 3.92340321396525, "grad_norm": 0.2456669956445694, "learning_rate": 1.7183130215369972e-05, "loss": 0.0287, "num_input_tokens_seen": 46286080, "step": 24050 }, { "epoch": 3.9242189411860675, "grad_norm": 0.06902839988470078, "learning_rate": 1.7173805587243292e-05, "loss": 0.0957, "num_input_tokens_seen": 46295824, "step": 24055 }, { "epoch": 3.9250346684068846, "grad_norm": 0.0211858619004488, "learning_rate": 1.7164482166014147e-05, "loss": 0.0016, "num_input_tokens_seen": 46305376, "step": 24060 }, { "epoch": 3.925850395627702, "grad_norm": 0.056660573929548264, "learning_rate": 1.7155159953120313e-05, "loss": 0.0024, "num_input_tokens_seen": 46315104, "step": 24065 }, { "epoch": 3.9266661228485193, "grad_norm": 0.14528074860572815, "learning_rate": 1.714583894999941e-05, "loss": 0.31, "num_input_tokens_seen": 46324624, "step": 24070 }, { "epoch": 3.927481850069337, "grad_norm": 0.017731698229908943, "learning_rate": 1.7136519158088826e-05, "loss": 0.0022, "num_input_tokens_seen": 46334080, "step": 24075 }, { "epoch": 3.928297577290154, "grad_norm": 25.751537322998047, "learning_rate": 1.712720057882581e-05, "loss": 0.1074, "num_input_tokens_seen": 46344608, "step": 24080 }, { "epoch": 3.9291133045109716, "grad_norm": 0.036074765026569366, "learning_rate": 1.7117883213647413e-05, "loss": 0.0048, "num_input_tokens_seen": 46354288, "step": 24085 }, { "epoch": 3.9299290317317888, "grad_norm": 4.23414421081543, "learning_rate": 1.710856706399046e-05, "loss": 0.0763, "num_input_tokens_seen": 46363984, "step": 24090 }, { "epoch": 3.9307447589526063, "grad_norm": 0.8619965314865112, "learning_rate": 1.7099252131291648e-05, "loss": 0.004, "num_input_tokens_seen": 46374208, "step": 24095 }, { "epoch": 3.9315604861734235, "grad_norm": 0.054474372416734695, "learning_rate": 1.708993841698744e-05, "loss": 0.078, "num_input_tokens_seen": 46384544, "step": 24100 }, { "epoch": 3.932376213394241, "grad_norm": 85.84513854980469, "learning_rate": 1.7080625922514132e-05, "loss": 0.0933, "num_input_tokens_seen": 46394592, "step": 24105 }, { "epoch": 3.933191940615058, "grad_norm": 0.03263372182846069, "learning_rate": 1.7071314649307836e-05, "loss": 0.0536, "num_input_tokens_seen": 46404864, "step": 24110 }, { "epoch": 3.9340076678358757, "grad_norm": 0.05418926849961281, "learning_rate": 1.7062004598804448e-05, "loss": 0.1214, "num_input_tokens_seen": 46415312, "step": 24115 }, { "epoch": 3.934823395056693, "grad_norm": 0.18630270659923553, "learning_rate": 1.7052695772439702e-05, "loss": 0.0657, "num_input_tokens_seen": 46425360, "step": 24120 }, { "epoch": 3.9356391222775104, "grad_norm": 0.0683775246143341, "learning_rate": 1.7043388171649154e-05, "loss": 0.0769, "num_input_tokens_seen": 46434960, "step": 24125 }, { "epoch": 3.936454849498328, "grad_norm": 0.05967056751251221, "learning_rate": 1.7034081797868127e-05, "loss": 0.0026, "num_input_tokens_seen": 46445424, "step": 24130 }, { "epoch": 3.937270576719145, "grad_norm": 42.790687561035156, "learning_rate": 1.70247766525318e-05, "loss": 0.1627, "num_input_tokens_seen": 46455504, "step": 24135 }, { "epoch": 3.9380863039399623, "grad_norm": 0.09888812154531479, "learning_rate": 1.701547273707514e-05, "loss": 0.0018, "num_input_tokens_seen": 46463696, "step": 24140 }, { "epoch": 3.93890203116078, "grad_norm": 0.20845076441764832, "learning_rate": 1.7006170052932916e-05, "loss": 0.0476, "num_input_tokens_seen": 46473424, "step": 24145 }, { "epoch": 3.9397177583815974, "grad_norm": 0.09587772935628891, "learning_rate": 1.6996868601539735e-05, "loss": 0.0043, "num_input_tokens_seen": 46482768, "step": 24150 }, { "epoch": 3.9405334856024146, "grad_norm": 0.024844219908118248, "learning_rate": 1.6987568384329977e-05, "loss": 0.0033, "num_input_tokens_seen": 46491216, "step": 24155 }, { "epoch": 3.9413492128232317, "grad_norm": 13.605527877807617, "learning_rate": 1.6978269402737866e-05, "loss": 0.1896, "num_input_tokens_seen": 46501120, "step": 24160 }, { "epoch": 3.9421649400440493, "grad_norm": 0.1087789386510849, "learning_rate": 1.696897165819743e-05, "loss": 0.0022, "num_input_tokens_seen": 46512080, "step": 24165 }, { "epoch": 3.942980667264867, "grad_norm": 0.03651588410139084, "learning_rate": 1.6959675152142487e-05, "loss": 0.0089, "num_input_tokens_seen": 46520960, "step": 24170 }, { "epoch": 3.943796394485684, "grad_norm": 0.8023672699928284, "learning_rate": 1.6950379886006667e-05, "loss": 0.0044, "num_input_tokens_seen": 46531184, "step": 24175 }, { "epoch": 3.944612121706501, "grad_norm": 0.022965652868151665, "learning_rate": 1.6941085861223438e-05, "loss": 0.0624, "num_input_tokens_seen": 46541008, "step": 24180 }, { "epoch": 3.9454278489273187, "grad_norm": 5.563466548919678, "learning_rate": 1.6931793079226034e-05, "loss": 0.062, "num_input_tokens_seen": 46550832, "step": 24185 }, { "epoch": 3.9462435761481363, "grad_norm": 4.368497371673584, "learning_rate": 1.692250154144754e-05, "loss": 0.1405, "num_input_tokens_seen": 46560576, "step": 24190 }, { "epoch": 3.9470593033689534, "grad_norm": 3.488481044769287, "learning_rate": 1.6913211249320807e-05, "loss": 0.1524, "num_input_tokens_seen": 46569312, "step": 24195 }, { "epoch": 3.9478750305897705, "grad_norm": 0.02807697467505932, "learning_rate": 1.6903922204278522e-05, "loss": 0.0641, "num_input_tokens_seen": 46576928, "step": 24200 }, { "epoch": 3.9478750305897705, "eval_loss": 0.2038833200931549, "eval_runtime": 80.9065, "eval_samples_per_second": 33.681, "eval_steps_per_second": 16.847, "num_input_tokens_seen": 46576928, "step": 24200 }, { "epoch": 3.948690757810588, "grad_norm": 0.024746917188167572, "learning_rate": 1.6894634407753186e-05, "loss": 0.0715, "num_input_tokens_seen": 46588144, "step": 24205 }, { "epoch": 3.9495064850314057, "grad_norm": 0.06978381425142288, "learning_rate": 1.6885347861177077e-05, "loss": 0.0024, "num_input_tokens_seen": 46597680, "step": 24210 }, { "epoch": 3.950322212252223, "grad_norm": 4.501682281494141, "learning_rate": 1.6876062565982298e-05, "loss": 0.0688, "num_input_tokens_seen": 46608272, "step": 24215 }, { "epoch": 3.9511379394730404, "grad_norm": 3.5782864093780518, "learning_rate": 1.6866778523600774e-05, "loss": 0.0942, "num_input_tokens_seen": 46618416, "step": 24220 }, { "epoch": 3.9519536666938575, "grad_norm": 4.167619228363037, "learning_rate": 1.6857495735464195e-05, "loss": 0.0778, "num_input_tokens_seen": 46627824, "step": 24225 }, { "epoch": 3.952769393914675, "grad_norm": 4.194647789001465, "learning_rate": 1.6848214203004115e-05, "loss": 0.0618, "num_input_tokens_seen": 46637680, "step": 24230 }, { "epoch": 3.953585121135492, "grad_norm": 0.1633656620979309, "learning_rate": 1.6838933927651835e-05, "loss": 0.0033, "num_input_tokens_seen": 46647056, "step": 24235 }, { "epoch": 3.95440084835631, "grad_norm": 3.164591073989868, "learning_rate": 1.6829654910838506e-05, "loss": 0.0131, "num_input_tokens_seen": 46656672, "step": 24240 }, { "epoch": 3.955216575577127, "grad_norm": 0.018441790714859962, "learning_rate": 1.6820377153995065e-05, "loss": 0.0098, "num_input_tokens_seen": 46666848, "step": 24245 }, { "epoch": 3.9560323027979445, "grad_norm": 4.026503562927246, "learning_rate": 1.681110065855226e-05, "loss": 0.221, "num_input_tokens_seen": 46676192, "step": 24250 }, { "epoch": 3.9568480300187616, "grad_norm": 0.03441628813743591, "learning_rate": 1.6801825425940642e-05, "loss": 0.0371, "num_input_tokens_seen": 46685216, "step": 24255 }, { "epoch": 3.957663757239579, "grad_norm": 3.1989409923553467, "learning_rate": 1.679255145759056e-05, "loss": 0.1321, "num_input_tokens_seen": 46694352, "step": 24260 }, { "epoch": 3.9584794844603963, "grad_norm": 0.10141424834728241, "learning_rate": 1.6783278754932187e-05, "loss": 0.0375, "num_input_tokens_seen": 46704416, "step": 24265 }, { "epoch": 3.959295211681214, "grad_norm": 0.042829498648643494, "learning_rate": 1.6774007319395496e-05, "loss": 0.0184, "num_input_tokens_seen": 46713744, "step": 24270 }, { "epoch": 3.960110938902031, "grad_norm": 0.08499446511268616, "learning_rate": 1.6764737152410243e-05, "loss": 0.0771, "num_input_tokens_seen": 46722896, "step": 24275 }, { "epoch": 3.9609266661228486, "grad_norm": 4.533848285675049, "learning_rate": 1.6755468255406016e-05, "loss": 0.1907, "num_input_tokens_seen": 46732320, "step": 24280 }, { "epoch": 3.9617423933436657, "grad_norm": 1.8037093877792358, "learning_rate": 1.674620062981219e-05, "loss": 0.0067, "num_input_tokens_seen": 46741504, "step": 24285 }, { "epoch": 3.9625581205644833, "grad_norm": 17.512836456298828, "learning_rate": 1.6736934277057947e-05, "loss": 0.0331, "num_input_tokens_seen": 46750656, "step": 24290 }, { "epoch": 3.9633738477853004, "grad_norm": 0.05266428366303444, "learning_rate": 1.6727669198572286e-05, "loss": 0.0043, "num_input_tokens_seen": 46760608, "step": 24295 }, { "epoch": 3.964189575006118, "grad_norm": 0.016441915184259415, "learning_rate": 1.6718405395783984e-05, "loss": 0.0338, "num_input_tokens_seen": 46770288, "step": 24300 }, { "epoch": 3.965005302226935, "grad_norm": 0.25239428877830505, "learning_rate": 1.6709142870121643e-05, "loss": 0.1299, "num_input_tokens_seen": 46779024, "step": 24305 }, { "epoch": 3.9658210294477527, "grad_norm": 0.11472759395837784, "learning_rate": 1.669988162301367e-05, "loss": 0.0183, "num_input_tokens_seen": 46789648, "step": 24310 }, { "epoch": 3.96663675666857, "grad_norm": 16.50849723815918, "learning_rate": 1.6690621655888243e-05, "loss": 0.1044, "num_input_tokens_seen": 46800224, "step": 24315 }, { "epoch": 3.9674524838893874, "grad_norm": 0.11051449924707413, "learning_rate": 1.6681362970173386e-05, "loss": 0.0026, "num_input_tokens_seen": 46809264, "step": 24320 }, { "epoch": 3.968268211110205, "grad_norm": 0.07795634120702744, "learning_rate": 1.6672105567296904e-05, "loss": 0.1385, "num_input_tokens_seen": 46816656, "step": 24325 }, { "epoch": 3.969083938331022, "grad_norm": 0.16109685599803925, "learning_rate": 1.666284944868639e-05, "loss": 0.0034, "num_input_tokens_seen": 46828224, "step": 24330 }, { "epoch": 3.9698996655518393, "grad_norm": 0.742036759853363, "learning_rate": 1.665359461576927e-05, "loss": 0.0115, "num_input_tokens_seen": 46838032, "step": 24335 }, { "epoch": 3.970715392772657, "grad_norm": 0.1102018877863884, "learning_rate": 1.6644341069972736e-05, "loss": 0.0599, "num_input_tokens_seen": 46847696, "step": 24340 }, { "epoch": 3.9715311199934744, "grad_norm": 11.664041519165039, "learning_rate": 1.6635088812723813e-05, "loss": 0.1187, "num_input_tokens_seen": 46857232, "step": 24345 }, { "epoch": 3.9723468472142915, "grad_norm": 0.5128690004348755, "learning_rate": 1.6625837845449328e-05, "loss": 0.0026, "num_input_tokens_seen": 46867040, "step": 24350 }, { "epoch": 3.9731625744351087, "grad_norm": 0.12635549902915955, "learning_rate": 1.6616588169575874e-05, "loss": 0.0025, "num_input_tokens_seen": 46876704, "step": 24355 }, { "epoch": 3.9739783016559262, "grad_norm": 4.196481227874756, "learning_rate": 1.6607339786529878e-05, "loss": 0.0641, "num_input_tokens_seen": 46887216, "step": 24360 }, { "epoch": 3.974794028876744, "grad_norm": 0.027192248031497, "learning_rate": 1.659809269773756e-05, "loss": 0.0019, "num_input_tokens_seen": 46896288, "step": 24365 }, { "epoch": 3.975609756097561, "grad_norm": 21.22047996520996, "learning_rate": 1.658884690462493e-05, "loss": 0.1313, "num_input_tokens_seen": 46906176, "step": 24370 }, { "epoch": 3.976425483318378, "grad_norm": 0.08497146517038345, "learning_rate": 1.6579602408617813e-05, "loss": 0.0013, "num_input_tokens_seen": 46916768, "step": 24375 }, { "epoch": 3.9772412105391957, "grad_norm": 0.11507779359817505, "learning_rate": 1.657035921114181e-05, "loss": 0.2353, "num_input_tokens_seen": 46926224, "step": 24380 }, { "epoch": 3.9780569377600132, "grad_norm": 0.01416892558336258, "learning_rate": 1.656111731362236e-05, "loss": 0.0011, "num_input_tokens_seen": 46935280, "step": 24385 }, { "epoch": 3.9788726649808304, "grad_norm": 0.8340739011764526, "learning_rate": 1.6551876717484666e-05, "loss": 0.0037, "num_input_tokens_seen": 46945616, "step": 24390 }, { "epoch": 3.9796883922016475, "grad_norm": 0.2509841024875641, "learning_rate": 1.6542637424153752e-05, "loss": 0.201, "num_input_tokens_seen": 46955920, "step": 24395 }, { "epoch": 3.980504119422465, "grad_norm": 0.1071416512131691, "learning_rate": 1.6533399435054418e-05, "loss": 0.1402, "num_input_tokens_seen": 46965120, "step": 24400 }, { "epoch": 3.980504119422465, "eval_loss": 0.21300624310970306, "eval_runtime": 80.826, "eval_samples_per_second": 33.714, "eval_steps_per_second": 16.863, "num_input_tokens_seen": 46965120, "step": 24400 }, { "epoch": 3.9813198466432826, "grad_norm": 11.579100608825684, "learning_rate": 1.6524162751611304e-05, "loss": 0.2035, "num_input_tokens_seen": 46976032, "step": 24405 }, { "epoch": 3.9821355738640998, "grad_norm": 0.094705730676651, "learning_rate": 1.6514927375248796e-05, "loss": 0.0447, "num_input_tokens_seen": 46984528, "step": 24410 }, { "epoch": 3.9829513010849174, "grad_norm": 0.016481125727295876, "learning_rate": 1.6505693307391127e-05, "loss": 0.052, "num_input_tokens_seen": 46993936, "step": 24415 }, { "epoch": 3.9837670283057345, "grad_norm": 0.4847278594970703, "learning_rate": 1.6496460549462288e-05, "loss": 0.0574, "num_input_tokens_seen": 47003776, "step": 24420 }, { "epoch": 3.984582755526552, "grad_norm": 4.57187557220459, "learning_rate": 1.6487229102886097e-05, "loss": 0.2164, "num_input_tokens_seen": 47013088, "step": 24425 }, { "epoch": 3.985398482747369, "grad_norm": 0.08142723888158798, "learning_rate": 1.6477998969086155e-05, "loss": 0.0818, "num_input_tokens_seen": 47022256, "step": 24430 }, { "epoch": 3.9862142099681868, "grad_norm": 0.31529226899147034, "learning_rate": 1.646877014948587e-05, "loss": 0.0048, "num_input_tokens_seen": 47031648, "step": 24435 }, { "epoch": 3.987029937189004, "grad_norm": 28.93999671936035, "learning_rate": 1.6459542645508433e-05, "loss": 0.0919, "num_input_tokens_seen": 47041984, "step": 24440 }, { "epoch": 3.9878456644098215, "grad_norm": 0.04300645366311073, "learning_rate": 1.6450316458576852e-05, "loss": 0.0992, "num_input_tokens_seen": 47051264, "step": 24445 }, { "epoch": 3.9886613916306386, "grad_norm": 6.638238430023193, "learning_rate": 1.6441091590113912e-05, "loss": 0.0717, "num_input_tokens_seen": 47060240, "step": 24450 }, { "epoch": 3.989477118851456, "grad_norm": 4.9952569007873535, "learning_rate": 1.6431868041542213e-05, "loss": 0.0826, "num_input_tokens_seen": 47069392, "step": 24455 }, { "epoch": 3.9902928460722733, "grad_norm": 0.08905123174190521, "learning_rate": 1.6422645814284123e-05, "loss": 0.0056, "num_input_tokens_seen": 47078688, "step": 24460 }, { "epoch": 3.991108573293091, "grad_norm": 0.03752795234322548, "learning_rate": 1.6413424909761846e-05, "loss": 0.0277, "num_input_tokens_seen": 47088400, "step": 24465 }, { "epoch": 3.991924300513908, "grad_norm": 0.055017776787281036, "learning_rate": 1.640420532939736e-05, "loss": 0.0019, "num_input_tokens_seen": 47098304, "step": 24470 }, { "epoch": 3.9927400277347256, "grad_norm": 0.5369076728820801, "learning_rate": 1.639498707461242e-05, "loss": 0.005, "num_input_tokens_seen": 47107872, "step": 24475 }, { "epoch": 3.9935557549555427, "grad_norm": 7.4578328132629395, "learning_rate": 1.6385770146828614e-05, "loss": 0.0441, "num_input_tokens_seen": 47117664, "step": 24480 }, { "epoch": 3.9943714821763603, "grad_norm": 0.3262813985347748, "learning_rate": 1.637655454746731e-05, "loss": 0.0029, "num_input_tokens_seen": 47126080, "step": 24485 }, { "epoch": 3.9951872093971774, "grad_norm": 0.01661662943661213, "learning_rate": 1.6367340277949658e-05, "loss": 0.0406, "num_input_tokens_seen": 47135600, "step": 24490 }, { "epoch": 3.996002936617995, "grad_norm": 0.04780857264995575, "learning_rate": 1.635812733969663e-05, "loss": 0.0024, "num_input_tokens_seen": 47145696, "step": 24495 }, { "epoch": 3.9968186638388126, "grad_norm": 0.09613076597452164, "learning_rate": 1.634891573412896e-05, "loss": 0.0668, "num_input_tokens_seen": 47156032, "step": 24500 }, { "epoch": 3.9976343910596297, "grad_norm": 2.8151471614837646, "learning_rate": 1.6339705462667196e-05, "loss": 0.0035, "num_input_tokens_seen": 47165424, "step": 24505 }, { "epoch": 3.998450118280447, "grad_norm": 1.1662797927856445, "learning_rate": 1.633049652673169e-05, "loss": 0.1197, "num_input_tokens_seen": 47174544, "step": 24510 }, { "epoch": 3.9992658455012644, "grad_norm": 0.19224657118320465, "learning_rate": 1.632128892774256e-05, "loss": 0.0497, "num_input_tokens_seen": 47184320, "step": 24515 }, { "epoch": 4.0, "grad_norm": 0.038961734622716904, "learning_rate": 1.6312082667119737e-05, "loss": 0.0014, "num_input_tokens_seen": 47193024, "step": 24520 }, { "epoch": 4.000815727220817, "grad_norm": 0.03520209714770317, "learning_rate": 1.630287774628296e-05, "loss": 0.0405, "num_input_tokens_seen": 47203712, "step": 24525 }, { "epoch": 4.001631454441635, "grad_norm": 0.1216912493109703, "learning_rate": 1.6293674166651718e-05, "loss": 0.0022, "num_input_tokens_seen": 47214032, "step": 24530 }, { "epoch": 4.002447181662452, "grad_norm": 0.053527992218732834, "learning_rate": 1.6284471929645338e-05, "loss": 0.0012, "num_input_tokens_seen": 47221840, "step": 24535 }, { "epoch": 4.003262908883269, "grad_norm": 0.07751041650772095, "learning_rate": 1.627527103668291e-05, "loss": 0.0315, "num_input_tokens_seen": 47229776, "step": 24540 }, { "epoch": 4.0040786361040865, "grad_norm": 0.01812962256371975, "learning_rate": 1.6266071489183327e-05, "loss": 0.0158, "num_input_tokens_seen": 47240128, "step": 24545 }, { "epoch": 4.004894363324905, "grad_norm": 0.37402328848838806, "learning_rate": 1.6256873288565283e-05, "loss": 0.003, "num_input_tokens_seen": 47250112, "step": 24550 }, { "epoch": 4.005710090545722, "grad_norm": 0.048578135669231415, "learning_rate": 1.6247676436247245e-05, "loss": 0.0073, "num_input_tokens_seen": 47259568, "step": 24555 }, { "epoch": 4.006525817766539, "grad_norm": 0.26142391562461853, "learning_rate": 1.6238480933647486e-05, "loss": 0.0046, "num_input_tokens_seen": 47268240, "step": 24560 }, { "epoch": 4.007341544987356, "grad_norm": 0.060112565755844116, "learning_rate": 1.6229286782184083e-05, "loss": 0.002, "num_input_tokens_seen": 47277920, "step": 24565 }, { "epoch": 4.008157272208174, "grad_norm": 4.441196918487549, "learning_rate": 1.622009398327487e-05, "loss": 0.0463, "num_input_tokens_seen": 47288704, "step": 24570 }, { "epoch": 4.008972999428991, "grad_norm": 9.833135604858398, "learning_rate": 1.6210902538337502e-05, "loss": 0.029, "num_input_tokens_seen": 47299136, "step": 24575 }, { "epoch": 4.009788726649808, "grad_norm": 0.025285400450229645, "learning_rate": 1.6201712448789413e-05, "loss": 0.0004, "num_input_tokens_seen": 47308784, "step": 24580 }, { "epoch": 4.010604453870625, "grad_norm": 0.9714885354042053, "learning_rate": 1.6192523716047827e-05, "loss": 0.0061, "num_input_tokens_seen": 47318864, "step": 24585 }, { "epoch": 4.011420181091443, "grad_norm": 0.959635853767395, "learning_rate": 1.6183336341529776e-05, "loss": 0.0021, "num_input_tokens_seen": 47327840, "step": 24590 }, { "epoch": 4.0122359083122605, "grad_norm": 0.4139183461666107, "learning_rate": 1.6174150326652047e-05, "loss": 0.0621, "num_input_tokens_seen": 47337280, "step": 24595 }, { "epoch": 4.013051635533078, "grad_norm": 0.021040674299001694, "learning_rate": 1.6164965672831256e-05, "loss": 0.0794, "num_input_tokens_seen": 47347920, "step": 24600 }, { "epoch": 4.013051635533078, "eval_loss": 0.24521394073963165, "eval_runtime": 80.8766, "eval_samples_per_second": 33.693, "eval_steps_per_second": 16.853, "num_input_tokens_seen": 47347920, "step": 24600 }, { "epoch": 4.013867362753895, "grad_norm": 0.1859603226184845, "learning_rate": 1.6155782381483784e-05, "loss": 0.0018, "num_input_tokens_seen": 47358416, "step": 24605 }, { "epoch": 4.014683089974713, "grad_norm": 0.01376344170421362, "learning_rate": 1.6146600454025813e-05, "loss": 0.0031, "num_input_tokens_seen": 47367632, "step": 24610 }, { "epoch": 4.01549881719553, "grad_norm": 0.018431531265378, "learning_rate": 1.6137419891873317e-05, "loss": 0.0006, "num_input_tokens_seen": 47377488, "step": 24615 }, { "epoch": 4.016314544416347, "grad_norm": 0.03586997464299202, "learning_rate": 1.6128240696442038e-05, "loss": 0.0012, "num_input_tokens_seen": 47386464, "step": 24620 }, { "epoch": 4.017130271637164, "grad_norm": 0.23163391649723053, "learning_rate": 1.611906286914753e-05, "loss": 0.0019, "num_input_tokens_seen": 47395904, "step": 24625 }, { "epoch": 4.017945998857982, "grad_norm": 0.005785262677818537, "learning_rate": 1.6109886411405144e-05, "loss": 0.0128, "num_input_tokens_seen": 47405760, "step": 24630 }, { "epoch": 4.018761726078799, "grad_norm": 0.006688904482871294, "learning_rate": 1.6100711324629985e-05, "loss": 0.0864, "num_input_tokens_seen": 47416688, "step": 24635 }, { "epoch": 4.0195774532996165, "grad_norm": 0.03737286478281021, "learning_rate": 1.609153761023698e-05, "loss": 0.0005, "num_input_tokens_seen": 47427024, "step": 24640 }, { "epoch": 4.020393180520434, "grad_norm": 0.02640603482723236, "learning_rate": 1.608236526964083e-05, "loss": 0.0014, "num_input_tokens_seen": 47436880, "step": 24645 }, { "epoch": 4.021208907741252, "grad_norm": 0.001717281760647893, "learning_rate": 1.607319430425601e-05, "loss": 0.0578, "num_input_tokens_seen": 47445840, "step": 24650 }, { "epoch": 4.022024634962069, "grad_norm": 0.1321447640657425, "learning_rate": 1.606402471549682e-05, "loss": 0.0011, "num_input_tokens_seen": 47454832, "step": 24655 }, { "epoch": 4.022840362182886, "grad_norm": 0.008719339035451412, "learning_rate": 1.6054856504777312e-05, "loss": 0.002, "num_input_tokens_seen": 47464496, "step": 24660 }, { "epoch": 4.023656089403703, "grad_norm": 0.937724232673645, "learning_rate": 1.6045689673511334e-05, "loss": 0.0007, "num_input_tokens_seen": 47474512, "step": 24665 }, { "epoch": 4.024471816624521, "grad_norm": 0.00032544086570851505, "learning_rate": 1.6036524223112548e-05, "loss": 0.1045, "num_input_tokens_seen": 47484752, "step": 24670 }, { "epoch": 4.025287543845338, "grad_norm": 0.05838370323181152, "learning_rate": 1.602736015499436e-05, "loss": 0.0008, "num_input_tokens_seen": 47494592, "step": 24675 }, { "epoch": 4.026103271066155, "grad_norm": 0.03847470134496689, "learning_rate": 1.601819747057e-05, "loss": 0.0017, "num_input_tokens_seen": 47504128, "step": 24680 }, { "epoch": 4.026918998286972, "grad_norm": 0.004498833324760199, "learning_rate": 1.6009036171252465e-05, "loss": 0.0009, "num_input_tokens_seen": 47514896, "step": 24685 }, { "epoch": 4.02773472550779, "grad_norm": 0.026205789297819138, "learning_rate": 1.599987625845453e-05, "loss": 0.0006, "num_input_tokens_seen": 47526160, "step": 24690 }, { "epoch": 4.028550452728608, "grad_norm": 0.0018192871939390898, "learning_rate": 1.599071773358879e-05, "loss": 0.0006, "num_input_tokens_seen": 47535744, "step": 24695 }, { "epoch": 4.029366179949425, "grad_norm": 0.027693338692188263, "learning_rate": 1.598156059806758e-05, "loss": 0.0034, "num_input_tokens_seen": 47545840, "step": 24700 }, { "epoch": 4.030181907170242, "grad_norm": 0.040869057178497314, "learning_rate": 1.5972404853303062e-05, "loss": 0.0021, "num_input_tokens_seen": 47555680, "step": 24705 }, { "epoch": 4.03099763439106, "grad_norm": 0.0015580900944769382, "learning_rate": 1.5963250500707172e-05, "loss": 0.0007, "num_input_tokens_seen": 47565200, "step": 24710 }, { "epoch": 4.031813361611877, "grad_norm": 0.0033802727703005075, "learning_rate": 1.5954097541691612e-05, "loss": 0.0016, "num_input_tokens_seen": 47574112, "step": 24715 }, { "epoch": 4.032629088832694, "grad_norm": 0.002068422269076109, "learning_rate": 1.5944945977667884e-05, "loss": 0.0002, "num_input_tokens_seen": 47583664, "step": 24720 }, { "epoch": 4.033444816053512, "grad_norm": 0.02074592560529709, "learning_rate": 1.593579581004729e-05, "loss": 0.0663, "num_input_tokens_seen": 47592880, "step": 24725 }, { "epoch": 4.034260543274329, "grad_norm": 2.037975549697876, "learning_rate": 1.592664704024088e-05, "loss": 0.0961, "num_input_tokens_seen": 47601744, "step": 24730 }, { "epoch": 4.035076270495146, "grad_norm": 0.10853776335716248, "learning_rate": 1.591749966965953e-05, "loss": 0.0473, "num_input_tokens_seen": 47612464, "step": 24735 }, { "epoch": 4.0358919977159635, "grad_norm": 0.003803413128480315, "learning_rate": 1.5908353699713856e-05, "loss": 0.0002, "num_input_tokens_seen": 47622368, "step": 24740 }, { "epoch": 4.0367077249367815, "grad_norm": 0.004953471943736076, "learning_rate": 1.5899209131814298e-05, "loss": 0.0004, "num_input_tokens_seen": 47631488, "step": 24745 }, { "epoch": 4.037523452157599, "grad_norm": 20.023988723754883, "learning_rate": 1.5890065967371067e-05, "loss": 0.013, "num_input_tokens_seen": 47642400, "step": 24750 }, { "epoch": 4.038339179378416, "grad_norm": 0.06802979111671448, "learning_rate": 1.5880924207794144e-05, "loss": 0.069, "num_input_tokens_seen": 47650224, "step": 24755 }, { "epoch": 4.039154906599233, "grad_norm": 0.21095427870750427, "learning_rate": 1.5871783854493298e-05, "loss": 0.0007, "num_input_tokens_seen": 47661280, "step": 24760 }, { "epoch": 4.039970633820051, "grad_norm": 0.054634880274534225, "learning_rate": 1.5862644908878106e-05, "loss": 0.0014, "num_input_tokens_seen": 47671152, "step": 24765 }, { "epoch": 4.040786361040868, "grad_norm": 0.05222489312291145, "learning_rate": 1.5853507372357885e-05, "loss": 0.0771, "num_input_tokens_seen": 47682336, "step": 24770 }, { "epoch": 4.041602088261685, "grad_norm": 0.008566876873373985, "learning_rate": 1.5844371246341776e-05, "loss": 0.0011, "num_input_tokens_seen": 47691424, "step": 24775 }, { "epoch": 4.042417815482502, "grad_norm": 0.03620561584830284, "learning_rate": 1.5835236532238674e-05, "loss": 0.0003, "num_input_tokens_seen": 47700608, "step": 24780 }, { "epoch": 4.04323354270332, "grad_norm": 0.009106768295168877, "learning_rate": 1.582610323145727e-05, "loss": 0.0034, "num_input_tokens_seen": 47710288, "step": 24785 }, { "epoch": 4.0440492699241375, "grad_norm": 0.0014110731426626444, "learning_rate": 1.5816971345406035e-05, "loss": 0.0492, "num_input_tokens_seen": 47720528, "step": 24790 }, { "epoch": 4.044864997144955, "grad_norm": 0.1264573484659195, "learning_rate": 1.5807840875493225e-05, "loss": 0.0637, "num_input_tokens_seen": 47730912, "step": 24795 }, { "epoch": 4.045680724365772, "grad_norm": 0.1316433548927307, "learning_rate": 1.5798711823126854e-05, "loss": 0.0025, "num_input_tokens_seen": 47741360, "step": 24800 }, { "epoch": 4.045680724365772, "eval_loss": 0.28094929456710815, "eval_runtime": 80.8081, "eval_samples_per_second": 33.722, "eval_steps_per_second": 16.867, "num_input_tokens_seen": 47741360, "step": 24800 }, { "epoch": 4.04649645158659, "grad_norm": 0.0322575643658638, "learning_rate": 1.578958418971477e-05, "loss": 0.0007, "num_input_tokens_seen": 47750032, "step": 24805 }, { "epoch": 4.047312178807407, "grad_norm": 0.02060464769601822, "learning_rate": 1.578045797666453e-05, "loss": 0.1573, "num_input_tokens_seen": 47759648, "step": 24810 }, { "epoch": 4.048127906028224, "grad_norm": 0.18679948151111603, "learning_rate": 1.5771333185383548e-05, "loss": 0.0614, "num_input_tokens_seen": 47769680, "step": 24815 }, { "epoch": 4.048943633249041, "grad_norm": 0.17369987070560455, "learning_rate": 1.576220981727895e-05, "loss": 0.001, "num_input_tokens_seen": 47780448, "step": 24820 }, { "epoch": 4.049759360469859, "grad_norm": 0.032723069190979004, "learning_rate": 1.575308787375769e-05, "loss": 0.0006, "num_input_tokens_seen": 47788704, "step": 24825 }, { "epoch": 4.050575087690676, "grad_norm": 0.012539477087557316, "learning_rate": 1.5743967356226492e-05, "loss": 0.001, "num_input_tokens_seen": 47799168, "step": 24830 }, { "epoch": 4.051390814911493, "grad_norm": 0.009959574788808823, "learning_rate": 1.5734848266091835e-05, "loss": 0.0801, "num_input_tokens_seen": 47810176, "step": 24835 }, { "epoch": 4.052206542132311, "grad_norm": 0.0822744369506836, "learning_rate": 1.572573060476001e-05, "loss": 0.0256, "num_input_tokens_seen": 47819728, "step": 24840 }, { "epoch": 4.053022269353129, "grad_norm": 0.021424904465675354, "learning_rate": 1.5716614373637085e-05, "loss": 0.001, "num_input_tokens_seen": 47829264, "step": 24845 }, { "epoch": 4.053837996573946, "grad_norm": 0.0635121613740921, "learning_rate": 1.570749957412887e-05, "loss": 0.0003, "num_input_tokens_seen": 47838288, "step": 24850 }, { "epoch": 4.054653723794763, "grad_norm": 0.011067149229347706, "learning_rate": 1.5698386207641013e-05, "loss": 0.0398, "num_input_tokens_seen": 47848000, "step": 24855 }, { "epoch": 4.05546945101558, "grad_norm": 0.027723047882318497, "learning_rate": 1.5689274275578884e-05, "loss": 0.0032, "num_input_tokens_seen": 47858656, "step": 24860 }, { "epoch": 4.056285178236398, "grad_norm": 0.0363859124481678, "learning_rate": 1.5680163779347667e-05, "loss": 0.0394, "num_input_tokens_seen": 47868592, "step": 24865 }, { "epoch": 4.057100905457215, "grad_norm": 0.4519597589969635, "learning_rate": 1.5671054720352327e-05, "loss": 0.0012, "num_input_tokens_seen": 47878112, "step": 24870 }, { "epoch": 4.057916632678032, "grad_norm": 16.602123260498047, "learning_rate": 1.566194709999757e-05, "loss": 0.0555, "num_input_tokens_seen": 47888736, "step": 24875 }, { "epoch": 4.058732359898849, "grad_norm": 0.19579163193702698, "learning_rate": 1.5652840919687933e-05, "loss": 0.0004, "num_input_tokens_seen": 47897968, "step": 24880 }, { "epoch": 4.059548087119667, "grad_norm": 0.016514956951141357, "learning_rate": 1.5643736180827676e-05, "loss": 0.0002, "num_input_tokens_seen": 47906912, "step": 24885 }, { "epoch": 4.0603638143404845, "grad_norm": 0.3894067704677582, "learning_rate": 1.5634632884820878e-05, "loss": 0.0016, "num_input_tokens_seen": 47916656, "step": 24890 }, { "epoch": 4.061179541561302, "grad_norm": 0.0437079593539238, "learning_rate": 1.5625531033071395e-05, "loss": 0.0115, "num_input_tokens_seen": 47926688, "step": 24895 }, { "epoch": 4.06199526878212, "grad_norm": 31.4813289642334, "learning_rate": 1.5616430626982828e-05, "loss": 0.0873, "num_input_tokens_seen": 47936656, "step": 24900 }, { "epoch": 4.062810996002937, "grad_norm": 0.01585392653942108, "learning_rate": 1.5607331667958575e-05, "loss": 0.0392, "num_input_tokens_seen": 47945760, "step": 24905 }, { "epoch": 4.063626723223754, "grad_norm": 0.02878882922232151, "learning_rate": 1.5598234157401824e-05, "loss": 0.0005, "num_input_tokens_seen": 47955328, "step": 24910 }, { "epoch": 4.064442450444571, "grad_norm": 0.01308412291109562, "learning_rate": 1.5589138096715503e-05, "loss": 0.065, "num_input_tokens_seen": 47964512, "step": 24915 }, { "epoch": 4.065258177665389, "grad_norm": 0.011524206958711147, "learning_rate": 1.5580043487302365e-05, "loss": 0.1193, "num_input_tokens_seen": 47973600, "step": 24920 }, { "epoch": 4.066073904886206, "grad_norm": 0.008461885154247284, "learning_rate": 1.5570950330564888e-05, "loss": 0.1218, "num_input_tokens_seen": 47982256, "step": 24925 }, { "epoch": 4.066889632107023, "grad_norm": 0.004268927033990622, "learning_rate": 1.5561858627905367e-05, "loss": 0.0004, "num_input_tokens_seen": 47992736, "step": 24930 }, { "epoch": 4.0677053593278405, "grad_norm": 0.028923628851771355, "learning_rate": 1.5552768380725857e-05, "loss": 0.0004, "num_input_tokens_seen": 48002816, "step": 24935 }, { "epoch": 4.0685210865486585, "grad_norm": 0.08614329248666763, "learning_rate": 1.5543679590428183e-05, "loss": 0.0005, "num_input_tokens_seen": 48011792, "step": 24940 }, { "epoch": 4.069336813769476, "grad_norm": 0.04956847429275513, "learning_rate": 1.5534592258413943e-05, "loss": 0.002, "num_input_tokens_seen": 48021296, "step": 24945 }, { "epoch": 4.070152540990293, "grad_norm": 0.04258593171834946, "learning_rate": 1.5525506386084538e-05, "loss": 0.0004, "num_input_tokens_seen": 48030528, "step": 24950 }, { "epoch": 4.07096826821111, "grad_norm": 0.23239527642726898, "learning_rate": 1.55164219748411e-05, "loss": 0.0019, "num_input_tokens_seen": 48041312, "step": 24955 }, { "epoch": 4.071783995431928, "grad_norm": 0.5418606996536255, "learning_rate": 1.550733902608459e-05, "loss": 0.0008, "num_input_tokens_seen": 48051056, "step": 24960 }, { "epoch": 4.072599722652745, "grad_norm": 0.009863296523690224, "learning_rate": 1.549825754121568e-05, "loss": 0.0059, "num_input_tokens_seen": 48060560, "step": 24965 }, { "epoch": 4.073415449873562, "grad_norm": 0.03339235112071037, "learning_rate": 1.5489177521634864e-05, "loss": 0.0026, "num_input_tokens_seen": 48070368, "step": 24970 }, { "epoch": 4.074231177094379, "grad_norm": 48.5400390625, "learning_rate": 1.5480098968742402e-05, "loss": 0.0576, "num_input_tokens_seen": 48079968, "step": 24975 }, { "epoch": 4.075046904315197, "grad_norm": 0.0018725974950939417, "learning_rate": 1.5471021883938304e-05, "loss": 0.0005, "num_input_tokens_seen": 48090000, "step": 24980 }, { "epoch": 4.0758626315360145, "grad_norm": 0.050902511924505234, "learning_rate": 1.546194626862238e-05, "loss": 0.001, "num_input_tokens_seen": 48100672, "step": 24985 }, { "epoch": 4.076678358756832, "grad_norm": 0.3964039981365204, "learning_rate": 1.5452872124194216e-05, "loss": 0.0008, "num_input_tokens_seen": 48111088, "step": 24990 }, { "epoch": 4.077494085977649, "grad_norm": 0.010497546754777431, "learning_rate": 1.5443799452053136e-05, "loss": 0.0015, "num_input_tokens_seen": 48121360, "step": 24995 }, { "epoch": 4.078309813198467, "grad_norm": 0.0035301283933222294, "learning_rate": 1.543472825359828e-05, "loss": 0.0002, "num_input_tokens_seen": 48131120, "step": 25000 }, { "epoch": 4.078309813198467, "eval_loss": 0.30152493715286255, "eval_runtime": 80.802, "eval_samples_per_second": 33.724, "eval_steps_per_second": 16.868, "num_input_tokens_seen": 48131120, "step": 25000 }, { "epoch": 4.079125540419284, "grad_norm": 0.0049644592218101025, "learning_rate": 1.5425658530228522e-05, "loss": 0.0009, "num_input_tokens_seen": 48140752, "step": 25005 }, { "epoch": 4.079941267640101, "grad_norm": 0.007858151569962502, "learning_rate": 1.5416590283342546e-05, "loss": 0.0008, "num_input_tokens_seen": 48150208, "step": 25010 }, { "epoch": 4.080756994860918, "grad_norm": 0.001734524848870933, "learning_rate": 1.5407523514338783e-05, "loss": 0.0001, "num_input_tokens_seen": 48160288, "step": 25015 }, { "epoch": 4.081572722081736, "grad_norm": 0.004718372598290443, "learning_rate": 1.539845822461543e-05, "loss": 0.0003, "num_input_tokens_seen": 48169744, "step": 25020 }, { "epoch": 4.082388449302553, "grad_norm": 34.24982452392578, "learning_rate": 1.538939441557048e-05, "loss": 0.0846, "num_input_tokens_seen": 48178528, "step": 25025 }, { "epoch": 4.08320417652337, "grad_norm": 0.005440251901745796, "learning_rate": 1.5380332088601696e-05, "loss": 0.116, "num_input_tokens_seen": 48187376, "step": 25030 }, { "epoch": 4.0840199037441876, "grad_norm": 0.08615681529045105, "learning_rate": 1.537127124510658e-05, "loss": 0.0002, "num_input_tokens_seen": 48198176, "step": 25035 }, { "epoch": 4.084835630965006, "grad_norm": 0.0166386179625988, "learning_rate": 1.5362211886482457e-05, "loss": 0.0652, "num_input_tokens_seen": 48207760, "step": 25040 }, { "epoch": 4.085651358185823, "grad_norm": 0.010905846953392029, "learning_rate": 1.5353154014126363e-05, "loss": 0.0671, "num_input_tokens_seen": 48217376, "step": 25045 }, { "epoch": 4.08646708540664, "grad_norm": 0.08945515751838684, "learning_rate": 1.534409762943515e-05, "loss": 0.0009, "num_input_tokens_seen": 48226304, "step": 25050 }, { "epoch": 4.087282812627457, "grad_norm": 0.00025104766245931387, "learning_rate": 1.5335042733805438e-05, "loss": 0.0005, "num_input_tokens_seen": 48235712, "step": 25055 }, { "epoch": 4.088098539848275, "grad_norm": 0.028660941869020462, "learning_rate": 1.532598932863358e-05, "loss": 0.0511, "num_input_tokens_seen": 48245328, "step": 25060 }, { "epoch": 4.088914267069092, "grad_norm": 0.0369504913687706, "learning_rate": 1.531693741531574e-05, "loss": 0.0004, "num_input_tokens_seen": 48255680, "step": 25065 }, { "epoch": 4.089729994289909, "grad_norm": 0.06170933321118355, "learning_rate": 1.5307886995247844e-05, "loss": 0.001, "num_input_tokens_seen": 48264272, "step": 25070 }, { "epoch": 4.090545721510727, "grad_norm": 0.0005244009080342948, "learning_rate": 1.529883806982557e-05, "loss": 0.0028, "num_input_tokens_seen": 48273600, "step": 25075 }, { "epoch": 4.091361448731544, "grad_norm": 0.0023459920194000006, "learning_rate": 1.5289790640444376e-05, "loss": 0.0005, "num_input_tokens_seen": 48285040, "step": 25080 }, { "epoch": 4.0921771759523615, "grad_norm": 0.0012872301740571856, "learning_rate": 1.5280744708499494e-05, "loss": 0.0001, "num_input_tokens_seen": 48294784, "step": 25085 }, { "epoch": 4.092992903173179, "grad_norm": 0.010459679178893566, "learning_rate": 1.527170027538591e-05, "loss": 0.001, "num_input_tokens_seen": 48304640, "step": 25090 }, { "epoch": 4.093808630393997, "grad_norm": 0.004987538326531649, "learning_rate": 1.5262657342498407e-05, "loss": 0.0053, "num_input_tokens_seen": 48314976, "step": 25095 }, { "epoch": 4.094624357614814, "grad_norm": 0.11365418136119843, "learning_rate": 1.52536159112315e-05, "loss": 0.0014, "num_input_tokens_seen": 48323808, "step": 25100 }, { "epoch": 4.095440084835631, "grad_norm": 0.025314530357718468, "learning_rate": 1.5244575982979497e-05, "loss": 0.0005, "num_input_tokens_seen": 48332832, "step": 25105 }, { "epoch": 4.096255812056448, "grad_norm": 0.017930861562490463, "learning_rate": 1.5235537559136487e-05, "loss": 0.0002, "num_input_tokens_seen": 48341696, "step": 25110 }, { "epoch": 4.097071539277266, "grad_norm": 0.0022739567793905735, "learning_rate": 1.5226500641096286e-05, "loss": 0.0006, "num_input_tokens_seen": 48350400, "step": 25115 }, { "epoch": 4.097887266498083, "grad_norm": 0.001938443980179727, "learning_rate": 1.5217465230252509e-05, "loss": 0.0001, "num_input_tokens_seen": 48361424, "step": 25120 }, { "epoch": 4.0987029937189, "grad_norm": 24.894914627075195, "learning_rate": 1.5208431327998523e-05, "loss": 0.167, "num_input_tokens_seen": 48370064, "step": 25125 }, { "epoch": 4.0995187209397175, "grad_norm": 0.005086456425487995, "learning_rate": 1.5199398935727477e-05, "loss": 0.0018, "num_input_tokens_seen": 48380784, "step": 25130 }, { "epoch": 4.1003344481605355, "grad_norm": 0.1125529408454895, "learning_rate": 1.5190368054832282e-05, "loss": 0.0009, "num_input_tokens_seen": 48390480, "step": 25135 }, { "epoch": 4.101150175381353, "grad_norm": 0.041917331516742706, "learning_rate": 1.5181338686705601e-05, "loss": 0.0006, "num_input_tokens_seen": 48400720, "step": 25140 }, { "epoch": 4.10196590260217, "grad_norm": 0.007162280846387148, "learning_rate": 1.5172310832739889e-05, "loss": 0.0003, "num_input_tokens_seen": 48410208, "step": 25145 }, { "epoch": 4.102781629822987, "grad_norm": 0.22528891265392303, "learning_rate": 1.5163284494327346e-05, "loss": 0.0003, "num_input_tokens_seen": 48419776, "step": 25150 }, { "epoch": 4.103597357043805, "grad_norm": 0.0004449961124919355, "learning_rate": 1.5154259672859952e-05, "loss": 0.0024, "num_input_tokens_seen": 48428624, "step": 25155 }, { "epoch": 4.104413084264622, "grad_norm": 0.013554660603404045, "learning_rate": 1.5145236369729452e-05, "loss": 0.0001, "num_input_tokens_seen": 48437680, "step": 25160 }, { "epoch": 4.105228811485439, "grad_norm": 0.27067843079566956, "learning_rate": 1.5136214586327335e-05, "loss": 0.0029, "num_input_tokens_seen": 48446560, "step": 25165 }, { "epoch": 4.106044538706256, "grad_norm": 0.0032543863635510206, "learning_rate": 1.5127194324044885e-05, "loss": 0.0002, "num_input_tokens_seen": 48455328, "step": 25170 }, { "epoch": 4.106860265927074, "grad_norm": 39.17627716064453, "learning_rate": 1.5118175584273148e-05, "loss": 0.0873, "num_input_tokens_seen": 48464176, "step": 25175 }, { "epoch": 4.1076759931478914, "grad_norm": 0.09623521566390991, "learning_rate": 1.5109158368402909e-05, "loss": 0.0738, "num_input_tokens_seen": 48474096, "step": 25180 }, { "epoch": 4.108491720368709, "grad_norm": 0.02058033086359501, "learning_rate": 1.5100142677824753e-05, "loss": 0.0002, "num_input_tokens_seen": 48484144, "step": 25185 }, { "epoch": 4.109307447589526, "grad_norm": 0.03881360962986946, "learning_rate": 1.509112851392901e-05, "loss": 0.0286, "num_input_tokens_seen": 48493200, "step": 25190 }, { "epoch": 4.110123174810344, "grad_norm": 0.004550425335764885, "learning_rate": 1.5082115878105763e-05, "loss": 0.0983, "num_input_tokens_seen": 48503104, "step": 25195 }, { "epoch": 4.110938902031161, "grad_norm": 8.09701919555664, "learning_rate": 1.5073104771744892e-05, "loss": 0.0777, "num_input_tokens_seen": 48513200, "step": 25200 }, { "epoch": 4.110938902031161, "eval_loss": 0.34385573863983154, "eval_runtime": 80.8034, "eval_samples_per_second": 33.724, "eval_steps_per_second": 16.868, "num_input_tokens_seen": 48513200, "step": 25200 }, { "epoch": 4.111754629251978, "grad_norm": 0.006142204627394676, "learning_rate": 1.5064095196236006e-05, "loss": 0.0712, "num_input_tokens_seen": 48521232, "step": 25205 }, { "epoch": 4.112570356472795, "grad_norm": 0.036868032068014145, "learning_rate": 1.50550871529685e-05, "loss": 0.0712, "num_input_tokens_seen": 48531776, "step": 25210 }, { "epoch": 4.113386083693613, "grad_norm": 0.002552848309278488, "learning_rate": 1.5046080643331546e-05, "loss": 0.0, "num_input_tokens_seen": 48539856, "step": 25215 }, { "epoch": 4.11420181091443, "grad_norm": 0.02722635678946972, "learning_rate": 1.5037075668714028e-05, "loss": 0.0001, "num_input_tokens_seen": 48550928, "step": 25220 }, { "epoch": 4.115017538135247, "grad_norm": 0.031839147210121155, "learning_rate": 1.5028072230504656e-05, "loss": 0.0, "num_input_tokens_seen": 48560000, "step": 25225 }, { "epoch": 4.1158332653560645, "grad_norm": 0.24002070724964142, "learning_rate": 1.5019070330091861e-05, "loss": 0.0476, "num_input_tokens_seen": 48569328, "step": 25230 }, { "epoch": 4.1166489925768825, "grad_norm": 0.009369546547532082, "learning_rate": 1.5010069968863843e-05, "loss": 0.0003, "num_input_tokens_seen": 48579296, "step": 25235 }, { "epoch": 4.1174647197977, "grad_norm": 0.002563368296250701, "learning_rate": 1.5001071148208584e-05, "loss": 0.001, "num_input_tokens_seen": 48589536, "step": 25240 }, { "epoch": 4.118280447018517, "grad_norm": 0.03323278948664665, "learning_rate": 1.49920738695138e-05, "loss": 0.0003, "num_input_tokens_seen": 48599552, "step": 25245 }, { "epoch": 4.119096174239334, "grad_norm": 0.002293582074344158, "learning_rate": 1.4983078134166995e-05, "loss": 0.0002, "num_input_tokens_seen": 48609776, "step": 25250 }, { "epoch": 4.119911901460152, "grad_norm": 0.0017176505643874407, "learning_rate": 1.4974083943555428e-05, "loss": 0.045, "num_input_tokens_seen": 48620016, "step": 25255 }, { "epoch": 4.120727628680969, "grad_norm": 0.0028694237116724253, "learning_rate": 1.496509129906611e-05, "loss": 0.0005, "num_input_tokens_seen": 48630576, "step": 25260 }, { "epoch": 4.121543355901786, "grad_norm": 0.011396225541830063, "learning_rate": 1.4956100202085809e-05, "loss": 0.0374, "num_input_tokens_seen": 48638896, "step": 25265 }, { "epoch": 4.122359083122603, "grad_norm": 0.0035782770719379187, "learning_rate": 1.4947110654001093e-05, "loss": 0.0003, "num_input_tokens_seen": 48649088, "step": 25270 }, { "epoch": 4.123174810343421, "grad_norm": 0.004682477563619614, "learning_rate": 1.4938122656198234e-05, "loss": 0.0002, "num_input_tokens_seen": 48658000, "step": 25275 }, { "epoch": 4.1239905375642385, "grad_norm": 0.0041271718218922615, "learning_rate": 1.4929136210063316e-05, "loss": 0.0001, "num_input_tokens_seen": 48667216, "step": 25280 }, { "epoch": 4.124806264785056, "grad_norm": 0.02926722727715969, "learning_rate": 1.4920151316982146e-05, "loss": 0.0001, "num_input_tokens_seen": 48676240, "step": 25285 }, { "epoch": 4.125621992005874, "grad_norm": 0.008742683567106724, "learning_rate": 1.4911167978340312e-05, "loss": 0.1107, "num_input_tokens_seen": 48685488, "step": 25290 }, { "epoch": 4.126437719226691, "grad_norm": 0.011970361694693565, "learning_rate": 1.4902186195523166e-05, "loss": 0.0002, "num_input_tokens_seen": 48695808, "step": 25295 }, { "epoch": 4.127253446447508, "grad_norm": 0.008450365625321865, "learning_rate": 1.4893205969915805e-05, "loss": 0.0003, "num_input_tokens_seen": 48705840, "step": 25300 }, { "epoch": 4.128069173668325, "grad_norm": 0.00044616463128477335, "learning_rate": 1.4884227302903086e-05, "loss": 0.0431, "num_input_tokens_seen": 48716000, "step": 25305 }, { "epoch": 4.128884900889143, "grad_norm": 0.0017520278925076127, "learning_rate": 1.4875250195869653e-05, "loss": 0.0992, "num_input_tokens_seen": 48723920, "step": 25310 }, { "epoch": 4.12970062810996, "grad_norm": 7.41831111907959, "learning_rate": 1.4866274650199862e-05, "loss": 0.1372, "num_input_tokens_seen": 48733248, "step": 25315 }, { "epoch": 4.130516355330777, "grad_norm": 0.0008146843174472451, "learning_rate": 1.485730066727788e-05, "loss": 0.0012, "num_input_tokens_seen": 48743296, "step": 25320 }, { "epoch": 4.1313320825515945, "grad_norm": 0.040522944182157516, "learning_rate": 1.4848328248487586e-05, "loss": 0.0005, "num_input_tokens_seen": 48752224, "step": 25325 }, { "epoch": 4.1321478097724125, "grad_norm": 17.124252319335938, "learning_rate": 1.4839357395212656e-05, "loss": 0.0888, "num_input_tokens_seen": 48761152, "step": 25330 }, { "epoch": 4.13296353699323, "grad_norm": 0.01254356186836958, "learning_rate": 1.4830388108836502e-05, "loss": 0.0005, "num_input_tokens_seen": 48770768, "step": 25335 }, { "epoch": 4.133779264214047, "grad_norm": 0.009608396328985691, "learning_rate": 1.4821420390742299e-05, "loss": 0.0003, "num_input_tokens_seen": 48780496, "step": 25340 }, { "epoch": 4.134594991434864, "grad_norm": 0.025942504405975342, "learning_rate": 1.4812454242312979e-05, "loss": 0.0001, "num_input_tokens_seen": 48789808, "step": 25345 }, { "epoch": 4.135410718655682, "grad_norm": 0.03256136178970337, "learning_rate": 1.4803489664931253e-05, "loss": 0.0009, "num_input_tokens_seen": 48798976, "step": 25350 }, { "epoch": 4.136226445876499, "grad_norm": 0.006307831034064293, "learning_rate": 1.4794526659979544e-05, "loss": 0.0001, "num_input_tokens_seen": 48807808, "step": 25355 }, { "epoch": 4.137042173097316, "grad_norm": 6.784500598907471, "learning_rate": 1.4785565228840086e-05, "loss": 0.1227, "num_input_tokens_seen": 48816672, "step": 25360 }, { "epoch": 4.137857900318133, "grad_norm": 0.00750777218490839, "learning_rate": 1.4776605372894819e-05, "loss": 0.0007, "num_input_tokens_seen": 48826720, "step": 25365 }, { "epoch": 4.138673627538951, "grad_norm": 0.018230482935905457, "learning_rate": 1.4767647093525488e-05, "loss": 0.0002, "num_input_tokens_seen": 48836464, "step": 25370 }, { "epoch": 4.139489354759768, "grad_norm": 0.279835969209671, "learning_rate": 1.4758690392113566e-05, "loss": 0.0008, "num_input_tokens_seen": 48845808, "step": 25375 }, { "epoch": 4.1403050819805856, "grad_norm": 0.194243386387825, "learning_rate": 1.4749735270040276e-05, "loss": 0.0005, "num_input_tokens_seen": 48855328, "step": 25380 }, { "epoch": 4.141120809201403, "grad_norm": 0.0275235828012228, "learning_rate": 1.4740781728686623e-05, "loss": 0.0001, "num_input_tokens_seen": 48865504, "step": 25385 }, { "epoch": 4.141936536422221, "grad_norm": 0.3329564929008484, "learning_rate": 1.4731829769433358e-05, "loss": 0.0006, "num_input_tokens_seen": 48874016, "step": 25390 }, { "epoch": 4.142752263643038, "grad_norm": 0.21243129670619965, "learning_rate": 1.4722879393660976e-05, "loss": 0.0179, "num_input_tokens_seen": 48884336, "step": 25395 }, { "epoch": 4.143567990863855, "grad_norm": 1.0618553161621094, "learning_rate": 1.4713930602749748e-05, "loss": 0.0003, "num_input_tokens_seen": 48894496, "step": 25400 }, { "epoch": 4.143567990863855, "eval_loss": 0.3471934497356415, "eval_runtime": 80.9146, "eval_samples_per_second": 33.678, "eval_steps_per_second": 16.845, "num_input_tokens_seen": 48894496, "step": 25400 }, { "epoch": 4.144383718084672, "grad_norm": 0.4986124634742737, "learning_rate": 1.470498339807968e-05, "loss": 0.0006, "num_input_tokens_seen": 48903408, "step": 25405 }, { "epoch": 4.14519944530549, "grad_norm": 0.01012337300926447, "learning_rate": 1.4696037781030542e-05, "loss": 0.1551, "num_input_tokens_seen": 48912624, "step": 25410 }, { "epoch": 4.146015172526307, "grad_norm": 0.001453577191568911, "learning_rate": 1.4687093752981876e-05, "loss": 0.0002, "num_input_tokens_seen": 48922544, "step": 25415 }, { "epoch": 4.146830899747124, "grad_norm": 0.006849515251815319, "learning_rate": 1.4678151315312943e-05, "loss": 0.0001, "num_input_tokens_seen": 48931984, "step": 25420 }, { "epoch": 4.1476466269679415, "grad_norm": 0.0021802112460136414, "learning_rate": 1.4669210469402789e-05, "loss": 0.0027, "num_input_tokens_seen": 48941872, "step": 25425 }, { "epoch": 4.1484623541887595, "grad_norm": 0.0004160350072197616, "learning_rate": 1.4660271216630218e-05, "loss": 0.0002, "num_input_tokens_seen": 48952576, "step": 25430 }, { "epoch": 4.149278081409577, "grad_norm": 0.06035448610782623, "learning_rate": 1.4651333558373748e-05, "loss": 0.0003, "num_input_tokens_seen": 48962560, "step": 25435 }, { "epoch": 4.150093808630394, "grad_norm": 0.03161610662937164, "learning_rate": 1.4642397496011707e-05, "loss": 0.1335, "num_input_tokens_seen": 48971136, "step": 25440 }, { "epoch": 4.150909535851211, "grad_norm": 0.003473455784842372, "learning_rate": 1.4633463030922129e-05, "loss": 0.0001, "num_input_tokens_seen": 48980144, "step": 25445 }, { "epoch": 4.151725263072029, "grad_norm": 0.012153146788477898, "learning_rate": 1.462453016448282e-05, "loss": 0.0001, "num_input_tokens_seen": 48989856, "step": 25450 }, { "epoch": 4.152540990292846, "grad_norm": 0.0022665075957775116, "learning_rate": 1.4615598898071354e-05, "loss": 0.0003, "num_input_tokens_seen": 49000528, "step": 25455 }, { "epoch": 4.153356717513663, "grad_norm": 14.563397407531738, "learning_rate": 1.4606669233065026e-05, "loss": 0.0977, "num_input_tokens_seen": 49009744, "step": 25460 }, { "epoch": 4.154172444734481, "grad_norm": 0.005002403166145086, "learning_rate": 1.4597741170840914e-05, "loss": 0.0002, "num_input_tokens_seen": 49018816, "step": 25465 }, { "epoch": 4.154988171955298, "grad_norm": 0.024871278554201126, "learning_rate": 1.4588814712775853e-05, "loss": 0.0005, "num_input_tokens_seen": 49028912, "step": 25470 }, { "epoch": 4.1558038991761155, "grad_norm": 0.014005659148097038, "learning_rate": 1.4579889860246382e-05, "loss": 0.0003, "num_input_tokens_seen": 49039440, "step": 25475 }, { "epoch": 4.156619626396933, "grad_norm": 0.0020793324802070856, "learning_rate": 1.457096661462885e-05, "loss": 0.0003, "num_input_tokens_seen": 49050176, "step": 25480 }, { "epoch": 4.157435353617751, "grad_norm": 0.08226029574871063, "learning_rate": 1.4562044977299322e-05, "loss": 0.0011, "num_input_tokens_seen": 49059504, "step": 25485 }, { "epoch": 4.158251080838568, "grad_norm": 0.0007905582315288484, "learning_rate": 1.4553124949633623e-05, "loss": 0.0342, "num_input_tokens_seen": 49069280, "step": 25490 }, { "epoch": 4.159066808059385, "grad_norm": 0.0014207548229023814, "learning_rate": 1.4544206533007354e-05, "loss": 0.0001, "num_input_tokens_seen": 49077360, "step": 25495 }, { "epoch": 4.159882535280202, "grad_norm": 0.021081652492284775, "learning_rate": 1.4535289728795821e-05, "loss": 0.0061, "num_input_tokens_seen": 49087312, "step": 25500 }, { "epoch": 4.16069826250102, "grad_norm": 0.061852194368839264, "learning_rate": 1.4526374538374132e-05, "loss": 0.002, "num_input_tokens_seen": 49097776, "step": 25505 }, { "epoch": 4.161513989721837, "grad_norm": 0.04086807370185852, "learning_rate": 1.4517460963117097e-05, "loss": 0.0002, "num_input_tokens_seen": 49107024, "step": 25510 }, { "epoch": 4.162329716942654, "grad_norm": 0.04808320105075836, "learning_rate": 1.4508549004399314e-05, "loss": 0.0006, "num_input_tokens_seen": 49116400, "step": 25515 }, { "epoch": 4.163145444163471, "grad_norm": 0.1742749810218811, "learning_rate": 1.449963866359513e-05, "loss": 0.0004, "num_input_tokens_seen": 49126048, "step": 25520 }, { "epoch": 4.1639611713842895, "grad_norm": 12.550145149230957, "learning_rate": 1.4490729942078607e-05, "loss": 0.0981, "num_input_tokens_seen": 49136176, "step": 25525 }, { "epoch": 4.164776898605107, "grad_norm": 0.0005936824018135667, "learning_rate": 1.4481822841223608e-05, "loss": 0.0752, "num_input_tokens_seen": 49145952, "step": 25530 }, { "epoch": 4.165592625825924, "grad_norm": 0.025811195373535156, "learning_rate": 1.4472917362403704e-05, "loss": 0.0177, "num_input_tokens_seen": 49155536, "step": 25535 }, { "epoch": 4.166408353046741, "grad_norm": 0.0010026388335973024, "learning_rate": 1.4464013506992224e-05, "loss": 0.0001, "num_input_tokens_seen": 49164624, "step": 25540 }, { "epoch": 4.167224080267559, "grad_norm": 0.001719532534480095, "learning_rate": 1.4455111276362277e-05, "loss": 0.0008, "num_input_tokens_seen": 49173552, "step": 25545 }, { "epoch": 4.168039807488376, "grad_norm": 0.3652609586715698, "learning_rate": 1.4446210671886676e-05, "loss": 0.0288, "num_input_tokens_seen": 49182240, "step": 25550 }, { "epoch": 4.168855534709193, "grad_norm": 0.1573914885520935, "learning_rate": 1.4437311694938015e-05, "loss": 0.0004, "num_input_tokens_seen": 49192784, "step": 25555 }, { "epoch": 4.16967126193001, "grad_norm": 0.009031890891492367, "learning_rate": 1.442841434688864e-05, "loss": 0.0386, "num_input_tokens_seen": 49202992, "step": 25560 }, { "epoch": 4.170486989150828, "grad_norm": 0.028110604733228683, "learning_rate": 1.4419518629110615e-05, "loss": 0.0006, "num_input_tokens_seen": 49213648, "step": 25565 }, { "epoch": 4.171302716371645, "grad_norm": 0.0352906733751297, "learning_rate": 1.4410624542975778e-05, "loss": 0.0001, "num_input_tokens_seen": 49221568, "step": 25570 }, { "epoch": 4.1721184435924625, "grad_norm": 0.004607397131621838, "learning_rate": 1.4401732089855724e-05, "loss": 0.0001, "num_input_tokens_seen": 49231568, "step": 25575 }, { "epoch": 4.17293417081328, "grad_norm": 0.11152482032775879, "learning_rate": 1.4392841271121754e-05, "loss": 0.0006, "num_input_tokens_seen": 49241472, "step": 25580 }, { "epoch": 4.173749898034098, "grad_norm": 0.05409608408808708, "learning_rate": 1.438395208814497e-05, "loss": 0.0171, "num_input_tokens_seen": 49250816, "step": 25585 }, { "epoch": 4.174565625254915, "grad_norm": 0.0016910518752411008, "learning_rate": 1.4375064542296174e-05, "loss": 0.0001, "num_input_tokens_seen": 49259792, "step": 25590 }, { "epoch": 4.175381352475732, "grad_norm": 0.02315112203359604, "learning_rate": 1.4366178634945946e-05, "loss": 0.0001, "num_input_tokens_seen": 49270624, "step": 25595 }, { "epoch": 4.176197079696549, "grad_norm": 0.018210064619779587, "learning_rate": 1.4357294367464616e-05, "loss": 0.059, "num_input_tokens_seen": 49280736, "step": 25600 }, { "epoch": 4.176197079696549, "eval_loss": 0.34600019454956055, "eval_runtime": 80.7076, "eval_samples_per_second": 33.764, "eval_steps_per_second": 16.888, "num_input_tokens_seen": 49280736, "step": 25600 }, { "epoch": 4.177012806917367, "grad_norm": 0.0013478301698341966, "learning_rate": 1.434841174122224e-05, "loss": 0.0001, "num_input_tokens_seen": 49289696, "step": 25605 }, { "epoch": 4.177828534138184, "grad_norm": 0.8006335496902466, "learning_rate": 1.4339530757588615e-05, "loss": 0.0016, "num_input_tokens_seen": 49299648, "step": 25610 }, { "epoch": 4.178644261359001, "grad_norm": 6.429629802703857, "learning_rate": 1.433065141793333e-05, "loss": 0.0044, "num_input_tokens_seen": 49309136, "step": 25615 }, { "epoch": 4.1794599885798185, "grad_norm": 0.07878509163856506, "learning_rate": 1.4321773723625665e-05, "loss": 0.002, "num_input_tokens_seen": 49318960, "step": 25620 }, { "epoch": 4.1802757158006365, "grad_norm": 0.0010459866607561707, "learning_rate": 1.4312897676034693e-05, "loss": 0.1691, "num_input_tokens_seen": 49328624, "step": 25625 }, { "epoch": 4.181091443021454, "grad_norm": 0.056737545877695084, "learning_rate": 1.4304023276529188e-05, "loss": 0.0007, "num_input_tokens_seen": 49339472, "step": 25630 }, { "epoch": 4.181907170242271, "grad_norm": 0.03182247281074524, "learning_rate": 1.4295150526477712e-05, "loss": 0.0692, "num_input_tokens_seen": 49348704, "step": 25635 }, { "epoch": 4.182722897463089, "grad_norm": 0.1388068050146103, "learning_rate": 1.4286279427248562e-05, "loss": 0.0535, "num_input_tokens_seen": 49357600, "step": 25640 }, { "epoch": 4.183538624683906, "grad_norm": 0.007649614941328764, "learning_rate": 1.4277409980209747e-05, "loss": 0.1285, "num_input_tokens_seen": 49366896, "step": 25645 }, { "epoch": 4.184354351904723, "grad_norm": 0.03364101052284241, "learning_rate": 1.4268542186729061e-05, "loss": 0.0005, "num_input_tokens_seen": 49375792, "step": 25650 }, { "epoch": 4.18517007912554, "grad_norm": 0.008624524809420109, "learning_rate": 1.4259676048174043e-05, "loss": 0.0242, "num_input_tokens_seen": 49385328, "step": 25655 }, { "epoch": 4.185985806346358, "grad_norm": 0.0030206735245883465, "learning_rate": 1.4250811565911937e-05, "loss": 0.0292, "num_input_tokens_seen": 49394640, "step": 25660 }, { "epoch": 4.186801533567175, "grad_norm": 0.004434315022081137, "learning_rate": 1.4241948741309782e-05, "loss": 0.0003, "num_input_tokens_seen": 49404000, "step": 25665 }, { "epoch": 4.1876172607879925, "grad_norm": 0.006189236883074045, "learning_rate": 1.4233087575734317e-05, "loss": 0.0074, "num_input_tokens_seen": 49412400, "step": 25670 }, { "epoch": 4.18843298800881, "grad_norm": 0.0016688453033566475, "learning_rate": 1.422422807055206e-05, "loss": 0.0028, "num_input_tokens_seen": 49423504, "step": 25675 }, { "epoch": 4.189248715229628, "grad_norm": 0.3669392168521881, "learning_rate": 1.4215370227129243e-05, "loss": 0.0006, "num_input_tokens_seen": 49432816, "step": 25680 }, { "epoch": 4.190064442450445, "grad_norm": 0.12147662043571472, "learning_rate": 1.4206514046831876e-05, "loss": 0.0006, "num_input_tokens_seen": 49443440, "step": 25685 }, { "epoch": 4.190880169671262, "grad_norm": 0.004125605802983046, "learning_rate": 1.419765953102567e-05, "loss": 0.115, "num_input_tokens_seen": 49452928, "step": 25690 }, { "epoch": 4.191695896892079, "grad_norm": 0.22745098173618317, "learning_rate": 1.4188806681076125e-05, "loss": 0.0256, "num_input_tokens_seen": 49462480, "step": 25695 }, { "epoch": 4.192511624112897, "grad_norm": 0.00023403146769851446, "learning_rate": 1.4179955498348443e-05, "loss": 0.001, "num_input_tokens_seen": 49471344, "step": 25700 }, { "epoch": 4.193327351333714, "grad_norm": 0.018868349492549896, "learning_rate": 1.4171105984207605e-05, "loss": 0.0004, "num_input_tokens_seen": 49480880, "step": 25705 }, { "epoch": 4.194143078554531, "grad_norm": 0.007115568034350872, "learning_rate": 1.4162258140018304e-05, "loss": 0.0002, "num_input_tokens_seen": 49491520, "step": 25710 }, { "epoch": 4.194958805775348, "grad_norm": 0.005087337456643581, "learning_rate": 1.4153411967144986e-05, "loss": 0.0029, "num_input_tokens_seen": 49501584, "step": 25715 }, { "epoch": 4.195774532996166, "grad_norm": 0.0017590692732483149, "learning_rate": 1.4144567466951864e-05, "loss": 0.0613, "num_input_tokens_seen": 49510944, "step": 25720 }, { "epoch": 4.196590260216984, "grad_norm": 0.024553196504712105, "learning_rate": 1.4135724640802844e-05, "loss": 0.0002, "num_input_tokens_seen": 49520272, "step": 25725 }, { "epoch": 4.197405987437801, "grad_norm": 0.08808231353759766, "learning_rate": 1.4126883490061615e-05, "loss": 0.0103, "num_input_tokens_seen": 49530544, "step": 25730 }, { "epoch": 4.198221714658618, "grad_norm": 0.0026220111176371574, "learning_rate": 1.4118044016091603e-05, "loss": 0.0001, "num_input_tokens_seen": 49540560, "step": 25735 }, { "epoch": 4.199037441879436, "grad_norm": 0.0006192632718011737, "learning_rate": 1.410920622025594e-05, "loss": 0.0002, "num_input_tokens_seen": 49548928, "step": 25740 }, { "epoch": 4.199853169100253, "grad_norm": 0.024585548788309097, "learning_rate": 1.4100370103917554e-05, "loss": 0.0671, "num_input_tokens_seen": 49557744, "step": 25745 }, { "epoch": 4.20066889632107, "grad_norm": 0.050641462206840515, "learning_rate": 1.409153566843907e-05, "loss": 0.0003, "num_input_tokens_seen": 49567376, "step": 25750 }, { "epoch": 4.201484623541887, "grad_norm": 0.0032806340605020523, "learning_rate": 1.408270291518286e-05, "loss": 0.0002, "num_input_tokens_seen": 49577632, "step": 25755 }, { "epoch": 4.202300350762705, "grad_norm": 0.001166615984402597, "learning_rate": 1.407387184551107e-05, "loss": 0.0205, "num_input_tokens_seen": 49587600, "step": 25760 }, { "epoch": 4.203116077983522, "grad_norm": 0.022229013964533806, "learning_rate": 1.4065042460785532e-05, "loss": 0.001, "num_input_tokens_seen": 49596480, "step": 25765 }, { "epoch": 4.2039318052043395, "grad_norm": 0.005687819328159094, "learning_rate": 1.405621476236787e-05, "loss": 0.1191, "num_input_tokens_seen": 49604896, "step": 25770 }, { "epoch": 4.204747532425157, "grad_norm": 0.035757072269916534, "learning_rate": 1.4047388751619423e-05, "loss": 0.0001, "num_input_tokens_seen": 49615248, "step": 25775 }, { "epoch": 4.205563259645975, "grad_norm": 0.0005072953645139933, "learning_rate": 1.4038564429901264e-05, "loss": 0.0835, "num_input_tokens_seen": 49625136, "step": 25780 }, { "epoch": 4.206378986866792, "grad_norm": 0.003972779028117657, "learning_rate": 1.4029741798574227e-05, "loss": 0.0054, "num_input_tokens_seen": 49634240, "step": 25785 }, { "epoch": 4.207194714087609, "grad_norm": 0.008946002461016178, "learning_rate": 1.402092085899886e-05, "loss": 0.0003, "num_input_tokens_seen": 49643456, "step": 25790 }, { "epoch": 4.208010441308426, "grad_norm": 0.01701081171631813, "learning_rate": 1.4012101612535464e-05, "loss": 0.0005, "num_input_tokens_seen": 49652896, "step": 25795 }, { "epoch": 4.208826168529244, "grad_norm": 1.3958977460861206, "learning_rate": 1.4003284060544092e-05, "loss": 0.0571, "num_input_tokens_seen": 49662304, "step": 25800 }, { "epoch": 4.208826168529244, "eval_loss": 0.3478054702281952, "eval_runtime": 80.7375, "eval_samples_per_second": 33.751, "eval_steps_per_second": 16.882, "num_input_tokens_seen": 49662304, "step": 25800 }, { "epoch": 4.209641895750061, "grad_norm": 0.006040541455149651, "learning_rate": 1.3994468204384504e-05, "loss": 0.0002, "num_input_tokens_seen": 49671504, "step": 25805 }, { "epoch": 4.210457622970878, "grad_norm": 0.004502724856138229, "learning_rate": 1.398565404541622e-05, "loss": 0.1403, "num_input_tokens_seen": 49681504, "step": 25810 }, { "epoch": 4.211273350191696, "grad_norm": 0.0006544798379763961, "learning_rate": 1.3976841584998513e-05, "loss": 0.0492, "num_input_tokens_seen": 49691616, "step": 25815 }, { "epoch": 4.2120890774125135, "grad_norm": 0.00337960128672421, "learning_rate": 1.3968030824490352e-05, "loss": 0.0486, "num_input_tokens_seen": 49700848, "step": 25820 }, { "epoch": 4.212904804633331, "grad_norm": 0.01426099892705679, "learning_rate": 1.3959221765250469e-05, "loss": 0.0001, "num_input_tokens_seen": 49711344, "step": 25825 }, { "epoch": 4.213720531854148, "grad_norm": 0.0074837701395154, "learning_rate": 1.3950414408637343e-05, "loss": 0.0021, "num_input_tokens_seen": 49719952, "step": 25830 }, { "epoch": 4.214536259074965, "grad_norm": 0.04813743755221367, "learning_rate": 1.3941608756009166e-05, "loss": 0.0003, "num_input_tokens_seen": 49728832, "step": 25835 }, { "epoch": 4.215351986295783, "grad_norm": 0.0016241914127022028, "learning_rate": 1.3932804808723898e-05, "loss": 0.0561, "num_input_tokens_seen": 49739264, "step": 25840 }, { "epoch": 4.2161677135166, "grad_norm": 0.0014190970687195659, "learning_rate": 1.3924002568139194e-05, "loss": 0.0, "num_input_tokens_seen": 49748720, "step": 25845 }, { "epoch": 4.216983440737417, "grad_norm": 0.009176288731396198, "learning_rate": 1.3915202035612485e-05, "loss": 0.0001, "num_input_tokens_seen": 49757920, "step": 25850 }, { "epoch": 4.217799167958235, "grad_norm": 8.907415390014648, "learning_rate": 1.3906403212500935e-05, "loss": 0.1883, "num_input_tokens_seen": 49767808, "step": 25855 }, { "epoch": 4.218614895179052, "grad_norm": 0.01592724584043026, "learning_rate": 1.3897606100161409e-05, "loss": 0.0092, "num_input_tokens_seen": 49777952, "step": 25860 }, { "epoch": 4.219430622399869, "grad_norm": 0.1833149641752243, "learning_rate": 1.388881069995055e-05, "loss": 0.0005, "num_input_tokens_seen": 49788928, "step": 25865 }, { "epoch": 4.220246349620687, "grad_norm": 0.07290448248386383, "learning_rate": 1.3880017013224708e-05, "loss": 0.0004, "num_input_tokens_seen": 49799008, "step": 25870 }, { "epoch": 4.221062076841505, "grad_norm": 231.2193603515625, "learning_rate": 1.3871225041339984e-05, "loss": 0.0168, "num_input_tokens_seen": 49808976, "step": 25875 }, { "epoch": 4.221877804062322, "grad_norm": 0.07989279925823212, "learning_rate": 1.386243478565222e-05, "loss": 0.0006, "num_input_tokens_seen": 49819456, "step": 25880 }, { "epoch": 4.222693531283139, "grad_norm": 0.07538764178752899, "learning_rate": 1.3853646247516966e-05, "loss": 0.0004, "num_input_tokens_seen": 49829152, "step": 25885 }, { "epoch": 4.223509258503956, "grad_norm": 0.00047165140858851373, "learning_rate": 1.3844859428289545e-05, "loss": 0.0003, "num_input_tokens_seen": 49838336, "step": 25890 }, { "epoch": 4.224324985724774, "grad_norm": 0.016114406287670135, "learning_rate": 1.3836074329324984e-05, "loss": 0.0053, "num_input_tokens_seen": 49848048, "step": 25895 }, { "epoch": 4.225140712945591, "grad_norm": 0.3649333417415619, "learning_rate": 1.3827290951978044e-05, "loss": 0.0147, "num_input_tokens_seen": 49858880, "step": 25900 }, { "epoch": 4.225956440166408, "grad_norm": 0.009507891722023487, "learning_rate": 1.381850929760326e-05, "loss": 0.0002, "num_input_tokens_seen": 49868992, "step": 25905 }, { "epoch": 4.226772167387225, "grad_norm": 0.010153367184102535, "learning_rate": 1.3809729367554842e-05, "loss": 0.0001, "num_input_tokens_seen": 49878032, "step": 25910 }, { "epoch": 4.227587894608043, "grad_norm": 0.015174352563917637, "learning_rate": 1.3800951163186784e-05, "loss": 0.1404, "num_input_tokens_seen": 49887792, "step": 25915 }, { "epoch": 4.2284036218288605, "grad_norm": 0.001228676992468536, "learning_rate": 1.3792174685852801e-05, "loss": 0.001, "num_input_tokens_seen": 49897392, "step": 25920 }, { "epoch": 4.229219349049678, "grad_norm": 15.817127227783203, "learning_rate": 1.378339993690632e-05, "loss": 0.0044, "num_input_tokens_seen": 49907600, "step": 25925 }, { "epoch": 4.230035076270495, "grad_norm": 0.1138000339269638, "learning_rate": 1.3774626917700523e-05, "loss": 0.0012, "num_input_tokens_seen": 49918480, "step": 25930 }, { "epoch": 4.230850803491313, "grad_norm": 0.002131738467141986, "learning_rate": 1.3765855629588334e-05, "loss": 0.0582, "num_input_tokens_seen": 49928576, "step": 25935 }, { "epoch": 4.23166653071213, "grad_norm": 0.007961717434227467, "learning_rate": 1.3757086073922374e-05, "loss": 0.0177, "num_input_tokens_seen": 49937888, "step": 25940 }, { "epoch": 4.232482257932947, "grad_norm": 0.35454070568084717, "learning_rate": 1.3748318252055038e-05, "loss": 0.0019, "num_input_tokens_seen": 49946496, "step": 25945 }, { "epoch": 4.233297985153764, "grad_norm": 0.0314023457467556, "learning_rate": 1.3739552165338416e-05, "loss": 0.0005, "num_input_tokens_seen": 49956528, "step": 25950 }, { "epoch": 4.234113712374582, "grad_norm": 0.40387022495269775, "learning_rate": 1.3730787815124354e-05, "loss": 0.0013, "num_input_tokens_seen": 49966080, "step": 25955 }, { "epoch": 4.234929439595399, "grad_norm": 0.1928439438343048, "learning_rate": 1.3722025202764443e-05, "loss": 0.0005, "num_input_tokens_seen": 49975680, "step": 25960 }, { "epoch": 4.2357451668162165, "grad_norm": 0.0045936210080981255, "learning_rate": 1.371326432960997e-05, "loss": 0.0001, "num_input_tokens_seen": 49984560, "step": 25965 }, { "epoch": 4.236560894037034, "grad_norm": 0.0078263645991683, "learning_rate": 1.3704505197011969e-05, "loss": 0.0002, "num_input_tokens_seen": 49992448, "step": 25970 }, { "epoch": 4.237376621257852, "grad_norm": 0.03715648874640465, "learning_rate": 1.3695747806321224e-05, "loss": 0.0006, "num_input_tokens_seen": 50003184, "step": 25975 }, { "epoch": 4.238192348478669, "grad_norm": 0.00435200659558177, "learning_rate": 1.3686992158888212e-05, "loss": 0.0002, "num_input_tokens_seen": 50013104, "step": 25980 }, { "epoch": 4.239008075699486, "grad_norm": 5.730128288269043, "learning_rate": 1.367823825606319e-05, "loss": 0.2253, "num_input_tokens_seen": 50021680, "step": 25985 }, { "epoch": 4.239823802920303, "grad_norm": 0.027173927053809166, "learning_rate": 1.36694860991961e-05, "loss": 0.0002, "num_input_tokens_seen": 50030752, "step": 25990 }, { "epoch": 4.240639530141121, "grad_norm": 0.016446787863969803, "learning_rate": 1.3660735689636636e-05, "loss": 0.0002, "num_input_tokens_seen": 50039664, "step": 25995 }, { "epoch": 4.241455257361938, "grad_norm": 0.0018797438824549317, "learning_rate": 1.365198702873424e-05, "loss": 0.0, "num_input_tokens_seen": 50049312, "step": 26000 }, { "epoch": 4.241455257361938, "eval_loss": 0.33171892166137695, "eval_runtime": 80.8481, "eval_samples_per_second": 33.705, "eval_steps_per_second": 16.859, "num_input_tokens_seen": 50049312, "step": 26000 }, { "epoch": 4.242270984582755, "grad_norm": 0.005326115991920233, "learning_rate": 1.364324011783804e-05, "loss": 0.0001, "num_input_tokens_seen": 50058128, "step": 26005 }, { "epoch": 4.243086711803572, "grad_norm": 0.01111272070556879, "learning_rate": 1.3634494958296934e-05, "loss": 0.0236, "num_input_tokens_seen": 50068784, "step": 26010 }, { "epoch": 4.2439024390243905, "grad_norm": 0.03375871479511261, "learning_rate": 1.3625751551459542e-05, "loss": 0.0002, "num_input_tokens_seen": 50078960, "step": 26015 }, { "epoch": 4.244718166245208, "grad_norm": 0.013628837652504444, "learning_rate": 1.3617009898674188e-05, "loss": 0.0004, "num_input_tokens_seen": 50088928, "step": 26020 }, { "epoch": 4.245533893466025, "grad_norm": 0.004901840351521969, "learning_rate": 1.3608270001288967e-05, "loss": 0.0001, "num_input_tokens_seen": 50098240, "step": 26025 }, { "epoch": 4.246349620686843, "grad_norm": 0.08040036261081696, "learning_rate": 1.359953186065166e-05, "loss": 0.0003, "num_input_tokens_seen": 50107440, "step": 26030 }, { "epoch": 4.24716534790766, "grad_norm": 0.0005498180398717523, "learning_rate": 1.3590795478109814e-05, "loss": 0.0002, "num_input_tokens_seen": 50116448, "step": 26035 }, { "epoch": 4.247981075128477, "grad_norm": 0.012785360217094421, "learning_rate": 1.3582060855010675e-05, "loss": 0.0002, "num_input_tokens_seen": 50126432, "step": 26040 }, { "epoch": 4.248796802349294, "grad_norm": 0.041918836534023285, "learning_rate": 1.3573327992701245e-05, "loss": 0.0001, "num_input_tokens_seen": 50135744, "step": 26045 }, { "epoch": 4.249612529570112, "grad_norm": 18.211400985717773, "learning_rate": 1.356459689252823e-05, "loss": 0.1114, "num_input_tokens_seen": 50146608, "step": 26050 }, { "epoch": 4.250428256790929, "grad_norm": 24.82187843322754, "learning_rate": 1.3555867555838087e-05, "loss": 0.1589, "num_input_tokens_seen": 50156752, "step": 26055 }, { "epoch": 4.251243984011746, "grad_norm": 0.003291226690635085, "learning_rate": 1.3547139983976975e-05, "loss": 0.0001, "num_input_tokens_seen": 50166112, "step": 26060 }, { "epoch": 4.2520597112325635, "grad_norm": 0.017770759761333466, "learning_rate": 1.3538414178290815e-05, "loss": 0.0003, "num_input_tokens_seen": 50176800, "step": 26065 }, { "epoch": 4.252875438453382, "grad_norm": 0.00564453424885869, "learning_rate": 1.3529690140125209e-05, "loss": 0.0002, "num_input_tokens_seen": 50186576, "step": 26070 }, { "epoch": 4.253691165674199, "grad_norm": 0.02476654388010502, "learning_rate": 1.352096787082553e-05, "loss": 0.0006, "num_input_tokens_seen": 50197280, "step": 26075 }, { "epoch": 4.254506892895016, "grad_norm": 0.058327946811914444, "learning_rate": 1.3512247371736871e-05, "loss": 0.0015, "num_input_tokens_seen": 50205840, "step": 26080 }, { "epoch": 4.255322620115833, "grad_norm": 0.006647058296948671, "learning_rate": 1.3503528644204022e-05, "loss": 0.0001, "num_input_tokens_seen": 50216320, "step": 26085 }, { "epoch": 4.256138347336651, "grad_norm": 1.850220799446106, "learning_rate": 1.349481168957153e-05, "loss": 0.0008, "num_input_tokens_seen": 50225584, "step": 26090 }, { "epoch": 4.256954074557468, "grad_norm": 0.09686274826526642, "learning_rate": 1.3486096509183665e-05, "loss": 0.0004, "num_input_tokens_seen": 50233808, "step": 26095 }, { "epoch": 4.257769801778285, "grad_norm": 0.0048599280416965485, "learning_rate": 1.3477383104384406e-05, "loss": 0.0001, "num_input_tokens_seen": 50244672, "step": 26100 }, { "epoch": 4.258585528999102, "grad_norm": 0.01210116595029831, "learning_rate": 1.3468671476517481e-05, "loss": 0.0835, "num_input_tokens_seen": 50254288, "step": 26105 }, { "epoch": 4.25940125621992, "grad_norm": 0.007289690896868706, "learning_rate": 1.3459961626926326e-05, "loss": 0.0002, "num_input_tokens_seen": 50263760, "step": 26110 }, { "epoch": 4.2602169834407375, "grad_norm": 0.002110244007781148, "learning_rate": 1.3451253556954101e-05, "loss": 0.013, "num_input_tokens_seen": 50273120, "step": 26115 }, { "epoch": 4.261032710661555, "grad_norm": 0.0003979688335675746, "learning_rate": 1.3442547267943717e-05, "loss": 0.0004, "num_input_tokens_seen": 50281872, "step": 26120 }, { "epoch": 4.261848437882372, "grad_norm": 0.001792142167687416, "learning_rate": 1.3433842761237774e-05, "loss": 0.0814, "num_input_tokens_seen": 50290464, "step": 26125 }, { "epoch": 4.26266416510319, "grad_norm": 0.004740442615002394, "learning_rate": 1.3425140038178639e-05, "loss": 0.0001, "num_input_tokens_seen": 50299200, "step": 26130 }, { "epoch": 4.263479892324007, "grad_norm": 0.03809071332216263, "learning_rate": 1.3416439100108358e-05, "loss": 0.0001, "num_input_tokens_seen": 50308576, "step": 26135 }, { "epoch": 4.264295619544824, "grad_norm": 0.024625711143016815, "learning_rate": 1.3407739948368734e-05, "loss": 0.0002, "num_input_tokens_seen": 50317968, "step": 26140 }, { "epoch": 4.265111346765641, "grad_norm": 0.014175823889672756, "learning_rate": 1.3399042584301298e-05, "loss": 0.0001, "num_input_tokens_seen": 50325792, "step": 26145 }, { "epoch": 4.265927073986459, "grad_norm": 0.0024534829426556826, "learning_rate": 1.3390347009247272e-05, "loss": 0.0002, "num_input_tokens_seen": 50335984, "step": 26150 }, { "epoch": 4.266742801207276, "grad_norm": 34.10519790649414, "learning_rate": 1.3381653224547635e-05, "loss": 0.1463, "num_input_tokens_seen": 50346048, "step": 26155 }, { "epoch": 4.2675585284280935, "grad_norm": 6.553363800048828, "learning_rate": 1.3372961231543086e-05, "loss": 0.0631, "num_input_tokens_seen": 50354400, "step": 26160 }, { "epoch": 4.268374255648911, "grad_norm": 0.03131799027323723, "learning_rate": 1.3364271031574016e-05, "loss": 0.1624, "num_input_tokens_seen": 50365248, "step": 26165 }, { "epoch": 4.269189982869729, "grad_norm": 0.015517237596213818, "learning_rate": 1.335558262598059e-05, "loss": 0.0002, "num_input_tokens_seen": 50376128, "step": 26170 }, { "epoch": 4.270005710090546, "grad_norm": 0.0029579047113656998, "learning_rate": 1.3346896016102645e-05, "loss": 0.0001, "num_input_tokens_seen": 50385168, "step": 26175 }, { "epoch": 4.270821437311363, "grad_norm": 4.485101222991943, "learning_rate": 1.3338211203279788e-05, "loss": 0.128, "num_input_tokens_seen": 50394160, "step": 26180 }, { "epoch": 4.27163716453218, "grad_norm": 0.0018814982613548636, "learning_rate": 1.3329528188851303e-05, "loss": 0.0003, "num_input_tokens_seen": 50404288, "step": 26185 }, { "epoch": 4.272452891752998, "grad_norm": 9.847282409667969, "learning_rate": 1.3320846974156242e-05, "loss": 0.1106, "num_input_tokens_seen": 50414160, "step": 26190 }, { "epoch": 4.273268618973815, "grad_norm": 0.14477947354316711, "learning_rate": 1.3312167560533337e-05, "loss": 0.0003, "num_input_tokens_seen": 50423760, "step": 26195 }, { "epoch": 4.274084346194632, "grad_norm": 0.02611338533461094, "learning_rate": 1.3303489949321082e-05, "loss": 0.0008, "num_input_tokens_seen": 50433008, "step": 26200 }, { "epoch": 4.274084346194632, "eval_loss": 0.2933138310909271, "eval_runtime": 80.8531, "eval_samples_per_second": 33.703, "eval_steps_per_second": 16.858, "num_input_tokens_seen": 50433008, "step": 26200 }, { "epoch": 4.27490007341545, "grad_norm": 0.07208452373743057, "learning_rate": 1.3294814141857653e-05, "loss": 0.0861, "num_input_tokens_seen": 50442320, "step": 26205 }, { "epoch": 4.275715800636267, "grad_norm": 0.20082610845565796, "learning_rate": 1.3286140139480992e-05, "loss": 0.0009, "num_input_tokens_seen": 50451120, "step": 26210 }, { "epoch": 4.276531527857085, "grad_norm": 0.023116081953048706, "learning_rate": 1.3277467943528719e-05, "loss": 0.0942, "num_input_tokens_seen": 50461024, "step": 26215 }, { "epoch": 4.277347255077902, "grad_norm": 0.7903038859367371, "learning_rate": 1.3268797555338203e-05, "loss": 0.117, "num_input_tokens_seen": 50470672, "step": 26220 }, { "epoch": 4.27816298229872, "grad_norm": 0.04493061453104019, "learning_rate": 1.3260128976246533e-05, "loss": 0.0007, "num_input_tokens_seen": 50480304, "step": 26225 }, { "epoch": 4.278978709519537, "grad_norm": 0.1617674082517624, "learning_rate": 1.32514622075905e-05, "loss": 0.002, "num_input_tokens_seen": 50489328, "step": 26230 }, { "epoch": 4.279794436740354, "grad_norm": 0.01991272158920765, "learning_rate": 1.3242797250706638e-05, "loss": 0.0346, "num_input_tokens_seen": 50498944, "step": 26235 }, { "epoch": 4.280610163961171, "grad_norm": 0.0017542754067108035, "learning_rate": 1.3234134106931195e-05, "loss": 0.0083, "num_input_tokens_seen": 50509168, "step": 26240 }, { "epoch": 4.281425891181989, "grad_norm": 0.012650713324546814, "learning_rate": 1.322547277760013e-05, "loss": 0.0002, "num_input_tokens_seen": 50519248, "step": 26245 }, { "epoch": 4.282241618402806, "grad_norm": 0.015573319979012012, "learning_rate": 1.3216813264049132e-05, "loss": 0.0001, "num_input_tokens_seen": 50528816, "step": 26250 }, { "epoch": 4.283057345623623, "grad_norm": 0.012198296375572681, "learning_rate": 1.32081555676136e-05, "loss": 0.0008, "num_input_tokens_seen": 50538528, "step": 26255 }, { "epoch": 4.2838730728444405, "grad_norm": 0.2845352292060852, "learning_rate": 1.3199499689628674e-05, "loss": 0.0011, "num_input_tokens_seen": 50548608, "step": 26260 }, { "epoch": 4.2846888000652585, "grad_norm": 0.002538378583267331, "learning_rate": 1.3190845631429192e-05, "loss": 0.0004, "num_input_tokens_seen": 50558128, "step": 26265 }, { "epoch": 4.285504527286076, "grad_norm": 0.2780515253543854, "learning_rate": 1.3182193394349704e-05, "loss": 0.0026, "num_input_tokens_seen": 50568304, "step": 26270 }, { "epoch": 4.286320254506893, "grad_norm": 0.8973797559738159, "learning_rate": 1.3173542979724507e-05, "loss": 0.0016, "num_input_tokens_seen": 50577232, "step": 26275 }, { "epoch": 4.28713598172771, "grad_norm": 0.042790528386831284, "learning_rate": 1.3164894388887617e-05, "loss": 0.0009, "num_input_tokens_seen": 50586448, "step": 26280 }, { "epoch": 4.287951708948528, "grad_norm": 0.07429895550012589, "learning_rate": 1.3156247623172727e-05, "loss": 0.0499, "num_input_tokens_seen": 50596640, "step": 26285 }, { "epoch": 4.288767436169345, "grad_norm": 0.061249181628227234, "learning_rate": 1.3147602683913302e-05, "loss": 0.0005, "num_input_tokens_seen": 50606784, "step": 26290 }, { "epoch": 4.289583163390162, "grad_norm": 9.751968383789062, "learning_rate": 1.3138959572442481e-05, "loss": 0.1147, "num_input_tokens_seen": 50615696, "step": 26295 }, { "epoch": 4.290398890610979, "grad_norm": 0.04981352388858795, "learning_rate": 1.3130318290093146e-05, "loss": 0.0025, "num_input_tokens_seen": 50624176, "step": 26300 }, { "epoch": 4.291214617831797, "grad_norm": 0.08974776417016983, "learning_rate": 1.3121678838197909e-05, "loss": 0.0006, "num_input_tokens_seen": 50634432, "step": 26305 }, { "epoch": 4.2920303450526145, "grad_norm": 0.004823956172913313, "learning_rate": 1.3113041218089056e-05, "loss": 0.0002, "num_input_tokens_seen": 50644256, "step": 26310 }, { "epoch": 4.292846072273432, "grad_norm": 0.07111804932355881, "learning_rate": 1.3104405431098626e-05, "loss": 0.0006, "num_input_tokens_seen": 50653760, "step": 26315 }, { "epoch": 4.293661799494249, "grad_norm": 0.09377999603748322, "learning_rate": 1.3095771478558377e-05, "loss": 0.0009, "num_input_tokens_seen": 50663264, "step": 26320 }, { "epoch": 4.294477526715067, "grad_norm": 0.03317088261246681, "learning_rate": 1.3087139361799766e-05, "loss": 0.0005, "num_input_tokens_seen": 50672704, "step": 26325 }, { "epoch": 4.295293253935884, "grad_norm": 0.005880612414330244, "learning_rate": 1.3078509082153964e-05, "loss": 0.0004, "num_input_tokens_seen": 50682160, "step": 26330 }, { "epoch": 4.296108981156701, "grad_norm": 0.0016174317570403218, "learning_rate": 1.3069880640951885e-05, "loss": 0.1173, "num_input_tokens_seen": 50690960, "step": 26335 }, { "epoch": 4.296924708377518, "grad_norm": 0.0017025875858962536, "learning_rate": 1.3061254039524123e-05, "loss": 0.0755, "num_input_tokens_seen": 50700496, "step": 26340 }, { "epoch": 4.297740435598336, "grad_norm": 0.01713685877621174, "learning_rate": 1.3052629279201028e-05, "loss": 0.1825, "num_input_tokens_seen": 50710480, "step": 26345 }, { "epoch": 4.298556162819153, "grad_norm": 0.013798116706311703, "learning_rate": 1.3044006361312633e-05, "loss": 0.0007, "num_input_tokens_seen": 50720528, "step": 26350 }, { "epoch": 4.2993718900399704, "grad_norm": 0.008771366439759731, "learning_rate": 1.30353852871887e-05, "loss": 0.0661, "num_input_tokens_seen": 50728880, "step": 26355 }, { "epoch": 4.300187617260788, "grad_norm": 0.020943783223628998, "learning_rate": 1.302676605815873e-05, "loss": 0.0015, "num_input_tokens_seen": 50738960, "step": 26360 }, { "epoch": 4.301003344481606, "grad_norm": 0.020093023777008057, "learning_rate": 1.3018148675551884e-05, "loss": 0.0867, "num_input_tokens_seen": 50748496, "step": 26365 }, { "epoch": 4.301819071702423, "grad_norm": 10.702857971191406, "learning_rate": 1.3009533140697094e-05, "loss": 0.0836, "num_input_tokens_seen": 50757904, "step": 26370 }, { "epoch": 4.30263479892324, "grad_norm": 0.023129086941480637, "learning_rate": 1.3000919454922966e-05, "loss": 0.2654, "num_input_tokens_seen": 50767328, "step": 26375 }, { "epoch": 4.303450526144058, "grad_norm": 0.023271890357136726, "learning_rate": 1.299230761955785e-05, "loss": 0.013, "num_input_tokens_seen": 50778288, "step": 26380 }, { "epoch": 4.304266253364875, "grad_norm": 0.11564366519451141, "learning_rate": 1.2983697635929807e-05, "loss": 0.0054, "num_input_tokens_seen": 50787888, "step": 26385 }, { "epoch": 4.305081980585692, "grad_norm": 36.21962356567383, "learning_rate": 1.2975089505366584e-05, "loss": 0.1406, "num_input_tokens_seen": 50796336, "step": 26390 }, { "epoch": 4.305897707806509, "grad_norm": 0.2771676480770111, "learning_rate": 1.2966483229195683e-05, "loss": 0.002, "num_input_tokens_seen": 50805968, "step": 26395 }, { "epoch": 4.306713435027326, "grad_norm": 0.04493905231356621, "learning_rate": 1.2957878808744283e-05, "loss": 0.0881, "num_input_tokens_seen": 50815824, "step": 26400 }, { "epoch": 4.306713435027326, "eval_loss": 0.2597741484642029, "eval_runtime": 80.7478, "eval_samples_per_second": 33.747, "eval_steps_per_second": 16.88, "num_input_tokens_seen": 50815824, "step": 26400 }, { "epoch": 4.307529162248144, "grad_norm": 0.020193159580230713, "learning_rate": 1.294927624533931e-05, "loss": 0.0588, "num_input_tokens_seen": 50825488, "step": 26405 }, { "epoch": 4.3083448894689615, "grad_norm": 0.03382895141839981, "learning_rate": 1.2940675540307378e-05, "loss": 0.0009, "num_input_tokens_seen": 50835568, "step": 26410 }, { "epoch": 4.309160616689779, "grad_norm": 0.05363347753882408, "learning_rate": 1.2932076694974814e-05, "loss": 0.0739, "num_input_tokens_seen": 50844688, "step": 26415 }, { "epoch": 4.309976343910597, "grad_norm": 0.1053803414106369, "learning_rate": 1.2923479710667682e-05, "loss": 0.0011, "num_input_tokens_seen": 50853760, "step": 26420 }, { "epoch": 4.310792071131414, "grad_norm": 0.024718323722481728, "learning_rate": 1.2914884588711751e-05, "loss": 0.001, "num_input_tokens_seen": 50862960, "step": 26425 }, { "epoch": 4.311607798352231, "grad_norm": 5.792600154876709, "learning_rate": 1.2906291330432475e-05, "loss": 0.0798, "num_input_tokens_seen": 50872288, "step": 26430 }, { "epoch": 4.312423525573048, "grad_norm": 0.04693234711885452, "learning_rate": 1.2897699937155055e-05, "loss": 0.052, "num_input_tokens_seen": 50882560, "step": 26435 }, { "epoch": 4.313239252793866, "grad_norm": 4.753458499908447, "learning_rate": 1.2889110410204403e-05, "loss": 0.0138, "num_input_tokens_seen": 50891968, "step": 26440 }, { "epoch": 4.314054980014683, "grad_norm": 0.009324286133050919, "learning_rate": 1.2880522750905111e-05, "loss": 0.0699, "num_input_tokens_seen": 50902160, "step": 26445 }, { "epoch": 4.3148707072355, "grad_norm": 3.6827433109283447, "learning_rate": 1.2871936960581523e-05, "loss": 0.0384, "num_input_tokens_seen": 50911872, "step": 26450 }, { "epoch": 4.3156864344563175, "grad_norm": 0.03175116702914238, "learning_rate": 1.2863353040557658e-05, "loss": 0.0702, "num_input_tokens_seen": 50920832, "step": 26455 }, { "epoch": 4.3165021616771355, "grad_norm": 1.8437674045562744, "learning_rate": 1.2854770992157273e-05, "loss": 0.0021, "num_input_tokens_seen": 50930592, "step": 26460 }, { "epoch": 4.317317888897953, "grad_norm": 0.13195478916168213, "learning_rate": 1.2846190816703835e-05, "loss": 0.0204, "num_input_tokens_seen": 50940272, "step": 26465 }, { "epoch": 4.31813361611877, "grad_norm": 0.006395094562321901, "learning_rate": 1.2837612515520498e-05, "loss": 0.0005, "num_input_tokens_seen": 50950016, "step": 26470 }, { "epoch": 4.318949343339587, "grad_norm": 0.02099006064236164, "learning_rate": 1.2829036089930163e-05, "loss": 0.0012, "num_input_tokens_seen": 50960336, "step": 26475 }, { "epoch": 4.319765070560405, "grad_norm": 0.15499219298362732, "learning_rate": 1.2820461541255412e-05, "loss": 0.0011, "num_input_tokens_seen": 50969680, "step": 26480 }, { "epoch": 4.320580797781222, "grad_norm": 0.020830053836107254, "learning_rate": 1.2811888870818543e-05, "loss": 0.0421, "num_input_tokens_seen": 50978800, "step": 26485 }, { "epoch": 4.321396525002039, "grad_norm": 0.044102706015110016, "learning_rate": 1.2803318079941581e-05, "loss": 0.0027, "num_input_tokens_seen": 50988640, "step": 26490 }, { "epoch": 4.322212252222856, "grad_norm": 0.10092943906784058, "learning_rate": 1.2794749169946235e-05, "loss": 0.0022, "num_input_tokens_seen": 50998720, "step": 26495 }, { "epoch": 4.323027979443674, "grad_norm": 0.005928037688136101, "learning_rate": 1.2786182142153952e-05, "loss": 0.073, "num_input_tokens_seen": 51009456, "step": 26500 }, { "epoch": 4.3238437066644915, "grad_norm": 0.02094883657991886, "learning_rate": 1.2777616997885878e-05, "loss": 0.0022, "num_input_tokens_seen": 51019328, "step": 26505 }, { "epoch": 4.324659433885309, "grad_norm": 0.018969262018799782, "learning_rate": 1.2769053738462847e-05, "loss": 0.0009, "num_input_tokens_seen": 51028784, "step": 26510 }, { "epoch": 4.325475161106126, "grad_norm": 0.16617494821548462, "learning_rate": 1.2760492365205434e-05, "loss": 0.002, "num_input_tokens_seen": 51039696, "step": 26515 }, { "epoch": 4.326290888326944, "grad_norm": 0.012671218253672123, "learning_rate": 1.2751932879433919e-05, "loss": 0.0002, "num_input_tokens_seen": 51048496, "step": 26520 }, { "epoch": 4.327106615547761, "grad_norm": 3.1654796600341797, "learning_rate": 1.2743375282468267e-05, "loss": 0.0094, "num_input_tokens_seen": 51057648, "step": 26525 }, { "epoch": 4.327922342768578, "grad_norm": 0.21880324184894562, "learning_rate": 1.2734819575628182e-05, "loss": 0.0006, "num_input_tokens_seen": 51066304, "step": 26530 }, { "epoch": 4.328738069989395, "grad_norm": 0.034358780831098557, "learning_rate": 1.2726265760233039e-05, "loss": 0.0134, "num_input_tokens_seen": 51075360, "step": 26535 }, { "epoch": 4.329553797210213, "grad_norm": 0.054923079907894135, "learning_rate": 1.271771383760197e-05, "loss": 0.0785, "num_input_tokens_seen": 51085296, "step": 26540 }, { "epoch": 4.33036952443103, "grad_norm": 0.011564402841031551, "learning_rate": 1.2709163809053764e-05, "loss": 0.0006, "num_input_tokens_seen": 51096064, "step": 26545 }, { "epoch": 4.331185251651847, "grad_norm": 0.028378145769238472, "learning_rate": 1.2700615675906963e-05, "loss": 0.0004, "num_input_tokens_seen": 51105392, "step": 26550 }, { "epoch": 4.332000978872665, "grad_norm": 0.014359741471707821, "learning_rate": 1.269206943947978e-05, "loss": 0.0002, "num_input_tokens_seen": 51115136, "step": 26555 }, { "epoch": 4.332816706093483, "grad_norm": 0.03838399797677994, "learning_rate": 1.2683525101090177e-05, "loss": 0.0233, "num_input_tokens_seen": 51123984, "step": 26560 }, { "epoch": 4.3336324333143, "grad_norm": 0.08003851026296616, "learning_rate": 1.2674982662055765e-05, "loss": 0.0021, "num_input_tokens_seen": 51134208, "step": 26565 }, { "epoch": 4.334448160535117, "grad_norm": 0.021664729341864586, "learning_rate": 1.2666442123693922e-05, "loss": 0.0003, "num_input_tokens_seen": 51143504, "step": 26570 }, { "epoch": 4.335263887755934, "grad_norm": 0.04180941730737686, "learning_rate": 1.265790348732169e-05, "loss": 0.0559, "num_input_tokens_seen": 51152928, "step": 26575 }, { "epoch": 4.336079614976752, "grad_norm": 0.004951280541718006, "learning_rate": 1.264936675425584e-05, "loss": 0.0286, "num_input_tokens_seen": 51162512, "step": 26580 }, { "epoch": 4.336895342197569, "grad_norm": 0.06744127720594406, "learning_rate": 1.2640831925812852e-05, "loss": 0.0016, "num_input_tokens_seen": 51172704, "step": 26585 }, { "epoch": 4.337711069418386, "grad_norm": 0.03823000565171242, "learning_rate": 1.263229900330889e-05, "loss": 0.0003, "num_input_tokens_seen": 51181488, "step": 26590 }, { "epoch": 4.338526796639204, "grad_norm": 0.7452120780944824, "learning_rate": 1.2623767988059843e-05, "loss": 0.0315, "num_input_tokens_seen": 51191648, "step": 26595 }, { "epoch": 4.339342523860021, "grad_norm": 0.048186302185058594, "learning_rate": 1.2615238881381309e-05, "loss": 0.0003, "num_input_tokens_seen": 51200224, "step": 26600 }, { "epoch": 4.339342523860021, "eval_loss": 0.3100602626800537, "eval_runtime": 80.7373, "eval_samples_per_second": 33.751, "eval_steps_per_second": 16.882, "num_input_tokens_seen": 51200224, "step": 26600 }, { "epoch": 4.3401582510808385, "grad_norm": 0.0020642331801354885, "learning_rate": 1.2606711684588568e-05, "loss": 0.0004, "num_input_tokens_seen": 51210992, "step": 26605 }, { "epoch": 4.340973978301656, "grad_norm": 4.522244930267334, "learning_rate": 1.2598186398996636e-05, "loss": 0.0098, "num_input_tokens_seen": 51220400, "step": 26610 }, { "epoch": 4.341789705522474, "grad_norm": 4.09148645401001, "learning_rate": 1.2589663025920207e-05, "loss": 0.0557, "num_input_tokens_seen": 51229600, "step": 26615 }, { "epoch": 4.342605432743291, "grad_norm": 0.0334138460457325, "learning_rate": 1.2581141566673705e-05, "loss": 0.0227, "num_input_tokens_seen": 51239584, "step": 26620 }, { "epoch": 4.343421159964108, "grad_norm": 0.03333883360028267, "learning_rate": 1.257262202257124e-05, "loss": 0.0024, "num_input_tokens_seen": 51250352, "step": 26625 }, { "epoch": 4.344236887184925, "grad_norm": 0.19517117738723755, "learning_rate": 1.2564104394926618e-05, "loss": 0.0211, "num_input_tokens_seen": 51261056, "step": 26630 }, { "epoch": 4.345052614405743, "grad_norm": 0.0008747159736230969, "learning_rate": 1.2555588685053383e-05, "loss": 0.0002, "num_input_tokens_seen": 51271472, "step": 26635 }, { "epoch": 4.34586834162656, "grad_norm": 0.012038808315992355, "learning_rate": 1.2547074894264762e-05, "loss": 0.0111, "num_input_tokens_seen": 51281168, "step": 26640 }, { "epoch": 4.346684068847377, "grad_norm": 0.06703027337789536, "learning_rate": 1.2538563023873679e-05, "loss": 0.0005, "num_input_tokens_seen": 51290512, "step": 26645 }, { "epoch": 4.3474997960681945, "grad_norm": 0.03354346752166748, "learning_rate": 1.2530053075192789e-05, "loss": 0.0002, "num_input_tokens_seen": 51300096, "step": 26650 }, { "epoch": 4.3483155232890125, "grad_norm": 0.00613150792196393, "learning_rate": 1.252154504953441e-05, "loss": 0.0002, "num_input_tokens_seen": 51310592, "step": 26655 }, { "epoch": 4.34913125050983, "grad_norm": 0.002316959435120225, "learning_rate": 1.25130389482106e-05, "loss": 0.0001, "num_input_tokens_seen": 51320112, "step": 26660 }, { "epoch": 4.349946977730647, "grad_norm": 0.004145472776144743, "learning_rate": 1.2504534772533116e-05, "loss": 0.1606, "num_input_tokens_seen": 51329680, "step": 26665 }, { "epoch": 4.350762704951464, "grad_norm": 0.01335993967950344, "learning_rate": 1.2496032523813387e-05, "loss": 0.073, "num_input_tokens_seen": 51339472, "step": 26670 }, { "epoch": 4.351578432172282, "grad_norm": 0.09671194106340408, "learning_rate": 1.2487532203362576e-05, "loss": 0.0592, "num_input_tokens_seen": 51348784, "step": 26675 }, { "epoch": 4.352394159393099, "grad_norm": 0.18848970532417297, "learning_rate": 1.247903381249155e-05, "loss": 0.0964, "num_input_tokens_seen": 51359104, "step": 26680 }, { "epoch": 4.353209886613916, "grad_norm": 13.316473007202148, "learning_rate": 1.2470537352510853e-05, "loss": 0.153, "num_input_tokens_seen": 51369360, "step": 26685 }, { "epoch": 4.354025613834733, "grad_norm": 0.03320087119936943, "learning_rate": 1.2462042824730758e-05, "loss": 0.0003, "num_input_tokens_seen": 51379184, "step": 26690 }, { "epoch": 4.354841341055551, "grad_norm": 14.01892375946045, "learning_rate": 1.245355023046122e-05, "loss": 0.0183, "num_input_tokens_seen": 51388240, "step": 26695 }, { "epoch": 4.3556570682763684, "grad_norm": 3.771258592605591, "learning_rate": 1.2445059571011896e-05, "loss": 0.0033, "num_input_tokens_seen": 51398192, "step": 26700 }, { "epoch": 4.356472795497186, "grad_norm": 0.06576807796955109, "learning_rate": 1.2436570847692173e-05, "loss": 0.0105, "num_input_tokens_seen": 51408064, "step": 26705 }, { "epoch": 4.357288522718003, "grad_norm": 0.02893972396850586, "learning_rate": 1.2428084061811096e-05, "loss": 0.0656, "num_input_tokens_seen": 51417088, "step": 26710 }, { "epoch": 4.358104249938821, "grad_norm": 0.10869091749191284, "learning_rate": 1.2419599214677447e-05, "loss": 0.0402, "num_input_tokens_seen": 51426624, "step": 26715 }, { "epoch": 4.358919977159638, "grad_norm": 4.7330803871154785, "learning_rate": 1.2411116307599702e-05, "loss": 0.0652, "num_input_tokens_seen": 51436416, "step": 26720 }, { "epoch": 4.359735704380455, "grad_norm": 0.19756150245666504, "learning_rate": 1.2402635341886016e-05, "loss": 0.0007, "num_input_tokens_seen": 51444832, "step": 26725 }, { "epoch": 4.360551431601272, "grad_norm": 0.14302070438861847, "learning_rate": 1.2394156318844278e-05, "loss": 0.0798, "num_input_tokens_seen": 51454624, "step": 26730 }, { "epoch": 4.36136715882209, "grad_norm": 0.022031238302588463, "learning_rate": 1.2385679239782039e-05, "loss": 0.0008, "num_input_tokens_seen": 51463520, "step": 26735 }, { "epoch": 4.362182886042907, "grad_norm": 0.07698238641023636, "learning_rate": 1.2377204106006585e-05, "loss": 0.0857, "num_input_tokens_seen": 51473760, "step": 26740 }, { "epoch": 4.362998613263724, "grad_norm": 0.01059255562722683, "learning_rate": 1.2368730918824891e-05, "loss": 0.0012, "num_input_tokens_seen": 51483760, "step": 26745 }, { "epoch": 4.3638143404845415, "grad_norm": 0.028460761532187462, "learning_rate": 1.236025967954362e-05, "loss": 0.0003, "num_input_tokens_seen": 51492240, "step": 26750 }, { "epoch": 4.3646300677053596, "grad_norm": 0.3287648558616638, "learning_rate": 1.2351790389469153e-05, "loss": 0.0009, "num_input_tokens_seen": 51502336, "step": 26755 }, { "epoch": 4.365445794926177, "grad_norm": 0.027642013505101204, "learning_rate": 1.234332304990755e-05, "loss": 0.0002, "num_input_tokens_seen": 51510960, "step": 26760 }, { "epoch": 4.366261522146994, "grad_norm": 0.006307921838015318, "learning_rate": 1.2334857662164593e-05, "loss": 0.001, "num_input_tokens_seen": 51520352, "step": 26765 }, { "epoch": 4.367077249367812, "grad_norm": 0.007151339203119278, "learning_rate": 1.2326394227545743e-05, "loss": 0.0067, "num_input_tokens_seen": 51530432, "step": 26770 }, { "epoch": 4.367892976588629, "grad_norm": 0.08536624163389206, "learning_rate": 1.2317932747356162e-05, "loss": 0.0907, "num_input_tokens_seen": 51539456, "step": 26775 }, { "epoch": 4.368708703809446, "grad_norm": 1.287213683128357, "learning_rate": 1.2309473222900726e-05, "loss": 0.0007, "num_input_tokens_seen": 51550240, "step": 26780 }, { "epoch": 4.369524431030263, "grad_norm": 0.0034736585803329945, "learning_rate": 1.2301015655484006e-05, "loss": 0.0882, "num_input_tokens_seen": 51559920, "step": 26785 }, { "epoch": 4.370340158251081, "grad_norm": 0.0012373024364933372, "learning_rate": 1.2292560046410245e-05, "loss": 0.0358, "num_input_tokens_seen": 51568432, "step": 26790 }, { "epoch": 4.371155885471898, "grad_norm": 0.007708060089498758, "learning_rate": 1.228410639698343e-05, "loss": 0.0011, "num_input_tokens_seen": 51577184, "step": 26795 }, { "epoch": 4.3719716126927155, "grad_norm": 0.023401573300361633, "learning_rate": 1.2275654708507195e-05, "loss": 0.0034, "num_input_tokens_seen": 51585680, "step": 26800 }, { "epoch": 4.3719716126927155, "eval_loss": 0.2956853210926056, "eval_runtime": 80.6965, "eval_samples_per_second": 33.769, "eval_steps_per_second": 16.89, "num_input_tokens_seen": 51585680, "step": 26800 }, { "epoch": 4.372787339913533, "grad_norm": 0.1410549432039261, "learning_rate": 1.2267204982284908e-05, "loss": 0.0243, "num_input_tokens_seen": 51595104, "step": 26805 }, { "epoch": 4.373603067134351, "grad_norm": 0.022220106795430183, "learning_rate": 1.2258757219619635e-05, "loss": 0.0489, "num_input_tokens_seen": 51605376, "step": 26810 }, { "epoch": 4.374418794355168, "grad_norm": 25.683921813964844, "learning_rate": 1.2250311421814104e-05, "loss": 0.0377, "num_input_tokens_seen": 51614352, "step": 26815 }, { "epoch": 4.375234521575985, "grad_norm": 0.0010744687169790268, "learning_rate": 1.2241867590170772e-05, "loss": 0.0006, "num_input_tokens_seen": 51624832, "step": 26820 }, { "epoch": 4.376050248796802, "grad_norm": 0.02825832925736904, "learning_rate": 1.2233425725991799e-05, "loss": 0.0008, "num_input_tokens_seen": 51633616, "step": 26825 }, { "epoch": 4.37686597601762, "grad_norm": 0.0012699214275926352, "learning_rate": 1.2224985830579003e-05, "loss": 0.115, "num_input_tokens_seen": 51643120, "step": 26830 }, { "epoch": 4.377681703238437, "grad_norm": 0.10168243944644928, "learning_rate": 1.2216547905233944e-05, "loss": 0.0006, "num_input_tokens_seen": 51653376, "step": 26835 }, { "epoch": 4.378497430459254, "grad_norm": 42.91822052001953, "learning_rate": 1.2208111951257842e-05, "loss": 0.073, "num_input_tokens_seen": 51662848, "step": 26840 }, { "epoch": 4.3793131576800715, "grad_norm": 12.029683113098145, "learning_rate": 1.2199677969951622e-05, "loss": 0.0356, "num_input_tokens_seen": 51673312, "step": 26845 }, { "epoch": 4.3801288849008895, "grad_norm": 0.08704328536987305, "learning_rate": 1.2191245962615927e-05, "loss": 0.0343, "num_input_tokens_seen": 51683328, "step": 26850 }, { "epoch": 4.380944612121707, "grad_norm": 0.07342653721570969, "learning_rate": 1.218281593055106e-05, "loss": 0.001, "num_input_tokens_seen": 51692432, "step": 26855 }, { "epoch": 4.381760339342524, "grad_norm": 5.496770858764648, "learning_rate": 1.217438787505705e-05, "loss": 0.0573, "num_input_tokens_seen": 51701760, "step": 26860 }, { "epoch": 4.382576066563341, "grad_norm": 0.07875459641218185, "learning_rate": 1.2165961797433615e-05, "loss": 0.0485, "num_input_tokens_seen": 51712000, "step": 26865 }, { "epoch": 4.383391793784159, "grad_norm": 0.04703235998749733, "learning_rate": 1.215753769898014e-05, "loss": 0.0002, "num_input_tokens_seen": 51721536, "step": 26870 }, { "epoch": 4.384207521004976, "grad_norm": 0.005180924665182829, "learning_rate": 1.2149115580995755e-05, "loss": 0.0007, "num_input_tokens_seen": 51730224, "step": 26875 }, { "epoch": 4.385023248225793, "grad_norm": 0.004441073630005121, "learning_rate": 1.2140695444779227e-05, "loss": 0.0003, "num_input_tokens_seen": 51739376, "step": 26880 }, { "epoch": 4.38583897544661, "grad_norm": 0.04048759490251541, "learning_rate": 1.2132277291629066e-05, "loss": 0.0023, "num_input_tokens_seen": 51749728, "step": 26885 }, { "epoch": 4.386654702667428, "grad_norm": 0.002435484901070595, "learning_rate": 1.2123861122843458e-05, "loss": 0.0013, "num_input_tokens_seen": 51759456, "step": 26890 }, { "epoch": 4.387470429888245, "grad_norm": 0.7838939428329468, "learning_rate": 1.2115446939720271e-05, "loss": 0.0005, "num_input_tokens_seen": 51769984, "step": 26895 }, { "epoch": 4.388286157109063, "grad_norm": 0.03935825452208519, "learning_rate": 1.210703474355708e-05, "loss": 0.0039, "num_input_tokens_seen": 51779344, "step": 26900 }, { "epoch": 4.38910188432988, "grad_norm": 0.004889271222054958, "learning_rate": 1.2098624535651164e-05, "loss": 0.0126, "num_input_tokens_seen": 51789792, "step": 26905 }, { "epoch": 4.389917611550698, "grad_norm": 0.028974387794733047, "learning_rate": 1.2090216317299477e-05, "loss": 0.0005, "num_input_tokens_seen": 51799136, "step": 26910 }, { "epoch": 4.390733338771515, "grad_norm": 0.006588068790733814, "learning_rate": 1.2081810089798668e-05, "loss": 0.0003, "num_input_tokens_seen": 51809168, "step": 26915 }, { "epoch": 4.391549065992332, "grad_norm": 0.012297599576413631, "learning_rate": 1.2073405854445072e-05, "loss": 0.0123, "num_input_tokens_seen": 51818160, "step": 26920 }, { "epoch": 4.392364793213149, "grad_norm": 0.32700884342193604, "learning_rate": 1.206500361253474e-05, "loss": 0.0008, "num_input_tokens_seen": 51827280, "step": 26925 }, { "epoch": 4.393180520433967, "grad_norm": 0.028834694996476173, "learning_rate": 1.2056603365363409e-05, "loss": 0.0002, "num_input_tokens_seen": 51836048, "step": 26930 }, { "epoch": 4.393996247654784, "grad_norm": 0.070265032351017, "learning_rate": 1.2048205114226487e-05, "loss": 0.0798, "num_input_tokens_seen": 51845680, "step": 26935 }, { "epoch": 4.394811974875601, "grad_norm": 0.6492206454277039, "learning_rate": 1.2039808860419102e-05, "loss": 0.0476, "num_input_tokens_seen": 51855552, "step": 26940 }, { "epoch": 4.395627702096419, "grad_norm": 0.007049871608614922, "learning_rate": 1.2031414605236066e-05, "loss": 0.0002, "num_input_tokens_seen": 51865856, "step": 26945 }, { "epoch": 4.3964434293172365, "grad_norm": 0.0112222945317626, "learning_rate": 1.2023022349971862e-05, "loss": 0.061, "num_input_tokens_seen": 51874112, "step": 26950 }, { "epoch": 4.397259156538054, "grad_norm": 0.003557237097993493, "learning_rate": 1.20146320959207e-05, "loss": 0.0006, "num_input_tokens_seen": 51883296, "step": 26955 }, { "epoch": 4.398074883758871, "grad_norm": 0.05301665514707565, "learning_rate": 1.2006243844376445e-05, "loss": 0.0005, "num_input_tokens_seen": 51893776, "step": 26960 }, { "epoch": 4.398890610979688, "grad_norm": 0.32285434007644653, "learning_rate": 1.1997857596632678e-05, "loss": 0.0008, "num_input_tokens_seen": 51903776, "step": 26965 }, { "epoch": 4.399706338200506, "grad_norm": 3.9339439868927, "learning_rate": 1.1989473353982672e-05, "loss": 0.0042, "num_input_tokens_seen": 51911888, "step": 26970 }, { "epoch": 4.400522065421323, "grad_norm": 0.45930227637290955, "learning_rate": 1.198109111771937e-05, "loss": 0.0603, "num_input_tokens_seen": 51921280, "step": 26975 }, { "epoch": 4.40133779264214, "grad_norm": 0.020496224984526634, "learning_rate": 1.197271088913543e-05, "loss": 0.0004, "num_input_tokens_seen": 51930896, "step": 26980 }, { "epoch": 4.402153519862958, "grad_norm": 0.05934469774365425, "learning_rate": 1.1964332669523182e-05, "loss": 0.0617, "num_input_tokens_seen": 51940560, "step": 26985 }, { "epoch": 4.402969247083775, "grad_norm": 0.008876089006662369, "learning_rate": 1.1955956460174645e-05, "loss": 0.065, "num_input_tokens_seen": 51950016, "step": 26990 }, { "epoch": 4.4037849743045925, "grad_norm": 0.004935589153319597, "learning_rate": 1.1947582262381552e-05, "loss": 0.1264, "num_input_tokens_seen": 51959568, "step": 26995 }, { "epoch": 4.40460070152541, "grad_norm": 2.803917646408081, "learning_rate": 1.1939210077435293e-05, "loss": 0.0809, "num_input_tokens_seen": 51969184, "step": 27000 }, { "epoch": 4.40460070152541, "eval_loss": 0.2965461015701294, "eval_runtime": 80.8144, "eval_samples_per_second": 33.719, "eval_steps_per_second": 16.866, "num_input_tokens_seen": 51969184, "step": 27000 }, { "epoch": 4.405416428746228, "grad_norm": 0.08864413946866989, "learning_rate": 1.193083990662697e-05, "loss": 0.0007, "num_input_tokens_seen": 51978944, "step": 27005 }, { "epoch": 4.406232155967045, "grad_norm": 0.00794893503189087, "learning_rate": 1.192247175124738e-05, "loss": 0.0004, "num_input_tokens_seen": 51989328, "step": 27010 }, { "epoch": 4.407047883187862, "grad_norm": 0.04537788778543472, "learning_rate": 1.191410561258698e-05, "loss": 0.0037, "num_input_tokens_seen": 52000128, "step": 27015 }, { "epoch": 4.407863610408679, "grad_norm": 0.05901098996400833, "learning_rate": 1.1905741491935944e-05, "loss": 0.0003, "num_input_tokens_seen": 52010432, "step": 27020 }, { "epoch": 4.408679337629497, "grad_norm": 0.06703463196754456, "learning_rate": 1.1897379390584129e-05, "loss": 0.0755, "num_input_tokens_seen": 52019696, "step": 27025 }, { "epoch": 4.409495064850314, "grad_norm": 55.30966567993164, "learning_rate": 1.1889019309821062e-05, "loss": 0.1275, "num_input_tokens_seen": 52028384, "step": 27030 }, { "epoch": 4.410310792071131, "grad_norm": 17.22942352294922, "learning_rate": 1.188066125093599e-05, "loss": 0.0342, "num_input_tokens_seen": 52038720, "step": 27035 }, { "epoch": 4.411126519291948, "grad_norm": 0.06161477789282799, "learning_rate": 1.1872305215217811e-05, "loss": 0.0009, "num_input_tokens_seen": 52047856, "step": 27040 }, { "epoch": 4.4119422465127665, "grad_norm": 0.01575692556798458, "learning_rate": 1.186395120395514e-05, "loss": 0.0003, "num_input_tokens_seen": 52057520, "step": 27045 }, { "epoch": 4.412757973733584, "grad_norm": 0.27089226245880127, "learning_rate": 1.1855599218436283e-05, "loss": 0.0008, "num_input_tokens_seen": 52065984, "step": 27050 }, { "epoch": 4.413573700954401, "grad_norm": 1.7325830459594727, "learning_rate": 1.1847249259949209e-05, "loss": 0.0024, "num_input_tokens_seen": 52076000, "step": 27055 }, { "epoch": 4.414389428175218, "grad_norm": 0.04089543595910072, "learning_rate": 1.1838901329781574e-05, "loss": 0.0002, "num_input_tokens_seen": 52085424, "step": 27060 }, { "epoch": 4.415205155396036, "grad_norm": 0.008485835045576096, "learning_rate": 1.1830555429220758e-05, "loss": 0.0005, "num_input_tokens_seen": 52094384, "step": 27065 }, { "epoch": 4.416020882616853, "grad_norm": 0.16975021362304688, "learning_rate": 1.1822211559553784e-05, "loss": 0.0026, "num_input_tokens_seen": 52104304, "step": 27070 }, { "epoch": 4.41683660983767, "grad_norm": 0.013496994972229004, "learning_rate": 1.18138697220674e-05, "loss": 0.0757, "num_input_tokens_seen": 52114928, "step": 27075 }, { "epoch": 4.417652337058487, "grad_norm": 0.021854763850569725, "learning_rate": 1.1805529918048e-05, "loss": 0.031, "num_input_tokens_seen": 52125456, "step": 27080 }, { "epoch": 4.418468064279305, "grad_norm": 0.0023511468898504972, "learning_rate": 1.1797192148781702e-05, "loss": 0.0007, "num_input_tokens_seen": 52134512, "step": 27085 }, { "epoch": 4.419283791500122, "grad_norm": 0.0012605227530002594, "learning_rate": 1.1788856415554297e-05, "loss": 0.0017, "num_input_tokens_seen": 52144816, "step": 27090 }, { "epoch": 4.4200995187209395, "grad_norm": 0.04577717185020447, "learning_rate": 1.1780522719651249e-05, "loss": 0.0005, "num_input_tokens_seen": 52153936, "step": 27095 }, { "epoch": 4.420915245941757, "grad_norm": 0.04085297882556915, "learning_rate": 1.1772191062357721e-05, "loss": 0.0009, "num_input_tokens_seen": 52164032, "step": 27100 }, { "epoch": 4.421730973162575, "grad_norm": 57.03218078613281, "learning_rate": 1.1763861444958573e-05, "loss": 0.1691, "num_input_tokens_seen": 52174160, "step": 27105 }, { "epoch": 4.422546700383392, "grad_norm": 0.004571868572384119, "learning_rate": 1.1755533868738317e-05, "loss": 0.0019, "num_input_tokens_seen": 52185008, "step": 27110 }, { "epoch": 4.423362427604209, "grad_norm": 7.581484317779541, "learning_rate": 1.1747208334981185e-05, "loss": 0.0883, "num_input_tokens_seen": 52194432, "step": 27115 }, { "epoch": 4.424178154825027, "grad_norm": 0.014398230239748955, "learning_rate": 1.1738884844971067e-05, "loss": 0.1209, "num_input_tokens_seen": 52204176, "step": 27120 }, { "epoch": 4.424993882045844, "grad_norm": 0.35722947120666504, "learning_rate": 1.1730563399991563e-05, "loss": 0.0538, "num_input_tokens_seen": 52214048, "step": 27125 }, { "epoch": 4.425809609266661, "grad_norm": 0.028922293335199356, "learning_rate": 1.1722244001325938e-05, "loss": 0.0005, "num_input_tokens_seen": 52223872, "step": 27130 }, { "epoch": 4.426625336487478, "grad_norm": 0.01724296249449253, "learning_rate": 1.1713926650257137e-05, "loss": 0.0003, "num_input_tokens_seen": 52233376, "step": 27135 }, { "epoch": 4.4274410637082955, "grad_norm": 0.03567345812916756, "learning_rate": 1.170561134806781e-05, "loss": 0.0006, "num_input_tokens_seen": 52243424, "step": 27140 }, { "epoch": 4.4282567909291135, "grad_norm": 0.0026640091091394424, "learning_rate": 1.1697298096040287e-05, "loss": 0.0006, "num_input_tokens_seen": 52253408, "step": 27145 }, { "epoch": 4.429072518149931, "grad_norm": 0.0045626116916537285, "learning_rate": 1.1688986895456567e-05, "loss": 0.0005, "num_input_tokens_seen": 52261808, "step": 27150 }, { "epoch": 4.429888245370748, "grad_norm": 18.008512496948242, "learning_rate": 1.1680677747598349e-05, "loss": 0.0829, "num_input_tokens_seen": 52271120, "step": 27155 }, { "epoch": 4.430703972591566, "grad_norm": 0.0018856257665902376, "learning_rate": 1.1672370653746995e-05, "loss": 0.0004, "num_input_tokens_seen": 52280896, "step": 27160 }, { "epoch": 4.431519699812383, "grad_norm": 0.019878484308719635, "learning_rate": 1.166406561518357e-05, "loss": 0.0009, "num_input_tokens_seen": 52289920, "step": 27165 }, { "epoch": 4.4323354270332, "grad_norm": 7.366093158721924, "learning_rate": 1.1655762633188826e-05, "loss": 0.0671, "num_input_tokens_seen": 52299664, "step": 27170 }, { "epoch": 4.433151154254017, "grad_norm": 0.005082122515887022, "learning_rate": 1.1647461709043172e-05, "loss": 0.001, "num_input_tokens_seen": 52310336, "step": 27175 }, { "epoch": 4.433966881474835, "grad_norm": 0.0040321326814591885, "learning_rate": 1.1639162844026722e-05, "loss": 0.0994, "num_input_tokens_seen": 52321552, "step": 27180 }, { "epoch": 4.434782608695652, "grad_norm": 0.08862625807523727, "learning_rate": 1.163086603941927e-05, "loss": 0.0018, "num_input_tokens_seen": 52332352, "step": 27185 }, { "epoch": 4.4355983359164695, "grad_norm": 0.011311000213027, "learning_rate": 1.1622571296500273e-05, "loss": 0.0016, "num_input_tokens_seen": 52342816, "step": 27190 }, { "epoch": 4.436414063137287, "grad_norm": 0.0016275796806439757, "learning_rate": 1.1614278616548904e-05, "loss": 0.0001, "num_input_tokens_seen": 52353376, "step": 27195 }, { "epoch": 4.437229790358105, "grad_norm": 0.011404612101614475, "learning_rate": 1.1605988000843986e-05, "loss": 0.0639, "num_input_tokens_seen": 52363216, "step": 27200 }, { "epoch": 4.437229790358105, "eval_loss": 0.31266775727272034, "eval_runtime": 80.8336, "eval_samples_per_second": 33.711, "eval_steps_per_second": 16.862, "num_input_tokens_seen": 52363216, "step": 27200 }, { "epoch": 4.438045517578922, "grad_norm": 0.0016413782723248005, "learning_rate": 1.1597699450664028e-05, "loss": 0.0005, "num_input_tokens_seen": 52371984, "step": 27205 }, { "epoch": 4.438861244799739, "grad_norm": 0.05604039132595062, "learning_rate": 1.1589412967287252e-05, "loss": 0.0922, "num_input_tokens_seen": 52380368, "step": 27210 }, { "epoch": 4.439676972020556, "grad_norm": 0.04327915608882904, "learning_rate": 1.1581128551991514e-05, "loss": 0.0004, "num_input_tokens_seen": 52389744, "step": 27215 }, { "epoch": 4.440492699241374, "grad_norm": 13.137255668640137, "learning_rate": 1.1572846206054383e-05, "loss": 0.0624, "num_input_tokens_seen": 52398784, "step": 27220 }, { "epoch": 4.441308426462191, "grad_norm": 0.10487014055252075, "learning_rate": 1.1564565930753113e-05, "loss": 0.0806, "num_input_tokens_seen": 52409248, "step": 27225 }, { "epoch": 4.442124153683008, "grad_norm": 13.543889999389648, "learning_rate": 1.1556287727364606e-05, "loss": 0.0663, "num_input_tokens_seen": 52419376, "step": 27230 }, { "epoch": 4.442939880903825, "grad_norm": 0.06203262135386467, "learning_rate": 1.1548011597165489e-05, "loss": 0.1111, "num_input_tokens_seen": 52429152, "step": 27235 }, { "epoch": 4.443755608124643, "grad_norm": 6.730424404144287, "learning_rate": 1.1539737541432019e-05, "loss": 0.1147, "num_input_tokens_seen": 52437760, "step": 27240 }, { "epoch": 4.444571335345461, "grad_norm": 0.0218763817101717, "learning_rate": 1.1531465561440174e-05, "loss": 0.1255, "num_input_tokens_seen": 52446736, "step": 27245 }, { "epoch": 4.445387062566278, "grad_norm": 0.1511482149362564, "learning_rate": 1.1523195658465605e-05, "loss": 0.002, "num_input_tokens_seen": 52456032, "step": 27250 }, { "epoch": 4.446202789787095, "grad_norm": 0.06124471127986908, "learning_rate": 1.1514927833783618e-05, "loss": 0.0561, "num_input_tokens_seen": 52465312, "step": 27255 }, { "epoch": 4.447018517007913, "grad_norm": 0.02266511134803295, "learning_rate": 1.150666208866922e-05, "loss": 0.0007, "num_input_tokens_seen": 52474944, "step": 27260 }, { "epoch": 4.44783424422873, "grad_norm": 0.17816798388957977, "learning_rate": 1.1498398424397106e-05, "loss": 0.0015, "num_input_tokens_seen": 52484352, "step": 27265 }, { "epoch": 4.448649971449547, "grad_norm": 0.5943220257759094, "learning_rate": 1.1490136842241628e-05, "loss": 0.0016, "num_input_tokens_seen": 52494464, "step": 27270 }, { "epoch": 4.449465698670364, "grad_norm": 0.06546531617641449, "learning_rate": 1.1481877343476813e-05, "loss": 0.0006, "num_input_tokens_seen": 52503024, "step": 27275 }, { "epoch": 4.450281425891182, "grad_norm": 0.05388158559799194, "learning_rate": 1.14736199293764e-05, "loss": 0.0007, "num_input_tokens_seen": 52512896, "step": 27280 }, { "epoch": 4.451097153111999, "grad_norm": 0.025695666670799255, "learning_rate": 1.1465364601213771e-05, "loss": 0.1013, "num_input_tokens_seen": 52523072, "step": 27285 }, { "epoch": 4.4519128803328165, "grad_norm": 25.730215072631836, "learning_rate": 1.1457111360262012e-05, "loss": 0.0805, "num_input_tokens_seen": 52532016, "step": 27290 }, { "epoch": 4.4527286075536345, "grad_norm": 0.03942424803972244, "learning_rate": 1.1448860207793869e-05, "loss": 0.0008, "num_input_tokens_seen": 52541744, "step": 27295 }, { "epoch": 4.453544334774452, "grad_norm": 0.02362404204905033, "learning_rate": 1.144061114508177e-05, "loss": 0.059, "num_input_tokens_seen": 52551520, "step": 27300 }, { "epoch": 4.454360061995269, "grad_norm": 0.0014905639691278338, "learning_rate": 1.1432364173397842e-05, "loss": 0.0005, "num_input_tokens_seen": 52560320, "step": 27305 }, { "epoch": 4.455175789216086, "grad_norm": 0.03810155391693115, "learning_rate": 1.1424119294013852e-05, "loss": 0.0005, "num_input_tokens_seen": 52569408, "step": 27310 }, { "epoch": 4.455991516436903, "grad_norm": 0.034818969666957855, "learning_rate": 1.1415876508201279e-05, "loss": 0.0012, "num_input_tokens_seen": 52579632, "step": 27315 }, { "epoch": 4.456807243657721, "grad_norm": 0.015210271812975407, "learning_rate": 1.140763581723125e-05, "loss": 0.0008, "num_input_tokens_seen": 52588336, "step": 27320 }, { "epoch": 4.457622970878538, "grad_norm": 0.006266670301556587, "learning_rate": 1.1399397222374588e-05, "loss": 0.0008, "num_input_tokens_seen": 52597840, "step": 27325 }, { "epoch": 4.458438698099355, "grad_norm": 0.019938131794333458, "learning_rate": 1.1391160724901804e-05, "loss": 0.0013, "num_input_tokens_seen": 52608624, "step": 27330 }, { "epoch": 4.459254425320173, "grad_norm": 0.0973593071103096, "learning_rate": 1.138292632608304e-05, "loss": 0.0013, "num_input_tokens_seen": 52617328, "step": 27335 }, { "epoch": 4.4600701525409905, "grad_norm": 0.045006413012742996, "learning_rate": 1.1374694027188174e-05, "loss": 0.0018, "num_input_tokens_seen": 52626912, "step": 27340 }, { "epoch": 4.460885879761808, "grad_norm": 0.051531385630369186, "learning_rate": 1.1366463829486711e-05, "loss": 0.0003, "num_input_tokens_seen": 52635184, "step": 27345 }, { "epoch": 4.461701606982625, "grad_norm": 0.0028297114185988903, "learning_rate": 1.1358235734247849e-05, "loss": 0.055, "num_input_tokens_seen": 52643920, "step": 27350 }, { "epoch": 4.462517334203443, "grad_norm": 0.0024091554805636406, "learning_rate": 1.1350009742740478e-05, "loss": 0.053, "num_input_tokens_seen": 52652880, "step": 27355 }, { "epoch": 4.46333306142426, "grad_norm": 0.349821001291275, "learning_rate": 1.134178585623313e-05, "loss": 0.1384, "num_input_tokens_seen": 52662848, "step": 27360 }, { "epoch": 4.464148788645077, "grad_norm": 11.847404479980469, "learning_rate": 1.1333564075994047e-05, "loss": 0.0374, "num_input_tokens_seen": 52671616, "step": 27365 }, { "epoch": 4.464964515865894, "grad_norm": 0.06208675727248192, "learning_rate": 1.1325344403291133e-05, "loss": 0.0004, "num_input_tokens_seen": 52680992, "step": 27370 }, { "epoch": 4.465780243086712, "grad_norm": 0.00231663859449327, "learning_rate": 1.1317126839391951e-05, "loss": 0.0005, "num_input_tokens_seen": 52690128, "step": 27375 }, { "epoch": 4.466595970307529, "grad_norm": 0.12235964834690094, "learning_rate": 1.1308911385563766e-05, "loss": 0.0482, "num_input_tokens_seen": 52698144, "step": 27380 }, { "epoch": 4.467411697528346, "grad_norm": 0.003220735117793083, "learning_rate": 1.1300698043073494e-05, "loss": 0.0045, "num_input_tokens_seen": 52706720, "step": 27385 }, { "epoch": 4.468227424749164, "grad_norm": 0.019529301673173904, "learning_rate": 1.1292486813187736e-05, "loss": 0.0003, "num_input_tokens_seen": 52717040, "step": 27390 }, { "epoch": 4.469043151969982, "grad_norm": 0.29283371567726135, "learning_rate": 1.1284277697172782e-05, "loss": 0.0048, "num_input_tokens_seen": 52727280, "step": 27395 }, { "epoch": 4.469858879190799, "grad_norm": 8.523221015930176, "learning_rate": 1.127607069629456e-05, "loss": 0.0635, "num_input_tokens_seen": 52737552, "step": 27400 }, { "epoch": 4.469858879190799, "eval_loss": 0.30705398321151733, "eval_runtime": 80.8571, "eval_samples_per_second": 33.701, "eval_steps_per_second": 16.857, "num_input_tokens_seen": 52737552, "step": 27400 }, { "epoch": 4.470674606411616, "grad_norm": 0.011841274797916412, "learning_rate": 1.1267865811818701e-05, "loss": 0.0015, "num_input_tokens_seen": 52747440, "step": 27405 }, { "epoch": 4.471490333632433, "grad_norm": 0.0014679136220365763, "learning_rate": 1.1259663045010513e-05, "loss": 0.0146, "num_input_tokens_seen": 52756400, "step": 27410 }, { "epoch": 4.472306060853251, "grad_norm": 0.003348392201587558, "learning_rate": 1.1251462397134957e-05, "loss": 0.1024, "num_input_tokens_seen": 52766176, "step": 27415 }, { "epoch": 4.473121788074068, "grad_norm": 0.00517160864546895, "learning_rate": 1.1243263869456664e-05, "loss": 0.0001, "num_input_tokens_seen": 52775952, "step": 27420 }, { "epoch": 4.473937515294885, "grad_norm": 0.0016489183763042092, "learning_rate": 1.1235067463239967e-05, "loss": 0.0007, "num_input_tokens_seen": 52784800, "step": 27425 }, { "epoch": 4.474753242515702, "grad_norm": 0.015711165964603424, "learning_rate": 1.122687317974884e-05, "loss": 0.0003, "num_input_tokens_seen": 52792928, "step": 27430 }, { "epoch": 4.47556896973652, "grad_norm": 0.1817866414785385, "learning_rate": 1.1218681020246963e-05, "loss": 0.0024, "num_input_tokens_seen": 52802624, "step": 27435 }, { "epoch": 4.4763846969573375, "grad_norm": 0.00916200876235962, "learning_rate": 1.1210490985997652e-05, "loss": 0.0004, "num_input_tokens_seen": 52812832, "step": 27440 }, { "epoch": 4.477200424178155, "grad_norm": 0.02062566950917244, "learning_rate": 1.1202303078263917e-05, "loss": 0.0014, "num_input_tokens_seen": 52821392, "step": 27445 }, { "epoch": 4.478016151398972, "grad_norm": 0.055763598531484604, "learning_rate": 1.1194117298308451e-05, "loss": 0.0005, "num_input_tokens_seen": 52830912, "step": 27450 }, { "epoch": 4.47883187861979, "grad_norm": 10.963383674621582, "learning_rate": 1.1185933647393585e-05, "loss": 0.1572, "num_input_tokens_seen": 52840752, "step": 27455 }, { "epoch": 4.479647605840607, "grad_norm": 0.007039172574877739, "learning_rate": 1.1177752126781354e-05, "loss": 0.0609, "num_input_tokens_seen": 52850720, "step": 27460 }, { "epoch": 4.480463333061424, "grad_norm": 0.0014556105015799403, "learning_rate": 1.1169572737733441e-05, "loss": 0.0007, "num_input_tokens_seen": 52860848, "step": 27465 }, { "epoch": 4.481279060282241, "grad_norm": 19.080047607421875, "learning_rate": 1.1161395481511216e-05, "loss": 0.0959, "num_input_tokens_seen": 52870336, "step": 27470 }, { "epoch": 4.482094787503059, "grad_norm": 1.7914551496505737, "learning_rate": 1.1153220359375722e-05, "loss": 0.0015, "num_input_tokens_seen": 52879584, "step": 27475 }, { "epoch": 4.482910514723876, "grad_norm": 0.007308828178793192, "learning_rate": 1.114504737258765e-05, "loss": 0.0006, "num_input_tokens_seen": 52888608, "step": 27480 }, { "epoch": 4.4837262419446935, "grad_norm": 0.030152933672070503, "learning_rate": 1.1136876522407393e-05, "loss": 0.0002, "num_input_tokens_seen": 52896928, "step": 27485 }, { "epoch": 4.484541969165511, "grad_norm": 0.009968513622879982, "learning_rate": 1.1128707810094985e-05, "loss": 0.0021, "num_input_tokens_seen": 52905680, "step": 27490 }, { "epoch": 4.485357696386329, "grad_norm": 1.176693320274353, "learning_rate": 1.1120541236910157e-05, "loss": 0.0545, "num_input_tokens_seen": 52914448, "step": 27495 }, { "epoch": 4.486173423607146, "grad_norm": 0.02881661057472229, "learning_rate": 1.111237680411229e-05, "loss": 0.0013, "num_input_tokens_seen": 52923904, "step": 27500 }, { "epoch": 4.486989150827963, "grad_norm": 0.016001030802726746, "learning_rate": 1.1104214512960433e-05, "loss": 0.0002, "num_input_tokens_seen": 52932832, "step": 27505 }, { "epoch": 4.487804878048781, "grad_norm": 1.0003025531768799, "learning_rate": 1.1096054364713327e-05, "loss": 0.0025, "num_input_tokens_seen": 52941440, "step": 27510 }, { "epoch": 4.488620605269598, "grad_norm": 0.10789652168750763, "learning_rate": 1.1087896360629371e-05, "loss": 0.13, "num_input_tokens_seen": 52949488, "step": 27515 }, { "epoch": 4.489436332490415, "grad_norm": 0.04929972067475319, "learning_rate": 1.107974050196662e-05, "loss": 0.0004, "num_input_tokens_seen": 52959600, "step": 27520 }, { "epoch": 4.490252059711232, "grad_norm": 0.008675298653542995, "learning_rate": 1.1071586789982816e-05, "loss": 0.1203, "num_input_tokens_seen": 52970464, "step": 27525 }, { "epoch": 4.49106778693205, "grad_norm": 0.023729365319013596, "learning_rate": 1.1063435225935373e-05, "loss": 0.0004, "num_input_tokens_seen": 52979104, "step": 27530 }, { "epoch": 4.4918835141528675, "grad_norm": 0.008539385162293911, "learning_rate": 1.1055285811081348e-05, "loss": 0.0015, "num_input_tokens_seen": 52989296, "step": 27535 }, { "epoch": 4.492699241373685, "grad_norm": 0.08213286101818085, "learning_rate": 1.1047138546677499e-05, "loss": 0.0004, "num_input_tokens_seen": 52998352, "step": 27540 }, { "epoch": 4.493514968594502, "grad_norm": 0.009643204510211945, "learning_rate": 1.1038993433980219e-05, "loss": 0.0152, "num_input_tokens_seen": 53007632, "step": 27545 }, { "epoch": 4.49433069581532, "grad_norm": 0.041353657841682434, "learning_rate": 1.1030850474245597e-05, "loss": 0.0652, "num_input_tokens_seen": 53016992, "step": 27550 }, { "epoch": 4.495146423036137, "grad_norm": 56.41768264770508, "learning_rate": 1.102270966872939e-05, "loss": 0.0464, "num_input_tokens_seen": 53026032, "step": 27555 }, { "epoch": 4.495962150256954, "grad_norm": 0.01682122051715851, "learning_rate": 1.1014571018687e-05, "loss": 0.1826, "num_input_tokens_seen": 53035296, "step": 27560 }, { "epoch": 4.496777877477771, "grad_norm": 0.006626329384744167, "learning_rate": 1.1006434525373502e-05, "loss": 0.0006, "num_input_tokens_seen": 53044560, "step": 27565 }, { "epoch": 4.497593604698589, "grad_norm": 0.0320235937833786, "learning_rate": 1.0998300190043664e-05, "loss": 0.0002, "num_input_tokens_seen": 53054992, "step": 27570 }, { "epoch": 4.498409331919406, "grad_norm": 0.06648106127977371, "learning_rate": 1.0990168013951882e-05, "loss": 0.0715, "num_input_tokens_seen": 53065456, "step": 27575 }, { "epoch": 4.499225059140223, "grad_norm": 0.0061197299510240555, "learning_rate": 1.0982037998352263e-05, "loss": 0.0003, "num_input_tokens_seen": 53074800, "step": 27580 }, { "epoch": 4.5000407863610405, "grad_norm": 0.061199869960546494, "learning_rate": 1.0973910144498534e-05, "loss": 0.034, "num_input_tokens_seen": 53084912, "step": 27585 }, { "epoch": 4.500856513581859, "grad_norm": 0.008911139331758022, "learning_rate": 1.0965784453644123e-05, "loss": 0.0003, "num_input_tokens_seen": 53094288, "step": 27590 }, { "epoch": 4.501672240802676, "grad_norm": 15.988982200622559, "learning_rate": 1.0957660927042127e-05, "loss": 0.024, "num_input_tokens_seen": 53103616, "step": 27595 }, { "epoch": 4.502487968023493, "grad_norm": 2.400892972946167, "learning_rate": 1.094953956594527e-05, "loss": 0.0045, "num_input_tokens_seen": 53112128, "step": 27600 }, { "epoch": 4.502487968023493, "eval_loss": 0.30987748503685, "eval_runtime": 80.658, "eval_samples_per_second": 33.785, "eval_steps_per_second": 16.899, "num_input_tokens_seen": 53112128, "step": 27600 }, { "epoch": 4.50330369524431, "grad_norm": 0.04038749262690544, "learning_rate": 1.0941420371605981e-05, "loss": 0.0002, "num_input_tokens_seen": 53122464, "step": 27605 }, { "epoch": 4.504119422465128, "grad_norm": 0.3109874427318573, "learning_rate": 1.0933303345276354e-05, "loss": 0.0011, "num_input_tokens_seen": 53132464, "step": 27610 }, { "epoch": 4.504935149685945, "grad_norm": 1.4290670156478882, "learning_rate": 1.0925188488208112e-05, "loss": 0.2077, "num_input_tokens_seen": 53142384, "step": 27615 }, { "epoch": 4.505750876906762, "grad_norm": 0.0432475283741951, "learning_rate": 1.0917075801652694e-05, "loss": 0.0055, "num_input_tokens_seen": 53151984, "step": 27620 }, { "epoch": 4.506566604127579, "grad_norm": 0.3086196780204773, "learning_rate": 1.0908965286861151e-05, "loss": 0.0007, "num_input_tokens_seen": 53160896, "step": 27625 }, { "epoch": 4.507382331348397, "grad_norm": 6.639530181884766, "learning_rate": 1.090085694508425e-05, "loss": 0.0589, "num_input_tokens_seen": 53169792, "step": 27630 }, { "epoch": 4.5081980585692145, "grad_norm": 0.02406887337565422, "learning_rate": 1.089275077757238e-05, "loss": 0.0014, "num_input_tokens_seen": 53178656, "step": 27635 }, { "epoch": 4.509013785790032, "grad_norm": 0.0226896982640028, "learning_rate": 1.0884646785575633e-05, "loss": 0.0006, "num_input_tokens_seen": 53186880, "step": 27640 }, { "epoch": 4.50982951301085, "grad_norm": 0.028864959254860878, "learning_rate": 1.0876544970343728e-05, "loss": 0.0755, "num_input_tokens_seen": 53195440, "step": 27645 }, { "epoch": 4.510645240231667, "grad_norm": 45.3610954284668, "learning_rate": 1.0868445333126082e-05, "loss": 0.0902, "num_input_tokens_seen": 53205008, "step": 27650 }, { "epoch": 4.511460967452484, "grad_norm": 1.206969141960144, "learning_rate": 1.0860347875171745e-05, "loss": 0.0026, "num_input_tokens_seen": 53213536, "step": 27655 }, { "epoch": 4.512276694673301, "grad_norm": 0.0038382208440452814, "learning_rate": 1.0852252597729465e-05, "loss": 0.0003, "num_input_tokens_seen": 53222080, "step": 27660 }, { "epoch": 4.513092421894118, "grad_norm": 0.0003795329248532653, "learning_rate": 1.0844159502047615e-05, "loss": 0.0003, "num_input_tokens_seen": 53232000, "step": 27665 }, { "epoch": 4.513908149114936, "grad_norm": 0.04083646088838577, "learning_rate": 1.0836068589374265e-05, "loss": 0.0002, "num_input_tokens_seen": 53240944, "step": 27670 }, { "epoch": 4.514723876335753, "grad_norm": 0.017296355217695236, "learning_rate": 1.0827979860957144e-05, "loss": 0.0017, "num_input_tokens_seen": 53252144, "step": 27675 }, { "epoch": 4.5155396035565705, "grad_norm": 0.0026963448617607355, "learning_rate": 1.0819893318043615e-05, "loss": 0.0001, "num_input_tokens_seen": 53261360, "step": 27680 }, { "epoch": 4.5163553307773885, "grad_norm": 0.0013875265140086412, "learning_rate": 1.0811808961880734e-05, "loss": 0.0003, "num_input_tokens_seen": 53271712, "step": 27685 }, { "epoch": 4.517171057998206, "grad_norm": 0.006388282869011164, "learning_rate": 1.080372679371522e-05, "loss": 0.0045, "num_input_tokens_seen": 53279984, "step": 27690 }, { "epoch": 4.517986785219023, "grad_norm": 1.375949501991272, "learning_rate": 1.0795646814793428e-05, "loss": 0.0034, "num_input_tokens_seen": 53288880, "step": 27695 }, { "epoch": 4.51880251243984, "grad_norm": 0.1493152529001236, "learning_rate": 1.078756902636141e-05, "loss": 0.0007, "num_input_tokens_seen": 53298272, "step": 27700 }, { "epoch": 4.519618239660657, "grad_norm": 0.0022660507820546627, "learning_rate": 1.077949342966485e-05, "loss": 0.0001, "num_input_tokens_seen": 53307600, "step": 27705 }, { "epoch": 4.520433966881475, "grad_norm": 0.0036345673725008965, "learning_rate": 1.0771420025949103e-05, "loss": 0.0302, "num_input_tokens_seen": 53315968, "step": 27710 }, { "epoch": 4.521249694102292, "grad_norm": 0.005443000700324774, "learning_rate": 1.0763348816459204e-05, "loss": 0.0002, "num_input_tokens_seen": 53324864, "step": 27715 }, { "epoch": 4.522065421323109, "grad_norm": 0.004841735120862722, "learning_rate": 1.0755279802439816e-05, "loss": 0.0006, "num_input_tokens_seen": 53335360, "step": 27720 }, { "epoch": 4.522881148543927, "grad_norm": 0.005408790893852711, "learning_rate": 1.0747212985135293e-05, "loss": 0.059, "num_input_tokens_seen": 53344400, "step": 27725 }, { "epoch": 4.523696875764744, "grad_norm": 0.01719633862376213, "learning_rate": 1.073914836578965e-05, "loss": 0.0001, "num_input_tokens_seen": 53353904, "step": 27730 }, { "epoch": 4.524512602985562, "grad_norm": 0.13878685235977173, "learning_rate": 1.0731085945646529e-05, "loss": 0.0646, "num_input_tokens_seen": 53363408, "step": 27735 }, { "epoch": 4.525328330206379, "grad_norm": 0.11846911907196045, "learning_rate": 1.0723025725949285e-05, "loss": 0.0007, "num_input_tokens_seen": 53373520, "step": 27740 }, { "epoch": 4.526144057427197, "grad_norm": 10.024771690368652, "learning_rate": 1.0714967707940875e-05, "loss": 0.0692, "num_input_tokens_seen": 53384256, "step": 27745 }, { "epoch": 4.526959784648014, "grad_norm": 0.004513456020504236, "learning_rate": 1.0706911892863963e-05, "loss": 0.0005, "num_input_tokens_seen": 53393328, "step": 27750 }, { "epoch": 4.527775511868831, "grad_norm": 21.407913208007812, "learning_rate": 1.0698858281960866e-05, "loss": 0.1024, "num_input_tokens_seen": 53402464, "step": 27755 }, { "epoch": 4.528591239089648, "grad_norm": 7.185456275939941, "learning_rate": 1.069080687647353e-05, "loss": 0.0922, "num_input_tokens_seen": 53412448, "step": 27760 }, { "epoch": 4.529406966310466, "grad_norm": 0.004113584291189909, "learning_rate": 1.0682757677643596e-05, "loss": 0.0918, "num_input_tokens_seen": 53421200, "step": 27765 }, { "epoch": 4.530222693531283, "grad_norm": 0.015671275556087494, "learning_rate": 1.0674710686712359e-05, "loss": 0.1218, "num_input_tokens_seen": 53430704, "step": 27770 }, { "epoch": 4.5310384207521, "grad_norm": 0.0064125582575798035, "learning_rate": 1.0666665904920756e-05, "loss": 0.0004, "num_input_tokens_seen": 53441264, "step": 27775 }, { "epoch": 4.5318541479729175, "grad_norm": 0.028578242287039757, "learning_rate": 1.0658623333509385e-05, "loss": 0.0004, "num_input_tokens_seen": 53451344, "step": 27780 }, { "epoch": 4.5326698751937355, "grad_norm": 0.0403490774333477, "learning_rate": 1.0650582973718532e-05, "loss": 0.0004, "num_input_tokens_seen": 53460576, "step": 27785 }, { "epoch": 4.533485602414553, "grad_norm": 9.025498390197754, "learning_rate": 1.0642544826788098e-05, "loss": 0.1264, "num_input_tokens_seen": 53469968, "step": 27790 }, { "epoch": 4.53430132963537, "grad_norm": 0.010753270238637924, "learning_rate": 1.063450889395769e-05, "loss": 0.0004, "num_input_tokens_seen": 53479840, "step": 27795 }, { "epoch": 4.535117056856187, "grad_norm": 0.03198220580816269, "learning_rate": 1.062647517646653e-05, "loss": 0.053, "num_input_tokens_seen": 53489200, "step": 27800 }, { "epoch": 4.535117056856187, "eval_loss": 0.29390978813171387, "eval_runtime": 80.7627, "eval_samples_per_second": 33.741, "eval_steps_per_second": 16.877, "num_input_tokens_seen": 53489200, "step": 27800 }, { "epoch": 4.535932784077005, "grad_norm": 0.033159129321575165, "learning_rate": 1.0618443675553527e-05, "loss": 0.0015, "num_input_tokens_seen": 53498336, "step": 27805 }, { "epoch": 4.536748511297822, "grad_norm": 0.004897782579064369, "learning_rate": 1.0610414392457247e-05, "loss": 0.0012, "num_input_tokens_seen": 53506784, "step": 27810 }, { "epoch": 4.537564238518639, "grad_norm": 0.005513990297913551, "learning_rate": 1.0602387328415888e-05, "loss": 0.0009, "num_input_tokens_seen": 53517936, "step": 27815 }, { "epoch": 4.538379965739456, "grad_norm": 0.005346968770027161, "learning_rate": 1.0594362484667347e-05, "loss": 0.0545, "num_input_tokens_seen": 53528432, "step": 27820 }, { "epoch": 4.539195692960274, "grad_norm": 8.873652458190918, "learning_rate": 1.0586339862449132e-05, "loss": 0.1314, "num_input_tokens_seen": 53537824, "step": 27825 }, { "epoch": 4.5400114201810915, "grad_norm": 0.05593280494213104, "learning_rate": 1.0578319462998445e-05, "loss": 0.0005, "num_input_tokens_seen": 53546608, "step": 27830 }, { "epoch": 4.540827147401909, "grad_norm": 0.0021466645412147045, "learning_rate": 1.057030128755214e-05, "loss": 0.0003, "num_input_tokens_seen": 53556896, "step": 27835 }, { "epoch": 4.541642874622726, "grad_norm": 0.025468943640589714, "learning_rate": 1.0562285337346703e-05, "loss": 0.0754, "num_input_tokens_seen": 53565328, "step": 27840 }, { "epoch": 4.542458601843544, "grad_norm": 0.0066618481650948524, "learning_rate": 1.0554271613618308e-05, "loss": 0.0692, "num_input_tokens_seen": 53574272, "step": 27845 }, { "epoch": 4.543274329064361, "grad_norm": 0.020041652023792267, "learning_rate": 1.054626011760276e-05, "loss": 0.0474, "num_input_tokens_seen": 53583680, "step": 27850 }, { "epoch": 4.544090056285178, "grad_norm": 8.167065620422363, "learning_rate": 1.0538250850535549e-05, "loss": 0.0838, "num_input_tokens_seen": 53593232, "step": 27855 }, { "epoch": 4.544905783505996, "grad_norm": 0.011120961047708988, "learning_rate": 1.0530243813651794e-05, "loss": 0.0003, "num_input_tokens_seen": 53603248, "step": 27860 }, { "epoch": 4.545721510726813, "grad_norm": 0.10083500295877457, "learning_rate": 1.0522239008186271e-05, "loss": 0.0006, "num_input_tokens_seen": 53612656, "step": 27865 }, { "epoch": 4.54653723794763, "grad_norm": 0.036288660019636154, "learning_rate": 1.0514236435373434e-05, "loss": 0.0004, "num_input_tokens_seen": 53622240, "step": 27870 }, { "epoch": 4.5473529651684474, "grad_norm": 0.01678488776087761, "learning_rate": 1.0506236096447386e-05, "loss": 0.0008, "num_input_tokens_seen": 53633776, "step": 27875 }, { "epoch": 4.548168692389265, "grad_norm": 0.0308346226811409, "learning_rate": 1.049823799264186e-05, "loss": 0.001, "num_input_tokens_seen": 53641872, "step": 27880 }, { "epoch": 4.548984419610083, "grad_norm": 0.30307668447494507, "learning_rate": 1.049024212519028e-05, "loss": 0.0014, "num_input_tokens_seen": 53652256, "step": 27885 }, { "epoch": 4.5498001468309, "grad_norm": 10.468829154968262, "learning_rate": 1.0482248495325713e-05, "loss": 0.1147, "num_input_tokens_seen": 53661648, "step": 27890 }, { "epoch": 4.550615874051717, "grad_norm": 0.09121499210596085, "learning_rate": 1.047425710428086e-05, "loss": 0.0435, "num_input_tokens_seen": 53671072, "step": 27895 }, { "epoch": 4.551431601272535, "grad_norm": 0.013291721232235432, "learning_rate": 1.0466267953288114e-05, "loss": 0.0322, "num_input_tokens_seen": 53680592, "step": 27900 }, { "epoch": 4.552247328493352, "grad_norm": 0.5619160532951355, "learning_rate": 1.0458281043579482e-05, "loss": 0.124, "num_input_tokens_seen": 53689568, "step": 27905 }, { "epoch": 4.553063055714169, "grad_norm": 0.054128993302583694, "learning_rate": 1.0450296376386657e-05, "loss": 0.0054, "num_input_tokens_seen": 53700096, "step": 27910 }, { "epoch": 4.553878782934986, "grad_norm": 0.0062533216550946236, "learning_rate": 1.044231395294098e-05, "loss": 0.0013, "num_input_tokens_seen": 53709776, "step": 27915 }, { "epoch": 4.554694510155803, "grad_norm": 0.003335413755849004, "learning_rate": 1.0434333774473435e-05, "loss": 0.0005, "num_input_tokens_seen": 53718640, "step": 27920 }, { "epoch": 4.555510237376621, "grad_norm": 16.001440048217773, "learning_rate": 1.0426355842214657e-05, "loss": 0.0615, "num_input_tokens_seen": 53727536, "step": 27925 }, { "epoch": 4.5563259645974385, "grad_norm": 0.022818157449364662, "learning_rate": 1.0418380157394963e-05, "loss": 0.0003, "num_input_tokens_seen": 53738224, "step": 27930 }, { "epoch": 4.557141691818256, "grad_norm": 0.10587790608406067, "learning_rate": 1.0410406721244281e-05, "loss": 0.0033, "num_input_tokens_seen": 53747168, "step": 27935 }, { "epoch": 4.557957419039074, "grad_norm": 0.023672958835959435, "learning_rate": 1.0402435534992238e-05, "loss": 0.0004, "num_input_tokens_seen": 53756656, "step": 27940 }, { "epoch": 4.558773146259891, "grad_norm": 0.003456614911556244, "learning_rate": 1.0394466599868071e-05, "loss": 0.0003, "num_input_tokens_seen": 53764208, "step": 27945 }, { "epoch": 4.559588873480708, "grad_norm": 5.794954299926758, "learning_rate": 1.0386499917100697e-05, "loss": 0.0136, "num_input_tokens_seen": 53773872, "step": 27950 }, { "epoch": 4.560404600701525, "grad_norm": 0.047837864607572556, "learning_rate": 1.0378535487918692e-05, "loss": 0.0006, "num_input_tokens_seen": 53783536, "step": 27955 }, { "epoch": 4.561220327922343, "grad_norm": 0.027019627392292023, "learning_rate": 1.037057331355025e-05, "loss": 0.0678, "num_input_tokens_seen": 53793872, "step": 27960 }, { "epoch": 4.56203605514316, "grad_norm": 0.013781754299998283, "learning_rate": 1.0362613395223247e-05, "loss": 0.0004, "num_input_tokens_seen": 53803200, "step": 27965 }, { "epoch": 4.562851782363977, "grad_norm": 0.023078085854649544, "learning_rate": 1.0354655734165212e-05, "loss": 0.0003, "num_input_tokens_seen": 53812416, "step": 27970 }, { "epoch": 4.5636675095847945, "grad_norm": 2.1715927124023438, "learning_rate": 1.03467003316033e-05, "loss": 0.0801, "num_input_tokens_seen": 53822352, "step": 27975 }, { "epoch": 4.5644832368056125, "grad_norm": 0.015285207889974117, "learning_rate": 1.033874718876435e-05, "loss": 0.096, "num_input_tokens_seen": 53832448, "step": 27980 }, { "epoch": 4.56529896402643, "grad_norm": 0.00229967525228858, "learning_rate": 1.0330796306874818e-05, "loss": 0.07, "num_input_tokens_seen": 53841264, "step": 27985 }, { "epoch": 4.566114691247247, "grad_norm": 0.005794718861579895, "learning_rate": 1.032284768716085e-05, "loss": 0.0029, "num_input_tokens_seen": 53850960, "step": 27990 }, { "epoch": 4.566930418468064, "grad_norm": 0.0550277903676033, "learning_rate": 1.0314901330848206e-05, "loss": 0.0365, "num_input_tokens_seen": 53860592, "step": 27995 }, { "epoch": 4.567746145688882, "grad_norm": 0.07197088748216629, "learning_rate": 1.030695723916233e-05, "loss": 0.011, "num_input_tokens_seen": 53870832, "step": 28000 }, { "epoch": 4.567746145688882, "eval_loss": 0.28987744450569153, "eval_runtime": 80.766, "eval_samples_per_second": 33.739, "eval_steps_per_second": 16.876, "num_input_tokens_seen": 53870832, "step": 28000 }, { "epoch": 4.568561872909699, "grad_norm": 0.007324399892240763, "learning_rate": 1.0299015413328289e-05, "loss": 0.0007, "num_input_tokens_seen": 53879408, "step": 28005 }, { "epoch": 4.569377600130516, "grad_norm": 7.437551975250244, "learning_rate": 1.0291075854570809e-05, "loss": 0.0865, "num_input_tokens_seen": 53889104, "step": 28010 }, { "epoch": 4.570193327351333, "grad_norm": 0.8852342963218689, "learning_rate": 1.0283138564114275e-05, "loss": 0.0015, "num_input_tokens_seen": 53899424, "step": 28015 }, { "epoch": 4.571009054572151, "grad_norm": 0.06898311525583267, "learning_rate": 1.027520354318273e-05, "loss": 0.0166, "num_input_tokens_seen": 53909056, "step": 28020 }, { "epoch": 4.5718247817929685, "grad_norm": 0.13384276628494263, "learning_rate": 1.0267270792999828e-05, "loss": 0.0012, "num_input_tokens_seen": 53919744, "step": 28025 }, { "epoch": 4.572640509013786, "grad_norm": 0.026798294857144356, "learning_rate": 1.0259340314788919e-05, "loss": 0.001, "num_input_tokens_seen": 53929888, "step": 28030 }, { "epoch": 4.573456236234604, "grad_norm": 0.004744985140860081, "learning_rate": 1.0251412109772979e-05, "loss": 0.0002, "num_input_tokens_seen": 53940144, "step": 28035 }, { "epoch": 4.574271963455421, "grad_norm": 0.020340129733085632, "learning_rate": 1.0243486179174627e-05, "loss": 0.0555, "num_input_tokens_seen": 53949200, "step": 28040 }, { "epoch": 4.575087690676238, "grad_norm": 0.027749815955758095, "learning_rate": 1.0235562524216158e-05, "loss": 0.001, "num_input_tokens_seen": 53958256, "step": 28045 }, { "epoch": 4.575903417897055, "grad_norm": 0.009875953197479248, "learning_rate": 1.022764114611948e-05, "loss": 0.0049, "num_input_tokens_seen": 53967744, "step": 28050 }, { "epoch": 4.576719145117872, "grad_norm": 0.06990326195955276, "learning_rate": 1.0219722046106178e-05, "loss": 0.0018, "num_input_tokens_seen": 53976096, "step": 28055 }, { "epoch": 4.57753487233869, "grad_norm": 0.0032160787377506495, "learning_rate": 1.0211805225397486e-05, "loss": 0.0798, "num_input_tokens_seen": 53986784, "step": 28060 }, { "epoch": 4.578350599559507, "grad_norm": 0.0965803787112236, "learning_rate": 1.020389068521426e-05, "loss": 0.175, "num_input_tokens_seen": 53996640, "step": 28065 }, { "epoch": 4.579166326780324, "grad_norm": 0.0012839229311794043, "learning_rate": 1.0195978426777039e-05, "loss": 0.0001, "num_input_tokens_seen": 54005072, "step": 28070 }, { "epoch": 4.5799820540011424, "grad_norm": 0.07737034559249878, "learning_rate": 1.0188068451305982e-05, "loss": 0.0004, "num_input_tokens_seen": 54015072, "step": 28075 }, { "epoch": 4.58079778122196, "grad_norm": 0.06341926753520966, "learning_rate": 1.0180160760020902e-05, "loss": 0.0656, "num_input_tokens_seen": 54023424, "step": 28080 }, { "epoch": 4.581613508442777, "grad_norm": 0.0039006827864795923, "learning_rate": 1.0172255354141278e-05, "loss": 0.113, "num_input_tokens_seen": 54031712, "step": 28085 }, { "epoch": 4.582429235663594, "grad_norm": 1.8715381622314453, "learning_rate": 1.0164352234886205e-05, "loss": 0.0026, "num_input_tokens_seen": 54040672, "step": 28090 }, { "epoch": 4.583244962884411, "grad_norm": 0.02623968943953514, "learning_rate": 1.0156451403474454e-05, "loss": 0.001, "num_input_tokens_seen": 54052528, "step": 28095 }, { "epoch": 4.584060690105229, "grad_norm": 0.0031933931168168783, "learning_rate": 1.0148552861124443e-05, "loss": 0.0014, "num_input_tokens_seen": 54061328, "step": 28100 }, { "epoch": 4.584876417326046, "grad_norm": 23.665897369384766, "learning_rate": 1.0140656609054205e-05, "loss": 0.0338, "num_input_tokens_seen": 54071216, "step": 28105 }, { "epoch": 4.585692144546863, "grad_norm": 0.3940642178058624, "learning_rate": 1.0132762648481455e-05, "loss": 0.0012, "num_input_tokens_seen": 54081216, "step": 28110 }, { "epoch": 4.586507871767681, "grad_norm": 77.94605255126953, "learning_rate": 1.0124870980623543e-05, "loss": 0.0435, "num_input_tokens_seen": 54092064, "step": 28115 }, { "epoch": 4.587323598988498, "grad_norm": 0.005735392682254314, "learning_rate": 1.0116981606697453e-05, "loss": 0.0007, "num_input_tokens_seen": 54102112, "step": 28120 }, { "epoch": 4.5881393262093155, "grad_norm": 0.015021804720163345, "learning_rate": 1.0109094527919838e-05, "loss": 0.1356, "num_input_tokens_seen": 54112656, "step": 28125 }, { "epoch": 4.588955053430133, "grad_norm": 0.02084164321422577, "learning_rate": 1.010120974550697e-05, "loss": 0.0003, "num_input_tokens_seen": 54122256, "step": 28130 }, { "epoch": 4.589770780650951, "grad_norm": 0.0420987531542778, "learning_rate": 1.0093327260674795e-05, "loss": 0.0307, "num_input_tokens_seen": 54131984, "step": 28135 }, { "epoch": 4.590586507871768, "grad_norm": 6.020059585571289, "learning_rate": 1.0085447074638878e-05, "loss": 0.0615, "num_input_tokens_seen": 54142000, "step": 28140 }, { "epoch": 4.591402235092585, "grad_norm": 0.032579611986875534, "learning_rate": 1.0077569188614461e-05, "loss": 0.0537, "num_input_tokens_seen": 54151120, "step": 28145 }, { "epoch": 4.592217962313402, "grad_norm": 0.12453698366880417, "learning_rate": 1.0069693603816393e-05, "loss": 0.1538, "num_input_tokens_seen": 54161328, "step": 28150 }, { "epoch": 4.59303368953422, "grad_norm": 0.052957527339458466, "learning_rate": 1.0061820321459204e-05, "loss": 0.0336, "num_input_tokens_seen": 54171760, "step": 28155 }, { "epoch": 4.593849416755037, "grad_norm": 0.0018845285521820188, "learning_rate": 1.0053949342757038e-05, "loss": 0.0012, "num_input_tokens_seen": 54181872, "step": 28160 }, { "epoch": 4.594665143975854, "grad_norm": 0.14406506717205048, "learning_rate": 1.0046080668923717e-05, "loss": 0.0862, "num_input_tokens_seen": 54192608, "step": 28165 }, { "epoch": 4.5954808711966715, "grad_norm": 0.02756720967590809, "learning_rate": 1.003821430117267e-05, "loss": 0.0004, "num_input_tokens_seen": 54202816, "step": 28170 }, { "epoch": 4.5962965984174895, "grad_norm": 0.004292333498597145, "learning_rate": 1.0030350240716999e-05, "loss": 0.0007, "num_input_tokens_seen": 54212816, "step": 28175 }, { "epoch": 4.597112325638307, "grad_norm": 0.020274657756090164, "learning_rate": 1.0022488488769449e-05, "loss": 0.0032, "num_input_tokens_seen": 54222032, "step": 28180 }, { "epoch": 4.597928052859124, "grad_norm": 0.0051503777503967285, "learning_rate": 1.0014629046542387e-05, "loss": 0.1109, "num_input_tokens_seen": 54231872, "step": 28185 }, { "epoch": 4.598743780079941, "grad_norm": 0.017263909801840782, "learning_rate": 1.0006771915247842e-05, "loss": 0.0089, "num_input_tokens_seen": 54241296, "step": 28190 }, { "epoch": 4.599559507300759, "grad_norm": 0.9095161557197571, "learning_rate": 9.998917096097495e-06, "loss": 0.0011, "num_input_tokens_seen": 54251024, "step": 28195 }, { "epoch": 4.600375234521576, "grad_norm": 0.017255932092666626, "learning_rate": 9.991064590302638e-06, "loss": 0.1026, "num_input_tokens_seen": 54260848, "step": 28200 }, { "epoch": 4.600375234521576, "eval_loss": 0.2872140109539032, "eval_runtime": 80.7412, "eval_samples_per_second": 33.75, "eval_steps_per_second": 16.881, "num_input_tokens_seen": 54260848, "step": 28200 }, { "epoch": 4.601190961742393, "grad_norm": 0.019199803471565247, "learning_rate": 9.983214399074241e-06, "loss": 0.1035, "num_input_tokens_seen": 54270624, "step": 28205 }, { "epoch": 4.602006688963211, "grad_norm": 0.8882226347923279, "learning_rate": 9.975366523622893e-06, "loss": 0.0022, "num_input_tokens_seen": 54280048, "step": 28210 }, { "epoch": 4.602822416184028, "grad_norm": 0.013333875685930252, "learning_rate": 9.967520965158841e-06, "loss": 0.0637, "num_input_tokens_seen": 54290096, "step": 28215 }, { "epoch": 4.6036381434048455, "grad_norm": 0.027498068287968636, "learning_rate": 9.95967772489197e-06, "loss": 0.0862, "num_input_tokens_seen": 54299120, "step": 28220 }, { "epoch": 4.604453870625663, "grad_norm": 0.045408666133880615, "learning_rate": 9.951836804031794e-06, "loss": 0.0004, "num_input_tokens_seen": 54309840, "step": 28225 }, { "epoch": 4.60526959784648, "grad_norm": 0.014682456851005554, "learning_rate": 9.943998203787489e-06, "loss": 0.0025, "num_input_tokens_seen": 54319200, "step": 28230 }, { "epoch": 4.606085325067298, "grad_norm": 13.171574592590332, "learning_rate": 9.936161925367874e-06, "loss": 0.0268, "num_input_tokens_seen": 54327520, "step": 28235 }, { "epoch": 4.606901052288115, "grad_norm": 0.1415117084980011, "learning_rate": 9.928327969981386e-06, "loss": 0.0009, "num_input_tokens_seen": 54337792, "step": 28240 }, { "epoch": 4.607716779508932, "grad_norm": 0.3344447612762451, "learning_rate": 9.920496338836135e-06, "loss": 0.006, "num_input_tokens_seen": 54348080, "step": 28245 }, { "epoch": 4.60853250672975, "grad_norm": 0.03085177205502987, "learning_rate": 9.912667033139844e-06, "loss": 0.0003, "num_input_tokens_seen": 54357680, "step": 28250 }, { "epoch": 4.609348233950567, "grad_norm": 0.005696450360119343, "learning_rate": 9.904840054099893e-06, "loss": 0.0005, "num_input_tokens_seen": 54367744, "step": 28255 }, { "epoch": 4.610163961171384, "grad_norm": 0.21314749121665955, "learning_rate": 9.897015402923312e-06, "loss": 0.0381, "num_input_tokens_seen": 54376320, "step": 28260 }, { "epoch": 4.610979688392201, "grad_norm": 0.0019701975397765636, "learning_rate": 9.889193080816744e-06, "loss": 0.1276, "num_input_tokens_seen": 54385648, "step": 28265 }, { "epoch": 4.6117954156130185, "grad_norm": 0.029932638630270958, "learning_rate": 9.881373088986498e-06, "loss": 0.0004, "num_input_tokens_seen": 54394256, "step": 28270 }, { "epoch": 4.6126111428338366, "grad_norm": 0.019936472177505493, "learning_rate": 9.873555428638523e-06, "loss": 0.093, "num_input_tokens_seen": 54403248, "step": 28275 }, { "epoch": 4.613426870054654, "grad_norm": 0.005364337470382452, "learning_rate": 9.865740100978383e-06, "loss": 0.0008, "num_input_tokens_seen": 54413248, "step": 28280 }, { "epoch": 4.614242597275471, "grad_norm": 0.08785591274499893, "learning_rate": 9.857927107211315e-06, "loss": 0.0008, "num_input_tokens_seen": 54424160, "step": 28285 }, { "epoch": 4.615058324496289, "grad_norm": 0.0077701653353869915, "learning_rate": 9.850116448542177e-06, "loss": 0.0009, "num_input_tokens_seen": 54433344, "step": 28290 }, { "epoch": 4.615874051717106, "grad_norm": 0.001606420730240643, "learning_rate": 9.842308126175457e-06, "loss": 0.0004, "num_input_tokens_seen": 54442320, "step": 28295 }, { "epoch": 4.616689778937923, "grad_norm": 0.004322581924498081, "learning_rate": 9.834502141315315e-06, "loss": 0.0001, "num_input_tokens_seen": 54451360, "step": 28300 }, { "epoch": 4.61750550615874, "grad_norm": 0.03625791147351265, "learning_rate": 9.82669849516552e-06, "loss": 0.0003, "num_input_tokens_seen": 54461216, "step": 28305 }, { "epoch": 4.618321233379558, "grad_norm": 0.0035595258232206106, "learning_rate": 9.818897188929493e-06, "loss": 0.0339, "num_input_tokens_seen": 54471824, "step": 28310 }, { "epoch": 4.619136960600375, "grad_norm": 0.03653612360358238, "learning_rate": 9.811098223810309e-06, "loss": 0.0414, "num_input_tokens_seen": 54481120, "step": 28315 }, { "epoch": 4.6199526878211925, "grad_norm": 25.781972885131836, "learning_rate": 9.803301601010641e-06, "loss": 0.0181, "num_input_tokens_seen": 54491712, "step": 28320 }, { "epoch": 4.62076841504201, "grad_norm": 0.03326942399144173, "learning_rate": 9.795507321732853e-06, "loss": 0.0007, "num_input_tokens_seen": 54501488, "step": 28325 }, { "epoch": 4.621584142262828, "grad_norm": 12.48935604095459, "learning_rate": 9.787715387178898e-06, "loss": 0.0703, "num_input_tokens_seen": 54512544, "step": 28330 }, { "epoch": 4.622399869483645, "grad_norm": 0.0037428592331707478, "learning_rate": 9.779925798550399e-06, "loss": 0.0012, "num_input_tokens_seen": 54522240, "step": 28335 }, { "epoch": 4.623215596704462, "grad_norm": 0.0026526041328907013, "learning_rate": 9.772138557048619e-06, "loss": 0.0006, "num_input_tokens_seen": 54531872, "step": 28340 }, { "epoch": 4.624031323925279, "grad_norm": 0.012956790626049042, "learning_rate": 9.764353663874426e-06, "loss": 0.0007, "num_input_tokens_seen": 54541264, "step": 28345 }, { "epoch": 4.624847051146097, "grad_norm": 0.0056615667417645454, "learning_rate": 9.756571120228375e-06, "loss": 0.001, "num_input_tokens_seen": 54549920, "step": 28350 }, { "epoch": 4.625662778366914, "grad_norm": 0.03286433219909668, "learning_rate": 9.748790927310605e-06, "loss": 0.0013, "num_input_tokens_seen": 54558832, "step": 28355 }, { "epoch": 4.626478505587731, "grad_norm": 11.80356502532959, "learning_rate": 9.741013086320946e-06, "loss": 0.1047, "num_input_tokens_seen": 54568864, "step": 28360 }, { "epoch": 4.6272942328085485, "grad_norm": 0.029570920392870903, "learning_rate": 9.733237598458821e-06, "loss": 0.0006, "num_input_tokens_seen": 54578144, "step": 28365 }, { "epoch": 4.6281099600293665, "grad_norm": 0.0166916586458683, "learning_rate": 9.725464464923308e-06, "loss": 0.0003, "num_input_tokens_seen": 54588400, "step": 28370 }, { "epoch": 4.628925687250184, "grad_norm": 0.002456363756209612, "learning_rate": 9.717693686913123e-06, "loss": 0.0716, "num_input_tokens_seen": 54598464, "step": 28375 }, { "epoch": 4.629741414471001, "grad_norm": 0.011751753278076649, "learning_rate": 9.709925265626632e-06, "loss": 0.0006, "num_input_tokens_seen": 54608528, "step": 28380 }, { "epoch": 4.630557141691818, "grad_norm": 0.03366609290242195, "learning_rate": 9.702159202261801e-06, "loss": 0.0574, "num_input_tokens_seen": 54618032, "step": 28385 }, { "epoch": 4.631372868912636, "grad_norm": 0.04889577999711037, "learning_rate": 9.694395498016268e-06, "loss": 0.0003, "num_input_tokens_seen": 54628608, "step": 28390 }, { "epoch": 4.632188596133453, "grad_norm": 0.005756722763180733, "learning_rate": 9.686634154087298e-06, "loss": 0.0005, "num_input_tokens_seen": 54638432, "step": 28395 }, { "epoch": 4.63300432335427, "grad_norm": 0.017934581264853477, "learning_rate": 9.678875171671776e-06, "loss": 0.1132, "num_input_tokens_seen": 54647840, "step": 28400 }, { "epoch": 4.63300432335427, "eval_loss": 0.30900079011917114, "eval_runtime": 80.9586, "eval_samples_per_second": 33.659, "eval_steps_per_second": 16.836, "num_input_tokens_seen": 54647840, "step": 28400 }, { "epoch": 4.633820050575087, "grad_norm": 0.004680936224758625, "learning_rate": 9.671118551966246e-06, "loss": 0.0005, "num_input_tokens_seen": 54656816, "step": 28405 }, { "epoch": 4.634635777795905, "grad_norm": 6.1630377769470215, "learning_rate": 9.66336429616686e-06, "loss": 0.059, "num_input_tokens_seen": 54666112, "step": 28410 }, { "epoch": 4.635451505016722, "grad_norm": 0.024090241640806198, "learning_rate": 9.655612405469436e-06, "loss": 0.0007, "num_input_tokens_seen": 54675440, "step": 28415 }, { "epoch": 4.63626723223754, "grad_norm": 0.004379494581371546, "learning_rate": 9.647862881069413e-06, "loss": 0.0005, "num_input_tokens_seen": 54685392, "step": 28420 }, { "epoch": 4.637082959458358, "grad_norm": 0.09437837451696396, "learning_rate": 9.640115724161855e-06, "loss": 0.0006, "num_input_tokens_seen": 54696320, "step": 28425 }, { "epoch": 4.637898686679175, "grad_norm": 0.011149146594107151, "learning_rate": 9.632370935941483e-06, "loss": 0.0006, "num_input_tokens_seen": 54705904, "step": 28430 }, { "epoch": 4.638714413899992, "grad_norm": 0.018548443913459778, "learning_rate": 9.624628517602634e-06, "loss": 0.0003, "num_input_tokens_seen": 54716016, "step": 28435 }, { "epoch": 4.639530141120809, "grad_norm": 0.006921978201717138, "learning_rate": 9.61688847033928e-06, "loss": 0.0024, "num_input_tokens_seen": 54725968, "step": 28440 }, { "epoch": 4.640345868341626, "grad_norm": 0.018930716440081596, "learning_rate": 9.609150795345051e-06, "loss": 0.1235, "num_input_tokens_seen": 54735152, "step": 28445 }, { "epoch": 4.641161595562444, "grad_norm": 0.049646906554698944, "learning_rate": 9.601415493813171e-06, "loss": 0.0004, "num_input_tokens_seen": 54744192, "step": 28450 }, { "epoch": 4.641977322783261, "grad_norm": 10.558459281921387, "learning_rate": 9.593682566936533e-06, "loss": 0.0928, "num_input_tokens_seen": 54753728, "step": 28455 }, { "epoch": 4.642793050004078, "grad_norm": 6.604983806610107, "learning_rate": 9.58595201590766e-06, "loss": 0.1213, "num_input_tokens_seen": 54761808, "step": 28460 }, { "epoch": 4.643608777224896, "grad_norm": 0.26986798644065857, "learning_rate": 9.578223841918681e-06, "loss": 0.0893, "num_input_tokens_seen": 54771936, "step": 28465 }, { "epoch": 4.6444245044457135, "grad_norm": 98.97413635253906, "learning_rate": 9.570498046161389e-06, "loss": 0.1654, "num_input_tokens_seen": 54782032, "step": 28470 }, { "epoch": 4.645240231666531, "grad_norm": 0.014331972226500511, "learning_rate": 9.562774629827206e-06, "loss": 0.0021, "num_input_tokens_seen": 54792112, "step": 28475 }, { "epoch": 4.646055958887348, "grad_norm": 0.005218111909925938, "learning_rate": 9.555053594107163e-06, "loss": 0.0817, "num_input_tokens_seen": 54802160, "step": 28480 }, { "epoch": 4.646871686108166, "grad_norm": 0.029786240309476852, "learning_rate": 9.547334940191957e-06, "loss": 0.1712, "num_input_tokens_seen": 54810352, "step": 28485 }, { "epoch": 4.647687413328983, "grad_norm": 0.2902251183986664, "learning_rate": 9.539618669271886e-06, "loss": 0.1437, "num_input_tokens_seen": 54819616, "step": 28490 }, { "epoch": 4.6485031405498, "grad_norm": 0.034235600382089615, "learning_rate": 9.531904782536904e-06, "loss": 0.0006, "num_input_tokens_seen": 54829600, "step": 28495 }, { "epoch": 4.649318867770617, "grad_norm": 0.20147447288036346, "learning_rate": 9.524193281176597e-06, "loss": 0.0581, "num_input_tokens_seen": 54839872, "step": 28500 }, { "epoch": 4.650134594991435, "grad_norm": 0.06900875270366669, "learning_rate": 9.516484166380165e-06, "loss": 0.0794, "num_input_tokens_seen": 54849568, "step": 28505 }, { "epoch": 4.650950322212252, "grad_norm": 0.0270688496530056, "learning_rate": 9.508777439336447e-06, "loss": 0.0005, "num_input_tokens_seen": 54860304, "step": 28510 }, { "epoch": 4.6517660494330695, "grad_norm": 0.02159814164042473, "learning_rate": 9.50107310123393e-06, "loss": 0.0006, "num_input_tokens_seen": 54870512, "step": 28515 }, { "epoch": 4.652581776653887, "grad_norm": 0.03803844377398491, "learning_rate": 9.493371153260702e-06, "loss": 0.0012, "num_input_tokens_seen": 54879856, "step": 28520 }, { "epoch": 4.653397503874705, "grad_norm": 0.09692366421222687, "learning_rate": 9.485671596604523e-06, "loss": 0.0923, "num_input_tokens_seen": 54888720, "step": 28525 }, { "epoch": 4.654213231095522, "grad_norm": 0.010651948861777782, "learning_rate": 9.477974432452738e-06, "loss": 0.066, "num_input_tokens_seen": 54898352, "step": 28530 }, { "epoch": 4.655028958316339, "grad_norm": 0.05008753389120102, "learning_rate": 9.470279661992356e-06, "loss": 0.0055, "num_input_tokens_seen": 54908288, "step": 28535 }, { "epoch": 4.655844685537156, "grad_norm": 0.04944637790322304, "learning_rate": 9.462587286410021e-06, "loss": 0.0007, "num_input_tokens_seen": 54917840, "step": 28540 }, { "epoch": 4.656660412757974, "grad_norm": 0.006949511356651783, "learning_rate": 9.454897306891972e-06, "loss": 0.0004, "num_input_tokens_seen": 54927968, "step": 28545 }, { "epoch": 4.657476139978791, "grad_norm": 0.02847341075539589, "learning_rate": 9.44720972462411e-06, "loss": 0.1412, "num_input_tokens_seen": 54938096, "step": 28550 }, { "epoch": 4.658291867199608, "grad_norm": 0.04616566747426987, "learning_rate": 9.439524540791964e-06, "loss": 0.2182, "num_input_tokens_seen": 54948256, "step": 28555 }, { "epoch": 4.659107594420425, "grad_norm": 0.19187518954277039, "learning_rate": 9.431841756580673e-06, "loss": 0.0477, "num_input_tokens_seen": 54957920, "step": 28560 }, { "epoch": 4.6599233216412435, "grad_norm": 0.009280417114496231, "learning_rate": 9.42416137317503e-06, "loss": 0.0577, "num_input_tokens_seen": 54967296, "step": 28565 }, { "epoch": 4.660739048862061, "grad_norm": 0.23973330855369568, "learning_rate": 9.416483391759437e-06, "loss": 0.0013, "num_input_tokens_seen": 54977376, "step": 28570 }, { "epoch": 4.661554776082878, "grad_norm": 0.03102610819041729, "learning_rate": 9.408807813517945e-06, "loss": 0.0007, "num_input_tokens_seen": 54986688, "step": 28575 }, { "epoch": 4.662370503303695, "grad_norm": 0.033251840621232986, "learning_rate": 9.401134639634221e-06, "loss": 0.1211, "num_input_tokens_seen": 54996000, "step": 28580 }, { "epoch": 4.663186230524513, "grad_norm": 0.06074964627623558, "learning_rate": 9.393463871291555e-06, "loss": 0.0751, "num_input_tokens_seen": 55006080, "step": 28585 }, { "epoch": 4.66400195774533, "grad_norm": 0.046129532158374786, "learning_rate": 9.385795509672881e-06, "loss": 0.0006, "num_input_tokens_seen": 55015440, "step": 28590 }, { "epoch": 4.664817684966147, "grad_norm": 0.02401670441031456, "learning_rate": 9.378129555960771e-06, "loss": 0.0012, "num_input_tokens_seen": 55025440, "step": 28595 }, { "epoch": 4.665633412186965, "grad_norm": 0.269252747297287, "learning_rate": 9.370466011337392e-06, "loss": 0.0013, "num_input_tokens_seen": 55035376, "step": 28600 }, { "epoch": 4.665633412186965, "eval_loss": 0.259128212928772, "eval_runtime": 80.7375, "eval_samples_per_second": 33.751, "eval_steps_per_second": 16.882, "num_input_tokens_seen": 55035376, "step": 28600 }, { "epoch": 4.666449139407782, "grad_norm": 0.03438712656497955, "learning_rate": 9.362804876984573e-06, "loss": 0.0008, "num_input_tokens_seen": 55046064, "step": 28605 }, { "epoch": 4.667264866628599, "grad_norm": 85.65486145019531, "learning_rate": 9.355146154083747e-06, "loss": 0.0474, "num_input_tokens_seen": 55056288, "step": 28610 }, { "epoch": 4.6680805938494165, "grad_norm": 1.8175673484802246, "learning_rate": 9.347489843815987e-06, "loss": 0.06, "num_input_tokens_seen": 55066032, "step": 28615 }, { "epoch": 4.668896321070234, "grad_norm": 0.3266720175743103, "learning_rate": 9.339835947362002e-06, "loss": 0.0016, "num_input_tokens_seen": 55075088, "step": 28620 }, { "epoch": 4.669712048291052, "grad_norm": 55.290767669677734, "learning_rate": 9.332184465902105e-06, "loss": 0.0492, "num_input_tokens_seen": 55084896, "step": 28625 }, { "epoch": 4.670527775511869, "grad_norm": 0.010591632686555386, "learning_rate": 9.324535400616266e-06, "loss": 0.0007, "num_input_tokens_seen": 55095040, "step": 28630 }, { "epoch": 4.671343502732686, "grad_norm": 0.025258971378207207, "learning_rate": 9.31688875268405e-06, "loss": 0.0009, "num_input_tokens_seen": 55104736, "step": 28635 }, { "epoch": 4.672159229953504, "grad_norm": 0.009966680780053139, "learning_rate": 9.309244523284674e-06, "loss": 0.0015, "num_input_tokens_seen": 55115184, "step": 28640 }, { "epoch": 4.672974957174321, "grad_norm": 0.034626077860593796, "learning_rate": 9.301602713596982e-06, "loss": 0.0518, "num_input_tokens_seen": 55124112, "step": 28645 }, { "epoch": 4.673790684395138, "grad_norm": 0.02018824778497219, "learning_rate": 9.293963324799432e-06, "loss": 0.0376, "num_input_tokens_seen": 55133568, "step": 28650 }, { "epoch": 4.674606411615955, "grad_norm": 0.01787395589053631, "learning_rate": 9.286326358070104e-06, "loss": 0.0004, "num_input_tokens_seen": 55142992, "step": 28655 }, { "epoch": 4.6754221388367725, "grad_norm": 0.0003034610999748111, "learning_rate": 9.278691814586729e-06, "loss": 0.0694, "num_input_tokens_seen": 55152400, "step": 28660 }, { "epoch": 4.6762378660575905, "grad_norm": 9.912178993225098, "learning_rate": 9.271059695526635e-06, "loss": 0.0106, "num_input_tokens_seen": 55162080, "step": 28665 }, { "epoch": 4.677053593278408, "grad_norm": 0.02315036579966545, "learning_rate": 9.263430002066805e-06, "loss": 0.0005, "num_input_tokens_seen": 55171104, "step": 28670 }, { "epoch": 4.677869320499225, "grad_norm": 0.033830676227808, "learning_rate": 9.25580273538382e-06, "loss": 0.0004, "num_input_tokens_seen": 55180544, "step": 28675 }, { "epoch": 4.678685047720043, "grad_norm": 0.10170482099056244, "learning_rate": 9.248177896653907e-06, "loss": 0.001, "num_input_tokens_seen": 55190032, "step": 28680 }, { "epoch": 4.67950077494086, "grad_norm": 1.8015334606170654, "learning_rate": 9.240555487052918e-06, "loss": 0.2217, "num_input_tokens_seen": 55199024, "step": 28685 }, { "epoch": 4.680316502161677, "grad_norm": 7.346368789672852, "learning_rate": 9.232935507756313e-06, "loss": 0.1517, "num_input_tokens_seen": 55208944, "step": 28690 }, { "epoch": 4.681132229382494, "grad_norm": 9.573346138000488, "learning_rate": 9.225317959939193e-06, "loss": 0.1509, "num_input_tokens_seen": 55218256, "step": 28695 }, { "epoch": 4.681947956603312, "grad_norm": 0.019017398357391357, "learning_rate": 9.217702844776287e-06, "loss": 0.037, "num_input_tokens_seen": 55226736, "step": 28700 }, { "epoch": 4.682763683824129, "grad_norm": 0.349160373210907, "learning_rate": 9.210090163441929e-06, "loss": 0.0825, "num_input_tokens_seen": 55234896, "step": 28705 }, { "epoch": 4.6835794110449465, "grad_norm": 0.06276968866586685, "learning_rate": 9.202479917110105e-06, "loss": 0.0013, "num_input_tokens_seen": 55244544, "step": 28710 }, { "epoch": 4.684395138265764, "grad_norm": 0.8126015067100525, "learning_rate": 9.194872106954392e-06, "loss": 0.0039, "num_input_tokens_seen": 55255136, "step": 28715 }, { "epoch": 4.685210865486582, "grad_norm": 0.34795519709587097, "learning_rate": 9.187266734148029e-06, "loss": 0.0011, "num_input_tokens_seen": 55264752, "step": 28720 }, { "epoch": 4.686026592707399, "grad_norm": 6.419296741485596, "learning_rate": 9.179663799863849e-06, "loss": 0.0495, "num_input_tokens_seen": 55274464, "step": 28725 }, { "epoch": 4.686842319928216, "grad_norm": 0.04036049172282219, "learning_rate": 9.172063305274317e-06, "loss": 0.0019, "num_input_tokens_seen": 55284880, "step": 28730 }, { "epoch": 4.687658047149033, "grad_norm": 0.28777292370796204, "learning_rate": 9.164465251551527e-06, "loss": 0.002, "num_input_tokens_seen": 55294800, "step": 28735 }, { "epoch": 4.688473774369851, "grad_norm": 0.12812559306621552, "learning_rate": 9.156869639867205e-06, "loss": 0.043, "num_input_tokens_seen": 55303280, "step": 28740 }, { "epoch": 4.689289501590668, "grad_norm": 0.033637337386608124, "learning_rate": 9.149276471392677e-06, "loss": 0.0005, "num_input_tokens_seen": 55312624, "step": 28745 }, { "epoch": 4.690105228811485, "grad_norm": 0.06139753386378288, "learning_rate": 9.141685747298914e-06, "loss": 0.001, "num_input_tokens_seen": 55322784, "step": 28750 }, { "epoch": 4.690920956032302, "grad_norm": 0.012414403259754181, "learning_rate": 9.13409746875649e-06, "loss": 0.0006, "num_input_tokens_seen": 55332960, "step": 28755 }, { "epoch": 4.69173668325312, "grad_norm": 0.007809093687683344, "learning_rate": 9.12651163693562e-06, "loss": 0.1309, "num_input_tokens_seen": 55342512, "step": 28760 }, { "epoch": 4.692552410473938, "grad_norm": 0.18114778399467468, "learning_rate": 9.11892825300614e-06, "loss": 0.0009, "num_input_tokens_seen": 55352016, "step": 28765 }, { "epoch": 4.693368137694755, "grad_norm": 0.020816074684262276, "learning_rate": 9.111347318137491e-06, "loss": 0.0006, "num_input_tokens_seen": 55362016, "step": 28770 }, { "epoch": 4.694183864915573, "grad_norm": 0.02948654629290104, "learning_rate": 9.103768833498755e-06, "loss": 0.0049, "num_input_tokens_seen": 55372592, "step": 28775 }, { "epoch": 4.69499959213639, "grad_norm": 0.01536643784493208, "learning_rate": 9.096192800258639e-06, "loss": 0.0007, "num_input_tokens_seen": 55382080, "step": 28780 }, { "epoch": 4.695815319357207, "grad_norm": 0.0029281838797032833, "learning_rate": 9.088619219585443e-06, "loss": 0.0003, "num_input_tokens_seen": 55392128, "step": 28785 }, { "epoch": 4.696631046578024, "grad_norm": 0.08683109283447266, "learning_rate": 9.081048092647127e-06, "loss": 0.0575, "num_input_tokens_seen": 55402176, "step": 28790 }, { "epoch": 4.697446773798841, "grad_norm": 0.08719871193170547, "learning_rate": 9.073479420611245e-06, "loss": 0.0009, "num_input_tokens_seen": 55411200, "step": 28795 }, { "epoch": 4.698262501019659, "grad_norm": 0.03341691195964813, "learning_rate": 9.065913204644974e-06, "loss": 0.0017, "num_input_tokens_seen": 55421296, "step": 28800 }, { "epoch": 4.698262501019659, "eval_loss": 0.2716529071331024, "eval_runtime": 80.8359, "eval_samples_per_second": 33.71, "eval_steps_per_second": 16.861, "num_input_tokens_seen": 55421296, "step": 28800 }, { "epoch": 4.699078228240476, "grad_norm": 0.0983612909913063, "learning_rate": 9.058349445915135e-06, "loss": 0.0019, "num_input_tokens_seen": 55430512, "step": 28805 }, { "epoch": 4.6998939554612935, "grad_norm": 0.027046553790569305, "learning_rate": 9.050788145588138e-06, "loss": 0.0007, "num_input_tokens_seen": 55440848, "step": 28810 }, { "epoch": 4.7007096826821115, "grad_norm": 0.009782582521438599, "learning_rate": 9.043229304830039e-06, "loss": 0.0006, "num_input_tokens_seen": 55450880, "step": 28815 }, { "epoch": 4.701525409902929, "grad_norm": 0.011152306571602821, "learning_rate": 9.035672924806515e-06, "loss": 0.0008, "num_input_tokens_seen": 55460880, "step": 28820 }, { "epoch": 4.702341137123746, "grad_norm": 0.017405563965439796, "learning_rate": 9.028119006682839e-06, "loss": 0.0165, "num_input_tokens_seen": 55469952, "step": 28825 }, { "epoch": 4.703156864344563, "grad_norm": 0.0191696397960186, "learning_rate": 9.020567551623935e-06, "loss": 0.0542, "num_input_tokens_seen": 55480256, "step": 28830 }, { "epoch": 4.70397259156538, "grad_norm": 0.026876812800765038, "learning_rate": 9.013018560794318e-06, "loss": 0.0004, "num_input_tokens_seen": 55489584, "step": 28835 }, { "epoch": 4.704788318786198, "grad_norm": 10.659164428710938, "learning_rate": 9.005472035358139e-06, "loss": 0.1458, "num_input_tokens_seen": 55499056, "step": 28840 }, { "epoch": 4.705604046007015, "grad_norm": 0.1415507048368454, "learning_rate": 8.997927976479185e-06, "loss": 0.0006, "num_input_tokens_seen": 55509264, "step": 28845 }, { "epoch": 4.706419773227832, "grad_norm": 0.023069972172379494, "learning_rate": 8.99038638532082e-06, "loss": 0.0008, "num_input_tokens_seen": 55518816, "step": 28850 }, { "epoch": 4.70723550044865, "grad_norm": 0.02178395912051201, "learning_rate": 8.982847263046065e-06, "loss": 0.0007, "num_input_tokens_seen": 55529024, "step": 28855 }, { "epoch": 4.7080512276694675, "grad_norm": 0.518253743648529, "learning_rate": 8.975310610817555e-06, "loss": 0.0008, "num_input_tokens_seen": 55538352, "step": 28860 }, { "epoch": 4.708866954890285, "grad_norm": 0.21255944669246674, "learning_rate": 8.967776429797528e-06, "loss": 0.0567, "num_input_tokens_seen": 55548016, "step": 28865 }, { "epoch": 4.709682682111102, "grad_norm": 0.6026161909103394, "learning_rate": 8.960244721147842e-06, "loss": 0.0009, "num_input_tokens_seen": 55559376, "step": 28870 }, { "epoch": 4.71049840933192, "grad_norm": 44.062129974365234, "learning_rate": 8.952715486029995e-06, "loss": 0.0699, "num_input_tokens_seen": 55568864, "step": 28875 }, { "epoch": 4.711314136552737, "grad_norm": 0.004613762721419334, "learning_rate": 8.945188725605075e-06, "loss": 0.1046, "num_input_tokens_seen": 55579360, "step": 28880 }, { "epoch": 4.712129863773554, "grad_norm": 0.01911233738064766, "learning_rate": 8.937664441033817e-06, "loss": 0.0006, "num_input_tokens_seen": 55588000, "step": 28885 }, { "epoch": 4.712945590994371, "grad_norm": 0.008965488523244858, "learning_rate": 8.930142633476549e-06, "loss": 0.0004, "num_input_tokens_seen": 55596576, "step": 28890 }, { "epoch": 4.713761318215189, "grad_norm": 9.603878021240234, "learning_rate": 8.92262330409323e-06, "loss": 0.0966, "num_input_tokens_seen": 55605856, "step": 28895 }, { "epoch": 4.714577045436006, "grad_norm": 0.08284248411655426, "learning_rate": 8.915106454043448e-06, "loss": 0.0359, "num_input_tokens_seen": 55615168, "step": 28900 }, { "epoch": 4.715392772656823, "grad_norm": 0.03334151953458786, "learning_rate": 8.90759208448638e-06, "loss": 0.0007, "num_input_tokens_seen": 55624816, "step": 28905 }, { "epoch": 4.716208499877641, "grad_norm": 0.001848289160989225, "learning_rate": 8.900080196580848e-06, "loss": 0.0003, "num_input_tokens_seen": 55633920, "step": 28910 }, { "epoch": 4.717024227098459, "grad_norm": 0.05591629445552826, "learning_rate": 8.892570791485267e-06, "loss": 0.0009, "num_input_tokens_seen": 55644912, "step": 28915 }, { "epoch": 4.717839954319276, "grad_norm": 0.04341546446084976, "learning_rate": 8.885063870357688e-06, "loss": 0.0007, "num_input_tokens_seen": 55653952, "step": 28920 }, { "epoch": 4.718655681540093, "grad_norm": 0.02483350783586502, "learning_rate": 8.87755943435578e-06, "loss": 0.0007, "num_input_tokens_seen": 55662800, "step": 28925 }, { "epoch": 4.71947140876091, "grad_norm": 0.8385888338088989, "learning_rate": 8.87005748463681e-06, "loss": 0.0025, "num_input_tokens_seen": 55672544, "step": 28930 }, { "epoch": 4.720287135981728, "grad_norm": 0.027877187356352806, "learning_rate": 8.862558022357681e-06, "loss": 0.0004, "num_input_tokens_seen": 55680560, "step": 28935 }, { "epoch": 4.721102863202545, "grad_norm": 0.020277835428714752, "learning_rate": 8.855061048674903e-06, "loss": 0.0008, "num_input_tokens_seen": 55689984, "step": 28940 }, { "epoch": 4.721918590423362, "grad_norm": 8.7194185256958, "learning_rate": 8.847566564744595e-06, "loss": 0.1053, "num_input_tokens_seen": 55698800, "step": 28945 }, { "epoch": 4.72273431764418, "grad_norm": 0.002894797595217824, "learning_rate": 8.840074571722512e-06, "loss": 0.0237, "num_input_tokens_seen": 55708512, "step": 28950 }, { "epoch": 4.723550044864997, "grad_norm": 0.015159351751208305, "learning_rate": 8.832585070764002e-06, "loss": 0.0006, "num_input_tokens_seen": 55719424, "step": 28955 }, { "epoch": 4.7243657720858145, "grad_norm": 0.16535009443759918, "learning_rate": 8.825098063024045e-06, "loss": 0.0007, "num_input_tokens_seen": 55729216, "step": 28960 }, { "epoch": 4.725181499306632, "grad_norm": 13.128049850463867, "learning_rate": 8.817613549657244e-06, "loss": 0.1673, "num_input_tokens_seen": 55738752, "step": 28965 }, { "epoch": 4.725997226527449, "grad_norm": 0.19567158818244934, "learning_rate": 8.810131531817783e-06, "loss": 0.0007, "num_input_tokens_seen": 55749232, "step": 28970 }, { "epoch": 4.726812953748267, "grad_norm": 0.08913015574216843, "learning_rate": 8.802652010659496e-06, "loss": 0.0006, "num_input_tokens_seen": 55758816, "step": 28975 }, { "epoch": 4.727628680969084, "grad_norm": 0.008128045126795769, "learning_rate": 8.795174987335827e-06, "loss": 0.0013, "num_input_tokens_seen": 55767728, "step": 28980 }, { "epoch": 4.728444408189901, "grad_norm": 0.02478039264678955, "learning_rate": 8.787700462999807e-06, "loss": 0.0024, "num_input_tokens_seen": 55778000, "step": 28985 }, { "epoch": 4.729260135410719, "grad_norm": 0.00515482435002923, "learning_rate": 8.780228438804122e-06, "loss": 0.0016, "num_input_tokens_seen": 55787664, "step": 28990 }, { "epoch": 4.730075862631536, "grad_norm": 0.13370366394519806, "learning_rate": 8.772758915901032e-06, "loss": 0.0905, "num_input_tokens_seen": 55797392, "step": 28995 }, { "epoch": 4.730891589852353, "grad_norm": 0.007124179974198341, "learning_rate": 8.765291895442443e-06, "loss": 0.0021, "num_input_tokens_seen": 55807776, "step": 29000 }, { "epoch": 4.730891589852353, "eval_loss": 0.31047356128692627, "eval_runtime": 80.7454, "eval_samples_per_second": 33.748, "eval_steps_per_second": 16.88, "num_input_tokens_seen": 55807776, "step": 29000 }, { "epoch": 4.7317073170731705, "grad_norm": 0.04904992878437042, "learning_rate": 8.75782737857987e-06, "loss": 0.0004, "num_input_tokens_seen": 55818000, "step": 29005 }, { "epoch": 4.732523044293988, "grad_norm": 0.0528557188808918, "learning_rate": 8.750365366464425e-06, "loss": 0.0006, "num_input_tokens_seen": 55827584, "step": 29010 }, { "epoch": 4.733338771514806, "grad_norm": 0.3885667622089386, "learning_rate": 8.742905860246838e-06, "loss": 0.0018, "num_input_tokens_seen": 55835744, "step": 29015 }, { "epoch": 4.734154498735623, "grad_norm": 0.009611825458705425, "learning_rate": 8.735448861077478e-06, "loss": 0.0009, "num_input_tokens_seen": 55845472, "step": 29020 }, { "epoch": 4.73497022595644, "grad_norm": 0.1568276286125183, "learning_rate": 8.727994370106288e-06, "loss": 0.0168, "num_input_tokens_seen": 55854848, "step": 29025 }, { "epoch": 4.735785953177258, "grad_norm": 0.05704958364367485, "learning_rate": 8.720542388482861e-06, "loss": 0.004, "num_input_tokens_seen": 55863984, "step": 29030 }, { "epoch": 4.736601680398075, "grad_norm": 0.029354475438594818, "learning_rate": 8.71309291735637e-06, "loss": 0.0107, "num_input_tokens_seen": 55872624, "step": 29035 }, { "epoch": 4.737417407618892, "grad_norm": 0.028034843504428864, "learning_rate": 8.705645957875621e-06, "loss": 0.1092, "num_input_tokens_seen": 55882576, "step": 29040 }, { "epoch": 4.738233134839709, "grad_norm": 0.2190214842557907, "learning_rate": 8.698201511189048e-06, "loss": 0.0009, "num_input_tokens_seen": 55892496, "step": 29045 }, { "epoch": 4.739048862060527, "grad_norm": 0.2966507077217102, "learning_rate": 8.690759578444649e-06, "loss": 0.0005, "num_input_tokens_seen": 55900656, "step": 29050 }, { "epoch": 4.7398645892813445, "grad_norm": 0.01810399815440178, "learning_rate": 8.68332016079008e-06, "loss": 0.0002, "num_input_tokens_seen": 55909824, "step": 29055 }, { "epoch": 4.740680316502162, "grad_norm": 0.0012309656012803316, "learning_rate": 8.6758832593726e-06, "loss": 0.0631, "num_input_tokens_seen": 55920576, "step": 29060 }, { "epoch": 4.741496043722979, "grad_norm": 0.041393790394067764, "learning_rate": 8.668448875339053e-06, "loss": 0.0303, "num_input_tokens_seen": 55929408, "step": 29065 }, { "epoch": 4.742311770943797, "grad_norm": 0.030745988711714745, "learning_rate": 8.661017009835933e-06, "loss": 0.0008, "num_input_tokens_seen": 55938928, "step": 29070 }, { "epoch": 4.743127498164614, "grad_norm": 26.729082107543945, "learning_rate": 8.653587664009311e-06, "loss": 0.0338, "num_input_tokens_seen": 55947312, "step": 29075 }, { "epoch": 4.743943225385431, "grad_norm": 0.07619114965200424, "learning_rate": 8.646160839004902e-06, "loss": 0.0008, "num_input_tokens_seen": 55956912, "step": 29080 }, { "epoch": 4.744758952606248, "grad_norm": 17.525814056396484, "learning_rate": 8.638736535967998e-06, "loss": 0.0836, "num_input_tokens_seen": 55966912, "step": 29085 }, { "epoch": 4.745574679827066, "grad_norm": 0.0030876286327838898, "learning_rate": 8.631314756043535e-06, "loss": 0.0005, "num_input_tokens_seen": 55977264, "step": 29090 }, { "epoch": 4.746390407047883, "grad_norm": 0.003545728512108326, "learning_rate": 8.62389550037603e-06, "loss": 0.0002, "num_input_tokens_seen": 55986000, "step": 29095 }, { "epoch": 4.7472061342687, "grad_norm": 0.02296288311481476, "learning_rate": 8.616478770109646e-06, "loss": 0.0002, "num_input_tokens_seen": 55995184, "step": 29100 }, { "epoch": 4.7480218614895175, "grad_norm": 0.1510000079870224, "learning_rate": 8.609064566388111e-06, "loss": 0.0004, "num_input_tokens_seen": 56005712, "step": 29105 }, { "epoch": 4.748837588710336, "grad_norm": 0.004878553561866283, "learning_rate": 8.601652890354815e-06, "loss": 0.1046, "num_input_tokens_seen": 56014992, "step": 29110 }, { "epoch": 4.749653315931153, "grad_norm": 0.25840044021606445, "learning_rate": 8.594243743152705e-06, "loss": 0.0006, "num_input_tokens_seen": 56024288, "step": 29115 }, { "epoch": 4.75046904315197, "grad_norm": 0.004449308849871159, "learning_rate": 8.58683712592438e-06, "loss": 0.0001, "num_input_tokens_seen": 56034880, "step": 29120 }, { "epoch": 4.751284770372787, "grad_norm": 0.008717807941138744, "learning_rate": 8.579433039812037e-06, "loss": 0.0411, "num_input_tokens_seen": 56044896, "step": 29125 }, { "epoch": 4.752100497593605, "grad_norm": 0.016124319285154343, "learning_rate": 8.572031485957466e-06, "loss": 0.0001, "num_input_tokens_seen": 56053792, "step": 29130 }, { "epoch": 4.752916224814422, "grad_norm": 90.76868438720703, "learning_rate": 8.564632465502084e-06, "loss": 0.2008, "num_input_tokens_seen": 56063344, "step": 29135 }, { "epoch": 4.753731952035239, "grad_norm": 0.03148235008120537, "learning_rate": 8.557235979586928e-06, "loss": 0.0013, "num_input_tokens_seen": 56073008, "step": 29140 }, { "epoch": 4.754547679256056, "grad_norm": 0.012854045256972313, "learning_rate": 8.549842029352606e-06, "loss": 0.0002, "num_input_tokens_seen": 56083952, "step": 29145 }, { "epoch": 4.755363406476874, "grad_norm": 0.022190285846590996, "learning_rate": 8.542450615939376e-06, "loss": 0.032, "num_input_tokens_seen": 56094928, "step": 29150 }, { "epoch": 4.7561791336976915, "grad_norm": 6.399202823638916, "learning_rate": 8.535061740487082e-06, "loss": 0.0783, "num_input_tokens_seen": 56104352, "step": 29155 }, { "epoch": 4.756994860918509, "grad_norm": 0.12659777700901031, "learning_rate": 8.527675404135168e-06, "loss": 0.0008, "num_input_tokens_seen": 56113712, "step": 29160 }, { "epoch": 4.757810588139327, "grad_norm": 0.020050736144185066, "learning_rate": 8.520291608022724e-06, "loss": 0.0001, "num_input_tokens_seen": 56121776, "step": 29165 }, { "epoch": 4.758626315360144, "grad_norm": 0.003042595461010933, "learning_rate": 8.512910353288398e-06, "loss": 0.055, "num_input_tokens_seen": 56131744, "step": 29170 }, { "epoch": 4.759442042580961, "grad_norm": 0.00048571417573839426, "learning_rate": 8.505531641070486e-06, "loss": 0.0207, "num_input_tokens_seen": 56142160, "step": 29175 }, { "epoch": 4.760257769801778, "grad_norm": 0.029999980702996254, "learning_rate": 8.498155472506885e-06, "loss": 0.0013, "num_input_tokens_seen": 56151488, "step": 29180 }, { "epoch": 4.761073497022595, "grad_norm": 0.04854559898376465, "learning_rate": 8.49078184873508e-06, "loss": 0.0135, "num_input_tokens_seen": 56161376, "step": 29185 }, { "epoch": 4.761889224243413, "grad_norm": 46.10480499267578, "learning_rate": 8.483410770892188e-06, "loss": 0.1196, "num_input_tokens_seen": 56170288, "step": 29190 }, { "epoch": 4.76270495146423, "grad_norm": 0.054263219237327576, "learning_rate": 8.476042240114909e-06, "loss": 0.091, "num_input_tokens_seen": 56180048, "step": 29195 }, { "epoch": 4.7635206786850475, "grad_norm": 0.0036651398986577988, "learning_rate": 8.468676257539568e-06, "loss": 0.0026, "num_input_tokens_seen": 56188960, "step": 29200 }, { "epoch": 4.7635206786850475, "eval_loss": 0.3253553509712219, "eval_runtime": 80.7562, "eval_samples_per_second": 33.744, "eval_steps_per_second": 16.878, "num_input_tokens_seen": 56188960, "step": 29200 }, { "epoch": 4.7643364059058655, "grad_norm": 0.020418278872966766, "learning_rate": 8.4613128243021e-06, "loss": 0.0007, "num_input_tokens_seen": 56198912, "step": 29205 }, { "epoch": 4.765152133126683, "grad_norm": 0.8531571626663208, "learning_rate": 8.453951941538028e-06, "loss": 0.126, "num_input_tokens_seen": 56207792, "step": 29210 }, { "epoch": 4.7659678603475, "grad_norm": 0.6241617202758789, "learning_rate": 8.446593610382495e-06, "loss": 0.0638, "num_input_tokens_seen": 56216784, "step": 29215 }, { "epoch": 4.766783587568317, "grad_norm": 0.01545191090553999, "learning_rate": 8.439237831970259e-06, "loss": 0.0007, "num_input_tokens_seen": 56226576, "step": 29220 }, { "epoch": 4.767599314789135, "grad_norm": 0.011823173612356186, "learning_rate": 8.431884607435667e-06, "loss": 0.1036, "num_input_tokens_seen": 56236032, "step": 29225 }, { "epoch": 4.768415042009952, "grad_norm": 0.4259433150291443, "learning_rate": 8.424533937912665e-06, "loss": 0.0006, "num_input_tokens_seen": 56245152, "step": 29230 }, { "epoch": 4.769230769230769, "grad_norm": 123.6908187866211, "learning_rate": 8.41718582453484e-06, "loss": 0.1592, "num_input_tokens_seen": 56255920, "step": 29235 }, { "epoch": 4.770046496451586, "grad_norm": 0.0033435076475143433, "learning_rate": 8.409840268435346e-06, "loss": 0.0007, "num_input_tokens_seen": 56264864, "step": 29240 }, { "epoch": 4.770862223672404, "grad_norm": 19.928808212280273, "learning_rate": 8.402497270746976e-06, "loss": 0.1003, "num_input_tokens_seen": 56274496, "step": 29245 }, { "epoch": 4.771677950893221, "grad_norm": 0.01566491834819317, "learning_rate": 8.395156832602095e-06, "loss": 0.0005, "num_input_tokens_seen": 56283328, "step": 29250 }, { "epoch": 4.772493678114039, "grad_norm": 0.0015666184481233358, "learning_rate": 8.387818955132707e-06, "loss": 0.0003, "num_input_tokens_seen": 56292992, "step": 29255 }, { "epoch": 4.773309405334856, "grad_norm": 0.08166109025478363, "learning_rate": 8.38048363947039e-06, "loss": 0.0007, "num_input_tokens_seen": 56301648, "step": 29260 }, { "epoch": 4.774125132555674, "grad_norm": 0.025661462917923927, "learning_rate": 8.373150886746351e-06, "loss": 0.0003, "num_input_tokens_seen": 56310560, "step": 29265 }, { "epoch": 4.774940859776491, "grad_norm": 15.516839981079102, "learning_rate": 8.365820698091397e-06, "loss": 0.0027, "num_input_tokens_seen": 56321632, "step": 29270 }, { "epoch": 4.775756586997308, "grad_norm": 0.017042865976691246, "learning_rate": 8.358493074635922e-06, "loss": 0.0005, "num_input_tokens_seen": 56331568, "step": 29275 }, { "epoch": 4.776572314218125, "grad_norm": 0.0015472343657165766, "learning_rate": 8.351168017509948e-06, "loss": 0.0195, "num_input_tokens_seen": 56342016, "step": 29280 }, { "epoch": 4.777388041438943, "grad_norm": 0.005922304932028055, "learning_rate": 8.343845527843094e-06, "loss": 0.0007, "num_input_tokens_seen": 56351584, "step": 29285 }, { "epoch": 4.77820376865976, "grad_norm": 0.025167757645249367, "learning_rate": 8.336525606764566e-06, "loss": 0.0837, "num_input_tokens_seen": 56360480, "step": 29290 }, { "epoch": 4.779019495880577, "grad_norm": 0.005387045908719301, "learning_rate": 8.329208255403204e-06, "loss": 0.0008, "num_input_tokens_seen": 56370160, "step": 29295 }, { "epoch": 4.7798352231013945, "grad_norm": 0.003966575488448143, "learning_rate": 8.321893474887426e-06, "loss": 0.0018, "num_input_tokens_seen": 56379488, "step": 29300 }, { "epoch": 4.7806509503222125, "grad_norm": 0.010913270525634289, "learning_rate": 8.31458126634526e-06, "loss": 0.0712, "num_input_tokens_seen": 56388800, "step": 29305 }, { "epoch": 4.78146667754303, "grad_norm": 0.014997359365224838, "learning_rate": 8.30727163090435e-06, "loss": 0.0513, "num_input_tokens_seen": 56398816, "step": 29310 }, { "epoch": 4.782282404763847, "grad_norm": 0.00017243092588614672, "learning_rate": 8.29996456969192e-06, "loss": 0.0481, "num_input_tokens_seen": 56409152, "step": 29315 }, { "epoch": 4.783098131984664, "grad_norm": 0.1713588386774063, "learning_rate": 8.292660083834818e-06, "loss": 0.0513, "num_input_tokens_seen": 56417632, "step": 29320 }, { "epoch": 4.783913859205482, "grad_norm": 0.08407799899578094, "learning_rate": 8.2853581744595e-06, "loss": 0.0596, "num_input_tokens_seen": 56426144, "step": 29325 }, { "epoch": 4.784729586426299, "grad_norm": 78.71282958984375, "learning_rate": 8.278058842691991e-06, "loss": 0.0792, "num_input_tokens_seen": 56436736, "step": 29330 }, { "epoch": 4.785545313647116, "grad_norm": 0.011736671440303326, "learning_rate": 8.27076208965796e-06, "loss": 0.0254, "num_input_tokens_seen": 56446208, "step": 29335 }, { "epoch": 4.786361040867934, "grad_norm": 15.012602806091309, "learning_rate": 8.263467916482637e-06, "loss": 0.1842, "num_input_tokens_seen": 56457056, "step": 29340 }, { "epoch": 4.787176768088751, "grad_norm": 11.406736373901367, "learning_rate": 8.256176324290885e-06, "loss": 0.0736, "num_input_tokens_seen": 56468032, "step": 29345 }, { "epoch": 4.7879924953095685, "grad_norm": 0.07155075669288635, "learning_rate": 8.248887314207168e-06, "loss": 0.1676, "num_input_tokens_seen": 56478768, "step": 29350 }, { "epoch": 4.788808222530386, "grad_norm": 0.007790207397192717, "learning_rate": 8.24160088735553e-06, "loss": 0.0901, "num_input_tokens_seen": 56488624, "step": 29355 }, { "epoch": 4.789623949751203, "grad_norm": 0.03544297441840172, "learning_rate": 8.234317044859629e-06, "loss": 0.0437, "num_input_tokens_seen": 56498832, "step": 29360 }, { "epoch": 4.790439676972021, "grad_norm": 0.4646582305431366, "learning_rate": 8.227035787842744e-06, "loss": 0.0158, "num_input_tokens_seen": 56507920, "step": 29365 }, { "epoch": 4.791255404192838, "grad_norm": 0.24693550169467926, "learning_rate": 8.219757117427721e-06, "loss": 0.0609, "num_input_tokens_seen": 56518464, "step": 29370 }, { "epoch": 4.792071131413655, "grad_norm": 0.20216777920722961, "learning_rate": 8.212481034737014e-06, "loss": 0.0014, "num_input_tokens_seen": 56527456, "step": 29375 }, { "epoch": 4.792886858634473, "grad_norm": 0.02828468568623066, "learning_rate": 8.205207540892707e-06, "loss": 0.0356, "num_input_tokens_seen": 56537280, "step": 29380 }, { "epoch": 4.79370258585529, "grad_norm": 0.07167916744947433, "learning_rate": 8.197936637016442e-06, "loss": 0.004, "num_input_tokens_seen": 56547168, "step": 29385 }, { "epoch": 4.794518313076107, "grad_norm": 0.021421857178211212, "learning_rate": 8.190668324229508e-06, "loss": 0.0028, "num_input_tokens_seen": 56557264, "step": 29390 }, { "epoch": 4.7953340402969244, "grad_norm": 0.04208027571439743, "learning_rate": 8.183402603652749e-06, "loss": 0.0003, "num_input_tokens_seen": 56568656, "step": 29395 }, { "epoch": 4.796149767517742, "grad_norm": 0.007369279861450195, "learning_rate": 8.176139476406635e-06, "loss": 0.0, "num_input_tokens_seen": 56576864, "step": 29400 }, { "epoch": 4.796149767517742, "eval_loss": 0.3105483651161194, "eval_runtime": 80.8614, "eval_samples_per_second": 33.7, "eval_steps_per_second": 16.856, "num_input_tokens_seen": 56576864, "step": 29400 }, { "epoch": 4.79696549473856, "grad_norm": 0.08120331913232803, "learning_rate": 8.16887894361125e-06, "loss": 0.0002, "num_input_tokens_seen": 56585440, "step": 29405 }, { "epoch": 4.797781221959377, "grad_norm": 2.864596366882324, "learning_rate": 8.161621006386233e-06, "loss": 0.0056, "num_input_tokens_seen": 56594992, "step": 29410 }, { "epoch": 4.798596949180194, "grad_norm": 0.0031943577341735363, "learning_rate": 8.154365665850869e-06, "loss": 0.1377, "num_input_tokens_seen": 56603232, "step": 29415 }, { "epoch": 4.799412676401012, "grad_norm": 0.008128971792757511, "learning_rate": 8.147112923124005e-06, "loss": 0.0779, "num_input_tokens_seen": 56613200, "step": 29420 }, { "epoch": 4.800228403621829, "grad_norm": 0.025436338037252426, "learning_rate": 8.13986277932412e-06, "loss": 0.0005, "num_input_tokens_seen": 56623504, "step": 29425 }, { "epoch": 4.801044130842646, "grad_norm": 0.006279785186052322, "learning_rate": 8.132615235569277e-06, "loss": 0.0002, "num_input_tokens_seen": 56632480, "step": 29430 }, { "epoch": 4.801859858063463, "grad_norm": 0.014893677085638046, "learning_rate": 8.125370292977124e-06, "loss": 0.1125, "num_input_tokens_seen": 56640848, "step": 29435 }, { "epoch": 4.802675585284281, "grad_norm": 12.21967887878418, "learning_rate": 8.118127952664944e-06, "loss": 0.0648, "num_input_tokens_seen": 56650496, "step": 29440 }, { "epoch": 4.803491312505098, "grad_norm": 0.00926548894494772, "learning_rate": 8.110888215749574e-06, "loss": 0.0004, "num_input_tokens_seen": 56659424, "step": 29445 }, { "epoch": 4.8043070397259156, "grad_norm": 0.035291705280542374, "learning_rate": 8.10365108334749e-06, "loss": 0.0003, "num_input_tokens_seen": 56669632, "step": 29450 }, { "epoch": 4.805122766946733, "grad_norm": 0.004358710255473852, "learning_rate": 8.096416556574743e-06, "loss": 0.0026, "num_input_tokens_seen": 56679872, "step": 29455 }, { "epoch": 4.805938494167551, "grad_norm": 0.061044979840517044, "learning_rate": 8.08918463654698e-06, "loss": 0.0778, "num_input_tokens_seen": 56688912, "step": 29460 }, { "epoch": 4.806754221388368, "grad_norm": 0.0156325064599514, "learning_rate": 8.081955324379458e-06, "loss": 0.0016, "num_input_tokens_seen": 56699184, "step": 29465 }, { "epoch": 4.807569948609185, "grad_norm": 0.029088687151670456, "learning_rate": 8.074728621187039e-06, "loss": 0.0001, "num_input_tokens_seen": 56710160, "step": 29470 }, { "epoch": 4.808385675830002, "grad_norm": 1.599493145942688, "learning_rate": 8.067504528084158e-06, "loss": 0.0803, "num_input_tokens_seen": 56719888, "step": 29475 }, { "epoch": 4.80920140305082, "grad_norm": 0.012869654223322868, "learning_rate": 8.060283046184861e-06, "loss": 0.085, "num_input_tokens_seen": 56729488, "step": 29480 }, { "epoch": 4.810017130271637, "grad_norm": 0.0015180150512605906, "learning_rate": 8.053064176602806e-06, "loss": 0.0006, "num_input_tokens_seen": 56739840, "step": 29485 }, { "epoch": 4.810832857492454, "grad_norm": 0.02481578104197979, "learning_rate": 8.045847920451216e-06, "loss": 0.0004, "num_input_tokens_seen": 56748816, "step": 29490 }, { "epoch": 4.8116485847132715, "grad_norm": 0.22711904346942902, "learning_rate": 8.038634278842944e-06, "loss": 0.0017, "num_input_tokens_seen": 56758736, "step": 29495 }, { "epoch": 4.8124643119340895, "grad_norm": 0.2297838032245636, "learning_rate": 8.031423252890408e-06, "loss": 0.0005, "num_input_tokens_seen": 56768720, "step": 29500 }, { "epoch": 4.813280039154907, "grad_norm": 0.057855457067489624, "learning_rate": 8.024214843705646e-06, "loss": 0.0005, "num_input_tokens_seen": 56778208, "step": 29505 }, { "epoch": 4.814095766375724, "grad_norm": 0.012439632788300514, "learning_rate": 8.017009052400295e-06, "loss": 0.0935, "num_input_tokens_seen": 56787744, "step": 29510 }, { "epoch": 4.814911493596542, "grad_norm": 0.02094334550201893, "learning_rate": 8.00980588008557e-06, "loss": 0.1589, "num_input_tokens_seen": 56798048, "step": 29515 }, { "epoch": 4.815727220817359, "grad_norm": 0.126938134431839, "learning_rate": 8.002605327872282e-06, "loss": 0.0322, "num_input_tokens_seen": 56806848, "step": 29520 }, { "epoch": 4.816542948038176, "grad_norm": 0.0035665780305862427, "learning_rate": 7.995407396870862e-06, "loss": 0.0018, "num_input_tokens_seen": 56816416, "step": 29525 }, { "epoch": 4.817358675258993, "grad_norm": 7.204583168029785, "learning_rate": 7.988212088191307e-06, "loss": 0.0041, "num_input_tokens_seen": 56825744, "step": 29530 }, { "epoch": 4.81817440247981, "grad_norm": 5.94457483291626, "learning_rate": 7.98101940294324e-06, "loss": 0.1069, "num_input_tokens_seen": 56836384, "step": 29535 }, { "epoch": 4.818990129700628, "grad_norm": 0.06332418322563171, "learning_rate": 7.973829342235847e-06, "loss": 0.0003, "num_input_tokens_seen": 56846096, "step": 29540 }, { "epoch": 4.8198058569214455, "grad_norm": 0.012492476962506771, "learning_rate": 7.966641907177936e-06, "loss": 0.0006, "num_input_tokens_seen": 56855312, "step": 29545 }, { "epoch": 4.820621584142263, "grad_norm": 0.007072919048368931, "learning_rate": 7.959457098877901e-06, "loss": 0.0002, "num_input_tokens_seen": 56864416, "step": 29550 }, { "epoch": 4.821437311363081, "grad_norm": 0.02890246920287609, "learning_rate": 7.952274918443719e-06, "loss": 0.0062, "num_input_tokens_seen": 56873888, "step": 29555 }, { "epoch": 4.822253038583898, "grad_norm": 0.01757948286831379, "learning_rate": 7.945095366982983e-06, "loss": 0.0002, "num_input_tokens_seen": 56883696, "step": 29560 }, { "epoch": 4.823068765804715, "grad_norm": 0.0575508251786232, "learning_rate": 7.937918445602871e-06, "loss": 0.1003, "num_input_tokens_seen": 56893120, "step": 29565 }, { "epoch": 4.823884493025532, "grad_norm": 9.754446029663086, "learning_rate": 7.930744155410145e-06, "loss": 0.1932, "num_input_tokens_seen": 56901952, "step": 29570 }, { "epoch": 4.824700220246349, "grad_norm": 0.03639482334256172, "learning_rate": 7.923572497511181e-06, "loss": 0.0008, "num_input_tokens_seen": 56911632, "step": 29575 }, { "epoch": 4.825515947467167, "grad_norm": 4.320417881011963, "learning_rate": 7.916403473011927e-06, "loss": 0.0191, "num_input_tokens_seen": 56921072, "step": 29580 }, { "epoch": 4.826331674687984, "grad_norm": 0.01678074337542057, "learning_rate": 7.909237083017953e-06, "loss": 0.0044, "num_input_tokens_seen": 56930272, "step": 29585 }, { "epoch": 4.827147401908801, "grad_norm": 0.052319157868623734, "learning_rate": 7.902073328634389e-06, "loss": 0.0683, "num_input_tokens_seen": 56940416, "step": 29590 }, { "epoch": 4.8279631291296194, "grad_norm": 0.008065586909651756, "learning_rate": 7.894912210965987e-06, "loss": 0.0018, "num_input_tokens_seen": 56949584, "step": 29595 }, { "epoch": 4.828778856350437, "grad_norm": 0.048863429576158524, "learning_rate": 7.887753731117075e-06, "loss": 0.0004, "num_input_tokens_seen": 56959888, "step": 29600 }, { "epoch": 4.828778856350437, "eval_loss": 0.29312625527381897, "eval_runtime": 80.8646, "eval_samples_per_second": 33.698, "eval_steps_per_second": 16.855, "num_input_tokens_seen": 56959888, "step": 29600 }, { "epoch": 4.829594583571254, "grad_norm": 0.015248341485857964, "learning_rate": 7.880597890191587e-06, "loss": 0.0001, "num_input_tokens_seen": 56968656, "step": 29605 }, { "epoch": 4.830410310792071, "grad_norm": 0.0647365152835846, "learning_rate": 7.873444689293036e-06, "loss": 0.0006, "num_input_tokens_seen": 56978640, "step": 29610 }, { "epoch": 4.831226038012889, "grad_norm": 0.08112752437591553, "learning_rate": 7.866294129524548e-06, "loss": 0.1001, "num_input_tokens_seen": 56987712, "step": 29615 }, { "epoch": 4.832041765233706, "grad_norm": 28.979320526123047, "learning_rate": 7.859146211988811e-06, "loss": 0.0899, "num_input_tokens_seen": 56997392, "step": 29620 }, { "epoch": 4.832857492454523, "grad_norm": 0.04024634137749672, "learning_rate": 7.852000937788134e-06, "loss": 0.0011, "num_input_tokens_seen": 57008048, "step": 29625 }, { "epoch": 4.83367321967534, "grad_norm": 0.06969992071390152, "learning_rate": 7.844858308024416e-06, "loss": 0.0014, "num_input_tokens_seen": 57018064, "step": 29630 }, { "epoch": 4.834488946896158, "grad_norm": 1.6985028982162476, "learning_rate": 7.837718323799122e-06, "loss": 0.0011, "num_input_tokens_seen": 57027248, "step": 29635 }, { "epoch": 4.835304674116975, "grad_norm": 0.04491100087761879, "learning_rate": 7.83058098621334e-06, "loss": 0.0007, "num_input_tokens_seen": 57036032, "step": 29640 }, { "epoch": 4.8361204013377925, "grad_norm": 0.010894042439758778, "learning_rate": 7.823446296367739e-06, "loss": 0.0012, "num_input_tokens_seen": 57046720, "step": 29645 }, { "epoch": 4.83693612855861, "grad_norm": 0.009290440008044243, "learning_rate": 7.81631425536257e-06, "loss": 0.0472, "num_input_tokens_seen": 57057200, "step": 29650 }, { "epoch": 4.837751855779428, "grad_norm": 0.07362882047891617, "learning_rate": 7.809184864297689e-06, "loss": 0.0477, "num_input_tokens_seen": 57067168, "step": 29655 }, { "epoch": 4.838567583000245, "grad_norm": 32.15049362182617, "learning_rate": 7.802058124272532e-06, "loss": 0.0272, "num_input_tokens_seen": 57075584, "step": 29660 }, { "epoch": 4.839383310221062, "grad_norm": 0.007652977481484413, "learning_rate": 7.79493403638614e-06, "loss": 0.0001, "num_input_tokens_seen": 57086032, "step": 29665 }, { "epoch": 4.840199037441879, "grad_norm": 0.05592644587159157, "learning_rate": 7.787812601737132e-06, "loss": 0.0028, "num_input_tokens_seen": 57095824, "step": 29670 }, { "epoch": 4.841014764662697, "grad_norm": 0.04916062578558922, "learning_rate": 7.780693821423715e-06, "loss": 0.0005, "num_input_tokens_seen": 57105936, "step": 29675 }, { "epoch": 4.841830491883514, "grad_norm": 0.04308092221617699, "learning_rate": 7.773577696543705e-06, "loss": 0.0288, "num_input_tokens_seen": 57117280, "step": 29680 }, { "epoch": 4.842646219104331, "grad_norm": 0.03800622746348381, "learning_rate": 7.7664642281945e-06, "loss": 0.0002, "num_input_tokens_seen": 57127488, "step": 29685 }, { "epoch": 4.8434619463251485, "grad_norm": 0.004517769906669855, "learning_rate": 7.759353417473072e-06, "loss": 0.1044, "num_input_tokens_seen": 57137760, "step": 29690 }, { "epoch": 4.8442776735459665, "grad_norm": 0.09173062443733215, "learning_rate": 7.752245265476016e-06, "loss": 0.0012, "num_input_tokens_seen": 57148432, "step": 29695 }, { "epoch": 4.845093400766784, "grad_norm": 0.02735959365963936, "learning_rate": 7.745139773299481e-06, "loss": 0.0003, "num_input_tokens_seen": 57158352, "step": 29700 }, { "epoch": 4.845909127987601, "grad_norm": 0.010929577052593231, "learning_rate": 7.738036942039232e-06, "loss": 0.0037, "num_input_tokens_seen": 57167968, "step": 29705 }, { "epoch": 4.846724855208418, "grad_norm": 0.0028815127443522215, "learning_rate": 7.73093677279062e-06, "loss": 0.0749, "num_input_tokens_seen": 57177536, "step": 29710 }, { "epoch": 4.847540582429236, "grad_norm": 0.523450493812561, "learning_rate": 7.72383926664857e-06, "loss": 0.001, "num_input_tokens_seen": 57187344, "step": 29715 }, { "epoch": 4.848356309650053, "grad_norm": 0.0550629198551178, "learning_rate": 7.716744424707606e-06, "loss": 0.0008, "num_input_tokens_seen": 57196320, "step": 29720 }, { "epoch": 4.84917203687087, "grad_norm": 0.009666386991739273, "learning_rate": 7.709652248061858e-06, "loss": 0.0011, "num_input_tokens_seen": 57205472, "step": 29725 }, { "epoch": 4.849987764091688, "grad_norm": 0.01644756644964218, "learning_rate": 7.702562737805017e-06, "loss": 0.0007, "num_input_tokens_seen": 57215984, "step": 29730 }, { "epoch": 4.850803491312505, "grad_norm": 0.012515813112258911, "learning_rate": 7.695475895030365e-06, "loss": 0.0357, "num_input_tokens_seen": 57224288, "step": 29735 }, { "epoch": 4.8516192185333225, "grad_norm": 0.43464529514312744, "learning_rate": 7.6883917208308e-06, "loss": 0.0012, "num_input_tokens_seen": 57233248, "step": 29740 }, { "epoch": 4.85243494575414, "grad_norm": 0.00815972127020359, "learning_rate": 7.681310216298778e-06, "loss": 0.0004, "num_input_tokens_seen": 57243456, "step": 29745 }, { "epoch": 4.853250672974957, "grad_norm": 0.14130081236362457, "learning_rate": 7.674231382526367e-06, "loss": 0.0004, "num_input_tokens_seen": 57253408, "step": 29750 }, { "epoch": 4.854066400195775, "grad_norm": 0.0014583980664610863, "learning_rate": 7.667155220605198e-06, "loss": 0.0021, "num_input_tokens_seen": 57263072, "step": 29755 }, { "epoch": 4.854882127416592, "grad_norm": 0.007433898281306028, "learning_rate": 7.660081731626515e-06, "loss": 0.1532, "num_input_tokens_seen": 57273552, "step": 29760 }, { "epoch": 4.855697854637409, "grad_norm": 0.016502344980835915, "learning_rate": 7.653010916681141e-06, "loss": 0.0128, "num_input_tokens_seen": 57282752, "step": 29765 }, { "epoch": 4.856513581858227, "grad_norm": 0.01138320378959179, "learning_rate": 7.645942776859472e-06, "loss": 0.0703, "num_input_tokens_seen": 57291872, "step": 29770 }, { "epoch": 4.857329309079044, "grad_norm": 0.003510087262839079, "learning_rate": 7.63887731325152e-06, "loss": 0.0004, "num_input_tokens_seen": 57300496, "step": 29775 }, { "epoch": 4.858145036299861, "grad_norm": 0.020087111741304398, "learning_rate": 7.63181452694685e-06, "loss": 0.0006, "num_input_tokens_seen": 57309616, "step": 29780 }, { "epoch": 4.858960763520678, "grad_norm": 0.0804608091711998, "learning_rate": 7.624754419034644e-06, "loss": 0.0018, "num_input_tokens_seen": 57318752, "step": 29785 }, { "epoch": 4.859776490741496, "grad_norm": 0.09834223240613937, "learning_rate": 7.6176969906036645e-06, "loss": 0.0003, "num_input_tokens_seen": 57328608, "step": 29790 }, { "epoch": 4.8605922179623136, "grad_norm": 0.002311103045940399, "learning_rate": 7.610642242742242e-06, "loss": 0.1093, "num_input_tokens_seen": 57337552, "step": 29795 }, { "epoch": 4.861407945183131, "grad_norm": 0.20928291976451874, "learning_rate": 7.603590176538322e-06, "loss": 0.002, "num_input_tokens_seen": 57347776, "step": 29800 }, { "epoch": 4.861407945183131, "eval_loss": 0.3229916989803314, "eval_runtime": 80.8301, "eval_samples_per_second": 33.713, "eval_steps_per_second": 16.863, "num_input_tokens_seen": 57347776, "step": 29800 }, { "epoch": 4.862223672403948, "grad_norm": 6.289823055267334, "learning_rate": 7.596540793079404e-06, "loss": 0.0259, "num_input_tokens_seen": 57358224, "step": 29805 }, { "epoch": 4.863039399624766, "grad_norm": 0.010515241883695126, "learning_rate": 7.5894940934526125e-06, "loss": 0.0002, "num_input_tokens_seen": 57366768, "step": 29810 }, { "epoch": 4.863855126845583, "grad_norm": 22.3367919921875, "learning_rate": 7.582450078744621e-06, "loss": 0.0927, "num_input_tokens_seen": 57377024, "step": 29815 }, { "epoch": 4.8646708540664, "grad_norm": 0.0017649008659645915, "learning_rate": 7.575408750041707e-06, "loss": 0.0899, "num_input_tokens_seen": 57386976, "step": 29820 }, { "epoch": 4.865486581287217, "grad_norm": 0.007401398383080959, "learning_rate": 7.568370108429732e-06, "loss": 0.0002, "num_input_tokens_seen": 57396480, "step": 29825 }, { "epoch": 4.866302308508035, "grad_norm": 0.03225269913673401, "learning_rate": 7.561334154994154e-06, "loss": 0.0471, "num_input_tokens_seen": 57407072, "step": 29830 }, { "epoch": 4.867118035728852, "grad_norm": 0.024082940071821213, "learning_rate": 7.55430089081999e-06, "loss": 0.0879, "num_input_tokens_seen": 57416768, "step": 29835 }, { "epoch": 4.8679337629496695, "grad_norm": 0.17066335678100586, "learning_rate": 7.547270316991864e-06, "loss": 0.0693, "num_input_tokens_seen": 57426256, "step": 29840 }, { "epoch": 4.868749490170487, "grad_norm": 0.005452870391309261, "learning_rate": 7.5402424345939884e-06, "loss": 0.0152, "num_input_tokens_seen": 57436480, "step": 29845 }, { "epoch": 4.869565217391305, "grad_norm": 0.47496742010116577, "learning_rate": 7.533217244710133e-06, "loss": 0.0843, "num_input_tokens_seen": 57445200, "step": 29850 }, { "epoch": 4.870380944612122, "grad_norm": 0.07944869995117188, "learning_rate": 7.52619474842369e-06, "loss": 0.0006, "num_input_tokens_seen": 57454704, "step": 29855 }, { "epoch": 4.871196671832939, "grad_norm": 0.008232678286731243, "learning_rate": 7.519174946817597e-06, "loss": 0.0002, "num_input_tokens_seen": 57464864, "step": 29860 }, { "epoch": 4.872012399053756, "grad_norm": 0.035411395132541656, "learning_rate": 7.512157840974407e-06, "loss": 0.0002, "num_input_tokens_seen": 57474880, "step": 29865 }, { "epoch": 4.872828126274574, "grad_norm": 0.02575721964240074, "learning_rate": 7.5051434319762496e-06, "loss": 0.0006, "num_input_tokens_seen": 57483408, "step": 29870 }, { "epoch": 4.873643853495391, "grad_norm": 0.14267893135547638, "learning_rate": 7.498131720904822e-06, "loss": 0.0942, "num_input_tokens_seen": 57492800, "step": 29875 }, { "epoch": 4.874459580716208, "grad_norm": 0.007001079153269529, "learning_rate": 7.491122708841433e-06, "loss": 0.1977, "num_input_tokens_seen": 57502720, "step": 29880 }, { "epoch": 4.8752753079370255, "grad_norm": 0.06415156275033951, "learning_rate": 7.4841163968669524e-06, "loss": 0.0005, "num_input_tokens_seen": 57512272, "step": 29885 }, { "epoch": 4.8760910351578435, "grad_norm": 10.003375053405762, "learning_rate": 7.4771127860618355e-06, "loss": 0.1864, "num_input_tokens_seen": 57520496, "step": 29890 }, { "epoch": 4.876906762378661, "grad_norm": 0.07899028807878494, "learning_rate": 7.470111877506139e-06, "loss": 0.0008, "num_input_tokens_seen": 57528912, "step": 29895 }, { "epoch": 4.877722489599478, "grad_norm": 0.0712471604347229, "learning_rate": 7.463113672279479e-06, "loss": 0.0008, "num_input_tokens_seen": 57538048, "step": 29900 }, { "epoch": 4.878538216820296, "grad_norm": 0.027211831882596016, "learning_rate": 7.456118171461071e-06, "loss": 0.0009, "num_input_tokens_seen": 57547104, "step": 29905 }, { "epoch": 4.879353944041113, "grad_norm": 0.47711458802223206, "learning_rate": 7.449125376129721e-06, "loss": 0.0009, "num_input_tokens_seen": 57556352, "step": 29910 }, { "epoch": 4.88016967126193, "grad_norm": 0.07494336366653442, "learning_rate": 7.442135287363788e-06, "loss": 0.0003, "num_input_tokens_seen": 57566032, "step": 29915 }, { "epoch": 4.880985398482747, "grad_norm": 0.004289598204195499, "learning_rate": 7.435147906241247e-06, "loss": 0.0002, "num_input_tokens_seen": 57576320, "step": 29920 }, { "epoch": 4.881801125703564, "grad_norm": 0.24137970805168152, "learning_rate": 7.428163233839624e-06, "loss": 0.0009, "num_input_tokens_seen": 57584960, "step": 29925 }, { "epoch": 4.882616852924382, "grad_norm": 0.015219365246593952, "learning_rate": 7.4211812712360525e-06, "loss": 0.0701, "num_input_tokens_seen": 57594624, "step": 29930 }, { "epoch": 4.883432580145199, "grad_norm": 23.56155776977539, "learning_rate": 7.4142020195072464e-06, "loss": 0.1055, "num_input_tokens_seen": 57603424, "step": 29935 }, { "epoch": 4.884248307366017, "grad_norm": 14.840946197509766, "learning_rate": 7.407225479729479e-06, "loss": 0.0067, "num_input_tokens_seen": 57612944, "step": 29940 }, { "epoch": 4.885064034586835, "grad_norm": 0.014012876898050308, "learning_rate": 7.400251652978632e-06, "loss": 0.0002, "num_input_tokens_seen": 57623024, "step": 29945 }, { "epoch": 4.885879761807652, "grad_norm": 0.23813122510910034, "learning_rate": 7.393280540330147e-06, "loss": 0.0335, "num_input_tokens_seen": 57633120, "step": 29950 }, { "epoch": 4.886695489028469, "grad_norm": 1.6561859846115112, "learning_rate": 7.386312142859069e-06, "loss": 0.0044, "num_input_tokens_seen": 57641552, "step": 29955 }, { "epoch": 4.887511216249286, "grad_norm": 0.007501145359128714, "learning_rate": 7.379346461640008e-06, "loss": 0.0027, "num_input_tokens_seen": 57651888, "step": 29960 }, { "epoch": 4.888326943470103, "grad_norm": 10.022849082946777, "learning_rate": 7.372383497747149e-06, "loss": 0.0822, "num_input_tokens_seen": 57660416, "step": 29965 }, { "epoch": 4.889142670690921, "grad_norm": 0.027278967201709747, "learning_rate": 7.3654232522542775e-06, "loss": 0.0004, "num_input_tokens_seen": 57669904, "step": 29970 }, { "epoch": 4.889958397911738, "grad_norm": 0.007543955929577351, "learning_rate": 7.358465726234756e-06, "loss": 0.0325, "num_input_tokens_seen": 57678800, "step": 29975 }, { "epoch": 4.890774125132555, "grad_norm": 0.0028285353910177946, "learning_rate": 7.351510920761512e-06, "loss": 0.0003, "num_input_tokens_seen": 57688768, "step": 29980 }, { "epoch": 4.891589852353373, "grad_norm": 0.02037070132791996, "learning_rate": 7.344558836907067e-06, "loss": 0.0004, "num_input_tokens_seen": 57698960, "step": 29985 }, { "epoch": 4.8924055795741905, "grad_norm": 0.011889494024217129, "learning_rate": 7.3376094757435285e-06, "loss": 0.0005, "num_input_tokens_seen": 57708896, "step": 29990 }, { "epoch": 4.893221306795008, "grad_norm": 0.7071089744567871, "learning_rate": 7.330662838342561e-06, "loss": 0.0099, "num_input_tokens_seen": 57718464, "step": 29995 }, { "epoch": 4.894037034015825, "grad_norm": 0.000754126813262701, "learning_rate": 7.323718925775438e-06, "loss": 0.0001, "num_input_tokens_seen": 57727072, "step": 30000 }, { "epoch": 4.894037034015825, "eval_loss": 0.31458866596221924, "eval_runtime": 80.9175, "eval_samples_per_second": 33.676, "eval_steps_per_second": 16.844, "num_input_tokens_seen": 57727072, "step": 30000 }, { "epoch": 4.894852761236643, "grad_norm": 0.0017322037601843476, "learning_rate": 7.316777739112985e-06, "loss": 0.1091, "num_input_tokens_seen": 57734992, "step": 30005 }, { "epoch": 4.89566848845746, "grad_norm": 0.03434410318732262, "learning_rate": 7.309839279425626e-06, "loss": 0.002, "num_input_tokens_seen": 57745408, "step": 30010 }, { "epoch": 4.896484215678277, "grad_norm": 0.02855721116065979, "learning_rate": 7.302903547783366e-06, "loss": 0.0004, "num_input_tokens_seen": 57755888, "step": 30015 }, { "epoch": 4.897299942899094, "grad_norm": 0.045346569269895554, "learning_rate": 7.2959705452557644e-06, "loss": 0.0871, "num_input_tokens_seen": 57766992, "step": 30020 }, { "epoch": 4.898115670119912, "grad_norm": 0.029444752261042595, "learning_rate": 7.289040272911996e-06, "loss": 0.0005, "num_input_tokens_seen": 57777712, "step": 30025 }, { "epoch": 4.898931397340729, "grad_norm": 0.045236628502607346, "learning_rate": 7.282112731820789e-06, "loss": 0.0833, "num_input_tokens_seen": 57788448, "step": 30030 }, { "epoch": 4.8997471245615465, "grad_norm": 0.031414877623319626, "learning_rate": 7.275187923050447e-06, "loss": 0.0002, "num_input_tokens_seen": 57798192, "step": 30035 }, { "epoch": 4.900562851782364, "grad_norm": 0.004227069206535816, "learning_rate": 7.268265847668879e-06, "loss": 0.0008, "num_input_tokens_seen": 57808400, "step": 30040 }, { "epoch": 4.901378579003182, "grad_norm": 0.06504824012517929, "learning_rate": 7.261346506743538e-06, "loss": 0.0143, "num_input_tokens_seen": 57818720, "step": 30045 }, { "epoch": 4.902194306223999, "grad_norm": 0.012660208158195019, "learning_rate": 7.254429901341486e-06, "loss": 0.1486, "num_input_tokens_seen": 57828032, "step": 30050 }, { "epoch": 4.903010033444816, "grad_norm": 0.5867745876312256, "learning_rate": 7.247516032529356e-06, "loss": 0.011, "num_input_tokens_seen": 57836992, "step": 30055 }, { "epoch": 4.903825760665633, "grad_norm": 0.018725300207734108, "learning_rate": 7.240604901373338e-06, "loss": 0.0009, "num_input_tokens_seen": 57846080, "step": 30060 }, { "epoch": 4.904641487886451, "grad_norm": 0.04146686568856239, "learning_rate": 7.233696508939223e-06, "loss": 0.027, "num_input_tokens_seen": 57855776, "step": 30065 }, { "epoch": 4.905457215107268, "grad_norm": 0.02956628054380417, "learning_rate": 7.226790856292376e-06, "loss": 0.0006, "num_input_tokens_seen": 57865872, "step": 30070 }, { "epoch": 4.906272942328085, "grad_norm": 0.014584064483642578, "learning_rate": 7.219887944497727e-06, "loss": 0.0006, "num_input_tokens_seen": 57875728, "step": 30075 }, { "epoch": 4.907088669548903, "grad_norm": 0.02280486561357975, "learning_rate": 7.2129877746198e-06, "loss": 0.0002, "num_input_tokens_seen": 57884688, "step": 30080 }, { "epoch": 4.9079043967697205, "grad_norm": 0.0007078213384374976, "learning_rate": 7.20609034772268e-06, "loss": 0.0821, "num_input_tokens_seen": 57894944, "step": 30085 }, { "epoch": 4.908720123990538, "grad_norm": 0.0005087494500912726, "learning_rate": 7.19919566487004e-06, "loss": 0.1919, "num_input_tokens_seen": 57904448, "step": 30090 }, { "epoch": 4.909535851211355, "grad_norm": 13.423680305480957, "learning_rate": 7.192303727125132e-06, "loss": 0.0031, "num_input_tokens_seen": 57914096, "step": 30095 }, { "epoch": 4.910351578432172, "grad_norm": 0.045143112540245056, "learning_rate": 7.185414535550777e-06, "loss": 0.0461, "num_input_tokens_seen": 57924352, "step": 30100 }, { "epoch": 4.91116730565299, "grad_norm": 21.370059967041016, "learning_rate": 7.178528091209363e-06, "loss": 0.0715, "num_input_tokens_seen": 57935024, "step": 30105 }, { "epoch": 4.911983032873807, "grad_norm": 0.7813360691070557, "learning_rate": 7.171644395162888e-06, "loss": 0.0016, "num_input_tokens_seen": 57945696, "step": 30110 }, { "epoch": 4.912798760094624, "grad_norm": 0.009014819748699665, "learning_rate": 7.164763448472881e-06, "loss": 0.0544, "num_input_tokens_seen": 57955984, "step": 30115 }, { "epoch": 4.913614487315442, "grad_norm": 0.006391158327460289, "learning_rate": 7.157885252200491e-06, "loss": 0.0005, "num_input_tokens_seen": 57965856, "step": 30120 }, { "epoch": 4.914430214536259, "grad_norm": 13.62974739074707, "learning_rate": 7.151009807406403e-06, "loss": 0.1112, "num_input_tokens_seen": 57975328, "step": 30125 }, { "epoch": 4.915245941757076, "grad_norm": 7.233188152313232, "learning_rate": 7.144137115150909e-06, "loss": 0.0115, "num_input_tokens_seen": 57986352, "step": 30130 }, { "epoch": 4.9160616689778935, "grad_norm": 8.882387161254883, "learning_rate": 7.1372671764938725e-06, "loss": 0.0108, "num_input_tokens_seen": 57997008, "step": 30135 }, { "epoch": 4.916877396198711, "grad_norm": 0.2165774255990982, "learning_rate": 7.130399992494705e-06, "loss": 0.0493, "num_input_tokens_seen": 58005376, "step": 30140 }, { "epoch": 4.917693123419529, "grad_norm": 0.04220229387283325, "learning_rate": 7.123535564212419e-06, "loss": 0.0012, "num_input_tokens_seen": 58015488, "step": 30145 }, { "epoch": 4.918508850640346, "grad_norm": 0.0045854924246668816, "learning_rate": 7.116673892705611e-06, "loss": 0.14, "num_input_tokens_seen": 58025536, "step": 30150 }, { "epoch": 4.919324577861163, "grad_norm": 0.09481789916753769, "learning_rate": 7.109814979032415e-06, "loss": 0.0361, "num_input_tokens_seen": 58035760, "step": 30155 }, { "epoch": 4.920140305081981, "grad_norm": 0.1104186549782753, "learning_rate": 7.102958824250577e-06, "loss": 0.2172, "num_input_tokens_seen": 58044944, "step": 30160 }, { "epoch": 4.920956032302798, "grad_norm": 0.02952788583934307, "learning_rate": 7.096105429417393e-06, "loss": 0.0471, "num_input_tokens_seen": 58054672, "step": 30165 }, { "epoch": 4.921771759523615, "grad_norm": 0.07203978300094604, "learning_rate": 7.0892547955897506e-06, "loss": 0.0011, "num_input_tokens_seen": 58063248, "step": 30170 }, { "epoch": 4.922587486744432, "grad_norm": 0.2615008056163788, "learning_rate": 7.0824069238241e-06, "loss": 0.0044, "num_input_tokens_seen": 58073168, "step": 30175 }, { "epoch": 4.92340321396525, "grad_norm": 0.0038685959298163652, "learning_rate": 7.075561815176462e-06, "loss": 0.1065, "num_input_tokens_seen": 58082112, "step": 30180 }, { "epoch": 4.9242189411860675, "grad_norm": 6.455053806304932, "learning_rate": 7.068719470702445e-06, "loss": 0.1087, "num_input_tokens_seen": 58091280, "step": 30185 }, { "epoch": 4.925034668406885, "grad_norm": 0.06187400966882706, "learning_rate": 7.061879891457229e-06, "loss": 0.001, "num_input_tokens_seen": 58101456, "step": 30190 }, { "epoch": 4.925850395627702, "grad_norm": 0.045471400022506714, "learning_rate": 7.0550430784955515e-06, "loss": 0.0845, "num_input_tokens_seen": 58110432, "step": 30195 }, { "epoch": 4.92666612284852, "grad_norm": 0.08027292788028717, "learning_rate": 7.048209032871752e-06, "loss": 0.0019, "num_input_tokens_seen": 58119904, "step": 30200 }, { "epoch": 4.92666612284852, "eval_loss": 0.2772332429885864, "eval_runtime": 80.7569, "eval_samples_per_second": 33.743, "eval_steps_per_second": 16.878, "num_input_tokens_seen": 58119904, "step": 30200 }, { "epoch": 4.927481850069337, "grad_norm": 0.06089227274060249, "learning_rate": 7.0413777556397055e-06, "loss": 0.0037, "num_input_tokens_seen": 58129072, "step": 30205 }, { "epoch": 4.928297577290154, "grad_norm": 0.025963740423321724, "learning_rate": 7.0345492478528925e-06, "loss": 0.0008, "num_input_tokens_seen": 58138832, "step": 30210 }, { "epoch": 4.929113304510971, "grad_norm": 0.021789852529764175, "learning_rate": 7.02772351056436e-06, "loss": 0.0005, "num_input_tokens_seen": 58148592, "step": 30215 }, { "epoch": 4.929929031731789, "grad_norm": 10.208946228027344, "learning_rate": 7.020900544826709e-06, "loss": 0.0738, "num_input_tokens_seen": 58158576, "step": 30220 }, { "epoch": 4.930744758952606, "grad_norm": 6.250285625457764, "learning_rate": 7.014080351692134e-06, "loss": 0.1231, "num_input_tokens_seen": 58168384, "step": 30225 }, { "epoch": 4.9315604861734235, "grad_norm": 21.557445526123047, "learning_rate": 7.0072629322124024e-06, "loss": 0.0492, "num_input_tokens_seen": 58177856, "step": 30230 }, { "epoch": 4.932376213394241, "grad_norm": 0.016319647431373596, "learning_rate": 7.000448287438827e-06, "loss": 0.0012, "num_input_tokens_seen": 58187584, "step": 30235 }, { "epoch": 4.933191940615059, "grad_norm": 0.0012522232718765736, "learning_rate": 6.993636418422331e-06, "loss": 0.1283, "num_input_tokens_seen": 58197312, "step": 30240 }, { "epoch": 4.934007667835876, "grad_norm": 0.00965146254748106, "learning_rate": 6.986827326213383e-06, "loss": 0.0655, "num_input_tokens_seen": 58205520, "step": 30245 }, { "epoch": 4.934823395056693, "grad_norm": 0.16983255743980408, "learning_rate": 6.9800210118620205e-06, "loss": 0.0693, "num_input_tokens_seen": 58214000, "step": 30250 }, { "epoch": 4.935639122277511, "grad_norm": 6.728057384490967, "learning_rate": 6.973217476417876e-06, "loss": 0.0592, "num_input_tokens_seen": 58224400, "step": 30255 }, { "epoch": 4.936454849498328, "grad_norm": 0.03857949748635292, "learning_rate": 6.96641672093013e-06, "loss": 0.1248, "num_input_tokens_seen": 58234720, "step": 30260 }, { "epoch": 4.937270576719145, "grad_norm": 0.05814433842897415, "learning_rate": 6.95961874644755e-06, "loss": 0.0295, "num_input_tokens_seen": 58244928, "step": 30265 }, { "epoch": 4.938086303939962, "grad_norm": 74.25923156738281, "learning_rate": 6.952823554018476e-06, "loss": 0.0155, "num_input_tokens_seen": 58254912, "step": 30270 }, { "epoch": 4.938902031160779, "grad_norm": 24.68526268005371, "learning_rate": 6.946031144690798e-06, "loss": 0.0191, "num_input_tokens_seen": 58263712, "step": 30275 }, { "epoch": 4.939717758381597, "grad_norm": 0.03048804961144924, "learning_rate": 6.939241519512005e-06, "loss": 0.0027, "num_input_tokens_seen": 58273504, "step": 30280 }, { "epoch": 4.940533485602415, "grad_norm": 0.02598307840526104, "learning_rate": 6.932454679529129e-06, "loss": 0.0255, "num_input_tokens_seen": 58283808, "step": 30285 }, { "epoch": 4.941349212823232, "grad_norm": 0.013296595774590969, "learning_rate": 6.925670625788791e-06, "loss": 0.0366, "num_input_tokens_seen": 58294192, "step": 30290 }, { "epoch": 4.94216494004405, "grad_norm": 0.02714889496564865, "learning_rate": 6.918889359337186e-06, "loss": 0.099, "num_input_tokens_seen": 58304512, "step": 30295 }, { "epoch": 4.942980667264867, "grad_norm": 0.05835993215441704, "learning_rate": 6.912110881220058e-06, "loss": 0.0003, "num_input_tokens_seen": 58313760, "step": 30300 }, { "epoch": 4.943796394485684, "grad_norm": 0.6184003949165344, "learning_rate": 6.905335192482735e-06, "loss": 0.0039, "num_input_tokens_seen": 58323328, "step": 30305 }, { "epoch": 4.944612121706501, "grad_norm": 0.4750020503997803, "learning_rate": 6.8985622941701275e-06, "loss": 0.0466, "num_input_tokens_seen": 58333872, "step": 30310 }, { "epoch": 4.945427848927318, "grad_norm": 0.04125474765896797, "learning_rate": 6.89179218732669e-06, "loss": 0.0007, "num_input_tokens_seen": 58342624, "step": 30315 }, { "epoch": 4.946243576148136, "grad_norm": 21.737855911254883, "learning_rate": 6.8850248729964595e-06, "loss": 0.0132, "num_input_tokens_seen": 58352272, "step": 30320 }, { "epoch": 4.947059303368953, "grad_norm": 0.06596808135509491, "learning_rate": 6.8782603522230314e-06, "loss": 0.0006, "num_input_tokens_seen": 58360704, "step": 30325 }, { "epoch": 4.9478750305897705, "grad_norm": 0.012718032114207745, "learning_rate": 6.871498626049591e-06, "loss": 0.0114, "num_input_tokens_seen": 58370464, "step": 30330 }, { "epoch": 4.9486907578105885, "grad_norm": 0.15737119317054749, "learning_rate": 6.8647396955188875e-06, "loss": 0.1579, "num_input_tokens_seen": 58380688, "step": 30335 }, { "epoch": 4.949506485031406, "grad_norm": 0.013747339136898518, "learning_rate": 6.857983561673218e-06, "loss": 0.0003, "num_input_tokens_seen": 58389168, "step": 30340 }, { "epoch": 4.950322212252223, "grad_norm": 0.024790838360786438, "learning_rate": 6.851230225554467e-06, "loss": 0.0167, "num_input_tokens_seen": 58398336, "step": 30345 }, { "epoch": 4.95113793947304, "grad_norm": 0.4110797047615051, "learning_rate": 6.8444796882040946e-06, "loss": 0.0021, "num_input_tokens_seen": 58407472, "step": 30350 }, { "epoch": 4.951953666693858, "grad_norm": 35.162445068359375, "learning_rate": 6.837731950663106e-06, "loss": 0.0383, "num_input_tokens_seen": 58415936, "step": 30355 }, { "epoch": 4.952769393914675, "grad_norm": 7.273758888244629, "learning_rate": 6.830987013972098e-06, "loss": 0.1121, "num_input_tokens_seen": 58425920, "step": 30360 }, { "epoch": 4.953585121135492, "grad_norm": 0.17323897778987885, "learning_rate": 6.82424487917121e-06, "loss": 0.0224, "num_input_tokens_seen": 58434608, "step": 30365 }, { "epoch": 4.954400848356309, "grad_norm": 0.1595456600189209, "learning_rate": 6.8175055473001735e-06, "loss": 0.001, "num_input_tokens_seen": 58444288, "step": 30370 }, { "epoch": 4.955216575577127, "grad_norm": 0.012859274633228779, "learning_rate": 6.8107690193982855e-06, "loss": 0.0011, "num_input_tokens_seen": 58454256, "step": 30375 }, { "epoch": 4.9560323027979445, "grad_norm": 27.508962631225586, "learning_rate": 6.804035296504385e-06, "loss": 0.0185, "num_input_tokens_seen": 58464288, "step": 30380 }, { "epoch": 4.956848030018762, "grad_norm": 0.002101593418046832, "learning_rate": 6.797304379656916e-06, "loss": 0.0006, "num_input_tokens_seen": 58474720, "step": 30385 }, { "epoch": 4.957663757239579, "grad_norm": 0.03400321304798126, "learning_rate": 6.790576269893861e-06, "loss": 0.0069, "num_input_tokens_seen": 58485472, "step": 30390 }, { "epoch": 4.958479484460397, "grad_norm": 1.2019349336624146, "learning_rate": 6.783850968252772e-06, "loss": 0.0011, "num_input_tokens_seen": 58494768, "step": 30395 }, { "epoch": 4.959295211681214, "grad_norm": 0.060442861169576645, "learning_rate": 6.777128475770789e-06, "loss": 0.0009, "num_input_tokens_seen": 58503776, "step": 30400 }, { "epoch": 4.959295211681214, "eval_loss": 0.2871687114238739, "eval_runtime": 80.7474, "eval_samples_per_second": 33.747, "eval_steps_per_second": 16.88, "num_input_tokens_seen": 58503776, "step": 30400 }, { "epoch": 4.960110938902031, "grad_norm": 0.03324951231479645, "learning_rate": 6.77040879348459e-06, "loss": 0.0005, "num_input_tokens_seen": 58514288, "step": 30405 }, { "epoch": 4.960926666122848, "grad_norm": 0.04882033169269562, "learning_rate": 6.763691922430443e-06, "loss": 0.0003, "num_input_tokens_seen": 58524288, "step": 30410 }, { "epoch": 4.961742393343666, "grad_norm": 0.06984736025333405, "learning_rate": 6.756977863644178e-06, "loss": 0.1112, "num_input_tokens_seen": 58534400, "step": 30415 }, { "epoch": 4.962558120564483, "grad_norm": 0.016764091327786446, "learning_rate": 6.7502666181611804e-06, "loss": 0.0005, "num_input_tokens_seen": 58543424, "step": 30420 }, { "epoch": 4.9633738477853, "grad_norm": 7.532958030700684, "learning_rate": 6.743558187016405e-06, "loss": 0.0704, "num_input_tokens_seen": 58553072, "step": 30425 }, { "epoch": 4.964189575006118, "grad_norm": 0.012507195584475994, "learning_rate": 6.7368525712443925e-06, "loss": 0.067, "num_input_tokens_seen": 58562384, "step": 30430 }, { "epoch": 4.965005302226936, "grad_norm": 0.05107172578573227, "learning_rate": 6.7301497718792155e-06, "loss": 0.0005, "num_input_tokens_seen": 58572256, "step": 30435 }, { "epoch": 4.965821029447753, "grad_norm": 14.083772659301758, "learning_rate": 6.723449789954544e-06, "loss": 0.0359, "num_input_tokens_seen": 58582000, "step": 30440 }, { "epoch": 4.96663675666857, "grad_norm": 0.007018223870545626, "learning_rate": 6.716752626503586e-06, "loss": 0.0983, "num_input_tokens_seen": 58590848, "step": 30445 }, { "epoch": 4.967452483889387, "grad_norm": 0.008265236392617226, "learning_rate": 6.710058282559131e-06, "loss": 0.0005, "num_input_tokens_seen": 58602192, "step": 30450 }, { "epoch": 4.968268211110205, "grad_norm": 0.01607561856508255, "learning_rate": 6.703366759153545e-06, "loss": 0.0007, "num_input_tokens_seen": 58611264, "step": 30455 }, { "epoch": 4.969083938331022, "grad_norm": 0.012047852389514446, "learning_rate": 6.6966780573187335e-06, "loss": 0.0002, "num_input_tokens_seen": 58620720, "step": 30460 }, { "epoch": 4.969899665551839, "grad_norm": 0.010841209441423416, "learning_rate": 6.689992178086174e-06, "loss": 0.0011, "num_input_tokens_seen": 58631104, "step": 30465 }, { "epoch": 4.970715392772657, "grad_norm": 0.05733518302440643, "learning_rate": 6.683309122486925e-06, "loss": 0.1266, "num_input_tokens_seen": 58640672, "step": 30470 }, { "epoch": 4.971531119993474, "grad_norm": 0.01532884780317545, "learning_rate": 6.676628891551584e-06, "loss": 0.0653, "num_input_tokens_seen": 58650480, "step": 30475 }, { "epoch": 4.9723468472142915, "grad_norm": 0.061444658786058426, "learning_rate": 6.6699514863103385e-06, "loss": 0.0003, "num_input_tokens_seen": 58661280, "step": 30480 }, { "epoch": 4.973162574435109, "grad_norm": 0.020258828997612, "learning_rate": 6.663276907792921e-06, "loss": 0.0599, "num_input_tokens_seen": 58670832, "step": 30485 }, { "epoch": 4.973978301655926, "grad_norm": 0.018063649535179138, "learning_rate": 6.656605157028634e-06, "loss": 0.09, "num_input_tokens_seen": 58679776, "step": 30490 }, { "epoch": 4.974794028876744, "grad_norm": 0.014849094673991203, "learning_rate": 6.649936235046358e-06, "loss": 0.0056, "num_input_tokens_seen": 58688880, "step": 30495 }, { "epoch": 4.975609756097561, "grad_norm": 5.800947189331055, "learning_rate": 6.643270142874508e-06, "loss": 0.0253, "num_input_tokens_seen": 58698480, "step": 30500 }, { "epoch": 4.976425483318378, "grad_norm": 0.0051178704015910625, "learning_rate": 6.636606881541094e-06, "loss": 0.0003, "num_input_tokens_seen": 58706944, "step": 30505 }, { "epoch": 4.977241210539196, "grad_norm": 0.00138578109908849, "learning_rate": 6.629946452073662e-06, "loss": 0.0008, "num_input_tokens_seen": 58716368, "step": 30510 }, { "epoch": 4.978056937760013, "grad_norm": 0.1947837471961975, "learning_rate": 6.6232888554993375e-06, "loss": 0.0005, "num_input_tokens_seen": 58725392, "step": 30515 }, { "epoch": 4.97887266498083, "grad_norm": 10.312009811401367, "learning_rate": 6.616634092844817e-06, "loss": 0.2031, "num_input_tokens_seen": 58735152, "step": 30520 }, { "epoch": 4.9796883922016475, "grad_norm": 0.2626264989376068, "learning_rate": 6.609982165136331e-06, "loss": 0.0012, "num_input_tokens_seen": 58744224, "step": 30525 }, { "epoch": 4.9805041194224655, "grad_norm": 0.0028049293905496597, "learning_rate": 6.603333073399706e-06, "loss": 0.0881, "num_input_tokens_seen": 58753552, "step": 30530 }, { "epoch": 4.981319846643283, "grad_norm": 0.0643429085612297, "learning_rate": 6.596686818660308e-06, "loss": 0.0009, "num_input_tokens_seen": 58764160, "step": 30535 }, { "epoch": 4.9821355738641, "grad_norm": 0.293415367603302, "learning_rate": 6.590043401943066e-06, "loss": 0.0004, "num_input_tokens_seen": 58774000, "step": 30540 }, { "epoch": 4.982951301084917, "grad_norm": 0.5484192967414856, "learning_rate": 6.583402824272494e-06, "loss": 0.0019, "num_input_tokens_seen": 58784320, "step": 30545 }, { "epoch": 4.983767028305735, "grad_norm": 0.010343950241804123, "learning_rate": 6.576765086672634e-06, "loss": 0.0007, "num_input_tokens_seen": 58794032, "step": 30550 }, { "epoch": 4.984582755526552, "grad_norm": 0.0012514879927039146, "learning_rate": 6.57013019016712e-06, "loss": 0.0006, "num_input_tokens_seen": 58803808, "step": 30555 }, { "epoch": 4.985398482747369, "grad_norm": 0.012246488593518734, "learning_rate": 6.563498135779142e-06, "loss": 0.0009, "num_input_tokens_seen": 58812704, "step": 30560 }, { "epoch": 4.986214209968186, "grad_norm": 0.10214684903621674, "learning_rate": 6.556868924531431e-06, "loss": 0.0006, "num_input_tokens_seen": 58822160, "step": 30565 }, { "epoch": 4.987029937189004, "grad_norm": 0.07233995944261551, "learning_rate": 6.550242557446304e-06, "loss": 0.0007, "num_input_tokens_seen": 58833168, "step": 30570 }, { "epoch": 4.9878456644098215, "grad_norm": 0.008411375805735588, "learning_rate": 6.543619035545634e-06, "loss": 0.0003, "num_input_tokens_seen": 58843072, "step": 30575 }, { "epoch": 4.988661391630639, "grad_norm": 0.026670951396226883, "learning_rate": 6.53699835985084e-06, "loss": 0.0503, "num_input_tokens_seen": 58852832, "step": 30580 }, { "epoch": 4.989477118851456, "grad_norm": 0.037765685468912125, "learning_rate": 6.530380531382927e-06, "loss": 0.031, "num_input_tokens_seen": 58862160, "step": 30585 }, { "epoch": 4.990292846072274, "grad_norm": 0.24176208674907684, "learning_rate": 6.523765551162433e-06, "loss": 0.0017, "num_input_tokens_seen": 58872320, "step": 30590 }, { "epoch": 4.991108573293091, "grad_norm": 0.0030055276583880186, "learning_rate": 6.517153420209476e-06, "loss": 0.0006, "num_input_tokens_seen": 58882432, "step": 30595 }, { "epoch": 4.991924300513908, "grad_norm": 0.0291853379458189, "learning_rate": 6.510544139543739e-06, "loss": 0.0006, "num_input_tokens_seen": 58892528, "step": 30600 }, { "epoch": 4.991924300513908, "eval_loss": 0.2907790243625641, "eval_runtime": 80.8164, "eval_samples_per_second": 33.718, "eval_steps_per_second": 16.865, "num_input_tokens_seen": 58892528, "step": 30600 }, { "epoch": 4.992740027734725, "grad_norm": 0.016013478860259056, "learning_rate": 6.503937710184452e-06, "loss": 0.0011, "num_input_tokens_seen": 58901888, "step": 30605 }, { "epoch": 4.993555754955543, "grad_norm": 0.007794367149472237, "learning_rate": 6.4973341331503954e-06, "loss": 0.0004, "num_input_tokens_seen": 58911040, "step": 30610 }, { "epoch": 4.99437148217636, "grad_norm": 0.9587812423706055, "learning_rate": 6.490733409459942e-06, "loss": 0.1623, "num_input_tokens_seen": 58920976, "step": 30615 }, { "epoch": 4.995187209397177, "grad_norm": 0.02157323993742466, "learning_rate": 6.484135540130995e-06, "loss": 0.1114, "num_input_tokens_seen": 58930176, "step": 30620 }, { "epoch": 4.9960029366179945, "grad_norm": 5.63030481338501, "learning_rate": 6.4775405261810364e-06, "loss": 0.1429, "num_input_tokens_seen": 58938928, "step": 30625 }, { "epoch": 4.996818663838813, "grad_norm": 0.03717818483710289, "learning_rate": 6.470948368627092e-06, "loss": 0.0978, "num_input_tokens_seen": 58949632, "step": 30630 }, { "epoch": 4.99763439105963, "grad_norm": 0.002182401018217206, "learning_rate": 6.464359068485756e-06, "loss": 0.0001, "num_input_tokens_seen": 58958944, "step": 30635 }, { "epoch": 4.998450118280447, "grad_norm": 0.09002501517534256, "learning_rate": 6.457772626773195e-06, "loss": 0.0083, "num_input_tokens_seen": 58967296, "step": 30640 }, { "epoch": 4.999265845501265, "grad_norm": 0.026380367577075958, "learning_rate": 6.451189044505104e-06, "loss": 0.0003, "num_input_tokens_seen": 58978032, "step": 30645 }, { "epoch": 5.0, "grad_norm": 0.03750684857368469, "learning_rate": 6.44460832269676e-06, "loss": 0.0013, "num_input_tokens_seen": 58987136, "step": 30650 }, { "epoch": 5.000815727220817, "grad_norm": 0.109971784055233, "learning_rate": 6.438030462363001e-06, "loss": 0.0008, "num_input_tokens_seen": 58997088, "step": 30655 }, { "epoch": 5.001631454441635, "grad_norm": 0.11415304988622665, "learning_rate": 6.431455464518205e-06, "loss": 0.0005, "num_input_tokens_seen": 59006448, "step": 30660 }, { "epoch": 5.002447181662452, "grad_norm": 0.06036347150802612, "learning_rate": 6.424883330176326e-06, "loss": 0.0288, "num_input_tokens_seen": 59016400, "step": 30665 }, { "epoch": 5.003262908883269, "grad_norm": 0.009359806776046753, "learning_rate": 6.418314060350864e-06, "loss": 0.0002, "num_input_tokens_seen": 59026784, "step": 30670 }, { "epoch": 5.0040786361040865, "grad_norm": 0.008684606291353703, "learning_rate": 6.4117476560548895e-06, "loss": 0.0004, "num_input_tokens_seen": 59036960, "step": 30675 }, { "epoch": 5.004894363324905, "grad_norm": 0.008994431234896183, "learning_rate": 6.405184118301016e-06, "loss": 0.059, "num_input_tokens_seen": 59046400, "step": 30680 }, { "epoch": 5.005710090545722, "grad_norm": 0.020929235965013504, "learning_rate": 6.398623448101434e-06, "loss": 0.0005, "num_input_tokens_seen": 59056208, "step": 30685 }, { "epoch": 5.006525817766539, "grad_norm": 0.01047689002007246, "learning_rate": 6.392065646467871e-06, "loss": 0.0001, "num_input_tokens_seen": 59065392, "step": 30690 }, { "epoch": 5.007341544987356, "grad_norm": 16.03750991821289, "learning_rate": 6.385510714411632e-06, "loss": 0.055, "num_input_tokens_seen": 59076560, "step": 30695 }, { "epoch": 5.008157272208174, "grad_norm": 0.011042055673897266, "learning_rate": 6.378958652943559e-06, "loss": 0.0003, "num_input_tokens_seen": 59085824, "step": 30700 }, { "epoch": 5.008972999428991, "grad_norm": 0.008968928828835487, "learning_rate": 6.3724094630740776e-06, "loss": 0.0002, "num_input_tokens_seen": 59095840, "step": 30705 }, { "epoch": 5.009788726649808, "grad_norm": 0.09539271891117096, "learning_rate": 6.365863145813136e-06, "loss": 0.0511, "num_input_tokens_seen": 59104816, "step": 30710 }, { "epoch": 5.010604453870625, "grad_norm": 0.07469534873962402, "learning_rate": 6.359319702170269e-06, "loss": 0.0008, "num_input_tokens_seen": 59113680, "step": 30715 }, { "epoch": 5.011420181091443, "grad_norm": 0.017290238291025162, "learning_rate": 6.352779133154566e-06, "loss": 0.001, "num_input_tokens_seen": 59122144, "step": 30720 }, { "epoch": 5.0122359083122605, "grad_norm": 0.1714223325252533, "learning_rate": 6.346241439774648e-06, "loss": 0.0005, "num_input_tokens_seen": 59132240, "step": 30725 }, { "epoch": 5.013051635533078, "grad_norm": 0.0034069567918777466, "learning_rate": 6.339706623038716e-06, "loss": 0.0093, "num_input_tokens_seen": 59141792, "step": 30730 }, { "epoch": 5.013867362753895, "grad_norm": 0.026847992092370987, "learning_rate": 6.333174683954532e-06, "loss": 0.0006, "num_input_tokens_seen": 59151728, "step": 30735 }, { "epoch": 5.014683089974713, "grad_norm": 0.01422700472176075, "learning_rate": 6.326645623529387e-06, "loss": 0.0003, "num_input_tokens_seen": 59161472, "step": 30740 }, { "epoch": 5.01549881719553, "grad_norm": 0.014845471829175949, "learning_rate": 6.320119442770156e-06, "loss": 0.0004, "num_input_tokens_seen": 59170672, "step": 30745 }, { "epoch": 5.016314544416347, "grad_norm": 0.04366480931639671, "learning_rate": 6.313596142683254e-06, "loss": 0.0005, "num_input_tokens_seen": 59180784, "step": 30750 }, { "epoch": 5.017130271637164, "grad_norm": 0.0247783362865448, "learning_rate": 6.307075724274647e-06, "loss": 0.0329, "num_input_tokens_seen": 59190624, "step": 30755 }, { "epoch": 5.017945998857982, "grad_norm": 0.0043134610168635845, "learning_rate": 6.300558188549882e-06, "loss": 0.0014, "num_input_tokens_seen": 59200672, "step": 30760 }, { "epoch": 5.018761726078799, "grad_norm": 0.0699254721403122, "learning_rate": 6.29404353651403e-06, "loss": 0.1448, "num_input_tokens_seen": 59211680, "step": 30765 }, { "epoch": 5.0195774532996165, "grad_norm": 0.0008285525254905224, "learning_rate": 6.287531769171737e-06, "loss": 0.0004, "num_input_tokens_seen": 59220528, "step": 30770 }, { "epoch": 5.020393180520434, "grad_norm": 0.007265446707606316, "learning_rate": 6.2810228875272045e-06, "loss": 0.0002, "num_input_tokens_seen": 59230688, "step": 30775 }, { "epoch": 5.021208907741252, "grad_norm": 0.004445031750947237, "learning_rate": 6.274516892584179e-06, "loss": 0.0004, "num_input_tokens_seen": 59239568, "step": 30780 }, { "epoch": 5.022024634962069, "grad_norm": 0.04526481777429581, "learning_rate": 6.268013785345969e-06, "loss": 0.1358, "num_input_tokens_seen": 59249376, "step": 30785 }, { "epoch": 5.022840362182886, "grad_norm": 19.317230224609375, "learning_rate": 6.26151356681543e-06, "loss": 0.0353, "num_input_tokens_seen": 59258864, "step": 30790 }, { "epoch": 5.023656089403703, "grad_norm": 0.0077144680544734, "learning_rate": 6.255016237994981e-06, "loss": 0.0003, "num_input_tokens_seen": 59270192, "step": 30795 }, { "epoch": 5.024471816624521, "grad_norm": 0.15244203805923462, "learning_rate": 6.248521799886603e-06, "loss": 0.0736, "num_input_tokens_seen": 59278112, "step": 30800 }, { "epoch": 5.024471816624521, "eval_loss": 0.2989204525947571, "eval_runtime": 80.695, "eval_samples_per_second": 33.769, "eval_steps_per_second": 16.891, "num_input_tokens_seen": 59278112, "step": 30800 }, { "epoch": 5.025287543845338, "grad_norm": 0.10966271162033081, "learning_rate": 6.242030253491798e-06, "loss": 0.0007, "num_input_tokens_seen": 59288064, "step": 30805 }, { "epoch": 5.026103271066155, "grad_norm": 0.02611847221851349, "learning_rate": 6.235541599811656e-06, "loss": 0.0003, "num_input_tokens_seen": 59296688, "step": 30810 }, { "epoch": 5.026918998286972, "grad_norm": 0.006320343352854252, "learning_rate": 6.229055839846814e-06, "loss": 0.0003, "num_input_tokens_seen": 59306880, "step": 30815 }, { "epoch": 5.02773472550779, "grad_norm": 0.014552047476172447, "learning_rate": 6.222572974597455e-06, "loss": 0.0004, "num_input_tokens_seen": 59315184, "step": 30820 }, { "epoch": 5.028550452728608, "grad_norm": 0.009076819755136967, "learning_rate": 6.216093005063306e-06, "loss": 0.0002, "num_input_tokens_seen": 59324208, "step": 30825 }, { "epoch": 5.029366179949425, "grad_norm": 0.021181363612413406, "learning_rate": 6.209615932243678e-06, "loss": 0.001, "num_input_tokens_seen": 59333712, "step": 30830 }, { "epoch": 5.030181907170242, "grad_norm": 0.2079966515302658, "learning_rate": 6.203141757137399e-06, "loss": 0.0004, "num_input_tokens_seen": 59341360, "step": 30835 }, { "epoch": 5.03099763439106, "grad_norm": 0.002395458286628127, "learning_rate": 6.196670480742886e-06, "loss": 0.1127, "num_input_tokens_seen": 59351184, "step": 30840 }, { "epoch": 5.031813361611877, "grad_norm": 0.22538582980632782, "learning_rate": 6.190202104058074e-06, "loss": 0.001, "num_input_tokens_seen": 59361520, "step": 30845 }, { "epoch": 5.032629088832694, "grad_norm": 0.020961783826351166, "learning_rate": 6.183736628080475e-06, "loss": 0.0206, "num_input_tokens_seen": 59370736, "step": 30850 }, { "epoch": 5.033444816053512, "grad_norm": 0.013540329411625862, "learning_rate": 6.177274053807155e-06, "loss": 0.0692, "num_input_tokens_seen": 59379920, "step": 30855 }, { "epoch": 5.034260543274329, "grad_norm": 0.060746245086193085, "learning_rate": 6.170814382234713e-06, "loss": 0.0003, "num_input_tokens_seen": 59389552, "step": 30860 }, { "epoch": 5.035076270495146, "grad_norm": 4.060926914215088, "learning_rate": 6.16435761435932e-06, "loss": 0.0338, "num_input_tokens_seen": 59399792, "step": 30865 }, { "epoch": 5.0358919977159635, "grad_norm": 0.05258561298251152, "learning_rate": 6.157903751176681e-06, "loss": 0.0005, "num_input_tokens_seen": 59409216, "step": 30870 }, { "epoch": 5.0367077249367815, "grad_norm": 0.008724750950932503, "learning_rate": 6.151452793682066e-06, "loss": 0.0004, "num_input_tokens_seen": 59419760, "step": 30875 }, { "epoch": 5.037523452157599, "grad_norm": 0.03288612514734268, "learning_rate": 6.145004742870305e-06, "loss": 0.0491, "num_input_tokens_seen": 59428928, "step": 30880 }, { "epoch": 5.038339179378416, "grad_norm": 0.009054508991539478, "learning_rate": 6.138559599735752e-06, "loss": 0.0002, "num_input_tokens_seen": 59439440, "step": 30885 }, { "epoch": 5.039154906599233, "grad_norm": 0.007498360704630613, "learning_rate": 6.132117365272344e-06, "loss": 0.0045, "num_input_tokens_seen": 59448544, "step": 30890 }, { "epoch": 5.039970633820051, "grad_norm": 0.16031551361083984, "learning_rate": 6.125678040473545e-06, "loss": 0.0008, "num_input_tokens_seen": 59457904, "step": 30895 }, { "epoch": 5.040786361040868, "grad_norm": 0.012220565229654312, "learning_rate": 6.1192416263323755e-06, "loss": 0.0003, "num_input_tokens_seen": 59466768, "step": 30900 }, { "epoch": 5.041602088261685, "grad_norm": 0.1529027819633484, "learning_rate": 6.112808123841424e-06, "loss": 0.001, "num_input_tokens_seen": 59476176, "step": 30905 }, { "epoch": 5.042417815482502, "grad_norm": 0.007863990031182766, "learning_rate": 6.106377533992805e-06, "loss": 0.0012, "num_input_tokens_seen": 59485728, "step": 30910 }, { "epoch": 5.04323354270332, "grad_norm": 0.04850643873214722, "learning_rate": 6.099949857778204e-06, "loss": 0.0004, "num_input_tokens_seen": 59495120, "step": 30915 }, { "epoch": 5.0440492699241375, "grad_norm": 0.016838911920785904, "learning_rate": 6.093525096188852e-06, "loss": 0.0005, "num_input_tokens_seen": 59505328, "step": 30920 }, { "epoch": 5.044864997144955, "grad_norm": 0.005308972205966711, "learning_rate": 6.087103250215518e-06, "loss": 0.0006, "num_input_tokens_seen": 59515232, "step": 30925 }, { "epoch": 5.045680724365772, "grad_norm": 0.12629809975624084, "learning_rate": 6.080684320848537e-06, "loss": 0.0004, "num_input_tokens_seen": 59524640, "step": 30930 }, { "epoch": 5.04649645158659, "grad_norm": 0.018715862184762955, "learning_rate": 6.074268309077794e-06, "loss": 0.0786, "num_input_tokens_seen": 59534624, "step": 30935 }, { "epoch": 5.047312178807407, "grad_norm": 0.0058837709948420525, "learning_rate": 6.067855215892709e-06, "loss": 0.0012, "num_input_tokens_seen": 59544128, "step": 30940 }, { "epoch": 5.048127906028224, "grad_norm": 0.05799185484647751, "learning_rate": 6.061445042282271e-06, "loss": 0.0054, "num_input_tokens_seen": 59555344, "step": 30945 }, { "epoch": 5.048943633249041, "grad_norm": 0.35772955417633057, "learning_rate": 6.055037789234999e-06, "loss": 0.0643, "num_input_tokens_seen": 59564832, "step": 30950 }, { "epoch": 5.049759360469859, "grad_norm": 0.007400171365588903, "learning_rate": 6.048633457738975e-06, "loss": 0.0002, "num_input_tokens_seen": 59574176, "step": 30955 }, { "epoch": 5.050575087690676, "grad_norm": 0.004096610005944967, "learning_rate": 6.042232048781837e-06, "loss": 0.0002, "num_input_tokens_seen": 59583504, "step": 30960 }, { "epoch": 5.051390814911493, "grad_norm": 0.01353007648140192, "learning_rate": 6.035833563350757e-06, "loss": 0.0002, "num_input_tokens_seen": 59594816, "step": 30965 }, { "epoch": 5.052206542132311, "grad_norm": 0.02594667486846447, "learning_rate": 6.0294380024324525e-06, "loss": 0.0002, "num_input_tokens_seen": 59604288, "step": 30970 }, { "epoch": 5.053022269353129, "grad_norm": 17.996803283691406, "learning_rate": 6.023045367013213e-06, "loss": 0.029, "num_input_tokens_seen": 59614960, "step": 30975 }, { "epoch": 5.053837996573946, "grad_norm": 0.14007024466991425, "learning_rate": 6.016655658078851e-06, "loss": 0.0014, "num_input_tokens_seen": 59624720, "step": 30980 }, { "epoch": 5.054653723794763, "grad_norm": 0.08318176865577698, "learning_rate": 6.010268876614753e-06, "loss": 0.0006, "num_input_tokens_seen": 59634224, "step": 30985 }, { "epoch": 5.05546945101558, "grad_norm": 0.0348329097032547, "learning_rate": 6.0038850236058266e-06, "loss": 0.0021, "num_input_tokens_seen": 59643664, "step": 30990 }, { "epoch": 5.056285178236398, "grad_norm": 0.009365054778754711, "learning_rate": 5.997504100036549e-06, "loss": 0.0002, "num_input_tokens_seen": 59654096, "step": 30995 }, { "epoch": 5.057100905457215, "grad_norm": 0.2098425030708313, "learning_rate": 5.991126106890949e-06, "loss": 0.0017, "num_input_tokens_seen": 59663264, "step": 31000 }, { "epoch": 5.057100905457215, "eval_loss": 0.3235328495502472, "eval_runtime": 80.75, "eval_samples_per_second": 33.746, "eval_steps_per_second": 16.879, "num_input_tokens_seen": 59663264, "step": 31000 }, { "epoch": 5.057916632678032, "grad_norm": 0.007061836775392294, "learning_rate": 5.984751045152576e-06, "loss": 0.0836, "num_input_tokens_seen": 59672192, "step": 31005 }, { "epoch": 5.058732359898849, "grad_norm": 0.08194391429424286, "learning_rate": 5.978378915804553e-06, "loss": 0.0011, "num_input_tokens_seen": 59682496, "step": 31010 }, { "epoch": 5.059548087119667, "grad_norm": 0.010696902871131897, "learning_rate": 5.972009719829547e-06, "loss": 0.0016, "num_input_tokens_seen": 59692608, "step": 31015 }, { "epoch": 5.0603638143404845, "grad_norm": 0.10578663647174835, "learning_rate": 5.965643458209755e-06, "loss": 0.0236, "num_input_tokens_seen": 59702304, "step": 31020 }, { "epoch": 5.061179541561302, "grad_norm": 0.33080533146858215, "learning_rate": 5.95928013192695e-06, "loss": 0.0007, "num_input_tokens_seen": 59710688, "step": 31025 }, { "epoch": 5.06199526878212, "grad_norm": 0.017366528511047363, "learning_rate": 5.952919741962423e-06, "loss": 0.0004, "num_input_tokens_seen": 59719696, "step": 31030 }, { "epoch": 5.062810996002937, "grad_norm": 0.004564769100397825, "learning_rate": 5.946562289297042e-06, "loss": 0.0002, "num_input_tokens_seen": 59730336, "step": 31035 }, { "epoch": 5.063626723223754, "grad_norm": 0.08805099129676819, "learning_rate": 5.9402077749111855e-06, "loss": 0.059, "num_input_tokens_seen": 59738304, "step": 31040 }, { "epoch": 5.064442450444571, "grad_norm": 0.020983580499887466, "learning_rate": 5.933856199784821e-06, "loss": 0.0002, "num_input_tokens_seen": 59748032, "step": 31045 }, { "epoch": 5.065258177665389, "grad_norm": 0.005441619083285332, "learning_rate": 5.927507564897419e-06, "loss": 0.0034, "num_input_tokens_seen": 59757712, "step": 31050 }, { "epoch": 5.066073904886206, "grad_norm": 0.007705654017627239, "learning_rate": 5.9211618712280395e-06, "loss": 0.0001, "num_input_tokens_seen": 59767632, "step": 31055 }, { "epoch": 5.066889632107023, "grad_norm": 0.0026555333752185106, "learning_rate": 5.914819119755255e-06, "loss": 0.0004, "num_input_tokens_seen": 59777104, "step": 31060 }, { "epoch": 5.0677053593278405, "grad_norm": 0.008944607339799404, "learning_rate": 5.908479311457205e-06, "loss": 0.0001, "num_input_tokens_seen": 59787040, "step": 31065 }, { "epoch": 5.0685210865486585, "grad_norm": 0.014516103081405163, "learning_rate": 5.902142447311559e-06, "loss": 0.0004, "num_input_tokens_seen": 59796688, "step": 31070 }, { "epoch": 5.069336813769476, "grad_norm": 0.060582585632801056, "learning_rate": 5.895808528295546e-06, "loss": 0.0002, "num_input_tokens_seen": 59806720, "step": 31075 }, { "epoch": 5.070152540990293, "grad_norm": 0.02949884906411171, "learning_rate": 5.889477555385941e-06, "loss": 0.0005, "num_input_tokens_seen": 59816432, "step": 31080 }, { "epoch": 5.07096826821111, "grad_norm": 0.021018370985984802, "learning_rate": 5.883149529559051e-06, "loss": 0.0002, "num_input_tokens_seen": 59825552, "step": 31085 }, { "epoch": 5.071783995431928, "grad_norm": 0.014025885611772537, "learning_rate": 5.876824451790738e-06, "loss": 0.0001, "num_input_tokens_seen": 59835040, "step": 31090 }, { "epoch": 5.072599722652745, "grad_norm": 0.0313572995364666, "learning_rate": 5.87050232305642e-06, "loss": 0.0958, "num_input_tokens_seen": 59844928, "step": 31095 }, { "epoch": 5.073415449873562, "grad_norm": 0.01302633248269558, "learning_rate": 5.864183144331034e-06, "loss": 0.0012, "num_input_tokens_seen": 59855408, "step": 31100 }, { "epoch": 5.074231177094379, "grad_norm": 0.011921796016395092, "learning_rate": 5.857866916589089e-06, "loss": 0.0734, "num_input_tokens_seen": 59865776, "step": 31105 }, { "epoch": 5.075046904315197, "grad_norm": 0.026202088221907616, "learning_rate": 5.8515536408046216e-06, "loss": 0.0003, "num_input_tokens_seen": 59875584, "step": 31110 }, { "epoch": 5.0758626315360145, "grad_norm": 0.038228005170822144, "learning_rate": 5.845243317951208e-06, "loss": 0.0003, "num_input_tokens_seen": 59884928, "step": 31115 }, { "epoch": 5.076678358756832, "grad_norm": 0.10030128806829453, "learning_rate": 5.838935949001997e-06, "loss": 0.0014, "num_input_tokens_seen": 59893936, "step": 31120 }, { "epoch": 5.077494085977649, "grad_norm": 0.0007789502269588411, "learning_rate": 5.8326315349296476e-06, "loss": 0.0001, "num_input_tokens_seen": 59903120, "step": 31125 }, { "epoch": 5.078309813198467, "grad_norm": 0.00466663995757699, "learning_rate": 5.826330076706396e-06, "loss": 0.0004, "num_input_tokens_seen": 59911824, "step": 31130 }, { "epoch": 5.079125540419284, "grad_norm": 0.003268468426540494, "learning_rate": 5.820031575303988e-06, "loss": 0.0001, "num_input_tokens_seen": 59921168, "step": 31135 }, { "epoch": 5.079941267640101, "grad_norm": 0.009664093144237995, "learning_rate": 5.813736031693745e-06, "loss": 0.0001, "num_input_tokens_seen": 59930144, "step": 31140 }, { "epoch": 5.080756994860918, "grad_norm": 0.007931915111839771, "learning_rate": 5.807443446846522e-06, "loss": 0.0003, "num_input_tokens_seen": 59940800, "step": 31145 }, { "epoch": 5.081572722081736, "grad_norm": 0.01763410121202469, "learning_rate": 5.801153821732699e-06, "loss": 0.0001, "num_input_tokens_seen": 59950048, "step": 31150 }, { "epoch": 5.082388449302553, "grad_norm": 0.010083198547363281, "learning_rate": 5.794867157322229e-06, "loss": 0.0001, "num_input_tokens_seen": 59959584, "step": 31155 }, { "epoch": 5.08320417652337, "grad_norm": 0.0438731387257576, "learning_rate": 5.788583454584593e-06, "loss": 0.0005, "num_input_tokens_seen": 59969216, "step": 31160 }, { "epoch": 5.0840199037441876, "grad_norm": 2.881692409515381, "learning_rate": 5.7823027144888075e-06, "loss": 0.0456, "num_input_tokens_seen": 59977616, "step": 31165 }, { "epoch": 5.084835630965006, "grad_norm": 0.0038481371011584997, "learning_rate": 5.776024938003455e-06, "loss": 0.0178, "num_input_tokens_seen": 59986768, "step": 31170 }, { "epoch": 5.085651358185823, "grad_norm": 9.64509105682373, "learning_rate": 5.7697501260966345e-06, "loss": 0.0732, "num_input_tokens_seen": 59997792, "step": 31175 }, { "epoch": 5.08646708540664, "grad_norm": 0.03111070953309536, "learning_rate": 5.7634782797360145e-06, "loss": 0.0571, "num_input_tokens_seen": 60007920, "step": 31180 }, { "epoch": 5.087282812627457, "grad_norm": 0.11465128511190414, "learning_rate": 5.757209399888777e-06, "loss": 0.0116, "num_input_tokens_seen": 60016752, "step": 31185 }, { "epoch": 5.088098539848275, "grad_norm": 0.015142572112381458, "learning_rate": 5.750943487521679e-06, "loss": 0.0478, "num_input_tokens_seen": 60026752, "step": 31190 }, { "epoch": 5.088914267069092, "grad_norm": 0.005200257990509272, "learning_rate": 5.744680543600986e-06, "loss": 0.088, "num_input_tokens_seen": 60036560, "step": 31195 }, { "epoch": 5.089729994289909, "grad_norm": 0.0008678357116878033, "learning_rate": 5.738420569092537e-06, "loss": 0.0006, "num_input_tokens_seen": 60047056, "step": 31200 }, { "epoch": 5.089729994289909, "eval_loss": 0.33572661876678467, "eval_runtime": 80.7056, "eval_samples_per_second": 33.765, "eval_steps_per_second": 16.889, "num_input_tokens_seen": 60047056, "step": 31200 }, { "epoch": 5.090545721510727, "grad_norm": 0.05431924760341644, "learning_rate": 5.732163564961684e-06, "loss": 0.0003, "num_input_tokens_seen": 60057440, "step": 31205 }, { "epoch": 5.091361448731544, "grad_norm": 0.00012081401655450463, "learning_rate": 5.725909532173354e-06, "loss": 0.0002, "num_input_tokens_seen": 60066672, "step": 31210 }, { "epoch": 5.0921771759523615, "grad_norm": 0.07571234554052353, "learning_rate": 5.719658471691977e-06, "loss": 0.055, "num_input_tokens_seen": 60077088, "step": 31215 }, { "epoch": 5.092992903173179, "grad_norm": 0.05824246630072594, "learning_rate": 5.71341038448156e-06, "loss": 0.0005, "num_input_tokens_seen": 60087248, "step": 31220 }, { "epoch": 5.093808630393997, "grad_norm": 0.1019153818488121, "learning_rate": 5.707165271505635e-06, "loss": 0.001, "num_input_tokens_seen": 60097344, "step": 31225 }, { "epoch": 5.094624357614814, "grad_norm": 0.007632746826857328, "learning_rate": 5.700923133727271e-06, "loss": 0.0013, "num_input_tokens_seen": 60107488, "step": 31230 }, { "epoch": 5.095440084835631, "grad_norm": 17.778562545776367, "learning_rate": 5.694683972109083e-06, "loss": 0.0856, "num_input_tokens_seen": 60116960, "step": 31235 }, { "epoch": 5.096255812056448, "grad_norm": 0.2742181122303009, "learning_rate": 5.688447787613241e-06, "loss": 0.0008, "num_input_tokens_seen": 60126160, "step": 31240 }, { "epoch": 5.097071539277266, "grad_norm": 0.03183436021208763, "learning_rate": 5.6822145812014285e-06, "loss": 0.0652, "num_input_tokens_seen": 60135872, "step": 31245 }, { "epoch": 5.097887266498083, "grad_norm": 0.002240631263703108, "learning_rate": 5.675984353834896e-06, "loss": 0.0004, "num_input_tokens_seen": 60145184, "step": 31250 }, { "epoch": 5.0987029937189, "grad_norm": 0.03552916646003723, "learning_rate": 5.66975710647441e-06, "loss": 0.0002, "num_input_tokens_seen": 60155664, "step": 31255 }, { "epoch": 5.0995187209397175, "grad_norm": 0.01464246865361929, "learning_rate": 5.663532840080304e-06, "loss": 0.0001, "num_input_tokens_seen": 60165680, "step": 31260 }, { "epoch": 5.1003344481605355, "grad_norm": 0.011792853474617004, "learning_rate": 5.6573115556124325e-06, "loss": 0.0001, "num_input_tokens_seen": 60175008, "step": 31265 }, { "epoch": 5.101150175381353, "grad_norm": 0.001099567161872983, "learning_rate": 5.651093254030185e-06, "loss": 0.0001, "num_input_tokens_seen": 60183776, "step": 31270 }, { "epoch": 5.10196590260217, "grad_norm": 1.54733407497406, "learning_rate": 5.644877936292514e-06, "loss": 0.0698, "num_input_tokens_seen": 60192928, "step": 31275 }, { "epoch": 5.102781629822987, "grad_norm": 0.026692837476730347, "learning_rate": 5.638665603357901e-06, "loss": 0.0007, "num_input_tokens_seen": 60202288, "step": 31280 }, { "epoch": 5.103597357043805, "grad_norm": 0.015945566818118095, "learning_rate": 5.632456256184357e-06, "loss": 0.0001, "num_input_tokens_seen": 60211872, "step": 31285 }, { "epoch": 5.104413084264622, "grad_norm": 0.02065522037446499, "learning_rate": 5.626249895729452e-06, "loss": 0.0004, "num_input_tokens_seen": 60220112, "step": 31290 }, { "epoch": 5.105228811485439, "grad_norm": 15.62655258178711, "learning_rate": 5.620046522950273e-06, "loss": 0.018, "num_input_tokens_seen": 60229856, "step": 31295 }, { "epoch": 5.106044538706256, "grad_norm": 6.504356861114502, "learning_rate": 5.613846138803464e-06, "loss": 0.0691, "num_input_tokens_seen": 60239248, "step": 31300 }, { "epoch": 5.106860265927074, "grad_norm": 0.12094593793153763, "learning_rate": 5.607648744245206e-06, "loss": 0.0005, "num_input_tokens_seen": 60249168, "step": 31305 }, { "epoch": 5.1076759931478914, "grad_norm": 0.7142212390899658, "learning_rate": 5.601454340231207e-06, "loss": 0.0011, "num_input_tokens_seen": 60258896, "step": 31310 }, { "epoch": 5.108491720368709, "grad_norm": 0.009344909340143204, "learning_rate": 5.595262927716724e-06, "loss": 0.0002, "num_input_tokens_seen": 60268400, "step": 31315 }, { "epoch": 5.109307447589526, "grad_norm": 0.02646021917462349, "learning_rate": 5.589074507656561e-06, "loss": 0.0044, "num_input_tokens_seen": 60278784, "step": 31320 }, { "epoch": 5.110123174810344, "grad_norm": 0.09445858746767044, "learning_rate": 5.582889081005044e-06, "loss": 0.0006, "num_input_tokens_seen": 60289472, "step": 31325 }, { "epoch": 5.110938902031161, "grad_norm": 27.14116096496582, "learning_rate": 5.5767066487160316e-06, "loss": 0.113, "num_input_tokens_seen": 60298432, "step": 31330 }, { "epoch": 5.111754629251978, "grad_norm": 0.06367482990026474, "learning_rate": 5.570527211742949e-06, "loss": 0.0004, "num_input_tokens_seen": 60308352, "step": 31335 }, { "epoch": 5.112570356472795, "grad_norm": 0.004792692139744759, "learning_rate": 5.564350771038731e-06, "loss": 0.0001, "num_input_tokens_seen": 60317600, "step": 31340 }, { "epoch": 5.113386083693613, "grad_norm": 0.5919497609138489, "learning_rate": 5.558177327555875e-06, "loss": 0.0006, "num_input_tokens_seen": 60327232, "step": 31345 }, { "epoch": 5.11420181091443, "grad_norm": 0.20375493168830872, "learning_rate": 5.552006882246388e-06, "loss": 0.0004, "num_input_tokens_seen": 60337664, "step": 31350 }, { "epoch": 5.115017538135247, "grad_norm": 23.40076446533203, "learning_rate": 5.545839436061839e-06, "loss": 0.0974, "num_input_tokens_seen": 60347760, "step": 31355 }, { "epoch": 5.1158332653560645, "grad_norm": 0.032386261969804764, "learning_rate": 5.539674989953331e-06, "loss": 0.027, "num_input_tokens_seen": 60358272, "step": 31360 }, { "epoch": 5.1166489925768825, "grad_norm": 0.006595219485461712, "learning_rate": 5.533513544871488e-06, "loss": 0.0003, "num_input_tokens_seen": 60368208, "step": 31365 }, { "epoch": 5.1174647197977, "grad_norm": 18.554197311401367, "learning_rate": 5.527355101766493e-06, "loss": 0.0064, "num_input_tokens_seen": 60377216, "step": 31370 }, { "epoch": 5.118280447018517, "grad_norm": 21.046842575073242, "learning_rate": 5.521199661588044e-06, "loss": 0.1457, "num_input_tokens_seen": 60385888, "step": 31375 }, { "epoch": 5.119096174239334, "grad_norm": 0.12710025906562805, "learning_rate": 5.5150472252853944e-06, "loss": 0.0068, "num_input_tokens_seen": 60395456, "step": 31380 }, { "epoch": 5.119911901460152, "grad_norm": 0.15666991472244263, "learning_rate": 5.50889779380733e-06, "loss": 0.0416, "num_input_tokens_seen": 60404944, "step": 31385 }, { "epoch": 5.120727628680969, "grad_norm": 0.1276642084121704, "learning_rate": 5.5027513681021605e-06, "loss": 0.0002, "num_input_tokens_seen": 60414816, "step": 31390 }, { "epoch": 5.121543355901786, "grad_norm": 0.0001097495696740225, "learning_rate": 5.4966079491177545e-06, "loss": 0.0004, "num_input_tokens_seen": 60424448, "step": 31395 }, { "epoch": 5.122359083122603, "grad_norm": 0.005146495532244444, "learning_rate": 5.490467537801491e-06, "loss": 0.0002, "num_input_tokens_seen": 60433680, "step": 31400 }, { "epoch": 5.122359083122603, "eval_loss": 0.34602808952331543, "eval_runtime": 80.7496, "eval_samples_per_second": 33.746, "eval_steps_per_second": 16.879, "num_input_tokens_seen": 60433680, "step": 31400 }, { "epoch": 5.123174810343421, "grad_norm": 0.001180388149805367, "learning_rate": 5.484330135100313e-06, "loss": 0.0001, "num_input_tokens_seen": 60442608, "step": 31405 }, { "epoch": 5.1239905375642385, "grad_norm": 0.01248217560350895, "learning_rate": 5.4781957419606785e-06, "loss": 0.0243, "num_input_tokens_seen": 60452880, "step": 31410 }, { "epoch": 5.124806264785056, "grad_norm": 0.005029503721743822, "learning_rate": 5.472064359328577e-06, "loss": 0.0001, "num_input_tokens_seen": 60463152, "step": 31415 }, { "epoch": 5.125621992005874, "grad_norm": 0.028627023100852966, "learning_rate": 5.4659359881495565e-06, "loss": 0.0001, "num_input_tokens_seen": 60473280, "step": 31420 }, { "epoch": 5.126437719226691, "grad_norm": 0.004175513982772827, "learning_rate": 5.4598106293686916e-06, "loss": 0.0001, "num_input_tokens_seen": 60482816, "step": 31425 }, { "epoch": 5.127253446447508, "grad_norm": 6.0385260581970215, "learning_rate": 5.45368828393058e-06, "loss": 0.0571, "num_input_tokens_seen": 60490704, "step": 31430 }, { "epoch": 5.128069173668325, "grad_norm": 0.313060998916626, "learning_rate": 5.44756895277937e-06, "loss": 0.0006, "num_input_tokens_seen": 60500032, "step": 31435 }, { "epoch": 5.128884900889143, "grad_norm": 0.02029760554432869, "learning_rate": 5.441452636858746e-06, "loss": 0.121, "num_input_tokens_seen": 60508752, "step": 31440 }, { "epoch": 5.12970062810996, "grad_norm": 0.006272973958402872, "learning_rate": 5.435339337111905e-06, "loss": 0.1349, "num_input_tokens_seen": 60518352, "step": 31445 }, { "epoch": 5.130516355330777, "grad_norm": 0.0012091598473489285, "learning_rate": 5.42922905448161e-06, "loss": 0.0006, "num_input_tokens_seen": 60527248, "step": 31450 }, { "epoch": 5.1313320825515945, "grad_norm": 0.0005065032746642828, "learning_rate": 5.423121789910129e-06, "loss": 0.0457, "num_input_tokens_seen": 60535472, "step": 31455 }, { "epoch": 5.1321478097724125, "grad_norm": 0.008354428224265575, "learning_rate": 5.417017544339287e-06, "loss": 0.0005, "num_input_tokens_seen": 60542848, "step": 31460 }, { "epoch": 5.13296353699323, "grad_norm": 0.7439097166061401, "learning_rate": 5.410916318710443e-06, "loss": 0.0007, "num_input_tokens_seen": 60552976, "step": 31465 }, { "epoch": 5.133779264214047, "grad_norm": 0.003909389488399029, "learning_rate": 5.404818113964466e-06, "loss": 0.0236, "num_input_tokens_seen": 60562144, "step": 31470 }, { "epoch": 5.134594991434864, "grad_norm": 0.007251961622387171, "learning_rate": 5.398722931041792e-06, "loss": 0.0004, "num_input_tokens_seen": 60572768, "step": 31475 }, { "epoch": 5.135410718655682, "grad_norm": 0.5276185870170593, "learning_rate": 5.392630770882367e-06, "loss": 0.0008, "num_input_tokens_seen": 60583472, "step": 31480 }, { "epoch": 5.136226445876499, "grad_norm": 0.041863374412059784, "learning_rate": 5.3865416344256705e-06, "loss": 0.0001, "num_input_tokens_seen": 60592832, "step": 31485 }, { "epoch": 5.137042173097316, "grad_norm": 0.0211684200912714, "learning_rate": 5.380455522610742e-06, "loss": 0.0002, "num_input_tokens_seen": 60601184, "step": 31490 }, { "epoch": 5.137857900318133, "grad_norm": 8.357906341552734, "learning_rate": 5.374372436376116e-06, "loss": 0.0417, "num_input_tokens_seen": 60610512, "step": 31495 }, { "epoch": 5.138673627538951, "grad_norm": 0.002328182104974985, "learning_rate": 5.368292376659895e-06, "loss": 0.11, "num_input_tokens_seen": 60620048, "step": 31500 }, { "epoch": 5.139489354759768, "grad_norm": 0.006947928573936224, "learning_rate": 5.362215344399701e-06, "loss": 0.0001, "num_input_tokens_seen": 60628768, "step": 31505 }, { "epoch": 5.1403050819805856, "grad_norm": 0.19723552465438843, "learning_rate": 5.356141340532678e-06, "loss": 0.0006, "num_input_tokens_seen": 60638416, "step": 31510 }, { "epoch": 5.141120809201403, "grad_norm": 0.0045617143623530865, "learning_rate": 5.350070365995522e-06, "loss": 0.0001, "num_input_tokens_seen": 60648224, "step": 31515 }, { "epoch": 5.141936536422221, "grad_norm": 0.09247495234012604, "learning_rate": 5.344002421724459e-06, "loss": 0.0009, "num_input_tokens_seen": 60658624, "step": 31520 }, { "epoch": 5.142752263643038, "grad_norm": 0.031111828982830048, "learning_rate": 5.337937508655228e-06, "loss": 0.0001, "num_input_tokens_seen": 60667984, "step": 31525 }, { "epoch": 5.143567990863855, "grad_norm": 0.15885606408119202, "learning_rate": 5.331875627723126e-06, "loss": 0.0003, "num_input_tokens_seen": 60676800, "step": 31530 }, { "epoch": 5.144383718084672, "grad_norm": 0.0007702323491685092, "learning_rate": 5.325816779862963e-06, "loss": 0.0, "num_input_tokens_seen": 60686928, "step": 31535 }, { "epoch": 5.14519944530549, "grad_norm": 0.03233535215258598, "learning_rate": 5.319760966009102e-06, "loss": 0.0001, "num_input_tokens_seen": 60696048, "step": 31540 }, { "epoch": 5.146015172526307, "grad_norm": 0.046683989465236664, "learning_rate": 5.3137081870954096e-06, "loss": 0.0003, "num_input_tokens_seen": 60706400, "step": 31545 }, { "epoch": 5.146830899747124, "grad_norm": 0.16779664158821106, "learning_rate": 5.307658444055313e-06, "loss": 0.1213, "num_input_tokens_seen": 60716288, "step": 31550 }, { "epoch": 5.1476466269679415, "grad_norm": 0.003246632171794772, "learning_rate": 5.301611737821749e-06, "loss": 0.0009, "num_input_tokens_seen": 60726256, "step": 31555 }, { "epoch": 5.1484623541887595, "grad_norm": 0.03836369886994362, "learning_rate": 5.295568069327206e-06, "loss": 0.0001, "num_input_tokens_seen": 60735360, "step": 31560 }, { "epoch": 5.149278081409577, "grad_norm": 0.008251724764704704, "learning_rate": 5.289527439503683e-06, "loss": 0.0002, "num_input_tokens_seen": 60744432, "step": 31565 }, { "epoch": 5.150093808630394, "grad_norm": 0.0013715712120756507, "learning_rate": 5.28348984928273e-06, "loss": 0.0002, "num_input_tokens_seen": 60753536, "step": 31570 }, { "epoch": 5.150909535851211, "grad_norm": 0.13735976815223694, "learning_rate": 5.27745529959541e-06, "loss": 0.0004, "num_input_tokens_seen": 60764112, "step": 31575 }, { "epoch": 5.151725263072029, "grad_norm": 0.07265564799308777, "learning_rate": 5.271423791372335e-06, "loss": 0.0002, "num_input_tokens_seen": 60772080, "step": 31580 }, { "epoch": 5.152540990292846, "grad_norm": 0.00021953573741484433, "learning_rate": 5.26539532554364e-06, "loss": 0.0001, "num_input_tokens_seen": 60781712, "step": 31585 }, { "epoch": 5.153356717513663, "grad_norm": 3.286869525909424, "learning_rate": 5.25936990303898e-06, "loss": 0.0109, "num_input_tokens_seen": 60791104, "step": 31590 }, { "epoch": 5.154172444734481, "grad_norm": 0.05356381833553314, "learning_rate": 5.253347524787555e-06, "loss": 0.0005, "num_input_tokens_seen": 60800720, "step": 31595 }, { "epoch": 5.154988171955298, "grad_norm": 0.0255911685526371, "learning_rate": 5.2473281917181035e-06, "loss": 0.0001, "num_input_tokens_seen": 60809376, "step": 31600 }, { "epoch": 5.154988171955298, "eval_loss": 0.3556883633136749, "eval_runtime": 80.7919, "eval_samples_per_second": 33.729, "eval_steps_per_second": 16.871, "num_input_tokens_seen": 60809376, "step": 31600 }, { "epoch": 5.1558038991761155, "grad_norm": 0.008608200587332249, "learning_rate": 5.241311904758864e-06, "loss": 0.0, "num_input_tokens_seen": 60819392, "step": 31605 }, { "epoch": 5.156619626396933, "grad_norm": 0.01548311673104763, "learning_rate": 5.23529866483764e-06, "loss": 0.0001, "num_input_tokens_seen": 60827888, "step": 31610 }, { "epoch": 5.157435353617751, "grad_norm": 0.0016424654750153422, "learning_rate": 5.229288472881732e-06, "loss": 0.0002, "num_input_tokens_seen": 60836384, "step": 31615 }, { "epoch": 5.158251080838568, "grad_norm": 0.0005490146577358246, "learning_rate": 5.2232813298180025e-06, "loss": 0.0, "num_input_tokens_seen": 60845296, "step": 31620 }, { "epoch": 5.159066808059385, "grad_norm": 0.0009154727449640632, "learning_rate": 5.217277236572824e-06, "loss": 0.0011, "num_input_tokens_seen": 60854768, "step": 31625 }, { "epoch": 5.159882535280202, "grad_norm": 0.05769462138414383, "learning_rate": 5.211276194072093e-06, "loss": 0.0412, "num_input_tokens_seen": 60864208, "step": 31630 }, { "epoch": 5.16069826250102, "grad_norm": 0.001404673676006496, "learning_rate": 5.205278203241254e-06, "loss": 0.0002, "num_input_tokens_seen": 60874400, "step": 31635 }, { "epoch": 5.161513989721837, "grad_norm": 0.13272343575954437, "learning_rate": 5.199283265005278e-06, "loss": 0.0002, "num_input_tokens_seen": 60884176, "step": 31640 }, { "epoch": 5.162329716942654, "grad_norm": 0.0036563489120453596, "learning_rate": 5.193291380288648e-06, "loss": 0.0009, "num_input_tokens_seen": 60894448, "step": 31645 }, { "epoch": 5.163145444163471, "grad_norm": 0.031116578727960587, "learning_rate": 5.1873025500153995e-06, "loss": 0.0004, "num_input_tokens_seen": 60904816, "step": 31650 }, { "epoch": 5.1639611713842895, "grad_norm": 0.003918944858014584, "learning_rate": 5.181316775109071e-06, "loss": 0.0003, "num_input_tokens_seen": 60914416, "step": 31655 }, { "epoch": 5.164776898605107, "grad_norm": 0.005959631875157356, "learning_rate": 5.1753340564927564e-06, "loss": 0.0002, "num_input_tokens_seen": 60922800, "step": 31660 }, { "epoch": 5.165592625825924, "grad_norm": 0.040252115577459335, "learning_rate": 5.169354395089068e-06, "loss": 0.0007, "num_input_tokens_seen": 60932432, "step": 31665 }, { "epoch": 5.166408353046741, "grad_norm": 0.0463593564927578, "learning_rate": 5.1633777918201346e-06, "loss": 0.0006, "num_input_tokens_seen": 60942816, "step": 31670 }, { "epoch": 5.167224080267559, "grad_norm": 0.0005338263581506908, "learning_rate": 5.157404247607625e-06, "loss": 0.0001, "num_input_tokens_seen": 60951264, "step": 31675 }, { "epoch": 5.168039807488376, "grad_norm": 0.015120682306587696, "learning_rate": 5.1514337633727454e-06, "loss": 0.0008, "num_input_tokens_seen": 60959440, "step": 31680 }, { "epoch": 5.168855534709193, "grad_norm": 0.0014297321904450655, "learning_rate": 5.145466340036206e-06, "loss": 0.0, "num_input_tokens_seen": 60969552, "step": 31685 }, { "epoch": 5.16967126193001, "grad_norm": 0.26616552472114563, "learning_rate": 5.139501978518274e-06, "loss": 0.0003, "num_input_tokens_seen": 60980432, "step": 31690 }, { "epoch": 5.170486989150828, "grad_norm": 2.349416971206665, "learning_rate": 5.133540679738716e-06, "loss": 0.0012, "num_input_tokens_seen": 60988704, "step": 31695 }, { "epoch": 5.171302716371645, "grad_norm": 0.014238925650715828, "learning_rate": 5.127582444616838e-06, "loss": 0.0002, "num_input_tokens_seen": 60997536, "step": 31700 }, { "epoch": 5.1721184435924625, "grad_norm": 0.049626484513282776, "learning_rate": 5.121627274071486e-06, "loss": 0.0856, "num_input_tokens_seen": 61007856, "step": 31705 }, { "epoch": 5.17293417081328, "grad_norm": 0.002059418009594083, "learning_rate": 5.115675169021009e-06, "loss": 0.0002, "num_input_tokens_seen": 61017600, "step": 31710 }, { "epoch": 5.173749898034098, "grad_norm": 0.012166374363005161, "learning_rate": 5.1097261303832994e-06, "loss": 0.0002, "num_input_tokens_seen": 61027568, "step": 31715 }, { "epoch": 5.174565625254915, "grad_norm": 96.7936782836914, "learning_rate": 5.103780159075788e-06, "loss": 0.0322, "num_input_tokens_seen": 61036672, "step": 31720 }, { "epoch": 5.175381352475732, "grad_norm": 0.004624402616173029, "learning_rate": 5.0978372560154e-06, "loss": 0.0005, "num_input_tokens_seen": 61045232, "step": 31725 }, { "epoch": 5.176197079696549, "grad_norm": 0.0018520912854000926, "learning_rate": 5.091897422118619e-06, "loss": 0.0, "num_input_tokens_seen": 61055024, "step": 31730 }, { "epoch": 5.177012806917367, "grad_norm": 0.013141282834112644, "learning_rate": 5.0859606583014305e-06, "loss": 0.0012, "num_input_tokens_seen": 61064624, "step": 31735 }, { "epoch": 5.177828534138184, "grad_norm": 0.07051275670528412, "learning_rate": 5.080026965479365e-06, "loss": 0.0002, "num_input_tokens_seen": 61074624, "step": 31740 }, { "epoch": 5.178644261359001, "grad_norm": 0.3106083869934082, "learning_rate": 5.074096344567475e-06, "loss": 0.0006, "num_input_tokens_seen": 61082928, "step": 31745 }, { "epoch": 5.1794599885798185, "grad_norm": 0.008861517533659935, "learning_rate": 5.0681687964803294e-06, "loss": 0.0337, "num_input_tokens_seen": 61091632, "step": 31750 }, { "epoch": 5.1802757158006365, "grad_norm": 11.43989086151123, "learning_rate": 5.06224432213204e-06, "loss": 0.0489, "num_input_tokens_seen": 61099520, "step": 31755 }, { "epoch": 5.181091443021454, "grad_norm": 0.0034374231472611427, "learning_rate": 5.056322922436224e-06, "loss": 0.0001, "num_input_tokens_seen": 61108784, "step": 31760 }, { "epoch": 5.181907170242271, "grad_norm": 0.01702452078461647, "learning_rate": 5.0504045983060465e-06, "loss": 0.0355, "num_input_tokens_seen": 61117712, "step": 31765 }, { "epoch": 5.182722897463089, "grad_norm": 0.010423209518194199, "learning_rate": 5.044489350654183e-06, "loss": 0.0576, "num_input_tokens_seen": 61126384, "step": 31770 }, { "epoch": 5.183538624683906, "grad_norm": 38.047813415527344, "learning_rate": 5.038577180392831e-06, "loss": 0.0414, "num_input_tokens_seen": 61135744, "step": 31775 }, { "epoch": 5.184354351904723, "grad_norm": 0.020376551896333694, "learning_rate": 5.032668088433729e-06, "loss": 0.0001, "num_input_tokens_seen": 61144864, "step": 31780 }, { "epoch": 5.18517007912554, "grad_norm": 4.8226423263549805, "learning_rate": 5.02676207568814e-06, "loss": 0.1322, "num_input_tokens_seen": 61154688, "step": 31785 }, { "epoch": 5.185985806346358, "grad_norm": 0.040360480546951294, "learning_rate": 5.02085914306683e-06, "loss": 0.0119, "num_input_tokens_seen": 61165184, "step": 31790 }, { "epoch": 5.186801533567175, "grad_norm": 0.007339680101722479, "learning_rate": 5.014959291480123e-06, "loss": 0.0001, "num_input_tokens_seen": 61176512, "step": 31795 }, { "epoch": 5.1876172607879925, "grad_norm": 0.1397785097360611, "learning_rate": 5.009062521837835e-06, "loss": 0.0002, "num_input_tokens_seen": 61186608, "step": 31800 }, { "epoch": 5.1876172607879925, "eval_loss": 0.35577070713043213, "eval_runtime": 80.7611, "eval_samples_per_second": 33.741, "eval_steps_per_second": 16.877, "num_input_tokens_seen": 61186608, "step": 31800 }, { "epoch": 5.18843298800881, "grad_norm": 1.9541652202606201, "learning_rate": 5.003168835049324e-06, "loss": 0.0008, "num_input_tokens_seen": 61196400, "step": 31805 }, { "epoch": 5.189248715229628, "grad_norm": 0.05294448882341385, "learning_rate": 4.997278232023483e-06, "loss": 0.0002, "num_input_tokens_seen": 61205504, "step": 31810 }, { "epoch": 5.190064442450445, "grad_norm": 0.002193066058680415, "learning_rate": 4.9913907136687036e-06, "loss": 0.0001, "num_input_tokens_seen": 61214400, "step": 31815 }, { "epoch": 5.190880169671262, "grad_norm": 0.04312419146299362, "learning_rate": 4.985506280892918e-06, "loss": 0.0356, "num_input_tokens_seen": 61223904, "step": 31820 }, { "epoch": 5.191695896892079, "grad_norm": 0.00527886301279068, "learning_rate": 4.979624934603589e-06, "loss": 0.0017, "num_input_tokens_seen": 61232928, "step": 31825 }, { "epoch": 5.192511624112897, "grad_norm": 0.004812545143067837, "learning_rate": 4.97374667570768e-06, "loss": 0.0002, "num_input_tokens_seen": 61242816, "step": 31830 }, { "epoch": 5.193327351333714, "grad_norm": 0.0751553475856781, "learning_rate": 4.967871505111704e-06, "loss": 0.0022, "num_input_tokens_seen": 61252512, "step": 31835 }, { "epoch": 5.194143078554531, "grad_norm": 0.7826909422874451, "learning_rate": 4.961999423721686e-06, "loss": 0.0002, "num_input_tokens_seen": 61262016, "step": 31840 }, { "epoch": 5.194958805775348, "grad_norm": 0.03573130443692207, "learning_rate": 4.956130432443159e-06, "loss": 0.0002, "num_input_tokens_seen": 61271904, "step": 31845 }, { "epoch": 5.195774532996166, "grad_norm": 0.16864635050296783, "learning_rate": 4.950264532181215e-06, "loss": 0.0005, "num_input_tokens_seen": 61279952, "step": 31850 }, { "epoch": 5.196590260216984, "grad_norm": 0.030119972303509712, "learning_rate": 4.944401723840433e-06, "loss": 0.1338, "num_input_tokens_seen": 61290272, "step": 31855 }, { "epoch": 5.197405987437801, "grad_norm": 0.05130394548177719, "learning_rate": 4.938542008324942e-06, "loss": 0.0002, "num_input_tokens_seen": 61299264, "step": 31860 }, { "epoch": 5.198221714658618, "grad_norm": 0.003444673726335168, "learning_rate": 4.9326853865383855e-06, "loss": 0.0002, "num_input_tokens_seen": 61308032, "step": 31865 }, { "epoch": 5.199037441879436, "grad_norm": 0.03635618835687637, "learning_rate": 4.926831859383918e-06, "loss": 0.0066, "num_input_tokens_seen": 61318000, "step": 31870 }, { "epoch": 5.199853169100253, "grad_norm": 0.3084987998008728, "learning_rate": 4.92098142776424e-06, "loss": 0.0009, "num_input_tokens_seen": 61327072, "step": 31875 }, { "epoch": 5.20066889632107, "grad_norm": 0.1060485914349556, "learning_rate": 4.91513409258155e-06, "loss": 0.0031, "num_input_tokens_seen": 61337456, "step": 31880 }, { "epoch": 5.201484623541887, "grad_norm": 0.001194066135212779, "learning_rate": 4.909289854737581e-06, "loss": 0.0003, "num_input_tokens_seen": 61345952, "step": 31885 }, { "epoch": 5.202300350762705, "grad_norm": 0.0016710680210962892, "learning_rate": 4.903448715133602e-06, "loss": 0.0001, "num_input_tokens_seen": 61355584, "step": 31890 }, { "epoch": 5.203116077983522, "grad_norm": 0.0068269274197518826, "learning_rate": 4.897610674670372e-06, "loss": 0.0003, "num_input_tokens_seen": 61364512, "step": 31895 }, { "epoch": 5.2039318052043395, "grad_norm": 0.00483560748398304, "learning_rate": 4.8917757342482e-06, "loss": 0.0305, "num_input_tokens_seen": 61373072, "step": 31900 }, { "epoch": 5.204747532425157, "grad_norm": 0.4956607520580292, "learning_rate": 4.885943894766909e-06, "loss": 0.0009, "num_input_tokens_seen": 61382768, "step": 31905 }, { "epoch": 5.205563259645975, "grad_norm": 0.00323199643753469, "learning_rate": 4.880115157125842e-06, "loss": 0.0569, "num_input_tokens_seen": 61393824, "step": 31910 }, { "epoch": 5.206378986866792, "grad_norm": 0.19240090250968933, "learning_rate": 4.874289522223857e-06, "loss": 0.0004, "num_input_tokens_seen": 61403952, "step": 31915 }, { "epoch": 5.207194714087609, "grad_norm": 9.819807052612305, "learning_rate": 4.868466990959339e-06, "loss": 0.027, "num_input_tokens_seen": 61412896, "step": 31920 }, { "epoch": 5.208010441308426, "grad_norm": 13.415438652038574, "learning_rate": 4.8626475642301964e-06, "loss": 0.1009, "num_input_tokens_seen": 61422160, "step": 31925 }, { "epoch": 5.208826168529244, "grad_norm": 0.007434983272105455, "learning_rate": 4.856831242933871e-06, "loss": 0.0004, "num_input_tokens_seen": 61433728, "step": 31930 }, { "epoch": 5.209641895750061, "grad_norm": 0.0013489011907950044, "learning_rate": 4.851018027967294e-06, "loss": 0.0001, "num_input_tokens_seen": 61443136, "step": 31935 }, { "epoch": 5.210457622970878, "grad_norm": 0.00497178640216589, "learning_rate": 4.845207920226946e-06, "loss": 0.0001, "num_input_tokens_seen": 61453088, "step": 31940 }, { "epoch": 5.211273350191696, "grad_norm": 0.014202209189534187, "learning_rate": 4.839400920608825e-06, "loss": 0.0002, "num_input_tokens_seen": 61462992, "step": 31945 }, { "epoch": 5.2120890774125135, "grad_norm": 0.02629457786679268, "learning_rate": 4.83359703000843e-06, "loss": 0.0043, "num_input_tokens_seen": 61472608, "step": 31950 }, { "epoch": 5.212904804633331, "grad_norm": 0.0025140675716102123, "learning_rate": 4.827796249320804e-06, "loss": 0.1025, "num_input_tokens_seen": 61482896, "step": 31955 }, { "epoch": 5.213720531854148, "grad_norm": 0.1737983226776123, "learning_rate": 4.82199857944049e-06, "loss": 0.0004, "num_input_tokens_seen": 61491088, "step": 31960 }, { "epoch": 5.214536259074965, "grad_norm": 0.008348588831722736, "learning_rate": 4.8162040212615695e-06, "loss": 0.0002, "num_input_tokens_seen": 61500880, "step": 31965 }, { "epoch": 5.215351986295783, "grad_norm": 0.03195154666900635, "learning_rate": 4.810412575677639e-06, "loss": 0.0002, "num_input_tokens_seen": 61510848, "step": 31970 }, { "epoch": 5.2161677135166, "grad_norm": 0.0018608031095936894, "learning_rate": 4.804624243581801e-06, "loss": 0.0498, "num_input_tokens_seen": 61520192, "step": 31975 }, { "epoch": 5.216983440737417, "grad_norm": 0.004157386254519224, "learning_rate": 4.798839025866703e-06, "loss": 0.0007, "num_input_tokens_seen": 61530480, "step": 31980 }, { "epoch": 5.217799167958235, "grad_norm": 0.004353375174105167, "learning_rate": 4.793056923424491e-06, "loss": 0.0107, "num_input_tokens_seen": 61538704, "step": 31985 }, { "epoch": 5.218614895179052, "grad_norm": 0.010558788664638996, "learning_rate": 4.78727793714683e-06, "loss": 0.0001, "num_input_tokens_seen": 61547792, "step": 31990 }, { "epoch": 5.219430622399869, "grad_norm": 0.012029417790472507, "learning_rate": 4.7815020679249285e-06, "loss": 0.0107, "num_input_tokens_seen": 61556976, "step": 31995 }, { "epoch": 5.220246349620687, "grad_norm": 0.02164536528289318, "learning_rate": 4.775729316649483e-06, "loss": 0.0001, "num_input_tokens_seen": 61567504, "step": 32000 }, { "epoch": 5.220246349620687, "eval_loss": 0.3639886975288391, "eval_runtime": 80.7134, "eval_samples_per_second": 33.761, "eval_steps_per_second": 16.887, "num_input_tokens_seen": 61567504, "step": 32000 }, { "epoch": 5.221062076841505, "grad_norm": 0.05325957387685776, "learning_rate": 4.769959684210728e-06, "loss": 0.0007, "num_input_tokens_seen": 61576880, "step": 32005 }, { "epoch": 5.221877804062322, "grad_norm": 1.292445182800293, "learning_rate": 4.764193171498426e-06, "loss": 0.0025, "num_input_tokens_seen": 61587680, "step": 32010 }, { "epoch": 5.222693531283139, "grad_norm": 0.24046987295150757, "learning_rate": 4.75842977940183e-06, "loss": 0.047, "num_input_tokens_seen": 61597088, "step": 32015 }, { "epoch": 5.223509258503956, "grad_norm": 0.010601782239973545, "learning_rate": 4.752669508809729e-06, "loss": 0.0002, "num_input_tokens_seen": 61607760, "step": 32020 }, { "epoch": 5.224324985724774, "grad_norm": 0.0029557007364928722, "learning_rate": 4.746912360610445e-06, "loss": 0.0002, "num_input_tokens_seen": 61617840, "step": 32025 }, { "epoch": 5.225140712945591, "grad_norm": 0.027795694768428802, "learning_rate": 4.741158335691781e-06, "loss": 0.0284, "num_input_tokens_seen": 61627232, "step": 32030 }, { "epoch": 5.225956440166408, "grad_norm": 0.0005824988475069404, "learning_rate": 4.7354074349410994e-06, "loss": 0.0003, "num_input_tokens_seen": 61636512, "step": 32035 }, { "epoch": 5.226772167387225, "grad_norm": 0.0024116470012813807, "learning_rate": 4.729659659245245e-06, "loss": 0.0001, "num_input_tokens_seen": 61647504, "step": 32040 }, { "epoch": 5.227587894608043, "grad_norm": 0.0635918378829956, "learning_rate": 4.723915009490601e-06, "loss": 0.0002, "num_input_tokens_seen": 61656528, "step": 32045 }, { "epoch": 5.2284036218288605, "grad_norm": 0.01668734848499298, "learning_rate": 4.718173486563077e-06, "loss": 0.1338, "num_input_tokens_seen": 61666336, "step": 32050 }, { "epoch": 5.229219349049678, "grad_norm": 0.066070057451725, "learning_rate": 4.71243509134808e-06, "loss": 0.0754, "num_input_tokens_seen": 61675888, "step": 32055 }, { "epoch": 5.230035076270495, "grad_norm": 0.00562182255089283, "learning_rate": 4.706699824730532e-06, "loss": 0.0589, "num_input_tokens_seen": 61685920, "step": 32060 }, { "epoch": 5.230850803491313, "grad_norm": 0.008274692110717297, "learning_rate": 4.700967687594901e-06, "loss": 0.1147, "num_input_tokens_seen": 61696720, "step": 32065 }, { "epoch": 5.23166653071213, "grad_norm": 0.003936741966754198, "learning_rate": 4.69523868082514e-06, "loss": 0.0002, "num_input_tokens_seen": 61707632, "step": 32070 }, { "epoch": 5.232482257932947, "grad_norm": 0.09083535522222519, "learning_rate": 4.689512805304747e-06, "loss": 0.0002, "num_input_tokens_seen": 61717408, "step": 32075 }, { "epoch": 5.233297985153764, "grad_norm": 0.6071617007255554, "learning_rate": 4.683790061916707e-06, "loss": 0.0673, "num_input_tokens_seen": 61727232, "step": 32080 }, { "epoch": 5.234113712374582, "grad_norm": 0.005463482346385717, "learning_rate": 4.678070451543551e-06, "loss": 0.0066, "num_input_tokens_seen": 61737648, "step": 32085 }, { "epoch": 5.234929439595399, "grad_norm": 0.005753290839493275, "learning_rate": 4.6723539750673204e-06, "loss": 0.0001, "num_input_tokens_seen": 61746640, "step": 32090 }, { "epoch": 5.2357451668162165, "grad_norm": 0.002765883458778262, "learning_rate": 4.666640633369551e-06, "loss": 0.0004, "num_input_tokens_seen": 61755536, "step": 32095 }, { "epoch": 5.236560894037034, "grad_norm": 0.03238053247332573, "learning_rate": 4.660930427331323e-06, "loss": 0.0001, "num_input_tokens_seen": 61765712, "step": 32100 }, { "epoch": 5.237376621257852, "grad_norm": 0.0043946364894509315, "learning_rate": 4.6552233578332244e-06, "loss": 0.0001, "num_input_tokens_seen": 61776784, "step": 32105 }, { "epoch": 5.238192348478669, "grad_norm": 0.011095008812844753, "learning_rate": 4.649519425755347e-06, "loss": 0.0166, "num_input_tokens_seen": 61786336, "step": 32110 }, { "epoch": 5.239008075699486, "grad_norm": 0.010188309475779533, "learning_rate": 4.64381863197732e-06, "loss": 0.0006, "num_input_tokens_seen": 61795328, "step": 32115 }, { "epoch": 5.239823802920303, "grad_norm": 0.0004541216476354748, "learning_rate": 4.638120977378269e-06, "loss": 0.0589, "num_input_tokens_seen": 61805600, "step": 32120 }, { "epoch": 5.240639530141121, "grad_norm": 0.08475473523139954, "learning_rate": 4.632426462836848e-06, "loss": 0.0611, "num_input_tokens_seen": 61815056, "step": 32125 }, { "epoch": 5.241455257361938, "grad_norm": 0.004806050099432468, "learning_rate": 4.626735089231224e-06, "loss": 0.0002, "num_input_tokens_seen": 61825024, "step": 32130 }, { "epoch": 5.242270984582755, "grad_norm": 8.129851341247559, "learning_rate": 4.621046857439068e-06, "loss": 0.1493, "num_input_tokens_seen": 61834544, "step": 32135 }, { "epoch": 5.243086711803572, "grad_norm": 0.017049185931682587, "learning_rate": 4.615361768337587e-06, "loss": 0.0001, "num_input_tokens_seen": 61844816, "step": 32140 }, { "epoch": 5.2439024390243905, "grad_norm": 0.0057350336574018, "learning_rate": 4.6096798228034946e-06, "loss": 0.0006, "num_input_tokens_seen": 61854400, "step": 32145 }, { "epoch": 5.244718166245208, "grad_norm": 0.0006228458951227367, "learning_rate": 4.604001021713008e-06, "loss": 0.0, "num_input_tokens_seen": 61863824, "step": 32150 }, { "epoch": 5.245533893466025, "grad_norm": 0.03804752230644226, "learning_rate": 4.598325365941883e-06, "loss": 0.0082, "num_input_tokens_seen": 61873152, "step": 32155 }, { "epoch": 5.246349620686843, "grad_norm": 16.21784019470215, "learning_rate": 4.5926528563653645e-06, "loss": 0.0629, "num_input_tokens_seen": 61882480, "step": 32160 }, { "epoch": 5.24716534790766, "grad_norm": 0.03372616693377495, "learning_rate": 4.5869834938582295e-06, "loss": 0.0001, "num_input_tokens_seen": 61893024, "step": 32165 }, { "epoch": 5.247981075128477, "grad_norm": 0.10133808106184006, "learning_rate": 4.581317279294772e-06, "loss": 0.0031, "num_input_tokens_seen": 61902208, "step": 32170 }, { "epoch": 5.248796802349294, "grad_norm": 0.03072618506848812, "learning_rate": 4.57565421354878e-06, "loss": 0.0168, "num_input_tokens_seen": 61910800, "step": 32175 }, { "epoch": 5.249612529570112, "grad_norm": 1.0557531118392944, "learning_rate": 4.569994297493579e-06, "loss": 0.0024, "num_input_tokens_seen": 61919776, "step": 32180 }, { "epoch": 5.250428256790929, "grad_norm": 0.010178273543715477, "learning_rate": 4.564337532002002e-06, "loss": 0.0008, "num_input_tokens_seen": 61930528, "step": 32185 }, { "epoch": 5.251243984011746, "grad_norm": 10.134964942932129, "learning_rate": 4.55868391794638e-06, "loss": 0.0715, "num_input_tokens_seen": 61940416, "step": 32190 }, { "epoch": 5.2520597112325635, "grad_norm": 0.012641996145248413, "learning_rate": 4.553033456198588e-06, "loss": 0.0047, "num_input_tokens_seen": 61949904, "step": 32195 }, { "epoch": 5.252875438453382, "grad_norm": 0.022256236523389816, "learning_rate": 4.54738614762999e-06, "loss": 0.0008, "num_input_tokens_seen": 61958976, "step": 32200 }, { "epoch": 5.252875438453382, "eval_loss": 0.3641124665737152, "eval_runtime": 80.8133, "eval_samples_per_second": 33.72, "eval_steps_per_second": 16.866, "num_input_tokens_seen": 61958976, "step": 32200 }, { "epoch": 5.253691165674199, "grad_norm": 0.02747037634253502, "learning_rate": 4.541741993111465e-06, "loss": 0.1192, "num_input_tokens_seen": 61967648, "step": 32205 }, { "epoch": 5.254506892895016, "grad_norm": 0.0028730365447700024, "learning_rate": 4.536100993513423e-06, "loss": 0.0712, "num_input_tokens_seen": 61975904, "step": 32210 }, { "epoch": 5.255322620115833, "grad_norm": 0.005925539415329695, "learning_rate": 4.530463149705768e-06, "loss": 0.0002, "num_input_tokens_seen": 61985200, "step": 32215 }, { "epoch": 5.256138347336651, "grad_norm": 0.025952132418751717, "learning_rate": 4.524828462557934e-06, "loss": 0.0001, "num_input_tokens_seen": 61995184, "step": 32220 }, { "epoch": 5.256954074557468, "grad_norm": 0.004850981757044792, "learning_rate": 4.5191969329388625e-06, "loss": 0.0003, "num_input_tokens_seen": 62005168, "step": 32225 }, { "epoch": 5.257769801778285, "grad_norm": 0.2970608174800873, "learning_rate": 4.5135685617169965e-06, "loss": 0.0003, "num_input_tokens_seen": 62015536, "step": 32230 }, { "epoch": 5.258585528999102, "grad_norm": 0.03423557057976723, "learning_rate": 4.507943349760313e-06, "loss": 0.0006, "num_input_tokens_seen": 62025520, "step": 32235 }, { "epoch": 5.25940125621992, "grad_norm": 9.608044624328613, "learning_rate": 4.502321297936277e-06, "loss": 0.1022, "num_input_tokens_seen": 62035488, "step": 32240 }, { "epoch": 5.2602169834407375, "grad_norm": 0.014089486561715603, "learning_rate": 4.496702407111888e-06, "loss": 0.1099, "num_input_tokens_seen": 62046032, "step": 32245 }, { "epoch": 5.261032710661555, "grad_norm": 0.14215205609798431, "learning_rate": 4.491086678153653e-06, "loss": 0.0001, "num_input_tokens_seen": 62054960, "step": 32250 }, { "epoch": 5.261848437882372, "grad_norm": 0.03170173987746239, "learning_rate": 4.485474111927579e-06, "loss": 0.0002, "num_input_tokens_seen": 62063744, "step": 32255 }, { "epoch": 5.26266416510319, "grad_norm": 0.008835511282086372, "learning_rate": 4.479864709299197e-06, "loss": 0.0919, "num_input_tokens_seen": 62075056, "step": 32260 }, { "epoch": 5.263479892324007, "grad_norm": 0.2061685025691986, "learning_rate": 4.474258471133555e-06, "loss": 0.0004, "num_input_tokens_seen": 62084224, "step": 32265 }, { "epoch": 5.264295619544824, "grad_norm": 0.005085598677396774, "learning_rate": 4.4686553982952014e-06, "loss": 0.0, "num_input_tokens_seen": 62094032, "step": 32270 }, { "epoch": 5.265111346765641, "grad_norm": 0.010521462187170982, "learning_rate": 4.463055491648191e-06, "loss": 0.0065, "num_input_tokens_seen": 62103312, "step": 32275 }, { "epoch": 5.265927073986459, "grad_norm": 0.005243151914328337, "learning_rate": 4.457458752056112e-06, "loss": 0.001, "num_input_tokens_seen": 62113152, "step": 32280 }, { "epoch": 5.266742801207276, "grad_norm": 0.016903940588235855, "learning_rate": 4.451865180382042e-06, "loss": 0.0002, "num_input_tokens_seen": 62122496, "step": 32285 }, { "epoch": 5.2675585284280935, "grad_norm": 0.002340092556551099, "learning_rate": 4.4462747774885936e-06, "loss": 0.123, "num_input_tokens_seen": 62132880, "step": 32290 }, { "epoch": 5.268374255648911, "grad_norm": 0.03518616035580635, "learning_rate": 4.440687544237859e-06, "loss": 0.0775, "num_input_tokens_seen": 62143488, "step": 32295 }, { "epoch": 5.269189982869729, "grad_norm": 0.012534652836620808, "learning_rate": 4.435103481491471e-06, "loss": 0.0481, "num_input_tokens_seen": 62151056, "step": 32300 }, { "epoch": 5.270005710090546, "grad_norm": 0.032477255910634995, "learning_rate": 4.429522590110569e-06, "loss": 0.0002, "num_input_tokens_seen": 62161104, "step": 32305 }, { "epoch": 5.270821437311363, "grad_norm": 0.041361913084983826, "learning_rate": 4.423944870955779e-06, "loss": 0.001, "num_input_tokens_seen": 62171408, "step": 32310 }, { "epoch": 5.27163716453218, "grad_norm": 0.0012467066990211606, "learning_rate": 4.418370324887272e-06, "loss": 0.0, "num_input_tokens_seen": 62180880, "step": 32315 }, { "epoch": 5.272452891752998, "grad_norm": 0.002477791626006365, "learning_rate": 4.412798952764699e-06, "loss": 0.0005, "num_input_tokens_seen": 62190240, "step": 32320 }, { "epoch": 5.273268618973815, "grad_norm": 0.006955038756132126, "learning_rate": 4.407230755447245e-06, "loss": 0.0002, "num_input_tokens_seen": 62200704, "step": 32325 }, { "epoch": 5.274084346194632, "grad_norm": 0.008691894821822643, "learning_rate": 4.401665733793598e-06, "loss": 0.0002, "num_input_tokens_seen": 62210320, "step": 32330 }, { "epoch": 5.27490007341545, "grad_norm": 0.0009038517018780112, "learning_rate": 4.3961038886619425e-06, "loss": 0.0002, "num_input_tokens_seen": 62219936, "step": 32335 }, { "epoch": 5.275715800636267, "grad_norm": 0.002432470442727208, "learning_rate": 4.39054522091e-06, "loss": 0.0001, "num_input_tokens_seen": 62230432, "step": 32340 }, { "epoch": 5.276531527857085, "grad_norm": 0.03569301962852478, "learning_rate": 4.384989731394979e-06, "loss": 0.001, "num_input_tokens_seen": 62239760, "step": 32345 }, { "epoch": 5.277347255077902, "grad_norm": 0.00402039522305131, "learning_rate": 4.379437420973598e-06, "loss": 0.0002, "num_input_tokens_seen": 62248064, "step": 32350 }, { "epoch": 5.27816298229872, "grad_norm": 0.00030607881490141153, "learning_rate": 4.373888290502107e-06, "loss": 0.0001, "num_input_tokens_seen": 62258896, "step": 32355 }, { "epoch": 5.278978709519537, "grad_norm": 0.08175937086343765, "learning_rate": 4.36834234083624e-06, "loss": 0.0002, "num_input_tokens_seen": 62269856, "step": 32360 }, { "epoch": 5.279794436740354, "grad_norm": 0.022769903764128685, "learning_rate": 4.362799572831258e-06, "loss": 0.0007, "num_input_tokens_seen": 62279824, "step": 32365 }, { "epoch": 5.280610163961171, "grad_norm": 0.07180264592170715, "learning_rate": 4.35725998734193e-06, "loss": 0.0021, "num_input_tokens_seen": 62288896, "step": 32370 }, { "epoch": 5.281425891181989, "grad_norm": 0.0013312598457559943, "learning_rate": 4.3517235852225195e-06, "loss": 0.0001, "num_input_tokens_seen": 62297312, "step": 32375 }, { "epoch": 5.282241618402806, "grad_norm": 0.015356938354671001, "learning_rate": 4.346190367326822e-06, "loss": 0.0, "num_input_tokens_seen": 62307136, "step": 32380 }, { "epoch": 5.283057345623623, "grad_norm": 0.119141586124897, "learning_rate": 4.340660334508115e-06, "loss": 0.0006, "num_input_tokens_seen": 62316320, "step": 32385 }, { "epoch": 5.2838730728444405, "grad_norm": 0.006686741951853037, "learning_rate": 4.335133487619206e-06, "loss": 0.0001, "num_input_tokens_seen": 62326288, "step": 32390 }, { "epoch": 5.2846888000652585, "grad_norm": 0.002165985992178321, "learning_rate": 4.329609827512409e-06, "loss": 0.0001, "num_input_tokens_seen": 62335936, "step": 32395 }, { "epoch": 5.285504527286076, "grad_norm": 0.0013928285334259272, "learning_rate": 4.324089355039531e-06, "loss": 0.0005, "num_input_tokens_seen": 62346176, "step": 32400 }, { "epoch": 5.285504527286076, "eval_loss": 0.36028480529785156, "eval_runtime": 80.7953, "eval_samples_per_second": 33.727, "eval_steps_per_second": 16.87, "num_input_tokens_seen": 62346176, "step": 32400 }, { "epoch": 5.286320254506893, "grad_norm": 0.12780433893203735, "learning_rate": 4.3185720710519075e-06, "loss": 0.0858, "num_input_tokens_seen": 62356384, "step": 32405 }, { "epoch": 5.28713598172771, "grad_norm": 0.05583399534225464, "learning_rate": 4.3130579764003724e-06, "loss": 0.0128, "num_input_tokens_seen": 62366400, "step": 32410 }, { "epoch": 5.287951708948528, "grad_norm": 0.040230643004179, "learning_rate": 4.307547071935267e-06, "loss": 0.0471, "num_input_tokens_seen": 62375216, "step": 32415 }, { "epoch": 5.288767436169345, "grad_norm": 0.007947935722768307, "learning_rate": 4.302039358506435e-06, "loss": 0.0001, "num_input_tokens_seen": 62384240, "step": 32420 }, { "epoch": 5.289583163390162, "grad_norm": 0.005036250688135624, "learning_rate": 4.296534836963245e-06, "loss": 0.0001, "num_input_tokens_seen": 62393520, "step": 32425 }, { "epoch": 5.290398890610979, "grad_norm": 0.0007263360312208533, "learning_rate": 4.291033508154555e-06, "loss": 0.0412, "num_input_tokens_seen": 62403296, "step": 32430 }, { "epoch": 5.291214617831797, "grad_norm": 0.0010393490083515644, "learning_rate": 4.285535372928748e-06, "loss": 0.0002, "num_input_tokens_seen": 62413040, "step": 32435 }, { "epoch": 5.2920303450526145, "grad_norm": 25.958534240722656, "learning_rate": 4.280040432133695e-06, "loss": 0.1638, "num_input_tokens_seen": 62423056, "step": 32440 }, { "epoch": 5.292846072273432, "grad_norm": 0.18869133293628693, "learning_rate": 4.274548686616789e-06, "loss": 0.0008, "num_input_tokens_seen": 62433344, "step": 32445 }, { "epoch": 5.293661799494249, "grad_norm": 0.0011404735269024968, "learning_rate": 4.2690601372249364e-06, "loss": 0.0655, "num_input_tokens_seen": 62442224, "step": 32450 }, { "epoch": 5.294477526715067, "grad_norm": 0.16967542469501495, "learning_rate": 4.263574784804525e-06, "loss": 0.0005, "num_input_tokens_seen": 62452880, "step": 32455 }, { "epoch": 5.295293253935884, "grad_norm": 0.0014702285407111049, "learning_rate": 4.258092630201479e-06, "loss": 0.0001, "num_input_tokens_seen": 62461824, "step": 32460 }, { "epoch": 5.296108981156701, "grad_norm": 0.0010399918537586927, "learning_rate": 4.252613674261202e-06, "loss": 0.0412, "num_input_tokens_seen": 62471136, "step": 32465 }, { "epoch": 5.296924708377518, "grad_norm": 0.002469992730766535, "learning_rate": 4.2471379178286224e-06, "loss": 0.0005, "num_input_tokens_seen": 62480480, "step": 32470 }, { "epoch": 5.297740435598336, "grad_norm": 0.003746332135051489, "learning_rate": 4.241665361748181e-06, "loss": 0.0002, "num_input_tokens_seen": 62489504, "step": 32475 }, { "epoch": 5.298556162819153, "grad_norm": 0.0013862766791135073, "learning_rate": 4.2361960068637994e-06, "loss": 0.0001, "num_input_tokens_seen": 62498032, "step": 32480 }, { "epoch": 5.2993718900399704, "grad_norm": 0.011705989949405193, "learning_rate": 4.230729854018933e-06, "loss": 0.0001, "num_input_tokens_seen": 62506448, "step": 32485 }, { "epoch": 5.300187617260788, "grad_norm": 0.022226454690098763, "learning_rate": 4.225266904056521e-06, "loss": 0.0001, "num_input_tokens_seen": 62516704, "step": 32490 }, { "epoch": 5.301003344481606, "grad_norm": 0.005176488310098648, "learning_rate": 4.21980715781903e-06, "loss": 0.0001, "num_input_tokens_seen": 62526896, "step": 32495 }, { "epoch": 5.301819071702423, "grad_norm": 4.240273952484131, "learning_rate": 4.214350616148416e-06, "loss": 0.0123, "num_input_tokens_seen": 62537984, "step": 32500 }, { "epoch": 5.30263479892324, "grad_norm": 0.005632300395518541, "learning_rate": 4.20889727988614e-06, "loss": 0.0127, "num_input_tokens_seen": 62548144, "step": 32505 }, { "epoch": 5.303450526144058, "grad_norm": 0.013292194344103336, "learning_rate": 4.20344714987318e-06, "loss": 0.0251, "num_input_tokens_seen": 62557216, "step": 32510 }, { "epoch": 5.304266253364875, "grad_norm": 0.06619030237197876, "learning_rate": 4.198000226950022e-06, "loss": 0.0002, "num_input_tokens_seen": 62568016, "step": 32515 }, { "epoch": 5.305081980585692, "grad_norm": 0.004339316394180059, "learning_rate": 4.192556511956635e-06, "loss": 0.0, "num_input_tokens_seen": 62576336, "step": 32520 }, { "epoch": 5.305897707806509, "grad_norm": 0.010643201880156994, "learning_rate": 4.18711600573252e-06, "loss": 0.0176, "num_input_tokens_seen": 62586768, "step": 32525 }, { "epoch": 5.306713435027326, "grad_norm": 0.0054487017914652824, "learning_rate": 4.181678709116671e-06, "loss": 0.0004, "num_input_tokens_seen": 62596448, "step": 32530 }, { "epoch": 5.307529162248144, "grad_norm": 0.00658324034884572, "learning_rate": 4.1762446229475785e-06, "loss": 0.0002, "num_input_tokens_seen": 62606160, "step": 32535 }, { "epoch": 5.3083448894689615, "grad_norm": 0.06975750625133514, "learning_rate": 4.17081374806326e-06, "loss": 0.0002, "num_input_tokens_seen": 62615600, "step": 32540 }, { "epoch": 5.309160616689779, "grad_norm": 0.0025979545898735523, "learning_rate": 4.165386085301212e-06, "loss": 0.1212, "num_input_tokens_seen": 62624912, "step": 32545 }, { "epoch": 5.309976343910597, "grad_norm": 0.022754691541194916, "learning_rate": 4.1599616354984525e-06, "loss": 0.0001, "num_input_tokens_seen": 62634400, "step": 32550 }, { "epoch": 5.310792071131414, "grad_norm": 0.004314853809773922, "learning_rate": 4.154540399491508e-06, "loss": 0.0002, "num_input_tokens_seen": 62643904, "step": 32555 }, { "epoch": 5.311607798352231, "grad_norm": 0.00013482071517501026, "learning_rate": 4.149122378116394e-06, "loss": 0.0002, "num_input_tokens_seen": 62653504, "step": 32560 }, { "epoch": 5.312423525573048, "grad_norm": 12.05527114868164, "learning_rate": 4.14370757220863e-06, "loss": 0.0028, "num_input_tokens_seen": 62663168, "step": 32565 }, { "epoch": 5.313239252793866, "grad_norm": 0.03337531536817551, "learning_rate": 4.138295982603263e-06, "loss": 0.0003, "num_input_tokens_seen": 62673104, "step": 32570 }, { "epoch": 5.314054980014683, "grad_norm": 0.08874347805976868, "learning_rate": 4.132887610134814e-06, "loss": 0.0005, "num_input_tokens_seen": 62683728, "step": 32575 }, { "epoch": 5.3148707072355, "grad_norm": 0.056108128279447556, "learning_rate": 4.127482455637335e-06, "loss": 0.0001, "num_input_tokens_seen": 62693648, "step": 32580 }, { "epoch": 5.3156864344563175, "grad_norm": 0.01327851414680481, "learning_rate": 4.1220805199443545e-06, "loss": 0.096, "num_input_tokens_seen": 62703536, "step": 32585 }, { "epoch": 5.3165021616771355, "grad_norm": 0.0008694272837601602, "learning_rate": 4.116681803888925e-06, "loss": 0.0004, "num_input_tokens_seen": 62714048, "step": 32590 }, { "epoch": 5.317317888897953, "grad_norm": 9.485426902770996, "learning_rate": 4.111286308303605e-06, "loss": 0.1139, "num_input_tokens_seen": 62723792, "step": 32595 }, { "epoch": 5.31813361611877, "grad_norm": 14.048677444458008, "learning_rate": 4.105894034020433e-06, "loss": 0.0829, "num_input_tokens_seen": 62734064, "step": 32600 }, { "epoch": 5.31813361611877, "eval_loss": 0.3541121482849121, "eval_runtime": 80.8252, "eval_samples_per_second": 33.715, "eval_steps_per_second": 16.864, "num_input_tokens_seen": 62734064, "step": 32600 }, { "epoch": 5.318949343339587, "grad_norm": 0.006289370357990265, "learning_rate": 4.100504981870975e-06, "loss": 0.0021, "num_input_tokens_seen": 62744320, "step": 32605 }, { "epoch": 5.319765070560405, "grad_norm": 0.12818866968154907, "learning_rate": 4.0951191526862915e-06, "loss": 0.0002, "num_input_tokens_seen": 62755072, "step": 32610 }, { "epoch": 5.320580797781222, "grad_norm": 0.08197928220033646, "learning_rate": 4.089736547296938e-06, "loss": 0.0003, "num_input_tokens_seen": 62765504, "step": 32615 }, { "epoch": 5.321396525002039, "grad_norm": 0.015050219371914864, "learning_rate": 4.08435716653299e-06, "loss": 0.0002, "num_input_tokens_seen": 62775504, "step": 32620 }, { "epoch": 5.322212252222856, "grad_norm": 0.00040734256617724895, "learning_rate": 4.0789810112240005e-06, "loss": 0.0015, "num_input_tokens_seen": 62785984, "step": 32625 }, { "epoch": 5.323027979443674, "grad_norm": 0.027248157188296318, "learning_rate": 4.073608082199057e-06, "loss": 0.0033, "num_input_tokens_seen": 62794304, "step": 32630 }, { "epoch": 5.3238437066644915, "grad_norm": 0.028996996581554413, "learning_rate": 4.068238380286718e-06, "loss": 0.0002, "num_input_tokens_seen": 62804944, "step": 32635 }, { "epoch": 5.324659433885309, "grad_norm": 0.05575619637966156, "learning_rate": 4.062871906315072e-06, "loss": 0.0004, "num_input_tokens_seen": 62815312, "step": 32640 }, { "epoch": 5.325475161106126, "grad_norm": 0.0015492454404011369, "learning_rate": 4.057508661111686e-06, "loss": 0.0001, "num_input_tokens_seen": 62825168, "step": 32645 }, { "epoch": 5.326290888326944, "grad_norm": 0.21539302170276642, "learning_rate": 4.052148645503648e-06, "loss": 0.0033, "num_input_tokens_seen": 62834976, "step": 32650 }, { "epoch": 5.327106615547761, "grad_norm": 0.002111805137246847, "learning_rate": 4.046791860317531e-06, "loss": 0.018, "num_input_tokens_seen": 62844256, "step": 32655 }, { "epoch": 5.327922342768578, "grad_norm": 0.02427043952047825, "learning_rate": 4.041438306379431e-06, "loss": 0.0004, "num_input_tokens_seen": 62853152, "step": 32660 }, { "epoch": 5.328738069989395, "grad_norm": 0.026529202237725258, "learning_rate": 4.036087984514916e-06, "loss": 0.0256, "num_input_tokens_seen": 62863248, "step": 32665 }, { "epoch": 5.329553797210213, "grad_norm": 0.003112578531727195, "learning_rate": 4.030740895549084e-06, "loss": 0.0002, "num_input_tokens_seen": 62872144, "step": 32670 }, { "epoch": 5.33036952443103, "grad_norm": 0.01585078425705433, "learning_rate": 4.025397040306531e-06, "loss": 0.0004, "num_input_tokens_seen": 62881248, "step": 32675 }, { "epoch": 5.331185251651847, "grad_norm": 15.348610877990723, "learning_rate": 4.0200564196113285e-06, "loss": 0.0471, "num_input_tokens_seen": 62891616, "step": 32680 }, { "epoch": 5.332000978872665, "grad_norm": 0.001434203120879829, "learning_rate": 4.014719034287079e-06, "loss": 0.0003, "num_input_tokens_seen": 62901184, "step": 32685 }, { "epoch": 5.332816706093483, "grad_norm": 0.003019321709871292, "learning_rate": 4.0093848851568775e-06, "loss": 0.0001, "num_input_tokens_seen": 62910080, "step": 32690 }, { "epoch": 5.3336324333143, "grad_norm": 0.011905476450920105, "learning_rate": 4.004053973043304e-06, "loss": 0.0005, "num_input_tokens_seen": 62921520, "step": 32695 }, { "epoch": 5.334448160535117, "grad_norm": 0.01798173598945141, "learning_rate": 3.998726298768465e-06, "loss": 0.0046, "num_input_tokens_seen": 62932480, "step": 32700 }, { "epoch": 5.335263887755934, "grad_norm": 0.029471293091773987, "learning_rate": 3.99340186315395e-06, "loss": 0.0301, "num_input_tokens_seen": 62943168, "step": 32705 }, { "epoch": 5.336079614976752, "grad_norm": 0.00032936883508227766, "learning_rate": 3.988080667020849e-06, "loss": 0.0004, "num_input_tokens_seen": 62952848, "step": 32710 }, { "epoch": 5.336895342197569, "grad_norm": 0.0026015909388661385, "learning_rate": 3.982762711189766e-06, "loss": 0.0671, "num_input_tokens_seen": 62962656, "step": 32715 }, { "epoch": 5.337711069418386, "grad_norm": 4.54944372177124, "learning_rate": 3.977447996480785e-06, "loss": 0.0278, "num_input_tokens_seen": 62971840, "step": 32720 }, { "epoch": 5.338526796639204, "grad_norm": 7.976385116577148, "learning_rate": 3.97213652371351e-06, "loss": 0.0177, "num_input_tokens_seen": 62980816, "step": 32725 }, { "epoch": 5.339342523860021, "grad_norm": 0.0005177623825147748, "learning_rate": 3.966828293707042e-06, "loss": 0.0, "num_input_tokens_seen": 62990720, "step": 32730 }, { "epoch": 5.3401582510808385, "grad_norm": 0.0020791569259017706, "learning_rate": 3.961523307279963e-06, "loss": 0.0, "num_input_tokens_seen": 62999152, "step": 32735 }, { "epoch": 5.340973978301656, "grad_norm": 0.0325995571911335, "learning_rate": 3.956221565250382e-06, "loss": 0.0001, "num_input_tokens_seen": 63009040, "step": 32740 }, { "epoch": 5.341789705522474, "grad_norm": 0.2758241891860962, "learning_rate": 3.950923068435883e-06, "loss": 0.0012, "num_input_tokens_seen": 63018048, "step": 32745 }, { "epoch": 5.342605432743291, "grad_norm": 0.1919012814760208, "learning_rate": 3.945627817653566e-06, "loss": 0.0142, "num_input_tokens_seen": 63027216, "step": 32750 }, { "epoch": 5.343421159964108, "grad_norm": 0.001726810005493462, "learning_rate": 3.9403358137200335e-06, "loss": 0.0001, "num_input_tokens_seen": 63036912, "step": 32755 }, { "epoch": 5.344236887184925, "grad_norm": 0.0015647229738533497, "learning_rate": 3.9350470574513605e-06, "loss": 0.0001, "num_input_tokens_seen": 63046928, "step": 32760 }, { "epoch": 5.345052614405743, "grad_norm": 0.09294306486845016, "learning_rate": 3.9297615496631525e-06, "loss": 0.0003, "num_input_tokens_seen": 63056144, "step": 32765 }, { "epoch": 5.34586834162656, "grad_norm": 0.0017677794676274061, "learning_rate": 3.924479291170505e-06, "loss": 0.0003, "num_input_tokens_seen": 63066208, "step": 32770 }, { "epoch": 5.346684068847377, "grad_norm": 0.003918840084224939, "learning_rate": 3.919200282788002e-06, "loss": 0.0622, "num_input_tokens_seen": 63076080, "step": 32775 }, { "epoch": 5.3474997960681945, "grad_norm": 0.14685074985027313, "learning_rate": 3.913924525329726e-06, "loss": 0.0008, "num_input_tokens_seen": 63087168, "step": 32780 }, { "epoch": 5.3483155232890125, "grad_norm": 0.06924079358577728, "learning_rate": 3.908652019609279e-06, "loss": 0.0002, "num_input_tokens_seen": 63097792, "step": 32785 }, { "epoch": 5.34913125050983, "grad_norm": 7.948786735534668, "learning_rate": 3.9033827664397364e-06, "loss": 0.0082, "num_input_tokens_seen": 63106496, "step": 32790 }, { "epoch": 5.349946977730647, "grad_norm": 0.04737003892660141, "learning_rate": 3.898116766633694e-06, "loss": 0.0001, "num_input_tokens_seen": 63116016, "step": 32795 }, { "epoch": 5.350762704951464, "grad_norm": 0.0038628431502729654, "learning_rate": 3.8928540210032225e-06, "loss": 0.0001, "num_input_tokens_seen": 63124752, "step": 32800 }, { "epoch": 5.350762704951464, "eval_loss": 0.3812498450279236, "eval_runtime": 80.7587, "eval_samples_per_second": 33.743, "eval_steps_per_second": 16.877, "num_input_tokens_seen": 63124752, "step": 32800 }, { "epoch": 5.351578432172282, "grad_norm": 0.001275920309126377, "learning_rate": 3.887594530359909e-06, "loss": 0.0005, "num_input_tokens_seen": 63134736, "step": 32805 }, { "epoch": 5.352394159393099, "grad_norm": 0.007162844762206078, "learning_rate": 3.88233829551484e-06, "loss": 0.0001, "num_input_tokens_seen": 63143984, "step": 32810 }, { "epoch": 5.353209886613916, "grad_norm": 0.006862602662295103, "learning_rate": 3.877085317278581e-06, "loss": 0.0001, "num_input_tokens_seen": 63154128, "step": 32815 }, { "epoch": 5.354025613834733, "grad_norm": 0.0004920942010357976, "learning_rate": 3.87183559646122e-06, "loss": 0.0028, "num_input_tokens_seen": 63163712, "step": 32820 }, { "epoch": 5.354841341055551, "grad_norm": 0.0068746148608624935, "learning_rate": 3.866589133872317e-06, "loss": 0.0003, "num_input_tokens_seen": 63173120, "step": 32825 }, { "epoch": 5.3556570682763684, "grad_norm": 0.1483575850725174, "learning_rate": 3.861345930320948e-06, "loss": 0.0027, "num_input_tokens_seen": 63183232, "step": 32830 }, { "epoch": 5.356472795497186, "grad_norm": 0.08666719496250153, "learning_rate": 3.856105986615688e-06, "loss": 0.0002, "num_input_tokens_seen": 63192304, "step": 32835 }, { "epoch": 5.357288522718003, "grad_norm": 0.0011205135378986597, "learning_rate": 3.850869303564589e-06, "loss": 0.0014, "num_input_tokens_seen": 63202560, "step": 32840 }, { "epoch": 5.358104249938821, "grad_norm": 0.032581232488155365, "learning_rate": 3.845635881975226e-06, "loss": 0.0004, "num_input_tokens_seen": 63211856, "step": 32845 }, { "epoch": 5.358919977159638, "grad_norm": 374.92657470703125, "learning_rate": 3.840405722654647e-06, "loss": 0.0858, "num_input_tokens_seen": 63221344, "step": 32850 }, { "epoch": 5.359735704380455, "grad_norm": 0.0011887390865013003, "learning_rate": 3.835178826409419e-06, "loss": 0.0001, "num_input_tokens_seen": 63231056, "step": 32855 }, { "epoch": 5.360551431601272, "grad_norm": 0.003893299726769328, "learning_rate": 3.8299551940455895e-06, "loss": 0.0001, "num_input_tokens_seen": 63241472, "step": 32860 }, { "epoch": 5.36136715882209, "grad_norm": 0.0067984564229846, "learning_rate": 3.824734826368703e-06, "loss": 0.0019, "num_input_tokens_seen": 63251120, "step": 32865 }, { "epoch": 5.362182886042907, "grad_norm": 0.0011222645407542586, "learning_rate": 3.819517724183813e-06, "loss": 0.0001, "num_input_tokens_seen": 63261152, "step": 32870 }, { "epoch": 5.362998613263724, "grad_norm": 0.0020816721953451633, "learning_rate": 3.8143038882954648e-06, "loss": 0.0002, "num_input_tokens_seen": 63272096, "step": 32875 }, { "epoch": 5.3638143404845415, "grad_norm": 0.00796410720795393, "learning_rate": 3.8090933195076867e-06, "loss": 0.0001, "num_input_tokens_seen": 63280832, "step": 32880 }, { "epoch": 5.3646300677053596, "grad_norm": 0.019257135689258575, "learning_rate": 3.8038860186240198e-06, "loss": 0.0002, "num_input_tokens_seen": 63290992, "step": 32885 }, { "epoch": 5.365445794926177, "grad_norm": 0.0012631048448383808, "learning_rate": 3.7986819864475026e-06, "loss": 0.0938, "num_input_tokens_seen": 63300656, "step": 32890 }, { "epoch": 5.366261522146994, "grad_norm": 0.035268284380435944, "learning_rate": 3.793481223780651e-06, "loss": 0.0019, "num_input_tokens_seen": 63309712, "step": 32895 }, { "epoch": 5.367077249367812, "grad_norm": 0.02851678803563118, "learning_rate": 3.788283731425496e-06, "loss": 0.0, "num_input_tokens_seen": 63319664, "step": 32900 }, { "epoch": 5.367892976588629, "grad_norm": 1.653497576713562, "learning_rate": 3.7830895101835488e-06, "loss": 0.0006, "num_input_tokens_seen": 63330624, "step": 32905 }, { "epoch": 5.368708703809446, "grad_norm": 0.01054456364363432, "learning_rate": 3.7778985608558274e-06, "loss": 0.0002, "num_input_tokens_seen": 63340384, "step": 32910 }, { "epoch": 5.369524431030263, "grad_norm": 0.003501955419778824, "learning_rate": 3.7727108842428443e-06, "loss": 0.1272, "num_input_tokens_seen": 63349504, "step": 32915 }, { "epoch": 5.370340158251081, "grad_norm": 0.0001624563883524388, "learning_rate": 3.7675264811446065e-06, "loss": 0.0026, "num_input_tokens_seen": 63358640, "step": 32920 }, { "epoch": 5.371155885471898, "grad_norm": 0.32521769404411316, "learning_rate": 3.7623453523605994e-06, "loss": 0.0005, "num_input_tokens_seen": 63368448, "step": 32925 }, { "epoch": 5.3719716126927155, "grad_norm": 0.16029514372348785, "learning_rate": 3.757167498689834e-06, "loss": 0.0007, "num_input_tokens_seen": 63377376, "step": 32930 }, { "epoch": 5.372787339913533, "grad_norm": 0.002865029498934746, "learning_rate": 3.7519929209307914e-06, "loss": 0.0008, "num_input_tokens_seen": 63386784, "step": 32935 }, { "epoch": 5.373603067134351, "grad_norm": 0.00210030865855515, "learning_rate": 3.746821619881463e-06, "loss": 0.0002, "num_input_tokens_seen": 63396720, "step": 32940 }, { "epoch": 5.374418794355168, "grad_norm": 0.07213758677244186, "learning_rate": 3.74165359633932e-06, "loss": 0.0001, "num_input_tokens_seen": 63405152, "step": 32945 }, { "epoch": 5.375234521575985, "grad_norm": 0.007487455382943153, "learning_rate": 3.736488851101341e-06, "loss": 0.092, "num_input_tokens_seen": 63415280, "step": 32950 }, { "epoch": 5.376050248796802, "grad_norm": 0.0026096610818058252, "learning_rate": 3.7313273849640035e-06, "loss": 0.0, "num_input_tokens_seen": 63424864, "step": 32955 }, { "epoch": 5.37686597601762, "grad_norm": 0.020738739520311356, "learning_rate": 3.7261691987232533e-06, "loss": 0.0001, "num_input_tokens_seen": 63436464, "step": 32960 }, { "epoch": 5.377681703238437, "grad_norm": 0.17758384346961975, "learning_rate": 3.7210142931745575e-06, "loss": 0.0004, "num_input_tokens_seen": 63445488, "step": 32965 }, { "epoch": 5.378497430459254, "grad_norm": 0.012546426616609097, "learning_rate": 3.7158626691128712e-06, "loss": 0.0, "num_input_tokens_seen": 63455472, "step": 32970 }, { "epoch": 5.3793131576800715, "grad_norm": 0.017964476719498634, "learning_rate": 3.710714327332629e-06, "loss": 0.0001, "num_input_tokens_seen": 63466464, "step": 32975 }, { "epoch": 5.3801288849008895, "grad_norm": 0.0008757150499150157, "learning_rate": 3.7055692686277815e-06, "loss": 0.0, "num_input_tokens_seen": 63477856, "step": 32980 }, { "epoch": 5.380944612121707, "grad_norm": 0.013606959953904152, "learning_rate": 3.70042749379175e-06, "loss": 0.0001, "num_input_tokens_seen": 63487984, "step": 32985 }, { "epoch": 5.381760339342524, "grad_norm": 0.0013093161396682262, "learning_rate": 3.6952890036174693e-06, "loss": 0.0001, "num_input_tokens_seen": 63497888, "step": 32990 }, { "epoch": 5.382576066563341, "grad_norm": 0.0019114958122372627, "learning_rate": 3.690153798897353e-06, "loss": 0.0, "num_input_tokens_seen": 63507920, "step": 32995 }, { "epoch": 5.383391793784159, "grad_norm": 0.002366473898291588, "learning_rate": 3.6850218804233225e-06, "loss": 0.0588, "num_input_tokens_seen": 63517792, "step": 33000 }, { "epoch": 5.383391793784159, "eval_loss": 0.4010331332683563, "eval_runtime": 80.7248, "eval_samples_per_second": 33.757, "eval_steps_per_second": 16.885, "num_input_tokens_seen": 63517792, "step": 33000 }, { "epoch": 5.384207521004976, "grad_norm": 0.0030121030285954475, "learning_rate": 3.679893248986779e-06, "loss": 0.0001, "num_input_tokens_seen": 63527312, "step": 33005 }, { "epoch": 5.385023248225793, "grad_norm": 0.0004834984429180622, "learning_rate": 3.6747679053786147e-06, "loss": 0.0001, "num_input_tokens_seen": 63536768, "step": 33010 }, { "epoch": 5.38583897544661, "grad_norm": 2.290647268295288, "learning_rate": 3.669645850389228e-06, "loss": 0.0009, "num_input_tokens_seen": 63545520, "step": 33015 }, { "epoch": 5.386654702667428, "grad_norm": 0.0018787257140502334, "learning_rate": 3.664527084808514e-06, "loss": 0.0, "num_input_tokens_seen": 63555120, "step": 33020 }, { "epoch": 5.387470429888245, "grad_norm": 0.004616338759660721, "learning_rate": 3.6594116094258337e-06, "loss": 0.0166, "num_input_tokens_seen": 63564496, "step": 33025 }, { "epoch": 5.388286157109063, "grad_norm": 0.01382819376885891, "learning_rate": 3.6542994250300665e-06, "loss": 0.0268, "num_input_tokens_seen": 63574736, "step": 33030 }, { "epoch": 5.38910188432988, "grad_norm": 0.15674805641174316, "learning_rate": 3.6491905324095825e-06, "loss": 0.0001, "num_input_tokens_seen": 63584976, "step": 33035 }, { "epoch": 5.389917611550698, "grad_norm": 0.003859263379126787, "learning_rate": 3.644084932352221e-06, "loss": 0.0002, "num_input_tokens_seen": 63594528, "step": 33040 }, { "epoch": 5.390733338771515, "grad_norm": 0.004134413320571184, "learning_rate": 3.6389826256453457e-06, "loss": 0.0373, "num_input_tokens_seen": 63604992, "step": 33045 }, { "epoch": 5.391549065992332, "grad_norm": 0.005377678200602531, "learning_rate": 3.633883613075781e-06, "loss": 0.0002, "num_input_tokens_seen": 63614048, "step": 33050 }, { "epoch": 5.392364793213149, "grad_norm": 0.01855410262942314, "learning_rate": 3.6287878954298693e-06, "loss": 0.0005, "num_input_tokens_seen": 63624112, "step": 33055 }, { "epoch": 5.393180520433967, "grad_norm": 0.0021042190492153168, "learning_rate": 3.6236954734934354e-06, "loss": 0.0002, "num_input_tokens_seen": 63633504, "step": 33060 }, { "epoch": 5.393996247654784, "grad_norm": 0.005044254474341869, "learning_rate": 3.618606348051784e-06, "loss": 0.0001, "num_input_tokens_seen": 63641520, "step": 33065 }, { "epoch": 5.394811974875601, "grad_norm": 0.2817225456237793, "learning_rate": 3.6135205198897376e-06, "loss": 0.0001, "num_input_tokens_seen": 63651280, "step": 33070 }, { "epoch": 5.395627702096419, "grad_norm": 0.012955715879797935, "learning_rate": 3.6084379897915854e-06, "loss": 0.0001, "num_input_tokens_seen": 63661536, "step": 33075 }, { "epoch": 5.3964434293172365, "grad_norm": 0.005116111133247614, "learning_rate": 3.6033587585411115e-06, "loss": 0.0011, "num_input_tokens_seen": 63670064, "step": 33080 }, { "epoch": 5.397259156538054, "grad_norm": 0.0004562589747365564, "learning_rate": 3.5982828269216117e-06, "loss": 0.0, "num_input_tokens_seen": 63680096, "step": 33085 }, { "epoch": 5.398074883758871, "grad_norm": 0.024795349687337875, "learning_rate": 3.593210195715843e-06, "loss": 0.0099, "num_input_tokens_seen": 63690208, "step": 33090 }, { "epoch": 5.398890610979688, "grad_norm": 0.001208803034387529, "learning_rate": 3.5881408657060773e-06, "loss": 0.0013, "num_input_tokens_seen": 63700576, "step": 33095 }, { "epoch": 5.399706338200506, "grad_norm": 0.012328486889600754, "learning_rate": 3.583074837674075e-06, "loss": 0.065, "num_input_tokens_seen": 63709216, "step": 33100 }, { "epoch": 5.400522065421323, "grad_norm": 0.00021860384731553495, "learning_rate": 3.578012112401069e-06, "loss": 0.0001, "num_input_tokens_seen": 63718320, "step": 33105 }, { "epoch": 5.40133779264214, "grad_norm": 0.11453940719366074, "learning_rate": 3.5729526906677996e-06, "loss": 0.0009, "num_input_tokens_seen": 63728128, "step": 33110 }, { "epoch": 5.402153519862958, "grad_norm": 0.011209359392523766, "learning_rate": 3.5678965732545007e-06, "loss": 0.0001, "num_input_tokens_seen": 63737472, "step": 33115 }, { "epoch": 5.402969247083775, "grad_norm": 0.002786609809845686, "learning_rate": 3.562843760940876e-06, "loss": 0.0001, "num_input_tokens_seen": 63745968, "step": 33120 }, { "epoch": 5.4037849743045925, "grad_norm": 0.012430094182491302, "learning_rate": 3.5577942545061473e-06, "loss": 0.0009, "num_input_tokens_seen": 63755056, "step": 33125 }, { "epoch": 5.40460070152541, "grad_norm": 18.275733947753906, "learning_rate": 3.5527480547289967e-06, "loss": 0.1232, "num_input_tokens_seen": 63764832, "step": 33130 }, { "epoch": 5.405416428746228, "grad_norm": 0.04171654209494591, "learning_rate": 3.547705162387624e-06, "loss": 0.0003, "num_input_tokens_seen": 63774064, "step": 33135 }, { "epoch": 5.406232155967045, "grad_norm": 0.09281235188245773, "learning_rate": 3.542665578259699e-06, "loss": 0.1211, "num_input_tokens_seen": 63783680, "step": 33140 }, { "epoch": 5.407047883187862, "grad_norm": 0.0004186280712019652, "learning_rate": 3.5376293031223945e-06, "loss": 0.0042, "num_input_tokens_seen": 63793744, "step": 33145 }, { "epoch": 5.407863610408679, "grad_norm": 0.026076607406139374, "learning_rate": 3.5325963377523614e-06, "loss": 0.0158, "num_input_tokens_seen": 63801264, "step": 33150 }, { "epoch": 5.408679337629497, "grad_norm": 5.07736622239463e-05, "learning_rate": 3.5275666829257536e-06, "loss": 0.0004, "num_input_tokens_seen": 63810608, "step": 33155 }, { "epoch": 5.409495064850314, "grad_norm": 0.0013768846401944757, "learning_rate": 3.5225403394181955e-06, "loss": 0.0001, "num_input_tokens_seen": 63819760, "step": 33160 }, { "epoch": 5.410310792071131, "grad_norm": 0.014616752974689007, "learning_rate": 3.517517308004828e-06, "loss": 0.0, "num_input_tokens_seen": 63829040, "step": 33165 }, { "epoch": 5.411126519291948, "grad_norm": 0.000618262100033462, "learning_rate": 3.512497589460251e-06, "loss": 0.0509, "num_input_tokens_seen": 63838608, "step": 33170 }, { "epoch": 5.4119422465127665, "grad_norm": 0.009115577675402164, "learning_rate": 3.5074811845585727e-06, "loss": 0.0, "num_input_tokens_seen": 63847296, "step": 33175 }, { "epoch": 5.412757973733584, "grad_norm": 0.012148477137088776, "learning_rate": 3.5024680940733937e-06, "loss": 0.0003, "num_input_tokens_seen": 63857456, "step": 33180 }, { "epoch": 5.413573700954401, "grad_norm": 0.01759098842740059, "learning_rate": 3.4974583187777852e-06, "loss": 0.0001, "num_input_tokens_seen": 63866848, "step": 33185 }, { "epoch": 5.414389428175218, "grad_norm": 0.0013702609576284885, "learning_rate": 3.4924518594443204e-06, "loss": 0.0251, "num_input_tokens_seen": 63877488, "step": 33190 }, { "epoch": 5.415205155396036, "grad_norm": 0.006177018396556377, "learning_rate": 3.4874487168450682e-06, "loss": 0.0002, "num_input_tokens_seen": 63885824, "step": 33195 }, { "epoch": 5.416020882616853, "grad_norm": 0.001938137225806713, "learning_rate": 3.482448891751558e-06, "loss": 0.0015, "num_input_tokens_seen": 63894896, "step": 33200 }, { "epoch": 5.416020882616853, "eval_loss": 0.4077255129814148, "eval_runtime": 80.9319, "eval_samples_per_second": 33.67, "eval_steps_per_second": 16.841, "num_input_tokens_seen": 63894896, "step": 33200 }, { "epoch": 5.41683660983767, "grad_norm": 0.019440537318587303, "learning_rate": 3.477452384934843e-06, "loss": 0.0001, "num_input_tokens_seen": 63904192, "step": 33205 }, { "epoch": 5.417652337058487, "grad_norm": 0.00201999070122838, "learning_rate": 3.472459197165434e-06, "loss": 0.0001, "num_input_tokens_seen": 63913296, "step": 33210 }, { "epoch": 5.418468064279305, "grad_norm": 1.1092650890350342, "learning_rate": 3.4674693292133518e-06, "loss": 0.0016, "num_input_tokens_seen": 63922624, "step": 33215 }, { "epoch": 5.419283791500122, "grad_norm": 0.00020981530542485416, "learning_rate": 3.4624827818480977e-06, "loss": 0.0001, "num_input_tokens_seen": 63930640, "step": 33220 }, { "epoch": 5.4200995187209395, "grad_norm": 0.0226138886064291, "learning_rate": 3.4574995558386474e-06, "loss": 0.0002, "num_input_tokens_seen": 63940528, "step": 33225 }, { "epoch": 5.420915245941757, "grad_norm": 0.00652636494487524, "learning_rate": 3.452519651953487e-06, "loss": 0.0011, "num_input_tokens_seen": 63951168, "step": 33230 }, { "epoch": 5.421730973162575, "grad_norm": 0.02918778546154499, "learning_rate": 3.447543070960585e-06, "loss": 0.0014, "num_input_tokens_seen": 63960320, "step": 33235 }, { "epoch": 5.422546700383392, "grad_norm": 0.32687464356422424, "learning_rate": 3.4425698136273778e-06, "loss": 0.0007, "num_input_tokens_seen": 63969888, "step": 33240 }, { "epoch": 5.423362427604209, "grad_norm": 0.03110630437731743, "learning_rate": 3.437599880720821e-06, "loss": 0.0001, "num_input_tokens_seen": 63979104, "step": 33245 }, { "epoch": 5.424178154825027, "grad_norm": 0.01887807995080948, "learning_rate": 3.4326332730073267e-06, "loss": 0.0008, "num_input_tokens_seen": 63988864, "step": 33250 }, { "epoch": 5.424993882045844, "grad_norm": 0.00016420021711383015, "learning_rate": 3.427669991252813e-06, "loss": 0.0815, "num_input_tokens_seen": 63999488, "step": 33255 }, { "epoch": 5.425809609266661, "grad_norm": 0.008022270165383816, "learning_rate": 3.42271003622269e-06, "loss": 0.0005, "num_input_tokens_seen": 64009568, "step": 33260 }, { "epoch": 5.426625336487478, "grad_norm": 0.22186827659606934, "learning_rate": 3.4177534086818286e-06, "loss": 0.0003, "num_input_tokens_seen": 64018704, "step": 33265 }, { "epoch": 5.4274410637082955, "grad_norm": 0.009726271033287048, "learning_rate": 3.412800109394612e-06, "loss": 0.0, "num_input_tokens_seen": 64028560, "step": 33270 }, { "epoch": 5.4282567909291135, "grad_norm": 0.00022326133330352604, "learning_rate": 3.4078501391249044e-06, "loss": 0.0, "num_input_tokens_seen": 64038416, "step": 33275 }, { "epoch": 5.429072518149931, "grad_norm": 0.00609755190089345, "learning_rate": 3.4029034986360453e-06, "loss": 0.0, "num_input_tokens_seen": 64047712, "step": 33280 }, { "epoch": 5.429888245370748, "grad_norm": 0.0002025630819844082, "learning_rate": 3.397960188690877e-06, "loss": 0.0, "num_input_tokens_seen": 64057024, "step": 33285 }, { "epoch": 5.430703972591566, "grad_norm": 0.0007289884379133582, "learning_rate": 3.393020210051717e-06, "loss": 0.0001, "num_input_tokens_seen": 64065648, "step": 33290 }, { "epoch": 5.431519699812383, "grad_norm": 0.007055410649627447, "learning_rate": 3.3880835634803655e-06, "loss": 0.0002, "num_input_tokens_seen": 64075248, "step": 33295 }, { "epoch": 5.4323354270332, "grad_norm": 0.00020816327014472336, "learning_rate": 3.383150249738126e-06, "loss": 0.0001, "num_input_tokens_seen": 64084368, "step": 33300 }, { "epoch": 5.433151154254017, "grad_norm": 0.0002809951838571578, "learning_rate": 3.3782202695857663e-06, "loss": 0.0003, "num_input_tokens_seen": 64093424, "step": 33305 }, { "epoch": 5.433966881474835, "grad_norm": 0.015901871025562286, "learning_rate": 3.373293623783558e-06, "loss": 0.0001, "num_input_tokens_seen": 64104384, "step": 33310 }, { "epoch": 5.434782608695652, "grad_norm": 0.0024868641048669815, "learning_rate": 3.368370313091257e-06, "loss": 0.0, "num_input_tokens_seen": 64113664, "step": 33315 }, { "epoch": 5.4355983359164695, "grad_norm": 0.052766717970371246, "learning_rate": 3.363450338268087e-06, "loss": 0.0005, "num_input_tokens_seen": 64123392, "step": 33320 }, { "epoch": 5.436414063137287, "grad_norm": 0.00024491079966537654, "learning_rate": 3.358533700072783e-06, "loss": 0.0003, "num_input_tokens_seen": 64132352, "step": 33325 }, { "epoch": 5.437229790358105, "grad_norm": 0.0002623642794787884, "learning_rate": 3.3536203992635377e-06, "loss": 0.0027, "num_input_tokens_seen": 64142416, "step": 33330 }, { "epoch": 5.438045517578922, "grad_norm": 0.011265553534030914, "learning_rate": 3.348710436598057e-06, "loss": 0.0001, "num_input_tokens_seen": 64152256, "step": 33335 }, { "epoch": 5.438861244799739, "grad_norm": 0.03879385441541672, "learning_rate": 3.3438038128335155e-06, "loss": 0.0001, "num_input_tokens_seen": 64161856, "step": 33340 }, { "epoch": 5.439676972020556, "grad_norm": 0.0026825256645679474, "learning_rate": 3.338900528726571e-06, "loss": 0.0001, "num_input_tokens_seen": 64171952, "step": 33345 }, { "epoch": 5.440492699241374, "grad_norm": 0.010695229284465313, "learning_rate": 3.3340005850333812e-06, "loss": 0.0002, "num_input_tokens_seen": 64182176, "step": 33350 }, { "epoch": 5.441308426462191, "grad_norm": 0.0003130580298602581, "learning_rate": 3.329103982509568e-06, "loss": 0.0, "num_input_tokens_seen": 64192016, "step": 33355 }, { "epoch": 5.442124153683008, "grad_norm": 0.000151906642713584, "learning_rate": 3.324210721910259e-06, "loss": 0.0001, "num_input_tokens_seen": 64201392, "step": 33360 }, { "epoch": 5.442939880903825, "grad_norm": 0.0002633333788253367, "learning_rate": 3.319320803990053e-06, "loss": 0.0001, "num_input_tokens_seen": 64211360, "step": 33365 }, { "epoch": 5.443755608124643, "grad_norm": 0.0010696336394175887, "learning_rate": 3.3144342295030274e-06, "loss": 0.0, "num_input_tokens_seen": 64221264, "step": 33370 }, { "epoch": 5.444571335345461, "grad_norm": 14.505257606506348, "learning_rate": 3.309550999202765e-06, "loss": 0.0284, "num_input_tokens_seen": 64231440, "step": 33375 }, { "epoch": 5.445387062566278, "grad_norm": 0.01787298358976841, "learning_rate": 3.3046711138423197e-06, "loss": 0.0, "num_input_tokens_seen": 64239520, "step": 33380 }, { "epoch": 5.446202789787095, "grad_norm": 0.11632764339447021, "learning_rate": 3.2997945741742255e-06, "loss": 0.0008, "num_input_tokens_seen": 64250096, "step": 33385 }, { "epoch": 5.447018517007913, "grad_norm": 0.16833464801311493, "learning_rate": 3.2949213809505082e-06, "loss": 0.0032, "num_input_tokens_seen": 64259184, "step": 33390 }, { "epoch": 5.44783424422873, "grad_norm": 15.603408813476562, "learning_rate": 3.2900515349226834e-06, "loss": 0.0252, "num_input_tokens_seen": 64268912, "step": 33395 }, { "epoch": 5.448649971449547, "grad_norm": 0.009904598817229271, "learning_rate": 3.285185036841731e-06, "loss": 0.0001, "num_input_tokens_seen": 64277584, "step": 33400 }, { "epoch": 5.448649971449547, "eval_loss": 0.42310479283332825, "eval_runtime": 80.8151, "eval_samples_per_second": 33.719, "eval_steps_per_second": 16.866, "num_input_tokens_seen": 64277584, "step": 33400 }, { "epoch": 5.449465698670364, "grad_norm": 0.005556174088269472, "learning_rate": 3.2803218874581377e-06, "loss": 0.0, "num_input_tokens_seen": 64286688, "step": 33405 }, { "epoch": 5.450281425891182, "grad_norm": 5.364911079406738, "learning_rate": 3.2754620875218494e-06, "loss": 0.0016, "num_input_tokens_seen": 64296240, "step": 33410 }, { "epoch": 5.451097153111999, "grad_norm": 0.0015947637148201466, "learning_rate": 3.2706056377823146e-06, "loss": 0.0001, "num_input_tokens_seen": 64305248, "step": 33415 }, { "epoch": 5.4519128803328165, "grad_norm": 0.0019168434664607048, "learning_rate": 3.2657525389884647e-06, "loss": 0.0, "num_input_tokens_seen": 64315008, "step": 33420 }, { "epoch": 5.4527286075536345, "grad_norm": 0.031532783061265945, "learning_rate": 3.260902791888698e-06, "loss": 0.0002, "num_input_tokens_seen": 64324480, "step": 33425 }, { "epoch": 5.453544334774452, "grad_norm": 0.018109135329723358, "learning_rate": 3.2560563972309166e-06, "loss": 0.0013, "num_input_tokens_seen": 64334096, "step": 33430 }, { "epoch": 5.454360061995269, "grad_norm": 0.2716256082057953, "learning_rate": 3.251213355762489e-06, "loss": 0.0009, "num_input_tokens_seen": 64344528, "step": 33435 }, { "epoch": 5.455175789216086, "grad_norm": 20.03949546813965, "learning_rate": 3.2463736682302707e-06, "loss": 0.0551, "num_input_tokens_seen": 64354160, "step": 33440 }, { "epoch": 5.455991516436903, "grad_norm": 0.2864861488342285, "learning_rate": 3.2415373353806124e-06, "loss": 0.0004, "num_input_tokens_seen": 64363360, "step": 33445 }, { "epoch": 5.456807243657721, "grad_norm": 0.02676519937813282, "learning_rate": 3.236704357959322e-06, "loss": 0.0011, "num_input_tokens_seen": 64372336, "step": 33450 }, { "epoch": 5.457622970878538, "grad_norm": 0.006676219403743744, "learning_rate": 3.2318747367117154e-06, "loss": 0.0001, "num_input_tokens_seen": 64381488, "step": 33455 }, { "epoch": 5.458438698099355, "grad_norm": 9.478784704697318e-06, "learning_rate": 3.227048472382585e-06, "loss": 0.0001, "num_input_tokens_seen": 64390272, "step": 33460 }, { "epoch": 5.459254425320173, "grad_norm": 0.0002661396865732968, "learning_rate": 3.2222255657161915e-06, "loss": 0.0001, "num_input_tokens_seen": 64399344, "step": 33465 }, { "epoch": 5.4600701525409905, "grad_norm": 0.06320412456989288, "learning_rate": 3.2174060174562924e-06, "loss": 0.0001, "num_input_tokens_seen": 64409488, "step": 33470 }, { "epoch": 5.460885879761808, "grad_norm": 0.0013156307395547628, "learning_rate": 3.2125898283461298e-06, "loss": 0.0, "num_input_tokens_seen": 64419456, "step": 33475 }, { "epoch": 5.461701606982625, "grad_norm": 0.0005471643526107073, "learning_rate": 3.207776999128406e-06, "loss": 0.0, "num_input_tokens_seen": 64429296, "step": 33480 }, { "epoch": 5.462517334203443, "grad_norm": 0.0024335719645023346, "learning_rate": 3.202967530545331e-06, "loss": 0.0001, "num_input_tokens_seen": 64439744, "step": 33485 }, { "epoch": 5.46333306142426, "grad_norm": 0.09151674062013626, "learning_rate": 3.1981614233385778e-06, "loss": 0.0089, "num_input_tokens_seen": 64450048, "step": 33490 }, { "epoch": 5.464148788645077, "grad_norm": 0.00021175014262553304, "learning_rate": 3.1933586782493115e-06, "loss": 0.0001, "num_input_tokens_seen": 64458128, "step": 33495 }, { "epoch": 5.464964515865894, "grad_norm": 0.0014677282888442278, "learning_rate": 3.188559296018184e-06, "loss": 0.0027, "num_input_tokens_seen": 64468112, "step": 33500 }, { "epoch": 5.465780243086712, "grad_norm": 0.005369419697672129, "learning_rate": 3.1837632773853098e-06, "loss": 0.0001, "num_input_tokens_seen": 64477232, "step": 33505 }, { "epoch": 5.466595970307529, "grad_norm": 0.010499611496925354, "learning_rate": 3.178970623090294e-06, "loss": 0.0007, "num_input_tokens_seen": 64486512, "step": 33510 }, { "epoch": 5.467411697528346, "grad_norm": 0.00016271141066681594, "learning_rate": 3.174181333872234e-06, "loss": 0.0053, "num_input_tokens_seen": 64495696, "step": 33515 }, { "epoch": 5.468227424749164, "grad_norm": 0.0026279082521796227, "learning_rate": 3.169395410469686e-06, "loss": 0.0001, "num_input_tokens_seen": 64505440, "step": 33520 }, { "epoch": 5.469043151969982, "grad_norm": 0.00027902520378120244, "learning_rate": 3.164612853620713e-06, "loss": 0.0, "num_input_tokens_seen": 64515728, "step": 33525 }, { "epoch": 5.469858879190799, "grad_norm": 9.93676803773269e-06, "learning_rate": 3.1598336640628333e-06, "loss": 0.0, "num_input_tokens_seen": 64525616, "step": 33530 }, { "epoch": 5.470674606411616, "grad_norm": 0.002300194464623928, "learning_rate": 3.155057842533063e-06, "loss": 0.0955, "num_input_tokens_seen": 64535504, "step": 33535 }, { "epoch": 5.471490333632433, "grad_norm": 0.0023925104178488255, "learning_rate": 3.1502853897678984e-06, "loss": 0.0163, "num_input_tokens_seen": 64544864, "step": 33540 }, { "epoch": 5.472306060853251, "grad_norm": 2.8869040761492215e-05, "learning_rate": 3.1455163065033017e-06, "loss": 0.0048, "num_input_tokens_seen": 64553504, "step": 33545 }, { "epoch": 5.473121788074068, "grad_norm": 0.012242693454027176, "learning_rate": 3.140750593474734e-06, "loss": 0.0001, "num_input_tokens_seen": 64562112, "step": 33550 }, { "epoch": 5.473937515294885, "grad_norm": 9.545345306396484, "learning_rate": 3.1359882514171294e-06, "loss": 0.0163, "num_input_tokens_seen": 64572576, "step": 33555 }, { "epoch": 5.474753242515702, "grad_norm": 7.622049808502197, "learning_rate": 3.1312292810648903e-06, "loss": 0.0418, "num_input_tokens_seen": 64583456, "step": 33560 }, { "epoch": 5.47556896973652, "grad_norm": 0.00010797900176839903, "learning_rate": 3.1264736831519204e-06, "loss": 0.0, "num_input_tokens_seen": 64593824, "step": 33565 }, { "epoch": 5.4763846969573375, "grad_norm": 0.15678802132606506, "learning_rate": 3.1217214584115863e-06, "loss": 0.0001, "num_input_tokens_seen": 64603968, "step": 33570 }, { "epoch": 5.477200424178155, "grad_norm": 0.005461752414703369, "learning_rate": 3.116972607576746e-06, "loss": 0.0002, "num_input_tokens_seen": 64613808, "step": 33575 }, { "epoch": 5.478016151398972, "grad_norm": 0.0008356762118637562, "learning_rate": 3.1122271313797303e-06, "loss": 0.0002, "num_input_tokens_seen": 64623952, "step": 33580 }, { "epoch": 5.47883187861979, "grad_norm": 0.003084291936829686, "learning_rate": 3.107485030552343e-06, "loss": 0.0809, "num_input_tokens_seen": 64633872, "step": 33585 }, { "epoch": 5.479647605840607, "grad_norm": 0.0010755832772701979, "learning_rate": 3.1027463058258848e-06, "loss": 0.0, "num_input_tokens_seen": 64642800, "step": 33590 }, { "epoch": 5.480463333061424, "grad_norm": 0.0012425872264429927, "learning_rate": 3.0980109579311273e-06, "loss": 0.0, "num_input_tokens_seen": 64651808, "step": 33595 }, { "epoch": 5.481279060282241, "grad_norm": 0.020338676869869232, "learning_rate": 3.093278987598314e-06, "loss": 0.0001, "num_input_tokens_seen": 64661856, "step": 33600 }, { "epoch": 5.481279060282241, "eval_loss": 0.4215812385082245, "eval_runtime": 80.7645, "eval_samples_per_second": 33.74, "eval_steps_per_second": 16.876, "num_input_tokens_seen": 64661856, "step": 33600 }, { "epoch": 5.482094787503059, "grad_norm": 0.0007696390966884792, "learning_rate": 3.0885503955571826e-06, "loss": 0.0001, "num_input_tokens_seen": 64671856, "step": 33605 }, { "epoch": 5.482910514723876, "grad_norm": 0.004453849978744984, "learning_rate": 3.0838251825369313e-06, "loss": 0.0, "num_input_tokens_seen": 64681632, "step": 33610 }, { "epoch": 5.4837262419446935, "grad_norm": 0.0025470457039773464, "learning_rate": 3.0791033492662517e-06, "loss": 0.0001, "num_input_tokens_seen": 64691728, "step": 33615 }, { "epoch": 5.484541969165511, "grad_norm": 0.01645263098180294, "learning_rate": 3.0743848964733203e-06, "loss": 0.0003, "num_input_tokens_seen": 64702208, "step": 33620 }, { "epoch": 5.485357696386329, "grad_norm": 0.0012543300399556756, "learning_rate": 3.0696698248857625e-06, "loss": 0.0002, "num_input_tokens_seen": 64712208, "step": 33625 }, { "epoch": 5.486173423607146, "grad_norm": 0.0013507569674402475, "learning_rate": 3.0649581352307192e-06, "loss": 0.0008, "num_input_tokens_seen": 64722080, "step": 33630 }, { "epoch": 5.486989150827963, "grad_norm": 0.0006617690669372678, "learning_rate": 3.060249828234776e-06, "loss": 0.0001, "num_input_tokens_seen": 64731984, "step": 33635 }, { "epoch": 5.487804878048781, "grad_norm": 0.008415188640356064, "learning_rate": 3.055544904624025e-06, "loss": 0.0002, "num_input_tokens_seen": 64741936, "step": 33640 }, { "epoch": 5.488620605269598, "grad_norm": 0.022019170224666595, "learning_rate": 3.050843365124026e-06, "loss": 0.0001, "num_input_tokens_seen": 64752048, "step": 33645 }, { "epoch": 5.489436332490415, "grad_norm": 0.04020630195736885, "learning_rate": 3.0461452104598083e-06, "loss": 0.0001, "num_input_tokens_seen": 64761600, "step": 33650 }, { "epoch": 5.490252059711232, "grad_norm": 0.002845930401235819, "learning_rate": 3.0414504413558836e-06, "loss": 0.0007, "num_input_tokens_seen": 64770816, "step": 33655 }, { "epoch": 5.49106778693205, "grad_norm": 0.005156876053661108, "learning_rate": 3.0367590585362564e-06, "loss": 0.0001, "num_input_tokens_seen": 64779536, "step": 33660 }, { "epoch": 5.4918835141528675, "grad_norm": 4.6490389649989083e-05, "learning_rate": 3.0320710627243813e-06, "loss": 0.0001, "num_input_tokens_seen": 64789664, "step": 33665 }, { "epoch": 5.492699241373685, "grad_norm": 0.014034437015652657, "learning_rate": 3.027386454643222e-06, "loss": 0.0676, "num_input_tokens_seen": 64799760, "step": 33670 }, { "epoch": 5.493514968594502, "grad_norm": 0.020551519468426704, "learning_rate": 3.0227052350151914e-06, "loss": 0.0, "num_input_tokens_seen": 64809264, "step": 33675 }, { "epoch": 5.49433069581532, "grad_norm": 10.151928901672363, "learning_rate": 3.0180274045621957e-06, "loss": 0.097, "num_input_tokens_seen": 64819152, "step": 33680 }, { "epoch": 5.495146423036137, "grad_norm": 0.009812862612307072, "learning_rate": 3.013352964005625e-06, "loss": 0.1313, "num_input_tokens_seen": 64827968, "step": 33685 }, { "epoch": 5.495962150256954, "grad_norm": 0.0016806575004011393, "learning_rate": 3.0086819140663218e-06, "loss": 0.0, "num_input_tokens_seen": 64837088, "step": 33690 }, { "epoch": 5.496777877477771, "grad_norm": 0.0025584385730326176, "learning_rate": 3.0040142554646265e-06, "loss": 0.0001, "num_input_tokens_seen": 64846336, "step": 33695 }, { "epoch": 5.497593604698589, "grad_norm": 0.00585075281560421, "learning_rate": 2.999349988920361e-06, "loss": 0.0005, "num_input_tokens_seen": 64856224, "step": 33700 }, { "epoch": 5.498409331919406, "grad_norm": 0.00010067458060802892, "learning_rate": 2.994689115152796e-06, "loss": 0.0191, "num_input_tokens_seen": 64865392, "step": 33705 }, { "epoch": 5.499225059140223, "grad_norm": 34.646759033203125, "learning_rate": 2.9900316348807105e-06, "loss": 0.0065, "num_input_tokens_seen": 64874848, "step": 33710 }, { "epoch": 5.5000407863610405, "grad_norm": 0.058772943913936615, "learning_rate": 2.985377548822338e-06, "loss": 0.0001, "num_input_tokens_seen": 64884496, "step": 33715 }, { "epoch": 5.500856513581859, "grad_norm": 0.0013226147275418043, "learning_rate": 2.980726857695404e-06, "loss": 0.0002, "num_input_tokens_seen": 64894464, "step": 33720 }, { "epoch": 5.501672240802676, "grad_norm": 0.0005453057819977403, "learning_rate": 2.9760795622171017e-06, "loss": 0.0001, "num_input_tokens_seen": 64902192, "step": 33725 }, { "epoch": 5.502487968023493, "grad_norm": 0.00026049846201203763, "learning_rate": 2.971435663104094e-06, "loss": 0.0589, "num_input_tokens_seen": 64910208, "step": 33730 }, { "epoch": 5.50330369524431, "grad_norm": 0.0008856600616127253, "learning_rate": 2.9667951610725385e-06, "loss": 0.0001, "num_input_tokens_seen": 64918928, "step": 33735 }, { "epoch": 5.504119422465128, "grad_norm": 0.012150133959949017, "learning_rate": 2.9621580568380575e-06, "loss": 0.0006, "num_input_tokens_seen": 64928320, "step": 33740 }, { "epoch": 5.504935149685945, "grad_norm": 0.0015535899437963963, "learning_rate": 2.9575243511157453e-06, "loss": 0.0, "num_input_tokens_seen": 64938304, "step": 33745 }, { "epoch": 5.505750876906762, "grad_norm": 0.002739539835602045, "learning_rate": 2.952894044620186e-06, "loss": 0.0, "num_input_tokens_seen": 64948464, "step": 33750 }, { "epoch": 5.506566604127579, "grad_norm": 0.03209034726023674, "learning_rate": 2.948267138065419e-06, "loss": 0.002, "num_input_tokens_seen": 64957856, "step": 33755 }, { "epoch": 5.507382331348397, "grad_norm": 0.016420898959040642, "learning_rate": 2.943643632164983e-06, "loss": 0.0002, "num_input_tokens_seen": 64966464, "step": 33760 }, { "epoch": 5.5081980585692145, "grad_norm": 0.003687257179990411, "learning_rate": 2.939023527631879e-06, "loss": 0.0002, "num_input_tokens_seen": 64976512, "step": 33765 }, { "epoch": 5.509013785790032, "grad_norm": 0.020152857527136803, "learning_rate": 2.934406825178576e-06, "loss": 0.0001, "num_input_tokens_seen": 64986880, "step": 33770 }, { "epoch": 5.50982951301085, "grad_norm": 0.008755396120250225, "learning_rate": 2.9297935255170357e-06, "loss": 0.0, "num_input_tokens_seen": 64995648, "step": 33775 }, { "epoch": 5.510645240231667, "grad_norm": 0.0003069050144404173, "learning_rate": 2.925183629358691e-06, "loss": 0.0001, "num_input_tokens_seen": 65005568, "step": 33780 }, { "epoch": 5.511460967452484, "grad_norm": 0.011406437493860722, "learning_rate": 2.9205771374144346e-06, "loss": 0.0728, "num_input_tokens_seen": 65014432, "step": 33785 }, { "epoch": 5.512276694673301, "grad_norm": 9.596352174412459e-05, "learning_rate": 2.915974050394657e-06, "loss": 0.0, "num_input_tokens_seen": 65023872, "step": 33790 }, { "epoch": 5.513092421894118, "grad_norm": 7.702846527099609, "learning_rate": 2.9113743690092067e-06, "loss": 0.0509, "num_input_tokens_seen": 65033280, "step": 33795 }, { "epoch": 5.513908149114936, "grad_norm": 0.0004860398476012051, "learning_rate": 2.906778093967402e-06, "loss": 0.0, "num_input_tokens_seen": 65043136, "step": 33800 }, { "epoch": 5.513908149114936, "eval_loss": 0.42182645201683044, "eval_runtime": 80.7491, "eval_samples_per_second": 33.747, "eval_steps_per_second": 16.879, "num_input_tokens_seen": 65043136, "step": 33800 }, { "epoch": 5.514723876335753, "grad_norm": 0.009953792206943035, "learning_rate": 2.9021852259780656e-06, "loss": 0.0, "num_input_tokens_seen": 65053504, "step": 33805 }, { "epoch": 5.5155396035565705, "grad_norm": 27.711772918701172, "learning_rate": 2.8975957657494583e-06, "loss": 0.1008, "num_input_tokens_seen": 65063120, "step": 33810 }, { "epoch": 5.5163553307773885, "grad_norm": 0.0008719742181710899, "learning_rate": 2.8930097139893417e-06, "loss": 0.0022, "num_input_tokens_seen": 65073824, "step": 33815 }, { "epoch": 5.517171057998206, "grad_norm": 0.0166680458933115, "learning_rate": 2.888427071404945e-06, "loss": 0.0002, "num_input_tokens_seen": 65083952, "step": 33820 }, { "epoch": 5.517986785219023, "grad_norm": 0.000595647725276649, "learning_rate": 2.8838478387029606e-06, "loss": 0.0001, "num_input_tokens_seen": 65094704, "step": 33825 }, { "epoch": 5.51880251243984, "grad_norm": 0.0013531037839129567, "learning_rate": 2.8792720165895737e-06, "loss": 0.0, "num_input_tokens_seen": 65105392, "step": 33830 }, { "epoch": 5.519618239660657, "grad_norm": 4.069205284118652, "learning_rate": 2.874699605770423e-06, "loss": 0.0053, "num_input_tokens_seen": 65114976, "step": 33835 }, { "epoch": 5.520433966881475, "grad_norm": 0.006763690151274204, "learning_rate": 2.8701306069506383e-06, "loss": 0.032, "num_input_tokens_seen": 65125664, "step": 33840 }, { "epoch": 5.521249694102292, "grad_norm": 0.02374289184808731, "learning_rate": 2.8655650208348178e-06, "loss": 0.0004, "num_input_tokens_seen": 65135888, "step": 33845 }, { "epoch": 5.522065421323109, "grad_norm": 0.006765235681086779, "learning_rate": 2.8610028481270257e-06, "loss": 0.0002, "num_input_tokens_seen": 65146208, "step": 33850 }, { "epoch": 5.522881148543927, "grad_norm": 1.8263603448867798, "learning_rate": 2.856444089530813e-06, "loss": 0.0022, "num_input_tokens_seen": 65156512, "step": 33855 }, { "epoch": 5.523696875764744, "grad_norm": 0.40857192873954773, "learning_rate": 2.8518887457491955e-06, "loss": 0.013, "num_input_tokens_seen": 65166656, "step": 33860 }, { "epoch": 5.524512602985562, "grad_norm": 0.5487334728240967, "learning_rate": 2.8473368174846666e-06, "loss": 0.0001, "num_input_tokens_seen": 65175664, "step": 33865 }, { "epoch": 5.525328330206379, "grad_norm": 0.006981448270380497, "learning_rate": 2.842788305439184e-06, "loss": 0.0, "num_input_tokens_seen": 65186400, "step": 33870 }, { "epoch": 5.526144057427197, "grad_norm": 0.009066244587302208, "learning_rate": 2.8382432103141925e-06, "loss": 0.0002, "num_input_tokens_seen": 65195664, "step": 33875 }, { "epoch": 5.526959784648014, "grad_norm": 0.00019251613412052393, "learning_rate": 2.833701532810598e-06, "loss": 0.0025, "num_input_tokens_seen": 65205408, "step": 33880 }, { "epoch": 5.527775511868831, "grad_norm": 0.0036231253761798143, "learning_rate": 2.8291632736287877e-06, "loss": 0.0001, "num_input_tokens_seen": 65216448, "step": 33885 }, { "epoch": 5.528591239089648, "grad_norm": 0.019062833860516548, "learning_rate": 2.824628433468615e-06, "loss": 0.0, "num_input_tokens_seen": 65226480, "step": 33890 }, { "epoch": 5.529406966310466, "grad_norm": 0.00013812539691571146, "learning_rate": 2.8200970130294073e-06, "loss": 0.0003, "num_input_tokens_seen": 65235328, "step": 33895 }, { "epoch": 5.530222693531283, "grad_norm": 0.12373921275138855, "learning_rate": 2.8155690130099775e-06, "loss": 0.0001, "num_input_tokens_seen": 65243792, "step": 33900 }, { "epoch": 5.5310384207521, "grad_norm": 0.0001367343938909471, "learning_rate": 2.8110444341085895e-06, "loss": 0.0652, "num_input_tokens_seen": 65254624, "step": 33905 }, { "epoch": 5.5318541479729175, "grad_norm": 0.005033377557992935, "learning_rate": 2.806523277022996e-06, "loss": 0.0, "num_input_tokens_seen": 65263872, "step": 33910 }, { "epoch": 5.5326698751937355, "grad_norm": 0.0016023525968194008, "learning_rate": 2.802005542450409e-06, "loss": 0.0, "num_input_tokens_seen": 65272480, "step": 33915 }, { "epoch": 5.533485602414553, "grad_norm": 0.0935104638338089, "learning_rate": 2.797491231087526e-06, "loss": 0.0001, "num_input_tokens_seen": 65281328, "step": 33920 }, { "epoch": 5.53430132963537, "grad_norm": 0.596838116645813, "learning_rate": 2.7929803436305137e-06, "loss": 0.0006, "num_input_tokens_seen": 65291344, "step": 33925 }, { "epoch": 5.535117056856187, "grad_norm": 0.08731511235237122, "learning_rate": 2.788472880774998e-06, "loss": 0.0001, "num_input_tokens_seen": 65300304, "step": 33930 }, { "epoch": 5.535932784077005, "grad_norm": 0.056846268475055695, "learning_rate": 2.7839688432160977e-06, "loss": 0.0001, "num_input_tokens_seen": 65310272, "step": 33935 }, { "epoch": 5.536748511297822, "grad_norm": 0.0048723104409873486, "learning_rate": 2.779468231648383e-06, "loss": 0.0011, "num_input_tokens_seen": 65320928, "step": 33940 }, { "epoch": 5.537564238518639, "grad_norm": 3.3654427528381348, "learning_rate": 2.774971046765906e-06, "loss": 0.0025, "num_input_tokens_seen": 65330592, "step": 33945 }, { "epoch": 5.538379965739456, "grad_norm": 0.012199044227600098, "learning_rate": 2.770477289262194e-06, "loss": 0.0, "num_input_tokens_seen": 65341744, "step": 33950 }, { "epoch": 5.539195692960274, "grad_norm": 0.005223495420068502, "learning_rate": 2.765986959830233e-06, "loss": 0.0001, "num_input_tokens_seen": 65349856, "step": 33955 }, { "epoch": 5.5400114201810915, "grad_norm": 0.02612699754536152, "learning_rate": 2.761500059162492e-06, "loss": 0.0002, "num_input_tokens_seen": 65357968, "step": 33960 }, { "epoch": 5.540827147401909, "grad_norm": 0.003287092549726367, "learning_rate": 2.757016587950914e-06, "loss": 0.0005, "num_input_tokens_seen": 65368368, "step": 33965 }, { "epoch": 5.541642874622726, "grad_norm": 0.0069119688123464584, "learning_rate": 2.752536546886897e-06, "loss": 0.0548, "num_input_tokens_seen": 65378064, "step": 33970 }, { "epoch": 5.542458601843544, "grad_norm": 0.00041147141018882394, "learning_rate": 2.7480599366613234e-06, "loss": 0.0001, "num_input_tokens_seen": 65388016, "step": 33975 }, { "epoch": 5.543274329064361, "grad_norm": 0.00101081607863307, "learning_rate": 2.7435867579645473e-06, "loss": 0.0, "num_input_tokens_seen": 65399056, "step": 33980 }, { "epoch": 5.544090056285178, "grad_norm": 0.00015848509792704135, "learning_rate": 2.739117011486378e-06, "loss": 0.0, "num_input_tokens_seen": 65408576, "step": 33985 }, { "epoch": 5.544905783505996, "grad_norm": 0.0036316802725195885, "learning_rate": 2.7346506979161216e-06, "loss": 0.0001, "num_input_tokens_seen": 65419584, "step": 33990 }, { "epoch": 5.545721510726813, "grad_norm": 0.03614652901887894, "learning_rate": 2.7301878179425227e-06, "loss": 0.0001, "num_input_tokens_seen": 65429952, "step": 33995 }, { "epoch": 5.54653723794763, "grad_norm": 0.00035824961378239095, "learning_rate": 2.7257283722538244e-06, "loss": 0.0, "num_input_tokens_seen": 65439360, "step": 34000 }, { "epoch": 5.54653723794763, "eval_loss": 0.42869704961776733, "eval_runtime": 80.8539, "eval_samples_per_second": 33.703, "eval_steps_per_second": 16.858, "num_input_tokens_seen": 65439360, "step": 34000 }, { "epoch": 5.5473529651684474, "grad_norm": 0.045987553894519806, "learning_rate": 2.7212723615377326e-06, "loss": 0.0002, "num_input_tokens_seen": 65449056, "step": 34005 }, { "epoch": 5.548168692389265, "grad_norm": 0.028905315324664116, "learning_rate": 2.7168197864814145e-06, "loss": 0.0004, "num_input_tokens_seen": 65459152, "step": 34010 }, { "epoch": 5.548984419610083, "grad_norm": 0.0036819924134761095, "learning_rate": 2.712370647771509e-06, "loss": 0.0034, "num_input_tokens_seen": 65468800, "step": 34015 }, { "epoch": 5.5498001468309, "grad_norm": 0.08211256563663483, "learning_rate": 2.707924946094137e-06, "loss": 0.0002, "num_input_tokens_seen": 65477872, "step": 34020 }, { "epoch": 5.550615874051717, "grad_norm": 10.649937629699707, "learning_rate": 2.7034826821348723e-06, "loss": 0.0917, "num_input_tokens_seen": 65486512, "step": 34025 }, { "epoch": 5.551431601272535, "grad_norm": 0.0005225877393968403, "learning_rate": 2.6990438565787786e-06, "loss": 0.0, "num_input_tokens_seen": 65494816, "step": 34030 }, { "epoch": 5.552247328493352, "grad_norm": 0.04222217574715614, "learning_rate": 2.6946084701103714e-06, "loss": 0.0001, "num_input_tokens_seen": 65504528, "step": 34035 }, { "epoch": 5.553063055714169, "grad_norm": 0.0026972019113600254, "learning_rate": 2.6901765234136428e-06, "loss": 0.0, "num_input_tokens_seen": 65514032, "step": 34040 }, { "epoch": 5.553878782934986, "grad_norm": 0.0014253114350140095, "learning_rate": 2.685748017172063e-06, "loss": 0.0, "num_input_tokens_seen": 65523632, "step": 34045 }, { "epoch": 5.554694510155803, "grad_norm": 0.0176312904804945, "learning_rate": 2.681322952068549e-06, "loss": 0.0001, "num_input_tokens_seen": 65532800, "step": 34050 }, { "epoch": 5.555510237376621, "grad_norm": 0.0008212112006731331, "learning_rate": 2.6769013287855137e-06, "loss": 0.0001, "num_input_tokens_seen": 65541712, "step": 34055 }, { "epoch": 5.5563259645974385, "grad_norm": 0.0004734042740892619, "learning_rate": 2.6724831480048286e-06, "loss": 0.0001, "num_input_tokens_seen": 65551968, "step": 34060 }, { "epoch": 5.557141691818256, "grad_norm": 0.0026510965544730425, "learning_rate": 2.66806841040782e-06, "loss": 0.0079, "num_input_tokens_seen": 65562880, "step": 34065 }, { "epoch": 5.557957419039074, "grad_norm": 0.0009585294756107032, "learning_rate": 2.6636571166753083e-06, "loss": 0.015, "num_input_tokens_seen": 65573552, "step": 34070 }, { "epoch": 5.558773146259891, "grad_norm": 0.012464320287108421, "learning_rate": 2.6592492674875598e-06, "loss": 0.0, "num_input_tokens_seen": 65582688, "step": 34075 }, { "epoch": 5.559588873480708, "grad_norm": 177.4068603515625, "learning_rate": 2.6548448635243305e-06, "loss": 0.0877, "num_input_tokens_seen": 65592320, "step": 34080 }, { "epoch": 5.560404600701525, "grad_norm": 0.6023803949356079, "learning_rate": 2.650443905464828e-06, "loss": 0.0007, "num_input_tokens_seen": 65602112, "step": 34085 }, { "epoch": 5.561220327922343, "grad_norm": 0.0003488771617412567, "learning_rate": 2.646046393987739e-06, "loss": 0.0004, "num_input_tokens_seen": 65611328, "step": 34090 }, { "epoch": 5.56203605514316, "grad_norm": 0.0002726101956795901, "learning_rate": 2.64165232977121e-06, "loss": 0.0773, "num_input_tokens_seen": 65620992, "step": 34095 }, { "epoch": 5.562851782363977, "grad_norm": 0.01144715677946806, "learning_rate": 2.6372617134928695e-06, "loss": 0.0, "num_input_tokens_seen": 65631200, "step": 34100 }, { "epoch": 5.5636675095847945, "grad_norm": 0.25471824407577515, "learning_rate": 2.6328745458297943e-06, "loss": 0.1856, "num_input_tokens_seen": 65639168, "step": 34105 }, { "epoch": 5.5644832368056125, "grad_norm": 17.431575775146484, "learning_rate": 2.6284908274585546e-06, "loss": 0.0808, "num_input_tokens_seen": 65647872, "step": 34110 }, { "epoch": 5.56529896402643, "grad_norm": 0.004181327298283577, "learning_rate": 2.6241105590551595e-06, "loss": 0.0002, "num_input_tokens_seen": 65656096, "step": 34115 }, { "epoch": 5.566114691247247, "grad_norm": 0.0026758587919175625, "learning_rate": 2.6197337412951105e-06, "loss": 0.0004, "num_input_tokens_seen": 65666528, "step": 34120 }, { "epoch": 5.566930418468064, "grad_norm": 0.29070863127708435, "learning_rate": 2.6153603748533705e-06, "loss": 0.0004, "num_input_tokens_seen": 65676240, "step": 34125 }, { "epoch": 5.567746145688882, "grad_norm": 0.8110777735710144, "learning_rate": 2.6109904604043585e-06, "loss": 0.0014, "num_input_tokens_seen": 65685424, "step": 34130 }, { "epoch": 5.568561872909699, "grad_norm": 0.0032734719570726156, "learning_rate": 2.6066239986219765e-06, "loss": 0.0, "num_input_tokens_seen": 65694768, "step": 34135 }, { "epoch": 5.569377600130516, "grad_norm": 0.056428201496601105, "learning_rate": 2.602260990179592e-06, "loss": 0.1023, "num_input_tokens_seen": 65704560, "step": 34140 }, { "epoch": 5.570193327351333, "grad_norm": 0.00010690031194826588, "learning_rate": 2.5979014357500248e-06, "loss": 0.0, "num_input_tokens_seen": 65714336, "step": 34145 }, { "epoch": 5.571009054572151, "grad_norm": 0.0023269320372492075, "learning_rate": 2.5935453360055844e-06, "loss": 0.0008, "num_input_tokens_seen": 65724176, "step": 34150 }, { "epoch": 5.5718247817929685, "grad_norm": 3.6386878490448, "learning_rate": 2.5891926916180283e-06, "loss": 0.0017, "num_input_tokens_seen": 65734048, "step": 34155 }, { "epoch": 5.572640509013786, "grad_norm": 0.002915432211011648, "learning_rate": 2.5848435032585883e-06, "loss": 0.0, "num_input_tokens_seen": 65743312, "step": 34160 }, { "epoch": 5.573456236234604, "grad_norm": 0.09645067155361176, "learning_rate": 2.58049777159797e-06, "loss": 0.0001, "num_input_tokens_seen": 65752368, "step": 34165 }, { "epoch": 5.574271963455421, "grad_norm": 0.0008067213348113, "learning_rate": 2.576155497306332e-06, "loss": 0.0001, "num_input_tokens_seen": 65761696, "step": 34170 }, { "epoch": 5.575087690676238, "grad_norm": 0.0005445702117867768, "learning_rate": 2.57181668105331e-06, "loss": 0.0671, "num_input_tokens_seen": 65771904, "step": 34175 }, { "epoch": 5.575903417897055, "grad_norm": 0.0031569949351251125, "learning_rate": 2.567481323508014e-06, "loss": 0.0005, "num_input_tokens_seen": 65780624, "step": 34180 }, { "epoch": 5.576719145117872, "grad_norm": 0.00029859167989343405, "learning_rate": 2.5631494253389954e-06, "loss": 0.0002, "num_input_tokens_seen": 65790640, "step": 34185 }, { "epoch": 5.57753487233869, "grad_norm": 17.99020767211914, "learning_rate": 2.5588209872142997e-06, "loss": 0.0742, "num_input_tokens_seen": 65800976, "step": 34190 }, { "epoch": 5.578350599559507, "grad_norm": 0.02916739508509636, "learning_rate": 2.5544960098014186e-06, "loss": 0.0007, "num_input_tokens_seen": 65810448, "step": 34195 }, { "epoch": 5.579166326780324, "grad_norm": 0.011232004500925541, "learning_rate": 2.550174493767318e-06, "loss": 0.0002, "num_input_tokens_seen": 65819600, "step": 34200 }, { "epoch": 5.579166326780324, "eval_loss": 0.4232428967952728, "eval_runtime": 80.7033, "eval_samples_per_second": 33.766, "eval_steps_per_second": 16.889, "num_input_tokens_seen": 65819600, "step": 34200 }, { "epoch": 5.5799820540011424, "grad_norm": 0.008586568757891655, "learning_rate": 2.545856439778438e-06, "loss": 0.0001, "num_input_tokens_seen": 65829056, "step": 34205 }, { "epoch": 5.58079778122196, "grad_norm": 0.0025846073403954506, "learning_rate": 2.541541848500667e-06, "loss": 0.0, "num_input_tokens_seen": 65839104, "step": 34210 }, { "epoch": 5.581613508442777, "grad_norm": 0.005546148866415024, "learning_rate": 2.5372307205993733e-06, "loss": 0.0, "num_input_tokens_seen": 65848048, "step": 34215 }, { "epoch": 5.582429235663594, "grad_norm": 0.017224598675966263, "learning_rate": 2.5329230567393917e-06, "loss": 0.0, "num_input_tokens_seen": 65857312, "step": 34220 }, { "epoch": 5.583244962884411, "grad_norm": 0.001593920635059476, "learning_rate": 2.5286188575850164e-06, "loss": 0.0001, "num_input_tokens_seen": 65867664, "step": 34225 }, { "epoch": 5.584060690105229, "grad_norm": 0.08144696801900864, "learning_rate": 2.5243181237999984e-06, "loss": 0.0005, "num_input_tokens_seen": 65877312, "step": 34230 }, { "epoch": 5.584876417326046, "grad_norm": 0.0006195150781422853, "learning_rate": 2.520020856047578e-06, "loss": 0.0006, "num_input_tokens_seen": 65886320, "step": 34235 }, { "epoch": 5.585692144546863, "grad_norm": 0.01700947992503643, "learning_rate": 2.515727054990438e-06, "loss": 0.0003, "num_input_tokens_seen": 65897552, "step": 34240 }, { "epoch": 5.586507871767681, "grad_norm": 0.009301673620939255, "learning_rate": 2.511436721290747e-06, "loss": 0.0007, "num_input_tokens_seen": 65907616, "step": 34245 }, { "epoch": 5.587323598988498, "grad_norm": 0.006484384648501873, "learning_rate": 2.5071498556101164e-06, "loss": 0.0, "num_input_tokens_seen": 65916528, "step": 34250 }, { "epoch": 5.5881393262093155, "grad_norm": 0.001854095607995987, "learning_rate": 2.5028664586096485e-06, "loss": 0.0001, "num_input_tokens_seen": 65925696, "step": 34255 }, { "epoch": 5.588955053430133, "grad_norm": 0.008942035026848316, "learning_rate": 2.498586530949881e-06, "loss": 0.0, "num_input_tokens_seen": 65935888, "step": 34260 }, { "epoch": 5.589770780650951, "grad_norm": 0.012764967978000641, "learning_rate": 2.4943100732908427e-06, "loss": 0.0001, "num_input_tokens_seen": 65943296, "step": 34265 }, { "epoch": 5.590586507871768, "grad_norm": 0.0015375122893601656, "learning_rate": 2.4900370862920188e-06, "loss": 0.0005, "num_input_tokens_seen": 65953344, "step": 34270 }, { "epoch": 5.591402235092585, "grad_norm": 0.0014671117532998323, "learning_rate": 2.4857675706123518e-06, "loss": 0.2063, "num_input_tokens_seen": 65962736, "step": 34275 }, { "epoch": 5.592217962313402, "grad_norm": 0.034251619130373, "learning_rate": 2.4815015269102543e-06, "loss": 0.0001, "num_input_tokens_seen": 65971808, "step": 34280 }, { "epoch": 5.59303368953422, "grad_norm": 0.07976346462965012, "learning_rate": 2.477238955843611e-06, "loss": 0.0001, "num_input_tokens_seen": 65980784, "step": 34285 }, { "epoch": 5.593849416755037, "grad_norm": 0.015306593850255013, "learning_rate": 2.4729798580697573e-06, "loss": 0.0087, "num_input_tokens_seen": 65991104, "step": 34290 }, { "epoch": 5.594665143975854, "grad_norm": 0.0022859808523207903, "learning_rate": 2.4687242342455034e-06, "loss": 0.043, "num_input_tokens_seen": 66000288, "step": 34295 }, { "epoch": 5.5954808711966715, "grad_norm": 0.00019661519036162645, "learning_rate": 2.4644720850271196e-06, "loss": 0.0, "num_input_tokens_seen": 66009760, "step": 34300 }, { "epoch": 5.5962965984174895, "grad_norm": 0.0006686635897494853, "learning_rate": 2.4602234110703364e-06, "loss": 0.0015, "num_input_tokens_seen": 66019488, "step": 34305 }, { "epoch": 5.597112325638307, "grad_norm": 0.000335508375428617, "learning_rate": 2.4559782130303576e-06, "loss": 0.0, "num_input_tokens_seen": 66028656, "step": 34310 }, { "epoch": 5.597928052859124, "grad_norm": 0.04062376171350479, "learning_rate": 2.451736491561843e-06, "loss": 0.0012, "num_input_tokens_seen": 66037616, "step": 34315 }, { "epoch": 5.598743780079941, "grad_norm": 0.0001334114494966343, "learning_rate": 2.4474982473189163e-06, "loss": 0.0001, "num_input_tokens_seen": 66048224, "step": 34320 }, { "epoch": 5.599559507300759, "grad_norm": 0.0032424945384263992, "learning_rate": 2.4432634809551796e-06, "loss": 0.0529, "num_input_tokens_seen": 66056656, "step": 34325 }, { "epoch": 5.600375234521576, "grad_norm": 0.0010142929386347532, "learning_rate": 2.439032193123675e-06, "loss": 0.0002, "num_input_tokens_seen": 66067040, "step": 34330 }, { "epoch": 5.601190961742393, "grad_norm": 0.002271645935252309, "learning_rate": 2.4348043844769297e-06, "loss": 0.0609, "num_input_tokens_seen": 66076816, "step": 34335 }, { "epoch": 5.602006688963211, "grad_norm": 0.00220847362652421, "learning_rate": 2.4305800556669146e-06, "loss": 0.0, "num_input_tokens_seen": 66085328, "step": 34340 }, { "epoch": 5.602822416184028, "grad_norm": 0.002726208418607712, "learning_rate": 2.426359207345083e-06, "loss": 0.0001, "num_input_tokens_seen": 66095408, "step": 34345 }, { "epoch": 5.6036381434048455, "grad_norm": 0.0268446896225214, "learning_rate": 2.4221418401623396e-06, "loss": 0.0001, "num_input_tokens_seen": 66104544, "step": 34350 }, { "epoch": 5.604453870625663, "grad_norm": 0.10239791125059128, "learning_rate": 2.4179279547690557e-06, "loss": 0.0004, "num_input_tokens_seen": 66113792, "step": 34355 }, { "epoch": 5.60526959784648, "grad_norm": 0.013039198704063892, "learning_rate": 2.413717551815062e-06, "loss": 0.0, "num_input_tokens_seen": 66122992, "step": 34360 }, { "epoch": 5.606085325067298, "grad_norm": 0.8140729665756226, "learning_rate": 2.409510631949666e-06, "loss": 0.0003, "num_input_tokens_seen": 66132992, "step": 34365 }, { "epoch": 5.606901052288115, "grad_norm": 30.34385871887207, "learning_rate": 2.405307195821618e-06, "loss": 0.1605, "num_input_tokens_seen": 66142992, "step": 34370 }, { "epoch": 5.607716779508932, "grad_norm": 0.026619957759976387, "learning_rate": 2.4011072440791372e-06, "loss": 0.063, "num_input_tokens_seen": 66151680, "step": 34375 }, { "epoch": 5.60853250672975, "grad_norm": 0.0007275252137333155, "learning_rate": 2.3969107773699233e-06, "loss": 0.0, "num_input_tokens_seen": 66161232, "step": 34380 }, { "epoch": 5.609348233950567, "grad_norm": 0.007574179675430059, "learning_rate": 2.3927177963411096e-06, "loss": 0.0001, "num_input_tokens_seen": 66171040, "step": 34385 }, { "epoch": 5.610163961171384, "grad_norm": 0.00025116815231740475, "learning_rate": 2.3885283016393144e-06, "loss": 0.0001, "num_input_tokens_seen": 66181152, "step": 34390 }, { "epoch": 5.610979688392201, "grad_norm": 0.006175695918500423, "learning_rate": 2.3843422939106076e-06, "loss": 0.0004, "num_input_tokens_seen": 66190528, "step": 34395 }, { "epoch": 5.6117954156130185, "grad_norm": 0.0003165092784911394, "learning_rate": 2.380159773800525e-06, "loss": 0.0002, "num_input_tokens_seen": 66199376, "step": 34400 }, { "epoch": 5.6117954156130185, "eval_loss": 0.43207570910453796, "eval_runtime": 80.8645, "eval_samples_per_second": 33.698, "eval_steps_per_second": 16.855, "num_input_tokens_seen": 66199376, "step": 34400 }, { "epoch": 5.6126111428338366, "grad_norm": 27.08694839477539, "learning_rate": 2.3759807419540675e-06, "loss": 0.1126, "num_input_tokens_seen": 66208752, "step": 34405 }, { "epoch": 5.613426870054654, "grad_norm": 0.0004606327274814248, "learning_rate": 2.3718051990156835e-06, "loss": 0.0003, "num_input_tokens_seen": 66217152, "step": 34410 }, { "epoch": 5.614242597275471, "grad_norm": 1.2059663534164429, "learning_rate": 2.367633145629311e-06, "loss": 0.0005, "num_input_tokens_seen": 66227520, "step": 34415 }, { "epoch": 5.615058324496289, "grad_norm": 6.813849176978692e-05, "learning_rate": 2.363464582438316e-06, "loss": 0.0001, "num_input_tokens_seen": 66236816, "step": 34420 }, { "epoch": 5.615874051717106, "grad_norm": 0.06286948174238205, "learning_rate": 2.3592995100855526e-06, "loss": 0.0001, "num_input_tokens_seen": 66246416, "step": 34425 }, { "epoch": 5.616689778937923, "grad_norm": 0.00048539412091486156, "learning_rate": 2.3551379292133273e-06, "loss": 0.0, "num_input_tokens_seen": 66255920, "step": 34430 }, { "epoch": 5.61750550615874, "grad_norm": 0.12515218555927277, "learning_rate": 2.3509798404634047e-06, "loss": 0.0001, "num_input_tokens_seen": 66265856, "step": 34435 }, { "epoch": 5.618321233379558, "grad_norm": 0.004960473161190748, "learning_rate": 2.346825244477019e-06, "loss": 0.0006, "num_input_tokens_seen": 66274656, "step": 34440 }, { "epoch": 5.619136960600375, "grad_norm": 0.11532081663608551, "learning_rate": 2.3426741418948545e-06, "loss": 0.0003, "num_input_tokens_seen": 66285424, "step": 34445 }, { "epoch": 5.6199526878211925, "grad_norm": 0.016643423587083817, "learning_rate": 2.3385265333570715e-06, "loss": 0.0001, "num_input_tokens_seen": 66295456, "step": 34450 }, { "epoch": 5.62076841504201, "grad_norm": 0.05104627087712288, "learning_rate": 2.334382419503278e-06, "loss": 0.0001, "num_input_tokens_seen": 66304928, "step": 34455 }, { "epoch": 5.621584142262828, "grad_norm": 0.003202929627150297, "learning_rate": 2.3302418009725465e-06, "loss": 0.0235, "num_input_tokens_seen": 66314000, "step": 34460 }, { "epoch": 5.622399869483645, "grad_norm": 0.023122865706682205, "learning_rate": 2.326104678403415e-06, "loss": 0.0355, "num_input_tokens_seen": 66324448, "step": 34465 }, { "epoch": 5.623215596704462, "grad_norm": 0.06286220997571945, "learning_rate": 2.321971052433883e-06, "loss": 0.0002, "num_input_tokens_seen": 66334256, "step": 34470 }, { "epoch": 5.624031323925279, "grad_norm": 0.0007961923838593066, "learning_rate": 2.3178409237014004e-06, "loss": 0.0001, "num_input_tokens_seen": 66343728, "step": 34475 }, { "epoch": 5.624847051146097, "grad_norm": 0.04530813917517662, "learning_rate": 2.313714292842889e-06, "loss": 0.0069, "num_input_tokens_seen": 66354800, "step": 34480 }, { "epoch": 5.625662778366914, "grad_norm": 0.0016849528765305877, "learning_rate": 2.309591160494734e-06, "loss": 0.0, "num_input_tokens_seen": 66364256, "step": 34485 }, { "epoch": 5.626478505587731, "grad_norm": 0.0020988532342016697, "learning_rate": 2.305471527292763e-06, "loss": 0.0, "num_input_tokens_seen": 66371792, "step": 34490 }, { "epoch": 5.6272942328085485, "grad_norm": 0.00030670425621792674, "learning_rate": 2.3013553938722817e-06, "loss": 0.0, "num_input_tokens_seen": 66381760, "step": 34495 }, { "epoch": 5.6281099600293665, "grad_norm": 0.0006370666669681668, "learning_rate": 2.297242760868043e-06, "loss": 0.1563, "num_input_tokens_seen": 66391184, "step": 34500 }, { "epoch": 5.628925687250184, "grad_norm": 0.19363515079021454, "learning_rate": 2.2931336289142735e-06, "loss": 0.0001, "num_input_tokens_seen": 66400816, "step": 34505 }, { "epoch": 5.629741414471001, "grad_norm": 0.0011998030822724104, "learning_rate": 2.289027998644655e-06, "loss": 0.0001, "num_input_tokens_seen": 66408848, "step": 34510 }, { "epoch": 5.630557141691818, "grad_norm": 0.020161299034953117, "learning_rate": 2.2849258706923228e-06, "loss": 0.0001, "num_input_tokens_seen": 66417792, "step": 34515 }, { "epoch": 5.631372868912636, "grad_norm": 0.0001485033571952954, "learning_rate": 2.2808272456898705e-06, "loss": 0.0, "num_input_tokens_seen": 66427408, "step": 34520 }, { "epoch": 5.632188596133453, "grad_norm": 0.00304078939370811, "learning_rate": 2.2767321242693707e-06, "loss": 0.0004, "num_input_tokens_seen": 66437824, "step": 34525 }, { "epoch": 5.63300432335427, "grad_norm": 0.10399207472801208, "learning_rate": 2.272640507062329e-06, "loss": 0.0002, "num_input_tokens_seen": 66446448, "step": 34530 }, { "epoch": 5.633820050575087, "grad_norm": 0.01899954304099083, "learning_rate": 2.2685523946997382e-06, "loss": 0.0, "num_input_tokens_seen": 66457488, "step": 34535 }, { "epoch": 5.634635777795905, "grad_norm": 0.002442468423396349, "learning_rate": 2.2644677878120245e-06, "loss": 0.0002, "num_input_tokens_seen": 66467744, "step": 34540 }, { "epoch": 5.635451505016722, "grad_norm": 0.015205609612166882, "learning_rate": 2.2603866870290897e-06, "loss": 0.0001, "num_input_tokens_seen": 66477776, "step": 34545 }, { "epoch": 5.63626723223754, "grad_norm": 0.0007120220689103007, "learning_rate": 2.256309092980294e-06, "loss": 0.0, "num_input_tokens_seen": 66488608, "step": 34550 }, { "epoch": 5.637082959458358, "grad_norm": 0.01625460758805275, "learning_rate": 2.252235006294448e-06, "loss": 0.0001, "num_input_tokens_seen": 66498544, "step": 34555 }, { "epoch": 5.637898686679175, "grad_norm": 0.005663703195750713, "learning_rate": 2.2481644275998333e-06, "loss": 0.0001, "num_input_tokens_seen": 66507776, "step": 34560 }, { "epoch": 5.638714413899992, "grad_norm": 0.06078421324491501, "learning_rate": 2.2440973575241832e-06, "loss": 0.0001, "num_input_tokens_seen": 66517792, "step": 34565 }, { "epoch": 5.639530141120809, "grad_norm": 0.00598903140053153, "learning_rate": 2.240033796694685e-06, "loss": 0.0001, "num_input_tokens_seen": 66528128, "step": 34570 }, { "epoch": 5.640345868341626, "grad_norm": 0.0027125696651637554, "learning_rate": 2.235973745737999e-06, "loss": 0.165, "num_input_tokens_seen": 66537984, "step": 34575 }, { "epoch": 5.641161595562444, "grad_norm": 0.0006385915330611169, "learning_rate": 2.2319172052802263e-06, "loss": 0.0, "num_input_tokens_seen": 66547904, "step": 34580 }, { "epoch": 5.641977322783261, "grad_norm": 0.0076702916994690895, "learning_rate": 2.2278641759469477e-06, "loss": 0.0001, "num_input_tokens_seen": 66556720, "step": 34585 }, { "epoch": 5.642793050004078, "grad_norm": 0.05705852434039116, "learning_rate": 2.2238146583631825e-06, "loss": 0.0002, "num_input_tokens_seen": 66565488, "step": 34590 }, { "epoch": 5.643608777224896, "grad_norm": 0.0005720813060179353, "learning_rate": 2.2197686531534256e-06, "loss": 0.0004, "num_input_tokens_seen": 66574256, "step": 34595 }, { "epoch": 5.6444245044457135, "grad_norm": 0.0005420442903414369, "learning_rate": 2.2157261609416087e-06, "loss": 0.0015, "num_input_tokens_seen": 66583936, "step": 34600 }, { "epoch": 5.6444245044457135, "eval_loss": 0.43429622054100037, "eval_runtime": 80.826, "eval_samples_per_second": 33.714, "eval_steps_per_second": 16.863, "num_input_tokens_seen": 66583936, "step": 34600 }, { "epoch": 5.645240231666531, "grad_norm": 0.00031163968378677964, "learning_rate": 2.211687182351149e-06, "loss": 0.0734, "num_input_tokens_seen": 66594032, "step": 34605 }, { "epoch": 5.646055958887348, "grad_norm": 0.0014913288177922368, "learning_rate": 2.2076517180048993e-06, "loss": 0.0096, "num_input_tokens_seen": 66603968, "step": 34610 }, { "epoch": 5.646871686108166, "grad_norm": 0.0035075952764600515, "learning_rate": 2.2036197685251834e-06, "loss": 0.0876, "num_input_tokens_seen": 66613360, "step": 34615 }, { "epoch": 5.647687413328983, "grad_norm": 0.0017381318612024188, "learning_rate": 2.199591334533771e-06, "loss": 0.0449, "num_input_tokens_seen": 66622496, "step": 34620 }, { "epoch": 5.6485031405498, "grad_norm": 5.071132659912109, "learning_rate": 2.1955664166519036e-06, "loss": 0.0017, "num_input_tokens_seen": 66632544, "step": 34625 }, { "epoch": 5.649318867770617, "grad_norm": 0.005006643943488598, "learning_rate": 2.1915450155002793e-06, "loss": 0.0001, "num_input_tokens_seen": 66642688, "step": 34630 }, { "epoch": 5.650134594991435, "grad_norm": 0.05387745797634125, "learning_rate": 2.187527131699038e-06, "loss": 0.0001, "num_input_tokens_seen": 66653168, "step": 34635 }, { "epoch": 5.650950322212252, "grad_norm": 0.0016984869726002216, "learning_rate": 2.18351276586779e-06, "loss": 0.0005, "num_input_tokens_seen": 66662848, "step": 34640 }, { "epoch": 5.6517660494330695, "grad_norm": 0.004280671011656523, "learning_rate": 2.1795019186256092e-06, "loss": 0.0, "num_input_tokens_seen": 66672528, "step": 34645 }, { "epoch": 5.652581776653887, "grad_norm": 0.0008302482892759144, "learning_rate": 2.1754945905910094e-06, "loss": 0.0, "num_input_tokens_seen": 66682320, "step": 34650 }, { "epoch": 5.653397503874705, "grad_norm": 0.019747748970985413, "learning_rate": 2.171490782381977e-06, "loss": 0.0067, "num_input_tokens_seen": 66691520, "step": 34655 }, { "epoch": 5.654213231095522, "grad_norm": 0.0053664217703044415, "learning_rate": 2.1674904946159425e-06, "loss": 0.0, "num_input_tokens_seen": 66700416, "step": 34660 }, { "epoch": 5.655028958316339, "grad_norm": 0.0015641620848327875, "learning_rate": 2.16349372790981e-06, "loss": 0.0001, "num_input_tokens_seen": 66710224, "step": 34665 }, { "epoch": 5.655844685537156, "grad_norm": 0.00244863610714674, "learning_rate": 2.159500482879928e-06, "loss": 0.0163, "num_input_tokens_seen": 66719072, "step": 34670 }, { "epoch": 5.656660412757974, "grad_norm": 0.004047784488648176, "learning_rate": 2.155510760142096e-06, "loss": 0.0001, "num_input_tokens_seen": 66728400, "step": 34675 }, { "epoch": 5.657476139978791, "grad_norm": 0.14833295345306396, "learning_rate": 2.151524560311588e-06, "loss": 0.0003, "num_input_tokens_seen": 66737824, "step": 34680 }, { "epoch": 5.658291867199608, "grad_norm": 0.0008342315559275448, "learning_rate": 2.147541884003129e-06, "loss": 0.0, "num_input_tokens_seen": 66746160, "step": 34685 }, { "epoch": 5.659107594420425, "grad_norm": 0.003410294186323881, "learning_rate": 2.1435627318308895e-06, "loss": 0.0003, "num_input_tokens_seen": 66756640, "step": 34690 }, { "epoch": 5.6599233216412435, "grad_norm": 0.0020529816392809153, "learning_rate": 2.139587104408511e-06, "loss": 0.0001, "num_input_tokens_seen": 66766496, "step": 34695 }, { "epoch": 5.660739048862061, "grad_norm": 0.0003370612394064665, "learning_rate": 2.1356150023490783e-06, "loss": 0.0001, "num_input_tokens_seen": 66775520, "step": 34700 }, { "epoch": 5.661554776082878, "grad_norm": 0.0025209097657352686, "learning_rate": 2.1316464262651464e-06, "loss": 0.0, "num_input_tokens_seen": 66786624, "step": 34705 }, { "epoch": 5.662370503303695, "grad_norm": 0.0014183612074702978, "learning_rate": 2.1276813767687224e-06, "loss": 0.0002, "num_input_tokens_seen": 66796528, "step": 34710 }, { "epoch": 5.663186230524513, "grad_norm": 0.0002339152997592464, "learning_rate": 2.123719854471254e-06, "loss": 0.0003, "num_input_tokens_seen": 66806336, "step": 34715 }, { "epoch": 5.66400195774533, "grad_norm": 0.002822627779096365, "learning_rate": 2.119761859983668e-06, "loss": 0.0, "num_input_tokens_seen": 66815728, "step": 34720 }, { "epoch": 5.664817684966147, "grad_norm": 0.0027950629591941833, "learning_rate": 2.1158073939163386e-06, "loss": 0.0004, "num_input_tokens_seen": 66825392, "step": 34725 }, { "epoch": 5.665633412186965, "grad_norm": 0.053809355944395065, "learning_rate": 2.111856456879088e-06, "loss": 0.0629, "num_input_tokens_seen": 66834624, "step": 34730 }, { "epoch": 5.666449139407782, "grad_norm": 0.004736187402158976, "learning_rate": 2.1079090494811993e-06, "loss": 0.0001, "num_input_tokens_seen": 66843856, "step": 34735 }, { "epoch": 5.667264866628599, "grad_norm": 0.3527417778968811, "learning_rate": 2.103965172331418e-06, "loss": 0.0003, "num_input_tokens_seen": 66853040, "step": 34740 }, { "epoch": 5.6680805938494165, "grad_norm": 0.0026544767897576094, "learning_rate": 2.100024826037933e-06, "loss": 0.0, "num_input_tokens_seen": 66862240, "step": 34745 }, { "epoch": 5.668896321070234, "grad_norm": 0.00019730794883798808, "learning_rate": 2.0960880112084027e-06, "loss": 0.1752, "num_input_tokens_seen": 66872128, "step": 34750 }, { "epoch": 5.669712048291052, "grad_norm": 0.0007762488094158471, "learning_rate": 2.092154728449927e-06, "loss": 0.0001, "num_input_tokens_seen": 66881344, "step": 34755 }, { "epoch": 5.670527775511869, "grad_norm": 0.019815238192677498, "learning_rate": 2.0882249783690687e-06, "loss": 0.0, "num_input_tokens_seen": 66890656, "step": 34760 }, { "epoch": 5.671343502732686, "grad_norm": 0.0016778809949755669, "learning_rate": 2.084298761571851e-06, "loss": 0.1042, "num_input_tokens_seen": 66900720, "step": 34765 }, { "epoch": 5.672159229953504, "grad_norm": 0.004625736735761166, "learning_rate": 2.080376078663737e-06, "loss": 0.0, "num_input_tokens_seen": 66910624, "step": 34770 }, { "epoch": 5.672974957174321, "grad_norm": 0.0011276863515377045, "learning_rate": 2.0764569302496593e-06, "loss": 0.1396, "num_input_tokens_seen": 66921440, "step": 34775 }, { "epoch": 5.673790684395138, "grad_norm": 8.216682181227952e-05, "learning_rate": 2.0725413169339957e-06, "loss": 0.045, "num_input_tokens_seen": 66931808, "step": 34780 }, { "epoch": 5.674606411615955, "grad_norm": 0.004087279085069895, "learning_rate": 2.068629239320588e-06, "loss": 0.0, "num_input_tokens_seen": 66940576, "step": 34785 }, { "epoch": 5.6754221388367725, "grad_norm": 0.0001060845170286484, "learning_rate": 2.064720698012726e-06, "loss": 0.0014, "num_input_tokens_seen": 66950304, "step": 34790 }, { "epoch": 5.6762378660575905, "grad_norm": 0.001785245374776423, "learning_rate": 2.0608156936131522e-06, "loss": 0.0, "num_input_tokens_seen": 66959792, "step": 34795 }, { "epoch": 5.677053593278408, "grad_norm": 0.015223968774080276, "learning_rate": 2.056914226724074e-06, "loss": 0.0, "num_input_tokens_seen": 66968960, "step": 34800 }, { "epoch": 5.677053593278408, "eval_loss": 0.4236864447593689, "eval_runtime": 80.6708, "eval_samples_per_second": 33.779, "eval_steps_per_second": 16.896, "num_input_tokens_seen": 66968960, "step": 34800 }, { "epoch": 5.677869320499225, "grad_norm": 0.0003819885023403913, "learning_rate": 2.0530162979471385e-06, "loss": 0.0191, "num_input_tokens_seen": 66978448, "step": 34805 }, { "epoch": 5.678685047720043, "grad_norm": 0.006558103021234274, "learning_rate": 2.0491219078834667e-06, "loss": 0.019, "num_input_tokens_seen": 66990528, "step": 34810 }, { "epoch": 5.67950077494086, "grad_norm": 0.000506250245962292, "learning_rate": 2.045231057133612e-06, "loss": 0.0794, "num_input_tokens_seen": 67000368, "step": 34815 }, { "epoch": 5.680316502161677, "grad_norm": 0.02343917079269886, "learning_rate": 2.0413437462975944e-06, "loss": 0.1084, "num_input_tokens_seen": 67009280, "step": 34820 }, { "epoch": 5.681132229382494, "grad_norm": 0.0004453643341548741, "learning_rate": 2.0374599759748843e-06, "loss": 0.0, "num_input_tokens_seen": 67018960, "step": 34825 }, { "epoch": 5.681947956603312, "grad_norm": 0.018644122406840324, "learning_rate": 2.033579746764419e-06, "loss": 0.0, "num_input_tokens_seen": 67027024, "step": 34830 }, { "epoch": 5.682763683824129, "grad_norm": 0.01832987554371357, "learning_rate": 2.029703059264565e-06, "loss": 0.0786, "num_input_tokens_seen": 67037456, "step": 34835 }, { "epoch": 5.6835794110449465, "grad_norm": 0.000741181371267885, "learning_rate": 2.02582991407316e-06, "loss": 0.0772, "num_input_tokens_seen": 67047088, "step": 34840 }, { "epoch": 5.684395138265764, "grad_norm": 0.28008800745010376, "learning_rate": 2.0219603117874992e-06, "loss": 0.0002, "num_input_tokens_seen": 67056208, "step": 34845 }, { "epoch": 5.685210865486582, "grad_norm": 0.0002070197369903326, "learning_rate": 2.0180942530043156e-06, "loss": 0.0917, "num_input_tokens_seen": 67065344, "step": 34850 }, { "epoch": 5.686026592707399, "grad_norm": 0.0020272487308830023, "learning_rate": 2.0142317383198107e-06, "loss": 0.0006, "num_input_tokens_seen": 67075344, "step": 34855 }, { "epoch": 5.686842319928216, "grad_norm": 0.013620263896882534, "learning_rate": 2.0103727683296243e-06, "loss": 0.0, "num_input_tokens_seen": 67085776, "step": 34860 }, { "epoch": 5.687658047149033, "grad_norm": 0.03826957941055298, "learning_rate": 2.0065173436288636e-06, "loss": 0.0001, "num_input_tokens_seen": 67094448, "step": 34865 }, { "epoch": 5.688473774369851, "grad_norm": 51.858917236328125, "learning_rate": 2.002665464812087e-06, "loss": 0.0815, "num_input_tokens_seen": 67105088, "step": 34870 }, { "epoch": 5.689289501590668, "grad_norm": 0.004135549999773502, "learning_rate": 1.998817132473291e-06, "loss": 0.0001, "num_input_tokens_seen": 67115760, "step": 34875 }, { "epoch": 5.690105228811485, "grad_norm": 0.0028280962724238634, "learning_rate": 1.9949723472059507e-06, "loss": 0.1438, "num_input_tokens_seen": 67125280, "step": 34880 }, { "epoch": 5.690920956032302, "grad_norm": 0.8142094612121582, "learning_rate": 1.9911311096029726e-06, "loss": 0.0017, "num_input_tokens_seen": 67136096, "step": 34885 }, { "epoch": 5.69173668325312, "grad_norm": 0.6228238344192505, "learning_rate": 1.9872934202567224e-06, "loss": 0.0023, "num_input_tokens_seen": 67145056, "step": 34890 }, { "epoch": 5.692552410473938, "grad_norm": 0.005021282006055117, "learning_rate": 1.9834592797590257e-06, "loss": 0.0412, "num_input_tokens_seen": 67154656, "step": 34895 }, { "epoch": 5.693368137694755, "grad_norm": 0.05628666281700134, "learning_rate": 1.979628688701149e-06, "loss": 0.0001, "num_input_tokens_seen": 67164656, "step": 34900 }, { "epoch": 5.694183864915573, "grad_norm": 0.010501094162464142, "learning_rate": 1.9758016476738193e-06, "loss": 0.0, "num_input_tokens_seen": 67176400, "step": 34905 }, { "epoch": 5.69499959213639, "grad_norm": 0.000197743545868434, "learning_rate": 1.971978157267221e-06, "loss": 0.1461, "num_input_tokens_seen": 67187376, "step": 34910 }, { "epoch": 5.695815319357207, "grad_norm": 0.5284741520881653, "learning_rate": 1.968158218070973e-06, "loss": 0.0007, "num_input_tokens_seen": 67197072, "step": 34915 }, { "epoch": 5.696631046578024, "grad_norm": 0.00023639520804863423, "learning_rate": 1.9643418306741682e-06, "loss": 0.0001, "num_input_tokens_seen": 67207328, "step": 34920 }, { "epoch": 5.697446773798841, "grad_norm": 0.0073517560958862305, "learning_rate": 1.9605289956653337e-06, "loss": 0.0, "num_input_tokens_seen": 67216768, "step": 34925 }, { "epoch": 5.698262501019659, "grad_norm": 0.0009272058377973735, "learning_rate": 1.9567197136324626e-06, "loss": 0.0028, "num_input_tokens_seen": 67226704, "step": 34930 }, { "epoch": 5.699078228240476, "grad_norm": 0.006661150138825178, "learning_rate": 1.9529139851629935e-06, "loss": 0.0, "num_input_tokens_seen": 67237584, "step": 34935 }, { "epoch": 5.6998939554612935, "grad_norm": 0.00011081222328357399, "learning_rate": 1.949111810843812e-06, "loss": 0.0001, "num_input_tokens_seen": 67247264, "step": 34940 }, { "epoch": 5.7007096826821115, "grad_norm": 0.00012545309436973184, "learning_rate": 1.9453131912612694e-06, "loss": 0.0001, "num_input_tokens_seen": 67257040, "step": 34945 }, { "epoch": 5.701525409902929, "grad_norm": 0.011492426507174969, "learning_rate": 1.941518127001149e-06, "loss": 0.0, "num_input_tokens_seen": 67267888, "step": 34950 }, { "epoch": 5.702341137123746, "grad_norm": 0.0005460525862872601, "learning_rate": 1.9377266186487107e-06, "loss": 0.0, "num_input_tokens_seen": 67277328, "step": 34955 }, { "epoch": 5.703156864344563, "grad_norm": 0.005582297686487436, "learning_rate": 1.9339386667886483e-06, "loss": 0.0, "num_input_tokens_seen": 67287712, "step": 34960 }, { "epoch": 5.70397259156538, "grad_norm": 0.0047483500093221664, "learning_rate": 1.9301542720051024e-06, "loss": 0.0, "num_input_tokens_seen": 67296864, "step": 34965 }, { "epoch": 5.704788318786198, "grad_norm": 0.001000634511001408, "learning_rate": 1.926373434881684e-06, "loss": 0.0004, "num_input_tokens_seen": 67305008, "step": 34970 }, { "epoch": 5.705604046007015, "grad_norm": 0.0016289422055706382, "learning_rate": 1.9225961560014468e-06, "loss": 0.0, "num_input_tokens_seen": 67313264, "step": 34975 }, { "epoch": 5.706419773227832, "grad_norm": 0.002368168206885457, "learning_rate": 1.918822435946885e-06, "loss": 0.0001, "num_input_tokens_seen": 67324128, "step": 34980 }, { "epoch": 5.70723550044865, "grad_norm": 0.0017208023928105831, "learning_rate": 1.915052275299961e-06, "loss": 0.0002, "num_input_tokens_seen": 67333056, "step": 34985 }, { "epoch": 5.7080512276694675, "grad_norm": 0.001414272584952414, "learning_rate": 1.9112856746420854e-06, "loss": 0.0, "num_input_tokens_seen": 67343120, "step": 34990 }, { "epoch": 5.708866954890285, "grad_norm": 0.0035479955840855837, "learning_rate": 1.907522634554104e-06, "loss": 0.0002, "num_input_tokens_seen": 67352720, "step": 34995 }, { "epoch": 5.709682682111102, "grad_norm": 0.0023573189973831177, "learning_rate": 1.9037631556163337e-06, "loss": 0.0001, "num_input_tokens_seen": 67361344, "step": 35000 }, { "epoch": 5.709682682111102, "eval_loss": 0.4073232114315033, "eval_runtime": 80.9234, "eval_samples_per_second": 33.674, "eval_steps_per_second": 16.843, "num_input_tokens_seen": 67361344, "step": 35000 }, { "epoch": 5.71049840933192, "grad_norm": 0.03243082016706467, "learning_rate": 1.9000072384085272e-06, "loss": 0.0001, "num_input_tokens_seen": 67370688, "step": 35005 }, { "epoch": 5.711314136552737, "grad_norm": 0.0021515660919249058, "learning_rate": 1.8962548835098987e-06, "loss": 0.1837, "num_input_tokens_seen": 67378512, "step": 35010 }, { "epoch": 5.712129863773554, "grad_norm": 0.004951545502990484, "learning_rate": 1.8925060914991077e-06, "loss": 0.0001, "num_input_tokens_seen": 67387536, "step": 35015 }, { "epoch": 5.712945590994371, "grad_norm": 0.009782088920474052, "learning_rate": 1.888760862954264e-06, "loss": 0.0, "num_input_tokens_seen": 67396752, "step": 35020 }, { "epoch": 5.713761318215189, "grad_norm": 0.042023301124572754, "learning_rate": 1.8850191984529309e-06, "loss": 0.0019, "num_input_tokens_seen": 67405632, "step": 35025 }, { "epoch": 5.714577045436006, "grad_norm": 0.17388562858104706, "learning_rate": 1.8812810985721186e-06, "loss": 0.0005, "num_input_tokens_seen": 67416496, "step": 35030 }, { "epoch": 5.715392772656823, "grad_norm": 0.0022425351198762655, "learning_rate": 1.8775465638882856e-06, "loss": 0.0001, "num_input_tokens_seen": 67425696, "step": 35035 }, { "epoch": 5.716208499877641, "grad_norm": 0.014727742411196232, "learning_rate": 1.8738155949773517e-06, "loss": 0.0, "num_input_tokens_seen": 67434704, "step": 35040 }, { "epoch": 5.717024227098459, "grad_norm": 0.00011154614912811667, "learning_rate": 1.8700881924146707e-06, "loss": 0.0005, "num_input_tokens_seen": 67443088, "step": 35045 }, { "epoch": 5.717839954319276, "grad_norm": 0.004393255803734064, "learning_rate": 1.8663643567750577e-06, "loss": 0.01, "num_input_tokens_seen": 67453984, "step": 35050 }, { "epoch": 5.718655681540093, "grad_norm": 0.0012673974270001054, "learning_rate": 1.8626440886327813e-06, "loss": 0.0, "num_input_tokens_seen": 67463584, "step": 35055 }, { "epoch": 5.71947140876091, "grad_norm": 0.2525550425052643, "learning_rate": 1.8589273885615432e-06, "loss": 0.0005, "num_input_tokens_seen": 67474416, "step": 35060 }, { "epoch": 5.720287135981728, "grad_norm": 0.003647042904049158, "learning_rate": 1.8552142571345133e-06, "loss": 0.0001, "num_input_tokens_seen": 67483920, "step": 35065 }, { "epoch": 5.721102863202545, "grad_norm": 0.0006760430405847728, "learning_rate": 1.8515046949243025e-06, "loss": 0.0, "num_input_tokens_seen": 67493824, "step": 35070 }, { "epoch": 5.721918590423362, "grad_norm": 0.03109326958656311, "learning_rate": 1.8477987025029674e-06, "loss": 0.0001, "num_input_tokens_seen": 67504336, "step": 35075 }, { "epoch": 5.72273431764418, "grad_norm": 0.01954587735235691, "learning_rate": 1.8440962804420232e-06, "loss": 0.0005, "num_input_tokens_seen": 67513920, "step": 35080 }, { "epoch": 5.723550044864997, "grad_norm": 0.001956944353878498, "learning_rate": 1.8403974293124265e-06, "loss": 0.0001, "num_input_tokens_seen": 67522608, "step": 35085 }, { "epoch": 5.7243657720858145, "grad_norm": 0.0022122461814433336, "learning_rate": 1.8367021496845854e-06, "loss": 0.0059, "num_input_tokens_seen": 67533040, "step": 35090 }, { "epoch": 5.725181499306632, "grad_norm": 2.9726661523454823e-05, "learning_rate": 1.8330104421283662e-06, "loss": 0.0778, "num_input_tokens_seen": 67542624, "step": 35095 }, { "epoch": 5.725997226527449, "grad_norm": 0.016950929537415504, "learning_rate": 1.8293223072130717e-06, "loss": 0.0, "num_input_tokens_seen": 67552720, "step": 35100 }, { "epoch": 5.726812953748267, "grad_norm": 0.003820565063506365, "learning_rate": 1.8256377455074525e-06, "loss": 0.0004, "num_input_tokens_seen": 67562576, "step": 35105 }, { "epoch": 5.727628680969084, "grad_norm": 0.0015769504243507981, "learning_rate": 1.8219567575797263e-06, "loss": 0.0003, "num_input_tokens_seen": 67572512, "step": 35110 }, { "epoch": 5.728444408189901, "grad_norm": 0.0007263029692694545, "learning_rate": 1.8182793439975365e-06, "loss": 0.0, "num_input_tokens_seen": 67582240, "step": 35115 }, { "epoch": 5.729260135410719, "grad_norm": 0.006716802716255188, "learning_rate": 1.8146055053279958e-06, "loss": 0.0059, "num_input_tokens_seen": 67592096, "step": 35120 }, { "epoch": 5.730075862631536, "grad_norm": 0.006368524394929409, "learning_rate": 1.8109352421376486e-06, "loss": 0.069, "num_input_tokens_seen": 67600672, "step": 35125 }, { "epoch": 5.730891589852353, "grad_norm": 0.5532575845718384, "learning_rate": 1.8072685549924972e-06, "loss": 0.0006, "num_input_tokens_seen": 67610816, "step": 35130 }, { "epoch": 5.7317073170731705, "grad_norm": 0.22028623521327972, "learning_rate": 1.8036054444579982e-06, "loss": 0.0, "num_input_tokens_seen": 67619984, "step": 35135 }, { "epoch": 5.732523044293988, "grad_norm": 6.934244447620586e-05, "learning_rate": 1.7999459110990407e-06, "loss": 0.0001, "num_input_tokens_seen": 67631120, "step": 35140 }, { "epoch": 5.733338771514806, "grad_norm": 0.00014358977205120027, "learning_rate": 1.7962899554799712e-06, "loss": 0.0001, "num_input_tokens_seen": 67641232, "step": 35145 }, { "epoch": 5.734154498735623, "grad_norm": 0.0039031922351568937, "learning_rate": 1.7926375781645937e-06, "loss": 0.0001, "num_input_tokens_seen": 67651568, "step": 35150 }, { "epoch": 5.73497022595644, "grad_norm": 0.0003916658170055598, "learning_rate": 1.7889887797161359e-06, "loss": 0.0, "num_input_tokens_seen": 67662560, "step": 35155 }, { "epoch": 5.735785953177258, "grad_norm": 15.491474151611328, "learning_rate": 1.7853435606973028e-06, "loss": 0.0602, "num_input_tokens_seen": 67672400, "step": 35160 }, { "epoch": 5.736601680398075, "grad_norm": 0.0042279870249331, "learning_rate": 1.781701921670223e-06, "loss": 0.0346, "num_input_tokens_seen": 67682272, "step": 35165 }, { "epoch": 5.737417407618892, "grad_norm": 7.408232340821996e-05, "learning_rate": 1.7780638631964886e-06, "loss": 0.0, "num_input_tokens_seen": 67692128, "step": 35170 }, { "epoch": 5.738233134839709, "grad_norm": 0.016666816547513008, "learning_rate": 1.7744293858371314e-06, "loss": 0.0001, "num_input_tokens_seen": 67700880, "step": 35175 }, { "epoch": 5.739048862060527, "grad_norm": 7.398685455322266, "learning_rate": 1.770798490152631e-06, "loss": 0.0176, "num_input_tokens_seen": 67710432, "step": 35180 }, { "epoch": 5.7398645892813445, "grad_norm": 0.00027188556850887835, "learning_rate": 1.767171176702917e-06, "loss": 0.0003, "num_input_tokens_seen": 67719680, "step": 35185 }, { "epoch": 5.740680316502162, "grad_norm": 0.003641789546236396, "learning_rate": 1.7635474460473755e-06, "loss": 0.0001, "num_input_tokens_seen": 67728400, "step": 35190 }, { "epoch": 5.741496043722979, "grad_norm": 0.00014994417142588645, "learning_rate": 1.7599272987448206e-06, "loss": 0.0001, "num_input_tokens_seen": 67736960, "step": 35195 }, { "epoch": 5.742311770943797, "grad_norm": 0.000802311347797513, "learning_rate": 1.7563107353535362e-06, "loss": 0.0024, "num_input_tokens_seen": 67746288, "step": 35200 }, { "epoch": 5.742311770943797, "eval_loss": 0.4150191843509674, "eval_runtime": 80.8831, "eval_samples_per_second": 33.691, "eval_steps_per_second": 16.851, "num_input_tokens_seen": 67746288, "step": 35200 }, { "epoch": 5.743127498164614, "grad_norm": 0.00034051266266033053, "learning_rate": 1.7526977564312263e-06, "loss": 0.0, "num_input_tokens_seen": 67756384, "step": 35205 }, { "epoch": 5.743943225385431, "grad_norm": 0.0010620758403092623, "learning_rate": 1.7490883625350701e-06, "loss": 0.0771, "num_input_tokens_seen": 67767360, "step": 35210 }, { "epoch": 5.744758952606248, "grad_norm": 0.0563153475522995, "learning_rate": 1.7454825542216807e-06, "loss": 0.0001, "num_input_tokens_seen": 67777456, "step": 35215 }, { "epoch": 5.745574679827066, "grad_norm": 0.07867935299873352, "learning_rate": 1.7418803320471105e-06, "loss": 0.0001, "num_input_tokens_seen": 67787472, "step": 35220 }, { "epoch": 5.746390407047883, "grad_norm": 4.198339462280273, "learning_rate": 1.7382816965668737e-06, "loss": 0.0028, "num_input_tokens_seen": 67797360, "step": 35225 }, { "epoch": 5.7472061342687, "grad_norm": 0.006921941414475441, "learning_rate": 1.7346866483359285e-06, "loss": 0.0981, "num_input_tokens_seen": 67807184, "step": 35230 }, { "epoch": 5.7480218614895175, "grad_norm": 0.0004688818589784205, "learning_rate": 1.7310951879086657e-06, "loss": 0.0007, "num_input_tokens_seen": 67817840, "step": 35235 }, { "epoch": 5.748837588710336, "grad_norm": 0.00028417102294042706, "learning_rate": 1.7275073158389471e-06, "loss": 0.0, "num_input_tokens_seen": 67827136, "step": 35240 }, { "epoch": 5.749653315931153, "grad_norm": 0.026444949209690094, "learning_rate": 1.723923032680061e-06, "loss": 0.0001, "num_input_tokens_seen": 67837280, "step": 35245 }, { "epoch": 5.75046904315197, "grad_norm": 0.00011674673442030326, "learning_rate": 1.7203423389847428e-06, "loss": 0.175, "num_input_tokens_seen": 67845888, "step": 35250 }, { "epoch": 5.751284770372787, "grad_norm": 0.001441081170924008, "learning_rate": 1.7167652353051928e-06, "loss": 0.0002, "num_input_tokens_seen": 67856640, "step": 35255 }, { "epoch": 5.752100497593605, "grad_norm": 57.650569915771484, "learning_rate": 1.7131917221930333e-06, "loss": 0.0918, "num_input_tokens_seen": 67865872, "step": 35260 }, { "epoch": 5.752916224814422, "grad_norm": 0.023072123527526855, "learning_rate": 1.7096218001993513e-06, "loss": 0.0001, "num_input_tokens_seen": 67875104, "step": 35265 }, { "epoch": 5.753731952035239, "grad_norm": 0.06030203774571419, "learning_rate": 1.706055469874676e-06, "loss": 0.0002, "num_input_tokens_seen": 67883520, "step": 35270 }, { "epoch": 5.754547679256056, "grad_norm": 0.03420248627662659, "learning_rate": 1.702492731768976e-06, "loss": 0.0, "num_input_tokens_seen": 67893632, "step": 35275 }, { "epoch": 5.755363406476874, "grad_norm": 0.00019640072423499078, "learning_rate": 1.6989335864316724e-06, "loss": 0.0003, "num_input_tokens_seen": 67902912, "step": 35280 }, { "epoch": 5.7561791336976915, "grad_norm": 0.00024461885914206505, "learning_rate": 1.6953780344116265e-06, "loss": 0.0, "num_input_tokens_seen": 67912960, "step": 35285 }, { "epoch": 5.756994860918509, "grad_norm": 0.002109758323058486, "learning_rate": 1.6918260762571497e-06, "loss": 0.045, "num_input_tokens_seen": 67923456, "step": 35290 }, { "epoch": 5.757810588139327, "grad_norm": 0.001155008445493877, "learning_rate": 1.6882777125160093e-06, "loss": 0.0, "num_input_tokens_seen": 67932544, "step": 35295 }, { "epoch": 5.758626315360144, "grad_norm": 13.537282943725586, "learning_rate": 1.6847329437353899e-06, "loss": 0.0571, "num_input_tokens_seen": 67941888, "step": 35300 }, { "epoch": 5.759442042580961, "grad_norm": 0.02160668559372425, "learning_rate": 1.6811917704619511e-06, "loss": 0.0672, "num_input_tokens_seen": 67950944, "step": 35305 }, { "epoch": 5.760257769801778, "grad_norm": 0.002867442788556218, "learning_rate": 1.67765419324179e-06, "loss": 0.0004, "num_input_tokens_seen": 67959216, "step": 35310 }, { "epoch": 5.761073497022595, "grad_norm": 0.0077324616722762585, "learning_rate": 1.6741202126204364e-06, "loss": 0.002, "num_input_tokens_seen": 67968912, "step": 35315 }, { "epoch": 5.761889224243413, "grad_norm": 73.47087097167969, "learning_rate": 1.6705898291428767e-06, "loss": 0.0508, "num_input_tokens_seen": 67978224, "step": 35320 }, { "epoch": 5.76270495146423, "grad_norm": 0.007945810444653034, "learning_rate": 1.6670630433535395e-06, "loss": 0.001, "num_input_tokens_seen": 67988352, "step": 35325 }, { "epoch": 5.7635206786850475, "grad_norm": 0.0009149198303930461, "learning_rate": 1.6635398557962979e-06, "loss": 0.0, "num_input_tokens_seen": 67996544, "step": 35330 }, { "epoch": 5.7643364059058655, "grad_norm": 0.00091234763385728, "learning_rate": 1.660020267014481e-06, "loss": 0.0, "num_input_tokens_seen": 68006208, "step": 35335 }, { "epoch": 5.765152133126683, "grad_norm": 0.035022564232349396, "learning_rate": 1.6565042775508438e-06, "loss": 0.0015, "num_input_tokens_seen": 68015392, "step": 35340 }, { "epoch": 5.7659678603475, "grad_norm": 0.07924036681652069, "learning_rate": 1.6529918879475997e-06, "loss": 0.0192, "num_input_tokens_seen": 68025632, "step": 35345 }, { "epoch": 5.766783587568317, "grad_norm": 0.0013157548382878304, "learning_rate": 1.6494830987464043e-06, "loss": 0.0, "num_input_tokens_seen": 68035104, "step": 35350 }, { "epoch": 5.767599314789135, "grad_norm": 0.005417463835328817, "learning_rate": 1.6459779104883555e-06, "loss": 0.1272, "num_input_tokens_seen": 68043504, "step": 35355 }, { "epoch": 5.768415042009952, "grad_norm": 0.0002775737957563251, "learning_rate": 1.6424763237140013e-06, "loss": 0.0025, "num_input_tokens_seen": 68052288, "step": 35360 }, { "epoch": 5.769230769230769, "grad_norm": 0.00838181097060442, "learning_rate": 1.6389783389633207e-06, "loss": 0.0002, "num_input_tokens_seen": 68062592, "step": 35365 }, { "epoch": 5.770046496451586, "grad_norm": 0.03316216543316841, "learning_rate": 1.6354839567757546e-06, "loss": 0.0153, "num_input_tokens_seen": 68072432, "step": 35370 }, { "epoch": 5.770862223672404, "grad_norm": 87.18426513671875, "learning_rate": 1.6319931776901831e-06, "loss": 0.0204, "num_input_tokens_seen": 68082816, "step": 35375 }, { "epoch": 5.771677950893221, "grad_norm": 0.08024285733699799, "learning_rate": 1.6285060022449229e-06, "loss": 0.0008, "num_input_tokens_seen": 68092512, "step": 35380 }, { "epoch": 5.772493678114039, "grad_norm": 0.00040631930460222065, "learning_rate": 1.6250224309777434e-06, "loss": 0.0508, "num_input_tokens_seen": 68103472, "step": 35385 }, { "epoch": 5.773309405334856, "grad_norm": 0.13054972887039185, "learning_rate": 1.6215424644258515e-06, "loss": 0.0037, "num_input_tokens_seen": 68111536, "step": 35390 }, { "epoch": 5.774125132555674, "grad_norm": 0.000645759457256645, "learning_rate": 1.6180661031259036e-06, "loss": 0.0002, "num_input_tokens_seen": 68121568, "step": 35395 }, { "epoch": 5.774940859776491, "grad_norm": 12.332720756530762, "learning_rate": 1.614593347613999e-06, "loss": 0.0138, "num_input_tokens_seen": 68131952, "step": 35400 }, { "epoch": 5.774940859776491, "eval_loss": 0.4070012867450714, "eval_runtime": 80.8156, "eval_samples_per_second": 33.719, "eval_steps_per_second": 16.866, "num_input_tokens_seen": 68131952, "step": 35400 }, { "epoch": 5.775756586997308, "grad_norm": 11.689114570617676, "learning_rate": 1.6111241984256758e-06, "loss": 0.1366, "num_input_tokens_seen": 68141456, "step": 35405 }, { "epoch": 5.776572314218125, "grad_norm": 0.0002820535155478865, "learning_rate": 1.6076586560959257e-06, "loss": 0.0003, "num_input_tokens_seen": 68151312, "step": 35410 }, { "epoch": 5.777388041438943, "grad_norm": 0.0030065798200666904, "learning_rate": 1.604196721159182e-06, "loss": 0.0005, "num_input_tokens_seen": 68160224, "step": 35415 }, { "epoch": 5.77820376865976, "grad_norm": 0.016216041520237923, "learning_rate": 1.6007383941493092e-06, "loss": 0.0002, "num_input_tokens_seen": 68170112, "step": 35420 }, { "epoch": 5.779019495880577, "grad_norm": 0.00025091334828175604, "learning_rate": 1.5972836755996285e-06, "loss": 0.0795, "num_input_tokens_seen": 68178400, "step": 35425 }, { "epoch": 5.7798352231013945, "grad_norm": 0.0008684471831656992, "learning_rate": 1.5938325660429076e-06, "loss": 0.0002, "num_input_tokens_seen": 68188208, "step": 35430 }, { "epoch": 5.7806509503222125, "grad_norm": 0.04678504541516304, "learning_rate": 1.5903850660113378e-06, "loss": 0.0054, "num_input_tokens_seen": 68197728, "step": 35435 }, { "epoch": 5.78146667754303, "grad_norm": 0.002001761691644788, "learning_rate": 1.5869411760365826e-06, "loss": 0.0, "num_input_tokens_seen": 68206768, "step": 35440 }, { "epoch": 5.782282404763847, "grad_norm": 0.008107738569378853, "learning_rate": 1.58350089664972e-06, "loss": 0.0, "num_input_tokens_seen": 68216432, "step": 35445 }, { "epoch": 5.783098131984664, "grad_norm": 0.0020015649497509003, "learning_rate": 1.5800642283812865e-06, "loss": 0.0001, "num_input_tokens_seen": 68226512, "step": 35450 }, { "epoch": 5.783913859205482, "grad_norm": 0.004109067842364311, "learning_rate": 1.5766311717612698e-06, "loss": 0.0, "num_input_tokens_seen": 68236448, "step": 35455 }, { "epoch": 5.784729586426299, "grad_norm": 0.0008403575629927218, "learning_rate": 1.5732017273190818e-06, "loss": 0.0001, "num_input_tokens_seen": 68246624, "step": 35460 }, { "epoch": 5.785545313647116, "grad_norm": 0.003677025903016329, "learning_rate": 1.5697758955835806e-06, "loss": 0.0005, "num_input_tokens_seen": 68256688, "step": 35465 }, { "epoch": 5.786361040867934, "grad_norm": 0.0006052427343092859, "learning_rate": 1.566353677083085e-06, "loss": 0.0005, "num_input_tokens_seen": 68266160, "step": 35470 }, { "epoch": 5.787176768088751, "grad_norm": 0.0018337809015065432, "learning_rate": 1.562935072345334e-06, "loss": 0.0001, "num_input_tokens_seen": 68275552, "step": 35475 }, { "epoch": 5.7879924953095685, "grad_norm": 0.0015688358107581735, "learning_rate": 1.5595200818975281e-06, "loss": 0.0009, "num_input_tokens_seen": 68284576, "step": 35480 }, { "epoch": 5.788808222530386, "grad_norm": 0.004171093460172415, "learning_rate": 1.5561087062662905e-06, "loss": 0.0096, "num_input_tokens_seen": 68293680, "step": 35485 }, { "epoch": 5.789623949751203, "grad_norm": 0.014494583941996098, "learning_rate": 1.5527009459777087e-06, "loss": 0.0609, "num_input_tokens_seen": 68303056, "step": 35490 }, { "epoch": 5.790439676972021, "grad_norm": 5.361377239227295, "learning_rate": 1.5492968015572984e-06, "loss": 0.0017, "num_input_tokens_seen": 68314208, "step": 35495 }, { "epoch": 5.791255404192838, "grad_norm": 0.0012403653236106038, "learning_rate": 1.5458962735300203e-06, "loss": 0.0003, "num_input_tokens_seen": 68325040, "step": 35500 }, { "epoch": 5.792071131413655, "grad_norm": 0.013984354212880135, "learning_rate": 1.54249936242028e-06, "loss": 0.0, "num_input_tokens_seen": 68334320, "step": 35505 }, { "epoch": 5.792886858634473, "grad_norm": 4.1172003746032715, "learning_rate": 1.5391060687519222e-06, "loss": 0.0046, "num_input_tokens_seen": 68344112, "step": 35510 }, { "epoch": 5.79370258585529, "grad_norm": 0.0004085782275069505, "learning_rate": 1.5357163930482367e-06, "loss": 0.0001, "num_input_tokens_seen": 68354176, "step": 35515 }, { "epoch": 5.794518313076107, "grad_norm": 0.013784394599497318, "learning_rate": 1.532330335831955e-06, "loss": 0.0002, "num_input_tokens_seen": 68364016, "step": 35520 }, { "epoch": 5.7953340402969244, "grad_norm": 0.6755465269088745, "learning_rate": 1.5289478976252491e-06, "loss": 0.0007, "num_input_tokens_seen": 68372544, "step": 35525 }, { "epoch": 5.796149767517742, "grad_norm": 0.05875735357403755, "learning_rate": 1.5255690789497345e-06, "loss": 0.1002, "num_input_tokens_seen": 68381936, "step": 35530 }, { "epoch": 5.79696549473856, "grad_norm": 0.0013826297363266349, "learning_rate": 1.5221938803264641e-06, "loss": 0.0001, "num_input_tokens_seen": 68392272, "step": 35535 }, { "epoch": 5.797781221959377, "grad_norm": 0.0002276280865771696, "learning_rate": 1.518822302275938e-06, "loss": 0.0115, "num_input_tokens_seen": 68402000, "step": 35540 }, { "epoch": 5.798596949180194, "grad_norm": 0.00120931351557374, "learning_rate": 1.5154543453180958e-06, "loss": 0.0011, "num_input_tokens_seen": 68410896, "step": 35545 }, { "epoch": 5.799412676401012, "grad_norm": 0.001087793381884694, "learning_rate": 1.5120900099723167e-06, "loss": 0.0127, "num_input_tokens_seen": 68420944, "step": 35550 }, { "epoch": 5.800228403621829, "grad_norm": 0.004828981589525938, "learning_rate": 1.5087292967574273e-06, "loss": 0.0008, "num_input_tokens_seen": 68429664, "step": 35555 }, { "epoch": 5.801044130842646, "grad_norm": 0.033584192395210266, "learning_rate": 1.5053722061916908e-06, "loss": 0.0006, "num_input_tokens_seen": 68439200, "step": 35560 }, { "epoch": 5.801859858063463, "grad_norm": 0.007758800871670246, "learning_rate": 1.5020187387928124e-06, "loss": 0.0489, "num_input_tokens_seen": 68447856, "step": 35565 }, { "epoch": 5.802675585284281, "grad_norm": 0.006724435370415449, "learning_rate": 1.4986688950779343e-06, "loss": 0.0001, "num_input_tokens_seen": 68456720, "step": 35570 }, { "epoch": 5.803491312505098, "grad_norm": 0.00045448497985489666, "learning_rate": 1.495322675563654e-06, "loss": 0.0072, "num_input_tokens_seen": 68466800, "step": 35575 }, { "epoch": 5.8043070397259156, "grad_norm": 0.004532771185040474, "learning_rate": 1.4919800807659922e-06, "loss": 0.061, "num_input_tokens_seen": 68476496, "step": 35580 }, { "epoch": 5.805122766946733, "grad_norm": 0.08929018676280975, "learning_rate": 1.4886411112004255e-06, "loss": 0.0001, "num_input_tokens_seen": 68486432, "step": 35585 }, { "epoch": 5.805938494167551, "grad_norm": 0.02685840241611004, "learning_rate": 1.4853057673818588e-06, "loss": 0.0001, "num_input_tokens_seen": 68495568, "step": 35590 }, { "epoch": 5.806754221388368, "grad_norm": 0.01207764446735382, "learning_rate": 1.481974049824647e-06, "loss": 0.1403, "num_input_tokens_seen": 68504480, "step": 35595 }, { "epoch": 5.807569948609185, "grad_norm": 0.008362608030438423, "learning_rate": 1.4786459590425849e-06, "loss": 0.0, "num_input_tokens_seen": 68514656, "step": 35600 }, { "epoch": 5.807569948609185, "eval_loss": 0.4128159284591675, "eval_runtime": 80.8354, "eval_samples_per_second": 33.71, "eval_steps_per_second": 16.861, "num_input_tokens_seen": 68514656, "step": 35600 }, { "epoch": 5.808385675830002, "grad_norm": 0.01793845184147358, "learning_rate": 1.4753214955489036e-06, "loss": 0.0001, "num_input_tokens_seen": 68524416, "step": 35605 }, { "epoch": 5.80920140305082, "grad_norm": 0.02436726912856102, "learning_rate": 1.4720006598562737e-06, "loss": 0.0836, "num_input_tokens_seen": 68534848, "step": 35610 }, { "epoch": 5.810017130271637, "grad_norm": 0.005390414036810398, "learning_rate": 1.4686834524768185e-06, "loss": 0.0488, "num_input_tokens_seen": 68545344, "step": 35615 }, { "epoch": 5.810832857492454, "grad_norm": 0.00013437285088002682, "learning_rate": 1.4653698739220844e-06, "loss": 0.0, "num_input_tokens_seen": 68553952, "step": 35620 }, { "epoch": 5.8116485847132715, "grad_norm": 0.11392287164926529, "learning_rate": 1.4620599247030715e-06, "loss": 0.0001, "num_input_tokens_seen": 68564416, "step": 35625 }, { "epoch": 5.8124643119340895, "grad_norm": 0.05169738084077835, "learning_rate": 1.4587536053302125e-06, "loss": 0.0003, "num_input_tokens_seen": 68573360, "step": 35630 }, { "epoch": 5.813280039154907, "grad_norm": 0.1717449277639389, "learning_rate": 1.4554509163133862e-06, "loss": 0.0003, "num_input_tokens_seen": 68581856, "step": 35635 }, { "epoch": 5.814095766375724, "grad_norm": 0.00437556579709053, "learning_rate": 1.4521518581619098e-06, "loss": 0.0002, "num_input_tokens_seen": 68591456, "step": 35640 }, { "epoch": 5.814911493596542, "grad_norm": 0.018583793193101883, "learning_rate": 1.4488564313845348e-06, "loss": 0.0876, "num_input_tokens_seen": 68600400, "step": 35645 }, { "epoch": 5.815727220817359, "grad_norm": 0.02636311948299408, "learning_rate": 1.4455646364894603e-06, "loss": 0.0002, "num_input_tokens_seen": 68609504, "step": 35650 }, { "epoch": 5.816542948038176, "grad_norm": 0.03219575807452202, "learning_rate": 1.4422764739843247e-06, "loss": 0.0002, "num_input_tokens_seen": 68618256, "step": 35655 }, { "epoch": 5.817358675258993, "grad_norm": 0.0009431563084945083, "learning_rate": 1.4389919443762e-06, "loss": 0.0001, "num_input_tokens_seen": 68626992, "step": 35660 }, { "epoch": 5.81817440247981, "grad_norm": 0.007610204163938761, "learning_rate": 1.4357110481716063e-06, "loss": 0.0007, "num_input_tokens_seen": 68638032, "step": 35665 }, { "epoch": 5.818990129700628, "grad_norm": 0.017003916203975677, "learning_rate": 1.4324337858764941e-06, "loss": 0.0728, "num_input_tokens_seen": 68646784, "step": 35670 }, { "epoch": 5.8198058569214455, "grad_norm": 0.018243273720145226, "learning_rate": 1.4291601579962622e-06, "loss": 0.0005, "num_input_tokens_seen": 68656560, "step": 35675 }, { "epoch": 5.820621584142263, "grad_norm": 0.007134707178920507, "learning_rate": 1.42589016503574e-06, "loss": 0.0004, "num_input_tokens_seen": 68666272, "step": 35680 }, { "epoch": 5.821437311363081, "grad_norm": 0.0010278299450874329, "learning_rate": 1.4226238074992099e-06, "loss": 0.0691, "num_input_tokens_seen": 68677104, "step": 35685 }, { "epoch": 5.822253038583898, "grad_norm": 0.006524089258164167, "learning_rate": 1.4193610858903778e-06, "loss": 0.0005, "num_input_tokens_seen": 68686176, "step": 35690 }, { "epoch": 5.823068765804715, "grad_norm": 0.012002323754131794, "learning_rate": 1.416102000712402e-06, "loss": 0.0001, "num_input_tokens_seen": 68696352, "step": 35695 }, { "epoch": 5.823884493025532, "grad_norm": 0.14333637058734894, "learning_rate": 1.4128465524678668e-06, "loss": 0.0001, "num_input_tokens_seen": 68703376, "step": 35700 }, { "epoch": 5.824700220246349, "grad_norm": 0.0052132634446024895, "learning_rate": 1.4095947416588124e-06, "loss": 0.0001, "num_input_tokens_seen": 68713296, "step": 35705 }, { "epoch": 5.825515947467167, "grad_norm": 0.1278913915157318, "learning_rate": 1.4063465687866983e-06, "loss": 0.0001, "num_input_tokens_seen": 68724608, "step": 35710 }, { "epoch": 5.826331674687984, "grad_norm": 0.10133115947246552, "learning_rate": 1.4031020343524438e-06, "loss": 0.0002, "num_input_tokens_seen": 68734736, "step": 35715 }, { "epoch": 5.827147401908801, "grad_norm": 0.0007220886182039976, "learning_rate": 1.3998611388563926e-06, "loss": 0.0, "num_input_tokens_seen": 68745744, "step": 35720 }, { "epoch": 5.8279631291296194, "grad_norm": 0.07524703443050385, "learning_rate": 1.3966238827983314e-06, "loss": 0.0002, "num_input_tokens_seen": 68756720, "step": 35725 }, { "epoch": 5.828778856350437, "grad_norm": 0.05090766400098801, "learning_rate": 1.393390266677483e-06, "loss": 0.0019, "num_input_tokens_seen": 68765376, "step": 35730 }, { "epoch": 5.829594583571254, "grad_norm": 0.011276155710220337, "learning_rate": 1.3901602909925204e-06, "loss": 0.0, "num_input_tokens_seen": 68776352, "step": 35735 }, { "epoch": 5.830410310792071, "grad_norm": 0.032255325466394424, "learning_rate": 1.3869339562415373e-06, "loss": 0.001, "num_input_tokens_seen": 68786368, "step": 35740 }, { "epoch": 5.831226038012889, "grad_norm": 0.006121320184320211, "learning_rate": 1.38371126292208e-06, "loss": 0.0, "num_input_tokens_seen": 68796800, "step": 35745 }, { "epoch": 5.832041765233706, "grad_norm": 30.339590072631836, "learning_rate": 1.3804922115311286e-06, "loss": 0.0672, "num_input_tokens_seen": 68806240, "step": 35750 }, { "epoch": 5.832857492454523, "grad_norm": 0.27656418085098267, "learning_rate": 1.3772768025650945e-06, "loss": 0.0003, "num_input_tokens_seen": 68815184, "step": 35755 }, { "epoch": 5.83367321967534, "grad_norm": 0.0032046320848166943, "learning_rate": 1.3740650365198448e-06, "loss": 0.0002, "num_input_tokens_seen": 68825488, "step": 35760 }, { "epoch": 5.834488946896158, "grad_norm": 0.3569280505180359, "learning_rate": 1.3708569138906612e-06, "loss": 0.001, "num_input_tokens_seen": 68834560, "step": 35765 }, { "epoch": 5.835304674116975, "grad_norm": 0.009556412696838379, "learning_rate": 1.367652435172287e-06, "loss": 0.0003, "num_input_tokens_seen": 68845072, "step": 35770 }, { "epoch": 5.8361204013377925, "grad_norm": 0.022116603329777718, "learning_rate": 1.364451600858893e-06, "loss": 0.0002, "num_input_tokens_seen": 68855392, "step": 35775 }, { "epoch": 5.83693612855861, "grad_norm": 0.010362355969846249, "learning_rate": 1.3612544114440823e-06, "loss": 0.0006, "num_input_tokens_seen": 68866128, "step": 35780 }, { "epoch": 5.837751855779428, "grad_norm": 0.021388670429587364, "learning_rate": 1.3580608674209072e-06, "loss": 0.0001, "num_input_tokens_seen": 68875280, "step": 35785 }, { "epoch": 5.838567583000245, "grad_norm": 0.2831452488899231, "learning_rate": 1.3548709692818434e-06, "loss": 0.0033, "num_input_tokens_seen": 68883088, "step": 35790 }, { "epoch": 5.839383310221062, "grad_norm": 0.7247372269630432, "learning_rate": 1.3516847175188223e-06, "loss": 0.0072, "num_input_tokens_seen": 68894112, "step": 35795 }, { "epoch": 5.840199037441879, "grad_norm": 0.07774128019809723, "learning_rate": 1.348502112623204e-06, "loss": 0.0003, "num_input_tokens_seen": 68904544, "step": 35800 }, { "epoch": 5.840199037441879, "eval_loss": 0.41697007417678833, "eval_runtime": 80.7808, "eval_samples_per_second": 33.733, "eval_steps_per_second": 16.873, "num_input_tokens_seen": 68904544, "step": 35800 }, { "epoch": 5.841014764662697, "grad_norm": 0.0006544351927004755, "learning_rate": 1.3453231550857787e-06, "loss": 0.15, "num_input_tokens_seen": 68914592, "step": 35805 }, { "epoch": 5.841830491883514, "grad_norm": 0.01029796339571476, "learning_rate": 1.3421478453967878e-06, "loss": 0.0001, "num_input_tokens_seen": 68924256, "step": 35810 }, { "epoch": 5.842646219104331, "grad_norm": 0.012277159839868546, "learning_rate": 1.3389761840459065e-06, "loss": 0.0001, "num_input_tokens_seen": 68934032, "step": 35815 }, { "epoch": 5.8434619463251485, "grad_norm": 0.00045035689254291356, "learning_rate": 1.3358081715222376e-06, "loss": 0.0, "num_input_tokens_seen": 68943408, "step": 35820 }, { "epoch": 5.8442776735459665, "grad_norm": 0.19453823566436768, "learning_rate": 1.3326438083143295e-06, "loss": 0.0003, "num_input_tokens_seen": 68952304, "step": 35825 }, { "epoch": 5.845093400766784, "grad_norm": 0.034052032977342606, "learning_rate": 1.3294830949101723e-06, "loss": 0.0001, "num_input_tokens_seen": 68961488, "step": 35830 }, { "epoch": 5.845909127987601, "grad_norm": 0.0013797353021800518, "learning_rate": 1.3263260317971815e-06, "loss": 0.098, "num_input_tokens_seen": 68971008, "step": 35835 }, { "epoch": 5.846724855208418, "grad_norm": 0.05214092880487442, "learning_rate": 1.3231726194622208e-06, "loss": 0.0001, "num_input_tokens_seen": 68979376, "step": 35840 }, { "epoch": 5.847540582429236, "grad_norm": 0.030942775309085846, "learning_rate": 1.3200228583915814e-06, "loss": 0.0001, "num_input_tokens_seen": 68988640, "step": 35845 }, { "epoch": 5.848356309650053, "grad_norm": 0.022891614586114883, "learning_rate": 1.3168767490709971e-06, "loss": 0.0016, "num_input_tokens_seen": 68997440, "step": 35850 }, { "epoch": 5.84917203687087, "grad_norm": 0.011679514311254025, "learning_rate": 1.3137342919856437e-06, "loss": 0.0001, "num_input_tokens_seen": 69006896, "step": 35855 }, { "epoch": 5.849987764091688, "grad_norm": 0.02894728071987629, "learning_rate": 1.310595487620117e-06, "loss": 0.0877, "num_input_tokens_seen": 69016464, "step": 35860 }, { "epoch": 5.850803491312505, "grad_norm": 0.052372466772794724, "learning_rate": 1.3074603364584715e-06, "loss": 0.1919, "num_input_tokens_seen": 69025424, "step": 35865 }, { "epoch": 5.8516192185333225, "grad_norm": 0.1823301762342453, "learning_rate": 1.3043288389841758e-06, "loss": 0.0338, "num_input_tokens_seen": 69035440, "step": 35870 }, { "epoch": 5.85243494575414, "grad_norm": 0.004164736717939377, "learning_rate": 1.3012009956801546e-06, "loss": 0.0005, "num_input_tokens_seen": 69044128, "step": 35875 }, { "epoch": 5.853250672974957, "grad_norm": 0.03332589939236641, "learning_rate": 1.2980768070287586e-06, "loss": 0.0008, "num_input_tokens_seen": 69054752, "step": 35880 }, { "epoch": 5.854066400195775, "grad_norm": 0.005793763790279627, "learning_rate": 1.2949562735117716e-06, "loss": 0.0001, "num_input_tokens_seen": 69063216, "step": 35885 }, { "epoch": 5.854882127416592, "grad_norm": 0.0005968720652163029, "learning_rate": 1.291839395610428e-06, "loss": 0.0, "num_input_tokens_seen": 69072288, "step": 35890 }, { "epoch": 5.855697854637409, "grad_norm": 23.337541580200195, "learning_rate": 1.2887261738053852e-06, "loss": 0.0027, "num_input_tokens_seen": 69082304, "step": 35895 }, { "epoch": 5.856513581858227, "grad_norm": 0.0027366974391043186, "learning_rate": 1.2856166085767396e-06, "loss": 0.0, "num_input_tokens_seen": 69092528, "step": 35900 }, { "epoch": 5.857329309079044, "grad_norm": 0.28557440638542175, "learning_rate": 1.2825107004040272e-06, "loss": 0.0006, "num_input_tokens_seen": 69100800, "step": 35905 }, { "epoch": 5.858145036299861, "grad_norm": 6.871426105499268, "learning_rate": 1.2794084497662146e-06, "loss": 0.0301, "num_input_tokens_seen": 69110192, "step": 35910 }, { "epoch": 5.858960763520678, "grad_norm": 0.0021685000974684954, "learning_rate": 1.276309857141711e-06, "loss": 0.0, "num_input_tokens_seen": 69119808, "step": 35915 }, { "epoch": 5.859776490741496, "grad_norm": 0.0015514480182901025, "learning_rate": 1.273214923008359e-06, "loss": 0.0, "num_input_tokens_seen": 69130016, "step": 35920 }, { "epoch": 5.8605922179623136, "grad_norm": 0.024640263989567757, "learning_rate": 1.2701236478434352e-06, "loss": 0.0003, "num_input_tokens_seen": 69138272, "step": 35925 }, { "epoch": 5.861407945183131, "grad_norm": 0.00047124424600042403, "learning_rate": 1.2670360321236502e-06, "loss": 0.0001, "num_input_tokens_seen": 69148512, "step": 35930 }, { "epoch": 5.862223672403948, "grad_norm": 0.010630698874592781, "learning_rate": 1.2639520763251617e-06, "loss": 0.0609, "num_input_tokens_seen": 69158784, "step": 35935 }, { "epoch": 5.863039399624766, "grad_norm": 0.002444202546030283, "learning_rate": 1.2608717809235448e-06, "loss": 0.0002, "num_input_tokens_seen": 69168912, "step": 35940 }, { "epoch": 5.863855126845583, "grad_norm": 0.0014985863817855716, "learning_rate": 1.2577951463938282e-06, "loss": 0.0002, "num_input_tokens_seen": 69180240, "step": 35945 }, { "epoch": 5.8646708540664, "grad_norm": 0.11506868898868561, "learning_rate": 1.2547221732104569e-06, "loss": 0.0008, "num_input_tokens_seen": 69190752, "step": 35950 }, { "epoch": 5.865486581287217, "grad_norm": 0.004304185044020414, "learning_rate": 1.25165286184733e-06, "loss": 0.0, "num_input_tokens_seen": 69199968, "step": 35955 }, { "epoch": 5.866302308508035, "grad_norm": 0.02520868368446827, "learning_rate": 1.248587212777777e-06, "loss": 0.0001, "num_input_tokens_seen": 69209552, "step": 35960 }, { "epoch": 5.867118035728852, "grad_norm": 0.012443888932466507, "learning_rate": 1.2455252264745532e-06, "loss": 0.0007, "num_input_tokens_seen": 69218976, "step": 35965 }, { "epoch": 5.8679337629496695, "grad_norm": 0.001557537936605513, "learning_rate": 1.2424669034098528e-06, "loss": 0.0002, "num_input_tokens_seen": 69228976, "step": 35970 }, { "epoch": 5.868749490170487, "grad_norm": 0.0008785572135820985, "learning_rate": 1.2394122440553185e-06, "loss": 0.0001, "num_input_tokens_seen": 69238416, "step": 35975 }, { "epoch": 5.869565217391305, "grad_norm": 0.016476793214678764, "learning_rate": 1.2363612488820037e-06, "loss": 0.0002, "num_input_tokens_seen": 69248032, "step": 35980 }, { "epoch": 5.870380944612122, "grad_norm": 0.003485153429210186, "learning_rate": 1.2333139183604208e-06, "loss": 0.0003, "num_input_tokens_seen": 69257328, "step": 35985 }, { "epoch": 5.871196671832939, "grad_norm": 0.008910558186471462, "learning_rate": 1.2302702529604998e-06, "loss": 0.0037, "num_input_tokens_seen": 69267056, "step": 35990 }, { "epoch": 5.872012399053756, "grad_norm": 0.0035690502263605595, "learning_rate": 1.227230253151615e-06, "loss": 0.0001, "num_input_tokens_seen": 69277760, "step": 35995 }, { "epoch": 5.872828126274574, "grad_norm": 0.014318207278847694, "learning_rate": 1.2241939194025748e-06, "loss": 0.0001, "num_input_tokens_seen": 69286320, "step": 36000 }, { "epoch": 5.872828126274574, "eval_loss": 0.4163505733013153, "eval_runtime": 80.909, "eval_samples_per_second": 33.68, "eval_steps_per_second": 16.846, "num_input_tokens_seen": 69286320, "step": 36000 }, { "epoch": 5.873643853495391, "grad_norm": 0.0008287746459245682, "learning_rate": 1.2211612521816156e-06, "loss": 0.0001, "num_input_tokens_seen": 69296240, "step": 36005 }, { "epoch": 5.874459580716208, "grad_norm": 0.0012523513287305832, "learning_rate": 1.2181322519564137e-06, "loss": 0.0, "num_input_tokens_seen": 69306032, "step": 36010 }, { "epoch": 5.8752753079370255, "grad_norm": 0.002570911543443799, "learning_rate": 1.2151069191940839e-06, "loss": 0.0, "num_input_tokens_seen": 69317312, "step": 36015 }, { "epoch": 5.8760910351578435, "grad_norm": 0.006763566751033068, "learning_rate": 1.2120852543611644e-06, "loss": 0.0001, "num_input_tokens_seen": 69327952, "step": 36020 }, { "epoch": 5.876906762378661, "grad_norm": 0.0007292142254300416, "learning_rate": 1.2090672579236379e-06, "loss": 0.0006, "num_input_tokens_seen": 69337728, "step": 36025 }, { "epoch": 5.877722489599478, "grad_norm": 0.020395882427692413, "learning_rate": 1.2060529303469126e-06, "loss": 0.0001, "num_input_tokens_seen": 69346736, "step": 36030 }, { "epoch": 5.878538216820296, "grad_norm": 0.002720719203352928, "learning_rate": 1.2030422720958445e-06, "loss": 0.0922, "num_input_tokens_seen": 69355456, "step": 36035 }, { "epoch": 5.879353944041113, "grad_norm": 0.04322320222854614, "learning_rate": 1.200035283634704e-06, "loss": 0.0001, "num_input_tokens_seen": 69366448, "step": 36040 }, { "epoch": 5.88016967126193, "grad_norm": 0.00019082985818386078, "learning_rate": 1.1970319654272144e-06, "loss": 0.0002, "num_input_tokens_seen": 69375488, "step": 36045 }, { "epoch": 5.880985398482747, "grad_norm": 0.006643225904554129, "learning_rate": 1.1940323179365192e-06, "loss": 0.0, "num_input_tokens_seen": 69384240, "step": 36050 }, { "epoch": 5.881801125703564, "grad_norm": 0.014796133153140545, "learning_rate": 1.1910363416252095e-06, "loss": 0.0222, "num_input_tokens_seen": 69393120, "step": 36055 }, { "epoch": 5.882616852924382, "grad_norm": 0.0002822483947966248, "learning_rate": 1.1880440369552964e-06, "loss": 0.045, "num_input_tokens_seen": 69402928, "step": 36060 }, { "epoch": 5.883432580145199, "grad_norm": 0.007039160467684269, "learning_rate": 1.1850554043882328e-06, "loss": 0.0001, "num_input_tokens_seen": 69412992, "step": 36065 }, { "epoch": 5.884248307366017, "grad_norm": 8.387018897337839e-05, "learning_rate": 1.1820704443849028e-06, "loss": 0.0017, "num_input_tokens_seen": 69423200, "step": 36070 }, { "epoch": 5.885064034586835, "grad_norm": 0.0034641576930880547, "learning_rate": 1.1790891574056219e-06, "loss": 0.0001, "num_input_tokens_seen": 69433168, "step": 36075 }, { "epoch": 5.885879761807652, "grad_norm": 0.34729644656181335, "learning_rate": 1.1761115439101523e-06, "loss": 0.0007, "num_input_tokens_seen": 69442976, "step": 36080 }, { "epoch": 5.886695489028469, "grad_norm": 0.0011519661638885736, "learning_rate": 1.1731376043576659e-06, "loss": 0.0, "num_input_tokens_seen": 69452960, "step": 36085 }, { "epoch": 5.887511216249286, "grad_norm": 0.0008063346613198519, "learning_rate": 1.1701673392067875e-06, "loss": 0.0003, "num_input_tokens_seen": 69462080, "step": 36090 }, { "epoch": 5.888326943470103, "grad_norm": 0.029690222814679146, "learning_rate": 1.1672007489155757e-06, "loss": 0.0001, "num_input_tokens_seen": 69471616, "step": 36095 }, { "epoch": 5.889142670690921, "grad_norm": 0.0011345328530296683, "learning_rate": 1.164237833941506e-06, "loss": 0.0, "num_input_tokens_seen": 69482048, "step": 36100 }, { "epoch": 5.889958397911738, "grad_norm": 0.00012335307837929577, "learning_rate": 1.1612785947415022e-06, "loss": 0.0002, "num_input_tokens_seen": 69492144, "step": 36105 }, { "epoch": 5.890774125132555, "grad_norm": 0.0030078624840825796, "learning_rate": 1.1583230317719185e-06, "loss": 0.0373, "num_input_tokens_seen": 69502256, "step": 36110 }, { "epoch": 5.891589852353373, "grad_norm": 0.0005217280704528093, "learning_rate": 1.1553711454885318e-06, "loss": 0.0, "num_input_tokens_seen": 69511472, "step": 36115 }, { "epoch": 5.8924055795741905, "grad_norm": 0.005548741668462753, "learning_rate": 1.152422936346567e-06, "loss": 0.0001, "num_input_tokens_seen": 69520768, "step": 36120 }, { "epoch": 5.893221306795008, "grad_norm": 19.757701873779297, "learning_rate": 1.1494784048006718e-06, "loss": 0.0251, "num_input_tokens_seen": 69530304, "step": 36125 }, { "epoch": 5.894037034015825, "grad_norm": 0.00945267267525196, "learning_rate": 1.1465375513049326e-06, "loss": 0.0, "num_input_tokens_seen": 69539872, "step": 36130 }, { "epoch": 5.894852761236643, "grad_norm": 0.001253066468052566, "learning_rate": 1.1436003763128616e-06, "loss": 0.0001, "num_input_tokens_seen": 69550224, "step": 36135 }, { "epoch": 5.89566848845746, "grad_norm": 0.00039400518289767206, "learning_rate": 1.1406668802774106e-06, "loss": 0.0001, "num_input_tokens_seen": 69559936, "step": 36140 }, { "epoch": 5.896484215678277, "grad_norm": 0.0005062380223535001, "learning_rate": 1.137737063650965e-06, "loss": 0.0, "num_input_tokens_seen": 69568848, "step": 36145 }, { "epoch": 5.897299942899094, "grad_norm": 0.022498145699501038, "learning_rate": 1.1348109268853323e-06, "loss": 0.0204, "num_input_tokens_seen": 69579824, "step": 36150 }, { "epoch": 5.898115670119912, "grad_norm": 0.0007638476672582328, "learning_rate": 1.1318884704317634e-06, "loss": 0.0, "num_input_tokens_seen": 69590336, "step": 36155 }, { "epoch": 5.898931397340729, "grad_norm": 0.0013432144187390804, "learning_rate": 1.1289696947409417e-06, "loss": 0.0001, "num_input_tokens_seen": 69601056, "step": 36160 }, { "epoch": 5.8997471245615465, "grad_norm": 0.004882112145423889, "learning_rate": 1.126054600262974e-06, "loss": 0.0001, "num_input_tokens_seen": 69609984, "step": 36165 }, { "epoch": 5.900562851782364, "grad_norm": 0.006230381783097982, "learning_rate": 1.1231431874474064e-06, "loss": 0.0, "num_input_tokens_seen": 69619248, "step": 36170 }, { "epoch": 5.901378579003182, "grad_norm": 0.00036118554999120533, "learning_rate": 1.12023545674321e-06, "loss": 0.0, "num_input_tokens_seen": 69628304, "step": 36175 }, { "epoch": 5.902194306223999, "grad_norm": 0.002203789073973894, "learning_rate": 1.117331408598804e-06, "loss": 0.0, "num_input_tokens_seen": 69637184, "step": 36180 }, { "epoch": 5.903010033444816, "grad_norm": 0.03624836727976799, "learning_rate": 1.1144310434620191e-06, "loss": 0.0, "num_input_tokens_seen": 69647664, "step": 36185 }, { "epoch": 5.903825760665633, "grad_norm": 0.013137330301105976, "learning_rate": 1.1115343617801365e-06, "loss": 0.0018, "num_input_tokens_seen": 69657728, "step": 36190 }, { "epoch": 5.904641487886451, "grad_norm": 0.0021411972120404243, "learning_rate": 1.1086413639998515e-06, "loss": 0.0037, "num_input_tokens_seen": 69667952, "step": 36195 }, { "epoch": 5.905457215107268, "grad_norm": 0.002752948785200715, "learning_rate": 1.1057520505673103e-06, "loss": 0.1023, "num_input_tokens_seen": 69676640, "step": 36200 }, { "epoch": 5.905457215107268, "eval_loss": 0.4185701906681061, "eval_runtime": 80.7856, "eval_samples_per_second": 33.731, "eval_steps_per_second": 16.872, "num_input_tokens_seen": 69676640, "step": 36200 }, { "epoch": 5.906272942328085, "grad_norm": 7.729828212177381e-05, "learning_rate": 1.1028664219280727e-06, "loss": 0.0001, "num_input_tokens_seen": 69686592, "step": 36205 }, { "epoch": 5.907088669548903, "grad_norm": 0.054397813975811005, "learning_rate": 1.0999844785271468e-06, "loss": 0.0001, "num_input_tokens_seen": 69695824, "step": 36210 }, { "epoch": 5.9079043967697205, "grad_norm": 26.673362731933594, "learning_rate": 1.097106220808955e-06, "loss": 0.0369, "num_input_tokens_seen": 69704240, "step": 36215 }, { "epoch": 5.908720123990538, "grad_norm": 0.027054961770772934, "learning_rate": 1.0942316492173698e-06, "loss": 0.0, "num_input_tokens_seen": 69713584, "step": 36220 }, { "epoch": 5.909535851211355, "grad_norm": 8.016550418687984e-05, "learning_rate": 1.0913607641956841e-06, "loss": 0.0034, "num_input_tokens_seen": 69723328, "step": 36225 }, { "epoch": 5.910351578432172, "grad_norm": 0.00032200542045757174, "learning_rate": 1.0884935661866213e-06, "loss": 0.0, "num_input_tokens_seen": 69732736, "step": 36230 }, { "epoch": 5.91116730565299, "grad_norm": 0.019485782831907272, "learning_rate": 1.0856300556323418e-06, "loss": 0.002, "num_input_tokens_seen": 69742528, "step": 36235 }, { "epoch": 5.911983032873807, "grad_norm": 0.008767404593527317, "learning_rate": 1.0827702329744365e-06, "loss": 0.001, "num_input_tokens_seen": 69751232, "step": 36240 }, { "epoch": 5.912798760094624, "grad_norm": 0.007466188166290522, "learning_rate": 1.0799140986539197e-06, "loss": 0.0001, "num_input_tokens_seen": 69761984, "step": 36245 }, { "epoch": 5.913614487315442, "grad_norm": 0.2192697674036026, "learning_rate": 1.0770616531112526e-06, "loss": 0.0003, "num_input_tokens_seen": 69771408, "step": 36250 }, { "epoch": 5.914430214536259, "grad_norm": 2.7514741304912604e-05, "learning_rate": 1.0742128967863085e-06, "loss": 0.0017, "num_input_tokens_seen": 69779920, "step": 36255 }, { "epoch": 5.915245941757076, "grad_norm": 0.001689986907877028, "learning_rate": 1.071367830118411e-06, "loss": 0.0, "num_input_tokens_seen": 69789648, "step": 36260 }, { "epoch": 5.9160616689778935, "grad_norm": 0.024721331894397736, "learning_rate": 1.068526453546298e-06, "loss": 0.0002, "num_input_tokens_seen": 69799936, "step": 36265 }, { "epoch": 5.916877396198711, "grad_norm": 0.004334015306085348, "learning_rate": 1.0656887675081467e-06, "loss": 0.0, "num_input_tokens_seen": 69810096, "step": 36270 }, { "epoch": 5.917693123419529, "grad_norm": 0.003990250173956156, "learning_rate": 1.0628547724415628e-06, "loss": 0.0, "num_input_tokens_seen": 69818256, "step": 36275 }, { "epoch": 5.918508850640346, "grad_norm": 17.622440338134766, "learning_rate": 1.0600244687835881e-06, "loss": 0.0961, "num_input_tokens_seen": 69827568, "step": 36280 }, { "epoch": 5.919324577861163, "grad_norm": 0.01530389953404665, "learning_rate": 1.0571978569706876e-06, "loss": 0.0001, "num_input_tokens_seen": 69836944, "step": 36285 }, { "epoch": 5.920140305081981, "grad_norm": 8.22303295135498, "learning_rate": 1.0543749374387652e-06, "loss": 0.0109, "num_input_tokens_seen": 69845680, "step": 36290 }, { "epoch": 5.920956032302798, "grad_norm": 0.005711407400667667, "learning_rate": 1.051555710623142e-06, "loss": 0.1085, "num_input_tokens_seen": 69855840, "step": 36295 }, { "epoch": 5.921771759523615, "grad_norm": 0.003789416281506419, "learning_rate": 1.0487401769585847e-06, "loss": 0.0001, "num_input_tokens_seen": 69866448, "step": 36300 }, { "epoch": 5.922587486744432, "grad_norm": 0.10992971062660217, "learning_rate": 1.0459283368792845e-06, "loss": 0.0003, "num_input_tokens_seen": 69875440, "step": 36305 }, { "epoch": 5.92340321396525, "grad_norm": 0.003301407443359494, "learning_rate": 1.043120190818858e-06, "loss": 0.0, "num_input_tokens_seen": 69885456, "step": 36310 }, { "epoch": 5.9242189411860675, "grad_norm": 0.030404970049858093, "learning_rate": 1.0403157392103596e-06, "loss": 0.0043, "num_input_tokens_seen": 69895600, "step": 36315 }, { "epoch": 5.925034668406885, "grad_norm": 0.002404683269560337, "learning_rate": 1.0375149824862735e-06, "loss": 0.019, "num_input_tokens_seen": 69905680, "step": 36320 }, { "epoch": 5.925850395627702, "grad_norm": 0.010823442600667477, "learning_rate": 1.034717921078507e-06, "loss": 0.0041, "num_input_tokens_seen": 69915840, "step": 36325 }, { "epoch": 5.92666612284852, "grad_norm": 0.005141077563166618, "learning_rate": 1.0319245554184009e-06, "loss": 0.0006, "num_input_tokens_seen": 69924384, "step": 36330 }, { "epoch": 5.927481850069337, "grad_norm": 0.00464071799069643, "learning_rate": 1.0291348859367361e-06, "loss": 0.0151, "num_input_tokens_seen": 69934272, "step": 36335 }, { "epoch": 5.928297577290154, "grad_norm": 0.8827366828918457, "learning_rate": 1.0263489130637016e-06, "loss": 0.0008, "num_input_tokens_seen": 69944336, "step": 36340 }, { "epoch": 5.929113304510971, "grad_norm": 9.406994819641113, "learning_rate": 1.0235666372289427e-06, "loss": 0.0052, "num_input_tokens_seen": 69953744, "step": 36345 }, { "epoch": 5.929929031731789, "grad_norm": 0.015848549082875252, "learning_rate": 1.0207880588615076e-06, "loss": 0.0001, "num_input_tokens_seen": 69962224, "step": 36350 }, { "epoch": 5.930744758952606, "grad_norm": 0.008051946759223938, "learning_rate": 1.0180131783898984e-06, "loss": 0.0, "num_input_tokens_seen": 69972400, "step": 36355 }, { "epoch": 5.9315604861734235, "grad_norm": 0.004730635322630405, "learning_rate": 1.0152419962420362e-06, "loss": 0.0, "num_input_tokens_seen": 69982448, "step": 36360 }, { "epoch": 5.932376213394241, "grad_norm": 0.0003212848387192935, "learning_rate": 1.0124745128452685e-06, "loss": 0.0, "num_input_tokens_seen": 69991968, "step": 36365 }, { "epoch": 5.933191940615059, "grad_norm": 0.0005786723340861499, "learning_rate": 1.0097107286263758e-06, "loss": 0.0, "num_input_tokens_seen": 70000880, "step": 36370 }, { "epoch": 5.934007667835876, "grad_norm": 0.004499983508139849, "learning_rate": 1.00695064401157e-06, "loss": 0.0002, "num_input_tokens_seen": 70009136, "step": 36375 }, { "epoch": 5.934823395056693, "grad_norm": 0.02237872965633869, "learning_rate": 1.0041942594264886e-06, "loss": 0.0002, "num_input_tokens_seen": 70018368, "step": 36380 }, { "epoch": 5.935639122277511, "grad_norm": 0.002402003388851881, "learning_rate": 1.001441575296208e-06, "loss": 0.0, "num_input_tokens_seen": 70027024, "step": 36385 }, { "epoch": 5.936454849498328, "grad_norm": 37.73929214477539, "learning_rate": 9.986925920452139e-07, "loss": 0.1755, "num_input_tokens_seen": 70037536, "step": 36390 }, { "epoch": 5.937270576719145, "grad_norm": 0.0002513858489692211, "learning_rate": 9.959473100974475e-07, "loss": 0.0001, "num_input_tokens_seen": 70047360, "step": 36395 }, { "epoch": 5.938086303939962, "grad_norm": 0.01707611232995987, "learning_rate": 9.932057298762564e-07, "loss": 0.0001, "num_input_tokens_seen": 70057024, "step": 36400 }, { "epoch": 5.938086303939962, "eval_loss": 0.42632344365119934, "eval_runtime": 80.9213, "eval_samples_per_second": 33.675, "eval_steps_per_second": 16.844, "num_input_tokens_seen": 70057024, "step": 36400 }, { "epoch": 5.938902031160779, "grad_norm": 0.05336268991231918, "learning_rate": 9.90467851804433e-07, "loss": 0.0, "num_input_tokens_seen": 70067024, "step": 36405 }, { "epoch": 5.939717758381597, "grad_norm": 0.001477487268857658, "learning_rate": 9.877336763041895e-07, "loss": 0.0816, "num_input_tokens_seen": 70075968, "step": 36410 }, { "epoch": 5.940533485602415, "grad_norm": 0.008978289552032948, "learning_rate": 9.850032037971662e-07, "loss": 0.0001, "num_input_tokens_seen": 70085328, "step": 36415 }, { "epoch": 5.941349212823232, "grad_norm": 7.29666805267334, "learning_rate": 9.822764347044406e-07, "loss": 0.048, "num_input_tokens_seen": 70094592, "step": 36420 }, { "epoch": 5.94216494004405, "grad_norm": 17.945383071899414, "learning_rate": 9.795533694465175e-07, "loss": 0.0805, "num_input_tokens_seen": 70104880, "step": 36425 }, { "epoch": 5.942980667264867, "grad_norm": 0.004881592467427254, "learning_rate": 9.768340084433197e-07, "loss": 0.0, "num_input_tokens_seen": 70111904, "step": 36430 }, { "epoch": 5.943796394485684, "grad_norm": 0.0006939506274648011, "learning_rate": 9.741183521142143e-07, "loss": 0.0017, "num_input_tokens_seen": 70120224, "step": 36435 }, { "epoch": 5.944612121706501, "grad_norm": 0.00026958595844917, "learning_rate": 9.714064008779889e-07, "loss": 0.0, "num_input_tokens_seen": 70129552, "step": 36440 }, { "epoch": 5.945427848927318, "grad_norm": 0.0005477212835103273, "learning_rate": 9.686981551528584e-07, "loss": 0.0, "num_input_tokens_seen": 70140112, "step": 36445 }, { "epoch": 5.946243576148136, "grad_norm": 0.010925560258328915, "learning_rate": 9.65993615356467e-07, "loss": 0.0001, "num_input_tokens_seen": 70150624, "step": 36450 }, { "epoch": 5.947059303368953, "grad_norm": 16.94795799255371, "learning_rate": 9.632927819058917e-07, "loss": 0.1274, "num_input_tokens_seen": 70159120, "step": 36455 }, { "epoch": 5.9478750305897705, "grad_norm": 0.0013974921312183142, "learning_rate": 9.605956552176305e-07, "loss": 0.0, "num_input_tokens_seen": 70168752, "step": 36460 }, { "epoch": 5.9486907578105885, "grad_norm": 0.006974565330892801, "learning_rate": 9.579022357076223e-07, "loss": 0.0, "num_input_tokens_seen": 70177824, "step": 36465 }, { "epoch": 5.949506485031406, "grad_norm": 0.017130674794316292, "learning_rate": 9.552125237912158e-07, "loss": 0.0001, "num_input_tokens_seen": 70186688, "step": 36470 }, { "epoch": 5.950322212252223, "grad_norm": 0.0007349959923885763, "learning_rate": 9.525265198832096e-07, "loss": 0.0, "num_input_tokens_seen": 70196384, "step": 36475 }, { "epoch": 5.95113793947304, "grad_norm": 0.010469692759215832, "learning_rate": 9.498442243978112e-07, "loss": 0.0001, "num_input_tokens_seen": 70205584, "step": 36480 }, { "epoch": 5.951953666693858, "grad_norm": 6.797251262469217e-05, "learning_rate": 9.471656377486649e-07, "loss": 0.0011, "num_input_tokens_seen": 70215600, "step": 36485 }, { "epoch": 5.952769393914675, "grad_norm": 0.015789875760674477, "learning_rate": 9.444907603488456e-07, "loss": 0.1459, "num_input_tokens_seen": 70225408, "step": 36490 }, { "epoch": 5.953585121135492, "grad_norm": 16.77529525756836, "learning_rate": 9.418195926108514e-07, "loss": 0.0954, "num_input_tokens_seen": 70234592, "step": 36495 }, { "epoch": 5.954400848356309, "grad_norm": 1.1809891475422774e-05, "learning_rate": 9.391521349466053e-07, "loss": 0.0006, "num_input_tokens_seen": 70243024, "step": 36500 }, { "epoch": 5.955216575577127, "grad_norm": 0.008567470125854015, "learning_rate": 9.364883877674758e-07, "loss": 0.0001, "num_input_tokens_seen": 70252672, "step": 36505 }, { "epoch": 5.9560323027979445, "grad_norm": 0.015186447650194168, "learning_rate": 9.33828351484231e-07, "loss": 0.0001, "num_input_tokens_seen": 70262160, "step": 36510 }, { "epoch": 5.956848030018762, "grad_norm": 0.01005709357559681, "learning_rate": 9.311720265070906e-07, "loss": 0.0002, "num_input_tokens_seen": 70270304, "step": 36515 }, { "epoch": 5.957663757239579, "grad_norm": 8.262692426797003e-05, "learning_rate": 9.285194132456931e-07, "loss": 0.0528, "num_input_tokens_seen": 70280592, "step": 36520 }, { "epoch": 5.958479484460397, "grad_norm": 0.006421425845474005, "learning_rate": 9.258705121091032e-07, "loss": 0.0814, "num_input_tokens_seen": 70288800, "step": 36525 }, { "epoch": 5.959295211681214, "grad_norm": 0.0011423748219385743, "learning_rate": 9.232253235058136e-07, "loss": 0.059, "num_input_tokens_seen": 70298608, "step": 36530 }, { "epoch": 5.960110938902031, "grad_norm": 0.009473681449890137, "learning_rate": 9.205838478437478e-07, "loss": 0.0, "num_input_tokens_seen": 70309456, "step": 36535 }, { "epoch": 5.960926666122848, "grad_norm": 0.17331618070602417, "learning_rate": 9.179460855302524e-07, "loss": 0.0529, "num_input_tokens_seen": 70319168, "step": 36540 }, { "epoch": 5.961742393343666, "grad_norm": 0.0003173246805090457, "learning_rate": 9.153120369721046e-07, "loss": 0.0, "num_input_tokens_seen": 70328832, "step": 36545 }, { "epoch": 5.962558120564483, "grad_norm": 0.05057176575064659, "learning_rate": 9.126817025755103e-07, "loss": 0.0002, "num_input_tokens_seen": 70337232, "step": 36550 }, { "epoch": 5.9633738477853, "grad_norm": 3.727425792021677e-05, "learning_rate": 9.100550827460947e-07, "loss": 0.0002, "num_input_tokens_seen": 70347504, "step": 36555 }, { "epoch": 5.964189575006118, "grad_norm": 0.09858830273151398, "learning_rate": 9.0743217788892e-07, "loss": 0.0178, "num_input_tokens_seen": 70355600, "step": 36560 }, { "epoch": 5.965005302226936, "grad_norm": 0.004284251015633345, "learning_rate": 9.048129884084683e-07, "loss": 0.0001, "num_input_tokens_seen": 70364768, "step": 36565 }, { "epoch": 5.965821029447753, "grad_norm": 1.0039558410644531, "learning_rate": 9.021975147086553e-07, "loss": 0.0022, "num_input_tokens_seen": 70374816, "step": 36570 }, { "epoch": 5.96663675666857, "grad_norm": 0.009357760660350323, "learning_rate": 8.995857571928141e-07, "loss": 0.0001, "num_input_tokens_seen": 70385120, "step": 36575 }, { "epoch": 5.967452483889387, "grad_norm": 0.001704140449874103, "learning_rate": 8.969777162637139e-07, "loss": 0.0002, "num_input_tokens_seen": 70394816, "step": 36580 }, { "epoch": 5.968268211110205, "grad_norm": 0.04503907263278961, "learning_rate": 8.943733923235525e-07, "loss": 0.0002, "num_input_tokens_seen": 70404000, "step": 36585 }, { "epoch": 5.969083938331022, "grad_norm": 0.00015239229833241552, "learning_rate": 8.917727857739394e-07, "loss": 0.0, "num_input_tokens_seen": 70413008, "step": 36590 }, { "epoch": 5.969899665551839, "grad_norm": 0.00370673555880785, "learning_rate": 8.891758970159258e-07, "loss": 0.0, "num_input_tokens_seen": 70422720, "step": 36595 }, { "epoch": 5.970715392772657, "grad_norm": 0.0012943295296281576, "learning_rate": 8.86582726449986e-07, "loss": 0.0, "num_input_tokens_seen": 70432848, "step": 36600 }, { "epoch": 5.970715392772657, "eval_loss": 0.4205092191696167, "eval_runtime": 80.7951, "eval_samples_per_second": 33.727, "eval_steps_per_second": 16.87, "num_input_tokens_seen": 70432848, "step": 36600 }, { "epoch": 5.971531119993474, "grad_norm": 0.00554748484864831, "learning_rate": 8.839932744760165e-07, "loss": 0.0005, "num_input_tokens_seen": 70442848, "step": 36605 }, { "epoch": 5.9723468472142915, "grad_norm": 0.033921435475349426, "learning_rate": 8.814075414933482e-07, "loss": 0.0, "num_input_tokens_seen": 70452112, "step": 36610 }, { "epoch": 5.973162574435109, "grad_norm": 0.0005889647291041911, "learning_rate": 8.788255279007257e-07, "loss": 0.0022, "num_input_tokens_seen": 70461680, "step": 36615 }, { "epoch": 5.973978301655926, "grad_norm": 0.009340082295238972, "learning_rate": 8.762472340963362e-07, "loss": 0.0002, "num_input_tokens_seen": 70471680, "step": 36620 }, { "epoch": 5.974794028876744, "grad_norm": 0.0019879706669598818, "learning_rate": 8.736726604777811e-07, "loss": 0.0024, "num_input_tokens_seen": 70480176, "step": 36625 }, { "epoch": 5.975609756097561, "grad_norm": 0.0054903835989534855, "learning_rate": 8.711018074420901e-07, "loss": 0.0001, "num_input_tokens_seen": 70490528, "step": 36630 }, { "epoch": 5.976425483318378, "grad_norm": 0.014695270918309689, "learning_rate": 8.685346753857209e-07, "loss": 0.0001, "num_input_tokens_seen": 70500896, "step": 36635 }, { "epoch": 5.977241210539196, "grad_norm": 0.007761961780488491, "learning_rate": 8.659712647045654e-07, "loss": 0.0002, "num_input_tokens_seen": 70510656, "step": 36640 }, { "epoch": 5.978056937760013, "grad_norm": 0.03636031970381737, "learning_rate": 8.634115757939209e-07, "loss": 0.0001, "num_input_tokens_seen": 70520176, "step": 36645 }, { "epoch": 5.97887266498083, "grad_norm": 0.23160231113433838, "learning_rate": 8.608556090485387e-07, "loss": 0.0001, "num_input_tokens_seen": 70529808, "step": 36650 }, { "epoch": 5.9796883922016475, "grad_norm": 0.00794603768736124, "learning_rate": 8.583033648625671e-07, "loss": 0.0001, "num_input_tokens_seen": 70540064, "step": 36655 }, { "epoch": 5.9805041194224655, "grad_norm": 0.00044677068945020437, "learning_rate": 8.557548436295998e-07, "loss": 0.0, "num_input_tokens_seen": 70550352, "step": 36660 }, { "epoch": 5.981319846643283, "grad_norm": 0.10896940529346466, "learning_rate": 8.532100457426556e-07, "loss": 0.0003, "num_input_tokens_seen": 70559856, "step": 36665 }, { "epoch": 5.9821355738641, "grad_norm": 0.001767410314641893, "learning_rate": 8.506689715941679e-07, "loss": 0.0, "num_input_tokens_seen": 70568048, "step": 36670 }, { "epoch": 5.982951301084917, "grad_norm": 0.005785114597529173, "learning_rate": 8.481316215760011e-07, "loss": 0.0149, "num_input_tokens_seen": 70578640, "step": 36675 }, { "epoch": 5.983767028305735, "grad_norm": 0.14954037964344025, "learning_rate": 8.455979960794558e-07, "loss": 0.0027, "num_input_tokens_seen": 70587232, "step": 36680 }, { "epoch": 5.984582755526552, "grad_norm": 0.02177238091826439, "learning_rate": 8.430680954952364e-07, "loss": 0.0, "num_input_tokens_seen": 70597088, "step": 36685 }, { "epoch": 5.985398482747369, "grad_norm": 0.026733502745628357, "learning_rate": 8.405419202134974e-07, "loss": 0.1191, "num_input_tokens_seen": 70606992, "step": 36690 }, { "epoch": 5.986214209968186, "grad_norm": 0.00021599663887172937, "learning_rate": 8.380194706237993e-07, "loss": 0.047, "num_input_tokens_seen": 70617440, "step": 36695 }, { "epoch": 5.987029937189004, "grad_norm": 0.00501007167622447, "learning_rate": 8.355007471151366e-07, "loss": 0.0001, "num_input_tokens_seen": 70627088, "step": 36700 }, { "epoch": 5.9878456644098215, "grad_norm": 0.003056610468775034, "learning_rate": 8.329857500759292e-07, "loss": 0.0001, "num_input_tokens_seen": 70637296, "step": 36705 }, { "epoch": 5.988661391630639, "grad_norm": 0.0007404976640827954, "learning_rate": 8.304744798940194e-07, "loss": 0.0007, "num_input_tokens_seen": 70647104, "step": 36710 }, { "epoch": 5.989477118851456, "grad_norm": 63.47469711303711, "learning_rate": 8.279669369566756e-07, "loss": 0.069, "num_input_tokens_seen": 70656080, "step": 36715 }, { "epoch": 5.990292846072274, "grad_norm": 0.0828835666179657, "learning_rate": 8.254631216505993e-07, "loss": 0.0001, "num_input_tokens_seen": 70664832, "step": 36720 }, { "epoch": 5.991108573293091, "grad_norm": 0.0010661112610250711, "learning_rate": 8.229630343619038e-07, "loss": 0.0, "num_input_tokens_seen": 70674400, "step": 36725 }, { "epoch": 5.991924300513908, "grad_norm": 0.0035801085177809, "learning_rate": 8.204666754761392e-07, "loss": 0.0, "num_input_tokens_seen": 70684848, "step": 36730 }, { "epoch": 5.992740027734725, "grad_norm": 0.009451458230614662, "learning_rate": 8.179740453782669e-07, "loss": 0.0001, "num_input_tokens_seen": 70694224, "step": 36735 }, { "epoch": 5.993555754955543, "grad_norm": 0.02582218125462532, "learning_rate": 8.154851444526907e-07, "loss": 0.0691, "num_input_tokens_seen": 70705168, "step": 36740 }, { "epoch": 5.99437148217636, "grad_norm": 0.0009088531951420009, "learning_rate": 8.129999730832283e-07, "loss": 0.0003, "num_input_tokens_seen": 70714576, "step": 36745 }, { "epoch": 5.995187209397177, "grad_norm": 0.08148504793643951, "learning_rate": 8.105185316531178e-07, "loss": 0.0393, "num_input_tokens_seen": 70722832, "step": 36750 }, { "epoch": 5.9960029366179945, "grad_norm": 0.24808913469314575, "learning_rate": 8.08040820545039e-07, "loss": 0.0003, "num_input_tokens_seen": 70732944, "step": 36755 }, { "epoch": 5.996818663838813, "grad_norm": 0.07057999819517136, "learning_rate": 8.055668401410782e-07, "loss": 0.0671, "num_input_tokens_seen": 70742704, "step": 36760 }, { "epoch": 5.99763439105963, "grad_norm": 0.0771535262465477, "learning_rate": 8.030965908227578e-07, "loss": 0.0001, "num_input_tokens_seen": 70752240, "step": 36765 }, { "epoch": 5.998450118280447, "grad_norm": 0.13539473712444305, "learning_rate": 8.006300729710203e-07, "loss": 0.0002, "num_input_tokens_seen": 70761120, "step": 36770 }, { "epoch": 5.999265845501265, "grad_norm": 0.00044025806710124016, "learning_rate": 7.981672869662337e-07, "loss": 0.0001, "num_input_tokens_seen": 70769664, "step": 36775 }, { "epoch": 6.0, "grad_norm": 0.006823807954788208, "learning_rate": 7.957082331881888e-07, "loss": 0.0001, "num_input_tokens_seen": 70778304, "step": 36780 }, { "epoch": 6.000815727220817, "grad_norm": 0.0017661770107224584, "learning_rate": 7.932529120161069e-07, "loss": 0.0, "num_input_tokens_seen": 70788096, "step": 36785 }, { "epoch": 6.001631454441635, "grad_norm": 0.001681513269431889, "learning_rate": 7.908013238286243e-07, "loss": 0.0, "num_input_tokens_seen": 70798704, "step": 36790 }, { "epoch": 6.002447181662452, "grad_norm": 0.0022882609628140926, "learning_rate": 7.883534690038136e-07, "loss": 0.0002, "num_input_tokens_seen": 70809536, "step": 36795 }, { "epoch": 6.003262908883269, "grad_norm": 0.007127041928470135, "learning_rate": 7.859093479191559e-07, "loss": 0.0001, "num_input_tokens_seen": 70819440, "step": 36800 }, { "epoch": 6.003262908883269, "eval_loss": 0.4214898943901062, "eval_runtime": 80.9613, "eval_samples_per_second": 33.658, "eval_steps_per_second": 16.835, "num_input_tokens_seen": 70819440, "step": 36800 }, { "epoch": 6.0040786361040865, "grad_norm": 0.5960737466812134, "learning_rate": 7.834689609515722e-07, "loss": 0.0002, "num_input_tokens_seen": 70828736, "step": 36805 }, { "epoch": 6.004894363324905, "grad_norm": 0.007006845436990261, "learning_rate": 7.810323084774002e-07, "loss": 0.0006, "num_input_tokens_seen": 70838128, "step": 36810 }, { "epoch": 6.005710090545722, "grad_norm": 0.00028880289755761623, "learning_rate": 7.785993908723976e-07, "loss": 0.0001, "num_input_tokens_seen": 70847104, "step": 36815 }, { "epoch": 6.006525817766539, "grad_norm": 0.01138211227953434, "learning_rate": 7.761702085117534e-07, "loss": 0.0002, "num_input_tokens_seen": 70855520, "step": 36820 }, { "epoch": 6.007341544987356, "grad_norm": 0.0016879892209544778, "learning_rate": 7.737447617700844e-07, "loss": 0.0, "num_input_tokens_seen": 70865984, "step": 36825 }, { "epoch": 6.008157272208174, "grad_norm": 0.003876954782754183, "learning_rate": 7.713230510214136e-07, "loss": 0.0001, "num_input_tokens_seen": 70874800, "step": 36830 }, { "epoch": 6.008972999428991, "grad_norm": 0.00526140071451664, "learning_rate": 7.689050766392092e-07, "loss": 0.1205, "num_input_tokens_seen": 70882912, "step": 36835 }, { "epoch": 6.009788726649808, "grad_norm": 0.0035982097033411264, "learning_rate": 7.664908389963477e-07, "loss": 0.0, "num_input_tokens_seen": 70892960, "step": 36840 }, { "epoch": 6.010604453870625, "grad_norm": 0.008816607296466827, "learning_rate": 7.64080338465134e-07, "loss": 0.0004, "num_input_tokens_seen": 70901744, "step": 36845 }, { "epoch": 6.011420181091443, "grad_norm": 0.0062200878746807575, "learning_rate": 7.616735754173043e-07, "loss": 0.0002, "num_input_tokens_seen": 70911872, "step": 36850 }, { "epoch": 6.0122359083122605, "grad_norm": 0.0010891271522268653, "learning_rate": 7.592705502240005e-07, "loss": 0.0001, "num_input_tokens_seen": 70921696, "step": 36855 }, { "epoch": 6.013051635533078, "grad_norm": 0.04124739021062851, "learning_rate": 7.568712632558095e-07, "loss": 0.0354, "num_input_tokens_seen": 70932000, "step": 36860 }, { "epoch": 6.013867362753895, "grad_norm": 0.005176685284823179, "learning_rate": 7.544757148827297e-07, "loss": 0.0284, "num_input_tokens_seen": 70942496, "step": 36865 }, { "epoch": 6.014683089974713, "grad_norm": 0.008814333006739616, "learning_rate": 7.520839054741797e-07, "loss": 0.0, "num_input_tokens_seen": 70953536, "step": 36870 }, { "epoch": 6.01549881719553, "grad_norm": 0.01046714186668396, "learning_rate": 7.496958353990113e-07, "loss": 0.0, "num_input_tokens_seen": 70963952, "step": 36875 }, { "epoch": 6.016314544416347, "grad_norm": 0.22599689662456512, "learning_rate": 7.473115050254941e-07, "loss": 0.0002, "num_input_tokens_seen": 70974400, "step": 36880 }, { "epoch": 6.017130271637164, "grad_norm": 0.0008166134357452393, "learning_rate": 7.449309147213173e-07, "loss": 0.0001, "num_input_tokens_seen": 70985296, "step": 36885 }, { "epoch": 6.017945998857982, "grad_norm": 0.018062371760606766, "learning_rate": 7.425540648536067e-07, "loss": 0.0001, "num_input_tokens_seen": 70993104, "step": 36890 }, { "epoch": 6.018761726078799, "grad_norm": 0.0006012567318975925, "learning_rate": 7.40180955788894e-07, "loss": 0.0, "num_input_tokens_seen": 71002528, "step": 36895 }, { "epoch": 6.0195774532996165, "grad_norm": 0.0002186358906328678, "learning_rate": 7.378115878931474e-07, "loss": 0.0001, "num_input_tokens_seen": 71011712, "step": 36900 }, { "epoch": 6.020393180520434, "grad_norm": 9.738749940879643e-05, "learning_rate": 7.354459615317527e-07, "loss": 0.0002, "num_input_tokens_seen": 71020528, "step": 36905 }, { "epoch": 6.021208907741252, "grad_norm": 0.0002457669179420918, "learning_rate": 7.33084077069518e-07, "loss": 0.094, "num_input_tokens_seen": 71029680, "step": 36910 }, { "epoch": 6.022024634962069, "grad_norm": 0.0005698423483408988, "learning_rate": 7.307259348706768e-07, "loss": 0.0, "num_input_tokens_seen": 71037536, "step": 36915 }, { "epoch": 6.022840362182886, "grad_norm": 0.002722004661336541, "learning_rate": 7.283715352988801e-07, "loss": 0.0, "num_input_tokens_seen": 71046480, "step": 36920 }, { "epoch": 6.023656089403703, "grad_norm": 34.087528228759766, "learning_rate": 7.260208787172068e-07, "loss": 0.0025, "num_input_tokens_seen": 71056592, "step": 36925 }, { "epoch": 6.024471816624521, "grad_norm": 0.002382394392043352, "learning_rate": 7.23673965488167e-07, "loss": 0.0001, "num_input_tokens_seen": 71065952, "step": 36930 }, { "epoch": 6.025287543845338, "grad_norm": 0.0001588907471159473, "learning_rate": 7.213307959736709e-07, "loss": 0.0001, "num_input_tokens_seen": 71076672, "step": 36935 }, { "epoch": 6.026103271066155, "grad_norm": 0.004486395511776209, "learning_rate": 7.189913705350715e-07, "loss": 0.0, "num_input_tokens_seen": 71085760, "step": 36940 }, { "epoch": 6.026918998286972, "grad_norm": 0.013770767487585545, "learning_rate": 7.166556895331411e-07, "loss": 0.0, "num_input_tokens_seen": 71095920, "step": 36945 }, { "epoch": 6.02773472550779, "grad_norm": 0.0029596167150884867, "learning_rate": 7.143237533280639e-07, "loss": 0.0004, "num_input_tokens_seen": 71105264, "step": 36950 }, { "epoch": 6.028550452728608, "grad_norm": 0.0005928972386755049, "learning_rate": 7.119955622794578e-07, "loss": 0.0001, "num_input_tokens_seen": 71114736, "step": 36955 }, { "epoch": 6.029366179949425, "grad_norm": 0.10230780392885208, "learning_rate": 7.096711167463577e-07, "loss": 0.0163, "num_input_tokens_seen": 71124736, "step": 36960 }, { "epoch": 6.030181907170242, "grad_norm": 0.005246656481176615, "learning_rate": 7.073504170872213e-07, "loss": 0.0006, "num_input_tokens_seen": 71135136, "step": 36965 }, { "epoch": 6.03099763439106, "grad_norm": 0.000749376427847892, "learning_rate": 7.05033463659932e-07, "loss": 0.0001, "num_input_tokens_seen": 71143680, "step": 36970 }, { "epoch": 6.031813361611877, "grad_norm": 27.575695037841797, "learning_rate": 7.027202568217928e-07, "loss": 0.0026, "num_input_tokens_seen": 71154176, "step": 36975 }, { "epoch": 6.032629088832694, "grad_norm": 0.012653502635657787, "learning_rate": 7.004107969295293e-07, "loss": 0.0002, "num_input_tokens_seen": 71163104, "step": 36980 }, { "epoch": 6.033444816053512, "grad_norm": 0.04394585266709328, "learning_rate": 6.9810508433929e-07, "loss": 0.0001, "num_input_tokens_seen": 71173616, "step": 36985 }, { "epoch": 6.034260543274329, "grad_norm": 0.0005478697130456567, "learning_rate": 6.958031194066406e-07, "loss": 0.0001, "num_input_tokens_seen": 71182832, "step": 36990 }, { "epoch": 6.035076270495146, "grad_norm": 0.014896316453814507, "learning_rate": 6.935049024865776e-07, "loss": 0.0097, "num_input_tokens_seen": 71193280, "step": 36995 }, { "epoch": 6.0358919977159635, "grad_norm": 0.0794442668557167, "learning_rate": 6.912104339335118e-07, "loss": 0.0001, "num_input_tokens_seen": 71203008, "step": 37000 }, { "epoch": 6.0358919977159635, "eval_loss": 0.42486536502838135, "eval_runtime": 80.7679, "eval_samples_per_second": 33.739, "eval_steps_per_second": 16.876, "num_input_tokens_seen": 71203008, "step": 37000 }, { "epoch": 6.0367077249367815, "grad_norm": 0.02585206739604473, "learning_rate": 6.889197141012799e-07, "loss": 0.0001, "num_input_tokens_seen": 71210928, "step": 37005 }, { "epoch": 6.037523452157599, "grad_norm": 0.002483068499714136, "learning_rate": 6.866327433431435e-07, "loss": 0.0001, "num_input_tokens_seen": 71220064, "step": 37010 }, { "epoch": 6.038339179378416, "grad_norm": 0.011107205413281918, "learning_rate": 6.843495220117735e-07, "loss": 0.0001, "num_input_tokens_seen": 71230384, "step": 37015 }, { "epoch": 6.039154906599233, "grad_norm": 0.00010904530790867284, "learning_rate": 6.820700504592798e-07, "loss": 0.0, "num_input_tokens_seen": 71240672, "step": 37020 }, { "epoch": 6.039970633820051, "grad_norm": 0.000826433242764324, "learning_rate": 6.797943290371839e-07, "loss": 0.0449, "num_input_tokens_seen": 71250528, "step": 37025 }, { "epoch": 6.040786361040868, "grad_norm": 0.005403764545917511, "learning_rate": 6.775223580964274e-07, "loss": 0.0049, "num_input_tokens_seen": 71262000, "step": 37030 }, { "epoch": 6.041602088261685, "grad_norm": 0.008191421627998352, "learning_rate": 6.7525413798738e-07, "loss": 0.0001, "num_input_tokens_seen": 71272176, "step": 37035 }, { "epoch": 6.042417815482502, "grad_norm": 0.0013018219033256173, "learning_rate": 6.729896690598259e-07, "loss": 0.0, "num_input_tokens_seen": 71280784, "step": 37040 }, { "epoch": 6.04323354270332, "grad_norm": 0.020403224974870682, "learning_rate": 6.707289516629772e-07, "loss": 0.0, "num_input_tokens_seen": 71290576, "step": 37045 }, { "epoch": 6.0440492699241375, "grad_norm": 0.0004913843586109579, "learning_rate": 6.684719861454692e-07, "loss": 0.0096, "num_input_tokens_seen": 71301216, "step": 37050 }, { "epoch": 6.044864997144955, "grad_norm": 1.8243537851958536e-05, "learning_rate": 6.662187728553481e-07, "loss": 0.0001, "num_input_tokens_seen": 71310096, "step": 37055 }, { "epoch": 6.045680724365772, "grad_norm": 0.0006171854911372066, "learning_rate": 6.639693121400892e-07, "loss": 0.0, "num_input_tokens_seen": 71321200, "step": 37060 }, { "epoch": 6.04649645158659, "grad_norm": 0.05402669683098793, "learning_rate": 6.617236043465868e-07, "loss": 0.0002, "num_input_tokens_seen": 71331552, "step": 37065 }, { "epoch": 6.047312178807407, "grad_norm": 0.41084057092666626, "learning_rate": 6.594816498211587e-07, "loss": 0.0007, "num_input_tokens_seen": 71340672, "step": 37070 }, { "epoch": 6.048127906028224, "grad_norm": 0.030768757686018944, "learning_rate": 6.572434489095447e-07, "loss": 0.0, "num_input_tokens_seen": 71350688, "step": 37075 }, { "epoch": 6.048943633249041, "grad_norm": 0.009608573280274868, "learning_rate": 6.550090019568994e-07, "loss": 0.0004, "num_input_tokens_seen": 71360800, "step": 37080 }, { "epoch": 6.049759360469859, "grad_norm": 0.041162628680467606, "learning_rate": 6.527783093078027e-07, "loss": 0.0001, "num_input_tokens_seen": 71371344, "step": 37085 }, { "epoch": 6.050575087690676, "grad_norm": 0.2687203288078308, "learning_rate": 6.5055137130626e-07, "loss": 0.0551, "num_input_tokens_seen": 71380912, "step": 37090 }, { "epoch": 6.051390814911493, "grad_norm": 0.029169725254178047, "learning_rate": 6.483281882956854e-07, "loss": 0.0, "num_input_tokens_seen": 71391504, "step": 37095 }, { "epoch": 6.052206542132311, "grad_norm": 0.003794929012656212, "learning_rate": 6.461087606189298e-07, "loss": 0.0001, "num_input_tokens_seen": 71400112, "step": 37100 }, { "epoch": 6.053022269353129, "grad_norm": 0.0007314307731576264, "learning_rate": 6.438930886182554e-07, "loss": 0.0589, "num_input_tokens_seen": 71408832, "step": 37105 }, { "epoch": 6.053837996573946, "grad_norm": 0.06002825126051903, "learning_rate": 6.416811726353417e-07, "loss": 0.0002, "num_input_tokens_seen": 71418096, "step": 37110 }, { "epoch": 6.054653723794763, "grad_norm": 0.0016451451228931546, "learning_rate": 6.394730130112991e-07, "loss": 0.0001, "num_input_tokens_seen": 71427056, "step": 37115 }, { "epoch": 6.05546945101558, "grad_norm": 0.1709650307893753, "learning_rate": 6.372686100866471e-07, "loss": 0.0002, "num_input_tokens_seen": 71436272, "step": 37120 }, { "epoch": 6.056285178236398, "grad_norm": 0.0018742403481155634, "learning_rate": 6.350679642013413e-07, "loss": 0.0, "num_input_tokens_seen": 71445808, "step": 37125 }, { "epoch": 6.057100905457215, "grad_norm": 0.028838029131293297, "learning_rate": 6.328710756947437e-07, "loss": 0.0001, "num_input_tokens_seen": 71455408, "step": 37130 }, { "epoch": 6.057916632678032, "grad_norm": 0.01598254404962063, "learning_rate": 6.306779449056416e-07, "loss": 0.0001, "num_input_tokens_seen": 71464960, "step": 37135 }, { "epoch": 6.058732359898849, "grad_norm": 0.00038534513441845775, "learning_rate": 6.284885721722422e-07, "loss": 0.0, "num_input_tokens_seen": 71475696, "step": 37140 }, { "epoch": 6.059548087119667, "grad_norm": 0.0007755607948638499, "learning_rate": 6.26302957832181e-07, "loss": 0.0, "num_input_tokens_seen": 71486368, "step": 37145 }, { "epoch": 6.0603638143404845, "grad_norm": 0.002187174977734685, "learning_rate": 6.241211022224997e-07, "loss": 0.0002, "num_input_tokens_seen": 71496368, "step": 37150 }, { "epoch": 6.061179541561302, "grad_norm": 0.018522346392273903, "learning_rate": 6.219430056796732e-07, "loss": 0.0, "num_input_tokens_seen": 71505712, "step": 37155 }, { "epoch": 6.06199526878212, "grad_norm": 0.001072386628948152, "learning_rate": 6.19768668539586e-07, "loss": 0.0001, "num_input_tokens_seen": 71515968, "step": 37160 }, { "epoch": 6.062810996002937, "grad_norm": 0.05065721273422241, "learning_rate": 6.175980911375528e-07, "loss": 0.0918, "num_input_tokens_seen": 71525824, "step": 37165 }, { "epoch": 6.063626723223754, "grad_norm": 0.1062140241265297, "learning_rate": 6.154312738083034e-07, "loss": 0.0002, "num_input_tokens_seen": 71535712, "step": 37170 }, { "epoch": 6.064442450444571, "grad_norm": 0.0007074464811012149, "learning_rate": 6.132682168859843e-07, "loss": 0.0002, "num_input_tokens_seen": 71544224, "step": 37175 }, { "epoch": 6.065258177665389, "grad_norm": 37.38218307495117, "learning_rate": 6.111089207041704e-07, "loss": 0.123, "num_input_tokens_seen": 71552992, "step": 37180 }, { "epoch": 6.066073904886206, "grad_norm": 0.010236413218080997, "learning_rate": 6.089533855958507e-07, "loss": 0.0001, "num_input_tokens_seen": 71562176, "step": 37185 }, { "epoch": 6.066889632107023, "grad_norm": 0.0003965416399296373, "learning_rate": 6.068016118934372e-07, "loss": 0.0, "num_input_tokens_seen": 71570256, "step": 37190 }, { "epoch": 6.0677053593278405, "grad_norm": 0.0005695985746569932, "learning_rate": 6.04653599928759e-07, "loss": 0.0, "num_input_tokens_seen": 71579392, "step": 37195 }, { "epoch": 6.0685210865486585, "grad_norm": 0.000742063857614994, "learning_rate": 6.025093500330675e-07, "loss": 0.1001, "num_input_tokens_seen": 71588672, "step": 37200 }, { "epoch": 6.0685210865486585, "eval_loss": 0.42549702525138855, "eval_runtime": 80.7239, "eval_samples_per_second": 33.757, "eval_steps_per_second": 16.885, "num_input_tokens_seen": 71588672, "step": 37200 }, { "epoch": 6.069336813769476, "grad_norm": 0.005558207165449858, "learning_rate": 6.003688625370291e-07, "loss": 0.0, "num_input_tokens_seen": 71597856, "step": 37205 }, { "epoch": 6.070152540990293, "grad_norm": 0.1775227040052414, "learning_rate": 5.982321377707406e-07, "loss": 0.0003, "num_input_tokens_seen": 71607360, "step": 37210 }, { "epoch": 6.07096826821111, "grad_norm": 0.00492906803265214, "learning_rate": 5.96099176063708e-07, "loss": 0.0001, "num_input_tokens_seen": 71616368, "step": 37215 }, { "epoch": 6.071783995431928, "grad_norm": 0.012307398021221161, "learning_rate": 5.93969977744857e-07, "loss": 0.0001, "num_input_tokens_seen": 71626864, "step": 37220 }, { "epoch": 6.072599722652745, "grad_norm": 0.004265491850674152, "learning_rate": 5.918445431425445e-07, "loss": 0.0001, "num_input_tokens_seen": 71635472, "step": 37225 }, { "epoch": 6.073415449873562, "grad_norm": 0.0056029632687568665, "learning_rate": 5.897228725845333e-07, "loss": 0.0003, "num_input_tokens_seen": 71645168, "step": 37230 }, { "epoch": 6.074231177094379, "grad_norm": 0.0509447306394577, "learning_rate": 5.876049663980171e-07, "loss": 0.0001, "num_input_tokens_seen": 71653776, "step": 37235 }, { "epoch": 6.075046904315197, "grad_norm": 0.0010131974704563618, "learning_rate": 5.854908249095959e-07, "loss": 0.0289, "num_input_tokens_seen": 71662368, "step": 37240 }, { "epoch": 6.0758626315360145, "grad_norm": 0.3204793632030487, "learning_rate": 5.833804484453031e-07, "loss": 0.0004, "num_input_tokens_seen": 71671520, "step": 37245 }, { "epoch": 6.076678358756832, "grad_norm": 0.0007030792185105383, "learning_rate": 5.81273837330587e-07, "loss": 0.0002, "num_input_tokens_seen": 71681344, "step": 37250 }, { "epoch": 6.077494085977649, "grad_norm": 0.003919497597962618, "learning_rate": 5.791709918903071e-07, "loss": 0.0002, "num_input_tokens_seen": 71690624, "step": 37255 }, { "epoch": 6.078309813198467, "grad_norm": 0.21143034100532532, "learning_rate": 5.770719124487483e-07, "loss": 0.0001, "num_input_tokens_seen": 71700592, "step": 37260 }, { "epoch": 6.079125540419284, "grad_norm": 0.004179174080491066, "learning_rate": 5.749765993296241e-07, "loss": 0.0002, "num_input_tokens_seen": 71710112, "step": 37265 }, { "epoch": 6.079941267640101, "grad_norm": 0.1817561239004135, "learning_rate": 5.728850528560509e-07, "loss": 0.0003, "num_input_tokens_seen": 71720464, "step": 37270 }, { "epoch": 6.080756994860918, "grad_norm": 0.0012729656882584095, "learning_rate": 5.707972733505707e-07, "loss": 0.0, "num_input_tokens_seen": 71730240, "step": 37275 }, { "epoch": 6.081572722081736, "grad_norm": 0.014026719145476818, "learning_rate": 5.687132611351509e-07, "loss": 0.0001, "num_input_tokens_seen": 71739888, "step": 37280 }, { "epoch": 6.082388449302553, "grad_norm": 0.005108939949423075, "learning_rate": 5.666330165311651e-07, "loss": 0.0001, "num_input_tokens_seen": 71749552, "step": 37285 }, { "epoch": 6.08320417652337, "grad_norm": 0.019166449084877968, "learning_rate": 5.645565398594204e-07, "loss": 0.0, "num_input_tokens_seen": 71759424, "step": 37290 }, { "epoch": 6.0840199037441876, "grad_norm": 0.03266797587275505, "learning_rate": 5.624838314401304e-07, "loss": 0.0003, "num_input_tokens_seen": 71768080, "step": 37295 }, { "epoch": 6.084835630965006, "grad_norm": 0.013285050168633461, "learning_rate": 5.604148915929336e-07, "loss": 0.0002, "num_input_tokens_seen": 71778208, "step": 37300 }, { "epoch": 6.085651358185823, "grad_norm": 0.019582552835345268, "learning_rate": 5.583497206368887e-07, "loss": 0.0002, "num_input_tokens_seen": 71787552, "step": 37305 }, { "epoch": 6.08646708540664, "grad_norm": 0.0010262492578476667, "learning_rate": 5.562883188904688e-07, "loss": 0.0223, "num_input_tokens_seen": 71797888, "step": 37310 }, { "epoch": 6.087282812627457, "grad_norm": 0.002160652307793498, "learning_rate": 5.542306866715724e-07, "loss": 0.0, "num_input_tokens_seen": 71807616, "step": 37315 }, { "epoch": 6.088098539848275, "grad_norm": 0.8067982792854309, "learning_rate": 5.52176824297504e-07, "loss": 0.0014, "num_input_tokens_seen": 71817984, "step": 37320 }, { "epoch": 6.088914267069092, "grad_norm": 0.00955216120928526, "learning_rate": 5.501267320850018e-07, "loss": 0.0002, "num_input_tokens_seen": 71826768, "step": 37325 }, { "epoch": 6.089729994289909, "grad_norm": 0.027539145201444626, "learning_rate": 5.480804103502157e-07, "loss": 0.0001, "num_input_tokens_seen": 71837328, "step": 37330 }, { "epoch": 6.090545721510727, "grad_norm": 0.1999334990978241, "learning_rate": 5.460378594087101e-07, "loss": 0.0004, "num_input_tokens_seen": 71846528, "step": 37335 }, { "epoch": 6.091361448731544, "grad_norm": 0.019167210906744003, "learning_rate": 5.439990795754773e-07, "loss": 0.0, "num_input_tokens_seen": 71855280, "step": 37340 }, { "epoch": 6.0921771759523615, "grad_norm": 0.0012152515118941665, "learning_rate": 5.419640711649188e-07, "loss": 0.0, "num_input_tokens_seen": 71864656, "step": 37345 }, { "epoch": 6.092992903173179, "grad_norm": 9.29888992686756e-05, "learning_rate": 5.399328344908583e-07, "loss": 0.0, "num_input_tokens_seen": 71874400, "step": 37350 }, { "epoch": 6.093808630393997, "grad_norm": 0.0008154423558153212, "learning_rate": 5.379053698665399e-07, "loss": 0.0, "num_input_tokens_seen": 71883984, "step": 37355 }, { "epoch": 6.094624357614814, "grad_norm": 0.00031481878249906003, "learning_rate": 5.358816776046216e-07, "loss": 0.0004, "num_input_tokens_seen": 71893856, "step": 37360 }, { "epoch": 6.095440084835631, "grad_norm": 0.13200505077838898, "learning_rate": 5.338617580171817e-07, "loss": 0.1048, "num_input_tokens_seen": 71904368, "step": 37365 }, { "epoch": 6.096255812056448, "grad_norm": 0.02744072489440441, "learning_rate": 5.318456114157239e-07, "loss": 0.0003, "num_input_tokens_seen": 71913408, "step": 37370 }, { "epoch": 6.097071539277266, "grad_norm": 20.14779281616211, "learning_rate": 5.298332381111576e-07, "loss": 0.0878, "num_input_tokens_seen": 71923296, "step": 37375 }, { "epoch": 6.097887266498083, "grad_norm": 0.017371058464050293, "learning_rate": 5.27824638413818e-07, "loss": 0.0, "num_input_tokens_seen": 71934032, "step": 37380 }, { "epoch": 6.0987029937189, "grad_norm": 0.029726896435022354, "learning_rate": 5.258198126334546e-07, "loss": 0.0691, "num_input_tokens_seen": 71943488, "step": 37385 }, { "epoch": 6.0995187209397175, "grad_norm": 0.02405744418501854, "learning_rate": 5.238187610792367e-07, "loss": 0.0005, "num_input_tokens_seen": 71953616, "step": 37390 }, { "epoch": 6.1003344481605355, "grad_norm": 0.03689601644873619, "learning_rate": 5.218214840597563e-07, "loss": 0.0001, "num_input_tokens_seen": 71962928, "step": 37395 }, { "epoch": 6.101150175381353, "grad_norm": 0.00022739746782463044, "learning_rate": 5.198279818830115e-07, "loss": 0.0, "num_input_tokens_seen": 71972608, "step": 37400 }, { "epoch": 6.101150175381353, "eval_loss": 0.42471539974212646, "eval_runtime": 80.8559, "eval_samples_per_second": 33.702, "eval_steps_per_second": 16.857, "num_input_tokens_seen": 71972608, "step": 37400 }, { "epoch": 6.10196590260217, "grad_norm": 0.0015579232713207603, "learning_rate": 5.178382548564287e-07, "loss": 0.0002, "num_input_tokens_seen": 71981168, "step": 37405 }, { "epoch": 6.102781629822987, "grad_norm": 0.002350447466596961, "learning_rate": 5.15852303286854e-07, "loss": 0.0001, "num_input_tokens_seen": 71991280, "step": 37410 }, { "epoch": 6.103597357043805, "grad_norm": 0.019029982388019562, "learning_rate": 5.138701274805396e-07, "loss": 0.0001, "num_input_tokens_seen": 72001664, "step": 37415 }, { "epoch": 6.104413084264622, "grad_norm": 0.561782717704773, "learning_rate": 5.118917277431606e-07, "loss": 0.001, "num_input_tokens_seen": 72010560, "step": 37420 }, { "epoch": 6.105228811485439, "grad_norm": 4.940182770951651e-05, "learning_rate": 5.099171043798145e-07, "loss": 0.0001, "num_input_tokens_seen": 72019392, "step": 37425 }, { "epoch": 6.106044538706256, "grad_norm": 0.005670933518558741, "learning_rate": 5.079462576950133e-07, "loss": 0.0, "num_input_tokens_seen": 72028224, "step": 37430 }, { "epoch": 6.106860265927074, "grad_norm": 0.05800432711839676, "learning_rate": 5.059791879926862e-07, "loss": 0.0001, "num_input_tokens_seen": 72037968, "step": 37435 }, { "epoch": 6.1076759931478914, "grad_norm": 0.0002777784538920969, "learning_rate": 5.040158955761793e-07, "loss": 0.0002, "num_input_tokens_seen": 72047888, "step": 37440 }, { "epoch": 6.108491720368709, "grad_norm": 0.024548660963773727, "learning_rate": 5.020563807482559e-07, "loss": 0.0509, "num_input_tokens_seen": 72056208, "step": 37445 }, { "epoch": 6.109307447589526, "grad_norm": 0.23261651396751404, "learning_rate": 5.001006438110995e-07, "loss": 0.0003, "num_input_tokens_seen": 72064944, "step": 37450 }, { "epoch": 6.110123174810344, "grad_norm": 0.0016602360410615802, "learning_rate": 4.981486850663075e-07, "loss": 0.0001, "num_input_tokens_seen": 72074448, "step": 37455 }, { "epoch": 6.110938902031161, "grad_norm": 2.851459264755249, "learning_rate": 4.962005048149005e-07, "loss": 0.0028, "num_input_tokens_seen": 72084352, "step": 37460 }, { "epoch": 6.111754629251978, "grad_norm": 0.0058348774909973145, "learning_rate": 4.942561033573073e-07, "loss": 0.0002, "num_input_tokens_seen": 72093776, "step": 37465 }, { "epoch": 6.112570356472795, "grad_norm": 8.540109634399414, "learning_rate": 4.923154809933827e-07, "loss": 0.0629, "num_input_tokens_seen": 72102208, "step": 37470 }, { "epoch": 6.113386083693613, "grad_norm": 0.08757565170526505, "learning_rate": 4.903786380223957e-07, "loss": 0.0002, "num_input_tokens_seen": 72111488, "step": 37475 }, { "epoch": 6.11420181091443, "grad_norm": 0.020683126524090767, "learning_rate": 4.884455747430266e-07, "loss": 0.0001, "num_input_tokens_seen": 72121968, "step": 37480 }, { "epoch": 6.115017538135247, "grad_norm": 0.0003131729317829013, "learning_rate": 4.865162914533816e-07, "loss": 0.0003, "num_input_tokens_seen": 72132832, "step": 37485 }, { "epoch": 6.1158332653560645, "grad_norm": 0.0009984896751120687, "learning_rate": 4.845907884509809e-07, "loss": 0.0002, "num_input_tokens_seen": 72142448, "step": 37490 }, { "epoch": 6.1166489925768825, "grad_norm": 0.0006180782220326364, "learning_rate": 4.82669066032762e-07, "loss": 0.0003, "num_input_tokens_seen": 72151504, "step": 37495 }, { "epoch": 6.1174647197977, "grad_norm": 0.0015614940784871578, "learning_rate": 4.807511244950768e-07, "loss": 0.0005, "num_input_tokens_seen": 72161472, "step": 37500 }, { "epoch": 6.118280447018517, "grad_norm": 0.0009746617288328707, "learning_rate": 4.788369641336943e-07, "loss": 0.0004, "num_input_tokens_seen": 72171056, "step": 37505 }, { "epoch": 6.119096174239334, "grad_norm": 0.01848367042839527, "learning_rate": 4.769265852438032e-07, "loss": 0.0, "num_input_tokens_seen": 72180816, "step": 37510 }, { "epoch": 6.119911901460152, "grad_norm": 0.004535168409347534, "learning_rate": 4.750199881200124e-07, "loss": 0.0001, "num_input_tokens_seen": 72190800, "step": 37515 }, { "epoch": 6.120727628680969, "grad_norm": 0.0007056644535623491, "learning_rate": 4.7311717305633664e-07, "loss": 0.0, "num_input_tokens_seen": 72200224, "step": 37520 }, { "epoch": 6.121543355901786, "grad_norm": 0.038176026195287704, "learning_rate": 4.7121814034621623e-07, "loss": 0.0856, "num_input_tokens_seen": 72210912, "step": 37525 }, { "epoch": 6.122359083122603, "grad_norm": 0.012397459708154202, "learning_rate": 4.693228902825114e-07, "loss": 0.0001, "num_input_tokens_seen": 72220960, "step": 37530 }, { "epoch": 6.123174810343421, "grad_norm": 0.10591011494398117, "learning_rate": 4.6743142315748277e-07, "loss": 0.0002, "num_input_tokens_seen": 72231696, "step": 37535 }, { "epoch": 6.1239905375642385, "grad_norm": 0.015863994136452675, "learning_rate": 4.655437392628276e-07, "loss": 0.0001, "num_input_tokens_seen": 72240736, "step": 37540 }, { "epoch": 6.124806264785056, "grad_norm": 0.0013600196689367294, "learning_rate": 4.636598388896463e-07, "loss": 0.0003, "num_input_tokens_seen": 72250736, "step": 37545 }, { "epoch": 6.125621992005874, "grad_norm": 0.0022489719558507204, "learning_rate": 4.6177972232845925e-07, "loss": 0.0, "num_input_tokens_seen": 72260272, "step": 37550 }, { "epoch": 6.126437719226691, "grad_norm": 0.0033004307188093662, "learning_rate": 4.5990338986920953e-07, "loss": 0.0001, "num_input_tokens_seen": 72270432, "step": 37555 }, { "epoch": 6.127253446447508, "grad_norm": 0.0003805548185482621, "learning_rate": 4.5803084180124633e-07, "loss": 0.0, "num_input_tokens_seen": 72279840, "step": 37560 }, { "epoch": 6.128069173668325, "grad_norm": 0.021376099437475204, "learning_rate": 4.561620784133386e-07, "loss": 0.0, "num_input_tokens_seen": 72290784, "step": 37565 }, { "epoch": 6.128884900889143, "grad_norm": 0.0008299924666061997, "learning_rate": 4.5429709999367796e-07, "loss": 0.0, "num_input_tokens_seen": 72301952, "step": 37570 }, { "epoch": 6.12970062810996, "grad_norm": 0.35629817843437195, "learning_rate": 4.5243590682986223e-07, "loss": 0.0006, "num_input_tokens_seen": 72310704, "step": 37575 }, { "epoch": 6.130516355330777, "grad_norm": 4.572857051243773e-06, "learning_rate": 4.5057849920891735e-07, "loss": 0.0009, "num_input_tokens_seen": 72319808, "step": 37580 }, { "epoch": 6.1313320825515945, "grad_norm": 0.02266744337975979, "learning_rate": 4.487248774172698e-07, "loss": 0.0001, "num_input_tokens_seen": 72329776, "step": 37585 }, { "epoch": 6.1321478097724125, "grad_norm": 0.006646823137998581, "learning_rate": 4.4687504174077965e-07, "loss": 0.0003, "num_input_tokens_seen": 72339424, "step": 37590 }, { "epoch": 6.13296353699323, "grad_norm": 6.674238681793213, "learning_rate": 4.450289924647133e-07, "loss": 0.0053, "num_input_tokens_seen": 72348624, "step": 37595 }, { "epoch": 6.133779264214047, "grad_norm": 0.005763204302638769, "learning_rate": 4.431867298737513e-07, "loss": 0.0, "num_input_tokens_seen": 72358032, "step": 37600 }, { "epoch": 6.133779264214047, "eval_loss": 0.4275158941745758, "eval_runtime": 80.8961, "eval_samples_per_second": 33.685, "eval_steps_per_second": 16.849, "num_input_tokens_seen": 72358032, "step": 37600 }, { "epoch": 6.134594991434864, "grad_norm": 0.0009897038107737899, "learning_rate": 4.41348254251997e-07, "loss": 0.0627, "num_input_tokens_seen": 72366992, "step": 37605 }, { "epoch": 6.135410718655682, "grad_norm": 0.0012603654759004712, "learning_rate": 4.395135658829652e-07, "loss": 0.0, "num_input_tokens_seen": 72376112, "step": 37610 }, { "epoch": 6.136226445876499, "grad_norm": 0.018857544288039207, "learning_rate": 4.376826650495852e-07, "loss": 0.0001, "num_input_tokens_seen": 72386656, "step": 37615 }, { "epoch": 6.137042173097316, "grad_norm": 0.060737937688827515, "learning_rate": 4.358555520342117e-07, "loss": 0.0, "num_input_tokens_seen": 72395536, "step": 37620 }, { "epoch": 6.137857900318133, "grad_norm": 0.025484254583716393, "learning_rate": 4.3403222711860257e-07, "loss": 0.0003, "num_input_tokens_seen": 72405408, "step": 37625 }, { "epoch": 6.138673627538951, "grad_norm": 0.007928840816020966, "learning_rate": 4.3221269058394133e-07, "loss": 0.0, "num_input_tokens_seen": 72416240, "step": 37630 }, { "epoch": 6.139489354759768, "grad_norm": 0.0013771243393421173, "learning_rate": 4.303969427108173e-07, "loss": 0.0001, "num_input_tokens_seen": 72426896, "step": 37635 }, { "epoch": 6.1403050819805856, "grad_norm": 3.679261862998828e-05, "learning_rate": 4.2858498377924825e-07, "loss": 0.0, "num_input_tokens_seen": 72437952, "step": 37640 }, { "epoch": 6.141120809201403, "grad_norm": 0.0001613422209629789, "learning_rate": 4.267768140686579e-07, "loss": 0.0003, "num_input_tokens_seen": 72447552, "step": 37645 }, { "epoch": 6.141936536422221, "grad_norm": 0.764238178730011, "learning_rate": 4.2497243385788975e-07, "loss": 0.0005, "num_input_tokens_seen": 72456800, "step": 37650 }, { "epoch": 6.142752263643038, "grad_norm": 0.013893110677599907, "learning_rate": 4.231718434251991e-07, "loss": 0.0001, "num_input_tokens_seen": 72467520, "step": 37655 }, { "epoch": 6.143567990863855, "grad_norm": 0.189430832862854, "learning_rate": 4.213750430482666e-07, "loss": 0.0002, "num_input_tokens_seen": 72476816, "step": 37660 }, { "epoch": 6.144383718084672, "grad_norm": 0.011111067607998848, "learning_rate": 4.1958203300417054e-07, "loss": 0.0001, "num_input_tokens_seen": 72486208, "step": 37665 }, { "epoch": 6.14519944530549, "grad_norm": 0.0009204242960549891, "learning_rate": 4.177928135694259e-07, "loss": 0.0002, "num_input_tokens_seen": 72496896, "step": 37670 }, { "epoch": 6.146015172526307, "grad_norm": 0.012642601504921913, "learning_rate": 4.1600738501994807e-07, "loss": 0.0, "num_input_tokens_seen": 72506816, "step": 37675 }, { "epoch": 6.146830899747124, "grad_norm": 0.000895906996447593, "learning_rate": 4.1422574763107237e-07, "loss": 0.0, "num_input_tokens_seen": 72516528, "step": 37680 }, { "epoch": 6.1476466269679415, "grad_norm": 0.001189414062537253, "learning_rate": 4.124479016775512e-07, "loss": 0.0002, "num_input_tokens_seen": 72526656, "step": 37685 }, { "epoch": 6.1484623541887595, "grad_norm": 0.011999026872217655, "learning_rate": 4.106738474335514e-07, "loss": 0.0001, "num_input_tokens_seen": 72536320, "step": 37690 }, { "epoch": 6.149278081409577, "grad_norm": 0.011131152510643005, "learning_rate": 4.089035851726486e-07, "loss": 0.0001, "num_input_tokens_seen": 72545392, "step": 37695 }, { "epoch": 6.150093808630394, "grad_norm": 0.0019825228955596685, "learning_rate": 4.0713711516784937e-07, "loss": 0.0, "num_input_tokens_seen": 72555840, "step": 37700 }, { "epoch": 6.150909535851211, "grad_norm": 0.016121987253427505, "learning_rate": 4.05374437691558e-07, "loss": 0.0001, "num_input_tokens_seen": 72565232, "step": 37705 }, { "epoch": 6.151725263072029, "grad_norm": 0.14221343398094177, "learning_rate": 4.036155530156044e-07, "loss": 0.0001, "num_input_tokens_seen": 72574944, "step": 37710 }, { "epoch": 6.152540990292846, "grad_norm": 0.015314286574721336, "learning_rate": 4.018604614112298e-07, "loss": 0.0003, "num_input_tokens_seen": 72585088, "step": 37715 }, { "epoch": 6.153356717513663, "grad_norm": 0.008472204208374023, "learning_rate": 4.0010916314908996e-07, "loss": 0.0021, "num_input_tokens_seen": 72595920, "step": 37720 }, { "epoch": 6.154172444734481, "grad_norm": 0.03270888328552246, "learning_rate": 3.983616584992578e-07, "loss": 0.0176, "num_input_tokens_seen": 72605872, "step": 37725 }, { "epoch": 6.154988171955298, "grad_norm": 0.0005438809748739004, "learning_rate": 3.9661794773122595e-07, "loss": 0.0, "num_input_tokens_seen": 72615328, "step": 37730 }, { "epoch": 6.1558038991761155, "grad_norm": 0.00027201726334169507, "learning_rate": 3.9487803111388777e-07, "loss": 0.0001, "num_input_tokens_seen": 72625168, "step": 37735 }, { "epoch": 6.156619626396933, "grad_norm": 0.02234921231865883, "learning_rate": 3.9314190891556747e-07, "loss": 0.0, "num_input_tokens_seen": 72634128, "step": 37740 }, { "epoch": 6.157435353617751, "grad_norm": 0.3198321461677551, "learning_rate": 3.914095814039925e-07, "loss": 0.0004, "num_input_tokens_seen": 72643712, "step": 37745 }, { "epoch": 6.158251080838568, "grad_norm": 0.30486732721328735, "learning_rate": 3.896810488463104e-07, "loss": 0.0003, "num_input_tokens_seen": 72653984, "step": 37750 }, { "epoch": 6.159066808059385, "grad_norm": 0.007602867670357227, "learning_rate": 3.8795631150908565e-07, "loss": 0.0, "num_input_tokens_seen": 72664272, "step": 37755 }, { "epoch": 6.159882535280202, "grad_norm": 0.0048869457095861435, "learning_rate": 3.862353696582888e-07, "loss": 0.0001, "num_input_tokens_seen": 72673232, "step": 37760 }, { "epoch": 6.16069826250102, "grad_norm": 3.774412834900431e-05, "learning_rate": 3.8451822355931313e-07, "loss": 0.0001, "num_input_tokens_seen": 72682848, "step": 37765 }, { "epoch": 6.161513989721837, "grad_norm": 0.0003592060529626906, "learning_rate": 3.82804873476969e-07, "loss": 0.0006, "num_input_tokens_seen": 72693344, "step": 37770 }, { "epoch": 6.162329716942654, "grad_norm": 0.019628573209047318, "learning_rate": 3.810953196754702e-07, "loss": 0.0, "num_input_tokens_seen": 72702448, "step": 37775 }, { "epoch": 6.163145444163471, "grad_norm": 0.004069392569363117, "learning_rate": 3.793895624184529e-07, "loss": 0.0, "num_input_tokens_seen": 72712208, "step": 37780 }, { "epoch": 6.1639611713842895, "grad_norm": 0.20643098652362823, "learning_rate": 3.776876019689679e-07, "loss": 0.0003, "num_input_tokens_seen": 72722064, "step": 37785 }, { "epoch": 6.164776898605107, "grad_norm": 2.1042048931121826, "learning_rate": 3.7598943858947743e-07, "loss": 0.0034, "num_input_tokens_seen": 72730864, "step": 37790 }, { "epoch": 6.165592625825924, "grad_norm": 0.21159300208091736, "learning_rate": 3.742950725418637e-07, "loss": 0.0007, "num_input_tokens_seen": 72740576, "step": 37795 }, { "epoch": 6.166408353046741, "grad_norm": 0.0030823908746242523, "learning_rate": 3.726045040874093e-07, "loss": 0.0205, "num_input_tokens_seen": 72749840, "step": 37800 }, { "epoch": 6.166408353046741, "eval_loss": 0.43063226342201233, "eval_runtime": 80.8932, "eval_samples_per_second": 33.686, "eval_steps_per_second": 16.849, "num_input_tokens_seen": 72749840, "step": 37800 }, { "epoch": 6.167224080267559, "grad_norm": 0.00038093156763352454, "learning_rate": 3.709177334868308e-07, "loss": 0.0001, "num_input_tokens_seen": 72759584, "step": 37805 }, { "epoch": 6.168039807488376, "grad_norm": 0.012056652456521988, "learning_rate": 3.692347610002478e-07, "loss": 0.0002, "num_input_tokens_seen": 72768576, "step": 37810 }, { "epoch": 6.168855534709193, "grad_norm": 0.012469054199755192, "learning_rate": 3.675555868871916e-07, "loss": 0.0, "num_input_tokens_seen": 72777216, "step": 37815 }, { "epoch": 6.16967126193001, "grad_norm": 0.01621723361313343, "learning_rate": 3.658802114066162e-07, "loss": 0.0002, "num_input_tokens_seen": 72787584, "step": 37820 }, { "epoch": 6.170486989150828, "grad_norm": 0.0009080781601369381, "learning_rate": 3.6420863481688437e-07, "loss": 0.0001, "num_input_tokens_seen": 72797328, "step": 37825 }, { "epoch": 6.171302716371645, "grad_norm": 0.002025323687121272, "learning_rate": 3.625408573757705e-07, "loss": 0.0, "num_input_tokens_seen": 72807056, "step": 37830 }, { "epoch": 6.1721184435924625, "grad_norm": 0.09320291131734848, "learning_rate": 3.608768793404743e-07, "loss": 0.0002, "num_input_tokens_seen": 72816288, "step": 37835 }, { "epoch": 6.17293417081328, "grad_norm": 0.0014322957722470164, "learning_rate": 3.592167009675934e-07, "loss": 0.0, "num_input_tokens_seen": 72825888, "step": 37840 }, { "epoch": 6.173749898034098, "grad_norm": 0.00019626734138000757, "learning_rate": 3.575603225131563e-07, "loss": 0.0, "num_input_tokens_seen": 72835232, "step": 37845 }, { "epoch": 6.174565625254915, "grad_norm": 0.000513850711286068, "learning_rate": 3.55907744232592e-07, "loss": 0.0489, "num_input_tokens_seen": 72845168, "step": 37850 }, { "epoch": 6.175381352475732, "grad_norm": 0.005138747859746218, "learning_rate": 3.5425896638075217e-07, "loss": 0.0001, "num_input_tokens_seen": 72854240, "step": 37855 }, { "epoch": 6.176197079696549, "grad_norm": 0.046071551740169525, "learning_rate": 3.5261398921189736e-07, "loss": 0.0, "num_input_tokens_seen": 72863024, "step": 37860 }, { "epoch": 6.177012806917367, "grad_norm": 0.006841407623142004, "learning_rate": 3.509728129797024e-07, "loss": 0.0004, "num_input_tokens_seen": 72873056, "step": 37865 }, { "epoch": 6.177828534138184, "grad_norm": 0.0008780325297266245, "learning_rate": 3.4933543793725656e-07, "loss": 0.0005, "num_input_tokens_seen": 72882768, "step": 37870 }, { "epoch": 6.178644261359001, "grad_norm": 0.004809126257896423, "learning_rate": 3.4770186433707163e-07, "loss": 0.0001, "num_input_tokens_seen": 72892832, "step": 37875 }, { "epoch": 6.1794599885798185, "grad_norm": 0.030070152133703232, "learning_rate": 3.4607209243105453e-07, "loss": 0.0002, "num_input_tokens_seen": 72902480, "step": 37880 }, { "epoch": 6.1802757158006365, "grad_norm": 0.04537274315953255, "learning_rate": 3.444461224705431e-07, "loss": 0.0001, "num_input_tokens_seen": 72912208, "step": 37885 }, { "epoch": 6.181091443021454, "grad_norm": 0.04185444489121437, "learning_rate": 3.4282395470628116e-07, "loss": 0.0001, "num_input_tokens_seen": 72920992, "step": 37890 }, { "epoch": 6.181907170242271, "grad_norm": 0.006838340312242508, "learning_rate": 3.4120558938842417e-07, "loss": 0.0002, "num_input_tokens_seen": 72930880, "step": 37895 }, { "epoch": 6.182722897463089, "grad_norm": 0.00019899892504327, "learning_rate": 3.395910267665503e-07, "loss": 0.0, "num_input_tokens_seen": 72940512, "step": 37900 }, { "epoch": 6.183538624683906, "grad_norm": 0.03334561735391617, "learning_rate": 3.3798026708964094e-07, "loss": 0.0001, "num_input_tokens_seen": 72950144, "step": 37905 }, { "epoch": 6.184354351904723, "grad_norm": 0.0037364677991718054, "learning_rate": 3.3637331060609456e-07, "loss": 0.0001, "num_input_tokens_seen": 72960832, "step": 37910 }, { "epoch": 6.18517007912554, "grad_norm": 0.0005289844702929258, "learning_rate": 3.3477015756372966e-07, "loss": 0.0455, "num_input_tokens_seen": 72969392, "step": 37915 }, { "epoch": 6.185985806346358, "grad_norm": 0.22829431295394897, "learning_rate": 3.3317080820976785e-07, "loss": 0.0166, "num_input_tokens_seen": 72978512, "step": 37920 }, { "epoch": 6.186801533567175, "grad_norm": 0.019450142979621887, "learning_rate": 3.315752627908508e-07, "loss": 0.0001, "num_input_tokens_seen": 72988032, "step": 37925 }, { "epoch": 6.1876172607879925, "grad_norm": 0.007784212473779917, "learning_rate": 3.299835215530317e-07, "loss": 0.0005, "num_input_tokens_seen": 72997360, "step": 37930 }, { "epoch": 6.18843298800881, "grad_norm": 0.026570003479719162, "learning_rate": 3.2839558474177245e-07, "loss": 0.0, "num_input_tokens_seen": 73007088, "step": 37935 }, { "epoch": 6.189248715229628, "grad_norm": 0.005044455174356699, "learning_rate": 3.2681145260196056e-07, "loss": 0.0003, "num_input_tokens_seen": 73016064, "step": 37940 }, { "epoch": 6.190064442450445, "grad_norm": 0.0028974120505154133, "learning_rate": 3.252311253778839e-07, "loss": 0.0002, "num_input_tokens_seen": 73026656, "step": 37945 }, { "epoch": 6.190880169671262, "grad_norm": 0.8925571441650391, "learning_rate": 3.2365460331325034e-07, "loss": 0.0005, "num_input_tokens_seen": 73037088, "step": 37950 }, { "epoch": 6.191695896892079, "grad_norm": 0.00017623142048250884, "learning_rate": 3.2208188665117934e-07, "loss": 0.0, "num_input_tokens_seen": 73046544, "step": 37955 }, { "epoch": 6.192511624112897, "grad_norm": 0.009228883311152458, "learning_rate": 3.205129756342018e-07, "loss": 0.0002, "num_input_tokens_seen": 73054640, "step": 37960 }, { "epoch": 6.193327351333714, "grad_norm": 0.0027208079118281603, "learning_rate": 3.189478705042659e-07, "loss": 0.0007, "num_input_tokens_seen": 73062160, "step": 37965 }, { "epoch": 6.194143078554531, "grad_norm": 0.041423603892326355, "learning_rate": 3.173865715027341e-07, "loss": 0.0001, "num_input_tokens_seen": 73071152, "step": 37970 }, { "epoch": 6.194958805775348, "grad_norm": 0.00032420112984254956, "learning_rate": 3.158290788703694e-07, "loss": 0.0003, "num_input_tokens_seen": 73079808, "step": 37975 }, { "epoch": 6.195774532996166, "grad_norm": 0.011294618248939514, "learning_rate": 3.1427539284736297e-07, "loss": 0.0, "num_input_tokens_seen": 73090320, "step": 37980 }, { "epoch": 6.196590260216984, "grad_norm": 11.844620704650879, "learning_rate": 3.127255136733093e-07, "loss": 0.0374, "num_input_tokens_seen": 73099200, "step": 37985 }, { "epoch": 6.197405987437801, "grad_norm": 0.09276025742292404, "learning_rate": 3.1117944158722544e-07, "loss": 0.0001, "num_input_tokens_seen": 73109456, "step": 37990 }, { "epoch": 6.198221714658618, "grad_norm": 0.02068670652806759, "learning_rate": 3.0963717682752635e-07, "loss": 0.0116, "num_input_tokens_seen": 73118912, "step": 37995 }, { "epoch": 6.199037441879436, "grad_norm": 0.0005831625894643366, "learning_rate": 3.080987196320578e-07, "loss": 0.0, "num_input_tokens_seen": 73128448, "step": 38000 }, { "epoch": 6.199037441879436, "eval_loss": 0.43132612109184265, "eval_runtime": 80.7682, "eval_samples_per_second": 33.739, "eval_steps_per_second": 16.875, "num_input_tokens_seen": 73128448, "step": 38000 }, { "epoch": 6.199853169100253, "grad_norm": 0.012328514829277992, "learning_rate": 3.065640702380607e-07, "loss": 0.0001, "num_input_tokens_seen": 73138224, "step": 38005 }, { "epoch": 6.20066889632107, "grad_norm": 0.00012192685971967876, "learning_rate": 3.050332288822011e-07, "loss": 0.0, "num_input_tokens_seen": 73147328, "step": 38010 }, { "epoch": 6.201484623541887, "grad_norm": 0.005297380965203047, "learning_rate": 3.035061958005542e-07, "loss": 0.0001, "num_input_tokens_seen": 73158528, "step": 38015 }, { "epoch": 6.202300350762705, "grad_norm": 0.00033935357350856066, "learning_rate": 3.019829712286093e-07, "loss": 0.0001, "num_input_tokens_seen": 73169184, "step": 38020 }, { "epoch": 6.203116077983522, "grad_norm": 0.004501851741224527, "learning_rate": 3.004635554012647e-07, "loss": 0.003, "num_input_tokens_seen": 73178384, "step": 38025 }, { "epoch": 6.2039318052043395, "grad_norm": 0.04251653701066971, "learning_rate": 2.9894794855283017e-07, "loss": 0.0001, "num_input_tokens_seen": 73188864, "step": 38030 }, { "epoch": 6.204747532425157, "grad_norm": 0.0010082449298352003, "learning_rate": 2.9743615091703816e-07, "loss": 0.0, "num_input_tokens_seen": 73197872, "step": 38035 }, { "epoch": 6.205563259645975, "grad_norm": 0.00010766710329335183, "learning_rate": 2.959281627270216e-07, "loss": 0.0, "num_input_tokens_seen": 73206752, "step": 38040 }, { "epoch": 6.206378986866792, "grad_norm": 0.006396548822522163, "learning_rate": 2.944239842153362e-07, "loss": 0.0, "num_input_tokens_seen": 73216480, "step": 38045 }, { "epoch": 6.207194714087609, "grad_norm": 0.033965814858675, "learning_rate": 2.929236156139381e-07, "loss": 0.0191, "num_input_tokens_seen": 73225552, "step": 38050 }, { "epoch": 6.208010441308426, "grad_norm": 0.0001619336981093511, "learning_rate": 2.9142705715420883e-07, "loss": 0.0, "num_input_tokens_seen": 73235264, "step": 38055 }, { "epoch": 6.208826168529244, "grad_norm": 0.0008583989110775292, "learning_rate": 2.8993430906693595e-07, "loss": 0.0001, "num_input_tokens_seen": 73245216, "step": 38060 }, { "epoch": 6.209641895750061, "grad_norm": 0.022865796461701393, "learning_rate": 2.88445371582316e-07, "loss": 0.0001, "num_input_tokens_seen": 73254512, "step": 38065 }, { "epoch": 6.210457622970878, "grad_norm": 0.0033078710548579693, "learning_rate": 2.8696024492996796e-07, "loss": 0.0, "num_input_tokens_seen": 73263824, "step": 38070 }, { "epoch": 6.211273350191696, "grad_norm": 0.0008394717006012797, "learning_rate": 2.854789293389115e-07, "loss": 0.0, "num_input_tokens_seen": 73273232, "step": 38075 }, { "epoch": 6.2120890774125135, "grad_norm": 26.766817092895508, "learning_rate": 2.8400142503758606e-07, "loss": 0.1444, "num_input_tokens_seen": 73282976, "step": 38080 }, { "epoch": 6.212904804633331, "grad_norm": 0.07352882623672485, "learning_rate": 2.8252773225384276e-07, "loss": 0.0002, "num_input_tokens_seen": 73293360, "step": 38085 }, { "epoch": 6.213720531854148, "grad_norm": 0.0038054136093705893, "learning_rate": 2.8105785121494143e-07, "loss": 0.0007, "num_input_tokens_seen": 73304032, "step": 38090 }, { "epoch": 6.214536259074965, "grad_norm": 0.0032637575641274452, "learning_rate": 2.795917821475563e-07, "loss": 0.0, "num_input_tokens_seen": 73313424, "step": 38095 }, { "epoch": 6.215351986295783, "grad_norm": 0.0002779765927698463, "learning_rate": 2.78129525277776e-07, "loss": 0.0, "num_input_tokens_seen": 73323152, "step": 38100 }, { "epoch": 6.2161677135166, "grad_norm": 0.004255656152963638, "learning_rate": 2.766710808310952e-07, "loss": 0.0877, "num_input_tokens_seen": 73332736, "step": 38105 }, { "epoch": 6.216983440737417, "grad_norm": 5.022926416131668e-05, "learning_rate": 2.7521644903242827e-07, "loss": 0.1026, "num_input_tokens_seen": 73342512, "step": 38110 }, { "epoch": 6.217799167958235, "grad_norm": 0.3161628842353821, "learning_rate": 2.7376563010609593e-07, "loss": 0.0339, "num_input_tokens_seen": 73351424, "step": 38115 }, { "epoch": 6.218614895179052, "grad_norm": 5.251771290204488e-05, "learning_rate": 2.72318624275833e-07, "loss": 0.0, "num_input_tokens_seen": 73360976, "step": 38120 }, { "epoch": 6.219430622399869, "grad_norm": 0.0012282413663342595, "learning_rate": 2.7087543176478324e-07, "loss": 0.069, "num_input_tokens_seen": 73371184, "step": 38125 }, { "epoch": 6.220246349620687, "grad_norm": 0.0011438218643888831, "learning_rate": 2.694360527955103e-07, "loss": 0.0, "num_input_tokens_seen": 73380128, "step": 38130 }, { "epoch": 6.221062076841505, "grad_norm": 0.7103810906410217, "learning_rate": 2.680004875899811e-07, "loss": 0.001, "num_input_tokens_seen": 73390528, "step": 38135 }, { "epoch": 6.221877804062322, "grad_norm": 0.28679919242858887, "learning_rate": 2.665687363695768e-07, "loss": 0.0004, "num_input_tokens_seen": 73399680, "step": 38140 }, { "epoch": 6.222693531283139, "grad_norm": 35.854736328125, "learning_rate": 2.6514079935509584e-07, "loss": 0.065, "num_input_tokens_seen": 73409408, "step": 38145 }, { "epoch": 6.223509258503956, "grad_norm": 0.0018989169038832188, "learning_rate": 2.6371667676673983e-07, "loss": 0.001, "num_input_tokens_seen": 73419296, "step": 38150 }, { "epoch": 6.224324985724774, "grad_norm": 0.0005773166776634753, "learning_rate": 2.6229636882412755e-07, "loss": 0.0649, "num_input_tokens_seen": 73429152, "step": 38155 }, { "epoch": 6.225140712945591, "grad_norm": 0.3438016474246979, "learning_rate": 2.6087987574628935e-07, "loss": 0.0001, "num_input_tokens_seen": 73439040, "step": 38160 }, { "epoch": 6.225956440166408, "grad_norm": 0.1980399489402771, "learning_rate": 2.5946719775166437e-07, "loss": 0.0002, "num_input_tokens_seen": 73448112, "step": 38165 }, { "epoch": 6.226772167387225, "grad_norm": 0.05778291076421738, "learning_rate": 2.5805833505810616e-07, "loss": 0.0961, "num_input_tokens_seen": 73458480, "step": 38170 }, { "epoch": 6.227587894608043, "grad_norm": 0.009439200162887573, "learning_rate": 2.566532878828798e-07, "loss": 0.0007, "num_input_tokens_seen": 73468800, "step": 38175 }, { "epoch": 6.2284036218288605, "grad_norm": 0.0027965412009507418, "learning_rate": 2.552520564426619e-07, "loss": 0.0, "num_input_tokens_seen": 73479216, "step": 38180 }, { "epoch": 6.229219349049678, "grad_norm": 0.001868593622930348, "learning_rate": 2.5385464095353803e-07, "loss": 0.0, "num_input_tokens_seen": 73488416, "step": 38185 }, { "epoch": 6.230035076270495, "grad_norm": 0.01312166266143322, "learning_rate": 2.5246104163100804e-07, "loss": 0.1545, "num_input_tokens_seen": 73497872, "step": 38190 }, { "epoch": 6.230850803491313, "grad_norm": 0.102153480052948, "learning_rate": 2.510712586899833e-07, "loss": 0.0001, "num_input_tokens_seen": 73508160, "step": 38195 }, { "epoch": 6.23166653071213, "grad_norm": 0.010422631166875362, "learning_rate": 2.4968529234478124e-07, "loss": 0.0489, "num_input_tokens_seen": 73518048, "step": 38200 }, { "epoch": 6.23166653071213, "eval_loss": 0.42790207266807556, "eval_runtime": 80.9529, "eval_samples_per_second": 33.662, "eval_steps_per_second": 16.837, "num_input_tokens_seen": 73518048, "step": 38200 }, { "epoch": 6.232482257932947, "grad_norm": 0.015803473070263863, "learning_rate": 2.483031428091448e-07, "loss": 0.0002, "num_input_tokens_seen": 73527520, "step": 38205 }, { "epoch": 6.233297985153764, "grad_norm": 2.6475634513190016e-05, "learning_rate": 2.469248102962091e-07, "loss": 0.0, "num_input_tokens_seen": 73537616, "step": 38210 }, { "epoch": 6.234113712374582, "grad_norm": 2.8936452508787625e-05, "learning_rate": 2.4555029501853455e-07, "loss": 0.0589, "num_input_tokens_seen": 73547584, "step": 38215 }, { "epoch": 6.234929439595399, "grad_norm": 0.0056885662488639355, "learning_rate": 2.441795971880906e-07, "loss": 0.0005, "num_input_tokens_seen": 73557472, "step": 38220 }, { "epoch": 6.2357451668162165, "grad_norm": 0.007597015239298344, "learning_rate": 2.4281271701625255e-07, "loss": 0.0, "num_input_tokens_seen": 73567200, "step": 38225 }, { "epoch": 6.236560894037034, "grad_norm": 0.01771087571978569, "learning_rate": 2.4144965471381007e-07, "loss": 0.0, "num_input_tokens_seen": 73577520, "step": 38230 }, { "epoch": 6.237376621257852, "grad_norm": 0.000918201229069382, "learning_rate": 2.400904104909674e-07, "loss": 0.0, "num_input_tokens_seen": 73587488, "step": 38235 }, { "epoch": 6.238192348478669, "grad_norm": 0.002090360736474395, "learning_rate": 2.3873498455733725e-07, "loss": 0.0001, "num_input_tokens_seen": 73596992, "step": 38240 }, { "epoch": 6.239008075699486, "grad_norm": 0.004264052491635084, "learning_rate": 2.3738337712194137e-07, "loss": 0.0002, "num_input_tokens_seen": 73606176, "step": 38245 }, { "epoch": 6.239823802920303, "grad_norm": 0.0005753637524321675, "learning_rate": 2.3603558839321305e-07, "loss": 0.0001, "num_input_tokens_seen": 73616192, "step": 38250 }, { "epoch": 6.240639530141121, "grad_norm": 0.15240179002285004, "learning_rate": 2.3469161857900267e-07, "loss": 0.0006, "num_input_tokens_seen": 73625840, "step": 38255 }, { "epoch": 6.241455257361938, "grad_norm": 0.06360458582639694, "learning_rate": 2.3335146788656393e-07, "loss": 0.0651, "num_input_tokens_seen": 73636272, "step": 38260 }, { "epoch": 6.242270984582755, "grad_norm": 0.004018230829387903, "learning_rate": 2.3201513652256757e-07, "loss": 0.0002, "num_input_tokens_seen": 73645728, "step": 38265 }, { "epoch": 6.243086711803572, "grad_norm": 0.22967888414859772, "learning_rate": 2.3068262469308766e-07, "loss": 0.0002, "num_input_tokens_seen": 73654352, "step": 38270 }, { "epoch": 6.2439024390243905, "grad_norm": 0.0009506201604381204, "learning_rate": 2.2935393260362093e-07, "loss": 0.0001, "num_input_tokens_seen": 73664336, "step": 38275 }, { "epoch": 6.244718166245208, "grad_norm": 0.00016358945867978036, "learning_rate": 2.2802906045906458e-07, "loss": 0.0001, "num_input_tokens_seen": 73673168, "step": 38280 }, { "epoch": 6.245533893466025, "grad_norm": 0.0005773482262156904, "learning_rate": 2.2670800846373018e-07, "loss": 0.0001, "num_input_tokens_seen": 73681296, "step": 38285 }, { "epoch": 6.246349620686843, "grad_norm": 0.004640878178179264, "learning_rate": 2.2539077682134367e-07, "loss": 0.0284, "num_input_tokens_seen": 73691808, "step": 38290 }, { "epoch": 6.24716534790766, "grad_norm": 0.017230065539479256, "learning_rate": 2.2407736573503423e-07, "loss": 0.0, "num_input_tokens_seen": 73701904, "step": 38295 }, { "epoch": 6.247981075128477, "grad_norm": 0.008947390131652355, "learning_rate": 2.2276777540735093e-07, "loss": 0.0001, "num_input_tokens_seen": 73712496, "step": 38300 }, { "epoch": 6.248796802349294, "grad_norm": 0.00046550086699426174, "learning_rate": 2.2146200604024613e-07, "loss": 0.0001, "num_input_tokens_seen": 73723040, "step": 38305 }, { "epoch": 6.249612529570112, "grad_norm": 0.0011918229283764958, "learning_rate": 2.2016005783508375e-07, "loss": 0.0, "num_input_tokens_seen": 73732192, "step": 38310 }, { "epoch": 6.250428256790929, "grad_norm": 0.0013637581141665578, "learning_rate": 2.1886193099264763e-07, "loss": 0.0457, "num_input_tokens_seen": 73741536, "step": 38315 }, { "epoch": 6.251243984011746, "grad_norm": 1.7337129975203425e-05, "learning_rate": 2.175676257131165e-07, "loss": 0.0011, "num_input_tokens_seen": 73752432, "step": 38320 }, { "epoch": 6.2520597112325635, "grad_norm": 0.0069271293468773365, "learning_rate": 2.162771421960974e-07, "loss": 0.0001, "num_input_tokens_seen": 73762176, "step": 38325 }, { "epoch": 6.252875438453382, "grad_norm": 0.0009327371371909976, "learning_rate": 2.1499048064059224e-07, "loss": 0.0, "num_input_tokens_seen": 73771904, "step": 38330 }, { "epoch": 6.253691165674199, "grad_norm": 0.004393650218844414, "learning_rate": 2.1370764124502285e-07, "loss": 0.0, "num_input_tokens_seen": 73779840, "step": 38335 }, { "epoch": 6.254506892895016, "grad_norm": 0.2524547278881073, "learning_rate": 2.1242862420721988e-07, "loss": 0.0003, "num_input_tokens_seen": 73790368, "step": 38340 }, { "epoch": 6.255322620115833, "grad_norm": 17.634403228759766, "learning_rate": 2.1115342972442276e-07, "loss": 0.1429, "num_input_tokens_seen": 73801872, "step": 38345 }, { "epoch": 6.256138347336651, "grad_norm": 0.1925666779279709, "learning_rate": 2.0988205799328252e-07, "loss": 0.0002, "num_input_tokens_seen": 73810944, "step": 38350 }, { "epoch": 6.256954074557468, "grad_norm": 0.00020445536938495934, "learning_rate": 2.0861450920986182e-07, "loss": 0.0, "num_input_tokens_seen": 73820944, "step": 38355 }, { "epoch": 6.257769801778285, "grad_norm": 0.01049358956515789, "learning_rate": 2.07350783569632e-07, "loss": 0.0, "num_input_tokens_seen": 73831216, "step": 38360 }, { "epoch": 6.258585528999102, "grad_norm": 0.01721576415002346, "learning_rate": 2.060908812674761e-07, "loss": 0.0469, "num_input_tokens_seen": 73841024, "step": 38365 }, { "epoch": 6.25940125621992, "grad_norm": 0.0688231810927391, "learning_rate": 2.0483480249768317e-07, "loss": 0.0002, "num_input_tokens_seen": 73850672, "step": 38370 }, { "epoch": 6.2602169834407375, "grad_norm": 0.05295912176370621, "learning_rate": 2.035825474539621e-07, "loss": 0.0004, "num_input_tokens_seen": 73860928, "step": 38375 }, { "epoch": 6.261032710661555, "grad_norm": 0.0017493595369160175, "learning_rate": 2.0233411632942235e-07, "loss": 0.0014, "num_input_tokens_seen": 73869824, "step": 38380 }, { "epoch": 6.261848437882372, "grad_norm": 0.038942378014326096, "learning_rate": 2.0108950931658764e-07, "loss": 0.0001, "num_input_tokens_seen": 73880880, "step": 38385 }, { "epoch": 6.26266416510319, "grad_norm": 0.009261072613298893, "learning_rate": 1.998487266073934e-07, "loss": 0.0014, "num_input_tokens_seen": 73890352, "step": 38390 }, { "epoch": 6.263479892324007, "grad_norm": 0.08335316181182861, "learning_rate": 1.986117683931865e-07, "loss": 0.002, "num_input_tokens_seen": 73900624, "step": 38395 }, { "epoch": 6.264295619544824, "grad_norm": 0.000517318956553936, "learning_rate": 1.9737863486471442e-07, "loss": 0.0027, "num_input_tokens_seen": 73911328, "step": 38400 }, { "epoch": 6.264295619544824, "eval_loss": 0.43132084608078003, "eval_runtime": 81.0195, "eval_samples_per_second": 33.634, "eval_steps_per_second": 16.823, "num_input_tokens_seen": 73911328, "step": 38400 }, { "epoch": 6.265111346765641, "grad_norm": 0.10452427715063095, "learning_rate": 1.9614932621215e-07, "loss": 0.0001, "num_input_tokens_seen": 73919952, "step": 38405 }, { "epoch": 6.265927073986459, "grad_norm": 0.01603737473487854, "learning_rate": 1.9492384262506102e-07, "loss": 0.0549, "num_input_tokens_seen": 73928752, "step": 38410 }, { "epoch": 6.266742801207276, "grad_norm": 0.030529038980603218, "learning_rate": 1.9370218429243524e-07, "loss": 0.0, "num_input_tokens_seen": 73938816, "step": 38415 }, { "epoch": 6.2675585284280935, "grad_norm": 15.628142356872559, "learning_rate": 1.9248435140267197e-07, "loss": 0.0633, "num_input_tokens_seen": 73948752, "step": 38420 }, { "epoch": 6.268374255648911, "grad_norm": 0.10009320080280304, "learning_rate": 1.9127034414356814e-07, "loss": 0.0002, "num_input_tokens_seen": 73957808, "step": 38425 }, { "epoch": 6.269189982869729, "grad_norm": 0.0018680905923247337, "learning_rate": 1.9006016270234627e-07, "loss": 0.1023, "num_input_tokens_seen": 73966640, "step": 38430 }, { "epoch": 6.270005710090546, "grad_norm": 8.943584442138672, "learning_rate": 1.888538072656293e-07, "loss": 0.049, "num_input_tokens_seen": 73977776, "step": 38435 }, { "epoch": 6.270821437311363, "grad_norm": 0.0061329626478254795, "learning_rate": 1.8765127801944893e-07, "loss": 0.0, "num_input_tokens_seen": 73986848, "step": 38440 }, { "epoch": 6.27163716453218, "grad_norm": 0.0006664322572760284, "learning_rate": 1.8645257514925406e-07, "loss": 0.0168, "num_input_tokens_seen": 73996672, "step": 38445 }, { "epoch": 6.272452891752998, "grad_norm": 0.004873103462159634, "learning_rate": 1.8525769883989685e-07, "loss": 0.0, "num_input_tokens_seen": 74006656, "step": 38450 }, { "epoch": 6.273268618973815, "grad_norm": 0.04373672232031822, "learning_rate": 1.8406664927564654e-07, "loss": 0.0001, "num_input_tokens_seen": 74015248, "step": 38455 }, { "epoch": 6.274084346194632, "grad_norm": 0.0020727806258946657, "learning_rate": 1.8287942664017566e-07, "loss": 0.0, "num_input_tokens_seen": 74024848, "step": 38460 }, { "epoch": 6.27490007341545, "grad_norm": 0.012044518254697323, "learning_rate": 1.8169603111656552e-07, "loss": 0.1751, "num_input_tokens_seen": 74033632, "step": 38465 }, { "epoch": 6.275715800636267, "grad_norm": 0.04660499468445778, "learning_rate": 1.805164628873146e-07, "loss": 0.0001, "num_input_tokens_seen": 74043120, "step": 38470 }, { "epoch": 6.276531527857085, "grad_norm": 0.0003341910778544843, "learning_rate": 1.793407221343274e-07, "loss": 0.0001, "num_input_tokens_seen": 74052656, "step": 38475 }, { "epoch": 6.277347255077902, "grad_norm": 0.040517840534448624, "learning_rate": 1.781688090389172e-07, "loss": 0.0002, "num_input_tokens_seen": 74060864, "step": 38480 }, { "epoch": 6.27816298229872, "grad_norm": 0.019437633454799652, "learning_rate": 1.770007237818061e-07, "loss": 0.0001, "num_input_tokens_seen": 74071360, "step": 38485 }, { "epoch": 6.278978709519537, "grad_norm": 0.050211332738399506, "learning_rate": 1.7583646654313059e-07, "loss": 0.0163, "num_input_tokens_seen": 74080736, "step": 38490 }, { "epoch": 6.279794436740354, "grad_norm": 0.055938366800546646, "learning_rate": 1.7467603750242757e-07, "loss": 0.0001, "num_input_tokens_seen": 74090400, "step": 38495 }, { "epoch": 6.280610163961171, "grad_norm": 0.0002876278304029256, "learning_rate": 1.7351943683865944e-07, "loss": 0.0002, "num_input_tokens_seen": 74101312, "step": 38500 }, { "epoch": 6.281425891181989, "grad_norm": 0.0012514765840023756, "learning_rate": 1.723666647301808e-07, "loss": 0.0, "num_input_tokens_seen": 74110832, "step": 38505 }, { "epoch": 6.282241618402806, "grad_norm": 0.027702588587999344, "learning_rate": 1.712177213547661e-07, "loss": 0.0003, "num_input_tokens_seen": 74119520, "step": 38510 }, { "epoch": 6.283057345623623, "grad_norm": 0.011245502158999443, "learning_rate": 1.7007260688959581e-07, "loss": 0.0, "num_input_tokens_seen": 74129584, "step": 38515 }, { "epoch": 6.2838730728444405, "grad_norm": 0.0009693837491795421, "learning_rate": 1.68931321511262e-07, "loss": 0.0, "num_input_tokens_seen": 74138784, "step": 38520 }, { "epoch": 6.2846888000652585, "grad_norm": 0.0020931106992065907, "learning_rate": 1.6779386539576835e-07, "loss": 0.0, "num_input_tokens_seen": 74147776, "step": 38525 }, { "epoch": 6.285504527286076, "grad_norm": 0.12587322294712067, "learning_rate": 1.666602387185162e-07, "loss": 0.0002, "num_input_tokens_seen": 74157008, "step": 38530 }, { "epoch": 6.286320254506893, "grad_norm": 0.3880302309989929, "learning_rate": 1.655304416543352e-07, "loss": 0.0014, "num_input_tokens_seen": 74165888, "step": 38535 }, { "epoch": 6.28713598172771, "grad_norm": 0.002088974928483367, "learning_rate": 1.6440447437744698e-07, "loss": 0.0001, "num_input_tokens_seen": 74175232, "step": 38540 }, { "epoch": 6.287951708948528, "grad_norm": 0.047933295369148254, "learning_rate": 1.6328233706149332e-07, "loss": 0.047, "num_input_tokens_seen": 74185360, "step": 38545 }, { "epoch": 6.288767436169345, "grad_norm": 0.0003854745300486684, "learning_rate": 1.6216402987951906e-07, "loss": 0.0003, "num_input_tokens_seen": 74194368, "step": 38550 }, { "epoch": 6.289583163390162, "grad_norm": 0.02094748429954052, "learning_rate": 1.6104955300398627e-07, "loss": 0.0001, "num_input_tokens_seen": 74204448, "step": 38555 }, { "epoch": 6.290398890610979, "grad_norm": 0.0056582167744636536, "learning_rate": 1.5993890660675748e-07, "loss": 0.0, "num_input_tokens_seen": 74214272, "step": 38560 }, { "epoch": 6.291214617831797, "grad_norm": 0.00039423079579137266, "learning_rate": 1.5883209085910678e-07, "loss": 0.0001, "num_input_tokens_seen": 74224416, "step": 38565 }, { "epoch": 6.2920303450526145, "grad_norm": 0.015253539197146893, "learning_rate": 1.5772910593172264e-07, "loss": 0.0, "num_input_tokens_seen": 74233920, "step": 38570 }, { "epoch": 6.292846072273432, "grad_norm": 0.10662376880645752, "learning_rate": 1.5662995199469954e-07, "loss": 0.0007, "num_input_tokens_seen": 74243600, "step": 38575 }, { "epoch": 6.293661799494249, "grad_norm": 0.048305220901966095, "learning_rate": 1.5553462921753802e-07, "loss": 0.0001, "num_input_tokens_seen": 74252464, "step": 38580 }, { "epoch": 6.294477526715067, "grad_norm": 0.00844282004982233, "learning_rate": 1.544431377691502e-07, "loss": 0.0138, "num_input_tokens_seen": 74261504, "step": 38585 }, { "epoch": 6.295293253935884, "grad_norm": 0.007357417605817318, "learning_rate": 1.5335547781785975e-07, "loss": 0.0004, "num_input_tokens_seen": 74271296, "step": 38590 }, { "epoch": 6.296108981156701, "grad_norm": 0.027828512713313103, "learning_rate": 1.5227164953139917e-07, "loss": 0.0001, "num_input_tokens_seen": 74283088, "step": 38595 }, { "epoch": 6.296924708377518, "grad_norm": 0.12729643285274506, "learning_rate": 1.511916530769042e-07, "loss": 0.0022, "num_input_tokens_seen": 74293168, "step": 38600 }, { "epoch": 6.296924708377518, "eval_loss": 0.43207141757011414, "eval_runtime": 80.8812, "eval_samples_per_second": 33.691, "eval_steps_per_second": 16.852, "num_input_tokens_seen": 74293168, "step": 38600 }, { "epoch": 6.297740435598336, "grad_norm": 0.021399831399321556, "learning_rate": 1.5011548862092773e-07, "loss": 0.0, "num_input_tokens_seen": 74302128, "step": 38605 }, { "epoch": 6.298556162819153, "grad_norm": 1.1982929706573486, "learning_rate": 1.490431563294231e-07, "loss": 0.0009, "num_input_tokens_seen": 74312512, "step": 38610 }, { "epoch": 6.2993718900399704, "grad_norm": 0.04214861989021301, "learning_rate": 1.4797465636776365e-07, "loss": 0.0204, "num_input_tokens_seen": 74322560, "step": 38615 }, { "epoch": 6.300187617260788, "grad_norm": 0.14999078214168549, "learning_rate": 1.4690998890072027e-07, "loss": 0.0002, "num_input_tokens_seen": 74333008, "step": 38620 }, { "epoch": 6.301003344481606, "grad_norm": 0.001557877636514604, "learning_rate": 1.4584915409248112e-07, "loss": 0.0, "num_input_tokens_seen": 74341984, "step": 38625 }, { "epoch": 6.301819071702423, "grad_norm": 0.014056497253477573, "learning_rate": 1.4479215210663754e-07, "loss": 0.0, "num_input_tokens_seen": 74351888, "step": 38630 }, { "epoch": 6.30263479892324, "grad_norm": 5.71000337600708, "learning_rate": 1.4373898310619528e-07, "loss": 0.0031, "num_input_tokens_seen": 74360832, "step": 38635 }, { "epoch": 6.303450526144058, "grad_norm": 0.006764468736946583, "learning_rate": 1.4268964725356604e-07, "loss": 0.0009, "num_input_tokens_seen": 74370576, "step": 38640 }, { "epoch": 6.304266253364875, "grad_norm": 0.0010803836630657315, "learning_rate": 1.4164414471056764e-07, "loss": 0.0, "num_input_tokens_seen": 74379120, "step": 38645 }, { "epoch": 6.305081980585692, "grad_norm": 0.010560018010437489, "learning_rate": 1.4060247563843497e-07, "loss": 0.0001, "num_input_tokens_seen": 74388704, "step": 38650 }, { "epoch": 6.305897707806509, "grad_norm": 0.007082704920321703, "learning_rate": 1.3956464019780068e-07, "loss": 0.0, "num_input_tokens_seen": 74397840, "step": 38655 }, { "epoch": 6.306713435027326, "grad_norm": 0.00020454195328056812, "learning_rate": 1.385306385487145e-07, "loss": 0.0002, "num_input_tokens_seen": 74406608, "step": 38660 }, { "epoch": 6.307529162248144, "grad_norm": 0.0009234490571543574, "learning_rate": 1.3750047085063222e-07, "loss": 0.0002, "num_input_tokens_seen": 74416336, "step": 38665 }, { "epoch": 6.3083448894689615, "grad_norm": 0.015144089236855507, "learning_rate": 1.3647413726242119e-07, "loss": 0.0, "num_input_tokens_seen": 74425520, "step": 38670 }, { "epoch": 6.309160616689779, "grad_norm": 0.002970809582620859, "learning_rate": 1.3545163794235205e-07, "loss": 0.0, "num_input_tokens_seen": 74435584, "step": 38675 }, { "epoch": 6.309976343910597, "grad_norm": 0.004480911418795586, "learning_rate": 1.3443297304810698e-07, "loss": 0.0, "num_input_tokens_seen": 74445120, "step": 38680 }, { "epoch": 6.310792071131414, "grad_norm": 0.0016552393790334463, "learning_rate": 1.3341814273677977e-07, "loss": 0.0, "num_input_tokens_seen": 74453584, "step": 38685 }, { "epoch": 6.311607798352231, "grad_norm": 0.0015006755711510777, "learning_rate": 1.324071471648647e-07, "loss": 0.0508, "num_input_tokens_seen": 74463088, "step": 38690 }, { "epoch": 6.312423525573048, "grad_norm": 0.012080833315849304, "learning_rate": 1.3139998648827312e-07, "loss": 0.0001, "num_input_tokens_seen": 74472640, "step": 38695 }, { "epoch": 6.313239252793866, "grad_norm": 0.020458057522773743, "learning_rate": 1.3039666086232526e-07, "loss": 0.0001, "num_input_tokens_seen": 74483376, "step": 38700 }, { "epoch": 6.314054980014683, "grad_norm": 0.005747103597968817, "learning_rate": 1.2939717044174183e-07, "loss": 0.0, "num_input_tokens_seen": 74492144, "step": 38705 }, { "epoch": 6.3148707072355, "grad_norm": 0.0024281542282551527, "learning_rate": 1.284015153806578e-07, "loss": 0.0001, "num_input_tokens_seen": 74499600, "step": 38710 }, { "epoch": 6.3156864344563175, "grad_norm": 0.0003351231571286917, "learning_rate": 1.274096958326171e-07, "loss": 0.0, "num_input_tokens_seen": 74509200, "step": 38715 }, { "epoch": 6.3165021616771355, "grad_norm": 0.0004856616724282503, "learning_rate": 1.2642171195056952e-07, "loss": 0.0, "num_input_tokens_seen": 74517664, "step": 38720 }, { "epoch": 6.317317888897953, "grad_norm": 0.08418016880750656, "learning_rate": 1.2543756388687377e-07, "loss": 0.0002, "num_input_tokens_seen": 74527152, "step": 38725 }, { "epoch": 6.31813361611877, "grad_norm": 0.00023966707522049546, "learning_rate": 1.2445725179330014e-07, "loss": 0.0001, "num_input_tokens_seen": 74537024, "step": 38730 }, { "epoch": 6.318949343339587, "grad_norm": 12.337958335876465, "learning_rate": 1.2348077582102212e-07, "loss": 0.0392, "num_input_tokens_seen": 74546800, "step": 38735 }, { "epoch": 6.319765070560405, "grad_norm": 0.06317112594842911, "learning_rate": 1.2250813612062762e-07, "loss": 0.0002, "num_input_tokens_seen": 74556496, "step": 38740 }, { "epoch": 6.320580797781222, "grad_norm": 0.001131101860664785, "learning_rate": 1.215393328421105e-07, "loss": 0.0001, "num_input_tokens_seen": 74565808, "step": 38745 }, { "epoch": 6.321396525002039, "grad_norm": 0.0005767687107436359, "learning_rate": 1.2057436613486796e-07, "loss": 0.0001, "num_input_tokens_seen": 74574560, "step": 38750 }, { "epoch": 6.322212252222856, "grad_norm": 0.004494583699852228, "learning_rate": 1.1961323614771424e-07, "loss": 0.0002, "num_input_tokens_seen": 74584384, "step": 38755 }, { "epoch": 6.323027979443674, "grad_norm": 0.0006484885234385729, "learning_rate": 1.1865594302886418e-07, "loss": 0.0004, "num_input_tokens_seen": 74593808, "step": 38760 }, { "epoch": 6.3238437066644915, "grad_norm": 0.0016054888255894184, "learning_rate": 1.1770248692594687e-07, "loss": 0.0003, "num_input_tokens_seen": 74602256, "step": 38765 }, { "epoch": 6.324659433885309, "grad_norm": 0.0003994106373284012, "learning_rate": 1.167528679859975e-07, "loss": 0.049, "num_input_tokens_seen": 74611376, "step": 38770 }, { "epoch": 6.325475161106126, "grad_norm": 0.1121457889676094, "learning_rate": 1.1580708635545446e-07, "loss": 0.0002, "num_input_tokens_seen": 74621088, "step": 38775 }, { "epoch": 6.326290888326944, "grad_norm": 0.0004211502964608371, "learning_rate": 1.1486514218017885e-07, "loss": 0.0001, "num_input_tokens_seen": 74631120, "step": 38780 }, { "epoch": 6.327106615547761, "grad_norm": 0.00223659910261631, "learning_rate": 1.1392703560542117e-07, "loss": 0.0, "num_input_tokens_seen": 74640768, "step": 38785 }, { "epoch": 6.327922342768578, "grad_norm": 0.250339150428772, "learning_rate": 1.129927667758518e-07, "loss": 0.0002, "num_input_tokens_seen": 74650480, "step": 38790 }, { "epoch": 6.328738069989395, "grad_norm": 0.0010227874154224992, "learning_rate": 1.1206233583554992e-07, "loss": 0.0, "num_input_tokens_seen": 74660368, "step": 38795 }, { "epoch": 6.329553797210213, "grad_norm": 0.07325878739356995, "learning_rate": 1.1113574292799523e-07, "loss": 0.0001, "num_input_tokens_seen": 74668864, "step": 38800 }, { "epoch": 6.329553797210213, "eval_loss": 0.43222880363464355, "eval_runtime": 80.8361, "eval_samples_per_second": 33.71, "eval_steps_per_second": 16.861, "num_input_tokens_seen": 74668864, "step": 38800 }, { "epoch": 6.33036952443103, "grad_norm": 0.05192551761865616, "learning_rate": 1.1021298819608449e-07, "loss": 0.0001, "num_input_tokens_seen": 74678944, "step": 38805 }, { "epoch": 6.331185251651847, "grad_norm": 0.0005548546905629337, "learning_rate": 1.0929407178211226e-07, "loss": 0.0001, "num_input_tokens_seen": 74688112, "step": 38810 }, { "epoch": 6.332000978872665, "grad_norm": 0.0012906170450150967, "learning_rate": 1.0837899382779293e-07, "loss": 0.0, "num_input_tokens_seen": 74697840, "step": 38815 }, { "epoch": 6.332816706093483, "grad_norm": 0.002072409028187394, "learning_rate": 1.0746775447423862e-07, "loss": 0.0691, "num_input_tokens_seen": 74707040, "step": 38820 }, { "epoch": 6.3336324333143, "grad_norm": 0.22494277358055115, "learning_rate": 1.0656035386197583e-07, "loss": 0.0002, "num_input_tokens_seen": 74717936, "step": 38825 }, { "epoch": 6.334448160535117, "grad_norm": 0.0006147457170300186, "learning_rate": 1.0565679213093982e-07, "loss": 0.0, "num_input_tokens_seen": 74727856, "step": 38830 }, { "epoch": 6.335263887755934, "grad_norm": 0.0012053807731717825, "learning_rate": 1.0475706942046638e-07, "loss": 0.0, "num_input_tokens_seen": 74739008, "step": 38835 }, { "epoch": 6.336079614976752, "grad_norm": 0.02828344888985157, "learning_rate": 1.0386118586930282e-07, "loss": 0.0012, "num_input_tokens_seen": 74748544, "step": 38840 }, { "epoch": 6.336895342197569, "grad_norm": 0.006490842904895544, "learning_rate": 1.0296914161561367e-07, "loss": 0.0002, "num_input_tokens_seen": 74759920, "step": 38845 }, { "epoch": 6.337711069418386, "grad_norm": 0.008045039139688015, "learning_rate": 1.0208093679695552e-07, "loss": 0.0, "num_input_tokens_seen": 74770832, "step": 38850 }, { "epoch": 6.338526796639204, "grad_norm": 0.012326308526098728, "learning_rate": 1.0119657155030493e-07, "loss": 0.0, "num_input_tokens_seen": 74781008, "step": 38855 }, { "epoch": 6.339342523860021, "grad_norm": 0.00018347649893257767, "learning_rate": 1.003160460120417e-07, "loss": 0.0002, "num_input_tokens_seen": 74791392, "step": 38860 }, { "epoch": 6.3401582510808385, "grad_norm": 0.0007413696148432791, "learning_rate": 9.943936031795165e-08, "loss": 0.1084, "num_input_tokens_seen": 74801888, "step": 38865 }, { "epoch": 6.340973978301656, "grad_norm": 0.03414979204535484, "learning_rate": 9.856651460323219e-08, "loss": 0.0001, "num_input_tokens_seen": 74811872, "step": 38870 }, { "epoch": 6.341789705522474, "grad_norm": 0.008377674967050552, "learning_rate": 9.769750900248953e-08, "loss": 0.0002, "num_input_tokens_seen": 74821312, "step": 38875 }, { "epoch": 6.342605432743291, "grad_norm": 0.00023008491552900523, "learning_rate": 9.683234364973038e-08, "loss": 0.0018, "num_input_tokens_seen": 74829152, "step": 38880 }, { "epoch": 6.343421159964108, "grad_norm": 0.008498262614011765, "learning_rate": 9.597101867837854e-08, "loss": 0.0711, "num_input_tokens_seen": 74839424, "step": 38885 }, { "epoch": 6.344236887184925, "grad_norm": 0.019187068566679955, "learning_rate": 9.511353422125835e-08, "loss": 0.0, "num_input_tokens_seen": 74848800, "step": 38890 }, { "epoch": 6.345052614405743, "grad_norm": 0.02537946216762066, "learning_rate": 9.42598904106029e-08, "loss": 0.0001, "num_input_tokens_seen": 74859744, "step": 38895 }, { "epoch": 6.34586834162656, "grad_norm": 0.06825371086597443, "learning_rate": 9.341008737806245e-08, "loss": 0.0001, "num_input_tokens_seen": 74869168, "step": 38900 }, { "epoch": 6.346684068847377, "grad_norm": 0.010381617583334446, "learning_rate": 9.256412525467661e-08, "loss": 0.0001, "num_input_tokens_seen": 74879104, "step": 38905 }, { "epoch": 6.3474997960681945, "grad_norm": 0.013218936510384083, "learning_rate": 9.172200417091326e-08, "loss": 0.0001, "num_input_tokens_seen": 74889440, "step": 38910 }, { "epoch": 6.3483155232890125, "grad_norm": 0.002103060483932495, "learning_rate": 9.088372425663239e-08, "loss": 0.015, "num_input_tokens_seen": 74898160, "step": 38915 }, { "epoch": 6.34913125050983, "grad_norm": 0.007399910595268011, "learning_rate": 9.004928564110837e-08, "loss": 0.0998, "num_input_tokens_seen": 74907456, "step": 38920 }, { "epoch": 6.349946977730647, "grad_norm": 0.00274618505500257, "learning_rate": 8.92186884530244e-08, "loss": 0.0001, "num_input_tokens_seen": 74916752, "step": 38925 }, { "epoch": 6.350762704951464, "grad_norm": 0.020511247217655182, "learning_rate": 8.83919328204641e-08, "loss": 0.0, "num_input_tokens_seen": 74926864, "step": 38930 }, { "epoch": 6.351578432172282, "grad_norm": 0.0051707117818295956, "learning_rate": 8.756901887093105e-08, "loss": 0.0098, "num_input_tokens_seen": 74936976, "step": 38935 }, { "epoch": 6.352394159393099, "grad_norm": 0.0021117492578923702, "learning_rate": 8.674994673132098e-08, "loss": 0.0001, "num_input_tokens_seen": 74947168, "step": 38940 }, { "epoch": 6.353209886613916, "grad_norm": 0.00014916194777470082, "learning_rate": 8.593471652794949e-08, "loss": 0.0, "num_input_tokens_seen": 74957056, "step": 38945 }, { "epoch": 6.354025613834733, "grad_norm": 11.474272727966309, "learning_rate": 8.512332838653548e-08, "loss": 0.0285, "num_input_tokens_seen": 74965824, "step": 38950 }, { "epoch": 6.354841341055551, "grad_norm": 0.09001975506544113, "learning_rate": 8.431578243220106e-08, "loss": 0.0001, "num_input_tokens_seen": 74974528, "step": 38955 }, { "epoch": 6.3556570682763684, "grad_norm": 0.056070778518915176, "learning_rate": 8.351207878948552e-08, "loss": 0.065, "num_input_tokens_seen": 74983248, "step": 38960 }, { "epoch": 6.356472795497186, "grad_norm": 0.20149976015090942, "learning_rate": 8.271221758232583e-08, "loss": 0.0004, "num_input_tokens_seen": 74993056, "step": 38965 }, { "epoch": 6.357288522718003, "grad_norm": 0.015607823617756367, "learning_rate": 8.191619893407332e-08, "loss": 0.0001, "num_input_tokens_seen": 75003616, "step": 38970 }, { "epoch": 6.358104249938821, "grad_norm": 0.0007378182490356266, "learning_rate": 8.112402296748534e-08, "loss": 0.0, "num_input_tokens_seen": 75012576, "step": 38975 }, { "epoch": 6.358919977159638, "grad_norm": 0.07907837629318237, "learning_rate": 8.033568980471973e-08, "loss": 0.0002, "num_input_tokens_seen": 75020416, "step": 38980 }, { "epoch": 6.359735704380455, "grad_norm": 0.009372622705996037, "learning_rate": 7.955119956735146e-08, "loss": 0.001, "num_input_tokens_seen": 75030560, "step": 38985 }, { "epoch": 6.360551431601272, "grad_norm": 0.005973889492452145, "learning_rate": 7.877055237636155e-08, "loss": 0.0007, "num_input_tokens_seen": 75040880, "step": 38990 }, { "epoch": 6.36136715882209, "grad_norm": 0.0005555851384997368, "learning_rate": 7.79937483521287e-08, "loss": 0.0, "num_input_tokens_seen": 75050464, "step": 38995 }, { "epoch": 6.362182886042907, "grad_norm": 0.0020480486564338207, "learning_rate": 7.722078761444873e-08, "loss": 0.0815, "num_input_tokens_seen": 75058640, "step": 39000 }, { "epoch": 6.362182886042907, "eval_loss": 0.43038803339004517, "eval_runtime": 80.8962, "eval_samples_per_second": 33.685, "eval_steps_per_second": 16.849, "num_input_tokens_seen": 75058640, "step": 39000 }, { "epoch": 6.362998613263724, "grad_norm": 0.0021881572902202606, "learning_rate": 7.645167028252631e-08, "loss": 0.0001, "num_input_tokens_seen": 75068624, "step": 39005 }, { "epoch": 6.3638143404845415, "grad_norm": 0.0004536547348834574, "learning_rate": 7.568639647496379e-08, "loss": 0.0139, "num_input_tokens_seen": 75078016, "step": 39010 }, { "epoch": 6.3646300677053596, "grad_norm": 0.00515685835853219, "learning_rate": 7.492496630977508e-08, "loss": 0.0, "num_input_tokens_seen": 75088048, "step": 39015 }, { "epoch": 6.365445794926177, "grad_norm": 0.0002489391481503844, "learning_rate": 7.416737990438571e-08, "loss": 0.0001, "num_input_tokens_seen": 75098080, "step": 39020 }, { "epoch": 6.366261522146994, "grad_norm": 0.03279557451605797, "learning_rate": 7.341363737562445e-08, "loss": 0.0671, "num_input_tokens_seen": 75106912, "step": 39025 }, { "epoch": 6.367077249367812, "grad_norm": 0.0017679744632914662, "learning_rate": 7.266373883972887e-08, "loss": 0.0, "num_input_tokens_seen": 75116416, "step": 39030 }, { "epoch": 6.367892976588629, "grad_norm": 0.00030085915932431817, "learning_rate": 7.191768441233981e-08, "loss": 0.0, "num_input_tokens_seen": 75126352, "step": 39035 }, { "epoch": 6.368708703809446, "grad_norm": 0.013345484621822834, "learning_rate": 7.11754742085069e-08, "loss": 0.0002, "num_input_tokens_seen": 75134768, "step": 39040 }, { "epoch": 6.369524431030263, "grad_norm": 0.05360725522041321, "learning_rate": 7.043710834269413e-08, "loss": 0.0001, "num_input_tokens_seen": 75144512, "step": 39045 }, { "epoch": 6.370340158251081, "grad_norm": 0.012186066247522831, "learning_rate": 6.970258692876319e-08, "loss": 0.0096, "num_input_tokens_seen": 75152752, "step": 39050 }, { "epoch": 6.371155885471898, "grad_norm": 0.0006762922857888043, "learning_rate": 6.897191007998738e-08, "loss": 0.0, "num_input_tokens_seen": 75162736, "step": 39055 }, { "epoch": 6.3719716126927155, "grad_norm": 0.11676666885614395, "learning_rate": 6.824507790904599e-08, "loss": 0.0002, "num_input_tokens_seen": 75172544, "step": 39060 }, { "epoch": 6.372787339913533, "grad_norm": 0.0032813926227390766, "learning_rate": 6.752209052802439e-08, "loss": 0.0001, "num_input_tokens_seen": 75182800, "step": 39065 }, { "epoch": 6.373603067134351, "grad_norm": 0.005844974424690008, "learning_rate": 6.680294804841946e-08, "loss": 0.0001, "num_input_tokens_seen": 75192144, "step": 39070 }, { "epoch": 6.374418794355168, "grad_norm": 0.00034117489121854305, "learning_rate": 6.608765058112865e-08, "loss": 0.0, "num_input_tokens_seen": 75201312, "step": 39075 }, { "epoch": 6.375234521575985, "grad_norm": 0.029377294704318047, "learning_rate": 6.537619823646368e-08, "loss": 0.0, "num_input_tokens_seen": 75211440, "step": 39080 }, { "epoch": 6.376050248796802, "grad_norm": 0.0036276967730373144, "learning_rate": 6.466859112413404e-08, "loss": 0.0, "num_input_tokens_seen": 75220704, "step": 39085 }, { "epoch": 6.37686597601762, "grad_norm": 0.005256518721580505, "learning_rate": 6.39648293532663e-08, "loss": 0.0, "num_input_tokens_seen": 75230368, "step": 39090 }, { "epoch": 6.377681703238437, "grad_norm": 0.04324124753475189, "learning_rate": 6.32649130323848e-08, "loss": 0.0005, "num_input_tokens_seen": 75238832, "step": 39095 }, { "epoch": 6.378497430459254, "grad_norm": 7.187277515185997e-05, "learning_rate": 6.256884226943094e-08, "loss": 0.0, "num_input_tokens_seen": 75247280, "step": 39100 }, { "epoch": 6.3793131576800715, "grad_norm": 0.0005827744607813656, "learning_rate": 6.187661717174386e-08, "loss": 0.0, "num_input_tokens_seen": 75257568, "step": 39105 }, { "epoch": 6.3801288849008895, "grad_norm": 0.041603896766901016, "learning_rate": 6.118823784607708e-08, "loss": 0.0, "num_input_tokens_seen": 75266160, "step": 39110 }, { "epoch": 6.380944612121707, "grad_norm": 0.0008327624527737498, "learning_rate": 6.050370439858178e-08, "loss": 0.0, "num_input_tokens_seen": 75275376, "step": 39115 }, { "epoch": 6.381760339342524, "grad_norm": 0.003377774031832814, "learning_rate": 5.98230169348235e-08, "loss": 0.0001, "num_input_tokens_seen": 75284880, "step": 39120 }, { "epoch": 6.382576066563341, "grad_norm": 0.012812762521207333, "learning_rate": 5.914617555977664e-08, "loss": 0.0003, "num_input_tokens_seen": 75294544, "step": 39125 }, { "epoch": 6.383391793784159, "grad_norm": 0.016691578552126884, "learning_rate": 5.8473180377816017e-08, "loss": 0.0001, "num_input_tokens_seen": 75305056, "step": 39130 }, { "epoch": 6.384207521004976, "grad_norm": 0.0002118917036568746, "learning_rate": 5.780403149272251e-08, "loss": 0.0001, "num_input_tokens_seen": 75315792, "step": 39135 }, { "epoch": 6.385023248225793, "grad_norm": 0.0010576039785519242, "learning_rate": 5.7138729007694126e-08, "loss": 0.0001, "num_input_tokens_seen": 75325824, "step": 39140 }, { "epoch": 6.38583897544661, "grad_norm": 0.0007199023384600878, "learning_rate": 5.64772730253238e-08, "loss": 0.0001, "num_input_tokens_seen": 75336112, "step": 39145 }, { "epoch": 6.386654702667428, "grad_norm": 0.028712818399071693, "learning_rate": 5.5819663647618814e-08, "loss": 0.0001, "num_input_tokens_seen": 75345536, "step": 39150 }, { "epoch": 6.387470429888245, "grad_norm": 2.0710000171675347e-05, "learning_rate": 5.5165900975989723e-08, "loss": 0.0009, "num_input_tokens_seen": 75355232, "step": 39155 }, { "epoch": 6.388286157109063, "grad_norm": 18.163393020629883, "learning_rate": 5.451598511125311e-08, "loss": 0.0014, "num_input_tokens_seen": 75364368, "step": 39160 }, { "epoch": 6.38910188432988, "grad_norm": 0.019846059381961823, "learning_rate": 5.3869916153637124e-08, "loss": 0.0002, "num_input_tokens_seen": 75374464, "step": 39165 }, { "epoch": 6.389917611550698, "grad_norm": 0.0012517140712589025, "learning_rate": 5.322769420277318e-08, "loss": 0.0, "num_input_tokens_seen": 75384224, "step": 39170 }, { "epoch": 6.390733338771515, "grad_norm": 0.004302258137613535, "learning_rate": 5.258931935769873e-08, "loss": 0.0001, "num_input_tokens_seen": 75394144, "step": 39175 }, { "epoch": 6.391549065992332, "grad_norm": 0.006362768821418285, "learning_rate": 5.19547917168628e-08, "loss": 0.0, "num_input_tokens_seen": 75403552, "step": 39180 }, { "epoch": 6.392364793213149, "grad_norm": 0.3587648272514343, "learning_rate": 5.13241113781121e-08, "loss": 0.0004, "num_input_tokens_seen": 75413424, "step": 39185 }, { "epoch": 6.393180520433967, "grad_norm": 0.004127536900341511, "learning_rate": 5.0697278438707755e-08, "loss": 0.0002, "num_input_tokens_seen": 75421776, "step": 39190 }, { "epoch": 6.393996247654784, "grad_norm": 0.0029909999575465918, "learning_rate": 5.0074292995316854e-08, "loss": 0.0691, "num_input_tokens_seen": 75431552, "step": 39195 }, { "epoch": 6.394811974875601, "grad_norm": 0.012615897692739964, "learning_rate": 4.945515514400978e-08, "loss": 0.0, "num_input_tokens_seen": 75440784, "step": 39200 }, { "epoch": 6.394811974875601, "eval_loss": 0.4322725534439087, "eval_runtime": 80.8184, "eval_samples_per_second": 33.718, "eval_steps_per_second": 16.865, "num_input_tokens_seen": 75440784, "step": 39200 }, { "epoch": 6.395627702096419, "grad_norm": 15.903099060058594, "learning_rate": 4.883986498026571e-08, "loss": 0.0691, "num_input_tokens_seen": 75449888, "step": 39205 }, { "epoch": 6.3964434293172365, "grad_norm": 0.20059248805046082, "learning_rate": 4.822842259896987e-08, "loss": 0.0005, "num_input_tokens_seen": 75459856, "step": 39210 }, { "epoch": 6.397259156538054, "grad_norm": 5.773574230261147e-05, "learning_rate": 4.762082809441626e-08, "loss": 0.0, "num_input_tokens_seen": 75469872, "step": 39215 }, { "epoch": 6.398074883758871, "grad_norm": 0.0006738292868249118, "learning_rate": 4.7017081560302156e-08, "loss": 0.0, "num_input_tokens_seen": 75479680, "step": 39220 }, { "epoch": 6.398890610979688, "grad_norm": 0.0016268420731648803, "learning_rate": 4.6417183089730866e-08, "loss": 0.0008, "num_input_tokens_seen": 75488208, "step": 39225 }, { "epoch": 6.399706338200506, "grad_norm": 0.002153501147404313, "learning_rate": 4.5821132775217265e-08, "loss": 0.0001, "num_input_tokens_seen": 75496928, "step": 39230 }, { "epoch": 6.400522065421323, "grad_norm": 0.0041078124195337296, "learning_rate": 4.5228930708679504e-08, "loss": 0.0, "num_input_tokens_seen": 75505184, "step": 39235 }, { "epoch": 6.40133779264214, "grad_norm": 0.00017611452494747937, "learning_rate": 4.464057698144175e-08, "loss": 0.0, "num_input_tokens_seen": 75514464, "step": 39240 }, { "epoch": 6.402153519862958, "grad_norm": 0.000216047847061418, "learning_rate": 4.4056071684236974e-08, "loss": 0.0001, "num_input_tokens_seen": 75523584, "step": 39245 }, { "epoch": 6.402969247083775, "grad_norm": 0.019391216337680817, "learning_rate": 4.347541490719864e-08, "loss": 0.0773, "num_input_tokens_seen": 75533280, "step": 39250 }, { "epoch": 6.4037849743045925, "grad_norm": 0.0021015095990151167, "learning_rate": 4.2898606739877336e-08, "loss": 0.0001, "num_input_tokens_seen": 75543536, "step": 39255 }, { "epoch": 6.40460070152541, "grad_norm": 0.0029628605116158724, "learning_rate": 4.232564727122135e-08, "loss": 0.0, "num_input_tokens_seen": 75552992, "step": 39260 }, { "epoch": 6.405416428746228, "grad_norm": 0.014030426740646362, "learning_rate": 4.1756536589585004e-08, "loss": 0.0018, "num_input_tokens_seen": 75562112, "step": 39265 }, { "epoch": 6.406232155967045, "grad_norm": 0.02493157982826233, "learning_rate": 4.119127478273976e-08, "loss": 0.0001, "num_input_tokens_seen": 75571840, "step": 39270 }, { "epoch": 6.407047883187862, "grad_norm": 6.063130058464594e-05, "learning_rate": 4.062986193784923e-08, "loss": 0.0, "num_input_tokens_seen": 75580048, "step": 39275 }, { "epoch": 6.407863610408679, "grad_norm": 0.0025714049115777016, "learning_rate": 4.007229814149416e-08, "loss": 0.0, "num_input_tokens_seen": 75589552, "step": 39280 }, { "epoch": 6.408679337629497, "grad_norm": 0.026193276047706604, "learning_rate": 3.951858347965576e-08, "loss": 0.1048, "num_input_tokens_seen": 75597440, "step": 39285 }, { "epoch": 6.409495064850314, "grad_norm": 0.010306963697075844, "learning_rate": 3.896871803772684e-08, "loss": 0.0, "num_input_tokens_seen": 75607232, "step": 39290 }, { "epoch": 6.410310792071131, "grad_norm": 0.10911701619625092, "learning_rate": 3.842270190050068e-08, "loss": 0.0001, "num_input_tokens_seen": 75616464, "step": 39295 }, { "epoch": 6.411126519291948, "grad_norm": 0.0112397326156497, "learning_rate": 3.7880535152179376e-08, "loss": 0.1314, "num_input_tokens_seen": 75625520, "step": 39300 }, { "epoch": 6.4119422465127665, "grad_norm": 0.00023864346439950168, "learning_rate": 3.734221787637382e-08, "loss": 0.0001, "num_input_tokens_seen": 75634480, "step": 39305 }, { "epoch": 6.412757973733584, "grad_norm": 0.005558731500059366, "learning_rate": 3.680775015609817e-08, "loss": 0.0001, "num_input_tokens_seen": 75645504, "step": 39310 }, { "epoch": 6.413573700954401, "grad_norm": 0.47004592418670654, "learning_rate": 3.627713207377537e-08, "loss": 0.0004, "num_input_tokens_seen": 75654736, "step": 39315 }, { "epoch": 6.414389428175218, "grad_norm": 0.0003401571011636406, "learning_rate": 3.575036371123164e-08, "loss": 0.0006, "num_input_tokens_seen": 75665584, "step": 39320 }, { "epoch": 6.415205155396036, "grad_norm": 0.001447474118322134, "learning_rate": 3.5227445149704776e-08, "loss": 0.0002, "num_input_tokens_seen": 75674896, "step": 39325 }, { "epoch": 6.416020882616853, "grad_norm": 0.07769839465618134, "learning_rate": 3.470837646983027e-08, "loss": 0.0262, "num_input_tokens_seen": 75684992, "step": 39330 }, { "epoch": 6.41683660983767, "grad_norm": 0.1606525182723999, "learning_rate": 3.419315775165799e-08, "loss": 0.0002, "num_input_tokens_seen": 75694816, "step": 39335 }, { "epoch": 6.417652337058487, "grad_norm": 0.004300068598240614, "learning_rate": 3.368178907464103e-08, "loss": 0.0, "num_input_tokens_seen": 75705248, "step": 39340 }, { "epoch": 6.418468064279305, "grad_norm": 0.011182623915374279, "learning_rate": 3.317427051763855e-08, "loss": 0.0002, "num_input_tokens_seen": 75714608, "step": 39345 }, { "epoch": 6.419283791500122, "grad_norm": 0.003666088916361332, "learning_rate": 3.267060215891571e-08, "loss": 0.0001, "num_input_tokens_seen": 75725152, "step": 39350 }, { "epoch": 6.4200995187209395, "grad_norm": 0.00047670636558905244, "learning_rate": 3.217078407614649e-08, "loss": 0.0413, "num_input_tokens_seen": 75733744, "step": 39355 }, { "epoch": 6.420915245941757, "grad_norm": 0.053861938416957855, "learning_rate": 3.1674816346405345e-08, "loss": 0.0004, "num_input_tokens_seen": 75742560, "step": 39360 }, { "epoch": 6.421730973162575, "grad_norm": 0.008911073207855225, "learning_rate": 3.11826990461811e-08, "loss": 0.0, "num_input_tokens_seen": 75753744, "step": 39365 }, { "epoch": 6.422546700383392, "grad_norm": 0.012877212837338448, "learning_rate": 3.069443225136304e-08, "loss": 0.0001, "num_input_tokens_seen": 75764064, "step": 39370 }, { "epoch": 6.423362427604209, "grad_norm": 0.13550816476345062, "learning_rate": 3.021001603724372e-08, "loss": 0.0002, "num_input_tokens_seen": 75774416, "step": 39375 }, { "epoch": 6.424178154825027, "grad_norm": 0.01373336836695671, "learning_rate": 2.9729450478532818e-08, "loss": 0.0, "num_input_tokens_seen": 75784112, "step": 39380 }, { "epoch": 6.424993882045844, "grad_norm": 0.0031185182742774487, "learning_rate": 2.9252735649337726e-08, "loss": 0.0001, "num_input_tokens_seen": 75793904, "step": 39385 }, { "epoch": 6.425809609266661, "grad_norm": 0.0065710521303117275, "learning_rate": 2.8779871623171863e-08, "loss": 0.0, "num_input_tokens_seen": 75803088, "step": 39390 }, { "epoch": 6.426625336487478, "grad_norm": 0.00012323036207817495, "learning_rate": 2.8310858472957448e-08, "loss": 0.0, "num_input_tokens_seen": 75812512, "step": 39395 }, { "epoch": 6.4274410637082955, "grad_norm": 1.2371742725372314, "learning_rate": 2.784569627101996e-08, "loss": 0.0003, "num_input_tokens_seen": 75822528, "step": 39400 }, { "epoch": 6.4274410637082955, "eval_loss": 0.43134525418281555, "eval_runtime": 80.8128, "eval_samples_per_second": 33.72, "eval_steps_per_second": 16.866, "num_input_tokens_seen": 75822528, "step": 39400 }, { "epoch": 6.4282567909291135, "grad_norm": 0.05749901756644249, "learning_rate": 2.738438508909924e-08, "loss": 0.0001, "num_input_tokens_seen": 75832176, "step": 39405 }, { "epoch": 6.429072518149931, "grad_norm": 0.0005124294548295438, "learning_rate": 2.692692499833005e-08, "loss": 0.069, "num_input_tokens_seen": 75842256, "step": 39410 }, { "epoch": 6.429888245370748, "grad_norm": 0.6346259117126465, "learning_rate": 2.647331606926151e-08, "loss": 0.0619, "num_input_tokens_seen": 75851920, "step": 39415 }, { "epoch": 6.430703972591566, "grad_norm": 0.00014135574747342616, "learning_rate": 2.6023558371843225e-08, "loss": 0.0001, "num_input_tokens_seen": 75860928, "step": 39420 }, { "epoch": 6.431519699812383, "grad_norm": 0.006119697820395231, "learning_rate": 2.557765197543638e-08, "loss": 0.0, "num_input_tokens_seen": 75871200, "step": 39425 }, { "epoch": 6.4323354270332, "grad_norm": 0.016580328345298767, "learning_rate": 2.513559694880263e-08, "loss": 0.0001, "num_input_tokens_seen": 75879296, "step": 39430 }, { "epoch": 6.433151154254017, "grad_norm": 0.006659103091806173, "learning_rate": 2.469739336011523e-08, "loss": 0.0001, "num_input_tokens_seen": 75887344, "step": 39435 }, { "epoch": 6.433966881474835, "grad_norm": 0.001438735518604517, "learning_rate": 2.4263041276947894e-08, "loss": 0.0, "num_input_tokens_seen": 75895312, "step": 39440 }, { "epoch": 6.434782608695652, "grad_norm": 0.016210846602916718, "learning_rate": 2.3832540766283164e-08, "loss": 0.0214, "num_input_tokens_seen": 75904752, "step": 39445 }, { "epoch": 6.4355983359164695, "grad_norm": 0.0051102046854794025, "learning_rate": 2.3405891894512366e-08, "loss": 0.0003, "num_input_tokens_seen": 75912960, "step": 39450 }, { "epoch": 6.436414063137287, "grad_norm": 0.002722666598856449, "learning_rate": 2.29830947274301e-08, "loss": 0.0, "num_input_tokens_seen": 75923440, "step": 39455 }, { "epoch": 6.437229790358105, "grad_norm": 0.013269281946122646, "learning_rate": 2.2564149330231432e-08, "loss": 0.0001, "num_input_tokens_seen": 75934720, "step": 39460 }, { "epoch": 6.438045517578922, "grad_norm": 0.010242369957268238, "learning_rate": 2.2149055767528572e-08, "loss": 0.0097, "num_input_tokens_seen": 75943648, "step": 39465 }, { "epoch": 6.438861244799739, "grad_norm": 0.0032678861171007156, "learning_rate": 2.1737814103334197e-08, "loss": 0.0001, "num_input_tokens_seen": 75952736, "step": 39470 }, { "epoch": 6.439676972020556, "grad_norm": 0.003274139016866684, "learning_rate": 2.1330424401064253e-08, "loss": 0.0, "num_input_tokens_seen": 75961376, "step": 39475 }, { "epoch": 6.440492699241374, "grad_norm": 0.049566201865673065, "learning_rate": 2.092688672354348e-08, "loss": 0.0001, "num_input_tokens_seen": 75971040, "step": 39480 }, { "epoch": 6.441308426462191, "grad_norm": 0.0002966089523397386, "learning_rate": 2.0527201133005435e-08, "loss": 0.0001, "num_input_tokens_seen": 75979136, "step": 39485 }, { "epoch": 6.442124153683008, "grad_norm": 0.019986407831311226, "learning_rate": 2.0131367691084148e-08, "loss": 0.0001, "num_input_tokens_seen": 75988176, "step": 39490 }, { "epoch": 6.442939880903825, "grad_norm": 0.014278525486588478, "learning_rate": 1.9739386458819675e-08, "loss": 0.0001, "num_input_tokens_seen": 75997360, "step": 39495 }, { "epoch": 6.443755608124643, "grad_norm": 0.1277991533279419, "learning_rate": 1.9351257496666442e-08, "loss": 0.0004, "num_input_tokens_seen": 76007952, "step": 39500 }, { "epoch": 6.444571335345461, "grad_norm": 0.32554343342781067, "learning_rate": 1.896698086447657e-08, "loss": 0.0003, "num_input_tokens_seen": 76019344, "step": 39505 }, { "epoch": 6.445387062566278, "grad_norm": 0.005455636885017157, "learning_rate": 1.8586556621505436e-08, "loss": 0.0, "num_input_tokens_seen": 76027248, "step": 39510 }, { "epoch": 6.446202789787095, "grad_norm": 0.004166078753769398, "learning_rate": 1.820998482642833e-08, "loss": 0.0205, "num_input_tokens_seen": 76037808, "step": 39515 }, { "epoch": 6.447018517007913, "grad_norm": 0.0011211010860279202, "learning_rate": 1.7837265537309912e-08, "loss": 0.0001, "num_input_tokens_seen": 76046928, "step": 39520 }, { "epoch": 6.44783424422873, "grad_norm": 0.0008550810744054615, "learning_rate": 1.7468398811629206e-08, "loss": 0.0, "num_input_tokens_seen": 76056928, "step": 39525 }, { "epoch": 6.448649971449547, "grad_norm": 1.9580775499343872, "learning_rate": 1.710338470627404e-08, "loss": 0.1476, "num_input_tokens_seen": 76066656, "step": 39530 }, { "epoch": 6.449465698670364, "grad_norm": 0.0001763892942108214, "learning_rate": 1.6742223277529945e-08, "loss": 0.0, "num_input_tokens_seen": 76076256, "step": 39535 }, { "epoch": 6.450281425891182, "grad_norm": 0.00456434627994895, "learning_rate": 1.6384914581094036e-08, "loss": 0.0001, "num_input_tokens_seen": 76083840, "step": 39540 }, { "epoch": 6.451097153111999, "grad_norm": 0.0030059944838285446, "learning_rate": 1.6031458672069455e-08, "loss": 0.0, "num_input_tokens_seen": 76093440, "step": 39545 }, { "epoch": 6.4519128803328165, "grad_norm": 0.001377565204165876, "learning_rate": 1.5681855604962602e-08, "loss": 0.0001, "num_input_tokens_seen": 76103008, "step": 39550 }, { "epoch": 6.4527286075536345, "grad_norm": 0.1083531379699707, "learning_rate": 1.5336105433683135e-08, "loss": 0.0003, "num_input_tokens_seen": 76111200, "step": 39555 }, { "epoch": 6.453544334774452, "grad_norm": 0.01929358020424843, "learning_rate": 1.499420821155506e-08, "loss": 0.0549, "num_input_tokens_seen": 76120384, "step": 39560 }, { "epoch": 6.454360061995269, "grad_norm": 0.03612871095538139, "learning_rate": 1.4656163991302874e-08, "loss": 0.0002, "num_input_tokens_seen": 76130464, "step": 39565 }, { "epoch": 6.455175789216086, "grad_norm": 0.3573758602142334, "learning_rate": 1.4321972825051544e-08, "loss": 0.0002, "num_input_tokens_seen": 76140464, "step": 39570 }, { "epoch": 6.455991516436903, "grad_norm": 0.005629109218716621, "learning_rate": 1.3991634764345951e-08, "loss": 0.0, "num_input_tokens_seen": 76150736, "step": 39575 }, { "epoch": 6.456807243657721, "grad_norm": 0.0056963954120874405, "learning_rate": 1.3665149860120352e-08, "loss": 0.0, "num_input_tokens_seen": 76160512, "step": 39580 }, { "epoch": 6.457622970878538, "grad_norm": 0.0007689961930736899, "learning_rate": 1.3342518162728912e-08, "loss": 0.0, "num_input_tokens_seen": 76169312, "step": 39585 }, { "epoch": 6.458438698099355, "grad_norm": 0.0008042078698053956, "learning_rate": 1.30237397219235e-08, "loss": 0.0, "num_input_tokens_seen": 76179264, "step": 39590 }, { "epoch": 6.459254425320173, "grad_norm": 0.2355177402496338, "learning_rate": 1.2708814586862016e-08, "loss": 0.024, "num_input_tokens_seen": 76188512, "step": 39595 }, { "epoch": 6.4600701525409905, "grad_norm": 0.017542066052556038, "learning_rate": 1.2397742806111168e-08, "loss": 0.0001, "num_input_tokens_seen": 76198368, "step": 39600 }, { "epoch": 6.4600701525409905, "eval_loss": 0.43227317929267883, "eval_runtime": 81.0114, "eval_samples_per_second": 33.637, "eval_steps_per_second": 16.825, "num_input_tokens_seen": 76198368, "step": 39600 }, { "epoch": 6.460885879761808, "grad_norm": 0.015693770721554756, "learning_rate": 1.209052442764369e-08, "loss": 0.0001, "num_input_tokens_seen": 76209312, "step": 39605 }, { "epoch": 6.461701606982625, "grad_norm": 0.0035692632663995028, "learning_rate": 1.17871594988328e-08, "loss": 0.0162, "num_input_tokens_seen": 76218528, "step": 39610 }, { "epoch": 6.462517334203443, "grad_norm": 0.0003850652137771249, "learning_rate": 1.1487648066466072e-08, "loss": 0.0028, "num_input_tokens_seen": 76229056, "step": 39615 }, { "epoch": 6.46333306142426, "grad_norm": 0.004574996419250965, "learning_rate": 1.1191990176728784e-08, "loss": 0.0001, "num_input_tokens_seen": 76238912, "step": 39620 }, { "epoch": 6.464148788645077, "grad_norm": 0.0008568131597712636, "learning_rate": 1.0900185875215018e-08, "loss": 0.0002, "num_input_tokens_seen": 76249632, "step": 39625 }, { "epoch": 6.464964515865894, "grad_norm": 0.003104947740212083, "learning_rate": 1.0612235206924891e-08, "loss": 0.0002, "num_input_tokens_seen": 76258144, "step": 39630 }, { "epoch": 6.465780243086712, "grad_norm": 0.0003952247789129615, "learning_rate": 1.0328138216264549e-08, "loss": 0.0001, "num_input_tokens_seen": 76267360, "step": 39635 }, { "epoch": 6.466595970307529, "grad_norm": 11.972979545593262, "learning_rate": 1.004789494704339e-08, "loss": 0.1334, "num_input_tokens_seen": 76276000, "step": 39640 }, { "epoch": 6.467411697528346, "grad_norm": 0.0012468339409679174, "learning_rate": 9.771505442482397e-09, "loss": 0.055, "num_input_tokens_seen": 76285744, "step": 39645 }, { "epoch": 6.468227424749164, "grad_norm": 0.01752065308392048, "learning_rate": 9.498969745200259e-09, "loss": 0.0001, "num_input_tokens_seen": 76295280, "step": 39650 }, { "epoch": 6.469043151969982, "grad_norm": 0.004381328821182251, "learning_rate": 9.230287897230017e-09, "loss": 0.0006, "num_input_tokens_seen": 76304304, "step": 39655 }, { "epoch": 6.469858879190799, "grad_norm": 0.011510418727993965, "learning_rate": 8.965459940002419e-09, "loss": 0.0002, "num_input_tokens_seen": 76313600, "step": 39660 }, { "epoch": 6.470674606411616, "grad_norm": 0.00038195119122974575, "learning_rate": 8.704485914357019e-09, "loss": 0.0001, "num_input_tokens_seen": 76322496, "step": 39665 }, { "epoch": 6.471490333632433, "grad_norm": 0.015948470681905746, "learning_rate": 8.447365860539402e-09, "loss": 0.0001, "num_input_tokens_seen": 76331520, "step": 39670 }, { "epoch": 6.472306060853251, "grad_norm": 0.04745955020189285, "learning_rate": 8.194099818201184e-09, "loss": 0.0001, "num_input_tokens_seen": 76341344, "step": 39675 }, { "epoch": 6.473121788074068, "grad_norm": 0.0005192036041989923, "learning_rate": 7.944687826400011e-09, "loss": 0.0793, "num_input_tokens_seen": 76351664, "step": 39680 }, { "epoch": 6.473937515294885, "grad_norm": 0.08478375524282455, "learning_rate": 7.699129923599557e-09, "loss": 0.0002, "num_input_tokens_seen": 76360352, "step": 39685 }, { "epoch": 6.474753242515702, "grad_norm": 0.000786377873737365, "learning_rate": 7.457426147663982e-09, "loss": 0.0, "num_input_tokens_seen": 76370240, "step": 39690 }, { "epoch": 6.47556896973652, "grad_norm": 0.0012536863796412945, "learning_rate": 7.219576535871797e-09, "loss": 0.0002, "num_input_tokens_seen": 76379024, "step": 39695 }, { "epoch": 6.4763846969573375, "grad_norm": 0.23984064161777496, "learning_rate": 6.985581124896445e-09, "loss": 0.0004, "num_input_tokens_seen": 76388832, "step": 39700 }, { "epoch": 6.477200424178155, "grad_norm": 0.10173244029283524, "learning_rate": 6.755439950828501e-09, "loss": 0.0002, "num_input_tokens_seen": 76398400, "step": 39705 }, { "epoch": 6.478016151398972, "grad_norm": 0.015316109172999859, "learning_rate": 6.5291530491562444e-09, "loss": 0.0, "num_input_tokens_seen": 76408144, "step": 39710 }, { "epoch": 6.47883187861979, "grad_norm": 0.0023654436226934195, "learning_rate": 6.3067204547739845e-09, "loss": 0.0002, "num_input_tokens_seen": 76416624, "step": 39715 }, { "epoch": 6.479647605840607, "grad_norm": 0.0008601440349593759, "learning_rate": 6.088142201987612e-09, "loss": 0.0001, "num_input_tokens_seen": 76426912, "step": 39720 }, { "epoch": 6.480463333061424, "grad_norm": 0.0432768352329731, "learning_rate": 5.873418324503499e-09, "loss": 0.0002, "num_input_tokens_seen": 76437568, "step": 39725 }, { "epoch": 6.481279060282241, "grad_norm": 0.008610392920672894, "learning_rate": 5.6625488554340465e-09, "loss": 0.0, "num_input_tokens_seen": 76447328, "step": 39730 }, { "epoch": 6.482094787503059, "grad_norm": 0.0004242167924530804, "learning_rate": 5.455533827297688e-09, "loss": 0.0, "num_input_tokens_seen": 76457584, "step": 39735 }, { "epoch": 6.482910514723876, "grad_norm": 8.815022468566895, "learning_rate": 5.252373272018885e-09, "loss": 0.0053, "num_input_tokens_seen": 76467152, "step": 39740 }, { "epoch": 6.4837262419446935, "grad_norm": 0.005693368148058653, "learning_rate": 5.053067220925356e-09, "loss": 0.0072, "num_input_tokens_seen": 76478304, "step": 39745 }, { "epoch": 6.484541969165511, "grad_norm": 0.00042055395897477865, "learning_rate": 4.857615704759177e-09, "loss": 0.0, "num_input_tokens_seen": 76486464, "step": 39750 }, { "epoch": 6.485357696386329, "grad_norm": 0.0057978578843176365, "learning_rate": 4.666018753654577e-09, "loss": 0.0001, "num_input_tokens_seen": 76496176, "step": 39755 }, { "epoch": 6.486173423607146, "grad_norm": 0.005742598325014114, "learning_rate": 4.478276397162917e-09, "loss": 0.0, "num_input_tokens_seen": 76505120, "step": 39760 }, { "epoch": 6.486989150827963, "grad_norm": 0.009149398654699326, "learning_rate": 4.294388664233262e-09, "loss": 0.0005, "num_input_tokens_seen": 76513968, "step": 39765 }, { "epoch": 6.487804878048781, "grad_norm": 0.003378194058313966, "learning_rate": 4.114355583223484e-09, "loss": 0.0001, "num_input_tokens_seen": 76522912, "step": 39770 }, { "epoch": 6.488620605269598, "grad_norm": 0.052299823611974716, "learning_rate": 3.9381771818974845e-09, "loss": 0.0035, "num_input_tokens_seen": 76533248, "step": 39775 }, { "epoch": 6.489436332490415, "grad_norm": 0.00012234528549015522, "learning_rate": 3.765853487427973e-09, "loss": 0.0, "num_input_tokens_seen": 76543376, "step": 39780 }, { "epoch": 6.490252059711232, "grad_norm": 0.0015794542850926518, "learning_rate": 3.5973845263825857e-09, "loss": 0.0, "num_input_tokens_seen": 76552096, "step": 39785 }, { "epoch": 6.49106778693205, "grad_norm": 0.005627244710922241, "learning_rate": 3.4327703247488684e-09, "loss": 0.0001, "num_input_tokens_seen": 76560832, "step": 39790 }, { "epoch": 6.4918835141528675, "grad_norm": 0.001920433365739882, "learning_rate": 3.2720109079037443e-09, "loss": 0.0, "num_input_tokens_seen": 76570480, "step": 39795 }, { "epoch": 6.492699241373685, "grad_norm": 8.884011185728014e-05, "learning_rate": 3.1151063006468193e-09, "loss": 0.0001, "num_input_tokens_seen": 76581104, "step": 39800 }, { "epoch": 6.492699241373685, "eval_loss": 0.4304603636264801, "eval_runtime": 80.9385, "eval_samples_per_second": 33.668, "eval_steps_per_second": 16.84, "num_input_tokens_seen": 76581104, "step": 39800 }, { "epoch": 6.493514968594502, "grad_norm": 0.002416192786768079, "learning_rate": 2.962056527169854e-09, "loss": 0.0, "num_input_tokens_seen": 76590896, "step": 39805 }, { "epoch": 6.49433069581532, "grad_norm": 0.00035022362135350704, "learning_rate": 2.8128616110761898e-09, "loss": 0.0109, "num_input_tokens_seen": 76600464, "step": 39810 }, { "epoch": 6.495146423036137, "grad_norm": 0.14222073554992676, "learning_rate": 2.6675215753724223e-09, "loss": 0.0002, "num_input_tokens_seen": 76610864, "step": 39815 }, { "epoch": 6.495962150256954, "grad_norm": 0.0045479666441679, "learning_rate": 2.5260364424739557e-09, "loss": 0.0, "num_input_tokens_seen": 76620688, "step": 39820 }, { "epoch": 6.496777877477771, "grad_norm": 0.008560976013541222, "learning_rate": 2.3884062341994475e-09, "loss": 0.0, "num_input_tokens_seen": 76630128, "step": 39825 }, { "epoch": 6.497593604698589, "grad_norm": 0.04120331630110741, "learning_rate": 2.25463097177081e-09, "loss": 0.0006, "num_input_tokens_seen": 76638128, "step": 39830 }, { "epoch": 6.498409331919406, "grad_norm": 0.005657477770000696, "learning_rate": 2.1247106758215397e-09, "loss": 0.0001, "num_input_tokens_seen": 76646032, "step": 39835 }, { "epoch": 6.499225059140223, "grad_norm": 3.9727066905470565e-05, "learning_rate": 1.998645366382834e-09, "loss": 0.0005, "num_input_tokens_seen": 76656768, "step": 39840 }, { "epoch": 6.5000407863610405, "grad_norm": 0.004963970277458429, "learning_rate": 1.876435062897475e-09, "loss": 0.0002, "num_input_tokens_seen": 76667984, "step": 39845 }, { "epoch": 6.500856513581859, "grad_norm": 0.0004239265399519354, "learning_rate": 1.758079784211497e-09, "loss": 0.0002, "num_input_tokens_seen": 76677392, "step": 39850 }, { "epoch": 6.501672240802676, "grad_norm": 0.0017939461395144463, "learning_rate": 1.6435795485797434e-09, "loss": 0.043, "num_input_tokens_seen": 76687072, "step": 39855 }, { "epoch": 6.502487968023493, "grad_norm": 0.02101399376988411, "learning_rate": 1.5329343736547596e-09, "loss": 0.0001, "num_input_tokens_seen": 76696976, "step": 39860 }, { "epoch": 6.50330369524431, "grad_norm": 0.010752837173640728, "learning_rate": 1.4261442765006739e-09, "loss": 0.0001, "num_input_tokens_seen": 76706512, "step": 39865 }, { "epoch": 6.504119422465128, "grad_norm": 0.0036498054396361113, "learning_rate": 1.3232092735876445e-09, "loss": 0.0, "num_input_tokens_seen": 76716496, "step": 39870 }, { "epoch": 6.504935149685945, "grad_norm": 0.0025586928240954876, "learning_rate": 1.2241293807918607e-09, "loss": 0.0001, "num_input_tokens_seen": 76725744, "step": 39875 }, { "epoch": 6.505750876906762, "grad_norm": 0.0008163147722370923, "learning_rate": 1.128904613387216e-09, "loss": 0.0001, "num_input_tokens_seen": 76736368, "step": 39880 }, { "epoch": 6.506566604127579, "grad_norm": 0.0901479721069336, "learning_rate": 1.0375349860591853e-09, "loss": 0.0002, "num_input_tokens_seen": 76747024, "step": 39885 }, { "epoch": 6.507382331348397, "grad_norm": 0.00039940711576491594, "learning_rate": 9.5002051290205e-10, "loss": 0.0, "num_input_tokens_seen": 76754944, "step": 39890 }, { "epoch": 6.5081980585692145, "grad_norm": 3.3765722037060186e-05, "learning_rate": 8.663612074077954e-10, "loss": 0.0, "num_input_tokens_seen": 76764368, "step": 39895 }, { "epoch": 6.509013785790032, "grad_norm": 0.05935891345143318, "learning_rate": 7.865570824799884e-10, "loss": 0.0001, "num_input_tokens_seen": 76774384, "step": 39900 }, { "epoch": 6.50982951301085, "grad_norm": 0.0003302493132650852, "learning_rate": 7.106081504254514e-10, "loss": 0.0, "num_input_tokens_seen": 76784208, "step": 39905 }, { "epoch": 6.510645240231667, "grad_norm": 0.0007298489799723029, "learning_rate": 6.385144229570372e-10, "loss": 0.0, "num_input_tokens_seen": 76793600, "step": 39910 }, { "epoch": 6.511460967452484, "grad_norm": 0.00030683496152050793, "learning_rate": 5.70275911190854e-10, "loss": 0.0138, "num_input_tokens_seen": 76802192, "step": 39915 }, { "epoch": 6.512276694673301, "grad_norm": 0.006773993372917175, "learning_rate": 5.058926256490403e-10, "loss": 0.0001, "num_input_tokens_seen": 76811264, "step": 39920 }, { "epoch": 6.513092421894118, "grad_norm": 2.6595800591167063e-05, "learning_rate": 4.4536457626254134e-10, "loss": 0.0, "num_input_tokens_seen": 76820224, "step": 39925 }, { "epoch": 6.513908149114936, "grad_norm": 6.791690248064697e-05, "learning_rate": 3.88691772365557e-10, "loss": 0.0, "num_input_tokens_seen": 76829824, "step": 39930 }, { "epoch": 6.514723876335753, "grad_norm": 0.0013166024582460523, "learning_rate": 3.358742226955425e-10, "loss": 0.0, "num_input_tokens_seen": 76838992, "step": 39935 }, { "epoch": 6.5155396035565705, "grad_norm": 0.024361034855246544, "learning_rate": 2.8691193539875925e-10, "loss": 0.0163, "num_input_tokens_seen": 76848464, "step": 39940 }, { "epoch": 6.5163553307773885, "grad_norm": 0.4554716646671295, "learning_rate": 2.418049180274995e-10, "loss": 0.0004, "num_input_tokens_seen": 76858288, "step": 39945 }, { "epoch": 6.517171057998206, "grad_norm": 0.009627443738281727, "learning_rate": 2.005531775373104e-10, "loss": 0.1063, "num_input_tokens_seen": 76867280, "step": 39950 }, { "epoch": 6.517986785219023, "grad_norm": 0.010694829747080803, "learning_rate": 1.6315672028699435e-10, "loss": 0.0177, "num_input_tokens_seen": 76876992, "step": 39955 }, { "epoch": 6.51880251243984, "grad_norm": 0.0006582719506695867, "learning_rate": 1.2961555204693555e-10, "loss": 0.0001, "num_input_tokens_seen": 76887008, "step": 39960 }, { "epoch": 6.519618239660657, "grad_norm": 0.012019234709441662, "learning_rate": 9.992967798799768e-11, "loss": 0.0002, "num_input_tokens_seen": 76896288, "step": 39965 }, { "epoch": 6.520433966881475, "grad_norm": 0.0004875215236097574, "learning_rate": 7.409910268707521e-11, "loss": 0.0, "num_input_tokens_seen": 76906288, "step": 39970 }, { "epoch": 6.521249694102292, "grad_norm": 0.017597844824194908, "learning_rate": 5.212383012986877e-11, "loss": 0.0003, "num_input_tokens_seen": 76914000, "step": 39975 }, { "epoch": 6.522065421323109, "grad_norm": 0.013650244101881981, "learning_rate": 3.400386370533415e-11, "loss": 0.0, "num_input_tokens_seen": 76923104, "step": 39980 }, { "epoch": 6.522881148543927, "grad_norm": 0.001450541545636952, "learning_rate": 1.9739206205682258e-11, "loss": 0.0252, "num_input_tokens_seen": 76933424, "step": 39985 }, { "epoch": 6.523696875764744, "grad_norm": 0.002987588755786419, "learning_rate": 9.329859829154685e-12, "loss": 0.001, "num_input_tokens_seen": 76943616, "step": 39990 }, { "epoch": 6.524512602985562, "grad_norm": 0.13603216409683228, "learning_rate": 2.7758261855748148e-12, "loss": 0.0005, "num_input_tokens_seen": 76953296, "step": 39995 }, { "epoch": 6.525328330206379, "grad_norm": 0.0351916179060936, "learning_rate": 7.710628524559838e-14, "loss": 0.0001, "num_input_tokens_seen": 76963024, "step": 40000 }, { "epoch": 6.525328330206379, "eval_loss": 0.42914924025535583, "eval_runtime": 80.7775, "eval_samples_per_second": 33.735, "eval_steps_per_second": 16.874, "num_input_tokens_seen": 76963024, "step": 40000 }, { "epoch": 6.525328330206379, "num_input_tokens_seen": 76963024, "step": 40000, "total_flos": 3.2436319396082074e+17, "train_loss": 0.07350978353808856, "train_runtime": 33114.2032, "train_samples_per_second": 4.832, "train_steps_per_second": 1.208 } ], "logging_steps": 5, "max_steps": 40000, "num_input_tokens_seen": 76963024, "num_train_epochs": 7, "save_steps": 200, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 3.2436319396082074e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }