diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,65844 @@ +{ + "best_global_step": 37800, + "best_metric": 0.49408820271492004, + "best_model_checkpoint": "saves/ia3/mistral-7b-instruct/train_record_1745950254/checkpoint-37800", + "epoch": 1.2803277639075603, + "eval_steps": 200, + "global_step": 40000, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.00016004097048844505, + "grad_norm": 10.01242446899414, + "learning_rate": 4.999999876629946e-05, + "loss": 4.7588, + "num_input_tokens_seen": 7696, + "step": 5 + }, + { + "epoch": 0.0003200819409768901, + "grad_norm": 5.193350791931152, + "learning_rate": 4.999999375439123e-05, + "loss": 4.4944, + "num_input_tokens_seen": 15264, + "step": 10 + }, + { + "epoch": 0.0004801229114653351, + "grad_norm": 19.06399917602539, + "learning_rate": 4.9999984887169785e-05, + "loss": 3.0838, + "num_input_tokens_seen": 23168, + "step": 15 + }, + { + "epoch": 0.0006401638819537802, + "grad_norm": 17.543014526367188, + "learning_rate": 4.9999972164636506e-05, + "loss": 3.4841, + "num_input_tokens_seen": 30928, + "step": 20 + }, + { + "epoch": 0.0008002048524422252, + "grad_norm": 8.948269844055176, + "learning_rate": 4.999995558679334e-05, + "loss": 3.4071, + "num_input_tokens_seen": 38400, + "step": 25 + }, + { + "epoch": 0.0009602458229306702, + "grad_norm": 10.509188652038574, + "learning_rate": 4.999993515364287e-05, + "loss": 5.2599, + "num_input_tokens_seen": 46592, + "step": 30 + }, + { + "epoch": 0.0011202867934191152, + "grad_norm": 16.99201774597168, + "learning_rate": 4.999991086518822e-05, + "loss": 4.4605, + "num_input_tokens_seen": 54256, + "step": 35 + }, + { + "epoch": 0.0012803277639075604, + "grad_norm": 8.616759300231934, + "learning_rate": 4.999988272143315e-05, + "loss": 3.314, + "num_input_tokens_seen": 62000, + "step": 40 + }, + { + "epoch": 0.0014403687343960054, + "grad_norm": 17.70408058166504, + "learning_rate": 4.999985072238199e-05, + "loss": 4.7484, + "num_input_tokens_seen": 69712, + "step": 45 + }, + { + "epoch": 0.0016004097048844504, + "grad_norm": 6.372953414916992, + "learning_rate": 4.999981486803969e-05, + "loss": 3.3694, + "num_input_tokens_seen": 77056, + "step": 50 + }, + { + "epoch": 0.0017604506753728954, + "grad_norm": 15.68087387084961, + "learning_rate": 4.999977515841176e-05, + "loss": 3.0911, + "num_input_tokens_seen": 84736, + "step": 55 + }, + { + "epoch": 0.0019204916458613404, + "grad_norm": 7.74759578704834, + "learning_rate": 4.9999731593504344e-05, + "loss": 4.7994, + "num_input_tokens_seen": 92192, + "step": 60 + }, + { + "epoch": 0.0020805326163497854, + "grad_norm": 10.348034858703613, + "learning_rate": 4.999968417332415e-05, + "loss": 2.9374, + "num_input_tokens_seen": 99936, + "step": 65 + }, + { + "epoch": 0.0022405735868382304, + "grad_norm": 6.791186809539795, + "learning_rate": 4.999963289787848e-05, + "loss": 2.3165, + "num_input_tokens_seen": 108016, + "step": 70 + }, + { + "epoch": 0.002400614557326676, + "grad_norm": 6.738121032714844, + "learning_rate": 4.999957776717526e-05, + "loss": 1.7202, + "num_input_tokens_seen": 115408, + "step": 75 + }, + { + "epoch": 0.002560655527815121, + "grad_norm": 5.398982524871826, + "learning_rate": 4.9999518781222984e-05, + "loss": 1.3302, + "num_input_tokens_seen": 122912, + "step": 80 + }, + { + "epoch": 0.002720696498303566, + "grad_norm": 5.035694122314453, + "learning_rate": 4.9999455940030746e-05, + "loss": 2.6427, + "num_input_tokens_seen": 130960, + "step": 85 + }, + { + "epoch": 0.002880737468792011, + "grad_norm": 5.1332502365112305, + "learning_rate": 4.999938924360824e-05, + "loss": 2.0842, + "num_input_tokens_seen": 138800, + "step": 90 + }, + { + "epoch": 0.003040778439280456, + "grad_norm": 6.531647205352783, + "learning_rate": 4.999931869196575e-05, + "loss": 2.2955, + "num_input_tokens_seen": 146816, + "step": 95 + }, + { + "epoch": 0.003200819409768901, + "grad_norm": 2.9172205924987793, + "learning_rate": 4.999924428511416e-05, + "loss": 2.2762, + "num_input_tokens_seen": 154528, + "step": 100 + }, + { + "epoch": 0.003360860380257346, + "grad_norm": 3.0506885051727295, + "learning_rate": 4.999916602306494e-05, + "loss": 1.7255, + "num_input_tokens_seen": 162336, + "step": 105 + }, + { + "epoch": 0.003520901350745791, + "grad_norm": 2.125943899154663, + "learning_rate": 4.999908390583016e-05, + "loss": 0.8693, + "num_input_tokens_seen": 169984, + "step": 110 + }, + { + "epoch": 0.003680942321234236, + "grad_norm": 3.693803310394287, + "learning_rate": 4.999899793342247e-05, + "loss": 1.9365, + "num_input_tokens_seen": 177840, + "step": 115 + }, + { + "epoch": 0.003840983291722681, + "grad_norm": 4.745045185089111, + "learning_rate": 4.999890810585516e-05, + "loss": 2.368, + "num_input_tokens_seen": 185280, + "step": 120 + }, + { + "epoch": 0.004001024262211126, + "grad_norm": 4.962631702423096, + "learning_rate": 4.999881442314206e-05, + "loss": 1.6402, + "num_input_tokens_seen": 192688, + "step": 125 + }, + { + "epoch": 0.004161065232699571, + "grad_norm": 7.916421890258789, + "learning_rate": 4.9998716885297617e-05, + "loss": 1.3085, + "num_input_tokens_seen": 200192, + "step": 130 + }, + { + "epoch": 0.004321106203188016, + "grad_norm": 4.663506984710693, + "learning_rate": 4.999861549233688e-05, + "loss": 1.8731, + "num_input_tokens_seen": 207808, + "step": 135 + }, + { + "epoch": 0.004481147173676461, + "grad_norm": 6.527729034423828, + "learning_rate": 4.999851024427548e-05, + "loss": 2.4475, + "num_input_tokens_seen": 215088, + "step": 140 + }, + { + "epoch": 0.004641188144164906, + "grad_norm": 6.063356399536133, + "learning_rate": 4.999840114112965e-05, + "loss": 1.7387, + "num_input_tokens_seen": 222352, + "step": 145 + }, + { + "epoch": 0.004801229114653352, + "grad_norm": 5.421499729156494, + "learning_rate": 4.999828818291621e-05, + "loss": 1.6335, + "num_input_tokens_seen": 229904, + "step": 150 + }, + { + "epoch": 0.004961270085141797, + "grad_norm": 4.515336990356445, + "learning_rate": 4.999817136965259e-05, + "loss": 1.752, + "num_input_tokens_seen": 237088, + "step": 155 + }, + { + "epoch": 0.005121311055630242, + "grad_norm": 5.148532867431641, + "learning_rate": 4.9998050701356794e-05, + "loss": 1.8466, + "num_input_tokens_seen": 244288, + "step": 160 + }, + { + "epoch": 0.005281352026118687, + "grad_norm": 4.20058012008667, + "learning_rate": 4.999792617804744e-05, + "loss": 1.9445, + "num_input_tokens_seen": 252080, + "step": 165 + }, + { + "epoch": 0.005441392996607132, + "grad_norm": 6.673853397369385, + "learning_rate": 4.9997797799743724e-05, + "loss": 1.4664, + "num_input_tokens_seen": 259616, + "step": 170 + }, + { + "epoch": 0.005601433967095577, + "grad_norm": 4.822836399078369, + "learning_rate": 4.999766556646545e-05, + "loss": 1.314, + "num_input_tokens_seen": 267696, + "step": 175 + }, + { + "epoch": 0.005761474937584022, + "grad_norm": 4.849950313568115, + "learning_rate": 4.9997529478232996e-05, + "loss": 1.1727, + "num_input_tokens_seen": 275424, + "step": 180 + }, + { + "epoch": 0.005921515908072467, + "grad_norm": 3.6646735668182373, + "learning_rate": 4.9997389535067365e-05, + "loss": 1.4723, + "num_input_tokens_seen": 283280, + "step": 185 + }, + { + "epoch": 0.006081556878560912, + "grad_norm": 5.454855918884277, + "learning_rate": 4.999724573699012e-05, + "loss": 1.7534, + "num_input_tokens_seen": 291136, + "step": 190 + }, + { + "epoch": 0.006241597849049357, + "grad_norm": 12.758896827697754, + "learning_rate": 4.9997098084023457e-05, + "loss": 1.762, + "num_input_tokens_seen": 298960, + "step": 195 + }, + { + "epoch": 0.006401638819537802, + "grad_norm": 1.750335693359375, + "learning_rate": 4.999694657619013e-05, + "loss": 0.8115, + "num_input_tokens_seen": 306656, + "step": 200 + }, + { + "epoch": 0.006401638819537802, + "eval_loss": 1.5857535600662231, + "eval_runtime": 557.5128, + "eval_samples_per_second": 24.907, + "eval_steps_per_second": 12.454, + "num_input_tokens_seen": 306656, + "step": 200 + }, + { + "epoch": 0.006561679790026247, + "grad_norm": 6.26470422744751, + "learning_rate": 4.999679121351352e-05, + "loss": 1.6697, + "num_input_tokens_seen": 314032, + "step": 205 + }, + { + "epoch": 0.006721720760514692, + "grad_norm": 3.183344602584839, + "learning_rate": 4.9996631996017565e-05, + "loss": 1.2557, + "num_input_tokens_seen": 321968, + "step": 210 + }, + { + "epoch": 0.006881761731003137, + "grad_norm": 7.393520355224609, + "learning_rate": 4.9996468923726835e-05, + "loss": 0.7266, + "num_input_tokens_seen": 329600, + "step": 215 + }, + { + "epoch": 0.007041802701491582, + "grad_norm": 2.4141907691955566, + "learning_rate": 4.999630199666647e-05, + "loss": 1.3516, + "num_input_tokens_seen": 336864, + "step": 220 + }, + { + "epoch": 0.007201843671980027, + "grad_norm": 3.184197187423706, + "learning_rate": 4.999613121486222e-05, + "loss": 0.8843, + "num_input_tokens_seen": 344544, + "step": 225 + }, + { + "epoch": 0.007361884642468472, + "grad_norm": 6.633127212524414, + "learning_rate": 4.999595657834041e-05, + "loss": 1.4582, + "num_input_tokens_seen": 352224, + "step": 230 + }, + { + "epoch": 0.007521925612956917, + "grad_norm": 2.111218214035034, + "learning_rate": 4.999577808712798e-05, + "loss": 2.3024, + "num_input_tokens_seen": 359680, + "step": 235 + }, + { + "epoch": 0.007681966583445362, + "grad_norm": 2.5455076694488525, + "learning_rate": 4.999559574125244e-05, + "loss": 1.4034, + "num_input_tokens_seen": 367008, + "step": 240 + }, + { + "epoch": 0.007842007553933807, + "grad_norm": 1.381466031074524, + "learning_rate": 4.9995409540741934e-05, + "loss": 1.6338, + "num_input_tokens_seen": 374768, + "step": 245 + }, + { + "epoch": 0.008002048524422252, + "grad_norm": 6.325981140136719, + "learning_rate": 4.999521948562516e-05, + "loss": 1.3238, + "num_input_tokens_seen": 382096, + "step": 250 + }, + { + "epoch": 0.008162089494910697, + "grad_norm": 2.925565004348755, + "learning_rate": 4.999502557593143e-05, + "loss": 1.4591, + "num_input_tokens_seen": 389648, + "step": 255 + }, + { + "epoch": 0.008322130465399142, + "grad_norm": 2.258669376373291, + "learning_rate": 4.999482781169066e-05, + "loss": 1.5647, + "num_input_tokens_seen": 396992, + "step": 260 + }, + { + "epoch": 0.008482171435887587, + "grad_norm": 4.067651271820068, + "learning_rate": 4.9994626192933324e-05, + "loss": 1.4, + "num_input_tokens_seen": 404432, + "step": 265 + }, + { + "epoch": 0.008642212406376032, + "grad_norm": 3.896963119506836, + "learning_rate": 4.999442071969054e-05, + "loss": 1.2512, + "num_input_tokens_seen": 411824, + "step": 270 + }, + { + "epoch": 0.008802253376864477, + "grad_norm": 5.286522388458252, + "learning_rate": 4.999421139199397e-05, + "loss": 1.4775, + "num_input_tokens_seen": 419680, + "step": 275 + }, + { + "epoch": 0.008962294347352922, + "grad_norm": 3.6899430751800537, + "learning_rate": 4.999399820987592e-05, + "loss": 1.2613, + "num_input_tokens_seen": 427040, + "step": 280 + }, + { + "epoch": 0.009122335317841367, + "grad_norm": 4.498958587646484, + "learning_rate": 4.999378117336924e-05, + "loss": 1.057, + "num_input_tokens_seen": 434800, + "step": 285 + }, + { + "epoch": 0.009282376288329812, + "grad_norm": 1.7503211498260498, + "learning_rate": 4.9993560282507415e-05, + "loss": 1.3483, + "num_input_tokens_seen": 442256, + "step": 290 + }, + { + "epoch": 0.009442417258818258, + "grad_norm": 2.566305637359619, + "learning_rate": 4.9993335537324495e-05, + "loss": 1.164, + "num_input_tokens_seen": 449936, + "step": 295 + }, + { + "epoch": 0.009602458229306703, + "grad_norm": 3.7911529541015625, + "learning_rate": 4.999310693785516e-05, + "loss": 1.1641, + "num_input_tokens_seen": 457200, + "step": 300 + }, + { + "epoch": 0.009762499199795148, + "grad_norm": 1.9935815334320068, + "learning_rate": 4.9992874484134653e-05, + "loss": 0.8205, + "num_input_tokens_seen": 464848, + "step": 305 + }, + { + "epoch": 0.009922540170283593, + "grad_norm": 3.420297145843506, + "learning_rate": 4.999263817619882e-05, + "loss": 1.4838, + "num_input_tokens_seen": 472400, + "step": 310 + }, + { + "epoch": 0.010082581140772038, + "grad_norm": 3.066006660461426, + "learning_rate": 4.9992398014084105e-05, + "loss": 1.3454, + "num_input_tokens_seen": 479552, + "step": 315 + }, + { + "epoch": 0.010242622111260483, + "grad_norm": 2.7864933013916016, + "learning_rate": 4.999215399782754e-05, + "loss": 1.4325, + "num_input_tokens_seen": 486688, + "step": 320 + }, + { + "epoch": 0.010402663081748928, + "grad_norm": 1.6280431747436523, + "learning_rate": 4.999190612746675e-05, + "loss": 1.1503, + "num_input_tokens_seen": 494192, + "step": 325 + }, + { + "epoch": 0.010562704052237373, + "grad_norm": 1.6240603923797607, + "learning_rate": 4.999165440303998e-05, + "loss": 1.2267, + "num_input_tokens_seen": 501424, + "step": 330 + }, + { + "epoch": 0.010722745022725818, + "grad_norm": 4.077406406402588, + "learning_rate": 4.999139882458603e-05, + "loss": 1.0886, + "num_input_tokens_seen": 508592, + "step": 335 + }, + { + "epoch": 0.010882785993214263, + "grad_norm": 8.10267162322998, + "learning_rate": 4.9991139392144314e-05, + "loss": 1.224, + "num_input_tokens_seen": 516128, + "step": 340 + }, + { + "epoch": 0.011042826963702708, + "grad_norm": 3.6765384674072266, + "learning_rate": 4.999087610575485e-05, + "loss": 0.9542, + "num_input_tokens_seen": 523568, + "step": 345 + }, + { + "epoch": 0.011202867934191153, + "grad_norm": 3.1798064708709717, + "learning_rate": 4.999060896545824e-05, + "loss": 1.3639, + "num_input_tokens_seen": 531360, + "step": 350 + }, + { + "epoch": 0.011362908904679598, + "grad_norm": 3.8282735347747803, + "learning_rate": 4.999033797129568e-05, + "loss": 1.2266, + "num_input_tokens_seen": 539392, + "step": 355 + }, + { + "epoch": 0.011522949875168043, + "grad_norm": 4.848293781280518, + "learning_rate": 4.999006312330894e-05, + "loss": 1.4184, + "num_input_tokens_seen": 547168, + "step": 360 + }, + { + "epoch": 0.011682990845656488, + "grad_norm": 3.6730473041534424, + "learning_rate": 4.998978442154043e-05, + "loss": 1.2441, + "num_input_tokens_seen": 554896, + "step": 365 + }, + { + "epoch": 0.011843031816144933, + "grad_norm": 2.781184673309326, + "learning_rate": 4.9989501866033125e-05, + "loss": 1.5538, + "num_input_tokens_seen": 561984, + "step": 370 + }, + { + "epoch": 0.012003072786633378, + "grad_norm": 5.691532135009766, + "learning_rate": 4.998921545683059e-05, + "loss": 0.8029, + "num_input_tokens_seen": 569760, + "step": 375 + }, + { + "epoch": 0.012163113757121823, + "grad_norm": 2.3803822994232178, + "learning_rate": 4.9988925193976996e-05, + "loss": 1.3562, + "num_input_tokens_seen": 577200, + "step": 380 + }, + { + "epoch": 0.012323154727610268, + "grad_norm": 3.0285041332244873, + "learning_rate": 4.998863107751711e-05, + "loss": 0.8533, + "num_input_tokens_seen": 584864, + "step": 385 + }, + { + "epoch": 0.012483195698098713, + "grad_norm": 1.9115033149719238, + "learning_rate": 4.998833310749629e-05, + "loss": 1.2375, + "num_input_tokens_seen": 592208, + "step": 390 + }, + { + "epoch": 0.012643236668587158, + "grad_norm": 4.9323296546936035, + "learning_rate": 4.998803128396047e-05, + "loss": 1.294, + "num_input_tokens_seen": 600432, + "step": 395 + }, + { + "epoch": 0.012803277639075603, + "grad_norm": 2.0581071376800537, + "learning_rate": 4.9987725606956215e-05, + "loss": 1.7881, + "num_input_tokens_seen": 608000, + "step": 400 + }, + { + "epoch": 0.012803277639075603, + "eval_loss": 1.1619302034378052, + "eval_runtime": 558.1467, + "eval_samples_per_second": 24.879, + "eval_steps_per_second": 12.439, + "num_input_tokens_seen": 608000, + "step": 400 + }, + { + "epoch": 0.012963318609564048, + "grad_norm": 2.3770320415496826, + "learning_rate": 4.998741607653066e-05, + "loss": 1.1357, + "num_input_tokens_seen": 615392, + "step": 405 + }, + { + "epoch": 0.013123359580052493, + "grad_norm": 2.0840580463409424, + "learning_rate": 4.9987102692731523e-05, + "loss": 0.9292, + "num_input_tokens_seen": 623072, + "step": 410 + }, + { + "epoch": 0.013283400550540938, + "grad_norm": 3.901731014251709, + "learning_rate": 4.9986785455607157e-05, + "loss": 0.8809, + "num_input_tokens_seen": 630704, + "step": 415 + }, + { + "epoch": 0.013443441521029383, + "grad_norm": 3.217191219329834, + "learning_rate": 4.9986464365206456e-05, + "loss": 1.0542, + "num_input_tokens_seen": 638176, + "step": 420 + }, + { + "epoch": 0.013603482491517828, + "grad_norm": 4.30696439743042, + "learning_rate": 4.9986139421578956e-05, + "loss": 0.8431, + "num_input_tokens_seen": 645488, + "step": 425 + }, + { + "epoch": 0.013763523462006273, + "grad_norm": 6.916722774505615, + "learning_rate": 4.998581062477477e-05, + "loss": 0.9667, + "num_input_tokens_seen": 653280, + "step": 430 + }, + { + "epoch": 0.013923564432494718, + "grad_norm": 3.49442982673645, + "learning_rate": 4.998547797484458e-05, + "loss": 1.3155, + "num_input_tokens_seen": 660768, + "step": 435 + }, + { + "epoch": 0.014083605402983163, + "grad_norm": 3.4313900470733643, + "learning_rate": 4.9985141471839706e-05, + "loss": 1.3797, + "num_input_tokens_seen": 668304, + "step": 440 + }, + { + "epoch": 0.014243646373471608, + "grad_norm": 1.7349777221679688, + "learning_rate": 4.998480111581203e-05, + "loss": 0.8229, + "num_input_tokens_seen": 675840, + "step": 445 + }, + { + "epoch": 0.014403687343960053, + "grad_norm": 2.2968337535858154, + "learning_rate": 4.998445690681405e-05, + "loss": 1.1371, + "num_input_tokens_seen": 683120, + "step": 450 + }, + { + "epoch": 0.014563728314448498, + "grad_norm": 1.937695026397705, + "learning_rate": 4.9984108844898834e-05, + "loss": 1.3206, + "num_input_tokens_seen": 690880, + "step": 455 + }, + { + "epoch": 0.014723769284936943, + "grad_norm": 1.560514211654663, + "learning_rate": 4.9983756930120076e-05, + "loss": 1.0426, + "num_input_tokens_seen": 698992, + "step": 460 + }, + { + "epoch": 0.014883810255425388, + "grad_norm": 3.9114527702331543, + "learning_rate": 4.9983401162532025e-05, + "loss": 0.9359, + "num_input_tokens_seen": 707040, + "step": 465 + }, + { + "epoch": 0.015043851225913833, + "grad_norm": 2.5483570098876953, + "learning_rate": 4.998304154218955e-05, + "loss": 1.0475, + "num_input_tokens_seen": 715040, + "step": 470 + }, + { + "epoch": 0.015203892196402278, + "grad_norm": 4.075963497161865, + "learning_rate": 4.998267806914812e-05, + "loss": 0.8347, + "num_input_tokens_seen": 723456, + "step": 475 + }, + { + "epoch": 0.015363933166890723, + "grad_norm": 3.85520339012146, + "learning_rate": 4.998231074346378e-05, + "loss": 0.6205, + "num_input_tokens_seen": 730640, + "step": 480 + }, + { + "epoch": 0.015523974137379168, + "grad_norm": 2.1799368858337402, + "learning_rate": 4.998193956519317e-05, + "loss": 1.0752, + "num_input_tokens_seen": 738096, + "step": 485 + }, + { + "epoch": 0.015684015107867613, + "grad_norm": 2.144212484359741, + "learning_rate": 4.9981564534393545e-05, + "loss": 0.9296, + "num_input_tokens_seen": 745568, + "step": 490 + }, + { + "epoch": 0.01584405607835606, + "grad_norm": 3.816544532775879, + "learning_rate": 4.998118565112272e-05, + "loss": 1.2985, + "num_input_tokens_seen": 753152, + "step": 495 + }, + { + "epoch": 0.016004097048844503, + "grad_norm": 4.927980899810791, + "learning_rate": 4.998080291543914e-05, + "loss": 1.2954, + "num_input_tokens_seen": 760432, + "step": 500 + }, + { + "epoch": 0.01616413801933295, + "grad_norm": 1.3879120349884033, + "learning_rate": 4.9980416327401826e-05, + "loss": 1.0703, + "num_input_tokens_seen": 767968, + "step": 505 + }, + { + "epoch": 0.016324178989821393, + "grad_norm": 1.798303484916687, + "learning_rate": 4.998002588707038e-05, + "loss": 1.0416, + "num_input_tokens_seen": 776032, + "step": 510 + }, + { + "epoch": 0.01648421996030984, + "grad_norm": 2.0076215267181396, + "learning_rate": 4.997963159450503e-05, + "loss": 0.8314, + "num_input_tokens_seen": 784016, + "step": 515 + }, + { + "epoch": 0.016644260930798283, + "grad_norm": 2.5197644233703613, + "learning_rate": 4.9979233449766575e-05, + "loss": 1.0234, + "num_input_tokens_seen": 792464, + "step": 520 + }, + { + "epoch": 0.01680430190128673, + "grad_norm": 3.5475265979766846, + "learning_rate": 4.997883145291641e-05, + "loss": 1.2653, + "num_input_tokens_seen": 799968, + "step": 525 + }, + { + "epoch": 0.016964342871775173, + "grad_norm": 2.2246124744415283, + "learning_rate": 4.9978425604016536e-05, + "loss": 1.0265, + "num_input_tokens_seen": 807776, + "step": 530 + }, + { + "epoch": 0.01712438384226362, + "grad_norm": 4.612852096557617, + "learning_rate": 4.9978015903129536e-05, + "loss": 1.2003, + "num_input_tokens_seen": 815584, + "step": 535 + }, + { + "epoch": 0.017284424812752063, + "grad_norm": 5.253464698791504, + "learning_rate": 4.997760235031859e-05, + "loss": 1.6823, + "num_input_tokens_seen": 823104, + "step": 540 + }, + { + "epoch": 0.01744446578324051, + "grad_norm": 2.41931414604187, + "learning_rate": 4.9977184945647473e-05, + "loss": 0.9945, + "num_input_tokens_seen": 830768, + "step": 545 + }, + { + "epoch": 0.017604506753728953, + "grad_norm": 2.9875924587249756, + "learning_rate": 4.997676368918055e-05, + "loss": 1.1034, + "num_input_tokens_seen": 838384, + "step": 550 + }, + { + "epoch": 0.0177645477242174, + "grad_norm": 3.074810028076172, + "learning_rate": 4.9976338580982794e-05, + "loss": 1.0054, + "num_input_tokens_seen": 845584, + "step": 555 + }, + { + "epoch": 0.017924588694705843, + "grad_norm": 1.7997236251831055, + "learning_rate": 4.9975909621119755e-05, + "loss": 1.2012, + "num_input_tokens_seen": 853760, + "step": 560 + }, + { + "epoch": 0.01808462966519429, + "grad_norm": 2.341209888458252, + "learning_rate": 4.997547680965758e-05, + "loss": 0.8441, + "num_input_tokens_seen": 861696, + "step": 565 + }, + { + "epoch": 0.018244670635682733, + "grad_norm": 1.9608770608901978, + "learning_rate": 4.997504014666302e-05, + "loss": 0.9762, + "num_input_tokens_seen": 869120, + "step": 570 + }, + { + "epoch": 0.01840471160617118, + "grad_norm": 2.9394819736480713, + "learning_rate": 4.997459963220342e-05, + "loss": 1.2098, + "num_input_tokens_seen": 876672, + "step": 575 + }, + { + "epoch": 0.018564752576659623, + "grad_norm": 1.4544752836227417, + "learning_rate": 4.997415526634671e-05, + "loss": 0.9151, + "num_input_tokens_seen": 884352, + "step": 580 + }, + { + "epoch": 0.01872479354714807, + "grad_norm": 1.790677547454834, + "learning_rate": 4.99737070491614e-05, + "loss": 1.2198, + "num_input_tokens_seen": 892176, + "step": 585 + }, + { + "epoch": 0.018884834517636517, + "grad_norm": 3.3706741333007812, + "learning_rate": 4.997325498071663e-05, + "loss": 0.7666, + "num_input_tokens_seen": 899584, + "step": 590 + }, + { + "epoch": 0.01904487548812496, + "grad_norm": 5.2199201583862305, + "learning_rate": 4.997279906108211e-05, + "loss": 0.9487, + "num_input_tokens_seen": 907472, + "step": 595 + }, + { + "epoch": 0.019204916458613407, + "grad_norm": 2.146742343902588, + "learning_rate": 4.9972339290328155e-05, + "loss": 0.9015, + "num_input_tokens_seen": 914496, + "step": 600 + }, + { + "epoch": 0.019204916458613407, + "eval_loss": 0.9876173138618469, + "eval_runtime": 561.7335, + "eval_samples_per_second": 24.72, + "eval_steps_per_second": 12.36, + "num_input_tokens_seen": 914496, + "step": 600 + }, + { + "epoch": 0.01936495742910185, + "grad_norm": 2.2437663078308105, + "learning_rate": 4.9971875668525646e-05, + "loss": 0.9269, + "num_input_tokens_seen": 922064, + "step": 605 + }, + { + "epoch": 0.019524998399590297, + "grad_norm": 2.1510448455810547, + "learning_rate": 4.997140819574609e-05, + "loss": 1.2778, + "num_input_tokens_seen": 930480, + "step": 610 + }, + { + "epoch": 0.01968503937007874, + "grad_norm": 2.162353038787842, + "learning_rate": 4.997093687206159e-05, + "loss": 0.8322, + "num_input_tokens_seen": 937776, + "step": 615 + }, + { + "epoch": 0.019845080340567187, + "grad_norm": 1.8398215770721436, + "learning_rate": 4.997046169754482e-05, + "loss": 0.8462, + "num_input_tokens_seen": 945296, + "step": 620 + }, + { + "epoch": 0.02000512131105563, + "grad_norm": 1.5070343017578125, + "learning_rate": 4.996998267226905e-05, + "loss": 0.844, + "num_input_tokens_seen": 952880, + "step": 625 + }, + { + "epoch": 0.020165162281544077, + "grad_norm": 1.8440190553665161, + "learning_rate": 4.996949979630817e-05, + "loss": 0.6508, + "num_input_tokens_seen": 960464, + "step": 630 + }, + { + "epoch": 0.02032520325203252, + "grad_norm": 1.8594907522201538, + "learning_rate": 4.996901306973663e-05, + "loss": 1.0339, + "num_input_tokens_seen": 967680, + "step": 635 + }, + { + "epoch": 0.020485244222520967, + "grad_norm": 3.5400309562683105, + "learning_rate": 4.996852249262949e-05, + "loss": 0.8278, + "num_input_tokens_seen": 975600, + "step": 640 + }, + { + "epoch": 0.02064528519300941, + "grad_norm": 3.58797025680542, + "learning_rate": 4.996802806506241e-05, + "loss": 0.9997, + "num_input_tokens_seen": 983424, + "step": 645 + }, + { + "epoch": 0.020805326163497857, + "grad_norm": 3.4604649543762207, + "learning_rate": 4.996752978711164e-05, + "loss": 0.974, + "num_input_tokens_seen": 991168, + "step": 650 + }, + { + "epoch": 0.0209653671339863, + "grad_norm": 2.9648900032043457, + "learning_rate": 4.996702765885401e-05, + "loss": 1.124, + "num_input_tokens_seen": 999088, + "step": 655 + }, + { + "epoch": 0.021125408104474747, + "grad_norm": 2.95387864112854, + "learning_rate": 4.9966521680366964e-05, + "loss": 0.6761, + "num_input_tokens_seen": 1006224, + "step": 660 + }, + { + "epoch": 0.02128544907496319, + "grad_norm": 1.3770196437835693, + "learning_rate": 4.9966011851728524e-05, + "loss": 1.119, + "num_input_tokens_seen": 1013984, + "step": 665 + }, + { + "epoch": 0.021445490045451637, + "grad_norm": 3.297236204147339, + "learning_rate": 4.996549817301731e-05, + "loss": 0.9518, + "num_input_tokens_seen": 1021808, + "step": 670 + }, + { + "epoch": 0.02160553101594008, + "grad_norm": 2.6880807876586914, + "learning_rate": 4.9964980644312544e-05, + "loss": 0.7317, + "num_input_tokens_seen": 1029168, + "step": 675 + }, + { + "epoch": 0.021765571986428527, + "grad_norm": 1.9097691774368286, + "learning_rate": 4.996445926569403e-05, + "loss": 0.9965, + "num_input_tokens_seen": 1036704, + "step": 680 + }, + { + "epoch": 0.02192561295691697, + "grad_norm": 2.432004451751709, + "learning_rate": 4.996393403724218e-05, + "loss": 0.8078, + "num_input_tokens_seen": 1044032, + "step": 685 + }, + { + "epoch": 0.022085653927405417, + "grad_norm": 2.3165552616119385, + "learning_rate": 4.9963404959037985e-05, + "loss": 0.9325, + "num_input_tokens_seen": 1052352, + "step": 690 + }, + { + "epoch": 0.02224569489789386, + "grad_norm": 2.987881660461426, + "learning_rate": 4.996287203116303e-05, + "loss": 1.0882, + "num_input_tokens_seen": 1059872, + "step": 695 + }, + { + "epoch": 0.022405735868382307, + "grad_norm": 2.5928432941436768, + "learning_rate": 4.996233525369951e-05, + "loss": 0.967, + "num_input_tokens_seen": 1067008, + "step": 700 + }, + { + "epoch": 0.02256577683887075, + "grad_norm": 2.77951979637146, + "learning_rate": 4.99617946267302e-05, + "loss": 0.8314, + "num_input_tokens_seen": 1074288, + "step": 705 + }, + { + "epoch": 0.022725817809359197, + "grad_norm": 0.922175943851471, + "learning_rate": 4.996125015033846e-05, + "loss": 0.7353, + "num_input_tokens_seen": 1081792, + "step": 710 + }, + { + "epoch": 0.02288585877984764, + "grad_norm": 2.5809390544891357, + "learning_rate": 4.996070182460827e-05, + "loss": 0.8979, + "num_input_tokens_seen": 1089712, + "step": 715 + }, + { + "epoch": 0.023045899750336087, + "grad_norm": 2.335158109664917, + "learning_rate": 4.996014964962418e-05, + "loss": 1.0485, + "num_input_tokens_seen": 1097392, + "step": 720 + }, + { + "epoch": 0.02320594072082453, + "grad_norm": 1.5486377477645874, + "learning_rate": 4.9959593625471344e-05, + "loss": 0.9775, + "num_input_tokens_seen": 1105344, + "step": 725 + }, + { + "epoch": 0.023365981691312977, + "grad_norm": 1.9421513080596924, + "learning_rate": 4.995903375223552e-05, + "loss": 0.9788, + "num_input_tokens_seen": 1113072, + "step": 730 + }, + { + "epoch": 0.02352602266180142, + "grad_norm": 2.987614154815674, + "learning_rate": 4.995847003000302e-05, + "loss": 0.861, + "num_input_tokens_seen": 1120640, + "step": 735 + }, + { + "epoch": 0.023686063632289867, + "grad_norm": 2.0790579319000244, + "learning_rate": 4.9957902458860804e-05, + "loss": 1.1541, + "num_input_tokens_seen": 1127840, + "step": 740 + }, + { + "epoch": 0.02384610460277831, + "grad_norm": 2.2145261764526367, + "learning_rate": 4.995733103889639e-05, + "loss": 0.8692, + "num_input_tokens_seen": 1135184, + "step": 745 + }, + { + "epoch": 0.024006145573266757, + "grad_norm": 1.7522001266479492, + "learning_rate": 4.99567557701979e-05, + "loss": 0.9998, + "num_input_tokens_seen": 1142816, + "step": 750 + }, + { + "epoch": 0.0241661865437552, + "grad_norm": 2.5133213996887207, + "learning_rate": 4.995617665285403e-05, + "loss": 0.845, + "num_input_tokens_seen": 1150288, + "step": 755 + }, + { + "epoch": 0.024326227514243647, + "grad_norm": 1.0021804571151733, + "learning_rate": 4.99555936869541e-05, + "loss": 0.704, + "num_input_tokens_seen": 1157952, + "step": 760 + }, + { + "epoch": 0.02448626848473209, + "grad_norm": 0.9390085935592651, + "learning_rate": 4.995500687258803e-05, + "loss": 0.9316, + "num_input_tokens_seen": 1165808, + "step": 765 + }, + { + "epoch": 0.024646309455220537, + "grad_norm": 2.4175314903259277, + "learning_rate": 4.995441620984628e-05, + "loss": 0.5936, + "num_input_tokens_seen": 1173216, + "step": 770 + }, + { + "epoch": 0.02480635042570898, + "grad_norm": 1.0162566900253296, + "learning_rate": 4.995382169881996e-05, + "loss": 0.6828, + "num_input_tokens_seen": 1180448, + "step": 775 + }, + { + "epoch": 0.024966391396197427, + "grad_norm": 3.004122018814087, + "learning_rate": 4.9953223339600755e-05, + "loss": 0.8181, + "num_input_tokens_seen": 1188176, + "step": 780 + }, + { + "epoch": 0.025126432366685873, + "grad_norm": 1.470341444015503, + "learning_rate": 4.995262113228091e-05, + "loss": 0.7704, + "num_input_tokens_seen": 1195584, + "step": 785 + }, + { + "epoch": 0.025286473337174317, + "grad_norm": 2.2060582637786865, + "learning_rate": 4.995201507695332e-05, + "loss": 0.7819, + "num_input_tokens_seen": 1203072, + "step": 790 + }, + { + "epoch": 0.025446514307662763, + "grad_norm": 2.308779239654541, + "learning_rate": 4.995140517371144e-05, + "loss": 0.6987, + "num_input_tokens_seen": 1210720, + "step": 795 + }, + { + "epoch": 0.025606555278151207, + "grad_norm": 2.119197130203247, + "learning_rate": 4.995079142264932e-05, + "loss": 0.932, + "num_input_tokens_seen": 1218512, + "step": 800 + }, + { + "epoch": 0.025606555278151207, + "eval_loss": 0.8972654938697815, + "eval_runtime": 558.8864, + "eval_samples_per_second": 24.846, + "eval_steps_per_second": 12.423, + "num_input_tokens_seen": 1218512, + "step": 800 + }, + { + "epoch": 0.025766596248639653, + "grad_norm": 1.9359214305877686, + "learning_rate": 4.995017382386162e-05, + "loss": 0.9016, + "num_input_tokens_seen": 1226304, + "step": 805 + }, + { + "epoch": 0.025926637219128097, + "grad_norm": 1.8055107593536377, + "learning_rate": 4.994955237744356e-05, + "loss": 0.605, + "num_input_tokens_seen": 1234016, + "step": 810 + }, + { + "epoch": 0.026086678189616543, + "grad_norm": 1.2635116577148438, + "learning_rate": 4.994892708349101e-05, + "loss": 0.949, + "num_input_tokens_seen": 1241376, + "step": 815 + }, + { + "epoch": 0.026246719160104987, + "grad_norm": 2.578148126602173, + "learning_rate": 4.994829794210035e-05, + "loss": 0.7682, + "num_input_tokens_seen": 1248416, + "step": 820 + }, + { + "epoch": 0.026406760130593433, + "grad_norm": 1.1067172288894653, + "learning_rate": 4.994766495336864e-05, + "loss": 0.7281, + "num_input_tokens_seen": 1256368, + "step": 825 + }, + { + "epoch": 0.026566801101081877, + "grad_norm": 1.7212575674057007, + "learning_rate": 4.994702811739348e-05, + "loss": 1.0467, + "num_input_tokens_seen": 1264240, + "step": 830 + }, + { + "epoch": 0.026726842071570323, + "grad_norm": 1.3689614534378052, + "learning_rate": 4.994638743427308e-05, + "loss": 0.8241, + "num_input_tokens_seen": 1272096, + "step": 835 + }, + { + "epoch": 0.026886883042058767, + "grad_norm": 1.6953729391098022, + "learning_rate": 4.994574290410624e-05, + "loss": 1.1662, + "num_input_tokens_seen": 1279744, + "step": 840 + }, + { + "epoch": 0.027046924012547213, + "grad_norm": 0.7912226915359497, + "learning_rate": 4.9945094526992364e-05, + "loss": 0.6655, + "num_input_tokens_seen": 1286896, + "step": 845 + }, + { + "epoch": 0.027206964983035656, + "grad_norm": 2.0062530040740967, + "learning_rate": 4.994444230303142e-05, + "loss": 0.8006, + "num_input_tokens_seen": 1294560, + "step": 850 + }, + { + "epoch": 0.027367005953524103, + "grad_norm": 1.591726541519165, + "learning_rate": 4.994378623232402e-05, + "loss": 0.8042, + "num_input_tokens_seen": 1302048, + "step": 855 + }, + { + "epoch": 0.027527046924012546, + "grad_norm": 1.7495880126953125, + "learning_rate": 4.99431263149713e-05, + "loss": 1.1768, + "num_input_tokens_seen": 1309504, + "step": 860 + }, + { + "epoch": 0.027687087894500993, + "grad_norm": 1.2567874193191528, + "learning_rate": 4.9942462551075056e-05, + "loss": 1.0262, + "num_input_tokens_seen": 1317120, + "step": 865 + }, + { + "epoch": 0.027847128864989436, + "grad_norm": 2.553496837615967, + "learning_rate": 4.994179494073764e-05, + "loss": 1.2095, + "num_input_tokens_seen": 1324832, + "step": 870 + }, + { + "epoch": 0.028007169835477883, + "grad_norm": 1.1391057968139648, + "learning_rate": 4.9941123484062e-05, + "loss": 0.7984, + "num_input_tokens_seen": 1332704, + "step": 875 + }, + { + "epoch": 0.028167210805966326, + "grad_norm": 2.454726457595825, + "learning_rate": 4.99404481811517e-05, + "loss": 0.9221, + "num_input_tokens_seen": 1339952, + "step": 880 + }, + { + "epoch": 0.028327251776454773, + "grad_norm": 2.8456435203552246, + "learning_rate": 4.9939769032110864e-05, + "loss": 0.7904, + "num_input_tokens_seen": 1347200, + "step": 885 + }, + { + "epoch": 0.028487292746943216, + "grad_norm": 1.5795212984085083, + "learning_rate": 4.993908603704423e-05, + "loss": 0.8187, + "num_input_tokens_seen": 1354288, + "step": 890 + }, + { + "epoch": 0.028647333717431663, + "grad_norm": 1.7874056100845337, + "learning_rate": 4.9938399196057126e-05, + "loss": 0.6578, + "num_input_tokens_seen": 1361824, + "step": 895 + }, + { + "epoch": 0.028807374687920106, + "grad_norm": 1.781019687652588, + "learning_rate": 4.993770850925547e-05, + "loss": 0.8229, + "num_input_tokens_seen": 1369456, + "step": 900 + }, + { + "epoch": 0.028967415658408553, + "grad_norm": 1.8877140283584595, + "learning_rate": 4.993701397674577e-05, + "loss": 1.1041, + "num_input_tokens_seen": 1376848, + "step": 905 + }, + { + "epoch": 0.029127456628896996, + "grad_norm": 2.037485122680664, + "learning_rate": 4.993631559863515e-05, + "loss": 0.7373, + "num_input_tokens_seen": 1384576, + "step": 910 + }, + { + "epoch": 0.029287497599385443, + "grad_norm": 0.8151519298553467, + "learning_rate": 4.9935613375031283e-05, + "loss": 0.6737, + "num_input_tokens_seen": 1391872, + "step": 915 + }, + { + "epoch": 0.029447538569873886, + "grad_norm": 2.281546115875244, + "learning_rate": 4.993490730604248e-05, + "loss": 0.7808, + "num_input_tokens_seen": 1399520, + "step": 920 + }, + { + "epoch": 0.029607579540362333, + "grad_norm": 1.4026532173156738, + "learning_rate": 4.993419739177761e-05, + "loss": 0.9965, + "num_input_tokens_seen": 1407424, + "step": 925 + }, + { + "epoch": 0.029767620510850776, + "grad_norm": 1.0741691589355469, + "learning_rate": 4.9933483632346164e-05, + "loss": 0.7617, + "num_input_tokens_seen": 1415312, + "step": 930 + }, + { + "epoch": 0.029927661481339223, + "grad_norm": 1.673317790031433, + "learning_rate": 4.993276602785821e-05, + "loss": 0.8206, + "num_input_tokens_seen": 1422592, + "step": 935 + }, + { + "epoch": 0.030087702451827666, + "grad_norm": 2.0570428371429443, + "learning_rate": 4.993204457842441e-05, + "loss": 0.7959, + "num_input_tokens_seen": 1429744, + "step": 940 + }, + { + "epoch": 0.030247743422316113, + "grad_norm": 2.1242997646331787, + "learning_rate": 4.993131928415602e-05, + "loss": 1.2791, + "num_input_tokens_seen": 1436832, + "step": 945 + }, + { + "epoch": 0.030407784392804556, + "grad_norm": 1.1686103343963623, + "learning_rate": 4.993059014516489e-05, + "loss": 0.6668, + "num_input_tokens_seen": 1444224, + "step": 950 + }, + { + "epoch": 0.030567825363293003, + "grad_norm": 4.764544486999512, + "learning_rate": 4.9929857161563464e-05, + "loss": 1.0138, + "num_input_tokens_seen": 1452928, + "step": 955 + }, + { + "epoch": 0.030727866333781446, + "grad_norm": 0.9199651479721069, + "learning_rate": 4.992912033346477e-05, + "loss": 0.7681, + "num_input_tokens_seen": 1460688, + "step": 960 + }, + { + "epoch": 0.030887907304269893, + "grad_norm": 1.5518690347671509, + "learning_rate": 4.992837966098245e-05, + "loss": 0.7993, + "num_input_tokens_seen": 1468528, + "step": 965 + }, + { + "epoch": 0.031047948274758336, + "grad_norm": 1.2777084112167358, + "learning_rate": 4.992763514423071e-05, + "loss": 0.9223, + "num_input_tokens_seen": 1476272, + "step": 970 + }, + { + "epoch": 0.031207989245246783, + "grad_norm": 1.3910199403762817, + "learning_rate": 4.992688678332437e-05, + "loss": 0.7891, + "num_input_tokens_seen": 1483600, + "step": 975 + }, + { + "epoch": 0.031368030215735226, + "grad_norm": 2.7203726768493652, + "learning_rate": 4.992613457837884e-05, + "loss": 0.8621, + "num_input_tokens_seen": 1490976, + "step": 980 + }, + { + "epoch": 0.03152807118622367, + "grad_norm": 1.0623961687088013, + "learning_rate": 4.992537852951011e-05, + "loss": 0.8893, + "num_input_tokens_seen": 1498672, + "step": 985 + }, + { + "epoch": 0.03168811215671212, + "grad_norm": 1.7467058897018433, + "learning_rate": 4.9924618636834785e-05, + "loss": 0.8569, + "num_input_tokens_seen": 1506704, + "step": 990 + }, + { + "epoch": 0.03184815312720057, + "grad_norm": 1.3380464315414429, + "learning_rate": 4.9923854900470046e-05, + "loss": 0.8427, + "num_input_tokens_seen": 1514144, + "step": 995 + }, + { + "epoch": 0.032008194097689006, + "grad_norm": 2.208622455596924, + "learning_rate": 4.992308732053367e-05, + "loss": 0.9304, + "num_input_tokens_seen": 1521872, + "step": 1000 + }, + { + "epoch": 0.032008194097689006, + "eval_loss": 0.8335565328598022, + "eval_runtime": 558.4937, + "eval_samples_per_second": 24.863, + "eval_steps_per_second": 12.432, + "num_input_tokens_seen": 1521872, + "step": 1000 + }, + { + "epoch": 0.03216823506817745, + "grad_norm": 3.206652879714966, + "learning_rate": 4.992231589714402e-05, + "loss": 0.755, + "num_input_tokens_seen": 1529456, + "step": 1005 + }, + { + "epoch": 0.0323282760386659, + "grad_norm": 2.341054677963257, + "learning_rate": 4.992154063042007e-05, + "loss": 0.8588, + "num_input_tokens_seen": 1537168, + "step": 1010 + }, + { + "epoch": 0.03248831700915435, + "grad_norm": 1.0668998956680298, + "learning_rate": 4.992076152048136e-05, + "loss": 0.6001, + "num_input_tokens_seen": 1544384, + "step": 1015 + }, + { + "epoch": 0.032648357979642786, + "grad_norm": 1.858285903930664, + "learning_rate": 4.991997856744807e-05, + "loss": 1.0006, + "num_input_tokens_seen": 1551856, + "step": 1020 + }, + { + "epoch": 0.03280839895013123, + "grad_norm": 1.6690374612808228, + "learning_rate": 4.9919191771440905e-05, + "loss": 0.8955, + "num_input_tokens_seen": 1559152, + "step": 1025 + }, + { + "epoch": 0.03296843992061968, + "grad_norm": 1.0571908950805664, + "learning_rate": 4.991840113258122e-05, + "loss": 0.664, + "num_input_tokens_seen": 1566800, + "step": 1030 + }, + { + "epoch": 0.03312848089110813, + "grad_norm": 2.072415828704834, + "learning_rate": 4.9917606650990933e-05, + "loss": 0.6651, + "num_input_tokens_seen": 1574656, + "step": 1035 + }, + { + "epoch": 0.033288521861596566, + "grad_norm": 3.080324172973633, + "learning_rate": 4.9916808326792566e-05, + "loss": 0.8811, + "num_input_tokens_seen": 1582688, + "step": 1040 + }, + { + "epoch": 0.03344856283208501, + "grad_norm": 1.7415390014648438, + "learning_rate": 4.9916006160109235e-05, + "loss": 0.866, + "num_input_tokens_seen": 1590064, + "step": 1045 + }, + { + "epoch": 0.03360860380257346, + "grad_norm": 2.4193646907806396, + "learning_rate": 4.991520015106464e-05, + "loss": 1.1476, + "num_input_tokens_seen": 1597872, + "step": 1050 + }, + { + "epoch": 0.03376864477306191, + "grad_norm": 1.4254292249679565, + "learning_rate": 4.991439029978308e-05, + "loss": 0.624, + "num_input_tokens_seen": 1605264, + "step": 1055 + }, + { + "epoch": 0.033928685743550346, + "grad_norm": 1.1801148653030396, + "learning_rate": 4.9913576606389434e-05, + "loss": 0.8321, + "num_input_tokens_seen": 1612528, + "step": 1060 + }, + { + "epoch": 0.03408872671403879, + "grad_norm": 1.6129988431930542, + "learning_rate": 4.991275907100919e-05, + "loss": 0.8689, + "num_input_tokens_seen": 1619936, + "step": 1065 + }, + { + "epoch": 0.03424876768452724, + "grad_norm": 1.676758050918579, + "learning_rate": 4.9911937693768434e-05, + "loss": 0.6225, + "num_input_tokens_seen": 1627488, + "step": 1070 + }, + { + "epoch": 0.03440880865501569, + "grad_norm": 0.7639420032501221, + "learning_rate": 4.991111247479382e-05, + "loss": 0.827, + "num_input_tokens_seen": 1635552, + "step": 1075 + }, + { + "epoch": 0.034568849625504126, + "grad_norm": 1.8965637683868408, + "learning_rate": 4.9910283414212605e-05, + "loss": 0.667, + "num_input_tokens_seen": 1643200, + "step": 1080 + }, + { + "epoch": 0.03472889059599257, + "grad_norm": 2.7052338123321533, + "learning_rate": 4.990945051215265e-05, + "loss": 0.9135, + "num_input_tokens_seen": 1650816, + "step": 1085 + }, + { + "epoch": 0.03488893156648102, + "grad_norm": 1.5858641862869263, + "learning_rate": 4.99086137687424e-05, + "loss": 0.8071, + "num_input_tokens_seen": 1658048, + "step": 1090 + }, + { + "epoch": 0.03504897253696947, + "grad_norm": 0.956683874130249, + "learning_rate": 4.9907773184110874e-05, + "loss": 0.6818, + "num_input_tokens_seen": 1665712, + "step": 1095 + }, + { + "epoch": 0.035209013507457906, + "grad_norm": 1.285988688468933, + "learning_rate": 4.9906928758387715e-05, + "loss": 0.692, + "num_input_tokens_seen": 1673920, + "step": 1100 + }, + { + "epoch": 0.03536905447794635, + "grad_norm": 1.185224175453186, + "learning_rate": 4.9906080491703146e-05, + "loss": 0.9163, + "num_input_tokens_seen": 1681664, + "step": 1105 + }, + { + "epoch": 0.0355290954484348, + "grad_norm": 2.5889346599578857, + "learning_rate": 4.990522838418797e-05, + "loss": 0.8389, + "num_input_tokens_seen": 1688880, + "step": 1110 + }, + { + "epoch": 0.03568913641892325, + "grad_norm": 2.5011978149414062, + "learning_rate": 4.9904372435973604e-05, + "loss": 0.854, + "num_input_tokens_seen": 1696352, + "step": 1115 + }, + { + "epoch": 0.035849177389411686, + "grad_norm": 1.8850194215774536, + "learning_rate": 4.990351264719203e-05, + "loss": 0.7508, + "num_input_tokens_seen": 1704288, + "step": 1120 + }, + { + "epoch": 0.03600921835990013, + "grad_norm": 1.3049675226211548, + "learning_rate": 4.990264901797586e-05, + "loss": 0.6121, + "num_input_tokens_seen": 1711824, + "step": 1125 + }, + { + "epoch": 0.03616925933038858, + "grad_norm": 2.1335537433624268, + "learning_rate": 4.990178154845826e-05, + "loss": 0.9845, + "num_input_tokens_seen": 1719488, + "step": 1130 + }, + { + "epoch": 0.03632930030087703, + "grad_norm": 1.3001476526260376, + "learning_rate": 4.9900910238773014e-05, + "loss": 0.9913, + "num_input_tokens_seen": 1726656, + "step": 1135 + }, + { + "epoch": 0.036489341271365466, + "grad_norm": 1.6774969100952148, + "learning_rate": 4.990003508905448e-05, + "loss": 0.7678, + "num_input_tokens_seen": 1734192, + "step": 1140 + }, + { + "epoch": 0.03664938224185391, + "grad_norm": 1.2213107347488403, + "learning_rate": 4.989915609943763e-05, + "loss": 0.8079, + "num_input_tokens_seen": 1741472, + "step": 1145 + }, + { + "epoch": 0.03680942321234236, + "grad_norm": 1.7149819135665894, + "learning_rate": 4.9898273270058e-05, + "loss": 0.5698, + "num_input_tokens_seen": 1749136, + "step": 1150 + }, + { + "epoch": 0.03696946418283081, + "grad_norm": 1.7412751913070679, + "learning_rate": 4.989738660105174e-05, + "loss": 0.8898, + "num_input_tokens_seen": 1756416, + "step": 1155 + }, + { + "epoch": 0.037129505153319246, + "grad_norm": 0.6843770146369934, + "learning_rate": 4.989649609255559e-05, + "loss": 0.5911, + "num_input_tokens_seen": 1764224, + "step": 1160 + }, + { + "epoch": 0.03728954612380769, + "grad_norm": 0.6784584522247314, + "learning_rate": 4.989560174470687e-05, + "loss": 0.444, + "num_input_tokens_seen": 1771728, + "step": 1165 + }, + { + "epoch": 0.03744958709429614, + "grad_norm": 1.2379224300384521, + "learning_rate": 4.989470355764351e-05, + "loss": 0.6531, + "num_input_tokens_seen": 1778928, + "step": 1170 + }, + { + "epoch": 0.03760962806478459, + "grad_norm": 0.893855631351471, + "learning_rate": 4.9893801531504e-05, + "loss": 0.5867, + "num_input_tokens_seen": 1786416, + "step": 1175 + }, + { + "epoch": 0.03776966903527303, + "grad_norm": 1.086013674736023, + "learning_rate": 4.9892895666427475e-05, + "loss": 0.9766, + "num_input_tokens_seen": 1794064, + "step": 1180 + }, + { + "epoch": 0.03792971000576147, + "grad_norm": 2.6469979286193848, + "learning_rate": 4.9891985962553606e-05, + "loss": 0.7513, + "num_input_tokens_seen": 1801568, + "step": 1185 + }, + { + "epoch": 0.03808975097624992, + "grad_norm": 1.3699606657028198, + "learning_rate": 4.989107242002269e-05, + "loss": 0.5782, + "num_input_tokens_seen": 1809104, + "step": 1190 + }, + { + "epoch": 0.038249791946738367, + "grad_norm": 2.160222291946411, + "learning_rate": 4.989015503897561e-05, + "loss": 0.8715, + "num_input_tokens_seen": 1816880, + "step": 1195 + }, + { + "epoch": 0.03840983291722681, + "grad_norm": 2.133535623550415, + "learning_rate": 4.988923381955383e-05, + "loss": 0.8057, + "num_input_tokens_seen": 1824176, + "step": 1200 + }, + { + "epoch": 0.03840983291722681, + "eval_loss": 0.7928249835968018, + "eval_runtime": 557.0757, + "eval_samples_per_second": 24.927, + "eval_steps_per_second": 12.463, + "num_input_tokens_seen": 1824176, + "step": 1200 + }, + { + "epoch": 0.03856987388771525, + "grad_norm": 1.5663657188415527, + "learning_rate": 4.988830876189942e-05, + "loss": 0.6444, + "num_input_tokens_seen": 1831840, + "step": 1205 + }, + { + "epoch": 0.0387299148582037, + "grad_norm": 2.9660449028015137, + "learning_rate": 4.988737986615503e-05, + "loss": 0.7189, + "num_input_tokens_seen": 1839360, + "step": 1210 + }, + { + "epoch": 0.038889955828692147, + "grad_norm": 0.984630823135376, + "learning_rate": 4.988644713246391e-05, + "loss": 0.8773, + "num_input_tokens_seen": 1846848, + "step": 1215 + }, + { + "epoch": 0.03904999679918059, + "grad_norm": 3.0257816314697266, + "learning_rate": 4.988551056096991e-05, + "loss": 0.8488, + "num_input_tokens_seen": 1854336, + "step": 1220 + }, + { + "epoch": 0.03921003776966903, + "grad_norm": 1.322302222251892, + "learning_rate": 4.988457015181743e-05, + "loss": 0.6316, + "num_input_tokens_seen": 1862144, + "step": 1225 + }, + { + "epoch": 0.03937007874015748, + "grad_norm": 1.8838233947753906, + "learning_rate": 4.988362590515153e-05, + "loss": 0.533, + "num_input_tokens_seen": 1869840, + "step": 1230 + }, + { + "epoch": 0.039530119710645927, + "grad_norm": 1.0223157405853271, + "learning_rate": 4.9882677821117805e-05, + "loss": 0.9243, + "num_input_tokens_seen": 1877632, + "step": 1235 + }, + { + "epoch": 0.03969016068113437, + "grad_norm": 1.1609458923339844, + "learning_rate": 4.988172589986246e-05, + "loss": 0.6809, + "num_input_tokens_seen": 1884912, + "step": 1240 + }, + { + "epoch": 0.03985020165162281, + "grad_norm": 3.0800514221191406, + "learning_rate": 4.9880770141532304e-05, + "loss": 0.755, + "num_input_tokens_seen": 1892368, + "step": 1245 + }, + { + "epoch": 0.04001024262211126, + "grad_norm": 1.113801121711731, + "learning_rate": 4.987981054627472e-05, + "loss": 0.9486, + "num_input_tokens_seen": 1899888, + "step": 1250 + }, + { + "epoch": 0.040170283592599706, + "grad_norm": 1.1060446500778198, + "learning_rate": 4.987884711423769e-05, + "loss": 0.6861, + "num_input_tokens_seen": 1907472, + "step": 1255 + }, + { + "epoch": 0.04033032456308815, + "grad_norm": 1.6990107297897339, + "learning_rate": 4.9877879845569784e-05, + "loss": 0.656, + "num_input_tokens_seen": 1915360, + "step": 1260 + }, + { + "epoch": 0.04049036553357659, + "grad_norm": 1.4322165250778198, + "learning_rate": 4.9876908740420175e-05, + "loss": 0.6079, + "num_input_tokens_seen": 1923104, + "step": 1265 + }, + { + "epoch": 0.04065040650406504, + "grad_norm": 1.2743746042251587, + "learning_rate": 4.987593379893861e-05, + "loss": 0.5705, + "num_input_tokens_seen": 1930368, + "step": 1270 + }, + { + "epoch": 0.040810447474553486, + "grad_norm": 1.6982921361923218, + "learning_rate": 4.987495502127545e-05, + "loss": 0.7355, + "num_input_tokens_seen": 1938032, + "step": 1275 + }, + { + "epoch": 0.04097048844504193, + "grad_norm": 1.3005919456481934, + "learning_rate": 4.987397240758162e-05, + "loss": 0.7805, + "num_input_tokens_seen": 1945408, + "step": 1280 + }, + { + "epoch": 0.04113052941553037, + "grad_norm": 1.3845946788787842, + "learning_rate": 4.9872985958008664e-05, + "loss": 0.7244, + "num_input_tokens_seen": 1953376, + "step": 1285 + }, + { + "epoch": 0.04129057038601882, + "grad_norm": 1.6271398067474365, + "learning_rate": 4.987199567270871e-05, + "loss": 0.788, + "num_input_tokens_seen": 1961024, + "step": 1290 + }, + { + "epoch": 0.041450611356507266, + "grad_norm": 2.136662244796753, + "learning_rate": 4.9871001551834444e-05, + "loss": 0.5385, + "num_input_tokens_seen": 1968608, + "step": 1295 + }, + { + "epoch": 0.04161065232699571, + "grad_norm": 2.8038103580474854, + "learning_rate": 4.98700035955392e-05, + "loss": 1.0798, + "num_input_tokens_seen": 1975968, + "step": 1300 + }, + { + "epoch": 0.04177069329748415, + "grad_norm": 1.5076278448104858, + "learning_rate": 4.986900180397686e-05, + "loss": 0.7842, + "num_input_tokens_seen": 1983664, + "step": 1305 + }, + { + "epoch": 0.0419307342679726, + "grad_norm": 1.049155354499817, + "learning_rate": 4.9867996177301926e-05, + "loss": 0.8062, + "num_input_tokens_seen": 1990896, + "step": 1310 + }, + { + "epoch": 0.042090775238461046, + "grad_norm": 1.463449478149414, + "learning_rate": 4.9866986715669464e-05, + "loss": 0.7675, + "num_input_tokens_seen": 1998496, + "step": 1315 + }, + { + "epoch": 0.04225081620894949, + "grad_norm": 2.2500691413879395, + "learning_rate": 4.9865973419235155e-05, + "loss": 0.6767, + "num_input_tokens_seen": 2006784, + "step": 1320 + }, + { + "epoch": 0.04241085717943793, + "grad_norm": 1.8245335817337036, + "learning_rate": 4.986495628815526e-05, + "loss": 0.7749, + "num_input_tokens_seen": 2014512, + "step": 1325 + }, + { + "epoch": 0.04257089814992638, + "grad_norm": 2.751232624053955, + "learning_rate": 4.986393532258663e-05, + "loss": 0.5864, + "num_input_tokens_seen": 2021936, + "step": 1330 + }, + { + "epoch": 0.042730939120414826, + "grad_norm": 1.1088682413101196, + "learning_rate": 4.986291052268671e-05, + "loss": 0.9221, + "num_input_tokens_seen": 2029824, + "step": 1335 + }, + { + "epoch": 0.04289098009090327, + "grad_norm": 0.9636037349700928, + "learning_rate": 4.986188188861355e-05, + "loss": 0.6011, + "num_input_tokens_seen": 2036976, + "step": 1340 + }, + { + "epoch": 0.04305102106139171, + "grad_norm": 1.1649733781814575, + "learning_rate": 4.9860849420525766e-05, + "loss": 0.9188, + "num_input_tokens_seen": 2044384, + "step": 1345 + }, + { + "epoch": 0.04321106203188016, + "grad_norm": 1.9049980640411377, + "learning_rate": 4.9859813118582575e-05, + "loss": 0.654, + "num_input_tokens_seen": 2052368, + "step": 1350 + }, + { + "epoch": 0.043371103002368606, + "grad_norm": 1.0249799489974976, + "learning_rate": 4.98587729829438e-05, + "loss": 0.7784, + "num_input_tokens_seen": 2059696, + "step": 1355 + }, + { + "epoch": 0.04353114397285705, + "grad_norm": 1.4978278875350952, + "learning_rate": 4.985772901376983e-05, + "loss": 0.4694, + "num_input_tokens_seen": 2067008, + "step": 1360 + }, + { + "epoch": 0.0436911849433455, + "grad_norm": 1.6264230012893677, + "learning_rate": 4.9856681211221666e-05, + "loss": 0.908, + "num_input_tokens_seen": 2074176, + "step": 1365 + }, + { + "epoch": 0.04385122591383394, + "grad_norm": 2.0693411827087402, + "learning_rate": 4.985562957546089e-05, + "loss": 0.885, + "num_input_tokens_seen": 2081840, + "step": 1370 + }, + { + "epoch": 0.044011266884322386, + "grad_norm": 1.2779563665390015, + "learning_rate": 4.9854574106649686e-05, + "loss": 0.7551, + "num_input_tokens_seen": 2089296, + "step": 1375 + }, + { + "epoch": 0.04417130785481083, + "grad_norm": 1.4726965427398682, + "learning_rate": 4.985351480495081e-05, + "loss": 0.4797, + "num_input_tokens_seen": 2097328, + "step": 1380 + }, + { + "epoch": 0.04433134882529928, + "grad_norm": 0.9137635827064514, + "learning_rate": 4.985245167052762e-05, + "loss": 0.6166, + "num_input_tokens_seen": 2104800, + "step": 1385 + }, + { + "epoch": 0.04449138979578772, + "grad_norm": 1.4751914739608765, + "learning_rate": 4.9851384703544066e-05, + "loss": 0.7601, + "num_input_tokens_seen": 2112224, + "step": 1390 + }, + { + "epoch": 0.044651430766276166, + "grad_norm": 1.218609094619751, + "learning_rate": 4.985031390416469e-05, + "loss": 0.7642, + "num_input_tokens_seen": 2119712, + "step": 1395 + }, + { + "epoch": 0.04481147173676461, + "grad_norm": 2.2845797538757324, + "learning_rate": 4.984923927255461e-05, + "loss": 0.7164, + "num_input_tokens_seen": 2127280, + "step": 1400 + }, + { + "epoch": 0.04481147173676461, + "eval_loss": 0.7638634443283081, + "eval_runtime": 560.1161, + "eval_samples_per_second": 24.791, + "eval_steps_per_second": 12.396, + "num_input_tokens_seen": 2127280, + "step": 1400 + }, + { + "epoch": 0.04497151270725306, + "grad_norm": 2.015341281890869, + "learning_rate": 4.984816080887958e-05, + "loss": 0.6653, + "num_input_tokens_seen": 2134576, + "step": 1405 + }, + { + "epoch": 0.0451315536777415, + "grad_norm": 1.4943286180496216, + "learning_rate": 4.9847078513305875e-05, + "loss": 0.7899, + "num_input_tokens_seen": 2142176, + "step": 1410 + }, + { + "epoch": 0.045291594648229946, + "grad_norm": 1.3919262886047363, + "learning_rate": 4.984599238600043e-05, + "loss": 0.5856, + "num_input_tokens_seen": 2149376, + "step": 1415 + }, + { + "epoch": 0.04545163561871839, + "grad_norm": 2.316739559173584, + "learning_rate": 4.9844902427130716e-05, + "loss": 0.4997, + "num_input_tokens_seen": 2156752, + "step": 1420 + }, + { + "epoch": 0.04561167658920684, + "grad_norm": 1.5163812637329102, + "learning_rate": 4.984380863686482e-05, + "loss": 0.6107, + "num_input_tokens_seen": 2163920, + "step": 1425 + }, + { + "epoch": 0.04577171755969528, + "grad_norm": 1.3973745107650757, + "learning_rate": 4.984271101537143e-05, + "loss": 0.5372, + "num_input_tokens_seen": 2171840, + "step": 1430 + }, + { + "epoch": 0.045931758530183726, + "grad_norm": 1.4885987043380737, + "learning_rate": 4.9841609562819816e-05, + "loss": 0.6282, + "num_input_tokens_seen": 2179136, + "step": 1435 + }, + { + "epoch": 0.04609179950067217, + "grad_norm": 2.397493362426758, + "learning_rate": 4.984050427937983e-05, + "loss": 0.6079, + "num_input_tokens_seen": 2186656, + "step": 1440 + }, + { + "epoch": 0.04625184047116062, + "grad_norm": 2.2574288845062256, + "learning_rate": 4.983939516522191e-05, + "loss": 0.6002, + "num_input_tokens_seen": 2194640, + "step": 1445 + }, + { + "epoch": 0.04641188144164906, + "grad_norm": 1.211082100868225, + "learning_rate": 4.983828222051711e-05, + "loss": 0.7734, + "num_input_tokens_seen": 2202480, + "step": 1450 + }, + { + "epoch": 0.046571922412137506, + "grad_norm": 0.8040353655815125, + "learning_rate": 4.983716544543705e-05, + "loss": 0.5544, + "num_input_tokens_seen": 2209952, + "step": 1455 + }, + { + "epoch": 0.04673196338262595, + "grad_norm": 1.8620191812515259, + "learning_rate": 4.983604484015395e-05, + "loss": 0.9031, + "num_input_tokens_seen": 2217376, + "step": 1460 + }, + { + "epoch": 0.0468920043531144, + "grad_norm": 1.5877915620803833, + "learning_rate": 4.983492040484064e-05, + "loss": 0.669, + "num_input_tokens_seen": 2224912, + "step": 1465 + }, + { + "epoch": 0.04705204532360284, + "grad_norm": 0.9423280954360962, + "learning_rate": 4.98337921396705e-05, + "loss": 0.4486, + "num_input_tokens_seen": 2232160, + "step": 1470 + }, + { + "epoch": 0.047212086294091286, + "grad_norm": 1.864829421043396, + "learning_rate": 4.983266004481753e-05, + "loss": 0.986, + "num_input_tokens_seen": 2239616, + "step": 1475 + }, + { + "epoch": 0.04737212726457973, + "grad_norm": 1.0964481830596924, + "learning_rate": 4.9831524120456316e-05, + "loss": 0.6694, + "num_input_tokens_seen": 2247648, + "step": 1480 + }, + { + "epoch": 0.04753216823506818, + "grad_norm": 0.8187300562858582, + "learning_rate": 4.9830384366762026e-05, + "loss": 0.593, + "num_input_tokens_seen": 2254800, + "step": 1485 + }, + { + "epoch": 0.04769220920555662, + "grad_norm": 1.0399445295333862, + "learning_rate": 4.9829240783910436e-05, + "loss": 0.6608, + "num_input_tokens_seen": 2262112, + "step": 1490 + }, + { + "epoch": 0.047852250176045066, + "grad_norm": 3.084890365600586, + "learning_rate": 4.982809337207789e-05, + "loss": 0.8711, + "num_input_tokens_seen": 2269440, + "step": 1495 + }, + { + "epoch": 0.04801229114653351, + "grad_norm": 1.9291945695877075, + "learning_rate": 4.9826942131441337e-05, + "loss": 0.828, + "num_input_tokens_seen": 2276976, + "step": 1500 + }, + { + "epoch": 0.04817233211702196, + "grad_norm": 0.8415671586990356, + "learning_rate": 4.9825787062178315e-05, + "loss": 0.7074, + "num_input_tokens_seen": 2284960, + "step": 1505 + }, + { + "epoch": 0.0483323730875104, + "grad_norm": 1.354689121246338, + "learning_rate": 4.9824628164466945e-05, + "loss": 0.7863, + "num_input_tokens_seen": 2292256, + "step": 1510 + }, + { + "epoch": 0.048492414057998846, + "grad_norm": 1.4506409168243408, + "learning_rate": 4.982346543848595e-05, + "loss": 0.7616, + "num_input_tokens_seen": 2299424, + "step": 1515 + }, + { + "epoch": 0.04865245502848729, + "grad_norm": 1.6359193325042725, + "learning_rate": 4.9822298884414626e-05, + "loss": 0.5073, + "num_input_tokens_seen": 2306512, + "step": 1520 + }, + { + "epoch": 0.04881249599897574, + "grad_norm": 1.3733700513839722, + "learning_rate": 4.982112850243288e-05, + "loss": 0.8305, + "num_input_tokens_seen": 2314768, + "step": 1525 + }, + { + "epoch": 0.04897253696946418, + "grad_norm": 1.908233880996704, + "learning_rate": 4.98199542927212e-05, + "loss": 0.8627, + "num_input_tokens_seen": 2321744, + "step": 1530 + }, + { + "epoch": 0.049132577939952626, + "grad_norm": 1.4428960084915161, + "learning_rate": 4.981877625546066e-05, + "loss": 0.6811, + "num_input_tokens_seen": 2329568, + "step": 1535 + }, + { + "epoch": 0.04929261891044107, + "grad_norm": 1.1684117317199707, + "learning_rate": 4.981759439083293e-05, + "loss": 0.7429, + "num_input_tokens_seen": 2336720, + "step": 1540 + }, + { + "epoch": 0.04945265988092952, + "grad_norm": 1.9972771406173706, + "learning_rate": 4.981640869902027e-05, + "loss": 0.5941, + "num_input_tokens_seen": 2344336, + "step": 1545 + }, + { + "epoch": 0.04961270085141796, + "grad_norm": 1.3218647241592407, + "learning_rate": 4.9815219180205517e-05, + "loss": 0.6112, + "num_input_tokens_seen": 2351904, + "step": 1550 + }, + { + "epoch": 0.049772741821906406, + "grad_norm": 1.2618064880371094, + "learning_rate": 4.9814025834572126e-05, + "loss": 0.5963, + "num_input_tokens_seen": 2359584, + "step": 1555 + }, + { + "epoch": 0.04993278279239485, + "grad_norm": 1.2737226486206055, + "learning_rate": 4.981282866230411e-05, + "loss": 0.7888, + "num_input_tokens_seen": 2366944, + "step": 1560 + }, + { + "epoch": 0.0500928237628833, + "grad_norm": 1.019113540649414, + "learning_rate": 4.981162766358611e-05, + "loss": 0.7165, + "num_input_tokens_seen": 2374544, + "step": 1565 + }, + { + "epoch": 0.05025286473337175, + "grad_norm": 1.4670021533966064, + "learning_rate": 4.9810422838603316e-05, + "loss": 0.7257, + "num_input_tokens_seen": 2382048, + "step": 1570 + }, + { + "epoch": 0.050412905703860186, + "grad_norm": 1.3955533504486084, + "learning_rate": 4.9809214187541533e-05, + "loss": 0.5952, + "num_input_tokens_seen": 2390224, + "step": 1575 + }, + { + "epoch": 0.05057294667434863, + "grad_norm": 1.8718990087509155, + "learning_rate": 4.980800171058715e-05, + "loss": 0.716, + "num_input_tokens_seen": 2397856, + "step": 1580 + }, + { + "epoch": 0.05073298764483708, + "grad_norm": 1.5863887071609497, + "learning_rate": 4.980678540792715e-05, + "loss": 0.8013, + "num_input_tokens_seen": 2405648, + "step": 1585 + }, + { + "epoch": 0.050893028615325527, + "grad_norm": 0.42810407280921936, + "learning_rate": 4.980556527974909e-05, + "loss": 0.688, + "num_input_tokens_seen": 2414048, + "step": 1590 + }, + { + "epoch": 0.051053069585813966, + "grad_norm": 2.246030807495117, + "learning_rate": 4.980434132624114e-05, + "loss": 0.8053, + "num_input_tokens_seen": 2421776, + "step": 1595 + }, + { + "epoch": 0.05121311055630241, + "grad_norm": 0.9319208264350891, + "learning_rate": 4.980311354759205e-05, + "loss": 0.7748, + "num_input_tokens_seen": 2429264, + "step": 1600 + }, + { + "epoch": 0.05121311055630241, + "eval_loss": 0.7435381412506104, + "eval_runtime": 561.6194, + "eval_samples_per_second": 24.725, + "eval_steps_per_second": 12.362, + "num_input_tokens_seen": 2429264, + "step": 1600 + }, + { + "epoch": 0.05137315152679086, + "grad_norm": 1.8753480911254883, + "learning_rate": 4.980188194399116e-05, + "loss": 0.8378, + "num_input_tokens_seen": 2436912, + "step": 1605 + }, + { + "epoch": 0.051533192497279307, + "grad_norm": 1.0900638103485107, + "learning_rate": 4.9800646515628384e-05, + "loss": 0.5617, + "num_input_tokens_seen": 2444976, + "step": 1610 + }, + { + "epoch": 0.051693233467767746, + "grad_norm": 1.9576271772384644, + "learning_rate": 4.979940726269426e-05, + "loss": 0.6426, + "num_input_tokens_seen": 2452608, + "step": 1615 + }, + { + "epoch": 0.05185327443825619, + "grad_norm": 1.6444047689437866, + "learning_rate": 4.979816418537988e-05, + "loss": 0.7466, + "num_input_tokens_seen": 2460256, + "step": 1620 + }, + { + "epoch": 0.05201331540874464, + "grad_norm": 2.022545099258423, + "learning_rate": 4.979691728387696e-05, + "loss": 0.8761, + "num_input_tokens_seen": 2467664, + "step": 1625 + }, + { + "epoch": 0.052173356379233086, + "grad_norm": 1.456283450126648, + "learning_rate": 4.979566655837776e-05, + "loss": 0.7557, + "num_input_tokens_seen": 2475600, + "step": 1630 + }, + { + "epoch": 0.052333397349721526, + "grad_norm": 1.1050891876220703, + "learning_rate": 4.9794412009075184e-05, + "loss": 0.6215, + "num_input_tokens_seen": 2483024, + "step": 1635 + }, + { + "epoch": 0.05249343832020997, + "grad_norm": 0.4970279932022095, + "learning_rate": 4.979315363616269e-05, + "loss": 0.5771, + "num_input_tokens_seen": 2490640, + "step": 1640 + }, + { + "epoch": 0.05265347929069842, + "grad_norm": 1.389966368675232, + "learning_rate": 4.979189143983434e-05, + "loss": 0.5836, + "num_input_tokens_seen": 2498176, + "step": 1645 + }, + { + "epoch": 0.052813520261186866, + "grad_norm": 1.3919246196746826, + "learning_rate": 4.979062542028478e-05, + "loss": 0.6701, + "num_input_tokens_seen": 2505664, + "step": 1650 + }, + { + "epoch": 0.052973561231675306, + "grad_norm": 1.6933093070983887, + "learning_rate": 4.978935557770923e-05, + "loss": 0.6979, + "num_input_tokens_seen": 2513664, + "step": 1655 + }, + { + "epoch": 0.05313360220216375, + "grad_norm": 1.1605689525604248, + "learning_rate": 4.978808191230353e-05, + "loss": 0.7613, + "num_input_tokens_seen": 2521168, + "step": 1660 + }, + { + "epoch": 0.0532936431726522, + "grad_norm": 1.1663812398910522, + "learning_rate": 4.9786804424264085e-05, + "loss": 0.7723, + "num_input_tokens_seen": 2529072, + "step": 1665 + }, + { + "epoch": 0.053453684143140646, + "grad_norm": 0.9593190550804138, + "learning_rate": 4.978552311378792e-05, + "loss": 0.6627, + "num_input_tokens_seen": 2536336, + "step": 1670 + }, + { + "epoch": 0.053613725113629086, + "grad_norm": 1.4924570322036743, + "learning_rate": 4.978423798107261e-05, + "loss": 0.579, + "num_input_tokens_seen": 2544080, + "step": 1675 + }, + { + "epoch": 0.05377376608411753, + "grad_norm": 1.3037962913513184, + "learning_rate": 4.978294902631635e-05, + "loss": 0.4295, + "num_input_tokens_seen": 2552000, + "step": 1680 + }, + { + "epoch": 0.05393380705460598, + "grad_norm": 1.5921368598937988, + "learning_rate": 4.9781656249717914e-05, + "loss": 1.0094, + "num_input_tokens_seen": 2559920, + "step": 1685 + }, + { + "epoch": 0.054093848025094426, + "grad_norm": 1.5157859325408936, + "learning_rate": 4.9780359651476645e-05, + "loss": 0.572, + "num_input_tokens_seen": 2567456, + "step": 1690 + }, + { + "epoch": 0.054253888995582866, + "grad_norm": 0.7663775682449341, + "learning_rate": 4.977905923179251e-05, + "loss": 0.5333, + "num_input_tokens_seen": 2575184, + "step": 1695 + }, + { + "epoch": 0.05441392996607131, + "grad_norm": 3.180962324142456, + "learning_rate": 4.977775499086606e-05, + "loss": 1.3871, + "num_input_tokens_seen": 2583536, + "step": 1700 + }, + { + "epoch": 0.05457397093655976, + "grad_norm": 0.8067048192024231, + "learning_rate": 4.97764469288984e-05, + "loss": 0.4374, + "num_input_tokens_seen": 2591264, + "step": 1705 + }, + { + "epoch": 0.054734011907048206, + "grad_norm": 2.257223129272461, + "learning_rate": 4.977513504609127e-05, + "loss": 0.6892, + "num_input_tokens_seen": 2599216, + "step": 1710 + }, + { + "epoch": 0.054894052877536646, + "grad_norm": 1.505334496498108, + "learning_rate": 4.9773819342646965e-05, + "loss": 0.7694, + "num_input_tokens_seen": 2606896, + "step": 1715 + }, + { + "epoch": 0.05505409384802509, + "grad_norm": 1.6683048009872437, + "learning_rate": 4.97724998187684e-05, + "loss": 0.8437, + "num_input_tokens_seen": 2615312, + "step": 1720 + }, + { + "epoch": 0.05521413481851354, + "grad_norm": 1.042114496231079, + "learning_rate": 4.9771176474659045e-05, + "loss": 0.5278, + "num_input_tokens_seen": 2622672, + "step": 1725 + }, + { + "epoch": 0.055374175789001986, + "grad_norm": 0.7056401968002319, + "learning_rate": 4.976984931052299e-05, + "loss": 0.5478, + "num_input_tokens_seen": 2630384, + "step": 1730 + }, + { + "epoch": 0.055534216759490426, + "grad_norm": 0.8735511898994446, + "learning_rate": 4.976851832656489e-05, + "loss": 0.4841, + "num_input_tokens_seen": 2638064, + "step": 1735 + }, + { + "epoch": 0.05569425772997887, + "grad_norm": 1.4991618394851685, + "learning_rate": 4.9767183522990004e-05, + "loss": 0.4904, + "num_input_tokens_seen": 2645648, + "step": 1740 + }, + { + "epoch": 0.05585429870046732, + "grad_norm": 1.206368327140808, + "learning_rate": 4.9765844900004176e-05, + "loss": 0.6767, + "num_input_tokens_seen": 2653696, + "step": 1745 + }, + { + "epoch": 0.056014339670955766, + "grad_norm": 2.1058995723724365, + "learning_rate": 4.9764502457813834e-05, + "loss": 0.7382, + "num_input_tokens_seen": 2660960, + "step": 1750 + }, + { + "epoch": 0.05617438064144421, + "grad_norm": 0.7663532495498657, + "learning_rate": 4.9763156196626005e-05, + "loss": 0.5501, + "num_input_tokens_seen": 2668528, + "step": 1755 + }, + { + "epoch": 0.05633442161193265, + "grad_norm": 2.6473493576049805, + "learning_rate": 4.97618061166483e-05, + "loss": 0.7845, + "num_input_tokens_seen": 2676112, + "step": 1760 + }, + { + "epoch": 0.0564944625824211, + "grad_norm": 2.1389119625091553, + "learning_rate": 4.9760452218088915e-05, + "loss": 0.7844, + "num_input_tokens_seen": 2683328, + "step": 1765 + }, + { + "epoch": 0.056654503552909546, + "grad_norm": 2.214531660079956, + "learning_rate": 4.975909450115663e-05, + "loss": 0.8992, + "num_input_tokens_seen": 2691008, + "step": 1770 + }, + { + "epoch": 0.05681454452339799, + "grad_norm": 1.6330480575561523, + "learning_rate": 4.975773296606084e-05, + "loss": 0.799, + "num_input_tokens_seen": 2698528, + "step": 1775 + }, + { + "epoch": 0.05697458549388643, + "grad_norm": 1.6369048357009888, + "learning_rate": 4.97563676130115e-05, + "loss": 0.7281, + "num_input_tokens_seen": 2706656, + "step": 1780 + }, + { + "epoch": 0.05713462646437488, + "grad_norm": 0.9995440244674683, + "learning_rate": 4.9754998442219166e-05, + "loss": 0.5487, + "num_input_tokens_seen": 2714336, + "step": 1785 + }, + { + "epoch": 0.057294667434863326, + "grad_norm": 0.7598903179168701, + "learning_rate": 4.9753625453894984e-05, + "loss": 0.5335, + "num_input_tokens_seen": 2722080, + "step": 1790 + }, + { + "epoch": 0.05745470840535177, + "grad_norm": 1.3214234113693237, + "learning_rate": 4.975224864825068e-05, + "loss": 0.9489, + "num_input_tokens_seen": 2729088, + "step": 1795 + }, + { + "epoch": 0.05761474937584021, + "grad_norm": 1.373632550239563, + "learning_rate": 4.9750868025498576e-05, + "loss": 0.5404, + "num_input_tokens_seen": 2736704, + "step": 1800 + }, + { + "epoch": 0.05761474937584021, + "eval_loss": 0.7271351218223572, + "eval_runtime": 559.3339, + "eval_samples_per_second": 24.826, + "eval_steps_per_second": 12.413, + "num_input_tokens_seen": 2736704, + "step": 1800 + }, + { + "epoch": 0.05777479034632866, + "grad_norm": 2.1200175285339355, + "learning_rate": 4.974948358585158e-05, + "loss": 0.7274, + "num_input_tokens_seen": 2743984, + "step": 1805 + }, + { + "epoch": 0.057934831316817106, + "grad_norm": 1.0403584241867065, + "learning_rate": 4.9748095329523205e-05, + "loss": 0.7236, + "num_input_tokens_seen": 2751504, + "step": 1810 + }, + { + "epoch": 0.05809487228730555, + "grad_norm": 0.9181123375892639, + "learning_rate": 4.974670325672752e-05, + "loss": 0.598, + "num_input_tokens_seen": 2759232, + "step": 1815 + }, + { + "epoch": 0.05825491325779399, + "grad_norm": 1.080022931098938, + "learning_rate": 4.974530736767921e-05, + "loss": 0.6557, + "num_input_tokens_seen": 2766720, + "step": 1820 + }, + { + "epoch": 0.05841495422828244, + "grad_norm": 0.6316962242126465, + "learning_rate": 4.9743907662593524e-05, + "loss": 0.5621, + "num_input_tokens_seen": 2773904, + "step": 1825 + }, + { + "epoch": 0.058574995198770886, + "grad_norm": 1.8168349266052246, + "learning_rate": 4.974250414168633e-05, + "loss": 0.6534, + "num_input_tokens_seen": 2781680, + "step": 1830 + }, + { + "epoch": 0.05873503616925933, + "grad_norm": 2.172159433364868, + "learning_rate": 4.974109680517407e-05, + "loss": 0.7675, + "num_input_tokens_seen": 2789376, + "step": 1835 + }, + { + "epoch": 0.05889507713974777, + "grad_norm": 1.4563441276550293, + "learning_rate": 4.973968565327376e-05, + "loss": 0.7137, + "num_input_tokens_seen": 2797008, + "step": 1840 + }, + { + "epoch": 0.05905511811023622, + "grad_norm": 1.3277214765548706, + "learning_rate": 4.973827068620303e-05, + "loss": 0.6361, + "num_input_tokens_seen": 2804832, + "step": 1845 + }, + { + "epoch": 0.059215159080724666, + "grad_norm": 1.778009057044983, + "learning_rate": 4.973685190418008e-05, + "loss": 0.6652, + "num_input_tokens_seen": 2812272, + "step": 1850 + }, + { + "epoch": 0.05937520005121311, + "grad_norm": 1.4020510911941528, + "learning_rate": 4.97354293074237e-05, + "loss": 0.5382, + "num_input_tokens_seen": 2819344, + "step": 1855 + }, + { + "epoch": 0.05953524102170155, + "grad_norm": 1.5278948545455933, + "learning_rate": 4.9734002896153276e-05, + "loss": 0.5811, + "num_input_tokens_seen": 2826944, + "step": 1860 + }, + { + "epoch": 0.05969528199219, + "grad_norm": 0.9363411068916321, + "learning_rate": 4.973257267058877e-05, + "loss": 0.4883, + "num_input_tokens_seen": 2834032, + "step": 1865 + }, + { + "epoch": 0.059855322962678446, + "grad_norm": 1.745566725730896, + "learning_rate": 4.973113863095076e-05, + "loss": 0.6478, + "num_input_tokens_seen": 2842224, + "step": 1870 + }, + { + "epoch": 0.06001536393316689, + "grad_norm": 1.8293681144714355, + "learning_rate": 4.9729700777460384e-05, + "loss": 0.9619, + "num_input_tokens_seen": 2849632, + "step": 1875 + }, + { + "epoch": 0.06017540490365533, + "grad_norm": 1.3044594526290894, + "learning_rate": 4.972825911033937e-05, + "loss": 0.5364, + "num_input_tokens_seen": 2857856, + "step": 1880 + }, + { + "epoch": 0.06033544587414378, + "grad_norm": 1.1253299713134766, + "learning_rate": 4.9726813629810056e-05, + "loss": 0.6707, + "num_input_tokens_seen": 2865168, + "step": 1885 + }, + { + "epoch": 0.060495486844632226, + "grad_norm": 2.3690848350524902, + "learning_rate": 4.9725364336095326e-05, + "loss": 0.8968, + "num_input_tokens_seen": 2872608, + "step": 1890 + }, + { + "epoch": 0.06065552781512067, + "grad_norm": 1.8372266292572021, + "learning_rate": 4.972391122941871e-05, + "loss": 0.5757, + "num_input_tokens_seen": 2879776, + "step": 1895 + }, + { + "epoch": 0.06081556878560911, + "grad_norm": 1.4678893089294434, + "learning_rate": 4.972245431000428e-05, + "loss": 0.7202, + "num_input_tokens_seen": 2887648, + "step": 1900 + }, + { + "epoch": 0.06097560975609756, + "grad_norm": 1.1638588905334473, + "learning_rate": 4.972099357807671e-05, + "loss": 0.7702, + "num_input_tokens_seen": 2895056, + "step": 1905 + }, + { + "epoch": 0.061135650726586006, + "grad_norm": 0.6891557574272156, + "learning_rate": 4.971952903386127e-05, + "loss": 0.4249, + "num_input_tokens_seen": 2902672, + "step": 1910 + }, + { + "epoch": 0.06129569169707445, + "grad_norm": 0.5572391152381897, + "learning_rate": 4.971806067758381e-05, + "loss": 0.6325, + "num_input_tokens_seen": 2910336, + "step": 1915 + }, + { + "epoch": 0.06145573266756289, + "grad_norm": 1.0609261989593506, + "learning_rate": 4.971658850947076e-05, + "loss": 0.6885, + "num_input_tokens_seen": 2917728, + "step": 1920 + }, + { + "epoch": 0.06161577363805134, + "grad_norm": 1.2347691059112549, + "learning_rate": 4.9715112529749165e-05, + "loss": 0.8125, + "num_input_tokens_seen": 2924864, + "step": 1925 + }, + { + "epoch": 0.061775814608539786, + "grad_norm": 1.212631344795227, + "learning_rate": 4.9713632738646624e-05, + "loss": 0.9562, + "num_input_tokens_seen": 2932592, + "step": 1930 + }, + { + "epoch": 0.06193585557902823, + "grad_norm": 1.2359366416931152, + "learning_rate": 4.971214913639134e-05, + "loss": 0.6733, + "num_input_tokens_seen": 2940384, + "step": 1935 + }, + { + "epoch": 0.06209589654951667, + "grad_norm": 0.7703158855438232, + "learning_rate": 4.9710661723212104e-05, + "loss": 0.4534, + "num_input_tokens_seen": 2948176, + "step": 1940 + }, + { + "epoch": 0.06225593752000512, + "grad_norm": 1.5711071491241455, + "learning_rate": 4.9709170499338295e-05, + "loss": 0.8308, + "num_input_tokens_seen": 2955520, + "step": 1945 + }, + { + "epoch": 0.062415978490493566, + "grad_norm": 1.2245604991912842, + "learning_rate": 4.9707675464999895e-05, + "loss": 0.5978, + "num_input_tokens_seen": 2963216, + "step": 1950 + }, + { + "epoch": 0.06257601946098201, + "grad_norm": 0.8701627850532532, + "learning_rate": 4.970617662042743e-05, + "loss": 0.4844, + "num_input_tokens_seen": 2970992, + "step": 1955 + }, + { + "epoch": 0.06273606043147045, + "grad_norm": 1.4940787553787231, + "learning_rate": 4.970467396585206e-05, + "loss": 0.8063, + "num_input_tokens_seen": 2978480, + "step": 1960 + }, + { + "epoch": 0.0628961014019589, + "grad_norm": 1.3265186548233032, + "learning_rate": 4.97031675015055e-05, + "loss": 0.621, + "num_input_tokens_seen": 2985664, + "step": 1965 + }, + { + "epoch": 0.06305614237244735, + "grad_norm": 0.6673404574394226, + "learning_rate": 4.9701657227620075e-05, + "loss": 0.4918, + "num_input_tokens_seen": 2995104, + "step": 1970 + }, + { + "epoch": 0.06321618334293579, + "grad_norm": 1.2663886547088623, + "learning_rate": 4.9700143144428685e-05, + "loss": 0.7268, + "num_input_tokens_seen": 3002992, + "step": 1975 + }, + { + "epoch": 0.06337622431342424, + "grad_norm": 0.8947978615760803, + "learning_rate": 4.969862525216482e-05, + "loss": 0.809, + "num_input_tokens_seen": 3010640, + "step": 1980 + }, + { + "epoch": 0.06353626528391268, + "grad_norm": 1.5146124362945557, + "learning_rate": 4.9697103551062556e-05, + "loss": 0.5073, + "num_input_tokens_seen": 3018304, + "step": 1985 + }, + { + "epoch": 0.06369630625440113, + "grad_norm": 2.6080613136291504, + "learning_rate": 4.9695578041356565e-05, + "loss": 0.8506, + "num_input_tokens_seen": 3026176, + "step": 1990 + }, + { + "epoch": 0.06385634722488957, + "grad_norm": 1.770978331565857, + "learning_rate": 4.969404872328209e-05, + "loss": 0.7308, + "num_input_tokens_seen": 3033600, + "step": 1995 + }, + { + "epoch": 0.06401638819537801, + "grad_norm": 0.8471187353134155, + "learning_rate": 4.969251559707498e-05, + "loss": 0.6634, + "num_input_tokens_seen": 3041072, + "step": 2000 + }, + { + "epoch": 0.06401638819537801, + "eval_loss": 0.710352897644043, + "eval_runtime": 557.7628, + "eval_samples_per_second": 24.896, + "eval_steps_per_second": 12.448, + "num_input_tokens_seen": 3041072, + "step": 2000 + }, + { + "epoch": 0.06417642916586647, + "grad_norm": 0.9771375060081482, + "learning_rate": 4.9690978662971674e-05, + "loss": 0.8289, + "num_input_tokens_seen": 3048960, + "step": 2005 + }, + { + "epoch": 0.0643364701363549, + "grad_norm": 2.339676856994629, + "learning_rate": 4.968943792120916e-05, + "loss": 1.1824, + "num_input_tokens_seen": 3056560, + "step": 2010 + }, + { + "epoch": 0.06449651110684335, + "grad_norm": 0.6018199324607849, + "learning_rate": 4.9687893372025046e-05, + "loss": 0.6644, + "num_input_tokens_seen": 3063664, + "step": 2015 + }, + { + "epoch": 0.0646565520773318, + "grad_norm": 1.251231074333191, + "learning_rate": 4.9686345015657535e-05, + "loss": 0.6913, + "num_input_tokens_seen": 3070992, + "step": 2020 + }, + { + "epoch": 0.06481659304782024, + "grad_norm": 9.561107635498047, + "learning_rate": 4.968479285234538e-05, + "loss": 0.7678, + "num_input_tokens_seen": 3078768, + "step": 2025 + }, + { + "epoch": 0.0649766340183087, + "grad_norm": 0.8474399447441101, + "learning_rate": 4.9683236882327974e-05, + "loss": 0.7194, + "num_input_tokens_seen": 3086336, + "step": 2030 + }, + { + "epoch": 0.06513667498879713, + "grad_norm": 1.3668711185455322, + "learning_rate": 4.968167710584526e-05, + "loss": 0.8077, + "num_input_tokens_seen": 3093776, + "step": 2035 + }, + { + "epoch": 0.06529671595928557, + "grad_norm": 2.5661864280700684, + "learning_rate": 4.968011352313775e-05, + "loss": 1.0129, + "num_input_tokens_seen": 3101152, + "step": 2040 + }, + { + "epoch": 0.06545675692977403, + "grad_norm": 0.7437636852264404, + "learning_rate": 4.967854613444659e-05, + "loss": 0.5876, + "num_input_tokens_seen": 3108912, + "step": 2045 + }, + { + "epoch": 0.06561679790026247, + "grad_norm": 1.0835976600646973, + "learning_rate": 4.967697494001349e-05, + "loss": 0.6761, + "num_input_tokens_seen": 3117120, + "step": 2050 + }, + { + "epoch": 0.0657768388707509, + "grad_norm": 0.8186072707176208, + "learning_rate": 4.9675399940080736e-05, + "loss": 0.7281, + "num_input_tokens_seen": 3124752, + "step": 2055 + }, + { + "epoch": 0.06593687984123936, + "grad_norm": 1.449580430984497, + "learning_rate": 4.9673821134891226e-05, + "loss": 0.7857, + "num_input_tokens_seen": 3132720, + "step": 2060 + }, + { + "epoch": 0.0660969208117278, + "grad_norm": 0.7937209010124207, + "learning_rate": 4.967223852468842e-05, + "loss": 0.7462, + "num_input_tokens_seen": 3139712, + "step": 2065 + }, + { + "epoch": 0.06625696178221625, + "grad_norm": 1.0111793279647827, + "learning_rate": 4.967065210971639e-05, + "loss": 0.7077, + "num_input_tokens_seen": 3147040, + "step": 2070 + }, + { + "epoch": 0.06641700275270469, + "grad_norm": 1.4386954307556152, + "learning_rate": 4.966906189021977e-05, + "loss": 0.6555, + "num_input_tokens_seen": 3154752, + "step": 2075 + }, + { + "epoch": 0.06657704372319313, + "grad_norm": 0.9681957364082336, + "learning_rate": 4.966746786644379e-05, + "loss": 0.6336, + "num_input_tokens_seen": 3162608, + "step": 2080 + }, + { + "epoch": 0.06673708469368159, + "grad_norm": 1.6117857694625854, + "learning_rate": 4.966587003863429e-05, + "loss": 0.6298, + "num_input_tokens_seen": 3170240, + "step": 2085 + }, + { + "epoch": 0.06689712566417003, + "grad_norm": 0.9707816243171692, + "learning_rate": 4.966426840703765e-05, + "loss": 0.6092, + "num_input_tokens_seen": 3178432, + "step": 2090 + }, + { + "epoch": 0.06705716663465847, + "grad_norm": 1.8204681873321533, + "learning_rate": 4.9662662971900875e-05, + "loss": 0.8483, + "num_input_tokens_seen": 3185408, + "step": 2095 + }, + { + "epoch": 0.06721720760514692, + "grad_norm": 1.324836254119873, + "learning_rate": 4.9661053733471534e-05, + "loss": 0.6497, + "num_input_tokens_seen": 3192928, + "step": 2100 + }, + { + "epoch": 0.06737724857563536, + "grad_norm": 1.627789855003357, + "learning_rate": 4.965944069199781e-05, + "loss": 0.7972, + "num_input_tokens_seen": 3200320, + "step": 2105 + }, + { + "epoch": 0.06753728954612381, + "grad_norm": 1.451941728591919, + "learning_rate": 4.965782384772842e-05, + "loss": 0.5882, + "num_input_tokens_seen": 3208096, + "step": 2110 + }, + { + "epoch": 0.06769733051661225, + "grad_norm": 1.5565869808197021, + "learning_rate": 4.9656203200912734e-05, + "loss": 0.8078, + "num_input_tokens_seen": 3216048, + "step": 2115 + }, + { + "epoch": 0.06785737148710069, + "grad_norm": 1.403233528137207, + "learning_rate": 4.965457875180067e-05, + "loss": 0.6944, + "num_input_tokens_seen": 3223536, + "step": 2120 + }, + { + "epoch": 0.06801741245758915, + "grad_norm": 1.7262367010116577, + "learning_rate": 4.9652950500642724e-05, + "loss": 0.7636, + "num_input_tokens_seen": 3231296, + "step": 2125 + }, + { + "epoch": 0.06817745342807759, + "grad_norm": 1.6894452571868896, + "learning_rate": 4.965131844769001e-05, + "loss": 0.8181, + "num_input_tokens_seen": 3238768, + "step": 2130 + }, + { + "epoch": 0.06833749439856603, + "grad_norm": 1.4443962574005127, + "learning_rate": 4.96496825931942e-05, + "loss": 0.5872, + "num_input_tokens_seen": 3246624, + "step": 2135 + }, + { + "epoch": 0.06849753536905448, + "grad_norm": 1.2869515419006348, + "learning_rate": 4.9648042937407566e-05, + "loss": 0.5401, + "num_input_tokens_seen": 3254384, + "step": 2140 + }, + { + "epoch": 0.06865757633954292, + "grad_norm": 0.698890209197998, + "learning_rate": 4.964639948058297e-05, + "loss": 0.612, + "num_input_tokens_seen": 3261680, + "step": 2145 + }, + { + "epoch": 0.06881761731003137, + "grad_norm": 1.0994969606399536, + "learning_rate": 4.9644752222973846e-05, + "loss": 0.859, + "num_input_tokens_seen": 3269184, + "step": 2150 + }, + { + "epoch": 0.06897765828051981, + "grad_norm": 1.757907748222351, + "learning_rate": 4.964310116483422e-05, + "loss": 0.7945, + "num_input_tokens_seen": 3276512, + "step": 2155 + }, + { + "epoch": 0.06913769925100825, + "grad_norm": 0.7746456265449524, + "learning_rate": 4.964144630641872e-05, + "loss": 0.5785, + "num_input_tokens_seen": 3283760, + "step": 2160 + }, + { + "epoch": 0.0692977402214967, + "grad_norm": 0.6792855858802795, + "learning_rate": 4.9639787647982525e-05, + "loss": 0.5062, + "num_input_tokens_seen": 3291616, + "step": 2165 + }, + { + "epoch": 0.06945778119198515, + "grad_norm": 0.5447894334793091, + "learning_rate": 4.963812518978143e-05, + "loss": 0.8386, + "num_input_tokens_seen": 3299056, + "step": 2170 + }, + { + "epoch": 0.0696178221624736, + "grad_norm": 2.4981489181518555, + "learning_rate": 4.963645893207182e-05, + "loss": 0.6584, + "num_input_tokens_seen": 3306768, + "step": 2175 + }, + { + "epoch": 0.06977786313296204, + "grad_norm": 1.322155475616455, + "learning_rate": 4.963478887511063e-05, + "loss": 0.9064, + "num_input_tokens_seen": 3314144, + "step": 2180 + }, + { + "epoch": 0.06993790410345048, + "grad_norm": 3.546212911605835, + "learning_rate": 4.963311501915542e-05, + "loss": 0.6306, + "num_input_tokens_seen": 3321472, + "step": 2185 + }, + { + "epoch": 0.07009794507393893, + "grad_norm": 1.8576436042785645, + "learning_rate": 4.963143736446432e-05, + "loss": 0.6432, + "num_input_tokens_seen": 3328560, + "step": 2190 + }, + { + "epoch": 0.07025798604442737, + "grad_norm": 0.7259161472320557, + "learning_rate": 4.962975591129603e-05, + "loss": 0.5711, + "num_input_tokens_seen": 3335728, + "step": 2195 + }, + { + "epoch": 0.07041802701491581, + "grad_norm": 2.3840646743774414, + "learning_rate": 4.962807065990986e-05, + "loss": 1.0893, + "num_input_tokens_seen": 3343472, + "step": 2200 + }, + { + "epoch": 0.07041802701491581, + "eval_loss": 0.6959899067878723, + "eval_runtime": 557.7537, + "eval_samples_per_second": 24.896, + "eval_steps_per_second": 12.448, + "num_input_tokens_seen": 3343472, + "step": 2200 + }, + { + "epoch": 0.07057806798540427, + "grad_norm": 2.1421287059783936, + "learning_rate": 4.9626381610565714e-05, + "loss": 0.6603, + "num_input_tokens_seen": 3351088, + "step": 2205 + }, + { + "epoch": 0.0707381089558927, + "grad_norm": 1.2309461832046509, + "learning_rate": 4.9624688763524043e-05, + "loss": 0.6881, + "num_input_tokens_seen": 3358448, + "step": 2210 + }, + { + "epoch": 0.07089814992638116, + "grad_norm": 0.698651134967804, + "learning_rate": 4.962299211904591e-05, + "loss": 0.6326, + "num_input_tokens_seen": 3366048, + "step": 2215 + }, + { + "epoch": 0.0710581908968696, + "grad_norm": 1.0310192108154297, + "learning_rate": 4.962129167739296e-05, + "loss": 0.8235, + "num_input_tokens_seen": 3373616, + "step": 2220 + }, + { + "epoch": 0.07121823186735804, + "grad_norm": 1.1600013971328735, + "learning_rate": 4.961958743882742e-05, + "loss": 0.6762, + "num_input_tokens_seen": 3381392, + "step": 2225 + }, + { + "epoch": 0.0713782728378465, + "grad_norm": 0.6028136014938354, + "learning_rate": 4.961787940361211e-05, + "loss": 0.4891, + "num_input_tokens_seen": 3389440, + "step": 2230 + }, + { + "epoch": 0.07153831380833493, + "grad_norm": 1.1262248754501343, + "learning_rate": 4.961616757201043e-05, + "loss": 0.7558, + "num_input_tokens_seen": 3396944, + "step": 2235 + }, + { + "epoch": 0.07169835477882337, + "grad_norm": 1.4422204494476318, + "learning_rate": 4.961445194428637e-05, + "loss": 0.6424, + "num_input_tokens_seen": 3404080, + "step": 2240 + }, + { + "epoch": 0.07185839574931183, + "grad_norm": 1.0507590770721436, + "learning_rate": 4.9612732520704486e-05, + "loss": 0.5838, + "num_input_tokens_seen": 3411936, + "step": 2245 + }, + { + "epoch": 0.07201843671980027, + "grad_norm": 0.7182942032814026, + "learning_rate": 4.961100930152994e-05, + "loss": 0.7582, + "num_input_tokens_seen": 3418976, + "step": 2250 + }, + { + "epoch": 0.07217847769028872, + "grad_norm": 1.6605020761489868, + "learning_rate": 4.960928228702849e-05, + "loss": 0.632, + "num_input_tokens_seen": 3426624, + "step": 2255 + }, + { + "epoch": 0.07233851866077716, + "grad_norm": 1.350947618484497, + "learning_rate": 4.960755147746645e-05, + "loss": 0.5416, + "num_input_tokens_seen": 3434288, + "step": 2260 + }, + { + "epoch": 0.0724985596312656, + "grad_norm": 1.0351059436798096, + "learning_rate": 4.9605816873110736e-05, + "loss": 0.7028, + "num_input_tokens_seen": 3441472, + "step": 2265 + }, + { + "epoch": 0.07265860060175405, + "grad_norm": 2.315316915512085, + "learning_rate": 4.960407847422883e-05, + "loss": 0.7108, + "num_input_tokens_seen": 3449152, + "step": 2270 + }, + { + "epoch": 0.07281864157224249, + "grad_norm": 1.0269198417663574, + "learning_rate": 4.960233628108885e-05, + "loss": 0.768, + "num_input_tokens_seen": 3456560, + "step": 2275 + }, + { + "epoch": 0.07297868254273093, + "grad_norm": 0.8483147025108337, + "learning_rate": 4.960059029395942e-05, + "loss": 0.5954, + "num_input_tokens_seen": 3465200, + "step": 2280 + }, + { + "epoch": 0.07313872351321939, + "grad_norm": 1.2037514448165894, + "learning_rate": 4.959884051310983e-05, + "loss": 0.5992, + "num_input_tokens_seen": 3472976, + "step": 2285 + }, + { + "epoch": 0.07329876448370783, + "grad_norm": 1.4120204448699951, + "learning_rate": 4.959708693880991e-05, + "loss": 0.6447, + "num_input_tokens_seen": 3480496, + "step": 2290 + }, + { + "epoch": 0.07345880545419628, + "grad_norm": 1.0643653869628906, + "learning_rate": 4.9595329571330074e-05, + "loss": 0.7846, + "num_input_tokens_seen": 3488432, + "step": 2295 + }, + { + "epoch": 0.07361884642468472, + "grad_norm": 1.8689343929290771, + "learning_rate": 4.9593568410941326e-05, + "loss": 0.8772, + "num_input_tokens_seen": 3496192, + "step": 2300 + }, + { + "epoch": 0.07377888739517316, + "grad_norm": 1.89470636844635, + "learning_rate": 4.959180345791528e-05, + "loss": 0.6194, + "num_input_tokens_seen": 3503568, + "step": 2305 + }, + { + "epoch": 0.07393892836566161, + "grad_norm": 1.2733662128448486, + "learning_rate": 4.9590034712524086e-05, + "loss": 0.8171, + "num_input_tokens_seen": 3511840, + "step": 2310 + }, + { + "epoch": 0.07409896933615005, + "grad_norm": 1.307349443435669, + "learning_rate": 4.958826217504053e-05, + "loss": 0.5137, + "num_input_tokens_seen": 3519616, + "step": 2315 + }, + { + "epoch": 0.07425901030663849, + "grad_norm": 0.5898303985595703, + "learning_rate": 4.958648584573795e-05, + "loss": 0.4654, + "num_input_tokens_seen": 3527488, + "step": 2320 + }, + { + "epoch": 0.07441905127712695, + "grad_norm": 1.2374564409255981, + "learning_rate": 4.958470572489028e-05, + "loss": 0.6348, + "num_input_tokens_seen": 3535040, + "step": 2325 + }, + { + "epoch": 0.07457909224761539, + "grad_norm": 0.6548512578010559, + "learning_rate": 4.958292181277203e-05, + "loss": 0.6198, + "num_input_tokens_seen": 3542896, + "step": 2330 + }, + { + "epoch": 0.07473913321810384, + "grad_norm": 0.72239750623703, + "learning_rate": 4.958113410965832e-05, + "loss": 0.4818, + "num_input_tokens_seen": 3550080, + "step": 2335 + }, + { + "epoch": 0.07489917418859228, + "grad_norm": 1.5098949670791626, + "learning_rate": 4.957934261582481e-05, + "loss": 0.7095, + "num_input_tokens_seen": 3557216, + "step": 2340 + }, + { + "epoch": 0.07505921515908072, + "grad_norm": 1.0559002161026, + "learning_rate": 4.95775473315478e-05, + "loss": 0.7759, + "num_input_tokens_seen": 3564496, + "step": 2345 + }, + { + "epoch": 0.07521925612956917, + "grad_norm": 1.0072203874588013, + "learning_rate": 4.9575748257104124e-05, + "loss": 0.5999, + "num_input_tokens_seen": 3572096, + "step": 2350 + }, + { + "epoch": 0.07537929710005761, + "grad_norm": 0.8195162415504456, + "learning_rate": 4.9573945392771224e-05, + "loss": 0.5672, + "num_input_tokens_seen": 3579488, + "step": 2355 + }, + { + "epoch": 0.07553933807054607, + "grad_norm": 0.8449749946594238, + "learning_rate": 4.9572138738827134e-05, + "loss": 0.5766, + "num_input_tokens_seen": 3587184, + "step": 2360 + }, + { + "epoch": 0.0756993790410345, + "grad_norm": 1.7060869932174683, + "learning_rate": 4.957032829555046e-05, + "loss": 0.8002, + "num_input_tokens_seen": 3594624, + "step": 2365 + }, + { + "epoch": 0.07585942001152295, + "grad_norm": 1.1193716526031494, + "learning_rate": 4.956851406322039e-05, + "loss": 0.7182, + "num_input_tokens_seen": 3602144, + "step": 2370 + }, + { + "epoch": 0.0760194609820114, + "grad_norm": 0.8636143207550049, + "learning_rate": 4.9566696042116704e-05, + "loss": 0.5716, + "num_input_tokens_seen": 3609504, + "step": 2375 + }, + { + "epoch": 0.07617950195249984, + "grad_norm": 0.8968456983566284, + "learning_rate": 4.9564874232519766e-05, + "loss": 0.5584, + "num_input_tokens_seen": 3618256, + "step": 2380 + }, + { + "epoch": 0.07633954292298828, + "grad_norm": 0.8655830025672913, + "learning_rate": 4.9563048634710516e-05, + "loss": 0.5903, + "num_input_tokens_seen": 3626672, + "step": 2385 + }, + { + "epoch": 0.07649958389347673, + "grad_norm": 0.9095003604888916, + "learning_rate": 4.956121924897049e-05, + "loss": 0.6135, + "num_input_tokens_seen": 3634240, + "step": 2390 + }, + { + "epoch": 0.07665962486396517, + "grad_norm": 1.2573186159133911, + "learning_rate": 4.955938607558181e-05, + "loss": 0.6842, + "num_input_tokens_seen": 3642240, + "step": 2395 + }, + { + "epoch": 0.07681966583445363, + "grad_norm": 0.8020550608634949, + "learning_rate": 4.955754911482715e-05, + "loss": 0.5612, + "num_input_tokens_seen": 3649488, + "step": 2400 + }, + { + "epoch": 0.07681966583445363, + "eval_loss": 0.6852848529815674, + "eval_runtime": 561.7759, + "eval_samples_per_second": 24.718, + "eval_steps_per_second": 12.359, + "num_input_tokens_seen": 3649488, + "step": 2400 + }, + { + "epoch": 0.07697970680494207, + "grad_norm": 1.6815658807754517, + "learning_rate": 4.9555708366989804e-05, + "loss": 0.7224, + "num_input_tokens_seen": 3656704, + "step": 2405 + }, + { + "epoch": 0.0771397477754305, + "grad_norm": 1.1589627265930176, + "learning_rate": 4.9553863832353655e-05, + "loss": 0.7337, + "num_input_tokens_seen": 3664208, + "step": 2410 + }, + { + "epoch": 0.07729978874591896, + "grad_norm": 0.8496106863021851, + "learning_rate": 4.955201551120313e-05, + "loss": 0.5816, + "num_input_tokens_seen": 3671600, + "step": 2415 + }, + { + "epoch": 0.0774598297164074, + "grad_norm": 0.6047427654266357, + "learning_rate": 4.955016340382328e-05, + "loss": 0.504, + "num_input_tokens_seen": 3679552, + "step": 2420 + }, + { + "epoch": 0.07761987068689584, + "grad_norm": 1.8450944423675537, + "learning_rate": 4.954830751049972e-05, + "loss": 0.6051, + "num_input_tokens_seen": 3687328, + "step": 2425 + }, + { + "epoch": 0.07777991165738429, + "grad_norm": 1.83318030834198, + "learning_rate": 4.954644783151864e-05, + "loss": 0.6749, + "num_input_tokens_seen": 3694608, + "step": 2430 + }, + { + "epoch": 0.07793995262787273, + "grad_norm": 1.9011383056640625, + "learning_rate": 4.954458436716684e-05, + "loss": 0.529, + "num_input_tokens_seen": 3702848, + "step": 2435 + }, + { + "epoch": 0.07809999359836119, + "grad_norm": 1.0794055461883545, + "learning_rate": 4.954271711773168e-05, + "loss": 0.5378, + "num_input_tokens_seen": 3710192, + "step": 2440 + }, + { + "epoch": 0.07826003456884963, + "grad_norm": 0.6897259950637817, + "learning_rate": 4.9540846083501115e-05, + "loss": 0.6227, + "num_input_tokens_seen": 3717856, + "step": 2445 + }, + { + "epoch": 0.07842007553933807, + "grad_norm": 0.8993026614189148, + "learning_rate": 4.953897126476369e-05, + "loss": 0.7907, + "num_input_tokens_seen": 3725008, + "step": 2450 + }, + { + "epoch": 0.07858011650982652, + "grad_norm": 1.2294617891311646, + "learning_rate": 4.9537092661808514e-05, + "loss": 0.7692, + "num_input_tokens_seen": 3732592, + "step": 2455 + }, + { + "epoch": 0.07874015748031496, + "grad_norm": 1.2163927555084229, + "learning_rate": 4.9535210274925306e-05, + "loss": 0.7814, + "num_input_tokens_seen": 3740480, + "step": 2460 + }, + { + "epoch": 0.0789001984508034, + "grad_norm": 1.0437994003295898, + "learning_rate": 4.953332410440435e-05, + "loss": 0.6426, + "num_input_tokens_seen": 3748256, + "step": 2465 + }, + { + "epoch": 0.07906023942129185, + "grad_norm": 2.415705442428589, + "learning_rate": 4.9531434150536496e-05, + "loss": 0.8421, + "num_input_tokens_seen": 3756080, + "step": 2470 + }, + { + "epoch": 0.07922028039178029, + "grad_norm": 1.0448119640350342, + "learning_rate": 4.952954041361322e-05, + "loss": 0.6704, + "num_input_tokens_seen": 3763936, + "step": 2475 + }, + { + "epoch": 0.07938032136226875, + "grad_norm": 1.09918212890625, + "learning_rate": 4.952764289392655e-05, + "loss": 0.558, + "num_input_tokens_seen": 3771072, + "step": 2480 + }, + { + "epoch": 0.07954036233275719, + "grad_norm": 1.4919615983963013, + "learning_rate": 4.952574159176912e-05, + "loss": 0.5489, + "num_input_tokens_seen": 3778448, + "step": 2485 + }, + { + "epoch": 0.07970040330324563, + "grad_norm": 1.0320907831192017, + "learning_rate": 4.952383650743413e-05, + "loss": 0.7132, + "num_input_tokens_seen": 3786352, + "step": 2490 + }, + { + "epoch": 0.07986044427373408, + "grad_norm": 1.6366517543792725, + "learning_rate": 4.952192764121536e-05, + "loss": 0.6181, + "num_input_tokens_seen": 3794352, + "step": 2495 + }, + { + "epoch": 0.08002048524422252, + "grad_norm": 0.8165029287338257, + "learning_rate": 4.9520014993407185e-05, + "loss": 0.6317, + "num_input_tokens_seen": 3801824, + "step": 2500 + }, + { + "epoch": 0.08018052621471096, + "grad_norm": 0.7319425344467163, + "learning_rate": 4.951809856430456e-05, + "loss": 0.5435, + "num_input_tokens_seen": 3809232, + "step": 2505 + }, + { + "epoch": 0.08034056718519941, + "grad_norm": 1.315706729888916, + "learning_rate": 4.951617835420303e-05, + "loss": 0.6579, + "num_input_tokens_seen": 3816864, + "step": 2510 + }, + { + "epoch": 0.08050060815568785, + "grad_norm": 2.2982349395751953, + "learning_rate": 4.951425436339869e-05, + "loss": 0.5994, + "num_input_tokens_seen": 3824368, + "step": 2515 + }, + { + "epoch": 0.0806606491261763, + "grad_norm": 1.0307234525680542, + "learning_rate": 4.9512326592188274e-05, + "loss": 0.6332, + "num_input_tokens_seen": 3831760, + "step": 2520 + }, + { + "epoch": 0.08082069009666475, + "grad_norm": 1.462273359298706, + "learning_rate": 4.9510395040869054e-05, + "loss": 0.6334, + "num_input_tokens_seen": 3839120, + "step": 2525 + }, + { + "epoch": 0.08098073106715319, + "grad_norm": 0.8158721923828125, + "learning_rate": 4.9508459709738905e-05, + "loss": 0.8163, + "num_input_tokens_seen": 3846704, + "step": 2530 + }, + { + "epoch": 0.08114077203764164, + "grad_norm": 1.3491960763931274, + "learning_rate": 4.950652059909627e-05, + "loss": 0.5736, + "num_input_tokens_seen": 3855184, + "step": 2535 + }, + { + "epoch": 0.08130081300813008, + "grad_norm": 0.9270776510238647, + "learning_rate": 4.95045777092402e-05, + "loss": 0.529, + "num_input_tokens_seen": 3862768, + "step": 2540 + }, + { + "epoch": 0.08146085397861853, + "grad_norm": 1.317438006401062, + "learning_rate": 4.950263104047031e-05, + "loss": 0.7322, + "num_input_tokens_seen": 3869984, + "step": 2545 + }, + { + "epoch": 0.08162089494910697, + "grad_norm": 1.2313737869262695, + "learning_rate": 4.9500680593086775e-05, + "loss": 0.6435, + "num_input_tokens_seen": 3877392, + "step": 2550 + }, + { + "epoch": 0.08178093591959541, + "grad_norm": 0.7292083501815796, + "learning_rate": 4.94987263673904e-05, + "loss": 0.6425, + "num_input_tokens_seen": 3884688, + "step": 2555 + }, + { + "epoch": 0.08194097689008387, + "grad_norm": 0.8170074820518494, + "learning_rate": 4.949676836368256e-05, + "loss": 0.697, + "num_input_tokens_seen": 3892592, + "step": 2560 + }, + { + "epoch": 0.0821010178605723, + "grad_norm": 1.4472042322158813, + "learning_rate": 4.949480658226518e-05, + "loss": 0.6413, + "num_input_tokens_seen": 3899856, + "step": 2565 + }, + { + "epoch": 0.08226105883106075, + "grad_norm": 2.2161316871643066, + "learning_rate": 4.949284102344082e-05, + "loss": 0.8432, + "num_input_tokens_seen": 3907648, + "step": 2570 + }, + { + "epoch": 0.0824210998015492, + "grad_norm": 1.296159029006958, + "learning_rate": 4.9490871687512565e-05, + "loss": 0.6152, + "num_input_tokens_seen": 3915408, + "step": 2575 + }, + { + "epoch": 0.08258114077203764, + "grad_norm": 2.019073486328125, + "learning_rate": 4.948889857478413e-05, + "loss": 0.8636, + "num_input_tokens_seen": 3923600, + "step": 2580 + }, + { + "epoch": 0.0827411817425261, + "grad_norm": 0.8166035413742065, + "learning_rate": 4.948692168555978e-05, + "loss": 0.5456, + "num_input_tokens_seen": 3931152, + "step": 2585 + }, + { + "epoch": 0.08290122271301453, + "grad_norm": 1.2364286184310913, + "learning_rate": 4.94849410201444e-05, + "loss": 0.6533, + "num_input_tokens_seen": 3938784, + "step": 2590 + }, + { + "epoch": 0.08306126368350297, + "grad_norm": 1.3864548206329346, + "learning_rate": 4.948295657884341e-05, + "loss": 0.6446, + "num_input_tokens_seen": 3946240, + "step": 2595 + }, + { + "epoch": 0.08322130465399143, + "grad_norm": 0.8176770210266113, + "learning_rate": 4.9480968361962835e-05, + "loss": 0.5582, + "num_input_tokens_seen": 3953824, + "step": 2600 + }, + { + "epoch": 0.08322130465399143, + "eval_loss": 0.6774986386299133, + "eval_runtime": 559.9122, + "eval_samples_per_second": 24.8, + "eval_steps_per_second": 12.4, + "num_input_tokens_seen": 3953824, + "step": 2600 + }, + { + "epoch": 0.08338134562447987, + "grad_norm": 1.0834664106369019, + "learning_rate": 4.9478976369809305e-05, + "loss": 0.7718, + "num_input_tokens_seen": 3961328, + "step": 2605 + }, + { + "epoch": 0.0835413865949683, + "grad_norm": 0.9409216046333313, + "learning_rate": 4.947698060268999e-05, + "loss": 0.6089, + "num_input_tokens_seen": 3969120, + "step": 2610 + }, + { + "epoch": 0.08370142756545676, + "grad_norm": 0.7750560641288757, + "learning_rate": 4.9474981060912665e-05, + "loss": 0.4662, + "num_input_tokens_seen": 3977184, + "step": 2615 + }, + { + "epoch": 0.0838614685359452, + "grad_norm": 1.0298025608062744, + "learning_rate": 4.94729777447857e-05, + "loss": 0.4771, + "num_input_tokens_seen": 3984752, + "step": 2620 + }, + { + "epoch": 0.08402150950643365, + "grad_norm": 0.5456575155258179, + "learning_rate": 4.947097065461801e-05, + "loss": 0.5406, + "num_input_tokens_seen": 3992656, + "step": 2625 + }, + { + "epoch": 0.08418155047692209, + "grad_norm": 1.1720476150512695, + "learning_rate": 4.9468959790719125e-05, + "loss": 0.6805, + "num_input_tokens_seen": 4000288, + "step": 2630 + }, + { + "epoch": 0.08434159144741053, + "grad_norm": 2.0215601921081543, + "learning_rate": 4.9466945153399146e-05, + "loss": 0.7877, + "num_input_tokens_seen": 4007472, + "step": 2635 + }, + { + "epoch": 0.08450163241789899, + "grad_norm": 1.9195985794067383, + "learning_rate": 4.9464926742968755e-05, + "loss": 0.7466, + "num_input_tokens_seen": 4015840, + "step": 2640 + }, + { + "epoch": 0.08466167338838743, + "grad_norm": 2.9836692810058594, + "learning_rate": 4.946290455973921e-05, + "loss": 0.6152, + "num_input_tokens_seen": 4023664, + "step": 2645 + }, + { + "epoch": 0.08482171435887587, + "grad_norm": 0.7797253131866455, + "learning_rate": 4.9460878604022365e-05, + "loss": 0.6819, + "num_input_tokens_seen": 4031488, + "step": 2650 + }, + { + "epoch": 0.08498175532936432, + "grad_norm": 1.0177887678146362, + "learning_rate": 4.945884887613065e-05, + "loss": 0.7069, + "num_input_tokens_seen": 4038640, + "step": 2655 + }, + { + "epoch": 0.08514179629985276, + "grad_norm": 1.0664386749267578, + "learning_rate": 4.9456815376377055e-05, + "loss": 0.6302, + "num_input_tokens_seen": 4046064, + "step": 2660 + }, + { + "epoch": 0.08530183727034121, + "grad_norm": 1.226686954498291, + "learning_rate": 4.9454778105075195e-05, + "loss": 0.6267, + "num_input_tokens_seen": 4053152, + "step": 2665 + }, + { + "epoch": 0.08546187824082965, + "grad_norm": 1.4226435422897339, + "learning_rate": 4.945273706253924e-05, + "loss": 0.6981, + "num_input_tokens_seen": 4060816, + "step": 2670 + }, + { + "epoch": 0.08562191921131809, + "grad_norm": 0.806235671043396, + "learning_rate": 4.9450692249083925e-05, + "loss": 0.5071, + "num_input_tokens_seen": 4068608, + "step": 2675 + }, + { + "epoch": 0.08578196018180655, + "grad_norm": 2.02567720413208, + "learning_rate": 4.9448643665024605e-05, + "loss": 0.6234, + "num_input_tokens_seen": 4076240, + "step": 2680 + }, + { + "epoch": 0.08594200115229499, + "grad_norm": 1.5321677923202515, + "learning_rate": 4.944659131067719e-05, + "loss": 0.3864, + "num_input_tokens_seen": 4083792, + "step": 2685 + }, + { + "epoch": 0.08610204212278343, + "grad_norm": 0.7286571264266968, + "learning_rate": 4.944453518635818e-05, + "loss": 0.5446, + "num_input_tokens_seen": 4091040, + "step": 2690 + }, + { + "epoch": 0.08626208309327188, + "grad_norm": 1.2578409910202026, + "learning_rate": 4.944247529238465e-05, + "loss": 0.6808, + "num_input_tokens_seen": 4099072, + "step": 2695 + }, + { + "epoch": 0.08642212406376032, + "grad_norm": 2.043022632598877, + "learning_rate": 4.944041162907427e-05, + "loss": 0.5876, + "num_input_tokens_seen": 4106368, + "step": 2700 + }, + { + "epoch": 0.08658216503424877, + "grad_norm": 1.829338788986206, + "learning_rate": 4.943834419674529e-05, + "loss": 0.7016, + "num_input_tokens_seen": 4113936, + "step": 2705 + }, + { + "epoch": 0.08674220600473721, + "grad_norm": 1.358072280883789, + "learning_rate": 4.9436272995716506e-05, + "loss": 0.4417, + "num_input_tokens_seen": 4121264, + "step": 2710 + }, + { + "epoch": 0.08690224697522565, + "grad_norm": 1.3615583181381226, + "learning_rate": 4.943419802630735e-05, + "loss": 0.5889, + "num_input_tokens_seen": 4128624, + "step": 2715 + }, + { + "epoch": 0.0870622879457141, + "grad_norm": 0.47475719451904297, + "learning_rate": 4.94321192888378e-05, + "loss": 0.6895, + "num_input_tokens_seen": 4136608, + "step": 2720 + }, + { + "epoch": 0.08722232891620255, + "grad_norm": 1.1903080940246582, + "learning_rate": 4.943003678362842e-05, + "loss": 0.7964, + "num_input_tokens_seen": 4144400, + "step": 2725 + }, + { + "epoch": 0.087382369886691, + "grad_norm": 0.952420175075531, + "learning_rate": 4.942795051100036e-05, + "loss": 0.5962, + "num_input_tokens_seen": 4152224, + "step": 2730 + }, + { + "epoch": 0.08754241085717944, + "grad_norm": 1.140627145767212, + "learning_rate": 4.942586047127536e-05, + "loss": 0.5024, + "num_input_tokens_seen": 4159392, + "step": 2735 + }, + { + "epoch": 0.08770245182766788, + "grad_norm": 1.3068759441375732, + "learning_rate": 4.942376666477571e-05, + "loss": 0.5828, + "num_input_tokens_seen": 4167536, + "step": 2740 + }, + { + "epoch": 0.08786249279815633, + "grad_norm": 1.6635386943817139, + "learning_rate": 4.9421669091824304e-05, + "loss": 0.67, + "num_input_tokens_seen": 4174688, + "step": 2745 + }, + { + "epoch": 0.08802253376864477, + "grad_norm": 0.4766179323196411, + "learning_rate": 4.9419567752744634e-05, + "loss": 0.4326, + "num_input_tokens_seen": 4182656, + "step": 2750 + }, + { + "epoch": 0.08818257473913321, + "grad_norm": 1.4008114337921143, + "learning_rate": 4.941746264786074e-05, + "loss": 0.6996, + "num_input_tokens_seen": 4189872, + "step": 2755 + }, + { + "epoch": 0.08834261570962167, + "grad_norm": 1.6392241716384888, + "learning_rate": 4.9415353777497254e-05, + "loss": 0.5666, + "num_input_tokens_seen": 4197360, + "step": 2760 + }, + { + "epoch": 0.0885026566801101, + "grad_norm": 1.1861194372177124, + "learning_rate": 4.9413241141979394e-05, + "loss": 0.6054, + "num_input_tokens_seen": 4204928, + "step": 2765 + }, + { + "epoch": 0.08866269765059856, + "grad_norm": 1.4490841627120972, + "learning_rate": 4.9411124741632956e-05, + "loss": 0.5757, + "num_input_tokens_seen": 4212432, + "step": 2770 + }, + { + "epoch": 0.088822738621087, + "grad_norm": 1.010413408279419, + "learning_rate": 4.940900457678431e-05, + "loss": 0.5999, + "num_input_tokens_seen": 4219936, + "step": 2775 + }, + { + "epoch": 0.08898277959157544, + "grad_norm": 1.167784333229065, + "learning_rate": 4.9406880647760425e-05, + "loss": 0.6055, + "num_input_tokens_seen": 4227392, + "step": 2780 + }, + { + "epoch": 0.08914282056206389, + "grad_norm": 1.980897307395935, + "learning_rate": 4.9404752954888824e-05, + "loss": 0.7177, + "num_input_tokens_seen": 4234720, + "step": 2785 + }, + { + "epoch": 0.08930286153255233, + "grad_norm": 1.0197569131851196, + "learning_rate": 4.940262149849762e-05, + "loss": 0.5591, + "num_input_tokens_seen": 4242464, + "step": 2790 + }, + { + "epoch": 0.08946290250304077, + "grad_norm": 0.6827260851860046, + "learning_rate": 4.9400486278915526e-05, + "loss": 0.6495, + "num_input_tokens_seen": 4250272, + "step": 2795 + }, + { + "epoch": 0.08962294347352923, + "grad_norm": 0.6823484897613525, + "learning_rate": 4.939834729647181e-05, + "loss": 0.5903, + "num_input_tokens_seen": 4257904, + "step": 2800 + }, + { + "epoch": 0.08962294347352923, + "eval_loss": 0.6691160798072815, + "eval_runtime": 559.7397, + "eval_samples_per_second": 24.808, + "eval_steps_per_second": 12.404, + "num_input_tokens_seen": 4257904, + "step": 2800 + }, + { + "epoch": 0.08978298444401767, + "grad_norm": 1.9426416158676147, + "learning_rate": 4.9396204551496326e-05, + "loss": 0.6405, + "num_input_tokens_seen": 4265392, + "step": 2805 + }, + { + "epoch": 0.08994302541450612, + "grad_norm": 0.8621007800102234, + "learning_rate": 4.939405804431952e-05, + "loss": 0.6709, + "num_input_tokens_seen": 4272944, + "step": 2810 + }, + { + "epoch": 0.09010306638499456, + "grad_norm": 0.5417982935905457, + "learning_rate": 4.9391907775272414e-05, + "loss": 0.4889, + "num_input_tokens_seen": 4280272, + "step": 2815 + }, + { + "epoch": 0.090263107355483, + "grad_norm": 0.997646689414978, + "learning_rate": 4.9389753744686604e-05, + "loss": 0.5769, + "num_input_tokens_seen": 4287920, + "step": 2820 + }, + { + "epoch": 0.09042314832597145, + "grad_norm": 0.8363180756568909, + "learning_rate": 4.938759595289426e-05, + "loss": 0.5838, + "num_input_tokens_seen": 4295424, + "step": 2825 + }, + { + "epoch": 0.09058318929645989, + "grad_norm": 1.104830265045166, + "learning_rate": 4.938543440022815e-05, + "loss": 0.725, + "num_input_tokens_seen": 4303408, + "step": 2830 + }, + { + "epoch": 0.09074323026694833, + "grad_norm": 1.1268222332000732, + "learning_rate": 4.938326908702161e-05, + "loss": 0.7548, + "num_input_tokens_seen": 4310896, + "step": 2835 + }, + { + "epoch": 0.09090327123743679, + "grad_norm": 1.0747884511947632, + "learning_rate": 4.9381100013608554e-05, + "loss": 0.7567, + "num_input_tokens_seen": 4318208, + "step": 2840 + }, + { + "epoch": 0.09106331220792523, + "grad_norm": 0.7930203080177307, + "learning_rate": 4.9378927180323485e-05, + "loss": 0.6435, + "num_input_tokens_seen": 4325760, + "step": 2845 + }, + { + "epoch": 0.09122335317841368, + "grad_norm": 0.8280487656593323, + "learning_rate": 4.937675058750148e-05, + "loss": 0.6085, + "num_input_tokens_seen": 4333216, + "step": 2850 + }, + { + "epoch": 0.09138339414890212, + "grad_norm": 0.5594574213027954, + "learning_rate": 4.937457023547819e-05, + "loss": 0.6508, + "num_input_tokens_seen": 4340752, + "step": 2855 + }, + { + "epoch": 0.09154343511939056, + "grad_norm": 3.1955902576446533, + "learning_rate": 4.9372386124589876e-05, + "loss": 0.8165, + "num_input_tokens_seen": 4347984, + "step": 2860 + }, + { + "epoch": 0.09170347608987901, + "grad_norm": 1.1743874549865723, + "learning_rate": 4.937019825517333e-05, + "loss": 0.6534, + "num_input_tokens_seen": 4355952, + "step": 2865 + }, + { + "epoch": 0.09186351706036745, + "grad_norm": 0.5182885527610779, + "learning_rate": 4.9368006627565954e-05, + "loss": 0.6558, + "num_input_tokens_seen": 4363456, + "step": 2870 + }, + { + "epoch": 0.09202355803085589, + "grad_norm": 0.8887224197387695, + "learning_rate": 4.936581124210573e-05, + "loss": 0.4438, + "num_input_tokens_seen": 4371136, + "step": 2875 + }, + { + "epoch": 0.09218359900134435, + "grad_norm": 1.8663504123687744, + "learning_rate": 4.9363612099131216e-05, + "loss": 0.6029, + "num_input_tokens_seen": 4378192, + "step": 2880 + }, + { + "epoch": 0.09234363997183279, + "grad_norm": 0.7953860759735107, + "learning_rate": 4.936140919898155e-05, + "loss": 0.7283, + "num_input_tokens_seen": 4386144, + "step": 2885 + }, + { + "epoch": 0.09250368094232124, + "grad_norm": 1.0848835706710815, + "learning_rate": 4.9359202541996426e-05, + "loss": 0.736, + "num_input_tokens_seen": 4393264, + "step": 2890 + }, + { + "epoch": 0.09266372191280968, + "grad_norm": 1.6836477518081665, + "learning_rate": 4.935699212851616e-05, + "loss": 0.4923, + "num_input_tokens_seen": 4400672, + "step": 2895 + }, + { + "epoch": 0.09282376288329812, + "grad_norm": 1.044900894165039, + "learning_rate": 4.935477795888162e-05, + "loss": 0.551, + "num_input_tokens_seen": 4407920, + "step": 2900 + }, + { + "epoch": 0.09298380385378657, + "grad_norm": 0.9487115144729614, + "learning_rate": 4.935256003343426e-05, + "loss": 0.8809, + "num_input_tokens_seen": 4415536, + "step": 2905 + }, + { + "epoch": 0.09314384482427501, + "grad_norm": 1.2644648551940918, + "learning_rate": 4.93503383525161e-05, + "loss": 0.6783, + "num_input_tokens_seen": 4423312, + "step": 2910 + }, + { + "epoch": 0.09330388579476345, + "grad_norm": 0.9119481444358826, + "learning_rate": 4.934811291646977e-05, + "loss": 0.5185, + "num_input_tokens_seen": 4430624, + "step": 2915 + }, + { + "epoch": 0.0934639267652519, + "grad_norm": 1.849015474319458, + "learning_rate": 4.934588372563845e-05, + "loss": 0.5526, + "num_input_tokens_seen": 4438176, + "step": 2920 + }, + { + "epoch": 0.09362396773574035, + "grad_norm": 3.3929638862609863, + "learning_rate": 4.93436507803659e-05, + "loss": 0.5708, + "num_input_tokens_seen": 4445376, + "step": 2925 + }, + { + "epoch": 0.0937840087062288, + "grad_norm": 1.2073572874069214, + "learning_rate": 4.934141408099649e-05, + "loss": 0.5124, + "num_input_tokens_seen": 4452560, + "step": 2930 + }, + { + "epoch": 0.09394404967671724, + "grad_norm": 1.4691801071166992, + "learning_rate": 4.9339173627875135e-05, + "loss": 0.647, + "num_input_tokens_seen": 4459920, + "step": 2935 + }, + { + "epoch": 0.09410409064720568, + "grad_norm": 1.0668058395385742, + "learning_rate": 4.9336929421347335e-05, + "loss": 0.5231, + "num_input_tokens_seen": 4467296, + "step": 2940 + }, + { + "epoch": 0.09426413161769413, + "grad_norm": 0.9110621213912964, + "learning_rate": 4.933468146175918e-05, + "loss": 0.6132, + "num_input_tokens_seen": 4475088, + "step": 2945 + }, + { + "epoch": 0.09442417258818257, + "grad_norm": 1.1095468997955322, + "learning_rate": 4.933242974945734e-05, + "loss": 0.6616, + "num_input_tokens_seen": 4482688, + "step": 2950 + }, + { + "epoch": 0.09458421355867103, + "grad_norm": 1.2849664688110352, + "learning_rate": 4.933017428478906e-05, + "loss": 0.625, + "num_input_tokens_seen": 4490208, + "step": 2955 + }, + { + "epoch": 0.09474425452915947, + "grad_norm": 0.8971932530403137, + "learning_rate": 4.932791506810214e-05, + "loss": 0.5026, + "num_input_tokens_seen": 4497664, + "step": 2960 + }, + { + "epoch": 0.0949042954996479, + "grad_norm": 1.1097588539123535, + "learning_rate": 4.932565209974499e-05, + "loss": 0.9326, + "num_input_tokens_seen": 4505488, + "step": 2965 + }, + { + "epoch": 0.09506433647013636, + "grad_norm": 1.0436105728149414, + "learning_rate": 4.93233853800666e-05, + "loss": 0.6109, + "num_input_tokens_seen": 4513008, + "step": 2970 + }, + { + "epoch": 0.0952243774406248, + "grad_norm": 0.8983567357063293, + "learning_rate": 4.932111490941651e-05, + "loss": 0.573, + "num_input_tokens_seen": 4520992, + "step": 2975 + }, + { + "epoch": 0.09538441841111324, + "grad_norm": 1.5102367401123047, + "learning_rate": 4.9318840688144876e-05, + "loss": 0.5658, + "num_input_tokens_seen": 4529088, + "step": 2980 + }, + { + "epoch": 0.09554445938160169, + "grad_norm": 0.9365469813346863, + "learning_rate": 4.9316562716602387e-05, + "loss": 0.589, + "num_input_tokens_seen": 4537264, + "step": 2985 + }, + { + "epoch": 0.09570450035209013, + "grad_norm": 0.7899389863014221, + "learning_rate": 4.9314280995140346e-05, + "loss": 0.7311, + "num_input_tokens_seen": 4544976, + "step": 2990 + }, + { + "epoch": 0.09586454132257859, + "grad_norm": 0.4060385227203369, + "learning_rate": 4.931199552411063e-05, + "loss": 0.6974, + "num_input_tokens_seen": 4552320, + "step": 2995 + }, + { + "epoch": 0.09602458229306703, + "grad_norm": 1.5618573427200317, + "learning_rate": 4.930970630386568e-05, + "loss": 0.6387, + "num_input_tokens_seen": 4560128, + "step": 3000 + }, + { + "epoch": 0.09602458229306703, + "eval_loss": 0.6626488566398621, + "eval_runtime": 557.4102, + "eval_samples_per_second": 24.912, + "eval_steps_per_second": 12.456, + "num_input_tokens_seen": 4560128, + "step": 3000 + }, + { + "epoch": 0.09618462326355547, + "grad_norm": 1.9712647199630737, + "learning_rate": 4.9307413334758524e-05, + "loss": 0.7168, + "num_input_tokens_seen": 4567760, + "step": 3005 + }, + { + "epoch": 0.09634466423404392, + "grad_norm": 0.7315365672111511, + "learning_rate": 4.930511661714276e-05, + "loss": 0.6147, + "num_input_tokens_seen": 4575184, + "step": 3010 + }, + { + "epoch": 0.09650470520453236, + "grad_norm": 0.7971432209014893, + "learning_rate": 4.9302816151372576e-05, + "loss": 0.5571, + "num_input_tokens_seen": 4583296, + "step": 3015 + }, + { + "epoch": 0.0966647461750208, + "grad_norm": 0.7919638156890869, + "learning_rate": 4.930051193780274e-05, + "loss": 0.6031, + "num_input_tokens_seen": 4590800, + "step": 3020 + }, + { + "epoch": 0.09682478714550925, + "grad_norm": 1.0093653202056885, + "learning_rate": 4.929820397678858e-05, + "loss": 0.6714, + "num_input_tokens_seen": 4598192, + "step": 3025 + }, + { + "epoch": 0.09698482811599769, + "grad_norm": 0.6401515007019043, + "learning_rate": 4.9295892268686015e-05, + "loss": 0.6121, + "num_input_tokens_seen": 4605648, + "step": 3030 + }, + { + "epoch": 0.09714486908648615, + "grad_norm": 0.842907726764679, + "learning_rate": 4.9293576813851536e-05, + "loss": 0.4466, + "num_input_tokens_seen": 4613424, + "step": 3035 + }, + { + "epoch": 0.09730491005697459, + "grad_norm": 0.7773587107658386, + "learning_rate": 4.929125761264223e-05, + "loss": 0.6508, + "num_input_tokens_seen": 4620832, + "step": 3040 + }, + { + "epoch": 0.09746495102746303, + "grad_norm": 0.8349104523658752, + "learning_rate": 4.928893466541573e-05, + "loss": 0.5204, + "num_input_tokens_seen": 4628496, + "step": 3045 + }, + { + "epoch": 0.09762499199795148, + "grad_norm": 2.200345277786255, + "learning_rate": 4.928660797253027e-05, + "loss": 0.9241, + "num_input_tokens_seen": 4635888, + "step": 3050 + }, + { + "epoch": 0.09778503296843992, + "grad_norm": 1.1222355365753174, + "learning_rate": 4.928427753434467e-05, + "loss": 0.542, + "num_input_tokens_seen": 4643072, + "step": 3055 + }, + { + "epoch": 0.09794507393892836, + "grad_norm": 0.8426560759544373, + "learning_rate": 4.9281943351218286e-05, + "loss": 0.7268, + "num_input_tokens_seen": 4650848, + "step": 3060 + }, + { + "epoch": 0.09810511490941681, + "grad_norm": 0.8598178029060364, + "learning_rate": 4.9279605423511095e-05, + "loss": 0.6075, + "num_input_tokens_seen": 4658432, + "step": 3065 + }, + { + "epoch": 0.09826515587990525, + "grad_norm": 1.035646915435791, + "learning_rate": 4.927726375158363e-05, + "loss": 0.6329, + "num_input_tokens_seen": 4665456, + "step": 3070 + }, + { + "epoch": 0.0984251968503937, + "grad_norm": 1.7721261978149414, + "learning_rate": 4.9274918335797004e-05, + "loss": 0.6156, + "num_input_tokens_seen": 4672896, + "step": 3075 + }, + { + "epoch": 0.09858523782088215, + "grad_norm": 0.7773606777191162, + "learning_rate": 4.927256917651292e-05, + "loss": 0.5482, + "num_input_tokens_seen": 4680544, + "step": 3080 + }, + { + "epoch": 0.09874527879137059, + "grad_norm": 1.0569629669189453, + "learning_rate": 4.927021627409364e-05, + "loss": 0.5529, + "num_input_tokens_seen": 4688048, + "step": 3085 + }, + { + "epoch": 0.09890531976185904, + "grad_norm": 1.078602910041809, + "learning_rate": 4.9267859628902005e-05, + "loss": 0.8406, + "num_input_tokens_seen": 4695328, + "step": 3090 + }, + { + "epoch": 0.09906536073234748, + "grad_norm": 1.3732478618621826, + "learning_rate": 4.9265499241301454e-05, + "loss": 0.6558, + "num_input_tokens_seen": 4702912, + "step": 3095 + }, + { + "epoch": 0.09922540170283592, + "grad_norm": 1.2809128761291504, + "learning_rate": 4.926313511165598e-05, + "loss": 0.5044, + "num_input_tokens_seen": 4710416, + "step": 3100 + }, + { + "epoch": 0.09938544267332437, + "grad_norm": 1.2744991779327393, + "learning_rate": 4.926076724033016e-05, + "loss": 0.7269, + "num_input_tokens_seen": 4718480, + "step": 3105 + }, + { + "epoch": 0.09954548364381281, + "grad_norm": 3.876368284225464, + "learning_rate": 4.9258395627689146e-05, + "loss": 0.7616, + "num_input_tokens_seen": 4725888, + "step": 3110 + }, + { + "epoch": 0.09970552461430127, + "grad_norm": 1.8352584838867188, + "learning_rate": 4.925602027409868e-05, + "loss": 0.7062, + "num_input_tokens_seen": 4734000, + "step": 3115 + }, + { + "epoch": 0.0998655655847897, + "grad_norm": 0.8262115120887756, + "learning_rate": 4.925364117992507e-05, + "loss": 0.5667, + "num_input_tokens_seen": 4741440, + "step": 3120 + }, + { + "epoch": 0.10002560655527815, + "grad_norm": 1.210296630859375, + "learning_rate": 4.92512583455352e-05, + "loss": 0.6254, + "num_input_tokens_seen": 4749184, + "step": 3125 + }, + { + "epoch": 0.1001856475257666, + "grad_norm": 1.722274661064148, + "learning_rate": 4.9248871771296536e-05, + "loss": 0.7461, + "num_input_tokens_seen": 4756736, + "step": 3130 + }, + { + "epoch": 0.10034568849625504, + "grad_norm": 0.7942022085189819, + "learning_rate": 4.924648145757711e-05, + "loss": 0.5504, + "num_input_tokens_seen": 4764128, + "step": 3135 + }, + { + "epoch": 0.1005057294667435, + "grad_norm": 0.6593976616859436, + "learning_rate": 4.924408740474554e-05, + "loss": 0.4852, + "num_input_tokens_seen": 4772064, + "step": 3140 + }, + { + "epoch": 0.10066577043723193, + "grad_norm": 1.2369623184204102, + "learning_rate": 4.924168961317103e-05, + "loss": 0.6833, + "num_input_tokens_seen": 4779824, + "step": 3145 + }, + { + "epoch": 0.10082581140772037, + "grad_norm": 1.9277011156082153, + "learning_rate": 4.9239288083223334e-05, + "loss": 0.5351, + "num_input_tokens_seen": 4787984, + "step": 3150 + }, + { + "epoch": 0.10098585237820883, + "grad_norm": 0.808043360710144, + "learning_rate": 4.9236882815272803e-05, + "loss": 0.7307, + "num_input_tokens_seen": 4796032, + "step": 3155 + }, + { + "epoch": 0.10114589334869727, + "grad_norm": 0.9451218247413635, + "learning_rate": 4.9234473809690365e-05, + "loss": 0.6223, + "num_input_tokens_seen": 4803712, + "step": 3160 + }, + { + "epoch": 0.1013059343191857, + "grad_norm": 1.0269356966018677, + "learning_rate": 4.923206106684752e-05, + "loss": 0.6297, + "num_input_tokens_seen": 4811632, + "step": 3165 + }, + { + "epoch": 0.10146597528967416, + "grad_norm": 1.1476318836212158, + "learning_rate": 4.922964458711634e-05, + "loss": 0.535, + "num_input_tokens_seen": 4819248, + "step": 3170 + }, + { + "epoch": 0.1016260162601626, + "grad_norm": 0.5486667156219482, + "learning_rate": 4.9227224370869474e-05, + "loss": 0.6421, + "num_input_tokens_seen": 4826752, + "step": 3175 + }, + { + "epoch": 0.10178605723065105, + "grad_norm": 1.215318202972412, + "learning_rate": 4.9224800418480155e-05, + "loss": 0.5435, + "num_input_tokens_seen": 4834832, + "step": 3180 + }, + { + "epoch": 0.10194609820113949, + "grad_norm": 1.0406358242034912, + "learning_rate": 4.9222372730322176e-05, + "loss": 0.5816, + "num_input_tokens_seen": 4842736, + "step": 3185 + }, + { + "epoch": 0.10210613917162793, + "grad_norm": 0.9366384744644165, + "learning_rate": 4.921994130676993e-05, + "loss": 0.7122, + "num_input_tokens_seen": 4850400, + "step": 3190 + }, + { + "epoch": 0.10226618014211639, + "grad_norm": 1.5716423988342285, + "learning_rate": 4.9217506148198366e-05, + "loss": 0.4857, + "num_input_tokens_seen": 4857888, + "step": 3195 + }, + { + "epoch": 0.10242622111260483, + "grad_norm": 0.9801841974258423, + "learning_rate": 4.921506725498302e-05, + "loss": 0.5909, + "num_input_tokens_seen": 4865424, + "step": 3200 + }, + { + "epoch": 0.10242622111260483, + "eval_loss": 0.6562412977218628, + "eval_runtime": 557.9765, + "eval_samples_per_second": 24.886, + "eval_steps_per_second": 12.443, + "num_input_tokens_seen": 4865424, + "step": 3200 + }, + { + "epoch": 0.10258626208309327, + "grad_norm": 0.755262017250061, + "learning_rate": 4.9212624627499994e-05, + "loss": 0.3954, + "num_input_tokens_seen": 4872912, + "step": 3205 + }, + { + "epoch": 0.10274630305358172, + "grad_norm": 0.9039387702941895, + "learning_rate": 4.921017826612597e-05, + "loss": 0.5637, + "num_input_tokens_seen": 4880640, + "step": 3210 + }, + { + "epoch": 0.10290634402407016, + "grad_norm": 0.8064532279968262, + "learning_rate": 4.9207728171238223e-05, + "loss": 0.5181, + "num_input_tokens_seen": 4888672, + "step": 3215 + }, + { + "epoch": 0.10306638499455861, + "grad_norm": 2.840531349182129, + "learning_rate": 4.920527434321458e-05, + "loss": 0.6544, + "num_input_tokens_seen": 4896560, + "step": 3220 + }, + { + "epoch": 0.10322642596504705, + "grad_norm": 0.5805821418762207, + "learning_rate": 4.920281678243345e-05, + "loss": 0.634, + "num_input_tokens_seen": 4903968, + "step": 3225 + }, + { + "epoch": 0.10338646693553549, + "grad_norm": 1.148572564125061, + "learning_rate": 4.920035548927381e-05, + "loss": 0.6686, + "num_input_tokens_seen": 4911520, + "step": 3230 + }, + { + "epoch": 0.10354650790602395, + "grad_norm": 1.0593619346618652, + "learning_rate": 4.919789046411525e-05, + "loss": 0.9145, + "num_input_tokens_seen": 4919008, + "step": 3235 + }, + { + "epoch": 0.10370654887651239, + "grad_norm": 1.2881971597671509, + "learning_rate": 4.919542170733787e-05, + "loss": 0.7286, + "num_input_tokens_seen": 4926448, + "step": 3240 + }, + { + "epoch": 0.10386658984700083, + "grad_norm": 0.7800453901290894, + "learning_rate": 4.919294921932242e-05, + "loss": 0.6347, + "num_input_tokens_seen": 4934112, + "step": 3245 + }, + { + "epoch": 0.10402663081748928, + "grad_norm": 1.3310350179672241, + "learning_rate": 4.919047300045016e-05, + "loss": 0.6684, + "num_input_tokens_seen": 4941712, + "step": 3250 + }, + { + "epoch": 0.10418667178797772, + "grad_norm": 1.460924744606018, + "learning_rate": 4.918799305110299e-05, + "loss": 0.8116, + "num_input_tokens_seen": 4949088, + "step": 3255 + }, + { + "epoch": 0.10434671275846617, + "grad_norm": 1.1152271032333374, + "learning_rate": 4.918550937166331e-05, + "loss": 0.6019, + "num_input_tokens_seen": 4956656, + "step": 3260 + }, + { + "epoch": 0.10450675372895461, + "grad_norm": 0.8659340739250183, + "learning_rate": 4.918302196251415e-05, + "loss": 0.6346, + "num_input_tokens_seen": 4964160, + "step": 3265 + }, + { + "epoch": 0.10466679469944305, + "grad_norm": 0.9172273874282837, + "learning_rate": 4.91805308240391e-05, + "loss": 0.6592, + "num_input_tokens_seen": 4971584, + "step": 3270 + }, + { + "epoch": 0.1048268356699315, + "grad_norm": 0.9049335718154907, + "learning_rate": 4.9178035956622326e-05, + "loss": 0.5911, + "num_input_tokens_seen": 4979168, + "step": 3275 + }, + { + "epoch": 0.10498687664041995, + "grad_norm": 0.4858700931072235, + "learning_rate": 4.917553736064857e-05, + "loss": 0.5839, + "num_input_tokens_seen": 4986800, + "step": 3280 + }, + { + "epoch": 0.10514691761090839, + "grad_norm": 0.9532343149185181, + "learning_rate": 4.917303503650314e-05, + "loss": 0.5807, + "num_input_tokens_seen": 4993760, + "step": 3285 + }, + { + "epoch": 0.10530695858139684, + "grad_norm": 1.1181119680404663, + "learning_rate": 4.917052898457194e-05, + "loss": 0.657, + "num_input_tokens_seen": 5001120, + "step": 3290 + }, + { + "epoch": 0.10546699955188528, + "grad_norm": 0.8373987674713135, + "learning_rate": 4.916801920524141e-05, + "loss": 0.7836, + "num_input_tokens_seen": 5008848, + "step": 3295 + }, + { + "epoch": 0.10562704052237373, + "grad_norm": 0.6870822906494141, + "learning_rate": 4.916550569889862e-05, + "loss": 0.7124, + "num_input_tokens_seen": 5016528, + "step": 3300 + }, + { + "epoch": 0.10578708149286217, + "grad_norm": 0.7166744470596313, + "learning_rate": 4.916298846593116e-05, + "loss": 0.7074, + "num_input_tokens_seen": 5023984, + "step": 3305 + }, + { + "epoch": 0.10594712246335061, + "grad_norm": 1.455855369567871, + "learning_rate": 4.916046750672722e-05, + "loss": 0.7056, + "num_input_tokens_seen": 5031120, + "step": 3310 + }, + { + "epoch": 0.10610716343383907, + "grad_norm": 1.1060985326766968, + "learning_rate": 4.915794282167559e-05, + "loss": 0.6544, + "num_input_tokens_seen": 5038784, + "step": 3315 + }, + { + "epoch": 0.1062672044043275, + "grad_norm": 1.0893803834915161, + "learning_rate": 4.915541441116558e-05, + "loss": 0.6173, + "num_input_tokens_seen": 5046992, + "step": 3320 + }, + { + "epoch": 0.10642724537481596, + "grad_norm": 1.1891427040100098, + "learning_rate": 4.915288227558711e-05, + "loss": 0.5816, + "num_input_tokens_seen": 5054576, + "step": 3325 + }, + { + "epoch": 0.1065872863453044, + "grad_norm": 0.8833869099617004, + "learning_rate": 4.915034641533066e-05, + "loss": 0.6664, + "num_input_tokens_seen": 5062032, + "step": 3330 + }, + { + "epoch": 0.10674732731579284, + "grad_norm": 1.056496024131775, + "learning_rate": 4.914780683078731e-05, + "loss": 0.4736, + "num_input_tokens_seen": 5069472, + "step": 3335 + }, + { + "epoch": 0.10690736828628129, + "grad_norm": 1.9408706426620483, + "learning_rate": 4.9145263522348695e-05, + "loss": 0.9102, + "num_input_tokens_seen": 5077232, + "step": 3340 + }, + { + "epoch": 0.10706740925676973, + "grad_norm": 0.9604783058166504, + "learning_rate": 4.9142716490407e-05, + "loss": 0.4756, + "num_input_tokens_seen": 5085232, + "step": 3345 + }, + { + "epoch": 0.10722745022725817, + "grad_norm": 0.9406498670578003, + "learning_rate": 4.914016573535504e-05, + "loss": 0.599, + "num_input_tokens_seen": 5093584, + "step": 3350 + }, + { + "epoch": 0.10738749119774663, + "grad_norm": 0.6419205069541931, + "learning_rate": 4.9137611257586154e-05, + "loss": 0.5535, + "num_input_tokens_seen": 5101040, + "step": 3355 + }, + { + "epoch": 0.10754753216823507, + "grad_norm": 1.1089245080947876, + "learning_rate": 4.9135053057494274e-05, + "loss": 0.6042, + "num_input_tokens_seen": 5108576, + "step": 3360 + }, + { + "epoch": 0.10770757313872352, + "grad_norm": 1.0419074296951294, + "learning_rate": 4.913249113547392e-05, + "loss": 0.4958, + "num_input_tokens_seen": 5116064, + "step": 3365 + }, + { + "epoch": 0.10786761410921196, + "grad_norm": 0.8760762810707092, + "learning_rate": 4.912992549192016e-05, + "loss": 0.478, + "num_input_tokens_seen": 5123568, + "step": 3370 + }, + { + "epoch": 0.1080276550797004, + "grad_norm": 1.593248963356018, + "learning_rate": 4.9127356127228665e-05, + "loss": 1.0025, + "num_input_tokens_seen": 5131264, + "step": 3375 + }, + { + "epoch": 0.10818769605018885, + "grad_norm": 2.4648003578186035, + "learning_rate": 4.912478304179564e-05, + "loss": 0.7202, + "num_input_tokens_seen": 5138784, + "step": 3380 + }, + { + "epoch": 0.10834773702067729, + "grad_norm": 1.7377219200134277, + "learning_rate": 4.9122206236017896e-05, + "loss": 0.5781, + "num_input_tokens_seen": 5146128, + "step": 3385 + }, + { + "epoch": 0.10850777799116573, + "grad_norm": 1.5799615383148193, + "learning_rate": 4.911962571029282e-05, + "loss": 0.7677, + "num_input_tokens_seen": 5153664, + "step": 3390 + }, + { + "epoch": 0.10866781896165419, + "grad_norm": 0.7500333786010742, + "learning_rate": 4.9117041465018353e-05, + "loss": 0.5558, + "num_input_tokens_seen": 5160832, + "step": 3395 + }, + { + "epoch": 0.10882785993214263, + "grad_norm": 0.8144181370735168, + "learning_rate": 4.911445350059302e-05, + "loss": 0.6038, + "num_input_tokens_seen": 5169024, + "step": 3400 + }, + { + "epoch": 0.10882785993214263, + "eval_loss": 0.6508111953735352, + "eval_runtime": 562.4854, + "eval_samples_per_second": 24.687, + "eval_steps_per_second": 12.343, + "num_input_tokens_seen": 5169024, + "step": 3400 + }, + { + "epoch": 0.10898790090263108, + "grad_norm": 0.6357820630073547, + "learning_rate": 4.9111861817415905e-05, + "loss": 0.6895, + "num_input_tokens_seen": 5176480, + "step": 3405 + }, + { + "epoch": 0.10914794187311952, + "grad_norm": 1.3349807262420654, + "learning_rate": 4.91092664158867e-05, + "loss": 0.6536, + "num_input_tokens_seen": 5183744, + "step": 3410 + }, + { + "epoch": 0.10930798284360796, + "grad_norm": 2.2460858821868896, + "learning_rate": 4.910666729640563e-05, + "loss": 0.934, + "num_input_tokens_seen": 5191776, + "step": 3415 + }, + { + "epoch": 0.10946802381409641, + "grad_norm": 0.9225485324859619, + "learning_rate": 4.910406445937353e-05, + "loss": 0.5862, + "num_input_tokens_seen": 5199408, + "step": 3420 + }, + { + "epoch": 0.10962806478458485, + "grad_norm": 1.5031101703643799, + "learning_rate": 4.9101457905191774e-05, + "loss": 0.7023, + "num_input_tokens_seen": 5207328, + "step": 3425 + }, + { + "epoch": 0.10978810575507329, + "grad_norm": 0.7876431345939636, + "learning_rate": 4.909884763426233e-05, + "loss": 0.6317, + "num_input_tokens_seen": 5214720, + "step": 3430 + }, + { + "epoch": 0.10994814672556175, + "grad_norm": 1.0854765176773071, + "learning_rate": 4.9096233646987736e-05, + "loss": 0.745, + "num_input_tokens_seen": 5223360, + "step": 3435 + }, + { + "epoch": 0.11010818769605019, + "grad_norm": 1.6316241025924683, + "learning_rate": 4.9093615943771104e-05, + "loss": 0.8822, + "num_input_tokens_seen": 5231792, + "step": 3440 + }, + { + "epoch": 0.11026822866653864, + "grad_norm": 2.530510663986206, + "learning_rate": 4.909099452501611e-05, + "loss": 0.7337, + "num_input_tokens_seen": 5238864, + "step": 3445 + }, + { + "epoch": 0.11042826963702708, + "grad_norm": 1.117913007736206, + "learning_rate": 4.908836939112702e-05, + "loss": 0.6767, + "num_input_tokens_seen": 5246448, + "step": 3450 + }, + { + "epoch": 0.11058831060751552, + "grad_norm": 0.4648854434490204, + "learning_rate": 4.908574054250865e-05, + "loss": 0.382, + "num_input_tokens_seen": 5254256, + "step": 3455 + }, + { + "epoch": 0.11074835157800397, + "grad_norm": 0.6582554578781128, + "learning_rate": 4.9083107979566414e-05, + "loss": 0.4719, + "num_input_tokens_seen": 5261952, + "step": 3460 + }, + { + "epoch": 0.11090839254849241, + "grad_norm": 0.8473914861679077, + "learning_rate": 4.908047170270628e-05, + "loss": 0.5436, + "num_input_tokens_seen": 5269920, + "step": 3465 + }, + { + "epoch": 0.11106843351898085, + "grad_norm": 0.6684194207191467, + "learning_rate": 4.9077831712334784e-05, + "loss": 0.6957, + "num_input_tokens_seen": 5277056, + "step": 3470 + }, + { + "epoch": 0.1112284744894693, + "grad_norm": 0.8911676406860352, + "learning_rate": 4.907518800885907e-05, + "loss": 0.5367, + "num_input_tokens_seen": 5284544, + "step": 3475 + }, + { + "epoch": 0.11138851545995775, + "grad_norm": 0.8050469160079956, + "learning_rate": 4.907254059268681e-05, + "loss": 0.786, + "num_input_tokens_seen": 5292400, + "step": 3480 + }, + { + "epoch": 0.1115485564304462, + "grad_norm": 0.5424482822418213, + "learning_rate": 4.906988946422628e-05, + "loss": 0.6754, + "num_input_tokens_seen": 5300032, + "step": 3485 + }, + { + "epoch": 0.11170859740093464, + "grad_norm": 0.9828670024871826, + "learning_rate": 4.9067234623886315e-05, + "loss": 0.6063, + "num_input_tokens_seen": 5307680, + "step": 3490 + }, + { + "epoch": 0.11186863837142308, + "grad_norm": 0.4401840269565582, + "learning_rate": 4.9064576072076316e-05, + "loss": 0.5254, + "num_input_tokens_seen": 5315408, + "step": 3495 + }, + { + "epoch": 0.11202867934191153, + "grad_norm": 1.0345547199249268, + "learning_rate": 4.906191380920628e-05, + "loss": 0.536, + "num_input_tokens_seen": 5323392, + "step": 3500 + }, + { + "epoch": 0.11218872031239997, + "grad_norm": 0.841712236404419, + "learning_rate": 4.905924783568675e-05, + "loss": 0.7682, + "num_input_tokens_seen": 5331392, + "step": 3505 + }, + { + "epoch": 0.11234876128288843, + "grad_norm": 0.56590735912323, + "learning_rate": 4.905657815192886e-05, + "loss": 0.6901, + "num_input_tokens_seen": 5338976, + "step": 3510 + }, + { + "epoch": 0.11250880225337687, + "grad_norm": 1.0578559637069702, + "learning_rate": 4.90539047583443e-05, + "loss": 0.5091, + "num_input_tokens_seen": 5346528, + "step": 3515 + }, + { + "epoch": 0.1126688432238653, + "grad_norm": 2.5697193145751953, + "learning_rate": 4.905122765534534e-05, + "loss": 0.7738, + "num_input_tokens_seen": 5353664, + "step": 3520 + }, + { + "epoch": 0.11282888419435376, + "grad_norm": 1.4484422206878662, + "learning_rate": 4.9048546843344846e-05, + "loss": 0.7523, + "num_input_tokens_seen": 5362240, + "step": 3525 + }, + { + "epoch": 0.1129889251648422, + "grad_norm": 1.3049415349960327, + "learning_rate": 4.9045862322756206e-05, + "loss": 0.7128, + "num_input_tokens_seen": 5369920, + "step": 3530 + }, + { + "epoch": 0.11314896613533064, + "grad_norm": 1.425256371498108, + "learning_rate": 4.904317409399342e-05, + "loss": 0.6992, + "num_input_tokens_seen": 5377536, + "step": 3535 + }, + { + "epoch": 0.11330900710581909, + "grad_norm": 0.9161001443862915, + "learning_rate": 4.904048215747104e-05, + "loss": 0.5942, + "num_input_tokens_seen": 5384768, + "step": 3540 + }, + { + "epoch": 0.11346904807630753, + "grad_norm": 1.5137766599655151, + "learning_rate": 4.90377865136042e-05, + "loss": 0.8521, + "num_input_tokens_seen": 5395728, + "step": 3545 + }, + { + "epoch": 0.11362908904679599, + "grad_norm": 0.7787429094314575, + "learning_rate": 4.90350871628086e-05, + "loss": 0.6831, + "num_input_tokens_seen": 5403008, + "step": 3550 + }, + { + "epoch": 0.11378913001728443, + "grad_norm": 0.6379167437553406, + "learning_rate": 4.903238410550052e-05, + "loss": 0.5709, + "num_input_tokens_seen": 5410768, + "step": 3555 + }, + { + "epoch": 0.11394917098777287, + "grad_norm": 0.5977700352668762, + "learning_rate": 4.90296773420968e-05, + "loss": 0.5889, + "num_input_tokens_seen": 5418128, + "step": 3560 + }, + { + "epoch": 0.11410921195826132, + "grad_norm": 1.4209635257720947, + "learning_rate": 4.902696687301486e-05, + "loss": 0.6095, + "num_input_tokens_seen": 5425760, + "step": 3565 + }, + { + "epoch": 0.11426925292874976, + "grad_norm": 1.2601799964904785, + "learning_rate": 4.902425269867268e-05, + "loss": 0.6398, + "num_input_tokens_seen": 5433072, + "step": 3570 + }, + { + "epoch": 0.1144292938992382, + "grad_norm": 0.533749520778656, + "learning_rate": 4.902153481948883e-05, + "loss": 0.6237, + "num_input_tokens_seen": 5440384, + "step": 3575 + }, + { + "epoch": 0.11458933486972665, + "grad_norm": 1.7099257707595825, + "learning_rate": 4.901881323588244e-05, + "loss": 0.7282, + "num_input_tokens_seen": 5447760, + "step": 3580 + }, + { + "epoch": 0.11474937584021509, + "grad_norm": 1.2420058250427246, + "learning_rate": 4.90160879482732e-05, + "loss": 0.7292, + "num_input_tokens_seen": 5455360, + "step": 3585 + }, + { + "epoch": 0.11490941681070355, + "grad_norm": 0.6476201415061951, + "learning_rate": 4.9013358957081405e-05, + "loss": 0.6748, + "num_input_tokens_seen": 5462960, + "step": 3590 + }, + { + "epoch": 0.11506945778119199, + "grad_norm": 1.5280050039291382, + "learning_rate": 4.901062626272789e-05, + "loss": 0.7108, + "num_input_tokens_seen": 5470288, + "step": 3595 + }, + { + "epoch": 0.11522949875168043, + "grad_norm": 1.7465983629226685, + "learning_rate": 4.900788986563406e-05, + "loss": 0.4876, + "num_input_tokens_seen": 5477808, + "step": 3600 + }, + { + "epoch": 0.11522949875168043, + "eval_loss": 0.644148588180542, + "eval_runtime": 559.9917, + "eval_samples_per_second": 24.797, + "eval_steps_per_second": 12.398, + "num_input_tokens_seen": 5477808, + "step": 3600 + }, + { + "epoch": 0.11538953972216888, + "grad_norm": 0.8776022791862488, + "learning_rate": 4.9005149766221915e-05, + "loss": 0.6051, + "num_input_tokens_seen": 5485536, + "step": 3605 + }, + { + "epoch": 0.11554958069265732, + "grad_norm": 0.6347915530204773, + "learning_rate": 4.9002405964914e-05, + "loss": 0.5257, + "num_input_tokens_seen": 5493632, + "step": 3610 + }, + { + "epoch": 0.11570962166314576, + "grad_norm": 1.7645628452301025, + "learning_rate": 4.899965846213346e-05, + "loss": 0.5407, + "num_input_tokens_seen": 5501200, + "step": 3615 + }, + { + "epoch": 0.11586966263363421, + "grad_norm": 2.0732455253601074, + "learning_rate": 4.899690725830399e-05, + "loss": 0.6273, + "num_input_tokens_seen": 5509088, + "step": 3620 + }, + { + "epoch": 0.11602970360412265, + "grad_norm": 0.6791529655456543, + "learning_rate": 4.899415235384985e-05, + "loss": 0.6532, + "num_input_tokens_seen": 5517024, + "step": 3625 + }, + { + "epoch": 0.1161897445746111, + "grad_norm": 0.6580854654312134, + "learning_rate": 4.899139374919589e-05, + "loss": 0.5791, + "num_input_tokens_seen": 5524320, + "step": 3630 + }, + { + "epoch": 0.11634978554509955, + "grad_norm": 1.373570442199707, + "learning_rate": 4.898863144476752e-05, + "loss": 0.6505, + "num_input_tokens_seen": 5532528, + "step": 3635 + }, + { + "epoch": 0.11650982651558799, + "grad_norm": 1.1487774848937988, + "learning_rate": 4.898586544099072e-05, + "loss": 0.6097, + "num_input_tokens_seen": 5540112, + "step": 3640 + }, + { + "epoch": 0.11666986748607644, + "grad_norm": 0.8668302297592163, + "learning_rate": 4.898309573829204e-05, + "loss": 0.5691, + "num_input_tokens_seen": 5547568, + "step": 3645 + }, + { + "epoch": 0.11682990845656488, + "grad_norm": 1.1509904861450195, + "learning_rate": 4.898032233709862e-05, + "loss": 0.5556, + "num_input_tokens_seen": 5555600, + "step": 3650 + }, + { + "epoch": 0.11698994942705332, + "grad_norm": 1.4000558853149414, + "learning_rate": 4.8977545237838123e-05, + "loss": 0.7334, + "num_input_tokens_seen": 5562880, + "step": 3655 + }, + { + "epoch": 0.11714999039754177, + "grad_norm": 1.2417837381362915, + "learning_rate": 4.8974764440938836e-05, + "loss": 0.5442, + "num_input_tokens_seen": 5570160, + "step": 3660 + }, + { + "epoch": 0.11731003136803021, + "grad_norm": 0.9305194616317749, + "learning_rate": 4.897197994682959e-05, + "loss": 0.5083, + "num_input_tokens_seen": 5577808, + "step": 3665 + }, + { + "epoch": 0.11747007233851867, + "grad_norm": 0.7630409002304077, + "learning_rate": 4.8969191755939786e-05, + "loss": 0.5342, + "num_input_tokens_seen": 5585952, + "step": 3670 + }, + { + "epoch": 0.1176301133090071, + "grad_norm": 0.9291691780090332, + "learning_rate": 4.8966399868699396e-05, + "loss": 0.8311, + "num_input_tokens_seen": 5593824, + "step": 3675 + }, + { + "epoch": 0.11779015427949555, + "grad_norm": 0.7527806162834167, + "learning_rate": 4.8963604285538965e-05, + "loss": 0.5404, + "num_input_tokens_seen": 5601904, + "step": 3680 + }, + { + "epoch": 0.117950195249984, + "grad_norm": 0.7207104563713074, + "learning_rate": 4.8960805006889604e-05, + "loss": 0.5496, + "num_input_tokens_seen": 5610576, + "step": 3685 + }, + { + "epoch": 0.11811023622047244, + "grad_norm": 0.5043500065803528, + "learning_rate": 4.8958002033183004e-05, + "loss": 0.5191, + "num_input_tokens_seen": 5618416, + "step": 3690 + }, + { + "epoch": 0.11827027719096088, + "grad_norm": 1.2624499797821045, + "learning_rate": 4.8955195364851414e-05, + "loss": 0.6421, + "num_input_tokens_seen": 5626496, + "step": 3695 + }, + { + "epoch": 0.11843031816144933, + "grad_norm": 0.49169686436653137, + "learning_rate": 4.895238500232766e-05, + "loss": 0.6783, + "num_input_tokens_seen": 5633856, + "step": 3700 + }, + { + "epoch": 0.11859035913193777, + "grad_norm": 0.7290477156639099, + "learning_rate": 4.8949570946045143e-05, + "loss": 0.5954, + "num_input_tokens_seen": 5641104, + "step": 3705 + }, + { + "epoch": 0.11875040010242623, + "grad_norm": 0.8549315929412842, + "learning_rate": 4.89467531964378e-05, + "loss": 0.4586, + "num_input_tokens_seen": 5648288, + "step": 3710 + }, + { + "epoch": 0.11891044107291467, + "grad_norm": 1.053047776222229, + "learning_rate": 4.894393175394019e-05, + "loss": 0.5818, + "num_input_tokens_seen": 5656016, + "step": 3715 + }, + { + "epoch": 0.1190704820434031, + "grad_norm": 1.6291310787200928, + "learning_rate": 4.8941106618987406e-05, + "loss": 0.6525, + "num_input_tokens_seen": 5663472, + "step": 3720 + }, + { + "epoch": 0.11923052301389156, + "grad_norm": 0.9998893141746521, + "learning_rate": 4.893827779201512e-05, + "loss": 0.6516, + "num_input_tokens_seen": 5671248, + "step": 3725 + }, + { + "epoch": 0.11939056398438, + "grad_norm": 1.1796724796295166, + "learning_rate": 4.893544527345957e-05, + "loss": 0.6239, + "num_input_tokens_seen": 5678384, + "step": 3730 + }, + { + "epoch": 0.11955060495486845, + "grad_norm": 0.8463683724403381, + "learning_rate": 4.8932609063757563e-05, + "loss": 0.433, + "num_input_tokens_seen": 5686192, + "step": 3735 + }, + { + "epoch": 0.11971064592535689, + "grad_norm": 1.4423551559448242, + "learning_rate": 4.8929769163346484e-05, + "loss": 0.6527, + "num_input_tokens_seen": 5694016, + "step": 3740 + }, + { + "epoch": 0.11987068689584533, + "grad_norm": 1.0632685422897339, + "learning_rate": 4.892692557266429e-05, + "loss": 0.5213, + "num_input_tokens_seen": 5701568, + "step": 3745 + }, + { + "epoch": 0.12003072786633379, + "grad_norm": 0.5637761950492859, + "learning_rate": 4.8924078292149464e-05, + "loss": 0.6676, + "num_input_tokens_seen": 5709600, + "step": 3750 + }, + { + "epoch": 0.12019076883682223, + "grad_norm": 0.768598735332489, + "learning_rate": 4.892122732224114e-05, + "loss": 0.611, + "num_input_tokens_seen": 5717408, + "step": 3755 + }, + { + "epoch": 0.12035080980731067, + "grad_norm": 1.2737919092178345, + "learning_rate": 4.8918372663378944e-05, + "loss": 0.7149, + "num_input_tokens_seen": 5725088, + "step": 3760 + }, + { + "epoch": 0.12051085077779912, + "grad_norm": 1.0466620922088623, + "learning_rate": 4.89155143160031e-05, + "loss": 1.1155, + "num_input_tokens_seen": 5733232, + "step": 3765 + }, + { + "epoch": 0.12067089174828756, + "grad_norm": 1.0855050086975098, + "learning_rate": 4.891265228055441e-05, + "loss": 0.7016, + "num_input_tokens_seen": 5740336, + "step": 3770 + }, + { + "epoch": 0.12083093271877601, + "grad_norm": 1.2598704099655151, + "learning_rate": 4.890978655747424e-05, + "loss": 0.6461, + "num_input_tokens_seen": 5748000, + "step": 3775 + }, + { + "epoch": 0.12099097368926445, + "grad_norm": 1.752120852470398, + "learning_rate": 4.89069171472045e-05, + "loss": 0.7009, + "num_input_tokens_seen": 5755488, + "step": 3780 + }, + { + "epoch": 0.12115101465975289, + "grad_norm": 0.9451640248298645, + "learning_rate": 4.890404405018772e-05, + "loss": 0.5598, + "num_input_tokens_seen": 5762736, + "step": 3785 + }, + { + "epoch": 0.12131105563024135, + "grad_norm": 0.786924421787262, + "learning_rate": 4.8901167266866934e-05, + "loss": 0.5267, + "num_input_tokens_seen": 5770064, + "step": 3790 + }, + { + "epoch": 0.12147109660072979, + "grad_norm": 0.8308252096176147, + "learning_rate": 4.88982867976858e-05, + "loss": 0.6244, + "num_input_tokens_seen": 5777616, + "step": 3795 + }, + { + "epoch": 0.12163113757121823, + "grad_norm": 1.7096140384674072, + "learning_rate": 4.889540264308852e-05, + "loss": 0.7048, + "num_input_tokens_seen": 5785408, + "step": 3800 + }, + { + "epoch": 0.12163113757121823, + "eval_loss": 0.6393736004829407, + "eval_runtime": 559.2389, + "eval_samples_per_second": 24.83, + "eval_steps_per_second": 12.415, + "num_input_tokens_seen": 5785408, + "step": 3800 + }, + { + "epoch": 0.12179117854170668, + "grad_norm": 1.106424331665039, + "learning_rate": 4.889251480351986e-05, + "loss": 0.666, + "num_input_tokens_seen": 5792880, + "step": 3805 + }, + { + "epoch": 0.12195121951219512, + "grad_norm": 1.5878840684890747, + "learning_rate": 4.888962327942517e-05, + "loss": 0.6321, + "num_input_tokens_seen": 5800432, + "step": 3810 + }, + { + "epoch": 0.12211126048268357, + "grad_norm": 0.5413963198661804, + "learning_rate": 4.8886728071250356e-05, + "loss": 0.4909, + "num_input_tokens_seen": 5808064, + "step": 3815 + }, + { + "epoch": 0.12227130145317201, + "grad_norm": 1.462110161781311, + "learning_rate": 4.8883829179441884e-05, + "loss": 0.5808, + "num_input_tokens_seen": 5815936, + "step": 3820 + }, + { + "epoch": 0.12243134242366045, + "grad_norm": 0.5704402327537537, + "learning_rate": 4.888092660444682e-05, + "loss": 0.626, + "num_input_tokens_seen": 5823264, + "step": 3825 + }, + { + "epoch": 0.1225913833941489, + "grad_norm": 0.8582004308700562, + "learning_rate": 4.887802034671276e-05, + "loss": 0.4899, + "num_input_tokens_seen": 5830960, + "step": 3830 + }, + { + "epoch": 0.12275142436463735, + "grad_norm": 0.866195797920227, + "learning_rate": 4.88751104066879e-05, + "loss": 0.5676, + "num_input_tokens_seen": 5839568, + "step": 3835 + }, + { + "epoch": 0.12291146533512579, + "grad_norm": 2.022921562194824, + "learning_rate": 4.887219678482098e-05, + "loss": 0.5942, + "num_input_tokens_seen": 5847024, + "step": 3840 + }, + { + "epoch": 0.12307150630561424, + "grad_norm": 0.957573413848877, + "learning_rate": 4.8869279481561316e-05, + "loss": 0.558, + "num_input_tokens_seen": 5854608, + "step": 3845 + }, + { + "epoch": 0.12323154727610268, + "grad_norm": 0.7141245603561401, + "learning_rate": 4.88663584973588e-05, + "loss": 0.4082, + "num_input_tokens_seen": 5861952, + "step": 3850 + }, + { + "epoch": 0.12339158824659113, + "grad_norm": 1.1146990060806274, + "learning_rate": 4.8863433832663874e-05, + "loss": 0.5224, + "num_input_tokens_seen": 5869232, + "step": 3855 + }, + { + "epoch": 0.12355162921707957, + "grad_norm": 2.753453016281128, + "learning_rate": 4.886050548792757e-05, + "loss": 0.7209, + "num_input_tokens_seen": 5876720, + "step": 3860 + }, + { + "epoch": 0.12371167018756801, + "grad_norm": 1.1533968448638916, + "learning_rate": 4.8857573463601465e-05, + "loss": 0.5562, + "num_input_tokens_seen": 5884160, + "step": 3865 + }, + { + "epoch": 0.12387171115805647, + "grad_norm": 1.2452203035354614, + "learning_rate": 4.885463776013772e-05, + "loss": 0.7575, + "num_input_tokens_seen": 5892048, + "step": 3870 + }, + { + "epoch": 0.1240317521285449, + "grad_norm": 1.6826434135437012, + "learning_rate": 4.8851698377989056e-05, + "loss": 0.6732, + "num_input_tokens_seen": 5899520, + "step": 3875 + }, + { + "epoch": 0.12419179309903335, + "grad_norm": 0.8104444742202759, + "learning_rate": 4.884875531760876e-05, + "loss": 0.5994, + "num_input_tokens_seen": 5906496, + "step": 3880 + }, + { + "epoch": 0.1243518340695218, + "grad_norm": 1.092221975326538, + "learning_rate": 4.88458085794507e-05, + "loss": 0.6566, + "num_input_tokens_seen": 5914032, + "step": 3885 + }, + { + "epoch": 0.12451187504001024, + "grad_norm": 1.1017259359359741, + "learning_rate": 4.884285816396929e-05, + "loss": 0.9221, + "num_input_tokens_seen": 5921648, + "step": 3890 + }, + { + "epoch": 0.12467191601049869, + "grad_norm": 0.813228964805603, + "learning_rate": 4.8839904071619526e-05, + "loss": 0.5201, + "num_input_tokens_seen": 5929456, + "step": 3895 + }, + { + "epoch": 0.12483195698098713, + "grad_norm": 0.83821702003479, + "learning_rate": 4.8836946302856955e-05, + "loss": 0.5325, + "num_input_tokens_seen": 5937328, + "step": 3900 + }, + { + "epoch": 0.12499199795147557, + "grad_norm": 0.767419695854187, + "learning_rate": 4.8833984858137715e-05, + "loss": 0.6027, + "num_input_tokens_seen": 5944944, + "step": 3905 + }, + { + "epoch": 0.12515203892196403, + "grad_norm": 0.5647637248039246, + "learning_rate": 4.8831019737918494e-05, + "loss": 0.6212, + "num_input_tokens_seen": 5953296, + "step": 3910 + }, + { + "epoch": 0.12531207989245247, + "grad_norm": 1.7678123712539673, + "learning_rate": 4.882805094265655e-05, + "loss": 0.64, + "num_input_tokens_seen": 5961152, + "step": 3915 + }, + { + "epoch": 0.1254721208629409, + "grad_norm": 1.0179332494735718, + "learning_rate": 4.8825078472809706e-05, + "loss": 0.6933, + "num_input_tokens_seen": 5968448, + "step": 3920 + }, + { + "epoch": 0.12563216183342935, + "grad_norm": 0.6071088910102844, + "learning_rate": 4.882210232883635e-05, + "loss": 0.58, + "num_input_tokens_seen": 5975936, + "step": 3925 + }, + { + "epoch": 0.1257922028039178, + "grad_norm": 0.9118504524230957, + "learning_rate": 4.881912251119546e-05, + "loss": 0.5681, + "num_input_tokens_seen": 5983328, + "step": 3930 + }, + { + "epoch": 0.12595224377440625, + "grad_norm": 0.535110354423523, + "learning_rate": 4.881613902034654e-05, + "loss": 0.576, + "num_input_tokens_seen": 5991168, + "step": 3935 + }, + { + "epoch": 0.1261122847448947, + "grad_norm": 1.0623037815093994, + "learning_rate": 4.88131518567497e-05, + "loss": 0.5875, + "num_input_tokens_seen": 5999088, + "step": 3940 + }, + { + "epoch": 0.12627232571538313, + "grad_norm": 0.7299715876579285, + "learning_rate": 4.881016102086558e-05, + "loss": 0.6131, + "num_input_tokens_seen": 6006320, + "step": 3945 + }, + { + "epoch": 0.12643236668587157, + "grad_norm": 1.5713506937026978, + "learning_rate": 4.8807166513155425e-05, + "loss": 0.7205, + "num_input_tokens_seen": 6014496, + "step": 3950 + }, + { + "epoch": 0.12659240765636004, + "grad_norm": 1.6739890575408936, + "learning_rate": 4.8804168334081004e-05, + "loss": 0.5816, + "num_input_tokens_seen": 6022192, + "step": 3955 + }, + { + "epoch": 0.12675244862684848, + "grad_norm": 1.401807188987732, + "learning_rate": 4.880116648410468e-05, + "loss": 0.6816, + "num_input_tokens_seen": 6029808, + "step": 3960 + }, + { + "epoch": 0.12691248959733692, + "grad_norm": 0.6883082389831543, + "learning_rate": 4.879816096368939e-05, + "loss": 0.5731, + "num_input_tokens_seen": 6037344, + "step": 3965 + }, + { + "epoch": 0.12707253056782536, + "grad_norm": 1.193999171257019, + "learning_rate": 4.879515177329861e-05, + "loss": 0.4965, + "num_input_tokens_seen": 6045248, + "step": 3970 + }, + { + "epoch": 0.1272325715383138, + "grad_norm": 1.3120501041412354, + "learning_rate": 4.8792138913396394e-05, + "loss": 0.7204, + "num_input_tokens_seen": 6053088, + "step": 3975 + }, + { + "epoch": 0.12739261250880227, + "grad_norm": 1.0619449615478516, + "learning_rate": 4.8789122384447374e-05, + "loss": 0.4467, + "num_input_tokens_seen": 6060784, + "step": 3980 + }, + { + "epoch": 0.1275526534792907, + "grad_norm": 0.9130566716194153, + "learning_rate": 4.878610218691673e-05, + "loss": 0.6665, + "num_input_tokens_seen": 6068432, + "step": 3985 + }, + { + "epoch": 0.12771269444977915, + "grad_norm": 2.0766713619232178, + "learning_rate": 4.87830783212702e-05, + "loss": 0.6411, + "num_input_tokens_seen": 6076176, + "step": 3990 + }, + { + "epoch": 0.12787273542026759, + "grad_norm": 0.7549774646759033, + "learning_rate": 4.878005078797413e-05, + "loss": 0.7824, + "num_input_tokens_seen": 6083776, + "step": 3995 + }, + { + "epoch": 0.12803277639075603, + "grad_norm": 1.118322730064392, + "learning_rate": 4.877701958749539e-05, + "loss": 0.5233, + "num_input_tokens_seen": 6091152, + "step": 4000 + }, + { + "epoch": 0.12803277639075603, + "eval_loss": 0.6348217725753784, + "eval_runtime": 557.4641, + "eval_samples_per_second": 24.909, + "eval_steps_per_second": 12.455, + "num_input_tokens_seen": 6091152, + "step": 4000 + }, + { + "epoch": 0.12819281736124447, + "grad_norm": 0.6995099782943726, + "learning_rate": 4.877398472030142e-05, + "loss": 0.5898, + "num_input_tokens_seen": 6098752, + "step": 4005 + }, + { + "epoch": 0.12835285833173293, + "grad_norm": 0.922370433807373, + "learning_rate": 4.877094618686024e-05, + "loss": 0.5739, + "num_input_tokens_seen": 6106144, + "step": 4010 + }, + { + "epoch": 0.12851289930222137, + "grad_norm": 0.9083023071289062, + "learning_rate": 4.876790398764045e-05, + "loss": 0.5901, + "num_input_tokens_seen": 6113440, + "step": 4015 + }, + { + "epoch": 0.1286729402727098, + "grad_norm": 1.6451681852340698, + "learning_rate": 4.8764858123111167e-05, + "loss": 0.6684, + "num_input_tokens_seen": 6121520, + "step": 4020 + }, + { + "epoch": 0.12883298124319825, + "grad_norm": 0.9024233222007751, + "learning_rate": 4.876180859374212e-05, + "loss": 0.6822, + "num_input_tokens_seen": 6128736, + "step": 4025 + }, + { + "epoch": 0.1289930222136867, + "grad_norm": 0.8361315727233887, + "learning_rate": 4.875875540000357e-05, + "loss": 0.5302, + "num_input_tokens_seen": 6136288, + "step": 4030 + }, + { + "epoch": 0.12915306318417516, + "grad_norm": 0.5274816155433655, + "learning_rate": 4.8755698542366376e-05, + "loss": 0.8094, + "num_input_tokens_seen": 6143568, + "step": 4035 + }, + { + "epoch": 0.1293131041546636, + "grad_norm": 1.5162042379379272, + "learning_rate": 4.875263802130193e-05, + "loss": 0.687, + "num_input_tokens_seen": 6150528, + "step": 4040 + }, + { + "epoch": 0.12947314512515204, + "grad_norm": 0.9229976534843445, + "learning_rate": 4.8749573837282207e-05, + "loss": 0.4564, + "num_input_tokens_seen": 6157840, + "step": 4045 + }, + { + "epoch": 0.12963318609564048, + "grad_norm": 1.613142728805542, + "learning_rate": 4.874650599077974e-05, + "loss": 0.6628, + "num_input_tokens_seen": 6164800, + "step": 4050 + }, + { + "epoch": 0.12979322706612892, + "grad_norm": 0.6491219401359558, + "learning_rate": 4.874343448226764e-05, + "loss": 0.5353, + "num_input_tokens_seen": 6172416, + "step": 4055 + }, + { + "epoch": 0.1299532680366174, + "grad_norm": 0.9546117186546326, + "learning_rate": 4.874035931221955e-05, + "loss": 0.4679, + "num_input_tokens_seen": 6179680, + "step": 4060 + }, + { + "epoch": 0.13011330900710583, + "grad_norm": 0.7317907214164734, + "learning_rate": 4.8737280481109724e-05, + "loss": 0.4562, + "num_input_tokens_seen": 6187744, + "step": 4065 + }, + { + "epoch": 0.13027334997759427, + "grad_norm": 0.9662302732467651, + "learning_rate": 4.873419798941294e-05, + "loss": 0.6758, + "num_input_tokens_seen": 6194960, + "step": 4070 + }, + { + "epoch": 0.1304333909480827, + "grad_norm": 1.0322964191436768, + "learning_rate": 4.873111183760458e-05, + "loss": 0.5126, + "num_input_tokens_seen": 6202848, + "step": 4075 + }, + { + "epoch": 0.13059343191857115, + "grad_norm": 0.8224974870681763, + "learning_rate": 4.8728022026160537e-05, + "loss": 0.6025, + "num_input_tokens_seen": 6210256, + "step": 4080 + }, + { + "epoch": 0.13075347288905959, + "grad_norm": 0.9157331585884094, + "learning_rate": 4.872492855555732e-05, + "loss": 0.6019, + "num_input_tokens_seen": 6217536, + "step": 4085 + }, + { + "epoch": 0.13091351385954805, + "grad_norm": 1.001854658126831, + "learning_rate": 4.8721831426271956e-05, + "loss": 0.4736, + "num_input_tokens_seen": 6225088, + "step": 4090 + }, + { + "epoch": 0.1310735548300365, + "grad_norm": 0.9502314329147339, + "learning_rate": 4.87187306387821e-05, + "loss": 0.5924, + "num_input_tokens_seen": 6232688, + "step": 4095 + }, + { + "epoch": 0.13123359580052493, + "grad_norm": 0.6673247218132019, + "learning_rate": 4.87156261935659e-05, + "loss": 0.5672, + "num_input_tokens_seen": 6240464, + "step": 4100 + }, + { + "epoch": 0.13139363677101337, + "grad_norm": 2.831580877304077, + "learning_rate": 4.871251809110211e-05, + "loss": 0.6741, + "num_input_tokens_seen": 6247632, + "step": 4105 + }, + { + "epoch": 0.1315536777415018, + "grad_norm": 0.6784937381744385, + "learning_rate": 4.8709406331870044e-05, + "loss": 0.6039, + "num_input_tokens_seen": 6255600, + "step": 4110 + }, + { + "epoch": 0.13171371871199028, + "grad_norm": 1.3386235237121582, + "learning_rate": 4.8706290916349574e-05, + "loss": 0.6842, + "num_input_tokens_seen": 6262896, + "step": 4115 + }, + { + "epoch": 0.13187375968247872, + "grad_norm": 0.31453797221183777, + "learning_rate": 4.8703171845021134e-05, + "loss": 0.5304, + "num_input_tokens_seen": 6270336, + "step": 4120 + }, + { + "epoch": 0.13203380065296716, + "grad_norm": 0.8607007265090942, + "learning_rate": 4.870004911836572e-05, + "loss": 0.94, + "num_input_tokens_seen": 6278112, + "step": 4125 + }, + { + "epoch": 0.1321938416234556, + "grad_norm": 0.4591667652130127, + "learning_rate": 4.869692273686489e-05, + "loss": 0.78, + "num_input_tokens_seen": 6285552, + "step": 4130 + }, + { + "epoch": 0.13235388259394404, + "grad_norm": 0.5752595067024231, + "learning_rate": 4.869379270100079e-05, + "loss": 0.3973, + "num_input_tokens_seen": 6293616, + "step": 4135 + }, + { + "epoch": 0.1325139235644325, + "grad_norm": 1.2379658222198486, + "learning_rate": 4.86906590112561e-05, + "loss": 0.5545, + "num_input_tokens_seen": 6301296, + "step": 4140 + }, + { + "epoch": 0.13267396453492095, + "grad_norm": 1.0152920484542847, + "learning_rate": 4.8687521668114064e-05, + "loss": 0.748, + "num_input_tokens_seen": 6308960, + "step": 4145 + }, + { + "epoch": 0.13283400550540939, + "grad_norm": 0.6682627201080322, + "learning_rate": 4.868438067205853e-05, + "loss": 0.8006, + "num_input_tokens_seen": 6316448, + "step": 4150 + }, + { + "epoch": 0.13299404647589783, + "grad_norm": 1.0796047449111938, + "learning_rate": 4.8681236023573844e-05, + "loss": 0.6349, + "num_input_tokens_seen": 6324416, + "step": 4155 + }, + { + "epoch": 0.13315408744638627, + "grad_norm": 0.6413664817810059, + "learning_rate": 4.867808772314497e-05, + "loss": 0.5929, + "num_input_tokens_seen": 6331728, + "step": 4160 + }, + { + "epoch": 0.13331412841687473, + "grad_norm": 0.46329566836357117, + "learning_rate": 4.867493577125741e-05, + "loss": 0.4932, + "num_input_tokens_seen": 6339056, + "step": 4165 + }, + { + "epoch": 0.13347416938736317, + "grad_norm": 0.5000333786010742, + "learning_rate": 4.867178016839725e-05, + "loss": 0.3495, + "num_input_tokens_seen": 6346992, + "step": 4170 + }, + { + "epoch": 0.1336342103578516, + "grad_norm": 1.191885232925415, + "learning_rate": 4.8668620915051094e-05, + "loss": 0.5322, + "num_input_tokens_seen": 6354272, + "step": 4175 + }, + { + "epoch": 0.13379425132834005, + "grad_norm": 0.6250351071357727, + "learning_rate": 4.866545801170616e-05, + "loss": 0.5797, + "num_input_tokens_seen": 6361776, + "step": 4180 + }, + { + "epoch": 0.1339542922988285, + "grad_norm": 0.6623409390449524, + "learning_rate": 4.86622914588502e-05, + "loss": 0.6375, + "num_input_tokens_seen": 6370048, + "step": 4185 + }, + { + "epoch": 0.13411433326931693, + "grad_norm": 1.7987505197525024, + "learning_rate": 4.865912125697154e-05, + "loss": 0.7134, + "num_input_tokens_seen": 6377696, + "step": 4190 + }, + { + "epoch": 0.1342743742398054, + "grad_norm": 1.2534252405166626, + "learning_rate": 4.865594740655907e-05, + "loss": 0.554, + "num_input_tokens_seen": 6385616, + "step": 4195 + }, + { + "epoch": 0.13443441521029384, + "grad_norm": 0.6230606436729431, + "learning_rate": 4.865276990810222e-05, + "loss": 0.7181, + "num_input_tokens_seen": 6393712, + "step": 4200 + }, + { + "epoch": 0.13443441521029384, + "eval_loss": 0.6323497295379639, + "eval_runtime": 559.4736, + "eval_samples_per_second": 24.82, + "eval_steps_per_second": 12.41, + "num_input_tokens_seen": 6393712, + "step": 4200 + }, + { + "epoch": 0.13459445618078228, + "grad_norm": 1.1673624515533447, + "learning_rate": 4.8649588762091016e-05, + "loss": 0.6728, + "num_input_tokens_seen": 6401024, + "step": 4205 + }, + { + "epoch": 0.13475449715127072, + "grad_norm": 1.1493229866027832, + "learning_rate": 4.8646403969016016e-05, + "loss": 0.5852, + "num_input_tokens_seen": 6408960, + "step": 4210 + }, + { + "epoch": 0.13491453812175916, + "grad_norm": 2.4547626972198486, + "learning_rate": 4.864321552936838e-05, + "loss": 0.7033, + "num_input_tokens_seen": 6416256, + "step": 4215 + }, + { + "epoch": 0.13507457909224763, + "grad_norm": 0.5172017812728882, + "learning_rate": 4.864002344363978e-05, + "loss": 0.546, + "num_input_tokens_seen": 6424096, + "step": 4220 + }, + { + "epoch": 0.13523462006273607, + "grad_norm": 0.8167120218276978, + "learning_rate": 4.863682771232248e-05, + "loss": 0.6052, + "num_input_tokens_seen": 6431808, + "step": 4225 + }, + { + "epoch": 0.1353946610332245, + "grad_norm": 1.5623259544372559, + "learning_rate": 4.8633628335909324e-05, + "loss": 0.5112, + "num_input_tokens_seen": 6439696, + "step": 4230 + }, + { + "epoch": 0.13555470200371295, + "grad_norm": 1.0275593996047974, + "learning_rate": 4.8630425314893676e-05, + "loss": 0.6431, + "num_input_tokens_seen": 6447728, + "step": 4235 + }, + { + "epoch": 0.13571474297420139, + "grad_norm": 1.057497501373291, + "learning_rate": 4.862721864976948e-05, + "loss": 0.5683, + "num_input_tokens_seen": 6455312, + "step": 4240 + }, + { + "epoch": 0.13587478394468985, + "grad_norm": 0.5051935315132141, + "learning_rate": 4.862400834103125e-05, + "loss": 0.6278, + "num_input_tokens_seen": 6463104, + "step": 4245 + }, + { + "epoch": 0.1360348249151783, + "grad_norm": 1.3239825963974, + "learning_rate": 4.862079438917406e-05, + "loss": 0.756, + "num_input_tokens_seen": 6470768, + "step": 4250 + }, + { + "epoch": 0.13619486588566673, + "grad_norm": 0.8671619892120361, + "learning_rate": 4.8617576794693536e-05, + "loss": 0.5999, + "num_input_tokens_seen": 6478144, + "step": 4255 + }, + { + "epoch": 0.13635490685615517, + "grad_norm": 1.2272650003433228, + "learning_rate": 4.8614355558085875e-05, + "loss": 0.5351, + "num_input_tokens_seen": 6486608, + "step": 4260 + }, + { + "epoch": 0.1365149478266436, + "grad_norm": 0.6238626837730408, + "learning_rate": 4.861113067984783e-05, + "loss": 0.5495, + "num_input_tokens_seen": 6494080, + "step": 4265 + }, + { + "epoch": 0.13667498879713205, + "grad_norm": 1.1085374355316162, + "learning_rate": 4.860790216047671e-05, + "loss": 0.4642, + "num_input_tokens_seen": 6501152, + "step": 4270 + }, + { + "epoch": 0.13683502976762052, + "grad_norm": 0.8489047288894653, + "learning_rate": 4.860467000047041e-05, + "loss": 0.5679, + "num_input_tokens_seen": 6508352, + "step": 4275 + }, + { + "epoch": 0.13699507073810896, + "grad_norm": 0.6011355519294739, + "learning_rate": 4.860143420032737e-05, + "loss": 0.458, + "num_input_tokens_seen": 6516048, + "step": 4280 + }, + { + "epoch": 0.1371551117085974, + "grad_norm": 0.7629432678222656, + "learning_rate": 4.859819476054657e-05, + "loss": 0.6337, + "num_input_tokens_seen": 6523488, + "step": 4285 + }, + { + "epoch": 0.13731515267908584, + "grad_norm": 1.2264838218688965, + "learning_rate": 4.859495168162758e-05, + "loss": 0.4829, + "num_input_tokens_seen": 6531264, + "step": 4290 + }, + { + "epoch": 0.13747519364957428, + "grad_norm": 1.1300344467163086, + "learning_rate": 4.859170496407054e-05, + "loss": 0.7206, + "num_input_tokens_seen": 6539088, + "step": 4295 + }, + { + "epoch": 0.13763523462006275, + "grad_norm": 1.067683219909668, + "learning_rate": 4.8588454608376114e-05, + "loss": 0.5285, + "num_input_tokens_seen": 6546368, + "step": 4300 + }, + { + "epoch": 0.1377952755905512, + "grad_norm": 0.9639940857887268, + "learning_rate": 4.8585200615045555e-05, + "loss": 0.5462, + "num_input_tokens_seen": 6554192, + "step": 4305 + }, + { + "epoch": 0.13795531656103963, + "grad_norm": 0.8050921559333801, + "learning_rate": 4.8581942984580674e-05, + "loss": 0.5761, + "num_input_tokens_seen": 6561952, + "step": 4310 + }, + { + "epoch": 0.13811535753152807, + "grad_norm": 1.0407971143722534, + "learning_rate": 4.857868171748384e-05, + "loss": 0.6014, + "num_input_tokens_seen": 6569648, + "step": 4315 + }, + { + "epoch": 0.1382753985020165, + "grad_norm": 0.6363945007324219, + "learning_rate": 4.8575416814257976e-05, + "loss": 0.6078, + "num_input_tokens_seen": 6577552, + "step": 4320 + }, + { + "epoch": 0.13843543947250497, + "grad_norm": 1.2990272045135498, + "learning_rate": 4.857214827540657e-05, + "loss": 0.6689, + "num_input_tokens_seen": 6585792, + "step": 4325 + }, + { + "epoch": 0.1385954804429934, + "grad_norm": 1.2083371877670288, + "learning_rate": 4.856887610143367e-05, + "loss": 0.7605, + "num_input_tokens_seen": 6593952, + "step": 4330 + }, + { + "epoch": 0.13875552141348185, + "grad_norm": 0.9039738178253174, + "learning_rate": 4.8565600292843896e-05, + "loss": 0.5491, + "num_input_tokens_seen": 6601216, + "step": 4335 + }, + { + "epoch": 0.1389155623839703, + "grad_norm": 0.9137353301048279, + "learning_rate": 4.856232085014241e-05, + "loss": 0.7146, + "num_input_tokens_seen": 6609168, + "step": 4340 + }, + { + "epoch": 0.13907560335445873, + "grad_norm": 1.6805943250656128, + "learning_rate": 4.855903777383495e-05, + "loss": 0.6647, + "num_input_tokens_seen": 6616576, + "step": 4345 + }, + { + "epoch": 0.1392356443249472, + "grad_norm": 1.0678280591964722, + "learning_rate": 4.85557510644278e-05, + "loss": 0.4938, + "num_input_tokens_seen": 6624048, + "step": 4350 + }, + { + "epoch": 0.13939568529543564, + "grad_norm": 1.1492457389831543, + "learning_rate": 4.855246072242782e-05, + "loss": 0.7828, + "num_input_tokens_seen": 6631824, + "step": 4355 + }, + { + "epoch": 0.13955572626592408, + "grad_norm": 0.8767714500427246, + "learning_rate": 4.8549166748342414e-05, + "loss": 0.5554, + "num_input_tokens_seen": 6639456, + "step": 4360 + }, + { + "epoch": 0.13971576723641252, + "grad_norm": 0.9297212958335876, + "learning_rate": 4.8545869142679556e-05, + "loss": 0.5935, + "num_input_tokens_seen": 6647248, + "step": 4365 + }, + { + "epoch": 0.13987580820690096, + "grad_norm": 0.7626619935035706, + "learning_rate": 4.8542567905947776e-05, + "loss": 0.5807, + "num_input_tokens_seen": 6656000, + "step": 4370 + }, + { + "epoch": 0.1400358491773894, + "grad_norm": 0.5468161106109619, + "learning_rate": 4.853926303865618e-05, + "loss": 0.6338, + "num_input_tokens_seen": 6663584, + "step": 4375 + }, + { + "epoch": 0.14019589014787787, + "grad_norm": 1.493271827697754, + "learning_rate": 4.853595454131441e-05, + "loss": 0.6894, + "num_input_tokens_seen": 6670848, + "step": 4380 + }, + { + "epoch": 0.1403559311183663, + "grad_norm": 1.022782802581787, + "learning_rate": 4.8532642414432674e-05, + "loss": 0.8086, + "num_input_tokens_seen": 6678416, + "step": 4385 + }, + { + "epoch": 0.14051597208885475, + "grad_norm": 0.6208885312080383, + "learning_rate": 4.8529326658521754e-05, + "loss": 0.6302, + "num_input_tokens_seen": 6685792, + "step": 4390 + }, + { + "epoch": 0.14067601305934319, + "grad_norm": 1.221459150314331, + "learning_rate": 4.8526007274092965e-05, + "loss": 0.5544, + "num_input_tokens_seen": 6693232, + "step": 4395 + }, + { + "epoch": 0.14083605402983163, + "grad_norm": 1.383186936378479, + "learning_rate": 4.852268426165822e-05, + "loss": 0.6783, + "num_input_tokens_seen": 6700592, + "step": 4400 + }, + { + "epoch": 0.14083605402983163, + "eval_loss": 0.6270559430122375, + "eval_runtime": 561.2703, + "eval_samples_per_second": 24.74, + "eval_steps_per_second": 12.37, + "num_input_tokens_seen": 6700592, + "step": 4400 + }, + { + "epoch": 0.1409960950003201, + "grad_norm": 0.8412292003631592, + "learning_rate": 4.851935762172995e-05, + "loss": 0.3992, + "num_input_tokens_seen": 6708416, + "step": 4405 + }, + { + "epoch": 0.14115613597080853, + "grad_norm": 1.2841655015945435, + "learning_rate": 4.8516027354821175e-05, + "loss": 0.5373, + "num_input_tokens_seen": 6716144, + "step": 4410 + }, + { + "epoch": 0.14131617694129697, + "grad_norm": 1.3710732460021973, + "learning_rate": 4.851269346144546e-05, + "loss": 0.6738, + "num_input_tokens_seen": 6724064, + "step": 4415 + }, + { + "epoch": 0.1414762179117854, + "grad_norm": 0.8963541984558105, + "learning_rate": 4.850935594211693e-05, + "loss": 0.6474, + "num_input_tokens_seen": 6732384, + "step": 4420 + }, + { + "epoch": 0.14163625888227385, + "grad_norm": 0.4725821614265442, + "learning_rate": 4.850601479735029e-05, + "loss": 0.5809, + "num_input_tokens_seen": 6740032, + "step": 4425 + }, + { + "epoch": 0.14179629985276232, + "grad_norm": 0.7195135951042175, + "learning_rate": 4.850267002766076e-05, + "loss": 0.4511, + "num_input_tokens_seen": 6747536, + "step": 4430 + }, + { + "epoch": 0.14195634082325076, + "grad_norm": 0.584216296672821, + "learning_rate": 4.849932163356417e-05, + "loss": 0.4101, + "num_input_tokens_seen": 6755024, + "step": 4435 + }, + { + "epoch": 0.1421163817937392, + "grad_norm": 1.43464994430542, + "learning_rate": 4.8495969615576864e-05, + "loss": 0.7414, + "num_input_tokens_seen": 6762304, + "step": 4440 + }, + { + "epoch": 0.14227642276422764, + "grad_norm": 1.2070200443267822, + "learning_rate": 4.849261397421577e-05, + "loss": 0.695, + "num_input_tokens_seen": 6769536, + "step": 4445 + }, + { + "epoch": 0.14243646373471608, + "grad_norm": 0.6594811081886292, + "learning_rate": 4.848925470999839e-05, + "loss": 0.6262, + "num_input_tokens_seen": 6777040, + "step": 4450 + }, + { + "epoch": 0.14259650470520452, + "grad_norm": 1.141030192375183, + "learning_rate": 4.848589182344273e-05, + "loss": 0.4829, + "num_input_tokens_seen": 6784544, + "step": 4455 + }, + { + "epoch": 0.142756545675693, + "grad_norm": 1.0249994993209839, + "learning_rate": 4.848252531506742e-05, + "loss": 0.6279, + "num_input_tokens_seen": 6792416, + "step": 4460 + }, + { + "epoch": 0.14291658664618143, + "grad_norm": 0.7724733948707581, + "learning_rate": 4.847915518539161e-05, + "loss": 0.5599, + "num_input_tokens_seen": 6800352, + "step": 4465 + }, + { + "epoch": 0.14307662761666987, + "grad_norm": 0.8655611872673035, + "learning_rate": 4.847578143493501e-05, + "loss": 0.5143, + "num_input_tokens_seen": 6807968, + "step": 4470 + }, + { + "epoch": 0.1432366685871583, + "grad_norm": 0.9364143013954163, + "learning_rate": 4.847240406421789e-05, + "loss": 0.5734, + "num_input_tokens_seen": 6815440, + "step": 4475 + }, + { + "epoch": 0.14339670955764675, + "grad_norm": 2.182478189468384, + "learning_rate": 4.84690230737611e-05, + "loss": 0.6756, + "num_input_tokens_seen": 6823008, + "step": 4480 + }, + { + "epoch": 0.1435567505281352, + "grad_norm": 0.3931407928466797, + "learning_rate": 4.846563846408602e-05, + "loss": 0.4575, + "num_input_tokens_seen": 6830160, + "step": 4485 + }, + { + "epoch": 0.14371679149862365, + "grad_norm": 0.530615508556366, + "learning_rate": 4.84622502357146e-05, + "loss": 0.6075, + "num_input_tokens_seen": 6838304, + "step": 4490 + }, + { + "epoch": 0.1438768324691121, + "grad_norm": 1.3562556505203247, + "learning_rate": 4.8458858389169345e-05, + "loss": 0.6074, + "num_input_tokens_seen": 6846048, + "step": 4495 + }, + { + "epoch": 0.14403687343960053, + "grad_norm": 0.5069437623023987, + "learning_rate": 4.8455462924973334e-05, + "loss": 0.5769, + "num_input_tokens_seen": 6854640, + "step": 4500 + }, + { + "epoch": 0.14419691441008897, + "grad_norm": 0.8891128301620483, + "learning_rate": 4.845206384365018e-05, + "loss": 0.5411, + "num_input_tokens_seen": 6862016, + "step": 4505 + }, + { + "epoch": 0.14435695538057744, + "grad_norm": 1.5650057792663574, + "learning_rate": 4.844866114572405e-05, + "loss": 0.7981, + "num_input_tokens_seen": 6869344, + "step": 4510 + }, + { + "epoch": 0.14451699635106588, + "grad_norm": 0.48146653175354004, + "learning_rate": 4.8445254831719706e-05, + "loss": 0.4461, + "num_input_tokens_seen": 6876800, + "step": 4515 + }, + { + "epoch": 0.14467703732155432, + "grad_norm": 0.9456567168235779, + "learning_rate": 4.8441844902162434e-05, + "loss": 0.5675, + "num_input_tokens_seen": 6884352, + "step": 4520 + }, + { + "epoch": 0.14483707829204276, + "grad_norm": 0.565869152545929, + "learning_rate": 4.843843135757809e-05, + "loss": 0.5551, + "num_input_tokens_seen": 6892320, + "step": 4525 + }, + { + "epoch": 0.1449971192625312, + "grad_norm": 0.5409795641899109, + "learning_rate": 4.843501419849308e-05, + "loss": 0.7772, + "num_input_tokens_seen": 6899776, + "step": 4530 + }, + { + "epoch": 0.14515716023301967, + "grad_norm": 1.2320817708969116, + "learning_rate": 4.8431593425434386e-05, + "loss": 0.5686, + "num_input_tokens_seen": 6907984, + "step": 4535 + }, + { + "epoch": 0.1453172012035081, + "grad_norm": 0.9911248683929443, + "learning_rate": 4.8428169038929526e-05, + "loss": 0.436, + "num_input_tokens_seen": 6915504, + "step": 4540 + }, + { + "epoch": 0.14547724217399655, + "grad_norm": 0.9177669882774353, + "learning_rate": 4.8424741039506575e-05, + "loss": 0.5866, + "num_input_tokens_seen": 6922848, + "step": 4545 + }, + { + "epoch": 0.14563728314448499, + "grad_norm": 0.7093709111213684, + "learning_rate": 4.842130942769419e-05, + "loss": 0.609, + "num_input_tokens_seen": 6930336, + "step": 4550 + }, + { + "epoch": 0.14579732411497343, + "grad_norm": 0.7458511590957642, + "learning_rate": 4.841787420402156e-05, + "loss": 0.6561, + "num_input_tokens_seen": 6937696, + "step": 4555 + }, + { + "epoch": 0.14595736508546187, + "grad_norm": 0.7952134013175964, + "learning_rate": 4.841443536901844e-05, + "loss": 0.5702, + "num_input_tokens_seen": 6945088, + "step": 4560 + }, + { + "epoch": 0.14611740605595033, + "grad_norm": 0.9709889888763428, + "learning_rate": 4.841099292321514e-05, + "loss": 0.4417, + "num_input_tokens_seen": 6952480, + "step": 4565 + }, + { + "epoch": 0.14627744702643877, + "grad_norm": 0.5156117677688599, + "learning_rate": 4.8407546867142525e-05, + "loss": 0.5944, + "num_input_tokens_seen": 6959808, + "step": 4570 + }, + { + "epoch": 0.1464374879969272, + "grad_norm": 0.5721389055252075, + "learning_rate": 4.840409720133203e-05, + "loss": 0.6393, + "num_input_tokens_seen": 6967648, + "step": 4575 + }, + { + "epoch": 0.14659752896741565, + "grad_norm": 0.7314876317977905, + "learning_rate": 4.8400643926315634e-05, + "loss": 0.5939, + "num_input_tokens_seen": 6975456, + "step": 4580 + }, + { + "epoch": 0.1467575699379041, + "grad_norm": 0.5930989980697632, + "learning_rate": 4.839718704262587e-05, + "loss": 0.532, + "num_input_tokens_seen": 6983184, + "step": 4585 + }, + { + "epoch": 0.14691761090839256, + "grad_norm": 1.1171725988388062, + "learning_rate": 4.839372655079585e-05, + "loss": 0.5218, + "num_input_tokens_seen": 6990816, + "step": 4590 + }, + { + "epoch": 0.147077651878881, + "grad_norm": 0.8167979121208191, + "learning_rate": 4.83902624513592e-05, + "loss": 0.6099, + "num_input_tokens_seen": 6998224, + "step": 4595 + }, + { + "epoch": 0.14723769284936944, + "grad_norm": 1.1761811971664429, + "learning_rate": 4.838679474485014e-05, + "loss": 0.5696, + "num_input_tokens_seen": 7006048, + "step": 4600 + }, + { + "epoch": 0.14723769284936944, + "eval_loss": 0.6254128813743591, + "eval_runtime": 559.3937, + "eval_samples_per_second": 24.823, + "eval_steps_per_second": 12.412, + "num_input_tokens_seen": 7006048, + "step": 4600 + }, + { + "epoch": 0.14739773381985788, + "grad_norm": 1.5796098709106445, + "learning_rate": 4.838332343180343e-05, + "loss": 0.6495, + "num_input_tokens_seen": 7013328, + "step": 4605 + }, + { + "epoch": 0.14755777479034632, + "grad_norm": 0.9940457940101624, + "learning_rate": 4.83798485127544e-05, + "loss": 0.5377, + "num_input_tokens_seen": 7021280, + "step": 4610 + }, + { + "epoch": 0.1477178157608348, + "grad_norm": 0.9453684687614441, + "learning_rate": 4.837636998823892e-05, + "loss": 0.7647, + "num_input_tokens_seen": 7028496, + "step": 4615 + }, + { + "epoch": 0.14787785673132323, + "grad_norm": 0.8582830429077148, + "learning_rate": 4.8372887858793414e-05, + "loss": 0.4789, + "num_input_tokens_seen": 7036224, + "step": 4620 + }, + { + "epoch": 0.14803789770181167, + "grad_norm": 0.6817514300346375, + "learning_rate": 4.836940212495489e-05, + "loss": 0.5175, + "num_input_tokens_seen": 7043952, + "step": 4625 + }, + { + "epoch": 0.1481979386723001, + "grad_norm": 0.7442020773887634, + "learning_rate": 4.836591278726087e-05, + "loss": 0.5662, + "num_input_tokens_seen": 7052000, + "step": 4630 + }, + { + "epoch": 0.14835797964278855, + "grad_norm": 0.5279739499092102, + "learning_rate": 4.836241984624947e-05, + "loss": 0.887, + "num_input_tokens_seen": 7059072, + "step": 4635 + }, + { + "epoch": 0.14851802061327699, + "grad_norm": 1.052008867263794, + "learning_rate": 4.8358923302459336e-05, + "loss": 0.5161, + "num_input_tokens_seen": 7066496, + "step": 4640 + }, + { + "epoch": 0.14867806158376545, + "grad_norm": 1.1052714586257935, + "learning_rate": 4.835542315642968e-05, + "loss": 0.6268, + "num_input_tokens_seen": 7074224, + "step": 4645 + }, + { + "epoch": 0.1488381025542539, + "grad_norm": 1.5589959621429443, + "learning_rate": 4.8351919408700274e-05, + "loss": 0.5645, + "num_input_tokens_seen": 7081664, + "step": 4650 + }, + { + "epoch": 0.14899814352474233, + "grad_norm": 0.9529060125350952, + "learning_rate": 4.834841205981144e-05, + "loss": 0.6391, + "num_input_tokens_seen": 7088736, + "step": 4655 + }, + { + "epoch": 0.14915818449523077, + "grad_norm": 0.7903347015380859, + "learning_rate": 4.8344901110304054e-05, + "loss": 0.5816, + "num_input_tokens_seen": 7096816, + "step": 4660 + }, + { + "epoch": 0.1493182254657192, + "grad_norm": 0.6745935082435608, + "learning_rate": 4.8341386560719534e-05, + "loss": 0.6972, + "num_input_tokens_seen": 7103904, + "step": 4665 + }, + { + "epoch": 0.14947826643620768, + "grad_norm": 1.071433186531067, + "learning_rate": 4.833786841159989e-05, + "loss": 0.8478, + "num_input_tokens_seen": 7111216, + "step": 4670 + }, + { + "epoch": 0.14963830740669612, + "grad_norm": 0.6074586510658264, + "learning_rate": 4.833434666348765e-05, + "loss": 0.5198, + "num_input_tokens_seen": 7119088, + "step": 4675 + }, + { + "epoch": 0.14979834837718456, + "grad_norm": 0.6597241163253784, + "learning_rate": 4.833082131692592e-05, + "loss": 0.5995, + "num_input_tokens_seen": 7126528, + "step": 4680 + }, + { + "epoch": 0.149958389347673, + "grad_norm": 1.1274795532226562, + "learning_rate": 4.832729237245835e-05, + "loss": 0.6977, + "num_input_tokens_seen": 7133776, + "step": 4685 + }, + { + "epoch": 0.15011843031816144, + "grad_norm": 1.014112114906311, + "learning_rate": 4.8323759830629145e-05, + "loss": 0.666, + "num_input_tokens_seen": 7141248, + "step": 4690 + }, + { + "epoch": 0.1502784712886499, + "grad_norm": 1.0374388694763184, + "learning_rate": 4.8320223691983066e-05, + "loss": 0.7498, + "num_input_tokens_seen": 7149168, + "step": 4695 + }, + { + "epoch": 0.15043851225913835, + "grad_norm": 0.9771149754524231, + "learning_rate": 4.831668395706544e-05, + "loss": 0.5464, + "num_input_tokens_seen": 7156096, + "step": 4700 + }, + { + "epoch": 0.15059855322962679, + "grad_norm": 1.5932278633117676, + "learning_rate": 4.8313140626422125e-05, + "loss": 0.539, + "num_input_tokens_seen": 7163392, + "step": 4705 + }, + { + "epoch": 0.15075859420011523, + "grad_norm": 0.8848294019699097, + "learning_rate": 4.830959370059956e-05, + "loss": 0.6339, + "num_input_tokens_seen": 7170864, + "step": 4710 + }, + { + "epoch": 0.15091863517060367, + "grad_norm": 1.654638648033142, + "learning_rate": 4.830604318014472e-05, + "loss": 0.6029, + "num_input_tokens_seen": 7178256, + "step": 4715 + }, + { + "epoch": 0.15107867614109213, + "grad_norm": 1.0658844709396362, + "learning_rate": 4.830248906560514e-05, + "loss": 0.5594, + "num_input_tokens_seen": 7186272, + "step": 4720 + }, + { + "epoch": 0.15123871711158057, + "grad_norm": 0.39618685841560364, + "learning_rate": 4.829893135752891e-05, + "loss": 0.592, + "num_input_tokens_seen": 7194416, + "step": 4725 + }, + { + "epoch": 0.151398758082069, + "grad_norm": 0.6195219159126282, + "learning_rate": 4.829537005646466e-05, + "loss": 0.5091, + "num_input_tokens_seen": 7202112, + "step": 4730 + }, + { + "epoch": 0.15155879905255745, + "grad_norm": 0.8221493363380432, + "learning_rate": 4.8291805162961615e-05, + "loss": 0.6783, + "num_input_tokens_seen": 7209840, + "step": 4735 + }, + { + "epoch": 0.1517188400230459, + "grad_norm": 1.0576956272125244, + "learning_rate": 4.82882366775695e-05, + "loss": 0.5361, + "num_input_tokens_seen": 7217120, + "step": 4740 + }, + { + "epoch": 0.15187888099353433, + "grad_norm": 0.7131669521331787, + "learning_rate": 4.828466460083864e-05, + "loss": 0.5079, + "num_input_tokens_seen": 7224560, + "step": 4745 + }, + { + "epoch": 0.1520389219640228, + "grad_norm": 0.5890803933143616, + "learning_rate": 4.8281088933319877e-05, + "loss": 0.7168, + "num_input_tokens_seen": 7231856, + "step": 4750 + }, + { + "epoch": 0.15219896293451124, + "grad_norm": 2.3038129806518555, + "learning_rate": 4.827750967556464e-05, + "loss": 0.7909, + "num_input_tokens_seen": 7239312, + "step": 4755 + }, + { + "epoch": 0.15235900390499968, + "grad_norm": 1.0196114778518677, + "learning_rate": 4.827392682812488e-05, + "loss": 0.7961, + "num_input_tokens_seen": 7246608, + "step": 4760 + }, + { + "epoch": 0.15251904487548812, + "grad_norm": 1.0511764287948608, + "learning_rate": 4.827034039155312e-05, + "loss": 0.6379, + "num_input_tokens_seen": 7254112, + "step": 4765 + }, + { + "epoch": 0.15267908584597656, + "grad_norm": 1.141928791999817, + "learning_rate": 4.8266750366402445e-05, + "loss": 0.5736, + "num_input_tokens_seen": 7261456, + "step": 4770 + }, + { + "epoch": 0.15283912681646503, + "grad_norm": 1.0800381898880005, + "learning_rate": 4.8263156753226476e-05, + "loss": 0.6649, + "num_input_tokens_seen": 7268800, + "step": 4775 + }, + { + "epoch": 0.15299916778695347, + "grad_norm": 1.269799828529358, + "learning_rate": 4.8259559552579394e-05, + "loss": 0.6082, + "num_input_tokens_seen": 7276400, + "step": 4780 + }, + { + "epoch": 0.1531592087574419, + "grad_norm": 0.849112868309021, + "learning_rate": 4.825595876501593e-05, + "loss": 0.4709, + "num_input_tokens_seen": 7284288, + "step": 4785 + }, + { + "epoch": 0.15331924972793035, + "grad_norm": 0.7207449078559875, + "learning_rate": 4.825235439109137e-05, + "loss": 0.5481, + "num_input_tokens_seen": 7292000, + "step": 4790 + }, + { + "epoch": 0.15347929069841879, + "grad_norm": 1.1235378980636597, + "learning_rate": 4.824874643136156e-05, + "loss": 0.5658, + "num_input_tokens_seen": 7300048, + "step": 4795 + }, + { + "epoch": 0.15363933166890725, + "grad_norm": 1.5008130073547363, + "learning_rate": 4.824513488638288e-05, + "loss": 0.5197, + "num_input_tokens_seen": 7307424, + "step": 4800 + }, + { + "epoch": 0.15363933166890725, + "eval_loss": 0.6217425465583801, + "eval_runtime": 558.6708, + "eval_samples_per_second": 24.855, + "eval_steps_per_second": 12.428, + "num_input_tokens_seen": 7307424, + "step": 4800 + }, + { + "epoch": 0.1537993726393957, + "grad_norm": 0.8533058762550354, + "learning_rate": 4.8241519756712293e-05, + "loss": 0.6886, + "num_input_tokens_seen": 7314960, + "step": 4805 + }, + { + "epoch": 0.15395941360988413, + "grad_norm": 0.8054823279380798, + "learning_rate": 4.8237901042907285e-05, + "loss": 0.5019, + "num_input_tokens_seen": 7323232, + "step": 4810 + }, + { + "epoch": 0.15411945458037257, + "grad_norm": 0.5511018633842468, + "learning_rate": 4.823427874552591e-05, + "loss": 0.5989, + "num_input_tokens_seen": 7330800, + "step": 4815 + }, + { + "epoch": 0.154279495550861, + "grad_norm": 0.7198414206504822, + "learning_rate": 4.823065286512677e-05, + "loss": 0.6026, + "num_input_tokens_seen": 7338272, + "step": 4820 + }, + { + "epoch": 0.15443953652134945, + "grad_norm": 0.9098744988441467, + "learning_rate": 4.8227023402269025e-05, + "loss": 0.587, + "num_input_tokens_seen": 7345952, + "step": 4825 + }, + { + "epoch": 0.15459957749183792, + "grad_norm": 0.5816830396652222, + "learning_rate": 4.822339035751239e-05, + "loss": 0.4847, + "num_input_tokens_seen": 7353488, + "step": 4830 + }, + { + "epoch": 0.15475961846232636, + "grad_norm": 1.081634283065796, + "learning_rate": 4.8219753731417104e-05, + "loss": 0.8075, + "num_input_tokens_seen": 7361072, + "step": 4835 + }, + { + "epoch": 0.1549196594328148, + "grad_norm": 1.2762314081192017, + "learning_rate": 4.821611352454401e-05, + "loss": 0.6676, + "num_input_tokens_seen": 7368640, + "step": 4840 + }, + { + "epoch": 0.15507970040330324, + "grad_norm": 0.9490680694580078, + "learning_rate": 4.8212469737454444e-05, + "loss": 0.8572, + "num_input_tokens_seen": 7376256, + "step": 4845 + }, + { + "epoch": 0.15523974137379168, + "grad_norm": 0.5494232177734375, + "learning_rate": 4.820882237071035e-05, + "loss": 0.6179, + "num_input_tokens_seen": 7383536, + "step": 4850 + }, + { + "epoch": 0.15539978234428015, + "grad_norm": 1.3120840787887573, + "learning_rate": 4.820517142487417e-05, + "loss": 0.5677, + "num_input_tokens_seen": 7391520, + "step": 4855 + }, + { + "epoch": 0.15555982331476859, + "grad_norm": 0.8417726755142212, + "learning_rate": 4.8201516900508956e-05, + "loss": 0.5413, + "num_input_tokens_seen": 7398784, + "step": 4860 + }, + { + "epoch": 0.15571986428525703, + "grad_norm": 0.5399201512336731, + "learning_rate": 4.819785879817827e-05, + "loss": 0.5488, + "num_input_tokens_seen": 7406304, + "step": 4865 + }, + { + "epoch": 0.15587990525574547, + "grad_norm": 0.3606429994106293, + "learning_rate": 4.8194197118446226e-05, + "loss": 0.4032, + "num_input_tokens_seen": 7413952, + "step": 4870 + }, + { + "epoch": 0.1560399462262339, + "grad_norm": 0.9390645027160645, + "learning_rate": 4.819053186187752e-05, + "loss": 0.6481, + "num_input_tokens_seen": 7421456, + "step": 4875 + }, + { + "epoch": 0.15619998719672237, + "grad_norm": 1.3513855934143066, + "learning_rate": 4.818686302903736e-05, + "loss": 0.6164, + "num_input_tokens_seen": 7428752, + "step": 4880 + }, + { + "epoch": 0.1563600281672108, + "grad_norm": 1.1530189514160156, + "learning_rate": 4.818319062049154e-05, + "loss": 0.5242, + "num_input_tokens_seen": 7436656, + "step": 4885 + }, + { + "epoch": 0.15652006913769925, + "grad_norm": 0.7275683283805847, + "learning_rate": 4.817951463680639e-05, + "loss": 0.6562, + "num_input_tokens_seen": 7445232, + "step": 4890 + }, + { + "epoch": 0.1566801101081877, + "grad_norm": 1.377921223640442, + "learning_rate": 4.817583507854879e-05, + "loss": 0.5776, + "num_input_tokens_seen": 7452800, + "step": 4895 + }, + { + "epoch": 0.15684015107867613, + "grad_norm": 0.43485525250434875, + "learning_rate": 4.817215194628617e-05, + "loss": 0.5769, + "num_input_tokens_seen": 7460800, + "step": 4900 + }, + { + "epoch": 0.1570001920491646, + "grad_norm": 1.4191079139709473, + "learning_rate": 4.816846524058653e-05, + "loss": 0.8023, + "num_input_tokens_seen": 7468480, + "step": 4905 + }, + { + "epoch": 0.15716023301965304, + "grad_norm": 2.3911917209625244, + "learning_rate": 4.816477496201839e-05, + "loss": 0.6705, + "num_input_tokens_seen": 7476064, + "step": 4910 + }, + { + "epoch": 0.15732027399014148, + "grad_norm": 0.7021450400352478, + "learning_rate": 4.8161081111150845e-05, + "loss": 0.817, + "num_input_tokens_seen": 7483408, + "step": 4915 + }, + { + "epoch": 0.15748031496062992, + "grad_norm": 1.622652292251587, + "learning_rate": 4.815738368855354e-05, + "loss": 0.6862, + "num_input_tokens_seen": 7490784, + "step": 4920 + }, + { + "epoch": 0.15764035593111836, + "grad_norm": 0.3302921652793884, + "learning_rate": 4.815368269479664e-05, + "loss": 0.3954, + "num_input_tokens_seen": 7498416, + "step": 4925 + }, + { + "epoch": 0.1578003969016068, + "grad_norm": 1.3277217149734497, + "learning_rate": 4.814997813045092e-05, + "loss": 0.5573, + "num_input_tokens_seen": 7506480, + "step": 4930 + }, + { + "epoch": 0.15796043787209527, + "grad_norm": 0.5098202228546143, + "learning_rate": 4.814626999608764e-05, + "loss": 0.5981, + "num_input_tokens_seen": 7514288, + "step": 4935 + }, + { + "epoch": 0.1581204788425837, + "grad_norm": 1.1685487031936646, + "learning_rate": 4.814255829227865e-05, + "loss": 0.7587, + "num_input_tokens_seen": 7521520, + "step": 4940 + }, + { + "epoch": 0.15828051981307215, + "grad_norm": 0.7818238735198975, + "learning_rate": 4.813884301959635e-05, + "loss": 0.6129, + "num_input_tokens_seen": 7529472, + "step": 4945 + }, + { + "epoch": 0.15844056078356059, + "grad_norm": 0.9823248982429504, + "learning_rate": 4.813512417861368e-05, + "loss": 0.7234, + "num_input_tokens_seen": 7537312, + "step": 4950 + }, + { + "epoch": 0.15860060175404903, + "grad_norm": 1.4347654581069946, + "learning_rate": 4.813140176990411e-05, + "loss": 0.5526, + "num_input_tokens_seen": 7544720, + "step": 4955 + }, + { + "epoch": 0.1587606427245375, + "grad_norm": 2.4383692741394043, + "learning_rate": 4.8127675794041714e-05, + "loss": 0.6869, + "num_input_tokens_seen": 7552064, + "step": 4960 + }, + { + "epoch": 0.15892068369502593, + "grad_norm": 0.5758613348007202, + "learning_rate": 4.812394625160107e-05, + "loss": 0.7706, + "num_input_tokens_seen": 7560128, + "step": 4965 + }, + { + "epoch": 0.15908072466551437, + "grad_norm": 0.6747694611549377, + "learning_rate": 4.812021314315732e-05, + "loss": 0.7, + "num_input_tokens_seen": 7567440, + "step": 4970 + }, + { + "epoch": 0.1592407656360028, + "grad_norm": 1.231058955192566, + "learning_rate": 4.811647646928616e-05, + "loss": 0.6133, + "num_input_tokens_seen": 7575552, + "step": 4975 + }, + { + "epoch": 0.15940080660649125, + "grad_norm": 0.5769082903862, + "learning_rate": 4.8112736230563814e-05, + "loss": 0.6361, + "num_input_tokens_seen": 7583424, + "step": 4980 + }, + { + "epoch": 0.15956084757697972, + "grad_norm": 1.0580310821533203, + "learning_rate": 4.81089924275671e-05, + "loss": 0.8838, + "num_input_tokens_seen": 7591024, + "step": 4985 + }, + { + "epoch": 0.15972088854746816, + "grad_norm": 0.7649575471878052, + "learning_rate": 4.810524506087335e-05, + "loss": 0.5658, + "num_input_tokens_seen": 7598720, + "step": 4990 + }, + { + "epoch": 0.1598809295179566, + "grad_norm": 0.5255262851715088, + "learning_rate": 4.810149413106044e-05, + "loss": 0.4512, + "num_input_tokens_seen": 7606592, + "step": 4995 + }, + { + "epoch": 0.16004097048844504, + "grad_norm": 0.8957213163375854, + "learning_rate": 4.809773963870684e-05, + "loss": 0.6421, + "num_input_tokens_seen": 7613952, + "step": 5000 + }, + { + "epoch": 0.16004097048844504, + "eval_loss": 0.6179686784744263, + "eval_runtime": 557.7605, + "eval_samples_per_second": 24.896, + "eval_steps_per_second": 12.448, + "num_input_tokens_seen": 7613952, + "step": 5000 + }, + { + "epoch": 0.16020101145893348, + "grad_norm": 1.2529364824295044, + "learning_rate": 4.809398158439151e-05, + "loss": 0.6124, + "num_input_tokens_seen": 7621936, + "step": 5005 + }, + { + "epoch": 0.16036105242942192, + "grad_norm": 0.9663079380989075, + "learning_rate": 4.8090219968694005e-05, + "loss": 0.6565, + "num_input_tokens_seen": 7629808, + "step": 5010 + }, + { + "epoch": 0.1605210933999104, + "grad_norm": 0.6907167434692383, + "learning_rate": 4.808645479219442e-05, + "loss": 0.5474, + "num_input_tokens_seen": 7637168, + "step": 5015 + }, + { + "epoch": 0.16068113437039883, + "grad_norm": 0.545076310634613, + "learning_rate": 4.8082686055473375e-05, + "loss": 0.5986, + "num_input_tokens_seen": 7645664, + "step": 5020 + }, + { + "epoch": 0.16084117534088727, + "grad_norm": 1.7195569276809692, + "learning_rate": 4.8078913759112066e-05, + "loss": 0.5689, + "num_input_tokens_seen": 7653104, + "step": 5025 + }, + { + "epoch": 0.1610012163113757, + "grad_norm": 1.2938737869262695, + "learning_rate": 4.807513790369223e-05, + "loss": 0.6136, + "num_input_tokens_seen": 7660912, + "step": 5030 + }, + { + "epoch": 0.16116125728186415, + "grad_norm": 1.0403988361358643, + "learning_rate": 4.8071358489796145e-05, + "loss": 0.6066, + "num_input_tokens_seen": 7668608, + "step": 5035 + }, + { + "epoch": 0.1613212982523526, + "grad_norm": 1.1329249143600464, + "learning_rate": 4.806757551800665e-05, + "loss": 0.8098, + "num_input_tokens_seen": 7676256, + "step": 5040 + }, + { + "epoch": 0.16148133922284105, + "grad_norm": 1.0564786195755005, + "learning_rate": 4.806378898890713e-05, + "loss": 0.7625, + "num_input_tokens_seen": 7683424, + "step": 5045 + }, + { + "epoch": 0.1616413801933295, + "grad_norm": 2.883589744567871, + "learning_rate": 4.80599989030815e-05, + "loss": 0.9942, + "num_input_tokens_seen": 7691120, + "step": 5050 + }, + { + "epoch": 0.16180142116381793, + "grad_norm": 0.5661562085151672, + "learning_rate": 4.805620526111426e-05, + "loss": 0.5564, + "num_input_tokens_seen": 7698704, + "step": 5055 + }, + { + "epoch": 0.16196146213430637, + "grad_norm": 0.4893054962158203, + "learning_rate": 4.805240806359042e-05, + "loss": 0.5013, + "num_input_tokens_seen": 7706512, + "step": 5060 + }, + { + "epoch": 0.16212150310479484, + "grad_norm": 0.7611250877380371, + "learning_rate": 4.804860731109557e-05, + "loss": 0.4052, + "num_input_tokens_seen": 7714000, + "step": 5065 + }, + { + "epoch": 0.16228154407528328, + "grad_norm": 0.9929341077804565, + "learning_rate": 4.804480300421581e-05, + "loss": 0.5244, + "num_input_tokens_seen": 7721744, + "step": 5070 + }, + { + "epoch": 0.16244158504577172, + "grad_norm": 2.1925418376922607, + "learning_rate": 4.804099514353784e-05, + "loss": 0.7745, + "num_input_tokens_seen": 7729536, + "step": 5075 + }, + { + "epoch": 0.16260162601626016, + "grad_norm": 0.7927454113960266, + "learning_rate": 4.8037183729648867e-05, + "loss": 0.5356, + "num_input_tokens_seen": 7737168, + "step": 5080 + }, + { + "epoch": 0.1627616669867486, + "grad_norm": 0.5405261516571045, + "learning_rate": 4.803336876313666e-05, + "loss": 0.489, + "num_input_tokens_seen": 7744864, + "step": 5085 + }, + { + "epoch": 0.16292170795723707, + "grad_norm": 0.775251030921936, + "learning_rate": 4.802955024458953e-05, + "loss": 0.5221, + "num_input_tokens_seen": 7752576, + "step": 5090 + }, + { + "epoch": 0.1630817489277255, + "grad_norm": 0.8078415989875793, + "learning_rate": 4.802572817459634e-05, + "loss": 0.5664, + "num_input_tokens_seen": 7759808, + "step": 5095 + }, + { + "epoch": 0.16324178989821395, + "grad_norm": 1.2434852123260498, + "learning_rate": 4.802190255374651e-05, + "loss": 0.6156, + "num_input_tokens_seen": 7767888, + "step": 5100 + }, + { + "epoch": 0.16340183086870239, + "grad_norm": 0.7584583759307861, + "learning_rate": 4.801807338263e-05, + "loss": 0.4346, + "num_input_tokens_seen": 7775168, + "step": 5105 + }, + { + "epoch": 0.16356187183919083, + "grad_norm": 0.9604533314704895, + "learning_rate": 4.8014240661837306e-05, + "loss": 0.6027, + "num_input_tokens_seen": 7783056, + "step": 5110 + }, + { + "epoch": 0.16372191280967927, + "grad_norm": 0.9093446135520935, + "learning_rate": 4.80104043919595e-05, + "loss": 0.4314, + "num_input_tokens_seen": 7790656, + "step": 5115 + }, + { + "epoch": 0.16388195378016773, + "grad_norm": 1.2013072967529297, + "learning_rate": 4.800656457358815e-05, + "loss": 0.6743, + "num_input_tokens_seen": 7798224, + "step": 5120 + }, + { + "epoch": 0.16404199475065617, + "grad_norm": 1.0272585153579712, + "learning_rate": 4.800272120731544e-05, + "loss": 0.6481, + "num_input_tokens_seen": 7805344, + "step": 5125 + }, + { + "epoch": 0.1642020357211446, + "grad_norm": 0.7708238959312439, + "learning_rate": 4.799887429373404e-05, + "loss": 0.5218, + "num_input_tokens_seen": 7813456, + "step": 5130 + }, + { + "epoch": 0.16436207669163305, + "grad_norm": 0.5229723453521729, + "learning_rate": 4.79950238334372e-05, + "loss": 0.4789, + "num_input_tokens_seen": 7821600, + "step": 5135 + }, + { + "epoch": 0.1645221176621215, + "grad_norm": 1.8400288820266724, + "learning_rate": 4.799116982701872e-05, + "loss": 0.8052, + "num_input_tokens_seen": 7829104, + "step": 5140 + }, + { + "epoch": 0.16468215863260996, + "grad_norm": 0.6893177032470703, + "learning_rate": 4.7987312275072926e-05, + "loss": 0.7648, + "num_input_tokens_seen": 7836528, + "step": 5145 + }, + { + "epoch": 0.1648421996030984, + "grad_norm": 0.47316184639930725, + "learning_rate": 4.79834511781947e-05, + "loss": 0.4698, + "num_input_tokens_seen": 7844528, + "step": 5150 + }, + { + "epoch": 0.16500224057358684, + "grad_norm": 0.42362716794013977, + "learning_rate": 4.797958653697947e-05, + "loss": 0.4771, + "num_input_tokens_seen": 7852432, + "step": 5155 + }, + { + "epoch": 0.16516228154407528, + "grad_norm": 0.3352607190608978, + "learning_rate": 4.7975718352023225e-05, + "loss": 0.518, + "num_input_tokens_seen": 7860016, + "step": 5160 + }, + { + "epoch": 0.16532232251456372, + "grad_norm": 1.597568392753601, + "learning_rate": 4.7971846623922476e-05, + "loss": 0.6948, + "num_input_tokens_seen": 7868000, + "step": 5165 + }, + { + "epoch": 0.1654823634850522, + "grad_norm": 0.3884619176387787, + "learning_rate": 4.7967971353274294e-05, + "loss": 0.5857, + "num_input_tokens_seen": 7875600, + "step": 5170 + }, + { + "epoch": 0.16564240445554063, + "grad_norm": 0.5140993595123291, + "learning_rate": 4.79640925406763e-05, + "loss": 0.4981, + "num_input_tokens_seen": 7882880, + "step": 5175 + }, + { + "epoch": 0.16580244542602907, + "grad_norm": 1.3076881170272827, + "learning_rate": 4.796021018672664e-05, + "loss": 0.5467, + "num_input_tokens_seen": 7890064, + "step": 5180 + }, + { + "epoch": 0.1659624863965175, + "grad_norm": 1.0799822807312012, + "learning_rate": 4.795632429202405e-05, + "loss": 0.6695, + "num_input_tokens_seen": 7897776, + "step": 5185 + }, + { + "epoch": 0.16612252736700595, + "grad_norm": 0.9639331102371216, + "learning_rate": 4.795243485716775e-05, + "loss": 0.3944, + "num_input_tokens_seen": 7905200, + "step": 5190 + }, + { + "epoch": 0.16628256833749439, + "grad_norm": 0.9454852342605591, + "learning_rate": 4.794854188275757e-05, + "loss": 0.4637, + "num_input_tokens_seen": 7913056, + "step": 5195 + }, + { + "epoch": 0.16644260930798285, + "grad_norm": 0.7996925711631775, + "learning_rate": 4.794464536939384e-05, + "loss": 0.5504, + "num_input_tokens_seen": 7920784, + "step": 5200 + }, + { + "epoch": 0.16644260930798285, + "eval_loss": 0.6143309473991394, + "eval_runtime": 561.5226, + "eval_samples_per_second": 24.729, + "eval_steps_per_second": 12.365, + "num_input_tokens_seen": 7920784, + "step": 5200 + }, + { + "epoch": 0.1666026502784713, + "grad_norm": 0.5535822510719299, + "learning_rate": 4.794074531767745e-05, + "loss": 0.6392, + "num_input_tokens_seen": 7928192, + "step": 5205 + }, + { + "epoch": 0.16676269124895973, + "grad_norm": 0.6790224313735962, + "learning_rate": 4.7936841728209834e-05, + "loss": 0.6553, + "num_input_tokens_seen": 7935920, + "step": 5210 + }, + { + "epoch": 0.16692273221944817, + "grad_norm": 0.8560389876365662, + "learning_rate": 4.7932934601593e-05, + "loss": 0.6613, + "num_input_tokens_seen": 7943312, + "step": 5215 + }, + { + "epoch": 0.1670827731899366, + "grad_norm": 0.74202960729599, + "learning_rate": 4.792902393842943e-05, + "loss": 0.4152, + "num_input_tokens_seen": 7951024, + "step": 5220 + }, + { + "epoch": 0.16724281416042508, + "grad_norm": 1.181990623474121, + "learning_rate": 4.792510973932225e-05, + "loss": 0.6556, + "num_input_tokens_seen": 7958576, + "step": 5225 + }, + { + "epoch": 0.16740285513091352, + "grad_norm": 1.0074570178985596, + "learning_rate": 4.7921192004875036e-05, + "loss": 0.6056, + "num_input_tokens_seen": 7966704, + "step": 5230 + }, + { + "epoch": 0.16756289610140196, + "grad_norm": 0.7350451946258545, + "learning_rate": 4.791727073569198e-05, + "loss": 0.5735, + "num_input_tokens_seen": 7973840, + "step": 5235 + }, + { + "epoch": 0.1677229370718904, + "grad_norm": 0.9782227873802185, + "learning_rate": 4.7913345932377775e-05, + "loss": 0.5135, + "num_input_tokens_seen": 7981824, + "step": 5240 + }, + { + "epoch": 0.16788297804237884, + "grad_norm": 0.6113510131835938, + "learning_rate": 4.790941759553769e-05, + "loss": 0.479, + "num_input_tokens_seen": 7989200, + "step": 5245 + }, + { + "epoch": 0.1680430190128673, + "grad_norm": 0.7468065619468689, + "learning_rate": 4.79054857257775e-05, + "loss": 0.5921, + "num_input_tokens_seen": 7996384, + "step": 5250 + }, + { + "epoch": 0.16820305998335575, + "grad_norm": 0.8270837664604187, + "learning_rate": 4.790155032370357e-05, + "loss": 0.7032, + "num_input_tokens_seen": 8003536, + "step": 5255 + }, + { + "epoch": 0.16836310095384419, + "grad_norm": 1.0575464963912964, + "learning_rate": 4.789761138992278e-05, + "loss": 0.6158, + "num_input_tokens_seen": 8011216, + "step": 5260 + }, + { + "epoch": 0.16852314192433263, + "grad_norm": 0.9285179972648621, + "learning_rate": 4.7893668925042565e-05, + "loss": 0.6247, + "num_input_tokens_seen": 8018656, + "step": 5265 + }, + { + "epoch": 0.16868318289482107, + "grad_norm": 0.6677285432815552, + "learning_rate": 4.78897229296709e-05, + "loss": 0.9327, + "num_input_tokens_seen": 8026128, + "step": 5270 + }, + { + "epoch": 0.16884322386530953, + "grad_norm": 1.0974512100219727, + "learning_rate": 4.7885773404416315e-05, + "loss": 0.5636, + "num_input_tokens_seen": 8034352, + "step": 5275 + }, + { + "epoch": 0.16900326483579797, + "grad_norm": 1.7108746767044067, + "learning_rate": 4.788182034988786e-05, + "loss": 0.6709, + "num_input_tokens_seen": 8042128, + "step": 5280 + }, + { + "epoch": 0.1691633058062864, + "grad_norm": 0.9558054804801941, + "learning_rate": 4.787786376669516e-05, + "loss": 0.5421, + "num_input_tokens_seen": 8049520, + "step": 5285 + }, + { + "epoch": 0.16932334677677485, + "grad_norm": 0.4644889235496521, + "learning_rate": 4.787390365544837e-05, + "loss": 0.3966, + "num_input_tokens_seen": 8057216, + "step": 5290 + }, + { + "epoch": 0.1694833877472633, + "grad_norm": 1.038769245147705, + "learning_rate": 4.786994001675818e-05, + "loss": 0.5823, + "num_input_tokens_seen": 8064576, + "step": 5295 + }, + { + "epoch": 0.16964342871775173, + "grad_norm": 1.131738543510437, + "learning_rate": 4.786597285123584e-05, + "loss": 0.5976, + "num_input_tokens_seen": 8072320, + "step": 5300 + }, + { + "epoch": 0.1698034696882402, + "grad_norm": 1.0056308507919312, + "learning_rate": 4.7862002159493135e-05, + "loss": 0.7196, + "num_input_tokens_seen": 8079584, + "step": 5305 + }, + { + "epoch": 0.16996351065872864, + "grad_norm": 0.5996102094650269, + "learning_rate": 4.785802794214239e-05, + "loss": 0.5833, + "num_input_tokens_seen": 8087392, + "step": 5310 + }, + { + "epoch": 0.17012355162921708, + "grad_norm": 0.9655048847198486, + "learning_rate": 4.7854050199796495e-05, + "loss": 0.4649, + "num_input_tokens_seen": 8095184, + "step": 5315 + }, + { + "epoch": 0.17028359259970552, + "grad_norm": 0.8809915781021118, + "learning_rate": 4.7850068933068845e-05, + "loss": 0.4968, + "num_input_tokens_seen": 8102544, + "step": 5320 + }, + { + "epoch": 0.17044363357019396, + "grad_norm": 0.791230320930481, + "learning_rate": 4.7846084142573425e-05, + "loss": 0.7961, + "num_input_tokens_seen": 8109872, + "step": 5325 + }, + { + "epoch": 0.17060367454068243, + "grad_norm": 1.117175817489624, + "learning_rate": 4.7842095828924725e-05, + "loss": 0.5648, + "num_input_tokens_seen": 8117376, + "step": 5330 + }, + { + "epoch": 0.17076371551117087, + "grad_norm": 0.7215973138809204, + "learning_rate": 4.783810399273779e-05, + "loss": 0.6981, + "num_input_tokens_seen": 8125184, + "step": 5335 + }, + { + "epoch": 0.1709237564816593, + "grad_norm": 1.4907022714614868, + "learning_rate": 4.7834108634628226e-05, + "loss": 0.6651, + "num_input_tokens_seen": 8132560, + "step": 5340 + }, + { + "epoch": 0.17108379745214775, + "grad_norm": 1.4294975996017456, + "learning_rate": 4.783010975521216e-05, + "loss": 0.705, + "num_input_tokens_seen": 8140032, + "step": 5345 + }, + { + "epoch": 0.17124383842263619, + "grad_norm": 1.0775128602981567, + "learning_rate": 4.782610735510626e-05, + "loss": 0.7288, + "num_input_tokens_seen": 8147680, + "step": 5350 + }, + { + "epoch": 0.17140387939312465, + "grad_norm": 0.7669017314910889, + "learning_rate": 4.782210143492776e-05, + "loss": 0.5899, + "num_input_tokens_seen": 8155408, + "step": 5355 + }, + { + "epoch": 0.1715639203636131, + "grad_norm": 1.071877121925354, + "learning_rate": 4.781809199529442e-05, + "loss": 0.7677, + "num_input_tokens_seen": 8162624, + "step": 5360 + }, + { + "epoch": 0.17172396133410153, + "grad_norm": 0.8805832266807556, + "learning_rate": 4.781407903682454e-05, + "loss": 0.5333, + "num_input_tokens_seen": 8170288, + "step": 5365 + }, + { + "epoch": 0.17188400230458997, + "grad_norm": 0.8880531191825867, + "learning_rate": 4.781006256013698e-05, + "loss": 0.5933, + "num_input_tokens_seen": 8177952, + "step": 5370 + }, + { + "epoch": 0.1720440432750784, + "grad_norm": 1.3639700412750244, + "learning_rate": 4.7806042565851115e-05, + "loss": 0.6098, + "num_input_tokens_seen": 8185264, + "step": 5375 + }, + { + "epoch": 0.17220408424556685, + "grad_norm": 0.6166390776634216, + "learning_rate": 4.7802019054586895e-05, + "loss": 0.6181, + "num_input_tokens_seen": 8192784, + "step": 5380 + }, + { + "epoch": 0.17236412521605532, + "grad_norm": 0.5984594821929932, + "learning_rate": 4.779799202696479e-05, + "loss": 0.4746, + "num_input_tokens_seen": 8200304, + "step": 5385 + }, + { + "epoch": 0.17252416618654376, + "grad_norm": 0.6210222840309143, + "learning_rate": 4.779396148360581e-05, + "loss": 0.5852, + "num_input_tokens_seen": 8208384, + "step": 5390 + }, + { + "epoch": 0.1726842071570322, + "grad_norm": 0.7095164656639099, + "learning_rate": 4.7789927425131517e-05, + "loss": 0.5408, + "num_input_tokens_seen": 8216256, + "step": 5395 + }, + { + "epoch": 0.17284424812752064, + "grad_norm": 2.036269187927246, + "learning_rate": 4.778588985216403e-05, + "loss": 0.6105, + "num_input_tokens_seen": 8223472, + "step": 5400 + }, + { + "epoch": 0.17284424812752064, + "eval_loss": 0.6103848218917847, + "eval_runtime": 561.5154, + "eval_samples_per_second": 24.73, + "eval_steps_per_second": 12.365, + "num_input_tokens_seen": 8223472, + "step": 5400 + }, + { + "epoch": 0.17300428909800908, + "grad_norm": 0.8363615274429321, + "learning_rate": 4.778184876532598e-05, + "loss": 0.5368, + "num_input_tokens_seen": 8230848, + "step": 5405 + }, + { + "epoch": 0.17316433006849755, + "grad_norm": 0.8621445894241333, + "learning_rate": 4.7777804165240556e-05, + "loss": 0.5441, + "num_input_tokens_seen": 8238080, + "step": 5410 + }, + { + "epoch": 0.17332437103898599, + "grad_norm": 0.5930768847465515, + "learning_rate": 4.7773756052531485e-05, + "loss": 0.6151, + "num_input_tokens_seen": 8246208, + "step": 5415 + }, + { + "epoch": 0.17348441200947443, + "grad_norm": 0.6462142467498779, + "learning_rate": 4.7769704427823035e-05, + "loss": 0.4667, + "num_input_tokens_seen": 8253408, + "step": 5420 + }, + { + "epoch": 0.17364445297996287, + "grad_norm": 0.46295493841171265, + "learning_rate": 4.776564929174003e-05, + "loss": 0.5095, + "num_input_tokens_seen": 8260880, + "step": 5425 + }, + { + "epoch": 0.1738044939504513, + "grad_norm": 0.8023867011070251, + "learning_rate": 4.7761590644907806e-05, + "loss": 0.5976, + "num_input_tokens_seen": 8268384, + "step": 5430 + }, + { + "epoch": 0.17396453492093977, + "grad_norm": 1.1030511856079102, + "learning_rate": 4.7757528487952263e-05, + "loss": 0.5949, + "num_input_tokens_seen": 8276240, + "step": 5435 + }, + { + "epoch": 0.1741245758914282, + "grad_norm": 0.9605992436408997, + "learning_rate": 4.7753462821499836e-05, + "loss": 0.6799, + "num_input_tokens_seen": 8283712, + "step": 5440 + }, + { + "epoch": 0.17428461686191665, + "grad_norm": 0.44389626383781433, + "learning_rate": 4.774939364617751e-05, + "loss": 0.7063, + "num_input_tokens_seen": 8291104, + "step": 5445 + }, + { + "epoch": 0.1744446578324051, + "grad_norm": 0.7829863429069519, + "learning_rate": 4.7745320962612795e-05, + "loss": 0.8202, + "num_input_tokens_seen": 8298784, + "step": 5450 + }, + { + "epoch": 0.17460469880289353, + "grad_norm": 0.650851309299469, + "learning_rate": 4.7741244771433756e-05, + "loss": 0.5389, + "num_input_tokens_seen": 8306336, + "step": 5455 + }, + { + "epoch": 0.174764739773382, + "grad_norm": 0.43437692523002625, + "learning_rate": 4.7737165073268985e-05, + "loss": 0.6032, + "num_input_tokens_seen": 8314080, + "step": 5460 + }, + { + "epoch": 0.17492478074387044, + "grad_norm": 0.6433952450752258, + "learning_rate": 4.7733081868747626e-05, + "loss": 0.3945, + "num_input_tokens_seen": 8321600, + "step": 5465 + }, + { + "epoch": 0.17508482171435888, + "grad_norm": 0.6473081707954407, + "learning_rate": 4.772899515849936e-05, + "loss": 0.6313, + "num_input_tokens_seen": 8329344, + "step": 5470 + }, + { + "epoch": 0.17524486268484732, + "grad_norm": 1.0878773927688599, + "learning_rate": 4.7724904943154414e-05, + "loss": 0.7225, + "num_input_tokens_seen": 8336912, + "step": 5475 + }, + { + "epoch": 0.17540490365533576, + "grad_norm": 1.1305088996887207, + "learning_rate": 4.772081122334354e-05, + "loss": 0.7323, + "num_input_tokens_seen": 8344688, + "step": 5480 + }, + { + "epoch": 0.1755649446258242, + "grad_norm": 1.0516875982284546, + "learning_rate": 4.771671399969806e-05, + "loss": 0.6733, + "num_input_tokens_seen": 8352400, + "step": 5485 + }, + { + "epoch": 0.17572498559631267, + "grad_norm": 0.6219242811203003, + "learning_rate": 4.7712613272849794e-05, + "loss": 0.6044, + "num_input_tokens_seen": 8360480, + "step": 5490 + }, + { + "epoch": 0.1758850265668011, + "grad_norm": 0.47516462206840515, + "learning_rate": 4.770850904343114e-05, + "loss": 0.4918, + "num_input_tokens_seen": 8368416, + "step": 5495 + }, + { + "epoch": 0.17604506753728955, + "grad_norm": 0.5815261006355286, + "learning_rate": 4.770440131207502e-05, + "loss": 0.6174, + "num_input_tokens_seen": 8376128, + "step": 5500 + }, + { + "epoch": 0.17620510850777799, + "grad_norm": 0.9317788481712341, + "learning_rate": 4.7700290079414896e-05, + "loss": 0.6754, + "num_input_tokens_seen": 8383696, + "step": 5505 + }, + { + "epoch": 0.17636514947826643, + "grad_norm": 1.041383147239685, + "learning_rate": 4.769617534608477e-05, + "loss": 0.4016, + "num_input_tokens_seen": 8391360, + "step": 5510 + }, + { + "epoch": 0.1765251904487549, + "grad_norm": 0.7336726188659668, + "learning_rate": 4.7692057112719193e-05, + "loss": 0.6254, + "num_input_tokens_seen": 8399216, + "step": 5515 + }, + { + "epoch": 0.17668523141924333, + "grad_norm": 0.9840289950370789, + "learning_rate": 4.7687935379953234e-05, + "loss": 0.5893, + "num_input_tokens_seen": 8406592, + "step": 5520 + }, + { + "epoch": 0.17684527238973177, + "grad_norm": 0.8670046925544739, + "learning_rate": 4.7683810148422534e-05, + "loss": 0.7096, + "num_input_tokens_seen": 8414384, + "step": 5525 + }, + { + "epoch": 0.1770053133602202, + "grad_norm": 1.299365520477295, + "learning_rate": 4.767968141876324e-05, + "loss": 0.6673, + "num_input_tokens_seen": 8422224, + "step": 5530 + }, + { + "epoch": 0.17716535433070865, + "grad_norm": 0.5431571006774902, + "learning_rate": 4.767554919161207e-05, + "loss": 0.5804, + "num_input_tokens_seen": 8429584, + "step": 5535 + }, + { + "epoch": 0.17732539530119712, + "grad_norm": 0.9183301329612732, + "learning_rate": 4.767141346760624e-05, + "loss": 0.4567, + "num_input_tokens_seen": 8436944, + "step": 5540 + }, + { + "epoch": 0.17748543627168556, + "grad_norm": 0.7556785345077515, + "learning_rate": 4.766727424738356e-05, + "loss": 0.4585, + "num_input_tokens_seen": 8444752, + "step": 5545 + }, + { + "epoch": 0.177645477242174, + "grad_norm": 0.7161464095115662, + "learning_rate": 4.7663131531582325e-05, + "loss": 0.4955, + "num_input_tokens_seen": 8451984, + "step": 5550 + }, + { + "epoch": 0.17780551821266244, + "grad_norm": 1.1192426681518555, + "learning_rate": 4.765898532084142e-05, + "loss": 0.5985, + "num_input_tokens_seen": 8459360, + "step": 5555 + }, + { + "epoch": 0.17796555918315088, + "grad_norm": 0.8458656072616577, + "learning_rate": 4.765483561580022e-05, + "loss": 0.652, + "num_input_tokens_seen": 8467024, + "step": 5560 + }, + { + "epoch": 0.17812560015363932, + "grad_norm": 0.8968318104743958, + "learning_rate": 4.7650682417098666e-05, + "loss": 0.7171, + "num_input_tokens_seen": 8474976, + "step": 5565 + }, + { + "epoch": 0.17828564112412779, + "grad_norm": 0.597501814365387, + "learning_rate": 4.7646525725377244e-05, + "loss": 0.5252, + "num_input_tokens_seen": 8482016, + "step": 5570 + }, + { + "epoch": 0.17844568209461623, + "grad_norm": 0.7350776195526123, + "learning_rate": 4.764236554127696e-05, + "loss": 0.664, + "num_input_tokens_seen": 8489904, + "step": 5575 + }, + { + "epoch": 0.17860572306510467, + "grad_norm": 0.9129493236541748, + "learning_rate": 4.7638201865439356e-05, + "loss": 0.5094, + "num_input_tokens_seen": 8497216, + "step": 5580 + }, + { + "epoch": 0.1787657640355931, + "grad_norm": 1.3745465278625488, + "learning_rate": 4.7634034698506545e-05, + "loss": 0.5666, + "num_input_tokens_seen": 8504976, + "step": 5585 + }, + { + "epoch": 0.17892580500608155, + "grad_norm": 0.6541405916213989, + "learning_rate": 4.762986404112115e-05, + "loss": 0.6027, + "num_input_tokens_seen": 8512384, + "step": 5590 + }, + { + "epoch": 0.17908584597657, + "grad_norm": 0.7667807340621948, + "learning_rate": 4.762568989392633e-05, + "loss": 0.5117, + "num_input_tokens_seen": 8520032, + "step": 5595 + }, + { + "epoch": 0.17924588694705845, + "grad_norm": 0.7867023944854736, + "learning_rate": 4.76215122575658e-05, + "loss": 0.5743, + "num_input_tokens_seen": 8527408, + "step": 5600 + }, + { + "epoch": 0.17924588694705845, + "eval_loss": 0.6073886752128601, + "eval_runtime": 560.2877, + "eval_samples_per_second": 24.784, + "eval_steps_per_second": 12.392, + "num_input_tokens_seen": 8527408, + "step": 5600 + }, + { + "epoch": 0.1794059279175469, + "grad_norm": 0.8464555740356445, + "learning_rate": 4.7617331132683795e-05, + "loss": 0.8647, + "num_input_tokens_seen": 8534672, + "step": 5605 + }, + { + "epoch": 0.17956596888803533, + "grad_norm": 1.148964285850525, + "learning_rate": 4.7613146519925105e-05, + "loss": 0.6428, + "num_input_tokens_seen": 8542224, + "step": 5610 + }, + { + "epoch": 0.17972600985852377, + "grad_norm": 0.7826103568077087, + "learning_rate": 4.7608958419935045e-05, + "loss": 0.4785, + "num_input_tokens_seen": 8550112, + "step": 5615 + }, + { + "epoch": 0.17988605082901224, + "grad_norm": 1.9742746353149414, + "learning_rate": 4.760476683335948e-05, + "loss": 0.7148, + "num_input_tokens_seen": 8557488, + "step": 5620 + }, + { + "epoch": 0.18004609179950068, + "grad_norm": 0.6249325275421143, + "learning_rate": 4.760057176084479e-05, + "loss": 0.6906, + "num_input_tokens_seen": 8564912, + "step": 5625 + }, + { + "epoch": 0.18020613276998912, + "grad_norm": 0.7349624633789062, + "learning_rate": 4.759637320303793e-05, + "loss": 0.465, + "num_input_tokens_seen": 8572192, + "step": 5630 + }, + { + "epoch": 0.18036617374047756, + "grad_norm": 1.4884079694747925, + "learning_rate": 4.759217116058635e-05, + "loss": 0.54, + "num_input_tokens_seen": 8579600, + "step": 5635 + }, + { + "epoch": 0.180526214710966, + "grad_norm": 0.7094123959541321, + "learning_rate": 4.758796563413807e-05, + "loss": 0.6311, + "num_input_tokens_seen": 8587568, + "step": 5640 + }, + { + "epoch": 0.18068625568145444, + "grad_norm": 0.8674600720405579, + "learning_rate": 4.758375662434163e-05, + "loss": 0.5759, + "num_input_tokens_seen": 8595024, + "step": 5645 + }, + { + "epoch": 0.1808462966519429, + "grad_norm": 0.762087881565094, + "learning_rate": 4.7579544131846114e-05, + "loss": 0.6691, + "num_input_tokens_seen": 8602928, + "step": 5650 + }, + { + "epoch": 0.18100633762243135, + "grad_norm": 0.6998353004455566, + "learning_rate": 4.757532815730114e-05, + "loss": 0.3987, + "num_input_tokens_seen": 8610544, + "step": 5655 + }, + { + "epoch": 0.18116637859291979, + "grad_norm": 0.9850977063179016, + "learning_rate": 4.7571108701356865e-05, + "loss": 0.4372, + "num_input_tokens_seen": 8617808, + "step": 5660 + }, + { + "epoch": 0.18132641956340823, + "grad_norm": 1.1198621988296509, + "learning_rate": 4.756688576466398e-05, + "loss": 0.713, + "num_input_tokens_seen": 8625664, + "step": 5665 + }, + { + "epoch": 0.18148646053389667, + "grad_norm": 0.9034255146980286, + "learning_rate": 4.756265934787372e-05, + "loss": 0.4694, + "num_input_tokens_seen": 8633648, + "step": 5670 + }, + { + "epoch": 0.18164650150438513, + "grad_norm": 0.5718774795532227, + "learning_rate": 4.755842945163785e-05, + "loss": 0.6221, + "num_input_tokens_seen": 8641952, + "step": 5675 + }, + { + "epoch": 0.18180654247487357, + "grad_norm": 0.7156134247779846, + "learning_rate": 4.755419607660867e-05, + "loss": 0.6572, + "num_input_tokens_seen": 8649840, + "step": 5680 + }, + { + "epoch": 0.181966583445362, + "grad_norm": 1.115200400352478, + "learning_rate": 4.7549959223439016e-05, + "loss": 0.5693, + "num_input_tokens_seen": 8657168, + "step": 5685 + }, + { + "epoch": 0.18212662441585045, + "grad_norm": 1.0532550811767578, + "learning_rate": 4.754571889278228e-05, + "loss": 0.7198, + "num_input_tokens_seen": 8664832, + "step": 5690 + }, + { + "epoch": 0.1822866653863389, + "grad_norm": 0.6227928400039673, + "learning_rate": 4.754147508529235e-05, + "loss": 0.5342, + "num_input_tokens_seen": 8672064, + "step": 5695 + }, + { + "epoch": 0.18244670635682736, + "grad_norm": 0.6002780795097351, + "learning_rate": 4.75372278016237e-05, + "loss": 0.6151, + "num_input_tokens_seen": 8679936, + "step": 5700 + }, + { + "epoch": 0.1826067473273158, + "grad_norm": 0.6857936978340149, + "learning_rate": 4.753297704243129e-05, + "loss": 0.435, + "num_input_tokens_seen": 8687376, + "step": 5705 + }, + { + "epoch": 0.18276678829780424, + "grad_norm": 0.6759807467460632, + "learning_rate": 4.752872280837066e-05, + "loss": 0.5659, + "num_input_tokens_seen": 8695344, + "step": 5710 + }, + { + "epoch": 0.18292682926829268, + "grad_norm": 1.031193494796753, + "learning_rate": 4.752446510009786e-05, + "loss": 0.6667, + "num_input_tokens_seen": 8703392, + "step": 5715 + }, + { + "epoch": 0.18308687023878112, + "grad_norm": 1.0164926052093506, + "learning_rate": 4.7520203918269476e-05, + "loss": 0.6921, + "num_input_tokens_seen": 8711200, + "step": 5720 + }, + { + "epoch": 0.1832469112092696, + "grad_norm": 1.0394277572631836, + "learning_rate": 4.751593926354265e-05, + "loss": 0.5364, + "num_input_tokens_seen": 8718896, + "step": 5725 + }, + { + "epoch": 0.18340695217975803, + "grad_norm": 0.7230845093727112, + "learning_rate": 4.751167113657503e-05, + "loss": 0.6473, + "num_input_tokens_seen": 8726480, + "step": 5730 + }, + { + "epoch": 0.18356699315024647, + "grad_norm": 0.4391130208969116, + "learning_rate": 4.7507399538024834e-05, + "loss": 0.386, + "num_input_tokens_seen": 8734016, + "step": 5735 + }, + { + "epoch": 0.1837270341207349, + "grad_norm": 1.651195764541626, + "learning_rate": 4.750312446855077e-05, + "loss": 0.7056, + "num_input_tokens_seen": 8741872, + "step": 5740 + }, + { + "epoch": 0.18388707509122335, + "grad_norm": 1.3892611265182495, + "learning_rate": 4.749884592881212e-05, + "loss": 0.597, + "num_input_tokens_seen": 8749280, + "step": 5745 + }, + { + "epoch": 0.18404711606171179, + "grad_norm": 1.0489732027053833, + "learning_rate": 4.74945639194687e-05, + "loss": 0.6665, + "num_input_tokens_seen": 8756864, + "step": 5750 + }, + { + "epoch": 0.18420715703220025, + "grad_norm": 0.5563551783561707, + "learning_rate": 4.749027844118083e-05, + "loss": 0.3973, + "num_input_tokens_seen": 8764448, + "step": 5755 + }, + { + "epoch": 0.1843671980026887, + "grad_norm": 0.916999876499176, + "learning_rate": 4.7485989494609395e-05, + "loss": 0.5801, + "num_input_tokens_seen": 8772128, + "step": 5760 + }, + { + "epoch": 0.18452723897317713, + "grad_norm": 0.8170648217201233, + "learning_rate": 4.748169708041581e-05, + "loss": 0.6381, + "num_input_tokens_seen": 8779680, + "step": 5765 + }, + { + "epoch": 0.18468727994366557, + "grad_norm": 1.4762773513793945, + "learning_rate": 4.7477401199262004e-05, + "loss": 0.661, + "num_input_tokens_seen": 8787392, + "step": 5770 + }, + { + "epoch": 0.184847320914154, + "grad_norm": 1.04656183719635, + "learning_rate": 4.747310185181048e-05, + "loss": 0.5682, + "num_input_tokens_seen": 8795536, + "step": 5775 + }, + { + "epoch": 0.18500736188464248, + "grad_norm": 1.0406602621078491, + "learning_rate": 4.746879903872422e-05, + "loss": 0.5578, + "num_input_tokens_seen": 8803424, + "step": 5780 + }, + { + "epoch": 0.18516740285513092, + "grad_norm": 0.8977932929992676, + "learning_rate": 4.746449276066679e-05, + "loss": 0.5542, + "num_input_tokens_seen": 8811232, + "step": 5785 + }, + { + "epoch": 0.18532744382561936, + "grad_norm": 0.6623252630233765, + "learning_rate": 4.746018301830227e-05, + "loss": 0.4364, + "num_input_tokens_seen": 8818624, + "step": 5790 + }, + { + "epoch": 0.1854874847961078, + "grad_norm": 0.7269492149353027, + "learning_rate": 4.7455869812295275e-05, + "loss": 0.6295, + "num_input_tokens_seen": 8826672, + "step": 5795 + }, + { + "epoch": 0.18564752576659624, + "grad_norm": 0.7576043605804443, + "learning_rate": 4.7451553143310964e-05, + "loss": 0.5296, + "num_input_tokens_seen": 8834416, + "step": 5800 + }, + { + "epoch": 0.18564752576659624, + "eval_loss": 0.6054133772850037, + "eval_runtime": 557.7668, + "eval_samples_per_second": 24.896, + "eval_steps_per_second": 12.448, + "num_input_tokens_seen": 8834416, + "step": 5800 + }, + { + "epoch": 0.1858075667370847, + "grad_norm": 0.6478068828582764, + "learning_rate": 4.744723301201501e-05, + "loss": 0.4765, + "num_input_tokens_seen": 8841872, + "step": 5805 + }, + { + "epoch": 0.18596760770757315, + "grad_norm": 0.7143697142601013, + "learning_rate": 4.744290941907364e-05, + "loss": 0.5459, + "num_input_tokens_seen": 8850320, + "step": 5810 + }, + { + "epoch": 0.18612764867806159, + "grad_norm": 0.6518799662590027, + "learning_rate": 4.7438582365153594e-05, + "loss": 0.4314, + "num_input_tokens_seen": 8858160, + "step": 5815 + }, + { + "epoch": 0.18628768964855003, + "grad_norm": 1.5027365684509277, + "learning_rate": 4.743425185092217e-05, + "loss": 0.4767, + "num_input_tokens_seen": 8865632, + "step": 5820 + }, + { + "epoch": 0.18644773061903847, + "grad_norm": 0.42759010195732117, + "learning_rate": 4.742991787704719e-05, + "loss": 0.5715, + "num_input_tokens_seen": 8873376, + "step": 5825 + }, + { + "epoch": 0.1866077715895269, + "grad_norm": 0.5651483535766602, + "learning_rate": 4.7425580444196994e-05, + "loss": 0.4232, + "num_input_tokens_seen": 8880832, + "step": 5830 + }, + { + "epoch": 0.18676781256001537, + "grad_norm": 0.5552348494529724, + "learning_rate": 4.742123955304048e-05, + "loss": 0.4251, + "num_input_tokens_seen": 8888912, + "step": 5835 + }, + { + "epoch": 0.1869278535305038, + "grad_norm": 1.990808367729187, + "learning_rate": 4.741689520424706e-05, + "loss": 0.6792, + "num_input_tokens_seen": 8896224, + "step": 5840 + }, + { + "epoch": 0.18708789450099225, + "grad_norm": 1.0441073179244995, + "learning_rate": 4.741254739848669e-05, + "loss": 0.5976, + "num_input_tokens_seen": 8904080, + "step": 5845 + }, + { + "epoch": 0.1872479354714807, + "grad_norm": 0.9496373534202576, + "learning_rate": 4.740819613642987e-05, + "loss": 0.5164, + "num_input_tokens_seen": 8912128, + "step": 5850 + }, + { + "epoch": 0.18740797644196913, + "grad_norm": 0.7937702536582947, + "learning_rate": 4.74038414187476e-05, + "loss": 0.5345, + "num_input_tokens_seen": 8919648, + "step": 5855 + }, + { + "epoch": 0.1875680174124576, + "grad_norm": 1.3435649871826172, + "learning_rate": 4.739948324611144e-05, + "loss": 0.5538, + "num_input_tokens_seen": 8927456, + "step": 5860 + }, + { + "epoch": 0.18772805838294604, + "grad_norm": 0.9741174578666687, + "learning_rate": 4.7395121619193465e-05, + "loss": 0.5097, + "num_input_tokens_seen": 8934800, + "step": 5865 + }, + { + "epoch": 0.18788809935343448, + "grad_norm": 0.7925089001655579, + "learning_rate": 4.7390756538666313e-05, + "loss": 0.6235, + "num_input_tokens_seen": 8942208, + "step": 5870 + }, + { + "epoch": 0.18804814032392292, + "grad_norm": 0.7623099684715271, + "learning_rate": 4.738638800520311e-05, + "loss": 0.5403, + "num_input_tokens_seen": 8949680, + "step": 5875 + }, + { + "epoch": 0.18820818129441136, + "grad_norm": 0.5435269474983215, + "learning_rate": 4.738201601947757e-05, + "loss": 0.4624, + "num_input_tokens_seen": 8957216, + "step": 5880 + }, + { + "epoch": 0.18836822226489983, + "grad_norm": 0.6789757609367371, + "learning_rate": 4.7377640582163876e-05, + "loss": 0.4529, + "num_input_tokens_seen": 8964480, + "step": 5885 + }, + { + "epoch": 0.18852826323538827, + "grad_norm": 0.5819696187973022, + "learning_rate": 4.7373261693936786e-05, + "loss": 0.6189, + "num_input_tokens_seen": 8971696, + "step": 5890 + }, + { + "epoch": 0.1886883042058767, + "grad_norm": 1.2457680702209473, + "learning_rate": 4.7368879355471595e-05, + "loss": 0.6289, + "num_input_tokens_seen": 8979136, + "step": 5895 + }, + { + "epoch": 0.18884834517636515, + "grad_norm": 0.7879301309585571, + "learning_rate": 4.736449356744409e-05, + "loss": 0.4714, + "num_input_tokens_seen": 8986736, + "step": 5900 + }, + { + "epoch": 0.18900838614685359, + "grad_norm": 2.8044047355651855, + "learning_rate": 4.736010433053064e-05, + "loss": 0.7218, + "num_input_tokens_seen": 8994304, + "step": 5905 + }, + { + "epoch": 0.18916842711734205, + "grad_norm": 1.8353251218795776, + "learning_rate": 4.73557116454081e-05, + "loss": 0.6515, + "num_input_tokens_seen": 9002272, + "step": 5910 + }, + { + "epoch": 0.1893284680878305, + "grad_norm": 0.8367384672164917, + "learning_rate": 4.735131551275389e-05, + "loss": 0.5665, + "num_input_tokens_seen": 9009568, + "step": 5915 + }, + { + "epoch": 0.18948850905831893, + "grad_norm": 0.7404084205627441, + "learning_rate": 4.734691593324594e-05, + "loss": 0.6982, + "num_input_tokens_seen": 9016576, + "step": 5920 + }, + { + "epoch": 0.18964855002880737, + "grad_norm": 1.4331055879592896, + "learning_rate": 4.734251290756272e-05, + "loss": 0.7837, + "num_input_tokens_seen": 9024144, + "step": 5925 + }, + { + "epoch": 0.1898085909992958, + "grad_norm": 0.6531298756599426, + "learning_rate": 4.7338106436383246e-05, + "loss": 0.6121, + "num_input_tokens_seen": 9031520, + "step": 5930 + }, + { + "epoch": 0.18996863196978425, + "grad_norm": 1.3634753227233887, + "learning_rate": 4.733369652038703e-05, + "loss": 0.687, + "num_input_tokens_seen": 9039152, + "step": 5935 + }, + { + "epoch": 0.19012867294027272, + "grad_norm": 1.2352230548858643, + "learning_rate": 4.7329283160254156e-05, + "loss": 0.53, + "num_input_tokens_seen": 9047056, + "step": 5940 + }, + { + "epoch": 0.19028871391076116, + "grad_norm": 0.9940985441207886, + "learning_rate": 4.732486635666521e-05, + "loss": 0.5752, + "num_input_tokens_seen": 9055168, + "step": 5945 + }, + { + "epoch": 0.1904487548812496, + "grad_norm": 0.8911786079406738, + "learning_rate": 4.732044611030132e-05, + "loss": 0.8138, + "num_input_tokens_seen": 9062496, + "step": 5950 + }, + { + "epoch": 0.19060879585173804, + "grad_norm": 0.8451092839241028, + "learning_rate": 4.731602242184414e-05, + "loss": 0.5999, + "num_input_tokens_seen": 9070288, + "step": 5955 + }, + { + "epoch": 0.19076883682222648, + "grad_norm": 0.47476914525032043, + "learning_rate": 4.7311595291975864e-05, + "loss": 0.4019, + "num_input_tokens_seen": 9077520, + "step": 5960 + }, + { + "epoch": 0.19092887779271495, + "grad_norm": 1.2385011911392212, + "learning_rate": 4.7307164721379216e-05, + "loss": 0.7542, + "num_input_tokens_seen": 9085136, + "step": 5965 + }, + { + "epoch": 0.19108891876320339, + "grad_norm": 0.8916173577308655, + "learning_rate": 4.730273071073743e-05, + "loss": 0.4236, + "num_input_tokens_seen": 9092544, + "step": 5970 + }, + { + "epoch": 0.19124895973369183, + "grad_norm": 0.5229238867759705, + "learning_rate": 4.729829326073429e-05, + "loss": 0.3748, + "num_input_tokens_seen": 9100400, + "step": 5975 + }, + { + "epoch": 0.19140900070418027, + "grad_norm": 1.2763566970825195, + "learning_rate": 4.7293852372054126e-05, + "loss": 0.6422, + "num_input_tokens_seen": 9108144, + "step": 5980 + }, + { + "epoch": 0.1915690416746687, + "grad_norm": 0.9347240328788757, + "learning_rate": 4.728940804538176e-05, + "loss": 0.6616, + "num_input_tokens_seen": 9115712, + "step": 5985 + }, + { + "epoch": 0.19172908264515717, + "grad_norm": 0.7781647443771362, + "learning_rate": 4.7284960281402556e-05, + "loss": 0.638, + "num_input_tokens_seen": 9122848, + "step": 5990 + }, + { + "epoch": 0.1918891236156456, + "grad_norm": 1.0724449157714844, + "learning_rate": 4.728050908080244e-05, + "loss": 0.5873, + "num_input_tokens_seen": 9130208, + "step": 5995 + }, + { + "epoch": 0.19204916458613405, + "grad_norm": 0.8275885581970215, + "learning_rate": 4.727605444426782e-05, + "loss": 0.4067, + "num_input_tokens_seen": 9138176, + "step": 6000 + }, + { + "epoch": 0.19204916458613405, + "eval_loss": 0.603553295135498, + "eval_runtime": 558.4052, + "eval_samples_per_second": 24.867, + "eval_steps_per_second": 12.434, + "num_input_tokens_seen": 9138176, + "step": 6000 + }, + { + "epoch": 0.1922092055566225, + "grad_norm": 1.1243342161178589, + "learning_rate": 4.727159637248567e-05, + "loss": 0.412, + "num_input_tokens_seen": 9145536, + "step": 6005 + }, + { + "epoch": 0.19236924652711093, + "grad_norm": 0.6974411606788635, + "learning_rate": 4.7267134866143474e-05, + "loss": 0.6502, + "num_input_tokens_seen": 9153408, + "step": 6010 + }, + { + "epoch": 0.19252928749759937, + "grad_norm": 0.9157295823097229, + "learning_rate": 4.726266992592926e-05, + "loss": 0.5849, + "num_input_tokens_seen": 9161072, + "step": 6015 + }, + { + "epoch": 0.19268932846808784, + "grad_norm": 0.5512407422065735, + "learning_rate": 4.725820155253157e-05, + "loss": 0.5359, + "num_input_tokens_seen": 9168656, + "step": 6020 + }, + { + "epoch": 0.19284936943857628, + "grad_norm": 0.8851261138916016, + "learning_rate": 4.725372974663948e-05, + "loss": 0.4817, + "num_input_tokens_seen": 9176560, + "step": 6025 + }, + { + "epoch": 0.19300941040906472, + "grad_norm": 0.7167941927909851, + "learning_rate": 4.724925450894262e-05, + "loss": 0.4975, + "num_input_tokens_seen": 9184240, + "step": 6030 + }, + { + "epoch": 0.19316945137955316, + "grad_norm": 0.5800012350082397, + "learning_rate": 4.72447758401311e-05, + "loss": 0.4913, + "num_input_tokens_seen": 9191424, + "step": 6035 + }, + { + "epoch": 0.1933294923500416, + "grad_norm": 1.509982705116272, + "learning_rate": 4.7240293740895616e-05, + "loss": 0.7778, + "num_input_tokens_seen": 9199344, + "step": 6040 + }, + { + "epoch": 0.19348953332053007, + "grad_norm": 0.4647118151187897, + "learning_rate": 4.723580821192733e-05, + "loss": 0.555, + "num_input_tokens_seen": 9206864, + "step": 6045 + }, + { + "epoch": 0.1936495742910185, + "grad_norm": 1.7939932346343994, + "learning_rate": 4.7231319253917996e-05, + "loss": 0.6071, + "num_input_tokens_seen": 9214480, + "step": 6050 + }, + { + "epoch": 0.19380961526150695, + "grad_norm": 0.6668114066123962, + "learning_rate": 4.722682686755986e-05, + "loss": 0.439, + "num_input_tokens_seen": 9222096, + "step": 6055 + }, + { + "epoch": 0.19396965623199539, + "grad_norm": 0.7463449835777283, + "learning_rate": 4.722233105354569e-05, + "loss": 0.6303, + "num_input_tokens_seen": 9229840, + "step": 6060 + }, + { + "epoch": 0.19412969720248383, + "grad_norm": 0.6398652791976929, + "learning_rate": 4.7217831812568815e-05, + "loss": 0.5365, + "num_input_tokens_seen": 9237968, + "step": 6065 + }, + { + "epoch": 0.1942897381729723, + "grad_norm": 0.874140739440918, + "learning_rate": 4.721332914532307e-05, + "loss": 0.7586, + "num_input_tokens_seen": 9245680, + "step": 6070 + }, + { + "epoch": 0.19444977914346073, + "grad_norm": 1.442766547203064, + "learning_rate": 4.720882305250281e-05, + "loss": 0.6567, + "num_input_tokens_seen": 9252832, + "step": 6075 + }, + { + "epoch": 0.19460982011394917, + "grad_norm": 0.7545415759086609, + "learning_rate": 4.720431353480295e-05, + "loss": 0.4658, + "num_input_tokens_seen": 9260448, + "step": 6080 + }, + { + "epoch": 0.1947698610844376, + "grad_norm": 1.3247406482696533, + "learning_rate": 4.719980059291891e-05, + "loss": 0.7039, + "num_input_tokens_seen": 9267840, + "step": 6085 + }, + { + "epoch": 0.19492990205492605, + "grad_norm": 0.855976402759552, + "learning_rate": 4.7195284227546634e-05, + "loss": 0.4958, + "num_input_tokens_seen": 9275328, + "step": 6090 + }, + { + "epoch": 0.19508994302541452, + "grad_norm": 0.7427745461463928, + "learning_rate": 4.7190764439382604e-05, + "loss": 0.471, + "num_input_tokens_seen": 9282880, + "step": 6095 + }, + { + "epoch": 0.19524998399590296, + "grad_norm": 0.9876025319099426, + "learning_rate": 4.7186241229123826e-05, + "loss": 0.6577, + "num_input_tokens_seen": 9290720, + "step": 6100 + }, + { + "epoch": 0.1954100249663914, + "grad_norm": 0.4379621148109436, + "learning_rate": 4.718171459746785e-05, + "loss": 0.4789, + "num_input_tokens_seen": 9298944, + "step": 6105 + }, + { + "epoch": 0.19557006593687984, + "grad_norm": 0.6737650036811829, + "learning_rate": 4.717718454511273e-05, + "loss": 0.5551, + "num_input_tokens_seen": 9306608, + "step": 6110 + }, + { + "epoch": 0.19573010690736828, + "grad_norm": 0.9843308329582214, + "learning_rate": 4.7172651072757056e-05, + "loss": 0.5452, + "num_input_tokens_seen": 9314400, + "step": 6115 + }, + { + "epoch": 0.19589014787785672, + "grad_norm": 0.8916650414466858, + "learning_rate": 4.7168114181099945e-05, + "loss": 0.7392, + "num_input_tokens_seen": 9322000, + "step": 6120 + }, + { + "epoch": 0.19605018884834519, + "grad_norm": 0.7281366586685181, + "learning_rate": 4.716357387084105e-05, + "loss": 0.4247, + "num_input_tokens_seen": 9329680, + "step": 6125 + }, + { + "epoch": 0.19621022981883363, + "grad_norm": 0.8171167373657227, + "learning_rate": 4.715903014268054e-05, + "loss": 0.6012, + "num_input_tokens_seen": 9337696, + "step": 6130 + }, + { + "epoch": 0.19637027078932207, + "grad_norm": 2.939704179763794, + "learning_rate": 4.715448299731911e-05, + "loss": 0.6342, + "num_input_tokens_seen": 9344848, + "step": 6135 + }, + { + "epoch": 0.1965303117598105, + "grad_norm": 0.6098791360855103, + "learning_rate": 4.7149932435457986e-05, + "loss": 0.5343, + "num_input_tokens_seen": 9352608, + "step": 6140 + }, + { + "epoch": 0.19669035273029895, + "grad_norm": 0.8546373248100281, + "learning_rate": 4.714537845779894e-05, + "loss": 0.5266, + "num_input_tokens_seen": 9360784, + "step": 6145 + }, + { + "epoch": 0.1968503937007874, + "grad_norm": 0.6442511677742004, + "learning_rate": 4.714082106504423e-05, + "loss": 0.4133, + "num_input_tokens_seen": 9368656, + "step": 6150 + }, + { + "epoch": 0.19701043467127585, + "grad_norm": 0.9990066885948181, + "learning_rate": 4.713626025789667e-05, + "loss": 0.5031, + "num_input_tokens_seen": 9375984, + "step": 6155 + }, + { + "epoch": 0.1971704756417643, + "grad_norm": 0.6688899993896484, + "learning_rate": 4.7131696037059606e-05, + "loss": 0.5335, + "num_input_tokens_seen": 9383856, + "step": 6160 + }, + { + "epoch": 0.19733051661225273, + "grad_norm": 0.5302935838699341, + "learning_rate": 4.712712840323689e-05, + "loss": 0.5007, + "num_input_tokens_seen": 9391680, + "step": 6165 + }, + { + "epoch": 0.19749055758274117, + "grad_norm": 1.5097975730895996, + "learning_rate": 4.71225573571329e-05, + "loss": 0.5724, + "num_input_tokens_seen": 9399344, + "step": 6170 + }, + { + "epoch": 0.19765059855322964, + "grad_norm": 0.9301640391349792, + "learning_rate": 4.711798289945256e-05, + "loss": 0.4867, + "num_input_tokens_seen": 9407072, + "step": 6175 + }, + { + "epoch": 0.19781063952371808, + "grad_norm": 0.9491104483604431, + "learning_rate": 4.71134050309013e-05, + "loss": 0.4479, + "num_input_tokens_seen": 9414384, + "step": 6180 + }, + { + "epoch": 0.19797068049420652, + "grad_norm": 0.5411805510520935, + "learning_rate": 4.710882375218509e-05, + "loss": 0.5916, + "num_input_tokens_seen": 9421824, + "step": 6185 + }, + { + "epoch": 0.19813072146469496, + "grad_norm": 0.5017980933189392, + "learning_rate": 4.7104239064010424e-05, + "loss": 0.4203, + "num_input_tokens_seen": 9429424, + "step": 6190 + }, + { + "epoch": 0.1982907624351834, + "grad_norm": 1.5545010566711426, + "learning_rate": 4.709965096708432e-05, + "loss": 0.5052, + "num_input_tokens_seen": 9437552, + "step": 6195 + }, + { + "epoch": 0.19845080340567184, + "grad_norm": 1.031404972076416, + "learning_rate": 4.709505946211431e-05, + "loss": 0.7168, + "num_input_tokens_seen": 9444880, + "step": 6200 + }, + { + "epoch": 0.19845080340567184, + "eval_loss": 0.6011281609535217, + "eval_runtime": 563.4397, + "eval_samples_per_second": 24.645, + "eval_steps_per_second": 12.323, + "num_input_tokens_seen": 9444880, + "step": 6200 + }, + { + "epoch": 0.1986108443761603, + "grad_norm": 1.4848867654800415, + "learning_rate": 4.709046454980846e-05, + "loss": 0.6724, + "num_input_tokens_seen": 9452672, + "step": 6205 + }, + { + "epoch": 0.19877088534664875, + "grad_norm": 0.6500815749168396, + "learning_rate": 4.708586623087538e-05, + "loss": 0.6768, + "num_input_tokens_seen": 9459872, + "step": 6210 + }, + { + "epoch": 0.19893092631713719, + "grad_norm": 0.6337331533432007, + "learning_rate": 4.708126450602418e-05, + "loss": 0.5372, + "num_input_tokens_seen": 9467376, + "step": 6215 + }, + { + "epoch": 0.19909096728762563, + "grad_norm": 1.4158320426940918, + "learning_rate": 4.7076659375964495e-05, + "loss": 0.6469, + "num_input_tokens_seen": 9475296, + "step": 6220 + }, + { + "epoch": 0.19925100825811407, + "grad_norm": 0.6336644291877747, + "learning_rate": 4.707205084140651e-05, + "loss": 0.5285, + "num_input_tokens_seen": 9482768, + "step": 6225 + }, + { + "epoch": 0.19941104922860253, + "grad_norm": 0.6833570003509521, + "learning_rate": 4.7067438903060904e-05, + "loss": 0.5279, + "num_input_tokens_seen": 9490304, + "step": 6230 + }, + { + "epoch": 0.19957109019909097, + "grad_norm": 0.9926424026489258, + "learning_rate": 4.70628235616389e-05, + "loss": 0.7247, + "num_input_tokens_seen": 9497664, + "step": 6235 + }, + { + "epoch": 0.1997311311695794, + "grad_norm": 1.0350083112716675, + "learning_rate": 4.7058204817852256e-05, + "loss": 0.602, + "num_input_tokens_seen": 9505952, + "step": 6240 + }, + { + "epoch": 0.19989117214006785, + "grad_norm": 0.7331730723381042, + "learning_rate": 4.705358267241322e-05, + "loss": 0.7372, + "num_input_tokens_seen": 9513408, + "step": 6245 + }, + { + "epoch": 0.2000512131105563, + "grad_norm": 0.5373892188072205, + "learning_rate": 4.704895712603459e-05, + "loss": 0.5601, + "num_input_tokens_seen": 9520624, + "step": 6250 + }, + { + "epoch": 0.20021125408104476, + "grad_norm": 1.011770486831665, + "learning_rate": 4.704432817942969e-05, + "loss": 0.6074, + "num_input_tokens_seen": 9528480, + "step": 6255 + }, + { + "epoch": 0.2003712950515332, + "grad_norm": 1.9441444873809814, + "learning_rate": 4.703969583331236e-05, + "loss": 0.5635, + "num_input_tokens_seen": 9535840, + "step": 6260 + }, + { + "epoch": 0.20053133602202164, + "grad_norm": 0.6714699268341064, + "learning_rate": 4.7035060088396965e-05, + "loss": 0.5589, + "num_input_tokens_seen": 9543696, + "step": 6265 + }, + { + "epoch": 0.20069137699251008, + "grad_norm": 0.4773525893688202, + "learning_rate": 4.703042094539839e-05, + "loss": 0.3837, + "num_input_tokens_seen": 9550864, + "step": 6270 + }, + { + "epoch": 0.20085141796299852, + "grad_norm": 0.7360977530479431, + "learning_rate": 4.702577840503206e-05, + "loss": 0.445, + "num_input_tokens_seen": 9558144, + "step": 6275 + }, + { + "epoch": 0.201011458933487, + "grad_norm": 0.6567984223365784, + "learning_rate": 4.70211324680139e-05, + "loss": 0.4314, + "num_input_tokens_seen": 9565632, + "step": 6280 + }, + { + "epoch": 0.20117149990397543, + "grad_norm": 1.717024564743042, + "learning_rate": 4.7016483135060386e-05, + "loss": 0.4417, + "num_input_tokens_seen": 9573392, + "step": 6285 + }, + { + "epoch": 0.20133154087446387, + "grad_norm": 0.7733656764030457, + "learning_rate": 4.701183040688849e-05, + "loss": 0.6847, + "num_input_tokens_seen": 9580720, + "step": 6290 + }, + { + "epoch": 0.2014915818449523, + "grad_norm": 0.826812207698822, + "learning_rate": 4.700717428421573e-05, + "loss": 0.6116, + "num_input_tokens_seen": 9588288, + "step": 6295 + }, + { + "epoch": 0.20165162281544075, + "grad_norm": 1.05031418800354, + "learning_rate": 4.700251476776014e-05, + "loss": 0.4075, + "num_input_tokens_seen": 9595440, + "step": 6300 + }, + { + "epoch": 0.20181166378592919, + "grad_norm": 0.9531649947166443, + "learning_rate": 4.699785185824026e-05, + "loss": 0.4428, + "num_input_tokens_seen": 9603280, + "step": 6305 + }, + { + "epoch": 0.20197170475641765, + "grad_norm": 0.7923893332481384, + "learning_rate": 4.699318555637519e-05, + "loss": 0.7067, + "num_input_tokens_seen": 9610912, + "step": 6310 + }, + { + "epoch": 0.2021317457269061, + "grad_norm": 0.8564261198043823, + "learning_rate": 4.6988515862884525e-05, + "loss": 0.7099, + "num_input_tokens_seen": 9618528, + "step": 6315 + }, + { + "epoch": 0.20229178669739453, + "grad_norm": 0.8754181265830994, + "learning_rate": 4.698384277848838e-05, + "loss": 0.7511, + "num_input_tokens_seen": 9625984, + "step": 6320 + }, + { + "epoch": 0.20245182766788297, + "grad_norm": 0.7175393104553223, + "learning_rate": 4.6979166303907425e-05, + "loss": 0.5031, + "num_input_tokens_seen": 9633824, + "step": 6325 + }, + { + "epoch": 0.2026118686383714, + "grad_norm": 0.6281511783599854, + "learning_rate": 4.697448643986281e-05, + "loss": 0.5166, + "num_input_tokens_seen": 9641424, + "step": 6330 + }, + { + "epoch": 0.20277190960885988, + "grad_norm": 0.9350003004074097, + "learning_rate": 4.696980318707624e-05, + "loss": 0.7913, + "num_input_tokens_seen": 9649008, + "step": 6335 + }, + { + "epoch": 0.20293195057934832, + "grad_norm": 1.2866401672363281, + "learning_rate": 4.6965116546269924e-05, + "loss": 0.623, + "num_input_tokens_seen": 9656448, + "step": 6340 + }, + { + "epoch": 0.20309199154983676, + "grad_norm": 1.1215614080429077, + "learning_rate": 4.6960426518166615e-05, + "loss": 0.7788, + "num_input_tokens_seen": 9664064, + "step": 6345 + }, + { + "epoch": 0.2032520325203252, + "grad_norm": 0.9085702896118164, + "learning_rate": 4.6955733103489556e-05, + "loss": 0.5805, + "num_input_tokens_seen": 9671552, + "step": 6350 + }, + { + "epoch": 0.20341207349081364, + "grad_norm": 1.3073610067367554, + "learning_rate": 4.695103630296255e-05, + "loss": 0.8103, + "num_input_tokens_seen": 9679392, + "step": 6355 + }, + { + "epoch": 0.2035721144613021, + "grad_norm": 1.1363242864608765, + "learning_rate": 4.694633611730988e-05, + "loss": 0.5552, + "num_input_tokens_seen": 9686704, + "step": 6360 + }, + { + "epoch": 0.20373215543179055, + "grad_norm": 0.6821451187133789, + "learning_rate": 4.694163254725639e-05, + "loss": 0.599, + "num_input_tokens_seen": 9694096, + "step": 6365 + }, + { + "epoch": 0.20389219640227899, + "grad_norm": 0.6142565011978149, + "learning_rate": 4.693692559352743e-05, + "loss": 0.4362, + "num_input_tokens_seen": 9701376, + "step": 6370 + }, + { + "epoch": 0.20405223737276743, + "grad_norm": 0.7395855784416199, + "learning_rate": 4.693221525684886e-05, + "loss": 0.5239, + "num_input_tokens_seen": 9709200, + "step": 6375 + }, + { + "epoch": 0.20421227834325587, + "grad_norm": 0.5686452984809875, + "learning_rate": 4.6927501537947084e-05, + "loss": 0.4978, + "num_input_tokens_seen": 9716896, + "step": 6380 + }, + { + "epoch": 0.2043723193137443, + "grad_norm": 1.2746598720550537, + "learning_rate": 4.692278443754901e-05, + "loss": 0.7249, + "num_input_tokens_seen": 9724192, + "step": 6385 + }, + { + "epoch": 0.20453236028423277, + "grad_norm": 1.1285096406936646, + "learning_rate": 4.691806395638208e-05, + "loss": 0.7438, + "num_input_tokens_seen": 9731968, + "step": 6390 + }, + { + "epoch": 0.2046924012547212, + "grad_norm": 0.9483610987663269, + "learning_rate": 4.6913340095174255e-05, + "loss": 0.4918, + "num_input_tokens_seen": 9739328, + "step": 6395 + }, + { + "epoch": 0.20485244222520965, + "grad_norm": 0.9611797332763672, + "learning_rate": 4.690861285465399e-05, + "loss": 0.461, + "num_input_tokens_seen": 9746832, + "step": 6400 + }, + { + "epoch": 0.20485244222520965, + "eval_loss": 0.5975978374481201, + "eval_runtime": 560.2767, + "eval_samples_per_second": 24.784, + "eval_steps_per_second": 12.392, + "num_input_tokens_seen": 9746832, + "step": 6400 + }, + { + "epoch": 0.2050124831956981, + "grad_norm": 0.7784366011619568, + "learning_rate": 4.690388223555031e-05, + "loss": 0.5426, + "num_input_tokens_seen": 9753936, + "step": 6405 + }, + { + "epoch": 0.20517252416618653, + "grad_norm": 0.7368760108947754, + "learning_rate": 4.689914823859273e-05, + "loss": 0.6805, + "num_input_tokens_seen": 9761728, + "step": 6410 + }, + { + "epoch": 0.205332565136675, + "grad_norm": 1.3245866298675537, + "learning_rate": 4.689441086451129e-05, + "loss": 0.5966, + "num_input_tokens_seen": 9769584, + "step": 6415 + }, + { + "epoch": 0.20549260610716344, + "grad_norm": 1.032783031463623, + "learning_rate": 4.688967011403655e-05, + "loss": 0.3991, + "num_input_tokens_seen": 9777264, + "step": 6420 + }, + { + "epoch": 0.20565264707765188, + "grad_norm": 0.5898444056510925, + "learning_rate": 4.68849259878996e-05, + "loss": 0.4647, + "num_input_tokens_seen": 9784528, + "step": 6425 + }, + { + "epoch": 0.20581268804814032, + "grad_norm": 0.6088990569114685, + "learning_rate": 4.6880178486832036e-05, + "loss": 0.5271, + "num_input_tokens_seen": 9792032, + "step": 6430 + }, + { + "epoch": 0.20597272901862876, + "grad_norm": 0.77117520570755, + "learning_rate": 4.687542761156598e-05, + "loss": 0.5285, + "num_input_tokens_seen": 9799216, + "step": 6435 + }, + { + "epoch": 0.20613276998911723, + "grad_norm": 1.225519061088562, + "learning_rate": 4.6870673362834096e-05, + "loss": 0.6904, + "num_input_tokens_seen": 9807152, + "step": 6440 + }, + { + "epoch": 0.20629281095960567, + "grad_norm": 0.8555203080177307, + "learning_rate": 4.6865915741369526e-05, + "loss": 0.54, + "num_input_tokens_seen": 9814448, + "step": 6445 + }, + { + "epoch": 0.2064528519300941, + "grad_norm": 0.6018242239952087, + "learning_rate": 4.686115474790597e-05, + "loss": 0.5462, + "num_input_tokens_seen": 9822400, + "step": 6450 + }, + { + "epoch": 0.20661289290058255, + "grad_norm": 0.7401317358016968, + "learning_rate": 4.685639038317762e-05, + "loss": 0.7003, + "num_input_tokens_seen": 9830112, + "step": 6455 + }, + { + "epoch": 0.20677293387107099, + "grad_norm": 0.9829227328300476, + "learning_rate": 4.685162264791921e-05, + "loss": 0.5874, + "num_input_tokens_seen": 9837520, + "step": 6460 + }, + { + "epoch": 0.20693297484155945, + "grad_norm": 0.6660376787185669, + "learning_rate": 4.684685154286599e-05, + "loss": 0.5145, + "num_input_tokens_seen": 9845344, + "step": 6465 + }, + { + "epoch": 0.2070930158120479, + "grad_norm": 0.8337805271148682, + "learning_rate": 4.684207706875371e-05, + "loss": 0.6316, + "num_input_tokens_seen": 9852656, + "step": 6470 + }, + { + "epoch": 0.20725305678253633, + "grad_norm": 1.1122114658355713, + "learning_rate": 4.683729922631866e-05, + "loss": 0.6598, + "num_input_tokens_seen": 9859904, + "step": 6475 + }, + { + "epoch": 0.20741309775302477, + "grad_norm": 1.0401208400726318, + "learning_rate": 4.683251801629765e-05, + "loss": 0.5492, + "num_input_tokens_seen": 9867328, + "step": 6480 + }, + { + "epoch": 0.2075731387235132, + "grad_norm": 0.5794508457183838, + "learning_rate": 4.6827733439428e-05, + "loss": 0.567, + "num_input_tokens_seen": 9874864, + "step": 6485 + }, + { + "epoch": 0.20773317969400165, + "grad_norm": 0.609871506690979, + "learning_rate": 4.682294549644754e-05, + "loss": 0.7545, + "num_input_tokens_seen": 9882576, + "step": 6490 + }, + { + "epoch": 0.20789322066449012, + "grad_norm": 1.2407195568084717, + "learning_rate": 4.681815418809464e-05, + "loss": 0.7163, + "num_input_tokens_seen": 9890464, + "step": 6495 + }, + { + "epoch": 0.20805326163497856, + "grad_norm": 1.0366379022598267, + "learning_rate": 4.681335951510819e-05, + "loss": 0.503, + "num_input_tokens_seen": 9897856, + "step": 6500 + }, + { + "epoch": 0.208213302605467, + "grad_norm": 1.1761130094528198, + "learning_rate": 4.6808561478227576e-05, + "loss": 0.6471, + "num_input_tokens_seen": 9905632, + "step": 6505 + }, + { + "epoch": 0.20837334357595544, + "grad_norm": 1.4301210641860962, + "learning_rate": 4.680376007819271e-05, + "loss": 0.676, + "num_input_tokens_seen": 9912816, + "step": 6510 + }, + { + "epoch": 0.20853338454644388, + "grad_norm": 1.1979323625564575, + "learning_rate": 4.679895531574405e-05, + "loss": 0.6989, + "num_input_tokens_seen": 9920176, + "step": 6515 + }, + { + "epoch": 0.20869342551693235, + "grad_norm": 1.355584740638733, + "learning_rate": 4.679414719162253e-05, + "loss": 0.5771, + "num_input_tokens_seen": 9927696, + "step": 6520 + }, + { + "epoch": 0.20885346648742079, + "grad_norm": 0.5972957015037537, + "learning_rate": 4.6789335706569635e-05, + "loss": 0.4672, + "num_input_tokens_seen": 9935104, + "step": 6525 + }, + { + "epoch": 0.20901350745790923, + "grad_norm": 0.7249810099601746, + "learning_rate": 4.678452086132734e-05, + "loss": 0.6899, + "num_input_tokens_seen": 9942528, + "step": 6530 + }, + { + "epoch": 0.20917354842839767, + "grad_norm": 0.5684406161308289, + "learning_rate": 4.677970265663818e-05, + "loss": 0.7085, + "num_input_tokens_seen": 9950096, + "step": 6535 + }, + { + "epoch": 0.2093335893988861, + "grad_norm": 1.0273245573043823, + "learning_rate": 4.677488109324517e-05, + "loss": 0.4942, + "num_input_tokens_seen": 9957808, + "step": 6540 + }, + { + "epoch": 0.20949363036937457, + "grad_norm": 0.7890669703483582, + "learning_rate": 4.6770056171891846e-05, + "loss": 0.4744, + "num_input_tokens_seen": 9966048, + "step": 6545 + }, + { + "epoch": 0.209653671339863, + "grad_norm": 1.3004094362258911, + "learning_rate": 4.6765227893322286e-05, + "loss": 0.5799, + "num_input_tokens_seen": 9973808, + "step": 6550 + }, + { + "epoch": 0.20981371231035145, + "grad_norm": 0.7599849700927734, + "learning_rate": 4.676039625828107e-05, + "loss": 0.5358, + "num_input_tokens_seen": 9981552, + "step": 6555 + }, + { + "epoch": 0.2099737532808399, + "grad_norm": 1.0172641277313232, + "learning_rate": 4.675556126751328e-05, + "loss": 0.7707, + "num_input_tokens_seen": 9989392, + "step": 6560 + }, + { + "epoch": 0.21013379425132833, + "grad_norm": 0.5911253094673157, + "learning_rate": 4.6750722921764556e-05, + "loss": 0.6233, + "num_input_tokens_seen": 9997312, + "step": 6565 + }, + { + "epoch": 0.21029383522181677, + "grad_norm": 1.0011309385299683, + "learning_rate": 4.674588122178102e-05, + "loss": 0.5379, + "num_input_tokens_seen": 10004768, + "step": 6570 + }, + { + "epoch": 0.21045387619230524, + "grad_norm": 0.49054577946662903, + "learning_rate": 4.674103616830931e-05, + "loss": 0.4309, + "num_input_tokens_seen": 10012384, + "step": 6575 + }, + { + "epoch": 0.21061391716279368, + "grad_norm": 1.7022589445114136, + "learning_rate": 4.673618776209663e-05, + "loss": 0.6841, + "num_input_tokens_seen": 10019680, + "step": 6580 + }, + { + "epoch": 0.21077395813328212, + "grad_norm": 0.8028180003166199, + "learning_rate": 4.673133600389063e-05, + "loss": 0.5701, + "num_input_tokens_seen": 10026896, + "step": 6585 + }, + { + "epoch": 0.21093399910377056, + "grad_norm": 1.3265849351882935, + "learning_rate": 4.672648089443953e-05, + "loss": 0.5426, + "num_input_tokens_seen": 10034288, + "step": 6590 + }, + { + "epoch": 0.211094040074259, + "grad_norm": 0.7649710178375244, + "learning_rate": 4.672162243449204e-05, + "loss": 0.4436, + "num_input_tokens_seen": 10041744, + "step": 6595 + }, + { + "epoch": 0.21125408104474747, + "grad_norm": 0.8049880266189575, + "learning_rate": 4.67167606247974e-05, + "loss": 0.7191, + "num_input_tokens_seen": 10049776, + "step": 6600 + }, + { + "epoch": 0.21125408104474747, + "eval_loss": 0.5959543585777283, + "eval_runtime": 559.6438, + "eval_samples_per_second": 24.812, + "eval_steps_per_second": 12.406, + "num_input_tokens_seen": 10049776, + "step": 6600 + }, + { + "epoch": 0.2114141220152359, + "grad_norm": 0.7298079133033752, + "learning_rate": 4.671189546610536e-05, + "loss": 0.603, + "num_input_tokens_seen": 10057616, + "step": 6605 + }, + { + "epoch": 0.21157416298572435, + "grad_norm": 1.0288366079330444, + "learning_rate": 4.67070269591662e-05, + "loss": 0.7905, + "num_input_tokens_seen": 10064800, + "step": 6610 + }, + { + "epoch": 0.21173420395621279, + "grad_norm": 0.9308603405952454, + "learning_rate": 4.670215510473068e-05, + "loss": 0.7606, + "num_input_tokens_seen": 10072320, + "step": 6615 + }, + { + "epoch": 0.21189424492670123, + "grad_norm": 0.9251461625099182, + "learning_rate": 4.669727990355013e-05, + "loss": 0.4962, + "num_input_tokens_seen": 10079696, + "step": 6620 + }, + { + "epoch": 0.2120542858971897, + "grad_norm": 0.3983932435512543, + "learning_rate": 4.669240135637635e-05, + "loss": 0.5856, + "num_input_tokens_seen": 10087024, + "step": 6625 + }, + { + "epoch": 0.21221432686767813, + "grad_norm": 1.3703784942626953, + "learning_rate": 4.6687519463961675e-05, + "loss": 0.8415, + "num_input_tokens_seen": 10094352, + "step": 6630 + }, + { + "epoch": 0.21237436783816657, + "grad_norm": 0.9022024869918823, + "learning_rate": 4.668263422705896e-05, + "loss": 0.3934, + "num_input_tokens_seen": 10101696, + "step": 6635 + }, + { + "epoch": 0.212534408808655, + "grad_norm": 1.1210694313049316, + "learning_rate": 4.667774564642156e-05, + "loss": 0.5478, + "num_input_tokens_seen": 10109472, + "step": 6640 + }, + { + "epoch": 0.21269444977914345, + "grad_norm": 0.940606415271759, + "learning_rate": 4.6672853722803365e-05, + "loss": 0.7603, + "num_input_tokens_seen": 10116992, + "step": 6645 + }, + { + "epoch": 0.21285449074963192, + "grad_norm": 0.8884837031364441, + "learning_rate": 4.666795845695877e-05, + "loss": 0.5357, + "num_input_tokens_seen": 10124496, + "step": 6650 + }, + { + "epoch": 0.21301453172012036, + "grad_norm": 0.5148705840110779, + "learning_rate": 4.666305984964269e-05, + "loss": 0.5754, + "num_input_tokens_seen": 10132272, + "step": 6655 + }, + { + "epoch": 0.2131745726906088, + "grad_norm": 0.7751322984695435, + "learning_rate": 4.6658157901610535e-05, + "loss": 0.6134, + "num_input_tokens_seen": 10139808, + "step": 6660 + }, + { + "epoch": 0.21333461366109724, + "grad_norm": 0.7671756744384766, + "learning_rate": 4.665325261361826e-05, + "loss": 0.6065, + "num_input_tokens_seen": 10146944, + "step": 6665 + }, + { + "epoch": 0.21349465463158568, + "grad_norm": 0.82252037525177, + "learning_rate": 4.664834398642232e-05, + "loss": 0.6979, + "num_input_tokens_seen": 10154000, + "step": 6670 + }, + { + "epoch": 0.21365469560207412, + "grad_norm": 0.5486531257629395, + "learning_rate": 4.6643432020779686e-05, + "loss": 0.532, + "num_input_tokens_seen": 10162032, + "step": 6675 + }, + { + "epoch": 0.21381473657256259, + "grad_norm": 1.3844165802001953, + "learning_rate": 4.663851671744786e-05, + "loss": 0.6339, + "num_input_tokens_seen": 10170272, + "step": 6680 + }, + { + "epoch": 0.21397477754305103, + "grad_norm": 0.6186041235923767, + "learning_rate": 4.6633598077184815e-05, + "loss": 0.5535, + "num_input_tokens_seen": 10177552, + "step": 6685 + }, + { + "epoch": 0.21413481851353947, + "grad_norm": 1.1645474433898926, + "learning_rate": 4.662867610074908e-05, + "loss": 0.5685, + "num_input_tokens_seen": 10185104, + "step": 6690 + }, + { + "epoch": 0.2142948594840279, + "grad_norm": 1.0970690250396729, + "learning_rate": 4.6623750788899696e-05, + "loss": 0.6911, + "num_input_tokens_seen": 10192672, + "step": 6695 + }, + { + "epoch": 0.21445490045451635, + "grad_norm": 0.6880583763122559, + "learning_rate": 4.6618822142396195e-05, + "loss": 0.713, + "num_input_tokens_seen": 10200144, + "step": 6700 + }, + { + "epoch": 0.2146149414250048, + "grad_norm": 0.8420742154121399, + "learning_rate": 4.661389016199864e-05, + "loss": 0.4788, + "num_input_tokens_seen": 10208000, + "step": 6705 + }, + { + "epoch": 0.21477498239549325, + "grad_norm": 0.9001253843307495, + "learning_rate": 4.660895484846761e-05, + "loss": 0.4414, + "num_input_tokens_seen": 10215840, + "step": 6710 + }, + { + "epoch": 0.2149350233659817, + "grad_norm": 0.8753502368927002, + "learning_rate": 4.660401620256418e-05, + "loss": 0.6521, + "num_input_tokens_seen": 10223472, + "step": 6715 + }, + { + "epoch": 0.21509506433647013, + "grad_norm": 0.40078845620155334, + "learning_rate": 4.659907422504997e-05, + "loss": 0.5325, + "num_input_tokens_seen": 10231520, + "step": 6720 + }, + { + "epoch": 0.21525510530695857, + "grad_norm": 0.8776347637176514, + "learning_rate": 4.6594128916687074e-05, + "loss": 0.5275, + "num_input_tokens_seen": 10238848, + "step": 6725 + }, + { + "epoch": 0.21541514627744704, + "grad_norm": 0.5489394664764404, + "learning_rate": 4.658918027823813e-05, + "loss": 0.3283, + "num_input_tokens_seen": 10246352, + "step": 6730 + }, + { + "epoch": 0.21557518724793548, + "grad_norm": 0.938763439655304, + "learning_rate": 4.658422831046628e-05, + "loss": 0.655, + "num_input_tokens_seen": 10254048, + "step": 6735 + }, + { + "epoch": 0.21573522821842392, + "grad_norm": 0.5266051888465881, + "learning_rate": 4.657927301413518e-05, + "loss": 0.5325, + "num_input_tokens_seen": 10261312, + "step": 6740 + }, + { + "epoch": 0.21589526918891236, + "grad_norm": 0.8076333999633789, + "learning_rate": 4.657431439000901e-05, + "loss": 0.4648, + "num_input_tokens_seen": 10268880, + "step": 6745 + }, + { + "epoch": 0.2160553101594008, + "grad_norm": 1.5385829210281372, + "learning_rate": 4.656935243885243e-05, + "loss": 0.599, + "num_input_tokens_seen": 10275856, + "step": 6750 + }, + { + "epoch": 0.21621535112988924, + "grad_norm": 0.7362918853759766, + "learning_rate": 4.656438716143066e-05, + "loss": 0.4292, + "num_input_tokens_seen": 10283696, + "step": 6755 + }, + { + "epoch": 0.2163753921003777, + "grad_norm": 0.4664347767829895, + "learning_rate": 4.6559418558509384e-05, + "loss": 0.62, + "num_input_tokens_seen": 10291328, + "step": 6760 + }, + { + "epoch": 0.21653543307086615, + "grad_norm": 0.5586669445037842, + "learning_rate": 4.6554446630854833e-05, + "loss": 0.3822, + "num_input_tokens_seen": 10298912, + "step": 6765 + }, + { + "epoch": 0.21669547404135459, + "grad_norm": 1.337626576423645, + "learning_rate": 4.654947137923374e-05, + "loss": 0.6502, + "num_input_tokens_seen": 10306288, + "step": 6770 + }, + { + "epoch": 0.21685551501184303, + "grad_norm": 1.0953319072723389, + "learning_rate": 4.654449280441335e-05, + "loss": 0.6064, + "num_input_tokens_seen": 10314400, + "step": 6775 + }, + { + "epoch": 0.21701555598233147, + "grad_norm": 0.9161373376846313, + "learning_rate": 4.653951090716143e-05, + "loss": 0.4963, + "num_input_tokens_seen": 10322240, + "step": 6780 + }, + { + "epoch": 0.21717559695281993, + "grad_norm": 0.6976571083068848, + "learning_rate": 4.653452568824625e-05, + "loss": 0.3588, + "num_input_tokens_seen": 10329744, + "step": 6785 + }, + { + "epoch": 0.21733563792330837, + "grad_norm": 0.7078070640563965, + "learning_rate": 4.6529537148436585e-05, + "loss": 0.4435, + "num_input_tokens_seen": 10337200, + "step": 6790 + }, + { + "epoch": 0.2174956788937968, + "grad_norm": 1.5676697492599487, + "learning_rate": 4.6524545288501734e-05, + "loss": 0.6237, + "num_input_tokens_seen": 10344704, + "step": 6795 + }, + { + "epoch": 0.21765571986428525, + "grad_norm": 0.8753778338432312, + "learning_rate": 4.6519550109211506e-05, + "loss": 0.4289, + "num_input_tokens_seen": 10352672, + "step": 6800 + }, + { + "epoch": 0.21765571986428525, + "eval_loss": 0.5925070643424988, + "eval_runtime": 557.707, + "eval_samples_per_second": 24.898, + "eval_steps_per_second": 12.449, + "num_input_tokens_seen": 10352672, + "step": 6800 + }, + { + "epoch": 0.2178157608347737, + "grad_norm": 0.5150783061981201, + "learning_rate": 4.651455161133622e-05, + "loss": 0.5442, + "num_input_tokens_seen": 10359824, + "step": 6805 + }, + { + "epoch": 0.21797580180526216, + "grad_norm": 0.6421305537223816, + "learning_rate": 4.6509549795646704e-05, + "loss": 0.5417, + "num_input_tokens_seen": 10367440, + "step": 6810 + }, + { + "epoch": 0.2181358427757506, + "grad_norm": 0.5854849219322205, + "learning_rate": 4.6504544662914306e-05, + "loss": 0.4962, + "num_input_tokens_seen": 10374976, + "step": 6815 + }, + { + "epoch": 0.21829588374623904, + "grad_norm": 0.45090046525001526, + "learning_rate": 4.6499536213910876e-05, + "loss": 0.5159, + "num_input_tokens_seen": 10382464, + "step": 6820 + }, + { + "epoch": 0.21845592471672748, + "grad_norm": 0.5407193303108215, + "learning_rate": 4.6494524449408786e-05, + "loss": 0.5396, + "num_input_tokens_seen": 10389888, + "step": 6825 + }, + { + "epoch": 0.21861596568721592, + "grad_norm": 0.4367115795612335, + "learning_rate": 4.6489509370180903e-05, + "loss": 0.4735, + "num_input_tokens_seen": 10397344, + "step": 6830 + }, + { + "epoch": 0.21877600665770439, + "grad_norm": 1.5306373834609985, + "learning_rate": 4.648449097700063e-05, + "loss": 0.4966, + "num_input_tokens_seen": 10405520, + "step": 6835 + }, + { + "epoch": 0.21893604762819283, + "grad_norm": 1.0457130670547485, + "learning_rate": 4.647946927064185e-05, + "loss": 0.6547, + "num_input_tokens_seen": 10412992, + "step": 6840 + }, + { + "epoch": 0.21909608859868127, + "grad_norm": 1.1099305152893066, + "learning_rate": 4.647444425187898e-05, + "loss": 0.5061, + "num_input_tokens_seen": 10420560, + "step": 6845 + }, + { + "epoch": 0.2192561295691697, + "grad_norm": 1.959218144416809, + "learning_rate": 4.646941592148695e-05, + "loss": 0.7686, + "num_input_tokens_seen": 10428368, + "step": 6850 + }, + { + "epoch": 0.21941617053965815, + "grad_norm": 1.7112497091293335, + "learning_rate": 4.646438428024117e-05, + "loss": 0.6301, + "num_input_tokens_seen": 10436320, + "step": 6855 + }, + { + "epoch": 0.21957621151014659, + "grad_norm": 1.3627498149871826, + "learning_rate": 4.64593493289176e-05, + "loss": 0.6069, + "num_input_tokens_seen": 10443616, + "step": 6860 + }, + { + "epoch": 0.21973625248063505, + "grad_norm": 1.1702673435211182, + "learning_rate": 4.64543110682927e-05, + "loss": 0.4117, + "num_input_tokens_seen": 10450688, + "step": 6865 + }, + { + "epoch": 0.2198962934511235, + "grad_norm": 1.2049485445022583, + "learning_rate": 4.644926949914341e-05, + "loss": 0.6219, + "num_input_tokens_seen": 10458480, + "step": 6870 + }, + { + "epoch": 0.22005633442161193, + "grad_norm": 0.47470933198928833, + "learning_rate": 4.644422462224722e-05, + "loss": 0.5759, + "num_input_tokens_seen": 10466080, + "step": 6875 + }, + { + "epoch": 0.22021637539210037, + "grad_norm": 0.8911974430084229, + "learning_rate": 4.643917643838211e-05, + "loss": 0.6047, + "num_input_tokens_seen": 10473872, + "step": 6880 + }, + { + "epoch": 0.2203764163625888, + "grad_norm": 1.0070406198501587, + "learning_rate": 4.6434124948326564e-05, + "loss": 0.5534, + "num_input_tokens_seen": 10481632, + "step": 6885 + }, + { + "epoch": 0.22053645733307728, + "grad_norm": 0.7700351476669312, + "learning_rate": 4.6429070152859594e-05, + "loss": 0.4492, + "num_input_tokens_seen": 10489104, + "step": 6890 + }, + { + "epoch": 0.22069649830356572, + "grad_norm": 1.025607705116272, + "learning_rate": 4.6424012052760714e-05, + "loss": 0.5453, + "num_input_tokens_seen": 10496880, + "step": 6895 + }, + { + "epoch": 0.22085653927405416, + "grad_norm": 0.8947519659996033, + "learning_rate": 4.6418950648809945e-05, + "loss": 0.537, + "num_input_tokens_seen": 10504144, + "step": 6900 + }, + { + "epoch": 0.2210165802445426, + "grad_norm": 1.2132172584533691, + "learning_rate": 4.641388594178782e-05, + "loss": 0.5671, + "num_input_tokens_seen": 10512096, + "step": 6905 + }, + { + "epoch": 0.22117662121503104, + "grad_norm": 0.46018341183662415, + "learning_rate": 4.640881793247538e-05, + "loss": 0.6063, + "num_input_tokens_seen": 10519424, + "step": 6910 + }, + { + "epoch": 0.2213366621855195, + "grad_norm": 1.0054885149002075, + "learning_rate": 4.6403746621654173e-05, + "loss": 0.6073, + "num_input_tokens_seen": 10527120, + "step": 6915 + }, + { + "epoch": 0.22149670315600795, + "grad_norm": 0.6715860366821289, + "learning_rate": 4.639867201010626e-05, + "loss": 0.4984, + "num_input_tokens_seen": 10534416, + "step": 6920 + }, + { + "epoch": 0.22165674412649639, + "grad_norm": 0.7697739601135254, + "learning_rate": 4.6393594098614204e-05, + "loss": 0.6768, + "num_input_tokens_seen": 10542048, + "step": 6925 + }, + { + "epoch": 0.22181678509698483, + "grad_norm": 1.386242389678955, + "learning_rate": 4.63885128879611e-05, + "loss": 0.7636, + "num_input_tokens_seen": 10549712, + "step": 6930 + }, + { + "epoch": 0.22197682606747327, + "grad_norm": 0.7215406894683838, + "learning_rate": 4.638342837893052e-05, + "loss": 0.5143, + "num_input_tokens_seen": 10557696, + "step": 6935 + }, + { + "epoch": 0.2221368670379617, + "grad_norm": 1.0435311794281006, + "learning_rate": 4.6378340572306565e-05, + "loss": 0.526, + "num_input_tokens_seen": 10565552, + "step": 6940 + }, + { + "epoch": 0.22229690800845017, + "grad_norm": 1.457741141319275, + "learning_rate": 4.6373249468873833e-05, + "loss": 0.5953, + "num_input_tokens_seen": 10573152, + "step": 6945 + }, + { + "epoch": 0.2224569489789386, + "grad_norm": 1.5771821737289429, + "learning_rate": 4.636815506941744e-05, + "loss": 0.502, + "num_input_tokens_seen": 10580704, + "step": 6950 + }, + { + "epoch": 0.22261698994942705, + "grad_norm": 0.956475019454956, + "learning_rate": 4.6363057374723004e-05, + "loss": 0.5076, + "num_input_tokens_seen": 10588112, + "step": 6955 + }, + { + "epoch": 0.2227770309199155, + "grad_norm": 0.7736781239509583, + "learning_rate": 4.635795638557666e-05, + "loss": 0.4506, + "num_input_tokens_seen": 10595312, + "step": 6960 + }, + { + "epoch": 0.22293707189040393, + "grad_norm": 0.43515899777412415, + "learning_rate": 4.635285210276504e-05, + "loss": 0.6302, + "num_input_tokens_seen": 10603264, + "step": 6965 + }, + { + "epoch": 0.2230971128608924, + "grad_norm": 0.6443701982498169, + "learning_rate": 4.6347744527075295e-05, + "loss": 0.4789, + "num_input_tokens_seen": 10610816, + "step": 6970 + }, + { + "epoch": 0.22325715383138084, + "grad_norm": 1.3230254650115967, + "learning_rate": 4.634263365929506e-05, + "loss": 0.5968, + "num_input_tokens_seen": 10618096, + "step": 6975 + }, + { + "epoch": 0.22341719480186928, + "grad_norm": 1.2190051078796387, + "learning_rate": 4.6337519500212515e-05, + "loss": 0.4438, + "num_input_tokens_seen": 10625168, + "step": 6980 + }, + { + "epoch": 0.22357723577235772, + "grad_norm": 0.7341931462287903, + "learning_rate": 4.633240205061632e-05, + "loss": 0.5609, + "num_input_tokens_seen": 10632496, + "step": 6985 + }, + { + "epoch": 0.22373727674284616, + "grad_norm": 0.6002076268196106, + "learning_rate": 4.632728131129565e-05, + "loss": 0.416, + "num_input_tokens_seen": 10640016, + "step": 6990 + }, + { + "epoch": 0.22389731771333463, + "grad_norm": 1.3966522216796875, + "learning_rate": 4.632215728304018e-05, + "loss": 0.7361, + "num_input_tokens_seen": 10647744, + "step": 6995 + }, + { + "epoch": 0.22405735868382307, + "grad_norm": 0.4457888603210449, + "learning_rate": 4.63170299666401e-05, + "loss": 0.4285, + "num_input_tokens_seen": 10655360, + "step": 7000 + }, + { + "epoch": 0.22405735868382307, + "eval_loss": 0.5903034210205078, + "eval_runtime": 560.8834, + "eval_samples_per_second": 24.757, + "eval_steps_per_second": 12.379, + "num_input_tokens_seen": 10655360, + "step": 7000 + }, + { + "epoch": 0.2242173996543115, + "grad_norm": 1.155311107635498, + "learning_rate": 4.631189936288612e-05, + "loss": 0.5694, + "num_input_tokens_seen": 10662672, + "step": 7005 + }, + { + "epoch": 0.22437744062479995, + "grad_norm": 0.7725893259048462, + "learning_rate": 4.630676547256944e-05, + "loss": 0.4767, + "num_input_tokens_seen": 10670096, + "step": 7010 + }, + { + "epoch": 0.22453748159528839, + "grad_norm": 0.9227953553199768, + "learning_rate": 4.630162829648176e-05, + "loss": 0.5286, + "num_input_tokens_seen": 10678096, + "step": 7015 + }, + { + "epoch": 0.22469752256577685, + "grad_norm": 0.7940772175788879, + "learning_rate": 4.629648783541531e-05, + "loss": 0.6756, + "num_input_tokens_seen": 10685728, + "step": 7020 + }, + { + "epoch": 0.2248575635362653, + "grad_norm": 1.4991520643234253, + "learning_rate": 4.6291344090162804e-05, + "loss": 0.5432, + "num_input_tokens_seen": 10694064, + "step": 7025 + }, + { + "epoch": 0.22501760450675373, + "grad_norm": 1.8165603876113892, + "learning_rate": 4.628619706151748e-05, + "loss": 0.7617, + "num_input_tokens_seen": 10701600, + "step": 7030 + }, + { + "epoch": 0.22517764547724217, + "grad_norm": 0.759185791015625, + "learning_rate": 4.628104675027306e-05, + "loss": 0.435, + "num_input_tokens_seen": 10709168, + "step": 7035 + }, + { + "epoch": 0.2253376864477306, + "grad_norm": 1.1036840677261353, + "learning_rate": 4.6275893157223805e-05, + "loss": 0.5432, + "num_input_tokens_seen": 10717136, + "step": 7040 + }, + { + "epoch": 0.22549772741821905, + "grad_norm": 0.5456637144088745, + "learning_rate": 4.627073628316445e-05, + "loss": 0.4676, + "num_input_tokens_seen": 10725104, + "step": 7045 + }, + { + "epoch": 0.22565776838870752, + "grad_norm": 0.6687513589859009, + "learning_rate": 4.626557612889026e-05, + "loss": 0.4515, + "num_input_tokens_seen": 10732416, + "step": 7050 + }, + { + "epoch": 0.22581780935919596, + "grad_norm": 0.7775974869728088, + "learning_rate": 4.626041269519699e-05, + "loss": 0.4982, + "num_input_tokens_seen": 10739872, + "step": 7055 + }, + { + "epoch": 0.2259778503296844, + "grad_norm": 0.6791682243347168, + "learning_rate": 4.6255245982880905e-05, + "loss": 0.5764, + "num_input_tokens_seen": 10747600, + "step": 7060 + }, + { + "epoch": 0.22613789130017284, + "grad_norm": 0.45900294184684753, + "learning_rate": 4.625007599273879e-05, + "loss": 0.4525, + "num_input_tokens_seen": 10755408, + "step": 7065 + }, + { + "epoch": 0.22629793227066128, + "grad_norm": 1.0050876140594482, + "learning_rate": 4.6244902725567895e-05, + "loss": 0.4795, + "num_input_tokens_seen": 10763168, + "step": 7070 + }, + { + "epoch": 0.22645797324114975, + "grad_norm": 1.1286191940307617, + "learning_rate": 4.6239726182166024e-05, + "loss": 0.6553, + "num_input_tokens_seen": 10770304, + "step": 7075 + }, + { + "epoch": 0.22661801421163819, + "grad_norm": 0.8003321886062622, + "learning_rate": 4.623454636333147e-05, + "loss": 0.5893, + "num_input_tokens_seen": 10777888, + "step": 7080 + }, + { + "epoch": 0.22677805518212663, + "grad_norm": 0.5903188586235046, + "learning_rate": 4.622936326986301e-05, + "loss": 0.4695, + "num_input_tokens_seen": 10785376, + "step": 7085 + }, + { + "epoch": 0.22693809615261507, + "grad_norm": 0.8501474857330322, + "learning_rate": 4.6224176902559946e-05, + "loss": 0.4723, + "num_input_tokens_seen": 10793200, + "step": 7090 + }, + { + "epoch": 0.2270981371231035, + "grad_norm": 0.9826895594596863, + "learning_rate": 4.621898726222209e-05, + "loss": 0.4636, + "num_input_tokens_seen": 10800592, + "step": 7095 + }, + { + "epoch": 0.22725817809359197, + "grad_norm": 0.8003636002540588, + "learning_rate": 4.6213794349649744e-05, + "loss": 0.3843, + "num_input_tokens_seen": 10808528, + "step": 7100 + }, + { + "epoch": 0.2274182190640804, + "grad_norm": 0.8699879050254822, + "learning_rate": 4.6208598165643715e-05, + "loss": 0.5867, + "num_input_tokens_seen": 10815792, + "step": 7105 + }, + { + "epoch": 0.22757826003456885, + "grad_norm": 0.5492406487464905, + "learning_rate": 4.620339871100533e-05, + "loss": 0.5331, + "num_input_tokens_seen": 10823312, + "step": 7110 + }, + { + "epoch": 0.2277383010050573, + "grad_norm": 1.0736478567123413, + "learning_rate": 4.6198195986536394e-05, + "loss": 0.716, + "num_input_tokens_seen": 10830768, + "step": 7115 + }, + { + "epoch": 0.22789834197554573, + "grad_norm": 0.9929563403129578, + "learning_rate": 4.619298999303926e-05, + "loss": 0.5786, + "num_input_tokens_seen": 10838448, + "step": 7120 + }, + { + "epoch": 0.22805838294603417, + "grad_norm": 0.6287401914596558, + "learning_rate": 4.618778073131673e-05, + "loss": 0.4863, + "num_input_tokens_seen": 10846272, + "step": 7125 + }, + { + "epoch": 0.22821842391652264, + "grad_norm": 0.7692697048187256, + "learning_rate": 4.618256820217215e-05, + "loss": 0.7292, + "num_input_tokens_seen": 10853472, + "step": 7130 + }, + { + "epoch": 0.22837846488701108, + "grad_norm": 0.9424571990966797, + "learning_rate": 4.617735240640936e-05, + "loss": 0.4059, + "num_input_tokens_seen": 10860864, + "step": 7135 + }, + { + "epoch": 0.22853850585749952, + "grad_norm": 0.7671375870704651, + "learning_rate": 4.6172133344832705e-05, + "loss": 0.653, + "num_input_tokens_seen": 10868736, + "step": 7140 + }, + { + "epoch": 0.22869854682798796, + "grad_norm": 0.9138753414154053, + "learning_rate": 4.6166911018247004e-05, + "loss": 0.5725, + "num_input_tokens_seen": 10876320, + "step": 7145 + }, + { + "epoch": 0.2288585877984764, + "grad_norm": 0.9469364285469055, + "learning_rate": 4.616168542745764e-05, + "loss": 0.678, + "num_input_tokens_seen": 10883744, + "step": 7150 + }, + { + "epoch": 0.22901862876896487, + "grad_norm": 2.2541005611419678, + "learning_rate": 4.6156456573270446e-05, + "loss": 0.5681, + "num_input_tokens_seen": 10892304, + "step": 7155 + }, + { + "epoch": 0.2291786697394533, + "grad_norm": 1.7159606218338013, + "learning_rate": 4.615122445649177e-05, + "loss": 0.6695, + "num_input_tokens_seen": 10899808, + "step": 7160 + }, + { + "epoch": 0.22933871070994175, + "grad_norm": 0.5276926159858704, + "learning_rate": 4.6145989077928486e-05, + "loss": 0.5275, + "num_input_tokens_seen": 10907152, + "step": 7165 + }, + { + "epoch": 0.22949875168043019, + "grad_norm": 1.329194188117981, + "learning_rate": 4.6140750438387953e-05, + "loss": 0.5444, + "num_input_tokens_seen": 10914912, + "step": 7170 + }, + { + "epoch": 0.22965879265091863, + "grad_norm": 0.9011327624320984, + "learning_rate": 4.613550853867803e-05, + "loss": 0.4492, + "num_input_tokens_seen": 10922640, + "step": 7175 + }, + { + "epoch": 0.2298188336214071, + "grad_norm": 1.3555457592010498, + "learning_rate": 4.613026337960708e-05, + "loss": 0.5513, + "num_input_tokens_seen": 10930416, + "step": 7180 + }, + { + "epoch": 0.22997887459189553, + "grad_norm": 0.9139378070831299, + "learning_rate": 4.612501496198398e-05, + "loss": 0.6505, + "num_input_tokens_seen": 10938272, + "step": 7185 + }, + { + "epoch": 0.23013891556238397, + "grad_norm": 1.362973928451538, + "learning_rate": 4.61197632866181e-05, + "loss": 0.4803, + "num_input_tokens_seen": 10945936, + "step": 7190 + }, + { + "epoch": 0.2302989565328724, + "grad_norm": 1.2961748838424683, + "learning_rate": 4.611450835431931e-05, + "loss": 0.6692, + "num_input_tokens_seen": 10953600, + "step": 7195 + }, + { + "epoch": 0.23045899750336085, + "grad_norm": 1.0778976678848267, + "learning_rate": 4.6109250165898e-05, + "loss": 0.6494, + "num_input_tokens_seen": 10961872, + "step": 7200 + }, + { + "epoch": 0.23045899750336085, + "eval_loss": 0.5881946682929993, + "eval_runtime": 563.19, + "eval_samples_per_second": 24.656, + "eval_steps_per_second": 12.328, + "num_input_tokens_seen": 10961872, + "step": 7200 + }, + { + "epoch": 0.2306190384738493, + "grad_norm": 0.6386324167251587, + "learning_rate": 4.610398872216503e-05, + "loss": 0.4594, + "num_input_tokens_seen": 10969920, + "step": 7205 + }, + { + "epoch": 0.23077907944433776, + "grad_norm": 0.6154646873474121, + "learning_rate": 4.6098724023931796e-05, + "loss": 0.3942, + "num_input_tokens_seen": 10977632, + "step": 7210 + }, + { + "epoch": 0.2309391204148262, + "grad_norm": 1.0100737810134888, + "learning_rate": 4.609345607201017e-05, + "loss": 0.4929, + "num_input_tokens_seen": 10985184, + "step": 7215 + }, + { + "epoch": 0.23109916138531464, + "grad_norm": 1.406501293182373, + "learning_rate": 4.608818486721254e-05, + "loss": 0.4974, + "num_input_tokens_seen": 10992464, + "step": 7220 + }, + { + "epoch": 0.23125920235580308, + "grad_norm": 1.3209435939788818, + "learning_rate": 4.608291041035179e-05, + "loss": 0.5623, + "num_input_tokens_seen": 10999936, + "step": 7225 + }, + { + "epoch": 0.23141924332629152, + "grad_norm": 1.6173430681228638, + "learning_rate": 4.607763270224132e-05, + "loss": 0.5107, + "num_input_tokens_seen": 11007648, + "step": 7230 + }, + { + "epoch": 0.23157928429677999, + "grad_norm": 1.516273856163025, + "learning_rate": 4.6072351743695e-05, + "loss": 0.5112, + "num_input_tokens_seen": 11015168, + "step": 7235 + }, + { + "epoch": 0.23173932526726843, + "grad_norm": 1.322184443473816, + "learning_rate": 4.606706753552723e-05, + "loss": 0.5473, + "num_input_tokens_seen": 11022944, + "step": 7240 + }, + { + "epoch": 0.23189936623775687, + "grad_norm": 0.6504122018814087, + "learning_rate": 4.6061780078552906e-05, + "loss": 0.4957, + "num_input_tokens_seen": 11030384, + "step": 7245 + }, + { + "epoch": 0.2320594072082453, + "grad_norm": 0.555223822593689, + "learning_rate": 4.605648937358742e-05, + "loss": 0.567, + "num_input_tokens_seen": 11038080, + "step": 7250 + }, + { + "epoch": 0.23221944817873375, + "grad_norm": 0.7712476253509521, + "learning_rate": 4.605119542144665e-05, + "loss": 0.6105, + "num_input_tokens_seen": 11045728, + "step": 7255 + }, + { + "epoch": 0.2323794891492222, + "grad_norm": 0.6425728797912598, + "learning_rate": 4.604589822294701e-05, + "loss": 0.4608, + "num_input_tokens_seen": 11053200, + "step": 7260 + }, + { + "epoch": 0.23253953011971065, + "grad_norm": 0.9567745923995972, + "learning_rate": 4.604059777890537e-05, + "loss": 0.5576, + "num_input_tokens_seen": 11061264, + "step": 7265 + }, + { + "epoch": 0.2326995710901991, + "grad_norm": 0.6859607696533203, + "learning_rate": 4.6035294090139145e-05, + "loss": 0.4232, + "num_input_tokens_seen": 11069424, + "step": 7270 + }, + { + "epoch": 0.23285961206068753, + "grad_norm": 0.693625271320343, + "learning_rate": 4.6029987157466226e-05, + "loss": 0.5262, + "num_input_tokens_seen": 11077072, + "step": 7275 + }, + { + "epoch": 0.23301965303117597, + "grad_norm": 1.221483826637268, + "learning_rate": 4.602467698170502e-05, + "loss": 0.6266, + "num_input_tokens_seen": 11085232, + "step": 7280 + }, + { + "epoch": 0.23317969400166444, + "grad_norm": 0.6756184101104736, + "learning_rate": 4.601936356367439e-05, + "loss": 0.5529, + "num_input_tokens_seen": 11092976, + "step": 7285 + }, + { + "epoch": 0.23333973497215288, + "grad_norm": 1.726340651512146, + "learning_rate": 4.601404690419377e-05, + "loss": 0.5834, + "num_input_tokens_seen": 11100720, + "step": 7290 + }, + { + "epoch": 0.23349977594264132, + "grad_norm": 0.8472786545753479, + "learning_rate": 4.600872700408303e-05, + "loss": 0.6855, + "num_input_tokens_seen": 11108128, + "step": 7295 + }, + { + "epoch": 0.23365981691312976, + "grad_norm": 1.244470238685608, + "learning_rate": 4.600340386416258e-05, + "loss": 0.7551, + "num_input_tokens_seen": 11115312, + "step": 7300 + }, + { + "epoch": 0.2338198578836182, + "grad_norm": 1.140711784362793, + "learning_rate": 4.5998077485253296e-05, + "loss": 0.5611, + "num_input_tokens_seen": 11122928, + "step": 7305 + }, + { + "epoch": 0.23397989885410664, + "grad_norm": 1.2659752368927002, + "learning_rate": 4.59927478681766e-05, + "loss": 0.7874, + "num_input_tokens_seen": 11130448, + "step": 7310 + }, + { + "epoch": 0.2341399398245951, + "grad_norm": 1.108168125152588, + "learning_rate": 4.5987415013754366e-05, + "loss": 0.6735, + "num_input_tokens_seen": 11138800, + "step": 7315 + }, + { + "epoch": 0.23429998079508355, + "grad_norm": 0.4618755280971527, + "learning_rate": 4.598207892280899e-05, + "loss": 0.5478, + "num_input_tokens_seen": 11146240, + "step": 7320 + }, + { + "epoch": 0.23446002176557199, + "grad_norm": 1.374781608581543, + "learning_rate": 4.597673959616337e-05, + "loss": 0.6036, + "num_input_tokens_seen": 11153648, + "step": 7325 + }, + { + "epoch": 0.23462006273606043, + "grad_norm": 1.0626747608184814, + "learning_rate": 4.597139703464089e-05, + "loss": 0.52, + "num_input_tokens_seen": 11161312, + "step": 7330 + }, + { + "epoch": 0.23478010370654886, + "grad_norm": 0.3776513338088989, + "learning_rate": 4.596605123906545e-05, + "loss": 0.5226, + "num_input_tokens_seen": 11168656, + "step": 7335 + }, + { + "epoch": 0.23494014467703733, + "grad_norm": 0.6737067103385925, + "learning_rate": 4.596070221026143e-05, + "loss": 0.4556, + "num_input_tokens_seen": 11175840, + "step": 7340 + }, + { + "epoch": 0.23510018564752577, + "grad_norm": 1.0530197620391846, + "learning_rate": 4.595534994905372e-05, + "loss": 0.5293, + "num_input_tokens_seen": 11183296, + "step": 7345 + }, + { + "epoch": 0.2352602266180142, + "grad_norm": 1.5087146759033203, + "learning_rate": 4.594999445626771e-05, + "loss": 0.5835, + "num_input_tokens_seen": 11190720, + "step": 7350 + }, + { + "epoch": 0.23542026758850265, + "grad_norm": 0.8073695302009583, + "learning_rate": 4.5944635732729276e-05, + "loss": 0.6772, + "num_input_tokens_seen": 11197872, + "step": 7355 + }, + { + "epoch": 0.2355803085589911, + "grad_norm": 0.7782508730888367, + "learning_rate": 4.5939273779264804e-05, + "loss": 0.6478, + "num_input_tokens_seen": 11205472, + "step": 7360 + }, + { + "epoch": 0.23574034952947956, + "grad_norm": 1.602340817451477, + "learning_rate": 4.593390859670118e-05, + "loss": 0.7186, + "num_input_tokens_seen": 11213008, + "step": 7365 + }, + { + "epoch": 0.235900390499968, + "grad_norm": 1.548722505569458, + "learning_rate": 4.5928540185865776e-05, + "loss": 0.5274, + "num_input_tokens_seen": 11220528, + "step": 7370 + }, + { + "epoch": 0.23606043147045644, + "grad_norm": 1.2151705026626587, + "learning_rate": 4.592316854758648e-05, + "loss": 0.7148, + "num_input_tokens_seen": 11227920, + "step": 7375 + }, + { + "epoch": 0.23622047244094488, + "grad_norm": 0.6732231378555298, + "learning_rate": 4.5917793682691646e-05, + "loss": 0.5301, + "num_input_tokens_seen": 11235328, + "step": 7380 + }, + { + "epoch": 0.23638051341143332, + "grad_norm": 0.43421560525894165, + "learning_rate": 4.5912415592010164e-05, + "loss": 0.532, + "num_input_tokens_seen": 11242496, + "step": 7385 + }, + { + "epoch": 0.23654055438192176, + "grad_norm": 0.9298310875892639, + "learning_rate": 4.5907034276371386e-05, + "loss": 0.7177, + "num_input_tokens_seen": 11249552, + "step": 7390 + }, + { + "epoch": 0.23670059535241023, + "grad_norm": 0.5620187520980835, + "learning_rate": 4.5901649736605196e-05, + "loss": 0.3373, + "num_input_tokens_seen": 11257760, + "step": 7395 + }, + { + "epoch": 0.23686063632289867, + "grad_norm": 0.646039605140686, + "learning_rate": 4.589626197354195e-05, + "loss": 0.5142, + "num_input_tokens_seen": 11265584, + "step": 7400 + }, + { + "epoch": 0.23686063632289867, + "eval_loss": 0.5867252349853516, + "eval_runtime": 560.16, + "eval_samples_per_second": 24.789, + "eval_steps_per_second": 12.395, + "num_input_tokens_seen": 11265584, + "step": 7400 + }, + { + "epoch": 0.2370206772933871, + "grad_norm": 1.4174046516418457, + "learning_rate": 4.5890870988012504e-05, + "loss": 0.5035, + "num_input_tokens_seen": 11272768, + "step": 7405 + }, + { + "epoch": 0.23718071826387555, + "grad_norm": 0.9069093465805054, + "learning_rate": 4.5885476780848226e-05, + "loss": 0.7707, + "num_input_tokens_seen": 11280928, + "step": 7410 + }, + { + "epoch": 0.23734075923436398, + "grad_norm": 0.7916245460510254, + "learning_rate": 4.5880079352880964e-05, + "loss": 0.4177, + "num_input_tokens_seen": 11288576, + "step": 7415 + }, + { + "epoch": 0.23750080020485245, + "grad_norm": 0.7692219018936157, + "learning_rate": 4.5874678704943065e-05, + "loss": 0.5573, + "num_input_tokens_seen": 11296416, + "step": 7420 + }, + { + "epoch": 0.2376608411753409, + "grad_norm": 1.381445288658142, + "learning_rate": 4.5869274837867394e-05, + "loss": 0.7238, + "num_input_tokens_seen": 11303824, + "step": 7425 + }, + { + "epoch": 0.23782088214582933, + "grad_norm": 1.2720359563827515, + "learning_rate": 4.5863867752487275e-05, + "loss": 0.461, + "num_input_tokens_seen": 11311104, + "step": 7430 + }, + { + "epoch": 0.23798092311631777, + "grad_norm": 0.7390636205673218, + "learning_rate": 4.5858457449636554e-05, + "loss": 0.4209, + "num_input_tokens_seen": 11319296, + "step": 7435 + }, + { + "epoch": 0.2381409640868062, + "grad_norm": 0.5132645964622498, + "learning_rate": 4.5853043930149574e-05, + "loss": 0.6093, + "num_input_tokens_seen": 11326848, + "step": 7440 + }, + { + "epoch": 0.23830100505729468, + "grad_norm": 0.8615275621414185, + "learning_rate": 4.584762719486117e-05, + "loss": 0.6391, + "num_input_tokens_seen": 11334288, + "step": 7445 + }, + { + "epoch": 0.23846104602778312, + "grad_norm": 1.7003568410873413, + "learning_rate": 4.584220724460665e-05, + "loss": 0.6552, + "num_input_tokens_seen": 11341568, + "step": 7450 + }, + { + "epoch": 0.23862108699827156, + "grad_norm": 0.6587767004966736, + "learning_rate": 4.5836784080221865e-05, + "loss": 0.4895, + "num_input_tokens_seen": 11348944, + "step": 7455 + }, + { + "epoch": 0.23878112796876, + "grad_norm": 0.5426783561706543, + "learning_rate": 4.583135770254312e-05, + "loss": 0.5787, + "num_input_tokens_seen": 11357008, + "step": 7460 + }, + { + "epoch": 0.23894116893924844, + "grad_norm": 0.8802905678749084, + "learning_rate": 4.5825928112407236e-05, + "loss": 0.5091, + "num_input_tokens_seen": 11364608, + "step": 7465 + }, + { + "epoch": 0.2391012099097369, + "grad_norm": 0.5879753828048706, + "learning_rate": 4.582049531065152e-05, + "loss": 0.5432, + "num_input_tokens_seen": 11372480, + "step": 7470 + }, + { + "epoch": 0.23926125088022535, + "grad_norm": 1.0821019411087036, + "learning_rate": 4.5815059298113783e-05, + "loss": 0.5146, + "num_input_tokens_seen": 11379952, + "step": 7475 + }, + { + "epoch": 0.23942129185071379, + "grad_norm": 0.7071691751480103, + "learning_rate": 4.580962007563232e-05, + "loss": 0.5271, + "num_input_tokens_seen": 11387744, + "step": 7480 + }, + { + "epoch": 0.23958133282120223, + "grad_norm": 0.6058462262153625, + "learning_rate": 4.5804177644045935e-05, + "loss": 0.5017, + "num_input_tokens_seen": 11395376, + "step": 7485 + }, + { + "epoch": 0.23974137379169067, + "grad_norm": 0.7756891846656799, + "learning_rate": 4.579873200419391e-05, + "loss": 0.514, + "num_input_tokens_seen": 11403024, + "step": 7490 + }, + { + "epoch": 0.2399014147621791, + "grad_norm": 0.6251128315925598, + "learning_rate": 4.5793283156916046e-05, + "loss": 0.757, + "num_input_tokens_seen": 11410368, + "step": 7495 + }, + { + "epoch": 0.24006145573266757, + "grad_norm": 1.2918413877487183, + "learning_rate": 4.578783110305261e-05, + "loss": 0.5775, + "num_input_tokens_seen": 11418432, + "step": 7500 + }, + { + "epoch": 0.240221496703156, + "grad_norm": 1.998584270477295, + "learning_rate": 4.578237584344438e-05, + "loss": 0.5179, + "num_input_tokens_seen": 11426112, + "step": 7505 + }, + { + "epoch": 0.24038153767364445, + "grad_norm": 0.6153678297996521, + "learning_rate": 4.577691737893263e-05, + "loss": 0.6027, + "num_input_tokens_seen": 11433648, + "step": 7510 + }, + { + "epoch": 0.2405415786441329, + "grad_norm": 0.6478773355484009, + "learning_rate": 4.577145571035912e-05, + "loss": 0.537, + "num_input_tokens_seen": 11441184, + "step": 7515 + }, + { + "epoch": 0.24070161961462133, + "grad_norm": 0.6089341640472412, + "learning_rate": 4.576599083856611e-05, + "loss": 0.5453, + "num_input_tokens_seen": 11448768, + "step": 7520 + }, + { + "epoch": 0.2408616605851098, + "grad_norm": 0.4321328401565552, + "learning_rate": 4.576052276439635e-05, + "loss": 0.4976, + "num_input_tokens_seen": 11456112, + "step": 7525 + }, + { + "epoch": 0.24102170155559824, + "grad_norm": 0.9484442472457886, + "learning_rate": 4.575505148869308e-05, + "loss": 0.5342, + "num_input_tokens_seen": 11463696, + "step": 7530 + }, + { + "epoch": 0.24118174252608668, + "grad_norm": 0.6148832440376282, + "learning_rate": 4.574957701230006e-05, + "loss": 0.4986, + "num_input_tokens_seen": 11471648, + "step": 7535 + }, + { + "epoch": 0.24134178349657512, + "grad_norm": 0.992352306842804, + "learning_rate": 4.57440993360615e-05, + "loss": 0.4159, + "num_input_tokens_seen": 11479760, + "step": 7540 + }, + { + "epoch": 0.24150182446706356, + "grad_norm": 0.4656471014022827, + "learning_rate": 4.5738618460822134e-05, + "loss": 0.5135, + "num_input_tokens_seen": 11487440, + "step": 7545 + }, + { + "epoch": 0.24166186543755203, + "grad_norm": 0.5173009037971497, + "learning_rate": 4.573313438742719e-05, + "loss": 0.589, + "num_input_tokens_seen": 11495296, + "step": 7550 + }, + { + "epoch": 0.24182190640804047, + "grad_norm": 1.1084269285202026, + "learning_rate": 4.5727647116722374e-05, + "loss": 0.4431, + "num_input_tokens_seen": 11503312, + "step": 7555 + }, + { + "epoch": 0.2419819473785289, + "grad_norm": 1.0903600454330444, + "learning_rate": 4.5722156649553884e-05, + "loss": 0.6168, + "num_input_tokens_seen": 11511136, + "step": 7560 + }, + { + "epoch": 0.24214198834901735, + "grad_norm": 0.8676537871360779, + "learning_rate": 4.571666298676843e-05, + "loss": 0.5619, + "num_input_tokens_seen": 11518352, + "step": 7565 + }, + { + "epoch": 0.24230202931950579, + "grad_norm": 3.2293293476104736, + "learning_rate": 4.571116612921321e-05, + "loss": 0.6327, + "num_input_tokens_seen": 11525648, + "step": 7570 + }, + { + "epoch": 0.24246207028999422, + "grad_norm": 0.9315980076789856, + "learning_rate": 4.57056660777359e-05, + "loss": 0.5151, + "num_input_tokens_seen": 11533296, + "step": 7575 + }, + { + "epoch": 0.2426221112604827, + "grad_norm": 1.008633017539978, + "learning_rate": 4.5700162833184666e-05, + "loss": 0.5027, + "num_input_tokens_seen": 11541312, + "step": 7580 + }, + { + "epoch": 0.24278215223097113, + "grad_norm": 0.48139169812202454, + "learning_rate": 4.5694656396408195e-05, + "loss": 0.3728, + "num_input_tokens_seen": 11549136, + "step": 7585 + }, + { + "epoch": 0.24294219320145957, + "grad_norm": 0.41830405592918396, + "learning_rate": 4.5689146768255646e-05, + "loss": 0.5126, + "num_input_tokens_seen": 11556480, + "step": 7590 + }, + { + "epoch": 0.243102234171948, + "grad_norm": 0.8585829734802246, + "learning_rate": 4.568363394957667e-05, + "loss": 0.603, + "num_input_tokens_seen": 11564432, + "step": 7595 + }, + { + "epoch": 0.24326227514243645, + "grad_norm": 0.9507288336753845, + "learning_rate": 4.567811794122141e-05, + "loss": 0.6324, + "num_input_tokens_seen": 11571872, + "step": 7600 + }, + { + "epoch": 0.24326227514243645, + "eval_loss": 0.5849308967590332, + "eval_runtime": 558.2056, + "eval_samples_per_second": 24.876, + "eval_steps_per_second": 12.438, + "num_input_tokens_seen": 11571872, + "step": 7600 + }, + { + "epoch": 0.24342231611292492, + "grad_norm": 1.5527912378311157, + "learning_rate": 4.56725987440405e-05, + "loss": 0.6384, + "num_input_tokens_seen": 11579728, + "step": 7605 + }, + { + "epoch": 0.24358235708341336, + "grad_norm": 0.49759647250175476, + "learning_rate": 4.566707635888508e-05, + "loss": 0.618, + "num_input_tokens_seen": 11587376, + "step": 7610 + }, + { + "epoch": 0.2437423980539018, + "grad_norm": 0.9339687824249268, + "learning_rate": 4.566155078660677e-05, + "loss": 0.4531, + "num_input_tokens_seen": 11595344, + "step": 7615 + }, + { + "epoch": 0.24390243902439024, + "grad_norm": 1.2929613590240479, + "learning_rate": 4.565602202805768e-05, + "loss": 0.6455, + "num_input_tokens_seen": 11602816, + "step": 7620 + }, + { + "epoch": 0.24406247999487868, + "grad_norm": 0.640177845954895, + "learning_rate": 4.56504900840904e-05, + "loss": 0.5298, + "num_input_tokens_seen": 11610976, + "step": 7625 + }, + { + "epoch": 0.24422252096536715, + "grad_norm": 0.9480876922607422, + "learning_rate": 4.564495495555805e-05, + "loss": 0.5772, + "num_input_tokens_seen": 11618320, + "step": 7630 + }, + { + "epoch": 0.24438256193585559, + "grad_norm": 0.8721270561218262, + "learning_rate": 4.5639416643314204e-05, + "loss": 0.4913, + "num_input_tokens_seen": 11625760, + "step": 7635 + }, + { + "epoch": 0.24454260290634403, + "grad_norm": 1.198887825012207, + "learning_rate": 4.5633875148212946e-05, + "loss": 0.7196, + "num_input_tokens_seen": 11633600, + "step": 7640 + }, + { + "epoch": 0.24470264387683247, + "grad_norm": 0.8347474932670593, + "learning_rate": 4.562833047110883e-05, + "loss": 0.6877, + "num_input_tokens_seen": 11641664, + "step": 7645 + }, + { + "epoch": 0.2448626848473209, + "grad_norm": 0.44610029458999634, + "learning_rate": 4.5622782612856923e-05, + "loss": 0.4547, + "num_input_tokens_seen": 11649392, + "step": 7650 + }, + { + "epoch": 0.24502272581780937, + "grad_norm": 1.1177787780761719, + "learning_rate": 4.561723157431278e-05, + "loss": 0.5314, + "num_input_tokens_seen": 11657056, + "step": 7655 + }, + { + "epoch": 0.2451827667882978, + "grad_norm": 0.8463836908340454, + "learning_rate": 4.5611677356332435e-05, + "loss": 0.4927, + "num_input_tokens_seen": 11664864, + "step": 7660 + }, + { + "epoch": 0.24534280775878625, + "grad_norm": 0.7779257297515869, + "learning_rate": 4.560611995977242e-05, + "loss": 0.3666, + "num_input_tokens_seen": 11672528, + "step": 7665 + }, + { + "epoch": 0.2455028487292747, + "grad_norm": 0.8962683081626892, + "learning_rate": 4.560055938548975e-05, + "loss": 0.5243, + "num_input_tokens_seen": 11679760, + "step": 7670 + }, + { + "epoch": 0.24566288969976313, + "grad_norm": 0.6146455407142639, + "learning_rate": 4.5594995634341944e-05, + "loss": 0.452, + "num_input_tokens_seen": 11687792, + "step": 7675 + }, + { + "epoch": 0.24582293067025157, + "grad_norm": 1.4697914123535156, + "learning_rate": 4.5589428707187e-05, + "loss": 0.6042, + "num_input_tokens_seen": 11695072, + "step": 7680 + }, + { + "epoch": 0.24598297164074004, + "grad_norm": 0.6150116324424744, + "learning_rate": 4.55838586048834e-05, + "loss": 0.6114, + "num_input_tokens_seen": 11702160, + "step": 7685 + }, + { + "epoch": 0.24614301261122848, + "grad_norm": 0.8411738872528076, + "learning_rate": 4.557828532829013e-05, + "loss": 0.5398, + "num_input_tokens_seen": 11709696, + "step": 7690 + }, + { + "epoch": 0.24630305358171692, + "grad_norm": 0.7127346396446228, + "learning_rate": 4.557270887826667e-05, + "loss": 0.4978, + "num_input_tokens_seen": 11717376, + "step": 7695 + }, + { + "epoch": 0.24646309455220536, + "grad_norm": 0.35730963945388794, + "learning_rate": 4.556712925567296e-05, + "loss": 0.4999, + "num_input_tokens_seen": 11725264, + "step": 7700 + }, + { + "epoch": 0.2466231355226938, + "grad_norm": 0.497885525226593, + "learning_rate": 4.5561546461369454e-05, + "loss": 0.4717, + "num_input_tokens_seen": 11733232, + "step": 7705 + }, + { + "epoch": 0.24678317649318227, + "grad_norm": 1.1125736236572266, + "learning_rate": 4.55559604962171e-05, + "loss": 0.4758, + "num_input_tokens_seen": 11741504, + "step": 7710 + }, + { + "epoch": 0.2469432174636707, + "grad_norm": 0.6268452405929565, + "learning_rate": 4.55503713610773e-05, + "loss": 0.5538, + "num_input_tokens_seen": 11749088, + "step": 7715 + }, + { + "epoch": 0.24710325843415915, + "grad_norm": 0.8956989645957947, + "learning_rate": 4.5544779056812e-05, + "loss": 0.4958, + "num_input_tokens_seen": 11756576, + "step": 7720 + }, + { + "epoch": 0.24726329940464759, + "grad_norm": 0.6975457072257996, + "learning_rate": 4.553918358428358e-05, + "loss": 0.5205, + "num_input_tokens_seen": 11764080, + "step": 7725 + }, + { + "epoch": 0.24742334037513602, + "grad_norm": 0.9252241253852844, + "learning_rate": 4.553358494435494e-05, + "loss": 0.5788, + "num_input_tokens_seen": 11771456, + "step": 7730 + }, + { + "epoch": 0.2475833813456245, + "grad_norm": 0.8005891442298889, + "learning_rate": 4.5527983137889464e-05, + "loss": 0.637, + "num_input_tokens_seen": 11779200, + "step": 7735 + }, + { + "epoch": 0.24774342231611293, + "grad_norm": 0.6644135117530823, + "learning_rate": 4.5522378165751015e-05, + "loss": 0.6622, + "num_input_tokens_seen": 11787456, + "step": 7740 + }, + { + "epoch": 0.24790346328660137, + "grad_norm": 0.7034630179405212, + "learning_rate": 4.5516770028803954e-05, + "loss": 0.5788, + "num_input_tokens_seen": 11795056, + "step": 7745 + }, + { + "epoch": 0.2480635042570898, + "grad_norm": 1.1243295669555664, + "learning_rate": 4.5511158727913116e-05, + "loss": 0.7849, + "num_input_tokens_seen": 11802720, + "step": 7750 + }, + { + "epoch": 0.24822354522757825, + "grad_norm": 1.3726931810379028, + "learning_rate": 4.5505544263943856e-05, + "loss": 0.6532, + "num_input_tokens_seen": 11810928, + "step": 7755 + }, + { + "epoch": 0.2483835861980667, + "grad_norm": 0.8018686771392822, + "learning_rate": 4.549992663776197e-05, + "loss": 0.4327, + "num_input_tokens_seen": 11818432, + "step": 7760 + }, + { + "epoch": 0.24854362716855516, + "grad_norm": 1.3945149183273315, + "learning_rate": 4.5494305850233786e-05, + "loss": 0.7961, + "num_input_tokens_seen": 11825584, + "step": 7765 + }, + { + "epoch": 0.2487036681390436, + "grad_norm": 0.7019472718238831, + "learning_rate": 4.5488681902226094e-05, + "loss": 0.5652, + "num_input_tokens_seen": 11833360, + "step": 7770 + }, + { + "epoch": 0.24886370910953204, + "grad_norm": 0.46597912907600403, + "learning_rate": 4.5483054794606174e-05, + "loss": 0.6858, + "num_input_tokens_seen": 11841728, + "step": 7775 + }, + { + "epoch": 0.24902375008002048, + "grad_norm": 0.8962759971618652, + "learning_rate": 4.547742452824179e-05, + "loss": 0.4083, + "num_input_tokens_seen": 11848960, + "step": 7780 + }, + { + "epoch": 0.24918379105050892, + "grad_norm": 0.3523496091365814, + "learning_rate": 4.5471791104001215e-05, + "loss": 0.4588, + "num_input_tokens_seen": 11856560, + "step": 7785 + }, + { + "epoch": 0.24934383202099739, + "grad_norm": 0.6096436381340027, + "learning_rate": 4.546615452275319e-05, + "loss": 0.5731, + "num_input_tokens_seen": 11864256, + "step": 7790 + }, + { + "epoch": 0.24950387299148583, + "grad_norm": 0.8625327348709106, + "learning_rate": 4.5460514785366944e-05, + "loss": 0.6897, + "num_input_tokens_seen": 11871552, + "step": 7795 + }, + { + "epoch": 0.24966391396197427, + "grad_norm": 0.7565982937812805, + "learning_rate": 4.545487189271219e-05, + "loss": 0.5091, + "num_input_tokens_seen": 11879344, + "step": 7800 + }, + { + "epoch": 0.24966391396197427, + "eval_loss": 0.5822526812553406, + "eval_runtime": 558.183, + "eval_samples_per_second": 24.877, + "eval_steps_per_second": 12.439, + "num_input_tokens_seen": 11879344, + "step": 7800 + }, + { + "epoch": 0.2498239549324627, + "grad_norm": 1.5681618452072144, + "learning_rate": 4.544922584565914e-05, + "loss": 0.6334, + "num_input_tokens_seen": 11886496, + "step": 7805 + }, + { + "epoch": 0.24998399590295114, + "grad_norm": 1.0839897394180298, + "learning_rate": 4.544357664507848e-05, + "loss": 0.4004, + "num_input_tokens_seen": 11894288, + "step": 7810 + }, + { + "epoch": 0.2501440368734396, + "grad_norm": 0.5368139147758484, + "learning_rate": 4.54379242918414e-05, + "loss": 0.4267, + "num_input_tokens_seen": 11902304, + "step": 7815 + }, + { + "epoch": 0.25030407784392805, + "grad_norm": 1.0167032480239868, + "learning_rate": 4.543226878681955e-05, + "loss": 0.6976, + "num_input_tokens_seen": 11909728, + "step": 7820 + }, + { + "epoch": 0.2504641188144165, + "grad_norm": 0.7250860333442688, + "learning_rate": 4.5426610130885087e-05, + "loss": 0.3703, + "num_input_tokens_seen": 11916848, + "step": 7825 + }, + { + "epoch": 0.25062415978490493, + "grad_norm": 0.8458815813064575, + "learning_rate": 4.542094832491064e-05, + "loss": 0.6313, + "num_input_tokens_seen": 11924432, + "step": 7830 + }, + { + "epoch": 0.25078420075539337, + "grad_norm": 0.43027395009994507, + "learning_rate": 4.541528336976934e-05, + "loss": 0.5456, + "num_input_tokens_seen": 11931856, + "step": 7835 + }, + { + "epoch": 0.2509442417258818, + "grad_norm": 0.48603370785713196, + "learning_rate": 4.540961526633479e-05, + "loss": 0.4587, + "num_input_tokens_seen": 11939680, + "step": 7840 + }, + { + "epoch": 0.25110428269637025, + "grad_norm": 1.2126914262771606, + "learning_rate": 4.540394401548108e-05, + "loss": 0.54, + "num_input_tokens_seen": 11947312, + "step": 7845 + }, + { + "epoch": 0.2512643236668587, + "grad_norm": 0.9912428259849548, + "learning_rate": 4.539826961808279e-05, + "loss": 0.7176, + "num_input_tokens_seen": 11954976, + "step": 7850 + }, + { + "epoch": 0.2514243646373472, + "grad_norm": 1.6131412982940674, + "learning_rate": 4.5392592075014994e-05, + "loss": 0.6902, + "num_input_tokens_seen": 11963056, + "step": 7855 + }, + { + "epoch": 0.2515844056078356, + "grad_norm": 0.41015902161598206, + "learning_rate": 4.538691138715322e-05, + "loss": 0.4589, + "num_input_tokens_seen": 11970672, + "step": 7860 + }, + { + "epoch": 0.25174444657832407, + "grad_norm": 0.6491456627845764, + "learning_rate": 4.5381227555373516e-05, + "loss": 0.4755, + "num_input_tokens_seen": 11978144, + "step": 7865 + }, + { + "epoch": 0.2519044875488125, + "grad_norm": 0.9429036378860474, + "learning_rate": 4.537554058055239e-05, + "loss": 0.5336, + "num_input_tokens_seen": 11985776, + "step": 7870 + }, + { + "epoch": 0.25206452851930095, + "grad_norm": 0.7056849002838135, + "learning_rate": 4.5369850463566865e-05, + "loss": 0.6329, + "num_input_tokens_seen": 11993216, + "step": 7875 + }, + { + "epoch": 0.2522245694897894, + "grad_norm": 0.9282968044281006, + "learning_rate": 4.5364157205294404e-05, + "loss": 0.58, + "num_input_tokens_seen": 12000768, + "step": 7880 + }, + { + "epoch": 0.2523846104602778, + "grad_norm": 0.6812275052070618, + "learning_rate": 4.5358460806612996e-05, + "loss": 0.7519, + "num_input_tokens_seen": 12008912, + "step": 7885 + }, + { + "epoch": 0.25254465143076626, + "grad_norm": 0.44572713971138, + "learning_rate": 4.535276126840109e-05, + "loss": 0.555, + "num_input_tokens_seen": 12016640, + "step": 7890 + }, + { + "epoch": 0.2527046924012547, + "grad_norm": 0.7012948989868164, + "learning_rate": 4.5347058591537626e-05, + "loss": 0.8187, + "num_input_tokens_seen": 12024192, + "step": 7895 + }, + { + "epoch": 0.25286473337174314, + "grad_norm": 1.1499890089035034, + "learning_rate": 4.534135277690203e-05, + "loss": 0.5691, + "num_input_tokens_seen": 12032592, + "step": 7900 + }, + { + "epoch": 0.2530247743422316, + "grad_norm": 0.6975510120391846, + "learning_rate": 4.533564382537421e-05, + "loss": 0.4531, + "num_input_tokens_seen": 12039728, + "step": 7905 + }, + { + "epoch": 0.2531848153127201, + "grad_norm": 0.7239655256271362, + "learning_rate": 4.532993173783456e-05, + "loss": 0.4693, + "num_input_tokens_seen": 12047200, + "step": 7910 + }, + { + "epoch": 0.2533448562832085, + "grad_norm": 0.39928388595581055, + "learning_rate": 4.5324216515163954e-05, + "loss": 0.4047, + "num_input_tokens_seen": 12054848, + "step": 7915 + }, + { + "epoch": 0.25350489725369696, + "grad_norm": 0.8353431820869446, + "learning_rate": 4.531849815824375e-05, + "loss": 0.5829, + "num_input_tokens_seen": 12062720, + "step": 7920 + }, + { + "epoch": 0.2536649382241854, + "grad_norm": 0.9775884747505188, + "learning_rate": 4.5312776667955795e-05, + "loss": 0.5827, + "num_input_tokens_seen": 12070272, + "step": 7925 + }, + { + "epoch": 0.25382497919467384, + "grad_norm": 1.166042447090149, + "learning_rate": 4.5307052045182405e-05, + "loss": 0.6504, + "num_input_tokens_seen": 12077536, + "step": 7930 + }, + { + "epoch": 0.2539850201651623, + "grad_norm": 1.5319970846176147, + "learning_rate": 4.53013242908064e-05, + "loss": 0.4904, + "num_input_tokens_seen": 12084848, + "step": 7935 + }, + { + "epoch": 0.2541450611356507, + "grad_norm": 0.9523910880088806, + "learning_rate": 4.529559340571107e-05, + "loss": 0.4647, + "num_input_tokens_seen": 12091936, + "step": 7940 + }, + { + "epoch": 0.25430510210613916, + "grad_norm": 1.4283541440963745, + "learning_rate": 4.528985939078018e-05, + "loss": 0.7521, + "num_input_tokens_seen": 12098944, + "step": 7945 + }, + { + "epoch": 0.2544651430766276, + "grad_norm": 1.0261784791946411, + "learning_rate": 4.5284122246898e-05, + "loss": 0.6464, + "num_input_tokens_seen": 12107088, + "step": 7950 + }, + { + "epoch": 0.25462518404711604, + "grad_norm": 0.8268322348594666, + "learning_rate": 4.527838197494926e-05, + "loss": 0.4933, + "num_input_tokens_seen": 12114784, + "step": 7955 + }, + { + "epoch": 0.25478522501760453, + "grad_norm": 1.0799767971038818, + "learning_rate": 4.527263857581918e-05, + "loss": 0.6359, + "num_input_tokens_seen": 12122416, + "step": 7960 + }, + { + "epoch": 0.254945265988093, + "grad_norm": 0.7917869091033936, + "learning_rate": 4.526689205039347e-05, + "loss": 0.6362, + "num_input_tokens_seen": 12129648, + "step": 7965 + }, + { + "epoch": 0.2551053069585814, + "grad_norm": 1.0684185028076172, + "learning_rate": 4.5261142399558324e-05, + "loss": 0.3785, + "num_input_tokens_seen": 12137056, + "step": 7970 + }, + { + "epoch": 0.25526534792906985, + "grad_norm": 0.6616789102554321, + "learning_rate": 4.525538962420041e-05, + "loss": 0.5163, + "num_input_tokens_seen": 12144784, + "step": 7975 + }, + { + "epoch": 0.2554253888995583, + "grad_norm": 1.0208892822265625, + "learning_rate": 4.524963372520685e-05, + "loss": 0.5449, + "num_input_tokens_seen": 12152352, + "step": 7980 + }, + { + "epoch": 0.25558542987004673, + "grad_norm": 0.574067234992981, + "learning_rate": 4.524387470346531e-05, + "loss": 0.5206, + "num_input_tokens_seen": 12159824, + "step": 7985 + }, + { + "epoch": 0.25574547084053517, + "grad_norm": 1.1225420236587524, + "learning_rate": 4.5238112559863885e-05, + "loss": 0.6036, + "num_input_tokens_seen": 12167504, + "step": 7990 + }, + { + "epoch": 0.2559055118110236, + "grad_norm": 0.8730553984642029, + "learning_rate": 4.5232347295291175e-05, + "loss": 0.5009, + "num_input_tokens_seen": 12174912, + "step": 7995 + }, + { + "epoch": 0.25606555278151205, + "grad_norm": 1.1253060102462769, + "learning_rate": 4.522657891063626e-05, + "loss": 0.7383, + "num_input_tokens_seen": 12182400, + "step": 8000 + }, + { + "epoch": 0.25606555278151205, + "eval_loss": 0.5800753235816956, + "eval_runtime": 565.4287, + "eval_samples_per_second": 24.558, + "eval_steps_per_second": 12.279, + "num_input_tokens_seen": 12182400, + "step": 8000 + }, + { + "epoch": 0.2562255937520005, + "grad_norm": 0.8641526103019714, + "learning_rate": 4.52208074067887e-05, + "loss": 0.4449, + "num_input_tokens_seen": 12190160, + "step": 8005 + }, + { + "epoch": 0.25638563472248893, + "grad_norm": 0.6658448576927185, + "learning_rate": 4.5215032784638516e-05, + "loss": 0.435, + "num_input_tokens_seen": 12197584, + "step": 8010 + }, + { + "epoch": 0.2565456756929774, + "grad_norm": 0.8667029738426208, + "learning_rate": 4.5209255045076245e-05, + "loss": 0.6033, + "num_input_tokens_seen": 12205072, + "step": 8015 + }, + { + "epoch": 0.25670571666346587, + "grad_norm": 0.5872107744216919, + "learning_rate": 4.5203474188992875e-05, + "loss": 0.4931, + "num_input_tokens_seen": 12212752, + "step": 8020 + }, + { + "epoch": 0.2568657576339543, + "grad_norm": 0.7862933278083801, + "learning_rate": 4.51976902172799e-05, + "loss": 0.4758, + "num_input_tokens_seen": 12220160, + "step": 8025 + }, + { + "epoch": 0.25702579860444275, + "grad_norm": 1.2240885496139526, + "learning_rate": 4.519190313082927e-05, + "loss": 0.6871, + "num_input_tokens_seen": 12227632, + "step": 8030 + }, + { + "epoch": 0.2571858395749312, + "grad_norm": 0.5240903496742249, + "learning_rate": 4.518611293053343e-05, + "loss": 0.4039, + "num_input_tokens_seen": 12235456, + "step": 8035 + }, + { + "epoch": 0.2573458805454196, + "grad_norm": 1.077811598777771, + "learning_rate": 4.51803196172853e-05, + "loss": 0.4905, + "num_input_tokens_seen": 12243040, + "step": 8040 + }, + { + "epoch": 0.25750592151590807, + "grad_norm": 1.125274419784546, + "learning_rate": 4.517452319197828e-05, + "loss": 0.5104, + "num_input_tokens_seen": 12250912, + "step": 8045 + }, + { + "epoch": 0.2576659624863965, + "grad_norm": 0.809076189994812, + "learning_rate": 4.5168723655506265e-05, + "loss": 0.707, + "num_input_tokens_seen": 12258592, + "step": 8050 + }, + { + "epoch": 0.25782600345688494, + "grad_norm": 1.0810153484344482, + "learning_rate": 4.51629210087636e-05, + "loss": 0.661, + "num_input_tokens_seen": 12266096, + "step": 8055 + }, + { + "epoch": 0.2579860444273734, + "grad_norm": 1.3949825763702393, + "learning_rate": 4.515711525264513e-05, + "loss": 0.3569, + "num_input_tokens_seen": 12274048, + "step": 8060 + }, + { + "epoch": 0.2581460853978619, + "grad_norm": 0.7202088832855225, + "learning_rate": 4.5151306388046175e-05, + "loss": 0.6742, + "num_input_tokens_seen": 12281536, + "step": 8065 + }, + { + "epoch": 0.2583061263683503, + "grad_norm": 0.5660973787307739, + "learning_rate": 4.514549441586255e-05, + "loss": 0.4522, + "num_input_tokens_seen": 12289408, + "step": 8070 + }, + { + "epoch": 0.25846616733883876, + "grad_norm": 0.687716543674469, + "learning_rate": 4.513967933699051e-05, + "loss": 0.5859, + "num_input_tokens_seen": 12297008, + "step": 8075 + }, + { + "epoch": 0.2586262083093272, + "grad_norm": 0.8817722797393799, + "learning_rate": 4.513386115232684e-05, + "loss": 0.4955, + "num_input_tokens_seen": 12304768, + "step": 8080 + }, + { + "epoch": 0.25878624927981564, + "grad_norm": 0.8443325757980347, + "learning_rate": 4.5128039862768745e-05, + "loss": 0.7881, + "num_input_tokens_seen": 12312112, + "step": 8085 + }, + { + "epoch": 0.2589462902503041, + "grad_norm": 0.47000652551651, + "learning_rate": 4.512221546921397e-05, + "loss": 0.5113, + "num_input_tokens_seen": 12319760, + "step": 8090 + }, + { + "epoch": 0.2591063312207925, + "grad_norm": 1.1888982057571411, + "learning_rate": 4.5116387972560694e-05, + "loss": 0.4608, + "num_input_tokens_seen": 12327312, + "step": 8095 + }, + { + "epoch": 0.25926637219128096, + "grad_norm": 0.7963992953300476, + "learning_rate": 4.511055737370759e-05, + "loss": 0.3995, + "num_input_tokens_seen": 12334560, + "step": 8100 + }, + { + "epoch": 0.2594264131617694, + "grad_norm": 0.6334463357925415, + "learning_rate": 4.510472367355383e-05, + "loss": 0.458, + "num_input_tokens_seen": 12342112, + "step": 8105 + }, + { + "epoch": 0.25958645413225784, + "grad_norm": 1.7499722242355347, + "learning_rate": 4.509888687299901e-05, + "loss": 0.6823, + "num_input_tokens_seen": 12349456, + "step": 8110 + }, + { + "epoch": 0.2597464951027463, + "grad_norm": 0.7856096029281616, + "learning_rate": 4.5093046972943266e-05, + "loss": 0.4224, + "num_input_tokens_seen": 12357776, + "step": 8115 + }, + { + "epoch": 0.2599065360732348, + "grad_norm": 0.7117695212364197, + "learning_rate": 4.508720397428717e-05, + "loss": 0.3778, + "num_input_tokens_seen": 12365296, + "step": 8120 + }, + { + "epoch": 0.2600665770437232, + "grad_norm": 0.9492992758750916, + "learning_rate": 4.508135787793178e-05, + "loss": 0.6402, + "num_input_tokens_seen": 12372752, + "step": 8125 + }, + { + "epoch": 0.26022661801421165, + "grad_norm": 0.7829934358596802, + "learning_rate": 4.5075508684778664e-05, + "loss": 0.6437, + "num_input_tokens_seen": 12380608, + "step": 8130 + }, + { + "epoch": 0.2603866589847001, + "grad_norm": 1.0993305444717407, + "learning_rate": 4.506965639572982e-05, + "loss": 0.6432, + "num_input_tokens_seen": 12388432, + "step": 8135 + }, + { + "epoch": 0.26054669995518853, + "grad_norm": 0.7487428188323975, + "learning_rate": 4.506380101168774e-05, + "loss": 0.5508, + "num_input_tokens_seen": 12395680, + "step": 8140 + }, + { + "epoch": 0.26070674092567697, + "grad_norm": 1.1936514377593994, + "learning_rate": 4.505794253355542e-05, + "loss": 0.3809, + "num_input_tokens_seen": 12403616, + "step": 8145 + }, + { + "epoch": 0.2608667818961654, + "grad_norm": 0.7402029037475586, + "learning_rate": 4.5052080962236286e-05, + "loss": 0.6738, + "num_input_tokens_seen": 12411072, + "step": 8150 + }, + { + "epoch": 0.26102682286665385, + "grad_norm": 1.0958597660064697, + "learning_rate": 4.504621629863428e-05, + "loss": 0.546, + "num_input_tokens_seen": 12418256, + "step": 8155 + }, + { + "epoch": 0.2611868638371423, + "grad_norm": 0.5441081523895264, + "learning_rate": 4.504034854365381e-05, + "loss": 0.5378, + "num_input_tokens_seen": 12425632, + "step": 8160 + }, + { + "epoch": 0.26134690480763073, + "grad_norm": 0.7940401434898376, + "learning_rate": 4.503447769819974e-05, + "loss": 0.5043, + "num_input_tokens_seen": 12432736, + "step": 8165 + }, + { + "epoch": 0.26150694577811917, + "grad_norm": 0.8621566891670227, + "learning_rate": 4.502860376317745e-05, + "loss": 0.5045, + "num_input_tokens_seen": 12440464, + "step": 8170 + }, + { + "epoch": 0.26166698674860767, + "grad_norm": 0.8728803396224976, + "learning_rate": 4.502272673949276e-05, + "loss": 0.5089, + "num_input_tokens_seen": 12447792, + "step": 8175 + }, + { + "epoch": 0.2618270277190961, + "grad_norm": 0.7877692580223083, + "learning_rate": 4.501684662805199e-05, + "loss": 0.6157, + "num_input_tokens_seen": 12454848, + "step": 8180 + }, + { + "epoch": 0.26198706868958455, + "grad_norm": 0.8561547994613647, + "learning_rate": 4.5010963429761924e-05, + "loss": 0.5515, + "num_input_tokens_seen": 12462384, + "step": 8185 + }, + { + "epoch": 0.262147109660073, + "grad_norm": 0.6085007786750793, + "learning_rate": 4.500507714552982e-05, + "loss": 0.5338, + "num_input_tokens_seen": 12470288, + "step": 8190 + }, + { + "epoch": 0.2623071506305614, + "grad_norm": 0.6689333319664001, + "learning_rate": 4.499918777626342e-05, + "loss": 0.569, + "num_input_tokens_seen": 12477424, + "step": 8195 + }, + { + "epoch": 0.26246719160104987, + "grad_norm": 0.4274224638938904, + "learning_rate": 4.499329532287093e-05, + "loss": 0.4138, + "num_input_tokens_seen": 12484784, + "step": 8200 + }, + { + "epoch": 0.26246719160104987, + "eval_loss": 0.5771757364273071, + "eval_runtime": 559.8373, + "eval_samples_per_second": 24.804, + "eval_steps_per_second": 12.402, + "num_input_tokens_seen": 12484784, + "step": 8200 + }, + { + "epoch": 0.2626272325715383, + "grad_norm": 1.1063411235809326, + "learning_rate": 4.4987399786261064e-05, + "loss": 0.6642, + "num_input_tokens_seen": 12492144, + "step": 8205 + }, + { + "epoch": 0.26278727354202674, + "grad_norm": 0.8674256205558777, + "learning_rate": 4.498150116734297e-05, + "loss": 0.5905, + "num_input_tokens_seen": 12499104, + "step": 8210 + }, + { + "epoch": 0.2629473145125152, + "grad_norm": 3.019615650177002, + "learning_rate": 4.4975599467026294e-05, + "loss": 0.5901, + "num_input_tokens_seen": 12506944, + "step": 8215 + }, + { + "epoch": 0.2631073554830036, + "grad_norm": 0.8276799917221069, + "learning_rate": 4.496969468622114e-05, + "loss": 0.4616, + "num_input_tokens_seen": 12514224, + "step": 8220 + }, + { + "epoch": 0.2632673964534921, + "grad_norm": 0.6250467300415039, + "learning_rate": 4.496378682583813e-05, + "loss": 0.4705, + "num_input_tokens_seen": 12521632, + "step": 8225 + }, + { + "epoch": 0.26342743742398056, + "grad_norm": 0.6577309370040894, + "learning_rate": 4.495787588678829e-05, + "loss": 0.4954, + "num_input_tokens_seen": 12529248, + "step": 8230 + }, + { + "epoch": 0.263587478394469, + "grad_norm": 0.5690023899078369, + "learning_rate": 4.4951961869983196e-05, + "loss": 0.4973, + "num_input_tokens_seen": 12536704, + "step": 8235 + }, + { + "epoch": 0.26374751936495744, + "grad_norm": 0.5642849206924438, + "learning_rate": 4.494604477633485e-05, + "loss": 0.509, + "num_input_tokens_seen": 12544192, + "step": 8240 + }, + { + "epoch": 0.2639075603354459, + "grad_norm": 1.0393099784851074, + "learning_rate": 4.4940124606755734e-05, + "loss": 0.6096, + "num_input_tokens_seen": 12552096, + "step": 8245 + }, + { + "epoch": 0.2640676013059343, + "grad_norm": 0.4451197683811188, + "learning_rate": 4.493420136215882e-05, + "loss": 0.6469, + "num_input_tokens_seen": 12559680, + "step": 8250 + }, + { + "epoch": 0.26422764227642276, + "grad_norm": 1.1914526224136353, + "learning_rate": 4.492827504345756e-05, + "loss": 0.55, + "num_input_tokens_seen": 12567824, + "step": 8255 + }, + { + "epoch": 0.2643876832469112, + "grad_norm": 0.40392744541168213, + "learning_rate": 4.492234565156584e-05, + "loss": 0.4417, + "num_input_tokens_seen": 12575104, + "step": 8260 + }, + { + "epoch": 0.26454772421739964, + "grad_norm": 0.5877530574798584, + "learning_rate": 4.491641318739807e-05, + "loss": 0.4687, + "num_input_tokens_seen": 12582752, + "step": 8265 + }, + { + "epoch": 0.2647077651878881, + "grad_norm": 0.6423861384391785, + "learning_rate": 4.4910477651869096e-05, + "loss": 0.4509, + "num_input_tokens_seen": 12589936, + "step": 8270 + }, + { + "epoch": 0.2648678061583765, + "grad_norm": 0.6637521982192993, + "learning_rate": 4.4904539045894254e-05, + "loss": 0.452, + "num_input_tokens_seen": 12597456, + "step": 8275 + }, + { + "epoch": 0.265027847128865, + "grad_norm": 0.7475348114967346, + "learning_rate": 4.4898597370389364e-05, + "loss": 0.4777, + "num_input_tokens_seen": 12605456, + "step": 8280 + }, + { + "epoch": 0.26518788809935345, + "grad_norm": 0.8799005746841431, + "learning_rate": 4.489265262627069e-05, + "loss": 0.4478, + "num_input_tokens_seen": 12612608, + "step": 8285 + }, + { + "epoch": 0.2653479290698419, + "grad_norm": 0.9167254567146301, + "learning_rate": 4.488670481445499e-05, + "loss": 0.5237, + "num_input_tokens_seen": 12620208, + "step": 8290 + }, + { + "epoch": 0.26550797004033033, + "grad_norm": 1.702927827835083, + "learning_rate": 4.488075393585951e-05, + "loss": 0.7116, + "num_input_tokens_seen": 12627760, + "step": 8295 + }, + { + "epoch": 0.26566801101081877, + "grad_norm": 0.5988535284996033, + "learning_rate": 4.487479999140193e-05, + "loss": 0.4469, + "num_input_tokens_seen": 12635728, + "step": 8300 + }, + { + "epoch": 0.2658280519813072, + "grad_norm": 0.5595522522926331, + "learning_rate": 4.4868842982000425e-05, + "loss": 0.5059, + "num_input_tokens_seen": 12642864, + "step": 8305 + }, + { + "epoch": 0.26598809295179565, + "grad_norm": 1.1601983308792114, + "learning_rate": 4.486288290857365e-05, + "loss": 0.6924, + "num_input_tokens_seen": 12650352, + "step": 8310 + }, + { + "epoch": 0.2661481339222841, + "grad_norm": 0.43594086170196533, + "learning_rate": 4.4856919772040715e-05, + "loss": 0.3608, + "num_input_tokens_seen": 12657776, + "step": 8315 + }, + { + "epoch": 0.26630817489277253, + "grad_norm": 0.8270493745803833, + "learning_rate": 4.485095357332122e-05, + "loss": 0.6025, + "num_input_tokens_seen": 12665120, + "step": 8320 + }, + { + "epoch": 0.26646821586326097, + "grad_norm": 1.4563571214675903, + "learning_rate": 4.484498431333521e-05, + "loss": 0.6651, + "num_input_tokens_seen": 12672944, + "step": 8325 + }, + { + "epoch": 0.26662825683374947, + "grad_norm": 0.9445053935050964, + "learning_rate": 4.4839011993003245e-05, + "loss": 0.5391, + "num_input_tokens_seen": 12680736, + "step": 8330 + }, + { + "epoch": 0.2667882978042379, + "grad_norm": 1.3226991891860962, + "learning_rate": 4.4833036613246305e-05, + "loss": 0.5704, + "num_input_tokens_seen": 12688576, + "step": 8335 + }, + { + "epoch": 0.26694833877472635, + "grad_norm": 0.5745265483856201, + "learning_rate": 4.482705817498589e-05, + "loss": 0.4425, + "num_input_tokens_seen": 12696000, + "step": 8340 + }, + { + "epoch": 0.2671083797452148, + "grad_norm": 1.0061084032058716, + "learning_rate": 4.4821076679143934e-05, + "loss": 0.4493, + "num_input_tokens_seen": 12703408, + "step": 8345 + }, + { + "epoch": 0.2672684207157032, + "grad_norm": 0.5987309217453003, + "learning_rate": 4.481509212664288e-05, + "loss": 0.3898, + "num_input_tokens_seen": 12712336, + "step": 8350 + }, + { + "epoch": 0.26742846168619167, + "grad_norm": 1.5281832218170166, + "learning_rate": 4.480910451840559e-05, + "loss": 0.7227, + "num_input_tokens_seen": 12719728, + "step": 8355 + }, + { + "epoch": 0.2675885026566801, + "grad_norm": 0.8607586026191711, + "learning_rate": 4.480311385535546e-05, + "loss": 0.6812, + "num_input_tokens_seen": 12727344, + "step": 8360 + }, + { + "epoch": 0.26774854362716854, + "grad_norm": 0.9321521520614624, + "learning_rate": 4.47971201384163e-05, + "loss": 0.5241, + "num_input_tokens_seen": 12734592, + "step": 8365 + }, + { + "epoch": 0.267908584597657, + "grad_norm": 1.1461739540100098, + "learning_rate": 4.4791123368512446e-05, + "loss": 0.493, + "num_input_tokens_seen": 12742000, + "step": 8370 + }, + { + "epoch": 0.2680686255681454, + "grad_norm": 0.9727409482002258, + "learning_rate": 4.478512354656864e-05, + "loss": 0.7127, + "num_input_tokens_seen": 12749200, + "step": 8375 + }, + { + "epoch": 0.26822866653863386, + "grad_norm": 0.873784601688385, + "learning_rate": 4.477912067351016e-05, + "loss": 0.5398, + "num_input_tokens_seen": 12756880, + "step": 8380 + }, + { + "epoch": 0.26838870750912236, + "grad_norm": 1.1484571695327759, + "learning_rate": 4.477311475026271e-05, + "loss": 0.6627, + "num_input_tokens_seen": 12764064, + "step": 8385 + }, + { + "epoch": 0.2685487484796108, + "grad_norm": 0.6338135004043579, + "learning_rate": 4.476710577775248e-05, + "loss": 0.42, + "num_input_tokens_seen": 12771856, + "step": 8390 + }, + { + "epoch": 0.26870878945009924, + "grad_norm": 0.9122999906539917, + "learning_rate": 4.476109375690612e-05, + "loss": 0.5623, + "num_input_tokens_seen": 12779296, + "step": 8395 + }, + { + "epoch": 0.2688688304205877, + "grad_norm": 0.8229836821556091, + "learning_rate": 4.4755078688650784e-05, + "loss": 0.5738, + "num_input_tokens_seen": 12786640, + "step": 8400 + }, + { + "epoch": 0.2688688304205877, + "eval_loss": 0.5762130618095398, + "eval_runtime": 560.1093, + "eval_samples_per_second": 24.792, + "eval_steps_per_second": 12.396, + "num_input_tokens_seen": 12786640, + "step": 8400 + }, + { + "epoch": 0.2690288713910761, + "grad_norm": 0.9646776914596558, + "learning_rate": 4.474906057391406e-05, + "loss": 0.4553, + "num_input_tokens_seen": 12794352, + "step": 8405 + }, + { + "epoch": 0.26918891236156456, + "grad_norm": 1.0871381759643555, + "learning_rate": 4.4743039413624e-05, + "loss": 0.4478, + "num_input_tokens_seen": 12802096, + "step": 8410 + }, + { + "epoch": 0.269348953332053, + "grad_norm": 1.9578900337219238, + "learning_rate": 4.473701520870916e-05, + "loss": 0.5418, + "num_input_tokens_seen": 12809248, + "step": 8415 + }, + { + "epoch": 0.26950899430254144, + "grad_norm": 0.8078241944313049, + "learning_rate": 4.4730987960098544e-05, + "loss": 0.4933, + "num_input_tokens_seen": 12816320, + "step": 8420 + }, + { + "epoch": 0.2696690352730299, + "grad_norm": 2.825792074203491, + "learning_rate": 4.4724957668721635e-05, + "loss": 0.7073, + "num_input_tokens_seen": 12823632, + "step": 8425 + }, + { + "epoch": 0.2698290762435183, + "grad_norm": 1.9016330242156982, + "learning_rate": 4.471892433550836e-05, + "loss": 0.5649, + "num_input_tokens_seen": 12831248, + "step": 8430 + }, + { + "epoch": 0.2699891172140068, + "grad_norm": 1.4931353330612183, + "learning_rate": 4.471288796138916e-05, + "loss": 0.5945, + "num_input_tokens_seen": 12838928, + "step": 8435 + }, + { + "epoch": 0.27014915818449525, + "grad_norm": 0.4753241240978241, + "learning_rate": 4.470684854729491e-05, + "loss": 0.5724, + "num_input_tokens_seen": 12846512, + "step": 8440 + }, + { + "epoch": 0.2703091991549837, + "grad_norm": 0.6860668659210205, + "learning_rate": 4.4700806094156955e-05, + "loss": 0.5103, + "num_input_tokens_seen": 12854720, + "step": 8445 + }, + { + "epoch": 0.27046924012547213, + "grad_norm": 0.5609471797943115, + "learning_rate": 4.469476060290713e-05, + "loss": 0.464, + "num_input_tokens_seen": 12862272, + "step": 8450 + }, + { + "epoch": 0.27062928109596057, + "grad_norm": 0.7803241014480591, + "learning_rate": 4.468871207447772e-05, + "loss": 0.4205, + "num_input_tokens_seen": 12869792, + "step": 8455 + }, + { + "epoch": 0.270789322066449, + "grad_norm": 1.1592233180999756, + "learning_rate": 4.4682660509801486e-05, + "loss": 0.5939, + "num_input_tokens_seen": 12877056, + "step": 8460 + }, + { + "epoch": 0.27094936303693745, + "grad_norm": 0.6525108218193054, + "learning_rate": 4.467660590981165e-05, + "loss": 0.5406, + "num_input_tokens_seen": 12884464, + "step": 8465 + }, + { + "epoch": 0.2711094040074259, + "grad_norm": 1.4751461744308472, + "learning_rate": 4.467054827544191e-05, + "loss": 0.5421, + "num_input_tokens_seen": 12891664, + "step": 8470 + }, + { + "epoch": 0.27126944497791433, + "grad_norm": 0.8943687081336975, + "learning_rate": 4.4664487607626434e-05, + "loss": 0.575, + "num_input_tokens_seen": 12899376, + "step": 8475 + }, + { + "epoch": 0.27142948594840277, + "grad_norm": 0.6453889608383179, + "learning_rate": 4.4658423907299845e-05, + "loss": 0.5045, + "num_input_tokens_seen": 12907008, + "step": 8480 + }, + { + "epoch": 0.2715895269188912, + "grad_norm": 1.9384816884994507, + "learning_rate": 4.465235717539725e-05, + "loss": 0.7137, + "num_input_tokens_seen": 12914336, + "step": 8485 + }, + { + "epoch": 0.2717495678893797, + "grad_norm": 0.8234109878540039, + "learning_rate": 4.464628741285421e-05, + "loss": 0.664, + "num_input_tokens_seen": 12921664, + "step": 8490 + }, + { + "epoch": 0.27190960885986815, + "grad_norm": 1.0984275341033936, + "learning_rate": 4.4640214620606754e-05, + "loss": 0.6045, + "num_input_tokens_seen": 12928736, + "step": 8495 + }, + { + "epoch": 0.2720696498303566, + "grad_norm": 0.7237162590026855, + "learning_rate": 4.46341387995914e-05, + "loss": 0.4638, + "num_input_tokens_seen": 12936400, + "step": 8500 + }, + { + "epoch": 0.272229690800845, + "grad_norm": 0.6783583164215088, + "learning_rate": 4.4628059950745106e-05, + "loss": 0.7255, + "num_input_tokens_seen": 12944352, + "step": 8505 + }, + { + "epoch": 0.27238973177133347, + "grad_norm": 0.41578856110572815, + "learning_rate": 4.4621978075005297e-05, + "loss": 0.3916, + "num_input_tokens_seen": 12952384, + "step": 8510 + }, + { + "epoch": 0.2725497727418219, + "grad_norm": 1.2311104536056519, + "learning_rate": 4.461589317330989e-05, + "loss": 0.5037, + "num_input_tokens_seen": 12960704, + "step": 8515 + }, + { + "epoch": 0.27270981371231034, + "grad_norm": 1.267075777053833, + "learning_rate": 4.460980524659724e-05, + "loss": 0.5376, + "num_input_tokens_seen": 12968336, + "step": 8520 + }, + { + "epoch": 0.2728698546827988, + "grad_norm": 0.799390971660614, + "learning_rate": 4.46037142958062e-05, + "loss": 0.6529, + "num_input_tokens_seen": 12975888, + "step": 8525 + }, + { + "epoch": 0.2730298956532872, + "grad_norm": 1.083345890045166, + "learning_rate": 4.4597620321876046e-05, + "loss": 0.6276, + "num_input_tokens_seen": 12983456, + "step": 8530 + }, + { + "epoch": 0.27318993662377566, + "grad_norm": 1.3910913467407227, + "learning_rate": 4.459152332574656e-05, + "loss": 0.6327, + "num_input_tokens_seen": 12991008, + "step": 8535 + }, + { + "epoch": 0.2733499775942641, + "grad_norm": 0.79203861951828, + "learning_rate": 4.4585423308357985e-05, + "loss": 0.5672, + "num_input_tokens_seen": 12998336, + "step": 8540 + }, + { + "epoch": 0.2735100185647526, + "grad_norm": 0.7776643633842468, + "learning_rate": 4.457932027065102e-05, + "loss": 0.4482, + "num_input_tokens_seen": 13006160, + "step": 8545 + }, + { + "epoch": 0.27367005953524104, + "grad_norm": 0.8966943025588989, + "learning_rate": 4.45732142135668e-05, + "loss": 0.5665, + "num_input_tokens_seen": 13014048, + "step": 8550 + }, + { + "epoch": 0.2738301005057295, + "grad_norm": 0.9222797751426697, + "learning_rate": 4.4567105138046986e-05, + "loss": 0.4864, + "num_input_tokens_seen": 13021616, + "step": 8555 + }, + { + "epoch": 0.2739901414762179, + "grad_norm": 1.2835613489151, + "learning_rate": 4.456099304503365e-05, + "loss": 0.4545, + "num_input_tokens_seen": 13029312, + "step": 8560 + }, + { + "epoch": 0.27415018244670636, + "grad_norm": 1.1428675651550293, + "learning_rate": 4.455487793546939e-05, + "loss": 0.5946, + "num_input_tokens_seen": 13036944, + "step": 8565 + }, + { + "epoch": 0.2743102234171948, + "grad_norm": 1.5462177991867065, + "learning_rate": 4.454875981029719e-05, + "loss": 0.5578, + "num_input_tokens_seen": 13044512, + "step": 8570 + }, + { + "epoch": 0.27447026438768324, + "grad_norm": 1.3482439517974854, + "learning_rate": 4.454263867046057e-05, + "loss": 0.422, + "num_input_tokens_seen": 13051648, + "step": 8575 + }, + { + "epoch": 0.2746303053581717, + "grad_norm": 0.874725341796875, + "learning_rate": 4.4536514516903484e-05, + "loss": 0.5045, + "num_input_tokens_seen": 13058672, + "step": 8580 + }, + { + "epoch": 0.2747903463286601, + "grad_norm": 0.6210857629776001, + "learning_rate": 4.453038735057034e-05, + "loss": 0.6446, + "num_input_tokens_seen": 13066064, + "step": 8585 + }, + { + "epoch": 0.27495038729914856, + "grad_norm": 1.0030909776687622, + "learning_rate": 4.4524257172406034e-05, + "loss": 0.4568, + "num_input_tokens_seen": 13073616, + "step": 8590 + }, + { + "epoch": 0.27511042826963705, + "grad_norm": 0.795568585395813, + "learning_rate": 4.451812398335592e-05, + "loss": 0.4046, + "num_input_tokens_seen": 13080832, + "step": 8595 + }, + { + "epoch": 0.2752704692401255, + "grad_norm": 0.7200952768325806, + "learning_rate": 4.4511987784365805e-05, + "loss": 0.6092, + "num_input_tokens_seen": 13088384, + "step": 8600 + }, + { + "epoch": 0.2752704692401255, + "eval_loss": 0.5741302371025085, + "eval_runtime": 557.9168, + "eval_samples_per_second": 24.889, + "eval_steps_per_second": 12.445, + "num_input_tokens_seen": 13088384, + "step": 8600 + }, + { + "epoch": 0.27543051021061393, + "grad_norm": 0.6262813806533813, + "learning_rate": 4.450584857638197e-05, + "loss": 0.5318, + "num_input_tokens_seen": 13095536, + "step": 8605 + }, + { + "epoch": 0.2755905511811024, + "grad_norm": 0.8745664954185486, + "learning_rate": 4.449970636035116e-05, + "loss": 0.4908, + "num_input_tokens_seen": 13103104, + "step": 8610 + }, + { + "epoch": 0.2757505921515908, + "grad_norm": 0.6957159638404846, + "learning_rate": 4.4493561137220574e-05, + "loss": 0.5096, + "num_input_tokens_seen": 13110976, + "step": 8615 + }, + { + "epoch": 0.27591063312207925, + "grad_norm": 0.5485116839408875, + "learning_rate": 4.44874129079379e-05, + "loss": 0.3885, + "num_input_tokens_seen": 13118480, + "step": 8620 + }, + { + "epoch": 0.2760706740925677, + "grad_norm": 0.8169776797294617, + "learning_rate": 4.4481261673451255e-05, + "loss": 0.5284, + "num_input_tokens_seen": 13126000, + "step": 8625 + }, + { + "epoch": 0.27623071506305613, + "grad_norm": 1.229607343673706, + "learning_rate": 4.4475107434709245e-05, + "loss": 0.5535, + "num_input_tokens_seen": 13133632, + "step": 8630 + }, + { + "epoch": 0.27639075603354457, + "grad_norm": 0.8783779740333557, + "learning_rate": 4.446895019266093e-05, + "loss": 0.541, + "num_input_tokens_seen": 13141184, + "step": 8635 + }, + { + "epoch": 0.276550797004033, + "grad_norm": 0.7364345788955688, + "learning_rate": 4.446278994825583e-05, + "loss": 0.7205, + "num_input_tokens_seen": 13148368, + "step": 8640 + }, + { + "epoch": 0.27671083797452145, + "grad_norm": 1.076692819595337, + "learning_rate": 4.445662670244394e-05, + "loss": 0.5805, + "num_input_tokens_seen": 13155968, + "step": 8645 + }, + { + "epoch": 0.27687087894500995, + "grad_norm": 1.0341124534606934, + "learning_rate": 4.44504604561757e-05, + "loss": 0.5313, + "num_input_tokens_seen": 13163792, + "step": 8650 + }, + { + "epoch": 0.2770309199154984, + "grad_norm": 0.5489898324012756, + "learning_rate": 4.4444291210402035e-05, + "loss": 0.4552, + "num_input_tokens_seen": 13170912, + "step": 8655 + }, + { + "epoch": 0.2771909608859868, + "grad_norm": 0.5914074778556824, + "learning_rate": 4.443811896607431e-05, + "loss": 0.6645, + "num_input_tokens_seen": 13178864, + "step": 8660 + }, + { + "epoch": 0.27735100185647527, + "grad_norm": 1.0648936033248901, + "learning_rate": 4.443194372414436e-05, + "loss": 0.4309, + "num_input_tokens_seen": 13186400, + "step": 8665 + }, + { + "epoch": 0.2775110428269637, + "grad_norm": 1.0045795440673828, + "learning_rate": 4.442576548556449e-05, + "loss": 0.4547, + "num_input_tokens_seen": 13194256, + "step": 8670 + }, + { + "epoch": 0.27767108379745215, + "grad_norm": 0.9157430529594421, + "learning_rate": 4.441958425128747e-05, + "loss": 0.7994, + "num_input_tokens_seen": 13202096, + "step": 8675 + }, + { + "epoch": 0.2778311247679406, + "grad_norm": 0.8569799065589905, + "learning_rate": 4.4413400022266515e-05, + "loss": 0.6716, + "num_input_tokens_seen": 13209968, + "step": 8680 + }, + { + "epoch": 0.277991165738429, + "grad_norm": 1.1717560291290283, + "learning_rate": 4.4407212799455313e-05, + "loss": 0.6623, + "num_input_tokens_seen": 13217920, + "step": 8685 + }, + { + "epoch": 0.27815120670891746, + "grad_norm": 1.0461314916610718, + "learning_rate": 4.4401022583808003e-05, + "loss": 0.478, + "num_input_tokens_seen": 13225536, + "step": 8690 + }, + { + "epoch": 0.2783112476794059, + "grad_norm": 0.7402077317237854, + "learning_rate": 4.439482937627921e-05, + "loss": 0.4252, + "num_input_tokens_seen": 13233232, + "step": 8695 + }, + { + "epoch": 0.2784712886498944, + "grad_norm": 1.8554868698120117, + "learning_rate": 4.4388633177824004e-05, + "loss": 0.5529, + "num_input_tokens_seen": 13240320, + "step": 8700 + }, + { + "epoch": 0.27863132962038284, + "grad_norm": 0.9104053378105164, + "learning_rate": 4.4382433989397895e-05, + "loss": 0.616, + "num_input_tokens_seen": 13247840, + "step": 8705 + }, + { + "epoch": 0.2787913705908713, + "grad_norm": 1.1045072078704834, + "learning_rate": 4.4376231811956895e-05, + "loss": 0.5995, + "num_input_tokens_seen": 13255648, + "step": 8710 + }, + { + "epoch": 0.2789514115613597, + "grad_norm": 0.8246548771858215, + "learning_rate": 4.437002664645745e-05, + "loss": 0.6106, + "num_input_tokens_seen": 13263584, + "step": 8715 + }, + { + "epoch": 0.27911145253184816, + "grad_norm": 1.6801934242248535, + "learning_rate": 4.436381849385649e-05, + "loss": 0.6123, + "num_input_tokens_seen": 13271152, + "step": 8720 + }, + { + "epoch": 0.2792714935023366, + "grad_norm": 1.180335521697998, + "learning_rate": 4.435760735511136e-05, + "loss": 0.5678, + "num_input_tokens_seen": 13279088, + "step": 8725 + }, + { + "epoch": 0.27943153447282504, + "grad_norm": 0.8517492413520813, + "learning_rate": 4.435139323117992e-05, + "loss": 0.556, + "num_input_tokens_seen": 13286608, + "step": 8730 + }, + { + "epoch": 0.2795915754433135, + "grad_norm": 1.4867802858352661, + "learning_rate": 4.434517612302046e-05, + "loss": 0.4553, + "num_input_tokens_seen": 13293952, + "step": 8735 + }, + { + "epoch": 0.2797516164138019, + "grad_norm": 0.44219380617141724, + "learning_rate": 4.433895603159174e-05, + "loss": 0.4727, + "num_input_tokens_seen": 13301504, + "step": 8740 + }, + { + "epoch": 0.27991165738429036, + "grad_norm": 0.7909645438194275, + "learning_rate": 4.433273295785296e-05, + "loss": 0.5053, + "num_input_tokens_seen": 13309168, + "step": 8745 + }, + { + "epoch": 0.2800716983547788, + "grad_norm": 0.5429316759109497, + "learning_rate": 4.432650690276382e-05, + "loss": 0.4043, + "num_input_tokens_seen": 13316960, + "step": 8750 + }, + { + "epoch": 0.2802317393252673, + "grad_norm": 1.1871819496154785, + "learning_rate": 4.4320277867284435e-05, + "loss": 0.4572, + "num_input_tokens_seen": 13325040, + "step": 8755 + }, + { + "epoch": 0.28039178029575573, + "grad_norm": 0.9656745195388794, + "learning_rate": 4.431404585237541e-05, + "loss": 0.6274, + "num_input_tokens_seen": 13332608, + "step": 8760 + }, + { + "epoch": 0.2805518212662442, + "grad_norm": 0.5171418190002441, + "learning_rate": 4.43078108589978e-05, + "loss": 0.504, + "num_input_tokens_seen": 13339952, + "step": 8765 + }, + { + "epoch": 0.2807118622367326, + "grad_norm": 0.6744235157966614, + "learning_rate": 4.4301572888113116e-05, + "loss": 0.551, + "num_input_tokens_seen": 13347536, + "step": 8770 + }, + { + "epoch": 0.28087190320722105, + "grad_norm": 0.6140366196632385, + "learning_rate": 4.4295331940683337e-05, + "loss": 0.498, + "num_input_tokens_seen": 13355424, + "step": 8775 + }, + { + "epoch": 0.2810319441777095, + "grad_norm": 0.6820389032363892, + "learning_rate": 4.428908801767089e-05, + "loss": 0.5232, + "num_input_tokens_seen": 13363408, + "step": 8780 + }, + { + "epoch": 0.28119198514819793, + "grad_norm": 0.9952068328857422, + "learning_rate": 4.428284112003868e-05, + "loss": 0.5819, + "num_input_tokens_seen": 13370608, + "step": 8785 + }, + { + "epoch": 0.28135202611868637, + "grad_norm": 0.7564281225204468, + "learning_rate": 4.4276591248750033e-05, + "loss": 0.5819, + "num_input_tokens_seen": 13378880, + "step": 8790 + }, + { + "epoch": 0.2815120670891748, + "grad_norm": 0.7071586847305298, + "learning_rate": 4.4270338404768774e-05, + "loss": 0.5355, + "num_input_tokens_seen": 13386992, + "step": 8795 + }, + { + "epoch": 0.28167210805966325, + "grad_norm": 1.222283124923706, + "learning_rate": 4.426408258905917e-05, + "loss": 0.626, + "num_input_tokens_seen": 13394624, + "step": 8800 + }, + { + "epoch": 0.28167210805966325, + "eval_loss": 0.5724745988845825, + "eval_runtime": 559.0042, + "eval_samples_per_second": 24.841, + "eval_steps_per_second": 12.42, + "num_input_tokens_seen": 13394624, + "step": 8800 + }, + { + "epoch": 0.28183214903015175, + "grad_norm": 0.6163823008537292, + "learning_rate": 4.425782380258594e-05, + "loss": 0.5315, + "num_input_tokens_seen": 13402112, + "step": 8805 + }, + { + "epoch": 0.2819921900006402, + "grad_norm": 0.973550021648407, + "learning_rate": 4.425156204631427e-05, + "loss": 0.5429, + "num_input_tokens_seen": 13409680, + "step": 8810 + }, + { + "epoch": 0.2821522309711286, + "grad_norm": 0.7415067553520203, + "learning_rate": 4.424529732120981e-05, + "loss": 0.5033, + "num_input_tokens_seen": 13417104, + "step": 8815 + }, + { + "epoch": 0.28231227194161707, + "grad_norm": 1.2485142946243286, + "learning_rate": 4.423902962823864e-05, + "loss": 0.8256, + "num_input_tokens_seen": 13424944, + "step": 8820 + }, + { + "epoch": 0.2824723129121055, + "grad_norm": 0.7119803428649902, + "learning_rate": 4.423275896836733e-05, + "loss": 0.4998, + "num_input_tokens_seen": 13433072, + "step": 8825 + }, + { + "epoch": 0.28263235388259395, + "grad_norm": 0.7286111116409302, + "learning_rate": 4.42264853425629e-05, + "loss": 0.5666, + "num_input_tokens_seen": 13440816, + "step": 8830 + }, + { + "epoch": 0.2827923948530824, + "grad_norm": 0.9478473663330078, + "learning_rate": 4.4220208751792816e-05, + "loss": 0.4386, + "num_input_tokens_seen": 13448704, + "step": 8835 + }, + { + "epoch": 0.2829524358235708, + "grad_norm": 0.94295734167099, + "learning_rate": 4.421392919702499e-05, + "loss": 0.5039, + "num_input_tokens_seen": 13456128, + "step": 8840 + }, + { + "epoch": 0.28311247679405926, + "grad_norm": 0.9946351647377014, + "learning_rate": 4.4207646679227846e-05, + "loss": 0.7324, + "num_input_tokens_seen": 13463840, + "step": 8845 + }, + { + "epoch": 0.2832725177645477, + "grad_norm": 1.3361215591430664, + "learning_rate": 4.42013611993702e-05, + "loss": 0.5578, + "num_input_tokens_seen": 13471088, + "step": 8850 + }, + { + "epoch": 0.28343255873503614, + "grad_norm": 0.36846745014190674, + "learning_rate": 4.419507275842135e-05, + "loss": 0.4757, + "num_input_tokens_seen": 13478496, + "step": 8855 + }, + { + "epoch": 0.28359259970552464, + "grad_norm": 1.297744631767273, + "learning_rate": 4.418878135735106e-05, + "loss": 0.7052, + "num_input_tokens_seen": 13486544, + "step": 8860 + }, + { + "epoch": 0.2837526406760131, + "grad_norm": 0.8965442180633545, + "learning_rate": 4.418248699712955e-05, + "loss": 0.6868, + "num_input_tokens_seen": 13493696, + "step": 8865 + }, + { + "epoch": 0.2839126816465015, + "grad_norm": 0.6514096856117249, + "learning_rate": 4.417618967872748e-05, + "loss": 0.56, + "num_input_tokens_seen": 13501008, + "step": 8870 + }, + { + "epoch": 0.28407272261698996, + "grad_norm": 1.532496690750122, + "learning_rate": 4.4169889403115985e-05, + "loss": 0.6033, + "num_input_tokens_seen": 13508672, + "step": 8875 + }, + { + "epoch": 0.2842327635874784, + "grad_norm": 0.7316017746925354, + "learning_rate": 4.4163586171266627e-05, + "loss": 0.5542, + "num_input_tokens_seen": 13516192, + "step": 8880 + }, + { + "epoch": 0.28439280455796684, + "grad_norm": 0.4125514626502991, + "learning_rate": 4.415727998415147e-05, + "loss": 0.6241, + "num_input_tokens_seen": 13524048, + "step": 8885 + }, + { + "epoch": 0.2845528455284553, + "grad_norm": 0.674460232257843, + "learning_rate": 4.4150970842742985e-05, + "loss": 0.3681, + "num_input_tokens_seen": 13531632, + "step": 8890 + }, + { + "epoch": 0.2847128864989437, + "grad_norm": 1.1655364036560059, + "learning_rate": 4.4144658748014134e-05, + "loss": 0.6066, + "num_input_tokens_seen": 13539264, + "step": 8895 + }, + { + "epoch": 0.28487292746943216, + "grad_norm": 0.8450012803077698, + "learning_rate": 4.413834370093831e-05, + "loss": 0.4934, + "num_input_tokens_seen": 13546992, + "step": 8900 + }, + { + "epoch": 0.2850329684399206, + "grad_norm": 0.8032479286193848, + "learning_rate": 4.413202570248939e-05, + "loss": 0.7252, + "num_input_tokens_seen": 13555088, + "step": 8905 + }, + { + "epoch": 0.28519300941040904, + "grad_norm": 2.9566378593444824, + "learning_rate": 4.412570475364167e-05, + "loss": 0.4845, + "num_input_tokens_seen": 13562624, + "step": 8910 + }, + { + "epoch": 0.28535305038089753, + "grad_norm": 0.863731324672699, + "learning_rate": 4.411938085536994e-05, + "loss": 0.6956, + "num_input_tokens_seen": 13570304, + "step": 8915 + }, + { + "epoch": 0.285513091351386, + "grad_norm": 1.1983084678649902, + "learning_rate": 4.41130540086494e-05, + "loss": 0.5663, + "num_input_tokens_seen": 13578080, + "step": 8920 + }, + { + "epoch": 0.2856731323218744, + "grad_norm": 0.7481193542480469, + "learning_rate": 4.4106724214455754e-05, + "loss": 0.4847, + "num_input_tokens_seen": 13585584, + "step": 8925 + }, + { + "epoch": 0.28583317329236285, + "grad_norm": 0.9489485025405884, + "learning_rate": 4.4100391473765115e-05, + "loss": 0.522, + "num_input_tokens_seen": 13592816, + "step": 8930 + }, + { + "epoch": 0.2859932142628513, + "grad_norm": 0.8080632090568542, + "learning_rate": 4.409405578755408e-05, + "loss": 0.5106, + "num_input_tokens_seen": 13600448, + "step": 8935 + }, + { + "epoch": 0.28615325523333973, + "grad_norm": 1.0294984579086304, + "learning_rate": 4.4087717156799705e-05, + "loss": 0.5395, + "num_input_tokens_seen": 13607776, + "step": 8940 + }, + { + "epoch": 0.28631329620382817, + "grad_norm": 1.0644735097885132, + "learning_rate": 4.408137558247946e-05, + "loss": 0.6349, + "num_input_tokens_seen": 13615456, + "step": 8945 + }, + { + "epoch": 0.2864733371743166, + "grad_norm": 0.819896936416626, + "learning_rate": 4.4075031065571306e-05, + "loss": 0.4588, + "num_input_tokens_seen": 13623568, + "step": 8950 + }, + { + "epoch": 0.28663337814480505, + "grad_norm": 0.4382733702659607, + "learning_rate": 4.406868360705366e-05, + "loss": 0.4413, + "num_input_tokens_seen": 13631024, + "step": 8955 + }, + { + "epoch": 0.2867934191152935, + "grad_norm": 0.8409963846206665, + "learning_rate": 4.406233320790536e-05, + "loss": 0.3346, + "num_input_tokens_seen": 13638576, + "step": 8960 + }, + { + "epoch": 0.286953460085782, + "grad_norm": 1.4014946222305298, + "learning_rate": 4.4055979869105734e-05, + "loss": 0.6171, + "num_input_tokens_seen": 13646048, + "step": 8965 + }, + { + "epoch": 0.2871135010562704, + "grad_norm": 1.1555300951004028, + "learning_rate": 4.404962359163454e-05, + "loss": 0.5503, + "num_input_tokens_seen": 13654064, + "step": 8970 + }, + { + "epoch": 0.28727354202675887, + "grad_norm": 0.7552679181098938, + "learning_rate": 4.404326437647199e-05, + "loss": 0.4675, + "num_input_tokens_seen": 13661584, + "step": 8975 + }, + { + "epoch": 0.2874335829972473, + "grad_norm": 0.9763898253440857, + "learning_rate": 4.403690222459877e-05, + "loss": 0.5946, + "num_input_tokens_seen": 13669184, + "step": 8980 + }, + { + "epoch": 0.28759362396773575, + "grad_norm": 0.474920392036438, + "learning_rate": 4.4030537136995984e-05, + "loss": 0.6307, + "num_input_tokens_seen": 13676400, + "step": 8985 + }, + { + "epoch": 0.2877536649382242, + "grad_norm": 1.0752623081207275, + "learning_rate": 4.402416911464523e-05, + "loss": 0.5996, + "num_input_tokens_seen": 13683984, + "step": 8990 + }, + { + "epoch": 0.2879137059087126, + "grad_norm": 0.6227189898490906, + "learning_rate": 4.4017798158528516e-05, + "loss": 0.4683, + "num_input_tokens_seen": 13691856, + "step": 8995 + }, + { + "epoch": 0.28807374687920106, + "grad_norm": 0.8842900395393372, + "learning_rate": 4.401142426962834e-05, + "loss": 0.5067, + "num_input_tokens_seen": 13699248, + "step": 9000 + }, + { + "epoch": 0.28807374687920106, + "eval_loss": 0.569955050945282, + "eval_runtime": 562.9193, + "eval_samples_per_second": 24.668, + "eval_steps_per_second": 12.334, + "num_input_tokens_seen": 13699248, + "step": 9000 + }, + { + "epoch": 0.2882337878496895, + "grad_norm": 0.8991791009902954, + "learning_rate": 4.400504744892763e-05, + "loss": 0.71, + "num_input_tokens_seen": 13707120, + "step": 9005 + }, + { + "epoch": 0.28839382882017794, + "grad_norm": 0.7271087765693665, + "learning_rate": 4.399866769740975e-05, + "loss": 0.4596, + "num_input_tokens_seen": 13714768, + "step": 9010 + }, + { + "epoch": 0.2885538697906664, + "grad_norm": 0.7473287582397461, + "learning_rate": 4.399228501605859e-05, + "loss": 0.5336, + "num_input_tokens_seen": 13722976, + "step": 9015 + }, + { + "epoch": 0.2887139107611549, + "grad_norm": 1.1741743087768555, + "learning_rate": 4.398589940585839e-05, + "loss": 0.6601, + "num_input_tokens_seen": 13730496, + "step": 9020 + }, + { + "epoch": 0.2888739517316433, + "grad_norm": 0.8068690896034241, + "learning_rate": 4.3979510867793917e-05, + "loss": 0.5353, + "num_input_tokens_seen": 13738512, + "step": 9025 + }, + { + "epoch": 0.28903399270213176, + "grad_norm": 0.7641463279724121, + "learning_rate": 4.3973119402850346e-05, + "loss": 0.5972, + "num_input_tokens_seen": 13745952, + "step": 9030 + }, + { + "epoch": 0.2891940336726202, + "grad_norm": 0.5966415405273438, + "learning_rate": 4.396672501201334e-05, + "loss": 0.3404, + "num_input_tokens_seen": 13753584, + "step": 9035 + }, + { + "epoch": 0.28935407464310864, + "grad_norm": 0.5701924562454224, + "learning_rate": 4.396032769626899e-05, + "loss": 0.4793, + "num_input_tokens_seen": 13761216, + "step": 9040 + }, + { + "epoch": 0.2895141156135971, + "grad_norm": 0.8420502543449402, + "learning_rate": 4.395392745660384e-05, + "loss": 0.4857, + "num_input_tokens_seen": 13768752, + "step": 9045 + }, + { + "epoch": 0.2896741565840855, + "grad_norm": 1.3793011903762817, + "learning_rate": 4.394752429400488e-05, + "loss": 0.4629, + "num_input_tokens_seen": 13776320, + "step": 9050 + }, + { + "epoch": 0.28983419755457396, + "grad_norm": 0.6919432878494263, + "learning_rate": 4.394111820945957e-05, + "loss": 0.5471, + "num_input_tokens_seen": 13784192, + "step": 9055 + }, + { + "epoch": 0.2899942385250624, + "grad_norm": 4.273997783660889, + "learning_rate": 4.393470920395579e-05, + "loss": 0.536, + "num_input_tokens_seen": 13791456, + "step": 9060 + }, + { + "epoch": 0.29015427949555084, + "grad_norm": 1.1908382177352905, + "learning_rate": 4.392829727848192e-05, + "loss": 0.674, + "num_input_tokens_seen": 13798960, + "step": 9065 + }, + { + "epoch": 0.29031432046603933, + "grad_norm": 0.566279947757721, + "learning_rate": 4.392188243402673e-05, + "loss": 0.6288, + "num_input_tokens_seen": 13806848, + "step": 9070 + }, + { + "epoch": 0.2904743614365278, + "grad_norm": 0.7450540661811829, + "learning_rate": 4.391546467157949e-05, + "loss": 0.5055, + "num_input_tokens_seen": 13814464, + "step": 9075 + }, + { + "epoch": 0.2906344024070162, + "grad_norm": 0.8209052085876465, + "learning_rate": 4.390904399212988e-05, + "loss": 0.5295, + "num_input_tokens_seen": 13821728, + "step": 9080 + }, + { + "epoch": 0.29079444337750465, + "grad_norm": 1.008332371711731, + "learning_rate": 4.390262039666807e-05, + "loss": 0.5365, + "num_input_tokens_seen": 13829456, + "step": 9085 + }, + { + "epoch": 0.2909544843479931, + "grad_norm": 1.300512433052063, + "learning_rate": 4.389619388618464e-05, + "loss": 0.6425, + "num_input_tokens_seen": 13837088, + "step": 9090 + }, + { + "epoch": 0.29111452531848153, + "grad_norm": 1.0757070779800415, + "learning_rate": 4.3889764461670655e-05, + "loss": 0.5414, + "num_input_tokens_seen": 13844432, + "step": 9095 + }, + { + "epoch": 0.29127456628896997, + "grad_norm": 0.44507455825805664, + "learning_rate": 4.38833321241176e-05, + "loss": 0.5781, + "num_input_tokens_seen": 13851824, + "step": 9100 + }, + { + "epoch": 0.2914346072594584, + "grad_norm": 0.8652309775352478, + "learning_rate": 4.3876896874517434e-05, + "loss": 0.4043, + "num_input_tokens_seen": 13859568, + "step": 9105 + }, + { + "epoch": 0.29159464822994685, + "grad_norm": 0.534821093082428, + "learning_rate": 4.3870458713862554e-05, + "loss": 0.5138, + "num_input_tokens_seen": 13867360, + "step": 9110 + }, + { + "epoch": 0.2917546892004353, + "grad_norm": 0.6799042224884033, + "learning_rate": 4.386401764314579e-05, + "loss": 0.7301, + "num_input_tokens_seen": 13875344, + "step": 9115 + }, + { + "epoch": 0.29191473017092373, + "grad_norm": 1.2197638750076294, + "learning_rate": 4.385757366336045e-05, + "loss": 0.4706, + "num_input_tokens_seen": 13882752, + "step": 9120 + }, + { + "epoch": 0.2920747711414122, + "grad_norm": 0.9556190371513367, + "learning_rate": 4.385112677550027e-05, + "loss": 0.6295, + "num_input_tokens_seen": 13890336, + "step": 9125 + }, + { + "epoch": 0.29223481211190067, + "grad_norm": 1.1709924936294556, + "learning_rate": 4.384467698055945e-05, + "loss": 0.5547, + "num_input_tokens_seen": 13898080, + "step": 9130 + }, + { + "epoch": 0.2923948530823891, + "grad_norm": 0.7546523213386536, + "learning_rate": 4.383822427953261e-05, + "loss": 0.5745, + "num_input_tokens_seen": 13905728, + "step": 9135 + }, + { + "epoch": 0.29255489405287755, + "grad_norm": 2.1406521797180176, + "learning_rate": 4.3831768673414864e-05, + "loss": 0.6965, + "num_input_tokens_seen": 13913568, + "step": 9140 + }, + { + "epoch": 0.292714935023366, + "grad_norm": 1.3032124042510986, + "learning_rate": 4.382531016320173e-05, + "loss": 0.5653, + "num_input_tokens_seen": 13920784, + "step": 9145 + }, + { + "epoch": 0.2928749759938544, + "grad_norm": 0.8347233533859253, + "learning_rate": 4.3818848749889184e-05, + "loss": 0.7045, + "num_input_tokens_seen": 13928480, + "step": 9150 + }, + { + "epoch": 0.29303501696434286, + "grad_norm": 1.0316052436828613, + "learning_rate": 4.381238443447368e-05, + "loss": 0.6055, + "num_input_tokens_seen": 13936128, + "step": 9155 + }, + { + "epoch": 0.2931950579348313, + "grad_norm": 1.0392228364944458, + "learning_rate": 4.380591721795208e-05, + "loss": 0.5409, + "num_input_tokens_seen": 13943872, + "step": 9160 + }, + { + "epoch": 0.29335509890531974, + "grad_norm": 0.6549797058105469, + "learning_rate": 4.3799447101321723e-05, + "loss": 0.5705, + "num_input_tokens_seen": 13951488, + "step": 9165 + }, + { + "epoch": 0.2935151398758082, + "grad_norm": 0.49519336223602295, + "learning_rate": 4.379297408558036e-05, + "loss": 0.5335, + "num_input_tokens_seen": 13959152, + "step": 9170 + }, + { + "epoch": 0.2936751808462967, + "grad_norm": 1.572889804840088, + "learning_rate": 4.378649817172624e-05, + "loss": 0.5554, + "num_input_tokens_seen": 13966480, + "step": 9175 + }, + { + "epoch": 0.2938352218167851, + "grad_norm": 0.8264721632003784, + "learning_rate": 4.378001936075801e-05, + "loss": 0.5724, + "num_input_tokens_seen": 13973968, + "step": 9180 + }, + { + "epoch": 0.29399526278727356, + "grad_norm": 0.9408347010612488, + "learning_rate": 4.377353765367479e-05, + "loss": 0.6337, + "num_input_tokens_seen": 13981136, + "step": 9185 + }, + { + "epoch": 0.294155303757762, + "grad_norm": 0.7309188842773438, + "learning_rate": 4.376705305147614e-05, + "loss": 0.6052, + "num_input_tokens_seen": 13988352, + "step": 9190 + }, + { + "epoch": 0.29431534472825044, + "grad_norm": 0.9942716360092163, + "learning_rate": 4.376056555516206e-05, + "loss": 0.4858, + "num_input_tokens_seen": 13996016, + "step": 9195 + }, + { + "epoch": 0.2944753856987389, + "grad_norm": 0.49650248885154724, + "learning_rate": 4.375407516573302e-05, + "loss": 0.5356, + "num_input_tokens_seen": 14003792, + "step": 9200 + }, + { + "epoch": 0.2944753856987389, + "eval_loss": 0.5686150789260864, + "eval_runtime": 560.0508, + "eval_samples_per_second": 24.794, + "eval_steps_per_second": 12.397, + "num_input_tokens_seen": 14003792, + "step": 9200 + }, + { + "epoch": 0.2946354266692273, + "grad_norm": 0.6541647911071777, + "learning_rate": 4.3747581884189913e-05, + "loss": 0.5893, + "num_input_tokens_seen": 14011136, + "step": 9205 + }, + { + "epoch": 0.29479546763971576, + "grad_norm": 1.0910307168960571, + "learning_rate": 4.374108571153408e-05, + "loss": 0.6063, + "num_input_tokens_seen": 14018640, + "step": 9210 + }, + { + "epoch": 0.2949555086102042, + "grad_norm": 0.5971367955207825, + "learning_rate": 4.3734586648767316e-05, + "loss": 0.5117, + "num_input_tokens_seen": 14026320, + "step": 9215 + }, + { + "epoch": 0.29511554958069264, + "grad_norm": 1.2953505516052246, + "learning_rate": 4.372808469689186e-05, + "loss": 0.6743, + "num_input_tokens_seen": 14033952, + "step": 9220 + }, + { + "epoch": 0.2952755905511811, + "grad_norm": 1.205655574798584, + "learning_rate": 4.372157985691039e-05, + "loss": 0.568, + "num_input_tokens_seen": 14041408, + "step": 9225 + }, + { + "epoch": 0.2954356315216696, + "grad_norm": 1.2748241424560547, + "learning_rate": 4.371507212982603e-05, + "loss": 0.5399, + "num_input_tokens_seen": 14048896, + "step": 9230 + }, + { + "epoch": 0.295595672492158, + "grad_norm": 1.292056679725647, + "learning_rate": 4.370856151664236e-05, + "loss": 0.5027, + "num_input_tokens_seen": 14056208, + "step": 9235 + }, + { + "epoch": 0.29575571346264645, + "grad_norm": 0.7308512926101685, + "learning_rate": 4.3702048018363404e-05, + "loss": 0.6601, + "num_input_tokens_seen": 14063984, + "step": 9240 + }, + { + "epoch": 0.2959157544331349, + "grad_norm": 0.7311374545097351, + "learning_rate": 4.369553163599362e-05, + "loss": 0.6189, + "num_input_tokens_seen": 14071744, + "step": 9245 + }, + { + "epoch": 0.29607579540362333, + "grad_norm": 0.44780871272087097, + "learning_rate": 4.3689012370537904e-05, + "loss": 0.428, + "num_input_tokens_seen": 14079728, + "step": 9250 + }, + { + "epoch": 0.29623583637411177, + "grad_norm": 1.5105477571487427, + "learning_rate": 4.368249022300164e-05, + "loss": 0.7788, + "num_input_tokens_seen": 14086720, + "step": 9255 + }, + { + "epoch": 0.2963958773446002, + "grad_norm": 0.5101797580718994, + "learning_rate": 4.367596519439059e-05, + "loss": 0.4434, + "num_input_tokens_seen": 14094544, + "step": 9260 + }, + { + "epoch": 0.29655591831508865, + "grad_norm": 1.28072190284729, + "learning_rate": 4.366943728571101e-05, + "loss": 0.4887, + "num_input_tokens_seen": 14101648, + "step": 9265 + }, + { + "epoch": 0.2967159592855771, + "grad_norm": 0.9940551519393921, + "learning_rate": 4.366290649796959e-05, + "loss": 0.5311, + "num_input_tokens_seen": 14109520, + "step": 9270 + }, + { + "epoch": 0.29687600025606553, + "grad_norm": 0.7147066593170166, + "learning_rate": 4.3656372832173456e-05, + "loss": 0.4055, + "num_input_tokens_seen": 14116880, + "step": 9275 + }, + { + "epoch": 0.29703604122655397, + "grad_norm": 0.8207569122314453, + "learning_rate": 4.364983628933017e-05, + "loss": 0.7869, + "num_input_tokens_seen": 14124048, + "step": 9280 + }, + { + "epoch": 0.29719608219704247, + "grad_norm": 1.425110936164856, + "learning_rate": 4.364329687044777e-05, + "loss": 0.6777, + "num_input_tokens_seen": 14131984, + "step": 9285 + }, + { + "epoch": 0.2973561231675309, + "grad_norm": 0.6921384334564209, + "learning_rate": 4.36367545765347e-05, + "loss": 0.6203, + "num_input_tokens_seen": 14139152, + "step": 9290 + }, + { + "epoch": 0.29751616413801935, + "grad_norm": 2.3670034408569336, + "learning_rate": 4.363020940859988e-05, + "loss": 0.6319, + "num_input_tokens_seen": 14146320, + "step": 9295 + }, + { + "epoch": 0.2976762051085078, + "grad_norm": 1.0279783010482788, + "learning_rate": 4.362366136765263e-05, + "loss": 0.5714, + "num_input_tokens_seen": 14154032, + "step": 9300 + }, + { + "epoch": 0.2978362460789962, + "grad_norm": 1.421208381652832, + "learning_rate": 4.361711045470278e-05, + "loss": 0.6582, + "num_input_tokens_seen": 14161648, + "step": 9305 + }, + { + "epoch": 0.29799628704948466, + "grad_norm": 0.37606722116470337, + "learning_rate": 4.3610556670760524e-05, + "loss": 0.3313, + "num_input_tokens_seen": 14168816, + "step": 9310 + }, + { + "epoch": 0.2981563280199731, + "grad_norm": 0.9738573431968689, + "learning_rate": 4.360400001683657e-05, + "loss": 0.5777, + "num_input_tokens_seen": 14176944, + "step": 9315 + }, + { + "epoch": 0.29831636899046154, + "grad_norm": 0.5278303027153015, + "learning_rate": 4.3597440493942e-05, + "loss": 0.5787, + "num_input_tokens_seen": 14184208, + "step": 9320 + }, + { + "epoch": 0.29847640996095, + "grad_norm": 0.9135903716087341, + "learning_rate": 4.3590878103088405e-05, + "loss": 0.4974, + "num_input_tokens_seen": 14191648, + "step": 9325 + }, + { + "epoch": 0.2986364509314384, + "grad_norm": 0.6159066557884216, + "learning_rate": 4.358431284528779e-05, + "loss": 0.5039, + "num_input_tokens_seen": 14199504, + "step": 9330 + }, + { + "epoch": 0.2987964919019269, + "grad_norm": 1.3312151432037354, + "learning_rate": 4.357774472155257e-05, + "loss": 0.5351, + "num_input_tokens_seen": 14207280, + "step": 9335 + }, + { + "epoch": 0.29895653287241536, + "grad_norm": 1.1541322469711304, + "learning_rate": 4.3571173732895664e-05, + "loss": 0.5326, + "num_input_tokens_seen": 14214656, + "step": 9340 + }, + { + "epoch": 0.2991165738429038, + "grad_norm": 0.8360676169395447, + "learning_rate": 4.356459988033039e-05, + "loss": 0.5976, + "num_input_tokens_seen": 14222336, + "step": 9345 + }, + { + "epoch": 0.29927661481339224, + "grad_norm": 0.5276663899421692, + "learning_rate": 4.355802316487051e-05, + "loss": 0.494, + "num_input_tokens_seen": 14229552, + "step": 9350 + }, + { + "epoch": 0.2994366557838807, + "grad_norm": 0.9377007484436035, + "learning_rate": 4.355144358753025e-05, + "loss": 0.4476, + "num_input_tokens_seen": 14236944, + "step": 9355 + }, + { + "epoch": 0.2995966967543691, + "grad_norm": 1.0308853387832642, + "learning_rate": 4.354486114932425e-05, + "loss": 0.4263, + "num_input_tokens_seen": 14244336, + "step": 9360 + }, + { + "epoch": 0.29975673772485756, + "grad_norm": 0.7501993775367737, + "learning_rate": 4.353827585126762e-05, + "loss": 0.4713, + "num_input_tokens_seen": 14251712, + "step": 9365 + }, + { + "epoch": 0.299916778695346, + "grad_norm": 0.7476397156715393, + "learning_rate": 4.353168769437588e-05, + "loss": 0.5688, + "num_input_tokens_seen": 14259696, + "step": 9370 + }, + { + "epoch": 0.30007681966583444, + "grad_norm": 1.0594271421432495, + "learning_rate": 4.3525096679665014e-05, + "loss": 0.6356, + "num_input_tokens_seen": 14267248, + "step": 9375 + }, + { + "epoch": 0.3002368606363229, + "grad_norm": 0.4274722635746002, + "learning_rate": 4.351850280815144e-05, + "loss": 0.4447, + "num_input_tokens_seen": 14275552, + "step": 9380 + }, + { + "epoch": 0.3003969016068113, + "grad_norm": 0.39791756868362427, + "learning_rate": 4.3511906080852014e-05, + "loss": 0.3851, + "num_input_tokens_seen": 14283504, + "step": 9385 + }, + { + "epoch": 0.3005569425772998, + "grad_norm": 0.744077205657959, + "learning_rate": 4.350530649878404e-05, + "loss": 0.5521, + "num_input_tokens_seen": 14291056, + "step": 9390 + }, + { + "epoch": 0.30071698354778825, + "grad_norm": 0.7878570556640625, + "learning_rate": 4.3498704062965246e-05, + "loss": 0.5206, + "num_input_tokens_seen": 14298800, + "step": 9395 + }, + { + "epoch": 0.3008770245182767, + "grad_norm": 0.5880170464515686, + "learning_rate": 4.3492098774413815e-05, + "loss": 0.7087, + "num_input_tokens_seen": 14305968, + "step": 9400 + }, + { + "epoch": 0.3008770245182767, + "eval_loss": 0.5670424699783325, + "eval_runtime": 559.5423, + "eval_samples_per_second": 24.817, + "eval_steps_per_second": 12.408, + "num_input_tokens_seen": 14305968, + "step": 9400 + }, + { + "epoch": 0.30103706548876513, + "grad_norm": 0.49274978041648865, + "learning_rate": 4.3485490634148375e-05, + "loss": 0.3524, + "num_input_tokens_seen": 14313248, + "step": 9405 + }, + { + "epoch": 0.30119710645925357, + "grad_norm": 0.8516497015953064, + "learning_rate": 4.347887964318797e-05, + "loss": 0.6754, + "num_input_tokens_seen": 14320720, + "step": 9410 + }, + { + "epoch": 0.301357147429742, + "grad_norm": 2.02329421043396, + "learning_rate": 4.34722658025521e-05, + "loss": 0.7181, + "num_input_tokens_seen": 14328656, + "step": 9415 + }, + { + "epoch": 0.30151718840023045, + "grad_norm": 1.340470314025879, + "learning_rate": 4.346564911326071e-05, + "loss": 0.5261, + "num_input_tokens_seen": 14336240, + "step": 9420 + }, + { + "epoch": 0.3016772293707189, + "grad_norm": 0.6087207198143005, + "learning_rate": 4.345902957633418e-05, + "loss": 0.5334, + "num_input_tokens_seen": 14343936, + "step": 9425 + }, + { + "epoch": 0.30183727034120733, + "grad_norm": 0.9085726141929626, + "learning_rate": 4.345240719279331e-05, + "loss": 0.4784, + "num_input_tokens_seen": 14351664, + "step": 9430 + }, + { + "epoch": 0.30199731131169577, + "grad_norm": 0.8086707592010498, + "learning_rate": 4.3445781963659374e-05, + "loss": 0.5461, + "num_input_tokens_seen": 14358768, + "step": 9435 + }, + { + "epoch": 0.30215735228218427, + "grad_norm": 0.7759921550750732, + "learning_rate": 4.3439153889954045e-05, + "loss": 0.5289, + "num_input_tokens_seen": 14366336, + "step": 9440 + }, + { + "epoch": 0.3023173932526727, + "grad_norm": 0.7072399258613586, + "learning_rate": 4.343252297269946e-05, + "loss": 0.3565, + "num_input_tokens_seen": 14374048, + "step": 9445 + }, + { + "epoch": 0.30247743422316115, + "grad_norm": 0.8271534442901611, + "learning_rate": 4.342588921291821e-05, + "loss": 0.5742, + "num_input_tokens_seen": 14381680, + "step": 9450 + }, + { + "epoch": 0.3026374751936496, + "grad_norm": 0.9400429725646973, + "learning_rate": 4.341925261163328e-05, + "loss": 0.5166, + "num_input_tokens_seen": 14389376, + "step": 9455 + }, + { + "epoch": 0.302797516164138, + "grad_norm": 0.9940510392189026, + "learning_rate": 4.341261316986813e-05, + "loss": 0.5803, + "num_input_tokens_seen": 14396720, + "step": 9460 + }, + { + "epoch": 0.30295755713462647, + "grad_norm": 0.9207686185836792, + "learning_rate": 4.340597088864664e-05, + "loss": 0.658, + "num_input_tokens_seen": 14404496, + "step": 9465 + }, + { + "epoch": 0.3031175981051149, + "grad_norm": 0.865799605846405, + "learning_rate": 4.339932576899313e-05, + "loss": 0.4387, + "num_input_tokens_seen": 14412416, + "step": 9470 + }, + { + "epoch": 0.30327763907560334, + "grad_norm": 0.9265658855438232, + "learning_rate": 4.3392677811932375e-05, + "loss": 0.6016, + "num_input_tokens_seen": 14420016, + "step": 9475 + }, + { + "epoch": 0.3034376800460918, + "grad_norm": 0.4952104687690735, + "learning_rate": 4.338602701848956e-05, + "loss": 0.4263, + "num_input_tokens_seen": 14427664, + "step": 9480 + }, + { + "epoch": 0.3035977210165802, + "grad_norm": 0.7012612819671631, + "learning_rate": 4.337937338969033e-05, + "loss": 0.5728, + "num_input_tokens_seen": 14435024, + "step": 9485 + }, + { + "epoch": 0.30375776198706866, + "grad_norm": 0.7299774289131165, + "learning_rate": 4.337271692656075e-05, + "loss": 0.3952, + "num_input_tokens_seen": 14443040, + "step": 9490 + }, + { + "epoch": 0.30391780295755716, + "grad_norm": 0.8587384223937988, + "learning_rate": 4.336605763012733e-05, + "loss": 0.4179, + "num_input_tokens_seen": 14451168, + "step": 9495 + }, + { + "epoch": 0.3040778439280456, + "grad_norm": 1.033046841621399, + "learning_rate": 4.3359395501417026e-05, + "loss": 0.5649, + "num_input_tokens_seen": 14458608, + "step": 9500 + }, + { + "epoch": 0.30423788489853404, + "grad_norm": 0.5985113382339478, + "learning_rate": 4.335273054145722e-05, + "loss": 0.7472, + "num_input_tokens_seen": 14465648, + "step": 9505 + }, + { + "epoch": 0.3043979258690225, + "grad_norm": 0.9471729397773743, + "learning_rate": 4.334606275127572e-05, + "loss": 0.6967, + "num_input_tokens_seen": 14472784, + "step": 9510 + }, + { + "epoch": 0.3045579668395109, + "grad_norm": 0.46219998598098755, + "learning_rate": 4.33393921319008e-05, + "loss": 0.513, + "num_input_tokens_seen": 14479920, + "step": 9515 + }, + { + "epoch": 0.30471800780999936, + "grad_norm": 0.4969165325164795, + "learning_rate": 4.3332718684361146e-05, + "loss": 0.4641, + "num_input_tokens_seen": 14487376, + "step": 9520 + }, + { + "epoch": 0.3048780487804878, + "grad_norm": 0.855856716632843, + "learning_rate": 4.332604240968588e-05, + "loss": 0.3902, + "num_input_tokens_seen": 14495104, + "step": 9525 + }, + { + "epoch": 0.30503808975097624, + "grad_norm": 0.19919775426387787, + "learning_rate": 4.331936330890459e-05, + "loss": 0.3589, + "num_input_tokens_seen": 14502944, + "step": 9530 + }, + { + "epoch": 0.3051981307214647, + "grad_norm": 0.9006764888763428, + "learning_rate": 4.331268138304725e-05, + "loss": 0.4738, + "num_input_tokens_seen": 14509984, + "step": 9535 + }, + { + "epoch": 0.3053581716919531, + "grad_norm": 0.838519811630249, + "learning_rate": 4.330599663314431e-05, + "loss": 0.4394, + "num_input_tokens_seen": 14517296, + "step": 9540 + }, + { + "epoch": 0.3055182126624416, + "grad_norm": 0.6926352381706238, + "learning_rate": 4.329930906022665e-05, + "loss": 0.5683, + "num_input_tokens_seen": 14524896, + "step": 9545 + }, + { + "epoch": 0.30567825363293005, + "grad_norm": 0.6464629769325256, + "learning_rate": 4.3292618665325564e-05, + "loss": 0.3741, + "num_input_tokens_seen": 14532336, + "step": 9550 + }, + { + "epoch": 0.3058382946034185, + "grad_norm": 1.6954089403152466, + "learning_rate": 4.3285925449472796e-05, + "loss": 0.6312, + "num_input_tokens_seen": 14539856, + "step": 9555 + }, + { + "epoch": 0.30599833557390693, + "grad_norm": 0.8927382230758667, + "learning_rate": 4.327922941370054e-05, + "loss": 0.5723, + "num_input_tokens_seen": 14547568, + "step": 9560 + }, + { + "epoch": 0.30615837654439537, + "grad_norm": 0.627816915512085, + "learning_rate": 4.3272530559041384e-05, + "loss": 0.5351, + "num_input_tokens_seen": 14555360, + "step": 9565 + }, + { + "epoch": 0.3063184175148838, + "grad_norm": 0.7036879658699036, + "learning_rate": 4.32658288865284e-05, + "loss": 0.5164, + "num_input_tokens_seen": 14562912, + "step": 9570 + }, + { + "epoch": 0.30647845848537225, + "grad_norm": 0.9171373248100281, + "learning_rate": 4.325912439719505e-05, + "loss": 0.5119, + "num_input_tokens_seen": 14570096, + "step": 9575 + }, + { + "epoch": 0.3066384994558607, + "grad_norm": 1.3307253122329712, + "learning_rate": 4.3252417092075266e-05, + "loss": 0.6485, + "num_input_tokens_seen": 14577184, + "step": 9580 + }, + { + "epoch": 0.30679854042634913, + "grad_norm": 0.7963880896568298, + "learning_rate": 4.3245706972203385e-05, + "loss": 0.5581, + "num_input_tokens_seen": 14584688, + "step": 9585 + }, + { + "epoch": 0.30695858139683757, + "grad_norm": 0.3627079725265503, + "learning_rate": 4.323899403861421e-05, + "loss": 0.4594, + "num_input_tokens_seen": 14592016, + "step": 9590 + }, + { + "epoch": 0.307118622367326, + "grad_norm": 0.9992244839668274, + "learning_rate": 4.3232278292342935e-05, + "loss": 0.5193, + "num_input_tokens_seen": 14599856, + "step": 9595 + }, + { + "epoch": 0.3072786633378145, + "grad_norm": 0.954164445400238, + "learning_rate": 4.322555973442524e-05, + "loss": 0.5085, + "num_input_tokens_seen": 14607328, + "step": 9600 + }, + { + "epoch": 0.3072786633378145, + "eval_loss": 0.5656592845916748, + "eval_runtime": 558.0134, + "eval_samples_per_second": 24.885, + "eval_steps_per_second": 12.442, + "num_input_tokens_seen": 14607328, + "step": 9600 + }, + { + "epoch": 0.30743870430830295, + "grad_norm": 0.584834635257721, + "learning_rate": 4.3218838365897184e-05, + "loss": 0.4598, + "num_input_tokens_seen": 14614688, + "step": 9605 + }, + { + "epoch": 0.3075987452787914, + "grad_norm": 0.6987232565879822, + "learning_rate": 4.3212114187795306e-05, + "loss": 0.3977, + "num_input_tokens_seen": 14622464, + "step": 9610 + }, + { + "epoch": 0.3077587862492798, + "grad_norm": 1.0568631887435913, + "learning_rate": 4.320538720115656e-05, + "loss": 0.564, + "num_input_tokens_seen": 14629984, + "step": 9615 + }, + { + "epoch": 0.30791882721976827, + "grad_norm": 0.9638930559158325, + "learning_rate": 4.319865740701831e-05, + "loss": 0.8882, + "num_input_tokens_seen": 14637536, + "step": 9620 + }, + { + "epoch": 0.3080788681902567, + "grad_norm": 0.8006612658500671, + "learning_rate": 4.3191924806418396e-05, + "loss": 0.5071, + "num_input_tokens_seen": 14645232, + "step": 9625 + }, + { + "epoch": 0.30823890916074514, + "grad_norm": 1.244044303894043, + "learning_rate": 4.318518940039507e-05, + "loss": 0.6247, + "num_input_tokens_seen": 14652304, + "step": 9630 + }, + { + "epoch": 0.3083989501312336, + "grad_norm": 0.8918948173522949, + "learning_rate": 4.3178451189987e-05, + "loss": 0.4182, + "num_input_tokens_seen": 14660192, + "step": 9635 + }, + { + "epoch": 0.308558991101722, + "grad_norm": 0.8024488091468811, + "learning_rate": 4.3171710176233315e-05, + "loss": 0.4923, + "num_input_tokens_seen": 14668016, + "step": 9640 + }, + { + "epoch": 0.30871903207221046, + "grad_norm": 0.7691476941108704, + "learning_rate": 4.316496636017355e-05, + "loss": 0.5402, + "num_input_tokens_seen": 14675680, + "step": 9645 + }, + { + "epoch": 0.3088790730426989, + "grad_norm": 1.610133171081543, + "learning_rate": 4.315821974284771e-05, + "loss": 0.75, + "num_input_tokens_seen": 14683504, + "step": 9650 + }, + { + "epoch": 0.3090391140131874, + "grad_norm": 0.5662567019462585, + "learning_rate": 4.315147032529619e-05, + "loss": 0.4595, + "num_input_tokens_seen": 14691440, + "step": 9655 + }, + { + "epoch": 0.30919915498367584, + "grad_norm": 0.7975409030914307, + "learning_rate": 4.3144718108559845e-05, + "loss": 0.4865, + "num_input_tokens_seen": 14699296, + "step": 9660 + }, + { + "epoch": 0.3093591959541643, + "grad_norm": 0.7136954665184021, + "learning_rate": 4.3137963093679945e-05, + "loss": 0.5334, + "num_input_tokens_seen": 14707440, + "step": 9665 + }, + { + "epoch": 0.3095192369246527, + "grad_norm": 0.8750830888748169, + "learning_rate": 4.31312052816982e-05, + "loss": 0.5408, + "num_input_tokens_seen": 14715104, + "step": 9670 + }, + { + "epoch": 0.30967927789514116, + "grad_norm": 1.4059062004089355, + "learning_rate": 4.312444467365675e-05, + "loss": 0.6073, + "num_input_tokens_seen": 14722928, + "step": 9675 + }, + { + "epoch": 0.3098393188656296, + "grad_norm": 0.5411925911903381, + "learning_rate": 4.311768127059816e-05, + "loss": 0.4542, + "num_input_tokens_seen": 14730096, + "step": 9680 + }, + { + "epoch": 0.30999935983611804, + "grad_norm": 0.5061872005462646, + "learning_rate": 4.3110915073565444e-05, + "loss": 0.3847, + "num_input_tokens_seen": 14737504, + "step": 9685 + }, + { + "epoch": 0.3101594008066065, + "grad_norm": 2.301029682159424, + "learning_rate": 4.310414608360203e-05, + "loss": 0.7325, + "num_input_tokens_seen": 14745024, + "step": 9690 + }, + { + "epoch": 0.3103194417770949, + "grad_norm": 1.824466347694397, + "learning_rate": 4.309737430175177e-05, + "loss": 0.5112, + "num_input_tokens_seen": 14752624, + "step": 9695 + }, + { + "epoch": 0.31047948274758336, + "grad_norm": 0.9042192697525024, + "learning_rate": 4.309059972905897e-05, + "loss": 0.6501, + "num_input_tokens_seen": 14760080, + "step": 9700 + }, + { + "epoch": 0.31063952371807185, + "grad_norm": 1.720211148262024, + "learning_rate": 4.308382236656836e-05, + "loss": 0.5926, + "num_input_tokens_seen": 14767616, + "step": 9705 + }, + { + "epoch": 0.3107995646885603, + "grad_norm": 1.223905324935913, + "learning_rate": 4.307704221532507e-05, + "loss": 0.534, + "num_input_tokens_seen": 14775104, + "step": 9710 + }, + { + "epoch": 0.31095960565904873, + "grad_norm": 0.8553587198257446, + "learning_rate": 4.307025927637471e-05, + "loss": 0.6656, + "num_input_tokens_seen": 14782816, + "step": 9715 + }, + { + "epoch": 0.31111964662953717, + "grad_norm": 0.9554088711738586, + "learning_rate": 4.306347355076328e-05, + "loss": 0.4637, + "num_input_tokens_seen": 14790192, + "step": 9720 + }, + { + "epoch": 0.3112796876000256, + "grad_norm": 1.4517223834991455, + "learning_rate": 4.305668503953724e-05, + "loss": 0.5909, + "num_input_tokens_seen": 14797680, + "step": 9725 + }, + { + "epoch": 0.31143972857051405, + "grad_norm": 1.1911193132400513, + "learning_rate": 4.3049893743743436e-05, + "loss": 0.584, + "num_input_tokens_seen": 14805776, + "step": 9730 + }, + { + "epoch": 0.3115997695410025, + "grad_norm": 1.2276214361190796, + "learning_rate": 4.304309966442919e-05, + "loss": 0.5612, + "num_input_tokens_seen": 14813472, + "step": 9735 + }, + { + "epoch": 0.31175981051149093, + "grad_norm": 0.9003334045410156, + "learning_rate": 4.303630280264224e-05, + "loss": 0.5366, + "num_input_tokens_seen": 14821328, + "step": 9740 + }, + { + "epoch": 0.31191985148197937, + "grad_norm": 0.5846673846244812, + "learning_rate": 4.302950315943074e-05, + "loss": 0.5678, + "num_input_tokens_seen": 14828880, + "step": 9745 + }, + { + "epoch": 0.3120798924524678, + "grad_norm": 0.8288050889968872, + "learning_rate": 4.3022700735843275e-05, + "loss": 0.4638, + "num_input_tokens_seen": 14836112, + "step": 9750 + }, + { + "epoch": 0.31223993342295625, + "grad_norm": 1.300505518913269, + "learning_rate": 4.301589553292887e-05, + "loss": 0.6024, + "num_input_tokens_seen": 14843328, + "step": 9755 + }, + { + "epoch": 0.31239997439344475, + "grad_norm": 0.6344118118286133, + "learning_rate": 4.300908755173697e-05, + "loss": 0.4376, + "num_input_tokens_seen": 14851072, + "step": 9760 + }, + { + "epoch": 0.3125600153639332, + "grad_norm": 0.7525883316993713, + "learning_rate": 4.300227679331745e-05, + "loss": 0.4978, + "num_input_tokens_seen": 14858688, + "step": 9765 + }, + { + "epoch": 0.3127200563344216, + "grad_norm": 0.4367482364177704, + "learning_rate": 4.299546325872063e-05, + "loss": 0.3986, + "num_input_tokens_seen": 14866256, + "step": 9770 + }, + { + "epoch": 0.31288009730491007, + "grad_norm": 1.1421462297439575, + "learning_rate": 4.2988646948997225e-05, + "loss": 0.4496, + "num_input_tokens_seen": 14873920, + "step": 9775 + }, + { + "epoch": 0.3130401382753985, + "grad_norm": 0.8628638982772827, + "learning_rate": 4.29818278651984e-05, + "loss": 0.5284, + "num_input_tokens_seen": 14881232, + "step": 9780 + }, + { + "epoch": 0.31320017924588694, + "grad_norm": 0.5468716025352478, + "learning_rate": 4.297500600837574e-05, + "loss": 0.6136, + "num_input_tokens_seen": 14888896, + "step": 9785 + }, + { + "epoch": 0.3133602202163754, + "grad_norm": 1.0618879795074463, + "learning_rate": 4.2968181379581276e-05, + "loss": 0.5477, + "num_input_tokens_seen": 14896448, + "step": 9790 + }, + { + "epoch": 0.3135202611868638, + "grad_norm": 0.6335957050323486, + "learning_rate": 4.296135397986743e-05, + "loss": 0.5414, + "num_input_tokens_seen": 14903680, + "step": 9795 + }, + { + "epoch": 0.31368030215735226, + "grad_norm": 0.8202107548713684, + "learning_rate": 4.295452381028709e-05, + "loss": 0.5123, + "num_input_tokens_seen": 14911184, + "step": 9800 + }, + { + "epoch": 0.31368030215735226, + "eval_loss": 0.5638614296913147, + "eval_runtime": 560.6626, + "eval_samples_per_second": 24.767, + "eval_steps_per_second": 12.384, + "num_input_tokens_seen": 14911184, + "step": 9800 + }, + { + "epoch": 0.3138403431278407, + "grad_norm": 0.7084894180297852, + "learning_rate": 4.294769087189354e-05, + "loss": 0.4584, + "num_input_tokens_seen": 14919136, + "step": 9805 + }, + { + "epoch": 0.3140003840983292, + "grad_norm": 0.9831363558769226, + "learning_rate": 4.294085516574052e-05, + "loss": 0.5727, + "num_input_tokens_seen": 14927088, + "step": 9810 + }, + { + "epoch": 0.31416042506881764, + "grad_norm": 0.7459443211555481, + "learning_rate": 4.2934016692882176e-05, + "loss": 0.361, + "num_input_tokens_seen": 14934688, + "step": 9815 + }, + { + "epoch": 0.3143204660393061, + "grad_norm": 1.076149344444275, + "learning_rate": 4.292717545437308e-05, + "loss": 0.3808, + "num_input_tokens_seen": 14941984, + "step": 9820 + }, + { + "epoch": 0.3144805070097945, + "grad_norm": 0.7600417137145996, + "learning_rate": 4.292033145126825e-05, + "loss": 0.5349, + "num_input_tokens_seen": 14949184, + "step": 9825 + }, + { + "epoch": 0.31464054798028296, + "grad_norm": 1.1327463388442993, + "learning_rate": 4.29134846846231e-05, + "loss": 0.6219, + "num_input_tokens_seen": 14956384, + "step": 9830 + }, + { + "epoch": 0.3148005889507714, + "grad_norm": 0.5630513429641724, + "learning_rate": 4.29066351554935e-05, + "loss": 0.4418, + "num_input_tokens_seen": 14963840, + "step": 9835 + }, + { + "epoch": 0.31496062992125984, + "grad_norm": 0.9916123151779175, + "learning_rate": 4.289978286493574e-05, + "loss": 0.4882, + "num_input_tokens_seen": 14971152, + "step": 9840 + }, + { + "epoch": 0.3151206708917483, + "grad_norm": 1.3942610025405884, + "learning_rate": 4.28929278140065e-05, + "loss": 0.5942, + "num_input_tokens_seen": 14978848, + "step": 9845 + }, + { + "epoch": 0.3152807118622367, + "grad_norm": 0.872172474861145, + "learning_rate": 4.288607000376295e-05, + "loss": 0.5203, + "num_input_tokens_seen": 14986096, + "step": 9850 + }, + { + "epoch": 0.31544075283272516, + "grad_norm": 0.8795726299285889, + "learning_rate": 4.2879209435262624e-05, + "loss": 0.5821, + "num_input_tokens_seen": 14994368, + "step": 9855 + }, + { + "epoch": 0.3156007938032136, + "grad_norm": 0.43863117694854736, + "learning_rate": 4.287234610956353e-05, + "loss": 0.3863, + "num_input_tokens_seen": 15002368, + "step": 9860 + }, + { + "epoch": 0.3157608347737021, + "grad_norm": 1.3118253946304321, + "learning_rate": 4.2865480027724056e-05, + "loss": 0.5558, + "num_input_tokens_seen": 15009904, + "step": 9865 + }, + { + "epoch": 0.31592087574419053, + "grad_norm": 0.5520841479301453, + "learning_rate": 4.285861119080306e-05, + "loss": 0.6172, + "num_input_tokens_seen": 15017728, + "step": 9870 + }, + { + "epoch": 0.31608091671467897, + "grad_norm": 1.116693377494812, + "learning_rate": 4.2851739599859784e-05, + "loss": 0.6058, + "num_input_tokens_seen": 15025136, + "step": 9875 + }, + { + "epoch": 0.3162409576851674, + "grad_norm": 0.5180211663246155, + "learning_rate": 4.2844865255953934e-05, + "loss": 0.4774, + "num_input_tokens_seen": 15032832, + "step": 9880 + }, + { + "epoch": 0.31640099865565585, + "grad_norm": 0.8648021817207336, + "learning_rate": 4.2837988160145605e-05, + "loss": 0.7657, + "num_input_tokens_seen": 15040368, + "step": 9885 + }, + { + "epoch": 0.3165610396261443, + "grad_norm": 0.7965812683105469, + "learning_rate": 4.2831108313495336e-05, + "loss": 0.5014, + "num_input_tokens_seen": 15048464, + "step": 9890 + }, + { + "epoch": 0.31672108059663273, + "grad_norm": 1.0512056350708008, + "learning_rate": 4.282422571706408e-05, + "loss": 0.3831, + "num_input_tokens_seen": 15056880, + "step": 9895 + }, + { + "epoch": 0.31688112156712117, + "grad_norm": 0.9262704253196716, + "learning_rate": 4.281734037191323e-05, + "loss": 0.4395, + "num_input_tokens_seen": 15064032, + "step": 9900 + }, + { + "epoch": 0.3170411625376096, + "grad_norm": 0.566419243812561, + "learning_rate": 4.281045227910459e-05, + "loss": 0.4062, + "num_input_tokens_seen": 15071616, + "step": 9905 + }, + { + "epoch": 0.31720120350809805, + "grad_norm": 0.6511299014091492, + "learning_rate": 4.280356143970038e-05, + "loss": 0.4508, + "num_input_tokens_seen": 15079152, + "step": 9910 + }, + { + "epoch": 0.3173612444785865, + "grad_norm": 1.1279901266098022, + "learning_rate": 4.279666785476327e-05, + "loss": 0.5582, + "num_input_tokens_seen": 15087184, + "step": 9915 + }, + { + "epoch": 0.317521285449075, + "grad_norm": 0.678996205329895, + "learning_rate": 4.2789771525356325e-05, + "loss": 0.5603, + "num_input_tokens_seen": 15095536, + "step": 9920 + }, + { + "epoch": 0.3176813264195634, + "grad_norm": 1.0226565599441528, + "learning_rate": 4.2782872452543056e-05, + "loss": 0.6744, + "num_input_tokens_seen": 15103072, + "step": 9925 + }, + { + "epoch": 0.31784136739005187, + "grad_norm": 1.1178308725357056, + "learning_rate": 4.2775970637387376e-05, + "loss": 0.6188, + "num_input_tokens_seen": 15110768, + "step": 9930 + }, + { + "epoch": 0.3180014083605403, + "grad_norm": 0.7823948264122009, + "learning_rate": 4.276906608095363e-05, + "loss": 0.4273, + "num_input_tokens_seen": 15118464, + "step": 9935 + }, + { + "epoch": 0.31816144933102875, + "grad_norm": 1.5005146265029907, + "learning_rate": 4.276215878430661e-05, + "loss": 0.564, + "num_input_tokens_seen": 15125696, + "step": 9940 + }, + { + "epoch": 0.3183214903015172, + "grad_norm": 1.3257962465286255, + "learning_rate": 4.275524874851149e-05, + "loss": 0.6042, + "num_input_tokens_seen": 15132928, + "step": 9945 + }, + { + "epoch": 0.3184815312720056, + "grad_norm": 1.3359168767929077, + "learning_rate": 4.274833597463388e-05, + "loss": 0.5402, + "num_input_tokens_seen": 15140688, + "step": 9950 + }, + { + "epoch": 0.31864157224249406, + "grad_norm": 1.3763236999511719, + "learning_rate": 4.2741420463739824e-05, + "loss": 0.5509, + "num_input_tokens_seen": 15148672, + "step": 9955 + }, + { + "epoch": 0.3188016132129825, + "grad_norm": 0.9595286846160889, + "learning_rate": 4.273450221689578e-05, + "loss": 0.5293, + "num_input_tokens_seen": 15156304, + "step": 9960 + }, + { + "epoch": 0.31896165418347094, + "grad_norm": 1.5481151342391968, + "learning_rate": 4.272758123516863e-05, + "loss": 0.5796, + "num_input_tokens_seen": 15163904, + "step": 9965 + }, + { + "epoch": 0.31912169515395944, + "grad_norm": 1.2876557111740112, + "learning_rate": 4.272065751962567e-05, + "loss": 0.6077, + "num_input_tokens_seen": 15171888, + "step": 9970 + }, + { + "epoch": 0.3192817361244479, + "grad_norm": 1.1846214532852173, + "learning_rate": 4.271373107133464e-05, + "loss": 0.5813, + "num_input_tokens_seen": 15179472, + "step": 9975 + }, + { + "epoch": 0.3194417770949363, + "grad_norm": 0.8495386838912964, + "learning_rate": 4.270680189136366e-05, + "loss": 0.5172, + "num_input_tokens_seen": 15187360, + "step": 9980 + }, + { + "epoch": 0.31960181806542476, + "grad_norm": 1.4887921810150146, + "learning_rate": 4.269986998078132e-05, + "loss": 0.584, + "num_input_tokens_seen": 15194544, + "step": 9985 + }, + { + "epoch": 0.3197618590359132, + "grad_norm": 1.0409082174301147, + "learning_rate": 4.2692935340656595e-05, + "loss": 0.5185, + "num_input_tokens_seen": 15202416, + "step": 9990 + }, + { + "epoch": 0.31992190000640164, + "grad_norm": 0.46998000144958496, + "learning_rate": 4.26859979720589e-05, + "loss": 0.4857, + "num_input_tokens_seen": 15210224, + "step": 9995 + }, + { + "epoch": 0.3200819409768901, + "grad_norm": 0.7414172291755676, + "learning_rate": 4.267905787605806e-05, + "loss": 0.5891, + "num_input_tokens_seen": 15217456, + "step": 10000 + }, + { + "epoch": 0.3200819409768901, + "eval_loss": 0.5625746250152588, + "eval_runtime": 563.6936, + "eval_samples_per_second": 24.634, + "eval_steps_per_second": 12.317, + "num_input_tokens_seen": 15217456, + "step": 10000 + }, + { + "epoch": 0.3202419819473785, + "grad_norm": 0.5844265222549438, + "learning_rate": 4.267211505372433e-05, + "loss": 0.5766, + "num_input_tokens_seen": 15225152, + "step": 10005 + }, + { + "epoch": 0.32040202291786696, + "grad_norm": 0.6834073662757874, + "learning_rate": 4.266516950612837e-05, + "loss": 0.4871, + "num_input_tokens_seen": 15232608, + "step": 10010 + }, + { + "epoch": 0.3205620638883554, + "grad_norm": 0.750968873500824, + "learning_rate": 4.265822123434128e-05, + "loss": 0.5039, + "num_input_tokens_seen": 15240528, + "step": 10015 + }, + { + "epoch": 0.32072210485884384, + "grad_norm": 1.0366311073303223, + "learning_rate": 4.265127023943457e-05, + "loss": 0.5207, + "num_input_tokens_seen": 15248384, + "step": 10020 + }, + { + "epoch": 0.32088214582933233, + "grad_norm": 1.1341584920883179, + "learning_rate": 4.2644316522480176e-05, + "loss": 0.4918, + "num_input_tokens_seen": 15256128, + "step": 10025 + }, + { + "epoch": 0.3210421867998208, + "grad_norm": 1.0409873723983765, + "learning_rate": 4.263736008455044e-05, + "loss": 0.636, + "num_input_tokens_seen": 15263488, + "step": 10030 + }, + { + "epoch": 0.3212022277703092, + "grad_norm": 0.8501798510551453, + "learning_rate": 4.2630400926718125e-05, + "loss": 0.6166, + "num_input_tokens_seen": 15271936, + "step": 10035 + }, + { + "epoch": 0.32136226874079765, + "grad_norm": 1.5590245723724365, + "learning_rate": 4.262343905005644e-05, + "loss": 0.524, + "num_input_tokens_seen": 15279456, + "step": 10040 + }, + { + "epoch": 0.3215223097112861, + "grad_norm": 0.6874428987503052, + "learning_rate": 4.261647445563897e-05, + "loss": 0.519, + "num_input_tokens_seen": 15287712, + "step": 10045 + }, + { + "epoch": 0.32168235068177453, + "grad_norm": 3.2253739833831787, + "learning_rate": 4.260950714453976e-05, + "loss": 0.6213, + "num_input_tokens_seen": 15294928, + "step": 10050 + }, + { + "epoch": 0.32184239165226297, + "grad_norm": 0.6839028596878052, + "learning_rate": 4.2602537117833266e-05, + "loss": 0.4898, + "num_input_tokens_seen": 15302080, + "step": 10055 + }, + { + "epoch": 0.3220024326227514, + "grad_norm": 0.48160576820373535, + "learning_rate": 4.259556437659433e-05, + "loss": 0.5075, + "num_input_tokens_seen": 15309392, + "step": 10060 + }, + { + "epoch": 0.32216247359323985, + "grad_norm": 1.3761417865753174, + "learning_rate": 4.258858892189825e-05, + "loss": 0.5512, + "num_input_tokens_seen": 15316832, + "step": 10065 + }, + { + "epoch": 0.3223225145637283, + "grad_norm": 0.7871055603027344, + "learning_rate": 4.2581610754820725e-05, + "loss": 0.5111, + "num_input_tokens_seen": 15324640, + "step": 10070 + }, + { + "epoch": 0.3224825555342168, + "grad_norm": 0.8712334036827087, + "learning_rate": 4.2574629876437876e-05, + "loss": 0.8086, + "num_input_tokens_seen": 15331872, + "step": 10075 + }, + { + "epoch": 0.3226425965047052, + "grad_norm": 0.6656718850135803, + "learning_rate": 4.256764628782625e-05, + "loss": 0.4485, + "num_input_tokens_seen": 15339456, + "step": 10080 + }, + { + "epoch": 0.32280263747519367, + "grad_norm": 0.6506297588348389, + "learning_rate": 4.256065999006279e-05, + "loss": 0.7124, + "num_input_tokens_seen": 15346688, + "step": 10085 + }, + { + "epoch": 0.3229626784456821, + "grad_norm": 1.09325110912323, + "learning_rate": 4.2553670984224885e-05, + "loss": 0.5827, + "num_input_tokens_seen": 15354000, + "step": 10090 + }, + { + "epoch": 0.32312271941617055, + "grad_norm": 0.7297424077987671, + "learning_rate": 4.254667927139032e-05, + "loss": 0.4726, + "num_input_tokens_seen": 15361456, + "step": 10095 + }, + { + "epoch": 0.323282760386659, + "grad_norm": 0.5142780542373657, + "learning_rate": 4.2539684852637295e-05, + "loss": 0.7342, + "num_input_tokens_seen": 15369376, + "step": 10100 + }, + { + "epoch": 0.3234428013571474, + "grad_norm": 0.7250069975852966, + "learning_rate": 4.253268772904446e-05, + "loss": 0.4722, + "num_input_tokens_seen": 15376624, + "step": 10105 + }, + { + "epoch": 0.32360284232763586, + "grad_norm": 1.2726829051971436, + "learning_rate": 4.252568790169085e-05, + "loss": 0.5779, + "num_input_tokens_seen": 15384736, + "step": 10110 + }, + { + "epoch": 0.3237628832981243, + "grad_norm": 1.4055935144424438, + "learning_rate": 4.251868537165592e-05, + "loss": 0.5147, + "num_input_tokens_seen": 15392512, + "step": 10115 + }, + { + "epoch": 0.32392292426861274, + "grad_norm": 0.8379647135734558, + "learning_rate": 4.251168014001955e-05, + "loss": 0.6618, + "num_input_tokens_seen": 15399712, + "step": 10120 + }, + { + "epoch": 0.3240829652391012, + "grad_norm": 0.4955224096775055, + "learning_rate": 4.250467220786204e-05, + "loss": 0.4617, + "num_input_tokens_seen": 15407280, + "step": 10125 + }, + { + "epoch": 0.3242430062095897, + "grad_norm": 0.8797943592071533, + "learning_rate": 4.249766157626409e-05, + "loss": 0.5708, + "num_input_tokens_seen": 15415216, + "step": 10130 + }, + { + "epoch": 0.3244030471800781, + "grad_norm": 0.6671698093414307, + "learning_rate": 4.249064824630684e-05, + "loss": 0.5627, + "num_input_tokens_seen": 15422528, + "step": 10135 + }, + { + "epoch": 0.32456308815056656, + "grad_norm": 0.8234943747520447, + "learning_rate": 4.248363221907183e-05, + "loss": 0.525, + "num_input_tokens_seen": 15430320, + "step": 10140 + }, + { + "epoch": 0.324723129121055, + "grad_norm": 0.8474070429801941, + "learning_rate": 4.2476613495641026e-05, + "loss": 0.4901, + "num_input_tokens_seen": 15437680, + "step": 10145 + }, + { + "epoch": 0.32488317009154344, + "grad_norm": 0.46365392208099365, + "learning_rate": 4.246959207709679e-05, + "loss": 0.4652, + "num_input_tokens_seen": 15445056, + "step": 10150 + }, + { + "epoch": 0.3250432110620319, + "grad_norm": 0.6546779274940491, + "learning_rate": 4.246256796452192e-05, + "loss": 0.3571, + "num_input_tokens_seen": 15452480, + "step": 10155 + }, + { + "epoch": 0.3252032520325203, + "grad_norm": 1.0205620527267456, + "learning_rate": 4.245554115899962e-05, + "loss": 0.46, + "num_input_tokens_seen": 15459904, + "step": 10160 + }, + { + "epoch": 0.32536329300300876, + "grad_norm": 0.9734955430030823, + "learning_rate": 4.2448511661613514e-05, + "loss": 0.621, + "num_input_tokens_seen": 15467488, + "step": 10165 + }, + { + "epoch": 0.3255233339734972, + "grad_norm": 1.1332288980484009, + "learning_rate": 4.2441479473447635e-05, + "loss": 0.5034, + "num_input_tokens_seen": 15475376, + "step": 10170 + }, + { + "epoch": 0.32568337494398564, + "grad_norm": 0.8220967054367065, + "learning_rate": 4.243444459558644e-05, + "loss": 0.5975, + "num_input_tokens_seen": 15483520, + "step": 10175 + }, + { + "epoch": 0.32584341591447413, + "grad_norm": 1.1436412334442139, + "learning_rate": 4.24274070291148e-05, + "loss": 0.5353, + "num_input_tokens_seen": 15490944, + "step": 10180 + }, + { + "epoch": 0.3260034568849626, + "grad_norm": 0.561218798160553, + "learning_rate": 4.242036677511798e-05, + "loss": 0.4672, + "num_input_tokens_seen": 15498224, + "step": 10185 + }, + { + "epoch": 0.326163497855451, + "grad_norm": 0.5845983624458313, + "learning_rate": 4.241332383468169e-05, + "loss": 0.7, + "num_input_tokens_seen": 15506192, + "step": 10190 + }, + { + "epoch": 0.32632353882593945, + "grad_norm": 0.5840350985527039, + "learning_rate": 4.2406278208892034e-05, + "loss": 0.4629, + "num_input_tokens_seen": 15513936, + "step": 10195 + }, + { + "epoch": 0.3264835797964279, + "grad_norm": 0.45218080282211304, + "learning_rate": 4.2399229898835536e-05, + "loss": 0.624, + "num_input_tokens_seen": 15521408, + "step": 10200 + }, + { + "epoch": 0.3264835797964279, + "eval_loss": 0.5614072680473328, + "eval_runtime": 560.7509, + "eval_samples_per_second": 24.763, + "eval_steps_per_second": 12.382, + "num_input_tokens_seen": 15521408, + "step": 10200 + }, + { + "epoch": 0.32664362076691633, + "grad_norm": 0.6741349101066589, + "learning_rate": 4.239217890559914e-05, + "loss": 0.4727, + "num_input_tokens_seen": 15529088, + "step": 10205 + }, + { + "epoch": 0.32680366173740477, + "grad_norm": 1.1546159982681274, + "learning_rate": 4.238512523027019e-05, + "loss": 0.5886, + "num_input_tokens_seen": 15536672, + "step": 10210 + }, + { + "epoch": 0.3269637027078932, + "grad_norm": 0.5789036154747009, + "learning_rate": 4.237806887393645e-05, + "loss": 0.4491, + "num_input_tokens_seen": 15544400, + "step": 10215 + }, + { + "epoch": 0.32712374367838165, + "grad_norm": 0.5278497934341431, + "learning_rate": 4.237100983768611e-05, + "loss": 0.4798, + "num_input_tokens_seen": 15552288, + "step": 10220 + }, + { + "epoch": 0.3272837846488701, + "grad_norm": 0.9363940954208374, + "learning_rate": 4.2363948122607756e-05, + "loss": 0.5649, + "num_input_tokens_seen": 15559680, + "step": 10225 + }, + { + "epoch": 0.32744382561935853, + "grad_norm": 1.2886009216308594, + "learning_rate": 4.235688372979039e-05, + "loss": 0.5871, + "num_input_tokens_seen": 15566768, + "step": 10230 + }, + { + "epoch": 0.327603866589847, + "grad_norm": 1.2535945177078247, + "learning_rate": 4.234981666032343e-05, + "loss": 0.6055, + "num_input_tokens_seen": 15574448, + "step": 10235 + }, + { + "epoch": 0.32776390756033547, + "grad_norm": 0.5370572805404663, + "learning_rate": 4.2342746915296704e-05, + "loss": 0.4661, + "num_input_tokens_seen": 15582560, + "step": 10240 + }, + { + "epoch": 0.3279239485308239, + "grad_norm": 0.5461995005607605, + "learning_rate": 4.233567449580047e-05, + "loss": 0.4002, + "num_input_tokens_seen": 15590096, + "step": 10245 + }, + { + "epoch": 0.32808398950131235, + "grad_norm": 0.5592359304428101, + "learning_rate": 4.232859940292537e-05, + "loss": 0.3828, + "num_input_tokens_seen": 15597232, + "step": 10250 + }, + { + "epoch": 0.3282440304718008, + "grad_norm": 0.690271258354187, + "learning_rate": 4.232152163776248e-05, + "loss": 0.6168, + "num_input_tokens_seen": 15605312, + "step": 10255 + }, + { + "epoch": 0.3284040714422892, + "grad_norm": 1.0638247728347778, + "learning_rate": 4.231444120140328e-05, + "loss": 0.653, + "num_input_tokens_seen": 15612768, + "step": 10260 + }, + { + "epoch": 0.32856411241277766, + "grad_norm": 0.9047242999076843, + "learning_rate": 4.230735809493967e-05, + "loss": 0.6849, + "num_input_tokens_seen": 15620432, + "step": 10265 + }, + { + "epoch": 0.3287241533832661, + "grad_norm": 0.9307171702384949, + "learning_rate": 4.2300272319463926e-05, + "loss": 0.5451, + "num_input_tokens_seen": 15628208, + "step": 10270 + }, + { + "epoch": 0.32888419435375454, + "grad_norm": 0.6032224297523499, + "learning_rate": 4.2293183876068786e-05, + "loss": 0.4688, + "num_input_tokens_seen": 15635360, + "step": 10275 + }, + { + "epoch": 0.329044235324243, + "grad_norm": 0.7336350083351135, + "learning_rate": 4.228609276584737e-05, + "loss": 0.4269, + "num_input_tokens_seen": 15643216, + "step": 10280 + }, + { + "epoch": 0.3292042762947314, + "grad_norm": 0.8285170793533325, + "learning_rate": 4.227899898989323e-05, + "loss": 0.4185, + "num_input_tokens_seen": 15650944, + "step": 10285 + }, + { + "epoch": 0.3293643172652199, + "grad_norm": 0.6845345497131348, + "learning_rate": 4.2271902549300293e-05, + "loss": 0.5424, + "num_input_tokens_seen": 15658864, + "step": 10290 + }, + { + "epoch": 0.32952435823570836, + "grad_norm": 0.6764189600944519, + "learning_rate": 4.226480344516294e-05, + "loss": 0.5517, + "num_input_tokens_seen": 15666400, + "step": 10295 + }, + { + "epoch": 0.3296843992061968, + "grad_norm": 0.9235125184059143, + "learning_rate": 4.2257701678575925e-05, + "loss": 0.4717, + "num_input_tokens_seen": 15674256, + "step": 10300 + }, + { + "epoch": 0.32984444017668524, + "grad_norm": 0.8494036197662354, + "learning_rate": 4.225059725063444e-05, + "loss": 0.5613, + "num_input_tokens_seen": 15681840, + "step": 10305 + }, + { + "epoch": 0.3300044811471737, + "grad_norm": 0.7710822820663452, + "learning_rate": 4.2243490162434074e-05, + "loss": 0.43, + "num_input_tokens_seen": 15689312, + "step": 10310 + }, + { + "epoch": 0.3301645221176621, + "grad_norm": 0.8317019939422607, + "learning_rate": 4.223638041507083e-05, + "loss": 0.4184, + "num_input_tokens_seen": 15697120, + "step": 10315 + }, + { + "epoch": 0.33032456308815056, + "grad_norm": 2.198622703552246, + "learning_rate": 4.2229268009641124e-05, + "loss": 0.5673, + "num_input_tokens_seen": 15704400, + "step": 10320 + }, + { + "epoch": 0.330484604058639, + "grad_norm": 0.4740374684333801, + "learning_rate": 4.222215294724177e-05, + "loss": 0.5396, + "num_input_tokens_seen": 15711568, + "step": 10325 + }, + { + "epoch": 0.33064464502912744, + "grad_norm": 0.9407172203063965, + "learning_rate": 4.2215035228970005e-05, + "loss": 0.4377, + "num_input_tokens_seen": 15719696, + "step": 10330 + }, + { + "epoch": 0.3308046859996159, + "grad_norm": 0.7884413003921509, + "learning_rate": 4.2207914855923464e-05, + "loss": 0.4386, + "num_input_tokens_seen": 15727104, + "step": 10335 + }, + { + "epoch": 0.3309647269701044, + "grad_norm": 1.1659858226776123, + "learning_rate": 4.220079182920021e-05, + "loss": 0.5457, + "num_input_tokens_seen": 15734416, + "step": 10340 + }, + { + "epoch": 0.3311247679405928, + "grad_norm": 0.8774840831756592, + "learning_rate": 4.2193666149898705e-05, + "loss": 0.4816, + "num_input_tokens_seen": 15741824, + "step": 10345 + }, + { + "epoch": 0.33128480891108125, + "grad_norm": 1.2862240076065063, + "learning_rate": 4.21865378191178e-05, + "loss": 0.485, + "num_input_tokens_seen": 15749792, + "step": 10350 + }, + { + "epoch": 0.3314448498815697, + "grad_norm": 0.8154283165931702, + "learning_rate": 4.217940683795678e-05, + "loss": 0.6523, + "num_input_tokens_seen": 15757440, + "step": 10355 + }, + { + "epoch": 0.33160489085205813, + "grad_norm": 0.8457381725311279, + "learning_rate": 4.217227320751534e-05, + "loss": 0.6284, + "num_input_tokens_seen": 15765024, + "step": 10360 + }, + { + "epoch": 0.33176493182254657, + "grad_norm": 0.8690844774246216, + "learning_rate": 4.216513692889358e-05, + "loss": 0.453, + "num_input_tokens_seen": 15772304, + "step": 10365 + }, + { + "epoch": 0.331924972793035, + "grad_norm": 0.7506985664367676, + "learning_rate": 4.215799800319199e-05, + "loss": 0.5157, + "num_input_tokens_seen": 15780112, + "step": 10370 + }, + { + "epoch": 0.33208501376352345, + "grad_norm": 1.0376948118209839, + "learning_rate": 4.2150856431511485e-05, + "loss": 0.6429, + "num_input_tokens_seen": 15787648, + "step": 10375 + }, + { + "epoch": 0.3322450547340119, + "grad_norm": 0.9537982940673828, + "learning_rate": 4.214371221495339e-05, + "loss": 0.6085, + "num_input_tokens_seen": 15794704, + "step": 10380 + }, + { + "epoch": 0.33240509570450033, + "grad_norm": 0.9246692657470703, + "learning_rate": 4.213656535461942e-05, + "loss": 0.5582, + "num_input_tokens_seen": 15802368, + "step": 10385 + }, + { + "epoch": 0.33256513667498877, + "grad_norm": 0.4558405578136444, + "learning_rate": 4.2129415851611734e-05, + "loss": 0.5873, + "num_input_tokens_seen": 15809952, + "step": 10390 + }, + { + "epoch": 0.33272517764547727, + "grad_norm": 0.522418200969696, + "learning_rate": 4.2122263707032855e-05, + "loss": 0.3961, + "num_input_tokens_seen": 15817760, + "step": 10395 + }, + { + "epoch": 0.3328852186159657, + "grad_norm": 0.9836182594299316, + "learning_rate": 4.211510892198574e-05, + "loss": 0.5232, + "num_input_tokens_seen": 15825776, + "step": 10400 + }, + { + "epoch": 0.3328852186159657, + "eval_loss": 0.5594174861907959, + "eval_runtime": 559.2376, + "eval_samples_per_second": 24.83, + "eval_steps_per_second": 12.415, + "num_input_tokens_seen": 15825776, + "step": 10400 + }, + { + "epoch": 0.33304525958645415, + "grad_norm": 0.9638216495513916, + "learning_rate": 4.210795149757375e-05, + "loss": 0.538, + "num_input_tokens_seen": 15833088, + "step": 10405 + }, + { + "epoch": 0.3332053005569426, + "grad_norm": 0.7479906678199768, + "learning_rate": 4.210079143490065e-05, + "loss": 0.4383, + "num_input_tokens_seen": 15840960, + "step": 10410 + }, + { + "epoch": 0.333365341527431, + "grad_norm": 0.9833614230155945, + "learning_rate": 4.2093628735070604e-05, + "loss": 0.4818, + "num_input_tokens_seen": 15848768, + "step": 10415 + }, + { + "epoch": 0.33352538249791946, + "grad_norm": 1.577903151512146, + "learning_rate": 4.208646339918819e-05, + "loss": 0.5593, + "num_input_tokens_seen": 15856320, + "step": 10420 + }, + { + "epoch": 0.3336854234684079, + "grad_norm": 1.557826280593872, + "learning_rate": 4.2079295428358414e-05, + "loss": 0.5793, + "num_input_tokens_seen": 15863728, + "step": 10425 + }, + { + "epoch": 0.33384546443889634, + "grad_norm": 1.3290960788726807, + "learning_rate": 4.207212482368664e-05, + "loss": 0.6832, + "num_input_tokens_seen": 15871488, + "step": 10430 + }, + { + "epoch": 0.3340055054093848, + "grad_norm": 0.552210807800293, + "learning_rate": 4.206495158627867e-05, + "loss": 0.4906, + "num_input_tokens_seen": 15878528, + "step": 10435 + }, + { + "epoch": 0.3341655463798732, + "grad_norm": 0.6019265055656433, + "learning_rate": 4.205777571724073e-05, + "loss": 0.5282, + "num_input_tokens_seen": 15886304, + "step": 10440 + }, + { + "epoch": 0.3343255873503617, + "grad_norm": 0.5756690502166748, + "learning_rate": 4.20505972176794e-05, + "loss": 0.4992, + "num_input_tokens_seen": 15893952, + "step": 10445 + }, + { + "epoch": 0.33448562832085016, + "grad_norm": 1.0742101669311523, + "learning_rate": 4.204341608870171e-05, + "loss": 0.4878, + "num_input_tokens_seen": 15901456, + "step": 10450 + }, + { + "epoch": 0.3346456692913386, + "grad_norm": 0.8466203212738037, + "learning_rate": 4.203623233141508e-05, + "loss": 0.5804, + "num_input_tokens_seen": 15908944, + "step": 10455 + }, + { + "epoch": 0.33480571026182704, + "grad_norm": 0.7608335018157959, + "learning_rate": 4.2029045946927334e-05, + "loss": 0.5057, + "num_input_tokens_seen": 15916336, + "step": 10460 + }, + { + "epoch": 0.3349657512323155, + "grad_norm": 1.5741499662399292, + "learning_rate": 4.20218569363467e-05, + "loss": 0.5715, + "num_input_tokens_seen": 15924160, + "step": 10465 + }, + { + "epoch": 0.3351257922028039, + "grad_norm": 0.5804333090782166, + "learning_rate": 4.2014665300781834e-05, + "loss": 0.4669, + "num_input_tokens_seen": 15932208, + "step": 10470 + }, + { + "epoch": 0.33528583317329236, + "grad_norm": 0.7613980174064636, + "learning_rate": 4.200747104134174e-05, + "loss": 0.5231, + "num_input_tokens_seen": 15939952, + "step": 10475 + }, + { + "epoch": 0.3354458741437808, + "grad_norm": 0.5562503337860107, + "learning_rate": 4.200027415913588e-05, + "loss": 0.6648, + "num_input_tokens_seen": 15947520, + "step": 10480 + }, + { + "epoch": 0.33560591511426924, + "grad_norm": 1.358333945274353, + "learning_rate": 4.1993074655274126e-05, + "loss": 0.5429, + "num_input_tokens_seen": 15955712, + "step": 10485 + }, + { + "epoch": 0.3357659560847577, + "grad_norm": 0.5844719409942627, + "learning_rate": 4.198587253086669e-05, + "loss": 0.4664, + "num_input_tokens_seen": 15963792, + "step": 10490 + }, + { + "epoch": 0.3359259970552461, + "grad_norm": 0.5432619452476501, + "learning_rate": 4.197866778702426e-05, + "loss": 0.5611, + "num_input_tokens_seen": 15971424, + "step": 10495 + }, + { + "epoch": 0.3360860380257346, + "grad_norm": 0.5433865189552307, + "learning_rate": 4.197146042485789e-05, + "loss": 0.5134, + "num_input_tokens_seen": 15978928, + "step": 10500 + }, + { + "epoch": 0.33624607899622305, + "grad_norm": 0.6963246464729309, + "learning_rate": 4.1964250445479046e-05, + "loss": 0.6966, + "num_input_tokens_seen": 15986592, + "step": 10505 + }, + { + "epoch": 0.3364061199667115, + "grad_norm": 1.0416139364242554, + "learning_rate": 4.19570378499996e-05, + "loss": 0.5904, + "num_input_tokens_seen": 15994864, + "step": 10510 + }, + { + "epoch": 0.33656616093719993, + "grad_norm": 0.9859049320220947, + "learning_rate": 4.194982263953182e-05, + "loss": 0.4808, + "num_input_tokens_seen": 16002144, + "step": 10515 + }, + { + "epoch": 0.33672620190768837, + "grad_norm": 0.6316773891448975, + "learning_rate": 4.194260481518838e-05, + "loss": 0.6101, + "num_input_tokens_seen": 16009360, + "step": 10520 + }, + { + "epoch": 0.3368862428781768, + "grad_norm": 0.4928540885448456, + "learning_rate": 4.1935384378082366e-05, + "loss": 0.5594, + "num_input_tokens_seen": 16016976, + "step": 10525 + }, + { + "epoch": 0.33704628384866525, + "grad_norm": 0.7071569561958313, + "learning_rate": 4.1928161329327267e-05, + "loss": 0.4274, + "num_input_tokens_seen": 16025248, + "step": 10530 + }, + { + "epoch": 0.3372063248191537, + "grad_norm": 0.647937536239624, + "learning_rate": 4.1920935670036945e-05, + "loss": 0.4129, + "num_input_tokens_seen": 16033360, + "step": 10535 + }, + { + "epoch": 0.33736636578964213, + "grad_norm": 1.3723642826080322, + "learning_rate": 4.1913707401325705e-05, + "loss": 0.4814, + "num_input_tokens_seen": 16040992, + "step": 10540 + }, + { + "epoch": 0.33752640676013057, + "grad_norm": 0.7957385182380676, + "learning_rate": 4.1906476524308235e-05, + "loss": 0.42, + "num_input_tokens_seen": 16048480, + "step": 10545 + }, + { + "epoch": 0.33768644773061907, + "grad_norm": 1.0943312644958496, + "learning_rate": 4.189924304009962e-05, + "loss": 0.6501, + "num_input_tokens_seen": 16056288, + "step": 10550 + }, + { + "epoch": 0.3378464887011075, + "grad_norm": 0.8387945890426636, + "learning_rate": 4.189200694981537e-05, + "loss": 0.5574, + "num_input_tokens_seen": 16063744, + "step": 10555 + }, + { + "epoch": 0.33800652967159595, + "grad_norm": 0.4415355622768402, + "learning_rate": 4.188476825457136e-05, + "loss": 0.3807, + "num_input_tokens_seen": 16071440, + "step": 10560 + }, + { + "epoch": 0.3381665706420844, + "grad_norm": 0.9616689085960388, + "learning_rate": 4.18775269554839e-05, + "loss": 0.5687, + "num_input_tokens_seen": 16078928, + "step": 10565 + }, + { + "epoch": 0.3383266116125728, + "grad_norm": 1.0661051273345947, + "learning_rate": 4.187028305366969e-05, + "loss": 0.4406, + "num_input_tokens_seen": 16086992, + "step": 10570 + }, + { + "epoch": 0.33848665258306126, + "grad_norm": 0.7305091619491577, + "learning_rate": 4.1863036550245824e-05, + "loss": 0.4857, + "num_input_tokens_seen": 16094832, + "step": 10575 + }, + { + "epoch": 0.3386466935535497, + "grad_norm": 0.8685897588729858, + "learning_rate": 4.1855787446329806e-05, + "loss": 0.4306, + "num_input_tokens_seen": 16102576, + "step": 10580 + }, + { + "epoch": 0.33880673452403814, + "grad_norm": 0.8396205306053162, + "learning_rate": 4.184853574303955e-05, + "loss": 0.6005, + "num_input_tokens_seen": 16109888, + "step": 10585 + }, + { + "epoch": 0.3389667754945266, + "grad_norm": 0.9156930446624756, + "learning_rate": 4.184128144149334e-05, + "loss": 0.6117, + "num_input_tokens_seen": 16117616, + "step": 10590 + }, + { + "epoch": 0.339126816465015, + "grad_norm": 0.5003269910812378, + "learning_rate": 4.1834024542809896e-05, + "loss": 0.4891, + "num_input_tokens_seen": 16124960, + "step": 10595 + }, + { + "epoch": 0.33928685743550346, + "grad_norm": 0.6119455695152283, + "learning_rate": 4.1826765048108315e-05, + "loss": 0.6099, + "num_input_tokens_seen": 16132208, + "step": 10600 + }, + { + "epoch": 0.33928685743550346, + "eval_loss": 0.5583112239837646, + "eval_runtime": 558.3096, + "eval_samples_per_second": 24.872, + "eval_steps_per_second": 12.436, + "num_input_tokens_seen": 16132208, + "step": 10600 + }, + { + "epoch": 0.33944689840599196, + "grad_norm": 1.1144686937332153, + "learning_rate": 4.181950295850811e-05, + "loss": 0.5249, + "num_input_tokens_seen": 16139440, + "step": 10605 + }, + { + "epoch": 0.3396069393764804, + "grad_norm": 1.1618061065673828, + "learning_rate": 4.181223827512918e-05, + "loss": 0.5346, + "num_input_tokens_seen": 16146832, + "step": 10610 + }, + { + "epoch": 0.33976698034696884, + "grad_norm": 0.6622742414474487, + "learning_rate": 4.180497099909183e-05, + "loss": 0.5495, + "num_input_tokens_seen": 16154416, + "step": 10615 + }, + { + "epoch": 0.3399270213174573, + "grad_norm": 1.3280125856399536, + "learning_rate": 4.179770113151677e-05, + "loss": 0.5835, + "num_input_tokens_seen": 16161936, + "step": 10620 + }, + { + "epoch": 0.3400870622879457, + "grad_norm": 1.1587566137313843, + "learning_rate": 4.179042867352511e-05, + "loss": 0.4387, + "num_input_tokens_seen": 16169600, + "step": 10625 + }, + { + "epoch": 0.34024710325843416, + "grad_norm": 1.3389910459518433, + "learning_rate": 4.1783153626238334e-05, + "loss": 0.6973, + "num_input_tokens_seen": 16176992, + "step": 10630 + }, + { + "epoch": 0.3404071442289226, + "grad_norm": 0.8962140679359436, + "learning_rate": 4.177587599077836e-05, + "loss": 0.812, + "num_input_tokens_seen": 16184704, + "step": 10635 + }, + { + "epoch": 0.34056718519941104, + "grad_norm": 0.5128433704376221, + "learning_rate": 4.1768595768267494e-05, + "loss": 0.6192, + "num_input_tokens_seen": 16192784, + "step": 10640 + }, + { + "epoch": 0.3407272261698995, + "grad_norm": 0.5267797708511353, + "learning_rate": 4.176131295982843e-05, + "loss": 0.3615, + "num_input_tokens_seen": 16199904, + "step": 10645 + }, + { + "epoch": 0.3408872671403879, + "grad_norm": 1.3292505741119385, + "learning_rate": 4.1754027566584276e-05, + "loss": 0.5448, + "num_input_tokens_seen": 16207456, + "step": 10650 + }, + { + "epoch": 0.34104730811087636, + "grad_norm": 0.8925579786300659, + "learning_rate": 4.174673958965852e-05, + "loss": 0.5711, + "num_input_tokens_seen": 16215504, + "step": 10655 + }, + { + "epoch": 0.34120734908136485, + "grad_norm": 1.4057978391647339, + "learning_rate": 4.173944903017507e-05, + "loss": 0.5112, + "num_input_tokens_seen": 16223008, + "step": 10660 + }, + { + "epoch": 0.3413673900518533, + "grad_norm": 0.7157074809074402, + "learning_rate": 4.173215588925822e-05, + "loss": 0.5557, + "num_input_tokens_seen": 16230288, + "step": 10665 + }, + { + "epoch": 0.34152743102234173, + "grad_norm": 0.5510793328285217, + "learning_rate": 4.172486016803266e-05, + "loss": 0.5431, + "num_input_tokens_seen": 16237984, + "step": 10670 + }, + { + "epoch": 0.34168747199283017, + "grad_norm": 0.5322859883308411, + "learning_rate": 4.171756186762349e-05, + "loss": 0.5007, + "num_input_tokens_seen": 16245632, + "step": 10675 + }, + { + "epoch": 0.3418475129633186, + "grad_norm": 0.49249792098999023, + "learning_rate": 4.171026098915619e-05, + "loss": 0.615, + "num_input_tokens_seen": 16253008, + "step": 10680 + }, + { + "epoch": 0.34200755393380705, + "grad_norm": 0.7024332880973816, + "learning_rate": 4.170295753375665e-05, + "loss": 0.6069, + "num_input_tokens_seen": 16260528, + "step": 10685 + }, + { + "epoch": 0.3421675949042955, + "grad_norm": 0.7451575398445129, + "learning_rate": 4.169565150255117e-05, + "loss": 0.3783, + "num_input_tokens_seen": 16268656, + "step": 10690 + }, + { + "epoch": 0.34232763587478393, + "grad_norm": 0.6069150567054749, + "learning_rate": 4.16883428966664e-05, + "loss": 0.426, + "num_input_tokens_seen": 16276160, + "step": 10695 + }, + { + "epoch": 0.34248767684527237, + "grad_norm": 1.3756279945373535, + "learning_rate": 4.168103171722944e-05, + "loss": 0.6835, + "num_input_tokens_seen": 16283568, + "step": 10700 + }, + { + "epoch": 0.3426477178157608, + "grad_norm": 2.227795124053955, + "learning_rate": 4.167371796536777e-05, + "loss": 0.6071, + "num_input_tokens_seen": 16291456, + "step": 10705 + }, + { + "epoch": 0.3428077587862493, + "grad_norm": 1.001258373260498, + "learning_rate": 4.166640164220924e-05, + "loss": 0.4772, + "num_input_tokens_seen": 16299504, + "step": 10710 + }, + { + "epoch": 0.34296779975673775, + "grad_norm": 0.92508465051651, + "learning_rate": 4.1659082748882144e-05, + "loss": 0.6097, + "num_input_tokens_seen": 16307152, + "step": 10715 + }, + { + "epoch": 0.3431278407272262, + "grad_norm": 0.40093302726745605, + "learning_rate": 4.1651761286515135e-05, + "loss": 0.6648, + "num_input_tokens_seen": 16314752, + "step": 10720 + }, + { + "epoch": 0.3432878816977146, + "grad_norm": 0.8413928151130676, + "learning_rate": 4.164443725623728e-05, + "loss": 0.6161, + "num_input_tokens_seen": 16321824, + "step": 10725 + }, + { + "epoch": 0.34344792266820307, + "grad_norm": 0.5026217699050903, + "learning_rate": 4.163711065917802e-05, + "loss": 0.4614, + "num_input_tokens_seen": 16329472, + "step": 10730 + }, + { + "epoch": 0.3436079636386915, + "grad_norm": 0.675053596496582, + "learning_rate": 4.1629781496467234e-05, + "loss": 0.4884, + "num_input_tokens_seen": 16337328, + "step": 10735 + }, + { + "epoch": 0.34376800460917994, + "grad_norm": 0.6429606676101685, + "learning_rate": 4.1622449769235164e-05, + "loss": 0.6497, + "num_input_tokens_seen": 16344896, + "step": 10740 + }, + { + "epoch": 0.3439280455796684, + "grad_norm": 0.9005915522575378, + "learning_rate": 4.161511547861243e-05, + "loss": 0.4598, + "num_input_tokens_seen": 16352624, + "step": 10745 + }, + { + "epoch": 0.3440880865501568, + "grad_norm": 0.6346790194511414, + "learning_rate": 4.1607778625730104e-05, + "loss": 0.3941, + "num_input_tokens_seen": 16360224, + "step": 10750 + }, + { + "epoch": 0.34424812752064526, + "grad_norm": 0.71433424949646, + "learning_rate": 4.160043921171961e-05, + "loss": 0.4364, + "num_input_tokens_seen": 16368192, + "step": 10755 + }, + { + "epoch": 0.3444081684911337, + "grad_norm": 1.9666386842727661, + "learning_rate": 4.159309723771276e-05, + "loss": 0.5761, + "num_input_tokens_seen": 16375744, + "step": 10760 + }, + { + "epoch": 0.3445682094616222, + "grad_norm": 1.196095585823059, + "learning_rate": 4.158575270484181e-05, + "loss": 0.6097, + "num_input_tokens_seen": 16382752, + "step": 10765 + }, + { + "epoch": 0.34472825043211064, + "grad_norm": 0.5928419828414917, + "learning_rate": 4.157840561423936e-05, + "loss": 0.7831, + "num_input_tokens_seen": 16389952, + "step": 10770 + }, + { + "epoch": 0.3448882914025991, + "grad_norm": 0.41283345222473145, + "learning_rate": 4.1571055967038416e-05, + "loss": 0.4418, + "num_input_tokens_seen": 16397168, + "step": 10775 + }, + { + "epoch": 0.3450483323730875, + "grad_norm": 0.6470280289649963, + "learning_rate": 4.156370376437241e-05, + "loss": 0.3625, + "num_input_tokens_seen": 16404624, + "step": 10780 + }, + { + "epoch": 0.34520837334357596, + "grad_norm": 0.9870709180831909, + "learning_rate": 4.155634900737513e-05, + "loss": 0.5489, + "num_input_tokens_seen": 16412032, + "step": 10785 + }, + { + "epoch": 0.3453684143140644, + "grad_norm": 0.7049941420555115, + "learning_rate": 4.1548991697180764e-05, + "loss": 0.4528, + "num_input_tokens_seen": 16419104, + "step": 10790 + }, + { + "epoch": 0.34552845528455284, + "grad_norm": 1.025650143623352, + "learning_rate": 4.1541631834923914e-05, + "loss": 0.6814, + "num_input_tokens_seen": 16426528, + "step": 10795 + }, + { + "epoch": 0.3456884962550413, + "grad_norm": 0.8142889738082886, + "learning_rate": 4.153426942173956e-05, + "loss": 0.5191, + "num_input_tokens_seen": 16434512, + "step": 10800 + }, + { + "epoch": 0.3456884962550413, + "eval_loss": 0.5565073490142822, + "eval_runtime": 561.8576, + "eval_samples_per_second": 24.714, + "eval_steps_per_second": 12.357, + "num_input_tokens_seen": 16434512, + "step": 10800 + }, + { + "epoch": 0.3458485372255297, + "grad_norm": 1.2437843084335327, + "learning_rate": 4.152690445876308e-05, + "loss": 0.569, + "num_input_tokens_seen": 16442496, + "step": 10805 + }, + { + "epoch": 0.34600857819601816, + "grad_norm": 1.1902377605438232, + "learning_rate": 4.1519536947130245e-05, + "loss": 0.5183, + "num_input_tokens_seen": 16450432, + "step": 10810 + }, + { + "epoch": 0.34616861916650665, + "grad_norm": 1.4892574548721313, + "learning_rate": 4.151216688797722e-05, + "loss": 0.6926, + "num_input_tokens_seen": 16457664, + "step": 10815 + }, + { + "epoch": 0.3463286601369951, + "grad_norm": 0.5312864780426025, + "learning_rate": 4.150479428244054e-05, + "loss": 0.569, + "num_input_tokens_seen": 16464864, + "step": 10820 + }, + { + "epoch": 0.34648870110748353, + "grad_norm": 0.7329182028770447, + "learning_rate": 4.1497419131657176e-05, + "loss": 0.409, + "num_input_tokens_seen": 16472112, + "step": 10825 + }, + { + "epoch": 0.34664874207797197, + "grad_norm": 0.7040711045265198, + "learning_rate": 4.149004143676447e-05, + "loss": 0.4865, + "num_input_tokens_seen": 16479680, + "step": 10830 + }, + { + "epoch": 0.3468087830484604, + "grad_norm": 1.217433214187622, + "learning_rate": 4.148266119890015e-05, + "loss": 0.5892, + "num_input_tokens_seen": 16487632, + "step": 10835 + }, + { + "epoch": 0.34696882401894885, + "grad_norm": 0.891021728515625, + "learning_rate": 4.1475278419202324e-05, + "loss": 0.474, + "num_input_tokens_seen": 16495424, + "step": 10840 + }, + { + "epoch": 0.3471288649894373, + "grad_norm": 0.6100066304206848, + "learning_rate": 4.146789309880953e-05, + "loss": 0.5012, + "num_input_tokens_seen": 16502896, + "step": 10845 + }, + { + "epoch": 0.34728890595992573, + "grad_norm": 0.6673109531402588, + "learning_rate": 4.146050523886068e-05, + "loss": 0.9028, + "num_input_tokens_seen": 16510704, + "step": 10850 + }, + { + "epoch": 0.34744894693041417, + "grad_norm": 0.6317422389984131, + "learning_rate": 4.1453114840495055e-05, + "loss": 0.5947, + "num_input_tokens_seen": 16517984, + "step": 10855 + }, + { + "epoch": 0.3476089879009026, + "grad_norm": 0.7593173980712891, + "learning_rate": 4.1445721904852364e-05, + "loss": 0.4635, + "num_input_tokens_seen": 16525696, + "step": 10860 + }, + { + "epoch": 0.34776902887139105, + "grad_norm": 0.7136397361755371, + "learning_rate": 4.143832643307269e-05, + "loss": 0.5809, + "num_input_tokens_seen": 16532784, + "step": 10865 + }, + { + "epoch": 0.34792906984187955, + "grad_norm": 0.9660236835479736, + "learning_rate": 4.1430928426296503e-05, + "loss": 0.545, + "num_input_tokens_seen": 16540880, + "step": 10870 + }, + { + "epoch": 0.348089110812368, + "grad_norm": 0.37641093134880066, + "learning_rate": 4.142352788566466e-05, + "loss": 0.4449, + "num_input_tokens_seen": 16548960, + "step": 10875 + }, + { + "epoch": 0.3482491517828564, + "grad_norm": 1.5475209951400757, + "learning_rate": 4.1416124812318424e-05, + "loss": 0.5239, + "num_input_tokens_seen": 16556224, + "step": 10880 + }, + { + "epoch": 0.34840919275334487, + "grad_norm": 1.3160513639450073, + "learning_rate": 4.1408719207399453e-05, + "loss": 0.4521, + "num_input_tokens_seen": 16563856, + "step": 10885 + }, + { + "epoch": 0.3485692337238333, + "grad_norm": 1.972055435180664, + "learning_rate": 4.140131107204978e-05, + "loss": 0.5008, + "num_input_tokens_seen": 16571696, + "step": 10890 + }, + { + "epoch": 0.34872927469432174, + "grad_norm": 0.9014578461647034, + "learning_rate": 4.139390040741182e-05, + "loss": 0.5197, + "num_input_tokens_seen": 16579232, + "step": 10895 + }, + { + "epoch": 0.3488893156648102, + "grad_norm": 0.3170199394226074, + "learning_rate": 4.1386487214628396e-05, + "loss": 0.3451, + "num_input_tokens_seen": 16586544, + "step": 10900 + }, + { + "epoch": 0.3490493566352986, + "grad_norm": 1.5557868480682373, + "learning_rate": 4.137907149484272e-05, + "loss": 0.642, + "num_input_tokens_seen": 16594464, + "step": 10905 + }, + { + "epoch": 0.34920939760578706, + "grad_norm": 0.5882148742675781, + "learning_rate": 4.137165324919839e-05, + "loss": 0.5546, + "num_input_tokens_seen": 16601712, + "step": 10910 + }, + { + "epoch": 0.3493694385762755, + "grad_norm": 1.1828954219818115, + "learning_rate": 4.136423247883939e-05, + "loss": 0.5309, + "num_input_tokens_seen": 16609552, + "step": 10915 + }, + { + "epoch": 0.349529479546764, + "grad_norm": 0.6679588556289673, + "learning_rate": 4.135680918491009e-05, + "loss": 0.4278, + "num_input_tokens_seen": 16617152, + "step": 10920 + }, + { + "epoch": 0.34968952051725244, + "grad_norm": 1.1295506954193115, + "learning_rate": 4.1349383368555265e-05, + "loss": 0.5843, + "num_input_tokens_seen": 16624912, + "step": 10925 + }, + { + "epoch": 0.3498495614877409, + "grad_norm": 0.460429847240448, + "learning_rate": 4.1341955030920065e-05, + "loss": 0.4532, + "num_input_tokens_seen": 16632208, + "step": 10930 + }, + { + "epoch": 0.3500096024582293, + "grad_norm": 0.7850174903869629, + "learning_rate": 4.1334524173150036e-05, + "loss": 0.5173, + "num_input_tokens_seen": 16640224, + "step": 10935 + }, + { + "epoch": 0.35016964342871776, + "grad_norm": 0.5581339597702026, + "learning_rate": 4.13270907963911e-05, + "loss": 0.576, + "num_input_tokens_seen": 16647584, + "step": 10940 + }, + { + "epoch": 0.3503296843992062, + "grad_norm": 1.0274943113327026, + "learning_rate": 4.131965490178959e-05, + "loss": 0.534, + "num_input_tokens_seen": 16655344, + "step": 10945 + }, + { + "epoch": 0.35048972536969464, + "grad_norm": 0.9877519011497498, + "learning_rate": 4.131221649049222e-05, + "loss": 0.9132, + "num_input_tokens_seen": 16662800, + "step": 10950 + }, + { + "epoch": 0.3506497663401831, + "grad_norm": 0.693709671497345, + "learning_rate": 4.130477556364606e-05, + "loss": 0.4677, + "num_input_tokens_seen": 16670192, + "step": 10955 + }, + { + "epoch": 0.3508098073106715, + "grad_norm": 0.8757088780403137, + "learning_rate": 4.129733212239861e-05, + "loss": 0.4029, + "num_input_tokens_seen": 16677712, + "step": 10960 + }, + { + "epoch": 0.35096984828115996, + "grad_norm": 0.968875527381897, + "learning_rate": 4.128988616789774e-05, + "loss": 0.509, + "num_input_tokens_seen": 16685488, + "step": 10965 + }, + { + "epoch": 0.3511298892516484, + "grad_norm": 2.865173816680908, + "learning_rate": 4.1282437701291724e-05, + "loss": 0.5022, + "num_input_tokens_seen": 16693216, + "step": 10970 + }, + { + "epoch": 0.3512899302221369, + "grad_norm": 0.5369619131088257, + "learning_rate": 4.1274986723729184e-05, + "loss": 0.4935, + "num_input_tokens_seen": 16700832, + "step": 10975 + }, + { + "epoch": 0.35144997119262533, + "grad_norm": 0.850152850151062, + "learning_rate": 4.126753323635917e-05, + "loss": 0.5132, + "num_input_tokens_seen": 16708064, + "step": 10980 + }, + { + "epoch": 0.35161001216311377, + "grad_norm": 1.1963754892349243, + "learning_rate": 4.12600772403311e-05, + "loss": 0.6972, + "num_input_tokens_seen": 16715536, + "step": 10985 + }, + { + "epoch": 0.3517700531336022, + "grad_norm": 1.7489068508148193, + "learning_rate": 4.125261873679479e-05, + "loss": 0.7216, + "num_input_tokens_seen": 16723104, + "step": 10990 + }, + { + "epoch": 0.35193009410409065, + "grad_norm": 1.363040566444397, + "learning_rate": 4.124515772690042e-05, + "loss": 0.7869, + "num_input_tokens_seen": 16730592, + "step": 10995 + }, + { + "epoch": 0.3520901350745791, + "grad_norm": 0.944866418838501, + "learning_rate": 4.123769421179858e-05, + "loss": 0.576, + "num_input_tokens_seen": 16737888, + "step": 11000 + }, + { + "epoch": 0.3520901350745791, + "eval_loss": 0.5549571514129639, + "eval_runtime": 561.0407, + "eval_samples_per_second": 24.75, + "eval_steps_per_second": 12.375, + "num_input_tokens_seen": 16737888, + "step": 11000 + }, + { + "epoch": 0.35225017604506753, + "grad_norm": 0.865654468536377, + "learning_rate": 4.1230228192640236e-05, + "loss": 0.4193, + "num_input_tokens_seen": 16745472, + "step": 11005 + }, + { + "epoch": 0.35241021701555597, + "grad_norm": 0.5954025983810425, + "learning_rate": 4.122275967057675e-05, + "loss": 0.4877, + "num_input_tokens_seen": 16752864, + "step": 11010 + }, + { + "epoch": 0.3525702579860444, + "grad_norm": 0.846481204032898, + "learning_rate": 4.1215288646759846e-05, + "loss": 0.513, + "num_input_tokens_seen": 16760512, + "step": 11015 + }, + { + "epoch": 0.35273029895653285, + "grad_norm": 0.5578429698944092, + "learning_rate": 4.120781512234166e-05, + "loss": 0.4672, + "num_input_tokens_seen": 16768128, + "step": 11020 + }, + { + "epoch": 0.3528903399270213, + "grad_norm": 1.1883680820465088, + "learning_rate": 4.120033909847471e-05, + "loss": 0.5534, + "num_input_tokens_seen": 16775568, + "step": 11025 + }, + { + "epoch": 0.3530503808975098, + "grad_norm": 1.1905035972595215, + "learning_rate": 4.119286057631187e-05, + "loss": 0.6025, + "num_input_tokens_seen": 16783504, + "step": 11030 + }, + { + "epoch": 0.3532104218679982, + "grad_norm": 0.739244282245636, + "learning_rate": 4.118537955700646e-05, + "loss": 0.452, + "num_input_tokens_seen": 16791040, + "step": 11035 + }, + { + "epoch": 0.35337046283848667, + "grad_norm": 0.5729882121086121, + "learning_rate": 4.11778960417121e-05, + "loss": 0.5245, + "num_input_tokens_seen": 16798320, + "step": 11040 + }, + { + "epoch": 0.3535305038089751, + "grad_norm": 0.952694296836853, + "learning_rate": 4.117041003158288e-05, + "loss": 0.5195, + "num_input_tokens_seen": 16806016, + "step": 11045 + }, + { + "epoch": 0.35369054477946354, + "grad_norm": 1.7036032676696777, + "learning_rate": 4.1162921527773215e-05, + "loss": 0.6434, + "num_input_tokens_seen": 16813392, + "step": 11050 + }, + { + "epoch": 0.353850585749952, + "grad_norm": 0.9683284759521484, + "learning_rate": 4.115543053143794e-05, + "loss": 0.7038, + "num_input_tokens_seen": 16821056, + "step": 11055 + }, + { + "epoch": 0.3540106267204404, + "grad_norm": 0.6277300119400024, + "learning_rate": 4.114793704373226e-05, + "loss": 0.6051, + "num_input_tokens_seen": 16828576, + "step": 11060 + }, + { + "epoch": 0.35417066769092886, + "grad_norm": 0.7459772825241089, + "learning_rate": 4.114044106581175e-05, + "loss": 0.6842, + "num_input_tokens_seen": 16836288, + "step": 11065 + }, + { + "epoch": 0.3543307086614173, + "grad_norm": 0.9972091913223267, + "learning_rate": 4.11329425988324e-05, + "loss": 0.4543, + "num_input_tokens_seen": 16843808, + "step": 11070 + }, + { + "epoch": 0.35449074963190574, + "grad_norm": 0.7051295042037964, + "learning_rate": 4.112544164395056e-05, + "loss": 0.5128, + "num_input_tokens_seen": 16850752, + "step": 11075 + }, + { + "epoch": 0.35465079060239424, + "grad_norm": 0.697877824306488, + "learning_rate": 4.111793820232297e-05, + "loss": 0.6371, + "num_input_tokens_seen": 16858800, + "step": 11080 + }, + { + "epoch": 0.3548108315728827, + "grad_norm": 0.9367096424102783, + "learning_rate": 4.1110432275106767e-05, + "loss": 0.467, + "num_input_tokens_seen": 16866368, + "step": 11085 + }, + { + "epoch": 0.3549708725433711, + "grad_norm": 0.6109790205955505, + "learning_rate": 4.110292386345944e-05, + "loss": 0.5575, + "num_input_tokens_seen": 16874016, + "step": 11090 + }, + { + "epoch": 0.35513091351385956, + "grad_norm": 0.6204933524131775, + "learning_rate": 4.109541296853891e-05, + "loss": 0.3976, + "num_input_tokens_seen": 16881200, + "step": 11095 + }, + { + "epoch": 0.355290954484348, + "grad_norm": 1.032313585281372, + "learning_rate": 4.108789959150341e-05, + "loss": 0.5381, + "num_input_tokens_seen": 16888752, + "step": 11100 + }, + { + "epoch": 0.35545099545483644, + "grad_norm": 0.6130318641662598, + "learning_rate": 4.108038373351163e-05, + "loss": 0.4812, + "num_input_tokens_seen": 16896112, + "step": 11105 + }, + { + "epoch": 0.3556110364253249, + "grad_norm": 0.48100078105926514, + "learning_rate": 4.10728653957226e-05, + "loss": 0.4251, + "num_input_tokens_seen": 16903760, + "step": 11110 + }, + { + "epoch": 0.3557710773958133, + "grad_norm": 0.24897941946983337, + "learning_rate": 4.106534457929575e-05, + "loss": 0.5066, + "num_input_tokens_seen": 16912240, + "step": 11115 + }, + { + "epoch": 0.35593111836630176, + "grad_norm": 1.2763686180114746, + "learning_rate": 4.105782128539086e-05, + "loss": 0.4026, + "num_input_tokens_seen": 16920288, + "step": 11120 + }, + { + "epoch": 0.3560911593367902, + "grad_norm": 0.7897359132766724, + "learning_rate": 4.1050295515168144e-05, + "loss": 0.4462, + "num_input_tokens_seen": 16927584, + "step": 11125 + }, + { + "epoch": 0.35625120030727864, + "grad_norm": 0.4576791822910309, + "learning_rate": 4.1042767269788155e-05, + "loss": 0.5789, + "num_input_tokens_seen": 16935040, + "step": 11130 + }, + { + "epoch": 0.35641124127776713, + "grad_norm": 0.9550888538360596, + "learning_rate": 4.103523655041185e-05, + "loss": 0.5081, + "num_input_tokens_seen": 16942256, + "step": 11135 + }, + { + "epoch": 0.35657128224825557, + "grad_norm": 0.8047630786895752, + "learning_rate": 4.102770335820055e-05, + "loss": 0.6003, + "num_input_tokens_seen": 16950112, + "step": 11140 + }, + { + "epoch": 0.356731323218744, + "grad_norm": 0.5341451168060303, + "learning_rate": 4.1020167694315984e-05, + "loss": 0.4627, + "num_input_tokens_seen": 16957888, + "step": 11145 + }, + { + "epoch": 0.35689136418923245, + "grad_norm": 1.0073071718215942, + "learning_rate": 4.101262955992023e-05, + "loss": 0.613, + "num_input_tokens_seen": 16965552, + "step": 11150 + }, + { + "epoch": 0.3570514051597209, + "grad_norm": 0.6285186409950256, + "learning_rate": 4.100508895617578e-05, + "loss": 0.4695, + "num_input_tokens_seen": 16973104, + "step": 11155 + }, + { + "epoch": 0.35721144613020933, + "grad_norm": 0.8517255187034607, + "learning_rate": 4.099754588424547e-05, + "loss": 0.5413, + "num_input_tokens_seen": 16980416, + "step": 11160 + }, + { + "epoch": 0.35737148710069777, + "grad_norm": 0.5118817687034607, + "learning_rate": 4.0990000345292546e-05, + "loss": 0.5971, + "num_input_tokens_seen": 16988144, + "step": 11165 + }, + { + "epoch": 0.3575315280711862, + "grad_norm": 0.7015628814697266, + "learning_rate": 4.098245234048064e-05, + "loss": 0.6201, + "num_input_tokens_seen": 16995216, + "step": 11170 + }, + { + "epoch": 0.35769156904167465, + "grad_norm": 0.9746518135070801, + "learning_rate": 4.0974901870973726e-05, + "loss": 0.5698, + "num_input_tokens_seen": 17002768, + "step": 11175 + }, + { + "epoch": 0.3578516100121631, + "grad_norm": 1.3427969217300415, + "learning_rate": 4.096734893793619e-05, + "loss": 0.6225, + "num_input_tokens_seen": 17010432, + "step": 11180 + }, + { + "epoch": 0.3580116509826516, + "grad_norm": 0.3818061351776123, + "learning_rate": 4.095979354253279e-05, + "loss": 0.3767, + "num_input_tokens_seen": 17018048, + "step": 11185 + }, + { + "epoch": 0.35817169195314, + "grad_norm": 1.3278558254241943, + "learning_rate": 4.0952235685928656e-05, + "loss": 0.5221, + "num_input_tokens_seen": 17025472, + "step": 11190 + }, + { + "epoch": 0.35833173292362847, + "grad_norm": 0.7124447822570801, + "learning_rate": 4.094467536928932e-05, + "loss": 0.5252, + "num_input_tokens_seen": 17033184, + "step": 11195 + }, + { + "epoch": 0.3584917738941169, + "grad_norm": 0.6544645428657532, + "learning_rate": 4.093711259378067e-05, + "loss": 0.5733, + "num_input_tokens_seen": 17040512, + "step": 11200 + }, + { + "epoch": 0.3584917738941169, + "eval_loss": 0.5527748465538025, + "eval_runtime": 560.3145, + "eval_samples_per_second": 24.783, + "eval_steps_per_second": 12.391, + "num_input_tokens_seen": 17040512, + "step": 11200 + }, + { + "epoch": 0.35865181486460534, + "grad_norm": 0.915723979473114, + "learning_rate": 4.092954736056897e-05, + "loss": 0.5725, + "num_input_tokens_seen": 17048304, + "step": 11205 + }, + { + "epoch": 0.3588118558350938, + "grad_norm": 0.5868799090385437, + "learning_rate": 4.09219796708209e-05, + "loss": 0.324, + "num_input_tokens_seen": 17055792, + "step": 11210 + }, + { + "epoch": 0.3589718968055822, + "grad_norm": 1.0405552387237549, + "learning_rate": 4.0914409525703464e-05, + "loss": 0.5446, + "num_input_tokens_seen": 17063568, + "step": 11215 + }, + { + "epoch": 0.35913193777607066, + "grad_norm": 0.9703460931777954, + "learning_rate": 4.090683692638408e-05, + "loss": 0.4971, + "num_input_tokens_seen": 17070848, + "step": 11220 + }, + { + "epoch": 0.3592919787465591, + "grad_norm": 1.0399309396743774, + "learning_rate": 4.089926187403056e-05, + "loss": 0.5228, + "num_input_tokens_seen": 17078416, + "step": 11225 + }, + { + "epoch": 0.35945201971704754, + "grad_norm": 0.656028151512146, + "learning_rate": 4.0891684369811044e-05, + "loss": 0.4481, + "num_input_tokens_seen": 17085392, + "step": 11230 + }, + { + "epoch": 0.359612060687536, + "grad_norm": 0.3493669629096985, + "learning_rate": 4.0884104414894107e-05, + "loss": 0.4364, + "num_input_tokens_seen": 17093504, + "step": 11235 + }, + { + "epoch": 0.3597721016580245, + "grad_norm": 0.9248633980751038, + "learning_rate": 4.087652201044864e-05, + "loss": 0.5338, + "num_input_tokens_seen": 17100976, + "step": 11240 + }, + { + "epoch": 0.3599321426285129, + "grad_norm": 0.8657652735710144, + "learning_rate": 4.086893715764397e-05, + "loss": 0.5911, + "num_input_tokens_seen": 17108752, + "step": 11245 + }, + { + "epoch": 0.36009218359900136, + "grad_norm": 0.8320249915122986, + "learning_rate": 4.086134985764977e-05, + "loss": 0.5253, + "num_input_tokens_seen": 17116352, + "step": 11250 + }, + { + "epoch": 0.3602522245694898, + "grad_norm": 0.7890768647193909, + "learning_rate": 4.0853760111636085e-05, + "loss": 0.6657, + "num_input_tokens_seen": 17123760, + "step": 11255 + }, + { + "epoch": 0.36041226553997824, + "grad_norm": 0.47703924775123596, + "learning_rate": 4.084616792077337e-05, + "loss": 0.4723, + "num_input_tokens_seen": 17131264, + "step": 11260 + }, + { + "epoch": 0.3605723065104667, + "grad_norm": 1.0918323993682861, + "learning_rate": 4.083857328623243e-05, + "loss": 0.4901, + "num_input_tokens_seen": 17139424, + "step": 11265 + }, + { + "epoch": 0.3607323474809551, + "grad_norm": 0.4435534179210663, + "learning_rate": 4.083097620918444e-05, + "loss": 0.5052, + "num_input_tokens_seen": 17146992, + "step": 11270 + }, + { + "epoch": 0.36089238845144356, + "grad_norm": 0.6453854441642761, + "learning_rate": 4.082337669080097e-05, + "loss": 0.6062, + "num_input_tokens_seen": 17154208, + "step": 11275 + }, + { + "epoch": 0.361052429421932, + "grad_norm": 0.7366889715194702, + "learning_rate": 4.081577473225398e-05, + "loss": 0.4991, + "num_input_tokens_seen": 17161664, + "step": 11280 + }, + { + "epoch": 0.36121247039242044, + "grad_norm": 1.7490429878234863, + "learning_rate": 4.080817033471577e-05, + "loss": 0.6655, + "num_input_tokens_seen": 17169392, + "step": 11285 + }, + { + "epoch": 0.3613725113629089, + "grad_norm": 1.611175775527954, + "learning_rate": 4.080056349935903e-05, + "loss": 0.3782, + "num_input_tokens_seen": 17177296, + "step": 11290 + }, + { + "epoch": 0.3615325523333974, + "grad_norm": 0.5216576457023621, + "learning_rate": 4.079295422735684e-05, + "loss": 0.4639, + "num_input_tokens_seen": 17184688, + "step": 11295 + }, + { + "epoch": 0.3616925933038858, + "grad_norm": 0.7675060629844666, + "learning_rate": 4.078534251988264e-05, + "loss": 0.7468, + "num_input_tokens_seen": 17192768, + "step": 11300 + }, + { + "epoch": 0.36185263427437425, + "grad_norm": 1.0421199798583984, + "learning_rate": 4.077772837811025e-05, + "loss": 0.5027, + "num_input_tokens_seen": 17200240, + "step": 11305 + }, + { + "epoch": 0.3620126752448627, + "grad_norm": 0.4254607558250427, + "learning_rate": 4.0770111803213874e-05, + "loss": 0.43, + "num_input_tokens_seen": 17207792, + "step": 11310 + }, + { + "epoch": 0.36217271621535113, + "grad_norm": 0.6710538864135742, + "learning_rate": 4.076249279636807e-05, + "loss": 0.5884, + "num_input_tokens_seen": 17215520, + "step": 11315 + }, + { + "epoch": 0.36233275718583957, + "grad_norm": 0.6708403825759888, + "learning_rate": 4.075487135874781e-05, + "loss": 0.7566, + "num_input_tokens_seen": 17222592, + "step": 11320 + }, + { + "epoch": 0.362492798156328, + "grad_norm": 0.8582777976989746, + "learning_rate": 4.074724749152837e-05, + "loss": 0.5147, + "num_input_tokens_seen": 17229888, + "step": 11325 + }, + { + "epoch": 0.36265283912681645, + "grad_norm": 1.3950384855270386, + "learning_rate": 4.07396211958855e-05, + "loss": 0.7088, + "num_input_tokens_seen": 17237232, + "step": 11330 + }, + { + "epoch": 0.3628128800973049, + "grad_norm": 1.434332013130188, + "learning_rate": 4.073199247299523e-05, + "loss": 0.5572, + "num_input_tokens_seen": 17244640, + "step": 11335 + }, + { + "epoch": 0.36297292106779333, + "grad_norm": 0.666824996471405, + "learning_rate": 4.072436132403403e-05, + "loss": 0.3893, + "num_input_tokens_seen": 17252384, + "step": 11340 + }, + { + "epoch": 0.3631329620382818, + "grad_norm": 1.013649582862854, + "learning_rate": 4.0716727750178704e-05, + "loss": 0.5016, + "num_input_tokens_seen": 17260384, + "step": 11345 + }, + { + "epoch": 0.36329300300877027, + "grad_norm": 0.6707114577293396, + "learning_rate": 4.0709091752606455e-05, + "loss": 0.4339, + "num_input_tokens_seen": 17268448, + "step": 11350 + }, + { + "epoch": 0.3634530439792587, + "grad_norm": 1.2839422225952148, + "learning_rate": 4.070145333249484e-05, + "loss": 0.4959, + "num_input_tokens_seen": 17275728, + "step": 11355 + }, + { + "epoch": 0.36361308494974715, + "grad_norm": 0.6088213920593262, + "learning_rate": 4.069381249102181e-05, + "loss": 0.4368, + "num_input_tokens_seen": 17283568, + "step": 11360 + }, + { + "epoch": 0.3637731259202356, + "grad_norm": 1.154662013053894, + "learning_rate": 4.0686169229365665e-05, + "loss": 0.5609, + "num_input_tokens_seen": 17290848, + "step": 11365 + }, + { + "epoch": 0.363933166890724, + "grad_norm": 0.6254565715789795, + "learning_rate": 4.067852354870511e-05, + "loss": 0.6836, + "num_input_tokens_seen": 17298480, + "step": 11370 + }, + { + "epoch": 0.36409320786121246, + "grad_norm": 0.8237834572792053, + "learning_rate": 4.067087545021919e-05, + "loss": 0.3848, + "num_input_tokens_seen": 17305712, + "step": 11375 + }, + { + "epoch": 0.3642532488317009, + "grad_norm": 0.8285568356513977, + "learning_rate": 4.066322493508734e-05, + "loss": 0.6694, + "num_input_tokens_seen": 17313024, + "step": 11380 + }, + { + "epoch": 0.36441328980218934, + "grad_norm": 0.8227092027664185, + "learning_rate": 4.065557200448937e-05, + "loss": 0.51, + "num_input_tokens_seen": 17320384, + "step": 11385 + }, + { + "epoch": 0.3645733307726778, + "grad_norm": 0.7557495832443237, + "learning_rate": 4.064791665960546e-05, + "loss": 0.748, + "num_input_tokens_seen": 17329040, + "step": 11390 + }, + { + "epoch": 0.3647333717431662, + "grad_norm": 1.0527865886688232, + "learning_rate": 4.064025890161615e-05, + "loss": 0.4709, + "num_input_tokens_seen": 17336192, + "step": 11395 + }, + { + "epoch": 0.3648934127136547, + "grad_norm": 1.6474411487579346, + "learning_rate": 4.0632598731702373e-05, + "loss": 0.7007, + "num_input_tokens_seen": 17343664, + "step": 11400 + }, + { + "epoch": 0.3648934127136547, + "eval_loss": 0.5517261028289795, + "eval_runtime": 557.7938, + "eval_samples_per_second": 24.895, + "eval_steps_per_second": 12.447, + "num_input_tokens_seen": 17343664, + "step": 11400 + }, + { + "epoch": 0.36505345368414316, + "grad_norm": 1.343946099281311, + "learning_rate": 4.0624936151045426e-05, + "loss": 0.4043, + "num_input_tokens_seen": 17351040, + "step": 11405 + }, + { + "epoch": 0.3652134946546316, + "grad_norm": 0.9958085417747498, + "learning_rate": 4.061727116082696e-05, + "loss": 0.5061, + "num_input_tokens_seen": 17358544, + "step": 11410 + }, + { + "epoch": 0.36537353562512004, + "grad_norm": 2.1082239151000977, + "learning_rate": 4.060960376222903e-05, + "loss": 0.6098, + "num_input_tokens_seen": 17366704, + "step": 11415 + }, + { + "epoch": 0.3655335765956085, + "grad_norm": 1.8424911499023438, + "learning_rate": 4.0601933956434034e-05, + "loss": 0.586, + "num_input_tokens_seen": 17374912, + "step": 11420 + }, + { + "epoch": 0.3656936175660969, + "grad_norm": 0.8904283046722412, + "learning_rate": 4.059426174462476e-05, + "loss": 0.4815, + "num_input_tokens_seen": 17382176, + "step": 11425 + }, + { + "epoch": 0.36585365853658536, + "grad_norm": 0.8634864687919617, + "learning_rate": 4.058658712798435e-05, + "loss": 0.5748, + "num_input_tokens_seen": 17389728, + "step": 11430 + }, + { + "epoch": 0.3660136995070738, + "grad_norm": 0.658440113067627, + "learning_rate": 4.0578910107696336e-05, + "loss": 0.3884, + "num_input_tokens_seen": 17397712, + "step": 11435 + }, + { + "epoch": 0.36617374047756224, + "grad_norm": 0.637340784072876, + "learning_rate": 4.05712306849446e-05, + "loss": 0.4643, + "num_input_tokens_seen": 17404848, + "step": 11440 + }, + { + "epoch": 0.3663337814480507, + "grad_norm": 1.677551507949829, + "learning_rate": 4.0563548860913415e-05, + "loss": 0.5946, + "num_input_tokens_seen": 17412128, + "step": 11445 + }, + { + "epoch": 0.3664938224185392, + "grad_norm": 0.44986218214035034, + "learning_rate": 4.0555864636787414e-05, + "loss": 0.4277, + "num_input_tokens_seen": 17419760, + "step": 11450 + }, + { + "epoch": 0.3666538633890276, + "grad_norm": 1.817347764968872, + "learning_rate": 4.054817801375159e-05, + "loss": 0.721, + "num_input_tokens_seen": 17427648, + "step": 11455 + }, + { + "epoch": 0.36681390435951605, + "grad_norm": 0.7340410351753235, + "learning_rate": 4.054048899299134e-05, + "loss": 0.5706, + "num_input_tokens_seen": 17435152, + "step": 11460 + }, + { + "epoch": 0.3669739453300045, + "grad_norm": 1.2371625900268555, + "learning_rate": 4.0532797575692385e-05, + "loss": 0.6852, + "num_input_tokens_seen": 17442816, + "step": 11465 + }, + { + "epoch": 0.36713398630049293, + "grad_norm": 0.5327403545379639, + "learning_rate": 4.052510376304085e-05, + "loss": 0.4803, + "num_input_tokens_seen": 17450752, + "step": 11470 + }, + { + "epoch": 0.36729402727098137, + "grad_norm": 0.7859582901000977, + "learning_rate": 4.051740755622321e-05, + "loss": 0.6416, + "num_input_tokens_seen": 17458160, + "step": 11475 + }, + { + "epoch": 0.3674540682414698, + "grad_norm": 1.0451785326004028, + "learning_rate": 4.050970895642632e-05, + "loss": 0.598, + "num_input_tokens_seen": 17465824, + "step": 11480 + }, + { + "epoch": 0.36761410921195825, + "grad_norm": 0.5210081934928894, + "learning_rate": 4.050200796483741e-05, + "loss": 0.523, + "num_input_tokens_seen": 17473984, + "step": 11485 + }, + { + "epoch": 0.3677741501824467, + "grad_norm": 0.7352765798568726, + "learning_rate": 4.049430458264405e-05, + "loss": 0.7185, + "num_input_tokens_seen": 17481600, + "step": 11490 + }, + { + "epoch": 0.36793419115293513, + "grad_norm": 1.250636339187622, + "learning_rate": 4.048659881103422e-05, + "loss": 0.544, + "num_input_tokens_seen": 17489008, + "step": 11495 + }, + { + "epoch": 0.36809423212342357, + "grad_norm": 0.6971928477287292, + "learning_rate": 4.0478890651196235e-05, + "loss": 0.4769, + "num_input_tokens_seen": 17496576, + "step": 11500 + }, + { + "epoch": 0.36825427309391207, + "grad_norm": 1.7726234197616577, + "learning_rate": 4.047118010431879e-05, + "loss": 0.475, + "num_input_tokens_seen": 17503664, + "step": 11505 + }, + { + "epoch": 0.3684143140644005, + "grad_norm": 0.7211388945579529, + "learning_rate": 4.046346717159094e-05, + "loss": 0.4318, + "num_input_tokens_seen": 17511376, + "step": 11510 + }, + { + "epoch": 0.36857435503488895, + "grad_norm": 1.8667330741882324, + "learning_rate": 4.045575185420214e-05, + "loss": 0.7417, + "num_input_tokens_seen": 17518944, + "step": 11515 + }, + { + "epoch": 0.3687343960053774, + "grad_norm": 1.1732436418533325, + "learning_rate": 4.0448034153342165e-05, + "loss": 0.6535, + "num_input_tokens_seen": 17525808, + "step": 11520 + }, + { + "epoch": 0.3688944369758658, + "grad_norm": 0.500050961971283, + "learning_rate": 4.0440314070201194e-05, + "loss": 0.6204, + "num_input_tokens_seen": 17532976, + "step": 11525 + }, + { + "epoch": 0.36905447794635426, + "grad_norm": 1.0156567096710205, + "learning_rate": 4.043259160596976e-05, + "loss": 0.4894, + "num_input_tokens_seen": 17539888, + "step": 11530 + }, + { + "epoch": 0.3692145189168427, + "grad_norm": 1.959834098815918, + "learning_rate": 4.0424866761838767e-05, + "loss": 0.9251, + "num_input_tokens_seen": 17547200, + "step": 11535 + }, + { + "epoch": 0.36937455988733114, + "grad_norm": 0.96611487865448, + "learning_rate": 4.041713953899948e-05, + "loss": 0.4445, + "num_input_tokens_seen": 17554992, + "step": 11540 + }, + { + "epoch": 0.3695346008578196, + "grad_norm": 0.6140342354774475, + "learning_rate": 4.0409409938643515e-05, + "loss": 0.4525, + "num_input_tokens_seen": 17562512, + "step": 11545 + }, + { + "epoch": 0.369694641828308, + "grad_norm": 0.6732484102249146, + "learning_rate": 4.0401677961962904e-05, + "loss": 0.4822, + "num_input_tokens_seen": 17570048, + "step": 11550 + }, + { + "epoch": 0.3698546827987965, + "grad_norm": 0.3743967115879059, + "learning_rate": 4.039394361015001e-05, + "loss": 0.4427, + "num_input_tokens_seen": 17577520, + "step": 11555 + }, + { + "epoch": 0.37001472376928496, + "grad_norm": 0.935262143611908, + "learning_rate": 4.038620688439755e-05, + "loss": 0.6388, + "num_input_tokens_seen": 17584976, + "step": 11560 + }, + { + "epoch": 0.3701747647397734, + "grad_norm": 1.1118290424346924, + "learning_rate": 4.037846778589862e-05, + "loss": 0.7161, + "num_input_tokens_seen": 17592688, + "step": 11565 + }, + { + "epoch": 0.37033480571026184, + "grad_norm": 0.897808313369751, + "learning_rate": 4.0370726315846715e-05, + "loss": 0.4952, + "num_input_tokens_seen": 17600272, + "step": 11570 + }, + { + "epoch": 0.3704948466807503, + "grad_norm": 0.8329852223396301, + "learning_rate": 4.036298247543565e-05, + "loss": 0.6297, + "num_input_tokens_seen": 17607520, + "step": 11575 + }, + { + "epoch": 0.3706548876512387, + "grad_norm": 0.449349582195282, + "learning_rate": 4.035523626585962e-05, + "loss": 0.5679, + "num_input_tokens_seen": 17615536, + "step": 11580 + }, + { + "epoch": 0.37081492862172716, + "grad_norm": 0.5767116546630859, + "learning_rate": 4.0347487688313194e-05, + "loss": 0.5781, + "num_input_tokens_seen": 17622944, + "step": 11585 + }, + { + "epoch": 0.3709749695922156, + "grad_norm": 1.152858853340149, + "learning_rate": 4.0339736743991296e-05, + "loss": 0.5629, + "num_input_tokens_seen": 17630464, + "step": 11590 + }, + { + "epoch": 0.37113501056270404, + "grad_norm": 0.8208476901054382, + "learning_rate": 4.0331983434089227e-05, + "loss": 0.4967, + "num_input_tokens_seen": 17638368, + "step": 11595 + }, + { + "epoch": 0.3712950515331925, + "grad_norm": 2.2660512924194336, + "learning_rate": 4.032422775980264e-05, + "loss": 0.7309, + "num_input_tokens_seen": 17646512, + "step": 11600 + }, + { + "epoch": 0.3712950515331925, + "eval_loss": 0.5503464937210083, + "eval_runtime": 558.9166, + "eval_samples_per_second": 24.844, + "eval_steps_per_second": 12.422, + "num_input_tokens_seen": 17646512, + "step": 11600 + }, + { + "epoch": 0.3714550925036809, + "grad_norm": 0.5664612054824829, + "learning_rate": 4.031646972232754e-05, + "loss": 0.4545, + "num_input_tokens_seen": 17653840, + "step": 11605 + }, + { + "epoch": 0.3716151334741694, + "grad_norm": 0.9455744028091431, + "learning_rate": 4.0308709322860344e-05, + "loss": 0.5236, + "num_input_tokens_seen": 17661520, + "step": 11610 + }, + { + "epoch": 0.37177517444465785, + "grad_norm": 0.6934634447097778, + "learning_rate": 4.0300946562597784e-05, + "loss": 0.5835, + "num_input_tokens_seen": 17669248, + "step": 11615 + }, + { + "epoch": 0.3719352154151463, + "grad_norm": 0.501186192035675, + "learning_rate": 4.029318144273698e-05, + "loss": 0.3445, + "num_input_tokens_seen": 17676480, + "step": 11620 + }, + { + "epoch": 0.37209525638563473, + "grad_norm": 0.8624425530433655, + "learning_rate": 4.0285413964475415e-05, + "loss": 0.369, + "num_input_tokens_seen": 17684000, + "step": 11625 + }, + { + "epoch": 0.37225529735612317, + "grad_norm": 1.2054930925369263, + "learning_rate": 4.0277644129010927e-05, + "loss": 0.6962, + "num_input_tokens_seen": 17691488, + "step": 11630 + }, + { + "epoch": 0.3724153383266116, + "grad_norm": 0.6346865296363831, + "learning_rate": 4.0269871937541724e-05, + "loss": 0.5229, + "num_input_tokens_seen": 17699200, + "step": 11635 + }, + { + "epoch": 0.37257537929710005, + "grad_norm": 0.5696830153465271, + "learning_rate": 4.026209739126637e-05, + "loss": 0.4993, + "num_input_tokens_seen": 17706624, + "step": 11640 + }, + { + "epoch": 0.3727354202675885, + "grad_norm": 0.9186151623725891, + "learning_rate": 4.025432049138381e-05, + "loss": 0.4179, + "num_input_tokens_seen": 17714416, + "step": 11645 + }, + { + "epoch": 0.37289546123807693, + "grad_norm": 0.7531718611717224, + "learning_rate": 4.0246541239093325e-05, + "loss": 0.5225, + "num_input_tokens_seen": 17722064, + "step": 11650 + }, + { + "epoch": 0.37305550220856537, + "grad_norm": 0.9668963551521301, + "learning_rate": 4.023875963559459e-05, + "loss": 0.4512, + "num_input_tokens_seen": 17729904, + "step": 11655 + }, + { + "epoch": 0.3732155431790538, + "grad_norm": 0.8548779487609863, + "learning_rate": 4.023097568208761e-05, + "loss": 0.5706, + "num_input_tokens_seen": 17737504, + "step": 11660 + }, + { + "epoch": 0.3733755841495423, + "grad_norm": 0.7149280905723572, + "learning_rate": 4.022318937977277e-05, + "loss": 0.4809, + "num_input_tokens_seen": 17745408, + "step": 11665 + }, + { + "epoch": 0.37353562512003075, + "grad_norm": 0.603229284286499, + "learning_rate": 4.021540072985084e-05, + "loss": 0.4561, + "num_input_tokens_seen": 17752656, + "step": 11670 + }, + { + "epoch": 0.3736956660905192, + "grad_norm": 0.5577490329742432, + "learning_rate": 4.020760973352289e-05, + "loss": 0.5563, + "num_input_tokens_seen": 17760576, + "step": 11675 + }, + { + "epoch": 0.3738557070610076, + "grad_norm": 0.7058331370353699, + "learning_rate": 4.019981639199042e-05, + "loss": 0.5192, + "num_input_tokens_seen": 17767456, + "step": 11680 + }, + { + "epoch": 0.37401574803149606, + "grad_norm": 0.9523742198944092, + "learning_rate": 4.0192020706455245e-05, + "loss": 0.5002, + "num_input_tokens_seen": 17775120, + "step": 11685 + }, + { + "epoch": 0.3741757890019845, + "grad_norm": 0.7829042673110962, + "learning_rate": 4.018422267811956e-05, + "loss": 0.7675, + "num_input_tokens_seen": 17782368, + "step": 11690 + }, + { + "epoch": 0.37433582997247294, + "grad_norm": 1.1260873079299927, + "learning_rate": 4.017642230818592e-05, + "loss": 0.4717, + "num_input_tokens_seen": 17790208, + "step": 11695 + }, + { + "epoch": 0.3744958709429614, + "grad_norm": 1.386512041091919, + "learning_rate": 4.0168619597857246e-05, + "loss": 0.556, + "num_input_tokens_seen": 17797872, + "step": 11700 + }, + { + "epoch": 0.3746559119134498, + "grad_norm": 0.9487335681915283, + "learning_rate": 4.016081454833681e-05, + "loss": 0.6134, + "num_input_tokens_seen": 17805248, + "step": 11705 + }, + { + "epoch": 0.37481595288393826, + "grad_norm": 0.8051625490188599, + "learning_rate": 4.0153007160828245e-05, + "loss": 0.5925, + "num_input_tokens_seen": 17813088, + "step": 11710 + }, + { + "epoch": 0.37497599385442676, + "grad_norm": 2.0102267265319824, + "learning_rate": 4.0145197436535555e-05, + "loss": 0.6001, + "num_input_tokens_seen": 17820672, + "step": 11715 + }, + { + "epoch": 0.3751360348249152, + "grad_norm": 1.4646944999694824, + "learning_rate": 4.0137385376663095e-05, + "loss": 0.5565, + "num_input_tokens_seen": 17828176, + "step": 11720 + }, + { + "epoch": 0.37529607579540364, + "grad_norm": 0.5734401941299438, + "learning_rate": 4.012957098241558e-05, + "loss": 0.4872, + "num_input_tokens_seen": 17836112, + "step": 11725 + }, + { + "epoch": 0.3754561167658921, + "grad_norm": 0.625264048576355, + "learning_rate": 4.0121754254998076e-05, + "loss": 0.5355, + "num_input_tokens_seen": 17843456, + "step": 11730 + }, + { + "epoch": 0.3756161577363805, + "grad_norm": 1.2408050298690796, + "learning_rate": 4.011393519561606e-05, + "loss": 0.4441, + "num_input_tokens_seen": 17851056, + "step": 11735 + }, + { + "epoch": 0.37577619870686896, + "grad_norm": 0.7999330163002014, + "learning_rate": 4.010611380547529e-05, + "loss": 0.5632, + "num_input_tokens_seen": 17858880, + "step": 11740 + }, + { + "epoch": 0.3759362396773574, + "grad_norm": 1.2431484460830688, + "learning_rate": 4.009829008578192e-05, + "loss": 0.4338, + "num_input_tokens_seen": 17866304, + "step": 11745 + }, + { + "epoch": 0.37609628064784584, + "grad_norm": 0.5967097878456116, + "learning_rate": 4.00904640377425e-05, + "loss": 0.5008, + "num_input_tokens_seen": 17873760, + "step": 11750 + }, + { + "epoch": 0.3762563216183343, + "grad_norm": 1.037456750869751, + "learning_rate": 4.0082635662563886e-05, + "loss": 0.6651, + "num_input_tokens_seen": 17881696, + "step": 11755 + }, + { + "epoch": 0.3764163625888227, + "grad_norm": 1.0629241466522217, + "learning_rate": 4.007480496145331e-05, + "loss": 0.7726, + "num_input_tokens_seen": 17889712, + "step": 11760 + }, + { + "epoch": 0.37657640355931116, + "grad_norm": 0.5061311721801758, + "learning_rate": 4.006697193561837e-05, + "loss": 0.5434, + "num_input_tokens_seen": 17897040, + "step": 11765 + }, + { + "epoch": 0.37673644452979965, + "grad_norm": 1.2021136283874512, + "learning_rate": 4.005913658626701e-05, + "loss": 0.575, + "num_input_tokens_seen": 17904480, + "step": 11770 + }, + { + "epoch": 0.3768964855002881, + "grad_norm": 0.7116298675537109, + "learning_rate": 4.005129891460754e-05, + "loss": 0.631, + "num_input_tokens_seen": 17911872, + "step": 11775 + }, + { + "epoch": 0.37705652647077653, + "grad_norm": 0.3428499698638916, + "learning_rate": 4.004345892184864e-05, + "loss": 0.3467, + "num_input_tokens_seen": 17919584, + "step": 11780 + }, + { + "epoch": 0.37721656744126497, + "grad_norm": 3.4800267219543457, + "learning_rate": 4.003561660919932e-05, + "loss": 0.4893, + "num_input_tokens_seen": 17927424, + "step": 11785 + }, + { + "epoch": 0.3773766084117534, + "grad_norm": 0.6550127267837524, + "learning_rate": 4.002777197786897e-05, + "loss": 0.4832, + "num_input_tokens_seen": 17934864, + "step": 11790 + }, + { + "epoch": 0.37753664938224185, + "grad_norm": 0.8023700714111328, + "learning_rate": 4.0019925029067326e-05, + "loss": 0.4816, + "num_input_tokens_seen": 17942416, + "step": 11795 + }, + { + "epoch": 0.3776966903527303, + "grad_norm": 0.5051056146621704, + "learning_rate": 4.0012075764004495e-05, + "loss": 0.44, + "num_input_tokens_seen": 17949824, + "step": 11800 + }, + { + "epoch": 0.3776966903527303, + "eval_loss": 0.5488370656967163, + "eval_runtime": 562.9886, + "eval_samples_per_second": 24.665, + "eval_steps_per_second": 12.332, + "num_input_tokens_seen": 17949824, + "step": 11800 + }, + { + "epoch": 0.37785673132321873, + "grad_norm": 0.5437909960746765, + "learning_rate": 4.000422418389094e-05, + "loss": 0.4057, + "num_input_tokens_seen": 17957248, + "step": 11805 + }, + { + "epoch": 0.37801677229370717, + "grad_norm": 1.021309494972229, + "learning_rate": 3.999637028993744e-05, + "loss": 0.592, + "num_input_tokens_seen": 17965200, + "step": 11810 + }, + { + "epoch": 0.3781768132641956, + "grad_norm": 1.3017667531967163, + "learning_rate": 3.99885140833552e-05, + "loss": 0.4758, + "num_input_tokens_seen": 17972272, + "step": 11815 + }, + { + "epoch": 0.3783368542346841, + "grad_norm": 1.2444844245910645, + "learning_rate": 3.998065556535572e-05, + "loss": 0.5513, + "num_input_tokens_seen": 17979632, + "step": 11820 + }, + { + "epoch": 0.37849689520517255, + "grad_norm": 0.4093603789806366, + "learning_rate": 3.9972794737150895e-05, + "loss": 0.5181, + "num_input_tokens_seen": 17987392, + "step": 11825 + }, + { + "epoch": 0.378656936175661, + "grad_norm": 0.6576096415519714, + "learning_rate": 3.996493159995297e-05, + "loss": 0.4059, + "num_input_tokens_seen": 17994992, + "step": 11830 + }, + { + "epoch": 0.3788169771461494, + "grad_norm": 0.8157045841217041, + "learning_rate": 3.995706615497453e-05, + "loss": 0.6227, + "num_input_tokens_seen": 18002352, + "step": 11835 + }, + { + "epoch": 0.37897701811663786, + "grad_norm": 2.0282607078552246, + "learning_rate": 3.994919840342852e-05, + "loss": 0.5798, + "num_input_tokens_seen": 18010208, + "step": 11840 + }, + { + "epoch": 0.3791370590871263, + "grad_norm": 0.9961637258529663, + "learning_rate": 3.994132834652825e-05, + "loss": 0.4237, + "num_input_tokens_seen": 18017936, + "step": 11845 + }, + { + "epoch": 0.37929710005761474, + "grad_norm": 1.022152304649353, + "learning_rate": 3.99334559854874e-05, + "loss": 0.7025, + "num_input_tokens_seen": 18025648, + "step": 11850 + }, + { + "epoch": 0.3794571410281032, + "grad_norm": 0.6404092907905579, + "learning_rate": 3.9925581321519955e-05, + "loss": 0.552, + "num_input_tokens_seen": 18033056, + "step": 11855 + }, + { + "epoch": 0.3796171819985916, + "grad_norm": 0.6829541325569153, + "learning_rate": 3.991770435584031e-05, + "loss": 0.4713, + "num_input_tokens_seen": 18040384, + "step": 11860 + }, + { + "epoch": 0.37977722296908006, + "grad_norm": 0.8595101237297058, + "learning_rate": 3.990982508966319e-05, + "loss": 0.3633, + "num_input_tokens_seen": 18047888, + "step": 11865 + }, + { + "epoch": 0.3799372639395685, + "grad_norm": 0.8672055006027222, + "learning_rate": 3.990194352420367e-05, + "loss": 0.6697, + "num_input_tokens_seen": 18055536, + "step": 11870 + }, + { + "epoch": 0.380097304910057, + "grad_norm": 0.6201447248458862, + "learning_rate": 3.9894059660677184e-05, + "loss": 0.4845, + "num_input_tokens_seen": 18062656, + "step": 11875 + }, + { + "epoch": 0.38025734588054544, + "grad_norm": 0.7188119292259216, + "learning_rate": 3.9886173500299526e-05, + "loss": 0.6216, + "num_input_tokens_seen": 18070256, + "step": 11880 + }, + { + "epoch": 0.3804173868510339, + "grad_norm": 0.9758583307266235, + "learning_rate": 3.987828504428685e-05, + "loss": 0.4577, + "num_input_tokens_seen": 18077600, + "step": 11885 + }, + { + "epoch": 0.3805774278215223, + "grad_norm": 1.5808885097503662, + "learning_rate": 3.987039429385565e-05, + "loss": 0.6267, + "num_input_tokens_seen": 18085120, + "step": 11890 + }, + { + "epoch": 0.38073746879201076, + "grad_norm": 1.651623010635376, + "learning_rate": 3.986250125022277e-05, + "loss": 0.4572, + "num_input_tokens_seen": 18092912, + "step": 11895 + }, + { + "epoch": 0.3808975097624992, + "grad_norm": 0.724887490272522, + "learning_rate": 3.985460591460544e-05, + "loss": 0.4702, + "num_input_tokens_seen": 18099952, + "step": 11900 + }, + { + "epoch": 0.38105755073298764, + "grad_norm": 0.8347872495651245, + "learning_rate": 3.984670828822118e-05, + "loss": 0.4881, + "num_input_tokens_seen": 18107856, + "step": 11905 + }, + { + "epoch": 0.3812175917034761, + "grad_norm": 1.7153046131134033, + "learning_rate": 3.983880837228794e-05, + "loss": 0.6299, + "num_input_tokens_seen": 18116000, + "step": 11910 + }, + { + "epoch": 0.3813776326739645, + "grad_norm": 0.8470510840415955, + "learning_rate": 3.983090616802396e-05, + "loss": 0.5653, + "num_input_tokens_seen": 18123888, + "step": 11915 + }, + { + "epoch": 0.38153767364445296, + "grad_norm": 0.6446157693862915, + "learning_rate": 3.982300167664788e-05, + "loss": 0.4102, + "num_input_tokens_seen": 18131296, + "step": 11920 + }, + { + "epoch": 0.38169771461494145, + "grad_norm": 0.7880420088768005, + "learning_rate": 3.981509489937868e-05, + "loss": 0.415, + "num_input_tokens_seen": 18138592, + "step": 11925 + }, + { + "epoch": 0.3818577555854299, + "grad_norm": 0.819622814655304, + "learning_rate": 3.9807185837435643e-05, + "loss": 0.6384, + "num_input_tokens_seen": 18146256, + "step": 11930 + }, + { + "epoch": 0.38201779655591833, + "grad_norm": 0.6786606311798096, + "learning_rate": 3.9799274492038484e-05, + "loss": 0.5591, + "num_input_tokens_seen": 18153392, + "step": 11935 + }, + { + "epoch": 0.38217783752640677, + "grad_norm": 0.41580435633659363, + "learning_rate": 3.979136086440722e-05, + "loss": 0.4621, + "num_input_tokens_seen": 18161200, + "step": 11940 + }, + { + "epoch": 0.3823378784968952, + "grad_norm": 0.8089686632156372, + "learning_rate": 3.9783444955762226e-05, + "loss": 0.4117, + "num_input_tokens_seen": 18169424, + "step": 11945 + }, + { + "epoch": 0.38249791946738365, + "grad_norm": 0.9517388343811035, + "learning_rate": 3.977552676732424e-05, + "loss": 0.4482, + "num_input_tokens_seen": 18177232, + "step": 11950 + }, + { + "epoch": 0.3826579604378721, + "grad_norm": 0.7463553547859192, + "learning_rate": 3.976760630031435e-05, + "loss": 0.3684, + "num_input_tokens_seen": 18184480, + "step": 11955 + }, + { + "epoch": 0.38281800140836053, + "grad_norm": 0.7089853286743164, + "learning_rate": 3.975968355595398e-05, + "loss": 0.6374, + "num_input_tokens_seen": 18191984, + "step": 11960 + }, + { + "epoch": 0.38297804237884897, + "grad_norm": 0.4805748164653778, + "learning_rate": 3.9751758535464935e-05, + "loss": 0.4106, + "num_input_tokens_seen": 18199280, + "step": 11965 + }, + { + "epoch": 0.3831380833493374, + "grad_norm": 0.970484733581543, + "learning_rate": 3.9743831240069326e-05, + "loss": 0.4509, + "num_input_tokens_seen": 18207024, + "step": 11970 + }, + { + "epoch": 0.38329812431982585, + "grad_norm": 0.5895483493804932, + "learning_rate": 3.9735901670989675e-05, + "loss": 0.4039, + "num_input_tokens_seen": 18214368, + "step": 11975 + }, + { + "epoch": 0.38345816529031435, + "grad_norm": 1.8571817874908447, + "learning_rate": 3.97279698294488e-05, + "loss": 0.5324, + "num_input_tokens_seen": 18221904, + "step": 11980 + }, + { + "epoch": 0.3836182062608028, + "grad_norm": 0.7186763286590576, + "learning_rate": 3.9720035716669876e-05, + "loss": 0.6139, + "num_input_tokens_seen": 18229472, + "step": 11985 + }, + { + "epoch": 0.3837782472312912, + "grad_norm": 0.821086049079895, + "learning_rate": 3.9712099333876474e-05, + "loss": 0.4421, + "num_input_tokens_seen": 18236960, + "step": 11990 + }, + { + "epoch": 0.38393828820177966, + "grad_norm": 0.6690552830696106, + "learning_rate": 3.9704160682292475e-05, + "loss": 0.4853, + "num_input_tokens_seen": 18244352, + "step": 11995 + }, + { + "epoch": 0.3840983291722681, + "grad_norm": 1.5097336769104004, + "learning_rate": 3.9696219763142106e-05, + "loss": 0.5858, + "num_input_tokens_seen": 18252224, + "step": 12000 + }, + { + "epoch": 0.3840983291722681, + "eval_loss": 0.5479918122291565, + "eval_runtime": 560.4234, + "eval_samples_per_second": 24.778, + "eval_steps_per_second": 12.389, + "num_input_tokens_seen": 18252224, + "step": 12000 + }, + { + "epoch": 0.38425837014275654, + "grad_norm": 0.7644066214561462, + "learning_rate": 3.968827657764997e-05, + "loss": 0.6102, + "num_input_tokens_seen": 18259792, + "step": 12005 + }, + { + "epoch": 0.384418411113245, + "grad_norm": 0.6639662384986877, + "learning_rate": 3.9680331127041e-05, + "loss": 0.5921, + "num_input_tokens_seen": 18267248, + "step": 12010 + }, + { + "epoch": 0.3845784520837334, + "grad_norm": 0.5806403756141663, + "learning_rate": 3.9672383412540495e-05, + "loss": 0.6753, + "num_input_tokens_seen": 18274352, + "step": 12015 + }, + { + "epoch": 0.38473849305422186, + "grad_norm": 0.6900522708892822, + "learning_rate": 3.966443343537407e-05, + "loss": 0.5839, + "num_input_tokens_seen": 18282080, + "step": 12020 + }, + { + "epoch": 0.3848985340247103, + "grad_norm": 0.6659555435180664, + "learning_rate": 3.965648119676772e-05, + "loss": 0.3998, + "num_input_tokens_seen": 18290144, + "step": 12025 + }, + { + "epoch": 0.38505857499519874, + "grad_norm": 0.9297739267349243, + "learning_rate": 3.96485266979478e-05, + "loss": 0.5976, + "num_input_tokens_seen": 18297600, + "step": 12030 + }, + { + "epoch": 0.38521861596568724, + "grad_norm": 1.0759315490722656, + "learning_rate": 3.9640569940140974e-05, + "loss": 0.379, + "num_input_tokens_seen": 18305008, + "step": 12035 + }, + { + "epoch": 0.3853786569361757, + "grad_norm": 0.5395075082778931, + "learning_rate": 3.963261092457428e-05, + "loss": 0.5207, + "num_input_tokens_seen": 18312976, + "step": 12040 + }, + { + "epoch": 0.3855386979066641, + "grad_norm": 1.712165117263794, + "learning_rate": 3.962464965247509e-05, + "loss": 0.7878, + "num_input_tokens_seen": 18320576, + "step": 12045 + }, + { + "epoch": 0.38569873887715256, + "grad_norm": 0.6429418921470642, + "learning_rate": 3.9616686125071135e-05, + "loss": 0.483, + "num_input_tokens_seen": 18328256, + "step": 12050 + }, + { + "epoch": 0.385858779847641, + "grad_norm": 0.8570760488510132, + "learning_rate": 3.9608720343590506e-05, + "loss": 0.3004, + "num_input_tokens_seen": 18336080, + "step": 12055 + }, + { + "epoch": 0.38601882081812944, + "grad_norm": 0.7988244891166687, + "learning_rate": 3.960075230926161e-05, + "loss": 0.4216, + "num_input_tokens_seen": 18343440, + "step": 12060 + }, + { + "epoch": 0.3861788617886179, + "grad_norm": 0.621871829032898, + "learning_rate": 3.959278202331322e-05, + "loss": 0.4649, + "num_input_tokens_seen": 18351456, + "step": 12065 + }, + { + "epoch": 0.3863389027591063, + "grad_norm": 0.46421852707862854, + "learning_rate": 3.958480948697446e-05, + "loss": 0.4867, + "num_input_tokens_seen": 18359440, + "step": 12070 + }, + { + "epoch": 0.38649894372959476, + "grad_norm": 0.7685534954071045, + "learning_rate": 3.95768347014748e-05, + "loss": 0.5865, + "num_input_tokens_seen": 18366640, + "step": 12075 + }, + { + "epoch": 0.3866589847000832, + "grad_norm": 0.7507293224334717, + "learning_rate": 3.956885766804404e-05, + "loss": 0.4656, + "num_input_tokens_seen": 18374288, + "step": 12080 + }, + { + "epoch": 0.3868190256705717, + "grad_norm": 0.8003084659576416, + "learning_rate": 3.956087838791235e-05, + "loss": 0.4379, + "num_input_tokens_seen": 18381728, + "step": 12085 + }, + { + "epoch": 0.38697906664106013, + "grad_norm": 0.6481128334999084, + "learning_rate": 3.955289686231022e-05, + "loss": 0.439, + "num_input_tokens_seen": 18389264, + "step": 12090 + }, + { + "epoch": 0.38713910761154857, + "grad_norm": 1.0872316360473633, + "learning_rate": 3.9544913092468504e-05, + "loss": 0.5935, + "num_input_tokens_seen": 18396384, + "step": 12095 + }, + { + "epoch": 0.387299148582037, + "grad_norm": 0.8163420557975769, + "learning_rate": 3.9536927079618425e-05, + "loss": 0.406, + "num_input_tokens_seen": 18403584, + "step": 12100 + }, + { + "epoch": 0.38745918955252545, + "grad_norm": 1.590104341506958, + "learning_rate": 3.9528938824991494e-05, + "loss": 0.7183, + "num_input_tokens_seen": 18411328, + "step": 12105 + }, + { + "epoch": 0.3876192305230139, + "grad_norm": 1.4211269617080688, + "learning_rate": 3.952094832981962e-05, + "loss": 0.4248, + "num_input_tokens_seen": 18418784, + "step": 12110 + }, + { + "epoch": 0.38777927149350233, + "grad_norm": 0.8752005100250244, + "learning_rate": 3.951295559533503e-05, + "loss": 0.5513, + "num_input_tokens_seen": 18425984, + "step": 12115 + }, + { + "epoch": 0.38793931246399077, + "grad_norm": 0.9070929288864136, + "learning_rate": 3.95049606227703e-05, + "loss": 0.3511, + "num_input_tokens_seen": 18433344, + "step": 12120 + }, + { + "epoch": 0.3880993534344792, + "grad_norm": 0.6894732713699341, + "learning_rate": 3.949696341335838e-05, + "loss": 0.5456, + "num_input_tokens_seen": 18441424, + "step": 12125 + }, + { + "epoch": 0.38825939440496765, + "grad_norm": 1.0248112678527832, + "learning_rate": 3.9488963968332503e-05, + "loss": 0.4719, + "num_input_tokens_seen": 18448512, + "step": 12130 + }, + { + "epoch": 0.3884194353754561, + "grad_norm": 1.0098198652267456, + "learning_rate": 3.948096228892631e-05, + "loss": 0.6965, + "num_input_tokens_seen": 18456272, + "step": 12135 + }, + { + "epoch": 0.3885794763459446, + "grad_norm": 1.3473283052444458, + "learning_rate": 3.947295837637375e-05, + "loss": 0.6453, + "num_input_tokens_seen": 18463584, + "step": 12140 + }, + { + "epoch": 0.388739517316433, + "grad_norm": 1.3147177696228027, + "learning_rate": 3.9464952231909135e-05, + "loss": 0.636, + "num_input_tokens_seen": 18470848, + "step": 12145 + }, + { + "epoch": 0.38889955828692147, + "grad_norm": 0.7112852931022644, + "learning_rate": 3.945694385676711e-05, + "loss": 0.4985, + "num_input_tokens_seen": 18478640, + "step": 12150 + }, + { + "epoch": 0.3890595992574099, + "grad_norm": 0.604865550994873, + "learning_rate": 3.944893325218265e-05, + "loss": 0.4963, + "num_input_tokens_seen": 18486832, + "step": 12155 + }, + { + "epoch": 0.38921964022789834, + "grad_norm": 1.1375908851623535, + "learning_rate": 3.944092041939112e-05, + "loss": 0.5839, + "num_input_tokens_seen": 18494960, + "step": 12160 + }, + { + "epoch": 0.3893796811983868, + "grad_norm": 1.3244391679763794, + "learning_rate": 3.943290535962818e-05, + "loss": 0.5397, + "num_input_tokens_seen": 18502448, + "step": 12165 + }, + { + "epoch": 0.3895397221688752, + "grad_norm": 2.1433639526367188, + "learning_rate": 3.942488807412985e-05, + "loss": 0.5396, + "num_input_tokens_seen": 18509872, + "step": 12170 + }, + { + "epoch": 0.38969976313936366, + "grad_norm": 0.5488122701644897, + "learning_rate": 3.941686856413251e-05, + "loss": 0.4125, + "num_input_tokens_seen": 18517536, + "step": 12175 + }, + { + "epoch": 0.3898598041098521, + "grad_norm": 0.48031237721443176, + "learning_rate": 3.9408846830872874e-05, + "loss": 0.4249, + "num_input_tokens_seen": 18525120, + "step": 12180 + }, + { + "epoch": 0.39001984508034054, + "grad_norm": 0.6709063053131104, + "learning_rate": 3.940082287558798e-05, + "loss": 0.4663, + "num_input_tokens_seen": 18532752, + "step": 12185 + }, + { + "epoch": 0.39017988605082904, + "grad_norm": 2.0552945137023926, + "learning_rate": 3.939279669951522e-05, + "loss": 0.8926, + "num_input_tokens_seen": 18543648, + "step": 12190 + }, + { + "epoch": 0.3903399270213175, + "grad_norm": 0.9890773296356201, + "learning_rate": 3.938476830389234e-05, + "loss": 0.6335, + "num_input_tokens_seen": 18551344, + "step": 12195 + }, + { + "epoch": 0.3904999679918059, + "grad_norm": 0.9011390805244446, + "learning_rate": 3.937673768995742e-05, + "loss": 0.5043, + "num_input_tokens_seen": 18558800, + "step": 12200 + }, + { + "epoch": 0.3904999679918059, + "eval_loss": 0.5467565059661865, + "eval_runtime": 559.2117, + "eval_samples_per_second": 24.831, + "eval_steps_per_second": 12.416, + "num_input_tokens_seen": 18558800, + "step": 12200 + }, + { + "epoch": 0.39066000896229436, + "grad_norm": 0.7501817941665649, + "learning_rate": 3.936870485894888e-05, + "loss": 0.5533, + "num_input_tokens_seen": 18566144, + "step": 12205 + }, + { + "epoch": 0.3908200499327828, + "grad_norm": 1.0292774438858032, + "learning_rate": 3.9360669812105475e-05, + "loss": 0.5454, + "num_input_tokens_seen": 18573712, + "step": 12210 + }, + { + "epoch": 0.39098009090327124, + "grad_norm": 0.8793210387229919, + "learning_rate": 3.9352632550666325e-05, + "loss": 0.4772, + "num_input_tokens_seen": 18581264, + "step": 12215 + }, + { + "epoch": 0.3911401318737597, + "grad_norm": 0.6384241580963135, + "learning_rate": 3.9344593075870866e-05, + "loss": 0.4357, + "num_input_tokens_seen": 18588976, + "step": 12220 + }, + { + "epoch": 0.3913001728442481, + "grad_norm": 1.2322181463241577, + "learning_rate": 3.933655138895889e-05, + "loss": 0.5589, + "num_input_tokens_seen": 18596672, + "step": 12225 + }, + { + "epoch": 0.39146021381473656, + "grad_norm": 0.5108723640441895, + "learning_rate": 3.932850749117053e-05, + "loss": 0.6968, + "num_input_tokens_seen": 18604416, + "step": 12230 + }, + { + "epoch": 0.391620254785225, + "grad_norm": 0.6832753419876099, + "learning_rate": 3.932046138374624e-05, + "loss": 0.5129, + "num_input_tokens_seen": 18612016, + "step": 12235 + }, + { + "epoch": 0.39178029575571344, + "grad_norm": 0.5528329610824585, + "learning_rate": 3.9312413067926854e-05, + "loss": 0.4368, + "num_input_tokens_seen": 18619552, + "step": 12240 + }, + { + "epoch": 0.39194033672620193, + "grad_norm": 0.501373827457428, + "learning_rate": 3.9304362544953506e-05, + "loss": 0.5853, + "num_input_tokens_seen": 18627168, + "step": 12245 + }, + { + "epoch": 0.39210037769669037, + "grad_norm": 0.8011793494224548, + "learning_rate": 3.929630981606769e-05, + "loss": 0.5156, + "num_input_tokens_seen": 18634560, + "step": 12250 + }, + { + "epoch": 0.3922604186671788, + "grad_norm": 0.7525154948234558, + "learning_rate": 3.928825488251124e-05, + "loss": 0.4647, + "num_input_tokens_seen": 18642448, + "step": 12255 + }, + { + "epoch": 0.39242045963766725, + "grad_norm": 1.0241366624832153, + "learning_rate": 3.9280197745526344e-05, + "loss": 0.4955, + "num_input_tokens_seen": 18650016, + "step": 12260 + }, + { + "epoch": 0.3925805006081557, + "grad_norm": 1.015647530555725, + "learning_rate": 3.9272138406355495e-05, + "loss": 0.5209, + "num_input_tokens_seen": 18657632, + "step": 12265 + }, + { + "epoch": 0.39274054157864413, + "grad_norm": 0.5984101891517639, + "learning_rate": 3.926407686624154e-05, + "loss": 0.5912, + "num_input_tokens_seen": 18665168, + "step": 12270 + }, + { + "epoch": 0.39290058254913257, + "grad_norm": 0.9156619310379028, + "learning_rate": 3.9256013126427684e-05, + "loss": 0.6134, + "num_input_tokens_seen": 18672448, + "step": 12275 + }, + { + "epoch": 0.393060623519621, + "grad_norm": 1.4016727209091187, + "learning_rate": 3.9247947188157455e-05, + "loss": 0.316, + "num_input_tokens_seen": 18680688, + "step": 12280 + }, + { + "epoch": 0.39322066449010945, + "grad_norm": 1.7576743364334106, + "learning_rate": 3.9239879052674715e-05, + "loss": 0.7378, + "num_input_tokens_seen": 18688464, + "step": 12285 + }, + { + "epoch": 0.3933807054605979, + "grad_norm": 1.1142983436584473, + "learning_rate": 3.9231808721223673e-05, + "loss": 0.5312, + "num_input_tokens_seen": 18696128, + "step": 12290 + }, + { + "epoch": 0.3935407464310864, + "grad_norm": 0.8795405626296997, + "learning_rate": 3.9223736195048886e-05, + "loss": 0.4421, + "num_input_tokens_seen": 18704016, + "step": 12295 + }, + { + "epoch": 0.3937007874015748, + "grad_norm": 1.3608801364898682, + "learning_rate": 3.921566147539523e-05, + "loss": 0.6116, + "num_input_tokens_seen": 18711488, + "step": 12300 + }, + { + "epoch": 0.39386082837206327, + "grad_norm": 0.8204751014709473, + "learning_rate": 3.920758456350792e-05, + "loss": 0.4178, + "num_input_tokens_seen": 18719136, + "step": 12305 + }, + { + "epoch": 0.3940208693425517, + "grad_norm": 0.9759113192558289, + "learning_rate": 3.919950546063253e-05, + "loss": 0.5721, + "num_input_tokens_seen": 18726928, + "step": 12310 + }, + { + "epoch": 0.39418091031304014, + "grad_norm": 1.0225117206573486, + "learning_rate": 3.919142416801496e-05, + "loss": 0.4495, + "num_input_tokens_seen": 18734240, + "step": 12315 + }, + { + "epoch": 0.3943409512835286, + "grad_norm": 1.114543080329895, + "learning_rate": 3.918334068690144e-05, + "loss": 0.5098, + "num_input_tokens_seen": 18742128, + "step": 12320 + }, + { + "epoch": 0.394500992254017, + "grad_norm": 1.0369749069213867, + "learning_rate": 3.917525501853855e-05, + "loss": 0.3729, + "num_input_tokens_seen": 18749632, + "step": 12325 + }, + { + "epoch": 0.39466103322450546, + "grad_norm": 0.5672351121902466, + "learning_rate": 3.916716716417319e-05, + "loss": 0.4243, + "num_input_tokens_seen": 18757040, + "step": 12330 + }, + { + "epoch": 0.3948210741949939, + "grad_norm": 1.0278828144073486, + "learning_rate": 3.915907712505263e-05, + "loss": 0.4116, + "num_input_tokens_seen": 18764256, + "step": 12335 + }, + { + "epoch": 0.39498111516548234, + "grad_norm": 0.7498008012771606, + "learning_rate": 3.915098490242444e-05, + "loss": 0.5553, + "num_input_tokens_seen": 18771616, + "step": 12340 + }, + { + "epoch": 0.3951411561359708, + "grad_norm": 0.9348205924034119, + "learning_rate": 3.914289049753654e-05, + "loss": 0.6555, + "num_input_tokens_seen": 18779200, + "step": 12345 + }, + { + "epoch": 0.3953011971064593, + "grad_norm": 1.0781066417694092, + "learning_rate": 3.913479391163719e-05, + "loss": 0.5425, + "num_input_tokens_seen": 18787392, + "step": 12350 + }, + { + "epoch": 0.3954612380769477, + "grad_norm": 0.8369196653366089, + "learning_rate": 3.9126695145975e-05, + "loss": 0.441, + "num_input_tokens_seen": 18795216, + "step": 12355 + }, + { + "epoch": 0.39562127904743616, + "grad_norm": 0.8472725749015808, + "learning_rate": 3.911859420179889e-05, + "loss": 0.7122, + "num_input_tokens_seen": 18802608, + "step": 12360 + }, + { + "epoch": 0.3957813200179246, + "grad_norm": 0.5659597516059875, + "learning_rate": 3.911049108035813e-05, + "loss": 0.414, + "num_input_tokens_seen": 18809920, + "step": 12365 + }, + { + "epoch": 0.39594136098841304, + "grad_norm": 0.5716736912727356, + "learning_rate": 3.910238578290232e-05, + "loss": 0.3805, + "num_input_tokens_seen": 18817136, + "step": 12370 + }, + { + "epoch": 0.3961014019589015, + "grad_norm": 0.7623263597488403, + "learning_rate": 3.90942783106814e-05, + "loss": 0.377, + "num_input_tokens_seen": 18824624, + "step": 12375 + }, + { + "epoch": 0.3962614429293899, + "grad_norm": 0.554165780544281, + "learning_rate": 3.908616866494564e-05, + "loss": 0.5403, + "num_input_tokens_seen": 18832272, + "step": 12380 + }, + { + "epoch": 0.39642148389987836, + "grad_norm": 1.2473504543304443, + "learning_rate": 3.907805684694566e-05, + "loss": 0.7055, + "num_input_tokens_seen": 18840528, + "step": 12385 + }, + { + "epoch": 0.3965815248703668, + "grad_norm": 1.0429811477661133, + "learning_rate": 3.90699428579324e-05, + "loss": 0.5393, + "num_input_tokens_seen": 18848448, + "step": 12390 + }, + { + "epoch": 0.39674156584085524, + "grad_norm": 0.90384840965271, + "learning_rate": 3.906182669915713e-05, + "loss": 0.5946, + "num_input_tokens_seen": 18856416, + "step": 12395 + }, + { + "epoch": 0.3969016068113437, + "grad_norm": 1.2717218399047852, + "learning_rate": 3.9053708371871476e-05, + "loss": 0.7766, + "num_input_tokens_seen": 18863792, + "step": 12400 + }, + { + "epoch": 0.3969016068113437, + "eval_loss": 0.5456917881965637, + "eval_runtime": 558.3318, + "eval_samples_per_second": 24.871, + "eval_steps_per_second": 12.435, + "num_input_tokens_seen": 18863792, + "step": 12400 + }, + { + "epoch": 0.39706164778183217, + "grad_norm": 0.5629054307937622, + "learning_rate": 3.904558787732738e-05, + "loss": 0.5183, + "num_input_tokens_seen": 18871824, + "step": 12405 + }, + { + "epoch": 0.3972216887523206, + "grad_norm": 0.891152560710907, + "learning_rate": 3.9037465216777135e-05, + "loss": 0.4045, + "num_input_tokens_seen": 18879264, + "step": 12410 + }, + { + "epoch": 0.39738172972280905, + "grad_norm": 1.7127469778060913, + "learning_rate": 3.902934039147334e-05, + "loss": 0.6708, + "num_input_tokens_seen": 18886416, + "step": 12415 + }, + { + "epoch": 0.3975417706932975, + "grad_norm": 0.6176135540008545, + "learning_rate": 3.902121340266894e-05, + "loss": 0.3639, + "num_input_tokens_seen": 18893984, + "step": 12420 + }, + { + "epoch": 0.39770181166378593, + "grad_norm": 0.6396914720535278, + "learning_rate": 3.9013084251617246e-05, + "loss": 0.4302, + "num_input_tokens_seen": 18901280, + "step": 12425 + }, + { + "epoch": 0.39786185263427437, + "grad_norm": 0.7147884964942932, + "learning_rate": 3.9004952939571865e-05, + "loss": 0.5741, + "num_input_tokens_seen": 18909008, + "step": 12430 + }, + { + "epoch": 0.3980218936047628, + "grad_norm": 0.6381645202636719, + "learning_rate": 3.899681946778673e-05, + "loss": 0.6219, + "num_input_tokens_seen": 18916672, + "step": 12435 + }, + { + "epoch": 0.39818193457525125, + "grad_norm": 0.765694797039032, + "learning_rate": 3.898868383751615e-05, + "loss": 0.4044, + "num_input_tokens_seen": 18924304, + "step": 12440 + }, + { + "epoch": 0.3983419755457397, + "grad_norm": 0.6608931422233582, + "learning_rate": 3.8980546050014724e-05, + "loss": 0.4218, + "num_input_tokens_seen": 18932288, + "step": 12445 + }, + { + "epoch": 0.39850201651622813, + "grad_norm": 1.1821199655532837, + "learning_rate": 3.897240610653741e-05, + "loss": 0.6273, + "num_input_tokens_seen": 18939408, + "step": 12450 + }, + { + "epoch": 0.3986620574867166, + "grad_norm": 0.779896080493927, + "learning_rate": 3.896426400833948e-05, + "loss": 0.3541, + "num_input_tokens_seen": 18946720, + "step": 12455 + }, + { + "epoch": 0.39882209845720507, + "grad_norm": 0.508592426776886, + "learning_rate": 3.895611975667656e-05, + "loss": 0.456, + "num_input_tokens_seen": 18954576, + "step": 12460 + }, + { + "epoch": 0.3989821394276935, + "grad_norm": 1.271781086921692, + "learning_rate": 3.8947973352804584e-05, + "loss": 0.7924, + "num_input_tokens_seen": 18962512, + "step": 12465 + }, + { + "epoch": 0.39914218039818194, + "grad_norm": 0.5226426720619202, + "learning_rate": 3.893982479797984e-05, + "loss": 0.5043, + "num_input_tokens_seen": 18970384, + "step": 12470 + }, + { + "epoch": 0.3993022213686704, + "grad_norm": 0.5746468901634216, + "learning_rate": 3.8931674093458926e-05, + "loss": 0.3998, + "num_input_tokens_seen": 18977792, + "step": 12475 + }, + { + "epoch": 0.3994622623391588, + "grad_norm": 0.7647683620452881, + "learning_rate": 3.89235212404988e-05, + "loss": 0.4552, + "num_input_tokens_seen": 18985088, + "step": 12480 + }, + { + "epoch": 0.39962230330964726, + "grad_norm": 0.6520121693611145, + "learning_rate": 3.891536624035672e-05, + "loss": 0.3096, + "num_input_tokens_seen": 18992880, + "step": 12485 + }, + { + "epoch": 0.3997823442801357, + "grad_norm": 0.6628361940383911, + "learning_rate": 3.8907209094290295e-05, + "loss": 0.5068, + "num_input_tokens_seen": 19001408, + "step": 12490 + }, + { + "epoch": 0.39994238525062414, + "grad_norm": 0.5844176411628723, + "learning_rate": 3.8899049803557466e-05, + "loss": 0.5923, + "num_input_tokens_seen": 19009232, + "step": 12495 + }, + { + "epoch": 0.4001024262211126, + "grad_norm": 0.5365328192710876, + "learning_rate": 3.889088836941648e-05, + "loss": 0.4767, + "num_input_tokens_seen": 19016384, + "step": 12500 + }, + { + "epoch": 0.400262467191601, + "grad_norm": 0.47933197021484375, + "learning_rate": 3.8882724793125946e-05, + "loss": 0.4086, + "num_input_tokens_seen": 19023952, + "step": 12505 + }, + { + "epoch": 0.4004225081620895, + "grad_norm": 0.9909186363220215, + "learning_rate": 3.8874559075944794e-05, + "loss": 0.6608, + "num_input_tokens_seen": 19031408, + "step": 12510 + }, + { + "epoch": 0.40058254913257796, + "grad_norm": 0.6415579319000244, + "learning_rate": 3.886639121913227e-05, + "loss": 0.6321, + "num_input_tokens_seen": 19038384, + "step": 12515 + }, + { + "epoch": 0.4007425901030664, + "grad_norm": 0.7571452260017395, + "learning_rate": 3.885822122394797e-05, + "loss": 0.6486, + "num_input_tokens_seen": 19046032, + "step": 12520 + }, + { + "epoch": 0.40090263107355484, + "grad_norm": 0.5507268905639648, + "learning_rate": 3.8850049091651794e-05, + "loss": 0.4698, + "num_input_tokens_seen": 19053760, + "step": 12525 + }, + { + "epoch": 0.4010626720440433, + "grad_norm": 0.4625451862812042, + "learning_rate": 3.8841874823504e-05, + "loss": 0.4302, + "num_input_tokens_seen": 19061168, + "step": 12530 + }, + { + "epoch": 0.4012227130145317, + "grad_norm": 1.0476484298706055, + "learning_rate": 3.8833698420765157e-05, + "loss": 0.4681, + "num_input_tokens_seen": 19068544, + "step": 12535 + }, + { + "epoch": 0.40138275398502016, + "grad_norm": 1.1496784687042236, + "learning_rate": 3.882551988469618e-05, + "loss": 0.6298, + "num_input_tokens_seen": 19075792, + "step": 12540 + }, + { + "epoch": 0.4015427949555086, + "grad_norm": 0.5102055072784424, + "learning_rate": 3.881733921655829e-05, + "loss": 0.5392, + "num_input_tokens_seen": 19082832, + "step": 12545 + }, + { + "epoch": 0.40170283592599704, + "grad_norm": 0.6293399333953857, + "learning_rate": 3.8809156417613054e-05, + "loss": 0.472, + "num_input_tokens_seen": 19090736, + "step": 12550 + }, + { + "epoch": 0.4018628768964855, + "grad_norm": 0.4415757656097412, + "learning_rate": 3.8800971489122364e-05, + "loss": 0.3863, + "num_input_tokens_seen": 19098496, + "step": 12555 + }, + { + "epoch": 0.402022917866974, + "grad_norm": 0.5553296208381653, + "learning_rate": 3.8792784432348434e-05, + "loss": 0.3707, + "num_input_tokens_seen": 19105872, + "step": 12560 + }, + { + "epoch": 0.4021829588374624, + "grad_norm": 0.9519236087799072, + "learning_rate": 3.878459524855381e-05, + "loss": 0.5935, + "num_input_tokens_seen": 19113328, + "step": 12565 + }, + { + "epoch": 0.40234299980795085, + "grad_norm": 0.7833003997802734, + "learning_rate": 3.8776403939001384e-05, + "loss": 0.3907, + "num_input_tokens_seen": 19120976, + "step": 12570 + }, + { + "epoch": 0.4025030407784393, + "grad_norm": 1.1458886861801147, + "learning_rate": 3.876821050495433e-05, + "loss": 0.4956, + "num_input_tokens_seen": 19128160, + "step": 12575 + }, + { + "epoch": 0.40266308174892773, + "grad_norm": 0.4733797013759613, + "learning_rate": 3.87600149476762e-05, + "loss": 0.3834, + "num_input_tokens_seen": 19135568, + "step": 12580 + }, + { + "epoch": 0.40282312271941617, + "grad_norm": 0.5720117688179016, + "learning_rate": 3.8751817268430843e-05, + "loss": 0.3827, + "num_input_tokens_seen": 19143040, + "step": 12585 + }, + { + "epoch": 0.4029831636899046, + "grad_norm": 1.4856139421463013, + "learning_rate": 3.8743617468482464e-05, + "loss": 0.5057, + "num_input_tokens_seen": 19150256, + "step": 12590 + }, + { + "epoch": 0.40314320466039305, + "grad_norm": 0.4621201157569885, + "learning_rate": 3.8735415549095535e-05, + "loss": 0.4502, + "num_input_tokens_seen": 19157920, + "step": 12595 + }, + { + "epoch": 0.4033032456308815, + "grad_norm": 0.5861800312995911, + "learning_rate": 3.8727211511534934e-05, + "loss": 0.4641, + "num_input_tokens_seen": 19165216, + "step": 12600 + }, + { + "epoch": 0.4033032456308815, + "eval_loss": 0.5443362593650818, + "eval_runtime": 560.7649, + "eval_samples_per_second": 24.763, + "eval_steps_per_second": 12.381, + "num_input_tokens_seen": 19165216, + "step": 12600 + }, + { + "epoch": 0.40346328660136993, + "grad_norm": 1.622915506362915, + "learning_rate": 3.8719005357065804e-05, + "loss": 0.6549, + "num_input_tokens_seen": 19172768, + "step": 12605 + }, + { + "epoch": 0.40362332757185837, + "grad_norm": 0.7498023509979248, + "learning_rate": 3.8710797086953645e-05, + "loss": 0.4866, + "num_input_tokens_seen": 19180240, + "step": 12610 + }, + { + "epoch": 0.40378336854234687, + "grad_norm": 1.2851135730743408, + "learning_rate": 3.870258670246427e-05, + "loss": 0.5684, + "num_input_tokens_seen": 19187472, + "step": 12615 + }, + { + "epoch": 0.4039434095128353, + "grad_norm": 0.8570791482925415, + "learning_rate": 3.869437420486384e-05, + "loss": 0.5253, + "num_input_tokens_seen": 19194768, + "step": 12620 + }, + { + "epoch": 0.40410345048332375, + "grad_norm": 0.7527147531509399, + "learning_rate": 3.8686159595418805e-05, + "loss": 0.3826, + "num_input_tokens_seen": 19202608, + "step": 12625 + }, + { + "epoch": 0.4042634914538122, + "grad_norm": 0.6882185339927673, + "learning_rate": 3.867794287539597e-05, + "loss": 0.6241, + "num_input_tokens_seen": 19210672, + "step": 12630 + }, + { + "epoch": 0.4044235324243006, + "grad_norm": 0.60903400182724, + "learning_rate": 3.866972404606245e-05, + "loss": 0.6386, + "num_input_tokens_seen": 19218320, + "step": 12635 + }, + { + "epoch": 0.40458357339478906, + "grad_norm": 0.3944767713546753, + "learning_rate": 3.866150310868571e-05, + "loss": 0.3552, + "num_input_tokens_seen": 19225536, + "step": 12640 + }, + { + "epoch": 0.4047436143652775, + "grad_norm": 3.061568021774292, + "learning_rate": 3.8653280064533506e-05, + "loss": 0.5125, + "num_input_tokens_seen": 19233248, + "step": 12645 + }, + { + "epoch": 0.40490365533576594, + "grad_norm": 0.702171802520752, + "learning_rate": 3.864505491487394e-05, + "loss": 0.3818, + "num_input_tokens_seen": 19240864, + "step": 12650 + }, + { + "epoch": 0.4050636963062544, + "grad_norm": 1.382615327835083, + "learning_rate": 3.8636827660975414e-05, + "loss": 0.52, + "num_input_tokens_seen": 19248384, + "step": 12655 + }, + { + "epoch": 0.4052237372767428, + "grad_norm": 0.946197509765625, + "learning_rate": 3.862859830410671e-05, + "loss": 0.449, + "num_input_tokens_seen": 19255744, + "step": 12660 + }, + { + "epoch": 0.4053837782472313, + "grad_norm": 1.096174716949463, + "learning_rate": 3.862036684553688e-05, + "loss": 0.5703, + "num_input_tokens_seen": 19263008, + "step": 12665 + }, + { + "epoch": 0.40554381921771976, + "grad_norm": 0.6818562746047974, + "learning_rate": 3.8612133286535314e-05, + "loss": 0.3961, + "num_input_tokens_seen": 19270368, + "step": 12670 + }, + { + "epoch": 0.4057038601882082, + "grad_norm": 1.4893484115600586, + "learning_rate": 3.860389762837173e-05, + "loss": 0.726, + "num_input_tokens_seen": 19278016, + "step": 12675 + }, + { + "epoch": 0.40586390115869664, + "grad_norm": 0.38410520553588867, + "learning_rate": 3.859565987231618e-05, + "loss": 0.3739, + "num_input_tokens_seen": 19285728, + "step": 12680 + }, + { + "epoch": 0.4060239421291851, + "grad_norm": 0.3986121416091919, + "learning_rate": 3.858742001963902e-05, + "loss": 0.4659, + "num_input_tokens_seen": 19293312, + "step": 12685 + }, + { + "epoch": 0.4061839830996735, + "grad_norm": 0.3261299729347229, + "learning_rate": 3.857917807161094e-05, + "loss": 0.4754, + "num_input_tokens_seen": 19300560, + "step": 12690 + }, + { + "epoch": 0.40634402407016196, + "grad_norm": 1.00960111618042, + "learning_rate": 3.857093402950296e-05, + "loss": 0.4759, + "num_input_tokens_seen": 19308432, + "step": 12695 + }, + { + "epoch": 0.4065040650406504, + "grad_norm": 0.4816964864730835, + "learning_rate": 3.8562687894586414e-05, + "loss": 0.477, + "num_input_tokens_seen": 19316592, + "step": 12700 + }, + { + "epoch": 0.40666410601113884, + "grad_norm": 1.1690806150436401, + "learning_rate": 3.8554439668132946e-05, + "loss": 0.4896, + "num_input_tokens_seen": 19324256, + "step": 12705 + }, + { + "epoch": 0.4068241469816273, + "grad_norm": 0.6374971866607666, + "learning_rate": 3.854618935141455e-05, + "loss": 0.7366, + "num_input_tokens_seen": 19331792, + "step": 12710 + }, + { + "epoch": 0.4069841879521157, + "grad_norm": 1.147200345993042, + "learning_rate": 3.8537936945703525e-05, + "loss": 0.4742, + "num_input_tokens_seen": 19339424, + "step": 12715 + }, + { + "epoch": 0.4071442289226042, + "grad_norm": 1.024038553237915, + "learning_rate": 3.852968245227249e-05, + "loss": 0.6205, + "num_input_tokens_seen": 19346672, + "step": 12720 + }, + { + "epoch": 0.40730426989309265, + "grad_norm": 1.1728405952453613, + "learning_rate": 3.85214258723944e-05, + "loss": 0.5466, + "num_input_tokens_seen": 19354512, + "step": 12725 + }, + { + "epoch": 0.4074643108635811, + "grad_norm": 1.0197765827178955, + "learning_rate": 3.8513167207342524e-05, + "loss": 0.5837, + "num_input_tokens_seen": 19361568, + "step": 12730 + }, + { + "epoch": 0.40762435183406953, + "grad_norm": 0.2988959848880768, + "learning_rate": 3.850490645839044e-05, + "loss": 0.432, + "num_input_tokens_seen": 19369296, + "step": 12735 + }, + { + "epoch": 0.40778439280455797, + "grad_norm": 0.7109142541885376, + "learning_rate": 3.849664362681207e-05, + "loss": 0.591, + "num_input_tokens_seen": 19376896, + "step": 12740 + }, + { + "epoch": 0.4079444337750464, + "grad_norm": 1.1199548244476318, + "learning_rate": 3.848837871388165e-05, + "loss": 0.5563, + "num_input_tokens_seen": 19384416, + "step": 12745 + }, + { + "epoch": 0.40810447474553485, + "grad_norm": 0.7714565992355347, + "learning_rate": 3.848011172087371e-05, + "loss": 0.5042, + "num_input_tokens_seen": 19392240, + "step": 12750 + }, + { + "epoch": 0.4082645157160233, + "grad_norm": 0.8118323087692261, + "learning_rate": 3.847184264906315e-05, + "loss": 0.4287, + "num_input_tokens_seen": 19400160, + "step": 12755 + }, + { + "epoch": 0.40842455668651173, + "grad_norm": 0.9011208415031433, + "learning_rate": 3.846357149972516e-05, + "loss": 0.3782, + "num_input_tokens_seen": 19407600, + "step": 12760 + }, + { + "epoch": 0.40858459765700017, + "grad_norm": 0.8468871116638184, + "learning_rate": 3.8455298274135246e-05, + "loss": 0.5264, + "num_input_tokens_seen": 19414992, + "step": 12765 + }, + { + "epoch": 0.4087446386274886, + "grad_norm": 0.4801170229911804, + "learning_rate": 3.8447022973569254e-05, + "loss": 0.5293, + "num_input_tokens_seen": 19422656, + "step": 12770 + }, + { + "epoch": 0.4089046795979771, + "grad_norm": 0.7376617789268494, + "learning_rate": 3.843874559930332e-05, + "loss": 0.5754, + "num_input_tokens_seen": 19429936, + "step": 12775 + }, + { + "epoch": 0.40906472056846555, + "grad_norm": 0.34936872124671936, + "learning_rate": 3.843046615261394e-05, + "loss": 0.6235, + "num_input_tokens_seen": 19437472, + "step": 12780 + }, + { + "epoch": 0.409224761538954, + "grad_norm": 0.8121026158332825, + "learning_rate": 3.842218463477791e-05, + "loss": 0.6619, + "num_input_tokens_seen": 19445248, + "step": 12785 + }, + { + "epoch": 0.4093848025094424, + "grad_norm": 0.959132194519043, + "learning_rate": 3.841390104707233e-05, + "loss": 0.5085, + "num_input_tokens_seen": 19452784, + "step": 12790 + }, + { + "epoch": 0.40954484347993086, + "grad_norm": 0.9932464957237244, + "learning_rate": 3.8405615390774643e-05, + "loss": 0.4518, + "num_input_tokens_seen": 19460256, + "step": 12795 + }, + { + "epoch": 0.4097048844504193, + "grad_norm": 0.6109856367111206, + "learning_rate": 3.839732766716259e-05, + "loss": 0.5321, + "num_input_tokens_seen": 19467712, + "step": 12800 + }, + { + "epoch": 0.4097048844504193, + "eval_loss": 0.5429202914237976, + "eval_runtime": 561.1845, + "eval_samples_per_second": 24.744, + "eval_steps_per_second": 12.372, + "num_input_tokens_seen": 19467712, + "step": 12800 + }, + { + "epoch": 0.40986492542090774, + "grad_norm": 2.063464641571045, + "learning_rate": 3.838903787751425e-05, + "loss": 0.6406, + "num_input_tokens_seen": 19475360, + "step": 12805 + }, + { + "epoch": 0.4100249663913962, + "grad_norm": 0.529048502445221, + "learning_rate": 3.838074602310802e-05, + "loss": 0.3949, + "num_input_tokens_seen": 19482400, + "step": 12810 + }, + { + "epoch": 0.4101850073618846, + "grad_norm": 0.5769007205963135, + "learning_rate": 3.837245210522258e-05, + "loss": 0.4392, + "num_input_tokens_seen": 19490016, + "step": 12815 + }, + { + "epoch": 0.41034504833237306, + "grad_norm": 1.216174840927124, + "learning_rate": 3.8364156125136996e-05, + "loss": 0.4437, + "num_input_tokens_seen": 19497296, + "step": 12820 + }, + { + "epoch": 0.41050508930286156, + "grad_norm": 0.6896365284919739, + "learning_rate": 3.835585808413059e-05, + "loss": 0.5187, + "num_input_tokens_seen": 19504832, + "step": 12825 + }, + { + "epoch": 0.41066513027335, + "grad_norm": 0.8895873427391052, + "learning_rate": 3.8347557983483024e-05, + "loss": 0.5907, + "num_input_tokens_seen": 19512304, + "step": 12830 + }, + { + "epoch": 0.41082517124383844, + "grad_norm": 0.6330602169036865, + "learning_rate": 3.833925582447428e-05, + "loss": 0.615, + "num_input_tokens_seen": 19520176, + "step": 12835 + }, + { + "epoch": 0.4109852122143269, + "grad_norm": 0.4301898181438446, + "learning_rate": 3.8330951608384656e-05, + "loss": 0.5115, + "num_input_tokens_seen": 19527984, + "step": 12840 + }, + { + "epoch": 0.4111452531848153, + "grad_norm": 1.1042356491088867, + "learning_rate": 3.832264533649477e-05, + "loss": 0.514, + "num_input_tokens_seen": 19535168, + "step": 12845 + }, + { + "epoch": 0.41130529415530376, + "grad_norm": 0.8696669340133667, + "learning_rate": 3.8314337010085555e-05, + "loss": 0.4293, + "num_input_tokens_seen": 19542624, + "step": 12850 + }, + { + "epoch": 0.4114653351257922, + "grad_norm": 0.6799540519714355, + "learning_rate": 3.830602663043824e-05, + "loss": 0.5885, + "num_input_tokens_seen": 19549824, + "step": 12855 + }, + { + "epoch": 0.41162537609628064, + "grad_norm": 0.2988796830177307, + "learning_rate": 3.8297714198834414e-05, + "loss": 0.5021, + "num_input_tokens_seen": 19557120, + "step": 12860 + }, + { + "epoch": 0.4117854170667691, + "grad_norm": 1.2220302820205688, + "learning_rate": 3.828939971655595e-05, + "loss": 0.6221, + "num_input_tokens_seen": 19564880, + "step": 12865 + }, + { + "epoch": 0.4119454580372575, + "grad_norm": 1.0964481830596924, + "learning_rate": 3.828108318488505e-05, + "loss": 0.6408, + "num_input_tokens_seen": 19572272, + "step": 12870 + }, + { + "epoch": 0.41210549900774596, + "grad_norm": 0.34112638235092163, + "learning_rate": 3.8272764605104216e-05, + "loss": 0.4198, + "num_input_tokens_seen": 19579824, + "step": 12875 + }, + { + "epoch": 0.41226553997823445, + "grad_norm": 0.7087218761444092, + "learning_rate": 3.826444397849628e-05, + "loss": 0.5769, + "num_input_tokens_seen": 19587712, + "step": 12880 + }, + { + "epoch": 0.4124255809487229, + "grad_norm": 1.0908405780792236, + "learning_rate": 3.825612130634439e-05, + "loss": 0.5094, + "num_input_tokens_seen": 19594720, + "step": 12885 + }, + { + "epoch": 0.41258562191921133, + "grad_norm": 0.6860224008560181, + "learning_rate": 3.824779658993202e-05, + "loss": 0.4797, + "num_input_tokens_seen": 19602336, + "step": 12890 + }, + { + "epoch": 0.41274566288969977, + "grad_norm": 0.3166504502296448, + "learning_rate": 3.823946983054292e-05, + "loss": 0.5293, + "num_input_tokens_seen": 19609760, + "step": 12895 + }, + { + "epoch": 0.4129057038601882, + "grad_norm": 0.5046604871749878, + "learning_rate": 3.82311410294612e-05, + "loss": 0.4616, + "num_input_tokens_seen": 19617136, + "step": 12900 + }, + { + "epoch": 0.41306574483067665, + "grad_norm": 0.6848369836807251, + "learning_rate": 3.822281018797127e-05, + "loss": 0.4123, + "num_input_tokens_seen": 19624960, + "step": 12905 + }, + { + "epoch": 0.4132257858011651, + "grad_norm": 0.923821210861206, + "learning_rate": 3.821447730735783e-05, + "loss": 0.5641, + "num_input_tokens_seen": 19632176, + "step": 12910 + }, + { + "epoch": 0.41338582677165353, + "grad_norm": 0.7640395760536194, + "learning_rate": 3.820614238890592e-05, + "loss": 0.4998, + "num_input_tokens_seen": 19639248, + "step": 12915 + }, + { + "epoch": 0.41354586774214197, + "grad_norm": 1.3248111009597778, + "learning_rate": 3.819780543390091e-05, + "loss": 0.5388, + "num_input_tokens_seen": 19647280, + "step": 12920 + }, + { + "epoch": 0.4137059087126304, + "grad_norm": 0.6908413767814636, + "learning_rate": 3.818946644362844e-05, + "loss": 0.4885, + "num_input_tokens_seen": 19654256, + "step": 12925 + }, + { + "epoch": 0.4138659496831189, + "grad_norm": 0.6096101999282837, + "learning_rate": 3.81811254193745e-05, + "loss": 0.3928, + "num_input_tokens_seen": 19662336, + "step": 12930 + }, + { + "epoch": 0.41402599065360735, + "grad_norm": 0.3314949870109558, + "learning_rate": 3.8172782362425366e-05, + "loss": 0.4591, + "num_input_tokens_seen": 19670032, + "step": 12935 + }, + { + "epoch": 0.4141860316240958, + "grad_norm": 0.6430590152740479, + "learning_rate": 3.816443727406765e-05, + "loss": 0.4566, + "num_input_tokens_seen": 19677520, + "step": 12940 + }, + { + "epoch": 0.4143460725945842, + "grad_norm": 1.371550440788269, + "learning_rate": 3.815609015558829e-05, + "loss": 0.5784, + "num_input_tokens_seen": 19685184, + "step": 12945 + }, + { + "epoch": 0.41450611356507266, + "grad_norm": 0.5637928247451782, + "learning_rate": 3.814774100827448e-05, + "loss": 0.3977, + "num_input_tokens_seen": 19692608, + "step": 12950 + }, + { + "epoch": 0.4146661545355611, + "grad_norm": 0.9519808292388916, + "learning_rate": 3.813938983341379e-05, + "loss": 0.4572, + "num_input_tokens_seen": 19700272, + "step": 12955 + }, + { + "epoch": 0.41482619550604954, + "grad_norm": 0.8902126550674438, + "learning_rate": 3.813103663229407e-05, + "loss": 0.5966, + "num_input_tokens_seen": 19708144, + "step": 12960 + }, + { + "epoch": 0.414986236476538, + "grad_norm": 0.32155513763427734, + "learning_rate": 3.812268140620349e-05, + "loss": 0.5334, + "num_input_tokens_seen": 19715872, + "step": 12965 + }, + { + "epoch": 0.4151462774470264, + "grad_norm": 0.9090753197669983, + "learning_rate": 3.811432415643051e-05, + "loss": 0.6845, + "num_input_tokens_seen": 19723520, + "step": 12970 + }, + { + "epoch": 0.41530631841751486, + "grad_norm": 0.5957728624343872, + "learning_rate": 3.8105964884263954e-05, + "loss": 0.5986, + "num_input_tokens_seen": 19731104, + "step": 12975 + }, + { + "epoch": 0.4154663593880033, + "grad_norm": 1.1498334407806396, + "learning_rate": 3.809760359099291e-05, + "loss": 0.4294, + "num_input_tokens_seen": 19738576, + "step": 12980 + }, + { + "epoch": 0.4156264003584918, + "grad_norm": 0.816080629825592, + "learning_rate": 3.8089240277906804e-05, + "loss": 0.4285, + "num_input_tokens_seen": 19746048, + "step": 12985 + }, + { + "epoch": 0.41578644132898024, + "grad_norm": 2.7005038261413574, + "learning_rate": 3.808087494629535e-05, + "loss": 0.8766, + "num_input_tokens_seen": 19753808, + "step": 12990 + }, + { + "epoch": 0.4159464822994687, + "grad_norm": 0.9052961468696594, + "learning_rate": 3.8072507597448595e-05, + "loss": 0.5452, + "num_input_tokens_seen": 19760976, + "step": 12995 + }, + { + "epoch": 0.4161065232699571, + "grad_norm": 0.837519109249115, + "learning_rate": 3.806413823265689e-05, + "loss": 0.5, + "num_input_tokens_seen": 19768464, + "step": 13000 + }, + { + "epoch": 0.4161065232699571, + "eval_loss": 0.5417889952659607, + "eval_runtime": 561.412, + "eval_samples_per_second": 24.734, + "eval_steps_per_second": 12.367, + "num_input_tokens_seen": 19768464, + "step": 13000 + }, + { + "epoch": 0.41626656424044556, + "grad_norm": 1.19558584690094, + "learning_rate": 3.805576685321089e-05, + "loss": 0.6044, + "num_input_tokens_seen": 19776368, + "step": 13005 + }, + { + "epoch": 0.416426605210934, + "grad_norm": 1.058181643486023, + "learning_rate": 3.804739346040158e-05, + "loss": 0.445, + "num_input_tokens_seen": 19784000, + "step": 13010 + }, + { + "epoch": 0.41658664618142244, + "grad_norm": 1.8482639789581299, + "learning_rate": 3.8039018055520234e-05, + "loss": 0.5398, + "num_input_tokens_seen": 19792144, + "step": 13015 + }, + { + "epoch": 0.4167466871519109, + "grad_norm": 0.6072773337364197, + "learning_rate": 3.803064063985844e-05, + "loss": 0.399, + "num_input_tokens_seen": 19799600, + "step": 13020 + }, + { + "epoch": 0.4169067281223993, + "grad_norm": 0.40749746561050415, + "learning_rate": 3.802226121470811e-05, + "loss": 0.4462, + "num_input_tokens_seen": 19806848, + "step": 13025 + }, + { + "epoch": 0.41706676909288776, + "grad_norm": 0.7275246977806091, + "learning_rate": 3.801387978136145e-05, + "loss": 0.4709, + "num_input_tokens_seen": 19814528, + "step": 13030 + }, + { + "epoch": 0.4172268100633762, + "grad_norm": 0.8508867621421814, + "learning_rate": 3.800549634111099e-05, + "loss": 0.7089, + "num_input_tokens_seen": 19822384, + "step": 13035 + }, + { + "epoch": 0.4173868510338647, + "grad_norm": 2.3427982330322266, + "learning_rate": 3.799711089524955e-05, + "loss": 0.7518, + "num_input_tokens_seen": 19830336, + "step": 13040 + }, + { + "epoch": 0.41754689200435313, + "grad_norm": 0.5763950943946838, + "learning_rate": 3.7988723445070285e-05, + "loss": 0.576, + "num_input_tokens_seen": 19838000, + "step": 13045 + }, + { + "epoch": 0.41770693297484157, + "grad_norm": 0.35932672023773193, + "learning_rate": 3.798033399186663e-05, + "loss": 0.4759, + "num_input_tokens_seen": 19845536, + "step": 13050 + }, + { + "epoch": 0.41786697394533, + "grad_norm": 0.7642999291419983, + "learning_rate": 3.797194253693237e-05, + "loss": 0.4712, + "num_input_tokens_seen": 19853648, + "step": 13055 + }, + { + "epoch": 0.41802701491581845, + "grad_norm": 0.5781455636024475, + "learning_rate": 3.796354908156153e-05, + "loss": 0.4842, + "num_input_tokens_seen": 19861744, + "step": 13060 + }, + { + "epoch": 0.4181870558863069, + "grad_norm": 1.4282143115997314, + "learning_rate": 3.795515362704853e-05, + "loss": 0.5235, + "num_input_tokens_seen": 19869520, + "step": 13065 + }, + { + "epoch": 0.41834709685679533, + "grad_norm": 1.7606725692749023, + "learning_rate": 3.794675617468803e-05, + "loss": 0.6542, + "num_input_tokens_seen": 19876768, + "step": 13070 + }, + { + "epoch": 0.41850713782728377, + "grad_norm": 0.7744954824447632, + "learning_rate": 3.793835672577503e-05, + "loss": 0.4667, + "num_input_tokens_seen": 19883952, + "step": 13075 + }, + { + "epoch": 0.4186671787977722, + "grad_norm": 0.357397198677063, + "learning_rate": 3.7929955281604826e-05, + "loss": 0.5363, + "num_input_tokens_seen": 19891472, + "step": 13080 + }, + { + "epoch": 0.41882721976826065, + "grad_norm": 0.6028355360031128, + "learning_rate": 3.7921551843473036e-05, + "loss": 0.5844, + "num_input_tokens_seen": 19898864, + "step": 13085 + }, + { + "epoch": 0.41898726073874915, + "grad_norm": 1.0323868989944458, + "learning_rate": 3.791314641267557e-05, + "loss": 0.485, + "num_input_tokens_seen": 19906272, + "step": 13090 + }, + { + "epoch": 0.4191473017092376, + "grad_norm": 0.7331724762916565, + "learning_rate": 3.790473899050864e-05, + "loss": 0.6155, + "num_input_tokens_seen": 19913248, + "step": 13095 + }, + { + "epoch": 0.419307342679726, + "grad_norm": 0.8745256066322327, + "learning_rate": 3.7896329578268794e-05, + "loss": 0.5034, + "num_input_tokens_seen": 19921216, + "step": 13100 + }, + { + "epoch": 0.41946738365021446, + "grad_norm": 0.9741412997245789, + "learning_rate": 3.7887918177252855e-05, + "loss": 0.7499, + "num_input_tokens_seen": 19928816, + "step": 13105 + }, + { + "epoch": 0.4196274246207029, + "grad_norm": 0.9297001361846924, + "learning_rate": 3.787950478875798e-05, + "loss": 0.4276, + "num_input_tokens_seen": 19936448, + "step": 13110 + }, + { + "epoch": 0.41978746559119134, + "grad_norm": 0.7240337133407593, + "learning_rate": 3.787108941408162e-05, + "loss": 0.6104, + "num_input_tokens_seen": 19944160, + "step": 13115 + }, + { + "epoch": 0.4199475065616798, + "grad_norm": 1.299573302268982, + "learning_rate": 3.786267205452151e-05, + "loss": 0.4415, + "num_input_tokens_seen": 19951856, + "step": 13120 + }, + { + "epoch": 0.4201075475321682, + "grad_norm": 1.2315860986709595, + "learning_rate": 3.785425271137573e-05, + "loss": 0.569, + "num_input_tokens_seen": 19959344, + "step": 13125 + }, + { + "epoch": 0.42026758850265666, + "grad_norm": 1.0096107721328735, + "learning_rate": 3.7845831385942655e-05, + "loss": 0.5496, + "num_input_tokens_seen": 19966832, + "step": 13130 + }, + { + "epoch": 0.4204276294731451, + "grad_norm": 1.389374852180481, + "learning_rate": 3.7837408079520944e-05, + "loss": 0.7926, + "num_input_tokens_seen": 19974176, + "step": 13135 + }, + { + "epoch": 0.42058767044363354, + "grad_norm": 1.17063307762146, + "learning_rate": 3.782898279340957e-05, + "loss": 0.744, + "num_input_tokens_seen": 19981456, + "step": 13140 + }, + { + "epoch": 0.42074771141412204, + "grad_norm": 0.6536832451820374, + "learning_rate": 3.782055552890784e-05, + "loss": 0.6282, + "num_input_tokens_seen": 19988688, + "step": 13145 + }, + { + "epoch": 0.4209077523846105, + "grad_norm": 0.4611707329750061, + "learning_rate": 3.781212628731534e-05, + "loss": 0.4834, + "num_input_tokens_seen": 19995728, + "step": 13150 + }, + { + "epoch": 0.4210677933550989, + "grad_norm": 2.6718521118164062, + "learning_rate": 3.7803695069931946e-05, + "loss": 0.619, + "num_input_tokens_seen": 20003344, + "step": 13155 + }, + { + "epoch": 0.42122783432558736, + "grad_norm": 0.4966900944709778, + "learning_rate": 3.779526187805789e-05, + "loss": 0.5123, + "num_input_tokens_seen": 20010880, + "step": 13160 + }, + { + "epoch": 0.4213878752960758, + "grad_norm": 0.776640772819519, + "learning_rate": 3.778682671299364e-05, + "loss": 0.5627, + "num_input_tokens_seen": 20018560, + "step": 13165 + }, + { + "epoch": 0.42154791626656424, + "grad_norm": 0.8585749268531799, + "learning_rate": 3.777838957604003e-05, + "loss": 0.5066, + "num_input_tokens_seen": 20026464, + "step": 13170 + }, + { + "epoch": 0.4217079572370527, + "grad_norm": 0.46114593744277954, + "learning_rate": 3.776995046849816e-05, + "loss": 0.8936, + "num_input_tokens_seen": 20033904, + "step": 13175 + }, + { + "epoch": 0.4218679982075411, + "grad_norm": 0.6801918148994446, + "learning_rate": 3.776150939166945e-05, + "loss": 0.4602, + "num_input_tokens_seen": 20041584, + "step": 13180 + }, + { + "epoch": 0.42202803917802956, + "grad_norm": 0.4013632535934448, + "learning_rate": 3.775306634685562e-05, + "loss": 0.4113, + "num_input_tokens_seen": 20049328, + "step": 13185 + }, + { + "epoch": 0.422188080148518, + "grad_norm": 0.5271806716918945, + "learning_rate": 3.7744621335358696e-05, + "loss": 0.3871, + "num_input_tokens_seen": 20056736, + "step": 13190 + }, + { + "epoch": 0.4223481211190065, + "grad_norm": 1.106436848640442, + "learning_rate": 3.7736174358481e-05, + "loss": 0.4591, + "num_input_tokens_seen": 20063888, + "step": 13195 + }, + { + "epoch": 0.42250816208949493, + "grad_norm": 0.8747396469116211, + "learning_rate": 3.7727725417525175e-05, + "loss": 0.3868, + "num_input_tokens_seen": 20071264, + "step": 13200 + }, + { + "epoch": 0.42250816208949493, + "eval_loss": 0.540848433971405, + "eval_runtime": 561.4205, + "eval_samples_per_second": 24.734, + "eval_steps_per_second": 12.367, + "num_input_tokens_seen": 20071264, + "step": 13200 + }, + { + "epoch": 0.42266820305998337, + "grad_norm": 0.5074828863143921, + "learning_rate": 3.771927451379414e-05, + "loss": 0.4659, + "num_input_tokens_seen": 20079728, + "step": 13205 + }, + { + "epoch": 0.4228282440304718, + "grad_norm": 0.661638081073761, + "learning_rate": 3.7710821648591135e-05, + "loss": 0.6587, + "num_input_tokens_seen": 20086960, + "step": 13210 + }, + { + "epoch": 0.42298828500096025, + "grad_norm": 0.5903226733207703, + "learning_rate": 3.7702366823219694e-05, + "loss": 0.3874, + "num_input_tokens_seen": 20094752, + "step": 13215 + }, + { + "epoch": 0.4231483259714487, + "grad_norm": 0.5783836841583252, + "learning_rate": 3.769391003898366e-05, + "loss": 0.4465, + "num_input_tokens_seen": 20102400, + "step": 13220 + }, + { + "epoch": 0.42330836694193713, + "grad_norm": 2.279792308807373, + "learning_rate": 3.768545129718718e-05, + "loss": 0.562, + "num_input_tokens_seen": 20110048, + "step": 13225 + }, + { + "epoch": 0.42346840791242557, + "grad_norm": 1.8464689254760742, + "learning_rate": 3.7676990599134686e-05, + "loss": 0.7101, + "num_input_tokens_seen": 20117904, + "step": 13230 + }, + { + "epoch": 0.423628448882914, + "grad_norm": 1.1741039752960205, + "learning_rate": 3.766852794613095e-05, + "loss": 0.5952, + "num_input_tokens_seen": 20125296, + "step": 13235 + }, + { + "epoch": 0.42378848985340245, + "grad_norm": 1.0000983476638794, + "learning_rate": 3.766006333948099e-05, + "loss": 0.4551, + "num_input_tokens_seen": 20133168, + "step": 13240 + }, + { + "epoch": 0.4239485308238909, + "grad_norm": 0.7223187685012817, + "learning_rate": 3.765159678049017e-05, + "loss": 0.5528, + "num_input_tokens_seen": 20140640, + "step": 13245 + }, + { + "epoch": 0.4241085717943794, + "grad_norm": 1.0028419494628906, + "learning_rate": 3.7643128270464134e-05, + "loss": 0.6752, + "num_input_tokens_seen": 20148768, + "step": 13250 + }, + { + "epoch": 0.4242686127648678, + "grad_norm": 0.8659103512763977, + "learning_rate": 3.763465781070884e-05, + "loss": 0.574, + "num_input_tokens_seen": 20156848, + "step": 13255 + }, + { + "epoch": 0.42442865373535626, + "grad_norm": 0.9292460680007935, + "learning_rate": 3.762618540253052e-05, + "loss": 0.6167, + "num_input_tokens_seen": 20165264, + "step": 13260 + }, + { + "epoch": 0.4245886947058447, + "grad_norm": 0.9585785865783691, + "learning_rate": 3.761771104723576e-05, + "loss": 0.788, + "num_input_tokens_seen": 20172816, + "step": 13265 + }, + { + "epoch": 0.42474873567633314, + "grad_norm": 0.6154818534851074, + "learning_rate": 3.7609234746131386e-05, + "loss": 0.5202, + "num_input_tokens_seen": 20180560, + "step": 13270 + }, + { + "epoch": 0.4249087766468216, + "grad_norm": 1.0741509199142456, + "learning_rate": 3.7600756500524556e-05, + "loss": 0.562, + "num_input_tokens_seen": 20188016, + "step": 13275 + }, + { + "epoch": 0.42506881761731, + "grad_norm": 0.6585306525230408, + "learning_rate": 3.759227631172271e-05, + "loss": 0.6899, + "num_input_tokens_seen": 20195344, + "step": 13280 + }, + { + "epoch": 0.42522885858779846, + "grad_norm": 0.4954287111759186, + "learning_rate": 3.758379418103363e-05, + "loss": 0.3312, + "num_input_tokens_seen": 20202896, + "step": 13285 + }, + { + "epoch": 0.4253888995582869, + "grad_norm": 0.9306230545043945, + "learning_rate": 3.757531010976534e-05, + "loss": 0.3362, + "num_input_tokens_seen": 20210832, + "step": 13290 + }, + { + "epoch": 0.42554894052877534, + "grad_norm": 0.663874089717865, + "learning_rate": 3.75668240992262e-05, + "loss": 0.4976, + "num_input_tokens_seen": 20218688, + "step": 13295 + }, + { + "epoch": 0.42570898149926384, + "grad_norm": 0.7807934880256653, + "learning_rate": 3.7558336150724865e-05, + "loss": 0.5095, + "num_input_tokens_seen": 20226848, + "step": 13300 + }, + { + "epoch": 0.4258690224697523, + "grad_norm": 0.5211499333381653, + "learning_rate": 3.754984626557028e-05, + "loss": 0.3611, + "num_input_tokens_seen": 20234400, + "step": 13305 + }, + { + "epoch": 0.4260290634402407, + "grad_norm": 1.4757206439971924, + "learning_rate": 3.754135444507168e-05, + "loss": 0.8185, + "num_input_tokens_seen": 20241824, + "step": 13310 + }, + { + "epoch": 0.42618910441072916, + "grad_norm": 0.8075459003448486, + "learning_rate": 3.753286069053863e-05, + "loss": 0.5839, + "num_input_tokens_seen": 20249040, + "step": 13315 + }, + { + "epoch": 0.4263491453812176, + "grad_norm": 0.6918811798095703, + "learning_rate": 3.7524365003280945e-05, + "loss": 0.6316, + "num_input_tokens_seen": 20256912, + "step": 13320 + }, + { + "epoch": 0.42650918635170604, + "grad_norm": 1.1522384881973267, + "learning_rate": 3.75158673846088e-05, + "loss": 0.4035, + "num_input_tokens_seen": 20264080, + "step": 13325 + }, + { + "epoch": 0.4266692273221945, + "grad_norm": 0.4635078012943268, + "learning_rate": 3.750736783583262e-05, + "loss": 0.3659, + "num_input_tokens_seen": 20271840, + "step": 13330 + }, + { + "epoch": 0.4268292682926829, + "grad_norm": 1.1220953464508057, + "learning_rate": 3.7498866358263144e-05, + "loss": 0.6192, + "num_input_tokens_seen": 20279296, + "step": 13335 + }, + { + "epoch": 0.42698930926317136, + "grad_norm": 1.159253478050232, + "learning_rate": 3.74903629532114e-05, + "loss": 0.4381, + "num_input_tokens_seen": 20286656, + "step": 13340 + }, + { + "epoch": 0.4271493502336598, + "grad_norm": 0.5627667903900146, + "learning_rate": 3.748185762198873e-05, + "loss": 0.4837, + "num_input_tokens_seen": 20293936, + "step": 13345 + }, + { + "epoch": 0.42730939120414824, + "grad_norm": 0.6462261080741882, + "learning_rate": 3.747335036590676e-05, + "loss": 0.5386, + "num_input_tokens_seen": 20301360, + "step": 13350 + }, + { + "epoch": 0.42746943217463673, + "grad_norm": 1.2254078388214111, + "learning_rate": 3.7464841186277405e-05, + "loss": 0.5794, + "num_input_tokens_seen": 20308608, + "step": 13355 + }, + { + "epoch": 0.42762947314512517, + "grad_norm": 0.8216466903686523, + "learning_rate": 3.7456330084412896e-05, + "loss": 0.5005, + "num_input_tokens_seen": 20316304, + "step": 13360 + }, + { + "epoch": 0.4277895141156136, + "grad_norm": 0.8077788949012756, + "learning_rate": 3.744781706162576e-05, + "loss": 0.373, + "num_input_tokens_seen": 20323936, + "step": 13365 + }, + { + "epoch": 0.42794955508610205, + "grad_norm": 1.517069697380066, + "learning_rate": 3.743930211922879e-05, + "loss": 0.518, + "num_input_tokens_seen": 20331408, + "step": 13370 + }, + { + "epoch": 0.4281095960565905, + "grad_norm": 0.6677455902099609, + "learning_rate": 3.743078525853513e-05, + "loss": 0.5185, + "num_input_tokens_seen": 20338752, + "step": 13375 + }, + { + "epoch": 0.42826963702707893, + "grad_norm": 1.1325074434280396, + "learning_rate": 3.7422266480858154e-05, + "loss": 0.4496, + "num_input_tokens_seen": 20346176, + "step": 13380 + }, + { + "epoch": 0.42842967799756737, + "grad_norm": 1.1778560876846313, + "learning_rate": 3.741374578751158e-05, + "loss": 0.6932, + "num_input_tokens_seen": 20353440, + "step": 13385 + }, + { + "epoch": 0.4285897189680558, + "grad_norm": 0.9160352945327759, + "learning_rate": 3.740522317980941e-05, + "loss": 0.5127, + "num_input_tokens_seen": 20361424, + "step": 13390 + }, + { + "epoch": 0.42874975993854425, + "grad_norm": 1.2157405614852905, + "learning_rate": 3.739669865906593e-05, + "loss": 0.5676, + "num_input_tokens_seen": 20369040, + "step": 13395 + }, + { + "epoch": 0.4289098009090327, + "grad_norm": 1.644790768623352, + "learning_rate": 3.738817222659573e-05, + "loss": 0.5129, + "num_input_tokens_seen": 20376832, + "step": 13400 + }, + { + "epoch": 0.4289098009090327, + "eval_loss": 0.5402436852455139, + "eval_runtime": 561.5093, + "eval_samples_per_second": 24.73, + "eval_steps_per_second": 12.365, + "num_input_tokens_seen": 20376832, + "step": 13400 + }, + { + "epoch": 0.42906984187952113, + "grad_norm": 0.9208719730377197, + "learning_rate": 3.73796438837137e-05, + "loss": 0.3838, + "num_input_tokens_seen": 20384656, + "step": 13405 + }, + { + "epoch": 0.4292298828500096, + "grad_norm": 0.995880126953125, + "learning_rate": 3.7371113631735e-05, + "loss": 0.6554, + "num_input_tokens_seen": 20392096, + "step": 13410 + }, + { + "epoch": 0.42938992382049807, + "grad_norm": 0.8039676547050476, + "learning_rate": 3.736258147197512e-05, + "loss": 0.6934, + "num_input_tokens_seen": 20399808, + "step": 13415 + }, + { + "epoch": 0.4295499647909865, + "grad_norm": 0.9792742729187012, + "learning_rate": 3.735404740574981e-05, + "loss": 0.4562, + "num_input_tokens_seen": 20407712, + "step": 13420 + }, + { + "epoch": 0.42971000576147494, + "grad_norm": 0.7000641226768494, + "learning_rate": 3.7345511434375145e-05, + "loss": 0.4497, + "num_input_tokens_seen": 20415168, + "step": 13425 + }, + { + "epoch": 0.4298700467319634, + "grad_norm": 1.8308149576187134, + "learning_rate": 3.733697355916748e-05, + "loss": 0.8365, + "num_input_tokens_seen": 20423296, + "step": 13430 + }, + { + "epoch": 0.4300300877024518, + "grad_norm": 0.8076421618461609, + "learning_rate": 3.732843378144345e-05, + "loss": 0.4866, + "num_input_tokens_seen": 20431168, + "step": 13435 + }, + { + "epoch": 0.43019012867294026, + "grad_norm": 0.5675035715103149, + "learning_rate": 3.7319892102519995e-05, + "loss": 0.3977, + "num_input_tokens_seen": 20439216, + "step": 13440 + }, + { + "epoch": 0.4303501696434287, + "grad_norm": 0.7078480124473572, + "learning_rate": 3.731134852371436e-05, + "loss": 0.4107, + "num_input_tokens_seen": 20446768, + "step": 13445 + }, + { + "epoch": 0.43051021061391714, + "grad_norm": 0.6776353716850281, + "learning_rate": 3.730280304634408e-05, + "loss": 0.6001, + "num_input_tokens_seen": 20454368, + "step": 13450 + }, + { + "epoch": 0.4306702515844056, + "grad_norm": 0.7461414337158203, + "learning_rate": 3.729425567172696e-05, + "loss": 0.5521, + "num_input_tokens_seen": 20461904, + "step": 13455 + }, + { + "epoch": 0.4308302925548941, + "grad_norm": 0.7747664451599121, + "learning_rate": 3.728570640118111e-05, + "loss": 0.7722, + "num_input_tokens_seen": 20469392, + "step": 13460 + }, + { + "epoch": 0.4309903335253825, + "grad_norm": 0.6205732226371765, + "learning_rate": 3.727715523602494e-05, + "loss": 0.3491, + "num_input_tokens_seen": 20476768, + "step": 13465 + }, + { + "epoch": 0.43115037449587096, + "grad_norm": 0.875254213809967, + "learning_rate": 3.726860217757715e-05, + "loss": 0.4945, + "num_input_tokens_seen": 20484368, + "step": 13470 + }, + { + "epoch": 0.4313104154663594, + "grad_norm": 1.190262794494629, + "learning_rate": 3.726004722715673e-05, + "loss": 0.82, + "num_input_tokens_seen": 20492528, + "step": 13475 + }, + { + "epoch": 0.43147045643684784, + "grad_norm": 0.8502631187438965, + "learning_rate": 3.725149038608296e-05, + "loss": 0.4027, + "num_input_tokens_seen": 20500256, + "step": 13480 + }, + { + "epoch": 0.4316304974073363, + "grad_norm": 0.8247888684272766, + "learning_rate": 3.7242931655675404e-05, + "loss": 0.4592, + "num_input_tokens_seen": 20508480, + "step": 13485 + }, + { + "epoch": 0.4317905383778247, + "grad_norm": 1.4360607862472534, + "learning_rate": 3.7234371037253937e-05, + "loss": 0.4602, + "num_input_tokens_seen": 20516352, + "step": 13490 + }, + { + "epoch": 0.43195057934831316, + "grad_norm": 1.1788021326065063, + "learning_rate": 3.7225808532138705e-05, + "loss": 0.7512, + "num_input_tokens_seen": 20524352, + "step": 13495 + }, + { + "epoch": 0.4321106203188016, + "grad_norm": 0.7085981965065002, + "learning_rate": 3.721724414165016e-05, + "loss": 0.3426, + "num_input_tokens_seen": 20531712, + "step": 13500 + }, + { + "epoch": 0.43227066128929004, + "grad_norm": 0.9163821935653687, + "learning_rate": 3.720867786710904e-05, + "loss": 0.4927, + "num_input_tokens_seen": 20539200, + "step": 13505 + }, + { + "epoch": 0.4324307022597785, + "grad_norm": 0.5501469373703003, + "learning_rate": 3.7200109709836366e-05, + "loss": 0.3687, + "num_input_tokens_seen": 20546528, + "step": 13510 + }, + { + "epoch": 0.43259074323026697, + "grad_norm": 1.2447539567947388, + "learning_rate": 3.7191539671153465e-05, + "loss": 0.4777, + "num_input_tokens_seen": 20554128, + "step": 13515 + }, + { + "epoch": 0.4327507842007554, + "grad_norm": 1.2988418340682983, + "learning_rate": 3.718296775238193e-05, + "loss": 0.3704, + "num_input_tokens_seen": 20561936, + "step": 13520 + }, + { + "epoch": 0.43291082517124385, + "grad_norm": 1.348722219467163, + "learning_rate": 3.7174393954843675e-05, + "loss": 0.5926, + "num_input_tokens_seen": 20569648, + "step": 13525 + }, + { + "epoch": 0.4330708661417323, + "grad_norm": 0.6186180710792542, + "learning_rate": 3.716581827986087e-05, + "loss": 0.558, + "num_input_tokens_seen": 20577456, + "step": 13530 + }, + { + "epoch": 0.43323090711222073, + "grad_norm": 0.644038200378418, + "learning_rate": 3.7157240728756004e-05, + "loss": 0.4673, + "num_input_tokens_seen": 20584816, + "step": 13535 + }, + { + "epoch": 0.43339094808270917, + "grad_norm": 1.2676453590393066, + "learning_rate": 3.714866130285184e-05, + "loss": 0.5132, + "num_input_tokens_seen": 20592768, + "step": 13540 + }, + { + "epoch": 0.4335509890531976, + "grad_norm": 0.523043155670166, + "learning_rate": 3.714008000347143e-05, + "loss": 0.496, + "num_input_tokens_seen": 20600800, + "step": 13545 + }, + { + "epoch": 0.43371103002368605, + "grad_norm": 1.0504682064056396, + "learning_rate": 3.7131496831938126e-05, + "loss": 0.5812, + "num_input_tokens_seen": 20608576, + "step": 13550 + }, + { + "epoch": 0.4338710709941745, + "grad_norm": 0.4995371699333191, + "learning_rate": 3.7122911789575565e-05, + "loss": 0.5555, + "num_input_tokens_seen": 20616464, + "step": 13555 + }, + { + "epoch": 0.43403111196466293, + "grad_norm": 0.7261728644371033, + "learning_rate": 3.711432487770765e-05, + "loss": 0.5393, + "num_input_tokens_seen": 20623984, + "step": 13560 + }, + { + "epoch": 0.4341911529351514, + "grad_norm": 1.5514426231384277, + "learning_rate": 3.710573609765861e-05, + "loss": 0.5626, + "num_input_tokens_seen": 20631056, + "step": 13565 + }, + { + "epoch": 0.43435119390563987, + "grad_norm": 0.7297419309616089, + "learning_rate": 3.709714545075292e-05, + "loss": 0.517, + "num_input_tokens_seen": 20638624, + "step": 13570 + }, + { + "epoch": 0.4345112348761283, + "grad_norm": 0.46535053849220276, + "learning_rate": 3.708855293831538e-05, + "loss": 0.451, + "num_input_tokens_seen": 20646272, + "step": 13575 + }, + { + "epoch": 0.43467127584661674, + "grad_norm": 1.5041556358337402, + "learning_rate": 3.707995856167107e-05, + "loss": 0.651, + "num_input_tokens_seen": 20653888, + "step": 13580 + }, + { + "epoch": 0.4348313168171052, + "grad_norm": 0.53989177942276, + "learning_rate": 3.707136232214534e-05, + "loss": 0.673, + "num_input_tokens_seen": 20661408, + "step": 13585 + }, + { + "epoch": 0.4349913577875936, + "grad_norm": 0.1920476257801056, + "learning_rate": 3.7062764221063844e-05, + "loss": 0.4259, + "num_input_tokens_seen": 20668896, + "step": 13590 + }, + { + "epoch": 0.43515139875808206, + "grad_norm": 0.6542088985443115, + "learning_rate": 3.705416425975252e-05, + "loss": 0.3899, + "num_input_tokens_seen": 20676208, + "step": 13595 + }, + { + "epoch": 0.4353114397285705, + "grad_norm": 0.4551272690296173, + "learning_rate": 3.704556243953758e-05, + "loss": 0.5026, + "num_input_tokens_seen": 20683904, + "step": 13600 + }, + { + "epoch": 0.4353114397285705, + "eval_loss": 0.5394640564918518, + "eval_runtime": 561.4004, + "eval_samples_per_second": 24.735, + "eval_steps_per_second": 12.367, + "num_input_tokens_seen": 20683904, + "step": 13600 + }, + { + "epoch": 0.43547148069905894, + "grad_norm": 0.7028627991676331, + "learning_rate": 3.7036958761745535e-05, + "loss": 0.5119, + "num_input_tokens_seen": 20691344, + "step": 13605 + }, + { + "epoch": 0.4356315216695474, + "grad_norm": 1.0654277801513672, + "learning_rate": 3.702835322770318e-05, + "loss": 0.3632, + "num_input_tokens_seen": 20698432, + "step": 13610 + }, + { + "epoch": 0.4357915626400358, + "grad_norm": 1.3791730403900146, + "learning_rate": 3.701974583873761e-05, + "loss": 0.5772, + "num_input_tokens_seen": 20706096, + "step": 13615 + }, + { + "epoch": 0.4359516036105243, + "grad_norm": 1.7290290594100952, + "learning_rate": 3.701113659617618e-05, + "loss": 0.4607, + "num_input_tokens_seen": 20713584, + "step": 13620 + }, + { + "epoch": 0.43611164458101276, + "grad_norm": 0.7025564908981323, + "learning_rate": 3.7002525501346535e-05, + "loss": 0.6778, + "num_input_tokens_seen": 20720976, + "step": 13625 + }, + { + "epoch": 0.4362716855515012, + "grad_norm": 0.7914828062057495, + "learning_rate": 3.699391255557664e-05, + "loss": 0.5087, + "num_input_tokens_seen": 20728192, + "step": 13630 + }, + { + "epoch": 0.43643172652198964, + "grad_norm": 0.7143948674201965, + "learning_rate": 3.69852977601947e-05, + "loss": 0.4778, + "num_input_tokens_seen": 20736272, + "step": 13635 + }, + { + "epoch": 0.4365917674924781, + "grad_norm": 0.746439516544342, + "learning_rate": 3.697668111652922e-05, + "loss": 0.5344, + "num_input_tokens_seen": 20743904, + "step": 13640 + }, + { + "epoch": 0.4367518084629665, + "grad_norm": 0.5899429321289062, + "learning_rate": 3.6968062625909005e-05, + "loss": 0.5663, + "num_input_tokens_seen": 20752112, + "step": 13645 + }, + { + "epoch": 0.43691184943345496, + "grad_norm": 1.0353317260742188, + "learning_rate": 3.6959442289663135e-05, + "loss": 0.6767, + "num_input_tokens_seen": 20760096, + "step": 13650 + }, + { + "epoch": 0.4370718904039434, + "grad_norm": 0.4649255871772766, + "learning_rate": 3.695082010912098e-05, + "loss": 0.4716, + "num_input_tokens_seen": 20767616, + "step": 13655 + }, + { + "epoch": 0.43723193137443184, + "grad_norm": 1.045267939567566, + "learning_rate": 3.694219608561217e-05, + "loss": 0.486, + "num_input_tokens_seen": 20775072, + "step": 13660 + }, + { + "epoch": 0.4373919723449203, + "grad_norm": 0.7275721430778503, + "learning_rate": 3.693357022046665e-05, + "loss": 0.4304, + "num_input_tokens_seen": 20782752, + "step": 13665 + }, + { + "epoch": 0.43755201331540877, + "grad_norm": 0.5532507300376892, + "learning_rate": 3.6924942515014644e-05, + "loss": 0.489, + "num_input_tokens_seen": 20790672, + "step": 13670 + }, + { + "epoch": 0.4377120542858972, + "grad_norm": 0.5078639388084412, + "learning_rate": 3.691631297058664e-05, + "loss": 0.5274, + "num_input_tokens_seen": 20798112, + "step": 13675 + }, + { + "epoch": 0.43787209525638565, + "grad_norm": 0.39008164405822754, + "learning_rate": 3.6907681588513424e-05, + "loss": 0.4301, + "num_input_tokens_seen": 20805760, + "step": 13680 + }, + { + "epoch": 0.4380321362268741, + "grad_norm": 0.9505752325057983, + "learning_rate": 3.689904837012606e-05, + "loss": 0.4529, + "num_input_tokens_seen": 20813344, + "step": 13685 + }, + { + "epoch": 0.43819217719736253, + "grad_norm": 1.2401570081710815, + "learning_rate": 3.689041331675591e-05, + "loss": 0.6604, + "num_input_tokens_seen": 20821136, + "step": 13690 + }, + { + "epoch": 0.43835221816785097, + "grad_norm": 0.2888434827327728, + "learning_rate": 3.688177642973461e-05, + "loss": 0.4132, + "num_input_tokens_seen": 20828672, + "step": 13695 + }, + { + "epoch": 0.4385122591383394, + "grad_norm": 0.8883697390556335, + "learning_rate": 3.687313771039406e-05, + "loss": 0.6122, + "num_input_tokens_seen": 20836112, + "step": 13700 + }, + { + "epoch": 0.43867230010882785, + "grad_norm": 1.0498020648956299, + "learning_rate": 3.686449716006647e-05, + "loss": 0.5524, + "num_input_tokens_seen": 20843712, + "step": 13705 + }, + { + "epoch": 0.4388323410793163, + "grad_norm": 0.9153304696083069, + "learning_rate": 3.685585478008432e-05, + "loss": 0.4436, + "num_input_tokens_seen": 20851792, + "step": 13710 + }, + { + "epoch": 0.43899238204980473, + "grad_norm": 1.1285327672958374, + "learning_rate": 3.6847210571780364e-05, + "loss": 0.6104, + "num_input_tokens_seen": 20859424, + "step": 13715 + }, + { + "epoch": 0.43915242302029317, + "grad_norm": 0.702360212802887, + "learning_rate": 3.683856453648767e-05, + "loss": 0.4796, + "num_input_tokens_seen": 20867104, + "step": 13720 + }, + { + "epoch": 0.43931246399078167, + "grad_norm": 0.46130818128585815, + "learning_rate": 3.682991667553954e-05, + "loss": 0.5499, + "num_input_tokens_seen": 20874624, + "step": 13725 + }, + { + "epoch": 0.4394725049612701, + "grad_norm": 0.5666249394416809, + "learning_rate": 3.6821266990269606e-05, + "loss": 0.5078, + "num_input_tokens_seen": 20882304, + "step": 13730 + }, + { + "epoch": 0.43963254593175854, + "grad_norm": 1.0890138149261475, + "learning_rate": 3.681261548201174e-05, + "loss": 0.5328, + "num_input_tokens_seen": 20889712, + "step": 13735 + }, + { + "epoch": 0.439792586902247, + "grad_norm": 0.6005021929740906, + "learning_rate": 3.6803962152100125e-05, + "loss": 0.5221, + "num_input_tokens_seen": 20897536, + "step": 13740 + }, + { + "epoch": 0.4399526278727354, + "grad_norm": 0.6486762762069702, + "learning_rate": 3.67953070018692e-05, + "loss": 0.4878, + "num_input_tokens_seen": 20905008, + "step": 13745 + }, + { + "epoch": 0.44011266884322386, + "grad_norm": 0.6683205962181091, + "learning_rate": 3.678665003265371e-05, + "loss": 0.5894, + "num_input_tokens_seen": 20912032, + "step": 13750 + }, + { + "epoch": 0.4402727098137123, + "grad_norm": 1.1050945520401, + "learning_rate": 3.677799124578867e-05, + "loss": 0.5087, + "num_input_tokens_seen": 20919248, + "step": 13755 + }, + { + "epoch": 0.44043275078420074, + "grad_norm": 1.203328013420105, + "learning_rate": 3.676933064260937e-05, + "loss": 0.4826, + "num_input_tokens_seen": 20927072, + "step": 13760 + }, + { + "epoch": 0.4405927917546892, + "grad_norm": 0.8117390275001526, + "learning_rate": 3.6760668224451365e-05, + "loss": 0.5119, + "num_input_tokens_seen": 20934720, + "step": 13765 + }, + { + "epoch": 0.4407528327251776, + "grad_norm": 0.94798344373703, + "learning_rate": 3.675200399265054e-05, + "loss": 0.4508, + "num_input_tokens_seen": 20942544, + "step": 13770 + }, + { + "epoch": 0.44091287369566606, + "grad_norm": 0.5794587135314941, + "learning_rate": 3.6743337948543014e-05, + "loss": 0.6099, + "num_input_tokens_seen": 20950080, + "step": 13775 + }, + { + "epoch": 0.44107291466615456, + "grad_norm": 0.7672349810600281, + "learning_rate": 3.6734670093465204e-05, + "loss": 0.4956, + "num_input_tokens_seen": 20957536, + "step": 13780 + }, + { + "epoch": 0.441232955636643, + "grad_norm": 0.525059700012207, + "learning_rate": 3.672600042875379e-05, + "loss": 0.421, + "num_input_tokens_seen": 20965376, + "step": 13785 + }, + { + "epoch": 0.44139299660713144, + "grad_norm": 0.9631186723709106, + "learning_rate": 3.671732895574575e-05, + "loss": 0.5495, + "num_input_tokens_seen": 20973104, + "step": 13790 + }, + { + "epoch": 0.4415530375776199, + "grad_norm": 1.060053825378418, + "learning_rate": 3.670865567577834e-05, + "loss": 0.7034, + "num_input_tokens_seen": 20980432, + "step": 13795 + }, + { + "epoch": 0.4417130785481083, + "grad_norm": 0.75816410779953, + "learning_rate": 3.669998059018909e-05, + "loss": 0.6542, + "num_input_tokens_seen": 20987920, + "step": 13800 + }, + { + "epoch": 0.4417130785481083, + "eval_loss": 0.5380936861038208, + "eval_runtime": 561.5177, + "eval_samples_per_second": 24.729, + "eval_steps_per_second": 12.365, + "num_input_tokens_seen": 20987920, + "step": 13800 + }, + { + "epoch": 0.44187311951859676, + "grad_norm": 1.0361214876174927, + "learning_rate": 3.6691303700315796e-05, + "loss": 0.5283, + "num_input_tokens_seen": 20996240, + "step": 13805 + }, + { + "epoch": 0.4420331604890852, + "grad_norm": 1.203102707862854, + "learning_rate": 3.668262500749655e-05, + "loss": 0.4419, + "num_input_tokens_seen": 21003856, + "step": 13810 + }, + { + "epoch": 0.44219320145957364, + "grad_norm": 1.0426716804504395, + "learning_rate": 3.667394451306971e-05, + "loss": 0.5728, + "num_input_tokens_seen": 21011552, + "step": 13815 + }, + { + "epoch": 0.4423532424300621, + "grad_norm": 0.7651621699333191, + "learning_rate": 3.666526221837393e-05, + "loss": 0.3975, + "num_input_tokens_seen": 21019248, + "step": 13820 + }, + { + "epoch": 0.4425132834005505, + "grad_norm": 1.6430522203445435, + "learning_rate": 3.665657812474812e-05, + "loss": 0.712, + "num_input_tokens_seen": 21026656, + "step": 13825 + }, + { + "epoch": 0.442673324371039, + "grad_norm": 1.1726151704788208, + "learning_rate": 3.664789223353147e-05, + "loss": 0.4955, + "num_input_tokens_seen": 21034320, + "step": 13830 + }, + { + "epoch": 0.44283336534152745, + "grad_norm": 3.918382167816162, + "learning_rate": 3.663920454606347e-05, + "loss": 0.5328, + "num_input_tokens_seen": 21041952, + "step": 13835 + }, + { + "epoch": 0.4429934063120159, + "grad_norm": 0.48549190163612366, + "learning_rate": 3.6630515063683856e-05, + "loss": 0.5106, + "num_input_tokens_seen": 21049824, + "step": 13840 + }, + { + "epoch": 0.44315344728250433, + "grad_norm": 0.48189765214920044, + "learning_rate": 3.662182378773267e-05, + "loss": 0.359, + "num_input_tokens_seen": 21057392, + "step": 13845 + }, + { + "epoch": 0.44331348825299277, + "grad_norm": 0.7063305974006653, + "learning_rate": 3.66131307195502e-05, + "loss": 0.4398, + "num_input_tokens_seen": 21065152, + "step": 13850 + }, + { + "epoch": 0.4434735292234812, + "grad_norm": 0.9527907371520996, + "learning_rate": 3.6604435860477034e-05, + "loss": 0.5624, + "num_input_tokens_seen": 21072816, + "step": 13855 + }, + { + "epoch": 0.44363357019396965, + "grad_norm": 0.6583807468414307, + "learning_rate": 3.6595739211854025e-05, + "loss": 0.5126, + "num_input_tokens_seen": 21080448, + "step": 13860 + }, + { + "epoch": 0.4437936111644581, + "grad_norm": 0.8479244709014893, + "learning_rate": 3.658704077502231e-05, + "loss": 0.5833, + "num_input_tokens_seen": 21087952, + "step": 13865 + }, + { + "epoch": 0.44395365213494653, + "grad_norm": 0.522047758102417, + "learning_rate": 3.65783405513233e-05, + "loss": 0.5327, + "num_input_tokens_seen": 21095200, + "step": 13870 + }, + { + "epoch": 0.44411369310543497, + "grad_norm": 0.5019508004188538, + "learning_rate": 3.656963854209867e-05, + "loss": 0.4488, + "num_input_tokens_seen": 21102480, + "step": 13875 + }, + { + "epoch": 0.4442737340759234, + "grad_norm": 1.0481548309326172, + "learning_rate": 3.656093474869038e-05, + "loss": 0.5745, + "num_input_tokens_seen": 21110128, + "step": 13880 + }, + { + "epoch": 0.4444337750464119, + "grad_norm": 6.044252395629883, + "learning_rate": 3.655222917244068e-05, + "loss": 0.4494, + "num_input_tokens_seen": 21117856, + "step": 13885 + }, + { + "epoch": 0.44459381601690035, + "grad_norm": 0.7491482496261597, + "learning_rate": 3.6543521814692054e-05, + "loss": 0.5657, + "num_input_tokens_seen": 21125424, + "step": 13890 + }, + { + "epoch": 0.4447538569873888, + "grad_norm": 0.48635220527648926, + "learning_rate": 3.653481267678731e-05, + "loss": 0.4779, + "num_input_tokens_seen": 21132928, + "step": 13895 + }, + { + "epoch": 0.4449138979578772, + "grad_norm": 1.511081337928772, + "learning_rate": 3.652610176006949e-05, + "loss": 0.4793, + "num_input_tokens_seen": 21140480, + "step": 13900 + }, + { + "epoch": 0.44507393892836566, + "grad_norm": 0.5017912983894348, + "learning_rate": 3.6517389065881925e-05, + "loss": 0.4795, + "num_input_tokens_seen": 21147824, + "step": 13905 + }, + { + "epoch": 0.4452339798988541, + "grad_norm": 0.9356496334075928, + "learning_rate": 3.650867459556824e-05, + "loss": 0.5903, + "num_input_tokens_seen": 21155168, + "step": 13910 + }, + { + "epoch": 0.44539402086934254, + "grad_norm": 0.408646821975708, + "learning_rate": 3.64999583504723e-05, + "loss": 0.5535, + "num_input_tokens_seen": 21162896, + "step": 13915 + }, + { + "epoch": 0.445554061839831, + "grad_norm": 0.5209044814109802, + "learning_rate": 3.649124033193827e-05, + "loss": 0.546, + "num_input_tokens_seen": 21170688, + "step": 13920 + }, + { + "epoch": 0.4457141028103194, + "grad_norm": 0.5732250809669495, + "learning_rate": 3.648252054131057e-05, + "loss": 0.4951, + "num_input_tokens_seen": 21178432, + "step": 13925 + }, + { + "epoch": 0.44587414378080786, + "grad_norm": 0.6067366600036621, + "learning_rate": 3.647379897993391e-05, + "loss": 0.6099, + "num_input_tokens_seen": 21186400, + "step": 13930 + }, + { + "epoch": 0.44603418475129636, + "grad_norm": 2.1648623943328857, + "learning_rate": 3.646507564915325e-05, + "loss": 0.3347, + "num_input_tokens_seen": 21193904, + "step": 13935 + }, + { + "epoch": 0.4461942257217848, + "grad_norm": 0.4855308532714844, + "learning_rate": 3.645635055031385e-05, + "loss": 0.6368, + "num_input_tokens_seen": 21201872, + "step": 13940 + }, + { + "epoch": 0.44635426669227324, + "grad_norm": 1.221873164176941, + "learning_rate": 3.6447623684761224e-05, + "loss": 0.5318, + "num_input_tokens_seen": 21209088, + "step": 13945 + }, + { + "epoch": 0.4465143076627617, + "grad_norm": 0.7185245752334595, + "learning_rate": 3.643889505384117e-05, + "loss": 0.4367, + "num_input_tokens_seen": 21216512, + "step": 13950 + }, + { + "epoch": 0.4466743486332501, + "grad_norm": 0.8013274073600769, + "learning_rate": 3.6430164658899744e-05, + "loss": 0.5299, + "num_input_tokens_seen": 21224624, + "step": 13955 + }, + { + "epoch": 0.44683438960373856, + "grad_norm": 0.9584154486656189, + "learning_rate": 3.642143250128329e-05, + "loss": 0.4954, + "num_input_tokens_seen": 21232480, + "step": 13960 + }, + { + "epoch": 0.446994430574227, + "grad_norm": 0.7388973832130432, + "learning_rate": 3.641269858233841e-05, + "loss": 0.5881, + "num_input_tokens_seen": 21239840, + "step": 13965 + }, + { + "epoch": 0.44715447154471544, + "grad_norm": 0.9411572813987732, + "learning_rate": 3.640396290341199e-05, + "loss": 0.4, + "num_input_tokens_seen": 21247648, + "step": 13970 + }, + { + "epoch": 0.4473145125152039, + "grad_norm": 0.9295431971549988, + "learning_rate": 3.639522546585118e-05, + "loss": 0.4881, + "num_input_tokens_seen": 21255424, + "step": 13975 + }, + { + "epoch": 0.4474745534856923, + "grad_norm": 0.5316512584686279, + "learning_rate": 3.6386486271003404e-05, + "loss": 0.422, + "num_input_tokens_seen": 21263344, + "step": 13980 + }, + { + "epoch": 0.44763459445618076, + "grad_norm": 1.0765831470489502, + "learning_rate": 3.6377745320216346e-05, + "loss": 0.5649, + "num_input_tokens_seen": 21270848, + "step": 13985 + }, + { + "epoch": 0.44779463542666925, + "grad_norm": 0.8008197546005249, + "learning_rate": 3.636900261483798e-05, + "loss": 0.4751, + "num_input_tokens_seen": 21278352, + "step": 13990 + }, + { + "epoch": 0.4479546763971577, + "grad_norm": 0.793803870677948, + "learning_rate": 3.636025815621654e-05, + "loss": 0.6158, + "num_input_tokens_seen": 21285872, + "step": 13995 + }, + { + "epoch": 0.44811471736764613, + "grad_norm": 1.411299467086792, + "learning_rate": 3.635151194570054e-05, + "loss": 0.4867, + "num_input_tokens_seen": 21293728, + "step": 14000 + }, + { + "epoch": 0.44811471736764613, + "eval_loss": 0.5364681482315063, + "eval_runtime": 561.4152, + "eval_samples_per_second": 24.734, + "eval_steps_per_second": 12.367, + "num_input_tokens_seen": 21293728, + "step": 14000 + }, + { + "epoch": 0.44827475833813457, + "grad_norm": 0.6193768382072449, + "learning_rate": 3.634276398463873e-05, + "loss": 0.4695, + "num_input_tokens_seen": 21301264, + "step": 14005 + }, + { + "epoch": 0.448434799308623, + "grad_norm": 0.7949881553649902, + "learning_rate": 3.633401427438018e-05, + "loss": 0.4267, + "num_input_tokens_seen": 21309056, + "step": 14010 + }, + { + "epoch": 0.44859484027911145, + "grad_norm": 0.4055198132991791, + "learning_rate": 3.63252628162742e-05, + "loss": 0.3996, + "num_input_tokens_seen": 21316688, + "step": 14015 + }, + { + "epoch": 0.4487548812495999, + "grad_norm": 0.9707213640213013, + "learning_rate": 3.6316509611670364e-05, + "loss": 0.6836, + "num_input_tokens_seen": 21324288, + "step": 14020 + }, + { + "epoch": 0.44891492222008833, + "grad_norm": 0.7678858041763306, + "learning_rate": 3.630775466191854e-05, + "loss": 0.4165, + "num_input_tokens_seen": 21331680, + "step": 14025 + }, + { + "epoch": 0.44907496319057677, + "grad_norm": 0.6214866638183594, + "learning_rate": 3.629899796836884e-05, + "loss": 0.5785, + "num_input_tokens_seen": 21339168, + "step": 14030 + }, + { + "epoch": 0.4492350041610652, + "grad_norm": 0.642310380935669, + "learning_rate": 3.6290239532371666e-05, + "loss": 0.4465, + "num_input_tokens_seen": 21346736, + "step": 14035 + }, + { + "epoch": 0.4493950451315537, + "grad_norm": 0.736098051071167, + "learning_rate": 3.628147935527767e-05, + "loss": 0.6166, + "num_input_tokens_seen": 21354816, + "step": 14040 + }, + { + "epoch": 0.44955508610204215, + "grad_norm": 0.8371527194976807, + "learning_rate": 3.627271743843779e-05, + "loss": 0.4161, + "num_input_tokens_seen": 21362272, + "step": 14045 + }, + { + "epoch": 0.4497151270725306, + "grad_norm": 0.49000313878059387, + "learning_rate": 3.626395378320321e-05, + "loss": 0.568, + "num_input_tokens_seen": 21369792, + "step": 14050 + }, + { + "epoch": 0.449875168043019, + "grad_norm": 1.5800663232803345, + "learning_rate": 3.625518839092541e-05, + "loss": 0.5999, + "num_input_tokens_seen": 21377472, + "step": 14055 + }, + { + "epoch": 0.45003520901350746, + "grad_norm": 0.6305127143859863, + "learning_rate": 3.624642126295612e-05, + "loss": 0.4556, + "num_input_tokens_seen": 21384912, + "step": 14060 + }, + { + "epoch": 0.4501952499839959, + "grad_norm": 0.6693720817565918, + "learning_rate": 3.6237652400647345e-05, + "loss": 0.5066, + "num_input_tokens_seen": 21392592, + "step": 14065 + }, + { + "epoch": 0.45035529095448434, + "grad_norm": 1.1033930778503418, + "learning_rate": 3.622888180535134e-05, + "loss": 0.4786, + "num_input_tokens_seen": 21400688, + "step": 14070 + }, + { + "epoch": 0.4505153319249728, + "grad_norm": 0.3659643232822418, + "learning_rate": 3.6220109478420655e-05, + "loss": 0.3828, + "num_input_tokens_seen": 21408144, + "step": 14075 + }, + { + "epoch": 0.4506753728954612, + "grad_norm": 0.6624975204467773, + "learning_rate": 3.6211335421208084e-05, + "loss": 0.5485, + "num_input_tokens_seen": 21415568, + "step": 14080 + }, + { + "epoch": 0.45083541386594966, + "grad_norm": 0.8990398049354553, + "learning_rate": 3.62025596350667e-05, + "loss": 0.4436, + "num_input_tokens_seen": 21422688, + "step": 14085 + }, + { + "epoch": 0.4509954548364381, + "grad_norm": 0.6777291297912598, + "learning_rate": 3.619378212134984e-05, + "loss": 0.5253, + "num_input_tokens_seen": 21430592, + "step": 14090 + }, + { + "epoch": 0.4511554958069266, + "grad_norm": 0.7381248474121094, + "learning_rate": 3.618500288141111e-05, + "loss": 0.3036, + "num_input_tokens_seen": 21438000, + "step": 14095 + }, + { + "epoch": 0.45131553677741504, + "grad_norm": 1.5790371894836426, + "learning_rate": 3.617622191660438e-05, + "loss": 0.68, + "num_input_tokens_seen": 21446112, + "step": 14100 + }, + { + "epoch": 0.4514755777479035, + "grad_norm": 0.6994732022285461, + "learning_rate": 3.616743922828377e-05, + "loss": 0.4098, + "num_input_tokens_seen": 21454192, + "step": 14105 + }, + { + "epoch": 0.4516356187183919, + "grad_norm": 0.8714670538902283, + "learning_rate": 3.615865481780371e-05, + "loss": 0.4519, + "num_input_tokens_seen": 21461840, + "step": 14110 + }, + { + "epoch": 0.45179565968888036, + "grad_norm": 0.5248286724090576, + "learning_rate": 3.614986868651883e-05, + "loss": 0.4566, + "num_input_tokens_seen": 21469392, + "step": 14115 + }, + { + "epoch": 0.4519557006593688, + "grad_norm": 0.7147777676582336, + "learning_rate": 3.614108083578409e-05, + "loss": 0.4934, + "num_input_tokens_seen": 21476688, + "step": 14120 + }, + { + "epoch": 0.45211574162985724, + "grad_norm": 0.4478195905685425, + "learning_rate": 3.613229126695467e-05, + "loss": 0.5388, + "num_input_tokens_seen": 21484752, + "step": 14125 + }, + { + "epoch": 0.4522757826003457, + "grad_norm": 0.9245674014091492, + "learning_rate": 3.612349998138605e-05, + "loss": 0.5777, + "num_input_tokens_seen": 21492256, + "step": 14130 + }, + { + "epoch": 0.4524358235708341, + "grad_norm": 1.8398997783660889, + "learning_rate": 3.6114706980433946e-05, + "loss": 0.5707, + "num_input_tokens_seen": 21499792, + "step": 14135 + }, + { + "epoch": 0.45259586454132256, + "grad_norm": 0.7227407097816467, + "learning_rate": 3.610591226545435e-05, + "loss": 0.5815, + "num_input_tokens_seen": 21507840, + "step": 14140 + }, + { + "epoch": 0.452755905511811, + "grad_norm": 0.9763618111610413, + "learning_rate": 3.6097115837803505e-05, + "loss": 0.5704, + "num_input_tokens_seen": 21515488, + "step": 14145 + }, + { + "epoch": 0.4529159464822995, + "grad_norm": 1.0236083269119263, + "learning_rate": 3.608831769883795e-05, + "loss": 0.682, + "num_input_tokens_seen": 21523664, + "step": 14150 + }, + { + "epoch": 0.45307598745278793, + "grad_norm": 0.7835014462471008, + "learning_rate": 3.607951784991446e-05, + "loss": 0.6115, + "num_input_tokens_seen": 21531424, + "step": 14155 + }, + { + "epoch": 0.45323602842327637, + "grad_norm": 0.6463878154754639, + "learning_rate": 3.6070716292390085e-05, + "loss": 0.3176, + "num_input_tokens_seen": 21538800, + "step": 14160 + }, + { + "epoch": 0.4533960693937648, + "grad_norm": 0.8257312178611755, + "learning_rate": 3.606191302762213e-05, + "loss": 0.6347, + "num_input_tokens_seen": 21545920, + "step": 14165 + }, + { + "epoch": 0.45355611036425325, + "grad_norm": 0.4362197816371918, + "learning_rate": 3.605310805696818e-05, + "loss": 0.3539, + "num_input_tokens_seen": 21553248, + "step": 14170 + }, + { + "epoch": 0.4537161513347417, + "grad_norm": 0.5497269630432129, + "learning_rate": 3.6044301381786067e-05, + "loss": 0.3526, + "num_input_tokens_seen": 21560880, + "step": 14175 + }, + { + "epoch": 0.45387619230523013, + "grad_norm": 0.9035243391990662, + "learning_rate": 3.6035493003433883e-05, + "loss": 0.6146, + "num_input_tokens_seen": 21568576, + "step": 14180 + }, + { + "epoch": 0.45403623327571857, + "grad_norm": 0.7833364605903625, + "learning_rate": 3.6026682923269994e-05, + "loss": 0.5764, + "num_input_tokens_seen": 21576112, + "step": 14185 + }, + { + "epoch": 0.454196274246207, + "grad_norm": 1.1456466913223267, + "learning_rate": 3.6017871142653034e-05, + "loss": 0.5965, + "num_input_tokens_seen": 21584160, + "step": 14190 + }, + { + "epoch": 0.45435631521669545, + "grad_norm": 0.5262420177459717, + "learning_rate": 3.600905766294189e-05, + "loss": 0.4884, + "num_input_tokens_seen": 21592176, + "step": 14195 + }, + { + "epoch": 0.45451635618718395, + "grad_norm": 0.678762674331665, + "learning_rate": 3.60002424854957e-05, + "loss": 0.5284, + "num_input_tokens_seen": 21600160, + "step": 14200 + }, + { + "epoch": 0.45451635618718395, + "eval_loss": 0.5356020331382751, + "eval_runtime": 561.4764, + "eval_samples_per_second": 24.731, + "eval_steps_per_second": 12.366, + "num_input_tokens_seen": 21600160, + "step": 14200 + }, + { + "epoch": 0.4546763971576724, + "grad_norm": 0.38681405782699585, + "learning_rate": 3.5991425611673876e-05, + "loss": 0.4948, + "num_input_tokens_seen": 21607552, + "step": 14205 + }, + { + "epoch": 0.4548364381281608, + "grad_norm": 0.7021393179893494, + "learning_rate": 3.5982607042836105e-05, + "loss": 0.4406, + "num_input_tokens_seen": 21614992, + "step": 14210 + }, + { + "epoch": 0.45499647909864926, + "grad_norm": 1.4544216394424438, + "learning_rate": 3.597378678034231e-05, + "loss": 0.4245, + "num_input_tokens_seen": 21622416, + "step": 14215 + }, + { + "epoch": 0.4551565200691377, + "grad_norm": 1.0705076456069946, + "learning_rate": 3.596496482555269e-05, + "loss": 0.4367, + "num_input_tokens_seen": 21630240, + "step": 14220 + }, + { + "epoch": 0.45531656103962614, + "grad_norm": 1.0374680757522583, + "learning_rate": 3.595614117982769e-05, + "loss": 0.5673, + "num_input_tokens_seen": 21637888, + "step": 14225 + }, + { + "epoch": 0.4554766020101146, + "grad_norm": 0.8500940203666687, + "learning_rate": 3.594731584452805e-05, + "loss": 0.4116, + "num_input_tokens_seen": 21645776, + "step": 14230 + }, + { + "epoch": 0.455636642980603, + "grad_norm": 0.6904891133308411, + "learning_rate": 3.593848882101472e-05, + "loss": 0.5958, + "num_input_tokens_seen": 21653424, + "step": 14235 + }, + { + "epoch": 0.45579668395109146, + "grad_norm": 0.9496667385101318, + "learning_rate": 3.592966011064896e-05, + "loss": 0.5209, + "num_input_tokens_seen": 21661264, + "step": 14240 + }, + { + "epoch": 0.4559567249215799, + "grad_norm": 0.977471649646759, + "learning_rate": 3.592082971479226e-05, + "loss": 0.346, + "num_input_tokens_seen": 21669344, + "step": 14245 + }, + { + "epoch": 0.45611676589206834, + "grad_norm": 1.400998830795288, + "learning_rate": 3.5911997634806385e-05, + "loss": 0.5847, + "num_input_tokens_seen": 21676992, + "step": 14250 + }, + { + "epoch": 0.45627680686255684, + "grad_norm": 1.928209900856018, + "learning_rate": 3.5903163872053336e-05, + "loss": 0.598, + "num_input_tokens_seen": 21684640, + "step": 14255 + }, + { + "epoch": 0.4564368478330453, + "grad_norm": 0.9168707132339478, + "learning_rate": 3.58943284278954e-05, + "loss": 0.6351, + "num_input_tokens_seen": 21691808, + "step": 14260 + }, + { + "epoch": 0.4565968888035337, + "grad_norm": 0.9512290954589844, + "learning_rate": 3.588549130369512e-05, + "loss": 0.429, + "num_input_tokens_seen": 21699696, + "step": 14265 + }, + { + "epoch": 0.45675692977402216, + "grad_norm": 0.7337655425071716, + "learning_rate": 3.5876652500815274e-05, + "loss": 0.2927, + "num_input_tokens_seen": 21707440, + "step": 14270 + }, + { + "epoch": 0.4569169707445106, + "grad_norm": 0.6477466225624084, + "learning_rate": 3.586781202061894e-05, + "loss": 0.3812, + "num_input_tokens_seen": 21715088, + "step": 14275 + }, + { + "epoch": 0.45707701171499904, + "grad_norm": 0.5981266498565674, + "learning_rate": 3.585896986446942e-05, + "loss": 0.49, + "num_input_tokens_seen": 21722384, + "step": 14280 + }, + { + "epoch": 0.4572370526854875, + "grad_norm": 0.599774956703186, + "learning_rate": 3.585012603373028e-05, + "loss": 0.6437, + "num_input_tokens_seen": 21729840, + "step": 14285 + }, + { + "epoch": 0.4573970936559759, + "grad_norm": 0.706308126449585, + "learning_rate": 3.584128052976535e-05, + "loss": 0.6285, + "num_input_tokens_seen": 21737104, + "step": 14290 + }, + { + "epoch": 0.45755713462646436, + "grad_norm": 0.4027482867240906, + "learning_rate": 3.5832433353938724e-05, + "loss": 0.407, + "num_input_tokens_seen": 21744848, + "step": 14295 + }, + { + "epoch": 0.4577171755969528, + "grad_norm": 0.9633843898773193, + "learning_rate": 3.5823584507614746e-05, + "loss": 0.6712, + "num_input_tokens_seen": 21752432, + "step": 14300 + }, + { + "epoch": 0.4578772165674413, + "grad_norm": 1.825414776802063, + "learning_rate": 3.581473399215802e-05, + "loss": 0.5107, + "num_input_tokens_seen": 21760288, + "step": 14305 + }, + { + "epoch": 0.45803725753792973, + "grad_norm": 0.7842458486557007, + "learning_rate": 3.580588180893341e-05, + "loss": 0.6426, + "num_input_tokens_seen": 21768144, + "step": 14310 + }, + { + "epoch": 0.45819729850841817, + "grad_norm": 0.5895811319351196, + "learning_rate": 3.579702795930602e-05, + "loss": 0.5146, + "num_input_tokens_seen": 21776208, + "step": 14315 + }, + { + "epoch": 0.4583573394789066, + "grad_norm": 1.157383918762207, + "learning_rate": 3.578817244464125e-05, + "loss": 0.649, + "num_input_tokens_seen": 21783872, + "step": 14320 + }, + { + "epoch": 0.45851738044939505, + "grad_norm": 0.4553072452545166, + "learning_rate": 3.577931526630471e-05, + "loss": 0.4465, + "num_input_tokens_seen": 21792144, + "step": 14325 + }, + { + "epoch": 0.4586774214198835, + "grad_norm": 0.7018762826919556, + "learning_rate": 3.577045642566229e-05, + "loss": 0.4771, + "num_input_tokens_seen": 21799408, + "step": 14330 + }, + { + "epoch": 0.45883746239037193, + "grad_norm": 0.8182679414749146, + "learning_rate": 3.576159592408014e-05, + "loss": 0.4969, + "num_input_tokens_seen": 21806960, + "step": 14335 + }, + { + "epoch": 0.45899750336086037, + "grad_norm": 0.5945990085601807, + "learning_rate": 3.575273376292466e-05, + "loss": 0.5284, + "num_input_tokens_seen": 21814048, + "step": 14340 + }, + { + "epoch": 0.4591575443313488, + "grad_norm": 1.0729963779449463, + "learning_rate": 3.574386994356251e-05, + "loss": 0.5507, + "num_input_tokens_seen": 21821648, + "step": 14345 + }, + { + "epoch": 0.45931758530183725, + "grad_norm": 0.9806506037712097, + "learning_rate": 3.573500446736059e-05, + "loss": 0.6358, + "num_input_tokens_seen": 21828816, + "step": 14350 + }, + { + "epoch": 0.4594776262723257, + "grad_norm": 0.9115224480628967, + "learning_rate": 3.5726137335686094e-05, + "loss": 0.493, + "num_input_tokens_seen": 21837024, + "step": 14355 + }, + { + "epoch": 0.4596376672428142, + "grad_norm": 0.5087637901306152, + "learning_rate": 3.571726854990642e-05, + "loss": 0.4738, + "num_input_tokens_seen": 21844800, + "step": 14360 + }, + { + "epoch": 0.4597977082133026, + "grad_norm": 1.1161201000213623, + "learning_rate": 3.570839811138925e-05, + "loss": 0.5898, + "num_input_tokens_seen": 21852352, + "step": 14365 + }, + { + "epoch": 0.45995774918379106, + "grad_norm": 0.9914954304695129, + "learning_rate": 3.569952602150252e-05, + "loss": 0.716, + "num_input_tokens_seen": 21860064, + "step": 14370 + }, + { + "epoch": 0.4601177901542795, + "grad_norm": 2.358882188796997, + "learning_rate": 3.569065228161442e-05, + "loss": 0.5426, + "num_input_tokens_seen": 21867632, + "step": 14375 + }, + { + "epoch": 0.46027783112476794, + "grad_norm": 0.439048171043396, + "learning_rate": 3.5681776893093395e-05, + "loss": 0.4379, + "num_input_tokens_seen": 21875440, + "step": 14380 + }, + { + "epoch": 0.4604378720952564, + "grad_norm": 1.1896010637283325, + "learning_rate": 3.5672899857308134e-05, + "loss": 0.6036, + "num_input_tokens_seen": 21882736, + "step": 14385 + }, + { + "epoch": 0.4605979130657448, + "grad_norm": 0.614375650882721, + "learning_rate": 3.566402117562759e-05, + "loss": 0.4948, + "num_input_tokens_seen": 21890720, + "step": 14390 + }, + { + "epoch": 0.46075795403623326, + "grad_norm": 1.1841224431991577, + "learning_rate": 3.565514084942097e-05, + "loss": 0.6263, + "num_input_tokens_seen": 21898848, + "step": 14395 + }, + { + "epoch": 0.4609179950067217, + "grad_norm": 0.6967141628265381, + "learning_rate": 3.564625888005773e-05, + "loss": 0.4407, + "num_input_tokens_seen": 21906528, + "step": 14400 + }, + { + "epoch": 0.4609179950067217, + "eval_loss": 0.5345343351364136, + "eval_runtime": 561.0978, + "eval_samples_per_second": 24.748, + "eval_steps_per_second": 12.374, + "num_input_tokens_seen": 21906528, + "step": 14400 + }, + { + "epoch": 0.46107803597721014, + "grad_norm": 0.5084050893783569, + "learning_rate": 3.563737526890759e-05, + "loss": 0.3646, + "num_input_tokens_seen": 21914016, + "step": 14405 + }, + { + "epoch": 0.4612380769476986, + "grad_norm": 0.7169420123100281, + "learning_rate": 3.562849001734049e-05, + "loss": 0.6087, + "num_input_tokens_seen": 21921664, + "step": 14410 + }, + { + "epoch": 0.4613981179181871, + "grad_norm": 1.1740858554840088, + "learning_rate": 3.561960312672667e-05, + "loss": 0.5813, + "num_input_tokens_seen": 21928816, + "step": 14415 + }, + { + "epoch": 0.4615581588886755, + "grad_norm": 0.6015251874923706, + "learning_rate": 3.5610714598436596e-05, + "loss": 0.7107, + "num_input_tokens_seen": 21936384, + "step": 14420 + }, + { + "epoch": 0.46171819985916396, + "grad_norm": 1.1279683113098145, + "learning_rate": 3.5601824433840986e-05, + "loss": 0.3827, + "num_input_tokens_seen": 21943856, + "step": 14425 + }, + { + "epoch": 0.4618782408296524, + "grad_norm": 0.9316025376319885, + "learning_rate": 3.559293263431082e-05, + "loss": 0.667, + "num_input_tokens_seen": 21951248, + "step": 14430 + }, + { + "epoch": 0.46203828180014084, + "grad_norm": 0.5197045803070068, + "learning_rate": 3.558403920121732e-05, + "loss": 0.4217, + "num_input_tokens_seen": 21958768, + "step": 14435 + }, + { + "epoch": 0.4621983227706293, + "grad_norm": 0.7033841013908386, + "learning_rate": 3.557514413593197e-05, + "loss": 0.4268, + "num_input_tokens_seen": 21966160, + "step": 14440 + }, + { + "epoch": 0.4623583637411177, + "grad_norm": 0.3905976116657257, + "learning_rate": 3.55662474398265e-05, + "loss": 0.427, + "num_input_tokens_seen": 21974048, + "step": 14445 + }, + { + "epoch": 0.46251840471160616, + "grad_norm": 0.44435304403305054, + "learning_rate": 3.555734911427288e-05, + "loss": 0.4026, + "num_input_tokens_seen": 21981616, + "step": 14450 + }, + { + "epoch": 0.4626784456820946, + "grad_norm": 1.420907974243164, + "learning_rate": 3.5548449160643363e-05, + "loss": 0.4865, + "num_input_tokens_seen": 21989296, + "step": 14455 + }, + { + "epoch": 0.46283848665258304, + "grad_norm": 1.2709108591079712, + "learning_rate": 3.553954758031043e-05, + "loss": 0.6037, + "num_input_tokens_seen": 21996960, + "step": 14460 + }, + { + "epoch": 0.46299852762307153, + "grad_norm": 1.6061983108520508, + "learning_rate": 3.5530644374646815e-05, + "loss": 0.6375, + "num_input_tokens_seen": 22004352, + "step": 14465 + }, + { + "epoch": 0.46315856859355997, + "grad_norm": 0.7390985488891602, + "learning_rate": 3.552173954502549e-05, + "loss": 0.3943, + "num_input_tokens_seen": 22011856, + "step": 14470 + }, + { + "epoch": 0.4633186095640484, + "grad_norm": 0.6833860874176025, + "learning_rate": 3.55128330928197e-05, + "loss": 0.729, + "num_input_tokens_seen": 22019136, + "step": 14475 + }, + { + "epoch": 0.46347865053453685, + "grad_norm": 1.1599355936050415, + "learning_rate": 3.550392501940294e-05, + "loss": 0.5136, + "num_input_tokens_seen": 22026464, + "step": 14480 + }, + { + "epoch": 0.4636386915050253, + "grad_norm": 1.1800490617752075, + "learning_rate": 3.5495015326148945e-05, + "loss": 0.5451, + "num_input_tokens_seen": 22034064, + "step": 14485 + }, + { + "epoch": 0.46379873247551373, + "grad_norm": 0.821732759475708, + "learning_rate": 3.548610401443169e-05, + "loss": 0.5804, + "num_input_tokens_seen": 22041504, + "step": 14490 + }, + { + "epoch": 0.46395877344600217, + "grad_norm": 0.5453208088874817, + "learning_rate": 3.547719108562543e-05, + "loss": 0.4342, + "num_input_tokens_seen": 22048784, + "step": 14495 + }, + { + "epoch": 0.4641188144164906, + "grad_norm": 1.2954705953598022, + "learning_rate": 3.546827654110464e-05, + "loss": 0.715, + "num_input_tokens_seen": 22056016, + "step": 14500 + }, + { + "epoch": 0.46427885538697905, + "grad_norm": 0.6601961851119995, + "learning_rate": 3.545936038224405e-05, + "loss": 0.3374, + "num_input_tokens_seen": 22063520, + "step": 14505 + }, + { + "epoch": 0.4644388963574675, + "grad_norm": 0.950874924659729, + "learning_rate": 3.545044261041864e-05, + "loss": 0.3425, + "num_input_tokens_seen": 22071168, + "step": 14510 + }, + { + "epoch": 0.46459893732795593, + "grad_norm": 0.33140575885772705, + "learning_rate": 3.5441523227003657e-05, + "loss": 0.5792, + "num_input_tokens_seen": 22079408, + "step": 14515 + }, + { + "epoch": 0.4647589782984444, + "grad_norm": 0.7382060289382935, + "learning_rate": 3.543260223337459e-05, + "loss": 0.3779, + "num_input_tokens_seen": 22086848, + "step": 14520 + }, + { + "epoch": 0.46491901926893286, + "grad_norm": 0.9868638515472412, + "learning_rate": 3.542367963090714e-05, + "loss": 0.5115, + "num_input_tokens_seen": 22094576, + "step": 14525 + }, + { + "epoch": 0.4650790602394213, + "grad_norm": 1.2402455806732178, + "learning_rate": 3.5414755420977295e-05, + "loss": 0.6087, + "num_input_tokens_seen": 22102336, + "step": 14530 + }, + { + "epoch": 0.46523910120990974, + "grad_norm": 1.5917073488235474, + "learning_rate": 3.54058296049613e-05, + "loss": 0.6316, + "num_input_tokens_seen": 22110128, + "step": 14535 + }, + { + "epoch": 0.4653991421803982, + "grad_norm": 0.4831082224845886, + "learning_rate": 3.53969021842356e-05, + "loss": 0.5666, + "num_input_tokens_seen": 22117760, + "step": 14540 + }, + { + "epoch": 0.4655591831508866, + "grad_norm": 0.7243170142173767, + "learning_rate": 3.5387973160176926e-05, + "loss": 0.4883, + "num_input_tokens_seen": 22125360, + "step": 14545 + }, + { + "epoch": 0.46571922412137506, + "grad_norm": 0.5065599083900452, + "learning_rate": 3.537904253416224e-05, + "loss": 0.4766, + "num_input_tokens_seen": 22133008, + "step": 14550 + }, + { + "epoch": 0.4658792650918635, + "grad_norm": 1.5126302242279053, + "learning_rate": 3.537011030756878e-05, + "loss": 0.782, + "num_input_tokens_seen": 22143952, + "step": 14555 + }, + { + "epoch": 0.46603930606235194, + "grad_norm": 1.1081652641296387, + "learning_rate": 3.536117648177399e-05, + "loss": 0.7228, + "num_input_tokens_seen": 22151728, + "step": 14560 + }, + { + "epoch": 0.4661993470328404, + "grad_norm": 1.1407417058944702, + "learning_rate": 3.535224105815558e-05, + "loss": 0.5474, + "num_input_tokens_seen": 22159248, + "step": 14565 + }, + { + "epoch": 0.4663593880033289, + "grad_norm": 0.5741057395935059, + "learning_rate": 3.5343304038091494e-05, + "loss": 0.5307, + "num_input_tokens_seen": 22166880, + "step": 14570 + }, + { + "epoch": 0.4665194289738173, + "grad_norm": 0.5695005059242249, + "learning_rate": 3.5334365422959955e-05, + "loss": 0.5483, + "num_input_tokens_seen": 22174496, + "step": 14575 + }, + { + "epoch": 0.46667946994430576, + "grad_norm": 0.4780291020870209, + "learning_rate": 3.5325425214139396e-05, + "loss": 0.448, + "num_input_tokens_seen": 22182064, + "step": 14580 + }, + { + "epoch": 0.4668395109147942, + "grad_norm": 0.9551646709442139, + "learning_rate": 3.531648341300851e-05, + "loss": 0.655, + "num_input_tokens_seen": 22190000, + "step": 14585 + }, + { + "epoch": 0.46699955188528264, + "grad_norm": 0.5952916145324707, + "learning_rate": 3.530754002094623e-05, + "loss": 0.4502, + "num_input_tokens_seen": 22197552, + "step": 14590 + }, + { + "epoch": 0.4671595928557711, + "grad_norm": 0.8325368762016296, + "learning_rate": 3.529859503933175e-05, + "loss": 0.4718, + "num_input_tokens_seen": 22205408, + "step": 14595 + }, + { + "epoch": 0.4673196338262595, + "grad_norm": 0.64788419008255, + "learning_rate": 3.52896484695445e-05, + "loss": 0.4864, + "num_input_tokens_seen": 22212960, + "step": 14600 + }, + { + "epoch": 0.4673196338262595, + "eval_loss": 0.533869206905365, + "eval_runtime": 561.1875, + "eval_samples_per_second": 24.744, + "eval_steps_per_second": 12.372, + "num_input_tokens_seen": 22212960, + "step": 14600 + }, + { + "epoch": 0.46747967479674796, + "grad_norm": 0.5561193823814392, + "learning_rate": 3.528070031296414e-05, + "loss": 0.7149, + "num_input_tokens_seen": 22221136, + "step": 14605 + }, + { + "epoch": 0.4676397157672364, + "grad_norm": 0.5527142882347107, + "learning_rate": 3.5271750570970605e-05, + "loss": 0.6399, + "num_input_tokens_seen": 22228992, + "step": 14610 + }, + { + "epoch": 0.46779975673772484, + "grad_norm": 1.0689257383346558, + "learning_rate": 3.526279924494405e-05, + "loss": 0.6217, + "num_input_tokens_seen": 22236864, + "step": 14615 + }, + { + "epoch": 0.4679597977082133, + "grad_norm": 0.9634710550308228, + "learning_rate": 3.5253846336264874e-05, + "loss": 0.3999, + "num_input_tokens_seen": 22244416, + "step": 14620 + }, + { + "epoch": 0.46811983867870177, + "grad_norm": 0.6490592956542969, + "learning_rate": 3.5244891846313736e-05, + "loss": 0.3559, + "num_input_tokens_seen": 22251712, + "step": 14625 + }, + { + "epoch": 0.4682798796491902, + "grad_norm": 0.516569972038269, + "learning_rate": 3.5235935776471527e-05, + "loss": 0.5879, + "num_input_tokens_seen": 22259120, + "step": 14630 + }, + { + "epoch": 0.46843992061967865, + "grad_norm": 0.9371219277381897, + "learning_rate": 3.522697812811939e-05, + "loss": 0.4886, + "num_input_tokens_seen": 22266912, + "step": 14635 + }, + { + "epoch": 0.4685999615901671, + "grad_norm": 1.3651434183120728, + "learning_rate": 3.521801890263871e-05, + "loss": 0.6868, + "num_input_tokens_seen": 22274352, + "step": 14640 + }, + { + "epoch": 0.46876000256065553, + "grad_norm": 0.9815301895141602, + "learning_rate": 3.5209058101411114e-05, + "loss": 0.5317, + "num_input_tokens_seen": 22282288, + "step": 14645 + }, + { + "epoch": 0.46892004353114397, + "grad_norm": 0.49843713641166687, + "learning_rate": 3.520009572581845e-05, + "loss": 0.4083, + "num_input_tokens_seen": 22289888, + "step": 14650 + }, + { + "epoch": 0.4690800845016324, + "grad_norm": 0.5346230864524841, + "learning_rate": 3.519113177724285e-05, + "loss": 0.3982, + "num_input_tokens_seen": 22297728, + "step": 14655 + }, + { + "epoch": 0.46924012547212085, + "grad_norm": 1.362634539604187, + "learning_rate": 3.5182166257066656e-05, + "loss": 0.5911, + "num_input_tokens_seen": 22305728, + "step": 14660 + }, + { + "epoch": 0.4694001664426093, + "grad_norm": 1.1132986545562744, + "learning_rate": 3.517319916667247e-05, + "loss": 0.5454, + "num_input_tokens_seen": 22313136, + "step": 14665 + }, + { + "epoch": 0.46956020741309773, + "grad_norm": 0.48812347650527954, + "learning_rate": 3.516423050744313e-05, + "loss": 0.4235, + "num_input_tokens_seen": 22320800, + "step": 14670 + }, + { + "epoch": 0.4697202483835862, + "grad_norm": 0.6941549777984619, + "learning_rate": 3.5155260280761704e-05, + "loss": 0.6141, + "num_input_tokens_seen": 22328144, + "step": 14675 + }, + { + "epoch": 0.46988028935407467, + "grad_norm": 1.012425184249878, + "learning_rate": 3.514628848801154e-05, + "loss": 0.5441, + "num_input_tokens_seen": 22335648, + "step": 14680 + }, + { + "epoch": 0.4700403303245631, + "grad_norm": 1.388817548751831, + "learning_rate": 3.5137315130576174e-05, + "loss": 0.6164, + "num_input_tokens_seen": 22343632, + "step": 14685 + }, + { + "epoch": 0.47020037129505154, + "grad_norm": 1.1670794486999512, + "learning_rate": 3.512834020983942e-05, + "loss": 0.5844, + "num_input_tokens_seen": 22351104, + "step": 14690 + }, + { + "epoch": 0.47036041226554, + "grad_norm": 0.5022416114807129, + "learning_rate": 3.5119363727185334e-05, + "loss": 0.4236, + "num_input_tokens_seen": 22358416, + "step": 14695 + }, + { + "epoch": 0.4705204532360284, + "grad_norm": 0.6720594167709351, + "learning_rate": 3.511038568399819e-05, + "loss": 0.4769, + "num_input_tokens_seen": 22366352, + "step": 14700 + }, + { + "epoch": 0.47068049420651686, + "grad_norm": 0.6557324528694153, + "learning_rate": 3.510140608166251e-05, + "loss": 0.4796, + "num_input_tokens_seen": 22374272, + "step": 14705 + }, + { + "epoch": 0.4708405351770053, + "grad_norm": 0.6925322413444519, + "learning_rate": 3.509242492156308e-05, + "loss": 0.5234, + "num_input_tokens_seen": 22381744, + "step": 14710 + }, + { + "epoch": 0.47100057614749374, + "grad_norm": 0.5873704552650452, + "learning_rate": 3.5083442205084896e-05, + "loss": 0.4857, + "num_input_tokens_seen": 22388880, + "step": 14715 + }, + { + "epoch": 0.4711606171179822, + "grad_norm": 0.8520143628120422, + "learning_rate": 3.507445793361321e-05, + "loss": 0.3814, + "num_input_tokens_seen": 22396528, + "step": 14720 + }, + { + "epoch": 0.4713206580884706, + "grad_norm": 0.4877108931541443, + "learning_rate": 3.5065472108533505e-05, + "loss": 0.3634, + "num_input_tokens_seen": 22404064, + "step": 14725 + }, + { + "epoch": 0.4714806990589591, + "grad_norm": 1.0708534717559814, + "learning_rate": 3.5056484731231504e-05, + "loss": 0.4302, + "num_input_tokens_seen": 22411744, + "step": 14730 + }, + { + "epoch": 0.47164074002944756, + "grad_norm": 1.4063130617141724, + "learning_rate": 3.504749580309319e-05, + "loss": 0.5759, + "num_input_tokens_seen": 22419792, + "step": 14735 + }, + { + "epoch": 0.471800780999936, + "grad_norm": 1.382854700088501, + "learning_rate": 3.5038505325504753e-05, + "loss": 0.5752, + "num_input_tokens_seen": 22427488, + "step": 14740 + }, + { + "epoch": 0.47196082197042444, + "grad_norm": 0.9170624017715454, + "learning_rate": 3.502951329985264e-05, + "loss": 0.5366, + "num_input_tokens_seen": 22435184, + "step": 14745 + }, + { + "epoch": 0.4721208629409129, + "grad_norm": 0.5855825543403625, + "learning_rate": 3.502051972752354e-05, + "loss": 0.4488, + "num_input_tokens_seen": 22442816, + "step": 14750 + }, + { + "epoch": 0.4722809039114013, + "grad_norm": 1.5055549144744873, + "learning_rate": 3.5011524609904374e-05, + "loss": 0.594, + "num_input_tokens_seen": 22450656, + "step": 14755 + }, + { + "epoch": 0.47244094488188976, + "grad_norm": 0.5585252642631531, + "learning_rate": 3.50025279483823e-05, + "loss": 0.5776, + "num_input_tokens_seen": 22458528, + "step": 14760 + }, + { + "epoch": 0.4726009858523782, + "grad_norm": 1.1384453773498535, + "learning_rate": 3.499352974434472e-05, + "loss": 0.3768, + "num_input_tokens_seen": 22466320, + "step": 14765 + }, + { + "epoch": 0.47276102682286664, + "grad_norm": 1.1482281684875488, + "learning_rate": 3.498452999917926e-05, + "loss": 0.5515, + "num_input_tokens_seen": 22473792, + "step": 14770 + }, + { + "epoch": 0.4729210677933551, + "grad_norm": 0.6059805154800415, + "learning_rate": 3.4975528714273795e-05, + "loss": 0.6574, + "num_input_tokens_seen": 22481360, + "step": 14775 + }, + { + "epoch": 0.4730811087638435, + "grad_norm": 0.856903076171875, + "learning_rate": 3.4966525891016454e-05, + "loss": 0.6477, + "num_input_tokens_seen": 22488848, + "step": 14780 + }, + { + "epoch": 0.473241149734332, + "grad_norm": 0.7179682850837708, + "learning_rate": 3.495752153079557e-05, + "loss": 0.3929, + "num_input_tokens_seen": 22496080, + "step": 14785 + }, + { + "epoch": 0.47340119070482045, + "grad_norm": 0.9205502271652222, + "learning_rate": 3.494851563499974e-05, + "loss": 0.45, + "num_input_tokens_seen": 22503328, + "step": 14790 + }, + { + "epoch": 0.4735612316753089, + "grad_norm": 0.5413113832473755, + "learning_rate": 3.493950820501777e-05, + "loss": 0.6975, + "num_input_tokens_seen": 22510624, + "step": 14795 + }, + { + "epoch": 0.47372127264579733, + "grad_norm": 0.4410202205181122, + "learning_rate": 3.493049924223872e-05, + "loss": 0.465, + "num_input_tokens_seen": 22518592, + "step": 14800 + }, + { + "epoch": 0.47372127264579733, + "eval_loss": 0.5332948565483093, + "eval_runtime": 561.1264, + "eval_samples_per_second": 24.747, + "eval_steps_per_second": 12.373, + "num_input_tokens_seen": 22518592, + "step": 14800 + }, + { + "epoch": 0.47388131361628577, + "grad_norm": 0.703149139881134, + "learning_rate": 3.49214887480519e-05, + "loss": 0.3992, + "num_input_tokens_seen": 22526096, + "step": 14805 + }, + { + "epoch": 0.4740413545867742, + "grad_norm": 1.2640787363052368, + "learning_rate": 3.4912476723846834e-05, + "loss": 0.4403, + "num_input_tokens_seen": 22533984, + "step": 14810 + }, + { + "epoch": 0.47420139555726265, + "grad_norm": 0.6635756492614746, + "learning_rate": 3.490346317101328e-05, + "loss": 0.3904, + "num_input_tokens_seen": 22541312, + "step": 14815 + }, + { + "epoch": 0.4743614365277511, + "grad_norm": 0.7267002463340759, + "learning_rate": 3.4894448090941266e-05, + "loss": 0.4196, + "num_input_tokens_seen": 22549024, + "step": 14820 + }, + { + "epoch": 0.47452147749823953, + "grad_norm": 0.5621160864830017, + "learning_rate": 3.488543148502101e-05, + "loss": 0.5848, + "num_input_tokens_seen": 22556224, + "step": 14825 + }, + { + "epoch": 0.47468151846872797, + "grad_norm": 0.8941195607185364, + "learning_rate": 3.487641335464299e-05, + "loss": 0.3825, + "num_input_tokens_seen": 22563552, + "step": 14830 + }, + { + "epoch": 0.47484155943921647, + "grad_norm": 0.6432763338088989, + "learning_rate": 3.4867393701197914e-05, + "loss": 0.5315, + "num_input_tokens_seen": 22571056, + "step": 14835 + }, + { + "epoch": 0.4750016004097049, + "grad_norm": 1.2754031419754028, + "learning_rate": 3.485837252607673e-05, + "loss": 0.464, + "num_input_tokens_seen": 22578512, + "step": 14840 + }, + { + "epoch": 0.47516164138019334, + "grad_norm": 1.0060758590698242, + "learning_rate": 3.4849349830670615e-05, + "loss": 0.4866, + "num_input_tokens_seen": 22586272, + "step": 14845 + }, + { + "epoch": 0.4753216823506818, + "grad_norm": 0.779677152633667, + "learning_rate": 3.4840325616370976e-05, + "loss": 0.5193, + "num_input_tokens_seen": 22593824, + "step": 14850 + }, + { + "epoch": 0.4754817233211702, + "grad_norm": 0.6249027848243713, + "learning_rate": 3.483129988456947e-05, + "loss": 0.3886, + "num_input_tokens_seen": 22601616, + "step": 14855 + }, + { + "epoch": 0.47564176429165866, + "grad_norm": 0.6564319729804993, + "learning_rate": 3.482227263665797e-05, + "loss": 0.4285, + "num_input_tokens_seen": 22609536, + "step": 14860 + }, + { + "epoch": 0.4758018052621471, + "grad_norm": 0.41057267785072327, + "learning_rate": 3.48132438740286e-05, + "loss": 0.464, + "num_input_tokens_seen": 22616960, + "step": 14865 + }, + { + "epoch": 0.47596184623263554, + "grad_norm": 0.28808119893074036, + "learning_rate": 3.48042135980737e-05, + "loss": 0.2776, + "num_input_tokens_seen": 22624352, + "step": 14870 + }, + { + "epoch": 0.476121887203124, + "grad_norm": 0.8175050020217896, + "learning_rate": 3.479518181018586e-05, + "loss": 0.5566, + "num_input_tokens_seen": 22631968, + "step": 14875 + }, + { + "epoch": 0.4762819281736124, + "grad_norm": 0.9796851873397827, + "learning_rate": 3.4786148511757886e-05, + "loss": 0.6967, + "num_input_tokens_seen": 22639536, + "step": 14880 + }, + { + "epoch": 0.47644196914410086, + "grad_norm": 1.039734959602356, + "learning_rate": 3.477711370418284e-05, + "loss": 0.4662, + "num_input_tokens_seen": 22647440, + "step": 14885 + }, + { + "epoch": 0.47660201011458936, + "grad_norm": 0.6727138161659241, + "learning_rate": 3.476807738885399e-05, + "loss": 0.4585, + "num_input_tokens_seen": 22654960, + "step": 14890 + }, + { + "epoch": 0.4767620510850778, + "grad_norm": 0.46171942353248596, + "learning_rate": 3.475903956716485e-05, + "loss": 0.4339, + "num_input_tokens_seen": 22662208, + "step": 14895 + }, + { + "epoch": 0.47692209205556624, + "grad_norm": 1.6079398393630981, + "learning_rate": 3.475000024050917e-05, + "loss": 0.4725, + "num_input_tokens_seen": 22669424, + "step": 14900 + }, + { + "epoch": 0.4770821330260547, + "grad_norm": 1.1246964931488037, + "learning_rate": 3.4740959410280926e-05, + "loss": 0.5135, + "num_input_tokens_seen": 22677088, + "step": 14905 + }, + { + "epoch": 0.4772421739965431, + "grad_norm": 0.9085621237754822, + "learning_rate": 3.4731917077874324e-05, + "loss": 0.5828, + "num_input_tokens_seen": 22684528, + "step": 14910 + }, + { + "epoch": 0.47740221496703156, + "grad_norm": 0.5058877468109131, + "learning_rate": 3.4722873244683816e-05, + "loss": 0.5135, + "num_input_tokens_seen": 22691840, + "step": 14915 + }, + { + "epoch": 0.47756225593752, + "grad_norm": 1.2134699821472168, + "learning_rate": 3.4713827912104065e-05, + "loss": 0.5699, + "num_input_tokens_seen": 22699440, + "step": 14920 + }, + { + "epoch": 0.47772229690800844, + "grad_norm": 0.5420910120010376, + "learning_rate": 3.470478108152998e-05, + "loss": 0.4154, + "num_input_tokens_seen": 22707280, + "step": 14925 + }, + { + "epoch": 0.4778823378784969, + "grad_norm": 0.41360175609588623, + "learning_rate": 3.4695732754356695e-05, + "loss": 0.5053, + "num_input_tokens_seen": 22714976, + "step": 14930 + }, + { + "epoch": 0.4780423788489853, + "grad_norm": 0.7165433168411255, + "learning_rate": 3.4686682931979576e-05, + "loss": 0.4673, + "num_input_tokens_seen": 22722192, + "step": 14935 + }, + { + "epoch": 0.4782024198194738, + "grad_norm": 0.7022182941436768, + "learning_rate": 3.467763161579422e-05, + "loss": 0.4636, + "num_input_tokens_seen": 22730032, + "step": 14940 + }, + { + "epoch": 0.47836246078996225, + "grad_norm": 0.7270776629447937, + "learning_rate": 3.466857880719645e-05, + "loss": 0.4797, + "num_input_tokens_seen": 22738128, + "step": 14945 + }, + { + "epoch": 0.4785225017604507, + "grad_norm": 2.082087516784668, + "learning_rate": 3.465952450758233e-05, + "loss": 0.5032, + "num_input_tokens_seen": 22745280, + "step": 14950 + }, + { + "epoch": 0.47868254273093913, + "grad_norm": 0.3414320945739746, + "learning_rate": 3.4650468718348126e-05, + "loss": 0.6482, + "num_input_tokens_seen": 22753104, + "step": 14955 + }, + { + "epoch": 0.47884258370142757, + "grad_norm": 0.9150534868240356, + "learning_rate": 3.464141144089038e-05, + "loss": 0.6323, + "num_input_tokens_seen": 22760688, + "step": 14960 + }, + { + "epoch": 0.479002624671916, + "grad_norm": 0.5373393893241882, + "learning_rate": 3.463235267660583e-05, + "loss": 0.4871, + "num_input_tokens_seen": 22768160, + "step": 14965 + }, + { + "epoch": 0.47916266564240445, + "grad_norm": 0.6325352191925049, + "learning_rate": 3.462329242689145e-05, + "loss": 0.3796, + "num_input_tokens_seen": 22775712, + "step": 14970 + }, + { + "epoch": 0.4793227066128929, + "grad_norm": 1.037460446357727, + "learning_rate": 3.461423069314444e-05, + "loss": 0.4173, + "num_input_tokens_seen": 22783120, + "step": 14975 + }, + { + "epoch": 0.47948274758338133, + "grad_norm": 0.9208448529243469, + "learning_rate": 3.460516747676224e-05, + "loss": 0.4549, + "num_input_tokens_seen": 22790656, + "step": 14980 + }, + { + "epoch": 0.47964278855386977, + "grad_norm": 0.7294980883598328, + "learning_rate": 3.459610277914251e-05, + "loss": 0.5556, + "num_input_tokens_seen": 22798208, + "step": 14985 + }, + { + "epoch": 0.4798028295243582, + "grad_norm": 0.6741784811019897, + "learning_rate": 3.458703660168314e-05, + "loss": 0.5811, + "num_input_tokens_seen": 22806032, + "step": 14990 + }, + { + "epoch": 0.4799628704948467, + "grad_norm": 1.3963512182235718, + "learning_rate": 3.457796894578224e-05, + "loss": 0.4194, + "num_input_tokens_seen": 22813584, + "step": 14995 + }, + { + "epoch": 0.48012291146533514, + "grad_norm": 0.9518356323242188, + "learning_rate": 3.456889981283817e-05, + "loss": 0.5099, + "num_input_tokens_seen": 22821472, + "step": 15000 + }, + { + "epoch": 0.48012291146533514, + "eval_loss": 0.5327737331390381, + "eval_runtime": 561.1842, + "eval_samples_per_second": 24.744, + "eval_steps_per_second": 12.372, + "num_input_tokens_seen": 22821472, + "step": 15000 + }, + { + "epoch": 0.4802829524358236, + "grad_norm": 0.36937642097473145, + "learning_rate": 3.45598292042495e-05, + "loss": 0.6225, + "num_input_tokens_seen": 22829216, + "step": 15005 + }, + { + "epoch": 0.480442993406312, + "grad_norm": 0.9168657064437866, + "learning_rate": 3.4550757121415035e-05, + "loss": 0.4168, + "num_input_tokens_seen": 22836352, + "step": 15010 + }, + { + "epoch": 0.48060303437680046, + "grad_norm": 1.4564446210861206, + "learning_rate": 3.454168356573378e-05, + "loss": 0.6455, + "num_input_tokens_seen": 22843296, + "step": 15015 + }, + { + "epoch": 0.4807630753472889, + "grad_norm": 0.6965118646621704, + "learning_rate": 3.453260853860503e-05, + "loss": 0.48, + "num_input_tokens_seen": 22851184, + "step": 15020 + }, + { + "epoch": 0.48092311631777734, + "grad_norm": 0.9315576553344727, + "learning_rate": 3.452353204142824e-05, + "loss": 0.525, + "num_input_tokens_seen": 22858640, + "step": 15025 + }, + { + "epoch": 0.4810831572882658, + "grad_norm": 0.8811416029930115, + "learning_rate": 3.4514454075603136e-05, + "loss": 0.4809, + "num_input_tokens_seen": 22865984, + "step": 15030 + }, + { + "epoch": 0.4812431982587542, + "grad_norm": 1.0374784469604492, + "learning_rate": 3.450537464252964e-05, + "loss": 0.5807, + "num_input_tokens_seen": 22874096, + "step": 15035 + }, + { + "epoch": 0.48140323922924266, + "grad_norm": 1.7273118495941162, + "learning_rate": 3.4496293743607925e-05, + "loss": 0.6435, + "num_input_tokens_seen": 22881360, + "step": 15040 + }, + { + "epoch": 0.48156328019973116, + "grad_norm": 0.7049059867858887, + "learning_rate": 3.448721138023838e-05, + "loss": 0.3892, + "num_input_tokens_seen": 22888912, + "step": 15045 + }, + { + "epoch": 0.4817233211702196, + "grad_norm": 0.8301808834075928, + "learning_rate": 3.447812755382162e-05, + "loss": 0.6913, + "num_input_tokens_seen": 22896304, + "step": 15050 + }, + { + "epoch": 0.48188336214070804, + "grad_norm": 1.109824776649475, + "learning_rate": 3.446904226575847e-05, + "loss": 0.5356, + "num_input_tokens_seen": 22903392, + "step": 15055 + }, + { + "epoch": 0.4820434031111965, + "grad_norm": 1.2561209201812744, + "learning_rate": 3.445995551745002e-05, + "loss": 0.7168, + "num_input_tokens_seen": 22911360, + "step": 15060 + }, + { + "epoch": 0.4822034440816849, + "grad_norm": 0.6806330680847168, + "learning_rate": 3.445086731029753e-05, + "loss": 0.6367, + "num_input_tokens_seen": 22919024, + "step": 15065 + }, + { + "epoch": 0.48236348505217336, + "grad_norm": 1.4063421487808228, + "learning_rate": 3.444177764570255e-05, + "loss": 0.5105, + "num_input_tokens_seen": 22926624, + "step": 15070 + }, + { + "epoch": 0.4825235260226618, + "grad_norm": 1.0704753398895264, + "learning_rate": 3.44326865250668e-05, + "loss": 0.6209, + "num_input_tokens_seen": 22934112, + "step": 15075 + }, + { + "epoch": 0.48268356699315024, + "grad_norm": 0.5276826024055481, + "learning_rate": 3.442359394979225e-05, + "loss": 0.5417, + "num_input_tokens_seen": 22941456, + "step": 15080 + }, + { + "epoch": 0.4828436079636387, + "grad_norm": 0.7733486294746399, + "learning_rate": 3.441449992128108e-05, + "loss": 0.4938, + "num_input_tokens_seen": 22949360, + "step": 15085 + }, + { + "epoch": 0.4830036489341271, + "grad_norm": 1.002587080001831, + "learning_rate": 3.440540444093573e-05, + "loss": 0.4172, + "num_input_tokens_seen": 22956880, + "step": 15090 + }, + { + "epoch": 0.48316368990461556, + "grad_norm": 1.3971632719039917, + "learning_rate": 3.43963075101588e-05, + "loss": 0.5837, + "num_input_tokens_seen": 22964624, + "step": 15095 + }, + { + "epoch": 0.48332373087510405, + "grad_norm": 0.4215671122074127, + "learning_rate": 3.438720913035318e-05, + "loss": 0.3524, + "num_input_tokens_seen": 22972048, + "step": 15100 + }, + { + "epoch": 0.4834837718455925, + "grad_norm": 0.680838406085968, + "learning_rate": 3.437810930292195e-05, + "loss": 0.4138, + "num_input_tokens_seen": 22979840, + "step": 15105 + }, + { + "epoch": 0.48364381281608093, + "grad_norm": 0.628270149230957, + "learning_rate": 3.43690080292684e-05, + "loss": 0.3777, + "num_input_tokens_seen": 22987264, + "step": 15110 + }, + { + "epoch": 0.48380385378656937, + "grad_norm": 0.9177565574645996, + "learning_rate": 3.435990531079608e-05, + "loss": 0.388, + "num_input_tokens_seen": 22994496, + "step": 15115 + }, + { + "epoch": 0.4839638947570578, + "grad_norm": 0.4701400101184845, + "learning_rate": 3.435080114890874e-05, + "loss": 0.5108, + "num_input_tokens_seen": 23002064, + "step": 15120 + }, + { + "epoch": 0.48412393572754625, + "grad_norm": 0.7196186780929565, + "learning_rate": 3.434169554501035e-05, + "loss": 0.4619, + "num_input_tokens_seen": 23009680, + "step": 15125 + }, + { + "epoch": 0.4842839766980347, + "grad_norm": 0.9729064702987671, + "learning_rate": 3.433258850050511e-05, + "loss": 0.6906, + "num_input_tokens_seen": 23017472, + "step": 15130 + }, + { + "epoch": 0.48444401766852313, + "grad_norm": 1.3274049758911133, + "learning_rate": 3.4323480016797446e-05, + "loss": 0.4312, + "num_input_tokens_seen": 23025200, + "step": 15135 + }, + { + "epoch": 0.48460405863901157, + "grad_norm": 2.1496963500976562, + "learning_rate": 3.4314370095291995e-05, + "loss": 0.6969, + "num_input_tokens_seen": 23033168, + "step": 15140 + }, + { + "epoch": 0.4847640996095, + "grad_norm": 0.4213067889213562, + "learning_rate": 3.430525873739363e-05, + "loss": 0.4983, + "num_input_tokens_seen": 23041008, + "step": 15145 + }, + { + "epoch": 0.48492414057998845, + "grad_norm": 1.1471304893493652, + "learning_rate": 3.429614594450743e-05, + "loss": 0.4811, + "num_input_tokens_seen": 23049248, + "step": 15150 + }, + { + "epoch": 0.48508418155047694, + "grad_norm": 1.367727279663086, + "learning_rate": 3.428703171803869e-05, + "loss": 0.6272, + "num_input_tokens_seen": 23057216, + "step": 15155 + }, + { + "epoch": 0.4852442225209654, + "grad_norm": 0.49910128116607666, + "learning_rate": 3.4277916059392964e-05, + "loss": 0.3397, + "num_input_tokens_seen": 23064880, + "step": 15160 + }, + { + "epoch": 0.4854042634914538, + "grad_norm": 0.9847099184989929, + "learning_rate": 3.426879896997598e-05, + "loss": 0.4871, + "num_input_tokens_seen": 23072832, + "step": 15165 + }, + { + "epoch": 0.48556430446194226, + "grad_norm": 0.7547823786735535, + "learning_rate": 3.425968045119372e-05, + "loss": 0.3793, + "num_input_tokens_seen": 23080352, + "step": 15170 + }, + { + "epoch": 0.4857243454324307, + "grad_norm": 0.46101701259613037, + "learning_rate": 3.425056050445237e-05, + "loss": 0.5399, + "num_input_tokens_seen": 23087984, + "step": 15175 + }, + { + "epoch": 0.48588438640291914, + "grad_norm": 0.7559738159179688, + "learning_rate": 3.4241439131158336e-05, + "loss": 0.3986, + "num_input_tokens_seen": 23095760, + "step": 15180 + }, + { + "epoch": 0.4860444273734076, + "grad_norm": 0.6268016695976257, + "learning_rate": 3.423231633271825e-05, + "loss": 0.4684, + "num_input_tokens_seen": 23103184, + "step": 15185 + }, + { + "epoch": 0.486204468343896, + "grad_norm": 0.5946706533432007, + "learning_rate": 3.4223192110538985e-05, + "loss": 0.3797, + "num_input_tokens_seen": 23110336, + "step": 15190 + }, + { + "epoch": 0.48636450931438446, + "grad_norm": 1.241722822189331, + "learning_rate": 3.4214066466027575e-05, + "loss": 0.4759, + "num_input_tokens_seen": 23117632, + "step": 15195 + }, + { + "epoch": 0.4865245502848729, + "grad_norm": 0.39216890931129456, + "learning_rate": 3.4204939400591325e-05, + "loss": 0.5401, + "num_input_tokens_seen": 23124960, + "step": 15200 + }, + { + "epoch": 0.4865245502848729, + "eval_loss": 0.5316406488418579, + "eval_runtime": 561.3967, + "eval_samples_per_second": 24.735, + "eval_steps_per_second": 12.367, + "num_input_tokens_seen": 23124960, + "step": 15200 + }, + { + "epoch": 0.4866845912553614, + "grad_norm": 0.5215521454811096, + "learning_rate": 3.419581091563775e-05, + "loss": 0.5191, + "num_input_tokens_seen": 23132352, + "step": 15205 + }, + { + "epoch": 0.48684463222584984, + "grad_norm": 0.746721088886261, + "learning_rate": 3.418668101257456e-05, + "loss": 0.4117, + "num_input_tokens_seen": 23140272, + "step": 15210 + }, + { + "epoch": 0.4870046731963383, + "grad_norm": 0.5225956439971924, + "learning_rate": 3.417754969280971e-05, + "loss": 0.4961, + "num_input_tokens_seen": 23148048, + "step": 15215 + }, + { + "epoch": 0.4871647141668267, + "grad_norm": 1.5652813911437988, + "learning_rate": 3.416841695775137e-05, + "loss": 0.5276, + "num_input_tokens_seen": 23155504, + "step": 15220 + }, + { + "epoch": 0.48732475513731516, + "grad_norm": 1.4467716217041016, + "learning_rate": 3.415928280880792e-05, + "loss": 0.4621, + "num_input_tokens_seen": 23162976, + "step": 15225 + }, + { + "epoch": 0.4874847961078036, + "grad_norm": 0.9513192772865295, + "learning_rate": 3.4150147247387965e-05, + "loss": 0.5344, + "num_input_tokens_seen": 23170224, + "step": 15230 + }, + { + "epoch": 0.48764483707829204, + "grad_norm": 0.556121289730072, + "learning_rate": 3.4141010274900306e-05, + "loss": 0.3721, + "num_input_tokens_seen": 23177664, + "step": 15235 + }, + { + "epoch": 0.4878048780487805, + "grad_norm": 1.5378021001815796, + "learning_rate": 3.413187189275399e-05, + "loss": 0.5495, + "num_input_tokens_seen": 23185088, + "step": 15240 + }, + { + "epoch": 0.4879649190192689, + "grad_norm": 0.538789689540863, + "learning_rate": 3.4122732102358265e-05, + "loss": 0.4096, + "num_input_tokens_seen": 23192592, + "step": 15245 + }, + { + "epoch": 0.48812495998975736, + "grad_norm": 0.7818472981452942, + "learning_rate": 3.411359090512261e-05, + "loss": 0.5347, + "num_input_tokens_seen": 23199952, + "step": 15250 + }, + { + "epoch": 0.4882850009602458, + "grad_norm": 0.8325831890106201, + "learning_rate": 3.410444830245672e-05, + "loss": 0.5822, + "num_input_tokens_seen": 23207472, + "step": 15255 + }, + { + "epoch": 0.4884450419307343, + "grad_norm": 0.7739691138267517, + "learning_rate": 3.409530429577048e-05, + "loss": 0.6261, + "num_input_tokens_seen": 23215024, + "step": 15260 + }, + { + "epoch": 0.48860508290122273, + "grad_norm": 0.6747664213180542, + "learning_rate": 3.408615888647402e-05, + "loss": 0.4552, + "num_input_tokens_seen": 23222368, + "step": 15265 + }, + { + "epoch": 0.48876512387171117, + "grad_norm": 0.4834403991699219, + "learning_rate": 3.4077012075977675e-05, + "loss": 0.4031, + "num_input_tokens_seen": 23230032, + "step": 15270 + }, + { + "epoch": 0.4889251648421996, + "grad_norm": 0.9065139889717102, + "learning_rate": 3.4067863865692e-05, + "loss": 0.5907, + "num_input_tokens_seen": 23237648, + "step": 15275 + }, + { + "epoch": 0.48908520581268805, + "grad_norm": 1.025230050086975, + "learning_rate": 3.4058714257027755e-05, + "loss": 0.6213, + "num_input_tokens_seen": 23245472, + "step": 15280 + }, + { + "epoch": 0.4892452467831765, + "grad_norm": 0.8450905680656433, + "learning_rate": 3.404956325139594e-05, + "loss": 0.6061, + "num_input_tokens_seen": 23252944, + "step": 15285 + }, + { + "epoch": 0.48940528775366493, + "grad_norm": 0.8850245475769043, + "learning_rate": 3.404041085020775e-05, + "loss": 0.6782, + "num_input_tokens_seen": 23260576, + "step": 15290 + }, + { + "epoch": 0.48956532872415337, + "grad_norm": 1.3543845415115356, + "learning_rate": 3.403125705487459e-05, + "loss": 0.6027, + "num_input_tokens_seen": 23267968, + "step": 15295 + }, + { + "epoch": 0.4897253696946418, + "grad_norm": 0.9706786274909973, + "learning_rate": 3.402210186680811e-05, + "loss": 0.4641, + "num_input_tokens_seen": 23275968, + "step": 15300 + }, + { + "epoch": 0.48988541066513025, + "grad_norm": 0.5821167826652527, + "learning_rate": 3.4012945287420137e-05, + "loss": 0.5868, + "num_input_tokens_seen": 23283616, + "step": 15305 + }, + { + "epoch": 0.49004545163561875, + "grad_norm": 0.895984411239624, + "learning_rate": 3.400378731812274e-05, + "loss": 0.5526, + "num_input_tokens_seen": 23292304, + "step": 15310 + }, + { + "epoch": 0.4902054926061072, + "grad_norm": 1.3244333267211914, + "learning_rate": 3.399462796032817e-05, + "loss": 0.5351, + "num_input_tokens_seen": 23299504, + "step": 15315 + }, + { + "epoch": 0.4903655335765956, + "grad_norm": 0.8505727648735046, + "learning_rate": 3.3985467215448954e-05, + "loss": 0.5876, + "num_input_tokens_seen": 23307040, + "step": 15320 + }, + { + "epoch": 0.49052557454708406, + "grad_norm": 1.818678855895996, + "learning_rate": 3.3976305084897776e-05, + "loss": 0.4875, + "num_input_tokens_seen": 23314112, + "step": 15325 + }, + { + "epoch": 0.4906856155175725, + "grad_norm": 0.7704419493675232, + "learning_rate": 3.3967141570087544e-05, + "loss": 0.3957, + "num_input_tokens_seen": 23321648, + "step": 15330 + }, + { + "epoch": 0.49084565648806094, + "grad_norm": 0.6472753882408142, + "learning_rate": 3.39579766724314e-05, + "loss": 0.4757, + "num_input_tokens_seen": 23328976, + "step": 15335 + }, + { + "epoch": 0.4910056974585494, + "grad_norm": 1.353062391281128, + "learning_rate": 3.3948810393342677e-05, + "loss": 0.5669, + "num_input_tokens_seen": 23336416, + "step": 15340 + }, + { + "epoch": 0.4911657384290378, + "grad_norm": 0.8625170588493347, + "learning_rate": 3.3939642734234936e-05, + "loss": 0.4628, + "num_input_tokens_seen": 23344496, + "step": 15345 + }, + { + "epoch": 0.49132577939952626, + "grad_norm": 0.7673386335372925, + "learning_rate": 3.393047369652194e-05, + "loss": 0.4799, + "num_input_tokens_seen": 23352512, + "step": 15350 + }, + { + "epoch": 0.4914858203700147, + "grad_norm": 0.4428482949733734, + "learning_rate": 3.3921303281617664e-05, + "loss": 0.398, + "num_input_tokens_seen": 23360064, + "step": 15355 + }, + { + "epoch": 0.49164586134050314, + "grad_norm": 0.7637485861778259, + "learning_rate": 3.391213149093632e-05, + "loss": 0.4951, + "num_input_tokens_seen": 23368016, + "step": 15360 + }, + { + "epoch": 0.49180590231099164, + "grad_norm": 0.8828474283218384, + "learning_rate": 3.3902958325892303e-05, + "loss": 0.5516, + "num_input_tokens_seen": 23375632, + "step": 15365 + }, + { + "epoch": 0.4919659432814801, + "grad_norm": 0.9710733890533447, + "learning_rate": 3.389378378790023e-05, + "loss": 0.6753, + "num_input_tokens_seen": 23383840, + "step": 15370 + }, + { + "epoch": 0.4921259842519685, + "grad_norm": 1.0193946361541748, + "learning_rate": 3.388460787837493e-05, + "loss": 0.6013, + "num_input_tokens_seen": 23391264, + "step": 15375 + }, + { + "epoch": 0.49228602522245696, + "grad_norm": 0.8761424422264099, + "learning_rate": 3.387543059873145e-05, + "loss": 0.3355, + "num_input_tokens_seen": 23399216, + "step": 15380 + }, + { + "epoch": 0.4924460661929454, + "grad_norm": 0.8955491185188293, + "learning_rate": 3.386625195038503e-05, + "loss": 0.5107, + "num_input_tokens_seen": 23407264, + "step": 15385 + }, + { + "epoch": 0.49260610716343384, + "grad_norm": 1.067106008529663, + "learning_rate": 3.3857071934751136e-05, + "loss": 0.5668, + "num_input_tokens_seen": 23414480, + "step": 15390 + }, + { + "epoch": 0.4927661481339223, + "grad_norm": 1.020918846130371, + "learning_rate": 3.384789055324544e-05, + "loss": 0.4355, + "num_input_tokens_seen": 23421728, + "step": 15395 + }, + { + "epoch": 0.4929261891044107, + "grad_norm": 0.5651504993438721, + "learning_rate": 3.3838707807283843e-05, + "loss": 0.507, + "num_input_tokens_seen": 23428832, + "step": 15400 + }, + { + "epoch": 0.4929261891044107, + "eval_loss": 0.5298699140548706, + "eval_runtime": 561.1614, + "eval_samples_per_second": 24.745, + "eval_steps_per_second": 12.373, + "num_input_tokens_seen": 23428832, + "step": 15400 + }, + { + "epoch": 0.49308623007489916, + "grad_norm": 0.7762631773948669, + "learning_rate": 3.382952369828243e-05, + "loss": 0.3733, + "num_input_tokens_seen": 23436560, + "step": 15405 + }, + { + "epoch": 0.4932462710453876, + "grad_norm": 1.5437211990356445, + "learning_rate": 3.38203382276575e-05, + "loss": 0.502, + "num_input_tokens_seen": 23443728, + "step": 15410 + }, + { + "epoch": 0.4934063120158761, + "grad_norm": 0.7481032609939575, + "learning_rate": 3.381115139682557e-05, + "loss": 0.4484, + "num_input_tokens_seen": 23451936, + "step": 15415 + }, + { + "epoch": 0.49356635298636453, + "grad_norm": 0.6675187945365906, + "learning_rate": 3.3801963207203366e-05, + "loss": 0.6956, + "num_input_tokens_seen": 23459328, + "step": 15420 + }, + { + "epoch": 0.49372639395685297, + "grad_norm": 1.042624592781067, + "learning_rate": 3.379277366020782e-05, + "loss": 0.7374, + "num_input_tokens_seen": 23467600, + "step": 15425 + }, + { + "epoch": 0.4938864349273414, + "grad_norm": 1.074240803718567, + "learning_rate": 3.3783582757256085e-05, + "loss": 0.612, + "num_input_tokens_seen": 23475120, + "step": 15430 + }, + { + "epoch": 0.49404647589782985, + "grad_norm": 0.6517713665962219, + "learning_rate": 3.3774390499765504e-05, + "loss": 0.6322, + "num_input_tokens_seen": 23482992, + "step": 15435 + }, + { + "epoch": 0.4942065168683183, + "grad_norm": 0.6871440410614014, + "learning_rate": 3.376519688915364e-05, + "loss": 0.4539, + "num_input_tokens_seen": 23490304, + "step": 15440 + }, + { + "epoch": 0.49436655783880673, + "grad_norm": 1.0537896156311035, + "learning_rate": 3.3756001926838273e-05, + "loss": 0.5265, + "num_input_tokens_seen": 23497696, + "step": 15445 + }, + { + "epoch": 0.49452659880929517, + "grad_norm": 1.7204759120941162, + "learning_rate": 3.374680561423737e-05, + "loss": 0.5355, + "num_input_tokens_seen": 23504944, + "step": 15450 + }, + { + "epoch": 0.4946866397797836, + "grad_norm": 0.624626100063324, + "learning_rate": 3.373760795276912e-05, + "loss": 0.5144, + "num_input_tokens_seen": 23512432, + "step": 15455 + }, + { + "epoch": 0.49484668075027205, + "grad_norm": 0.8881427049636841, + "learning_rate": 3.372840894385192e-05, + "loss": 0.4404, + "num_input_tokens_seen": 23520128, + "step": 15460 + }, + { + "epoch": 0.4950067217207605, + "grad_norm": 2.3784215450286865, + "learning_rate": 3.3719208588904375e-05, + "loss": 0.6731, + "num_input_tokens_seen": 23527312, + "step": 15465 + }, + { + "epoch": 0.495166762691249, + "grad_norm": 0.4561968445777893, + "learning_rate": 3.371000688934529e-05, + "loss": 0.5918, + "num_input_tokens_seen": 23534944, + "step": 15470 + }, + { + "epoch": 0.4953268036617374, + "grad_norm": 1.4829413890838623, + "learning_rate": 3.370080384659369e-05, + "loss": 0.589, + "num_input_tokens_seen": 23542672, + "step": 15475 + }, + { + "epoch": 0.49548684463222586, + "grad_norm": 1.114821434020996, + "learning_rate": 3.36915994620688e-05, + "loss": 0.5121, + "num_input_tokens_seen": 23550352, + "step": 15480 + }, + { + "epoch": 0.4956468856027143, + "grad_norm": 0.3418019115924835, + "learning_rate": 3.3682393737190035e-05, + "loss": 0.5808, + "num_input_tokens_seen": 23558048, + "step": 15485 + }, + { + "epoch": 0.49580692657320274, + "grad_norm": 1.1091737747192383, + "learning_rate": 3.3673186673377054e-05, + "loss": 0.4831, + "num_input_tokens_seen": 23565472, + "step": 15490 + }, + { + "epoch": 0.4959669675436912, + "grad_norm": 1.3531497716903687, + "learning_rate": 3.366397827204969e-05, + "loss": 0.6451, + "num_input_tokens_seen": 23573504, + "step": 15495 + }, + { + "epoch": 0.4961270085141796, + "grad_norm": 0.6623345017433167, + "learning_rate": 3.3654768534628e-05, + "loss": 0.6353, + "num_input_tokens_seen": 23580656, + "step": 15500 + }, + { + "epoch": 0.49628704948466806, + "grad_norm": 0.8816599249839783, + "learning_rate": 3.3645557462532245e-05, + "loss": 0.4343, + "num_input_tokens_seen": 23587696, + "step": 15505 + }, + { + "epoch": 0.4964470904551565, + "grad_norm": 1.0086513757705688, + "learning_rate": 3.363634505718288e-05, + "loss": 0.6043, + "num_input_tokens_seen": 23595168, + "step": 15510 + }, + { + "epoch": 0.49660713142564494, + "grad_norm": 0.6683899760246277, + "learning_rate": 3.362713132000057e-05, + "loss": 0.4096, + "num_input_tokens_seen": 23602704, + "step": 15515 + }, + { + "epoch": 0.4967671723961334, + "grad_norm": 0.38546136021614075, + "learning_rate": 3.36179162524062e-05, + "loss": 0.5477, + "num_input_tokens_seen": 23610816, + "step": 15520 + }, + { + "epoch": 0.4969272133666219, + "grad_norm": 0.47721201181411743, + "learning_rate": 3.3608699855820846e-05, + "loss": 0.4415, + "num_input_tokens_seen": 23618704, + "step": 15525 + }, + { + "epoch": 0.4970872543371103, + "grad_norm": 0.7592159509658813, + "learning_rate": 3.359948213166578e-05, + "loss": 0.6115, + "num_input_tokens_seen": 23626944, + "step": 15530 + }, + { + "epoch": 0.49724729530759876, + "grad_norm": 0.9546903371810913, + "learning_rate": 3.359026308136252e-05, + "loss": 0.4448, + "num_input_tokens_seen": 23634528, + "step": 15535 + }, + { + "epoch": 0.4974073362780872, + "grad_norm": 1.1531126499176025, + "learning_rate": 3.358104270633272e-05, + "loss": 0.5373, + "num_input_tokens_seen": 23642480, + "step": 15540 + }, + { + "epoch": 0.49756737724857564, + "grad_norm": 0.702346920967102, + "learning_rate": 3.357182100799831e-05, + "loss": 0.4719, + "num_input_tokens_seen": 23650016, + "step": 15545 + }, + { + "epoch": 0.4977274182190641, + "grad_norm": 0.37754538655281067, + "learning_rate": 3.3562597987781384e-05, + "loss": 0.4539, + "num_input_tokens_seen": 23657424, + "step": 15550 + }, + { + "epoch": 0.4978874591895525, + "grad_norm": 0.530649721622467, + "learning_rate": 3.355337364710424e-05, + "loss": 0.5252, + "num_input_tokens_seen": 23665328, + "step": 15555 + }, + { + "epoch": 0.49804750016004096, + "grad_norm": 0.9262380003929138, + "learning_rate": 3.354414798738939e-05, + "loss": 0.4982, + "num_input_tokens_seen": 23673152, + "step": 15560 + }, + { + "epoch": 0.4982075411305294, + "grad_norm": 0.8276793956756592, + "learning_rate": 3.353492101005955e-05, + "loss": 0.569, + "num_input_tokens_seen": 23680640, + "step": 15565 + }, + { + "epoch": 0.49836758210101784, + "grad_norm": 0.6836730241775513, + "learning_rate": 3.352569271653763e-05, + "loss": 0.5288, + "num_input_tokens_seen": 23688384, + "step": 15570 + }, + { + "epoch": 0.49852762307150633, + "grad_norm": 0.7871518731117249, + "learning_rate": 3.351646310824675e-05, + "loss": 0.6665, + "num_input_tokens_seen": 23695856, + "step": 15575 + }, + { + "epoch": 0.49868766404199477, + "grad_norm": 0.6566817760467529, + "learning_rate": 3.350723218661023e-05, + "loss": 0.4782, + "num_input_tokens_seen": 23703872, + "step": 15580 + }, + { + "epoch": 0.4988477050124832, + "grad_norm": 0.684680163860321, + "learning_rate": 3.349799995305162e-05, + "loss": 0.5696, + "num_input_tokens_seen": 23711728, + "step": 15585 + }, + { + "epoch": 0.49900774598297165, + "grad_norm": 0.9654720425605774, + "learning_rate": 3.348876640899461e-05, + "loss": 0.5658, + "num_input_tokens_seen": 23719360, + "step": 15590 + }, + { + "epoch": 0.4991677869534601, + "grad_norm": 0.43445703387260437, + "learning_rate": 3.3479531555863144e-05, + "loss": 0.3906, + "num_input_tokens_seen": 23727088, + "step": 15595 + }, + { + "epoch": 0.49932782792394853, + "grad_norm": 0.5513219237327576, + "learning_rate": 3.3470295395081344e-05, + "loss": 0.4591, + "num_input_tokens_seen": 23734320, + "step": 15600 + }, + { + "epoch": 0.49932782792394853, + "eval_loss": 0.529191792011261, + "eval_runtime": 561.5086, + "eval_samples_per_second": 24.73, + "eval_steps_per_second": 12.365, + "num_input_tokens_seen": 23734320, + "step": 15600 + }, + { + "epoch": 0.49948786889443697, + "grad_norm": 0.713694155216217, + "learning_rate": 3.3461057928073556e-05, + "loss": 0.4982, + "num_input_tokens_seen": 23742240, + "step": 15605 + }, + { + "epoch": 0.4996479098649254, + "grad_norm": 1.0603212118148804, + "learning_rate": 3.345181915626431e-05, + "loss": 0.5437, + "num_input_tokens_seen": 23750080, + "step": 15610 + }, + { + "epoch": 0.49980795083541385, + "grad_norm": 1.3562030792236328, + "learning_rate": 3.344257908107834e-05, + "loss": 0.5338, + "num_input_tokens_seen": 23757584, + "step": 15615 + }, + { + "epoch": 0.4999679918059023, + "grad_norm": 0.6467585563659668, + "learning_rate": 3.343333770394058e-05, + "loss": 0.4215, + "num_input_tokens_seen": 23764928, + "step": 15620 + }, + { + "epoch": 0.5001280327763907, + "grad_norm": 0.9843539595603943, + "learning_rate": 3.342409502627616e-05, + "loss": 0.5224, + "num_input_tokens_seen": 23772208, + "step": 15625 + }, + { + "epoch": 0.5002880737468792, + "grad_norm": 1.0263110399246216, + "learning_rate": 3.341485104951043e-05, + "loss": 0.8271, + "num_input_tokens_seen": 23780080, + "step": 15630 + }, + { + "epoch": 0.5004481147173676, + "grad_norm": 0.8653048872947693, + "learning_rate": 3.340560577506892e-05, + "loss": 0.5738, + "num_input_tokens_seen": 23787008, + "step": 15635 + }, + { + "epoch": 0.5006081556878561, + "grad_norm": 0.6694958209991455, + "learning_rate": 3.339635920437735e-05, + "loss": 0.5757, + "num_input_tokens_seen": 23794464, + "step": 15640 + }, + { + "epoch": 0.5007681966583445, + "grad_norm": 0.6924029588699341, + "learning_rate": 3.338711133886169e-05, + "loss": 0.4512, + "num_input_tokens_seen": 23802112, + "step": 15645 + }, + { + "epoch": 0.500928237628833, + "grad_norm": 1.6908196210861206, + "learning_rate": 3.3377862179948064e-05, + "loss": 0.5927, + "num_input_tokens_seen": 23809584, + "step": 15650 + }, + { + "epoch": 0.5010882785993215, + "grad_norm": 0.7920040488243103, + "learning_rate": 3.336861172906281e-05, + "loss": 0.5269, + "num_input_tokens_seen": 23817040, + "step": 15655 + }, + { + "epoch": 0.5012483195698099, + "grad_norm": 0.5222511291503906, + "learning_rate": 3.335935998763245e-05, + "loss": 0.5256, + "num_input_tokens_seen": 23824864, + "step": 15660 + }, + { + "epoch": 0.5014083605402984, + "grad_norm": 0.665894627571106, + "learning_rate": 3.3350106957083744e-05, + "loss": 0.4818, + "num_input_tokens_seen": 23832608, + "step": 15665 + }, + { + "epoch": 0.5015684015107867, + "grad_norm": 1.2761954069137573, + "learning_rate": 3.33408526388436e-05, + "loss": 0.6057, + "num_input_tokens_seen": 23840112, + "step": 15670 + }, + { + "epoch": 0.5017284424812752, + "grad_norm": 0.7393139600753784, + "learning_rate": 3.3331597034339166e-05, + "loss": 0.5056, + "num_input_tokens_seen": 23847760, + "step": 15675 + }, + { + "epoch": 0.5018884834517636, + "grad_norm": 1.3130526542663574, + "learning_rate": 3.3322340144997764e-05, + "loss": 0.6035, + "num_input_tokens_seen": 23855120, + "step": 15680 + }, + { + "epoch": 0.5020485244222521, + "grad_norm": 0.8638205528259277, + "learning_rate": 3.331308197224693e-05, + "loss": 0.477, + "num_input_tokens_seen": 23863440, + "step": 15685 + }, + { + "epoch": 0.5022085653927405, + "grad_norm": 1.2630351781845093, + "learning_rate": 3.330382251751438e-05, + "loss": 0.5194, + "num_input_tokens_seen": 23871072, + "step": 15690 + }, + { + "epoch": 0.502368606363229, + "grad_norm": 0.9046270847320557, + "learning_rate": 3.3294561782228054e-05, + "loss": 0.4866, + "num_input_tokens_seen": 23878608, + "step": 15695 + }, + { + "epoch": 0.5025286473337174, + "grad_norm": 0.6973585486412048, + "learning_rate": 3.328529976781607e-05, + "loss": 0.4797, + "num_input_tokens_seen": 23886848, + "step": 15700 + }, + { + "epoch": 0.5026886883042059, + "grad_norm": 0.6037237048149109, + "learning_rate": 3.327603647570673e-05, + "loss": 0.4972, + "num_input_tokens_seen": 23894512, + "step": 15705 + }, + { + "epoch": 0.5028487292746944, + "grad_norm": 1.123068928718567, + "learning_rate": 3.326677190732857e-05, + "loss": 0.3753, + "num_input_tokens_seen": 23902224, + "step": 15710 + }, + { + "epoch": 0.5030087702451828, + "grad_norm": 0.4679146409034729, + "learning_rate": 3.325750606411029e-05, + "loss": 0.3981, + "num_input_tokens_seen": 23910304, + "step": 15715 + }, + { + "epoch": 0.5031688112156713, + "grad_norm": 0.7260696291923523, + "learning_rate": 3.3248238947480804e-05, + "loss": 0.4799, + "num_input_tokens_seen": 23917840, + "step": 15720 + }, + { + "epoch": 0.5033288521861596, + "grad_norm": 0.5507386922836304, + "learning_rate": 3.323897055886922e-05, + "loss": 0.5113, + "num_input_tokens_seen": 23925440, + "step": 15725 + }, + { + "epoch": 0.5034888931566481, + "grad_norm": 0.9637209177017212, + "learning_rate": 3.322970089970484e-05, + "loss": 0.4273, + "num_input_tokens_seen": 23932976, + "step": 15730 + }, + { + "epoch": 0.5036489341271365, + "grad_norm": 0.7828580141067505, + "learning_rate": 3.3220429971417165e-05, + "loss": 0.5584, + "num_input_tokens_seen": 23940544, + "step": 15735 + }, + { + "epoch": 0.503808975097625, + "grad_norm": 0.44130080938339233, + "learning_rate": 3.321115777543588e-05, + "loss": 0.3669, + "num_input_tokens_seen": 23947744, + "step": 15740 + }, + { + "epoch": 0.5039690160681134, + "grad_norm": 1.3999075889587402, + "learning_rate": 3.320188431319088e-05, + "loss": 0.6691, + "num_input_tokens_seen": 23955792, + "step": 15745 + }, + { + "epoch": 0.5041290570386019, + "grad_norm": 0.5805653929710388, + "learning_rate": 3.319260958611224e-05, + "loss": 0.5499, + "num_input_tokens_seen": 23962896, + "step": 15750 + }, + { + "epoch": 0.5042890980090903, + "grad_norm": 1.3490875959396362, + "learning_rate": 3.3183333595630256e-05, + "loss": 0.4884, + "num_input_tokens_seen": 23970240, + "step": 15755 + }, + { + "epoch": 0.5044491389795788, + "grad_norm": 1.05503511428833, + "learning_rate": 3.317405634317538e-05, + "loss": 0.529, + "num_input_tokens_seen": 23977776, + "step": 15760 + }, + { + "epoch": 0.5046091799500673, + "grad_norm": 0.343244731426239, + "learning_rate": 3.3164777830178315e-05, + "loss": 0.4363, + "num_input_tokens_seen": 23984832, + "step": 15765 + }, + { + "epoch": 0.5047692209205557, + "grad_norm": 0.8185353875160217, + "learning_rate": 3.315549805806989e-05, + "loss": 0.6345, + "num_input_tokens_seen": 23992048, + "step": 15770 + }, + { + "epoch": 0.5049292618910441, + "grad_norm": 0.9687768816947937, + "learning_rate": 3.314621702828118e-05, + "loss": 0.4844, + "num_input_tokens_seen": 23999568, + "step": 15775 + }, + { + "epoch": 0.5050893028615325, + "grad_norm": 0.602663516998291, + "learning_rate": 3.313693474224342e-05, + "loss": 0.4013, + "num_input_tokens_seen": 24007312, + "step": 15780 + }, + { + "epoch": 0.505249343832021, + "grad_norm": 0.906125545501709, + "learning_rate": 3.312765120138809e-05, + "loss": 0.4441, + "num_input_tokens_seen": 24015152, + "step": 15785 + }, + { + "epoch": 0.5054093848025094, + "grad_norm": 0.6404649615287781, + "learning_rate": 3.311836640714679e-05, + "loss": 0.8163, + "num_input_tokens_seen": 24022512, + "step": 15790 + }, + { + "epoch": 0.5055694257729979, + "grad_norm": 0.4490694999694824, + "learning_rate": 3.310908036095137e-05, + "loss": 0.4788, + "num_input_tokens_seen": 24030112, + "step": 15795 + }, + { + "epoch": 0.5057294667434863, + "grad_norm": 1.1948635578155518, + "learning_rate": 3.309979306423386e-05, + "loss": 0.6651, + "num_input_tokens_seen": 24037968, + "step": 15800 + }, + { + "epoch": 0.5057294667434863, + "eval_loss": 0.5285232663154602, + "eval_runtime": 561.2541, + "eval_samples_per_second": 24.741, + "eval_steps_per_second": 12.371, + "num_input_tokens_seen": 24037968, + "step": 15800 + }, + { + "epoch": 0.5058895077139748, + "grad_norm": 0.30594250559806824, + "learning_rate": 3.309050451842647e-05, + "loss": 0.3581, + "num_input_tokens_seen": 24045824, + "step": 15805 + }, + { + "epoch": 0.5060495486844632, + "grad_norm": 1.1319363117218018, + "learning_rate": 3.3081214724961604e-05, + "loss": 0.4304, + "num_input_tokens_seen": 24053472, + "step": 15810 + }, + { + "epoch": 0.5062095896549517, + "grad_norm": 0.40031716227531433, + "learning_rate": 3.307192368527188e-05, + "loss": 0.3708, + "num_input_tokens_seen": 24061440, + "step": 15815 + }, + { + "epoch": 0.5063696306254402, + "grad_norm": 0.4398922920227051, + "learning_rate": 3.306263140079008e-05, + "loss": 0.2877, + "num_input_tokens_seen": 24068800, + "step": 15820 + }, + { + "epoch": 0.5065296715959285, + "grad_norm": 0.8155357241630554, + "learning_rate": 3.30533378729492e-05, + "loss": 0.39, + "num_input_tokens_seen": 24076736, + "step": 15825 + }, + { + "epoch": 0.506689712566417, + "grad_norm": 0.8976965546607971, + "learning_rate": 3.304404310318242e-05, + "loss": 0.4779, + "num_input_tokens_seen": 24083744, + "step": 15830 + }, + { + "epoch": 0.5068497535369054, + "grad_norm": 0.41253647208213806, + "learning_rate": 3.3034747092923105e-05, + "loss": 0.3239, + "num_input_tokens_seen": 24091664, + "step": 15835 + }, + { + "epoch": 0.5070097945073939, + "grad_norm": 0.8407290577888489, + "learning_rate": 3.3025449843604806e-05, + "loss": 0.4532, + "num_input_tokens_seen": 24099216, + "step": 15840 + }, + { + "epoch": 0.5071698354778823, + "grad_norm": 1.0709810256958008, + "learning_rate": 3.30161513566613e-05, + "loss": 0.4473, + "num_input_tokens_seen": 24107344, + "step": 15845 + }, + { + "epoch": 0.5073298764483708, + "grad_norm": 0.6019288301467896, + "learning_rate": 3.3006851633526506e-05, + "loss": 0.4835, + "num_input_tokens_seen": 24115024, + "step": 15850 + }, + { + "epoch": 0.5074899174188592, + "grad_norm": 0.7992231249809265, + "learning_rate": 3.2997550675634584e-05, + "loss": 0.5249, + "num_input_tokens_seen": 24122544, + "step": 15855 + }, + { + "epoch": 0.5076499583893477, + "grad_norm": 0.9489024877548218, + "learning_rate": 3.2988248484419825e-05, + "loss": 0.4562, + "num_input_tokens_seen": 24130160, + "step": 15860 + }, + { + "epoch": 0.5078099993598362, + "grad_norm": 0.8065629601478577, + "learning_rate": 3.2978945061316776e-05, + "loss": 0.5422, + "num_input_tokens_seen": 24137296, + "step": 15865 + }, + { + "epoch": 0.5079700403303246, + "grad_norm": 0.6113823652267456, + "learning_rate": 3.296964040776013e-05, + "loss": 0.5579, + "num_input_tokens_seen": 24144352, + "step": 15870 + }, + { + "epoch": 0.508130081300813, + "grad_norm": 0.7285082340240479, + "learning_rate": 3.296033452518478e-05, + "loss": 0.44, + "num_input_tokens_seen": 24151696, + "step": 15875 + }, + { + "epoch": 0.5082901222713014, + "grad_norm": 0.9217434525489807, + "learning_rate": 3.2951027415025806e-05, + "loss": 0.4828, + "num_input_tokens_seen": 24159328, + "step": 15880 + }, + { + "epoch": 0.5084501632417899, + "grad_norm": 0.8478432297706604, + "learning_rate": 3.294171907871849e-05, + "loss": 0.5053, + "num_input_tokens_seen": 24166848, + "step": 15885 + }, + { + "epoch": 0.5086102042122783, + "grad_norm": 0.6077796220779419, + "learning_rate": 3.293240951769828e-05, + "loss": 0.5021, + "num_input_tokens_seen": 24174432, + "step": 15890 + }, + { + "epoch": 0.5087702451827668, + "grad_norm": 0.7759031653404236, + "learning_rate": 3.2923098733400846e-05, + "loss": 0.5759, + "num_input_tokens_seen": 24182736, + "step": 15895 + }, + { + "epoch": 0.5089302861532552, + "grad_norm": 0.39249372482299805, + "learning_rate": 3.291378672726202e-05, + "loss": 0.4256, + "num_input_tokens_seen": 24190112, + "step": 15900 + }, + { + "epoch": 0.5090903271237437, + "grad_norm": 0.5462753176689148, + "learning_rate": 3.2904473500717824e-05, + "loss": 0.6599, + "num_input_tokens_seen": 24198128, + "step": 15905 + }, + { + "epoch": 0.5092503680942321, + "grad_norm": 0.7785064578056335, + "learning_rate": 3.289515905520449e-05, + "loss": 0.4594, + "num_input_tokens_seen": 24205584, + "step": 15910 + }, + { + "epoch": 0.5094104090647206, + "grad_norm": 1.5017439126968384, + "learning_rate": 3.288584339215841e-05, + "loss": 0.5987, + "num_input_tokens_seen": 24213024, + "step": 15915 + }, + { + "epoch": 0.5095704500352091, + "grad_norm": 0.5025779604911804, + "learning_rate": 3.287652651301617e-05, + "loss": 0.4501, + "num_input_tokens_seen": 24220624, + "step": 15920 + }, + { + "epoch": 0.5097304910056975, + "grad_norm": 0.7989525198936462, + "learning_rate": 3.286720841921457e-05, + "loss": 0.3606, + "num_input_tokens_seen": 24227968, + "step": 15925 + }, + { + "epoch": 0.509890531976186, + "grad_norm": 1.2629679441452026, + "learning_rate": 3.285788911219056e-05, + "loss": 0.5731, + "num_input_tokens_seen": 24236032, + "step": 15930 + }, + { + "epoch": 0.5100505729466743, + "grad_norm": 0.7447479963302612, + "learning_rate": 3.284856859338131e-05, + "loss": 0.6004, + "num_input_tokens_seen": 24243856, + "step": 15935 + }, + { + "epoch": 0.5102106139171628, + "grad_norm": 0.4816516637802124, + "learning_rate": 3.283924686422414e-05, + "loss": 0.4807, + "num_input_tokens_seen": 24251536, + "step": 15940 + }, + { + "epoch": 0.5103706548876512, + "grad_norm": 1.0379526615142822, + "learning_rate": 3.282992392615659e-05, + "loss": 0.3945, + "num_input_tokens_seen": 24259440, + "step": 15945 + }, + { + "epoch": 0.5105306958581397, + "grad_norm": 1.164549708366394, + "learning_rate": 3.282059978061638e-05, + "loss": 0.696, + "num_input_tokens_seen": 24266912, + "step": 15950 + }, + { + "epoch": 0.5106907368286281, + "grad_norm": 0.5239722728729248, + "learning_rate": 3.28112744290414e-05, + "loss": 0.6279, + "num_input_tokens_seen": 24275392, + "step": 15955 + }, + { + "epoch": 0.5108507777991166, + "grad_norm": 1.4050588607788086, + "learning_rate": 3.280194787286974e-05, + "loss": 0.5908, + "num_input_tokens_seen": 24283392, + "step": 15960 + }, + { + "epoch": 0.511010818769605, + "grad_norm": 1.1252968311309814, + "learning_rate": 3.2792620113539674e-05, + "loss": 0.3828, + "num_input_tokens_seen": 24290720, + "step": 15965 + }, + { + "epoch": 0.5111708597400935, + "grad_norm": 0.4926842749118805, + "learning_rate": 3.278329115248966e-05, + "loss": 0.5027, + "num_input_tokens_seen": 24298656, + "step": 15970 + }, + { + "epoch": 0.511330900710582, + "grad_norm": 0.8869339227676392, + "learning_rate": 3.277396099115834e-05, + "loss": 0.4531, + "num_input_tokens_seen": 24306192, + "step": 15975 + }, + { + "epoch": 0.5114909416810703, + "grad_norm": 0.4450492262840271, + "learning_rate": 3.276462963098454e-05, + "loss": 0.4108, + "num_input_tokens_seen": 24313760, + "step": 15980 + }, + { + "epoch": 0.5116509826515588, + "grad_norm": 0.8358228206634521, + "learning_rate": 3.275529707340728e-05, + "loss": 0.5277, + "num_input_tokens_seen": 24321120, + "step": 15985 + }, + { + "epoch": 0.5118110236220472, + "grad_norm": 0.9734163880348206, + "learning_rate": 3.274596331986574e-05, + "loss": 0.486, + "num_input_tokens_seen": 24328320, + "step": 15990 + }, + { + "epoch": 0.5119710645925357, + "grad_norm": 0.8134451508522034, + "learning_rate": 3.273662837179932e-05, + "loss": 0.3749, + "num_input_tokens_seen": 24336128, + "step": 15995 + }, + { + "epoch": 0.5121311055630241, + "grad_norm": 0.5272929668426514, + "learning_rate": 3.272729223064758e-05, + "loss": 0.5211, + "num_input_tokens_seen": 24344064, + "step": 16000 + }, + { + "epoch": 0.5121311055630241, + "eval_loss": 0.5284156203269958, + "eval_runtime": 561.4873, + "eval_samples_per_second": 24.731, + "eval_steps_per_second": 12.365, + "num_input_tokens_seen": 24344064, + "step": 16000 + }, + { + "epoch": 0.5122911465335126, + "grad_norm": 0.8715248107910156, + "learning_rate": 3.2717954897850264e-05, + "loss": 0.4946, + "num_input_tokens_seen": 24351264, + "step": 16005 + }, + { + "epoch": 0.512451187504001, + "grad_norm": 0.5955188870429993, + "learning_rate": 3.270861637484733e-05, + "loss": 0.4168, + "num_input_tokens_seen": 24359024, + "step": 16010 + }, + { + "epoch": 0.5126112284744895, + "grad_norm": 0.5834535360336304, + "learning_rate": 3.2699276663078867e-05, + "loss": 0.3531, + "num_input_tokens_seen": 24367568, + "step": 16015 + }, + { + "epoch": 0.5127712694449779, + "grad_norm": 0.3775244951248169, + "learning_rate": 3.268993576398519e-05, + "loss": 0.4686, + "num_input_tokens_seen": 24375296, + "step": 16020 + }, + { + "epoch": 0.5129313104154664, + "grad_norm": 0.6535698175430298, + "learning_rate": 3.268059367900678e-05, + "loss": 0.5098, + "num_input_tokens_seen": 24382592, + "step": 16025 + }, + { + "epoch": 0.5130913513859549, + "grad_norm": 0.8663296699523926, + "learning_rate": 3.26712504095843e-05, + "loss": 0.5145, + "num_input_tokens_seen": 24389808, + "step": 16030 + }, + { + "epoch": 0.5132513923564432, + "grad_norm": 1.1189180612564087, + "learning_rate": 3.2661905957158615e-05, + "loss": 0.4982, + "num_input_tokens_seen": 24397040, + "step": 16035 + }, + { + "epoch": 0.5134114333269317, + "grad_norm": 1.7260717153549194, + "learning_rate": 3.2652560323170734e-05, + "loss": 0.5927, + "num_input_tokens_seen": 24404864, + "step": 16040 + }, + { + "epoch": 0.5135714742974201, + "grad_norm": 0.8285902142524719, + "learning_rate": 3.264321350906189e-05, + "loss": 0.5076, + "num_input_tokens_seen": 24412352, + "step": 16045 + }, + { + "epoch": 0.5137315152679086, + "grad_norm": 0.7690265774726868, + "learning_rate": 3.263386551627346e-05, + "loss": 0.5733, + "num_input_tokens_seen": 24419472, + "step": 16050 + }, + { + "epoch": 0.513891556238397, + "grad_norm": 1.1433550119400024, + "learning_rate": 3.2624516346247055e-05, + "loss": 0.6857, + "num_input_tokens_seen": 24426960, + "step": 16055 + }, + { + "epoch": 0.5140515972088855, + "grad_norm": 1.3578342199325562, + "learning_rate": 3.2615166000424404e-05, + "loss": 0.3947, + "num_input_tokens_seen": 24434912, + "step": 16060 + }, + { + "epoch": 0.5142116381793739, + "grad_norm": 0.9389097690582275, + "learning_rate": 3.260581448024745e-05, + "loss": 0.5012, + "num_input_tokens_seen": 24441920, + "step": 16065 + }, + { + "epoch": 0.5143716791498624, + "grad_norm": 0.6510707139968872, + "learning_rate": 3.2596461787158335e-05, + "loss": 0.5288, + "num_input_tokens_seen": 24449248, + "step": 16070 + }, + { + "epoch": 0.5145317201203508, + "grad_norm": 1.3173905611038208, + "learning_rate": 3.258710792259934e-05, + "loss": 0.5387, + "num_input_tokens_seen": 24457232, + "step": 16075 + }, + { + "epoch": 0.5146917610908393, + "grad_norm": 1.0494741201400757, + "learning_rate": 3.257775288801296e-05, + "loss": 0.5394, + "num_input_tokens_seen": 24465168, + "step": 16080 + }, + { + "epoch": 0.5148518020613277, + "grad_norm": 0.7352792620658875, + "learning_rate": 3.256839668484186e-05, + "loss": 0.4927, + "num_input_tokens_seen": 24472208, + "step": 16085 + }, + { + "epoch": 0.5150118430318161, + "grad_norm": 0.5958560705184937, + "learning_rate": 3.255903931452888e-05, + "loss": 0.4406, + "num_input_tokens_seen": 24479648, + "step": 16090 + }, + { + "epoch": 0.5151718840023046, + "grad_norm": 0.6318920850753784, + "learning_rate": 3.2549680778517045e-05, + "loss": 0.3262, + "num_input_tokens_seen": 24487472, + "step": 16095 + }, + { + "epoch": 0.515331924972793, + "grad_norm": 0.8057692646980286, + "learning_rate": 3.2540321078249556e-05, + "loss": 0.4773, + "num_input_tokens_seen": 24494592, + "step": 16100 + }, + { + "epoch": 0.5154919659432815, + "grad_norm": 1.050693154335022, + "learning_rate": 3.2530960215169795e-05, + "loss": 0.4715, + "num_input_tokens_seen": 24501936, + "step": 16105 + }, + { + "epoch": 0.5156520069137699, + "grad_norm": 0.9866595268249512, + "learning_rate": 3.2521598190721345e-05, + "loss": 0.5853, + "num_input_tokens_seen": 24510224, + "step": 16110 + }, + { + "epoch": 0.5158120478842584, + "grad_norm": 0.8758458495140076, + "learning_rate": 3.251223500634792e-05, + "loss": 0.5141, + "num_input_tokens_seen": 24518096, + "step": 16115 + }, + { + "epoch": 0.5159720888547468, + "grad_norm": 0.643241286277771, + "learning_rate": 3.2502870663493445e-05, + "loss": 0.4682, + "num_input_tokens_seen": 24525680, + "step": 16120 + }, + { + "epoch": 0.5161321298252353, + "grad_norm": 0.8690606355667114, + "learning_rate": 3.249350516360203e-05, + "loss": 0.4619, + "num_input_tokens_seen": 24533008, + "step": 16125 + }, + { + "epoch": 0.5162921707957238, + "grad_norm": 0.8514220714569092, + "learning_rate": 3.248413850811797e-05, + "loss": 0.5447, + "num_input_tokens_seen": 24540608, + "step": 16130 + }, + { + "epoch": 0.5164522117662121, + "grad_norm": 0.7314367890357971, + "learning_rate": 3.2474770698485677e-05, + "loss": 0.5304, + "num_input_tokens_seen": 24548272, + "step": 16135 + }, + { + "epoch": 0.5166122527367006, + "grad_norm": 0.6434447765350342, + "learning_rate": 3.246540173614983e-05, + "loss": 0.5362, + "num_input_tokens_seen": 24555536, + "step": 16140 + }, + { + "epoch": 0.516772293707189, + "grad_norm": 0.5522851347923279, + "learning_rate": 3.2456031622555197e-05, + "loss": 0.5122, + "num_input_tokens_seen": 24562960, + "step": 16145 + }, + { + "epoch": 0.5169323346776775, + "grad_norm": 1.3261579275131226, + "learning_rate": 3.2446660359146794e-05, + "loss": 0.6132, + "num_input_tokens_seen": 24570464, + "step": 16150 + }, + { + "epoch": 0.5170923756481659, + "grad_norm": 0.6127949357032776, + "learning_rate": 3.2437287947369786e-05, + "loss": 0.3929, + "num_input_tokens_seen": 24577952, + "step": 16155 + }, + { + "epoch": 0.5172524166186544, + "grad_norm": 0.8190617561340332, + "learning_rate": 3.2427914388669525e-05, + "loss": 0.5153, + "num_input_tokens_seen": 24585872, + "step": 16160 + }, + { + "epoch": 0.5174124575891428, + "grad_norm": 0.8416407108306885, + "learning_rate": 3.241853968449151e-05, + "loss": 0.4369, + "num_input_tokens_seen": 24594448, + "step": 16165 + }, + { + "epoch": 0.5175724985596313, + "grad_norm": 0.9695448279380798, + "learning_rate": 3.240916383628144e-05, + "loss": 0.5095, + "num_input_tokens_seen": 24602496, + "step": 16170 + }, + { + "epoch": 0.5177325395301197, + "grad_norm": 1.7063876390457153, + "learning_rate": 3.239978684548521e-05, + "loss": 0.4878, + "num_input_tokens_seen": 24609984, + "step": 16175 + }, + { + "epoch": 0.5178925805006082, + "grad_norm": 1.0846927165985107, + "learning_rate": 3.239040871354885e-05, + "loss": 0.4956, + "num_input_tokens_seen": 24617088, + "step": 16180 + }, + { + "epoch": 0.5180526214710967, + "grad_norm": 0.728522777557373, + "learning_rate": 3.2381029441918596e-05, + "loss": 0.5521, + "num_input_tokens_seen": 24625376, + "step": 16185 + }, + { + "epoch": 0.518212662441585, + "grad_norm": 0.6137141585350037, + "learning_rate": 3.2371649032040845e-05, + "loss": 0.4828, + "num_input_tokens_seen": 24632816, + "step": 16190 + }, + { + "epoch": 0.5183727034120735, + "grad_norm": 0.6145040988922119, + "learning_rate": 3.2362267485362174e-05, + "loss": 0.4378, + "num_input_tokens_seen": 24640656, + "step": 16195 + }, + { + "epoch": 0.5185327443825619, + "grad_norm": 3.1921136379241943, + "learning_rate": 3.235288480332934e-05, + "loss": 0.4678, + "num_input_tokens_seen": 24647968, + "step": 16200 + }, + { + "epoch": 0.5185327443825619, + "eval_loss": 0.5273978114128113, + "eval_runtime": 561.3169, + "eval_samples_per_second": 24.738, + "eval_steps_per_second": 12.369, + "num_input_tokens_seen": 24647968, + "step": 16200 + }, + { + "epoch": 0.5186927853530504, + "grad_norm": 0.9834450483322144, + "learning_rate": 3.234350098738927e-05, + "loss": 0.4497, + "num_input_tokens_seen": 24655888, + "step": 16205 + }, + { + "epoch": 0.5188528263235388, + "grad_norm": 0.530600905418396, + "learning_rate": 3.233411603898906e-05, + "loss": 0.4932, + "num_input_tokens_seen": 24663360, + "step": 16210 + }, + { + "epoch": 0.5190128672940273, + "grad_norm": 0.7122696042060852, + "learning_rate": 3.232472995957599e-05, + "loss": 0.6343, + "num_input_tokens_seen": 24671104, + "step": 16215 + }, + { + "epoch": 0.5191729082645157, + "grad_norm": 0.7666733264923096, + "learning_rate": 3.231534275059751e-05, + "loss": 0.5248, + "num_input_tokens_seen": 24678816, + "step": 16220 + }, + { + "epoch": 0.5193329492350042, + "grad_norm": 1.2616404294967651, + "learning_rate": 3.230595441350125e-05, + "loss": 0.5687, + "num_input_tokens_seen": 24686656, + "step": 16225 + }, + { + "epoch": 0.5194929902054926, + "grad_norm": 0.86185622215271, + "learning_rate": 3.2296564949735e-05, + "loss": 0.4938, + "num_input_tokens_seen": 24694000, + "step": 16230 + }, + { + "epoch": 0.519653031175981, + "grad_norm": 1.1847591400146484, + "learning_rate": 3.228717436074675e-05, + "loss": 0.5493, + "num_input_tokens_seen": 24702048, + "step": 16235 + }, + { + "epoch": 0.5198130721464695, + "grad_norm": 0.5651764273643494, + "learning_rate": 3.227778264798463e-05, + "loss": 0.5571, + "num_input_tokens_seen": 24709696, + "step": 16240 + }, + { + "epoch": 0.5199731131169579, + "grad_norm": 0.8169847130775452, + "learning_rate": 3.226838981289698e-05, + "loss": 0.4222, + "num_input_tokens_seen": 24717408, + "step": 16245 + }, + { + "epoch": 0.5201331540874464, + "grad_norm": 0.8335725665092468, + "learning_rate": 3.225899585693227e-05, + "loss": 0.5059, + "num_input_tokens_seen": 24724848, + "step": 16250 + }, + { + "epoch": 0.5202931950579348, + "grad_norm": 0.5953288674354553, + "learning_rate": 3.224960078153918e-05, + "loss": 0.5895, + "num_input_tokens_seen": 24732368, + "step": 16255 + }, + { + "epoch": 0.5204532360284233, + "grad_norm": 0.36903905868530273, + "learning_rate": 3.224020458816655e-05, + "loss": 0.5384, + "num_input_tokens_seen": 24739888, + "step": 16260 + }, + { + "epoch": 0.5206132769989117, + "grad_norm": 1.336954951286316, + "learning_rate": 3.223080727826337e-05, + "loss": 0.5493, + "num_input_tokens_seen": 24746960, + "step": 16265 + }, + { + "epoch": 0.5207733179694002, + "grad_norm": 0.734536349773407, + "learning_rate": 3.222140885327885e-05, + "loss": 0.6983, + "num_input_tokens_seen": 24754784, + "step": 16270 + }, + { + "epoch": 0.5209333589398886, + "grad_norm": 0.4864906072616577, + "learning_rate": 3.221200931466234e-05, + "loss": 0.5462, + "num_input_tokens_seen": 24762816, + "step": 16275 + }, + { + "epoch": 0.5210933999103771, + "grad_norm": 1.971429705619812, + "learning_rate": 3.220260866386336e-05, + "loss": 0.7906, + "num_input_tokens_seen": 24770400, + "step": 16280 + }, + { + "epoch": 0.5212534408808654, + "grad_norm": 0.8119062185287476, + "learning_rate": 3.21932069023316e-05, + "loss": 0.5533, + "num_input_tokens_seen": 24777888, + "step": 16285 + }, + { + "epoch": 0.5214134818513539, + "grad_norm": 0.5041411519050598, + "learning_rate": 3.218380403151695e-05, + "loss": 0.4064, + "num_input_tokens_seen": 24785248, + "step": 16290 + }, + { + "epoch": 0.5215735228218424, + "grad_norm": 0.8472173810005188, + "learning_rate": 3.217440005286943e-05, + "loss": 0.4491, + "num_input_tokens_seen": 24793120, + "step": 16295 + }, + { + "epoch": 0.5217335637923308, + "grad_norm": 0.7151390910148621, + "learning_rate": 3.216499496783928e-05, + "loss": 0.4215, + "num_input_tokens_seen": 24801040, + "step": 16300 + }, + { + "epoch": 0.5218936047628193, + "grad_norm": 0.839257001876831, + "learning_rate": 3.2155588777876856e-05, + "loss": 0.4402, + "num_input_tokens_seen": 24808448, + "step": 16305 + }, + { + "epoch": 0.5220536457333077, + "grad_norm": 0.5085650086402893, + "learning_rate": 3.214618148443273e-05, + "loss": 0.622, + "num_input_tokens_seen": 24815888, + "step": 16310 + }, + { + "epoch": 0.5222136867037962, + "grad_norm": 0.756403923034668, + "learning_rate": 3.2136773088957595e-05, + "loss": 0.4385, + "num_input_tokens_seen": 24823248, + "step": 16315 + }, + { + "epoch": 0.5223737276742846, + "grad_norm": 0.6861956119537354, + "learning_rate": 3.2127363592902374e-05, + "loss": 0.6495, + "num_input_tokens_seen": 24831072, + "step": 16320 + }, + { + "epoch": 0.5225337686447731, + "grad_norm": 0.517824649810791, + "learning_rate": 3.211795299771812e-05, + "loss": 0.4926, + "num_input_tokens_seen": 24839552, + "step": 16325 + }, + { + "epoch": 0.5226938096152615, + "grad_norm": 0.7451983690261841, + "learning_rate": 3.210854130485605e-05, + "loss": 0.4322, + "num_input_tokens_seen": 24846848, + "step": 16330 + }, + { + "epoch": 0.52285385058575, + "grad_norm": 0.39841723442077637, + "learning_rate": 3.209912851576759e-05, + "loss": 0.4082, + "num_input_tokens_seen": 24854464, + "step": 16335 + }, + { + "epoch": 0.5230138915562383, + "grad_norm": 0.902009129524231, + "learning_rate": 3.208971463190431e-05, + "loss": 0.5706, + "num_input_tokens_seen": 24862016, + "step": 16340 + }, + { + "epoch": 0.5231739325267268, + "grad_norm": 1.0485527515411377, + "learning_rate": 3.208029965471793e-05, + "loss": 0.5053, + "num_input_tokens_seen": 24868912, + "step": 16345 + }, + { + "epoch": 0.5233339734972153, + "grad_norm": 1.4206064939498901, + "learning_rate": 3.2070883585660364e-05, + "loss": 0.6003, + "num_input_tokens_seen": 24876208, + "step": 16350 + }, + { + "epoch": 0.5234940144677037, + "grad_norm": 0.98345947265625, + "learning_rate": 3.20614664261837e-05, + "loss": 0.4323, + "num_input_tokens_seen": 24883360, + "step": 16355 + }, + { + "epoch": 0.5236540554381922, + "grad_norm": 1.1366468667984009, + "learning_rate": 3.205204817774016e-05, + "loss": 0.5025, + "num_input_tokens_seen": 24891056, + "step": 16360 + }, + { + "epoch": 0.5238140964086806, + "grad_norm": 0.8625043034553528, + "learning_rate": 3.204262884178218e-05, + "loss": 0.5308, + "num_input_tokens_seen": 24898224, + "step": 16365 + }, + { + "epoch": 0.5239741373791691, + "grad_norm": 0.7812259197235107, + "learning_rate": 3.2033208419762314e-05, + "loss": 0.5351, + "num_input_tokens_seen": 24905616, + "step": 16370 + }, + { + "epoch": 0.5241341783496575, + "grad_norm": 0.6644781231880188, + "learning_rate": 3.2023786913133344e-05, + "loss": 0.5978, + "num_input_tokens_seen": 24913888, + "step": 16375 + }, + { + "epoch": 0.524294219320146, + "grad_norm": 0.6647806763648987, + "learning_rate": 3.201436432334816e-05, + "loss": 0.5107, + "num_input_tokens_seen": 24921632, + "step": 16380 + }, + { + "epoch": 0.5244542602906344, + "grad_norm": 0.9681705236434937, + "learning_rate": 3.2004940651859844e-05, + "loss": 0.4439, + "num_input_tokens_seen": 24929840, + "step": 16385 + }, + { + "epoch": 0.5246143012611229, + "grad_norm": 1.2515941858291626, + "learning_rate": 3.1995515900121655e-05, + "loss": 0.3886, + "num_input_tokens_seen": 24937424, + "step": 16390 + }, + { + "epoch": 0.5247743422316113, + "grad_norm": 1.256091594696045, + "learning_rate": 3.1986090069587e-05, + "loss": 0.6466, + "num_input_tokens_seen": 24945168, + "step": 16395 + }, + { + "epoch": 0.5249343832020997, + "grad_norm": 1.9088337421417236, + "learning_rate": 3.1976663161709466e-05, + "loss": 0.4696, + "num_input_tokens_seen": 24952928, + "step": 16400 + }, + { + "epoch": 0.5249343832020997, + "eval_loss": 0.5261112451553345, + "eval_runtime": 561.4132, + "eval_samples_per_second": 24.734, + "eval_steps_per_second": 12.367, + "num_input_tokens_seen": 24952928, + "step": 16400 + }, + { + "epoch": 0.5250944241725882, + "grad_norm": 1.0632678270339966, + "learning_rate": 3.196723517794279e-05, + "loss": 0.5489, + "num_input_tokens_seen": 24960752, + "step": 16405 + }, + { + "epoch": 0.5252544651430766, + "grad_norm": 0.8600162267684937, + "learning_rate": 3.19578061197409e-05, + "loss": 0.418, + "num_input_tokens_seen": 24967824, + "step": 16410 + }, + { + "epoch": 0.5254145061135651, + "grad_norm": 0.7652732133865356, + "learning_rate": 3.194837598855787e-05, + "loss": 0.5338, + "num_input_tokens_seen": 24975600, + "step": 16415 + }, + { + "epoch": 0.5255745470840535, + "grad_norm": 1.3545476198196411, + "learning_rate": 3.193894478584794e-05, + "loss": 0.4696, + "num_input_tokens_seen": 24983136, + "step": 16420 + }, + { + "epoch": 0.525734588054542, + "grad_norm": 0.6946269273757935, + "learning_rate": 3.192951251306553e-05, + "loss": 0.4629, + "num_input_tokens_seen": 24991024, + "step": 16425 + }, + { + "epoch": 0.5258946290250304, + "grad_norm": 0.8633112907409668, + "learning_rate": 3.192007917166521e-05, + "loss": 0.5714, + "num_input_tokens_seen": 24998800, + "step": 16430 + }, + { + "epoch": 0.5260546699955189, + "grad_norm": 1.0283259153366089, + "learning_rate": 3.191064476310171e-05, + "loss": 0.7086, + "num_input_tokens_seen": 25006384, + "step": 16435 + }, + { + "epoch": 0.5262147109660072, + "grad_norm": 0.5429224967956543, + "learning_rate": 3.1901209288829944e-05, + "loss": 0.3327, + "num_input_tokens_seen": 25014064, + "step": 16440 + }, + { + "epoch": 0.5263747519364957, + "grad_norm": 0.5778437256813049, + "learning_rate": 3.1891772750304985e-05, + "loss": 0.5956, + "num_input_tokens_seen": 25021440, + "step": 16445 + }, + { + "epoch": 0.5265347929069842, + "grad_norm": 0.6459833979606628, + "learning_rate": 3.188233514898206e-05, + "loss": 0.4914, + "num_input_tokens_seen": 25029712, + "step": 16450 + }, + { + "epoch": 0.5266948338774726, + "grad_norm": 1.0101240873336792, + "learning_rate": 3.187289648631657e-05, + "loss": 0.4852, + "num_input_tokens_seen": 25036960, + "step": 16455 + }, + { + "epoch": 0.5268548748479611, + "grad_norm": 0.7913943529129028, + "learning_rate": 3.186345676376406e-05, + "loss": 0.3946, + "num_input_tokens_seen": 25044704, + "step": 16460 + }, + { + "epoch": 0.5270149158184495, + "grad_norm": 0.6162383556365967, + "learning_rate": 3.1854015982780275e-05, + "loss": 0.508, + "num_input_tokens_seen": 25052144, + "step": 16465 + }, + { + "epoch": 0.527174956788938, + "grad_norm": 0.5893405079841614, + "learning_rate": 3.1844574144821084e-05, + "loss": 0.3906, + "num_input_tokens_seen": 25059904, + "step": 16470 + }, + { + "epoch": 0.5273349977594264, + "grad_norm": 0.9897502064704895, + "learning_rate": 3.1835131251342554e-05, + "loss": 0.4524, + "num_input_tokens_seen": 25067552, + "step": 16475 + }, + { + "epoch": 0.5274950387299149, + "grad_norm": 0.6454185843467712, + "learning_rate": 3.182568730380089e-05, + "loss": 0.5166, + "num_input_tokens_seen": 25075552, + "step": 16480 + }, + { + "epoch": 0.5276550797004033, + "grad_norm": 0.44959887862205505, + "learning_rate": 3.181624230365245e-05, + "loss": 0.3682, + "num_input_tokens_seen": 25083472, + "step": 16485 + }, + { + "epoch": 0.5278151206708918, + "grad_norm": 0.9277119636535645, + "learning_rate": 3.180679625235381e-05, + "loss": 0.4844, + "num_input_tokens_seen": 25090800, + "step": 16490 + }, + { + "epoch": 0.5279751616413801, + "grad_norm": 0.6804785132408142, + "learning_rate": 3.1797349151361646e-05, + "loss": 0.4505, + "num_input_tokens_seen": 25098272, + "step": 16495 + }, + { + "epoch": 0.5281352026118686, + "grad_norm": 0.8007342219352722, + "learning_rate": 3.178790100213281e-05, + "loss": 0.4982, + "num_input_tokens_seen": 25106000, + "step": 16500 + }, + { + "epoch": 0.5282952435823571, + "grad_norm": 0.8672520518302917, + "learning_rate": 3.1778451806124346e-05, + "loss": 0.4665, + "num_input_tokens_seen": 25113744, + "step": 16505 + }, + { + "epoch": 0.5284552845528455, + "grad_norm": 1.185323715209961, + "learning_rate": 3.176900156479342e-05, + "loss": 0.4976, + "num_input_tokens_seen": 25121488, + "step": 16510 + }, + { + "epoch": 0.528615325523334, + "grad_norm": 0.6730177998542786, + "learning_rate": 3.17595502795974e-05, + "loss": 0.4306, + "num_input_tokens_seen": 25129456, + "step": 16515 + }, + { + "epoch": 0.5287753664938224, + "grad_norm": 0.4507526755332947, + "learning_rate": 3.175009795199377e-05, + "loss": 0.4809, + "num_input_tokens_seen": 25137264, + "step": 16520 + }, + { + "epoch": 0.5289354074643109, + "grad_norm": 0.5852676033973694, + "learning_rate": 3.1740644583440224e-05, + "loss": 0.419, + "num_input_tokens_seen": 25144464, + "step": 16525 + }, + { + "epoch": 0.5290954484347993, + "grad_norm": 0.7127730846405029, + "learning_rate": 3.173119017539457e-05, + "loss": 0.5524, + "num_input_tokens_seen": 25151312, + "step": 16530 + }, + { + "epoch": 0.5292554894052878, + "grad_norm": 0.5841956734657288, + "learning_rate": 3.172173472931479e-05, + "loss": 0.4228, + "num_input_tokens_seen": 25159088, + "step": 16535 + }, + { + "epoch": 0.5294155303757762, + "grad_norm": 0.9974806308746338, + "learning_rate": 3.1712278246659055e-05, + "loss": 0.5668, + "num_input_tokens_seen": 25167088, + "step": 16540 + }, + { + "epoch": 0.5295755713462647, + "grad_norm": 0.5831204652786255, + "learning_rate": 3.170282072888566e-05, + "loss": 0.4948, + "num_input_tokens_seen": 25175008, + "step": 16545 + }, + { + "epoch": 0.529735612316753, + "grad_norm": 0.5679012537002563, + "learning_rate": 3.169336217745307e-05, + "loss": 0.5392, + "num_input_tokens_seen": 25182480, + "step": 16550 + }, + { + "epoch": 0.5298956532872415, + "grad_norm": 0.7028695344924927, + "learning_rate": 3.1683902593819924e-05, + "loss": 0.4595, + "num_input_tokens_seen": 25190352, + "step": 16555 + }, + { + "epoch": 0.53005569425773, + "grad_norm": 0.4796917736530304, + "learning_rate": 3.1674441979445e-05, + "loss": 0.4097, + "num_input_tokens_seen": 25197872, + "step": 16560 + }, + { + "epoch": 0.5302157352282184, + "grad_norm": 0.4097890257835388, + "learning_rate": 3.166498033578725e-05, + "loss": 0.4944, + "num_input_tokens_seen": 25205792, + "step": 16565 + }, + { + "epoch": 0.5303757761987069, + "grad_norm": 1.562151312828064, + "learning_rate": 3.165551766430578e-05, + "loss": 0.5524, + "num_input_tokens_seen": 25212976, + "step": 16570 + }, + { + "epoch": 0.5305358171691953, + "grad_norm": 0.7587957382202148, + "learning_rate": 3.164605396645984e-05, + "loss": 0.5046, + "num_input_tokens_seen": 25220512, + "step": 16575 + }, + { + "epoch": 0.5306958581396838, + "grad_norm": 0.5703162550926208, + "learning_rate": 3.163658924370886e-05, + "loss": 0.5924, + "num_input_tokens_seen": 25227584, + "step": 16580 + }, + { + "epoch": 0.5308558991101722, + "grad_norm": 0.44803279638290405, + "learning_rate": 3.1627123497512415e-05, + "loss": 0.6081, + "num_input_tokens_seen": 25235344, + "step": 16585 + }, + { + "epoch": 0.5310159400806607, + "grad_norm": 0.5226919651031494, + "learning_rate": 3.1617656729330245e-05, + "loss": 0.4008, + "num_input_tokens_seen": 25242512, + "step": 16590 + }, + { + "epoch": 0.531175981051149, + "grad_norm": 0.9983401298522949, + "learning_rate": 3.1608188940622255e-05, + "loss": 0.6745, + "num_input_tokens_seen": 25250096, + "step": 16595 + }, + { + "epoch": 0.5313360220216375, + "grad_norm": 0.9192315936088562, + "learning_rate": 3.159872013284847e-05, + "loss": 0.4875, + "num_input_tokens_seen": 25257664, + "step": 16600 + }, + { + "epoch": 0.5313360220216375, + "eval_loss": 0.52545166015625, + "eval_runtime": 561.4183, + "eval_samples_per_second": 24.734, + "eval_steps_per_second": 12.367, + "num_input_tokens_seen": 25257664, + "step": 16600 + }, + { + "epoch": 0.531496062992126, + "grad_norm": 0.6724062561988831, + "learning_rate": 3.1589250307469134e-05, + "loss": 0.5867, + "num_input_tokens_seen": 25265056, + "step": 16605 + }, + { + "epoch": 0.5316561039626144, + "grad_norm": 1.4089206457138062, + "learning_rate": 3.1579779465944586e-05, + "loss": 0.5696, + "num_input_tokens_seen": 25272656, + "step": 16610 + }, + { + "epoch": 0.5318161449331029, + "grad_norm": 2.906001329421997, + "learning_rate": 3.1570307609735363e-05, + "loss": 0.7587, + "num_input_tokens_seen": 25281040, + "step": 16615 + }, + { + "epoch": 0.5319761859035913, + "grad_norm": 0.7754539847373962, + "learning_rate": 3.156083474030213e-05, + "loss": 0.5504, + "num_input_tokens_seen": 25288448, + "step": 16620 + }, + { + "epoch": 0.5321362268740798, + "grad_norm": 0.8113704323768616, + "learning_rate": 3.155136085910573e-05, + "loss": 0.7245, + "num_input_tokens_seen": 25295984, + "step": 16625 + }, + { + "epoch": 0.5322962678445682, + "grad_norm": 0.9845893979072571, + "learning_rate": 3.154188596760717e-05, + "loss": 0.5436, + "num_input_tokens_seen": 25303648, + "step": 16630 + }, + { + "epoch": 0.5324563088150567, + "grad_norm": 0.7974210381507874, + "learning_rate": 3.153241006726757e-05, + "loss": 0.3934, + "num_input_tokens_seen": 25311152, + "step": 16635 + }, + { + "epoch": 0.5326163497855451, + "grad_norm": 0.5023690462112427, + "learning_rate": 3.152293315954825e-05, + "loss": 0.3824, + "num_input_tokens_seen": 25318800, + "step": 16640 + }, + { + "epoch": 0.5327763907560336, + "grad_norm": 0.9948844909667969, + "learning_rate": 3.1513455245910666e-05, + "loss": 0.6357, + "num_input_tokens_seen": 25326592, + "step": 16645 + }, + { + "epoch": 0.5329364317265219, + "grad_norm": 0.8764356970787048, + "learning_rate": 3.150397632781643e-05, + "loss": 0.4008, + "num_input_tokens_seen": 25334160, + "step": 16650 + }, + { + "epoch": 0.5330964726970104, + "grad_norm": 0.4923551380634308, + "learning_rate": 3.149449640672731e-05, + "loss": 0.5302, + "num_input_tokens_seen": 25341856, + "step": 16655 + }, + { + "epoch": 0.5332565136674989, + "grad_norm": 0.7728623151779175, + "learning_rate": 3.148501548410523e-05, + "loss": 0.4711, + "num_input_tokens_seen": 25349248, + "step": 16660 + }, + { + "epoch": 0.5334165546379873, + "grad_norm": 0.5481140613555908, + "learning_rate": 3.1475533561412256e-05, + "loss": 0.4714, + "num_input_tokens_seen": 25356608, + "step": 16665 + }, + { + "epoch": 0.5335765956084758, + "grad_norm": 0.798921525478363, + "learning_rate": 3.146605064011065e-05, + "loss": 0.346, + "num_input_tokens_seen": 25363888, + "step": 16670 + }, + { + "epoch": 0.5337366365789642, + "grad_norm": 0.3786429166793823, + "learning_rate": 3.145656672166277e-05, + "loss": 0.4405, + "num_input_tokens_seen": 25371392, + "step": 16675 + }, + { + "epoch": 0.5338966775494527, + "grad_norm": 0.6004661917686462, + "learning_rate": 3.144708180753116e-05, + "loss": 0.517, + "num_input_tokens_seen": 25379008, + "step": 16680 + }, + { + "epoch": 0.5340567185199411, + "grad_norm": 1.063419222831726, + "learning_rate": 3.143759589917851e-05, + "loss": 0.7875, + "num_input_tokens_seen": 25386544, + "step": 16685 + }, + { + "epoch": 0.5342167594904296, + "grad_norm": 1.0438799858093262, + "learning_rate": 3.142810899806768e-05, + "loss": 0.7468, + "num_input_tokens_seen": 25394288, + "step": 16690 + }, + { + "epoch": 0.534376800460918, + "grad_norm": 0.7803182005882263, + "learning_rate": 3.141862110566166e-05, + "loss": 0.5259, + "num_input_tokens_seen": 25402368, + "step": 16695 + }, + { + "epoch": 0.5345368414314065, + "grad_norm": 0.45741620659828186, + "learning_rate": 3.1409132223423606e-05, + "loss": 0.3876, + "num_input_tokens_seen": 25410048, + "step": 16700 + }, + { + "epoch": 0.5346968824018948, + "grad_norm": 0.6154036521911621, + "learning_rate": 3.139964235281682e-05, + "loss": 0.4565, + "num_input_tokens_seen": 25417680, + "step": 16705 + }, + { + "epoch": 0.5348569233723833, + "grad_norm": 0.7223185300827026, + "learning_rate": 3.139015149530476e-05, + "loss": 0.458, + "num_input_tokens_seen": 25425088, + "step": 16710 + }, + { + "epoch": 0.5350169643428718, + "grad_norm": 0.9825800657272339, + "learning_rate": 3.1380659652351034e-05, + "loss": 0.5377, + "num_input_tokens_seen": 25432528, + "step": 16715 + }, + { + "epoch": 0.5351770053133602, + "grad_norm": 0.5738939642906189, + "learning_rate": 3.137116682541941e-05, + "loss": 0.5212, + "num_input_tokens_seen": 25440224, + "step": 16720 + }, + { + "epoch": 0.5353370462838487, + "grad_norm": 0.9146373867988586, + "learning_rate": 3.136167301597379e-05, + "loss": 0.4238, + "num_input_tokens_seen": 25448144, + "step": 16725 + }, + { + "epoch": 0.5354970872543371, + "grad_norm": 0.5655967593193054, + "learning_rate": 3.1352178225478254e-05, + "loss": 0.4591, + "num_input_tokens_seen": 25455520, + "step": 16730 + }, + { + "epoch": 0.5356571282248256, + "grad_norm": 1.1096091270446777, + "learning_rate": 3.1342682455396996e-05, + "loss": 0.5837, + "num_input_tokens_seen": 25462864, + "step": 16735 + }, + { + "epoch": 0.535817169195314, + "grad_norm": 0.8107653260231018, + "learning_rate": 3.133318570719441e-05, + "loss": 0.6128, + "num_input_tokens_seen": 25470848, + "step": 16740 + }, + { + "epoch": 0.5359772101658025, + "grad_norm": 1.3119375705718994, + "learning_rate": 3.132368798233499e-05, + "loss": 0.3762, + "num_input_tokens_seen": 25477840, + "step": 16745 + }, + { + "epoch": 0.5361372511362908, + "grad_norm": 0.3552458882331848, + "learning_rate": 3.131418928228342e-05, + "loss": 0.5499, + "num_input_tokens_seen": 25485072, + "step": 16750 + }, + { + "epoch": 0.5362972921067793, + "grad_norm": 0.3871796727180481, + "learning_rate": 3.1304689608504514e-05, + "loss": 0.4153, + "num_input_tokens_seen": 25492448, + "step": 16755 + }, + { + "epoch": 0.5364573330772677, + "grad_norm": 0.8774692416191101, + "learning_rate": 3.129518896246324e-05, + "loss": 0.6635, + "num_input_tokens_seen": 25500096, + "step": 16760 + }, + { + "epoch": 0.5366173740477562, + "grad_norm": 0.3720734417438507, + "learning_rate": 3.128568734562472e-05, + "loss": 0.5745, + "num_input_tokens_seen": 25507792, + "step": 16765 + }, + { + "epoch": 0.5367774150182447, + "grad_norm": 0.57358717918396, + "learning_rate": 3.127618475945421e-05, + "loss": 0.5471, + "num_input_tokens_seen": 25515344, + "step": 16770 + }, + { + "epoch": 0.5369374559887331, + "grad_norm": 0.8401272296905518, + "learning_rate": 3.126668120541715e-05, + "loss": 0.4161, + "num_input_tokens_seen": 25522848, + "step": 16775 + }, + { + "epoch": 0.5370974969592216, + "grad_norm": 0.7045642733573914, + "learning_rate": 3.1257176684979096e-05, + "loss": 0.5451, + "num_input_tokens_seen": 25530848, + "step": 16780 + }, + { + "epoch": 0.53725753792971, + "grad_norm": 0.7999702095985413, + "learning_rate": 3.124767119960576e-05, + "loss": 0.5413, + "num_input_tokens_seen": 25538848, + "step": 16785 + }, + { + "epoch": 0.5374175789001985, + "grad_norm": 1.1575446128845215, + "learning_rate": 3.123816475076301e-05, + "loss": 0.5226, + "num_input_tokens_seen": 25546304, + "step": 16790 + }, + { + "epoch": 0.5375776198706869, + "grad_norm": 1.1642879247665405, + "learning_rate": 3.122865733991687e-05, + "loss": 0.4356, + "num_input_tokens_seen": 25553968, + "step": 16795 + }, + { + "epoch": 0.5377376608411754, + "grad_norm": 1.226412057876587, + "learning_rate": 3.1219148968533486e-05, + "loss": 0.441, + "num_input_tokens_seen": 25561392, + "step": 16800 + }, + { + "epoch": 0.5377376608411754, + "eval_loss": 0.5249937176704407, + "eval_runtime": 561.4408, + "eval_samples_per_second": 24.733, + "eval_steps_per_second": 12.366, + "num_input_tokens_seen": 25561392, + "step": 16800 + }, + { + "epoch": 0.5378977018116637, + "grad_norm": 0.9040122032165527, + "learning_rate": 3.120963963807918e-05, + "loss": 0.4727, + "num_input_tokens_seen": 25568848, + "step": 16805 + }, + { + "epoch": 0.5380577427821522, + "grad_norm": 0.9974980354309082, + "learning_rate": 3.12001293500204e-05, + "loss": 0.4386, + "num_input_tokens_seen": 25576144, + "step": 16810 + }, + { + "epoch": 0.5382177837526406, + "grad_norm": 0.46089112758636475, + "learning_rate": 3.1190618105823765e-05, + "loss": 0.3819, + "num_input_tokens_seen": 25583520, + "step": 16815 + }, + { + "epoch": 0.5383778247231291, + "grad_norm": 0.6122909784317017, + "learning_rate": 3.118110590695603e-05, + "loss": 0.4214, + "num_input_tokens_seen": 25590880, + "step": 16820 + }, + { + "epoch": 0.5385378656936176, + "grad_norm": 0.8445742726325989, + "learning_rate": 3.117159275488407e-05, + "loss": 0.3989, + "num_input_tokens_seen": 25598272, + "step": 16825 + }, + { + "epoch": 0.538697906664106, + "grad_norm": 1.0492616891860962, + "learning_rate": 3.1162078651074956e-05, + "loss": 0.4825, + "num_input_tokens_seen": 25606096, + "step": 16830 + }, + { + "epoch": 0.5388579476345945, + "grad_norm": 0.7243807315826416, + "learning_rate": 3.1152563596995885e-05, + "loss": 0.5949, + "num_input_tokens_seen": 25613872, + "step": 16835 + }, + { + "epoch": 0.5390179886050829, + "grad_norm": 0.6361618638038635, + "learning_rate": 3.1143047594114186e-05, + "loss": 0.4125, + "num_input_tokens_seen": 25621264, + "step": 16840 + }, + { + "epoch": 0.5391780295755714, + "grad_norm": 0.9498146176338196, + "learning_rate": 3.113353064389734e-05, + "loss": 0.4171, + "num_input_tokens_seen": 25628944, + "step": 16845 + }, + { + "epoch": 0.5393380705460598, + "grad_norm": 1.939011812210083, + "learning_rate": 3.1124012747812993e-05, + "loss": 0.6349, + "num_input_tokens_seen": 25635936, + "step": 16850 + }, + { + "epoch": 0.5394981115165483, + "grad_norm": 0.4442337155342102, + "learning_rate": 3.1114493907328936e-05, + "loss": 0.4759, + "num_input_tokens_seen": 25643632, + "step": 16855 + }, + { + "epoch": 0.5396581524870366, + "grad_norm": 0.6514855623245239, + "learning_rate": 3.110497412391306e-05, + "loss": 0.6797, + "num_input_tokens_seen": 25651792, + "step": 16860 + }, + { + "epoch": 0.5398181934575251, + "grad_norm": 0.8464357852935791, + "learning_rate": 3.1095453399033466e-05, + "loss": 0.6268, + "num_input_tokens_seen": 25659024, + "step": 16865 + }, + { + "epoch": 0.5399782344280136, + "grad_norm": 1.0771589279174805, + "learning_rate": 3.108593173415835e-05, + "loss": 0.4954, + "num_input_tokens_seen": 25666784, + "step": 16870 + }, + { + "epoch": 0.540138275398502, + "grad_norm": 0.9056537747383118, + "learning_rate": 3.107640913075609e-05, + "loss": 0.5519, + "num_input_tokens_seen": 25675216, + "step": 16875 + }, + { + "epoch": 0.5402983163689905, + "grad_norm": 1.165709376335144, + "learning_rate": 3.106688559029517e-05, + "loss": 0.5493, + "num_input_tokens_seen": 25683024, + "step": 16880 + }, + { + "epoch": 0.5404583573394789, + "grad_norm": 1.1127506494522095, + "learning_rate": 3.105736111424425e-05, + "loss": 0.517, + "num_input_tokens_seen": 25690272, + "step": 16885 + }, + { + "epoch": 0.5406183983099674, + "grad_norm": 0.8024916648864746, + "learning_rate": 3.1047835704072136e-05, + "loss": 0.5932, + "num_input_tokens_seen": 25698224, + "step": 16890 + }, + { + "epoch": 0.5407784392804558, + "grad_norm": 0.6997218728065491, + "learning_rate": 3.103830936124775e-05, + "loss": 0.4861, + "num_input_tokens_seen": 25705840, + "step": 16895 + }, + { + "epoch": 0.5409384802509443, + "grad_norm": 1.1384679079055786, + "learning_rate": 3.102878208724018e-05, + "loss": 0.6041, + "num_input_tokens_seen": 25713328, + "step": 16900 + }, + { + "epoch": 0.5410985212214326, + "grad_norm": 0.39429280161857605, + "learning_rate": 3.101925388351865e-05, + "loss": 0.489, + "num_input_tokens_seen": 25720656, + "step": 16905 + }, + { + "epoch": 0.5412585621919211, + "grad_norm": 0.648672342300415, + "learning_rate": 3.1009724751552515e-05, + "loss": 0.4821, + "num_input_tokens_seen": 25728000, + "step": 16910 + }, + { + "epoch": 0.5414186031624095, + "grad_norm": 0.8804163336753845, + "learning_rate": 3.100019469281131e-05, + "loss": 0.6625, + "num_input_tokens_seen": 25735472, + "step": 16915 + }, + { + "epoch": 0.541578644132898, + "grad_norm": 1.0227948427200317, + "learning_rate": 3.0990663708764685e-05, + "loss": 0.578, + "num_input_tokens_seen": 25743056, + "step": 16920 + }, + { + "epoch": 0.5417386851033865, + "grad_norm": 0.880757212638855, + "learning_rate": 3.098113180088243e-05, + "loss": 0.482, + "num_input_tokens_seen": 25750384, + "step": 16925 + }, + { + "epoch": 0.5418987260738749, + "grad_norm": 1.1271820068359375, + "learning_rate": 3.097159897063448e-05, + "loss": 0.4246, + "num_input_tokens_seen": 25757760, + "step": 16930 + }, + { + "epoch": 0.5420587670443634, + "grad_norm": 0.9231168031692505, + "learning_rate": 3.096206521949094e-05, + "loss": 0.6724, + "num_input_tokens_seen": 25765264, + "step": 16935 + }, + { + "epoch": 0.5422188080148518, + "grad_norm": 0.8450976014137268, + "learning_rate": 3.0952530548922006e-05, + "loss": 0.4006, + "num_input_tokens_seen": 25772912, + "step": 16940 + }, + { + "epoch": 0.5423788489853403, + "grad_norm": 0.6855316758155823, + "learning_rate": 3.0942994960398064e-05, + "loss": 0.3561, + "num_input_tokens_seen": 25780496, + "step": 16945 + }, + { + "epoch": 0.5425388899558287, + "grad_norm": 0.4019949734210968, + "learning_rate": 3.093345845538961e-05, + "loss": 0.4193, + "num_input_tokens_seen": 25788192, + "step": 16950 + }, + { + "epoch": 0.5426989309263172, + "grad_norm": 0.5372663736343384, + "learning_rate": 3.09239210353673e-05, + "loss": 0.4646, + "num_input_tokens_seen": 25795872, + "step": 16955 + }, + { + "epoch": 0.5428589718968055, + "grad_norm": 1.0153917074203491, + "learning_rate": 3.0914382701801926e-05, + "loss": 0.446, + "num_input_tokens_seen": 25803520, + "step": 16960 + }, + { + "epoch": 0.543019012867294, + "grad_norm": 0.7613310217857361, + "learning_rate": 3.090484345616441e-05, + "loss": 0.5757, + "num_input_tokens_seen": 25810880, + "step": 16965 + }, + { + "epoch": 0.5431790538377824, + "grad_norm": 0.5613976120948792, + "learning_rate": 3.0895303299925825e-05, + "loss": 0.5004, + "num_input_tokens_seen": 25818176, + "step": 16970 + }, + { + "epoch": 0.5433390948082709, + "grad_norm": 1.26933753490448, + "learning_rate": 3.0885762234557393e-05, + "loss": 0.5268, + "num_input_tokens_seen": 25825808, + "step": 16975 + }, + { + "epoch": 0.5434991357787594, + "grad_norm": 0.7438509464263916, + "learning_rate": 3.087622026153045e-05, + "loss": 0.444, + "num_input_tokens_seen": 25833536, + "step": 16980 + }, + { + "epoch": 0.5436591767492478, + "grad_norm": 1.1198461055755615, + "learning_rate": 3.086667738231651e-05, + "loss": 0.6388, + "num_input_tokens_seen": 25841040, + "step": 16985 + }, + { + "epoch": 0.5438192177197363, + "grad_norm": 1.018491506576538, + "learning_rate": 3.085713359838718e-05, + "loss": 0.5842, + "num_input_tokens_seen": 25848512, + "step": 16990 + }, + { + "epoch": 0.5439792586902247, + "grad_norm": 0.6677848696708679, + "learning_rate": 3.084758891121425e-05, + "loss": 0.4519, + "num_input_tokens_seen": 25855584, + "step": 16995 + }, + { + "epoch": 0.5441392996607132, + "grad_norm": 0.4062381982803345, + "learning_rate": 3.083804332226963e-05, + "loss": 0.4989, + "num_input_tokens_seen": 25862992, + "step": 17000 + }, + { + "epoch": 0.5441392996607132, + "eval_loss": 0.5235934853553772, + "eval_runtime": 561.4662, + "eval_samples_per_second": 24.732, + "eval_steps_per_second": 12.366, + "num_input_tokens_seen": 25862992, + "step": 17000 + }, + { + "epoch": 0.5442993406312016, + "grad_norm": 1.474153995513916, + "learning_rate": 3.082849683302536e-05, + "loss": 0.6389, + "num_input_tokens_seen": 25870688, + "step": 17005 + }, + { + "epoch": 0.54445938160169, + "grad_norm": 0.6948064565658569, + "learning_rate": 3.081894944495363e-05, + "loss": 0.5863, + "num_input_tokens_seen": 25878528, + "step": 17010 + }, + { + "epoch": 0.5446194225721784, + "grad_norm": 1.2132936716079712, + "learning_rate": 3.080940115952677e-05, + "loss": 0.7387, + "num_input_tokens_seen": 25886352, + "step": 17015 + }, + { + "epoch": 0.5447794635426669, + "grad_norm": 0.7700939774513245, + "learning_rate": 3.0799851978217245e-05, + "loss": 0.6899, + "num_input_tokens_seen": 25894144, + "step": 17020 + }, + { + "epoch": 0.5449395045131553, + "grad_norm": 0.7291902899742126, + "learning_rate": 3.0790301902497666e-05, + "loss": 0.6036, + "num_input_tokens_seen": 25902368, + "step": 17025 + }, + { + "epoch": 0.5450995454836438, + "grad_norm": 0.8537243604660034, + "learning_rate": 3.078075093384076e-05, + "loss": 0.5598, + "num_input_tokens_seen": 25910272, + "step": 17030 + }, + { + "epoch": 0.5452595864541323, + "grad_norm": 0.9963390231132507, + "learning_rate": 3.077119907371942e-05, + "loss": 0.571, + "num_input_tokens_seen": 25917744, + "step": 17035 + }, + { + "epoch": 0.5454196274246207, + "grad_norm": 1.0965653657913208, + "learning_rate": 3.076164632360666e-05, + "loss": 0.5791, + "num_input_tokens_seen": 25925520, + "step": 17040 + }, + { + "epoch": 0.5455796683951092, + "grad_norm": 0.4920889139175415, + "learning_rate": 3.075209268497563e-05, + "loss": 0.3666, + "num_input_tokens_seen": 25933456, + "step": 17045 + }, + { + "epoch": 0.5457397093655976, + "grad_norm": 1.1392287015914917, + "learning_rate": 3.074253815929961e-05, + "loss": 0.5256, + "num_input_tokens_seen": 25941104, + "step": 17050 + }, + { + "epoch": 0.5458997503360861, + "grad_norm": 0.7651338577270508, + "learning_rate": 3.0732982748052054e-05, + "loss": 0.5477, + "num_input_tokens_seen": 25948592, + "step": 17055 + }, + { + "epoch": 0.5460597913065744, + "grad_norm": 0.9450258612632751, + "learning_rate": 3.072342645270651e-05, + "loss": 0.5653, + "num_input_tokens_seen": 25956176, + "step": 17060 + }, + { + "epoch": 0.5462198322770629, + "grad_norm": 1.2397162914276123, + "learning_rate": 3.071386927473668e-05, + "loss": 0.4552, + "num_input_tokens_seen": 25963952, + "step": 17065 + }, + { + "epoch": 0.5463798732475513, + "grad_norm": 0.48593196272850037, + "learning_rate": 3.0704311215616404e-05, + "loss": 0.3331, + "num_input_tokens_seen": 25971152, + "step": 17070 + }, + { + "epoch": 0.5465399142180398, + "grad_norm": 0.7689840793609619, + "learning_rate": 3.0694752276819656e-05, + "loss": 0.391, + "num_input_tokens_seen": 25978256, + "step": 17075 + }, + { + "epoch": 0.5466999551885282, + "grad_norm": 1.2097203731536865, + "learning_rate": 3.068519245982054e-05, + "loss": 0.5896, + "num_input_tokens_seen": 25985520, + "step": 17080 + }, + { + "epoch": 0.5468599961590167, + "grad_norm": 0.7438316941261292, + "learning_rate": 3.0675631766093304e-05, + "loss": 0.6204, + "num_input_tokens_seen": 25993408, + "step": 17085 + }, + { + "epoch": 0.5470200371295052, + "grad_norm": 0.6304094791412354, + "learning_rate": 3.066607019711232e-05, + "loss": 0.4542, + "num_input_tokens_seen": 26000672, + "step": 17090 + }, + { + "epoch": 0.5471800780999936, + "grad_norm": 0.7600059509277344, + "learning_rate": 3.065650775435211e-05, + "loss": 0.3867, + "num_input_tokens_seen": 26008112, + "step": 17095 + }, + { + "epoch": 0.5473401190704821, + "grad_norm": 0.6319849491119385, + "learning_rate": 3.0646944439287326e-05, + "loss": 0.6037, + "num_input_tokens_seen": 26015408, + "step": 17100 + }, + { + "epoch": 0.5475001600409705, + "grad_norm": 1.0364644527435303, + "learning_rate": 3.0637380253392736e-05, + "loss": 0.569, + "num_input_tokens_seen": 26022688, + "step": 17105 + }, + { + "epoch": 0.547660201011459, + "grad_norm": 0.5161800980567932, + "learning_rate": 3.062781519814327e-05, + "loss": 0.3451, + "num_input_tokens_seen": 26029792, + "step": 17110 + }, + { + "epoch": 0.5478202419819473, + "grad_norm": 0.7143962979316711, + "learning_rate": 3.0618249275013985e-05, + "loss": 0.431, + "num_input_tokens_seen": 26037104, + "step": 17115 + }, + { + "epoch": 0.5479802829524358, + "grad_norm": 1.302582859992981, + "learning_rate": 3.060868248548005e-05, + "loss": 0.7169, + "num_input_tokens_seen": 26044864, + "step": 17120 + }, + { + "epoch": 0.5481403239229242, + "grad_norm": 1.0068238973617554, + "learning_rate": 3.0599114831016796e-05, + "loss": 0.5222, + "num_input_tokens_seen": 26052560, + "step": 17125 + }, + { + "epoch": 0.5483003648934127, + "grad_norm": 1.3697035312652588, + "learning_rate": 3.0589546313099666e-05, + "loss": 0.6084, + "num_input_tokens_seen": 26060528, + "step": 17130 + }, + { + "epoch": 0.5484604058639012, + "grad_norm": 0.7387868165969849, + "learning_rate": 3.0579976933204255e-05, + "loss": 0.5309, + "num_input_tokens_seen": 26067888, + "step": 17135 + }, + { + "epoch": 0.5486204468343896, + "grad_norm": 0.9845479130744934, + "learning_rate": 3.0570406692806284e-05, + "loss": 0.6465, + "num_input_tokens_seen": 26076000, + "step": 17140 + }, + { + "epoch": 0.5487804878048781, + "grad_norm": 0.5609056949615479, + "learning_rate": 3.05608355933816e-05, + "loss": 0.4521, + "num_input_tokens_seen": 26083936, + "step": 17145 + }, + { + "epoch": 0.5489405287753665, + "grad_norm": 1.2380053997039795, + "learning_rate": 3.055126363640618e-05, + "loss": 0.5727, + "num_input_tokens_seen": 26091296, + "step": 17150 + }, + { + "epoch": 0.549100569745855, + "grad_norm": 0.6620471477508545, + "learning_rate": 3.0541690823356146e-05, + "loss": 0.4614, + "num_input_tokens_seen": 26099280, + "step": 17155 + }, + { + "epoch": 0.5492606107163434, + "grad_norm": 0.5242370367050171, + "learning_rate": 3.053211715570775e-05, + "loss": 0.3596, + "num_input_tokens_seen": 26107328, + "step": 17160 + }, + { + "epoch": 0.5494206516868319, + "grad_norm": 1.2418138980865479, + "learning_rate": 3.052254263493736e-05, + "loss": 0.6599, + "num_input_tokens_seen": 26115488, + "step": 17165 + }, + { + "epoch": 0.5495806926573202, + "grad_norm": 0.9534025192260742, + "learning_rate": 3.0512967262521498e-05, + "loss": 0.4113, + "num_input_tokens_seen": 26123584, + "step": 17170 + }, + { + "epoch": 0.5497407336278087, + "grad_norm": 1.0661674737930298, + "learning_rate": 3.0503391039936803e-05, + "loss": 0.5414, + "num_input_tokens_seen": 26131040, + "step": 17175 + }, + { + "epoch": 0.5499007745982971, + "grad_norm": 0.7137097120285034, + "learning_rate": 3.0493813968660056e-05, + "loss": 0.372, + "num_input_tokens_seen": 26138800, + "step": 17180 + }, + { + "epoch": 0.5500608155687856, + "grad_norm": 0.9267966747283936, + "learning_rate": 3.0484236050168153e-05, + "loss": 0.5782, + "num_input_tokens_seen": 26146576, + "step": 17185 + }, + { + "epoch": 0.5502208565392741, + "grad_norm": 0.5974714159965515, + "learning_rate": 3.0474657285938123e-05, + "loss": 0.6356, + "num_input_tokens_seen": 26154240, + "step": 17190 + }, + { + "epoch": 0.5503808975097625, + "grad_norm": 0.5518897771835327, + "learning_rate": 3.046507767744715e-05, + "loss": 0.5033, + "num_input_tokens_seen": 26161664, + "step": 17195 + }, + { + "epoch": 0.550540938480251, + "grad_norm": 0.4982105791568756, + "learning_rate": 3.045549722617252e-05, + "loss": 0.4765, + "num_input_tokens_seen": 26169264, + "step": 17200 + }, + { + "epoch": 0.550540938480251, + "eval_loss": 0.5230038166046143, + "eval_runtime": 561.3434, + "eval_samples_per_second": 24.737, + "eval_steps_per_second": 12.369, + "num_input_tokens_seen": 26169264, + "step": 17200 + }, + { + "epoch": 0.5507009794507394, + "grad_norm": 0.811053991317749, + "learning_rate": 3.0445915933591658e-05, + "loss": 0.5112, + "num_input_tokens_seen": 26177008, + "step": 17205 + }, + { + "epoch": 0.5508610204212279, + "grad_norm": 0.5341818332672119, + "learning_rate": 3.0436333801182114e-05, + "loss": 0.4856, + "num_input_tokens_seen": 26184864, + "step": 17210 + }, + { + "epoch": 0.5510210613917162, + "grad_norm": 2.3821399211883545, + "learning_rate": 3.0426750830421596e-05, + "loss": 0.5388, + "num_input_tokens_seen": 26192208, + "step": 17215 + }, + { + "epoch": 0.5511811023622047, + "grad_norm": 0.4910202920436859, + "learning_rate": 3.0417167022787897e-05, + "loss": 0.5062, + "num_input_tokens_seen": 26199888, + "step": 17220 + }, + { + "epoch": 0.5513411433326931, + "grad_norm": 1.5613458156585693, + "learning_rate": 3.0407582379758966e-05, + "loss": 0.5564, + "num_input_tokens_seen": 26207216, + "step": 17225 + }, + { + "epoch": 0.5515011843031816, + "grad_norm": 0.9628223776817322, + "learning_rate": 3.039799690281287e-05, + "loss": 0.5954, + "num_input_tokens_seen": 26214464, + "step": 17230 + }, + { + "epoch": 0.55166122527367, + "grad_norm": 1.7418248653411865, + "learning_rate": 3.0388410593427823e-05, + "loss": 0.8677, + "num_input_tokens_seen": 26221952, + "step": 17235 + }, + { + "epoch": 0.5518212662441585, + "grad_norm": 0.4758302867412567, + "learning_rate": 3.0378823453082146e-05, + "loss": 0.6183, + "num_input_tokens_seen": 26229392, + "step": 17240 + }, + { + "epoch": 0.551981307214647, + "grad_norm": 0.48755013942718506, + "learning_rate": 3.03692354832543e-05, + "loss": 0.3661, + "num_input_tokens_seen": 26236784, + "step": 17245 + }, + { + "epoch": 0.5521413481851354, + "grad_norm": 1.4774218797683716, + "learning_rate": 3.0359646685422865e-05, + "loss": 0.5654, + "num_input_tokens_seen": 26244192, + "step": 17250 + }, + { + "epoch": 0.5523013891556239, + "grad_norm": 0.6243028044700623, + "learning_rate": 3.035005706106656e-05, + "loss": 0.4543, + "num_input_tokens_seen": 26251712, + "step": 17255 + }, + { + "epoch": 0.5524614301261123, + "grad_norm": 0.6695688962936401, + "learning_rate": 3.034046661166422e-05, + "loss": 0.7771, + "num_input_tokens_seen": 26258864, + "step": 17260 + }, + { + "epoch": 0.5526214710966008, + "grad_norm": 0.5259593725204468, + "learning_rate": 3.033087533869482e-05, + "loss": 0.2938, + "num_input_tokens_seen": 26266288, + "step": 17265 + }, + { + "epoch": 0.5527815120670891, + "grad_norm": 0.8221638202667236, + "learning_rate": 3.0321283243637444e-05, + "loss": 0.4946, + "num_input_tokens_seen": 26273936, + "step": 17270 + }, + { + "epoch": 0.5529415530375776, + "grad_norm": 1.289249300956726, + "learning_rate": 3.0311690327971326e-05, + "loss": 0.616, + "num_input_tokens_seen": 26281456, + "step": 17275 + }, + { + "epoch": 0.553101594008066, + "grad_norm": 0.7697759866714478, + "learning_rate": 3.030209659317581e-05, + "loss": 0.6034, + "num_input_tokens_seen": 26289008, + "step": 17280 + }, + { + "epoch": 0.5532616349785545, + "grad_norm": 0.7480800151824951, + "learning_rate": 3.0292502040730362e-05, + "loss": 0.5107, + "num_input_tokens_seen": 26296384, + "step": 17285 + }, + { + "epoch": 0.5534216759490429, + "grad_norm": 1.6461429595947266, + "learning_rate": 3.0282906672114597e-05, + "loss": 0.3886, + "num_input_tokens_seen": 26303920, + "step": 17290 + }, + { + "epoch": 0.5535817169195314, + "grad_norm": 1.1699097156524658, + "learning_rate": 3.027331048880823e-05, + "loss": 0.5574, + "num_input_tokens_seen": 26311520, + "step": 17295 + }, + { + "epoch": 0.5537417578900199, + "grad_norm": 0.5872467160224915, + "learning_rate": 3.0263713492291123e-05, + "loss": 0.4064, + "num_input_tokens_seen": 26319648, + "step": 17300 + }, + { + "epoch": 0.5539017988605083, + "grad_norm": 1.0055359601974487, + "learning_rate": 3.0254115684043242e-05, + "loss": 0.492, + "num_input_tokens_seen": 26327360, + "step": 17305 + }, + { + "epoch": 0.5540618398309968, + "grad_norm": 0.5772802829742432, + "learning_rate": 3.024451706554469e-05, + "loss": 0.4468, + "num_input_tokens_seen": 26335088, + "step": 17310 + }, + { + "epoch": 0.5542218808014852, + "grad_norm": 0.45271366834640503, + "learning_rate": 3.0234917638275705e-05, + "loss": 0.4586, + "num_input_tokens_seen": 26342544, + "step": 17315 + }, + { + "epoch": 0.5543819217719737, + "grad_norm": 0.7187497019767761, + "learning_rate": 3.0225317403716635e-05, + "loss": 0.4833, + "num_input_tokens_seen": 26350080, + "step": 17320 + }, + { + "epoch": 0.554541962742462, + "grad_norm": 0.7550798654556274, + "learning_rate": 3.0215716363347956e-05, + "loss": 0.5552, + "num_input_tokens_seen": 26357424, + "step": 17325 + }, + { + "epoch": 0.5547020037129505, + "grad_norm": 1.4538177251815796, + "learning_rate": 3.0206114518650275e-05, + "loss": 0.5136, + "num_input_tokens_seen": 26364944, + "step": 17330 + }, + { + "epoch": 0.5548620446834389, + "grad_norm": 0.8261287808418274, + "learning_rate": 3.0196511871104304e-05, + "loss": 0.4926, + "num_input_tokens_seen": 26372560, + "step": 17335 + }, + { + "epoch": 0.5550220856539274, + "grad_norm": 1.0842742919921875, + "learning_rate": 3.01869084221909e-05, + "loss": 0.6407, + "num_input_tokens_seen": 26379808, + "step": 17340 + }, + { + "epoch": 0.5551821266244159, + "grad_norm": 0.9140076637268066, + "learning_rate": 3.0177304173391037e-05, + "loss": 0.4964, + "num_input_tokens_seen": 26387792, + "step": 17345 + }, + { + "epoch": 0.5553421675949043, + "grad_norm": 0.8454163670539856, + "learning_rate": 3.01676991261858e-05, + "loss": 0.4462, + "num_input_tokens_seen": 26395008, + "step": 17350 + }, + { + "epoch": 0.5555022085653928, + "grad_norm": 0.5620718002319336, + "learning_rate": 3.015809328205642e-05, + "loss": 0.5282, + "num_input_tokens_seen": 26402432, + "step": 17355 + }, + { + "epoch": 0.5556622495358812, + "grad_norm": 0.7585345506668091, + "learning_rate": 3.0148486642484248e-05, + "loss": 0.4433, + "num_input_tokens_seen": 26410336, + "step": 17360 + }, + { + "epoch": 0.5558222905063697, + "grad_norm": 0.3759196102619171, + "learning_rate": 3.0138879208950722e-05, + "loss": 0.3609, + "num_input_tokens_seen": 26417360, + "step": 17365 + }, + { + "epoch": 0.555982331476858, + "grad_norm": 0.6649003028869629, + "learning_rate": 3.012927098293744e-05, + "loss": 0.5675, + "num_input_tokens_seen": 26425136, + "step": 17370 + }, + { + "epoch": 0.5561423724473465, + "grad_norm": 0.7392808198928833, + "learning_rate": 3.0119661965926123e-05, + "loss": 0.7037, + "num_input_tokens_seen": 26433168, + "step": 17375 + }, + { + "epoch": 0.5563024134178349, + "grad_norm": 0.662887454032898, + "learning_rate": 3.0110052159398587e-05, + "loss": 0.5556, + "num_input_tokens_seen": 26440672, + "step": 17380 + }, + { + "epoch": 0.5564624543883234, + "grad_norm": 0.5119553804397583, + "learning_rate": 3.0100441564836802e-05, + "loss": 0.5531, + "num_input_tokens_seen": 26448112, + "step": 17385 + }, + { + "epoch": 0.5566224953588118, + "grad_norm": 1.4204100370407104, + "learning_rate": 3.0090830183722817e-05, + "loss": 0.4774, + "num_input_tokens_seen": 26456016, + "step": 17390 + }, + { + "epoch": 0.5567825363293003, + "grad_norm": 0.5114992260932922, + "learning_rate": 3.0081218017538852e-05, + "loss": 0.3671, + "num_input_tokens_seen": 26463648, + "step": 17395 + }, + { + "epoch": 0.5569425772997888, + "grad_norm": 5.287968635559082, + "learning_rate": 3.0071605067767212e-05, + "loss": 0.5298, + "num_input_tokens_seen": 26471584, + "step": 17400 + }, + { + "epoch": 0.5569425772997888, + "eval_loss": 0.5215440392494202, + "eval_runtime": 561.474, + "eval_samples_per_second": 24.731, + "eval_steps_per_second": 12.366, + "num_input_tokens_seen": 26471584, + "step": 17400 + }, + { + "epoch": 0.5571026182702772, + "grad_norm": 1.061439037322998, + "learning_rate": 3.006199133589034e-05, + "loss": 0.4897, + "num_input_tokens_seen": 26479360, + "step": 17405 + }, + { + "epoch": 0.5572626592407657, + "grad_norm": 1.6215932369232178, + "learning_rate": 3.005237682339079e-05, + "loss": 0.6037, + "num_input_tokens_seen": 26486896, + "step": 17410 + }, + { + "epoch": 0.5574227002112541, + "grad_norm": 1.62063467502594, + "learning_rate": 3.0042761531751228e-05, + "loss": 0.5217, + "num_input_tokens_seen": 26494448, + "step": 17415 + }, + { + "epoch": 0.5575827411817426, + "grad_norm": 1.7906768321990967, + "learning_rate": 3.0033145462454482e-05, + "loss": 0.5223, + "num_input_tokens_seen": 26501856, + "step": 17420 + }, + { + "epoch": 0.5577427821522309, + "grad_norm": 0.6543260812759399, + "learning_rate": 3.002352861698345e-05, + "loss": 0.5284, + "num_input_tokens_seen": 26509200, + "step": 17425 + }, + { + "epoch": 0.5579028231227194, + "grad_norm": 0.8761099576950073, + "learning_rate": 3.0013910996821178e-05, + "loss": 0.4588, + "num_input_tokens_seen": 26516496, + "step": 17430 + }, + { + "epoch": 0.5580628640932078, + "grad_norm": 1.3561540842056274, + "learning_rate": 3.0004292603450817e-05, + "loss": 0.5034, + "num_input_tokens_seen": 26523888, + "step": 17435 + }, + { + "epoch": 0.5582229050636963, + "grad_norm": 0.8551333546638489, + "learning_rate": 2.9994673438355653e-05, + "loss": 0.4446, + "num_input_tokens_seen": 26531600, + "step": 17440 + }, + { + "epoch": 0.5583829460341847, + "grad_norm": 1.704360008239746, + "learning_rate": 2.9985053503019078e-05, + "loss": 0.7355, + "num_input_tokens_seen": 26538784, + "step": 17445 + }, + { + "epoch": 0.5585429870046732, + "grad_norm": 0.6665220260620117, + "learning_rate": 2.99754327989246e-05, + "loss": 0.4741, + "num_input_tokens_seen": 26546416, + "step": 17450 + }, + { + "epoch": 0.5587030279751617, + "grad_norm": 0.5790572166442871, + "learning_rate": 2.9965811327555864e-05, + "loss": 0.4427, + "num_input_tokens_seen": 26553840, + "step": 17455 + }, + { + "epoch": 0.5588630689456501, + "grad_norm": 0.9037244915962219, + "learning_rate": 2.995618909039662e-05, + "loss": 0.4662, + "num_input_tokens_seen": 26561312, + "step": 17460 + }, + { + "epoch": 0.5590231099161386, + "grad_norm": 0.642407238483429, + "learning_rate": 2.9946566088930727e-05, + "loss": 0.4887, + "num_input_tokens_seen": 26568816, + "step": 17465 + }, + { + "epoch": 0.559183150886627, + "grad_norm": 0.9194790124893188, + "learning_rate": 2.9936942324642192e-05, + "loss": 0.617, + "num_input_tokens_seen": 26575824, + "step": 17470 + }, + { + "epoch": 0.5593431918571155, + "grad_norm": 1.6331552267074585, + "learning_rate": 2.9927317799015097e-05, + "loss": 0.4376, + "num_input_tokens_seen": 26583520, + "step": 17475 + }, + { + "epoch": 0.5595032328276038, + "grad_norm": 0.6235457062721252, + "learning_rate": 2.9917692513533685e-05, + "loss": 0.5302, + "num_input_tokens_seen": 26591264, + "step": 17480 + }, + { + "epoch": 0.5596632737980923, + "grad_norm": 0.6238104104995728, + "learning_rate": 2.990806646968229e-05, + "loss": 0.5057, + "num_input_tokens_seen": 26598896, + "step": 17485 + }, + { + "epoch": 0.5598233147685807, + "grad_norm": 0.753558874130249, + "learning_rate": 2.989843966894536e-05, + "loss": 0.5454, + "num_input_tokens_seen": 26606512, + "step": 17490 + }, + { + "epoch": 0.5599833557390692, + "grad_norm": 0.637832522392273, + "learning_rate": 2.9888812112807472e-05, + "loss": 0.3658, + "num_input_tokens_seen": 26614496, + "step": 17495 + }, + { + "epoch": 0.5601433967095576, + "grad_norm": 0.6582695245742798, + "learning_rate": 2.987918380275333e-05, + "loss": 0.5051, + "num_input_tokens_seen": 26622096, + "step": 17500 + }, + { + "epoch": 0.5603034376800461, + "grad_norm": 0.7968776822090149, + "learning_rate": 2.9869554740267724e-05, + "loss": 0.5988, + "num_input_tokens_seen": 26630064, + "step": 17505 + }, + { + "epoch": 0.5604634786505346, + "grad_norm": 1.0623432397842407, + "learning_rate": 2.9859924926835585e-05, + "loss": 0.3888, + "num_input_tokens_seen": 26638032, + "step": 17510 + }, + { + "epoch": 0.560623519621023, + "grad_norm": 1.7343509197235107, + "learning_rate": 2.9850294363941944e-05, + "loss": 0.5623, + "num_input_tokens_seen": 26645872, + "step": 17515 + }, + { + "epoch": 0.5607835605915115, + "grad_norm": 0.6298938989639282, + "learning_rate": 2.9840663053071967e-05, + "loss": 0.5199, + "num_input_tokens_seen": 26653120, + "step": 17520 + }, + { + "epoch": 0.5609436015619998, + "grad_norm": 0.5816067457199097, + "learning_rate": 2.983103099571091e-05, + "loss": 0.377, + "num_input_tokens_seen": 26660976, + "step": 17525 + }, + { + "epoch": 0.5611036425324883, + "grad_norm": 0.43230241537094116, + "learning_rate": 2.9821398193344164e-05, + "loss": 0.4819, + "num_input_tokens_seen": 26669424, + "step": 17530 + }, + { + "epoch": 0.5612636835029767, + "grad_norm": 1.03272545337677, + "learning_rate": 2.9811764647457226e-05, + "loss": 0.53, + "num_input_tokens_seen": 26676864, + "step": 17535 + }, + { + "epoch": 0.5614237244734652, + "grad_norm": 1.644360899925232, + "learning_rate": 2.9802130359535714e-05, + "loss": 0.5962, + "num_input_tokens_seen": 26684416, + "step": 17540 + }, + { + "epoch": 0.5615837654439536, + "grad_norm": 1.1195341348648071, + "learning_rate": 2.979249533106535e-05, + "loss": 0.3663, + "num_input_tokens_seen": 26692192, + "step": 17545 + }, + { + "epoch": 0.5617438064144421, + "grad_norm": 0.5765323638916016, + "learning_rate": 2.9782859563531986e-05, + "loss": 0.4108, + "num_input_tokens_seen": 26699456, + "step": 17550 + }, + { + "epoch": 0.5619038473849305, + "grad_norm": 0.3776504695415497, + "learning_rate": 2.977322305842156e-05, + "loss": 0.3984, + "num_input_tokens_seen": 26706800, + "step": 17555 + }, + { + "epoch": 0.562063888355419, + "grad_norm": 1.1721901893615723, + "learning_rate": 2.9763585817220162e-05, + "loss": 0.5934, + "num_input_tokens_seen": 26713840, + "step": 17560 + }, + { + "epoch": 0.5622239293259075, + "grad_norm": 0.8769797682762146, + "learning_rate": 2.975394784141397e-05, + "loss": 0.4613, + "num_input_tokens_seen": 26721008, + "step": 17565 + }, + { + "epoch": 0.5623839702963959, + "grad_norm": 1.2603940963745117, + "learning_rate": 2.974430913248928e-05, + "loss": 0.5759, + "num_input_tokens_seen": 26728048, + "step": 17570 + }, + { + "epoch": 0.5625440112668844, + "grad_norm": 1.248365044593811, + "learning_rate": 2.9734669691932497e-05, + "loss": 0.4944, + "num_input_tokens_seen": 26735392, + "step": 17575 + }, + { + "epoch": 0.5627040522373727, + "grad_norm": 0.9459635615348816, + "learning_rate": 2.9725029521230147e-05, + "loss": 0.5761, + "num_input_tokens_seen": 26743120, + "step": 17580 + }, + { + "epoch": 0.5628640932078612, + "grad_norm": 0.7117725610733032, + "learning_rate": 2.9715388621868873e-05, + "loss": 0.7545, + "num_input_tokens_seen": 26750672, + "step": 17585 + }, + { + "epoch": 0.5630241341783496, + "grad_norm": 0.8722570538520813, + "learning_rate": 2.970574699533541e-05, + "loss": 0.442, + "num_input_tokens_seen": 26758352, + "step": 17590 + }, + { + "epoch": 0.5631841751488381, + "grad_norm": 0.4610704779624939, + "learning_rate": 2.969610464311662e-05, + "loss": 0.342, + "num_input_tokens_seen": 26766064, + "step": 17595 + }, + { + "epoch": 0.5633442161193265, + "grad_norm": 0.7156997919082642, + "learning_rate": 2.9686461566699487e-05, + "loss": 0.4642, + "num_input_tokens_seen": 26773280, + "step": 17600 + }, + { + "epoch": 0.5633442161193265, + "eval_loss": 0.5212998986244202, + "eval_runtime": 561.4062, + "eval_samples_per_second": 24.734, + "eval_steps_per_second": 12.367, + "num_input_tokens_seen": 26773280, + "step": 17600 + }, + { + "epoch": 0.563504257089815, + "grad_norm": 0.4861135184764862, + "learning_rate": 2.9676817767571086e-05, + "loss": 0.5674, + "num_input_tokens_seen": 26780672, + "step": 17605 + }, + { + "epoch": 0.5636642980603035, + "grad_norm": 0.6716658473014832, + "learning_rate": 2.966717324721861e-05, + "loss": 0.5465, + "num_input_tokens_seen": 26788448, + "step": 17610 + }, + { + "epoch": 0.5638243390307919, + "grad_norm": 1.5050830841064453, + "learning_rate": 2.9657528007129366e-05, + "loss": 0.5939, + "num_input_tokens_seen": 26795936, + "step": 17615 + }, + { + "epoch": 0.5639843800012804, + "grad_norm": 0.3472190499305725, + "learning_rate": 2.9647882048790777e-05, + "loss": 0.34, + "num_input_tokens_seen": 26803744, + "step": 17620 + }, + { + "epoch": 0.5641444209717688, + "grad_norm": 1.032755732536316, + "learning_rate": 2.963823537369037e-05, + "loss": 0.4835, + "num_input_tokens_seen": 26811168, + "step": 17625 + }, + { + "epoch": 0.5643044619422573, + "grad_norm": 0.5664652585983276, + "learning_rate": 2.9628587983315775e-05, + "loss": 0.6921, + "num_input_tokens_seen": 26818688, + "step": 17630 + }, + { + "epoch": 0.5644645029127456, + "grad_norm": 0.9017260074615479, + "learning_rate": 2.9618939879154746e-05, + "loss": 0.5936, + "num_input_tokens_seen": 26826032, + "step": 17635 + }, + { + "epoch": 0.5646245438832341, + "grad_norm": 0.7592001557350159, + "learning_rate": 2.9609291062695143e-05, + "loss": 0.4599, + "num_input_tokens_seen": 26833600, + "step": 17640 + }, + { + "epoch": 0.5647845848537225, + "grad_norm": 0.6767595410346985, + "learning_rate": 2.9599641535424938e-05, + "loss": 0.3786, + "num_input_tokens_seen": 26841008, + "step": 17645 + }, + { + "epoch": 0.564944625824211, + "grad_norm": 0.8629952073097229, + "learning_rate": 2.9589991298832202e-05, + "loss": 0.5104, + "num_input_tokens_seen": 26848832, + "step": 17650 + }, + { + "epoch": 0.5651046667946994, + "grad_norm": 0.7031072378158569, + "learning_rate": 2.958034035440513e-05, + "loss": 0.4622, + "num_input_tokens_seen": 26856592, + "step": 17655 + }, + { + "epoch": 0.5652647077651879, + "grad_norm": 1.6774407625198364, + "learning_rate": 2.957068870363201e-05, + "loss": 0.516, + "num_input_tokens_seen": 26863840, + "step": 17660 + }, + { + "epoch": 0.5654247487356764, + "grad_norm": 1.0140832662582397, + "learning_rate": 2.956103634800126e-05, + "loss": 0.6148, + "num_input_tokens_seen": 26871296, + "step": 17665 + }, + { + "epoch": 0.5655847897061648, + "grad_norm": 0.7052072882652283, + "learning_rate": 2.9551383289001384e-05, + "loss": 0.5142, + "num_input_tokens_seen": 26878816, + "step": 17670 + }, + { + "epoch": 0.5657448306766533, + "grad_norm": 1.2436379194259644, + "learning_rate": 2.9541729528121005e-05, + "loss": 0.5225, + "num_input_tokens_seen": 26887008, + "step": 17675 + }, + { + "epoch": 0.5659048716471416, + "grad_norm": 0.8082709908485413, + "learning_rate": 2.9532075066848856e-05, + "loss": 0.6156, + "num_input_tokens_seen": 26894608, + "step": 17680 + }, + { + "epoch": 0.5660649126176301, + "grad_norm": 0.6174049377441406, + "learning_rate": 2.9522419906673786e-05, + "loss": 0.5497, + "num_input_tokens_seen": 26902288, + "step": 17685 + }, + { + "epoch": 0.5662249535881185, + "grad_norm": 0.6158568859100342, + "learning_rate": 2.951276404908474e-05, + "loss": 0.4734, + "num_input_tokens_seen": 26910368, + "step": 17690 + }, + { + "epoch": 0.566384994558607, + "grad_norm": 0.5944452285766602, + "learning_rate": 2.9503107495570752e-05, + "loss": 0.4417, + "num_input_tokens_seen": 26917728, + "step": 17695 + }, + { + "epoch": 0.5665450355290954, + "grad_norm": 0.5105090737342834, + "learning_rate": 2.9493450247621003e-05, + "loss": 0.4753, + "num_input_tokens_seen": 26925152, + "step": 17700 + }, + { + "epoch": 0.5667050764995839, + "grad_norm": 0.8233117461204529, + "learning_rate": 2.948379230672476e-05, + "loss": 0.5062, + "num_input_tokens_seen": 26933008, + "step": 17705 + }, + { + "epoch": 0.5668651174700723, + "grad_norm": 0.6463912725448608, + "learning_rate": 2.9474133674371396e-05, + "loss": 0.4967, + "num_input_tokens_seen": 26940672, + "step": 17710 + }, + { + "epoch": 0.5670251584405608, + "grad_norm": 0.8022679090499878, + "learning_rate": 2.9464474352050387e-05, + "loss": 0.5184, + "num_input_tokens_seen": 26948352, + "step": 17715 + }, + { + "epoch": 0.5671851994110493, + "grad_norm": 0.5136876702308655, + "learning_rate": 2.9454814341251336e-05, + "loss": 0.593, + "num_input_tokens_seen": 26955824, + "step": 17720 + }, + { + "epoch": 0.5673452403815377, + "grad_norm": 1.3323568105697632, + "learning_rate": 2.9445153643463942e-05, + "loss": 0.6592, + "num_input_tokens_seen": 26963696, + "step": 17725 + }, + { + "epoch": 0.5675052813520262, + "grad_norm": 1.1788712739944458, + "learning_rate": 2.943549226017798e-05, + "loss": 0.4946, + "num_input_tokens_seen": 26971136, + "step": 17730 + }, + { + "epoch": 0.5676653223225145, + "grad_norm": 0.776615560054779, + "learning_rate": 2.942583019288337e-05, + "loss": 0.5486, + "num_input_tokens_seen": 26978704, + "step": 17735 + }, + { + "epoch": 0.567825363293003, + "grad_norm": 1.2540011405944824, + "learning_rate": 2.9416167443070132e-05, + "loss": 0.5845, + "num_input_tokens_seen": 26985968, + "step": 17740 + }, + { + "epoch": 0.5679854042634914, + "grad_norm": 0.3016582727432251, + "learning_rate": 2.9406504012228375e-05, + "loss": 0.5402, + "num_input_tokens_seen": 26993616, + "step": 17745 + }, + { + "epoch": 0.5681454452339799, + "grad_norm": 1.0090464353561401, + "learning_rate": 2.939683990184832e-05, + "loss": 0.6586, + "num_input_tokens_seen": 27000880, + "step": 17750 + }, + { + "epoch": 0.5683054862044683, + "grad_norm": 0.6814385652542114, + "learning_rate": 2.93871751134203e-05, + "loss": 0.4098, + "num_input_tokens_seen": 27008624, + "step": 17755 + }, + { + "epoch": 0.5684655271749568, + "grad_norm": 0.7805646061897278, + "learning_rate": 2.9377509648434752e-05, + "loss": 0.5083, + "num_input_tokens_seen": 27015760, + "step": 17760 + }, + { + "epoch": 0.5686255681454452, + "grad_norm": 0.5121346116065979, + "learning_rate": 2.9367843508382203e-05, + "loss": 0.3247, + "num_input_tokens_seen": 27023296, + "step": 17765 + }, + { + "epoch": 0.5687856091159337, + "grad_norm": 0.8102079629898071, + "learning_rate": 2.9358176694753293e-05, + "loss": 0.5152, + "num_input_tokens_seen": 27031552, + "step": 17770 + }, + { + "epoch": 0.5689456500864222, + "grad_norm": 1.0740329027175903, + "learning_rate": 2.9348509209038766e-05, + "loss": 0.4225, + "num_input_tokens_seen": 27039328, + "step": 17775 + }, + { + "epoch": 0.5691056910569106, + "grad_norm": 1.1147881746292114, + "learning_rate": 2.933884105272947e-05, + "loss": 0.6062, + "num_input_tokens_seen": 27047392, + "step": 17780 + }, + { + "epoch": 0.569265732027399, + "grad_norm": 0.6552900671958923, + "learning_rate": 2.9329172227316366e-05, + "loss": 0.4143, + "num_input_tokens_seen": 27054864, + "step": 17785 + }, + { + "epoch": 0.5694257729978874, + "grad_norm": 0.6982048153877258, + "learning_rate": 2.93195027342905e-05, + "loss": 0.3728, + "num_input_tokens_seen": 27062256, + "step": 17790 + }, + { + "epoch": 0.5695858139683759, + "grad_norm": 0.7117248177528381, + "learning_rate": 2.9309832575143024e-05, + "loss": 0.6267, + "num_input_tokens_seen": 27069696, + "step": 17795 + }, + { + "epoch": 0.5697458549388643, + "grad_norm": 1.0846290588378906, + "learning_rate": 2.930016175136521e-05, + "loss": 0.422, + "num_input_tokens_seen": 27077600, + "step": 17800 + }, + { + "epoch": 0.5697458549388643, + "eval_loss": 0.5204117298126221, + "eval_runtime": 561.3215, + "eval_samples_per_second": 24.738, + "eval_steps_per_second": 12.369, + "num_input_tokens_seen": 27077600, + "step": 17800 + }, + { + "epoch": 0.5699058959093528, + "grad_norm": 0.6409925818443298, + "learning_rate": 2.9290490264448412e-05, + "loss": 0.6169, + "num_input_tokens_seen": 27085008, + "step": 17805 + }, + { + "epoch": 0.5700659368798412, + "grad_norm": 1.025760531425476, + "learning_rate": 2.9280818115884094e-05, + "loss": 0.5267, + "num_input_tokens_seen": 27092848, + "step": 17810 + }, + { + "epoch": 0.5702259778503297, + "grad_norm": 0.6067785024642944, + "learning_rate": 2.9271145307163828e-05, + "loss": 0.4454, + "num_input_tokens_seen": 27100208, + "step": 17815 + }, + { + "epoch": 0.5703860188208181, + "grad_norm": 0.6557331085205078, + "learning_rate": 2.9261471839779287e-05, + "loss": 0.545, + "num_input_tokens_seen": 27107952, + "step": 17820 + }, + { + "epoch": 0.5705460597913066, + "grad_norm": 0.5046424865722656, + "learning_rate": 2.925179771522223e-05, + "loss": 0.4612, + "num_input_tokens_seen": 27114768, + "step": 17825 + }, + { + "epoch": 0.5707061007617951, + "grad_norm": 1.013304352760315, + "learning_rate": 2.9242122934984535e-05, + "loss": 0.4547, + "num_input_tokens_seen": 27122256, + "step": 17830 + }, + { + "epoch": 0.5708661417322834, + "grad_norm": 0.6608114838600159, + "learning_rate": 2.9232447500558176e-05, + "loss": 0.4364, + "num_input_tokens_seen": 27130304, + "step": 17835 + }, + { + "epoch": 0.571026182702772, + "grad_norm": 1.4935487508773804, + "learning_rate": 2.9222771413435225e-05, + "loss": 0.5119, + "num_input_tokens_seen": 27137888, + "step": 17840 + }, + { + "epoch": 0.5711862236732603, + "grad_norm": 1.3342375755310059, + "learning_rate": 2.9213094675107848e-05, + "loss": 0.5451, + "num_input_tokens_seen": 27145472, + "step": 17845 + }, + { + "epoch": 0.5713462646437488, + "grad_norm": 1.6413230895996094, + "learning_rate": 2.9203417287068335e-05, + "loss": 0.5715, + "num_input_tokens_seen": 27152864, + "step": 17850 + }, + { + "epoch": 0.5715063056142372, + "grad_norm": 0.5551530718803406, + "learning_rate": 2.9193739250809042e-05, + "loss": 0.351, + "num_input_tokens_seen": 27160336, + "step": 17855 + }, + { + "epoch": 0.5716663465847257, + "grad_norm": 0.6111531853675842, + "learning_rate": 2.9184060567822463e-05, + "loss": 0.491, + "num_input_tokens_seen": 27167856, + "step": 17860 + }, + { + "epoch": 0.5718263875552141, + "grad_norm": 1.2405461072921753, + "learning_rate": 2.9174381239601166e-05, + "loss": 0.4607, + "num_input_tokens_seen": 27175968, + "step": 17865 + }, + { + "epoch": 0.5719864285257026, + "grad_norm": 0.534383237361908, + "learning_rate": 2.916470126763783e-05, + "loss": 0.4177, + "num_input_tokens_seen": 27183408, + "step": 17870 + }, + { + "epoch": 0.5721464694961911, + "grad_norm": 1.157355785369873, + "learning_rate": 2.9155020653425203e-05, + "loss": 0.4823, + "num_input_tokens_seen": 27190832, + "step": 17875 + }, + { + "epoch": 0.5723065104666795, + "grad_norm": 0.5013614892959595, + "learning_rate": 2.9145339398456184e-05, + "loss": 0.3444, + "num_input_tokens_seen": 27198288, + "step": 17880 + }, + { + "epoch": 0.572466551437168, + "grad_norm": 1.4075263738632202, + "learning_rate": 2.913565750422374e-05, + "loss": 0.3517, + "num_input_tokens_seen": 27205744, + "step": 17885 + }, + { + "epoch": 0.5726265924076563, + "grad_norm": 0.6762638092041016, + "learning_rate": 2.9125974972220938e-05, + "loss": 0.4796, + "num_input_tokens_seen": 27213200, + "step": 17890 + }, + { + "epoch": 0.5727866333781448, + "grad_norm": 0.997982919216156, + "learning_rate": 2.9116291803940932e-05, + "loss": 0.5811, + "num_input_tokens_seen": 27221056, + "step": 17895 + }, + { + "epoch": 0.5729466743486332, + "grad_norm": 0.8849591016769409, + "learning_rate": 2.910660800087701e-05, + "loss": 0.5324, + "num_input_tokens_seen": 27228256, + "step": 17900 + }, + { + "epoch": 0.5731067153191217, + "grad_norm": 1.1278818845748901, + "learning_rate": 2.909692356452254e-05, + "loss": 0.6158, + "num_input_tokens_seen": 27235936, + "step": 17905 + }, + { + "epoch": 0.5732667562896101, + "grad_norm": 1.1441855430603027, + "learning_rate": 2.9087238496370962e-05, + "loss": 0.5346, + "num_input_tokens_seen": 27243824, + "step": 17910 + }, + { + "epoch": 0.5734267972600986, + "grad_norm": 1.6751362085342407, + "learning_rate": 2.907755279791583e-05, + "loss": 0.5684, + "num_input_tokens_seen": 27251616, + "step": 17915 + }, + { + "epoch": 0.573586838230587, + "grad_norm": 0.7647731304168701, + "learning_rate": 2.906786647065083e-05, + "loss": 0.5114, + "num_input_tokens_seen": 27259424, + "step": 17920 + }, + { + "epoch": 0.5737468792010755, + "grad_norm": 1.2915581464767456, + "learning_rate": 2.9058179516069695e-05, + "loss": 0.4719, + "num_input_tokens_seen": 27266976, + "step": 17925 + }, + { + "epoch": 0.573906920171564, + "grad_norm": 0.41826868057250977, + "learning_rate": 2.9048491935666282e-05, + "loss": 0.5171, + "num_input_tokens_seen": 27274624, + "step": 17930 + }, + { + "epoch": 0.5740669611420524, + "grad_norm": 0.6317813992500305, + "learning_rate": 2.9038803730934534e-05, + "loss": 0.4047, + "num_input_tokens_seen": 27282352, + "step": 17935 + }, + { + "epoch": 0.5742270021125409, + "grad_norm": 0.5372638702392578, + "learning_rate": 2.9029114903368503e-05, + "loss": 0.4487, + "num_input_tokens_seen": 27290288, + "step": 17940 + }, + { + "epoch": 0.5743870430830292, + "grad_norm": 1.0830533504486084, + "learning_rate": 2.9019425454462318e-05, + "loss": 0.5303, + "num_input_tokens_seen": 27297728, + "step": 17945 + }, + { + "epoch": 0.5745470840535177, + "grad_norm": 0.7972970008850098, + "learning_rate": 2.9009735385710212e-05, + "loss": 0.5507, + "num_input_tokens_seen": 27306256, + "step": 17950 + }, + { + "epoch": 0.5747071250240061, + "grad_norm": 1.0849360227584839, + "learning_rate": 2.900004469860652e-05, + "loss": 0.4803, + "num_input_tokens_seen": 27313728, + "step": 17955 + }, + { + "epoch": 0.5748671659944946, + "grad_norm": 1.4424376487731934, + "learning_rate": 2.8990353394645668e-05, + "loss": 0.7386, + "num_input_tokens_seen": 27321584, + "step": 17960 + }, + { + "epoch": 0.575027206964983, + "grad_norm": 0.8427581191062927, + "learning_rate": 2.8980661475322186e-05, + "loss": 0.4002, + "num_input_tokens_seen": 27328624, + "step": 17965 + }, + { + "epoch": 0.5751872479354715, + "grad_norm": 0.6640461683273315, + "learning_rate": 2.897096894213067e-05, + "loss": 0.4337, + "num_input_tokens_seen": 27336144, + "step": 17970 + }, + { + "epoch": 0.5753472889059599, + "grad_norm": 0.5926553010940552, + "learning_rate": 2.8961275796565845e-05, + "loss": 0.3732, + "num_input_tokens_seen": 27343696, + "step": 17975 + }, + { + "epoch": 0.5755073298764484, + "grad_norm": 1.5418015718460083, + "learning_rate": 2.8951582040122517e-05, + "loss": 0.4609, + "num_input_tokens_seen": 27351008, + "step": 17980 + }, + { + "epoch": 0.5756673708469369, + "grad_norm": 0.8170880675315857, + "learning_rate": 2.894188767429557e-05, + "loss": 0.4777, + "num_input_tokens_seen": 27358240, + "step": 17985 + }, + { + "epoch": 0.5758274118174252, + "grad_norm": 0.5448799729347229, + "learning_rate": 2.8932192700580014e-05, + "loss": 0.4618, + "num_input_tokens_seen": 27366032, + "step": 17990 + }, + { + "epoch": 0.5759874527879137, + "grad_norm": 0.4548736810684204, + "learning_rate": 2.8922497120470916e-05, + "loss": 0.4938, + "num_input_tokens_seen": 27374000, + "step": 17995 + }, + { + "epoch": 0.5761474937584021, + "grad_norm": 1.0885438919067383, + "learning_rate": 2.891280093546348e-05, + "loss": 0.6037, + "num_input_tokens_seen": 27381664, + "step": 18000 + }, + { + "epoch": 0.5761474937584021, + "eval_loss": 0.5197606086730957, + "eval_runtime": 560.9638, + "eval_samples_per_second": 24.754, + "eval_steps_per_second": 12.377, + "num_input_tokens_seen": 27381664, + "step": 18000 + }, + { + "epoch": 0.5763075347288906, + "grad_norm": 0.5126599669456482, + "learning_rate": 2.890310414705297e-05, + "loss": 0.4761, + "num_input_tokens_seen": 27390096, + "step": 18005 + }, + { + "epoch": 0.576467575699379, + "grad_norm": 0.35252296924591064, + "learning_rate": 2.8893406756734742e-05, + "loss": 0.3974, + "num_input_tokens_seen": 27398112, + "step": 18010 + }, + { + "epoch": 0.5766276166698675, + "grad_norm": 0.5761852860450745, + "learning_rate": 2.888370876600427e-05, + "loss": 0.5269, + "num_input_tokens_seen": 27406384, + "step": 18015 + }, + { + "epoch": 0.5767876576403559, + "grad_norm": 0.5234588980674744, + "learning_rate": 2.8874010176357104e-05, + "loss": 0.5035, + "num_input_tokens_seen": 27414592, + "step": 18020 + }, + { + "epoch": 0.5769476986108444, + "grad_norm": 0.7099037766456604, + "learning_rate": 2.886431098928888e-05, + "loss": 0.4831, + "num_input_tokens_seen": 27421952, + "step": 18025 + }, + { + "epoch": 0.5771077395813328, + "grad_norm": 0.40129753947257996, + "learning_rate": 2.885461120629534e-05, + "loss": 0.3945, + "num_input_tokens_seen": 27429440, + "step": 18030 + }, + { + "epoch": 0.5772677805518213, + "grad_norm": 0.45894086360931396, + "learning_rate": 2.8844910828872317e-05, + "loss": 0.5561, + "num_input_tokens_seen": 27437856, + "step": 18035 + }, + { + "epoch": 0.5774278215223098, + "grad_norm": 0.49594202637672424, + "learning_rate": 2.8835209858515715e-05, + "loss": 0.4459, + "num_input_tokens_seen": 27445984, + "step": 18040 + }, + { + "epoch": 0.5775878624927981, + "grad_norm": 1.8940144777297974, + "learning_rate": 2.8825508296721566e-05, + "loss": 0.7237, + "num_input_tokens_seen": 27453600, + "step": 18045 + }, + { + "epoch": 0.5777479034632866, + "grad_norm": 0.7243181467056274, + "learning_rate": 2.881580614498596e-05, + "loss": 0.434, + "num_input_tokens_seen": 27461072, + "step": 18050 + }, + { + "epoch": 0.577907944433775, + "grad_norm": 0.4172305166721344, + "learning_rate": 2.8806103404805103e-05, + "loss": 0.5432, + "num_input_tokens_seen": 27468352, + "step": 18055 + }, + { + "epoch": 0.5780679854042635, + "grad_norm": 1.7070907354354858, + "learning_rate": 2.8796400077675257e-05, + "loss": 0.6827, + "num_input_tokens_seen": 27476048, + "step": 18060 + }, + { + "epoch": 0.5782280263747519, + "grad_norm": 0.47954505681991577, + "learning_rate": 2.8786696165092812e-05, + "loss": 0.4851, + "num_input_tokens_seen": 27483376, + "step": 18065 + }, + { + "epoch": 0.5783880673452404, + "grad_norm": 1.4720733165740967, + "learning_rate": 2.8776991668554236e-05, + "loss": 0.6271, + "num_input_tokens_seen": 27491216, + "step": 18070 + }, + { + "epoch": 0.5785481083157288, + "grad_norm": 0.6004696488380432, + "learning_rate": 2.876728658955608e-05, + "loss": 0.3836, + "num_input_tokens_seen": 27499024, + "step": 18075 + }, + { + "epoch": 0.5787081492862173, + "grad_norm": 0.8693124651908875, + "learning_rate": 2.8757580929594986e-05, + "loss": 0.5136, + "num_input_tokens_seen": 27506352, + "step": 18080 + }, + { + "epoch": 0.5788681902567057, + "grad_norm": 1.0921911001205444, + "learning_rate": 2.87478746901677e-05, + "loss": 0.4471, + "num_input_tokens_seen": 27514240, + "step": 18085 + }, + { + "epoch": 0.5790282312271942, + "grad_norm": 1.3405179977416992, + "learning_rate": 2.873816787277103e-05, + "loss": 0.6234, + "num_input_tokens_seen": 27521168, + "step": 18090 + }, + { + "epoch": 0.5791882721976827, + "grad_norm": 0.7074670195579529, + "learning_rate": 2.8728460478901903e-05, + "loss": 0.5945, + "num_input_tokens_seen": 27528528, + "step": 18095 + }, + { + "epoch": 0.579348313168171, + "grad_norm": 0.5553057789802551, + "learning_rate": 2.8718752510057307e-05, + "loss": 0.4205, + "num_input_tokens_seen": 27536048, + "step": 18100 + }, + { + "epoch": 0.5795083541386595, + "grad_norm": 0.5012366771697998, + "learning_rate": 2.870904396773435e-05, + "loss": 0.4812, + "num_input_tokens_seen": 27543872, + "step": 18105 + }, + { + "epoch": 0.5796683951091479, + "grad_norm": 0.9110991954803467, + "learning_rate": 2.86993348534302e-05, + "loss": 0.6308, + "num_input_tokens_seen": 27551040, + "step": 18110 + }, + { + "epoch": 0.5798284360796364, + "grad_norm": 0.551802396774292, + "learning_rate": 2.868962516864212e-05, + "loss": 0.5002, + "num_input_tokens_seen": 27558288, + "step": 18115 + }, + { + "epoch": 0.5799884770501248, + "grad_norm": 1.2993792295455933, + "learning_rate": 2.8679914914867477e-05, + "loss": 0.4329, + "num_input_tokens_seen": 27565952, + "step": 18120 + }, + { + "epoch": 0.5801485180206133, + "grad_norm": 0.916602611541748, + "learning_rate": 2.8670204093603713e-05, + "loss": 0.5808, + "num_input_tokens_seen": 27573568, + "step": 18125 + }, + { + "epoch": 0.5803085589911017, + "grad_norm": 0.8526074886322021, + "learning_rate": 2.8660492706348357e-05, + "loss": 0.3838, + "num_input_tokens_seen": 27581072, + "step": 18130 + }, + { + "epoch": 0.5804685999615902, + "grad_norm": 0.7965792417526245, + "learning_rate": 2.8650780754599022e-05, + "loss": 0.5545, + "num_input_tokens_seen": 27588576, + "step": 18135 + }, + { + "epoch": 0.5806286409320787, + "grad_norm": 0.8733725547790527, + "learning_rate": 2.8641068239853407e-05, + "loss": 0.4654, + "num_input_tokens_seen": 27596000, + "step": 18140 + }, + { + "epoch": 0.580788681902567, + "grad_norm": 0.7138229608535767, + "learning_rate": 2.863135516360932e-05, + "loss": 0.5075, + "num_input_tokens_seen": 27603424, + "step": 18145 + }, + { + "epoch": 0.5809487228730555, + "grad_norm": 0.6748948097229004, + "learning_rate": 2.8621641527364633e-05, + "loss": 0.8119, + "num_input_tokens_seen": 27610800, + "step": 18150 + }, + { + "epoch": 0.5811087638435439, + "grad_norm": 0.3175811171531677, + "learning_rate": 2.8611927332617313e-05, + "loss": 0.4753, + "num_input_tokens_seen": 27618560, + "step": 18155 + }, + { + "epoch": 0.5812688048140324, + "grad_norm": 0.8727008700370789, + "learning_rate": 2.8602212580865405e-05, + "loss": 0.4351, + "num_input_tokens_seen": 27626800, + "step": 18160 + }, + { + "epoch": 0.5814288457845208, + "grad_norm": 1.2571477890014648, + "learning_rate": 2.859249727360705e-05, + "loss": 0.618, + "num_input_tokens_seen": 27634784, + "step": 18165 + }, + { + "epoch": 0.5815888867550093, + "grad_norm": 1.4242786169052124, + "learning_rate": 2.8582781412340465e-05, + "loss": 0.52, + "num_input_tokens_seen": 27641920, + "step": 18170 + }, + { + "epoch": 0.5817489277254977, + "grad_norm": 0.42911189794540405, + "learning_rate": 2.857306499856397e-05, + "loss": 0.3929, + "num_input_tokens_seen": 27649936, + "step": 18175 + }, + { + "epoch": 0.5819089686959862, + "grad_norm": 0.6192184686660767, + "learning_rate": 2.856334803377594e-05, + "loss": 0.4138, + "num_input_tokens_seen": 27658032, + "step": 18180 + }, + { + "epoch": 0.5820690096664746, + "grad_norm": 0.30603525042533875, + "learning_rate": 2.8553630519474867e-05, + "loss": 0.5141, + "num_input_tokens_seen": 27665168, + "step": 18185 + }, + { + "epoch": 0.5822290506369631, + "grad_norm": 0.6894403100013733, + "learning_rate": 2.8543912457159317e-05, + "loss": 0.6095, + "num_input_tokens_seen": 27672784, + "step": 18190 + }, + { + "epoch": 0.5823890916074516, + "grad_norm": 1.5349748134613037, + "learning_rate": 2.853419384832792e-05, + "loss": 0.4804, + "num_input_tokens_seen": 27680368, + "step": 18195 + }, + { + "epoch": 0.5825491325779399, + "grad_norm": 0.2896096110343933, + "learning_rate": 2.8524474694479423e-05, + "loss": 0.5, + "num_input_tokens_seen": 27688016, + "step": 18200 + }, + { + "epoch": 0.5825491325779399, + "eval_loss": 0.5191994905471802, + "eval_runtime": 561.1324, + "eval_samples_per_second": 24.746, + "eval_steps_per_second": 12.373, + "num_input_tokens_seen": 27688016, + "step": 18200 + }, + { + "epoch": 0.5827091735484284, + "grad_norm": 1.0074666738510132, + "learning_rate": 2.851475499711264e-05, + "loss": 0.5924, + "num_input_tokens_seen": 27695760, + "step": 18205 + }, + { + "epoch": 0.5828692145189168, + "grad_norm": 0.5145854353904724, + "learning_rate": 2.8505034757726468e-05, + "loss": 0.4798, + "num_input_tokens_seen": 27703072, + "step": 18210 + }, + { + "epoch": 0.5830292554894053, + "grad_norm": 0.4106283187866211, + "learning_rate": 2.8495313977819886e-05, + "loss": 0.5512, + "num_input_tokens_seen": 27710368, + "step": 18215 + }, + { + "epoch": 0.5831892964598937, + "grad_norm": 0.6049672961235046, + "learning_rate": 2.8485592658891956e-05, + "loss": 0.4706, + "num_input_tokens_seen": 27718080, + "step": 18220 + }, + { + "epoch": 0.5833493374303822, + "grad_norm": 0.834518551826477, + "learning_rate": 2.8475870802441844e-05, + "loss": 0.4614, + "num_input_tokens_seen": 27726416, + "step": 18225 + }, + { + "epoch": 0.5835093784008706, + "grad_norm": 0.4199415445327759, + "learning_rate": 2.8466148409968774e-05, + "loss": 0.4781, + "num_input_tokens_seen": 27733840, + "step": 18230 + }, + { + "epoch": 0.5836694193713591, + "grad_norm": 1.5899797677993774, + "learning_rate": 2.8456425482972067e-05, + "loss": 0.4954, + "num_input_tokens_seen": 27741344, + "step": 18235 + }, + { + "epoch": 0.5838294603418475, + "grad_norm": 0.7484381198883057, + "learning_rate": 2.84467020229511e-05, + "loss": 0.4224, + "num_input_tokens_seen": 27749056, + "step": 18240 + }, + { + "epoch": 0.583989501312336, + "grad_norm": 0.9072278738021851, + "learning_rate": 2.8436978031405375e-05, + "loss": 0.4206, + "num_input_tokens_seen": 27757104, + "step": 18245 + }, + { + "epoch": 0.5841495422828245, + "grad_norm": 0.9533885717391968, + "learning_rate": 2.842725350983445e-05, + "loss": 0.4101, + "num_input_tokens_seen": 27764800, + "step": 18250 + }, + { + "epoch": 0.5843095832533128, + "grad_norm": 0.7634472250938416, + "learning_rate": 2.8417528459737957e-05, + "loss": 0.4296, + "num_input_tokens_seen": 27772128, + "step": 18255 + }, + { + "epoch": 0.5844696242238013, + "grad_norm": 1.998674988746643, + "learning_rate": 2.8407802882615624e-05, + "loss": 0.5496, + "num_input_tokens_seen": 27779360, + "step": 18260 + }, + { + "epoch": 0.5846296651942897, + "grad_norm": 0.6680245995521545, + "learning_rate": 2.8398076779967277e-05, + "loss": 0.5269, + "num_input_tokens_seen": 27787056, + "step": 18265 + }, + { + "epoch": 0.5847897061647782, + "grad_norm": 1.9329348802566528, + "learning_rate": 2.8388350153292774e-05, + "loss": 0.5605, + "num_input_tokens_seen": 27794128, + "step": 18270 + }, + { + "epoch": 0.5849497471352666, + "grad_norm": 0.8032154440879822, + "learning_rate": 2.8378623004092103e-05, + "loss": 0.7553, + "num_input_tokens_seen": 27801776, + "step": 18275 + }, + { + "epoch": 0.5851097881057551, + "grad_norm": 0.7504251599311829, + "learning_rate": 2.8368895333865302e-05, + "loss": 0.4609, + "num_input_tokens_seen": 27809168, + "step": 18280 + }, + { + "epoch": 0.5852698290762435, + "grad_norm": 0.4068552851676941, + "learning_rate": 2.835916714411251e-05, + "loss": 0.3588, + "num_input_tokens_seen": 27816640, + "step": 18285 + }, + { + "epoch": 0.585429870046732, + "grad_norm": 0.7936936020851135, + "learning_rate": 2.8349438436333926e-05, + "loss": 0.6346, + "num_input_tokens_seen": 27823744, + "step": 18290 + }, + { + "epoch": 0.5855899110172204, + "grad_norm": 1.0236104726791382, + "learning_rate": 2.833970921202984e-05, + "loss": 0.3542, + "num_input_tokens_seen": 27831264, + "step": 18295 + }, + { + "epoch": 0.5857499519877089, + "grad_norm": 0.7645183801651001, + "learning_rate": 2.8329979472700628e-05, + "loss": 0.4388, + "num_input_tokens_seen": 27838768, + "step": 18300 + }, + { + "epoch": 0.5859099929581973, + "grad_norm": 0.6334354877471924, + "learning_rate": 2.832024921984674e-05, + "loss": 0.5372, + "num_input_tokens_seen": 27846672, + "step": 18305 + }, + { + "epoch": 0.5860700339286857, + "grad_norm": 0.6670653223991394, + "learning_rate": 2.8310518454968693e-05, + "loss": 0.6152, + "num_input_tokens_seen": 27854160, + "step": 18310 + }, + { + "epoch": 0.5862300748991742, + "grad_norm": 1.067909598350525, + "learning_rate": 2.8300787179567095e-05, + "loss": 0.4914, + "num_input_tokens_seen": 27861792, + "step": 18315 + }, + { + "epoch": 0.5863901158696626, + "grad_norm": 0.6943708658218384, + "learning_rate": 2.8291055395142636e-05, + "loss": 0.418, + "num_input_tokens_seen": 27869536, + "step": 18320 + }, + { + "epoch": 0.5865501568401511, + "grad_norm": 0.42834028601646423, + "learning_rate": 2.8281323103196073e-05, + "loss": 0.4545, + "num_input_tokens_seen": 27876848, + "step": 18325 + }, + { + "epoch": 0.5867101978106395, + "grad_norm": 0.6386202573776245, + "learning_rate": 2.8271590305228256e-05, + "loss": 0.5881, + "num_input_tokens_seen": 27884752, + "step": 18330 + }, + { + "epoch": 0.586870238781128, + "grad_norm": 0.8913755416870117, + "learning_rate": 2.82618570027401e-05, + "loss": 0.4434, + "num_input_tokens_seen": 27892464, + "step": 18335 + }, + { + "epoch": 0.5870302797516164, + "grad_norm": 1.1044241189956665, + "learning_rate": 2.8252123197232604e-05, + "loss": 0.4066, + "num_input_tokens_seen": 27900368, + "step": 18340 + }, + { + "epoch": 0.5871903207221049, + "grad_norm": 1.5744978189468384, + "learning_rate": 2.8242388890206843e-05, + "loss": 0.4512, + "num_input_tokens_seen": 27908032, + "step": 18345 + }, + { + "epoch": 0.5873503616925934, + "grad_norm": 1.6448332071304321, + "learning_rate": 2.8232654083163967e-05, + "loss": 0.6258, + "num_input_tokens_seen": 27915856, + "step": 18350 + }, + { + "epoch": 0.5875104026630817, + "grad_norm": 0.8840616345405579, + "learning_rate": 2.822291877760521e-05, + "loss": 0.5006, + "num_input_tokens_seen": 27923344, + "step": 18355 + }, + { + "epoch": 0.5876704436335702, + "grad_norm": 0.6125129461288452, + "learning_rate": 2.8213182975031864e-05, + "loss": 0.3233, + "num_input_tokens_seen": 27930912, + "step": 18360 + }, + { + "epoch": 0.5878304846040586, + "grad_norm": 0.469102680683136, + "learning_rate": 2.8203446676945337e-05, + "loss": 0.534, + "num_input_tokens_seen": 27939104, + "step": 18365 + }, + { + "epoch": 0.5879905255745471, + "grad_norm": 0.7382650375366211, + "learning_rate": 2.8193709884847075e-05, + "loss": 0.4401, + "num_input_tokens_seen": 27946048, + "step": 18370 + }, + { + "epoch": 0.5881505665450355, + "grad_norm": 0.7470505833625793, + "learning_rate": 2.8183972600238605e-05, + "loss": 0.4922, + "num_input_tokens_seen": 27953856, + "step": 18375 + }, + { + "epoch": 0.588310607515524, + "grad_norm": 0.4681538939476013, + "learning_rate": 2.817423482462156e-05, + "loss": 0.3837, + "num_input_tokens_seen": 27961216, + "step": 18380 + }, + { + "epoch": 0.5884706484860124, + "grad_norm": 0.5635206699371338, + "learning_rate": 2.8164496559497605e-05, + "loss": 0.4599, + "num_input_tokens_seen": 27969136, + "step": 18385 + }, + { + "epoch": 0.5886306894565009, + "grad_norm": 1.121106743812561, + "learning_rate": 2.815475780636852e-05, + "loss": 0.578, + "num_input_tokens_seen": 27977008, + "step": 18390 + }, + { + "epoch": 0.5887907304269893, + "grad_norm": 0.4341549575328827, + "learning_rate": 2.814501856673613e-05, + "loss": 0.3977, + "num_input_tokens_seen": 27984864, + "step": 18395 + }, + { + "epoch": 0.5889507713974778, + "grad_norm": 0.8311530947685242, + "learning_rate": 2.8135278842102353e-05, + "loss": 0.5292, + "num_input_tokens_seen": 27992592, + "step": 18400 + }, + { + "epoch": 0.5889507713974778, + "eval_loss": 0.5183470845222473, + "eval_runtime": 561.0333, + "eval_samples_per_second": 24.751, + "eval_steps_per_second": 12.375, + "num_input_tokens_seen": 27992592, + "step": 18400 + }, + { + "epoch": 0.5891108123679663, + "grad_norm": 0.833889901638031, + "learning_rate": 2.8125538633969183e-05, + "loss": 0.5136, + "num_input_tokens_seen": 28000208, + "step": 18405 + }, + { + "epoch": 0.5892708533384546, + "grad_norm": 1.0334945917129517, + "learning_rate": 2.8115797943838677e-05, + "loss": 0.508, + "num_input_tokens_seen": 28007616, + "step": 18410 + }, + { + "epoch": 0.5894308943089431, + "grad_norm": 2.1451776027679443, + "learning_rate": 2.810605677321298e-05, + "loss": 0.623, + "num_input_tokens_seen": 28015504, + "step": 18415 + }, + { + "epoch": 0.5895909352794315, + "grad_norm": 0.4834805130958557, + "learning_rate": 2.809631512359428e-05, + "loss": 0.3422, + "num_input_tokens_seen": 28023184, + "step": 18420 + }, + { + "epoch": 0.58975097624992, + "grad_norm": 0.47866082191467285, + "learning_rate": 2.8086572996484884e-05, + "loss": 0.3198, + "num_input_tokens_seen": 28030752, + "step": 18425 + }, + { + "epoch": 0.5899110172204084, + "grad_norm": 1.4954882860183716, + "learning_rate": 2.8076830393387143e-05, + "loss": 0.6582, + "num_input_tokens_seen": 28038192, + "step": 18430 + }, + { + "epoch": 0.5900710581908969, + "grad_norm": 0.5361660122871399, + "learning_rate": 2.8067087315803497e-05, + "loss": 0.3991, + "num_input_tokens_seen": 28045792, + "step": 18435 + }, + { + "epoch": 0.5902310991613853, + "grad_norm": 0.7306966781616211, + "learning_rate": 2.8057343765236433e-05, + "loss": 0.5438, + "num_input_tokens_seen": 28053232, + "step": 18440 + }, + { + "epoch": 0.5903911401318738, + "grad_norm": 1.1529734134674072, + "learning_rate": 2.804759974318854e-05, + "loss": 0.5237, + "num_input_tokens_seen": 28061632, + "step": 18445 + }, + { + "epoch": 0.5905511811023622, + "grad_norm": 0.5818424224853516, + "learning_rate": 2.8037855251162482e-05, + "loss": 0.3978, + "num_input_tokens_seen": 28069200, + "step": 18450 + }, + { + "epoch": 0.5907112220728507, + "grad_norm": 0.6568143963813782, + "learning_rate": 2.802811029066096e-05, + "loss": 0.456, + "num_input_tokens_seen": 28076656, + "step": 18455 + }, + { + "epoch": 0.5908712630433391, + "grad_norm": 0.9115195274353027, + "learning_rate": 2.8018364863186764e-05, + "loss": 0.4956, + "num_input_tokens_seen": 28084752, + "step": 18460 + }, + { + "epoch": 0.5910313040138275, + "grad_norm": 1.1101466417312622, + "learning_rate": 2.800861897024279e-05, + "loss": 0.5538, + "num_input_tokens_seen": 28092256, + "step": 18465 + }, + { + "epoch": 0.591191344984316, + "grad_norm": 0.567314624786377, + "learning_rate": 2.799887261333196e-05, + "loss": 0.4742, + "num_input_tokens_seen": 28099776, + "step": 18470 + }, + { + "epoch": 0.5913513859548044, + "grad_norm": 0.4305395483970642, + "learning_rate": 2.798912579395728e-05, + "loss": 0.4534, + "num_input_tokens_seen": 28107520, + "step": 18475 + }, + { + "epoch": 0.5915114269252929, + "grad_norm": 0.5702937841415405, + "learning_rate": 2.797937851362185e-05, + "loss": 0.3797, + "num_input_tokens_seen": 28115232, + "step": 18480 + }, + { + "epoch": 0.5916714678957813, + "grad_norm": 0.6999801397323608, + "learning_rate": 2.7969630773828802e-05, + "loss": 0.4982, + "num_input_tokens_seen": 28122976, + "step": 18485 + }, + { + "epoch": 0.5918315088662698, + "grad_norm": 0.967430591583252, + "learning_rate": 2.7959882576081382e-05, + "loss": 0.4169, + "num_input_tokens_seen": 28130176, + "step": 18490 + }, + { + "epoch": 0.5919915498367582, + "grad_norm": 0.6649240255355835, + "learning_rate": 2.795013392188286e-05, + "loss": 0.4066, + "num_input_tokens_seen": 28137536, + "step": 18495 + }, + { + "epoch": 0.5921515908072467, + "grad_norm": 0.6030566692352295, + "learning_rate": 2.7940384812736614e-05, + "loss": 0.4407, + "num_input_tokens_seen": 28145328, + "step": 18500 + }, + { + "epoch": 0.592311631777735, + "grad_norm": 0.7069063186645508, + "learning_rate": 2.7930635250146087e-05, + "loss": 0.4803, + "num_input_tokens_seen": 28152688, + "step": 18505 + }, + { + "epoch": 0.5924716727482235, + "grad_norm": 0.854673445224762, + "learning_rate": 2.792088523561477e-05, + "loss": 0.4096, + "num_input_tokens_seen": 28160272, + "step": 18510 + }, + { + "epoch": 0.592631713718712, + "grad_norm": 1.377485752105713, + "learning_rate": 2.7911134770646246e-05, + "loss": 0.4386, + "num_input_tokens_seen": 28168480, + "step": 18515 + }, + { + "epoch": 0.5927917546892004, + "grad_norm": 0.8165595531463623, + "learning_rate": 2.7901383856744157e-05, + "loss": 0.4843, + "num_input_tokens_seen": 28176272, + "step": 18520 + }, + { + "epoch": 0.5929517956596889, + "grad_norm": 1.2512216567993164, + "learning_rate": 2.7891632495412217e-05, + "loss": 0.5308, + "num_input_tokens_seen": 28183920, + "step": 18525 + }, + { + "epoch": 0.5931118366301773, + "grad_norm": 0.7157017588615417, + "learning_rate": 2.7881880688154205e-05, + "loss": 0.4343, + "num_input_tokens_seen": 28191344, + "step": 18530 + }, + { + "epoch": 0.5932718776006658, + "grad_norm": 0.6601340174674988, + "learning_rate": 2.7872128436473977e-05, + "loss": 0.432, + "num_input_tokens_seen": 28199072, + "step": 18535 + }, + { + "epoch": 0.5934319185711542, + "grad_norm": 0.42725670337677, + "learning_rate": 2.7862375741875448e-05, + "loss": 0.3585, + "num_input_tokens_seen": 28206560, + "step": 18540 + }, + { + "epoch": 0.5935919595416427, + "grad_norm": 0.6970852613449097, + "learning_rate": 2.785262260586261e-05, + "loss": 0.6283, + "num_input_tokens_seen": 28214304, + "step": 18545 + }, + { + "epoch": 0.5937520005121311, + "grad_norm": 0.7956437468528748, + "learning_rate": 2.7842869029939517e-05, + "loss": 0.6199, + "num_input_tokens_seen": 28221472, + "step": 18550 + }, + { + "epoch": 0.5939120414826196, + "grad_norm": 1.1310324668884277, + "learning_rate": 2.7833115015610296e-05, + "loss": 0.4848, + "num_input_tokens_seen": 28229136, + "step": 18555 + }, + { + "epoch": 0.5940720824531079, + "grad_norm": 0.6531811356544495, + "learning_rate": 2.7823360564379136e-05, + "loss": 0.5315, + "num_input_tokens_seen": 28236464, + "step": 18560 + }, + { + "epoch": 0.5942321234235964, + "grad_norm": 0.8455192446708679, + "learning_rate": 2.7813605677750297e-05, + "loss": 0.5542, + "num_input_tokens_seen": 28243872, + "step": 18565 + }, + { + "epoch": 0.5943921643940849, + "grad_norm": 0.6490011811256409, + "learning_rate": 2.7803850357228102e-05, + "loss": 0.4707, + "num_input_tokens_seen": 28251296, + "step": 18570 + }, + { + "epoch": 0.5945522053645733, + "grad_norm": 0.47826048731803894, + "learning_rate": 2.779409460431695e-05, + "loss": 0.6259, + "num_input_tokens_seen": 28258816, + "step": 18575 + }, + { + "epoch": 0.5947122463350618, + "grad_norm": 0.5760512948036194, + "learning_rate": 2.778433842052129e-05, + "loss": 0.4178, + "num_input_tokens_seen": 28266720, + "step": 18580 + }, + { + "epoch": 0.5948722873055502, + "grad_norm": 1.7869962453842163, + "learning_rate": 2.7774581807345664e-05, + "loss": 0.5567, + "num_input_tokens_seen": 28274944, + "step": 18585 + }, + { + "epoch": 0.5950323282760387, + "grad_norm": 0.8767525553703308, + "learning_rate": 2.776482476629465e-05, + "loss": 0.6383, + "num_input_tokens_seen": 28282640, + "step": 18590 + }, + { + "epoch": 0.5951923692465271, + "grad_norm": 0.7711523175239563, + "learning_rate": 2.7755067298872924e-05, + "loss": 0.6163, + "num_input_tokens_seen": 28290400, + "step": 18595 + }, + { + "epoch": 0.5953524102170156, + "grad_norm": 0.7971335053443909, + "learning_rate": 2.774530940658518e-05, + "loss": 0.3928, + "num_input_tokens_seen": 28297696, + "step": 18600 + }, + { + "epoch": 0.5953524102170156, + "eval_loss": 0.5178605318069458, + "eval_runtime": 561.9056, + "eval_samples_per_second": 24.712, + "eval_steps_per_second": 12.356, + "num_input_tokens_seen": 28297696, + "step": 18600 + }, + { + "epoch": 0.595512451187504, + "grad_norm": 0.8041240572929382, + "learning_rate": 2.7735551090936236e-05, + "loss": 0.6498, + "num_input_tokens_seen": 28305632, + "step": 18605 + }, + { + "epoch": 0.5956724921579925, + "grad_norm": 0.7099907398223877, + "learning_rate": 2.7725792353430934e-05, + "loss": 0.4527, + "num_input_tokens_seen": 28313008, + "step": 18610 + }, + { + "epoch": 0.595832533128481, + "grad_norm": 0.7300171256065369, + "learning_rate": 2.77160331955742e-05, + "loss": 0.3795, + "num_input_tokens_seen": 28321008, + "step": 18615 + }, + { + "epoch": 0.5959925740989693, + "grad_norm": 0.7712820768356323, + "learning_rate": 2.7706273618871008e-05, + "loss": 0.4691, + "num_input_tokens_seen": 28328768, + "step": 18620 + }, + { + "epoch": 0.5961526150694578, + "grad_norm": 1.102015733718872, + "learning_rate": 2.769651362482642e-05, + "loss": 0.4459, + "num_input_tokens_seen": 28336352, + "step": 18625 + }, + { + "epoch": 0.5963126560399462, + "grad_norm": 0.6272205710411072, + "learning_rate": 2.768675321494555e-05, + "loss": 0.7738, + "num_input_tokens_seen": 28344416, + "step": 18630 + }, + { + "epoch": 0.5964726970104347, + "grad_norm": 1.2642261981964111, + "learning_rate": 2.7676992390733565e-05, + "loss": 0.5529, + "num_input_tokens_seen": 28351744, + "step": 18635 + }, + { + "epoch": 0.5966327379809231, + "grad_norm": 1.2677276134490967, + "learning_rate": 2.766723115369571e-05, + "loss": 0.4223, + "num_input_tokens_seen": 28358880, + "step": 18640 + }, + { + "epoch": 0.5967927789514116, + "grad_norm": 0.932051956653595, + "learning_rate": 2.765746950533729e-05, + "loss": 0.5789, + "num_input_tokens_seen": 28366304, + "step": 18645 + }, + { + "epoch": 0.5969528199219, + "grad_norm": 0.4547455608844757, + "learning_rate": 2.7647707447163684e-05, + "loss": 0.3235, + "num_input_tokens_seen": 28373504, + "step": 18650 + }, + { + "epoch": 0.5971128608923885, + "grad_norm": 0.5753258466720581, + "learning_rate": 2.7637944980680315e-05, + "loss": 0.3995, + "num_input_tokens_seen": 28380976, + "step": 18655 + }, + { + "epoch": 0.5972729018628768, + "grad_norm": 0.7102798223495483, + "learning_rate": 2.762818210739268e-05, + "loss": 0.5141, + "num_input_tokens_seen": 28388336, + "step": 18660 + }, + { + "epoch": 0.5974329428333653, + "grad_norm": 1.0671049356460571, + "learning_rate": 2.7618418828806332e-05, + "loss": 0.4651, + "num_input_tokens_seen": 28395792, + "step": 18665 + }, + { + "epoch": 0.5975929838038538, + "grad_norm": 0.7752060890197754, + "learning_rate": 2.76086551464269e-05, + "loss": 0.4042, + "num_input_tokens_seen": 28403072, + "step": 18670 + }, + { + "epoch": 0.5977530247743422, + "grad_norm": 0.5113054513931274, + "learning_rate": 2.759889106176006e-05, + "loss": 0.6449, + "num_input_tokens_seen": 28410800, + "step": 18675 + }, + { + "epoch": 0.5979130657448307, + "grad_norm": 0.8806492686271667, + "learning_rate": 2.758912657631156e-05, + "loss": 0.418, + "num_input_tokens_seen": 28418512, + "step": 18680 + }, + { + "epoch": 0.5980731067153191, + "grad_norm": 0.38753461837768555, + "learning_rate": 2.7579361691587198e-05, + "loss": 0.4771, + "num_input_tokens_seen": 28426224, + "step": 18685 + }, + { + "epoch": 0.5982331476858076, + "grad_norm": 1.3007533550262451, + "learning_rate": 2.756959640909285e-05, + "loss": 0.5519, + "num_input_tokens_seen": 28433776, + "step": 18690 + }, + { + "epoch": 0.598393188656296, + "grad_norm": 0.6578150391578674, + "learning_rate": 2.7559830730334452e-05, + "loss": 0.3423, + "num_input_tokens_seen": 28441504, + "step": 18695 + }, + { + "epoch": 0.5985532296267845, + "grad_norm": 0.670873761177063, + "learning_rate": 2.7550064656817988e-05, + "loss": 0.502, + "num_input_tokens_seen": 28449504, + "step": 18700 + }, + { + "epoch": 0.5987132705972729, + "grad_norm": 0.4731259047985077, + "learning_rate": 2.7540298190049503e-05, + "loss": 0.4725, + "num_input_tokens_seen": 28456864, + "step": 18705 + }, + { + "epoch": 0.5988733115677614, + "grad_norm": 0.9102723002433777, + "learning_rate": 2.7530531331535107e-05, + "loss": 0.6395, + "num_input_tokens_seen": 28464240, + "step": 18710 + }, + { + "epoch": 0.5990333525382497, + "grad_norm": 1.5668542385101318, + "learning_rate": 2.752076408278099e-05, + "loss": 0.4486, + "num_input_tokens_seen": 28472880, + "step": 18715 + }, + { + "epoch": 0.5991933935087382, + "grad_norm": 0.8540098071098328, + "learning_rate": 2.751099644529337e-05, + "loss": 0.6583, + "num_input_tokens_seen": 28480672, + "step": 18720 + }, + { + "epoch": 0.5993534344792267, + "grad_norm": 0.9558483958244324, + "learning_rate": 2.7501228420578533e-05, + "loss": 0.6007, + "num_input_tokens_seen": 28488432, + "step": 18725 + }, + { + "epoch": 0.5995134754497151, + "grad_norm": 0.916144609451294, + "learning_rate": 2.7491460010142857e-05, + "loss": 0.4611, + "num_input_tokens_seen": 28495840, + "step": 18730 + }, + { + "epoch": 0.5996735164202036, + "grad_norm": 0.7876863479614258, + "learning_rate": 2.7481691215492727e-05, + "loss": 0.6418, + "num_input_tokens_seen": 28503680, + "step": 18735 + }, + { + "epoch": 0.599833557390692, + "grad_norm": 0.7881147265434265, + "learning_rate": 2.747192203813463e-05, + "loss": 0.6157, + "num_input_tokens_seen": 28511264, + "step": 18740 + }, + { + "epoch": 0.5999935983611805, + "grad_norm": 0.5796400904655457, + "learning_rate": 2.7462152479575087e-05, + "loss": 0.5437, + "num_input_tokens_seen": 28518416, + "step": 18745 + }, + { + "epoch": 0.6001536393316689, + "grad_norm": 0.42832422256469727, + "learning_rate": 2.7452382541320697e-05, + "loss": 0.3719, + "num_input_tokens_seen": 28525984, + "step": 18750 + }, + { + "epoch": 0.6003136803021574, + "grad_norm": 0.47644951939582825, + "learning_rate": 2.7442612224878096e-05, + "loss": 0.5237, + "num_input_tokens_seen": 28533536, + "step": 18755 + }, + { + "epoch": 0.6004737212726458, + "grad_norm": 0.7756120562553406, + "learning_rate": 2.7432841531753994e-05, + "loss": 0.4223, + "num_input_tokens_seen": 28541280, + "step": 18760 + }, + { + "epoch": 0.6006337622431343, + "grad_norm": 0.3415312170982361, + "learning_rate": 2.7423070463455147e-05, + "loss": 0.4001, + "num_input_tokens_seen": 28548960, + "step": 18765 + }, + { + "epoch": 0.6007938032136226, + "grad_norm": 0.4832223951816559, + "learning_rate": 2.7413299021488397e-05, + "loss": 0.552, + "num_input_tokens_seen": 28556368, + "step": 18770 + }, + { + "epoch": 0.6009538441841111, + "grad_norm": 0.8754349946975708, + "learning_rate": 2.7403527207360615e-05, + "loss": 0.4162, + "num_input_tokens_seen": 28564176, + "step": 18775 + }, + { + "epoch": 0.6011138851545996, + "grad_norm": 0.7604027986526489, + "learning_rate": 2.7393755022578722e-05, + "loss": 0.7268, + "num_input_tokens_seen": 28572176, + "step": 18780 + }, + { + "epoch": 0.601273926125088, + "grad_norm": 1.2365772724151611, + "learning_rate": 2.7383982468649714e-05, + "loss": 0.5312, + "num_input_tokens_seen": 28580304, + "step": 18785 + }, + { + "epoch": 0.6014339670955765, + "grad_norm": 1.4822925329208374, + "learning_rate": 2.7374209547080665e-05, + "loss": 0.5644, + "num_input_tokens_seen": 28588560, + "step": 18790 + }, + { + "epoch": 0.6015940080660649, + "grad_norm": 0.4620225131511688, + "learning_rate": 2.7364436259378663e-05, + "loss": 0.4567, + "num_input_tokens_seen": 28595712, + "step": 18795 + }, + { + "epoch": 0.6017540490365534, + "grad_norm": 1.6580572128295898, + "learning_rate": 2.735466260705088e-05, + "loss": 0.4373, + "num_input_tokens_seen": 28603264, + "step": 18800 + }, + { + "epoch": 0.6017540490365534, + "eval_loss": 0.517553985118866, + "eval_runtime": 560.7798, + "eval_samples_per_second": 24.762, + "eval_steps_per_second": 12.381, + "num_input_tokens_seen": 28603264, + "step": 18800 + }, + { + "epoch": 0.6019140900070418, + "grad_norm": 1.4906089305877686, + "learning_rate": 2.7344888591604524e-05, + "loss": 0.7722, + "num_input_tokens_seen": 28610912, + "step": 18805 + }, + { + "epoch": 0.6020741309775303, + "grad_norm": 0.8567203879356384, + "learning_rate": 2.7335114214546893e-05, + "loss": 0.5699, + "num_input_tokens_seen": 28618512, + "step": 18810 + }, + { + "epoch": 0.6022341719480186, + "grad_norm": 1.12086820602417, + "learning_rate": 2.7325339477385293e-05, + "loss": 0.3713, + "num_input_tokens_seen": 28625984, + "step": 18815 + }, + { + "epoch": 0.6023942129185071, + "grad_norm": 1.0331326723098755, + "learning_rate": 2.7315564381627128e-05, + "loss": 0.3927, + "num_input_tokens_seen": 28633680, + "step": 18820 + }, + { + "epoch": 0.6025542538889955, + "grad_norm": 1.2121564149856567, + "learning_rate": 2.7305788928779835e-05, + "loss": 0.541, + "num_input_tokens_seen": 28641248, + "step": 18825 + }, + { + "epoch": 0.602714294859484, + "grad_norm": 0.5847721695899963, + "learning_rate": 2.729601312035091e-05, + "loss": 0.4785, + "num_input_tokens_seen": 28648640, + "step": 18830 + }, + { + "epoch": 0.6028743358299725, + "grad_norm": 1.1336063146591187, + "learning_rate": 2.7286236957847915e-05, + "loss": 0.6815, + "num_input_tokens_seen": 28656288, + "step": 18835 + }, + { + "epoch": 0.6030343768004609, + "grad_norm": 0.60164874792099, + "learning_rate": 2.7276460442778446e-05, + "loss": 0.4311, + "num_input_tokens_seen": 28664208, + "step": 18840 + }, + { + "epoch": 0.6031944177709494, + "grad_norm": 0.926084578037262, + "learning_rate": 2.726668357665017e-05, + "loss": 0.4948, + "num_input_tokens_seen": 28672272, + "step": 18845 + }, + { + "epoch": 0.6033544587414378, + "grad_norm": 0.6077949404716492, + "learning_rate": 2.7256906360970808e-05, + "loss": 0.5693, + "num_input_tokens_seen": 28679584, + "step": 18850 + }, + { + "epoch": 0.6035144997119263, + "grad_norm": 0.6903151869773865, + "learning_rate": 2.7247128797248117e-05, + "loss": 0.4887, + "num_input_tokens_seen": 28687152, + "step": 18855 + }, + { + "epoch": 0.6036745406824147, + "grad_norm": 0.7983115315437317, + "learning_rate": 2.7237350886989925e-05, + "loss": 0.7396, + "num_input_tokens_seen": 28695296, + "step": 18860 + }, + { + "epoch": 0.6038345816529032, + "grad_norm": 0.7533659338951111, + "learning_rate": 2.7227572631704107e-05, + "loss": 0.4478, + "num_input_tokens_seen": 28703024, + "step": 18865 + }, + { + "epoch": 0.6039946226233915, + "grad_norm": 1.490218162536621, + "learning_rate": 2.7217794032898596e-05, + "loss": 0.6024, + "num_input_tokens_seen": 28710576, + "step": 18870 + }, + { + "epoch": 0.60415466359388, + "grad_norm": 0.9784407019615173, + "learning_rate": 2.7208015092081384e-05, + "loss": 0.4461, + "num_input_tokens_seen": 28717920, + "step": 18875 + }, + { + "epoch": 0.6043147045643685, + "grad_norm": 0.42024049162864685, + "learning_rate": 2.719823581076049e-05, + "loss": 0.4757, + "num_input_tokens_seen": 28725472, + "step": 18880 + }, + { + "epoch": 0.6044747455348569, + "grad_norm": 0.4224807024002075, + "learning_rate": 2.718845619044401e-05, + "loss": 0.3335, + "num_input_tokens_seen": 28732944, + "step": 18885 + }, + { + "epoch": 0.6046347865053454, + "grad_norm": 0.8409940600395203, + "learning_rate": 2.7178676232640088e-05, + "loss": 0.5074, + "num_input_tokens_seen": 28740768, + "step": 18890 + }, + { + "epoch": 0.6047948274758338, + "grad_norm": 0.8269007205963135, + "learning_rate": 2.716889593885691e-05, + "loss": 0.492, + "num_input_tokens_seen": 28748576, + "step": 18895 + }, + { + "epoch": 0.6049548684463223, + "grad_norm": 0.4957093298435211, + "learning_rate": 2.7159115310602716e-05, + "loss": 0.5421, + "num_input_tokens_seen": 28756928, + "step": 18900 + }, + { + "epoch": 0.6051149094168107, + "grad_norm": 1.3509021997451782, + "learning_rate": 2.7149334349385814e-05, + "loss": 0.5237, + "num_input_tokens_seen": 28764448, + "step": 18905 + }, + { + "epoch": 0.6052749503872992, + "grad_norm": 0.6604598760604858, + "learning_rate": 2.713955305671454e-05, + "loss": 0.5119, + "num_input_tokens_seen": 28771616, + "step": 18910 + }, + { + "epoch": 0.6054349913577876, + "grad_norm": 0.815462052822113, + "learning_rate": 2.71297714340973e-05, + "loss": 0.5366, + "num_input_tokens_seen": 28779168, + "step": 18915 + }, + { + "epoch": 0.605595032328276, + "grad_norm": 0.9865866303443909, + "learning_rate": 2.7119989483042545e-05, + "loss": 0.7309, + "num_input_tokens_seen": 28786608, + "step": 18920 + }, + { + "epoch": 0.6057550732987644, + "grad_norm": 0.6895275712013245, + "learning_rate": 2.7110207205058768e-05, + "loss": 0.6505, + "num_input_tokens_seen": 28794320, + "step": 18925 + }, + { + "epoch": 0.6059151142692529, + "grad_norm": 0.6800060868263245, + "learning_rate": 2.7100424601654517e-05, + "loss": 0.5242, + "num_input_tokens_seen": 28801888, + "step": 18930 + }, + { + "epoch": 0.6060751552397414, + "grad_norm": 0.5220323801040649, + "learning_rate": 2.7090641674338403e-05, + "loss": 0.4779, + "num_input_tokens_seen": 28809536, + "step": 18935 + }, + { + "epoch": 0.6062351962102298, + "grad_norm": 0.8101303577423096, + "learning_rate": 2.7080858424619072e-05, + "loss": 0.4708, + "num_input_tokens_seen": 28817280, + "step": 18940 + }, + { + "epoch": 0.6063952371807183, + "grad_norm": 0.7232475280761719, + "learning_rate": 2.707107485400521e-05, + "loss": 0.4796, + "num_input_tokens_seen": 28825008, + "step": 18945 + }, + { + "epoch": 0.6065552781512067, + "grad_norm": 0.6328175663948059, + "learning_rate": 2.7061290964005586e-05, + "loss": 0.4357, + "num_input_tokens_seen": 28832672, + "step": 18950 + }, + { + "epoch": 0.6067153191216952, + "grad_norm": 0.8663139939308167, + "learning_rate": 2.7051506756129e-05, + "loss": 0.5259, + "num_input_tokens_seen": 28840096, + "step": 18955 + }, + { + "epoch": 0.6068753600921836, + "grad_norm": 0.6187529563903809, + "learning_rate": 2.704172223188428e-05, + "loss": 0.6683, + "num_input_tokens_seen": 28848368, + "step": 18960 + }, + { + "epoch": 0.6070354010626721, + "grad_norm": 1.1989248991012573, + "learning_rate": 2.7031937392780334e-05, + "loss": 0.5041, + "num_input_tokens_seen": 28855792, + "step": 18965 + }, + { + "epoch": 0.6071954420331604, + "grad_norm": 0.8786941766738892, + "learning_rate": 2.702215224032611e-05, + "loss": 0.6054, + "num_input_tokens_seen": 28864208, + "step": 18970 + }, + { + "epoch": 0.6073554830036489, + "grad_norm": 1.095348834991455, + "learning_rate": 2.70123667760306e-05, + "loss": 0.5019, + "num_input_tokens_seen": 28871888, + "step": 18975 + }, + { + "epoch": 0.6075155239741373, + "grad_norm": 0.6759410500526428, + "learning_rate": 2.7002581001402845e-05, + "loss": 0.7083, + "num_input_tokens_seen": 28879744, + "step": 18980 + }, + { + "epoch": 0.6076755649446258, + "grad_norm": 1.0713411569595337, + "learning_rate": 2.6992794917951923e-05, + "loss": 0.5048, + "num_input_tokens_seen": 28886928, + "step": 18985 + }, + { + "epoch": 0.6078356059151143, + "grad_norm": 1.8954956531524658, + "learning_rate": 2.6983008527187e-05, + "loss": 0.4931, + "num_input_tokens_seen": 28894944, + "step": 18990 + }, + { + "epoch": 0.6079956468856027, + "grad_norm": 1.0200731754302979, + "learning_rate": 2.697322183061723e-05, + "loss": 0.6874, + "num_input_tokens_seen": 28902080, + "step": 18995 + }, + { + "epoch": 0.6081556878560912, + "grad_norm": 1.577252984046936, + "learning_rate": 2.696343482975186e-05, + "loss": 0.5141, + "num_input_tokens_seen": 28910096, + "step": 19000 + }, + { + "epoch": 0.6081556878560912, + "eval_loss": 0.516933023929596, + "eval_runtime": 560.216, + "eval_samples_per_second": 24.787, + "eval_steps_per_second": 12.393, + "num_input_tokens_seen": 28910096, + "step": 19000 + }, + { + "epoch": 0.6083157288265796, + "grad_norm": 0.56952965259552, + "learning_rate": 2.695364752610016e-05, + "loss": 0.3733, + "num_input_tokens_seen": 28917440, + "step": 19005 + }, + { + "epoch": 0.6084757697970681, + "grad_norm": 0.7459151148796082, + "learning_rate": 2.6943859921171467e-05, + "loss": 0.5667, + "num_input_tokens_seen": 28924832, + "step": 19010 + }, + { + "epoch": 0.6086358107675565, + "grad_norm": 0.8670946955680847, + "learning_rate": 2.6934072016475143e-05, + "loss": 0.7458, + "num_input_tokens_seen": 28933168, + "step": 19015 + }, + { + "epoch": 0.608795851738045, + "grad_norm": 0.6920768618583679, + "learning_rate": 2.6924283813520606e-05, + "loss": 0.4464, + "num_input_tokens_seen": 28941344, + "step": 19020 + }, + { + "epoch": 0.6089558927085333, + "grad_norm": 0.8611333966255188, + "learning_rate": 2.691449531381733e-05, + "loss": 0.6123, + "num_input_tokens_seen": 28948976, + "step": 19025 + }, + { + "epoch": 0.6091159336790218, + "grad_norm": 0.7195848226547241, + "learning_rate": 2.6904706518874816e-05, + "loss": 0.6223, + "num_input_tokens_seen": 28956384, + "step": 19030 + }, + { + "epoch": 0.6092759746495102, + "grad_norm": 0.6941114664077759, + "learning_rate": 2.6894917430202615e-05, + "loss": 0.4089, + "num_input_tokens_seen": 28963872, + "step": 19035 + }, + { + "epoch": 0.6094360156199987, + "grad_norm": 0.5196229815483093, + "learning_rate": 2.6885128049310343e-05, + "loss": 0.5976, + "num_input_tokens_seen": 28971392, + "step": 19040 + }, + { + "epoch": 0.6095960565904872, + "grad_norm": 0.545373797416687, + "learning_rate": 2.687533837770762e-05, + "loss": 0.48, + "num_input_tokens_seen": 28978896, + "step": 19045 + }, + { + "epoch": 0.6097560975609756, + "grad_norm": 0.9409043788909912, + "learning_rate": 2.6865548416904162e-05, + "loss": 0.5939, + "num_input_tokens_seen": 28986672, + "step": 19050 + }, + { + "epoch": 0.6099161385314641, + "grad_norm": 0.6502018570899963, + "learning_rate": 2.68557581684097e-05, + "loss": 0.3676, + "num_input_tokens_seen": 28994528, + "step": 19055 + }, + { + "epoch": 0.6100761795019525, + "grad_norm": 0.6093831062316895, + "learning_rate": 2.6845967633733998e-05, + "loss": 0.3718, + "num_input_tokens_seen": 29002144, + "step": 19060 + }, + { + "epoch": 0.610236220472441, + "grad_norm": 0.6538158655166626, + "learning_rate": 2.683617681438689e-05, + "loss": 0.3488, + "num_input_tokens_seen": 29009600, + "step": 19065 + }, + { + "epoch": 0.6103962614429294, + "grad_norm": 0.5759629607200623, + "learning_rate": 2.682638571187825e-05, + "loss": 0.4759, + "num_input_tokens_seen": 29017200, + "step": 19070 + }, + { + "epoch": 0.6105563024134179, + "grad_norm": 0.6334283947944641, + "learning_rate": 2.6816594327717976e-05, + "loss": 0.5573, + "num_input_tokens_seen": 29024448, + "step": 19075 + }, + { + "epoch": 0.6107163433839062, + "grad_norm": 1.3365546464920044, + "learning_rate": 2.680680266341603e-05, + "loss": 0.5145, + "num_input_tokens_seen": 29032064, + "step": 19080 + }, + { + "epoch": 0.6108763843543947, + "grad_norm": 0.903786838054657, + "learning_rate": 2.67970107204824e-05, + "loss": 0.4963, + "num_input_tokens_seen": 29039376, + "step": 19085 + }, + { + "epoch": 0.6110364253248832, + "grad_norm": 2.1280453205108643, + "learning_rate": 2.6787218500427142e-05, + "loss": 0.5533, + "num_input_tokens_seen": 29047296, + "step": 19090 + }, + { + "epoch": 0.6111964662953716, + "grad_norm": 1.485452651977539, + "learning_rate": 2.6777426004760332e-05, + "loss": 0.5303, + "num_input_tokens_seen": 29055632, + "step": 19095 + }, + { + "epoch": 0.6113565072658601, + "grad_norm": 0.48127806186676025, + "learning_rate": 2.6767633234992094e-05, + "loss": 0.4675, + "num_input_tokens_seen": 29063824, + "step": 19100 + }, + { + "epoch": 0.6115165482363485, + "grad_norm": 1.3364897966384888, + "learning_rate": 2.6757840192632598e-05, + "loss": 0.5919, + "num_input_tokens_seen": 29071552, + "step": 19105 + }, + { + "epoch": 0.611676589206837, + "grad_norm": 0.9298954606056213, + "learning_rate": 2.6748046879192052e-05, + "loss": 0.4697, + "num_input_tokens_seen": 29079296, + "step": 19110 + }, + { + "epoch": 0.6118366301773254, + "grad_norm": 0.8763457536697388, + "learning_rate": 2.673825329618071e-05, + "loss": 0.595, + "num_input_tokens_seen": 29086848, + "step": 19115 + }, + { + "epoch": 0.6119966711478139, + "grad_norm": 1.1203268766403198, + "learning_rate": 2.6728459445108866e-05, + "loss": 0.4557, + "num_input_tokens_seen": 29094704, + "step": 19120 + }, + { + "epoch": 0.6121567121183022, + "grad_norm": 0.6121047735214233, + "learning_rate": 2.6718665327486854e-05, + "loss": 0.4537, + "num_input_tokens_seen": 29102752, + "step": 19125 + }, + { + "epoch": 0.6123167530887907, + "grad_norm": 1.3739875555038452, + "learning_rate": 2.6708870944825048e-05, + "loss": 0.6316, + "num_input_tokens_seen": 29110528, + "step": 19130 + }, + { + "epoch": 0.6124767940592791, + "grad_norm": 0.8088451027870178, + "learning_rate": 2.6699076298633874e-05, + "loss": 0.5935, + "num_input_tokens_seen": 29118400, + "step": 19135 + }, + { + "epoch": 0.6126368350297676, + "grad_norm": 0.7024485468864441, + "learning_rate": 2.6689281390423788e-05, + "loss": 0.4638, + "num_input_tokens_seen": 29126400, + "step": 19140 + }, + { + "epoch": 0.6127968760002561, + "grad_norm": 1.4154118299484253, + "learning_rate": 2.667948622170527e-05, + "loss": 0.4642, + "num_input_tokens_seen": 29134176, + "step": 19145 + }, + { + "epoch": 0.6129569169707445, + "grad_norm": 0.5456987023353577, + "learning_rate": 2.6669690793988873e-05, + "loss": 0.4781, + "num_input_tokens_seen": 29141424, + "step": 19150 + }, + { + "epoch": 0.613116957941233, + "grad_norm": 0.4611887037754059, + "learning_rate": 2.665989510878518e-05, + "loss": 0.3903, + "num_input_tokens_seen": 29149120, + "step": 19155 + }, + { + "epoch": 0.6132769989117214, + "grad_norm": 0.9664293527603149, + "learning_rate": 2.6650099167604793e-05, + "loss": 0.4187, + "num_input_tokens_seen": 29157344, + "step": 19160 + }, + { + "epoch": 0.6134370398822099, + "grad_norm": 0.4008118212223053, + "learning_rate": 2.6640302971958376e-05, + "loss": 0.4575, + "num_input_tokens_seen": 29164208, + "step": 19165 + }, + { + "epoch": 0.6135970808526983, + "grad_norm": 1.8786156177520752, + "learning_rate": 2.6630506523356635e-05, + "loss": 0.5946, + "num_input_tokens_seen": 29171680, + "step": 19170 + }, + { + "epoch": 0.6137571218231868, + "grad_norm": 1.2616944313049316, + "learning_rate": 2.6620709823310297e-05, + "loss": 0.6316, + "num_input_tokens_seen": 29180288, + "step": 19175 + }, + { + "epoch": 0.6139171627936751, + "grad_norm": 0.49041101336479187, + "learning_rate": 2.661091287333014e-05, + "loss": 0.3352, + "num_input_tokens_seen": 29187792, + "step": 19180 + }, + { + "epoch": 0.6140772037641636, + "grad_norm": 0.6078844666481018, + "learning_rate": 2.660111567492696e-05, + "loss": 0.6491, + "num_input_tokens_seen": 29195616, + "step": 19185 + }, + { + "epoch": 0.614237244734652, + "grad_norm": 0.6152863502502441, + "learning_rate": 2.6591318229611635e-05, + "loss": 0.3886, + "num_input_tokens_seen": 29202992, + "step": 19190 + }, + { + "epoch": 0.6143972857051405, + "grad_norm": 0.4031214416027069, + "learning_rate": 2.6581520538895037e-05, + "loss": 0.6309, + "num_input_tokens_seen": 29210736, + "step": 19195 + }, + { + "epoch": 0.614557326675629, + "grad_norm": 1.096476674079895, + "learning_rate": 2.6571722604288102e-05, + "loss": 0.562, + "num_input_tokens_seen": 29218624, + "step": 19200 + }, + { + "epoch": 0.614557326675629, + "eval_loss": 0.5162514448165894, + "eval_runtime": 560.2977, + "eval_samples_per_second": 24.783, + "eval_steps_per_second": 12.392, + "num_input_tokens_seen": 29218624, + "step": 19200 + }, + { + "epoch": 0.6147173676461174, + "grad_norm": 1.5048482418060303, + "learning_rate": 2.656192442730179e-05, + "loss": 0.5943, + "num_input_tokens_seen": 29226512, + "step": 19205 + }, + { + "epoch": 0.6148774086166059, + "grad_norm": 0.7627814412117004, + "learning_rate": 2.6552126009447098e-05, + "loss": 0.471, + "num_input_tokens_seen": 29234224, + "step": 19210 + }, + { + "epoch": 0.6150374495870943, + "grad_norm": 0.6448109149932861, + "learning_rate": 2.654232735223507e-05, + "loss": 0.559, + "num_input_tokens_seen": 29242064, + "step": 19215 + }, + { + "epoch": 0.6151974905575828, + "grad_norm": 1.5012966394424438, + "learning_rate": 2.6532528457176787e-05, + "loss": 0.5371, + "num_input_tokens_seen": 29249936, + "step": 19220 + }, + { + "epoch": 0.6153575315280712, + "grad_norm": 2.8480336666107178, + "learning_rate": 2.6522729325783348e-05, + "loss": 0.5593, + "num_input_tokens_seen": 29257568, + "step": 19225 + }, + { + "epoch": 0.6155175724985597, + "grad_norm": 0.4931820333003998, + "learning_rate": 2.6512929959565914e-05, + "loss": 0.4757, + "num_input_tokens_seen": 29264864, + "step": 19230 + }, + { + "epoch": 0.615677613469048, + "grad_norm": 0.7447841763496399, + "learning_rate": 2.6503130360035673e-05, + "loss": 0.7866, + "num_input_tokens_seen": 29272256, + "step": 19235 + }, + { + "epoch": 0.6158376544395365, + "grad_norm": 1.678725004196167, + "learning_rate": 2.6493330528703835e-05, + "loss": 0.4227, + "num_input_tokens_seen": 29279584, + "step": 19240 + }, + { + "epoch": 0.6159976954100249, + "grad_norm": 1.2069655656814575, + "learning_rate": 2.648353046708167e-05, + "loss": 0.5653, + "num_input_tokens_seen": 29287120, + "step": 19245 + }, + { + "epoch": 0.6161577363805134, + "grad_norm": 0.5089053511619568, + "learning_rate": 2.647373017668046e-05, + "loss": 0.6142, + "num_input_tokens_seen": 29294512, + "step": 19250 + }, + { + "epoch": 0.6163177773510019, + "grad_norm": 0.960587203502655, + "learning_rate": 2.6463929659011537e-05, + "loss": 0.4803, + "num_input_tokens_seen": 29302080, + "step": 19255 + }, + { + "epoch": 0.6164778183214903, + "grad_norm": 1.1925920248031616, + "learning_rate": 2.6454128915586262e-05, + "loss": 0.4481, + "num_input_tokens_seen": 29309600, + "step": 19260 + }, + { + "epoch": 0.6166378592919788, + "grad_norm": 1.186102032661438, + "learning_rate": 2.6444327947916036e-05, + "loss": 0.5667, + "num_input_tokens_seen": 29317328, + "step": 19265 + }, + { + "epoch": 0.6167979002624672, + "grad_norm": 0.7163619995117188, + "learning_rate": 2.6434526757512292e-05, + "loss": 0.477, + "num_input_tokens_seen": 29324768, + "step": 19270 + }, + { + "epoch": 0.6169579412329557, + "grad_norm": 1.706009864807129, + "learning_rate": 2.6424725345886486e-05, + "loss": 0.5432, + "num_input_tokens_seen": 29332336, + "step": 19275 + }, + { + "epoch": 0.617117982203444, + "grad_norm": 0.9040642976760864, + "learning_rate": 2.641492371455014e-05, + "loss": 0.6028, + "num_input_tokens_seen": 29340192, + "step": 19280 + }, + { + "epoch": 0.6172780231739325, + "grad_norm": 1.141499400138855, + "learning_rate": 2.640512186501477e-05, + "loss": 0.5531, + "num_input_tokens_seen": 29347216, + "step": 19285 + }, + { + "epoch": 0.6174380641444209, + "grad_norm": 0.5464344024658203, + "learning_rate": 2.639531979879195e-05, + "loss": 0.3599, + "num_input_tokens_seen": 29354816, + "step": 19290 + }, + { + "epoch": 0.6175981051149094, + "grad_norm": 0.5185766220092773, + "learning_rate": 2.638551751739328e-05, + "loss": 0.5355, + "num_input_tokens_seen": 29362208, + "step": 19295 + }, + { + "epoch": 0.6177581460853978, + "grad_norm": 0.7831389307975769, + "learning_rate": 2.6375715022330404e-05, + "loss": 0.7113, + "num_input_tokens_seen": 29369360, + "step": 19300 + }, + { + "epoch": 0.6179181870558863, + "grad_norm": 0.6621671915054321, + "learning_rate": 2.6365912315114976e-05, + "loss": 0.3979, + "num_input_tokens_seen": 29377328, + "step": 19305 + }, + { + "epoch": 0.6180782280263748, + "grad_norm": 0.8131052851676941, + "learning_rate": 2.6356109397258704e-05, + "loss": 0.4297, + "num_input_tokens_seen": 29385344, + "step": 19310 + }, + { + "epoch": 0.6182382689968632, + "grad_norm": 0.8247477412223816, + "learning_rate": 2.6346306270273325e-05, + "loss": 0.5239, + "num_input_tokens_seen": 29393792, + "step": 19315 + }, + { + "epoch": 0.6183983099673517, + "grad_norm": 0.547673761844635, + "learning_rate": 2.6336502935670608e-05, + "loss": 0.5311, + "num_input_tokens_seen": 29401584, + "step": 19320 + }, + { + "epoch": 0.6185583509378401, + "grad_norm": 3.0719027519226074, + "learning_rate": 2.6326699394962333e-05, + "loss": 0.8594, + "num_input_tokens_seen": 29409600, + "step": 19325 + }, + { + "epoch": 0.6187183919083286, + "grad_norm": 1.304166555404663, + "learning_rate": 2.6316895649660334e-05, + "loss": 0.6006, + "num_input_tokens_seen": 29417280, + "step": 19330 + }, + { + "epoch": 0.6188784328788169, + "grad_norm": 1.0435717105865479, + "learning_rate": 2.6307091701276486e-05, + "loss": 0.4932, + "num_input_tokens_seen": 29425328, + "step": 19335 + }, + { + "epoch": 0.6190384738493054, + "grad_norm": 0.9618701338768005, + "learning_rate": 2.629728755132267e-05, + "loss": 0.4862, + "num_input_tokens_seen": 29432928, + "step": 19340 + }, + { + "epoch": 0.6191985148197938, + "grad_norm": 1.0201139450073242, + "learning_rate": 2.628748320131081e-05, + "loss": 0.485, + "num_input_tokens_seen": 29440992, + "step": 19345 + }, + { + "epoch": 0.6193585557902823, + "grad_norm": 1.8321254253387451, + "learning_rate": 2.6277678652752856e-05, + "loss": 0.7062, + "num_input_tokens_seen": 29448960, + "step": 19350 + }, + { + "epoch": 0.6195185967607708, + "grad_norm": 0.7015206813812256, + "learning_rate": 2.6267873907160807e-05, + "loss": 0.5442, + "num_input_tokens_seen": 29456368, + "step": 19355 + }, + { + "epoch": 0.6196786377312592, + "grad_norm": 0.6490680575370789, + "learning_rate": 2.6258068966046668e-05, + "loss": 0.5653, + "num_input_tokens_seen": 29463728, + "step": 19360 + }, + { + "epoch": 0.6198386787017477, + "grad_norm": 1.1025195121765137, + "learning_rate": 2.6248263830922475e-05, + "loss": 0.7357, + "num_input_tokens_seen": 29470672, + "step": 19365 + }, + { + "epoch": 0.6199987196722361, + "grad_norm": 0.48530831933021545, + "learning_rate": 2.6238458503300318e-05, + "loss": 0.403, + "num_input_tokens_seen": 29477728, + "step": 19370 + }, + { + "epoch": 0.6201587606427246, + "grad_norm": 0.6886855959892273, + "learning_rate": 2.6228652984692292e-05, + "loss": 0.4129, + "num_input_tokens_seen": 29484960, + "step": 19375 + }, + { + "epoch": 0.620318801613213, + "grad_norm": 0.6914053559303284, + "learning_rate": 2.621884727661054e-05, + "loss": 0.5053, + "num_input_tokens_seen": 29492256, + "step": 19380 + }, + { + "epoch": 0.6204788425837015, + "grad_norm": 1.0817285776138306, + "learning_rate": 2.6209041380567222e-05, + "loss": 0.635, + "num_input_tokens_seen": 29499632, + "step": 19385 + }, + { + "epoch": 0.6206388835541898, + "grad_norm": 0.5609902143478394, + "learning_rate": 2.6199235298074527e-05, + "loss": 0.449, + "num_input_tokens_seen": 29507024, + "step": 19390 + }, + { + "epoch": 0.6207989245246783, + "grad_norm": 0.8436936140060425, + "learning_rate": 2.618942903064468e-05, + "loss": 0.3849, + "num_input_tokens_seen": 29514352, + "step": 19395 + }, + { + "epoch": 0.6209589654951667, + "grad_norm": 0.2897673547267914, + "learning_rate": 2.6179622579789932e-05, + "loss": 0.4593, + "num_input_tokens_seen": 29521824, + "step": 19400 + }, + { + "epoch": 0.6209589654951667, + "eval_loss": 0.5155550837516785, + "eval_runtime": 560.4741, + "eval_samples_per_second": 24.775, + "eval_steps_per_second": 12.388, + "num_input_tokens_seen": 29521824, + "step": 19400 + }, + { + "epoch": 0.6211190064656552, + "grad_norm": 0.7350932359695435, + "learning_rate": 2.6169815947022553e-05, + "loss": 0.5603, + "num_input_tokens_seen": 29529536, + "step": 19405 + }, + { + "epoch": 0.6212790474361437, + "grad_norm": 0.895028829574585, + "learning_rate": 2.6160009133854853e-05, + "loss": 0.3779, + "num_input_tokens_seen": 29537552, + "step": 19410 + }, + { + "epoch": 0.6214390884066321, + "grad_norm": 0.5730772018432617, + "learning_rate": 2.6150202141799168e-05, + "loss": 0.4689, + "num_input_tokens_seen": 29545008, + "step": 19415 + }, + { + "epoch": 0.6215991293771206, + "grad_norm": 0.5746307969093323, + "learning_rate": 2.614039497236786e-05, + "loss": 0.5032, + "num_input_tokens_seen": 29552608, + "step": 19420 + }, + { + "epoch": 0.621759170347609, + "grad_norm": 0.9090111255645752, + "learning_rate": 2.6130587627073315e-05, + "loss": 0.4849, + "num_input_tokens_seen": 29560256, + "step": 19425 + }, + { + "epoch": 0.6219192113180975, + "grad_norm": 0.6857851147651672, + "learning_rate": 2.6120780107427956e-05, + "loss": 0.3868, + "num_input_tokens_seen": 29568048, + "step": 19430 + }, + { + "epoch": 0.6220792522885858, + "grad_norm": 1.125645399093628, + "learning_rate": 2.6110972414944214e-05, + "loss": 0.3862, + "num_input_tokens_seen": 29575696, + "step": 19435 + }, + { + "epoch": 0.6222392932590743, + "grad_norm": 0.47701123356819153, + "learning_rate": 2.6101164551134565e-05, + "loss": 0.5189, + "num_input_tokens_seen": 29583152, + "step": 19440 + }, + { + "epoch": 0.6223993342295627, + "grad_norm": 0.5833953619003296, + "learning_rate": 2.6091356517511505e-05, + "loss": 0.4893, + "num_input_tokens_seen": 29590272, + "step": 19445 + }, + { + "epoch": 0.6225593752000512, + "grad_norm": 0.5491629242897034, + "learning_rate": 2.608154831558755e-05, + "loss": 0.4347, + "num_input_tokens_seen": 29598000, + "step": 19450 + }, + { + "epoch": 0.6227194161705396, + "grad_norm": 0.792246401309967, + "learning_rate": 2.607173994687526e-05, + "loss": 0.6357, + "num_input_tokens_seen": 29605648, + "step": 19455 + }, + { + "epoch": 0.6228794571410281, + "grad_norm": 0.3697579801082611, + "learning_rate": 2.6061931412887196e-05, + "loss": 0.519, + "num_input_tokens_seen": 29613056, + "step": 19460 + }, + { + "epoch": 0.6230394981115166, + "grad_norm": 0.4765170216560364, + "learning_rate": 2.6052122715135973e-05, + "loss": 0.4388, + "num_input_tokens_seen": 29620896, + "step": 19465 + }, + { + "epoch": 0.623199539082005, + "grad_norm": 1.93509840965271, + "learning_rate": 2.60423138551342e-05, + "loss": 0.4873, + "num_input_tokens_seen": 29628592, + "step": 19470 + }, + { + "epoch": 0.6233595800524935, + "grad_norm": 0.7846272587776184, + "learning_rate": 2.6032504834394527e-05, + "loss": 0.5121, + "num_input_tokens_seen": 29636272, + "step": 19475 + }, + { + "epoch": 0.6235196210229819, + "grad_norm": 0.7216150760650635, + "learning_rate": 2.602269565442964e-05, + "loss": 0.5587, + "num_input_tokens_seen": 29643840, + "step": 19480 + }, + { + "epoch": 0.6236796619934704, + "grad_norm": 0.6572282314300537, + "learning_rate": 2.6012886316752227e-05, + "loss": 0.4698, + "num_input_tokens_seen": 29651200, + "step": 19485 + }, + { + "epoch": 0.6238397029639587, + "grad_norm": 0.4949203133583069, + "learning_rate": 2.6003076822875018e-05, + "loss": 0.5201, + "num_input_tokens_seen": 29658704, + "step": 19490 + }, + { + "epoch": 0.6239997439344472, + "grad_norm": 0.9146119952201843, + "learning_rate": 2.5993267174310755e-05, + "loss": 0.4751, + "num_input_tokens_seen": 29665856, + "step": 19495 + }, + { + "epoch": 0.6241597849049356, + "grad_norm": 0.4747484624385834, + "learning_rate": 2.5983457372572218e-05, + "loss": 0.3328, + "num_input_tokens_seen": 29673296, + "step": 19500 + }, + { + "epoch": 0.6243198258754241, + "grad_norm": 0.5953177809715271, + "learning_rate": 2.597364741917219e-05, + "loss": 0.4149, + "num_input_tokens_seen": 29681072, + "step": 19505 + }, + { + "epoch": 0.6244798668459125, + "grad_norm": 0.6055886745452881, + "learning_rate": 2.5963837315623492e-05, + "loss": 0.5732, + "num_input_tokens_seen": 29687952, + "step": 19510 + }, + { + "epoch": 0.624639907816401, + "grad_norm": 0.6878328323364258, + "learning_rate": 2.595402706343897e-05, + "loss": 0.381, + "num_input_tokens_seen": 29695840, + "step": 19515 + }, + { + "epoch": 0.6247999487868895, + "grad_norm": 1.0562396049499512, + "learning_rate": 2.594421666413148e-05, + "loss": 0.588, + "num_input_tokens_seen": 29702944, + "step": 19520 + }, + { + "epoch": 0.6249599897573779, + "grad_norm": 0.6216396689414978, + "learning_rate": 2.5934406119213928e-05, + "loss": 0.4483, + "num_input_tokens_seen": 29711184, + "step": 19525 + }, + { + "epoch": 0.6251200307278664, + "grad_norm": 0.5166635513305664, + "learning_rate": 2.5924595430199193e-05, + "loss": 0.5039, + "num_input_tokens_seen": 29719008, + "step": 19530 + }, + { + "epoch": 0.6252800716983548, + "grad_norm": 0.5055813193321228, + "learning_rate": 2.5914784598600238e-05, + "loss": 0.6166, + "num_input_tokens_seen": 29726496, + "step": 19535 + }, + { + "epoch": 0.6254401126688433, + "grad_norm": 1.88589346408844, + "learning_rate": 2.5904973625930002e-05, + "loss": 0.4172, + "num_input_tokens_seen": 29734816, + "step": 19540 + }, + { + "epoch": 0.6256001536393316, + "grad_norm": 0.7869886755943298, + "learning_rate": 2.5895162513701456e-05, + "loss": 0.3977, + "num_input_tokens_seen": 29742032, + "step": 19545 + }, + { + "epoch": 0.6257601946098201, + "grad_norm": 0.35729357600212097, + "learning_rate": 2.5885351263427593e-05, + "loss": 0.2449, + "num_input_tokens_seen": 29750048, + "step": 19550 + }, + { + "epoch": 0.6259202355803085, + "grad_norm": 0.45554620027542114, + "learning_rate": 2.5875539876621448e-05, + "loss": 0.4108, + "num_input_tokens_seen": 29757808, + "step": 19555 + }, + { + "epoch": 0.626080276550797, + "grad_norm": 0.45330610871315, + "learning_rate": 2.586572835479605e-05, + "loss": 0.3359, + "num_input_tokens_seen": 29764944, + "step": 19560 + }, + { + "epoch": 0.6262403175212854, + "grad_norm": 0.5138568878173828, + "learning_rate": 2.585591669946446e-05, + "loss": 0.5535, + "num_input_tokens_seen": 29772656, + "step": 19565 + }, + { + "epoch": 0.6264003584917739, + "grad_norm": 0.6002029180526733, + "learning_rate": 2.5846104912139756e-05, + "loss": 0.5141, + "num_input_tokens_seen": 29780448, + "step": 19570 + }, + { + "epoch": 0.6265603994622624, + "grad_norm": 0.7393360733985901, + "learning_rate": 2.583629299433505e-05, + "loss": 0.3631, + "num_input_tokens_seen": 29787904, + "step": 19575 + }, + { + "epoch": 0.6267204404327508, + "grad_norm": 0.821391761302948, + "learning_rate": 2.582648094756345e-05, + "loss": 0.5428, + "num_input_tokens_seen": 29795056, + "step": 19580 + }, + { + "epoch": 0.6268804814032393, + "grad_norm": 0.7630119323730469, + "learning_rate": 2.5816668773338098e-05, + "loss": 0.5886, + "num_input_tokens_seen": 29802640, + "step": 19585 + }, + { + "epoch": 0.6270405223737276, + "grad_norm": 1.3078815937042236, + "learning_rate": 2.580685647317216e-05, + "loss": 0.5401, + "num_input_tokens_seen": 29810416, + "step": 19590 + }, + { + "epoch": 0.6272005633442161, + "grad_norm": 2.2403669357299805, + "learning_rate": 2.5797044048578818e-05, + "loss": 0.3287, + "num_input_tokens_seen": 29818112, + "step": 19595 + }, + { + "epoch": 0.6273606043147045, + "grad_norm": 0.5479199886322021, + "learning_rate": 2.5787231501071262e-05, + "loss": 0.5216, + "num_input_tokens_seen": 29825872, + "step": 19600 + }, + { + "epoch": 0.6273606043147045, + "eval_loss": 0.5154410600662231, + "eval_runtime": 560.3363, + "eval_samples_per_second": 24.782, + "eval_steps_per_second": 12.391, + "num_input_tokens_seen": 29825872, + "step": 19600 + }, + { + "epoch": 0.627520645285193, + "grad_norm": 0.7580716609954834, + "learning_rate": 2.577741883216272e-05, + "loss": 0.5361, + "num_input_tokens_seen": 29833424, + "step": 19605 + }, + { + "epoch": 0.6276806862556814, + "grad_norm": 0.974260151386261, + "learning_rate": 2.576760604336642e-05, + "loss": 0.539, + "num_input_tokens_seen": 29840720, + "step": 19610 + }, + { + "epoch": 0.6278407272261699, + "grad_norm": 0.6878101229667664, + "learning_rate": 2.575779313619563e-05, + "loss": 0.4785, + "num_input_tokens_seen": 29847968, + "step": 19615 + }, + { + "epoch": 0.6280007681966584, + "grad_norm": 0.6676644682884216, + "learning_rate": 2.5747980112163605e-05, + "loss": 0.4003, + "num_input_tokens_seen": 29855232, + "step": 19620 + }, + { + "epoch": 0.6281608091671468, + "grad_norm": 0.44286423921585083, + "learning_rate": 2.5738166972783656e-05, + "loss": 0.4147, + "num_input_tokens_seen": 29862752, + "step": 19625 + }, + { + "epoch": 0.6283208501376353, + "grad_norm": 1.014763355255127, + "learning_rate": 2.5728353719569075e-05, + "loss": 0.5694, + "num_input_tokens_seen": 29870624, + "step": 19630 + }, + { + "epoch": 0.6284808911081237, + "grad_norm": 0.5245571732521057, + "learning_rate": 2.57185403540332e-05, + "loss": 0.4111, + "num_input_tokens_seen": 29878096, + "step": 19635 + }, + { + "epoch": 0.6286409320786122, + "grad_norm": 0.5148958563804626, + "learning_rate": 2.5708726877689375e-05, + "loss": 0.4905, + "num_input_tokens_seen": 29885392, + "step": 19640 + }, + { + "epoch": 0.6288009730491005, + "grad_norm": 0.5871463418006897, + "learning_rate": 2.5698913292050964e-05, + "loss": 0.4398, + "num_input_tokens_seen": 29892800, + "step": 19645 + }, + { + "epoch": 0.628961014019589, + "grad_norm": 1.298873782157898, + "learning_rate": 2.568909959863133e-05, + "loss": 0.5385, + "num_input_tokens_seen": 29900224, + "step": 19650 + }, + { + "epoch": 0.6291210549900774, + "grad_norm": 0.5667327046394348, + "learning_rate": 2.5679285798943887e-05, + "loss": 0.5121, + "num_input_tokens_seen": 29908016, + "step": 19655 + }, + { + "epoch": 0.6292810959605659, + "grad_norm": 1.0833271741867065, + "learning_rate": 2.5669471894502035e-05, + "loss": 0.5528, + "num_input_tokens_seen": 29914896, + "step": 19660 + }, + { + "epoch": 0.6294411369310543, + "grad_norm": 0.8431330919265747, + "learning_rate": 2.56596578868192e-05, + "loss": 0.5877, + "num_input_tokens_seen": 29922240, + "step": 19665 + }, + { + "epoch": 0.6296011779015428, + "grad_norm": 1.0654703378677368, + "learning_rate": 2.564984377740883e-05, + "loss": 0.498, + "num_input_tokens_seen": 29930032, + "step": 19670 + }, + { + "epoch": 0.6297612188720313, + "grad_norm": 0.6743608713150024, + "learning_rate": 2.564002956778438e-05, + "loss": 0.4358, + "num_input_tokens_seen": 29937536, + "step": 19675 + }, + { + "epoch": 0.6299212598425197, + "grad_norm": 1.433366060256958, + "learning_rate": 2.563021525945934e-05, + "loss": 0.5801, + "num_input_tokens_seen": 29944720, + "step": 19680 + }, + { + "epoch": 0.6300813008130082, + "grad_norm": 1.4752240180969238, + "learning_rate": 2.562040085394718e-05, + "loss": 0.5964, + "num_input_tokens_seen": 29951984, + "step": 19685 + }, + { + "epoch": 0.6302413417834966, + "grad_norm": 0.7046123147010803, + "learning_rate": 2.56105863527614e-05, + "loss": 0.3942, + "num_input_tokens_seen": 29959616, + "step": 19690 + }, + { + "epoch": 0.630401382753985, + "grad_norm": 0.6197466254234314, + "learning_rate": 2.5600771757415548e-05, + "loss": 0.5735, + "num_input_tokens_seen": 29967200, + "step": 19695 + }, + { + "epoch": 0.6305614237244734, + "grad_norm": 0.8038577437400818, + "learning_rate": 2.5590957069423134e-05, + "loss": 0.5589, + "num_input_tokens_seen": 29974720, + "step": 19700 + }, + { + "epoch": 0.6307214646949619, + "grad_norm": 0.5527219772338867, + "learning_rate": 2.5581142290297716e-05, + "loss": 0.2341, + "num_input_tokens_seen": 29982320, + "step": 19705 + }, + { + "epoch": 0.6308815056654503, + "grad_norm": 0.3090932071208954, + "learning_rate": 2.557132742155285e-05, + "loss": 0.3939, + "num_input_tokens_seen": 29989648, + "step": 19710 + }, + { + "epoch": 0.6310415466359388, + "grad_norm": 1.4506784677505493, + "learning_rate": 2.556151246470212e-05, + "loss": 0.4713, + "num_input_tokens_seen": 29997600, + "step": 19715 + }, + { + "epoch": 0.6312015876064272, + "grad_norm": 0.46531352400779724, + "learning_rate": 2.5551697421259114e-05, + "loss": 0.4719, + "num_input_tokens_seen": 30005312, + "step": 19720 + }, + { + "epoch": 0.6313616285769157, + "grad_norm": 0.577396810054779, + "learning_rate": 2.554188229273743e-05, + "loss": 0.4128, + "num_input_tokens_seen": 30012928, + "step": 19725 + }, + { + "epoch": 0.6315216695474042, + "grad_norm": 0.38224083185195923, + "learning_rate": 2.5532067080650678e-05, + "loss": 0.4882, + "num_input_tokens_seen": 30020656, + "step": 19730 + }, + { + "epoch": 0.6316817105178926, + "grad_norm": 1.1932907104492188, + "learning_rate": 2.55222517865125e-05, + "loss": 0.491, + "num_input_tokens_seen": 30028496, + "step": 19735 + }, + { + "epoch": 0.6318417514883811, + "grad_norm": 0.46713659167289734, + "learning_rate": 2.5512436411836538e-05, + "loss": 0.3267, + "num_input_tokens_seen": 30037184, + "step": 19740 + }, + { + "epoch": 0.6320017924588694, + "grad_norm": 0.5124878883361816, + "learning_rate": 2.5502620958136443e-05, + "loss": 0.5421, + "num_input_tokens_seen": 30044656, + "step": 19745 + }, + { + "epoch": 0.6321618334293579, + "grad_norm": 0.8572183847427368, + "learning_rate": 2.5492805426925874e-05, + "loss": 0.5029, + "num_input_tokens_seen": 30051936, + "step": 19750 + }, + { + "epoch": 0.6323218743998463, + "grad_norm": 0.8235751390457153, + "learning_rate": 2.5482989819718523e-05, + "loss": 0.5329, + "num_input_tokens_seen": 30060064, + "step": 19755 + }, + { + "epoch": 0.6324819153703348, + "grad_norm": 1.3571428060531616, + "learning_rate": 2.5473174138028065e-05, + "loss": 0.4813, + "num_input_tokens_seen": 30067568, + "step": 19760 + }, + { + "epoch": 0.6326419563408232, + "grad_norm": 0.4821469485759735, + "learning_rate": 2.5463358383368212e-05, + "loss": 0.5545, + "num_input_tokens_seen": 30074784, + "step": 19765 + }, + { + "epoch": 0.6328019973113117, + "grad_norm": 0.6280540823936462, + "learning_rate": 2.545354255725267e-05, + "loss": 0.5111, + "num_input_tokens_seen": 30082384, + "step": 19770 + }, + { + "epoch": 0.6329620382818001, + "grad_norm": 0.8585426211357117, + "learning_rate": 2.5443726661195165e-05, + "loss": 0.6499, + "num_input_tokens_seen": 30089808, + "step": 19775 + }, + { + "epoch": 0.6331220792522886, + "grad_norm": 0.5466024279594421, + "learning_rate": 2.543391069670944e-05, + "loss": 0.4601, + "num_input_tokens_seen": 30097584, + "step": 19780 + }, + { + "epoch": 0.6332821202227771, + "grad_norm": 0.7045871615409851, + "learning_rate": 2.5424094665309228e-05, + "loss": 0.5281, + "num_input_tokens_seen": 30105264, + "step": 19785 + }, + { + "epoch": 0.6334421611932655, + "grad_norm": 0.9428433775901794, + "learning_rate": 2.5414278568508292e-05, + "loss": 0.6276, + "num_input_tokens_seen": 30113360, + "step": 19790 + }, + { + "epoch": 0.633602202163754, + "grad_norm": 0.7673274278640747, + "learning_rate": 2.540446240782039e-05, + "loss": 0.4646, + "num_input_tokens_seen": 30121232, + "step": 19795 + }, + { + "epoch": 0.6337622431342423, + "grad_norm": 0.7234642505645752, + "learning_rate": 2.5394646184759307e-05, + "loss": 0.4665, + "num_input_tokens_seen": 30128800, + "step": 19800 + }, + { + "epoch": 0.6337622431342423, + "eval_loss": 0.5146429538726807, + "eval_runtime": 560.3482, + "eval_samples_per_second": 24.781, + "eval_steps_per_second": 12.391, + "num_input_tokens_seen": 30128800, + "step": 19800 + }, + { + "epoch": 0.6339222841047308, + "grad_norm": 0.9278604984283447, + "learning_rate": 2.538482990083882e-05, + "loss": 0.4667, + "num_input_tokens_seen": 30136272, + "step": 19805 + }, + { + "epoch": 0.6340823250752192, + "grad_norm": 0.4919201135635376, + "learning_rate": 2.5375013557572725e-05, + "loss": 0.3114, + "num_input_tokens_seen": 30143280, + "step": 19810 + }, + { + "epoch": 0.6342423660457077, + "grad_norm": 0.5463968515396118, + "learning_rate": 2.536519715647483e-05, + "loss": 0.3561, + "num_input_tokens_seen": 30150992, + "step": 19815 + }, + { + "epoch": 0.6344024070161961, + "grad_norm": 1.0248537063598633, + "learning_rate": 2.535538069905894e-05, + "loss": 0.4922, + "num_input_tokens_seen": 30158448, + "step": 19820 + }, + { + "epoch": 0.6345624479866846, + "grad_norm": 1.2100462913513184, + "learning_rate": 2.534556418683888e-05, + "loss": 0.5384, + "num_input_tokens_seen": 30165696, + "step": 19825 + }, + { + "epoch": 0.634722488957173, + "grad_norm": 0.6754421591758728, + "learning_rate": 2.5335747621328486e-05, + "loss": 0.4225, + "num_input_tokens_seen": 30173760, + "step": 19830 + }, + { + "epoch": 0.6348825299276615, + "grad_norm": 1.2082200050354004, + "learning_rate": 2.5325931004041586e-05, + "loss": 0.4886, + "num_input_tokens_seen": 30181104, + "step": 19835 + }, + { + "epoch": 0.63504257089815, + "grad_norm": 0.5573235750198364, + "learning_rate": 2.5316114336492032e-05, + "loss": 0.4966, + "num_input_tokens_seen": 30188320, + "step": 19840 + }, + { + "epoch": 0.6352026118686384, + "grad_norm": 0.620451807975769, + "learning_rate": 2.530629762019367e-05, + "loss": 0.5404, + "num_input_tokens_seen": 30196144, + "step": 19845 + }, + { + "epoch": 0.6353626528391269, + "grad_norm": 0.8587017059326172, + "learning_rate": 2.5296480856660364e-05, + "loss": 0.5752, + "num_input_tokens_seen": 30203680, + "step": 19850 + }, + { + "epoch": 0.6355226938096152, + "grad_norm": 2.1676487922668457, + "learning_rate": 2.528666404740599e-05, + "loss": 0.6731, + "num_input_tokens_seen": 30211360, + "step": 19855 + }, + { + "epoch": 0.6356827347801037, + "grad_norm": 0.5222254395484924, + "learning_rate": 2.527684719394442e-05, + "loss": 0.3598, + "num_input_tokens_seen": 30218560, + "step": 19860 + }, + { + "epoch": 0.6358427757505921, + "grad_norm": 0.9383366107940674, + "learning_rate": 2.526703029778953e-05, + "loss": 0.3777, + "num_input_tokens_seen": 30226912, + "step": 19865 + }, + { + "epoch": 0.6360028167210806, + "grad_norm": 0.6402169466018677, + "learning_rate": 2.5257213360455208e-05, + "loss": 0.5626, + "num_input_tokens_seen": 30234128, + "step": 19870 + }, + { + "epoch": 0.636162857691569, + "grad_norm": 0.6390893459320068, + "learning_rate": 2.5247396383455353e-05, + "loss": 0.7214, + "num_input_tokens_seen": 30241248, + "step": 19875 + }, + { + "epoch": 0.6363228986620575, + "grad_norm": 1.1777695417404175, + "learning_rate": 2.523757936830387e-05, + "loss": 0.6146, + "num_input_tokens_seen": 30248752, + "step": 19880 + }, + { + "epoch": 0.636482939632546, + "grad_norm": 0.68865567445755, + "learning_rate": 2.5227762316514662e-05, + "loss": 0.4639, + "num_input_tokens_seen": 30256080, + "step": 19885 + }, + { + "epoch": 0.6366429806030344, + "grad_norm": 0.566741406917572, + "learning_rate": 2.5217945229601648e-05, + "loss": 0.5178, + "num_input_tokens_seen": 30263568, + "step": 19890 + }, + { + "epoch": 0.6368030215735229, + "grad_norm": 0.7086255550384521, + "learning_rate": 2.5208128109078738e-05, + "loss": 0.4054, + "num_input_tokens_seen": 30271184, + "step": 19895 + }, + { + "epoch": 0.6369630625440112, + "grad_norm": 0.7298460602760315, + "learning_rate": 2.5198310956459853e-05, + "loss": 0.5209, + "num_input_tokens_seen": 30278704, + "step": 19900 + }, + { + "epoch": 0.6371231035144997, + "grad_norm": 1.2972067594528198, + "learning_rate": 2.518849377325893e-05, + "loss": 0.5934, + "num_input_tokens_seen": 30286432, + "step": 19905 + }, + { + "epoch": 0.6372831444849881, + "grad_norm": 0.8949582576751709, + "learning_rate": 2.51786765609899e-05, + "loss": 0.7717, + "num_input_tokens_seen": 30293888, + "step": 19910 + }, + { + "epoch": 0.6374431854554766, + "grad_norm": 0.3844914138317108, + "learning_rate": 2.5168859321166694e-05, + "loss": 0.3454, + "num_input_tokens_seen": 30301264, + "step": 19915 + }, + { + "epoch": 0.637603226425965, + "grad_norm": 0.5302860140800476, + "learning_rate": 2.515904205530326e-05, + "loss": 0.3978, + "num_input_tokens_seen": 30309008, + "step": 19920 + }, + { + "epoch": 0.6377632673964535, + "grad_norm": 0.6664412021636963, + "learning_rate": 2.514922476491355e-05, + "loss": 0.4135, + "num_input_tokens_seen": 30316432, + "step": 19925 + }, + { + "epoch": 0.6379233083669419, + "grad_norm": 0.6049566268920898, + "learning_rate": 2.51394074515115e-05, + "loss": 0.425, + "num_input_tokens_seen": 30324112, + "step": 19930 + }, + { + "epoch": 0.6380833493374304, + "grad_norm": 0.7523671388626099, + "learning_rate": 2.5129590116611067e-05, + "loss": 0.6382, + "num_input_tokens_seen": 30331920, + "step": 19935 + }, + { + "epoch": 0.6382433903079189, + "grad_norm": 0.8008710741996765, + "learning_rate": 2.5119772761726212e-05, + "loss": 0.4257, + "num_input_tokens_seen": 30339504, + "step": 19940 + }, + { + "epoch": 0.6384034312784073, + "grad_norm": 0.6693298816680908, + "learning_rate": 2.5109955388370893e-05, + "loss": 0.3772, + "num_input_tokens_seen": 30347568, + "step": 19945 + }, + { + "epoch": 0.6385634722488958, + "grad_norm": 0.6073232889175415, + "learning_rate": 2.510013799805907e-05, + "loss": 0.3826, + "num_input_tokens_seen": 30355168, + "step": 19950 + }, + { + "epoch": 0.6387235132193841, + "grad_norm": 0.5140669345855713, + "learning_rate": 2.5090320592304706e-05, + "loss": 0.5262, + "num_input_tokens_seen": 30363120, + "step": 19955 + }, + { + "epoch": 0.6388835541898726, + "grad_norm": 0.9947265386581421, + "learning_rate": 2.5080503172621777e-05, + "loss": 0.491, + "num_input_tokens_seen": 30370544, + "step": 19960 + }, + { + "epoch": 0.639043595160361, + "grad_norm": 0.780404806137085, + "learning_rate": 2.5070685740524246e-05, + "loss": 0.4519, + "num_input_tokens_seen": 30378192, + "step": 19965 + }, + { + "epoch": 0.6392036361308495, + "grad_norm": 1.1170179843902588, + "learning_rate": 2.5060868297526084e-05, + "loss": 0.539, + "num_input_tokens_seen": 30385728, + "step": 19970 + }, + { + "epoch": 0.6393636771013379, + "grad_norm": 0.8204863667488098, + "learning_rate": 2.5051050845141267e-05, + "loss": 0.5315, + "num_input_tokens_seen": 30394032, + "step": 19975 + }, + { + "epoch": 0.6395237180718264, + "grad_norm": 0.4600532054901123, + "learning_rate": 2.5041233384883765e-05, + "loss": 0.4595, + "num_input_tokens_seen": 30401728, + "step": 19980 + }, + { + "epoch": 0.6396837590423148, + "grad_norm": 0.44674646854400635, + "learning_rate": 2.5031415918267564e-05, + "loss": 0.4429, + "num_input_tokens_seen": 30409312, + "step": 19985 + }, + { + "epoch": 0.6398438000128033, + "grad_norm": 0.8826571106910706, + "learning_rate": 2.5021598446806626e-05, + "loss": 0.4226, + "num_input_tokens_seen": 30416576, + "step": 19990 + }, + { + "epoch": 0.6400038409832918, + "grad_norm": 1.2192069292068481, + "learning_rate": 2.5011780972014937e-05, + "loss": 0.6649, + "num_input_tokens_seen": 30424256, + "step": 19995 + }, + { + "epoch": 0.6401638819537802, + "grad_norm": 0.3827016055583954, + "learning_rate": 2.5001963495406478e-05, + "loss": 0.3965, + "num_input_tokens_seen": 30432080, + "step": 20000 + }, + { + "epoch": 0.6401638819537802, + "eval_loss": 0.5141545534133911, + "eval_runtime": 560.5668, + "eval_samples_per_second": 24.771, + "eval_steps_per_second": 12.386, + "num_input_tokens_seen": 30432080, + "step": 20000 + }, + { + "epoch": 0.6403239229242687, + "grad_norm": 0.3314017355442047, + "learning_rate": 2.499214601849522e-05, + "loss": 0.2817, + "num_input_tokens_seen": 30439712, + "step": 20005 + }, + { + "epoch": 0.640483963894757, + "grad_norm": 0.4009249210357666, + "learning_rate": 2.4982328542795148e-05, + "loss": 0.3602, + "num_input_tokens_seen": 30447088, + "step": 20010 + }, + { + "epoch": 0.6406440048652455, + "grad_norm": 0.5794426798820496, + "learning_rate": 2.497251106982024e-05, + "loss": 0.324, + "num_input_tokens_seen": 30454752, + "step": 20015 + }, + { + "epoch": 0.6408040458357339, + "grad_norm": 0.36331745982170105, + "learning_rate": 2.4962693601084458e-05, + "loss": 0.4164, + "num_input_tokens_seen": 30462128, + "step": 20020 + }, + { + "epoch": 0.6409640868062224, + "grad_norm": 0.8090078234672546, + "learning_rate": 2.4952876138101794e-05, + "loss": 0.5449, + "num_input_tokens_seen": 30469760, + "step": 20025 + }, + { + "epoch": 0.6411241277767108, + "grad_norm": 0.7777200937271118, + "learning_rate": 2.4943058682386233e-05, + "loss": 0.4508, + "num_input_tokens_seen": 30477552, + "step": 20030 + }, + { + "epoch": 0.6412841687471993, + "grad_norm": 1.061418056488037, + "learning_rate": 2.493324123545173e-05, + "loss": 0.7251, + "num_input_tokens_seen": 30484912, + "step": 20035 + }, + { + "epoch": 0.6414442097176877, + "grad_norm": 1.4296246767044067, + "learning_rate": 2.4923423798812272e-05, + "loss": 0.5795, + "num_input_tokens_seen": 30492272, + "step": 20040 + }, + { + "epoch": 0.6416042506881762, + "grad_norm": 0.7559539675712585, + "learning_rate": 2.4913606373981825e-05, + "loss": 0.4833, + "num_input_tokens_seen": 30499984, + "step": 20045 + }, + { + "epoch": 0.6417642916586647, + "grad_norm": 0.696273148059845, + "learning_rate": 2.4903788962474357e-05, + "loss": 0.4596, + "num_input_tokens_seen": 30507456, + "step": 20050 + }, + { + "epoch": 0.641924332629153, + "grad_norm": 0.6349233388900757, + "learning_rate": 2.489397156580385e-05, + "loss": 0.5469, + "num_input_tokens_seen": 30514960, + "step": 20055 + }, + { + "epoch": 0.6420843735996415, + "grad_norm": 0.7534859776496887, + "learning_rate": 2.4884154185484246e-05, + "loss": 0.4371, + "num_input_tokens_seen": 30522192, + "step": 20060 + }, + { + "epoch": 0.6422444145701299, + "grad_norm": 0.9380366206169128, + "learning_rate": 2.4874336823029526e-05, + "loss": 0.8417, + "num_input_tokens_seen": 30530224, + "step": 20065 + }, + { + "epoch": 0.6424044555406184, + "grad_norm": 0.969047486782074, + "learning_rate": 2.4864519479953656e-05, + "loss": 0.5534, + "num_input_tokens_seen": 30537824, + "step": 20070 + }, + { + "epoch": 0.6425644965111068, + "grad_norm": 0.9842394590377808, + "learning_rate": 2.485470215777058e-05, + "loss": 0.5927, + "num_input_tokens_seen": 30545168, + "step": 20075 + }, + { + "epoch": 0.6427245374815953, + "grad_norm": 0.850658118724823, + "learning_rate": 2.4844884857994258e-05, + "loss": 0.5341, + "num_input_tokens_seen": 30552864, + "step": 20080 + }, + { + "epoch": 0.6428845784520837, + "grad_norm": 0.5791357755661011, + "learning_rate": 2.4835067582138638e-05, + "loss": 0.4476, + "num_input_tokens_seen": 30560880, + "step": 20085 + }, + { + "epoch": 0.6430446194225722, + "grad_norm": 0.9200384020805359, + "learning_rate": 2.4825250331717666e-05, + "loss": 0.568, + "num_input_tokens_seen": 30568576, + "step": 20090 + }, + { + "epoch": 0.6432046603930607, + "grad_norm": 0.7877505421638489, + "learning_rate": 2.4815433108245298e-05, + "loss": 0.5969, + "num_input_tokens_seen": 30576208, + "step": 20095 + }, + { + "epoch": 0.6433647013635491, + "grad_norm": 0.5380582809448242, + "learning_rate": 2.4805615913235456e-05, + "loss": 0.4972, + "num_input_tokens_seen": 30583776, + "step": 20100 + }, + { + "epoch": 0.6435247423340376, + "grad_norm": 0.8199526071548462, + "learning_rate": 2.479579874820208e-05, + "loss": 0.6281, + "num_input_tokens_seen": 30591328, + "step": 20105 + }, + { + "epoch": 0.6436847833045259, + "grad_norm": 1.318984866142273, + "learning_rate": 2.4785981614659115e-05, + "loss": 0.6355, + "num_input_tokens_seen": 30598528, + "step": 20110 + }, + { + "epoch": 0.6438448242750144, + "grad_norm": 1.118703007698059, + "learning_rate": 2.477616451412047e-05, + "loss": 0.6978, + "num_input_tokens_seen": 30607216, + "step": 20115 + }, + { + "epoch": 0.6440048652455028, + "grad_norm": 2.649280309677124, + "learning_rate": 2.476634744810007e-05, + "loss": 0.4892, + "num_input_tokens_seen": 30614832, + "step": 20120 + }, + { + "epoch": 0.6441649062159913, + "grad_norm": 0.7260949611663818, + "learning_rate": 2.475653041811183e-05, + "loss": 0.4101, + "num_input_tokens_seen": 30622336, + "step": 20125 + }, + { + "epoch": 0.6443249471864797, + "grad_norm": 0.47214579582214355, + "learning_rate": 2.4746713425669652e-05, + "loss": 0.3508, + "num_input_tokens_seen": 30630560, + "step": 20130 + }, + { + "epoch": 0.6444849881569682, + "grad_norm": 2.2958195209503174, + "learning_rate": 2.4736896472287458e-05, + "loss": 0.4351, + "num_input_tokens_seen": 30637952, + "step": 20135 + }, + { + "epoch": 0.6446450291274566, + "grad_norm": 0.5763238668441772, + "learning_rate": 2.4727079559479124e-05, + "loss": 0.5095, + "num_input_tokens_seen": 30646240, + "step": 20140 + }, + { + "epoch": 0.6448050700979451, + "grad_norm": 1.8151342868804932, + "learning_rate": 2.4717262688758557e-05, + "loss": 0.7, + "num_input_tokens_seen": 30654096, + "step": 20145 + }, + { + "epoch": 0.6449651110684336, + "grad_norm": 0.7270373106002808, + "learning_rate": 2.4707445861639637e-05, + "loss": 0.6059, + "num_input_tokens_seen": 30661552, + "step": 20150 + }, + { + "epoch": 0.645125152038922, + "grad_norm": 0.6886783838272095, + "learning_rate": 2.4697629079636244e-05, + "loss": 0.5232, + "num_input_tokens_seen": 30668896, + "step": 20155 + }, + { + "epoch": 0.6452851930094105, + "grad_norm": 1.6188323497772217, + "learning_rate": 2.4687812344262244e-05, + "loss": 0.7796, + "num_input_tokens_seen": 30677168, + "step": 20160 + }, + { + "epoch": 0.6454452339798988, + "grad_norm": 1.0555676221847534, + "learning_rate": 2.46779956570315e-05, + "loss": 0.3938, + "num_input_tokens_seen": 30685024, + "step": 20165 + }, + { + "epoch": 0.6456052749503873, + "grad_norm": 0.6462313532829285, + "learning_rate": 2.466817901945787e-05, + "loss": 0.4214, + "num_input_tokens_seen": 30692336, + "step": 20170 + }, + { + "epoch": 0.6457653159208757, + "grad_norm": 1.3040111064910889, + "learning_rate": 2.4658362433055217e-05, + "loss": 0.5131, + "num_input_tokens_seen": 30699552, + "step": 20175 + }, + { + "epoch": 0.6459253568913642, + "grad_norm": 0.5753178000450134, + "learning_rate": 2.4648545899337356e-05, + "loss": 0.4397, + "num_input_tokens_seen": 30706976, + "step": 20180 + }, + { + "epoch": 0.6460853978618526, + "grad_norm": 0.5029861927032471, + "learning_rate": 2.4638729419818143e-05, + "loss": 0.4511, + "num_input_tokens_seen": 30714352, + "step": 20185 + }, + { + "epoch": 0.6462454388323411, + "grad_norm": 0.6508108973503113, + "learning_rate": 2.46289129960114e-05, + "loss": 0.5439, + "num_input_tokens_seen": 30721952, + "step": 20190 + }, + { + "epoch": 0.6464054798028295, + "grad_norm": 0.6831299066543579, + "learning_rate": 2.4619096629430924e-05, + "loss": 0.6225, + "num_input_tokens_seen": 30730096, + "step": 20195 + }, + { + "epoch": 0.646565520773318, + "grad_norm": 0.6858999729156494, + "learning_rate": 2.4609280321590543e-05, + "loss": 0.463, + "num_input_tokens_seen": 30737872, + "step": 20200 + }, + { + "epoch": 0.646565520773318, + "eval_loss": 0.5128901600837708, + "eval_runtime": 560.2385, + "eval_samples_per_second": 24.786, + "eval_steps_per_second": 12.393, + "num_input_tokens_seen": 30737872, + "step": 20200 + }, + { + "epoch": 0.6467255617438065, + "grad_norm": 1.0180680751800537, + "learning_rate": 2.4599464074004037e-05, + "loss": 0.5626, + "num_input_tokens_seen": 30745504, + "step": 20205 + }, + { + "epoch": 0.6468856027142948, + "grad_norm": 0.7039761543273926, + "learning_rate": 2.4589647888185204e-05, + "loss": 0.5335, + "num_input_tokens_seen": 30753584, + "step": 20210 + }, + { + "epoch": 0.6470456436847833, + "grad_norm": 0.8906511664390564, + "learning_rate": 2.4579831765647836e-05, + "loss": 0.5008, + "num_input_tokens_seen": 30760800, + "step": 20215 + }, + { + "epoch": 0.6472056846552717, + "grad_norm": 0.5761902332305908, + "learning_rate": 2.4570015707905676e-05, + "loss": 0.5198, + "num_input_tokens_seen": 30768624, + "step": 20220 + }, + { + "epoch": 0.6473657256257602, + "grad_norm": 0.4074558615684509, + "learning_rate": 2.4560199716472508e-05, + "loss": 0.4551, + "num_input_tokens_seen": 30776128, + "step": 20225 + }, + { + "epoch": 0.6475257665962486, + "grad_norm": 0.530333936214447, + "learning_rate": 2.455038379286207e-05, + "loss": 0.7448, + "num_input_tokens_seen": 30783872, + "step": 20230 + }, + { + "epoch": 0.6476858075667371, + "grad_norm": 1.23623788356781, + "learning_rate": 2.4540567938588095e-05, + "loss": 0.5562, + "num_input_tokens_seen": 30791520, + "step": 20235 + }, + { + "epoch": 0.6478458485372255, + "grad_norm": 0.972292959690094, + "learning_rate": 2.4530752155164328e-05, + "loss": 0.6485, + "num_input_tokens_seen": 30799312, + "step": 20240 + }, + { + "epoch": 0.648005889507714, + "grad_norm": 0.9032193422317505, + "learning_rate": 2.4520936444104463e-05, + "loss": 0.499, + "num_input_tokens_seen": 30806592, + "step": 20245 + }, + { + "epoch": 0.6481659304782024, + "grad_norm": 0.5856055021286011, + "learning_rate": 2.4511120806922218e-05, + "loss": 0.3664, + "num_input_tokens_seen": 30814528, + "step": 20250 + }, + { + "epoch": 0.6483259714486909, + "grad_norm": 0.595859706401825, + "learning_rate": 2.45013052451313e-05, + "loss": 0.3721, + "num_input_tokens_seen": 30822320, + "step": 20255 + }, + { + "epoch": 0.6484860124191794, + "grad_norm": 0.35783129930496216, + "learning_rate": 2.4491489760245376e-05, + "loss": 0.4058, + "num_input_tokens_seen": 30830144, + "step": 20260 + }, + { + "epoch": 0.6486460533896677, + "grad_norm": 0.6333772540092468, + "learning_rate": 2.4481674353778115e-05, + "loss": 0.5276, + "num_input_tokens_seen": 30837680, + "step": 20265 + }, + { + "epoch": 0.6488060943601562, + "grad_norm": 0.6650328040122986, + "learning_rate": 2.447185902724319e-05, + "loss": 0.3926, + "num_input_tokens_seen": 30845392, + "step": 20270 + }, + { + "epoch": 0.6489661353306446, + "grad_norm": 1.074828863143921, + "learning_rate": 2.4462043782154233e-05, + "loss": 0.4433, + "num_input_tokens_seen": 30852464, + "step": 20275 + }, + { + "epoch": 0.6491261763011331, + "grad_norm": 0.4574130177497864, + "learning_rate": 2.4452228620024895e-05, + "loss": 0.3716, + "num_input_tokens_seen": 30859712, + "step": 20280 + }, + { + "epoch": 0.6492862172716215, + "grad_norm": 0.6622092723846436, + "learning_rate": 2.4442413542368776e-05, + "loss": 0.5896, + "num_input_tokens_seen": 30867152, + "step": 20285 + }, + { + "epoch": 0.64944625824211, + "grad_norm": 0.9408570528030396, + "learning_rate": 2.4432598550699502e-05, + "loss": 0.5127, + "num_input_tokens_seen": 30874944, + "step": 20290 + }, + { + "epoch": 0.6496062992125984, + "grad_norm": 0.9271009564399719, + "learning_rate": 2.4422783646530663e-05, + "loss": 0.5122, + "num_input_tokens_seen": 30882432, + "step": 20295 + }, + { + "epoch": 0.6497663401830869, + "grad_norm": 0.6549489498138428, + "learning_rate": 2.441296883137584e-05, + "loss": 0.4843, + "num_input_tokens_seen": 30889968, + "step": 20300 + }, + { + "epoch": 0.6499263811535753, + "grad_norm": 0.557341992855072, + "learning_rate": 2.4403154106748592e-05, + "loss": 0.5242, + "num_input_tokens_seen": 30897504, + "step": 20305 + }, + { + "epoch": 0.6500864221240638, + "grad_norm": 0.5427817702293396, + "learning_rate": 2.4393339474162494e-05, + "loss": 0.4431, + "num_input_tokens_seen": 30905120, + "step": 20310 + }, + { + "epoch": 0.6502464630945523, + "grad_norm": 0.5350136160850525, + "learning_rate": 2.4383524935131062e-05, + "loss": 0.4303, + "num_input_tokens_seen": 30912992, + "step": 20315 + }, + { + "epoch": 0.6504065040650406, + "grad_norm": 0.9049208164215088, + "learning_rate": 2.437371049116784e-05, + "loss": 0.5499, + "num_input_tokens_seen": 30920704, + "step": 20320 + }, + { + "epoch": 0.6505665450355291, + "grad_norm": 0.5244518518447876, + "learning_rate": 2.436389614378632e-05, + "loss": 0.3782, + "num_input_tokens_seen": 30928064, + "step": 20325 + }, + { + "epoch": 0.6507265860060175, + "grad_norm": 0.6494530439376831, + "learning_rate": 2.435408189450002e-05, + "loss": 0.4689, + "num_input_tokens_seen": 30935584, + "step": 20330 + }, + { + "epoch": 0.650886626976506, + "grad_norm": 0.9323655962944031, + "learning_rate": 2.4344267744822406e-05, + "loss": 0.4819, + "num_input_tokens_seen": 30943264, + "step": 20335 + }, + { + "epoch": 0.6510466679469944, + "grad_norm": 0.9937655329704285, + "learning_rate": 2.4334453696266944e-05, + "loss": 0.4853, + "num_input_tokens_seen": 30950512, + "step": 20340 + }, + { + "epoch": 0.6512067089174829, + "grad_norm": 0.5827783942222595, + "learning_rate": 2.432463975034708e-05, + "loss": 0.4228, + "num_input_tokens_seen": 30957776, + "step": 20345 + }, + { + "epoch": 0.6513667498879713, + "grad_norm": 0.7815657258033752, + "learning_rate": 2.4314825908576265e-05, + "loss": 0.4953, + "num_input_tokens_seen": 30964912, + "step": 20350 + }, + { + "epoch": 0.6515267908584598, + "grad_norm": 1.1565767526626587, + "learning_rate": 2.4305012172467897e-05, + "loss": 0.5852, + "num_input_tokens_seen": 30972464, + "step": 20355 + }, + { + "epoch": 0.6516868318289483, + "grad_norm": 1.0740466117858887, + "learning_rate": 2.4295198543535393e-05, + "loss": 0.4761, + "num_input_tokens_seen": 30979968, + "step": 20360 + }, + { + "epoch": 0.6518468727994366, + "grad_norm": 0.5373172760009766, + "learning_rate": 2.4285385023292124e-05, + "loss": 0.3778, + "num_input_tokens_seen": 30987456, + "step": 20365 + }, + { + "epoch": 0.6520069137699251, + "grad_norm": 1.002793312072754, + "learning_rate": 2.427557161325147e-05, + "loss": 0.501, + "num_input_tokens_seen": 30995184, + "step": 20370 + }, + { + "epoch": 0.6521669547404135, + "grad_norm": 0.7686530947685242, + "learning_rate": 2.4265758314926778e-05, + "loss": 0.4107, + "num_input_tokens_seen": 31003024, + "step": 20375 + }, + { + "epoch": 0.652326995710902, + "grad_norm": 0.6177448034286499, + "learning_rate": 2.4255945129831373e-05, + "loss": 0.6288, + "num_input_tokens_seen": 31010608, + "step": 20380 + }, + { + "epoch": 0.6524870366813904, + "grad_norm": 0.45050790905952454, + "learning_rate": 2.4246132059478578e-05, + "loss": 0.665, + "num_input_tokens_seen": 31018128, + "step": 20385 + }, + { + "epoch": 0.6526470776518789, + "grad_norm": 1.6011016368865967, + "learning_rate": 2.4236319105381706e-05, + "loss": 0.6947, + "num_input_tokens_seen": 31026656, + "step": 20390 + }, + { + "epoch": 0.6528071186223673, + "grad_norm": 0.2926540970802307, + "learning_rate": 2.422650626905401e-05, + "loss": 0.4031, + "num_input_tokens_seen": 31034448, + "step": 20395 + }, + { + "epoch": 0.6529671595928558, + "grad_norm": 1.6223244667053223, + "learning_rate": 2.4216693552008785e-05, + "loss": 0.6503, + "num_input_tokens_seen": 31041328, + "step": 20400 + }, + { + "epoch": 0.6529671595928558, + "eval_loss": 0.5124738812446594, + "eval_runtime": 560.5956, + "eval_samples_per_second": 24.77, + "eval_steps_per_second": 12.385, + "num_input_tokens_seen": 31041328, + "step": 20400 + }, + { + "epoch": 0.6531272005633442, + "grad_norm": 0.6891874670982361, + "learning_rate": 2.4206880955759247e-05, + "loss": 0.6274, + "num_input_tokens_seen": 31048688, + "step": 20405 + }, + { + "epoch": 0.6532872415338327, + "grad_norm": 0.7611523866653442, + "learning_rate": 2.419706848181863e-05, + "loss": 0.4055, + "num_input_tokens_seen": 31056304, + "step": 20410 + }, + { + "epoch": 0.6534472825043212, + "grad_norm": 0.5109581351280212, + "learning_rate": 2.4187256131700153e-05, + "loss": 0.5441, + "num_input_tokens_seen": 31064144, + "step": 20415 + }, + { + "epoch": 0.6536073234748095, + "grad_norm": 1.2305067777633667, + "learning_rate": 2.4177443906916985e-05, + "loss": 0.5068, + "num_input_tokens_seen": 31073008, + "step": 20420 + }, + { + "epoch": 0.653767364445298, + "grad_norm": 0.44451943039894104, + "learning_rate": 2.4167631808982303e-05, + "loss": 0.4954, + "num_input_tokens_seen": 31081200, + "step": 20425 + }, + { + "epoch": 0.6539274054157864, + "grad_norm": 1.0134879350662231, + "learning_rate": 2.4157819839409264e-05, + "loss": 0.6276, + "num_input_tokens_seen": 31088640, + "step": 20430 + }, + { + "epoch": 0.6540874463862749, + "grad_norm": 0.6471532583236694, + "learning_rate": 2.414800799971098e-05, + "loss": 0.5304, + "num_input_tokens_seen": 31096544, + "step": 20435 + }, + { + "epoch": 0.6542474873567633, + "grad_norm": 0.4777883291244507, + "learning_rate": 2.4138196291400582e-05, + "loss": 0.4267, + "num_input_tokens_seen": 31104032, + "step": 20440 + }, + { + "epoch": 0.6544075283272518, + "grad_norm": 0.5195966958999634, + "learning_rate": 2.412838471599114e-05, + "loss": 0.4956, + "num_input_tokens_seen": 31111808, + "step": 20445 + }, + { + "epoch": 0.6545675692977402, + "grad_norm": 0.8042178750038147, + "learning_rate": 2.411857327499572e-05, + "loss": 0.5568, + "num_input_tokens_seen": 31119088, + "step": 20450 + }, + { + "epoch": 0.6547276102682287, + "grad_norm": 1.293122410774231, + "learning_rate": 2.410876196992739e-05, + "loss": 0.5232, + "num_input_tokens_seen": 31126448, + "step": 20455 + }, + { + "epoch": 0.6548876512387171, + "grad_norm": 1.1315381526947021, + "learning_rate": 2.4098950802299156e-05, + "loss": 0.5791, + "num_input_tokens_seen": 31134096, + "step": 20460 + }, + { + "epoch": 0.6550476922092056, + "grad_norm": 1.515467643737793, + "learning_rate": 2.4089139773624027e-05, + "loss": 0.5371, + "num_input_tokens_seen": 31141392, + "step": 20465 + }, + { + "epoch": 0.655207733179694, + "grad_norm": 0.6259340047836304, + "learning_rate": 2.4079328885415007e-05, + "loss": 0.5237, + "num_input_tokens_seen": 31149280, + "step": 20470 + }, + { + "epoch": 0.6553677741501824, + "grad_norm": 0.5189145803451538, + "learning_rate": 2.4069518139185036e-05, + "loss": 0.4733, + "num_input_tokens_seen": 31156528, + "step": 20475 + }, + { + "epoch": 0.6555278151206709, + "grad_norm": 0.5752387046813965, + "learning_rate": 2.405970753644706e-05, + "loss": 0.4389, + "num_input_tokens_seen": 31163952, + "step": 20480 + }, + { + "epoch": 0.6556878560911593, + "grad_norm": 0.4798070192337036, + "learning_rate": 2.4049897078714e-05, + "loss": 0.3931, + "num_input_tokens_seen": 31171760, + "step": 20485 + }, + { + "epoch": 0.6558478970616478, + "grad_norm": 1.2291454076766968, + "learning_rate": 2.404008676749874e-05, + "loss": 0.4271, + "num_input_tokens_seen": 31179088, + "step": 20490 + }, + { + "epoch": 0.6560079380321362, + "grad_norm": 0.5741247534751892, + "learning_rate": 2.403027660431418e-05, + "loss": 0.3946, + "num_input_tokens_seen": 31186448, + "step": 20495 + }, + { + "epoch": 0.6561679790026247, + "grad_norm": 0.8153184056282043, + "learning_rate": 2.402046659067314e-05, + "loss": 0.4965, + "num_input_tokens_seen": 31194032, + "step": 20500 + }, + { + "epoch": 0.6563280199731131, + "grad_norm": 0.6346098780632019, + "learning_rate": 2.401065672808847e-05, + "loss": 0.4863, + "num_input_tokens_seen": 31201312, + "step": 20505 + }, + { + "epoch": 0.6564880609436016, + "grad_norm": 0.8284209370613098, + "learning_rate": 2.400084701807296e-05, + "loss": 0.4287, + "num_input_tokens_seen": 31208720, + "step": 20510 + }, + { + "epoch": 0.65664810191409, + "grad_norm": 0.8916441202163696, + "learning_rate": 2.39910374621394e-05, + "loss": 0.4518, + "num_input_tokens_seen": 31216176, + "step": 20515 + }, + { + "epoch": 0.6568081428845784, + "grad_norm": 0.5070757269859314, + "learning_rate": 2.3981228061800544e-05, + "loss": 0.5453, + "num_input_tokens_seen": 31223376, + "step": 20520 + }, + { + "epoch": 0.656968183855067, + "grad_norm": 1.2794830799102783, + "learning_rate": 2.3971418818569115e-05, + "loss": 0.4963, + "num_input_tokens_seen": 31230864, + "step": 20525 + }, + { + "epoch": 0.6571282248255553, + "grad_norm": 1.3651893138885498, + "learning_rate": 2.3961609733957832e-05, + "loss": 0.4521, + "num_input_tokens_seen": 31238272, + "step": 20530 + }, + { + "epoch": 0.6572882657960438, + "grad_norm": 0.7452698945999146, + "learning_rate": 2.395180080947939e-05, + "loss": 0.4527, + "num_input_tokens_seen": 31245824, + "step": 20535 + }, + { + "epoch": 0.6574483067665322, + "grad_norm": 0.668324887752533, + "learning_rate": 2.394199204664642e-05, + "loss": 0.4299, + "num_input_tokens_seen": 31253216, + "step": 20540 + }, + { + "epoch": 0.6576083477370207, + "grad_norm": 0.4890843331813812, + "learning_rate": 2.3932183446971583e-05, + "loss": 0.7616, + "num_input_tokens_seen": 31261040, + "step": 20545 + }, + { + "epoch": 0.6577683887075091, + "grad_norm": 0.6108789443969727, + "learning_rate": 2.3922375011967473e-05, + "loss": 0.5055, + "num_input_tokens_seen": 31268880, + "step": 20550 + }, + { + "epoch": 0.6579284296779976, + "grad_norm": 0.8907095789909363, + "learning_rate": 2.3912566743146676e-05, + "loss": 0.5405, + "num_input_tokens_seen": 31276432, + "step": 20555 + }, + { + "epoch": 0.658088470648486, + "grad_norm": 0.5622972846031189, + "learning_rate": 2.390275864202176e-05, + "loss": 0.4262, + "num_input_tokens_seen": 31283984, + "step": 20560 + }, + { + "epoch": 0.6582485116189745, + "grad_norm": 0.49914705753326416, + "learning_rate": 2.3892950710105243e-05, + "loss": 0.3842, + "num_input_tokens_seen": 31291296, + "step": 20565 + }, + { + "epoch": 0.6584085525894628, + "grad_norm": 1.2584542036056519, + "learning_rate": 2.3883142948909635e-05, + "loss": 0.4545, + "num_input_tokens_seen": 31298720, + "step": 20570 + }, + { + "epoch": 0.6585685935599513, + "grad_norm": 1.0563548803329468, + "learning_rate": 2.3873335359947433e-05, + "loss": 0.3978, + "num_input_tokens_seen": 31306224, + "step": 20575 + }, + { + "epoch": 0.6587286345304398, + "grad_norm": 0.7999202013015747, + "learning_rate": 2.3863527944731066e-05, + "loss": 0.4862, + "num_input_tokens_seen": 31313808, + "step": 20580 + }, + { + "epoch": 0.6588886755009282, + "grad_norm": 1.162878155708313, + "learning_rate": 2.385372070477298e-05, + "loss": 0.6773, + "num_input_tokens_seen": 31321104, + "step": 20585 + }, + { + "epoch": 0.6590487164714167, + "grad_norm": 1.1264503002166748, + "learning_rate": 2.384391364158556e-05, + "loss": 0.5765, + "num_input_tokens_seen": 31328896, + "step": 20590 + }, + { + "epoch": 0.6592087574419051, + "grad_norm": 1.3764489889144897, + "learning_rate": 2.3834106756681185e-05, + "loss": 0.384, + "num_input_tokens_seen": 31336320, + "step": 20595 + }, + { + "epoch": 0.6593687984123936, + "grad_norm": 0.978012204170227, + "learning_rate": 2.3824300051572206e-05, + "loss": 0.4899, + "num_input_tokens_seen": 31344080, + "step": 20600 + }, + { + "epoch": 0.6593687984123936, + "eval_loss": 0.5114625692367554, + "eval_runtime": 561.2177, + "eval_samples_per_second": 24.743, + "eval_steps_per_second": 12.371, + "num_input_tokens_seen": 31344080, + "step": 20600 + }, + { + "epoch": 0.659528839382882, + "grad_norm": 1.0503337383270264, + "learning_rate": 2.3814493527770923e-05, + "loss": 0.593, + "num_input_tokens_seen": 31351504, + "step": 20605 + }, + { + "epoch": 0.6596888803533705, + "grad_norm": 0.6177509427070618, + "learning_rate": 2.3804687186789637e-05, + "loss": 0.4447, + "num_input_tokens_seen": 31359328, + "step": 20610 + }, + { + "epoch": 0.6598489213238589, + "grad_norm": 1.4545267820358276, + "learning_rate": 2.379488103014062e-05, + "loss": 0.4266, + "num_input_tokens_seen": 31366848, + "step": 20615 + }, + { + "epoch": 0.6600089622943474, + "grad_norm": 0.41802018880844116, + "learning_rate": 2.3785075059336086e-05, + "loss": 0.371, + "num_input_tokens_seen": 31374720, + "step": 20620 + }, + { + "epoch": 0.6601690032648359, + "grad_norm": 0.502812385559082, + "learning_rate": 2.3775269275888248e-05, + "loss": 0.5134, + "num_input_tokens_seen": 31382416, + "step": 20625 + }, + { + "epoch": 0.6603290442353242, + "grad_norm": 1.0473954677581787, + "learning_rate": 2.3765463681309274e-05, + "loss": 0.6657, + "num_input_tokens_seen": 31390032, + "step": 20630 + }, + { + "epoch": 0.6604890852058127, + "grad_norm": 0.7596583962440491, + "learning_rate": 2.3755658277111313e-05, + "loss": 0.4901, + "num_input_tokens_seen": 31397520, + "step": 20635 + }, + { + "epoch": 0.6606491261763011, + "grad_norm": 1.2341090440750122, + "learning_rate": 2.374585306480649e-05, + "loss": 0.4816, + "num_input_tokens_seen": 31404944, + "step": 20640 + }, + { + "epoch": 0.6608091671467896, + "grad_norm": 0.7533516883850098, + "learning_rate": 2.3736048045906877e-05, + "loss": 0.4647, + "num_input_tokens_seen": 31412160, + "step": 20645 + }, + { + "epoch": 0.660969208117278, + "grad_norm": 1.0397535562515259, + "learning_rate": 2.372624322192454e-05, + "loss": 0.5293, + "num_input_tokens_seen": 31419184, + "step": 20650 + }, + { + "epoch": 0.6611292490877665, + "grad_norm": 0.8558357357978821, + "learning_rate": 2.3716438594371516e-05, + "loss": 0.6435, + "num_input_tokens_seen": 31426944, + "step": 20655 + }, + { + "epoch": 0.6612892900582549, + "grad_norm": 1.2184416055679321, + "learning_rate": 2.3706634164759784e-05, + "loss": 0.5873, + "num_input_tokens_seen": 31434384, + "step": 20660 + }, + { + "epoch": 0.6614493310287434, + "grad_norm": 0.9463518857955933, + "learning_rate": 2.3696829934601323e-05, + "loss": 0.5632, + "num_input_tokens_seen": 31441808, + "step": 20665 + }, + { + "epoch": 0.6616093719992318, + "grad_norm": 1.351554274559021, + "learning_rate": 2.3687025905408053e-05, + "loss": 0.4002, + "num_input_tokens_seen": 31449632, + "step": 20670 + }, + { + "epoch": 0.6617694129697202, + "grad_norm": 0.7897055149078369, + "learning_rate": 2.3677222078691886e-05, + "loss": 0.6002, + "num_input_tokens_seen": 31456992, + "step": 20675 + }, + { + "epoch": 0.6619294539402087, + "grad_norm": 0.5556943416595459, + "learning_rate": 2.366741845596471e-05, + "loss": 0.3866, + "num_input_tokens_seen": 31464592, + "step": 20680 + }, + { + "epoch": 0.6620894949106971, + "grad_norm": 0.9903510808944702, + "learning_rate": 2.3657615038738343e-05, + "loss": 0.5391, + "num_input_tokens_seen": 31472336, + "step": 20685 + }, + { + "epoch": 0.6622495358811856, + "grad_norm": 0.675367534160614, + "learning_rate": 2.3647811828524614e-05, + "loss": 0.4896, + "num_input_tokens_seen": 31479872, + "step": 20690 + }, + { + "epoch": 0.662409576851674, + "grad_norm": 0.5221633315086365, + "learning_rate": 2.363800882683529e-05, + "loss": 0.4526, + "num_input_tokens_seen": 31487456, + "step": 20695 + }, + { + "epoch": 0.6625696178221625, + "grad_norm": 0.9789991974830627, + "learning_rate": 2.3628206035182125e-05, + "loss": 0.4639, + "num_input_tokens_seen": 31494816, + "step": 20700 + }, + { + "epoch": 0.6627296587926509, + "grad_norm": 0.5252173542976379, + "learning_rate": 2.361840345507683e-05, + "loss": 0.4989, + "num_input_tokens_seen": 31502160, + "step": 20705 + }, + { + "epoch": 0.6628896997631394, + "grad_norm": 0.7043821811676025, + "learning_rate": 2.3608601088031073e-05, + "loss": 0.6202, + "num_input_tokens_seen": 31509952, + "step": 20710 + }, + { + "epoch": 0.6630497407336278, + "grad_norm": 1.4008703231811523, + "learning_rate": 2.3598798935556516e-05, + "loss": 0.6755, + "num_input_tokens_seen": 31517760, + "step": 20715 + }, + { + "epoch": 0.6632097817041163, + "grad_norm": 0.47050195932388306, + "learning_rate": 2.3588996999164784e-05, + "loss": 0.4837, + "num_input_tokens_seen": 31525664, + "step": 20720 + }, + { + "epoch": 0.6633698226746046, + "grad_norm": 0.5396744608879089, + "learning_rate": 2.3579195280367434e-05, + "loss": 0.4231, + "num_input_tokens_seen": 31533424, + "step": 20725 + }, + { + "epoch": 0.6635298636450931, + "grad_norm": 0.7467303276062012, + "learning_rate": 2.356939378067603e-05, + "loss": 0.4742, + "num_input_tokens_seen": 31541088, + "step": 20730 + }, + { + "epoch": 0.6636899046155816, + "grad_norm": 0.4916883111000061, + "learning_rate": 2.3559592501602092e-05, + "loss": 0.5013, + "num_input_tokens_seen": 31548496, + "step": 20735 + }, + { + "epoch": 0.66384994558607, + "grad_norm": 0.6180232167243958, + "learning_rate": 2.3549791444657076e-05, + "loss": 0.4273, + "num_input_tokens_seen": 31555936, + "step": 20740 + }, + { + "epoch": 0.6640099865565585, + "grad_norm": 0.5748370289802551, + "learning_rate": 2.353999061135246e-05, + "loss": 0.3489, + "num_input_tokens_seen": 31563680, + "step": 20745 + }, + { + "epoch": 0.6641700275270469, + "grad_norm": 0.7388957142829895, + "learning_rate": 2.3530190003199626e-05, + "loss": 0.5673, + "num_input_tokens_seen": 31571296, + "step": 20750 + }, + { + "epoch": 0.6643300684975354, + "grad_norm": 0.46445775032043457, + "learning_rate": 2.3520389621709965e-05, + "loss": 0.3434, + "num_input_tokens_seen": 31579136, + "step": 20755 + }, + { + "epoch": 0.6644901094680238, + "grad_norm": 1.1654376983642578, + "learning_rate": 2.351058946839483e-05, + "loss": 0.5211, + "num_input_tokens_seen": 31586752, + "step": 20760 + }, + { + "epoch": 0.6646501504385123, + "grad_norm": 1.0124104022979736, + "learning_rate": 2.350078954476551e-05, + "loss": 0.5611, + "num_input_tokens_seen": 31593744, + "step": 20765 + }, + { + "epoch": 0.6648101914090007, + "grad_norm": 0.7143092751502991, + "learning_rate": 2.3490989852333272e-05, + "loss": 0.4266, + "num_input_tokens_seen": 31601648, + "step": 20770 + }, + { + "epoch": 0.6649702323794892, + "grad_norm": 0.5983901023864746, + "learning_rate": 2.3481190392609377e-05, + "loss": 0.5209, + "num_input_tokens_seen": 31608928, + "step": 20775 + }, + { + "epoch": 0.6651302733499775, + "grad_norm": 0.9271571040153503, + "learning_rate": 2.3471391167105e-05, + "loss": 0.3863, + "num_input_tokens_seen": 31616880, + "step": 20780 + }, + { + "epoch": 0.665290314320466, + "grad_norm": 0.8530615568161011, + "learning_rate": 2.3461592177331325e-05, + "loss": 0.5247, + "num_input_tokens_seen": 31624320, + "step": 20785 + }, + { + "epoch": 0.6654503552909545, + "grad_norm": 1.2287979125976562, + "learning_rate": 2.345179342479946e-05, + "loss": 0.4372, + "num_input_tokens_seen": 31631760, + "step": 20790 + }, + { + "epoch": 0.6656103962614429, + "grad_norm": 0.8560863137245178, + "learning_rate": 2.3441994911020503e-05, + "loss": 0.452, + "num_input_tokens_seen": 31639520, + "step": 20795 + }, + { + "epoch": 0.6657704372319314, + "grad_norm": 0.9310333728790283, + "learning_rate": 2.3432196637505522e-05, + "loss": 0.5516, + "num_input_tokens_seen": 31646576, + "step": 20800 + }, + { + "epoch": 0.6657704372319314, + "eval_loss": 0.5113004446029663, + "eval_runtime": 559.4328, + "eval_samples_per_second": 24.822, + "eval_steps_per_second": 12.411, + "num_input_tokens_seen": 31646576, + "step": 20800 + }, + { + "epoch": 0.6659304782024198, + "grad_norm": 0.5997359752655029, + "learning_rate": 2.3422398605765515e-05, + "loss": 0.5573, + "num_input_tokens_seen": 31654016, + "step": 20805 + }, + { + "epoch": 0.6660905191729083, + "grad_norm": 1.0377233028411865, + "learning_rate": 2.3412600817311462e-05, + "loss": 0.647, + "num_input_tokens_seen": 31662464, + "step": 20810 + }, + { + "epoch": 0.6662505601433967, + "grad_norm": 1.0401285886764526, + "learning_rate": 2.3402803273654326e-05, + "loss": 0.4953, + "num_input_tokens_seen": 31669840, + "step": 20815 + }, + { + "epoch": 0.6664106011138852, + "grad_norm": 0.9076739549636841, + "learning_rate": 2.3393005976304983e-05, + "loss": 0.3905, + "num_input_tokens_seen": 31677712, + "step": 20820 + }, + { + "epoch": 0.6665706420843736, + "grad_norm": 0.6734322905540466, + "learning_rate": 2.338320892677432e-05, + "loss": 0.6213, + "num_input_tokens_seen": 31685312, + "step": 20825 + }, + { + "epoch": 0.666730683054862, + "grad_norm": 0.6059896945953369, + "learning_rate": 2.3373412126573155e-05, + "loss": 0.7104, + "num_input_tokens_seen": 31692688, + "step": 20830 + }, + { + "epoch": 0.6668907240253504, + "grad_norm": 0.4342031180858612, + "learning_rate": 2.3363615577212285e-05, + "loss": 0.4579, + "num_input_tokens_seen": 31700480, + "step": 20835 + }, + { + "epoch": 0.6670507649958389, + "grad_norm": 1.8723201751708984, + "learning_rate": 2.3353819280202455e-05, + "loss": 0.5316, + "num_input_tokens_seen": 31708256, + "step": 20840 + }, + { + "epoch": 0.6672108059663274, + "grad_norm": 0.8544442653656006, + "learning_rate": 2.334402323705438e-05, + "loss": 0.4361, + "num_input_tokens_seen": 31715792, + "step": 20845 + }, + { + "epoch": 0.6673708469368158, + "grad_norm": 0.5581718683242798, + "learning_rate": 2.3334227449278725e-05, + "loss": 0.379, + "num_input_tokens_seen": 31723296, + "step": 20850 + }, + { + "epoch": 0.6675308879073043, + "grad_norm": 1.6277536153793335, + "learning_rate": 2.3324431918386143e-05, + "loss": 0.5661, + "num_input_tokens_seen": 31730976, + "step": 20855 + }, + { + "epoch": 0.6676909288777927, + "grad_norm": 1.344457983970642, + "learning_rate": 2.3314636645887207e-05, + "loss": 0.4487, + "num_input_tokens_seen": 31738752, + "step": 20860 + }, + { + "epoch": 0.6678509698482812, + "grad_norm": 0.4468259811401367, + "learning_rate": 2.3304841633292487e-05, + "loss": 0.4468, + "num_input_tokens_seen": 31745984, + "step": 20865 + }, + { + "epoch": 0.6680110108187696, + "grad_norm": 0.8063430190086365, + "learning_rate": 2.329504688211248e-05, + "loss": 0.4635, + "num_input_tokens_seen": 31753328, + "step": 20870 + }, + { + "epoch": 0.6681710517892581, + "grad_norm": 0.6300115585327148, + "learning_rate": 2.3285252393857677e-05, + "loss": 0.4631, + "num_input_tokens_seen": 31761008, + "step": 20875 + }, + { + "epoch": 0.6683310927597464, + "grad_norm": 0.6437951922416687, + "learning_rate": 2.327545817003851e-05, + "loss": 0.4974, + "num_input_tokens_seen": 31768368, + "step": 20880 + }, + { + "epoch": 0.6684911337302349, + "grad_norm": 0.9188230633735657, + "learning_rate": 2.326566421216535e-05, + "loss": 0.4401, + "num_input_tokens_seen": 31775808, + "step": 20885 + }, + { + "epoch": 0.6686511747007234, + "grad_norm": 0.6263755559921265, + "learning_rate": 2.3255870521748565e-05, + "loss": 0.431, + "num_input_tokens_seen": 31783312, + "step": 20890 + }, + { + "epoch": 0.6688112156712118, + "grad_norm": 0.4329671561717987, + "learning_rate": 2.3246077100298474e-05, + "loss": 0.5027, + "num_input_tokens_seen": 31790752, + "step": 20895 + }, + { + "epoch": 0.6689712566417003, + "grad_norm": 0.858172595500946, + "learning_rate": 2.3236283949325328e-05, + "loss": 0.547, + "num_input_tokens_seen": 31797840, + "step": 20900 + }, + { + "epoch": 0.6691312976121887, + "grad_norm": 0.5715430974960327, + "learning_rate": 2.3226491070339368e-05, + "loss": 0.5193, + "num_input_tokens_seen": 31805632, + "step": 20905 + }, + { + "epoch": 0.6692913385826772, + "grad_norm": 0.8334700465202332, + "learning_rate": 2.3216698464850762e-05, + "loss": 0.5694, + "num_input_tokens_seen": 31813488, + "step": 20910 + }, + { + "epoch": 0.6694513795531656, + "grad_norm": 1.0989428758621216, + "learning_rate": 2.320690613436967e-05, + "loss": 0.5586, + "num_input_tokens_seen": 31821104, + "step": 20915 + }, + { + "epoch": 0.6696114205236541, + "grad_norm": 0.5173330307006836, + "learning_rate": 2.3197114080406192e-05, + "loss": 0.4388, + "num_input_tokens_seen": 31828592, + "step": 20920 + }, + { + "epoch": 0.6697714614941425, + "grad_norm": 1.2433356046676636, + "learning_rate": 2.3187322304470365e-05, + "loss": 0.7323, + "num_input_tokens_seen": 31836528, + "step": 20925 + }, + { + "epoch": 0.669931502464631, + "grad_norm": 0.5946447253227234, + "learning_rate": 2.3177530808072222e-05, + "loss": 0.4359, + "num_input_tokens_seen": 31844336, + "step": 20930 + }, + { + "epoch": 0.6700915434351193, + "grad_norm": 0.9708590507507324, + "learning_rate": 2.316773959272174e-05, + "loss": 0.6632, + "num_input_tokens_seen": 31852000, + "step": 20935 + }, + { + "epoch": 0.6702515844056078, + "grad_norm": 0.768923819065094, + "learning_rate": 2.3157948659928823e-05, + "loss": 0.4349, + "num_input_tokens_seen": 31859520, + "step": 20940 + }, + { + "epoch": 0.6704116253760963, + "grad_norm": 0.31925976276397705, + "learning_rate": 2.3148158011203388e-05, + "loss": 0.3016, + "num_input_tokens_seen": 31867024, + "step": 20945 + }, + { + "epoch": 0.6705716663465847, + "grad_norm": 0.8433496356010437, + "learning_rate": 2.3138367648055253e-05, + "loss": 0.6865, + "num_input_tokens_seen": 31874576, + "step": 20950 + }, + { + "epoch": 0.6707317073170732, + "grad_norm": 0.7332851886749268, + "learning_rate": 2.312857757199422e-05, + "loss": 0.5306, + "num_input_tokens_seen": 31882304, + "step": 20955 + }, + { + "epoch": 0.6708917482875616, + "grad_norm": 0.8099440336227417, + "learning_rate": 2.3118787784530048e-05, + "loss": 0.4303, + "num_input_tokens_seen": 31890000, + "step": 20960 + }, + { + "epoch": 0.6710517892580501, + "grad_norm": 0.6765058636665344, + "learning_rate": 2.310899828717243e-05, + "loss": 0.5687, + "num_input_tokens_seen": 31897920, + "step": 20965 + }, + { + "epoch": 0.6712118302285385, + "grad_norm": 0.6546658277511597, + "learning_rate": 2.309920908143104e-05, + "loss": 0.6036, + "num_input_tokens_seen": 31905728, + "step": 20970 + }, + { + "epoch": 0.671371871199027, + "grad_norm": 1.3445335626602173, + "learning_rate": 2.308942016881551e-05, + "loss": 0.464, + "num_input_tokens_seen": 31913440, + "step": 20975 + }, + { + "epoch": 0.6715319121695154, + "grad_norm": 0.6104726195335388, + "learning_rate": 2.307963155083539e-05, + "loss": 0.4191, + "num_input_tokens_seen": 31920784, + "step": 20980 + }, + { + "epoch": 0.6716919531400039, + "grad_norm": 0.5414733290672302, + "learning_rate": 2.306984322900022e-05, + "loss": 0.4927, + "num_input_tokens_seen": 31928848, + "step": 20985 + }, + { + "epoch": 0.6718519941104922, + "grad_norm": 0.3659231960773468, + "learning_rate": 2.3060055204819482e-05, + "loss": 0.4667, + "num_input_tokens_seen": 31936464, + "step": 20990 + }, + { + "epoch": 0.6720120350809807, + "grad_norm": 1.0606653690338135, + "learning_rate": 2.3050267479802604e-05, + "loss": 0.436, + "num_input_tokens_seen": 31944256, + "step": 20995 + }, + { + "epoch": 0.6721720760514692, + "grad_norm": 0.9462403059005737, + "learning_rate": 2.304048005545899e-05, + "loss": 0.4349, + "num_input_tokens_seen": 31951744, + "step": 21000 + }, + { + "epoch": 0.6721720760514692, + "eval_loss": 0.5107414126396179, + "eval_runtime": 558.9079, + "eval_samples_per_second": 24.845, + "eval_steps_per_second": 12.422, + "num_input_tokens_seen": 31951744, + "step": 21000 + }, + { + "epoch": 0.6723321170219576, + "grad_norm": 0.7711098790168762, + "learning_rate": 2.3030692933297972e-05, + "loss": 0.4627, + "num_input_tokens_seen": 31959664, + "step": 21005 + }, + { + "epoch": 0.6724921579924461, + "grad_norm": 0.34671735763549805, + "learning_rate": 2.3020906114828843e-05, + "loss": 0.5716, + "num_input_tokens_seen": 31967296, + "step": 21010 + }, + { + "epoch": 0.6726521989629345, + "grad_norm": 0.8552422523498535, + "learning_rate": 2.301111960156088e-05, + "loss": 0.2798, + "num_input_tokens_seen": 31975616, + "step": 21015 + }, + { + "epoch": 0.672812239933423, + "grad_norm": 0.8483350872993469, + "learning_rate": 2.300133339500326e-05, + "loss": 0.5169, + "num_input_tokens_seen": 31983520, + "step": 21020 + }, + { + "epoch": 0.6729722809039114, + "grad_norm": 0.8075284957885742, + "learning_rate": 2.2991547496665148e-05, + "loss": 0.4358, + "num_input_tokens_seen": 31990672, + "step": 21025 + }, + { + "epoch": 0.6731323218743999, + "grad_norm": 0.45466798543930054, + "learning_rate": 2.298176190805565e-05, + "loss": 0.6171, + "num_input_tokens_seen": 31998160, + "step": 21030 + }, + { + "epoch": 0.6732923628448882, + "grad_norm": 0.620844304561615, + "learning_rate": 2.2971976630683826e-05, + "loss": 0.5001, + "num_input_tokens_seen": 32005952, + "step": 21035 + }, + { + "epoch": 0.6734524038153767, + "grad_norm": 1.6539885997772217, + "learning_rate": 2.29621916660587e-05, + "loss": 0.657, + "num_input_tokens_seen": 32013488, + "step": 21040 + }, + { + "epoch": 0.6736124447858651, + "grad_norm": 0.7977017164230347, + "learning_rate": 2.295240701568922e-05, + "loss": 0.4952, + "num_input_tokens_seen": 32021184, + "step": 21045 + }, + { + "epoch": 0.6737724857563536, + "grad_norm": 1.0467742681503296, + "learning_rate": 2.2942622681084312e-05, + "loss": 0.3723, + "num_input_tokens_seen": 32028960, + "step": 21050 + }, + { + "epoch": 0.6739325267268421, + "grad_norm": 1.273114800453186, + "learning_rate": 2.293283866375284e-05, + "loss": 0.5838, + "num_input_tokens_seen": 32036240, + "step": 21055 + }, + { + "epoch": 0.6740925676973305, + "grad_norm": 0.6735864877700806, + "learning_rate": 2.2923054965203627e-05, + "loss": 0.6234, + "num_input_tokens_seen": 32044544, + "step": 21060 + }, + { + "epoch": 0.674252608667819, + "grad_norm": 0.40463078022003174, + "learning_rate": 2.2913271586945443e-05, + "loss": 0.4945, + "num_input_tokens_seen": 32051712, + "step": 21065 + }, + { + "epoch": 0.6744126496383074, + "grad_norm": 0.5007460117340088, + "learning_rate": 2.290348853048699e-05, + "loss": 0.3478, + "num_input_tokens_seen": 32059584, + "step": 21070 + }, + { + "epoch": 0.6745726906087959, + "grad_norm": 0.6307410001754761, + "learning_rate": 2.2893705797336956e-05, + "loss": 0.5068, + "num_input_tokens_seen": 32067152, + "step": 21075 + }, + { + "epoch": 0.6747327315792843, + "grad_norm": 0.7103579044342041, + "learning_rate": 2.288392338900397e-05, + "loss": 0.5095, + "num_input_tokens_seen": 32074944, + "step": 21080 + }, + { + "epoch": 0.6748927725497728, + "grad_norm": 0.8517158031463623, + "learning_rate": 2.2874141306996576e-05, + "loss": 0.409, + "num_input_tokens_seen": 32082976, + "step": 21085 + }, + { + "epoch": 0.6750528135202611, + "grad_norm": 0.8128587007522583, + "learning_rate": 2.2864359552823312e-05, + "loss": 0.342, + "num_input_tokens_seen": 32090640, + "step": 21090 + }, + { + "epoch": 0.6752128544907496, + "grad_norm": 0.7859005928039551, + "learning_rate": 2.2854578127992648e-05, + "loss": 0.5129, + "num_input_tokens_seen": 32098544, + "step": 21095 + }, + { + "epoch": 0.6753728954612381, + "grad_norm": 0.8801397681236267, + "learning_rate": 2.2844797034012988e-05, + "loss": 0.435, + "num_input_tokens_seen": 32106032, + "step": 21100 + }, + { + "epoch": 0.6755329364317265, + "grad_norm": 1.0812106132507324, + "learning_rate": 2.2835016272392722e-05, + "loss": 0.4028, + "num_input_tokens_seen": 32114064, + "step": 21105 + }, + { + "epoch": 0.675692977402215, + "grad_norm": 0.8798759579658508, + "learning_rate": 2.2825235844640142e-05, + "loss": 0.4889, + "num_input_tokens_seen": 32121104, + "step": 21110 + }, + { + "epoch": 0.6758530183727034, + "grad_norm": 0.7603171467781067, + "learning_rate": 2.2815455752263522e-05, + "loss": 0.516, + "num_input_tokens_seen": 32129008, + "step": 21115 + }, + { + "epoch": 0.6760130593431919, + "grad_norm": 0.6676321029663086, + "learning_rate": 2.2805675996771092e-05, + "loss": 0.6074, + "num_input_tokens_seen": 32136800, + "step": 21120 + }, + { + "epoch": 0.6761731003136803, + "grad_norm": 0.7026374340057373, + "learning_rate": 2.2795896579670987e-05, + "loss": 0.3808, + "num_input_tokens_seen": 32144432, + "step": 21125 + }, + { + "epoch": 0.6763331412841688, + "grad_norm": 0.69007807970047, + "learning_rate": 2.2786117502471337e-05, + "loss": 0.4441, + "num_input_tokens_seen": 32152416, + "step": 21130 + }, + { + "epoch": 0.6764931822546572, + "grad_norm": 1.0959550142288208, + "learning_rate": 2.2776338766680185e-05, + "loss": 0.4481, + "num_input_tokens_seen": 32159824, + "step": 21135 + }, + { + "epoch": 0.6766532232251457, + "grad_norm": 0.823154091835022, + "learning_rate": 2.2766560373805533e-05, + "loss": 0.7373, + "num_input_tokens_seen": 32167376, + "step": 21140 + }, + { + "epoch": 0.676813264195634, + "grad_norm": 0.6788070797920227, + "learning_rate": 2.2756782325355353e-05, + "loss": 0.3959, + "num_input_tokens_seen": 32175552, + "step": 21145 + }, + { + "epoch": 0.6769733051661225, + "grad_norm": 0.43499451875686646, + "learning_rate": 2.2747004622837514e-05, + "loss": 0.3771, + "num_input_tokens_seen": 32182864, + "step": 21150 + }, + { + "epoch": 0.677133346136611, + "grad_norm": 0.3533773422241211, + "learning_rate": 2.2737227267759878e-05, + "loss": 0.6083, + "num_input_tokens_seen": 32189952, + "step": 21155 + }, + { + "epoch": 0.6772933871070994, + "grad_norm": 0.7799853086471558, + "learning_rate": 2.272745026163024e-05, + "loss": 0.7138, + "num_input_tokens_seen": 32197456, + "step": 21160 + }, + { + "epoch": 0.6774534280775879, + "grad_norm": 1.298700213432312, + "learning_rate": 2.271767360595633e-05, + "loss": 0.5855, + "num_input_tokens_seen": 32204656, + "step": 21165 + }, + { + "epoch": 0.6776134690480763, + "grad_norm": 0.8354967832565308, + "learning_rate": 2.270789730224583e-05, + "loss": 0.4115, + "num_input_tokens_seen": 32212048, + "step": 21170 + }, + { + "epoch": 0.6777735100185648, + "grad_norm": 0.5662662982940674, + "learning_rate": 2.2698121352006367e-05, + "loss": 0.3968, + "num_input_tokens_seen": 32219200, + "step": 21175 + }, + { + "epoch": 0.6779335509890532, + "grad_norm": 1.2385767698287964, + "learning_rate": 2.2688345756745517e-05, + "loss": 0.5995, + "num_input_tokens_seen": 32226336, + "step": 21180 + }, + { + "epoch": 0.6780935919595417, + "grad_norm": 1.182218074798584, + "learning_rate": 2.267857051797081e-05, + "loss": 0.4319, + "num_input_tokens_seen": 32234208, + "step": 21185 + }, + { + "epoch": 0.67825363293003, + "grad_norm": 0.6666486263275146, + "learning_rate": 2.2668795637189695e-05, + "loss": 0.4679, + "num_input_tokens_seen": 32242400, + "step": 21190 + }, + { + "epoch": 0.6784136739005185, + "grad_norm": 0.5331123471260071, + "learning_rate": 2.2659021115909586e-05, + "loss": 0.5438, + "num_input_tokens_seen": 32250000, + "step": 21195 + }, + { + "epoch": 0.6785737148710069, + "grad_norm": 0.6424439549446106, + "learning_rate": 2.2649246955637847e-05, + "loss": 0.4682, + "num_input_tokens_seen": 32257664, + "step": 21200 + }, + { + "epoch": 0.6785737148710069, + "eval_loss": 0.5105786919593811, + "eval_runtime": 558.8262, + "eval_samples_per_second": 24.849, + "eval_steps_per_second": 12.424, + "num_input_tokens_seen": 32257664, + "step": 21200 + }, + { + "epoch": 0.6787337558414954, + "grad_norm": 1.2377007007598877, + "learning_rate": 2.2639473157881766e-05, + "loss": 0.4802, + "num_input_tokens_seen": 32265056, + "step": 21205 + }, + { + "epoch": 0.6788937968119839, + "grad_norm": 0.763427197933197, + "learning_rate": 2.2629699724148594e-05, + "loss": 0.8831, + "num_input_tokens_seen": 32272512, + "step": 21210 + }, + { + "epoch": 0.6790538377824723, + "grad_norm": 0.8819624781608582, + "learning_rate": 2.26199266559455e-05, + "loss": 0.5235, + "num_input_tokens_seen": 32280704, + "step": 21215 + }, + { + "epoch": 0.6792138787529608, + "grad_norm": 1.0942943096160889, + "learning_rate": 2.2610153954779625e-05, + "loss": 0.4088, + "num_input_tokens_seen": 32288816, + "step": 21220 + }, + { + "epoch": 0.6793739197234492, + "grad_norm": 1.7093944549560547, + "learning_rate": 2.2600381622158056e-05, + "loss": 0.7626, + "num_input_tokens_seen": 32296608, + "step": 21225 + }, + { + "epoch": 0.6795339606939377, + "grad_norm": 0.8502948880195618, + "learning_rate": 2.2590609659587783e-05, + "loss": 0.4931, + "num_input_tokens_seen": 32304320, + "step": 21230 + }, + { + "epoch": 0.6796940016644261, + "grad_norm": 0.883773684501648, + "learning_rate": 2.2580838068575787e-05, + "loss": 0.4402, + "num_input_tokens_seen": 32311984, + "step": 21235 + }, + { + "epoch": 0.6798540426349146, + "grad_norm": 0.9956501722335815, + "learning_rate": 2.257106685062896e-05, + "loss": 0.7088, + "num_input_tokens_seen": 32319728, + "step": 21240 + }, + { + "epoch": 0.6800140836054029, + "grad_norm": 1.0724225044250488, + "learning_rate": 2.256129600725415e-05, + "loss": 0.5562, + "num_input_tokens_seen": 32327456, + "step": 21245 + }, + { + "epoch": 0.6801741245758914, + "grad_norm": 0.7689825892448425, + "learning_rate": 2.2551525539958145e-05, + "loss": 0.5975, + "num_input_tokens_seen": 32334912, + "step": 21250 + }, + { + "epoch": 0.6803341655463798, + "grad_norm": 0.7234827876091003, + "learning_rate": 2.2541755450247663e-05, + "loss": 0.4902, + "num_input_tokens_seen": 32342512, + "step": 21255 + }, + { + "epoch": 0.6804942065168683, + "grad_norm": 0.8552404642105103, + "learning_rate": 2.2531985739629382e-05, + "loss": 0.4443, + "num_input_tokens_seen": 32349936, + "step": 21260 + }, + { + "epoch": 0.6806542474873568, + "grad_norm": 0.38554227352142334, + "learning_rate": 2.2522216409609924e-05, + "loss": 0.4081, + "num_input_tokens_seen": 32357808, + "step": 21265 + }, + { + "epoch": 0.6808142884578452, + "grad_norm": 0.6574006676673889, + "learning_rate": 2.2512447461695826e-05, + "loss": 0.4958, + "num_input_tokens_seen": 32365472, + "step": 21270 + }, + { + "epoch": 0.6809743294283337, + "grad_norm": 0.488569974899292, + "learning_rate": 2.2502678897393593e-05, + "loss": 0.4591, + "num_input_tokens_seen": 32373120, + "step": 21275 + }, + { + "epoch": 0.6811343703988221, + "grad_norm": 0.8761528730392456, + "learning_rate": 2.2492910718209665e-05, + "loss": 0.7583, + "num_input_tokens_seen": 32380736, + "step": 21280 + }, + { + "epoch": 0.6812944113693106, + "grad_norm": 0.7135965824127197, + "learning_rate": 2.2483142925650398e-05, + "loss": 0.5033, + "num_input_tokens_seen": 32388304, + "step": 21285 + }, + { + "epoch": 0.681454452339799, + "grad_norm": 0.5612173676490784, + "learning_rate": 2.247337552122213e-05, + "loss": 0.6541, + "num_input_tokens_seen": 32395312, + "step": 21290 + }, + { + "epoch": 0.6816144933102875, + "grad_norm": 0.4730812907218933, + "learning_rate": 2.24636085064311e-05, + "loss": 0.4018, + "num_input_tokens_seen": 32402864, + "step": 21295 + }, + { + "epoch": 0.6817745342807758, + "grad_norm": 0.6141682267189026, + "learning_rate": 2.245384188278351e-05, + "loss": 0.503, + "num_input_tokens_seen": 32410208, + "step": 21300 + }, + { + "epoch": 0.6819345752512643, + "grad_norm": 0.7513027191162109, + "learning_rate": 2.2444075651785513e-05, + "loss": 0.4394, + "num_input_tokens_seen": 32417264, + "step": 21305 + }, + { + "epoch": 0.6820946162217527, + "grad_norm": 0.4810510277748108, + "learning_rate": 2.243430981494316e-05, + "loss": 0.3833, + "num_input_tokens_seen": 32424736, + "step": 21310 + }, + { + "epoch": 0.6822546571922412, + "grad_norm": 0.5554569363594055, + "learning_rate": 2.2424544373762475e-05, + "loss": 0.4557, + "num_input_tokens_seen": 32432432, + "step": 21315 + }, + { + "epoch": 0.6824146981627297, + "grad_norm": 0.534913182258606, + "learning_rate": 2.2414779329749418e-05, + "loss": 0.3294, + "num_input_tokens_seen": 32439824, + "step": 21320 + }, + { + "epoch": 0.6825747391332181, + "grad_norm": 0.726302981376648, + "learning_rate": 2.2405014684409873e-05, + "loss": 0.4434, + "num_input_tokens_seen": 32447728, + "step": 21325 + }, + { + "epoch": 0.6827347801037066, + "grad_norm": 1.542841911315918, + "learning_rate": 2.239525043924968e-05, + "loss": 0.4251, + "num_input_tokens_seen": 32455152, + "step": 21330 + }, + { + "epoch": 0.682894821074195, + "grad_norm": 0.6294319033622742, + "learning_rate": 2.2385486595774592e-05, + "loss": 0.49, + "num_input_tokens_seen": 32462704, + "step": 21335 + }, + { + "epoch": 0.6830548620446835, + "grad_norm": 1.0318762063980103, + "learning_rate": 2.237572315549033e-05, + "loss": 0.6611, + "num_input_tokens_seen": 32470080, + "step": 21340 + }, + { + "epoch": 0.6832149030151718, + "grad_norm": 0.8054423928260803, + "learning_rate": 2.2365960119902545e-05, + "loss": 0.4396, + "num_input_tokens_seen": 32477360, + "step": 21345 + }, + { + "epoch": 0.6833749439856603, + "grad_norm": 1.4071639776229858, + "learning_rate": 2.2356197490516806e-05, + "loss": 0.4573, + "num_input_tokens_seen": 32484784, + "step": 21350 + }, + { + "epoch": 0.6835349849561487, + "grad_norm": 0.6196765899658203, + "learning_rate": 2.234643526883863e-05, + "loss": 0.3111, + "num_input_tokens_seen": 32492384, + "step": 21355 + }, + { + "epoch": 0.6836950259266372, + "grad_norm": 0.6537829041481018, + "learning_rate": 2.2336673456373497e-05, + "loss": 0.3953, + "num_input_tokens_seen": 32500288, + "step": 21360 + }, + { + "epoch": 0.6838550668971257, + "grad_norm": 1.5610288381576538, + "learning_rate": 2.2326912054626772e-05, + "loss": 0.4349, + "num_input_tokens_seen": 32507984, + "step": 21365 + }, + { + "epoch": 0.6840151078676141, + "grad_norm": 0.5629582405090332, + "learning_rate": 2.2317151065103813e-05, + "loss": 0.4097, + "num_input_tokens_seen": 32515664, + "step": 21370 + }, + { + "epoch": 0.6841751488381026, + "grad_norm": 0.7878997921943665, + "learning_rate": 2.2307390489309865e-05, + "loss": 0.4242, + "num_input_tokens_seen": 32523104, + "step": 21375 + }, + { + "epoch": 0.684335189808591, + "grad_norm": 0.7861080765724182, + "learning_rate": 2.2297630328750146e-05, + "loss": 0.5894, + "num_input_tokens_seen": 32530944, + "step": 21380 + }, + { + "epoch": 0.6844952307790795, + "grad_norm": 1.0403785705566406, + "learning_rate": 2.228787058492979e-05, + "loss": 0.3478, + "num_input_tokens_seen": 32538496, + "step": 21385 + }, + { + "epoch": 0.6846552717495679, + "grad_norm": 0.6690950989723206, + "learning_rate": 2.2278111259353875e-05, + "loss": 0.5696, + "num_input_tokens_seen": 32546224, + "step": 21390 + }, + { + "epoch": 0.6848153127200564, + "grad_norm": 0.34248629212379456, + "learning_rate": 2.2268352353527395e-05, + "loss": 0.3629, + "num_input_tokens_seen": 32553888, + "step": 21395 + }, + { + "epoch": 0.6849753536905447, + "grad_norm": 1.124806523323059, + "learning_rate": 2.225859386895533e-05, + "loss": 0.6893, + "num_input_tokens_seen": 32561408, + "step": 21400 + }, + { + "epoch": 0.6849753536905447, + "eval_loss": 0.5100002288818359, + "eval_runtime": 558.6661, + "eval_samples_per_second": 24.856, + "eval_steps_per_second": 12.428, + "num_input_tokens_seen": 32561408, + "step": 21400 + }, + { + "epoch": 0.6851353946610332, + "grad_norm": 0.7770601511001587, + "learning_rate": 2.2248835807142525e-05, + "loss": 0.4179, + "num_input_tokens_seen": 32569488, + "step": 21405 + }, + { + "epoch": 0.6852954356315216, + "grad_norm": 0.49725255370140076, + "learning_rate": 2.2239078169593826e-05, + "loss": 0.4697, + "num_input_tokens_seen": 32577312, + "step": 21410 + }, + { + "epoch": 0.6854554766020101, + "grad_norm": 3.374659299850464, + "learning_rate": 2.222932095781396e-05, + "loss": 0.3893, + "num_input_tokens_seen": 32585808, + "step": 21415 + }, + { + "epoch": 0.6856155175724986, + "grad_norm": 0.9437829256057739, + "learning_rate": 2.221956417330762e-05, + "loss": 0.3183, + "num_input_tokens_seen": 32593424, + "step": 21420 + }, + { + "epoch": 0.685775558542987, + "grad_norm": 0.719283938407898, + "learning_rate": 2.2209807817579438e-05, + "loss": 0.4664, + "num_input_tokens_seen": 32601568, + "step": 21425 + }, + { + "epoch": 0.6859355995134755, + "grad_norm": 0.844899594783783, + "learning_rate": 2.220005189213394e-05, + "loss": 0.5302, + "num_input_tokens_seen": 32609024, + "step": 21430 + }, + { + "epoch": 0.6860956404839639, + "grad_norm": 1.2877808809280396, + "learning_rate": 2.2190296398475624e-05, + "loss": 0.3346, + "num_input_tokens_seen": 32616656, + "step": 21435 + }, + { + "epoch": 0.6862556814544524, + "grad_norm": 1.29386305809021, + "learning_rate": 2.2180541338108926e-05, + "loss": 0.5662, + "num_input_tokens_seen": 32624080, + "step": 21440 + }, + { + "epoch": 0.6864157224249408, + "grad_norm": 0.8303503394126892, + "learning_rate": 2.2170786712538176e-05, + "loss": 0.6424, + "num_input_tokens_seen": 32631616, + "step": 21445 + }, + { + "epoch": 0.6865757633954293, + "grad_norm": 0.428276002407074, + "learning_rate": 2.216103252326768e-05, + "loss": 0.3646, + "num_input_tokens_seen": 32639392, + "step": 21450 + }, + { + "epoch": 0.6867358043659176, + "grad_norm": 0.6053194999694824, + "learning_rate": 2.2151278771801635e-05, + "loss": 0.4238, + "num_input_tokens_seen": 32646784, + "step": 21455 + }, + { + "epoch": 0.6868958453364061, + "grad_norm": 0.8479558825492859, + "learning_rate": 2.21415254596442e-05, + "loss": 0.3368, + "num_input_tokens_seen": 32654704, + "step": 21460 + }, + { + "epoch": 0.6870558863068945, + "grad_norm": 0.6701275706291199, + "learning_rate": 2.213177258829947e-05, + "loss": 0.3698, + "num_input_tokens_seen": 32662160, + "step": 21465 + }, + { + "epoch": 0.687215927277383, + "grad_norm": 1.6046463251113892, + "learning_rate": 2.2122020159271445e-05, + "loss": 0.5431, + "num_input_tokens_seen": 32670080, + "step": 21470 + }, + { + "epoch": 0.6873759682478715, + "grad_norm": 1.1787571907043457, + "learning_rate": 2.2112268174064075e-05, + "loss": 0.4374, + "num_input_tokens_seen": 32677744, + "step": 21475 + }, + { + "epoch": 0.6875360092183599, + "grad_norm": 0.9196311831474304, + "learning_rate": 2.2102516634181253e-05, + "loss": 0.4054, + "num_input_tokens_seen": 32685344, + "step": 21480 + }, + { + "epoch": 0.6876960501888484, + "grad_norm": 0.3635399043560028, + "learning_rate": 2.209276554112677e-05, + "loss": 0.4772, + "num_input_tokens_seen": 32693072, + "step": 21485 + }, + { + "epoch": 0.6878560911593368, + "grad_norm": 0.6910290718078613, + "learning_rate": 2.2083014896404384e-05, + "loss": 0.6623, + "num_input_tokens_seen": 32700880, + "step": 21490 + }, + { + "epoch": 0.6880161321298253, + "grad_norm": 0.8085564374923706, + "learning_rate": 2.207326470151775e-05, + "loss": 0.465, + "num_input_tokens_seen": 32708208, + "step": 21495 + }, + { + "epoch": 0.6881761731003136, + "grad_norm": 0.7749375104904175, + "learning_rate": 2.2063514957970477e-05, + "loss": 0.4596, + "num_input_tokens_seen": 32715568, + "step": 21500 + }, + { + "epoch": 0.6883362140708021, + "grad_norm": 0.7135915756225586, + "learning_rate": 2.205376566726611e-05, + "loss": 0.5042, + "num_input_tokens_seen": 32723472, + "step": 21505 + }, + { + "epoch": 0.6884962550412905, + "grad_norm": 0.7738295197486877, + "learning_rate": 2.204401683090809e-05, + "loss": 0.4374, + "num_input_tokens_seen": 32731120, + "step": 21510 + }, + { + "epoch": 0.688656296011779, + "grad_norm": 0.7445336580276489, + "learning_rate": 2.203426845039982e-05, + "loss": 0.5072, + "num_input_tokens_seen": 32738880, + "step": 21515 + }, + { + "epoch": 0.6888163369822674, + "grad_norm": 0.6308954954147339, + "learning_rate": 2.202452052724464e-05, + "loss": 0.4779, + "num_input_tokens_seen": 32746560, + "step": 21520 + }, + { + "epoch": 0.6889763779527559, + "grad_norm": 0.626699686050415, + "learning_rate": 2.2014773062945777e-05, + "loss": 0.5962, + "num_input_tokens_seen": 32753840, + "step": 21525 + }, + { + "epoch": 0.6891364189232444, + "grad_norm": 0.8012779355049133, + "learning_rate": 2.2005026059006427e-05, + "loss": 0.5529, + "num_input_tokens_seen": 32761392, + "step": 21530 + }, + { + "epoch": 0.6892964598937328, + "grad_norm": 0.647818922996521, + "learning_rate": 2.1995279516929695e-05, + "loss": 0.4331, + "num_input_tokens_seen": 32768800, + "step": 21535 + }, + { + "epoch": 0.6894565008642213, + "grad_norm": 0.4133044183254242, + "learning_rate": 2.1985533438218613e-05, + "loss": 0.4319, + "num_input_tokens_seen": 32776544, + "step": 21540 + }, + { + "epoch": 0.6896165418347097, + "grad_norm": 2.03057861328125, + "learning_rate": 2.197578782437617e-05, + "loss": 0.3798, + "num_input_tokens_seen": 32784048, + "step": 21545 + }, + { + "epoch": 0.6897765828051982, + "grad_norm": 1.8037998676300049, + "learning_rate": 2.196604267690524e-05, + "loss": 0.4791, + "num_input_tokens_seen": 32791296, + "step": 21550 + }, + { + "epoch": 0.6899366237756865, + "grad_norm": 1.77540123462677, + "learning_rate": 2.195629799730865e-05, + "loss": 0.6137, + "num_input_tokens_seen": 32799008, + "step": 21555 + }, + { + "epoch": 0.690096664746175, + "grad_norm": 0.7993898391723633, + "learning_rate": 2.1946553787089173e-05, + "loss": 0.5601, + "num_input_tokens_seen": 32806272, + "step": 21560 + }, + { + "epoch": 0.6902567057166634, + "grad_norm": 1.525353193283081, + "learning_rate": 2.193681004774947e-05, + "loss": 0.3153, + "num_input_tokens_seen": 32813968, + "step": 21565 + }, + { + "epoch": 0.6904167466871519, + "grad_norm": 1.4006807804107666, + "learning_rate": 2.1927066780792154e-05, + "loss": 0.5762, + "num_input_tokens_seen": 32821968, + "step": 21570 + }, + { + "epoch": 0.6905767876576403, + "grad_norm": 0.8311100602149963, + "learning_rate": 2.191732398771975e-05, + "loss": 0.367, + "num_input_tokens_seen": 32829776, + "step": 21575 + }, + { + "epoch": 0.6907368286281288, + "grad_norm": 1.2398216724395752, + "learning_rate": 2.1907581670034725e-05, + "loss": 0.4907, + "num_input_tokens_seen": 32837664, + "step": 21580 + }, + { + "epoch": 0.6908968695986173, + "grad_norm": 0.5960164070129395, + "learning_rate": 2.189783982923948e-05, + "loss": 0.3878, + "num_input_tokens_seen": 32845456, + "step": 21585 + }, + { + "epoch": 0.6910569105691057, + "grad_norm": 0.7744295597076416, + "learning_rate": 2.1888098466836303e-05, + "loss": 0.4858, + "num_input_tokens_seen": 32853136, + "step": 21590 + }, + { + "epoch": 0.6912169515395942, + "grad_norm": 0.6607256531715393, + "learning_rate": 2.1878357584327457e-05, + "loss": 0.3938, + "num_input_tokens_seen": 32860816, + "step": 21595 + }, + { + "epoch": 0.6913769925100826, + "grad_norm": 0.43947017192840576, + "learning_rate": 2.1868617183215103e-05, + "loss": 0.3469, + "num_input_tokens_seen": 32868640, + "step": 21600 + }, + { + "epoch": 0.6913769925100826, + "eval_loss": 0.5095346570014954, + "eval_runtime": 558.8297, + "eval_samples_per_second": 24.848, + "eval_steps_per_second": 12.424, + "num_input_tokens_seen": 32868640, + "step": 21600 + }, + { + "epoch": 0.691537033480571, + "grad_norm": 1.1455069780349731, + "learning_rate": 2.1858877265001327e-05, + "loss": 0.4549, + "num_input_tokens_seen": 32876048, + "step": 21605 + }, + { + "epoch": 0.6916970744510594, + "grad_norm": 0.6036530137062073, + "learning_rate": 2.184913783118816e-05, + "loss": 0.454, + "num_input_tokens_seen": 32883472, + "step": 21610 + }, + { + "epoch": 0.6918571154215479, + "grad_norm": 0.5247396230697632, + "learning_rate": 2.1839398883277522e-05, + "loss": 0.4835, + "num_input_tokens_seen": 32890960, + "step": 21615 + }, + { + "epoch": 0.6920171563920363, + "grad_norm": 0.7282567024230957, + "learning_rate": 2.182966042277129e-05, + "loss": 0.4801, + "num_input_tokens_seen": 32898672, + "step": 21620 + }, + { + "epoch": 0.6921771973625248, + "grad_norm": 0.8771766424179077, + "learning_rate": 2.181992245117128e-05, + "loss": 0.5636, + "num_input_tokens_seen": 32906512, + "step": 21625 + }, + { + "epoch": 0.6923372383330133, + "grad_norm": 0.5630145072937012, + "learning_rate": 2.181018496997918e-05, + "loss": 0.3616, + "num_input_tokens_seen": 32914240, + "step": 21630 + }, + { + "epoch": 0.6924972793035017, + "grad_norm": 0.6761271357536316, + "learning_rate": 2.1800447980696648e-05, + "loss": 0.508, + "num_input_tokens_seen": 32921616, + "step": 21635 + }, + { + "epoch": 0.6926573202739902, + "grad_norm": 0.7338886260986328, + "learning_rate": 2.1790711484825248e-05, + "loss": 0.3856, + "num_input_tokens_seen": 32929248, + "step": 21640 + }, + { + "epoch": 0.6928173612444786, + "grad_norm": 0.758660078048706, + "learning_rate": 2.178097548386646e-05, + "loss": 0.5074, + "num_input_tokens_seen": 32937008, + "step": 21645 + }, + { + "epoch": 0.6929774022149671, + "grad_norm": 0.41755411028862, + "learning_rate": 2.1771239979321712e-05, + "loss": 0.329, + "num_input_tokens_seen": 32944528, + "step": 21650 + }, + { + "epoch": 0.6931374431854554, + "grad_norm": 1.0807839632034302, + "learning_rate": 2.1761504972692327e-05, + "loss": 0.5308, + "num_input_tokens_seen": 32952240, + "step": 21655 + }, + { + "epoch": 0.6932974841559439, + "grad_norm": 0.6753992438316345, + "learning_rate": 2.1751770465479572e-05, + "loss": 0.3675, + "num_input_tokens_seen": 32959984, + "step": 21660 + }, + { + "epoch": 0.6934575251264323, + "grad_norm": 0.9615740180015564, + "learning_rate": 2.174203645918464e-05, + "loss": 0.4955, + "num_input_tokens_seen": 32967504, + "step": 21665 + }, + { + "epoch": 0.6936175660969208, + "grad_norm": 0.6708840131759644, + "learning_rate": 2.1732302955308624e-05, + "loss": 0.4413, + "num_input_tokens_seen": 32974976, + "step": 21670 + }, + { + "epoch": 0.6937776070674092, + "grad_norm": 1.363632082939148, + "learning_rate": 2.172256995535255e-05, + "loss": 0.4893, + "num_input_tokens_seen": 32982816, + "step": 21675 + }, + { + "epoch": 0.6939376480378977, + "grad_norm": 1.0437501668930054, + "learning_rate": 2.171283746081739e-05, + "loss": 0.4408, + "num_input_tokens_seen": 32990432, + "step": 21680 + }, + { + "epoch": 0.6940976890083862, + "grad_norm": 1.140380859375, + "learning_rate": 2.1703105473203988e-05, + "loss": 0.4113, + "num_input_tokens_seen": 32998000, + "step": 21685 + }, + { + "epoch": 0.6942577299788746, + "grad_norm": 0.6001764535903931, + "learning_rate": 2.1693373994013168e-05, + "loss": 0.4746, + "num_input_tokens_seen": 33005824, + "step": 21690 + }, + { + "epoch": 0.6944177709493631, + "grad_norm": 0.6294049620628357, + "learning_rate": 2.168364302474562e-05, + "loss": 0.6341, + "num_input_tokens_seen": 33013200, + "step": 21695 + }, + { + "epoch": 0.6945778119198515, + "grad_norm": 0.8919858932495117, + "learning_rate": 2.167391256690199e-05, + "loss": 0.4991, + "num_input_tokens_seen": 33020592, + "step": 21700 + }, + { + "epoch": 0.69473785289034, + "grad_norm": 0.5622825622558594, + "learning_rate": 2.1664182621982855e-05, + "loss": 0.4192, + "num_input_tokens_seen": 33028032, + "step": 21705 + }, + { + "epoch": 0.6948978938608283, + "grad_norm": 0.8135715126991272, + "learning_rate": 2.1654453191488673e-05, + "loss": 0.5078, + "num_input_tokens_seen": 33035392, + "step": 21710 + }, + { + "epoch": 0.6950579348313168, + "grad_norm": 0.4900316894054413, + "learning_rate": 2.1644724276919846e-05, + "loss": 0.4244, + "num_input_tokens_seen": 33042800, + "step": 21715 + }, + { + "epoch": 0.6952179758018052, + "grad_norm": 0.3570006489753723, + "learning_rate": 2.1634995879776715e-05, + "loss": 0.4906, + "num_input_tokens_seen": 33050064, + "step": 21720 + }, + { + "epoch": 0.6953780167722937, + "grad_norm": 0.8089873790740967, + "learning_rate": 2.162526800155949e-05, + "loss": 0.427, + "num_input_tokens_seen": 33057312, + "step": 21725 + }, + { + "epoch": 0.6955380577427821, + "grad_norm": 0.47706979513168335, + "learning_rate": 2.1615540643768363e-05, + "loss": 0.4238, + "num_input_tokens_seen": 33066000, + "step": 21730 + }, + { + "epoch": 0.6956980987132706, + "grad_norm": 0.5217772722244263, + "learning_rate": 2.160581380790339e-05, + "loss": 0.4692, + "num_input_tokens_seen": 33073952, + "step": 21735 + }, + { + "epoch": 0.6958581396837591, + "grad_norm": 0.789079487323761, + "learning_rate": 2.1596087495464586e-05, + "loss": 0.4909, + "num_input_tokens_seen": 33082192, + "step": 21740 + }, + { + "epoch": 0.6960181806542475, + "grad_norm": 1.2621755599975586, + "learning_rate": 2.1586361707951866e-05, + "loss": 0.5749, + "num_input_tokens_seen": 33090272, + "step": 21745 + }, + { + "epoch": 0.696178221624736, + "grad_norm": 0.9076764583587646, + "learning_rate": 2.157663644686507e-05, + "loss": 0.594, + "num_input_tokens_seen": 33097856, + "step": 21750 + }, + { + "epoch": 0.6963382625952244, + "grad_norm": 0.45582884550094604, + "learning_rate": 2.156691171370396e-05, + "loss": 0.3295, + "num_input_tokens_seen": 33105792, + "step": 21755 + }, + { + "epoch": 0.6964983035657129, + "grad_norm": 0.6847209334373474, + "learning_rate": 2.1557187509968195e-05, + "loss": 0.4333, + "num_input_tokens_seen": 33113520, + "step": 21760 + }, + { + "epoch": 0.6966583445362012, + "grad_norm": 0.857729971408844, + "learning_rate": 2.1547463837157382e-05, + "loss": 0.6298, + "num_input_tokens_seen": 33120800, + "step": 21765 + }, + { + "epoch": 0.6968183855066897, + "grad_norm": 0.6709650754928589, + "learning_rate": 2.1537740696771045e-05, + "loss": 0.4465, + "num_input_tokens_seen": 33128576, + "step": 21770 + }, + { + "epoch": 0.6969784264771781, + "grad_norm": 0.48401156067848206, + "learning_rate": 2.1528018090308587e-05, + "loss": 0.3888, + "num_input_tokens_seen": 33136608, + "step": 21775 + }, + { + "epoch": 0.6971384674476666, + "grad_norm": 0.5311548113822937, + "learning_rate": 2.151829601926938e-05, + "loss": 0.4937, + "num_input_tokens_seen": 33144224, + "step": 21780 + }, + { + "epoch": 0.697298508418155, + "grad_norm": 0.5295307636260986, + "learning_rate": 2.1508574485152684e-05, + "loss": 0.4011, + "num_input_tokens_seen": 33151680, + "step": 21785 + }, + { + "epoch": 0.6974585493886435, + "grad_norm": 0.6235116124153137, + "learning_rate": 2.1498853489457667e-05, + "loss": 0.6063, + "num_input_tokens_seen": 33159072, + "step": 21790 + }, + { + "epoch": 0.697618590359132, + "grad_norm": 0.7789545059204102, + "learning_rate": 2.1489133033683455e-05, + "loss": 0.4956, + "num_input_tokens_seen": 33166848, + "step": 21795 + }, + { + "epoch": 0.6977786313296204, + "grad_norm": 0.4486400783061981, + "learning_rate": 2.1479413119329038e-05, + "loss": 0.343, + "num_input_tokens_seen": 33175008, + "step": 21800 + }, + { + "epoch": 0.6977786313296204, + "eval_loss": 0.508938729763031, + "eval_runtime": 558.7445, + "eval_samples_per_second": 24.852, + "eval_steps_per_second": 12.426, + "num_input_tokens_seen": 33175008, + "step": 21800 + }, + { + "epoch": 0.6979386723001089, + "grad_norm": 0.6973618268966675, + "learning_rate": 2.1469693747893355e-05, + "loss": 0.3875, + "num_input_tokens_seen": 33182464, + "step": 21805 + }, + { + "epoch": 0.6980987132705972, + "grad_norm": 0.5364859700202942, + "learning_rate": 2.1459974920875274e-05, + "loss": 0.4416, + "num_input_tokens_seen": 33189936, + "step": 21810 + }, + { + "epoch": 0.6982587542410857, + "grad_norm": 0.43393102288246155, + "learning_rate": 2.145025663977354e-05, + "loss": 0.4396, + "num_input_tokens_seen": 33197472, + "step": 21815 + }, + { + "epoch": 0.6984187952115741, + "grad_norm": 1.9290997982025146, + "learning_rate": 2.1440538906086844e-05, + "loss": 0.6342, + "num_input_tokens_seen": 33204576, + "step": 21820 + }, + { + "epoch": 0.6985788361820626, + "grad_norm": 0.7487742304801941, + "learning_rate": 2.1430821721313782e-05, + "loss": 0.3206, + "num_input_tokens_seen": 33212352, + "step": 21825 + }, + { + "epoch": 0.698738877152551, + "grad_norm": 0.7494029402732849, + "learning_rate": 2.142110508695286e-05, + "loss": 0.5286, + "num_input_tokens_seen": 33220496, + "step": 21830 + }, + { + "epoch": 0.6988989181230395, + "grad_norm": 0.49205490946769714, + "learning_rate": 2.1411389004502515e-05, + "loss": 0.5643, + "num_input_tokens_seen": 33228176, + "step": 21835 + }, + { + "epoch": 0.699058959093528, + "grad_norm": 0.7967429757118225, + "learning_rate": 2.140167347546107e-05, + "loss": 0.5243, + "num_input_tokens_seen": 33235712, + "step": 21840 + }, + { + "epoch": 0.6992190000640164, + "grad_norm": 0.8874650001525879, + "learning_rate": 2.1391958501326793e-05, + "loss": 0.5116, + "num_input_tokens_seen": 33243488, + "step": 21845 + }, + { + "epoch": 0.6993790410345049, + "grad_norm": 1.245413899421692, + "learning_rate": 2.1382244083597873e-05, + "loss": 0.4206, + "num_input_tokens_seen": 33250704, + "step": 21850 + }, + { + "epoch": 0.6995390820049933, + "grad_norm": 0.778044581413269, + "learning_rate": 2.137253022377237e-05, + "loss": 0.4686, + "num_input_tokens_seen": 33257856, + "step": 21855 + }, + { + "epoch": 0.6996991229754818, + "grad_norm": 1.0468624830245972, + "learning_rate": 2.136281692334829e-05, + "loss": 0.513, + "num_input_tokens_seen": 33266320, + "step": 21860 + }, + { + "epoch": 0.6998591639459701, + "grad_norm": 0.7178104519844055, + "learning_rate": 2.135310418382356e-05, + "loss": 0.4204, + "num_input_tokens_seen": 33274064, + "step": 21865 + }, + { + "epoch": 0.7000192049164586, + "grad_norm": 0.6467424631118774, + "learning_rate": 2.134339200669598e-05, + "loss": 0.5056, + "num_input_tokens_seen": 33281616, + "step": 21870 + }, + { + "epoch": 0.700179245886947, + "grad_norm": 0.6376796364784241, + "learning_rate": 2.133368039346331e-05, + "loss": 0.5131, + "num_input_tokens_seen": 33289600, + "step": 21875 + }, + { + "epoch": 0.7003392868574355, + "grad_norm": 1.30674409866333, + "learning_rate": 2.1323969345623195e-05, + "loss": 0.5673, + "num_input_tokens_seen": 33297392, + "step": 21880 + }, + { + "epoch": 0.7004993278279239, + "grad_norm": 2.351201057434082, + "learning_rate": 2.1314258864673207e-05, + "loss": 0.3851, + "num_input_tokens_seen": 33304688, + "step": 21885 + }, + { + "epoch": 0.7006593687984124, + "grad_norm": 0.6786066889762878, + "learning_rate": 2.130454895211082e-05, + "loss": 0.4601, + "num_input_tokens_seen": 33312224, + "step": 21890 + }, + { + "epoch": 0.7008194097689009, + "grad_norm": 1.2529340982437134, + "learning_rate": 2.129483960943342e-05, + "loss": 0.4364, + "num_input_tokens_seen": 33319760, + "step": 21895 + }, + { + "epoch": 0.7009794507393893, + "grad_norm": 0.5238481163978577, + "learning_rate": 2.128513083813831e-05, + "loss": 0.4696, + "num_input_tokens_seen": 33327952, + "step": 21900 + }, + { + "epoch": 0.7011394917098778, + "grad_norm": 0.6787087917327881, + "learning_rate": 2.1275422639722724e-05, + "loss": 0.5395, + "num_input_tokens_seen": 33335664, + "step": 21905 + }, + { + "epoch": 0.7012995326803662, + "grad_norm": 0.6752010583877563, + "learning_rate": 2.126571501568376e-05, + "loss": 0.5284, + "num_input_tokens_seen": 33343872, + "step": 21910 + }, + { + "epoch": 0.7014595736508547, + "grad_norm": 0.5545326471328735, + "learning_rate": 2.1256007967518478e-05, + "loss": 0.3817, + "num_input_tokens_seen": 33351424, + "step": 21915 + }, + { + "epoch": 0.701619614621343, + "grad_norm": 0.6473385691642761, + "learning_rate": 2.124630149672381e-05, + "loss": 0.588, + "num_input_tokens_seen": 33359328, + "step": 21920 + }, + { + "epoch": 0.7017796555918315, + "grad_norm": 0.6198776960372925, + "learning_rate": 2.1236595604796624e-05, + "loss": 0.368, + "num_input_tokens_seen": 33366896, + "step": 21925 + }, + { + "epoch": 0.7019396965623199, + "grad_norm": 0.7023319602012634, + "learning_rate": 2.1226890293233693e-05, + "loss": 0.5439, + "num_input_tokens_seen": 33374048, + "step": 21930 + }, + { + "epoch": 0.7020997375328084, + "grad_norm": 1.893319010734558, + "learning_rate": 2.1217185563531694e-05, + "loss": 0.5603, + "num_input_tokens_seen": 33380944, + "step": 21935 + }, + { + "epoch": 0.7022597785032968, + "grad_norm": 0.7007039189338684, + "learning_rate": 2.120748141718721e-05, + "loss": 0.4897, + "num_input_tokens_seen": 33388592, + "step": 21940 + }, + { + "epoch": 0.7024198194737853, + "grad_norm": 0.7518835067749023, + "learning_rate": 2.1197777855696765e-05, + "loss": 0.458, + "num_input_tokens_seen": 33396272, + "step": 21945 + }, + { + "epoch": 0.7025798604442738, + "grad_norm": 0.741400957107544, + "learning_rate": 2.1188074880556746e-05, + "loss": 0.4459, + "num_input_tokens_seen": 33403632, + "step": 21950 + }, + { + "epoch": 0.7027399014147622, + "grad_norm": 0.9099222421646118, + "learning_rate": 2.1178372493263495e-05, + "loss": 0.5022, + "num_input_tokens_seen": 33411264, + "step": 21955 + }, + { + "epoch": 0.7028999423852507, + "grad_norm": 0.5396514534950256, + "learning_rate": 2.116867069531322e-05, + "loss": 0.3769, + "num_input_tokens_seen": 33419248, + "step": 21960 + }, + { + "epoch": 0.703059983355739, + "grad_norm": 0.6250927448272705, + "learning_rate": 2.1158969488202073e-05, + "loss": 0.3826, + "num_input_tokens_seen": 33426528, + "step": 21965 + }, + { + "epoch": 0.7032200243262275, + "grad_norm": 1.367824673652649, + "learning_rate": 2.114926887342611e-05, + "loss": 0.4674, + "num_input_tokens_seen": 33433984, + "step": 21970 + }, + { + "epoch": 0.7033800652967159, + "grad_norm": 1.3906915187835693, + "learning_rate": 2.113956885248127e-05, + "loss": 0.5953, + "num_input_tokens_seen": 33441952, + "step": 21975 + }, + { + "epoch": 0.7035401062672044, + "grad_norm": 1.4051212072372437, + "learning_rate": 2.112986942686342e-05, + "loss": 0.5203, + "num_input_tokens_seen": 33450288, + "step": 21980 + }, + { + "epoch": 0.7037001472376928, + "grad_norm": 0.7336240410804749, + "learning_rate": 2.112017059806835e-05, + "loss": 0.4376, + "num_input_tokens_seen": 33457808, + "step": 21985 + }, + { + "epoch": 0.7038601882081813, + "grad_norm": 0.5867712497711182, + "learning_rate": 2.1110472367591724e-05, + "loss": 0.5469, + "num_input_tokens_seen": 33465440, + "step": 21990 + }, + { + "epoch": 0.7040202291786697, + "grad_norm": 0.7584975957870483, + "learning_rate": 2.1100774736929145e-05, + "loss": 0.4999, + "num_input_tokens_seen": 33473008, + "step": 21995 + }, + { + "epoch": 0.7041802701491582, + "grad_norm": 0.3723980784416199, + "learning_rate": 2.10910777075761e-05, + "loss": 0.532, + "num_input_tokens_seen": 33481296, + "step": 22000 + }, + { + "epoch": 0.7041802701491582, + "eval_loss": 0.5087840557098389, + "eval_runtime": 558.8322, + "eval_samples_per_second": 24.848, + "eval_steps_per_second": 12.424, + "num_input_tokens_seen": 33481296, + "step": 22000 + }, + { + "epoch": 0.7043403111196467, + "grad_norm": 0.5569709539413452, + "learning_rate": 2.108138128102799e-05, + "loss": 0.6821, + "num_input_tokens_seen": 33488928, + "step": 22005 + }, + { + "epoch": 0.7045003520901351, + "grad_norm": 0.8305103778839111, + "learning_rate": 2.107168545878014e-05, + "loss": 0.5598, + "num_input_tokens_seen": 33496256, + "step": 22010 + }, + { + "epoch": 0.7046603930606236, + "grad_norm": 0.9459194540977478, + "learning_rate": 2.106199024232775e-05, + "loss": 0.5987, + "num_input_tokens_seen": 33503968, + "step": 22015 + }, + { + "epoch": 0.7048204340311119, + "grad_norm": 0.8983486890792847, + "learning_rate": 2.105229563316595e-05, + "loss": 0.5714, + "num_input_tokens_seen": 33512288, + "step": 22020 + }, + { + "epoch": 0.7049804750016004, + "grad_norm": 0.39597830176353455, + "learning_rate": 2.1042601632789784e-05, + "loss": 0.4577, + "num_input_tokens_seen": 33520480, + "step": 22025 + }, + { + "epoch": 0.7051405159720888, + "grad_norm": 0.4968871474266052, + "learning_rate": 2.103290824269417e-05, + "loss": 0.4865, + "num_input_tokens_seen": 33527792, + "step": 22030 + }, + { + "epoch": 0.7053005569425773, + "grad_norm": 1.0776945352554321, + "learning_rate": 2.1023215464373965e-05, + "loss": 0.6245, + "num_input_tokens_seen": 33535184, + "step": 22035 + }, + { + "epoch": 0.7054605979130657, + "grad_norm": 0.90141361951828, + "learning_rate": 2.1013523299323908e-05, + "loss": 0.488, + "num_input_tokens_seen": 33542112, + "step": 22040 + }, + { + "epoch": 0.7056206388835542, + "grad_norm": 0.5857739448547363, + "learning_rate": 2.1003831749038654e-05, + "loss": 0.4891, + "num_input_tokens_seen": 33549824, + "step": 22045 + }, + { + "epoch": 0.7057806798540426, + "grad_norm": 0.5840986967086792, + "learning_rate": 2.099414081501277e-05, + "loss": 0.3932, + "num_input_tokens_seen": 33556960, + "step": 22050 + }, + { + "epoch": 0.7059407208245311, + "grad_norm": 0.7652472257614136, + "learning_rate": 2.09844504987407e-05, + "loss": 0.347, + "num_input_tokens_seen": 33564736, + "step": 22055 + }, + { + "epoch": 0.7061007617950196, + "grad_norm": 0.8119410872459412, + "learning_rate": 2.097476080171683e-05, + "loss": 0.4802, + "num_input_tokens_seen": 33572528, + "step": 22060 + }, + { + "epoch": 0.706260802765508, + "grad_norm": 0.6021588444709778, + "learning_rate": 2.0965071725435436e-05, + "loss": 0.4763, + "num_input_tokens_seen": 33579680, + "step": 22065 + }, + { + "epoch": 0.7064208437359965, + "grad_norm": 0.6844158172607422, + "learning_rate": 2.0955383271390684e-05, + "loss": 0.4978, + "num_input_tokens_seen": 33587696, + "step": 22070 + }, + { + "epoch": 0.7065808847064848, + "grad_norm": 0.31636935472488403, + "learning_rate": 2.094569544107666e-05, + "loss": 0.3444, + "num_input_tokens_seen": 33595104, + "step": 22075 + }, + { + "epoch": 0.7067409256769733, + "grad_norm": 1.1929073333740234, + "learning_rate": 2.093600823598735e-05, + "loss": 0.492, + "num_input_tokens_seen": 33602560, + "step": 22080 + }, + { + "epoch": 0.7069009666474617, + "grad_norm": 2.9995815753936768, + "learning_rate": 2.092632165761663e-05, + "loss": 0.5246, + "num_input_tokens_seen": 33610160, + "step": 22085 + }, + { + "epoch": 0.7070610076179502, + "grad_norm": 1.7518563270568848, + "learning_rate": 2.091663570745832e-05, + "loss": 0.6329, + "num_input_tokens_seen": 33618512, + "step": 22090 + }, + { + "epoch": 0.7072210485884386, + "grad_norm": 1.1094433069229126, + "learning_rate": 2.0906950387006086e-05, + "loss": 0.5164, + "num_input_tokens_seen": 33626176, + "step": 22095 + }, + { + "epoch": 0.7073810895589271, + "grad_norm": 0.533163845539093, + "learning_rate": 2.0897265697753543e-05, + "loss": 0.357, + "num_input_tokens_seen": 33633728, + "step": 22100 + }, + { + "epoch": 0.7075411305294156, + "grad_norm": 0.6957634091377258, + "learning_rate": 2.088758164119419e-05, + "loss": 0.2849, + "num_input_tokens_seen": 33640992, + "step": 22105 + }, + { + "epoch": 0.707701171499904, + "grad_norm": 0.7685115933418274, + "learning_rate": 2.0877898218821428e-05, + "loss": 0.351, + "num_input_tokens_seen": 33648672, + "step": 22110 + }, + { + "epoch": 0.7078612124703925, + "grad_norm": 0.5350859761238098, + "learning_rate": 2.0868215432128565e-05, + "loss": 0.5331, + "num_input_tokens_seen": 33655792, + "step": 22115 + }, + { + "epoch": 0.7080212534408808, + "grad_norm": 0.9112591743469238, + "learning_rate": 2.0858533282608796e-05, + "loss": 0.4451, + "num_input_tokens_seen": 33663200, + "step": 22120 + }, + { + "epoch": 0.7081812944113693, + "grad_norm": 0.9768216013908386, + "learning_rate": 2.084885177175524e-05, + "loss": 0.5701, + "num_input_tokens_seen": 33670640, + "step": 22125 + }, + { + "epoch": 0.7083413353818577, + "grad_norm": 1.3054734468460083, + "learning_rate": 2.0839170901060917e-05, + "loss": 0.5634, + "num_input_tokens_seen": 33678480, + "step": 22130 + }, + { + "epoch": 0.7085013763523462, + "grad_norm": 0.8375928997993469, + "learning_rate": 2.082949067201872e-05, + "loss": 0.5383, + "num_input_tokens_seen": 33686016, + "step": 22135 + }, + { + "epoch": 0.7086614173228346, + "grad_norm": 0.5180899500846863, + "learning_rate": 2.0819811086121475e-05, + "loss": 0.3675, + "num_input_tokens_seen": 33693312, + "step": 22140 + }, + { + "epoch": 0.7088214582933231, + "grad_norm": 0.9198427796363831, + "learning_rate": 2.08101321448619e-05, + "loss": 0.4713, + "num_input_tokens_seen": 33700704, + "step": 22145 + }, + { + "epoch": 0.7089814992638115, + "grad_norm": 0.8218144178390503, + "learning_rate": 2.080045384973259e-05, + "loss": 0.6064, + "num_input_tokens_seen": 33707936, + "step": 22150 + }, + { + "epoch": 0.7091415402343, + "grad_norm": 0.4695010185241699, + "learning_rate": 2.0790776202226082e-05, + "loss": 0.5696, + "num_input_tokens_seen": 33715616, + "step": 22155 + }, + { + "epoch": 0.7093015812047885, + "grad_norm": 0.9382839798927307, + "learning_rate": 2.078109920383477e-05, + "loss": 0.3912, + "num_input_tokens_seen": 33722976, + "step": 22160 + }, + { + "epoch": 0.7094616221752769, + "grad_norm": 0.6483880281448364, + "learning_rate": 2.0771422856050978e-05, + "loss": 0.5175, + "num_input_tokens_seen": 33730880, + "step": 22165 + }, + { + "epoch": 0.7096216631457654, + "grad_norm": 0.9729284048080444, + "learning_rate": 2.076174716036693e-05, + "loss": 0.5376, + "num_input_tokens_seen": 33738064, + "step": 22170 + }, + { + "epoch": 0.7097817041162537, + "grad_norm": 1.1776556968688965, + "learning_rate": 2.075207211827472e-05, + "loss": 0.4716, + "num_input_tokens_seen": 33745584, + "step": 22175 + }, + { + "epoch": 0.7099417450867422, + "grad_norm": 0.31739452481269836, + "learning_rate": 2.074239773126638e-05, + "loss": 0.2597, + "num_input_tokens_seen": 33753376, + "step": 22180 + }, + { + "epoch": 0.7101017860572306, + "grad_norm": 1.7560615539550781, + "learning_rate": 2.073272400083382e-05, + "loss": 0.4132, + "num_input_tokens_seen": 33760688, + "step": 22185 + }, + { + "epoch": 0.7102618270277191, + "grad_norm": 0.9819889068603516, + "learning_rate": 2.072305092846883e-05, + "loss": 0.4773, + "num_input_tokens_seen": 33767904, + "step": 22190 + }, + { + "epoch": 0.7104218679982075, + "grad_norm": 0.441845178604126, + "learning_rate": 2.0713378515663152e-05, + "loss": 0.3973, + "num_input_tokens_seen": 33775264, + "step": 22195 + }, + { + "epoch": 0.710581908968696, + "grad_norm": 0.6232631206512451, + "learning_rate": 2.070370676390836e-05, + "loss": 0.3321, + "num_input_tokens_seen": 33782240, + "step": 22200 + }, + { + "epoch": 0.710581908968696, + "eval_loss": 0.508085310459137, + "eval_runtime": 559.2254, + "eval_samples_per_second": 24.831, + "eval_steps_per_second": 12.415, + "num_input_tokens_seen": 33782240, + "step": 22200 + }, + { + "epoch": 0.7107419499391844, + "grad_norm": 0.8496631979942322, + "learning_rate": 2.0694035674695974e-05, + "loss": 0.438, + "num_input_tokens_seen": 33790016, + "step": 22205 + }, + { + "epoch": 0.7109019909096729, + "grad_norm": 1.5657490491867065, + "learning_rate": 2.0684365249517416e-05, + "loss": 0.5387, + "num_input_tokens_seen": 33797184, + "step": 22210 + }, + { + "epoch": 0.7110620318801614, + "grad_norm": 0.49097493290901184, + "learning_rate": 2.067469548986396e-05, + "loss": 0.5889, + "num_input_tokens_seen": 33804672, + "step": 22215 + }, + { + "epoch": 0.7112220728506498, + "grad_norm": 0.6243382096290588, + "learning_rate": 2.066502639722681e-05, + "loss": 0.4357, + "num_input_tokens_seen": 33812240, + "step": 22220 + }, + { + "epoch": 0.7113821138211383, + "grad_norm": 1.176383376121521, + "learning_rate": 2.065535797309708e-05, + "loss": 0.5595, + "num_input_tokens_seen": 33820032, + "step": 22225 + }, + { + "epoch": 0.7115421547916266, + "grad_norm": 0.9640659093856812, + "learning_rate": 2.0645690218965736e-05, + "loss": 0.5381, + "num_input_tokens_seen": 33827792, + "step": 22230 + }, + { + "epoch": 0.7117021957621151, + "grad_norm": 1.1132491827011108, + "learning_rate": 2.063602313632369e-05, + "loss": 0.5211, + "num_input_tokens_seen": 33835184, + "step": 22235 + }, + { + "epoch": 0.7118622367326035, + "grad_norm": 0.5882983803749084, + "learning_rate": 2.0626356726661704e-05, + "loss": 0.4088, + "num_input_tokens_seen": 33843328, + "step": 22240 + }, + { + "epoch": 0.712022277703092, + "grad_norm": 0.9131066203117371, + "learning_rate": 2.0616690991470477e-05, + "loss": 0.4138, + "num_input_tokens_seen": 33850704, + "step": 22245 + }, + { + "epoch": 0.7121823186735804, + "grad_norm": 1.1170657873153687, + "learning_rate": 2.0607025932240595e-05, + "loss": 0.5756, + "num_input_tokens_seen": 33857776, + "step": 22250 + }, + { + "epoch": 0.7123423596440689, + "grad_norm": 0.5590413212776184, + "learning_rate": 2.059736155046251e-05, + "loss": 0.4661, + "num_input_tokens_seen": 33865744, + "step": 22255 + }, + { + "epoch": 0.7125024006145573, + "grad_norm": 0.45041972398757935, + "learning_rate": 2.0587697847626603e-05, + "loss": 0.394, + "num_input_tokens_seen": 33873088, + "step": 22260 + }, + { + "epoch": 0.7126624415850458, + "grad_norm": 0.5104215741157532, + "learning_rate": 2.057803482522314e-05, + "loss": 0.2919, + "num_input_tokens_seen": 33880736, + "step": 22265 + }, + { + "epoch": 0.7128224825555343, + "grad_norm": 0.5167564749717712, + "learning_rate": 2.056837248474227e-05, + "loss": 0.4751, + "num_input_tokens_seen": 33888960, + "step": 22270 + }, + { + "epoch": 0.7129825235260226, + "grad_norm": 0.74423748254776, + "learning_rate": 2.0558710827674064e-05, + "loss": 0.3673, + "num_input_tokens_seen": 33896288, + "step": 22275 + }, + { + "epoch": 0.7131425644965111, + "grad_norm": 0.8668872117996216, + "learning_rate": 2.054904985550845e-05, + "loss": 0.582, + "num_input_tokens_seen": 33903760, + "step": 22280 + }, + { + "epoch": 0.7133026054669995, + "grad_norm": 0.5482052564620972, + "learning_rate": 2.0539389569735287e-05, + "loss": 0.5203, + "num_input_tokens_seen": 33911200, + "step": 22285 + }, + { + "epoch": 0.713462646437488, + "grad_norm": 1.1756465435028076, + "learning_rate": 2.052972997184431e-05, + "loss": 0.7594, + "num_input_tokens_seen": 33918416, + "step": 22290 + }, + { + "epoch": 0.7136226874079764, + "grad_norm": 0.6611672639846802, + "learning_rate": 2.0520071063325146e-05, + "loss": 0.3921, + "num_input_tokens_seen": 33925808, + "step": 22295 + }, + { + "epoch": 0.7137827283784649, + "grad_norm": 0.9804672598838806, + "learning_rate": 2.051041284566732e-05, + "loss": 0.6768, + "num_input_tokens_seen": 33933152, + "step": 22300 + }, + { + "epoch": 0.7139427693489533, + "grad_norm": 0.7207421660423279, + "learning_rate": 2.050075532036026e-05, + "loss": 0.4666, + "num_input_tokens_seen": 33940464, + "step": 22305 + }, + { + "epoch": 0.7141028103194418, + "grad_norm": 0.5301462411880493, + "learning_rate": 2.0491098488893264e-05, + "loss": 0.5785, + "num_input_tokens_seen": 33948528, + "step": 22310 + }, + { + "epoch": 0.7142628512899302, + "grad_norm": 0.7764976024627686, + "learning_rate": 2.0481442352755546e-05, + "loss": 0.4747, + "num_input_tokens_seen": 33956048, + "step": 22315 + }, + { + "epoch": 0.7144228922604187, + "grad_norm": 0.7575382590293884, + "learning_rate": 2.0471786913436198e-05, + "loss": 0.4612, + "num_input_tokens_seen": 33963712, + "step": 22320 + }, + { + "epoch": 0.7145829332309072, + "grad_norm": 0.8149582147598267, + "learning_rate": 2.0462132172424218e-05, + "loss": 0.4456, + "num_input_tokens_seen": 33971344, + "step": 22325 + }, + { + "epoch": 0.7147429742013955, + "grad_norm": 0.4574781656265259, + "learning_rate": 2.0452478131208484e-05, + "loss": 0.5855, + "num_input_tokens_seen": 33978800, + "step": 22330 + }, + { + "epoch": 0.714903015171884, + "grad_norm": 1.114096999168396, + "learning_rate": 2.0442824791277765e-05, + "loss": 0.5349, + "num_input_tokens_seen": 33986160, + "step": 22335 + }, + { + "epoch": 0.7150630561423724, + "grad_norm": 0.4753103256225586, + "learning_rate": 2.0433172154120727e-05, + "loss": 0.3724, + "num_input_tokens_seen": 33994240, + "step": 22340 + }, + { + "epoch": 0.7152230971128609, + "grad_norm": 1.67015540599823, + "learning_rate": 2.0423520221225947e-05, + "loss": 0.509, + "num_input_tokens_seen": 34001680, + "step": 22345 + }, + { + "epoch": 0.7153831380833493, + "grad_norm": 0.786612331867218, + "learning_rate": 2.0413868994081848e-05, + "loss": 0.4788, + "num_input_tokens_seen": 34009552, + "step": 22350 + }, + { + "epoch": 0.7155431790538378, + "grad_norm": 0.6738588809967041, + "learning_rate": 2.0404218474176795e-05, + "loss": 0.4739, + "num_input_tokens_seen": 34017696, + "step": 22355 + }, + { + "epoch": 0.7157032200243262, + "grad_norm": 0.6828126907348633, + "learning_rate": 2.0394568662999002e-05, + "loss": 0.4966, + "num_input_tokens_seen": 34025008, + "step": 22360 + }, + { + "epoch": 0.7158632609948147, + "grad_norm": 0.7581323981285095, + "learning_rate": 2.0384919562036593e-05, + "loss": 0.4427, + "num_input_tokens_seen": 34033040, + "step": 22365 + }, + { + "epoch": 0.7160233019653032, + "grad_norm": 0.6633213758468628, + "learning_rate": 2.0375271172777593e-05, + "loss": 0.3603, + "num_input_tokens_seen": 34041424, + "step": 22370 + }, + { + "epoch": 0.7161833429357916, + "grad_norm": 0.8406950235366821, + "learning_rate": 2.0365623496709885e-05, + "loss": 0.4352, + "num_input_tokens_seen": 34048704, + "step": 22375 + }, + { + "epoch": 0.71634338390628, + "grad_norm": 0.6588326096534729, + "learning_rate": 2.0355976535321283e-05, + "loss": 0.4609, + "num_input_tokens_seen": 34056752, + "step": 22380 + }, + { + "epoch": 0.7165034248767684, + "grad_norm": 0.5673519372940063, + "learning_rate": 2.034633029009945e-05, + "loss": 0.3902, + "num_input_tokens_seen": 34064224, + "step": 22385 + }, + { + "epoch": 0.7166634658472569, + "grad_norm": 1.0354102849960327, + "learning_rate": 2.0336684762531972e-05, + "loss": 0.5973, + "num_input_tokens_seen": 34072576, + "step": 22390 + }, + { + "epoch": 0.7168235068177453, + "grad_norm": 1.4829163551330566, + "learning_rate": 2.032703995410631e-05, + "loss": 0.484, + "num_input_tokens_seen": 34080464, + "step": 22395 + }, + { + "epoch": 0.7169835477882338, + "grad_norm": 1.0543146133422852, + "learning_rate": 2.031739586630981e-05, + "loss": 0.6552, + "num_input_tokens_seen": 34088112, + "step": 22400 + }, + { + "epoch": 0.7169835477882338, + "eval_loss": 0.5079393982887268, + "eval_runtime": 559.0359, + "eval_samples_per_second": 24.839, + "eval_steps_per_second": 12.42, + "num_input_tokens_seen": 34088112, + "step": 22400 + }, + { + "epoch": 0.7171435887587222, + "grad_norm": 1.194173812866211, + "learning_rate": 2.0307752500629707e-05, + "loss": 0.593, + "num_input_tokens_seen": 34096032, + "step": 22405 + }, + { + "epoch": 0.7173036297292107, + "grad_norm": 0.5574042201042175, + "learning_rate": 2.0298109858553144e-05, + "loss": 0.5449, + "num_input_tokens_seen": 34103696, + "step": 22410 + }, + { + "epoch": 0.7174636706996991, + "grad_norm": 0.9770538210868835, + "learning_rate": 2.028846794156712e-05, + "loss": 0.5677, + "num_input_tokens_seen": 34112016, + "step": 22415 + }, + { + "epoch": 0.7176237116701876, + "grad_norm": 0.33129018545150757, + "learning_rate": 2.027882675115856e-05, + "loss": 0.3723, + "num_input_tokens_seen": 34120016, + "step": 22420 + }, + { + "epoch": 0.7177837526406761, + "grad_norm": 0.5127713680267334, + "learning_rate": 2.026918628881423e-05, + "loss": 0.4992, + "num_input_tokens_seen": 34127872, + "step": 22425 + }, + { + "epoch": 0.7179437936111644, + "grad_norm": 0.4131616950035095, + "learning_rate": 2.0259546556020833e-05, + "loss": 0.3352, + "num_input_tokens_seen": 34135312, + "step": 22430 + }, + { + "epoch": 0.7181038345816529, + "grad_norm": 0.51494961977005, + "learning_rate": 2.024990755426493e-05, + "loss": 0.4302, + "num_input_tokens_seen": 34143152, + "step": 22435 + }, + { + "epoch": 0.7182638755521413, + "grad_norm": 0.48748406767845154, + "learning_rate": 2.0240269285032975e-05, + "loss": 0.4245, + "num_input_tokens_seen": 34150672, + "step": 22440 + }, + { + "epoch": 0.7184239165226298, + "grad_norm": 0.4486156404018402, + "learning_rate": 2.0230631749811306e-05, + "loss": 0.4156, + "num_input_tokens_seen": 34158512, + "step": 22445 + }, + { + "epoch": 0.7185839574931182, + "grad_norm": 1.2713998556137085, + "learning_rate": 2.0220994950086162e-05, + "loss": 0.3924, + "num_input_tokens_seen": 34166016, + "step": 22450 + }, + { + "epoch": 0.7187439984636067, + "grad_norm": 0.7770536541938782, + "learning_rate": 2.021135888734365e-05, + "loss": 0.5115, + "num_input_tokens_seen": 34173792, + "step": 22455 + }, + { + "epoch": 0.7189040394340951, + "grad_norm": 0.92894047498703, + "learning_rate": 2.0201723563069783e-05, + "loss": 0.4056, + "num_input_tokens_seen": 34181280, + "step": 22460 + }, + { + "epoch": 0.7190640804045836, + "grad_norm": 0.8487383127212524, + "learning_rate": 2.0192088978750433e-05, + "loss": 0.5324, + "num_input_tokens_seen": 34188496, + "step": 22465 + }, + { + "epoch": 0.719224121375072, + "grad_norm": 0.8831579089164734, + "learning_rate": 2.0182455135871385e-05, + "loss": 0.5044, + "num_input_tokens_seen": 34195808, + "step": 22470 + }, + { + "epoch": 0.7193841623455605, + "grad_norm": 0.7652484774589539, + "learning_rate": 2.0172822035918305e-05, + "loss": 0.4128, + "num_input_tokens_seen": 34203376, + "step": 22475 + }, + { + "epoch": 0.719544203316049, + "grad_norm": 0.7657535076141357, + "learning_rate": 2.016318968037671e-05, + "loss": 0.3746, + "num_input_tokens_seen": 34210720, + "step": 22480 + }, + { + "epoch": 0.7197042442865373, + "grad_norm": 1.240238904953003, + "learning_rate": 2.015355807073206e-05, + "loss": 0.5835, + "num_input_tokens_seen": 34218592, + "step": 22485 + }, + { + "epoch": 0.7198642852570258, + "grad_norm": 0.6178722977638245, + "learning_rate": 2.0143927208469664e-05, + "loss": 0.2784, + "num_input_tokens_seen": 34225936, + "step": 22490 + }, + { + "epoch": 0.7200243262275142, + "grad_norm": 0.5641149282455444, + "learning_rate": 2.0134297095074708e-05, + "loss": 0.5236, + "num_input_tokens_seen": 34233648, + "step": 22495 + }, + { + "epoch": 0.7201843671980027, + "grad_norm": 0.8188678622245789, + "learning_rate": 2.0124667732032297e-05, + "loss": 0.6196, + "num_input_tokens_seen": 34241472, + "step": 22500 + }, + { + "epoch": 0.7203444081684911, + "grad_norm": 0.5442745685577393, + "learning_rate": 2.011503912082738e-05, + "loss": 0.4895, + "num_input_tokens_seen": 34248768, + "step": 22505 + }, + { + "epoch": 0.7205044491389796, + "grad_norm": 0.6356115937232971, + "learning_rate": 2.0105411262944823e-05, + "loss": 0.2897, + "num_input_tokens_seen": 34256176, + "step": 22510 + }, + { + "epoch": 0.720664490109468, + "grad_norm": 0.61234450340271, + "learning_rate": 2.0095784159869366e-05, + "loss": 0.4108, + "num_input_tokens_seen": 34263424, + "step": 22515 + }, + { + "epoch": 0.7208245310799565, + "grad_norm": 1.1303139925003052, + "learning_rate": 2.0086157813085608e-05, + "loss": 0.4722, + "num_input_tokens_seen": 34270960, + "step": 22520 + }, + { + "epoch": 0.7209845720504449, + "grad_norm": 0.3914656341075897, + "learning_rate": 2.0076532224078068e-05, + "loss": 0.2434, + "num_input_tokens_seen": 34278320, + "step": 22525 + }, + { + "epoch": 0.7211446130209334, + "grad_norm": 0.7200520634651184, + "learning_rate": 2.0066907394331142e-05, + "loss": 0.4748, + "num_input_tokens_seen": 34285872, + "step": 22530 + }, + { + "epoch": 0.7213046539914219, + "grad_norm": 0.7079151272773743, + "learning_rate": 2.0057283325329077e-05, + "loss": 0.4008, + "num_input_tokens_seen": 34293104, + "step": 22535 + }, + { + "epoch": 0.7214646949619102, + "grad_norm": 0.4602554738521576, + "learning_rate": 2.0047660018556047e-05, + "loss": 0.4971, + "num_input_tokens_seen": 34300336, + "step": 22540 + }, + { + "epoch": 0.7216247359323987, + "grad_norm": 1.5303434133529663, + "learning_rate": 2.0038037475496075e-05, + "loss": 0.6245, + "num_input_tokens_seen": 34308000, + "step": 22545 + }, + { + "epoch": 0.7217847769028871, + "grad_norm": 0.8987612128257751, + "learning_rate": 2.0028415697633073e-05, + "loss": 0.3495, + "num_input_tokens_seen": 34314976, + "step": 22550 + }, + { + "epoch": 0.7219448178733756, + "grad_norm": 0.8892344236373901, + "learning_rate": 2.0018794686450858e-05, + "loss": 0.4683, + "num_input_tokens_seen": 34321904, + "step": 22555 + }, + { + "epoch": 0.722104858843864, + "grad_norm": 0.5452123880386353, + "learning_rate": 2.0009174443433088e-05, + "loss": 0.3591, + "num_input_tokens_seen": 34329568, + "step": 22560 + }, + { + "epoch": 0.7222648998143525, + "grad_norm": 0.7064860463142395, + "learning_rate": 1.999955497006334e-05, + "loss": 0.5363, + "num_input_tokens_seen": 34337168, + "step": 22565 + }, + { + "epoch": 0.7224249407848409, + "grad_norm": 1.6636414527893066, + "learning_rate": 1.9989936267825067e-05, + "loss": 0.6548, + "num_input_tokens_seen": 34344656, + "step": 22570 + }, + { + "epoch": 0.7225849817553294, + "grad_norm": 1.1661372184753418, + "learning_rate": 1.9980318338201572e-05, + "loss": 0.4826, + "num_input_tokens_seen": 34352704, + "step": 22575 + }, + { + "epoch": 0.7227450227258178, + "grad_norm": 0.6838195323944092, + "learning_rate": 1.997070118267607e-05, + "loss": 0.4868, + "num_input_tokens_seen": 34360144, + "step": 22580 + }, + { + "epoch": 0.7229050636963062, + "grad_norm": 0.716143012046814, + "learning_rate": 1.9961084802731654e-05, + "loss": 0.5796, + "num_input_tokens_seen": 34367888, + "step": 22585 + }, + { + "epoch": 0.7230651046667947, + "grad_norm": 0.6396598219871521, + "learning_rate": 1.9951469199851273e-05, + "loss": 0.3457, + "num_input_tokens_seen": 34375200, + "step": 22590 + }, + { + "epoch": 0.7232251456372831, + "grad_norm": 0.620939314365387, + "learning_rate": 1.99418543755178e-05, + "loss": 0.3658, + "num_input_tokens_seen": 34383040, + "step": 22595 + }, + { + "epoch": 0.7233851866077716, + "grad_norm": 0.7441362142562866, + "learning_rate": 1.9932240331213936e-05, + "loss": 0.5395, + "num_input_tokens_seen": 34390944, + "step": 22600 + }, + { + "epoch": 0.7233851866077716, + "eval_loss": 0.5078697204589844, + "eval_runtime": 559.1154, + "eval_samples_per_second": 24.836, + "eval_steps_per_second": 12.418, + "num_input_tokens_seen": 34390944, + "step": 22600 + }, + { + "epoch": 0.72354522757826, + "grad_norm": 1.389042615890503, + "learning_rate": 1.9922627068422297e-05, + "loss": 0.5219, + "num_input_tokens_seen": 34398528, + "step": 22605 + }, + { + "epoch": 0.7237052685487485, + "grad_norm": 0.7514331936836243, + "learning_rate": 1.991301458862538e-05, + "loss": 0.5089, + "num_input_tokens_seen": 34406400, + "step": 22610 + }, + { + "epoch": 0.7238653095192369, + "grad_norm": 0.7377894520759583, + "learning_rate": 1.9903402893305536e-05, + "loss": 0.4845, + "num_input_tokens_seen": 34413568, + "step": 22615 + }, + { + "epoch": 0.7240253504897254, + "grad_norm": 0.639964759349823, + "learning_rate": 1.9893791983945016e-05, + "loss": 0.4845, + "num_input_tokens_seen": 34421248, + "step": 22620 + }, + { + "epoch": 0.7241853914602138, + "grad_norm": 1.2581744194030762, + "learning_rate": 1.988418186202594e-05, + "loss": 0.4317, + "num_input_tokens_seen": 34428800, + "step": 22625 + }, + { + "epoch": 0.7243454324307023, + "grad_norm": 2.324370861053467, + "learning_rate": 1.98745725290303e-05, + "loss": 0.701, + "num_input_tokens_seen": 34436592, + "step": 22630 + }, + { + "epoch": 0.7245054734011908, + "grad_norm": 0.8177740573883057, + "learning_rate": 1.986496398644e-05, + "loss": 0.4104, + "num_input_tokens_seen": 34444368, + "step": 22635 + }, + { + "epoch": 0.7246655143716791, + "grad_norm": 0.9811644554138184, + "learning_rate": 1.9855356235736777e-05, + "loss": 0.5304, + "num_input_tokens_seen": 34452400, + "step": 22640 + }, + { + "epoch": 0.7248255553421676, + "grad_norm": 0.9892751574516296, + "learning_rate": 1.9845749278402277e-05, + "loss": 0.5182, + "num_input_tokens_seen": 34459632, + "step": 22645 + }, + { + "epoch": 0.724985596312656, + "grad_norm": 0.6585812568664551, + "learning_rate": 1.9836143115918006e-05, + "loss": 0.5558, + "num_input_tokens_seen": 34467504, + "step": 22650 + }, + { + "epoch": 0.7251456372831445, + "grad_norm": 0.7478508949279785, + "learning_rate": 1.9826537749765367e-05, + "loss": 0.6307, + "num_input_tokens_seen": 34475312, + "step": 22655 + }, + { + "epoch": 0.7253056782536329, + "grad_norm": 0.7143306732177734, + "learning_rate": 1.9816933181425625e-05, + "loss": 0.3582, + "num_input_tokens_seen": 34482576, + "step": 22660 + }, + { + "epoch": 0.7254657192241214, + "grad_norm": 0.5083871483802795, + "learning_rate": 1.9807329412379903e-05, + "loss": 0.5576, + "num_input_tokens_seen": 34490192, + "step": 22665 + }, + { + "epoch": 0.7256257601946098, + "grad_norm": 0.6366752982139587, + "learning_rate": 1.9797726444109247e-05, + "loss": 0.4129, + "num_input_tokens_seen": 34497696, + "step": 22670 + }, + { + "epoch": 0.7257858011650983, + "grad_norm": 0.8245328068733215, + "learning_rate": 1.9788124278094557e-05, + "loss": 0.3668, + "num_input_tokens_seen": 34505312, + "step": 22675 + }, + { + "epoch": 0.7259458421355867, + "grad_norm": 0.5816821455955505, + "learning_rate": 1.9778522915816594e-05, + "loss": 0.3282, + "num_input_tokens_seen": 34512768, + "step": 22680 + }, + { + "epoch": 0.7261058831060752, + "grad_norm": 1.594311237335205, + "learning_rate": 1.9768922358756014e-05, + "loss": 0.5936, + "num_input_tokens_seen": 34520048, + "step": 22685 + }, + { + "epoch": 0.7262659240765637, + "grad_norm": 0.6435288190841675, + "learning_rate": 1.9759322608393353e-05, + "loss": 0.3969, + "num_input_tokens_seen": 34527552, + "step": 22690 + }, + { + "epoch": 0.726425965047052, + "grad_norm": 0.998565137386322, + "learning_rate": 1.9749723666208992e-05, + "loss": 0.6097, + "num_input_tokens_seen": 34535056, + "step": 22695 + }, + { + "epoch": 0.7265860060175405, + "grad_norm": 0.7619953751564026, + "learning_rate": 1.9740125533683235e-05, + "loss": 0.5635, + "num_input_tokens_seen": 34542768, + "step": 22700 + }, + { + "epoch": 0.7267460469880289, + "grad_norm": 1.096339464187622, + "learning_rate": 1.9730528212296208e-05, + "loss": 0.6855, + "num_input_tokens_seen": 34550400, + "step": 22705 + }, + { + "epoch": 0.7269060879585174, + "grad_norm": 1.5223677158355713, + "learning_rate": 1.9720931703527945e-05, + "loss": 0.5095, + "num_input_tokens_seen": 34557856, + "step": 22710 + }, + { + "epoch": 0.7270661289290058, + "grad_norm": 1.354120135307312, + "learning_rate": 1.9711336008858373e-05, + "loss": 0.6294, + "num_input_tokens_seen": 34565600, + "step": 22715 + }, + { + "epoch": 0.7272261698994943, + "grad_norm": 1.312329888343811, + "learning_rate": 1.9701741129767233e-05, + "loss": 0.8478, + "num_input_tokens_seen": 34573008, + "step": 22720 + }, + { + "epoch": 0.7273862108699827, + "grad_norm": 0.5174224376678467, + "learning_rate": 1.9692147067734202e-05, + "loss": 0.3712, + "num_input_tokens_seen": 34580432, + "step": 22725 + }, + { + "epoch": 0.7275462518404712, + "grad_norm": 1.0201789140701294, + "learning_rate": 1.96825538242388e-05, + "loss": 0.3789, + "num_input_tokens_seen": 34588608, + "step": 22730 + }, + { + "epoch": 0.7277062928109596, + "grad_norm": 0.9448941946029663, + "learning_rate": 1.967296140076041e-05, + "loss": 0.5164, + "num_input_tokens_seen": 34596656, + "step": 22735 + }, + { + "epoch": 0.727866333781448, + "grad_norm": 1.2309643030166626, + "learning_rate": 1.966336979877833e-05, + "loss": 0.5417, + "num_input_tokens_seen": 34604128, + "step": 22740 + }, + { + "epoch": 0.7280263747519365, + "grad_norm": 6.8389739990234375, + "learning_rate": 1.9653779019771678e-05, + "loss": 0.4291, + "num_input_tokens_seen": 34611744, + "step": 22745 + }, + { + "epoch": 0.7281864157224249, + "grad_norm": 1.8634779453277588, + "learning_rate": 1.9644189065219488e-05, + "loss": 0.5338, + "num_input_tokens_seen": 34619440, + "step": 22750 + }, + { + "epoch": 0.7283464566929134, + "grad_norm": 0.9900765419006348, + "learning_rate": 1.9634599936600655e-05, + "loss": 0.5732, + "num_input_tokens_seen": 34626784, + "step": 22755 + }, + { + "epoch": 0.7285064976634018, + "grad_norm": 0.736758828163147, + "learning_rate": 1.9625011635393935e-05, + "loss": 0.584, + "num_input_tokens_seen": 34634464, + "step": 22760 + }, + { + "epoch": 0.7286665386338903, + "grad_norm": 0.49907130002975464, + "learning_rate": 1.9615424163077963e-05, + "loss": 0.4509, + "num_input_tokens_seen": 34641792, + "step": 22765 + }, + { + "epoch": 0.7288265796043787, + "grad_norm": 0.529428243637085, + "learning_rate": 1.9605837521131263e-05, + "loss": 0.3784, + "num_input_tokens_seen": 34649712, + "step": 22770 + }, + { + "epoch": 0.7289866205748672, + "grad_norm": 0.6586135625839233, + "learning_rate": 1.9596251711032192e-05, + "loss": 0.4883, + "num_input_tokens_seen": 34657232, + "step": 22775 + }, + { + "epoch": 0.7291466615453556, + "grad_norm": 0.6529535055160522, + "learning_rate": 1.958666673425903e-05, + "loss": 0.4518, + "num_input_tokens_seen": 34665600, + "step": 22780 + }, + { + "epoch": 0.7293067025158441, + "grad_norm": 0.5122243762016296, + "learning_rate": 1.957708259228987e-05, + "loss": 0.4254, + "num_input_tokens_seen": 34672976, + "step": 22785 + }, + { + "epoch": 0.7294667434863324, + "grad_norm": 0.7060846090316772, + "learning_rate": 1.956749928660273e-05, + "loss": 0.3785, + "num_input_tokens_seen": 34681104, + "step": 22790 + }, + { + "epoch": 0.7296267844568209, + "grad_norm": 0.999360203742981, + "learning_rate": 1.955791681867547e-05, + "loss": 0.3896, + "num_input_tokens_seen": 34688880, + "step": 22795 + }, + { + "epoch": 0.7297868254273094, + "grad_norm": 0.4424268305301666, + "learning_rate": 1.9548335189985824e-05, + "loss": 0.4636, + "num_input_tokens_seen": 34696480, + "step": 22800 + }, + { + "epoch": 0.7297868254273094, + "eval_loss": 0.5070478320121765, + "eval_runtime": 559.1062, + "eval_samples_per_second": 24.836, + "eval_steps_per_second": 12.418, + "num_input_tokens_seen": 34696480, + "step": 22800 + }, + { + "epoch": 0.7299468663977978, + "grad_norm": 1.768338680267334, + "learning_rate": 1.9538754402011396e-05, + "loss": 0.5087, + "num_input_tokens_seen": 34704096, + "step": 22805 + }, + { + "epoch": 0.7301069073682863, + "grad_norm": 1.8089261054992676, + "learning_rate": 1.952917445622968e-05, + "loss": 0.4969, + "num_input_tokens_seen": 34711888, + "step": 22810 + }, + { + "epoch": 0.7302669483387747, + "grad_norm": 0.5067367553710938, + "learning_rate": 1.9519595354118005e-05, + "loss": 0.4365, + "num_input_tokens_seen": 34718880, + "step": 22815 + }, + { + "epoch": 0.7304269893092632, + "grad_norm": 0.788169801235199, + "learning_rate": 1.951001709715361e-05, + "loss": 0.4317, + "num_input_tokens_seen": 34725952, + "step": 22820 + }, + { + "epoch": 0.7305870302797516, + "grad_norm": 0.4445969760417938, + "learning_rate": 1.9500439686813556e-05, + "loss": 0.3223, + "num_input_tokens_seen": 34733184, + "step": 22825 + }, + { + "epoch": 0.7307470712502401, + "grad_norm": 0.4945024847984314, + "learning_rate": 1.949086312457482e-05, + "loss": 0.4693, + "num_input_tokens_seen": 34740688, + "step": 22830 + }, + { + "epoch": 0.7309071122207285, + "grad_norm": 0.616067111492157, + "learning_rate": 1.9481287411914223e-05, + "loss": 0.5393, + "num_input_tokens_seen": 34748176, + "step": 22835 + }, + { + "epoch": 0.731067153191217, + "grad_norm": 1.030326008796692, + "learning_rate": 1.9471712550308457e-05, + "loss": 0.6471, + "num_input_tokens_seen": 34755904, + "step": 22840 + }, + { + "epoch": 0.7312271941617055, + "grad_norm": 0.6437064409255981, + "learning_rate": 1.946213854123409e-05, + "loss": 0.3765, + "num_input_tokens_seen": 34763760, + "step": 22845 + }, + { + "epoch": 0.7313872351321938, + "grad_norm": 0.8549103140830994, + "learning_rate": 1.9452565386167554e-05, + "loss": 0.5357, + "num_input_tokens_seen": 34771440, + "step": 22850 + }, + { + "epoch": 0.7315472761026823, + "grad_norm": 0.7707763314247131, + "learning_rate": 1.9442993086585142e-05, + "loss": 0.5188, + "num_input_tokens_seen": 34778768, + "step": 22855 + }, + { + "epoch": 0.7317073170731707, + "grad_norm": 0.8189672231674194, + "learning_rate": 1.9433421643963043e-05, + "loss": 0.4203, + "num_input_tokens_seen": 34786544, + "step": 22860 + }, + { + "epoch": 0.7318673580436592, + "grad_norm": 0.5449572801589966, + "learning_rate": 1.942385105977727e-05, + "loss": 0.3599, + "num_input_tokens_seen": 34793920, + "step": 22865 + }, + { + "epoch": 0.7320273990141476, + "grad_norm": 0.8688105344772339, + "learning_rate": 1.9414281335503743e-05, + "loss": 0.5468, + "num_input_tokens_seen": 34801344, + "step": 22870 + }, + { + "epoch": 0.7321874399846361, + "grad_norm": 0.5548898577690125, + "learning_rate": 1.9404712472618232e-05, + "loss": 0.3772, + "num_input_tokens_seen": 34808912, + "step": 22875 + }, + { + "epoch": 0.7323474809551245, + "grad_norm": 0.7992070317268372, + "learning_rate": 1.939514447259636e-05, + "loss": 0.4236, + "num_input_tokens_seen": 34816624, + "step": 22880 + }, + { + "epoch": 0.732507521925613, + "grad_norm": 0.5693809390068054, + "learning_rate": 1.938557733691365e-05, + "loss": 0.5593, + "num_input_tokens_seen": 34823904, + "step": 22885 + }, + { + "epoch": 0.7326675628961014, + "grad_norm": 0.7734145522117615, + "learning_rate": 1.9376011067045476e-05, + "loss": 0.3547, + "num_input_tokens_seen": 34831520, + "step": 22890 + }, + { + "epoch": 0.7328276038665898, + "grad_norm": 0.6480451822280884, + "learning_rate": 1.9366445664467065e-05, + "loss": 0.5632, + "num_input_tokens_seen": 34839088, + "step": 22895 + }, + { + "epoch": 0.7329876448370783, + "grad_norm": 1.2981826066970825, + "learning_rate": 1.9356881130653533e-05, + "loss": 0.5131, + "num_input_tokens_seen": 34847120, + "step": 22900 + }, + { + "epoch": 0.7331476858075667, + "grad_norm": 0.5689300894737244, + "learning_rate": 1.9347317467079846e-05, + "loss": 0.4287, + "num_input_tokens_seen": 34854448, + "step": 22905 + }, + { + "epoch": 0.7333077267780552, + "grad_norm": 0.6801109910011292, + "learning_rate": 1.9337754675220836e-05, + "loss": 0.4389, + "num_input_tokens_seen": 34861840, + "step": 22910 + }, + { + "epoch": 0.7334677677485436, + "grad_norm": 0.47588372230529785, + "learning_rate": 1.9328192756551218e-05, + "loss": 0.4663, + "num_input_tokens_seen": 34869344, + "step": 22915 + }, + { + "epoch": 0.7336278087190321, + "grad_norm": 0.96588134765625, + "learning_rate": 1.931863171254555e-05, + "loss": 0.4953, + "num_input_tokens_seen": 34877152, + "step": 22920 + }, + { + "epoch": 0.7337878496895205, + "grad_norm": 0.9175223112106323, + "learning_rate": 1.930907154467826e-05, + "loss": 0.4677, + "num_input_tokens_seen": 34884624, + "step": 22925 + }, + { + "epoch": 0.733947890660009, + "grad_norm": 0.6887897849082947, + "learning_rate": 1.9299512254423673e-05, + "loss": 0.5474, + "num_input_tokens_seen": 34892096, + "step": 22930 + }, + { + "epoch": 0.7341079316304974, + "grad_norm": 0.6480000019073486, + "learning_rate": 1.9289953843255914e-05, + "loss": 0.5558, + "num_input_tokens_seen": 34899728, + "step": 22935 + }, + { + "epoch": 0.7342679726009859, + "grad_norm": 1.7181469202041626, + "learning_rate": 1.9280396312649048e-05, + "loss": 0.441, + "num_input_tokens_seen": 34907520, + "step": 22940 + }, + { + "epoch": 0.7344280135714742, + "grad_norm": 0.7467706799507141, + "learning_rate": 1.9270839664076936e-05, + "loss": 0.3287, + "num_input_tokens_seen": 34915008, + "step": 22945 + }, + { + "epoch": 0.7345880545419627, + "grad_norm": 0.46681976318359375, + "learning_rate": 1.9261283899013345e-05, + "loss": 0.4669, + "num_input_tokens_seen": 34922736, + "step": 22950 + }, + { + "epoch": 0.7347480955124512, + "grad_norm": 0.7186853885650635, + "learning_rate": 1.92517290189319e-05, + "loss": 0.4434, + "num_input_tokens_seen": 34930448, + "step": 22955 + }, + { + "epoch": 0.7349081364829396, + "grad_norm": 2.0451436042785645, + "learning_rate": 1.924217502530607e-05, + "loss": 0.7429, + "num_input_tokens_seen": 34937568, + "step": 22960 + }, + { + "epoch": 0.7350681774534281, + "grad_norm": 0.6653398275375366, + "learning_rate": 1.9232621919609207e-05, + "loss": 0.4833, + "num_input_tokens_seen": 34944864, + "step": 22965 + }, + { + "epoch": 0.7352282184239165, + "grad_norm": 0.5848659873008728, + "learning_rate": 1.9223069703314534e-05, + "loss": 0.4897, + "num_input_tokens_seen": 34952192, + "step": 22970 + }, + { + "epoch": 0.735388259394405, + "grad_norm": 0.5007315278053284, + "learning_rate": 1.92135183778951e-05, + "loss": 0.3544, + "num_input_tokens_seen": 34959792, + "step": 22975 + }, + { + "epoch": 0.7355483003648934, + "grad_norm": 0.9800735712051392, + "learning_rate": 1.9203967944823857e-05, + "loss": 0.4103, + "num_input_tokens_seen": 34967440, + "step": 22980 + }, + { + "epoch": 0.7357083413353819, + "grad_norm": 0.9132712483406067, + "learning_rate": 1.9194418405573588e-05, + "loss": 0.5308, + "num_input_tokens_seen": 34974848, + "step": 22985 + }, + { + "epoch": 0.7358683823058703, + "grad_norm": 0.5992487668991089, + "learning_rate": 1.9184869761616954e-05, + "loss": 0.5219, + "num_input_tokens_seen": 34982816, + "step": 22990 + }, + { + "epoch": 0.7360284232763588, + "grad_norm": 0.46083196997642517, + "learning_rate": 1.9175322014426495e-05, + "loss": 0.5539, + "num_input_tokens_seen": 34990896, + "step": 22995 + }, + { + "epoch": 0.7361884642468471, + "grad_norm": 0.8258150219917297, + "learning_rate": 1.9165775165474565e-05, + "loss": 0.5396, + "num_input_tokens_seen": 34998048, + "step": 23000 + }, + { + "epoch": 0.7361884642468471, + "eval_loss": 0.5067988038063049, + "eval_runtime": 559.2262, + "eval_samples_per_second": 24.831, + "eval_steps_per_second": 12.415, + "num_input_tokens_seen": 34998048, + "step": 23000 + }, + { + "epoch": 0.7363485052173356, + "grad_norm": 0.7802894115447998, + "learning_rate": 1.9156229216233434e-05, + "loss": 0.57, + "num_input_tokens_seen": 35005376, + "step": 23005 + }, + { + "epoch": 0.7365085461878241, + "grad_norm": 0.5691549181938171, + "learning_rate": 1.9146684168175184e-05, + "loss": 0.5996, + "num_input_tokens_seen": 35012880, + "step": 23010 + }, + { + "epoch": 0.7366685871583125, + "grad_norm": 0.6220206618309021, + "learning_rate": 1.9137140022771796e-05, + "loss": 0.4999, + "num_input_tokens_seen": 35020736, + "step": 23015 + }, + { + "epoch": 0.736828628128801, + "grad_norm": 0.6062358617782593, + "learning_rate": 1.9127596781495103e-05, + "loss": 0.5829, + "num_input_tokens_seen": 35028704, + "step": 23020 + }, + { + "epoch": 0.7369886690992894, + "grad_norm": 0.7758865356445312, + "learning_rate": 1.9118054445816767e-05, + "loss": 0.4913, + "num_input_tokens_seen": 35036720, + "step": 23025 + }, + { + "epoch": 0.7371487100697779, + "grad_norm": 0.48847445845603943, + "learning_rate": 1.9108513017208356e-05, + "loss": 0.3793, + "num_input_tokens_seen": 35043904, + "step": 23030 + }, + { + "epoch": 0.7373087510402663, + "grad_norm": 0.6822643876075745, + "learning_rate": 1.9098972497141287e-05, + "loss": 0.3822, + "num_input_tokens_seen": 35053200, + "step": 23035 + }, + { + "epoch": 0.7374687920107548, + "grad_norm": 0.478767454624176, + "learning_rate": 1.9089432887086806e-05, + "loss": 0.3627, + "num_input_tokens_seen": 35060624, + "step": 23040 + }, + { + "epoch": 0.7376288329812432, + "grad_norm": 0.7281224131584167, + "learning_rate": 1.9079894188516056e-05, + "loss": 0.4274, + "num_input_tokens_seen": 35068640, + "step": 23045 + }, + { + "epoch": 0.7377888739517316, + "grad_norm": 0.3620925545692444, + "learning_rate": 1.907035640290002e-05, + "loss": 0.3472, + "num_input_tokens_seen": 35076064, + "step": 23050 + }, + { + "epoch": 0.73794891492222, + "grad_norm": 0.5882554650306702, + "learning_rate": 1.9060819531709534e-05, + "loss": 0.5589, + "num_input_tokens_seen": 35083152, + "step": 23055 + }, + { + "epoch": 0.7381089558927085, + "grad_norm": 1.2953888177871704, + "learning_rate": 1.9051283576415325e-05, + "loss": 0.4407, + "num_input_tokens_seen": 35090464, + "step": 23060 + }, + { + "epoch": 0.738268996863197, + "grad_norm": 0.6065852642059326, + "learning_rate": 1.904174853848793e-05, + "loss": 0.5156, + "num_input_tokens_seen": 35098128, + "step": 23065 + }, + { + "epoch": 0.7384290378336854, + "grad_norm": 0.5611146688461304, + "learning_rate": 1.903221441939779e-05, + "loss": 0.4006, + "num_input_tokens_seen": 35105632, + "step": 23070 + }, + { + "epoch": 0.7385890788041739, + "grad_norm": 0.6618630886077881, + "learning_rate": 1.9022681220615194e-05, + "loss": 0.3576, + "num_input_tokens_seen": 35113520, + "step": 23075 + }, + { + "epoch": 0.7387491197746623, + "grad_norm": 0.9452515840530396, + "learning_rate": 1.9013148943610255e-05, + "loss": 0.4008, + "num_input_tokens_seen": 35121024, + "step": 23080 + }, + { + "epoch": 0.7389091607451508, + "grad_norm": 0.519498348236084, + "learning_rate": 1.9003617589852998e-05, + "loss": 0.4191, + "num_input_tokens_seen": 35128496, + "step": 23085 + }, + { + "epoch": 0.7390692017156392, + "grad_norm": 1.0853321552276611, + "learning_rate": 1.899408716081326e-05, + "loss": 0.6501, + "num_input_tokens_seen": 35135984, + "step": 23090 + }, + { + "epoch": 0.7392292426861277, + "grad_norm": 0.3500109016895294, + "learning_rate": 1.898455765796075e-05, + "loss": 0.4188, + "num_input_tokens_seen": 35143968, + "step": 23095 + }, + { + "epoch": 0.739389283656616, + "grad_norm": 0.3543640077114105, + "learning_rate": 1.8975029082765053e-05, + "loss": 0.5563, + "num_input_tokens_seen": 35150864, + "step": 23100 + }, + { + "epoch": 0.7395493246271045, + "grad_norm": 0.8756570816040039, + "learning_rate": 1.8965501436695577e-05, + "loss": 0.6043, + "num_input_tokens_seen": 35158240, + "step": 23105 + }, + { + "epoch": 0.739709365597593, + "grad_norm": 0.6916809678077698, + "learning_rate": 1.895597472122161e-05, + "loss": 0.3255, + "num_input_tokens_seen": 35165616, + "step": 23110 + }, + { + "epoch": 0.7398694065680814, + "grad_norm": 0.9635496139526367, + "learning_rate": 1.894644893781231e-05, + "loss": 0.6263, + "num_input_tokens_seen": 35173072, + "step": 23115 + }, + { + "epoch": 0.7400294475385699, + "grad_norm": 0.6447659730911255, + "learning_rate": 1.893692408793665e-05, + "loss": 0.4494, + "num_input_tokens_seen": 35180304, + "step": 23120 + }, + { + "epoch": 0.7401894885090583, + "grad_norm": 0.5809710025787354, + "learning_rate": 1.8927400173063493e-05, + "loss": 0.4401, + "num_input_tokens_seen": 35188064, + "step": 23125 + }, + { + "epoch": 0.7403495294795468, + "grad_norm": 0.7495164275169373, + "learning_rate": 1.891787719466154e-05, + "loss": 0.3697, + "num_input_tokens_seen": 35195472, + "step": 23130 + }, + { + "epoch": 0.7405095704500352, + "grad_norm": 2.589221715927124, + "learning_rate": 1.8908355154199346e-05, + "loss": 0.6596, + "num_input_tokens_seen": 35203072, + "step": 23135 + }, + { + "epoch": 0.7406696114205237, + "grad_norm": 0.5327308177947998, + "learning_rate": 1.8898834053145357e-05, + "loss": 0.5529, + "num_input_tokens_seen": 35210896, + "step": 23140 + }, + { + "epoch": 0.7408296523910121, + "grad_norm": 0.523716390132904, + "learning_rate": 1.8889313892967813e-05, + "loss": 0.4211, + "num_input_tokens_seen": 35218544, + "step": 23145 + }, + { + "epoch": 0.7409896933615006, + "grad_norm": 0.5895228981971741, + "learning_rate": 1.8879794675134863e-05, + "loss": 0.3275, + "num_input_tokens_seen": 35226240, + "step": 23150 + }, + { + "epoch": 0.7411497343319889, + "grad_norm": 1.0082416534423828, + "learning_rate": 1.8870276401114494e-05, + "loss": 0.5327, + "num_input_tokens_seen": 35233680, + "step": 23155 + }, + { + "epoch": 0.7413097753024774, + "grad_norm": 0.7210562229156494, + "learning_rate": 1.886075907237453e-05, + "loss": 0.2997, + "num_input_tokens_seen": 35240912, + "step": 23160 + }, + { + "epoch": 0.7414698162729659, + "grad_norm": 1.3574168682098389, + "learning_rate": 1.8851242690382672e-05, + "loss": 0.6042, + "num_input_tokens_seen": 35248336, + "step": 23165 + }, + { + "epoch": 0.7416298572434543, + "grad_norm": 0.9164567589759827, + "learning_rate": 1.884172725660645e-05, + "loss": 0.5965, + "num_input_tokens_seen": 35256240, + "step": 23170 + }, + { + "epoch": 0.7417898982139428, + "grad_norm": 0.39944395422935486, + "learning_rate": 1.8832212772513277e-05, + "loss": 0.338, + "num_input_tokens_seen": 35263696, + "step": 23175 + }, + { + "epoch": 0.7419499391844312, + "grad_norm": 1.111043095588684, + "learning_rate": 1.8822699239570414e-05, + "loss": 0.4138, + "num_input_tokens_seen": 35271504, + "step": 23180 + }, + { + "epoch": 0.7421099801549197, + "grad_norm": 0.8084381222724915, + "learning_rate": 1.8813186659244943e-05, + "loss": 0.4759, + "num_input_tokens_seen": 35279168, + "step": 23185 + }, + { + "epoch": 0.7422700211254081, + "grad_norm": 0.6055072546005249, + "learning_rate": 1.880367503300385e-05, + "loss": 0.4686, + "num_input_tokens_seen": 35286752, + "step": 23190 + }, + { + "epoch": 0.7424300620958966, + "grad_norm": 0.7199418544769287, + "learning_rate": 1.8794164362313927e-05, + "loss": 0.5587, + "num_input_tokens_seen": 35294720, + "step": 23195 + }, + { + "epoch": 0.742590103066385, + "grad_norm": 0.7869528532028198, + "learning_rate": 1.878465464864185e-05, + "loss": 0.5457, + "num_input_tokens_seen": 35302272, + "step": 23200 + }, + { + "epoch": 0.742590103066385, + "eval_loss": 0.5063515305519104, + "eval_runtime": 559.3819, + "eval_samples_per_second": 24.824, + "eval_steps_per_second": 12.412, + "num_input_tokens_seen": 35302272, + "step": 23200 + }, + { + "epoch": 0.7427501440368734, + "grad_norm": 0.6222087144851685, + "learning_rate": 1.877514589345414e-05, + "loss": 0.4818, + "num_input_tokens_seen": 35310176, + "step": 23205 + }, + { + "epoch": 0.7429101850073618, + "grad_norm": 10.212325096130371, + "learning_rate": 1.876563809821715e-05, + "loss": 0.5428, + "num_input_tokens_seen": 35317776, + "step": 23210 + }, + { + "epoch": 0.7430702259778503, + "grad_norm": 1.838914394378662, + "learning_rate": 1.8756131264397106e-05, + "loss": 0.4911, + "num_input_tokens_seen": 35325472, + "step": 23215 + }, + { + "epoch": 0.7432302669483388, + "grad_norm": 0.8189935684204102, + "learning_rate": 1.87466253934601e-05, + "loss": 0.5714, + "num_input_tokens_seen": 35333440, + "step": 23220 + }, + { + "epoch": 0.7433903079188272, + "grad_norm": 1.0189424753189087, + "learning_rate": 1.8737120486872033e-05, + "loss": 0.5167, + "num_input_tokens_seen": 35341296, + "step": 23225 + }, + { + "epoch": 0.7435503488893157, + "grad_norm": 0.7306773066520691, + "learning_rate": 1.8727616546098696e-05, + "loss": 0.5801, + "num_input_tokens_seen": 35348656, + "step": 23230 + }, + { + "epoch": 0.7437103898598041, + "grad_norm": 1.4543614387512207, + "learning_rate": 1.8718113572605716e-05, + "loss": 0.6629, + "num_input_tokens_seen": 35356320, + "step": 23235 + }, + { + "epoch": 0.7438704308302926, + "grad_norm": 0.8098775744438171, + "learning_rate": 1.8708611567858554e-05, + "loss": 0.3798, + "num_input_tokens_seen": 35363728, + "step": 23240 + }, + { + "epoch": 0.744030471800781, + "grad_norm": 1.072810411453247, + "learning_rate": 1.8699110533322565e-05, + "loss": 0.6034, + "num_input_tokens_seen": 35372144, + "step": 23245 + }, + { + "epoch": 0.7441905127712695, + "grad_norm": 0.4859139621257782, + "learning_rate": 1.8689610470462897e-05, + "loss": 0.6541, + "num_input_tokens_seen": 35379840, + "step": 23250 + }, + { + "epoch": 0.7443505537417578, + "grad_norm": 0.7499716877937317, + "learning_rate": 1.8680111380744604e-05, + "loss": 0.3046, + "num_input_tokens_seen": 35387296, + "step": 23255 + }, + { + "epoch": 0.7445105947122463, + "grad_norm": 0.5871492028236389, + "learning_rate": 1.8670613265632564e-05, + "loss": 0.5411, + "num_input_tokens_seen": 35394256, + "step": 23260 + }, + { + "epoch": 0.7446706356827347, + "grad_norm": 0.9380208253860474, + "learning_rate": 1.866111612659149e-05, + "loss": 0.458, + "num_input_tokens_seen": 35401712, + "step": 23265 + }, + { + "epoch": 0.7448306766532232, + "grad_norm": 0.7052749991416931, + "learning_rate": 1.8651619965085967e-05, + "loss": 0.4756, + "num_input_tokens_seen": 35409392, + "step": 23270 + }, + { + "epoch": 0.7449907176237117, + "grad_norm": 0.8742642998695374, + "learning_rate": 1.8642124782580433e-05, + "loss": 0.4395, + "num_input_tokens_seen": 35417408, + "step": 23275 + }, + { + "epoch": 0.7451507585942001, + "grad_norm": 0.5010367035865784, + "learning_rate": 1.8632630580539144e-05, + "loss": 0.3884, + "num_input_tokens_seen": 35425440, + "step": 23280 + }, + { + "epoch": 0.7453107995646886, + "grad_norm": 0.3630715012550354, + "learning_rate": 1.862313736042625e-05, + "loss": 0.4524, + "num_input_tokens_seen": 35433504, + "step": 23285 + }, + { + "epoch": 0.745470840535177, + "grad_norm": 0.8668965101242065, + "learning_rate": 1.8613645123705703e-05, + "loss": 0.6184, + "num_input_tokens_seen": 35440816, + "step": 23290 + }, + { + "epoch": 0.7456308815056655, + "grad_norm": 1.1174739599227905, + "learning_rate": 1.8604153871841328e-05, + "loss": 0.7928, + "num_input_tokens_seen": 35448496, + "step": 23295 + }, + { + "epoch": 0.7457909224761539, + "grad_norm": 1.1760879755020142, + "learning_rate": 1.859466360629682e-05, + "loss": 0.6884, + "num_input_tokens_seen": 35456304, + "step": 23300 + }, + { + "epoch": 0.7459509634466424, + "grad_norm": 1.515073537826538, + "learning_rate": 1.8585174328535666e-05, + "loss": 0.6275, + "num_input_tokens_seen": 35464256, + "step": 23305 + }, + { + "epoch": 0.7461110044171307, + "grad_norm": 1.3840444087982178, + "learning_rate": 1.857568604002124e-05, + "loss": 0.5614, + "num_input_tokens_seen": 35472224, + "step": 23310 + }, + { + "epoch": 0.7462710453876192, + "grad_norm": 0.5197117924690247, + "learning_rate": 1.8566198742216774e-05, + "loss": 0.4445, + "num_input_tokens_seen": 35479808, + "step": 23315 + }, + { + "epoch": 0.7464310863581076, + "grad_norm": 0.67009037733078, + "learning_rate": 1.85567124365853e-05, + "loss": 0.3461, + "num_input_tokens_seen": 35487536, + "step": 23320 + }, + { + "epoch": 0.7465911273285961, + "grad_norm": 0.6499744653701782, + "learning_rate": 1.854722712458975e-05, + "loss": 0.4605, + "num_input_tokens_seen": 35495200, + "step": 23325 + }, + { + "epoch": 0.7467511682990846, + "grad_norm": 0.695098876953125, + "learning_rate": 1.853774280769286e-05, + "loss": 0.5143, + "num_input_tokens_seen": 35502800, + "step": 23330 + }, + { + "epoch": 0.746911209269573, + "grad_norm": 0.5179017782211304, + "learning_rate": 1.852825948735724e-05, + "loss": 0.4166, + "num_input_tokens_seen": 35511264, + "step": 23335 + }, + { + "epoch": 0.7470712502400615, + "grad_norm": 1.2466646432876587, + "learning_rate": 1.851877716504534e-05, + "loss": 0.6018, + "num_input_tokens_seen": 35519120, + "step": 23340 + }, + { + "epoch": 0.7472312912105499, + "grad_norm": 0.6421826481819153, + "learning_rate": 1.8509295842219448e-05, + "loss": 0.5075, + "num_input_tokens_seen": 35526176, + "step": 23345 + }, + { + "epoch": 0.7473913321810384, + "grad_norm": 0.6112781167030334, + "learning_rate": 1.8499815520341697e-05, + "loss": 0.3664, + "num_input_tokens_seen": 35534176, + "step": 23350 + }, + { + "epoch": 0.7475513731515268, + "grad_norm": 0.41167011857032776, + "learning_rate": 1.8490336200874094e-05, + "loss": 0.3882, + "num_input_tokens_seen": 35541456, + "step": 23355 + }, + { + "epoch": 0.7477114141220152, + "grad_norm": 0.523912250995636, + "learning_rate": 1.848085788527844e-05, + "loss": 0.5203, + "num_input_tokens_seen": 35549072, + "step": 23360 + }, + { + "epoch": 0.7478714550925036, + "grad_norm": 0.8054673671722412, + "learning_rate": 1.847138057501644e-05, + "loss": 0.5158, + "num_input_tokens_seen": 35556432, + "step": 23365 + }, + { + "epoch": 0.7480314960629921, + "grad_norm": 0.6679694652557373, + "learning_rate": 1.8461904271549582e-05, + "loss": 0.485, + "num_input_tokens_seen": 35564576, + "step": 23370 + }, + { + "epoch": 0.7481915370334806, + "grad_norm": 1.17446768283844, + "learning_rate": 1.845242897633926e-05, + "loss": 0.4555, + "num_input_tokens_seen": 35572256, + "step": 23375 + }, + { + "epoch": 0.748351578003969, + "grad_norm": 2.6675949096679688, + "learning_rate": 1.844295469084667e-05, + "loss": 0.5501, + "num_input_tokens_seen": 35579968, + "step": 23380 + }, + { + "epoch": 0.7485116189744575, + "grad_norm": 0.8290353417396545, + "learning_rate": 1.843348141653286e-05, + "loss": 0.4658, + "num_input_tokens_seen": 35587552, + "step": 23385 + }, + { + "epoch": 0.7486716599449459, + "grad_norm": 1.0369168519973755, + "learning_rate": 1.842400915485874e-05, + "loss": 0.4618, + "num_input_tokens_seen": 35595264, + "step": 23390 + }, + { + "epoch": 0.7488317009154344, + "grad_norm": 0.3131342828273773, + "learning_rate": 1.8414537907285053e-05, + "loss": 0.4379, + "num_input_tokens_seen": 35603280, + "step": 23395 + }, + { + "epoch": 0.7489917418859228, + "grad_norm": 0.8890591263771057, + "learning_rate": 1.840506767527237e-05, + "loss": 0.5577, + "num_input_tokens_seen": 35611216, + "step": 23400 + }, + { + "epoch": 0.7489917418859228, + "eval_loss": 0.5055484175682068, + "eval_runtime": 559.3894, + "eval_samples_per_second": 24.823, + "eval_steps_per_second": 12.412, + "num_input_tokens_seen": 35611216, + "step": 23400 + }, + { + "epoch": 0.7491517828564113, + "grad_norm": 1.0276323556900024, + "learning_rate": 1.8395598460281137e-05, + "loss": 0.7007, + "num_input_tokens_seen": 35618816, + "step": 23405 + }, + { + "epoch": 0.7493118238268996, + "grad_norm": 0.5391591787338257, + "learning_rate": 1.838613026377161e-05, + "loss": 0.3644, + "num_input_tokens_seen": 35626512, + "step": 23410 + }, + { + "epoch": 0.7494718647973881, + "grad_norm": 0.956787109375, + "learning_rate": 1.8376663087203917e-05, + "loss": 0.5303, + "num_input_tokens_seen": 35633712, + "step": 23415 + }, + { + "epoch": 0.7496319057678765, + "grad_norm": 1.0348973274230957, + "learning_rate": 1.8367196932038014e-05, + "loss": 0.5999, + "num_input_tokens_seen": 35641680, + "step": 23420 + }, + { + "epoch": 0.749791946738365, + "grad_norm": 0.730469822883606, + "learning_rate": 1.8357731799733686e-05, + "loss": 0.5029, + "num_input_tokens_seen": 35649488, + "step": 23425 + }, + { + "epoch": 0.7499519877088535, + "grad_norm": 0.8088845610618591, + "learning_rate": 1.8348267691750586e-05, + "loss": 0.5794, + "num_input_tokens_seen": 35656688, + "step": 23430 + }, + { + "epoch": 0.7501120286793419, + "grad_norm": 0.6831542253494263, + "learning_rate": 1.833880460954821e-05, + "loss": 0.5586, + "num_input_tokens_seen": 35664176, + "step": 23435 + }, + { + "epoch": 0.7502720696498304, + "grad_norm": 0.7882079482078552, + "learning_rate": 1.8329342554585866e-05, + "loss": 0.458, + "num_input_tokens_seen": 35671856, + "step": 23440 + }, + { + "epoch": 0.7504321106203188, + "grad_norm": 0.510707676410675, + "learning_rate": 1.8319881528322735e-05, + "loss": 0.5026, + "num_input_tokens_seen": 35679472, + "step": 23445 + }, + { + "epoch": 0.7505921515908073, + "grad_norm": 0.5669872164726257, + "learning_rate": 1.8310421532217815e-05, + "loss": 0.4325, + "num_input_tokens_seen": 35687056, + "step": 23450 + }, + { + "epoch": 0.7507521925612957, + "grad_norm": 0.5733433365821838, + "learning_rate": 1.8300962567729958e-05, + "loss": 0.5078, + "num_input_tokens_seen": 35695024, + "step": 23455 + }, + { + "epoch": 0.7509122335317842, + "grad_norm": 1.251330852508545, + "learning_rate": 1.8291504636317866e-05, + "loss": 0.6065, + "num_input_tokens_seen": 35702128, + "step": 23460 + }, + { + "epoch": 0.7510722745022725, + "grad_norm": 1.1638785600662231, + "learning_rate": 1.8282047739440055e-05, + "loss": 0.4821, + "num_input_tokens_seen": 35710176, + "step": 23465 + }, + { + "epoch": 0.751232315472761, + "grad_norm": 0.7223319411277771, + "learning_rate": 1.8272591878554903e-05, + "loss": 0.392, + "num_input_tokens_seen": 35717936, + "step": 23470 + }, + { + "epoch": 0.7513923564432494, + "grad_norm": 1.4137544631958008, + "learning_rate": 1.8263137055120638e-05, + "loss": 0.4968, + "num_input_tokens_seen": 35725776, + "step": 23475 + }, + { + "epoch": 0.7515523974137379, + "grad_norm": 0.6384099721908569, + "learning_rate": 1.8253683270595295e-05, + "loss": 0.5506, + "num_input_tokens_seen": 35733456, + "step": 23480 + }, + { + "epoch": 0.7517124383842264, + "grad_norm": 0.6321364045143127, + "learning_rate": 1.824423052643677e-05, + "loss": 0.5804, + "num_input_tokens_seen": 35740832, + "step": 23485 + }, + { + "epoch": 0.7518724793547148, + "grad_norm": 0.680597186088562, + "learning_rate": 1.82347788241028e-05, + "loss": 0.5735, + "num_input_tokens_seen": 35748592, + "step": 23490 + }, + { + "epoch": 0.7520325203252033, + "grad_norm": 0.5505670309066772, + "learning_rate": 1.8225328165050942e-05, + "loss": 0.4872, + "num_input_tokens_seen": 35756656, + "step": 23495 + }, + { + "epoch": 0.7521925612956917, + "grad_norm": 0.6812238693237305, + "learning_rate": 1.821587855073863e-05, + "loss": 0.4268, + "num_input_tokens_seen": 35765168, + "step": 23500 + }, + { + "epoch": 0.7523526022661802, + "grad_norm": 0.6410269737243652, + "learning_rate": 1.8206429982623086e-05, + "loss": 0.47, + "num_input_tokens_seen": 35772464, + "step": 23505 + }, + { + "epoch": 0.7525126432366686, + "grad_norm": 0.6651862263679504, + "learning_rate": 1.8196982462161416e-05, + "loss": 0.4832, + "num_input_tokens_seen": 35779744, + "step": 23510 + }, + { + "epoch": 0.752672684207157, + "grad_norm": 1.0552464723587036, + "learning_rate": 1.818753599081055e-05, + "loss": 0.52, + "num_input_tokens_seen": 35787360, + "step": 23515 + }, + { + "epoch": 0.7528327251776454, + "grad_norm": 0.39766398072242737, + "learning_rate": 1.817809057002724e-05, + "loss": 0.5443, + "num_input_tokens_seen": 35794496, + "step": 23520 + }, + { + "epoch": 0.7529927661481339, + "grad_norm": 1.0537245273590088, + "learning_rate": 1.8168646201268096e-05, + "loss": 0.4952, + "num_input_tokens_seen": 35802320, + "step": 23525 + }, + { + "epoch": 0.7531528071186223, + "grad_norm": 1.070565104484558, + "learning_rate": 1.8159202885989557e-05, + "loss": 0.4373, + "num_input_tokens_seen": 35809968, + "step": 23530 + }, + { + "epoch": 0.7533128480891108, + "grad_norm": 0.7242513298988342, + "learning_rate": 1.814976062564789e-05, + "loss": 0.5153, + "num_input_tokens_seen": 35818336, + "step": 23535 + }, + { + "epoch": 0.7534728890595993, + "grad_norm": 0.9006926417350769, + "learning_rate": 1.8140319421699234e-05, + "loss": 0.3484, + "num_input_tokens_seen": 35826208, + "step": 23540 + }, + { + "epoch": 0.7536329300300877, + "grad_norm": 0.5251461863517761, + "learning_rate": 1.8130879275599515e-05, + "loss": 0.358, + "num_input_tokens_seen": 35833920, + "step": 23545 + }, + { + "epoch": 0.7537929710005762, + "grad_norm": 1.5441032648086548, + "learning_rate": 1.8121440188804544e-05, + "loss": 0.5909, + "num_input_tokens_seen": 35841424, + "step": 23550 + }, + { + "epoch": 0.7539530119710646, + "grad_norm": 0.889721691608429, + "learning_rate": 1.811200216276993e-05, + "loss": 0.5664, + "num_input_tokens_seen": 35849424, + "step": 23555 + }, + { + "epoch": 0.7541130529415531, + "grad_norm": 0.5564526915550232, + "learning_rate": 1.810256519895115e-05, + "loss": 0.4165, + "num_input_tokens_seen": 35857104, + "step": 23560 + }, + { + "epoch": 0.7542730939120414, + "grad_norm": 0.539888322353363, + "learning_rate": 1.8093129298803494e-05, + "loss": 0.3851, + "num_input_tokens_seen": 35865024, + "step": 23565 + }, + { + "epoch": 0.7544331348825299, + "grad_norm": 0.8475088477134705, + "learning_rate": 1.808369446378209e-05, + "loss": 0.4979, + "num_input_tokens_seen": 35872736, + "step": 23570 + }, + { + "epoch": 0.7545931758530183, + "grad_norm": 0.8098413348197937, + "learning_rate": 1.8074260695341914e-05, + "loss": 0.3877, + "num_input_tokens_seen": 35880320, + "step": 23575 + }, + { + "epoch": 0.7547532168235068, + "grad_norm": 1.9364275932312012, + "learning_rate": 1.8064827994937782e-05, + "loss": 0.657, + "num_input_tokens_seen": 35887920, + "step": 23580 + }, + { + "epoch": 0.7549132577939953, + "grad_norm": 1.3611303567886353, + "learning_rate": 1.8055396364024317e-05, + "loss": 0.6405, + "num_input_tokens_seen": 35895424, + "step": 23585 + }, + { + "epoch": 0.7550732987644837, + "grad_norm": 0.4815181791782379, + "learning_rate": 1.804596580405601e-05, + "loss": 0.4325, + "num_input_tokens_seen": 35903440, + "step": 23590 + }, + { + "epoch": 0.7552333397349722, + "grad_norm": 1.2345718145370483, + "learning_rate": 1.8036536316487174e-05, + "loss": 0.5187, + "num_input_tokens_seen": 35910640, + "step": 23595 + }, + { + "epoch": 0.7553933807054606, + "grad_norm": 0.6549308896064758, + "learning_rate": 1.802710790277193e-05, + "loss": 0.6506, + "num_input_tokens_seen": 35918512, + "step": 23600 + }, + { + "epoch": 0.7553933807054606, + "eval_loss": 0.5050768852233887, + "eval_runtime": 559.3273, + "eval_samples_per_second": 24.826, + "eval_steps_per_second": 12.413, + "num_input_tokens_seen": 35918512, + "step": 23600 + }, + { + "epoch": 0.7555534216759491, + "grad_norm": 0.7605641484260559, + "learning_rate": 1.801768056436429e-05, + "loss": 0.4746, + "num_input_tokens_seen": 35925856, + "step": 23605 + }, + { + "epoch": 0.7557134626464375, + "grad_norm": 2.2607686519622803, + "learning_rate": 1.8008254302718035e-05, + "loss": 0.6902, + "num_input_tokens_seen": 35933728, + "step": 23610 + }, + { + "epoch": 0.755873503616926, + "grad_norm": 0.9816949963569641, + "learning_rate": 1.7998829119286837e-05, + "loss": 0.5387, + "num_input_tokens_seen": 35941488, + "step": 23615 + }, + { + "epoch": 0.7560335445874143, + "grad_norm": 0.9183422923088074, + "learning_rate": 1.798940501552418e-05, + "loss": 0.4735, + "num_input_tokens_seen": 35948896, + "step": 23620 + }, + { + "epoch": 0.7561935855579028, + "grad_norm": 1.45661461353302, + "learning_rate": 1.797998199288336e-05, + "loss": 0.3622, + "num_input_tokens_seen": 35955840, + "step": 23625 + }, + { + "epoch": 0.7563536265283912, + "grad_norm": 0.6084232330322266, + "learning_rate": 1.7970560052817543e-05, + "loss": 0.4192, + "num_input_tokens_seen": 35964176, + "step": 23630 + }, + { + "epoch": 0.7565136674988797, + "grad_norm": 0.6124838590621948, + "learning_rate": 1.7961139196779702e-05, + "loss": 0.3803, + "num_input_tokens_seen": 35971568, + "step": 23635 + }, + { + "epoch": 0.7566737084693682, + "grad_norm": 0.5496603846549988, + "learning_rate": 1.7951719426222647e-05, + "loss": 0.5447, + "num_input_tokens_seen": 35979072, + "step": 23640 + }, + { + "epoch": 0.7568337494398566, + "grad_norm": 1.047278881072998, + "learning_rate": 1.794230074259904e-05, + "loss": 0.6163, + "num_input_tokens_seen": 35986416, + "step": 23645 + }, + { + "epoch": 0.7569937904103451, + "grad_norm": 0.8120248913764954, + "learning_rate": 1.7932883147361336e-05, + "loss": 0.5598, + "num_input_tokens_seen": 35993984, + "step": 23650 + }, + { + "epoch": 0.7571538313808335, + "grad_norm": 0.7152830362319946, + "learning_rate": 1.7923466641961865e-05, + "loss": 0.4306, + "num_input_tokens_seen": 36001568, + "step": 23655 + }, + { + "epoch": 0.757313872351322, + "grad_norm": 0.5690480470657349, + "learning_rate": 1.791405122785278e-05, + "loss": 0.4131, + "num_input_tokens_seen": 36009424, + "step": 23660 + }, + { + "epoch": 0.7574739133218104, + "grad_norm": 0.7692463397979736, + "learning_rate": 1.7904636906486037e-05, + "loss": 0.6601, + "num_input_tokens_seen": 36016896, + "step": 23665 + }, + { + "epoch": 0.7576339542922989, + "grad_norm": 0.9861642122268677, + "learning_rate": 1.7895223679313448e-05, + "loss": 0.4263, + "num_input_tokens_seen": 36024384, + "step": 23670 + }, + { + "epoch": 0.7577939952627872, + "grad_norm": 1.0107579231262207, + "learning_rate": 1.7885811547786653e-05, + "loss": 0.5164, + "num_input_tokens_seen": 36031936, + "step": 23675 + }, + { + "epoch": 0.7579540362332757, + "grad_norm": 1.4012961387634277, + "learning_rate": 1.7876400513357115e-05, + "loss": 0.3973, + "num_input_tokens_seen": 36039808, + "step": 23680 + }, + { + "epoch": 0.7581140772037641, + "grad_norm": 0.6481572389602661, + "learning_rate": 1.7866990577476146e-05, + "loss": 0.3179, + "num_input_tokens_seen": 36047184, + "step": 23685 + }, + { + "epoch": 0.7582741181742526, + "grad_norm": 1.1174445152282715, + "learning_rate": 1.7857581741594863e-05, + "loss": 0.4655, + "num_input_tokens_seen": 36054608, + "step": 23690 + }, + { + "epoch": 0.7584341591447411, + "grad_norm": 0.7685070633888245, + "learning_rate": 1.7848174007164237e-05, + "loss": 0.3687, + "num_input_tokens_seen": 36062256, + "step": 23695 + }, + { + "epoch": 0.7585942001152295, + "grad_norm": 1.2179806232452393, + "learning_rate": 1.7838767375635052e-05, + "loss": 0.5587, + "num_input_tokens_seen": 36069824, + "step": 23700 + }, + { + "epoch": 0.758754241085718, + "grad_norm": 0.4295915365219116, + "learning_rate": 1.782936184845793e-05, + "loss": 0.5251, + "num_input_tokens_seen": 36077216, + "step": 23705 + }, + { + "epoch": 0.7589142820562064, + "grad_norm": 0.567284882068634, + "learning_rate": 1.7819957427083334e-05, + "loss": 0.5578, + "num_input_tokens_seen": 36084768, + "step": 23710 + }, + { + "epoch": 0.7590743230266949, + "grad_norm": 0.8057486414909363, + "learning_rate": 1.7810554112961516e-05, + "loss": 0.4901, + "num_input_tokens_seen": 36092416, + "step": 23715 + }, + { + "epoch": 0.7592343639971832, + "grad_norm": 0.5291479229927063, + "learning_rate": 1.7801151907542607e-05, + "loss": 0.4315, + "num_input_tokens_seen": 36099680, + "step": 23720 + }, + { + "epoch": 0.7593944049676717, + "grad_norm": 1.5091173648834229, + "learning_rate": 1.7791750812276547e-05, + "loss": 0.6369, + "num_input_tokens_seen": 36107360, + "step": 23725 + }, + { + "epoch": 0.7595544459381601, + "grad_norm": 0.6470710039138794, + "learning_rate": 1.778235082861309e-05, + "loss": 0.4235, + "num_input_tokens_seen": 36115424, + "step": 23730 + }, + { + "epoch": 0.7597144869086486, + "grad_norm": 0.5234347581863403, + "learning_rate": 1.777295195800184e-05, + "loss": 0.4583, + "num_input_tokens_seen": 36122544, + "step": 23735 + }, + { + "epoch": 0.759874527879137, + "grad_norm": 1.6579328775405884, + "learning_rate": 1.7763554201892215e-05, + "loss": 0.605, + "num_input_tokens_seen": 36130400, + "step": 23740 + }, + { + "epoch": 0.7600345688496255, + "grad_norm": 1.2465749979019165, + "learning_rate": 1.7754157561733476e-05, + "loss": 0.4064, + "num_input_tokens_seen": 36138272, + "step": 23745 + }, + { + "epoch": 0.760194609820114, + "grad_norm": 0.6974813938140869, + "learning_rate": 1.7744762038974702e-05, + "loss": 0.4165, + "num_input_tokens_seen": 36146128, + "step": 23750 + }, + { + "epoch": 0.7603546507906024, + "grad_norm": 0.8622053265571594, + "learning_rate": 1.7735367635064788e-05, + "loss": 0.4004, + "num_input_tokens_seen": 36153968, + "step": 23755 + }, + { + "epoch": 0.7605146917610909, + "grad_norm": 1.0864930152893066, + "learning_rate": 1.7725974351452474e-05, + "loss": 0.5226, + "num_input_tokens_seen": 36161488, + "step": 23760 + }, + { + "epoch": 0.7606747327315793, + "grad_norm": 1.0548430681228638, + "learning_rate": 1.771658218958634e-05, + "loss": 0.6664, + "num_input_tokens_seen": 36169296, + "step": 23765 + }, + { + "epoch": 0.7608347737020678, + "grad_norm": 0.5656983852386475, + "learning_rate": 1.770719115091475e-05, + "loss": 0.5052, + "num_input_tokens_seen": 36177232, + "step": 23770 + }, + { + "epoch": 0.7609948146725561, + "grad_norm": 0.8505110144615173, + "learning_rate": 1.7697801236885935e-05, + "loss": 0.5302, + "num_input_tokens_seen": 36185104, + "step": 23775 + }, + { + "epoch": 0.7611548556430446, + "grad_norm": 0.4323257803916931, + "learning_rate": 1.7688412448947944e-05, + "loss": 0.4178, + "num_input_tokens_seen": 36192640, + "step": 23780 + }, + { + "epoch": 0.761314896613533, + "grad_norm": 0.9091343283653259, + "learning_rate": 1.767902478854862e-05, + "loss": 0.5706, + "num_input_tokens_seen": 36201120, + "step": 23785 + }, + { + "epoch": 0.7614749375840215, + "grad_norm": 0.7816592454910278, + "learning_rate": 1.766963825713569e-05, + "loss": 0.4119, + "num_input_tokens_seen": 36208496, + "step": 23790 + }, + { + "epoch": 0.7616349785545099, + "grad_norm": 0.5525997281074524, + "learning_rate": 1.766025285615665e-05, + "loss": 0.5872, + "num_input_tokens_seen": 36216256, + "step": 23795 + }, + { + "epoch": 0.7617950195249984, + "grad_norm": 1.0787385702133179, + "learning_rate": 1.7650868587058854e-05, + "loss": 0.4636, + "num_input_tokens_seen": 36223840, + "step": 23800 + }, + { + "epoch": 0.7617950195249984, + "eval_loss": 0.5045191049575806, + "eval_runtime": 559.1804, + "eval_samples_per_second": 24.833, + "eval_steps_per_second": 12.416, + "num_input_tokens_seen": 36223840, + "step": 23800 + }, + { + "epoch": 0.7619550604954869, + "grad_norm": 0.48503345251083374, + "learning_rate": 1.7641485451289484e-05, + "loss": 0.5804, + "num_input_tokens_seen": 36231360, + "step": 23805 + }, + { + "epoch": 0.7621151014659753, + "grad_norm": 0.7112874984741211, + "learning_rate": 1.7632103450295534e-05, + "loss": 0.4986, + "num_input_tokens_seen": 36239472, + "step": 23810 + }, + { + "epoch": 0.7622751424364638, + "grad_norm": 0.633533775806427, + "learning_rate": 1.762272258552381e-05, + "loss": 0.509, + "num_input_tokens_seen": 36247280, + "step": 23815 + }, + { + "epoch": 0.7624351834069522, + "grad_norm": 0.8456358909606934, + "learning_rate": 1.7613342858420988e-05, + "loss": 0.5763, + "num_input_tokens_seen": 36254704, + "step": 23820 + }, + { + "epoch": 0.7625952243774407, + "grad_norm": 0.4244774580001831, + "learning_rate": 1.760396427043351e-05, + "loss": 0.4222, + "num_input_tokens_seen": 36262160, + "step": 23825 + }, + { + "epoch": 0.762755265347929, + "grad_norm": 0.6800742745399475, + "learning_rate": 1.7594586823007696e-05, + "loss": 0.5387, + "num_input_tokens_seen": 36269280, + "step": 23830 + }, + { + "epoch": 0.7629153063184175, + "grad_norm": 0.6527806520462036, + "learning_rate": 1.7585210517589646e-05, + "loss": 0.5171, + "num_input_tokens_seen": 36277360, + "step": 23835 + }, + { + "epoch": 0.7630753472889059, + "grad_norm": 0.31675389409065247, + "learning_rate": 1.7575835355625314e-05, + "loss": 0.3671, + "num_input_tokens_seen": 36285520, + "step": 23840 + }, + { + "epoch": 0.7632353882593944, + "grad_norm": 0.9230730533599854, + "learning_rate": 1.756646133856048e-05, + "loss": 0.4755, + "num_input_tokens_seen": 36293376, + "step": 23845 + }, + { + "epoch": 0.7633954292298829, + "grad_norm": 0.2455289661884308, + "learning_rate": 1.7557088467840714e-05, + "loss": 0.325, + "num_input_tokens_seen": 36301488, + "step": 23850 + }, + { + "epoch": 0.7635554702003713, + "grad_norm": 0.8814258575439453, + "learning_rate": 1.7547716744911438e-05, + "loss": 0.3583, + "num_input_tokens_seen": 36309136, + "step": 23855 + }, + { + "epoch": 0.7637155111708598, + "grad_norm": 0.7634087204933167, + "learning_rate": 1.7538346171217902e-05, + "loss": 0.4677, + "num_input_tokens_seen": 36316880, + "step": 23860 + }, + { + "epoch": 0.7638755521413482, + "grad_norm": 1.1457477807998657, + "learning_rate": 1.7528976748205146e-05, + "loss": 0.4307, + "num_input_tokens_seen": 36324608, + "step": 23865 + }, + { + "epoch": 0.7640355931118367, + "grad_norm": 0.7305124998092651, + "learning_rate": 1.751960847731807e-05, + "loss": 0.4663, + "num_input_tokens_seen": 36331920, + "step": 23870 + }, + { + "epoch": 0.764195634082325, + "grad_norm": 0.4917221963405609, + "learning_rate": 1.7510241360001362e-05, + "loss": 0.3937, + "num_input_tokens_seen": 36339184, + "step": 23875 + }, + { + "epoch": 0.7643556750528135, + "grad_norm": 1.1315544843673706, + "learning_rate": 1.7500875397699562e-05, + "loss": 0.4031, + "num_input_tokens_seen": 36346928, + "step": 23880 + }, + { + "epoch": 0.7645157160233019, + "grad_norm": 0.7518841624259949, + "learning_rate": 1.7491510591857015e-05, + "loss": 0.5461, + "num_input_tokens_seen": 36354816, + "step": 23885 + }, + { + "epoch": 0.7646757569937904, + "grad_norm": 1.1769381761550903, + "learning_rate": 1.7482146943917896e-05, + "loss": 0.451, + "num_input_tokens_seen": 36363104, + "step": 23890 + }, + { + "epoch": 0.7648357979642788, + "grad_norm": 0.8256083726882935, + "learning_rate": 1.7472784455326185e-05, + "loss": 0.5144, + "num_input_tokens_seen": 36370656, + "step": 23895 + }, + { + "epoch": 0.7649958389347673, + "grad_norm": 1.1645013093948364, + "learning_rate": 1.746342312752572e-05, + "loss": 0.4977, + "num_input_tokens_seen": 36378480, + "step": 23900 + }, + { + "epoch": 0.7651558799052558, + "grad_norm": 0.6135721802711487, + "learning_rate": 1.74540629619601e-05, + "loss": 0.464, + "num_input_tokens_seen": 36386192, + "step": 23905 + }, + { + "epoch": 0.7653159208757442, + "grad_norm": 0.4553830325603485, + "learning_rate": 1.7444703960072815e-05, + "loss": 0.4206, + "num_input_tokens_seen": 36393440, + "step": 23910 + }, + { + "epoch": 0.7654759618462327, + "grad_norm": 2.394111394882202, + "learning_rate": 1.7435346123307118e-05, + "loss": 0.8989, + "num_input_tokens_seen": 36400912, + "step": 23915 + }, + { + "epoch": 0.7656360028167211, + "grad_norm": 0.537976861000061, + "learning_rate": 1.742598945310611e-05, + "loss": 0.4884, + "num_input_tokens_seen": 36408432, + "step": 23920 + }, + { + "epoch": 0.7657960437872096, + "grad_norm": 1.0358763933181763, + "learning_rate": 1.741663395091272e-05, + "loss": 0.4761, + "num_input_tokens_seen": 36415536, + "step": 23925 + }, + { + "epoch": 0.7659560847576979, + "grad_norm": 0.9858996272087097, + "learning_rate": 1.7407279618169657e-05, + "loss": 0.5188, + "num_input_tokens_seen": 36422912, + "step": 23930 + }, + { + "epoch": 0.7661161257281864, + "grad_norm": 1.2762885093688965, + "learning_rate": 1.73979264563195e-05, + "loss": 0.5251, + "num_input_tokens_seen": 36430080, + "step": 23935 + }, + { + "epoch": 0.7662761666986748, + "grad_norm": 1.3185791969299316, + "learning_rate": 1.7388574466804625e-05, + "loss": 0.6173, + "num_input_tokens_seen": 36437712, + "step": 23940 + }, + { + "epoch": 0.7664362076691633, + "grad_norm": 0.7425370216369629, + "learning_rate": 1.7379223651067207e-05, + "loss": 0.3126, + "num_input_tokens_seen": 36445200, + "step": 23945 + }, + { + "epoch": 0.7665962486396517, + "grad_norm": 1.1479415893554688, + "learning_rate": 1.736987401054928e-05, + "loss": 0.5257, + "num_input_tokens_seen": 36452864, + "step": 23950 + }, + { + "epoch": 0.7667562896101402, + "grad_norm": 0.9599429965019226, + "learning_rate": 1.736052554669266e-05, + "loss": 0.5217, + "num_input_tokens_seen": 36460208, + "step": 23955 + }, + { + "epoch": 0.7669163305806287, + "grad_norm": 0.5185725688934326, + "learning_rate": 1.7351178260939007e-05, + "loss": 0.5342, + "num_input_tokens_seen": 36467664, + "step": 23960 + }, + { + "epoch": 0.7670763715511171, + "grad_norm": 1.00529944896698, + "learning_rate": 1.7341832154729794e-05, + "loss": 0.4483, + "num_input_tokens_seen": 36475408, + "step": 23965 + }, + { + "epoch": 0.7672364125216056, + "grad_norm": 1.1461365222930908, + "learning_rate": 1.7332487229506286e-05, + "loss": 0.5268, + "num_input_tokens_seen": 36483344, + "step": 23970 + }, + { + "epoch": 0.767396453492094, + "grad_norm": 0.9552193284034729, + "learning_rate": 1.732314348670961e-05, + "loss": 0.4447, + "num_input_tokens_seen": 36491120, + "step": 23975 + }, + { + "epoch": 0.7675564944625825, + "grad_norm": 0.6899494528770447, + "learning_rate": 1.7313800927780686e-05, + "loss": 0.4534, + "num_input_tokens_seen": 36498384, + "step": 23980 + }, + { + "epoch": 0.7677165354330708, + "grad_norm": 0.9876281023025513, + "learning_rate": 1.7304459554160245e-05, + "loss": 0.6089, + "num_input_tokens_seen": 36506208, + "step": 23985 + }, + { + "epoch": 0.7678765764035593, + "grad_norm": 0.9276464581489563, + "learning_rate": 1.7295119367288853e-05, + "loss": 0.4605, + "num_input_tokens_seen": 36513664, + "step": 23990 + }, + { + "epoch": 0.7680366173740477, + "grad_norm": 1.1826783418655396, + "learning_rate": 1.728578036860688e-05, + "loss": 0.4963, + "num_input_tokens_seen": 36521088, + "step": 23995 + }, + { + "epoch": 0.7681966583445362, + "grad_norm": 0.9505362510681152, + "learning_rate": 1.7276442559554513e-05, + "loss": 0.5077, + "num_input_tokens_seen": 36528496, + "step": 24000 + }, + { + "epoch": 0.7681966583445362, + "eval_loss": 0.5044520497322083, + "eval_runtime": 559.0932, + "eval_samples_per_second": 24.837, + "eval_steps_per_second": 12.418, + "num_input_tokens_seen": 36528496, + "step": 24000 + }, + { + "epoch": 0.7683566993150246, + "grad_norm": 1.2564585208892822, + "learning_rate": 1.726710594157177e-05, + "loss": 0.3753, + "num_input_tokens_seen": 36535840, + "step": 24005 + }, + { + "epoch": 0.7685167402855131, + "grad_norm": 0.5303597450256348, + "learning_rate": 1.725777051609846e-05, + "loss": 0.3796, + "num_input_tokens_seen": 36543328, + "step": 24010 + }, + { + "epoch": 0.7686767812560016, + "grad_norm": 1.0016661882400513, + "learning_rate": 1.7248436284574228e-05, + "loss": 0.489, + "num_input_tokens_seen": 36550800, + "step": 24015 + }, + { + "epoch": 0.76883682222649, + "grad_norm": 0.7073590755462646, + "learning_rate": 1.723910324843855e-05, + "loss": 0.4092, + "num_input_tokens_seen": 36558592, + "step": 24020 + }, + { + "epoch": 0.7689968631969785, + "grad_norm": 0.6817220449447632, + "learning_rate": 1.722977140913067e-05, + "loss": 0.3806, + "num_input_tokens_seen": 36566560, + "step": 24025 + }, + { + "epoch": 0.7691569041674668, + "grad_norm": 0.6730770468711853, + "learning_rate": 1.7220440768089688e-05, + "loss": 0.4887, + "num_input_tokens_seen": 36573920, + "step": 24030 + }, + { + "epoch": 0.7693169451379553, + "grad_norm": 0.7354686260223389, + "learning_rate": 1.7211111326754505e-05, + "loss": 0.4646, + "num_input_tokens_seen": 36581680, + "step": 24035 + }, + { + "epoch": 0.7694769861084437, + "grad_norm": 0.348868191242218, + "learning_rate": 1.720178308656383e-05, + "loss": 0.5427, + "num_input_tokens_seen": 36588832, + "step": 24040 + }, + { + "epoch": 0.7696370270789322, + "grad_norm": 0.5836193561553955, + "learning_rate": 1.719245604895621e-05, + "loss": 0.3877, + "num_input_tokens_seen": 36596336, + "step": 24045 + }, + { + "epoch": 0.7697970680494206, + "grad_norm": 1.119518756866455, + "learning_rate": 1.7183130215369972e-05, + "loss": 0.4838, + "num_input_tokens_seen": 36603712, + "step": 24050 + }, + { + "epoch": 0.7699571090199091, + "grad_norm": 0.9721588492393494, + "learning_rate": 1.7173805587243292e-05, + "loss": 0.5108, + "num_input_tokens_seen": 36611200, + "step": 24055 + }, + { + "epoch": 0.7701171499903975, + "grad_norm": 0.6854874491691589, + "learning_rate": 1.7164482166014147e-05, + "loss": 0.4786, + "num_input_tokens_seen": 36618976, + "step": 24060 + }, + { + "epoch": 0.770277190960886, + "grad_norm": 0.5847227573394775, + "learning_rate": 1.7155159953120313e-05, + "loss": 0.534, + "num_input_tokens_seen": 36626400, + "step": 24065 + }, + { + "epoch": 0.7704372319313745, + "grad_norm": 0.8883029818534851, + "learning_rate": 1.714583894999941e-05, + "loss": 0.5011, + "num_input_tokens_seen": 36633712, + "step": 24070 + }, + { + "epoch": 0.7705972729018629, + "grad_norm": 1.5436885356903076, + "learning_rate": 1.7136519158088826e-05, + "loss": 0.5416, + "num_input_tokens_seen": 36641088, + "step": 24075 + }, + { + "epoch": 0.7707573138723514, + "grad_norm": 0.7574918270111084, + "learning_rate": 1.712720057882581e-05, + "loss": 0.424, + "num_input_tokens_seen": 36648512, + "step": 24080 + }, + { + "epoch": 0.7709173548428397, + "grad_norm": 0.9911123514175415, + "learning_rate": 1.7117883213647413e-05, + "loss": 0.4543, + "num_input_tokens_seen": 36655888, + "step": 24085 + }, + { + "epoch": 0.7710773958133282, + "grad_norm": 0.3812101185321808, + "learning_rate": 1.710856706399046e-05, + "loss": 0.4401, + "num_input_tokens_seen": 36663712, + "step": 24090 + }, + { + "epoch": 0.7712374367838166, + "grad_norm": 1.466168761253357, + "learning_rate": 1.7099252131291648e-05, + "loss": 0.6021, + "num_input_tokens_seen": 36671440, + "step": 24095 + }, + { + "epoch": 0.7713974777543051, + "grad_norm": 0.8415486812591553, + "learning_rate": 1.708993841698744e-05, + "loss": 0.5323, + "num_input_tokens_seen": 36678608, + "step": 24100 + }, + { + "epoch": 0.7715575187247935, + "grad_norm": 1.020872712135315, + "learning_rate": 1.7080625922514132e-05, + "loss": 0.8553, + "num_input_tokens_seen": 36686080, + "step": 24105 + }, + { + "epoch": 0.771717559695282, + "grad_norm": 0.8565654158592224, + "learning_rate": 1.7071314649307836e-05, + "loss": 0.4779, + "num_input_tokens_seen": 36693344, + "step": 24110 + }, + { + "epoch": 0.7718776006657705, + "grad_norm": 0.8813328146934509, + "learning_rate": 1.7062004598804448e-05, + "loss": 0.4181, + "num_input_tokens_seen": 36701168, + "step": 24115 + }, + { + "epoch": 0.7720376416362589, + "grad_norm": 0.806854248046875, + "learning_rate": 1.7052695772439702e-05, + "loss": 0.4461, + "num_input_tokens_seen": 36708912, + "step": 24120 + }, + { + "epoch": 0.7721976826067474, + "grad_norm": 0.6404240727424622, + "learning_rate": 1.7043388171649154e-05, + "loss": 0.4614, + "num_input_tokens_seen": 36716528, + "step": 24125 + }, + { + "epoch": 0.7723577235772358, + "grad_norm": 1.0114810466766357, + "learning_rate": 1.7034081797868127e-05, + "loss": 0.4734, + "num_input_tokens_seen": 36725136, + "step": 24130 + }, + { + "epoch": 0.7725177645477243, + "grad_norm": 1.010589361190796, + "learning_rate": 1.70247766525318e-05, + "loss": 0.6542, + "num_input_tokens_seen": 36732896, + "step": 24135 + }, + { + "epoch": 0.7726778055182126, + "grad_norm": 0.7223961353302002, + "learning_rate": 1.701547273707514e-05, + "loss": 0.6216, + "num_input_tokens_seen": 36740672, + "step": 24140 + }, + { + "epoch": 0.7728378464887011, + "grad_norm": 0.7855631113052368, + "learning_rate": 1.7006170052932916e-05, + "loss": 0.5252, + "num_input_tokens_seen": 36748688, + "step": 24145 + }, + { + "epoch": 0.7729978874591895, + "grad_norm": 0.9565868377685547, + "learning_rate": 1.6996868601539735e-05, + "loss": 0.5184, + "num_input_tokens_seen": 36755952, + "step": 24150 + }, + { + "epoch": 0.773157928429678, + "grad_norm": 0.4981763958930969, + "learning_rate": 1.6987568384329977e-05, + "loss": 0.418, + "num_input_tokens_seen": 36763696, + "step": 24155 + }, + { + "epoch": 0.7733179694001664, + "grad_norm": 0.6332204341888428, + "learning_rate": 1.6978269402737866e-05, + "loss": 0.5429, + "num_input_tokens_seen": 36771024, + "step": 24160 + }, + { + "epoch": 0.7734780103706549, + "grad_norm": 1.6622356176376343, + "learning_rate": 1.696897165819743e-05, + "loss": 0.5349, + "num_input_tokens_seen": 36778832, + "step": 24165 + }, + { + "epoch": 0.7736380513411434, + "grad_norm": 0.5288935899734497, + "learning_rate": 1.6959675152142487e-05, + "loss": 0.552, + "num_input_tokens_seen": 36786416, + "step": 24170 + }, + { + "epoch": 0.7737980923116318, + "grad_norm": 0.5188202857971191, + "learning_rate": 1.6950379886006667e-05, + "loss": 0.3953, + "num_input_tokens_seen": 36793648, + "step": 24175 + }, + { + "epoch": 0.7739581332821203, + "grad_norm": 0.4848053753376007, + "learning_rate": 1.6941085861223438e-05, + "loss": 0.3865, + "num_input_tokens_seen": 36801344, + "step": 24180 + }, + { + "epoch": 0.7741181742526086, + "grad_norm": 0.4722879230976105, + "learning_rate": 1.6931793079226034e-05, + "loss": 0.4126, + "num_input_tokens_seen": 36809584, + "step": 24185 + }, + { + "epoch": 0.7742782152230971, + "grad_norm": 0.3934342563152313, + "learning_rate": 1.692250154144754e-05, + "loss": 0.2733, + "num_input_tokens_seen": 36816816, + "step": 24190 + }, + { + "epoch": 0.7744382561935855, + "grad_norm": 0.6755506992340088, + "learning_rate": 1.6913211249320807e-05, + "loss": 0.3856, + "num_input_tokens_seen": 36824560, + "step": 24195 + }, + { + "epoch": 0.774598297164074, + "grad_norm": 0.4990006983280182, + "learning_rate": 1.6903922204278522e-05, + "loss": 0.4869, + "num_input_tokens_seen": 36832464, + "step": 24200 + }, + { + "epoch": 0.774598297164074, + "eval_loss": 0.5038062930107117, + "eval_runtime": 559.1339, + "eval_samples_per_second": 24.835, + "eval_steps_per_second": 12.417, + "num_input_tokens_seen": 36832464, + "step": 24200 + }, + { + "epoch": 0.7747583381345624, + "grad_norm": 0.6297587156295776, + "learning_rate": 1.6894634407753186e-05, + "loss": 0.4406, + "num_input_tokens_seen": 36839840, + "step": 24205 + }, + { + "epoch": 0.7749183791050509, + "grad_norm": 0.6537964940071106, + "learning_rate": 1.6885347861177077e-05, + "loss": 0.6954, + "num_input_tokens_seen": 36847728, + "step": 24210 + }, + { + "epoch": 0.7750784200755393, + "grad_norm": 0.6760811805725098, + "learning_rate": 1.6876062565982298e-05, + "loss": 0.335, + "num_input_tokens_seen": 36854976, + "step": 24215 + }, + { + "epoch": 0.7752384610460278, + "grad_norm": 0.5680379271507263, + "learning_rate": 1.6866778523600774e-05, + "loss": 0.3864, + "num_input_tokens_seen": 36862624, + "step": 24220 + }, + { + "epoch": 0.7753985020165163, + "grad_norm": 1.640418529510498, + "learning_rate": 1.6857495735464195e-05, + "loss": 0.5584, + "num_input_tokens_seen": 36870080, + "step": 24225 + }, + { + "epoch": 0.7755585429870047, + "grad_norm": 0.8970237374305725, + "learning_rate": 1.6848214203004115e-05, + "loss": 0.6668, + "num_input_tokens_seen": 36877600, + "step": 24230 + }, + { + "epoch": 0.7757185839574932, + "grad_norm": 1.106999158859253, + "learning_rate": 1.6838933927651835e-05, + "loss": 0.4715, + "num_input_tokens_seen": 36885616, + "step": 24235 + }, + { + "epoch": 0.7758786249279815, + "grad_norm": 0.923538088798523, + "learning_rate": 1.6829654910838506e-05, + "loss": 0.7217, + "num_input_tokens_seen": 36893392, + "step": 24240 + }, + { + "epoch": 0.77603866589847, + "grad_norm": 0.49733513593673706, + "learning_rate": 1.6820377153995065e-05, + "loss": 0.5085, + "num_input_tokens_seen": 36900880, + "step": 24245 + }, + { + "epoch": 0.7761987068689584, + "grad_norm": 0.9704430103302002, + "learning_rate": 1.681110065855226e-05, + "loss": 0.4053, + "num_input_tokens_seen": 36908336, + "step": 24250 + }, + { + "epoch": 0.7763587478394469, + "grad_norm": 0.7879246473312378, + "learning_rate": 1.6801825425940642e-05, + "loss": 0.3834, + "num_input_tokens_seen": 36915632, + "step": 24255 + }, + { + "epoch": 0.7765187888099353, + "grad_norm": 0.4759298861026764, + "learning_rate": 1.679255145759056e-05, + "loss": 0.4179, + "num_input_tokens_seen": 36923904, + "step": 24260 + }, + { + "epoch": 0.7766788297804238, + "grad_norm": 0.7068710923194885, + "learning_rate": 1.6783278754932187e-05, + "loss": 0.5234, + "num_input_tokens_seen": 36931728, + "step": 24265 + }, + { + "epoch": 0.7768388707509122, + "grad_norm": 0.3167804479598999, + "learning_rate": 1.6774007319395496e-05, + "loss": 0.4461, + "num_input_tokens_seen": 36939744, + "step": 24270 + }, + { + "epoch": 0.7769989117214007, + "grad_norm": 0.5992100834846497, + "learning_rate": 1.6764737152410243e-05, + "loss": 0.4076, + "num_input_tokens_seen": 36947648, + "step": 24275 + }, + { + "epoch": 0.7771589526918892, + "grad_norm": 0.8788628578186035, + "learning_rate": 1.6755468255406016e-05, + "loss": 0.3789, + "num_input_tokens_seen": 36955072, + "step": 24280 + }, + { + "epoch": 0.7773189936623776, + "grad_norm": 0.7546830177307129, + "learning_rate": 1.674620062981219e-05, + "loss": 0.3378, + "num_input_tokens_seen": 36962720, + "step": 24285 + }, + { + "epoch": 0.777479034632866, + "grad_norm": 0.6678546071052551, + "learning_rate": 1.6736934277057947e-05, + "loss": 0.5067, + "num_input_tokens_seen": 36970224, + "step": 24290 + }, + { + "epoch": 0.7776390756033544, + "grad_norm": 0.5686137676239014, + "learning_rate": 1.6727669198572286e-05, + "loss": 0.4483, + "num_input_tokens_seen": 36977872, + "step": 24295 + }, + { + "epoch": 0.7777991165738429, + "grad_norm": 0.8770976662635803, + "learning_rate": 1.6718405395783984e-05, + "loss": 0.4695, + "num_input_tokens_seen": 36985520, + "step": 24300 + }, + { + "epoch": 0.7779591575443313, + "grad_norm": 0.5451494455337524, + "learning_rate": 1.6709142870121643e-05, + "loss": 0.4201, + "num_input_tokens_seen": 36993168, + "step": 24305 + }, + { + "epoch": 0.7781191985148198, + "grad_norm": 0.6737242937088013, + "learning_rate": 1.669988162301367e-05, + "loss": 0.4049, + "num_input_tokens_seen": 37000784, + "step": 24310 + }, + { + "epoch": 0.7782792394853082, + "grad_norm": 0.9658076763153076, + "learning_rate": 1.6690621655888243e-05, + "loss": 0.5037, + "num_input_tokens_seen": 37008976, + "step": 24315 + }, + { + "epoch": 0.7784392804557967, + "grad_norm": 0.9672630429267883, + "learning_rate": 1.6681362970173386e-05, + "loss": 0.5369, + "num_input_tokens_seen": 37016624, + "step": 24320 + }, + { + "epoch": 0.7785993214262851, + "grad_norm": 0.589422345161438, + "learning_rate": 1.6672105567296904e-05, + "loss": 0.3835, + "num_input_tokens_seen": 37024528, + "step": 24325 + }, + { + "epoch": 0.7787593623967736, + "grad_norm": 0.5856667757034302, + "learning_rate": 1.666284944868639e-05, + "loss": 0.4712, + "num_input_tokens_seen": 37032192, + "step": 24330 + }, + { + "epoch": 0.7789194033672621, + "grad_norm": 0.5574761033058167, + "learning_rate": 1.665359461576927e-05, + "loss": 0.4404, + "num_input_tokens_seen": 37039456, + "step": 24335 + }, + { + "epoch": 0.7790794443377504, + "grad_norm": 0.5891138315200806, + "learning_rate": 1.6644341069972736e-05, + "loss": 0.5063, + "num_input_tokens_seen": 37047744, + "step": 24340 + }, + { + "epoch": 0.7792394853082389, + "grad_norm": 0.5205401182174683, + "learning_rate": 1.6635088812723813e-05, + "loss": 0.4492, + "num_input_tokens_seen": 37055152, + "step": 24345 + }, + { + "epoch": 0.7793995262787273, + "grad_norm": 0.531408429145813, + "learning_rate": 1.6625837845449328e-05, + "loss": 0.4883, + "num_input_tokens_seen": 37062736, + "step": 24350 + }, + { + "epoch": 0.7795595672492158, + "grad_norm": 0.4309960603713989, + "learning_rate": 1.6616588169575874e-05, + "loss": 0.4638, + "num_input_tokens_seen": 37070224, + "step": 24355 + }, + { + "epoch": 0.7797196082197042, + "grad_norm": 0.6821696162223816, + "learning_rate": 1.6607339786529878e-05, + "loss": 0.5228, + "num_input_tokens_seen": 37077456, + "step": 24360 + }, + { + "epoch": 0.7798796491901927, + "grad_norm": 0.5392578840255737, + "learning_rate": 1.659809269773756e-05, + "loss": 0.4171, + "num_input_tokens_seen": 37085760, + "step": 24365 + }, + { + "epoch": 0.7800396901606811, + "grad_norm": 0.9295234084129333, + "learning_rate": 1.658884690462493e-05, + "loss": 0.4162, + "num_input_tokens_seen": 37093296, + "step": 24370 + }, + { + "epoch": 0.7801997311311696, + "grad_norm": 1.3146556615829468, + "learning_rate": 1.6579602408617813e-05, + "loss": 0.5578, + "num_input_tokens_seen": 37100768, + "step": 24375 + }, + { + "epoch": 0.7803597721016581, + "grad_norm": 3.1600608825683594, + "learning_rate": 1.657035921114181e-05, + "loss": 0.4545, + "num_input_tokens_seen": 37108928, + "step": 24380 + }, + { + "epoch": 0.7805198130721465, + "grad_norm": 0.6085098385810852, + "learning_rate": 1.656111731362236e-05, + "loss": 0.5184, + "num_input_tokens_seen": 37116848, + "step": 24385 + }, + { + "epoch": 0.780679854042635, + "grad_norm": 0.7063145637512207, + "learning_rate": 1.6551876717484666e-05, + "loss": 0.4384, + "num_input_tokens_seen": 37124288, + "step": 24390 + }, + { + "epoch": 0.7808398950131233, + "grad_norm": 1.0379472970962524, + "learning_rate": 1.6542637424153752e-05, + "loss": 0.3481, + "num_input_tokens_seen": 37131984, + "step": 24395 + }, + { + "epoch": 0.7809999359836118, + "grad_norm": 0.6688527464866638, + "learning_rate": 1.6533399435054418e-05, + "loss": 0.4854, + "num_input_tokens_seen": 37139696, + "step": 24400 + }, + { + "epoch": 0.7809999359836118, + "eval_loss": 0.5034312009811401, + "eval_runtime": 558.9967, + "eval_samples_per_second": 24.841, + "eval_steps_per_second": 12.42, + "num_input_tokens_seen": 37139696, + "step": 24400 + }, + { + "epoch": 0.7811599769541002, + "grad_norm": 0.5756580233573914, + "learning_rate": 1.6524162751611304e-05, + "loss": 0.581, + "num_input_tokens_seen": 37146608, + "step": 24405 + }, + { + "epoch": 0.7813200179245887, + "grad_norm": 0.5849642753601074, + "learning_rate": 1.6514927375248796e-05, + "loss": 0.3242, + "num_input_tokens_seen": 37153936, + "step": 24410 + }, + { + "epoch": 0.7814800588950771, + "grad_norm": 1.2650704383850098, + "learning_rate": 1.6505693307391127e-05, + "loss": 0.4376, + "num_input_tokens_seen": 37161456, + "step": 24415 + }, + { + "epoch": 0.7816400998655656, + "grad_norm": 1.1607019901275635, + "learning_rate": 1.6496460549462288e-05, + "loss": 0.6345, + "num_input_tokens_seen": 37168880, + "step": 24420 + }, + { + "epoch": 0.781800140836054, + "grad_norm": 1.0178508758544922, + "learning_rate": 1.6487229102886097e-05, + "loss": 0.6719, + "num_input_tokens_seen": 37176608, + "step": 24425 + }, + { + "epoch": 0.7819601818065425, + "grad_norm": 0.8699253797531128, + "learning_rate": 1.6477998969086155e-05, + "loss": 0.4971, + "num_input_tokens_seen": 37184208, + "step": 24430 + }, + { + "epoch": 0.782120222777031, + "grad_norm": 0.5072495341300964, + "learning_rate": 1.646877014948587e-05, + "loss": 0.436, + "num_input_tokens_seen": 37192192, + "step": 24435 + }, + { + "epoch": 0.7822802637475194, + "grad_norm": 1.420283555984497, + "learning_rate": 1.6459542645508433e-05, + "loss": 0.3411, + "num_input_tokens_seen": 37199712, + "step": 24440 + }, + { + "epoch": 0.7824403047180079, + "grad_norm": 0.8019717335700989, + "learning_rate": 1.6450316458576852e-05, + "loss": 0.3965, + "num_input_tokens_seen": 37206912, + "step": 24445 + }, + { + "epoch": 0.7826003456884962, + "grad_norm": 0.6538688540458679, + "learning_rate": 1.6441091590113912e-05, + "loss": 0.4209, + "num_input_tokens_seen": 37214432, + "step": 24450 + }, + { + "epoch": 0.7827603866589847, + "grad_norm": 1.3015859127044678, + "learning_rate": 1.6431868041542213e-05, + "loss": 0.6051, + "num_input_tokens_seen": 37222208, + "step": 24455 + }, + { + "epoch": 0.7829204276294731, + "grad_norm": 0.6434647440910339, + "learning_rate": 1.6422645814284123e-05, + "loss": 0.5412, + "num_input_tokens_seen": 37229936, + "step": 24460 + }, + { + "epoch": 0.7830804685999616, + "grad_norm": 1.0387787818908691, + "learning_rate": 1.6413424909761846e-05, + "loss": 0.4366, + "num_input_tokens_seen": 37237376, + "step": 24465 + }, + { + "epoch": 0.78324050957045, + "grad_norm": 0.6799547672271729, + "learning_rate": 1.640420532939736e-05, + "loss": 0.444, + "num_input_tokens_seen": 37245136, + "step": 24470 + }, + { + "epoch": 0.7834005505409385, + "grad_norm": 0.7544641494750977, + "learning_rate": 1.639498707461242e-05, + "loss": 0.3994, + "num_input_tokens_seen": 37252736, + "step": 24475 + }, + { + "epoch": 0.7835605915114269, + "grad_norm": 0.6547905206680298, + "learning_rate": 1.6385770146828614e-05, + "loss": 0.3169, + "num_input_tokens_seen": 37260368, + "step": 24480 + }, + { + "epoch": 0.7837206324819154, + "grad_norm": 0.7238139510154724, + "learning_rate": 1.637655454746731e-05, + "loss": 0.5641, + "num_input_tokens_seen": 37268528, + "step": 24485 + }, + { + "epoch": 0.7838806734524039, + "grad_norm": 1.2439481019973755, + "learning_rate": 1.6367340277949658e-05, + "loss": 0.5064, + "num_input_tokens_seen": 37276432, + "step": 24490 + }, + { + "epoch": 0.7840407144228922, + "grad_norm": 0.5736567974090576, + "learning_rate": 1.635812733969663e-05, + "loss": 0.3231, + "num_input_tokens_seen": 37284224, + "step": 24495 + }, + { + "epoch": 0.7842007553933807, + "grad_norm": 0.7899514436721802, + "learning_rate": 1.634891573412896e-05, + "loss": 0.3677, + "num_input_tokens_seen": 37291536, + "step": 24500 + }, + { + "epoch": 0.7843607963638691, + "grad_norm": 0.6878347396850586, + "learning_rate": 1.6339705462667196e-05, + "loss": 0.4581, + "num_input_tokens_seen": 37298816, + "step": 24505 + }, + { + "epoch": 0.7845208373343576, + "grad_norm": 0.8458444476127625, + "learning_rate": 1.633049652673169e-05, + "loss": 0.5943, + "num_input_tokens_seen": 37306496, + "step": 24510 + }, + { + "epoch": 0.784680878304846, + "grad_norm": 0.5372876524925232, + "learning_rate": 1.632128892774256e-05, + "loss": 0.3094, + "num_input_tokens_seen": 37314016, + "step": 24515 + }, + { + "epoch": 0.7848409192753345, + "grad_norm": 0.61272132396698, + "learning_rate": 1.6312082667119737e-05, + "loss": 0.48, + "num_input_tokens_seen": 37321632, + "step": 24520 + }, + { + "epoch": 0.7850009602458229, + "grad_norm": 0.7799585461616516, + "learning_rate": 1.630287774628296e-05, + "loss": 0.4277, + "num_input_tokens_seen": 37329136, + "step": 24525 + }, + { + "epoch": 0.7851610012163114, + "grad_norm": 0.40579789876937866, + "learning_rate": 1.6293674166651718e-05, + "loss": 0.4762, + "num_input_tokens_seen": 37337168, + "step": 24530 + }, + { + "epoch": 0.7853210421867998, + "grad_norm": 0.36416253447532654, + "learning_rate": 1.6284471929645338e-05, + "loss": 0.3969, + "num_input_tokens_seen": 37344928, + "step": 24535 + }, + { + "epoch": 0.7854810831572883, + "grad_norm": 0.528310239315033, + "learning_rate": 1.627527103668291e-05, + "loss": 0.6048, + "num_input_tokens_seen": 37352432, + "step": 24540 + }, + { + "epoch": 0.7856411241277768, + "grad_norm": 0.8351765275001526, + "learning_rate": 1.6266071489183327e-05, + "loss": 0.5147, + "num_input_tokens_seen": 37359600, + "step": 24545 + }, + { + "epoch": 0.7858011650982651, + "grad_norm": 0.849382758140564, + "learning_rate": 1.6256873288565283e-05, + "loss": 0.6321, + "num_input_tokens_seen": 37366928, + "step": 24550 + }, + { + "epoch": 0.7859612060687536, + "grad_norm": 0.5211536884307861, + "learning_rate": 1.6247676436247245e-05, + "loss": 0.3338, + "num_input_tokens_seen": 37374400, + "step": 24555 + }, + { + "epoch": 0.786121247039242, + "grad_norm": 0.8986665606498718, + "learning_rate": 1.6238480933647486e-05, + "loss": 0.3929, + "num_input_tokens_seen": 37381696, + "step": 24560 + }, + { + "epoch": 0.7862812880097305, + "grad_norm": 0.5751513838768005, + "learning_rate": 1.6229286782184083e-05, + "loss": 0.5164, + "num_input_tokens_seen": 37388944, + "step": 24565 + }, + { + "epoch": 0.7864413289802189, + "grad_norm": 0.8175910711288452, + "learning_rate": 1.622009398327487e-05, + "loss": 0.4045, + "num_input_tokens_seen": 37396064, + "step": 24570 + }, + { + "epoch": 0.7866013699507074, + "grad_norm": 0.8300122618675232, + "learning_rate": 1.6210902538337502e-05, + "loss": 0.497, + "num_input_tokens_seen": 37404176, + "step": 24575 + }, + { + "epoch": 0.7867614109211958, + "grad_norm": 0.5055200457572937, + "learning_rate": 1.6201712448789413e-05, + "loss": 0.6269, + "num_input_tokens_seen": 37411328, + "step": 24580 + }, + { + "epoch": 0.7869214518916843, + "grad_norm": 1.6013551950454712, + "learning_rate": 1.6192523716047827e-05, + "loss": 0.5359, + "num_input_tokens_seen": 37418336, + "step": 24585 + }, + { + "epoch": 0.7870814928621728, + "grad_norm": 0.891038179397583, + "learning_rate": 1.6183336341529776e-05, + "loss": 0.4013, + "num_input_tokens_seen": 37425680, + "step": 24590 + }, + { + "epoch": 0.7872415338326612, + "grad_norm": 0.6331064701080322, + "learning_rate": 1.6174150326652047e-05, + "loss": 0.458, + "num_input_tokens_seen": 37433296, + "step": 24595 + }, + { + "epoch": 0.7874015748031497, + "grad_norm": 1.1487051248550415, + "learning_rate": 1.6164965672831256e-05, + "loss": 0.3727, + "num_input_tokens_seen": 37440752, + "step": 24600 + }, + { + "epoch": 0.7874015748031497, + "eval_loss": 0.5031501650810242, + "eval_runtime": 559.0191, + "eval_samples_per_second": 24.84, + "eval_steps_per_second": 12.42, + "num_input_tokens_seen": 37440752, + "step": 24600 + }, + { + "epoch": 0.787561615773638, + "grad_norm": 1.4069122076034546, + "learning_rate": 1.6155782381483784e-05, + "loss": 0.5786, + "num_input_tokens_seen": 37448224, + "step": 24605 + }, + { + "epoch": 0.7877216567441265, + "grad_norm": 0.7032915949821472, + "learning_rate": 1.6146600454025813e-05, + "loss": 0.5176, + "num_input_tokens_seen": 37456384, + "step": 24610 + }, + { + "epoch": 0.7878816977146149, + "grad_norm": 0.7704972624778748, + "learning_rate": 1.6137419891873317e-05, + "loss": 0.4401, + "num_input_tokens_seen": 37464352, + "step": 24615 + }, + { + "epoch": 0.7880417386851034, + "grad_norm": 0.46020543575286865, + "learning_rate": 1.6128240696442038e-05, + "loss": 0.4005, + "num_input_tokens_seen": 37471664, + "step": 24620 + }, + { + "epoch": 0.7882017796555918, + "grad_norm": 1.3106776475906372, + "learning_rate": 1.611906286914753e-05, + "loss": 0.6452, + "num_input_tokens_seen": 37479504, + "step": 24625 + }, + { + "epoch": 0.7883618206260803, + "grad_norm": 0.46133852005004883, + "learning_rate": 1.6109886411405144e-05, + "loss": 0.3168, + "num_input_tokens_seen": 37487472, + "step": 24630 + }, + { + "epoch": 0.7885218615965687, + "grad_norm": 0.5641418695449829, + "learning_rate": 1.6100711324629985e-05, + "loss": 0.3794, + "num_input_tokens_seen": 37495056, + "step": 24635 + }, + { + "epoch": 0.7886819025670572, + "grad_norm": 0.4280891418457031, + "learning_rate": 1.609153761023698e-05, + "loss": 0.4665, + "num_input_tokens_seen": 37502560, + "step": 24640 + }, + { + "epoch": 0.7888419435375457, + "grad_norm": 0.9962635040283203, + "learning_rate": 1.608236526964083e-05, + "loss": 0.547, + "num_input_tokens_seen": 37510080, + "step": 24645 + }, + { + "epoch": 0.789001984508034, + "grad_norm": 0.9289914965629578, + "learning_rate": 1.607319430425601e-05, + "loss": 0.5237, + "num_input_tokens_seen": 37518112, + "step": 24650 + }, + { + "epoch": 0.7891620254785225, + "grad_norm": 0.5601279735565186, + "learning_rate": 1.606402471549682e-05, + "loss": 0.3796, + "num_input_tokens_seen": 37525760, + "step": 24655 + }, + { + "epoch": 0.7893220664490109, + "grad_norm": 0.9874101281166077, + "learning_rate": 1.6054856504777312e-05, + "loss": 0.476, + "num_input_tokens_seen": 37533184, + "step": 24660 + }, + { + "epoch": 0.7894821074194994, + "grad_norm": 0.8252514600753784, + "learning_rate": 1.6045689673511334e-05, + "loss": 0.4631, + "num_input_tokens_seen": 37540400, + "step": 24665 + }, + { + "epoch": 0.7896421483899878, + "grad_norm": 0.7877565026283264, + "learning_rate": 1.6036524223112548e-05, + "loss": 0.5396, + "num_input_tokens_seen": 37547744, + "step": 24670 + }, + { + "epoch": 0.7898021893604763, + "grad_norm": 0.3264223635196686, + "learning_rate": 1.602736015499436e-05, + "loss": 0.3831, + "num_input_tokens_seen": 37555296, + "step": 24675 + }, + { + "epoch": 0.7899622303309647, + "grad_norm": 0.35128548741340637, + "learning_rate": 1.601819747057e-05, + "loss": 0.4838, + "num_input_tokens_seen": 37562576, + "step": 24680 + }, + { + "epoch": 0.7901222713014532, + "grad_norm": 1.744303584098816, + "learning_rate": 1.6009036171252465e-05, + "loss": 0.4592, + "num_input_tokens_seen": 37569904, + "step": 24685 + }, + { + "epoch": 0.7902823122719416, + "grad_norm": 0.5338437557220459, + "learning_rate": 1.599987625845453e-05, + "loss": 0.3737, + "num_input_tokens_seen": 37577456, + "step": 24690 + }, + { + "epoch": 0.7904423532424301, + "grad_norm": 0.8255091309547424, + "learning_rate": 1.599071773358879e-05, + "loss": 0.5439, + "num_input_tokens_seen": 37585504, + "step": 24695 + }, + { + "epoch": 0.7906023942129186, + "grad_norm": 1.3109936714172363, + "learning_rate": 1.598156059806758e-05, + "loss": 0.3969, + "num_input_tokens_seen": 37593392, + "step": 24700 + }, + { + "epoch": 0.7907624351834069, + "grad_norm": 1.3945949077606201, + "learning_rate": 1.5972404853303062e-05, + "loss": 0.4374, + "num_input_tokens_seen": 37600896, + "step": 24705 + }, + { + "epoch": 0.7909224761538954, + "grad_norm": 1.098273515701294, + "learning_rate": 1.5963250500707172e-05, + "loss": 0.6709, + "num_input_tokens_seen": 37608416, + "step": 24710 + }, + { + "epoch": 0.7910825171243838, + "grad_norm": 0.6362510323524475, + "learning_rate": 1.5954097541691612e-05, + "loss": 0.4166, + "num_input_tokens_seen": 37616064, + "step": 24715 + }, + { + "epoch": 0.7912425580948723, + "grad_norm": 0.6256474852561951, + "learning_rate": 1.5944945977667884e-05, + "loss": 0.4842, + "num_input_tokens_seen": 37623536, + "step": 24720 + }, + { + "epoch": 0.7914025990653607, + "grad_norm": 0.47606512904167175, + "learning_rate": 1.593579581004729e-05, + "loss": 0.4205, + "num_input_tokens_seen": 37631232, + "step": 24725 + }, + { + "epoch": 0.7915626400358492, + "grad_norm": 0.9491265416145325, + "learning_rate": 1.592664704024088e-05, + "loss": 0.4222, + "num_input_tokens_seen": 37638240, + "step": 24730 + }, + { + "epoch": 0.7917226810063376, + "grad_norm": 0.4808104336261749, + "learning_rate": 1.591749966965953e-05, + "loss": 0.4526, + "num_input_tokens_seen": 37646144, + "step": 24735 + }, + { + "epoch": 0.7918827219768261, + "grad_norm": 1.837319016456604, + "learning_rate": 1.5908353699713856e-05, + "loss": 0.6674, + "num_input_tokens_seen": 37653872, + "step": 24740 + }, + { + "epoch": 0.7920427629473145, + "grad_norm": 0.6063370704650879, + "learning_rate": 1.5899209131814298e-05, + "loss": 0.5461, + "num_input_tokens_seen": 37661536, + "step": 24745 + }, + { + "epoch": 0.792202803917803, + "grad_norm": 0.9707294702529907, + "learning_rate": 1.5890065967371067e-05, + "loss": 0.4338, + "num_input_tokens_seen": 37669296, + "step": 24750 + }, + { + "epoch": 0.7923628448882915, + "grad_norm": 1.0886709690093994, + "learning_rate": 1.5880924207794144e-05, + "loss": 0.3761, + "num_input_tokens_seen": 37677104, + "step": 24755 + }, + { + "epoch": 0.7925228858587798, + "grad_norm": 0.844352662563324, + "learning_rate": 1.5871783854493298e-05, + "loss": 0.5438, + "num_input_tokens_seen": 37684432, + "step": 24760 + }, + { + "epoch": 0.7926829268292683, + "grad_norm": 0.9118363261222839, + "learning_rate": 1.5862644908878106e-05, + "loss": 0.3992, + "num_input_tokens_seen": 37692112, + "step": 24765 + }, + { + "epoch": 0.7928429677997567, + "grad_norm": 0.6503305435180664, + "learning_rate": 1.5853507372357885e-05, + "loss": 0.7782, + "num_input_tokens_seen": 37699888, + "step": 24770 + }, + { + "epoch": 0.7930030087702452, + "grad_norm": 1.8689682483673096, + "learning_rate": 1.5844371246341776e-05, + "loss": 0.4961, + "num_input_tokens_seen": 37707008, + "step": 24775 + }, + { + "epoch": 0.7931630497407336, + "grad_norm": 0.6052921414375305, + "learning_rate": 1.5835236532238674e-05, + "loss": 0.4184, + "num_input_tokens_seen": 37714256, + "step": 24780 + }, + { + "epoch": 0.7933230907112221, + "grad_norm": 0.8842248320579529, + "learning_rate": 1.582610323145727e-05, + "loss": 0.3978, + "num_input_tokens_seen": 37721680, + "step": 24785 + }, + { + "epoch": 0.7934831316817105, + "grad_norm": 1.0044060945510864, + "learning_rate": 1.5816971345406035e-05, + "loss": 0.4601, + "num_input_tokens_seen": 37729200, + "step": 24790 + }, + { + "epoch": 0.793643172652199, + "grad_norm": 0.9299100637435913, + "learning_rate": 1.5807840875493225e-05, + "loss": 0.4488, + "num_input_tokens_seen": 37737072, + "step": 24795 + }, + { + "epoch": 0.7938032136226874, + "grad_norm": 0.9079821109771729, + "learning_rate": 1.5798711823126854e-05, + "loss": 0.6113, + "num_input_tokens_seen": 37744512, + "step": 24800 + }, + { + "epoch": 0.7938032136226874, + "eval_loss": 0.5028742551803589, + "eval_runtime": 558.6626, + "eval_samples_per_second": 24.856, + "eval_steps_per_second": 12.428, + "num_input_tokens_seen": 37744512, + "step": 24800 + }, + { + "epoch": 0.7939632545931758, + "grad_norm": 0.4078458845615387, + "learning_rate": 1.578958418971477e-05, + "loss": 0.3233, + "num_input_tokens_seen": 37752240, + "step": 24805 + }, + { + "epoch": 0.7941232955636643, + "grad_norm": 0.7672935128211975, + "learning_rate": 1.578045797666453e-05, + "loss": 0.454, + "num_input_tokens_seen": 37759376, + "step": 24810 + }, + { + "epoch": 0.7942833365341527, + "grad_norm": 0.6847561001777649, + "learning_rate": 1.5771333185383548e-05, + "loss": 0.6477, + "num_input_tokens_seen": 37766704, + "step": 24815 + }, + { + "epoch": 0.7944433775046412, + "grad_norm": 1.0746606588363647, + "learning_rate": 1.576220981727895e-05, + "loss": 0.6014, + "num_input_tokens_seen": 37774512, + "step": 24820 + }, + { + "epoch": 0.7946034184751296, + "grad_norm": 0.46660470962524414, + "learning_rate": 1.575308787375769e-05, + "loss": 0.7864, + "num_input_tokens_seen": 37781952, + "step": 24825 + }, + { + "epoch": 0.7947634594456181, + "grad_norm": 1.897411584854126, + "learning_rate": 1.5743967356226492e-05, + "loss": 0.6618, + "num_input_tokens_seen": 37790032, + "step": 24830 + }, + { + "epoch": 0.7949235004161065, + "grad_norm": 1.0899310111999512, + "learning_rate": 1.5734848266091835e-05, + "loss": 0.6143, + "num_input_tokens_seen": 37797616, + "step": 24835 + }, + { + "epoch": 0.795083541386595, + "grad_norm": 1.477084994316101, + "learning_rate": 1.572573060476001e-05, + "loss": 0.3485, + "num_input_tokens_seen": 37805200, + "step": 24840 + }, + { + "epoch": 0.7952435823570834, + "grad_norm": 0.5014597177505493, + "learning_rate": 1.5716614373637085e-05, + "loss": 0.4647, + "num_input_tokens_seen": 37813104, + "step": 24845 + }, + { + "epoch": 0.7954036233275719, + "grad_norm": 0.7221612334251404, + "learning_rate": 1.570749957412887e-05, + "loss": 0.5867, + "num_input_tokens_seen": 37820912, + "step": 24850 + }, + { + "epoch": 0.7955636642980604, + "grad_norm": 0.6963078379631042, + "learning_rate": 1.5698386207641013e-05, + "loss": 0.4085, + "num_input_tokens_seen": 37828112, + "step": 24855 + }, + { + "epoch": 0.7957237052685487, + "grad_norm": 0.7739929556846619, + "learning_rate": 1.5689274275578884e-05, + "loss": 0.3171, + "num_input_tokens_seen": 37836128, + "step": 24860 + }, + { + "epoch": 0.7958837462390372, + "grad_norm": 0.6007003784179688, + "learning_rate": 1.5680163779347667e-05, + "loss": 0.3866, + "num_input_tokens_seen": 37843952, + "step": 24865 + }, + { + "epoch": 0.7960437872095256, + "grad_norm": 1.1197421550750732, + "learning_rate": 1.5671054720352327e-05, + "loss": 0.5342, + "num_input_tokens_seen": 37851840, + "step": 24870 + }, + { + "epoch": 0.7962038281800141, + "grad_norm": 1.0227628946304321, + "learning_rate": 1.566194709999757e-05, + "loss": 0.4477, + "num_input_tokens_seen": 37859232, + "step": 24875 + }, + { + "epoch": 0.7963638691505025, + "grad_norm": 1.3952676057815552, + "learning_rate": 1.5652840919687933e-05, + "loss": 0.5255, + "num_input_tokens_seen": 37866832, + "step": 24880 + }, + { + "epoch": 0.796523910120991, + "grad_norm": 0.7537621855735779, + "learning_rate": 1.5643736180827676e-05, + "loss": 0.4441, + "num_input_tokens_seen": 37874448, + "step": 24885 + }, + { + "epoch": 0.7966839510914794, + "grad_norm": 0.3390861749649048, + "learning_rate": 1.5634632884820878e-05, + "loss": 0.6385, + "num_input_tokens_seen": 37882304, + "step": 24890 + }, + { + "epoch": 0.7968439920619679, + "grad_norm": 0.5690947771072388, + "learning_rate": 1.5625531033071395e-05, + "loss": 0.4268, + "num_input_tokens_seen": 37889728, + "step": 24895 + }, + { + "epoch": 0.7970040330324563, + "grad_norm": 1.839322566986084, + "learning_rate": 1.5616430626982828e-05, + "loss": 0.6688, + "num_input_tokens_seen": 37896960, + "step": 24900 + }, + { + "epoch": 0.7971640740029448, + "grad_norm": 0.6350906491279602, + "learning_rate": 1.5607331667958575e-05, + "loss": 0.3674, + "num_input_tokens_seen": 37904864, + "step": 24905 + }, + { + "epoch": 0.7973241149734333, + "grad_norm": 0.3550928235054016, + "learning_rate": 1.5598234157401824e-05, + "loss": 0.3277, + "num_input_tokens_seen": 37912112, + "step": 24910 + }, + { + "epoch": 0.7974841559439216, + "grad_norm": 1.6402630805969238, + "learning_rate": 1.5589138096715503e-05, + "loss": 0.5447, + "num_input_tokens_seen": 37919664, + "step": 24915 + }, + { + "epoch": 0.7976441969144101, + "grad_norm": 0.6237583756446838, + "learning_rate": 1.5580043487302365e-05, + "loss": 0.3722, + "num_input_tokens_seen": 37927376, + "step": 24920 + }, + { + "epoch": 0.7978042378848985, + "grad_norm": 0.5395490527153015, + "learning_rate": 1.5570950330564888e-05, + "loss": 0.5126, + "num_input_tokens_seen": 37935072, + "step": 24925 + }, + { + "epoch": 0.797964278855387, + "grad_norm": 0.6502416133880615, + "learning_rate": 1.5561858627905367e-05, + "loss": 0.3658, + "num_input_tokens_seen": 37943136, + "step": 24930 + }, + { + "epoch": 0.7981243198258754, + "grad_norm": 0.6118649244308472, + "learning_rate": 1.5552768380725857e-05, + "loss": 0.3793, + "num_input_tokens_seen": 37950976, + "step": 24935 + }, + { + "epoch": 0.7982843607963639, + "grad_norm": 1.4598636627197266, + "learning_rate": 1.5543679590428183e-05, + "loss": 0.5284, + "num_input_tokens_seen": 37959728, + "step": 24940 + }, + { + "epoch": 0.7984444017668523, + "grad_norm": 0.4550531804561615, + "learning_rate": 1.5534592258413943e-05, + "loss": 0.405, + "num_input_tokens_seen": 37967168, + "step": 24945 + }, + { + "epoch": 0.7986044427373408, + "grad_norm": 0.4333422780036926, + "learning_rate": 1.5525506386084538e-05, + "loss": 0.3144, + "num_input_tokens_seen": 37974656, + "step": 24950 + }, + { + "epoch": 0.7987644837078292, + "grad_norm": 0.7042096853256226, + "learning_rate": 1.55164219748411e-05, + "loss": 0.3837, + "num_input_tokens_seen": 37982176, + "step": 24955 + }, + { + "epoch": 0.7989245246783176, + "grad_norm": 0.5218954086303711, + "learning_rate": 1.550733902608459e-05, + "loss": 0.4244, + "num_input_tokens_seen": 37989712, + "step": 24960 + }, + { + "epoch": 0.7990845656488061, + "grad_norm": 0.9809643626213074, + "learning_rate": 1.549825754121568e-05, + "loss": 0.3948, + "num_input_tokens_seen": 37997504, + "step": 24965 + }, + { + "epoch": 0.7992446066192945, + "grad_norm": 0.3603096306324005, + "learning_rate": 1.5489177521634864e-05, + "loss": 0.4297, + "num_input_tokens_seen": 38005184, + "step": 24970 + }, + { + "epoch": 0.799404647589783, + "grad_norm": 0.5699914693832397, + "learning_rate": 1.5480098968742402e-05, + "loss": 0.6522, + "num_input_tokens_seen": 38012896, + "step": 24975 + }, + { + "epoch": 0.7995646885602714, + "grad_norm": 1.023171305656433, + "learning_rate": 1.5471021883938304e-05, + "loss": 0.4137, + "num_input_tokens_seen": 38020656, + "step": 24980 + }, + { + "epoch": 0.7997247295307599, + "grad_norm": 0.9466267228126526, + "learning_rate": 1.546194626862238e-05, + "loss": 0.4626, + "num_input_tokens_seen": 38028096, + "step": 24985 + }, + { + "epoch": 0.7998847705012483, + "grad_norm": 0.868740439414978, + "learning_rate": 1.5452872124194216e-05, + "loss": 0.4682, + "num_input_tokens_seen": 38035808, + "step": 24990 + }, + { + "epoch": 0.8000448114717368, + "grad_norm": 0.5916742086410522, + "learning_rate": 1.5443799452053136e-05, + "loss": 0.4559, + "num_input_tokens_seen": 38043376, + "step": 24995 + }, + { + "epoch": 0.8002048524422252, + "grad_norm": 1.0272595882415771, + "learning_rate": 1.543472825359828e-05, + "loss": 0.4791, + "num_input_tokens_seen": 38050896, + "step": 25000 + }, + { + "epoch": 0.8002048524422252, + "eval_loss": 0.5027077794075012, + "eval_runtime": 558.5246, + "eval_samples_per_second": 24.862, + "eval_steps_per_second": 12.431, + "num_input_tokens_seen": 38050896, + "step": 25000 + }, + { + "epoch": 0.8003648934127137, + "grad_norm": 0.4817247986793518, + "learning_rate": 1.5425658530228522e-05, + "loss": 0.3876, + "num_input_tokens_seen": 38058944, + "step": 25005 + }, + { + "epoch": 0.800524934383202, + "grad_norm": 0.9943234324455261, + "learning_rate": 1.5416590283342546e-05, + "loss": 0.3767, + "num_input_tokens_seen": 38066544, + "step": 25010 + }, + { + "epoch": 0.8006849753536905, + "grad_norm": 0.6581643223762512, + "learning_rate": 1.5407523514338783e-05, + "loss": 0.5625, + "num_input_tokens_seen": 38074048, + "step": 25015 + }, + { + "epoch": 0.800845016324179, + "grad_norm": 0.7623013854026794, + "learning_rate": 1.539845822461543e-05, + "loss": 0.4391, + "num_input_tokens_seen": 38081664, + "step": 25020 + }, + { + "epoch": 0.8010050572946674, + "grad_norm": 0.8008696436882019, + "learning_rate": 1.538939441557048e-05, + "loss": 0.644, + "num_input_tokens_seen": 38089136, + "step": 25025 + }, + { + "epoch": 0.8011650982651559, + "grad_norm": 0.6115216016769409, + "learning_rate": 1.5380332088601696e-05, + "loss": 0.4189, + "num_input_tokens_seen": 38097120, + "step": 25030 + }, + { + "epoch": 0.8013251392356443, + "grad_norm": 0.7553244829177856, + "learning_rate": 1.537127124510658e-05, + "loss": 0.539, + "num_input_tokens_seen": 38104736, + "step": 25035 + }, + { + "epoch": 0.8014851802061328, + "grad_norm": 1.0701868534088135, + "learning_rate": 1.5362211886482457e-05, + "loss": 0.4368, + "num_input_tokens_seen": 38112480, + "step": 25040 + }, + { + "epoch": 0.8016452211766212, + "grad_norm": 1.2375870943069458, + "learning_rate": 1.5353154014126363e-05, + "loss": 0.5101, + "num_input_tokens_seen": 38120096, + "step": 25045 + }, + { + "epoch": 0.8018052621471097, + "grad_norm": 0.4189166724681854, + "learning_rate": 1.534409762943515e-05, + "loss": 0.4751, + "num_input_tokens_seen": 38127248, + "step": 25050 + }, + { + "epoch": 0.8019653031175981, + "grad_norm": 0.7172994613647461, + "learning_rate": 1.5335042733805438e-05, + "loss": 0.4479, + "num_input_tokens_seen": 38134464, + "step": 25055 + }, + { + "epoch": 0.8021253440880866, + "grad_norm": 0.5681036710739136, + "learning_rate": 1.532598932863358e-05, + "loss": 0.5517, + "num_input_tokens_seen": 38142112, + "step": 25060 + }, + { + "epoch": 0.8022853850585749, + "grad_norm": 1.0575804710388184, + "learning_rate": 1.531693741531574e-05, + "loss": 0.4716, + "num_input_tokens_seen": 38150272, + "step": 25065 + }, + { + "epoch": 0.8024454260290634, + "grad_norm": 0.8831023573875427, + "learning_rate": 1.5307886995247844e-05, + "loss": 0.4201, + "num_input_tokens_seen": 38157440, + "step": 25070 + }, + { + "epoch": 0.8026054669995519, + "grad_norm": 0.5988088250160217, + "learning_rate": 1.529883806982557e-05, + "loss": 0.3217, + "num_input_tokens_seen": 38164464, + "step": 25075 + }, + { + "epoch": 0.8027655079700403, + "grad_norm": 1.1132426261901855, + "learning_rate": 1.5289790640444376e-05, + "loss": 0.4434, + "num_input_tokens_seen": 38171952, + "step": 25080 + }, + { + "epoch": 0.8029255489405288, + "grad_norm": 0.3419592082500458, + "learning_rate": 1.5280744708499494e-05, + "loss": 0.5093, + "num_input_tokens_seen": 38179632, + "step": 25085 + }, + { + "epoch": 0.8030855899110172, + "grad_norm": 0.37156808376312256, + "learning_rate": 1.527170027538591e-05, + "loss": 0.4546, + "num_input_tokens_seen": 38187280, + "step": 25090 + }, + { + "epoch": 0.8032456308815057, + "grad_norm": 0.8176003694534302, + "learning_rate": 1.5262657342498407e-05, + "loss": 0.4978, + "num_input_tokens_seen": 38194448, + "step": 25095 + }, + { + "epoch": 0.8034056718519941, + "grad_norm": 0.45708924531936646, + "learning_rate": 1.52536159112315e-05, + "loss": 0.3036, + "num_input_tokens_seen": 38202208, + "step": 25100 + }, + { + "epoch": 0.8035657128224826, + "grad_norm": 0.9794596433639526, + "learning_rate": 1.5244575982979497e-05, + "loss": 0.5524, + "num_input_tokens_seen": 38209344, + "step": 25105 + }, + { + "epoch": 0.803725753792971, + "grad_norm": 0.6875475645065308, + "learning_rate": 1.5235537559136487e-05, + "loss": 0.482, + "num_input_tokens_seen": 38217216, + "step": 25110 + }, + { + "epoch": 0.8038857947634594, + "grad_norm": 0.3360922634601593, + "learning_rate": 1.5226500641096286e-05, + "loss": 0.4255, + "num_input_tokens_seen": 38224512, + "step": 25115 + }, + { + "epoch": 0.804045835733948, + "grad_norm": 1.876957654953003, + "learning_rate": 1.5217465230252509e-05, + "loss": 0.6139, + "num_input_tokens_seen": 38231808, + "step": 25120 + }, + { + "epoch": 0.8042058767044363, + "grad_norm": 0.8582999110221863, + "learning_rate": 1.5208431327998523e-05, + "loss": 0.6508, + "num_input_tokens_seen": 38239472, + "step": 25125 + }, + { + "epoch": 0.8043659176749248, + "grad_norm": 1.6926531791687012, + "learning_rate": 1.5199398935727477e-05, + "loss": 0.6064, + "num_input_tokens_seen": 38247184, + "step": 25130 + }, + { + "epoch": 0.8045259586454132, + "grad_norm": 0.8251970410346985, + "learning_rate": 1.5190368054832282e-05, + "loss": 0.5511, + "num_input_tokens_seen": 38254528, + "step": 25135 + }, + { + "epoch": 0.8046859996159017, + "grad_norm": 0.6335318684577942, + "learning_rate": 1.5181338686705601e-05, + "loss": 0.4702, + "num_input_tokens_seen": 38262240, + "step": 25140 + }, + { + "epoch": 0.8048460405863901, + "grad_norm": 0.8338454961776733, + "learning_rate": 1.5172310832739889e-05, + "loss": 0.631, + "num_input_tokens_seen": 38269792, + "step": 25145 + }, + { + "epoch": 0.8050060815568786, + "grad_norm": 0.7006444334983826, + "learning_rate": 1.5163284494327346e-05, + "loss": 0.3793, + "num_input_tokens_seen": 38277232, + "step": 25150 + }, + { + "epoch": 0.805166122527367, + "grad_norm": 0.9859421849250793, + "learning_rate": 1.5154259672859952e-05, + "loss": 0.6622, + "num_input_tokens_seen": 38285312, + "step": 25155 + }, + { + "epoch": 0.8053261634978555, + "grad_norm": 0.8544344902038574, + "learning_rate": 1.5145236369729452e-05, + "loss": 0.4683, + "num_input_tokens_seen": 38292736, + "step": 25160 + }, + { + "epoch": 0.8054862044683438, + "grad_norm": 0.719989538192749, + "learning_rate": 1.5136214586327335e-05, + "loss": 0.392, + "num_input_tokens_seen": 38300416, + "step": 25165 + }, + { + "epoch": 0.8056462454388323, + "grad_norm": 0.5860865712165833, + "learning_rate": 1.5127194324044885e-05, + "loss": 0.4667, + "num_input_tokens_seen": 38308144, + "step": 25170 + }, + { + "epoch": 0.8058062864093208, + "grad_norm": 1.2408465147018433, + "learning_rate": 1.5118175584273148e-05, + "loss": 0.5069, + "num_input_tokens_seen": 38315856, + "step": 25175 + }, + { + "epoch": 0.8059663273798092, + "grad_norm": 0.9716469645500183, + "learning_rate": 1.5109158368402909e-05, + "loss": 0.4764, + "num_input_tokens_seen": 38323168, + "step": 25180 + }, + { + "epoch": 0.8061263683502977, + "grad_norm": 0.7879971265792847, + "learning_rate": 1.5100142677824753e-05, + "loss": 0.5059, + "num_input_tokens_seen": 38330704, + "step": 25185 + }, + { + "epoch": 0.8062864093207861, + "grad_norm": 0.5190970301628113, + "learning_rate": 1.509112851392901e-05, + "loss": 0.4578, + "num_input_tokens_seen": 38338032, + "step": 25190 + }, + { + "epoch": 0.8064464502912746, + "grad_norm": 0.384522944688797, + "learning_rate": 1.5082115878105763e-05, + "loss": 0.4713, + "num_input_tokens_seen": 38345376, + "step": 25195 + }, + { + "epoch": 0.806606491261763, + "grad_norm": 0.8932282328605652, + "learning_rate": 1.5073104771744892e-05, + "loss": 0.4708, + "num_input_tokens_seen": 38353216, + "step": 25200 + }, + { + "epoch": 0.806606491261763, + "eval_loss": 0.5020519495010376, + "eval_runtime": 558.8356, + "eval_samples_per_second": 24.848, + "eval_steps_per_second": 12.424, + "num_input_tokens_seen": 38353216, + "step": 25200 + }, + { + "epoch": 0.8067665322322515, + "grad_norm": 0.6141695976257324, + "learning_rate": 1.5064095196236006e-05, + "loss": 0.3941, + "num_input_tokens_seen": 38360656, + "step": 25205 + }, + { + "epoch": 0.8069265732027399, + "grad_norm": 0.6750101447105408, + "learning_rate": 1.50550871529685e-05, + "loss": 0.4764, + "num_input_tokens_seen": 38368192, + "step": 25210 + }, + { + "epoch": 0.8070866141732284, + "grad_norm": 0.5409985184669495, + "learning_rate": 1.5046080643331546e-05, + "loss": 0.4621, + "num_input_tokens_seen": 38376080, + "step": 25215 + }, + { + "epoch": 0.8072466551437167, + "grad_norm": 0.4574755132198334, + "learning_rate": 1.5037075668714028e-05, + "loss": 0.5259, + "num_input_tokens_seen": 38384080, + "step": 25220 + }, + { + "epoch": 0.8074066961142052, + "grad_norm": 0.7051235437393188, + "learning_rate": 1.5028072230504656e-05, + "loss": 0.5569, + "num_input_tokens_seen": 38391472, + "step": 25225 + }, + { + "epoch": 0.8075667370846937, + "grad_norm": 1.2049747705459595, + "learning_rate": 1.5019070330091861e-05, + "loss": 0.4447, + "num_input_tokens_seen": 38398960, + "step": 25230 + }, + { + "epoch": 0.8077267780551821, + "grad_norm": 0.45668360590934753, + "learning_rate": 1.5010069968863843e-05, + "loss": 0.4231, + "num_input_tokens_seen": 38406736, + "step": 25235 + }, + { + "epoch": 0.8078868190256706, + "grad_norm": 0.5585614442825317, + "learning_rate": 1.5001071148208584e-05, + "loss": 0.3998, + "num_input_tokens_seen": 38414368, + "step": 25240 + }, + { + "epoch": 0.808046859996159, + "grad_norm": 0.7682858109474182, + "learning_rate": 1.49920738695138e-05, + "loss": 0.4455, + "num_input_tokens_seen": 38422320, + "step": 25245 + }, + { + "epoch": 0.8082069009666475, + "grad_norm": 3.385753631591797, + "learning_rate": 1.4983078134166995e-05, + "loss": 0.4849, + "num_input_tokens_seen": 38429696, + "step": 25250 + }, + { + "epoch": 0.8083669419371359, + "grad_norm": 0.653083086013794, + "learning_rate": 1.4974083943555428e-05, + "loss": 0.4053, + "num_input_tokens_seen": 38437184, + "step": 25255 + }, + { + "epoch": 0.8085269829076244, + "grad_norm": 0.6504998207092285, + "learning_rate": 1.496509129906611e-05, + "loss": 0.5964, + "num_input_tokens_seen": 38445104, + "step": 25260 + }, + { + "epoch": 0.8086870238781128, + "grad_norm": 0.4722305238246918, + "learning_rate": 1.4956100202085809e-05, + "loss": 0.4458, + "num_input_tokens_seen": 38452576, + "step": 25265 + }, + { + "epoch": 0.8088470648486012, + "grad_norm": 0.674004316329956, + "learning_rate": 1.4947110654001093e-05, + "loss": 0.384, + "num_input_tokens_seen": 38460016, + "step": 25270 + }, + { + "epoch": 0.8090071058190896, + "grad_norm": 1.1165906190872192, + "learning_rate": 1.4938122656198234e-05, + "loss": 0.5012, + "num_input_tokens_seen": 38468560, + "step": 25275 + }, + { + "epoch": 0.8091671467895781, + "grad_norm": 0.6000190377235413, + "learning_rate": 1.4929136210063316e-05, + "loss": 0.4799, + "num_input_tokens_seen": 38476448, + "step": 25280 + }, + { + "epoch": 0.8093271877600666, + "grad_norm": 0.4930257797241211, + "learning_rate": 1.4920151316982146e-05, + "loss": 0.2914, + "num_input_tokens_seen": 38484240, + "step": 25285 + }, + { + "epoch": 0.809487228730555, + "grad_norm": 1.0377743244171143, + "learning_rate": 1.4911167978340312e-05, + "loss": 0.4904, + "num_input_tokens_seen": 38492016, + "step": 25290 + }, + { + "epoch": 0.8096472697010435, + "grad_norm": 0.9329785704612732, + "learning_rate": 1.4902186195523166e-05, + "loss": 0.4494, + "num_input_tokens_seen": 38499152, + "step": 25295 + }, + { + "epoch": 0.8098073106715319, + "grad_norm": 0.9548522233963013, + "learning_rate": 1.4893205969915805e-05, + "loss": 0.4656, + "num_input_tokens_seen": 38506416, + "step": 25300 + }, + { + "epoch": 0.8099673516420204, + "grad_norm": 0.762322187423706, + "learning_rate": 1.4884227302903086e-05, + "loss": 0.4798, + "num_input_tokens_seen": 38513712, + "step": 25305 + }, + { + "epoch": 0.8101273926125088, + "grad_norm": 1.0149580240249634, + "learning_rate": 1.4875250195869653e-05, + "loss": 0.4257, + "num_input_tokens_seen": 38521216, + "step": 25310 + }, + { + "epoch": 0.8102874335829973, + "grad_norm": 1.205828309059143, + "learning_rate": 1.4866274650199862e-05, + "loss": 0.5292, + "num_input_tokens_seen": 38528496, + "step": 25315 + }, + { + "epoch": 0.8104474745534856, + "grad_norm": 0.7428536415100098, + "learning_rate": 1.485730066727788e-05, + "loss": 0.5095, + "num_input_tokens_seen": 38536096, + "step": 25320 + }, + { + "epoch": 0.8106075155239741, + "grad_norm": 1.2267565727233887, + "learning_rate": 1.4848328248487586e-05, + "loss": 0.5668, + "num_input_tokens_seen": 38543424, + "step": 25325 + }, + { + "epoch": 0.8107675564944626, + "grad_norm": 0.9918549656867981, + "learning_rate": 1.4839357395212656e-05, + "loss": 0.6369, + "num_input_tokens_seen": 38551120, + "step": 25330 + }, + { + "epoch": 0.810927597464951, + "grad_norm": 0.7663891911506653, + "learning_rate": 1.4830388108836502e-05, + "loss": 0.4801, + "num_input_tokens_seen": 38558976, + "step": 25335 + }, + { + "epoch": 0.8110876384354395, + "grad_norm": 0.7256098985671997, + "learning_rate": 1.4821420390742299e-05, + "loss": 0.6997, + "num_input_tokens_seen": 38566800, + "step": 25340 + }, + { + "epoch": 0.8112476794059279, + "grad_norm": 0.6347603797912598, + "learning_rate": 1.4812454242312979e-05, + "loss": 0.4091, + "num_input_tokens_seen": 38574336, + "step": 25345 + }, + { + "epoch": 0.8114077203764164, + "grad_norm": 1.2466553449630737, + "learning_rate": 1.4803489664931253e-05, + "loss": 0.5604, + "num_input_tokens_seen": 38581952, + "step": 25350 + }, + { + "epoch": 0.8115677613469048, + "grad_norm": 0.8248147964477539, + "learning_rate": 1.4794526659979544e-05, + "loss": 0.4812, + "num_input_tokens_seen": 38589648, + "step": 25355 + }, + { + "epoch": 0.8117278023173933, + "grad_norm": 0.9094799160957336, + "learning_rate": 1.4785565228840086e-05, + "loss": 0.3819, + "num_input_tokens_seen": 38597456, + "step": 25360 + }, + { + "epoch": 0.8118878432878817, + "grad_norm": 0.573052704334259, + "learning_rate": 1.4776605372894819e-05, + "loss": 0.4578, + "num_input_tokens_seen": 38605488, + "step": 25365 + }, + { + "epoch": 0.8120478842583702, + "grad_norm": 1.3082516193389893, + "learning_rate": 1.4767647093525488e-05, + "loss": 0.5681, + "num_input_tokens_seen": 38612768, + "step": 25370 + }, + { + "epoch": 0.8122079252288585, + "grad_norm": 1.2087496519088745, + "learning_rate": 1.4758690392113566e-05, + "loss": 0.5185, + "num_input_tokens_seen": 38620848, + "step": 25375 + }, + { + "epoch": 0.812367966199347, + "grad_norm": 0.671220600605011, + "learning_rate": 1.4749735270040276e-05, + "loss": 0.557, + "num_input_tokens_seen": 38628432, + "step": 25380 + }, + { + "epoch": 0.8125280071698355, + "grad_norm": 0.7720868587493896, + "learning_rate": 1.4740781728686623e-05, + "loss": 0.4437, + "num_input_tokens_seen": 38635776, + "step": 25385 + }, + { + "epoch": 0.8126880481403239, + "grad_norm": 0.41881152987480164, + "learning_rate": 1.4731829769433358e-05, + "loss": 0.4124, + "num_input_tokens_seen": 38643808, + "step": 25390 + }, + { + "epoch": 0.8128480891108124, + "grad_norm": 0.49981826543807983, + "learning_rate": 1.4722879393660976e-05, + "loss": 0.4978, + "num_input_tokens_seen": 38651568, + "step": 25395 + }, + { + "epoch": 0.8130081300813008, + "grad_norm": 0.5670068860054016, + "learning_rate": 1.4713930602749748e-05, + "loss": 0.5611, + "num_input_tokens_seen": 38659776, + "step": 25400 + }, + { + "epoch": 0.8130081300813008, + "eval_loss": 0.502097487449646, + "eval_runtime": 559.103, + "eval_samples_per_second": 24.836, + "eval_steps_per_second": 12.418, + "num_input_tokens_seen": 38659776, + "step": 25400 + }, + { + "epoch": 0.8131681710517893, + "grad_norm": 1.1054134368896484, + "learning_rate": 1.470498339807968e-05, + "loss": 0.5744, + "num_input_tokens_seen": 38667488, + "step": 25405 + }, + { + "epoch": 0.8133282120222777, + "grad_norm": 0.8061515092849731, + "learning_rate": 1.4696037781030542e-05, + "loss": 0.5365, + "num_input_tokens_seen": 38675024, + "step": 25410 + }, + { + "epoch": 0.8134882529927662, + "grad_norm": 1.5314544439315796, + "learning_rate": 1.4687093752981876e-05, + "loss": 0.5825, + "num_input_tokens_seen": 38682736, + "step": 25415 + }, + { + "epoch": 0.8136482939632546, + "grad_norm": 0.4935750365257263, + "learning_rate": 1.4678151315312943e-05, + "loss": 0.3672, + "num_input_tokens_seen": 38690384, + "step": 25420 + }, + { + "epoch": 0.813808334933743, + "grad_norm": 1.3760390281677246, + "learning_rate": 1.4669210469402789e-05, + "loss": 0.5076, + "num_input_tokens_seen": 38698048, + "step": 25425 + }, + { + "epoch": 0.8139683759042314, + "grad_norm": 1.0017576217651367, + "learning_rate": 1.4660271216630218e-05, + "loss": 0.4529, + "num_input_tokens_seen": 38706176, + "step": 25430 + }, + { + "epoch": 0.8141284168747199, + "grad_norm": 1.2113598585128784, + "learning_rate": 1.4651333558373748e-05, + "loss": 0.5359, + "num_input_tokens_seen": 38714128, + "step": 25435 + }, + { + "epoch": 0.8142884578452084, + "grad_norm": 0.4362415671348572, + "learning_rate": 1.4642397496011707e-05, + "loss": 0.3943, + "num_input_tokens_seen": 38721760, + "step": 25440 + }, + { + "epoch": 0.8144484988156968, + "grad_norm": 1.1527724266052246, + "learning_rate": 1.4633463030922129e-05, + "loss": 0.7778, + "num_input_tokens_seen": 38729328, + "step": 25445 + }, + { + "epoch": 0.8146085397861853, + "grad_norm": 0.8038321733474731, + "learning_rate": 1.462453016448282e-05, + "loss": 0.5255, + "num_input_tokens_seen": 38737216, + "step": 25450 + }, + { + "epoch": 0.8147685807566737, + "grad_norm": 0.8924669027328491, + "learning_rate": 1.4615598898071354e-05, + "loss": 0.4959, + "num_input_tokens_seen": 38744704, + "step": 25455 + }, + { + "epoch": 0.8149286217271622, + "grad_norm": 0.86458420753479, + "learning_rate": 1.4606669233065026e-05, + "loss": 0.4109, + "num_input_tokens_seen": 38752416, + "step": 25460 + }, + { + "epoch": 0.8150886626976506, + "grad_norm": 0.7121837139129639, + "learning_rate": 1.4597741170840914e-05, + "loss": 0.449, + "num_input_tokens_seen": 38759840, + "step": 25465 + }, + { + "epoch": 0.8152487036681391, + "grad_norm": 0.9779587984085083, + "learning_rate": 1.4588814712775853e-05, + "loss": 0.6195, + "num_input_tokens_seen": 38767440, + "step": 25470 + }, + { + "epoch": 0.8154087446386274, + "grad_norm": 0.6997586488723755, + "learning_rate": 1.4579889860246382e-05, + "loss": 0.4007, + "num_input_tokens_seen": 38774976, + "step": 25475 + }, + { + "epoch": 0.8155687856091159, + "grad_norm": 0.9313941597938538, + "learning_rate": 1.457096661462885e-05, + "loss": 0.4316, + "num_input_tokens_seen": 38782144, + "step": 25480 + }, + { + "epoch": 0.8157288265796043, + "grad_norm": 0.25289392471313477, + "learning_rate": 1.4562044977299322e-05, + "loss": 0.5446, + "num_input_tokens_seen": 38789600, + "step": 25485 + }, + { + "epoch": 0.8158888675500928, + "grad_norm": 0.8739418983459473, + "learning_rate": 1.4553124949633623e-05, + "loss": 0.5277, + "num_input_tokens_seen": 38797216, + "step": 25490 + }, + { + "epoch": 0.8160489085205813, + "grad_norm": 1.1957906484603882, + "learning_rate": 1.4544206533007354e-05, + "loss": 0.7102, + "num_input_tokens_seen": 38804640, + "step": 25495 + }, + { + "epoch": 0.8162089494910697, + "grad_norm": 1.1828010082244873, + "learning_rate": 1.4535289728795821e-05, + "loss": 0.4224, + "num_input_tokens_seen": 38812432, + "step": 25500 + }, + { + "epoch": 0.8163689904615582, + "grad_norm": 0.6792615652084351, + "learning_rate": 1.4526374538374132e-05, + "loss": 0.4622, + "num_input_tokens_seen": 38819792, + "step": 25505 + }, + { + "epoch": 0.8165290314320466, + "grad_norm": 1.8673491477966309, + "learning_rate": 1.4517460963117097e-05, + "loss": 0.5551, + "num_input_tokens_seen": 38827696, + "step": 25510 + }, + { + "epoch": 0.8166890724025351, + "grad_norm": 0.8161424398422241, + "learning_rate": 1.4508549004399314e-05, + "loss": 0.4275, + "num_input_tokens_seen": 38835152, + "step": 25515 + }, + { + "epoch": 0.8168491133730235, + "grad_norm": 0.7644457221031189, + "learning_rate": 1.449963866359513e-05, + "loss": 0.4262, + "num_input_tokens_seen": 38842688, + "step": 25520 + }, + { + "epoch": 0.817009154343512, + "grad_norm": 1.1019980907440186, + "learning_rate": 1.4490729942078607e-05, + "loss": 0.3871, + "num_input_tokens_seen": 38849904, + "step": 25525 + }, + { + "epoch": 0.8171691953140003, + "grad_norm": 0.3858011066913605, + "learning_rate": 1.4481822841223608e-05, + "loss": 0.4183, + "num_input_tokens_seen": 38857056, + "step": 25530 + }, + { + "epoch": 0.8173292362844888, + "grad_norm": 0.701706051826477, + "learning_rate": 1.4472917362403704e-05, + "loss": 0.3738, + "num_input_tokens_seen": 38864496, + "step": 25535 + }, + { + "epoch": 0.8174892772549772, + "grad_norm": 0.5106015801429749, + "learning_rate": 1.4464013506992224e-05, + "loss": 0.4747, + "num_input_tokens_seen": 38872224, + "step": 25540 + }, + { + "epoch": 0.8176493182254657, + "grad_norm": 0.728686511516571, + "learning_rate": 1.4455111276362277e-05, + "loss": 0.4624, + "num_input_tokens_seen": 38880240, + "step": 25545 + }, + { + "epoch": 0.8178093591959542, + "grad_norm": 0.7571744918823242, + "learning_rate": 1.4446210671886676e-05, + "loss": 0.3848, + "num_input_tokens_seen": 38887888, + "step": 25550 + }, + { + "epoch": 0.8179694001664426, + "grad_norm": 0.293334037065506, + "learning_rate": 1.4437311694938015e-05, + "loss": 0.4954, + "num_input_tokens_seen": 38895248, + "step": 25555 + }, + { + "epoch": 0.8181294411369311, + "grad_norm": 0.5315829515457153, + "learning_rate": 1.442841434688864e-05, + "loss": 0.3452, + "num_input_tokens_seen": 38902560, + "step": 25560 + }, + { + "epoch": 0.8182894821074195, + "grad_norm": 0.6936334371566772, + "learning_rate": 1.4419518629110615e-05, + "loss": 0.5575, + "num_input_tokens_seen": 38910032, + "step": 25565 + }, + { + "epoch": 0.818449523077908, + "grad_norm": 0.29350101947784424, + "learning_rate": 1.4410624542975778e-05, + "loss": 0.3541, + "num_input_tokens_seen": 38918576, + "step": 25570 + }, + { + "epoch": 0.8186095640483964, + "grad_norm": 0.6227320432662964, + "learning_rate": 1.4401732089855724e-05, + "loss": 0.4481, + "num_input_tokens_seen": 38925824, + "step": 25575 + }, + { + "epoch": 0.8187696050188848, + "grad_norm": 1.1745582818984985, + "learning_rate": 1.4392841271121754e-05, + "loss": 0.615, + "num_input_tokens_seen": 38932976, + "step": 25580 + }, + { + "epoch": 0.8189296459893732, + "grad_norm": 1.2213658094406128, + "learning_rate": 1.438395208814497e-05, + "loss": 0.4128, + "num_input_tokens_seen": 38941056, + "step": 25585 + }, + { + "epoch": 0.8190896869598617, + "grad_norm": 0.40904948115348816, + "learning_rate": 1.4375064542296174e-05, + "loss": 0.52, + "num_input_tokens_seen": 38948352, + "step": 25590 + }, + { + "epoch": 0.8192497279303502, + "grad_norm": 0.839720606803894, + "learning_rate": 1.4366178634945946e-05, + "loss": 0.4551, + "num_input_tokens_seen": 38956000, + "step": 25595 + }, + { + "epoch": 0.8194097689008386, + "grad_norm": 0.6347141861915588, + "learning_rate": 1.4357294367464616e-05, + "loss": 0.4683, + "num_input_tokens_seen": 38963712, + "step": 25600 + }, + { + "epoch": 0.8194097689008386, + "eval_loss": 0.5018065571784973, + "eval_runtime": 559.1417, + "eval_samples_per_second": 24.834, + "eval_steps_per_second": 12.417, + "num_input_tokens_seen": 38963712, + "step": 25600 + }, + { + "epoch": 0.8195698098713271, + "grad_norm": 0.9625940322875977, + "learning_rate": 1.434841174122224e-05, + "loss": 0.559, + "num_input_tokens_seen": 38971392, + "step": 25605 + }, + { + "epoch": 0.8197298508418155, + "grad_norm": 0.6661865711212158, + "learning_rate": 1.4339530757588615e-05, + "loss": 0.4987, + "num_input_tokens_seen": 38979584, + "step": 25610 + }, + { + "epoch": 0.819889891812304, + "grad_norm": 2.2462589740753174, + "learning_rate": 1.433065141793333e-05, + "loss": 0.568, + "num_input_tokens_seen": 38987344, + "step": 25615 + }, + { + "epoch": 0.8200499327827924, + "grad_norm": 0.8753699064254761, + "learning_rate": 1.4321773723625665e-05, + "loss": 0.5345, + "num_input_tokens_seen": 38994352, + "step": 25620 + }, + { + "epoch": 0.8202099737532809, + "grad_norm": 1.2696894407272339, + "learning_rate": 1.4312897676034693e-05, + "loss": 0.3953, + "num_input_tokens_seen": 39001808, + "step": 25625 + }, + { + "epoch": 0.8203700147237692, + "grad_norm": 0.35665228962898254, + "learning_rate": 1.4304023276529188e-05, + "loss": 0.3794, + "num_input_tokens_seen": 39009264, + "step": 25630 + }, + { + "epoch": 0.8205300556942577, + "grad_norm": 0.6704835295677185, + "learning_rate": 1.4295150526477712e-05, + "loss": 0.4853, + "num_input_tokens_seen": 39017056, + "step": 25635 + }, + { + "epoch": 0.8206900966647461, + "grad_norm": 0.4205530881881714, + "learning_rate": 1.4286279427248562e-05, + "loss": 0.4347, + "num_input_tokens_seen": 39024992, + "step": 25640 + }, + { + "epoch": 0.8208501376352346, + "grad_norm": 1.0699958801269531, + "learning_rate": 1.4277409980209747e-05, + "loss": 0.4177, + "num_input_tokens_seen": 39033152, + "step": 25645 + }, + { + "epoch": 0.8210101786057231, + "grad_norm": 0.34197482466697693, + "learning_rate": 1.4268542186729061e-05, + "loss": 0.4294, + "num_input_tokens_seen": 39040768, + "step": 25650 + }, + { + "epoch": 0.8211702195762115, + "grad_norm": 0.7042090892791748, + "learning_rate": 1.4259676048174043e-05, + "loss": 0.5236, + "num_input_tokens_seen": 39048144, + "step": 25655 + }, + { + "epoch": 0.8213302605467, + "grad_norm": 1.2295947074890137, + "learning_rate": 1.4250811565911937e-05, + "loss": 0.5504, + "num_input_tokens_seen": 39055328, + "step": 25660 + }, + { + "epoch": 0.8214903015171884, + "grad_norm": 1.0153353214263916, + "learning_rate": 1.4241948741309782e-05, + "loss": 0.4782, + "num_input_tokens_seen": 39062928, + "step": 25665 + }, + { + "epoch": 0.8216503424876769, + "grad_norm": 0.7554364800453186, + "learning_rate": 1.4233087575734317e-05, + "loss": 0.4746, + "num_input_tokens_seen": 39070128, + "step": 25670 + }, + { + "epoch": 0.8218103834581653, + "grad_norm": 0.38449037075042725, + "learning_rate": 1.422422807055206e-05, + "loss": 0.3542, + "num_input_tokens_seen": 39078000, + "step": 25675 + }, + { + "epoch": 0.8219704244286538, + "grad_norm": 1.3405203819274902, + "learning_rate": 1.4215370227129243e-05, + "loss": 0.4633, + "num_input_tokens_seen": 39085584, + "step": 25680 + }, + { + "epoch": 0.8221304653991421, + "grad_norm": 0.5745534896850586, + "learning_rate": 1.4206514046831876e-05, + "loss": 0.4619, + "num_input_tokens_seen": 39093248, + "step": 25685 + }, + { + "epoch": 0.8222905063696306, + "grad_norm": 0.4322667717933655, + "learning_rate": 1.419765953102567e-05, + "loss": 0.481, + "num_input_tokens_seen": 39100720, + "step": 25690 + }, + { + "epoch": 0.822450547340119, + "grad_norm": 0.9591973423957825, + "learning_rate": 1.4188806681076125e-05, + "loss": 0.5374, + "num_input_tokens_seen": 39108240, + "step": 25695 + }, + { + "epoch": 0.8226105883106075, + "grad_norm": 1.0737783908843994, + "learning_rate": 1.4179955498348443e-05, + "loss": 0.5306, + "num_input_tokens_seen": 39115664, + "step": 25700 + }, + { + "epoch": 0.822770629281096, + "grad_norm": 1.2597333192825317, + "learning_rate": 1.4171105984207605e-05, + "loss": 0.5817, + "num_input_tokens_seen": 39123248, + "step": 25705 + }, + { + "epoch": 0.8229306702515844, + "grad_norm": 0.6815688610076904, + "learning_rate": 1.4162258140018304e-05, + "loss": 0.4481, + "num_input_tokens_seen": 39131184, + "step": 25710 + }, + { + "epoch": 0.8230907112220729, + "grad_norm": 0.9635088443756104, + "learning_rate": 1.4153411967144986e-05, + "loss": 0.4488, + "num_input_tokens_seen": 39139104, + "step": 25715 + }, + { + "epoch": 0.8232507521925613, + "grad_norm": 0.959574818611145, + "learning_rate": 1.4144567466951864e-05, + "loss": 0.6234, + "num_input_tokens_seen": 39146320, + "step": 25720 + }, + { + "epoch": 0.8234107931630498, + "grad_norm": 0.6564578413963318, + "learning_rate": 1.4135724640802844e-05, + "loss": 0.586, + "num_input_tokens_seen": 39154176, + "step": 25725 + }, + { + "epoch": 0.8235708341335382, + "grad_norm": 0.9294217824935913, + "learning_rate": 1.4126883490061615e-05, + "loss": 0.5442, + "num_input_tokens_seen": 39161584, + "step": 25730 + }, + { + "epoch": 0.8237308751040266, + "grad_norm": 0.5229426622390747, + "learning_rate": 1.4118044016091603e-05, + "loss": 0.5059, + "num_input_tokens_seen": 39168784, + "step": 25735 + }, + { + "epoch": 0.823890916074515, + "grad_norm": 0.6865636706352234, + "learning_rate": 1.410920622025594e-05, + "loss": 0.4336, + "num_input_tokens_seen": 39176368, + "step": 25740 + }, + { + "epoch": 0.8240509570450035, + "grad_norm": 0.3958596885204315, + "learning_rate": 1.4100370103917554e-05, + "loss": 0.6382, + "num_input_tokens_seen": 39184096, + "step": 25745 + }, + { + "epoch": 0.8242109980154919, + "grad_norm": 0.5891345739364624, + "learning_rate": 1.409153566843907e-05, + "loss": 0.3693, + "num_input_tokens_seen": 39191536, + "step": 25750 + }, + { + "epoch": 0.8243710389859804, + "grad_norm": 0.7264537811279297, + "learning_rate": 1.408270291518286e-05, + "loss": 0.5044, + "num_input_tokens_seen": 39199472, + "step": 25755 + }, + { + "epoch": 0.8245310799564689, + "grad_norm": 1.4113980531692505, + "learning_rate": 1.407387184551107e-05, + "loss": 0.7019, + "num_input_tokens_seen": 39207056, + "step": 25760 + }, + { + "epoch": 0.8246911209269573, + "grad_norm": 0.7019463181495667, + "learning_rate": 1.4065042460785532e-05, + "loss": 0.3949, + "num_input_tokens_seen": 39214752, + "step": 25765 + }, + { + "epoch": 0.8248511618974458, + "grad_norm": 0.570361852645874, + "learning_rate": 1.405621476236787e-05, + "loss": 0.4687, + "num_input_tokens_seen": 39222480, + "step": 25770 + }, + { + "epoch": 0.8250112028679342, + "grad_norm": 1.1192169189453125, + "learning_rate": 1.4047388751619423e-05, + "loss": 0.5523, + "num_input_tokens_seen": 39229632, + "step": 25775 + }, + { + "epoch": 0.8251712438384227, + "grad_norm": 5.913112163543701, + "learning_rate": 1.4038564429901264e-05, + "loss": 0.3404, + "num_input_tokens_seen": 39237072, + "step": 25780 + }, + { + "epoch": 0.825331284808911, + "grad_norm": 0.6261961460113525, + "learning_rate": 1.4029741798574227e-05, + "loss": 0.3159, + "num_input_tokens_seen": 39244928, + "step": 25785 + }, + { + "epoch": 0.8254913257793995, + "grad_norm": 0.3169487714767456, + "learning_rate": 1.402092085899886e-05, + "loss": 0.2778, + "num_input_tokens_seen": 39252592, + "step": 25790 + }, + { + "epoch": 0.8256513667498879, + "grad_norm": 0.8279714584350586, + "learning_rate": 1.4012101612535464e-05, + "loss": 0.4694, + "num_input_tokens_seen": 39260688, + "step": 25795 + }, + { + "epoch": 0.8258114077203764, + "grad_norm": 0.5414356589317322, + "learning_rate": 1.4003284060544092e-05, + "loss": 0.4308, + "num_input_tokens_seen": 39269392, + "step": 25800 + }, + { + "epoch": 0.8258114077203764, + "eval_loss": 0.5014550685882568, + "eval_runtime": 558.9172, + "eval_samples_per_second": 24.844, + "eval_steps_per_second": 12.422, + "num_input_tokens_seen": 39269392, + "step": 25800 + }, + { + "epoch": 0.8259714486908648, + "grad_norm": 0.8710790276527405, + "learning_rate": 1.3994468204384504e-05, + "loss": 0.5316, + "num_input_tokens_seen": 39276672, + "step": 25805 + }, + { + "epoch": 0.8261314896613533, + "grad_norm": 0.37038248777389526, + "learning_rate": 1.398565404541622e-05, + "loss": 0.4204, + "num_input_tokens_seen": 39284128, + "step": 25810 + }, + { + "epoch": 0.8262915306318418, + "grad_norm": 0.947532594203949, + "learning_rate": 1.3976841584998513e-05, + "loss": 0.6158, + "num_input_tokens_seen": 39291312, + "step": 25815 + }, + { + "epoch": 0.8264515716023302, + "grad_norm": 0.5405230522155762, + "learning_rate": 1.3968030824490352e-05, + "loss": 0.3535, + "num_input_tokens_seen": 39298720, + "step": 25820 + }, + { + "epoch": 0.8266116125728187, + "grad_norm": 0.46682068705558777, + "learning_rate": 1.3959221765250469e-05, + "loss": 0.3427, + "num_input_tokens_seen": 39306112, + "step": 25825 + }, + { + "epoch": 0.8267716535433071, + "grad_norm": 0.6214560866355896, + "learning_rate": 1.3950414408637343e-05, + "loss": 0.3766, + "num_input_tokens_seen": 39313840, + "step": 25830 + }, + { + "epoch": 0.8269316945137956, + "grad_norm": 0.8190863132476807, + "learning_rate": 1.3941608756009166e-05, + "loss": 0.4509, + "num_input_tokens_seen": 39320784, + "step": 25835 + }, + { + "epoch": 0.8270917354842839, + "grad_norm": 1.3834182024002075, + "learning_rate": 1.3932804808723898e-05, + "loss": 0.4601, + "num_input_tokens_seen": 39328080, + "step": 25840 + }, + { + "epoch": 0.8272517764547724, + "grad_norm": 0.7036078572273254, + "learning_rate": 1.3924002568139194e-05, + "loss": 0.4576, + "num_input_tokens_seen": 39334960, + "step": 25845 + }, + { + "epoch": 0.8274118174252608, + "grad_norm": 0.8344947099685669, + "learning_rate": 1.3915202035612485e-05, + "loss": 0.581, + "num_input_tokens_seen": 39342432, + "step": 25850 + }, + { + "epoch": 0.8275718583957493, + "grad_norm": 0.6048165559768677, + "learning_rate": 1.3906403212500935e-05, + "loss": 0.5842, + "num_input_tokens_seen": 39349968, + "step": 25855 + }, + { + "epoch": 0.8277318993662378, + "grad_norm": 0.4911506474018097, + "learning_rate": 1.3897606100161409e-05, + "loss": 0.6659, + "num_input_tokens_seen": 39357472, + "step": 25860 + }, + { + "epoch": 0.8278919403367262, + "grad_norm": 0.642271876335144, + "learning_rate": 1.388881069995055e-05, + "loss": 0.4953, + "num_input_tokens_seen": 39365280, + "step": 25865 + }, + { + "epoch": 0.8280519813072147, + "grad_norm": 0.6209138631820679, + "learning_rate": 1.3880017013224708e-05, + "loss": 0.5734, + "num_input_tokens_seen": 39373968, + "step": 25870 + }, + { + "epoch": 0.8282120222777031, + "grad_norm": 0.8472746014595032, + "learning_rate": 1.3871225041339984e-05, + "loss": 0.4977, + "num_input_tokens_seen": 39381792, + "step": 25875 + }, + { + "epoch": 0.8283720632481916, + "grad_norm": 0.9388529658317566, + "learning_rate": 1.386243478565222e-05, + "loss": 0.518, + "num_input_tokens_seen": 39388464, + "step": 25880 + }, + { + "epoch": 0.82853210421868, + "grad_norm": 1.034156322479248, + "learning_rate": 1.3853646247516966e-05, + "loss": 0.4546, + "num_input_tokens_seen": 39396480, + "step": 25885 + }, + { + "epoch": 0.8286921451891684, + "grad_norm": 0.611189067363739, + "learning_rate": 1.3844859428289545e-05, + "loss": 0.4246, + "num_input_tokens_seen": 39404368, + "step": 25890 + }, + { + "epoch": 0.8288521861596568, + "grad_norm": 0.6274194121360779, + "learning_rate": 1.3836074329324984e-05, + "loss": 0.3884, + "num_input_tokens_seen": 39411472, + "step": 25895 + }, + { + "epoch": 0.8290122271301453, + "grad_norm": 0.5490933060646057, + "learning_rate": 1.3827290951978044e-05, + "loss": 0.3357, + "num_input_tokens_seen": 39418928, + "step": 25900 + }, + { + "epoch": 0.8291722681006337, + "grad_norm": 0.34552690386772156, + "learning_rate": 1.381850929760326e-05, + "loss": 0.4533, + "num_input_tokens_seen": 39427040, + "step": 25905 + }, + { + "epoch": 0.8293323090711222, + "grad_norm": 0.4875420928001404, + "learning_rate": 1.3809729367554842e-05, + "loss": 0.3501, + "num_input_tokens_seen": 39435040, + "step": 25910 + }, + { + "epoch": 0.8294923500416107, + "grad_norm": 0.2686924338340759, + "learning_rate": 1.3800951163186784e-05, + "loss": 0.5884, + "num_input_tokens_seen": 39442544, + "step": 25915 + }, + { + "epoch": 0.8296523910120991, + "grad_norm": 0.7824366092681885, + "learning_rate": 1.3792174685852801e-05, + "loss": 0.3039, + "num_input_tokens_seen": 39450240, + "step": 25920 + }, + { + "epoch": 0.8298124319825876, + "grad_norm": 0.7530903220176697, + "learning_rate": 1.378339993690632e-05, + "loss": 0.5139, + "num_input_tokens_seen": 39457744, + "step": 25925 + }, + { + "epoch": 0.829972472953076, + "grad_norm": 1.2046301364898682, + "learning_rate": 1.3774626917700523e-05, + "loss": 0.5709, + "num_input_tokens_seen": 39465472, + "step": 25930 + }, + { + "epoch": 0.8301325139235645, + "grad_norm": 0.3093004524707794, + "learning_rate": 1.3765855629588334e-05, + "loss": 0.4741, + "num_input_tokens_seen": 39472768, + "step": 25935 + }, + { + "epoch": 0.8302925548940528, + "grad_norm": 0.4062521159648895, + "learning_rate": 1.3757086073922374e-05, + "loss": 0.4674, + "num_input_tokens_seen": 39480432, + "step": 25940 + }, + { + "epoch": 0.8304525958645413, + "grad_norm": 0.5346865057945251, + "learning_rate": 1.3748318252055038e-05, + "loss": 0.3986, + "num_input_tokens_seen": 39488784, + "step": 25945 + }, + { + "epoch": 0.8306126368350297, + "grad_norm": 0.5050056576728821, + "learning_rate": 1.3739552165338416e-05, + "loss": 0.3819, + "num_input_tokens_seen": 39495904, + "step": 25950 + }, + { + "epoch": 0.8307726778055182, + "grad_norm": 1.0239022970199585, + "learning_rate": 1.3730787815124354e-05, + "loss": 0.6558, + "num_input_tokens_seen": 39503456, + "step": 25955 + }, + { + "epoch": 0.8309327187760066, + "grad_norm": 0.4767526388168335, + "learning_rate": 1.3722025202764443e-05, + "loss": 0.4559, + "num_input_tokens_seen": 39511696, + "step": 25960 + }, + { + "epoch": 0.8310927597464951, + "grad_norm": 1.9396461248397827, + "learning_rate": 1.371326432960997e-05, + "loss": 0.6021, + "num_input_tokens_seen": 39519104, + "step": 25965 + }, + { + "epoch": 0.8312528007169836, + "grad_norm": 0.8684155344963074, + "learning_rate": 1.3704505197011969e-05, + "loss": 0.5041, + "num_input_tokens_seen": 39526880, + "step": 25970 + }, + { + "epoch": 0.831412841687472, + "grad_norm": 0.5822229981422424, + "learning_rate": 1.3695747806321224e-05, + "loss": 0.4257, + "num_input_tokens_seen": 39534400, + "step": 25975 + }, + { + "epoch": 0.8315728826579605, + "grad_norm": 0.38492798805236816, + "learning_rate": 1.3686992158888212e-05, + "loss": 0.5156, + "num_input_tokens_seen": 39542112, + "step": 25980 + }, + { + "epoch": 0.8317329236284489, + "grad_norm": 0.47940173745155334, + "learning_rate": 1.367823825606319e-05, + "loss": 0.7055, + "num_input_tokens_seen": 39549680, + "step": 25985 + }, + { + "epoch": 0.8318929645989374, + "grad_norm": 0.7358474731445312, + "learning_rate": 1.36694860991961e-05, + "loss": 0.5074, + "num_input_tokens_seen": 39557440, + "step": 25990 + }, + { + "epoch": 0.8320530055694257, + "grad_norm": 1.1707054376602173, + "learning_rate": 1.3660735689636636e-05, + "loss": 0.4006, + "num_input_tokens_seen": 39564976, + "step": 25995 + }, + { + "epoch": 0.8322130465399142, + "grad_norm": 0.971899151802063, + "learning_rate": 1.365198702873424e-05, + "loss": 0.6569, + "num_input_tokens_seen": 39571936, + "step": 26000 + }, + { + "epoch": 0.8322130465399142, + "eval_loss": 0.5013381838798523, + "eval_runtime": 558.9904, + "eval_samples_per_second": 24.841, + "eval_steps_per_second": 12.421, + "num_input_tokens_seen": 39571936, + "step": 26000 + }, + { + "epoch": 0.8323730875104026, + "grad_norm": 0.7035961747169495, + "learning_rate": 1.364324011783804e-05, + "loss": 0.3777, + "num_input_tokens_seen": 39578800, + "step": 26005 + }, + { + "epoch": 0.8325331284808911, + "grad_norm": 0.8883458971977234, + "learning_rate": 1.3634494958296934e-05, + "loss": 0.5382, + "num_input_tokens_seen": 39586032, + "step": 26010 + }, + { + "epoch": 0.8326931694513795, + "grad_norm": 0.5701582431793213, + "learning_rate": 1.3625751551459542e-05, + "loss": 0.4238, + "num_input_tokens_seen": 39593440, + "step": 26015 + }, + { + "epoch": 0.832853210421868, + "grad_norm": 0.818908154964447, + "learning_rate": 1.3617009898674188e-05, + "loss": 0.6451, + "num_input_tokens_seen": 39600960, + "step": 26020 + }, + { + "epoch": 0.8330132513923565, + "grad_norm": 0.9686407446861267, + "learning_rate": 1.3608270001288967e-05, + "loss": 0.4116, + "num_input_tokens_seen": 39608640, + "step": 26025 + }, + { + "epoch": 0.8331732923628449, + "grad_norm": 1.350632667541504, + "learning_rate": 1.359953186065166e-05, + "loss": 0.5151, + "num_input_tokens_seen": 39616368, + "step": 26030 + }, + { + "epoch": 0.8333333333333334, + "grad_norm": 1.156790018081665, + "learning_rate": 1.3590795478109814e-05, + "loss": 0.4978, + "num_input_tokens_seen": 39623936, + "step": 26035 + }, + { + "epoch": 0.8334933743038218, + "grad_norm": 1.339256763458252, + "learning_rate": 1.3582060855010675e-05, + "loss": 0.4791, + "num_input_tokens_seen": 39631712, + "step": 26040 + }, + { + "epoch": 0.8336534152743102, + "grad_norm": 0.5588196516036987, + "learning_rate": 1.3573327992701245e-05, + "loss": 0.3496, + "num_input_tokens_seen": 39638992, + "step": 26045 + }, + { + "epoch": 0.8338134562447986, + "grad_norm": 0.4723484516143799, + "learning_rate": 1.356459689252823e-05, + "loss": 0.3803, + "num_input_tokens_seen": 39646608, + "step": 26050 + }, + { + "epoch": 0.8339734972152871, + "grad_norm": 0.6005216836929321, + "learning_rate": 1.3555867555838087e-05, + "loss": 0.5343, + "num_input_tokens_seen": 39654128, + "step": 26055 + }, + { + "epoch": 0.8341335381857755, + "grad_norm": 0.5994199514389038, + "learning_rate": 1.3547139983976975e-05, + "loss": 0.4936, + "num_input_tokens_seen": 39661888, + "step": 26060 + }, + { + "epoch": 0.834293579156264, + "grad_norm": 0.6019294261932373, + "learning_rate": 1.3538414178290815e-05, + "loss": 0.4012, + "num_input_tokens_seen": 39669648, + "step": 26065 + }, + { + "epoch": 0.8344536201267524, + "grad_norm": 0.8506550192832947, + "learning_rate": 1.3529690140125209e-05, + "loss": 0.4468, + "num_input_tokens_seen": 39677088, + "step": 26070 + }, + { + "epoch": 0.8346136610972409, + "grad_norm": 0.6993705034255981, + "learning_rate": 1.352096787082553e-05, + "loss": 0.5025, + "num_input_tokens_seen": 39684896, + "step": 26075 + }, + { + "epoch": 0.8347737020677294, + "grad_norm": 1.170447826385498, + "learning_rate": 1.3512247371736871e-05, + "loss": 0.5447, + "num_input_tokens_seen": 39692736, + "step": 26080 + }, + { + "epoch": 0.8349337430382178, + "grad_norm": 1.1207724809646606, + "learning_rate": 1.3503528644204022e-05, + "loss": 0.6192, + "num_input_tokens_seen": 39700000, + "step": 26085 + }, + { + "epoch": 0.8350937840087063, + "grad_norm": 0.6710749268531799, + "learning_rate": 1.349481168957153e-05, + "loss": 0.4984, + "num_input_tokens_seen": 39707856, + "step": 26090 + }, + { + "epoch": 0.8352538249791946, + "grad_norm": 1.1833306550979614, + "learning_rate": 1.3486096509183665e-05, + "loss": 0.6495, + "num_input_tokens_seen": 39715152, + "step": 26095 + }, + { + "epoch": 0.8354138659496831, + "grad_norm": 0.5954766869544983, + "learning_rate": 1.3477383104384406e-05, + "loss": 0.5563, + "num_input_tokens_seen": 39723296, + "step": 26100 + }, + { + "epoch": 0.8355739069201715, + "grad_norm": 0.4900154173374176, + "learning_rate": 1.3468671476517481e-05, + "loss": 0.2761, + "num_input_tokens_seen": 39731216, + "step": 26105 + }, + { + "epoch": 0.83573394789066, + "grad_norm": 1.000715970993042, + "learning_rate": 1.3459961626926326e-05, + "loss": 0.5844, + "num_input_tokens_seen": 39738736, + "step": 26110 + }, + { + "epoch": 0.8358939888611484, + "grad_norm": 0.45230644941329956, + "learning_rate": 1.3451253556954101e-05, + "loss": 0.402, + "num_input_tokens_seen": 39747088, + "step": 26115 + }, + { + "epoch": 0.8360540298316369, + "grad_norm": 0.5341519713401794, + "learning_rate": 1.3442547267943717e-05, + "loss": 0.401, + "num_input_tokens_seen": 39754640, + "step": 26120 + }, + { + "epoch": 0.8362140708021254, + "grad_norm": 0.7013290524482727, + "learning_rate": 1.3433842761237774e-05, + "loss": 0.6334, + "num_input_tokens_seen": 39762384, + "step": 26125 + }, + { + "epoch": 0.8363741117726138, + "grad_norm": 0.6881536841392517, + "learning_rate": 1.3425140038178639e-05, + "loss": 0.4089, + "num_input_tokens_seen": 39769824, + "step": 26130 + }, + { + "epoch": 0.8365341527431023, + "grad_norm": 0.6347928643226624, + "learning_rate": 1.3416439100108358e-05, + "loss": 0.4788, + "num_input_tokens_seen": 39777152, + "step": 26135 + }, + { + "epoch": 0.8366941937135907, + "grad_norm": 0.7235655784606934, + "learning_rate": 1.3407739948368734e-05, + "loss": 0.4105, + "num_input_tokens_seen": 39784800, + "step": 26140 + }, + { + "epoch": 0.8368542346840792, + "grad_norm": 0.7832708358764648, + "learning_rate": 1.3399042584301298e-05, + "loss": 0.355, + "num_input_tokens_seen": 39792544, + "step": 26145 + }, + { + "epoch": 0.8370142756545675, + "grad_norm": 0.5071791410446167, + "learning_rate": 1.3390347009247272e-05, + "loss": 0.4548, + "num_input_tokens_seen": 39799760, + "step": 26150 + }, + { + "epoch": 0.837174316625056, + "grad_norm": 0.5264939665794373, + "learning_rate": 1.3381653224547635e-05, + "loss": 0.4602, + "num_input_tokens_seen": 39808416, + "step": 26155 + }, + { + "epoch": 0.8373343575955444, + "grad_norm": 0.3975908160209656, + "learning_rate": 1.3372961231543086e-05, + "loss": 0.4243, + "num_input_tokens_seen": 39815568, + "step": 26160 + }, + { + "epoch": 0.8374943985660329, + "grad_norm": 0.8571106791496277, + "learning_rate": 1.3364271031574016e-05, + "loss": 0.4994, + "num_input_tokens_seen": 39823056, + "step": 26165 + }, + { + "epoch": 0.8376544395365213, + "grad_norm": 0.6202446222305298, + "learning_rate": 1.335558262598059e-05, + "loss": 0.4102, + "num_input_tokens_seen": 39830640, + "step": 26170 + }, + { + "epoch": 0.8378144805070098, + "grad_norm": 0.6231268048286438, + "learning_rate": 1.3346896016102645e-05, + "loss": 0.6379, + "num_input_tokens_seen": 39838256, + "step": 26175 + }, + { + "epoch": 0.8379745214774983, + "grad_norm": 0.3409198224544525, + "learning_rate": 1.3338211203279788e-05, + "loss": 0.3687, + "num_input_tokens_seen": 39845408, + "step": 26180 + }, + { + "epoch": 0.8381345624479867, + "grad_norm": 0.7445582151412964, + "learning_rate": 1.3329528188851303e-05, + "loss": 0.4322, + "num_input_tokens_seen": 39853104, + "step": 26185 + }, + { + "epoch": 0.8382946034184752, + "grad_norm": 0.7094312906265259, + "learning_rate": 1.3320846974156242e-05, + "loss": 0.4092, + "num_input_tokens_seen": 39860368, + "step": 26190 + }, + { + "epoch": 0.8384546443889636, + "grad_norm": 0.6035698652267456, + "learning_rate": 1.3312167560533337e-05, + "loss": 0.515, + "num_input_tokens_seen": 39867904, + "step": 26195 + }, + { + "epoch": 0.838614685359452, + "grad_norm": 0.8662488460540771, + "learning_rate": 1.3303489949321082e-05, + "loss": 0.4806, + "num_input_tokens_seen": 39875600, + "step": 26200 + }, + { + "epoch": 0.838614685359452, + "eval_loss": 0.5009545683860779, + "eval_runtime": 559.0254, + "eval_samples_per_second": 24.84, + "eval_steps_per_second": 12.42, + "num_input_tokens_seen": 39875600, + "step": 26200 + }, + { + "epoch": 0.8387747263299404, + "grad_norm": 0.7514047026634216, + "learning_rate": 1.3294814141857653e-05, + "loss": 0.4834, + "num_input_tokens_seen": 39883376, + "step": 26205 + }, + { + "epoch": 0.8389347673004289, + "grad_norm": 0.7877265810966492, + "learning_rate": 1.3286140139480992e-05, + "loss": 0.4728, + "num_input_tokens_seen": 39891200, + "step": 26210 + }, + { + "epoch": 0.8390948082709173, + "grad_norm": 1.1680580377578735, + "learning_rate": 1.3277467943528719e-05, + "loss": 0.6442, + "num_input_tokens_seen": 39898656, + "step": 26215 + }, + { + "epoch": 0.8392548492414058, + "grad_norm": 0.632597804069519, + "learning_rate": 1.3268797555338203e-05, + "loss": 0.5487, + "num_input_tokens_seen": 39906592, + "step": 26220 + }, + { + "epoch": 0.8394148902118942, + "grad_norm": 0.9112147092819214, + "learning_rate": 1.3260128976246533e-05, + "loss": 0.2443, + "num_input_tokens_seen": 39913952, + "step": 26225 + }, + { + "epoch": 0.8395749311823827, + "grad_norm": 0.7476250529289246, + "learning_rate": 1.32514622075905e-05, + "loss": 0.6, + "num_input_tokens_seen": 39921920, + "step": 26230 + }, + { + "epoch": 0.8397349721528712, + "grad_norm": 0.7618991732597351, + "learning_rate": 1.3242797250706638e-05, + "loss": 0.4031, + "num_input_tokens_seen": 39929456, + "step": 26235 + }, + { + "epoch": 0.8398950131233596, + "grad_norm": 0.5445509552955627, + "learning_rate": 1.3234134106931195e-05, + "loss": 0.588, + "num_input_tokens_seen": 39936800, + "step": 26240 + }, + { + "epoch": 0.8400550540938481, + "grad_norm": 1.2497798204421997, + "learning_rate": 1.322547277760013e-05, + "loss": 0.6578, + "num_input_tokens_seen": 39944752, + "step": 26245 + }, + { + "epoch": 0.8402150950643364, + "grad_norm": 0.45568373799324036, + "learning_rate": 1.3216813264049132e-05, + "loss": 0.5069, + "num_input_tokens_seen": 39952128, + "step": 26250 + }, + { + "epoch": 0.8403751360348249, + "grad_norm": 0.7172919511795044, + "learning_rate": 1.32081555676136e-05, + "loss": 0.4453, + "num_input_tokens_seen": 39959808, + "step": 26255 + }, + { + "epoch": 0.8405351770053133, + "grad_norm": 0.5663807988166809, + "learning_rate": 1.3199499689628674e-05, + "loss": 0.4025, + "num_input_tokens_seen": 39967328, + "step": 26260 + }, + { + "epoch": 0.8406952179758018, + "grad_norm": 1.4360038042068481, + "learning_rate": 1.3190845631429192e-05, + "loss": 0.5852, + "num_input_tokens_seen": 39975632, + "step": 26265 + }, + { + "epoch": 0.8408552589462902, + "grad_norm": 0.508573591709137, + "learning_rate": 1.3182193394349704e-05, + "loss": 0.2512, + "num_input_tokens_seen": 39982944, + "step": 26270 + }, + { + "epoch": 0.8410152999167787, + "grad_norm": 0.6333335041999817, + "learning_rate": 1.3173542979724507e-05, + "loss": 0.4572, + "num_input_tokens_seen": 39990528, + "step": 26275 + }, + { + "epoch": 0.8411753408872671, + "grad_norm": 0.7447213530540466, + "learning_rate": 1.3164894388887617e-05, + "loss": 0.426, + "num_input_tokens_seen": 39998192, + "step": 26280 + }, + { + "epoch": 0.8413353818577556, + "grad_norm": 0.7881432771682739, + "learning_rate": 1.3156247623172727e-05, + "loss": 0.3568, + "num_input_tokens_seen": 40005504, + "step": 26285 + }, + { + "epoch": 0.8414954228282441, + "grad_norm": 0.6476948261260986, + "learning_rate": 1.3147602683913302e-05, + "loss": 0.565, + "num_input_tokens_seen": 40012912, + "step": 26290 + }, + { + "epoch": 0.8416554637987325, + "grad_norm": 1.1696666479110718, + "learning_rate": 1.3138959572442481e-05, + "loss": 0.5718, + "num_input_tokens_seen": 40020288, + "step": 26295 + }, + { + "epoch": 0.841815504769221, + "grad_norm": 0.8255903720855713, + "learning_rate": 1.3130318290093146e-05, + "loss": 0.5599, + "num_input_tokens_seen": 40027792, + "step": 26300 + }, + { + "epoch": 0.8419755457397093, + "grad_norm": 0.6298211812973022, + "learning_rate": 1.3121678838197909e-05, + "loss": 0.6929, + "num_input_tokens_seen": 40035184, + "step": 26305 + }, + { + "epoch": 0.8421355867101978, + "grad_norm": 0.4883016347885132, + "learning_rate": 1.3113041218089056e-05, + "loss": 0.3649, + "num_input_tokens_seen": 40042848, + "step": 26310 + }, + { + "epoch": 0.8422956276806862, + "grad_norm": 0.7301226854324341, + "learning_rate": 1.3104405431098626e-05, + "loss": 0.4899, + "num_input_tokens_seen": 40050304, + "step": 26315 + }, + { + "epoch": 0.8424556686511747, + "grad_norm": 1.4513016939163208, + "learning_rate": 1.3095771478558377e-05, + "loss": 0.5536, + "num_input_tokens_seen": 40057696, + "step": 26320 + }, + { + "epoch": 0.8426157096216631, + "grad_norm": 0.8512584567070007, + "learning_rate": 1.3087139361799766e-05, + "loss": 0.456, + "num_input_tokens_seen": 40065200, + "step": 26325 + }, + { + "epoch": 0.8427757505921516, + "grad_norm": 0.9131436944007874, + "learning_rate": 1.3078509082153964e-05, + "loss": 0.5256, + "num_input_tokens_seen": 40072992, + "step": 26330 + }, + { + "epoch": 0.8429357915626401, + "grad_norm": 0.9691986441612244, + "learning_rate": 1.3069880640951885e-05, + "loss": 0.4191, + "num_input_tokens_seen": 40080544, + "step": 26335 + }, + { + "epoch": 0.8430958325331285, + "grad_norm": 0.4146193861961365, + "learning_rate": 1.3061254039524123e-05, + "loss": 0.4428, + "num_input_tokens_seen": 40087872, + "step": 26340 + }, + { + "epoch": 0.843255873503617, + "grad_norm": 0.5543784499168396, + "learning_rate": 1.3052629279201028e-05, + "loss": 0.3863, + "num_input_tokens_seen": 40095776, + "step": 26345 + }, + { + "epoch": 0.8434159144741054, + "grad_norm": 2.2977137565612793, + "learning_rate": 1.3044006361312633e-05, + "loss": 0.4867, + "num_input_tokens_seen": 40103456, + "step": 26350 + }, + { + "epoch": 0.8435759554445939, + "grad_norm": 0.9401383996009827, + "learning_rate": 1.30353852871887e-05, + "loss": 0.6473, + "num_input_tokens_seen": 40111136, + "step": 26355 + }, + { + "epoch": 0.8437359964150822, + "grad_norm": 0.6709264516830444, + "learning_rate": 1.302676605815873e-05, + "loss": 0.4527, + "num_input_tokens_seen": 40118864, + "step": 26360 + }, + { + "epoch": 0.8438960373855707, + "grad_norm": 0.6796517372131348, + "learning_rate": 1.3018148675551884e-05, + "loss": 0.6137, + "num_input_tokens_seen": 40127440, + "step": 26365 + }, + { + "epoch": 0.8440560783560591, + "grad_norm": 1.0936214923858643, + "learning_rate": 1.3009533140697094e-05, + "loss": 0.5994, + "num_input_tokens_seen": 40135328, + "step": 26370 + }, + { + "epoch": 0.8442161193265476, + "grad_norm": 0.4535386264324188, + "learning_rate": 1.3000919454922966e-05, + "loss": 0.4514, + "num_input_tokens_seen": 40142848, + "step": 26375 + }, + { + "epoch": 0.844376160297036, + "grad_norm": 0.34783700108528137, + "learning_rate": 1.299230761955785e-05, + "loss": 0.5454, + "num_input_tokens_seen": 40150976, + "step": 26380 + }, + { + "epoch": 0.8445362012675245, + "grad_norm": 0.5022891163825989, + "learning_rate": 1.2983697635929807e-05, + "loss": 0.5023, + "num_input_tokens_seen": 40158768, + "step": 26385 + }, + { + "epoch": 0.844696242238013, + "grad_norm": 0.6994764804840088, + "learning_rate": 1.2975089505366584e-05, + "loss": 0.5487, + "num_input_tokens_seen": 40166624, + "step": 26390 + }, + { + "epoch": 0.8448562832085014, + "grad_norm": 1.8261849880218506, + "learning_rate": 1.2966483229195683e-05, + "loss": 0.4211, + "num_input_tokens_seen": 40174176, + "step": 26395 + }, + { + "epoch": 0.8450163241789899, + "grad_norm": 0.9398272633552551, + "learning_rate": 1.2957878808744283e-05, + "loss": 0.6366, + "num_input_tokens_seen": 40181360, + "step": 26400 + }, + { + "epoch": 0.8450163241789899, + "eval_loss": 0.5006195306777954, + "eval_runtime": 559.1139, + "eval_samples_per_second": 24.836, + "eval_steps_per_second": 12.418, + "num_input_tokens_seen": 40181360, + "step": 26400 + }, + { + "epoch": 0.8451763651494782, + "grad_norm": 0.748626708984375, + "learning_rate": 1.294927624533931e-05, + "loss": 0.4186, + "num_input_tokens_seen": 40188736, + "step": 26405 + }, + { + "epoch": 0.8453364061199667, + "grad_norm": 0.806022047996521, + "learning_rate": 1.2940675540307378e-05, + "loss": 0.4996, + "num_input_tokens_seen": 40196032, + "step": 26410 + }, + { + "epoch": 0.8454964470904551, + "grad_norm": 1.0504236221313477, + "learning_rate": 1.2932076694974814e-05, + "loss": 0.4446, + "num_input_tokens_seen": 40204096, + "step": 26415 + }, + { + "epoch": 0.8456564880609436, + "grad_norm": 0.7848362326622009, + "learning_rate": 1.2923479710667682e-05, + "loss": 0.4139, + "num_input_tokens_seen": 40211696, + "step": 26420 + }, + { + "epoch": 0.845816529031432, + "grad_norm": 1.0254292488098145, + "learning_rate": 1.2914884588711751e-05, + "loss": 0.3814, + "num_input_tokens_seen": 40219168, + "step": 26425 + }, + { + "epoch": 0.8459765700019205, + "grad_norm": 0.7976954579353333, + "learning_rate": 1.2906291330432475e-05, + "loss": 0.5514, + "num_input_tokens_seen": 40226176, + "step": 26430 + }, + { + "epoch": 0.8461366109724089, + "grad_norm": 0.7255022525787354, + "learning_rate": 1.2897699937155055e-05, + "loss": 0.4381, + "num_input_tokens_seen": 40233552, + "step": 26435 + }, + { + "epoch": 0.8462966519428974, + "grad_norm": 0.42550545930862427, + "learning_rate": 1.2889110410204403e-05, + "loss": 0.4864, + "num_input_tokens_seen": 40241152, + "step": 26440 + }, + { + "epoch": 0.8464566929133859, + "grad_norm": 1.197648525238037, + "learning_rate": 1.2880522750905111e-05, + "loss": 0.5545, + "num_input_tokens_seen": 40248928, + "step": 26445 + }, + { + "epoch": 0.8466167338838743, + "grad_norm": 0.328965961933136, + "learning_rate": 1.2871936960581523e-05, + "loss": 0.545, + "num_input_tokens_seen": 40256336, + "step": 26450 + }, + { + "epoch": 0.8467767748543628, + "grad_norm": 1.2199136018753052, + "learning_rate": 1.2863353040557658e-05, + "loss": 0.5115, + "num_input_tokens_seen": 40263440, + "step": 26455 + }, + { + "epoch": 0.8469368158248511, + "grad_norm": 1.9516425132751465, + "learning_rate": 1.2854770992157273e-05, + "loss": 0.6413, + "num_input_tokens_seen": 40270880, + "step": 26460 + }, + { + "epoch": 0.8470968567953396, + "grad_norm": 0.5097949504852295, + "learning_rate": 1.2846190816703835e-05, + "loss": 0.501, + "num_input_tokens_seen": 40278112, + "step": 26465 + }, + { + "epoch": 0.847256897765828, + "grad_norm": 0.4293060004711151, + "learning_rate": 1.2837612515520498e-05, + "loss": 0.4139, + "num_input_tokens_seen": 40285440, + "step": 26470 + }, + { + "epoch": 0.8474169387363165, + "grad_norm": 1.1590739488601685, + "learning_rate": 1.2829036089930163e-05, + "loss": 0.616, + "num_input_tokens_seen": 40293776, + "step": 26475 + }, + { + "epoch": 0.8475769797068049, + "grad_norm": 1.1377424001693726, + "learning_rate": 1.2820461541255412e-05, + "loss": 0.5045, + "num_input_tokens_seen": 40301440, + "step": 26480 + }, + { + "epoch": 0.8477370206772934, + "grad_norm": 1.0712343454360962, + "learning_rate": 1.2811888870818543e-05, + "loss": 0.4512, + "num_input_tokens_seen": 40308912, + "step": 26485 + }, + { + "epoch": 0.8478970616477818, + "grad_norm": 0.9916813373565674, + "learning_rate": 1.2803318079941581e-05, + "loss": 0.4006, + "num_input_tokens_seen": 40316288, + "step": 26490 + }, + { + "epoch": 0.8480571026182703, + "grad_norm": 0.611518144607544, + "learning_rate": 1.2794749169946235e-05, + "loss": 0.5036, + "num_input_tokens_seen": 40323728, + "step": 26495 + }, + { + "epoch": 0.8482171435887588, + "grad_norm": 1.0971330404281616, + "learning_rate": 1.2786182142153952e-05, + "loss": 0.4174, + "num_input_tokens_seen": 40331072, + "step": 26500 + }, + { + "epoch": 0.8483771845592472, + "grad_norm": 1.0499869585037231, + "learning_rate": 1.2777616997885878e-05, + "loss": 0.4754, + "num_input_tokens_seen": 40338720, + "step": 26505 + }, + { + "epoch": 0.8485372255297357, + "grad_norm": 0.6350919008255005, + "learning_rate": 1.2769053738462847e-05, + "loss": 0.6133, + "num_input_tokens_seen": 40346544, + "step": 26510 + }, + { + "epoch": 0.848697266500224, + "grad_norm": 0.5796319842338562, + "learning_rate": 1.2760492365205434e-05, + "loss": 0.4547, + "num_input_tokens_seen": 40354192, + "step": 26515 + }, + { + "epoch": 0.8488573074707125, + "grad_norm": 1.1499720811843872, + "learning_rate": 1.2751932879433919e-05, + "loss": 0.8609, + "num_input_tokens_seen": 40361696, + "step": 26520 + }, + { + "epoch": 0.8490173484412009, + "grad_norm": 1.5590773820877075, + "learning_rate": 1.2743375282468267e-05, + "loss": 0.4672, + "num_input_tokens_seen": 40369296, + "step": 26525 + }, + { + "epoch": 0.8491773894116894, + "grad_norm": 1.2265032529830933, + "learning_rate": 1.2734819575628182e-05, + "loss": 0.4322, + "num_input_tokens_seen": 40376592, + "step": 26530 + }, + { + "epoch": 0.8493374303821778, + "grad_norm": 0.9138855934143066, + "learning_rate": 1.2726265760233039e-05, + "loss": 0.5282, + "num_input_tokens_seen": 40384080, + "step": 26535 + }, + { + "epoch": 0.8494974713526663, + "grad_norm": 0.6791107654571533, + "learning_rate": 1.271771383760197e-05, + "loss": 0.4393, + "num_input_tokens_seen": 40391872, + "step": 26540 + }, + { + "epoch": 0.8496575123231547, + "grad_norm": 0.41317254304885864, + "learning_rate": 1.2709163809053764e-05, + "loss": 0.4911, + "num_input_tokens_seen": 40399440, + "step": 26545 + }, + { + "epoch": 0.8498175532936432, + "grad_norm": 0.5680972933769226, + "learning_rate": 1.2700615675906963e-05, + "loss": 0.4095, + "num_input_tokens_seen": 40407328, + "step": 26550 + }, + { + "epoch": 0.8499775942641317, + "grad_norm": 0.5988373756408691, + "learning_rate": 1.269206943947978e-05, + "loss": 0.5079, + "num_input_tokens_seen": 40414944, + "step": 26555 + }, + { + "epoch": 0.85013763523462, + "grad_norm": 0.6795648336410522, + "learning_rate": 1.2683525101090177e-05, + "loss": 0.414, + "num_input_tokens_seen": 40422352, + "step": 26560 + }, + { + "epoch": 0.8502976762051085, + "grad_norm": 0.7209281325340271, + "learning_rate": 1.2674982662055765e-05, + "loss": 0.4289, + "num_input_tokens_seen": 40430384, + "step": 26565 + }, + { + "epoch": 0.8504577171755969, + "grad_norm": 0.5433172583580017, + "learning_rate": 1.2666442123693922e-05, + "loss": 0.4168, + "num_input_tokens_seen": 40437808, + "step": 26570 + }, + { + "epoch": 0.8506177581460854, + "grad_norm": 1.0219236612319946, + "learning_rate": 1.265790348732169e-05, + "loss": 0.3965, + "num_input_tokens_seen": 40445760, + "step": 26575 + }, + { + "epoch": 0.8507777991165738, + "grad_norm": 1.4297034740447998, + "learning_rate": 1.264936675425584e-05, + "loss": 0.4222, + "num_input_tokens_seen": 40453024, + "step": 26580 + }, + { + "epoch": 0.8509378400870623, + "grad_norm": 0.8747305274009705, + "learning_rate": 1.2640831925812852e-05, + "loss": 0.6194, + "num_input_tokens_seen": 40460624, + "step": 26585 + }, + { + "epoch": 0.8510978810575507, + "grad_norm": 0.9488397836685181, + "learning_rate": 1.263229900330889e-05, + "loss": 0.6263, + "num_input_tokens_seen": 40468256, + "step": 26590 + }, + { + "epoch": 0.8512579220280392, + "grad_norm": 0.42472049593925476, + "learning_rate": 1.2623767988059843e-05, + "loss": 0.3136, + "num_input_tokens_seen": 40475696, + "step": 26595 + }, + { + "epoch": 0.8514179629985277, + "grad_norm": 0.8067901730537415, + "learning_rate": 1.2615238881381309e-05, + "loss": 0.5174, + "num_input_tokens_seen": 40483344, + "step": 26600 + }, + { + "epoch": 0.8514179629985277, + "eval_loss": 0.49995502829551697, + "eval_runtime": 558.9354, + "eval_samples_per_second": 24.844, + "eval_steps_per_second": 12.422, + "num_input_tokens_seen": 40483344, + "step": 26600 + }, + { + "epoch": 0.8515780039690161, + "grad_norm": 0.4959373474121094, + "learning_rate": 1.2606711684588568e-05, + "loss": 0.6125, + "num_input_tokens_seen": 40490976, + "step": 26605 + }, + { + "epoch": 0.8517380449395046, + "grad_norm": 0.7444741129875183, + "learning_rate": 1.2598186398996636e-05, + "loss": 0.3825, + "num_input_tokens_seen": 40498352, + "step": 26610 + }, + { + "epoch": 0.8518980859099929, + "grad_norm": 0.899785578250885, + "learning_rate": 1.2589663025920207e-05, + "loss": 0.4896, + "num_input_tokens_seen": 40505696, + "step": 26615 + }, + { + "epoch": 0.8520581268804814, + "grad_norm": 0.747299313545227, + "learning_rate": 1.2581141566673705e-05, + "loss": 0.5189, + "num_input_tokens_seen": 40513200, + "step": 26620 + }, + { + "epoch": 0.8522181678509698, + "grad_norm": 0.5062180161476135, + "learning_rate": 1.257262202257124e-05, + "loss": 0.4816, + "num_input_tokens_seen": 40520752, + "step": 26625 + }, + { + "epoch": 0.8523782088214583, + "grad_norm": 1.0691739320755005, + "learning_rate": 1.2564104394926618e-05, + "loss": 0.6877, + "num_input_tokens_seen": 40528848, + "step": 26630 + }, + { + "epoch": 0.8525382497919467, + "grad_norm": 0.9646039605140686, + "learning_rate": 1.2555588685053383e-05, + "loss": 0.694, + "num_input_tokens_seen": 40536352, + "step": 26635 + }, + { + "epoch": 0.8526982907624352, + "grad_norm": 0.9169516563415527, + "learning_rate": 1.2547074894264762e-05, + "loss": 0.6409, + "num_input_tokens_seen": 40544448, + "step": 26640 + }, + { + "epoch": 0.8528583317329236, + "grad_norm": 0.9366081357002258, + "learning_rate": 1.2538563023873679e-05, + "loss": 0.4715, + "num_input_tokens_seen": 40552000, + "step": 26645 + }, + { + "epoch": 0.8530183727034121, + "grad_norm": 0.6559948921203613, + "learning_rate": 1.2530053075192789e-05, + "loss": 0.4619, + "num_input_tokens_seen": 40559664, + "step": 26650 + }, + { + "epoch": 0.8531784136739006, + "grad_norm": 0.662040650844574, + "learning_rate": 1.252154504953441e-05, + "loss": 0.5005, + "num_input_tokens_seen": 40567360, + "step": 26655 + }, + { + "epoch": 0.853338454644389, + "grad_norm": 0.6121042966842651, + "learning_rate": 1.25130389482106e-05, + "loss": 0.3629, + "num_input_tokens_seen": 40575056, + "step": 26660 + }, + { + "epoch": 0.8534984956148775, + "grad_norm": 1.239465594291687, + "learning_rate": 1.2504534772533116e-05, + "loss": 0.424, + "num_input_tokens_seen": 40582464, + "step": 26665 + }, + { + "epoch": 0.8536585365853658, + "grad_norm": 0.5823318958282471, + "learning_rate": 1.2496032523813387e-05, + "loss": 0.5367, + "num_input_tokens_seen": 40589872, + "step": 26670 + }, + { + "epoch": 0.8538185775558543, + "grad_norm": 0.8123833537101746, + "learning_rate": 1.2487532203362576e-05, + "loss": 0.7164, + "num_input_tokens_seen": 40597296, + "step": 26675 + }, + { + "epoch": 0.8539786185263427, + "grad_norm": 1.064243197441101, + "learning_rate": 1.247903381249155e-05, + "loss": 0.512, + "num_input_tokens_seen": 40605040, + "step": 26680 + }, + { + "epoch": 0.8541386594968312, + "grad_norm": 1.4013527631759644, + "learning_rate": 1.2470537352510853e-05, + "loss": 0.501, + "num_input_tokens_seen": 40613072, + "step": 26685 + }, + { + "epoch": 0.8542987004673196, + "grad_norm": 0.5635025501251221, + "learning_rate": 1.2462042824730758e-05, + "loss": 0.3485, + "num_input_tokens_seen": 40620576, + "step": 26690 + }, + { + "epoch": 0.8544587414378081, + "grad_norm": 1.523366093635559, + "learning_rate": 1.245355023046122e-05, + "loss": 0.6217, + "num_input_tokens_seen": 40628080, + "step": 26695 + }, + { + "epoch": 0.8546187824082965, + "grad_norm": 0.565075159072876, + "learning_rate": 1.2445059571011896e-05, + "loss": 0.3722, + "num_input_tokens_seen": 40635584, + "step": 26700 + }, + { + "epoch": 0.854778823378785, + "grad_norm": 1.0922062397003174, + "learning_rate": 1.2436570847692173e-05, + "loss": 0.5274, + "num_input_tokens_seen": 40643040, + "step": 26705 + }, + { + "epoch": 0.8549388643492735, + "grad_norm": 0.6647751927375793, + "learning_rate": 1.2428084061811096e-05, + "loss": 0.4674, + "num_input_tokens_seen": 40650320, + "step": 26710 + }, + { + "epoch": 0.8550989053197618, + "grad_norm": 0.6716829538345337, + "learning_rate": 1.2419599214677447e-05, + "loss": 0.4805, + "num_input_tokens_seen": 40657888, + "step": 26715 + }, + { + "epoch": 0.8552589462902503, + "grad_norm": 1.870820164680481, + "learning_rate": 1.2411116307599702e-05, + "loss": 0.6013, + "num_input_tokens_seen": 40665248, + "step": 26720 + }, + { + "epoch": 0.8554189872607387, + "grad_norm": 0.6938775777816772, + "learning_rate": 1.2402635341886016e-05, + "loss": 0.4424, + "num_input_tokens_seen": 40672992, + "step": 26725 + }, + { + "epoch": 0.8555790282312272, + "grad_norm": 0.6646104454994202, + "learning_rate": 1.2394156318844278e-05, + "loss": 0.4745, + "num_input_tokens_seen": 40680544, + "step": 26730 + }, + { + "epoch": 0.8557390692017156, + "grad_norm": 3.1664977073669434, + "learning_rate": 1.2385679239782039e-05, + "loss": 0.4564, + "num_input_tokens_seen": 40687632, + "step": 26735 + }, + { + "epoch": 0.8558991101722041, + "grad_norm": 1.3597532510757446, + "learning_rate": 1.2377204106006585e-05, + "loss": 0.5733, + "num_input_tokens_seen": 40695376, + "step": 26740 + }, + { + "epoch": 0.8560591511426925, + "grad_norm": 0.7459635734558105, + "learning_rate": 1.2368730918824891e-05, + "loss": 0.488, + "num_input_tokens_seen": 40702816, + "step": 26745 + }, + { + "epoch": 0.856219192113181, + "grad_norm": 0.399122953414917, + "learning_rate": 1.236025967954362e-05, + "loss": 0.6113, + "num_input_tokens_seen": 40710992, + "step": 26750 + }, + { + "epoch": 0.8563792330836694, + "grad_norm": 1.4507858753204346, + "learning_rate": 1.2351790389469153e-05, + "loss": 0.568, + "num_input_tokens_seen": 40718256, + "step": 26755 + }, + { + "epoch": 0.8565392740541579, + "grad_norm": 0.7485409379005432, + "learning_rate": 1.234332304990755e-05, + "loss": 0.4722, + "num_input_tokens_seen": 40725184, + "step": 26760 + }, + { + "epoch": 0.8566993150246464, + "grad_norm": 0.49680304527282715, + "learning_rate": 1.2334857662164593e-05, + "loss": 0.5758, + "num_input_tokens_seen": 40732800, + "step": 26765 + }, + { + "epoch": 0.8568593559951347, + "grad_norm": 0.6844622492790222, + "learning_rate": 1.2326394227545743e-05, + "loss": 0.5467, + "num_input_tokens_seen": 40740272, + "step": 26770 + }, + { + "epoch": 0.8570193969656232, + "grad_norm": 0.36988791823387146, + "learning_rate": 1.2317932747356162e-05, + "loss": 0.4449, + "num_input_tokens_seen": 40748288, + "step": 26775 + }, + { + "epoch": 0.8571794379361116, + "grad_norm": 0.6617534160614014, + "learning_rate": 1.2309473222900726e-05, + "loss": 0.5155, + "num_input_tokens_seen": 40755792, + "step": 26780 + }, + { + "epoch": 0.8573394789066001, + "grad_norm": 0.5201579928398132, + "learning_rate": 1.2301015655484006e-05, + "loss": 0.4708, + "num_input_tokens_seen": 40763264, + "step": 26785 + }, + { + "epoch": 0.8574995198770885, + "grad_norm": 0.6665192246437073, + "learning_rate": 1.2292560046410245e-05, + "loss": 0.3836, + "num_input_tokens_seen": 40771312, + "step": 26790 + }, + { + "epoch": 0.857659560847577, + "grad_norm": 0.923427164554596, + "learning_rate": 1.228410639698343e-05, + "loss": 0.4304, + "num_input_tokens_seen": 40779056, + "step": 26795 + }, + { + "epoch": 0.8578196018180654, + "grad_norm": 0.7490281462669373, + "learning_rate": 1.2275654708507195e-05, + "loss": 0.5153, + "num_input_tokens_seen": 40786576, + "step": 26800 + }, + { + "epoch": 0.8578196018180654, + "eval_loss": 0.4995971620082855, + "eval_runtime": 558.9147, + "eval_samples_per_second": 24.845, + "eval_steps_per_second": 12.422, + "num_input_tokens_seen": 40786576, + "step": 26800 + }, + { + "epoch": 0.8579796427885539, + "grad_norm": 0.7384370565414429, + "learning_rate": 1.2267204982284908e-05, + "loss": 0.5366, + "num_input_tokens_seen": 40793936, + "step": 26805 + }, + { + "epoch": 0.8581396837590423, + "grad_norm": 1.1975075006484985, + "learning_rate": 1.2258757219619635e-05, + "loss": 0.4535, + "num_input_tokens_seen": 40801536, + "step": 26810 + }, + { + "epoch": 0.8582997247295308, + "grad_norm": 0.7474463582038879, + "learning_rate": 1.2250311421814104e-05, + "loss": 0.4405, + "num_input_tokens_seen": 40809632, + "step": 26815 + }, + { + "epoch": 0.8584597657000193, + "grad_norm": 1.0278894901275635, + "learning_rate": 1.2241867590170772e-05, + "loss": 0.4392, + "num_input_tokens_seen": 40817408, + "step": 26820 + }, + { + "epoch": 0.8586198066705076, + "grad_norm": 0.7963740229606628, + "learning_rate": 1.2233425725991799e-05, + "loss": 0.7907, + "num_input_tokens_seen": 40825264, + "step": 26825 + }, + { + "epoch": 0.8587798476409961, + "grad_norm": 1.7146879434585571, + "learning_rate": 1.2224985830579003e-05, + "loss": 0.5394, + "num_input_tokens_seen": 40832832, + "step": 26830 + }, + { + "epoch": 0.8589398886114845, + "grad_norm": 1.257388949394226, + "learning_rate": 1.2216547905233944e-05, + "loss": 0.5828, + "num_input_tokens_seen": 40840240, + "step": 26835 + }, + { + "epoch": 0.859099929581973, + "grad_norm": 0.294565349817276, + "learning_rate": 1.2208111951257842e-05, + "loss": 0.4217, + "num_input_tokens_seen": 40848176, + "step": 26840 + }, + { + "epoch": 0.8592599705524614, + "grad_norm": 0.9089974761009216, + "learning_rate": 1.2199677969951622e-05, + "loss": 0.6675, + "num_input_tokens_seen": 40856336, + "step": 26845 + }, + { + "epoch": 0.8594200115229499, + "grad_norm": 0.5147054195404053, + "learning_rate": 1.2191245962615927e-05, + "loss": 0.3353, + "num_input_tokens_seen": 40864064, + "step": 26850 + }, + { + "epoch": 0.8595800524934383, + "grad_norm": 0.5541189312934875, + "learning_rate": 1.218281593055106e-05, + "loss": 0.4683, + "num_input_tokens_seen": 40871792, + "step": 26855 + }, + { + "epoch": 0.8597400934639268, + "grad_norm": 0.599851131439209, + "learning_rate": 1.217438787505705e-05, + "loss": 0.3775, + "num_input_tokens_seen": 40879072, + "step": 26860 + }, + { + "epoch": 0.8599001344344153, + "grad_norm": 1.3155285120010376, + "learning_rate": 1.2165961797433615e-05, + "loss": 0.494, + "num_input_tokens_seen": 40886880, + "step": 26865 + }, + { + "epoch": 0.8600601754049036, + "grad_norm": 0.6482808589935303, + "learning_rate": 1.215753769898014e-05, + "loss": 0.5402, + "num_input_tokens_seen": 40894064, + "step": 26870 + }, + { + "epoch": 0.8602202163753921, + "grad_norm": 0.6711449027061462, + "learning_rate": 1.2149115580995755e-05, + "loss": 0.6065, + "num_input_tokens_seen": 40901568, + "step": 26875 + }, + { + "epoch": 0.8603802573458805, + "grad_norm": 0.5038667917251587, + "learning_rate": 1.2140695444779227e-05, + "loss": 0.3787, + "num_input_tokens_seen": 40909360, + "step": 26880 + }, + { + "epoch": 0.860540298316369, + "grad_norm": 1.1647119522094727, + "learning_rate": 1.2132277291629066e-05, + "loss": 0.4602, + "num_input_tokens_seen": 40917376, + "step": 26885 + }, + { + "epoch": 0.8607003392868574, + "grad_norm": 0.5520946979522705, + "learning_rate": 1.2123861122843458e-05, + "loss": 0.4321, + "num_input_tokens_seen": 40925024, + "step": 26890 + }, + { + "epoch": 0.8608603802573459, + "grad_norm": 0.7747708559036255, + "learning_rate": 1.2115446939720271e-05, + "loss": 0.4807, + "num_input_tokens_seen": 40932992, + "step": 26895 + }, + { + "epoch": 0.8610204212278343, + "grad_norm": 0.7122434973716736, + "learning_rate": 1.210703474355708e-05, + "loss": 0.4147, + "num_input_tokens_seen": 40940496, + "step": 26900 + }, + { + "epoch": 0.8611804621983228, + "grad_norm": 0.4236508905887604, + "learning_rate": 1.2098624535651164e-05, + "loss": 0.4734, + "num_input_tokens_seen": 40947776, + "step": 26905 + }, + { + "epoch": 0.8613405031688112, + "grad_norm": 0.6654263734817505, + "learning_rate": 1.2090216317299477e-05, + "loss": 0.4818, + "num_input_tokens_seen": 40955472, + "step": 26910 + }, + { + "epoch": 0.8615005441392997, + "grad_norm": 0.8818615674972534, + "learning_rate": 1.2081810089798668e-05, + "loss": 0.6407, + "num_input_tokens_seen": 40962928, + "step": 26915 + }, + { + "epoch": 0.8616605851097882, + "grad_norm": 0.7864658832550049, + "learning_rate": 1.2073405854445072e-05, + "loss": 0.4522, + "num_input_tokens_seen": 40970640, + "step": 26920 + }, + { + "epoch": 0.8618206260802765, + "grad_norm": 0.31080177426338196, + "learning_rate": 1.206500361253474e-05, + "loss": 0.3508, + "num_input_tokens_seen": 40978224, + "step": 26925 + }, + { + "epoch": 0.861980667050765, + "grad_norm": 0.5423251390457153, + "learning_rate": 1.2056603365363409e-05, + "loss": 0.5561, + "num_input_tokens_seen": 40985600, + "step": 26930 + }, + { + "epoch": 0.8621407080212534, + "grad_norm": 0.7754898071289062, + "learning_rate": 1.2048205114226487e-05, + "loss": 0.5616, + "num_input_tokens_seen": 40993568, + "step": 26935 + }, + { + "epoch": 0.8623007489917419, + "grad_norm": 0.6607070565223694, + "learning_rate": 1.2039808860419102e-05, + "loss": 0.3241, + "num_input_tokens_seen": 41001648, + "step": 26940 + }, + { + "epoch": 0.8624607899622303, + "grad_norm": 0.48559674620628357, + "learning_rate": 1.2031414605236066e-05, + "loss": 0.4806, + "num_input_tokens_seen": 41009744, + "step": 26945 + }, + { + "epoch": 0.8626208309327188, + "grad_norm": 0.6956878304481506, + "learning_rate": 1.2023022349971862e-05, + "loss": 0.4107, + "num_input_tokens_seen": 41017440, + "step": 26950 + }, + { + "epoch": 0.8627808719032072, + "grad_norm": 0.5112590193748474, + "learning_rate": 1.20146320959207e-05, + "loss": 0.4737, + "num_input_tokens_seen": 41025264, + "step": 26955 + }, + { + "epoch": 0.8629409128736957, + "grad_norm": 0.579740047454834, + "learning_rate": 1.2006243844376445e-05, + "loss": 0.5776, + "num_input_tokens_seen": 41032656, + "step": 26960 + }, + { + "epoch": 0.8631009538441841, + "grad_norm": 0.869291365146637, + "learning_rate": 1.1997857596632678e-05, + "loss": 0.4434, + "num_input_tokens_seen": 41040528, + "step": 26965 + }, + { + "epoch": 0.8632609948146726, + "grad_norm": 0.5636051893234253, + "learning_rate": 1.1989473353982672e-05, + "loss": 0.4458, + "num_input_tokens_seen": 41047856, + "step": 26970 + }, + { + "epoch": 0.863421035785161, + "grad_norm": 0.5276069641113281, + "learning_rate": 1.198109111771937e-05, + "loss": 0.4577, + "num_input_tokens_seen": 41055536, + "step": 26975 + }, + { + "epoch": 0.8635810767556494, + "grad_norm": 0.5231412053108215, + "learning_rate": 1.197271088913543e-05, + "loss": 0.5075, + "num_input_tokens_seen": 41063328, + "step": 26980 + }, + { + "epoch": 0.8637411177261379, + "grad_norm": 0.6623637080192566, + "learning_rate": 1.1964332669523182e-05, + "loss": 0.476, + "num_input_tokens_seen": 41071408, + "step": 26985 + }, + { + "epoch": 0.8639011586966263, + "grad_norm": 0.8353164792060852, + "learning_rate": 1.1955956460174645e-05, + "loss": 0.483, + "num_input_tokens_seen": 41079040, + "step": 26990 + }, + { + "epoch": 0.8640611996671148, + "grad_norm": 0.5238384008407593, + "learning_rate": 1.1947582262381552e-05, + "loss": 0.4229, + "num_input_tokens_seen": 41086880, + "step": 26995 + }, + { + "epoch": 0.8642212406376032, + "grad_norm": 0.5516513586044312, + "learning_rate": 1.1939210077435293e-05, + "loss": 0.5619, + "num_input_tokens_seen": 41094912, + "step": 27000 + }, + { + "epoch": 0.8642212406376032, + "eval_loss": 0.4993181824684143, + "eval_runtime": 559.0665, + "eval_samples_per_second": 24.838, + "eval_steps_per_second": 12.419, + "num_input_tokens_seen": 41094912, + "step": 27000 + }, + { + "epoch": 0.8643812816080917, + "grad_norm": 0.6059340834617615, + "learning_rate": 1.193083990662697e-05, + "loss": 0.3728, + "num_input_tokens_seen": 41102128, + "step": 27005 + }, + { + "epoch": 0.8645413225785801, + "grad_norm": 0.5089453458786011, + "learning_rate": 1.192247175124738e-05, + "loss": 0.4894, + "num_input_tokens_seen": 41109104, + "step": 27010 + }, + { + "epoch": 0.8647013635490686, + "grad_norm": 1.1082758903503418, + "learning_rate": 1.191410561258698e-05, + "loss": 0.4764, + "num_input_tokens_seen": 41116336, + "step": 27015 + }, + { + "epoch": 0.864861404519557, + "grad_norm": 0.4191379249095917, + "learning_rate": 1.1905741491935944e-05, + "loss": 0.5069, + "num_input_tokens_seen": 41124032, + "step": 27020 + }, + { + "epoch": 0.8650214454900454, + "grad_norm": 0.9945096969604492, + "learning_rate": 1.1897379390584129e-05, + "loss": 0.3393, + "num_input_tokens_seen": 41131792, + "step": 27025 + }, + { + "epoch": 0.8651814864605339, + "grad_norm": 0.8307270407676697, + "learning_rate": 1.1889019309821062e-05, + "loss": 0.5145, + "num_input_tokens_seen": 41139472, + "step": 27030 + }, + { + "epoch": 0.8653415274310223, + "grad_norm": 1.0407419204711914, + "learning_rate": 1.188066125093599e-05, + "loss": 0.3911, + "num_input_tokens_seen": 41147264, + "step": 27035 + }, + { + "epoch": 0.8655015684015108, + "grad_norm": 0.5417180061340332, + "learning_rate": 1.1872305215217811e-05, + "loss": 0.3908, + "num_input_tokens_seen": 41154480, + "step": 27040 + }, + { + "epoch": 0.8656616093719992, + "grad_norm": 0.768436074256897, + "learning_rate": 1.186395120395514e-05, + "loss": 0.5119, + "num_input_tokens_seen": 41161840, + "step": 27045 + }, + { + "epoch": 0.8658216503424877, + "grad_norm": 0.6272039413452148, + "learning_rate": 1.1855599218436283e-05, + "loss": 0.3949, + "num_input_tokens_seen": 41169840, + "step": 27050 + }, + { + "epoch": 0.8659816913129761, + "grad_norm": 0.8363997936248779, + "learning_rate": 1.1847249259949209e-05, + "loss": 0.4313, + "num_input_tokens_seen": 41177216, + "step": 27055 + }, + { + "epoch": 0.8661417322834646, + "grad_norm": 0.9999217987060547, + "learning_rate": 1.1838901329781574e-05, + "loss": 0.6241, + "num_input_tokens_seen": 41184784, + "step": 27060 + }, + { + "epoch": 0.866301773253953, + "grad_norm": 0.6923297643661499, + "learning_rate": 1.1830555429220758e-05, + "loss": 0.6533, + "num_input_tokens_seen": 41191936, + "step": 27065 + }, + { + "epoch": 0.8664618142244415, + "grad_norm": 1.047684907913208, + "learning_rate": 1.1822211559553784e-05, + "loss": 0.4943, + "num_input_tokens_seen": 41199264, + "step": 27070 + }, + { + "epoch": 0.86662185519493, + "grad_norm": 0.8977012038230896, + "learning_rate": 1.18138697220674e-05, + "loss": 0.4945, + "num_input_tokens_seen": 41206848, + "step": 27075 + }, + { + "epoch": 0.8667818961654183, + "grad_norm": 0.9695370197296143, + "learning_rate": 1.1805529918048e-05, + "loss": 0.4689, + "num_input_tokens_seen": 41214160, + "step": 27080 + }, + { + "epoch": 0.8669419371359068, + "grad_norm": 1.0870201587677002, + "learning_rate": 1.1797192148781702e-05, + "loss": 0.3892, + "num_input_tokens_seen": 41221504, + "step": 27085 + }, + { + "epoch": 0.8671019781063952, + "grad_norm": 0.7840529680252075, + "learning_rate": 1.1788856415554297e-05, + "loss": 0.4173, + "num_input_tokens_seen": 41229008, + "step": 27090 + }, + { + "epoch": 0.8672620190768837, + "grad_norm": 1.0910364389419556, + "learning_rate": 1.1780522719651249e-05, + "loss": 0.5611, + "num_input_tokens_seen": 41237328, + "step": 27095 + }, + { + "epoch": 0.8674220600473721, + "grad_norm": 0.4516065716743469, + "learning_rate": 1.1772191062357721e-05, + "loss": 0.5642, + "num_input_tokens_seen": 41244720, + "step": 27100 + }, + { + "epoch": 0.8675821010178606, + "grad_norm": 0.6788187026977539, + "learning_rate": 1.1763861444958573e-05, + "loss": 0.4582, + "num_input_tokens_seen": 41251856, + "step": 27105 + }, + { + "epoch": 0.867742141988349, + "grad_norm": 0.3937881886959076, + "learning_rate": 1.1755533868738317e-05, + "loss": 0.4042, + "num_input_tokens_seen": 41259712, + "step": 27110 + }, + { + "epoch": 0.8679021829588375, + "grad_norm": 0.6200088858604431, + "learning_rate": 1.1747208334981185e-05, + "loss": 0.427, + "num_input_tokens_seen": 41267456, + "step": 27115 + }, + { + "epoch": 0.8680622239293259, + "grad_norm": 1.0228453874588013, + "learning_rate": 1.1738884844971067e-05, + "loss": 0.6654, + "num_input_tokens_seen": 41274976, + "step": 27120 + }, + { + "epoch": 0.8682222648998144, + "grad_norm": 1.104681372642517, + "learning_rate": 1.1730563399991563e-05, + "loss": 0.4495, + "num_input_tokens_seen": 41282656, + "step": 27125 + }, + { + "epoch": 0.8683823058703029, + "grad_norm": 1.5933563709259033, + "learning_rate": 1.1722244001325938e-05, + "loss": 0.582, + "num_input_tokens_seen": 41290864, + "step": 27130 + }, + { + "epoch": 0.8685423468407912, + "grad_norm": 0.711534321308136, + "learning_rate": 1.1713926650257137e-05, + "loss": 0.3974, + "num_input_tokens_seen": 41297744, + "step": 27135 + }, + { + "epoch": 0.8687023878112797, + "grad_norm": 0.6547474265098572, + "learning_rate": 1.170561134806781e-05, + "loss": 0.406, + "num_input_tokens_seen": 41305664, + "step": 27140 + }, + { + "epoch": 0.8688624287817681, + "grad_norm": 0.909949541091919, + "learning_rate": 1.1697298096040287e-05, + "loss": 0.5741, + "num_input_tokens_seen": 41313168, + "step": 27145 + }, + { + "epoch": 0.8690224697522566, + "grad_norm": 0.2925877273082733, + "learning_rate": 1.1688986895456567e-05, + "loss": 0.3678, + "num_input_tokens_seen": 41320720, + "step": 27150 + }, + { + "epoch": 0.869182510722745, + "grad_norm": 1.7273403406143188, + "learning_rate": 1.1680677747598349e-05, + "loss": 0.5619, + "num_input_tokens_seen": 41328272, + "step": 27155 + }, + { + "epoch": 0.8693425516932335, + "grad_norm": 0.5678044557571411, + "learning_rate": 1.1672370653746995e-05, + "loss": 0.5255, + "num_input_tokens_seen": 41336080, + "step": 27160 + }, + { + "epoch": 0.8695025926637219, + "grad_norm": 1.6169724464416504, + "learning_rate": 1.166406561518357e-05, + "loss": 0.8348, + "num_input_tokens_seen": 41343120, + "step": 27165 + }, + { + "epoch": 0.8696626336342104, + "grad_norm": 0.596406102180481, + "learning_rate": 1.1655762633188826e-05, + "loss": 0.4861, + "num_input_tokens_seen": 41350432, + "step": 27170 + }, + { + "epoch": 0.8698226746046988, + "grad_norm": 0.6591464877128601, + "learning_rate": 1.1647461709043172e-05, + "loss": 0.4459, + "num_input_tokens_seen": 41358496, + "step": 27175 + }, + { + "epoch": 0.8699827155751872, + "grad_norm": 0.7108458280563354, + "learning_rate": 1.1639162844026722e-05, + "loss": 0.3981, + "num_input_tokens_seen": 41366192, + "step": 27180 + }, + { + "epoch": 0.8701427565456757, + "grad_norm": 1.197798490524292, + "learning_rate": 1.163086603941927e-05, + "loss": 0.4848, + "num_input_tokens_seen": 41373360, + "step": 27185 + }, + { + "epoch": 0.8703027975161641, + "grad_norm": 0.8189484477043152, + "learning_rate": 1.1622571296500273e-05, + "loss": 0.5017, + "num_input_tokens_seen": 41380784, + "step": 27190 + }, + { + "epoch": 0.8704628384866526, + "grad_norm": 0.6820306181907654, + "learning_rate": 1.1614278616548904e-05, + "loss": 0.3764, + "num_input_tokens_seen": 41388208, + "step": 27195 + }, + { + "epoch": 0.870622879457141, + "grad_norm": 0.6322129964828491, + "learning_rate": 1.1605988000843986e-05, + "loss": 0.3401, + "num_input_tokens_seen": 41395648, + "step": 27200 + }, + { + "epoch": 0.870622879457141, + "eval_loss": 0.49904781579971313, + "eval_runtime": 559.1256, + "eval_samples_per_second": 24.835, + "eval_steps_per_second": 12.418, + "num_input_tokens_seen": 41395648, + "step": 27200 + }, + { + "epoch": 0.8707829204276295, + "grad_norm": 0.6818961501121521, + "learning_rate": 1.1597699450664028e-05, + "loss": 0.4915, + "num_input_tokens_seen": 41403040, + "step": 27205 + }, + { + "epoch": 0.8709429613981179, + "grad_norm": 1.6817642450332642, + "learning_rate": 1.1589412967287252e-05, + "loss": 0.53, + "num_input_tokens_seen": 41410592, + "step": 27210 + }, + { + "epoch": 0.8711030023686064, + "grad_norm": 0.7895181179046631, + "learning_rate": 1.1581128551991514e-05, + "loss": 0.3564, + "num_input_tokens_seen": 41418288, + "step": 27215 + }, + { + "epoch": 0.8712630433390948, + "grad_norm": 0.6045815944671631, + "learning_rate": 1.1572846206054383e-05, + "loss": 0.4516, + "num_input_tokens_seen": 41426144, + "step": 27220 + }, + { + "epoch": 0.8714230843095833, + "grad_norm": 0.8300838470458984, + "learning_rate": 1.1564565930753113e-05, + "loss": 0.4868, + "num_input_tokens_seen": 41433344, + "step": 27225 + }, + { + "epoch": 0.8715831252800716, + "grad_norm": 0.7785578966140747, + "learning_rate": 1.1556287727364606e-05, + "loss": 0.4394, + "num_input_tokens_seen": 41441152, + "step": 27230 + }, + { + "epoch": 0.8717431662505601, + "grad_norm": 1.293587565422058, + "learning_rate": 1.1548011597165489e-05, + "loss": 0.427, + "num_input_tokens_seen": 41449216, + "step": 27235 + }, + { + "epoch": 0.8719032072210486, + "grad_norm": 0.5499116778373718, + "learning_rate": 1.1539737541432019e-05, + "loss": 0.4553, + "num_input_tokens_seen": 41456832, + "step": 27240 + }, + { + "epoch": 0.872063248191537, + "grad_norm": 1.0094337463378906, + "learning_rate": 1.1531465561440174e-05, + "loss": 0.5441, + "num_input_tokens_seen": 41464208, + "step": 27245 + }, + { + "epoch": 0.8722232891620255, + "grad_norm": 1.3868216276168823, + "learning_rate": 1.1523195658465605e-05, + "loss": 0.4483, + "num_input_tokens_seen": 41471408, + "step": 27250 + }, + { + "epoch": 0.8723833301325139, + "grad_norm": 0.8922432065010071, + "learning_rate": 1.1514927833783618e-05, + "loss": 0.5867, + "num_input_tokens_seen": 41478640, + "step": 27255 + }, + { + "epoch": 0.8725433711030024, + "grad_norm": 0.6532539129257202, + "learning_rate": 1.150666208866922e-05, + "loss": 0.522, + "num_input_tokens_seen": 41486128, + "step": 27260 + }, + { + "epoch": 0.8727034120734908, + "grad_norm": 0.6947629451751709, + "learning_rate": 1.1498398424397106e-05, + "loss": 0.3825, + "num_input_tokens_seen": 41493408, + "step": 27265 + }, + { + "epoch": 0.8728634530439793, + "grad_norm": 1.4113081693649292, + "learning_rate": 1.1490136842241628e-05, + "loss": 0.6303, + "num_input_tokens_seen": 41500688, + "step": 27270 + }, + { + "epoch": 0.8730234940144677, + "grad_norm": 1.0180801153182983, + "learning_rate": 1.1481877343476813e-05, + "loss": 0.5962, + "num_input_tokens_seen": 41508336, + "step": 27275 + }, + { + "epoch": 0.8731835349849562, + "grad_norm": 0.49017781019210815, + "learning_rate": 1.14736199293764e-05, + "loss": 0.4881, + "num_input_tokens_seen": 41515920, + "step": 27280 + }, + { + "epoch": 0.8733435759554445, + "grad_norm": 0.6636616587638855, + "learning_rate": 1.1465364601213771e-05, + "loss": 0.5269, + "num_input_tokens_seen": 41523424, + "step": 27285 + }, + { + "epoch": 0.873503616925933, + "grad_norm": 0.5943916440010071, + "learning_rate": 1.1457111360262012e-05, + "loss": 0.5186, + "num_input_tokens_seen": 41530704, + "step": 27290 + }, + { + "epoch": 0.8736636578964215, + "grad_norm": 1.0380101203918457, + "learning_rate": 1.1448860207793869e-05, + "loss": 0.6157, + "num_input_tokens_seen": 41538144, + "step": 27295 + }, + { + "epoch": 0.8738236988669099, + "grad_norm": 0.3900953233242035, + "learning_rate": 1.144061114508177e-05, + "loss": 0.4206, + "num_input_tokens_seen": 41545680, + "step": 27300 + }, + { + "epoch": 0.8739837398373984, + "grad_norm": 0.7779468297958374, + "learning_rate": 1.1432364173397842e-05, + "loss": 0.5723, + "num_input_tokens_seen": 41553328, + "step": 27305 + }, + { + "epoch": 0.8741437808078868, + "grad_norm": 0.6879081726074219, + "learning_rate": 1.1424119294013852e-05, + "loss": 0.4388, + "num_input_tokens_seen": 41560768, + "step": 27310 + }, + { + "epoch": 0.8743038217783753, + "grad_norm": 0.7645517587661743, + "learning_rate": 1.1415876508201279e-05, + "loss": 0.4053, + "num_input_tokens_seen": 41568400, + "step": 27315 + }, + { + "epoch": 0.8744638627488637, + "grad_norm": 0.5907493829727173, + "learning_rate": 1.140763581723125e-05, + "loss": 0.4142, + "num_input_tokens_seen": 41576112, + "step": 27320 + }, + { + "epoch": 0.8746239037193522, + "grad_norm": 0.5828712582588196, + "learning_rate": 1.1399397222374588e-05, + "loss": 0.4504, + "num_input_tokens_seen": 41584080, + "step": 27325 + }, + { + "epoch": 0.8747839446898406, + "grad_norm": 0.761816143989563, + "learning_rate": 1.1391160724901804e-05, + "loss": 0.4234, + "num_input_tokens_seen": 41591376, + "step": 27330 + }, + { + "epoch": 0.874943985660329, + "grad_norm": 0.5583939552307129, + "learning_rate": 1.138292632608304e-05, + "loss": 0.3985, + "num_input_tokens_seen": 41598704, + "step": 27335 + }, + { + "epoch": 0.8751040266308175, + "grad_norm": 0.29322218894958496, + "learning_rate": 1.1374694027188174e-05, + "loss": 0.3859, + "num_input_tokens_seen": 41606320, + "step": 27340 + }, + { + "epoch": 0.8752640676013059, + "grad_norm": 1.7359230518341064, + "learning_rate": 1.1366463829486711e-05, + "loss": 0.5077, + "num_input_tokens_seen": 41613584, + "step": 27345 + }, + { + "epoch": 0.8754241085717944, + "grad_norm": 0.7047374248504639, + "learning_rate": 1.1358235734247849e-05, + "loss": 0.4044, + "num_input_tokens_seen": 41621024, + "step": 27350 + }, + { + "epoch": 0.8755841495422828, + "grad_norm": 0.9689246416091919, + "learning_rate": 1.1350009742740478e-05, + "loss": 0.4565, + "num_input_tokens_seen": 41628496, + "step": 27355 + }, + { + "epoch": 0.8757441905127713, + "grad_norm": 0.8926793336868286, + "learning_rate": 1.134178585623313e-05, + "loss": 0.3425, + "num_input_tokens_seen": 41635888, + "step": 27360 + }, + { + "epoch": 0.8759042314832597, + "grad_norm": 1.4333552122116089, + "learning_rate": 1.1333564075994047e-05, + "loss": 0.424, + "num_input_tokens_seen": 41643280, + "step": 27365 + }, + { + "epoch": 0.8760642724537482, + "grad_norm": 0.7334247827529907, + "learning_rate": 1.1325344403291133e-05, + "loss": 0.5458, + "num_input_tokens_seen": 41650720, + "step": 27370 + }, + { + "epoch": 0.8762243134242366, + "grad_norm": 0.5825796127319336, + "learning_rate": 1.1317126839391951e-05, + "loss": 0.4622, + "num_input_tokens_seen": 41657856, + "step": 27375 + }, + { + "epoch": 0.8763843543947251, + "grad_norm": 1.0992473363876343, + "learning_rate": 1.1308911385563766e-05, + "loss": 0.5259, + "num_input_tokens_seen": 41665648, + "step": 27380 + }, + { + "epoch": 0.8765443953652134, + "grad_norm": 0.6367810964584351, + "learning_rate": 1.1300698043073494e-05, + "loss": 0.4668, + "num_input_tokens_seen": 41673152, + "step": 27385 + }, + { + "epoch": 0.8767044363357019, + "grad_norm": 0.5711419582366943, + "learning_rate": 1.1292486813187736e-05, + "loss": 0.5245, + "num_input_tokens_seen": 41680544, + "step": 27390 + }, + { + "epoch": 0.8768644773061904, + "grad_norm": 1.2983318567276, + "learning_rate": 1.1284277697172782e-05, + "loss": 0.5728, + "num_input_tokens_seen": 41687920, + "step": 27395 + }, + { + "epoch": 0.8770245182766788, + "grad_norm": 1.0201544761657715, + "learning_rate": 1.127607069629456e-05, + "loss": 0.5198, + "num_input_tokens_seen": 41695584, + "step": 27400 + }, + { + "epoch": 0.8770245182766788, + "eval_loss": 0.4988884925842285, + "eval_runtime": 559.0333, + "eval_samples_per_second": 24.839, + "eval_steps_per_second": 12.42, + "num_input_tokens_seen": 41695584, + "step": 27400 + }, + { + "epoch": 0.8771845592471673, + "grad_norm": 0.6748443245887756, + "learning_rate": 1.1267865811818701e-05, + "loss": 0.5748, + "num_input_tokens_seen": 41703488, + "step": 27405 + }, + { + "epoch": 0.8773446002176557, + "grad_norm": 1.153497576713562, + "learning_rate": 1.1259663045010513e-05, + "loss": 0.7128, + "num_input_tokens_seen": 41711072, + "step": 27410 + }, + { + "epoch": 0.8775046411881442, + "grad_norm": 0.5927464365959167, + "learning_rate": 1.1251462397134957e-05, + "loss": 0.3765, + "num_input_tokens_seen": 41718800, + "step": 27415 + }, + { + "epoch": 0.8776646821586326, + "grad_norm": 0.9150522351264954, + "learning_rate": 1.1243263869456664e-05, + "loss": 0.4781, + "num_input_tokens_seen": 41726256, + "step": 27420 + }, + { + "epoch": 0.8778247231291211, + "grad_norm": 0.911018431186676, + "learning_rate": 1.1235067463239967e-05, + "loss": 0.4389, + "num_input_tokens_seen": 41733856, + "step": 27425 + }, + { + "epoch": 0.8779847640996095, + "grad_norm": 1.0295244455337524, + "learning_rate": 1.122687317974884e-05, + "loss": 0.6784, + "num_input_tokens_seen": 41741728, + "step": 27430 + }, + { + "epoch": 0.878144805070098, + "grad_norm": 0.9785176515579224, + "learning_rate": 1.1218681020246963e-05, + "loss": 0.5176, + "num_input_tokens_seen": 41749568, + "step": 27435 + }, + { + "epoch": 0.8783048460405863, + "grad_norm": 0.4178225100040436, + "learning_rate": 1.1210490985997652e-05, + "loss": 0.369, + "num_input_tokens_seen": 41757536, + "step": 27440 + }, + { + "epoch": 0.8784648870110748, + "grad_norm": 0.5311376452445984, + "learning_rate": 1.1202303078263917e-05, + "loss": 0.5658, + "num_input_tokens_seen": 41765200, + "step": 27445 + }, + { + "epoch": 0.8786249279815633, + "grad_norm": 0.5858943462371826, + "learning_rate": 1.1194117298308451e-05, + "loss": 0.4397, + "num_input_tokens_seen": 41772720, + "step": 27450 + }, + { + "epoch": 0.8787849689520517, + "grad_norm": 0.9536119699478149, + "learning_rate": 1.1185933647393585e-05, + "loss": 0.5512, + "num_input_tokens_seen": 41780224, + "step": 27455 + }, + { + "epoch": 0.8789450099225402, + "grad_norm": 0.46849191188812256, + "learning_rate": 1.1177752126781354e-05, + "loss": 0.5237, + "num_input_tokens_seen": 41787712, + "step": 27460 + }, + { + "epoch": 0.8791050508930286, + "grad_norm": 1.791765809059143, + "learning_rate": 1.1169572737733441e-05, + "loss": 0.6831, + "num_input_tokens_seen": 41795504, + "step": 27465 + }, + { + "epoch": 0.8792650918635171, + "grad_norm": 0.4927469789981842, + "learning_rate": 1.1161395481511216e-05, + "loss": 0.4145, + "num_input_tokens_seen": 41802976, + "step": 27470 + }, + { + "epoch": 0.8794251328340055, + "grad_norm": 0.7217821478843689, + "learning_rate": 1.1153220359375722e-05, + "loss": 0.5208, + "num_input_tokens_seen": 41810704, + "step": 27475 + }, + { + "epoch": 0.879585173804494, + "grad_norm": 0.7464678287506104, + "learning_rate": 1.114504737258765e-05, + "loss": 0.5828, + "num_input_tokens_seen": 41818592, + "step": 27480 + }, + { + "epoch": 0.8797452147749824, + "grad_norm": 1.2242157459259033, + "learning_rate": 1.1136876522407393e-05, + "loss": 0.4507, + "num_input_tokens_seen": 41826064, + "step": 27485 + }, + { + "epoch": 0.8799052557454708, + "grad_norm": 0.6948726773262024, + "learning_rate": 1.1128707810094985e-05, + "loss": 0.4508, + "num_input_tokens_seen": 41833312, + "step": 27490 + }, + { + "epoch": 0.8800652967159592, + "grad_norm": 0.36708566546440125, + "learning_rate": 1.1120541236910157e-05, + "loss": 0.5344, + "num_input_tokens_seen": 41840480, + "step": 27495 + }, + { + "epoch": 0.8802253376864477, + "grad_norm": 0.8070300817489624, + "learning_rate": 1.111237680411229e-05, + "loss": 0.4097, + "num_input_tokens_seen": 41847728, + "step": 27500 + }, + { + "epoch": 0.8803853786569362, + "grad_norm": 0.6483215689659119, + "learning_rate": 1.1104214512960433e-05, + "loss": 0.4623, + "num_input_tokens_seen": 41855680, + "step": 27505 + }, + { + "epoch": 0.8805454196274246, + "grad_norm": 0.6791028380393982, + "learning_rate": 1.1096054364713327e-05, + "loss": 0.5175, + "num_input_tokens_seen": 41863088, + "step": 27510 + }, + { + "epoch": 0.8807054605979131, + "grad_norm": 0.48254188895225525, + "learning_rate": 1.1087896360629371e-05, + "loss": 0.4117, + "num_input_tokens_seen": 41870864, + "step": 27515 + }, + { + "epoch": 0.8808655015684015, + "grad_norm": 0.732670783996582, + "learning_rate": 1.107974050196662e-05, + "loss": 0.4045, + "num_input_tokens_seen": 41878272, + "step": 27520 + }, + { + "epoch": 0.88102554253889, + "grad_norm": 0.6678426265716553, + "learning_rate": 1.1071586789982816e-05, + "loss": 0.497, + "num_input_tokens_seen": 41885792, + "step": 27525 + }, + { + "epoch": 0.8811855835093784, + "grad_norm": 1.0760811567306519, + "learning_rate": 1.1063435225935373e-05, + "loss": 0.6254, + "num_input_tokens_seen": 41893584, + "step": 27530 + }, + { + "epoch": 0.8813456244798669, + "grad_norm": 0.5404260754585266, + "learning_rate": 1.1055285811081348e-05, + "loss": 0.4528, + "num_input_tokens_seen": 41901104, + "step": 27535 + }, + { + "epoch": 0.8815056654503552, + "grad_norm": 0.6865019202232361, + "learning_rate": 1.1047138546677499e-05, + "loss": 0.4752, + "num_input_tokens_seen": 41909952, + "step": 27540 + }, + { + "epoch": 0.8816657064208437, + "grad_norm": 0.9503259062767029, + "learning_rate": 1.1038993433980219e-05, + "loss": 0.5557, + "num_input_tokens_seen": 41917488, + "step": 27545 + }, + { + "epoch": 0.8818257473913321, + "grad_norm": 0.9963448643684387, + "learning_rate": 1.1030850474245597e-05, + "loss": 0.646, + "num_input_tokens_seen": 41924624, + "step": 27550 + }, + { + "epoch": 0.8819857883618206, + "grad_norm": 0.539533257484436, + "learning_rate": 1.102270966872939e-05, + "loss": 0.5152, + "num_input_tokens_seen": 41932512, + "step": 27555 + }, + { + "epoch": 0.8821458293323091, + "grad_norm": 1.0207136869430542, + "learning_rate": 1.1014571018687e-05, + "loss": 0.4398, + "num_input_tokens_seen": 41940064, + "step": 27560 + }, + { + "epoch": 0.8823058703027975, + "grad_norm": 0.785761296749115, + "learning_rate": 1.1006434525373502e-05, + "loss": 0.5282, + "num_input_tokens_seen": 41948080, + "step": 27565 + }, + { + "epoch": 0.882465911273286, + "grad_norm": 1.736107349395752, + "learning_rate": 1.0998300190043664e-05, + "loss": 0.6013, + "num_input_tokens_seen": 41955744, + "step": 27570 + }, + { + "epoch": 0.8826259522437744, + "grad_norm": 0.9296143651008606, + "learning_rate": 1.0990168013951882e-05, + "loss": 0.4128, + "num_input_tokens_seen": 41963392, + "step": 27575 + }, + { + "epoch": 0.8827859932142629, + "grad_norm": 0.7032085657119751, + "learning_rate": 1.0982037998352263e-05, + "loss": 0.4947, + "num_input_tokens_seen": 41970704, + "step": 27580 + }, + { + "epoch": 0.8829460341847513, + "grad_norm": 1.4356889724731445, + "learning_rate": 1.0973910144498534e-05, + "loss": 0.5852, + "num_input_tokens_seen": 41978240, + "step": 27585 + }, + { + "epoch": 0.8831060751552398, + "grad_norm": 1.3632229566574097, + "learning_rate": 1.0965784453644123e-05, + "loss": 0.536, + "num_input_tokens_seen": 41986128, + "step": 27590 + }, + { + "epoch": 0.8832661161257281, + "grad_norm": 1.616295337677002, + "learning_rate": 1.0957660927042127e-05, + "loss": 0.4933, + "num_input_tokens_seen": 41993760, + "step": 27595 + }, + { + "epoch": 0.8834261570962166, + "grad_norm": 1.273374080657959, + "learning_rate": 1.094953956594527e-05, + "loss": 0.525, + "num_input_tokens_seen": 42001344, + "step": 27600 + }, + { + "epoch": 0.8834261570962166, + "eval_loss": 0.49849212169647217, + "eval_runtime": 558.8088, + "eval_samples_per_second": 24.849, + "eval_steps_per_second": 12.425, + "num_input_tokens_seen": 42001344, + "step": 27600 + }, + { + "epoch": 0.8835861980667051, + "grad_norm": 0.9656725525856018, + "learning_rate": 1.0941420371605981e-05, + "loss": 0.5404, + "num_input_tokens_seen": 42009120, + "step": 27605 + }, + { + "epoch": 0.8837462390371935, + "grad_norm": 0.5079985857009888, + "learning_rate": 1.0933303345276354e-05, + "loss": 0.4973, + "num_input_tokens_seen": 42016816, + "step": 27610 + }, + { + "epoch": 0.883906280007682, + "grad_norm": 1.0246691703796387, + "learning_rate": 1.0925188488208112e-05, + "loss": 0.6163, + "num_input_tokens_seen": 42024352, + "step": 27615 + }, + { + "epoch": 0.8840663209781704, + "grad_norm": 0.8192071914672852, + "learning_rate": 1.0917075801652694e-05, + "loss": 0.495, + "num_input_tokens_seen": 42032160, + "step": 27620 + }, + { + "epoch": 0.8842263619486589, + "grad_norm": 0.6036573052406311, + "learning_rate": 1.0908965286861151e-05, + "loss": 0.4634, + "num_input_tokens_seen": 42039136, + "step": 27625 + }, + { + "epoch": 0.8843864029191473, + "grad_norm": 0.4229575991630554, + "learning_rate": 1.090085694508425e-05, + "loss": 0.3755, + "num_input_tokens_seen": 42046736, + "step": 27630 + }, + { + "epoch": 0.8845464438896358, + "grad_norm": 1.118525505065918, + "learning_rate": 1.089275077757238e-05, + "loss": 0.4804, + "num_input_tokens_seen": 42053600, + "step": 27635 + }, + { + "epoch": 0.8847064848601242, + "grad_norm": 1.2246018648147583, + "learning_rate": 1.0884646785575633e-05, + "loss": 0.5275, + "num_input_tokens_seen": 42060944, + "step": 27640 + }, + { + "epoch": 0.8848665258306126, + "grad_norm": 1.1679214239120483, + "learning_rate": 1.0876544970343728e-05, + "loss": 0.3959, + "num_input_tokens_seen": 42068976, + "step": 27645 + }, + { + "epoch": 0.885026566801101, + "grad_norm": 0.8951038122177124, + "learning_rate": 1.0868445333126082e-05, + "loss": 0.5684, + "num_input_tokens_seen": 42076448, + "step": 27650 + }, + { + "epoch": 0.8851866077715895, + "grad_norm": 0.6355040669441223, + "learning_rate": 1.0860347875171745e-05, + "loss": 0.4429, + "num_input_tokens_seen": 42084048, + "step": 27655 + }, + { + "epoch": 0.885346648742078, + "grad_norm": 0.9590655565261841, + "learning_rate": 1.0852252597729465e-05, + "loss": 0.3639, + "num_input_tokens_seen": 42091536, + "step": 27660 + }, + { + "epoch": 0.8855066897125664, + "grad_norm": 1.006018042564392, + "learning_rate": 1.0844159502047615e-05, + "loss": 0.3674, + "num_input_tokens_seen": 42098960, + "step": 27665 + }, + { + "epoch": 0.8856667306830549, + "grad_norm": 0.8824678659439087, + "learning_rate": 1.0836068589374265e-05, + "loss": 0.5693, + "num_input_tokens_seen": 42107008, + "step": 27670 + }, + { + "epoch": 0.8858267716535433, + "grad_norm": 0.569274365901947, + "learning_rate": 1.0827979860957144e-05, + "loss": 0.5515, + "num_input_tokens_seen": 42114992, + "step": 27675 + }, + { + "epoch": 0.8859868126240318, + "grad_norm": 1.2833584547042847, + "learning_rate": 1.0819893318043615e-05, + "loss": 0.5637, + "num_input_tokens_seen": 42122800, + "step": 27680 + }, + { + "epoch": 0.8861468535945202, + "grad_norm": 0.7217593789100647, + "learning_rate": 1.0811808961880734e-05, + "loss": 0.4517, + "num_input_tokens_seen": 42130736, + "step": 27685 + }, + { + "epoch": 0.8863068945650087, + "grad_norm": 0.5068201422691345, + "learning_rate": 1.080372679371522e-05, + "loss": 0.3842, + "num_input_tokens_seen": 42138432, + "step": 27690 + }, + { + "epoch": 0.886466935535497, + "grad_norm": 1.0081958770751953, + "learning_rate": 1.0795646814793428e-05, + "loss": 0.4398, + "num_input_tokens_seen": 42145792, + "step": 27695 + }, + { + "epoch": 0.8866269765059855, + "grad_norm": 1.0205167531967163, + "learning_rate": 1.078756902636141e-05, + "loss": 0.4924, + "num_input_tokens_seen": 42153760, + "step": 27700 + }, + { + "epoch": 0.8867870174764739, + "grad_norm": 0.8706104159355164, + "learning_rate": 1.077949342966485e-05, + "loss": 0.481, + "num_input_tokens_seen": 42161520, + "step": 27705 + }, + { + "epoch": 0.8869470584469624, + "grad_norm": 1.307880163192749, + "learning_rate": 1.0771420025949103e-05, + "loss": 0.558, + "num_input_tokens_seen": 42170064, + "step": 27710 + }, + { + "epoch": 0.8871070994174509, + "grad_norm": 0.6447519063949585, + "learning_rate": 1.0763348816459204e-05, + "loss": 0.7217, + "num_input_tokens_seen": 42177600, + "step": 27715 + }, + { + "epoch": 0.8872671403879393, + "grad_norm": 1.00151526927948, + "learning_rate": 1.0755279802439816e-05, + "loss": 0.6883, + "num_input_tokens_seen": 42184960, + "step": 27720 + }, + { + "epoch": 0.8874271813584278, + "grad_norm": 0.47184526920318604, + "learning_rate": 1.0747212985135293e-05, + "loss": 0.4297, + "num_input_tokens_seen": 42192464, + "step": 27725 + }, + { + "epoch": 0.8875872223289162, + "grad_norm": 0.7759926319122314, + "learning_rate": 1.073914836578965e-05, + "loss": 0.4331, + "num_input_tokens_seen": 42200016, + "step": 27730 + }, + { + "epoch": 0.8877472632994047, + "grad_norm": 0.7799490094184875, + "learning_rate": 1.0731085945646529e-05, + "loss": 0.5512, + "num_input_tokens_seen": 42207696, + "step": 27735 + }, + { + "epoch": 0.8879073042698931, + "grad_norm": 0.44779860973358154, + "learning_rate": 1.0723025725949285e-05, + "loss": 0.3459, + "num_input_tokens_seen": 42215440, + "step": 27740 + }, + { + "epoch": 0.8880673452403816, + "grad_norm": 0.285317063331604, + "learning_rate": 1.0714967707940875e-05, + "loss": 0.3489, + "num_input_tokens_seen": 42223472, + "step": 27745 + }, + { + "epoch": 0.8882273862108699, + "grad_norm": 1.4369421005249023, + "learning_rate": 1.0706911892863963e-05, + "loss": 0.4195, + "num_input_tokens_seen": 42230976, + "step": 27750 + }, + { + "epoch": 0.8883874271813584, + "grad_norm": 1.4698125123977661, + "learning_rate": 1.0698858281960866e-05, + "loss": 0.5013, + "num_input_tokens_seen": 42238448, + "step": 27755 + }, + { + "epoch": 0.8885474681518468, + "grad_norm": 1.0616260766983032, + "learning_rate": 1.069080687647353e-05, + "loss": 0.4397, + "num_input_tokens_seen": 42245680, + "step": 27760 + }, + { + "epoch": 0.8887075091223353, + "grad_norm": 0.506776750087738, + "learning_rate": 1.0682757677643596e-05, + "loss": 0.533, + "num_input_tokens_seen": 42252912, + "step": 27765 + }, + { + "epoch": 0.8888675500928238, + "grad_norm": 1.071726679801941, + "learning_rate": 1.0674710686712359e-05, + "loss": 0.5228, + "num_input_tokens_seen": 42260512, + "step": 27770 + }, + { + "epoch": 0.8890275910633122, + "grad_norm": 0.7070693373680115, + "learning_rate": 1.0666665904920756e-05, + "loss": 0.381, + "num_input_tokens_seen": 42267536, + "step": 27775 + }, + { + "epoch": 0.8891876320338007, + "grad_norm": 1.3110524415969849, + "learning_rate": 1.0658623333509385e-05, + "loss": 0.5053, + "num_input_tokens_seen": 42274656, + "step": 27780 + }, + { + "epoch": 0.8893476730042891, + "grad_norm": 0.7562497854232788, + "learning_rate": 1.0650582973718532e-05, + "loss": 0.3562, + "num_input_tokens_seen": 42282176, + "step": 27785 + }, + { + "epoch": 0.8895077139747776, + "grad_norm": 0.6188444495201111, + "learning_rate": 1.0642544826788098e-05, + "loss": 0.5041, + "num_input_tokens_seen": 42290128, + "step": 27790 + }, + { + "epoch": 0.889667754945266, + "grad_norm": 0.5803350806236267, + "learning_rate": 1.063450889395769e-05, + "loss": 0.4835, + "num_input_tokens_seen": 42297824, + "step": 27795 + }, + { + "epoch": 0.8898277959157544, + "grad_norm": 0.8439061045646667, + "learning_rate": 1.062647517646653e-05, + "loss": 0.3574, + "num_input_tokens_seen": 42305104, + "step": 27800 + }, + { + "epoch": 0.8898277959157544, + "eval_loss": 0.4983082115650177, + "eval_runtime": 558.3038, + "eval_samples_per_second": 24.872, + "eval_steps_per_second": 12.436, + "num_input_tokens_seen": 42305104, + "step": 27800 + }, + { + "epoch": 0.8899878368862428, + "grad_norm": 0.643173336982727, + "learning_rate": 1.0618443675553527e-05, + "loss": 0.397, + "num_input_tokens_seen": 42312336, + "step": 27805 + }, + { + "epoch": 0.8901478778567313, + "grad_norm": 0.6433124542236328, + "learning_rate": 1.0610414392457247e-05, + "loss": 0.6726, + "num_input_tokens_seen": 42319600, + "step": 27810 + }, + { + "epoch": 0.8903079188272197, + "grad_norm": 1.312762975692749, + "learning_rate": 1.0602387328415888e-05, + "loss": 0.488, + "num_input_tokens_seen": 42327536, + "step": 27815 + }, + { + "epoch": 0.8904679597977082, + "grad_norm": 1.0853276252746582, + "learning_rate": 1.0594362484667347e-05, + "loss": 0.6623, + "num_input_tokens_seen": 42335024, + "step": 27820 + }, + { + "epoch": 0.8906280007681967, + "grad_norm": 0.6106882095336914, + "learning_rate": 1.0586339862449132e-05, + "loss": 0.5211, + "num_input_tokens_seen": 42342528, + "step": 27825 + }, + { + "epoch": 0.8907880417386851, + "grad_norm": 0.5210503935813904, + "learning_rate": 1.0578319462998445e-05, + "loss": 0.4752, + "num_input_tokens_seen": 42350096, + "step": 27830 + }, + { + "epoch": 0.8909480827091736, + "grad_norm": 0.931756317615509, + "learning_rate": 1.057030128755214e-05, + "loss": 0.5467, + "num_input_tokens_seen": 42357632, + "step": 27835 + }, + { + "epoch": 0.891108123679662, + "grad_norm": 1.7157357931137085, + "learning_rate": 1.0562285337346703e-05, + "loss": 0.8364, + "num_input_tokens_seen": 42365024, + "step": 27840 + }, + { + "epoch": 0.8912681646501505, + "grad_norm": 0.5955343842506409, + "learning_rate": 1.0554271613618308e-05, + "loss": 0.563, + "num_input_tokens_seen": 42372816, + "step": 27845 + }, + { + "epoch": 0.8914282056206388, + "grad_norm": 0.7119776606559753, + "learning_rate": 1.054626011760276e-05, + "loss": 0.5746, + "num_input_tokens_seen": 42380992, + "step": 27850 + }, + { + "epoch": 0.8915882465911273, + "grad_norm": 0.7841297388076782, + "learning_rate": 1.0538250850535549e-05, + "loss": 0.5291, + "num_input_tokens_seen": 42388064, + "step": 27855 + }, + { + "epoch": 0.8917482875616157, + "grad_norm": 0.5735011696815491, + "learning_rate": 1.0530243813651794e-05, + "loss": 0.4692, + "num_input_tokens_seen": 42395632, + "step": 27860 + }, + { + "epoch": 0.8919083285321042, + "grad_norm": 0.7517073750495911, + "learning_rate": 1.0522239008186271e-05, + "loss": 0.5185, + "num_input_tokens_seen": 42403184, + "step": 27865 + }, + { + "epoch": 0.8920683695025927, + "grad_norm": 0.5148018002510071, + "learning_rate": 1.0514236435373434e-05, + "loss": 0.4305, + "num_input_tokens_seen": 42410320, + "step": 27870 + }, + { + "epoch": 0.8922284104730811, + "grad_norm": 0.6916555762290955, + "learning_rate": 1.0506236096447386e-05, + "loss": 0.3479, + "num_input_tokens_seen": 42417744, + "step": 27875 + }, + { + "epoch": 0.8923884514435696, + "grad_norm": 0.45767825841903687, + "learning_rate": 1.049823799264186e-05, + "loss": 0.357, + "num_input_tokens_seen": 42425376, + "step": 27880 + }, + { + "epoch": 0.892548492414058, + "grad_norm": 1.6754827499389648, + "learning_rate": 1.049024212519028e-05, + "loss": 0.4469, + "num_input_tokens_seen": 42432800, + "step": 27885 + }, + { + "epoch": 0.8927085333845465, + "grad_norm": 1.0008858442306519, + "learning_rate": 1.0482248495325713e-05, + "loss": 0.4806, + "num_input_tokens_seen": 42440528, + "step": 27890 + }, + { + "epoch": 0.8928685743550349, + "grad_norm": 1.9618710279464722, + "learning_rate": 1.047425710428086e-05, + "loss": 0.5625, + "num_input_tokens_seen": 42447952, + "step": 27895 + }, + { + "epoch": 0.8930286153255234, + "grad_norm": 1.1132502555847168, + "learning_rate": 1.0466267953288114e-05, + "loss": 0.664, + "num_input_tokens_seen": 42455520, + "step": 27900 + }, + { + "epoch": 0.8931886562960117, + "grad_norm": 0.7841728329658508, + "learning_rate": 1.0458281043579482e-05, + "loss": 0.5695, + "num_input_tokens_seen": 42463232, + "step": 27905 + }, + { + "epoch": 0.8933486972665002, + "grad_norm": 1.3177887201309204, + "learning_rate": 1.0450296376386657e-05, + "loss": 0.7288, + "num_input_tokens_seen": 42470800, + "step": 27910 + }, + { + "epoch": 0.8935087382369886, + "grad_norm": 0.6882438063621521, + "learning_rate": 1.044231395294098e-05, + "loss": 0.4073, + "num_input_tokens_seen": 42478352, + "step": 27915 + }, + { + "epoch": 0.8936687792074771, + "grad_norm": 0.6104755401611328, + "learning_rate": 1.0434333774473435e-05, + "loss": 0.3365, + "num_input_tokens_seen": 42486336, + "step": 27920 + }, + { + "epoch": 0.8938288201779656, + "grad_norm": 0.590229868888855, + "learning_rate": 1.0426355842214657e-05, + "loss": 0.4375, + "num_input_tokens_seen": 42493376, + "step": 27925 + }, + { + "epoch": 0.893988861148454, + "grad_norm": 0.681689977645874, + "learning_rate": 1.0418380157394963e-05, + "loss": 0.4807, + "num_input_tokens_seen": 42501392, + "step": 27930 + }, + { + "epoch": 0.8941489021189425, + "grad_norm": 0.9751036763191223, + "learning_rate": 1.0410406721244281e-05, + "loss": 0.6283, + "num_input_tokens_seen": 42509120, + "step": 27935 + }, + { + "epoch": 0.8943089430894309, + "grad_norm": 0.9982411861419678, + "learning_rate": 1.0402435534992238e-05, + "loss": 0.5353, + "num_input_tokens_seen": 42516848, + "step": 27940 + }, + { + "epoch": 0.8944689840599194, + "grad_norm": 0.5678791999816895, + "learning_rate": 1.0394466599868071e-05, + "loss": 0.5385, + "num_input_tokens_seen": 42524576, + "step": 27945 + }, + { + "epoch": 0.8946290250304078, + "grad_norm": 0.77922123670578, + "learning_rate": 1.0386499917100697e-05, + "loss": 0.475, + "num_input_tokens_seen": 42532240, + "step": 27950 + }, + { + "epoch": 0.8947890660008962, + "grad_norm": 0.7105193734169006, + "learning_rate": 1.0378535487918692e-05, + "loss": 0.475, + "num_input_tokens_seen": 42539584, + "step": 27955 + }, + { + "epoch": 0.8949491069713846, + "grad_norm": 0.4326123297214508, + "learning_rate": 1.037057331355025e-05, + "loss": 0.4125, + "num_input_tokens_seen": 42547136, + "step": 27960 + }, + { + "epoch": 0.8951091479418731, + "grad_norm": 0.8049302697181702, + "learning_rate": 1.0362613395223247e-05, + "loss": 0.5986, + "num_input_tokens_seen": 42554592, + "step": 27965 + }, + { + "epoch": 0.8952691889123615, + "grad_norm": 0.7097154259681702, + "learning_rate": 1.0354655734165212e-05, + "loss": 0.4804, + "num_input_tokens_seen": 42562496, + "step": 27970 + }, + { + "epoch": 0.89542922988285, + "grad_norm": 0.6025387644767761, + "learning_rate": 1.03467003316033e-05, + "loss": 0.3537, + "num_input_tokens_seen": 42569520, + "step": 27975 + }, + { + "epoch": 0.8955892708533385, + "grad_norm": 0.8657096028327942, + "learning_rate": 1.033874718876435e-05, + "loss": 0.5713, + "num_input_tokens_seen": 42577184, + "step": 27980 + }, + { + "epoch": 0.8957493118238269, + "grad_norm": 0.8576495051383972, + "learning_rate": 1.0330796306874818e-05, + "loss": 0.4778, + "num_input_tokens_seen": 42584464, + "step": 27985 + }, + { + "epoch": 0.8959093527943154, + "grad_norm": 0.9831789135932922, + "learning_rate": 1.032284768716085e-05, + "loss": 0.543, + "num_input_tokens_seen": 42592320, + "step": 27990 + }, + { + "epoch": 0.8960693937648038, + "grad_norm": 0.8943485021591187, + "learning_rate": 1.0314901330848206e-05, + "loss": 0.5379, + "num_input_tokens_seen": 42599728, + "step": 27995 + }, + { + "epoch": 0.8962294347352923, + "grad_norm": 0.4608163833618164, + "learning_rate": 1.030695723916233e-05, + "loss": 0.4493, + "num_input_tokens_seen": 42606944, + "step": 28000 + }, + { + "epoch": 0.8962294347352923, + "eval_loss": 0.4981251358985901, + "eval_runtime": 557.0614, + "eval_samples_per_second": 24.927, + "eval_steps_per_second": 12.464, + "num_input_tokens_seen": 42606944, + "step": 28000 + }, + { + "epoch": 0.8963894757057806, + "grad_norm": 0.7836037278175354, + "learning_rate": 1.0299015413328289e-05, + "loss": 0.5988, + "num_input_tokens_seen": 42614464, + "step": 28005 + }, + { + "epoch": 0.8965495166762691, + "grad_norm": 1.6583863496780396, + "learning_rate": 1.0291075854570809e-05, + "loss": 0.6508, + "num_input_tokens_seen": 42622112, + "step": 28010 + }, + { + "epoch": 0.8967095576467575, + "grad_norm": 0.9862765669822693, + "learning_rate": 1.0283138564114275e-05, + "loss": 0.5592, + "num_input_tokens_seen": 42629584, + "step": 28015 + }, + { + "epoch": 0.896869598617246, + "grad_norm": 0.49284687638282776, + "learning_rate": 1.027520354318273e-05, + "loss": 0.3633, + "num_input_tokens_seen": 42637152, + "step": 28020 + }, + { + "epoch": 0.8970296395877344, + "grad_norm": 0.9839022755622864, + "learning_rate": 1.0267270792999828e-05, + "loss": 0.4842, + "num_input_tokens_seen": 42644592, + "step": 28025 + }, + { + "epoch": 0.8971896805582229, + "grad_norm": 0.6928106546401978, + "learning_rate": 1.0259340314788919e-05, + "loss": 0.4957, + "num_input_tokens_seen": 42651648, + "step": 28030 + }, + { + "epoch": 0.8973497215287114, + "grad_norm": 0.7809290885925293, + "learning_rate": 1.0251412109772979e-05, + "loss": 0.5632, + "num_input_tokens_seen": 42659328, + "step": 28035 + }, + { + "epoch": 0.8975097624991998, + "grad_norm": 0.692915141582489, + "learning_rate": 1.0243486179174627e-05, + "loss": 0.4417, + "num_input_tokens_seen": 42666848, + "step": 28040 + }, + { + "epoch": 0.8976698034696883, + "grad_norm": 1.3253587484359741, + "learning_rate": 1.0235562524216158e-05, + "loss": 0.5858, + "num_input_tokens_seen": 42675088, + "step": 28045 + }, + { + "epoch": 0.8978298444401767, + "grad_norm": 0.9105634689331055, + "learning_rate": 1.022764114611948e-05, + "loss": 0.691, + "num_input_tokens_seen": 42682608, + "step": 28050 + }, + { + "epoch": 0.8979898854106652, + "grad_norm": 0.9426195025444031, + "learning_rate": 1.0219722046106178e-05, + "loss": 0.4187, + "num_input_tokens_seen": 42689664, + "step": 28055 + }, + { + "epoch": 0.8981499263811535, + "grad_norm": 0.3307902216911316, + "learning_rate": 1.0211805225397486e-05, + "loss": 0.3648, + "num_input_tokens_seen": 42697376, + "step": 28060 + }, + { + "epoch": 0.898309967351642, + "grad_norm": 1.2331981658935547, + "learning_rate": 1.020389068521426e-05, + "loss": 0.5897, + "num_input_tokens_seen": 42704592, + "step": 28065 + }, + { + "epoch": 0.8984700083221304, + "grad_norm": 0.3295568525791168, + "learning_rate": 1.0195978426777039e-05, + "loss": 0.442, + "num_input_tokens_seen": 42712944, + "step": 28070 + }, + { + "epoch": 0.8986300492926189, + "grad_norm": 1.3321688175201416, + "learning_rate": 1.0188068451305982e-05, + "loss": 0.4635, + "num_input_tokens_seen": 42720320, + "step": 28075 + }, + { + "epoch": 0.8987900902631074, + "grad_norm": 0.6482740044593811, + "learning_rate": 1.0180160760020902e-05, + "loss": 0.4733, + "num_input_tokens_seen": 42727712, + "step": 28080 + }, + { + "epoch": 0.8989501312335958, + "grad_norm": 0.837283194065094, + "learning_rate": 1.0172255354141278e-05, + "loss": 0.4916, + "num_input_tokens_seen": 42735152, + "step": 28085 + }, + { + "epoch": 0.8991101722040843, + "grad_norm": 1.4476815462112427, + "learning_rate": 1.0164352234886205e-05, + "loss": 0.5631, + "num_input_tokens_seen": 42742720, + "step": 28090 + }, + { + "epoch": 0.8992702131745727, + "grad_norm": 1.1523327827453613, + "learning_rate": 1.0156451403474454e-05, + "loss": 0.496, + "num_input_tokens_seen": 42750448, + "step": 28095 + }, + { + "epoch": 0.8994302541450612, + "grad_norm": 0.7095789909362793, + "learning_rate": 1.0148552861124443e-05, + "loss": 0.4136, + "num_input_tokens_seen": 42758608, + "step": 28100 + }, + { + "epoch": 0.8995902951155496, + "grad_norm": 0.9402602910995483, + "learning_rate": 1.0140656609054205e-05, + "loss": 0.4831, + "num_input_tokens_seen": 42766416, + "step": 28105 + }, + { + "epoch": 0.899750336086038, + "grad_norm": 0.6958675384521484, + "learning_rate": 1.0132762648481455e-05, + "loss": 0.5103, + "num_input_tokens_seen": 42774096, + "step": 28110 + }, + { + "epoch": 0.8999103770565264, + "grad_norm": 1.5490777492523193, + "learning_rate": 1.0124870980623543e-05, + "loss": 0.4519, + "num_input_tokens_seen": 42781872, + "step": 28115 + }, + { + "epoch": 0.9000704180270149, + "grad_norm": 0.7877143621444702, + "learning_rate": 1.0116981606697453e-05, + "loss": 0.5665, + "num_input_tokens_seen": 42789152, + "step": 28120 + }, + { + "epoch": 0.9002304589975033, + "grad_norm": 0.731499969959259, + "learning_rate": 1.0109094527919838e-05, + "loss": 0.4306, + "num_input_tokens_seen": 42796512, + "step": 28125 + }, + { + "epoch": 0.9003904999679918, + "grad_norm": 0.6618577837944031, + "learning_rate": 1.010120974550697e-05, + "loss": 0.4889, + "num_input_tokens_seen": 42804112, + "step": 28130 + }, + { + "epoch": 0.9005505409384803, + "grad_norm": 0.838086724281311, + "learning_rate": 1.0093327260674795e-05, + "loss": 0.4301, + "num_input_tokens_seen": 42811888, + "step": 28135 + }, + { + "epoch": 0.9007105819089687, + "grad_norm": 0.7467448115348816, + "learning_rate": 1.0085447074638878e-05, + "loss": 0.3298, + "num_input_tokens_seen": 42819200, + "step": 28140 + }, + { + "epoch": 0.9008706228794572, + "grad_norm": 0.46722611784935, + "learning_rate": 1.0077569188614461e-05, + "loss": 0.3621, + "num_input_tokens_seen": 42826592, + "step": 28145 + }, + { + "epoch": 0.9010306638499456, + "grad_norm": 1.6689828634262085, + "learning_rate": 1.0069693603816393e-05, + "loss": 0.5308, + "num_input_tokens_seen": 42833888, + "step": 28150 + }, + { + "epoch": 0.9011907048204341, + "grad_norm": 0.8771463632583618, + "learning_rate": 1.0061820321459204e-05, + "loss": 0.3744, + "num_input_tokens_seen": 42841536, + "step": 28155 + }, + { + "epoch": 0.9013507457909224, + "grad_norm": 0.4702250361442566, + "learning_rate": 1.0053949342757038e-05, + "loss": 0.4264, + "num_input_tokens_seen": 42848624, + "step": 28160 + }, + { + "epoch": 0.9015107867614109, + "grad_norm": 0.7413252592086792, + "learning_rate": 1.0046080668923717e-05, + "loss": 0.5227, + "num_input_tokens_seen": 42856384, + "step": 28165 + }, + { + "epoch": 0.9016708277318993, + "grad_norm": 1.706555724143982, + "learning_rate": 1.003821430117267e-05, + "loss": 0.5732, + "num_input_tokens_seen": 42864272, + "step": 28170 + }, + { + "epoch": 0.9018308687023878, + "grad_norm": 1.1060619354248047, + "learning_rate": 1.0030350240716999e-05, + "loss": 0.4926, + "num_input_tokens_seen": 42871728, + "step": 28175 + }, + { + "epoch": 0.9019909096728762, + "grad_norm": 0.7785037159919739, + "learning_rate": 1.0022488488769449e-05, + "loss": 0.5531, + "num_input_tokens_seen": 42879008, + "step": 28180 + }, + { + "epoch": 0.9021509506433647, + "grad_norm": 0.7396772503852844, + "learning_rate": 1.0014629046542387e-05, + "loss": 0.4488, + "num_input_tokens_seen": 42886816, + "step": 28185 + }, + { + "epoch": 0.9023109916138532, + "grad_norm": 1.4030331373214722, + "learning_rate": 1.0006771915247842e-05, + "loss": 0.4696, + "num_input_tokens_seen": 42894288, + "step": 28190 + }, + { + "epoch": 0.9024710325843416, + "grad_norm": 0.9262598156929016, + "learning_rate": 9.998917096097495e-06, + "loss": 0.3591, + "num_input_tokens_seen": 42901776, + "step": 28195 + }, + { + "epoch": 0.9026310735548301, + "grad_norm": 0.616855263710022, + "learning_rate": 9.991064590302638e-06, + "loss": 0.4451, + "num_input_tokens_seen": 42909232, + "step": 28200 + }, + { + "epoch": 0.9026310735548301, + "eval_loss": 0.4979422986507416, + "eval_runtime": 557.0326, + "eval_samples_per_second": 24.929, + "eval_steps_per_second": 12.464, + "num_input_tokens_seen": 42909232, + "step": 28200 + }, + { + "epoch": 0.9027911145253185, + "grad_norm": 0.36280685663223267, + "learning_rate": 9.983214399074241e-06, + "loss": 0.3959, + "num_input_tokens_seen": 42916560, + "step": 28205 + }, + { + "epoch": 0.902951155495807, + "grad_norm": 0.7948787808418274, + "learning_rate": 9.975366523622893e-06, + "loss": 0.4547, + "num_input_tokens_seen": 42924192, + "step": 28210 + }, + { + "epoch": 0.9031111964662953, + "grad_norm": 1.368090033531189, + "learning_rate": 9.967520965158841e-06, + "loss": 0.5575, + "num_input_tokens_seen": 42932672, + "step": 28215 + }, + { + "epoch": 0.9032712374367838, + "grad_norm": 0.6466758251190186, + "learning_rate": 9.95967772489197e-06, + "loss": 0.5307, + "num_input_tokens_seen": 42940304, + "step": 28220 + }, + { + "epoch": 0.9034312784072722, + "grad_norm": 1.0562571287155151, + "learning_rate": 9.951836804031794e-06, + "loss": 0.3348, + "num_input_tokens_seen": 42947536, + "step": 28225 + }, + { + "epoch": 0.9035913193777607, + "grad_norm": 0.37582817673683167, + "learning_rate": 9.943998203787489e-06, + "loss": 0.4281, + "num_input_tokens_seen": 42954944, + "step": 28230 + }, + { + "epoch": 0.9037513603482491, + "grad_norm": 0.4534319341182709, + "learning_rate": 9.936161925367874e-06, + "loss": 0.4296, + "num_input_tokens_seen": 42962256, + "step": 28235 + }, + { + "epoch": 0.9039114013187376, + "grad_norm": 1.7790216207504272, + "learning_rate": 9.928327969981386e-06, + "loss": 0.5504, + "num_input_tokens_seen": 42969904, + "step": 28240 + }, + { + "epoch": 0.9040714422892261, + "grad_norm": 1.091599941253662, + "learning_rate": 9.920496338836135e-06, + "loss": 0.3125, + "num_input_tokens_seen": 42977312, + "step": 28245 + }, + { + "epoch": 0.9042314832597145, + "grad_norm": 0.5425643920898438, + "learning_rate": 9.912667033139844e-06, + "loss": 0.3806, + "num_input_tokens_seen": 42984816, + "step": 28250 + }, + { + "epoch": 0.904391524230203, + "grad_norm": 0.3713809847831726, + "learning_rate": 9.904840054099893e-06, + "loss": 0.4612, + "num_input_tokens_seen": 42993328, + "step": 28255 + }, + { + "epoch": 0.9045515652006914, + "grad_norm": 0.7629862427711487, + "learning_rate": 9.897015402923312e-06, + "loss": 0.3193, + "num_input_tokens_seen": 43000480, + "step": 28260 + }, + { + "epoch": 0.9047116061711798, + "grad_norm": 0.6700573563575745, + "learning_rate": 9.889193080816744e-06, + "loss": 0.6358, + "num_input_tokens_seen": 43008672, + "step": 28265 + }, + { + "epoch": 0.9048716471416682, + "grad_norm": 0.6481502652168274, + "learning_rate": 9.881373088986498e-06, + "loss": 0.4569, + "num_input_tokens_seen": 43016416, + "step": 28270 + }, + { + "epoch": 0.9050316881121567, + "grad_norm": 0.6933776140213013, + "learning_rate": 9.873555428638523e-06, + "loss": 0.5164, + "num_input_tokens_seen": 43024016, + "step": 28275 + }, + { + "epoch": 0.9051917290826451, + "grad_norm": 1.1414134502410889, + "learning_rate": 9.865740100978383e-06, + "loss": 0.6405, + "num_input_tokens_seen": 43031568, + "step": 28280 + }, + { + "epoch": 0.9053517700531336, + "grad_norm": 0.6980267763137817, + "learning_rate": 9.857927107211315e-06, + "loss": 0.4164, + "num_input_tokens_seen": 43039152, + "step": 28285 + }, + { + "epoch": 0.905511811023622, + "grad_norm": 0.947486162185669, + "learning_rate": 9.850116448542177e-06, + "loss": 0.4656, + "num_input_tokens_seen": 43046416, + "step": 28290 + }, + { + "epoch": 0.9056718519941105, + "grad_norm": 1.5586092472076416, + "learning_rate": 9.842308126175457e-06, + "loss": 0.5351, + "num_input_tokens_seen": 43054560, + "step": 28295 + }, + { + "epoch": 0.905831892964599, + "grad_norm": 0.5633066892623901, + "learning_rate": 9.834502141315315e-06, + "loss": 0.3818, + "num_input_tokens_seen": 43062496, + "step": 28300 + }, + { + "epoch": 0.9059919339350874, + "grad_norm": 0.4880208373069763, + "learning_rate": 9.82669849516552e-06, + "loss": 0.3961, + "num_input_tokens_seen": 43069936, + "step": 28305 + }, + { + "epoch": 0.9061519749055759, + "grad_norm": 0.9988816380500793, + "learning_rate": 9.818897188929493e-06, + "loss": 0.6159, + "num_input_tokens_seen": 43077472, + "step": 28310 + }, + { + "epoch": 0.9063120158760642, + "grad_norm": 0.5867146253585815, + "learning_rate": 9.811098223810309e-06, + "loss": 0.453, + "num_input_tokens_seen": 43085280, + "step": 28315 + }, + { + "epoch": 0.9064720568465527, + "grad_norm": 1.1690436601638794, + "learning_rate": 9.803301601010641e-06, + "loss": 0.5707, + "num_input_tokens_seen": 43092624, + "step": 28320 + }, + { + "epoch": 0.9066320978170411, + "grad_norm": 0.8717557787895203, + "learning_rate": 9.795507321732853e-06, + "loss": 0.4335, + "num_input_tokens_seen": 43099952, + "step": 28325 + }, + { + "epoch": 0.9067921387875296, + "grad_norm": 0.7702511548995972, + "learning_rate": 9.787715387178898e-06, + "loss": 0.5045, + "num_input_tokens_seen": 43107280, + "step": 28330 + }, + { + "epoch": 0.906952179758018, + "grad_norm": 0.8780905604362488, + "learning_rate": 9.779925798550399e-06, + "loss": 0.5546, + "num_input_tokens_seen": 43114816, + "step": 28335 + }, + { + "epoch": 0.9071122207285065, + "grad_norm": 0.6644898056983948, + "learning_rate": 9.772138557048619e-06, + "loss": 0.4458, + "num_input_tokens_seen": 43122240, + "step": 28340 + }, + { + "epoch": 0.907272261698995, + "grad_norm": 0.28651106357574463, + "learning_rate": 9.764353663874426e-06, + "loss": 0.3107, + "num_input_tokens_seen": 43129872, + "step": 28345 + }, + { + "epoch": 0.9074323026694834, + "grad_norm": 1.070500135421753, + "learning_rate": 9.756571120228375e-06, + "loss": 0.5052, + "num_input_tokens_seen": 43137648, + "step": 28350 + }, + { + "epoch": 0.9075923436399719, + "grad_norm": 0.5232598781585693, + "learning_rate": 9.748790927310605e-06, + "loss": 0.445, + "num_input_tokens_seen": 43145040, + "step": 28355 + }, + { + "epoch": 0.9077523846104603, + "grad_norm": 0.48519840836524963, + "learning_rate": 9.741013086320946e-06, + "loss": 0.2655, + "num_input_tokens_seen": 43152656, + "step": 28360 + }, + { + "epoch": 0.9079124255809488, + "grad_norm": 1.511988878250122, + "learning_rate": 9.733237598458821e-06, + "loss": 0.4405, + "num_input_tokens_seen": 43160512, + "step": 28365 + }, + { + "epoch": 0.9080724665514371, + "grad_norm": 2.0976734161376953, + "learning_rate": 9.725464464923308e-06, + "loss": 0.5598, + "num_input_tokens_seen": 43167968, + "step": 28370 + }, + { + "epoch": 0.9082325075219256, + "grad_norm": 0.48701971769332886, + "learning_rate": 9.717693686913123e-06, + "loss": 0.4918, + "num_input_tokens_seen": 43175472, + "step": 28375 + }, + { + "epoch": 0.908392548492414, + "grad_norm": 0.9328317046165466, + "learning_rate": 9.709925265626632e-06, + "loss": 0.4835, + "num_input_tokens_seen": 43182992, + "step": 28380 + }, + { + "epoch": 0.9085525894629025, + "grad_norm": 0.4534572958946228, + "learning_rate": 9.702159202261801e-06, + "loss": 0.397, + "num_input_tokens_seen": 43190224, + "step": 28385 + }, + { + "epoch": 0.9087126304333909, + "grad_norm": 0.7844505906105042, + "learning_rate": 9.694395498016268e-06, + "loss": 0.3949, + "num_input_tokens_seen": 43197792, + "step": 28390 + }, + { + "epoch": 0.9088726714038794, + "grad_norm": 0.6227028965950012, + "learning_rate": 9.686634154087298e-06, + "loss": 0.4341, + "num_input_tokens_seen": 43205120, + "step": 28395 + }, + { + "epoch": 0.9090327123743679, + "grad_norm": 0.886562705039978, + "learning_rate": 9.678875171671776e-06, + "loss": 0.4438, + "num_input_tokens_seen": 43212768, + "step": 28400 + }, + { + "epoch": 0.9090327123743679, + "eval_loss": 0.4979356825351715, + "eval_runtime": 557.2422, + "eval_samples_per_second": 24.919, + "eval_steps_per_second": 12.46, + "num_input_tokens_seen": 43212768, + "step": 28400 + }, + { + "epoch": 0.9091927533448563, + "grad_norm": 1.3806334733963013, + "learning_rate": 9.671118551966246e-06, + "loss": 0.5545, + "num_input_tokens_seen": 43220096, + "step": 28405 + }, + { + "epoch": 0.9093527943153448, + "grad_norm": 0.7376247644424438, + "learning_rate": 9.66336429616686e-06, + "loss": 0.4309, + "num_input_tokens_seen": 43227248, + "step": 28410 + }, + { + "epoch": 0.9095128352858332, + "grad_norm": 0.408652126789093, + "learning_rate": 9.655612405469436e-06, + "loss": 0.3263, + "num_input_tokens_seen": 43234928, + "step": 28415 + }, + { + "epoch": 0.9096728762563216, + "grad_norm": 0.592233419418335, + "learning_rate": 9.647862881069413e-06, + "loss": 0.5831, + "num_input_tokens_seen": 43242688, + "step": 28420 + }, + { + "epoch": 0.90983291722681, + "grad_norm": 1.1108858585357666, + "learning_rate": 9.640115724161855e-06, + "loss": 0.5143, + "num_input_tokens_seen": 43250816, + "step": 28425 + }, + { + "epoch": 0.9099929581972985, + "grad_norm": 1.1058998107910156, + "learning_rate": 9.632370935941483e-06, + "loss": 0.6236, + "num_input_tokens_seen": 43258800, + "step": 28430 + }, + { + "epoch": 0.9101529991677869, + "grad_norm": 0.948935866355896, + "learning_rate": 9.624628517602634e-06, + "loss": 0.5267, + "num_input_tokens_seen": 43266592, + "step": 28435 + }, + { + "epoch": 0.9103130401382754, + "grad_norm": 1.2529810667037964, + "learning_rate": 9.61688847033928e-06, + "loss": 0.5616, + "num_input_tokens_seen": 43274128, + "step": 28440 + }, + { + "epoch": 0.9104730811087638, + "grad_norm": 0.9170437455177307, + "learning_rate": 9.609150795345051e-06, + "loss": 0.3816, + "num_input_tokens_seen": 43281856, + "step": 28445 + }, + { + "epoch": 0.9106331220792523, + "grad_norm": 0.4642944633960724, + "learning_rate": 9.601415493813171e-06, + "loss": 0.3406, + "num_input_tokens_seen": 43289552, + "step": 28450 + }, + { + "epoch": 0.9107931630497408, + "grad_norm": 0.6830311417579651, + "learning_rate": 9.593682566936533e-06, + "loss": 0.3196, + "num_input_tokens_seen": 43297120, + "step": 28455 + }, + { + "epoch": 0.9109532040202292, + "grad_norm": 0.7888422012329102, + "learning_rate": 9.58595201590766e-06, + "loss": 0.4933, + "num_input_tokens_seen": 43305216, + "step": 28460 + }, + { + "epoch": 0.9111132449907177, + "grad_norm": 0.6979511380195618, + "learning_rate": 9.578223841918681e-06, + "loss": 0.6179, + "num_input_tokens_seen": 43312192, + "step": 28465 + }, + { + "epoch": 0.911273285961206, + "grad_norm": 1.310945987701416, + "learning_rate": 9.570498046161389e-06, + "loss": 0.3928, + "num_input_tokens_seen": 43319376, + "step": 28470 + }, + { + "epoch": 0.9114333269316945, + "grad_norm": 0.738650918006897, + "learning_rate": 9.562774629827206e-06, + "loss": 0.4182, + "num_input_tokens_seen": 43327152, + "step": 28475 + }, + { + "epoch": 0.9115933679021829, + "grad_norm": 0.6495643854141235, + "learning_rate": 9.555053594107163e-06, + "loss": 0.4352, + "num_input_tokens_seen": 43334800, + "step": 28480 + }, + { + "epoch": 0.9117534088726714, + "grad_norm": 1.7866361141204834, + "learning_rate": 9.547334940191957e-06, + "loss": 0.6141, + "num_input_tokens_seen": 43341936, + "step": 28485 + }, + { + "epoch": 0.9119134498431598, + "grad_norm": 0.4114820957183838, + "learning_rate": 9.539618669271886e-06, + "loss": 0.5464, + "num_input_tokens_seen": 43349984, + "step": 28490 + }, + { + "epoch": 0.9120734908136483, + "grad_norm": 0.6824632883071899, + "learning_rate": 9.531904782536904e-06, + "loss": 0.5297, + "num_input_tokens_seen": 43357360, + "step": 28495 + }, + { + "epoch": 0.9122335317841367, + "grad_norm": 0.6075016260147095, + "learning_rate": 9.524193281176597e-06, + "loss": 0.427, + "num_input_tokens_seen": 43365280, + "step": 28500 + }, + { + "epoch": 0.9123935727546252, + "grad_norm": 0.5059555172920227, + "learning_rate": 9.516484166380165e-06, + "loss": 0.4247, + "num_input_tokens_seen": 43372704, + "step": 28505 + }, + { + "epoch": 0.9125536137251137, + "grad_norm": 1.1343108415603638, + "learning_rate": 9.508777439336447e-06, + "loss": 0.4809, + "num_input_tokens_seen": 43380272, + "step": 28510 + }, + { + "epoch": 0.9127136546956021, + "grad_norm": 0.4375837445259094, + "learning_rate": 9.50107310123393e-06, + "loss": 0.392, + "num_input_tokens_seen": 43387888, + "step": 28515 + }, + { + "epoch": 0.9128736956660906, + "grad_norm": 0.44180113077163696, + "learning_rate": 9.493371153260702e-06, + "loss": 0.473, + "num_input_tokens_seen": 43395120, + "step": 28520 + }, + { + "epoch": 0.9130337366365789, + "grad_norm": 0.7179558277130127, + "learning_rate": 9.485671596604523e-06, + "loss": 0.4588, + "num_input_tokens_seen": 43403040, + "step": 28525 + }, + { + "epoch": 0.9131937776070674, + "grad_norm": 1.0951457023620605, + "learning_rate": 9.477974432452738e-06, + "loss": 0.5387, + "num_input_tokens_seen": 43410608, + "step": 28530 + }, + { + "epoch": 0.9133538185775558, + "grad_norm": 0.9593439698219299, + "learning_rate": 9.470279661992356e-06, + "loss": 0.3977, + "num_input_tokens_seen": 43418448, + "step": 28535 + }, + { + "epoch": 0.9135138595480443, + "grad_norm": 0.6557828783988953, + "learning_rate": 9.462587286410021e-06, + "loss": 0.4084, + "num_input_tokens_seen": 43426080, + "step": 28540 + }, + { + "epoch": 0.9136739005185327, + "grad_norm": 0.4137406349182129, + "learning_rate": 9.454897306891972e-06, + "loss": 0.4061, + "num_input_tokens_seen": 43433744, + "step": 28545 + }, + { + "epoch": 0.9138339414890212, + "grad_norm": 0.8747947216033936, + "learning_rate": 9.44720972462411e-06, + "loss": 0.3828, + "num_input_tokens_seen": 43440944, + "step": 28550 + }, + { + "epoch": 0.9139939824595096, + "grad_norm": 0.5619136095046997, + "learning_rate": 9.439524540791964e-06, + "loss": 0.5337, + "num_input_tokens_seen": 43448368, + "step": 28555 + }, + { + "epoch": 0.9141540234299981, + "grad_norm": 0.6958082318305969, + "learning_rate": 9.431841756580673e-06, + "loss": 0.3695, + "num_input_tokens_seen": 43456256, + "step": 28560 + }, + { + "epoch": 0.9143140644004866, + "grad_norm": 0.8230636119842529, + "learning_rate": 9.42416137317503e-06, + "loss": 0.6908, + "num_input_tokens_seen": 43464608, + "step": 28565 + }, + { + "epoch": 0.914474105370975, + "grad_norm": 1.2333492040634155, + "learning_rate": 9.416483391759437e-06, + "loss": 0.5869, + "num_input_tokens_seen": 43472272, + "step": 28570 + }, + { + "epoch": 0.9146341463414634, + "grad_norm": 0.9807189702987671, + "learning_rate": 9.408807813517945e-06, + "loss": 0.4568, + "num_input_tokens_seen": 43479584, + "step": 28575 + }, + { + "epoch": 0.9147941873119518, + "grad_norm": 0.5335193276405334, + "learning_rate": 9.401134639634221e-06, + "loss": 0.4603, + "num_input_tokens_seen": 43487136, + "step": 28580 + }, + { + "epoch": 0.9149542282824403, + "grad_norm": 0.40005382895469666, + "learning_rate": 9.393463871291555e-06, + "loss": 0.3145, + "num_input_tokens_seen": 43494768, + "step": 28585 + }, + { + "epoch": 0.9151142692529287, + "grad_norm": 0.5337772369384766, + "learning_rate": 9.385795509672881e-06, + "loss": 0.4308, + "num_input_tokens_seen": 43502240, + "step": 28590 + }, + { + "epoch": 0.9152743102234172, + "grad_norm": 0.9205461144447327, + "learning_rate": 9.378129555960771e-06, + "loss": 0.5754, + "num_input_tokens_seen": 43509504, + "step": 28595 + }, + { + "epoch": 0.9154343511939056, + "grad_norm": 0.9310091137886047, + "learning_rate": 9.370466011337392e-06, + "loss": 0.419, + "num_input_tokens_seen": 43517168, + "step": 28600 + }, + { + "epoch": 0.9154343511939056, + "eval_loss": 0.4978503882884979, + "eval_runtime": 557.3424, + "eval_samples_per_second": 24.915, + "eval_steps_per_second": 12.457, + "num_input_tokens_seen": 43517168, + "step": 28600 + }, + { + "epoch": 0.9155943921643941, + "grad_norm": 0.4880717396736145, + "learning_rate": 9.362804876984573e-06, + "loss": 0.3966, + "num_input_tokens_seen": 43524784, + "step": 28605 + }, + { + "epoch": 0.9157544331348826, + "grad_norm": 1.8209768533706665, + "learning_rate": 9.355146154083747e-06, + "loss": 0.5539, + "num_input_tokens_seen": 43531984, + "step": 28610 + }, + { + "epoch": 0.915914474105371, + "grad_norm": 0.4126402735710144, + "learning_rate": 9.347489843815987e-06, + "loss": 0.3685, + "num_input_tokens_seen": 43539968, + "step": 28615 + }, + { + "epoch": 0.9160745150758595, + "grad_norm": 1.0522167682647705, + "learning_rate": 9.339835947362002e-06, + "loss": 0.5567, + "num_input_tokens_seen": 43547520, + "step": 28620 + }, + { + "epoch": 0.9162345560463478, + "grad_norm": 0.5426998138427734, + "learning_rate": 9.332184465902105e-06, + "loss": 0.4233, + "num_input_tokens_seen": 43554960, + "step": 28625 + }, + { + "epoch": 0.9163945970168363, + "grad_norm": 0.6863012909889221, + "learning_rate": 9.324535400616266e-06, + "loss": 0.603, + "num_input_tokens_seen": 43562800, + "step": 28630 + }, + { + "epoch": 0.9165546379873247, + "grad_norm": 0.9534783363342285, + "learning_rate": 9.31688875268405e-06, + "loss": 0.4532, + "num_input_tokens_seen": 43570560, + "step": 28635 + }, + { + "epoch": 0.9167146789578132, + "grad_norm": 0.8407192826271057, + "learning_rate": 9.309244523284674e-06, + "loss": 0.4764, + "num_input_tokens_seen": 43578784, + "step": 28640 + }, + { + "epoch": 0.9168747199283016, + "grad_norm": 0.7808179259300232, + "learning_rate": 9.301602713596982e-06, + "loss": 0.5745, + "num_input_tokens_seen": 43586800, + "step": 28645 + }, + { + "epoch": 0.9170347608987901, + "grad_norm": 1.1528563499450684, + "learning_rate": 9.293963324799432e-06, + "loss": 0.4028, + "num_input_tokens_seen": 43594112, + "step": 28650 + }, + { + "epoch": 0.9171948018692785, + "grad_norm": 1.4259731769561768, + "learning_rate": 9.286326358070104e-06, + "loss": 0.5373, + "num_input_tokens_seen": 43601536, + "step": 28655 + }, + { + "epoch": 0.917354842839767, + "grad_norm": 1.3446890115737915, + "learning_rate": 9.278691814586729e-06, + "loss": 0.453, + "num_input_tokens_seen": 43609136, + "step": 28660 + }, + { + "epoch": 0.9175148838102555, + "grad_norm": 0.45734429359436035, + "learning_rate": 9.271059695526635e-06, + "loss": 0.3382, + "num_input_tokens_seen": 43616144, + "step": 28665 + }, + { + "epoch": 0.9176749247807439, + "grad_norm": 0.5109876990318298, + "learning_rate": 9.263430002066805e-06, + "loss": 0.4989, + "num_input_tokens_seen": 43623488, + "step": 28670 + }, + { + "epoch": 0.9178349657512324, + "grad_norm": 0.7841423749923706, + "learning_rate": 9.25580273538382e-06, + "loss": 0.3887, + "num_input_tokens_seen": 43631072, + "step": 28675 + }, + { + "epoch": 0.9179950067217207, + "grad_norm": 0.7781479954719543, + "learning_rate": 9.248177896653907e-06, + "loss": 0.6916, + "num_input_tokens_seen": 43638304, + "step": 28680 + }, + { + "epoch": 0.9181550476922092, + "grad_norm": 0.7817234992980957, + "learning_rate": 9.240555487052918e-06, + "loss": 0.7024, + "num_input_tokens_seen": 43646416, + "step": 28685 + }, + { + "epoch": 0.9183150886626976, + "grad_norm": 1.1915950775146484, + "learning_rate": 9.232935507756313e-06, + "loss": 0.4226, + "num_input_tokens_seen": 43653488, + "step": 28690 + }, + { + "epoch": 0.9184751296331861, + "grad_norm": 0.47160160541534424, + "learning_rate": 9.225317959939193e-06, + "loss": 0.3925, + "num_input_tokens_seen": 43661136, + "step": 28695 + }, + { + "epoch": 0.9186351706036745, + "grad_norm": 0.3975287675857544, + "learning_rate": 9.217702844776287e-06, + "loss": 0.4398, + "num_input_tokens_seen": 43668400, + "step": 28700 + }, + { + "epoch": 0.918795211574163, + "grad_norm": 0.5801938772201538, + "learning_rate": 9.210090163441929e-06, + "loss": 0.3532, + "num_input_tokens_seen": 43675920, + "step": 28705 + }, + { + "epoch": 0.9189552525446514, + "grad_norm": 1.1117807626724243, + "learning_rate": 9.202479917110105e-06, + "loss": 0.4636, + "num_input_tokens_seen": 43683456, + "step": 28710 + }, + { + "epoch": 0.9191152935151399, + "grad_norm": 0.6121049523353577, + "learning_rate": 9.194872106954392e-06, + "loss": 0.4726, + "num_input_tokens_seen": 43691376, + "step": 28715 + }, + { + "epoch": 0.9192753344856284, + "grad_norm": 1.0208743810653687, + "learning_rate": 9.187266734148029e-06, + "loss": 0.573, + "num_input_tokens_seen": 43699008, + "step": 28720 + }, + { + "epoch": 0.9194353754561168, + "grad_norm": 1.5856003761291504, + "learning_rate": 9.179663799863849e-06, + "loss": 0.5203, + "num_input_tokens_seen": 43706848, + "step": 28725 + }, + { + "epoch": 0.9195954164266052, + "grad_norm": 0.7735625505447388, + "learning_rate": 9.172063305274317e-06, + "loss": 0.3133, + "num_input_tokens_seen": 43714096, + "step": 28730 + }, + { + "epoch": 0.9197554573970936, + "grad_norm": 0.5784391760826111, + "learning_rate": 9.164465251551527e-06, + "loss": 0.4343, + "num_input_tokens_seen": 43721248, + "step": 28735 + }, + { + "epoch": 0.9199154983675821, + "grad_norm": 1.1552519798278809, + "learning_rate": 9.156869639867205e-06, + "loss": 0.5192, + "num_input_tokens_seen": 43728448, + "step": 28740 + }, + { + "epoch": 0.9200755393380705, + "grad_norm": 1.0659393072128296, + "learning_rate": 9.149276471392677e-06, + "loss": 0.4996, + "num_input_tokens_seen": 43735872, + "step": 28745 + }, + { + "epoch": 0.920235580308559, + "grad_norm": 0.611329972743988, + "learning_rate": 9.141685747298914e-06, + "loss": 0.5783, + "num_input_tokens_seen": 43743440, + "step": 28750 + }, + { + "epoch": 0.9203956212790474, + "grad_norm": 1.0551668405532837, + "learning_rate": 9.13409746875649e-06, + "loss": 0.4938, + "num_input_tokens_seen": 43750368, + "step": 28755 + }, + { + "epoch": 0.9205556622495359, + "grad_norm": 0.6663140058517456, + "learning_rate": 9.12651163693562e-06, + "loss": 0.5531, + "num_input_tokens_seen": 43758128, + "step": 28760 + }, + { + "epoch": 0.9207157032200243, + "grad_norm": 1.8782920837402344, + "learning_rate": 9.11892825300614e-06, + "loss": 0.563, + "num_input_tokens_seen": 43765520, + "step": 28765 + }, + { + "epoch": 0.9208757441905128, + "grad_norm": 0.590531587600708, + "learning_rate": 9.111347318137491e-06, + "loss": 0.4531, + "num_input_tokens_seen": 43773008, + "step": 28770 + }, + { + "epoch": 0.9210357851610013, + "grad_norm": 0.7478114366531372, + "learning_rate": 9.103768833498755e-06, + "loss": 0.4804, + "num_input_tokens_seen": 43780976, + "step": 28775 + }, + { + "epoch": 0.9211958261314896, + "grad_norm": 1.2396833896636963, + "learning_rate": 9.096192800258639e-06, + "loss": 0.4412, + "num_input_tokens_seen": 43789184, + "step": 28780 + }, + { + "epoch": 0.9213558671019781, + "grad_norm": 0.6566888093948364, + "learning_rate": 9.088619219585443e-06, + "loss": 0.519, + "num_input_tokens_seen": 43796672, + "step": 28785 + }, + { + "epoch": 0.9215159080724665, + "grad_norm": 0.5401328802108765, + "learning_rate": 9.081048092647127e-06, + "loss": 0.602, + "num_input_tokens_seen": 43804704, + "step": 28790 + }, + { + "epoch": 0.921675949042955, + "grad_norm": 0.7001001238822937, + "learning_rate": 9.073479420611245e-06, + "loss": 0.5797, + "num_input_tokens_seen": 43812816, + "step": 28795 + }, + { + "epoch": 0.9218359900134434, + "grad_norm": 1.2541199922561646, + "learning_rate": 9.065913204644974e-06, + "loss": 0.4622, + "num_input_tokens_seen": 43820288, + "step": 28800 + }, + { + "epoch": 0.9218359900134434, + "eval_loss": 0.49776753783226013, + "eval_runtime": 557.6284, + "eval_samples_per_second": 24.902, + "eval_steps_per_second": 12.451, + "num_input_tokens_seen": 43820288, + "step": 28800 + }, + { + "epoch": 0.9219960309839319, + "grad_norm": 0.8651156425476074, + "learning_rate": 9.058349445915135e-06, + "loss": 0.4236, + "num_input_tokens_seen": 43828336, + "step": 28805 + }, + { + "epoch": 0.9221560719544203, + "grad_norm": 1.0640366077423096, + "learning_rate": 9.050788145588138e-06, + "loss": 0.5638, + "num_input_tokens_seen": 43836032, + "step": 28810 + }, + { + "epoch": 0.9223161129249088, + "grad_norm": 1.4858715534210205, + "learning_rate": 9.043229304830039e-06, + "loss": 0.6893, + "num_input_tokens_seen": 43843600, + "step": 28815 + }, + { + "epoch": 0.9224761538953972, + "grad_norm": 1.0073672533035278, + "learning_rate": 9.035672924806515e-06, + "loss": 0.4587, + "num_input_tokens_seen": 43851104, + "step": 28820 + }, + { + "epoch": 0.9226361948658857, + "grad_norm": 0.902799129486084, + "learning_rate": 9.028119006682839e-06, + "loss": 0.4875, + "num_input_tokens_seen": 43859296, + "step": 28825 + }, + { + "epoch": 0.9227962358363742, + "grad_norm": 0.5133930444717407, + "learning_rate": 9.020567551623935e-06, + "loss": 0.4689, + "num_input_tokens_seen": 43866992, + "step": 28830 + }, + { + "epoch": 0.9229562768068625, + "grad_norm": 0.7638083100318909, + "learning_rate": 9.013018560794318e-06, + "loss": 0.4234, + "num_input_tokens_seen": 43874704, + "step": 28835 + }, + { + "epoch": 0.923116317777351, + "grad_norm": 1.2472878694534302, + "learning_rate": 9.005472035358139e-06, + "loss": 0.5775, + "num_input_tokens_seen": 43882672, + "step": 28840 + }, + { + "epoch": 0.9232763587478394, + "grad_norm": 0.7865089774131775, + "learning_rate": 8.997927976479185e-06, + "loss": 0.3812, + "num_input_tokens_seen": 43890032, + "step": 28845 + }, + { + "epoch": 0.9234363997183279, + "grad_norm": 1.022354245185852, + "learning_rate": 8.99038638532082e-06, + "loss": 0.4622, + "num_input_tokens_seen": 43897584, + "step": 28850 + }, + { + "epoch": 0.9235964406888163, + "grad_norm": 1.1298362016677856, + "learning_rate": 8.982847263046065e-06, + "loss": 0.4663, + "num_input_tokens_seen": 43905200, + "step": 28855 + }, + { + "epoch": 0.9237564816593048, + "grad_norm": 0.5016540884971619, + "learning_rate": 8.975310610817555e-06, + "loss": 0.6145, + "num_input_tokens_seen": 43912512, + "step": 28860 + }, + { + "epoch": 0.9239165226297932, + "grad_norm": 0.6834735870361328, + "learning_rate": 8.967776429797528e-06, + "loss": 0.5233, + "num_input_tokens_seen": 43920800, + "step": 28865 + }, + { + "epoch": 0.9240765636002817, + "grad_norm": 0.5180326104164124, + "learning_rate": 8.960244721147842e-06, + "loss": 0.2561, + "num_input_tokens_seen": 43928384, + "step": 28870 + }, + { + "epoch": 0.9242366045707702, + "grad_norm": 0.8400763273239136, + "learning_rate": 8.952715486029995e-06, + "loss": 0.4746, + "num_input_tokens_seen": 43936112, + "step": 28875 + }, + { + "epoch": 0.9243966455412586, + "grad_norm": 0.42211347818374634, + "learning_rate": 8.945188725605075e-06, + "loss": 0.3742, + "num_input_tokens_seen": 43943312, + "step": 28880 + }, + { + "epoch": 0.924556686511747, + "grad_norm": 0.6840474009513855, + "learning_rate": 8.937664441033817e-06, + "loss": 0.6375, + "num_input_tokens_seen": 43950768, + "step": 28885 + }, + { + "epoch": 0.9247167274822354, + "grad_norm": 0.4498637020587921, + "learning_rate": 8.930142633476549e-06, + "loss": 0.3719, + "num_input_tokens_seen": 43958512, + "step": 28890 + }, + { + "epoch": 0.9248767684527239, + "grad_norm": 3.328303337097168, + "learning_rate": 8.92262330409323e-06, + "loss": 0.5455, + "num_input_tokens_seen": 43966288, + "step": 28895 + }, + { + "epoch": 0.9250368094232123, + "grad_norm": 0.7282616496086121, + "learning_rate": 8.915106454043448e-06, + "loss": 0.4114, + "num_input_tokens_seen": 43974224, + "step": 28900 + }, + { + "epoch": 0.9251968503937008, + "grad_norm": 0.6042773127555847, + "learning_rate": 8.90759208448638e-06, + "loss": 0.3937, + "num_input_tokens_seen": 43981920, + "step": 28905 + }, + { + "epoch": 0.9253568913641892, + "grad_norm": 0.7671576738357544, + "learning_rate": 8.900080196580848e-06, + "loss": 0.5978, + "num_input_tokens_seen": 43989632, + "step": 28910 + }, + { + "epoch": 0.9255169323346777, + "grad_norm": 1.1805824041366577, + "learning_rate": 8.892570791485267e-06, + "loss": 0.5283, + "num_input_tokens_seen": 43997568, + "step": 28915 + }, + { + "epoch": 0.9256769733051661, + "grad_norm": 0.5794785618782043, + "learning_rate": 8.885063870357688e-06, + "loss": 0.5376, + "num_input_tokens_seen": 44005136, + "step": 28920 + }, + { + "epoch": 0.9258370142756546, + "grad_norm": 0.7703301906585693, + "learning_rate": 8.87755943435578e-06, + "loss": 0.4552, + "num_input_tokens_seen": 44013008, + "step": 28925 + }, + { + "epoch": 0.9259970552461431, + "grad_norm": 0.6666787266731262, + "learning_rate": 8.87005748463681e-06, + "loss": 0.5312, + "num_input_tokens_seen": 44020480, + "step": 28930 + }, + { + "epoch": 0.9261570962166314, + "grad_norm": 0.8592641353607178, + "learning_rate": 8.862558022357681e-06, + "loss": 0.4613, + "num_input_tokens_seen": 44027824, + "step": 28935 + }, + { + "epoch": 0.9263171371871199, + "grad_norm": 0.5823776125907898, + "learning_rate": 8.855061048674903e-06, + "loss": 0.4549, + "num_input_tokens_seen": 44035424, + "step": 28940 + }, + { + "epoch": 0.9264771781576083, + "grad_norm": 1.205589771270752, + "learning_rate": 8.847566564744595e-06, + "loss": 0.4419, + "num_input_tokens_seen": 44042976, + "step": 28945 + }, + { + "epoch": 0.9266372191280968, + "grad_norm": 0.9196025133132935, + "learning_rate": 8.840074571722512e-06, + "loss": 0.4599, + "num_input_tokens_seen": 44050640, + "step": 28950 + }, + { + "epoch": 0.9267972600985852, + "grad_norm": 0.6685994863510132, + "learning_rate": 8.832585070764002e-06, + "loss": 0.4221, + "num_input_tokens_seen": 44058048, + "step": 28955 + }, + { + "epoch": 0.9269573010690737, + "grad_norm": 1.3748940229415894, + "learning_rate": 8.825098063024045e-06, + "loss": 0.402, + "num_input_tokens_seen": 44065424, + "step": 28960 + }, + { + "epoch": 0.9271173420395621, + "grad_norm": 0.8298453688621521, + "learning_rate": 8.817613549657244e-06, + "loss": 0.4373, + "num_input_tokens_seen": 44072784, + "step": 28965 + }, + { + "epoch": 0.9272773830100506, + "grad_norm": 1.692057728767395, + "learning_rate": 8.810131531817783e-06, + "loss": 0.5199, + "num_input_tokens_seen": 44080480, + "step": 28970 + }, + { + "epoch": 0.927437423980539, + "grad_norm": 0.7711567878723145, + "learning_rate": 8.802652010659496e-06, + "loss": 0.4507, + "num_input_tokens_seen": 44087648, + "step": 28975 + }, + { + "epoch": 0.9275974649510275, + "grad_norm": 0.9237624406814575, + "learning_rate": 8.795174987335827e-06, + "loss": 0.4542, + "num_input_tokens_seen": 44095024, + "step": 28980 + }, + { + "epoch": 0.927757505921516, + "grad_norm": 0.6267331838607788, + "learning_rate": 8.787700462999807e-06, + "loss": 0.5644, + "num_input_tokens_seen": 44102464, + "step": 28985 + }, + { + "epoch": 0.9279175468920043, + "grad_norm": 1.1132241487503052, + "learning_rate": 8.780228438804122e-06, + "loss": 0.6351, + "num_input_tokens_seen": 44109840, + "step": 28990 + }, + { + "epoch": 0.9280775878624928, + "grad_norm": 0.5828704237937927, + "learning_rate": 8.772758915901032e-06, + "loss": 0.4525, + "num_input_tokens_seen": 44118016, + "step": 28995 + }, + { + "epoch": 0.9282376288329812, + "grad_norm": 0.6993043422698975, + "learning_rate": 8.765291895442443e-06, + "loss": 0.5337, + "num_input_tokens_seen": 44124992, + "step": 29000 + }, + { + "epoch": 0.9282376288329812, + "eval_loss": 0.4974127411842346, + "eval_runtime": 557.502, + "eval_samples_per_second": 24.908, + "eval_steps_per_second": 12.454, + "num_input_tokens_seen": 44124992, + "step": 29000 + }, + { + "epoch": 0.9283976698034697, + "grad_norm": 0.4774021506309509, + "learning_rate": 8.75782737857987e-06, + "loss": 0.5035, + "num_input_tokens_seen": 44132016, + "step": 29005 + }, + { + "epoch": 0.9285577107739581, + "grad_norm": 0.5718454122543335, + "learning_rate": 8.750365366464425e-06, + "loss": 0.4806, + "num_input_tokens_seen": 44139760, + "step": 29010 + }, + { + "epoch": 0.9287177517444466, + "grad_norm": 0.7056530714035034, + "learning_rate": 8.742905860246838e-06, + "loss": 0.3394, + "num_input_tokens_seen": 44147040, + "step": 29015 + }, + { + "epoch": 0.928877792714935, + "grad_norm": 0.6354730725288391, + "learning_rate": 8.735448861077478e-06, + "loss": 0.5934, + "num_input_tokens_seen": 44154880, + "step": 29020 + }, + { + "epoch": 0.9290378336854235, + "grad_norm": 0.8218173980712891, + "learning_rate": 8.727994370106288e-06, + "loss": 0.4988, + "num_input_tokens_seen": 44162672, + "step": 29025 + }, + { + "epoch": 0.9291978746559119, + "grad_norm": 1.055851936340332, + "learning_rate": 8.720542388482861e-06, + "loss": 0.5407, + "num_input_tokens_seen": 44170176, + "step": 29030 + }, + { + "epoch": 0.9293579156264004, + "grad_norm": 0.8551093339920044, + "learning_rate": 8.71309291735637e-06, + "loss": 0.5171, + "num_input_tokens_seen": 44178416, + "step": 29035 + }, + { + "epoch": 0.9295179565968889, + "grad_norm": 0.9811655879020691, + "learning_rate": 8.705645957875621e-06, + "loss": 0.5422, + "num_input_tokens_seen": 44186064, + "step": 29040 + }, + { + "epoch": 0.9296779975673772, + "grad_norm": 1.077210783958435, + "learning_rate": 8.698201511189048e-06, + "loss": 0.5001, + "num_input_tokens_seen": 44193424, + "step": 29045 + }, + { + "epoch": 0.9298380385378657, + "grad_norm": 0.8700220584869385, + "learning_rate": 8.690759578444649e-06, + "loss": 0.5665, + "num_input_tokens_seen": 44201168, + "step": 29050 + }, + { + "epoch": 0.9299980795083541, + "grad_norm": 1.2206529378890991, + "learning_rate": 8.68332016079008e-06, + "loss": 0.5879, + "num_input_tokens_seen": 44208368, + "step": 29055 + }, + { + "epoch": 0.9301581204788426, + "grad_norm": 0.7905001044273376, + "learning_rate": 8.6758832593726e-06, + "loss": 0.7762, + "num_input_tokens_seen": 44215376, + "step": 29060 + }, + { + "epoch": 0.930318161449331, + "grad_norm": 0.45279622077941895, + "learning_rate": 8.668448875339053e-06, + "loss": 0.3662, + "num_input_tokens_seen": 44222800, + "step": 29065 + }, + { + "epoch": 0.9304782024198195, + "grad_norm": 0.7287971377372742, + "learning_rate": 8.661017009835933e-06, + "loss": 0.4123, + "num_input_tokens_seen": 44230256, + "step": 29070 + }, + { + "epoch": 0.9306382433903079, + "grad_norm": 0.4298986494541168, + "learning_rate": 8.653587664009311e-06, + "loss": 0.4421, + "num_input_tokens_seen": 44238368, + "step": 29075 + }, + { + "epoch": 0.9307982843607964, + "grad_norm": 1.8214142322540283, + "learning_rate": 8.646160839004902e-06, + "loss": 0.5258, + "num_input_tokens_seen": 44246064, + "step": 29080 + }, + { + "epoch": 0.9309583253312849, + "grad_norm": 1.4714387655258179, + "learning_rate": 8.638736535967998e-06, + "loss": 0.5288, + "num_input_tokens_seen": 44253504, + "step": 29085 + }, + { + "epoch": 0.9311183663017732, + "grad_norm": 0.4289708733558655, + "learning_rate": 8.631314756043535e-06, + "loss": 0.4555, + "num_input_tokens_seen": 44260704, + "step": 29090 + }, + { + "epoch": 0.9312784072722617, + "grad_norm": 0.750105082988739, + "learning_rate": 8.62389550037603e-06, + "loss": 0.295, + "num_input_tokens_seen": 44268656, + "step": 29095 + }, + { + "epoch": 0.9314384482427501, + "grad_norm": 1.0169060230255127, + "learning_rate": 8.616478770109646e-06, + "loss": 0.3745, + "num_input_tokens_seen": 44275840, + "step": 29100 + }, + { + "epoch": 0.9315984892132386, + "grad_norm": 1.2237744331359863, + "learning_rate": 8.609064566388111e-06, + "loss": 0.4663, + "num_input_tokens_seen": 44283296, + "step": 29105 + }, + { + "epoch": 0.931758530183727, + "grad_norm": 0.8432574272155762, + "learning_rate": 8.601652890354815e-06, + "loss": 0.3892, + "num_input_tokens_seen": 44290608, + "step": 29110 + }, + { + "epoch": 0.9319185711542155, + "grad_norm": 1.6463618278503418, + "learning_rate": 8.594243743152705e-06, + "loss": 0.3988, + "num_input_tokens_seen": 44297936, + "step": 29115 + }, + { + "epoch": 0.9320786121247039, + "grad_norm": 0.8918399214744568, + "learning_rate": 8.58683712592438e-06, + "loss": 0.5351, + "num_input_tokens_seen": 44305904, + "step": 29120 + }, + { + "epoch": 0.9322386530951924, + "grad_norm": 0.3588482737541199, + "learning_rate": 8.579433039812037e-06, + "loss": 0.5206, + "num_input_tokens_seen": 44313776, + "step": 29125 + }, + { + "epoch": 0.9323986940656808, + "grad_norm": 0.8382155895233154, + "learning_rate": 8.572031485957466e-06, + "loss": 0.4847, + "num_input_tokens_seen": 44321152, + "step": 29130 + }, + { + "epoch": 0.9325587350361693, + "grad_norm": 0.7521788477897644, + "learning_rate": 8.564632465502084e-06, + "loss": 0.4154, + "num_input_tokens_seen": 44328800, + "step": 29135 + }, + { + "epoch": 0.9327187760066578, + "grad_norm": 1.4114257097244263, + "learning_rate": 8.557235979586928e-06, + "loss": 0.7173, + "num_input_tokens_seen": 44336496, + "step": 29140 + }, + { + "epoch": 0.9328788169771461, + "grad_norm": 0.7746953964233398, + "learning_rate": 8.549842029352606e-06, + "loss": 0.4986, + "num_input_tokens_seen": 44344144, + "step": 29145 + }, + { + "epoch": 0.9330388579476346, + "grad_norm": 2.939336061477661, + "learning_rate": 8.542450615939376e-06, + "loss": 0.4416, + "num_input_tokens_seen": 44352240, + "step": 29150 + }, + { + "epoch": 0.933198898918123, + "grad_norm": 1.0596802234649658, + "learning_rate": 8.535061740487082e-06, + "loss": 0.4835, + "num_input_tokens_seen": 44359632, + "step": 29155 + }, + { + "epoch": 0.9333589398886115, + "grad_norm": 0.8450761437416077, + "learning_rate": 8.527675404135168e-06, + "loss": 0.459, + "num_input_tokens_seen": 44367184, + "step": 29160 + }, + { + "epoch": 0.9335189808590999, + "grad_norm": 0.8807206749916077, + "learning_rate": 8.520291608022724e-06, + "loss": 0.7592, + "num_input_tokens_seen": 44375088, + "step": 29165 + }, + { + "epoch": 0.9336790218295884, + "grad_norm": 0.6834611892700195, + "learning_rate": 8.512910353288398e-06, + "loss": 0.7485, + "num_input_tokens_seen": 44382816, + "step": 29170 + }, + { + "epoch": 0.9338390628000768, + "grad_norm": 1.3160454034805298, + "learning_rate": 8.505531641070486e-06, + "loss": 0.3627, + "num_input_tokens_seen": 44390576, + "step": 29175 + }, + { + "epoch": 0.9339991037705653, + "grad_norm": 0.5691772699356079, + "learning_rate": 8.498155472506885e-06, + "loss": 0.3425, + "num_input_tokens_seen": 44398848, + "step": 29180 + }, + { + "epoch": 0.9341591447410537, + "grad_norm": 0.7460329532623291, + "learning_rate": 8.49078184873508e-06, + "loss": 0.5675, + "num_input_tokens_seen": 44406240, + "step": 29185 + }, + { + "epoch": 0.9343191857115422, + "grad_norm": 1.821981430053711, + "learning_rate": 8.483410770892188e-06, + "loss": 0.5934, + "num_input_tokens_seen": 44413840, + "step": 29190 + }, + { + "epoch": 0.9344792266820307, + "grad_norm": 0.8869585394859314, + "learning_rate": 8.476042240114909e-06, + "loss": 0.5843, + "num_input_tokens_seen": 44421504, + "step": 29195 + }, + { + "epoch": 0.934639267652519, + "grad_norm": 0.7018333673477173, + "learning_rate": 8.468676257539568e-06, + "loss": 0.5332, + "num_input_tokens_seen": 44428848, + "step": 29200 + }, + { + "epoch": 0.934639267652519, + "eval_loss": 0.4972049295902252, + "eval_runtime": 557.3481, + "eval_samples_per_second": 24.914, + "eval_steps_per_second": 12.457, + "num_input_tokens_seen": 44428848, + "step": 29200 + }, + { + "epoch": 0.9347993086230075, + "grad_norm": 0.5892936587333679, + "learning_rate": 8.4613128243021e-06, + "loss": 0.3747, + "num_input_tokens_seen": 44436848, + "step": 29205 + }, + { + "epoch": 0.9349593495934959, + "grad_norm": 0.3548818528652191, + "learning_rate": 8.453951941538028e-06, + "loss": 0.4623, + "num_input_tokens_seen": 44444720, + "step": 29210 + }, + { + "epoch": 0.9351193905639844, + "grad_norm": 0.7349144816398621, + "learning_rate": 8.446593610382495e-06, + "loss": 0.6256, + "num_input_tokens_seen": 44452736, + "step": 29215 + }, + { + "epoch": 0.9352794315344728, + "grad_norm": 0.9050276279449463, + "learning_rate": 8.439237831970259e-06, + "loss": 0.4816, + "num_input_tokens_seen": 44460656, + "step": 29220 + }, + { + "epoch": 0.9354394725049613, + "grad_norm": 0.9681312441825867, + "learning_rate": 8.431884607435667e-06, + "loss": 0.3475, + "num_input_tokens_seen": 44467744, + "step": 29225 + }, + { + "epoch": 0.9355995134754497, + "grad_norm": 0.7218356728553772, + "learning_rate": 8.424533937912665e-06, + "loss": 0.6052, + "num_input_tokens_seen": 44475392, + "step": 29230 + }, + { + "epoch": 0.9357595544459382, + "grad_norm": 0.6114490628242493, + "learning_rate": 8.41718582453484e-06, + "loss": 0.6665, + "num_input_tokens_seen": 44483008, + "step": 29235 + }, + { + "epoch": 0.9359195954164266, + "grad_norm": 0.564607560634613, + "learning_rate": 8.409840268435346e-06, + "loss": 0.6294, + "num_input_tokens_seen": 44490736, + "step": 29240 + }, + { + "epoch": 0.936079636386915, + "grad_norm": 1.1011184453964233, + "learning_rate": 8.402497270746976e-06, + "loss": 0.5891, + "num_input_tokens_seen": 44498656, + "step": 29245 + }, + { + "epoch": 0.9362396773574035, + "grad_norm": 0.9844199419021606, + "learning_rate": 8.395156832602095e-06, + "loss": 0.4519, + "num_input_tokens_seen": 44505952, + "step": 29250 + }, + { + "epoch": 0.9363997183278919, + "grad_norm": 0.6531586050987244, + "learning_rate": 8.387818955132707e-06, + "loss": 0.4421, + "num_input_tokens_seen": 44513248, + "step": 29255 + }, + { + "epoch": 0.9365597592983804, + "grad_norm": 0.7324082255363464, + "learning_rate": 8.38048363947039e-06, + "loss": 0.7567, + "num_input_tokens_seen": 44521184, + "step": 29260 + }, + { + "epoch": 0.9367198002688688, + "grad_norm": 1.2798951864242554, + "learning_rate": 8.373150886746351e-06, + "loss": 0.5425, + "num_input_tokens_seen": 44528656, + "step": 29265 + }, + { + "epoch": 0.9368798412393573, + "grad_norm": 0.8124381303787231, + "learning_rate": 8.365820698091397e-06, + "loss": 0.5383, + "num_input_tokens_seen": 44536704, + "step": 29270 + }, + { + "epoch": 0.9370398822098457, + "grad_norm": 1.363121747970581, + "learning_rate": 8.358493074635922e-06, + "loss": 0.5985, + "num_input_tokens_seen": 44544224, + "step": 29275 + }, + { + "epoch": 0.9371999231803342, + "grad_norm": 0.5642310380935669, + "learning_rate": 8.351168017509948e-06, + "loss": 0.3259, + "num_input_tokens_seen": 44551776, + "step": 29280 + }, + { + "epoch": 0.9373599641508226, + "grad_norm": 0.6328995823860168, + "learning_rate": 8.343845527843094e-06, + "loss": 0.4764, + "num_input_tokens_seen": 44559296, + "step": 29285 + }, + { + "epoch": 0.9375200051213111, + "grad_norm": 1.3942655324935913, + "learning_rate": 8.336525606764566e-06, + "loss": 0.6458, + "num_input_tokens_seen": 44566512, + "step": 29290 + }, + { + "epoch": 0.9376800460917994, + "grad_norm": 0.48211681842803955, + "learning_rate": 8.329208255403204e-06, + "loss": 0.4057, + "num_input_tokens_seen": 44573792, + "step": 29295 + }, + { + "epoch": 0.9378400870622879, + "grad_norm": 0.8972514867782593, + "learning_rate": 8.321893474887426e-06, + "loss": 0.555, + "num_input_tokens_seen": 44581488, + "step": 29300 + }, + { + "epoch": 0.9380001280327764, + "grad_norm": 0.49323365092277527, + "learning_rate": 8.31458126634526e-06, + "loss": 0.3439, + "num_input_tokens_seen": 44588928, + "step": 29305 + }, + { + "epoch": 0.9381601690032648, + "grad_norm": 1.0279252529144287, + "learning_rate": 8.30727163090435e-06, + "loss": 0.4488, + "num_input_tokens_seen": 44596416, + "step": 29310 + }, + { + "epoch": 0.9383202099737533, + "grad_norm": 0.46952277421951294, + "learning_rate": 8.29996456969192e-06, + "loss": 0.43, + "num_input_tokens_seen": 44603840, + "step": 29315 + }, + { + "epoch": 0.9384802509442417, + "grad_norm": 0.5652886033058167, + "learning_rate": 8.292660083834818e-06, + "loss": 0.4391, + "num_input_tokens_seen": 44611344, + "step": 29320 + }, + { + "epoch": 0.9386402919147302, + "grad_norm": 1.3339053392410278, + "learning_rate": 8.2853581744595e-06, + "loss": 0.5814, + "num_input_tokens_seen": 44619088, + "step": 29325 + }, + { + "epoch": 0.9388003328852186, + "grad_norm": 0.6252975463867188, + "learning_rate": 8.278058842691991e-06, + "loss": 0.5069, + "num_input_tokens_seen": 44626832, + "step": 29330 + }, + { + "epoch": 0.9389603738557071, + "grad_norm": 0.7137742042541504, + "learning_rate": 8.27076208965796e-06, + "loss": 0.4172, + "num_input_tokens_seen": 44634928, + "step": 29335 + }, + { + "epoch": 0.9391204148261955, + "grad_norm": 0.8505539298057556, + "learning_rate": 8.263467916482637e-06, + "loss": 0.4491, + "num_input_tokens_seen": 44642624, + "step": 29340 + }, + { + "epoch": 0.939280455796684, + "grad_norm": 0.6726827025413513, + "learning_rate": 8.256176324290885e-06, + "loss": 0.5761, + "num_input_tokens_seen": 44649904, + "step": 29345 + }, + { + "epoch": 0.9394404967671725, + "grad_norm": 0.9855346083641052, + "learning_rate": 8.248887314207168e-06, + "loss": 0.4673, + "num_input_tokens_seen": 44658816, + "step": 29350 + }, + { + "epoch": 0.9396005377376608, + "grad_norm": 0.5738513469696045, + "learning_rate": 8.24160088735553e-06, + "loss": 0.5147, + "num_input_tokens_seen": 44666784, + "step": 29355 + }, + { + "epoch": 0.9397605787081493, + "grad_norm": 0.5045621395111084, + "learning_rate": 8.234317044859629e-06, + "loss": 0.4279, + "num_input_tokens_seen": 44674944, + "step": 29360 + }, + { + "epoch": 0.9399206196786377, + "grad_norm": 0.8590540289878845, + "learning_rate": 8.227035787842744e-06, + "loss": 0.5287, + "num_input_tokens_seen": 44682352, + "step": 29365 + }, + { + "epoch": 0.9400806606491262, + "grad_norm": 1.7305827140808105, + "learning_rate": 8.219757117427721e-06, + "loss": 0.7624, + "num_input_tokens_seen": 44690320, + "step": 29370 + }, + { + "epoch": 0.9402407016196146, + "grad_norm": 0.4457116723060608, + "learning_rate": 8.212481034737014e-06, + "loss": 0.4643, + "num_input_tokens_seen": 44698192, + "step": 29375 + }, + { + "epoch": 0.9404007425901031, + "grad_norm": 4.278615474700928, + "learning_rate": 8.205207540892707e-06, + "loss": 0.4674, + "num_input_tokens_seen": 44705616, + "step": 29380 + }, + { + "epoch": 0.9405607835605915, + "grad_norm": 1.2552001476287842, + "learning_rate": 8.197936637016442e-06, + "loss": 0.6429, + "num_input_tokens_seen": 44712960, + "step": 29385 + }, + { + "epoch": 0.94072082453108, + "grad_norm": 0.5587794184684753, + "learning_rate": 8.190668324229508e-06, + "loss": 0.4925, + "num_input_tokens_seen": 44720384, + "step": 29390 + }, + { + "epoch": 0.9408808655015684, + "grad_norm": 0.6799207329750061, + "learning_rate": 8.183402603652749e-06, + "loss": 0.4437, + "num_input_tokens_seen": 44727968, + "step": 29395 + }, + { + "epoch": 0.9410409064720568, + "grad_norm": 1.0019447803497314, + "learning_rate": 8.176139476406635e-06, + "loss": 0.3471, + "num_input_tokens_seen": 44735232, + "step": 29400 + }, + { + "epoch": 0.9410409064720568, + "eval_loss": 0.49689826369285583, + "eval_runtime": 558.9061, + "eval_samples_per_second": 24.845, + "eval_steps_per_second": 12.422, + "num_input_tokens_seen": 44735232, + "step": 29400 + }, + { + "epoch": 0.9412009474425453, + "grad_norm": 0.690008282661438, + "learning_rate": 8.16887894361125e-06, + "loss": 0.3979, + "num_input_tokens_seen": 44743104, + "step": 29405 + }, + { + "epoch": 0.9413609884130337, + "grad_norm": 0.582870364189148, + "learning_rate": 8.161621006386233e-06, + "loss": 0.3817, + "num_input_tokens_seen": 44751104, + "step": 29410 + }, + { + "epoch": 0.9415210293835222, + "grad_norm": 1.8879958391189575, + "learning_rate": 8.154365665850869e-06, + "loss": 0.4604, + "num_input_tokens_seen": 44758592, + "step": 29415 + }, + { + "epoch": 0.9416810703540106, + "grad_norm": 0.631984531879425, + "learning_rate": 8.147112923124005e-06, + "loss": 0.4332, + "num_input_tokens_seen": 44766016, + "step": 29420 + }, + { + "epoch": 0.9418411113244991, + "grad_norm": 0.5652392506599426, + "learning_rate": 8.13986277932412e-06, + "loss": 0.5689, + "num_input_tokens_seen": 44773520, + "step": 29425 + }, + { + "epoch": 0.9420011522949875, + "grad_norm": 0.8550869226455688, + "learning_rate": 8.132615235569277e-06, + "loss": 0.3746, + "num_input_tokens_seen": 44780944, + "step": 29430 + }, + { + "epoch": 0.942161193265476, + "grad_norm": 0.6006888747215271, + "learning_rate": 8.125370292977124e-06, + "loss": 0.3716, + "num_input_tokens_seen": 44788640, + "step": 29435 + }, + { + "epoch": 0.9423212342359644, + "grad_norm": 0.6625158190727234, + "learning_rate": 8.118127952664944e-06, + "loss": 0.4764, + "num_input_tokens_seen": 44796048, + "step": 29440 + }, + { + "epoch": 0.9424812752064529, + "grad_norm": 0.5453199744224548, + "learning_rate": 8.110888215749574e-06, + "loss": 0.53, + "num_input_tokens_seen": 44803744, + "step": 29445 + }, + { + "epoch": 0.9426413161769412, + "grad_norm": 0.8739892840385437, + "learning_rate": 8.10365108334749e-06, + "loss": 0.429, + "num_input_tokens_seen": 44811344, + "step": 29450 + }, + { + "epoch": 0.9428013571474297, + "grad_norm": 1.098188042640686, + "learning_rate": 8.096416556574743e-06, + "loss": 0.447, + "num_input_tokens_seen": 44819120, + "step": 29455 + }, + { + "epoch": 0.9429613981179182, + "grad_norm": 0.37975290417671204, + "learning_rate": 8.08918463654698e-06, + "loss": 0.3237, + "num_input_tokens_seen": 44826464, + "step": 29460 + }, + { + "epoch": 0.9431214390884066, + "grad_norm": 0.6080119013786316, + "learning_rate": 8.081955324379458e-06, + "loss": 0.4954, + "num_input_tokens_seen": 44834160, + "step": 29465 + }, + { + "epoch": 0.9432814800588951, + "grad_norm": 0.45741507411003113, + "learning_rate": 8.074728621187039e-06, + "loss": 0.3692, + "num_input_tokens_seen": 44841920, + "step": 29470 + }, + { + "epoch": 0.9434415210293835, + "grad_norm": 0.8877120018005371, + "learning_rate": 8.067504528084158e-06, + "loss": 0.6296, + "num_input_tokens_seen": 44849616, + "step": 29475 + }, + { + "epoch": 0.943601561999872, + "grad_norm": 1.0111768245697021, + "learning_rate": 8.060283046184861e-06, + "loss": 0.5573, + "num_input_tokens_seen": 44857952, + "step": 29480 + }, + { + "epoch": 0.9437616029703604, + "grad_norm": 0.7405586242675781, + "learning_rate": 8.053064176602806e-06, + "loss": 0.3656, + "num_input_tokens_seen": 44865584, + "step": 29485 + }, + { + "epoch": 0.9439216439408489, + "grad_norm": 0.533115565776825, + "learning_rate": 8.045847920451216e-06, + "loss": 0.5695, + "num_input_tokens_seen": 44873456, + "step": 29490 + }, + { + "epoch": 0.9440816849113373, + "grad_norm": 1.0969241857528687, + "learning_rate": 8.038634278842944e-06, + "loss": 0.6433, + "num_input_tokens_seen": 44881280, + "step": 29495 + }, + { + "epoch": 0.9442417258818258, + "grad_norm": 0.6489683985710144, + "learning_rate": 8.031423252890408e-06, + "loss": 0.4675, + "num_input_tokens_seen": 44889424, + "step": 29500 + }, + { + "epoch": 0.9444017668523141, + "grad_norm": 1.5335006713867188, + "learning_rate": 8.024214843705646e-06, + "loss": 0.6087, + "num_input_tokens_seen": 44896592, + "step": 29505 + }, + { + "epoch": 0.9445618078228026, + "grad_norm": 0.7556756734848022, + "learning_rate": 8.017009052400295e-06, + "loss": 0.5791, + "num_input_tokens_seen": 44904672, + "step": 29510 + }, + { + "epoch": 0.9447218487932911, + "grad_norm": 0.5094907879829407, + "learning_rate": 8.00980588008557e-06, + "loss": 0.4617, + "num_input_tokens_seen": 44912096, + "step": 29515 + }, + { + "epoch": 0.9448818897637795, + "grad_norm": 0.6658592820167542, + "learning_rate": 8.002605327872282e-06, + "loss": 0.4774, + "num_input_tokens_seen": 44919664, + "step": 29520 + }, + { + "epoch": 0.945041930734268, + "grad_norm": 0.7483572959899902, + "learning_rate": 7.995407396870862e-06, + "loss": 0.4438, + "num_input_tokens_seen": 44927264, + "step": 29525 + }, + { + "epoch": 0.9452019717047564, + "grad_norm": 0.7344165444374084, + "learning_rate": 7.988212088191307e-06, + "loss": 0.4492, + "num_input_tokens_seen": 44935456, + "step": 29530 + }, + { + "epoch": 0.9453620126752449, + "grad_norm": 0.6630484461784363, + "learning_rate": 7.98101940294324e-06, + "loss": 0.3845, + "num_input_tokens_seen": 44942848, + "step": 29535 + }, + { + "epoch": 0.9455220536457333, + "grad_norm": 0.5224781036376953, + "learning_rate": 7.973829342235847e-06, + "loss": 0.3816, + "num_input_tokens_seen": 44950224, + "step": 29540 + }, + { + "epoch": 0.9456820946162218, + "grad_norm": 0.6114745736122131, + "learning_rate": 7.966641907177936e-06, + "loss": 0.3764, + "num_input_tokens_seen": 44957632, + "step": 29545 + }, + { + "epoch": 0.9458421355867102, + "grad_norm": 0.8756417632102966, + "learning_rate": 7.959457098877901e-06, + "loss": 0.6307, + "num_input_tokens_seen": 44965376, + "step": 29550 + }, + { + "epoch": 0.9460021765571986, + "grad_norm": 0.48890557885169983, + "learning_rate": 7.952274918443719e-06, + "loss": 0.4398, + "num_input_tokens_seen": 44972944, + "step": 29555 + }, + { + "epoch": 0.946162217527687, + "grad_norm": 0.7403753399848938, + "learning_rate": 7.945095366982983e-06, + "loss": 0.4763, + "num_input_tokens_seen": 44981120, + "step": 29560 + }, + { + "epoch": 0.9463222584981755, + "grad_norm": 0.8904264569282532, + "learning_rate": 7.937918445602871e-06, + "loss": 0.4713, + "num_input_tokens_seen": 44988768, + "step": 29565 + }, + { + "epoch": 0.946482299468664, + "grad_norm": 0.7461339235305786, + "learning_rate": 7.930744155410145e-06, + "loss": 0.4569, + "num_input_tokens_seen": 44996288, + "step": 29570 + }, + { + "epoch": 0.9466423404391524, + "grad_norm": 1.1346789598464966, + "learning_rate": 7.923572497511181e-06, + "loss": 0.6443, + "num_input_tokens_seen": 45003872, + "step": 29575 + }, + { + "epoch": 0.9468023814096409, + "grad_norm": 0.635968029499054, + "learning_rate": 7.916403473011927e-06, + "loss": 0.4861, + "num_input_tokens_seen": 45011088, + "step": 29580 + }, + { + "epoch": 0.9469624223801293, + "grad_norm": 0.31045883893966675, + "learning_rate": 7.909237083017953e-06, + "loss": 0.3328, + "num_input_tokens_seen": 45018464, + "step": 29585 + }, + { + "epoch": 0.9471224633506178, + "grad_norm": 0.8468340635299683, + "learning_rate": 7.902073328634389e-06, + "loss": 0.445, + "num_input_tokens_seen": 45026064, + "step": 29590 + }, + { + "epoch": 0.9472825043211062, + "grad_norm": 0.5231643319129944, + "learning_rate": 7.894912210965987e-06, + "loss": 0.4209, + "num_input_tokens_seen": 45033472, + "step": 29595 + }, + { + "epoch": 0.9474425452915947, + "grad_norm": 1.0811585187911987, + "learning_rate": 7.887753731117075e-06, + "loss": 0.5661, + "num_input_tokens_seen": 45040832, + "step": 29600 + }, + { + "epoch": 0.9474425452915947, + "eval_loss": 0.49673905968666077, + "eval_runtime": 557.4425, + "eval_samples_per_second": 24.91, + "eval_steps_per_second": 12.455, + "num_input_tokens_seen": 45040832, + "step": 29600 + }, + { + "epoch": 0.947602586262083, + "grad_norm": 0.6516578197479248, + "learning_rate": 7.880597890191587e-06, + "loss": 0.4413, + "num_input_tokens_seen": 45048000, + "step": 29605 + }, + { + "epoch": 0.9477626272325715, + "grad_norm": 0.45249125361442566, + "learning_rate": 7.873444689293036e-06, + "loss": 0.3908, + "num_input_tokens_seen": 45055584, + "step": 29610 + }, + { + "epoch": 0.94792266820306, + "grad_norm": 0.8483229875564575, + "learning_rate": 7.866294129524548e-06, + "loss": 0.4374, + "num_input_tokens_seen": 45063616, + "step": 29615 + }, + { + "epoch": 0.9480827091735484, + "grad_norm": 0.7361921072006226, + "learning_rate": 7.859146211988811e-06, + "loss": 0.493, + "num_input_tokens_seen": 45071120, + "step": 29620 + }, + { + "epoch": 0.9482427501440369, + "grad_norm": 0.5220005512237549, + "learning_rate": 7.852000937788134e-06, + "loss": 0.3477, + "num_input_tokens_seen": 45078544, + "step": 29625 + }, + { + "epoch": 0.9484027911145253, + "grad_norm": 0.5859923958778381, + "learning_rate": 7.844858308024416e-06, + "loss": 0.5188, + "num_input_tokens_seen": 45086160, + "step": 29630 + }, + { + "epoch": 0.9485628320850138, + "grad_norm": 0.5202769637107849, + "learning_rate": 7.837718323799122e-06, + "loss": 0.5151, + "num_input_tokens_seen": 45093568, + "step": 29635 + }, + { + "epoch": 0.9487228730555022, + "grad_norm": 0.8945116400718689, + "learning_rate": 7.83058098621334e-06, + "loss": 0.5762, + "num_input_tokens_seen": 45100832, + "step": 29640 + }, + { + "epoch": 0.9488829140259907, + "grad_norm": 0.7149078249931335, + "learning_rate": 7.823446296367739e-06, + "loss": 0.4951, + "num_input_tokens_seen": 45108384, + "step": 29645 + }, + { + "epoch": 0.9490429549964791, + "grad_norm": 0.6041957139968872, + "learning_rate": 7.81631425536257e-06, + "loss": 0.4312, + "num_input_tokens_seen": 45115840, + "step": 29650 + }, + { + "epoch": 0.9492029959669676, + "grad_norm": 2.245897054672241, + "learning_rate": 7.809184864297689e-06, + "loss": 0.4777, + "num_input_tokens_seen": 45123168, + "step": 29655 + }, + { + "epoch": 0.9493630369374559, + "grad_norm": 0.7073647975921631, + "learning_rate": 7.802058124272532e-06, + "loss": 0.5497, + "num_input_tokens_seen": 45130800, + "step": 29660 + }, + { + "epoch": 0.9495230779079444, + "grad_norm": 0.9350261092185974, + "learning_rate": 7.79493403638614e-06, + "loss": 0.5032, + "num_input_tokens_seen": 45138192, + "step": 29665 + }, + { + "epoch": 0.9496831188784329, + "grad_norm": 0.9251999855041504, + "learning_rate": 7.787812601737132e-06, + "loss": 0.4421, + "num_input_tokens_seen": 45145872, + "step": 29670 + }, + { + "epoch": 0.9498431598489213, + "grad_norm": 0.5106300115585327, + "learning_rate": 7.780693821423715e-06, + "loss": 0.4567, + "num_input_tokens_seen": 45153536, + "step": 29675 + }, + { + "epoch": 0.9500032008194098, + "grad_norm": 0.571054220199585, + "learning_rate": 7.773577696543705e-06, + "loss": 0.4484, + "num_input_tokens_seen": 45161536, + "step": 29680 + }, + { + "epoch": 0.9501632417898982, + "grad_norm": 1.22895348072052, + "learning_rate": 7.7664642281945e-06, + "loss": 0.5908, + "num_input_tokens_seen": 45169344, + "step": 29685 + }, + { + "epoch": 0.9503232827603867, + "grad_norm": 0.6325664520263672, + "learning_rate": 7.759353417473072e-06, + "loss": 0.4429, + "num_input_tokens_seen": 45176768, + "step": 29690 + }, + { + "epoch": 0.9504833237308751, + "grad_norm": 0.8052821159362793, + "learning_rate": 7.752245265476016e-06, + "loss": 0.4623, + "num_input_tokens_seen": 45184032, + "step": 29695 + }, + { + "epoch": 0.9506433647013636, + "grad_norm": 0.7074776291847229, + "learning_rate": 7.745139773299481e-06, + "loss": 0.616, + "num_input_tokens_seen": 45191712, + "step": 29700 + }, + { + "epoch": 0.950803405671852, + "grad_norm": 1.1075705289840698, + "learning_rate": 7.738036942039232e-06, + "loss": 0.5733, + "num_input_tokens_seen": 45199152, + "step": 29705 + }, + { + "epoch": 0.9509634466423404, + "grad_norm": 0.782466471195221, + "learning_rate": 7.73093677279062e-06, + "loss": 0.5477, + "num_input_tokens_seen": 45206752, + "step": 29710 + }, + { + "epoch": 0.9511234876128288, + "grad_norm": 0.8248505592346191, + "learning_rate": 7.72383926664857e-06, + "loss": 0.6104, + "num_input_tokens_seen": 45214096, + "step": 29715 + }, + { + "epoch": 0.9512835285833173, + "grad_norm": 0.7189908623695374, + "learning_rate": 7.716744424707606e-06, + "loss": 0.5666, + "num_input_tokens_seen": 45221968, + "step": 29720 + }, + { + "epoch": 0.9514435695538058, + "grad_norm": 0.47919201850891113, + "learning_rate": 7.709652248061858e-06, + "loss": 0.575, + "num_input_tokens_seen": 45229456, + "step": 29725 + }, + { + "epoch": 0.9516036105242942, + "grad_norm": 0.8278243541717529, + "learning_rate": 7.702562737805017e-06, + "loss": 0.5877, + "num_input_tokens_seen": 45236912, + "step": 29730 + }, + { + "epoch": 0.9517636514947827, + "grad_norm": 0.927588164806366, + "learning_rate": 7.695475895030365e-06, + "loss": 0.4474, + "num_input_tokens_seen": 45244352, + "step": 29735 + }, + { + "epoch": 0.9519236924652711, + "grad_norm": 1.1455960273742676, + "learning_rate": 7.6883917208308e-06, + "loss": 0.6894, + "num_input_tokens_seen": 45251904, + "step": 29740 + }, + { + "epoch": 0.9520837334357596, + "grad_norm": 0.7357909679412842, + "learning_rate": 7.681310216298778e-06, + "loss": 0.3733, + "num_input_tokens_seen": 45259424, + "step": 29745 + }, + { + "epoch": 0.952243774406248, + "grad_norm": 0.5775390863418579, + "learning_rate": 7.674231382526367e-06, + "loss": 0.3535, + "num_input_tokens_seen": 45266512, + "step": 29750 + }, + { + "epoch": 0.9524038153767365, + "grad_norm": 3.0182814598083496, + "learning_rate": 7.667155220605198e-06, + "loss": 0.5689, + "num_input_tokens_seen": 45274096, + "step": 29755 + }, + { + "epoch": 0.9525638563472248, + "grad_norm": 0.7915417551994324, + "learning_rate": 7.660081731626515e-06, + "loss": 0.4881, + "num_input_tokens_seen": 45281216, + "step": 29760 + }, + { + "epoch": 0.9527238973177133, + "grad_norm": 0.7014983892440796, + "learning_rate": 7.653010916681141e-06, + "loss": 0.3576, + "num_input_tokens_seen": 45288400, + "step": 29765 + }, + { + "epoch": 0.9528839382882017, + "grad_norm": 0.5551552176475525, + "learning_rate": 7.645942776859472e-06, + "loss": 0.5985, + "num_input_tokens_seen": 45296176, + "step": 29770 + }, + { + "epoch": 0.9530439792586902, + "grad_norm": 0.5874017477035522, + "learning_rate": 7.63887731325152e-06, + "loss": 0.5617, + "num_input_tokens_seen": 45304160, + "step": 29775 + }, + { + "epoch": 0.9532040202291787, + "grad_norm": 0.38111719489097595, + "learning_rate": 7.63181452694685e-06, + "loss": 0.4312, + "num_input_tokens_seen": 45311776, + "step": 29780 + }, + { + "epoch": 0.9533640611996671, + "grad_norm": 0.74106365442276, + "learning_rate": 7.624754419034644e-06, + "loss": 0.3636, + "num_input_tokens_seen": 45319792, + "step": 29785 + }, + { + "epoch": 0.9535241021701556, + "grad_norm": 0.5610349774360657, + "learning_rate": 7.6176969906036645e-06, + "loss": 0.4307, + "num_input_tokens_seen": 45327936, + "step": 29790 + }, + { + "epoch": 0.953684143140644, + "grad_norm": 0.5430832505226135, + "learning_rate": 7.610642242742242e-06, + "loss": 0.4505, + "num_input_tokens_seen": 45335568, + "step": 29795 + }, + { + "epoch": 0.9538441841111325, + "grad_norm": 1.2468189001083374, + "learning_rate": 7.603590176538322e-06, + "loss": 0.468, + "num_input_tokens_seen": 45343072, + "step": 29800 + }, + { + "epoch": 0.9538441841111325, + "eval_loss": 0.49677562713623047, + "eval_runtime": 557.3666, + "eval_samples_per_second": 24.914, + "eval_steps_per_second": 12.457, + "num_input_tokens_seen": 45343072, + "step": 29800 + }, + { + "epoch": 0.9540042250816209, + "grad_norm": 1.412567138671875, + "learning_rate": 7.596540793079404e-06, + "loss": 0.4634, + "num_input_tokens_seen": 45351168, + "step": 29805 + }, + { + "epoch": 0.9541642660521094, + "grad_norm": 0.7048571705818176, + "learning_rate": 7.5894940934526125e-06, + "loss": 0.5187, + "num_input_tokens_seen": 45358816, + "step": 29810 + }, + { + "epoch": 0.9543243070225977, + "grad_norm": 0.7749959230422974, + "learning_rate": 7.582450078744621e-06, + "loss": 0.4192, + "num_input_tokens_seen": 45366224, + "step": 29815 + }, + { + "epoch": 0.9544843479930862, + "grad_norm": 0.8869418501853943, + "learning_rate": 7.575408750041707e-06, + "loss": 0.5726, + "num_input_tokens_seen": 45373552, + "step": 29820 + }, + { + "epoch": 0.9546443889635747, + "grad_norm": 0.9969812035560608, + "learning_rate": 7.568370108429732e-06, + "loss": 0.5814, + "num_input_tokens_seen": 45380880, + "step": 29825 + }, + { + "epoch": 0.9548044299340631, + "grad_norm": 1.0204596519470215, + "learning_rate": 7.561334154994154e-06, + "loss": 0.5222, + "num_input_tokens_seen": 45388320, + "step": 29830 + }, + { + "epoch": 0.9549644709045516, + "grad_norm": 0.3177267611026764, + "learning_rate": 7.55430089081999e-06, + "loss": 0.4467, + "num_input_tokens_seen": 45395984, + "step": 29835 + }, + { + "epoch": 0.95512451187504, + "grad_norm": 0.6281551122665405, + "learning_rate": 7.547270316991864e-06, + "loss": 0.4323, + "num_input_tokens_seen": 45403760, + "step": 29840 + }, + { + "epoch": 0.9552845528455285, + "grad_norm": 1.0074028968811035, + "learning_rate": 7.5402424345939884e-06, + "loss": 0.5068, + "num_input_tokens_seen": 45411280, + "step": 29845 + }, + { + "epoch": 0.9554445938160169, + "grad_norm": 0.7501891255378723, + "learning_rate": 7.533217244710133e-06, + "loss": 0.4226, + "num_input_tokens_seen": 45418544, + "step": 29850 + }, + { + "epoch": 0.9556046347865054, + "grad_norm": 0.3573954701423645, + "learning_rate": 7.52619474842369e-06, + "loss": 0.4712, + "num_input_tokens_seen": 45425888, + "step": 29855 + }, + { + "epoch": 0.9557646757569938, + "grad_norm": 0.605556845664978, + "learning_rate": 7.519174946817597e-06, + "loss": 0.4238, + "num_input_tokens_seen": 45433680, + "step": 29860 + }, + { + "epoch": 0.9559247167274822, + "grad_norm": 0.5521661043167114, + "learning_rate": 7.512157840974407e-06, + "loss": 0.6489, + "num_input_tokens_seen": 45441056, + "step": 29865 + }, + { + "epoch": 0.9560847576979706, + "grad_norm": 0.8699871897697449, + "learning_rate": 7.5051434319762496e-06, + "loss": 0.4794, + "num_input_tokens_seen": 45448592, + "step": 29870 + }, + { + "epoch": 0.9562447986684591, + "grad_norm": 0.9849047064781189, + "learning_rate": 7.498131720904822e-06, + "loss": 0.529, + "num_input_tokens_seen": 45455888, + "step": 29875 + }, + { + "epoch": 0.9564048396389476, + "grad_norm": 0.4436878561973572, + "learning_rate": 7.491122708841433e-06, + "loss": 0.4279, + "num_input_tokens_seen": 45464096, + "step": 29880 + }, + { + "epoch": 0.956564880609436, + "grad_norm": 0.8613739013671875, + "learning_rate": 7.4841163968669524e-06, + "loss": 0.4391, + "num_input_tokens_seen": 45471552, + "step": 29885 + }, + { + "epoch": 0.9567249215799245, + "grad_norm": 1.0569862127304077, + "learning_rate": 7.4771127860618355e-06, + "loss": 0.4747, + "num_input_tokens_seen": 45479168, + "step": 29890 + }, + { + "epoch": 0.9568849625504129, + "grad_norm": 0.8872355818748474, + "learning_rate": 7.470111877506139e-06, + "loss": 0.3563, + "num_input_tokens_seen": 45487024, + "step": 29895 + }, + { + "epoch": 0.9570450035209014, + "grad_norm": 0.8797841668128967, + "learning_rate": 7.463113672279479e-06, + "loss": 0.5422, + "num_input_tokens_seen": 45494752, + "step": 29900 + }, + { + "epoch": 0.9572050444913898, + "grad_norm": 0.7582326531410217, + "learning_rate": 7.456118171461071e-06, + "loss": 0.4447, + "num_input_tokens_seen": 45502848, + "step": 29905 + }, + { + "epoch": 0.9573650854618783, + "grad_norm": 0.7412664890289307, + "learning_rate": 7.449125376129721e-06, + "loss": 0.5677, + "num_input_tokens_seen": 45510720, + "step": 29910 + }, + { + "epoch": 0.9575251264323666, + "grad_norm": 0.6934544444084167, + "learning_rate": 7.442135287363788e-06, + "loss": 0.497, + "num_input_tokens_seen": 45518496, + "step": 29915 + }, + { + "epoch": 0.9576851674028551, + "grad_norm": 1.0888954401016235, + "learning_rate": 7.435147906241247e-06, + "loss": 0.3437, + "num_input_tokens_seen": 45525968, + "step": 29920 + }, + { + "epoch": 0.9578452083733435, + "grad_norm": 0.6065466403961182, + "learning_rate": 7.428163233839624e-06, + "loss": 0.4797, + "num_input_tokens_seen": 45533312, + "step": 29925 + }, + { + "epoch": 0.958005249343832, + "grad_norm": 0.7683690190315247, + "learning_rate": 7.4211812712360525e-06, + "loss": 0.4627, + "num_input_tokens_seen": 45541392, + "step": 29930 + }, + { + "epoch": 0.9581652903143205, + "grad_norm": 0.6993369460105896, + "learning_rate": 7.4142020195072464e-06, + "loss": 0.5282, + "num_input_tokens_seen": 45548880, + "step": 29935 + }, + { + "epoch": 0.9583253312848089, + "grad_norm": 0.7449917197227478, + "learning_rate": 7.407225479729479e-06, + "loss": 0.4247, + "num_input_tokens_seen": 45556608, + "step": 29940 + }, + { + "epoch": 0.9584853722552974, + "grad_norm": 0.9126495718955994, + "learning_rate": 7.400251652978632e-06, + "loss": 0.6077, + "num_input_tokens_seen": 45564672, + "step": 29945 + }, + { + "epoch": 0.9586454132257858, + "grad_norm": 1.1204715967178345, + "learning_rate": 7.393280540330147e-06, + "loss": 0.5863, + "num_input_tokens_seen": 45572352, + "step": 29950 + }, + { + "epoch": 0.9588054541962743, + "grad_norm": 1.2472206354141235, + "learning_rate": 7.386312142859069e-06, + "loss": 0.4732, + "num_input_tokens_seen": 45580000, + "step": 29955 + }, + { + "epoch": 0.9589654951667627, + "grad_norm": 2.1093366146087646, + "learning_rate": 7.379346461640008e-06, + "loss": 0.6681, + "num_input_tokens_seen": 45587568, + "step": 29960 + }, + { + "epoch": 0.9591255361372512, + "grad_norm": 1.0532042980194092, + "learning_rate": 7.372383497747149e-06, + "loss": 0.5221, + "num_input_tokens_seen": 45595296, + "step": 29965 + }, + { + "epoch": 0.9592855771077395, + "grad_norm": 0.9212677478790283, + "learning_rate": 7.3654232522542775e-06, + "loss": 0.5206, + "num_input_tokens_seen": 45603136, + "step": 29970 + }, + { + "epoch": 0.959445618078228, + "grad_norm": 1.0738414525985718, + "learning_rate": 7.358465726234756e-06, + "loss": 0.3965, + "num_input_tokens_seen": 45610576, + "step": 29975 + }, + { + "epoch": 0.9596056590487164, + "grad_norm": 1.0540610551834106, + "learning_rate": 7.351510920761512e-06, + "loss": 0.4837, + "num_input_tokens_seen": 45618368, + "step": 29980 + }, + { + "epoch": 0.9597657000192049, + "grad_norm": 0.2689904272556305, + "learning_rate": 7.344558836907067e-06, + "loss": 0.4574, + "num_input_tokens_seen": 45626032, + "step": 29985 + }, + { + "epoch": 0.9599257409896934, + "grad_norm": 0.9054502844810486, + "learning_rate": 7.3376094757435285e-06, + "loss": 0.4786, + "num_input_tokens_seen": 45633312, + "step": 29990 + }, + { + "epoch": 0.9600857819601818, + "grad_norm": 0.908074676990509, + "learning_rate": 7.330662838342561e-06, + "loss": 0.4637, + "num_input_tokens_seen": 45640912, + "step": 29995 + }, + { + "epoch": 0.9602458229306703, + "grad_norm": 0.7466791272163391, + "learning_rate": 7.323718925775438e-06, + "loss": 0.3829, + "num_input_tokens_seen": 45648256, + "step": 30000 + }, + { + "epoch": 0.9602458229306703, + "eval_loss": 0.4965097904205322, + "eval_runtime": 557.4758, + "eval_samples_per_second": 24.909, + "eval_steps_per_second": 12.454, + "num_input_tokens_seen": 45648256, + "step": 30000 + }, + { + "epoch": 0.9604058639011587, + "grad_norm": 0.556466281414032, + "learning_rate": 7.316777739112985e-06, + "loss": 0.635, + "num_input_tokens_seen": 45656544, + "step": 30005 + }, + { + "epoch": 0.9605659048716472, + "grad_norm": 0.3171708583831787, + "learning_rate": 7.309839279425626e-06, + "loss": 0.501, + "num_input_tokens_seen": 45664336, + "step": 30010 + }, + { + "epoch": 0.9607259458421356, + "grad_norm": 0.5840229392051697, + "learning_rate": 7.302903547783366e-06, + "loss": 0.4793, + "num_input_tokens_seen": 45672112, + "step": 30015 + }, + { + "epoch": 0.960885986812624, + "grad_norm": 0.6930376887321472, + "learning_rate": 7.2959705452557644e-06, + "loss": 0.5581, + "num_input_tokens_seen": 45679552, + "step": 30020 + }, + { + "epoch": 0.9610460277831124, + "grad_norm": 0.7054030895233154, + "learning_rate": 7.289040272911996e-06, + "loss": 0.5533, + "num_input_tokens_seen": 45686928, + "step": 30025 + }, + { + "epoch": 0.9612060687536009, + "grad_norm": 0.5856296420097351, + "learning_rate": 7.282112731820789e-06, + "loss": 0.5007, + "num_input_tokens_seen": 45694400, + "step": 30030 + }, + { + "epoch": 0.9613661097240893, + "grad_norm": 0.6916295289993286, + "learning_rate": 7.275187923050447e-06, + "loss": 0.4935, + "num_input_tokens_seen": 45701952, + "step": 30035 + }, + { + "epoch": 0.9615261506945778, + "grad_norm": 0.6819443702697754, + "learning_rate": 7.268265847668879e-06, + "loss": 0.4865, + "num_input_tokens_seen": 45709712, + "step": 30040 + }, + { + "epoch": 0.9616861916650663, + "grad_norm": 1.039815902709961, + "learning_rate": 7.261346506743538e-06, + "loss": 0.3523, + "num_input_tokens_seen": 45717200, + "step": 30045 + }, + { + "epoch": 0.9618462326355547, + "grad_norm": 0.5752687454223633, + "learning_rate": 7.254429901341486e-06, + "loss": 0.5138, + "num_input_tokens_seen": 45724608, + "step": 30050 + }, + { + "epoch": 0.9620062736060432, + "grad_norm": 0.7512971758842468, + "learning_rate": 7.247516032529356e-06, + "loss": 0.5063, + "num_input_tokens_seen": 45731824, + "step": 30055 + }, + { + "epoch": 0.9621663145765316, + "grad_norm": 0.6770835518836975, + "learning_rate": 7.240604901373338e-06, + "loss": 0.4186, + "num_input_tokens_seen": 45739728, + "step": 30060 + }, + { + "epoch": 0.9623263555470201, + "grad_norm": 0.6091428995132446, + "learning_rate": 7.233696508939223e-06, + "loss": 0.4383, + "num_input_tokens_seen": 45747168, + "step": 30065 + }, + { + "epoch": 0.9624863965175084, + "grad_norm": 2.0407450199127197, + "learning_rate": 7.226790856292376e-06, + "loss": 0.6406, + "num_input_tokens_seen": 45754464, + "step": 30070 + }, + { + "epoch": 0.9626464374879969, + "grad_norm": 0.7347589135169983, + "learning_rate": 7.219887944497727e-06, + "loss": 0.5195, + "num_input_tokens_seen": 45762144, + "step": 30075 + }, + { + "epoch": 0.9628064784584853, + "grad_norm": 0.7128177881240845, + "learning_rate": 7.2129877746198e-06, + "loss": 0.4499, + "num_input_tokens_seen": 45769952, + "step": 30080 + }, + { + "epoch": 0.9629665194289738, + "grad_norm": 0.919877827167511, + "learning_rate": 7.20609034772268e-06, + "loss": 0.5085, + "num_input_tokens_seen": 45777376, + "step": 30085 + }, + { + "epoch": 0.9631265603994623, + "grad_norm": 0.9155752062797546, + "learning_rate": 7.19919566487004e-06, + "loss": 0.5083, + "num_input_tokens_seen": 45784976, + "step": 30090 + }, + { + "epoch": 0.9632866013699507, + "grad_norm": 0.7429288029670715, + "learning_rate": 7.192303727125132e-06, + "loss": 0.4718, + "num_input_tokens_seen": 45793168, + "step": 30095 + }, + { + "epoch": 0.9634466423404392, + "grad_norm": 0.7566495537757874, + "learning_rate": 7.185414535550777e-06, + "loss": 0.5564, + "num_input_tokens_seen": 45800800, + "step": 30100 + }, + { + "epoch": 0.9636066833109276, + "grad_norm": 1.0924369096755981, + "learning_rate": 7.178528091209363e-06, + "loss": 0.4276, + "num_input_tokens_seen": 45808432, + "step": 30105 + }, + { + "epoch": 0.9637667242814161, + "grad_norm": 1.0013478994369507, + "learning_rate": 7.171644395162888e-06, + "loss": 0.5343, + "num_input_tokens_seen": 45816896, + "step": 30110 + }, + { + "epoch": 0.9639267652519045, + "grad_norm": 0.5731616616249084, + "learning_rate": 7.164763448472881e-06, + "loss": 0.4713, + "num_input_tokens_seen": 45824352, + "step": 30115 + }, + { + "epoch": 0.964086806222393, + "grad_norm": 0.6028057336807251, + "learning_rate": 7.157885252200491e-06, + "loss": 0.5253, + "num_input_tokens_seen": 45831104, + "step": 30120 + }, + { + "epoch": 0.9642468471928813, + "grad_norm": 0.7317838072776794, + "learning_rate": 7.151009807406403e-06, + "loss": 0.4424, + "num_input_tokens_seen": 45838480, + "step": 30125 + }, + { + "epoch": 0.9644068881633698, + "grad_norm": 0.6548643708229065, + "learning_rate": 7.144137115150909e-06, + "loss": 0.375, + "num_input_tokens_seen": 45846128, + "step": 30130 + }, + { + "epoch": 0.9645669291338582, + "grad_norm": 0.9327361583709717, + "learning_rate": 7.1372671764938725e-06, + "loss": 0.5549, + "num_input_tokens_seen": 45854032, + "step": 30135 + }, + { + "epoch": 0.9647269701043467, + "grad_norm": 1.0463697910308838, + "learning_rate": 7.130399992494705e-06, + "loss": 0.4122, + "num_input_tokens_seen": 45861008, + "step": 30140 + }, + { + "epoch": 0.9648870110748352, + "grad_norm": 0.9898335933685303, + "learning_rate": 7.123535564212419e-06, + "loss": 0.3851, + "num_input_tokens_seen": 45868480, + "step": 30145 + }, + { + "epoch": 0.9650470520453236, + "grad_norm": 0.7221065759658813, + "learning_rate": 7.116673892705611e-06, + "loss": 0.413, + "num_input_tokens_seen": 45875552, + "step": 30150 + }, + { + "epoch": 0.9652070930158121, + "grad_norm": 1.1310160160064697, + "learning_rate": 7.109814979032415e-06, + "loss": 0.7399, + "num_input_tokens_seen": 45883312, + "step": 30155 + }, + { + "epoch": 0.9653671339863005, + "grad_norm": 0.46418312191963196, + "learning_rate": 7.102958824250577e-06, + "loss": 0.3562, + "num_input_tokens_seen": 45890976, + "step": 30160 + }, + { + "epoch": 0.965527174956789, + "grad_norm": 1.0772855281829834, + "learning_rate": 7.096105429417393e-06, + "loss": 0.4273, + "num_input_tokens_seen": 45898560, + "step": 30165 + }, + { + "epoch": 0.9656872159272774, + "grad_norm": 0.5926558971405029, + "learning_rate": 7.0892547955897506e-06, + "loss": 0.4807, + "num_input_tokens_seen": 45906352, + "step": 30170 + }, + { + "epoch": 0.9658472568977658, + "grad_norm": 0.8225716948509216, + "learning_rate": 7.0824069238241e-06, + "loss": 0.4842, + "num_input_tokens_seen": 45913264, + "step": 30175 + }, + { + "epoch": 0.9660072978682542, + "grad_norm": 0.46579059958457947, + "learning_rate": 7.075561815176462e-06, + "loss": 0.4172, + "num_input_tokens_seen": 45920880, + "step": 30180 + }, + { + "epoch": 0.9661673388387427, + "grad_norm": 0.3985949456691742, + "learning_rate": 7.068719470702445e-06, + "loss": 0.4803, + "num_input_tokens_seen": 45928688, + "step": 30185 + }, + { + "epoch": 0.9663273798092311, + "grad_norm": 0.580371618270874, + "learning_rate": 7.061879891457229e-06, + "loss": 0.3571, + "num_input_tokens_seen": 45936432, + "step": 30190 + }, + { + "epoch": 0.9664874207797196, + "grad_norm": 0.8447936773300171, + "learning_rate": 7.0550430784955515e-06, + "loss": 0.7106, + "num_input_tokens_seen": 45944096, + "step": 30195 + }, + { + "epoch": 0.9666474617502081, + "grad_norm": 0.9426921010017395, + "learning_rate": 7.048209032871752e-06, + "loss": 0.5386, + "num_input_tokens_seen": 45951600, + "step": 30200 + }, + { + "epoch": 0.9666474617502081, + "eval_loss": 0.4966777563095093, + "eval_runtime": 557.5168, + "eval_samples_per_second": 24.907, + "eval_steps_per_second": 12.453, + "num_input_tokens_seen": 45951600, + "step": 30200 + }, + { + "epoch": 0.9668075027206965, + "grad_norm": 0.576939046382904, + "learning_rate": 7.0413777556397055e-06, + "loss": 0.4791, + "num_input_tokens_seen": 45958880, + "step": 30205 + }, + { + "epoch": 0.966967543691185, + "grad_norm": 0.41991034150123596, + "learning_rate": 7.0345492478528925e-06, + "loss": 0.4696, + "num_input_tokens_seen": 45966320, + "step": 30210 + }, + { + "epoch": 0.9671275846616734, + "grad_norm": 1.012627363204956, + "learning_rate": 7.02772351056436e-06, + "loss": 0.4018, + "num_input_tokens_seen": 45973232, + "step": 30215 + }, + { + "epoch": 0.9672876256321619, + "grad_norm": 0.9915798306465149, + "learning_rate": 7.020900544826709e-06, + "loss": 0.6192, + "num_input_tokens_seen": 45980752, + "step": 30220 + }, + { + "epoch": 0.9674476666026502, + "grad_norm": 1.060697317123413, + "learning_rate": 7.014080351692134e-06, + "loss": 0.3344, + "num_input_tokens_seen": 45988192, + "step": 30225 + }, + { + "epoch": 0.9676077075731387, + "grad_norm": 0.8096356391906738, + "learning_rate": 7.0072629322124024e-06, + "loss": 0.3585, + "num_input_tokens_seen": 45996048, + "step": 30230 + }, + { + "epoch": 0.9677677485436271, + "grad_norm": 0.666655957698822, + "learning_rate": 7.000448287438827e-06, + "loss": 0.5816, + "num_input_tokens_seen": 46003488, + "step": 30235 + }, + { + "epoch": 0.9679277895141156, + "grad_norm": 1.1859153509140015, + "learning_rate": 6.993636418422331e-06, + "loss": 0.5965, + "num_input_tokens_seen": 46010864, + "step": 30240 + }, + { + "epoch": 0.968087830484604, + "grad_norm": 0.3885665535926819, + "learning_rate": 6.986827326213383e-06, + "loss": 0.5131, + "num_input_tokens_seen": 46018448, + "step": 30245 + }, + { + "epoch": 0.9682478714550925, + "grad_norm": 0.8662794232368469, + "learning_rate": 6.9800210118620205e-06, + "loss": 0.4235, + "num_input_tokens_seen": 46026912, + "step": 30250 + }, + { + "epoch": 0.968407912425581, + "grad_norm": 0.9780805706977844, + "learning_rate": 6.973217476417876e-06, + "loss": 0.5255, + "num_input_tokens_seen": 46034832, + "step": 30255 + }, + { + "epoch": 0.9685679533960694, + "grad_norm": 1.0257092714309692, + "learning_rate": 6.96641672093013e-06, + "loss": 0.546, + "num_input_tokens_seen": 46042336, + "step": 30260 + }, + { + "epoch": 0.9687279943665579, + "grad_norm": 0.6149593591690063, + "learning_rate": 6.95961874644755e-06, + "loss": 0.6064, + "num_input_tokens_seen": 46049552, + "step": 30265 + }, + { + "epoch": 0.9688880353370463, + "grad_norm": 0.9547048807144165, + "learning_rate": 6.952823554018476e-06, + "loss": 0.3527, + "num_input_tokens_seen": 46057088, + "step": 30270 + }, + { + "epoch": 0.9690480763075348, + "grad_norm": 0.5116984844207764, + "learning_rate": 6.946031144690798e-06, + "loss": 0.3668, + "num_input_tokens_seen": 46064672, + "step": 30275 + }, + { + "epoch": 0.9692081172780231, + "grad_norm": 0.4509184956550598, + "learning_rate": 6.939241519512005e-06, + "loss": 0.3864, + "num_input_tokens_seen": 46072192, + "step": 30280 + }, + { + "epoch": 0.9693681582485116, + "grad_norm": 0.8726211190223694, + "learning_rate": 6.932454679529129e-06, + "loss": 0.5673, + "num_input_tokens_seen": 46079424, + "step": 30285 + }, + { + "epoch": 0.969528199219, + "grad_norm": 0.79273521900177, + "learning_rate": 6.925670625788791e-06, + "loss": 0.4195, + "num_input_tokens_seen": 46087072, + "step": 30290 + }, + { + "epoch": 0.9696882401894885, + "grad_norm": 0.5323306322097778, + "learning_rate": 6.918889359337186e-06, + "loss": 0.4672, + "num_input_tokens_seen": 46094768, + "step": 30295 + }, + { + "epoch": 0.9698482811599769, + "grad_norm": 0.7106114029884338, + "learning_rate": 6.912110881220058e-06, + "loss": 0.3851, + "num_input_tokens_seen": 46102256, + "step": 30300 + }, + { + "epoch": 0.9700083221304654, + "grad_norm": 0.6109369993209839, + "learning_rate": 6.905335192482735e-06, + "loss": 0.4312, + "num_input_tokens_seen": 46110032, + "step": 30305 + }, + { + "epoch": 0.9701683631009539, + "grad_norm": 0.9266055226325989, + "learning_rate": 6.8985622941701275e-06, + "loss": 0.492, + "num_input_tokens_seen": 46117664, + "step": 30310 + }, + { + "epoch": 0.9703284040714423, + "grad_norm": 0.743146538734436, + "learning_rate": 6.89179218732669e-06, + "loss": 0.5187, + "num_input_tokens_seen": 46124736, + "step": 30315 + }, + { + "epoch": 0.9704884450419308, + "grad_norm": 0.5442156791687012, + "learning_rate": 6.8850248729964595e-06, + "loss": 0.4138, + "num_input_tokens_seen": 46132160, + "step": 30320 + }, + { + "epoch": 0.9706484860124192, + "grad_norm": 1.088718295097351, + "learning_rate": 6.8782603522230314e-06, + "loss": 0.3571, + "num_input_tokens_seen": 46139440, + "step": 30325 + }, + { + "epoch": 0.9708085269829076, + "grad_norm": 1.1042815446853638, + "learning_rate": 6.871498626049591e-06, + "loss": 0.2552, + "num_input_tokens_seen": 46146944, + "step": 30330 + }, + { + "epoch": 0.970968567953396, + "grad_norm": 1.3383787870407104, + "learning_rate": 6.8647396955188875e-06, + "loss": 0.5588, + "num_input_tokens_seen": 46153952, + "step": 30335 + }, + { + "epoch": 0.9711286089238845, + "grad_norm": 0.9320757389068604, + "learning_rate": 6.857983561673218e-06, + "loss": 0.4007, + "num_input_tokens_seen": 46161360, + "step": 30340 + }, + { + "epoch": 0.9712886498943729, + "grad_norm": 0.6272006034851074, + "learning_rate": 6.851230225554467e-06, + "loss": 0.4395, + "num_input_tokens_seen": 46168896, + "step": 30345 + }, + { + "epoch": 0.9714486908648614, + "grad_norm": 0.9039273262023926, + "learning_rate": 6.8444796882040946e-06, + "loss": 0.5155, + "num_input_tokens_seen": 46176352, + "step": 30350 + }, + { + "epoch": 0.9716087318353499, + "grad_norm": 0.8537362217903137, + "learning_rate": 6.837731950663106e-06, + "loss": 0.3575, + "num_input_tokens_seen": 46184000, + "step": 30355 + }, + { + "epoch": 0.9717687728058383, + "grad_norm": 0.3579242527484894, + "learning_rate": 6.830987013972098e-06, + "loss": 0.3931, + "num_input_tokens_seen": 46191616, + "step": 30360 + }, + { + "epoch": 0.9719288137763268, + "grad_norm": 0.6908368468284607, + "learning_rate": 6.82424487917121e-06, + "loss": 0.7025, + "num_input_tokens_seen": 46198656, + "step": 30365 + }, + { + "epoch": 0.9720888547468152, + "grad_norm": 0.8128472566604614, + "learning_rate": 6.8175055473001735e-06, + "loss": 0.5391, + "num_input_tokens_seen": 46206736, + "step": 30370 + }, + { + "epoch": 0.9722488957173037, + "grad_norm": 1.14053213596344, + "learning_rate": 6.8107690193982855e-06, + "loss": 0.4195, + "num_input_tokens_seen": 46214384, + "step": 30375 + }, + { + "epoch": 0.972408936687792, + "grad_norm": 0.5701867341995239, + "learning_rate": 6.804035296504385e-06, + "loss": 0.3056, + "num_input_tokens_seen": 46221936, + "step": 30380 + }, + { + "epoch": 0.9725689776582805, + "grad_norm": 0.6113330721855164, + "learning_rate": 6.797304379656916e-06, + "loss": 0.2809, + "num_input_tokens_seen": 46229328, + "step": 30385 + }, + { + "epoch": 0.9727290186287689, + "grad_norm": 0.9463574290275574, + "learning_rate": 6.790576269893861e-06, + "loss": 0.5778, + "num_input_tokens_seen": 46237040, + "step": 30390 + }, + { + "epoch": 0.9728890595992574, + "grad_norm": 0.6727426052093506, + "learning_rate": 6.783850968252772e-06, + "loss": 0.3913, + "num_input_tokens_seen": 46244848, + "step": 30395 + }, + { + "epoch": 0.9730491005697458, + "grad_norm": 0.7674655914306641, + "learning_rate": 6.777128475770789e-06, + "loss": 0.4994, + "num_input_tokens_seen": 46252880, + "step": 30400 + }, + { + "epoch": 0.9730491005697458, + "eval_loss": 0.4964067041873932, + "eval_runtime": 557.3665, + "eval_samples_per_second": 24.914, + "eval_steps_per_second": 12.457, + "num_input_tokens_seen": 46252880, + "step": 30400 + }, + { + "epoch": 0.9732091415402343, + "grad_norm": 0.7059398293495178, + "learning_rate": 6.77040879348459e-06, + "loss": 0.4803, + "num_input_tokens_seen": 46260400, + "step": 30405 + }, + { + "epoch": 0.9733691825107228, + "grad_norm": 0.4056241512298584, + "learning_rate": 6.763691922430443e-06, + "loss": 0.4171, + "num_input_tokens_seen": 46268288, + "step": 30410 + }, + { + "epoch": 0.9735292234812112, + "grad_norm": 0.677981436252594, + "learning_rate": 6.756977863644178e-06, + "loss": 0.5389, + "num_input_tokens_seen": 46276064, + "step": 30415 + }, + { + "epoch": 0.9736892644516997, + "grad_norm": 0.6719782948493958, + "learning_rate": 6.7502666181611804e-06, + "loss": 0.3704, + "num_input_tokens_seen": 46283472, + "step": 30420 + }, + { + "epoch": 0.9738493054221881, + "grad_norm": 0.8367710709571838, + "learning_rate": 6.743558187016405e-06, + "loss": 0.573, + "num_input_tokens_seen": 46291904, + "step": 30425 + }, + { + "epoch": 0.9740093463926766, + "grad_norm": 0.43815022706985474, + "learning_rate": 6.7368525712443925e-06, + "loss": 0.526, + "num_input_tokens_seen": 46299216, + "step": 30430 + }, + { + "epoch": 0.9741693873631649, + "grad_norm": 0.48055922985076904, + "learning_rate": 6.7301497718792155e-06, + "loss": 0.6054, + "num_input_tokens_seen": 46306832, + "step": 30435 + }, + { + "epoch": 0.9743294283336534, + "grad_norm": 0.8081449270248413, + "learning_rate": 6.723449789954544e-06, + "loss": 0.513, + "num_input_tokens_seen": 46314704, + "step": 30440 + }, + { + "epoch": 0.9744894693041418, + "grad_norm": 0.9237060546875, + "learning_rate": 6.716752626503586e-06, + "loss": 0.5072, + "num_input_tokens_seen": 46322272, + "step": 30445 + }, + { + "epoch": 0.9746495102746303, + "grad_norm": 1.3157522678375244, + "learning_rate": 6.710058282559131e-06, + "loss": 0.4973, + "num_input_tokens_seen": 46329888, + "step": 30450 + }, + { + "epoch": 0.9748095512451187, + "grad_norm": 0.7586143612861633, + "learning_rate": 6.703366759153545e-06, + "loss": 0.5357, + "num_input_tokens_seen": 46337280, + "step": 30455 + }, + { + "epoch": 0.9749695922156072, + "grad_norm": 0.6510346531867981, + "learning_rate": 6.6966780573187335e-06, + "loss": 0.349, + "num_input_tokens_seen": 46345008, + "step": 30460 + }, + { + "epoch": 0.9751296331860957, + "grad_norm": 0.8004162907600403, + "learning_rate": 6.689992178086174e-06, + "loss": 0.4649, + "num_input_tokens_seen": 46352432, + "step": 30465 + }, + { + "epoch": 0.9752896741565841, + "grad_norm": 0.6041718125343323, + "learning_rate": 6.683309122486925e-06, + "loss": 0.3778, + "num_input_tokens_seen": 46359776, + "step": 30470 + }, + { + "epoch": 0.9754497151270726, + "grad_norm": 1.055996060371399, + "learning_rate": 6.676628891551584e-06, + "loss": 0.5204, + "num_input_tokens_seen": 46367136, + "step": 30475 + }, + { + "epoch": 0.975609756097561, + "grad_norm": 0.6700268387794495, + "learning_rate": 6.6699514863103385e-06, + "loss": 0.4347, + "num_input_tokens_seen": 46374880, + "step": 30480 + }, + { + "epoch": 0.9757697970680494, + "grad_norm": 0.7534910440444946, + "learning_rate": 6.663276907792921e-06, + "loss": 0.5403, + "num_input_tokens_seen": 46382432, + "step": 30485 + }, + { + "epoch": 0.9759298380385378, + "grad_norm": 0.652637779712677, + "learning_rate": 6.656605157028634e-06, + "loss": 0.4115, + "num_input_tokens_seen": 46389792, + "step": 30490 + }, + { + "epoch": 0.9760898790090263, + "grad_norm": 0.898108959197998, + "learning_rate": 6.649936235046358e-06, + "loss": 0.4326, + "num_input_tokens_seen": 46397152, + "step": 30495 + }, + { + "epoch": 0.9762499199795147, + "grad_norm": 0.7656843066215515, + "learning_rate": 6.643270142874508e-06, + "loss": 0.3498, + "num_input_tokens_seen": 46404576, + "step": 30500 + }, + { + "epoch": 0.9764099609500032, + "grad_norm": 1.573516845703125, + "learning_rate": 6.636606881541094e-06, + "loss": 0.7105, + "num_input_tokens_seen": 46411712, + "step": 30505 + }, + { + "epoch": 0.9765700019204916, + "grad_norm": 0.799370527267456, + "learning_rate": 6.629946452073662e-06, + "loss": 0.7144, + "num_input_tokens_seen": 46419232, + "step": 30510 + }, + { + "epoch": 0.9767300428909801, + "grad_norm": 0.4324061870574951, + "learning_rate": 6.6232888554993375e-06, + "loss": 0.3284, + "num_input_tokens_seen": 46426496, + "step": 30515 + }, + { + "epoch": 0.9768900838614686, + "grad_norm": 0.4671030342578888, + "learning_rate": 6.616634092844817e-06, + "loss": 0.4015, + "num_input_tokens_seen": 46434096, + "step": 30520 + }, + { + "epoch": 0.977050124831957, + "grad_norm": 0.4761449098587036, + "learning_rate": 6.609982165136331e-06, + "loss": 0.6388, + "num_input_tokens_seen": 46441232, + "step": 30525 + }, + { + "epoch": 0.9772101658024455, + "grad_norm": 0.8976149559020996, + "learning_rate": 6.603333073399706e-06, + "loss": 0.5538, + "num_input_tokens_seen": 46449056, + "step": 30530 + }, + { + "epoch": 0.9773702067729338, + "grad_norm": 1.6047332286834717, + "learning_rate": 6.596686818660308e-06, + "loss": 0.4384, + "num_input_tokens_seen": 46456896, + "step": 30535 + }, + { + "epoch": 0.9775302477434223, + "grad_norm": 0.6116315126419067, + "learning_rate": 6.590043401943066e-06, + "loss": 0.3951, + "num_input_tokens_seen": 46464816, + "step": 30540 + }, + { + "epoch": 0.9776902887139107, + "grad_norm": 0.43427422642707825, + "learning_rate": 6.583402824272494e-06, + "loss": 0.4126, + "num_input_tokens_seen": 46472688, + "step": 30545 + }, + { + "epoch": 0.9778503296843992, + "grad_norm": 0.3750711679458618, + "learning_rate": 6.576765086672634e-06, + "loss": 0.4028, + "num_input_tokens_seen": 46480368, + "step": 30550 + }, + { + "epoch": 0.9780103706548876, + "grad_norm": 0.6632510423660278, + "learning_rate": 6.57013019016712e-06, + "loss": 0.5077, + "num_input_tokens_seen": 46487920, + "step": 30555 + }, + { + "epoch": 0.9781704116253761, + "grad_norm": 1.1216686964035034, + "learning_rate": 6.563498135779142e-06, + "loss": 0.3978, + "num_input_tokens_seen": 46496512, + "step": 30560 + }, + { + "epoch": 0.9783304525958645, + "grad_norm": 0.8793087601661682, + "learning_rate": 6.556868924531431e-06, + "loss": 0.3915, + "num_input_tokens_seen": 46504080, + "step": 30565 + }, + { + "epoch": 0.978490493566353, + "grad_norm": 0.46671420335769653, + "learning_rate": 6.550242557446304e-06, + "loss": 0.5133, + "num_input_tokens_seen": 46511824, + "step": 30570 + }, + { + "epoch": 0.9786505345368415, + "grad_norm": 1.2822400331497192, + "learning_rate": 6.543619035545634e-06, + "loss": 0.4325, + "num_input_tokens_seen": 46519072, + "step": 30575 + }, + { + "epoch": 0.9788105755073299, + "grad_norm": 0.9582037329673767, + "learning_rate": 6.53699835985084e-06, + "loss": 0.4892, + "num_input_tokens_seen": 46526496, + "step": 30580 + }, + { + "epoch": 0.9789706164778184, + "grad_norm": 1.1921606063842773, + "learning_rate": 6.530380531382927e-06, + "loss": 0.5357, + "num_input_tokens_seen": 46533888, + "step": 30585 + }, + { + "epoch": 0.9791306574483067, + "grad_norm": 0.9258164167404175, + "learning_rate": 6.523765551162433e-06, + "loss": 0.5356, + "num_input_tokens_seen": 46541264, + "step": 30590 + }, + { + "epoch": 0.9792906984187952, + "grad_norm": 0.9984666705131531, + "learning_rate": 6.517153420209476e-06, + "loss": 0.482, + "num_input_tokens_seen": 46549056, + "step": 30595 + }, + { + "epoch": 0.9794507393892836, + "grad_norm": 1.2347322702407837, + "learning_rate": 6.510544139543739e-06, + "loss": 0.5633, + "num_input_tokens_seen": 46556736, + "step": 30600 + }, + { + "epoch": 0.9794507393892836, + "eval_loss": 0.4962530732154846, + "eval_runtime": 557.613, + "eval_samples_per_second": 24.903, + "eval_steps_per_second": 12.451, + "num_input_tokens_seen": 46556736, + "step": 30600 + }, + { + "epoch": 0.9796107803597721, + "grad_norm": 0.928367555141449, + "learning_rate": 6.503937710184452e-06, + "loss": 0.5935, + "num_input_tokens_seen": 46564240, + "step": 30605 + }, + { + "epoch": 0.9797708213302605, + "grad_norm": 0.5841366648674011, + "learning_rate": 6.4973341331503954e-06, + "loss": 0.5981, + "num_input_tokens_seen": 46571600, + "step": 30610 + }, + { + "epoch": 0.979930862300749, + "grad_norm": 0.3601321578025818, + "learning_rate": 6.490733409459942e-06, + "loss": 0.3298, + "num_input_tokens_seen": 46579216, + "step": 30615 + }, + { + "epoch": 0.9800909032712375, + "grad_norm": 1.0215977430343628, + "learning_rate": 6.484135540130995e-06, + "loss": 0.4444, + "num_input_tokens_seen": 46587120, + "step": 30620 + }, + { + "epoch": 0.9802509442417259, + "grad_norm": 0.7029495239257812, + "learning_rate": 6.4775405261810364e-06, + "loss": 0.4337, + "num_input_tokens_seen": 46595040, + "step": 30625 + }, + { + "epoch": 0.9804109852122144, + "grad_norm": 0.5363652110099792, + "learning_rate": 6.470948368627092e-06, + "loss": 0.3896, + "num_input_tokens_seen": 46602800, + "step": 30630 + }, + { + "epoch": 0.9805710261827028, + "grad_norm": 0.5093280673027039, + "learning_rate": 6.464359068485756e-06, + "loss": 0.4048, + "num_input_tokens_seen": 46609856, + "step": 30635 + }, + { + "epoch": 0.9807310671531912, + "grad_norm": 0.7123279571533203, + "learning_rate": 6.457772626773195e-06, + "loss": 0.4967, + "num_input_tokens_seen": 46617504, + "step": 30640 + }, + { + "epoch": 0.9808911081236796, + "grad_norm": 0.60869961977005, + "learning_rate": 6.451189044505104e-06, + "loss": 0.5684, + "num_input_tokens_seen": 46624752, + "step": 30645 + }, + { + "epoch": 0.9810511490941681, + "grad_norm": 0.6498215794563293, + "learning_rate": 6.44460832269676e-06, + "loss": 0.4816, + "num_input_tokens_seen": 46632400, + "step": 30650 + }, + { + "epoch": 0.9812111900646565, + "grad_norm": 0.7313354015350342, + "learning_rate": 6.438030462363001e-06, + "loss": 0.4557, + "num_input_tokens_seen": 46639616, + "step": 30655 + }, + { + "epoch": 0.981371231035145, + "grad_norm": 1.025875449180603, + "learning_rate": 6.431455464518205e-06, + "loss": 0.4867, + "num_input_tokens_seen": 46647168, + "step": 30660 + }, + { + "epoch": 0.9815312720056334, + "grad_norm": 0.9564809203147888, + "learning_rate": 6.424883330176326e-06, + "loss": 0.2869, + "num_input_tokens_seen": 46654688, + "step": 30665 + }, + { + "epoch": 0.9816913129761219, + "grad_norm": 0.7570093870162964, + "learning_rate": 6.418314060350864e-06, + "loss": 0.6698, + "num_input_tokens_seen": 46662208, + "step": 30670 + }, + { + "epoch": 0.9818513539466104, + "grad_norm": 1.1689671277999878, + "learning_rate": 6.4117476560548895e-06, + "loss": 0.4792, + "num_input_tokens_seen": 46670160, + "step": 30675 + }, + { + "epoch": 0.9820113949170988, + "grad_norm": 0.9046589136123657, + "learning_rate": 6.405184118301016e-06, + "loss": 0.671, + "num_input_tokens_seen": 46677536, + "step": 30680 + }, + { + "epoch": 0.9821714358875873, + "grad_norm": 0.5210297703742981, + "learning_rate": 6.398623448101434e-06, + "loss": 0.423, + "num_input_tokens_seen": 46685168, + "step": 30685 + }, + { + "epoch": 0.9823314768580756, + "grad_norm": 0.3176429867744446, + "learning_rate": 6.392065646467871e-06, + "loss": 0.2553, + "num_input_tokens_seen": 46692208, + "step": 30690 + }, + { + "epoch": 0.9824915178285641, + "grad_norm": 0.7804811596870422, + "learning_rate": 6.385510714411632e-06, + "loss": 0.5159, + "num_input_tokens_seen": 46699584, + "step": 30695 + }, + { + "epoch": 0.9826515587990525, + "grad_norm": 0.7000617980957031, + "learning_rate": 6.378958652943559e-06, + "loss": 0.4829, + "num_input_tokens_seen": 46706928, + "step": 30700 + }, + { + "epoch": 0.982811599769541, + "grad_norm": 1.148136019706726, + "learning_rate": 6.3724094630740776e-06, + "loss": 0.5481, + "num_input_tokens_seen": 46714656, + "step": 30705 + }, + { + "epoch": 0.9829716407400294, + "grad_norm": 0.8307471871376038, + "learning_rate": 6.365863145813136e-06, + "loss": 0.4686, + "num_input_tokens_seen": 46722208, + "step": 30710 + }, + { + "epoch": 0.9831316817105179, + "grad_norm": 0.5772973299026489, + "learning_rate": 6.359319702170269e-06, + "loss": 0.4361, + "num_input_tokens_seen": 46729600, + "step": 30715 + }, + { + "epoch": 0.9832917226810063, + "grad_norm": 1.1412677764892578, + "learning_rate": 6.352779133154566e-06, + "loss": 0.5567, + "num_input_tokens_seen": 46737152, + "step": 30720 + }, + { + "epoch": 0.9834517636514948, + "grad_norm": 0.9687873125076294, + "learning_rate": 6.346241439774648e-06, + "loss": 0.4705, + "num_input_tokens_seen": 46744416, + "step": 30725 + }, + { + "epoch": 0.9836118046219833, + "grad_norm": 1.2870235443115234, + "learning_rate": 6.339706623038716e-06, + "loss": 0.4098, + "num_input_tokens_seen": 46751952, + "step": 30730 + }, + { + "epoch": 0.9837718455924717, + "grad_norm": 0.7850092053413391, + "learning_rate": 6.333174683954532e-06, + "loss": 0.4779, + "num_input_tokens_seen": 46759632, + "step": 30735 + }, + { + "epoch": 0.9839318865629602, + "grad_norm": 1.1113094091415405, + "learning_rate": 6.326645623529387e-06, + "loss": 0.5237, + "num_input_tokens_seen": 46767360, + "step": 30740 + }, + { + "epoch": 0.9840919275334485, + "grad_norm": 0.5258538722991943, + "learning_rate": 6.320119442770156e-06, + "loss": 0.3148, + "num_input_tokens_seen": 46774800, + "step": 30745 + }, + { + "epoch": 0.984251968503937, + "grad_norm": 0.3774808645248413, + "learning_rate": 6.313596142683254e-06, + "loss": 0.3113, + "num_input_tokens_seen": 46782512, + "step": 30750 + }, + { + "epoch": 0.9844120094744254, + "grad_norm": 0.808917224407196, + "learning_rate": 6.307075724274647e-06, + "loss": 0.3083, + "num_input_tokens_seen": 46790160, + "step": 30755 + }, + { + "epoch": 0.9845720504449139, + "grad_norm": 1.0330173969268799, + "learning_rate": 6.300558188549882e-06, + "loss": 0.711, + "num_input_tokens_seen": 46797792, + "step": 30760 + }, + { + "epoch": 0.9847320914154023, + "grad_norm": 1.103134274482727, + "learning_rate": 6.29404353651403e-06, + "loss": 0.4618, + "num_input_tokens_seen": 46805152, + "step": 30765 + }, + { + "epoch": 0.9848921323858908, + "grad_norm": 0.9342790842056274, + "learning_rate": 6.287531769171737e-06, + "loss": 0.4577, + "num_input_tokens_seen": 46812560, + "step": 30770 + }, + { + "epoch": 0.9850521733563792, + "grad_norm": 0.8365934491157532, + "learning_rate": 6.2810228875272045e-06, + "loss": 0.5004, + "num_input_tokens_seen": 46819872, + "step": 30775 + }, + { + "epoch": 0.9852122143268677, + "grad_norm": 1.4884332418441772, + "learning_rate": 6.274516892584179e-06, + "loss": 0.5676, + "num_input_tokens_seen": 46827360, + "step": 30780 + }, + { + "epoch": 0.9853722552973562, + "grad_norm": 2.2342066764831543, + "learning_rate": 6.268013785345969e-06, + "loss": 0.4653, + "num_input_tokens_seen": 46835296, + "step": 30785 + }, + { + "epoch": 0.9855322962678446, + "grad_norm": 0.4596932530403137, + "learning_rate": 6.26151356681543e-06, + "loss": 0.3636, + "num_input_tokens_seen": 46842656, + "step": 30790 + }, + { + "epoch": 0.985692337238333, + "grad_norm": 0.6068748831748962, + "learning_rate": 6.255016237994981e-06, + "loss": 0.4919, + "num_input_tokens_seen": 46850256, + "step": 30795 + }, + { + "epoch": 0.9858523782088214, + "grad_norm": 0.5698699355125427, + "learning_rate": 6.248521799886603e-06, + "loss": 0.6597, + "num_input_tokens_seen": 46858080, + "step": 30800 + }, + { + "epoch": 0.9858523782088214, + "eval_loss": 0.4962630271911621, + "eval_runtime": 557.4531, + "eval_samples_per_second": 24.91, + "eval_steps_per_second": 12.455, + "num_input_tokens_seen": 46858080, + "step": 30800 + }, + { + "epoch": 0.9860124191793099, + "grad_norm": 1.0866155624389648, + "learning_rate": 6.242030253491798e-06, + "loss": 0.3735, + "num_input_tokens_seen": 46865936, + "step": 30805 + }, + { + "epoch": 0.9861724601497983, + "grad_norm": 0.7663465738296509, + "learning_rate": 6.235541599811656e-06, + "loss": 0.4831, + "num_input_tokens_seen": 46873424, + "step": 30810 + }, + { + "epoch": 0.9863325011202868, + "grad_norm": 0.9046218991279602, + "learning_rate": 6.229055839846814e-06, + "loss": 0.5041, + "num_input_tokens_seen": 46880784, + "step": 30815 + }, + { + "epoch": 0.9864925420907752, + "grad_norm": 0.6378922462463379, + "learning_rate": 6.222572974597455e-06, + "loss": 0.3255, + "num_input_tokens_seen": 46888096, + "step": 30820 + }, + { + "epoch": 0.9866525830612637, + "grad_norm": 1.3857923746109009, + "learning_rate": 6.216093005063306e-06, + "loss": 0.6258, + "num_input_tokens_seen": 46895856, + "step": 30825 + }, + { + "epoch": 0.9868126240317522, + "grad_norm": 0.5752881765365601, + "learning_rate": 6.209615932243678e-06, + "loss": 0.3862, + "num_input_tokens_seen": 46903600, + "step": 30830 + }, + { + "epoch": 0.9869726650022406, + "grad_norm": 0.5217725038528442, + "learning_rate": 6.203141757137399e-06, + "loss": 0.3875, + "num_input_tokens_seen": 46911360, + "step": 30835 + }, + { + "epoch": 0.9871327059727291, + "grad_norm": 0.4695779085159302, + "learning_rate": 6.196670480742886e-06, + "loss": 0.3369, + "num_input_tokens_seen": 46918992, + "step": 30840 + }, + { + "epoch": 0.9872927469432174, + "grad_norm": 0.8215304017066956, + "learning_rate": 6.190202104058074e-06, + "loss": 0.4034, + "num_input_tokens_seen": 46926480, + "step": 30845 + }, + { + "epoch": 0.9874527879137059, + "grad_norm": 0.7261906266212463, + "learning_rate": 6.183736628080475e-06, + "loss": 0.3765, + "num_input_tokens_seen": 46934384, + "step": 30850 + }, + { + "epoch": 0.9876128288841943, + "grad_norm": 0.9563301801681519, + "learning_rate": 6.177274053807155e-06, + "loss": 0.551, + "num_input_tokens_seen": 46941632, + "step": 30855 + }, + { + "epoch": 0.9877728698546828, + "grad_norm": 0.5783276557922363, + "learning_rate": 6.170814382234713e-06, + "loss": 0.5826, + "num_input_tokens_seen": 46949296, + "step": 30860 + }, + { + "epoch": 0.9879329108251712, + "grad_norm": 1.281753420829773, + "learning_rate": 6.16435761435932e-06, + "loss": 0.7042, + "num_input_tokens_seen": 46956928, + "step": 30865 + }, + { + "epoch": 0.9880929517956597, + "grad_norm": 0.8516662120819092, + "learning_rate": 6.157903751176681e-06, + "loss": 0.4277, + "num_input_tokens_seen": 46964368, + "step": 30870 + }, + { + "epoch": 0.9882529927661481, + "grad_norm": 0.34893524646759033, + "learning_rate": 6.151452793682066e-06, + "loss": 0.4082, + "num_input_tokens_seen": 46971904, + "step": 30875 + }, + { + "epoch": 0.9884130337366366, + "grad_norm": 0.5805231332778931, + "learning_rate": 6.145004742870305e-06, + "loss": 0.5032, + "num_input_tokens_seen": 46979872, + "step": 30880 + }, + { + "epoch": 0.9885730747071251, + "grad_norm": 0.950454831123352, + "learning_rate": 6.138559599735752e-06, + "loss": 0.5949, + "num_input_tokens_seen": 46987232, + "step": 30885 + }, + { + "epoch": 0.9887331156776135, + "grad_norm": 0.8835535645484924, + "learning_rate": 6.132117365272344e-06, + "loss": 0.4237, + "num_input_tokens_seen": 46994864, + "step": 30890 + }, + { + "epoch": 0.988893156648102, + "grad_norm": 0.5826516151428223, + "learning_rate": 6.125678040473545e-06, + "loss": 0.4422, + "num_input_tokens_seen": 47002688, + "step": 30895 + }, + { + "epoch": 0.9890531976185903, + "grad_norm": 1.0172287225723267, + "learning_rate": 6.1192416263323755e-06, + "loss": 0.4449, + "num_input_tokens_seen": 47010592, + "step": 30900 + }, + { + "epoch": 0.9892132385890788, + "grad_norm": 0.9783677458763123, + "learning_rate": 6.112808123841424e-06, + "loss": 0.379, + "num_input_tokens_seen": 47018240, + "step": 30905 + }, + { + "epoch": 0.9893732795595672, + "grad_norm": 1.4780596494674683, + "learning_rate": 6.106377533992805e-06, + "loss": 0.5949, + "num_input_tokens_seen": 47025744, + "step": 30910 + }, + { + "epoch": 0.9895333205300557, + "grad_norm": 0.4713309109210968, + "learning_rate": 6.099949857778204e-06, + "loss": 0.3675, + "num_input_tokens_seen": 47033472, + "step": 30915 + }, + { + "epoch": 0.9896933615005441, + "grad_norm": 1.5806941986083984, + "learning_rate": 6.093525096188852e-06, + "loss": 0.5737, + "num_input_tokens_seen": 47040880, + "step": 30920 + }, + { + "epoch": 0.9898534024710326, + "grad_norm": 1.428882360458374, + "learning_rate": 6.087103250215518e-06, + "loss": 0.4896, + "num_input_tokens_seen": 47048352, + "step": 30925 + }, + { + "epoch": 0.990013443441521, + "grad_norm": 0.5489695072174072, + "learning_rate": 6.080684320848537e-06, + "loss": 0.4246, + "num_input_tokens_seen": 47055968, + "step": 30930 + }, + { + "epoch": 0.9901734844120095, + "grad_norm": 1.5872398614883423, + "learning_rate": 6.074268309077794e-06, + "loss": 0.5617, + "num_input_tokens_seen": 47063840, + "step": 30935 + }, + { + "epoch": 0.990333525382498, + "grad_norm": 0.494672954082489, + "learning_rate": 6.067855215892709e-06, + "loss": 0.4106, + "num_input_tokens_seen": 47071584, + "step": 30940 + }, + { + "epoch": 0.9904935663529864, + "grad_norm": 0.8765579462051392, + "learning_rate": 6.061445042282271e-06, + "loss": 0.6403, + "num_input_tokens_seen": 47079280, + "step": 30945 + }, + { + "epoch": 0.9906536073234748, + "grad_norm": 1.1827213764190674, + "learning_rate": 6.055037789234999e-06, + "loss": 0.5465, + "num_input_tokens_seen": 47087392, + "step": 30950 + }, + { + "epoch": 0.9908136482939632, + "grad_norm": 0.8438085317611694, + "learning_rate": 6.048633457738975e-06, + "loss": 0.7113, + "num_input_tokens_seen": 47095168, + "step": 30955 + }, + { + "epoch": 0.9909736892644517, + "grad_norm": 1.3424897193908691, + "learning_rate": 6.042232048781837e-06, + "loss": 0.4941, + "num_input_tokens_seen": 47102704, + "step": 30960 + }, + { + "epoch": 0.9911337302349401, + "grad_norm": 0.7349775433540344, + "learning_rate": 6.035833563350757e-06, + "loss": 0.553, + "num_input_tokens_seen": 47110272, + "step": 30965 + }, + { + "epoch": 0.9912937712054286, + "grad_norm": 0.4967513978481293, + "learning_rate": 6.0294380024324525e-06, + "loss": 0.3208, + "num_input_tokens_seen": 47118000, + "step": 30970 + }, + { + "epoch": 0.991453812175917, + "grad_norm": 0.6344567537307739, + "learning_rate": 6.023045367013213e-06, + "loss": 0.3212, + "num_input_tokens_seen": 47125312, + "step": 30975 + }, + { + "epoch": 0.9916138531464055, + "grad_norm": 0.902840256690979, + "learning_rate": 6.016655658078851e-06, + "loss": 0.4367, + "num_input_tokens_seen": 47132608, + "step": 30980 + }, + { + "epoch": 0.9917738941168939, + "grad_norm": 1.2206242084503174, + "learning_rate": 6.010268876614753e-06, + "loss": 0.5489, + "num_input_tokens_seen": 47140320, + "step": 30985 + }, + { + "epoch": 0.9919339350873824, + "grad_norm": 0.7447522878646851, + "learning_rate": 6.0038850236058266e-06, + "loss": 0.4658, + "num_input_tokens_seen": 47147984, + "step": 30990 + }, + { + "epoch": 0.9920939760578709, + "grad_norm": 0.5443540215492249, + "learning_rate": 5.997504100036549e-06, + "loss": 0.4683, + "num_input_tokens_seen": 47156224, + "step": 30995 + }, + { + "epoch": 0.9922540170283592, + "grad_norm": 0.7631616592407227, + "learning_rate": 5.991126106890949e-06, + "loss": 0.4899, + "num_input_tokens_seen": 47163856, + "step": 31000 + }, + { + "epoch": 0.9922540170283592, + "eval_loss": 0.4960794746875763, + "eval_runtime": 557.3041, + "eval_samples_per_second": 24.916, + "eval_steps_per_second": 12.458, + "num_input_tokens_seen": 47163856, + "step": 31000 + }, + { + "epoch": 0.9924140579988477, + "grad_norm": 1.0904395580291748, + "learning_rate": 5.984751045152576e-06, + "loss": 0.4883, + "num_input_tokens_seen": 47171952, + "step": 31005 + }, + { + "epoch": 0.9925740989693361, + "grad_norm": 0.5634860396385193, + "learning_rate": 5.978378915804553e-06, + "loss": 0.4706, + "num_input_tokens_seen": 47179968, + "step": 31010 + }, + { + "epoch": 0.9927341399398246, + "grad_norm": 0.36337411403656006, + "learning_rate": 5.972009719829547e-06, + "loss": 0.2963, + "num_input_tokens_seen": 47187648, + "step": 31015 + }, + { + "epoch": 0.992894180910313, + "grad_norm": 0.3969637155532837, + "learning_rate": 5.965643458209755e-06, + "loss": 0.5331, + "num_input_tokens_seen": 47195568, + "step": 31020 + }, + { + "epoch": 0.9930542218808015, + "grad_norm": 1.1749955415725708, + "learning_rate": 5.95928013192695e-06, + "loss": 0.5099, + "num_input_tokens_seen": 47202944, + "step": 31025 + }, + { + "epoch": 0.9932142628512899, + "grad_norm": 0.7526997327804565, + "learning_rate": 5.952919741962423e-06, + "loss": 0.4415, + "num_input_tokens_seen": 47210448, + "step": 31030 + }, + { + "epoch": 0.9933743038217784, + "grad_norm": 0.7372031211853027, + "learning_rate": 5.946562289297042e-06, + "loss": 0.4587, + "num_input_tokens_seen": 47217968, + "step": 31035 + }, + { + "epoch": 0.9935343447922668, + "grad_norm": 0.9213401675224304, + "learning_rate": 5.9402077749111855e-06, + "loss": 0.5861, + "num_input_tokens_seen": 47225664, + "step": 31040 + }, + { + "epoch": 0.9936943857627553, + "grad_norm": 0.8053582310676575, + "learning_rate": 5.933856199784821e-06, + "loss": 0.3746, + "num_input_tokens_seen": 47233584, + "step": 31045 + }, + { + "epoch": 0.9938544267332438, + "grad_norm": 0.7088239789009094, + "learning_rate": 5.927507564897419e-06, + "loss": 0.4814, + "num_input_tokens_seen": 47241520, + "step": 31050 + }, + { + "epoch": 0.9940144677037321, + "grad_norm": 0.7999728918075562, + "learning_rate": 5.9211618712280395e-06, + "loss": 0.4349, + "num_input_tokens_seen": 47248944, + "step": 31055 + }, + { + "epoch": 0.9941745086742206, + "grad_norm": 0.9085303544998169, + "learning_rate": 5.914819119755255e-06, + "loss": 0.3474, + "num_input_tokens_seen": 47256240, + "step": 31060 + }, + { + "epoch": 0.994334549644709, + "grad_norm": 0.3708161413669586, + "learning_rate": 5.908479311457205e-06, + "loss": 0.4095, + "num_input_tokens_seen": 47264288, + "step": 31065 + }, + { + "epoch": 0.9944945906151975, + "grad_norm": 1.015913963317871, + "learning_rate": 5.902142447311559e-06, + "loss": 0.566, + "num_input_tokens_seen": 47272032, + "step": 31070 + }, + { + "epoch": 0.9946546315856859, + "grad_norm": 0.6744931936264038, + "learning_rate": 5.895808528295546e-06, + "loss": 0.4463, + "num_input_tokens_seen": 47279376, + "step": 31075 + }, + { + "epoch": 0.9948146725561744, + "grad_norm": 0.48845428228378296, + "learning_rate": 5.889477555385941e-06, + "loss": 0.4736, + "num_input_tokens_seen": 47286752, + "step": 31080 + }, + { + "epoch": 0.9949747135266628, + "grad_norm": 0.9022527933120728, + "learning_rate": 5.883149529559051e-06, + "loss": 0.4547, + "num_input_tokens_seen": 47294672, + "step": 31085 + }, + { + "epoch": 0.9951347544971513, + "grad_norm": 0.9194526672363281, + "learning_rate": 5.876824451790738e-06, + "loss": 0.5153, + "num_input_tokens_seen": 47302448, + "step": 31090 + }, + { + "epoch": 0.9952947954676398, + "grad_norm": 0.34943604469299316, + "learning_rate": 5.87050232305642e-06, + "loss": 0.3403, + "num_input_tokens_seen": 47309936, + "step": 31095 + }, + { + "epoch": 0.9954548364381282, + "grad_norm": 0.4109703004360199, + "learning_rate": 5.864183144331034e-06, + "loss": 0.5985, + "num_input_tokens_seen": 47317728, + "step": 31100 + }, + { + "epoch": 0.9956148774086166, + "grad_norm": 0.5158684253692627, + "learning_rate": 5.857866916589089e-06, + "loss": 0.3477, + "num_input_tokens_seen": 47326128, + "step": 31105 + }, + { + "epoch": 0.995774918379105, + "grad_norm": 1.41402006149292, + "learning_rate": 5.8515536408046216e-06, + "loss": 0.7016, + "num_input_tokens_seen": 47333296, + "step": 31110 + }, + { + "epoch": 0.9959349593495935, + "grad_norm": 0.5486800074577332, + "learning_rate": 5.845243317951208e-06, + "loss": 0.4149, + "num_input_tokens_seen": 47341328, + "step": 31115 + }, + { + "epoch": 0.9960950003200819, + "grad_norm": 1.127213478088379, + "learning_rate": 5.838935949001997e-06, + "loss": 0.5208, + "num_input_tokens_seen": 47349296, + "step": 31120 + }, + { + "epoch": 0.9962550412905704, + "grad_norm": 0.6902269124984741, + "learning_rate": 5.8326315349296476e-06, + "loss": 0.4562, + "num_input_tokens_seen": 47356848, + "step": 31125 + }, + { + "epoch": 0.9964150822610588, + "grad_norm": 0.433366060256958, + "learning_rate": 5.826330076706396e-06, + "loss": 0.5162, + "num_input_tokens_seen": 47364464, + "step": 31130 + }, + { + "epoch": 0.9965751232315473, + "grad_norm": 1.0259515047073364, + "learning_rate": 5.820031575303988e-06, + "loss": 0.4317, + "num_input_tokens_seen": 47372160, + "step": 31135 + }, + { + "epoch": 0.9967351642020357, + "grad_norm": 0.8931521773338318, + "learning_rate": 5.813736031693745e-06, + "loss": 0.5896, + "num_input_tokens_seen": 47379552, + "step": 31140 + }, + { + "epoch": 0.9968952051725242, + "grad_norm": 0.9139845967292786, + "learning_rate": 5.807443446846522e-06, + "loss": 0.59, + "num_input_tokens_seen": 47387072, + "step": 31145 + }, + { + "epoch": 0.9970552461430127, + "grad_norm": 0.4334736168384552, + "learning_rate": 5.801153821732699e-06, + "loss": 0.438, + "num_input_tokens_seen": 47394320, + "step": 31150 + }, + { + "epoch": 0.997215287113501, + "grad_norm": 0.9472572803497314, + "learning_rate": 5.794867157322229e-06, + "loss": 0.535, + "num_input_tokens_seen": 47402800, + "step": 31155 + }, + { + "epoch": 0.9973753280839895, + "grad_norm": 1.0848214626312256, + "learning_rate": 5.788583454584593e-06, + "loss": 0.6788, + "num_input_tokens_seen": 47410288, + "step": 31160 + }, + { + "epoch": 0.9975353690544779, + "grad_norm": 0.5441232919692993, + "learning_rate": 5.7823027144888075e-06, + "loss": 0.4425, + "num_input_tokens_seen": 47417632, + "step": 31165 + }, + { + "epoch": 0.9976954100249664, + "grad_norm": 0.6657461524009705, + "learning_rate": 5.776024938003455e-06, + "loss": 0.4858, + "num_input_tokens_seen": 47424928, + "step": 31170 + }, + { + "epoch": 0.9978554509954548, + "grad_norm": 0.5829175114631653, + "learning_rate": 5.7697501260966345e-06, + "loss": 0.44, + "num_input_tokens_seen": 47432368, + "step": 31175 + }, + { + "epoch": 0.9980154919659433, + "grad_norm": 0.943316638469696, + "learning_rate": 5.7634782797360145e-06, + "loss": 0.4789, + "num_input_tokens_seen": 47439744, + "step": 31180 + }, + { + "epoch": 0.9981755329364317, + "grad_norm": 0.6540191769599915, + "learning_rate": 5.757209399888777e-06, + "loss": 0.4295, + "num_input_tokens_seen": 47447680, + "step": 31185 + }, + { + "epoch": 0.9983355739069202, + "grad_norm": 0.8383513689041138, + "learning_rate": 5.750943487521679e-06, + "loss": 0.4753, + "num_input_tokens_seen": 47455216, + "step": 31190 + }, + { + "epoch": 0.9984956148774086, + "grad_norm": 2.0408577919006348, + "learning_rate": 5.744680543600986e-06, + "loss": 0.4471, + "num_input_tokens_seen": 47462544, + "step": 31195 + }, + { + "epoch": 0.9986556558478971, + "grad_norm": 0.8669110536575317, + "learning_rate": 5.738420569092537e-06, + "loss": 0.4158, + "num_input_tokens_seen": 47469712, + "step": 31200 + }, + { + "epoch": 0.9986556558478971, + "eval_loss": 0.49612390995025635, + "eval_runtime": 557.0514, + "eval_samples_per_second": 24.928, + "eval_steps_per_second": 12.464, + "num_input_tokens_seen": 47469712, + "step": 31200 + }, + { + "epoch": 0.9988156968183856, + "grad_norm": 1.6941958665847778, + "learning_rate": 5.732163564961684e-06, + "loss": 0.5315, + "num_input_tokens_seen": 47477616, + "step": 31205 + }, + { + "epoch": 0.9989757377888739, + "grad_norm": 1.0546659231185913, + "learning_rate": 5.725909532173354e-06, + "loss": 0.4702, + "num_input_tokens_seen": 47484944, + "step": 31210 + }, + { + "epoch": 0.9991357787593624, + "grad_norm": 0.7073148488998413, + "learning_rate": 5.719658471691977e-06, + "loss": 0.4639, + "num_input_tokens_seen": 47493040, + "step": 31215 + }, + { + "epoch": 0.9992958197298508, + "grad_norm": 1.227396011352539, + "learning_rate": 5.71341038448156e-06, + "loss": 0.4417, + "num_input_tokens_seen": 47501088, + "step": 31220 + }, + { + "epoch": 0.9994558607003393, + "grad_norm": 1.538549780845642, + "learning_rate": 5.707165271505635e-06, + "loss": 0.532, + "num_input_tokens_seen": 47508752, + "step": 31225 + }, + { + "epoch": 0.9996159016708277, + "grad_norm": 0.6293674111366272, + "learning_rate": 5.700923133727271e-06, + "loss": 0.557, + "num_input_tokens_seen": 47517104, + "step": 31230 + }, + { + "epoch": 0.9997759426413162, + "grad_norm": 0.875503420829773, + "learning_rate": 5.694683972109083e-06, + "loss": 0.3917, + "num_input_tokens_seen": 47524480, + "step": 31235 + }, + { + "epoch": 0.9999359836118046, + "grad_norm": 0.6869008541107178, + "learning_rate": 5.688447787613241e-06, + "loss": 0.6268, + "num_input_tokens_seen": 47532400, + "step": 31240 + }, + { + "epoch": 1.000096024582293, + "grad_norm": 0.4987761080265045, + "learning_rate": 5.6822145812014285e-06, + "loss": 0.3363, + "num_input_tokens_seen": 47539664, + "step": 31245 + }, + { + "epoch": 1.0002560655527815, + "grad_norm": 0.7114173173904419, + "learning_rate": 5.675984353834896e-06, + "loss": 0.5886, + "num_input_tokens_seen": 47547280, + "step": 31250 + }, + { + "epoch": 1.00041610652327, + "grad_norm": 1.6206492185592651, + "learning_rate": 5.66975710647441e-06, + "loss": 0.5414, + "num_input_tokens_seen": 47554880, + "step": 31255 + }, + { + "epoch": 1.0005761474937584, + "grad_norm": 0.27504628896713257, + "learning_rate": 5.663532840080304e-06, + "loss": 0.4186, + "num_input_tokens_seen": 47562400, + "step": 31260 + }, + { + "epoch": 1.000736188464247, + "grad_norm": 0.6365491151809692, + "learning_rate": 5.6573115556124325e-06, + "loss": 0.4209, + "num_input_tokens_seen": 47570112, + "step": 31265 + }, + { + "epoch": 1.0008962294347352, + "grad_norm": 1.1979111433029175, + "learning_rate": 5.651093254030185e-06, + "loss": 0.5107, + "num_input_tokens_seen": 47577456, + "step": 31270 + }, + { + "epoch": 1.0010562704052237, + "grad_norm": 0.811820387840271, + "learning_rate": 5.644877936292514e-06, + "loss": 0.4546, + "num_input_tokens_seen": 47585328, + "step": 31275 + }, + { + "epoch": 1.0012163113757122, + "grad_norm": 0.994124174118042, + "learning_rate": 5.638665603357901e-06, + "loss": 0.4759, + "num_input_tokens_seen": 47592704, + "step": 31280 + }, + { + "epoch": 1.0013763523462007, + "grad_norm": 0.6916602253913879, + "learning_rate": 5.632456256184357e-06, + "loss": 0.3979, + "num_input_tokens_seen": 47600640, + "step": 31285 + }, + { + "epoch": 1.001536393316689, + "grad_norm": 0.9378623366355896, + "learning_rate": 5.626249895729452e-06, + "loss": 0.5472, + "num_input_tokens_seen": 47608448, + "step": 31290 + }, + { + "epoch": 1.0016964342871775, + "grad_norm": 1.0189539194107056, + "learning_rate": 5.620046522950273e-06, + "loss": 0.4701, + "num_input_tokens_seen": 47615744, + "step": 31295 + }, + { + "epoch": 1.001856475257666, + "grad_norm": 0.6783265471458435, + "learning_rate": 5.613846138803464e-06, + "loss": 0.5102, + "num_input_tokens_seen": 47623216, + "step": 31300 + }, + { + "epoch": 1.0020165162281545, + "grad_norm": 0.9231130480766296, + "learning_rate": 5.607648744245206e-06, + "loss": 0.4164, + "num_input_tokens_seen": 47630528, + "step": 31305 + }, + { + "epoch": 1.002176557198643, + "grad_norm": 0.7536009550094604, + "learning_rate": 5.601454340231207e-06, + "loss": 0.4386, + "num_input_tokens_seen": 47637920, + "step": 31310 + }, + { + "epoch": 1.0023365981691312, + "grad_norm": 0.6035728454589844, + "learning_rate": 5.595262927716724e-06, + "loss": 0.4676, + "num_input_tokens_seen": 47645632, + "step": 31315 + }, + { + "epoch": 1.0024966391396197, + "grad_norm": 1.2966033220291138, + "learning_rate": 5.589074507656561e-06, + "loss": 0.6061, + "num_input_tokens_seen": 47653152, + "step": 31320 + }, + { + "epoch": 1.0026566801101082, + "grad_norm": 0.5121551752090454, + "learning_rate": 5.582889081005044e-06, + "loss": 0.5077, + "num_input_tokens_seen": 47660624, + "step": 31325 + }, + { + "epoch": 1.0028167210805967, + "grad_norm": 0.563209056854248, + "learning_rate": 5.5767066487160316e-06, + "loss": 0.425, + "num_input_tokens_seen": 47667808, + "step": 31330 + }, + { + "epoch": 1.002976762051085, + "grad_norm": 0.6405938267707825, + "learning_rate": 5.570527211742949e-06, + "loss": 0.4114, + "num_input_tokens_seen": 47675136, + "step": 31335 + }, + { + "epoch": 1.0031368030215735, + "grad_norm": 1.6368590593338013, + "learning_rate": 5.564350771038731e-06, + "loss": 0.6676, + "num_input_tokens_seen": 47682896, + "step": 31340 + }, + { + "epoch": 1.003296843992062, + "grad_norm": 0.6445422768592834, + "learning_rate": 5.558177327555875e-06, + "loss": 0.4464, + "num_input_tokens_seen": 47690112, + "step": 31345 + }, + { + "epoch": 1.0034568849625505, + "grad_norm": 0.6754852533340454, + "learning_rate": 5.552006882246388e-06, + "loss": 0.3952, + "num_input_tokens_seen": 47697408, + "step": 31350 + }, + { + "epoch": 1.0036169259330388, + "grad_norm": 0.39299866557121277, + "learning_rate": 5.545839436061839e-06, + "loss": 0.4386, + "num_input_tokens_seen": 47705056, + "step": 31355 + }, + { + "epoch": 1.0037769669035272, + "grad_norm": 0.8395305871963501, + "learning_rate": 5.539674989953331e-06, + "loss": 0.5152, + "num_input_tokens_seen": 47712720, + "step": 31360 + }, + { + "epoch": 1.0039370078740157, + "grad_norm": 0.3998841941356659, + "learning_rate": 5.533513544871488e-06, + "loss": 0.3931, + "num_input_tokens_seen": 47720672, + "step": 31365 + }, + { + "epoch": 1.0040970488445042, + "grad_norm": 0.5431045889854431, + "learning_rate": 5.527355101766493e-06, + "loss": 0.2481, + "num_input_tokens_seen": 47728096, + "step": 31370 + }, + { + "epoch": 1.0042570898149927, + "grad_norm": 1.0041714906692505, + "learning_rate": 5.521199661588044e-06, + "loss": 0.6471, + "num_input_tokens_seen": 47735280, + "step": 31375 + }, + { + "epoch": 1.004417130785481, + "grad_norm": 0.6165810823440552, + "learning_rate": 5.5150472252853944e-06, + "loss": 0.3624, + "num_input_tokens_seen": 47742672, + "step": 31380 + }, + { + "epoch": 1.0045771717559695, + "grad_norm": 0.7818987369537354, + "learning_rate": 5.50889779380733e-06, + "loss": 0.3925, + "num_input_tokens_seen": 47749840, + "step": 31385 + }, + { + "epoch": 1.004737212726458, + "grad_norm": 1.3031142950057983, + "learning_rate": 5.5027513681021605e-06, + "loss": 0.4274, + "num_input_tokens_seen": 47757088, + "step": 31390 + }, + { + "epoch": 1.0048972536969465, + "grad_norm": 0.7345706820487976, + "learning_rate": 5.4966079491177545e-06, + "loss": 0.3736, + "num_input_tokens_seen": 47764736, + "step": 31395 + }, + { + "epoch": 1.0050572946674348, + "grad_norm": 1.0877605676651, + "learning_rate": 5.490467537801491e-06, + "loss": 0.379, + "num_input_tokens_seen": 47773184, + "step": 31400 + }, + { + "epoch": 1.0050572946674348, + "eval_loss": 0.49586036801338196, + "eval_runtime": 556.9984, + "eval_samples_per_second": 24.93, + "eval_steps_per_second": 12.465, + "num_input_tokens_seen": 47773184, + "step": 31400 + }, + { + "epoch": 1.0052173356379233, + "grad_norm": 0.6290103197097778, + "learning_rate": 5.484330135100313e-06, + "loss": 0.6461, + "num_input_tokens_seen": 47781488, + "step": 31405 + }, + { + "epoch": 1.0053773766084118, + "grad_norm": 0.8436869382858276, + "learning_rate": 5.4781957419606785e-06, + "loss": 0.3562, + "num_input_tokens_seen": 47788960, + "step": 31410 + }, + { + "epoch": 1.0055374175789002, + "grad_norm": 0.7143149375915527, + "learning_rate": 5.472064359328577e-06, + "loss": 0.4237, + "num_input_tokens_seen": 47797472, + "step": 31415 + }, + { + "epoch": 1.0056974585493887, + "grad_norm": 1.137811541557312, + "learning_rate": 5.4659359881495565e-06, + "loss": 0.5542, + "num_input_tokens_seen": 47805536, + "step": 31420 + }, + { + "epoch": 1.005857499519877, + "grad_norm": 0.7297395467758179, + "learning_rate": 5.4598106293686916e-06, + "loss": 0.5, + "num_input_tokens_seen": 47812928, + "step": 31425 + }, + { + "epoch": 1.0060175404903655, + "grad_norm": 1.3668065071105957, + "learning_rate": 5.45368828393058e-06, + "loss": 0.4212, + "num_input_tokens_seen": 47820352, + "step": 31430 + }, + { + "epoch": 1.006177581460854, + "grad_norm": 0.5566009879112244, + "learning_rate": 5.44756895277937e-06, + "loss": 0.4373, + "num_input_tokens_seen": 47827776, + "step": 31435 + }, + { + "epoch": 1.0063376224313425, + "grad_norm": 0.6452738642692566, + "learning_rate": 5.441452636858746e-06, + "loss": 0.483, + "num_input_tokens_seen": 47835264, + "step": 31440 + }, + { + "epoch": 1.0064976634018308, + "grad_norm": 0.8126090168952942, + "learning_rate": 5.435339337111905e-06, + "loss": 0.4587, + "num_input_tokens_seen": 47842864, + "step": 31445 + }, + { + "epoch": 1.0066577043723193, + "grad_norm": 0.4354894161224365, + "learning_rate": 5.42922905448161e-06, + "loss": 0.3629, + "num_input_tokens_seen": 47850480, + "step": 31450 + }, + { + "epoch": 1.0068177453428078, + "grad_norm": 0.8299148082733154, + "learning_rate": 5.423121789910129e-06, + "loss": 0.4634, + "num_input_tokens_seen": 47858240, + "step": 31455 + }, + { + "epoch": 1.0069777863132963, + "grad_norm": 1.2110167741775513, + "learning_rate": 5.417017544339287e-06, + "loss": 0.6422, + "num_input_tokens_seen": 47866512, + "step": 31460 + }, + { + "epoch": 1.0071378272837848, + "grad_norm": 0.532518208026886, + "learning_rate": 5.410916318710443e-06, + "loss": 0.6391, + "num_input_tokens_seen": 47874256, + "step": 31465 + }, + { + "epoch": 1.007297868254273, + "grad_norm": 0.7241614460945129, + "learning_rate": 5.404818113964466e-06, + "loss": 0.4429, + "num_input_tokens_seen": 47882064, + "step": 31470 + }, + { + "epoch": 1.0074579092247615, + "grad_norm": 0.6517507433891296, + "learning_rate": 5.398722931041792e-06, + "loss": 0.3391, + "num_input_tokens_seen": 47889648, + "step": 31475 + }, + { + "epoch": 1.00761795019525, + "grad_norm": 0.42260152101516724, + "learning_rate": 5.392630770882367e-06, + "loss": 0.4833, + "num_input_tokens_seen": 47897824, + "step": 31480 + }, + { + "epoch": 1.0077779911657385, + "grad_norm": 1.0424766540527344, + "learning_rate": 5.3865416344256705e-06, + "loss": 0.5596, + "num_input_tokens_seen": 47905632, + "step": 31485 + }, + { + "epoch": 1.0079380321362268, + "grad_norm": 1.131824254989624, + "learning_rate": 5.380455522610742e-06, + "loss": 0.4745, + "num_input_tokens_seen": 47913120, + "step": 31490 + }, + { + "epoch": 1.0080980731067153, + "grad_norm": 0.4581000506877899, + "learning_rate": 5.374372436376116e-06, + "loss": 0.3267, + "num_input_tokens_seen": 47920336, + "step": 31495 + }, + { + "epoch": 1.0082581140772038, + "grad_norm": 1.1940059661865234, + "learning_rate": 5.368292376659895e-06, + "loss": 0.4741, + "num_input_tokens_seen": 47927712, + "step": 31500 + }, + { + "epoch": 1.0084181550476923, + "grad_norm": 1.5530369281768799, + "learning_rate": 5.362215344399701e-06, + "loss": 0.4236, + "num_input_tokens_seen": 47935408, + "step": 31505 + }, + { + "epoch": 1.0085781960181806, + "grad_norm": 0.8621060252189636, + "learning_rate": 5.356141340532678e-06, + "loss": 0.431, + "num_input_tokens_seen": 47942800, + "step": 31510 + }, + { + "epoch": 1.008738236988669, + "grad_norm": 0.618839681148529, + "learning_rate": 5.350070365995522e-06, + "loss": 0.4477, + "num_input_tokens_seen": 47950832, + "step": 31515 + }, + { + "epoch": 1.0088982779591575, + "grad_norm": 0.8297393321990967, + "learning_rate": 5.344002421724459e-06, + "loss": 0.6657, + "num_input_tokens_seen": 47958176, + "step": 31520 + }, + { + "epoch": 1.009058318929646, + "grad_norm": 0.41607987880706787, + "learning_rate": 5.337937508655228e-06, + "loss": 0.4865, + "num_input_tokens_seen": 47965664, + "step": 31525 + }, + { + "epoch": 1.0092183599001345, + "grad_norm": 0.772571861743927, + "learning_rate": 5.331875627723126e-06, + "loss": 0.4902, + "num_input_tokens_seen": 47973184, + "step": 31530 + }, + { + "epoch": 1.0093784008706228, + "grad_norm": 0.7882866263389587, + "learning_rate": 5.325816779862963e-06, + "loss": 0.3938, + "num_input_tokens_seen": 47980592, + "step": 31535 + }, + { + "epoch": 1.0095384418411113, + "grad_norm": 0.6473761200904846, + "learning_rate": 5.319760966009102e-06, + "loss": 0.4257, + "num_input_tokens_seen": 47988288, + "step": 31540 + }, + { + "epoch": 1.0096984828115998, + "grad_norm": 0.8192795515060425, + "learning_rate": 5.3137081870954096e-06, + "loss": 0.4498, + "num_input_tokens_seen": 47995808, + "step": 31545 + }, + { + "epoch": 1.0098585237820883, + "grad_norm": 0.8295687437057495, + "learning_rate": 5.307658444055313e-06, + "loss": 0.5587, + "num_input_tokens_seen": 48003312, + "step": 31550 + }, + { + "epoch": 1.0100185647525766, + "grad_norm": 1.0785754919052124, + "learning_rate": 5.301611737821749e-06, + "loss": 0.5301, + "num_input_tokens_seen": 48011072, + "step": 31555 + }, + { + "epoch": 1.010178605723065, + "grad_norm": 0.25685709714889526, + "learning_rate": 5.295568069327206e-06, + "loss": 0.2865, + "num_input_tokens_seen": 48019040, + "step": 31560 + }, + { + "epoch": 1.0103386466935536, + "grad_norm": 0.5425376892089844, + "learning_rate": 5.289527439503683e-06, + "loss": 0.481, + "num_input_tokens_seen": 48026624, + "step": 31565 + }, + { + "epoch": 1.010498687664042, + "grad_norm": 1.583005428314209, + "learning_rate": 5.28348984928273e-06, + "loss": 0.4865, + "num_input_tokens_seen": 48034192, + "step": 31570 + }, + { + "epoch": 1.0106587286345305, + "grad_norm": 0.5762848854064941, + "learning_rate": 5.27745529959541e-06, + "loss": 0.2903, + "num_input_tokens_seen": 48041568, + "step": 31575 + }, + { + "epoch": 1.0108187696050188, + "grad_norm": 0.9306328296661377, + "learning_rate": 5.271423791372335e-06, + "loss": 0.4987, + "num_input_tokens_seen": 48049504, + "step": 31580 + }, + { + "epoch": 1.0109788105755073, + "grad_norm": 1.1345889568328857, + "learning_rate": 5.26539532554364e-06, + "loss": 0.443, + "num_input_tokens_seen": 48057200, + "step": 31585 + }, + { + "epoch": 1.0111388515459958, + "grad_norm": 0.7875816822052002, + "learning_rate": 5.25936990303898e-06, + "loss": 0.4725, + "num_input_tokens_seen": 48065264, + "step": 31590 + }, + { + "epoch": 1.0112988925164843, + "grad_norm": 0.6119101047515869, + "learning_rate": 5.253347524787555e-06, + "loss": 0.3763, + "num_input_tokens_seen": 48072912, + "step": 31595 + }, + { + "epoch": 1.0114589334869726, + "grad_norm": 0.6965429186820984, + "learning_rate": 5.2473281917181035e-06, + "loss": 0.5601, + "num_input_tokens_seen": 48080192, + "step": 31600 + }, + { + "epoch": 1.0114589334869726, + "eval_loss": 0.4957428574562073, + "eval_runtime": 556.1389, + "eval_samples_per_second": 24.969, + "eval_steps_per_second": 12.484, + "num_input_tokens_seen": 48080192, + "step": 31600 + }, + { + "epoch": 1.011618974457461, + "grad_norm": 0.5209363698959351, + "learning_rate": 5.241311904758864e-06, + "loss": 0.4073, + "num_input_tokens_seen": 48087776, + "step": 31605 + }, + { + "epoch": 1.0117790154279496, + "grad_norm": 1.1347285509109497, + "learning_rate": 5.23529866483764e-06, + "loss": 0.4703, + "num_input_tokens_seen": 48095216, + "step": 31610 + }, + { + "epoch": 1.011939056398438, + "grad_norm": 0.44157692790031433, + "learning_rate": 5.229288472881732e-06, + "loss": 0.4009, + "num_input_tokens_seen": 48103200, + "step": 31615 + }, + { + "epoch": 1.0120990973689263, + "grad_norm": 0.6929196119308472, + "learning_rate": 5.2232813298180025e-06, + "loss": 0.4299, + "num_input_tokens_seen": 48110944, + "step": 31620 + }, + { + "epoch": 1.0122591383394148, + "grad_norm": 1.4263964891433716, + "learning_rate": 5.217277236572824e-06, + "loss": 0.7131, + "num_input_tokens_seen": 48118768, + "step": 31625 + }, + { + "epoch": 1.0124191793099033, + "grad_norm": 0.6325758695602417, + "learning_rate": 5.211276194072093e-06, + "loss": 0.3745, + "num_input_tokens_seen": 48126448, + "step": 31630 + }, + { + "epoch": 1.0125792202803918, + "grad_norm": 1.062106966972351, + "learning_rate": 5.205278203241254e-06, + "loss": 0.3904, + "num_input_tokens_seen": 48133840, + "step": 31635 + }, + { + "epoch": 1.0127392612508803, + "grad_norm": 1.5396192073822021, + "learning_rate": 5.199283265005278e-06, + "loss": 0.5683, + "num_input_tokens_seen": 48140992, + "step": 31640 + }, + { + "epoch": 1.0128993022213686, + "grad_norm": 0.6989174485206604, + "learning_rate": 5.193291380288648e-06, + "loss": 0.4563, + "num_input_tokens_seen": 48148512, + "step": 31645 + }, + { + "epoch": 1.013059343191857, + "grad_norm": 0.626828670501709, + "learning_rate": 5.1873025500153995e-06, + "loss": 0.3937, + "num_input_tokens_seen": 48156624, + "step": 31650 + }, + { + "epoch": 1.0132193841623456, + "grad_norm": 0.8995130658149719, + "learning_rate": 5.181316775109071e-06, + "loss": 0.4283, + "num_input_tokens_seen": 48164000, + "step": 31655 + }, + { + "epoch": 1.013379425132834, + "grad_norm": 0.8687862157821655, + "learning_rate": 5.1753340564927564e-06, + "loss": 0.3801, + "num_input_tokens_seen": 48171808, + "step": 31660 + }, + { + "epoch": 1.0135394661033224, + "grad_norm": 0.4750725030899048, + "learning_rate": 5.169354395089068e-06, + "loss": 0.5326, + "num_input_tokens_seen": 48179024, + "step": 31665 + }, + { + "epoch": 1.0136995070738108, + "grad_norm": 0.6548845171928406, + "learning_rate": 5.1633777918201346e-06, + "loss": 0.4304, + "num_input_tokens_seen": 48186864, + "step": 31670 + }, + { + "epoch": 1.0138595480442993, + "grad_norm": 0.8422133922576904, + "learning_rate": 5.157404247607625e-06, + "loss": 0.37, + "num_input_tokens_seen": 48194080, + "step": 31675 + }, + { + "epoch": 1.0140195890147878, + "grad_norm": 0.5583165884017944, + "learning_rate": 5.1514337633727454e-06, + "loss": 0.4098, + "num_input_tokens_seen": 48201360, + "step": 31680 + }, + { + "epoch": 1.0141796299852763, + "grad_norm": 0.8381309509277344, + "learning_rate": 5.145466340036206e-06, + "loss": 0.5936, + "num_input_tokens_seen": 48208720, + "step": 31685 + }, + { + "epoch": 1.0143396709557646, + "grad_norm": 1.0072566270828247, + "learning_rate": 5.139501978518274e-06, + "loss": 0.744, + "num_input_tokens_seen": 48216000, + "step": 31690 + }, + { + "epoch": 1.014499711926253, + "grad_norm": 0.5724445581436157, + "learning_rate": 5.133540679738716e-06, + "loss": 0.5121, + "num_input_tokens_seen": 48223840, + "step": 31695 + }, + { + "epoch": 1.0146597528967416, + "grad_norm": 1.4876596927642822, + "learning_rate": 5.127582444616838e-06, + "loss": 0.5904, + "num_input_tokens_seen": 48231072, + "step": 31700 + }, + { + "epoch": 1.01481979386723, + "grad_norm": 1.4247429370880127, + "learning_rate": 5.121627274071486e-06, + "loss": 0.6227, + "num_input_tokens_seen": 48238400, + "step": 31705 + }, + { + "epoch": 1.0149798348377184, + "grad_norm": 1.320821762084961, + "learning_rate": 5.115675169021009e-06, + "loss": 0.6549, + "num_input_tokens_seen": 48246016, + "step": 31710 + }, + { + "epoch": 1.0151398758082069, + "grad_norm": 1.1278446912765503, + "learning_rate": 5.1097261303832994e-06, + "loss": 0.5296, + "num_input_tokens_seen": 48253968, + "step": 31715 + }, + { + "epoch": 1.0152999167786954, + "grad_norm": 0.41210058331489563, + "learning_rate": 5.103780159075788e-06, + "loss": 0.5199, + "num_input_tokens_seen": 48261472, + "step": 31720 + }, + { + "epoch": 1.0154599577491839, + "grad_norm": 0.9927553534507751, + "learning_rate": 5.0978372560154e-06, + "loss": 0.5803, + "num_input_tokens_seen": 48269360, + "step": 31725 + }, + { + "epoch": 1.0156199987196723, + "grad_norm": 2.1635689735412598, + "learning_rate": 5.091897422118619e-06, + "loss": 0.5058, + "num_input_tokens_seen": 48277296, + "step": 31730 + }, + { + "epoch": 1.0157800396901606, + "grad_norm": 0.3428053557872772, + "learning_rate": 5.0859606583014305e-06, + "loss": 0.2218, + "num_input_tokens_seen": 48284944, + "step": 31735 + }, + { + "epoch": 1.0159400806606491, + "grad_norm": 0.9431790709495544, + "learning_rate": 5.080026965479365e-06, + "loss": 0.5479, + "num_input_tokens_seen": 48292112, + "step": 31740 + }, + { + "epoch": 1.0161001216311376, + "grad_norm": 0.4208061397075653, + "learning_rate": 5.074096344567475e-06, + "loss": 0.5162, + "num_input_tokens_seen": 48299872, + "step": 31745 + }, + { + "epoch": 1.016260162601626, + "grad_norm": 0.5681877732276917, + "learning_rate": 5.0681687964803294e-06, + "loss": 0.4078, + "num_input_tokens_seen": 48307952, + "step": 31750 + }, + { + "epoch": 1.0164202035721144, + "grad_norm": 0.8517413139343262, + "learning_rate": 5.06224432213204e-06, + "loss": 0.3418, + "num_input_tokens_seen": 48315360, + "step": 31755 + }, + { + "epoch": 1.0165802445426029, + "grad_norm": 0.6212743520736694, + "learning_rate": 5.056322922436224e-06, + "loss": 0.4624, + "num_input_tokens_seen": 48322816, + "step": 31760 + }, + { + "epoch": 1.0167402855130914, + "grad_norm": 0.425913542509079, + "learning_rate": 5.0504045983060465e-06, + "loss": 0.5877, + "num_input_tokens_seen": 48330384, + "step": 31765 + }, + { + "epoch": 1.0169003264835799, + "grad_norm": 0.610493540763855, + "learning_rate": 5.044489350654183e-06, + "loss": 0.4644, + "num_input_tokens_seen": 48337840, + "step": 31770 + }, + { + "epoch": 1.0170603674540681, + "grad_norm": 0.39578014612197876, + "learning_rate": 5.038577180392831e-06, + "loss": 0.5692, + "num_input_tokens_seen": 48345808, + "step": 31775 + }, + { + "epoch": 1.0172204084245566, + "grad_norm": 0.6384095549583435, + "learning_rate": 5.032668088433729e-06, + "loss": 0.5152, + "num_input_tokens_seen": 48353008, + "step": 31780 + }, + { + "epoch": 1.0173804493950451, + "grad_norm": 1.3297555446624756, + "learning_rate": 5.02676207568814e-06, + "loss": 0.7342, + "num_input_tokens_seen": 48360960, + "step": 31785 + }, + { + "epoch": 1.0175404903655336, + "grad_norm": 0.5537871718406677, + "learning_rate": 5.02085914306683e-06, + "loss": 0.4237, + "num_input_tokens_seen": 48368256, + "step": 31790 + }, + { + "epoch": 1.0177005313360221, + "grad_norm": 0.4825536608695984, + "learning_rate": 5.014959291480123e-06, + "loss": 0.5105, + "num_input_tokens_seen": 48376704, + "step": 31795 + }, + { + "epoch": 1.0178605723065104, + "grad_norm": 0.9501863718032837, + "learning_rate": 5.009062521837835e-06, + "loss": 0.475, + "num_input_tokens_seen": 48384816, + "step": 31800 + }, + { + "epoch": 1.0178605723065104, + "eval_loss": 0.49555012583732605, + "eval_runtime": 557.2403, + "eval_samples_per_second": 24.919, + "eval_steps_per_second": 12.46, + "num_input_tokens_seen": 48384816, + "step": 31800 + }, + { + "epoch": 1.0180206132769989, + "grad_norm": 1.058863639831543, + "learning_rate": 5.003168835049324e-06, + "loss": 0.4875, + "num_input_tokens_seen": 48392336, + "step": 31805 + }, + { + "epoch": 1.0181806542474874, + "grad_norm": 0.8151993155479431, + "learning_rate": 4.997278232023483e-06, + "loss": 0.4438, + "num_input_tokens_seen": 48399824, + "step": 31810 + }, + { + "epoch": 1.0183406952179759, + "grad_norm": 1.585753321647644, + "learning_rate": 4.9913907136687036e-06, + "loss": 0.4266, + "num_input_tokens_seen": 48407952, + "step": 31815 + }, + { + "epoch": 1.0185007361884642, + "grad_norm": 1.0321500301361084, + "learning_rate": 4.985506280892918e-06, + "loss": 0.4293, + "num_input_tokens_seen": 48415904, + "step": 31820 + }, + { + "epoch": 1.0186607771589526, + "grad_norm": 0.620820164680481, + "learning_rate": 4.979624934603589e-06, + "loss": 0.5914, + "num_input_tokens_seen": 48423344, + "step": 31825 + }, + { + "epoch": 1.0188208181294411, + "grad_norm": 0.661278247833252, + "learning_rate": 4.97374667570768e-06, + "loss": 0.4568, + "num_input_tokens_seen": 48430704, + "step": 31830 + }, + { + "epoch": 1.0189808590999296, + "grad_norm": 0.4837711453437805, + "learning_rate": 4.967871505111704e-06, + "loss": 0.5691, + "num_input_tokens_seen": 48438496, + "step": 31835 + }, + { + "epoch": 1.0191409000704181, + "grad_norm": 0.5120259523391724, + "learning_rate": 4.961999423721686e-06, + "loss": 0.3685, + "num_input_tokens_seen": 48446128, + "step": 31840 + }, + { + "epoch": 1.0193009410409064, + "grad_norm": 0.4636346399784088, + "learning_rate": 4.956130432443159e-06, + "loss": 0.3353, + "num_input_tokens_seen": 48453408, + "step": 31845 + }, + { + "epoch": 1.019460982011395, + "grad_norm": 0.5570144057273865, + "learning_rate": 4.950264532181215e-06, + "loss": 0.5237, + "num_input_tokens_seen": 48460720, + "step": 31850 + }, + { + "epoch": 1.0196210229818834, + "grad_norm": 0.8720072507858276, + "learning_rate": 4.944401723840433e-06, + "loss": 0.3834, + "num_input_tokens_seen": 48467936, + "step": 31855 + }, + { + "epoch": 1.019781063952372, + "grad_norm": 1.5189058780670166, + "learning_rate": 4.938542008324942e-06, + "loss": 0.6728, + "num_input_tokens_seen": 48475696, + "step": 31860 + }, + { + "epoch": 1.0199411049228602, + "grad_norm": 0.8153279423713684, + "learning_rate": 4.9326853865383855e-06, + "loss": 0.5035, + "num_input_tokens_seen": 48483504, + "step": 31865 + }, + { + "epoch": 1.0201011458933487, + "grad_norm": 0.7715150713920593, + "learning_rate": 4.926831859383918e-06, + "loss": 0.4444, + "num_input_tokens_seen": 48490848, + "step": 31870 + }, + { + "epoch": 1.0202611868638372, + "grad_norm": 0.8049734234809875, + "learning_rate": 4.92098142776424e-06, + "loss": 0.4249, + "num_input_tokens_seen": 48498320, + "step": 31875 + }, + { + "epoch": 1.0204212278343257, + "grad_norm": 0.9832978248596191, + "learning_rate": 4.91513409258155e-06, + "loss": 0.6129, + "num_input_tokens_seen": 48505760, + "step": 31880 + }, + { + "epoch": 1.020581268804814, + "grad_norm": 0.8133461475372314, + "learning_rate": 4.909289854737581e-06, + "loss": 0.402, + "num_input_tokens_seen": 48513440, + "step": 31885 + }, + { + "epoch": 1.0207413097753024, + "grad_norm": 0.9131883382797241, + "learning_rate": 4.903448715133602e-06, + "loss": 0.5839, + "num_input_tokens_seen": 48521392, + "step": 31890 + }, + { + "epoch": 1.020901350745791, + "grad_norm": 0.7170103192329407, + "learning_rate": 4.897610674670372e-06, + "loss": 0.6345, + "num_input_tokens_seen": 48528624, + "step": 31895 + }, + { + "epoch": 1.0210613917162794, + "grad_norm": 0.5405590534210205, + "learning_rate": 4.8917757342482e-06, + "loss": 0.4833, + "num_input_tokens_seen": 48536176, + "step": 31900 + }, + { + "epoch": 1.021221432686768, + "grad_norm": 0.5958986282348633, + "learning_rate": 4.885943894766909e-06, + "loss": 0.3848, + "num_input_tokens_seen": 48544560, + "step": 31905 + }, + { + "epoch": 1.0213814736572562, + "grad_norm": 0.8116649985313416, + "learning_rate": 4.880115157125842e-06, + "loss": 0.4649, + "num_input_tokens_seen": 48552112, + "step": 31910 + }, + { + "epoch": 1.0215415146277447, + "grad_norm": 0.9627862572669983, + "learning_rate": 4.874289522223857e-06, + "loss": 0.4191, + "num_input_tokens_seen": 48560000, + "step": 31915 + }, + { + "epoch": 1.0217015555982332, + "grad_norm": 2.529254198074341, + "learning_rate": 4.868466990959339e-06, + "loss": 0.616, + "num_input_tokens_seen": 48568368, + "step": 31920 + }, + { + "epoch": 1.0218615965687217, + "grad_norm": 0.8505228161811829, + "learning_rate": 4.8626475642301964e-06, + "loss": 0.4741, + "num_input_tokens_seen": 48575520, + "step": 31925 + }, + { + "epoch": 1.02202163753921, + "grad_norm": 1.0463348627090454, + "learning_rate": 4.856831242933871e-06, + "loss": 0.6383, + "num_input_tokens_seen": 48583328, + "step": 31930 + }, + { + "epoch": 1.0221816785096984, + "grad_norm": 0.5032216906547546, + "learning_rate": 4.851018027967294e-06, + "loss": 0.4031, + "num_input_tokens_seen": 48590880, + "step": 31935 + }, + { + "epoch": 1.022341719480187, + "grad_norm": 0.357442706823349, + "learning_rate": 4.845207920226946e-06, + "loss": 0.4784, + "num_input_tokens_seen": 48598224, + "step": 31940 + }, + { + "epoch": 1.0225017604506754, + "grad_norm": 1.3176836967468262, + "learning_rate": 4.839400920608825e-06, + "loss": 0.6998, + "num_input_tokens_seen": 48605904, + "step": 31945 + }, + { + "epoch": 1.022661801421164, + "grad_norm": 0.3742789626121521, + "learning_rate": 4.83359703000843e-06, + "loss": 0.353, + "num_input_tokens_seen": 48613424, + "step": 31950 + }, + { + "epoch": 1.0228218423916522, + "grad_norm": 0.7214480638504028, + "learning_rate": 4.827796249320804e-06, + "loss": 0.3783, + "num_input_tokens_seen": 48620240, + "step": 31955 + }, + { + "epoch": 1.0229818833621407, + "grad_norm": 1.0902817249298096, + "learning_rate": 4.82199857944049e-06, + "loss": 0.4132, + "num_input_tokens_seen": 48627344, + "step": 31960 + }, + { + "epoch": 1.0231419243326292, + "grad_norm": 1.3306645154953003, + "learning_rate": 4.8162040212615695e-06, + "loss": 0.4502, + "num_input_tokens_seen": 48635136, + "step": 31965 + }, + { + "epoch": 1.0233019653031177, + "grad_norm": 0.4924475848674774, + "learning_rate": 4.810412575677639e-06, + "loss": 0.4418, + "num_input_tokens_seen": 48643248, + "step": 31970 + }, + { + "epoch": 1.023462006273606, + "grad_norm": 0.8038720488548279, + "learning_rate": 4.804624243581801e-06, + "loss": 0.2287, + "num_input_tokens_seen": 48650720, + "step": 31975 + }, + { + "epoch": 1.0236220472440944, + "grad_norm": 1.5727338790893555, + "learning_rate": 4.798839025866703e-06, + "loss": 0.5684, + "num_input_tokens_seen": 48658816, + "step": 31980 + }, + { + "epoch": 1.023782088214583, + "grad_norm": 0.4962688684463501, + "learning_rate": 4.793056923424491e-06, + "loss": 0.3156, + "num_input_tokens_seen": 48665968, + "step": 31985 + }, + { + "epoch": 1.0239421291850714, + "grad_norm": 0.6124109625816345, + "learning_rate": 4.78727793714683e-06, + "loss": 0.4542, + "num_input_tokens_seen": 48673280, + "step": 31990 + }, + { + "epoch": 1.02410217015556, + "grad_norm": 0.639606773853302, + "learning_rate": 4.7815020679249285e-06, + "loss": 0.3198, + "num_input_tokens_seen": 48681008, + "step": 31995 + }, + { + "epoch": 1.0242622111260482, + "grad_norm": 1.2259165048599243, + "learning_rate": 4.775729316649483e-06, + "loss": 0.5515, + "num_input_tokens_seen": 48688416, + "step": 32000 + }, + { + "epoch": 1.0242622111260482, + "eval_loss": 0.49556559324264526, + "eval_runtime": 557.3122, + "eval_samples_per_second": 24.916, + "eval_steps_per_second": 12.458, + "num_input_tokens_seen": 48688416, + "step": 32000 + }, + { + "epoch": 1.0244222520965367, + "grad_norm": 1.3918323516845703, + "learning_rate": 4.769959684210728e-06, + "loss": 0.603, + "num_input_tokens_seen": 48695856, + "step": 32005 + }, + { + "epoch": 1.0245822930670252, + "grad_norm": 0.8316125273704529, + "learning_rate": 4.764193171498426e-06, + "loss": 0.4247, + "num_input_tokens_seen": 48703264, + "step": 32010 + }, + { + "epoch": 1.0247423340375137, + "grad_norm": 1.2110182046890259, + "learning_rate": 4.75842977940183e-06, + "loss": 0.4227, + "num_input_tokens_seen": 48711072, + "step": 32015 + }, + { + "epoch": 1.024902375008002, + "grad_norm": 0.8098517060279846, + "learning_rate": 4.752669508809729e-06, + "loss": 0.3995, + "num_input_tokens_seen": 48718656, + "step": 32020 + }, + { + "epoch": 1.0250624159784905, + "grad_norm": 1.008904218673706, + "learning_rate": 4.746912360610445e-06, + "loss": 0.4205, + "num_input_tokens_seen": 48726080, + "step": 32025 + }, + { + "epoch": 1.025222456948979, + "grad_norm": 0.6961383819580078, + "learning_rate": 4.741158335691781e-06, + "loss": 0.463, + "num_input_tokens_seen": 48733552, + "step": 32030 + }, + { + "epoch": 1.0253824979194675, + "grad_norm": 0.7757290005683899, + "learning_rate": 4.7354074349410994e-06, + "loss": 0.4322, + "num_input_tokens_seen": 48741168, + "step": 32035 + }, + { + "epoch": 1.0255425388899557, + "grad_norm": 0.5988914370536804, + "learning_rate": 4.729659659245245e-06, + "loss": 0.4346, + "num_input_tokens_seen": 48748640, + "step": 32040 + }, + { + "epoch": 1.0257025798604442, + "grad_norm": 0.7278469204902649, + "learning_rate": 4.723915009490601e-06, + "loss": 0.4743, + "num_input_tokens_seen": 48756272, + "step": 32045 + }, + { + "epoch": 1.0258626208309327, + "grad_norm": 0.7251102328300476, + "learning_rate": 4.718173486563077e-06, + "loss": 0.3517, + "num_input_tokens_seen": 48763920, + "step": 32050 + }, + { + "epoch": 1.0260226618014212, + "grad_norm": 0.9786576628684998, + "learning_rate": 4.71243509134808e-06, + "loss": 0.4437, + "num_input_tokens_seen": 48771136, + "step": 32055 + }, + { + "epoch": 1.0261827027719097, + "grad_norm": 1.2037620544433594, + "learning_rate": 4.706699824730532e-06, + "loss": 0.6976, + "num_input_tokens_seen": 48779232, + "step": 32060 + }, + { + "epoch": 1.026342743742398, + "grad_norm": 0.3362762928009033, + "learning_rate": 4.700967687594901e-06, + "loss": 0.538, + "num_input_tokens_seen": 48786944, + "step": 32065 + }, + { + "epoch": 1.0265027847128865, + "grad_norm": 1.2829670906066895, + "learning_rate": 4.69523868082514e-06, + "loss": 0.4996, + "num_input_tokens_seen": 48794256, + "step": 32070 + }, + { + "epoch": 1.026662825683375, + "grad_norm": 0.6353484988212585, + "learning_rate": 4.689512805304747e-06, + "loss": 0.493, + "num_input_tokens_seen": 48801776, + "step": 32075 + }, + { + "epoch": 1.0268228666538635, + "grad_norm": 1.0384641885757446, + "learning_rate": 4.683790061916707e-06, + "loss": 0.4343, + "num_input_tokens_seen": 48809632, + "step": 32080 + }, + { + "epoch": 1.0269829076243517, + "grad_norm": 0.9300330877304077, + "learning_rate": 4.678070451543551e-06, + "loss": 0.529, + "num_input_tokens_seen": 48817072, + "step": 32085 + }, + { + "epoch": 1.0271429485948402, + "grad_norm": 0.7828682065010071, + "learning_rate": 4.6723539750673204e-06, + "loss": 0.4544, + "num_input_tokens_seen": 48824448, + "step": 32090 + }, + { + "epoch": 1.0273029895653287, + "grad_norm": 0.9471649527549744, + "learning_rate": 4.666640633369551e-06, + "loss": 0.5, + "num_input_tokens_seen": 48831744, + "step": 32095 + }, + { + "epoch": 1.0274630305358172, + "grad_norm": 0.6487364768981934, + "learning_rate": 4.660930427331323e-06, + "loss": 0.4172, + "num_input_tokens_seen": 48839232, + "step": 32100 + }, + { + "epoch": 1.0276230715063057, + "grad_norm": 0.9691084027290344, + "learning_rate": 4.6552233578332244e-06, + "loss": 0.3901, + "num_input_tokens_seen": 48847088, + "step": 32105 + }, + { + "epoch": 1.027783112476794, + "grad_norm": 1.6036911010742188, + "learning_rate": 4.649519425755347e-06, + "loss": 0.6154, + "num_input_tokens_seen": 48854480, + "step": 32110 + }, + { + "epoch": 1.0279431534472825, + "grad_norm": 0.9077528715133667, + "learning_rate": 4.64381863197732e-06, + "loss": 0.6471, + "num_input_tokens_seen": 48862416, + "step": 32115 + }, + { + "epoch": 1.028103194417771, + "grad_norm": 1.2241237163543701, + "learning_rate": 4.638120977378269e-06, + "loss": 0.629, + "num_input_tokens_seen": 48869792, + "step": 32120 + }, + { + "epoch": 1.0282632353882595, + "grad_norm": 0.8644647598266602, + "learning_rate": 4.632426462836848e-06, + "loss": 0.5213, + "num_input_tokens_seen": 48878128, + "step": 32125 + }, + { + "epoch": 1.0284232763587478, + "grad_norm": 1.1449179649353027, + "learning_rate": 4.626735089231224e-06, + "loss": 0.4835, + "num_input_tokens_seen": 48885344, + "step": 32130 + }, + { + "epoch": 1.0285833173292362, + "grad_norm": 0.9018648862838745, + "learning_rate": 4.621046857439068e-06, + "loss": 0.4467, + "num_input_tokens_seen": 48892880, + "step": 32135 + }, + { + "epoch": 1.0287433582997247, + "grad_norm": 0.7106832265853882, + "learning_rate": 4.615361768337587e-06, + "loss": 0.5447, + "num_input_tokens_seen": 48901840, + "step": 32140 + }, + { + "epoch": 1.0289033992702132, + "grad_norm": 1.3157252073287964, + "learning_rate": 4.6096798228034946e-06, + "loss": 0.6677, + "num_input_tokens_seen": 48909344, + "step": 32145 + }, + { + "epoch": 1.0290634402407015, + "grad_norm": 1.5456390380859375, + "learning_rate": 4.604001021713008e-06, + "loss": 0.5451, + "num_input_tokens_seen": 48917504, + "step": 32150 + }, + { + "epoch": 1.02922348121119, + "grad_norm": 1.17225980758667, + "learning_rate": 4.598325365941883e-06, + "loss": 0.5215, + "num_input_tokens_seen": 48924992, + "step": 32155 + }, + { + "epoch": 1.0293835221816785, + "grad_norm": 1.153385877609253, + "learning_rate": 4.5926528563653645e-06, + "loss": 0.4193, + "num_input_tokens_seen": 48932432, + "step": 32160 + }, + { + "epoch": 1.029543563152167, + "grad_norm": 0.5641108155250549, + "learning_rate": 4.5869834938582295e-06, + "loss": 0.5193, + "num_input_tokens_seen": 48940000, + "step": 32165 + }, + { + "epoch": 1.0297036041226555, + "grad_norm": 0.9260337352752686, + "learning_rate": 4.581317279294772e-06, + "loss": 0.5293, + "num_input_tokens_seen": 48947696, + "step": 32170 + }, + { + "epoch": 1.0298636450931438, + "grad_norm": 1.2180660963058472, + "learning_rate": 4.57565421354878e-06, + "loss": 0.5703, + "num_input_tokens_seen": 48955264, + "step": 32175 + }, + { + "epoch": 1.0300236860636323, + "grad_norm": 0.7960955500602722, + "learning_rate": 4.569994297493579e-06, + "loss": 0.4396, + "num_input_tokens_seen": 48963072, + "step": 32180 + }, + { + "epoch": 1.0301837270341208, + "grad_norm": 0.6085801720619202, + "learning_rate": 4.564337532002002e-06, + "loss": 0.4856, + "num_input_tokens_seen": 48970608, + "step": 32185 + }, + { + "epoch": 1.0303437680046093, + "grad_norm": 0.6065562963485718, + "learning_rate": 4.55868391794638e-06, + "loss": 0.4634, + "num_input_tokens_seen": 48978512, + "step": 32190 + }, + { + "epoch": 1.0305038089750975, + "grad_norm": 1.3335760831832886, + "learning_rate": 4.553033456198588e-06, + "loss": 0.7664, + "num_input_tokens_seen": 48985760, + "step": 32195 + }, + { + "epoch": 1.030663849945586, + "grad_norm": 0.7707436680793762, + "learning_rate": 4.54738614762999e-06, + "loss": 0.4853, + "num_input_tokens_seen": 48993296, + "step": 32200 + }, + { + "epoch": 1.030663849945586, + "eval_loss": 0.49517616629600525, + "eval_runtime": 556.2636, + "eval_samples_per_second": 24.963, + "eval_steps_per_second": 12.481, + "num_input_tokens_seen": 48993296, + "step": 32200 + }, + { + "epoch": 1.0308238909160745, + "grad_norm": 0.48615893721580505, + "learning_rate": 4.541741993111465e-06, + "loss": 0.4318, + "num_input_tokens_seen": 49000816, + "step": 32205 + }, + { + "epoch": 1.030983931886563, + "grad_norm": 0.702860951423645, + "learning_rate": 4.536100993513423e-06, + "loss": 0.5028, + "num_input_tokens_seen": 49008160, + "step": 32210 + }, + { + "epoch": 1.0311439728570515, + "grad_norm": 0.5114222764968872, + "learning_rate": 4.530463149705768e-06, + "loss": 0.5915, + "num_input_tokens_seen": 49015536, + "step": 32215 + }, + { + "epoch": 1.0313040138275398, + "grad_norm": 0.6059373617172241, + "learning_rate": 4.524828462557934e-06, + "loss": 0.4531, + "num_input_tokens_seen": 49023040, + "step": 32220 + }, + { + "epoch": 1.0314640547980283, + "grad_norm": 0.603505551815033, + "learning_rate": 4.5191969329388625e-06, + "loss": 0.6748, + "num_input_tokens_seen": 49030096, + "step": 32225 + }, + { + "epoch": 1.0316240957685168, + "grad_norm": 1.2809101343154907, + "learning_rate": 4.5135685617169965e-06, + "loss": 0.6442, + "num_input_tokens_seen": 49038208, + "step": 32230 + }, + { + "epoch": 1.0317841367390053, + "grad_norm": 0.6696087121963501, + "learning_rate": 4.507943349760313e-06, + "loss": 0.2794, + "num_input_tokens_seen": 49045376, + "step": 32235 + }, + { + "epoch": 1.0319441777094935, + "grad_norm": 0.9256532192230225, + "learning_rate": 4.502321297936277e-06, + "loss": 0.571, + "num_input_tokens_seen": 49053424, + "step": 32240 + }, + { + "epoch": 1.032104218679982, + "grad_norm": 0.5054640769958496, + "learning_rate": 4.496702407111888e-06, + "loss": 0.3808, + "num_input_tokens_seen": 49061872, + "step": 32245 + }, + { + "epoch": 1.0322642596504705, + "grad_norm": 0.8424472212791443, + "learning_rate": 4.491086678153653e-06, + "loss": 0.4324, + "num_input_tokens_seen": 49069504, + "step": 32250 + }, + { + "epoch": 1.032424300620959, + "grad_norm": 0.8740779161453247, + "learning_rate": 4.485474111927579e-06, + "loss": 0.5863, + "num_input_tokens_seen": 49076848, + "step": 32255 + }, + { + "epoch": 1.0325843415914475, + "grad_norm": 0.8276370167732239, + "learning_rate": 4.479864709299197e-06, + "loss": 0.3694, + "num_input_tokens_seen": 49084272, + "step": 32260 + }, + { + "epoch": 1.0327443825619358, + "grad_norm": 0.8881122469902039, + "learning_rate": 4.474258471133555e-06, + "loss": 0.4028, + "num_input_tokens_seen": 49092432, + "step": 32265 + }, + { + "epoch": 1.0329044235324243, + "grad_norm": 0.9106552004814148, + "learning_rate": 4.4686553982952014e-06, + "loss": 0.7044, + "num_input_tokens_seen": 49099760, + "step": 32270 + }, + { + "epoch": 1.0330644645029128, + "grad_norm": 0.5494552850723267, + "learning_rate": 4.463055491648191e-06, + "loss": 0.3907, + "num_input_tokens_seen": 49107312, + "step": 32275 + }, + { + "epoch": 1.0332245054734013, + "grad_norm": 1.3359967470169067, + "learning_rate": 4.457458752056112e-06, + "loss": 0.5037, + "num_input_tokens_seen": 49115232, + "step": 32280 + }, + { + "epoch": 1.0333845464438896, + "grad_norm": 1.8014732599258423, + "learning_rate": 4.451865180382042e-06, + "loss": 0.5922, + "num_input_tokens_seen": 49122816, + "step": 32285 + }, + { + "epoch": 1.033544587414378, + "grad_norm": 0.9893560409545898, + "learning_rate": 4.4462747774885936e-06, + "loss": 0.4975, + "num_input_tokens_seen": 49130352, + "step": 32290 + }, + { + "epoch": 1.0337046283848665, + "grad_norm": 0.5988892912864685, + "learning_rate": 4.440687544237859e-06, + "loss": 0.5839, + "num_input_tokens_seen": 49138064, + "step": 32295 + }, + { + "epoch": 1.033864669355355, + "grad_norm": 0.5527182221412659, + "learning_rate": 4.435103481491471e-06, + "loss": 0.3856, + "num_input_tokens_seen": 49145712, + "step": 32300 + }, + { + "epoch": 1.0340247103258433, + "grad_norm": 1.7719147205352783, + "learning_rate": 4.429522590110569e-06, + "loss": 0.5505, + "num_input_tokens_seen": 49153488, + "step": 32305 + }, + { + "epoch": 1.0341847512963318, + "grad_norm": 0.7806497812271118, + "learning_rate": 4.423944870955779e-06, + "loss": 0.4163, + "num_input_tokens_seen": 49161568, + "step": 32310 + }, + { + "epoch": 1.0343447922668203, + "grad_norm": 1.0130118131637573, + "learning_rate": 4.418370324887272e-06, + "loss": 0.7099, + "num_input_tokens_seen": 49169248, + "step": 32315 + }, + { + "epoch": 1.0345048332373088, + "grad_norm": 0.530307412147522, + "learning_rate": 4.412798952764699e-06, + "loss": 0.3959, + "num_input_tokens_seen": 49176592, + "step": 32320 + }, + { + "epoch": 1.0346648742077973, + "grad_norm": 0.45643869042396545, + "learning_rate": 4.407230755447245e-06, + "loss": 0.4507, + "num_input_tokens_seen": 49184656, + "step": 32325 + }, + { + "epoch": 1.0348249151782856, + "grad_norm": 0.4990514814853668, + "learning_rate": 4.401665733793598e-06, + "loss": 0.3877, + "num_input_tokens_seen": 49192048, + "step": 32330 + }, + { + "epoch": 1.034984956148774, + "grad_norm": 0.7210514545440674, + "learning_rate": 4.3961038886619425e-06, + "loss": 0.5649, + "num_input_tokens_seen": 49200080, + "step": 32335 + }, + { + "epoch": 1.0351449971192626, + "grad_norm": 0.7824261784553528, + "learning_rate": 4.39054522091e-06, + "loss": 0.6256, + "num_input_tokens_seen": 49207504, + "step": 32340 + }, + { + "epoch": 1.035305038089751, + "grad_norm": 0.9147293567657471, + "learning_rate": 4.384989731394979e-06, + "loss": 0.5757, + "num_input_tokens_seen": 49215136, + "step": 32345 + }, + { + "epoch": 1.0354650790602393, + "grad_norm": 1.1424895524978638, + "learning_rate": 4.379437420973598e-06, + "loss": 0.5283, + "num_input_tokens_seen": 49222768, + "step": 32350 + }, + { + "epoch": 1.0356251200307278, + "grad_norm": 0.7213343381881714, + "learning_rate": 4.373888290502107e-06, + "loss": 0.442, + "num_input_tokens_seen": 49230416, + "step": 32355 + }, + { + "epoch": 1.0357851610012163, + "grad_norm": 0.41358599066734314, + "learning_rate": 4.36834234083624e-06, + "loss": 0.3835, + "num_input_tokens_seen": 49237584, + "step": 32360 + }, + { + "epoch": 1.0359452019717048, + "grad_norm": 0.8503771424293518, + "learning_rate": 4.362799572831258e-06, + "loss": 0.5193, + "num_input_tokens_seen": 49244608, + "step": 32365 + }, + { + "epoch": 1.0361052429421933, + "grad_norm": 0.5401152968406677, + "learning_rate": 4.35725998734193e-06, + "loss": 0.4364, + "num_input_tokens_seen": 49252240, + "step": 32370 + }, + { + "epoch": 1.0362652839126816, + "grad_norm": 0.5571712255477905, + "learning_rate": 4.3517235852225195e-06, + "loss": 0.4197, + "num_input_tokens_seen": 49259648, + "step": 32375 + }, + { + "epoch": 1.03642532488317, + "grad_norm": 0.9325260519981384, + "learning_rate": 4.346190367326822e-06, + "loss": 0.4528, + "num_input_tokens_seen": 49267456, + "step": 32380 + }, + { + "epoch": 1.0365853658536586, + "grad_norm": 0.48087218403816223, + "learning_rate": 4.340660334508115e-06, + "loss": 0.5362, + "num_input_tokens_seen": 49274992, + "step": 32385 + }, + { + "epoch": 1.036745406824147, + "grad_norm": 0.3714747726917267, + "learning_rate": 4.335133487619206e-06, + "loss": 0.4453, + "num_input_tokens_seen": 49282416, + "step": 32390 + }, + { + "epoch": 1.0369054477946353, + "grad_norm": 0.44113701581954956, + "learning_rate": 4.329609827512409e-06, + "loss": 0.3981, + "num_input_tokens_seen": 49290896, + "step": 32395 + }, + { + "epoch": 1.0370654887651238, + "grad_norm": 0.4244166910648346, + "learning_rate": 4.324089355039531e-06, + "loss": 0.4677, + "num_input_tokens_seen": 49298128, + "step": 32400 + }, + { + "epoch": 1.0370654887651238, + "eval_loss": 0.4952358603477478, + "eval_runtime": 557.2652, + "eval_samples_per_second": 24.918, + "eval_steps_per_second": 12.459, + "num_input_tokens_seen": 49298128, + "step": 32400 + }, + { + "epoch": 1.0372255297356123, + "grad_norm": 0.6767312288284302, + "learning_rate": 4.3185720710519075e-06, + "loss": 0.3653, + "num_input_tokens_seen": 49305696, + "step": 32405 + }, + { + "epoch": 1.0373855707061008, + "grad_norm": 1.3049495220184326, + "learning_rate": 4.3130579764003724e-06, + "loss": 0.6156, + "num_input_tokens_seen": 49313200, + "step": 32410 + }, + { + "epoch": 1.037545611676589, + "grad_norm": 3.0579745769500732, + "learning_rate": 4.307547071935267e-06, + "loss": 0.4334, + "num_input_tokens_seen": 49320384, + "step": 32415 + }, + { + "epoch": 1.0377056526470776, + "grad_norm": 0.5714679956436157, + "learning_rate": 4.302039358506435e-06, + "loss": 0.5709, + "num_input_tokens_seen": 49327952, + "step": 32420 + }, + { + "epoch": 1.037865693617566, + "grad_norm": 0.43064334988594055, + "learning_rate": 4.296534836963245e-06, + "loss": 0.2777, + "num_input_tokens_seen": 49335872, + "step": 32425 + }, + { + "epoch": 1.0380257345880546, + "grad_norm": 0.9012437462806702, + "learning_rate": 4.291033508154555e-06, + "loss": 0.3833, + "num_input_tokens_seen": 49343840, + "step": 32430 + }, + { + "epoch": 1.038185775558543, + "grad_norm": 1.713500738143921, + "learning_rate": 4.285535372928748e-06, + "loss": 0.4059, + "num_input_tokens_seen": 49351568, + "step": 32435 + }, + { + "epoch": 1.0383458165290314, + "grad_norm": 0.7901766300201416, + "learning_rate": 4.280040432133695e-06, + "loss": 0.4065, + "num_input_tokens_seen": 49358832, + "step": 32440 + }, + { + "epoch": 1.0385058574995198, + "grad_norm": 0.5933013558387756, + "learning_rate": 4.274548686616789e-06, + "loss": 0.3622, + "num_input_tokens_seen": 49366656, + "step": 32445 + }, + { + "epoch": 1.0386658984700083, + "grad_norm": 0.6754314303398132, + "learning_rate": 4.2690601372249364e-06, + "loss": 0.4623, + "num_input_tokens_seen": 49374832, + "step": 32450 + }, + { + "epoch": 1.0388259394404968, + "grad_norm": 0.8614147305488586, + "learning_rate": 4.263574784804525e-06, + "loss": 0.4051, + "num_input_tokens_seen": 49382464, + "step": 32455 + }, + { + "epoch": 1.038985980410985, + "grad_norm": 0.7478454113006592, + "learning_rate": 4.258092630201479e-06, + "loss": 0.4151, + "num_input_tokens_seen": 49389808, + "step": 32460 + }, + { + "epoch": 1.0391460213814736, + "grad_norm": 0.8393138647079468, + "learning_rate": 4.252613674261202e-06, + "loss": 0.5543, + "num_input_tokens_seen": 49397984, + "step": 32465 + }, + { + "epoch": 1.039306062351962, + "grad_norm": 0.5899692177772522, + "learning_rate": 4.2471379178286224e-06, + "loss": 0.4397, + "num_input_tokens_seen": 49405616, + "step": 32470 + }, + { + "epoch": 1.0394661033224506, + "grad_norm": 1.334363341331482, + "learning_rate": 4.241665361748181e-06, + "loss": 0.5491, + "num_input_tokens_seen": 49413456, + "step": 32475 + }, + { + "epoch": 1.039626144292939, + "grad_norm": 0.46649834513664246, + "learning_rate": 4.2361960068637994e-06, + "loss": 0.4899, + "num_input_tokens_seen": 49420992, + "step": 32480 + }, + { + "epoch": 1.0397861852634274, + "grad_norm": 0.3960755467414856, + "learning_rate": 4.230729854018933e-06, + "loss": 0.3326, + "num_input_tokens_seen": 49428704, + "step": 32485 + }, + { + "epoch": 1.0399462262339159, + "grad_norm": 0.447793573141098, + "learning_rate": 4.225266904056521e-06, + "loss": 0.4101, + "num_input_tokens_seen": 49436416, + "step": 32490 + }, + { + "epoch": 1.0401062672044044, + "grad_norm": 0.4886610209941864, + "learning_rate": 4.21980715781903e-06, + "loss": 0.4847, + "num_input_tokens_seen": 49443776, + "step": 32495 + }, + { + "epoch": 1.0402663081748929, + "grad_norm": 1.453842043876648, + "learning_rate": 4.214350616148416e-06, + "loss": 0.4919, + "num_input_tokens_seen": 49450880, + "step": 32500 + }, + { + "epoch": 1.0404263491453811, + "grad_norm": 1.695494294166565, + "learning_rate": 4.20889727988614e-06, + "loss": 0.5935, + "num_input_tokens_seen": 49458560, + "step": 32505 + }, + { + "epoch": 1.0405863901158696, + "grad_norm": 1.4775941371917725, + "learning_rate": 4.20344714987318e-06, + "loss": 0.5247, + "num_input_tokens_seen": 49465936, + "step": 32510 + }, + { + "epoch": 1.0407464310863581, + "grad_norm": 0.46415066719055176, + "learning_rate": 4.198000226950022e-06, + "loss": 0.4003, + "num_input_tokens_seen": 49473296, + "step": 32515 + }, + { + "epoch": 1.0409064720568466, + "grad_norm": 0.9614671468734741, + "learning_rate": 4.192556511956635e-06, + "loss": 0.5207, + "num_input_tokens_seen": 49480672, + "step": 32520 + }, + { + "epoch": 1.041066513027335, + "grad_norm": 0.986075758934021, + "learning_rate": 4.18711600573252e-06, + "loss": 0.5161, + "num_input_tokens_seen": 49488400, + "step": 32525 + }, + { + "epoch": 1.0412265539978234, + "grad_norm": 1.6733758449554443, + "learning_rate": 4.181678709116671e-06, + "loss": 0.3575, + "num_input_tokens_seen": 49496672, + "step": 32530 + }, + { + "epoch": 1.0413865949683119, + "grad_norm": 0.6385301947593689, + "learning_rate": 4.1762446229475785e-06, + "loss": 0.5746, + "num_input_tokens_seen": 49504192, + "step": 32535 + }, + { + "epoch": 1.0415466359388004, + "grad_norm": 1.0496262311935425, + "learning_rate": 4.17081374806326e-06, + "loss": 0.5906, + "num_input_tokens_seen": 49511424, + "step": 32540 + }, + { + "epoch": 1.0417066769092889, + "grad_norm": 0.9748005270957947, + "learning_rate": 4.165386085301212e-06, + "loss": 0.3444, + "num_input_tokens_seen": 49518976, + "step": 32545 + }, + { + "epoch": 1.0418667178797771, + "grad_norm": 1.298679232597351, + "learning_rate": 4.1599616354984525e-06, + "loss": 0.4026, + "num_input_tokens_seen": 49526160, + "step": 32550 + }, + { + "epoch": 1.0420267588502656, + "grad_norm": 0.9359970092773438, + "learning_rate": 4.154540399491508e-06, + "loss": 0.4146, + "num_input_tokens_seen": 49533264, + "step": 32555 + }, + { + "epoch": 1.0421867998207541, + "grad_norm": 1.001994252204895, + "learning_rate": 4.149122378116394e-06, + "loss": 0.5228, + "num_input_tokens_seen": 49541184, + "step": 32560 + }, + { + "epoch": 1.0423468407912426, + "grad_norm": 0.790117621421814, + "learning_rate": 4.14370757220863e-06, + "loss": 0.5714, + "num_input_tokens_seen": 49548784, + "step": 32565 + }, + { + "epoch": 1.042506881761731, + "grad_norm": 0.7959557175636292, + "learning_rate": 4.138295982603263e-06, + "loss": 0.4972, + "num_input_tokens_seen": 49556576, + "step": 32570 + }, + { + "epoch": 1.0426669227322194, + "grad_norm": 0.8202748894691467, + "learning_rate": 4.132887610134814e-06, + "loss": 0.9639, + "num_input_tokens_seen": 49563952, + "step": 32575 + }, + { + "epoch": 1.0428269637027079, + "grad_norm": 0.8692677021026611, + "learning_rate": 4.127482455637335e-06, + "loss": 0.7387, + "num_input_tokens_seen": 49571536, + "step": 32580 + }, + { + "epoch": 1.0429870046731964, + "grad_norm": 0.6164880990982056, + "learning_rate": 4.1220805199443545e-06, + "loss": 0.4997, + "num_input_tokens_seen": 49578720, + "step": 32585 + }, + { + "epoch": 1.0431470456436849, + "grad_norm": 0.956673800945282, + "learning_rate": 4.116681803888925e-06, + "loss": 0.446, + "num_input_tokens_seen": 49586656, + "step": 32590 + }, + { + "epoch": 1.0433070866141732, + "grad_norm": 0.7893695831298828, + "learning_rate": 4.111286308303605e-06, + "loss": 0.4609, + "num_input_tokens_seen": 49594576, + "step": 32595 + }, + { + "epoch": 1.0434671275846616, + "grad_norm": 0.5590416789054871, + "learning_rate": 4.105894034020433e-06, + "loss": 0.2249, + "num_input_tokens_seen": 49602368, + "step": 32600 + }, + { + "epoch": 1.0434671275846616, + "eval_loss": 0.49524328112602234, + "eval_runtime": 557.3309, + "eval_samples_per_second": 24.915, + "eval_steps_per_second": 12.458, + "num_input_tokens_seen": 49602368, + "step": 32600 + }, + { + "epoch": 1.0436271685551501, + "grad_norm": 1.132723093032837, + "learning_rate": 4.100504981870975e-06, + "loss": 0.6029, + "num_input_tokens_seen": 49610144, + "step": 32605 + }, + { + "epoch": 1.0437872095256386, + "grad_norm": 0.6789646744728088, + "learning_rate": 4.0951191526862915e-06, + "loss": 0.4012, + "num_input_tokens_seen": 49617440, + "step": 32610 + }, + { + "epoch": 1.043947250496127, + "grad_norm": 0.8608438372612, + "learning_rate": 4.089736547296938e-06, + "loss": 0.4904, + "num_input_tokens_seen": 49624784, + "step": 32615 + }, + { + "epoch": 1.0441072914666154, + "grad_norm": 1.3007771968841553, + "learning_rate": 4.08435716653299e-06, + "loss": 0.5362, + "num_input_tokens_seen": 49631936, + "step": 32620 + }, + { + "epoch": 1.044267332437104, + "grad_norm": 0.7425694465637207, + "learning_rate": 4.0789810112240005e-06, + "loss": 0.4817, + "num_input_tokens_seen": 49639264, + "step": 32625 + }, + { + "epoch": 1.0444273734075924, + "grad_norm": 1.4064782857894897, + "learning_rate": 4.073608082199057e-06, + "loss": 0.4386, + "num_input_tokens_seen": 49646672, + "step": 32630 + }, + { + "epoch": 1.044587414378081, + "grad_norm": 0.5451858043670654, + "learning_rate": 4.068238380286718e-06, + "loss": 0.4501, + "num_input_tokens_seen": 49654016, + "step": 32635 + }, + { + "epoch": 1.0447474553485692, + "grad_norm": 1.7521635293960571, + "learning_rate": 4.062871906315072e-06, + "loss": 0.4799, + "num_input_tokens_seen": 49661200, + "step": 32640 + }, + { + "epoch": 1.0449074963190577, + "grad_norm": 0.7776521444320679, + "learning_rate": 4.057508661111686e-06, + "loss": 0.3795, + "num_input_tokens_seen": 49668816, + "step": 32645 + }, + { + "epoch": 1.0450675372895462, + "grad_norm": 0.44480666518211365, + "learning_rate": 4.052148645503648e-06, + "loss": 0.5985, + "num_input_tokens_seen": 49677024, + "step": 32650 + }, + { + "epoch": 1.0452275782600347, + "grad_norm": 0.892900824546814, + "learning_rate": 4.046791860317531e-06, + "loss": 0.4746, + "num_input_tokens_seen": 49685040, + "step": 32655 + }, + { + "epoch": 1.045387619230523, + "grad_norm": 0.6516596078872681, + "learning_rate": 4.041438306379431e-06, + "loss": 0.4426, + "num_input_tokens_seen": 49692880, + "step": 32660 + }, + { + "epoch": 1.0455476602010114, + "grad_norm": 0.5772596001625061, + "learning_rate": 4.036087984514916e-06, + "loss": 0.5214, + "num_input_tokens_seen": 49700912, + "step": 32665 + }, + { + "epoch": 1.0457077011715, + "grad_norm": 0.5854557156562805, + "learning_rate": 4.030740895549084e-06, + "loss": 0.4167, + "num_input_tokens_seen": 49708496, + "step": 32670 + }, + { + "epoch": 1.0458677421419884, + "grad_norm": 0.6286879181861877, + "learning_rate": 4.025397040306531e-06, + "loss": 0.4813, + "num_input_tokens_seen": 49715920, + "step": 32675 + }, + { + "epoch": 1.0460277831124767, + "grad_norm": 0.6368508338928223, + "learning_rate": 4.0200564196113285e-06, + "loss": 0.4113, + "num_input_tokens_seen": 49723808, + "step": 32680 + }, + { + "epoch": 1.0461878240829652, + "grad_norm": 1.0262724161148071, + "learning_rate": 4.014719034287079e-06, + "loss": 0.6348, + "num_input_tokens_seen": 49731936, + "step": 32685 + }, + { + "epoch": 1.0463478650534537, + "grad_norm": 1.1707850694656372, + "learning_rate": 4.0093848851568775e-06, + "loss": 0.5234, + "num_input_tokens_seen": 49739744, + "step": 32690 + }, + { + "epoch": 1.0465079060239422, + "grad_norm": 0.6955491900444031, + "learning_rate": 4.004053973043304e-06, + "loss": 0.4586, + "num_input_tokens_seen": 49747216, + "step": 32695 + }, + { + "epoch": 1.0466679469944307, + "grad_norm": 0.7567540407180786, + "learning_rate": 3.998726298768465e-06, + "loss": 0.5037, + "num_input_tokens_seen": 49754864, + "step": 32700 + }, + { + "epoch": 1.046827987964919, + "grad_norm": 0.9232239723205566, + "learning_rate": 3.99340186315395e-06, + "loss": 0.5372, + "num_input_tokens_seen": 49762192, + "step": 32705 + }, + { + "epoch": 1.0469880289354074, + "grad_norm": 0.583433985710144, + "learning_rate": 3.988080667020849e-06, + "loss": 0.3413, + "num_input_tokens_seen": 49770000, + "step": 32710 + }, + { + "epoch": 1.047148069905896, + "grad_norm": 0.7647145390510559, + "learning_rate": 3.982762711189766e-06, + "loss": 0.5171, + "num_input_tokens_seen": 49777824, + "step": 32715 + }, + { + "epoch": 1.0473081108763844, + "grad_norm": 1.224648356437683, + "learning_rate": 3.977447996480785e-06, + "loss": 0.4014, + "num_input_tokens_seen": 49785792, + "step": 32720 + }, + { + "epoch": 1.0474681518468727, + "grad_norm": 2.031275987625122, + "learning_rate": 3.97213652371351e-06, + "loss": 0.563, + "num_input_tokens_seen": 49794000, + "step": 32725 + }, + { + "epoch": 1.0476281928173612, + "grad_norm": 0.815695583820343, + "learning_rate": 3.966828293707042e-06, + "loss": 0.4247, + "num_input_tokens_seen": 49801872, + "step": 32730 + }, + { + "epoch": 1.0477882337878497, + "grad_norm": 1.9697695970535278, + "learning_rate": 3.961523307279963e-06, + "loss": 0.5134, + "num_input_tokens_seen": 49809456, + "step": 32735 + }, + { + "epoch": 1.0479482747583382, + "grad_norm": 0.6752428412437439, + "learning_rate": 3.956221565250382e-06, + "loss": 0.5315, + "num_input_tokens_seen": 49817488, + "step": 32740 + }, + { + "epoch": 1.0481083157288267, + "grad_norm": 0.4550950527191162, + "learning_rate": 3.950923068435883e-06, + "loss": 0.4352, + "num_input_tokens_seen": 49825696, + "step": 32745 + }, + { + "epoch": 1.048268356699315, + "grad_norm": 1.0083410739898682, + "learning_rate": 3.945627817653566e-06, + "loss": 0.5079, + "num_input_tokens_seen": 49833248, + "step": 32750 + }, + { + "epoch": 1.0484283976698034, + "grad_norm": 0.5169435739517212, + "learning_rate": 3.9403358137200335e-06, + "loss": 0.3477, + "num_input_tokens_seen": 49841344, + "step": 32755 + }, + { + "epoch": 1.048588438640292, + "grad_norm": 0.8119721412658691, + "learning_rate": 3.9350470574513605e-06, + "loss": 0.6035, + "num_input_tokens_seen": 49849296, + "step": 32760 + }, + { + "epoch": 1.0487484796107804, + "grad_norm": 0.8548628687858582, + "learning_rate": 3.9297615496631525e-06, + "loss": 0.3679, + "num_input_tokens_seen": 49857136, + "step": 32765 + }, + { + "epoch": 1.0489085205812687, + "grad_norm": 0.6411650776863098, + "learning_rate": 3.924479291170505e-06, + "loss": 0.4185, + "num_input_tokens_seen": 49864864, + "step": 32770 + }, + { + "epoch": 1.0490685615517572, + "grad_norm": 1.4716030359268188, + "learning_rate": 3.919200282788002e-06, + "loss": 0.4687, + "num_input_tokens_seen": 49872864, + "step": 32775 + }, + { + "epoch": 1.0492286025222457, + "grad_norm": 1.0885158777236938, + "learning_rate": 3.913924525329726e-06, + "loss": 0.465, + "num_input_tokens_seen": 49880368, + "step": 32780 + }, + { + "epoch": 1.0493886434927342, + "grad_norm": 0.9542534947395325, + "learning_rate": 3.908652019609279e-06, + "loss": 0.4922, + "num_input_tokens_seen": 49887824, + "step": 32785 + }, + { + "epoch": 1.0495486844632227, + "grad_norm": 1.0788148641586304, + "learning_rate": 3.9033827664397364e-06, + "loss": 0.5373, + "num_input_tokens_seen": 49895376, + "step": 32790 + }, + { + "epoch": 1.049708725433711, + "grad_norm": 0.656525194644928, + "learning_rate": 3.898116766633694e-06, + "loss": 0.5599, + "num_input_tokens_seen": 49902368, + "step": 32795 + }, + { + "epoch": 1.0498687664041995, + "grad_norm": 0.8040369153022766, + "learning_rate": 3.8928540210032225e-06, + "loss": 0.4776, + "num_input_tokens_seen": 49909872, + "step": 32800 + }, + { + "epoch": 1.0498687664041995, + "eval_loss": 0.4948614835739136, + "eval_runtime": 557.6879, + "eval_samples_per_second": 24.899, + "eval_steps_per_second": 12.45, + "num_input_tokens_seen": 49909872, + "step": 32800 + }, + { + "epoch": 1.050028807374688, + "grad_norm": 0.6759289503097534, + "learning_rate": 3.887594530359909e-06, + "loss": 0.3629, + "num_input_tokens_seen": 49917680, + "step": 32805 + }, + { + "epoch": 1.0501888483451765, + "grad_norm": 0.6306803822517395, + "learning_rate": 3.88233829551484e-06, + "loss": 0.4856, + "num_input_tokens_seen": 49925456, + "step": 32810 + }, + { + "epoch": 1.0503488893156647, + "grad_norm": 0.5388736128807068, + "learning_rate": 3.877085317278581e-06, + "loss": 0.4057, + "num_input_tokens_seen": 49933120, + "step": 32815 + }, + { + "epoch": 1.0505089302861532, + "grad_norm": 1.636193871498108, + "learning_rate": 3.87183559646122e-06, + "loss": 0.5364, + "num_input_tokens_seen": 49940656, + "step": 32820 + }, + { + "epoch": 1.0506689712566417, + "grad_norm": 0.9484972953796387, + "learning_rate": 3.866589133872317e-06, + "loss": 0.5907, + "num_input_tokens_seen": 49948576, + "step": 32825 + }, + { + "epoch": 1.0508290122271302, + "grad_norm": 1.0365580320358276, + "learning_rate": 3.861345930320948e-06, + "loss": 0.5058, + "num_input_tokens_seen": 49956080, + "step": 32830 + }, + { + "epoch": 1.0509890531976185, + "grad_norm": 1.124016284942627, + "learning_rate": 3.856105986615688e-06, + "loss": 0.5715, + "num_input_tokens_seen": 49964032, + "step": 32835 + }, + { + "epoch": 1.051149094168107, + "grad_norm": 0.623813807964325, + "learning_rate": 3.850869303564589e-06, + "loss": 0.3719, + "num_input_tokens_seen": 49971904, + "step": 32840 + }, + { + "epoch": 1.0513091351385955, + "grad_norm": 1.0121815204620361, + "learning_rate": 3.845635881975226e-06, + "loss": 0.6647, + "num_input_tokens_seen": 49980096, + "step": 32845 + }, + { + "epoch": 1.051469176109084, + "grad_norm": 0.7951675057411194, + "learning_rate": 3.840405722654647e-06, + "loss": 0.6849, + "num_input_tokens_seen": 49988336, + "step": 32850 + }, + { + "epoch": 1.0516292170795725, + "grad_norm": 1.3978056907653809, + "learning_rate": 3.835178826409419e-06, + "loss": 0.5119, + "num_input_tokens_seen": 49995904, + "step": 32855 + }, + { + "epoch": 1.0517892580500607, + "grad_norm": 0.8666081428527832, + "learning_rate": 3.8299551940455895e-06, + "loss": 0.3888, + "num_input_tokens_seen": 50003424, + "step": 32860 + }, + { + "epoch": 1.0519492990205492, + "grad_norm": 0.6705664992332458, + "learning_rate": 3.824734826368703e-06, + "loss": 0.3654, + "num_input_tokens_seen": 50011168, + "step": 32865 + }, + { + "epoch": 1.0521093399910377, + "grad_norm": 0.9219621419906616, + "learning_rate": 3.819517724183813e-06, + "loss": 0.3806, + "num_input_tokens_seen": 50018976, + "step": 32870 + }, + { + "epoch": 1.0522693809615262, + "grad_norm": 0.817934513092041, + "learning_rate": 3.8143038882954648e-06, + "loss": 0.5903, + "num_input_tokens_seen": 50026352, + "step": 32875 + }, + { + "epoch": 1.0524294219320145, + "grad_norm": 0.8566070199012756, + "learning_rate": 3.8090933195076867e-06, + "loss": 0.536, + "num_input_tokens_seen": 50034128, + "step": 32880 + }, + { + "epoch": 1.052589462902503, + "grad_norm": 0.7374052405357361, + "learning_rate": 3.8038860186240198e-06, + "loss": 0.4358, + "num_input_tokens_seen": 50042400, + "step": 32885 + }, + { + "epoch": 1.0527495038729915, + "grad_norm": 1.0742841958999634, + "learning_rate": 3.7986819864475026e-06, + "loss": 0.4937, + "num_input_tokens_seen": 50049392, + "step": 32890 + }, + { + "epoch": 1.05290954484348, + "grad_norm": 1.016737937927246, + "learning_rate": 3.793481223780651e-06, + "loss": 0.5504, + "num_input_tokens_seen": 50056752, + "step": 32895 + }, + { + "epoch": 1.0530695858139685, + "grad_norm": 1.3766710758209229, + "learning_rate": 3.788283731425496e-06, + "loss": 0.4776, + "num_input_tokens_seen": 50064400, + "step": 32900 + }, + { + "epoch": 1.0532296267844568, + "grad_norm": 0.6732779145240784, + "learning_rate": 3.7830895101835488e-06, + "loss": 0.3804, + "num_input_tokens_seen": 50072096, + "step": 32905 + }, + { + "epoch": 1.0533896677549452, + "grad_norm": 0.3032776415348053, + "learning_rate": 3.7778985608558274e-06, + "loss": 0.3756, + "num_input_tokens_seen": 50079872, + "step": 32910 + }, + { + "epoch": 1.0535497087254337, + "grad_norm": 1.656800389289856, + "learning_rate": 3.7727108842428443e-06, + "loss": 0.5798, + "num_input_tokens_seen": 50087152, + "step": 32915 + }, + { + "epoch": 1.0537097496959222, + "grad_norm": 1.5413793325424194, + "learning_rate": 3.7675264811446065e-06, + "loss": 0.4377, + "num_input_tokens_seen": 50094656, + "step": 32920 + }, + { + "epoch": 1.0538697906664105, + "grad_norm": 0.7500424385070801, + "learning_rate": 3.7623453523605994e-06, + "loss": 0.6834, + "num_input_tokens_seen": 50102000, + "step": 32925 + }, + { + "epoch": 1.054029831636899, + "grad_norm": 0.7411509156227112, + "learning_rate": 3.757167498689834e-06, + "loss": 0.4199, + "num_input_tokens_seen": 50109568, + "step": 32930 + }, + { + "epoch": 1.0541898726073875, + "grad_norm": 0.7838337421417236, + "learning_rate": 3.7519929209307914e-06, + "loss": 0.5081, + "num_input_tokens_seen": 50116832, + "step": 32935 + }, + { + "epoch": 1.054349913577876, + "grad_norm": 0.3791030943393707, + "learning_rate": 3.746821619881463e-06, + "loss": 0.4266, + "num_input_tokens_seen": 50124944, + "step": 32940 + }, + { + "epoch": 1.0545099545483643, + "grad_norm": 1.2719687223434448, + "learning_rate": 3.74165359633932e-06, + "loss": 0.3944, + "num_input_tokens_seen": 50132544, + "step": 32945 + }, + { + "epoch": 1.0546699955188528, + "grad_norm": 0.7095018029212952, + "learning_rate": 3.736488851101341e-06, + "loss": 0.6004, + "num_input_tokens_seen": 50140448, + "step": 32950 + }, + { + "epoch": 1.0548300364893413, + "grad_norm": 0.9139043688774109, + "learning_rate": 3.7313273849640035e-06, + "loss": 0.5676, + "num_input_tokens_seen": 50148384, + "step": 32955 + }, + { + "epoch": 1.0549900774598298, + "grad_norm": 0.9521812796592712, + "learning_rate": 3.7261691987232533e-06, + "loss": 0.5249, + "num_input_tokens_seen": 50156352, + "step": 32960 + }, + { + "epoch": 1.0551501184303183, + "grad_norm": 1.5175402164459229, + "learning_rate": 3.7210142931745575e-06, + "loss": 0.544, + "num_input_tokens_seen": 50163344, + "step": 32965 + }, + { + "epoch": 1.0553101594008065, + "grad_norm": 0.8001222610473633, + "learning_rate": 3.7158626691128712e-06, + "loss": 0.4867, + "num_input_tokens_seen": 50171408, + "step": 32970 + }, + { + "epoch": 1.055470200371295, + "grad_norm": 0.6075375080108643, + "learning_rate": 3.710714327332629e-06, + "loss": 0.4571, + "num_input_tokens_seen": 50178736, + "step": 32975 + }, + { + "epoch": 1.0556302413417835, + "grad_norm": 0.5460902452468872, + "learning_rate": 3.7055692686277815e-06, + "loss": 0.4244, + "num_input_tokens_seen": 50186240, + "step": 32980 + }, + { + "epoch": 1.055790282312272, + "grad_norm": 0.6824618577957153, + "learning_rate": 3.70042749379175e-06, + "loss": 0.4862, + "num_input_tokens_seen": 50195184, + "step": 32985 + }, + { + "epoch": 1.0559503232827603, + "grad_norm": 1.0554078817367554, + "learning_rate": 3.6952890036174693e-06, + "loss": 0.4497, + "num_input_tokens_seen": 50202656, + "step": 32990 + }, + { + "epoch": 1.0561103642532488, + "grad_norm": 0.5658578276634216, + "learning_rate": 3.690153798897353e-06, + "loss": 0.4523, + "num_input_tokens_seen": 50210144, + "step": 32995 + }, + { + "epoch": 1.0562704052237373, + "grad_norm": 0.6891281008720398, + "learning_rate": 3.6850218804233225e-06, + "loss": 0.6094, + "num_input_tokens_seen": 50217632, + "step": 33000 + }, + { + "epoch": 1.0562704052237373, + "eval_loss": 0.49489825963974, + "eval_runtime": 557.5102, + "eval_samples_per_second": 24.907, + "eval_steps_per_second": 12.454, + "num_input_tokens_seen": 50217632, + "step": 33000 + }, + { + "epoch": 1.0564304461942258, + "grad_norm": 0.3682490885257721, + "learning_rate": 3.679893248986779e-06, + "loss": 0.4656, + "num_input_tokens_seen": 50225040, + "step": 33005 + }, + { + "epoch": 1.0565904871647143, + "grad_norm": 0.42600247263908386, + "learning_rate": 3.6747679053786147e-06, + "loss": 0.4072, + "num_input_tokens_seen": 50232512, + "step": 33010 + }, + { + "epoch": 1.0567505281352025, + "grad_norm": 1.2984774112701416, + "learning_rate": 3.669645850389228e-06, + "loss": 0.567, + "num_input_tokens_seen": 50240432, + "step": 33015 + }, + { + "epoch": 1.056910569105691, + "grad_norm": 0.4513581395149231, + "learning_rate": 3.664527084808514e-06, + "loss": 0.5152, + "num_input_tokens_seen": 50247872, + "step": 33020 + }, + { + "epoch": 1.0570706100761795, + "grad_norm": 0.6777825355529785, + "learning_rate": 3.6594116094258337e-06, + "loss": 0.4667, + "num_input_tokens_seen": 50255984, + "step": 33025 + }, + { + "epoch": 1.057230651046668, + "grad_norm": 1.8678697347640991, + "learning_rate": 3.6542994250300665e-06, + "loss": 0.7522, + "num_input_tokens_seen": 50263184, + "step": 33030 + }, + { + "epoch": 1.0573906920171563, + "grad_norm": 0.5163565874099731, + "learning_rate": 3.6491905324095825e-06, + "loss": 0.4399, + "num_input_tokens_seen": 50270528, + "step": 33035 + }, + { + "epoch": 1.0575507329876448, + "grad_norm": 1.0965523719787598, + "learning_rate": 3.644084932352221e-06, + "loss": 0.4846, + "num_input_tokens_seen": 50277840, + "step": 33040 + }, + { + "epoch": 1.0577107739581333, + "grad_norm": 1.1398390531539917, + "learning_rate": 3.6389826256453457e-06, + "loss": 0.4792, + "num_input_tokens_seen": 50284976, + "step": 33045 + }, + { + "epoch": 1.0578708149286218, + "grad_norm": 0.5196439623832703, + "learning_rate": 3.633883613075781e-06, + "loss": 0.4377, + "num_input_tokens_seen": 50292304, + "step": 33050 + }, + { + "epoch": 1.0580308558991103, + "grad_norm": 0.37406203150749207, + "learning_rate": 3.6287878954298693e-06, + "loss": 0.5312, + "num_input_tokens_seen": 50299712, + "step": 33055 + }, + { + "epoch": 1.0581908968695986, + "grad_norm": 0.6296176314353943, + "learning_rate": 3.6236954734934354e-06, + "loss": 0.3813, + "num_input_tokens_seen": 50307456, + "step": 33060 + }, + { + "epoch": 1.058350937840087, + "grad_norm": 1.2439020872116089, + "learning_rate": 3.618606348051784e-06, + "loss": 0.4285, + "num_input_tokens_seen": 50314960, + "step": 33065 + }, + { + "epoch": 1.0585109788105755, + "grad_norm": 0.6756746172904968, + "learning_rate": 3.6135205198897376e-06, + "loss": 0.542, + "num_input_tokens_seen": 50322224, + "step": 33070 + }, + { + "epoch": 1.058671019781064, + "grad_norm": 0.9441954493522644, + "learning_rate": 3.6084379897915854e-06, + "loss": 0.5832, + "num_input_tokens_seen": 50329568, + "step": 33075 + }, + { + "epoch": 1.0588310607515523, + "grad_norm": 0.5718525648117065, + "learning_rate": 3.6033587585411115e-06, + "loss": 0.5283, + "num_input_tokens_seen": 50337136, + "step": 33080 + }, + { + "epoch": 1.0589911017220408, + "grad_norm": 0.725546658039093, + "learning_rate": 3.5982828269216117e-06, + "loss": 0.4993, + "num_input_tokens_seen": 50344336, + "step": 33085 + }, + { + "epoch": 1.0591511426925293, + "grad_norm": 0.6967185139656067, + "learning_rate": 3.593210195715843e-06, + "loss": 0.4312, + "num_input_tokens_seen": 50351792, + "step": 33090 + }, + { + "epoch": 1.0593111836630178, + "grad_norm": 0.3841772973537445, + "learning_rate": 3.5881408657060773e-06, + "loss": 0.3413, + "num_input_tokens_seen": 50359312, + "step": 33095 + }, + { + "epoch": 1.059471224633506, + "grad_norm": 0.6935784816741943, + "learning_rate": 3.583074837674075e-06, + "loss": 0.5597, + "num_input_tokens_seen": 50366976, + "step": 33100 + }, + { + "epoch": 1.0596312656039946, + "grad_norm": 0.9352164268493652, + "learning_rate": 3.578012112401069e-06, + "loss": 0.4751, + "num_input_tokens_seen": 50374224, + "step": 33105 + }, + { + "epoch": 1.059791306574483, + "grad_norm": 1.006023645401001, + "learning_rate": 3.5729526906677996e-06, + "loss": 0.5084, + "num_input_tokens_seen": 50382096, + "step": 33110 + }, + { + "epoch": 1.0599513475449716, + "grad_norm": 1.0520684719085693, + "learning_rate": 3.5678965732545007e-06, + "loss": 0.4362, + "num_input_tokens_seen": 50390432, + "step": 33115 + }, + { + "epoch": 1.06011138851546, + "grad_norm": 0.6040939092636108, + "learning_rate": 3.562843760940876e-06, + "loss": 0.3424, + "num_input_tokens_seen": 50398000, + "step": 33120 + }, + { + "epoch": 1.0602714294859483, + "grad_norm": 1.205391764640808, + "learning_rate": 3.5577942545061473e-06, + "loss": 0.4937, + "num_input_tokens_seen": 50405552, + "step": 33125 + }, + { + "epoch": 1.0604314704564368, + "grad_norm": 0.761845052242279, + "learning_rate": 3.5527480547289967e-06, + "loss": 0.3523, + "num_input_tokens_seen": 50412800, + "step": 33130 + }, + { + "epoch": 1.0605915114269253, + "grad_norm": 0.787491500377655, + "learning_rate": 3.547705162387624e-06, + "loss": 0.4618, + "num_input_tokens_seen": 50420208, + "step": 33135 + }, + { + "epoch": 1.0607515523974138, + "grad_norm": 0.6776275634765625, + "learning_rate": 3.542665578259699e-06, + "loss": 0.5439, + "num_input_tokens_seen": 50428944, + "step": 33140 + }, + { + "epoch": 1.060911593367902, + "grad_norm": 1.0092053413391113, + "learning_rate": 3.5376293031223945e-06, + "loss": 0.5831, + "num_input_tokens_seen": 50436336, + "step": 33145 + }, + { + "epoch": 1.0610716343383906, + "grad_norm": 1.070068359375, + "learning_rate": 3.5325963377523614e-06, + "loss": 0.7428, + "num_input_tokens_seen": 50443440, + "step": 33150 + }, + { + "epoch": 1.061231675308879, + "grad_norm": 0.9336593747138977, + "learning_rate": 3.5275666829257536e-06, + "loss": 0.6239, + "num_input_tokens_seen": 50451104, + "step": 33155 + }, + { + "epoch": 1.0613917162793676, + "grad_norm": 0.8425711989402771, + "learning_rate": 3.5225403394181955e-06, + "loss": 0.4302, + "num_input_tokens_seen": 50458832, + "step": 33160 + }, + { + "epoch": 1.061551757249856, + "grad_norm": 0.5504378080368042, + "learning_rate": 3.517517308004828e-06, + "loss": 0.4487, + "num_input_tokens_seen": 50466256, + "step": 33165 + }, + { + "epoch": 1.0617117982203443, + "grad_norm": 0.5040444135665894, + "learning_rate": 3.512497589460251e-06, + "loss": 0.3193, + "num_input_tokens_seen": 50474016, + "step": 33170 + }, + { + "epoch": 1.0618718391908328, + "grad_norm": 0.9207772016525269, + "learning_rate": 3.5074811845585727e-06, + "loss": 0.3732, + "num_input_tokens_seen": 50481600, + "step": 33175 + }, + { + "epoch": 1.0620318801613213, + "grad_norm": 0.7849960327148438, + "learning_rate": 3.5024680940733937e-06, + "loss": 0.4083, + "num_input_tokens_seen": 50489088, + "step": 33180 + }, + { + "epoch": 1.0621919211318098, + "grad_norm": 1.217895269393921, + "learning_rate": 3.4974583187777852e-06, + "loss": 0.4102, + "num_input_tokens_seen": 50496480, + "step": 33185 + }, + { + "epoch": 1.062351962102298, + "grad_norm": 1.0280841588974, + "learning_rate": 3.4924518594443204e-06, + "loss": 0.5777, + "num_input_tokens_seen": 50503664, + "step": 33190 + }, + { + "epoch": 1.0625120030727866, + "grad_norm": 0.6738953590393066, + "learning_rate": 3.4874487168450682e-06, + "loss": 0.4137, + "num_input_tokens_seen": 50511088, + "step": 33195 + }, + { + "epoch": 1.062672044043275, + "grad_norm": 0.4484426975250244, + "learning_rate": 3.482448891751558e-06, + "loss": 0.394, + "num_input_tokens_seen": 50518112, + "step": 33200 + }, + { + "epoch": 1.062672044043275, + "eval_loss": 0.4948088824748993, + "eval_runtime": 560.1132, + "eval_samples_per_second": 24.791, + "eval_steps_per_second": 12.396, + "num_input_tokens_seen": 50518112, + "step": 33200 + }, + { + "epoch": 1.0628320850137636, + "grad_norm": 1.164926528930664, + "learning_rate": 3.477452384934843e-06, + "loss": 0.6218, + "num_input_tokens_seen": 50525696, + "step": 33205 + }, + { + "epoch": 1.0629921259842519, + "grad_norm": 1.1275498867034912, + "learning_rate": 3.472459197165434e-06, + "loss": 0.6604, + "num_input_tokens_seen": 50533248, + "step": 33210 + }, + { + "epoch": 1.0631521669547404, + "grad_norm": 0.3651876449584961, + "learning_rate": 3.4674693292133518e-06, + "loss": 0.4544, + "num_input_tokens_seen": 50540432, + "step": 33215 + }, + { + "epoch": 1.0633122079252288, + "grad_norm": 0.45435622334480286, + "learning_rate": 3.4624827818480977e-06, + "loss": 0.3089, + "num_input_tokens_seen": 50548080, + "step": 33220 + }, + { + "epoch": 1.0634722488957173, + "grad_norm": 0.9009837508201599, + "learning_rate": 3.4574995558386474e-06, + "loss": 0.4821, + "num_input_tokens_seen": 50555616, + "step": 33225 + }, + { + "epoch": 1.0636322898662058, + "grad_norm": 1.8720651865005493, + "learning_rate": 3.452519651953487e-06, + "loss": 0.3403, + "num_input_tokens_seen": 50563328, + "step": 33230 + }, + { + "epoch": 1.0637923308366941, + "grad_norm": 1.0416566133499146, + "learning_rate": 3.447543070960585e-06, + "loss": 0.606, + "num_input_tokens_seen": 50570640, + "step": 33235 + }, + { + "epoch": 1.0639523718071826, + "grad_norm": 1.0939061641693115, + "learning_rate": 3.4425698136273778e-06, + "loss": 0.3621, + "num_input_tokens_seen": 50577872, + "step": 33240 + }, + { + "epoch": 1.064112412777671, + "grad_norm": 0.6957505941390991, + "learning_rate": 3.437599880720821e-06, + "loss": 0.3813, + "num_input_tokens_seen": 50585488, + "step": 33245 + }, + { + "epoch": 1.0642724537481596, + "grad_norm": 0.834675133228302, + "learning_rate": 3.4326332730073267e-06, + "loss": 0.5324, + "num_input_tokens_seen": 50593136, + "step": 33250 + }, + { + "epoch": 1.0644324947186479, + "grad_norm": 0.516710102558136, + "learning_rate": 3.427669991252813e-06, + "loss": 0.3263, + "num_input_tokens_seen": 50600592, + "step": 33255 + }, + { + "epoch": 1.0645925356891364, + "grad_norm": 1.3020869493484497, + "learning_rate": 3.42271003622269e-06, + "loss": 0.6592, + "num_input_tokens_seen": 50608544, + "step": 33260 + }, + { + "epoch": 1.0647525766596249, + "grad_norm": 0.4848802387714386, + "learning_rate": 3.4177534086818286e-06, + "loss": 0.391, + "num_input_tokens_seen": 50616176, + "step": 33265 + }, + { + "epoch": 1.0649126176301134, + "grad_norm": 1.0061906576156616, + "learning_rate": 3.412800109394612e-06, + "loss": 0.311, + "num_input_tokens_seen": 50623872, + "step": 33270 + }, + { + "epoch": 1.0650726586006019, + "grad_norm": 0.8999201059341431, + "learning_rate": 3.4078501391249044e-06, + "loss": 0.4905, + "num_input_tokens_seen": 50631584, + "step": 33275 + }, + { + "epoch": 1.0652326995710901, + "grad_norm": 0.7023124694824219, + "learning_rate": 3.4029034986360453e-06, + "loss": 0.5271, + "num_input_tokens_seen": 50639504, + "step": 33280 + }, + { + "epoch": 1.0653927405415786, + "grad_norm": 0.45717084407806396, + "learning_rate": 3.397960188690877e-06, + "loss": 0.3959, + "num_input_tokens_seen": 50647152, + "step": 33285 + }, + { + "epoch": 1.0655527815120671, + "grad_norm": 0.7940137982368469, + "learning_rate": 3.393020210051717e-06, + "loss": 0.5306, + "num_input_tokens_seen": 50654672, + "step": 33290 + }, + { + "epoch": 1.0657128224825556, + "grad_norm": 0.8097113370895386, + "learning_rate": 3.3880835634803655e-06, + "loss": 0.4688, + "num_input_tokens_seen": 50662608, + "step": 33295 + }, + { + "epoch": 1.0658728634530439, + "grad_norm": 0.6903155446052551, + "learning_rate": 3.383150249738126e-06, + "loss": 0.3769, + "num_input_tokens_seen": 50670032, + "step": 33300 + }, + { + "epoch": 1.0660329044235324, + "grad_norm": 0.7182335257530212, + "learning_rate": 3.3782202695857663e-06, + "loss": 0.348, + "num_input_tokens_seen": 50677408, + "step": 33305 + }, + { + "epoch": 1.0661929453940209, + "grad_norm": 0.5832404494285583, + "learning_rate": 3.373293623783558e-06, + "loss": 0.3974, + "num_input_tokens_seen": 50684464, + "step": 33310 + }, + { + "epoch": 1.0663529863645094, + "grad_norm": 0.5018178820610046, + "learning_rate": 3.368370313091257e-06, + "loss": 0.4691, + "num_input_tokens_seen": 50691952, + "step": 33315 + }, + { + "epoch": 1.0665130273349979, + "grad_norm": 0.8270835280418396, + "learning_rate": 3.363450338268087e-06, + "loss": 0.4808, + "num_input_tokens_seen": 50699936, + "step": 33320 + }, + { + "epoch": 1.0666730683054861, + "grad_norm": 1.2593090534210205, + "learning_rate": 3.358533700072783e-06, + "loss": 0.5651, + "num_input_tokens_seen": 50707696, + "step": 33325 + }, + { + "epoch": 1.0668331092759746, + "grad_norm": 0.5773563981056213, + "learning_rate": 3.3536203992635377e-06, + "loss": 0.4746, + "num_input_tokens_seen": 50715232, + "step": 33330 + }, + { + "epoch": 1.0669931502464631, + "grad_norm": 0.32453906536102295, + "learning_rate": 3.348710436598057e-06, + "loss": 0.529, + "num_input_tokens_seen": 50722896, + "step": 33335 + }, + { + "epoch": 1.0671531912169516, + "grad_norm": 0.3598928451538086, + "learning_rate": 3.3438038128335155e-06, + "loss": 0.3263, + "num_input_tokens_seen": 50730768, + "step": 33340 + }, + { + "epoch": 1.06731323218744, + "grad_norm": 0.99924236536026, + "learning_rate": 3.338900528726571e-06, + "loss": 0.3871, + "num_input_tokens_seen": 50738272, + "step": 33345 + }, + { + "epoch": 1.0674732731579284, + "grad_norm": 0.7011999487876892, + "learning_rate": 3.3340005850333812e-06, + "loss": 0.4779, + "num_input_tokens_seen": 50745936, + "step": 33350 + }, + { + "epoch": 1.0676333141284169, + "grad_norm": 0.5280786156654358, + "learning_rate": 3.329103982509568e-06, + "loss": 0.4504, + "num_input_tokens_seen": 50753520, + "step": 33355 + }, + { + "epoch": 1.0677933550989054, + "grad_norm": 0.7835909128189087, + "learning_rate": 3.324210721910259e-06, + "loss": 0.4572, + "num_input_tokens_seen": 50761248, + "step": 33360 + }, + { + "epoch": 1.0679533960693939, + "grad_norm": 0.2896532416343689, + "learning_rate": 3.319320803990053e-06, + "loss": 0.4696, + "num_input_tokens_seen": 50768944, + "step": 33365 + }, + { + "epoch": 1.0681134370398822, + "grad_norm": 0.5746126174926758, + "learning_rate": 3.3144342295030274e-06, + "loss": 0.3439, + "num_input_tokens_seen": 50776160, + "step": 33370 + }, + { + "epoch": 1.0682734780103706, + "grad_norm": 0.6011586785316467, + "learning_rate": 3.309550999202765e-06, + "loss": 0.5636, + "num_input_tokens_seen": 50783808, + "step": 33375 + }, + { + "epoch": 1.0684335189808591, + "grad_norm": 0.7637113332748413, + "learning_rate": 3.3046711138423197e-06, + "loss": 0.3406, + "num_input_tokens_seen": 50791328, + "step": 33380 + }, + { + "epoch": 1.0685935599513476, + "grad_norm": 0.6275800466537476, + "learning_rate": 3.2997945741742255e-06, + "loss": 0.4066, + "num_input_tokens_seen": 50798816, + "step": 33385 + }, + { + "epoch": 1.068753600921836, + "grad_norm": 0.7336696982383728, + "learning_rate": 3.2949213809505082e-06, + "loss": 0.3551, + "num_input_tokens_seen": 50805952, + "step": 33390 + }, + { + "epoch": 1.0689136418923244, + "grad_norm": 0.408801406621933, + "learning_rate": 3.2900515349226834e-06, + "loss": 0.378, + "num_input_tokens_seen": 50813584, + "step": 33395 + }, + { + "epoch": 1.069073682862813, + "grad_norm": 0.8080703616142273, + "learning_rate": 3.285185036841731e-06, + "loss": 0.4968, + "num_input_tokens_seen": 50821280, + "step": 33400 + }, + { + "epoch": 1.069073682862813, + "eval_loss": 0.4949004352092743, + "eval_runtime": 557.4402, + "eval_samples_per_second": 24.91, + "eval_steps_per_second": 12.455, + "num_input_tokens_seen": 50821280, + "step": 33400 + }, + { + "epoch": 1.0692337238333014, + "grad_norm": 1.1920619010925293, + "learning_rate": 3.2803218874581377e-06, + "loss": 0.4941, + "num_input_tokens_seen": 50828464, + "step": 33405 + }, + { + "epoch": 1.0693937648037897, + "grad_norm": 0.5845604538917542, + "learning_rate": 3.2754620875218494e-06, + "loss": 0.4036, + "num_input_tokens_seen": 50836128, + "step": 33410 + }, + { + "epoch": 1.0695538057742782, + "grad_norm": 0.900956392288208, + "learning_rate": 3.2706056377823146e-06, + "loss": 0.4962, + "num_input_tokens_seen": 50843936, + "step": 33415 + }, + { + "epoch": 1.0697138467447667, + "grad_norm": 1.057836890220642, + "learning_rate": 3.2657525389884647e-06, + "loss": 0.4818, + "num_input_tokens_seen": 50851648, + "step": 33420 + }, + { + "epoch": 1.0698738877152552, + "grad_norm": 0.681759238243103, + "learning_rate": 3.260902791888698e-06, + "loss": 0.4122, + "num_input_tokens_seen": 50859488, + "step": 33425 + }, + { + "epoch": 1.0700339286857437, + "grad_norm": 0.4086180627346039, + "learning_rate": 3.2560563972309166e-06, + "loss": 0.4401, + "num_input_tokens_seen": 50867968, + "step": 33430 + }, + { + "epoch": 1.070193969656232, + "grad_norm": 0.7586341500282288, + "learning_rate": 3.251213355762489e-06, + "loss": 0.4683, + "num_input_tokens_seen": 50875584, + "step": 33435 + }, + { + "epoch": 1.0703540106267204, + "grad_norm": 0.9337571263313293, + "learning_rate": 3.2463736682302707e-06, + "loss": 0.618, + "num_input_tokens_seen": 50883536, + "step": 33440 + }, + { + "epoch": 1.070514051597209, + "grad_norm": 0.6300714015960693, + "learning_rate": 3.2415373353806124e-06, + "loss": 0.5099, + "num_input_tokens_seen": 50891616, + "step": 33445 + }, + { + "epoch": 1.0706740925676974, + "grad_norm": 0.5023218393325806, + "learning_rate": 3.236704357959322e-06, + "loss": 0.3484, + "num_input_tokens_seen": 50899296, + "step": 33450 + }, + { + "epoch": 1.0708341335381857, + "grad_norm": 1.1965614557266235, + "learning_rate": 3.2318747367117154e-06, + "loss": 0.4088, + "num_input_tokens_seen": 50906880, + "step": 33455 + }, + { + "epoch": 1.0709941745086742, + "grad_norm": 1.245603084564209, + "learning_rate": 3.227048472382585e-06, + "loss": 0.7116, + "num_input_tokens_seen": 50914832, + "step": 33460 + }, + { + "epoch": 1.0711542154791627, + "grad_norm": 0.9898114204406738, + "learning_rate": 3.2222255657161915e-06, + "loss": 0.8189, + "num_input_tokens_seen": 50921936, + "step": 33465 + }, + { + "epoch": 1.0713142564496512, + "grad_norm": 0.6754611730575562, + "learning_rate": 3.2174060174562924e-06, + "loss": 0.4882, + "num_input_tokens_seen": 50930144, + "step": 33470 + }, + { + "epoch": 1.0714742974201394, + "grad_norm": 0.7227607369422913, + "learning_rate": 3.2125898283461298e-06, + "loss": 0.4006, + "num_input_tokens_seen": 50937536, + "step": 33475 + }, + { + "epoch": 1.071634338390628, + "grad_norm": 1.0764530897140503, + "learning_rate": 3.207776999128406e-06, + "loss": 0.4524, + "num_input_tokens_seen": 50945328, + "step": 33480 + }, + { + "epoch": 1.0717943793611164, + "grad_norm": 0.848405659198761, + "learning_rate": 3.202967530545331e-06, + "loss": 0.6567, + "num_input_tokens_seen": 50953104, + "step": 33485 + }, + { + "epoch": 1.071954420331605, + "grad_norm": 0.6887418627738953, + "learning_rate": 3.1981614233385778e-06, + "loss": 0.5041, + "num_input_tokens_seen": 50961360, + "step": 33490 + }, + { + "epoch": 1.0721144613020934, + "grad_norm": 0.6300256252288818, + "learning_rate": 3.1933586782493115e-06, + "loss": 0.3432, + "num_input_tokens_seen": 50968976, + "step": 33495 + }, + { + "epoch": 1.0722745022725817, + "grad_norm": 0.6643475294113159, + "learning_rate": 3.188559296018184e-06, + "loss": 0.4256, + "num_input_tokens_seen": 50976480, + "step": 33500 + }, + { + "epoch": 1.0724345432430702, + "grad_norm": 0.7973014116287231, + "learning_rate": 3.1837632773853098e-06, + "loss": 0.4695, + "num_input_tokens_seen": 50983632, + "step": 33505 + }, + { + "epoch": 1.0725945842135587, + "grad_norm": 1.1010456085205078, + "learning_rate": 3.178970623090294e-06, + "loss": 0.6122, + "num_input_tokens_seen": 50991168, + "step": 33510 + }, + { + "epoch": 1.0727546251840472, + "grad_norm": 0.7014950513839722, + "learning_rate": 3.174181333872234e-06, + "loss": 0.3996, + "num_input_tokens_seen": 50998368, + "step": 33515 + }, + { + "epoch": 1.0729146661545355, + "grad_norm": 0.4649421274662018, + "learning_rate": 3.169395410469686e-06, + "loss": 0.4028, + "num_input_tokens_seen": 51005680, + "step": 33520 + }, + { + "epoch": 1.073074707125024, + "grad_norm": 0.5818110704421997, + "learning_rate": 3.164612853620713e-06, + "loss": 0.5019, + "num_input_tokens_seen": 51013152, + "step": 33525 + }, + { + "epoch": 1.0732347480955124, + "grad_norm": 1.5416386127471924, + "learning_rate": 3.1598336640628333e-06, + "loss": 0.6991, + "num_input_tokens_seen": 51021104, + "step": 33530 + }, + { + "epoch": 1.073394789066001, + "grad_norm": 0.5466017127037048, + "learning_rate": 3.155057842533063e-06, + "loss": 0.3108, + "num_input_tokens_seen": 51028160, + "step": 33535 + }, + { + "epoch": 1.0735548300364894, + "grad_norm": 0.9873573184013367, + "learning_rate": 3.1502853897678984e-06, + "loss": 0.6066, + "num_input_tokens_seen": 51035456, + "step": 33540 + }, + { + "epoch": 1.0737148710069777, + "grad_norm": 0.9807007908821106, + "learning_rate": 3.1455163065033017e-06, + "loss": 0.5654, + "num_input_tokens_seen": 51043280, + "step": 33545 + }, + { + "epoch": 1.0738749119774662, + "grad_norm": 0.3968043923377991, + "learning_rate": 3.140750593474734e-06, + "loss": 0.368, + "num_input_tokens_seen": 51051536, + "step": 33550 + }, + { + "epoch": 1.0740349529479547, + "grad_norm": 0.434803307056427, + "learning_rate": 3.1359882514171294e-06, + "loss": 0.5082, + "num_input_tokens_seen": 51059104, + "step": 33555 + }, + { + "epoch": 1.0741949939184432, + "grad_norm": 0.7805004119873047, + "learning_rate": 3.1312292810648903e-06, + "loss": 0.4318, + "num_input_tokens_seen": 51066688, + "step": 33560 + }, + { + "epoch": 1.0743550348889315, + "grad_norm": 29.25476837158203, + "learning_rate": 3.1264736831519204e-06, + "loss": 0.6668, + "num_input_tokens_seen": 51074320, + "step": 33565 + }, + { + "epoch": 1.07451507585942, + "grad_norm": 0.5863790512084961, + "learning_rate": 3.1217214584115863e-06, + "loss": 0.4268, + "num_input_tokens_seen": 51082144, + "step": 33570 + }, + { + "epoch": 1.0746751168299085, + "grad_norm": 1.1931544542312622, + "learning_rate": 3.116972607576746e-06, + "loss": 0.4825, + "num_input_tokens_seen": 51089616, + "step": 33575 + }, + { + "epoch": 1.074835157800397, + "grad_norm": 1.091209888458252, + "learning_rate": 3.1122271313797303e-06, + "loss": 0.4491, + "num_input_tokens_seen": 51096976, + "step": 33580 + }, + { + "epoch": 1.0749951987708855, + "grad_norm": 0.6065694689750671, + "learning_rate": 3.107485030552343e-06, + "loss": 0.3937, + "num_input_tokens_seen": 51104320, + "step": 33585 + }, + { + "epoch": 1.0751552397413737, + "grad_norm": 0.5423709154129028, + "learning_rate": 3.1027463058258848e-06, + "loss": 0.3304, + "num_input_tokens_seen": 51112000, + "step": 33590 + }, + { + "epoch": 1.0753152807118622, + "grad_norm": 0.8045841455459595, + "learning_rate": 3.0980109579311273e-06, + "loss": 0.5605, + "num_input_tokens_seen": 51119392, + "step": 33595 + }, + { + "epoch": 1.0754753216823507, + "grad_norm": 0.7628235220909119, + "learning_rate": 3.093278987598314e-06, + "loss": 0.5216, + "num_input_tokens_seen": 51127232, + "step": 33600 + }, + { + "epoch": 1.0754753216823507, + "eval_loss": 0.49475330114364624, + "eval_runtime": 556.3875, + "eval_samples_per_second": 24.957, + "eval_steps_per_second": 12.479, + "num_input_tokens_seen": 51127232, + "step": 33600 + }, + { + "epoch": 1.0756353626528392, + "grad_norm": 0.4598935544490814, + "learning_rate": 3.0885503955571826e-06, + "loss": 0.5077, + "num_input_tokens_seen": 51134832, + "step": 33605 + }, + { + "epoch": 1.0757954036233275, + "grad_norm": 0.5868064761161804, + "learning_rate": 3.0838251825369313e-06, + "loss": 0.3958, + "num_input_tokens_seen": 51142672, + "step": 33610 + }, + { + "epoch": 1.075955444593816, + "grad_norm": 1.1750739812850952, + "learning_rate": 3.0791033492662517e-06, + "loss": 0.5255, + "num_input_tokens_seen": 51150544, + "step": 33615 + }, + { + "epoch": 1.0761154855643045, + "grad_norm": 0.7506466507911682, + "learning_rate": 3.0743848964733203e-06, + "loss": 0.4816, + "num_input_tokens_seen": 51158160, + "step": 33620 + }, + { + "epoch": 1.076275526534793, + "grad_norm": 0.3938434422016144, + "learning_rate": 3.0696698248857625e-06, + "loss": 0.3827, + "num_input_tokens_seen": 51165680, + "step": 33625 + }, + { + "epoch": 1.0764355675052815, + "grad_norm": 0.6311063170433044, + "learning_rate": 3.0649581352307192e-06, + "loss": 0.4665, + "num_input_tokens_seen": 51173216, + "step": 33630 + }, + { + "epoch": 1.0765956084757697, + "grad_norm": 1.0314486026763916, + "learning_rate": 3.060249828234776e-06, + "loss": 0.4319, + "num_input_tokens_seen": 51180816, + "step": 33635 + }, + { + "epoch": 1.0767556494462582, + "grad_norm": 1.2765454053878784, + "learning_rate": 3.055544904624025e-06, + "loss": 0.3394, + "num_input_tokens_seen": 51189088, + "step": 33640 + }, + { + "epoch": 1.0769156904167467, + "grad_norm": 0.6943764686584473, + "learning_rate": 3.050843365124026e-06, + "loss": 0.4537, + "num_input_tokens_seen": 51196432, + "step": 33645 + }, + { + "epoch": 1.0770757313872352, + "grad_norm": 0.6768283247947693, + "learning_rate": 3.0461452104598083e-06, + "loss": 0.4857, + "num_input_tokens_seen": 51204384, + "step": 33650 + }, + { + "epoch": 1.0772357723577235, + "grad_norm": 0.9304231405258179, + "learning_rate": 3.0414504413558836e-06, + "loss": 0.4386, + "num_input_tokens_seen": 51212656, + "step": 33655 + }, + { + "epoch": 1.077395813328212, + "grad_norm": 0.7623855471611023, + "learning_rate": 3.0367590585362564e-06, + "loss": 0.3836, + "num_input_tokens_seen": 51220416, + "step": 33660 + }, + { + "epoch": 1.0775558542987005, + "grad_norm": 1.0128535032272339, + "learning_rate": 3.0320710627243813e-06, + "loss": 0.404, + "num_input_tokens_seen": 51228944, + "step": 33665 + }, + { + "epoch": 1.077715895269189, + "grad_norm": 0.5511813759803772, + "learning_rate": 3.027386454643222e-06, + "loss": 0.4229, + "num_input_tokens_seen": 51236416, + "step": 33670 + }, + { + "epoch": 1.0778759362396773, + "grad_norm": 0.34353023767471313, + "learning_rate": 3.0227052350151914e-06, + "loss": 0.4032, + "num_input_tokens_seen": 51243904, + "step": 33675 + }, + { + "epoch": 1.0780359772101658, + "grad_norm": 0.46103358268737793, + "learning_rate": 3.0180274045621957e-06, + "loss": 0.3506, + "num_input_tokens_seen": 51251072, + "step": 33680 + }, + { + "epoch": 1.0781960181806542, + "grad_norm": 1.1836398839950562, + "learning_rate": 3.013352964005625e-06, + "loss": 0.4296, + "num_input_tokens_seen": 51259072, + "step": 33685 + }, + { + "epoch": 1.0783560591511427, + "grad_norm": 3.21876859664917, + "learning_rate": 3.0086819140663218e-06, + "loss": 0.4573, + "num_input_tokens_seen": 51266512, + "step": 33690 + }, + { + "epoch": 1.0785161001216312, + "grad_norm": 0.9231895804405212, + "learning_rate": 3.0040142554646265e-06, + "loss": 0.5764, + "num_input_tokens_seen": 51273840, + "step": 33695 + }, + { + "epoch": 1.0786761410921195, + "grad_norm": 0.9899126291275024, + "learning_rate": 2.999349988920361e-06, + "loss": 0.4618, + "num_input_tokens_seen": 51281888, + "step": 33700 + }, + { + "epoch": 1.078836182062608, + "grad_norm": 0.41833093762397766, + "learning_rate": 2.994689115152796e-06, + "loss": 0.3624, + "num_input_tokens_seen": 51289408, + "step": 33705 + }, + { + "epoch": 1.0789962230330965, + "grad_norm": 0.7591205835342407, + "learning_rate": 2.9900316348807105e-06, + "loss": 0.5223, + "num_input_tokens_seen": 51296912, + "step": 33710 + }, + { + "epoch": 1.079156264003585, + "grad_norm": 0.4678439795970917, + "learning_rate": 2.985377548822338e-06, + "loss": 0.5339, + "num_input_tokens_seen": 51304704, + "step": 33715 + }, + { + "epoch": 1.0793163049740733, + "grad_norm": 0.8622381091117859, + "learning_rate": 2.980726857695404e-06, + "loss": 0.432, + "num_input_tokens_seen": 51311872, + "step": 33720 + }, + { + "epoch": 1.0794763459445618, + "grad_norm": 0.7037981152534485, + "learning_rate": 2.9760795622171017e-06, + "loss": 0.4215, + "num_input_tokens_seen": 51318880, + "step": 33725 + }, + { + "epoch": 1.0796363869150503, + "grad_norm": 1.151924967765808, + "learning_rate": 2.971435663104094e-06, + "loss": 0.5461, + "num_input_tokens_seen": 51326528, + "step": 33730 + }, + { + "epoch": 1.0797964278855388, + "grad_norm": 0.6323187947273254, + "learning_rate": 2.9667951610725385e-06, + "loss": 0.3771, + "num_input_tokens_seen": 51333872, + "step": 33735 + }, + { + "epoch": 1.079956468856027, + "grad_norm": 0.4269039034843445, + "learning_rate": 2.9621580568380575e-06, + "loss": 0.488, + "num_input_tokens_seen": 51342176, + "step": 33740 + }, + { + "epoch": 1.0801165098265155, + "grad_norm": 0.8445913791656494, + "learning_rate": 2.9575243511157453e-06, + "loss": 0.435, + "num_input_tokens_seen": 51349616, + "step": 33745 + }, + { + "epoch": 1.080276550797004, + "grad_norm": 0.8769132494926453, + "learning_rate": 2.952894044620186e-06, + "loss": 0.5196, + "num_input_tokens_seen": 51356736, + "step": 33750 + }, + { + "epoch": 1.0804365917674925, + "grad_norm": 1.079931616783142, + "learning_rate": 2.948267138065419e-06, + "loss": 0.5129, + "num_input_tokens_seen": 51365120, + "step": 33755 + }, + { + "epoch": 1.080596632737981, + "grad_norm": 1.1865719556808472, + "learning_rate": 2.943643632164983e-06, + "loss": 0.5138, + "num_input_tokens_seen": 51373056, + "step": 33760 + }, + { + "epoch": 1.0807566737084693, + "grad_norm": 0.6316649913787842, + "learning_rate": 2.939023527631879e-06, + "loss": 0.3694, + "num_input_tokens_seen": 51380688, + "step": 33765 + }, + { + "epoch": 1.0809167146789578, + "grad_norm": 0.7644031047821045, + "learning_rate": 2.934406825178576e-06, + "loss": 0.4218, + "num_input_tokens_seen": 51388384, + "step": 33770 + }, + { + "epoch": 1.0810767556494463, + "grad_norm": 0.8549161553382874, + "learning_rate": 2.9297935255170357e-06, + "loss": 0.5542, + "num_input_tokens_seen": 51395968, + "step": 33775 + }, + { + "epoch": 1.0812367966199348, + "grad_norm": 0.6769658327102661, + "learning_rate": 2.925183629358691e-06, + "loss": 0.5158, + "num_input_tokens_seen": 51403248, + "step": 33780 + }, + { + "epoch": 1.081396837590423, + "grad_norm": 0.9975667595863342, + "learning_rate": 2.9205771374144346e-06, + "loss": 0.546, + "num_input_tokens_seen": 51411104, + "step": 33785 + }, + { + "epoch": 1.0815568785609115, + "grad_norm": 1.117472529411316, + "learning_rate": 2.915974050394657e-06, + "loss": 0.4647, + "num_input_tokens_seen": 51418912, + "step": 33790 + }, + { + "epoch": 1.0817169195314, + "grad_norm": 0.9747480750083923, + "learning_rate": 2.9113743690092067e-06, + "loss": 0.6648, + "num_input_tokens_seen": 51427168, + "step": 33795 + }, + { + "epoch": 1.0818769605018885, + "grad_norm": 0.703229546546936, + "learning_rate": 2.906778093967402e-06, + "loss": 0.5109, + "num_input_tokens_seen": 51435040, + "step": 33800 + }, + { + "epoch": 1.0818769605018885, + "eval_loss": 0.4945319890975952, + "eval_runtime": 557.5817, + "eval_samples_per_second": 24.904, + "eval_steps_per_second": 12.452, + "num_input_tokens_seen": 51435040, + "step": 33800 + }, + { + "epoch": 1.082037001472377, + "grad_norm": 0.6922637820243835, + "learning_rate": 2.9021852259780656e-06, + "loss": 0.541, + "num_input_tokens_seen": 51442224, + "step": 33805 + }, + { + "epoch": 1.0821970424428653, + "grad_norm": 0.9816201329231262, + "learning_rate": 2.8975957657494583e-06, + "loss": 0.5864, + "num_input_tokens_seen": 51450112, + "step": 33810 + }, + { + "epoch": 1.0823570834133538, + "grad_norm": 0.48372596502304077, + "learning_rate": 2.8930097139893417e-06, + "loss": 0.3852, + "num_input_tokens_seen": 51458000, + "step": 33815 + }, + { + "epoch": 1.0825171243838423, + "grad_norm": 1.0676666498184204, + "learning_rate": 2.888427071404945e-06, + "loss": 0.3285, + "num_input_tokens_seen": 51465440, + "step": 33820 + }, + { + "epoch": 1.0826771653543308, + "grad_norm": 0.6588717699050903, + "learning_rate": 2.8838478387029606e-06, + "loss": 0.7197, + "num_input_tokens_seen": 51473072, + "step": 33825 + }, + { + "epoch": 1.082837206324819, + "grad_norm": 0.7659130096435547, + "learning_rate": 2.8792720165895737e-06, + "loss": 0.4213, + "num_input_tokens_seen": 51480400, + "step": 33830 + }, + { + "epoch": 1.0829972472953076, + "grad_norm": 0.5101548433303833, + "learning_rate": 2.874699605770423e-06, + "loss": 0.4096, + "num_input_tokens_seen": 51488512, + "step": 33835 + }, + { + "epoch": 1.083157288265796, + "grad_norm": 0.4767008125782013, + "learning_rate": 2.8701306069506383e-06, + "loss": 0.3851, + "num_input_tokens_seen": 51495936, + "step": 33840 + }, + { + "epoch": 1.0833173292362845, + "grad_norm": 0.9356855154037476, + "learning_rate": 2.8655650208348178e-06, + "loss": 0.5152, + "num_input_tokens_seen": 51503568, + "step": 33845 + }, + { + "epoch": 1.083477370206773, + "grad_norm": 1.9294016361236572, + "learning_rate": 2.8610028481270257e-06, + "loss": 0.5179, + "num_input_tokens_seen": 51511056, + "step": 33850 + }, + { + "epoch": 1.0836374111772613, + "grad_norm": 1.1223523616790771, + "learning_rate": 2.856444089530813e-06, + "loss": 0.6372, + "num_input_tokens_seen": 51518864, + "step": 33855 + }, + { + "epoch": 1.0837974521477498, + "grad_norm": 1.3424129486083984, + "learning_rate": 2.8518887457491955e-06, + "loss": 0.6015, + "num_input_tokens_seen": 51526176, + "step": 33860 + }, + { + "epoch": 1.0839574931182383, + "grad_norm": 0.5809653997421265, + "learning_rate": 2.8473368174846666e-06, + "loss": 0.3927, + "num_input_tokens_seen": 51534080, + "step": 33865 + }, + { + "epoch": 1.0841175340887268, + "grad_norm": 0.8065551519393921, + "learning_rate": 2.842788305439184e-06, + "loss": 0.5759, + "num_input_tokens_seen": 51541904, + "step": 33870 + }, + { + "epoch": 1.084277575059215, + "grad_norm": 0.6231340169906616, + "learning_rate": 2.8382432103141925e-06, + "loss": 0.3873, + "num_input_tokens_seen": 51549536, + "step": 33875 + }, + { + "epoch": 1.0844376160297036, + "grad_norm": 0.5332165956497192, + "learning_rate": 2.833701532810598e-06, + "loss": 0.4611, + "num_input_tokens_seen": 51556992, + "step": 33880 + }, + { + "epoch": 1.084597657000192, + "grad_norm": 2.440246105194092, + "learning_rate": 2.8291632736287877e-06, + "loss": 0.4696, + "num_input_tokens_seen": 51564624, + "step": 33885 + }, + { + "epoch": 1.0847576979706806, + "grad_norm": 0.402132123708725, + "learning_rate": 2.824628433468615e-06, + "loss": 0.4209, + "num_input_tokens_seen": 51572576, + "step": 33890 + }, + { + "epoch": 1.084917738941169, + "grad_norm": 1.2935012578964233, + "learning_rate": 2.8200970130294073e-06, + "loss": 0.4523, + "num_input_tokens_seen": 51580736, + "step": 33895 + }, + { + "epoch": 1.0850777799116573, + "grad_norm": 0.9683690071105957, + "learning_rate": 2.8155690130099775e-06, + "loss": 0.5285, + "num_input_tokens_seen": 51588288, + "step": 33900 + }, + { + "epoch": 1.0852378208821458, + "grad_norm": 1.1657788753509521, + "learning_rate": 2.8110444341085895e-06, + "loss": 0.4353, + "num_input_tokens_seen": 51595936, + "step": 33905 + }, + { + "epoch": 1.0853978618526343, + "grad_norm": 0.44727030396461487, + "learning_rate": 2.806523277022996e-06, + "loss": 0.491, + "num_input_tokens_seen": 51603136, + "step": 33910 + }, + { + "epoch": 1.0855579028231228, + "grad_norm": 0.8592545986175537, + "learning_rate": 2.802005542450409e-06, + "loss": 0.474, + "num_input_tokens_seen": 51610352, + "step": 33915 + }, + { + "epoch": 1.085717943793611, + "grad_norm": 0.9166985750198364, + "learning_rate": 2.797491231087526e-06, + "loss": 0.6422, + "num_input_tokens_seen": 51617776, + "step": 33920 + }, + { + "epoch": 1.0858779847640996, + "grad_norm": 0.7380098700523376, + "learning_rate": 2.7929803436305137e-06, + "loss": 0.4131, + "num_input_tokens_seen": 51625200, + "step": 33925 + }, + { + "epoch": 1.086038025734588, + "grad_norm": 2.545482635498047, + "learning_rate": 2.788472880774998e-06, + "loss": 0.5227, + "num_input_tokens_seen": 51632752, + "step": 33930 + }, + { + "epoch": 1.0861980667050766, + "grad_norm": 0.7708050608634949, + "learning_rate": 2.7839688432160977e-06, + "loss": 0.4479, + "num_input_tokens_seen": 51640848, + "step": 33935 + }, + { + "epoch": 1.0863581076755648, + "grad_norm": 0.3868168294429779, + "learning_rate": 2.779468231648383e-06, + "loss": 0.4955, + "num_input_tokens_seen": 51648368, + "step": 33940 + }, + { + "epoch": 1.0865181486460533, + "grad_norm": 0.458212673664093, + "learning_rate": 2.774971046765906e-06, + "loss": 0.4349, + "num_input_tokens_seen": 51656352, + "step": 33945 + }, + { + "epoch": 1.0866781896165418, + "grad_norm": 0.9931617975234985, + "learning_rate": 2.770477289262194e-06, + "loss": 0.6383, + "num_input_tokens_seen": 51663520, + "step": 33950 + }, + { + "epoch": 1.0868382305870303, + "grad_norm": 0.37488481402397156, + "learning_rate": 2.765986959830233e-06, + "loss": 0.4399, + "num_input_tokens_seen": 51671552, + "step": 33955 + }, + { + "epoch": 1.0869982715575188, + "grad_norm": 0.9317154884338379, + "learning_rate": 2.761500059162492e-06, + "loss": 0.2974, + "num_input_tokens_seen": 51679072, + "step": 33960 + }, + { + "epoch": 1.087158312528007, + "grad_norm": 0.49784988164901733, + "learning_rate": 2.757016587950914e-06, + "loss": 0.3705, + "num_input_tokens_seen": 51686544, + "step": 33965 + }, + { + "epoch": 1.0873183534984956, + "grad_norm": 0.5112912058830261, + "learning_rate": 2.752536546886897e-06, + "loss": 0.4026, + "num_input_tokens_seen": 51694048, + "step": 33970 + }, + { + "epoch": 1.087478394468984, + "grad_norm": 0.7655388712882996, + "learning_rate": 2.7480599366613234e-06, + "loss": 0.4036, + "num_input_tokens_seen": 51701392, + "step": 33975 + }, + { + "epoch": 1.0876384354394726, + "grad_norm": 0.855472207069397, + "learning_rate": 2.7435867579645473e-06, + "loss": 0.5674, + "num_input_tokens_seen": 51709312, + "step": 33980 + }, + { + "epoch": 1.0877984764099609, + "grad_norm": 0.8532953262329102, + "learning_rate": 2.739117011486378e-06, + "loss": 0.4516, + "num_input_tokens_seen": 51716752, + "step": 33985 + }, + { + "epoch": 1.0879585173804494, + "grad_norm": 0.702150285243988, + "learning_rate": 2.7346506979161216e-06, + "loss": 0.5123, + "num_input_tokens_seen": 51724592, + "step": 33990 + }, + { + "epoch": 1.0881185583509378, + "grad_norm": 0.8381018042564392, + "learning_rate": 2.7301878179425227e-06, + "loss": 0.4206, + "num_input_tokens_seen": 51731904, + "step": 33995 + }, + { + "epoch": 1.0882785993214263, + "grad_norm": 1.64031183719635, + "learning_rate": 2.7257283722538244e-06, + "loss": 0.6351, + "num_input_tokens_seen": 51738784, + "step": 34000 + }, + { + "epoch": 1.0882785993214263, + "eval_loss": 0.4945540428161621, + "eval_runtime": 557.376, + "eval_samples_per_second": 24.913, + "eval_steps_per_second": 12.457, + "num_input_tokens_seen": 51738784, + "step": 34000 + }, + { + "epoch": 1.0884386402919146, + "grad_norm": 0.4989163875579834, + "learning_rate": 2.7212723615377326e-06, + "loss": 0.505, + "num_input_tokens_seen": 51746368, + "step": 34005 + }, + { + "epoch": 1.0885986812624031, + "grad_norm": 0.4867362678050995, + "learning_rate": 2.7168197864814145e-06, + "loss": 0.4371, + "num_input_tokens_seen": 51753936, + "step": 34010 + }, + { + "epoch": 1.0887587222328916, + "grad_norm": 0.6733598709106445, + "learning_rate": 2.712370647771509e-06, + "loss": 0.4463, + "num_input_tokens_seen": 51761264, + "step": 34015 + }, + { + "epoch": 1.08891876320338, + "grad_norm": 0.9694377779960632, + "learning_rate": 2.707924946094137e-06, + "loss": 0.5438, + "num_input_tokens_seen": 51768544, + "step": 34020 + }, + { + "epoch": 1.0890788041738686, + "grad_norm": 0.6768556237220764, + "learning_rate": 2.7034826821348723e-06, + "loss": 0.433, + "num_input_tokens_seen": 51776320, + "step": 34025 + }, + { + "epoch": 1.0892388451443569, + "grad_norm": 1.1705266237258911, + "learning_rate": 2.6990438565787786e-06, + "loss": 0.5171, + "num_input_tokens_seen": 51783952, + "step": 34030 + }, + { + "epoch": 1.0893988861148454, + "grad_norm": 0.6477959156036377, + "learning_rate": 2.6946084701103714e-06, + "loss": 0.3864, + "num_input_tokens_seen": 51791664, + "step": 34035 + }, + { + "epoch": 1.0895589270853339, + "grad_norm": 0.597176730632782, + "learning_rate": 2.6901765234136428e-06, + "loss": 0.4043, + "num_input_tokens_seen": 51800240, + "step": 34040 + }, + { + "epoch": 1.0897189680558224, + "grad_norm": 1.4157623052597046, + "learning_rate": 2.685748017172063e-06, + "loss": 0.6052, + "num_input_tokens_seen": 51807376, + "step": 34045 + }, + { + "epoch": 1.0898790090263106, + "grad_norm": 0.8179097175598145, + "learning_rate": 2.681322952068549e-06, + "loss": 0.4904, + "num_input_tokens_seen": 51815328, + "step": 34050 + }, + { + "epoch": 1.0900390499967991, + "grad_norm": 1.152701497077942, + "learning_rate": 2.6769013287855137e-06, + "loss": 0.4742, + "num_input_tokens_seen": 51822656, + "step": 34055 + }, + { + "epoch": 1.0901990909672876, + "grad_norm": 0.6716275215148926, + "learning_rate": 2.6724831480048286e-06, + "loss": 0.5035, + "num_input_tokens_seen": 51830320, + "step": 34060 + }, + { + "epoch": 1.0903591319377761, + "grad_norm": 1.290209412574768, + "learning_rate": 2.66806841040782e-06, + "loss": 0.4014, + "num_input_tokens_seen": 51838000, + "step": 34065 + }, + { + "epoch": 1.0905191729082646, + "grad_norm": 1.2815250158309937, + "learning_rate": 2.6636571166753083e-06, + "loss": 0.5044, + "num_input_tokens_seen": 51845760, + "step": 34070 + }, + { + "epoch": 1.0906792138787529, + "grad_norm": 1.0024446249008179, + "learning_rate": 2.6592492674875598e-06, + "loss": 0.5959, + "num_input_tokens_seen": 51853088, + "step": 34075 + }, + { + "epoch": 1.0908392548492414, + "grad_norm": 0.7665782570838928, + "learning_rate": 2.6548448635243305e-06, + "loss": 0.5789, + "num_input_tokens_seen": 51860640, + "step": 34080 + }, + { + "epoch": 1.0909992958197299, + "grad_norm": 1.0828230381011963, + "learning_rate": 2.650443905464828e-06, + "loss": 0.5875, + "num_input_tokens_seen": 51868096, + "step": 34085 + }, + { + "epoch": 1.0911593367902184, + "grad_norm": 0.7853445410728455, + "learning_rate": 2.646046393987739e-06, + "loss": 0.4634, + "num_input_tokens_seen": 51875200, + "step": 34090 + }, + { + "epoch": 1.0913193777607066, + "grad_norm": 0.4536905586719513, + "learning_rate": 2.64165232977121e-06, + "loss": 0.3774, + "num_input_tokens_seen": 51882528, + "step": 34095 + }, + { + "epoch": 1.0914794187311951, + "grad_norm": 0.5173072218894958, + "learning_rate": 2.6372617134928695e-06, + "loss": 0.3347, + "num_input_tokens_seen": 51889664, + "step": 34100 + }, + { + "epoch": 1.0916394597016836, + "grad_norm": 0.8506636619567871, + "learning_rate": 2.6328745458297943e-06, + "loss": 0.5357, + "num_input_tokens_seen": 51897104, + "step": 34105 + }, + { + "epoch": 1.0917995006721721, + "grad_norm": 0.42827045917510986, + "learning_rate": 2.6284908274585546e-06, + "loss": 0.3388, + "num_input_tokens_seen": 51904736, + "step": 34110 + }, + { + "epoch": 1.0919595416426606, + "grad_norm": 0.5981210470199585, + "learning_rate": 2.6241105590551595e-06, + "loss": 0.3609, + "num_input_tokens_seen": 51912272, + "step": 34115 + }, + { + "epoch": 1.092119582613149, + "grad_norm": 0.931155264377594, + "learning_rate": 2.6197337412951105e-06, + "loss": 0.4129, + "num_input_tokens_seen": 51919904, + "step": 34120 + }, + { + "epoch": 1.0922796235836374, + "grad_norm": 0.5943617820739746, + "learning_rate": 2.6153603748533705e-06, + "loss": 0.5686, + "num_input_tokens_seen": 51927392, + "step": 34125 + }, + { + "epoch": 1.0924396645541259, + "grad_norm": 0.6476207375526428, + "learning_rate": 2.6109904604043585e-06, + "loss": 0.5449, + "num_input_tokens_seen": 51935024, + "step": 34130 + }, + { + "epoch": 1.0925997055246144, + "grad_norm": 0.46377238631248474, + "learning_rate": 2.6066239986219765e-06, + "loss": 0.4352, + "num_input_tokens_seen": 51942224, + "step": 34135 + }, + { + "epoch": 1.0927597464951027, + "grad_norm": 0.7012153267860413, + "learning_rate": 2.602260990179592e-06, + "loss": 0.4724, + "num_input_tokens_seen": 51950208, + "step": 34140 + }, + { + "epoch": 1.0929197874655912, + "grad_norm": 0.6519737243652344, + "learning_rate": 2.5979014357500248e-06, + "loss": 0.5666, + "num_input_tokens_seen": 51957456, + "step": 34145 + }, + { + "epoch": 1.0930798284360796, + "grad_norm": 0.7541979551315308, + "learning_rate": 2.5935453360055844e-06, + "loss": 0.5534, + "num_input_tokens_seen": 51965168, + "step": 34150 + }, + { + "epoch": 1.0932398694065681, + "grad_norm": 0.607279360294342, + "learning_rate": 2.5891926916180283e-06, + "loss": 0.6232, + "num_input_tokens_seen": 51972304, + "step": 34155 + }, + { + "epoch": 1.0933999103770566, + "grad_norm": 0.6351111531257629, + "learning_rate": 2.5848435032585883e-06, + "loss": 0.4297, + "num_input_tokens_seen": 51980432, + "step": 34160 + }, + { + "epoch": 1.093559951347545, + "grad_norm": 0.6909335255622864, + "learning_rate": 2.58049777159797e-06, + "loss": 0.4191, + "num_input_tokens_seen": 51987840, + "step": 34165 + }, + { + "epoch": 1.0937199923180334, + "grad_norm": 0.37534859776496887, + "learning_rate": 2.576155497306332e-06, + "loss": 0.4709, + "num_input_tokens_seen": 51996192, + "step": 34170 + }, + { + "epoch": 1.093880033288522, + "grad_norm": 0.5172519087791443, + "learning_rate": 2.57181668105331e-06, + "loss": 0.3335, + "num_input_tokens_seen": 52004352, + "step": 34175 + }, + { + "epoch": 1.0940400742590104, + "grad_norm": 1.2970881462097168, + "learning_rate": 2.567481323508014e-06, + "loss": 0.4894, + "num_input_tokens_seen": 52011872, + "step": 34180 + }, + { + "epoch": 1.0942001152294987, + "grad_norm": 0.4131138324737549, + "learning_rate": 2.5631494253389954e-06, + "loss": 0.5403, + "num_input_tokens_seen": 52019328, + "step": 34185 + }, + { + "epoch": 1.0943601561999872, + "grad_norm": 1.6543763875961304, + "learning_rate": 2.5588209872142997e-06, + "loss": 0.5754, + "num_input_tokens_seen": 52027360, + "step": 34190 + }, + { + "epoch": 1.0945201971704757, + "grad_norm": 1.2166694402694702, + "learning_rate": 2.5544960098014186e-06, + "loss": 0.5324, + "num_input_tokens_seen": 52034832, + "step": 34195 + }, + { + "epoch": 1.0946802381409642, + "grad_norm": 0.6565656661987305, + "learning_rate": 2.550174493767318e-06, + "loss": 0.4113, + "num_input_tokens_seen": 52042176, + "step": 34200 + }, + { + "epoch": 1.0946802381409642, + "eval_loss": 0.4946494698524475, + "eval_runtime": 557.2213, + "eval_samples_per_second": 24.92, + "eval_steps_per_second": 12.46, + "num_input_tokens_seen": 52042176, + "step": 34200 + }, + { + "epoch": 1.0948402791114524, + "grad_norm": 1.209934115409851, + "learning_rate": 2.545856439778438e-06, + "loss": 0.5706, + "num_input_tokens_seen": 52049584, + "step": 34205 + }, + { + "epoch": 1.095000320081941, + "grad_norm": 0.41547462344169617, + "learning_rate": 2.541541848500667e-06, + "loss": 0.4126, + "num_input_tokens_seen": 52057168, + "step": 34210 + }, + { + "epoch": 1.0951603610524294, + "grad_norm": 1.5796293020248413, + "learning_rate": 2.5372307205993733e-06, + "loss": 0.4175, + "num_input_tokens_seen": 52064960, + "step": 34215 + }, + { + "epoch": 1.095320402022918, + "grad_norm": 0.6050800085067749, + "learning_rate": 2.5329230567393917e-06, + "loss": 0.3538, + "num_input_tokens_seen": 52072848, + "step": 34220 + }, + { + "epoch": 1.0954804429934064, + "grad_norm": 0.3730703592300415, + "learning_rate": 2.5286188575850164e-06, + "loss": 0.3418, + "num_input_tokens_seen": 52080608, + "step": 34225 + }, + { + "epoch": 1.0956404839638947, + "grad_norm": 0.5190412402153015, + "learning_rate": 2.5243181237999984e-06, + "loss": 0.5319, + "num_input_tokens_seen": 52088368, + "step": 34230 + }, + { + "epoch": 1.0958005249343832, + "grad_norm": 0.7215520739555359, + "learning_rate": 2.520020856047578e-06, + "loss": 0.4752, + "num_input_tokens_seen": 52096256, + "step": 34235 + }, + { + "epoch": 1.0959605659048717, + "grad_norm": 0.7388562560081482, + "learning_rate": 2.515727054990438e-06, + "loss": 0.4744, + "num_input_tokens_seen": 52103856, + "step": 34240 + }, + { + "epoch": 1.0961206068753602, + "grad_norm": 0.5623770356178284, + "learning_rate": 2.511436721290747e-06, + "loss": 0.3674, + "num_input_tokens_seen": 52112112, + "step": 34245 + }, + { + "epoch": 1.0962806478458484, + "grad_norm": 1.0663747787475586, + "learning_rate": 2.5071498556101164e-06, + "loss": 0.6012, + "num_input_tokens_seen": 52119776, + "step": 34250 + }, + { + "epoch": 1.096440688816337, + "grad_norm": 0.6193541884422302, + "learning_rate": 2.5028664586096485e-06, + "loss": 0.3577, + "num_input_tokens_seen": 52127200, + "step": 34255 + }, + { + "epoch": 1.0966007297868254, + "grad_norm": 0.6140197515487671, + "learning_rate": 2.498586530949881e-06, + "loss": 0.4486, + "num_input_tokens_seen": 52134928, + "step": 34260 + }, + { + "epoch": 1.096760770757314, + "grad_norm": 0.8198130130767822, + "learning_rate": 2.4943100732908427e-06, + "loss": 0.6844, + "num_input_tokens_seen": 52142544, + "step": 34265 + }, + { + "epoch": 1.0969208117278022, + "grad_norm": 0.8344619870185852, + "learning_rate": 2.4900370862920188e-06, + "loss": 0.4125, + "num_input_tokens_seen": 52150512, + "step": 34270 + }, + { + "epoch": 1.0970808526982907, + "grad_norm": 1.2262541055679321, + "learning_rate": 2.4857675706123518e-06, + "loss": 0.5147, + "num_input_tokens_seen": 52158144, + "step": 34275 + }, + { + "epoch": 1.0972408936687792, + "grad_norm": 0.536851167678833, + "learning_rate": 2.4815015269102543e-06, + "loss": 0.4071, + "num_input_tokens_seen": 52165824, + "step": 34280 + }, + { + "epoch": 1.0974009346392677, + "grad_norm": 0.6220750212669373, + "learning_rate": 2.477238955843611e-06, + "loss": 0.5273, + "num_input_tokens_seen": 52173728, + "step": 34285 + }, + { + "epoch": 1.0975609756097562, + "grad_norm": 0.7847750186920166, + "learning_rate": 2.4729798580697573e-06, + "loss": 0.3844, + "num_input_tokens_seen": 52181248, + "step": 34290 + }, + { + "epoch": 1.0977210165802445, + "grad_norm": 0.7294479608535767, + "learning_rate": 2.4687242342455034e-06, + "loss": 0.6366, + "num_input_tokens_seen": 52188976, + "step": 34295 + }, + { + "epoch": 1.097881057550733, + "grad_norm": 0.31180158257484436, + "learning_rate": 2.4644720850271196e-06, + "loss": 0.4573, + "num_input_tokens_seen": 52196496, + "step": 34300 + }, + { + "epoch": 1.0980410985212214, + "grad_norm": 0.6115036010742188, + "learning_rate": 2.4602234110703364e-06, + "loss": 0.4017, + "num_input_tokens_seen": 52203872, + "step": 34305 + }, + { + "epoch": 1.09820113949171, + "grad_norm": 0.6673750281333923, + "learning_rate": 2.4559782130303576e-06, + "loss": 0.3465, + "num_input_tokens_seen": 52211744, + "step": 34310 + }, + { + "epoch": 1.0983611804621982, + "grad_norm": 0.5658954381942749, + "learning_rate": 2.451736491561843e-06, + "loss": 0.5719, + "num_input_tokens_seen": 52219328, + "step": 34315 + }, + { + "epoch": 1.0985212214326867, + "grad_norm": 0.9170293807983398, + "learning_rate": 2.4474982473189163e-06, + "loss": 0.5268, + "num_input_tokens_seen": 52226608, + "step": 34320 + }, + { + "epoch": 1.0986812624031752, + "grad_norm": 1.5343844890594482, + "learning_rate": 2.4432634809551796e-06, + "loss": 0.5266, + "num_input_tokens_seen": 52234192, + "step": 34325 + }, + { + "epoch": 1.0988413033736637, + "grad_norm": 0.6876124143600464, + "learning_rate": 2.439032193123675e-06, + "loss": 0.4825, + "num_input_tokens_seen": 52241984, + "step": 34330 + }, + { + "epoch": 1.0990013443441522, + "grad_norm": 0.5662658214569092, + "learning_rate": 2.4348043844769297e-06, + "loss": 0.4047, + "num_input_tokens_seen": 52249824, + "step": 34335 + }, + { + "epoch": 1.0991613853146405, + "grad_norm": 0.6473657488822937, + "learning_rate": 2.4305800556669146e-06, + "loss": 0.5082, + "num_input_tokens_seen": 52257440, + "step": 34340 + }, + { + "epoch": 1.099321426285129, + "grad_norm": 0.9496975541114807, + "learning_rate": 2.426359207345083e-06, + "loss": 0.4753, + "num_input_tokens_seen": 52265040, + "step": 34345 + }, + { + "epoch": 1.0994814672556175, + "grad_norm": 0.7142330408096313, + "learning_rate": 2.4221418401623396e-06, + "loss": 0.5091, + "num_input_tokens_seen": 52273408, + "step": 34350 + }, + { + "epoch": 1.099641508226106, + "grad_norm": 0.905366837978363, + "learning_rate": 2.4179279547690557e-06, + "loss": 0.4717, + "num_input_tokens_seen": 52280976, + "step": 34355 + }, + { + "epoch": 1.0998015491965942, + "grad_norm": 0.7778168320655823, + "learning_rate": 2.413717551815062e-06, + "loss": 0.3817, + "num_input_tokens_seen": 52288336, + "step": 34360 + }, + { + "epoch": 1.0999615901670827, + "grad_norm": 0.4427150785923004, + "learning_rate": 2.409510631949666e-06, + "loss": 0.3244, + "num_input_tokens_seen": 52295568, + "step": 34365 + }, + { + "epoch": 1.1001216311375712, + "grad_norm": 0.6493459939956665, + "learning_rate": 2.405307195821618e-06, + "loss": 0.4836, + "num_input_tokens_seen": 52303040, + "step": 34370 + }, + { + "epoch": 1.1002816721080597, + "grad_norm": 0.6053269505500793, + "learning_rate": 2.4011072440791372e-06, + "loss": 0.5053, + "num_input_tokens_seen": 52310624, + "step": 34375 + }, + { + "epoch": 1.1004417130785482, + "grad_norm": 0.6085485816001892, + "learning_rate": 2.3969107773699233e-06, + "loss": 0.4519, + "num_input_tokens_seen": 52318288, + "step": 34380 + }, + { + "epoch": 1.1006017540490365, + "grad_norm": 0.31859922409057617, + "learning_rate": 2.3927177963411096e-06, + "loss": 0.3656, + "num_input_tokens_seen": 52325920, + "step": 34385 + }, + { + "epoch": 1.100761795019525, + "grad_norm": 0.9016819000244141, + "learning_rate": 2.3885283016393144e-06, + "loss": 0.3667, + "num_input_tokens_seen": 52333936, + "step": 34390 + }, + { + "epoch": 1.1009218359900135, + "grad_norm": 0.7280978560447693, + "learning_rate": 2.3843422939106076e-06, + "loss": 0.4198, + "num_input_tokens_seen": 52341504, + "step": 34395 + }, + { + "epoch": 1.101081876960502, + "grad_norm": 1.0110737085342407, + "learning_rate": 2.380159773800525e-06, + "loss": 0.5206, + "num_input_tokens_seen": 52348928, + "step": 34400 + }, + { + "epoch": 1.101081876960502, + "eval_loss": 0.4946358799934387, + "eval_runtime": 557.1539, + "eval_samples_per_second": 24.923, + "eval_steps_per_second": 12.462, + "num_input_tokens_seen": 52348928, + "step": 34400 + }, + { + "epoch": 1.1012419179309902, + "grad_norm": 0.5048511028289795, + "learning_rate": 2.3759807419540675e-06, + "loss": 0.3457, + "num_input_tokens_seen": 52356560, + "step": 34405 + }, + { + "epoch": 1.1014019589014787, + "grad_norm": 0.6163343787193298, + "learning_rate": 2.3718051990156835e-06, + "loss": 0.495, + "num_input_tokens_seen": 52363936, + "step": 34410 + }, + { + "epoch": 1.1015619998719672, + "grad_norm": 0.4035988748073578, + "learning_rate": 2.367633145629311e-06, + "loss": 0.3994, + "num_input_tokens_seen": 52371472, + "step": 34415 + }, + { + "epoch": 1.1017220408424557, + "grad_norm": 0.7080606818199158, + "learning_rate": 2.363464582438316e-06, + "loss": 0.6091, + "num_input_tokens_seen": 52378848, + "step": 34420 + }, + { + "epoch": 1.1018820818129442, + "grad_norm": 0.8947857618331909, + "learning_rate": 2.3592995100855526e-06, + "loss": 0.449, + "num_input_tokens_seen": 52386672, + "step": 34425 + }, + { + "epoch": 1.1020421227834325, + "grad_norm": 0.716458261013031, + "learning_rate": 2.3551379292133273e-06, + "loss": 0.4643, + "num_input_tokens_seen": 52393920, + "step": 34430 + }, + { + "epoch": 1.102202163753921, + "grad_norm": 0.9607234597206116, + "learning_rate": 2.3509798404634047e-06, + "loss": 0.3869, + "num_input_tokens_seen": 52401152, + "step": 34435 + }, + { + "epoch": 1.1023622047244095, + "grad_norm": 0.6158754229545593, + "learning_rate": 2.346825244477019e-06, + "loss": 0.406, + "num_input_tokens_seen": 52408608, + "step": 34440 + }, + { + "epoch": 1.102522245694898, + "grad_norm": 0.7664769291877747, + "learning_rate": 2.3426741418948545e-06, + "loss": 0.397, + "num_input_tokens_seen": 52416512, + "step": 34445 + }, + { + "epoch": 1.1026822866653863, + "grad_norm": 0.6008094549179077, + "learning_rate": 2.3385265333570715e-06, + "loss": 0.6786, + "num_input_tokens_seen": 52423808, + "step": 34450 + }, + { + "epoch": 1.1028423276358748, + "grad_norm": 0.5238674879074097, + "learning_rate": 2.334382419503278e-06, + "loss": 0.3628, + "num_input_tokens_seen": 52431248, + "step": 34455 + }, + { + "epoch": 1.1030023686063632, + "grad_norm": 1.6632304191589355, + "learning_rate": 2.3302418009725465e-06, + "loss": 0.4755, + "num_input_tokens_seen": 52439008, + "step": 34460 + }, + { + "epoch": 1.1031624095768517, + "grad_norm": 0.3834436535835266, + "learning_rate": 2.326104678403415e-06, + "loss": 0.4468, + "num_input_tokens_seen": 52447184, + "step": 34465 + }, + { + "epoch": 1.10332245054734, + "grad_norm": 0.48746851086616516, + "learning_rate": 2.321971052433883e-06, + "loss": 0.4291, + "num_input_tokens_seen": 52455088, + "step": 34470 + }, + { + "epoch": 1.1034824915178285, + "grad_norm": 0.6040600538253784, + "learning_rate": 2.3178409237014004e-06, + "loss": 0.7185, + "num_input_tokens_seen": 52462432, + "step": 34475 + }, + { + "epoch": 1.103642532488317, + "grad_norm": 0.7619737386703491, + "learning_rate": 2.313714292842889e-06, + "loss": 0.547, + "num_input_tokens_seen": 52470176, + "step": 34480 + }, + { + "epoch": 1.1038025734588055, + "grad_norm": 0.6912159323692322, + "learning_rate": 2.309591160494734e-06, + "loss": 0.5109, + "num_input_tokens_seen": 52477232, + "step": 34485 + }, + { + "epoch": 1.103962614429294, + "grad_norm": 0.7884219288825989, + "learning_rate": 2.305471527292763e-06, + "loss": 0.4874, + "num_input_tokens_seen": 52485104, + "step": 34490 + }, + { + "epoch": 1.1041226553997823, + "grad_norm": 0.4723868668079376, + "learning_rate": 2.3013553938722817e-06, + "loss": 0.4761, + "num_input_tokens_seen": 52492240, + "step": 34495 + }, + { + "epoch": 1.1042826963702708, + "grad_norm": 1.147530436515808, + "learning_rate": 2.297242760868043e-06, + "loss": 0.4644, + "num_input_tokens_seen": 52500064, + "step": 34500 + }, + { + "epoch": 1.1044427373407593, + "grad_norm": 0.5037840008735657, + "learning_rate": 2.2931336289142735e-06, + "loss": 0.4094, + "num_input_tokens_seen": 52507440, + "step": 34505 + }, + { + "epoch": 1.1046027783112478, + "grad_norm": 0.7224535942077637, + "learning_rate": 2.289027998644655e-06, + "loss": 0.4994, + "num_input_tokens_seen": 52515200, + "step": 34510 + }, + { + "epoch": 1.104762819281736, + "grad_norm": 0.8291861414909363, + "learning_rate": 2.2849258706923228e-06, + "loss": 0.4414, + "num_input_tokens_seen": 52523280, + "step": 34515 + }, + { + "epoch": 1.1049228602522245, + "grad_norm": 0.5341522097587585, + "learning_rate": 2.2808272456898705e-06, + "loss": 0.4733, + "num_input_tokens_seen": 52531184, + "step": 34520 + }, + { + "epoch": 1.105082901222713, + "grad_norm": 0.6387841105461121, + "learning_rate": 2.2767321242693707e-06, + "loss": 0.4364, + "num_input_tokens_seen": 52538192, + "step": 34525 + }, + { + "epoch": 1.1052429421932015, + "grad_norm": 0.6293898224830627, + "learning_rate": 2.272640507062329e-06, + "loss": 0.4309, + "num_input_tokens_seen": 52546176, + "step": 34530 + }, + { + "epoch": 1.1054029831636898, + "grad_norm": 1.0253115892410278, + "learning_rate": 2.2685523946997382e-06, + "loss": 0.4742, + "num_input_tokens_seen": 52553616, + "step": 34535 + }, + { + "epoch": 1.1055630241341783, + "grad_norm": 0.7496463656425476, + "learning_rate": 2.2644677878120245e-06, + "loss": 0.7594, + "num_input_tokens_seen": 52561408, + "step": 34540 + }, + { + "epoch": 1.1057230651046668, + "grad_norm": 0.6470986008644104, + "learning_rate": 2.2603866870290897e-06, + "loss": 0.4528, + "num_input_tokens_seen": 52568656, + "step": 34545 + }, + { + "epoch": 1.1058831060751553, + "grad_norm": 0.9406973719596863, + "learning_rate": 2.256309092980294e-06, + "loss": 0.4187, + "num_input_tokens_seen": 52576368, + "step": 34550 + }, + { + "epoch": 1.1060431470456438, + "grad_norm": 0.9979057312011719, + "learning_rate": 2.252235006294448e-06, + "loss": 0.3903, + "num_input_tokens_seen": 52583344, + "step": 34555 + }, + { + "epoch": 1.106203188016132, + "grad_norm": 0.7423200607299805, + "learning_rate": 2.2481644275998333e-06, + "loss": 0.3515, + "num_input_tokens_seen": 52591472, + "step": 34560 + }, + { + "epoch": 1.1063632289866205, + "grad_norm": 0.938506007194519, + "learning_rate": 2.2440973575241832e-06, + "loss": 0.4366, + "num_input_tokens_seen": 52599088, + "step": 34565 + }, + { + "epoch": 1.106523269957109, + "grad_norm": 0.7116999626159668, + "learning_rate": 2.240033796694685e-06, + "loss": 0.3952, + "num_input_tokens_seen": 52606304, + "step": 34570 + }, + { + "epoch": 1.1066833109275975, + "grad_norm": 1.1026928424835205, + "learning_rate": 2.235973745737999e-06, + "loss": 0.7485, + "num_input_tokens_seen": 52613680, + "step": 34575 + }, + { + "epoch": 1.1068433518980858, + "grad_norm": 0.6411988735198975, + "learning_rate": 2.2319172052802263e-06, + "loss": 0.4365, + "num_input_tokens_seen": 52621152, + "step": 34580 + }, + { + "epoch": 1.1070033928685743, + "grad_norm": 0.5776847004890442, + "learning_rate": 2.2278641759469477e-06, + "loss": 0.4672, + "num_input_tokens_seen": 52628912, + "step": 34585 + }, + { + "epoch": 1.1071634338390628, + "grad_norm": 1.1016134023666382, + "learning_rate": 2.2238146583631825e-06, + "loss": 0.3434, + "num_input_tokens_seen": 52636512, + "step": 34590 + }, + { + "epoch": 1.1073234748095513, + "grad_norm": 0.8717352151870728, + "learning_rate": 2.2197686531534256e-06, + "loss": 0.5635, + "num_input_tokens_seen": 52644016, + "step": 34595 + }, + { + "epoch": 1.1074835157800398, + "grad_norm": 1.0777885913848877, + "learning_rate": 2.2157261609416087e-06, + "loss": 0.5279, + "num_input_tokens_seen": 52651440, + "step": 34600 + }, + { + "epoch": 1.1074835157800398, + "eval_loss": 0.4945237934589386, + "eval_runtime": 556.9753, + "eval_samples_per_second": 24.931, + "eval_steps_per_second": 12.466, + "num_input_tokens_seen": 52651440, + "step": 34600 + }, + { + "epoch": 1.107643556750528, + "grad_norm": 0.4554215371608734, + "learning_rate": 2.211687182351149e-06, + "loss": 0.3525, + "num_input_tokens_seen": 52658960, + "step": 34605 + }, + { + "epoch": 1.1078035977210166, + "grad_norm": 1.1230967044830322, + "learning_rate": 2.2076517180048993e-06, + "loss": 0.5454, + "num_input_tokens_seen": 52666896, + "step": 34610 + }, + { + "epoch": 1.107963638691505, + "grad_norm": 0.9059692025184631, + "learning_rate": 2.2036197685251834e-06, + "loss": 0.4177, + "num_input_tokens_seen": 52674160, + "step": 34615 + }, + { + "epoch": 1.1081236796619935, + "grad_norm": 0.6393691301345825, + "learning_rate": 2.199591334533771e-06, + "loss": 0.4581, + "num_input_tokens_seen": 52681664, + "step": 34620 + }, + { + "epoch": 1.1082837206324818, + "grad_norm": 0.6993301510810852, + "learning_rate": 2.1955664166519036e-06, + "loss": 0.4656, + "num_input_tokens_seen": 52689568, + "step": 34625 + }, + { + "epoch": 1.1084437616029703, + "grad_norm": 0.8454827070236206, + "learning_rate": 2.1915450155002793e-06, + "loss": 0.5878, + "num_input_tokens_seen": 52696992, + "step": 34630 + }, + { + "epoch": 1.1086038025734588, + "grad_norm": 0.4071142077445984, + "learning_rate": 2.187527131699038e-06, + "loss": 0.3984, + "num_input_tokens_seen": 52704704, + "step": 34635 + }, + { + "epoch": 1.1087638435439473, + "grad_norm": 0.8138797283172607, + "learning_rate": 2.18351276586779e-06, + "loss": 0.4186, + "num_input_tokens_seen": 52712640, + "step": 34640 + }, + { + "epoch": 1.1089238845144358, + "grad_norm": 0.5421749353408813, + "learning_rate": 2.1795019186256092e-06, + "loss": 0.5081, + "num_input_tokens_seen": 52720384, + "step": 34645 + }, + { + "epoch": 1.109083925484924, + "grad_norm": 0.667878270149231, + "learning_rate": 2.1754945905910094e-06, + "loss": 0.4399, + "num_input_tokens_seen": 52728288, + "step": 34650 + }, + { + "epoch": 1.1092439664554126, + "grad_norm": 1.0467801094055176, + "learning_rate": 2.171490782381977e-06, + "loss": 0.4179, + "num_input_tokens_seen": 52736288, + "step": 34655 + }, + { + "epoch": 1.109404007425901, + "grad_norm": 2.3037071228027344, + "learning_rate": 2.1674904946159425e-06, + "loss": 0.4956, + "num_input_tokens_seen": 52744304, + "step": 34660 + }, + { + "epoch": 1.1095640483963896, + "grad_norm": 0.7282769083976746, + "learning_rate": 2.16349372790981e-06, + "loss": 0.2781, + "num_input_tokens_seen": 52752096, + "step": 34665 + }, + { + "epoch": 1.1097240893668778, + "grad_norm": 1.2020930051803589, + "learning_rate": 2.159500482879928e-06, + "loss": 0.4658, + "num_input_tokens_seen": 52760384, + "step": 34670 + }, + { + "epoch": 1.1098841303373663, + "grad_norm": 0.9317497611045837, + "learning_rate": 2.155510760142096e-06, + "loss": 0.4405, + "num_input_tokens_seen": 52768160, + "step": 34675 + }, + { + "epoch": 1.1100441713078548, + "grad_norm": 1.340794563293457, + "learning_rate": 2.151524560311588e-06, + "loss": 0.6134, + "num_input_tokens_seen": 52776288, + "step": 34680 + }, + { + "epoch": 1.1102042122783433, + "grad_norm": 0.5907966494560242, + "learning_rate": 2.147541884003129e-06, + "loss": 0.5575, + "num_input_tokens_seen": 52784016, + "step": 34685 + }, + { + "epoch": 1.1103642532488318, + "grad_norm": 0.7026211619377136, + "learning_rate": 2.1435627318308895e-06, + "loss": 0.5156, + "num_input_tokens_seen": 52791792, + "step": 34690 + }, + { + "epoch": 1.11052429421932, + "grad_norm": 0.5090404748916626, + "learning_rate": 2.139587104408511e-06, + "loss": 0.519, + "num_input_tokens_seen": 52799600, + "step": 34695 + }, + { + "epoch": 1.1106843351898086, + "grad_norm": 0.5309507846832275, + "learning_rate": 2.1356150023490783e-06, + "loss": 0.4202, + "num_input_tokens_seen": 52807728, + "step": 34700 + }, + { + "epoch": 1.110844376160297, + "grad_norm": 1.4897181987762451, + "learning_rate": 2.1316464262651464e-06, + "loss": 0.5238, + "num_input_tokens_seen": 52815184, + "step": 34705 + }, + { + "epoch": 1.1110044171307856, + "grad_norm": 1.0173274278640747, + "learning_rate": 2.1276813767687224e-06, + "loss": 0.4822, + "num_input_tokens_seen": 52822480, + "step": 34710 + }, + { + "epoch": 1.1111644581012738, + "grad_norm": 0.7122475504875183, + "learning_rate": 2.123719854471254e-06, + "loss": 0.4705, + "num_input_tokens_seen": 52829920, + "step": 34715 + }, + { + "epoch": 1.1113244990717623, + "grad_norm": 1.350516438484192, + "learning_rate": 2.119761859983668e-06, + "loss": 0.5276, + "num_input_tokens_seen": 52837424, + "step": 34720 + }, + { + "epoch": 1.1114845400422508, + "grad_norm": 0.736220121383667, + "learning_rate": 2.1158073939163386e-06, + "loss": 0.4612, + "num_input_tokens_seen": 52845376, + "step": 34725 + }, + { + "epoch": 1.1116445810127393, + "grad_norm": 0.5043531060218811, + "learning_rate": 2.111856456879088e-06, + "loss": 0.4602, + "num_input_tokens_seen": 52853024, + "step": 34730 + }, + { + "epoch": 1.1118046219832276, + "grad_norm": 1.2691054344177246, + "learning_rate": 2.1079090494811993e-06, + "loss": 0.4199, + "num_input_tokens_seen": 52860896, + "step": 34735 + }, + { + "epoch": 1.111964662953716, + "grad_norm": 0.9988275766372681, + "learning_rate": 2.103965172331418e-06, + "loss": 0.4502, + "num_input_tokens_seen": 52868384, + "step": 34740 + }, + { + "epoch": 1.1121247039242046, + "grad_norm": 1.3346505165100098, + "learning_rate": 2.100024826037933e-06, + "loss": 0.5348, + "num_input_tokens_seen": 52875728, + "step": 34745 + }, + { + "epoch": 1.112284744894693, + "grad_norm": 0.5999208688735962, + "learning_rate": 2.0960880112084027e-06, + "loss": 0.5754, + "num_input_tokens_seen": 52883424, + "step": 34750 + }, + { + "epoch": 1.1124447858651816, + "grad_norm": 1.258124828338623, + "learning_rate": 2.092154728449927e-06, + "loss": 0.5288, + "num_input_tokens_seen": 52890864, + "step": 34755 + }, + { + "epoch": 1.1126048268356699, + "grad_norm": 0.4018084406852722, + "learning_rate": 2.0882249783690687e-06, + "loss": 0.435, + "num_input_tokens_seen": 52898960, + "step": 34760 + }, + { + "epoch": 1.1127648678061584, + "grad_norm": 0.6900472044944763, + "learning_rate": 2.084298761571851e-06, + "loss": 0.4386, + "num_input_tokens_seen": 52906640, + "step": 34765 + }, + { + "epoch": 1.1129249087766468, + "grad_norm": 0.8260816335678101, + "learning_rate": 2.080376078663737e-06, + "loss": 0.478, + "num_input_tokens_seen": 52914336, + "step": 34770 + }, + { + "epoch": 1.1130849497471353, + "grad_norm": 0.6859257817268372, + "learning_rate": 2.0764569302496593e-06, + "loss": 0.3945, + "num_input_tokens_seen": 52921568, + "step": 34775 + }, + { + "epoch": 1.1132449907176236, + "grad_norm": 0.5763005614280701, + "learning_rate": 2.0725413169339957e-06, + "loss": 0.4239, + "num_input_tokens_seen": 52929344, + "step": 34780 + }, + { + "epoch": 1.1134050316881121, + "grad_norm": 0.46275684237480164, + "learning_rate": 2.068629239320588e-06, + "loss": 0.44, + "num_input_tokens_seen": 52937008, + "step": 34785 + }, + { + "epoch": 1.1135650726586006, + "grad_norm": 0.6083967089653015, + "learning_rate": 2.064720698012726e-06, + "loss": 0.4572, + "num_input_tokens_seen": 52944832, + "step": 34790 + }, + { + "epoch": 1.113725113629089, + "grad_norm": 0.2714424133300781, + "learning_rate": 2.0608156936131522e-06, + "loss": 0.5594, + "num_input_tokens_seen": 52952384, + "step": 34795 + }, + { + "epoch": 1.1138851545995774, + "grad_norm": 1.3387393951416016, + "learning_rate": 2.056914226724074e-06, + "loss": 0.4664, + "num_input_tokens_seen": 52960256, + "step": 34800 + }, + { + "epoch": 1.1138851545995774, + "eval_loss": 0.49441513419151306, + "eval_runtime": 557.0903, + "eval_samples_per_second": 24.926, + "eval_steps_per_second": 12.463, + "num_input_tokens_seen": 52960256, + "step": 34800 + }, + { + "epoch": 1.1140451955700659, + "grad_norm": 1.139603614807129, + "learning_rate": 2.0530162979471385e-06, + "loss": 0.4574, + "num_input_tokens_seen": 52967824, + "step": 34805 + }, + { + "epoch": 1.1142052365405544, + "grad_norm": 0.899115800857544, + "learning_rate": 2.0491219078834667e-06, + "loss": 0.4226, + "num_input_tokens_seen": 52975728, + "step": 34810 + }, + { + "epoch": 1.1143652775110429, + "grad_norm": 0.48991477489471436, + "learning_rate": 2.045231057133612e-06, + "loss": 0.3288, + "num_input_tokens_seen": 52983408, + "step": 34815 + }, + { + "epoch": 1.1145253184815314, + "grad_norm": 0.5148141980171204, + "learning_rate": 2.0413437462975944e-06, + "loss": 0.571, + "num_input_tokens_seen": 52991824, + "step": 34820 + }, + { + "epoch": 1.1146853594520196, + "grad_norm": 0.38641947507858276, + "learning_rate": 2.0374599759748843e-06, + "loss": 0.5151, + "num_input_tokens_seen": 52999120, + "step": 34825 + }, + { + "epoch": 1.1148454004225081, + "grad_norm": 0.4016680419445038, + "learning_rate": 2.033579746764419e-06, + "loss": 0.4788, + "num_input_tokens_seen": 53006880, + "step": 34830 + }, + { + "epoch": 1.1150054413929966, + "grad_norm": 1.792040467262268, + "learning_rate": 2.029703059264565e-06, + "loss": 0.6315, + "num_input_tokens_seen": 53014784, + "step": 34835 + }, + { + "epoch": 1.1151654823634851, + "grad_norm": 0.5857594013214111, + "learning_rate": 2.02582991407316e-06, + "loss": 0.4057, + "num_input_tokens_seen": 53022304, + "step": 34840 + }, + { + "epoch": 1.1153255233339734, + "grad_norm": 0.5957525968551636, + "learning_rate": 2.0219603117874992e-06, + "loss": 0.7009, + "num_input_tokens_seen": 53029888, + "step": 34845 + }, + { + "epoch": 1.1154855643044619, + "grad_norm": 0.596499502658844, + "learning_rate": 2.0180942530043156e-06, + "loss": 0.4113, + "num_input_tokens_seen": 53038160, + "step": 34850 + }, + { + "epoch": 1.1156456052749504, + "grad_norm": 1.217712640762329, + "learning_rate": 2.0142317383198107e-06, + "loss": 0.5328, + "num_input_tokens_seen": 53045840, + "step": 34855 + }, + { + "epoch": 1.1158056462454389, + "grad_norm": 0.7728743553161621, + "learning_rate": 2.0103727683296243e-06, + "loss": 0.5107, + "num_input_tokens_seen": 53054032, + "step": 34860 + }, + { + "epoch": 1.1159656872159274, + "grad_norm": 0.8732672333717346, + "learning_rate": 2.0065173436288636e-06, + "loss": 0.4215, + "num_input_tokens_seen": 53061552, + "step": 34865 + }, + { + "epoch": 1.1161257281864156, + "grad_norm": 0.8096289038658142, + "learning_rate": 2.002665464812087e-06, + "loss": 0.3595, + "num_input_tokens_seen": 53069280, + "step": 34870 + }, + { + "epoch": 1.1162857691569041, + "grad_norm": 0.5345891118049622, + "learning_rate": 1.998817132473291e-06, + "loss": 0.5168, + "num_input_tokens_seen": 53077136, + "step": 34875 + }, + { + "epoch": 1.1164458101273926, + "grad_norm": 1.0406900644302368, + "learning_rate": 1.9949723472059507e-06, + "loss": 0.3944, + "num_input_tokens_seen": 53084528, + "step": 34880 + }, + { + "epoch": 1.1166058510978811, + "grad_norm": 1.418250322341919, + "learning_rate": 1.9911311096029726e-06, + "loss": 0.595, + "num_input_tokens_seen": 53092688, + "step": 34885 + }, + { + "epoch": 1.1167658920683694, + "grad_norm": 1.01664400100708, + "learning_rate": 1.9872934202567224e-06, + "loss": 0.7324, + "num_input_tokens_seen": 53100144, + "step": 34890 + }, + { + "epoch": 1.116925933038858, + "grad_norm": 1.14387845993042, + "learning_rate": 1.9834592797590257e-06, + "loss": 0.4226, + "num_input_tokens_seen": 53107664, + "step": 34895 + }, + { + "epoch": 1.1170859740093464, + "grad_norm": 0.4671729505062103, + "learning_rate": 1.979628688701149e-06, + "loss": 0.4412, + "num_input_tokens_seen": 53115504, + "step": 34900 + }, + { + "epoch": 1.117246014979835, + "grad_norm": 2.7393338680267334, + "learning_rate": 1.9758016476738193e-06, + "loss": 0.5299, + "num_input_tokens_seen": 53122560, + "step": 34905 + }, + { + "epoch": 1.1174060559503234, + "grad_norm": 0.7019123435020447, + "learning_rate": 1.971978157267221e-06, + "loss": 0.515, + "num_input_tokens_seen": 53129936, + "step": 34910 + }, + { + "epoch": 1.1175660969208117, + "grad_norm": 0.4886298179626465, + "learning_rate": 1.968158218070973e-06, + "loss": 0.5105, + "num_input_tokens_seen": 53137792, + "step": 34915 + }, + { + "epoch": 1.1177261378913002, + "grad_norm": 0.6460680961608887, + "learning_rate": 1.9643418306741682e-06, + "loss": 0.5651, + "num_input_tokens_seen": 53145088, + "step": 34920 + }, + { + "epoch": 1.1178861788617886, + "grad_norm": 0.3865002393722534, + "learning_rate": 1.9605289956653337e-06, + "loss": 0.4745, + "num_input_tokens_seen": 53153280, + "step": 34925 + }, + { + "epoch": 1.1180462198322771, + "grad_norm": 0.8579115867614746, + "learning_rate": 1.9567197136324626e-06, + "loss": 0.3774, + "num_input_tokens_seen": 53160832, + "step": 34930 + }, + { + "epoch": 1.1182062608027654, + "grad_norm": 0.8611398935317993, + "learning_rate": 1.9529139851629935e-06, + "loss": 0.513, + "num_input_tokens_seen": 53168320, + "step": 34935 + }, + { + "epoch": 1.118366301773254, + "grad_norm": 0.7969039678573608, + "learning_rate": 1.949111810843812e-06, + "loss": 0.5671, + "num_input_tokens_seen": 53175392, + "step": 34940 + }, + { + "epoch": 1.1185263427437424, + "grad_norm": 0.738106369972229, + "learning_rate": 1.9453131912612694e-06, + "loss": 0.4423, + "num_input_tokens_seen": 53183040, + "step": 34945 + }, + { + "epoch": 1.118686383714231, + "grad_norm": 0.9231244921684265, + "learning_rate": 1.941518127001149e-06, + "loss": 0.4506, + "num_input_tokens_seen": 53190480, + "step": 34950 + }, + { + "epoch": 1.1188464246847194, + "grad_norm": 1.441756010055542, + "learning_rate": 1.9377266186487107e-06, + "loss": 0.6264, + "num_input_tokens_seen": 53197968, + "step": 34955 + }, + { + "epoch": 1.1190064656552077, + "grad_norm": 0.7538212537765503, + "learning_rate": 1.9339386667886483e-06, + "loss": 0.3901, + "num_input_tokens_seen": 53205424, + "step": 34960 + }, + { + "epoch": 1.1191665066256962, + "grad_norm": 0.904626190662384, + "learning_rate": 1.9301542720051024e-06, + "loss": 0.4343, + "num_input_tokens_seen": 53212560, + "step": 34965 + }, + { + "epoch": 1.1193265475961847, + "grad_norm": 0.7974050045013428, + "learning_rate": 1.926373434881684e-06, + "loss": 0.5697, + "num_input_tokens_seen": 53219968, + "step": 34970 + }, + { + "epoch": 1.1194865885666732, + "grad_norm": 0.546674907207489, + "learning_rate": 1.9225961560014468e-06, + "loss": 0.4949, + "num_input_tokens_seen": 53227744, + "step": 34975 + }, + { + "epoch": 1.1196466295371614, + "grad_norm": 0.6658270955085754, + "learning_rate": 1.918822435946885e-06, + "loss": 0.5424, + "num_input_tokens_seen": 53235136, + "step": 34980 + }, + { + "epoch": 1.11980667050765, + "grad_norm": 2.4944207668304443, + "learning_rate": 1.915052275299961e-06, + "loss": 0.4276, + "num_input_tokens_seen": 53242880, + "step": 34985 + }, + { + "epoch": 1.1199667114781384, + "grad_norm": 0.6302183866500854, + "learning_rate": 1.9112856746420854e-06, + "loss": 0.5378, + "num_input_tokens_seen": 53251088, + "step": 34990 + }, + { + "epoch": 1.120126752448627, + "grad_norm": 0.38534271717071533, + "learning_rate": 1.907522634554104e-06, + "loss": 0.3276, + "num_input_tokens_seen": 53258640, + "step": 34995 + }, + { + "epoch": 1.1202867934191152, + "grad_norm": 1.656530737876892, + "learning_rate": 1.9037631556163337e-06, + "loss": 0.4789, + "num_input_tokens_seen": 53265840, + "step": 35000 + }, + { + "epoch": 1.1202867934191152, + "eval_loss": 0.49451783299446106, + "eval_runtime": 557.4288, + "eval_samples_per_second": 24.911, + "eval_steps_per_second": 12.455, + "num_input_tokens_seen": 53265840, + "step": 35000 + }, + { + "epoch": 1.1204468343896037, + "grad_norm": 0.828628659248352, + "learning_rate": 1.9000072384085272e-06, + "loss": 0.5806, + "num_input_tokens_seen": 53274336, + "step": 35005 + }, + { + "epoch": 1.1206068753600922, + "grad_norm": 0.640177845954895, + "learning_rate": 1.8962548835098987e-06, + "loss": 0.5562, + "num_input_tokens_seen": 53281808, + "step": 35010 + }, + { + "epoch": 1.1207669163305807, + "grad_norm": 1.1724927425384521, + "learning_rate": 1.8925060914991077e-06, + "loss": 0.6548, + "num_input_tokens_seen": 53289376, + "step": 35015 + }, + { + "epoch": 1.1209269573010692, + "grad_norm": 0.6813307404518127, + "learning_rate": 1.888760862954264e-06, + "loss": 0.4822, + "num_input_tokens_seen": 53296736, + "step": 35020 + }, + { + "epoch": 1.1210869982715574, + "grad_norm": 0.3286261260509491, + "learning_rate": 1.8850191984529309e-06, + "loss": 0.3461, + "num_input_tokens_seen": 53304592, + "step": 35025 + }, + { + "epoch": 1.121247039242046, + "grad_norm": 0.9406117796897888, + "learning_rate": 1.8812810985721186e-06, + "loss": 0.556, + "num_input_tokens_seen": 53312656, + "step": 35030 + }, + { + "epoch": 1.1214070802125344, + "grad_norm": 0.5893839597702026, + "learning_rate": 1.8775465638882856e-06, + "loss": 0.393, + "num_input_tokens_seen": 53320080, + "step": 35035 + }, + { + "epoch": 1.121567121183023, + "grad_norm": 0.5285930037498474, + "learning_rate": 1.8738155949773517e-06, + "loss": 0.5533, + "num_input_tokens_seen": 53327712, + "step": 35040 + }, + { + "epoch": 1.1217271621535112, + "grad_norm": 1.0267802476882935, + "learning_rate": 1.8700881924146707e-06, + "loss": 0.5498, + "num_input_tokens_seen": 53335392, + "step": 35045 + }, + { + "epoch": 1.1218872031239997, + "grad_norm": 0.6332581043243408, + "learning_rate": 1.8663643567750577e-06, + "loss": 0.6139, + "num_input_tokens_seen": 53343104, + "step": 35050 + }, + { + "epoch": 1.1220472440944882, + "grad_norm": 0.45433178544044495, + "learning_rate": 1.8626440886327813e-06, + "loss": 0.6264, + "num_input_tokens_seen": 53350736, + "step": 35055 + }, + { + "epoch": 1.1222072850649767, + "grad_norm": 0.6143850684165955, + "learning_rate": 1.8589273885615432e-06, + "loss": 0.5368, + "num_input_tokens_seen": 53358768, + "step": 35060 + }, + { + "epoch": 1.122367326035465, + "grad_norm": 1.045348882675171, + "learning_rate": 1.8552142571345133e-06, + "loss": 0.437, + "num_input_tokens_seen": 53365696, + "step": 35065 + }, + { + "epoch": 1.1225273670059535, + "grad_norm": 0.9991343021392822, + "learning_rate": 1.8515046949243025e-06, + "loss": 0.5212, + "num_input_tokens_seen": 53373104, + "step": 35070 + }, + { + "epoch": 1.122687407976442, + "grad_norm": 0.36931145191192627, + "learning_rate": 1.8477987025029674e-06, + "loss": 0.3827, + "num_input_tokens_seen": 53380480, + "step": 35075 + }, + { + "epoch": 1.1228474489469304, + "grad_norm": 0.1965060532093048, + "learning_rate": 1.8440962804420232e-06, + "loss": 0.4114, + "num_input_tokens_seen": 53388576, + "step": 35080 + }, + { + "epoch": 1.123007489917419, + "grad_norm": 1.1245280504226685, + "learning_rate": 1.8403974293124265e-06, + "loss": 0.4732, + "num_input_tokens_seen": 53396784, + "step": 35085 + }, + { + "epoch": 1.1231675308879072, + "grad_norm": 0.38280701637268066, + "learning_rate": 1.8367021496845854e-06, + "loss": 0.6013, + "num_input_tokens_seen": 53404352, + "step": 35090 + }, + { + "epoch": 1.1233275718583957, + "grad_norm": 0.5747338533401489, + "learning_rate": 1.8330104421283662e-06, + "loss": 0.4665, + "num_input_tokens_seen": 53412208, + "step": 35095 + }, + { + "epoch": 1.1234876128288842, + "grad_norm": 1.0745623111724854, + "learning_rate": 1.8293223072130717e-06, + "loss": 0.4923, + "num_input_tokens_seen": 53419776, + "step": 35100 + }, + { + "epoch": 1.1236476537993727, + "grad_norm": 0.5604057312011719, + "learning_rate": 1.8256377455074525e-06, + "loss": 0.5793, + "num_input_tokens_seen": 53427248, + "step": 35105 + }, + { + "epoch": 1.123807694769861, + "grad_norm": 0.7284610271453857, + "learning_rate": 1.8219567575797263e-06, + "loss": 0.4683, + "num_input_tokens_seen": 53434480, + "step": 35110 + }, + { + "epoch": 1.1239677357403495, + "grad_norm": 0.6586262583732605, + "learning_rate": 1.8182793439975365e-06, + "loss": 0.3196, + "num_input_tokens_seen": 53442288, + "step": 35115 + }, + { + "epoch": 1.124127776710838, + "grad_norm": 1.032918095588684, + "learning_rate": 1.8146055053279958e-06, + "loss": 0.4554, + "num_input_tokens_seen": 53449728, + "step": 35120 + }, + { + "epoch": 1.1242878176813265, + "grad_norm": 0.6288018226623535, + "learning_rate": 1.8109352421376486e-06, + "loss": 0.3593, + "num_input_tokens_seen": 53456800, + "step": 35125 + }, + { + "epoch": 1.124447858651815, + "grad_norm": 1.606550693511963, + "learning_rate": 1.8072685549924972e-06, + "loss": 0.5627, + "num_input_tokens_seen": 53464576, + "step": 35130 + }, + { + "epoch": 1.1246078996223032, + "grad_norm": 0.9696722030639648, + "learning_rate": 1.8036054444579982e-06, + "loss": 0.4975, + "num_input_tokens_seen": 53472048, + "step": 35135 + }, + { + "epoch": 1.1247679405927917, + "grad_norm": 0.9575473070144653, + "learning_rate": 1.7999459110990407e-06, + "loss": 0.4809, + "num_input_tokens_seen": 53479856, + "step": 35140 + }, + { + "epoch": 1.1249279815632802, + "grad_norm": 0.9669926762580872, + "learning_rate": 1.7962899554799712e-06, + "loss": 0.5599, + "num_input_tokens_seen": 53487008, + "step": 35145 + }, + { + "epoch": 1.1250880225337687, + "grad_norm": 0.8753061890602112, + "learning_rate": 1.7926375781645937e-06, + "loss": 0.5621, + "num_input_tokens_seen": 53494768, + "step": 35150 + }, + { + "epoch": 1.125248063504257, + "grad_norm": 1.6821339130401611, + "learning_rate": 1.7889887797161359e-06, + "loss": 0.6017, + "num_input_tokens_seen": 53502304, + "step": 35155 + }, + { + "epoch": 1.1254081044747455, + "grad_norm": 0.6467716693878174, + "learning_rate": 1.7853435606973028e-06, + "loss": 0.5557, + "num_input_tokens_seen": 53509760, + "step": 35160 + }, + { + "epoch": 1.125568145445234, + "grad_norm": 1.0163726806640625, + "learning_rate": 1.781701921670223e-06, + "loss": 0.3935, + "num_input_tokens_seen": 53517600, + "step": 35165 + }, + { + "epoch": 1.1257281864157225, + "grad_norm": 1.1800357103347778, + "learning_rate": 1.7780638631964886e-06, + "loss": 0.4277, + "num_input_tokens_seen": 53525264, + "step": 35170 + }, + { + "epoch": 1.125888227386211, + "grad_norm": 0.38882628083229065, + "learning_rate": 1.7744293858371314e-06, + "loss": 0.3735, + "num_input_tokens_seen": 53532736, + "step": 35175 + }, + { + "epoch": 1.1260482683566992, + "grad_norm": 0.9079054594039917, + "learning_rate": 1.770798490152631e-06, + "loss": 0.5711, + "num_input_tokens_seen": 53540128, + "step": 35180 + }, + { + "epoch": 1.1262083093271877, + "grad_norm": 0.395527184009552, + "learning_rate": 1.767171176702917e-06, + "loss": 0.3402, + "num_input_tokens_seen": 53547888, + "step": 35185 + }, + { + "epoch": 1.1263683502976762, + "grad_norm": 0.6691354513168335, + "learning_rate": 1.7635474460473755e-06, + "loss": 0.5464, + "num_input_tokens_seen": 53555504, + "step": 35190 + }, + { + "epoch": 1.1265283912681647, + "grad_norm": 0.5707092881202698, + "learning_rate": 1.7599272987448206e-06, + "loss": 0.5002, + "num_input_tokens_seen": 53563232, + "step": 35195 + }, + { + "epoch": 1.126688432238653, + "grad_norm": 0.7930608987808228, + "learning_rate": 1.7563107353535362e-06, + "loss": 0.5124, + "num_input_tokens_seen": 53570848, + "step": 35200 + }, + { + "epoch": 1.126688432238653, + "eval_loss": 0.4944882392883301, + "eval_runtime": 557.3962, + "eval_samples_per_second": 24.912, + "eval_steps_per_second": 12.456, + "num_input_tokens_seen": 53570848, + "step": 35200 + }, + { + "epoch": 1.1268484732091415, + "grad_norm": 0.8943766355514526, + "learning_rate": 1.7526977564312263e-06, + "loss": 0.6131, + "num_input_tokens_seen": 53578608, + "step": 35205 + }, + { + "epoch": 1.12700851417963, + "grad_norm": 0.6137710809707642, + "learning_rate": 1.7490883625350701e-06, + "loss": 0.479, + "num_input_tokens_seen": 53585952, + "step": 35210 + }, + { + "epoch": 1.1271685551501185, + "grad_norm": 1.728399634361267, + "learning_rate": 1.7454825542216807e-06, + "loss": 0.5424, + "num_input_tokens_seen": 53592912, + "step": 35215 + }, + { + "epoch": 1.127328596120607, + "grad_norm": 0.44804686307907104, + "learning_rate": 1.7418803320471105e-06, + "loss": 0.5433, + "num_input_tokens_seen": 53600240, + "step": 35220 + }, + { + "epoch": 1.1274886370910953, + "grad_norm": 0.6562062501907349, + "learning_rate": 1.7382816965668737e-06, + "loss": 0.397, + "num_input_tokens_seen": 53607488, + "step": 35225 + }, + { + "epoch": 1.1276486780615838, + "grad_norm": 0.4801318347454071, + "learning_rate": 1.7346866483359285e-06, + "loss": 0.4112, + "num_input_tokens_seen": 53615008, + "step": 35230 + }, + { + "epoch": 1.1278087190320722, + "grad_norm": 1.080067753791809, + "learning_rate": 1.7310951879086657e-06, + "loss": 0.6092, + "num_input_tokens_seen": 53622272, + "step": 35235 + }, + { + "epoch": 1.1279687600025607, + "grad_norm": 0.9227017760276794, + "learning_rate": 1.7275073158389471e-06, + "loss": 0.4171, + "num_input_tokens_seen": 53629600, + "step": 35240 + }, + { + "epoch": 1.128128800973049, + "grad_norm": 0.5694355964660645, + "learning_rate": 1.723923032680061e-06, + "loss": 0.5422, + "num_input_tokens_seen": 53636784, + "step": 35245 + }, + { + "epoch": 1.1282888419435375, + "grad_norm": 0.9894905686378479, + "learning_rate": 1.7203423389847428e-06, + "loss": 0.5154, + "num_input_tokens_seen": 53643936, + "step": 35250 + }, + { + "epoch": 1.128448882914026, + "grad_norm": 0.47093093395233154, + "learning_rate": 1.7167652353051928e-06, + "loss": 0.5211, + "num_input_tokens_seen": 53652096, + "step": 35255 + }, + { + "epoch": 1.1286089238845145, + "grad_norm": 0.8029268383979797, + "learning_rate": 1.7131917221930333e-06, + "loss": 0.4604, + "num_input_tokens_seen": 53659568, + "step": 35260 + }, + { + "epoch": 1.128768964855003, + "grad_norm": 0.7824096083641052, + "learning_rate": 1.7096218001993513e-06, + "loss": 0.5071, + "num_input_tokens_seen": 53667248, + "step": 35265 + }, + { + "epoch": 1.1289290058254913, + "grad_norm": 0.6337498426437378, + "learning_rate": 1.706055469874676e-06, + "loss": 0.4429, + "num_input_tokens_seen": 53675312, + "step": 35270 + }, + { + "epoch": 1.1290890467959798, + "grad_norm": 0.6870077848434448, + "learning_rate": 1.702492731768976e-06, + "loss": 0.449, + "num_input_tokens_seen": 53683088, + "step": 35275 + }, + { + "epoch": 1.1292490877664683, + "grad_norm": 0.4609421193599701, + "learning_rate": 1.6989335864316724e-06, + "loss": 0.4891, + "num_input_tokens_seen": 53690336, + "step": 35280 + }, + { + "epoch": 1.1294091287369568, + "grad_norm": 1.138697862625122, + "learning_rate": 1.6953780344116265e-06, + "loss": 0.5553, + "num_input_tokens_seen": 53698000, + "step": 35285 + }, + { + "epoch": 1.129569169707445, + "grad_norm": 0.9433913230895996, + "learning_rate": 1.6918260762571497e-06, + "loss": 0.4121, + "num_input_tokens_seen": 53705456, + "step": 35290 + }, + { + "epoch": 1.1297292106779335, + "grad_norm": 0.5556656122207642, + "learning_rate": 1.6882777125160093e-06, + "loss": 0.4389, + "num_input_tokens_seen": 53713120, + "step": 35295 + }, + { + "epoch": 1.129889251648422, + "grad_norm": 0.7963104248046875, + "learning_rate": 1.6847329437353899e-06, + "loss": 0.3995, + "num_input_tokens_seen": 53721248, + "step": 35300 + }, + { + "epoch": 1.1300492926189105, + "grad_norm": 0.5020644068717957, + "learning_rate": 1.6811917704619511e-06, + "loss": 0.3612, + "num_input_tokens_seen": 53728288, + "step": 35305 + }, + { + "epoch": 1.1302093335893988, + "grad_norm": 0.6917355060577393, + "learning_rate": 1.67765419324179e-06, + "loss": 0.3601, + "num_input_tokens_seen": 53735680, + "step": 35310 + }, + { + "epoch": 1.1303693745598873, + "grad_norm": 0.6265881657600403, + "learning_rate": 1.6741202126204364e-06, + "loss": 0.5218, + "num_input_tokens_seen": 53743184, + "step": 35315 + }, + { + "epoch": 1.1305294155303758, + "grad_norm": 1.0528398752212524, + "learning_rate": 1.6705898291428767e-06, + "loss": 0.5951, + "num_input_tokens_seen": 53750896, + "step": 35320 + }, + { + "epoch": 1.1306894565008643, + "grad_norm": 0.5204370021820068, + "learning_rate": 1.6670630433535395e-06, + "loss": 0.557, + "num_input_tokens_seen": 53758656, + "step": 35325 + }, + { + "epoch": 1.1308494974713525, + "grad_norm": 1.7129054069519043, + "learning_rate": 1.6635398557962979e-06, + "loss": 0.5088, + "num_input_tokens_seen": 53766288, + "step": 35330 + }, + { + "epoch": 1.131009538441841, + "grad_norm": 0.529675304889679, + "learning_rate": 1.660020267014481e-06, + "loss": 0.4419, + "num_input_tokens_seen": 53774304, + "step": 35335 + }, + { + "epoch": 1.1311695794123295, + "grad_norm": 0.5687333941459656, + "learning_rate": 1.6565042775508438e-06, + "loss": 0.5236, + "num_input_tokens_seen": 53781616, + "step": 35340 + }, + { + "epoch": 1.131329620382818, + "grad_norm": 0.5759838819503784, + "learning_rate": 1.6529918879475997e-06, + "loss": 0.4278, + "num_input_tokens_seen": 53789632, + "step": 35345 + }, + { + "epoch": 1.1314896613533065, + "grad_norm": 0.820624589920044, + "learning_rate": 1.6494830987464043e-06, + "loss": 0.5462, + "num_input_tokens_seen": 53797824, + "step": 35350 + }, + { + "epoch": 1.1316497023237948, + "grad_norm": 0.47357475757598877, + "learning_rate": 1.6459779104883555e-06, + "loss": 0.3604, + "num_input_tokens_seen": 53805376, + "step": 35355 + }, + { + "epoch": 1.1318097432942833, + "grad_norm": 0.5226641893386841, + "learning_rate": 1.6424763237140013e-06, + "loss": 0.413, + "num_input_tokens_seen": 53813232, + "step": 35360 + }, + { + "epoch": 1.1319697842647718, + "grad_norm": 0.44168412685394287, + "learning_rate": 1.6389783389633207e-06, + "loss": 0.4044, + "num_input_tokens_seen": 53820928, + "step": 35365 + }, + { + "epoch": 1.1321298252352603, + "grad_norm": 1.471571683883667, + "learning_rate": 1.6354839567757546e-06, + "loss": 0.5701, + "num_input_tokens_seen": 53828528, + "step": 35370 + }, + { + "epoch": 1.1322898662057486, + "grad_norm": 1.4573315382003784, + "learning_rate": 1.6319931776901831e-06, + "loss": 0.3466, + "num_input_tokens_seen": 53835856, + "step": 35375 + }, + { + "epoch": 1.132449907176237, + "grad_norm": 1.2491835355758667, + "learning_rate": 1.6285060022449229e-06, + "loss": 0.453, + "num_input_tokens_seen": 53843312, + "step": 35380 + }, + { + "epoch": 1.1326099481467256, + "grad_norm": 0.8997844457626343, + "learning_rate": 1.6250224309777434e-06, + "loss": 0.5737, + "num_input_tokens_seen": 53850464, + "step": 35385 + }, + { + "epoch": 1.132769989117214, + "grad_norm": 0.6386669874191284, + "learning_rate": 1.6215424644258515e-06, + "loss": 0.5251, + "num_input_tokens_seen": 53857824, + "step": 35390 + }, + { + "epoch": 1.1329300300877025, + "grad_norm": 2.490657091140747, + "learning_rate": 1.6180661031259036e-06, + "loss": 0.6424, + "num_input_tokens_seen": 53865536, + "step": 35395 + }, + { + "epoch": 1.1330900710581908, + "grad_norm": 0.5199061036109924, + "learning_rate": 1.614593347613999e-06, + "loss": 0.3906, + "num_input_tokens_seen": 53873104, + "step": 35400 + }, + { + "epoch": 1.1330900710581908, + "eval_loss": 0.49439436197280884, + "eval_runtime": 556.673, + "eval_samples_per_second": 24.945, + "eval_steps_per_second": 12.472, + "num_input_tokens_seen": 53873104, + "step": 35400 + }, + { + "epoch": 1.1332501120286793, + "grad_norm": 0.9394104480743408, + "learning_rate": 1.6111241984256758e-06, + "loss": 0.3025, + "num_input_tokens_seen": 53880320, + "step": 35405 + }, + { + "epoch": 1.1334101529991678, + "grad_norm": 0.6878549456596375, + "learning_rate": 1.6076586560959257e-06, + "loss": 0.4414, + "num_input_tokens_seen": 53887616, + "step": 35410 + }, + { + "epoch": 1.1335701939696563, + "grad_norm": 0.4430363178253174, + "learning_rate": 1.604196721159182e-06, + "loss": 0.4237, + "num_input_tokens_seen": 53894816, + "step": 35415 + }, + { + "epoch": 1.1337302349401446, + "grad_norm": 0.34808221459388733, + "learning_rate": 1.6007383941493092e-06, + "loss": 0.6578, + "num_input_tokens_seen": 53902672, + "step": 35420 + }, + { + "epoch": 1.133890275910633, + "grad_norm": 0.6124733686447144, + "learning_rate": 1.5972836755996285e-06, + "loss": 0.4653, + "num_input_tokens_seen": 53910640, + "step": 35425 + }, + { + "epoch": 1.1340503168811216, + "grad_norm": 0.6975439190864563, + "learning_rate": 1.5938325660429076e-06, + "loss": 0.376, + "num_input_tokens_seen": 53918288, + "step": 35430 + }, + { + "epoch": 1.13421035785161, + "grad_norm": 0.6370826959609985, + "learning_rate": 1.5903850660113378e-06, + "loss": 0.4198, + "num_input_tokens_seen": 53925584, + "step": 35435 + }, + { + "epoch": 1.1343703988220986, + "grad_norm": 1.3958806991577148, + "learning_rate": 1.5869411760365826e-06, + "loss": 0.5938, + "num_input_tokens_seen": 53933392, + "step": 35440 + }, + { + "epoch": 1.1345304397925868, + "grad_norm": 1.0681569576263428, + "learning_rate": 1.58350089664972e-06, + "loss": 0.6356, + "num_input_tokens_seen": 53940832, + "step": 35445 + }, + { + "epoch": 1.1346904807630753, + "grad_norm": 0.45437759160995483, + "learning_rate": 1.5800642283812865e-06, + "loss": 0.4402, + "num_input_tokens_seen": 53948448, + "step": 35450 + }, + { + "epoch": 1.1348505217335638, + "grad_norm": 0.4933430850505829, + "learning_rate": 1.5766311717612698e-06, + "loss": 0.5825, + "num_input_tokens_seen": 53956192, + "step": 35455 + }, + { + "epoch": 1.1350105627040523, + "grad_norm": 0.39605605602264404, + "learning_rate": 1.5732017273190818e-06, + "loss": 0.593, + "num_input_tokens_seen": 53963488, + "step": 35460 + }, + { + "epoch": 1.1351706036745406, + "grad_norm": 0.668938934803009, + "learning_rate": 1.5697758955835806e-06, + "loss": 0.4228, + "num_input_tokens_seen": 53970896, + "step": 35465 + }, + { + "epoch": 1.135330644645029, + "grad_norm": 1.0184324979782104, + "learning_rate": 1.566353677083085e-06, + "loss": 0.6028, + "num_input_tokens_seen": 53979136, + "step": 35470 + }, + { + "epoch": 1.1354906856155176, + "grad_norm": 0.869257926940918, + "learning_rate": 1.562935072345334e-06, + "loss": 0.499, + "num_input_tokens_seen": 53987520, + "step": 35475 + }, + { + "epoch": 1.135650726586006, + "grad_norm": 0.524953305721283, + "learning_rate": 1.5595200818975281e-06, + "loss": 0.4814, + "num_input_tokens_seen": 53994960, + "step": 35480 + }, + { + "epoch": 1.1358107675564946, + "grad_norm": 1.1022979021072388, + "learning_rate": 1.5561087062662905e-06, + "loss": 0.5266, + "num_input_tokens_seen": 54002496, + "step": 35485 + }, + { + "epoch": 1.1359708085269828, + "grad_norm": 0.6524655222892761, + "learning_rate": 1.5527009459777087e-06, + "loss": 0.5878, + "num_input_tokens_seen": 54009664, + "step": 35490 + }, + { + "epoch": 1.1361308494974713, + "grad_norm": 0.8745843172073364, + "learning_rate": 1.5492968015572984e-06, + "loss": 0.678, + "num_input_tokens_seen": 54017280, + "step": 35495 + }, + { + "epoch": 1.1362908904679598, + "grad_norm": 0.29361864924430847, + "learning_rate": 1.5458962735300203e-06, + "loss": 0.3847, + "num_input_tokens_seen": 54025184, + "step": 35500 + }, + { + "epoch": 1.1364509314384483, + "grad_norm": 0.7487744688987732, + "learning_rate": 1.54249936242028e-06, + "loss": 0.5006, + "num_input_tokens_seen": 54033104, + "step": 35505 + }, + { + "epoch": 1.1366109724089366, + "grad_norm": 0.7388846278190613, + "learning_rate": 1.5391060687519222e-06, + "loss": 0.3008, + "num_input_tokens_seen": 54040416, + "step": 35510 + }, + { + "epoch": 1.136771013379425, + "grad_norm": 0.7706693410873413, + "learning_rate": 1.5357163930482367e-06, + "loss": 0.3879, + "num_input_tokens_seen": 54048112, + "step": 35515 + }, + { + "epoch": 1.1369310543499136, + "grad_norm": 0.5860503315925598, + "learning_rate": 1.532330335831955e-06, + "loss": 0.4361, + "num_input_tokens_seen": 54056256, + "step": 35520 + }, + { + "epoch": 1.137091095320402, + "grad_norm": 0.6038516163825989, + "learning_rate": 1.5289478976252491e-06, + "loss": 0.323, + "num_input_tokens_seen": 54063488, + "step": 35525 + }, + { + "epoch": 1.1372511362908906, + "grad_norm": 0.4848997890949249, + "learning_rate": 1.5255690789497345e-06, + "loss": 0.4652, + "num_input_tokens_seen": 54070992, + "step": 35530 + }, + { + "epoch": 1.1374111772613789, + "grad_norm": 0.8021281957626343, + "learning_rate": 1.5221938803264641e-06, + "loss": 0.4247, + "num_input_tokens_seen": 54078192, + "step": 35535 + }, + { + "epoch": 1.1375712182318674, + "grad_norm": 0.736685037612915, + "learning_rate": 1.518822302275938e-06, + "loss": 0.3657, + "num_input_tokens_seen": 54085856, + "step": 35540 + }, + { + "epoch": 1.1377312592023558, + "grad_norm": 0.6221773028373718, + "learning_rate": 1.5154543453180958e-06, + "loss": 0.4171, + "num_input_tokens_seen": 54093488, + "step": 35545 + }, + { + "epoch": 1.1378913001728443, + "grad_norm": 1.2043366432189941, + "learning_rate": 1.5120900099723167e-06, + "loss": 0.4893, + "num_input_tokens_seen": 54101056, + "step": 35550 + }, + { + "epoch": 1.1380513411433326, + "grad_norm": 0.5676681399345398, + "learning_rate": 1.5087292967574273e-06, + "loss": 0.4802, + "num_input_tokens_seen": 54108864, + "step": 35555 + }, + { + "epoch": 1.1382113821138211, + "grad_norm": 0.659120500087738, + "learning_rate": 1.5053722061916908e-06, + "loss": 0.4956, + "num_input_tokens_seen": 54116144, + "step": 35560 + }, + { + "epoch": 1.1383714230843096, + "grad_norm": 0.7066959142684937, + "learning_rate": 1.5020187387928124e-06, + "loss": 0.5742, + "num_input_tokens_seen": 54123632, + "step": 35565 + }, + { + "epoch": 1.138531464054798, + "grad_norm": 0.923302412033081, + "learning_rate": 1.4986688950779343e-06, + "loss": 0.547, + "num_input_tokens_seen": 54131344, + "step": 35570 + }, + { + "epoch": 1.1386915050252864, + "grad_norm": 0.7377892732620239, + "learning_rate": 1.495322675563654e-06, + "loss": 0.518, + "num_input_tokens_seen": 54139504, + "step": 35575 + }, + { + "epoch": 1.1388515459957749, + "grad_norm": 2.110609531402588, + "learning_rate": 1.4919800807659922e-06, + "loss": 0.3596, + "num_input_tokens_seen": 54148032, + "step": 35580 + }, + { + "epoch": 1.1390115869662634, + "grad_norm": 0.8454930782318115, + "learning_rate": 1.4886411112004255e-06, + "loss": 0.5521, + "num_input_tokens_seen": 54155776, + "step": 35585 + }, + { + "epoch": 1.1391716279367519, + "grad_norm": 0.5559260249137878, + "learning_rate": 1.4853057673818588e-06, + "loss": 0.2969, + "num_input_tokens_seen": 54163248, + "step": 35590 + }, + { + "epoch": 1.1393316689072401, + "grad_norm": 0.863146960735321, + "learning_rate": 1.481974049824647e-06, + "loss": 0.4352, + "num_input_tokens_seen": 54170800, + "step": 35595 + }, + { + "epoch": 1.1394917098777286, + "grad_norm": 1.3348777294158936, + "learning_rate": 1.4786459590425849e-06, + "loss": 0.4565, + "num_input_tokens_seen": 54178960, + "step": 35600 + }, + { + "epoch": 1.1394917098777286, + "eval_loss": 0.49437838792800903, + "eval_runtime": 532.8019, + "eval_samples_per_second": 26.062, + "eval_steps_per_second": 13.031, + "num_input_tokens_seen": 54178960, + "step": 35600 + }, + { + "epoch": 1.1396517508482171, + "grad_norm": 1.0474525690078735, + "learning_rate": 1.4753214955489036e-06, + "loss": 0.4399, + "num_input_tokens_seen": 54186400, + "step": 35605 + }, + { + "epoch": 1.1398117918187056, + "grad_norm": 0.613814651966095, + "learning_rate": 1.4720006598562737e-06, + "loss": 0.4223, + "num_input_tokens_seen": 54194336, + "step": 35610 + }, + { + "epoch": 1.1399718327891941, + "grad_norm": 0.8405812978744507, + "learning_rate": 1.4686834524768185e-06, + "loss": 0.4399, + "num_input_tokens_seen": 54201568, + "step": 35615 + }, + { + "epoch": 1.1401318737596824, + "grad_norm": 0.5757718086242676, + "learning_rate": 1.4653698739220844e-06, + "loss": 0.5263, + "num_input_tokens_seen": 54209136, + "step": 35620 + }, + { + "epoch": 1.1402919147301709, + "grad_norm": 0.814325213432312, + "learning_rate": 1.4620599247030715e-06, + "loss": 0.4227, + "num_input_tokens_seen": 54216880, + "step": 35625 + }, + { + "epoch": 1.1404519557006594, + "grad_norm": 0.6072387099266052, + "learning_rate": 1.4587536053302125e-06, + "loss": 0.4961, + "num_input_tokens_seen": 54224768, + "step": 35630 + }, + { + "epoch": 1.1406119966711479, + "grad_norm": 1.1659152507781982, + "learning_rate": 1.4554509163133862e-06, + "loss": 0.4819, + "num_input_tokens_seen": 54232416, + "step": 35635 + }, + { + "epoch": 1.1407720376416361, + "grad_norm": 0.8115803003311157, + "learning_rate": 1.4521518581619098e-06, + "loss": 0.5458, + "num_input_tokens_seen": 54240704, + "step": 35640 + }, + { + "epoch": 1.1409320786121246, + "grad_norm": 0.5216366052627563, + "learning_rate": 1.4488564313845348e-06, + "loss": 0.4868, + "num_input_tokens_seen": 54248496, + "step": 35645 + }, + { + "epoch": 1.1410921195826131, + "grad_norm": 1.4894198179244995, + "learning_rate": 1.4455646364894603e-06, + "loss": 0.6229, + "num_input_tokens_seen": 54257088, + "step": 35650 + }, + { + "epoch": 1.1412521605531016, + "grad_norm": 0.890421986579895, + "learning_rate": 1.4422764739843247e-06, + "loss": 0.4766, + "num_input_tokens_seen": 54264992, + "step": 35655 + }, + { + "epoch": 1.1414122015235901, + "grad_norm": 1.0130432844161987, + "learning_rate": 1.4389919443762e-06, + "loss": 0.4408, + "num_input_tokens_seen": 54272656, + "step": 35660 + }, + { + "epoch": 1.1415722424940784, + "grad_norm": 0.6078287363052368, + "learning_rate": 1.4357110481716063e-06, + "loss": 0.4254, + "num_input_tokens_seen": 54280672, + "step": 35665 + }, + { + "epoch": 1.141732283464567, + "grad_norm": 0.649196982383728, + "learning_rate": 1.4324337858764941e-06, + "loss": 0.3653, + "num_input_tokens_seen": 54288368, + "step": 35670 + }, + { + "epoch": 1.1418923244350554, + "grad_norm": 0.6401843428611755, + "learning_rate": 1.4291601579962622e-06, + "loss": 0.3736, + "num_input_tokens_seen": 54296016, + "step": 35675 + }, + { + "epoch": 1.142052365405544, + "grad_norm": 0.9844068884849548, + "learning_rate": 1.42589016503574e-06, + "loss": 0.7459, + "num_input_tokens_seen": 54303680, + "step": 35680 + }, + { + "epoch": 1.1422124063760322, + "grad_norm": 1.196803092956543, + "learning_rate": 1.4226238074992099e-06, + "loss": 0.5333, + "num_input_tokens_seen": 54311168, + "step": 35685 + }, + { + "epoch": 1.1423724473465207, + "grad_norm": 0.44425398111343384, + "learning_rate": 1.4193610858903778e-06, + "loss": 0.4096, + "num_input_tokens_seen": 54318688, + "step": 35690 + }, + { + "epoch": 1.1425324883170092, + "grad_norm": 1.5391755104064941, + "learning_rate": 1.416102000712402e-06, + "loss": 0.4291, + "num_input_tokens_seen": 54326368, + "step": 35695 + }, + { + "epoch": 1.1426925292874976, + "grad_norm": 0.5604559183120728, + "learning_rate": 1.4128465524678668e-06, + "loss": 0.5054, + "num_input_tokens_seen": 54334176, + "step": 35700 + }, + { + "epoch": 1.1428525702579861, + "grad_norm": 0.456371933221817, + "learning_rate": 1.4095947416588124e-06, + "loss": 0.5507, + "num_input_tokens_seen": 54341296, + "step": 35705 + }, + { + "epoch": 1.1430126112284744, + "grad_norm": 1.5874444246292114, + "learning_rate": 1.4063465687866983e-06, + "loss": 0.6985, + "num_input_tokens_seen": 54349504, + "step": 35710 + }, + { + "epoch": 1.143172652198963, + "grad_norm": 0.38922885060310364, + "learning_rate": 1.4031020343524438e-06, + "loss": 0.4863, + "num_input_tokens_seen": 54356992, + "step": 35715 + }, + { + "epoch": 1.1433326931694514, + "grad_norm": 0.9195135831832886, + "learning_rate": 1.3998611388563926e-06, + "loss": 0.4037, + "num_input_tokens_seen": 54364768, + "step": 35720 + }, + { + "epoch": 1.14349273413994, + "grad_norm": 0.7833739519119263, + "learning_rate": 1.3966238827983314e-06, + "loss": 0.5274, + "num_input_tokens_seen": 54372752, + "step": 35725 + }, + { + "epoch": 1.1436527751104282, + "grad_norm": 0.7352479696273804, + "learning_rate": 1.393390266677483e-06, + "loss": 0.5687, + "num_input_tokens_seen": 54380080, + "step": 35730 + }, + { + "epoch": 1.1438128160809167, + "grad_norm": 0.8166736364364624, + "learning_rate": 1.3901602909925204e-06, + "loss": 0.5359, + "num_input_tokens_seen": 54387376, + "step": 35735 + }, + { + "epoch": 1.1439728570514052, + "grad_norm": 0.7072277069091797, + "learning_rate": 1.3869339562415373e-06, + "loss": 0.5579, + "num_input_tokens_seen": 54394864, + "step": 35740 + }, + { + "epoch": 1.1441328980218937, + "grad_norm": 0.9925315976142883, + "learning_rate": 1.38371126292208e-06, + "loss": 0.5293, + "num_input_tokens_seen": 54402560, + "step": 35745 + }, + { + "epoch": 1.1442929389923822, + "grad_norm": 1.310701608657837, + "learning_rate": 1.3804922115311286e-06, + "loss": 0.6819, + "num_input_tokens_seen": 54409984, + "step": 35750 + }, + { + "epoch": 1.1444529799628704, + "grad_norm": 1.1697962284088135, + "learning_rate": 1.3772768025650945e-06, + "loss": 0.3802, + "num_input_tokens_seen": 54418224, + "step": 35755 + }, + { + "epoch": 1.144613020933359, + "grad_norm": 0.9324071407318115, + "learning_rate": 1.3740650365198448e-06, + "loss": 0.4782, + "num_input_tokens_seen": 54425712, + "step": 35760 + }, + { + "epoch": 1.1447730619038474, + "grad_norm": 0.5733529925346375, + "learning_rate": 1.3708569138906612e-06, + "loss": 0.5346, + "num_input_tokens_seen": 54433296, + "step": 35765 + }, + { + "epoch": 1.144933102874336, + "grad_norm": 1.1716411113739014, + "learning_rate": 1.367652435172287e-06, + "loss": 0.4736, + "num_input_tokens_seen": 54440624, + "step": 35770 + }, + { + "epoch": 1.1450931438448242, + "grad_norm": 0.8554869890213013, + "learning_rate": 1.364451600858893e-06, + "loss": 0.6627, + "num_input_tokens_seen": 54448496, + "step": 35775 + }, + { + "epoch": 1.1452531848153127, + "grad_norm": 0.7457057237625122, + "learning_rate": 1.3612544114440823e-06, + "loss": 0.3545, + "num_input_tokens_seen": 54456400, + "step": 35780 + }, + { + "epoch": 1.1454132257858012, + "grad_norm": 0.8873206377029419, + "learning_rate": 1.3580608674209072e-06, + "loss": 0.5357, + "num_input_tokens_seen": 54464032, + "step": 35785 + }, + { + "epoch": 1.1455732667562897, + "grad_norm": 0.9631444215774536, + "learning_rate": 1.3548709692818434e-06, + "loss": 0.5298, + "num_input_tokens_seen": 54471888, + "step": 35790 + }, + { + "epoch": 1.1457333077267782, + "grad_norm": 0.5972499847412109, + "learning_rate": 1.3516847175188223e-06, + "loss": 0.4265, + "num_input_tokens_seen": 54479600, + "step": 35795 + }, + { + "epoch": 1.1458933486972664, + "grad_norm": 1.1187467575073242, + "learning_rate": 1.348502112623204e-06, + "loss": 0.4616, + "num_input_tokens_seen": 54486752, + "step": 35800 + }, + { + "epoch": 1.1458933486972664, + "eval_loss": 0.49455875158309937, + "eval_runtime": 533.396, + "eval_samples_per_second": 26.033, + "eval_steps_per_second": 13.017, + "num_input_tokens_seen": 54486752, + "step": 35800 + }, + { + "epoch": 1.146053389667755, + "grad_norm": 0.7386647462844849, + "learning_rate": 1.3453231550857787e-06, + "loss": 0.3296, + "num_input_tokens_seen": 54494528, + "step": 35805 + }, + { + "epoch": 1.1462134306382434, + "grad_norm": 1.0117067098617554, + "learning_rate": 1.3421478453967878e-06, + "loss": 0.8111, + "num_input_tokens_seen": 54502224, + "step": 35810 + }, + { + "epoch": 1.146373471608732, + "grad_norm": 0.5075742602348328, + "learning_rate": 1.3389761840459065e-06, + "loss": 0.3817, + "num_input_tokens_seen": 54509200, + "step": 35815 + }, + { + "epoch": 1.1465335125792202, + "grad_norm": 0.4485440254211426, + "learning_rate": 1.3358081715222376e-06, + "loss": 0.4534, + "num_input_tokens_seen": 54516624, + "step": 35820 + }, + { + "epoch": 1.1466935535497087, + "grad_norm": 1.2142194509506226, + "learning_rate": 1.3326438083143295e-06, + "loss": 0.5749, + "num_input_tokens_seen": 54524048, + "step": 35825 + }, + { + "epoch": 1.1468535945201972, + "grad_norm": 0.7299299836158752, + "learning_rate": 1.3294830949101723e-06, + "loss": 0.4175, + "num_input_tokens_seen": 54531440, + "step": 35830 + }, + { + "epoch": 1.1470136354906857, + "grad_norm": 0.8424553275108337, + "learning_rate": 1.3263260317971815e-06, + "loss": 0.433, + "num_input_tokens_seen": 54539104, + "step": 35835 + }, + { + "epoch": 1.1471736764611742, + "grad_norm": 0.3044144809246063, + "learning_rate": 1.3231726194622208e-06, + "loss": 0.4315, + "num_input_tokens_seen": 54546768, + "step": 35840 + }, + { + "epoch": 1.1473337174316625, + "grad_norm": 0.5978630185127258, + "learning_rate": 1.3200228583915814e-06, + "loss": 0.4781, + "num_input_tokens_seen": 54554240, + "step": 35845 + }, + { + "epoch": 1.147493758402151, + "grad_norm": 1.8887205123901367, + "learning_rate": 1.3168767490709971e-06, + "loss": 0.6226, + "num_input_tokens_seen": 54561568, + "step": 35850 + }, + { + "epoch": 1.1476537993726394, + "grad_norm": 0.6131385564804077, + "learning_rate": 1.3137342919856437e-06, + "loss": 0.5219, + "num_input_tokens_seen": 54568880, + "step": 35855 + }, + { + "epoch": 1.1478138403431277, + "grad_norm": 0.4774203896522522, + "learning_rate": 1.310595487620117e-06, + "loss": 0.3586, + "num_input_tokens_seen": 54576592, + "step": 35860 + }, + { + "epoch": 1.1479738813136162, + "grad_norm": 0.7750375270843506, + "learning_rate": 1.3074603364584715e-06, + "loss": 0.483, + "num_input_tokens_seen": 54584352, + "step": 35865 + }, + { + "epoch": 1.1481339222841047, + "grad_norm": 0.7175117135047913, + "learning_rate": 1.3043288389841758e-06, + "loss": 0.5593, + "num_input_tokens_seen": 54591872, + "step": 35870 + }, + { + "epoch": 1.1482939632545932, + "grad_norm": 0.6814006567001343, + "learning_rate": 1.3012009956801546e-06, + "loss": 0.4494, + "num_input_tokens_seen": 54599648, + "step": 35875 + }, + { + "epoch": 1.1484540042250817, + "grad_norm": 0.6230738759040833, + "learning_rate": 1.2980768070287586e-06, + "loss": 0.6809, + "num_input_tokens_seen": 54607360, + "step": 35880 + }, + { + "epoch": 1.14861404519557, + "grad_norm": 0.5379196405410767, + "learning_rate": 1.2949562735117716e-06, + "loss": 0.3949, + "num_input_tokens_seen": 54614576, + "step": 35885 + }, + { + "epoch": 1.1487740861660585, + "grad_norm": 1.531001091003418, + "learning_rate": 1.291839395610428e-06, + "loss": 0.5673, + "num_input_tokens_seen": 54621936, + "step": 35890 + }, + { + "epoch": 1.148934127136547, + "grad_norm": 0.674045979976654, + "learning_rate": 1.2887261738053852e-06, + "loss": 0.3609, + "num_input_tokens_seen": 54629376, + "step": 35895 + }, + { + "epoch": 1.1490941681070355, + "grad_norm": 1.2635524272918701, + "learning_rate": 1.2856166085767396e-06, + "loss": 0.5763, + "num_input_tokens_seen": 54637152, + "step": 35900 + }, + { + "epoch": 1.1492542090775237, + "grad_norm": 0.8321492075920105, + "learning_rate": 1.2825107004040272e-06, + "loss": 0.3467, + "num_input_tokens_seen": 54644928, + "step": 35905 + }, + { + "epoch": 1.1494142500480122, + "grad_norm": 0.9278478026390076, + "learning_rate": 1.2794084497662146e-06, + "loss": 0.4257, + "num_input_tokens_seen": 54652672, + "step": 35910 + }, + { + "epoch": 1.1495742910185007, + "grad_norm": 0.5677594542503357, + "learning_rate": 1.276309857141711e-06, + "loss": 0.4013, + "num_input_tokens_seen": 54660416, + "step": 35915 + }, + { + "epoch": 1.1497343319889892, + "grad_norm": 0.7171262502670288, + "learning_rate": 1.273214923008359e-06, + "loss": 0.4574, + "num_input_tokens_seen": 54667872, + "step": 35920 + }, + { + "epoch": 1.1498943729594777, + "grad_norm": 1.0763248205184937, + "learning_rate": 1.2701236478434352e-06, + "loss": 0.4916, + "num_input_tokens_seen": 54675312, + "step": 35925 + }, + { + "epoch": 1.150054413929966, + "grad_norm": 0.42942944169044495, + "learning_rate": 1.2670360321236502e-06, + "loss": 0.3729, + "num_input_tokens_seen": 54682528, + "step": 35930 + }, + { + "epoch": 1.1502144549004545, + "grad_norm": 0.962393581867218, + "learning_rate": 1.2639520763251617e-06, + "loss": 0.3966, + "num_input_tokens_seen": 54690192, + "step": 35935 + }, + { + "epoch": 1.150374495870943, + "grad_norm": 0.5058780908584595, + "learning_rate": 1.2608717809235448e-06, + "loss": 0.4184, + "num_input_tokens_seen": 54697424, + "step": 35940 + }, + { + "epoch": 1.1505345368414315, + "grad_norm": 1.2017828226089478, + "learning_rate": 1.2577951463938282e-06, + "loss": 0.5642, + "num_input_tokens_seen": 54704752, + "step": 35945 + }, + { + "epoch": 1.1506945778119197, + "grad_norm": 0.83012455701828, + "learning_rate": 1.2547221732104569e-06, + "loss": 0.4435, + "num_input_tokens_seen": 54712080, + "step": 35950 + }, + { + "epoch": 1.1508546187824082, + "grad_norm": 0.8844779133796692, + "learning_rate": 1.25165286184733e-06, + "loss": 0.5687, + "num_input_tokens_seen": 54719744, + "step": 35955 + }, + { + "epoch": 1.1510146597528967, + "grad_norm": 0.7883891463279724, + "learning_rate": 1.248587212777777e-06, + "loss": 0.4603, + "num_input_tokens_seen": 54727200, + "step": 35960 + }, + { + "epoch": 1.1511747007233852, + "grad_norm": 0.6244942545890808, + "learning_rate": 1.2455252264745532e-06, + "loss": 0.3815, + "num_input_tokens_seen": 54734816, + "step": 35965 + }, + { + "epoch": 1.1513347416938737, + "grad_norm": 0.971721351146698, + "learning_rate": 1.2424669034098528e-06, + "loss": 0.4381, + "num_input_tokens_seen": 54742320, + "step": 35970 + }, + { + "epoch": 1.151494782664362, + "grad_norm": 1.7233678102493286, + "learning_rate": 1.2394122440553185e-06, + "loss": 0.5929, + "num_input_tokens_seen": 54749680, + "step": 35975 + }, + { + "epoch": 1.1516548236348505, + "grad_norm": 1.2318947315216064, + "learning_rate": 1.2363612488820037e-06, + "loss": 0.5761, + "num_input_tokens_seen": 54757216, + "step": 35980 + }, + { + "epoch": 1.151814864605339, + "grad_norm": 0.6976704597473145, + "learning_rate": 1.2333139183604208e-06, + "loss": 0.4189, + "num_input_tokens_seen": 54764688, + "step": 35985 + }, + { + "epoch": 1.1519749055758275, + "grad_norm": 1.0266841650009155, + "learning_rate": 1.2302702529604998e-06, + "loss": 0.6139, + "num_input_tokens_seen": 54772432, + "step": 35990 + }, + { + "epoch": 1.1521349465463158, + "grad_norm": 0.6083787679672241, + "learning_rate": 1.227230253151615e-06, + "loss": 0.4167, + "num_input_tokens_seen": 54780192, + "step": 35995 + }, + { + "epoch": 1.1522949875168043, + "grad_norm": 0.8445120453834534, + "learning_rate": 1.2241939194025748e-06, + "loss": 0.6993, + "num_input_tokens_seen": 54787456, + "step": 36000 + }, + { + "epoch": 1.1522949875168043, + "eval_loss": 0.4942479431629181, + "eval_runtime": 533.5975, + "eval_samples_per_second": 26.023, + "eval_steps_per_second": 13.012, + "num_input_tokens_seen": 54787456, + "step": 36000 + }, + { + "epoch": 1.1524550284872928, + "grad_norm": 0.5296769738197327, + "learning_rate": 1.2211612521816156e-06, + "loss": 0.5409, + "num_input_tokens_seen": 54794864, + "step": 36005 + }, + { + "epoch": 1.1526150694577812, + "grad_norm": 0.5094307065010071, + "learning_rate": 1.2181322519564137e-06, + "loss": 0.6464, + "num_input_tokens_seen": 54802672, + "step": 36010 + }, + { + "epoch": 1.1527751104282697, + "grad_norm": 0.7537270784378052, + "learning_rate": 1.2151069191940839e-06, + "loss": 0.4251, + "num_input_tokens_seen": 54811248, + "step": 36015 + }, + { + "epoch": 1.152935151398758, + "grad_norm": 1.3028062582015991, + "learning_rate": 1.2120852543611644e-06, + "loss": 0.4559, + "num_input_tokens_seen": 54818736, + "step": 36020 + }, + { + "epoch": 1.1530951923692465, + "grad_norm": 1.1633273363113403, + "learning_rate": 1.2090672579236379e-06, + "loss": 0.4837, + "num_input_tokens_seen": 54826080, + "step": 36025 + }, + { + "epoch": 1.153255233339735, + "grad_norm": 0.9465492367744446, + "learning_rate": 1.2060529303469126e-06, + "loss": 0.4409, + "num_input_tokens_seen": 54833456, + "step": 36030 + }, + { + "epoch": 1.1534152743102235, + "grad_norm": 0.5792670845985413, + "learning_rate": 1.2030422720958445e-06, + "loss": 0.4256, + "num_input_tokens_seen": 54841168, + "step": 36035 + }, + { + "epoch": 1.1535753152807118, + "grad_norm": 0.500637948513031, + "learning_rate": 1.200035283634704e-06, + "loss": 0.5663, + "num_input_tokens_seen": 54849024, + "step": 36040 + }, + { + "epoch": 1.1537353562512003, + "grad_norm": 0.7083380222320557, + "learning_rate": 1.1970319654272144e-06, + "loss": 0.5281, + "num_input_tokens_seen": 54856208, + "step": 36045 + }, + { + "epoch": 1.1538953972216888, + "grad_norm": 0.7166444063186646, + "learning_rate": 1.1940323179365192e-06, + "loss": 0.4573, + "num_input_tokens_seen": 54864528, + "step": 36050 + }, + { + "epoch": 1.1540554381921773, + "grad_norm": 1.5059553384780884, + "learning_rate": 1.1910363416252095e-06, + "loss": 0.5654, + "num_input_tokens_seen": 54871792, + "step": 36055 + }, + { + "epoch": 1.1542154791626658, + "grad_norm": 0.44765743613243103, + "learning_rate": 1.1880440369552964e-06, + "loss": 0.3515, + "num_input_tokens_seen": 54879520, + "step": 36060 + }, + { + "epoch": 1.154375520133154, + "grad_norm": 0.4531058967113495, + "learning_rate": 1.1850554043882328e-06, + "loss": 0.7138, + "num_input_tokens_seen": 54887168, + "step": 36065 + }, + { + "epoch": 1.1545355611036425, + "grad_norm": 1.072288990020752, + "learning_rate": 1.1820704443849028e-06, + "loss": 0.4958, + "num_input_tokens_seen": 54894992, + "step": 36070 + }, + { + "epoch": 1.154695602074131, + "grad_norm": 0.8962354063987732, + "learning_rate": 1.1790891574056219e-06, + "loss": 0.5023, + "num_input_tokens_seen": 54902112, + "step": 36075 + }, + { + "epoch": 1.1548556430446195, + "grad_norm": 0.7190349698066711, + "learning_rate": 1.1761115439101523e-06, + "loss": 0.4834, + "num_input_tokens_seen": 54909376, + "step": 36080 + }, + { + "epoch": 1.1550156840151078, + "grad_norm": 1.0282700061798096, + "learning_rate": 1.1731376043576659e-06, + "loss": 0.6029, + "num_input_tokens_seen": 54916560, + "step": 36085 + }, + { + "epoch": 1.1551757249855963, + "grad_norm": 0.8716580271720886, + "learning_rate": 1.1701673392067875e-06, + "loss": 0.4057, + "num_input_tokens_seen": 54924128, + "step": 36090 + }, + { + "epoch": 1.1553357659560848, + "grad_norm": 1.0161188840866089, + "learning_rate": 1.1672007489155757e-06, + "loss": 0.4169, + "num_input_tokens_seen": 54931808, + "step": 36095 + }, + { + "epoch": 1.1554958069265733, + "grad_norm": 0.5948909521102905, + "learning_rate": 1.164237833941506e-06, + "loss": 0.399, + "num_input_tokens_seen": 54939328, + "step": 36100 + }, + { + "epoch": 1.1556558478970618, + "grad_norm": 0.6736564040184021, + "learning_rate": 1.1612785947415022e-06, + "loss": 0.3238, + "num_input_tokens_seen": 54946464, + "step": 36105 + }, + { + "epoch": 1.15581588886755, + "grad_norm": 0.7652236223220825, + "learning_rate": 1.1583230317719185e-06, + "loss": 0.4696, + "num_input_tokens_seen": 54953936, + "step": 36110 + }, + { + "epoch": 1.1559759298380385, + "grad_norm": 0.5363258123397827, + "learning_rate": 1.1553711454885318e-06, + "loss": 0.3837, + "num_input_tokens_seen": 54961680, + "step": 36115 + }, + { + "epoch": 1.156135970808527, + "grad_norm": 0.6261836290359497, + "learning_rate": 1.152422936346567e-06, + "loss": 0.4293, + "num_input_tokens_seen": 54969040, + "step": 36120 + }, + { + "epoch": 1.1562960117790153, + "grad_norm": 0.7790428400039673, + "learning_rate": 1.1494784048006718e-06, + "loss": 0.4205, + "num_input_tokens_seen": 54976240, + "step": 36125 + }, + { + "epoch": 1.1564560527495038, + "grad_norm": 0.4097975790500641, + "learning_rate": 1.1465375513049326e-06, + "loss": 0.5812, + "num_input_tokens_seen": 54983728, + "step": 36130 + }, + { + "epoch": 1.1566160937199923, + "grad_norm": 0.793144941329956, + "learning_rate": 1.1436003763128616e-06, + "loss": 0.5078, + "num_input_tokens_seen": 54991024, + "step": 36135 + }, + { + "epoch": 1.1567761346904808, + "grad_norm": 0.7440252900123596, + "learning_rate": 1.1406668802774106e-06, + "loss": 0.4248, + "num_input_tokens_seen": 54998528, + "step": 36140 + }, + { + "epoch": 1.1569361756609693, + "grad_norm": 1.2448782920837402, + "learning_rate": 1.137737063650965e-06, + "loss": 0.5794, + "num_input_tokens_seen": 55006592, + "step": 36145 + }, + { + "epoch": 1.1570962166314576, + "grad_norm": 0.6494702100753784, + "learning_rate": 1.1348109268853323e-06, + "loss": 0.428, + "num_input_tokens_seen": 55014448, + "step": 36150 + }, + { + "epoch": 1.157256257601946, + "grad_norm": 0.8988195061683655, + "learning_rate": 1.1318884704317634e-06, + "loss": 0.559, + "num_input_tokens_seen": 55021968, + "step": 36155 + }, + { + "epoch": 1.1574162985724346, + "grad_norm": 1.177137017250061, + "learning_rate": 1.1289696947409417e-06, + "loss": 0.3997, + "num_input_tokens_seen": 55029568, + "step": 36160 + }, + { + "epoch": 1.157576339542923, + "grad_norm": 0.5524616837501526, + "learning_rate": 1.126054600262974e-06, + "loss": 0.4177, + "num_input_tokens_seen": 55037392, + "step": 36165 + }, + { + "epoch": 1.1577363805134113, + "grad_norm": 0.7163705229759216, + "learning_rate": 1.1231431874474064e-06, + "loss": 0.3545, + "num_input_tokens_seen": 55044768, + "step": 36170 + }, + { + "epoch": 1.1578964214838998, + "grad_norm": 0.4613659083843231, + "learning_rate": 1.12023545674321e-06, + "loss": 0.3751, + "num_input_tokens_seen": 55052368, + "step": 36175 + }, + { + "epoch": 1.1580564624543883, + "grad_norm": 0.5342206954956055, + "learning_rate": 1.117331408598804e-06, + "loss": 0.4578, + "num_input_tokens_seen": 55059632, + "step": 36180 + }, + { + "epoch": 1.1582165034248768, + "grad_norm": 0.7631288170814514, + "learning_rate": 1.1144310434620191e-06, + "loss": 0.4305, + "num_input_tokens_seen": 55067216, + "step": 36185 + }, + { + "epoch": 1.1583765443953653, + "grad_norm": 1.1516940593719482, + "learning_rate": 1.1115343617801365e-06, + "loss": 0.4574, + "num_input_tokens_seen": 55074992, + "step": 36190 + }, + { + "epoch": 1.1585365853658536, + "grad_norm": 1.1374174356460571, + "learning_rate": 1.1086413639998515e-06, + "loss": 0.5379, + "num_input_tokens_seen": 55082544, + "step": 36195 + }, + { + "epoch": 1.158696626336342, + "grad_norm": 0.5323073267936707, + "learning_rate": 1.1057520505673103e-06, + "loss": 0.3966, + "num_input_tokens_seen": 55090272, + "step": 36200 + }, + { + "epoch": 1.158696626336342, + "eval_loss": 0.49430230259895325, + "eval_runtime": 532.7161, + "eval_samples_per_second": 26.066, + "eval_steps_per_second": 13.033, + "num_input_tokens_seen": 55090272, + "step": 36200 + }, + { + "epoch": 1.1588566673068306, + "grad_norm": 0.7123360633850098, + "learning_rate": 1.1028664219280727e-06, + "loss": 0.6101, + "num_input_tokens_seen": 55098464, + "step": 36205 + }, + { + "epoch": 1.159016708277319, + "grad_norm": 1.0704387426376343, + "learning_rate": 1.0999844785271468e-06, + "loss": 0.5197, + "num_input_tokens_seen": 55105936, + "step": 36210 + }, + { + "epoch": 1.1591767492478073, + "grad_norm": 0.7444703578948975, + "learning_rate": 1.097106220808955e-06, + "loss": 0.4279, + "num_input_tokens_seen": 55113088, + "step": 36215 + }, + { + "epoch": 1.1593367902182958, + "grad_norm": 0.5216191411018372, + "learning_rate": 1.0942316492173698e-06, + "loss": 0.5493, + "num_input_tokens_seen": 55120576, + "step": 36220 + }, + { + "epoch": 1.1594968311887843, + "grad_norm": 0.5835695862770081, + "learning_rate": 1.0913607641956841e-06, + "loss": 0.3724, + "num_input_tokens_seen": 55128272, + "step": 36225 + }, + { + "epoch": 1.1596568721592728, + "grad_norm": 0.6502135396003723, + "learning_rate": 1.0884935661866213e-06, + "loss": 0.5001, + "num_input_tokens_seen": 55135552, + "step": 36230 + }, + { + "epoch": 1.1598169131297613, + "grad_norm": 0.697220504283905, + "learning_rate": 1.0856300556323418e-06, + "loss": 0.4211, + "num_input_tokens_seen": 55142800, + "step": 36235 + }, + { + "epoch": 1.1599769541002496, + "grad_norm": 0.45129814743995667, + "learning_rate": 1.0827702329744365e-06, + "loss": 0.4717, + "num_input_tokens_seen": 55150336, + "step": 36240 + }, + { + "epoch": 1.160136995070738, + "grad_norm": 0.6878434419631958, + "learning_rate": 1.0799140986539197e-06, + "loss": 0.5375, + "num_input_tokens_seen": 55158144, + "step": 36245 + }, + { + "epoch": 1.1602970360412266, + "grad_norm": 0.4909769594669342, + "learning_rate": 1.0770616531112526e-06, + "loss": 0.4242, + "num_input_tokens_seen": 55165424, + "step": 36250 + }, + { + "epoch": 1.160457077011715, + "grad_norm": 0.5143062472343445, + "learning_rate": 1.0742128967863085e-06, + "loss": 0.324, + "num_input_tokens_seen": 55173184, + "step": 36255 + }, + { + "epoch": 1.1606171179822034, + "grad_norm": 0.38587459921836853, + "learning_rate": 1.071367830118411e-06, + "loss": 0.4479, + "num_input_tokens_seen": 55181408, + "step": 36260 + }, + { + "epoch": 1.1607771589526918, + "grad_norm": 1.4146331548690796, + "learning_rate": 1.068526453546298e-06, + "loss": 0.5028, + "num_input_tokens_seen": 55188992, + "step": 36265 + }, + { + "epoch": 1.1609371999231803, + "grad_norm": 0.8701997995376587, + "learning_rate": 1.0656887675081467e-06, + "loss": 0.3302, + "num_input_tokens_seen": 55197088, + "step": 36270 + }, + { + "epoch": 1.1610972408936688, + "grad_norm": 0.6644944548606873, + "learning_rate": 1.0628547724415628e-06, + "loss": 0.4289, + "num_input_tokens_seen": 55204544, + "step": 36275 + }, + { + "epoch": 1.1612572818641573, + "grad_norm": 0.840222179889679, + "learning_rate": 1.0600244687835881e-06, + "loss": 0.4264, + "num_input_tokens_seen": 55211984, + "step": 36280 + }, + { + "epoch": 1.1614173228346456, + "grad_norm": 1.1610571146011353, + "learning_rate": 1.0571978569706876e-06, + "loss": 0.3984, + "num_input_tokens_seen": 55220080, + "step": 36285 + }, + { + "epoch": 1.161577363805134, + "grad_norm": 0.7200119495391846, + "learning_rate": 1.0543749374387652e-06, + "loss": 0.3831, + "num_input_tokens_seen": 55227616, + "step": 36290 + }, + { + "epoch": 1.1617374047756226, + "grad_norm": 1.1608792543411255, + "learning_rate": 1.051555710623142e-06, + "loss": 0.5835, + "num_input_tokens_seen": 55235648, + "step": 36295 + }, + { + "epoch": 1.161897445746111, + "grad_norm": 0.6958792209625244, + "learning_rate": 1.0487401769585847e-06, + "loss": 0.6031, + "num_input_tokens_seen": 55243536, + "step": 36300 + }, + { + "epoch": 1.1620574867165994, + "grad_norm": 0.8632726669311523, + "learning_rate": 1.0459283368792845e-06, + "loss": 0.484, + "num_input_tokens_seen": 55250944, + "step": 36305 + }, + { + "epoch": 1.1622175276870879, + "grad_norm": 1.1457436084747314, + "learning_rate": 1.043120190818858e-06, + "loss": 0.5089, + "num_input_tokens_seen": 55258704, + "step": 36310 + }, + { + "epoch": 1.1623775686575764, + "grad_norm": 0.6815905570983887, + "learning_rate": 1.0403157392103596e-06, + "loss": 0.4795, + "num_input_tokens_seen": 55266192, + "step": 36315 + }, + { + "epoch": 1.1625376096280648, + "grad_norm": 1.43303644657135, + "learning_rate": 1.0375149824862735e-06, + "loss": 0.4912, + "num_input_tokens_seen": 55273936, + "step": 36320 + }, + { + "epoch": 1.1626976505985533, + "grad_norm": 0.6225154995918274, + "learning_rate": 1.034717921078507e-06, + "loss": 0.4056, + "num_input_tokens_seen": 55281680, + "step": 36325 + }, + { + "epoch": 1.1628576915690416, + "grad_norm": 0.291517049074173, + "learning_rate": 1.0319245554184009e-06, + "loss": 0.4315, + "num_input_tokens_seen": 55289200, + "step": 36330 + }, + { + "epoch": 1.1630177325395301, + "grad_norm": 1.4461548328399658, + "learning_rate": 1.0291348859367361e-06, + "loss": 0.5464, + "num_input_tokens_seen": 55296400, + "step": 36335 + }, + { + "epoch": 1.1631777735100186, + "grad_norm": 0.5248846411705017, + "learning_rate": 1.0263489130637016e-06, + "loss": 0.5465, + "num_input_tokens_seen": 55303728, + "step": 36340 + }, + { + "epoch": 1.163337814480507, + "grad_norm": 1.336418867111206, + "learning_rate": 1.0235666372289427e-06, + "loss": 0.4704, + "num_input_tokens_seen": 55311312, + "step": 36345 + }, + { + "epoch": 1.1634978554509954, + "grad_norm": 0.8303534388542175, + "learning_rate": 1.0207880588615076e-06, + "loss": 0.4009, + "num_input_tokens_seen": 55318960, + "step": 36350 + }, + { + "epoch": 1.1636578964214839, + "grad_norm": 0.5989428162574768, + "learning_rate": 1.0180131783898984e-06, + "loss": 0.4359, + "num_input_tokens_seen": 55326816, + "step": 36355 + }, + { + "epoch": 1.1638179373919724, + "grad_norm": 0.9510461091995239, + "learning_rate": 1.0152419962420362e-06, + "loss": 0.5609, + "num_input_tokens_seen": 55334032, + "step": 36360 + }, + { + "epoch": 1.1639779783624609, + "grad_norm": 0.6799313426017761, + "learning_rate": 1.0124745128452685e-06, + "loss": 0.4642, + "num_input_tokens_seen": 55341552, + "step": 36365 + }, + { + "epoch": 1.1641380193329494, + "grad_norm": 1.9328629970550537, + "learning_rate": 1.0097107286263758e-06, + "loss": 0.5981, + "num_input_tokens_seen": 55348448, + "step": 36370 + }, + { + "epoch": 1.1642980603034376, + "grad_norm": 0.6841062307357788, + "learning_rate": 1.00695064401157e-06, + "loss": 0.4575, + "num_input_tokens_seen": 55355696, + "step": 36375 + }, + { + "epoch": 1.1644581012739261, + "grad_norm": 0.6842413544654846, + "learning_rate": 1.0041942594264886e-06, + "loss": 0.5234, + "num_input_tokens_seen": 55363456, + "step": 36380 + }, + { + "epoch": 1.1646181422444146, + "grad_norm": 0.906227707862854, + "learning_rate": 1.001441575296208e-06, + "loss": 0.4674, + "num_input_tokens_seen": 55371104, + "step": 36385 + }, + { + "epoch": 1.164778183214903, + "grad_norm": 0.5237711071968079, + "learning_rate": 9.986925920452139e-07, + "loss": 0.3798, + "num_input_tokens_seen": 55378640, + "step": 36390 + }, + { + "epoch": 1.1649382241853914, + "grad_norm": 0.9529891610145569, + "learning_rate": 9.959473100974475e-07, + "loss": 0.4761, + "num_input_tokens_seen": 55385824, + "step": 36395 + }, + { + "epoch": 1.1650982651558799, + "grad_norm": 1.2079890966415405, + "learning_rate": 9.932057298762564e-07, + "loss": 0.4756, + "num_input_tokens_seen": 55393072, + "step": 36400 + }, + { + "epoch": 1.1650982651558799, + "eval_loss": 0.49429166316986084, + "eval_runtime": 533.4773, + "eval_samples_per_second": 26.029, + "eval_steps_per_second": 13.015, + "num_input_tokens_seen": 55393072, + "step": 36400 + }, + { + "epoch": 1.1652583061263684, + "grad_norm": 0.9877011775970459, + "learning_rate": 9.90467851804433e-07, + "loss": 0.4516, + "num_input_tokens_seen": 55400736, + "step": 36405 + }, + { + "epoch": 1.1654183470968569, + "grad_norm": 0.5893441438674927, + "learning_rate": 9.877336763041895e-07, + "loss": 0.3525, + "num_input_tokens_seen": 55408352, + "step": 36410 + }, + { + "epoch": 1.1655783880673452, + "grad_norm": 1.3136757612228394, + "learning_rate": 9.850032037971662e-07, + "loss": 0.4769, + "num_input_tokens_seen": 55415792, + "step": 36415 + }, + { + "epoch": 1.1657384290378336, + "grad_norm": 0.7287224531173706, + "learning_rate": 9.822764347044406e-07, + "loss": 0.5806, + "num_input_tokens_seen": 55423232, + "step": 36420 + }, + { + "epoch": 1.1658984700083221, + "grad_norm": 0.5423542857170105, + "learning_rate": 9.795533694465175e-07, + "loss": 0.3627, + "num_input_tokens_seen": 55431360, + "step": 36425 + }, + { + "epoch": 1.1660585109788106, + "grad_norm": 1.096528172492981, + "learning_rate": 9.768340084433197e-07, + "loss": 0.4715, + "num_input_tokens_seen": 55438848, + "step": 36430 + }, + { + "epoch": 1.166218551949299, + "grad_norm": 0.7338435053825378, + "learning_rate": 9.741183521142143e-07, + "loss": 0.4826, + "num_input_tokens_seen": 55446448, + "step": 36435 + }, + { + "epoch": 1.1663785929197874, + "grad_norm": 0.6797317862510681, + "learning_rate": 9.714064008779889e-07, + "loss": 0.3421, + "num_input_tokens_seen": 55454368, + "step": 36440 + }, + { + "epoch": 1.166538633890276, + "grad_norm": 1.8147796392440796, + "learning_rate": 9.686981551528584e-07, + "loss": 0.6219, + "num_input_tokens_seen": 55461824, + "step": 36445 + }, + { + "epoch": 1.1666986748607644, + "grad_norm": 0.7060744762420654, + "learning_rate": 9.65993615356467e-07, + "loss": 0.3102, + "num_input_tokens_seen": 55469744, + "step": 36450 + }, + { + "epoch": 1.166858715831253, + "grad_norm": 0.7713801264762878, + "learning_rate": 9.632927819058917e-07, + "loss": 0.4537, + "num_input_tokens_seen": 55477328, + "step": 36455 + }, + { + "epoch": 1.1670187568017412, + "grad_norm": 0.9430317282676697, + "learning_rate": 9.605956552176305e-07, + "loss": 0.6302, + "num_input_tokens_seen": 55484640, + "step": 36460 + }, + { + "epoch": 1.1671787977722297, + "grad_norm": 0.5819905996322632, + "learning_rate": 9.579022357076223e-07, + "loss": 0.4621, + "num_input_tokens_seen": 55492256, + "step": 36465 + }, + { + "epoch": 1.1673388387427182, + "grad_norm": 0.2891734540462494, + "learning_rate": 9.552125237912158e-07, + "loss": 0.4552, + "num_input_tokens_seen": 55499888, + "step": 36470 + }, + { + "epoch": 1.1674988797132066, + "grad_norm": 0.6868339776992798, + "learning_rate": 9.525265198832096e-07, + "loss": 0.5857, + "num_input_tokens_seen": 55507584, + "step": 36475 + }, + { + "epoch": 1.167658920683695, + "grad_norm": 0.4250728189945221, + "learning_rate": 9.498442243978112e-07, + "loss": 0.4314, + "num_input_tokens_seen": 55515232, + "step": 36480 + }, + { + "epoch": 1.1678189616541834, + "grad_norm": 0.6989117860794067, + "learning_rate": 9.471656377486649e-07, + "loss": 0.4506, + "num_input_tokens_seen": 55522352, + "step": 36485 + }, + { + "epoch": 1.167979002624672, + "grad_norm": 1.100056529045105, + "learning_rate": 9.444907603488456e-07, + "loss": 0.4776, + "num_input_tokens_seen": 55529840, + "step": 36490 + }, + { + "epoch": 1.1681390435951604, + "grad_norm": 0.5791164636611938, + "learning_rate": 9.418195926108514e-07, + "loss": 0.3168, + "num_input_tokens_seen": 55537024, + "step": 36495 + }, + { + "epoch": 1.168299084565649, + "grad_norm": 1.0759344100952148, + "learning_rate": 9.391521349466053e-07, + "loss": 0.3962, + "num_input_tokens_seen": 55544816, + "step": 36500 + }, + { + "epoch": 1.1684591255361372, + "grad_norm": 1.0789626836776733, + "learning_rate": 9.364883877674758e-07, + "loss": 0.366, + "num_input_tokens_seen": 55552640, + "step": 36505 + }, + { + "epoch": 1.1686191665066257, + "grad_norm": 0.6476570963859558, + "learning_rate": 9.33828351484231e-07, + "loss": 0.4826, + "num_input_tokens_seen": 55560608, + "step": 36510 + }, + { + "epoch": 1.1687792074771142, + "grad_norm": 1.0605789422988892, + "learning_rate": 9.311720265070906e-07, + "loss": 0.6665, + "num_input_tokens_seen": 55568368, + "step": 36515 + }, + { + "epoch": 1.1689392484476027, + "grad_norm": 0.5519698262214661, + "learning_rate": 9.285194132456931e-07, + "loss": 0.4186, + "num_input_tokens_seen": 55575632, + "step": 36520 + }, + { + "epoch": 1.169099289418091, + "grad_norm": 1.5043110847473145, + "learning_rate": 9.258705121091032e-07, + "loss": 0.601, + "num_input_tokens_seen": 55582896, + "step": 36525 + }, + { + "epoch": 1.1692593303885794, + "grad_norm": 0.686950147151947, + "learning_rate": 9.232253235058136e-07, + "loss": 0.5267, + "num_input_tokens_seen": 55590176, + "step": 36530 + }, + { + "epoch": 1.169419371359068, + "grad_norm": 0.70945805311203, + "learning_rate": 9.205838478437478e-07, + "loss": 0.5659, + "num_input_tokens_seen": 55597936, + "step": 36535 + }, + { + "epoch": 1.1695794123295564, + "grad_norm": 0.7554287910461426, + "learning_rate": 9.179460855302524e-07, + "loss": 0.3529, + "num_input_tokens_seen": 55605360, + "step": 36540 + }, + { + "epoch": 1.169739453300045, + "grad_norm": 0.7836135029792786, + "learning_rate": 9.153120369721046e-07, + "loss": 0.5261, + "num_input_tokens_seen": 55613088, + "step": 36545 + }, + { + "epoch": 1.1698994942705332, + "grad_norm": 1.3639230728149414, + "learning_rate": 9.126817025755103e-07, + "loss": 0.5557, + "num_input_tokens_seen": 55620192, + "step": 36550 + }, + { + "epoch": 1.1700595352410217, + "grad_norm": 0.7788616418838501, + "learning_rate": 9.100550827460947e-07, + "loss": 0.3806, + "num_input_tokens_seen": 55627600, + "step": 36555 + }, + { + "epoch": 1.1702195762115102, + "grad_norm": 0.5084798336029053, + "learning_rate": 9.0743217788892e-07, + "loss": 0.4222, + "num_input_tokens_seen": 55635264, + "step": 36560 + }, + { + "epoch": 1.1703796171819987, + "grad_norm": 0.4659819006919861, + "learning_rate": 9.048129884084683e-07, + "loss": 0.4722, + "num_input_tokens_seen": 55642880, + "step": 36565 + }, + { + "epoch": 1.170539658152487, + "grad_norm": 0.739615797996521, + "learning_rate": 9.021975147086553e-07, + "loss": 0.4092, + "num_input_tokens_seen": 55650560, + "step": 36570 + }, + { + "epoch": 1.1706996991229754, + "grad_norm": 0.9696634411811829, + "learning_rate": 8.995857571928141e-07, + "loss": 0.5345, + "num_input_tokens_seen": 55658464, + "step": 36575 + }, + { + "epoch": 1.170859740093464, + "grad_norm": 0.6298927664756775, + "learning_rate": 8.969777162637139e-07, + "loss": 0.4418, + "num_input_tokens_seen": 55665760, + "step": 36580 + }, + { + "epoch": 1.1710197810639524, + "grad_norm": 0.8103079199790955, + "learning_rate": 8.943733923235525e-07, + "loss": 0.4672, + "num_input_tokens_seen": 55673632, + "step": 36585 + }, + { + "epoch": 1.171179822034441, + "grad_norm": 0.6148457527160645, + "learning_rate": 8.917727857739394e-07, + "loss": 0.4378, + "num_input_tokens_seen": 55681056, + "step": 36590 + }, + { + "epoch": 1.1713398630049292, + "grad_norm": 1.1612707376480103, + "learning_rate": 8.891758970159258e-07, + "loss": 0.4529, + "num_input_tokens_seen": 55688640, + "step": 36595 + }, + { + "epoch": 1.1714999039754177, + "grad_norm": 0.2640143930912018, + "learning_rate": 8.86582726449986e-07, + "loss": 0.3546, + "num_input_tokens_seen": 55696592, + "step": 36600 + }, + { + "epoch": 1.1714999039754177, + "eval_loss": 0.49419984221458435, + "eval_runtime": 533.5635, + "eval_samples_per_second": 26.025, + "eval_steps_per_second": 13.013, + "num_input_tokens_seen": 55696592, + "step": 36600 + }, + { + "epoch": 1.1716599449459062, + "grad_norm": 0.9179260730743408, + "learning_rate": 8.839932744760165e-07, + "loss": 0.3669, + "num_input_tokens_seen": 55704336, + "step": 36605 + }, + { + "epoch": 1.1718199859163947, + "grad_norm": 0.6677677035331726, + "learning_rate": 8.814075414933482e-07, + "loss": 0.5723, + "num_input_tokens_seen": 55711856, + "step": 36610 + }, + { + "epoch": 1.171980026886883, + "grad_norm": 0.8416438102722168, + "learning_rate": 8.788255279007257e-07, + "loss": 0.371, + "num_input_tokens_seen": 55719424, + "step": 36615 + }, + { + "epoch": 1.1721400678573715, + "grad_norm": 0.6503205299377441, + "learning_rate": 8.762472340963362e-07, + "loss": 0.4542, + "num_input_tokens_seen": 55726896, + "step": 36620 + }, + { + "epoch": 1.17230010882786, + "grad_norm": 0.4888763725757599, + "learning_rate": 8.736726604777811e-07, + "loss": 0.4831, + "num_input_tokens_seen": 55734080, + "step": 36625 + }, + { + "epoch": 1.1724601497983484, + "grad_norm": 0.6517277956008911, + "learning_rate": 8.711018074420901e-07, + "loss": 0.4719, + "num_input_tokens_seen": 55741584, + "step": 36630 + }, + { + "epoch": 1.172620190768837, + "grad_norm": 1.035988450050354, + "learning_rate": 8.685346753857209e-07, + "loss": 0.4744, + "num_input_tokens_seen": 55749472, + "step": 36635 + }, + { + "epoch": 1.1727802317393252, + "grad_norm": 1.1663810014724731, + "learning_rate": 8.659712647045654e-07, + "loss": 0.3437, + "num_input_tokens_seen": 55756864, + "step": 36640 + }, + { + "epoch": 1.1729402727098137, + "grad_norm": 0.9263721108436584, + "learning_rate": 8.634115757939209e-07, + "loss": 0.4713, + "num_input_tokens_seen": 55764480, + "step": 36645 + }, + { + "epoch": 1.1731003136803022, + "grad_norm": 0.48371192812919617, + "learning_rate": 8.608556090485387e-07, + "loss": 0.5562, + "num_input_tokens_seen": 55771968, + "step": 36650 + }, + { + "epoch": 1.1732603546507905, + "grad_norm": 0.5510023832321167, + "learning_rate": 8.583033648625671e-07, + "loss": 0.5088, + "num_input_tokens_seen": 55779840, + "step": 36655 + }, + { + "epoch": 1.173420395621279, + "grad_norm": 1.2293288707733154, + "learning_rate": 8.557548436295998e-07, + "loss": 0.4872, + "num_input_tokens_seen": 55787200, + "step": 36660 + }, + { + "epoch": 1.1735804365917675, + "grad_norm": 0.8125818371772766, + "learning_rate": 8.532100457426556e-07, + "loss": 0.435, + "num_input_tokens_seen": 55795088, + "step": 36665 + }, + { + "epoch": 1.173740477562256, + "grad_norm": 0.292939692735672, + "learning_rate": 8.506689715941679e-07, + "loss": 0.36, + "num_input_tokens_seen": 55802688, + "step": 36670 + }, + { + "epoch": 1.1739005185327445, + "grad_norm": 0.9023779034614563, + "learning_rate": 8.481316215760011e-07, + "loss": 0.5575, + "num_input_tokens_seen": 55809984, + "step": 36675 + }, + { + "epoch": 1.1740605595032327, + "grad_norm": 0.7107501029968262, + "learning_rate": 8.455979960794558e-07, + "loss": 0.6537, + "num_input_tokens_seen": 55817504, + "step": 36680 + }, + { + "epoch": 1.1742206004737212, + "grad_norm": 0.327676385641098, + "learning_rate": 8.430680954952364e-07, + "loss": 0.4402, + "num_input_tokens_seen": 55824784, + "step": 36685 + }, + { + "epoch": 1.1743806414442097, + "grad_norm": 0.7778353095054626, + "learning_rate": 8.405419202134974e-07, + "loss": 0.3585, + "num_input_tokens_seen": 55832656, + "step": 36690 + }, + { + "epoch": 1.1745406824146982, + "grad_norm": 1.0844368934631348, + "learning_rate": 8.380194706237993e-07, + "loss": 0.5216, + "num_input_tokens_seen": 55839984, + "step": 36695 + }, + { + "epoch": 1.1747007233851865, + "grad_norm": 0.9147434830665588, + "learning_rate": 8.355007471151366e-07, + "loss": 0.6069, + "num_input_tokens_seen": 55847424, + "step": 36700 + }, + { + "epoch": 1.174860764355675, + "grad_norm": 0.8171132206916809, + "learning_rate": 8.329857500759292e-07, + "loss": 0.4727, + "num_input_tokens_seen": 55854592, + "step": 36705 + }, + { + "epoch": 1.1750208053261635, + "grad_norm": 0.3992948830127716, + "learning_rate": 8.304744798940194e-07, + "loss": 0.4923, + "num_input_tokens_seen": 55862544, + "step": 36710 + }, + { + "epoch": 1.175180846296652, + "grad_norm": 0.7008938193321228, + "learning_rate": 8.279669369566756e-07, + "loss": 0.3225, + "num_input_tokens_seen": 55870032, + "step": 36715 + }, + { + "epoch": 1.1753408872671405, + "grad_norm": 0.6786944270133972, + "learning_rate": 8.254631216505993e-07, + "loss": 0.568, + "num_input_tokens_seen": 55878000, + "step": 36720 + }, + { + "epoch": 1.1755009282376288, + "grad_norm": 0.7325639128684998, + "learning_rate": 8.229630343619038e-07, + "loss": 0.5391, + "num_input_tokens_seen": 55885760, + "step": 36725 + }, + { + "epoch": 1.1756609692081172, + "grad_norm": 0.6951453685760498, + "learning_rate": 8.204666754761392e-07, + "loss": 0.4777, + "num_input_tokens_seen": 55893856, + "step": 36730 + }, + { + "epoch": 1.1758210101786057, + "grad_norm": 1.033803939819336, + "learning_rate": 8.179740453782669e-07, + "loss": 0.5663, + "num_input_tokens_seen": 55901520, + "step": 36735 + }, + { + "epoch": 1.1759810511490942, + "grad_norm": 0.6692439913749695, + "learning_rate": 8.154851444526907e-07, + "loss": 0.5298, + "num_input_tokens_seen": 55909296, + "step": 36740 + }, + { + "epoch": 1.1761410921195825, + "grad_norm": 0.4771471917629242, + "learning_rate": 8.129999730832283e-07, + "loss": 0.2794, + "num_input_tokens_seen": 55917376, + "step": 36745 + }, + { + "epoch": 1.176301133090071, + "grad_norm": 1.0904898643493652, + "learning_rate": 8.105185316531178e-07, + "loss": 0.3886, + "num_input_tokens_seen": 55925008, + "step": 36750 + }, + { + "epoch": 1.1764611740605595, + "grad_norm": 0.8837558627128601, + "learning_rate": 8.08040820545039e-07, + "loss": 0.4697, + "num_input_tokens_seen": 55932560, + "step": 36755 + }, + { + "epoch": 1.176621215031048, + "grad_norm": 0.9509218335151672, + "learning_rate": 8.055668401410782e-07, + "loss": 0.623, + "num_input_tokens_seen": 55940032, + "step": 36760 + }, + { + "epoch": 1.1767812560015365, + "grad_norm": 0.7794506549835205, + "learning_rate": 8.030965908227578e-07, + "loss": 0.4809, + "num_input_tokens_seen": 55947856, + "step": 36765 + }, + { + "epoch": 1.1769412969720248, + "grad_norm": 0.6818770170211792, + "learning_rate": 8.006300729710203e-07, + "loss": 0.3968, + "num_input_tokens_seen": 55956064, + "step": 36770 + }, + { + "epoch": 1.1771013379425133, + "grad_norm": 0.5373971462249756, + "learning_rate": 7.981672869662337e-07, + "loss": 0.4119, + "num_input_tokens_seen": 55963856, + "step": 36775 + }, + { + "epoch": 1.1772613789130018, + "grad_norm": 0.7143896222114563, + "learning_rate": 7.957082331881888e-07, + "loss": 0.3878, + "num_input_tokens_seen": 55971536, + "step": 36780 + }, + { + "epoch": 1.1774214198834903, + "grad_norm": 0.8991097807884216, + "learning_rate": 7.932529120161069e-07, + "loss": 0.6714, + "num_input_tokens_seen": 55978848, + "step": 36785 + }, + { + "epoch": 1.1775814608539785, + "grad_norm": 1.4101461172103882, + "learning_rate": 7.908013238286243e-07, + "loss": 0.4176, + "num_input_tokens_seen": 55986272, + "step": 36790 + }, + { + "epoch": 1.177741501824467, + "grad_norm": 1.7912465333938599, + "learning_rate": 7.883534690038136e-07, + "loss": 0.5499, + "num_input_tokens_seen": 55994128, + "step": 36795 + }, + { + "epoch": 1.1779015427949555, + "grad_norm": 0.8096499443054199, + "learning_rate": 7.859093479191559e-07, + "loss": 0.4819, + "num_input_tokens_seen": 56001936, + "step": 36800 + }, + { + "epoch": 1.1779015427949555, + "eval_loss": 0.4942837655544281, + "eval_runtime": 532.6948, + "eval_samples_per_second": 26.067, + "eval_steps_per_second": 13.034, + "num_input_tokens_seen": 56001936, + "step": 36800 + }, + { + "epoch": 1.178061583765444, + "grad_norm": 0.6613757014274597, + "learning_rate": 7.834689609515722e-07, + "loss": 0.4569, + "num_input_tokens_seen": 56009072, + "step": 36805 + }, + { + "epoch": 1.1782216247359325, + "grad_norm": 0.5660246014595032, + "learning_rate": 7.810323084774002e-07, + "loss": 0.3833, + "num_input_tokens_seen": 56016768, + "step": 36810 + }, + { + "epoch": 1.1783816657064208, + "grad_norm": 0.7462621927261353, + "learning_rate": 7.785993908723976e-07, + "loss": 0.5698, + "num_input_tokens_seen": 56024512, + "step": 36815 + }, + { + "epoch": 1.1785417066769093, + "grad_norm": 0.6644504070281982, + "learning_rate": 7.761702085117534e-07, + "loss": 0.4263, + "num_input_tokens_seen": 56032560, + "step": 36820 + }, + { + "epoch": 1.1787017476473978, + "grad_norm": 1.462558627128601, + "learning_rate": 7.737447617700844e-07, + "loss": 0.4034, + "num_input_tokens_seen": 56039424, + "step": 36825 + }, + { + "epoch": 1.1788617886178863, + "grad_norm": 0.7638331055641174, + "learning_rate": 7.713230510214136e-07, + "loss": 0.3843, + "num_input_tokens_seen": 56046928, + "step": 36830 + }, + { + "epoch": 1.1790218295883745, + "grad_norm": 0.547703206539154, + "learning_rate": 7.689050766392092e-07, + "loss": 0.3807, + "num_input_tokens_seen": 56054288, + "step": 36835 + }, + { + "epoch": 1.179181870558863, + "grad_norm": 0.7964333295822144, + "learning_rate": 7.664908389963477e-07, + "loss": 0.4087, + "num_input_tokens_seen": 56061712, + "step": 36840 + }, + { + "epoch": 1.1793419115293515, + "grad_norm": 0.4402085542678833, + "learning_rate": 7.64080338465134e-07, + "loss": 0.6514, + "num_input_tokens_seen": 56069984, + "step": 36845 + }, + { + "epoch": 1.17950195249984, + "grad_norm": 0.7149325013160706, + "learning_rate": 7.616735754173043e-07, + "loss": 0.5358, + "num_input_tokens_seen": 56078048, + "step": 36850 + }, + { + "epoch": 1.1796619934703285, + "grad_norm": 0.48671218752861023, + "learning_rate": 7.592705502240005e-07, + "loss": 0.41, + "num_input_tokens_seen": 56085840, + "step": 36855 + }, + { + "epoch": 1.1798220344408168, + "grad_norm": 0.5660630464553833, + "learning_rate": 7.568712632558095e-07, + "loss": 0.4244, + "num_input_tokens_seen": 56093648, + "step": 36860 + }, + { + "epoch": 1.1799820754113053, + "grad_norm": 0.48938891291618347, + "learning_rate": 7.544757148827297e-07, + "loss": 0.5414, + "num_input_tokens_seen": 56101344, + "step": 36865 + }, + { + "epoch": 1.1801421163817938, + "grad_norm": 0.43574175238609314, + "learning_rate": 7.520839054741797e-07, + "loss": 0.4978, + "num_input_tokens_seen": 56109104, + "step": 36870 + }, + { + "epoch": 1.1803021573522823, + "grad_norm": 0.8381606340408325, + "learning_rate": 7.496958353990113e-07, + "loss": 0.4042, + "num_input_tokens_seen": 56116512, + "step": 36875 + }, + { + "epoch": 1.1804621983227706, + "grad_norm": 1.0341092348098755, + "learning_rate": 7.473115050254941e-07, + "loss": 0.546, + "num_input_tokens_seen": 56124128, + "step": 36880 + }, + { + "epoch": 1.180622239293259, + "grad_norm": 1.424347996711731, + "learning_rate": 7.449309147213173e-07, + "loss": 0.5264, + "num_input_tokens_seen": 56131856, + "step": 36885 + }, + { + "epoch": 1.1807822802637475, + "grad_norm": 0.9184848666191101, + "learning_rate": 7.425540648536067e-07, + "loss": 0.5361, + "num_input_tokens_seen": 56139248, + "step": 36890 + }, + { + "epoch": 1.180942321234236, + "grad_norm": 0.6477321982383728, + "learning_rate": 7.40180955788894e-07, + "loss": 0.4225, + "num_input_tokens_seen": 56146608, + "step": 36895 + }, + { + "epoch": 1.1811023622047245, + "grad_norm": 0.7195053100585938, + "learning_rate": 7.378115878931474e-07, + "loss": 0.4623, + "num_input_tokens_seen": 56154016, + "step": 36900 + }, + { + "epoch": 1.1812624031752128, + "grad_norm": 0.9726659059524536, + "learning_rate": 7.354459615317527e-07, + "loss": 0.5461, + "num_input_tokens_seen": 56161040, + "step": 36905 + }, + { + "epoch": 1.1814224441457013, + "grad_norm": 0.8879455327987671, + "learning_rate": 7.33084077069518e-07, + "loss": 0.5758, + "num_input_tokens_seen": 56168672, + "step": 36910 + }, + { + "epoch": 1.1815824851161898, + "grad_norm": 1.420942783355713, + "learning_rate": 7.307259348706768e-07, + "loss": 0.4712, + "num_input_tokens_seen": 56176752, + "step": 36915 + }, + { + "epoch": 1.181742526086678, + "grad_norm": 0.8952575922012329, + "learning_rate": 7.283715352988801e-07, + "loss": 0.627, + "num_input_tokens_seen": 56185280, + "step": 36920 + }, + { + "epoch": 1.1819025670571666, + "grad_norm": 0.5724635124206543, + "learning_rate": 7.260208787172068e-07, + "loss": 0.4145, + "num_input_tokens_seen": 56192960, + "step": 36925 + }, + { + "epoch": 1.182062608027655, + "grad_norm": 1.2426904439926147, + "learning_rate": 7.23673965488167e-07, + "loss": 0.5095, + "num_input_tokens_seen": 56200720, + "step": 36930 + }, + { + "epoch": 1.1822226489981436, + "grad_norm": 0.43294045329093933, + "learning_rate": 7.213307959736709e-07, + "loss": 0.5642, + "num_input_tokens_seen": 56207872, + "step": 36935 + }, + { + "epoch": 1.182382689968632, + "grad_norm": 0.7593051791191101, + "learning_rate": 7.189913705350715e-07, + "loss": 0.4534, + "num_input_tokens_seen": 56215488, + "step": 36940 + }, + { + "epoch": 1.1825427309391203, + "grad_norm": 0.5888546705245972, + "learning_rate": 7.166556895331411e-07, + "loss": 0.4124, + "num_input_tokens_seen": 56223280, + "step": 36945 + }, + { + "epoch": 1.1827027719096088, + "grad_norm": 0.6167415976524353, + "learning_rate": 7.143237533280639e-07, + "loss": 0.5516, + "num_input_tokens_seen": 56230752, + "step": 36950 + }, + { + "epoch": 1.1828628128800973, + "grad_norm": 0.5512918829917908, + "learning_rate": 7.119955622794578e-07, + "loss": 0.3377, + "num_input_tokens_seen": 56237904, + "step": 36955 + }, + { + "epoch": 1.1830228538505858, + "grad_norm": 0.3568807542324066, + "learning_rate": 7.096711167463577e-07, + "loss": 0.3956, + "num_input_tokens_seen": 56245520, + "step": 36960 + }, + { + "epoch": 1.183182894821074, + "grad_norm": 1.0153710842132568, + "learning_rate": 7.073504170872213e-07, + "loss": 0.5611, + "num_input_tokens_seen": 56253328, + "step": 36965 + }, + { + "epoch": 1.1833429357915626, + "grad_norm": 0.3914947509765625, + "learning_rate": 7.05033463659932e-07, + "loss": 0.5499, + "num_input_tokens_seen": 56260640, + "step": 36970 + }, + { + "epoch": 1.183502976762051, + "grad_norm": 0.5242505669593811, + "learning_rate": 7.027202568217928e-07, + "loss": 0.3036, + "num_input_tokens_seen": 56268208, + "step": 36975 + }, + { + "epoch": 1.1836630177325396, + "grad_norm": 1.2899599075317383, + "learning_rate": 7.004107969295293e-07, + "loss": 0.4296, + "num_input_tokens_seen": 56276000, + "step": 36980 + }, + { + "epoch": 1.183823058703028, + "grad_norm": 0.5056478977203369, + "learning_rate": 6.9810508433929e-07, + "loss": 0.5346, + "num_input_tokens_seen": 56283984, + "step": 36985 + }, + { + "epoch": 1.1839830996735163, + "grad_norm": 0.6287028193473816, + "learning_rate": 6.958031194066406e-07, + "loss": 0.4187, + "num_input_tokens_seen": 56291920, + "step": 36990 + }, + { + "epoch": 1.1841431406440048, + "grad_norm": 0.8150428533554077, + "learning_rate": 6.935049024865776e-07, + "loss": 0.6703, + "num_input_tokens_seen": 56299344, + "step": 36995 + }, + { + "epoch": 1.1843031816144933, + "grad_norm": 0.9142974019050598, + "learning_rate": 6.912104339335118e-07, + "loss": 0.5585, + "num_input_tokens_seen": 56306928, + "step": 37000 + }, + { + "epoch": 1.1843031816144933, + "eval_loss": 0.4942500591278076, + "eval_runtime": 533.5517, + "eval_samples_per_second": 26.026, + "eval_steps_per_second": 13.013, + "num_input_tokens_seen": 56306928, + "step": 37000 + }, + { + "epoch": 1.1844632225849818, + "grad_norm": 0.5712901949882507, + "learning_rate": 6.889197141012799e-07, + "loss": 0.4427, + "num_input_tokens_seen": 56315232, + "step": 37005 + }, + { + "epoch": 1.18462326355547, + "grad_norm": 1.1630867719650269, + "learning_rate": 6.866327433431435e-07, + "loss": 0.7348, + "num_input_tokens_seen": 56323408, + "step": 37010 + }, + { + "epoch": 1.1847833045259586, + "grad_norm": 0.9625409841537476, + "learning_rate": 6.843495220117735e-07, + "loss": 0.4503, + "num_input_tokens_seen": 56331072, + "step": 37015 + }, + { + "epoch": 1.184943345496447, + "grad_norm": 1.0245192050933838, + "learning_rate": 6.820700504592798e-07, + "loss": 0.4974, + "num_input_tokens_seen": 56338640, + "step": 37020 + }, + { + "epoch": 1.1851033864669356, + "grad_norm": 1.629072904586792, + "learning_rate": 6.797943290371839e-07, + "loss": 0.3805, + "num_input_tokens_seen": 56346032, + "step": 37025 + }, + { + "epoch": 1.185263427437424, + "grad_norm": 0.8374184966087341, + "learning_rate": 6.775223580964274e-07, + "loss": 0.6054, + "num_input_tokens_seen": 56353888, + "step": 37030 + }, + { + "epoch": 1.1854234684079124, + "grad_norm": 0.7149990797042847, + "learning_rate": 6.7525413798738e-07, + "loss": 0.4118, + "num_input_tokens_seen": 56361664, + "step": 37035 + }, + { + "epoch": 1.1855835093784008, + "grad_norm": 0.5763672590255737, + "learning_rate": 6.729896690598259e-07, + "loss": 0.3987, + "num_input_tokens_seen": 56369120, + "step": 37040 + }, + { + "epoch": 1.1857435503488893, + "grad_norm": 0.5811710357666016, + "learning_rate": 6.707289516629772e-07, + "loss": 0.4069, + "num_input_tokens_seen": 56376336, + "step": 37045 + }, + { + "epoch": 1.1859035913193778, + "grad_norm": 0.6975969076156616, + "learning_rate": 6.684719861454692e-07, + "loss": 0.444, + "num_input_tokens_seen": 56383632, + "step": 37050 + }, + { + "epoch": 1.186063632289866, + "grad_norm": 1.6010279655456543, + "learning_rate": 6.662187728553481e-07, + "loss": 0.523, + "num_input_tokens_seen": 56391088, + "step": 37055 + }, + { + "epoch": 1.1862236732603546, + "grad_norm": 0.6010175943374634, + "learning_rate": 6.639693121400892e-07, + "loss": 0.6191, + "num_input_tokens_seen": 56398496, + "step": 37060 + }, + { + "epoch": 1.186383714230843, + "grad_norm": 0.6089956164360046, + "learning_rate": 6.617236043465868e-07, + "loss": 0.3652, + "num_input_tokens_seen": 56406192, + "step": 37065 + }, + { + "epoch": 1.1865437552013316, + "grad_norm": 0.39303499460220337, + "learning_rate": 6.594816498211587e-07, + "loss": 0.5874, + "num_input_tokens_seen": 56413584, + "step": 37070 + }, + { + "epoch": 1.18670379617182, + "grad_norm": 0.6531332731246948, + "learning_rate": 6.572434489095447e-07, + "loss": 0.5735, + "num_input_tokens_seen": 56421088, + "step": 37075 + }, + { + "epoch": 1.1868638371423084, + "grad_norm": 0.8611182570457458, + "learning_rate": 6.550090019568994e-07, + "loss": 0.4512, + "num_input_tokens_seen": 56428720, + "step": 37080 + }, + { + "epoch": 1.1870238781127969, + "grad_norm": 1.3653086423873901, + "learning_rate": 6.527783093078027e-07, + "loss": 0.4412, + "num_input_tokens_seen": 56437424, + "step": 37085 + }, + { + "epoch": 1.1871839190832854, + "grad_norm": 0.5966901779174805, + "learning_rate": 6.5055137130626e-07, + "loss": 0.4209, + "num_input_tokens_seen": 56445024, + "step": 37090 + }, + { + "epoch": 1.1873439600537739, + "grad_norm": 0.48019731044769287, + "learning_rate": 6.483281882956854e-07, + "loss": 0.4591, + "num_input_tokens_seen": 56452384, + "step": 37095 + }, + { + "epoch": 1.1875040010242621, + "grad_norm": 1.1129798889160156, + "learning_rate": 6.461087606189298e-07, + "loss": 0.4649, + "num_input_tokens_seen": 56460400, + "step": 37100 + }, + { + "epoch": 1.1876640419947506, + "grad_norm": 1.3234578371047974, + "learning_rate": 6.438930886182554e-07, + "loss": 0.5174, + "num_input_tokens_seen": 56468096, + "step": 37105 + }, + { + "epoch": 1.1878240829652391, + "grad_norm": 1.0124303102493286, + "learning_rate": 6.416811726353417e-07, + "loss": 0.3536, + "num_input_tokens_seen": 56475424, + "step": 37110 + }, + { + "epoch": 1.1879841239357276, + "grad_norm": 0.6710793972015381, + "learning_rate": 6.394730130112991e-07, + "loss": 0.5187, + "num_input_tokens_seen": 56483152, + "step": 37115 + }, + { + "epoch": 1.188144164906216, + "grad_norm": 1.3779584169387817, + "learning_rate": 6.372686100866471e-07, + "loss": 0.3815, + "num_input_tokens_seen": 56490672, + "step": 37120 + }, + { + "epoch": 1.1883042058767044, + "grad_norm": 0.6445643305778503, + "learning_rate": 6.350679642013413e-07, + "loss": 0.4132, + "num_input_tokens_seen": 56498160, + "step": 37125 + }, + { + "epoch": 1.1884642468471929, + "grad_norm": 1.1185872554779053, + "learning_rate": 6.328710756947437e-07, + "loss": 0.5164, + "num_input_tokens_seen": 56505664, + "step": 37130 + }, + { + "epoch": 1.1886242878176814, + "grad_norm": 1.092877984046936, + "learning_rate": 6.306779449056416e-07, + "loss": 0.5629, + "num_input_tokens_seen": 56513968, + "step": 37135 + }, + { + "epoch": 1.1887843287881699, + "grad_norm": 0.5654481053352356, + "learning_rate": 6.284885721722422e-07, + "loss": 0.4285, + "num_input_tokens_seen": 56521776, + "step": 37140 + }, + { + "epoch": 1.1889443697586581, + "grad_norm": 0.9617580771446228, + "learning_rate": 6.26302957832181e-07, + "loss": 0.4315, + "num_input_tokens_seen": 56529120, + "step": 37145 + }, + { + "epoch": 1.1891044107291466, + "grad_norm": 0.959706723690033, + "learning_rate": 6.241211022224997e-07, + "loss": 0.4948, + "num_input_tokens_seen": 56536320, + "step": 37150 + }, + { + "epoch": 1.1892644516996351, + "grad_norm": 0.7599544525146484, + "learning_rate": 6.219430056796732e-07, + "loss": 0.4477, + "num_input_tokens_seen": 56543440, + "step": 37155 + }, + { + "epoch": 1.1894244926701236, + "grad_norm": 0.7525361180305481, + "learning_rate": 6.19768668539586e-07, + "loss": 0.3801, + "num_input_tokens_seen": 56551136, + "step": 37160 + }, + { + "epoch": 1.1895845336406121, + "grad_norm": 0.5286897420883179, + "learning_rate": 6.175980911375528e-07, + "loss": 0.2818, + "num_input_tokens_seen": 56558832, + "step": 37165 + }, + { + "epoch": 1.1897445746111004, + "grad_norm": 0.6891841888427734, + "learning_rate": 6.154312738083034e-07, + "loss": 0.3357, + "num_input_tokens_seen": 56566416, + "step": 37170 + }, + { + "epoch": 1.1899046155815889, + "grad_norm": 0.48624059557914734, + "learning_rate": 6.132682168859843e-07, + "loss": 0.3182, + "num_input_tokens_seen": 56574064, + "step": 37175 + }, + { + "epoch": 1.1900646565520774, + "grad_norm": 0.5026324987411499, + "learning_rate": 6.111089207041704e-07, + "loss": 0.3535, + "num_input_tokens_seen": 56581904, + "step": 37180 + }, + { + "epoch": 1.1902246975225657, + "grad_norm": 1.150615930557251, + "learning_rate": 6.089533855958507e-07, + "loss": 0.5932, + "num_input_tokens_seen": 56589904, + "step": 37185 + }, + { + "epoch": 1.1903847384930542, + "grad_norm": 0.5666322112083435, + "learning_rate": 6.068016118934372e-07, + "loss": 0.633, + "num_input_tokens_seen": 56597600, + "step": 37190 + }, + { + "epoch": 1.1905447794635426, + "grad_norm": 0.38766857981681824, + "learning_rate": 6.04653599928759e-07, + "loss": 0.3388, + "num_input_tokens_seen": 56605408, + "step": 37195 + }, + { + "epoch": 1.1907048204340311, + "grad_norm": 0.7330759167671204, + "learning_rate": 6.025093500330675e-07, + "loss": 0.5079, + "num_input_tokens_seen": 56613088, + "step": 37200 + }, + { + "epoch": 1.1907048204340311, + "eval_loss": 0.4942490756511688, + "eval_runtime": 532.8709, + "eval_samples_per_second": 26.059, + "eval_steps_per_second": 13.029, + "num_input_tokens_seen": 56613088, + "step": 37200 + }, + { + "epoch": 1.1908648614045196, + "grad_norm": 1.013489842414856, + "learning_rate": 6.003688625370291e-07, + "loss": 0.4617, + "num_input_tokens_seen": 56620864, + "step": 37205 + }, + { + "epoch": 1.191024902375008, + "grad_norm": 0.6945508718490601, + "learning_rate": 5.982321377707406e-07, + "loss": 0.3077, + "num_input_tokens_seen": 56627984, + "step": 37210 + }, + { + "epoch": 1.1911849433454964, + "grad_norm": 0.4839029908180237, + "learning_rate": 5.96099176063708e-07, + "loss": 0.3807, + "num_input_tokens_seen": 56635632, + "step": 37215 + }, + { + "epoch": 1.191344984315985, + "grad_norm": 0.7349628210067749, + "learning_rate": 5.93969977744857e-07, + "loss": 0.6262, + "num_input_tokens_seen": 56643024, + "step": 37220 + }, + { + "epoch": 1.1915050252864734, + "grad_norm": 1.0847084522247314, + "learning_rate": 5.918445431425445e-07, + "loss": 0.4558, + "num_input_tokens_seen": 56650976, + "step": 37225 + }, + { + "epoch": 1.1916650662569617, + "grad_norm": 0.6623372435569763, + "learning_rate": 5.897228725845333e-07, + "loss": 0.4722, + "num_input_tokens_seen": 56658256, + "step": 37230 + }, + { + "epoch": 1.1918251072274502, + "grad_norm": 0.6627776622772217, + "learning_rate": 5.876049663980171e-07, + "loss": 0.3819, + "num_input_tokens_seen": 56665824, + "step": 37235 + }, + { + "epoch": 1.1919851481979387, + "grad_norm": 0.4933643341064453, + "learning_rate": 5.854908249095959e-07, + "loss": 0.351, + "num_input_tokens_seen": 56673200, + "step": 37240 + }, + { + "epoch": 1.1921451891684272, + "grad_norm": 0.46217235922813416, + "learning_rate": 5.833804484453031e-07, + "loss": 0.4753, + "num_input_tokens_seen": 56680880, + "step": 37245 + }, + { + "epoch": 1.1923052301389157, + "grad_norm": 0.8481001853942871, + "learning_rate": 5.81273837330587e-07, + "loss": 0.4972, + "num_input_tokens_seen": 56688784, + "step": 37250 + }, + { + "epoch": 1.192465271109404, + "grad_norm": 0.8250762820243835, + "learning_rate": 5.791709918903071e-07, + "loss": 0.3467, + "num_input_tokens_seen": 56696672, + "step": 37255 + }, + { + "epoch": 1.1926253120798924, + "grad_norm": 0.5813165307044983, + "learning_rate": 5.770719124487483e-07, + "loss": 0.7423, + "num_input_tokens_seen": 56703952, + "step": 37260 + }, + { + "epoch": 1.192785353050381, + "grad_norm": 0.9194161295890808, + "learning_rate": 5.749765993296241e-07, + "loss": 0.2911, + "num_input_tokens_seen": 56711296, + "step": 37265 + }, + { + "epoch": 1.1929453940208694, + "grad_norm": 1.2219247817993164, + "learning_rate": 5.728850528560509e-07, + "loss": 0.5125, + "num_input_tokens_seen": 56718368, + "step": 37270 + }, + { + "epoch": 1.1931054349913577, + "grad_norm": 0.7847950458526611, + "learning_rate": 5.707972733505707e-07, + "loss": 0.376, + "num_input_tokens_seen": 56725760, + "step": 37275 + }, + { + "epoch": 1.1932654759618462, + "grad_norm": 1.0156950950622559, + "learning_rate": 5.687132611351509e-07, + "loss": 0.4106, + "num_input_tokens_seen": 56733792, + "step": 37280 + }, + { + "epoch": 1.1934255169323347, + "grad_norm": 1.2065123319625854, + "learning_rate": 5.666330165311651e-07, + "loss": 0.4632, + "num_input_tokens_seen": 56741152, + "step": 37285 + }, + { + "epoch": 1.1935855579028232, + "grad_norm": 0.8257918953895569, + "learning_rate": 5.645565398594204e-07, + "loss": 0.4795, + "num_input_tokens_seen": 56748848, + "step": 37290 + }, + { + "epoch": 1.1937455988733117, + "grad_norm": 0.4779616594314575, + "learning_rate": 5.624838314401304e-07, + "loss": 0.3534, + "num_input_tokens_seen": 56756336, + "step": 37295 + }, + { + "epoch": 1.1939056398438, + "grad_norm": 0.5053879618644714, + "learning_rate": 5.604148915929336e-07, + "loss": 0.4228, + "num_input_tokens_seen": 56764512, + "step": 37300 + }, + { + "epoch": 1.1940656808142884, + "grad_norm": 0.7958428859710693, + "learning_rate": 5.583497206368887e-07, + "loss": 0.398, + "num_input_tokens_seen": 56771952, + "step": 37305 + }, + { + "epoch": 1.194225721784777, + "grad_norm": 1.0502934455871582, + "learning_rate": 5.562883188904688e-07, + "loss": 0.3944, + "num_input_tokens_seen": 56779504, + "step": 37310 + }, + { + "epoch": 1.1943857627552654, + "grad_norm": 0.712655782699585, + "learning_rate": 5.542306866715724e-07, + "loss": 0.5197, + "num_input_tokens_seen": 56787120, + "step": 37315 + }, + { + "epoch": 1.1945458037257537, + "grad_norm": 1.1179749965667725, + "learning_rate": 5.52176824297504e-07, + "loss": 0.4279, + "num_input_tokens_seen": 56795232, + "step": 37320 + }, + { + "epoch": 1.1947058446962422, + "grad_norm": 0.5002945065498352, + "learning_rate": 5.501267320850018e-07, + "loss": 0.505, + "num_input_tokens_seen": 56803168, + "step": 37325 + }, + { + "epoch": 1.1948658856667307, + "grad_norm": 0.5174474120140076, + "learning_rate": 5.480804103502157e-07, + "loss": 0.3816, + "num_input_tokens_seen": 56810480, + "step": 37330 + }, + { + "epoch": 1.1950259266372192, + "grad_norm": 0.815180242061615, + "learning_rate": 5.460378594087101e-07, + "loss": 0.4769, + "num_input_tokens_seen": 56818272, + "step": 37335 + }, + { + "epoch": 1.1951859676077077, + "grad_norm": 0.5225335359573364, + "learning_rate": 5.439990795754773e-07, + "loss": 0.4328, + "num_input_tokens_seen": 56825984, + "step": 37340 + }, + { + "epoch": 1.195346008578196, + "grad_norm": 0.4772447347640991, + "learning_rate": 5.419640711649188e-07, + "loss": 0.5358, + "num_input_tokens_seen": 56833920, + "step": 37345 + }, + { + "epoch": 1.1955060495486844, + "grad_norm": 1.2748161554336548, + "learning_rate": 5.399328344908583e-07, + "loss": 0.4813, + "num_input_tokens_seen": 56841792, + "step": 37350 + }, + { + "epoch": 1.195666090519173, + "grad_norm": 0.7918568849563599, + "learning_rate": 5.379053698665399e-07, + "loss": 0.5619, + "num_input_tokens_seen": 56849216, + "step": 37355 + }, + { + "epoch": 1.1958261314896614, + "grad_norm": 0.4694873094558716, + "learning_rate": 5.358816776046216e-07, + "loss": 0.6886, + "num_input_tokens_seen": 56856512, + "step": 37360 + }, + { + "epoch": 1.1959861724601497, + "grad_norm": 0.6336987614631653, + "learning_rate": 5.338617580171817e-07, + "loss": 0.4344, + "num_input_tokens_seen": 56864528, + "step": 37365 + }, + { + "epoch": 1.1961462134306382, + "grad_norm": 0.9032036066055298, + "learning_rate": 5.318456114157239e-07, + "loss": 0.3943, + "num_input_tokens_seen": 56872304, + "step": 37370 + }, + { + "epoch": 1.1963062544011267, + "grad_norm": 0.36835312843322754, + "learning_rate": 5.298332381111576e-07, + "loss": 0.2888, + "num_input_tokens_seen": 56879744, + "step": 37375 + }, + { + "epoch": 1.1964662953716152, + "grad_norm": 0.45319098234176636, + "learning_rate": 5.27824638413818e-07, + "loss": 0.5268, + "num_input_tokens_seen": 56887376, + "step": 37380 + }, + { + "epoch": 1.1966263363421037, + "grad_norm": 0.8480486869812012, + "learning_rate": 5.258198126334546e-07, + "loss": 0.4169, + "num_input_tokens_seen": 56894992, + "step": 37385 + }, + { + "epoch": 1.196786377312592, + "grad_norm": 0.9281548857688904, + "learning_rate": 5.238187610792367e-07, + "loss": 0.509, + "num_input_tokens_seen": 56902368, + "step": 37390 + }, + { + "epoch": 1.1969464182830805, + "grad_norm": 0.506151020526886, + "learning_rate": 5.218214840597563e-07, + "loss": 0.4276, + "num_input_tokens_seen": 56909888, + "step": 37395 + }, + { + "epoch": 1.197106459253569, + "grad_norm": 0.5681101679801941, + "learning_rate": 5.198279818830115e-07, + "loss": 0.6697, + "num_input_tokens_seen": 56917104, + "step": 37400 + }, + { + "epoch": 1.197106459253569, + "eval_loss": 0.49432340264320374, + "eval_runtime": 533.4408, + "eval_samples_per_second": 26.031, + "eval_steps_per_second": 13.016, + "num_input_tokens_seen": 56917104, + "step": 37400 + }, + { + "epoch": 1.1972665002240575, + "grad_norm": 0.6748425960540771, + "learning_rate": 5.178382548564287e-07, + "loss": 0.5121, + "num_input_tokens_seen": 56924592, + "step": 37405 + }, + { + "epoch": 1.1974265411945457, + "grad_norm": 0.4757738411426544, + "learning_rate": 5.15852303286854e-07, + "loss": 0.476, + "num_input_tokens_seen": 56932064, + "step": 37410 + }, + { + "epoch": 1.1975865821650342, + "grad_norm": 0.8286420106887817, + "learning_rate": 5.138701274805396e-07, + "loss": 0.4816, + "num_input_tokens_seen": 56939728, + "step": 37415 + }, + { + "epoch": 1.1977466231355227, + "grad_norm": 0.7762681245803833, + "learning_rate": 5.118917277431606e-07, + "loss": 0.3873, + "num_input_tokens_seen": 56947408, + "step": 37420 + }, + { + "epoch": 1.1979066641060112, + "grad_norm": 0.533150851726532, + "learning_rate": 5.099171043798145e-07, + "loss": 0.3561, + "num_input_tokens_seen": 56954912, + "step": 37425 + }, + { + "epoch": 1.1980667050764997, + "grad_norm": 1.2672967910766602, + "learning_rate": 5.079462576950133e-07, + "loss": 0.5178, + "num_input_tokens_seen": 56963120, + "step": 37430 + }, + { + "epoch": 1.198226746046988, + "grad_norm": 0.7996987700462341, + "learning_rate": 5.059791879926862e-07, + "loss": 0.3987, + "num_input_tokens_seen": 56970144, + "step": 37435 + }, + { + "epoch": 1.1983867870174765, + "grad_norm": 0.3591849207878113, + "learning_rate": 5.040158955761793e-07, + "loss": 0.3536, + "num_input_tokens_seen": 56977280, + "step": 37440 + }, + { + "epoch": 1.198546827987965, + "grad_norm": 0.3297499418258667, + "learning_rate": 5.020563807482559e-07, + "loss": 0.3919, + "num_input_tokens_seen": 56985248, + "step": 37445 + }, + { + "epoch": 1.1987068689584532, + "grad_norm": 0.5700517296791077, + "learning_rate": 5.001006438110995e-07, + "loss": 0.4389, + "num_input_tokens_seen": 56992416, + "step": 37450 + }, + { + "epoch": 1.1988669099289417, + "grad_norm": 0.4988510310649872, + "learning_rate": 4.981486850663075e-07, + "loss": 0.4008, + "num_input_tokens_seen": 57000208, + "step": 37455 + }, + { + "epoch": 1.1990269508994302, + "grad_norm": 0.33259594440460205, + "learning_rate": 4.962005048149005e-07, + "loss": 0.4428, + "num_input_tokens_seen": 57007776, + "step": 37460 + }, + { + "epoch": 1.1991869918699187, + "grad_norm": 0.7330555319786072, + "learning_rate": 4.942561033573073e-07, + "loss": 0.4402, + "num_input_tokens_seen": 57015344, + "step": 37465 + }, + { + "epoch": 1.1993470328404072, + "grad_norm": 1.3541673421859741, + "learning_rate": 4.923154809933827e-07, + "loss": 0.5187, + "num_input_tokens_seen": 57023168, + "step": 37470 + }, + { + "epoch": 1.1995070738108955, + "grad_norm": 1.252076268196106, + "learning_rate": 4.903786380223957e-07, + "loss": 0.4501, + "num_input_tokens_seen": 57031152, + "step": 37475 + }, + { + "epoch": 1.199667114781384, + "grad_norm": 0.38923004269599915, + "learning_rate": 4.884455747430266e-07, + "loss": 0.5881, + "num_input_tokens_seen": 57039088, + "step": 37480 + }, + { + "epoch": 1.1998271557518725, + "grad_norm": 0.4655034840106964, + "learning_rate": 4.865162914533816e-07, + "loss": 0.4825, + "num_input_tokens_seen": 57046336, + "step": 37485 + }, + { + "epoch": 1.199987196722361, + "grad_norm": 0.6779526472091675, + "learning_rate": 4.845907884509809e-07, + "loss": 0.5752, + "num_input_tokens_seen": 57054192, + "step": 37490 + }, + { + "epoch": 1.2001472376928493, + "grad_norm": 1.35663902759552, + "learning_rate": 4.82669066032762e-07, + "loss": 0.4997, + "num_input_tokens_seen": 57061312, + "step": 37495 + }, + { + "epoch": 1.2003072786633378, + "grad_norm": 0.4398830831050873, + "learning_rate": 4.807511244950768e-07, + "loss": 0.3635, + "num_input_tokens_seen": 57068736, + "step": 37500 + }, + { + "epoch": 1.2004673196338262, + "grad_norm": 0.45819148421287537, + "learning_rate": 4.788369641336943e-07, + "loss": 0.3624, + "num_input_tokens_seen": 57076816, + "step": 37505 + }, + { + "epoch": 1.2006273606043147, + "grad_norm": 0.7241436839103699, + "learning_rate": 4.769265852438032e-07, + "loss": 0.4919, + "num_input_tokens_seen": 57084368, + "step": 37510 + }, + { + "epoch": 1.2007874015748032, + "grad_norm": 0.3442750573158264, + "learning_rate": 4.750199881200124e-07, + "loss": 0.4452, + "num_input_tokens_seen": 57091872, + "step": 37515 + }, + { + "epoch": 1.2009474425452915, + "grad_norm": 0.7113109827041626, + "learning_rate": 4.7311717305633664e-07, + "loss": 0.5425, + "num_input_tokens_seen": 57099232, + "step": 37520 + }, + { + "epoch": 1.20110748351578, + "grad_norm": 1.0096521377563477, + "learning_rate": 4.7121814034621623e-07, + "loss": 0.6019, + "num_input_tokens_seen": 57106960, + "step": 37525 + }, + { + "epoch": 1.2012675244862685, + "grad_norm": 0.5085150599479675, + "learning_rate": 4.693228902825114e-07, + "loss": 0.4032, + "num_input_tokens_seen": 57114256, + "step": 37530 + }, + { + "epoch": 1.201427565456757, + "grad_norm": 0.6976354122161865, + "learning_rate": 4.6743142315748277e-07, + "loss": 0.3875, + "num_input_tokens_seen": 57121648, + "step": 37535 + }, + { + "epoch": 1.2015876064272453, + "grad_norm": 0.9274587631225586, + "learning_rate": 4.655437392628276e-07, + "loss": 0.6187, + "num_input_tokens_seen": 57129008, + "step": 37540 + }, + { + "epoch": 1.2017476473977338, + "grad_norm": 0.7635319232940674, + "learning_rate": 4.636598388896463e-07, + "loss": 0.6445, + "num_input_tokens_seen": 57136576, + "step": 37545 + }, + { + "epoch": 1.2019076883682223, + "grad_norm": 1.074972152709961, + "learning_rate": 4.6177972232845925e-07, + "loss": 0.4818, + "num_input_tokens_seen": 57144032, + "step": 37550 + }, + { + "epoch": 1.2020677293387108, + "grad_norm": 1.0323448181152344, + "learning_rate": 4.5990338986920953e-07, + "loss": 0.5341, + "num_input_tokens_seen": 57152400, + "step": 37555 + }, + { + "epoch": 1.2022277703091993, + "grad_norm": 1.0938262939453125, + "learning_rate": 4.5803084180124633e-07, + "loss": 0.4449, + "num_input_tokens_seen": 57160336, + "step": 37560 + }, + { + "epoch": 1.2023878112796875, + "grad_norm": 0.7381056547164917, + "learning_rate": 4.561620784133386e-07, + "loss": 0.4479, + "num_input_tokens_seen": 57167824, + "step": 37565 + }, + { + "epoch": 1.202547852250176, + "grad_norm": 0.7210086584091187, + "learning_rate": 4.5429709999367796e-07, + "loss": 0.3844, + "num_input_tokens_seen": 57175392, + "step": 37570 + }, + { + "epoch": 1.2027078932206645, + "grad_norm": 0.8005821108818054, + "learning_rate": 4.5243590682986223e-07, + "loss": 0.3578, + "num_input_tokens_seen": 57183040, + "step": 37575 + }, + { + "epoch": 1.202867934191153, + "grad_norm": 1.1155116558074951, + "learning_rate": 4.5057849920891735e-07, + "loss": 0.497, + "num_input_tokens_seen": 57190848, + "step": 37580 + }, + { + "epoch": 1.2030279751616413, + "grad_norm": 1.3421289920806885, + "learning_rate": 4.487248774172698e-07, + "loss": 0.4348, + "num_input_tokens_seen": 57198640, + "step": 37585 + }, + { + "epoch": 1.2031880161321298, + "grad_norm": 0.9098662734031677, + "learning_rate": 4.4687504174077965e-07, + "loss": 0.714, + "num_input_tokens_seen": 57210048, + "step": 37590 + }, + { + "epoch": 1.2033480571026183, + "grad_norm": 1.3382498025894165, + "learning_rate": 4.450289924647133e-07, + "loss": 0.4279, + "num_input_tokens_seen": 57217792, + "step": 37595 + }, + { + "epoch": 1.2035080980731068, + "grad_norm": 1.0817373991012573, + "learning_rate": 4.431867298737513e-07, + "loss": 0.5126, + "num_input_tokens_seen": 57225504, + "step": 37600 + }, + { + "epoch": 1.2035080980731068, + "eval_loss": 0.49425262212753296, + "eval_runtime": 539.1095, + "eval_samples_per_second": 25.757, + "eval_steps_per_second": 12.879, + "num_input_tokens_seen": 57225504, + "step": 37600 + }, + { + "epoch": 1.2036681390435953, + "grad_norm": 0.6505058407783508, + "learning_rate": 4.41348254251997e-07, + "loss": 0.4521, + "num_input_tokens_seen": 57232544, + "step": 37605 + }, + { + "epoch": 1.2038281800140835, + "grad_norm": 0.7405853867530823, + "learning_rate": 4.395135658829652e-07, + "loss": 0.4526, + "num_input_tokens_seen": 57240016, + "step": 37610 + }, + { + "epoch": 1.203988220984572, + "grad_norm": 0.4867990016937256, + "learning_rate": 4.376826650495852e-07, + "loss": 0.6296, + "num_input_tokens_seen": 57247504, + "step": 37615 + }, + { + "epoch": 1.2041482619550605, + "grad_norm": 0.7381411790847778, + "learning_rate": 4.358555520342117e-07, + "loss": 0.5351, + "num_input_tokens_seen": 57254576, + "step": 37620 + }, + { + "epoch": 1.204308302925549, + "grad_norm": 0.8983991742134094, + "learning_rate": 4.3403222711860257e-07, + "loss": 0.489, + "num_input_tokens_seen": 57261808, + "step": 37625 + }, + { + "epoch": 1.2044683438960373, + "grad_norm": 0.7522284984588623, + "learning_rate": 4.3221269058394133e-07, + "loss": 0.6482, + "num_input_tokens_seen": 57270096, + "step": 37630 + }, + { + "epoch": 1.2046283848665258, + "grad_norm": 0.9748446941375732, + "learning_rate": 4.303969427108173e-07, + "loss": 0.4253, + "num_input_tokens_seen": 57278176, + "step": 37635 + }, + { + "epoch": 1.2047884258370143, + "grad_norm": 1.0331884622573853, + "learning_rate": 4.2858498377924825e-07, + "loss": 0.5958, + "num_input_tokens_seen": 57285760, + "step": 37640 + }, + { + "epoch": 1.2049484668075028, + "grad_norm": 0.23655186593532562, + "learning_rate": 4.267768140686579e-07, + "loss": 0.3475, + "num_input_tokens_seen": 57293328, + "step": 37645 + }, + { + "epoch": 1.2051085077779913, + "grad_norm": 0.4534855782985687, + "learning_rate": 4.2497243385788975e-07, + "loss": 0.4027, + "num_input_tokens_seen": 57301584, + "step": 37650 + }, + { + "epoch": 1.2052685487484796, + "grad_norm": 0.5454785823822021, + "learning_rate": 4.231718434251991e-07, + "loss": 0.4247, + "num_input_tokens_seen": 57308912, + "step": 37655 + }, + { + "epoch": 1.205428589718968, + "grad_norm": 1.233317494392395, + "learning_rate": 4.213750430482666e-07, + "loss": 0.6192, + "num_input_tokens_seen": 57316416, + "step": 37660 + }, + { + "epoch": 1.2055886306894565, + "grad_norm": 0.6720459461212158, + "learning_rate": 4.1958203300417054e-07, + "loss": 0.5305, + "num_input_tokens_seen": 57323840, + "step": 37665 + }, + { + "epoch": 1.205748671659945, + "grad_norm": 0.7309826612472534, + "learning_rate": 4.177928135694259e-07, + "loss": 0.5227, + "num_input_tokens_seen": 57331472, + "step": 37670 + }, + { + "epoch": 1.2059087126304333, + "grad_norm": 0.5558325052261353, + "learning_rate": 4.1600738501994807e-07, + "loss": 0.4363, + "num_input_tokens_seen": 57339088, + "step": 37675 + }, + { + "epoch": 1.2060687536009218, + "grad_norm": 0.6472665071487427, + "learning_rate": 4.1422574763107237e-07, + "loss": 0.5631, + "num_input_tokens_seen": 57347712, + "step": 37680 + }, + { + "epoch": 1.2062287945714103, + "grad_norm": 1.0195790529251099, + "learning_rate": 4.124479016775512e-07, + "loss": 0.4139, + "num_input_tokens_seen": 57355104, + "step": 37685 + }, + { + "epoch": 1.2063888355418988, + "grad_norm": 0.9092127680778503, + "learning_rate": 4.106738474335514e-07, + "loss": 0.5493, + "num_input_tokens_seen": 57362752, + "step": 37690 + }, + { + "epoch": 1.2065488765123873, + "grad_norm": 0.9552814960479736, + "learning_rate": 4.089035851726486e-07, + "loss": 0.4333, + "num_input_tokens_seen": 57370768, + "step": 37695 + }, + { + "epoch": 1.2067089174828756, + "grad_norm": 0.9510065913200378, + "learning_rate": 4.0713711516784937e-07, + "loss": 0.4523, + "num_input_tokens_seen": 57378400, + "step": 37700 + }, + { + "epoch": 1.206868958453364, + "grad_norm": 0.6966593861579895, + "learning_rate": 4.05374437691558e-07, + "loss": 0.3957, + "num_input_tokens_seen": 57385888, + "step": 37705 + }, + { + "epoch": 1.2070289994238526, + "grad_norm": 1.575243592262268, + "learning_rate": 4.036155530156044e-07, + "loss": 0.5063, + "num_input_tokens_seen": 57393440, + "step": 37710 + }, + { + "epoch": 1.2071890403943408, + "grad_norm": 0.6994816660881042, + "learning_rate": 4.018604614112298e-07, + "loss": 0.4741, + "num_input_tokens_seen": 57400608, + "step": 37715 + }, + { + "epoch": 1.2073490813648293, + "grad_norm": 0.6427783370018005, + "learning_rate": 4.0010916314908996e-07, + "loss": 0.3845, + "num_input_tokens_seen": 57408592, + "step": 37720 + }, + { + "epoch": 1.2075091223353178, + "grad_norm": 0.4610055088996887, + "learning_rate": 3.983616584992578e-07, + "loss": 0.5159, + "num_input_tokens_seen": 57416160, + "step": 37725 + }, + { + "epoch": 1.2076691633058063, + "grad_norm": 0.47924038767814636, + "learning_rate": 3.9661794773122595e-07, + "loss": 0.4179, + "num_input_tokens_seen": 57423472, + "step": 37730 + }, + { + "epoch": 1.2078292042762948, + "grad_norm": 0.6788077354431152, + "learning_rate": 3.9487803111388777e-07, + "loss": 0.5057, + "num_input_tokens_seen": 57431280, + "step": 37735 + }, + { + "epoch": 1.207989245246783, + "grad_norm": 1.1368039846420288, + "learning_rate": 3.9314190891556747e-07, + "loss": 0.4064, + "num_input_tokens_seen": 57438768, + "step": 37740 + }, + { + "epoch": 1.2081492862172716, + "grad_norm": 0.4615834355354309, + "learning_rate": 3.914095814039925e-07, + "loss": 0.4339, + "num_input_tokens_seen": 57446320, + "step": 37745 + }, + { + "epoch": 1.20830932718776, + "grad_norm": 0.4975699782371521, + "learning_rate": 3.896810488463104e-07, + "loss": 0.4043, + "num_input_tokens_seen": 57453552, + "step": 37750 + }, + { + "epoch": 1.2084693681582486, + "grad_norm": 0.784580409526825, + "learning_rate": 3.8795631150908565e-07, + "loss": 0.3986, + "num_input_tokens_seen": 57461136, + "step": 37755 + }, + { + "epoch": 1.2086294091287368, + "grad_norm": 1.8225480318069458, + "learning_rate": 3.862353696582888e-07, + "loss": 0.5804, + "num_input_tokens_seen": 57469136, + "step": 37760 + }, + { + "epoch": 1.2087894500992253, + "grad_norm": 1.1559300422668457, + "learning_rate": 3.8451822355931313e-07, + "loss": 0.6598, + "num_input_tokens_seen": 57476448, + "step": 37765 + }, + { + "epoch": 1.2089494910697138, + "grad_norm": 0.5774423480033875, + "learning_rate": 3.82804873476969e-07, + "loss": 0.3835, + "num_input_tokens_seen": 57483680, + "step": 37770 + }, + { + "epoch": 1.2091095320402023, + "grad_norm": 1.4383794069290161, + "learning_rate": 3.810953196754702e-07, + "loss": 0.5542, + "num_input_tokens_seen": 57491312, + "step": 37775 + }, + { + "epoch": 1.2092695730106908, + "grad_norm": 0.9202820062637329, + "learning_rate": 3.793895624184529e-07, + "loss": 0.5028, + "num_input_tokens_seen": 57498672, + "step": 37780 + }, + { + "epoch": 1.209429613981179, + "grad_norm": 1.6338592767715454, + "learning_rate": 3.776876019689679e-07, + "loss": 0.5579, + "num_input_tokens_seen": 57505840, + "step": 37785 + }, + { + "epoch": 1.2095896549516676, + "grad_norm": 1.0163929462432861, + "learning_rate": 3.7598943858947743e-07, + "loss": 0.5559, + "num_input_tokens_seen": 57513456, + "step": 37790 + }, + { + "epoch": 1.209749695922156, + "grad_norm": 0.5057761669158936, + "learning_rate": 3.742950725418637e-07, + "loss": 0.445, + "num_input_tokens_seen": 57521168, + "step": 37795 + }, + { + "epoch": 1.2099097368926446, + "grad_norm": 0.8120536208152771, + "learning_rate": 3.726045040874093e-07, + "loss": 0.4852, + "num_input_tokens_seen": 57529136, + "step": 37800 + }, + { + "epoch": 1.2099097368926446, + "eval_loss": 0.49408820271492004, + "eval_runtime": 540.9339, + "eval_samples_per_second": 25.67, + "eval_steps_per_second": 12.835, + "num_input_tokens_seen": 57529136, + "step": 37800 + }, + { + "epoch": 1.2100697778631329, + "grad_norm": 0.5917366147041321, + "learning_rate": 3.709177334868308e-07, + "loss": 0.6234, + "num_input_tokens_seen": 57536672, + "step": 37805 + }, + { + "epoch": 1.2102298188336214, + "grad_norm": 1.146798849105835, + "learning_rate": 3.692347610002478e-07, + "loss": 0.5347, + "num_input_tokens_seen": 57543920, + "step": 37810 + }, + { + "epoch": 1.2103898598041098, + "grad_norm": 0.9930281043052673, + "learning_rate": 3.675555868871916e-07, + "loss": 0.545, + "num_input_tokens_seen": 57551520, + "step": 37815 + }, + { + "epoch": 1.2105499007745983, + "grad_norm": 0.9542350172996521, + "learning_rate": 3.658802114066162e-07, + "loss": 0.6152, + "num_input_tokens_seen": 57559120, + "step": 37820 + }, + { + "epoch": 1.2107099417450868, + "grad_norm": 0.6293383836746216, + "learning_rate": 3.6420863481688437e-07, + "loss": 0.3297, + "num_input_tokens_seen": 57566320, + "step": 37825 + }, + { + "epoch": 1.210869982715575, + "grad_norm": 0.67756187915802, + "learning_rate": 3.625408573757705e-07, + "loss": 0.4214, + "num_input_tokens_seen": 57573360, + "step": 37830 + }, + { + "epoch": 1.2110300236860636, + "grad_norm": 0.8727766871452332, + "learning_rate": 3.608768793404743e-07, + "loss": 0.6188, + "num_input_tokens_seen": 57581440, + "step": 37835 + }, + { + "epoch": 1.211190064656552, + "grad_norm": 2.6590425968170166, + "learning_rate": 3.592167009675934e-07, + "loss": 0.4863, + "num_input_tokens_seen": 57588848, + "step": 37840 + }, + { + "epoch": 1.2113501056270406, + "grad_norm": 0.5559795498847961, + "learning_rate": 3.575603225131563e-07, + "loss": 0.5861, + "num_input_tokens_seen": 57596640, + "step": 37845 + }, + { + "epoch": 1.2115101465975289, + "grad_norm": 0.8228672742843628, + "learning_rate": 3.55907744232592e-07, + "loss": 0.3684, + "num_input_tokens_seen": 57603856, + "step": 37850 + }, + { + "epoch": 1.2116701875680174, + "grad_norm": 0.6515946388244629, + "learning_rate": 3.5425896638075217e-07, + "loss": 0.3519, + "num_input_tokens_seen": 57611216, + "step": 37855 + }, + { + "epoch": 1.2118302285385059, + "grad_norm": 0.5935708284378052, + "learning_rate": 3.5261398921189736e-07, + "loss": 0.4521, + "num_input_tokens_seen": 57618560, + "step": 37860 + }, + { + "epoch": 1.2119902695089944, + "grad_norm": 1.0832186937332153, + "learning_rate": 3.509728129797024e-07, + "loss": 0.5786, + "num_input_tokens_seen": 57625744, + "step": 37865 + }, + { + "epoch": 1.2121503104794829, + "grad_norm": 0.7986687421798706, + "learning_rate": 3.4933543793725656e-07, + "loss": 0.4713, + "num_input_tokens_seen": 57633056, + "step": 37870 + }, + { + "epoch": 1.2123103514499711, + "grad_norm": 0.9085334539413452, + "learning_rate": 3.4770186433707163e-07, + "loss": 0.3543, + "num_input_tokens_seen": 57640496, + "step": 37875 + }, + { + "epoch": 1.2124703924204596, + "grad_norm": 0.811542809009552, + "learning_rate": 3.4607209243105453e-07, + "loss": 0.5448, + "num_input_tokens_seen": 57648336, + "step": 37880 + }, + { + "epoch": 1.2126304333909481, + "grad_norm": 0.8972316384315491, + "learning_rate": 3.444461224705431e-07, + "loss": 0.3551, + "num_input_tokens_seen": 57655856, + "step": 37885 + }, + { + "epoch": 1.2127904743614366, + "grad_norm": 0.4915730059146881, + "learning_rate": 3.4282395470628116e-07, + "loss": 0.3013, + "num_input_tokens_seen": 57663504, + "step": 37890 + }, + { + "epoch": 1.2129505153319249, + "grad_norm": 0.84966641664505, + "learning_rate": 3.4120558938842417e-07, + "loss": 0.4527, + "num_input_tokens_seen": 57671152, + "step": 37895 + }, + { + "epoch": 1.2131105563024134, + "grad_norm": 0.596085250377655, + "learning_rate": 3.395910267665503e-07, + "loss": 0.3132, + "num_input_tokens_seen": 57678896, + "step": 37900 + }, + { + "epoch": 1.2132705972729019, + "grad_norm": 0.4761301875114441, + "learning_rate": 3.3798026708964094e-07, + "loss": 0.612, + "num_input_tokens_seen": 57686080, + "step": 37905 + }, + { + "epoch": 1.2134306382433904, + "grad_norm": 0.8826370239257812, + "learning_rate": 3.3637331060609456e-07, + "loss": 0.4417, + "num_input_tokens_seen": 57693728, + "step": 37910 + }, + { + "epoch": 1.2135906792138789, + "grad_norm": 0.7909813523292542, + "learning_rate": 3.3477015756372966e-07, + "loss": 0.4439, + "num_input_tokens_seen": 57701712, + "step": 37915 + }, + { + "epoch": 1.2137507201843671, + "grad_norm": 0.6381686925888062, + "learning_rate": 3.3317080820976785e-07, + "loss": 0.3758, + "num_input_tokens_seen": 57709120, + "step": 37920 + }, + { + "epoch": 1.2139107611548556, + "grad_norm": 0.9517966508865356, + "learning_rate": 3.315752627908508e-07, + "loss": 0.4992, + "num_input_tokens_seen": 57716224, + "step": 37925 + }, + { + "epoch": 1.2140708021253441, + "grad_norm": 0.9329580664634705, + "learning_rate": 3.299835215530317e-07, + "loss": 0.6041, + "num_input_tokens_seen": 57723440, + "step": 37930 + }, + { + "epoch": 1.2142308430958326, + "grad_norm": 0.5342710614204407, + "learning_rate": 3.2839558474177245e-07, + "loss": 0.3692, + "num_input_tokens_seen": 57731072, + "step": 37935 + }, + { + "epoch": 1.214390884066321, + "grad_norm": 0.45099830627441406, + "learning_rate": 3.2681145260196056e-07, + "loss": 0.3515, + "num_input_tokens_seen": 57738688, + "step": 37940 + }, + { + "epoch": 1.2145509250368094, + "grad_norm": 0.5670714974403381, + "learning_rate": 3.252311253778839e-07, + "loss": 0.4977, + "num_input_tokens_seen": 57746928, + "step": 37945 + }, + { + "epoch": 1.2147109660072979, + "grad_norm": 0.5691476464271545, + "learning_rate": 3.2365460331325034e-07, + "loss": 0.5625, + "num_input_tokens_seen": 57754592, + "step": 37950 + }, + { + "epoch": 1.2148710069777864, + "grad_norm": 0.8322753310203552, + "learning_rate": 3.2208188665117934e-07, + "loss": 0.5247, + "num_input_tokens_seen": 57762608, + "step": 37955 + }, + { + "epoch": 1.2150310479482749, + "grad_norm": 0.919937252998352, + "learning_rate": 3.205129756342018e-07, + "loss": 0.4028, + "num_input_tokens_seen": 57770352, + "step": 37960 + }, + { + "epoch": 1.2151910889187632, + "grad_norm": 1.5389524698257446, + "learning_rate": 3.189478705042659e-07, + "loss": 0.3426, + "num_input_tokens_seen": 57777856, + "step": 37965 + }, + { + "epoch": 1.2153511298892516, + "grad_norm": 0.5778480768203735, + "learning_rate": 3.173865715027341e-07, + "loss": 0.5137, + "num_input_tokens_seen": 57785136, + "step": 37970 + }, + { + "epoch": 1.2155111708597401, + "grad_norm": 0.6278939247131348, + "learning_rate": 3.158290788703694e-07, + "loss": 0.4572, + "num_input_tokens_seen": 57792416, + "step": 37975 + }, + { + "epoch": 1.2156712118302284, + "grad_norm": 0.8284906148910522, + "learning_rate": 3.1427539284736297e-07, + "loss": 0.4207, + "num_input_tokens_seen": 57799824, + "step": 37980 + }, + { + "epoch": 1.215831252800717, + "grad_norm": 0.7360280156135559, + "learning_rate": 3.127255136733093e-07, + "loss": 0.5739, + "num_input_tokens_seen": 57807088, + "step": 37985 + }, + { + "epoch": 1.2159912937712054, + "grad_norm": 0.9727685451507568, + "learning_rate": 3.1117944158722544e-07, + "loss": 0.3773, + "num_input_tokens_seen": 57815088, + "step": 37990 + }, + { + "epoch": 1.216151334741694, + "grad_norm": 0.9013731479644775, + "learning_rate": 3.0963717682752635e-07, + "loss": 0.7038, + "num_input_tokens_seen": 57823232, + "step": 37995 + }, + { + "epoch": 1.2163113757121824, + "grad_norm": 0.4750463366508484, + "learning_rate": 3.080987196320578e-07, + "loss": 0.3548, + "num_input_tokens_seen": 57830720, + "step": 38000 + }, + { + "epoch": 1.2163113757121824, + "eval_loss": 0.4941426217556, + "eval_runtime": 583.5233, + "eval_samples_per_second": 23.797, + "eval_steps_per_second": 11.898, + "num_input_tokens_seen": 57830720, + "step": 38000 + }, + { + "epoch": 1.2164714166826707, + "grad_norm": 0.9168775081634521, + "learning_rate": 3.065640702380607e-07, + "loss": 0.5671, + "num_input_tokens_seen": 57838112, + "step": 38005 + }, + { + "epoch": 1.2166314576531592, + "grad_norm": 0.6800039410591125, + "learning_rate": 3.050332288822011e-07, + "loss": 0.5075, + "num_input_tokens_seen": 57845520, + "step": 38010 + }, + { + "epoch": 1.2167914986236477, + "grad_norm": 1.6661577224731445, + "learning_rate": 3.035061958005542e-07, + "loss": 0.405, + "num_input_tokens_seen": 57853248, + "step": 38015 + }, + { + "epoch": 1.2169515395941362, + "grad_norm": 0.7854710817337036, + "learning_rate": 3.019829712286093e-07, + "loss": 0.5733, + "num_input_tokens_seen": 57860800, + "step": 38020 + }, + { + "epoch": 1.2171115805646244, + "grad_norm": 1.0460466146469116, + "learning_rate": 3.004635554012647e-07, + "loss": 0.5543, + "num_input_tokens_seen": 57868016, + "step": 38025 + }, + { + "epoch": 1.217271621535113, + "grad_norm": 0.6460514068603516, + "learning_rate": 2.9894794855283017e-07, + "loss": 0.544, + "num_input_tokens_seen": 57875824, + "step": 38030 + }, + { + "epoch": 1.2174316625056014, + "grad_norm": 0.9793837666511536, + "learning_rate": 2.9743615091703816e-07, + "loss": 0.4714, + "num_input_tokens_seen": 57883472, + "step": 38035 + }, + { + "epoch": 1.21759170347609, + "grad_norm": 0.42688891291618347, + "learning_rate": 2.959281627270216e-07, + "loss": 0.3784, + "num_input_tokens_seen": 57891136, + "step": 38040 + }, + { + "epoch": 1.2177517444465784, + "grad_norm": 0.34113433957099915, + "learning_rate": 2.944239842153362e-07, + "loss": 0.4333, + "num_input_tokens_seen": 57898464, + "step": 38045 + }, + { + "epoch": 1.2179117854170667, + "grad_norm": 1.3895295858383179, + "learning_rate": 2.929236156139381e-07, + "loss": 0.4835, + "num_input_tokens_seen": 57906016, + "step": 38050 + }, + { + "epoch": 1.2180718263875552, + "grad_norm": 0.7003544569015503, + "learning_rate": 2.9142705715420883e-07, + "loss": 0.5504, + "num_input_tokens_seen": 57913616, + "step": 38055 + }, + { + "epoch": 1.2182318673580437, + "grad_norm": 0.7649917602539062, + "learning_rate": 2.8993430906693595e-07, + "loss": 0.4268, + "num_input_tokens_seen": 57921168, + "step": 38060 + }, + { + "epoch": 1.2183919083285322, + "grad_norm": 0.49119457602500916, + "learning_rate": 2.88445371582316e-07, + "loss": 0.5917, + "num_input_tokens_seen": 57928720, + "step": 38065 + }, + { + "epoch": 1.2185519492990204, + "grad_norm": 0.7201398015022278, + "learning_rate": 2.8696024492996796e-07, + "loss": 0.6205, + "num_input_tokens_seen": 57936064, + "step": 38070 + }, + { + "epoch": 1.218711990269509, + "grad_norm": 0.7953628301620483, + "learning_rate": 2.854789293389115e-07, + "loss": 0.435, + "num_input_tokens_seen": 57943760, + "step": 38075 + }, + { + "epoch": 1.2188720312399974, + "grad_norm": 0.6159888505935669, + "learning_rate": 2.8400142503758606e-07, + "loss": 0.4882, + "num_input_tokens_seen": 57951200, + "step": 38080 + }, + { + "epoch": 1.219032072210486, + "grad_norm": 1.4825868606567383, + "learning_rate": 2.8252773225384276e-07, + "loss": 0.4851, + "num_input_tokens_seen": 57958720, + "step": 38085 + }, + { + "epoch": 1.2191921131809744, + "grad_norm": 1.5526807308197021, + "learning_rate": 2.8105785121494143e-07, + "loss": 0.5042, + "num_input_tokens_seen": 57966416, + "step": 38090 + }, + { + "epoch": 1.2193521541514627, + "grad_norm": 0.36942094564437866, + "learning_rate": 2.795917821475563e-07, + "loss": 0.4419, + "num_input_tokens_seen": 57973984, + "step": 38095 + }, + { + "epoch": 1.2195121951219512, + "grad_norm": 0.528627872467041, + "learning_rate": 2.78129525277776e-07, + "loss": 0.4141, + "num_input_tokens_seen": 57982064, + "step": 38100 + }, + { + "epoch": 1.2196722360924397, + "grad_norm": 0.7449367046356201, + "learning_rate": 2.766710808310952e-07, + "loss": 0.4989, + "num_input_tokens_seen": 57990112, + "step": 38105 + }, + { + "epoch": 1.2198322770629282, + "grad_norm": 0.32872945070266724, + "learning_rate": 2.7521644903242827e-07, + "loss": 0.2295, + "num_input_tokens_seen": 57997424, + "step": 38110 + }, + { + "epoch": 1.2199923180334165, + "grad_norm": 0.9211053848266602, + "learning_rate": 2.7376563010609593e-07, + "loss": 0.4842, + "num_input_tokens_seen": 58005040, + "step": 38115 + }, + { + "epoch": 1.220152359003905, + "grad_norm": 1.4698588848114014, + "learning_rate": 2.72318624275833e-07, + "loss": 0.7413, + "num_input_tokens_seen": 58012496, + "step": 38120 + }, + { + "epoch": 1.2203123999743934, + "grad_norm": 0.6614349484443665, + "learning_rate": 2.7087543176478324e-07, + "loss": 0.3388, + "num_input_tokens_seen": 58020480, + "step": 38125 + }, + { + "epoch": 1.220472440944882, + "grad_norm": 0.5787258148193359, + "learning_rate": 2.694360527955103e-07, + "loss": 0.372, + "num_input_tokens_seen": 58028176, + "step": 38130 + }, + { + "epoch": 1.2206324819153704, + "grad_norm": 0.6069596409797668, + "learning_rate": 2.680004875899811e-07, + "loss": 0.3394, + "num_input_tokens_seen": 58036384, + "step": 38135 + }, + { + "epoch": 1.2207925228858587, + "grad_norm": 1.1342823505401611, + "learning_rate": 2.665687363695768e-07, + "loss": 0.4292, + "num_input_tokens_seen": 58043712, + "step": 38140 + }, + { + "epoch": 1.2209525638563472, + "grad_norm": 0.7234998345375061, + "learning_rate": 2.6514079935509584e-07, + "loss": 0.4679, + "num_input_tokens_seen": 58052752, + "step": 38145 + }, + { + "epoch": 1.2211126048268357, + "grad_norm": 0.539867103099823, + "learning_rate": 2.6371667676673983e-07, + "loss": 0.4638, + "num_input_tokens_seen": 58060576, + "step": 38150 + }, + { + "epoch": 1.2212726457973242, + "grad_norm": 0.5505018830299377, + "learning_rate": 2.6229636882412755e-07, + "loss": 0.4413, + "num_input_tokens_seen": 58068144, + "step": 38155 + }, + { + "epoch": 1.2214326867678125, + "grad_norm": 0.6418716311454773, + "learning_rate": 2.6087987574628935e-07, + "loss": 0.4806, + "num_input_tokens_seen": 58075936, + "step": 38160 + }, + { + "epoch": 1.221592727738301, + "grad_norm": 0.7546980381011963, + "learning_rate": 2.5946719775166437e-07, + "loss": 0.3586, + "num_input_tokens_seen": 58083232, + "step": 38165 + }, + { + "epoch": 1.2217527687087895, + "grad_norm": 1.1815811395645142, + "learning_rate": 2.5805833505810616e-07, + "loss": 0.5605, + "num_input_tokens_seen": 58090960, + "step": 38170 + }, + { + "epoch": 1.221912809679278, + "grad_norm": 1.0097259283065796, + "learning_rate": 2.566532878828798e-07, + "loss": 0.3675, + "num_input_tokens_seen": 58098480, + "step": 38175 + }, + { + "epoch": 1.2220728506497665, + "grad_norm": 0.8390321731567383, + "learning_rate": 2.552520564426619e-07, + "loss": 0.4322, + "num_input_tokens_seen": 58106016, + "step": 38180 + }, + { + "epoch": 1.2222328916202547, + "grad_norm": 0.6022911071777344, + "learning_rate": 2.5385464095353803e-07, + "loss": 0.3755, + "num_input_tokens_seen": 58113264, + "step": 38185 + }, + { + "epoch": 1.2223929325907432, + "grad_norm": 0.7268181443214417, + "learning_rate": 2.5246104163100804e-07, + "loss": 0.3708, + "num_input_tokens_seen": 58120512, + "step": 38190 + }, + { + "epoch": 1.2225529735612317, + "grad_norm": 0.661836564540863, + "learning_rate": 2.510712586899833e-07, + "loss": 0.3865, + "num_input_tokens_seen": 58127920, + "step": 38195 + }, + { + "epoch": 1.2227130145317202, + "grad_norm": 0.9175110459327698, + "learning_rate": 2.4968529234478124e-07, + "loss": 0.4162, + "num_input_tokens_seen": 58135872, + "step": 38200 + }, + { + "epoch": 1.2227130145317202, + "eval_loss": 0.49438804388046265, + "eval_runtime": 587.393, + "eval_samples_per_second": 23.64, + "eval_steps_per_second": 11.82, + "num_input_tokens_seen": 58135872, + "step": 38200 + }, + { + "epoch": 1.2228730555022085, + "grad_norm": 0.7055643200874329, + "learning_rate": 2.483031428091448e-07, + "loss": 0.4587, + "num_input_tokens_seen": 58143072, + "step": 38205 + }, + { + "epoch": 1.223033096472697, + "grad_norm": 0.46802058815956116, + "learning_rate": 2.469248102962091e-07, + "loss": 0.3669, + "num_input_tokens_seen": 58151392, + "step": 38210 + }, + { + "epoch": 1.2231931374431855, + "grad_norm": 0.7361202239990234, + "learning_rate": 2.4555029501853455e-07, + "loss": 0.5018, + "num_input_tokens_seen": 58159472, + "step": 38215 + }, + { + "epoch": 1.223353178413674, + "grad_norm": 0.6034494042396545, + "learning_rate": 2.441795971880906e-07, + "loss": 0.4791, + "num_input_tokens_seen": 58166688, + "step": 38220 + }, + { + "epoch": 1.2235132193841625, + "grad_norm": 0.7608619928359985, + "learning_rate": 2.4281271701625255e-07, + "loss": 0.4502, + "num_input_tokens_seen": 58174784, + "step": 38225 + }, + { + "epoch": 1.2236732603546507, + "grad_norm": 1.3780159950256348, + "learning_rate": 2.4144965471381007e-07, + "loss": 0.6092, + "num_input_tokens_seen": 58182432, + "step": 38230 + }, + { + "epoch": 1.2238333013251392, + "grad_norm": 0.9191195964813232, + "learning_rate": 2.400904104909674e-07, + "loss": 0.5188, + "num_input_tokens_seen": 58190096, + "step": 38235 + }, + { + "epoch": 1.2239933422956277, + "grad_norm": 0.8989391326904297, + "learning_rate": 2.3873498455733725e-07, + "loss": 0.5142, + "num_input_tokens_seen": 58197568, + "step": 38240 + }, + { + "epoch": 1.224153383266116, + "grad_norm": 0.7396336793899536, + "learning_rate": 2.3738337712194137e-07, + "loss": 0.5448, + "num_input_tokens_seen": 58205120, + "step": 38245 + }, + { + "epoch": 1.2243134242366045, + "grad_norm": 0.8555662631988525, + "learning_rate": 2.3603558839321305e-07, + "loss": 0.468, + "num_input_tokens_seen": 58212944, + "step": 38250 + }, + { + "epoch": 1.224473465207093, + "grad_norm": 0.5120391845703125, + "learning_rate": 2.3469161857900267e-07, + "loss": 0.522, + "num_input_tokens_seen": 58220464, + "step": 38255 + }, + { + "epoch": 1.2246335061775815, + "grad_norm": 0.48946765065193176, + "learning_rate": 2.3335146788656393e-07, + "loss": 0.5153, + "num_input_tokens_seen": 58227760, + "step": 38260 + }, + { + "epoch": 1.22479354714807, + "grad_norm": 0.8814501762390137, + "learning_rate": 2.3201513652256757e-07, + "loss": 0.5397, + "num_input_tokens_seen": 58235088, + "step": 38265 + }, + { + "epoch": 1.2249535881185583, + "grad_norm": 0.4902891516685486, + "learning_rate": 2.3068262469308766e-07, + "loss": 0.3397, + "num_input_tokens_seen": 58242896, + "step": 38270 + }, + { + "epoch": 1.2251136290890468, + "grad_norm": 0.879802405834198, + "learning_rate": 2.2935393260362093e-07, + "loss": 0.4912, + "num_input_tokens_seen": 58250144, + "step": 38275 + }, + { + "epoch": 1.2252736700595352, + "grad_norm": 0.49599605798721313, + "learning_rate": 2.2802906045906458e-07, + "loss": 0.6376, + "num_input_tokens_seen": 58258016, + "step": 38280 + }, + { + "epoch": 1.2254337110300237, + "grad_norm": 0.463986337184906, + "learning_rate": 2.2670800846373018e-07, + "loss": 0.4573, + "num_input_tokens_seen": 58265472, + "step": 38285 + }, + { + "epoch": 1.225593752000512, + "grad_norm": 0.5876167416572571, + "learning_rate": 2.2539077682134367e-07, + "loss": 0.4971, + "num_input_tokens_seen": 58272960, + "step": 38290 + }, + { + "epoch": 1.2257537929710005, + "grad_norm": 0.8919479250907898, + "learning_rate": 2.2407736573503423e-07, + "loss": 0.3296, + "num_input_tokens_seen": 58281488, + "step": 38295 + }, + { + "epoch": 1.225913833941489, + "grad_norm": 0.6665366291999817, + "learning_rate": 2.2276777540735093e-07, + "loss": 0.2941, + "num_input_tokens_seen": 58288832, + "step": 38300 + }, + { + "epoch": 1.2260738749119775, + "grad_norm": 0.8032627105712891, + "learning_rate": 2.2146200604024613e-07, + "loss": 0.5826, + "num_input_tokens_seen": 58296160, + "step": 38305 + }, + { + "epoch": 1.226233915882466, + "grad_norm": 0.8153811693191528, + "learning_rate": 2.2016005783508375e-07, + "loss": 0.4733, + "num_input_tokens_seen": 58303536, + "step": 38310 + }, + { + "epoch": 1.2263939568529543, + "grad_norm": 0.48079714179039, + "learning_rate": 2.1886193099264763e-07, + "loss": 0.4618, + "num_input_tokens_seen": 58311200, + "step": 38315 + }, + { + "epoch": 1.2265539978234428, + "grad_norm": 0.46981683373451233, + "learning_rate": 2.175676257131165e-07, + "loss": 0.3874, + "num_input_tokens_seen": 58319296, + "step": 38320 + }, + { + "epoch": 1.2267140387939313, + "grad_norm": 0.7585604190826416, + "learning_rate": 2.162771421960974e-07, + "loss": 0.5695, + "num_input_tokens_seen": 58327008, + "step": 38325 + }, + { + "epoch": 1.2268740797644198, + "grad_norm": 0.6189841032028198, + "learning_rate": 2.1499048064059224e-07, + "loss": 0.5098, + "num_input_tokens_seen": 58334768, + "step": 38330 + }, + { + "epoch": 1.227034120734908, + "grad_norm": 0.6767696142196655, + "learning_rate": 2.1370764124502285e-07, + "loss": 0.3888, + "num_input_tokens_seen": 58342432, + "step": 38335 + }, + { + "epoch": 1.2271941617053965, + "grad_norm": 0.6765265464782715, + "learning_rate": 2.1242862420721988e-07, + "loss": 0.3496, + "num_input_tokens_seen": 58349776, + "step": 38340 + }, + { + "epoch": 1.227354202675885, + "grad_norm": 0.7971556782722473, + "learning_rate": 2.1115342972442276e-07, + "loss": 0.3998, + "num_input_tokens_seen": 58357312, + "step": 38345 + }, + { + "epoch": 1.2275142436463735, + "grad_norm": 0.5360360741615295, + "learning_rate": 2.0988205799328252e-07, + "loss": 0.5677, + "num_input_tokens_seen": 58364800, + "step": 38350 + }, + { + "epoch": 1.227674284616862, + "grad_norm": 0.7403934001922607, + "learning_rate": 2.0861450920986182e-07, + "loss": 0.4577, + "num_input_tokens_seen": 58372800, + "step": 38355 + }, + { + "epoch": 1.2278343255873503, + "grad_norm": 0.8006895184516907, + "learning_rate": 2.07350783569632e-07, + "loss": 0.3922, + "num_input_tokens_seen": 58380384, + "step": 38360 + }, + { + "epoch": 1.2279943665578388, + "grad_norm": 0.6242028474807739, + "learning_rate": 2.060908812674761e-07, + "loss": 0.4258, + "num_input_tokens_seen": 58387760, + "step": 38365 + }, + { + "epoch": 1.2281544075283273, + "grad_norm": 0.5453587174415588, + "learning_rate": 2.0483480249768317e-07, + "loss": 0.3273, + "num_input_tokens_seen": 58395408, + "step": 38370 + }, + { + "epoch": 1.2283144484988158, + "grad_norm": 0.8774694800376892, + "learning_rate": 2.035825474539621e-07, + "loss": 0.7632, + "num_input_tokens_seen": 58403088, + "step": 38375 + }, + { + "epoch": 1.228474489469304, + "grad_norm": 0.26435336470603943, + "learning_rate": 2.0233411632942235e-07, + "loss": 0.4048, + "num_input_tokens_seen": 58410304, + "step": 38380 + }, + { + "epoch": 1.2286345304397925, + "grad_norm": 0.636542797088623, + "learning_rate": 2.0108950931658764e-07, + "loss": 0.47, + "num_input_tokens_seen": 58417472, + "step": 38385 + }, + { + "epoch": 1.228794571410281, + "grad_norm": 0.7513243556022644, + "learning_rate": 1.998487266073934e-07, + "loss": 0.43, + "num_input_tokens_seen": 58424496, + "step": 38390 + }, + { + "epoch": 1.2289546123807695, + "grad_norm": 0.735522985458374, + "learning_rate": 1.986117683931865e-07, + "loss": 0.4552, + "num_input_tokens_seen": 58431744, + "step": 38395 + }, + { + "epoch": 1.229114653351258, + "grad_norm": 1.0810816287994385, + "learning_rate": 1.9737863486471442e-07, + "loss": 0.4424, + "num_input_tokens_seen": 58439424, + "step": 38400 + }, + { + "epoch": 1.229114653351258, + "eval_loss": 0.4941679835319519, + "eval_runtime": 586.5373, + "eval_samples_per_second": 23.675, + "eval_steps_per_second": 11.837, + "num_input_tokens_seen": 58439424, + "step": 38400 + }, + { + "epoch": 1.2292746943217463, + "grad_norm": 1.2872860431671143, + "learning_rate": 1.9614932621215e-07, + "loss": 0.5649, + "num_input_tokens_seen": 58446992, + "step": 38405 + }, + { + "epoch": 1.2294347352922348, + "grad_norm": 0.4290251135826111, + "learning_rate": 1.9492384262506102e-07, + "loss": 0.3156, + "num_input_tokens_seen": 58454432, + "step": 38410 + }, + { + "epoch": 1.2295947762627233, + "grad_norm": 1.3200913667678833, + "learning_rate": 1.9370218429243524e-07, + "loss": 0.5435, + "num_input_tokens_seen": 58461888, + "step": 38415 + }, + { + "epoch": 1.2297548172332118, + "grad_norm": 0.6015681028366089, + "learning_rate": 1.9248435140267197e-07, + "loss": 0.5201, + "num_input_tokens_seen": 58469440, + "step": 38420 + }, + { + "epoch": 1.2299148582037, + "grad_norm": 0.5510584115982056, + "learning_rate": 1.9127034414356814e-07, + "loss": 0.4151, + "num_input_tokens_seen": 58476768, + "step": 38425 + }, + { + "epoch": 1.2300748991741886, + "grad_norm": 0.6535500288009644, + "learning_rate": 1.9006016270234627e-07, + "loss": 0.6317, + "num_input_tokens_seen": 58483904, + "step": 38430 + }, + { + "epoch": 1.230234940144677, + "grad_norm": 0.6988920569419861, + "learning_rate": 1.888538072656293e-07, + "loss": 0.5545, + "num_input_tokens_seen": 58491504, + "step": 38435 + }, + { + "epoch": 1.2303949811151655, + "grad_norm": 1.2187659740447998, + "learning_rate": 1.8765127801944893e-07, + "loss": 0.3731, + "num_input_tokens_seen": 58499136, + "step": 38440 + }, + { + "epoch": 1.230555022085654, + "grad_norm": 0.7151020169258118, + "learning_rate": 1.8645257514925406e-07, + "loss": 0.5821, + "num_input_tokens_seen": 58506592, + "step": 38445 + }, + { + "epoch": 1.2307150630561423, + "grad_norm": 0.6828793287277222, + "learning_rate": 1.8525769883989685e-07, + "loss": 0.4984, + "num_input_tokens_seen": 58514480, + "step": 38450 + }, + { + "epoch": 1.2308751040266308, + "grad_norm": 1.7336103916168213, + "learning_rate": 1.8406664927564654e-07, + "loss": 0.6306, + "num_input_tokens_seen": 58522208, + "step": 38455 + }, + { + "epoch": 1.2310351449971193, + "grad_norm": 0.8359353542327881, + "learning_rate": 1.8287942664017566e-07, + "loss": 0.4625, + "num_input_tokens_seen": 58529808, + "step": 38460 + }, + { + "epoch": 1.2311951859676078, + "grad_norm": 0.6648274660110474, + "learning_rate": 1.8169603111656552e-07, + "loss": 0.6164, + "num_input_tokens_seen": 58537552, + "step": 38465 + }, + { + "epoch": 1.231355226938096, + "grad_norm": 0.6761818528175354, + "learning_rate": 1.805164628873146e-07, + "loss": 0.5288, + "num_input_tokens_seen": 58545936, + "step": 38470 + }, + { + "epoch": 1.2315152679085846, + "grad_norm": 1.2326289415359497, + "learning_rate": 1.793407221343274e-07, + "loss": 0.7037, + "num_input_tokens_seen": 58553664, + "step": 38475 + }, + { + "epoch": 1.231675308879073, + "grad_norm": 0.7174332141876221, + "learning_rate": 1.781688090389172e-07, + "loss": 0.4379, + "num_input_tokens_seen": 58561408, + "step": 38480 + }, + { + "epoch": 1.2318353498495616, + "grad_norm": 0.7930138111114502, + "learning_rate": 1.770007237818061e-07, + "loss": 0.4569, + "num_input_tokens_seen": 58568896, + "step": 38485 + }, + { + "epoch": 1.23199539082005, + "grad_norm": 0.4562898576259613, + "learning_rate": 1.7583646654313059e-07, + "loss": 0.4564, + "num_input_tokens_seen": 58575984, + "step": 38490 + }, + { + "epoch": 1.2321554317905383, + "grad_norm": 0.8066756725311279, + "learning_rate": 1.7467603750242757e-07, + "loss": 0.3634, + "num_input_tokens_seen": 58583424, + "step": 38495 + }, + { + "epoch": 1.2323154727610268, + "grad_norm": 0.8928836584091187, + "learning_rate": 1.7351943683865944e-07, + "loss": 0.5166, + "num_input_tokens_seen": 58590336, + "step": 38500 + }, + { + "epoch": 1.2324755137315153, + "grad_norm": 0.6351864337921143, + "learning_rate": 1.723666647301808e-07, + "loss": 0.4483, + "num_input_tokens_seen": 58597664, + "step": 38505 + }, + { + "epoch": 1.2326355547020036, + "grad_norm": 0.6434081792831421, + "learning_rate": 1.712177213547661e-07, + "loss": 0.4152, + "num_input_tokens_seen": 58605168, + "step": 38510 + }, + { + "epoch": 1.232795595672492, + "grad_norm": 0.30581822991371155, + "learning_rate": 1.7007260688959581e-07, + "loss": 0.4702, + "num_input_tokens_seen": 58612928, + "step": 38515 + }, + { + "epoch": 1.2329556366429806, + "grad_norm": 0.9752845764160156, + "learning_rate": 1.68931321511262e-07, + "loss": 0.4876, + "num_input_tokens_seen": 58620320, + "step": 38520 + }, + { + "epoch": 1.233115677613469, + "grad_norm": 0.6479003429412842, + "learning_rate": 1.6779386539576835e-07, + "loss": 0.4805, + "num_input_tokens_seen": 58628336, + "step": 38525 + }, + { + "epoch": 1.2332757185839576, + "grad_norm": 0.6616200804710388, + "learning_rate": 1.666602387185162e-07, + "loss": 0.4433, + "num_input_tokens_seen": 58635984, + "step": 38530 + }, + { + "epoch": 1.2334357595544458, + "grad_norm": 0.49830323457717896, + "learning_rate": 1.655304416543352e-07, + "loss": 0.5113, + "num_input_tokens_seen": 58643776, + "step": 38535 + }, + { + "epoch": 1.2335958005249343, + "grad_norm": 0.635942816734314, + "learning_rate": 1.6440447437744698e-07, + "loss": 0.4807, + "num_input_tokens_seen": 58651280, + "step": 38540 + }, + { + "epoch": 1.2337558414954228, + "grad_norm": 1.0306850671768188, + "learning_rate": 1.6328233706149332e-07, + "loss": 0.4911, + "num_input_tokens_seen": 58658928, + "step": 38545 + }, + { + "epoch": 1.2339158824659113, + "grad_norm": 0.8582386374473572, + "learning_rate": 1.6216402987951906e-07, + "loss": 0.5301, + "num_input_tokens_seen": 58666048, + "step": 38550 + }, + { + "epoch": 1.2340759234363996, + "grad_norm": 0.6477484703063965, + "learning_rate": 1.6104955300398627e-07, + "loss": 0.3739, + "num_input_tokens_seen": 58673632, + "step": 38555 + }, + { + "epoch": 1.234235964406888, + "grad_norm": 0.8849284648895264, + "learning_rate": 1.5993890660675748e-07, + "loss": 0.4219, + "num_input_tokens_seen": 58681328, + "step": 38560 + }, + { + "epoch": 1.2343960053773766, + "grad_norm": 0.5035168528556824, + "learning_rate": 1.5883209085910678e-07, + "loss": 0.3118, + "num_input_tokens_seen": 58689424, + "step": 38565 + }, + { + "epoch": 1.234556046347865, + "grad_norm": 0.5783849954605103, + "learning_rate": 1.5772910593172264e-07, + "loss": 0.5461, + "num_input_tokens_seen": 58697248, + "step": 38570 + }, + { + "epoch": 1.2347160873183536, + "grad_norm": 0.8740996718406677, + "learning_rate": 1.5662995199469954e-07, + "loss": 0.527, + "num_input_tokens_seen": 58705264, + "step": 38575 + }, + { + "epoch": 1.2348761282888419, + "grad_norm": 1.215350866317749, + "learning_rate": 1.5553462921753802e-07, + "loss": 0.4531, + "num_input_tokens_seen": 58713200, + "step": 38580 + }, + { + "epoch": 1.2350361692593304, + "grad_norm": 0.6201362609863281, + "learning_rate": 1.544431377691502e-07, + "loss": 0.3742, + "num_input_tokens_seen": 58720704, + "step": 38585 + }, + { + "epoch": 1.2351962102298188, + "grad_norm": 0.6186583042144775, + "learning_rate": 1.5335547781785975e-07, + "loss": 0.3733, + "num_input_tokens_seen": 58728208, + "step": 38590 + }, + { + "epoch": 1.2353562512003073, + "grad_norm": 0.9074679613113403, + "learning_rate": 1.5227164953139917e-07, + "loss": 0.4185, + "num_input_tokens_seen": 58735808, + "step": 38595 + }, + { + "epoch": 1.2355162921707956, + "grad_norm": 0.6116739511489868, + "learning_rate": 1.511916530769042e-07, + "loss": 0.4559, + "num_input_tokens_seen": 58743296, + "step": 38600 + }, + { + "epoch": 1.2355162921707956, + "eval_loss": 0.4942282736301422, + "eval_runtime": 562.6418, + "eval_samples_per_second": 24.68, + "eval_steps_per_second": 12.34, + "num_input_tokens_seen": 58743296, + "step": 38600 + }, + { + "epoch": 1.2356763331412841, + "grad_norm": 0.4487040042877197, + "learning_rate": 1.5011548862092773e-07, + "loss": 0.5615, + "num_input_tokens_seen": 58751104, + "step": 38605 + }, + { + "epoch": 1.2358363741117726, + "grad_norm": 0.384289026260376, + "learning_rate": 1.490431563294231e-07, + "loss": 0.3605, + "num_input_tokens_seen": 58758928, + "step": 38610 + }, + { + "epoch": 1.235996415082261, + "grad_norm": 0.3776870369911194, + "learning_rate": 1.4797465636776365e-07, + "loss": 0.4831, + "num_input_tokens_seen": 58767056, + "step": 38615 + }, + { + "epoch": 1.2361564560527496, + "grad_norm": 2.7778613567352295, + "learning_rate": 1.4690998890072027e-07, + "loss": 0.5386, + "num_input_tokens_seen": 58774096, + "step": 38620 + }, + { + "epoch": 1.2363164970232379, + "grad_norm": 0.9624217748641968, + "learning_rate": 1.4584915409248112e-07, + "loss": 0.3139, + "num_input_tokens_seen": 58781584, + "step": 38625 + }, + { + "epoch": 1.2364765379937264, + "grad_norm": 0.4684661030769348, + "learning_rate": 1.4479215210663754e-07, + "loss": 0.741, + "num_input_tokens_seen": 58789536, + "step": 38630 + }, + { + "epoch": 1.2366365789642149, + "grad_norm": 1.3891732692718506, + "learning_rate": 1.4373898310619528e-07, + "loss": 0.4952, + "num_input_tokens_seen": 58797024, + "step": 38635 + }, + { + "epoch": 1.2367966199347034, + "grad_norm": 0.7542610764503479, + "learning_rate": 1.4268964725356604e-07, + "loss": 0.5168, + "num_input_tokens_seen": 58804400, + "step": 38640 + }, + { + "epoch": 1.2369566609051916, + "grad_norm": 0.6698740124702454, + "learning_rate": 1.4164414471056764e-07, + "loss": 0.4752, + "num_input_tokens_seen": 58812272, + "step": 38645 + }, + { + "epoch": 1.2371167018756801, + "grad_norm": 1.3354361057281494, + "learning_rate": 1.4060247563843497e-07, + "loss": 0.6487, + "num_input_tokens_seen": 58819952, + "step": 38650 + }, + { + "epoch": 1.2372767428461686, + "grad_norm": 0.4753945469856262, + "learning_rate": 1.3956464019780068e-07, + "loss": 0.3791, + "num_input_tokens_seen": 58827312, + "step": 38655 + }, + { + "epoch": 1.2374367838166571, + "grad_norm": 1.3449164628982544, + "learning_rate": 1.385306385487145e-07, + "loss": 0.4894, + "num_input_tokens_seen": 58834960, + "step": 38660 + }, + { + "epoch": 1.2375968247871456, + "grad_norm": 1.2734758853912354, + "learning_rate": 1.3750047085063222e-07, + "loss": 0.5424, + "num_input_tokens_seen": 58842432, + "step": 38665 + }, + { + "epoch": 1.2377568657576339, + "grad_norm": 0.8779078722000122, + "learning_rate": 1.3647413726242119e-07, + "loss": 0.5731, + "num_input_tokens_seen": 58849760, + "step": 38670 + }, + { + "epoch": 1.2379169067281224, + "grad_norm": 0.7764212489128113, + "learning_rate": 1.3545163794235205e-07, + "loss": 0.5472, + "num_input_tokens_seen": 58857344, + "step": 38675 + }, + { + "epoch": 1.2380769476986109, + "grad_norm": 0.3417578935623169, + "learning_rate": 1.3443297304810698e-07, + "loss": 0.4377, + "num_input_tokens_seen": 58864544, + "step": 38680 + }, + { + "epoch": 1.2382369886690994, + "grad_norm": 1.2174999713897705, + "learning_rate": 1.3341814273677977e-07, + "loss": 0.6053, + "num_input_tokens_seen": 58871744, + "step": 38685 + }, + { + "epoch": 1.2383970296395876, + "grad_norm": 1.8138056993484497, + "learning_rate": 1.324071471648647e-07, + "loss": 0.8481, + "num_input_tokens_seen": 58879440, + "step": 38690 + }, + { + "epoch": 1.2385570706100761, + "grad_norm": 0.3912934362888336, + "learning_rate": 1.3139998648827312e-07, + "loss": 0.5339, + "num_input_tokens_seen": 58887200, + "step": 38695 + }, + { + "epoch": 1.2387171115805646, + "grad_norm": 0.544904887676239, + "learning_rate": 1.3039666086232526e-07, + "loss": 0.4984, + "num_input_tokens_seen": 58894736, + "step": 38700 + }, + { + "epoch": 1.2388771525510531, + "grad_norm": 0.8213520050048828, + "learning_rate": 1.2939717044174183e-07, + "loss": 0.3012, + "num_input_tokens_seen": 58902416, + "step": 38705 + }, + { + "epoch": 1.2390371935215416, + "grad_norm": 0.44076526165008545, + "learning_rate": 1.284015153806578e-07, + "loss": 0.3788, + "num_input_tokens_seen": 58909728, + "step": 38710 + }, + { + "epoch": 1.23919723449203, + "grad_norm": 0.5023871064186096, + "learning_rate": 1.274096958326171e-07, + "loss": 0.5731, + "num_input_tokens_seen": 58917872, + "step": 38715 + }, + { + "epoch": 1.2393572754625184, + "grad_norm": 0.698128879070282, + "learning_rate": 1.2642171195056952e-07, + "loss": 0.4731, + "num_input_tokens_seen": 58925648, + "step": 38720 + }, + { + "epoch": 1.2395173164330069, + "grad_norm": 0.7997857928276062, + "learning_rate": 1.2543756388687377e-07, + "loss": 0.4137, + "num_input_tokens_seen": 58933536, + "step": 38725 + }, + { + "epoch": 1.2396773574034954, + "grad_norm": 1.8028781414031982, + "learning_rate": 1.2445725179330014e-07, + "loss": 0.5078, + "num_input_tokens_seen": 58941088, + "step": 38730 + }, + { + "epoch": 1.2398373983739837, + "grad_norm": 1.0122138261795044, + "learning_rate": 1.2348077582102212e-07, + "loss": 0.5563, + "num_input_tokens_seen": 58948432, + "step": 38735 + }, + { + "epoch": 1.2399974393444722, + "grad_norm": 1.0961980819702148, + "learning_rate": 1.2250813612062762e-07, + "loss": 0.5843, + "num_input_tokens_seen": 58955984, + "step": 38740 + }, + { + "epoch": 1.2401574803149606, + "grad_norm": 0.6354242563247681, + "learning_rate": 1.215393328421105e-07, + "loss": 0.5285, + "num_input_tokens_seen": 58963120, + "step": 38745 + }, + { + "epoch": 1.2403175212854491, + "grad_norm": 1.0413198471069336, + "learning_rate": 1.2057436613486796e-07, + "loss": 0.3123, + "num_input_tokens_seen": 58970768, + "step": 38750 + }, + { + "epoch": 1.2404775622559376, + "grad_norm": 0.5048874616622925, + "learning_rate": 1.1961323614771424e-07, + "loss": 0.3742, + "num_input_tokens_seen": 58978352, + "step": 38755 + }, + { + "epoch": 1.240637603226426, + "grad_norm": 0.9270750880241394, + "learning_rate": 1.1865594302886418e-07, + "loss": 0.3681, + "num_input_tokens_seen": 58985904, + "step": 38760 + }, + { + "epoch": 1.2407976441969144, + "grad_norm": 0.36088964343070984, + "learning_rate": 1.1770248692594687e-07, + "loss": 0.5285, + "num_input_tokens_seen": 58993408, + "step": 38765 + }, + { + "epoch": 1.240957685167403, + "grad_norm": 0.6993919014930725, + "learning_rate": 1.167528679859975e-07, + "loss": 0.4327, + "num_input_tokens_seen": 59001168, + "step": 38770 + }, + { + "epoch": 1.2411177261378914, + "grad_norm": 0.6737692952156067, + "learning_rate": 1.1580708635545446e-07, + "loss": 0.5469, + "num_input_tokens_seen": 59009152, + "step": 38775 + }, + { + "epoch": 1.2412777671083797, + "grad_norm": 1.2712591886520386, + "learning_rate": 1.1486514218017885e-07, + "loss": 0.4092, + "num_input_tokens_seen": 59016448, + "step": 38780 + }, + { + "epoch": 1.2414378080788682, + "grad_norm": 0.6024255752563477, + "learning_rate": 1.1392703560542117e-07, + "loss": 0.384, + "num_input_tokens_seen": 59024064, + "step": 38785 + }, + { + "epoch": 1.2415978490493567, + "grad_norm": 0.6238341927528381, + "learning_rate": 1.129927667758518e-07, + "loss": 0.5381, + "num_input_tokens_seen": 59031088, + "step": 38790 + }, + { + "epoch": 1.2417578900198452, + "grad_norm": 1.098850131034851, + "learning_rate": 1.1206233583554992e-07, + "loss": 0.4967, + "num_input_tokens_seen": 59038720, + "step": 38795 + }, + { + "epoch": 1.2419179309903334, + "grad_norm": 2.1525299549102783, + "learning_rate": 1.1113574292799523e-07, + "loss": 0.5134, + "num_input_tokens_seen": 59046368, + "step": 38800 + }, + { + "epoch": 1.2419179309903334, + "eval_loss": 0.49421513080596924, + "eval_runtime": 533.6774, + "eval_samples_per_second": 26.019, + "eval_steps_per_second": 13.01, + "num_input_tokens_seen": 59046368, + "step": 38800 + }, + { + "epoch": 1.242077971960822, + "grad_norm": 0.7365657687187195, + "learning_rate": 1.1021298819608449e-07, + "loss": 0.3898, + "num_input_tokens_seen": 59054176, + "step": 38805 + }, + { + "epoch": 1.2422380129313104, + "grad_norm": 0.7960294485092163, + "learning_rate": 1.0929407178211226e-07, + "loss": 0.5212, + "num_input_tokens_seen": 59061456, + "step": 38810 + }, + { + "epoch": 1.242398053901799, + "grad_norm": 0.4185481369495392, + "learning_rate": 1.0837899382779293e-07, + "loss": 0.6375, + "num_input_tokens_seen": 59068832, + "step": 38815 + }, + { + "epoch": 1.2425580948722872, + "grad_norm": 0.5909984111785889, + "learning_rate": 1.0746775447423862e-07, + "loss": 0.4976, + "num_input_tokens_seen": 59076320, + "step": 38820 + }, + { + "epoch": 1.2427181358427757, + "grad_norm": 0.9037697315216064, + "learning_rate": 1.0656035386197583e-07, + "loss": 0.5157, + "num_input_tokens_seen": 59084064, + "step": 38825 + }, + { + "epoch": 1.2428781768132642, + "grad_norm": 0.4624319076538086, + "learning_rate": 1.0565679213093982e-07, + "loss": 0.4445, + "num_input_tokens_seen": 59091808, + "step": 38830 + }, + { + "epoch": 1.2430382177837527, + "grad_norm": 0.7861763834953308, + "learning_rate": 1.0475706942046638e-07, + "loss": 0.3701, + "num_input_tokens_seen": 59098976, + "step": 38835 + }, + { + "epoch": 1.2431982587542412, + "grad_norm": 0.8317458629608154, + "learning_rate": 1.0386118586930282e-07, + "loss": 0.5032, + "num_input_tokens_seen": 59106496, + "step": 38840 + }, + { + "epoch": 1.2433582997247294, + "grad_norm": 0.9274567365646362, + "learning_rate": 1.0296914161561367e-07, + "loss": 0.5334, + "num_input_tokens_seen": 59113760, + "step": 38845 + }, + { + "epoch": 1.243518340695218, + "grad_norm": 0.6430602669715881, + "learning_rate": 1.0208093679695552e-07, + "loss": 0.4242, + "num_input_tokens_seen": 59121328, + "step": 38850 + }, + { + "epoch": 1.2436783816657064, + "grad_norm": 0.6834715604782104, + "learning_rate": 1.0119657155030493e-07, + "loss": 0.4496, + "num_input_tokens_seen": 59128496, + "step": 38855 + }, + { + "epoch": 1.243838422636195, + "grad_norm": 0.4201795160770416, + "learning_rate": 1.003160460120417e-07, + "loss": 0.4679, + "num_input_tokens_seen": 59136096, + "step": 38860 + }, + { + "epoch": 1.2439984636066832, + "grad_norm": 0.8802342414855957, + "learning_rate": 9.943936031795165e-08, + "loss": 0.4507, + "num_input_tokens_seen": 59143744, + "step": 38865 + }, + { + "epoch": 1.2441585045771717, + "grad_norm": 0.5554446578025818, + "learning_rate": 9.856651460323219e-08, + "loss": 0.4367, + "num_input_tokens_seen": 59151488, + "step": 38870 + }, + { + "epoch": 1.2443185455476602, + "grad_norm": 0.7407867312431335, + "learning_rate": 9.769750900248953e-08, + "loss": 0.4854, + "num_input_tokens_seen": 59159232, + "step": 38875 + }, + { + "epoch": 1.2444785865181487, + "grad_norm": 0.7709037661552429, + "learning_rate": 9.683234364973038e-08, + "loss": 0.4167, + "num_input_tokens_seen": 59167040, + "step": 38880 + }, + { + "epoch": 1.2446386274886372, + "grad_norm": 0.654590368270874, + "learning_rate": 9.597101867837854e-08, + "loss": 0.4805, + "num_input_tokens_seen": 59174320, + "step": 38885 + }, + { + "epoch": 1.2447986684591255, + "grad_norm": 1.1341031789779663, + "learning_rate": 9.511353422125835e-08, + "loss": 0.4809, + "num_input_tokens_seen": 59182400, + "step": 38890 + }, + { + "epoch": 1.244958709429614, + "grad_norm": 0.5589483380317688, + "learning_rate": 9.42598904106029e-08, + "loss": 0.4931, + "num_input_tokens_seen": 59190608, + "step": 38895 + }, + { + "epoch": 1.2451187504001024, + "grad_norm": 0.7441843748092651, + "learning_rate": 9.341008737806245e-08, + "loss": 0.3916, + "num_input_tokens_seen": 59198112, + "step": 38900 + }, + { + "epoch": 1.245278791370591, + "grad_norm": 0.9134421944618225, + "learning_rate": 9.256412525467661e-08, + "loss": 0.6117, + "num_input_tokens_seen": 59205696, + "step": 38905 + }, + { + "epoch": 1.2454388323410792, + "grad_norm": 0.8026689291000366, + "learning_rate": 9.172200417091326e-08, + "loss": 0.4208, + "num_input_tokens_seen": 59213664, + "step": 38910 + }, + { + "epoch": 1.2455988733115677, + "grad_norm": 0.38376379013061523, + "learning_rate": 9.088372425663239e-08, + "loss": 0.3858, + "num_input_tokens_seen": 59220992, + "step": 38915 + }, + { + "epoch": 1.2457589142820562, + "grad_norm": 0.7693675756454468, + "learning_rate": 9.004928564110837e-08, + "loss": 0.4945, + "num_input_tokens_seen": 59228512, + "step": 38920 + }, + { + "epoch": 1.2459189552525447, + "grad_norm": 0.5228487849235535, + "learning_rate": 8.92186884530244e-08, + "loss": 0.4127, + "num_input_tokens_seen": 59236160, + "step": 38925 + }, + { + "epoch": 1.2460789962230332, + "grad_norm": 0.5646858215332031, + "learning_rate": 8.83919328204641e-08, + "loss": 0.5023, + "num_input_tokens_seen": 59244032, + "step": 38930 + }, + { + "epoch": 1.2462390371935215, + "grad_norm": 1.1262710094451904, + "learning_rate": 8.756901887093105e-08, + "loss": 0.6144, + "num_input_tokens_seen": 59251808, + "step": 38935 + }, + { + "epoch": 1.24639907816401, + "grad_norm": 1.2475544214248657, + "learning_rate": 8.674994673132098e-08, + "loss": 0.3581, + "num_input_tokens_seen": 59259296, + "step": 38940 + }, + { + "epoch": 1.2465591191344985, + "grad_norm": 1.0060884952545166, + "learning_rate": 8.593471652794949e-08, + "loss": 0.5099, + "num_input_tokens_seen": 59266784, + "step": 38945 + }, + { + "epoch": 1.246719160104987, + "grad_norm": 0.6672375202178955, + "learning_rate": 8.512332838653548e-08, + "loss": 0.4131, + "num_input_tokens_seen": 59274912, + "step": 38950 + }, + { + "epoch": 1.2468792010754752, + "grad_norm": 0.6914503574371338, + "learning_rate": 8.431578243220106e-08, + "loss": 0.5594, + "num_input_tokens_seen": 59282192, + "step": 38955 + }, + { + "epoch": 1.2470392420459637, + "grad_norm": 0.5123249888420105, + "learning_rate": 8.351207878948552e-08, + "loss": 0.3742, + "num_input_tokens_seen": 59290096, + "step": 38960 + }, + { + "epoch": 1.2471992830164522, + "grad_norm": 0.536521852016449, + "learning_rate": 8.271221758232583e-08, + "loss": 0.3666, + "num_input_tokens_seen": 59298000, + "step": 38965 + }, + { + "epoch": 1.2473593239869407, + "grad_norm": 1.2077661752700806, + "learning_rate": 8.191619893407332e-08, + "loss": 0.5614, + "num_input_tokens_seen": 59305824, + "step": 38970 + }, + { + "epoch": 1.2475193649574292, + "grad_norm": 0.36623674631118774, + "learning_rate": 8.112402296748534e-08, + "loss": 0.5657, + "num_input_tokens_seen": 59313424, + "step": 38975 + }, + { + "epoch": 1.2476794059279175, + "grad_norm": 0.4891241490840912, + "learning_rate": 8.033568980471973e-08, + "loss": 0.3359, + "num_input_tokens_seen": 59321328, + "step": 38980 + }, + { + "epoch": 1.247839446898406, + "grad_norm": 0.5104494094848633, + "learning_rate": 7.955119956735146e-08, + "loss": 0.4698, + "num_input_tokens_seen": 59328848, + "step": 38985 + }, + { + "epoch": 1.2479994878688945, + "grad_norm": 0.5834977030754089, + "learning_rate": 7.877055237636155e-08, + "loss": 0.4819, + "num_input_tokens_seen": 59336688, + "step": 38990 + }, + { + "epoch": 1.248159528839383, + "grad_norm": 0.43096843361854553, + "learning_rate": 7.79937483521287e-08, + "loss": 0.4146, + "num_input_tokens_seen": 59343904, + "step": 38995 + }, + { + "epoch": 1.2483195698098712, + "grad_norm": 0.4978943169116974, + "learning_rate": 7.722078761444873e-08, + "loss": 0.4145, + "num_input_tokens_seen": 59351696, + "step": 39000 + }, + { + "epoch": 1.2483195698098712, + "eval_loss": 0.4941789507865906, + "eval_runtime": 533.7415, + "eval_samples_per_second": 26.016, + "eval_steps_per_second": 13.008, + "num_input_tokens_seen": 59351696, + "step": 39000 + }, + { + "epoch": 1.2484796107803597, + "grad_norm": 0.3916763961315155, + "learning_rate": 7.645167028252631e-08, + "loss": 0.4174, + "num_input_tokens_seen": 59359760, + "step": 39005 + }, + { + "epoch": 1.2486396517508482, + "grad_norm": 0.9264020919799805, + "learning_rate": 7.568639647496379e-08, + "loss": 0.3462, + "num_input_tokens_seen": 59367312, + "step": 39010 + }, + { + "epoch": 1.2487996927213367, + "grad_norm": 0.7746555209159851, + "learning_rate": 7.492496630977508e-08, + "loss": 0.4587, + "num_input_tokens_seen": 59375056, + "step": 39015 + }, + { + "epoch": 1.2489597336918252, + "grad_norm": 0.8502033948898315, + "learning_rate": 7.416737990438571e-08, + "loss": 0.6009, + "num_input_tokens_seen": 59382880, + "step": 39020 + }, + { + "epoch": 1.2491197746623135, + "grad_norm": 0.4181652367115021, + "learning_rate": 7.341363737562445e-08, + "loss": 0.3887, + "num_input_tokens_seen": 59390976, + "step": 39025 + }, + { + "epoch": 1.249279815632802, + "grad_norm": 0.8851715326309204, + "learning_rate": 7.266373883972887e-08, + "loss": 0.5282, + "num_input_tokens_seen": 59398992, + "step": 39030 + }, + { + "epoch": 1.2494398566032905, + "grad_norm": 0.5590914487838745, + "learning_rate": 7.191768441233981e-08, + "loss": 0.423, + "num_input_tokens_seen": 59407184, + "step": 39035 + }, + { + "epoch": 1.249599897573779, + "grad_norm": 0.6891692876815796, + "learning_rate": 7.11754742085069e-08, + "loss": 0.6351, + "num_input_tokens_seen": 59414752, + "step": 39040 + }, + { + "epoch": 1.2497599385442673, + "grad_norm": 0.5256860256195068, + "learning_rate": 7.043710834269413e-08, + "loss": 0.6255, + "num_input_tokens_seen": 59422048, + "step": 39045 + }, + { + "epoch": 1.2499199795147558, + "grad_norm": 0.6281445622444153, + "learning_rate": 6.970258692876319e-08, + "loss": 0.5677, + "num_input_tokens_seen": 59429760, + "step": 39050 + }, + { + "epoch": 1.2500800204852442, + "grad_norm": 0.5407058596611023, + "learning_rate": 6.897191007998738e-08, + "loss": 0.5029, + "num_input_tokens_seen": 59437248, + "step": 39055 + }, + { + "epoch": 1.2502400614557327, + "grad_norm": 0.42455464601516724, + "learning_rate": 6.824507790904599e-08, + "loss": 0.5029, + "num_input_tokens_seen": 59445248, + "step": 39060 + }, + { + "epoch": 1.2504001024262212, + "grad_norm": 0.7274391055107117, + "learning_rate": 6.752209052802439e-08, + "loss": 0.3646, + "num_input_tokens_seen": 59452912, + "step": 39065 + }, + { + "epoch": 1.2505601433967095, + "grad_norm": 0.6861937046051025, + "learning_rate": 6.680294804841946e-08, + "loss": 0.4337, + "num_input_tokens_seen": 59460896, + "step": 39070 + }, + { + "epoch": 1.250720184367198, + "grad_norm": 0.7569223046302795, + "learning_rate": 6.608765058112865e-08, + "loss": 0.4648, + "num_input_tokens_seen": 59468352, + "step": 39075 + }, + { + "epoch": 1.2508802253376865, + "grad_norm": 0.7938427329063416, + "learning_rate": 6.537619823646368e-08, + "loss": 0.3768, + "num_input_tokens_seen": 59476576, + "step": 39080 + }, + { + "epoch": 1.2510402663081748, + "grad_norm": 0.6487194299697876, + "learning_rate": 6.466859112413404e-08, + "loss": 0.3785, + "num_input_tokens_seen": 59484464, + "step": 39085 + }, + { + "epoch": 1.2512003072786633, + "grad_norm": 1.0623362064361572, + "learning_rate": 6.39648293532663e-08, + "loss": 0.467, + "num_input_tokens_seen": 59491664, + "step": 39090 + }, + { + "epoch": 1.2513603482491518, + "grad_norm": 0.671907365322113, + "learning_rate": 6.32649130323848e-08, + "loss": 0.5558, + "num_input_tokens_seen": 59499552, + "step": 39095 + }, + { + "epoch": 1.2515203892196403, + "grad_norm": 0.5494381189346313, + "learning_rate": 6.256884226943094e-08, + "loss": 0.6052, + "num_input_tokens_seen": 59506720, + "step": 39100 + }, + { + "epoch": 1.2516804301901288, + "grad_norm": 0.7933165431022644, + "learning_rate": 6.187661717174386e-08, + "loss": 0.3945, + "num_input_tokens_seen": 59514256, + "step": 39105 + }, + { + "epoch": 1.2518404711606173, + "grad_norm": 0.8860942721366882, + "learning_rate": 6.118823784607708e-08, + "loss": 0.4606, + "num_input_tokens_seen": 59521856, + "step": 39110 + }, + { + "epoch": 1.2520005121311055, + "grad_norm": 0.753858208656311, + "learning_rate": 6.050370439858178e-08, + "loss": 0.5206, + "num_input_tokens_seen": 59529872, + "step": 39115 + }, + { + "epoch": 1.252160553101594, + "grad_norm": 0.6991385221481323, + "learning_rate": 5.98230169348235e-08, + "loss": 0.5519, + "num_input_tokens_seen": 59537376, + "step": 39120 + }, + { + "epoch": 1.2523205940720825, + "grad_norm": 0.4986454248428345, + "learning_rate": 5.914617555977664e-08, + "loss": 0.5206, + "num_input_tokens_seen": 59545136, + "step": 39125 + }, + { + "epoch": 1.2524806350425708, + "grad_norm": 1.9988828897476196, + "learning_rate": 5.8473180377816017e-08, + "loss": 0.4732, + "num_input_tokens_seen": 59552608, + "step": 39130 + }, + { + "epoch": 1.2526406760130593, + "grad_norm": 0.4356997013092041, + "learning_rate": 5.780403149272251e-08, + "loss": 0.4719, + "num_input_tokens_seen": 59560160, + "step": 39135 + }, + { + "epoch": 1.2528007169835478, + "grad_norm": 0.6187626719474792, + "learning_rate": 5.7138729007694126e-08, + "loss": 0.4388, + "num_input_tokens_seen": 59567872, + "step": 39140 + }, + { + "epoch": 1.2529607579540363, + "grad_norm": 0.47993531823158264, + "learning_rate": 5.64772730253238e-08, + "loss": 0.4358, + "num_input_tokens_seen": 59575664, + "step": 39145 + }, + { + "epoch": 1.2531207989245248, + "grad_norm": 0.33714866638183594, + "learning_rate": 5.5819663647618814e-08, + "loss": 0.4159, + "num_input_tokens_seen": 59583280, + "step": 39150 + }, + { + "epoch": 1.253280839895013, + "grad_norm": 0.8962756991386414, + "learning_rate": 5.5165900975989723e-08, + "loss": 0.6043, + "num_input_tokens_seen": 59591312, + "step": 39155 + }, + { + "epoch": 1.2534408808655015, + "grad_norm": 0.3385743200778961, + "learning_rate": 5.451598511125311e-08, + "loss": 0.4544, + "num_input_tokens_seen": 59598784, + "step": 39160 + }, + { + "epoch": 1.25360092183599, + "grad_norm": 0.9758191704750061, + "learning_rate": 5.3869916153637124e-08, + "loss": 0.4827, + "num_input_tokens_seen": 59606288, + "step": 39165 + }, + { + "epoch": 1.2537609628064785, + "grad_norm": 1.2720075845718384, + "learning_rate": 5.322769420277318e-08, + "loss": 0.6523, + "num_input_tokens_seen": 59614160, + "step": 39170 + }, + { + "epoch": 1.2539210037769668, + "grad_norm": 1.4960057735443115, + "learning_rate": 5.258931935769873e-08, + "loss": 0.6343, + "num_input_tokens_seen": 59621360, + "step": 39175 + }, + { + "epoch": 1.2540810447474553, + "grad_norm": 0.860136091709137, + "learning_rate": 5.19547917168628e-08, + "loss": 0.5459, + "num_input_tokens_seen": 59628560, + "step": 39180 + }, + { + "epoch": 1.2542410857179438, + "grad_norm": 0.7591225504875183, + "learning_rate": 5.13241113781121e-08, + "loss": 0.3947, + "num_input_tokens_seen": 59635488, + "step": 39185 + }, + { + "epoch": 1.2544011266884323, + "grad_norm": 1.203080177307129, + "learning_rate": 5.0697278438707755e-08, + "loss": 0.621, + "num_input_tokens_seen": 59643072, + "step": 39190 + }, + { + "epoch": 1.2545611676589208, + "grad_norm": 0.5331307649612427, + "learning_rate": 5.0074292995316854e-08, + "loss": 0.5036, + "num_input_tokens_seen": 59650432, + "step": 39195 + }, + { + "epoch": 1.254721208629409, + "grad_norm": 0.5371875762939453, + "learning_rate": 4.945515514400978e-08, + "loss": 0.3358, + "num_input_tokens_seen": 59657728, + "step": 39200 + }, + { + "epoch": 1.254721208629409, + "eval_loss": 0.49419355392456055, + "eval_runtime": 532.7896, + "eval_samples_per_second": 26.063, + "eval_steps_per_second": 13.031, + "num_input_tokens_seen": 59657728, + "step": 39200 + }, + { + "epoch": 1.2548812495998976, + "grad_norm": 0.9600719213485718, + "learning_rate": 4.883986498026571e-08, + "loss": 0.4439, + "num_input_tokens_seen": 59665104, + "step": 39205 + }, + { + "epoch": 1.255041290570386, + "grad_norm": 0.5552607774734497, + "learning_rate": 4.822842259896987e-08, + "loss": 0.5043, + "num_input_tokens_seen": 59672464, + "step": 39210 + }, + { + "epoch": 1.2552013315408745, + "grad_norm": 0.8877492547035217, + "learning_rate": 4.762082809441626e-08, + "loss": 0.3612, + "num_input_tokens_seen": 59679904, + "step": 39215 + }, + { + "epoch": 1.2553613725113628, + "grad_norm": 0.9549881219863892, + "learning_rate": 4.7017081560302156e-08, + "loss": 0.6951, + "num_input_tokens_seen": 59687104, + "step": 39220 + }, + { + "epoch": 1.2555214134818513, + "grad_norm": 0.8195977210998535, + "learning_rate": 4.6417183089730866e-08, + "loss": 0.4767, + "num_input_tokens_seen": 59694608, + "step": 39225 + }, + { + "epoch": 1.2556814544523398, + "grad_norm": 1.163818597793579, + "learning_rate": 4.5821132775217265e-08, + "loss": 0.4893, + "num_input_tokens_seen": 59702176, + "step": 39230 + }, + { + "epoch": 1.2558414954228283, + "grad_norm": 0.63653564453125, + "learning_rate": 4.5228930708679504e-08, + "loss": 0.6, + "num_input_tokens_seen": 59709840, + "step": 39235 + }, + { + "epoch": 1.2560015363933168, + "grad_norm": 1.2240538597106934, + "learning_rate": 4.464057698144175e-08, + "loss": 0.4882, + "num_input_tokens_seen": 59716848, + "step": 39240 + }, + { + "epoch": 1.256161577363805, + "grad_norm": 1.6437573432922363, + "learning_rate": 4.4056071684236974e-08, + "loss": 0.3019, + "num_input_tokens_seen": 59724704, + "step": 39245 + }, + { + "epoch": 1.2563216183342936, + "grad_norm": 0.653701901435852, + "learning_rate": 4.347541490719864e-08, + "loss": 0.4839, + "num_input_tokens_seen": 59732208, + "step": 39250 + }, + { + "epoch": 1.256481659304782, + "grad_norm": 1.5313618183135986, + "learning_rate": 4.2898606739877336e-08, + "loss": 0.445, + "num_input_tokens_seen": 59740128, + "step": 39255 + }, + { + "epoch": 1.2566417002752703, + "grad_norm": 1.284820556640625, + "learning_rate": 4.232564727122135e-08, + "loss": 0.5147, + "num_input_tokens_seen": 59747952, + "step": 39260 + }, + { + "epoch": 1.2568017412457588, + "grad_norm": 0.9770386815071106, + "learning_rate": 4.1756536589585004e-08, + "loss": 0.4432, + "num_input_tokens_seen": 59755712, + "step": 39265 + }, + { + "epoch": 1.2569617822162473, + "grad_norm": 0.6086365580558777, + "learning_rate": 4.119127478273976e-08, + "loss": 0.5624, + "num_input_tokens_seen": 59763216, + "step": 39270 + }, + { + "epoch": 1.2571218231867358, + "grad_norm": 0.7430784702301025, + "learning_rate": 4.062986193784923e-08, + "loss": 0.6003, + "num_input_tokens_seen": 59770736, + "step": 39275 + }, + { + "epoch": 1.2572818641572243, + "grad_norm": 0.7547402381896973, + "learning_rate": 4.007229814149416e-08, + "loss": 0.5552, + "num_input_tokens_seen": 59778112, + "step": 39280 + }, + { + "epoch": 1.2574419051277128, + "grad_norm": 1.2399901151657104, + "learning_rate": 3.951858347965576e-08, + "loss": 0.4347, + "num_input_tokens_seen": 59785792, + "step": 39285 + }, + { + "epoch": 1.257601946098201, + "grad_norm": 0.49320170283317566, + "learning_rate": 3.896871803772684e-08, + "loss": 0.3788, + "num_input_tokens_seen": 59793168, + "step": 39290 + }, + { + "epoch": 1.2577619870686896, + "grad_norm": 0.7478191256523132, + "learning_rate": 3.842270190050068e-08, + "loss": 0.3827, + "num_input_tokens_seen": 59801264, + "step": 39295 + }, + { + "epoch": 1.257922028039178, + "grad_norm": 0.6046276688575745, + "learning_rate": 3.7880535152179376e-08, + "loss": 0.5289, + "num_input_tokens_seen": 59808944, + "step": 39300 + }, + { + "epoch": 1.2580820690096663, + "grad_norm": 0.6118161678314209, + "learning_rate": 3.734221787637382e-08, + "loss": 0.3069, + "num_input_tokens_seen": 59816400, + "step": 39305 + }, + { + "epoch": 1.2582421099801548, + "grad_norm": 0.9343571662902832, + "learning_rate": 3.680775015609817e-08, + "loss": 0.515, + "num_input_tokens_seen": 59824064, + "step": 39310 + }, + { + "epoch": 1.2584021509506433, + "grad_norm": 0.5609239935874939, + "learning_rate": 3.627713207377537e-08, + "loss": 0.4205, + "num_input_tokens_seen": 59831984, + "step": 39315 + }, + { + "epoch": 1.2585621919211318, + "grad_norm": 0.642283022403717, + "learning_rate": 3.575036371123164e-08, + "loss": 0.4685, + "num_input_tokens_seen": 59839376, + "step": 39320 + }, + { + "epoch": 1.2587222328916203, + "grad_norm": 1.5787476301193237, + "learning_rate": 3.5227445149704776e-08, + "loss": 0.4378, + "num_input_tokens_seen": 59847552, + "step": 39325 + }, + { + "epoch": 1.2588822738621088, + "grad_norm": 1.1417194604873657, + "learning_rate": 3.470837646983027e-08, + "loss": 0.5166, + "num_input_tokens_seen": 59854976, + "step": 39330 + }, + { + "epoch": 1.259042314832597, + "grad_norm": 1.2965853214263916, + "learning_rate": 3.419315775165799e-08, + "loss": 0.4067, + "num_input_tokens_seen": 59862544, + "step": 39335 + }, + { + "epoch": 1.2592023558030856, + "grad_norm": 0.7650753855705261, + "learning_rate": 3.368178907464103e-08, + "loss": 0.4615, + "num_input_tokens_seen": 59870064, + "step": 39340 + }, + { + "epoch": 1.259362396773574, + "grad_norm": 0.3403717279434204, + "learning_rate": 3.317427051763855e-08, + "loss": 0.4058, + "num_input_tokens_seen": 59877776, + "step": 39345 + }, + { + "epoch": 1.2595224377440624, + "grad_norm": 1.0914734601974487, + "learning_rate": 3.267060215891571e-08, + "loss": 0.4427, + "num_input_tokens_seen": 59884976, + "step": 39350 + }, + { + "epoch": 1.2596824787145509, + "grad_norm": 0.7790363430976868, + "learning_rate": 3.217078407614649e-08, + "loss": 0.3021, + "num_input_tokens_seen": 59892288, + "step": 39355 + }, + { + "epoch": 1.2598425196850394, + "grad_norm": 2.383838176727295, + "learning_rate": 3.1674816346405345e-08, + "loss": 0.864, + "num_input_tokens_seen": 59899728, + "step": 39360 + }, + { + "epoch": 1.2600025606555278, + "grad_norm": 0.9943415522575378, + "learning_rate": 3.11826990461811e-08, + "loss": 0.5172, + "num_input_tokens_seen": 59907072, + "step": 39365 + }, + { + "epoch": 1.2601626016260163, + "grad_norm": 0.5315088629722595, + "learning_rate": 3.069443225136304e-08, + "loss": 0.3769, + "num_input_tokens_seen": 59914448, + "step": 39370 + }, + { + "epoch": 1.2603226425965048, + "grad_norm": 0.6541149020195007, + "learning_rate": 3.021001603724372e-08, + "loss": 0.4624, + "num_input_tokens_seen": 59922032, + "step": 39375 + }, + { + "epoch": 1.2604826835669931, + "grad_norm": 0.559267520904541, + "learning_rate": 2.9729450478532818e-08, + "loss": 0.3988, + "num_input_tokens_seen": 59929888, + "step": 39380 + }, + { + "epoch": 1.2606427245374816, + "grad_norm": 0.9604151248931885, + "learning_rate": 2.9252735649337726e-08, + "loss": 0.4159, + "num_input_tokens_seen": 59937264, + "step": 39385 + }, + { + "epoch": 1.26080276550797, + "grad_norm": 0.5206674933433533, + "learning_rate": 2.8779871623171863e-08, + "loss": 0.4945, + "num_input_tokens_seen": 59944960, + "step": 39390 + }, + { + "epoch": 1.2609628064784584, + "grad_norm": 1.0575566291809082, + "learning_rate": 2.8310858472957448e-08, + "loss": 0.4167, + "num_input_tokens_seen": 59952560, + "step": 39395 + }, + { + "epoch": 1.2611228474489469, + "grad_norm": 0.3766443431377411, + "learning_rate": 2.784569627101996e-08, + "loss": 0.4092, + "num_input_tokens_seen": 59960256, + "step": 39400 + }, + { + "epoch": 1.2611228474489469, + "eval_loss": 0.4941261410713196, + "eval_runtime": 532.8509, + "eval_samples_per_second": 26.06, + "eval_steps_per_second": 13.03, + "num_input_tokens_seen": 59960256, + "step": 39400 + }, + { + "epoch": 1.2612828884194354, + "grad_norm": 0.3943004310131073, + "learning_rate": 2.738438508909924e-08, + "loss": 0.4527, + "num_input_tokens_seen": 59967504, + "step": 39405 + }, + { + "epoch": 1.2614429293899239, + "grad_norm": 0.7660521864891052, + "learning_rate": 2.692692499833005e-08, + "loss": 0.5693, + "num_input_tokens_seen": 59976048, + "step": 39410 + }, + { + "epoch": 1.2616029703604124, + "grad_norm": 0.5469510555267334, + "learning_rate": 2.647331606926151e-08, + "loss": 0.4426, + "num_input_tokens_seen": 59983792, + "step": 39415 + }, + { + "epoch": 1.2617630113309006, + "grad_norm": 0.6682594418525696, + "learning_rate": 2.6023558371843225e-08, + "loss": 0.5524, + "num_input_tokens_seen": 59991520, + "step": 39420 + }, + { + "epoch": 1.2619230523013891, + "grad_norm": 0.9578883647918701, + "learning_rate": 2.557765197543638e-08, + "loss": 0.5612, + "num_input_tokens_seen": 59999088, + "step": 39425 + }, + { + "epoch": 1.2620830932718776, + "grad_norm": 0.5657016038894653, + "learning_rate": 2.513559694880263e-08, + "loss": 0.4356, + "num_input_tokens_seen": 60006400, + "step": 39430 + }, + { + "epoch": 1.2622431342423661, + "grad_norm": 1.1851049661636353, + "learning_rate": 2.469739336011523e-08, + "loss": 0.5217, + "num_input_tokens_seen": 60013968, + "step": 39435 + }, + { + "epoch": 1.2624031752128544, + "grad_norm": 0.7071923017501831, + "learning_rate": 2.4263041276947894e-08, + "loss": 0.4359, + "num_input_tokens_seen": 60021248, + "step": 39440 + }, + { + "epoch": 1.2625632161833429, + "grad_norm": 1.1414533853530884, + "learning_rate": 2.3832540766283164e-08, + "loss": 0.6523, + "num_input_tokens_seen": 60028672, + "step": 39445 + }, + { + "epoch": 1.2627232571538314, + "grad_norm": 0.834797203540802, + "learning_rate": 2.3405891894512366e-08, + "loss": 0.3869, + "num_input_tokens_seen": 60036336, + "step": 39450 + }, + { + "epoch": 1.2628832981243199, + "grad_norm": 0.8581787347793579, + "learning_rate": 2.29830947274301e-08, + "loss": 0.4403, + "num_input_tokens_seen": 60044416, + "step": 39455 + }, + { + "epoch": 1.2630433390948084, + "grad_norm": 1.1683703660964966, + "learning_rate": 2.2564149330231432e-08, + "loss": 0.5175, + "num_input_tokens_seen": 60051776, + "step": 39460 + }, + { + "epoch": 1.2632033800652966, + "grad_norm": 0.6412839293479919, + "learning_rate": 2.2149055767528572e-08, + "loss": 0.3907, + "num_input_tokens_seen": 60058960, + "step": 39465 + }, + { + "epoch": 1.2633634210357851, + "grad_norm": 0.665162205696106, + "learning_rate": 2.1737814103334197e-08, + "loss": 0.468, + "num_input_tokens_seen": 60066688, + "step": 39470 + }, + { + "epoch": 1.2635234620062736, + "grad_norm": 0.75398188829422, + "learning_rate": 2.1330424401064253e-08, + "loss": 0.3811, + "num_input_tokens_seen": 60074048, + "step": 39475 + }, + { + "epoch": 1.2636835029767621, + "grad_norm": 1.2957924604415894, + "learning_rate": 2.092688672354348e-08, + "loss": 0.6811, + "num_input_tokens_seen": 60081552, + "step": 39480 + }, + { + "epoch": 1.2638435439472504, + "grad_norm": 0.6423112750053406, + "learning_rate": 2.0527201133005435e-08, + "loss": 0.4395, + "num_input_tokens_seen": 60089472, + "step": 39485 + }, + { + "epoch": 1.264003584917739, + "grad_norm": 0.3818736970424652, + "learning_rate": 2.0131367691084148e-08, + "loss": 0.4707, + "num_input_tokens_seen": 60097120, + "step": 39490 + }, + { + "epoch": 1.2641636258882274, + "grad_norm": 0.6290951371192932, + "learning_rate": 1.9739386458819675e-08, + "loss": 0.3687, + "num_input_tokens_seen": 60105024, + "step": 39495 + }, + { + "epoch": 1.2643236668587159, + "grad_norm": 1.0756795406341553, + "learning_rate": 1.9351257496666442e-08, + "loss": 0.3866, + "num_input_tokens_seen": 60113408, + "step": 39500 + }, + { + "epoch": 1.2644837078292044, + "grad_norm": 0.8422448635101318, + "learning_rate": 1.896698086447657e-08, + "loss": 0.4692, + "num_input_tokens_seen": 60121216, + "step": 39505 + }, + { + "epoch": 1.2646437487996927, + "grad_norm": 0.3586011528968811, + "learning_rate": 1.8586556621505436e-08, + "loss": 0.4397, + "num_input_tokens_seen": 60128688, + "step": 39510 + }, + { + "epoch": 1.2648037897701812, + "grad_norm": 0.41117656230926514, + "learning_rate": 1.820998482642833e-08, + "loss": 0.4034, + "num_input_tokens_seen": 60135904, + "step": 39515 + }, + { + "epoch": 1.2649638307406696, + "grad_norm": 0.8605940937995911, + "learning_rate": 1.7837265537309912e-08, + "loss": 0.397, + "num_input_tokens_seen": 60143696, + "step": 39520 + }, + { + "epoch": 1.265123871711158, + "grad_norm": 1.1266188621520996, + "learning_rate": 1.7468398811629206e-08, + "loss": 0.5366, + "num_input_tokens_seen": 60151280, + "step": 39525 + }, + { + "epoch": 1.2652839126816464, + "grad_norm": 0.7428194284439087, + "learning_rate": 1.710338470627404e-08, + "loss": 0.361, + "num_input_tokens_seen": 60158768, + "step": 39530 + }, + { + "epoch": 1.265443953652135, + "grad_norm": 0.5505408644676208, + "learning_rate": 1.6742223277529945e-08, + "loss": 0.7301, + "num_input_tokens_seen": 60165792, + "step": 39535 + }, + { + "epoch": 1.2656039946226234, + "grad_norm": 0.5568205118179321, + "learning_rate": 1.6384914581094036e-08, + "loss": 0.4109, + "num_input_tokens_seen": 60173712, + "step": 39540 + }, + { + "epoch": 1.265764035593112, + "grad_norm": 0.5543830990791321, + "learning_rate": 1.6031458672069455e-08, + "loss": 0.5, + "num_input_tokens_seen": 60181392, + "step": 39545 + }, + { + "epoch": 1.2659240765636004, + "grad_norm": 0.6526793241500854, + "learning_rate": 1.5681855604962602e-08, + "loss": 0.436, + "num_input_tokens_seen": 60188688, + "step": 39550 + }, + { + "epoch": 1.2660841175340887, + "grad_norm": 0.870580792427063, + "learning_rate": 1.5336105433683135e-08, + "loss": 0.652, + "num_input_tokens_seen": 60196480, + "step": 39555 + }, + { + "epoch": 1.2662441585045772, + "grad_norm": 1.0578991174697876, + "learning_rate": 1.499420821155506e-08, + "loss": 0.4101, + "num_input_tokens_seen": 60204256, + "step": 39560 + }, + { + "epoch": 1.2664041994750657, + "grad_norm": 0.6466372013092041, + "learning_rate": 1.4656163991302874e-08, + "loss": 0.2912, + "num_input_tokens_seen": 60211664, + "step": 39565 + }, + { + "epoch": 1.266564240445554, + "grad_norm": 0.3347301185131073, + "learning_rate": 1.4321972825051544e-08, + "loss": 0.4412, + "num_input_tokens_seen": 60219904, + "step": 39570 + }, + { + "epoch": 1.2667242814160424, + "grad_norm": 0.7771092057228088, + "learning_rate": 1.3991634764345951e-08, + "loss": 0.4049, + "num_input_tokens_seen": 60227456, + "step": 39575 + }, + { + "epoch": 1.266884322386531, + "grad_norm": 0.8969974517822266, + "learning_rate": 1.3665149860120352e-08, + "loss": 0.4346, + "num_input_tokens_seen": 60235504, + "step": 39580 + }, + { + "epoch": 1.2670443633570194, + "grad_norm": 0.7718308568000793, + "learning_rate": 1.3342518162728912e-08, + "loss": 0.3937, + "num_input_tokens_seen": 60243024, + "step": 39585 + }, + { + "epoch": 1.267204404327508, + "grad_norm": 0.4813597798347473, + "learning_rate": 1.30237397219235e-08, + "loss": 0.3499, + "num_input_tokens_seen": 60250656, + "step": 39590 + }, + { + "epoch": 1.2673644452979964, + "grad_norm": 1.774098515510559, + "learning_rate": 1.2708814586862016e-08, + "loss": 0.6706, + "num_input_tokens_seen": 60257904, + "step": 39595 + }, + { + "epoch": 1.2675244862684847, + "grad_norm": 0.46119382977485657, + "learning_rate": 1.2397742806111168e-08, + "loss": 0.3274, + "num_input_tokens_seen": 60265552, + "step": 39600 + }, + { + "epoch": 1.2675244862684847, + "eval_loss": 0.4941261410713196, + "eval_runtime": 532.7663, + "eval_samples_per_second": 26.064, + "eval_steps_per_second": 13.032, + "num_input_tokens_seen": 60265552, + "step": 39600 + }, + { + "epoch": 1.2676845272389732, + "grad_norm": 0.34976816177368164, + "learning_rate": 1.209052442764369e-08, + "loss": 0.3374, + "num_input_tokens_seen": 60273408, + "step": 39605 + }, + { + "epoch": 1.2678445682094617, + "grad_norm": 0.6173710823059082, + "learning_rate": 1.17871594988328e-08, + "loss": 0.5062, + "num_input_tokens_seen": 60280784, + "step": 39610 + }, + { + "epoch": 1.26800460917995, + "grad_norm": 0.5027252435684204, + "learning_rate": 1.1487648066466072e-08, + "loss": 0.3611, + "num_input_tokens_seen": 60288240, + "step": 39615 + }, + { + "epoch": 1.2681646501504384, + "grad_norm": 0.6729545593261719, + "learning_rate": 1.1191990176728784e-08, + "loss": 0.6453, + "num_input_tokens_seen": 60295504, + "step": 39620 + }, + { + "epoch": 1.268324691120927, + "grad_norm": 0.5245463848114014, + "learning_rate": 1.0900185875215018e-08, + "loss": 0.437, + "num_input_tokens_seen": 60302944, + "step": 39625 + }, + { + "epoch": 1.2684847320914154, + "grad_norm": 0.3794570863246918, + "learning_rate": 1.0612235206924891e-08, + "loss": 0.2737, + "num_input_tokens_seen": 60310416, + "step": 39630 + }, + { + "epoch": 1.268644773061904, + "grad_norm": 0.7972590923309326, + "learning_rate": 1.0328138216264549e-08, + "loss": 0.3417, + "num_input_tokens_seen": 60317984, + "step": 39635 + }, + { + "epoch": 1.2688048140323924, + "grad_norm": 1.3603054285049438, + "learning_rate": 1.004789494704339e-08, + "loss": 0.5801, + "num_input_tokens_seen": 60325600, + "step": 39640 + }, + { + "epoch": 1.2689648550028807, + "grad_norm": 1.2002609968185425, + "learning_rate": 9.771505442482397e-09, + "loss": 0.6833, + "num_input_tokens_seen": 60332912, + "step": 39645 + }, + { + "epoch": 1.2691248959733692, + "grad_norm": 0.4872257709503174, + "learning_rate": 9.498969745200259e-09, + "loss": 0.5499, + "num_input_tokens_seen": 60341056, + "step": 39650 + }, + { + "epoch": 1.2692849369438577, + "grad_norm": 0.6495395302772522, + "learning_rate": 9.230287897230017e-09, + "loss": 0.4715, + "num_input_tokens_seen": 60348512, + "step": 39655 + }, + { + "epoch": 1.269444977914346, + "grad_norm": 1.0378756523132324, + "learning_rate": 8.965459940002419e-09, + "loss": 0.4259, + "num_input_tokens_seen": 60356080, + "step": 39660 + }, + { + "epoch": 1.2696050188848345, + "grad_norm": 1.3022998571395874, + "learning_rate": 8.704485914357019e-09, + "loss": 0.5032, + "num_input_tokens_seen": 60363552, + "step": 39665 + }, + { + "epoch": 1.269765059855323, + "grad_norm": 0.7891783714294434, + "learning_rate": 8.447365860539402e-09, + "loss": 0.397, + "num_input_tokens_seen": 60371280, + "step": 39670 + }, + { + "epoch": 1.2699251008258114, + "grad_norm": 1.2375876903533936, + "learning_rate": 8.194099818201184e-09, + "loss": 0.5856, + "num_input_tokens_seen": 60379040, + "step": 39675 + }, + { + "epoch": 1.2700851417963, + "grad_norm": 1.0179442167282104, + "learning_rate": 7.944687826400011e-09, + "loss": 0.6218, + "num_input_tokens_seen": 60386528, + "step": 39680 + }, + { + "epoch": 1.2702451827667882, + "grad_norm": 0.44306135177612305, + "learning_rate": 7.699129923599557e-09, + "loss": 0.4863, + "num_input_tokens_seen": 60394224, + "step": 39685 + }, + { + "epoch": 1.2704052237372767, + "grad_norm": 0.7522984743118286, + "learning_rate": 7.457426147663982e-09, + "loss": 0.3782, + "num_input_tokens_seen": 60401792, + "step": 39690 + }, + { + "epoch": 1.2705652647077652, + "grad_norm": 0.5071145296096802, + "learning_rate": 7.219576535871797e-09, + "loss": 0.5273, + "num_input_tokens_seen": 60409024, + "step": 39695 + }, + { + "epoch": 1.2707253056782537, + "grad_norm": 1.6600602865219116, + "learning_rate": 6.985581124896445e-09, + "loss": 0.511, + "num_input_tokens_seen": 60416000, + "step": 39700 + }, + { + "epoch": 1.270885346648742, + "grad_norm": 1.1291747093200684, + "learning_rate": 6.755439950828501e-09, + "loss": 0.4961, + "num_input_tokens_seen": 60423808, + "step": 39705 + }, + { + "epoch": 1.2710453876192305, + "grad_norm": 0.8687453269958496, + "learning_rate": 6.5291530491562444e-09, + "loss": 0.5448, + "num_input_tokens_seen": 60431520, + "step": 39710 + }, + { + "epoch": 1.271205428589719, + "grad_norm": 0.6448242664337158, + "learning_rate": 6.3067204547739845e-09, + "loss": 0.6854, + "num_input_tokens_seen": 60438576, + "step": 39715 + }, + { + "epoch": 1.2713654695602075, + "grad_norm": 0.6766575574874878, + "learning_rate": 6.088142201987612e-09, + "loss": 0.5586, + "num_input_tokens_seen": 60446400, + "step": 39720 + }, + { + "epoch": 1.271525510530696, + "grad_norm": 0.818651556968689, + "learning_rate": 5.873418324503499e-09, + "loss": 0.534, + "num_input_tokens_seen": 60453472, + "step": 39725 + }, + { + "epoch": 1.2716855515011842, + "grad_norm": 1.038204550743103, + "learning_rate": 5.6625488554340465e-09, + "loss": 0.5117, + "num_input_tokens_seen": 60460576, + "step": 39730 + }, + { + "epoch": 1.2718455924716727, + "grad_norm": 0.6020045876502991, + "learning_rate": 5.455533827297688e-09, + "loss": 0.3857, + "num_input_tokens_seen": 60468416, + "step": 39735 + }, + { + "epoch": 1.2720056334421612, + "grad_norm": 0.7331849336624146, + "learning_rate": 5.252373272018885e-09, + "loss": 0.5418, + "num_input_tokens_seen": 60476128, + "step": 39740 + }, + { + "epoch": 1.2721656744126497, + "grad_norm": 0.9977552890777588, + "learning_rate": 5.053067220925356e-09, + "loss": 0.5357, + "num_input_tokens_seen": 60483776, + "step": 39745 + }, + { + "epoch": 1.272325715383138, + "grad_norm": 1.915889024734497, + "learning_rate": 4.857615704759177e-09, + "loss": 0.557, + "num_input_tokens_seen": 60491248, + "step": 39750 + }, + { + "epoch": 1.2724857563536265, + "grad_norm": 0.9867902994155884, + "learning_rate": 4.666018753654577e-09, + "loss": 0.3927, + "num_input_tokens_seen": 60498992, + "step": 39755 + }, + { + "epoch": 1.272645797324115, + "grad_norm": 1.9949606657028198, + "learning_rate": 4.478276397162917e-09, + "loss": 0.4745, + "num_input_tokens_seen": 60506496, + "step": 39760 + }, + { + "epoch": 1.2728058382946035, + "grad_norm": 1.09710693359375, + "learning_rate": 4.294388664233262e-09, + "loss": 0.5125, + "num_input_tokens_seen": 60513856, + "step": 39765 + }, + { + "epoch": 1.272965879265092, + "grad_norm": 0.9497220516204834, + "learning_rate": 4.114355583223484e-09, + "loss": 0.5183, + "num_input_tokens_seen": 60520832, + "step": 39770 + }, + { + "epoch": 1.2731259202355802, + "grad_norm": 0.8334598541259766, + "learning_rate": 3.9381771818974845e-09, + "loss": 0.4796, + "num_input_tokens_seen": 60528064, + "step": 39775 + }, + { + "epoch": 1.2732859612060687, + "grad_norm": 0.6566285490989685, + "learning_rate": 3.765853487427973e-09, + "loss": 0.4709, + "num_input_tokens_seen": 60535920, + "step": 39780 + }, + { + "epoch": 1.2734460021765572, + "grad_norm": 0.7993645668029785, + "learning_rate": 3.5973845263825857e-09, + "loss": 0.4565, + "num_input_tokens_seen": 60543776, + "step": 39785 + }, + { + "epoch": 1.2736060431470455, + "grad_norm": 0.6362599730491638, + "learning_rate": 3.4327703247488684e-09, + "loss": 0.4741, + "num_input_tokens_seen": 60551728, + "step": 39790 + }, + { + "epoch": 1.273766084117534, + "grad_norm": 0.6357427835464478, + "learning_rate": 3.2720109079037443e-09, + "loss": 0.4521, + "num_input_tokens_seen": 60559376, + "step": 39795 + }, + { + "epoch": 1.2739261250880225, + "grad_norm": 0.6090737581253052, + "learning_rate": 3.1151063006468193e-09, + "loss": 0.5532, + "num_input_tokens_seen": 60567296, + "step": 39800 + }, + { + "epoch": 1.2739261250880225, + "eval_loss": 0.4941261410713196, + "eval_runtime": 533.7882, + "eval_samples_per_second": 26.014, + "eval_steps_per_second": 13.007, + "num_input_tokens_seen": 60567296, + "step": 39800 + }, + { + "epoch": 1.274086166058511, + "grad_norm": 0.6478419899940491, + "learning_rate": 2.962056527169854e-09, + "loss": 0.7249, + "num_input_tokens_seen": 60574960, + "step": 39805 + }, + { + "epoch": 1.2742462070289995, + "grad_norm": 0.5582372546195984, + "learning_rate": 2.8128616110761898e-09, + "loss": 0.4993, + "num_input_tokens_seen": 60582544, + "step": 39810 + }, + { + "epoch": 1.274406247999488, + "grad_norm": 0.7331055402755737, + "learning_rate": 2.6675215753724223e-09, + "loss": 0.4706, + "num_input_tokens_seen": 60589536, + "step": 39815 + }, + { + "epoch": 1.2745662889699763, + "grad_norm": 1.1746819019317627, + "learning_rate": 2.5260364424739557e-09, + "loss": 0.5156, + "num_input_tokens_seen": 60597408, + "step": 39820 + }, + { + "epoch": 1.2747263299404648, + "grad_norm": 0.7763099670410156, + "learning_rate": 2.3884062341994475e-09, + "loss": 0.5791, + "num_input_tokens_seen": 60605056, + "step": 39825 + }, + { + "epoch": 1.2748863709109532, + "grad_norm": 0.5453096032142639, + "learning_rate": 2.25463097177081e-09, + "loss": 0.4739, + "num_input_tokens_seen": 60612912, + "step": 39830 + }, + { + "epoch": 1.2750464118814415, + "grad_norm": 1.0549918413162231, + "learning_rate": 2.1247106758215397e-09, + "loss": 0.3847, + "num_input_tokens_seen": 60620528, + "step": 39835 + }, + { + "epoch": 1.27520645285193, + "grad_norm": 0.7667837142944336, + "learning_rate": 1.998645366382834e-09, + "loss": 0.3403, + "num_input_tokens_seen": 60628464, + "step": 39840 + }, + { + "epoch": 1.2753664938224185, + "grad_norm": 0.8489201664924622, + "learning_rate": 1.876435062897475e-09, + "loss": 0.61, + "num_input_tokens_seen": 60636560, + "step": 39845 + }, + { + "epoch": 1.275526534792907, + "grad_norm": 0.7860630750656128, + "learning_rate": 1.758079784211497e-09, + "loss": 0.4145, + "num_input_tokens_seen": 60644256, + "step": 39850 + }, + { + "epoch": 1.2756865757633955, + "grad_norm": 0.8884113430976868, + "learning_rate": 1.6435795485797434e-09, + "loss": 0.3767, + "num_input_tokens_seen": 60651776, + "step": 39855 + }, + { + "epoch": 1.275846616733884, + "grad_norm": 0.7042926549911499, + "learning_rate": 1.5329343736547596e-09, + "loss": 0.5245, + "num_input_tokens_seen": 60659248, + "step": 39860 + }, + { + "epoch": 1.2760066577043723, + "grad_norm": 0.6575899124145508, + "learning_rate": 1.4261442765006739e-09, + "loss": 0.3565, + "num_input_tokens_seen": 60666448, + "step": 39865 + }, + { + "epoch": 1.2761666986748608, + "grad_norm": 0.6868808269500732, + "learning_rate": 1.3232092735876445e-09, + "loss": 0.4698, + "num_input_tokens_seen": 60674416, + "step": 39870 + }, + { + "epoch": 1.2763267396453493, + "grad_norm": 0.7052068114280701, + "learning_rate": 1.2241293807918607e-09, + "loss": 0.3379, + "num_input_tokens_seen": 60682288, + "step": 39875 + }, + { + "epoch": 1.2764867806158375, + "grad_norm": 1.2854713201522827, + "learning_rate": 1.128904613387216e-09, + "loss": 0.5358, + "num_input_tokens_seen": 60690736, + "step": 39880 + }, + { + "epoch": 1.276646821586326, + "grad_norm": 1.2108761072158813, + "learning_rate": 1.0375349860591853e-09, + "loss": 0.486, + "num_input_tokens_seen": 60698176, + "step": 39885 + }, + { + "epoch": 1.2768068625568145, + "grad_norm": 0.9437669515609741, + "learning_rate": 9.5002051290205e-10, + "loss": 0.3751, + "num_input_tokens_seen": 60705744, + "step": 39890 + }, + { + "epoch": 1.276966903527303, + "grad_norm": 0.6431964039802551, + "learning_rate": 8.663612074077954e-10, + "loss": 0.5697, + "num_input_tokens_seen": 60713024, + "step": 39895 + }, + { + "epoch": 1.2771269444977915, + "grad_norm": 0.6932287216186523, + "learning_rate": 7.865570824799884e-10, + "loss": 0.4348, + "num_input_tokens_seen": 60720928, + "step": 39900 + }, + { + "epoch": 1.27728698546828, + "grad_norm": 0.6251325607299805, + "learning_rate": 7.106081504254514e-10, + "loss": 0.3818, + "num_input_tokens_seen": 60728192, + "step": 39905 + }, + { + "epoch": 1.2774470264387683, + "grad_norm": 0.508618950843811, + "learning_rate": 6.385144229570372e-10, + "loss": 0.3227, + "num_input_tokens_seen": 60735520, + "step": 39910 + }, + { + "epoch": 1.2776070674092568, + "grad_norm": 0.6808803677558899, + "learning_rate": 5.70275911190854e-10, + "loss": 0.499, + "num_input_tokens_seen": 60743168, + "step": 39915 + }, + { + "epoch": 1.2777671083797453, + "grad_norm": 0.9249085783958435, + "learning_rate": 5.058926256490403e-10, + "loss": 0.4142, + "num_input_tokens_seen": 60750496, + "step": 39920 + }, + { + "epoch": 1.2779271493502335, + "grad_norm": 1.0031681060791016, + "learning_rate": 4.4536457626254134e-10, + "loss": 0.3717, + "num_input_tokens_seen": 60757872, + "step": 39925 + }, + { + "epoch": 1.278087190320722, + "grad_norm": 0.618643045425415, + "learning_rate": 3.88691772365557e-10, + "loss": 0.5995, + "num_input_tokens_seen": 60765376, + "step": 39930 + }, + { + "epoch": 1.2782472312912105, + "grad_norm": 0.4715254604816437, + "learning_rate": 3.358742226955425e-10, + "loss": 0.3923, + "num_input_tokens_seen": 60772912, + "step": 39935 + }, + { + "epoch": 1.278407272261699, + "grad_norm": 0.6556099057197571, + "learning_rate": 2.8691193539875925e-10, + "loss": 0.6435, + "num_input_tokens_seen": 60780528, + "step": 39940 + }, + { + "epoch": 1.2785673132321875, + "grad_norm": 1.8113346099853516, + "learning_rate": 2.418049180274995e-10, + "loss": 0.572, + "num_input_tokens_seen": 60788272, + "step": 39945 + }, + { + "epoch": 1.2787273542026758, + "grad_norm": 0.8897580504417419, + "learning_rate": 2.005531775373104e-10, + "loss": 0.5002, + "num_input_tokens_seen": 60795648, + "step": 39950 + }, + { + "epoch": 1.2788873951731643, + "grad_norm": 0.6359497308731079, + "learning_rate": 1.6315672028699435e-10, + "loss": 0.6333, + "num_input_tokens_seen": 60803360, + "step": 39955 + }, + { + "epoch": 1.2790474361436528, + "grad_norm": 0.4078029692173004, + "learning_rate": 1.2961555204693555e-10, + "loss": 0.4447, + "num_input_tokens_seen": 60810480, + "step": 39960 + }, + { + "epoch": 1.2792074771141413, + "grad_norm": 0.9092631340026855, + "learning_rate": 9.992967798799768e-11, + "loss": 0.4968, + "num_input_tokens_seen": 60818128, + "step": 39965 + }, + { + "epoch": 1.2793675180846296, + "grad_norm": 0.7089337706565857, + "learning_rate": 7.409910268707521e-11, + "loss": 0.4906, + "num_input_tokens_seen": 60825888, + "step": 39970 + }, + { + "epoch": 1.279527559055118, + "grad_norm": 0.549818754196167, + "learning_rate": 5.212383012986877e-11, + "loss": 0.4819, + "num_input_tokens_seen": 60833536, + "step": 39975 + }, + { + "epoch": 1.2796876000256066, + "grad_norm": 1.101468563079834, + "learning_rate": 3.400386370533415e-11, + "loss": 0.547, + "num_input_tokens_seen": 60840864, + "step": 39980 + }, + { + "epoch": 1.279847640996095, + "grad_norm": 0.3029453158378601, + "learning_rate": 1.9739206205682258e-11, + "loss": 0.386, + "num_input_tokens_seen": 60848336, + "step": 39985 + }, + { + "epoch": 1.2800076819665835, + "grad_norm": 1.2421717643737793, + "learning_rate": 9.329859829154685e-12, + "loss": 0.4371, + "num_input_tokens_seen": 60855632, + "step": 39990 + }, + { + "epoch": 1.2801677229370718, + "grad_norm": 0.5356718301773071, + "learning_rate": 2.7758261855748148e-12, + "loss": 0.3467, + "num_input_tokens_seen": 60862976, + "step": 39995 + }, + { + "epoch": 1.2803277639075603, + "grad_norm": 1.062281847000122, + "learning_rate": 7.710628524559838e-14, + "loss": 0.4226, + "num_input_tokens_seen": 60870656, + "step": 40000 + }, + { + "epoch": 1.2803277639075603, + "eval_loss": 0.4941261410713196, + "eval_runtime": 533.7591, + "eval_samples_per_second": 26.015, + "eval_steps_per_second": 13.008, + "num_input_tokens_seen": 60870656, + "step": 40000 + }, + { + "epoch": 1.2803277639075603, + "num_input_tokens_seen": 60870656, + "step": 40000, + "total_flos": 2.5982039689286124e+18, + "train_loss": 0.5386139680683613, + "train_runtime": 133028.9435, + "train_samples_per_second": 1.203, + "train_steps_per_second": 0.301 + } + ], + "logging_steps": 5, + "max_steps": 40000, + "num_input_tokens_seen": 60870656, + "num_train_epochs": 2, + "save_steps": 200, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 2.5982039689286124e+18, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}