{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 0.3, "eval_steps": 500, "global_step": 30000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001, "grad_norm": 20.30853843688965, "learning_rate": 2.97e-05, "loss": 6.3768, "num_input_tokens_seen": 6553600, "step": 100, "train_runtime": 67.7249, "train_tokens_per_second": 96767.992 }, { "epoch": 0.002, "grad_norm": 12.960590362548828, "learning_rate": 5.97e-05, "loss": 1.5751, "num_input_tokens_seen": 13107200, "step": 200, "train_runtime": 121.8001, "train_tokens_per_second": 107612.389 }, { "epoch": 0.003, "grad_norm": 4.234757900238037, "learning_rate": 8.969999999999998e-05, "loss": 1.3419, "num_input_tokens_seen": 19660800, "step": 300, "train_runtime": 175.7236, "train_tokens_per_second": 111884.791 }, { "epoch": 0.004, "grad_norm": 2.8189597129821777, "learning_rate": 0.0001197, "loss": 1.1467, "num_input_tokens_seen": 26214400, "step": 400, "train_runtime": 229.814, "train_tokens_per_second": 114067.895 }, { "epoch": 0.005, "grad_norm": 4.593051910400391, "learning_rate": 0.00014969999999999998, "loss": 1.02, "num_input_tokens_seen": 32768000, "step": 500, "train_runtime": 288.5437, "train_tokens_per_second": 113563.389 }, { "epoch": 0.006, "grad_norm": 1.745302677154541, "learning_rate": 0.00017969999999999998, "loss": 0.9381, "num_input_tokens_seen": 39321600, "step": 600, "train_runtime": 343.6994, "train_tokens_per_second": 114406.961 }, { "epoch": 0.007, "grad_norm": 2.180215358734131, "learning_rate": 0.00020969999999999997, "loss": 0.8912, "num_input_tokens_seen": 45875200, "step": 700, "train_runtime": 398.5625, "train_tokens_per_second": 115101.641 }, { "epoch": 0.008, "grad_norm": 1.5782365798950195, "learning_rate": 0.0002397, "loss": 0.8525, "num_input_tokens_seen": 52428800, "step": 800, "train_runtime": 453.9393, "train_tokens_per_second": 115497.369 }, { "epoch": 0.009, "grad_norm": 1.306846022605896, "learning_rate": 0.0002697, "loss": 0.8411, "num_input_tokens_seen": 58982400, "step": 900, "train_runtime": 508.8336, "train_tokens_per_second": 115916.867 }, { "epoch": 0.01, "grad_norm": 0.9485424160957336, "learning_rate": 0.00029969999999999997, "loss": 0.8207, "num_input_tokens_seen": 65536000, "step": 1000, "train_runtime": 569.5888, "train_tokens_per_second": 115058.451 }, { "epoch": 0.011, "grad_norm": 0.8611180186271667, "learning_rate": 0.00029999925978027874, "loss": 0.8158, "num_input_tokens_seen": 72089600, "step": 1100, "train_runtime": 623.5912, "train_tokens_per_second": 115603.946 }, { "epoch": 0.012, "grad_norm": 0.8997637033462524, "learning_rate": 0.0002999970091452017, "loss": 0.7948, "num_input_tokens_seen": 78643200, "step": 1200, "train_runtime": 679.5203, "train_tokens_per_second": 115733.415 }, { "epoch": 0.013, "grad_norm": 0.6279132962226868, "learning_rate": 0.00029999324804190795, "loss": 0.7852, "num_input_tokens_seen": 85196800, "step": 1300, "train_runtime": 739.7161, "train_tokens_per_second": 115174.993 }, { "epoch": 0.014, "grad_norm": 0.5959413051605225, "learning_rate": 0.0002999879765082716, "loss": 0.7736, "num_input_tokens_seen": 91750400, "step": 1400, "train_runtime": 794.2065, "train_tokens_per_second": 115524.612 }, { "epoch": 0.015, "grad_norm": 0.6557334065437317, "learning_rate": 0.000299981194597377, "loss": 0.7631, "num_input_tokens_seen": 98304000, "step": 1500, "train_runtime": 847.9465, "train_tokens_per_second": 115931.844 }, { "epoch": 0.016, "grad_norm": 0.7466850876808167, "learning_rate": 0.0002999729023775179, "loss": 0.7564, "num_input_tokens_seen": 104857600, "step": 1600, "train_runtime": 907.419, "train_tokens_per_second": 115555.874 }, { "epoch": 0.017, "grad_norm": 0.6613496541976929, "learning_rate": 0.0002999630999321969, "loss": 0.7664, "num_input_tokens_seen": 111411200, "step": 1700, "train_runtime": 957.6918, "train_tokens_per_second": 116333.047 }, { "epoch": 0.018, "grad_norm": 0.49886658787727356, "learning_rate": 0.00029995178736012443, "loss": 0.7577, "num_input_tokens_seen": 117964800, "step": 1800, "train_runtime": 1017.5123, "train_tokens_per_second": 115934.516 }, { "epoch": 0.019, "grad_norm": 0.5621941089630127, "learning_rate": 0.0002999389647752181, "loss": 0.743, "num_input_tokens_seen": 124518400, "step": 1900, "train_runtime": 1070.9552, "train_tokens_per_second": 116268.538 }, { "epoch": 0.02, "grad_norm": 0.46181556582450867, "learning_rate": 0.00029992463230660104, "loss": 0.7429, "num_input_tokens_seen": 131072000, "step": 2000, "train_runtime": 1130.6338, "train_tokens_per_second": 115927.894 }, { "epoch": 0.021, "grad_norm": 0.4571639895439148, "learning_rate": 0.00029990879009860117, "loss": 0.7349, "num_input_tokens_seen": 137625600, "step": 2100, "train_runtime": 1185.0559, "train_tokens_per_second": 116134.269 }, { "epoch": 0.022, "grad_norm": 0.5650346875190735, "learning_rate": 0.0002998914383107493, "loss": 0.7373, "num_input_tokens_seen": 144179200, "step": 2200, "train_runtime": 1239.4558, "train_tokens_per_second": 116324.603 }, { "epoch": 0.023, "grad_norm": 0.5640336871147156, "learning_rate": 0.0002998725771177778, "loss": 0.7383, "num_input_tokens_seen": 150732800, "step": 2300, "train_runtime": 1294.8284, "train_tokens_per_second": 116411.407 }, { "epoch": 0.024, "grad_norm": 0.4288141429424286, "learning_rate": 0.00029985220670961847, "loss": 0.7256, "num_input_tokens_seen": 157286400, "step": 2400, "train_runtime": 1353.9338, "train_tokens_per_second": 116169.935 }, { "epoch": 0.025, "grad_norm": 0.7532113194465637, "learning_rate": 0.0002998303272914014, "loss": 0.7301, "num_input_tokens_seen": 163840000, "step": 2500, "train_runtime": 1408.3918, "train_tokens_per_second": 116331.266 }, { "epoch": 0.026, "grad_norm": 0.42602330446243286, "learning_rate": 0.00029980693908345185, "loss": 0.7238, "num_input_tokens_seen": 170393600, "step": 2600, "train_runtime": 1461.9698, "train_tokens_per_second": 116550.699 }, { "epoch": 0.027, "grad_norm": 0.655644953250885, "learning_rate": 0.00029978204232128895, "loss": 0.7213, "num_input_tokens_seen": 176947200, "step": 2700, "train_runtime": 1523.0525, "train_tokens_per_second": 116179.321 }, { "epoch": 0.028, "grad_norm": 0.3962925970554352, "learning_rate": 0.0002997556372556227, "loss": 0.7238, "num_input_tokens_seen": 183500800, "step": 2800, "train_runtime": 1577.6914, "train_tokens_per_second": 116309.692 }, { "epoch": 0.029, "grad_norm": 0.44488221406936646, "learning_rate": 0.0002997277241523519, "loss": 0.7218, "num_input_tokens_seen": 190054400, "step": 2900, "train_runtime": 1631.4556, "train_tokens_per_second": 116493.763 }, { "epoch": 0.03, "grad_norm": 0.3886267840862274, "learning_rate": 0.00029969830329256125, "loss": 0.7091, "num_input_tokens_seen": 196608000, "step": 3000, "train_runtime": 1686.3399, "train_tokens_per_second": 116588.595 }, { "epoch": 0.031, "grad_norm": 0.46353062987327576, "learning_rate": 0.00029966737497251836, "loss": 0.7132, "num_input_tokens_seen": 203161600, "step": 3100, "train_runtime": 1745.3481, "train_tokens_per_second": 116401.765 }, { "epoch": 0.032, "grad_norm": 0.45476260781288147, "learning_rate": 0.0002996349395036711, "loss": 0.7099, "num_input_tokens_seen": 209715200, "step": 3200, "train_runtime": 1799.8354, "train_tokens_per_second": 116519.096 }, { "epoch": 0.033, "grad_norm": 0.5311458706855774, "learning_rate": 0.00029960099721264435, "loss": 0.7162, "num_input_tokens_seen": 216268800, "step": 3300, "train_runtime": 1854.9316, "train_tokens_per_second": 116591.252 }, { "epoch": 0.034, "grad_norm": 0.4546497166156769, "learning_rate": 0.0002995655484412365, "loss": 0.7046, "num_input_tokens_seen": 222822400, "step": 3400, "train_runtime": 1909.1868, "train_tokens_per_second": 116710.63 }, { "epoch": 0.035, "grad_norm": 0.6349391937255859, "learning_rate": 0.00029952859354641636, "loss": 0.7073, "num_input_tokens_seen": 229376000, "step": 3500, "train_runtime": 1969.0427, "train_tokens_per_second": 116491.125 }, { "epoch": 0.036, "grad_norm": 0.493557870388031, "learning_rate": 0.00029949013290031924, "loss": 0.7021, "num_input_tokens_seen": 235929600, "step": 3600, "train_runtime": 2023.4924, "train_tokens_per_second": 116595.251 }, { "epoch": 0.037, "grad_norm": 0.3123897314071655, "learning_rate": 0.00029945016689024353, "loss": 0.696, "num_input_tokens_seen": 242483200, "step": 3700, "train_runtime": 2077.3277, "train_tokens_per_second": 116728.432 }, { "epoch": 0.038, "grad_norm": 0.43099701404571533, "learning_rate": 0.0002994086959186464, "loss": 0.6997, "num_input_tokens_seen": 249036800, "step": 3800, "train_runtime": 2132.1624, "train_tokens_per_second": 116800.109 }, { "epoch": 0.039, "grad_norm": 0.3274790346622467, "learning_rate": 0.00029936572040314014, "loss": 0.7034, "num_input_tokens_seen": 255590400, "step": 3900, "train_runtime": 2193.0795, "train_tokens_per_second": 116544.063 }, { "epoch": 0.04, "grad_norm": 0.33723658323287964, "learning_rate": 0.0002993212407764877, "loss": 0.6977, "num_input_tokens_seen": 262144000, "step": 4000, "train_runtime": 2246.9274, "train_tokens_per_second": 116667.767 }, { "epoch": 0.041, "grad_norm": 0.42106112837791443, "learning_rate": 0.00029927525748659834, "loss": 0.7078, "num_input_tokens_seen": 268697600, "step": 4100, "train_runtime": 2302.3386, "train_tokens_per_second": 116706.38 }, { "epoch": 0.042, "grad_norm": 0.3567107915878296, "learning_rate": 0.0002992277709965234, "loss": 0.6967, "num_input_tokens_seen": 275251200, "step": 4200, "train_runtime": 2356.6313, "train_tokens_per_second": 116798.585 }, { "epoch": 0.043, "grad_norm": 0.4269777834415436, "learning_rate": 0.0002991787817844513, "loss": 0.6976, "num_input_tokens_seen": 281804800, "step": 4300, "train_runtime": 2417.993, "train_tokens_per_second": 116544.92 }, { "epoch": 0.044, "grad_norm": 0.4035900831222534, "learning_rate": 0.0002991282903437028, "loss": 0.6915, "num_input_tokens_seen": 288358400, "step": 4400, "train_runtime": 2471.1514, "train_tokens_per_second": 116689.896 }, { "epoch": 0.045, "grad_norm": 0.3197948634624481, "learning_rate": 0.0002990762971827262, "loss": 0.6905, "num_input_tokens_seen": 294912000, "step": 4500, "train_runtime": 2525.6119, "train_tokens_per_second": 116768.533 }, { "epoch": 0.046, "grad_norm": 0.3118247389793396, "learning_rate": 0.00029902280282509197, "loss": 0.6959, "num_input_tokens_seen": 301465600, "step": 4600, "train_runtime": 2587.1477, "train_tokens_per_second": 116524.311 }, { "epoch": 0.047, "grad_norm": 0.38194310665130615, "learning_rate": 0.0002989678078094878, "loss": 0.6812, "num_input_tokens_seen": 308019200, "step": 4700, "train_runtime": 2645.2255, "train_tokens_per_second": 116443.456 }, { "epoch": 0.048, "grad_norm": 0.5399206876754761, "learning_rate": 0.00029891131268971284, "loss": 0.6843, "num_input_tokens_seen": 314572800, "step": 4800, "train_runtime": 2702.2839, "train_tokens_per_second": 116409.974 }, { "epoch": 0.049, "grad_norm": 0.3325801491737366, "learning_rate": 0.0002988533180346723, "loss": 0.6803, "num_input_tokens_seen": 321126400, "step": 4900, "train_runtime": 2761.723, "train_tokens_per_second": 116277.558 }, { "epoch": 0.05, "grad_norm": 0.41960838437080383, "learning_rate": 0.0002987938244283717, "loss": 0.6772, "num_input_tokens_seen": 327680000, "step": 5000, "train_runtime": 2818.6212, "train_tokens_per_second": 116255.424 }, { "epoch": 0.051, "grad_norm": 0.40982258319854736, "learning_rate": 0.00029873283246991105, "loss": 0.6861, "num_input_tokens_seen": 334233600, "step": 5100, "train_runtime": 2873.5528, "train_tokens_per_second": 116313.715 }, { "epoch": 0.052, "grad_norm": 0.31062355637550354, "learning_rate": 0.0002986703427734787, "loss": 0.6852, "num_input_tokens_seen": 340787200, "step": 5200, "train_runtime": 2929.1895, "train_tokens_per_second": 116341.805 }, { "epoch": 0.053, "grad_norm": 0.3270035684108734, "learning_rate": 0.00029860635596834517, "loss": 0.6854, "num_input_tokens_seen": 347340800, "step": 5300, "train_runtime": 2983.4815, "train_tokens_per_second": 116421.301 }, { "epoch": 0.054, "grad_norm": 0.3153812289237976, "learning_rate": 0.0002985408726988569, "loss": 0.6826, "num_input_tokens_seen": 353894400, "step": 5400, "train_runtime": 3042.2104, "train_tokens_per_second": 116328.049 }, { "epoch": 0.055, "grad_norm": 0.328072190284729, "learning_rate": 0.0002984738936244296, "loss": 0.6796, "num_input_tokens_seen": 360448000, "step": 5500, "train_runtime": 3096.1295, "train_tokens_per_second": 116418.904 }, { "epoch": 0.056, "grad_norm": 0.32828816771507263, "learning_rate": 0.0002984054194195419, "loss": 0.6927, "num_input_tokens_seen": 367001600, "step": 5600, "train_runtime": 3151.7383, "train_tokens_per_second": 116444.184 }, { "epoch": 0.057, "grad_norm": 0.3481093645095825, "learning_rate": 0.0002983354507737283, "loss": 0.6876, "num_input_tokens_seen": 373555200, "step": 5700, "train_runtime": 3207.4, "train_tokens_per_second": 116466.67 }, { "epoch": 0.058, "grad_norm": 0.4054113030433655, "learning_rate": 0.00029826398839157215, "loss": 0.6868, "num_input_tokens_seen": 380108800, "step": 5800, "train_runtime": 3267.0971, "train_tokens_per_second": 116344.508 }, { "epoch": 0.059, "grad_norm": 0.3671651780605316, "learning_rate": 0.000298191032992699, "loss": 0.6818, "num_input_tokens_seen": 386662400, "step": 5900, "train_runtime": 3320.6147, "train_tokens_per_second": 116443.019 }, { "epoch": 0.06, "grad_norm": 0.2815225422382355, "learning_rate": 0.0002981165853117688, "loss": 0.6735, "num_input_tokens_seen": 393216000, "step": 6000, "train_runtime": 3380.4372, "train_tokens_per_second": 116321.049 }, { "epoch": 0.061, "grad_norm": 0.5300637483596802, "learning_rate": 0.000298040646098469, "loss": 0.6792, "num_input_tokens_seen": 399769600, "step": 6100, "train_runtime": 3434.9665, "train_tokens_per_second": 116382.388 }, { "epoch": 0.062, "grad_norm": 0.28088632225990295, "learning_rate": 0.0002979632161175064, "loss": 0.6763, "num_input_tokens_seen": 406323200, "step": 6200, "train_runtime": 3490.2084, "train_tokens_per_second": 116418.034 }, { "epoch": 0.063, "grad_norm": 0.5535532236099243, "learning_rate": 0.0002978842961486003, "loss": 0.6769, "num_input_tokens_seen": 412876800, "step": 6300, "train_runtime": 3587.6493, "train_tokens_per_second": 115082.821 }, { "epoch": 0.064, "grad_norm": 0.3341565728187561, "learning_rate": 0.0002978038869864738, "loss": 0.6734, "num_input_tokens_seen": 419430400, "step": 6400, "train_runtime": 3650.2458, "train_tokens_per_second": 114904.702 }, { "epoch": 0.065, "grad_norm": 0.34299543499946594, "learning_rate": 0.0002977219894408463, "loss": 0.671, "num_input_tokens_seen": 425984000, "step": 6500, "train_runtime": 3704.075, "train_tokens_per_second": 115004.151 }, { "epoch": 0.066, "grad_norm": 0.3468911349773407, "learning_rate": 0.0002976386043364251, "loss": 0.6799, "num_input_tokens_seen": 432537600, "step": 6600, "train_runtime": 3758.2332, "train_tokens_per_second": 115090.677 }, { "epoch": 0.067, "grad_norm": 0.29494985938072205, "learning_rate": 0.00029755373251289733, "loss": 0.6836, "num_input_tokens_seen": 439091200, "step": 6700, "train_runtime": 3818.8992, "train_tokens_per_second": 114978.475 }, { "epoch": 0.068, "grad_norm": 0.43209826946258545, "learning_rate": 0.0002974673748249213, "loss": 0.6721, "num_input_tokens_seen": 445644800, "step": 6800, "train_runtime": 3873.8852, "train_tokens_per_second": 115038.204 }, { "epoch": 0.069, "grad_norm": 0.36356812715530396, "learning_rate": 0.00029737953214211804, "loss": 0.6742, "num_input_tokens_seen": 452198400, "step": 6900, "train_runtime": 3928.8659, "train_tokens_per_second": 115096.42 }, { "epoch": 0.07, "grad_norm": 0.4425353705883026, "learning_rate": 0.0002972902053490623, "loss": 0.6691, "num_input_tokens_seen": 458752000, "step": 7000, "train_runtime": 3982.4622, "train_tokens_per_second": 115193.058 }, { "epoch": 0.071, "grad_norm": 0.3378838300704956, "learning_rate": 0.00029719939534527393, "loss": 0.6817, "num_input_tokens_seen": 465305600, "step": 7100, "train_runtime": 4038.9423, "train_tokens_per_second": 115204.814 }, { "epoch": 0.072, "grad_norm": 0.46528419852256775, "learning_rate": 0.00029710710304520866, "loss": 0.6731, "num_input_tokens_seen": 471859200, "step": 7200, "train_runtime": 4094.7269, "train_tokens_per_second": 115235.817 }, { "epoch": 0.073, "grad_norm": 0.45732709765434265, "learning_rate": 0.00029701332937824885, "loss": 0.6667, "num_input_tokens_seen": 478412800, "step": 7300, "train_runtime": 4155.1372, "train_tokens_per_second": 115137.666 }, { "epoch": 0.074, "grad_norm": 0.28316208720207214, "learning_rate": 0.0002969180752886944, "loss": 0.6735, "num_input_tokens_seen": 484966400, "step": 7400, "train_runtime": 4208.4074, "train_tokens_per_second": 115237.512 }, { "epoch": 0.075, "grad_norm": 0.35296332836151123, "learning_rate": 0.0002968213417357529, "loss": 0.674, "num_input_tokens_seen": 491520000, "step": 7500, "train_runtime": 4262.7857, "train_tokens_per_second": 115304.882 }, { "epoch": 0.076, "grad_norm": 0.37089261412620544, "learning_rate": 0.00029672312969353015, "loss": 0.6696, "num_input_tokens_seen": 498073600, "step": 7600, "train_runtime": 4324.696, "train_tokens_per_second": 115169.622 }, { "epoch": 0.077, "grad_norm": 0.3451579809188843, "learning_rate": 0.00029662344015102027, "loss": 0.669, "num_input_tokens_seen": 504627200, "step": 7700, "train_runtime": 4380.0469, "train_tokens_per_second": 115210.456 }, { "epoch": 0.078, "grad_norm": 0.2890200912952423, "learning_rate": 0.00029652227411209594, "loss": 0.6672, "num_input_tokens_seen": 511180800, "step": 7800, "train_runtime": 4433.6829, "train_tokens_per_second": 115294.849 }, { "epoch": 0.079, "grad_norm": 0.3042987287044525, "learning_rate": 0.0002964196325954979, "loss": 0.6642, "num_input_tokens_seen": 517734400, "step": 7900, "train_runtime": 4488.6662, "train_tokens_per_second": 115342.593 }, { "epoch": 0.08, "grad_norm": 0.34017419815063477, "learning_rate": 0.0002963155166348253, "loss": 0.6689, "num_input_tokens_seen": 524288000, "step": 8000, "train_runtime": 4548.362, "train_tokens_per_second": 115269.63 }, { "epoch": 0.081, "grad_norm": 0.5584019422531128, "learning_rate": 0.0002962099272785246, "loss": 0.6703, "num_input_tokens_seen": 530841600, "step": 8100, "train_runtime": 4605.2398, "train_tokens_per_second": 115269.047 }, { "epoch": 0.082, "grad_norm": 0.2589603364467621, "learning_rate": 0.0002961028655898794, "loss": 0.6634, "num_input_tokens_seen": 537395200, "step": 8200, "train_runtime": 4660.2455, "train_tokens_per_second": 115314.783 }, { "epoch": 0.083, "grad_norm": 0.41600939631462097, "learning_rate": 0.0002959943326469998, "loss": 0.6611, "num_input_tokens_seen": 543948800, "step": 8300, "train_runtime": 4714.0847, "train_tokens_per_second": 115388.0 }, { "epoch": 0.084, "grad_norm": 0.28658175468444824, "learning_rate": 0.0002958843295428112, "loss": 0.659, "num_input_tokens_seen": 550502400, "step": 8400, "train_runtime": 4768.8365, "train_tokens_per_second": 115437.466 }, { "epoch": 0.085, "grad_norm": 0.290019154548645, "learning_rate": 0.0002957728573850438, "loss": 0.661, "num_input_tokens_seen": 557056000, "step": 8500, "train_runtime": 4828.5019, "train_tokens_per_second": 115368.289 }, { "epoch": 0.086, "grad_norm": 0.3445529341697693, "learning_rate": 0.0002956599172962209, "loss": 0.6745, "num_input_tokens_seen": 563609600, "step": 8600, "train_runtime": 4883.6807, "train_tokens_per_second": 115406.727 }, { "epoch": 0.087, "grad_norm": 0.3387536406517029, "learning_rate": 0.0002955455104136479, "loss": 0.6596, "num_input_tokens_seen": 570163200, "step": 8700, "train_runtime": 4940.0949, "train_tokens_per_second": 115415.435 }, { "epoch": 0.088, "grad_norm": 0.3031691312789917, "learning_rate": 0.00029542963788940096, "loss": 0.658, "num_input_tokens_seen": 576716800, "step": 8800, "train_runtime": 4993.542, "train_tokens_per_second": 115492.529 }, { "epoch": 0.089, "grad_norm": 0.3859446048736572, "learning_rate": 0.00029531230089031505, "loss": 0.6684, "num_input_tokens_seen": 583270400, "step": 8900, "train_runtime": 5048.437, "train_tokens_per_second": 115534.848 }, { "epoch": 0.09, "grad_norm": 0.28773999214172363, "learning_rate": 0.0002951935005979724, "loss": 0.6585, "num_input_tokens_seen": 589824000, "step": 9000, "train_runtime": 5107.8168, "train_tokens_per_second": 115474.775 }, { "epoch": 0.091, "grad_norm": 0.3639370799064636, "learning_rate": 0.0002950732382086907, "loss": 0.6573, "num_input_tokens_seen": 596377600, "step": 9100, "train_runtime": 5164.5839, "train_tokens_per_second": 115474.473 }, { "epoch": 0.092, "grad_norm": 0.3325030505657196, "learning_rate": 0.0002949515149335108, "loss": 0.6634, "num_input_tokens_seen": 602931200, "step": 9200, "train_runtime": 5219.0432, "train_tokens_per_second": 115525.237 }, { "epoch": 0.093, "grad_norm": 0.3296424448490143, "learning_rate": 0.0002948283319981848, "loss": 0.65, "num_input_tokens_seen": 609484800, "step": 9300, "train_runtime": 5273.9916, "train_tokens_per_second": 115564.234 }, { "epoch": 0.094, "grad_norm": 0.323024719953537, "learning_rate": 0.00029470369064316354, "loss": 0.6533, "num_input_tokens_seen": 616038400, "step": 9400, "train_runtime": 5327.8342, "train_tokens_per_second": 115626.421 }, { "epoch": 0.095, "grad_norm": 0.31108859181404114, "learning_rate": 0.00029457759212358397, "loss": 0.6549, "num_input_tokens_seen": 622592000, "step": 9500, "train_runtime": 5386.4463, "train_tokens_per_second": 115584.927 }, { "epoch": 0.096, "grad_norm": 0.43114957213401794, "learning_rate": 0.00029445003770925686, "loss": 0.6518, "num_input_tokens_seen": 629145600, "step": 9600, "train_runtime": 5441.6464, "train_tokens_per_second": 115616.773 }, { "epoch": 0.097, "grad_norm": 0.3269326090812683, "learning_rate": 0.00029432102868465367, "loss": 0.6539, "num_input_tokens_seen": 635699200, "step": 9700, "train_runtime": 5497.0335, "train_tokens_per_second": 115644.046 }, { "epoch": 0.098, "grad_norm": 0.3119284510612488, "learning_rate": 0.0002941905663488939, "loss": 0.6525, "num_input_tokens_seen": 642252800, "step": 9800, "train_runtime": 5557.4204, "train_tokens_per_second": 115566.712 }, { "epoch": 0.099, "grad_norm": 0.30968910455703735, "learning_rate": 0.0002940586520157318, "loss": 0.6593, "num_input_tokens_seen": 648806400, "step": 9900, "train_runtime": 5612.1781, "train_tokens_per_second": 115606.881 }, { "epoch": 0.1, "grad_norm": 0.3167921006679535, "learning_rate": 0.00029392528701354325, "loss": 0.6516, "num_input_tokens_seen": 655360000, "step": 10000, "train_runtime": 5665.7095, "train_tokens_per_second": 115671.303 }, { "epoch": 0.101, "grad_norm": 0.3000660538673401, "learning_rate": 0.00029379047268531243, "loss": 0.6579, "num_input_tokens_seen": 661913600, "step": 10100, "train_runtime": 5721.9414, "train_tokens_per_second": 115679.899 }, { "epoch": 0.102, "grad_norm": 0.30189189314842224, "learning_rate": 0.00029365421038861795, "loss": 0.6594, "num_input_tokens_seen": 668467200, "step": 10200, "train_runtime": 5777.1263, "train_tokens_per_second": 115709.294 }, { "epoch": 0.103, "grad_norm": 0.2689358592033386, "learning_rate": 0.0002935165014956198, "loss": 0.6527, "num_input_tokens_seen": 675020800, "step": 10300, "train_runtime": 5832.7222, "train_tokens_per_second": 115729.977 }, { "epoch": 0.104, "grad_norm": 0.40012845396995544, "learning_rate": 0.0002933773473930448, "loss": 0.6442, "num_input_tokens_seen": 681574400, "step": 10400, "train_runtime": 5888.393, "train_tokens_per_second": 115748.796 }, { "epoch": 0.105, "grad_norm": 0.26425933837890625, "learning_rate": 0.0002932367494821734, "loss": 0.655, "num_input_tokens_seen": 688128000, "step": 10500, "train_runtime": 5949.719, "train_tokens_per_second": 115657.227 }, { "epoch": 0.106, "grad_norm": 0.3352152705192566, "learning_rate": 0.00029309470917882497, "loss": 0.6513, "num_input_tokens_seen": 694681600, "step": 10600, "train_runtime": 6005.1104, "train_tokens_per_second": 115681.738 }, { "epoch": 0.107, "grad_norm": 0.2809161841869354, "learning_rate": 0.0002929512279133437, "loss": 0.6544, "num_input_tokens_seen": 701235200, "step": 10700, "train_runtime": 6060.3088, "train_tokens_per_second": 115709.482 }, { "epoch": 0.108, "grad_norm": 0.29120925068855286, "learning_rate": 0.0002928063071305844, "loss": 0.6502, "num_input_tokens_seen": 707788800, "step": 10800, "train_runtime": 6114.2187, "train_tokens_per_second": 115761.119 }, { "epoch": 0.109, "grad_norm": 0.312067449092865, "learning_rate": 0.0002926599482898978, "loss": 0.6495, "num_input_tokens_seen": 714342400, "step": 10900, "train_runtime": 6169.1634, "train_tokens_per_second": 115792.426 }, { "epoch": 0.11, "grad_norm": 0.4553526043891907, "learning_rate": 0.00029251215286511573, "loss": 0.6508, "num_input_tokens_seen": 720896000, "step": 11000, "train_runtime": 6230.1722, "train_tokens_per_second": 115710.445 }, { "epoch": 0.111, "grad_norm": 0.2880023419857025, "learning_rate": 0.00029236292234453647, "loss": 0.6476, "num_input_tokens_seen": 727449600, "step": 11100, "train_runtime": 6284.648, "train_tokens_per_second": 115750.253 }, { "epoch": 0.112, "grad_norm": 0.3704841136932373, "learning_rate": 0.0002922122582309097, "loss": 0.6559, "num_input_tokens_seen": 734003200, "step": 11200, "train_runtime": 6340.8669, "train_tokens_per_second": 115757.547 }, { "epoch": 0.113, "grad_norm": 0.34455180168151855, "learning_rate": 0.0002920601620414215, "loss": 0.6482, "num_input_tokens_seen": 740556800, "step": 11300, "train_runtime": 6393.5327, "train_tokens_per_second": 115829.046 }, { "epoch": 0.114, "grad_norm": 0.3323168456554413, "learning_rate": 0.0002919066353076786, "loss": 0.6489, "num_input_tokens_seen": 747110400, "step": 11400, "train_runtime": 6449.7175, "train_tokens_per_second": 115836.144 }, { "epoch": 0.115, "grad_norm": 0.4480803608894348, "learning_rate": 0.00029175167957569366, "loss": 0.6492, "num_input_tokens_seen": 753664000, "step": 11500, "train_runtime": 6503.6523, "train_tokens_per_second": 115883.194 }, { "epoch": 0.116, "grad_norm": 0.2677996754646301, "learning_rate": 0.0002915952964058691, "loss": 0.6462, "num_input_tokens_seen": 760217600, "step": 11600, "train_runtime": 6565.0474, "train_tokens_per_second": 115797.731 }, { "epoch": 0.117, "grad_norm": 0.2838340699672699, "learning_rate": 0.00029143748737298173, "loss": 0.6574, "num_input_tokens_seen": 766771200, "step": 11700, "train_runtime": 6620.4342, "train_tokens_per_second": 115818.868 }, { "epoch": 0.118, "grad_norm": 0.3221312165260315, "learning_rate": 0.00029127825406616677, "loss": 0.6456, "num_input_tokens_seen": 773324800, "step": 11800, "train_runtime": 6676.1674, "train_tokens_per_second": 115833.645 }, { "epoch": 0.119, "grad_norm": 0.2825019657611847, "learning_rate": 0.0002911175980889019, "loss": 0.6526, "num_input_tokens_seen": 779878400, "step": 11900, "train_runtime": 6730.5574, "train_tokens_per_second": 115871.295 }, { "epoch": 0.12, "grad_norm": 0.38775157928466797, "learning_rate": 0.00029095552105899095, "loss": 0.6454, "num_input_tokens_seen": 786432000, "step": 12000, "train_runtime": 6784.0533, "train_tokens_per_second": 115923.617 }, { "epoch": 0.121, "grad_norm": 0.32168275117874146, "learning_rate": 0.0002907920246085478, "loss": 0.644, "num_input_tokens_seen": 792985600, "step": 12100, "train_runtime": 6844.3129, "train_tokens_per_second": 115860.512 }, { "epoch": 0.122, "grad_norm": 0.3085991442203522, "learning_rate": 0.00029062711038397996, "loss": 0.6609, "num_input_tokens_seen": 799539200, "step": 12200, "train_runtime": 6899.148, "train_tokens_per_second": 115889.557 }, { "epoch": 0.123, "grad_norm": 0.26165974140167236, "learning_rate": 0.00029046078004597175, "loss": 0.6436, "num_input_tokens_seen": 806092800, "step": 12300, "train_runtime": 6953.2129, "train_tokens_per_second": 115930.983 }, { "epoch": 0.124, "grad_norm": 0.5689886212348938, "learning_rate": 0.00029029303526946796, "loss": 0.6432, "num_input_tokens_seen": 812646400, "step": 12400, "train_runtime": 7008.0879, "train_tokens_per_second": 115958.363 }, { "epoch": 0.125, "grad_norm": 0.3152099847793579, "learning_rate": 0.0002901238777436565, "loss": 0.6391, "num_input_tokens_seen": 819200000, "step": 12500, "train_runtime": 7067.2869, "train_tokens_per_second": 115914.355 }, { "epoch": 0.126, "grad_norm": 0.28095683455467224, "learning_rate": 0.00028995330917195184, "loss": 0.645, "num_input_tokens_seen": 825753600, "step": 12600, "train_runtime": 7122.3035, "train_tokens_per_second": 115939.121 }, { "epoch": 0.127, "grad_norm": 0.271088570356369, "learning_rate": 0.00028978133127197765, "loss": 0.6452, "num_input_tokens_seen": 832307200, "step": 12700, "train_runtime": 7177.0079, "train_tokens_per_second": 115968.55 }, { "epoch": 0.128, "grad_norm": 0.3357028067111969, "learning_rate": 0.0002896079457755493, "loss": 0.6477, "num_input_tokens_seen": 838860800, "step": 12800, "train_runtime": 7232.8071, "train_tokens_per_second": 115979.977 }, { "epoch": 0.129, "grad_norm": 0.4945499897003174, "learning_rate": 0.000289433154428657, "loss": 0.6455, "num_input_tokens_seen": 845414400, "step": 12900, "train_runtime": 7287.7238, "train_tokens_per_second": 116005.274 }, { "epoch": 0.13, "grad_norm": 0.43597710132598877, "learning_rate": 0.0002892569589914476, "loss": 0.6422, "num_input_tokens_seen": 851968000, "step": 13000, "train_runtime": 7347.1365, "train_tokens_per_second": 115959.19 }, { "epoch": 0.131, "grad_norm": 0.2706185281276703, "learning_rate": 0.0002890793612382072, "loss": 0.6434, "num_input_tokens_seen": 858521600, "step": 13100, "train_runtime": 7401.7618, "train_tokens_per_second": 115988.818 }, { "epoch": 0.132, "grad_norm": 0.22793404757976532, "learning_rate": 0.0002889003629573432, "loss": 0.646, "num_input_tokens_seen": 865075200, "step": 13200, "train_runtime": 7457.327, "train_tokens_per_second": 116003.389 }, { "epoch": 0.133, "grad_norm": 0.2563342750072479, "learning_rate": 0.00028871996595136626, "loss": 0.6503, "num_input_tokens_seen": 871628800, "step": 13300, "train_runtime": 7511.6003, "train_tokens_per_second": 116037.697 }, { "epoch": 0.134, "grad_norm": 0.30800577998161316, "learning_rate": 0.0002885381720368723, "loss": 0.6378, "num_input_tokens_seen": 878182400, "step": 13400, "train_runtime": 7571.585, "train_tokens_per_second": 115983.959 }, { "epoch": 0.135, "grad_norm": 0.2970588505268097, "learning_rate": 0.000288354983044524, "loss": 0.6375, "num_input_tokens_seen": 884736000, "step": 13500, "train_runtime": 7625.682, "train_tokens_per_second": 116020.573 }, { "epoch": 0.136, "grad_norm": 0.24373579025268555, "learning_rate": 0.00028817040081903245, "loss": 0.6436, "num_input_tokens_seen": 891289600, "step": 13600, "train_runtime": 7679.3213, "train_tokens_per_second": 116063.591 }, { "epoch": 0.137, "grad_norm": 0.6528549194335938, "learning_rate": 0.00028798442721913867, "loss": 0.639, "num_input_tokens_seen": 897843200, "step": 13700, "train_runtime": 7735.7681, "train_tokens_per_second": 116063.872 }, { "epoch": 0.138, "grad_norm": 0.2750374376773834, "learning_rate": 0.00028779706411759465, "loss": 0.6362, "num_input_tokens_seen": 904396800, "step": 13800, "train_runtime": 7796.3703, "train_tokens_per_second": 116002.288 }, { "epoch": 0.139, "grad_norm": 0.3658313751220703, "learning_rate": 0.00028760831340114484, "loss": 0.6424, "num_input_tokens_seen": 910950400, "step": 13900, "train_runtime": 7851.8089, "train_tokens_per_second": 116017.903 }, { "epoch": 0.14, "grad_norm": 0.2838168740272522, "learning_rate": 0.00028741817697050683, "loss": 0.6424, "num_input_tokens_seen": 917504000, "step": 14000, "train_runtime": 7906.2377, "train_tokens_per_second": 116048.117 }, { "epoch": 0.141, "grad_norm": 0.30455613136291504, "learning_rate": 0.00028722665674035233, "loss": 0.637, "num_input_tokens_seen": 924057600, "step": 14100, "train_runtime": 7961.9367, "train_tokens_per_second": 116059.401 }, { "epoch": 0.142, "grad_norm": 0.3287622630596161, "learning_rate": 0.0002870337546392879, "loss": 0.6612, "num_input_tokens_seen": 930611200, "step": 14200, "train_runtime": 8016.7903, "train_tokens_per_second": 116082.767 }, { "epoch": 0.143, "grad_norm": 0.2602579593658447, "learning_rate": 0.00028683947260983576, "loss": 0.6433, "num_input_tokens_seen": 937164800, "step": 14300, "train_runtime": 8072.3352, "train_tokens_per_second": 116095.873 }, { "epoch": 0.144, "grad_norm": 0.4514355957508087, "learning_rate": 0.00028664381260841356, "loss": 0.6346, "num_input_tokens_seen": 943718400, "step": 14400, "train_runtime": 8133.1299, "train_tokens_per_second": 116033.853 }, { "epoch": 0.145, "grad_norm": 0.7613756060600281, "learning_rate": 0.0002864467766053154, "loss": 0.6383, "num_input_tokens_seen": 950272000, "step": 14500, "train_runtime": 8186.629, "train_tokens_per_second": 116076.103 }, { "epoch": 0.146, "grad_norm": 0.28044015169143677, "learning_rate": 0.00028624836658469165, "loss": 0.6363, "num_input_tokens_seen": 956825600, "step": 14600, "train_runtime": 8246.5064, "train_tokens_per_second": 116027.995 }, { "epoch": 0.147, "grad_norm": 0.3139192759990692, "learning_rate": 0.00028604858454452906, "loss": 0.65, "num_input_tokens_seen": 963379200, "step": 14700, "train_runtime": 8296.4517, "train_tokens_per_second": 116119.424 }, { "epoch": 0.148, "grad_norm": 0.33300819993019104, "learning_rate": 0.00028584743249663057, "loss": 0.6409, "num_input_tokens_seen": 969932800, "step": 14800, "train_runtime": 8351.7667, "train_tokens_per_second": 116135.045 }, { "epoch": 0.149, "grad_norm": 0.35637542605400085, "learning_rate": 0.000285644912466595, "loss": 0.6356, "num_input_tokens_seen": 976486400, "step": 14900, "train_runtime": 8412.3756, "train_tokens_per_second": 116077.366 }, { "epoch": 0.15, "grad_norm": 0.2518846392631531, "learning_rate": 0.00028544102649379684, "loss": 0.6435, "num_input_tokens_seen": 983040000, "step": 15000, "train_runtime": 8466.8259, "train_tokens_per_second": 116104.904 }, { "epoch": 0.151, "grad_norm": 0.2983591854572296, "learning_rate": 0.00028523577663136556, "loss": 0.638, "num_input_tokens_seen": 989593600, "step": 15100, "train_runtime": 8526.2848, "train_tokens_per_second": 116063.869 }, { "epoch": 0.152, "grad_norm": 0.26403748989105225, "learning_rate": 0.000285029164946165, "loss": 0.6439, "num_input_tokens_seen": 996147200, "step": 15200, "train_runtime": 8581.358, "train_tokens_per_second": 116082.7 }, { "epoch": 0.153, "grad_norm": 0.3652186393737793, "learning_rate": 0.0002848211935187725, "loss": 0.6499, "num_input_tokens_seen": 1002700800, "step": 15300, "train_runtime": 8637.0354, "train_tokens_per_second": 116093.167 }, { "epoch": 0.154, "grad_norm": 0.3816509246826172, "learning_rate": 0.0002846118644434581, "loss": 0.6355, "num_input_tokens_seen": 1009254400, "step": 15400, "train_runtime": 8691.2267, "train_tokens_per_second": 116123.355 }, { "epoch": 0.155, "grad_norm": 0.35392388701438904, "learning_rate": 0.00028440117982816326, "loss": 0.6286, "num_input_tokens_seen": 1015808000, "step": 15500, "train_runtime": 8745.8114, "train_tokens_per_second": 116147.943 }, { "epoch": 0.156, "grad_norm": 0.34099990129470825, "learning_rate": 0.0002841891417944796, "loss": 0.6396, "num_input_tokens_seen": 1022361600, "step": 15600, "train_runtime": 8805.4075, "train_tokens_per_second": 116106.108 }, { "epoch": 0.157, "grad_norm": 0.2872321605682373, "learning_rate": 0.0002839757524776279, "loss": 0.6436, "num_input_tokens_seen": 1028915200, "step": 15700, "train_runtime": 8860.6468, "train_tokens_per_second": 116121.907 }, { "epoch": 0.158, "grad_norm": 0.2647675573825836, "learning_rate": 0.0002837610140264361, "loss": 0.636, "num_input_tokens_seen": 1035468800, "step": 15800, "train_runtime": 8915.1622, "train_tokens_per_second": 116146.94 }, { "epoch": 0.159, "grad_norm": 0.28176337480545044, "learning_rate": 0.0002835449286033182, "loss": 0.6322, "num_input_tokens_seen": 1042022400, "step": 15900, "train_runtime": 8975.5504, "train_tokens_per_second": 116095.655 }, { "epoch": 0.16, "grad_norm": 0.31414464116096497, "learning_rate": 0.0002833274983842518, "loss": 0.6282, "num_input_tokens_seen": 1048576000, "step": 16000, "train_runtime": 9029.8157, "train_tokens_per_second": 116123.744 }, { "epoch": 0.161, "grad_norm": 0.3065328598022461, "learning_rate": 0.0002831087255587569, "loss": 0.6333, "num_input_tokens_seen": 1055129600, "step": 16100, "train_runtime": 9084.3619, "train_tokens_per_second": 116147.905 }, { "epoch": 0.162, "grad_norm": 0.28930503129959106, "learning_rate": 0.0002828886123298734, "loss": 0.6333, "num_input_tokens_seen": 1061683200, "step": 16200, "train_runtime": 9139.4895, "train_tokens_per_second": 116164.388 }, { "epoch": 0.163, "grad_norm": 0.28349098563194275, "learning_rate": 0.00028266716091413906, "loss": 0.6368, "num_input_tokens_seen": 1068236800, "step": 16300, "train_runtime": 9194.7322, "train_tokens_per_second": 116179.219 }, { "epoch": 0.164, "grad_norm": 0.2510078251361847, "learning_rate": 0.0002824443735415673, "loss": 0.6348, "num_input_tokens_seen": 1074790400, "step": 16400, "train_runtime": 9254.9047, "train_tokens_per_second": 116131.979 }, { "epoch": 0.165, "grad_norm": 0.35605230927467346, "learning_rate": 0.0002822202524556243, "loss": 0.639, "num_input_tokens_seen": 1081344000, "step": 16500, "train_runtime": 9310.1385, "train_tokens_per_second": 116146.929 }, { "epoch": 0.166, "grad_norm": 0.518841564655304, "learning_rate": 0.00028199479991320695, "loss": 0.6424, "num_input_tokens_seen": 1087897600, "step": 16600, "train_runtime": 9365.0545, "train_tokens_per_second": 116165.645 }, { "epoch": 0.167, "grad_norm": 0.2293197512626648, "learning_rate": 0.00028176801818461994, "loss": 0.6316, "num_input_tokens_seen": 1094451200, "step": 16700, "train_runtime": 9419.4884, "train_tokens_per_second": 116190.09 }, { "epoch": 0.168, "grad_norm": 0.2790850102901459, "learning_rate": 0.00028153990955355273, "loss": 0.636, "num_input_tokens_seen": 1101004800, "step": 16800, "train_runtime": 9476.046, "train_tokens_per_second": 116188.207 }, { "epoch": 0.169, "grad_norm": 0.2565983235836029, "learning_rate": 0.00028131047631705665, "loss": 0.6351, "num_input_tokens_seen": 1107558400, "step": 16900, "train_runtime": 9536.06, "train_tokens_per_second": 116144.236 }, { "epoch": 0.17, "grad_norm": 0.2592833638191223, "learning_rate": 0.00028107972078552187, "loss": 0.637, "num_input_tokens_seen": 1114112000, "step": 17000, "train_runtime": 9591.4651, "train_tokens_per_second": 116156.603 }, { "epoch": 0.171, "grad_norm": 0.2744060754776001, "learning_rate": 0.0002808476452826541, "loss": 0.6313, "num_input_tokens_seen": 1120665600, "step": 17100, "train_runtime": 9644.9448, "train_tokens_per_second": 116192.018 }, { "epoch": 0.172, "grad_norm": 0.2991725504398346, "learning_rate": 0.00028061425214545094, "loss": 0.6298, "num_input_tokens_seen": 1127219200, "step": 17200, "train_runtime": 9706.1229, "train_tokens_per_second": 116134.857 }, { "epoch": 0.173, "grad_norm": 0.3391658365726471, "learning_rate": 0.00028037954372417883, "loss": 0.6316, "num_input_tokens_seen": 1133772800, "step": 17300, "train_runtime": 9760.541, "train_tokens_per_second": 116158.807 }, { "epoch": 0.174, "grad_norm": 0.3975388705730438, "learning_rate": 0.0002801435223823488, "loss": 0.639, "num_input_tokens_seen": 1140326400, "step": 17400, "train_runtime": 9816.1595, "train_tokens_per_second": 116168.284 }, { "epoch": 0.175, "grad_norm": 0.3754902482032776, "learning_rate": 0.00027990619049669336, "loss": 0.6391, "num_input_tokens_seen": 1146880000, "step": 17500, "train_runtime": 9870.38, "train_tokens_per_second": 116194.108 }, { "epoch": 0.176, "grad_norm": 0.2742336094379425, "learning_rate": 0.00027966755045714177, "loss": 0.6425, "num_input_tokens_seen": 1153433600, "step": 17600, "train_runtime": 9924.9127, "train_tokens_per_second": 116215.994 }, { "epoch": 0.177, "grad_norm": 0.31107792258262634, "learning_rate": 0.00027942760466679673, "loss": 0.6312, "num_input_tokens_seen": 1159987200, "step": 17700, "train_runtime": 9985.7287, "train_tokens_per_second": 116164.502 }, { "epoch": 0.178, "grad_norm": 0.2781549096107483, "learning_rate": 0.00027918635554190956, "loss": 0.6447, "num_input_tokens_seen": 1166540800, "step": 17800, "train_runtime": 10039.8303, "train_tokens_per_second": 116191.287 }, { "epoch": 0.179, "grad_norm": 0.28636643290519714, "learning_rate": 0.00027894380551185636, "loss": 0.6383, "num_input_tokens_seen": 1173094400, "step": 17900, "train_runtime": 10096.4761, "train_tokens_per_second": 116188.498 }, { "epoch": 0.18, "grad_norm": 0.25008583068847656, "learning_rate": 0.00027869995701911314, "loss": 0.6286, "num_input_tokens_seen": 1179648000, "step": 18000, "train_runtime": 10150.5104, "train_tokens_per_second": 116215.634 }, { "epoch": 0.181, "grad_norm": 0.2600831985473633, "learning_rate": 0.0002784548125192316, "loss": 0.6265, "num_input_tokens_seen": 1186201600, "step": 18100, "train_runtime": 10205.6407, "train_tokens_per_second": 116229.998 }, { "epoch": 0.182, "grad_norm": 0.2586776614189148, "learning_rate": 0.0002782083744808141, "loss": 0.6292, "num_input_tokens_seen": 1192755200, "step": 18200, "train_runtime": 10267.3133, "train_tokens_per_second": 116170.138 }, { "epoch": 0.183, "grad_norm": 0.6283107399940491, "learning_rate": 0.000277960645385489, "loss": 0.6364, "num_input_tokens_seen": 1199308800, "step": 18300, "train_runtime": 10323.5921, "train_tokens_per_second": 116171.657 }, { "epoch": 0.184, "grad_norm": 0.2882465124130249, "learning_rate": 0.00027771162772788544, "loss": 0.6318, "num_input_tokens_seen": 1205862400, "step": 18400, "train_runtime": 10377.5746, "train_tokens_per_second": 116198.866 }, { "epoch": 0.185, "grad_norm": 0.3091796934604645, "learning_rate": 0.00027746132401560857, "loss": 0.6269, "num_input_tokens_seen": 1212416000, "step": 18500, "train_runtime": 10432.3665, "train_tokens_per_second": 116216.776 }, { "epoch": 0.186, "grad_norm": 0.280862957239151, "learning_rate": 0.0002772097367692139, "loss": 0.6323, "num_input_tokens_seen": 1218969600, "step": 18600, "train_runtime": 10487.2106, "train_tokens_per_second": 116233.921 }, { "epoch": 0.187, "grad_norm": 0.24253763258457184, "learning_rate": 0.00027695686852218226, "loss": 0.6374, "num_input_tokens_seen": 1225523200, "step": 18700, "train_runtime": 10547.3913, "train_tokens_per_second": 116192.067 }, { "epoch": 0.188, "grad_norm": 0.26167231798171997, "learning_rate": 0.00027670272182089416, "loss": 0.6357, "num_input_tokens_seen": 1232076800, "step": 18800, "train_runtime": 10602.8292, "train_tokens_per_second": 116202.645 }, { "epoch": 0.189, "grad_norm": 0.31137940287590027, "learning_rate": 0.0002764472992246039, "loss": 0.6305, "num_input_tokens_seen": 1238630400, "step": 18900, "train_runtime": 10657.9196, "train_tokens_per_second": 116216.903 }, { "epoch": 0.19, "grad_norm": 0.2601984441280365, "learning_rate": 0.0002761906033054143, "loss": 0.6301, "num_input_tokens_seen": 1245184000, "step": 19000, "train_runtime": 10712.1812, "train_tokens_per_second": 116240.005 }, { "epoch": 0.191, "grad_norm": 0.42287951707839966, "learning_rate": 0.00027593263664825045, "loss": 0.6324, "num_input_tokens_seen": 1251737600, "step": 19100, "train_runtime": 10766.7423, "train_tokens_per_second": 116259.642 }, { "epoch": 0.192, "grad_norm": 0.28982725739479065, "learning_rate": 0.00027567340185083363, "loss": 0.6291, "num_input_tokens_seen": 1258291200, "step": 19200, "train_runtime": 10826.785, "train_tokens_per_second": 116220.208 }, { "epoch": 0.193, "grad_norm": 0.2896488308906555, "learning_rate": 0.00027541290152365537, "loss": 0.6354, "num_input_tokens_seen": 1264844800, "step": 19300, "train_runtime": 10881.8628, "train_tokens_per_second": 116234.216 }, { "epoch": 0.194, "grad_norm": 0.27890026569366455, "learning_rate": 0.00027515113828995117, "loss": 0.6339, "num_input_tokens_seen": 1271398400, "step": 19400, "train_runtime": 10937.449, "train_tokens_per_second": 116242.682 }, { "epoch": 0.195, "grad_norm": 0.2991676926612854, "learning_rate": 0.00027488811478567374, "loss": 0.6282, "num_input_tokens_seen": 1277952000, "step": 19500, "train_runtime": 10991.5778, "train_tokens_per_second": 116266.475 }, { "epoch": 0.196, "grad_norm": 0.2899467349052429, "learning_rate": 0.0002746238336594671, "loss": 0.6264, "num_input_tokens_seen": 1284505600, "step": 19600, "train_runtime": 11050.8685, "train_tokens_per_second": 116235.714 }, { "epoch": 0.197, "grad_norm": 0.32908034324645996, "learning_rate": 0.00027435829757263894, "loss": 0.6321, "num_input_tokens_seen": 1291059200, "step": 19700, "train_runtime": 11105.6721, "train_tokens_per_second": 116252.234 }, { "epoch": 0.198, "grad_norm": 0.23919178545475006, "learning_rate": 0.0002740915091991349, "loss": 0.6342, "num_input_tokens_seen": 1297612800, "step": 19800, "train_runtime": 11160.1183, "train_tokens_per_second": 116272.316 }, { "epoch": 0.199, "grad_norm": 0.5094599723815918, "learning_rate": 0.0002738234712255109, "loss": 0.6317, "num_input_tokens_seen": 1304166400, "step": 19900, "train_runtime": 11221.0166, "train_tokens_per_second": 116225.333 }, { "epoch": 0.2, "grad_norm": 0.37415555119514465, "learning_rate": 0.00027355418635090635, "loss": 0.634, "num_input_tokens_seen": 1310720000, "step": 20000, "train_runtime": 11276.9767, "train_tokens_per_second": 116229.734 }, { "epoch": 0.201, "grad_norm": 0.39444148540496826, "learning_rate": 0.000273283657287017, "loss": 0.6272, "num_input_tokens_seen": 1317273600, "step": 20100, "train_runtime": 11332.3114, "train_tokens_per_second": 116240.505 }, { "epoch": 0.202, "grad_norm": 0.2751108705997467, "learning_rate": 0.00027301188675806745, "loss": 0.6385, "num_input_tokens_seen": 1323827200, "step": 20200, "train_runtime": 11387.621, "train_tokens_per_second": 116251.427 }, { "epoch": 0.203, "grad_norm": 0.2579997479915619, "learning_rate": 0.0002727388775007839, "loss": 0.6278, "num_input_tokens_seen": 1330380800, "step": 20300, "train_runtime": 11441.506, "train_tokens_per_second": 116276.721 }, { "epoch": 0.204, "grad_norm": 0.27806708216667175, "learning_rate": 0.0002724646322643666, "loss": 0.6292, "num_input_tokens_seen": 1336934400, "step": 20400, "train_runtime": 11496.0966, "train_tokens_per_second": 116294.638 }, { "epoch": 0.205, "grad_norm": 0.48101556301116943, "learning_rate": 0.000272189153810462, "loss": 0.6335, "num_input_tokens_seen": 1343488000, "step": 20500, "train_runtime": 11557.1454, "train_tokens_per_second": 116247.391 }, { "epoch": 0.206, "grad_norm": 0.24878458678722382, "learning_rate": 0.0002719124449131351, "loss": 0.6309, "num_input_tokens_seen": 1350041600, "step": 20600, "train_runtime": 11613.0753, "train_tokens_per_second": 116251.86 }, { "epoch": 0.207, "grad_norm": 0.2999299466609955, "learning_rate": 0.00027163450835884144, "loss": 0.627, "num_input_tokens_seen": 1356595200, "step": 20700, "train_runtime": 11667.4657, "train_tokens_per_second": 116271.625 }, { "epoch": 0.208, "grad_norm": 0.2567562162876129, "learning_rate": 0.00027135534694639894, "loss": 0.6326, "num_input_tokens_seen": 1363148800, "step": 20800, "train_runtime": 11723.8934, "train_tokens_per_second": 116271.0 }, { "epoch": 0.209, "grad_norm": 0.3484431207180023, "learning_rate": 0.00027107496348696003, "loss": 0.6356, "num_input_tokens_seen": 1369702400, "step": 20900, "train_runtime": 11777.6307, "train_tokens_per_second": 116296.939 }, { "epoch": 0.21, "grad_norm": 0.23648300766944885, "learning_rate": 0.00027079336080398296, "loss": 0.6256, "num_input_tokens_seen": 1376256000, "step": 21000, "train_runtime": 11838.2167, "train_tokens_per_second": 116255.348 }, { "epoch": 0.211, "grad_norm": 0.33549532294273376, "learning_rate": 0.00027051054173320366, "loss": 0.6271, "num_input_tokens_seen": 1382809600, "step": 21100, "train_runtime": 11893.4157, "train_tokens_per_second": 116266.819 }, { "epoch": 0.212, "grad_norm": 0.2517142593860626, "learning_rate": 0.000270226509122607, "loss": 0.6254, "num_input_tokens_seen": 1389363200, "step": 21200, "train_runtime": 11947.4932, "train_tokens_per_second": 116289.098 }, { "epoch": 0.213, "grad_norm": 0.2526894807815552, "learning_rate": 0.0002699412658323983, "loss": 0.633, "num_input_tokens_seen": 1395916800, "step": 21300, "train_runtime": 12002.0117, "train_tokens_per_second": 116306.902 }, { "epoch": 0.214, "grad_norm": 0.283974289894104, "learning_rate": 0.00026965481473497423, "loss": 0.6273, "num_input_tokens_seen": 1402470400, "step": 21400, "train_runtime": 12063.5743, "train_tokens_per_second": 116256.622 }, { "epoch": 0.215, "grad_norm": 0.24663467705249786, "learning_rate": 0.0002693671587148942, "loss": 0.6236, "num_input_tokens_seen": 1409024000, "step": 21500, "train_runtime": 12118.0936, "train_tokens_per_second": 116274.395 }, { "epoch": 0.216, "grad_norm": 0.30240553617477417, "learning_rate": 0.0002690783006688511, "loss": 0.6272, "num_input_tokens_seen": 1415577600, "step": 21600, "train_runtime": 12174.4024, "train_tokens_per_second": 116274.915 }, { "epoch": 0.217, "grad_norm": 0.2617557644844055, "learning_rate": 0.0002687882435056423, "loss": 0.6256, "num_input_tokens_seen": 1422131200, "step": 21700, "train_runtime": 12229.4977, "train_tokens_per_second": 116286.967 }, { "epoch": 0.218, "grad_norm": 0.3469904661178589, "learning_rate": 0.0002684969901461402, "loss": 0.634, "num_input_tokens_seen": 1428684800, "step": 21800, "train_runtime": 12284.4989, "train_tokens_per_second": 116299.803 }, { "epoch": 0.219, "grad_norm": 0.32695531845092773, "learning_rate": 0.000268204543523263, "loss": 0.6343, "num_input_tokens_seen": 1435238400, "step": 21900, "train_runtime": 12340.2707, "train_tokens_per_second": 116305.261 }, { "epoch": 0.22, "grad_norm": 0.2909862995147705, "learning_rate": 0.0002679109065819447, "loss": 0.6275, "num_input_tokens_seen": 1441792000, "step": 22000, "train_runtime": 12395.2144, "train_tokens_per_second": 116318.44 }, { "epoch": 0.221, "grad_norm": 0.32462117075920105, "learning_rate": 0.0002676160822791062, "loss": 0.6265, "num_input_tokens_seen": 1448345600, "step": 22100, "train_runtime": 12456.5575, "train_tokens_per_second": 116271.738 }, { "epoch": 0.222, "grad_norm": 0.32560595870018005, "learning_rate": 0.00026732007358362496, "loss": 0.6434, "num_input_tokens_seen": 1454899200, "step": 22200, "train_runtime": 12512.3995, "train_tokens_per_second": 116276.594 }, { "epoch": 0.223, "grad_norm": 0.2810288369655609, "learning_rate": 0.0002670228834763052, "loss": 0.6289, "num_input_tokens_seen": 1461452800, "step": 22300, "train_runtime": 12567.5646, "train_tokens_per_second": 116287.67 }, { "epoch": 0.224, "grad_norm": 0.2596298158168793, "learning_rate": 0.00026672451494984804, "loss": 0.6287, "num_input_tokens_seen": 1468006400, "step": 22400, "train_runtime": 12622.4346, "train_tokens_per_second": 116301.367 }, { "epoch": 0.225, "grad_norm": 0.2850113809108734, "learning_rate": 0.0002664249710088213, "loss": 0.6242, "num_input_tokens_seen": 1474560000, "step": 22500, "train_runtime": 12676.5798, "train_tokens_per_second": 116321.596 }, { "epoch": 0.226, "grad_norm": 0.27683302760124207, "learning_rate": 0.00026612425466962893, "loss": 0.6206, "num_input_tokens_seen": 1481113600, "step": 22600, "train_runtime": 12737.6237, "train_tokens_per_second": 116278.643 }, { "epoch": 0.227, "grad_norm": 0.239657461643219, "learning_rate": 0.00026582236896048134, "loss": 0.6225, "num_input_tokens_seen": 1487667200, "step": 22700, "train_runtime": 12792.2969, "train_tokens_per_second": 116293.987 }, { "epoch": 0.228, "grad_norm": 0.23831160366535187, "learning_rate": 0.00026551931692136413, "loss": 0.6239, "num_input_tokens_seen": 1494220800, "step": 22800, "train_runtime": 12846.1745, "train_tokens_per_second": 116316.402 }, { "epoch": 0.229, "grad_norm": 0.34589797258377075, "learning_rate": 0.00026521510160400804, "loss": 0.6247, "num_input_tokens_seen": 1500774400, "step": 22900, "train_runtime": 12900.799, "train_tokens_per_second": 116331.895 }, { "epoch": 0.23, "grad_norm": 0.2497359812259674, "learning_rate": 0.00026490972607185793, "loss": 0.62, "num_input_tokens_seen": 1507328000, "step": 23000, "train_runtime": 12961.1601, "train_tokens_per_second": 116295.763 }, { "epoch": 0.231, "grad_norm": 0.24375373125076294, "learning_rate": 0.0002646031934000421, "loss": 0.6315, "num_input_tokens_seen": 1513881600, "step": 23100, "train_runtime": 13015.5663, "train_tokens_per_second": 116313.156 }, { "epoch": 0.232, "grad_norm": 0.26220017671585083, "learning_rate": 0.00026429550667534095, "loss": 0.6285, "num_input_tokens_seen": 1520435200, "step": 23200, "train_runtime": 13070.9925, "train_tokens_per_second": 116321.327 }, { "epoch": 0.233, "grad_norm": 0.3479808568954468, "learning_rate": 0.0002639866689961565, "loss": 0.6212, "num_input_tokens_seen": 1526988800, "step": 23300, "train_runtime": 13127.7973, "train_tokens_per_second": 116317.213 }, { "epoch": 0.234, "grad_norm": 0.2807121276855469, "learning_rate": 0.00026367668347248083, "loss": 0.6233, "num_input_tokens_seen": 1533542400, "step": 23400, "train_runtime": 13181.874, "train_tokens_per_second": 116337.206 }, { "epoch": 0.235, "grad_norm": 0.22995448112487793, "learning_rate": 0.0002633655532258646, "loss": 0.6322, "num_input_tokens_seen": 1540096000, "step": 23500, "train_runtime": 13243.5947, "train_tokens_per_second": 116289.877 }, { "epoch": 0.236, "grad_norm": 0.2977929413318634, "learning_rate": 0.000263053281389386, "loss": 0.6255, "num_input_tokens_seen": 1546649600, "step": 23600, "train_runtime": 13297.1642, "train_tokens_per_second": 116314.244 }, { "epoch": 0.237, "grad_norm": 0.2991272211074829, "learning_rate": 0.0002627398711076189, "loss": 0.6178, "num_input_tokens_seen": 1553203200, "step": 23700, "train_runtime": 13352.7857, "train_tokens_per_second": 116320.537 }, { "epoch": 0.238, "grad_norm": 0.26155802607536316, "learning_rate": 0.0002624253255366014, "loss": 0.6173, "num_input_tokens_seen": 1559756800, "step": 23800, "train_runtime": 13406.4815, "train_tokens_per_second": 116343.486 }, { "epoch": 0.239, "grad_norm": 0.3786696493625641, "learning_rate": 0.0002621096478438039, "loss": 0.6275, "num_input_tokens_seen": 1566310400, "step": 23900, "train_runtime": 13461.9758, "train_tokens_per_second": 116350.707 }, { "epoch": 0.24, "grad_norm": 0.347703754901886, "learning_rate": 0.00026179284120809727, "loss": 0.6317, "num_input_tokens_seen": 1572864000, "step": 24000, "train_runtime": 13518.3676, "train_tokens_per_second": 116350.143 }, { "epoch": 0.241, "grad_norm": 0.2794933021068573, "learning_rate": 0.0002614749088197208, "loss": 0.6213, "num_input_tokens_seen": 1579417600, "step": 24100, "train_runtime": 13573.1619, "train_tokens_per_second": 116363.277 }, { "epoch": 0.242, "grad_norm": 0.4467460811138153, "learning_rate": 0.00026115585388025015, "loss": 0.618, "num_input_tokens_seen": 1585971200, "step": 24200, "train_runtime": 13633.9103, "train_tokens_per_second": 116325.483 }, { "epoch": 0.243, "grad_norm": 0.3274633288383484, "learning_rate": 0.00026083567960256493, "loss": 0.6198, "num_input_tokens_seen": 1592524800, "step": 24300, "train_runtime": 13690.8223, "train_tokens_per_second": 116320.61 }, { "epoch": 0.244, "grad_norm": 0.26471319794654846, "learning_rate": 0.00026051438921081667, "loss": 0.621, "num_input_tokens_seen": 1599078400, "step": 24400, "train_runtime": 13746.056, "train_tokens_per_second": 116329.979 }, { "epoch": 0.245, "grad_norm": 0.29486411809921265, "learning_rate": 0.00026019198594039595, "loss": 0.6275, "num_input_tokens_seen": 1605632000, "step": 24500, "train_runtime": 13802.5927, "train_tokens_per_second": 116328.29 }, { "epoch": 0.246, "grad_norm": 0.31198298931121826, "learning_rate": 0.00025986847303790026, "loss": 0.6232, "num_input_tokens_seen": 1612185600, "step": 24600, "train_runtime": 13856.7259, "train_tokens_per_second": 116346.792 }, { "epoch": 0.247, "grad_norm": 0.28400614857673645, "learning_rate": 0.00025954385376110076, "loss": 0.6213, "num_input_tokens_seen": 1618739200, "step": 24700, "train_runtime": 13911.4019, "train_tokens_per_second": 116360.609 }, { "epoch": 0.248, "grad_norm": 0.24376364052295685, "learning_rate": 0.00025921813137891005, "loss": 0.6279, "num_input_tokens_seen": 1625292800, "step": 24800, "train_runtime": 13968.2353, "train_tokens_per_second": 116356.345 }, { "epoch": 0.249, "grad_norm": 0.28743863105773926, "learning_rate": 0.000258891309171349, "loss": 0.6238, "num_input_tokens_seen": 1631846400, "step": 24900, "train_runtime": 14022.4846, "train_tokens_per_second": 116373.557 }, { "epoch": 0.25, "grad_norm": 0.33175599575042725, "learning_rate": 0.00025856339042951344, "loss": 0.6159, "num_input_tokens_seen": 1638400000, "step": 25000, "train_runtime": 14076.9493, "train_tokens_per_second": 116388.855 }, { "epoch": 0.251, "grad_norm": 0.31394195556640625, "learning_rate": 0.0002582343784555415, "loss": 0.6194, "num_input_tokens_seen": 1644953600, "step": 25100, "train_runtime": 14137.6518, "train_tokens_per_second": 116352.675 }, { "epoch": 0.252, "grad_norm": 0.2717108428478241, "learning_rate": 0.00025790427656258017, "loss": 0.63, "num_input_tokens_seen": 1651507200, "step": 25200, "train_runtime": 14191.6974, "train_tokens_per_second": 116371.365 }, { "epoch": 0.253, "grad_norm": 0.2907046377658844, "learning_rate": 0.00025757308807475185, "loss": 0.6214, "num_input_tokens_seen": 1658060800, "step": 25300, "train_runtime": 14247.6012, "train_tokens_per_second": 116374.734 }, { "epoch": 0.254, "grad_norm": 0.25791919231414795, "learning_rate": 0.00025724081632712086, "loss": 0.6197, "num_input_tokens_seen": 1664614400, "step": 25400, "train_runtime": 14302.1699, "train_tokens_per_second": 116388.94 }, { "epoch": 0.255, "grad_norm": 0.23399506509304047, "learning_rate": 0.0002569074646656601, "loss": 0.6145, "num_input_tokens_seen": 1671168000, "step": 25500, "train_runtime": 14362.8828, "train_tokens_per_second": 116353.243 }, { "epoch": 0.256, "grad_norm": 0.3080284297466278, "learning_rate": 0.00025657303644721695, "loss": 0.6288, "num_input_tokens_seen": 1677721600, "step": 25600, "train_runtime": 14418.2746, "train_tokens_per_second": 116360.775 }, { "epoch": 0.257, "grad_norm": 0.27764490246772766, "learning_rate": 0.00025623753503948004, "loss": 0.6302, "num_input_tokens_seen": 1684275200, "step": 25700, "train_runtime": 14472.1462, "train_tokens_per_second": 116380.471 }, { "epoch": 0.258, "grad_norm": 0.3530017137527466, "learning_rate": 0.00025590096382094475, "loss": 0.6205, "num_input_tokens_seen": 1690828800, "step": 25800, "train_runtime": 14527.093, "train_tokens_per_second": 116391.407 }, { "epoch": 0.259, "grad_norm": 0.28179532289505005, "learning_rate": 0.00025556332618087945, "loss": 0.6196, "num_input_tokens_seen": 1697382400, "step": 25900, "train_runtime": 14586.2883, "train_tokens_per_second": 116368.357 }, { "epoch": 0.26, "grad_norm": 0.27701336145401, "learning_rate": 0.00025522462551929155, "loss": 0.6244, "num_input_tokens_seen": 1703936000, "step": 26000, "train_runtime": 14641.6003, "train_tokens_per_second": 116376.35 }, { "epoch": 0.261, "grad_norm": 0.2609594464302063, "learning_rate": 0.00025488486524689283, "loss": 0.625, "num_input_tokens_seen": 1710489600, "step": 26100, "train_runtime": 14698.0992, "train_tokens_per_second": 116374.884 }, { "epoch": 0.262, "grad_norm": 0.2739205062389374, "learning_rate": 0.00025454404878506555, "loss": 0.6214, "num_input_tokens_seen": 1717043200, "step": 26200, "train_runtime": 14752.0923, "train_tokens_per_second": 116393.198 }, { "epoch": 0.263, "grad_norm": 0.27936309576034546, "learning_rate": 0.0002542021795658276, "loss": 0.6203, "num_input_tokens_seen": 1723596800, "step": 26300, "train_runtime": 14806.7928, "train_tokens_per_second": 116405.816 }, { "epoch": 0.264, "grad_norm": 0.24206426739692688, "learning_rate": 0.0002538592610317984, "loss": 0.622, "num_input_tokens_seen": 1730150400, "step": 26400, "train_runtime": 14868.0925, "train_tokens_per_second": 116366.669 }, { "epoch": 0.265, "grad_norm": 0.29860714077949524, "learning_rate": 0.00025351529663616355, "loss": 0.6244, "num_input_tokens_seen": 1736704000, "step": 26500, "train_runtime": 14923.7941, "train_tokens_per_second": 116371.479 }, { "epoch": 0.266, "grad_norm": 0.2928908169269562, "learning_rate": 0.00025317028984264087, "loss": 0.6183, "num_input_tokens_seen": 1743257600, "step": 26600, "train_runtime": 14978.8307, "train_tokens_per_second": 116381.421 }, { "epoch": 0.267, "grad_norm": 0.32871657609939575, "learning_rate": 0.0002528242441254448, "loss": 0.6163, "num_input_tokens_seen": 1749811200, "step": 26700, "train_runtime": 15032.6295, "train_tokens_per_second": 116400.873 }, { "epoch": 0.268, "grad_norm": 0.26268327236175537, "learning_rate": 0.000252477162969252, "loss": 0.6211, "num_input_tokens_seen": 1756364800, "step": 26800, "train_runtime": 15087.2644, "train_tokens_per_second": 116413.735 }, { "epoch": 0.269, "grad_norm": 0.248652845621109, "learning_rate": 0.00025212904986916584, "loss": 0.6231, "num_input_tokens_seen": 1762918400, "step": 26900, "train_runtime": 15142.5638, "train_tokens_per_second": 116421.395 }, { "epoch": 0.27, "grad_norm": 0.28964775800704956, "learning_rate": 0.00025177990833068133, "loss": 0.6233, "num_input_tokens_seen": 1769472000, "step": 27000, "train_runtime": 15204.3907, "train_tokens_per_second": 116379.014 }, { "epoch": 0.271, "grad_norm": 0.3479403555393219, "learning_rate": 0.0002514297418696499, "loss": 0.6136, "num_input_tokens_seen": 1776025600, "step": 27100, "train_runtime": 15259.9986, "train_tokens_per_second": 116384.388 }, { "epoch": 0.272, "grad_norm": 0.2727760076522827, "learning_rate": 0.0002510785540122439, "loss": 0.6215, "num_input_tokens_seen": 1782579200, "step": 27200, "train_runtime": 15313.7249, "train_tokens_per_second": 116404.024 }, { "epoch": 0.273, "grad_norm": 0.2885916531085968, "learning_rate": 0.0002507263482949212, "loss": 0.6269, "num_input_tokens_seen": 1789132800, "step": 27300, "train_runtime": 15370.4778, "train_tokens_per_second": 116400.598 }, { "epoch": 0.274, "grad_norm": 0.2588540017604828, "learning_rate": 0.0002503731282643894, "loss": 0.619, "num_input_tokens_seen": 1795686400, "step": 27400, "train_runtime": 15425.7714, "train_tokens_per_second": 116408.208 }, { "epoch": 0.275, "grad_norm": 0.29681408405303955, "learning_rate": 0.0002500188974775704, "loss": 0.6175, "num_input_tokens_seen": 1802240000, "step": 27500, "train_runtime": 15480.68, "train_tokens_per_second": 116418.658 }, { "epoch": 0.276, "grad_norm": 0.319444864988327, "learning_rate": 0.00024966365950156416, "loss": 0.6155, "num_input_tokens_seen": 1808793600, "step": 27600, "train_runtime": 15541.0108, "train_tokens_per_second": 116388.414 }, { "epoch": 0.277, "grad_norm": 0.3779368996620178, "learning_rate": 0.00024930741791361326, "loss": 0.6238, "num_input_tokens_seen": 1815347200, "step": 27700, "train_runtime": 15597.4681, "train_tokens_per_second": 116387.3 }, { "epoch": 0.278, "grad_norm": 0.4451746940612793, "learning_rate": 0.0002489501763010664, "loss": 0.6201, "num_input_tokens_seen": 1821900800, "step": 27800, "train_runtime": 15648.3014, "train_tokens_per_second": 116428.024 }, { "epoch": 0.279, "grad_norm": 0.35030296444892883, "learning_rate": 0.00024859193826134285, "loss": 0.6173, "num_input_tokens_seen": 1828454400, "step": 27900, "train_runtime": 15708.8593, "train_tokens_per_second": 116396.383 }, { "epoch": 0.28, "grad_norm": 0.2962912321090698, "learning_rate": 0.00024823270740189556, "loss": 0.6154, "num_input_tokens_seen": 1835008000, "step": 28000, "train_runtime": 15764.0565, "train_tokens_per_second": 116404.556 }, { "epoch": 0.281, "grad_norm": 0.2939753234386444, "learning_rate": 0.00024787248734017527, "loss": 0.6222, "num_input_tokens_seen": 1841561600, "step": 28100, "train_runtime": 15818.596, "train_tokens_per_second": 116417.513 }, { "epoch": 0.282, "grad_norm": 0.24470455944538116, "learning_rate": 0.0002475112817035941, "loss": 0.6238, "num_input_tokens_seen": 1848115200, "step": 28200, "train_runtime": 15872.9134, "train_tokens_per_second": 116432.01 }, { "epoch": 0.283, "grad_norm": 0.31226930022239685, "learning_rate": 0.0002471490941294887, "loss": 0.6212, "num_input_tokens_seen": 1854668800, "step": 28300, "train_runtime": 15933.0328, "train_tokens_per_second": 116404.003 }, { "epoch": 0.284, "grad_norm": 0.2585756182670593, "learning_rate": 0.000246785928265084, "loss": 0.6191, "num_input_tokens_seen": 1861222400, "step": 28400, "train_runtime": 15987.3279, "train_tokens_per_second": 116418.604 }, { "epoch": 0.285, "grad_norm": 0.4430062770843506, "learning_rate": 0.0002464217877674562, "loss": 0.6249, "num_input_tokens_seen": 1867776000, "step": 28500, "train_runtime": 16041.837, "train_tokens_per_second": 116431.553 }, { "epoch": 0.286, "grad_norm": 0.31980106234550476, "learning_rate": 0.0002460566763034961, "loss": 0.6303, "num_input_tokens_seen": 1874329600, "step": 28600, "train_runtime": 16098.6135, "train_tokens_per_second": 116428.014 }, { "epoch": 0.287, "grad_norm": 0.26904401183128357, "learning_rate": 0.00024569059754987196, "loss": 0.6214, "num_input_tokens_seen": 1880883200, "step": 28700, "train_runtime": 16159.9296, "train_tokens_per_second": 116391.794 }, { "epoch": 0.288, "grad_norm": 0.2635185420513153, "learning_rate": 0.00024532355519299296, "loss": 0.6181, "num_input_tokens_seen": 1887436800, "step": 28800, "train_runtime": 16214.468, "train_tokens_per_second": 116404.485 }, { "epoch": 0.289, "grad_norm": 1.0330250263214111, "learning_rate": 0.0002449555529289714, "loss": 0.6241, "num_input_tokens_seen": 1893990400, "step": 28900, "train_runtime": 16268.5971, "train_tokens_per_second": 116420.02 }, { "epoch": 0.29, "grad_norm": 0.3014289438724518, "learning_rate": 0.0002445865944635861, "loss": 0.6295, "num_input_tokens_seen": 1900544000, "step": 29000, "train_runtime": 16324.9935, "train_tokens_per_second": 116419.281 }, { "epoch": 0.291, "grad_norm": 0.2632603347301483, "learning_rate": 0.0002442166835122446, "loss": 0.6188, "num_input_tokens_seen": 1907097600, "step": 29100, "train_runtime": 16378.6859, "train_tokens_per_second": 116437.766 }, { "epoch": 0.292, "grad_norm": 0.22746869921684265, "learning_rate": 0.00024384582379994614, "loss": 0.6172, "num_input_tokens_seen": 1913651200, "step": 29200, "train_runtime": 16434.3593, "train_tokens_per_second": 116442.093 }, { "epoch": 0.293, "grad_norm": 0.26981279253959656, "learning_rate": 0.00024347401906124388, "loss": 0.6161, "num_input_tokens_seen": 1920204800, "step": 29300, "train_runtime": 16494.2752, "train_tokens_per_second": 116416.44 }, { "epoch": 0.294, "grad_norm": 0.30709710717201233, "learning_rate": 0.0002431012730402075, "loss": 0.6225, "num_input_tokens_seen": 1926758400, "step": 29400, "train_runtime": 16549.5618, "train_tokens_per_second": 116423.53 }, { "epoch": 0.295, "grad_norm": 0.27512043714523315, "learning_rate": 0.00024272758949038517, "loss": 0.6167, "num_input_tokens_seen": 1933312000, "step": 29500, "train_runtime": 16604.0285, "train_tokens_per_second": 116436.321 }, { "epoch": 0.296, "grad_norm": 0.33047741651535034, "learning_rate": 0.00024235297217476616, "loss": 0.6194, "num_input_tokens_seen": 1939865600, "step": 29600, "train_runtime": 16665.4033, "train_tokens_per_second": 116400.76 }, { "epoch": 0.297, "grad_norm": 0.23181872069835663, "learning_rate": 0.00024197742486574268, "loss": 0.6235, "num_input_tokens_seen": 1946419200, "step": 29700, "train_runtime": 16719.5543, "train_tokens_per_second": 116415.735 }, { "epoch": 0.298, "grad_norm": 0.25960394740104675, "learning_rate": 0.0002416009513450719, "loss": 0.6189, "num_input_tokens_seen": 1952972800, "step": 29800, "train_runtime": 16775.5144, "train_tokens_per_second": 116418.058 }, { "epoch": 0.299, "grad_norm": 0.3980496823787689, "learning_rate": 0.00024122355540383806, "loss": 0.6205, "num_input_tokens_seen": 1959526400, "step": 29900, "train_runtime": 16831.9075, "train_tokens_per_second": 116417.369 }, { "epoch": 0.3, "grad_norm": 0.274140328168869, "learning_rate": 0.00024084524084241405, "loss": 0.6137, "num_input_tokens_seen": 1966080000, "step": 30000, "train_runtime": 16885.9089, "train_tokens_per_second": 116433.176 } ], "logging_steps": 100, "max_steps": 100000, "num_input_tokens_seen": 1966080000, "num_train_epochs": 9223372036854775807, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.755848237056e+16, "train_batch_size": 256, "trial_name": null, "trial_params": null }