diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,61255 @@ +{ + "best_global_step": 28620, + "best_metric": 0.46405282616615295, + "best_model_checkpoint": "saves_multiple/bitfit/llama-3-8b-instruct/train_codealpacapy_42_1767887026/checkpoint-28620", + "epoch": 10.0, + "eval_steps": 1908, + "global_step": 38150, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.001310615989515072, + "grad_norm": 3.846224784851074, + "learning_rate": 5.2424639580602885e-08, + "loss": 1.3019, + "num_input_tokens_seen": 1376, + "step": 5 + }, + { + "epoch": 0.002621231979030144, + "grad_norm": 3.254185914993286, + "learning_rate": 1.1795543905635651e-07, + "loss": 1.0837, + "num_input_tokens_seen": 2624, + "step": 10 + }, + { + "epoch": 0.003931847968545216, + "grad_norm": 2.339458703994751, + "learning_rate": 1.8348623853211012e-07, + "loss": 1.5739, + "num_input_tokens_seen": 3728, + "step": 15 + }, + { + "epoch": 0.005242463958060288, + "grad_norm": 2.598907232284546, + "learning_rate": 2.490170380078637e-07, + "loss": 1.0047, + "num_input_tokens_seen": 4928, + "step": 20 + }, + { + "epoch": 0.00655307994757536, + "grad_norm": 3.50645112991333, + "learning_rate": 3.1454783748361734e-07, + "loss": 1.7238, + "num_input_tokens_seen": 5952, + "step": 25 + }, + { + "epoch": 0.007863695937090432, + "grad_norm": 3.7421765327453613, + "learning_rate": 3.8007863695937093e-07, + "loss": 1.2477, + "num_input_tokens_seen": 7248, + "step": 30 + }, + { + "epoch": 0.009174311926605505, + "grad_norm": 2.8655080795288086, + "learning_rate": 4.4560943643512453e-07, + "loss": 1.8127, + "num_input_tokens_seen": 8368, + "step": 35 + }, + { + "epoch": 0.010484927916120577, + "grad_norm": 6.733255863189697, + "learning_rate": 5.111402359108782e-07, + "loss": 1.2682, + "num_input_tokens_seen": 9696, + "step": 40 + }, + { + "epoch": 0.011795543905635648, + "grad_norm": 4.4938812255859375, + "learning_rate": 5.766710353866317e-07, + "loss": 1.1807, + "num_input_tokens_seen": 11248, + "step": 45 + }, + { + "epoch": 0.01310615989515072, + "grad_norm": 5.0903120040893555, + "learning_rate": 6.422018348623854e-07, + "loss": 1.3835, + "num_input_tokens_seen": 12480, + "step": 50 + }, + { + "epoch": 0.014416775884665793, + "grad_norm": 2.1385786533355713, + "learning_rate": 7.07732634338139e-07, + "loss": 1.2339, + "num_input_tokens_seen": 13632, + "step": 55 + }, + { + "epoch": 0.015727391874180863, + "grad_norm": 5.264211177825928, + "learning_rate": 7.732634338138926e-07, + "loss": 1.1147, + "num_input_tokens_seen": 15472, + "step": 60 + }, + { + "epoch": 0.01703800786369594, + "grad_norm": 4.055630207061768, + "learning_rate": 8.387942332896462e-07, + "loss": 1.3305, + "num_input_tokens_seen": 16544, + "step": 65 + }, + { + "epoch": 0.01834862385321101, + "grad_norm": 3.656439781188965, + "learning_rate": 9.043250327653998e-07, + "loss": 1.9867, + "num_input_tokens_seen": 17744, + "step": 70 + }, + { + "epoch": 0.019659239842726082, + "grad_norm": 2.4001808166503906, + "learning_rate": 9.698558322411533e-07, + "loss": 1.7979, + "num_input_tokens_seen": 18816, + "step": 75 + }, + { + "epoch": 0.020969855832241154, + "grad_norm": 2.9585587978363037, + "learning_rate": 1.035386631716907e-06, + "loss": 1.3662, + "num_input_tokens_seen": 20000, + "step": 80 + }, + { + "epoch": 0.022280471821756225, + "grad_norm": 4.001765727996826, + "learning_rate": 1.1009174311926608e-06, + "loss": 1.4613, + "num_input_tokens_seen": 21456, + "step": 85 + }, + { + "epoch": 0.023591087811271297, + "grad_norm": 3.6414175033569336, + "learning_rate": 1.1664482306684142e-06, + "loss": 1.8693, + "num_input_tokens_seen": 22544, + "step": 90 + }, + { + "epoch": 0.02490170380078637, + "grad_norm": 3.9390757083892822, + "learning_rate": 1.2319790301441677e-06, + "loss": 1.2346, + "num_input_tokens_seen": 23712, + "step": 95 + }, + { + "epoch": 0.02621231979030144, + "grad_norm": 2.8755433559417725, + "learning_rate": 1.2975098296199214e-06, + "loss": 1.0142, + "num_input_tokens_seen": 25056, + "step": 100 + }, + { + "epoch": 0.027522935779816515, + "grad_norm": 13.067869186401367, + "learning_rate": 1.363040629095675e-06, + "loss": 1.8748, + "num_input_tokens_seen": 26304, + "step": 105 + }, + { + "epoch": 0.028833551769331587, + "grad_norm": 2.8877787590026855, + "learning_rate": 1.4285714285714286e-06, + "loss": 1.5463, + "num_input_tokens_seen": 27216, + "step": 110 + }, + { + "epoch": 0.03014416775884666, + "grad_norm": 2.4005017280578613, + "learning_rate": 1.4941022280471821e-06, + "loss": 1.1346, + "num_input_tokens_seen": 28944, + "step": 115 + }, + { + "epoch": 0.03145478374836173, + "grad_norm": 1.663111925125122, + "learning_rate": 1.559633027522936e-06, + "loss": 1.2819, + "num_input_tokens_seen": 30368, + "step": 120 + }, + { + "epoch": 0.0327653997378768, + "grad_norm": 5.853882312774658, + "learning_rate": 1.6251638269986893e-06, + "loss": 1.2409, + "num_input_tokens_seen": 31408, + "step": 125 + }, + { + "epoch": 0.03407601572739188, + "grad_norm": 2.7691025733947754, + "learning_rate": 1.690694626474443e-06, + "loss": 1.0254, + "num_input_tokens_seen": 32576, + "step": 130 + }, + { + "epoch": 0.035386631716906945, + "grad_norm": 1.7569845914840698, + "learning_rate": 1.7562254259501965e-06, + "loss": 1.2749, + "num_input_tokens_seen": 33904, + "step": 135 + }, + { + "epoch": 0.03669724770642202, + "grad_norm": 14.174750328063965, + "learning_rate": 1.8217562254259502e-06, + "loss": 1.8093, + "num_input_tokens_seen": 35040, + "step": 140 + }, + { + "epoch": 0.03800786369593709, + "grad_norm": 4.061116695404053, + "learning_rate": 1.8872870249017041e-06, + "loss": 1.3049, + "num_input_tokens_seen": 36112, + "step": 145 + }, + { + "epoch": 0.039318479685452164, + "grad_norm": 7.5897440910339355, + "learning_rate": 1.9528178243774574e-06, + "loss": 1.695, + "num_input_tokens_seen": 37472, + "step": 150 + }, + { + "epoch": 0.04062909567496723, + "grad_norm": 6.924594402313232, + "learning_rate": 2.0183486238532113e-06, + "loss": 1.2835, + "num_input_tokens_seen": 38704, + "step": 155 + }, + { + "epoch": 0.04193971166448231, + "grad_norm": 4.211245536804199, + "learning_rate": 2.083879423328965e-06, + "loss": 1.0876, + "num_input_tokens_seen": 40000, + "step": 160 + }, + { + "epoch": 0.04325032765399738, + "grad_norm": 2.518361806869507, + "learning_rate": 2.1494102228047183e-06, + "loss": 3.7697, + "num_input_tokens_seen": 41248, + "step": 165 + }, + { + "epoch": 0.04456094364351245, + "grad_norm": 5.665794849395752, + "learning_rate": 2.2149410222804718e-06, + "loss": 1.4235, + "num_input_tokens_seen": 42272, + "step": 170 + }, + { + "epoch": 0.045871559633027525, + "grad_norm": 2.5097098350524902, + "learning_rate": 2.2804718217562257e-06, + "loss": 1.2972, + "num_input_tokens_seen": 43744, + "step": 175 + }, + { + "epoch": 0.047182175622542594, + "grad_norm": 7.952512264251709, + "learning_rate": 2.346002621231979e-06, + "loss": 1.4481, + "num_input_tokens_seen": 45296, + "step": 180 + }, + { + "epoch": 0.04849279161205767, + "grad_norm": 3.1425018310546875, + "learning_rate": 2.4115334207077327e-06, + "loss": 1.312, + "num_input_tokens_seen": 46352, + "step": 185 + }, + { + "epoch": 0.04980340760157274, + "grad_norm": 1.5122203826904297, + "learning_rate": 2.4770642201834866e-06, + "loss": 0.8633, + "num_input_tokens_seen": 48432, + "step": 190 + }, + { + "epoch": 0.05111402359108781, + "grad_norm": 9.691925048828125, + "learning_rate": 2.54259501965924e-06, + "loss": 1.279, + "num_input_tokens_seen": 49904, + "step": 195 + }, + { + "epoch": 0.05242463958060288, + "grad_norm": 3.0802063941955566, + "learning_rate": 2.6081258191349936e-06, + "loss": 1.4006, + "num_input_tokens_seen": 51200, + "step": 200 + }, + { + "epoch": 0.053735255570117955, + "grad_norm": 2.2185332775115967, + "learning_rate": 2.673656618610747e-06, + "loss": 1.1759, + "num_input_tokens_seen": 52544, + "step": 205 + }, + { + "epoch": 0.05504587155963303, + "grad_norm": 3.067477226257324, + "learning_rate": 2.739187418086501e-06, + "loss": 1.3775, + "num_input_tokens_seen": 53984, + "step": 210 + }, + { + "epoch": 0.0563564875491481, + "grad_norm": 2.2086341381073, + "learning_rate": 2.8047182175622545e-06, + "loss": 1.2139, + "num_input_tokens_seen": 55296, + "step": 215 + }, + { + "epoch": 0.057667103538663174, + "grad_norm": 2.6030097007751465, + "learning_rate": 2.870249017038008e-06, + "loss": 0.973, + "num_input_tokens_seen": 56480, + "step": 220 + }, + { + "epoch": 0.05897771952817824, + "grad_norm": 4.604193687438965, + "learning_rate": 2.935779816513762e-06, + "loss": 1.2668, + "num_input_tokens_seen": 57504, + "step": 225 + }, + { + "epoch": 0.06028833551769332, + "grad_norm": 1.3474212884902954, + "learning_rate": 3.0013106159895154e-06, + "loss": 0.8436, + "num_input_tokens_seen": 59312, + "step": 230 + }, + { + "epoch": 0.061598951507208385, + "grad_norm": 1.581684947013855, + "learning_rate": 3.066841415465269e-06, + "loss": 0.8479, + "num_input_tokens_seen": 61024, + "step": 235 + }, + { + "epoch": 0.06290956749672345, + "grad_norm": 2.574087381362915, + "learning_rate": 3.1323722149410228e-06, + "loss": 1.0611, + "num_input_tokens_seen": 62448, + "step": 240 + }, + { + "epoch": 0.06422018348623854, + "grad_norm": 1.907650351524353, + "learning_rate": 3.1979030144167763e-06, + "loss": 1.0752, + "num_input_tokens_seen": 64016, + "step": 245 + }, + { + "epoch": 0.0655307994757536, + "grad_norm": 12.28991985321045, + "learning_rate": 3.2634338138925293e-06, + "loss": 1.4371, + "num_input_tokens_seen": 65344, + "step": 250 + }, + { + "epoch": 0.06684141546526867, + "grad_norm": 4.6881585121154785, + "learning_rate": 3.328964613368283e-06, + "loss": 1.0877, + "num_input_tokens_seen": 66512, + "step": 255 + }, + { + "epoch": 0.06815203145478375, + "grad_norm": 1.1858348846435547, + "learning_rate": 3.394495412844037e-06, + "loss": 1.1306, + "num_input_tokens_seen": 67792, + "step": 260 + }, + { + "epoch": 0.06946264744429882, + "grad_norm": 2.206373691558838, + "learning_rate": 3.4600262123197906e-06, + "loss": 0.9708, + "num_input_tokens_seen": 69488, + "step": 265 + }, + { + "epoch": 0.07077326343381389, + "grad_norm": 1.6537030935287476, + "learning_rate": 3.5255570117955437e-06, + "loss": 1.6446, + "num_input_tokens_seen": 70912, + "step": 270 + }, + { + "epoch": 0.07208387942332896, + "grad_norm": 7.618088245391846, + "learning_rate": 3.591087811271298e-06, + "loss": 1.4261, + "num_input_tokens_seen": 72704, + "step": 275 + }, + { + "epoch": 0.07339449541284404, + "grad_norm": 1.9148473739624023, + "learning_rate": 3.6566186107470515e-06, + "loss": 1.248, + "num_input_tokens_seen": 73888, + "step": 280 + }, + { + "epoch": 0.07470511140235911, + "grad_norm": 1.4986572265625, + "learning_rate": 3.7221494102228046e-06, + "loss": 1.3604, + "num_input_tokens_seen": 75520, + "step": 285 + }, + { + "epoch": 0.07601572739187418, + "grad_norm": 6.730327129364014, + "learning_rate": 3.787680209698558e-06, + "loss": 1.6788, + "num_input_tokens_seen": 76672, + "step": 290 + }, + { + "epoch": 0.07732634338138926, + "grad_norm": 7.483281135559082, + "learning_rate": 3.853211009174312e-06, + "loss": 1.2469, + "num_input_tokens_seen": 77936, + "step": 295 + }, + { + "epoch": 0.07863695937090433, + "grad_norm": 2.496995449066162, + "learning_rate": 3.918741808650066e-06, + "loss": 1.3427, + "num_input_tokens_seen": 79040, + "step": 300 + }, + { + "epoch": 0.0799475753604194, + "grad_norm": 1.9096730947494507, + "learning_rate": 3.984272608125819e-06, + "loss": 0.8865, + "num_input_tokens_seen": 80880, + "step": 305 + }, + { + "epoch": 0.08125819134993446, + "grad_norm": 2.71871280670166, + "learning_rate": 4.049803407601573e-06, + "loss": 1.3419, + "num_input_tokens_seen": 82240, + "step": 310 + }, + { + "epoch": 0.08256880733944955, + "grad_norm": 2.390155553817749, + "learning_rate": 4.115334207077327e-06, + "loss": 1.1966, + "num_input_tokens_seen": 83568, + "step": 315 + }, + { + "epoch": 0.08387942332896461, + "grad_norm": 4.776840686798096, + "learning_rate": 4.18086500655308e-06, + "loss": 0.8945, + "num_input_tokens_seen": 85072, + "step": 320 + }, + { + "epoch": 0.08519003931847968, + "grad_norm": 1.5795719623565674, + "learning_rate": 4.246395806028834e-06, + "loss": 1.1378, + "num_input_tokens_seen": 86480, + "step": 325 + }, + { + "epoch": 0.08650065530799476, + "grad_norm": 4.540282726287842, + "learning_rate": 4.311926605504588e-06, + "loss": 0.8783, + "num_input_tokens_seen": 88304, + "step": 330 + }, + { + "epoch": 0.08781127129750983, + "grad_norm": 1.8548303842544556, + "learning_rate": 4.377457404980341e-06, + "loss": 1.2616, + "num_input_tokens_seen": 89536, + "step": 335 + }, + { + "epoch": 0.0891218872870249, + "grad_norm": 3.623392105102539, + "learning_rate": 4.442988204456095e-06, + "loss": 1.0107, + "num_input_tokens_seen": 90816, + "step": 340 + }, + { + "epoch": 0.09043250327653997, + "grad_norm": 2.529656171798706, + "learning_rate": 4.508519003931848e-06, + "loss": 1.0864, + "num_input_tokens_seen": 92128, + "step": 345 + }, + { + "epoch": 0.09174311926605505, + "grad_norm": 2.369105100631714, + "learning_rate": 4.574049803407602e-06, + "loss": 0.9431, + "num_input_tokens_seen": 93648, + "step": 350 + }, + { + "epoch": 0.09305373525557012, + "grad_norm": 1.5330578088760376, + "learning_rate": 4.639580602883356e-06, + "loss": 0.9618, + "num_input_tokens_seen": 95552, + "step": 355 + }, + { + "epoch": 0.09436435124508519, + "grad_norm": 2.724327325820923, + "learning_rate": 4.705111402359109e-06, + "loss": 1.2001, + "num_input_tokens_seen": 96640, + "step": 360 + }, + { + "epoch": 0.09567496723460026, + "grad_norm": 2.9173381328582764, + "learning_rate": 4.7706422018348626e-06, + "loss": 1.2289, + "num_input_tokens_seen": 98080, + "step": 365 + }, + { + "epoch": 0.09698558322411534, + "grad_norm": 2.145061492919922, + "learning_rate": 4.8361730013106165e-06, + "loss": 1.1443, + "num_input_tokens_seen": 99424, + "step": 370 + }, + { + "epoch": 0.0982961992136304, + "grad_norm": 4.89067268371582, + "learning_rate": 4.9017038007863695e-06, + "loss": 1.0205, + "num_input_tokens_seen": 100688, + "step": 375 + }, + { + "epoch": 0.09960681520314547, + "grad_norm": 3.211623430252075, + "learning_rate": 4.9672346002621235e-06, + "loss": 1.165, + "num_input_tokens_seen": 102256, + "step": 380 + }, + { + "epoch": 0.10091743119266056, + "grad_norm": 7.70582389831543, + "learning_rate": 5.032765399737877e-06, + "loss": 1.4703, + "num_input_tokens_seen": 103200, + "step": 385 + }, + { + "epoch": 0.10222804718217562, + "grad_norm": 1.4686022996902466, + "learning_rate": 5.0982961992136304e-06, + "loss": 1.3071, + "num_input_tokens_seen": 104512, + "step": 390 + }, + { + "epoch": 0.10353866317169069, + "grad_norm": 10.97958755493164, + "learning_rate": 5.163826998689384e-06, + "loss": 1.3595, + "num_input_tokens_seen": 105600, + "step": 395 + }, + { + "epoch": 0.10484927916120576, + "grad_norm": 1.9453681707382202, + "learning_rate": 5.229357798165138e-06, + "loss": 1.1708, + "num_input_tokens_seen": 107072, + "step": 400 + }, + { + "epoch": 0.10615989515072084, + "grad_norm": 8.922235488891602, + "learning_rate": 5.294888597640891e-06, + "loss": 1.7714, + "num_input_tokens_seen": 108000, + "step": 405 + }, + { + "epoch": 0.10747051114023591, + "grad_norm": 6.142666816711426, + "learning_rate": 5.360419397116645e-06, + "loss": 1.6009, + "num_input_tokens_seen": 109056, + "step": 410 + }, + { + "epoch": 0.10878112712975098, + "grad_norm": 5.79868745803833, + "learning_rate": 5.425950196592398e-06, + "loss": 1.1909, + "num_input_tokens_seen": 110080, + "step": 415 + }, + { + "epoch": 0.11009174311926606, + "grad_norm": 2.244628667831421, + "learning_rate": 5.491480996068152e-06, + "loss": 1.0042, + "num_input_tokens_seen": 111376, + "step": 420 + }, + { + "epoch": 0.11140235910878113, + "grad_norm": 2.628077745437622, + "learning_rate": 5.557011795543906e-06, + "loss": 1.3302, + "num_input_tokens_seen": 112816, + "step": 425 + }, + { + "epoch": 0.1127129750982962, + "grad_norm": 4.687661170959473, + "learning_rate": 5.622542595019659e-06, + "loss": 1.8225, + "num_input_tokens_seen": 113872, + "step": 430 + }, + { + "epoch": 0.11402359108781127, + "grad_norm": 3.6515042781829834, + "learning_rate": 5.688073394495413e-06, + "loss": 1.3042, + "num_input_tokens_seen": 114976, + "step": 435 + }, + { + "epoch": 0.11533420707732635, + "grad_norm": 6.468430995941162, + "learning_rate": 5.753604193971167e-06, + "loss": 1.1965, + "num_input_tokens_seen": 116608, + "step": 440 + }, + { + "epoch": 0.11664482306684142, + "grad_norm": 1.6323978900909424, + "learning_rate": 5.81913499344692e-06, + "loss": 0.9484, + "num_input_tokens_seen": 118064, + "step": 445 + }, + { + "epoch": 0.11795543905635648, + "grad_norm": 24.916282653808594, + "learning_rate": 5.884665792922674e-06, + "loss": 1.9031, + "num_input_tokens_seen": 119280, + "step": 450 + }, + { + "epoch": 0.11926605504587157, + "grad_norm": 2.8626625537872314, + "learning_rate": 5.950196592398428e-06, + "loss": 1.4533, + "num_input_tokens_seen": 120512, + "step": 455 + }, + { + "epoch": 0.12057667103538663, + "grad_norm": 3.5557332038879395, + "learning_rate": 6.015727391874181e-06, + "loss": 2.4936, + "num_input_tokens_seen": 121568, + "step": 460 + }, + { + "epoch": 0.1218872870249017, + "grad_norm": 5.3994855880737305, + "learning_rate": 6.081258191349935e-06, + "loss": 1.0119, + "num_input_tokens_seen": 122800, + "step": 465 + }, + { + "epoch": 0.12319790301441677, + "grad_norm": 4.495554447174072, + "learning_rate": 6.146788990825689e-06, + "loss": 1.1578, + "num_input_tokens_seen": 124256, + "step": 470 + }, + { + "epoch": 0.12450851900393185, + "grad_norm": 7.532297611236572, + "learning_rate": 6.212319790301442e-06, + "loss": 1.3075, + "num_input_tokens_seen": 125376, + "step": 475 + }, + { + "epoch": 0.1258191349934469, + "grad_norm": 5.466452598571777, + "learning_rate": 6.277850589777196e-06, + "loss": 1.2591, + "num_input_tokens_seen": 126544, + "step": 480 + }, + { + "epoch": 0.127129750982962, + "grad_norm": 3.010937213897705, + "learning_rate": 6.343381389252949e-06, + "loss": 1.2232, + "num_input_tokens_seen": 127648, + "step": 485 + }, + { + "epoch": 0.12844036697247707, + "grad_norm": 1.77584707736969, + "learning_rate": 6.408912188728703e-06, + "loss": 0.973, + "num_input_tokens_seen": 129360, + "step": 490 + }, + { + "epoch": 0.12975098296199214, + "grad_norm": 6.748624801635742, + "learning_rate": 6.474442988204456e-06, + "loss": 1.1493, + "num_input_tokens_seen": 130656, + "step": 495 + }, + { + "epoch": 0.1310615989515072, + "grad_norm": 5.36815881729126, + "learning_rate": 6.539973787680211e-06, + "loss": 1.1137, + "num_input_tokens_seen": 131888, + "step": 500 + }, + { + "epoch": 0.13237221494102228, + "grad_norm": 3.3732218742370605, + "learning_rate": 6.605504587155964e-06, + "loss": 0.7972, + "num_input_tokens_seen": 135792, + "step": 505 + }, + { + "epoch": 0.13368283093053734, + "grad_norm": 5.763914585113525, + "learning_rate": 6.671035386631718e-06, + "loss": 1.1163, + "num_input_tokens_seen": 136976, + "step": 510 + }, + { + "epoch": 0.1349934469200524, + "grad_norm": 7.237623691558838, + "learning_rate": 6.736566186107471e-06, + "loss": 1.1935, + "num_input_tokens_seen": 138080, + "step": 515 + }, + { + "epoch": 0.1363040629095675, + "grad_norm": 2.885622501373291, + "learning_rate": 6.8020969855832246e-06, + "loss": 0.9255, + "num_input_tokens_seen": 139504, + "step": 520 + }, + { + "epoch": 0.13761467889908258, + "grad_norm": 2.8283274173736572, + "learning_rate": 6.867627785058978e-06, + "loss": 0.9843, + "num_input_tokens_seen": 140992, + "step": 525 + }, + { + "epoch": 0.13892529488859764, + "grad_norm": 5.531304836273193, + "learning_rate": 6.933158584534731e-06, + "loss": 2.3689, + "num_input_tokens_seen": 142128, + "step": 530 + }, + { + "epoch": 0.1402359108781127, + "grad_norm": 3.4855892658233643, + "learning_rate": 6.9986893840104855e-06, + "loss": 0.9957, + "num_input_tokens_seen": 143264, + "step": 535 + }, + { + "epoch": 0.14154652686762778, + "grad_norm": 18.80222511291504, + "learning_rate": 7.064220183486239e-06, + "loss": 1.4764, + "num_input_tokens_seen": 145040, + "step": 540 + }, + { + "epoch": 0.14285714285714285, + "grad_norm": 6.341445446014404, + "learning_rate": 7.1297509829619924e-06, + "loss": 0.9933, + "num_input_tokens_seen": 147120, + "step": 545 + }, + { + "epoch": 0.14416775884665792, + "grad_norm": 3.294316530227661, + "learning_rate": 7.195281782437746e-06, + "loss": 1.4763, + "num_input_tokens_seen": 148544, + "step": 550 + }, + { + "epoch": 0.145478374836173, + "grad_norm": 2.8423612117767334, + "learning_rate": 7.260812581913499e-06, + "loss": 1.1525, + "num_input_tokens_seen": 149872, + "step": 555 + }, + { + "epoch": 0.14678899082568808, + "grad_norm": 5.413058280944824, + "learning_rate": 7.326343381389253e-06, + "loss": 1.0485, + "num_input_tokens_seen": 151360, + "step": 560 + }, + { + "epoch": 0.14809960681520315, + "grad_norm": 5.421134948730469, + "learning_rate": 7.391874180865006e-06, + "loss": 1.0327, + "num_input_tokens_seen": 152496, + "step": 565 + }, + { + "epoch": 0.14941022280471822, + "grad_norm": 7.7952046394348145, + "learning_rate": 7.457404980340761e-06, + "loss": 0.9173, + "num_input_tokens_seen": 154176, + "step": 570 + }, + { + "epoch": 0.15072083879423329, + "grad_norm": 3.868422508239746, + "learning_rate": 7.522935779816514e-06, + "loss": 0.7842, + "num_input_tokens_seen": 156704, + "step": 575 + }, + { + "epoch": 0.15203145478374835, + "grad_norm": 4.6714887619018555, + "learning_rate": 7.588466579292268e-06, + "loss": 1.1394, + "num_input_tokens_seen": 157712, + "step": 580 + }, + { + "epoch": 0.15334207077326342, + "grad_norm": 2.360036849975586, + "learning_rate": 7.653997378768021e-06, + "loss": 0.7267, + "num_input_tokens_seen": 159120, + "step": 585 + }, + { + "epoch": 0.15465268676277852, + "grad_norm": 5.994152545928955, + "learning_rate": 7.719528178243775e-06, + "loss": 1.0151, + "num_input_tokens_seen": 160144, + "step": 590 + }, + { + "epoch": 0.1559633027522936, + "grad_norm": 4.685830116271973, + "learning_rate": 7.785058977719529e-06, + "loss": 0.7124, + "num_input_tokens_seen": 161504, + "step": 595 + }, + { + "epoch": 0.15727391874180865, + "grad_norm": 4.997828006744385, + "learning_rate": 7.850589777195281e-06, + "loss": 1.3054, + "num_input_tokens_seen": 162704, + "step": 600 + }, + { + "epoch": 0.15858453473132372, + "grad_norm": 8.459344863891602, + "learning_rate": 7.916120576671037e-06, + "loss": 1.0856, + "num_input_tokens_seen": 163616, + "step": 605 + }, + { + "epoch": 0.1598951507208388, + "grad_norm": 3.40563702583313, + "learning_rate": 7.981651376146789e-06, + "loss": 0.722, + "num_input_tokens_seen": 165408, + "step": 610 + }, + { + "epoch": 0.16120576671035386, + "grad_norm": 1.5694941282272339, + "learning_rate": 8.047182175622543e-06, + "loss": 0.7731, + "num_input_tokens_seen": 167248, + "step": 615 + }, + { + "epoch": 0.16251638269986893, + "grad_norm": 2.676271915435791, + "learning_rate": 8.112712975098297e-06, + "loss": 0.7228, + "num_input_tokens_seen": 168800, + "step": 620 + }, + { + "epoch": 0.16382699868938402, + "grad_norm": 2.2293405532836914, + "learning_rate": 8.17824377457405e-06, + "loss": 0.8775, + "num_input_tokens_seen": 169984, + "step": 625 + }, + { + "epoch": 0.1651376146788991, + "grad_norm": 16.881282806396484, + "learning_rate": 8.243774574049803e-06, + "loss": 1.5016, + "num_input_tokens_seen": 171008, + "step": 630 + }, + { + "epoch": 0.16644823066841416, + "grad_norm": 3.1740939617156982, + "learning_rate": 8.309305373525557e-06, + "loss": 0.9511, + "num_input_tokens_seen": 172704, + "step": 635 + }, + { + "epoch": 0.16775884665792923, + "grad_norm": 2.131955146789551, + "learning_rate": 8.374836173001311e-06, + "loss": 1.3583, + "num_input_tokens_seen": 173920, + "step": 640 + }, + { + "epoch": 0.1690694626474443, + "grad_norm": 5.271409034729004, + "learning_rate": 8.440366972477065e-06, + "loss": 1.1971, + "num_input_tokens_seen": 174880, + "step": 645 + }, + { + "epoch": 0.17038007863695936, + "grad_norm": 3.865002155303955, + "learning_rate": 8.505897771952819e-06, + "loss": 1.0556, + "num_input_tokens_seen": 175984, + "step": 650 + }, + { + "epoch": 0.17169069462647443, + "grad_norm": 2.3796420097351074, + "learning_rate": 8.571428571428573e-06, + "loss": 1.8921, + "num_input_tokens_seen": 177184, + "step": 655 + }, + { + "epoch": 0.17300131061598953, + "grad_norm": 2.042616844177246, + "learning_rate": 8.636959370904325e-06, + "loss": 0.8475, + "num_input_tokens_seen": 178256, + "step": 660 + }, + { + "epoch": 0.1743119266055046, + "grad_norm": 24.40669059753418, + "learning_rate": 8.702490170380079e-06, + "loss": 1.4024, + "num_input_tokens_seen": 179552, + "step": 665 + }, + { + "epoch": 0.17562254259501967, + "grad_norm": 5.376147270202637, + "learning_rate": 8.768020969855833e-06, + "loss": 0.7908, + "num_input_tokens_seen": 180944, + "step": 670 + }, + { + "epoch": 0.17693315858453473, + "grad_norm": 3.1144649982452393, + "learning_rate": 8.833551769331587e-06, + "loss": 0.8539, + "num_input_tokens_seen": 182320, + "step": 675 + }, + { + "epoch": 0.1782437745740498, + "grad_norm": 1.6401352882385254, + "learning_rate": 8.89908256880734e-06, + "loss": 0.7349, + "num_input_tokens_seen": 183888, + "step": 680 + }, + { + "epoch": 0.17955439056356487, + "grad_norm": 3.6570894718170166, + "learning_rate": 8.964613368283094e-06, + "loss": 0.6101, + "num_input_tokens_seen": 186416, + "step": 685 + }, + { + "epoch": 0.18086500655307994, + "grad_norm": 7.9763503074646, + "learning_rate": 9.030144167758847e-06, + "loss": 0.7557, + "num_input_tokens_seen": 187760, + "step": 690 + }, + { + "epoch": 0.182175622542595, + "grad_norm": 3.9461050033569336, + "learning_rate": 9.0956749672346e-06, + "loss": 0.7521, + "num_input_tokens_seen": 189088, + "step": 695 + }, + { + "epoch": 0.1834862385321101, + "grad_norm": 4.23895263671875, + "learning_rate": 9.161205766710354e-06, + "loss": 0.9724, + "num_input_tokens_seen": 190528, + "step": 700 + }, + { + "epoch": 0.18479685452162517, + "grad_norm": 2.330246925354004, + "learning_rate": 9.226736566186107e-06, + "loss": 0.6913, + "num_input_tokens_seen": 191712, + "step": 705 + }, + { + "epoch": 0.18610747051114024, + "grad_norm": 2.3057668209075928, + "learning_rate": 9.29226736566186e-06, + "loss": 0.6596, + "num_input_tokens_seen": 193200, + "step": 710 + }, + { + "epoch": 0.1874180865006553, + "grad_norm": 4.193898677825928, + "learning_rate": 9.357798165137616e-06, + "loss": 0.7758, + "num_input_tokens_seen": 194752, + "step": 715 + }, + { + "epoch": 0.18872870249017037, + "grad_norm": 2.001544237136841, + "learning_rate": 9.423328964613368e-06, + "loss": 1.9828, + "num_input_tokens_seen": 195952, + "step": 720 + }, + { + "epoch": 0.19003931847968544, + "grad_norm": 44.31068420410156, + "learning_rate": 9.488859764089122e-06, + "loss": 1.3361, + "num_input_tokens_seen": 197008, + "step": 725 + }, + { + "epoch": 0.1913499344692005, + "grad_norm": 4.486158847808838, + "learning_rate": 9.554390563564876e-06, + "loss": 0.8131, + "num_input_tokens_seen": 198112, + "step": 730 + }, + { + "epoch": 0.1926605504587156, + "grad_norm": 1.6438493728637695, + "learning_rate": 9.619921363040628e-06, + "loss": 0.8507, + "num_input_tokens_seen": 199968, + "step": 735 + }, + { + "epoch": 0.19397116644823068, + "grad_norm": 5.3324785232543945, + "learning_rate": 9.685452162516382e-06, + "loss": 1.0968, + "num_input_tokens_seen": 201056, + "step": 740 + }, + { + "epoch": 0.19528178243774574, + "grad_norm": 11.962841033935547, + "learning_rate": 9.750982961992136e-06, + "loss": 0.8785, + "num_input_tokens_seen": 202192, + "step": 745 + }, + { + "epoch": 0.1965923984272608, + "grad_norm": 62.54576110839844, + "learning_rate": 9.81651376146789e-06, + "loss": 0.6166, + "num_input_tokens_seen": 203856, + "step": 750 + }, + { + "epoch": 0.19790301441677588, + "grad_norm": 2.1135246753692627, + "learning_rate": 9.882044560943644e-06, + "loss": 0.6381, + "num_input_tokens_seen": 206128, + "step": 755 + }, + { + "epoch": 0.19921363040629095, + "grad_norm": 4.32196569442749, + "learning_rate": 9.947575360419398e-06, + "loss": 0.7326, + "num_input_tokens_seen": 207392, + "step": 760 + }, + { + "epoch": 0.20052424639580602, + "grad_norm": 2.709244728088379, + "learning_rate": 1.0013106159895152e-05, + "loss": 0.9758, + "num_input_tokens_seen": 208880, + "step": 765 + }, + { + "epoch": 0.2018348623853211, + "grad_norm": 4.3247175216674805, + "learning_rate": 1.0078636959370904e-05, + "loss": 0.6994, + "num_input_tokens_seen": 210240, + "step": 770 + }, + { + "epoch": 0.20314547837483618, + "grad_norm": 7.593923568725586, + "learning_rate": 1.0144167758846658e-05, + "loss": 0.5848, + "num_input_tokens_seen": 211488, + "step": 775 + }, + { + "epoch": 0.20445609436435125, + "grad_norm": 2.215697765350342, + "learning_rate": 1.0209698558322412e-05, + "loss": 0.6479, + "num_input_tokens_seen": 212880, + "step": 780 + }, + { + "epoch": 0.20576671035386632, + "grad_norm": 3.3179214000701904, + "learning_rate": 1.0275229357798166e-05, + "loss": 0.808, + "num_input_tokens_seen": 213904, + "step": 785 + }, + { + "epoch": 0.20707732634338138, + "grad_norm": 2.6715049743652344, + "learning_rate": 1.034076015727392e-05, + "loss": 0.7652, + "num_input_tokens_seen": 215632, + "step": 790 + }, + { + "epoch": 0.20838794233289645, + "grad_norm": 7.10101318359375, + "learning_rate": 1.0406290956749674e-05, + "loss": 0.8782, + "num_input_tokens_seen": 216832, + "step": 795 + }, + { + "epoch": 0.20969855832241152, + "grad_norm": 4.105067729949951, + "learning_rate": 1.0471821756225426e-05, + "loss": 0.7608, + "num_input_tokens_seen": 217920, + "step": 800 + }, + { + "epoch": 0.21100917431192662, + "grad_norm": 8.440079689025879, + "learning_rate": 1.053735255570118e-05, + "loss": 0.7095, + "num_input_tokens_seen": 219600, + "step": 805 + }, + { + "epoch": 0.21231979030144169, + "grad_norm": 3.3597910404205322, + "learning_rate": 1.0602883355176934e-05, + "loss": 0.6632, + "num_input_tokens_seen": 220816, + "step": 810 + }, + { + "epoch": 0.21363040629095675, + "grad_norm": 19.218225479125977, + "learning_rate": 1.0668414154652686e-05, + "loss": 0.8523, + "num_input_tokens_seen": 222224, + "step": 815 + }, + { + "epoch": 0.21494102228047182, + "grad_norm": 2.8824307918548584, + "learning_rate": 1.0733944954128442e-05, + "loss": 0.6088, + "num_input_tokens_seen": 223968, + "step": 820 + }, + { + "epoch": 0.2162516382699869, + "grad_norm": 2.4520785808563232, + "learning_rate": 1.0799475753604196e-05, + "loss": 0.5564, + "num_input_tokens_seen": 225248, + "step": 825 + }, + { + "epoch": 0.21756225425950196, + "grad_norm": 3.0361359119415283, + "learning_rate": 1.0865006553079948e-05, + "loss": 0.7633, + "num_input_tokens_seen": 226528, + "step": 830 + }, + { + "epoch": 0.21887287024901703, + "grad_norm": 10.8978910446167, + "learning_rate": 1.0930537352555702e-05, + "loss": 0.5007, + "num_input_tokens_seen": 227568, + "step": 835 + }, + { + "epoch": 0.22018348623853212, + "grad_norm": 12.54814338684082, + "learning_rate": 1.0996068152031456e-05, + "loss": 0.559, + "num_input_tokens_seen": 228544, + "step": 840 + }, + { + "epoch": 0.2214941022280472, + "grad_norm": 1.4517916440963745, + "learning_rate": 1.1061598951507208e-05, + "loss": 0.5913, + "num_input_tokens_seen": 229712, + "step": 845 + }, + { + "epoch": 0.22280471821756226, + "grad_norm": 15.90205192565918, + "learning_rate": 1.1127129750982962e-05, + "loss": 0.7169, + "num_input_tokens_seen": 231120, + "step": 850 + }, + { + "epoch": 0.22411533420707733, + "grad_norm": 4.795571327209473, + "learning_rate": 1.1192660550458717e-05, + "loss": 0.6716, + "num_input_tokens_seen": 232368, + "step": 855 + }, + { + "epoch": 0.2254259501965924, + "grad_norm": 1.621397614479065, + "learning_rate": 1.125819134993447e-05, + "loss": 0.6908, + "num_input_tokens_seen": 233680, + "step": 860 + }, + { + "epoch": 0.22673656618610746, + "grad_norm": 4.189392566680908, + "learning_rate": 1.1323722149410223e-05, + "loss": 0.505, + "num_input_tokens_seen": 234896, + "step": 865 + }, + { + "epoch": 0.22804718217562253, + "grad_norm": 3.772437334060669, + "learning_rate": 1.1389252948885977e-05, + "loss": 0.4439, + "num_input_tokens_seen": 236800, + "step": 870 + }, + { + "epoch": 0.22935779816513763, + "grad_norm": 8.723319053649902, + "learning_rate": 1.145478374836173e-05, + "loss": 0.601, + "num_input_tokens_seen": 238272, + "step": 875 + }, + { + "epoch": 0.2306684141546527, + "grad_norm": 2.401005744934082, + "learning_rate": 1.1520314547837483e-05, + "loss": 0.7426, + "num_input_tokens_seen": 239408, + "step": 880 + }, + { + "epoch": 0.23197903014416776, + "grad_norm": 3.8032524585723877, + "learning_rate": 1.1585845347313237e-05, + "loss": 0.8322, + "num_input_tokens_seen": 240528, + "step": 885 + }, + { + "epoch": 0.23328964613368283, + "grad_norm": 3.4565324783325195, + "learning_rate": 1.1651376146788991e-05, + "loss": 0.5646, + "num_input_tokens_seen": 241696, + "step": 890 + }, + { + "epoch": 0.2346002621231979, + "grad_norm": 7.788290500640869, + "learning_rate": 1.1716906946264745e-05, + "loss": 0.5562, + "num_input_tokens_seen": 243728, + "step": 895 + }, + { + "epoch": 0.23591087811271297, + "grad_norm": 2.38454008102417, + "learning_rate": 1.1782437745740499e-05, + "loss": 0.7314, + "num_input_tokens_seen": 244912, + "step": 900 + }, + { + "epoch": 0.23722149410222804, + "grad_norm": 4.002533435821533, + "learning_rate": 1.1847968545216253e-05, + "loss": 0.5366, + "num_input_tokens_seen": 246016, + "step": 905 + }, + { + "epoch": 0.23853211009174313, + "grad_norm": 16.818267822265625, + "learning_rate": 1.1913499344692005e-05, + "loss": 0.597, + "num_input_tokens_seen": 247552, + "step": 910 + }, + { + "epoch": 0.2398427260812582, + "grad_norm": 2.5908501148223877, + "learning_rate": 1.197903014416776e-05, + "loss": 0.8315, + "num_input_tokens_seen": 248864, + "step": 915 + }, + { + "epoch": 0.24115334207077327, + "grad_norm": 2.812532663345337, + "learning_rate": 1.2044560943643513e-05, + "loss": 0.486, + "num_input_tokens_seen": 250304, + "step": 920 + }, + { + "epoch": 0.24246395806028834, + "grad_norm": 1.5494157075881958, + "learning_rate": 1.2110091743119267e-05, + "loss": 0.5463, + "num_input_tokens_seen": 251472, + "step": 925 + }, + { + "epoch": 0.2437745740498034, + "grad_norm": 3.4146125316619873, + "learning_rate": 1.2175622542595021e-05, + "loss": 0.6133, + "num_input_tokens_seen": 252624, + "step": 930 + }, + { + "epoch": 0.24508519003931847, + "grad_norm": 2.8869972229003906, + "learning_rate": 1.2241153342070775e-05, + "loss": 0.5824, + "num_input_tokens_seen": 253824, + "step": 935 + }, + { + "epoch": 0.24639580602883354, + "grad_norm": 3.6604082584381104, + "learning_rate": 1.2306684141546527e-05, + "loss": 0.6856, + "num_input_tokens_seen": 255184, + "step": 940 + }, + { + "epoch": 0.24770642201834864, + "grad_norm": 3.377242088317871, + "learning_rate": 1.2372214941022281e-05, + "loss": 0.4847, + "num_input_tokens_seen": 256640, + "step": 945 + }, + { + "epoch": 0.2490170380078637, + "grad_norm": 4.452298164367676, + "learning_rate": 1.2437745740498035e-05, + "loss": 0.7604, + "num_input_tokens_seen": 257840, + "step": 950 + }, + { + "epoch": 0.2503276539973788, + "grad_norm": 3.8565943241119385, + "learning_rate": 1.2503276539973787e-05, + "loss": 0.5509, + "num_input_tokens_seen": 259440, + "step": 955 + }, + { + "epoch": 0.2516382699868938, + "grad_norm": 4.669867992401123, + "learning_rate": 1.2568807339449543e-05, + "loss": 0.6799, + "num_input_tokens_seen": 260736, + "step": 960 + }, + { + "epoch": 0.2529488859764089, + "grad_norm": 8.861310005187988, + "learning_rate": 1.2634338138925295e-05, + "loss": 0.7417, + "num_input_tokens_seen": 261872, + "step": 965 + }, + { + "epoch": 0.254259501965924, + "grad_norm": 2.854640245437622, + "learning_rate": 1.2699868938401049e-05, + "loss": 1.1318, + "num_input_tokens_seen": 262736, + "step": 970 + }, + { + "epoch": 0.25557011795543905, + "grad_norm": 5.866477012634277, + "learning_rate": 1.2765399737876801e-05, + "loss": 0.5794, + "num_input_tokens_seen": 264064, + "step": 975 + }, + { + "epoch": 0.25688073394495414, + "grad_norm": 17.72063636779785, + "learning_rate": 1.2830930537352557e-05, + "loss": 0.7958, + "num_input_tokens_seen": 265536, + "step": 980 + }, + { + "epoch": 0.2581913499344692, + "grad_norm": 3.2275748252868652, + "learning_rate": 1.289646133682831e-05, + "loss": 0.6775, + "num_input_tokens_seen": 267296, + "step": 985 + }, + { + "epoch": 0.2595019659239843, + "grad_norm": 7.504053592681885, + "learning_rate": 1.2961992136304063e-05, + "loss": 0.762, + "num_input_tokens_seen": 268816, + "step": 990 + }, + { + "epoch": 0.2608125819134993, + "grad_norm": 11.147927284240723, + "learning_rate": 1.3027522935779818e-05, + "loss": 0.8429, + "num_input_tokens_seen": 270112, + "step": 995 + }, + { + "epoch": 0.2621231979030144, + "grad_norm": 2.4582149982452393, + "learning_rate": 1.309305373525557e-05, + "loss": 0.6889, + "num_input_tokens_seen": 271280, + "step": 1000 + }, + { + "epoch": 0.2634338138925295, + "grad_norm": 4.680461406707764, + "learning_rate": 1.3158584534731325e-05, + "loss": 0.6356, + "num_input_tokens_seen": 272304, + "step": 1005 + }, + { + "epoch": 0.26474442988204455, + "grad_norm": 7.048578262329102, + "learning_rate": 1.3224115334207077e-05, + "loss": 0.7831, + "num_input_tokens_seen": 273440, + "step": 1010 + }, + { + "epoch": 0.26605504587155965, + "grad_norm": 4.305698871612549, + "learning_rate": 1.328964613368283e-05, + "loss": 0.454, + "num_input_tokens_seen": 274736, + "step": 1015 + }, + { + "epoch": 0.2673656618610747, + "grad_norm": 2.348912000656128, + "learning_rate": 1.3355176933158586e-05, + "loss": 0.732, + "num_input_tokens_seen": 275984, + "step": 1020 + }, + { + "epoch": 0.2686762778505898, + "grad_norm": 2.179819107055664, + "learning_rate": 1.3420707732634339e-05, + "loss": 0.8033, + "num_input_tokens_seen": 277520, + "step": 1025 + }, + { + "epoch": 0.2699868938401048, + "grad_norm": 2.592933416366577, + "learning_rate": 1.3486238532110092e-05, + "loss": 0.5731, + "num_input_tokens_seen": 278928, + "step": 1030 + }, + { + "epoch": 0.2712975098296199, + "grad_norm": 8.50871753692627, + "learning_rate": 1.3551769331585845e-05, + "loss": 0.841, + "num_input_tokens_seen": 280480, + "step": 1035 + }, + { + "epoch": 0.272608125819135, + "grad_norm": 9.32423210144043, + "learning_rate": 1.36173001310616e-05, + "loss": 0.6013, + "num_input_tokens_seen": 281552, + "step": 1040 + }, + { + "epoch": 0.27391874180865006, + "grad_norm": 3.5568161010742188, + "learning_rate": 1.3682830930537352e-05, + "loss": 0.5993, + "num_input_tokens_seen": 282576, + "step": 1045 + }, + { + "epoch": 0.27522935779816515, + "grad_norm": 6.473196983337402, + "learning_rate": 1.3748361730013106e-05, + "loss": 0.5306, + "num_input_tokens_seen": 283824, + "step": 1050 + }, + { + "epoch": 0.2765399737876802, + "grad_norm": 2.7421183586120605, + "learning_rate": 1.3813892529488862e-05, + "loss": 0.665, + "num_input_tokens_seen": 285104, + "step": 1055 + }, + { + "epoch": 0.2778505897771953, + "grad_norm": 2.306835889816284, + "learning_rate": 1.3879423328964614e-05, + "loss": 0.5755, + "num_input_tokens_seen": 286560, + "step": 1060 + }, + { + "epoch": 0.27916120576671033, + "grad_norm": 21.754440307617188, + "learning_rate": 1.3944954128440368e-05, + "loss": 0.556, + "num_input_tokens_seen": 287888, + "step": 1065 + }, + { + "epoch": 0.2804718217562254, + "grad_norm": 4.792845726013184, + "learning_rate": 1.401048492791612e-05, + "loss": 0.5712, + "num_input_tokens_seen": 289216, + "step": 1070 + }, + { + "epoch": 0.2817824377457405, + "grad_norm": 1.573318362236023, + "learning_rate": 1.4076015727391876e-05, + "loss": 0.8578, + "num_input_tokens_seen": 290464, + "step": 1075 + }, + { + "epoch": 0.28309305373525556, + "grad_norm": 1.9565222263336182, + "learning_rate": 1.4141546526867626e-05, + "loss": 0.4946, + "num_input_tokens_seen": 291488, + "step": 1080 + }, + { + "epoch": 0.28440366972477066, + "grad_norm": 8.727396965026855, + "learning_rate": 1.4207077326343382e-05, + "loss": 0.735, + "num_input_tokens_seen": 293008, + "step": 1085 + }, + { + "epoch": 0.2857142857142857, + "grad_norm": 5.919317245483398, + "learning_rate": 1.4272608125819138e-05, + "loss": 0.4387, + "num_input_tokens_seen": 294560, + "step": 1090 + }, + { + "epoch": 0.2870249017038008, + "grad_norm": 3.4362902641296387, + "learning_rate": 1.4338138925294888e-05, + "loss": 0.4508, + "num_input_tokens_seen": 295824, + "step": 1095 + }, + { + "epoch": 0.28833551769331583, + "grad_norm": 3.793881416320801, + "learning_rate": 1.4403669724770644e-05, + "loss": 0.6034, + "num_input_tokens_seen": 297088, + "step": 1100 + }, + { + "epoch": 0.28964613368283093, + "grad_norm": 2.614691734313965, + "learning_rate": 1.4469200524246396e-05, + "loss": 0.6647, + "num_input_tokens_seen": 298128, + "step": 1105 + }, + { + "epoch": 0.290956749672346, + "grad_norm": 4.212225914001465, + "learning_rate": 1.453473132372215e-05, + "loss": 0.6422, + "num_input_tokens_seen": 299328, + "step": 1110 + }, + { + "epoch": 0.29226736566186107, + "grad_norm": 3.2870635986328125, + "learning_rate": 1.4600262123197902e-05, + "loss": 0.474, + "num_input_tokens_seen": 300720, + "step": 1115 + }, + { + "epoch": 0.29357798165137616, + "grad_norm": 12.251633644104004, + "learning_rate": 1.4665792922673658e-05, + "loss": 0.7106, + "num_input_tokens_seen": 302240, + "step": 1120 + }, + { + "epoch": 0.2948885976408912, + "grad_norm": 6.208831310272217, + "learning_rate": 1.4731323722149412e-05, + "loss": 0.8764, + "num_input_tokens_seen": 303440, + "step": 1125 + }, + { + "epoch": 0.2961992136304063, + "grad_norm": 3.3015620708465576, + "learning_rate": 1.4796854521625164e-05, + "loss": 0.6382, + "num_input_tokens_seen": 304704, + "step": 1130 + }, + { + "epoch": 0.29750982961992134, + "grad_norm": 4.440910816192627, + "learning_rate": 1.486238532110092e-05, + "loss": 0.6772, + "num_input_tokens_seen": 305984, + "step": 1135 + }, + { + "epoch": 0.29882044560943644, + "grad_norm": 2.1689233779907227, + "learning_rate": 1.4927916120576672e-05, + "loss": 0.5833, + "num_input_tokens_seen": 307296, + "step": 1140 + }, + { + "epoch": 0.30013106159895153, + "grad_norm": 17.993671417236328, + "learning_rate": 1.4993446920052426e-05, + "loss": 0.6833, + "num_input_tokens_seen": 308352, + "step": 1145 + }, + { + "epoch": 0.30144167758846657, + "grad_norm": 3.6551010608673096, + "learning_rate": 1.5058977719528178e-05, + "loss": 0.5611, + "num_input_tokens_seen": 309344, + "step": 1150 + }, + { + "epoch": 0.30275229357798167, + "grad_norm": 2.8117897510528564, + "learning_rate": 1.5124508519003932e-05, + "loss": 0.4859, + "num_input_tokens_seen": 310800, + "step": 1155 + }, + { + "epoch": 0.3040629095674967, + "grad_norm": 12.157155990600586, + "learning_rate": 1.5190039318479687e-05, + "loss": 0.8057, + "num_input_tokens_seen": 312528, + "step": 1160 + }, + { + "epoch": 0.3053735255570118, + "grad_norm": 2.3299484252929688, + "learning_rate": 1.525557011795544e-05, + "loss": 0.6515, + "num_input_tokens_seen": 313760, + "step": 1165 + }, + { + "epoch": 0.30668414154652685, + "grad_norm": 7.617490768432617, + "learning_rate": 1.5321100917431195e-05, + "loss": 0.6007, + "num_input_tokens_seen": 315136, + "step": 1170 + }, + { + "epoch": 0.30799475753604194, + "grad_norm": 3.9842963218688965, + "learning_rate": 1.5386631716906946e-05, + "loss": 0.5035, + "num_input_tokens_seen": 316256, + "step": 1175 + }, + { + "epoch": 0.30930537352555704, + "grad_norm": 3.059326648712158, + "learning_rate": 1.54521625163827e-05, + "loss": 0.6468, + "num_input_tokens_seen": 318416, + "step": 1180 + }, + { + "epoch": 0.3106159895150721, + "grad_norm": 3.545830488204956, + "learning_rate": 1.5517693315858454e-05, + "loss": 0.7404, + "num_input_tokens_seen": 319520, + "step": 1185 + }, + { + "epoch": 0.3119266055045872, + "grad_norm": 2.2521426677703857, + "learning_rate": 1.5583224115334208e-05, + "loss": 0.5237, + "num_input_tokens_seen": 321104, + "step": 1190 + }, + { + "epoch": 0.3132372214941022, + "grad_norm": 2.3522818088531494, + "learning_rate": 1.564875491480996e-05, + "loss": 0.7657, + "num_input_tokens_seen": 322912, + "step": 1195 + }, + { + "epoch": 0.3145478374836173, + "grad_norm": 2.222841501235962, + "learning_rate": 1.5714285714285715e-05, + "loss": 0.5862, + "num_input_tokens_seen": 324176, + "step": 1200 + }, + { + "epoch": 0.31585845347313235, + "grad_norm": 3.5261142253875732, + "learning_rate": 1.577981651376147e-05, + "loss": 0.6482, + "num_input_tokens_seen": 325168, + "step": 1205 + }, + { + "epoch": 0.31716906946264745, + "grad_norm": 3.775074005126953, + "learning_rate": 1.5845347313237223e-05, + "loss": 0.4767, + "num_input_tokens_seen": 326448, + "step": 1210 + }, + { + "epoch": 0.31847968545216254, + "grad_norm": 4.996596336364746, + "learning_rate": 1.5910878112712977e-05, + "loss": 0.4404, + "num_input_tokens_seen": 327872, + "step": 1215 + }, + { + "epoch": 0.3197903014416776, + "grad_norm": 7.50034236907959, + "learning_rate": 1.5976408912188728e-05, + "loss": 1.1635, + "num_input_tokens_seen": 328992, + "step": 1220 + }, + { + "epoch": 0.3211009174311927, + "grad_norm": 8.115758895874023, + "learning_rate": 1.604193971166448e-05, + "loss": 0.6725, + "num_input_tokens_seen": 330240, + "step": 1225 + }, + { + "epoch": 0.3224115334207077, + "grad_norm": 3.9682509899139404, + "learning_rate": 1.610747051114024e-05, + "loss": 0.7667, + "num_input_tokens_seen": 331696, + "step": 1230 + }, + { + "epoch": 0.3237221494102228, + "grad_norm": 4.420391082763672, + "learning_rate": 1.617300131061599e-05, + "loss": 0.6587, + "num_input_tokens_seen": 332688, + "step": 1235 + }, + { + "epoch": 0.32503276539973786, + "grad_norm": 8.135515213012695, + "learning_rate": 1.6238532110091743e-05, + "loss": 0.5869, + "num_input_tokens_seen": 334080, + "step": 1240 + }, + { + "epoch": 0.32634338138925295, + "grad_norm": 2.8545074462890625, + "learning_rate": 1.6304062909567497e-05, + "loss": 0.8142, + "num_input_tokens_seen": 335408, + "step": 1245 + }, + { + "epoch": 0.32765399737876805, + "grad_norm": 5.277438163757324, + "learning_rate": 1.636959370904325e-05, + "loss": 0.5955, + "num_input_tokens_seen": 336816, + "step": 1250 + }, + { + "epoch": 0.3289646133682831, + "grad_norm": 4.19973611831665, + "learning_rate": 1.6435124508519005e-05, + "loss": 0.5377, + "num_input_tokens_seen": 337984, + "step": 1255 + }, + { + "epoch": 0.3302752293577982, + "grad_norm": 2.189424753189087, + "learning_rate": 1.650065530799476e-05, + "loss": 0.5812, + "num_input_tokens_seen": 339504, + "step": 1260 + }, + { + "epoch": 0.3315858453473132, + "grad_norm": 4.466281890869141, + "learning_rate": 1.6566186107470513e-05, + "loss": 0.4981, + "num_input_tokens_seen": 340544, + "step": 1265 + }, + { + "epoch": 0.3328964613368283, + "grad_norm": 4.341704368591309, + "learning_rate": 1.6631716906946267e-05, + "loss": 0.5965, + "num_input_tokens_seen": 341488, + "step": 1270 + }, + { + "epoch": 0.33420707732634336, + "grad_norm": 2.0648202896118164, + "learning_rate": 1.669724770642202e-05, + "loss": 0.5534, + "num_input_tokens_seen": 342736, + "step": 1275 + }, + { + "epoch": 0.33551769331585846, + "grad_norm": 2.937502384185791, + "learning_rate": 1.676277850589777e-05, + "loss": 0.6023, + "num_input_tokens_seen": 344000, + "step": 1280 + }, + { + "epoch": 0.33682830930537355, + "grad_norm": 9.486555099487305, + "learning_rate": 1.682830930537353e-05, + "loss": 0.7386, + "num_input_tokens_seen": 345056, + "step": 1285 + }, + { + "epoch": 0.3381389252948886, + "grad_norm": 6.755458354949951, + "learning_rate": 1.689384010484928e-05, + "loss": 0.6564, + "num_input_tokens_seen": 346080, + "step": 1290 + }, + { + "epoch": 0.3394495412844037, + "grad_norm": 6.814481735229492, + "learning_rate": 1.6959370904325033e-05, + "loss": 0.694, + "num_input_tokens_seen": 347360, + "step": 1295 + }, + { + "epoch": 0.34076015727391873, + "grad_norm": 9.723411560058594, + "learning_rate": 1.702490170380079e-05, + "loss": 0.7216, + "num_input_tokens_seen": 348464, + "step": 1300 + }, + { + "epoch": 0.3420707732634338, + "grad_norm": 2.815275192260742, + "learning_rate": 1.709043250327654e-05, + "loss": 0.571, + "num_input_tokens_seen": 349952, + "step": 1305 + }, + { + "epoch": 0.34338138925294887, + "grad_norm": 2.5255422592163086, + "learning_rate": 1.7155963302752295e-05, + "loss": 0.7713, + "num_input_tokens_seen": 351296, + "step": 1310 + }, + { + "epoch": 0.34469200524246396, + "grad_norm": 1.3294346332550049, + "learning_rate": 1.722149410222805e-05, + "loss": 0.552, + "num_input_tokens_seen": 353232, + "step": 1315 + }, + { + "epoch": 0.34600262123197906, + "grad_norm": 3.2029519081115723, + "learning_rate": 1.7287024901703802e-05, + "loss": 0.4789, + "num_input_tokens_seen": 354640, + "step": 1320 + }, + { + "epoch": 0.3473132372214941, + "grad_norm": 13.984296798706055, + "learning_rate": 1.7352555701179553e-05, + "loss": 0.5216, + "num_input_tokens_seen": 355856, + "step": 1325 + }, + { + "epoch": 0.3486238532110092, + "grad_norm": 6.176485538482666, + "learning_rate": 1.741808650065531e-05, + "loss": 0.659, + "num_input_tokens_seen": 356992, + "step": 1330 + }, + { + "epoch": 0.34993446920052423, + "grad_norm": 5.0185227394104, + "learning_rate": 1.7483617300131064e-05, + "loss": 0.5779, + "num_input_tokens_seen": 358256, + "step": 1335 + }, + { + "epoch": 0.35124508519003933, + "grad_norm": 2.992762804031372, + "learning_rate": 1.7549148099606815e-05, + "loss": 0.628, + "num_input_tokens_seen": 359376, + "step": 1340 + }, + { + "epoch": 0.35255570117955437, + "grad_norm": 5.489147186279297, + "learning_rate": 1.7614678899082572e-05, + "loss": 0.5435, + "num_input_tokens_seen": 361056, + "step": 1345 + }, + { + "epoch": 0.35386631716906947, + "grad_norm": 7.373166561126709, + "learning_rate": 1.7680209698558323e-05, + "loss": 0.6277, + "num_input_tokens_seen": 362640, + "step": 1350 + }, + { + "epoch": 0.35517693315858456, + "grad_norm": 2.723423480987549, + "learning_rate": 1.7745740498034076e-05, + "loss": 0.6669, + "num_input_tokens_seen": 364256, + "step": 1355 + }, + { + "epoch": 0.3564875491480996, + "grad_norm": 4.253269195556641, + "learning_rate": 1.781127129750983e-05, + "loss": 0.6994, + "num_input_tokens_seen": 365344, + "step": 1360 + }, + { + "epoch": 0.3577981651376147, + "grad_norm": 7.307094097137451, + "learning_rate": 1.7876802096985584e-05, + "loss": 0.6085, + "num_input_tokens_seen": 366368, + "step": 1365 + }, + { + "epoch": 0.35910878112712974, + "grad_norm": 3.5286130905151367, + "learning_rate": 1.7942332896461335e-05, + "loss": 0.514, + "num_input_tokens_seen": 367920, + "step": 1370 + }, + { + "epoch": 0.36041939711664484, + "grad_norm": 2.1979024410247803, + "learning_rate": 1.8007863695937092e-05, + "loss": 0.6916, + "num_input_tokens_seen": 369104, + "step": 1375 + }, + { + "epoch": 0.3617300131061599, + "grad_norm": 2.188822031021118, + "learning_rate": 1.8073394495412846e-05, + "loss": 0.5604, + "num_input_tokens_seen": 370768, + "step": 1380 + }, + { + "epoch": 0.36304062909567497, + "grad_norm": 11.335149765014648, + "learning_rate": 1.8138925294888597e-05, + "loss": 0.5826, + "num_input_tokens_seen": 371952, + "step": 1385 + }, + { + "epoch": 0.36435124508519, + "grad_norm": 2.85606050491333, + "learning_rate": 1.8204456094364354e-05, + "loss": 0.7948, + "num_input_tokens_seen": 373104, + "step": 1390 + }, + { + "epoch": 0.3656618610747051, + "grad_norm": 3.589306116104126, + "learning_rate": 1.8269986893840104e-05, + "loss": 0.5782, + "num_input_tokens_seen": 374384, + "step": 1395 + }, + { + "epoch": 0.3669724770642202, + "grad_norm": 6.287440299987793, + "learning_rate": 1.833551769331586e-05, + "loss": 0.7293, + "num_input_tokens_seen": 375344, + "step": 1400 + }, + { + "epoch": 0.36828309305373524, + "grad_norm": 4.452231407165527, + "learning_rate": 1.8401048492791612e-05, + "loss": 0.5225, + "num_input_tokens_seen": 376960, + "step": 1405 + }, + { + "epoch": 0.36959370904325034, + "grad_norm": 4.26068115234375, + "learning_rate": 1.8466579292267366e-05, + "loss": 0.7043, + "num_input_tokens_seen": 378352, + "step": 1410 + }, + { + "epoch": 0.3709043250327654, + "grad_norm": 2.01566743850708, + "learning_rate": 1.853211009174312e-05, + "loss": 0.627, + "num_input_tokens_seen": 379488, + "step": 1415 + }, + { + "epoch": 0.3722149410222805, + "grad_norm": 4.144623279571533, + "learning_rate": 1.8597640891218874e-05, + "loss": 0.702, + "num_input_tokens_seen": 380752, + "step": 1420 + }, + { + "epoch": 0.3735255570117955, + "grad_norm": 5.356663227081299, + "learning_rate": 1.8663171690694628e-05, + "loss": 0.5308, + "num_input_tokens_seen": 382400, + "step": 1425 + }, + { + "epoch": 0.3748361730013106, + "grad_norm": 2.393522262573242, + "learning_rate": 1.872870249017038e-05, + "loss": 0.7334, + "num_input_tokens_seen": 383680, + "step": 1430 + }, + { + "epoch": 0.3761467889908257, + "grad_norm": 5.889967441558838, + "learning_rate": 1.8794233289646136e-05, + "loss": 0.5525, + "num_input_tokens_seen": 385088, + "step": 1435 + }, + { + "epoch": 0.37745740498034075, + "grad_norm": 12.11541748046875, + "learning_rate": 1.8859764089121886e-05, + "loss": 0.5962, + "num_input_tokens_seen": 386080, + "step": 1440 + }, + { + "epoch": 0.37876802096985585, + "grad_norm": 2.6202778816223145, + "learning_rate": 1.892529488859764e-05, + "loss": 0.7397, + "num_input_tokens_seen": 387440, + "step": 1445 + }, + { + "epoch": 0.3800786369593709, + "grad_norm": 9.354446411132812, + "learning_rate": 1.8990825688073397e-05, + "loss": 0.573, + "num_input_tokens_seen": 388336, + "step": 1450 + }, + { + "epoch": 0.381389252948886, + "grad_norm": 3.525128126144409, + "learning_rate": 1.9056356487549148e-05, + "loss": 0.5836, + "num_input_tokens_seen": 389728, + "step": 1455 + }, + { + "epoch": 0.382699868938401, + "grad_norm": 2.58384370803833, + "learning_rate": 1.9121887287024902e-05, + "loss": 0.5212, + "num_input_tokens_seen": 390864, + "step": 1460 + }, + { + "epoch": 0.3840104849279161, + "grad_norm": 2.093106985092163, + "learning_rate": 1.9187418086500656e-05, + "loss": 0.7438, + "num_input_tokens_seen": 392576, + "step": 1465 + }, + { + "epoch": 0.3853211009174312, + "grad_norm": 2.2800204753875732, + "learning_rate": 1.925294888597641e-05, + "loss": 0.4338, + "num_input_tokens_seen": 393808, + "step": 1470 + }, + { + "epoch": 0.38663171690694625, + "grad_norm": 2.3157079219818115, + "learning_rate": 1.9318479685452164e-05, + "loss": 0.5033, + "num_input_tokens_seen": 395312, + "step": 1475 + }, + { + "epoch": 0.38794233289646135, + "grad_norm": 4.910935401916504, + "learning_rate": 1.9384010484927918e-05, + "loss": 0.777, + "num_input_tokens_seen": 396864, + "step": 1480 + }, + { + "epoch": 0.3892529488859764, + "grad_norm": 8.051618576049805, + "learning_rate": 1.944954128440367e-05, + "loss": 0.609, + "num_input_tokens_seen": 398048, + "step": 1485 + }, + { + "epoch": 0.3905635648754915, + "grad_norm": 4.279458045959473, + "learning_rate": 1.9515072083879425e-05, + "loss": 0.6009, + "num_input_tokens_seen": 399040, + "step": 1490 + }, + { + "epoch": 0.3918741808650065, + "grad_norm": 3.9840071201324463, + "learning_rate": 1.958060288335518e-05, + "loss": 0.6186, + "num_input_tokens_seen": 400016, + "step": 1495 + }, + { + "epoch": 0.3931847968545216, + "grad_norm": 3.5082945823669434, + "learning_rate": 1.964613368283093e-05, + "loss": 0.6021, + "num_input_tokens_seen": 400992, + "step": 1500 + }, + { + "epoch": 0.3944954128440367, + "grad_norm": 4.162998199462891, + "learning_rate": 1.9711664482306684e-05, + "loss": 0.5423, + "num_input_tokens_seen": 402672, + "step": 1505 + }, + { + "epoch": 0.39580602883355176, + "grad_norm": 10.994751930236816, + "learning_rate": 1.9777195281782438e-05, + "loss": 0.6125, + "num_input_tokens_seen": 403664, + "step": 1510 + }, + { + "epoch": 0.39711664482306686, + "grad_norm": 19.227764129638672, + "learning_rate": 1.984272608125819e-05, + "loss": 0.4828, + "num_input_tokens_seen": 405024, + "step": 1515 + }, + { + "epoch": 0.3984272608125819, + "grad_norm": 9.503660202026367, + "learning_rate": 1.9908256880733945e-05, + "loss": 0.6125, + "num_input_tokens_seen": 406672, + "step": 1520 + }, + { + "epoch": 0.399737876802097, + "grad_norm": 9.269362449645996, + "learning_rate": 1.99737876802097e-05, + "loss": 0.4958, + "num_input_tokens_seen": 407840, + "step": 1525 + }, + { + "epoch": 0.40104849279161203, + "grad_norm": 7.678936958312988, + "learning_rate": 2.0039318479685453e-05, + "loss": 0.5342, + "num_input_tokens_seen": 408992, + "step": 1530 + }, + { + "epoch": 0.40235910878112713, + "grad_norm": 5.488819599151611, + "learning_rate": 2.0104849279161207e-05, + "loss": 0.4858, + "num_input_tokens_seen": 410240, + "step": 1535 + }, + { + "epoch": 0.4036697247706422, + "grad_norm": 1.8297392129898071, + "learning_rate": 2.017038007863696e-05, + "loss": 1.8382, + "num_input_tokens_seen": 411456, + "step": 1540 + }, + { + "epoch": 0.40498034076015726, + "grad_norm": 3.6354527473449707, + "learning_rate": 2.023591087811271e-05, + "loss": 0.6659, + "num_input_tokens_seen": 412496, + "step": 1545 + }, + { + "epoch": 0.40629095674967236, + "grad_norm": 3.4280247688293457, + "learning_rate": 2.030144167758847e-05, + "loss": 0.5698, + "num_input_tokens_seen": 413792, + "step": 1550 + }, + { + "epoch": 0.4076015727391874, + "grad_norm": 2.239339590072632, + "learning_rate": 2.0366972477064223e-05, + "loss": 0.5335, + "num_input_tokens_seen": 414880, + "step": 1555 + }, + { + "epoch": 0.4089121887287025, + "grad_norm": 2.477618932723999, + "learning_rate": 2.0432503276539973e-05, + "loss": 0.5319, + "num_input_tokens_seen": 416608, + "step": 1560 + }, + { + "epoch": 0.41022280471821754, + "grad_norm": 2.1741421222686768, + "learning_rate": 2.049803407601573e-05, + "loss": 0.4645, + "num_input_tokens_seen": 418336, + "step": 1565 + }, + { + "epoch": 0.41153342070773263, + "grad_norm": 3.7242610454559326, + "learning_rate": 2.056356487549148e-05, + "loss": 0.4515, + "num_input_tokens_seen": 419616, + "step": 1570 + }, + { + "epoch": 0.41284403669724773, + "grad_norm": 4.856847286224365, + "learning_rate": 2.0629095674967235e-05, + "loss": 0.7138, + "num_input_tokens_seen": 421088, + "step": 1575 + }, + { + "epoch": 0.41415465268676277, + "grad_norm": 7.142799377441406, + "learning_rate": 2.069462647444299e-05, + "loss": 0.5848, + "num_input_tokens_seen": 422352, + "step": 1580 + }, + { + "epoch": 0.41546526867627787, + "grad_norm": 10.2919340133667, + "learning_rate": 2.0760157273918743e-05, + "loss": 0.5916, + "num_input_tokens_seen": 423232, + "step": 1585 + }, + { + "epoch": 0.4167758846657929, + "grad_norm": 5.277079105377197, + "learning_rate": 2.0825688073394497e-05, + "loss": 0.4644, + "num_input_tokens_seen": 424768, + "step": 1590 + }, + { + "epoch": 0.418086500655308, + "grad_norm": 6.021617889404297, + "learning_rate": 2.089121887287025e-05, + "loss": 0.8739, + "num_input_tokens_seen": 425968, + "step": 1595 + }, + { + "epoch": 0.41939711664482304, + "grad_norm": 4.737050533294678, + "learning_rate": 2.0956749672346005e-05, + "loss": 0.6586, + "num_input_tokens_seen": 427216, + "step": 1600 + }, + { + "epoch": 0.42070773263433814, + "grad_norm": 2.3742330074310303, + "learning_rate": 2.1022280471821755e-05, + "loss": 0.6596, + "num_input_tokens_seen": 428656, + "step": 1605 + }, + { + "epoch": 0.42201834862385323, + "grad_norm": 9.704755783081055, + "learning_rate": 2.1087811271297513e-05, + "loss": 0.9066, + "num_input_tokens_seen": 429840, + "step": 1610 + }, + { + "epoch": 0.4233289646133683, + "grad_norm": 3.0565719604492188, + "learning_rate": 2.1153342070773263e-05, + "loss": 0.653, + "num_input_tokens_seen": 431104, + "step": 1615 + }, + { + "epoch": 0.42463958060288337, + "grad_norm": 2.2464468479156494, + "learning_rate": 2.1218872870249017e-05, + "loss": 0.3877, + "num_input_tokens_seen": 432768, + "step": 1620 + }, + { + "epoch": 0.4259501965923984, + "grad_norm": 5.296932697296143, + "learning_rate": 2.1284403669724774e-05, + "loss": 0.3949, + "num_input_tokens_seen": 434160, + "step": 1625 + }, + { + "epoch": 0.4272608125819135, + "grad_norm": 2.2421209812164307, + "learning_rate": 2.1349934469200525e-05, + "loss": 0.7281, + "num_input_tokens_seen": 435600, + "step": 1630 + }, + { + "epoch": 0.42857142857142855, + "grad_norm": 10.738851547241211, + "learning_rate": 2.141546526867628e-05, + "loss": 0.6849, + "num_input_tokens_seen": 437168, + "step": 1635 + }, + { + "epoch": 0.42988204456094364, + "grad_norm": 4.492898464202881, + "learning_rate": 2.1480996068152033e-05, + "loss": 0.7686, + "num_input_tokens_seen": 438080, + "step": 1640 + }, + { + "epoch": 0.43119266055045874, + "grad_norm": 2.7208640575408936, + "learning_rate": 2.1546526867627787e-05, + "loss": 0.4515, + "num_input_tokens_seen": 439728, + "step": 1645 + }, + { + "epoch": 0.4325032765399738, + "grad_norm": 26.380624771118164, + "learning_rate": 2.1612057667103537e-05, + "loss": 2.1613, + "num_input_tokens_seen": 440704, + "step": 1650 + }, + { + "epoch": 0.4338138925294889, + "grad_norm": 4.055468559265137, + "learning_rate": 2.1677588466579294e-05, + "loss": 0.6539, + "num_input_tokens_seen": 442160, + "step": 1655 + }, + { + "epoch": 0.4351245085190039, + "grad_norm": 1.4117684364318848, + "learning_rate": 2.1743119266055048e-05, + "loss": 0.4889, + "num_input_tokens_seen": 444480, + "step": 1660 + }, + { + "epoch": 0.436435124508519, + "grad_norm": 3.2511231899261475, + "learning_rate": 2.18086500655308e-05, + "loss": 0.5511, + "num_input_tokens_seen": 446128, + "step": 1665 + }, + { + "epoch": 0.43774574049803405, + "grad_norm": 5.170319080352783, + "learning_rate": 2.1874180865006556e-05, + "loss": 0.5549, + "num_input_tokens_seen": 447248, + "step": 1670 + }, + { + "epoch": 0.43905635648754915, + "grad_norm": 5.120906352996826, + "learning_rate": 2.1939711664482307e-05, + "loss": 0.6896, + "num_input_tokens_seen": 449392, + "step": 1675 + }, + { + "epoch": 0.44036697247706424, + "grad_norm": 1.388336181640625, + "learning_rate": 2.200524246395806e-05, + "loss": 0.7628, + "num_input_tokens_seen": 450688, + "step": 1680 + }, + { + "epoch": 0.4416775884665793, + "grad_norm": 9.989461898803711, + "learning_rate": 2.2070773263433814e-05, + "loss": 0.5147, + "num_input_tokens_seen": 451712, + "step": 1685 + }, + { + "epoch": 0.4429882044560944, + "grad_norm": 3.855394124984741, + "learning_rate": 2.213630406290957e-05, + "loss": 0.4918, + "num_input_tokens_seen": 452896, + "step": 1690 + }, + { + "epoch": 0.4442988204456094, + "grad_norm": 4.201488494873047, + "learning_rate": 2.2201834862385322e-05, + "loss": 0.6169, + "num_input_tokens_seen": 454544, + "step": 1695 + }, + { + "epoch": 0.4456094364351245, + "grad_norm": 8.127564430236816, + "learning_rate": 2.2267365661861076e-05, + "loss": 0.3792, + "num_input_tokens_seen": 455856, + "step": 1700 + }, + { + "epoch": 0.44692005242463956, + "grad_norm": 5.2498931884765625, + "learning_rate": 2.233289646133683e-05, + "loss": 0.6177, + "num_input_tokens_seen": 457200, + "step": 1705 + }, + { + "epoch": 0.44823066841415465, + "grad_norm": 9.607030868530273, + "learning_rate": 2.239842726081258e-05, + "loss": 0.6808, + "num_input_tokens_seen": 458448, + "step": 1710 + }, + { + "epoch": 0.44954128440366975, + "grad_norm": 4.274682521820068, + "learning_rate": 2.2463958060288338e-05, + "loss": 0.5534, + "num_input_tokens_seen": 459712, + "step": 1715 + }, + { + "epoch": 0.4508519003931848, + "grad_norm": 5.928576469421387, + "learning_rate": 2.252948885976409e-05, + "loss": 0.5579, + "num_input_tokens_seen": 460976, + "step": 1720 + }, + { + "epoch": 0.4521625163826999, + "grad_norm": 3.21767520904541, + "learning_rate": 2.2595019659239842e-05, + "loss": 0.5169, + "num_input_tokens_seen": 462176, + "step": 1725 + }, + { + "epoch": 0.4534731323722149, + "grad_norm": 1.9343457221984863, + "learning_rate": 2.26605504587156e-05, + "loss": 0.5262, + "num_input_tokens_seen": 463280, + "step": 1730 + }, + { + "epoch": 0.45478374836173, + "grad_norm": 7.810166358947754, + "learning_rate": 2.272608125819135e-05, + "loss": 0.6356, + "num_input_tokens_seen": 464448, + "step": 1735 + }, + { + "epoch": 0.45609436435124506, + "grad_norm": 3.5483086109161377, + "learning_rate": 2.2791612057667104e-05, + "loss": 0.7372, + "num_input_tokens_seen": 465248, + "step": 1740 + }, + { + "epoch": 0.45740498034076016, + "grad_norm": 2.797677993774414, + "learning_rate": 2.2857142857142858e-05, + "loss": 0.5883, + "num_input_tokens_seen": 466640, + "step": 1745 + }, + { + "epoch": 0.45871559633027525, + "grad_norm": 8.553635597229004, + "learning_rate": 2.2922673656618612e-05, + "loss": 0.4533, + "num_input_tokens_seen": 467872, + "step": 1750 + }, + { + "epoch": 0.4600262123197903, + "grad_norm": 2.7947206497192383, + "learning_rate": 2.2988204456094366e-05, + "loss": 0.4319, + "num_input_tokens_seen": 469344, + "step": 1755 + }, + { + "epoch": 0.4613368283093054, + "grad_norm": 4.975508213043213, + "learning_rate": 2.305373525557012e-05, + "loss": 0.9103, + "num_input_tokens_seen": 470416, + "step": 1760 + }, + { + "epoch": 0.46264744429882043, + "grad_norm": 4.240067481994629, + "learning_rate": 2.3119266055045874e-05, + "loss": 0.5567, + "num_input_tokens_seen": 471712, + "step": 1765 + }, + { + "epoch": 0.4639580602883355, + "grad_norm": 4.605864524841309, + "learning_rate": 2.3184796854521628e-05, + "loss": 0.6659, + "num_input_tokens_seen": 472944, + "step": 1770 + }, + { + "epoch": 0.46526867627785057, + "grad_norm": 1.9480764865875244, + "learning_rate": 2.325032765399738e-05, + "loss": 0.5742, + "num_input_tokens_seen": 474112, + "step": 1775 + }, + { + "epoch": 0.46657929226736566, + "grad_norm": 2.5360255241394043, + "learning_rate": 2.3315858453473132e-05, + "loss": 0.4403, + "num_input_tokens_seen": 475408, + "step": 1780 + }, + { + "epoch": 0.46788990825688076, + "grad_norm": 6.429988861083984, + "learning_rate": 2.3381389252948886e-05, + "loss": 0.5785, + "num_input_tokens_seen": 476624, + "step": 1785 + }, + { + "epoch": 0.4692005242463958, + "grad_norm": 2.5668535232543945, + "learning_rate": 2.344692005242464e-05, + "loss": 0.4503, + "num_input_tokens_seen": 478000, + "step": 1790 + }, + { + "epoch": 0.4705111402359109, + "grad_norm": 24.837604522705078, + "learning_rate": 2.3512450851900394e-05, + "loss": 0.6381, + "num_input_tokens_seen": 479072, + "step": 1795 + }, + { + "epoch": 0.47182175622542594, + "grad_norm": 2.4411888122558594, + "learning_rate": 2.3577981651376148e-05, + "loss": 0.6744, + "num_input_tokens_seen": 480160, + "step": 1800 + }, + { + "epoch": 0.47313237221494103, + "grad_norm": 2.1908984184265137, + "learning_rate": 2.36435124508519e-05, + "loss": 2.0963, + "num_input_tokens_seen": 482048, + "step": 1805 + }, + { + "epoch": 0.4744429882044561, + "grad_norm": 4.410700798034668, + "learning_rate": 2.3709043250327656e-05, + "loss": 0.4422, + "num_input_tokens_seen": 483184, + "step": 1810 + }, + { + "epoch": 0.47575360419397117, + "grad_norm": 5.100732803344727, + "learning_rate": 2.377457404980341e-05, + "loss": 0.6144, + "num_input_tokens_seen": 484608, + "step": 1815 + }, + { + "epoch": 0.47706422018348627, + "grad_norm": 3.740973949432373, + "learning_rate": 2.3840104849279163e-05, + "loss": 0.5998, + "num_input_tokens_seen": 485568, + "step": 1820 + }, + { + "epoch": 0.4783748361730013, + "grad_norm": 2.1813464164733887, + "learning_rate": 2.3905635648754914e-05, + "loss": 0.5242, + "num_input_tokens_seen": 487072, + "step": 1825 + }, + { + "epoch": 0.4796854521625164, + "grad_norm": 3.6211397647857666, + "learning_rate": 2.397116644823067e-05, + "loss": 0.6767, + "num_input_tokens_seen": 488480, + "step": 1830 + }, + { + "epoch": 0.48099606815203144, + "grad_norm": 6.580210208892822, + "learning_rate": 2.4036697247706425e-05, + "loss": 0.6906, + "num_input_tokens_seen": 489728, + "step": 1835 + }, + { + "epoch": 0.48230668414154654, + "grad_norm": 8.138657569885254, + "learning_rate": 2.4102228047182176e-05, + "loss": 0.4451, + "num_input_tokens_seen": 491808, + "step": 1840 + }, + { + "epoch": 0.4836173001310616, + "grad_norm": 3.082435131072998, + "learning_rate": 2.4167758846657933e-05, + "loss": 0.5185, + "num_input_tokens_seen": 493056, + "step": 1845 + }, + { + "epoch": 0.4849279161205767, + "grad_norm": 15.313329696655273, + "learning_rate": 2.4233289646133683e-05, + "loss": 0.539, + "num_input_tokens_seen": 494304, + "step": 1850 + }, + { + "epoch": 0.48623853211009177, + "grad_norm": 2.9238624572753906, + "learning_rate": 2.4298820445609437e-05, + "loss": 0.5576, + "num_input_tokens_seen": 496064, + "step": 1855 + }, + { + "epoch": 0.4875491480996068, + "grad_norm": 4.020905017852783, + "learning_rate": 2.436435124508519e-05, + "loss": 0.4833, + "num_input_tokens_seen": 497184, + "step": 1860 + }, + { + "epoch": 0.4888597640891219, + "grad_norm": 4.318493843078613, + "learning_rate": 2.4429882044560945e-05, + "loss": 0.5115, + "num_input_tokens_seen": 498352, + "step": 1865 + }, + { + "epoch": 0.49017038007863695, + "grad_norm": 2.0091683864593506, + "learning_rate": 2.44954128440367e-05, + "loss": 0.5449, + "num_input_tokens_seen": 499600, + "step": 1870 + }, + { + "epoch": 0.49148099606815204, + "grad_norm": 1.6362696886062622, + "learning_rate": 2.4560943643512453e-05, + "loss": 0.5156, + "num_input_tokens_seen": 501184, + "step": 1875 + }, + { + "epoch": 0.4927916120576671, + "grad_norm": 4.358371734619141, + "learning_rate": 2.4626474442988207e-05, + "loss": 0.5307, + "num_input_tokens_seen": 502496, + "step": 1880 + }, + { + "epoch": 0.4941022280471822, + "grad_norm": 3.058372735977173, + "learning_rate": 2.4692005242463957e-05, + "loss": 0.8345, + "num_input_tokens_seen": 503936, + "step": 1885 + }, + { + "epoch": 0.4954128440366973, + "grad_norm": 3.60732364654541, + "learning_rate": 2.4757536041939715e-05, + "loss": 0.5446, + "num_input_tokens_seen": 505008, + "step": 1890 + }, + { + "epoch": 0.4967234600262123, + "grad_norm": 11.88129711151123, + "learning_rate": 2.4823066841415465e-05, + "loss": 0.796, + "num_input_tokens_seen": 506272, + "step": 1895 + }, + { + "epoch": 0.4980340760157274, + "grad_norm": 9.882281303405762, + "learning_rate": 2.488859764089122e-05, + "loss": 0.691, + "num_input_tokens_seen": 507840, + "step": 1900 + }, + { + "epoch": 0.49934469200524245, + "grad_norm": 5.8761467933654785, + "learning_rate": 2.4954128440366977e-05, + "loss": 0.31, + "num_input_tokens_seen": 509088, + "step": 1905 + }, + { + "epoch": 0.5001310615989515, + "eval_loss": 0.542202353477478, + "eval_runtime": 18.1978, + "eval_samples_per_second": 46.599, + "eval_steps_per_second": 23.3, + "num_input_tokens_seen": 509696, + "step": 1908 + }, + { + "epoch": 0.5006553079947575, + "grad_norm": 6.951882839202881, + "learning_rate": 2.501965923984273e-05, + "loss": 0.4542, + "num_input_tokens_seen": 510176, + "step": 1910 + }, + { + "epoch": 0.5019659239842726, + "grad_norm": 2.847299337387085, + "learning_rate": 2.5085190039318478e-05, + "loss": 0.4988, + "num_input_tokens_seen": 511504, + "step": 1915 + }, + { + "epoch": 0.5032765399737876, + "grad_norm": 7.416575908660889, + "learning_rate": 2.5150720838794235e-05, + "loss": 0.5266, + "num_input_tokens_seen": 513104, + "step": 1920 + }, + { + "epoch": 0.5045871559633027, + "grad_norm": 3.062163829803467, + "learning_rate": 2.521625163826999e-05, + "loss": 0.3983, + "num_input_tokens_seen": 514624, + "step": 1925 + }, + { + "epoch": 0.5058977719528178, + "grad_norm": 4.941585540771484, + "learning_rate": 2.5281782437745743e-05, + "loss": 0.5054, + "num_input_tokens_seen": 516112, + "step": 1930 + }, + { + "epoch": 0.5072083879423329, + "grad_norm": 5.509173393249512, + "learning_rate": 2.5347313237221493e-05, + "loss": 0.5663, + "num_input_tokens_seen": 517392, + "step": 1935 + }, + { + "epoch": 0.508519003931848, + "grad_norm": 4.008965492248535, + "learning_rate": 2.5412844036697247e-05, + "loss": 0.8344, + "num_input_tokens_seen": 518928, + "step": 1940 + }, + { + "epoch": 0.509829619921363, + "grad_norm": 10.186779975891113, + "learning_rate": 2.5478374836173e-05, + "loss": 0.6282, + "num_input_tokens_seen": 520432, + "step": 1945 + }, + { + "epoch": 0.5111402359108781, + "grad_norm": 3.001925468444824, + "learning_rate": 2.554390563564876e-05, + "loss": 0.5453, + "num_input_tokens_seen": 521824, + "step": 1950 + }, + { + "epoch": 0.5124508519003932, + "grad_norm": 6.44737434387207, + "learning_rate": 2.5609436435124512e-05, + "loss": 0.5291, + "num_input_tokens_seen": 523056, + "step": 1955 + }, + { + "epoch": 0.5137614678899083, + "grad_norm": 4.664173126220703, + "learning_rate": 2.5674967234600263e-05, + "loss": 0.87, + "num_input_tokens_seen": 524176, + "step": 1960 + }, + { + "epoch": 0.5150720838794234, + "grad_norm": 6.643097877502441, + "learning_rate": 2.5740498034076017e-05, + "loss": 0.377, + "num_input_tokens_seen": 526304, + "step": 1965 + }, + { + "epoch": 0.5163826998689384, + "grad_norm": 8.044621467590332, + "learning_rate": 2.580602883355177e-05, + "loss": 0.577, + "num_input_tokens_seen": 527456, + "step": 1970 + }, + { + "epoch": 0.5176933158584535, + "grad_norm": 2.7140467166900635, + "learning_rate": 2.5871559633027525e-05, + "loss": 0.6028, + "num_input_tokens_seen": 529008, + "step": 1975 + }, + { + "epoch": 0.5190039318479686, + "grad_norm": 2.374124526977539, + "learning_rate": 2.5937090432503282e-05, + "loss": 0.5499, + "num_input_tokens_seen": 530432, + "step": 1980 + }, + { + "epoch": 0.5203145478374837, + "grad_norm": 0.34991654753685, + "learning_rate": 2.600262123197903e-05, + "loss": 0.3553, + "num_input_tokens_seen": 533360, + "step": 1985 + }, + { + "epoch": 0.5216251638269986, + "grad_norm": 5.177097797393799, + "learning_rate": 2.6068152031454783e-05, + "loss": 0.5345, + "num_input_tokens_seen": 534448, + "step": 1990 + }, + { + "epoch": 0.5229357798165137, + "grad_norm": 3.798830032348633, + "learning_rate": 2.613368283093054e-05, + "loss": 0.4868, + "num_input_tokens_seen": 535984, + "step": 1995 + }, + { + "epoch": 0.5242463958060288, + "grad_norm": 3.8111677169799805, + "learning_rate": 2.6199213630406294e-05, + "loss": 0.7519, + "num_input_tokens_seen": 537424, + "step": 2000 + }, + { + "epoch": 0.5255570117955439, + "grad_norm": 4.533329010009766, + "learning_rate": 2.6264744429882045e-05, + "loss": 0.4655, + "num_input_tokens_seen": 538944, + "step": 2005 + }, + { + "epoch": 0.526867627785059, + "grad_norm": 2.6256635189056396, + "learning_rate": 2.63302752293578e-05, + "loss": 0.4782, + "num_input_tokens_seen": 540416, + "step": 2010 + }, + { + "epoch": 0.528178243774574, + "grad_norm": 7.475856781005859, + "learning_rate": 2.6395806028833552e-05, + "loss": 0.4486, + "num_input_tokens_seen": 541776, + "step": 2015 + }, + { + "epoch": 0.5294888597640891, + "grad_norm": 3.3739025592803955, + "learning_rate": 2.6461336828309306e-05, + "loss": 0.6161, + "num_input_tokens_seen": 542960, + "step": 2020 + }, + { + "epoch": 0.5307994757536042, + "grad_norm": 6.266442775726318, + "learning_rate": 2.6526867627785064e-05, + "loss": 0.5884, + "num_input_tokens_seen": 544032, + "step": 2025 + }, + { + "epoch": 0.5321100917431193, + "grad_norm": 2.343364715576172, + "learning_rate": 2.659239842726081e-05, + "loss": 0.4948, + "num_input_tokens_seen": 545280, + "step": 2030 + }, + { + "epoch": 0.5334207077326344, + "grad_norm": 14.29284954071045, + "learning_rate": 2.6657929226736568e-05, + "loss": 0.517, + "num_input_tokens_seen": 546752, + "step": 2035 + }, + { + "epoch": 0.5347313237221494, + "grad_norm": 4.535431385040283, + "learning_rate": 2.6723460026212322e-05, + "loss": 0.5736, + "num_input_tokens_seen": 548528, + "step": 2040 + }, + { + "epoch": 0.5360419397116645, + "grad_norm": 1.7546025514602661, + "learning_rate": 2.6788990825688076e-05, + "loss": 0.4746, + "num_input_tokens_seen": 549792, + "step": 2045 + }, + { + "epoch": 0.5373525557011796, + "grad_norm": 2.978302478790283, + "learning_rate": 2.685452162516383e-05, + "loss": 0.5842, + "num_input_tokens_seen": 551088, + "step": 2050 + }, + { + "epoch": 0.5386631716906947, + "grad_norm": 3.3602712154388428, + "learning_rate": 2.692005242463958e-05, + "loss": 0.4951, + "num_input_tokens_seen": 552272, + "step": 2055 + }, + { + "epoch": 0.5399737876802096, + "grad_norm": 20.621910095214844, + "learning_rate": 2.6985583224115334e-05, + "loss": 0.5563, + "num_input_tokens_seen": 553680, + "step": 2060 + }, + { + "epoch": 0.5412844036697247, + "grad_norm": 5.132722854614258, + "learning_rate": 2.7051114023591088e-05, + "loss": 0.6143, + "num_input_tokens_seen": 555520, + "step": 2065 + }, + { + "epoch": 0.5425950196592398, + "grad_norm": 9.228644371032715, + "learning_rate": 2.7116644823066845e-05, + "loss": 0.5669, + "num_input_tokens_seen": 556672, + "step": 2070 + }, + { + "epoch": 0.5439056356487549, + "grad_norm": 2.939870595932007, + "learning_rate": 2.7182175622542593e-05, + "loss": 0.6473, + "num_input_tokens_seen": 558048, + "step": 2075 + }, + { + "epoch": 0.54521625163827, + "grad_norm": 4.210785865783691, + "learning_rate": 2.724770642201835e-05, + "loss": 1.0177, + "num_input_tokens_seen": 559120, + "step": 2080 + }, + { + "epoch": 0.546526867627785, + "grad_norm": 6.547249794006348, + "learning_rate": 2.7313237221494104e-05, + "loss": 0.5534, + "num_input_tokens_seen": 560320, + "step": 2085 + }, + { + "epoch": 0.5478374836173001, + "grad_norm": 5.810853004455566, + "learning_rate": 2.7378768020969858e-05, + "loss": 0.6314, + "num_input_tokens_seen": 561456, + "step": 2090 + }, + { + "epoch": 0.5491480996068152, + "grad_norm": 2.6524384021759033, + "learning_rate": 2.744429882044561e-05, + "loss": 0.5361, + "num_input_tokens_seen": 562512, + "step": 2095 + }, + { + "epoch": 0.5504587155963303, + "grad_norm": 4.1743879318237305, + "learning_rate": 2.7509829619921362e-05, + "loss": 0.6136, + "num_input_tokens_seen": 563840, + "step": 2100 + }, + { + "epoch": 0.5517693315858454, + "grad_norm": 3.7012386322021484, + "learning_rate": 2.7575360419397116e-05, + "loss": 0.4823, + "num_input_tokens_seen": 565328, + "step": 2105 + }, + { + "epoch": 0.5530799475753604, + "grad_norm": 6.485119819641113, + "learning_rate": 2.7640891218872873e-05, + "loss": 0.5885, + "num_input_tokens_seen": 566368, + "step": 2110 + }, + { + "epoch": 0.5543905635648755, + "grad_norm": 9.610931396484375, + "learning_rate": 2.7706422018348627e-05, + "loss": 0.7452, + "num_input_tokens_seen": 567280, + "step": 2115 + }, + { + "epoch": 0.5557011795543906, + "grad_norm": 1.5580191612243652, + "learning_rate": 2.777195281782438e-05, + "loss": 0.5515, + "num_input_tokens_seen": 568864, + "step": 2120 + }, + { + "epoch": 0.5570117955439057, + "grad_norm": 5.40121603012085, + "learning_rate": 2.7837483617300132e-05, + "loss": 0.4105, + "num_input_tokens_seen": 570112, + "step": 2125 + }, + { + "epoch": 0.5583224115334207, + "grad_norm": 10.918758392333984, + "learning_rate": 2.7903014416775886e-05, + "loss": 0.7046, + "num_input_tokens_seen": 571136, + "step": 2130 + }, + { + "epoch": 0.5596330275229358, + "grad_norm": 6.841586112976074, + "learning_rate": 2.796854521625164e-05, + "loss": 0.5665, + "num_input_tokens_seen": 572256, + "step": 2135 + }, + { + "epoch": 0.5609436435124509, + "grad_norm": 7.236339569091797, + "learning_rate": 2.8034076015727394e-05, + "loss": 0.4759, + "num_input_tokens_seen": 573936, + "step": 2140 + }, + { + "epoch": 0.562254259501966, + "grad_norm": 1.1559780836105347, + "learning_rate": 2.8099606815203144e-05, + "loss": 0.6233, + "num_input_tokens_seen": 575408, + "step": 2145 + }, + { + "epoch": 0.563564875491481, + "grad_norm": 6.198041915893555, + "learning_rate": 2.8165137614678898e-05, + "loss": 0.5291, + "num_input_tokens_seen": 577184, + "step": 2150 + }, + { + "epoch": 0.564875491480996, + "grad_norm": 4.751108169555664, + "learning_rate": 2.8230668414154655e-05, + "loss": 0.4034, + "num_input_tokens_seen": 578336, + "step": 2155 + }, + { + "epoch": 0.5661861074705111, + "grad_norm": 4.266983985900879, + "learning_rate": 2.829619921363041e-05, + "loss": 0.7563, + "num_input_tokens_seen": 579424, + "step": 2160 + }, + { + "epoch": 0.5674967234600262, + "grad_norm": 8.249197006225586, + "learning_rate": 2.8361730013106163e-05, + "loss": 0.6107, + "num_input_tokens_seen": 580576, + "step": 2165 + }, + { + "epoch": 0.5688073394495413, + "grad_norm": 2.855048179626465, + "learning_rate": 2.8427260812581914e-05, + "loss": 0.5983, + "num_input_tokens_seen": 581712, + "step": 2170 + }, + { + "epoch": 0.5701179554390564, + "grad_norm": 2.169517755508423, + "learning_rate": 2.8492791612057668e-05, + "loss": 0.6383, + "num_input_tokens_seen": 583088, + "step": 2175 + }, + { + "epoch": 0.5714285714285714, + "grad_norm": 1.837321162223816, + "learning_rate": 2.855832241153342e-05, + "loss": 0.4523, + "num_input_tokens_seen": 584576, + "step": 2180 + }, + { + "epoch": 0.5727391874180865, + "grad_norm": 3.8923916816711426, + "learning_rate": 2.862385321100918e-05, + "loss": 1.1727, + "num_input_tokens_seen": 585632, + "step": 2185 + }, + { + "epoch": 0.5740498034076016, + "grad_norm": 2.942000389099121, + "learning_rate": 2.8689384010484926e-05, + "loss": 0.6389, + "num_input_tokens_seen": 587056, + "step": 2190 + }, + { + "epoch": 0.5753604193971167, + "grad_norm": 12.592362403869629, + "learning_rate": 2.875491480996068e-05, + "loss": 0.6787, + "num_input_tokens_seen": 588320, + "step": 2195 + }, + { + "epoch": 0.5766710353866317, + "grad_norm": 6.390462875366211, + "learning_rate": 2.8820445609436437e-05, + "loss": 0.4534, + "num_input_tokens_seen": 589632, + "step": 2200 + }, + { + "epoch": 0.5779816513761468, + "grad_norm": 2.7148921489715576, + "learning_rate": 2.888597640891219e-05, + "loss": 0.5122, + "num_input_tokens_seen": 591104, + "step": 2205 + }, + { + "epoch": 0.5792922673656619, + "grad_norm": 1.4179942607879639, + "learning_rate": 2.8951507208387945e-05, + "loss": 0.5432, + "num_input_tokens_seen": 593024, + "step": 2210 + }, + { + "epoch": 0.580602883355177, + "grad_norm": 4.789599895477295, + "learning_rate": 2.9017038007863695e-05, + "loss": 0.419, + "num_input_tokens_seen": 594176, + "step": 2215 + }, + { + "epoch": 0.581913499344692, + "grad_norm": 3.5971128940582275, + "learning_rate": 2.908256880733945e-05, + "loss": 0.5843, + "num_input_tokens_seen": 595632, + "step": 2220 + }, + { + "epoch": 0.583224115334207, + "grad_norm": 3.3563992977142334, + "learning_rate": 2.9148099606815203e-05, + "loss": 0.6963, + "num_input_tokens_seen": 597280, + "step": 2225 + }, + { + "epoch": 0.5845347313237221, + "grad_norm": 4.857802391052246, + "learning_rate": 2.921363040629096e-05, + "loss": 0.4867, + "num_input_tokens_seen": 598752, + "step": 2230 + }, + { + "epoch": 0.5858453473132372, + "grad_norm": 3.6485657691955566, + "learning_rate": 2.9279161205766714e-05, + "loss": 0.5217, + "num_input_tokens_seen": 600144, + "step": 2235 + }, + { + "epoch": 0.5871559633027523, + "grad_norm": 3.208124876022339, + "learning_rate": 2.9344692005242465e-05, + "loss": 0.5176, + "num_input_tokens_seen": 601440, + "step": 2240 + }, + { + "epoch": 0.5884665792922673, + "grad_norm": 5.725273609161377, + "learning_rate": 2.941022280471822e-05, + "loss": 0.8561, + "num_input_tokens_seen": 602912, + "step": 2245 + }, + { + "epoch": 0.5897771952817824, + "grad_norm": 8.104081153869629, + "learning_rate": 2.9475753604193973e-05, + "loss": 0.5123, + "num_input_tokens_seen": 604192, + "step": 2250 + }, + { + "epoch": 0.5910878112712975, + "grad_norm": 3.1850180625915527, + "learning_rate": 2.9541284403669727e-05, + "loss": 0.4434, + "num_input_tokens_seen": 605408, + "step": 2255 + }, + { + "epoch": 0.5923984272608126, + "grad_norm": 5.308319091796875, + "learning_rate": 2.9606815203145477e-05, + "loss": 0.5531, + "num_input_tokens_seen": 606832, + "step": 2260 + }, + { + "epoch": 0.5937090432503277, + "grad_norm": 3.578338146209717, + "learning_rate": 2.967234600262123e-05, + "loss": 0.5928, + "num_input_tokens_seen": 608048, + "step": 2265 + }, + { + "epoch": 0.5950196592398427, + "grad_norm": 4.867231845855713, + "learning_rate": 2.9737876802096985e-05, + "loss": 0.5004, + "num_input_tokens_seen": 609008, + "step": 2270 + }, + { + "epoch": 0.5963302752293578, + "grad_norm": 2.5214738845825195, + "learning_rate": 2.9803407601572742e-05, + "loss": 0.3699, + "num_input_tokens_seen": 610464, + "step": 2275 + }, + { + "epoch": 0.5976408912188729, + "grad_norm": 4.2405219078063965, + "learning_rate": 2.9868938401048496e-05, + "loss": 0.4457, + "num_input_tokens_seen": 611568, + "step": 2280 + }, + { + "epoch": 0.598951507208388, + "grad_norm": 5.2951579093933105, + "learning_rate": 2.9934469200524247e-05, + "loss": 0.5964, + "num_input_tokens_seen": 612800, + "step": 2285 + }, + { + "epoch": 0.6002621231979031, + "grad_norm": 4.79838228225708, + "learning_rate": 3e-05, + "loss": 0.6153, + "num_input_tokens_seen": 614064, + "step": 2290 + }, + { + "epoch": 0.601572739187418, + "grad_norm": 5.39982795715332, + "learning_rate": 3.0065530799475755e-05, + "loss": 0.4756, + "num_input_tokens_seen": 615184, + "step": 2295 + }, + { + "epoch": 0.6028833551769331, + "grad_norm": 6.33749532699585, + "learning_rate": 3.013106159895151e-05, + "loss": 0.6931, + "num_input_tokens_seen": 616736, + "step": 2300 + }, + { + "epoch": 0.6041939711664482, + "grad_norm": 5.078553199768066, + "learning_rate": 3.0196592398427266e-05, + "loss": 0.4462, + "num_input_tokens_seen": 618064, + "step": 2305 + }, + { + "epoch": 0.6055045871559633, + "grad_norm": 1.9832507371902466, + "learning_rate": 3.0262123197903013e-05, + "loss": 0.5156, + "num_input_tokens_seen": 620448, + "step": 2310 + }, + { + "epoch": 0.6068152031454783, + "grad_norm": 5.675999641418457, + "learning_rate": 3.032765399737877e-05, + "loss": 0.4839, + "num_input_tokens_seen": 621728, + "step": 2315 + }, + { + "epoch": 0.6081258191349934, + "grad_norm": 2.7732155323028564, + "learning_rate": 3.0393184796854524e-05, + "loss": 0.5063, + "num_input_tokens_seen": 622864, + "step": 2320 + }, + { + "epoch": 0.6094364351245085, + "grad_norm": 3.8743813037872314, + "learning_rate": 3.0458715596330278e-05, + "loss": 0.5463, + "num_input_tokens_seen": 624352, + "step": 2325 + }, + { + "epoch": 0.6107470511140236, + "grad_norm": 9.679034233093262, + "learning_rate": 3.052424639580603e-05, + "loss": 0.5778, + "num_input_tokens_seen": 625568, + "step": 2330 + }, + { + "epoch": 0.6120576671035387, + "grad_norm": 3.6903884410858154, + "learning_rate": 3.058977719528178e-05, + "loss": 0.4138, + "num_input_tokens_seen": 627056, + "step": 2335 + }, + { + "epoch": 0.6133682830930537, + "grad_norm": 2.8224403858184814, + "learning_rate": 3.0655307994757537e-05, + "loss": 0.4716, + "num_input_tokens_seen": 628112, + "step": 2340 + }, + { + "epoch": 0.6146788990825688, + "grad_norm": 2.032667875289917, + "learning_rate": 3.0720838794233294e-05, + "loss": 0.4329, + "num_input_tokens_seen": 629424, + "step": 2345 + }, + { + "epoch": 0.6159895150720839, + "grad_norm": 2.532590627670288, + "learning_rate": 3.0786369593709044e-05, + "loss": 0.3475, + "num_input_tokens_seen": 630896, + "step": 2350 + }, + { + "epoch": 0.617300131061599, + "grad_norm": 2.6434149742126465, + "learning_rate": 3.0851900393184795e-05, + "loss": 0.591, + "num_input_tokens_seen": 632224, + "step": 2355 + }, + { + "epoch": 0.6186107470511141, + "grad_norm": 8.282522201538086, + "learning_rate": 3.091743119266055e-05, + "loss": 0.7656, + "num_input_tokens_seen": 633600, + "step": 2360 + }, + { + "epoch": 0.6199213630406291, + "grad_norm": 6.785233974456787, + "learning_rate": 3.09829619921363e-05, + "loss": 0.7432, + "num_input_tokens_seen": 634880, + "step": 2365 + }, + { + "epoch": 0.6212319790301442, + "grad_norm": 3.012680768966675, + "learning_rate": 3.104849279161206e-05, + "loss": 0.4372, + "num_input_tokens_seen": 636160, + "step": 2370 + }, + { + "epoch": 0.6225425950196593, + "grad_norm": 2.7701728343963623, + "learning_rate": 3.111402359108782e-05, + "loss": 0.4077, + "num_input_tokens_seen": 637440, + "step": 2375 + }, + { + "epoch": 0.6238532110091743, + "grad_norm": 7.189513683319092, + "learning_rate": 3.117955439056357e-05, + "loss": 0.4028, + "num_input_tokens_seen": 638672, + "step": 2380 + }, + { + "epoch": 0.6251638269986893, + "grad_norm": 3.0784056186676025, + "learning_rate": 3.124508519003932e-05, + "loss": 0.4652, + "num_input_tokens_seen": 639744, + "step": 2385 + }, + { + "epoch": 0.6264744429882044, + "grad_norm": 2.8524630069732666, + "learning_rate": 3.1310615989515076e-05, + "loss": 0.5974, + "num_input_tokens_seen": 641280, + "step": 2390 + }, + { + "epoch": 0.6277850589777195, + "grad_norm": 1.4591037034988403, + "learning_rate": 3.1376146788990826e-05, + "loss": 0.6764, + "num_input_tokens_seen": 642576, + "step": 2395 + }, + { + "epoch": 0.6290956749672346, + "grad_norm": 4.094770908355713, + "learning_rate": 3.144167758846658e-05, + "loss": 0.4613, + "num_input_tokens_seen": 643952, + "step": 2400 + }, + { + "epoch": 0.6304062909567497, + "grad_norm": 10.565229415893555, + "learning_rate": 3.1507208387942334e-05, + "loss": 0.7141, + "num_input_tokens_seen": 644992, + "step": 2405 + }, + { + "epoch": 0.6317169069462647, + "grad_norm": 3.0611679553985596, + "learning_rate": 3.1572739187418085e-05, + "loss": 0.4243, + "num_input_tokens_seen": 646080, + "step": 2410 + }, + { + "epoch": 0.6330275229357798, + "grad_norm": 4.857278347015381, + "learning_rate": 3.163826998689384e-05, + "loss": 0.6387, + "num_input_tokens_seen": 647344, + "step": 2415 + }, + { + "epoch": 0.6343381389252949, + "grad_norm": 4.1823649406433105, + "learning_rate": 3.17038007863696e-05, + "loss": 0.4768, + "num_input_tokens_seen": 648704, + "step": 2420 + }, + { + "epoch": 0.63564875491481, + "grad_norm": 2.181061029434204, + "learning_rate": 3.176933158584535e-05, + "loss": 0.5354, + "num_input_tokens_seen": 650000, + "step": 2425 + }, + { + "epoch": 0.6369593709043251, + "grad_norm": 2.736772060394287, + "learning_rate": 3.18348623853211e-05, + "loss": 0.5615, + "num_input_tokens_seen": 651840, + "step": 2430 + }, + { + "epoch": 0.6382699868938401, + "grad_norm": 6.12676477432251, + "learning_rate": 3.190039318479686e-05, + "loss": 0.5815, + "num_input_tokens_seen": 652784, + "step": 2435 + }, + { + "epoch": 0.6395806028833552, + "grad_norm": 6.800152778625488, + "learning_rate": 3.196592398427261e-05, + "loss": 0.7154, + "num_input_tokens_seen": 654176, + "step": 2440 + }, + { + "epoch": 0.6408912188728703, + "grad_norm": 4.328539848327637, + "learning_rate": 3.2031454783748365e-05, + "loss": 0.6286, + "num_input_tokens_seen": 655200, + "step": 2445 + }, + { + "epoch": 0.6422018348623854, + "grad_norm": 3.9988889694213867, + "learning_rate": 3.2096985583224116e-05, + "loss": 0.3235, + "num_input_tokens_seen": 656592, + "step": 2450 + }, + { + "epoch": 0.6435124508519003, + "grad_norm": 3.25407075881958, + "learning_rate": 3.216251638269987e-05, + "loss": 0.4398, + "num_input_tokens_seen": 657648, + "step": 2455 + }, + { + "epoch": 0.6448230668414154, + "grad_norm": 4.693026542663574, + "learning_rate": 3.2228047182175624e-05, + "loss": 0.3289, + "num_input_tokens_seen": 659136, + "step": 2460 + }, + { + "epoch": 0.6461336828309305, + "grad_norm": 2.111581325531006, + "learning_rate": 3.229357798165138e-05, + "loss": 0.5287, + "num_input_tokens_seen": 660624, + "step": 2465 + }, + { + "epoch": 0.6474442988204456, + "grad_norm": 3.2723419666290283, + "learning_rate": 3.235910878112713e-05, + "loss": 0.6776, + "num_input_tokens_seen": 661488, + "step": 2470 + }, + { + "epoch": 0.6487549148099607, + "grad_norm": 4.1780524253845215, + "learning_rate": 3.242463958060288e-05, + "loss": 0.4366, + "num_input_tokens_seen": 662688, + "step": 2475 + }, + { + "epoch": 0.6500655307994757, + "grad_norm": 2.6736724376678467, + "learning_rate": 3.249017038007864e-05, + "loss": 0.2688, + "num_input_tokens_seen": 663904, + "step": 2480 + }, + { + "epoch": 0.6513761467889908, + "grad_norm": 1.2296706438064575, + "learning_rate": 3.255570117955439e-05, + "loss": 0.5484, + "num_input_tokens_seen": 665360, + "step": 2485 + }, + { + "epoch": 0.6526867627785059, + "grad_norm": 3.8515188694000244, + "learning_rate": 3.262123197903015e-05, + "loss": 0.4574, + "num_input_tokens_seen": 666832, + "step": 2490 + }, + { + "epoch": 0.653997378768021, + "grad_norm": 6.453360557556152, + "learning_rate": 3.26867627785059e-05, + "loss": 0.6264, + "num_input_tokens_seen": 668112, + "step": 2495 + }, + { + "epoch": 0.6553079947575361, + "grad_norm": 8.789599418640137, + "learning_rate": 3.2752293577981655e-05, + "loss": 0.6931, + "num_input_tokens_seen": 669024, + "step": 2500 + }, + { + "epoch": 0.6566186107470511, + "grad_norm": 5.666074752807617, + "learning_rate": 3.2817824377457405e-05, + "loss": 0.4246, + "num_input_tokens_seen": 670192, + "step": 2505 + }, + { + "epoch": 0.6579292267365662, + "grad_norm": 3.6754987239837646, + "learning_rate": 3.288335517693316e-05, + "loss": 0.4032, + "num_input_tokens_seen": 671440, + "step": 2510 + }, + { + "epoch": 0.6592398427260813, + "grad_norm": 11.473990440368652, + "learning_rate": 3.294888597640891e-05, + "loss": 0.5367, + "num_input_tokens_seen": 672800, + "step": 2515 + }, + { + "epoch": 0.6605504587155964, + "grad_norm": 2.154189109802246, + "learning_rate": 3.3014416775884664e-05, + "loss": 0.5393, + "num_input_tokens_seen": 674016, + "step": 2520 + }, + { + "epoch": 0.6618610747051114, + "grad_norm": 6.014385223388672, + "learning_rate": 3.307994757536042e-05, + "loss": 0.3386, + "num_input_tokens_seen": 675104, + "step": 2525 + }, + { + "epoch": 0.6631716906946264, + "grad_norm": 2.722870349884033, + "learning_rate": 3.314547837483618e-05, + "loss": 0.3041, + "num_input_tokens_seen": 676288, + "step": 2530 + }, + { + "epoch": 0.6644823066841415, + "grad_norm": 14.239012718200684, + "learning_rate": 3.321100917431193e-05, + "loss": 0.5851, + "num_input_tokens_seen": 677424, + "step": 2535 + }, + { + "epoch": 0.6657929226736566, + "grad_norm": 4.004636287689209, + "learning_rate": 3.327653997378768e-05, + "loss": 0.5266, + "num_input_tokens_seen": 678400, + "step": 2540 + }, + { + "epoch": 0.6671035386631717, + "grad_norm": 11.515826225280762, + "learning_rate": 3.334207077326344e-05, + "loss": 0.4454, + "num_input_tokens_seen": 679392, + "step": 2545 + }, + { + "epoch": 0.6684141546526867, + "grad_norm": 4.839748859405518, + "learning_rate": 3.340760157273919e-05, + "loss": 0.6928, + "num_input_tokens_seen": 680544, + "step": 2550 + }, + { + "epoch": 0.6697247706422018, + "grad_norm": 4.508795738220215, + "learning_rate": 3.3473132372214945e-05, + "loss": 0.5211, + "num_input_tokens_seen": 682144, + "step": 2555 + }, + { + "epoch": 0.6710353866317169, + "grad_norm": 15.76900577545166, + "learning_rate": 3.3538663171690695e-05, + "loss": 0.4391, + "num_input_tokens_seen": 683344, + "step": 2560 + }, + { + "epoch": 0.672346002621232, + "grad_norm": 2.2405524253845215, + "learning_rate": 3.3604193971166446e-05, + "loss": 0.6359, + "num_input_tokens_seen": 684704, + "step": 2565 + }, + { + "epoch": 0.6736566186107471, + "grad_norm": 2.8745644092559814, + "learning_rate": 3.36697247706422e-05, + "loss": 0.5334, + "num_input_tokens_seen": 685776, + "step": 2570 + }, + { + "epoch": 0.6749672346002621, + "grad_norm": 2.5968010425567627, + "learning_rate": 3.373525557011796e-05, + "loss": 0.596, + "num_input_tokens_seen": 687584, + "step": 2575 + }, + { + "epoch": 0.6762778505897772, + "grad_norm": 3.3119561672210693, + "learning_rate": 3.380078636959371e-05, + "loss": 0.4729, + "num_input_tokens_seen": 688864, + "step": 2580 + }, + { + "epoch": 0.6775884665792923, + "grad_norm": 2.3446881771087646, + "learning_rate": 3.386631716906947e-05, + "loss": 0.6219, + "num_input_tokens_seen": 690512, + "step": 2585 + }, + { + "epoch": 0.6788990825688074, + "grad_norm": 3.6298539638519287, + "learning_rate": 3.393184796854522e-05, + "loss": 0.6831, + "num_input_tokens_seen": 691728, + "step": 2590 + }, + { + "epoch": 0.6802096985583224, + "grad_norm": 2.6999757289886475, + "learning_rate": 3.399737876802097e-05, + "loss": 0.5213, + "num_input_tokens_seen": 693280, + "step": 2595 + }, + { + "epoch": 0.6815203145478375, + "grad_norm": 4.894536972045898, + "learning_rate": 3.4062909567496726e-05, + "loss": 0.3026, + "num_input_tokens_seen": 694656, + "step": 2600 + }, + { + "epoch": 0.6828309305373526, + "grad_norm": 5.71895694732666, + "learning_rate": 3.4128440366972484e-05, + "loss": 0.4885, + "num_input_tokens_seen": 695712, + "step": 2605 + }, + { + "epoch": 0.6841415465268676, + "grad_norm": 3.2583277225494385, + "learning_rate": 3.419397116644823e-05, + "loss": 0.4651, + "num_input_tokens_seen": 697072, + "step": 2610 + }, + { + "epoch": 0.6854521625163827, + "grad_norm": 3.8966116905212402, + "learning_rate": 3.4259501965923985e-05, + "loss": 0.5264, + "num_input_tokens_seen": 698304, + "step": 2615 + }, + { + "epoch": 0.6867627785058977, + "grad_norm": 4.769927024841309, + "learning_rate": 3.432503276539974e-05, + "loss": 0.6624, + "num_input_tokens_seen": 699696, + "step": 2620 + }, + { + "epoch": 0.6880733944954128, + "grad_norm": 4.072166919708252, + "learning_rate": 3.439056356487549e-05, + "loss": 0.562, + "num_input_tokens_seen": 701056, + "step": 2625 + }, + { + "epoch": 0.6893840104849279, + "grad_norm": 8.221331596374512, + "learning_rate": 3.445609436435125e-05, + "loss": 0.5181, + "num_input_tokens_seen": 702080, + "step": 2630 + }, + { + "epoch": 0.690694626474443, + "grad_norm": 4.570629596710205, + "learning_rate": 3.4521625163827e-05, + "loss": 0.4576, + "num_input_tokens_seen": 703424, + "step": 2635 + }, + { + "epoch": 0.6920052424639581, + "grad_norm": 6.180621147155762, + "learning_rate": 3.458715596330275e-05, + "loss": 0.5545, + "num_input_tokens_seen": 704816, + "step": 2640 + }, + { + "epoch": 0.6933158584534731, + "grad_norm": 6.305951118469238, + "learning_rate": 3.465268676277851e-05, + "loss": 0.4266, + "num_input_tokens_seen": 706096, + "step": 2645 + }, + { + "epoch": 0.6946264744429882, + "grad_norm": 5.416182994842529, + "learning_rate": 3.4718217562254266e-05, + "loss": 0.4522, + "num_input_tokens_seen": 707104, + "step": 2650 + }, + { + "epoch": 0.6959370904325033, + "grad_norm": 2.4217171669006348, + "learning_rate": 3.4783748361730016e-05, + "loss": 0.6009, + "num_input_tokens_seen": 708224, + "step": 2655 + }, + { + "epoch": 0.6972477064220184, + "grad_norm": 3.1519691944122314, + "learning_rate": 3.484927916120577e-05, + "loss": 0.5975, + "num_input_tokens_seen": 709872, + "step": 2660 + }, + { + "epoch": 0.6985583224115334, + "grad_norm": 4.084074974060059, + "learning_rate": 3.4914809960681524e-05, + "loss": 0.6004, + "num_input_tokens_seen": 710976, + "step": 2665 + }, + { + "epoch": 0.6998689384010485, + "grad_norm": 3.2832725048065186, + "learning_rate": 3.4980340760157274e-05, + "loss": 0.6, + "num_input_tokens_seen": 712800, + "step": 2670 + }, + { + "epoch": 0.7011795543905636, + "grad_norm": 4.888327598571777, + "learning_rate": 3.504587155963303e-05, + "loss": 0.3777, + "num_input_tokens_seen": 714032, + "step": 2675 + }, + { + "epoch": 0.7024901703800787, + "grad_norm": 4.724072456359863, + "learning_rate": 3.511140235910878e-05, + "loss": 0.4711, + "num_input_tokens_seen": 715440, + "step": 2680 + }, + { + "epoch": 0.7038007863695938, + "grad_norm": 21.416221618652344, + "learning_rate": 3.517693315858453e-05, + "loss": 0.6358, + "num_input_tokens_seen": 716336, + "step": 2685 + }, + { + "epoch": 0.7051114023591087, + "grad_norm": 10.987424850463867, + "learning_rate": 3.524246395806029e-05, + "loss": 0.4379, + "num_input_tokens_seen": 718304, + "step": 2690 + }, + { + "epoch": 0.7064220183486238, + "grad_norm": 4.579024791717529, + "learning_rate": 3.530799475753605e-05, + "loss": 0.4041, + "num_input_tokens_seen": 719600, + "step": 2695 + }, + { + "epoch": 0.7077326343381389, + "grad_norm": 3.363926887512207, + "learning_rate": 3.53735255570118e-05, + "loss": 0.5366, + "num_input_tokens_seen": 721152, + "step": 2700 + }, + { + "epoch": 0.709043250327654, + "grad_norm": 4.475964546203613, + "learning_rate": 3.543905635648755e-05, + "loss": 0.4841, + "num_input_tokens_seen": 722560, + "step": 2705 + }, + { + "epoch": 0.7103538663171691, + "grad_norm": 1.562048077583313, + "learning_rate": 3.5504587155963306e-05, + "loss": 0.4993, + "num_input_tokens_seen": 724464, + "step": 2710 + }, + { + "epoch": 0.7116644823066841, + "grad_norm": 12.22320556640625, + "learning_rate": 3.5570117955439056e-05, + "loss": 0.5523, + "num_input_tokens_seen": 725648, + "step": 2715 + }, + { + "epoch": 0.7129750982961992, + "grad_norm": 4.659371376037598, + "learning_rate": 3.5635648754914814e-05, + "loss": 0.4788, + "num_input_tokens_seen": 726992, + "step": 2720 + }, + { + "epoch": 0.7142857142857143, + "grad_norm": 14.853114128112793, + "learning_rate": 3.570117955439057e-05, + "loss": 0.5979, + "num_input_tokens_seen": 728112, + "step": 2725 + }, + { + "epoch": 0.7155963302752294, + "grad_norm": 3.3155200481414795, + "learning_rate": 3.5766710353866315e-05, + "loss": 0.5107, + "num_input_tokens_seen": 729296, + "step": 2730 + }, + { + "epoch": 0.7169069462647444, + "grad_norm": 3.220916509628296, + "learning_rate": 3.583224115334207e-05, + "loss": 0.5203, + "num_input_tokens_seen": 730864, + "step": 2735 + }, + { + "epoch": 0.7182175622542595, + "grad_norm": 5.35984992980957, + "learning_rate": 3.589777195281783e-05, + "loss": 0.6524, + "num_input_tokens_seen": 731904, + "step": 2740 + }, + { + "epoch": 0.7195281782437746, + "grad_norm": 20.413835525512695, + "learning_rate": 3.596330275229358e-05, + "loss": 0.6801, + "num_input_tokens_seen": 733968, + "step": 2745 + }, + { + "epoch": 0.7208387942332897, + "grad_norm": 4.702981948852539, + "learning_rate": 3.602883355176933e-05, + "loss": 0.3773, + "num_input_tokens_seen": 735344, + "step": 2750 + }, + { + "epoch": 0.7221494102228048, + "grad_norm": 2.463870048522949, + "learning_rate": 3.609436435124509e-05, + "loss": 0.4158, + "num_input_tokens_seen": 736608, + "step": 2755 + }, + { + "epoch": 0.7234600262123198, + "grad_norm": 1.690972089767456, + "learning_rate": 3.615989515072084e-05, + "loss": 0.5182, + "num_input_tokens_seen": 737984, + "step": 2760 + }, + { + "epoch": 0.7247706422018348, + "grad_norm": 1.889755129814148, + "learning_rate": 3.6225425950196595e-05, + "loss": 0.5002, + "num_input_tokens_seen": 739296, + "step": 2765 + }, + { + "epoch": 0.7260812581913499, + "grad_norm": 4.19559383392334, + "learning_rate": 3.629095674967235e-05, + "loss": 0.5661, + "num_input_tokens_seen": 740176, + "step": 2770 + }, + { + "epoch": 0.727391874180865, + "grad_norm": 6.935085296630859, + "learning_rate": 3.6356487549148097e-05, + "loss": 0.8883, + "num_input_tokens_seen": 741648, + "step": 2775 + }, + { + "epoch": 0.72870249017038, + "grad_norm": 5.01033878326416, + "learning_rate": 3.6422018348623854e-05, + "loss": 0.4078, + "num_input_tokens_seen": 742928, + "step": 2780 + }, + { + "epoch": 0.7300131061598951, + "grad_norm": 3.3818366527557373, + "learning_rate": 3.648754914809961e-05, + "loss": 0.4928, + "num_input_tokens_seen": 744384, + "step": 2785 + }, + { + "epoch": 0.7313237221494102, + "grad_norm": 8.705668449401855, + "learning_rate": 3.655307994757536e-05, + "loss": 0.6054, + "num_input_tokens_seen": 745792, + "step": 2790 + }, + { + "epoch": 0.7326343381389253, + "grad_norm": 11.149552345275879, + "learning_rate": 3.661861074705111e-05, + "loss": 0.5264, + "num_input_tokens_seen": 747216, + "step": 2795 + }, + { + "epoch": 0.7339449541284404, + "grad_norm": 2.540930986404419, + "learning_rate": 3.668414154652687e-05, + "loss": 0.5144, + "num_input_tokens_seen": 749008, + "step": 2800 + }, + { + "epoch": 0.7352555701179554, + "grad_norm": 10.497570991516113, + "learning_rate": 3.674967234600262e-05, + "loss": 0.5528, + "num_input_tokens_seen": 750160, + "step": 2805 + }, + { + "epoch": 0.7365661861074705, + "grad_norm": 3.432769536972046, + "learning_rate": 3.681520314547838e-05, + "loss": 0.5259, + "num_input_tokens_seen": 750976, + "step": 2810 + }, + { + "epoch": 0.7378768020969856, + "grad_norm": 2.5150399208068848, + "learning_rate": 3.6880733944954135e-05, + "loss": 0.4245, + "num_input_tokens_seen": 752448, + "step": 2815 + }, + { + "epoch": 0.7391874180865007, + "grad_norm": 4.100630760192871, + "learning_rate": 3.694626474442988e-05, + "loss": 0.5285, + "num_input_tokens_seen": 754224, + "step": 2820 + }, + { + "epoch": 0.7404980340760158, + "grad_norm": 5.031260013580322, + "learning_rate": 3.7011795543905636e-05, + "loss": 0.5951, + "num_input_tokens_seen": 755296, + "step": 2825 + }, + { + "epoch": 0.7418086500655308, + "grad_norm": 1.733210563659668, + "learning_rate": 3.707732634338139e-05, + "loss": 0.4073, + "num_input_tokens_seen": 756960, + "step": 2830 + }, + { + "epoch": 0.7431192660550459, + "grad_norm": 4.547720909118652, + "learning_rate": 3.7142857142857143e-05, + "loss": 0.4319, + "num_input_tokens_seen": 758080, + "step": 2835 + }, + { + "epoch": 0.744429882044561, + "grad_norm": 2.9076178073883057, + "learning_rate": 3.72083879423329e-05, + "loss": 0.7278, + "num_input_tokens_seen": 759504, + "step": 2840 + }, + { + "epoch": 0.745740498034076, + "grad_norm": 4.386256694793701, + "learning_rate": 3.727391874180865e-05, + "loss": 0.6407, + "num_input_tokens_seen": 760608, + "step": 2845 + }, + { + "epoch": 0.747051114023591, + "grad_norm": 7.048031806945801, + "learning_rate": 3.73394495412844e-05, + "loss": 0.5664, + "num_input_tokens_seen": 761936, + "step": 2850 + }, + { + "epoch": 0.7483617300131061, + "grad_norm": 3.8189163208007812, + "learning_rate": 3.740498034076016e-05, + "loss": 0.4755, + "num_input_tokens_seen": 763104, + "step": 2855 + }, + { + "epoch": 0.7496723460026212, + "grad_norm": 3.1391735076904297, + "learning_rate": 3.7470511140235916e-05, + "loss": 0.3848, + "num_input_tokens_seen": 764096, + "step": 2860 + }, + { + "epoch": 0.7509829619921363, + "grad_norm": 5.55087947845459, + "learning_rate": 3.753604193971167e-05, + "loss": 0.3643, + "num_input_tokens_seen": 765232, + "step": 2865 + }, + { + "epoch": 0.7522935779816514, + "grad_norm": 2.0338683128356934, + "learning_rate": 3.760157273918742e-05, + "loss": 0.4824, + "num_input_tokens_seen": 766512, + "step": 2870 + }, + { + "epoch": 0.7536041939711664, + "grad_norm": 2.1641390323638916, + "learning_rate": 3.7667103538663175e-05, + "loss": 0.5063, + "num_input_tokens_seen": 767936, + "step": 2875 + }, + { + "epoch": 0.7549148099606815, + "grad_norm": 4.453091621398926, + "learning_rate": 3.7732634338138925e-05, + "loss": 0.5953, + "num_input_tokens_seen": 769632, + "step": 2880 + }, + { + "epoch": 0.7562254259501966, + "grad_norm": 4.251415252685547, + "learning_rate": 3.779816513761468e-05, + "loss": 0.4679, + "num_input_tokens_seen": 770896, + "step": 2885 + }, + { + "epoch": 0.7575360419397117, + "grad_norm": 3.5021400451660156, + "learning_rate": 3.786369593709043e-05, + "loss": 0.4517, + "num_input_tokens_seen": 772160, + "step": 2890 + }, + { + "epoch": 0.7588466579292268, + "grad_norm": 3.195491075515747, + "learning_rate": 3.7929226736566184e-05, + "loss": 0.4911, + "num_input_tokens_seen": 773408, + "step": 2895 + }, + { + "epoch": 0.7601572739187418, + "grad_norm": 4.404767990112305, + "learning_rate": 3.799475753604194e-05, + "loss": 0.5726, + "num_input_tokens_seen": 774400, + "step": 2900 + }, + { + "epoch": 0.7614678899082569, + "grad_norm": 2.9966256618499756, + "learning_rate": 3.80602883355177e-05, + "loss": 0.5671, + "num_input_tokens_seen": 775632, + "step": 2905 + }, + { + "epoch": 0.762778505897772, + "grad_norm": 6.604687213897705, + "learning_rate": 3.812581913499345e-05, + "loss": 0.5596, + "num_input_tokens_seen": 776960, + "step": 2910 + }, + { + "epoch": 0.7640891218872871, + "grad_norm": 5.6684064865112305, + "learning_rate": 3.81913499344692e-05, + "loss": 0.3948, + "num_input_tokens_seen": 778144, + "step": 2915 + }, + { + "epoch": 0.765399737876802, + "grad_norm": 9.726417541503906, + "learning_rate": 3.825688073394496e-05, + "loss": 0.5081, + "num_input_tokens_seen": 779584, + "step": 2920 + }, + { + "epoch": 0.7667103538663171, + "grad_norm": 7.665738582611084, + "learning_rate": 3.832241153342071e-05, + "loss": 0.4181, + "num_input_tokens_seen": 780752, + "step": 2925 + }, + { + "epoch": 0.7680209698558322, + "grad_norm": 3.7370412349700928, + "learning_rate": 3.8387942332896464e-05, + "loss": 0.6374, + "num_input_tokens_seen": 782112, + "step": 2930 + }, + { + "epoch": 0.7693315858453473, + "grad_norm": 24.001323699951172, + "learning_rate": 3.8453473132372215e-05, + "loss": 0.687, + "num_input_tokens_seen": 783024, + "step": 2935 + }, + { + "epoch": 0.7706422018348624, + "grad_norm": 1.9928960800170898, + "learning_rate": 3.851900393184797e-05, + "loss": 0.5013, + "num_input_tokens_seen": 784336, + "step": 2940 + }, + { + "epoch": 0.7719528178243774, + "grad_norm": 3.0472841262817383, + "learning_rate": 3.858453473132372e-05, + "loss": 0.4966, + "num_input_tokens_seen": 785616, + "step": 2945 + }, + { + "epoch": 0.7732634338138925, + "grad_norm": 3.2063612937927246, + "learning_rate": 3.865006553079948e-05, + "loss": 0.7092, + "num_input_tokens_seen": 786880, + "step": 2950 + }, + { + "epoch": 0.7745740498034076, + "grad_norm": 3.9423670768737793, + "learning_rate": 3.871559633027523e-05, + "loss": 0.4312, + "num_input_tokens_seen": 787968, + "step": 2955 + }, + { + "epoch": 0.7758846657929227, + "grad_norm": 10.579744338989258, + "learning_rate": 3.878112712975098e-05, + "loss": 0.6219, + "num_input_tokens_seen": 789040, + "step": 2960 + }, + { + "epoch": 0.7771952817824378, + "grad_norm": 4.243243217468262, + "learning_rate": 3.884665792922674e-05, + "loss": 0.596, + "num_input_tokens_seen": 790256, + "step": 2965 + }, + { + "epoch": 0.7785058977719528, + "grad_norm": 7.145074367523193, + "learning_rate": 3.891218872870249e-05, + "loss": 0.7812, + "num_input_tokens_seen": 791520, + "step": 2970 + }, + { + "epoch": 0.7798165137614679, + "grad_norm": 7.594117164611816, + "learning_rate": 3.8977719528178246e-05, + "loss": 0.8442, + "num_input_tokens_seen": 792912, + "step": 2975 + }, + { + "epoch": 0.781127129750983, + "grad_norm": 1.4554498195648193, + "learning_rate": 3.9043250327654004e-05, + "loss": 0.566, + "num_input_tokens_seen": 794848, + "step": 2980 + }, + { + "epoch": 0.7824377457404981, + "grad_norm": 5.736520290374756, + "learning_rate": 3.9108781127129754e-05, + "loss": 0.3373, + "num_input_tokens_seen": 796256, + "step": 2985 + }, + { + "epoch": 0.783748361730013, + "grad_norm": 4.675373077392578, + "learning_rate": 3.9174311926605505e-05, + "loss": 0.6252, + "num_input_tokens_seen": 797792, + "step": 2990 + }, + { + "epoch": 0.7850589777195282, + "grad_norm": 4.164758205413818, + "learning_rate": 3.923984272608126e-05, + "loss": 0.5754, + "num_input_tokens_seen": 798880, + "step": 2995 + }, + { + "epoch": 0.7863695937090432, + "grad_norm": 4.539575099945068, + "learning_rate": 3.930537352555701e-05, + "loss": 0.5707, + "num_input_tokens_seen": 799856, + "step": 3000 + }, + { + "epoch": 0.7876802096985583, + "grad_norm": 3.4115664958953857, + "learning_rate": 3.937090432503276e-05, + "loss": 0.4709, + "num_input_tokens_seen": 800992, + "step": 3005 + }, + { + "epoch": 0.7889908256880734, + "grad_norm": 3.507639169692993, + "learning_rate": 3.943643512450852e-05, + "loss": 0.4717, + "num_input_tokens_seen": 802432, + "step": 3010 + }, + { + "epoch": 0.7903014416775884, + "grad_norm": 3.301596164703369, + "learning_rate": 3.950196592398428e-05, + "loss": 0.5277, + "num_input_tokens_seen": 803504, + "step": 3015 + }, + { + "epoch": 0.7916120576671035, + "grad_norm": 4.0193986892700195, + "learning_rate": 3.956749672346003e-05, + "loss": 0.5372, + "num_input_tokens_seen": 804752, + "step": 3020 + }, + { + "epoch": 0.7929226736566186, + "grad_norm": 3.5412142276763916, + "learning_rate": 3.9633027522935785e-05, + "loss": 0.5986, + "num_input_tokens_seen": 805872, + "step": 3025 + }, + { + "epoch": 0.7942332896461337, + "grad_norm": 2.3779296875, + "learning_rate": 3.9698558322411536e-05, + "loss": 0.4872, + "num_input_tokens_seen": 807472, + "step": 3030 + }, + { + "epoch": 0.7955439056356488, + "grad_norm": 3.9345145225524902, + "learning_rate": 3.9764089121887286e-05, + "loss": 0.6172, + "num_input_tokens_seen": 809056, + "step": 3035 + }, + { + "epoch": 0.7968545216251638, + "grad_norm": 6.209622383117676, + "learning_rate": 3.9829619921363044e-05, + "loss": 0.4259, + "num_input_tokens_seen": 810400, + "step": 3040 + }, + { + "epoch": 0.7981651376146789, + "grad_norm": 2.187204599380493, + "learning_rate": 3.9895150720838794e-05, + "loss": 0.5499, + "num_input_tokens_seen": 811536, + "step": 3045 + }, + { + "epoch": 0.799475753604194, + "grad_norm": 2.2584378719329834, + "learning_rate": 3.996068152031455e-05, + "loss": 0.5087, + "num_input_tokens_seen": 813040, + "step": 3050 + }, + { + "epoch": 0.8007863695937091, + "grad_norm": 4.479312419891357, + "learning_rate": 4.00262123197903e-05, + "loss": 0.5597, + "num_input_tokens_seen": 814112, + "step": 3055 + }, + { + "epoch": 0.8020969855832241, + "grad_norm": 4.370864391326904, + "learning_rate": 4.009174311926606e-05, + "loss": 0.4419, + "num_input_tokens_seen": 816352, + "step": 3060 + }, + { + "epoch": 0.8034076015727392, + "grad_norm": 1.8822239637374878, + "learning_rate": 4.015727391874181e-05, + "loss": 0.4964, + "num_input_tokens_seen": 817648, + "step": 3065 + }, + { + "epoch": 0.8047182175622543, + "grad_norm": 2.737382411956787, + "learning_rate": 4.022280471821757e-05, + "loss": 0.4044, + "num_input_tokens_seen": 818848, + "step": 3070 + }, + { + "epoch": 0.8060288335517694, + "grad_norm": 6.540686130523682, + "learning_rate": 4.028833551769332e-05, + "loss": 0.6004, + "num_input_tokens_seen": 820304, + "step": 3075 + }, + { + "epoch": 0.8073394495412844, + "grad_norm": 4.031937599182129, + "learning_rate": 4.035386631716907e-05, + "loss": 0.4792, + "num_input_tokens_seen": 821344, + "step": 3080 + }, + { + "epoch": 0.8086500655307994, + "grad_norm": 3.082925319671631, + "learning_rate": 4.0419397116644826e-05, + "loss": 0.5953, + "num_input_tokens_seen": 822912, + "step": 3085 + }, + { + "epoch": 0.8099606815203145, + "grad_norm": 3.1688013076782227, + "learning_rate": 4.048492791612058e-05, + "loss": 0.5598, + "num_input_tokens_seen": 824144, + "step": 3090 + }, + { + "epoch": 0.8112712975098296, + "grad_norm": 4.848661422729492, + "learning_rate": 4.0550458715596333e-05, + "loss": 0.4778, + "num_input_tokens_seen": 825152, + "step": 3095 + }, + { + "epoch": 0.8125819134993447, + "grad_norm": 3.8179433345794678, + "learning_rate": 4.0615989515072084e-05, + "loss": 0.7468, + "num_input_tokens_seen": 826272, + "step": 3100 + }, + { + "epoch": 0.8138925294888598, + "grad_norm": 1.8588799238204956, + "learning_rate": 4.068152031454784e-05, + "loss": 0.7784, + "num_input_tokens_seen": 827680, + "step": 3105 + }, + { + "epoch": 0.8152031454783748, + "grad_norm": 7.13505744934082, + "learning_rate": 4.074705111402359e-05, + "loss": 0.5276, + "num_input_tokens_seen": 828800, + "step": 3110 + }, + { + "epoch": 0.8165137614678899, + "grad_norm": 11.689985275268555, + "learning_rate": 4.081258191349935e-05, + "loss": 0.6601, + "num_input_tokens_seen": 829792, + "step": 3115 + }, + { + "epoch": 0.817824377457405, + "grad_norm": 4.231538772583008, + "learning_rate": 4.08781127129751e-05, + "loss": 0.6467, + "num_input_tokens_seen": 830928, + "step": 3120 + }, + { + "epoch": 0.8191349934469201, + "grad_norm": 5.653774738311768, + "learning_rate": 4.094364351245085e-05, + "loss": 0.5433, + "num_input_tokens_seen": 832112, + "step": 3125 + }, + { + "epoch": 0.8204456094364351, + "grad_norm": 1.8984185457229614, + "learning_rate": 4.100917431192661e-05, + "loss": 0.4351, + "num_input_tokens_seen": 833344, + "step": 3130 + }, + { + "epoch": 0.8217562254259502, + "grad_norm": 12.164495468139648, + "learning_rate": 4.1074705111402365e-05, + "loss": 0.738, + "num_input_tokens_seen": 834624, + "step": 3135 + }, + { + "epoch": 0.8230668414154653, + "grad_norm": 8.599830627441406, + "learning_rate": 4.1140235910878115e-05, + "loss": 0.5488, + "num_input_tokens_seen": 835904, + "step": 3140 + }, + { + "epoch": 0.8243774574049804, + "grad_norm": 4.120616912841797, + "learning_rate": 4.1205766710353866e-05, + "loss": 0.489, + "num_input_tokens_seen": 837152, + "step": 3145 + }, + { + "epoch": 0.8256880733944955, + "grad_norm": 3.3423726558685303, + "learning_rate": 4.127129750982962e-05, + "loss": 0.5489, + "num_input_tokens_seen": 838768, + "step": 3150 + }, + { + "epoch": 0.8269986893840104, + "grad_norm": 6.321200370788574, + "learning_rate": 4.1336828309305374e-05, + "loss": 0.5348, + "num_input_tokens_seen": 839760, + "step": 3155 + }, + { + "epoch": 0.8283093053735255, + "grad_norm": 3.0220494270324707, + "learning_rate": 4.140235910878113e-05, + "loss": 0.5651, + "num_input_tokens_seen": 841056, + "step": 3160 + }, + { + "epoch": 0.8296199213630406, + "grad_norm": 6.284887790679932, + "learning_rate": 4.146788990825689e-05, + "loss": 0.6113, + "num_input_tokens_seen": 842432, + "step": 3165 + }, + { + "epoch": 0.8309305373525557, + "grad_norm": 11.082125663757324, + "learning_rate": 4.153342070773263e-05, + "loss": 0.4581, + "num_input_tokens_seen": 843680, + "step": 3170 + }, + { + "epoch": 0.8322411533420708, + "grad_norm": 3.329486131668091, + "learning_rate": 4.159895150720839e-05, + "loss": 0.4841, + "num_input_tokens_seen": 844704, + "step": 3175 + }, + { + "epoch": 0.8335517693315858, + "grad_norm": 2.6636805534362793, + "learning_rate": 4.1664482306684147e-05, + "loss": 0.6268, + "num_input_tokens_seen": 845888, + "step": 3180 + }, + { + "epoch": 0.8348623853211009, + "grad_norm": 5.612847328186035, + "learning_rate": 4.17300131061599e-05, + "loss": 0.3916, + "num_input_tokens_seen": 847232, + "step": 3185 + }, + { + "epoch": 0.836173001310616, + "grad_norm": 3.4173336029052734, + "learning_rate": 4.1795543905635654e-05, + "loss": 0.5407, + "num_input_tokens_seen": 848480, + "step": 3190 + }, + { + "epoch": 0.8374836173001311, + "grad_norm": 6.157424449920654, + "learning_rate": 4.1861074705111405e-05, + "loss": 0.3889, + "num_input_tokens_seen": 849984, + "step": 3195 + }, + { + "epoch": 0.8387942332896461, + "grad_norm": 7.973843097686768, + "learning_rate": 4.1926605504587155e-05, + "loss": 0.5291, + "num_input_tokens_seen": 851776, + "step": 3200 + }, + { + "epoch": 0.8401048492791612, + "grad_norm": 10.494187355041504, + "learning_rate": 4.199213630406291e-05, + "loss": 0.5407, + "num_input_tokens_seen": 852880, + "step": 3205 + }, + { + "epoch": 0.8414154652686763, + "grad_norm": 2.590644359588623, + "learning_rate": 4.205766710353867e-05, + "loss": 0.5796, + "num_input_tokens_seen": 854176, + "step": 3210 + }, + { + "epoch": 0.8427260812581914, + "grad_norm": 2.8315157890319824, + "learning_rate": 4.2123197903014414e-05, + "loss": 0.4203, + "num_input_tokens_seen": 855488, + "step": 3215 + }, + { + "epoch": 0.8440366972477065, + "grad_norm": 2.6199119091033936, + "learning_rate": 4.218872870249017e-05, + "loss": 0.3939, + "num_input_tokens_seen": 856544, + "step": 3220 + }, + { + "epoch": 0.8453473132372215, + "grad_norm": 2.970715284347534, + "learning_rate": 4.225425950196593e-05, + "loss": 0.5536, + "num_input_tokens_seen": 858112, + "step": 3225 + }, + { + "epoch": 0.8466579292267365, + "grad_norm": 1.5228052139282227, + "learning_rate": 4.231979030144168e-05, + "loss": 0.5914, + "num_input_tokens_seen": 859360, + "step": 3230 + }, + { + "epoch": 0.8479685452162516, + "grad_norm": 2.149010181427002, + "learning_rate": 4.2385321100917436e-05, + "loss": 0.4809, + "num_input_tokens_seen": 860624, + "step": 3235 + }, + { + "epoch": 0.8492791612057667, + "grad_norm": 2.6885218620300293, + "learning_rate": 4.245085190039319e-05, + "loss": 0.5514, + "num_input_tokens_seen": 861904, + "step": 3240 + }, + { + "epoch": 0.8505897771952818, + "grad_norm": 17.483585357666016, + "learning_rate": 4.251638269986894e-05, + "loss": 0.705, + "num_input_tokens_seen": 862976, + "step": 3245 + }, + { + "epoch": 0.8519003931847968, + "grad_norm": 2.7285780906677246, + "learning_rate": 4.2581913499344695e-05, + "loss": 0.4433, + "num_input_tokens_seen": 864432, + "step": 3250 + }, + { + "epoch": 0.8532110091743119, + "grad_norm": 2.514153480529785, + "learning_rate": 4.264744429882045e-05, + "loss": 0.6507, + "num_input_tokens_seen": 865856, + "step": 3255 + }, + { + "epoch": 0.854521625163827, + "grad_norm": 10.833983421325684, + "learning_rate": 4.27129750982962e-05, + "loss": 0.234, + "num_input_tokens_seen": 867120, + "step": 3260 + }, + { + "epoch": 0.8558322411533421, + "grad_norm": 4.1989545822143555, + "learning_rate": 4.277850589777195e-05, + "loss": 0.4936, + "num_input_tokens_seen": 868624, + "step": 3265 + }, + { + "epoch": 0.8571428571428571, + "grad_norm": 10.043166160583496, + "learning_rate": 4.284403669724771e-05, + "loss": 0.3679, + "num_input_tokens_seen": 869712, + "step": 3270 + }, + { + "epoch": 0.8584534731323722, + "grad_norm": 5.21294641494751, + "learning_rate": 4.290956749672346e-05, + "loss": 0.569, + "num_input_tokens_seen": 871392, + "step": 3275 + }, + { + "epoch": 0.8597640891218873, + "grad_norm": 2.367223024368286, + "learning_rate": 4.297509829619922e-05, + "loss": 0.5429, + "num_input_tokens_seen": 872640, + "step": 3280 + }, + { + "epoch": 0.8610747051114024, + "grad_norm": 6.284125804901123, + "learning_rate": 4.304062909567497e-05, + "loss": 0.4703, + "num_input_tokens_seen": 873680, + "step": 3285 + }, + { + "epoch": 0.8623853211009175, + "grad_norm": 4.9927873611450195, + "learning_rate": 4.310615989515072e-05, + "loss": 0.466, + "num_input_tokens_seen": 874688, + "step": 3290 + }, + { + "epoch": 0.8636959370904325, + "grad_norm": 3.148082733154297, + "learning_rate": 4.3171690694626476e-05, + "loss": 0.569, + "num_input_tokens_seen": 875984, + "step": 3295 + }, + { + "epoch": 0.8650065530799476, + "grad_norm": 6.023165702819824, + "learning_rate": 4.3237221494102234e-05, + "loss": 0.4427, + "num_input_tokens_seen": 876976, + "step": 3300 + }, + { + "epoch": 0.8663171690694627, + "grad_norm": 2.9244823455810547, + "learning_rate": 4.3302752293577984e-05, + "loss": 0.5621, + "num_input_tokens_seen": 878336, + "step": 3305 + }, + { + "epoch": 0.8676277850589778, + "grad_norm": 4.102625846862793, + "learning_rate": 4.3368283093053735e-05, + "loss": 0.4927, + "num_input_tokens_seen": 879408, + "step": 3310 + }, + { + "epoch": 0.8689384010484927, + "grad_norm": 2.2959845066070557, + "learning_rate": 4.343381389252949e-05, + "loss": 0.4088, + "num_input_tokens_seen": 880640, + "step": 3315 + }, + { + "epoch": 0.8702490170380078, + "grad_norm": 5.344770908355713, + "learning_rate": 4.349934469200524e-05, + "loss": 0.4922, + "num_input_tokens_seen": 882400, + "step": 3320 + }, + { + "epoch": 0.8715596330275229, + "grad_norm": 24.39225959777832, + "learning_rate": 4.3564875491481e-05, + "loss": 0.5475, + "num_input_tokens_seen": 883584, + "step": 3325 + }, + { + "epoch": 0.872870249017038, + "grad_norm": 2.023585557937622, + "learning_rate": 4.363040629095675e-05, + "loss": 0.5965, + "num_input_tokens_seen": 884736, + "step": 3330 + }, + { + "epoch": 0.8741808650065531, + "grad_norm": 4.714507102966309, + "learning_rate": 4.36959370904325e-05, + "loss": 0.4086, + "num_input_tokens_seen": 886064, + "step": 3335 + }, + { + "epoch": 0.8754914809960681, + "grad_norm": 2.4556283950805664, + "learning_rate": 4.376146788990826e-05, + "loss": 0.5156, + "num_input_tokens_seen": 887312, + "step": 3340 + }, + { + "epoch": 0.8768020969855832, + "grad_norm": 3.188958168029785, + "learning_rate": 4.3826998689384016e-05, + "loss": 0.7225, + "num_input_tokens_seen": 888496, + "step": 3345 + }, + { + "epoch": 0.8781127129750983, + "grad_norm": 14.082079887390137, + "learning_rate": 4.3892529488859766e-05, + "loss": 0.4376, + "num_input_tokens_seen": 890160, + "step": 3350 + }, + { + "epoch": 0.8794233289646134, + "grad_norm": 2.501422643661499, + "learning_rate": 4.395806028833552e-05, + "loss": 0.5082, + "num_input_tokens_seen": 891584, + "step": 3355 + }, + { + "epoch": 0.8807339449541285, + "grad_norm": 4.313608169555664, + "learning_rate": 4.4023591087811274e-05, + "loss": 0.4939, + "num_input_tokens_seen": 892768, + "step": 3360 + }, + { + "epoch": 0.8820445609436435, + "grad_norm": 2.0840256214141846, + "learning_rate": 4.4089121887287024e-05, + "loss": 0.4936, + "num_input_tokens_seen": 894432, + "step": 3365 + }, + { + "epoch": 0.8833551769331586, + "grad_norm": 6.203019618988037, + "learning_rate": 4.415465268676278e-05, + "loss": 0.7352, + "num_input_tokens_seen": 895456, + "step": 3370 + }, + { + "epoch": 0.8846657929226737, + "grad_norm": 4.284923553466797, + "learning_rate": 4.422018348623854e-05, + "loss": 0.9053, + "num_input_tokens_seen": 896896, + "step": 3375 + }, + { + "epoch": 0.8859764089121888, + "grad_norm": 4.1956987380981445, + "learning_rate": 4.428571428571428e-05, + "loss": 0.9173, + "num_input_tokens_seen": 898368, + "step": 3380 + }, + { + "epoch": 0.8872870249017037, + "grad_norm": 4.412883281707764, + "learning_rate": 4.435124508519004e-05, + "loss": 0.5191, + "num_input_tokens_seen": 900576, + "step": 3385 + }, + { + "epoch": 0.8885976408912188, + "grad_norm": 1.9427191019058228, + "learning_rate": 4.44167758846658e-05, + "loss": 0.5815, + "num_input_tokens_seen": 902000, + "step": 3390 + }, + { + "epoch": 0.8899082568807339, + "grad_norm": 2.7885067462921143, + "learning_rate": 4.448230668414155e-05, + "loss": 0.519, + "num_input_tokens_seen": 903008, + "step": 3395 + }, + { + "epoch": 0.891218872870249, + "grad_norm": 2.4800949096679688, + "learning_rate": 4.45478374836173e-05, + "loss": 0.4472, + "num_input_tokens_seen": 904192, + "step": 3400 + }, + { + "epoch": 0.8925294888597641, + "grad_norm": 5.888509273529053, + "learning_rate": 4.4613368283093056e-05, + "loss": 0.6236, + "num_input_tokens_seen": 905344, + "step": 3405 + }, + { + "epoch": 0.8938401048492791, + "grad_norm": 5.27817964553833, + "learning_rate": 4.4678899082568806e-05, + "loss": 0.5549, + "num_input_tokens_seen": 906544, + "step": 3410 + }, + { + "epoch": 0.8951507208387942, + "grad_norm": 1.8244706392288208, + "learning_rate": 4.4744429882044564e-05, + "loss": 0.4221, + "num_input_tokens_seen": 907696, + "step": 3415 + }, + { + "epoch": 0.8964613368283093, + "grad_norm": 3.2579026222229004, + "learning_rate": 4.480996068152032e-05, + "loss": 0.5082, + "num_input_tokens_seen": 908800, + "step": 3420 + }, + { + "epoch": 0.8977719528178244, + "grad_norm": 1.521405816078186, + "learning_rate": 4.487549148099607e-05, + "loss": 0.3498, + "num_input_tokens_seen": 910208, + "step": 3425 + }, + { + "epoch": 0.8990825688073395, + "grad_norm": 3.3038651943206787, + "learning_rate": 4.494102228047182e-05, + "loss": 0.5349, + "num_input_tokens_seen": 911536, + "step": 3430 + }, + { + "epoch": 0.9003931847968545, + "grad_norm": 5.1953935623168945, + "learning_rate": 4.500655307994758e-05, + "loss": 0.5023, + "num_input_tokens_seen": 913008, + "step": 3435 + }, + { + "epoch": 0.9017038007863696, + "grad_norm": 4.0715227127075195, + "learning_rate": 4.507208387942333e-05, + "loss": 0.5088, + "num_input_tokens_seen": 914288, + "step": 3440 + }, + { + "epoch": 0.9030144167758847, + "grad_norm": 2.3136215209960938, + "learning_rate": 4.513761467889909e-05, + "loss": 0.4805, + "num_input_tokens_seen": 915296, + "step": 3445 + }, + { + "epoch": 0.9043250327653998, + "grad_norm": 1.7530795335769653, + "learning_rate": 4.520314547837484e-05, + "loss": 0.4889, + "num_input_tokens_seen": 917024, + "step": 3450 + }, + { + "epoch": 0.9056356487549148, + "grad_norm": 4.194927215576172, + "learning_rate": 4.526867627785059e-05, + "loss": 0.907, + "num_input_tokens_seen": 918224, + "step": 3455 + }, + { + "epoch": 0.9069462647444299, + "grad_norm": 5.374001502990723, + "learning_rate": 4.5334207077326345e-05, + "loss": 0.4873, + "num_input_tokens_seen": 919552, + "step": 3460 + }, + { + "epoch": 0.908256880733945, + "grad_norm": 2.939671039581299, + "learning_rate": 4.53997378768021e-05, + "loss": 0.6868, + "num_input_tokens_seen": 921136, + "step": 3465 + }, + { + "epoch": 0.90956749672346, + "grad_norm": 8.824606895446777, + "learning_rate": 4.546526867627785e-05, + "loss": 0.631, + "num_input_tokens_seen": 922080, + "step": 3470 + }, + { + "epoch": 0.9108781127129751, + "grad_norm": 2.173858880996704, + "learning_rate": 4.5530799475753604e-05, + "loss": 0.542, + "num_input_tokens_seen": 923664, + "step": 3475 + }, + { + "epoch": 0.9121887287024901, + "grad_norm": 3.411517381668091, + "learning_rate": 4.559633027522936e-05, + "loss": 0.5009, + "num_input_tokens_seen": 925280, + "step": 3480 + }, + { + "epoch": 0.9134993446920052, + "grad_norm": 3.8556556701660156, + "learning_rate": 4.566186107470511e-05, + "loss": 0.6504, + "num_input_tokens_seen": 926608, + "step": 3485 + }, + { + "epoch": 0.9148099606815203, + "grad_norm": 34.033042907714844, + "learning_rate": 4.572739187418087e-05, + "loss": 0.6214, + "num_input_tokens_seen": 927696, + "step": 3490 + }, + { + "epoch": 0.9161205766710354, + "grad_norm": 4.1882710456848145, + "learning_rate": 4.579292267365662e-05, + "loss": 0.5026, + "num_input_tokens_seen": 928944, + "step": 3495 + }, + { + "epoch": 0.9174311926605505, + "grad_norm": 7.34973669052124, + "learning_rate": 4.585845347313238e-05, + "loss": 0.4244, + "num_input_tokens_seen": 929872, + "step": 3500 + }, + { + "epoch": 0.9187418086500655, + "grad_norm": 5.230820655822754, + "learning_rate": 4.592398427260813e-05, + "loss": 0.5049, + "num_input_tokens_seen": 931072, + "step": 3505 + }, + { + "epoch": 0.9200524246395806, + "grad_norm": 2.699226140975952, + "learning_rate": 4.5989515072083885e-05, + "loss": 0.4648, + "num_input_tokens_seen": 932240, + "step": 3510 + }, + { + "epoch": 0.9213630406290957, + "grad_norm": 3.8674843311309814, + "learning_rate": 4.6055045871559635e-05, + "loss": 0.5867, + "num_input_tokens_seen": 933536, + "step": 3515 + }, + { + "epoch": 0.9226736566186108, + "grad_norm": 1.3170865774154663, + "learning_rate": 4.6120576671035386e-05, + "loss": 0.5016, + "num_input_tokens_seen": 935088, + "step": 3520 + }, + { + "epoch": 0.9239842726081258, + "grad_norm": 3.653167963027954, + "learning_rate": 4.618610747051114e-05, + "loss": 0.7138, + "num_input_tokens_seen": 936736, + "step": 3525 + }, + { + "epoch": 0.9252948885976409, + "grad_norm": 7.83473539352417, + "learning_rate": 4.6251638269986893e-05, + "loss": 0.6021, + "num_input_tokens_seen": 937920, + "step": 3530 + }, + { + "epoch": 0.926605504587156, + "grad_norm": 3.028937578201294, + "learning_rate": 4.631716906946265e-05, + "loss": 0.3813, + "num_input_tokens_seen": 939216, + "step": 3535 + }, + { + "epoch": 0.927916120576671, + "grad_norm": 4.7050065994262695, + "learning_rate": 4.63826998689384e-05, + "loss": 0.8294, + "num_input_tokens_seen": 940320, + "step": 3540 + }, + { + "epoch": 0.9292267365661862, + "grad_norm": 5.618896484375, + "learning_rate": 4.644823066841416e-05, + "loss": 0.4463, + "num_input_tokens_seen": 942096, + "step": 3545 + }, + { + "epoch": 0.9305373525557011, + "grad_norm": 7.89493989944458, + "learning_rate": 4.651376146788991e-05, + "loss": 0.5434, + "num_input_tokens_seen": 943216, + "step": 3550 + }, + { + "epoch": 0.9318479685452162, + "grad_norm": 2.6541941165924072, + "learning_rate": 4.6579292267365666e-05, + "loss": 0.5085, + "num_input_tokens_seen": 944368, + "step": 3555 + }, + { + "epoch": 0.9331585845347313, + "grad_norm": 2.4785399436950684, + "learning_rate": 4.664482306684142e-05, + "loss": 0.4567, + "num_input_tokens_seen": 948080, + "step": 3560 + }, + { + "epoch": 0.9344692005242464, + "grad_norm": 4.839658260345459, + "learning_rate": 4.671035386631717e-05, + "loss": 0.4064, + "num_input_tokens_seen": 949280, + "step": 3565 + }, + { + "epoch": 0.9357798165137615, + "grad_norm": 69.17341613769531, + "learning_rate": 4.6775884665792925e-05, + "loss": 0.6331, + "num_input_tokens_seen": 950448, + "step": 3570 + }, + { + "epoch": 0.9370904325032765, + "grad_norm": 3.6433627605438232, + "learning_rate": 4.684141546526868e-05, + "loss": 0.6239, + "num_input_tokens_seen": 951600, + "step": 3575 + }, + { + "epoch": 0.9384010484927916, + "grad_norm": 2.4331092834472656, + "learning_rate": 4.690694626474443e-05, + "loss": 0.5982, + "num_input_tokens_seen": 952992, + "step": 3580 + }, + { + "epoch": 0.9397116644823067, + "grad_norm": 1.670027732849121, + "learning_rate": 4.697247706422019e-05, + "loss": 0.5285, + "num_input_tokens_seen": 955136, + "step": 3585 + }, + { + "epoch": 0.9410222804718218, + "grad_norm": 8.448213577270508, + "learning_rate": 4.703800786369594e-05, + "loss": 0.5412, + "num_input_tokens_seen": 956288, + "step": 3590 + }, + { + "epoch": 0.9423328964613368, + "grad_norm": 1.9383037090301514, + "learning_rate": 4.710353866317169e-05, + "loss": 0.4053, + "num_input_tokens_seen": 957568, + "step": 3595 + }, + { + "epoch": 0.9436435124508519, + "grad_norm": 3.1501169204711914, + "learning_rate": 4.716906946264745e-05, + "loss": 0.4515, + "num_input_tokens_seen": 959616, + "step": 3600 + }, + { + "epoch": 0.944954128440367, + "grad_norm": 4.755237579345703, + "learning_rate": 4.72346002621232e-05, + "loss": 0.4682, + "num_input_tokens_seen": 961280, + "step": 3605 + }, + { + "epoch": 0.9462647444298821, + "grad_norm": 10.261889457702637, + "learning_rate": 4.730013106159895e-05, + "loss": 0.4327, + "num_input_tokens_seen": 962560, + "step": 3610 + }, + { + "epoch": 0.9475753604193972, + "grad_norm": 2.3457529544830322, + "learning_rate": 4.7365661861074707e-05, + "loss": 0.3693, + "num_input_tokens_seen": 963632, + "step": 3615 + }, + { + "epoch": 0.9488859764089121, + "grad_norm": 3.722864866256714, + "learning_rate": 4.7431192660550464e-05, + "loss": 0.5701, + "num_input_tokens_seen": 965024, + "step": 3620 + }, + { + "epoch": 0.9501965923984272, + "grad_norm": 3.844956636428833, + "learning_rate": 4.7496723460026214e-05, + "loss": 0.4521, + "num_input_tokens_seen": 966192, + "step": 3625 + }, + { + "epoch": 0.9515072083879423, + "grad_norm": 2.4837608337402344, + "learning_rate": 4.756225425950197e-05, + "loss": 0.4777, + "num_input_tokens_seen": 967776, + "step": 3630 + }, + { + "epoch": 0.9528178243774574, + "grad_norm": 17.44171142578125, + "learning_rate": 4.762778505897772e-05, + "loss": 0.3727, + "num_input_tokens_seen": 969040, + "step": 3635 + }, + { + "epoch": 0.9541284403669725, + "grad_norm": 4.630838394165039, + "learning_rate": 4.769331585845347e-05, + "loss": 0.4824, + "num_input_tokens_seen": 970624, + "step": 3640 + }, + { + "epoch": 0.9554390563564875, + "grad_norm": 4.589667320251465, + "learning_rate": 4.775884665792923e-05, + "loss": 0.4746, + "num_input_tokens_seen": 972032, + "step": 3645 + }, + { + "epoch": 0.9567496723460026, + "grad_norm": 2.561443567276001, + "learning_rate": 4.782437745740499e-05, + "loss": 0.395, + "num_input_tokens_seen": 973440, + "step": 3650 + }, + { + "epoch": 0.9580602883355177, + "grad_norm": 4.3606462478637695, + "learning_rate": 4.788990825688074e-05, + "loss": 0.463, + "num_input_tokens_seen": 974784, + "step": 3655 + }, + { + "epoch": 0.9593709043250328, + "grad_norm": 6.098413944244385, + "learning_rate": 4.795543905635649e-05, + "loss": 0.4676, + "num_input_tokens_seen": 975728, + "step": 3660 + }, + { + "epoch": 0.9606815203145478, + "grad_norm": 1.163827657699585, + "learning_rate": 4.8020969855832246e-05, + "loss": 0.4358, + "num_input_tokens_seen": 977648, + "step": 3665 + }, + { + "epoch": 0.9619921363040629, + "grad_norm": 4.633989334106445, + "learning_rate": 4.8086500655307996e-05, + "loss": 0.4378, + "num_input_tokens_seen": 978944, + "step": 3670 + }, + { + "epoch": 0.963302752293578, + "grad_norm": 1.6289641857147217, + "learning_rate": 4.8152031454783754e-05, + "loss": 0.5447, + "num_input_tokens_seen": 979872, + "step": 3675 + }, + { + "epoch": 0.9646133682830931, + "grad_norm": 3.36187744140625, + "learning_rate": 4.8217562254259504e-05, + "loss": 0.456, + "num_input_tokens_seen": 981296, + "step": 3680 + }, + { + "epoch": 0.9659239842726082, + "grad_norm": 8.345768928527832, + "learning_rate": 4.8283093053735255e-05, + "loss": 0.7038, + "num_input_tokens_seen": 982272, + "step": 3685 + }, + { + "epoch": 0.9672346002621232, + "grad_norm": 3.0126147270202637, + "learning_rate": 4.834862385321101e-05, + "loss": 0.4498, + "num_input_tokens_seen": 983808, + "step": 3690 + }, + { + "epoch": 0.9685452162516383, + "grad_norm": 4.699247360229492, + "learning_rate": 4.841415465268677e-05, + "loss": 0.4843, + "num_input_tokens_seen": 984896, + "step": 3695 + }, + { + "epoch": 0.9698558322411533, + "grad_norm": 1.526923656463623, + "learning_rate": 4.847968545216252e-05, + "loss": 0.4137, + "num_input_tokens_seen": 986224, + "step": 3700 + }, + { + "epoch": 0.9711664482306684, + "grad_norm": 1.5035260915756226, + "learning_rate": 4.854521625163827e-05, + "loss": 0.5333, + "num_input_tokens_seen": 987552, + "step": 3705 + }, + { + "epoch": 0.9724770642201835, + "grad_norm": 6.157303333282471, + "learning_rate": 4.861074705111403e-05, + "loss": 0.5377, + "num_input_tokens_seen": 988688, + "step": 3710 + }, + { + "epoch": 0.9737876802096985, + "grad_norm": 3.062502861022949, + "learning_rate": 4.867627785058978e-05, + "loss": 0.52, + "num_input_tokens_seen": 990032, + "step": 3715 + }, + { + "epoch": 0.9750982961992136, + "grad_norm": 1.549845814704895, + "learning_rate": 4.8741808650065535e-05, + "loss": 0.3418, + "num_input_tokens_seen": 991520, + "step": 3720 + }, + { + "epoch": 0.9764089121887287, + "grad_norm": 2.4715237617492676, + "learning_rate": 4.880733944954129e-05, + "loss": 0.5682, + "num_input_tokens_seen": 992640, + "step": 3725 + }, + { + "epoch": 0.9777195281782438, + "grad_norm": 4.2892937660217285, + "learning_rate": 4.8872870249017036e-05, + "loss": 0.5885, + "num_input_tokens_seen": 993872, + "step": 3730 + }, + { + "epoch": 0.9790301441677588, + "grad_norm": 1.6444200277328491, + "learning_rate": 4.8938401048492794e-05, + "loss": 0.4464, + "num_input_tokens_seen": 995584, + "step": 3735 + }, + { + "epoch": 0.9803407601572739, + "grad_norm": 4.341414928436279, + "learning_rate": 4.900393184796855e-05, + "loss": 0.4612, + "num_input_tokens_seen": 996864, + "step": 3740 + }, + { + "epoch": 0.981651376146789, + "grad_norm": 10.876659393310547, + "learning_rate": 4.90694626474443e-05, + "loss": 0.4458, + "num_input_tokens_seen": 998096, + "step": 3745 + }, + { + "epoch": 0.9829619921363041, + "grad_norm": 2.85394549369812, + "learning_rate": 4.913499344692005e-05, + "loss": 0.4511, + "num_input_tokens_seen": 999760, + "step": 3750 + }, + { + "epoch": 0.9842726081258192, + "grad_norm": 3.9688026905059814, + "learning_rate": 4.920052424639581e-05, + "loss": 0.5198, + "num_input_tokens_seen": 1000784, + "step": 3755 + }, + { + "epoch": 0.9855832241153342, + "grad_norm": 3.827826738357544, + "learning_rate": 4.926605504587156e-05, + "loss": 0.4642, + "num_input_tokens_seen": 1002528, + "step": 3760 + }, + { + "epoch": 0.9868938401048493, + "grad_norm": 2.6589643955230713, + "learning_rate": 4.933158584534732e-05, + "loss": 0.395, + "num_input_tokens_seen": 1003888, + "step": 3765 + }, + { + "epoch": 0.9882044560943644, + "grad_norm": 2.6767003536224365, + "learning_rate": 4.9397116644823075e-05, + "loss": 0.371, + "num_input_tokens_seen": 1005168, + "step": 3770 + }, + { + "epoch": 0.9895150720838795, + "grad_norm": 2.378314971923828, + "learning_rate": 4.946264744429882e-05, + "loss": 0.3822, + "num_input_tokens_seen": 1006624, + "step": 3775 + }, + { + "epoch": 0.9908256880733946, + "grad_norm": 3.319126844406128, + "learning_rate": 4.9528178243774576e-05, + "loss": 0.4333, + "num_input_tokens_seen": 1008224, + "step": 3780 + }, + { + "epoch": 0.9921363040629095, + "grad_norm": 0.897281289100647, + "learning_rate": 4.959370904325033e-05, + "loss": 0.4908, + "num_input_tokens_seen": 1010368, + "step": 3785 + }, + { + "epoch": 0.9934469200524246, + "grad_norm": 3.6097183227539062, + "learning_rate": 4.9659239842726083e-05, + "loss": 0.4747, + "num_input_tokens_seen": 1011632, + "step": 3790 + }, + { + "epoch": 0.9947575360419397, + "grad_norm": 7.0994768142700195, + "learning_rate": 4.972477064220184e-05, + "loss": 0.5943, + "num_input_tokens_seen": 1012704, + "step": 3795 + }, + { + "epoch": 0.9960681520314548, + "grad_norm": 6.950616359710693, + "learning_rate": 4.979030144167759e-05, + "loss": 0.6483, + "num_input_tokens_seen": 1015728, + "step": 3800 + }, + { + "epoch": 0.9973787680209698, + "grad_norm": 3.018352746963501, + "learning_rate": 4.985583224115334e-05, + "loss": 0.4208, + "num_input_tokens_seen": 1017264, + "step": 3805 + }, + { + "epoch": 0.9986893840104849, + "grad_norm": 4.113101959228516, + "learning_rate": 4.99213630406291e-05, + "loss": 0.6753, + "num_input_tokens_seen": 1018160, + "step": 3810 + }, + { + "epoch": 1.0, + "grad_norm": 2.680549383163452, + "learning_rate": 4.9986893840104856e-05, + "loss": 0.4525, + "num_input_tokens_seen": 1019608, + "step": 3815 + }, + { + "epoch": 1.000262123197903, + "eval_loss": 0.49553799629211426, + "eval_runtime": 18.1967, + "eval_samples_per_second": 46.602, + "eval_steps_per_second": 23.301, + "num_input_tokens_seen": 1019912, + "step": 3816 + }, + { + "epoch": 1.001310615989515, + "grad_norm": 3.479198694229126, + "learning_rate": 4.9999998325613814e-05, + "loss": 0.4531, + "num_input_tokens_seen": 1021096, + "step": 3820 + }, + { + "epoch": 1.0026212319790302, + "grad_norm": 2.9229607582092285, + "learning_rate": 4.999999152342031e-05, + "loss": 0.5793, + "num_input_tokens_seen": 1022328, + "step": 3825 + }, + { + "epoch": 1.0039318479685453, + "grad_norm": 10.2268705368042, + "learning_rate": 4.999997948877178e-05, + "loss": 0.5688, + "num_input_tokens_seen": 1023608, + "step": 3830 + }, + { + "epoch": 1.0052424639580604, + "grad_norm": 5.792478561401367, + "learning_rate": 4.9999962221670737e-05, + "loss": 0.4207, + "num_input_tokens_seen": 1024920, + "step": 3835 + }, + { + "epoch": 1.0065530799475753, + "grad_norm": 2.0326218605041504, + "learning_rate": 4.9999939722120804e-05, + "loss": 0.6241, + "num_input_tokens_seen": 1026392, + "step": 3840 + }, + { + "epoch": 1.0078636959370904, + "grad_norm": 4.051065921783447, + "learning_rate": 4.9999911990126674e-05, + "loss": 0.5847, + "num_input_tokens_seen": 1027848, + "step": 3845 + }, + { + "epoch": 1.0091743119266054, + "grad_norm": 11.348526954650879, + "learning_rate": 4.999987902569416e-05, + "loss": 0.4757, + "num_input_tokens_seen": 1029032, + "step": 3850 + }, + { + "epoch": 1.0104849279161205, + "grad_norm": 7.797495365142822, + "learning_rate": 4.9999840828830167e-05, + "loss": 0.5261, + "num_input_tokens_seen": 1031672, + "step": 3855 + }, + { + "epoch": 1.0117955439056356, + "grad_norm": 6.00745153427124, + "learning_rate": 4.999979739954269e-05, + "loss": 0.5885, + "num_input_tokens_seen": 1033304, + "step": 3860 + }, + { + "epoch": 1.0131061598951507, + "grad_norm": 3.6262881755828857, + "learning_rate": 4.999974873784081e-05, + "loss": 0.8432, + "num_input_tokens_seen": 1034728, + "step": 3865 + }, + { + "epoch": 1.0144167758846658, + "grad_norm": 5.109577655792236, + "learning_rate": 4.9999694843734715e-05, + "loss": 0.5486, + "num_input_tokens_seen": 1035864, + "step": 3870 + }, + { + "epoch": 1.015727391874181, + "grad_norm": 2.330989360809326, + "learning_rate": 4.999963571723568e-05, + "loss": 0.5239, + "num_input_tokens_seen": 1037352, + "step": 3875 + }, + { + "epoch": 1.017038007863696, + "grad_norm": 5.146859169006348, + "learning_rate": 4.9999571358356095e-05, + "loss": 0.6148, + "num_input_tokens_seen": 1038904, + "step": 3880 + }, + { + "epoch": 1.018348623853211, + "grad_norm": 3.542444944381714, + "learning_rate": 4.9999501767109416e-05, + "loss": 1.1073, + "num_input_tokens_seen": 1039992, + "step": 3885 + }, + { + "epoch": 1.019659239842726, + "grad_norm": 0.9979768395423889, + "learning_rate": 4.999942694351021e-05, + "loss": 0.44, + "num_input_tokens_seen": 1041640, + "step": 3890 + }, + { + "epoch": 1.020969855832241, + "grad_norm": 1.61745285987854, + "learning_rate": 4.9999346887574146e-05, + "loss": 0.4639, + "num_input_tokens_seen": 1042744, + "step": 3895 + }, + { + "epoch": 1.0222804718217562, + "grad_norm": 2.4563300609588623, + "learning_rate": 4.999926159931797e-05, + "loss": 0.4661, + "num_input_tokens_seen": 1043752, + "step": 3900 + }, + { + "epoch": 1.0235910878112713, + "grad_norm": 2.706158399581909, + "learning_rate": 4.999917107875954e-05, + "loss": 0.3971, + "num_input_tokens_seen": 1045080, + "step": 3905 + }, + { + "epoch": 1.0249017038007864, + "grad_norm": 3.0223007202148438, + "learning_rate": 4.999907532591779e-05, + "loss": 0.3286, + "num_input_tokens_seen": 1046216, + "step": 3910 + }, + { + "epoch": 1.0262123197903015, + "grad_norm": 5.007085800170898, + "learning_rate": 4.9998974340812775e-05, + "loss": 0.5021, + "num_input_tokens_seen": 1047176, + "step": 3915 + }, + { + "epoch": 1.0275229357798166, + "grad_norm": 2.5549046993255615, + "learning_rate": 4.999886812346563e-05, + "loss": 0.3383, + "num_input_tokens_seen": 1048728, + "step": 3920 + }, + { + "epoch": 1.0288335517693317, + "grad_norm": 3.8153603076934814, + "learning_rate": 4.999875667389858e-05, + "loss": 0.5922, + "num_input_tokens_seen": 1050120, + "step": 3925 + }, + { + "epoch": 1.0301441677588468, + "grad_norm": 2.5426993370056152, + "learning_rate": 4.999863999213495e-05, + "loss": 0.45, + "num_input_tokens_seen": 1052008, + "step": 3930 + }, + { + "epoch": 1.0314547837483616, + "grad_norm": 4.055718898773193, + "learning_rate": 4.999851807819916e-05, + "loss": 0.3943, + "num_input_tokens_seen": 1053192, + "step": 3935 + }, + { + "epoch": 1.0327653997378767, + "grad_norm": 2.581129789352417, + "learning_rate": 4.9998390932116734e-05, + "loss": 0.4687, + "num_input_tokens_seen": 1054616, + "step": 3940 + }, + { + "epoch": 1.0340760157273918, + "grad_norm": 5.169582366943359, + "learning_rate": 4.9998258553914286e-05, + "loss": 0.5473, + "num_input_tokens_seen": 1055656, + "step": 3945 + }, + { + "epoch": 1.035386631716907, + "grad_norm": 6.518865585327148, + "learning_rate": 4.999812094361951e-05, + "loss": 0.5443, + "num_input_tokens_seen": 1056904, + "step": 3950 + }, + { + "epoch": 1.036697247706422, + "grad_norm": 1.6065930128097534, + "learning_rate": 4.9997978101261215e-05, + "loss": 0.5709, + "num_input_tokens_seen": 1058184, + "step": 3955 + }, + { + "epoch": 1.0380078636959371, + "grad_norm": 3.22346830368042, + "learning_rate": 4.9997830026869294e-05, + "loss": 0.4069, + "num_input_tokens_seen": 1059432, + "step": 3960 + }, + { + "epoch": 1.0393184796854522, + "grad_norm": 2.1141433715820312, + "learning_rate": 4.999767672047474e-05, + "loss": 0.4526, + "num_input_tokens_seen": 1060728, + "step": 3965 + }, + { + "epoch": 1.0406290956749673, + "grad_norm": 6.777366638183594, + "learning_rate": 4.999751818210965e-05, + "loss": 0.664, + "num_input_tokens_seen": 1061928, + "step": 3970 + }, + { + "epoch": 1.0419397116644824, + "grad_norm": 2.4557156562805176, + "learning_rate": 4.999735441180718e-05, + "loss": 0.5117, + "num_input_tokens_seen": 1063128, + "step": 3975 + }, + { + "epoch": 1.0432503276539973, + "grad_norm": 1.513266682624817, + "learning_rate": 4.999718540960164e-05, + "loss": 0.5323, + "num_input_tokens_seen": 1064744, + "step": 3980 + }, + { + "epoch": 1.0445609436435124, + "grad_norm": 5.948873996734619, + "learning_rate": 4.999701117552839e-05, + "loss": 0.5598, + "num_input_tokens_seen": 1065784, + "step": 3985 + }, + { + "epoch": 1.0458715596330275, + "grad_norm": 2.4618680477142334, + "learning_rate": 4.999683170962388e-05, + "loss": 0.5726, + "num_input_tokens_seen": 1066936, + "step": 3990 + }, + { + "epoch": 1.0471821756225426, + "grad_norm": 2.1388704776763916, + "learning_rate": 4.9996647011925685e-05, + "loss": 0.4404, + "num_input_tokens_seen": 1068120, + "step": 3995 + }, + { + "epoch": 1.0484927916120577, + "grad_norm": 5.067405700683594, + "learning_rate": 4.9996457082472465e-05, + "loss": 0.6758, + "num_input_tokens_seen": 1069400, + "step": 4000 + }, + { + "epoch": 1.0498034076015728, + "grad_norm": 6.863471031188965, + "learning_rate": 4.999626192130397e-05, + "loss": 0.6239, + "num_input_tokens_seen": 1070360, + "step": 4005 + }, + { + "epoch": 1.0511140235910879, + "grad_norm": 3.0839946269989014, + "learning_rate": 4.999606152846104e-05, + "loss": 0.4233, + "num_input_tokens_seen": 1071672, + "step": 4010 + }, + { + "epoch": 1.052424639580603, + "grad_norm": 3.9888100624084473, + "learning_rate": 4.9995855903985624e-05, + "loss": 0.4426, + "num_input_tokens_seen": 1073336, + "step": 4015 + }, + { + "epoch": 1.053735255570118, + "grad_norm": 4.280579090118408, + "learning_rate": 4.9995645047920756e-05, + "loss": 0.4411, + "num_input_tokens_seen": 1074744, + "step": 4020 + }, + { + "epoch": 1.0550458715596331, + "grad_norm": 2.38657546043396, + "learning_rate": 4.999542896031056e-05, + "loss": 0.7415, + "num_input_tokens_seen": 1075976, + "step": 4025 + }, + { + "epoch": 1.056356487549148, + "grad_norm": 2.598572254180908, + "learning_rate": 4.999520764120029e-05, + "loss": 0.4045, + "num_input_tokens_seen": 1077192, + "step": 4030 + }, + { + "epoch": 1.0576671035386631, + "grad_norm": 1.8743196725845337, + "learning_rate": 4.999498109063624e-05, + "loss": 0.3358, + "num_input_tokens_seen": 1078376, + "step": 4035 + }, + { + "epoch": 1.0589777195281782, + "grad_norm": 3.293060779571533, + "learning_rate": 4.999474930866583e-05, + "loss": 0.6361, + "num_input_tokens_seen": 1079592, + "step": 4040 + }, + { + "epoch": 1.0602883355176933, + "grad_norm": 3.766364336013794, + "learning_rate": 4.999451229533759e-05, + "loss": 0.5234, + "num_input_tokens_seen": 1080696, + "step": 4045 + }, + { + "epoch": 1.0615989515072084, + "grad_norm": 4.0759596824646, + "learning_rate": 4.99942700507011e-05, + "loss": 0.9637, + "num_input_tokens_seen": 1082456, + "step": 4050 + }, + { + "epoch": 1.0629095674967235, + "grad_norm": 1.7492636442184448, + "learning_rate": 4.9994022574807085e-05, + "loss": 0.5813, + "num_input_tokens_seen": 1084024, + "step": 4055 + }, + { + "epoch": 1.0642201834862386, + "grad_norm": 1.2447739839553833, + "learning_rate": 4.999376986770733e-05, + "loss": 0.5246, + "num_input_tokens_seen": 1085800, + "step": 4060 + }, + { + "epoch": 1.0655307994757537, + "grad_norm": 2.5789270401000977, + "learning_rate": 4.999351192945473e-05, + "loss": 0.7247, + "num_input_tokens_seen": 1086792, + "step": 4065 + }, + { + "epoch": 1.0668414154652686, + "grad_norm": 4.895508289337158, + "learning_rate": 4.999324876010326e-05, + "loss": 0.4684, + "num_input_tokens_seen": 1087992, + "step": 4070 + }, + { + "epoch": 1.0681520314547837, + "grad_norm": 1.058543086051941, + "learning_rate": 4.999298035970801e-05, + "loss": 0.4474, + "num_input_tokens_seen": 1089624, + "step": 4075 + }, + { + "epoch": 1.0694626474442988, + "grad_norm": 3.0782065391540527, + "learning_rate": 4.999270672832516e-05, + "loss": 0.4677, + "num_input_tokens_seen": 1090840, + "step": 4080 + }, + { + "epoch": 1.0707732634338138, + "grad_norm": 1.9176151752471924, + "learning_rate": 4.9992427866011974e-05, + "loss": 0.5537, + "num_input_tokens_seen": 1092136, + "step": 4085 + }, + { + "epoch": 1.072083879423329, + "grad_norm": 2.4691312313079834, + "learning_rate": 4.9992143772826826e-05, + "loss": 0.4433, + "num_input_tokens_seen": 1093208, + "step": 4090 + }, + { + "epoch": 1.073394495412844, + "grad_norm": 2.930676221847534, + "learning_rate": 4.9991854448829165e-05, + "loss": 0.5983, + "num_input_tokens_seen": 1094712, + "step": 4095 + }, + { + "epoch": 1.0747051114023591, + "grad_norm": 1.9187074899673462, + "learning_rate": 4.999155989407954e-05, + "loss": 0.6638, + "num_input_tokens_seen": 1096184, + "step": 4100 + }, + { + "epoch": 1.0760157273918742, + "grad_norm": 1.3057550191879272, + "learning_rate": 4.999126010863963e-05, + "loss": 0.3869, + "num_input_tokens_seen": 1097656, + "step": 4105 + }, + { + "epoch": 1.0773263433813893, + "grad_norm": 3.097501516342163, + "learning_rate": 4.999095509257214e-05, + "loss": 0.2901, + "num_input_tokens_seen": 1098888, + "step": 4110 + }, + { + "epoch": 1.0786369593709044, + "grad_norm": 5.893470764160156, + "learning_rate": 4.999064484594095e-05, + "loss": 0.4789, + "num_input_tokens_seen": 1100056, + "step": 4115 + }, + { + "epoch": 1.0799475753604193, + "grad_norm": 2.2010891437530518, + "learning_rate": 4.999032936881096e-05, + "loss": 0.5757, + "num_input_tokens_seen": 1101320, + "step": 4120 + }, + { + "epoch": 1.0812581913499344, + "grad_norm": 3.369310140609741, + "learning_rate": 4.999000866124822e-05, + "loss": 0.6987, + "num_input_tokens_seen": 1102280, + "step": 4125 + }, + { + "epoch": 1.0825688073394495, + "grad_norm": 3.4118709564208984, + "learning_rate": 4.9989682723319846e-05, + "loss": 0.5763, + "num_input_tokens_seen": 1103464, + "step": 4130 + }, + { + "epoch": 1.0838794233289646, + "grad_norm": 1.3418960571289062, + "learning_rate": 4.9989351555094055e-05, + "loss": 0.5434, + "num_input_tokens_seen": 1104824, + "step": 4135 + }, + { + "epoch": 1.0851900393184797, + "grad_norm": 2.8186588287353516, + "learning_rate": 4.998901515664017e-05, + "loss": 0.3689, + "num_input_tokens_seen": 1105944, + "step": 4140 + }, + { + "epoch": 1.0865006553079948, + "grad_norm": 2.2977657318115234, + "learning_rate": 4.998867352802858e-05, + "loss": 0.5907, + "num_input_tokens_seen": 1107880, + "step": 4145 + }, + { + "epoch": 1.0878112712975099, + "grad_norm": 2.3262741565704346, + "learning_rate": 4.9988326669330797e-05, + "loss": 0.6195, + "num_input_tokens_seen": 1109336, + "step": 4150 + }, + { + "epoch": 1.089121887287025, + "grad_norm": 2.0605993270874023, + "learning_rate": 4.998797458061942e-05, + "loss": 0.4529, + "num_input_tokens_seen": 1110648, + "step": 4155 + }, + { + "epoch": 1.09043250327654, + "grad_norm": 1.2215328216552734, + "learning_rate": 4.998761726196814e-05, + "loss": 0.4635, + "num_input_tokens_seen": 1112584, + "step": 4160 + }, + { + "epoch": 1.091743119266055, + "grad_norm": 4.143659591674805, + "learning_rate": 4.9987254713451754e-05, + "loss": 0.6, + "num_input_tokens_seen": 1113928, + "step": 4165 + }, + { + "epoch": 1.09305373525557, + "grad_norm": 3.549055814743042, + "learning_rate": 4.998688693514612e-05, + "loss": 0.4872, + "num_input_tokens_seen": 1115768, + "step": 4170 + }, + { + "epoch": 1.0943643512450851, + "grad_norm": 1.407109260559082, + "learning_rate": 4.998651392712823e-05, + "loss": 0.4247, + "num_input_tokens_seen": 1117048, + "step": 4175 + }, + { + "epoch": 1.0956749672346002, + "grad_norm": 5.637406826019287, + "learning_rate": 4.998613568947614e-05, + "loss": 0.3289, + "num_input_tokens_seen": 1118136, + "step": 4180 + }, + { + "epoch": 1.0969855832241153, + "grad_norm": 2.173771858215332, + "learning_rate": 4.998575222226903e-05, + "loss": 0.399, + "num_input_tokens_seen": 1119608, + "step": 4185 + }, + { + "epoch": 1.0982961992136304, + "grad_norm": 2.075049877166748, + "learning_rate": 4.9985363525587145e-05, + "loss": 0.5412, + "num_input_tokens_seen": 1121000, + "step": 4190 + }, + { + "epoch": 1.0996068152031455, + "grad_norm": 3.221026659011841, + "learning_rate": 4.998496959951185e-05, + "loss": 0.4993, + "num_input_tokens_seen": 1122600, + "step": 4195 + }, + { + "epoch": 1.1009174311926606, + "grad_norm": 4.072919845581055, + "learning_rate": 4.998457044412559e-05, + "loss": 0.6504, + "num_input_tokens_seen": 1123736, + "step": 4200 + }, + { + "epoch": 1.1022280471821757, + "grad_norm": 2.4355690479278564, + "learning_rate": 4.998416605951191e-05, + "loss": 0.8397, + "num_input_tokens_seen": 1125096, + "step": 4205 + }, + { + "epoch": 1.1035386631716908, + "grad_norm": 4.295064926147461, + "learning_rate": 4.998375644575543e-05, + "loss": 0.7239, + "num_input_tokens_seen": 1126408, + "step": 4210 + }, + { + "epoch": 1.1048492791612057, + "grad_norm": 5.247220516204834, + "learning_rate": 4.99833416029419e-05, + "loss": 0.4479, + "num_input_tokens_seen": 1127528, + "step": 4215 + }, + { + "epoch": 1.1061598951507208, + "grad_norm": 3.8446950912475586, + "learning_rate": 4.998292153115814e-05, + "loss": 0.4742, + "num_input_tokens_seen": 1128504, + "step": 4220 + }, + { + "epoch": 1.1074705111402359, + "grad_norm": 3.026477575302124, + "learning_rate": 4.9982496230492085e-05, + "loss": 0.5844, + "num_input_tokens_seen": 1129640, + "step": 4225 + }, + { + "epoch": 1.108781127129751, + "grad_norm": 2.6401944160461426, + "learning_rate": 4.998206570103271e-05, + "loss": 0.658, + "num_input_tokens_seen": 1130856, + "step": 4230 + }, + { + "epoch": 1.110091743119266, + "grad_norm": 2.4469778537750244, + "learning_rate": 4.998162994287017e-05, + "loss": 0.6792, + "num_input_tokens_seen": 1132424, + "step": 4235 + }, + { + "epoch": 1.1114023591087812, + "grad_norm": 3.1317026615142822, + "learning_rate": 4.998118895609564e-05, + "loss": 0.3274, + "num_input_tokens_seen": 1133896, + "step": 4240 + }, + { + "epoch": 1.1127129750982963, + "grad_norm": 1.7231117486953735, + "learning_rate": 4.9980742740801425e-05, + "loss": 0.7579, + "num_input_tokens_seen": 1135320, + "step": 4245 + }, + { + "epoch": 1.1140235910878113, + "grad_norm": 3.971670627593994, + "learning_rate": 4.998029129708092e-05, + "loss": 0.5606, + "num_input_tokens_seen": 1136520, + "step": 4250 + }, + { + "epoch": 1.1153342070773264, + "grad_norm": 1.0966918468475342, + "learning_rate": 4.997983462502861e-05, + "loss": 0.3633, + "num_input_tokens_seen": 1138120, + "step": 4255 + }, + { + "epoch": 1.1166448230668413, + "grad_norm": 12.056770324707031, + "learning_rate": 4.997937272474007e-05, + "loss": 0.5978, + "num_input_tokens_seen": 1139256, + "step": 4260 + }, + { + "epoch": 1.1179554390563564, + "grad_norm": 6.057975769042969, + "learning_rate": 4.9978905596311985e-05, + "loss": 0.4442, + "num_input_tokens_seen": 1140472, + "step": 4265 + }, + { + "epoch": 1.1192660550458715, + "grad_norm": 3.4972901344299316, + "learning_rate": 4.9978433239842115e-05, + "loss": 0.5606, + "num_input_tokens_seen": 1141736, + "step": 4270 + }, + { + "epoch": 1.1205766710353866, + "grad_norm": 3.1188626289367676, + "learning_rate": 4.997795565542933e-05, + "loss": 0.4745, + "num_input_tokens_seen": 1142888, + "step": 4275 + }, + { + "epoch": 1.1218872870249017, + "grad_norm": 1.7907289266586304, + "learning_rate": 4.997747284317358e-05, + "loss": 0.6111, + "num_input_tokens_seen": 1144344, + "step": 4280 + }, + { + "epoch": 1.1231979030144168, + "grad_norm": 1.6452300548553467, + "learning_rate": 4.997698480317593e-05, + "loss": 0.437, + "num_input_tokens_seen": 1145432, + "step": 4285 + }, + { + "epoch": 1.124508519003932, + "grad_norm": 3.93662691116333, + "learning_rate": 4.997649153553851e-05, + "loss": 0.4021, + "num_input_tokens_seen": 1146600, + "step": 4290 + }, + { + "epoch": 1.125819134993447, + "grad_norm": 1.0137161016464233, + "learning_rate": 4.9975993040364574e-05, + "loss": 0.4635, + "num_input_tokens_seen": 1148088, + "step": 4295 + }, + { + "epoch": 1.127129750982962, + "grad_norm": 3.4959936141967773, + "learning_rate": 4.9975489317758446e-05, + "loss": 0.4012, + "num_input_tokens_seen": 1149400, + "step": 4300 + }, + { + "epoch": 1.1284403669724772, + "grad_norm": 5.734461307525635, + "learning_rate": 4.9974980367825555e-05, + "loss": 0.3203, + "num_input_tokens_seen": 1150584, + "step": 4305 + }, + { + "epoch": 1.129750982961992, + "grad_norm": 3.6145377159118652, + "learning_rate": 4.997446619067243e-05, + "loss": 0.4351, + "num_input_tokens_seen": 1151704, + "step": 4310 + }, + { + "epoch": 1.1310615989515072, + "grad_norm": 10.673559188842773, + "learning_rate": 4.997394678640669e-05, + "loss": 0.454, + "num_input_tokens_seen": 1153512, + "step": 4315 + }, + { + "epoch": 1.1323722149410222, + "grad_norm": 4.221001625061035, + "learning_rate": 4.997342215513703e-05, + "loss": 0.6989, + "num_input_tokens_seen": 1154744, + "step": 4320 + }, + { + "epoch": 1.1336828309305373, + "grad_norm": 3.4163143634796143, + "learning_rate": 4.9972892296973263e-05, + "loss": 0.7095, + "num_input_tokens_seen": 1156600, + "step": 4325 + }, + { + "epoch": 1.1349934469200524, + "grad_norm": 2.602552890777588, + "learning_rate": 4.997235721202629e-05, + "loss": 0.6226, + "num_input_tokens_seen": 1157672, + "step": 4330 + }, + { + "epoch": 1.1363040629095675, + "grad_norm": 2.3680813312530518, + "learning_rate": 4.997181690040811e-05, + "loss": 0.6347, + "num_input_tokens_seen": 1159400, + "step": 4335 + }, + { + "epoch": 1.1376146788990826, + "grad_norm": 3.7697858810424805, + "learning_rate": 4.997127136223179e-05, + "loss": 0.4858, + "num_input_tokens_seen": 1160664, + "step": 4340 + }, + { + "epoch": 1.1389252948885977, + "grad_norm": 3.3237574100494385, + "learning_rate": 4.997072059761153e-05, + "loss": 0.4652, + "num_input_tokens_seen": 1161816, + "step": 4345 + }, + { + "epoch": 1.1402359108781126, + "grad_norm": 2.791456460952759, + "learning_rate": 4.997016460666258e-05, + "loss": 0.3321, + "num_input_tokens_seen": 1163480, + "step": 4350 + }, + { + "epoch": 1.1415465268676277, + "grad_norm": 2.6806931495666504, + "learning_rate": 4.996960338950134e-05, + "loss": 0.476, + "num_input_tokens_seen": 1164680, + "step": 4355 + }, + { + "epoch": 1.1428571428571428, + "grad_norm": 2.5392260551452637, + "learning_rate": 4.996903694624525e-05, + "loss": 0.4346, + "num_input_tokens_seen": 1166088, + "step": 4360 + }, + { + "epoch": 1.144167758846658, + "grad_norm": 1.9325629472732544, + "learning_rate": 4.996846527701287e-05, + "loss": 0.4825, + "num_input_tokens_seen": 1167304, + "step": 4365 + }, + { + "epoch": 1.145478374836173, + "grad_norm": 5.360953330993652, + "learning_rate": 4.996788838192384e-05, + "loss": 0.4081, + "num_input_tokens_seen": 1168472, + "step": 4370 + }, + { + "epoch": 1.146788990825688, + "grad_norm": 1.7758482694625854, + "learning_rate": 4.9967306261098924e-05, + "loss": 0.6087, + "num_input_tokens_seen": 1169960, + "step": 4375 + }, + { + "epoch": 1.1480996068152032, + "grad_norm": 3.2754499912261963, + "learning_rate": 4.996671891465994e-05, + "loss": 0.5952, + "num_input_tokens_seen": 1171768, + "step": 4380 + }, + { + "epoch": 1.1494102228047183, + "grad_norm": 5.799715518951416, + "learning_rate": 4.996612634272983e-05, + "loss": 0.5031, + "num_input_tokens_seen": 1173512, + "step": 4385 + }, + { + "epoch": 1.1507208387942334, + "grad_norm": 4.001332759857178, + "learning_rate": 4.996552854543262e-05, + "loss": 0.4445, + "num_input_tokens_seen": 1175128, + "step": 4390 + }, + { + "epoch": 1.1520314547837485, + "grad_norm": 4.166684627532959, + "learning_rate": 4.9964925522893416e-05, + "loss": 0.5223, + "num_input_tokens_seen": 1176360, + "step": 4395 + }, + { + "epoch": 1.1533420707732633, + "grad_norm": 5.553712844848633, + "learning_rate": 4.9964317275238435e-05, + "loss": 0.5571, + "num_input_tokens_seen": 1177704, + "step": 4400 + }, + { + "epoch": 1.1546526867627784, + "grad_norm": 3.190680980682373, + "learning_rate": 4.996370380259499e-05, + "loss": 0.5161, + "num_input_tokens_seen": 1179224, + "step": 4405 + }, + { + "epoch": 1.1559633027522935, + "grad_norm": 4.121308326721191, + "learning_rate": 4.996308510509147e-05, + "loss": 0.5136, + "num_input_tokens_seen": 1180504, + "step": 4410 + }, + { + "epoch": 1.1572739187418086, + "grad_norm": 6.118819236755371, + "learning_rate": 4.9962461182857366e-05, + "loss": 0.5537, + "num_input_tokens_seen": 1181816, + "step": 4415 + }, + { + "epoch": 1.1585845347313237, + "grad_norm": 4.230898857116699, + "learning_rate": 4.9961832036023275e-05, + "loss": 1.0873, + "num_input_tokens_seen": 1182840, + "step": 4420 + }, + { + "epoch": 1.1598951507208388, + "grad_norm": 2.594989776611328, + "learning_rate": 4.996119766472087e-05, + "loss": 0.4524, + "num_input_tokens_seen": 1184184, + "step": 4425 + }, + { + "epoch": 1.161205766710354, + "grad_norm": 1.9066647291183472, + "learning_rate": 4.996055806908292e-05, + "loss": 0.4934, + "num_input_tokens_seen": 1185928, + "step": 4430 + }, + { + "epoch": 1.162516382699869, + "grad_norm": 3.2752344608306885, + "learning_rate": 4.99599132492433e-05, + "loss": 0.4392, + "num_input_tokens_seen": 1187352, + "step": 4435 + }, + { + "epoch": 1.163826998689384, + "grad_norm": 6.960115909576416, + "learning_rate": 4.995926320533695e-05, + "loss": 0.4219, + "num_input_tokens_seen": 1188712, + "step": 4440 + }, + { + "epoch": 1.165137614678899, + "grad_norm": 1.0107840299606323, + "learning_rate": 4.995860793749995e-05, + "loss": 0.5517, + "num_input_tokens_seen": 1190840, + "step": 4445 + }, + { + "epoch": 1.166448230668414, + "grad_norm": 9.009607315063477, + "learning_rate": 4.995794744586942e-05, + "loss": 0.4066, + "num_input_tokens_seen": 1191912, + "step": 4450 + }, + { + "epoch": 1.1677588466579292, + "grad_norm": 4.659906387329102, + "learning_rate": 4.9957281730583635e-05, + "loss": 0.667, + "num_input_tokens_seen": 1193624, + "step": 4455 + }, + { + "epoch": 1.1690694626474443, + "grad_norm": 1.589707851409912, + "learning_rate": 4.9956610791781885e-05, + "loss": 0.5257, + "num_input_tokens_seen": 1195048, + "step": 4460 + }, + { + "epoch": 1.1703800786369594, + "grad_norm": 2.1175811290740967, + "learning_rate": 4.995593462960464e-05, + "loss": 0.5736, + "num_input_tokens_seen": 1196744, + "step": 4465 + }, + { + "epoch": 1.1716906946264745, + "grad_norm": 1.3649318218231201, + "learning_rate": 4.9955253244193375e-05, + "loss": 0.5627, + "num_input_tokens_seen": 1198504, + "step": 4470 + }, + { + "epoch": 1.1730013106159896, + "grad_norm": 0.9985455870628357, + "learning_rate": 4.9954566635690744e-05, + "loss": 0.607, + "num_input_tokens_seen": 1200088, + "step": 4475 + }, + { + "epoch": 1.1743119266055047, + "grad_norm": 5.8241424560546875, + "learning_rate": 4.995387480424043e-05, + "loss": 0.4394, + "num_input_tokens_seen": 1201416, + "step": 4480 + }, + { + "epoch": 1.1756225425950197, + "grad_norm": 1.7106698751449585, + "learning_rate": 4.995317774998723e-05, + "loss": 0.5143, + "num_input_tokens_seen": 1202568, + "step": 4485 + }, + { + "epoch": 1.1769331585845348, + "grad_norm": 14.254616737365723, + "learning_rate": 4.995247547307704e-05, + "loss": 0.6019, + "num_input_tokens_seen": 1203832, + "step": 4490 + }, + { + "epoch": 1.1782437745740497, + "grad_norm": 6.9849419593811035, + "learning_rate": 4.995176797365686e-05, + "loss": 0.4523, + "num_input_tokens_seen": 1204824, + "step": 4495 + }, + { + "epoch": 1.1795543905635648, + "grad_norm": 2.8676178455352783, + "learning_rate": 4.995105525187475e-05, + "loss": 0.4681, + "num_input_tokens_seen": 1205864, + "step": 4500 + }, + { + "epoch": 1.18086500655308, + "grad_norm": 2.1489784717559814, + "learning_rate": 4.9950337307879884e-05, + "loss": 0.7389, + "num_input_tokens_seen": 1207192, + "step": 4505 + }, + { + "epoch": 1.182175622542595, + "grad_norm": 2.0811171531677246, + "learning_rate": 4.994961414182254e-05, + "loss": 0.4528, + "num_input_tokens_seen": 1208280, + "step": 4510 + }, + { + "epoch": 1.18348623853211, + "grad_norm": 3.2702693939208984, + "learning_rate": 4.9948885753854056e-05, + "loss": 0.35, + "num_input_tokens_seen": 1210200, + "step": 4515 + }, + { + "epoch": 1.1847968545216252, + "grad_norm": 3.059018135070801, + "learning_rate": 4.99481521441269e-05, + "loss": 0.5764, + "num_input_tokens_seen": 1211944, + "step": 4520 + }, + { + "epoch": 1.1861074705111403, + "grad_norm": 2.294731855392456, + "learning_rate": 4.99474133127946e-05, + "loss": 0.4432, + "num_input_tokens_seen": 1213192, + "step": 4525 + }, + { + "epoch": 1.1874180865006554, + "grad_norm": 5.270437717437744, + "learning_rate": 4.99466692600118e-05, + "loss": 0.5912, + "num_input_tokens_seen": 1214520, + "step": 4530 + }, + { + "epoch": 1.1887287024901703, + "grad_norm": 5.280127048492432, + "learning_rate": 4.994591998593423e-05, + "loss": 0.572, + "num_input_tokens_seen": 1215448, + "step": 4535 + }, + { + "epoch": 1.1900393184796854, + "grad_norm": 2.682750940322876, + "learning_rate": 4.9945165490718714e-05, + "loss": 0.4912, + "num_input_tokens_seen": 1216472, + "step": 4540 + }, + { + "epoch": 1.1913499344692005, + "grad_norm": 4.31819486618042, + "learning_rate": 4.994440577452316e-05, + "loss": 0.8913, + "num_input_tokens_seen": 1217512, + "step": 4545 + }, + { + "epoch": 1.1926605504587156, + "grad_norm": 1.7925559282302856, + "learning_rate": 4.9943640837506574e-05, + "loss": 0.5063, + "num_input_tokens_seen": 1219336, + "step": 4550 + }, + { + "epoch": 1.1939711664482306, + "grad_norm": 4.562820911407471, + "learning_rate": 4.994287067982907e-05, + "loss": 0.4476, + "num_input_tokens_seen": 1220408, + "step": 4555 + }, + { + "epoch": 1.1952817824377457, + "grad_norm": 6.6192307472229, + "learning_rate": 4.994209530165183e-05, + "loss": 0.3825, + "num_input_tokens_seen": 1221656, + "step": 4560 + }, + { + "epoch": 1.1965923984272608, + "grad_norm": 2.3031606674194336, + "learning_rate": 4.9941314703137134e-05, + "loss": 0.4669, + "num_input_tokens_seen": 1222808, + "step": 4565 + }, + { + "epoch": 1.197903014416776, + "grad_norm": 3.428365707397461, + "learning_rate": 4.994052888444837e-05, + "loss": 0.5162, + "num_input_tokens_seen": 1223944, + "step": 4570 + }, + { + "epoch": 1.199213630406291, + "grad_norm": 3.3788013458251953, + "learning_rate": 4.993973784575e-05, + "loss": 0.7011, + "num_input_tokens_seen": 1225096, + "step": 4575 + }, + { + "epoch": 1.2005242463958061, + "grad_norm": 3.789381265640259, + "learning_rate": 4.9938941587207586e-05, + "loss": 0.4095, + "num_input_tokens_seen": 1226040, + "step": 4580 + }, + { + "epoch": 1.2018348623853212, + "grad_norm": 3.4525792598724365, + "learning_rate": 4.9938140108987795e-05, + "loss": 0.4426, + "num_input_tokens_seen": 1227128, + "step": 4585 + }, + { + "epoch": 1.203145478374836, + "grad_norm": 3.5445830821990967, + "learning_rate": 4.9937333411258366e-05, + "loss": 0.5772, + "num_input_tokens_seen": 1228200, + "step": 4590 + }, + { + "epoch": 1.2044560943643512, + "grad_norm": 1.9707504510879517, + "learning_rate": 4.993652149418815e-05, + "loss": 0.4031, + "num_input_tokens_seen": 1229304, + "step": 4595 + }, + { + "epoch": 1.2057667103538663, + "grad_norm": 6.338958740234375, + "learning_rate": 4.9935704357947065e-05, + "loss": 0.4212, + "num_input_tokens_seen": 1231080, + "step": 4600 + }, + { + "epoch": 1.2070773263433814, + "grad_norm": 5.914894104003906, + "learning_rate": 4.9934882002706154e-05, + "loss": 0.5674, + "num_input_tokens_seen": 1232248, + "step": 4605 + }, + { + "epoch": 1.2083879423328965, + "grad_norm": 4.9003682136535645, + "learning_rate": 4.9934054428637514e-05, + "loss": 0.492, + "num_input_tokens_seen": 1233224, + "step": 4610 + }, + { + "epoch": 1.2096985583224116, + "grad_norm": 4.358852863311768, + "learning_rate": 4.9933221635914365e-05, + "loss": 0.5388, + "num_input_tokens_seen": 1235032, + "step": 4615 + }, + { + "epoch": 1.2110091743119267, + "grad_norm": 3.148399591445923, + "learning_rate": 4.9932383624711014e-05, + "loss": 0.4013, + "num_input_tokens_seen": 1236232, + "step": 4620 + }, + { + "epoch": 1.2123197903014418, + "grad_norm": 3.865182399749756, + "learning_rate": 4.993154039520285e-05, + "loss": 0.6304, + "num_input_tokens_seen": 1237464, + "step": 4625 + }, + { + "epoch": 1.2136304062909566, + "grad_norm": 15.974211692810059, + "learning_rate": 4.9930691947566354e-05, + "loss": 0.6017, + "num_input_tokens_seen": 1238360, + "step": 4630 + }, + { + "epoch": 1.2149410222804717, + "grad_norm": 8.073601722717285, + "learning_rate": 4.992983828197911e-05, + "loss": 0.6961, + "num_input_tokens_seen": 1239816, + "step": 4635 + }, + { + "epoch": 1.2162516382699868, + "grad_norm": 9.032161712646484, + "learning_rate": 4.99289793986198e-05, + "loss": 0.6098, + "num_input_tokens_seen": 1240904, + "step": 4640 + }, + { + "epoch": 1.217562254259502, + "grad_norm": 3.7445027828216553, + "learning_rate": 4.992811529766816e-05, + "loss": 0.3998, + "num_input_tokens_seen": 1242088, + "step": 4645 + }, + { + "epoch": 1.218872870249017, + "grad_norm": 1.0137075185775757, + "learning_rate": 4.992724597930508e-05, + "loss": 0.4911, + "num_input_tokens_seen": 1244120, + "step": 4650 + }, + { + "epoch": 1.2201834862385321, + "grad_norm": 2.8173880577087402, + "learning_rate": 4.9926371443712474e-05, + "loss": 0.361, + "num_input_tokens_seen": 1245336, + "step": 4655 + }, + { + "epoch": 1.2214941022280472, + "grad_norm": 5.092093467712402, + "learning_rate": 4.9925491691073396e-05, + "loss": 0.4598, + "num_input_tokens_seen": 1246632, + "step": 4660 + }, + { + "epoch": 1.2228047182175623, + "grad_norm": 16.401535034179688, + "learning_rate": 4.9924606721571975e-05, + "loss": 0.7671, + "num_input_tokens_seen": 1247368, + "step": 4665 + }, + { + "epoch": 1.2241153342070774, + "grad_norm": 1.5698373317718506, + "learning_rate": 4.9923716535393434e-05, + "loss": 0.4831, + "num_input_tokens_seen": 1248488, + "step": 4670 + }, + { + "epoch": 1.2254259501965925, + "grad_norm": 6.324403285980225, + "learning_rate": 4.9922821132724086e-05, + "loss": 0.5311, + "num_input_tokens_seen": 1250264, + "step": 4675 + }, + { + "epoch": 1.2267365661861074, + "grad_norm": 2.1559176445007324, + "learning_rate": 4.992192051375134e-05, + "loss": 0.4767, + "num_input_tokens_seen": 1251592, + "step": 4680 + }, + { + "epoch": 1.2280471821756225, + "grad_norm": 1.3505057096481323, + "learning_rate": 4.992101467866369e-05, + "loss": 0.4645, + "num_input_tokens_seen": 1253112, + "step": 4685 + }, + { + "epoch": 1.2293577981651376, + "grad_norm": 4.193678379058838, + "learning_rate": 4.9920103627650734e-05, + "loss": 0.4145, + "num_input_tokens_seen": 1254760, + "step": 4690 + }, + { + "epoch": 1.2306684141546527, + "grad_norm": 5.467373847961426, + "learning_rate": 4.991918736090314e-05, + "loss": 0.4206, + "num_input_tokens_seen": 1255720, + "step": 4695 + }, + { + "epoch": 1.2319790301441678, + "grad_norm": 3.4040892124176025, + "learning_rate": 4.991826587861269e-05, + "loss": 0.5309, + "num_input_tokens_seen": 1256712, + "step": 4700 + }, + { + "epoch": 1.2332896461336829, + "grad_norm": 254.21617126464844, + "learning_rate": 4.991733918097225e-05, + "loss": 0.6753, + "num_input_tokens_seen": 1257800, + "step": 4705 + }, + { + "epoch": 1.234600262123198, + "grad_norm": 3.4555084705352783, + "learning_rate": 4.9916407268175775e-05, + "loss": 0.4511, + "num_input_tokens_seen": 1259080, + "step": 4710 + }, + { + "epoch": 1.235910878112713, + "grad_norm": 3.69730281829834, + "learning_rate": 4.991547014041831e-05, + "loss": 0.3808, + "num_input_tokens_seen": 1260424, + "step": 4715 + }, + { + "epoch": 1.237221494102228, + "grad_norm": 2.1723690032958984, + "learning_rate": 4.9914527797895995e-05, + "loss": 0.4572, + "num_input_tokens_seen": 1261736, + "step": 4720 + }, + { + "epoch": 1.238532110091743, + "grad_norm": 3.1363346576690674, + "learning_rate": 4.9913580240806054e-05, + "loss": 0.3133, + "num_input_tokens_seen": 1262808, + "step": 4725 + }, + { + "epoch": 1.2398427260812581, + "grad_norm": 3.864301919937134, + "learning_rate": 4.991262746934682e-05, + "loss": 0.4714, + "num_input_tokens_seen": 1264648, + "step": 4730 + }, + { + "epoch": 1.2411533420707732, + "grad_norm": 3.958303213119507, + "learning_rate": 4.991166948371771e-05, + "loss": 0.5756, + "num_input_tokens_seen": 1265896, + "step": 4735 + }, + { + "epoch": 1.2424639580602883, + "grad_norm": 5.356152057647705, + "learning_rate": 4.991070628411921e-05, + "loss": 0.5668, + "num_input_tokens_seen": 1267352, + "step": 4740 + }, + { + "epoch": 1.2437745740498034, + "grad_norm": 1.7217473983764648, + "learning_rate": 4.9909737870752935e-05, + "loss": 0.4768, + "num_input_tokens_seen": 1269000, + "step": 4745 + }, + { + "epoch": 1.2450851900393185, + "grad_norm": 2.17436146736145, + "learning_rate": 4.990876424382156e-05, + "loss": 0.4806, + "num_input_tokens_seen": 1270120, + "step": 4750 + }, + { + "epoch": 1.2463958060288336, + "grad_norm": 3.779226303100586, + "learning_rate": 4.9907785403528863e-05, + "loss": 0.4731, + "num_input_tokens_seen": 1271160, + "step": 4755 + }, + { + "epoch": 1.2477064220183487, + "grad_norm": 3.444239854812622, + "learning_rate": 4.9906801350079724e-05, + "loss": 0.4604, + "num_input_tokens_seen": 1272424, + "step": 4760 + }, + { + "epoch": 1.2490170380078638, + "grad_norm": 1.8856843709945679, + "learning_rate": 4.9905812083680105e-05, + "loss": 0.5701, + "num_input_tokens_seen": 1273768, + "step": 4765 + }, + { + "epoch": 1.2503276539973789, + "grad_norm": 2.422234296798706, + "learning_rate": 4.990481760453704e-05, + "loss": 0.5883, + "num_input_tokens_seen": 1275160, + "step": 4770 + }, + { + "epoch": 1.2516382699868938, + "grad_norm": 4.066681385040283, + "learning_rate": 4.9903817912858684e-05, + "loss": 0.6633, + "num_input_tokens_seen": 1276248, + "step": 4775 + }, + { + "epoch": 1.2529488859764089, + "grad_norm": 2.2595713138580322, + "learning_rate": 4.9902813008854274e-05, + "loss": 0.354, + "num_input_tokens_seen": 1277496, + "step": 4780 + }, + { + "epoch": 1.254259501965924, + "grad_norm": 3.3294970989227295, + "learning_rate": 4.990180289273413e-05, + "loss": 0.4574, + "num_input_tokens_seen": 1278584, + "step": 4785 + }, + { + "epoch": 1.255570117955439, + "grad_norm": 3.503775119781494, + "learning_rate": 4.990078756470966e-05, + "loss": 0.6033, + "num_input_tokens_seen": 1279704, + "step": 4790 + }, + { + "epoch": 1.2568807339449541, + "grad_norm": 8.247633934020996, + "learning_rate": 4.989976702499339e-05, + "loss": 0.5348, + "num_input_tokens_seen": 1280984, + "step": 4795 + }, + { + "epoch": 1.2581913499344692, + "grad_norm": 6.627811908721924, + "learning_rate": 4.98987412737989e-05, + "loss": 0.5988, + "num_input_tokens_seen": 1282072, + "step": 4800 + }, + { + "epoch": 1.2595019659239843, + "grad_norm": 3.273416757583618, + "learning_rate": 4.989771031134088e-05, + "loss": 0.5097, + "num_input_tokens_seen": 1283176, + "step": 4805 + }, + { + "epoch": 1.2608125819134992, + "grad_norm": 67.68443298339844, + "learning_rate": 4.989667413783511e-05, + "loss": 0.6309, + "num_input_tokens_seen": 1284600, + "step": 4810 + }, + { + "epoch": 1.2621231979030143, + "grad_norm": 3.667262315750122, + "learning_rate": 4.9895632753498465e-05, + "loss": 0.4841, + "num_input_tokens_seen": 1285992, + "step": 4815 + }, + { + "epoch": 1.2634338138925294, + "grad_norm": 2.3704512119293213, + "learning_rate": 4.98945861585489e-05, + "loss": 0.3631, + "num_input_tokens_seen": 1287768, + "step": 4820 + }, + { + "epoch": 1.2647444298820445, + "grad_norm": 2.6663498878479004, + "learning_rate": 4.989353435320547e-05, + "loss": 0.4151, + "num_input_tokens_seen": 1289240, + "step": 4825 + }, + { + "epoch": 1.2660550458715596, + "grad_norm": 2.2680089473724365, + "learning_rate": 4.9892477337688306e-05, + "loss": 0.55, + "num_input_tokens_seen": 1290392, + "step": 4830 + }, + { + "epoch": 1.2673656618610747, + "grad_norm": 5.851598262786865, + "learning_rate": 4.9891415112218654e-05, + "loss": 0.667, + "num_input_tokens_seen": 1291656, + "step": 4835 + }, + { + "epoch": 1.2686762778505898, + "grad_norm": 3.73793363571167, + "learning_rate": 4.9890347677018826e-05, + "loss": 0.475, + "num_input_tokens_seen": 1292936, + "step": 4840 + }, + { + "epoch": 1.2699868938401049, + "grad_norm": 2.2508456707000732, + "learning_rate": 4.9889275032312235e-05, + "loss": 0.4965, + "num_input_tokens_seen": 1294424, + "step": 4845 + }, + { + "epoch": 1.27129750982962, + "grad_norm": 1.499815821647644, + "learning_rate": 4.988819717832338e-05, + "loss": 0.3884, + "num_input_tokens_seen": 1295848, + "step": 4850 + }, + { + "epoch": 1.272608125819135, + "grad_norm": 10.406848907470703, + "learning_rate": 4.988711411527786e-05, + "loss": 0.6188, + "num_input_tokens_seen": 1296808, + "step": 4855 + }, + { + "epoch": 1.2739187418086502, + "grad_norm": 4.163819313049316, + "learning_rate": 4.988602584340236e-05, + "loss": 0.4227, + "num_input_tokens_seen": 1298440, + "step": 4860 + }, + { + "epoch": 1.2752293577981653, + "grad_norm": 2.547050952911377, + "learning_rate": 4.988493236292465e-05, + "loss": 0.5575, + "num_input_tokens_seen": 1299848, + "step": 4865 + }, + { + "epoch": 1.2765399737876801, + "grad_norm": 8.973435401916504, + "learning_rate": 4.98838336740736e-05, + "loss": 0.4959, + "num_input_tokens_seen": 1301224, + "step": 4870 + }, + { + "epoch": 1.2778505897771952, + "grad_norm": 2.274423837661743, + "learning_rate": 4.9882729777079154e-05, + "loss": 0.2963, + "num_input_tokens_seen": 1302776, + "step": 4875 + }, + { + "epoch": 1.2791612057667103, + "grad_norm": 1.6019290685653687, + "learning_rate": 4.9881620672172355e-05, + "loss": 0.4002, + "num_input_tokens_seen": 1304264, + "step": 4880 + }, + { + "epoch": 1.2804718217562254, + "grad_norm": 2.6488237380981445, + "learning_rate": 4.988050635958535e-05, + "loss": 0.6916, + "num_input_tokens_seen": 1305528, + "step": 4885 + }, + { + "epoch": 1.2817824377457405, + "grad_norm": 6.10117244720459, + "learning_rate": 4.987938683955135e-05, + "loss": 0.3344, + "num_input_tokens_seen": 1306616, + "step": 4890 + }, + { + "epoch": 1.2830930537352556, + "grad_norm": 3.225538730621338, + "learning_rate": 4.987826211230467e-05, + "loss": 0.8327, + "num_input_tokens_seen": 1307752, + "step": 4895 + }, + { + "epoch": 1.2844036697247707, + "grad_norm": 4.6357622146606445, + "learning_rate": 4.987713217808072e-05, + "loss": 0.4776, + "num_input_tokens_seen": 1309224, + "step": 4900 + }, + { + "epoch": 1.2857142857142856, + "grad_norm": 2.3829383850097656, + "learning_rate": 4.987599703711599e-05, + "loss": 0.3027, + "num_input_tokens_seen": 1310376, + "step": 4905 + }, + { + "epoch": 1.2870249017038007, + "grad_norm": 2.961498975753784, + "learning_rate": 4.9874856689648065e-05, + "loss": 0.5493, + "num_input_tokens_seen": 1311480, + "step": 4910 + }, + { + "epoch": 1.2883355176933158, + "grad_norm": 4.906671524047852, + "learning_rate": 4.9873711135915605e-05, + "loss": 0.8474, + "num_input_tokens_seen": 1312904, + "step": 4915 + }, + { + "epoch": 1.2896461336828309, + "grad_norm": 3.1629297733306885, + "learning_rate": 4.987256037615839e-05, + "loss": 0.3765, + "num_input_tokens_seen": 1314040, + "step": 4920 + }, + { + "epoch": 1.290956749672346, + "grad_norm": 6.228118419647217, + "learning_rate": 4.987140441061726e-05, + "loss": 0.798, + "num_input_tokens_seen": 1315112, + "step": 4925 + }, + { + "epoch": 1.292267365661861, + "grad_norm": 1.2081496715545654, + "learning_rate": 4.987024323953417e-05, + "loss": 0.505, + "num_input_tokens_seen": 1318312, + "step": 4930 + }, + { + "epoch": 1.2935779816513762, + "grad_norm": 2.651796817779541, + "learning_rate": 4.9869076863152134e-05, + "loss": 0.3459, + "num_input_tokens_seen": 1319192, + "step": 4935 + }, + { + "epoch": 1.2948885976408913, + "grad_norm": 3.0379021167755127, + "learning_rate": 4.986790528171529e-05, + "loss": 0.4324, + "num_input_tokens_seen": 1320616, + "step": 4940 + }, + { + "epoch": 1.2961992136304064, + "grad_norm": 1.6606534719467163, + "learning_rate": 4.986672849546883e-05, + "loss": 0.3933, + "num_input_tokens_seen": 1321768, + "step": 4945 + }, + { + "epoch": 1.2975098296199215, + "grad_norm": 2.654423713684082, + "learning_rate": 4.9865546504659063e-05, + "loss": 0.6465, + "num_input_tokens_seen": 1323272, + "step": 4950 + }, + { + "epoch": 1.2988204456094365, + "grad_norm": 2.7398107051849365, + "learning_rate": 4.986435930953338e-05, + "loss": 0.4204, + "num_input_tokens_seen": 1324456, + "step": 4955 + }, + { + "epoch": 1.3001310615989516, + "grad_norm": 6.435380458831787, + "learning_rate": 4.986316691034026e-05, + "loss": 0.3965, + "num_input_tokens_seen": 1325496, + "step": 4960 + }, + { + "epoch": 1.3014416775884665, + "grad_norm": 2.1770172119140625, + "learning_rate": 4.986196930732926e-05, + "loss": 0.4606, + "num_input_tokens_seen": 1326568, + "step": 4965 + }, + { + "epoch": 1.3027522935779816, + "grad_norm": 1.781927227973938, + "learning_rate": 4.986076650075104e-05, + "loss": 0.3405, + "num_input_tokens_seen": 1327672, + "step": 4970 + }, + { + "epoch": 1.3040629095674967, + "grad_norm": 2.0926504135131836, + "learning_rate": 4.985955849085735e-05, + "loss": 0.4425, + "num_input_tokens_seen": 1329032, + "step": 4975 + }, + { + "epoch": 1.3053735255570118, + "grad_norm": 4.0224738121032715, + "learning_rate": 4.9858345277901034e-05, + "loss": 0.4891, + "num_input_tokens_seen": 1330056, + "step": 4980 + }, + { + "epoch": 1.306684141546527, + "grad_norm": 1.799001932144165, + "learning_rate": 4.9857126862135984e-05, + "loss": 0.612, + "num_input_tokens_seen": 1331384, + "step": 4985 + }, + { + "epoch": 1.307994757536042, + "grad_norm": 1.6876479387283325, + "learning_rate": 4.985590324381724e-05, + "loss": 0.4665, + "num_input_tokens_seen": 1332728, + "step": 4990 + }, + { + "epoch": 1.309305373525557, + "grad_norm": 2.3358747959136963, + "learning_rate": 4.98546744232009e-05, + "loss": 0.4953, + "num_input_tokens_seen": 1334216, + "step": 4995 + }, + { + "epoch": 1.310615989515072, + "grad_norm": 2.90468430519104, + "learning_rate": 4.9853440400544136e-05, + "loss": 0.706, + "num_input_tokens_seen": 1335656, + "step": 5000 + }, + { + "epoch": 1.311926605504587, + "grad_norm": 1.2325845956802368, + "learning_rate": 4.985220117610525e-05, + "loss": 0.4843, + "num_input_tokens_seen": 1336952, + "step": 5005 + }, + { + "epoch": 1.3132372214941022, + "grad_norm": 2.257413148880005, + "learning_rate": 4.985095675014359e-05, + "loss": 0.5122, + "num_input_tokens_seen": 1337960, + "step": 5010 + }, + { + "epoch": 1.3145478374836173, + "grad_norm": 2.1283884048461914, + "learning_rate": 4.984970712291963e-05, + "loss": 0.5835, + "num_input_tokens_seen": 1339256, + "step": 5015 + }, + { + "epoch": 1.3158584534731324, + "grad_norm": 6.144309043884277, + "learning_rate": 4.98484522946949e-05, + "loss": 0.4461, + "num_input_tokens_seen": 1340840, + "step": 5020 + }, + { + "epoch": 1.3171690694626474, + "grad_norm": 3.6274302005767822, + "learning_rate": 4.984719226573205e-05, + "loss": 0.5655, + "num_input_tokens_seen": 1342216, + "step": 5025 + }, + { + "epoch": 1.3184796854521625, + "grad_norm": 3.9430692195892334, + "learning_rate": 4.984592703629478e-05, + "loss": 0.6924, + "num_input_tokens_seen": 1343672, + "step": 5030 + }, + { + "epoch": 1.3197903014416776, + "grad_norm": 3.1660656929016113, + "learning_rate": 4.9844656606647924e-05, + "loss": 0.6727, + "num_input_tokens_seen": 1345048, + "step": 5035 + }, + { + "epoch": 1.3211009174311927, + "grad_norm": 5.959998607635498, + "learning_rate": 4.984338097705736e-05, + "loss": 0.499, + "num_input_tokens_seen": 1346664, + "step": 5040 + }, + { + "epoch": 1.3224115334207078, + "grad_norm": 0.9007425308227539, + "learning_rate": 4.984210014779008e-05, + "loss": 0.5282, + "num_input_tokens_seen": 1348584, + "step": 5045 + }, + { + "epoch": 1.323722149410223, + "grad_norm": 2.957144021987915, + "learning_rate": 4.984081411911417e-05, + "loss": 0.4595, + "num_input_tokens_seen": 1349624, + "step": 5050 + }, + { + "epoch": 1.3250327653997378, + "grad_norm": 1.6422086954116821, + "learning_rate": 4.9839522891298784e-05, + "loss": 0.527, + "num_input_tokens_seen": 1350968, + "step": 5055 + }, + { + "epoch": 1.326343381389253, + "grad_norm": 4.144062042236328, + "learning_rate": 4.983822646461417e-05, + "loss": 0.712, + "num_input_tokens_seen": 1352152, + "step": 5060 + }, + { + "epoch": 1.327653997378768, + "grad_norm": 4.491000175476074, + "learning_rate": 4.9836924839331686e-05, + "loss": 0.6705, + "num_input_tokens_seen": 1353400, + "step": 5065 + }, + { + "epoch": 1.328964613368283, + "grad_norm": 4.141427040100098, + "learning_rate": 4.983561801572374e-05, + "loss": 0.6542, + "num_input_tokens_seen": 1354552, + "step": 5070 + }, + { + "epoch": 1.3302752293577982, + "grad_norm": 3.353884696960449, + "learning_rate": 4.983430599406386e-05, + "loss": 0.5149, + "num_input_tokens_seen": 1355912, + "step": 5075 + }, + { + "epoch": 1.3315858453473133, + "grad_norm": 5.218926429748535, + "learning_rate": 4.983298877462664e-05, + "loss": 0.8383, + "num_input_tokens_seen": 1357288, + "step": 5080 + }, + { + "epoch": 1.3328964613368284, + "grad_norm": 3.839869260787964, + "learning_rate": 4.983166635768778e-05, + "loss": 0.4622, + "num_input_tokens_seen": 1358456, + "step": 5085 + }, + { + "epoch": 1.3342070773263432, + "grad_norm": 1.9699840545654297, + "learning_rate": 4.9830338743524064e-05, + "loss": 0.6204, + "num_input_tokens_seen": 1360072, + "step": 5090 + }, + { + "epoch": 1.3355176933158583, + "grad_norm": 4.1153788566589355, + "learning_rate": 4.982900593241334e-05, + "loss": 0.6784, + "num_input_tokens_seen": 1361176, + "step": 5095 + }, + { + "epoch": 1.3368283093053734, + "grad_norm": 2.820051431655884, + "learning_rate": 4.982766792463458e-05, + "loss": 0.593, + "num_input_tokens_seen": 1362328, + "step": 5100 + }, + { + "epoch": 1.3381389252948885, + "grad_norm": 2.6360740661621094, + "learning_rate": 4.9826324720467834e-05, + "loss": 0.4286, + "num_input_tokens_seen": 1363992, + "step": 5105 + }, + { + "epoch": 1.3394495412844036, + "grad_norm": 2.718240976333618, + "learning_rate": 4.982497632019421e-05, + "loss": 0.3497, + "num_input_tokens_seen": 1365448, + "step": 5110 + }, + { + "epoch": 1.3407601572739187, + "grad_norm": 5.610764026641846, + "learning_rate": 4.9823622724095936e-05, + "loss": 0.5747, + "num_input_tokens_seen": 1366600, + "step": 5115 + }, + { + "epoch": 1.3420707732634338, + "grad_norm": 2.360482931137085, + "learning_rate": 4.982226393245632e-05, + "loss": 0.5584, + "num_input_tokens_seen": 1367816, + "step": 5120 + }, + { + "epoch": 1.343381389252949, + "grad_norm": 3.3222622871398926, + "learning_rate": 4.982089994555975e-05, + "loss": 0.3572, + "num_input_tokens_seen": 1369016, + "step": 5125 + }, + { + "epoch": 1.344692005242464, + "grad_norm": 4.021669387817383, + "learning_rate": 4.981953076369171e-05, + "loss": 0.796, + "num_input_tokens_seen": 1370360, + "step": 5130 + }, + { + "epoch": 1.3460026212319791, + "grad_norm": 1.1235102415084839, + "learning_rate": 4.981815638713877e-05, + "loss": 0.66, + "num_input_tokens_seen": 1371656, + "step": 5135 + }, + { + "epoch": 1.3473132372214942, + "grad_norm": 2.495849609375, + "learning_rate": 4.981677681618858e-05, + "loss": 0.5208, + "num_input_tokens_seen": 1372904, + "step": 5140 + }, + { + "epoch": 1.3486238532110093, + "grad_norm": 3.0730161666870117, + "learning_rate": 4.981539205112988e-05, + "loss": 0.3761, + "num_input_tokens_seen": 1375528, + "step": 5145 + }, + { + "epoch": 1.3499344692005242, + "grad_norm": 2.196380138397217, + "learning_rate": 4.98140020922525e-05, + "loss": 0.5335, + "num_input_tokens_seen": 1377032, + "step": 5150 + }, + { + "epoch": 1.3512450851900393, + "grad_norm": 2.443507671356201, + "learning_rate": 4.9812606939847356e-05, + "loss": 0.42, + "num_input_tokens_seen": 1378536, + "step": 5155 + }, + { + "epoch": 1.3525557011795544, + "grad_norm": 4.101146221160889, + "learning_rate": 4.981120659420646e-05, + "loss": 0.3946, + "num_input_tokens_seen": 1379816, + "step": 5160 + }, + { + "epoch": 1.3538663171690695, + "grad_norm": 8.25256633758545, + "learning_rate": 4.9809801055622887e-05, + "loss": 0.5144, + "num_input_tokens_seen": 1380936, + "step": 5165 + }, + { + "epoch": 1.3551769331585846, + "grad_norm": 1.4814704656600952, + "learning_rate": 4.980839032439082e-05, + "loss": 0.5061, + "num_input_tokens_seen": 1382152, + "step": 5170 + }, + { + "epoch": 1.3564875491480997, + "grad_norm": 3.9106192588806152, + "learning_rate": 4.980697440080553e-05, + "loss": 0.6181, + "num_input_tokens_seen": 1383160, + "step": 5175 + }, + { + "epoch": 1.3577981651376148, + "grad_norm": 2.8762896060943604, + "learning_rate": 4.980555328516335e-05, + "loss": 0.589, + "num_input_tokens_seen": 1384376, + "step": 5180 + }, + { + "epoch": 1.3591087811271296, + "grad_norm": 1.1207369565963745, + "learning_rate": 4.9804126977761747e-05, + "loss": 0.4532, + "num_input_tokens_seen": 1385800, + "step": 5185 + }, + { + "epoch": 1.3604193971166447, + "grad_norm": 3.892688274383545, + "learning_rate": 4.980269547889921e-05, + "loss": 0.4329, + "num_input_tokens_seen": 1386760, + "step": 5190 + }, + { + "epoch": 1.3617300131061598, + "grad_norm": 2.160311698913574, + "learning_rate": 4.980125878887537e-05, + "loss": 0.4732, + "num_input_tokens_seen": 1388008, + "step": 5195 + }, + { + "epoch": 1.363040629095675, + "grad_norm": 3.8244733810424805, + "learning_rate": 4.979981690799092e-05, + "loss": 0.4949, + "num_input_tokens_seen": 1389272, + "step": 5200 + }, + { + "epoch": 1.36435124508519, + "grad_norm": 7.656268119812012, + "learning_rate": 4.9798369836547644e-05, + "loss": 0.4337, + "num_input_tokens_seen": 1390952, + "step": 5205 + }, + { + "epoch": 1.365661861074705, + "grad_norm": 1.6119651794433594, + "learning_rate": 4.97969175748484e-05, + "loss": 0.4403, + "num_input_tokens_seen": 1392584, + "step": 5210 + }, + { + "epoch": 1.3669724770642202, + "grad_norm": 2.0880274772644043, + "learning_rate": 4.9795460123197166e-05, + "loss": 0.4204, + "num_input_tokens_seen": 1393896, + "step": 5215 + }, + { + "epoch": 1.3682830930537353, + "grad_norm": 4.171200275421143, + "learning_rate": 4.979399748189896e-05, + "loss": 0.4711, + "num_input_tokens_seen": 1394808, + "step": 5220 + }, + { + "epoch": 1.3695937090432504, + "grad_norm": 3.954911708831787, + "learning_rate": 4.979252965125993e-05, + "loss": 0.6281, + "num_input_tokens_seen": 1395992, + "step": 5225 + }, + { + "epoch": 1.3709043250327655, + "grad_norm": 8.555139541625977, + "learning_rate": 4.9791056631587276e-05, + "loss": 0.7328, + "num_input_tokens_seen": 1397048, + "step": 5230 + }, + { + "epoch": 1.3722149410222806, + "grad_norm": 5.875662803649902, + "learning_rate": 4.978957842318931e-05, + "loss": 0.4233, + "num_input_tokens_seen": 1398168, + "step": 5235 + }, + { + "epoch": 1.3735255570117955, + "grad_norm": 3.6526243686676025, + "learning_rate": 4.978809502637541e-05, + "loss": 0.7053, + "num_input_tokens_seen": 1399448, + "step": 5240 + }, + { + "epoch": 1.3748361730013106, + "grad_norm": 2.2703769207000732, + "learning_rate": 4.9786606441456054e-05, + "loss": 0.5529, + "num_input_tokens_seen": 1400680, + "step": 5245 + }, + { + "epoch": 1.3761467889908257, + "grad_norm": 4.648558139801025, + "learning_rate": 4.9785112668742796e-05, + "loss": 0.3431, + "num_input_tokens_seen": 1402232, + "step": 5250 + }, + { + "epoch": 1.3774574049803407, + "grad_norm": 2.5565059185028076, + "learning_rate": 4.978361370854828e-05, + "loss": 0.3069, + "num_input_tokens_seen": 1403640, + "step": 5255 + }, + { + "epoch": 1.3787680209698558, + "grad_norm": 4.994054317474365, + "learning_rate": 4.978210956118624e-05, + "loss": 0.3363, + "num_input_tokens_seen": 1404696, + "step": 5260 + }, + { + "epoch": 1.380078636959371, + "grad_norm": 6.166210174560547, + "learning_rate": 4.9780600226971486e-05, + "loss": 0.4926, + "num_input_tokens_seen": 1406168, + "step": 5265 + }, + { + "epoch": 1.381389252948886, + "grad_norm": 1.2562886476516724, + "learning_rate": 4.977908570621993e-05, + "loss": 0.4672, + "num_input_tokens_seen": 1407608, + "step": 5270 + }, + { + "epoch": 1.382699868938401, + "grad_norm": 3.271016836166382, + "learning_rate": 4.977756599924854e-05, + "loss": 0.5321, + "num_input_tokens_seen": 1408760, + "step": 5275 + }, + { + "epoch": 1.384010484927916, + "grad_norm": 63.962398529052734, + "learning_rate": 4.97760411063754e-05, + "loss": 0.4618, + "num_input_tokens_seen": 1409800, + "step": 5280 + }, + { + "epoch": 1.385321100917431, + "grad_norm": 3.1313059329986572, + "learning_rate": 4.977451102791968e-05, + "loss": 0.5378, + "num_input_tokens_seen": 1411096, + "step": 5285 + }, + { + "epoch": 1.3866317169069462, + "grad_norm": 1.176604986190796, + "learning_rate": 4.977297576420159e-05, + "loss": 0.5513, + "num_input_tokens_seen": 1413016, + "step": 5290 + }, + { + "epoch": 1.3879423328964613, + "grad_norm": 1.4792574644088745, + "learning_rate": 4.977143531554249e-05, + "loss": 0.5388, + "num_input_tokens_seen": 1414424, + "step": 5295 + }, + { + "epoch": 1.3892529488859764, + "grad_norm": 2.6324641704559326, + "learning_rate": 4.9769889682264774e-05, + "loss": 0.405, + "num_input_tokens_seen": 1415528, + "step": 5300 + }, + { + "epoch": 1.3905635648754915, + "grad_norm": 3.5557749271392822, + "learning_rate": 4.976833886469196e-05, + "loss": 0.4614, + "num_input_tokens_seen": 1416648, + "step": 5305 + }, + { + "epoch": 1.3918741808650066, + "grad_norm": 1.2878283262252808, + "learning_rate": 4.97667828631486e-05, + "loss": 0.5225, + "num_input_tokens_seen": 1418456, + "step": 5310 + }, + { + "epoch": 1.3931847968545217, + "grad_norm": 2.7889416217803955, + "learning_rate": 4.976522167796038e-05, + "loss": 0.4432, + "num_input_tokens_seen": 1420152, + "step": 5315 + }, + { + "epoch": 1.3944954128440368, + "grad_norm": 1.106245517730713, + "learning_rate": 4.9763655309454064e-05, + "loss": 0.4627, + "num_input_tokens_seen": 1422296, + "step": 5320 + }, + { + "epoch": 1.3958060288335519, + "grad_norm": 2.7735586166381836, + "learning_rate": 4.976208375795748e-05, + "loss": 0.4917, + "num_input_tokens_seen": 1423624, + "step": 5325 + }, + { + "epoch": 1.397116644823067, + "grad_norm": 3.2569377422332764, + "learning_rate": 4.9760507023799544e-05, + "loss": 0.3797, + "num_input_tokens_seen": 1424808, + "step": 5330 + }, + { + "epoch": 1.3984272608125818, + "grad_norm": 5.0525221824646, + "learning_rate": 4.9758925107310276e-05, + "loss": 0.9482, + "num_input_tokens_seen": 1426072, + "step": 5335 + }, + { + "epoch": 1.399737876802097, + "grad_norm": 4.987411975860596, + "learning_rate": 4.975733800882077e-05, + "loss": 0.5994, + "num_input_tokens_seen": 1427336, + "step": 5340 + }, + { + "epoch": 1.401048492791612, + "grad_norm": 1.376250147819519, + "learning_rate": 4.975574572866318e-05, + "loss": 0.4652, + "num_input_tokens_seen": 1428744, + "step": 5345 + }, + { + "epoch": 1.4023591087811271, + "grad_norm": 1.9612778425216675, + "learning_rate": 4.97541482671708e-05, + "loss": 0.5954, + "num_input_tokens_seen": 1430136, + "step": 5350 + }, + { + "epoch": 1.4036697247706422, + "grad_norm": 4.010923385620117, + "learning_rate": 4.975254562467794e-05, + "loss": 0.4812, + "num_input_tokens_seen": 1431368, + "step": 5355 + }, + { + "epoch": 1.4049803407601573, + "grad_norm": 4.700248718261719, + "learning_rate": 4.9750937801520064e-05, + "loss": 0.5614, + "num_input_tokens_seen": 1432680, + "step": 5360 + }, + { + "epoch": 1.4062909567496724, + "grad_norm": 2.403756856918335, + "learning_rate": 4.974932479803367e-05, + "loss": 0.5306, + "num_input_tokens_seen": 1433864, + "step": 5365 + }, + { + "epoch": 1.4076015727391873, + "grad_norm": 2.099158525466919, + "learning_rate": 4.974770661455636e-05, + "loss": 0.4734, + "num_input_tokens_seen": 1435064, + "step": 5370 + }, + { + "epoch": 1.4089121887287024, + "grad_norm": 4.965158462524414, + "learning_rate": 4.974608325142681e-05, + "loss": 0.457, + "num_input_tokens_seen": 1436328, + "step": 5375 + }, + { + "epoch": 1.4102228047182175, + "grad_norm": 2.540417432785034, + "learning_rate": 4.97444547089848e-05, + "loss": 0.5281, + "num_input_tokens_seen": 1437512, + "step": 5380 + }, + { + "epoch": 1.4115334207077326, + "grad_norm": 1.3547626733779907, + "learning_rate": 4.974282098757118e-05, + "loss": 0.5929, + "num_input_tokens_seen": 1439128, + "step": 5385 + }, + { + "epoch": 1.4128440366972477, + "grad_norm": 3.5174238681793213, + "learning_rate": 4.974118208752787e-05, + "loss": 0.4265, + "num_input_tokens_seen": 1440776, + "step": 5390 + }, + { + "epoch": 1.4141546526867628, + "grad_norm": 3.1637203693389893, + "learning_rate": 4.97395380091979e-05, + "loss": 0.4088, + "num_input_tokens_seen": 1442120, + "step": 5395 + }, + { + "epoch": 1.4154652686762779, + "grad_norm": 1.305765986442566, + "learning_rate": 4.9737888752925375e-05, + "loss": 0.4932, + "num_input_tokens_seen": 1443512, + "step": 5400 + }, + { + "epoch": 1.416775884665793, + "grad_norm": 2.482024908065796, + "learning_rate": 4.973623431905548e-05, + "loss": 0.5654, + "num_input_tokens_seen": 1444856, + "step": 5405 + }, + { + "epoch": 1.418086500655308, + "grad_norm": 2.1631808280944824, + "learning_rate": 4.973457470793447e-05, + "loss": 0.3657, + "num_input_tokens_seen": 1446232, + "step": 5410 + }, + { + "epoch": 1.4193971166448232, + "grad_norm": 2.0853159427642822, + "learning_rate": 4.973290991990972e-05, + "loss": 0.5706, + "num_input_tokens_seen": 1447384, + "step": 5415 + }, + { + "epoch": 1.4207077326343382, + "grad_norm": 1.8786041736602783, + "learning_rate": 4.973123995532966e-05, + "loss": 0.6728, + "num_input_tokens_seen": 1448744, + "step": 5420 + }, + { + "epoch": 1.4220183486238533, + "grad_norm": 5.752572059631348, + "learning_rate": 4.972956481454381e-05, + "loss": 0.4863, + "num_input_tokens_seen": 1451880, + "step": 5425 + }, + { + "epoch": 1.4233289646133682, + "grad_norm": 3.27070689201355, + "learning_rate": 4.972788449790277e-05, + "loss": 0.5236, + "num_input_tokens_seen": 1453432, + "step": 5430 + }, + { + "epoch": 1.4246395806028833, + "grad_norm": 2.580162286758423, + "learning_rate": 4.9726199005758234e-05, + "loss": 0.5521, + "num_input_tokens_seen": 1455272, + "step": 5435 + }, + { + "epoch": 1.4259501965923984, + "grad_norm": 4.188800811767578, + "learning_rate": 4.972450833846297e-05, + "loss": 0.4673, + "num_input_tokens_seen": 1456424, + "step": 5440 + }, + { + "epoch": 1.4272608125819135, + "grad_norm": 4.362132549285889, + "learning_rate": 4.972281249637083e-05, + "loss": 0.4791, + "num_input_tokens_seen": 1457576, + "step": 5445 + }, + { + "epoch": 1.4285714285714286, + "grad_norm": 3.5692903995513916, + "learning_rate": 4.9721111479836756e-05, + "loss": 0.4573, + "num_input_tokens_seen": 1459240, + "step": 5450 + }, + { + "epoch": 1.4298820445609437, + "grad_norm": 2.1606335639953613, + "learning_rate": 4.971940528921676e-05, + "loss": 0.7131, + "num_input_tokens_seen": 1460872, + "step": 5455 + }, + { + "epoch": 1.4311926605504588, + "grad_norm": 2.3678531646728516, + "learning_rate": 4.9717693924867944e-05, + "loss": 0.5002, + "num_input_tokens_seen": 1461896, + "step": 5460 + }, + { + "epoch": 1.4325032765399737, + "grad_norm": 4.343686580657959, + "learning_rate": 4.9715977387148504e-05, + "loss": 0.736, + "num_input_tokens_seen": 1463144, + "step": 5465 + }, + { + "epoch": 1.4338138925294888, + "grad_norm": 2.6443111896514893, + "learning_rate": 4.971425567641771e-05, + "loss": 0.6407, + "num_input_tokens_seen": 1464536, + "step": 5470 + }, + { + "epoch": 1.4351245085190039, + "grad_norm": 2.058511972427368, + "learning_rate": 4.97125287930359e-05, + "loss": 0.3969, + "num_input_tokens_seen": 1465640, + "step": 5475 + }, + { + "epoch": 1.436435124508519, + "grad_norm": 4.379739284515381, + "learning_rate": 4.971079673736451e-05, + "loss": 0.5459, + "num_input_tokens_seen": 1467144, + "step": 5480 + }, + { + "epoch": 1.437745740498034, + "grad_norm": 1.5105164051055908, + "learning_rate": 4.970905950976606e-05, + "loss": 0.5189, + "num_input_tokens_seen": 1468440, + "step": 5485 + }, + { + "epoch": 1.4390563564875491, + "grad_norm": 1.671203851699829, + "learning_rate": 4.970731711060415e-05, + "loss": 0.6242, + "num_input_tokens_seen": 1469848, + "step": 5490 + }, + { + "epoch": 1.4403669724770642, + "grad_norm": 1.9800145626068115, + "learning_rate": 4.970556954024346e-05, + "loss": 0.4499, + "num_input_tokens_seen": 1471448, + "step": 5495 + }, + { + "epoch": 1.4416775884665793, + "grad_norm": 2.1021335124969482, + "learning_rate": 4.970381679904975e-05, + "loss": 0.3866, + "num_input_tokens_seen": 1472552, + "step": 5500 + }, + { + "epoch": 1.4429882044560944, + "grad_norm": 1.8267673254013062, + "learning_rate": 4.970205888738988e-05, + "loss": 0.6135, + "num_input_tokens_seen": 1474424, + "step": 5505 + }, + { + "epoch": 1.4442988204456095, + "grad_norm": 2.7570786476135254, + "learning_rate": 4.970029580563176e-05, + "loss": 0.5388, + "num_input_tokens_seen": 1475544, + "step": 5510 + }, + { + "epoch": 1.4456094364351246, + "grad_norm": 2.155921220779419, + "learning_rate": 4.96985275541444e-05, + "loss": 0.7551, + "num_input_tokens_seen": 1476536, + "step": 5515 + }, + { + "epoch": 1.4469200524246395, + "grad_norm": 2.546236038208008, + "learning_rate": 4.969675413329791e-05, + "loss": 0.7202, + "num_input_tokens_seen": 1477656, + "step": 5520 + }, + { + "epoch": 1.4482306684141546, + "grad_norm": 4.914178848266602, + "learning_rate": 4.969497554346344e-05, + "loss": 0.4297, + "num_input_tokens_seen": 1478856, + "step": 5525 + }, + { + "epoch": 1.4495412844036697, + "grad_norm": 2.290057420730591, + "learning_rate": 4.969319178501327e-05, + "loss": 0.4899, + "num_input_tokens_seen": 1480232, + "step": 5530 + }, + { + "epoch": 1.4508519003931848, + "grad_norm": 1.6418383121490479, + "learning_rate": 4.969140285832072e-05, + "loss": 0.5161, + "num_input_tokens_seen": 1481960, + "step": 5535 + }, + { + "epoch": 1.4521625163826999, + "grad_norm": 1.7969509363174438, + "learning_rate": 4.968960876376022e-05, + "loss": 0.4134, + "num_input_tokens_seen": 1483256, + "step": 5540 + }, + { + "epoch": 1.453473132372215, + "grad_norm": 4.206799030303955, + "learning_rate": 4.9687809501707264e-05, + "loss": 0.426, + "num_input_tokens_seen": 1485400, + "step": 5545 + }, + { + "epoch": 1.45478374836173, + "grad_norm": 1.2263095378875732, + "learning_rate": 4.968600507253843e-05, + "loss": 0.3806, + "num_input_tokens_seen": 1486936, + "step": 5550 + }, + { + "epoch": 1.456094364351245, + "grad_norm": 2.298912763595581, + "learning_rate": 4.968419547663139e-05, + "loss": 0.4592, + "num_input_tokens_seen": 1488440, + "step": 5555 + }, + { + "epoch": 1.45740498034076, + "grad_norm": 4.316133975982666, + "learning_rate": 4.9682380714364897e-05, + "loss": 0.4966, + "num_input_tokens_seen": 1489960, + "step": 5560 + }, + { + "epoch": 1.4587155963302751, + "grad_norm": 2.090860366821289, + "learning_rate": 4.968056078611876e-05, + "loss": 0.6137, + "num_input_tokens_seen": 1491288, + "step": 5565 + }, + { + "epoch": 1.4600262123197902, + "grad_norm": 1.5496792793273926, + "learning_rate": 4.96787356922739e-05, + "loss": 0.5158, + "num_input_tokens_seen": 1492696, + "step": 5570 + }, + { + "epoch": 1.4613368283093053, + "grad_norm": 1.5149002075195312, + "learning_rate": 4.9676905433212296e-05, + "loss": 0.678, + "num_input_tokens_seen": 1494424, + "step": 5575 + }, + { + "epoch": 1.4626474442988204, + "grad_norm": 3.780266523361206, + "learning_rate": 4.967507000931702e-05, + "loss": 0.6351, + "num_input_tokens_seen": 1496136, + "step": 5580 + }, + { + "epoch": 1.4639580602883355, + "grad_norm": 2.248699188232422, + "learning_rate": 4.967322942097225e-05, + "loss": 0.2689, + "num_input_tokens_seen": 1497112, + "step": 5585 + }, + { + "epoch": 1.4652686762778506, + "grad_norm": 3.4716901779174805, + "learning_rate": 4.9671383668563167e-05, + "loss": 0.4937, + "num_input_tokens_seen": 1498456, + "step": 5590 + }, + { + "epoch": 1.4665792922673657, + "grad_norm": 11.66008472442627, + "learning_rate": 4.966953275247612e-05, + "loss": 0.4789, + "num_input_tokens_seen": 1499576, + "step": 5595 + }, + { + "epoch": 1.4678899082568808, + "grad_norm": 3.386140823364258, + "learning_rate": 4.966767667309849e-05, + "loss": 0.6179, + "num_input_tokens_seen": 1500536, + "step": 5600 + }, + { + "epoch": 1.469200524246396, + "grad_norm": 3.6100728511810303, + "learning_rate": 4.966581543081876e-05, + "loss": 0.5405, + "num_input_tokens_seen": 1501880, + "step": 5605 + }, + { + "epoch": 1.470511140235911, + "grad_norm": 4.028811454772949, + "learning_rate": 4.966394902602649e-05, + "loss": 0.4301, + "num_input_tokens_seen": 1503240, + "step": 5610 + }, + { + "epoch": 1.4718217562254259, + "grad_norm": 1.8648440837860107, + "learning_rate": 4.966207745911229e-05, + "loss": 0.5353, + "num_input_tokens_seen": 1504440, + "step": 5615 + }, + { + "epoch": 1.473132372214941, + "grad_norm": 2.453352928161621, + "learning_rate": 4.9660200730467894e-05, + "loss": 0.3979, + "num_input_tokens_seen": 1505928, + "step": 5620 + }, + { + "epoch": 1.474442988204456, + "grad_norm": 6.307033061981201, + "learning_rate": 4.9658318840486095e-05, + "loss": 0.5737, + "num_input_tokens_seen": 1507384, + "step": 5625 + }, + { + "epoch": 1.4757536041939712, + "grad_norm": 2.955228567123413, + "learning_rate": 4.9656431789560765e-05, + "loss": 0.5065, + "num_input_tokens_seen": 1508264, + "step": 5630 + }, + { + "epoch": 1.4770642201834863, + "grad_norm": 1.5923436880111694, + "learning_rate": 4.9654539578086865e-05, + "loss": 0.5031, + "num_input_tokens_seen": 1510024, + "step": 5635 + }, + { + "epoch": 1.4783748361730014, + "grad_norm": 4.7808756828308105, + "learning_rate": 4.965264220646044e-05, + "loss": 0.6555, + "num_input_tokens_seen": 1511240, + "step": 5640 + }, + { + "epoch": 1.4796854521625165, + "grad_norm": 1.4378057718276978, + "learning_rate": 4.965073967507859e-05, + "loss": 0.3737, + "num_input_tokens_seen": 1513128, + "step": 5645 + }, + { + "epoch": 1.4809960681520313, + "grad_norm": 2.0018532276153564, + "learning_rate": 4.9648831984339525e-05, + "loss": 0.4505, + "num_input_tokens_seen": 1514088, + "step": 5650 + }, + { + "epoch": 1.4823066841415464, + "grad_norm": 2.3433713912963867, + "learning_rate": 4.9646919134642513e-05, + "loss": 0.5171, + "num_input_tokens_seen": 1515560, + "step": 5655 + }, + { + "epoch": 1.4836173001310615, + "grad_norm": 2.0862865447998047, + "learning_rate": 4.964500112638791e-05, + "loss": 0.5604, + "num_input_tokens_seen": 1516488, + "step": 5660 + }, + { + "epoch": 1.4849279161205766, + "grad_norm": 1.558788537979126, + "learning_rate": 4.964307795997716e-05, + "loss": 0.5298, + "num_input_tokens_seen": 1518008, + "step": 5665 + }, + { + "epoch": 1.4862385321100917, + "grad_norm": 1.1311637163162231, + "learning_rate": 4.964114963581277e-05, + "loss": 0.4874, + "num_input_tokens_seen": 1519064, + "step": 5670 + }, + { + "epoch": 1.4875491480996068, + "grad_norm": 4.7496795654296875, + "learning_rate": 4.963921615429833e-05, + "loss": 0.4993, + "num_input_tokens_seen": 1520536, + "step": 5675 + }, + { + "epoch": 1.488859764089122, + "grad_norm": 3.4135820865631104, + "learning_rate": 4.963727751583853e-05, + "loss": 0.6647, + "num_input_tokens_seen": 1522408, + "step": 5680 + }, + { + "epoch": 1.490170380078637, + "grad_norm": 3.420074224472046, + "learning_rate": 4.963533372083912e-05, + "loss": 0.4918, + "num_input_tokens_seen": 1523528, + "step": 5685 + }, + { + "epoch": 1.491480996068152, + "grad_norm": 1.7694140672683716, + "learning_rate": 4.963338476970692e-05, + "loss": 0.4254, + "num_input_tokens_seen": 1525144, + "step": 5690 + }, + { + "epoch": 1.4927916120576672, + "grad_norm": 37.33271789550781, + "learning_rate": 4.963143066284986e-05, + "loss": 0.601, + "num_input_tokens_seen": 1526216, + "step": 5695 + }, + { + "epoch": 1.4941022280471823, + "grad_norm": 2.098437786102295, + "learning_rate": 4.962947140067691e-05, + "loss": 0.3702, + "num_input_tokens_seen": 1527272, + "step": 5700 + }, + { + "epoch": 1.4954128440366974, + "grad_norm": 6.478491306304932, + "learning_rate": 4.962750698359816e-05, + "loss": 0.4828, + "num_input_tokens_seen": 1528312, + "step": 5705 + }, + { + "epoch": 1.4967234600262123, + "grad_norm": 5.510324001312256, + "learning_rate": 4.962553741202475e-05, + "loss": 0.6449, + "num_input_tokens_seen": 1529432, + "step": 5710 + }, + { + "epoch": 1.4980340760157274, + "grad_norm": 2.0894205570220947, + "learning_rate": 4.962356268636891e-05, + "loss": 0.4604, + "num_input_tokens_seen": 1530888, + "step": 5715 + }, + { + "epoch": 1.4993446920052425, + "grad_norm": 4.926305770874023, + "learning_rate": 4.962158280704395e-05, + "loss": 0.3652, + "num_input_tokens_seen": 1532152, + "step": 5720 + }, + { + "epoch": 1.5003931847968546, + "eval_loss": 0.48007017374038696, + "eval_runtime": 18.204, + "eval_samples_per_second": 46.583, + "eval_steps_per_second": 23.292, + "num_input_tokens_seen": 1533144, + "step": 5724 + }, + { + "epoch": 1.5006553079947575, + "grad_norm": 1.493299126625061, + "learning_rate": 4.961959777446425e-05, + "loss": 0.5332, + "num_input_tokens_seen": 1533976, + "step": 5725 + }, + { + "epoch": 1.5019659239842726, + "grad_norm": 3.8490142822265625, + "learning_rate": 4.961760758904527e-05, + "loss": 0.6343, + "num_input_tokens_seen": 1534968, + "step": 5730 + }, + { + "epoch": 1.5032765399737875, + "grad_norm": 4.05137300491333, + "learning_rate": 4.961561225120356e-05, + "loss": 0.5493, + "num_input_tokens_seen": 1536200, + "step": 5735 + }, + { + "epoch": 1.5045871559633026, + "grad_norm": 1.6833194494247437, + "learning_rate": 4.9613611761356736e-05, + "loss": 0.4664, + "num_input_tokens_seen": 1537784, + "step": 5740 + }, + { + "epoch": 1.5058977719528177, + "grad_norm": 4.167136192321777, + "learning_rate": 4.96116061199235e-05, + "loss": 0.5262, + "num_input_tokens_seen": 1539192, + "step": 5745 + }, + { + "epoch": 1.5072083879423328, + "grad_norm": 5.153746604919434, + "learning_rate": 4.960959532732363e-05, + "loss": 0.6193, + "num_input_tokens_seen": 1540296, + "step": 5750 + }, + { + "epoch": 1.508519003931848, + "grad_norm": 2.0641517639160156, + "learning_rate": 4.9607579383977984e-05, + "loss": 0.3813, + "num_input_tokens_seen": 1541624, + "step": 5755 + }, + { + "epoch": 1.509829619921363, + "grad_norm": 1.6970394849777222, + "learning_rate": 4.9605558290308484e-05, + "loss": 0.488, + "num_input_tokens_seen": 1543128, + "step": 5760 + }, + { + "epoch": 1.511140235910878, + "grad_norm": 2.3446261882781982, + "learning_rate": 4.9603532046738154e-05, + "loss": 0.4228, + "num_input_tokens_seen": 1544440, + "step": 5765 + }, + { + "epoch": 1.5124508519003932, + "grad_norm": 2.219092845916748, + "learning_rate": 4.960150065369108e-05, + "loss": 0.3716, + "num_input_tokens_seen": 1545624, + "step": 5770 + }, + { + "epoch": 1.5137614678899083, + "grad_norm": 3.6507656574249268, + "learning_rate": 4.9599464111592417e-05, + "loss": 0.5654, + "num_input_tokens_seen": 1546696, + "step": 5775 + }, + { + "epoch": 1.5150720838794234, + "grad_norm": 2.3577306270599365, + "learning_rate": 4.959742242086843e-05, + "loss": 0.6025, + "num_input_tokens_seen": 1548088, + "step": 5780 + }, + { + "epoch": 1.5163826998689385, + "grad_norm": 2.9469153881073, + "learning_rate": 4.959537558194643e-05, + "loss": 0.4448, + "num_input_tokens_seen": 1549240, + "step": 5785 + }, + { + "epoch": 1.5176933158584536, + "grad_norm": 8.630414962768555, + "learning_rate": 4.959332359525482e-05, + "loss": 0.5724, + "num_input_tokens_seen": 1550600, + "step": 5790 + }, + { + "epoch": 1.5190039318479687, + "grad_norm": 3.1512510776519775, + "learning_rate": 4.9591266461223075e-05, + "loss": 0.4727, + "num_input_tokens_seen": 1552120, + "step": 5795 + }, + { + "epoch": 1.5203145478374838, + "grad_norm": 4.271605491638184, + "learning_rate": 4.958920418028175e-05, + "loss": 0.4841, + "num_input_tokens_seen": 1553224, + "step": 5800 + }, + { + "epoch": 1.5216251638269986, + "grad_norm": 2.929898500442505, + "learning_rate": 4.958713675286247e-05, + "loss": 0.3791, + "num_input_tokens_seen": 1554136, + "step": 5805 + }, + { + "epoch": 1.5229357798165137, + "grad_norm": 3.028090000152588, + "learning_rate": 4.9585064179397955e-05, + "loss": 0.3592, + "num_input_tokens_seen": 1556472, + "step": 5810 + }, + { + "epoch": 1.5242463958060288, + "grad_norm": 1.9691215753555298, + "learning_rate": 4.958298646032199e-05, + "loss": 0.4014, + "num_input_tokens_seen": 1557720, + "step": 5815 + }, + { + "epoch": 1.525557011795544, + "grad_norm": 2.369809627532959, + "learning_rate": 4.958090359606943e-05, + "loss": 0.5037, + "num_input_tokens_seen": 1558792, + "step": 5820 + }, + { + "epoch": 1.526867627785059, + "grad_norm": 4.09300422668457, + "learning_rate": 4.957881558707622e-05, + "loss": 0.4471, + "num_input_tokens_seen": 1559880, + "step": 5825 + }, + { + "epoch": 1.528178243774574, + "grad_norm": 8.85969352722168, + "learning_rate": 4.957672243377938e-05, + "loss": 0.5579, + "num_input_tokens_seen": 1561464, + "step": 5830 + }, + { + "epoch": 1.529488859764089, + "grad_norm": 2.072631359100342, + "learning_rate": 4.9574624136617e-05, + "loss": 0.6838, + "num_input_tokens_seen": 1562904, + "step": 5835 + }, + { + "epoch": 1.530799475753604, + "grad_norm": 2.881932258605957, + "learning_rate": 4.9572520696028245e-05, + "loss": 0.5006, + "num_input_tokens_seen": 1564360, + "step": 5840 + }, + { + "epoch": 1.5321100917431192, + "grad_norm": 1.8419080972671509, + "learning_rate": 4.9570412112453377e-05, + "loss": 0.4789, + "num_input_tokens_seen": 1565736, + "step": 5845 + }, + { + "epoch": 1.5334207077326343, + "grad_norm": 6.230313301086426, + "learning_rate": 4.95682983863337e-05, + "loss": 0.7665, + "num_input_tokens_seen": 1567080, + "step": 5850 + }, + { + "epoch": 1.5347313237221494, + "grad_norm": 3.3782801628112793, + "learning_rate": 4.9566179518111625e-05, + "loss": 0.5087, + "num_input_tokens_seen": 1568056, + "step": 5855 + }, + { + "epoch": 1.5360419397116645, + "grad_norm": 2.714571952819824, + "learning_rate": 4.9564055508230616e-05, + "loss": 0.4705, + "num_input_tokens_seen": 1569144, + "step": 5860 + }, + { + "epoch": 1.5373525557011796, + "grad_norm": 3.4294817447662354, + "learning_rate": 4.9561926357135235e-05, + "loss": 0.429, + "num_input_tokens_seen": 1570968, + "step": 5865 + }, + { + "epoch": 1.5386631716906947, + "grad_norm": 2.2576801776885986, + "learning_rate": 4.9559792065271116e-05, + "loss": 0.624, + "num_input_tokens_seen": 1572040, + "step": 5870 + }, + { + "epoch": 1.5399737876802098, + "grad_norm": 3.8934435844421387, + "learning_rate": 4.955765263308495e-05, + "loss": 0.374, + "num_input_tokens_seen": 1573128, + "step": 5875 + }, + { + "epoch": 1.5412844036697249, + "grad_norm": 2.296872854232788, + "learning_rate": 4.9555508061024514e-05, + "loss": 0.3408, + "num_input_tokens_seen": 1574552, + "step": 5880 + }, + { + "epoch": 1.54259501965924, + "grad_norm": 8.956052780151367, + "learning_rate": 4.955335834953867e-05, + "loss": 0.4069, + "num_input_tokens_seen": 1575576, + "step": 5885 + }, + { + "epoch": 1.543905635648755, + "grad_norm": 1.2642483711242676, + "learning_rate": 4.955120349907736e-05, + "loss": 0.4792, + "num_input_tokens_seen": 1577048, + "step": 5890 + }, + { + "epoch": 1.5452162516382701, + "grad_norm": 2.3852243423461914, + "learning_rate": 4.954904351009157e-05, + "loss": 0.5263, + "num_input_tokens_seen": 1578328, + "step": 5895 + }, + { + "epoch": 1.546526867627785, + "grad_norm": 1.3674627542495728, + "learning_rate": 4.954687838303339e-05, + "loss": 0.3514, + "num_input_tokens_seen": 1579992, + "step": 5900 + }, + { + "epoch": 1.5478374836173001, + "grad_norm": 1.7508867979049683, + "learning_rate": 4.9544708118355984e-05, + "loss": 0.5327, + "num_input_tokens_seen": 1581912, + "step": 5905 + }, + { + "epoch": 1.5491480996068152, + "grad_norm": 2.1193020343780518, + "learning_rate": 4.9542532716513576e-05, + "loss": 0.514, + "num_input_tokens_seen": 1583048, + "step": 5910 + }, + { + "epoch": 1.5504587155963303, + "grad_norm": 2.5414986610412598, + "learning_rate": 4.9540352177961466e-05, + "loss": 0.588, + "num_input_tokens_seen": 1584360, + "step": 5915 + }, + { + "epoch": 1.5517693315858454, + "grad_norm": 6.923000335693359, + "learning_rate": 4.953816650315606e-05, + "loss": 0.5353, + "num_input_tokens_seen": 1585496, + "step": 5920 + }, + { + "epoch": 1.5530799475753603, + "grad_norm": 5.527699947357178, + "learning_rate": 4.9535975692554805e-05, + "loss": 0.4781, + "num_input_tokens_seen": 1587704, + "step": 5925 + }, + { + "epoch": 1.5543905635648754, + "grad_norm": 2.924713134765625, + "learning_rate": 4.953377974661623e-05, + "loss": 0.5681, + "num_input_tokens_seen": 1588712, + "step": 5930 + }, + { + "epoch": 1.5557011795543905, + "grad_norm": 8.34090805053711, + "learning_rate": 4.953157866579994e-05, + "loss": 0.6743, + "num_input_tokens_seen": 1589960, + "step": 5935 + }, + { + "epoch": 1.5570117955439056, + "grad_norm": 4.873517036437988, + "learning_rate": 4.9529372450566624e-05, + "loss": 0.4602, + "num_input_tokens_seen": 1591544, + "step": 5940 + }, + { + "epoch": 1.5583224115334207, + "grad_norm": 1.495120644569397, + "learning_rate": 4.952716110137804e-05, + "loss": 0.3808, + "num_input_tokens_seen": 1593272, + "step": 5945 + }, + { + "epoch": 1.5596330275229358, + "grad_norm": 1.3717983961105347, + "learning_rate": 4.952494461869702e-05, + "loss": 0.4149, + "num_input_tokens_seen": 1594728, + "step": 5950 + }, + { + "epoch": 1.5609436435124509, + "grad_norm": 4.069150924682617, + "learning_rate": 4.952272300298747e-05, + "loss": 0.7711, + "num_input_tokens_seen": 1596312, + "step": 5955 + }, + { + "epoch": 1.562254259501966, + "grad_norm": 2.193009376525879, + "learning_rate": 4.952049625471436e-05, + "loss": 0.459, + "num_input_tokens_seen": 1597336, + "step": 5960 + }, + { + "epoch": 1.563564875491481, + "grad_norm": 1.5649629831314087, + "learning_rate": 4.951826437434375e-05, + "loss": 0.466, + "num_input_tokens_seen": 1598376, + "step": 5965 + }, + { + "epoch": 1.5648754914809961, + "grad_norm": 2.7942724227905273, + "learning_rate": 4.951602736234278e-05, + "loss": 0.4909, + "num_input_tokens_seen": 1599512, + "step": 5970 + }, + { + "epoch": 1.5661861074705112, + "grad_norm": 1.8230726718902588, + "learning_rate": 4.9513785219179634e-05, + "loss": 0.3714, + "num_input_tokens_seen": 1601000, + "step": 5975 + }, + { + "epoch": 1.5674967234600263, + "grad_norm": 5.410324573516846, + "learning_rate": 4.95115379453236e-05, + "loss": 0.563, + "num_input_tokens_seen": 1602248, + "step": 5980 + }, + { + "epoch": 1.5688073394495414, + "grad_norm": 2.5686323642730713, + "learning_rate": 4.950928554124503e-05, + "loss": 0.6358, + "num_input_tokens_seen": 1603816, + "step": 5985 + }, + { + "epoch": 1.5701179554390565, + "grad_norm": 5.137300491333008, + "learning_rate": 4.950702800741534e-05, + "loss": 0.6958, + "num_input_tokens_seen": 1604792, + "step": 5990 + }, + { + "epoch": 1.5714285714285714, + "grad_norm": 5.772006511688232, + "learning_rate": 4.950476534430703e-05, + "loss": 0.5678, + "num_input_tokens_seen": 1605736, + "step": 5995 + }, + { + "epoch": 1.5727391874180865, + "grad_norm": 2.503063678741455, + "learning_rate": 4.950249755239369e-05, + "loss": 0.3522, + "num_input_tokens_seen": 1607272, + "step": 6000 + }, + { + "epoch": 1.5740498034076016, + "grad_norm": 3.9764158725738525, + "learning_rate": 4.950022463214994e-05, + "loss": 0.5829, + "num_input_tokens_seen": 1608888, + "step": 6005 + }, + { + "epoch": 1.5753604193971167, + "grad_norm": 5.270267009735107, + "learning_rate": 4.949794658405151e-05, + "loss": 0.3868, + "num_input_tokens_seen": 1610344, + "step": 6010 + }, + { + "epoch": 1.5766710353866316, + "grad_norm": 2.412449359893799, + "learning_rate": 4.9495663408575185e-05, + "loss": 0.7301, + "num_input_tokens_seen": 1611736, + "step": 6015 + }, + { + "epoch": 1.5779816513761467, + "grad_norm": 2.258915662765503, + "learning_rate": 4.949337510619884e-05, + "loss": 0.3502, + "num_input_tokens_seen": 1613064, + "step": 6020 + }, + { + "epoch": 1.5792922673656618, + "grad_norm": 3.149928569793701, + "learning_rate": 4.94910816774014e-05, + "loss": 0.4219, + "num_input_tokens_seen": 1614328, + "step": 6025 + }, + { + "epoch": 1.5806028833551768, + "grad_norm": 4.962479591369629, + "learning_rate": 4.9488783122662886e-05, + "loss": 0.5039, + "num_input_tokens_seen": 1615368, + "step": 6030 + }, + { + "epoch": 1.581913499344692, + "grad_norm": 3.0633952617645264, + "learning_rate": 4.9486479442464374e-05, + "loss": 0.3815, + "num_input_tokens_seen": 1616744, + "step": 6035 + }, + { + "epoch": 1.583224115334207, + "grad_norm": 2.1740190982818604, + "learning_rate": 4.948417063728803e-05, + "loss": 0.4539, + "num_input_tokens_seen": 1617944, + "step": 6040 + }, + { + "epoch": 1.5845347313237221, + "grad_norm": 2.687119722366333, + "learning_rate": 4.948185670761707e-05, + "loss": 0.3628, + "num_input_tokens_seen": 1619208, + "step": 6045 + }, + { + "epoch": 1.5858453473132372, + "grad_norm": 2.987440586090088, + "learning_rate": 4.94795376539358e-05, + "loss": 0.3435, + "num_input_tokens_seen": 1620392, + "step": 6050 + }, + { + "epoch": 1.5871559633027523, + "grad_norm": 3.2457258701324463, + "learning_rate": 4.947721347672961e-05, + "loss": 0.6083, + "num_input_tokens_seen": 1621512, + "step": 6055 + }, + { + "epoch": 1.5884665792922674, + "grad_norm": 2.0165555477142334, + "learning_rate": 4.9474884176484924e-05, + "loss": 0.5707, + "num_input_tokens_seen": 1622664, + "step": 6060 + }, + { + "epoch": 1.5897771952817825, + "grad_norm": 2.0634353160858154, + "learning_rate": 4.947254975368927e-05, + "loss": 0.5609, + "num_input_tokens_seen": 1624184, + "step": 6065 + }, + { + "epoch": 1.5910878112712976, + "grad_norm": 1.5067812204360962, + "learning_rate": 4.947021020883124e-05, + "loss": 0.3915, + "num_input_tokens_seen": 1625640, + "step": 6070 + }, + { + "epoch": 1.5923984272608127, + "grad_norm": 3.641207218170166, + "learning_rate": 4.9467865542400494e-05, + "loss": 0.3808, + "num_input_tokens_seen": 1626904, + "step": 6075 + }, + { + "epoch": 1.5937090432503278, + "grad_norm": 2.315666913986206, + "learning_rate": 4.946551575488776e-05, + "loss": 0.5453, + "num_input_tokens_seen": 1628216, + "step": 6080 + }, + { + "epoch": 1.5950196592398427, + "grad_norm": 1.9076250791549683, + "learning_rate": 4.946316084678486e-05, + "loss": 0.514, + "num_input_tokens_seen": 1629432, + "step": 6085 + }, + { + "epoch": 1.5963302752293578, + "grad_norm": 2.880415439605713, + "learning_rate": 4.946080081858466e-05, + "loss": 0.4197, + "num_input_tokens_seen": 1630552, + "step": 6090 + }, + { + "epoch": 1.5976408912188729, + "grad_norm": 1.3849436044692993, + "learning_rate": 4.945843567078111e-05, + "loss": 0.6692, + "num_input_tokens_seen": 1632024, + "step": 6095 + }, + { + "epoch": 1.598951507208388, + "grad_norm": 2.6993408203125, + "learning_rate": 4.9456065403869246e-05, + "loss": 0.636, + "num_input_tokens_seen": 1633448, + "step": 6100 + }, + { + "epoch": 1.600262123197903, + "grad_norm": 5.274441242218018, + "learning_rate": 4.9453690018345144e-05, + "loss": 0.6284, + "num_input_tokens_seen": 1634488, + "step": 6105 + }, + { + "epoch": 1.601572739187418, + "grad_norm": 6.0873847007751465, + "learning_rate": 4.945130951470597e-05, + "loss": 0.6202, + "num_input_tokens_seen": 1635736, + "step": 6110 + }, + { + "epoch": 1.602883355176933, + "grad_norm": 2.6655383110046387, + "learning_rate": 4.944892389344997e-05, + "loss": 0.4953, + "num_input_tokens_seen": 1636872, + "step": 6115 + }, + { + "epoch": 1.6041939711664481, + "grad_norm": 2.498307466506958, + "learning_rate": 4.9446533155076444e-05, + "loss": 0.4639, + "num_input_tokens_seen": 1638408, + "step": 6120 + }, + { + "epoch": 1.6055045871559632, + "grad_norm": 2.6697962284088135, + "learning_rate": 4.9444137300085766e-05, + "loss": 0.4428, + "num_input_tokens_seen": 1640024, + "step": 6125 + }, + { + "epoch": 1.6068152031454783, + "grad_norm": 3.613124370574951, + "learning_rate": 4.944173632897938e-05, + "loss": 0.6278, + "num_input_tokens_seen": 1641192, + "step": 6130 + }, + { + "epoch": 1.6081258191349934, + "grad_norm": 1.983076572418213, + "learning_rate": 4.943933024225983e-05, + "loss": 0.5121, + "num_input_tokens_seen": 1642360, + "step": 6135 + }, + { + "epoch": 1.6094364351245085, + "grad_norm": 2.6314444541931152, + "learning_rate": 4.943691904043067e-05, + "loss": 0.3979, + "num_input_tokens_seen": 1643352, + "step": 6140 + }, + { + "epoch": 1.6107470511140236, + "grad_norm": 2.960883378982544, + "learning_rate": 4.943450272399659e-05, + "loss": 0.6016, + "num_input_tokens_seen": 1644392, + "step": 6145 + }, + { + "epoch": 1.6120576671035387, + "grad_norm": 5.203797340393066, + "learning_rate": 4.94320812934633e-05, + "loss": 0.2949, + "num_input_tokens_seen": 1645672, + "step": 6150 + }, + { + "epoch": 1.6133682830930538, + "grad_norm": 1.6112900972366333, + "learning_rate": 4.942965474933762e-05, + "loss": 0.4977, + "num_input_tokens_seen": 1647240, + "step": 6155 + }, + { + "epoch": 1.614678899082569, + "grad_norm": 6.319227695465088, + "learning_rate": 4.9427223092127406e-05, + "loss": 0.456, + "num_input_tokens_seen": 1648344, + "step": 6160 + }, + { + "epoch": 1.615989515072084, + "grad_norm": 3.1486258506774902, + "learning_rate": 4.942478632234161e-05, + "loss": 0.374, + "num_input_tokens_seen": 1649352, + "step": 6165 + }, + { + "epoch": 1.617300131061599, + "grad_norm": 3.5949764251708984, + "learning_rate": 4.942234444049024e-05, + "loss": 0.3529, + "num_input_tokens_seen": 1650600, + "step": 6170 + }, + { + "epoch": 1.6186107470511142, + "grad_norm": 3.989520311355591, + "learning_rate": 4.9419897447084375e-05, + "loss": 0.43, + "num_input_tokens_seen": 1652024, + "step": 6175 + }, + { + "epoch": 1.619921363040629, + "grad_norm": 2.390857219696045, + "learning_rate": 4.9417445342636166e-05, + "loss": 0.3307, + "num_input_tokens_seen": 1654008, + "step": 6180 + }, + { + "epoch": 1.6212319790301442, + "grad_norm": 6.197994232177734, + "learning_rate": 4.941498812765884e-05, + "loss": 0.3604, + "num_input_tokens_seen": 1655320, + "step": 6185 + }, + { + "epoch": 1.6225425950196593, + "grad_norm": 3.9531235694885254, + "learning_rate": 4.941252580266668e-05, + "loss": 0.4179, + "num_input_tokens_seen": 1656408, + "step": 6190 + }, + { + "epoch": 1.6238532110091743, + "grad_norm": 1.476270318031311, + "learning_rate": 4.941005836817506e-05, + "loss": 0.5025, + "num_input_tokens_seen": 1658072, + "step": 6195 + }, + { + "epoch": 1.6251638269986892, + "grad_norm": 2.655658721923828, + "learning_rate": 4.940758582470039e-05, + "loss": 0.3591, + "num_input_tokens_seen": 1659112, + "step": 6200 + }, + { + "epoch": 1.6264744429882043, + "grad_norm": 5.251587390899658, + "learning_rate": 4.940510817276019e-05, + "loss": 0.3961, + "num_input_tokens_seen": 1660152, + "step": 6205 + }, + { + "epoch": 1.6277850589777194, + "grad_norm": 13.969517707824707, + "learning_rate": 4.940262541287302e-05, + "loss": 0.3241, + "num_input_tokens_seen": 1661272, + "step": 6210 + }, + { + "epoch": 1.6290956749672345, + "grad_norm": 2.22263765335083, + "learning_rate": 4.94001375455585e-05, + "loss": 0.3595, + "num_input_tokens_seen": 1662632, + "step": 6215 + }, + { + "epoch": 1.6304062909567496, + "grad_norm": 7.058943748474121, + "learning_rate": 4.939764457133737e-05, + "loss": 0.4951, + "num_input_tokens_seen": 1664056, + "step": 6220 + }, + { + "epoch": 1.6317169069462647, + "grad_norm": 1.7457275390625, + "learning_rate": 4.939514649073137e-05, + "loss": 0.4644, + "num_input_tokens_seen": 1665048, + "step": 6225 + }, + { + "epoch": 1.6330275229357798, + "grad_norm": 4.504209518432617, + "learning_rate": 4.939264330426337e-05, + "loss": 0.3976, + "num_input_tokens_seen": 1666456, + "step": 6230 + }, + { + "epoch": 1.634338138925295, + "grad_norm": 3.9998488426208496, + "learning_rate": 4.9390135012457275e-05, + "loss": 0.4766, + "num_input_tokens_seen": 1667560, + "step": 6235 + }, + { + "epoch": 1.63564875491481, + "grad_norm": 6.640893459320068, + "learning_rate": 4.938762161583807e-05, + "loss": 0.7312, + "num_input_tokens_seen": 1668584, + "step": 6240 + }, + { + "epoch": 1.636959370904325, + "grad_norm": 1.703174352645874, + "learning_rate": 4.9385103114931784e-05, + "loss": 0.4833, + "num_input_tokens_seen": 1669912, + "step": 6245 + }, + { + "epoch": 1.6382699868938402, + "grad_norm": 1.798006534576416, + "learning_rate": 4.9382579510265556e-05, + "loss": 0.6128, + "num_input_tokens_seen": 1671736, + "step": 6250 + }, + { + "epoch": 1.6395806028833553, + "grad_norm": 1.1580629348754883, + "learning_rate": 4.9380050802367573e-05, + "loss": 0.3483, + "num_input_tokens_seen": 1672936, + "step": 6255 + }, + { + "epoch": 1.6408912188728704, + "grad_norm": 4.12657356262207, + "learning_rate": 4.937751699176708e-05, + "loss": 0.6406, + "num_input_tokens_seen": 1674152, + "step": 6260 + }, + { + "epoch": 1.6422018348623855, + "grad_norm": 1.9138715267181396, + "learning_rate": 4.93749780789944e-05, + "loss": 0.4603, + "num_input_tokens_seen": 1675496, + "step": 6265 + }, + { + "epoch": 1.6435124508519003, + "grad_norm": 16.26593017578125, + "learning_rate": 4.937243406458093e-05, + "loss": 0.5115, + "num_input_tokens_seen": 1676264, + "step": 6270 + }, + { + "epoch": 1.6448230668414154, + "grad_norm": 1.1129025220870972, + "learning_rate": 4.9369884949059115e-05, + "loss": 0.8237, + "num_input_tokens_seen": 1677624, + "step": 6275 + }, + { + "epoch": 1.6461336828309305, + "grad_norm": 2.2420899868011475, + "learning_rate": 4.936733073296249e-05, + "loss": 0.4905, + "num_input_tokens_seen": 1678840, + "step": 6280 + }, + { + "epoch": 1.6474442988204456, + "grad_norm": 2.968702554702759, + "learning_rate": 4.936477141682565e-05, + "loss": 0.563, + "num_input_tokens_seen": 1679928, + "step": 6285 + }, + { + "epoch": 1.6487549148099607, + "grad_norm": 15.791129112243652, + "learning_rate": 4.9362207001184246e-05, + "loss": 1.4148, + "num_input_tokens_seen": 1681320, + "step": 6290 + }, + { + "epoch": 1.6500655307994756, + "grad_norm": 9.713336944580078, + "learning_rate": 4.9359637486575016e-05, + "loss": 0.4715, + "num_input_tokens_seen": 1682696, + "step": 6295 + }, + { + "epoch": 1.6513761467889907, + "grad_norm": 6.91204833984375, + "learning_rate": 4.9357062873535746e-05, + "loss": 0.4525, + "num_input_tokens_seen": 1683864, + "step": 6300 + }, + { + "epoch": 1.6526867627785058, + "grad_norm": 4.408916473388672, + "learning_rate": 4.9354483162605305e-05, + "loss": 0.5871, + "num_input_tokens_seen": 1684904, + "step": 6305 + }, + { + "epoch": 1.6539973787680209, + "grad_norm": 2.420609951019287, + "learning_rate": 4.9351898354323625e-05, + "loss": 0.5871, + "num_input_tokens_seen": 1685976, + "step": 6310 + }, + { + "epoch": 1.655307994757536, + "grad_norm": 3.030914545059204, + "learning_rate": 4.934930844923168e-05, + "loss": 0.4884, + "num_input_tokens_seen": 1687496, + "step": 6315 + }, + { + "epoch": 1.656618610747051, + "grad_norm": 2.6351616382598877, + "learning_rate": 4.934671344787156e-05, + "loss": 0.6086, + "num_input_tokens_seen": 1688840, + "step": 6320 + }, + { + "epoch": 1.6579292267365662, + "grad_norm": 2.595705032348633, + "learning_rate": 4.9344113350786386e-05, + "loss": 0.3991, + "num_input_tokens_seen": 1690232, + "step": 6325 + }, + { + "epoch": 1.6592398427260813, + "grad_norm": 3.2952334880828857, + "learning_rate": 4.934150815852034e-05, + "loss": 0.5094, + "num_input_tokens_seen": 1691576, + "step": 6330 + }, + { + "epoch": 1.6605504587155964, + "grad_norm": 4.759393215179443, + "learning_rate": 4.933889787161871e-05, + "loss": 0.5664, + "num_input_tokens_seen": 1692648, + "step": 6335 + }, + { + "epoch": 1.6618610747051115, + "grad_norm": 2.0619521141052246, + "learning_rate": 4.9336282490627803e-05, + "loss": 0.7265, + "num_input_tokens_seen": 1694168, + "step": 6340 + }, + { + "epoch": 1.6631716906946266, + "grad_norm": 3.538313627243042, + "learning_rate": 4.9333662016095016e-05, + "loss": 0.5201, + "num_input_tokens_seen": 1695800, + "step": 6345 + }, + { + "epoch": 1.6644823066841417, + "grad_norm": 2.524843215942383, + "learning_rate": 4.9331036448568826e-05, + "loss": 0.3329, + "num_input_tokens_seen": 1696904, + "step": 6350 + }, + { + "epoch": 1.6657929226736568, + "grad_norm": 8.02358627319336, + "learning_rate": 4.932840578859873e-05, + "loss": 0.5561, + "num_input_tokens_seen": 1698232, + "step": 6355 + }, + { + "epoch": 1.6671035386631718, + "grad_norm": 4.144388675689697, + "learning_rate": 4.932577003673535e-05, + "loss": 0.4196, + "num_input_tokens_seen": 1699320, + "step": 6360 + }, + { + "epoch": 1.6684141546526867, + "grad_norm": 3.936276435852051, + "learning_rate": 4.932312919353034e-05, + "loss": 0.5408, + "num_input_tokens_seen": 1700568, + "step": 6365 + }, + { + "epoch": 1.6697247706422018, + "grad_norm": 1.305861473083496, + "learning_rate": 4.932048325953641e-05, + "loss": 0.7206, + "num_input_tokens_seen": 1702216, + "step": 6370 + }, + { + "epoch": 1.671035386631717, + "grad_norm": 2.296505928039551, + "learning_rate": 4.931783223530736e-05, + "loss": 0.4056, + "num_input_tokens_seen": 1703672, + "step": 6375 + }, + { + "epoch": 1.672346002621232, + "grad_norm": 1.2103568315505981, + "learning_rate": 4.9315176121398035e-05, + "loss": 0.4445, + "num_input_tokens_seen": 1705352, + "step": 6380 + }, + { + "epoch": 1.673656618610747, + "grad_norm": 3.048358917236328, + "learning_rate": 4.931251491836436e-05, + "loss": 0.4469, + "num_input_tokens_seen": 1706456, + "step": 6385 + }, + { + "epoch": 1.674967234600262, + "grad_norm": 2.4783008098602295, + "learning_rate": 4.930984862676332e-05, + "loss": 0.408, + "num_input_tokens_seen": 1707864, + "step": 6390 + }, + { + "epoch": 1.676277850589777, + "grad_norm": 10.277024269104004, + "learning_rate": 4.930717724715298e-05, + "loss": 0.464, + "num_input_tokens_seen": 1709096, + "step": 6395 + }, + { + "epoch": 1.6775884665792922, + "grad_norm": 1.4275455474853516, + "learning_rate": 4.930450078009243e-05, + "loss": 0.4427, + "num_input_tokens_seen": 1710504, + "step": 6400 + }, + { + "epoch": 1.6788990825688073, + "grad_norm": 1.6900349855422974, + "learning_rate": 4.930181922614186e-05, + "loss": 0.46, + "num_input_tokens_seen": 1711832, + "step": 6405 + }, + { + "epoch": 1.6802096985583224, + "grad_norm": 1.8512250185012817, + "learning_rate": 4.929913258586252e-05, + "loss": 0.3666, + "num_input_tokens_seen": 1713128, + "step": 6410 + }, + { + "epoch": 1.6815203145478375, + "grad_norm": 8.070503234863281, + "learning_rate": 4.9296440859816715e-05, + "loss": 0.5291, + "num_input_tokens_seen": 1714616, + "step": 6415 + }, + { + "epoch": 1.6828309305373526, + "grad_norm": 1.0526713132858276, + "learning_rate": 4.9293744048567825e-05, + "loss": 0.5773, + "num_input_tokens_seen": 1715928, + "step": 6420 + }, + { + "epoch": 1.6841415465268676, + "grad_norm": 2.640137195587158, + "learning_rate": 4.929104215268027e-05, + "loss": 0.7976, + "num_input_tokens_seen": 1717016, + "step": 6425 + }, + { + "epoch": 1.6854521625163827, + "grad_norm": 1.6156574487686157, + "learning_rate": 4.9288335172719575e-05, + "loss": 0.4302, + "num_input_tokens_seen": 1718392, + "step": 6430 + }, + { + "epoch": 1.6867627785058978, + "grad_norm": 2.5132789611816406, + "learning_rate": 4.9285623109252286e-05, + "loss": 0.5806, + "num_input_tokens_seen": 1719752, + "step": 6435 + }, + { + "epoch": 1.688073394495413, + "grad_norm": 7.134504318237305, + "learning_rate": 4.928290596284606e-05, + "loss": 0.4978, + "num_input_tokens_seen": 1721272, + "step": 6440 + }, + { + "epoch": 1.689384010484928, + "grad_norm": 1.9776498079299927, + "learning_rate": 4.928018373406955e-05, + "loss": 0.3566, + "num_input_tokens_seen": 1722552, + "step": 6445 + }, + { + "epoch": 1.6906946264744431, + "grad_norm": 6.44937801361084, + "learning_rate": 4.927745642349255e-05, + "loss": 0.4687, + "num_input_tokens_seen": 1723784, + "step": 6450 + }, + { + "epoch": 1.6920052424639582, + "grad_norm": 3.944725513458252, + "learning_rate": 4.927472403168587e-05, + "loss": 0.4538, + "num_input_tokens_seen": 1724936, + "step": 6455 + }, + { + "epoch": 1.693315858453473, + "grad_norm": 2.463048219680786, + "learning_rate": 4.92719865592214e-05, + "loss": 0.5476, + "num_input_tokens_seen": 1726664, + "step": 6460 + }, + { + "epoch": 1.6946264744429882, + "grad_norm": 1.7198262214660645, + "learning_rate": 4.9269244006672065e-05, + "loss": 0.3802, + "num_input_tokens_seen": 1728184, + "step": 6465 + }, + { + "epoch": 1.6959370904325033, + "grad_norm": 5.220691204071045, + "learning_rate": 4.9266496374611904e-05, + "loss": 0.6603, + "num_input_tokens_seen": 1729384, + "step": 6470 + }, + { + "epoch": 1.6972477064220184, + "grad_norm": 14.116911888122559, + "learning_rate": 4.926374366361598e-05, + "loss": 0.6446, + "num_input_tokens_seen": 1730904, + "step": 6475 + }, + { + "epoch": 1.6985583224115333, + "grad_norm": 4.879854679107666, + "learning_rate": 4.926098587426043e-05, + "loss": 0.5758, + "num_input_tokens_seen": 1732312, + "step": 6480 + }, + { + "epoch": 1.6998689384010484, + "grad_norm": 2.6405327320098877, + "learning_rate": 4.9258223007122454e-05, + "loss": 0.4405, + "num_input_tokens_seen": 1733288, + "step": 6485 + }, + { + "epoch": 1.7011795543905635, + "grad_norm": 1.159731388092041, + "learning_rate": 4.925545506278032e-05, + "loss": 0.6888, + "num_input_tokens_seen": 1736248, + "step": 6490 + }, + { + "epoch": 1.7024901703800785, + "grad_norm": 1.2968300580978394, + "learning_rate": 4.9252682041813356e-05, + "loss": 0.3405, + "num_input_tokens_seen": 1737320, + "step": 6495 + }, + { + "epoch": 1.7038007863695936, + "grad_norm": 2.142031669616699, + "learning_rate": 4.924990394480195e-05, + "loss": 0.6197, + "num_input_tokens_seen": 1738536, + "step": 6500 + }, + { + "epoch": 1.7051114023591087, + "grad_norm": 7.346621036529541, + "learning_rate": 4.9247120772327534e-05, + "loss": 0.4434, + "num_input_tokens_seen": 1739992, + "step": 6505 + }, + { + "epoch": 1.7064220183486238, + "grad_norm": 6.031692028045654, + "learning_rate": 4.924433252497264e-05, + "loss": 0.5056, + "num_input_tokens_seen": 1741608, + "step": 6510 + }, + { + "epoch": 1.707732634338139, + "grad_norm": 6.352177619934082, + "learning_rate": 4.924153920332084e-05, + "loss": 0.4149, + "num_input_tokens_seen": 1743448, + "step": 6515 + }, + { + "epoch": 1.709043250327654, + "grad_norm": 3.0235090255737305, + "learning_rate": 4.923874080795677e-05, + "loss": 0.3759, + "num_input_tokens_seen": 1744600, + "step": 6520 + }, + { + "epoch": 1.7103538663171691, + "grad_norm": 2.1900205612182617, + "learning_rate": 4.923593733946614e-05, + "loss": 0.6532, + "num_input_tokens_seen": 1745848, + "step": 6525 + }, + { + "epoch": 1.7116644823066842, + "grad_norm": 1.2958263158798218, + "learning_rate": 4.92331287984357e-05, + "loss": 0.5189, + "num_input_tokens_seen": 1747320, + "step": 6530 + }, + { + "epoch": 1.7129750982961993, + "grad_norm": 2.1008410453796387, + "learning_rate": 4.9230315185453266e-05, + "loss": 0.5267, + "num_input_tokens_seen": 1748328, + "step": 6535 + }, + { + "epoch": 1.7142857142857144, + "grad_norm": 3.021622657775879, + "learning_rate": 4.922749650110773e-05, + "loss": 0.5447, + "num_input_tokens_seen": 1749304, + "step": 6540 + }, + { + "epoch": 1.7155963302752295, + "grad_norm": 2.6895906925201416, + "learning_rate": 4.9224672745989045e-05, + "loss": 0.5449, + "num_input_tokens_seen": 1750744, + "step": 6545 + }, + { + "epoch": 1.7169069462647444, + "grad_norm": 2.722075939178467, + "learning_rate": 4.922184392068821e-05, + "loss": 0.5331, + "num_input_tokens_seen": 1752024, + "step": 6550 + }, + { + "epoch": 1.7182175622542595, + "grad_norm": 7.785636901855469, + "learning_rate": 4.921901002579729e-05, + "loss": 0.3854, + "num_input_tokens_seen": 1753032, + "step": 6555 + }, + { + "epoch": 1.7195281782437746, + "grad_norm": 5.487945556640625, + "learning_rate": 4.921617106190942e-05, + "loss": 0.459, + "num_input_tokens_seen": 1754136, + "step": 6560 + }, + { + "epoch": 1.7208387942332897, + "grad_norm": 3.8207056522369385, + "learning_rate": 4.9213327029618796e-05, + "loss": 0.4084, + "num_input_tokens_seen": 1755384, + "step": 6565 + }, + { + "epoch": 1.7221494102228048, + "grad_norm": 3.039620876312256, + "learning_rate": 4.921047792952066e-05, + "loss": 0.4044, + "num_input_tokens_seen": 1756456, + "step": 6570 + }, + { + "epoch": 1.7234600262123196, + "grad_norm": 1.1312164068222046, + "learning_rate": 4.920762376221132e-05, + "loss": 0.5311, + "num_input_tokens_seen": 1758104, + "step": 6575 + }, + { + "epoch": 1.7247706422018347, + "grad_norm": 2.236459493637085, + "learning_rate": 4.920476452828815e-05, + "loss": 0.5168, + "num_input_tokens_seen": 1759896, + "step": 6580 + }, + { + "epoch": 1.7260812581913498, + "grad_norm": 2.6559243202209473, + "learning_rate": 4.9201900228349594e-05, + "loss": 0.8161, + "num_input_tokens_seen": 1761032, + "step": 6585 + }, + { + "epoch": 1.727391874180865, + "grad_norm": 7.131217956542969, + "learning_rate": 4.9199030862995144e-05, + "loss": 0.4305, + "num_input_tokens_seen": 1763000, + "step": 6590 + }, + { + "epoch": 1.72870249017038, + "grad_norm": 2.1330316066741943, + "learning_rate": 4.919615643282533e-05, + "loss": 0.5488, + "num_input_tokens_seen": 1764200, + "step": 6595 + }, + { + "epoch": 1.7300131061598951, + "grad_norm": 2.996046543121338, + "learning_rate": 4.91932769384418e-05, + "loss": 0.7667, + "num_input_tokens_seen": 1765272, + "step": 6600 + }, + { + "epoch": 1.7313237221494102, + "grad_norm": 1.122695803642273, + "learning_rate": 4.91903923804472e-05, + "loss": 0.7295, + "num_input_tokens_seen": 1766776, + "step": 6605 + }, + { + "epoch": 1.7326343381389253, + "grad_norm": 1.2124378681182861, + "learning_rate": 4.918750275944528e-05, + "loss": 0.3102, + "num_input_tokens_seen": 1767960, + "step": 6610 + }, + { + "epoch": 1.7339449541284404, + "grad_norm": 2.1634061336517334, + "learning_rate": 4.9184608076040816e-05, + "loss": 0.5825, + "num_input_tokens_seen": 1769400, + "step": 6615 + }, + { + "epoch": 1.7352555701179555, + "grad_norm": 5.406585216522217, + "learning_rate": 4.918170833083967e-05, + "loss": 0.5561, + "num_input_tokens_seen": 1770536, + "step": 6620 + }, + { + "epoch": 1.7365661861074706, + "grad_norm": 1.8940449953079224, + "learning_rate": 4.9178803524448756e-05, + "loss": 0.6606, + "num_input_tokens_seen": 1771896, + "step": 6625 + }, + { + "epoch": 1.7378768020969857, + "grad_norm": 2.909837245941162, + "learning_rate": 4.9175893657476046e-05, + "loss": 0.3446, + "num_input_tokens_seen": 1772904, + "step": 6630 + }, + { + "epoch": 1.7391874180865008, + "grad_norm": 2.070648431777954, + "learning_rate": 4.917297873053056e-05, + "loss": 0.4607, + "num_input_tokens_seen": 1774552, + "step": 6635 + }, + { + "epoch": 1.7404980340760159, + "grad_norm": 2.8006279468536377, + "learning_rate": 4.91700587442224e-05, + "loss": 0.4355, + "num_input_tokens_seen": 1776296, + "step": 6640 + }, + { + "epoch": 1.7418086500655308, + "grad_norm": 15.331221580505371, + "learning_rate": 4.91671336991627e-05, + "loss": 0.5245, + "num_input_tokens_seen": 1777640, + "step": 6645 + }, + { + "epoch": 1.7431192660550459, + "grad_norm": 3.290631055831909, + "learning_rate": 4.916420359596368e-05, + "loss": 0.379, + "num_input_tokens_seen": 1778936, + "step": 6650 + }, + { + "epoch": 1.744429882044561, + "grad_norm": 2.509050130844116, + "learning_rate": 4.91612684352386e-05, + "loss": 0.5313, + "num_input_tokens_seen": 1780024, + "step": 6655 + }, + { + "epoch": 1.745740498034076, + "grad_norm": 4.240983486175537, + "learning_rate": 4.915832821760178e-05, + "loss": 0.4678, + "num_input_tokens_seen": 1781272, + "step": 6660 + }, + { + "epoch": 1.747051114023591, + "grad_norm": 7.283449172973633, + "learning_rate": 4.915538294366861e-05, + "loss": 0.4301, + "num_input_tokens_seen": 1782504, + "step": 6665 + }, + { + "epoch": 1.748361730013106, + "grad_norm": 8.188507080078125, + "learning_rate": 4.915243261405553e-05, + "loss": 0.6286, + "num_input_tokens_seen": 1783576, + "step": 6670 + }, + { + "epoch": 1.7496723460026211, + "grad_norm": 3.106013774871826, + "learning_rate": 4.9149477229380023e-05, + "loss": 0.6217, + "num_input_tokens_seen": 1784424, + "step": 6675 + }, + { + "epoch": 1.7509829619921362, + "grad_norm": 1.9573625326156616, + "learning_rate": 4.914651679026068e-05, + "loss": 0.5377, + "num_input_tokens_seen": 1785992, + "step": 6680 + }, + { + "epoch": 1.7522935779816513, + "grad_norm": 1.1211843490600586, + "learning_rate": 4.9143551297317076e-05, + "loss": 0.3004, + "num_input_tokens_seen": 1787272, + "step": 6685 + }, + { + "epoch": 1.7536041939711664, + "grad_norm": 3.616795539855957, + "learning_rate": 4.9140580751169906e-05, + "loss": 0.6162, + "num_input_tokens_seen": 1788664, + "step": 6690 + }, + { + "epoch": 1.7549148099606815, + "grad_norm": 4.107919216156006, + "learning_rate": 4.91376051524409e-05, + "loss": 0.3835, + "num_input_tokens_seen": 1789864, + "step": 6695 + }, + { + "epoch": 1.7562254259501966, + "grad_norm": 3.220630645751953, + "learning_rate": 4.913462450175285e-05, + "loss": 0.3842, + "num_input_tokens_seen": 1791352, + "step": 6700 + }, + { + "epoch": 1.7575360419397117, + "grad_norm": 1.11677885055542, + "learning_rate": 4.913163879972957e-05, + "loss": 0.6417, + "num_input_tokens_seen": 1792920, + "step": 6705 + }, + { + "epoch": 1.7588466579292268, + "grad_norm": 5.294989109039307, + "learning_rate": 4.9128648046996004e-05, + "loss": 0.3792, + "num_input_tokens_seen": 1794040, + "step": 6710 + }, + { + "epoch": 1.7601572739187419, + "grad_norm": 8.93106746673584, + "learning_rate": 4.912565224417809e-05, + "loss": 0.4478, + "num_input_tokens_seen": 1795224, + "step": 6715 + }, + { + "epoch": 1.761467889908257, + "grad_norm": 1.2902204990386963, + "learning_rate": 4.9122651391902855e-05, + "loss": 0.6725, + "num_input_tokens_seen": 1796744, + "step": 6720 + }, + { + "epoch": 1.762778505897772, + "grad_norm": 2.2522964477539062, + "learning_rate": 4.9119645490798347e-05, + "loss": 0.5135, + "num_input_tokens_seen": 1797896, + "step": 6725 + }, + { + "epoch": 1.7640891218872872, + "grad_norm": 1.1229990720748901, + "learning_rate": 4.911663454149372e-05, + "loss": 0.4129, + "num_input_tokens_seen": 1799448, + "step": 6730 + }, + { + "epoch": 1.765399737876802, + "grad_norm": 14.292961120605469, + "learning_rate": 4.911361854461916e-05, + "loss": 1.0516, + "num_input_tokens_seen": 1800584, + "step": 6735 + }, + { + "epoch": 1.7667103538663171, + "grad_norm": 3.7369561195373535, + "learning_rate": 4.91105975008059e-05, + "loss": 0.3897, + "num_input_tokens_seen": 1801960, + "step": 6740 + }, + { + "epoch": 1.7680209698558322, + "grad_norm": 2.73284649848938, + "learning_rate": 4.910757141068624e-05, + "loss": 0.446, + "num_input_tokens_seen": 1803336, + "step": 6745 + }, + { + "epoch": 1.7693315858453473, + "grad_norm": 6.061513900756836, + "learning_rate": 4.910454027489354e-05, + "loss": 0.3488, + "num_input_tokens_seen": 1804840, + "step": 6750 + }, + { + "epoch": 1.7706422018348624, + "grad_norm": 2.2962350845336914, + "learning_rate": 4.910150409406221e-05, + "loss": 0.537, + "num_input_tokens_seen": 1806200, + "step": 6755 + }, + { + "epoch": 1.7719528178243773, + "grad_norm": 2.3754806518554688, + "learning_rate": 4.909846286882772e-05, + "loss": 0.5367, + "num_input_tokens_seen": 1807656, + "step": 6760 + }, + { + "epoch": 1.7732634338138924, + "grad_norm": 3.6660516262054443, + "learning_rate": 4.909541659982659e-05, + "loss": 0.4715, + "num_input_tokens_seen": 1808792, + "step": 6765 + }, + { + "epoch": 1.7745740498034075, + "grad_norm": 3.890929937362671, + "learning_rate": 4.9092365287696396e-05, + "loss": 0.4168, + "num_input_tokens_seen": 1809752, + "step": 6770 + }, + { + "epoch": 1.7758846657929226, + "grad_norm": 3.543004035949707, + "learning_rate": 4.908930893307578e-05, + "loss": 0.5536, + "num_input_tokens_seen": 1810968, + "step": 6775 + }, + { + "epoch": 1.7771952817824377, + "grad_norm": 2.6885716915130615, + "learning_rate": 4.9086247536604425e-05, + "loss": 0.3142, + "num_input_tokens_seen": 1812488, + "step": 6780 + }, + { + "epoch": 1.7785058977719528, + "grad_norm": 1.7786428928375244, + "learning_rate": 4.908318109892308e-05, + "loss": 0.4124, + "num_input_tokens_seen": 1813880, + "step": 6785 + }, + { + "epoch": 1.7798165137614679, + "grad_norm": 1.0324296951293945, + "learning_rate": 4.9080109620673546e-05, + "loss": 0.4244, + "num_input_tokens_seen": 1815752, + "step": 6790 + }, + { + "epoch": 1.781127129750983, + "grad_norm": 3.7049708366394043, + "learning_rate": 4.9077033102498666e-05, + "loss": 0.6574, + "num_input_tokens_seen": 1817336, + "step": 6795 + }, + { + "epoch": 1.782437745740498, + "grad_norm": 2.3073012828826904, + "learning_rate": 4.907395154504236e-05, + "loss": 0.4439, + "num_input_tokens_seen": 1819096, + "step": 6800 + }, + { + "epoch": 1.7837483617300132, + "grad_norm": 2.3527863025665283, + "learning_rate": 4.9070864948949604e-05, + "loss": 0.5093, + "num_input_tokens_seen": 1821032, + "step": 6805 + }, + { + "epoch": 1.7850589777195283, + "grad_norm": 1.4147614240646362, + "learning_rate": 4.9067773314866396e-05, + "loss": 0.6695, + "num_input_tokens_seen": 1822280, + "step": 6810 + }, + { + "epoch": 1.7863695937090434, + "grad_norm": 1.9426976442337036, + "learning_rate": 4.906467664343982e-05, + "loss": 0.3724, + "num_input_tokens_seen": 1823592, + "step": 6815 + }, + { + "epoch": 1.7876802096985585, + "grad_norm": 3.616804361343384, + "learning_rate": 4.9061574935318006e-05, + "loss": 0.484, + "num_input_tokens_seen": 1825304, + "step": 6820 + }, + { + "epoch": 1.7889908256880735, + "grad_norm": 2.448251247406006, + "learning_rate": 4.905846819115013e-05, + "loss": 0.4513, + "num_input_tokens_seen": 1826776, + "step": 6825 + }, + { + "epoch": 1.7903014416775884, + "grad_norm": 2.659686326980591, + "learning_rate": 4.905535641158644e-05, + "loss": 0.3172, + "num_input_tokens_seen": 1828104, + "step": 6830 + }, + { + "epoch": 1.7916120576671035, + "grad_norm": 2.5670621395111084, + "learning_rate": 4.9052239597278206e-05, + "loss": 0.5095, + "num_input_tokens_seen": 1829512, + "step": 6835 + }, + { + "epoch": 1.7929226736566186, + "grad_norm": 1.7700787782669067, + "learning_rate": 4.904911774887779e-05, + "loss": 0.6438, + "num_input_tokens_seen": 1831000, + "step": 6840 + }, + { + "epoch": 1.7942332896461337, + "grad_norm": 4.561263084411621, + "learning_rate": 4.904599086703858e-05, + "loss": 0.4201, + "num_input_tokens_seen": 1832120, + "step": 6845 + }, + { + "epoch": 1.7955439056356488, + "grad_norm": 9.126361846923828, + "learning_rate": 4.904285895241503e-05, + "loss": 0.5442, + "num_input_tokens_seen": 1833880, + "step": 6850 + }, + { + "epoch": 1.7968545216251637, + "grad_norm": 4.329135417938232, + "learning_rate": 4.903972200566265e-05, + "loss": 0.4072, + "num_input_tokens_seen": 1835208, + "step": 6855 + }, + { + "epoch": 1.7981651376146788, + "grad_norm": 5.871296405792236, + "learning_rate": 4.903658002743798e-05, + "loss": 0.4519, + "num_input_tokens_seen": 1836728, + "step": 6860 + }, + { + "epoch": 1.7994757536041939, + "grad_norm": 8.517304420471191, + "learning_rate": 4.9033433018398654e-05, + "loss": 0.5636, + "num_input_tokens_seen": 1838216, + "step": 6865 + }, + { + "epoch": 1.800786369593709, + "grad_norm": 3.7503979206085205, + "learning_rate": 4.9030280979203314e-05, + "loss": 0.5182, + "num_input_tokens_seen": 1839976, + "step": 6870 + }, + { + "epoch": 1.802096985583224, + "grad_norm": 1.6646521091461182, + "learning_rate": 4.90271239105117e-05, + "loss": 0.6454, + "num_input_tokens_seen": 1841096, + "step": 6875 + }, + { + "epoch": 1.8034076015727392, + "grad_norm": 2.4411728382110596, + "learning_rate": 4.902396181298455e-05, + "loss": 0.8629, + "num_input_tokens_seen": 1842376, + "step": 6880 + }, + { + "epoch": 1.8047182175622543, + "grad_norm": 4.131200790405273, + "learning_rate": 4.902079468728371e-05, + "loss": 0.341, + "num_input_tokens_seen": 1843432, + "step": 6885 + }, + { + "epoch": 1.8060288335517694, + "grad_norm": 5.555706977844238, + "learning_rate": 4.901762253407205e-05, + "loss": 0.7188, + "num_input_tokens_seen": 1844632, + "step": 6890 + }, + { + "epoch": 1.8073394495412844, + "grad_norm": 3.899718999862671, + "learning_rate": 4.9014445354013496e-05, + "loss": 0.6385, + "num_input_tokens_seen": 1845848, + "step": 6895 + }, + { + "epoch": 1.8086500655307995, + "grad_norm": 2.4952287673950195, + "learning_rate": 4.901126314777301e-05, + "loss": 0.4837, + "num_input_tokens_seen": 1847112, + "step": 6900 + }, + { + "epoch": 1.8099606815203146, + "grad_norm": 2.238776206970215, + "learning_rate": 4.9008075916016646e-05, + "loss": 0.8558, + "num_input_tokens_seen": 1848184, + "step": 6905 + }, + { + "epoch": 1.8112712975098297, + "grad_norm": 1.1998934745788574, + "learning_rate": 4.900488365941147e-05, + "loss": 0.5736, + "num_input_tokens_seen": 1849528, + "step": 6910 + }, + { + "epoch": 1.8125819134993448, + "grad_norm": 0.8598081469535828, + "learning_rate": 4.900168637862563e-05, + "loss": 0.4111, + "num_input_tokens_seen": 1850520, + "step": 6915 + }, + { + "epoch": 1.81389252948886, + "grad_norm": 1.9939528703689575, + "learning_rate": 4.899848407432829e-05, + "loss": 0.3511, + "num_input_tokens_seen": 1851496, + "step": 6920 + }, + { + "epoch": 1.8152031454783748, + "grad_norm": 2.6129584312438965, + "learning_rate": 4.8995276747189714e-05, + "loss": 0.5892, + "num_input_tokens_seen": 1853000, + "step": 6925 + }, + { + "epoch": 1.81651376146789, + "grad_norm": 2.1178297996520996, + "learning_rate": 4.899206439788117e-05, + "loss": 0.3753, + "num_input_tokens_seen": 1854280, + "step": 6930 + }, + { + "epoch": 1.817824377457405, + "grad_norm": 2.894911289215088, + "learning_rate": 4.898884702707499e-05, + "loss": 0.3752, + "num_input_tokens_seen": 1855416, + "step": 6935 + }, + { + "epoch": 1.81913499344692, + "grad_norm": 0.8735739588737488, + "learning_rate": 4.898562463544459e-05, + "loss": 0.5175, + "num_input_tokens_seen": 1857320, + "step": 6940 + }, + { + "epoch": 1.820445609436435, + "grad_norm": 2.069521903991699, + "learning_rate": 4.89823972236644e-05, + "loss": 0.5441, + "num_input_tokens_seen": 1858808, + "step": 6945 + }, + { + "epoch": 1.82175622542595, + "grad_norm": 2.795410394668579, + "learning_rate": 4.89791647924099e-05, + "loss": 0.652, + "num_input_tokens_seen": 1860120, + "step": 6950 + }, + { + "epoch": 1.8230668414154652, + "grad_norm": 2.5543606281280518, + "learning_rate": 4.8975927342357646e-05, + "loss": 0.4411, + "num_input_tokens_seen": 1861656, + "step": 6955 + }, + { + "epoch": 1.8243774574049803, + "grad_norm": 1.8915706872940063, + "learning_rate": 4.897268487418523e-05, + "loss": 0.6883, + "num_input_tokens_seen": 1862664, + "step": 6960 + }, + { + "epoch": 1.8256880733944953, + "grad_norm": 2.9259891510009766, + "learning_rate": 4.896943738857128e-05, + "loss": 0.4568, + "num_input_tokens_seen": 1864008, + "step": 6965 + }, + { + "epoch": 1.8269986893840104, + "grad_norm": 2.4062418937683105, + "learning_rate": 4.896618488619551e-05, + "loss": 0.3543, + "num_input_tokens_seen": 1865128, + "step": 6970 + }, + { + "epoch": 1.8283093053735255, + "grad_norm": 3.5423130989074707, + "learning_rate": 4.8962927367738655e-05, + "loss": 0.3395, + "num_input_tokens_seen": 1866296, + "step": 6975 + }, + { + "epoch": 1.8296199213630406, + "grad_norm": 3.573193311691284, + "learning_rate": 4.89596648338825e-05, + "loss": 0.6567, + "num_input_tokens_seen": 1867496, + "step": 6980 + }, + { + "epoch": 1.8309305373525557, + "grad_norm": 2.9509410858154297, + "learning_rate": 4.89563972853099e-05, + "loss": 0.5909, + "num_input_tokens_seen": 1868728, + "step": 6985 + }, + { + "epoch": 1.8322411533420708, + "grad_norm": 8.050041198730469, + "learning_rate": 4.895312472270475e-05, + "loss": 0.6107, + "num_input_tokens_seen": 1869816, + "step": 6990 + }, + { + "epoch": 1.833551769331586, + "grad_norm": 1.5667234659194946, + "learning_rate": 4.894984714675197e-05, + "loss": 0.5259, + "num_input_tokens_seen": 1871672, + "step": 6995 + }, + { + "epoch": 1.834862385321101, + "grad_norm": 1.9715230464935303, + "learning_rate": 4.8946564558137566e-05, + "loss": 0.3331, + "num_input_tokens_seen": 1872888, + "step": 7000 + }, + { + "epoch": 1.8361730013106161, + "grad_norm": 2.351400375366211, + "learning_rate": 4.894327695754858e-05, + "loss": 0.449, + "num_input_tokens_seen": 1874088, + "step": 7005 + }, + { + "epoch": 1.8374836173001312, + "grad_norm": 1.5817878246307373, + "learning_rate": 4.893998434567309e-05, + "loss": 0.3682, + "num_input_tokens_seen": 1875176, + "step": 7010 + }, + { + "epoch": 1.838794233289646, + "grad_norm": 2.9611563682556152, + "learning_rate": 4.893668672320025e-05, + "loss": 0.5066, + "num_input_tokens_seen": 1876760, + "step": 7015 + }, + { + "epoch": 1.8401048492791612, + "grad_norm": 2.1727685928344727, + "learning_rate": 4.893338409082023e-05, + "loss": 0.4923, + "num_input_tokens_seen": 1877928, + "step": 7020 + }, + { + "epoch": 1.8414154652686763, + "grad_norm": 8.042097091674805, + "learning_rate": 4.8930076449224285e-05, + "loss": 0.5746, + "num_input_tokens_seen": 1879224, + "step": 7025 + }, + { + "epoch": 1.8427260812581914, + "grad_norm": 3.114877700805664, + "learning_rate": 4.892676379910468e-05, + "loss": 0.5469, + "num_input_tokens_seen": 1880376, + "step": 7030 + }, + { + "epoch": 1.8440366972477065, + "grad_norm": 1.5809359550476074, + "learning_rate": 4.8923446141154746e-05, + "loss": 0.345, + "num_input_tokens_seen": 1881800, + "step": 7035 + }, + { + "epoch": 1.8453473132372213, + "grad_norm": 1.467731237411499, + "learning_rate": 4.8920123476068894e-05, + "loss": 0.3119, + "num_input_tokens_seen": 1882888, + "step": 7040 + }, + { + "epoch": 1.8466579292267364, + "grad_norm": 4.926959037780762, + "learning_rate": 4.891679580454251e-05, + "loss": 0.5022, + "num_input_tokens_seen": 1884072, + "step": 7045 + }, + { + "epoch": 1.8479685452162515, + "grad_norm": 1.1303763389587402, + "learning_rate": 4.89134631272721e-05, + "loss": 0.4409, + "num_input_tokens_seen": 1885480, + "step": 7050 + }, + { + "epoch": 1.8492791612057666, + "grad_norm": 2.936983346939087, + "learning_rate": 4.891012544495517e-05, + "loss": 0.8605, + "num_input_tokens_seen": 1886568, + "step": 7055 + }, + { + "epoch": 1.8505897771952817, + "grad_norm": 0.9545637369155884, + "learning_rate": 4.890678275829029e-05, + "loss": 0.4179, + "num_input_tokens_seen": 1888136, + "step": 7060 + }, + { + "epoch": 1.8519003931847968, + "grad_norm": 2.078420877456665, + "learning_rate": 4.89034350679771e-05, + "loss": 0.5659, + "num_input_tokens_seen": 1890008, + "step": 7065 + }, + { + "epoch": 1.853211009174312, + "grad_norm": 5.014571666717529, + "learning_rate": 4.890008237471625e-05, + "loss": 0.2546, + "num_input_tokens_seen": 1891368, + "step": 7070 + }, + { + "epoch": 1.854521625163827, + "grad_norm": 2.469261407852173, + "learning_rate": 4.889672467920945e-05, + "loss": 0.3995, + "num_input_tokens_seen": 1892904, + "step": 7075 + }, + { + "epoch": 1.855832241153342, + "grad_norm": 2.8876354694366455, + "learning_rate": 4.889336198215947e-05, + "loss": 0.6931, + "num_input_tokens_seen": 1894024, + "step": 7080 + }, + { + "epoch": 1.8571428571428572, + "grad_norm": 2.008077621459961, + "learning_rate": 4.888999428427011e-05, + "loss": 0.3787, + "num_input_tokens_seen": 1895544, + "step": 7085 + }, + { + "epoch": 1.8584534731323723, + "grad_norm": 2.492664337158203, + "learning_rate": 4.8886621586246226e-05, + "loss": 0.5731, + "num_input_tokens_seen": 1896536, + "step": 7090 + }, + { + "epoch": 1.8597640891218874, + "grad_norm": 2.450340986251831, + "learning_rate": 4.8883243888793715e-05, + "loss": 0.5168, + "num_input_tokens_seen": 1897544, + "step": 7095 + }, + { + "epoch": 1.8610747051114025, + "grad_norm": 2.3027472496032715, + "learning_rate": 4.887986119261953e-05, + "loss": 0.517, + "num_input_tokens_seen": 1898632, + "step": 7100 + }, + { + "epoch": 1.8623853211009176, + "grad_norm": 5.164468288421631, + "learning_rate": 4.887647349843165e-05, + "loss": 0.3974, + "num_input_tokens_seen": 1900328, + "step": 7105 + }, + { + "epoch": 1.8636959370904325, + "grad_norm": 1.627133846282959, + "learning_rate": 4.887308080693913e-05, + "loss": 0.5925, + "num_input_tokens_seen": 1902360, + "step": 7110 + }, + { + "epoch": 1.8650065530799476, + "grad_norm": 1.7039512395858765, + "learning_rate": 4.886968311885204e-05, + "loss": 0.3713, + "num_input_tokens_seen": 1903912, + "step": 7115 + }, + { + "epoch": 1.8663171690694627, + "grad_norm": 2.546722412109375, + "learning_rate": 4.8866280434881516e-05, + "loss": 0.5447, + "num_input_tokens_seen": 1905768, + "step": 7120 + }, + { + "epoch": 1.8676277850589778, + "grad_norm": 1.9408930540084839, + "learning_rate": 4.886287275573974e-05, + "loss": 0.4877, + "num_input_tokens_seen": 1906872, + "step": 7125 + }, + { + "epoch": 1.8689384010484926, + "grad_norm": 1.8376281261444092, + "learning_rate": 4.885946008213993e-05, + "loss": 0.6307, + "num_input_tokens_seen": 1908184, + "step": 7130 + }, + { + "epoch": 1.8702490170380077, + "grad_norm": 4.655306339263916, + "learning_rate": 4.885604241479635e-05, + "loss": 0.4888, + "num_input_tokens_seen": 1909176, + "step": 7135 + }, + { + "epoch": 1.8715596330275228, + "grad_norm": 2.386751413345337, + "learning_rate": 4.885261975442431e-05, + "loss": 0.534, + "num_input_tokens_seen": 1911224, + "step": 7140 + }, + { + "epoch": 1.872870249017038, + "grad_norm": 1.8485662937164307, + "learning_rate": 4.884919210174017e-05, + "loss": 0.5448, + "num_input_tokens_seen": 1912552, + "step": 7145 + }, + { + "epoch": 1.874180865006553, + "grad_norm": 2.32991886138916, + "learning_rate": 4.8845759457461326e-05, + "loss": 0.5136, + "num_input_tokens_seen": 1914776, + "step": 7150 + }, + { + "epoch": 1.875491480996068, + "grad_norm": 2.053471326828003, + "learning_rate": 4.8842321822306235e-05, + "loss": 0.4982, + "num_input_tokens_seen": 1916056, + "step": 7155 + }, + { + "epoch": 1.8768020969855832, + "grad_norm": 1.1836693286895752, + "learning_rate": 4.8838879196994385e-05, + "loss": 0.386, + "num_input_tokens_seen": 1917400, + "step": 7160 + }, + { + "epoch": 1.8781127129750983, + "grad_norm": 2.0068347454071045, + "learning_rate": 4.88354315822463e-05, + "loss": 0.5275, + "num_input_tokens_seen": 1918488, + "step": 7165 + }, + { + "epoch": 1.8794233289646134, + "grad_norm": 1.815474510192871, + "learning_rate": 4.8831978978783585e-05, + "loss": 0.5546, + "num_input_tokens_seen": 1919928, + "step": 7170 + }, + { + "epoch": 1.8807339449541285, + "grad_norm": 1.0896104574203491, + "learning_rate": 4.882852138732884e-05, + "loss": 0.4427, + "num_input_tokens_seen": 1921304, + "step": 7175 + }, + { + "epoch": 1.8820445609436436, + "grad_norm": 2.5142838954925537, + "learning_rate": 4.882505880860574e-05, + "loss": 0.4527, + "num_input_tokens_seen": 1923016, + "step": 7180 + }, + { + "epoch": 1.8833551769331587, + "grad_norm": 3.860708475112915, + "learning_rate": 4.8821591243339e-05, + "loss": 0.7514, + "num_input_tokens_seen": 1924184, + "step": 7185 + }, + { + "epoch": 1.8846657929226738, + "grad_norm": 5.4011359214782715, + "learning_rate": 4.8818118692254375e-05, + "loss": 0.4543, + "num_input_tokens_seen": 1925240, + "step": 7190 + }, + { + "epoch": 1.8859764089121889, + "grad_norm": 2.3434953689575195, + "learning_rate": 4.881464115607865e-05, + "loss": 0.5782, + "num_input_tokens_seen": 1926328, + "step": 7195 + }, + { + "epoch": 1.8872870249017037, + "grad_norm": 6.338839530944824, + "learning_rate": 4.88111586355397e-05, + "loss": 0.4297, + "num_input_tokens_seen": 1927768, + "step": 7200 + }, + { + "epoch": 1.8885976408912188, + "grad_norm": 1.815219521522522, + "learning_rate": 4.880767113136637e-05, + "loss": 0.5726, + "num_input_tokens_seen": 1929672, + "step": 7205 + }, + { + "epoch": 1.889908256880734, + "grad_norm": 7.349604606628418, + "learning_rate": 4.880417864428862e-05, + "loss": 0.4605, + "num_input_tokens_seen": 1931752, + "step": 7210 + }, + { + "epoch": 1.891218872870249, + "grad_norm": 2.199416160583496, + "learning_rate": 4.880068117503741e-05, + "loss": 0.7044, + "num_input_tokens_seen": 1933160, + "step": 7215 + }, + { + "epoch": 1.8925294888597641, + "grad_norm": 1.3717944622039795, + "learning_rate": 4.879717872434475e-05, + "loss": 0.5167, + "num_input_tokens_seen": 1934440, + "step": 7220 + }, + { + "epoch": 1.893840104849279, + "grad_norm": 2.8082516193389893, + "learning_rate": 4.879367129294371e-05, + "loss": 0.6791, + "num_input_tokens_seen": 1935912, + "step": 7225 + }, + { + "epoch": 1.895150720838794, + "grad_norm": 2.9458844661712646, + "learning_rate": 4.879015888156837e-05, + "loss": 0.4383, + "num_input_tokens_seen": 1937064, + "step": 7230 + }, + { + "epoch": 1.8964613368283092, + "grad_norm": 3.0292932987213135, + "learning_rate": 4.8786641490953886e-05, + "loss": 0.4193, + "num_input_tokens_seen": 1938232, + "step": 7235 + }, + { + "epoch": 1.8977719528178243, + "grad_norm": 1.7168062925338745, + "learning_rate": 4.878311912183644e-05, + "loss": 0.5324, + "num_input_tokens_seen": 1939368, + "step": 7240 + }, + { + "epoch": 1.8990825688073394, + "grad_norm": 1.8307721614837646, + "learning_rate": 4.877959177495326e-05, + "loss": 0.5647, + "num_input_tokens_seen": 1940648, + "step": 7245 + }, + { + "epoch": 1.9003931847968545, + "grad_norm": 3.1926794052124023, + "learning_rate": 4.87760594510426e-05, + "loss": 0.5166, + "num_input_tokens_seen": 1942232, + "step": 7250 + }, + { + "epoch": 1.9017038007863696, + "grad_norm": 2.066420316696167, + "learning_rate": 4.877252215084379e-05, + "loss": 0.7003, + "num_input_tokens_seen": 1943656, + "step": 7255 + }, + { + "epoch": 1.9030144167758847, + "grad_norm": 5.403193473815918, + "learning_rate": 4.876897987509716e-05, + "loss": 0.6727, + "num_input_tokens_seen": 1944936, + "step": 7260 + }, + { + "epoch": 1.9043250327653998, + "grad_norm": 1.9212629795074463, + "learning_rate": 4.876543262454412e-05, + "loss": 0.3267, + "num_input_tokens_seen": 1945800, + "step": 7265 + }, + { + "epoch": 1.9056356487549149, + "grad_norm": 2.266000270843506, + "learning_rate": 4.8761880399927084e-05, + "loss": 0.4396, + "num_input_tokens_seen": 1947144, + "step": 7270 + }, + { + "epoch": 1.90694626474443, + "grad_norm": 1.9462178945541382, + "learning_rate": 4.875832320198954e-05, + "loss": 0.6192, + "num_input_tokens_seen": 1948360, + "step": 7275 + }, + { + "epoch": 1.908256880733945, + "grad_norm": 7.4512505531311035, + "learning_rate": 4.8754761031476005e-05, + "loss": 0.4607, + "num_input_tokens_seen": 1949336, + "step": 7280 + }, + { + "epoch": 1.9095674967234602, + "grad_norm": 2.947800874710083, + "learning_rate": 4.875119388913203e-05, + "loss": 0.6241, + "num_input_tokens_seen": 1950744, + "step": 7285 + }, + { + "epoch": 1.9108781127129753, + "grad_norm": 2.616353750228882, + "learning_rate": 4.874762177570422e-05, + "loss": 0.4512, + "num_input_tokens_seen": 1952056, + "step": 7290 + }, + { + "epoch": 1.9121887287024901, + "grad_norm": 2.0679898262023926, + "learning_rate": 4.874404469194019e-05, + "loss": 0.6992, + "num_input_tokens_seen": 1953272, + "step": 7295 + }, + { + "epoch": 1.9134993446920052, + "grad_norm": 3.313509464263916, + "learning_rate": 4.8740462638588644e-05, + "loss": 0.3943, + "num_input_tokens_seen": 1954728, + "step": 7300 + }, + { + "epoch": 1.9148099606815203, + "grad_norm": 2.6870265007019043, + "learning_rate": 4.873687561639929e-05, + "loss": 0.4278, + "num_input_tokens_seen": 1956600, + "step": 7305 + }, + { + "epoch": 1.9161205766710354, + "grad_norm": 3.863673686981201, + "learning_rate": 4.8733283626122875e-05, + "loss": 0.4349, + "num_input_tokens_seen": 1958008, + "step": 7310 + }, + { + "epoch": 1.9174311926605505, + "grad_norm": 1.6935484409332275, + "learning_rate": 4.872968666851121e-05, + "loss": 0.4696, + "num_input_tokens_seen": 1959464, + "step": 7315 + }, + { + "epoch": 1.9187418086500654, + "grad_norm": 14.030714988708496, + "learning_rate": 4.872608474431712e-05, + "loss": 0.6228, + "num_input_tokens_seen": 1961224, + "step": 7320 + }, + { + "epoch": 1.9200524246395805, + "grad_norm": 3.171588659286499, + "learning_rate": 4.87224778542945e-05, + "loss": 0.4539, + "num_input_tokens_seen": 1962280, + "step": 7325 + }, + { + "epoch": 1.9213630406290956, + "grad_norm": 2.7199060916900635, + "learning_rate": 4.871886599919825e-05, + "loss": 0.6149, + "num_input_tokens_seen": 1963336, + "step": 7330 + }, + { + "epoch": 1.9226736566186107, + "grad_norm": 3.354259729385376, + "learning_rate": 4.871524917978433e-05, + "loss": 0.3941, + "num_input_tokens_seen": 1964504, + "step": 7335 + }, + { + "epoch": 1.9239842726081258, + "grad_norm": 5.790065288543701, + "learning_rate": 4.8711627396809725e-05, + "loss": 0.531, + "num_input_tokens_seen": 1965976, + "step": 7340 + }, + { + "epoch": 1.9252948885976409, + "grad_norm": 0.9999198317527771, + "learning_rate": 4.870800065103248e-05, + "loss": 0.5716, + "num_input_tokens_seen": 1967144, + "step": 7345 + }, + { + "epoch": 1.926605504587156, + "grad_norm": 4.923386573791504, + "learning_rate": 4.870436894321167e-05, + "loss": 0.3266, + "num_input_tokens_seen": 1968904, + "step": 7350 + }, + { + "epoch": 1.927916120576671, + "grad_norm": 5.054831027984619, + "learning_rate": 4.87007322741074e-05, + "loss": 0.4249, + "num_input_tokens_seen": 1970008, + "step": 7355 + }, + { + "epoch": 1.9292267365661862, + "grad_norm": 0.8191948533058167, + "learning_rate": 4.8697090644480804e-05, + "loss": 0.4014, + "num_input_tokens_seen": 1971416, + "step": 7360 + }, + { + "epoch": 1.9305373525557012, + "grad_norm": 1.8359558582305908, + "learning_rate": 4.86934440550941e-05, + "loss": 0.4107, + "num_input_tokens_seen": 1972648, + "step": 7365 + }, + { + "epoch": 1.9318479685452163, + "grad_norm": 1.0171012878417969, + "learning_rate": 4.868979250671047e-05, + "loss": 0.4355, + "num_input_tokens_seen": 1973928, + "step": 7370 + }, + { + "epoch": 1.9331585845347314, + "grad_norm": 2.8427610397338867, + "learning_rate": 4.868613600009422e-05, + "loss": 0.2942, + "num_input_tokens_seen": 1975144, + "step": 7375 + }, + { + "epoch": 1.9344692005242465, + "grad_norm": 7.593844413757324, + "learning_rate": 4.8682474536010624e-05, + "loss": 0.5206, + "num_input_tokens_seen": 1976808, + "step": 7380 + }, + { + "epoch": 1.9357798165137616, + "grad_norm": 8.261125564575195, + "learning_rate": 4.8678808115226026e-05, + "loss": 0.5947, + "num_input_tokens_seen": 1978024, + "step": 7385 + }, + { + "epoch": 1.9370904325032765, + "grad_norm": 3.714458465576172, + "learning_rate": 4.867513673850781e-05, + "loss": 0.5994, + "num_input_tokens_seen": 1979048, + "step": 7390 + }, + { + "epoch": 1.9384010484927916, + "grad_norm": 22.94642448425293, + "learning_rate": 4.867146040662438e-05, + "loss": 0.5262, + "num_input_tokens_seen": 1980296, + "step": 7395 + }, + { + "epoch": 1.9397116644823067, + "grad_norm": 2.8015122413635254, + "learning_rate": 4.8667779120345185e-05, + "loss": 0.3975, + "num_input_tokens_seen": 1981592, + "step": 7400 + }, + { + "epoch": 1.9410222804718218, + "grad_norm": 2.3649559020996094, + "learning_rate": 4.866409288044072e-05, + "loss": 0.4882, + "num_input_tokens_seen": 1982584, + "step": 7405 + }, + { + "epoch": 1.9423328964613367, + "grad_norm": 2.413203239440918, + "learning_rate": 4.86604016876825e-05, + "loss": 0.4495, + "num_input_tokens_seen": 1984168, + "step": 7410 + }, + { + "epoch": 1.9436435124508518, + "grad_norm": 1.8091403245925903, + "learning_rate": 4.8656705542843085e-05, + "loss": 0.4462, + "num_input_tokens_seen": 1985528, + "step": 7415 + }, + { + "epoch": 1.9449541284403669, + "grad_norm": 2.4851231575012207, + "learning_rate": 4.865300444669608e-05, + "loss": 0.4707, + "num_input_tokens_seen": 1986936, + "step": 7420 + }, + { + "epoch": 1.946264744429882, + "grad_norm": 5.060756683349609, + "learning_rate": 4.864929840001612e-05, + "loss": 0.4181, + "num_input_tokens_seen": 1988072, + "step": 7425 + }, + { + "epoch": 1.947575360419397, + "grad_norm": 2.1802303791046143, + "learning_rate": 4.8645587403578855e-05, + "loss": 0.7194, + "num_input_tokens_seen": 1989256, + "step": 7430 + }, + { + "epoch": 1.9488859764089121, + "grad_norm": 1.2876379489898682, + "learning_rate": 4.864187145816101e-05, + "loss": 0.5175, + "num_input_tokens_seen": 1990632, + "step": 7435 + }, + { + "epoch": 1.9501965923984272, + "grad_norm": 1.6239171028137207, + "learning_rate": 4.863815056454032e-05, + "loss": 0.433, + "num_input_tokens_seen": 1991896, + "step": 7440 + }, + { + "epoch": 1.9515072083879423, + "grad_norm": 3.2995622158050537, + "learning_rate": 4.863442472349555e-05, + "loss": 0.4563, + "num_input_tokens_seen": 1993576, + "step": 7445 + }, + { + "epoch": 1.9528178243774574, + "grad_norm": 1.4959968328475952, + "learning_rate": 4.8630693935806536e-05, + "loss": 0.3856, + "num_input_tokens_seen": 1995288, + "step": 7450 + }, + { + "epoch": 1.9541284403669725, + "grad_norm": 2.060763359069824, + "learning_rate": 4.8626958202254106e-05, + "loss": 0.6984, + "num_input_tokens_seen": 1996392, + "step": 7455 + }, + { + "epoch": 1.9554390563564876, + "grad_norm": 2.338310718536377, + "learning_rate": 4.862321752362015e-05, + "loss": 0.4541, + "num_input_tokens_seen": 1998552, + "step": 7460 + }, + { + "epoch": 1.9567496723460027, + "grad_norm": 2.2459163665771484, + "learning_rate": 4.8619471900687583e-05, + "loss": 0.4755, + "num_input_tokens_seen": 1999688, + "step": 7465 + }, + { + "epoch": 1.9580602883355178, + "grad_norm": 1.399213194847107, + "learning_rate": 4.861572133424036e-05, + "loss": 0.4332, + "num_input_tokens_seen": 2000904, + "step": 7470 + }, + { + "epoch": 1.959370904325033, + "grad_norm": 3.3811533451080322, + "learning_rate": 4.8611965825063466e-05, + "loss": 0.506, + "num_input_tokens_seen": 2002024, + "step": 7475 + }, + { + "epoch": 1.9606815203145478, + "grad_norm": 3.1673130989074707, + "learning_rate": 4.8608205373942915e-05, + "loss": 0.4719, + "num_input_tokens_seen": 2003128, + "step": 7480 + }, + { + "epoch": 1.9619921363040629, + "grad_norm": 1.723362684249878, + "learning_rate": 4.860443998166578e-05, + "loss": 0.4899, + "num_input_tokens_seen": 2004360, + "step": 7485 + }, + { + "epoch": 1.963302752293578, + "grad_norm": 6.403108596801758, + "learning_rate": 4.860066964902014e-05, + "loss": 0.5016, + "num_input_tokens_seen": 2005352, + "step": 7490 + }, + { + "epoch": 1.964613368283093, + "grad_norm": 1.9018880128860474, + "learning_rate": 4.859689437679512e-05, + "loss": 0.4442, + "num_input_tokens_seen": 2006520, + "step": 7495 + }, + { + "epoch": 1.9659239842726082, + "grad_norm": 2.856417179107666, + "learning_rate": 4.859311416578088e-05, + "loss": 0.4078, + "num_input_tokens_seen": 2007512, + "step": 7500 + }, + { + "epoch": 1.967234600262123, + "grad_norm": 6.72622013092041, + "learning_rate": 4.8589329016768615e-05, + "loss": 0.8769, + "num_input_tokens_seen": 2008504, + "step": 7505 + }, + { + "epoch": 1.9685452162516381, + "grad_norm": 2.862217903137207, + "learning_rate": 4.8585538930550547e-05, + "loss": 0.5395, + "num_input_tokens_seen": 2009960, + "step": 7510 + }, + { + "epoch": 1.9698558322411532, + "grad_norm": 5.306510925292969, + "learning_rate": 4.858174390791993e-05, + "loss": 0.4469, + "num_input_tokens_seen": 2011208, + "step": 7515 + }, + { + "epoch": 1.9711664482306683, + "grad_norm": 2.48496413230896, + "learning_rate": 4.857794394967106e-05, + "loss": 0.4699, + "num_input_tokens_seen": 2012536, + "step": 7520 + }, + { + "epoch": 1.9724770642201834, + "grad_norm": 1.3882068395614624, + "learning_rate": 4.8574139056599256e-05, + "loss": 0.3005, + "num_input_tokens_seen": 2014120, + "step": 7525 + }, + { + "epoch": 1.9737876802096985, + "grad_norm": 2.70369815826416, + "learning_rate": 4.8570329229500884e-05, + "loss": 0.5019, + "num_input_tokens_seen": 2015416, + "step": 7530 + }, + { + "epoch": 1.9750982961992136, + "grad_norm": 1.8175967931747437, + "learning_rate": 4.8566514469173326e-05, + "loss": 0.6662, + "num_input_tokens_seen": 2016840, + "step": 7535 + }, + { + "epoch": 1.9764089121887287, + "grad_norm": 8.160014152526855, + "learning_rate": 4.8562694776415016e-05, + "loss": 0.5087, + "num_input_tokens_seen": 2018328, + "step": 7540 + }, + { + "epoch": 1.9777195281782438, + "grad_norm": 0.7801327705383301, + "learning_rate": 4.8558870152025404e-05, + "loss": 0.3018, + "num_input_tokens_seen": 2019896, + "step": 7545 + }, + { + "epoch": 1.979030144167759, + "grad_norm": 4.107687950134277, + "learning_rate": 4.855504059680497e-05, + "loss": 0.5372, + "num_input_tokens_seen": 2021144, + "step": 7550 + }, + { + "epoch": 1.980340760157274, + "grad_norm": 2.935389518737793, + "learning_rate": 4.855120611155524e-05, + "loss": 0.3392, + "num_input_tokens_seen": 2022200, + "step": 7555 + }, + { + "epoch": 1.981651376146789, + "grad_norm": 7.714248180389404, + "learning_rate": 4.854736669707876e-05, + "loss": 0.5946, + "num_input_tokens_seen": 2023176, + "step": 7560 + }, + { + "epoch": 1.9829619921363042, + "grad_norm": 5.797147274017334, + "learning_rate": 4.854352235417913e-05, + "loss": 0.6968, + "num_input_tokens_seen": 2024312, + "step": 7565 + }, + { + "epoch": 1.9842726081258193, + "grad_norm": 1.3522987365722656, + "learning_rate": 4.853967308366094e-05, + "loss": 0.5241, + "num_input_tokens_seen": 2025576, + "step": 7570 + }, + { + "epoch": 1.9855832241153342, + "grad_norm": 1.0288139581680298, + "learning_rate": 4.8535818886329855e-05, + "loss": 0.3558, + "num_input_tokens_seen": 2027640, + "step": 7575 + }, + { + "epoch": 1.9868938401048493, + "grad_norm": 3.8317983150482178, + "learning_rate": 4.853195976299254e-05, + "loss": 0.4994, + "num_input_tokens_seen": 2029064, + "step": 7580 + }, + { + "epoch": 1.9882044560943644, + "grad_norm": 5.565948963165283, + "learning_rate": 4.85280957144567e-05, + "loss": 0.4954, + "num_input_tokens_seen": 2030328, + "step": 7585 + }, + { + "epoch": 1.9895150720838795, + "grad_norm": 3.6888370513916016, + "learning_rate": 4.852422674153109e-05, + "loss": 0.4646, + "num_input_tokens_seen": 2031672, + "step": 7590 + }, + { + "epoch": 1.9908256880733946, + "grad_norm": 2.136166572570801, + "learning_rate": 4.852035284502547e-05, + "loss": 0.533, + "num_input_tokens_seen": 2032760, + "step": 7595 + }, + { + "epoch": 1.9921363040629094, + "grad_norm": 1.821739912033081, + "learning_rate": 4.851647402575063e-05, + "loss": 0.4308, + "num_input_tokens_seen": 2033960, + "step": 7600 + }, + { + "epoch": 1.9934469200524245, + "grad_norm": 2.7893450260162354, + "learning_rate": 4.851259028451842e-05, + "loss": 0.542, + "num_input_tokens_seen": 2035192, + "step": 7605 + }, + { + "epoch": 1.9947575360419396, + "grad_norm": 5.952236175537109, + "learning_rate": 4.850870162214168e-05, + "loss": 0.5259, + "num_input_tokens_seen": 2036584, + "step": 7610 + }, + { + "epoch": 1.9960681520314547, + "grad_norm": 3.513505458831787, + "learning_rate": 4.850480803943432e-05, + "loss": 0.446, + "num_input_tokens_seen": 2037768, + "step": 7615 + }, + { + "epoch": 1.9973787680209698, + "grad_norm": 1.778805136680603, + "learning_rate": 4.8500909537211246e-05, + "loss": 0.597, + "num_input_tokens_seen": 2039128, + "step": 7620 + }, + { + "epoch": 1.998689384010485, + "grad_norm": 2.034057140350342, + "learning_rate": 4.849700611628841e-05, + "loss": 0.3756, + "num_input_tokens_seen": 2040376, + "step": 7625 + }, + { + "epoch": 2.0, + "grad_norm": 2.911923408508301, + "learning_rate": 4.84930977774828e-05, + "loss": 0.3695, + "num_input_tokens_seen": 2041488, + "step": 7630 + }, + { + "epoch": 2.000524246395806, + "eval_loss": 0.47599396109580994, + "eval_runtime": 18.1879, + "eval_samples_per_second": 46.625, + "eval_steps_per_second": 23.312, + "num_input_tokens_seen": 2042048, + "step": 7632 + }, + { + "epoch": 2.001310615989515, + "grad_norm": 3.564469337463379, + "learning_rate": 4.8489184521612416e-05, + "loss": 0.5027, + "num_input_tokens_seen": 2043136, + "step": 7635 + }, + { + "epoch": 2.00262123197903, + "grad_norm": 1.647562861442566, + "learning_rate": 4.848526634949629e-05, + "loss": 0.5259, + "num_input_tokens_seen": 2044288, + "step": 7640 + }, + { + "epoch": 2.0039318479685453, + "grad_norm": 2.1187832355499268, + "learning_rate": 4.848134326195451e-05, + "loss": 0.3441, + "num_input_tokens_seen": 2045520, + "step": 7645 + }, + { + "epoch": 2.0052424639580604, + "grad_norm": 1.4609802961349487, + "learning_rate": 4.847741525980815e-05, + "loss": 0.58, + "num_input_tokens_seen": 2046784, + "step": 7650 + }, + { + "epoch": 2.0065530799475755, + "grad_norm": 1.0301036834716797, + "learning_rate": 4.8473482343879344e-05, + "loss": 0.6512, + "num_input_tokens_seen": 2048544, + "step": 7655 + }, + { + "epoch": 2.0078636959370906, + "grad_norm": 2.931572675704956, + "learning_rate": 4.8469544514991235e-05, + "loss": 0.4562, + "num_input_tokens_seen": 2049888, + "step": 7660 + }, + { + "epoch": 2.0091743119266057, + "grad_norm": 1.7635279893875122, + "learning_rate": 4.846560177396802e-05, + "loss": 0.5482, + "num_input_tokens_seen": 2051136, + "step": 7665 + }, + { + "epoch": 2.0104849279161208, + "grad_norm": 2.5424697399139404, + "learning_rate": 4.8461654121634905e-05, + "loss": 0.3916, + "num_input_tokens_seen": 2052528, + "step": 7670 + }, + { + "epoch": 2.011795543905636, + "grad_norm": 2.1159603595733643, + "learning_rate": 4.8457701558818106e-05, + "loss": 0.6164, + "num_input_tokens_seen": 2053488, + "step": 7675 + }, + { + "epoch": 2.0131061598951505, + "grad_norm": 1.3887609243392944, + "learning_rate": 4.845374408634491e-05, + "loss": 0.4311, + "num_input_tokens_seen": 2054896, + "step": 7680 + }, + { + "epoch": 2.0144167758846656, + "grad_norm": 2.6611459255218506, + "learning_rate": 4.84497817050436e-05, + "loss": 0.4598, + "num_input_tokens_seen": 2055952, + "step": 7685 + }, + { + "epoch": 2.0157273918741807, + "grad_norm": 3.969905376434326, + "learning_rate": 4.8445814415743504e-05, + "loss": 0.4685, + "num_input_tokens_seen": 2056912, + "step": 7690 + }, + { + "epoch": 2.017038007863696, + "grad_norm": 6.547998428344727, + "learning_rate": 4.844184221927495e-05, + "loss": 0.4942, + "num_input_tokens_seen": 2058048, + "step": 7695 + }, + { + "epoch": 2.018348623853211, + "grad_norm": 5.795790672302246, + "learning_rate": 4.843786511646933e-05, + "loss": 0.5116, + "num_input_tokens_seen": 2059424, + "step": 7700 + }, + { + "epoch": 2.019659239842726, + "grad_norm": 2.472323179244995, + "learning_rate": 4.843388310815904e-05, + "loss": 0.3877, + "num_input_tokens_seen": 2060464, + "step": 7705 + }, + { + "epoch": 2.020969855832241, + "grad_norm": 3.621018886566162, + "learning_rate": 4.8429896195177504e-05, + "loss": 0.4815, + "num_input_tokens_seen": 2061744, + "step": 7710 + }, + { + "epoch": 2.022280471821756, + "grad_norm": 1.4692233800888062, + "learning_rate": 4.842590437835918e-05, + "loss": 0.3773, + "num_input_tokens_seen": 2063344, + "step": 7715 + }, + { + "epoch": 2.0235910878112713, + "grad_norm": 2.205232620239258, + "learning_rate": 4.8421907658539545e-05, + "loss": 0.3915, + "num_input_tokens_seen": 2064640, + "step": 7720 + }, + { + "epoch": 2.0249017038007864, + "grad_norm": 3.168144464492798, + "learning_rate": 4.84179060365551e-05, + "loss": 0.6142, + "num_input_tokens_seen": 2065840, + "step": 7725 + }, + { + "epoch": 2.0262123197903015, + "grad_norm": 1.663018822669983, + "learning_rate": 4.84138995132434e-05, + "loss": 0.6083, + "num_input_tokens_seen": 2067184, + "step": 7730 + }, + { + "epoch": 2.0275229357798166, + "grad_norm": 2.081707715988159, + "learning_rate": 4.840988808944298e-05, + "loss": 0.5276, + "num_input_tokens_seen": 2068496, + "step": 7735 + }, + { + "epoch": 2.0288335517693317, + "grad_norm": 1.4348702430725098, + "learning_rate": 4.8405871765993433e-05, + "loss": 0.4758, + "num_input_tokens_seen": 2070096, + "step": 7740 + }, + { + "epoch": 2.0301441677588468, + "grad_norm": 1.7656530141830444, + "learning_rate": 4.840185054373536e-05, + "loss": 0.3706, + "num_input_tokens_seen": 2071456, + "step": 7745 + }, + { + "epoch": 2.031454783748362, + "grad_norm": 2.9900126457214355, + "learning_rate": 4.839782442351042e-05, + "loss": 0.4337, + "num_input_tokens_seen": 2072480, + "step": 7750 + }, + { + "epoch": 2.032765399737877, + "grad_norm": 3.5742204189300537, + "learning_rate": 4.839379340616124e-05, + "loss": 0.498, + "num_input_tokens_seen": 2073680, + "step": 7755 + }, + { + "epoch": 2.034076015727392, + "grad_norm": 2.90747332572937, + "learning_rate": 4.838975749253153e-05, + "loss": 0.3949, + "num_input_tokens_seen": 2074816, + "step": 7760 + }, + { + "epoch": 2.035386631716907, + "grad_norm": 2.349093437194824, + "learning_rate": 4.838571668346599e-05, + "loss": 0.4914, + "num_input_tokens_seen": 2076096, + "step": 7765 + }, + { + "epoch": 2.036697247706422, + "grad_norm": 1.2347830533981323, + "learning_rate": 4.838167097981036e-05, + "loss": 0.4993, + "num_input_tokens_seen": 2077488, + "step": 7770 + }, + { + "epoch": 2.038007863695937, + "grad_norm": 6.0153489112854, + "learning_rate": 4.8377620382411385e-05, + "loss": 0.4714, + "num_input_tokens_seen": 2078544, + "step": 7775 + }, + { + "epoch": 2.039318479685452, + "grad_norm": 2.049419403076172, + "learning_rate": 4.837356489211686e-05, + "loss": 0.3776, + "num_input_tokens_seen": 2079696, + "step": 7780 + }, + { + "epoch": 2.040629095674967, + "grad_norm": 1.3308981657028198, + "learning_rate": 4.836950450977559e-05, + "loss": 0.7174, + "num_input_tokens_seen": 2080896, + "step": 7785 + }, + { + "epoch": 2.041939711664482, + "grad_norm": 1.309061884880066, + "learning_rate": 4.836543923623741e-05, + "loss": 0.4815, + "num_input_tokens_seen": 2082432, + "step": 7790 + }, + { + "epoch": 2.0432503276539973, + "grad_norm": 1.7432855367660522, + "learning_rate": 4.8361369072353164e-05, + "loss": 0.4047, + "num_input_tokens_seen": 2083840, + "step": 7795 + }, + { + "epoch": 2.0445609436435124, + "grad_norm": 5.528487682342529, + "learning_rate": 4.835729401897474e-05, + "loss": 0.4898, + "num_input_tokens_seen": 2085424, + "step": 7800 + }, + { + "epoch": 2.0458715596330275, + "grad_norm": 1.1441868543624878, + "learning_rate": 4.835321407695503e-05, + "loss": 0.5554, + "num_input_tokens_seen": 2087584, + "step": 7805 + }, + { + "epoch": 2.0471821756225426, + "grad_norm": 5.229543685913086, + "learning_rate": 4.834912924714797e-05, + "loss": 0.6028, + "num_input_tokens_seen": 2088832, + "step": 7810 + }, + { + "epoch": 2.0484927916120577, + "grad_norm": 2.553816556930542, + "learning_rate": 4.834503953040851e-05, + "loss": 0.6564, + "num_input_tokens_seen": 2090352, + "step": 7815 + }, + { + "epoch": 2.0498034076015728, + "grad_norm": 2.7505648136138916, + "learning_rate": 4.83409449275926e-05, + "loss": 0.5171, + "num_input_tokens_seen": 2091392, + "step": 7820 + }, + { + "epoch": 2.051114023591088, + "grad_norm": 2.899912118911743, + "learning_rate": 4.8336845439557256e-05, + "loss": 0.2934, + "num_input_tokens_seen": 2092320, + "step": 7825 + }, + { + "epoch": 2.052424639580603, + "grad_norm": 2.103740692138672, + "learning_rate": 4.833274106716048e-05, + "loss": 0.4785, + "num_input_tokens_seen": 2093520, + "step": 7830 + }, + { + "epoch": 2.053735255570118, + "grad_norm": 8.869431495666504, + "learning_rate": 4.832863181126132e-05, + "loss": 0.434, + "num_input_tokens_seen": 2094752, + "step": 7835 + }, + { + "epoch": 2.055045871559633, + "grad_norm": 2.985997200012207, + "learning_rate": 4.832451767271983e-05, + "loss": 0.3757, + "num_input_tokens_seen": 2096304, + "step": 7840 + }, + { + "epoch": 2.0563564875491482, + "grad_norm": 1.491176962852478, + "learning_rate": 4.832039865239708e-05, + "loss": 0.6067, + "num_input_tokens_seen": 2098128, + "step": 7845 + }, + { + "epoch": 2.0576671035386633, + "grad_norm": 3.135634422302246, + "learning_rate": 4.83162747511552e-05, + "loss": 0.5041, + "num_input_tokens_seen": 2099376, + "step": 7850 + }, + { + "epoch": 2.0589777195281784, + "grad_norm": 2.524956703186035, + "learning_rate": 4.83121459698573e-05, + "loss": 0.6171, + "num_input_tokens_seen": 2100608, + "step": 7855 + }, + { + "epoch": 2.0602883355176935, + "grad_norm": 3.6311423778533936, + "learning_rate": 4.830801230936752e-05, + "loss": 0.5724, + "num_input_tokens_seen": 2101936, + "step": 7860 + }, + { + "epoch": 2.061598951507208, + "grad_norm": 2.3756494522094727, + "learning_rate": 4.830387377055105e-05, + "loss": 0.3302, + "num_input_tokens_seen": 2103344, + "step": 7865 + }, + { + "epoch": 2.0629095674967233, + "grad_norm": 2.903491258621216, + "learning_rate": 4.829973035427406e-05, + "loss": 0.625, + "num_input_tokens_seen": 2104640, + "step": 7870 + }, + { + "epoch": 2.0642201834862384, + "grad_norm": 3.7666664123535156, + "learning_rate": 4.8295582061403766e-05, + "loss": 1.8064, + "num_input_tokens_seen": 2105856, + "step": 7875 + }, + { + "epoch": 2.0655307994757535, + "grad_norm": 2.1437127590179443, + "learning_rate": 4.8291428892808396e-05, + "loss": 0.4071, + "num_input_tokens_seen": 2107056, + "step": 7880 + }, + { + "epoch": 2.0668414154652686, + "grad_norm": 3.7182962894439697, + "learning_rate": 4.8287270849357205e-05, + "loss": 0.4817, + "num_input_tokens_seen": 2108448, + "step": 7885 + }, + { + "epoch": 2.0681520314547837, + "grad_norm": 3.9150848388671875, + "learning_rate": 4.828310793192046e-05, + "loss": 0.5123, + "num_input_tokens_seen": 2109648, + "step": 7890 + }, + { + "epoch": 2.0694626474442988, + "grad_norm": 1.2458165884017944, + "learning_rate": 4.827894014136945e-05, + "loss": 0.5062, + "num_input_tokens_seen": 2111184, + "step": 7895 + }, + { + "epoch": 2.070773263433814, + "grad_norm": 2.315821886062622, + "learning_rate": 4.827476747857651e-05, + "loss": 0.3627, + "num_input_tokens_seen": 2112528, + "step": 7900 + }, + { + "epoch": 2.072083879423329, + "grad_norm": 1.1289631128311157, + "learning_rate": 4.827058994441494e-05, + "loss": 0.5409, + "num_input_tokens_seen": 2113952, + "step": 7905 + }, + { + "epoch": 2.073394495412844, + "grad_norm": 1.9383989572525024, + "learning_rate": 4.82664075397591e-05, + "loss": 0.5132, + "num_input_tokens_seen": 2115600, + "step": 7910 + }, + { + "epoch": 2.074705111402359, + "grad_norm": 3.1014602184295654, + "learning_rate": 4.826222026548437e-05, + "loss": 0.4917, + "num_input_tokens_seen": 2116576, + "step": 7915 + }, + { + "epoch": 2.0760157273918742, + "grad_norm": 1.4358183145523071, + "learning_rate": 4.825802812246713e-05, + "loss": 0.6257, + "num_input_tokens_seen": 2117920, + "step": 7920 + }, + { + "epoch": 2.0773263433813893, + "grad_norm": 1.0840983390808105, + "learning_rate": 4.8253831111584794e-05, + "loss": 0.4089, + "num_input_tokens_seen": 2119600, + "step": 7925 + }, + { + "epoch": 2.0786369593709044, + "grad_norm": 3.5995845794677734, + "learning_rate": 4.8249629233715776e-05, + "loss": 0.5127, + "num_input_tokens_seen": 2120704, + "step": 7930 + }, + { + "epoch": 2.0799475753604195, + "grad_norm": 1.7765165567398071, + "learning_rate": 4.824542248973954e-05, + "loss": 0.4011, + "num_input_tokens_seen": 2122352, + "step": 7935 + }, + { + "epoch": 2.0812581913499346, + "grad_norm": 2.0385987758636475, + "learning_rate": 4.824121088053654e-05, + "loss": 0.4105, + "num_input_tokens_seen": 2123728, + "step": 7940 + }, + { + "epoch": 2.0825688073394497, + "grad_norm": 1.8082096576690674, + "learning_rate": 4.8236994406988255e-05, + "loss": 0.4796, + "num_input_tokens_seen": 2125168, + "step": 7945 + }, + { + "epoch": 2.083879423328965, + "grad_norm": 2.3671562671661377, + "learning_rate": 4.82327730699772e-05, + "loss": 0.4409, + "num_input_tokens_seen": 2126224, + "step": 7950 + }, + { + "epoch": 2.08519003931848, + "grad_norm": 1.4852485656738281, + "learning_rate": 4.822854687038688e-05, + "loss": 0.5472, + "num_input_tokens_seen": 2127728, + "step": 7955 + }, + { + "epoch": 2.0865006553079946, + "grad_norm": 1.6623988151550293, + "learning_rate": 4.8224315809101836e-05, + "loss": 0.6078, + "num_input_tokens_seen": 2128944, + "step": 7960 + }, + { + "epoch": 2.0878112712975097, + "grad_norm": 2.225189208984375, + "learning_rate": 4.822007988700762e-05, + "loss": 0.7304, + "num_input_tokens_seen": 2130144, + "step": 7965 + }, + { + "epoch": 2.0891218872870247, + "grad_norm": 6.3239359855651855, + "learning_rate": 4.821583910499081e-05, + "loss": 0.4721, + "num_input_tokens_seen": 2131344, + "step": 7970 + }, + { + "epoch": 2.09043250327654, + "grad_norm": 3.5886149406433105, + "learning_rate": 4.8211593463938985e-05, + "loss": 0.4507, + "num_input_tokens_seen": 2132592, + "step": 7975 + }, + { + "epoch": 2.091743119266055, + "grad_norm": 1.691910982131958, + "learning_rate": 4.8207342964740755e-05, + "loss": 0.5664, + "num_input_tokens_seen": 2133872, + "step": 7980 + }, + { + "epoch": 2.09305373525557, + "grad_norm": 1.7533539533615112, + "learning_rate": 4.8203087608285744e-05, + "loss": 0.594, + "num_input_tokens_seen": 2135392, + "step": 7985 + }, + { + "epoch": 2.094364351245085, + "grad_norm": 2.942584276199341, + "learning_rate": 4.819882739546458e-05, + "loss": 0.339, + "num_input_tokens_seen": 2136720, + "step": 7990 + }, + { + "epoch": 2.0956749672346002, + "grad_norm": 2.68672251701355, + "learning_rate": 4.8194562327168946e-05, + "loss": 0.4494, + "num_input_tokens_seen": 2138336, + "step": 7995 + }, + { + "epoch": 2.0969855832241153, + "grad_norm": 2.7628610134124756, + "learning_rate": 4.819029240429148e-05, + "loss": 0.4557, + "num_input_tokens_seen": 2140208, + "step": 8000 + }, + { + "epoch": 2.0982961992136304, + "grad_norm": 1.9678056240081787, + "learning_rate": 4.818601762772588e-05, + "loss": 0.3879, + "num_input_tokens_seen": 2141488, + "step": 8005 + }, + { + "epoch": 2.0996068152031455, + "grad_norm": 1.294571876525879, + "learning_rate": 4.818173799836686e-05, + "loss": 0.3904, + "num_input_tokens_seen": 2142752, + "step": 8010 + }, + { + "epoch": 2.1009174311926606, + "grad_norm": 2.5748605728149414, + "learning_rate": 4.817745351711013e-05, + "loss": 0.4101, + "num_input_tokens_seen": 2144032, + "step": 8015 + }, + { + "epoch": 2.1022280471821757, + "grad_norm": 4.211734294891357, + "learning_rate": 4.817316418485243e-05, + "loss": 0.4426, + "num_input_tokens_seen": 2145584, + "step": 8020 + }, + { + "epoch": 2.103538663171691, + "grad_norm": 1.3719594478607178, + "learning_rate": 4.81688700024915e-05, + "loss": 0.6615, + "num_input_tokens_seen": 2146720, + "step": 8025 + }, + { + "epoch": 2.104849279161206, + "grad_norm": 2.5548667907714844, + "learning_rate": 4.816457097092611e-05, + "loss": 0.5577, + "num_input_tokens_seen": 2147888, + "step": 8030 + }, + { + "epoch": 2.106159895150721, + "grad_norm": 4.682997703552246, + "learning_rate": 4.816026709105605e-05, + "loss": 0.3287, + "num_input_tokens_seen": 2149120, + "step": 8035 + }, + { + "epoch": 2.107470511140236, + "grad_norm": 4.254117488861084, + "learning_rate": 4.81559583637821e-05, + "loss": 0.4331, + "num_input_tokens_seen": 2150240, + "step": 8040 + }, + { + "epoch": 2.108781127129751, + "grad_norm": 2.3760135173797607, + "learning_rate": 4.8151644790006077e-05, + "loss": 0.4282, + "num_input_tokens_seen": 2151376, + "step": 8045 + }, + { + "epoch": 2.1100917431192663, + "grad_norm": 2.322392225265503, + "learning_rate": 4.8147326370630796e-05, + "loss": 0.4211, + "num_input_tokens_seen": 2152512, + "step": 8050 + }, + { + "epoch": 2.111402359108781, + "grad_norm": 2.3731188774108887, + "learning_rate": 4.814300310656011e-05, + "loss": 0.5344, + "num_input_tokens_seen": 2153856, + "step": 8055 + }, + { + "epoch": 2.112712975098296, + "grad_norm": 1.3517898321151733, + "learning_rate": 4.8138674998698854e-05, + "loss": 0.4908, + "num_input_tokens_seen": 2155040, + "step": 8060 + }, + { + "epoch": 2.114023591087811, + "grad_norm": 2.1092326641082764, + "learning_rate": 4.8134342047952906e-05, + "loss": 0.4552, + "num_input_tokens_seen": 2156400, + "step": 8065 + }, + { + "epoch": 2.1153342070773262, + "grad_norm": 1.2329596281051636, + "learning_rate": 4.8130004255229146e-05, + "loss": 0.5006, + "num_input_tokens_seen": 2157856, + "step": 8070 + }, + { + "epoch": 2.1166448230668413, + "grad_norm": 1.4912598133087158, + "learning_rate": 4.8125661621435445e-05, + "loss": 0.6507, + "num_input_tokens_seen": 2158864, + "step": 8075 + }, + { + "epoch": 2.1179554390563564, + "grad_norm": 2.3075859546661377, + "learning_rate": 4.8121314147480736e-05, + "loss": 0.332, + "num_input_tokens_seen": 2161936, + "step": 8080 + }, + { + "epoch": 2.1192660550458715, + "grad_norm": 1.5610333681106567, + "learning_rate": 4.811696183427493e-05, + "loss": 0.4584, + "num_input_tokens_seen": 2163904, + "step": 8085 + }, + { + "epoch": 2.1205766710353866, + "grad_norm": 1.7475708723068237, + "learning_rate": 4.811260468272895e-05, + "loss": 0.5853, + "num_input_tokens_seen": 2165296, + "step": 8090 + }, + { + "epoch": 2.1218872870249017, + "grad_norm": 2.304172992706299, + "learning_rate": 4.810824269375475e-05, + "loss": 0.4304, + "num_input_tokens_seen": 2166912, + "step": 8095 + }, + { + "epoch": 2.123197903014417, + "grad_norm": 2.4857254028320312, + "learning_rate": 4.810387586826527e-05, + "loss": 0.3922, + "num_input_tokens_seen": 2168320, + "step": 8100 + }, + { + "epoch": 2.124508519003932, + "grad_norm": 3.265610456466675, + "learning_rate": 4.80995042071745e-05, + "loss": 0.4396, + "num_input_tokens_seen": 2169568, + "step": 8105 + }, + { + "epoch": 2.125819134993447, + "grad_norm": 1.111757755279541, + "learning_rate": 4.809512771139741e-05, + "loss": 0.4655, + "num_input_tokens_seen": 2170640, + "step": 8110 + }, + { + "epoch": 2.127129750982962, + "grad_norm": 5.426815509796143, + "learning_rate": 4.809074638184999e-05, + "loss": 0.4092, + "num_input_tokens_seen": 2171440, + "step": 8115 + }, + { + "epoch": 2.128440366972477, + "grad_norm": 2.146864652633667, + "learning_rate": 4.808636021944927e-05, + "loss": 0.4761, + "num_input_tokens_seen": 2172544, + "step": 8120 + }, + { + "epoch": 2.1297509829619923, + "grad_norm": 5.888105392456055, + "learning_rate": 4.808196922511323e-05, + "loss": 0.7945, + "num_input_tokens_seen": 2173696, + "step": 8125 + }, + { + "epoch": 2.1310615989515074, + "grad_norm": 4.012242794036865, + "learning_rate": 4.8077573399760914e-05, + "loss": 0.3948, + "num_input_tokens_seen": 2175120, + "step": 8130 + }, + { + "epoch": 2.1323722149410225, + "grad_norm": 3.602001667022705, + "learning_rate": 4.807317274431236e-05, + "loss": 0.4753, + "num_input_tokens_seen": 2176272, + "step": 8135 + }, + { + "epoch": 2.133682830930537, + "grad_norm": 2.2850303649902344, + "learning_rate": 4.806876725968862e-05, + "loss": 0.3779, + "num_input_tokens_seen": 2177456, + "step": 8140 + }, + { + "epoch": 2.134993446920052, + "grad_norm": 15.358515739440918, + "learning_rate": 4.806435694681175e-05, + "loss": 0.2517, + "num_input_tokens_seen": 2178528, + "step": 8145 + }, + { + "epoch": 2.1363040629095673, + "grad_norm": 4.475828170776367, + "learning_rate": 4.805994180660483e-05, + "loss": 0.5057, + "num_input_tokens_seen": 2179744, + "step": 8150 + }, + { + "epoch": 2.1376146788990824, + "grad_norm": 1.6570570468902588, + "learning_rate": 4.805552183999193e-05, + "loss": 1.4564, + "num_input_tokens_seen": 2180880, + "step": 8155 + }, + { + "epoch": 2.1389252948885975, + "grad_norm": 3.177985191345215, + "learning_rate": 4.8051097047898144e-05, + "loss": 0.5547, + "num_input_tokens_seen": 2182032, + "step": 8160 + }, + { + "epoch": 2.1402359108781126, + "grad_norm": 4.716887950897217, + "learning_rate": 4.8046667431249585e-05, + "loss": 0.5519, + "num_input_tokens_seen": 2183312, + "step": 8165 + }, + { + "epoch": 2.1415465268676277, + "grad_norm": 3.588005781173706, + "learning_rate": 4.804223299097334e-05, + "loss": 0.4055, + "num_input_tokens_seen": 2184752, + "step": 8170 + }, + { + "epoch": 2.142857142857143, + "grad_norm": 2.066441059112549, + "learning_rate": 4.803779372799755e-05, + "loss": 0.3413, + "num_input_tokens_seen": 2186240, + "step": 8175 + }, + { + "epoch": 2.144167758846658, + "grad_norm": 2.063551902770996, + "learning_rate": 4.803334964325134e-05, + "loss": 0.3944, + "num_input_tokens_seen": 2187568, + "step": 8180 + }, + { + "epoch": 2.145478374836173, + "grad_norm": 6.353747844696045, + "learning_rate": 4.802890073766485e-05, + "loss": 0.4114, + "num_input_tokens_seen": 2188848, + "step": 8185 + }, + { + "epoch": 2.146788990825688, + "grad_norm": 2.976156234741211, + "learning_rate": 4.802444701216923e-05, + "loss": 0.4217, + "num_input_tokens_seen": 2190304, + "step": 8190 + }, + { + "epoch": 2.148099606815203, + "grad_norm": 1.180292010307312, + "learning_rate": 4.801998846769662e-05, + "loss": 0.4521, + "num_input_tokens_seen": 2191952, + "step": 8195 + }, + { + "epoch": 2.1494102228047183, + "grad_norm": 13.654682159423828, + "learning_rate": 4.801552510518021e-05, + "loss": 1.5295, + "num_input_tokens_seen": 2192992, + "step": 8200 + }, + { + "epoch": 2.1507208387942334, + "grad_norm": 3.4151644706726074, + "learning_rate": 4.801105692555415e-05, + "loss": 0.4805, + "num_input_tokens_seen": 2194048, + "step": 8205 + }, + { + "epoch": 2.1520314547837485, + "grad_norm": 8.468122482299805, + "learning_rate": 4.800658392975366e-05, + "loss": 0.7215, + "num_input_tokens_seen": 2195072, + "step": 8210 + }, + { + "epoch": 2.1533420707732636, + "grad_norm": 1.9149341583251953, + "learning_rate": 4.800210611871488e-05, + "loss": 0.371, + "num_input_tokens_seen": 2196224, + "step": 8215 + }, + { + "epoch": 2.1546526867627787, + "grad_norm": 1.5821490287780762, + "learning_rate": 4.7997623493375054e-05, + "loss": 0.5522, + "num_input_tokens_seen": 2197520, + "step": 8220 + }, + { + "epoch": 2.1559633027522938, + "grad_norm": 1.8950865268707275, + "learning_rate": 4.7993136054672356e-05, + "loss": 0.4621, + "num_input_tokens_seen": 2198816, + "step": 8225 + }, + { + "epoch": 2.157273918741809, + "grad_norm": 1.9299427270889282, + "learning_rate": 4.798864380354601e-05, + "loss": 0.3738, + "num_input_tokens_seen": 2200496, + "step": 8230 + }, + { + "epoch": 2.1585845347313235, + "grad_norm": 12.173334121704102, + "learning_rate": 4.798414674093624e-05, + "loss": 0.4119, + "num_input_tokens_seen": 2201616, + "step": 8235 + }, + { + "epoch": 2.1598951507208386, + "grad_norm": 3.649383068084717, + "learning_rate": 4.7979644867784276e-05, + "loss": 0.3161, + "num_input_tokens_seen": 2203056, + "step": 8240 + }, + { + "epoch": 2.1612057667103537, + "grad_norm": 1.3928524255752563, + "learning_rate": 4.797513818503234e-05, + "loss": 0.4704, + "num_input_tokens_seen": 2204272, + "step": 8245 + }, + { + "epoch": 2.162516382699869, + "grad_norm": 1.9166818857192993, + "learning_rate": 4.797062669362369e-05, + "loss": 0.626, + "num_input_tokens_seen": 2206144, + "step": 8250 + }, + { + "epoch": 2.163826998689384, + "grad_norm": 3.8071165084838867, + "learning_rate": 4.796611039450255e-05, + "loss": 1.1279, + "num_input_tokens_seen": 2207072, + "step": 8255 + }, + { + "epoch": 2.165137614678899, + "grad_norm": 2.874732255935669, + "learning_rate": 4.796158928861419e-05, + "loss": 0.5468, + "num_input_tokens_seen": 2208480, + "step": 8260 + }, + { + "epoch": 2.166448230668414, + "grad_norm": 2.723475933074951, + "learning_rate": 4.7957063376904875e-05, + "loss": 0.5238, + "num_input_tokens_seen": 2209696, + "step": 8265 + }, + { + "epoch": 2.167758846657929, + "grad_norm": 1.842960238456726, + "learning_rate": 4.7952532660321854e-05, + "loss": 0.6834, + "num_input_tokens_seen": 2210944, + "step": 8270 + }, + { + "epoch": 2.1690694626474443, + "grad_norm": 1.5332003831863403, + "learning_rate": 4.794799713981341e-05, + "loss": 0.4662, + "num_input_tokens_seen": 2212144, + "step": 8275 + }, + { + "epoch": 2.1703800786369594, + "grad_norm": 1.7814244031906128, + "learning_rate": 4.7943456816328815e-05, + "loss": 0.3628, + "num_input_tokens_seen": 2213600, + "step": 8280 + }, + { + "epoch": 2.1716906946264745, + "grad_norm": 5.339340686798096, + "learning_rate": 4.7938911690818347e-05, + "loss": 0.4716, + "num_input_tokens_seen": 2214848, + "step": 8285 + }, + { + "epoch": 2.1730013106159896, + "grad_norm": 2.955348014831543, + "learning_rate": 4.79343617642333e-05, + "loss": 0.4634, + "num_input_tokens_seen": 2216256, + "step": 8290 + }, + { + "epoch": 2.1743119266055047, + "grad_norm": 2.4864439964294434, + "learning_rate": 4.792980703752597e-05, + "loss": 0.4191, + "num_input_tokens_seen": 2217616, + "step": 8295 + }, + { + "epoch": 2.1756225425950197, + "grad_norm": 2.144582986831665, + "learning_rate": 4.792524751164964e-05, + "loss": 0.9099, + "num_input_tokens_seen": 2218656, + "step": 8300 + }, + { + "epoch": 2.176933158584535, + "grad_norm": 6.070431232452393, + "learning_rate": 4.792068318755861e-05, + "loss": 0.5017, + "num_input_tokens_seen": 2219728, + "step": 8305 + }, + { + "epoch": 2.17824377457405, + "grad_norm": 1.133120059967041, + "learning_rate": 4.791611406620821e-05, + "loss": 0.4172, + "num_input_tokens_seen": 2221248, + "step": 8310 + }, + { + "epoch": 2.179554390563565, + "grad_norm": 2.2382233142852783, + "learning_rate": 4.7911540148554725e-05, + "loss": 0.353, + "num_input_tokens_seen": 2222336, + "step": 8315 + }, + { + "epoch": 2.18086500655308, + "grad_norm": 14.688268661499023, + "learning_rate": 4.790696143555546e-05, + "loss": 0.5035, + "num_input_tokens_seen": 2223392, + "step": 8320 + }, + { + "epoch": 2.1821756225425952, + "grad_norm": 2.11450457572937, + "learning_rate": 4.790237792816877e-05, + "loss": 0.3827, + "num_input_tokens_seen": 2224672, + "step": 8325 + }, + { + "epoch": 2.18348623853211, + "grad_norm": 1.1013692617416382, + "learning_rate": 4.789778962735394e-05, + "loss": 0.4287, + "num_input_tokens_seen": 2226272, + "step": 8330 + }, + { + "epoch": 2.184796854521625, + "grad_norm": 1.5747638940811157, + "learning_rate": 4.78931965340713e-05, + "loss": 0.4104, + "num_input_tokens_seen": 2227536, + "step": 8335 + }, + { + "epoch": 2.18610747051114, + "grad_norm": 1.5392813682556152, + "learning_rate": 4.78885986492822e-05, + "loss": 0.4338, + "num_input_tokens_seen": 2228736, + "step": 8340 + }, + { + "epoch": 2.187418086500655, + "grad_norm": 1.7379587888717651, + "learning_rate": 4.788399597394894e-05, + "loss": 0.6663, + "num_input_tokens_seen": 2230000, + "step": 8345 + }, + { + "epoch": 2.1887287024901703, + "grad_norm": 1.7493726015090942, + "learning_rate": 4.787938850903486e-05, + "loss": 0.5389, + "num_input_tokens_seen": 2231328, + "step": 8350 + }, + { + "epoch": 2.1900393184796854, + "grad_norm": 2.691771984100342, + "learning_rate": 4.78747762555043e-05, + "loss": 0.4094, + "num_input_tokens_seen": 2232608, + "step": 8355 + }, + { + "epoch": 2.1913499344692005, + "grad_norm": 1.9138704538345337, + "learning_rate": 4.7870159214322594e-05, + "loss": 0.5245, + "num_input_tokens_seen": 2234000, + "step": 8360 + }, + { + "epoch": 2.1926605504587156, + "grad_norm": 1.3855538368225098, + "learning_rate": 4.7865537386456084e-05, + "loss": 0.3146, + "num_input_tokens_seen": 2235312, + "step": 8365 + }, + { + "epoch": 2.1939711664482306, + "grad_norm": 1.9233882427215576, + "learning_rate": 4.7860910772872104e-05, + "loss": 0.5369, + "num_input_tokens_seen": 2236384, + "step": 8370 + }, + { + "epoch": 2.1952817824377457, + "grad_norm": 3.1481828689575195, + "learning_rate": 4.7856279374539e-05, + "loss": 0.2836, + "num_input_tokens_seen": 2237824, + "step": 8375 + }, + { + "epoch": 2.196592398427261, + "grad_norm": 3.9105100631713867, + "learning_rate": 4.785164319242612e-05, + "loss": 0.3788, + "num_input_tokens_seen": 2239120, + "step": 8380 + }, + { + "epoch": 2.197903014416776, + "grad_norm": 1.7040809392929077, + "learning_rate": 4.78470022275038e-05, + "loss": 0.5132, + "num_input_tokens_seen": 2240576, + "step": 8385 + }, + { + "epoch": 2.199213630406291, + "grad_norm": 2.898595094680786, + "learning_rate": 4.784235648074339e-05, + "loss": 0.4987, + "num_input_tokens_seen": 2242080, + "step": 8390 + }, + { + "epoch": 2.200524246395806, + "grad_norm": 2.4399449825286865, + "learning_rate": 4.783770595311724e-05, + "loss": 0.6186, + "num_input_tokens_seen": 2243648, + "step": 8395 + }, + { + "epoch": 2.2018348623853212, + "grad_norm": 0.9274146556854248, + "learning_rate": 4.78330506455987e-05, + "loss": 0.3931, + "num_input_tokens_seen": 2245072, + "step": 8400 + }, + { + "epoch": 2.2031454783748363, + "grad_norm": 4.655673027038574, + "learning_rate": 4.78283905591621e-05, + "loss": 0.5596, + "num_input_tokens_seen": 2246176, + "step": 8405 + }, + { + "epoch": 2.2044560943643514, + "grad_norm": 3.0477261543273926, + "learning_rate": 4.78237256947828e-05, + "loss": 0.4133, + "num_input_tokens_seen": 2248048, + "step": 8410 + }, + { + "epoch": 2.2057667103538665, + "grad_norm": 1.5345577001571655, + "learning_rate": 4.7819056053437165e-05, + "loss": 0.4581, + "num_input_tokens_seen": 2249360, + "step": 8415 + }, + { + "epoch": 2.2070773263433816, + "grad_norm": 1.8313344717025757, + "learning_rate": 4.7814381636102515e-05, + "loss": 0.7085, + "num_input_tokens_seen": 2251600, + "step": 8420 + }, + { + "epoch": 2.2083879423328963, + "grad_norm": 7.4522905349731445, + "learning_rate": 4.780970244375721e-05, + "loss": 0.7039, + "num_input_tokens_seen": 2252368, + "step": 8425 + }, + { + "epoch": 2.2096985583224114, + "grad_norm": 4.7604570388793945, + "learning_rate": 4.780501847738059e-05, + "loss": 0.3485, + "num_input_tokens_seen": 2253376, + "step": 8430 + }, + { + "epoch": 2.2110091743119265, + "grad_norm": 3.349076271057129, + "learning_rate": 4.7800329737953016e-05, + "loss": 0.4996, + "num_input_tokens_seen": 2254688, + "step": 8435 + }, + { + "epoch": 2.2123197903014415, + "grad_norm": 2.1537106037139893, + "learning_rate": 4.779563622645582e-05, + "loss": 0.5107, + "num_input_tokens_seen": 2255888, + "step": 8440 + }, + { + "epoch": 2.2136304062909566, + "grad_norm": 4.800303936004639, + "learning_rate": 4.779093794387135e-05, + "loss": 0.3972, + "num_input_tokens_seen": 2256992, + "step": 8445 + }, + { + "epoch": 2.2149410222804717, + "grad_norm": 2.2869319915771484, + "learning_rate": 4.778623489118295e-05, + "loss": 0.7009, + "num_input_tokens_seen": 2258128, + "step": 8450 + }, + { + "epoch": 2.216251638269987, + "grad_norm": 1.9233044385910034, + "learning_rate": 4.7781527069374957e-05, + "loss": 0.6246, + "num_input_tokens_seen": 2259440, + "step": 8455 + }, + { + "epoch": 2.217562254259502, + "grad_norm": 3.21962833404541, + "learning_rate": 4.777681447943271e-05, + "loss": 0.3958, + "num_input_tokens_seen": 2260512, + "step": 8460 + }, + { + "epoch": 2.218872870249017, + "grad_norm": 2.703718900680542, + "learning_rate": 4.777209712234255e-05, + "loss": 0.5778, + "num_input_tokens_seen": 2261824, + "step": 8465 + }, + { + "epoch": 2.220183486238532, + "grad_norm": 4.672186851501465, + "learning_rate": 4.776737499909181e-05, + "loss": 0.6054, + "num_input_tokens_seen": 2263136, + "step": 8470 + }, + { + "epoch": 2.221494102228047, + "grad_norm": 1.9996154308319092, + "learning_rate": 4.776264811066882e-05, + "loss": 0.5748, + "num_input_tokens_seen": 2264448, + "step": 8475 + }, + { + "epoch": 2.2228047182175623, + "grad_norm": 0.8898599743843079, + "learning_rate": 4.7757916458062905e-05, + "loss": 0.4024, + "num_input_tokens_seen": 2265632, + "step": 8480 + }, + { + "epoch": 2.2241153342070774, + "grad_norm": 2.5459649562835693, + "learning_rate": 4.7753180042264404e-05, + "loss": 0.5236, + "num_input_tokens_seen": 2266912, + "step": 8485 + }, + { + "epoch": 2.2254259501965925, + "grad_norm": 2.2212188243865967, + "learning_rate": 4.7748438864264636e-05, + "loss": 0.4421, + "num_input_tokens_seen": 2268144, + "step": 8490 + }, + { + "epoch": 2.2267365661861076, + "grad_norm": 2.7964818477630615, + "learning_rate": 4.774369292505592e-05, + "loss": 0.5311, + "num_input_tokens_seen": 2269344, + "step": 8495 + }, + { + "epoch": 2.2280471821756227, + "grad_norm": 1.8792396783828735, + "learning_rate": 4.773894222563157e-05, + "loss": 0.4978, + "num_input_tokens_seen": 2270720, + "step": 8500 + }, + { + "epoch": 2.229357798165138, + "grad_norm": 1.801830768585205, + "learning_rate": 4.7734186766985904e-05, + "loss": 0.5423, + "num_input_tokens_seen": 2272304, + "step": 8505 + }, + { + "epoch": 2.230668414154653, + "grad_norm": 1.7618169784545898, + "learning_rate": 4.772942655011423e-05, + "loss": 0.3911, + "num_input_tokens_seen": 2273728, + "step": 8510 + }, + { + "epoch": 2.231979030144168, + "grad_norm": 2.1464335918426514, + "learning_rate": 4.772466157601285e-05, + "loss": 0.4345, + "num_input_tokens_seen": 2275104, + "step": 8515 + }, + { + "epoch": 2.2332896461336826, + "grad_norm": 3.688323736190796, + "learning_rate": 4.771989184567907e-05, + "loss": 0.2943, + "num_input_tokens_seen": 2276096, + "step": 8520 + }, + { + "epoch": 2.2346002621231977, + "grad_norm": 1.7137174606323242, + "learning_rate": 4.7715117360111185e-05, + "loss": 0.5724, + "num_input_tokens_seen": 2277424, + "step": 8525 + }, + { + "epoch": 2.235910878112713, + "grad_norm": 4.487665176391602, + "learning_rate": 4.771033812030849e-05, + "loss": 0.6167, + "num_input_tokens_seen": 2278432, + "step": 8530 + }, + { + "epoch": 2.237221494102228, + "grad_norm": 4.663251876831055, + "learning_rate": 4.7705554127271254e-05, + "loss": 0.331, + "num_input_tokens_seen": 2279456, + "step": 8535 + }, + { + "epoch": 2.238532110091743, + "grad_norm": 2.6544156074523926, + "learning_rate": 4.770076538200078e-05, + "loss": 0.4625, + "num_input_tokens_seen": 2280896, + "step": 8540 + }, + { + "epoch": 2.239842726081258, + "grad_norm": 3.322988748550415, + "learning_rate": 4.769597188549932e-05, + "loss": 0.4516, + "num_input_tokens_seen": 2282064, + "step": 8545 + }, + { + "epoch": 2.241153342070773, + "grad_norm": 2.079111099243164, + "learning_rate": 4.769117363877018e-05, + "loss": 0.5636, + "num_input_tokens_seen": 2283232, + "step": 8550 + }, + { + "epoch": 2.2424639580602883, + "grad_norm": 4.069112300872803, + "learning_rate": 4.76863706428176e-05, + "loss": 0.5371, + "num_input_tokens_seen": 2284896, + "step": 8555 + }, + { + "epoch": 2.2437745740498034, + "grad_norm": 4.388924598693848, + "learning_rate": 4.7681562898646836e-05, + "loss": 0.4577, + "num_input_tokens_seen": 2286176, + "step": 8560 + }, + { + "epoch": 2.2450851900393185, + "grad_norm": 2.598841905593872, + "learning_rate": 4.767675040726416e-05, + "loss": 0.4067, + "num_input_tokens_seen": 2287984, + "step": 8565 + }, + { + "epoch": 2.2463958060288336, + "grad_norm": 4.846996784210205, + "learning_rate": 4.76719331696768e-05, + "loss": 0.3731, + "num_input_tokens_seen": 2289008, + "step": 8570 + }, + { + "epoch": 2.2477064220183487, + "grad_norm": 2.3520047664642334, + "learning_rate": 4.7667111186893e-05, + "loss": 1.5211, + "num_input_tokens_seen": 2290208, + "step": 8575 + }, + { + "epoch": 2.249017038007864, + "grad_norm": 5.412135601043701, + "learning_rate": 4.766228445992199e-05, + "loss": 0.4291, + "num_input_tokens_seen": 2291584, + "step": 8580 + }, + { + "epoch": 2.250327653997379, + "grad_norm": 5.0524702072143555, + "learning_rate": 4.7657452989774e-05, + "loss": 0.3092, + "num_input_tokens_seen": 2292528, + "step": 8585 + }, + { + "epoch": 2.251638269986894, + "grad_norm": 2.832069158554077, + "learning_rate": 4.765261677746026e-05, + "loss": 0.6718, + "num_input_tokens_seen": 2293936, + "step": 8590 + }, + { + "epoch": 2.252948885976409, + "grad_norm": 1.995400071144104, + "learning_rate": 4.764777582399296e-05, + "loss": 0.4144, + "num_input_tokens_seen": 2295104, + "step": 8595 + }, + { + "epoch": 2.254259501965924, + "grad_norm": 1.3416434526443481, + "learning_rate": 4.764293013038531e-05, + "loss": 0.5045, + "num_input_tokens_seen": 2296400, + "step": 8600 + }, + { + "epoch": 2.255570117955439, + "grad_norm": 3.0714972019195557, + "learning_rate": 4.7638079697651514e-05, + "loss": 0.4748, + "num_input_tokens_seen": 2297648, + "step": 8605 + }, + { + "epoch": 2.2568807339449544, + "grad_norm": 1.6846095323562622, + "learning_rate": 4.763322452680675e-05, + "loss": 0.2625, + "num_input_tokens_seen": 2299008, + "step": 8610 + }, + { + "epoch": 2.258191349934469, + "grad_norm": 2.9389424324035645, + "learning_rate": 4.76283646188672e-05, + "loss": 0.5526, + "num_input_tokens_seen": 2300368, + "step": 8615 + }, + { + "epoch": 2.259501965923984, + "grad_norm": 2.863435745239258, + "learning_rate": 4.7623499974850036e-05, + "loss": 0.6132, + "num_input_tokens_seen": 2301472, + "step": 8620 + }, + { + "epoch": 2.260812581913499, + "grad_norm": 8.19389820098877, + "learning_rate": 4.7618630595773424e-05, + "loss": 0.4144, + "num_input_tokens_seen": 2302880, + "step": 8625 + }, + { + "epoch": 2.2621231979030143, + "grad_norm": 1.2736130952835083, + "learning_rate": 4.7613756482656505e-05, + "loss": 1.0064, + "num_input_tokens_seen": 2304416, + "step": 8630 + }, + { + "epoch": 2.2634338138925294, + "grad_norm": 2.1778197288513184, + "learning_rate": 4.760887763651942e-05, + "loss": 0.4614, + "num_input_tokens_seen": 2305904, + "step": 8635 + }, + { + "epoch": 2.2647444298820445, + "grad_norm": 3.1024866104125977, + "learning_rate": 4.7603994058383315e-05, + "loss": 0.4023, + "num_input_tokens_seen": 2307312, + "step": 8640 + }, + { + "epoch": 2.2660550458715596, + "grad_norm": 13.914787292480469, + "learning_rate": 4.759910574927032e-05, + "loss": 0.6353, + "num_input_tokens_seen": 2308336, + "step": 8645 + }, + { + "epoch": 2.2673656618610747, + "grad_norm": 2.3023335933685303, + "learning_rate": 4.759421271020353e-05, + "loss": 0.5366, + "num_input_tokens_seen": 2309648, + "step": 8650 + }, + { + "epoch": 2.26867627785059, + "grad_norm": 1.2729350328445435, + "learning_rate": 4.758931494220707e-05, + "loss": 0.5731, + "num_input_tokens_seen": 2310784, + "step": 8655 + }, + { + "epoch": 2.269986893840105, + "grad_norm": 2.359269618988037, + "learning_rate": 4.7584412446306e-05, + "loss": 0.9007, + "num_input_tokens_seen": 2311616, + "step": 8660 + }, + { + "epoch": 2.27129750982962, + "grad_norm": 1.1600197553634644, + "learning_rate": 4.7579505223526446e-05, + "loss": 0.6205, + "num_input_tokens_seen": 2313072, + "step": 8665 + }, + { + "epoch": 2.272608125819135, + "grad_norm": 13.768017768859863, + "learning_rate": 4.757459327489546e-05, + "loss": 0.5678, + "num_input_tokens_seen": 2314160, + "step": 8670 + }, + { + "epoch": 2.27391874180865, + "grad_norm": 3.007603645324707, + "learning_rate": 4.756967660144111e-05, + "loss": 0.4351, + "num_input_tokens_seen": 2315744, + "step": 8675 + }, + { + "epoch": 2.2752293577981653, + "grad_norm": 0.9228350520133972, + "learning_rate": 4.756475520419244e-05, + "loss": 0.421, + "num_input_tokens_seen": 2317536, + "step": 8680 + }, + { + "epoch": 2.2765399737876804, + "grad_norm": 8.167559623718262, + "learning_rate": 4.7559829084179494e-05, + "loss": 0.479, + "num_input_tokens_seen": 2318656, + "step": 8685 + }, + { + "epoch": 2.2778505897771955, + "grad_norm": 3.600985288619995, + "learning_rate": 4.755489824243331e-05, + "loss": 0.7479, + "num_input_tokens_seen": 2320448, + "step": 8690 + }, + { + "epoch": 2.2791612057667106, + "grad_norm": 2.870551586151123, + "learning_rate": 4.754996267998589e-05, + "loss": 0.4415, + "num_input_tokens_seen": 2321680, + "step": 8695 + }, + { + "epoch": 2.280471821756225, + "grad_norm": 7.973691463470459, + "learning_rate": 4.754502239787024e-05, + "loss": 0.552, + "num_input_tokens_seen": 2323184, + "step": 8700 + }, + { + "epoch": 2.2817824377457407, + "grad_norm": 2.9065732955932617, + "learning_rate": 4.7540077397120365e-05, + "loss": 0.6441, + "num_input_tokens_seen": 2324464, + "step": 8705 + }, + { + "epoch": 2.2830930537352554, + "grad_norm": 4.558143138885498, + "learning_rate": 4.7535127678771233e-05, + "loss": 0.4832, + "num_input_tokens_seen": 2325344, + "step": 8710 + }, + { + "epoch": 2.2844036697247705, + "grad_norm": 2.439424514770508, + "learning_rate": 4.7530173243858814e-05, + "loss": 0.5589, + "num_input_tokens_seen": 2326592, + "step": 8715 + }, + { + "epoch": 2.2857142857142856, + "grad_norm": 2.7859890460968018, + "learning_rate": 4.7525214093420065e-05, + "loss": 0.3721, + "num_input_tokens_seen": 2327760, + "step": 8720 + }, + { + "epoch": 2.2870249017038007, + "grad_norm": 1.2331644296646118, + "learning_rate": 4.7520250228492934e-05, + "loss": 0.5242, + "num_input_tokens_seen": 2329488, + "step": 8725 + }, + { + "epoch": 2.288335517693316, + "grad_norm": 4.791596412658691, + "learning_rate": 4.751528165011633e-05, + "loss": 0.3321, + "num_input_tokens_seen": 2330848, + "step": 8730 + }, + { + "epoch": 2.289646133682831, + "grad_norm": 2.1896374225616455, + "learning_rate": 4.7510308359330194e-05, + "loss": 0.6436, + "num_input_tokens_seen": 2332432, + "step": 8735 + }, + { + "epoch": 2.290956749672346, + "grad_norm": 2.393986940383911, + "learning_rate": 4.7505330357175406e-05, + "loss": 0.7761, + "num_input_tokens_seen": 2333520, + "step": 8740 + }, + { + "epoch": 2.292267365661861, + "grad_norm": 1.169521689414978, + "learning_rate": 4.750034764469386e-05, + "loss": 0.5483, + "num_input_tokens_seen": 2334816, + "step": 8745 + }, + { + "epoch": 2.293577981651376, + "grad_norm": 2.8667752742767334, + "learning_rate": 4.749536022292842e-05, + "loss": 0.6053, + "num_input_tokens_seen": 2336096, + "step": 8750 + }, + { + "epoch": 2.2948885976408913, + "grad_norm": 3.945732355117798, + "learning_rate": 4.7490368092922964e-05, + "loss": 0.4442, + "num_input_tokens_seen": 2337728, + "step": 8755 + }, + { + "epoch": 2.2961992136304064, + "grad_norm": 0.8940362930297852, + "learning_rate": 4.748537125572233e-05, + "loss": 0.3432, + "num_input_tokens_seen": 2339808, + "step": 8760 + }, + { + "epoch": 2.2975098296199215, + "grad_norm": 2.121189594268799, + "learning_rate": 4.748036971237234e-05, + "loss": 0.3492, + "num_input_tokens_seen": 2340800, + "step": 8765 + }, + { + "epoch": 2.2988204456094365, + "grad_norm": 9.362908363342285, + "learning_rate": 4.747536346391981e-05, + "loss": 0.4072, + "num_input_tokens_seen": 2342208, + "step": 8770 + }, + { + "epoch": 2.3001310615989516, + "grad_norm": 2.2711496353149414, + "learning_rate": 4.747035251141254e-05, + "loss": 0.3608, + "num_input_tokens_seen": 2343264, + "step": 8775 + }, + { + "epoch": 2.3014416775884667, + "grad_norm": 7.567618370056152, + "learning_rate": 4.746533685589932e-05, + "loss": 0.5081, + "num_input_tokens_seen": 2344432, + "step": 8780 + }, + { + "epoch": 2.302752293577982, + "grad_norm": 2.0208818912506104, + "learning_rate": 4.746031649842991e-05, + "loss": 0.525, + "num_input_tokens_seen": 2345792, + "step": 8785 + }, + { + "epoch": 2.304062909567497, + "grad_norm": 5.413013935089111, + "learning_rate": 4.745529144005507e-05, + "loss": 0.4059, + "num_input_tokens_seen": 2346912, + "step": 8790 + }, + { + "epoch": 2.3053735255570116, + "grad_norm": 1.865442156791687, + "learning_rate": 4.745026168182654e-05, + "loss": 0.6613, + "num_input_tokens_seen": 2348368, + "step": 8795 + }, + { + "epoch": 2.3066841415465267, + "grad_norm": 2.5159804821014404, + "learning_rate": 4.744522722479701e-05, + "loss": 0.4625, + "num_input_tokens_seen": 2349584, + "step": 8800 + }, + { + "epoch": 2.3079947575360418, + "grad_norm": 7.450370788574219, + "learning_rate": 4.7440188070020217e-05, + "loss": 0.423, + "num_input_tokens_seen": 2350592, + "step": 8805 + }, + { + "epoch": 2.309305373525557, + "grad_norm": 2.609569787979126, + "learning_rate": 4.7435144218550834e-05, + "loss": 0.4169, + "num_input_tokens_seen": 2351744, + "step": 8810 + }, + { + "epoch": 2.310615989515072, + "grad_norm": 3.6855661869049072, + "learning_rate": 4.7430095671444526e-05, + "loss": 0.682, + "num_input_tokens_seen": 2352832, + "step": 8815 + }, + { + "epoch": 2.311926605504587, + "grad_norm": 1.9334899187088013, + "learning_rate": 4.742504242975795e-05, + "loss": 0.4406, + "num_input_tokens_seen": 2354160, + "step": 8820 + }, + { + "epoch": 2.313237221494102, + "grad_norm": 1.7300341129302979, + "learning_rate": 4.7419984494548744e-05, + "loss": 0.3354, + "num_input_tokens_seen": 2355360, + "step": 8825 + }, + { + "epoch": 2.3145478374836173, + "grad_norm": 1.7558518648147583, + "learning_rate": 4.7414921866875524e-05, + "loss": 0.486, + "num_input_tokens_seen": 2356528, + "step": 8830 + }, + { + "epoch": 2.3158584534731324, + "grad_norm": 5.340721130371094, + "learning_rate": 4.740985454779788e-05, + "loss": 0.5926, + "num_input_tokens_seen": 2357824, + "step": 8835 + }, + { + "epoch": 2.3171690694626474, + "grad_norm": 2.5842361450195312, + "learning_rate": 4.740478253837641e-05, + "loss": 0.3646, + "num_input_tokens_seen": 2359168, + "step": 8840 + }, + { + "epoch": 2.3184796854521625, + "grad_norm": 2.6657309532165527, + "learning_rate": 4.739970583967266e-05, + "loss": 0.6183, + "num_input_tokens_seen": 2360224, + "step": 8845 + }, + { + "epoch": 2.3197903014416776, + "grad_norm": 4.682046413421631, + "learning_rate": 4.739462445274919e-05, + "loss": 0.5574, + "num_input_tokens_seen": 2361424, + "step": 8850 + }, + { + "epoch": 2.3211009174311927, + "grad_norm": 2.787841320037842, + "learning_rate": 4.738953837866951e-05, + "loss": 0.5188, + "num_input_tokens_seen": 2362624, + "step": 8855 + }, + { + "epoch": 2.322411533420708, + "grad_norm": 1.3629138469696045, + "learning_rate": 4.738444761849813e-05, + "loss": 0.4662, + "num_input_tokens_seen": 2363808, + "step": 8860 + }, + { + "epoch": 2.323722149410223, + "grad_norm": 1.5494775772094727, + "learning_rate": 4.7379352173300553e-05, + "loss": 0.5665, + "num_input_tokens_seen": 2365056, + "step": 8865 + }, + { + "epoch": 2.325032765399738, + "grad_norm": 2.9474427700042725, + "learning_rate": 4.737425204414322e-05, + "loss": 0.7005, + "num_input_tokens_seen": 2366544, + "step": 8870 + }, + { + "epoch": 2.326343381389253, + "grad_norm": 2.0268454551696777, + "learning_rate": 4.73691472320936e-05, + "loss": 0.2509, + "num_input_tokens_seen": 2368448, + "step": 8875 + }, + { + "epoch": 2.327653997378768, + "grad_norm": 2.382610559463501, + "learning_rate": 4.736403773822012e-05, + "loss": 0.6353, + "num_input_tokens_seen": 2369568, + "step": 8880 + }, + { + "epoch": 2.3289646133682833, + "grad_norm": 2.318959951400757, + "learning_rate": 4.735892356359217e-05, + "loss": 0.279, + "num_input_tokens_seen": 2370672, + "step": 8885 + }, + { + "epoch": 2.330275229357798, + "grad_norm": 2.3841333389282227, + "learning_rate": 4.735380470928017e-05, + "loss": 0.3991, + "num_input_tokens_seen": 2373248, + "step": 8890 + }, + { + "epoch": 2.331585845347313, + "grad_norm": 4.938017845153809, + "learning_rate": 4.7348681176355456e-05, + "loss": 0.5953, + "num_input_tokens_seen": 2374400, + "step": 8895 + }, + { + "epoch": 2.332896461336828, + "grad_norm": 1.0391695499420166, + "learning_rate": 4.734355296589039e-05, + "loss": 0.4089, + "num_input_tokens_seen": 2376160, + "step": 8900 + }, + { + "epoch": 2.3342070773263432, + "grad_norm": 2.008049964904785, + "learning_rate": 4.73384200789583e-05, + "loss": 0.5966, + "num_input_tokens_seen": 2377936, + "step": 8905 + }, + { + "epoch": 2.3355176933158583, + "grad_norm": 2.9286208152770996, + "learning_rate": 4.733328251663348e-05, + "loss": 0.5311, + "num_input_tokens_seen": 2379264, + "step": 8910 + }, + { + "epoch": 2.3368283093053734, + "grad_norm": 1.5960283279418945, + "learning_rate": 4.732814027999122e-05, + "loss": 0.3059, + "num_input_tokens_seen": 2380928, + "step": 8915 + }, + { + "epoch": 2.3381389252948885, + "grad_norm": 2.0761544704437256, + "learning_rate": 4.7322993370107775e-05, + "loss": 0.6194, + "num_input_tokens_seen": 2382368, + "step": 8920 + }, + { + "epoch": 2.3394495412844036, + "grad_norm": 1.9234933853149414, + "learning_rate": 4.7317841788060394e-05, + "loss": 0.5358, + "num_input_tokens_seen": 2383712, + "step": 8925 + }, + { + "epoch": 2.3407601572739187, + "grad_norm": 4.192538738250732, + "learning_rate": 4.731268553492729e-05, + "loss": 0.3882, + "num_input_tokens_seen": 2384976, + "step": 8930 + }, + { + "epoch": 2.342070773263434, + "grad_norm": 4.208868980407715, + "learning_rate": 4.730752461178766e-05, + "loss": 0.4223, + "num_input_tokens_seen": 2386528, + "step": 8935 + }, + { + "epoch": 2.343381389252949, + "grad_norm": 3.837918281555176, + "learning_rate": 4.730235901972166e-05, + "loss": 0.4312, + "num_input_tokens_seen": 2388016, + "step": 8940 + }, + { + "epoch": 2.344692005242464, + "grad_norm": 3.2767539024353027, + "learning_rate": 4.729718875981046e-05, + "loss": 0.4664, + "num_input_tokens_seen": 2389360, + "step": 8945 + }, + { + "epoch": 2.346002621231979, + "grad_norm": 9.526091575622559, + "learning_rate": 4.729201383313618e-05, + "loss": 0.4407, + "num_input_tokens_seen": 2390432, + "step": 8950 + }, + { + "epoch": 2.347313237221494, + "grad_norm": 2.300060272216797, + "learning_rate": 4.7286834240781925e-05, + "loss": 0.5598, + "num_input_tokens_seen": 2391664, + "step": 8955 + }, + { + "epoch": 2.3486238532110093, + "grad_norm": 3.802366018295288, + "learning_rate": 4.728164998383177e-05, + "loss": 0.6136, + "num_input_tokens_seen": 2392752, + "step": 8960 + }, + { + "epoch": 2.3499344692005244, + "grad_norm": 2.173907995223999, + "learning_rate": 4.727646106337077e-05, + "loss": 0.5351, + "num_input_tokens_seen": 2393856, + "step": 8965 + }, + { + "epoch": 2.3512450851900395, + "grad_norm": 2.5995163917541504, + "learning_rate": 4.727126748048496e-05, + "loss": 0.4082, + "num_input_tokens_seen": 2395056, + "step": 8970 + }, + { + "epoch": 2.352555701179554, + "grad_norm": 2.208634853363037, + "learning_rate": 4.726606923626135e-05, + "loss": 0.432, + "num_input_tokens_seen": 2396864, + "step": 8975 + }, + { + "epoch": 2.3538663171690697, + "grad_norm": 2.7352826595306396, + "learning_rate": 4.726086633178792e-05, + "loss": 0.4336, + "num_input_tokens_seen": 2398048, + "step": 8980 + }, + { + "epoch": 2.3551769331585843, + "grad_norm": 2.8280041217803955, + "learning_rate": 4.7255658768153635e-05, + "loss": 0.5203, + "num_input_tokens_seen": 2399184, + "step": 8985 + }, + { + "epoch": 2.3564875491480994, + "grad_norm": 3.0276596546173096, + "learning_rate": 4.725044654644842e-05, + "loss": 0.4548, + "num_input_tokens_seen": 2400832, + "step": 8990 + }, + { + "epoch": 2.3577981651376145, + "grad_norm": 1.9883427619934082, + "learning_rate": 4.724522966776319e-05, + "loss": 0.5609, + "num_input_tokens_seen": 2402096, + "step": 8995 + }, + { + "epoch": 2.3591087811271296, + "grad_norm": 2.8984251022338867, + "learning_rate": 4.7240008133189836e-05, + "loss": 0.3424, + "num_input_tokens_seen": 2403296, + "step": 9000 + }, + { + "epoch": 2.3604193971166447, + "grad_norm": 9.120626449584961, + "learning_rate": 4.723478194382119e-05, + "loss": 0.4949, + "num_input_tokens_seen": 2404464, + "step": 9005 + }, + { + "epoch": 2.36173001310616, + "grad_norm": 2.839026689529419, + "learning_rate": 4.722955110075112e-05, + "loss": 0.567, + "num_input_tokens_seen": 2405632, + "step": 9010 + }, + { + "epoch": 2.363040629095675, + "grad_norm": 1.7725763320922852, + "learning_rate": 4.7224315605074403e-05, + "loss": 0.6626, + "num_input_tokens_seen": 2407312, + "step": 9015 + }, + { + "epoch": 2.36435124508519, + "grad_norm": 2.845207452774048, + "learning_rate": 4.7219075457886836e-05, + "loss": 0.4147, + "num_input_tokens_seen": 2409216, + "step": 9020 + }, + { + "epoch": 2.365661861074705, + "grad_norm": 6.043459892272949, + "learning_rate": 4.721383066028516e-05, + "loss": 0.3496, + "num_input_tokens_seen": 2410464, + "step": 9025 + }, + { + "epoch": 2.36697247706422, + "grad_norm": 1.4394112825393677, + "learning_rate": 4.720858121336712e-05, + "loss": 0.4212, + "num_input_tokens_seen": 2412080, + "step": 9030 + }, + { + "epoch": 2.3682830930537353, + "grad_norm": 1.9516761302947998, + "learning_rate": 4.72033271182314e-05, + "loss": 0.356, + "num_input_tokens_seen": 2413152, + "step": 9035 + }, + { + "epoch": 2.3695937090432504, + "grad_norm": 3.1289384365081787, + "learning_rate": 4.7198068375977675e-05, + "loss": 0.4491, + "num_input_tokens_seen": 2414576, + "step": 9040 + }, + { + "epoch": 2.3709043250327655, + "grad_norm": 2.128732919692993, + "learning_rate": 4.71928049877066e-05, + "loss": 0.5808, + "num_input_tokens_seen": 2415696, + "step": 9045 + }, + { + "epoch": 2.3722149410222806, + "grad_norm": 1.6383463144302368, + "learning_rate": 4.718753695451978e-05, + "loss": 0.5413, + "num_input_tokens_seen": 2417792, + "step": 9050 + }, + { + "epoch": 2.3735255570117957, + "grad_norm": 2.0223023891448975, + "learning_rate": 4.718226427751982e-05, + "loss": 0.4607, + "num_input_tokens_seen": 2418880, + "step": 9055 + }, + { + "epoch": 2.374836173001311, + "grad_norm": 3.3097875118255615, + "learning_rate": 4.7176986957810275e-05, + "loss": 0.4918, + "num_input_tokens_seen": 2420688, + "step": 9060 + }, + { + "epoch": 2.376146788990826, + "grad_norm": 1.80087411403656, + "learning_rate": 4.717170499649568e-05, + "loss": 0.6074, + "num_input_tokens_seen": 2422016, + "step": 9065 + }, + { + "epoch": 2.3774574049803405, + "grad_norm": 4.794306755065918, + "learning_rate": 4.7166418394681526e-05, + "loss": 0.3962, + "num_input_tokens_seen": 2423440, + "step": 9070 + }, + { + "epoch": 2.378768020969856, + "grad_norm": 1.5220136642456055, + "learning_rate": 4.716112715347432e-05, + "loss": 0.3449, + "num_input_tokens_seen": 2424864, + "step": 9075 + }, + { + "epoch": 2.3800786369593707, + "grad_norm": 1.9750566482543945, + "learning_rate": 4.715583127398148e-05, + "loss": 0.3258, + "num_input_tokens_seen": 2426064, + "step": 9080 + }, + { + "epoch": 2.381389252948886, + "grad_norm": 1.403501272201538, + "learning_rate": 4.715053075731144e-05, + "loss": 0.3153, + "num_input_tokens_seen": 2427376, + "step": 9085 + }, + { + "epoch": 2.382699868938401, + "grad_norm": 3.474046230316162, + "learning_rate": 4.714522560457359e-05, + "loss": 0.4361, + "num_input_tokens_seen": 2428608, + "step": 9090 + }, + { + "epoch": 2.384010484927916, + "grad_norm": 3.1808409690856934, + "learning_rate": 4.713991581687827e-05, + "loss": 0.3398, + "num_input_tokens_seen": 2429952, + "step": 9095 + }, + { + "epoch": 2.385321100917431, + "grad_norm": 2.194119691848755, + "learning_rate": 4.7134601395336836e-05, + "loss": 0.5138, + "num_input_tokens_seen": 2430976, + "step": 9100 + }, + { + "epoch": 2.386631716906946, + "grad_norm": 3.703394889831543, + "learning_rate": 4.712928234106158e-05, + "loss": 0.5603, + "num_input_tokens_seen": 2432592, + "step": 9105 + }, + { + "epoch": 2.3879423328964613, + "grad_norm": 2.0383126735687256, + "learning_rate": 4.712395865516576e-05, + "loss": 0.2272, + "num_input_tokens_seen": 2433680, + "step": 9110 + }, + { + "epoch": 2.3892529488859764, + "grad_norm": 16.0150203704834, + "learning_rate": 4.711863033876362e-05, + "loss": 0.5328, + "num_input_tokens_seen": 2434752, + "step": 9115 + }, + { + "epoch": 2.3905635648754915, + "grad_norm": 2.9239275455474854, + "learning_rate": 4.7113297392970365e-05, + "loss": 0.5529, + "num_input_tokens_seen": 2435904, + "step": 9120 + }, + { + "epoch": 2.3918741808650066, + "grad_norm": 3.2938270568847656, + "learning_rate": 4.710795981890218e-05, + "loss": 0.4902, + "num_input_tokens_seen": 2437280, + "step": 9125 + }, + { + "epoch": 2.3931847968545217, + "grad_norm": 5.574960231781006, + "learning_rate": 4.7102617617676203e-05, + "loss": 0.5216, + "num_input_tokens_seen": 2438400, + "step": 9130 + }, + { + "epoch": 2.3944954128440368, + "grad_norm": 1.721887230873108, + "learning_rate": 4.7097270790410555e-05, + "loss": 0.3701, + "num_input_tokens_seen": 2439872, + "step": 9135 + }, + { + "epoch": 2.395806028833552, + "grad_norm": 6.327475070953369, + "learning_rate": 4.709191933822431e-05, + "loss": 0.6882, + "num_input_tokens_seen": 2441568, + "step": 9140 + }, + { + "epoch": 2.397116644823067, + "grad_norm": 1.5779691934585571, + "learning_rate": 4.7086563262237514e-05, + "loss": 0.4477, + "num_input_tokens_seen": 2442944, + "step": 9145 + }, + { + "epoch": 2.398427260812582, + "grad_norm": 6.975078582763672, + "learning_rate": 4.7081202563571194e-05, + "loss": 0.3984, + "num_input_tokens_seen": 2444208, + "step": 9150 + }, + { + "epoch": 2.399737876802097, + "grad_norm": 1.4191399812698364, + "learning_rate": 4.707583724334732e-05, + "loss": 0.4933, + "num_input_tokens_seen": 2445840, + "step": 9155 + }, + { + "epoch": 2.4010484927916123, + "grad_norm": 2.146841526031494, + "learning_rate": 4.707046730268887e-05, + "loss": 0.4562, + "num_input_tokens_seen": 2446864, + "step": 9160 + }, + { + "epoch": 2.402359108781127, + "grad_norm": 1.6775916814804077, + "learning_rate": 4.706509274271974e-05, + "loss": 0.5963, + "num_input_tokens_seen": 2447968, + "step": 9165 + }, + { + "epoch": 2.4036697247706424, + "grad_norm": 1.9380345344543457, + "learning_rate": 4.705971356456483e-05, + "loss": 0.5952, + "num_input_tokens_seen": 2449088, + "step": 9170 + }, + { + "epoch": 2.404980340760157, + "grad_norm": 2.4522581100463867, + "learning_rate": 4.7054329769349984e-05, + "loss": 0.388, + "num_input_tokens_seen": 2450192, + "step": 9175 + }, + { + "epoch": 2.406290956749672, + "grad_norm": 1.65061616897583, + "learning_rate": 4.704894135820202e-05, + "loss": 0.4958, + "num_input_tokens_seen": 2451504, + "step": 9180 + }, + { + "epoch": 2.4076015727391873, + "grad_norm": 1.2901828289031982, + "learning_rate": 4.704354833224873e-05, + "loss": 0.3562, + "num_input_tokens_seen": 2453520, + "step": 9185 + }, + { + "epoch": 2.4089121887287024, + "grad_norm": 1.4416431188583374, + "learning_rate": 4.7038150692618864e-05, + "loss": 0.4065, + "num_input_tokens_seen": 2455648, + "step": 9190 + }, + { + "epoch": 2.4102228047182175, + "grad_norm": 18.121511459350586, + "learning_rate": 4.703274844044214e-05, + "loss": 0.4503, + "num_input_tokens_seen": 2456768, + "step": 9195 + }, + { + "epoch": 2.4115334207077326, + "grad_norm": 3.8629186153411865, + "learning_rate": 4.702734157684924e-05, + "loss": 0.3561, + "num_input_tokens_seen": 2458112, + "step": 9200 + }, + { + "epoch": 2.4128440366972477, + "grad_norm": 3.8540687561035156, + "learning_rate": 4.7021930102971796e-05, + "loss": 0.4029, + "num_input_tokens_seen": 2459504, + "step": 9205 + }, + { + "epoch": 2.4141546526867628, + "grad_norm": 2.205900192260742, + "learning_rate": 4.701651401994244e-05, + "loss": 0.4603, + "num_input_tokens_seen": 2460336, + "step": 9210 + }, + { + "epoch": 2.415465268676278, + "grad_norm": 1.2710529565811157, + "learning_rate": 4.701109332889474e-05, + "loss": 0.7872, + "num_input_tokens_seen": 2461664, + "step": 9215 + }, + { + "epoch": 2.416775884665793, + "grad_norm": 1.6113512516021729, + "learning_rate": 4.700566803096324e-05, + "loss": 0.3807, + "num_input_tokens_seen": 2462912, + "step": 9220 + }, + { + "epoch": 2.418086500655308, + "grad_norm": 3.532566547393799, + "learning_rate": 4.700023812728344e-05, + "loss": 0.435, + "num_input_tokens_seen": 2464304, + "step": 9225 + }, + { + "epoch": 2.419397116644823, + "grad_norm": 2.6019845008850098, + "learning_rate": 4.699480361899181e-05, + "loss": 0.4382, + "num_input_tokens_seen": 2465552, + "step": 9230 + }, + { + "epoch": 2.4207077326343382, + "grad_norm": 2.0630829334259033, + "learning_rate": 4.6989364507225795e-05, + "loss": 0.4334, + "num_input_tokens_seen": 2467024, + "step": 9235 + }, + { + "epoch": 2.4220183486238533, + "grad_norm": 2.3899528980255127, + "learning_rate": 4.698392079312378e-05, + "loss": 0.6059, + "num_input_tokens_seen": 2469072, + "step": 9240 + }, + { + "epoch": 2.4233289646133684, + "grad_norm": 1.893070936203003, + "learning_rate": 4.697847247782513e-05, + "loss": 0.3299, + "num_input_tokens_seen": 2470720, + "step": 9245 + }, + { + "epoch": 2.4246395806028835, + "grad_norm": 5.154554843902588, + "learning_rate": 4.6973019562470166e-05, + "loss": 0.5507, + "num_input_tokens_seen": 2471888, + "step": 9250 + }, + { + "epoch": 2.4259501965923986, + "grad_norm": 2.2770206928253174, + "learning_rate": 4.6967562048200176e-05, + "loss": 0.4423, + "num_input_tokens_seen": 2473216, + "step": 9255 + }, + { + "epoch": 2.4272608125819133, + "grad_norm": 4.143144130706787, + "learning_rate": 4.69620999361574e-05, + "loss": 0.5666, + "num_input_tokens_seen": 2474352, + "step": 9260 + }, + { + "epoch": 2.4285714285714284, + "grad_norm": 0.8670002222061157, + "learning_rate": 4.695663322748506e-05, + "loss": 0.5694, + "num_input_tokens_seen": 2476304, + "step": 9265 + }, + { + "epoch": 2.4298820445609435, + "grad_norm": 2.269063711166382, + "learning_rate": 4.6951161923327324e-05, + "loss": 0.4358, + "num_input_tokens_seen": 2477616, + "step": 9270 + }, + { + "epoch": 2.4311926605504586, + "grad_norm": 2.910360813140869, + "learning_rate": 4.694568602482933e-05, + "loss": 0.5323, + "num_input_tokens_seen": 2478864, + "step": 9275 + }, + { + "epoch": 2.4325032765399737, + "grad_norm": 3.937321424484253, + "learning_rate": 4.694020553313716e-05, + "loss": 0.524, + "num_input_tokens_seen": 2479952, + "step": 9280 + }, + { + "epoch": 2.4338138925294888, + "grad_norm": 1.7902629375457764, + "learning_rate": 4.693472044939788e-05, + "loss": 0.395, + "num_input_tokens_seen": 2480992, + "step": 9285 + }, + { + "epoch": 2.435124508519004, + "grad_norm": 2.555553436279297, + "learning_rate": 4.692923077475952e-05, + "loss": 0.3642, + "num_input_tokens_seen": 2482352, + "step": 9290 + }, + { + "epoch": 2.436435124508519, + "grad_norm": 3.033275842666626, + "learning_rate": 4.6923736510371054e-05, + "loss": 0.5238, + "num_input_tokens_seen": 2483744, + "step": 9295 + }, + { + "epoch": 2.437745740498034, + "grad_norm": 1.5723377466201782, + "learning_rate": 4.69182376573824e-05, + "loss": 0.5565, + "num_input_tokens_seen": 2486752, + "step": 9300 + }, + { + "epoch": 2.439056356487549, + "grad_norm": 2.50778865814209, + "learning_rate": 4.6912734216944486e-05, + "loss": 0.38, + "num_input_tokens_seen": 2487984, + "step": 9305 + }, + { + "epoch": 2.4403669724770642, + "grad_norm": 3.011585235595703, + "learning_rate": 4.6907226190209165e-05, + "loss": 0.5487, + "num_input_tokens_seen": 2489568, + "step": 9310 + }, + { + "epoch": 2.4416775884665793, + "grad_norm": 1.482019066810608, + "learning_rate": 4.6901713578329255e-05, + "loss": 0.4267, + "num_input_tokens_seen": 2490592, + "step": 9315 + }, + { + "epoch": 2.4429882044560944, + "grad_norm": 2.6391847133636475, + "learning_rate": 4.689619638245853e-05, + "loss": 0.4903, + "num_input_tokens_seen": 2492064, + "step": 9320 + }, + { + "epoch": 2.4442988204456095, + "grad_norm": 4.345822811126709, + "learning_rate": 4.6890674603751746e-05, + "loss": 0.5161, + "num_input_tokens_seen": 2493248, + "step": 9325 + }, + { + "epoch": 2.4456094364351246, + "grad_norm": 1.390170931816101, + "learning_rate": 4.688514824336459e-05, + "loss": 0.4433, + "num_input_tokens_seen": 2494416, + "step": 9330 + }, + { + "epoch": 2.4469200524246397, + "grad_norm": 2.502927780151367, + "learning_rate": 4.687961730245371e-05, + "loss": 0.5293, + "num_input_tokens_seen": 2495584, + "step": 9335 + }, + { + "epoch": 2.448230668414155, + "grad_norm": 2.4510414600372314, + "learning_rate": 4.687408178217674e-05, + "loss": 0.6339, + "num_input_tokens_seen": 2496816, + "step": 9340 + }, + { + "epoch": 2.44954128440367, + "grad_norm": 2.0319478511810303, + "learning_rate": 4.6868541683692244e-05, + "loss": 0.5561, + "num_input_tokens_seen": 2497952, + "step": 9345 + }, + { + "epoch": 2.450851900393185, + "grad_norm": 2.19222354888916, + "learning_rate": 4.6862997008159774e-05, + "loss": 0.3496, + "num_input_tokens_seen": 2499200, + "step": 9350 + }, + { + "epoch": 2.4521625163826997, + "grad_norm": 1.2410725355148315, + "learning_rate": 4.685744775673979e-05, + "loss": 0.6869, + "num_input_tokens_seen": 2500768, + "step": 9355 + }, + { + "epoch": 2.4534731323722148, + "grad_norm": 5.197871208190918, + "learning_rate": 4.685189393059377e-05, + "loss": 0.4809, + "num_input_tokens_seen": 2501984, + "step": 9360 + }, + { + "epoch": 2.45478374836173, + "grad_norm": 3.061840772628784, + "learning_rate": 4.68463355308841e-05, + "loss": 0.4986, + "num_input_tokens_seen": 2503008, + "step": 9365 + }, + { + "epoch": 2.456094364351245, + "grad_norm": 3.2579574584960938, + "learning_rate": 4.684077255877415e-05, + "loss": 0.4173, + "num_input_tokens_seen": 2504160, + "step": 9370 + }, + { + "epoch": 2.45740498034076, + "grad_norm": 2.2280266284942627, + "learning_rate": 4.6835205015428246e-05, + "loss": 0.4293, + "num_input_tokens_seen": 2505536, + "step": 9375 + }, + { + "epoch": 2.458715596330275, + "grad_norm": 2.1136300563812256, + "learning_rate": 4.682963290201166e-05, + "loss": 0.342, + "num_input_tokens_seen": 2506512, + "step": 9380 + }, + { + "epoch": 2.4600262123197902, + "grad_norm": 3.915313243865967, + "learning_rate": 4.682405621969062e-05, + "loss": 0.5404, + "num_input_tokens_seen": 2507632, + "step": 9385 + }, + { + "epoch": 2.4613368283093053, + "grad_norm": 2.772799491882324, + "learning_rate": 4.681847496963233e-05, + "loss": 0.6528, + "num_input_tokens_seen": 2508848, + "step": 9390 + }, + { + "epoch": 2.4626474442988204, + "grad_norm": 3.0750348567962646, + "learning_rate": 4.6812889153004925e-05, + "loss": 0.3976, + "num_input_tokens_seen": 2510144, + "step": 9395 + }, + { + "epoch": 2.4639580602883355, + "grad_norm": 3.9629788398742676, + "learning_rate": 4.6807298770977516e-05, + "loss": 0.5097, + "num_input_tokens_seen": 2511360, + "step": 9400 + }, + { + "epoch": 2.4652686762778506, + "grad_norm": 5.624678611755371, + "learning_rate": 4.6801703824720146e-05, + "loss": 0.899, + "num_input_tokens_seen": 2512208, + "step": 9405 + }, + { + "epoch": 2.4665792922673657, + "grad_norm": 2.083249568939209, + "learning_rate": 4.679610431540384e-05, + "loss": 0.5291, + "num_input_tokens_seen": 2513360, + "step": 9410 + }, + { + "epoch": 2.467889908256881, + "grad_norm": 14.196989059448242, + "learning_rate": 4.679050024420056e-05, + "loss": 0.5857, + "num_input_tokens_seen": 2514480, + "step": 9415 + }, + { + "epoch": 2.469200524246396, + "grad_norm": 4.614956378936768, + "learning_rate": 4.678489161228323e-05, + "loss": 0.6034, + "num_input_tokens_seen": 2515536, + "step": 9420 + }, + { + "epoch": 2.470511140235911, + "grad_norm": 2.11728835105896, + "learning_rate": 4.677927842082572e-05, + "loss": 0.5234, + "num_input_tokens_seen": 2517264, + "step": 9425 + }, + { + "epoch": 2.471821756225426, + "grad_norm": 6.735559463500977, + "learning_rate": 4.6773660671002885e-05, + "loss": 0.4555, + "num_input_tokens_seen": 2518496, + "step": 9430 + }, + { + "epoch": 2.473132372214941, + "grad_norm": 2.333526849746704, + "learning_rate": 4.676803836399048e-05, + "loss": 0.4428, + "num_input_tokens_seen": 2520224, + "step": 9435 + }, + { + "epoch": 2.474442988204456, + "grad_norm": 2.0071523189544678, + "learning_rate": 4.6762411500965265e-05, + "loss": 0.6961, + "num_input_tokens_seen": 2521328, + "step": 9440 + }, + { + "epoch": 2.4757536041939714, + "grad_norm": 4.044301986694336, + "learning_rate": 4.675678008310492e-05, + "loss": 0.3722, + "num_input_tokens_seen": 2522400, + "step": 9445 + }, + { + "epoch": 2.477064220183486, + "grad_norm": 2.3862476348876953, + "learning_rate": 4.6751144111588094e-05, + "loss": 0.4024, + "num_input_tokens_seen": 2523776, + "step": 9450 + }, + { + "epoch": 2.478374836173001, + "grad_norm": 2.8803091049194336, + "learning_rate": 4.6745503587594396e-05, + "loss": 0.4718, + "num_input_tokens_seen": 2525232, + "step": 9455 + }, + { + "epoch": 2.4796854521625162, + "grad_norm": 4.48162317276001, + "learning_rate": 4.6739858512304366e-05, + "loss": 0.6172, + "num_input_tokens_seen": 2526512, + "step": 9460 + }, + { + "epoch": 2.4809960681520313, + "grad_norm": 2.42187237739563, + "learning_rate": 4.673420888689952e-05, + "loss": 0.3559, + "num_input_tokens_seen": 2527488, + "step": 9465 + }, + { + "epoch": 2.4823066841415464, + "grad_norm": 3.840848684310913, + "learning_rate": 4.67285547125623e-05, + "loss": 0.5981, + "num_input_tokens_seen": 2528736, + "step": 9470 + }, + { + "epoch": 2.4836173001310615, + "grad_norm": 1.4237312078475952, + "learning_rate": 4.6722895990476126e-05, + "loss": 0.4166, + "num_input_tokens_seen": 2530208, + "step": 9475 + }, + { + "epoch": 2.4849279161205766, + "grad_norm": 4.84081506729126, + "learning_rate": 4.671723272182535e-05, + "loss": 0.3308, + "num_input_tokens_seen": 2532000, + "step": 9480 + }, + { + "epoch": 2.4862385321100917, + "grad_norm": 2.882488250732422, + "learning_rate": 4.67115649077953e-05, + "loss": 0.3811, + "num_input_tokens_seen": 2533664, + "step": 9485 + }, + { + "epoch": 2.487549148099607, + "grad_norm": 2.6416101455688477, + "learning_rate": 4.670589254957222e-05, + "loss": 0.502, + "num_input_tokens_seen": 2534768, + "step": 9490 + }, + { + "epoch": 2.488859764089122, + "grad_norm": 3.8151192665100098, + "learning_rate": 4.6700215648343336e-05, + "loss": 0.5866, + "num_input_tokens_seen": 2536176, + "step": 9495 + }, + { + "epoch": 2.490170380078637, + "grad_norm": 2.281278371810913, + "learning_rate": 4.669453420529682e-05, + "loss": 0.3997, + "num_input_tokens_seen": 2537600, + "step": 9500 + }, + { + "epoch": 2.491480996068152, + "grad_norm": 3.458307981491089, + "learning_rate": 4.668884822162177e-05, + "loss": 0.3771, + "num_input_tokens_seen": 2538880, + "step": 9505 + }, + { + "epoch": 2.492791612057667, + "grad_norm": 2.5229063034057617, + "learning_rate": 4.668315769850827e-05, + "loss": 0.4641, + "num_input_tokens_seen": 2540288, + "step": 9510 + }, + { + "epoch": 2.4941022280471823, + "grad_norm": 3.358595848083496, + "learning_rate": 4.667746263714733e-05, + "loss": 0.3943, + "num_input_tokens_seen": 2541472, + "step": 9515 + }, + { + "epoch": 2.4954128440366974, + "grad_norm": 3.0506341457366943, + "learning_rate": 4.667176303873091e-05, + "loss": 0.5339, + "num_input_tokens_seen": 2543040, + "step": 9520 + }, + { + "epoch": 2.4967234600262125, + "grad_norm": 2.4486844539642334, + "learning_rate": 4.666605890445193e-05, + "loss": 0.5202, + "num_input_tokens_seen": 2544512, + "step": 9525 + }, + { + "epoch": 2.4980340760157276, + "grad_norm": 1.6416093111038208, + "learning_rate": 4.6660350235504265e-05, + "loss": 0.3231, + "num_input_tokens_seen": 2545648, + "step": 9530 + }, + { + "epoch": 2.4993446920052422, + "grad_norm": 1.2806379795074463, + "learning_rate": 4.6654637033082715e-05, + "loss": 0.3746, + "num_input_tokens_seen": 2546752, + "step": 9535 + }, + { + "epoch": 2.5006553079947578, + "grad_norm": 1.456357479095459, + "learning_rate": 4.6648919298383055e-05, + "loss": 0.3682, + "num_input_tokens_seen": 2547808, + "step": 9540 + }, + { + "epoch": 2.5006553079947578, + "eval_loss": 0.47390642762184143, + "eval_runtime": 18.2024, + "eval_samples_per_second": 46.587, + "eval_steps_per_second": 23.294, + "num_input_tokens_seen": 2547808, + "step": 9540 + }, + { + "epoch": 2.5019659239842724, + "grad_norm": 1.4339556694030762, + "learning_rate": 4.664319703260199e-05, + "loss": 0.5239, + "num_input_tokens_seen": 2551568, + "step": 9545 + }, + { + "epoch": 2.5032765399737875, + "grad_norm": 1.658297061920166, + "learning_rate": 4.6637470236937186e-05, + "loss": 0.341, + "num_input_tokens_seen": 2552864, + "step": 9550 + }, + { + "epoch": 2.5045871559633026, + "grad_norm": 3.00276780128479, + "learning_rate": 4.663173891258724e-05, + "loss": 0.4931, + "num_input_tokens_seen": 2554080, + "step": 9555 + }, + { + "epoch": 2.5058977719528177, + "grad_norm": 3.3485894203186035, + "learning_rate": 4.662600306075172e-05, + "loss": 0.4662, + "num_input_tokens_seen": 2555424, + "step": 9560 + }, + { + "epoch": 2.507208387942333, + "grad_norm": 1.6323100328445435, + "learning_rate": 4.662026268263113e-05, + "loss": 0.4482, + "num_input_tokens_seen": 2556576, + "step": 9565 + }, + { + "epoch": 2.508519003931848, + "grad_norm": 7.006704807281494, + "learning_rate": 4.6614517779426917e-05, + "loss": 0.5289, + "num_input_tokens_seen": 2557792, + "step": 9570 + }, + { + "epoch": 2.509829619921363, + "grad_norm": 1.4573556184768677, + "learning_rate": 4.660876835234148e-05, + "loss": 0.4021, + "num_input_tokens_seen": 2558976, + "step": 9575 + }, + { + "epoch": 2.511140235910878, + "grad_norm": 2.877042531967163, + "learning_rate": 4.660301440257816e-05, + "loss": 0.5133, + "num_input_tokens_seen": 2560064, + "step": 9580 + }, + { + "epoch": 2.512450851900393, + "grad_norm": 3.4486331939697266, + "learning_rate": 4.659725593134125e-05, + "loss": 0.4258, + "num_input_tokens_seen": 2561120, + "step": 9585 + }, + { + "epoch": 2.5137614678899083, + "grad_norm": 1.8013842105865479, + "learning_rate": 4.6591492939836e-05, + "loss": 0.6028, + "num_input_tokens_seen": 2563056, + "step": 9590 + }, + { + "epoch": 2.5150720838794234, + "grad_norm": 1.8173805475234985, + "learning_rate": 4.6585725429268585e-05, + "loss": 0.4171, + "num_input_tokens_seen": 2564224, + "step": 9595 + }, + { + "epoch": 2.5163826998689385, + "grad_norm": 43.361305236816406, + "learning_rate": 4.657995340084613e-05, + "loss": 0.6543, + "num_input_tokens_seen": 2565344, + "step": 9600 + }, + { + "epoch": 2.5176933158584536, + "grad_norm": 2.923604965209961, + "learning_rate": 4.657417685577671e-05, + "loss": 0.4205, + "num_input_tokens_seen": 2566576, + "step": 9605 + }, + { + "epoch": 2.5190039318479687, + "grad_norm": 2.936169147491455, + "learning_rate": 4.656839579526936e-05, + "loss": 0.5891, + "num_input_tokens_seen": 2567632, + "step": 9610 + }, + { + "epoch": 2.5203145478374838, + "grad_norm": 1.4318242073059082, + "learning_rate": 4.6562610220534035e-05, + "loss": 0.4314, + "num_input_tokens_seen": 2569056, + "step": 9615 + }, + { + "epoch": 2.5216251638269984, + "grad_norm": 3.2988338470458984, + "learning_rate": 4.6556820132781646e-05, + "loss": 0.4638, + "num_input_tokens_seen": 2570240, + "step": 9620 + }, + { + "epoch": 2.522935779816514, + "grad_norm": 2.8062493801116943, + "learning_rate": 4.6551025533224056e-05, + "loss": 0.4463, + "num_input_tokens_seen": 2571264, + "step": 9625 + }, + { + "epoch": 2.5242463958060286, + "grad_norm": 1.5980292558670044, + "learning_rate": 4.654522642307405e-05, + "loss": 0.4382, + "num_input_tokens_seen": 2572752, + "step": 9630 + }, + { + "epoch": 2.525557011795544, + "grad_norm": 5.327779769897461, + "learning_rate": 4.653942280354539e-05, + "loss": 0.5267, + "num_input_tokens_seen": 2573680, + "step": 9635 + }, + { + "epoch": 2.526867627785059, + "grad_norm": 2.2721338272094727, + "learning_rate": 4.6533614675852755e-05, + "loss": 0.449, + "num_input_tokens_seen": 2574640, + "step": 9640 + }, + { + "epoch": 2.528178243774574, + "grad_norm": 2.2228710651397705, + "learning_rate": 4.652780204121177e-05, + "loss": 0.4939, + "num_input_tokens_seen": 2576000, + "step": 9645 + }, + { + "epoch": 2.529488859764089, + "grad_norm": 3.960564374923706, + "learning_rate": 4.652198490083901e-05, + "loss": 0.4314, + "num_input_tokens_seen": 2577184, + "step": 9650 + }, + { + "epoch": 2.530799475753604, + "grad_norm": 1.0383309125900269, + "learning_rate": 4.6516163255952004e-05, + "loss": 0.7648, + "num_input_tokens_seen": 2578592, + "step": 9655 + }, + { + "epoch": 2.532110091743119, + "grad_norm": 1.883036494255066, + "learning_rate": 4.651033710776921e-05, + "loss": 0.5333, + "num_input_tokens_seen": 2579952, + "step": 9660 + }, + { + "epoch": 2.5334207077326343, + "grad_norm": 2.6779873371124268, + "learning_rate": 4.6504506457510025e-05, + "loss": 0.5683, + "num_input_tokens_seen": 2581312, + "step": 9665 + }, + { + "epoch": 2.5347313237221494, + "grad_norm": 2.3193752765655518, + "learning_rate": 4.649867130639479e-05, + "loss": 0.5978, + "num_input_tokens_seen": 2583040, + "step": 9670 + }, + { + "epoch": 2.5360419397116645, + "grad_norm": 6.7927021980285645, + "learning_rate": 4.64928316556448e-05, + "loss": 0.4106, + "num_input_tokens_seen": 2584112, + "step": 9675 + }, + { + "epoch": 2.5373525557011796, + "grad_norm": 2.7136473655700684, + "learning_rate": 4.648698750648228e-05, + "loss": 0.5209, + "num_input_tokens_seen": 2585360, + "step": 9680 + }, + { + "epoch": 2.5386631716906947, + "grad_norm": 1.4671131372451782, + "learning_rate": 4.64811388601304e-05, + "loss": 0.4738, + "num_input_tokens_seen": 2586656, + "step": 9685 + }, + { + "epoch": 2.5399737876802098, + "grad_norm": 2.6025516986846924, + "learning_rate": 4.647528571781328e-05, + "loss": 0.6026, + "num_input_tokens_seen": 2587904, + "step": 9690 + }, + { + "epoch": 2.541284403669725, + "grad_norm": 4.088352203369141, + "learning_rate": 4.646942808075596e-05, + "loss": 0.5129, + "num_input_tokens_seen": 2589088, + "step": 9695 + }, + { + "epoch": 2.54259501965924, + "grad_norm": 1.992355465888977, + "learning_rate": 4.646356595018444e-05, + "loss": 0.5491, + "num_input_tokens_seen": 2590528, + "step": 9700 + }, + { + "epoch": 2.543905635648755, + "grad_norm": 3.334611177444458, + "learning_rate": 4.6457699327325655e-05, + "loss": 0.3115, + "num_input_tokens_seen": 2591744, + "step": 9705 + }, + { + "epoch": 2.54521625163827, + "grad_norm": 6.3444504737854, + "learning_rate": 4.645182821340747e-05, + "loss": 0.4078, + "num_input_tokens_seen": 2593344, + "step": 9710 + }, + { + "epoch": 2.546526867627785, + "grad_norm": 2.6036558151245117, + "learning_rate": 4.644595260965872e-05, + "loss": 0.4892, + "num_input_tokens_seen": 2594576, + "step": 9715 + }, + { + "epoch": 2.5478374836173003, + "grad_norm": 5.040546417236328, + "learning_rate": 4.644007251730913e-05, + "loss": 0.6332, + "num_input_tokens_seen": 2596304, + "step": 9720 + }, + { + "epoch": 2.549148099606815, + "grad_norm": 1.9487712383270264, + "learning_rate": 4.643418793758941e-05, + "loss": 0.4084, + "num_input_tokens_seen": 2597888, + "step": 9725 + }, + { + "epoch": 2.5504587155963305, + "grad_norm": 2.4822254180908203, + "learning_rate": 4.642829887173119e-05, + "loss": 0.6074, + "num_input_tokens_seen": 2599104, + "step": 9730 + }, + { + "epoch": 2.551769331585845, + "grad_norm": 2.2588813304901123, + "learning_rate": 4.642240532096705e-05, + "loss": 0.348, + "num_input_tokens_seen": 2600336, + "step": 9735 + }, + { + "epoch": 2.5530799475753603, + "grad_norm": 5.36337423324585, + "learning_rate": 4.641650728653049e-05, + "loss": 0.4275, + "num_input_tokens_seen": 2601808, + "step": 9740 + }, + { + "epoch": 2.5543905635648754, + "grad_norm": 1.0921777486801147, + "learning_rate": 4.641060476965595e-05, + "loss": 0.6845, + "num_input_tokens_seen": 2603152, + "step": 9745 + }, + { + "epoch": 2.5557011795543905, + "grad_norm": 3.845309019088745, + "learning_rate": 4.640469777157883e-05, + "loss": 0.518, + "num_input_tokens_seen": 2604320, + "step": 9750 + }, + { + "epoch": 2.5570117955439056, + "grad_norm": 1.7275298833847046, + "learning_rate": 4.6398786293535455e-05, + "loss": 0.5445, + "num_input_tokens_seen": 2605616, + "step": 9755 + }, + { + "epoch": 2.5583224115334207, + "grad_norm": 2.4588615894317627, + "learning_rate": 4.639287033676308e-05, + "loss": 0.8263, + "num_input_tokens_seen": 2607008, + "step": 9760 + }, + { + "epoch": 2.5596330275229358, + "grad_norm": 2.169058084487915, + "learning_rate": 4.638694990249991e-05, + "loss": 0.6215, + "num_input_tokens_seen": 2608096, + "step": 9765 + }, + { + "epoch": 2.560943643512451, + "grad_norm": 3.019418716430664, + "learning_rate": 4.638102499198508e-05, + "loss": 0.6055, + "num_input_tokens_seen": 2609344, + "step": 9770 + }, + { + "epoch": 2.562254259501966, + "grad_norm": 3.628521203994751, + "learning_rate": 4.637509560645865e-05, + "loss": 0.3552, + "num_input_tokens_seen": 2610880, + "step": 9775 + }, + { + "epoch": 2.563564875491481, + "grad_norm": 1.7184760570526123, + "learning_rate": 4.636916174716166e-05, + "loss": 0.4443, + "num_input_tokens_seen": 2611968, + "step": 9780 + }, + { + "epoch": 2.564875491480996, + "grad_norm": 1.3631263971328735, + "learning_rate": 4.636322341533603e-05, + "loss": 0.491, + "num_input_tokens_seen": 2613376, + "step": 9785 + }, + { + "epoch": 2.5661861074705112, + "grad_norm": 4.875913619995117, + "learning_rate": 4.6357280612224656e-05, + "loss": 0.476, + "num_input_tokens_seen": 2614352, + "step": 9790 + }, + { + "epoch": 2.5674967234600263, + "grad_norm": 2.2872798442840576, + "learning_rate": 4.635133333907135e-05, + "loss": 0.5216, + "num_input_tokens_seen": 2615600, + "step": 9795 + }, + { + "epoch": 2.5688073394495414, + "grad_norm": 2.387446403503418, + "learning_rate": 4.634538159712088e-05, + "loss": 0.4307, + "num_input_tokens_seen": 2617040, + "step": 9800 + }, + { + "epoch": 2.5701179554390565, + "grad_norm": 1.6203107833862305, + "learning_rate": 4.6339425387618906e-05, + "loss": 0.5633, + "num_input_tokens_seen": 2618432, + "step": 9805 + }, + { + "epoch": 2.571428571428571, + "grad_norm": 1.8895763158798218, + "learning_rate": 4.633346471181207e-05, + "loss": 0.3927, + "num_input_tokens_seen": 2619488, + "step": 9810 + }, + { + "epoch": 2.5727391874180867, + "grad_norm": 2.2336175441741943, + "learning_rate": 4.632749957094794e-05, + "loss": 0.5276, + "num_input_tokens_seen": 2620688, + "step": 9815 + }, + { + "epoch": 2.5740498034076014, + "grad_norm": 3.8014719486236572, + "learning_rate": 4.6321529966275e-05, + "loss": 0.5378, + "num_input_tokens_seen": 2622032, + "step": 9820 + }, + { + "epoch": 2.575360419397117, + "grad_norm": 3.9381728172302246, + "learning_rate": 4.6315555899042686e-05, + "loss": 0.4483, + "num_input_tokens_seen": 2623728, + "step": 9825 + }, + { + "epoch": 2.5766710353866316, + "grad_norm": 2.84721040725708, + "learning_rate": 4.630957737050134e-05, + "loss": 0.8422, + "num_input_tokens_seen": 2624896, + "step": 9830 + }, + { + "epoch": 2.5779816513761467, + "grad_norm": 2.6388561725616455, + "learning_rate": 4.630359438190227e-05, + "loss": 0.4825, + "num_input_tokens_seen": 2626240, + "step": 9835 + }, + { + "epoch": 2.5792922673656618, + "grad_norm": 8.09935474395752, + "learning_rate": 4.629760693449772e-05, + "loss": 0.5665, + "num_input_tokens_seen": 2627424, + "step": 9840 + }, + { + "epoch": 2.580602883355177, + "grad_norm": 4.606799602508545, + "learning_rate": 4.6291615029540826e-05, + "loss": 0.4464, + "num_input_tokens_seen": 2629040, + "step": 9845 + }, + { + "epoch": 2.581913499344692, + "grad_norm": 3.284276008605957, + "learning_rate": 4.62856186682857e-05, + "loss": 0.4786, + "num_input_tokens_seen": 2630320, + "step": 9850 + }, + { + "epoch": 2.583224115334207, + "grad_norm": 1.746683955192566, + "learning_rate": 4.627961785198736e-05, + "loss": 0.4428, + "num_input_tokens_seen": 2631920, + "step": 9855 + }, + { + "epoch": 2.584534731323722, + "grad_norm": 2.101895809173584, + "learning_rate": 4.627361258190178e-05, + "loss": 0.6503, + "num_input_tokens_seen": 2633520, + "step": 9860 + }, + { + "epoch": 2.5858453473132372, + "grad_norm": 7.370722770690918, + "learning_rate": 4.626760285928584e-05, + "loss": 0.5224, + "num_input_tokens_seen": 2634400, + "step": 9865 + }, + { + "epoch": 2.5871559633027523, + "grad_norm": 1.525562047958374, + "learning_rate": 4.626158868539737e-05, + "loss": 0.4182, + "num_input_tokens_seen": 2635680, + "step": 9870 + }, + { + "epoch": 2.5884665792922674, + "grad_norm": 1.9477524757385254, + "learning_rate": 4.625557006149512e-05, + "loss": 0.3762, + "num_input_tokens_seen": 2636800, + "step": 9875 + }, + { + "epoch": 2.5897771952817825, + "grad_norm": 2.560084819793701, + "learning_rate": 4.624954698883879e-05, + "loss": 0.624, + "num_input_tokens_seen": 2638000, + "step": 9880 + }, + { + "epoch": 2.5910878112712976, + "grad_norm": 1.7415542602539062, + "learning_rate": 4.624351946868899e-05, + "loss": 0.6101, + "num_input_tokens_seen": 2639184, + "step": 9885 + }, + { + "epoch": 2.5923984272608127, + "grad_norm": 4.219828128814697, + "learning_rate": 4.6237487502307265e-05, + "loss": 0.4209, + "num_input_tokens_seen": 2640672, + "step": 9890 + }, + { + "epoch": 2.593709043250328, + "grad_norm": 2.335310935974121, + "learning_rate": 4.623145109095611e-05, + "loss": 0.5087, + "num_input_tokens_seen": 2641824, + "step": 9895 + }, + { + "epoch": 2.595019659239843, + "grad_norm": 1.3869574069976807, + "learning_rate": 4.622541023589893e-05, + "loss": 0.4658, + "num_input_tokens_seen": 2643968, + "step": 9900 + }, + { + "epoch": 2.5963302752293576, + "grad_norm": 7.199443340301514, + "learning_rate": 4.621936493840005e-05, + "loss": 0.627, + "num_input_tokens_seen": 2645072, + "step": 9905 + }, + { + "epoch": 2.597640891218873, + "grad_norm": 3.209975481033325, + "learning_rate": 4.621331519972476e-05, + "loss": 0.8003, + "num_input_tokens_seen": 2646464, + "step": 9910 + }, + { + "epoch": 2.5989515072083877, + "grad_norm": 4.875996112823486, + "learning_rate": 4.6207261021139246e-05, + "loss": 0.517, + "num_input_tokens_seen": 2647568, + "step": 9915 + }, + { + "epoch": 2.6002621231979033, + "grad_norm": 2.239478826522827, + "learning_rate": 4.620120240391065e-05, + "loss": 0.5959, + "num_input_tokens_seen": 2648896, + "step": 9920 + }, + { + "epoch": 2.601572739187418, + "grad_norm": 1.95870041847229, + "learning_rate": 4.6195139349307024e-05, + "loss": 0.3721, + "num_input_tokens_seen": 2650064, + "step": 9925 + }, + { + "epoch": 2.602883355176933, + "grad_norm": 2.0122909545898438, + "learning_rate": 4.6189071858597355e-05, + "loss": 0.4628, + "num_input_tokens_seen": 2651808, + "step": 9930 + }, + { + "epoch": 2.604193971166448, + "grad_norm": 5.376462936401367, + "learning_rate": 4.6182999933051554e-05, + "loss": 0.4513, + "num_input_tokens_seen": 2653280, + "step": 9935 + }, + { + "epoch": 2.6055045871559632, + "grad_norm": 8.720441818237305, + "learning_rate": 4.617692357394047e-05, + "loss": 0.8358, + "num_input_tokens_seen": 2654272, + "step": 9940 + }, + { + "epoch": 2.6068152031454783, + "grad_norm": 3.3949718475341797, + "learning_rate": 4.617084278253587e-05, + "loss": 0.372, + "num_input_tokens_seen": 2655744, + "step": 9945 + }, + { + "epoch": 2.6081258191349934, + "grad_norm": 2.887650966644287, + "learning_rate": 4.616475756011046e-05, + "loss": 1.0178, + "num_input_tokens_seen": 2657872, + "step": 9950 + }, + { + "epoch": 2.6094364351245085, + "grad_norm": 2.8434898853302, + "learning_rate": 4.615866790793786e-05, + "loss": 0.3381, + "num_input_tokens_seen": 2659600, + "step": 9955 + }, + { + "epoch": 2.6107470511140236, + "grad_norm": 2.8811023235321045, + "learning_rate": 4.6152573827292636e-05, + "loss": 0.473, + "num_input_tokens_seen": 2660816, + "step": 9960 + }, + { + "epoch": 2.6120576671035387, + "grad_norm": 1.8470860719680786, + "learning_rate": 4.6146475319450244e-05, + "loss": 0.4242, + "num_input_tokens_seen": 2662592, + "step": 9965 + }, + { + "epoch": 2.613368283093054, + "grad_norm": 3.2684459686279297, + "learning_rate": 4.6140372385687114e-05, + "loss": 0.5227, + "num_input_tokens_seen": 2663664, + "step": 9970 + }, + { + "epoch": 2.614678899082569, + "grad_norm": 7.9690752029418945, + "learning_rate": 4.6134265027280574e-05, + "loss": 0.544, + "num_input_tokens_seen": 2665008, + "step": 9975 + }, + { + "epoch": 2.615989515072084, + "grad_norm": 4.956567764282227, + "learning_rate": 4.6128153245508876e-05, + "loss": 0.4788, + "num_input_tokens_seen": 2666176, + "step": 9980 + }, + { + "epoch": 2.617300131061599, + "grad_norm": 2.098365068435669, + "learning_rate": 4.612203704165121e-05, + "loss": 0.5604, + "num_input_tokens_seen": 2667360, + "step": 9985 + }, + { + "epoch": 2.618610747051114, + "grad_norm": 1.1200007200241089, + "learning_rate": 4.6115916416987686e-05, + "loss": 0.5114, + "num_input_tokens_seen": 2668784, + "step": 9990 + }, + { + "epoch": 2.6199213630406293, + "grad_norm": 3.9829859733581543, + "learning_rate": 4.610979137279935e-05, + "loss": 0.4042, + "num_input_tokens_seen": 2670384, + "step": 9995 + }, + { + "epoch": 2.621231979030144, + "grad_norm": 2.4384894371032715, + "learning_rate": 4.6103661910368146e-05, + "loss": 0.4444, + "num_input_tokens_seen": 2671712, + "step": 10000 + }, + { + "epoch": 2.6225425950196595, + "grad_norm": 7.423447132110596, + "learning_rate": 4.6097528030976964e-05, + "loss": 0.5554, + "num_input_tokens_seen": 2672784, + "step": 10005 + }, + { + "epoch": 2.623853211009174, + "grad_norm": 2.2750494480133057, + "learning_rate": 4.609138973590963e-05, + "loss": 0.3814, + "num_input_tokens_seen": 2674416, + "step": 10010 + }, + { + "epoch": 2.625163826998689, + "grad_norm": 1.902856469154358, + "learning_rate": 4.6085247026450866e-05, + "loss": 0.4149, + "num_input_tokens_seen": 2675376, + "step": 10015 + }, + { + "epoch": 2.6264744429882043, + "grad_norm": 2.003051519393921, + "learning_rate": 4.607909990388632e-05, + "loss": 0.3803, + "num_input_tokens_seen": 2677152, + "step": 10020 + }, + { + "epoch": 2.6277850589777194, + "grad_norm": 4.11643123626709, + "learning_rate": 4.607294836950261e-05, + "loss": 0.5986, + "num_input_tokens_seen": 2678352, + "step": 10025 + }, + { + "epoch": 2.6290956749672345, + "grad_norm": 3.4655590057373047, + "learning_rate": 4.606679242458719e-05, + "loss": 0.552, + "num_input_tokens_seen": 2679536, + "step": 10030 + }, + { + "epoch": 2.6304062909567496, + "grad_norm": 1.6018112897872925, + "learning_rate": 4.606063207042853e-05, + "loss": 0.3974, + "num_input_tokens_seen": 2680688, + "step": 10035 + }, + { + "epoch": 2.6317169069462647, + "grad_norm": 1.6286158561706543, + "learning_rate": 4.6054467308315975e-05, + "loss": 0.4361, + "num_input_tokens_seen": 2682272, + "step": 10040 + }, + { + "epoch": 2.63302752293578, + "grad_norm": 3.970273494720459, + "learning_rate": 4.604829813953978e-05, + "loss": 0.4789, + "num_input_tokens_seen": 2683920, + "step": 10045 + }, + { + "epoch": 2.634338138925295, + "grad_norm": 1.9320080280303955, + "learning_rate": 4.6042124565391155e-05, + "loss": 0.5527, + "num_input_tokens_seen": 2685632, + "step": 10050 + }, + { + "epoch": 2.63564875491481, + "grad_norm": 1.319705843925476, + "learning_rate": 4.603594658716221e-05, + "loss": 0.5384, + "num_input_tokens_seen": 2687120, + "step": 10055 + }, + { + "epoch": 2.636959370904325, + "grad_norm": 0.47626397013664246, + "learning_rate": 4.6029764206146006e-05, + "loss": 0.4262, + "num_input_tokens_seen": 2689840, + "step": 10060 + }, + { + "epoch": 2.63826998689384, + "grad_norm": 2.6232306957244873, + "learning_rate": 4.602357742363649e-05, + "loss": 0.4768, + "num_input_tokens_seen": 2691376, + "step": 10065 + }, + { + "epoch": 2.6395806028833553, + "grad_norm": 1.4834895133972168, + "learning_rate": 4.601738624092853e-05, + "loss": 0.5129, + "num_input_tokens_seen": 2692800, + "step": 10070 + }, + { + "epoch": 2.6408912188728704, + "grad_norm": 2.512925386428833, + "learning_rate": 4.601119065931796e-05, + "loss": 0.5438, + "num_input_tokens_seen": 2694560, + "step": 10075 + }, + { + "epoch": 2.6422018348623855, + "grad_norm": 2.769681453704834, + "learning_rate": 4.600499068010148e-05, + "loss": 0.3629, + "num_input_tokens_seen": 2696160, + "step": 10080 + }, + { + "epoch": 2.6435124508519, + "grad_norm": 2.425704002380371, + "learning_rate": 4.599878630457674e-05, + "loss": 0.4219, + "num_input_tokens_seen": 2697504, + "step": 10085 + }, + { + "epoch": 2.6448230668414157, + "grad_norm": 1.4670308828353882, + "learning_rate": 4.599257753404233e-05, + "loss": 0.4473, + "num_input_tokens_seen": 2698896, + "step": 10090 + }, + { + "epoch": 2.6461336828309303, + "grad_norm": 2.2214972972869873, + "learning_rate": 4.59863643697977e-05, + "loss": 0.6086, + "num_input_tokens_seen": 2700368, + "step": 10095 + }, + { + "epoch": 2.647444298820446, + "grad_norm": 1.7149795293807983, + "learning_rate": 4.5980146813143264e-05, + "loss": 0.67, + "num_input_tokens_seen": 2701488, + "step": 10100 + }, + { + "epoch": 2.6487549148099605, + "grad_norm": 2.4307682514190674, + "learning_rate": 4.597392486538036e-05, + "loss": 0.4992, + "num_input_tokens_seen": 2702752, + "step": 10105 + }, + { + "epoch": 2.6500655307994756, + "grad_norm": 2.1746985912323, + "learning_rate": 4.596769852781122e-05, + "loss": 0.5029, + "num_input_tokens_seen": 2703952, + "step": 10110 + }, + { + "epoch": 2.6513761467889907, + "grad_norm": 1.4726836681365967, + "learning_rate": 4.596146780173901e-05, + "loss": 0.7294, + "num_input_tokens_seen": 2705024, + "step": 10115 + }, + { + "epoch": 2.652686762778506, + "grad_norm": 4.971144676208496, + "learning_rate": 4.59552326884678e-05, + "loss": 0.7584, + "num_input_tokens_seen": 2706224, + "step": 10120 + }, + { + "epoch": 2.653997378768021, + "grad_norm": 1.3295507431030273, + "learning_rate": 4.5948993189302605e-05, + "loss": 0.2889, + "num_input_tokens_seen": 2707552, + "step": 10125 + }, + { + "epoch": 2.655307994757536, + "grad_norm": 1.77785325050354, + "learning_rate": 4.5942749305549325e-05, + "loss": 0.4997, + "num_input_tokens_seen": 2708880, + "step": 10130 + }, + { + "epoch": 2.656618610747051, + "grad_norm": 2.5494515895843506, + "learning_rate": 4.593650103851481e-05, + "loss": 0.4962, + "num_input_tokens_seen": 2710192, + "step": 10135 + }, + { + "epoch": 2.657929226736566, + "grad_norm": 4.323812484741211, + "learning_rate": 4.5930248389506794e-05, + "loss": 0.5173, + "num_input_tokens_seen": 2711472, + "step": 10140 + }, + { + "epoch": 2.6592398427260813, + "grad_norm": 2.2826671600341797, + "learning_rate": 4.592399135983396e-05, + "loss": 0.4772, + "num_input_tokens_seen": 2712736, + "step": 10145 + }, + { + "epoch": 2.6605504587155964, + "grad_norm": 6.012180328369141, + "learning_rate": 4.591772995080589e-05, + "loss": 0.4384, + "num_input_tokens_seen": 2714512, + "step": 10150 + }, + { + "epoch": 2.6618610747051115, + "grad_norm": 14.274515151977539, + "learning_rate": 4.591146416373308e-05, + "loss": 0.5084, + "num_input_tokens_seen": 2715648, + "step": 10155 + }, + { + "epoch": 2.6631716906946266, + "grad_norm": 2.794477939605713, + "learning_rate": 4.590519399992695e-05, + "loss": 0.3885, + "num_input_tokens_seen": 2716928, + "step": 10160 + }, + { + "epoch": 2.6644823066841417, + "grad_norm": 1.4099982976913452, + "learning_rate": 4.5898919460699845e-05, + "loss": 0.4853, + "num_input_tokens_seen": 2718048, + "step": 10165 + }, + { + "epoch": 2.6657929226736568, + "grad_norm": 1.9766072034835815, + "learning_rate": 4.589264054736501e-05, + "loss": 0.4364, + "num_input_tokens_seen": 2719600, + "step": 10170 + }, + { + "epoch": 2.667103538663172, + "grad_norm": 1.656069278717041, + "learning_rate": 4.58863572612366e-05, + "loss": 0.5868, + "num_input_tokens_seen": 2721168, + "step": 10175 + }, + { + "epoch": 2.6684141546526865, + "grad_norm": 4.460202693939209, + "learning_rate": 4.5880069603629705e-05, + "loss": 0.4468, + "num_input_tokens_seen": 2722320, + "step": 10180 + }, + { + "epoch": 2.669724770642202, + "grad_norm": 3.5915560722351074, + "learning_rate": 4.587377757586032e-05, + "loss": 0.7527, + "num_input_tokens_seen": 2723552, + "step": 10185 + }, + { + "epoch": 2.6710353866317167, + "grad_norm": 22.65911293029785, + "learning_rate": 4.586748117924536e-05, + "loss": 0.457, + "num_input_tokens_seen": 2724512, + "step": 10190 + }, + { + "epoch": 2.6723460026212322, + "grad_norm": 1.5070809125900269, + "learning_rate": 4.586118041510264e-05, + "loss": 0.4049, + "num_input_tokens_seen": 2726000, + "step": 10195 + }, + { + "epoch": 2.673656618610747, + "grad_norm": 1.6305148601531982, + "learning_rate": 4.585487528475091e-05, + "loss": 0.4829, + "num_input_tokens_seen": 2728544, + "step": 10200 + }, + { + "epoch": 2.674967234600262, + "grad_norm": 1.8434317111968994, + "learning_rate": 4.584856578950981e-05, + "loss": 0.5427, + "num_input_tokens_seen": 2729712, + "step": 10205 + }, + { + "epoch": 2.676277850589777, + "grad_norm": 1.5459243059158325, + "learning_rate": 4.584225193069992e-05, + "loss": 0.3611, + "num_input_tokens_seen": 2730832, + "step": 10210 + }, + { + "epoch": 2.677588466579292, + "grad_norm": 4.889645576477051, + "learning_rate": 4.583593370964271e-05, + "loss": 0.3375, + "num_input_tokens_seen": 2731968, + "step": 10215 + }, + { + "epoch": 2.6788990825688073, + "grad_norm": 2.3207244873046875, + "learning_rate": 4.582961112766058e-05, + "loss": 0.4102, + "num_input_tokens_seen": 2733280, + "step": 10220 + }, + { + "epoch": 2.6802096985583224, + "grad_norm": 1.808450698852539, + "learning_rate": 4.582328418607682e-05, + "loss": 0.4859, + "num_input_tokens_seen": 2734432, + "step": 10225 + }, + { + "epoch": 2.6815203145478375, + "grad_norm": 1.8084566593170166, + "learning_rate": 4.581695288621568e-05, + "loss": 0.4735, + "num_input_tokens_seen": 2735760, + "step": 10230 + }, + { + "epoch": 2.6828309305373526, + "grad_norm": 2.3234200477600098, + "learning_rate": 4.581061722940225e-05, + "loss": 0.4177, + "num_input_tokens_seen": 2737920, + "step": 10235 + }, + { + "epoch": 2.6841415465268676, + "grad_norm": 2.3827731609344482, + "learning_rate": 4.580427721696261e-05, + "loss": 0.4686, + "num_input_tokens_seen": 2739328, + "step": 10240 + }, + { + "epoch": 2.6854521625163827, + "grad_norm": 3.0731940269470215, + "learning_rate": 4.579793285022368e-05, + "loss": 0.3933, + "num_input_tokens_seen": 2740928, + "step": 10245 + }, + { + "epoch": 2.686762778505898, + "grad_norm": 4.113832473754883, + "learning_rate": 4.579158413051335e-05, + "loss": 0.5559, + "num_input_tokens_seen": 2742256, + "step": 10250 + }, + { + "epoch": 2.688073394495413, + "grad_norm": 3.1094930171966553, + "learning_rate": 4.578523105916038e-05, + "loss": 0.5487, + "num_input_tokens_seen": 2743408, + "step": 10255 + }, + { + "epoch": 2.689384010484928, + "grad_norm": 2.816028356552124, + "learning_rate": 4.577887363749447e-05, + "loss": 0.4203, + "num_input_tokens_seen": 2744464, + "step": 10260 + }, + { + "epoch": 2.690694626474443, + "grad_norm": 22.957931518554688, + "learning_rate": 4.57725118668462e-05, + "loss": 0.5278, + "num_input_tokens_seen": 2745392, + "step": 10265 + }, + { + "epoch": 2.6920052424639582, + "grad_norm": 2.403311014175415, + "learning_rate": 4.5766145748547104e-05, + "loss": 0.6668, + "num_input_tokens_seen": 2746848, + "step": 10270 + }, + { + "epoch": 2.693315858453473, + "grad_norm": 2.952394485473633, + "learning_rate": 4.575977528392957e-05, + "loss": 0.4211, + "num_input_tokens_seen": 2748256, + "step": 10275 + }, + { + "epoch": 2.6946264744429884, + "grad_norm": 2.195614814758301, + "learning_rate": 4.5753400474326946e-05, + "loss": 0.4865, + "num_input_tokens_seen": 2749808, + "step": 10280 + }, + { + "epoch": 2.695937090432503, + "grad_norm": 1.3577419519424438, + "learning_rate": 4.574702132107346e-05, + "loss": 0.3248, + "num_input_tokens_seen": 2751008, + "step": 10285 + }, + { + "epoch": 2.6972477064220186, + "grad_norm": 2.1854894161224365, + "learning_rate": 4.5740637825504265e-05, + "loss": 0.3977, + "num_input_tokens_seen": 2752656, + "step": 10290 + }, + { + "epoch": 2.6985583224115333, + "grad_norm": 1.9751794338226318, + "learning_rate": 4.5734249988955405e-05, + "loss": 0.4924, + "num_input_tokens_seen": 2754256, + "step": 10295 + }, + { + "epoch": 2.6998689384010484, + "grad_norm": 2.9684109687805176, + "learning_rate": 4.572785781276385e-05, + "loss": 0.5119, + "num_input_tokens_seen": 2755168, + "step": 10300 + }, + { + "epoch": 2.7011795543905635, + "grad_norm": 1.901380181312561, + "learning_rate": 4.5721461298267466e-05, + "loss": 0.4642, + "num_input_tokens_seen": 2756528, + "step": 10305 + }, + { + "epoch": 2.7024901703800785, + "grad_norm": 1.1800216436386108, + "learning_rate": 4.5715060446805044e-05, + "loss": 0.5062, + "num_input_tokens_seen": 2757872, + "step": 10310 + }, + { + "epoch": 2.7038007863695936, + "grad_norm": 2.182513952255249, + "learning_rate": 4.570865525971626e-05, + "loss": 0.3562, + "num_input_tokens_seen": 2758976, + "step": 10315 + }, + { + "epoch": 2.7051114023591087, + "grad_norm": 2.6239938735961914, + "learning_rate": 4.570224573834171e-05, + "loss": 0.5901, + "num_input_tokens_seen": 2760416, + "step": 10320 + }, + { + "epoch": 2.706422018348624, + "grad_norm": 2.6250088214874268, + "learning_rate": 4.569583188402289e-05, + "loss": 0.5629, + "num_input_tokens_seen": 2761568, + "step": 10325 + }, + { + "epoch": 2.707732634338139, + "grad_norm": 3.3835856914520264, + "learning_rate": 4.5689413698102226e-05, + "loss": 0.7292, + "num_input_tokens_seen": 2762960, + "step": 10330 + }, + { + "epoch": 2.709043250327654, + "grad_norm": 2.9972951412200928, + "learning_rate": 4.5682991181923024e-05, + "loss": 0.6019, + "num_input_tokens_seen": 2764336, + "step": 10335 + }, + { + "epoch": 2.710353866317169, + "grad_norm": 2.2751100063323975, + "learning_rate": 4.56765643368295e-05, + "loss": 0.4993, + "num_input_tokens_seen": 2765760, + "step": 10340 + }, + { + "epoch": 2.711664482306684, + "grad_norm": 8.37344741821289, + "learning_rate": 4.5670133164166795e-05, + "loss": 0.4182, + "num_input_tokens_seen": 2766752, + "step": 10345 + }, + { + "epoch": 2.7129750982961993, + "grad_norm": 1.9089339971542358, + "learning_rate": 4.566369766528093e-05, + "loss": 0.3728, + "num_input_tokens_seen": 2768192, + "step": 10350 + }, + { + "epoch": 2.7142857142857144, + "grad_norm": 3.0410733222961426, + "learning_rate": 4.565725784151884e-05, + "loss": 0.6498, + "num_input_tokens_seen": 2769568, + "step": 10355 + }, + { + "epoch": 2.7155963302752295, + "grad_norm": 4.3213605880737305, + "learning_rate": 4.565081369422839e-05, + "loss": 0.3239, + "num_input_tokens_seen": 2770624, + "step": 10360 + }, + { + "epoch": 2.7169069462647446, + "grad_norm": 1.647017478942871, + "learning_rate": 4.56443652247583e-05, + "loss": 0.5246, + "num_input_tokens_seen": 2771952, + "step": 10365 + }, + { + "epoch": 2.7182175622542593, + "grad_norm": 2.3277809619903564, + "learning_rate": 4.5637912434458254e-05, + "loss": 0.4776, + "num_input_tokens_seen": 2773184, + "step": 10370 + }, + { + "epoch": 2.719528178243775, + "grad_norm": 2.289469003677368, + "learning_rate": 4.563145532467878e-05, + "loss": 0.3821, + "num_input_tokens_seen": 2774480, + "step": 10375 + }, + { + "epoch": 2.7208387942332894, + "grad_norm": 2.6926333904266357, + "learning_rate": 4.562499389677137e-05, + "loss": 0.4755, + "num_input_tokens_seen": 2775600, + "step": 10380 + }, + { + "epoch": 2.722149410222805, + "grad_norm": 4.430717468261719, + "learning_rate": 4.561852815208837e-05, + "loss": 0.5544, + "num_input_tokens_seen": 2776784, + "step": 10385 + }, + { + "epoch": 2.7234600262123196, + "grad_norm": 1.6392115354537964, + "learning_rate": 4.561205809198305e-05, + "loss": 0.3492, + "num_input_tokens_seen": 2778064, + "step": 10390 + }, + { + "epoch": 2.7247706422018347, + "grad_norm": 2.36397385597229, + "learning_rate": 4.560558371780959e-05, + "loss": 0.7192, + "num_input_tokens_seen": 2778960, + "step": 10395 + }, + { + "epoch": 2.72608125819135, + "grad_norm": 1.6050035953521729, + "learning_rate": 4.559910503092306e-05, + "loss": 0.4928, + "num_input_tokens_seen": 2780816, + "step": 10400 + }, + { + "epoch": 2.727391874180865, + "grad_norm": 1.6741303205490112, + "learning_rate": 4.559262203267945e-05, + "loss": 0.3756, + "num_input_tokens_seen": 2782368, + "step": 10405 + }, + { + "epoch": 2.72870249017038, + "grad_norm": 1.6331120729446411, + "learning_rate": 4.5586134724435614e-05, + "loss": 0.505, + "num_input_tokens_seen": 2783520, + "step": 10410 + }, + { + "epoch": 2.730013106159895, + "grad_norm": 1.6170871257781982, + "learning_rate": 4.557964310754935e-05, + "loss": 0.6089, + "num_input_tokens_seen": 2784672, + "step": 10415 + }, + { + "epoch": 2.73132372214941, + "grad_norm": 3.6422555446624756, + "learning_rate": 4.557314718337935e-05, + "loss": 0.4892, + "num_input_tokens_seen": 2785888, + "step": 10420 + }, + { + "epoch": 2.7326343381389253, + "grad_norm": 3.3657760620117188, + "learning_rate": 4.556664695328519e-05, + "loss": 0.5523, + "num_input_tokens_seen": 2786944, + "step": 10425 + }, + { + "epoch": 2.7339449541284404, + "grad_norm": 4.561648368835449, + "learning_rate": 4.556014241862736e-05, + "loss": 0.5402, + "num_input_tokens_seen": 2788496, + "step": 10430 + }, + { + "epoch": 2.7352555701179555, + "grad_norm": 3.699413299560547, + "learning_rate": 4.555363358076725e-05, + "loss": 0.4861, + "num_input_tokens_seen": 2789424, + "step": 10435 + }, + { + "epoch": 2.7365661861074706, + "grad_norm": 1.7631144523620605, + "learning_rate": 4.5547120441067136e-05, + "loss": 0.8235, + "num_input_tokens_seen": 2792544, + "step": 10440 + }, + { + "epoch": 2.7378768020969857, + "grad_norm": 1.0908747911453247, + "learning_rate": 4.554060300089022e-05, + "loss": 0.46, + "num_input_tokens_seen": 2794112, + "step": 10445 + }, + { + "epoch": 2.739187418086501, + "grad_norm": 1.934845209121704, + "learning_rate": 4.553408126160059e-05, + "loss": 0.4645, + "num_input_tokens_seen": 2795376, + "step": 10450 + }, + { + "epoch": 2.740498034076016, + "grad_norm": 1.5523455142974854, + "learning_rate": 4.552755522456324e-05, + "loss": 0.5206, + "num_input_tokens_seen": 2796784, + "step": 10455 + }, + { + "epoch": 2.741808650065531, + "grad_norm": 2.0946848392486572, + "learning_rate": 4.552102489114404e-05, + "loss": 0.705, + "num_input_tokens_seen": 2798192, + "step": 10460 + }, + { + "epoch": 2.7431192660550456, + "grad_norm": 2.620021343231201, + "learning_rate": 4.551449026270979e-05, + "loss": 0.5089, + "num_input_tokens_seen": 2799504, + "step": 10465 + }, + { + "epoch": 2.744429882044561, + "grad_norm": 1.5701767206192017, + "learning_rate": 4.550795134062817e-05, + "loss": 0.4646, + "num_input_tokens_seen": 2800640, + "step": 10470 + }, + { + "epoch": 2.745740498034076, + "grad_norm": 3.15548038482666, + "learning_rate": 4.5501408126267784e-05, + "loss": 0.5683, + "num_input_tokens_seen": 2801632, + "step": 10475 + }, + { + "epoch": 2.747051114023591, + "grad_norm": 2.789214849472046, + "learning_rate": 4.5494860620998086e-05, + "loss": 0.6435, + "num_input_tokens_seen": 2803344, + "step": 10480 + }, + { + "epoch": 2.748361730013106, + "grad_norm": 1.5315790176391602, + "learning_rate": 4.548830882618948e-05, + "loss": 0.4481, + "num_input_tokens_seen": 2804608, + "step": 10485 + }, + { + "epoch": 2.749672346002621, + "grad_norm": 2.669574499130249, + "learning_rate": 4.5481752743213234e-05, + "loss": 0.3816, + "num_input_tokens_seen": 2805824, + "step": 10490 + }, + { + "epoch": 2.750982961992136, + "grad_norm": 1.5830835103988647, + "learning_rate": 4.5475192373441525e-05, + "loss": 0.3538, + "num_input_tokens_seen": 2807104, + "step": 10495 + }, + { + "epoch": 2.7522935779816513, + "grad_norm": 1.5842366218566895, + "learning_rate": 4.5468627718247436e-05, + "loss": 0.309, + "num_input_tokens_seen": 2808480, + "step": 10500 + }, + { + "epoch": 2.7536041939711664, + "grad_norm": 5.006476402282715, + "learning_rate": 4.546205877900494e-05, + "loss": 0.3458, + "num_input_tokens_seen": 2810672, + "step": 10505 + }, + { + "epoch": 2.7549148099606815, + "grad_norm": 4.330968856811523, + "learning_rate": 4.5455485557088885e-05, + "loss": 0.5629, + "num_input_tokens_seen": 2811888, + "step": 10510 + }, + { + "epoch": 2.7562254259501966, + "grad_norm": 1.9988409280776978, + "learning_rate": 4.544890805387505e-05, + "loss": 0.386, + "num_input_tokens_seen": 2813200, + "step": 10515 + }, + { + "epoch": 2.7575360419397117, + "grad_norm": 5.715256690979004, + "learning_rate": 4.544232627074009e-05, + "loss": 0.6241, + "num_input_tokens_seen": 2814688, + "step": 10520 + }, + { + "epoch": 2.758846657929227, + "grad_norm": 2.0248453617095947, + "learning_rate": 4.5435740209061574e-05, + "loss": 0.5122, + "num_input_tokens_seen": 2815952, + "step": 10525 + }, + { + "epoch": 2.760157273918742, + "grad_norm": 1.617701530456543, + "learning_rate": 4.5429149870217936e-05, + "loss": 0.428, + "num_input_tokens_seen": 2817296, + "step": 10530 + }, + { + "epoch": 2.761467889908257, + "grad_norm": 2.841848134994507, + "learning_rate": 4.5422555255588524e-05, + "loss": 0.4493, + "num_input_tokens_seen": 2819168, + "step": 10535 + }, + { + "epoch": 2.762778505897772, + "grad_norm": 2.224156618118286, + "learning_rate": 4.541595636655359e-05, + "loss": 0.4233, + "num_input_tokens_seen": 2820400, + "step": 10540 + }, + { + "epoch": 2.764089121887287, + "grad_norm": 5.240928649902344, + "learning_rate": 4.540935320449426e-05, + "loss": 0.2948, + "num_input_tokens_seen": 2821536, + "step": 10545 + }, + { + "epoch": 2.765399737876802, + "grad_norm": 1.603045105934143, + "learning_rate": 4.540274577079258e-05, + "loss": 0.4089, + "num_input_tokens_seen": 2822640, + "step": 10550 + }, + { + "epoch": 2.7667103538663174, + "grad_norm": 1.5635836124420166, + "learning_rate": 4.539613406683145e-05, + "loss": 0.5467, + "num_input_tokens_seen": 2824432, + "step": 10555 + }, + { + "epoch": 2.768020969855832, + "grad_norm": 2.9079196453094482, + "learning_rate": 4.538951809399471e-05, + "loss": 0.378, + "num_input_tokens_seen": 2825856, + "step": 10560 + }, + { + "epoch": 2.7693315858453476, + "grad_norm": 1.8626960515975952, + "learning_rate": 4.538289785366706e-05, + "loss": 0.6403, + "num_input_tokens_seen": 2826912, + "step": 10565 + }, + { + "epoch": 2.770642201834862, + "grad_norm": 4.413777828216553, + "learning_rate": 4.53762733472341e-05, + "loss": 0.7466, + "num_input_tokens_seen": 2828240, + "step": 10570 + }, + { + "epoch": 2.7719528178243773, + "grad_norm": 1.7432724237442017, + "learning_rate": 4.536964457608235e-05, + "loss": 0.513, + "num_input_tokens_seen": 2829440, + "step": 10575 + }, + { + "epoch": 2.7732634338138924, + "grad_norm": 1.8371424674987793, + "learning_rate": 4.5363011541599185e-05, + "loss": 0.3752, + "num_input_tokens_seen": 2830736, + "step": 10580 + }, + { + "epoch": 2.7745740498034075, + "grad_norm": 3.923424005508423, + "learning_rate": 4.535637424517288e-05, + "loss": 0.5447, + "num_input_tokens_seen": 2832272, + "step": 10585 + }, + { + "epoch": 2.7758846657929226, + "grad_norm": 2.0506699085235596, + "learning_rate": 4.5349732688192624e-05, + "loss": 0.6311, + "num_input_tokens_seen": 2834000, + "step": 10590 + }, + { + "epoch": 2.7771952817824377, + "grad_norm": 4.217064380645752, + "learning_rate": 4.534308687204848e-05, + "loss": 0.3589, + "num_input_tokens_seen": 2835312, + "step": 10595 + }, + { + "epoch": 2.778505897771953, + "grad_norm": 1.7261381149291992, + "learning_rate": 4.53364367981314e-05, + "loss": 0.5475, + "num_input_tokens_seen": 2836720, + "step": 10600 + }, + { + "epoch": 2.779816513761468, + "grad_norm": 2.515169382095337, + "learning_rate": 4.5329782467833236e-05, + "loss": 0.5041, + "num_input_tokens_seen": 2838160, + "step": 10605 + }, + { + "epoch": 2.781127129750983, + "grad_norm": 2.505087375640869, + "learning_rate": 4.532312388254674e-05, + "loss": 0.5371, + "num_input_tokens_seen": 2839120, + "step": 10610 + }, + { + "epoch": 2.782437745740498, + "grad_norm": 3.083786964416504, + "learning_rate": 4.5316461043665516e-05, + "loss": 0.4989, + "num_input_tokens_seen": 2840496, + "step": 10615 + }, + { + "epoch": 2.783748361730013, + "grad_norm": 1.841678500175476, + "learning_rate": 4.53097939525841e-05, + "loss": 0.3901, + "num_input_tokens_seen": 2841824, + "step": 10620 + }, + { + "epoch": 2.7850589777195283, + "grad_norm": 2.7215964794158936, + "learning_rate": 4.530312261069791e-05, + "loss": 0.6655, + "num_input_tokens_seen": 2843024, + "step": 10625 + }, + { + "epoch": 2.7863695937090434, + "grad_norm": 2.122166872024536, + "learning_rate": 4.5296447019403245e-05, + "loss": 0.5157, + "num_input_tokens_seen": 2844112, + "step": 10630 + }, + { + "epoch": 2.7876802096985585, + "grad_norm": 2.4318978786468506, + "learning_rate": 4.5289767180097275e-05, + "loss": 0.5038, + "num_input_tokens_seen": 2845424, + "step": 10635 + }, + { + "epoch": 2.7889908256880735, + "grad_norm": 3.6948297023773193, + "learning_rate": 4.528308309417809e-05, + "loss": 0.4014, + "num_input_tokens_seen": 2847056, + "step": 10640 + }, + { + "epoch": 2.790301441677588, + "grad_norm": 3.1528115272521973, + "learning_rate": 4.527639476304467e-05, + "loss": 0.3596, + "num_input_tokens_seen": 2848400, + "step": 10645 + }, + { + "epoch": 2.7916120576671037, + "grad_norm": 2.8896331787109375, + "learning_rate": 4.526970218809686e-05, + "loss": 0.5715, + "num_input_tokens_seen": 2849472, + "step": 10650 + }, + { + "epoch": 2.7929226736566184, + "grad_norm": 20.514074325561523, + "learning_rate": 4.52630053707354e-05, + "loss": 0.6169, + "num_input_tokens_seen": 2850544, + "step": 10655 + }, + { + "epoch": 2.794233289646134, + "grad_norm": 1.320465087890625, + "learning_rate": 4.525630431236193e-05, + "loss": 0.5277, + "num_input_tokens_seen": 2851696, + "step": 10660 + }, + { + "epoch": 2.7955439056356486, + "grad_norm": 1.6207399368286133, + "learning_rate": 4.524959901437897e-05, + "loss": 0.5087, + "num_input_tokens_seen": 2853056, + "step": 10665 + }, + { + "epoch": 2.7968545216251637, + "grad_norm": 1.7495367527008057, + "learning_rate": 4.5242889478189926e-05, + "loss": 0.486, + "num_input_tokens_seen": 2854688, + "step": 10670 + }, + { + "epoch": 2.7981651376146788, + "grad_norm": 4.932611465454102, + "learning_rate": 4.5236175705199094e-05, + "loss": 0.3688, + "num_input_tokens_seen": 2855680, + "step": 10675 + }, + { + "epoch": 2.799475753604194, + "grad_norm": 2.2119972705841064, + "learning_rate": 4.522945769681164e-05, + "loss": 0.5148, + "num_input_tokens_seen": 2856880, + "step": 10680 + }, + { + "epoch": 2.800786369593709, + "grad_norm": 2.335390567779541, + "learning_rate": 4.522273545443365e-05, + "loss": 0.8708, + "num_input_tokens_seen": 2858032, + "step": 10685 + }, + { + "epoch": 2.802096985583224, + "grad_norm": 3.1034085750579834, + "learning_rate": 4.521600897947208e-05, + "loss": 0.5494, + "num_input_tokens_seen": 2859120, + "step": 10690 + }, + { + "epoch": 2.803407601572739, + "grad_norm": 1.6087130308151245, + "learning_rate": 4.520927827333475e-05, + "loss": 0.3514, + "num_input_tokens_seen": 2860352, + "step": 10695 + }, + { + "epoch": 2.8047182175622543, + "grad_norm": 1.7872446775436401, + "learning_rate": 4.5202543337430405e-05, + "loss": 0.27, + "num_input_tokens_seen": 2861664, + "step": 10700 + }, + { + "epoch": 2.8060288335517694, + "grad_norm": 4.681520938873291, + "learning_rate": 4.519580417316863e-05, + "loss": 0.603, + "num_input_tokens_seen": 2863376, + "step": 10705 + }, + { + "epoch": 2.8073394495412844, + "grad_norm": 0.7479950785636902, + "learning_rate": 4.518906078195996e-05, + "loss": 0.4856, + "num_input_tokens_seen": 2865312, + "step": 10710 + }, + { + "epoch": 2.8086500655307995, + "grad_norm": 1.9851285219192505, + "learning_rate": 4.518231316521574e-05, + "loss": 0.2789, + "num_input_tokens_seen": 2866512, + "step": 10715 + }, + { + "epoch": 2.8099606815203146, + "grad_norm": 2.820859909057617, + "learning_rate": 4.517556132434825e-05, + "loss": 0.6319, + "num_input_tokens_seen": 2867744, + "step": 10720 + }, + { + "epoch": 2.8112712975098297, + "grad_norm": 1.6085742712020874, + "learning_rate": 4.5168805260770633e-05, + "loss": 0.4577, + "num_input_tokens_seen": 2868800, + "step": 10725 + }, + { + "epoch": 2.812581913499345, + "grad_norm": 2.012608766555786, + "learning_rate": 4.516204497589692e-05, + "loss": 0.4142, + "num_input_tokens_seen": 2869856, + "step": 10730 + }, + { + "epoch": 2.81389252948886, + "grad_norm": 1.122301697731018, + "learning_rate": 4.5155280471142036e-05, + "loss": 0.4134, + "num_input_tokens_seen": 2871728, + "step": 10735 + }, + { + "epoch": 2.8152031454783746, + "grad_norm": 1.203264594078064, + "learning_rate": 4.514851174792177e-05, + "loss": 0.5999, + "num_input_tokens_seen": 2872848, + "step": 10740 + }, + { + "epoch": 2.81651376146789, + "grad_norm": 2.0182106494903564, + "learning_rate": 4.514173880765282e-05, + "loss": 0.401, + "num_input_tokens_seen": 2874128, + "step": 10745 + }, + { + "epoch": 2.8178243774574048, + "grad_norm": 2.78372859954834, + "learning_rate": 4.513496165175273e-05, + "loss": 0.4584, + "num_input_tokens_seen": 2875344, + "step": 10750 + }, + { + "epoch": 2.8191349934469203, + "grad_norm": 0.8767490983009338, + "learning_rate": 4.512818028163996e-05, + "loss": 0.3089, + "num_input_tokens_seen": 2876800, + "step": 10755 + }, + { + "epoch": 2.820445609436435, + "grad_norm": 1.1107330322265625, + "learning_rate": 4.512139469873384e-05, + "loss": 0.4184, + "num_input_tokens_seen": 2878256, + "step": 10760 + }, + { + "epoch": 2.82175622542595, + "grad_norm": 1.5992916822433472, + "learning_rate": 4.511460490445457e-05, + "loss": 0.6235, + "num_input_tokens_seen": 2879584, + "step": 10765 + }, + { + "epoch": 2.823066841415465, + "grad_norm": 3.92269229888916, + "learning_rate": 4.510781090022325e-05, + "loss": 0.4464, + "num_input_tokens_seen": 2881552, + "step": 10770 + }, + { + "epoch": 2.8243774574049803, + "grad_norm": 2.0883240699768066, + "learning_rate": 4.5101012687461854e-05, + "loss": 0.4719, + "num_input_tokens_seen": 2882816, + "step": 10775 + }, + { + "epoch": 2.8256880733944953, + "grad_norm": 1.2291418313980103, + "learning_rate": 4.509421026759323e-05, + "loss": 0.3862, + "num_input_tokens_seen": 2883936, + "step": 10780 + }, + { + "epoch": 2.8269986893840104, + "grad_norm": 2.0311429500579834, + "learning_rate": 4.508740364204113e-05, + "loss": 0.7623, + "num_input_tokens_seen": 2885200, + "step": 10785 + }, + { + "epoch": 2.8283093053735255, + "grad_norm": 8.99860668182373, + "learning_rate": 4.508059281223015e-05, + "loss": 0.5296, + "num_input_tokens_seen": 2886464, + "step": 10790 + }, + { + "epoch": 2.8296199213630406, + "grad_norm": 1.9118987321853638, + "learning_rate": 4.507377777958578e-05, + "loss": 0.3899, + "num_input_tokens_seen": 2888272, + "step": 10795 + }, + { + "epoch": 2.8309305373525557, + "grad_norm": 1.4592528343200684, + "learning_rate": 4.5066958545534415e-05, + "loss": 0.4716, + "num_input_tokens_seen": 2889552, + "step": 10800 + }, + { + "epoch": 2.832241153342071, + "grad_norm": 1.9629768133163452, + "learning_rate": 4.50601351115033e-05, + "loss": 0.4695, + "num_input_tokens_seen": 2891248, + "step": 10805 + }, + { + "epoch": 2.833551769331586, + "grad_norm": 3.860483169555664, + "learning_rate": 4.5053307478920564e-05, + "loss": 0.4662, + "num_input_tokens_seen": 2892480, + "step": 10810 + }, + { + "epoch": 2.834862385321101, + "grad_norm": 3.717531204223633, + "learning_rate": 4.504647564921523e-05, + "loss": 0.4268, + "num_input_tokens_seen": 2893872, + "step": 10815 + }, + { + "epoch": 2.836173001310616, + "grad_norm": 3.472304344177246, + "learning_rate": 4.503963962381717e-05, + "loss": 0.6966, + "num_input_tokens_seen": 2894992, + "step": 10820 + }, + { + "epoch": 2.837483617300131, + "grad_norm": 2.601073980331421, + "learning_rate": 4.503279940415717e-05, + "loss": 0.435, + "num_input_tokens_seen": 2896304, + "step": 10825 + }, + { + "epoch": 2.8387942332896463, + "grad_norm": 3.078873872756958, + "learning_rate": 4.5025954991666865e-05, + "loss": 0.4597, + "num_input_tokens_seen": 2897456, + "step": 10830 + }, + { + "epoch": 2.840104849279161, + "grad_norm": 3.17694091796875, + "learning_rate": 4.501910638777878e-05, + "loss": 0.4752, + "num_input_tokens_seen": 2899328, + "step": 10835 + }, + { + "epoch": 2.8414154652686765, + "grad_norm": 4.385345935821533, + "learning_rate": 4.5012253593926324e-05, + "loss": 0.4496, + "num_input_tokens_seen": 2900400, + "step": 10840 + }, + { + "epoch": 2.842726081258191, + "grad_norm": 3.666724920272827, + "learning_rate": 4.5005396611543765e-05, + "loss": 0.4846, + "num_input_tokens_seen": 2901936, + "step": 10845 + }, + { + "epoch": 2.8440366972477067, + "grad_norm": 1.716086506843567, + "learning_rate": 4.499853544206626e-05, + "loss": 0.4508, + "num_input_tokens_seen": 2903344, + "step": 10850 + }, + { + "epoch": 2.8453473132372213, + "grad_norm": 2.9204514026641846, + "learning_rate": 4.499167008692985e-05, + "loss": 0.3899, + "num_input_tokens_seen": 2904784, + "step": 10855 + }, + { + "epoch": 2.8466579292267364, + "grad_norm": 1.6807935237884521, + "learning_rate": 4.4984800547571425e-05, + "loss": 0.5704, + "num_input_tokens_seen": 2906416, + "step": 10860 + }, + { + "epoch": 2.8479685452162515, + "grad_norm": 6.067851543426514, + "learning_rate": 4.497792682542878e-05, + "loss": 0.3562, + "num_input_tokens_seen": 2907728, + "step": 10865 + }, + { + "epoch": 2.8492791612057666, + "grad_norm": 2.301427125930786, + "learning_rate": 4.4971048921940575e-05, + "loss": 0.4579, + "num_input_tokens_seen": 2909296, + "step": 10870 + }, + { + "epoch": 2.8505897771952817, + "grad_norm": 2.2882072925567627, + "learning_rate": 4.496416683854633e-05, + "loss": 0.3952, + "num_input_tokens_seen": 2910320, + "step": 10875 + }, + { + "epoch": 2.851900393184797, + "grad_norm": 3.5585451126098633, + "learning_rate": 4.495728057668647e-05, + "loss": 0.4629, + "num_input_tokens_seen": 2911696, + "step": 10880 + }, + { + "epoch": 2.853211009174312, + "grad_norm": 3.6512436866760254, + "learning_rate": 4.4950390137802254e-05, + "loss": 0.5251, + "num_input_tokens_seen": 2913360, + "step": 10885 + }, + { + "epoch": 2.854521625163827, + "grad_norm": 3.3321218490600586, + "learning_rate": 4.4943495523335865e-05, + "loss": 0.3841, + "num_input_tokens_seen": 2914624, + "step": 10890 + }, + { + "epoch": 2.855832241153342, + "grad_norm": 2.7668230533599854, + "learning_rate": 4.493659673473032e-05, + "loss": 0.334, + "num_input_tokens_seen": 2916032, + "step": 10895 + }, + { + "epoch": 2.857142857142857, + "grad_norm": 2.8620681762695312, + "learning_rate": 4.492969377342952e-05, + "loss": 0.3667, + "num_input_tokens_seen": 2917488, + "step": 10900 + }, + { + "epoch": 2.8584534731323723, + "grad_norm": 1.5178301334381104, + "learning_rate": 4.4922786640878265e-05, + "loss": 0.5195, + "num_input_tokens_seen": 2919184, + "step": 10905 + }, + { + "epoch": 2.8597640891218874, + "grad_norm": 3.098388433456421, + "learning_rate": 4.491587533852218e-05, + "loss": 0.4827, + "num_input_tokens_seen": 2920816, + "step": 10910 + }, + { + "epoch": 2.8610747051114025, + "grad_norm": 3.5688517093658447, + "learning_rate": 4.49089598678078e-05, + "loss": 0.713, + "num_input_tokens_seen": 2921856, + "step": 10915 + }, + { + "epoch": 2.8623853211009176, + "grad_norm": 1.1795560121536255, + "learning_rate": 4.490204023018252e-05, + "loss": 0.404, + "num_input_tokens_seen": 2923232, + "step": 10920 + }, + { + "epoch": 2.8636959370904327, + "grad_norm": 1.8465912342071533, + "learning_rate": 4.4895116427094605e-05, + "loss": 0.4226, + "num_input_tokens_seen": 2924704, + "step": 10925 + }, + { + "epoch": 2.8650065530799473, + "grad_norm": 1.1754008531570435, + "learning_rate": 4.48881884599932e-05, + "loss": 0.394, + "num_input_tokens_seen": 2926160, + "step": 10930 + }, + { + "epoch": 2.866317169069463, + "grad_norm": 1.3917251825332642, + "learning_rate": 4.4881256330328314e-05, + "loss": 0.5861, + "num_input_tokens_seen": 2927520, + "step": 10935 + }, + { + "epoch": 2.8676277850589775, + "grad_norm": 3.87668776512146, + "learning_rate": 4.4874320039550834e-05, + "loss": 0.469, + "num_input_tokens_seen": 2928864, + "step": 10940 + }, + { + "epoch": 2.8689384010484926, + "grad_norm": 5.306812286376953, + "learning_rate": 4.48673795891125e-05, + "loss": 0.4452, + "num_input_tokens_seen": 2930208, + "step": 10945 + }, + { + "epoch": 2.8702490170380077, + "grad_norm": 1.8191885948181152, + "learning_rate": 4.4860434980465954e-05, + "loss": 0.4001, + "num_input_tokens_seen": 2931248, + "step": 10950 + }, + { + "epoch": 2.871559633027523, + "grad_norm": 1.6471731662750244, + "learning_rate": 4.485348621506468e-05, + "loss": 0.4747, + "num_input_tokens_seen": 2932448, + "step": 10955 + }, + { + "epoch": 2.872870249017038, + "grad_norm": 1.9866753816604614, + "learning_rate": 4.484653329436305e-05, + "loss": 0.5376, + "num_input_tokens_seen": 2934128, + "step": 10960 + }, + { + "epoch": 2.874180865006553, + "grad_norm": 19.272443771362305, + "learning_rate": 4.483957621981629e-05, + "loss": 1.2311, + "num_input_tokens_seen": 2935664, + "step": 10965 + }, + { + "epoch": 2.875491480996068, + "grad_norm": 1.7062511444091797, + "learning_rate": 4.483261499288051e-05, + "loss": 0.3927, + "num_input_tokens_seen": 2937024, + "step": 10970 + }, + { + "epoch": 2.876802096985583, + "grad_norm": 2.508085250854492, + "learning_rate": 4.482564961501268e-05, + "loss": 0.4526, + "num_input_tokens_seen": 2938320, + "step": 10975 + }, + { + "epoch": 2.8781127129750983, + "grad_norm": 3.0437278747558594, + "learning_rate": 4.481868008767063e-05, + "loss": 0.607, + "num_input_tokens_seen": 2939600, + "step": 10980 + }, + { + "epoch": 2.8794233289646134, + "grad_norm": 2.9499576091766357, + "learning_rate": 4.4811706412313094e-05, + "loss": 0.5124, + "num_input_tokens_seen": 2940912, + "step": 10985 + }, + { + "epoch": 2.8807339449541285, + "grad_norm": 2.7320213317871094, + "learning_rate": 4.4804728590399635e-05, + "loss": 0.555, + "num_input_tokens_seen": 2942320, + "step": 10990 + }, + { + "epoch": 2.8820445609436436, + "grad_norm": 12.866477012634277, + "learning_rate": 4.47977466233907e-05, + "loss": 0.552, + "num_input_tokens_seen": 2943808, + "step": 10995 + }, + { + "epoch": 2.8833551769331587, + "grad_norm": 3.231647491455078, + "learning_rate": 4.479076051274761e-05, + "loss": 0.4248, + "num_input_tokens_seen": 2945088, + "step": 11000 + }, + { + "epoch": 2.8846657929226738, + "grad_norm": 2.7546894550323486, + "learning_rate": 4.478377025993254e-05, + "loss": 0.4114, + "num_input_tokens_seen": 2946672, + "step": 11005 + }, + { + "epoch": 2.885976408912189, + "grad_norm": 1.9635006189346313, + "learning_rate": 4.477677586640854e-05, + "loss": 0.3829, + "num_input_tokens_seen": 2948032, + "step": 11010 + }, + { + "epoch": 2.8872870249017035, + "grad_norm": 4.011830806732178, + "learning_rate": 4.476977733363951e-05, + "loss": 0.4209, + "num_input_tokens_seen": 2949312, + "step": 11015 + }, + { + "epoch": 2.888597640891219, + "grad_norm": 2.0618956089019775, + "learning_rate": 4.4762774663090256e-05, + "loss": 0.3978, + "num_input_tokens_seen": 2950592, + "step": 11020 + }, + { + "epoch": 2.8899082568807337, + "grad_norm": 1.7478957176208496, + "learning_rate": 4.4755767856226405e-05, + "loss": 0.269, + "num_input_tokens_seen": 2951648, + "step": 11025 + }, + { + "epoch": 2.8912188728702493, + "grad_norm": 2.97165584564209, + "learning_rate": 4.474875691451448e-05, + "loss": 0.3358, + "num_input_tokens_seen": 2952992, + "step": 11030 + }, + { + "epoch": 2.892529488859764, + "grad_norm": 3.8978211879730225, + "learning_rate": 4.474174183942186e-05, + "loss": 0.5788, + "num_input_tokens_seen": 2954112, + "step": 11035 + }, + { + "epoch": 2.893840104849279, + "grad_norm": 1.041987657546997, + "learning_rate": 4.473472263241678e-05, + "loss": 0.5857, + "num_input_tokens_seen": 2955600, + "step": 11040 + }, + { + "epoch": 2.895150720838794, + "grad_norm": 2.9574146270751953, + "learning_rate": 4.472769929496835e-05, + "loss": 0.4378, + "num_input_tokens_seen": 2956656, + "step": 11045 + }, + { + "epoch": 2.896461336828309, + "grad_norm": 2.0535194873809814, + "learning_rate": 4.472067182854654e-05, + "loss": 0.4219, + "num_input_tokens_seen": 2957984, + "step": 11050 + }, + { + "epoch": 2.8977719528178243, + "grad_norm": 4.009273052215576, + "learning_rate": 4.4713640234622205e-05, + "loss": 0.8078, + "num_input_tokens_seen": 2959296, + "step": 11055 + }, + { + "epoch": 2.8990825688073394, + "grad_norm": 2.6250319480895996, + "learning_rate": 4.4706604514667025e-05, + "loss": 0.6162, + "num_input_tokens_seen": 2960512, + "step": 11060 + }, + { + "epoch": 2.9003931847968545, + "grad_norm": 3.1263394355773926, + "learning_rate": 4.469956467015357e-05, + "loss": 0.5002, + "num_input_tokens_seen": 2961616, + "step": 11065 + }, + { + "epoch": 2.9017038007863696, + "grad_norm": 2.137057304382324, + "learning_rate": 4.469252070255527e-05, + "loss": 0.4286, + "num_input_tokens_seen": 2962720, + "step": 11070 + }, + { + "epoch": 2.9030144167758847, + "grad_norm": 12.75526237487793, + "learning_rate": 4.4685472613346404e-05, + "loss": 0.8132, + "num_input_tokens_seen": 2963952, + "step": 11075 + }, + { + "epoch": 2.9043250327653998, + "grad_norm": 2.0231564044952393, + "learning_rate": 4.467842040400214e-05, + "loss": 0.4256, + "num_input_tokens_seen": 2965376, + "step": 11080 + }, + { + "epoch": 2.905635648754915, + "grad_norm": 1.5869437456130981, + "learning_rate": 4.467136407599849e-05, + "loss": 1.0067, + "num_input_tokens_seen": 2966416, + "step": 11085 + }, + { + "epoch": 2.90694626474443, + "grad_norm": 1.4180911779403687, + "learning_rate": 4.466430363081233e-05, + "loss": 0.632, + "num_input_tokens_seen": 2967568, + "step": 11090 + }, + { + "epoch": 2.908256880733945, + "grad_norm": 1.640818476676941, + "learning_rate": 4.465723906992139e-05, + "loss": 0.4264, + "num_input_tokens_seen": 2969120, + "step": 11095 + }, + { + "epoch": 2.90956749672346, + "grad_norm": 1.9054533243179321, + "learning_rate": 4.465017039480428e-05, + "loss": 0.5453, + "num_input_tokens_seen": 2970352, + "step": 11100 + }, + { + "epoch": 2.9108781127129753, + "grad_norm": 4.733019828796387, + "learning_rate": 4.464309760694047e-05, + "loss": 0.4029, + "num_input_tokens_seen": 2971936, + "step": 11105 + }, + { + "epoch": 2.91218872870249, + "grad_norm": 0.9008641242980957, + "learning_rate": 4.4636020707810254e-05, + "loss": 0.6491, + "num_input_tokens_seen": 2973792, + "step": 11110 + }, + { + "epoch": 2.9134993446920054, + "grad_norm": 6.505386829376221, + "learning_rate": 4.462893969889484e-05, + "loss": 0.5432, + "num_input_tokens_seen": 2974848, + "step": 11115 + }, + { + "epoch": 2.91480996068152, + "grad_norm": 1.1143999099731445, + "learning_rate": 4.462185458167626e-05, + "loss": 0.5094, + "num_input_tokens_seen": 2976336, + "step": 11120 + }, + { + "epoch": 2.9161205766710356, + "grad_norm": 3.0310773849487305, + "learning_rate": 4.4614765357637425e-05, + "loss": 0.5625, + "num_input_tokens_seen": 2977184, + "step": 11125 + }, + { + "epoch": 2.9174311926605503, + "grad_norm": 1.5992976427078247, + "learning_rate": 4.4607672028262094e-05, + "loss": 0.6119, + "num_input_tokens_seen": 2978384, + "step": 11130 + }, + { + "epoch": 2.9187418086500654, + "grad_norm": 1.925862431526184, + "learning_rate": 4.460057459503488e-05, + "loss": 0.9149, + "num_input_tokens_seen": 2979376, + "step": 11135 + }, + { + "epoch": 2.9200524246395805, + "grad_norm": 2.260990858078003, + "learning_rate": 4.4593473059441274e-05, + "loss": 0.5198, + "num_input_tokens_seen": 2981024, + "step": 11140 + }, + { + "epoch": 2.9213630406290956, + "grad_norm": 4.620781898498535, + "learning_rate": 4.458636742296761e-05, + "loss": 0.498, + "num_input_tokens_seen": 2982160, + "step": 11145 + }, + { + "epoch": 2.9226736566186107, + "grad_norm": 1.1274138689041138, + "learning_rate": 4.4579257687101097e-05, + "loss": 0.5918, + "num_input_tokens_seen": 2983344, + "step": 11150 + }, + { + "epoch": 2.9239842726081258, + "grad_norm": 1.428270697593689, + "learning_rate": 4.457214385332977e-05, + "loss": 0.5092, + "num_input_tokens_seen": 2984960, + "step": 11155 + }, + { + "epoch": 2.925294888597641, + "grad_norm": 2.4014663696289062, + "learning_rate": 4.456502592314256e-05, + "loss": 0.5678, + "num_input_tokens_seen": 2986480, + "step": 11160 + }, + { + "epoch": 2.926605504587156, + "grad_norm": 2.6413464546203613, + "learning_rate": 4.455790389802922e-05, + "loss": 0.4435, + "num_input_tokens_seen": 2987568, + "step": 11165 + }, + { + "epoch": 2.927916120576671, + "grad_norm": 1.5058637857437134, + "learning_rate": 4.4550777779480393e-05, + "loss": 0.7203, + "num_input_tokens_seen": 2988688, + "step": 11170 + }, + { + "epoch": 2.929226736566186, + "grad_norm": 1.707846760749817, + "learning_rate": 4.454364756898756e-05, + "loss": 0.4268, + "num_input_tokens_seen": 2990224, + "step": 11175 + }, + { + "epoch": 2.9305373525557012, + "grad_norm": 4.152074337005615, + "learning_rate": 4.453651326804305e-05, + "loss": 0.5259, + "num_input_tokens_seen": 2991376, + "step": 11180 + }, + { + "epoch": 2.9318479685452163, + "grad_norm": 5.045623302459717, + "learning_rate": 4.452937487814007e-05, + "loss": 0.3791, + "num_input_tokens_seen": 2992624, + "step": 11185 + }, + { + "epoch": 2.9331585845347314, + "grad_norm": 1.763620376586914, + "learning_rate": 4.452223240077269e-05, + "loss": 0.4222, + "num_input_tokens_seen": 2993808, + "step": 11190 + }, + { + "epoch": 2.9344692005242465, + "grad_norm": 3.1930127143859863, + "learning_rate": 4.451508583743578e-05, + "loss": 0.424, + "num_input_tokens_seen": 2995008, + "step": 11195 + }, + { + "epoch": 2.9357798165137616, + "grad_norm": 1.07848060131073, + "learning_rate": 4.450793518962514e-05, + "loss": 0.4454, + "num_input_tokens_seen": 2996832, + "step": 11200 + }, + { + "epoch": 2.9370904325032763, + "grad_norm": 18.87545394897461, + "learning_rate": 4.450078045883736e-05, + "loss": 0.4484, + "num_input_tokens_seen": 2998176, + "step": 11205 + }, + { + "epoch": 2.938401048492792, + "grad_norm": 1.732879638671875, + "learning_rate": 4.449362164656993e-05, + "loss": 0.6973, + "num_input_tokens_seen": 2999696, + "step": 11210 + }, + { + "epoch": 2.9397116644823065, + "grad_norm": 3.588994264602661, + "learning_rate": 4.448645875432117e-05, + "loss": 0.4011, + "num_input_tokens_seen": 3000960, + "step": 11215 + }, + { + "epoch": 2.941022280471822, + "grad_norm": 2.200674057006836, + "learning_rate": 4.447929178359026e-05, + "loss": 0.3895, + "num_input_tokens_seen": 3002480, + "step": 11220 + }, + { + "epoch": 2.9423328964613367, + "grad_norm": 1.5709534883499146, + "learning_rate": 4.4472120735877246e-05, + "loss": 0.6456, + "num_input_tokens_seen": 3003472, + "step": 11225 + }, + { + "epoch": 2.9436435124508518, + "grad_norm": 3.015183448791504, + "learning_rate": 4.4464945612683005e-05, + "loss": 0.4006, + "num_input_tokens_seen": 3004800, + "step": 11230 + }, + { + "epoch": 2.944954128440367, + "grad_norm": 11.075900077819824, + "learning_rate": 4.4457766415509275e-05, + "loss": 0.4602, + "num_input_tokens_seen": 3006144, + "step": 11235 + }, + { + "epoch": 2.946264744429882, + "grad_norm": 1.6623305082321167, + "learning_rate": 4.445058314585866e-05, + "loss": 0.4749, + "num_input_tokens_seen": 3007248, + "step": 11240 + }, + { + "epoch": 2.947575360419397, + "grad_norm": 1.883922815322876, + "learning_rate": 4.44433958052346e-05, + "loss": 0.5205, + "num_input_tokens_seen": 3008720, + "step": 11245 + }, + { + "epoch": 2.948885976408912, + "grad_norm": 1.826959252357483, + "learning_rate": 4.443620439514138e-05, + "loss": 0.5145, + "num_input_tokens_seen": 3009808, + "step": 11250 + }, + { + "epoch": 2.9501965923984272, + "grad_norm": 3.9475536346435547, + "learning_rate": 4.442900891708418e-05, + "loss": 0.4183, + "num_input_tokens_seen": 3011424, + "step": 11255 + }, + { + "epoch": 2.9515072083879423, + "grad_norm": 1.764565348625183, + "learning_rate": 4.442180937256898e-05, + "loss": 0.3155, + "num_input_tokens_seen": 3012912, + "step": 11260 + }, + { + "epoch": 2.9528178243774574, + "grad_norm": 2.1513495445251465, + "learning_rate": 4.4414605763102635e-05, + "loss": 0.5875, + "num_input_tokens_seen": 3014176, + "step": 11265 + }, + { + "epoch": 2.9541284403669725, + "grad_norm": 1.9770654439926147, + "learning_rate": 4.4407398090192846e-05, + "loss": 0.5003, + "num_input_tokens_seen": 3015408, + "step": 11270 + }, + { + "epoch": 2.9554390563564876, + "grad_norm": 3.090874433517456, + "learning_rate": 4.440018635534818e-05, + "loss": 0.6731, + "num_input_tokens_seen": 3016848, + "step": 11275 + }, + { + "epoch": 2.9567496723460027, + "grad_norm": 2.6312201023101807, + "learning_rate": 4.439297056007802e-05, + "loss": 0.4296, + "num_input_tokens_seen": 3018128, + "step": 11280 + }, + { + "epoch": 2.958060288335518, + "grad_norm": 1.783115029335022, + "learning_rate": 4.4385750705892634e-05, + "loss": 0.3698, + "num_input_tokens_seen": 3019408, + "step": 11285 + }, + { + "epoch": 2.959370904325033, + "grad_norm": 5.475338459014893, + "learning_rate": 4.437852679430313e-05, + "loss": 0.6066, + "num_input_tokens_seen": 3020576, + "step": 11290 + }, + { + "epoch": 2.960681520314548, + "grad_norm": 2.2062876224517822, + "learning_rate": 4.437129882682144e-05, + "loss": 0.3245, + "num_input_tokens_seen": 3022224, + "step": 11295 + }, + { + "epoch": 2.9619921363040627, + "grad_norm": 13.341873168945312, + "learning_rate": 4.436406680496038e-05, + "loss": 0.716, + "num_input_tokens_seen": 3023024, + "step": 11300 + }, + { + "epoch": 2.963302752293578, + "grad_norm": 3.207613706588745, + "learning_rate": 4.43568307302336e-05, + "loss": 0.4081, + "num_input_tokens_seen": 3024208, + "step": 11305 + }, + { + "epoch": 2.964613368283093, + "grad_norm": 1.01847505569458, + "learning_rate": 4.4349590604155586e-05, + "loss": 0.4909, + "num_input_tokens_seen": 3025680, + "step": 11310 + }, + { + "epoch": 2.9659239842726084, + "grad_norm": 4.47849178314209, + "learning_rate": 4.43423464282417e-05, + "loss": 0.4889, + "num_input_tokens_seen": 3027216, + "step": 11315 + }, + { + "epoch": 2.967234600262123, + "grad_norm": 2.3465209007263184, + "learning_rate": 4.4335098204008126e-05, + "loss": 0.3498, + "num_input_tokens_seen": 3028544, + "step": 11320 + }, + { + "epoch": 2.968545216251638, + "grad_norm": 7.180325984954834, + "learning_rate": 4.43278459329719e-05, + "loss": 0.4555, + "num_input_tokens_seen": 3029872, + "step": 11325 + }, + { + "epoch": 2.9698558322411532, + "grad_norm": 2.6988914012908936, + "learning_rate": 4.432058961665092e-05, + "loss": 0.4515, + "num_input_tokens_seen": 3031392, + "step": 11330 + }, + { + "epoch": 2.9711664482306683, + "grad_norm": 2.008669853210449, + "learning_rate": 4.4313329256563916e-05, + "loss": 0.5798, + "num_input_tokens_seen": 3032880, + "step": 11335 + }, + { + "epoch": 2.9724770642201834, + "grad_norm": 2.452549457550049, + "learning_rate": 4.430606485423048e-05, + "loss": 0.5687, + "num_input_tokens_seen": 3034080, + "step": 11340 + }, + { + "epoch": 2.9737876802096985, + "grad_norm": 3.13478946685791, + "learning_rate": 4.4298796411171015e-05, + "loss": 0.2999, + "num_input_tokens_seen": 3035472, + "step": 11345 + }, + { + "epoch": 2.9750982961992136, + "grad_norm": 2.088823080062866, + "learning_rate": 4.429152392890681e-05, + "loss": 0.5576, + "num_input_tokens_seen": 3036656, + "step": 11350 + }, + { + "epoch": 2.9764089121887287, + "grad_norm": 4.965540885925293, + "learning_rate": 4.428424740895998e-05, + "loss": 0.4073, + "num_input_tokens_seen": 3037872, + "step": 11355 + }, + { + "epoch": 2.977719528178244, + "grad_norm": 1.9139569997787476, + "learning_rate": 4.427696685285349e-05, + "loss": 0.5348, + "num_input_tokens_seen": 3039408, + "step": 11360 + }, + { + "epoch": 2.979030144167759, + "grad_norm": 5.205620765686035, + "learning_rate": 4.4269682262111145e-05, + "loss": 0.5703, + "num_input_tokens_seen": 3040768, + "step": 11365 + }, + { + "epoch": 2.980340760157274, + "grad_norm": 1.3152612447738647, + "learning_rate": 4.42623936382576e-05, + "loss": 0.4901, + "num_input_tokens_seen": 3042608, + "step": 11370 + }, + { + "epoch": 2.981651376146789, + "grad_norm": 52.26844787597656, + "learning_rate": 4.425510098281835e-05, + "loss": 0.6459, + "num_input_tokens_seen": 3044080, + "step": 11375 + }, + { + "epoch": 2.982961992136304, + "grad_norm": 1.9518824815750122, + "learning_rate": 4.424780429731973e-05, + "loss": 0.4267, + "num_input_tokens_seen": 3045568, + "step": 11380 + }, + { + "epoch": 2.9842726081258193, + "grad_norm": 2.5075523853302, + "learning_rate": 4.424050358328893e-05, + "loss": 0.5773, + "num_input_tokens_seen": 3046816, + "step": 11385 + }, + { + "epoch": 2.9855832241153344, + "grad_norm": 0.767970085144043, + "learning_rate": 4.423319884225398e-05, + "loss": 0.4237, + "num_input_tokens_seen": 3047888, + "step": 11390 + }, + { + "epoch": 2.986893840104849, + "grad_norm": 1.8207859992980957, + "learning_rate": 4.422589007574374e-05, + "loss": 0.6651, + "num_input_tokens_seen": 3049248, + "step": 11395 + }, + { + "epoch": 2.9882044560943646, + "grad_norm": 2.5170600414276123, + "learning_rate": 4.421857728528792e-05, + "loss": 0.4502, + "num_input_tokens_seen": 3050432, + "step": 11400 + }, + { + "epoch": 2.9895150720838792, + "grad_norm": 2.258277416229248, + "learning_rate": 4.421126047241709e-05, + "loss": 0.4471, + "num_input_tokens_seen": 3051504, + "step": 11405 + }, + { + "epoch": 2.9908256880733948, + "grad_norm": 2.1136908531188965, + "learning_rate": 4.420393963866263e-05, + "loss": 0.4688, + "num_input_tokens_seen": 3052912, + "step": 11410 + }, + { + "epoch": 2.9921363040629094, + "grad_norm": 3.495340585708618, + "learning_rate": 4.4196614785556795e-05, + "loss": 0.5631, + "num_input_tokens_seen": 3053888, + "step": 11415 + }, + { + "epoch": 2.9934469200524245, + "grad_norm": 4.824585437774658, + "learning_rate": 4.418928591463265e-05, + "loss": 0.5374, + "num_input_tokens_seen": 3055104, + "step": 11420 + }, + { + "epoch": 2.9947575360419396, + "grad_norm": 3.4825921058654785, + "learning_rate": 4.418195302742412e-05, + "loss": 0.5418, + "num_input_tokens_seen": 3056464, + "step": 11425 + }, + { + "epoch": 2.9960681520314547, + "grad_norm": 1.6084731817245483, + "learning_rate": 4.417461612546596e-05, + "loss": 0.5391, + "num_input_tokens_seen": 3057600, + "step": 11430 + }, + { + "epoch": 2.99737876802097, + "grad_norm": 4.006417751312256, + "learning_rate": 4.416727521029379e-05, + "loss": 0.3732, + "num_input_tokens_seen": 3058864, + "step": 11435 + }, + { + "epoch": 2.998689384010485, + "grad_norm": 0.9414385557174683, + "learning_rate": 4.415993028344403e-05, + "loss": 0.4277, + "num_input_tokens_seen": 3060208, + "step": 11440 + }, + { + "epoch": 3.0, + "grad_norm": 2.620678186416626, + "learning_rate": 4.415258134645396e-05, + "loss": 0.4507, + "num_input_tokens_seen": 3061400, + "step": 11445 + }, + { + "epoch": 3.000786369593709, + "eval_loss": 0.47063717246055603, + "eval_runtime": 18.2121, + "eval_samples_per_second": 46.562, + "eval_steps_per_second": 23.281, + "num_input_tokens_seen": 3062008, + "step": 11448 + }, + { + "epoch": 3.001310615989515, + "grad_norm": 2.5218098163604736, + "learning_rate": 4.414522840086172e-05, + "loss": 0.5379, + "num_input_tokens_seen": 3062600, + "step": 11450 + }, + { + "epoch": 3.00262123197903, + "grad_norm": 2.7438223361968994, + "learning_rate": 4.413787144820625e-05, + "loss": 0.3711, + "num_input_tokens_seen": 3063624, + "step": 11455 + }, + { + "epoch": 3.0039318479685453, + "grad_norm": 4.0030412673950195, + "learning_rate": 4.413051049002735e-05, + "loss": 0.4369, + "num_input_tokens_seen": 3064616, + "step": 11460 + }, + { + "epoch": 3.0052424639580604, + "grad_norm": 4.3358588218688965, + "learning_rate": 4.412314552786566e-05, + "loss": 0.4263, + "num_input_tokens_seen": 3065896, + "step": 11465 + }, + { + "epoch": 3.0065530799475755, + "grad_norm": 1.7853381633758545, + "learning_rate": 4.4115776563262656e-05, + "loss": 0.3355, + "num_input_tokens_seen": 3066968, + "step": 11470 + }, + { + "epoch": 3.0078636959370906, + "grad_norm": 2.1047019958496094, + "learning_rate": 4.410840359776065e-05, + "loss": 0.6096, + "num_input_tokens_seen": 3068024, + "step": 11475 + }, + { + "epoch": 3.0091743119266057, + "grad_norm": 2.9729182720184326, + "learning_rate": 4.410102663290278e-05, + "loss": 0.4483, + "num_input_tokens_seen": 3069032, + "step": 11480 + }, + { + "epoch": 3.0104849279161208, + "grad_norm": 1.8147703409194946, + "learning_rate": 4.4093645670233045e-05, + "loss": 0.5315, + "num_input_tokens_seen": 3070264, + "step": 11485 + }, + { + "epoch": 3.011795543905636, + "grad_norm": 2.8938639163970947, + "learning_rate": 4.4086260711296265e-05, + "loss": 0.5566, + "num_input_tokens_seen": 3071352, + "step": 11490 + }, + { + "epoch": 3.0131061598951505, + "grad_norm": 5.876029968261719, + "learning_rate": 4.407887175763809e-05, + "loss": 0.5431, + "num_input_tokens_seen": 3072664, + "step": 11495 + }, + { + "epoch": 3.0144167758846656, + "grad_norm": 2.3817248344421387, + "learning_rate": 4.407147881080503e-05, + "loss": 0.5516, + "num_input_tokens_seen": 3074072, + "step": 11500 + }, + { + "epoch": 3.0157273918741807, + "grad_norm": 2.688767194747925, + "learning_rate": 4.40640818723444e-05, + "loss": 0.4487, + "num_input_tokens_seen": 3076072, + "step": 11505 + }, + { + "epoch": 3.017038007863696, + "grad_norm": 2.0309014320373535, + "learning_rate": 4.4056680943804374e-05, + "loss": 0.5096, + "num_input_tokens_seen": 3078008, + "step": 11510 + }, + { + "epoch": 3.018348623853211, + "grad_norm": 1.8422725200653076, + "learning_rate": 4.404927602673396e-05, + "loss": 0.4671, + "num_input_tokens_seen": 3079256, + "step": 11515 + }, + { + "epoch": 3.019659239842726, + "grad_norm": 3.8840863704681396, + "learning_rate": 4.404186712268299e-05, + "loss": 0.5141, + "num_input_tokens_seen": 3080568, + "step": 11520 + }, + { + "epoch": 3.020969855832241, + "grad_norm": 4.2573628425598145, + "learning_rate": 4.4034454233202124e-05, + "loss": 0.6499, + "num_input_tokens_seen": 3081800, + "step": 11525 + }, + { + "epoch": 3.022280471821756, + "grad_norm": 2.9742815494537354, + "learning_rate": 4.4027037359842884e-05, + "loss": 0.3128, + "num_input_tokens_seen": 3083096, + "step": 11530 + }, + { + "epoch": 3.0235910878112713, + "grad_norm": 2.50455641746521, + "learning_rate": 4.401961650415759e-05, + "loss": 0.4168, + "num_input_tokens_seen": 3084408, + "step": 11535 + }, + { + "epoch": 3.0249017038007864, + "grad_norm": 1.6730417013168335, + "learning_rate": 4.401219166769943e-05, + "loss": 0.3069, + "num_input_tokens_seen": 3085864, + "step": 11540 + }, + { + "epoch": 3.0262123197903015, + "grad_norm": 10.543930053710938, + "learning_rate": 4.400476285202241e-05, + "loss": 0.7604, + "num_input_tokens_seen": 3086872, + "step": 11545 + }, + { + "epoch": 3.0275229357798166, + "grad_norm": 4.478282928466797, + "learning_rate": 4.399733005868136e-05, + "loss": 0.5199, + "num_input_tokens_seen": 3088104, + "step": 11550 + }, + { + "epoch": 3.0288335517693317, + "grad_norm": 3.1714465618133545, + "learning_rate": 4.3989893289231954e-05, + "loss": 0.569, + "num_input_tokens_seen": 3089432, + "step": 11555 + }, + { + "epoch": 3.0301441677588468, + "grad_norm": 1.8304469585418701, + "learning_rate": 4.398245254523069e-05, + "loss": 0.5212, + "num_input_tokens_seen": 3090872, + "step": 11560 + }, + { + "epoch": 3.031454783748362, + "grad_norm": 5.6258955001831055, + "learning_rate": 4.397500782823492e-05, + "loss": 0.3656, + "num_input_tokens_seen": 3092168, + "step": 11565 + }, + { + "epoch": 3.032765399737877, + "grad_norm": 2.169266939163208, + "learning_rate": 4.3967559139802785e-05, + "loss": 0.3381, + "num_input_tokens_seen": 3093128, + "step": 11570 + }, + { + "epoch": 3.034076015727392, + "grad_norm": 1.483998417854309, + "learning_rate": 4.39601064814933e-05, + "loss": 0.6608, + "num_input_tokens_seen": 3094408, + "step": 11575 + }, + { + "epoch": 3.035386631716907, + "grad_norm": 13.172284126281738, + "learning_rate": 4.395264985486629e-05, + "loss": 0.7046, + "num_input_tokens_seen": 3095832, + "step": 11580 + }, + { + "epoch": 3.036697247706422, + "grad_norm": 2.354586362838745, + "learning_rate": 4.394518926148241e-05, + "loss": 0.4798, + "num_input_tokens_seen": 3097160, + "step": 11585 + }, + { + "epoch": 3.038007863695937, + "grad_norm": 2.931272268295288, + "learning_rate": 4.393772470290315e-05, + "loss": 0.4988, + "num_input_tokens_seen": 3098344, + "step": 11590 + }, + { + "epoch": 3.039318479685452, + "grad_norm": 2.1314704418182373, + "learning_rate": 4.393025618069084e-05, + "loss": 0.5915, + "num_input_tokens_seen": 3099976, + "step": 11595 + }, + { + "epoch": 3.040629095674967, + "grad_norm": 2.9969382286071777, + "learning_rate": 4.3922783696408615e-05, + "loss": 0.5954, + "num_input_tokens_seen": 3101064, + "step": 11600 + }, + { + "epoch": 3.041939711664482, + "grad_norm": 2.267925262451172, + "learning_rate": 4.3915307251620456e-05, + "loss": 0.3781, + "num_input_tokens_seen": 3102248, + "step": 11605 + }, + { + "epoch": 3.0432503276539973, + "grad_norm": 1.5950441360473633, + "learning_rate": 4.390782684789118e-05, + "loss": 0.5705, + "num_input_tokens_seen": 3103816, + "step": 11610 + }, + { + "epoch": 3.0445609436435124, + "grad_norm": 2.802872657775879, + "learning_rate": 4.3900342486786405e-05, + "loss": 0.3162, + "num_input_tokens_seen": 3104968, + "step": 11615 + }, + { + "epoch": 3.0458715596330275, + "grad_norm": 2.9694435596466064, + "learning_rate": 4.389285416987261e-05, + "loss": 0.5935, + "num_input_tokens_seen": 3106056, + "step": 11620 + }, + { + "epoch": 3.0471821756225426, + "grad_norm": 2.4292242527008057, + "learning_rate": 4.388536189871708e-05, + "loss": 0.5748, + "num_input_tokens_seen": 3107288, + "step": 11625 + }, + { + "epoch": 3.0484927916120577, + "grad_norm": 1.9217276573181152, + "learning_rate": 4.387786567488794e-05, + "loss": 0.5019, + "num_input_tokens_seen": 3108680, + "step": 11630 + }, + { + "epoch": 3.0498034076015728, + "grad_norm": 2.4005157947540283, + "learning_rate": 4.387036549995413e-05, + "loss": 0.5768, + "num_input_tokens_seen": 3110056, + "step": 11635 + }, + { + "epoch": 3.051114023591088, + "grad_norm": 7.027963638305664, + "learning_rate": 4.386286137548543e-05, + "loss": 0.4321, + "num_input_tokens_seen": 3111400, + "step": 11640 + }, + { + "epoch": 3.052424639580603, + "grad_norm": 2.2534685134887695, + "learning_rate": 4.3855353303052436e-05, + "loss": 0.6, + "num_input_tokens_seen": 3112568, + "step": 11645 + }, + { + "epoch": 3.053735255570118, + "grad_norm": 3.8418421745300293, + "learning_rate": 4.384784128422657e-05, + "loss": 0.4725, + "num_input_tokens_seen": 3113784, + "step": 11650 + }, + { + "epoch": 3.055045871559633, + "grad_norm": 0.8599597811698914, + "learning_rate": 4.3840325320580086e-05, + "loss": 0.4974, + "num_input_tokens_seen": 3115000, + "step": 11655 + }, + { + "epoch": 3.0563564875491482, + "grad_norm": 1.6391912698745728, + "learning_rate": 4.383280541368608e-05, + "loss": 0.5167, + "num_input_tokens_seen": 3116056, + "step": 11660 + }, + { + "epoch": 3.0576671035386633, + "grad_norm": 2.915290355682373, + "learning_rate": 4.382528156511843e-05, + "loss": 0.3903, + "num_input_tokens_seen": 3117656, + "step": 11665 + }, + { + "epoch": 3.0589777195281784, + "grad_norm": 0.7350751757621765, + "learning_rate": 4.381775377645188e-05, + "loss": 0.3678, + "num_input_tokens_seen": 3118712, + "step": 11670 + }, + { + "epoch": 3.0602883355176935, + "grad_norm": 2.17161226272583, + "learning_rate": 4.381022204926198e-05, + "loss": 0.4161, + "num_input_tokens_seen": 3119688, + "step": 11675 + }, + { + "epoch": 3.061598951507208, + "grad_norm": 53.40374755859375, + "learning_rate": 4.38026863851251e-05, + "loss": 0.6697, + "num_input_tokens_seen": 3120776, + "step": 11680 + }, + { + "epoch": 3.0629095674967233, + "grad_norm": 2.1690268516540527, + "learning_rate": 4.379514678561846e-05, + "loss": 0.5196, + "num_input_tokens_seen": 3122088, + "step": 11685 + }, + { + "epoch": 3.0642201834862384, + "grad_norm": 2.628680467605591, + "learning_rate": 4.3787603252320065e-05, + "loss": 0.7311, + "num_input_tokens_seen": 3124136, + "step": 11690 + }, + { + "epoch": 3.0655307994757535, + "grad_norm": 1.0175745487213135, + "learning_rate": 4.378005578680877e-05, + "loss": 0.4369, + "num_input_tokens_seen": 3125512, + "step": 11695 + }, + { + "epoch": 3.0668414154652686, + "grad_norm": 1.609846591949463, + "learning_rate": 4.377250439066426e-05, + "loss": 0.5743, + "num_input_tokens_seen": 3127032, + "step": 11700 + }, + { + "epoch": 3.0681520314547837, + "grad_norm": 2.428391933441162, + "learning_rate": 4.3764949065467006e-05, + "loss": 0.3683, + "num_input_tokens_seen": 3129480, + "step": 11705 + }, + { + "epoch": 3.0694626474442988, + "grad_norm": 1.690409779548645, + "learning_rate": 4.375738981279834e-05, + "loss": 0.4517, + "num_input_tokens_seen": 3130984, + "step": 11710 + }, + { + "epoch": 3.070773263433814, + "grad_norm": 1.7125693559646606, + "learning_rate": 4.37498266342404e-05, + "loss": 0.5022, + "num_input_tokens_seen": 3132088, + "step": 11715 + }, + { + "epoch": 3.072083879423329, + "grad_norm": 0.8862412571907043, + "learning_rate": 4.3742259531376135e-05, + "loss": 0.3434, + "num_input_tokens_seen": 3133576, + "step": 11720 + }, + { + "epoch": 3.073394495412844, + "grad_norm": 5.6864495277404785, + "learning_rate": 4.373468850578933e-05, + "loss": 0.5255, + "num_input_tokens_seen": 3134792, + "step": 11725 + }, + { + "epoch": 3.074705111402359, + "grad_norm": 1.528953194618225, + "learning_rate": 4.37271135590646e-05, + "loss": 0.5819, + "num_input_tokens_seen": 3136392, + "step": 11730 + }, + { + "epoch": 3.0760157273918742, + "grad_norm": 8.850600242614746, + "learning_rate": 4.371953469278737e-05, + "loss": 0.4891, + "num_input_tokens_seen": 3137544, + "step": 11735 + }, + { + "epoch": 3.0773263433813893, + "grad_norm": 2.23791766166687, + "learning_rate": 4.371195190854386e-05, + "loss": 0.5621, + "num_input_tokens_seen": 3138712, + "step": 11740 + }, + { + "epoch": 3.0786369593709044, + "grad_norm": 4.1364946365356445, + "learning_rate": 4.370436520792115e-05, + "loss": 0.5947, + "num_input_tokens_seen": 3139608, + "step": 11745 + }, + { + "epoch": 3.0799475753604195, + "grad_norm": 4.065584659576416, + "learning_rate": 4.369677459250712e-05, + "loss": 0.4882, + "num_input_tokens_seen": 3140776, + "step": 11750 + }, + { + "epoch": 3.0812581913499346, + "grad_norm": 5.635682582855225, + "learning_rate": 4.3689180063890476e-05, + "loss": 0.5267, + "num_input_tokens_seen": 3141816, + "step": 11755 + }, + { + "epoch": 3.0825688073394497, + "grad_norm": 1.888471245765686, + "learning_rate": 4.368158162366073e-05, + "loss": 0.3549, + "num_input_tokens_seen": 3142920, + "step": 11760 + }, + { + "epoch": 3.083879423328965, + "grad_norm": 3.315988302230835, + "learning_rate": 4.367397927340824e-05, + "loss": 0.4422, + "num_input_tokens_seen": 3143944, + "step": 11765 + }, + { + "epoch": 3.08519003931848, + "grad_norm": 3.321099281311035, + "learning_rate": 4.366637301472415e-05, + "loss": 0.6285, + "num_input_tokens_seen": 3145384, + "step": 11770 + }, + { + "epoch": 3.0865006553079946, + "grad_norm": 0.7445850968360901, + "learning_rate": 4.365876284920044e-05, + "loss": 0.3426, + "num_input_tokens_seen": 3148808, + "step": 11775 + }, + { + "epoch": 3.0878112712975097, + "grad_norm": 2.618957281112671, + "learning_rate": 4.365114877842991e-05, + "loss": 0.5161, + "num_input_tokens_seen": 3150152, + "step": 11780 + }, + { + "epoch": 3.0891218872870247, + "grad_norm": 2.8424346446990967, + "learning_rate": 4.364353080400616e-05, + "loss": 0.4812, + "num_input_tokens_seen": 3151272, + "step": 11785 + }, + { + "epoch": 3.09043250327654, + "grad_norm": 3.278366804122925, + "learning_rate": 4.363590892752363e-05, + "loss": 0.4648, + "num_input_tokens_seen": 3152392, + "step": 11790 + }, + { + "epoch": 3.091743119266055, + "grad_norm": 7.270721912384033, + "learning_rate": 4.362828315057756e-05, + "loss": 0.3772, + "num_input_tokens_seen": 3153976, + "step": 11795 + }, + { + "epoch": 3.09305373525557, + "grad_norm": 2.267803907394409, + "learning_rate": 4.362065347476401e-05, + "loss": 0.3827, + "num_input_tokens_seen": 3155160, + "step": 11800 + }, + { + "epoch": 3.094364351245085, + "grad_norm": 8.734265327453613, + "learning_rate": 4.361301990167988e-05, + "loss": 0.3895, + "num_input_tokens_seen": 3156568, + "step": 11805 + }, + { + "epoch": 3.0956749672346002, + "grad_norm": 2.844081401824951, + "learning_rate": 4.360538243292282e-05, + "loss": 0.2966, + "num_input_tokens_seen": 3157672, + "step": 11810 + }, + { + "epoch": 3.0969855832241153, + "grad_norm": 2.807403087615967, + "learning_rate": 4.359774107009138e-05, + "loss": 0.4214, + "num_input_tokens_seen": 3159352, + "step": 11815 + }, + { + "epoch": 3.0982961992136304, + "grad_norm": 1.458642601966858, + "learning_rate": 4.3590095814784866e-05, + "loss": 0.4454, + "num_input_tokens_seen": 3160808, + "step": 11820 + }, + { + "epoch": 3.0996068152031455, + "grad_norm": 2.43841814994812, + "learning_rate": 4.3582446668603414e-05, + "loss": 0.4831, + "num_input_tokens_seen": 3162152, + "step": 11825 + }, + { + "epoch": 3.1009174311926606, + "grad_norm": 2.153698682785034, + "learning_rate": 4.357479363314799e-05, + "loss": 0.4809, + "num_input_tokens_seen": 3163752, + "step": 11830 + }, + { + "epoch": 3.1022280471821757, + "grad_norm": 22.289690017700195, + "learning_rate": 4.356713671002035e-05, + "loss": 0.5805, + "num_input_tokens_seen": 3165064, + "step": 11835 + }, + { + "epoch": 3.103538663171691, + "grad_norm": 5.357280731201172, + "learning_rate": 4.355947590082308e-05, + "loss": 0.4782, + "num_input_tokens_seen": 3166008, + "step": 11840 + }, + { + "epoch": 3.104849279161206, + "grad_norm": 2.1513171195983887, + "learning_rate": 4.355181120715957e-05, + "loss": 0.3528, + "num_input_tokens_seen": 3167592, + "step": 11845 + }, + { + "epoch": 3.106159895150721, + "grad_norm": 8.041136741638184, + "learning_rate": 4.3544142630634024e-05, + "loss": 0.5859, + "num_input_tokens_seen": 3168520, + "step": 11850 + }, + { + "epoch": 3.107470511140236, + "grad_norm": 2.2428672313690186, + "learning_rate": 4.3536470172851476e-05, + "loss": 0.4156, + "num_input_tokens_seen": 3169704, + "step": 11855 + }, + { + "epoch": 3.108781127129751, + "grad_norm": 4.347925662994385, + "learning_rate": 4.352879383541775e-05, + "loss": 0.4535, + "num_input_tokens_seen": 3170776, + "step": 11860 + }, + { + "epoch": 3.1100917431192663, + "grad_norm": 2.539013147354126, + "learning_rate": 4.352111361993948e-05, + "loss": 0.4291, + "num_input_tokens_seen": 3172056, + "step": 11865 + }, + { + "epoch": 3.111402359108781, + "grad_norm": 2.4569437503814697, + "learning_rate": 4.351342952802414e-05, + "loss": 0.5572, + "num_input_tokens_seen": 3173304, + "step": 11870 + }, + { + "epoch": 3.112712975098296, + "grad_norm": 4.736288547515869, + "learning_rate": 4.3505741561279984e-05, + "loss": 0.4765, + "num_input_tokens_seen": 3174536, + "step": 11875 + }, + { + "epoch": 3.114023591087811, + "grad_norm": 2.015221118927002, + "learning_rate": 4.349804972131609e-05, + "loss": 0.45, + "num_input_tokens_seen": 3175864, + "step": 11880 + }, + { + "epoch": 3.1153342070773262, + "grad_norm": 3.3426105976104736, + "learning_rate": 4.349035400974236e-05, + "loss": 0.4712, + "num_input_tokens_seen": 3177192, + "step": 11885 + }, + { + "epoch": 3.1166448230668413, + "grad_norm": 1.9879053831100464, + "learning_rate": 4.348265442816949e-05, + "loss": 0.4347, + "num_input_tokens_seen": 3178328, + "step": 11890 + }, + { + "epoch": 3.1179554390563564, + "grad_norm": 1.6967991590499878, + "learning_rate": 4.347495097820897e-05, + "loss": 0.3266, + "num_input_tokens_seen": 3179640, + "step": 11895 + }, + { + "epoch": 3.1192660550458715, + "grad_norm": 5.109768390655518, + "learning_rate": 4.346724366147314e-05, + "loss": 0.384, + "num_input_tokens_seen": 3180968, + "step": 11900 + }, + { + "epoch": 3.1205766710353866, + "grad_norm": 1.9356194734573364, + "learning_rate": 4.345953247957512e-05, + "loss": 0.3792, + "num_input_tokens_seen": 3182488, + "step": 11905 + }, + { + "epoch": 3.1218872870249017, + "grad_norm": 1.7088496685028076, + "learning_rate": 4.345181743412885e-05, + "loss": 0.3489, + "num_input_tokens_seen": 3183752, + "step": 11910 + }, + { + "epoch": 3.123197903014417, + "grad_norm": 2.491114377975464, + "learning_rate": 4.344409852674907e-05, + "loss": 0.3947, + "num_input_tokens_seen": 3185000, + "step": 11915 + }, + { + "epoch": 3.124508519003932, + "grad_norm": 4.803016185760498, + "learning_rate": 4.343637575905133e-05, + "loss": 0.4649, + "num_input_tokens_seen": 3186392, + "step": 11920 + }, + { + "epoch": 3.125819134993447, + "grad_norm": 2.6332271099090576, + "learning_rate": 4.342864913265201e-05, + "loss": 0.3749, + "num_input_tokens_seen": 3187848, + "step": 11925 + }, + { + "epoch": 3.127129750982962, + "grad_norm": 16.78589630126953, + "learning_rate": 4.3420918649168265e-05, + "loss": 0.3434, + "num_input_tokens_seen": 3188632, + "step": 11930 + }, + { + "epoch": 3.128440366972477, + "grad_norm": 1.8899744749069214, + "learning_rate": 4.341318431021808e-05, + "loss": 0.4556, + "num_input_tokens_seen": 3190376, + "step": 11935 + }, + { + "epoch": 3.1297509829619923, + "grad_norm": 2.687309980392456, + "learning_rate": 4.340544611742023e-05, + "loss": 0.5222, + "num_input_tokens_seen": 3191512, + "step": 11940 + }, + { + "epoch": 3.1310615989515074, + "grad_norm": 2.40851092338562, + "learning_rate": 4.339770407239431e-05, + "loss": 0.5146, + "num_input_tokens_seen": 3193304, + "step": 11945 + }, + { + "epoch": 3.1323722149410225, + "grad_norm": 1.7774899005889893, + "learning_rate": 4.3389958176760714e-05, + "loss": 0.4266, + "num_input_tokens_seen": 3194568, + "step": 11950 + }, + { + "epoch": 3.133682830930537, + "grad_norm": 2.7855770587921143, + "learning_rate": 4.3382208432140646e-05, + "loss": 0.7399, + "num_input_tokens_seen": 3195480, + "step": 11955 + }, + { + "epoch": 3.134993446920052, + "grad_norm": 2.022711753845215, + "learning_rate": 4.337445484015613e-05, + "loss": 0.5101, + "num_input_tokens_seen": 3197288, + "step": 11960 + }, + { + "epoch": 3.1363040629095673, + "grad_norm": 2.615882396697998, + "learning_rate": 4.3366697402429945e-05, + "loss": 0.421, + "num_input_tokens_seen": 3198680, + "step": 11965 + }, + { + "epoch": 3.1376146788990824, + "grad_norm": 1.505500078201294, + "learning_rate": 4.335893612058574e-05, + "loss": 0.4095, + "num_input_tokens_seen": 3199944, + "step": 11970 + }, + { + "epoch": 3.1389252948885975, + "grad_norm": 3.8080973625183105, + "learning_rate": 4.335117099624792e-05, + "loss": 0.4184, + "num_input_tokens_seen": 3201144, + "step": 11975 + }, + { + "epoch": 3.1402359108781126, + "grad_norm": 1.9526431560516357, + "learning_rate": 4.3343402031041726e-05, + "loss": 0.5001, + "num_input_tokens_seen": 3202552, + "step": 11980 + }, + { + "epoch": 3.1415465268676277, + "grad_norm": 1.6245023012161255, + "learning_rate": 4.333562922659318e-05, + "loss": 0.3684, + "num_input_tokens_seen": 3203576, + "step": 11985 + }, + { + "epoch": 3.142857142857143, + "grad_norm": 1.9317125082015991, + "learning_rate": 4.332785258452911e-05, + "loss": 0.4093, + "num_input_tokens_seen": 3205128, + "step": 11990 + }, + { + "epoch": 3.144167758846658, + "grad_norm": 2.0163192749023438, + "learning_rate": 4.332007210647717e-05, + "loss": 0.5495, + "num_input_tokens_seen": 3206744, + "step": 11995 + }, + { + "epoch": 3.145478374836173, + "grad_norm": 4.868887424468994, + "learning_rate": 4.3312287794065795e-05, + "loss": 0.9219, + "num_input_tokens_seen": 3207944, + "step": 12000 + }, + { + "epoch": 3.146788990825688, + "grad_norm": 2.119908571243286, + "learning_rate": 4.330449964892422e-05, + "loss": 0.453, + "num_input_tokens_seen": 3209272, + "step": 12005 + }, + { + "epoch": 3.148099606815203, + "grad_norm": 2.270817279815674, + "learning_rate": 4.32967076726825e-05, + "loss": 0.4979, + "num_input_tokens_seen": 3210808, + "step": 12010 + }, + { + "epoch": 3.1494102228047183, + "grad_norm": 0.981022834777832, + "learning_rate": 4.328891186697148e-05, + "loss": 0.3893, + "num_input_tokens_seen": 3212776, + "step": 12015 + }, + { + "epoch": 3.1507208387942334, + "grad_norm": 4.122296333312988, + "learning_rate": 4.328111223342281e-05, + "loss": 0.6501, + "num_input_tokens_seen": 3213784, + "step": 12020 + }, + { + "epoch": 3.1520314547837485, + "grad_norm": 3.390723466873169, + "learning_rate": 4.327330877366893e-05, + "loss": 0.6407, + "num_input_tokens_seen": 3214696, + "step": 12025 + }, + { + "epoch": 3.1533420707732636, + "grad_norm": 2.112835168838501, + "learning_rate": 4.3265501489343094e-05, + "loss": 0.3962, + "num_input_tokens_seen": 3216696, + "step": 12030 + }, + { + "epoch": 3.1546526867627787, + "grad_norm": 2.6116244792938232, + "learning_rate": 4.325769038207936e-05, + "loss": 0.3573, + "num_input_tokens_seen": 3218024, + "step": 12035 + }, + { + "epoch": 3.1559633027522938, + "grad_norm": 1.7705047130584717, + "learning_rate": 4.3249875453512586e-05, + "loss": 0.4652, + "num_input_tokens_seen": 3219288, + "step": 12040 + }, + { + "epoch": 3.157273918741809, + "grad_norm": 1.508462905883789, + "learning_rate": 4.32420567052784e-05, + "loss": 0.4091, + "num_input_tokens_seen": 3220184, + "step": 12045 + }, + { + "epoch": 3.1585845347313235, + "grad_norm": 4.6796159744262695, + "learning_rate": 4.323423413901327e-05, + "loss": 0.5613, + "num_input_tokens_seen": 3221320, + "step": 12050 + }, + { + "epoch": 3.1598951507208386, + "grad_norm": 2.7440078258514404, + "learning_rate": 4.322640775635445e-05, + "loss": 0.3654, + "num_input_tokens_seen": 3222952, + "step": 12055 + }, + { + "epoch": 3.1612057667103537, + "grad_norm": 1.8164374828338623, + "learning_rate": 4.321857755893996e-05, + "loss": 0.4765, + "num_input_tokens_seen": 3224680, + "step": 12060 + }, + { + "epoch": 3.162516382699869, + "grad_norm": 2.2999460697174072, + "learning_rate": 4.3210743548408684e-05, + "loss": 0.3935, + "num_input_tokens_seen": 3225848, + "step": 12065 + }, + { + "epoch": 3.163826998689384, + "grad_norm": 2.006662368774414, + "learning_rate": 4.3202905726400236e-05, + "loss": 0.4336, + "num_input_tokens_seen": 3226984, + "step": 12070 + }, + { + "epoch": 3.165137614678899, + "grad_norm": 1.5688161849975586, + "learning_rate": 4.3195064094555085e-05, + "loss": 0.459, + "num_input_tokens_seen": 3228312, + "step": 12075 + }, + { + "epoch": 3.166448230668414, + "grad_norm": 2.9253013134002686, + "learning_rate": 4.318721865451445e-05, + "loss": 0.5464, + "num_input_tokens_seen": 3229656, + "step": 12080 + }, + { + "epoch": 3.167758846657929, + "grad_norm": 3.190143346786499, + "learning_rate": 4.317936940792037e-05, + "loss": 0.5934, + "num_input_tokens_seen": 3230904, + "step": 12085 + }, + { + "epoch": 3.1690694626474443, + "grad_norm": 2.2496626377105713, + "learning_rate": 4.317151635641569e-05, + "loss": 0.3974, + "num_input_tokens_seen": 3231992, + "step": 12090 + }, + { + "epoch": 3.1703800786369594, + "grad_norm": 2.000920295715332, + "learning_rate": 4.316365950164404e-05, + "loss": 0.4826, + "num_input_tokens_seen": 3232808, + "step": 12095 + }, + { + "epoch": 3.1716906946264745, + "grad_norm": 1.8366782665252686, + "learning_rate": 4.3155798845249827e-05, + "loss": 0.536, + "num_input_tokens_seen": 3234264, + "step": 12100 + }, + { + "epoch": 3.1730013106159896, + "grad_norm": 5.4056925773620605, + "learning_rate": 4.3147934388878295e-05, + "loss": 0.5368, + "num_input_tokens_seen": 3235208, + "step": 12105 + }, + { + "epoch": 3.1743119266055047, + "grad_norm": 2.9151394367218018, + "learning_rate": 4.314006613417546e-05, + "loss": 0.4168, + "num_input_tokens_seen": 3236776, + "step": 12110 + }, + { + "epoch": 3.1756225425950197, + "grad_norm": 1.4958763122558594, + "learning_rate": 4.3132194082788116e-05, + "loss": 0.5072, + "num_input_tokens_seen": 3238120, + "step": 12115 + }, + { + "epoch": 3.176933158584535, + "grad_norm": 1.104042410850525, + "learning_rate": 4.312431823636388e-05, + "loss": 0.461, + "num_input_tokens_seen": 3239592, + "step": 12120 + }, + { + "epoch": 3.17824377457405, + "grad_norm": 1.3131264448165894, + "learning_rate": 4.311643859655116e-05, + "loss": 0.7413, + "num_input_tokens_seen": 3241480, + "step": 12125 + }, + { + "epoch": 3.179554390563565, + "grad_norm": 2.860849142074585, + "learning_rate": 4.3108555164999144e-05, + "loss": 0.3674, + "num_input_tokens_seen": 3242648, + "step": 12130 + }, + { + "epoch": 3.18086500655308, + "grad_norm": 1.4394760131835938, + "learning_rate": 4.310066794335782e-05, + "loss": 0.4798, + "num_input_tokens_seen": 3243944, + "step": 12135 + }, + { + "epoch": 3.1821756225425952, + "grad_norm": 1.2316566705703735, + "learning_rate": 4.309277693327797e-05, + "loss": 0.5163, + "num_input_tokens_seen": 3245240, + "step": 12140 + }, + { + "epoch": 3.18348623853211, + "grad_norm": 4.9461140632629395, + "learning_rate": 4.308488213641116e-05, + "loss": 0.4763, + "num_input_tokens_seen": 3246248, + "step": 12145 + }, + { + "epoch": 3.184796854521625, + "grad_norm": 4.062112808227539, + "learning_rate": 4.3076983554409784e-05, + "loss": 0.7988, + "num_input_tokens_seen": 3247176, + "step": 12150 + }, + { + "epoch": 3.18610747051114, + "grad_norm": 2.147085428237915, + "learning_rate": 4.306908118892698e-05, + "loss": 0.5185, + "num_input_tokens_seen": 3248232, + "step": 12155 + }, + { + "epoch": 3.187418086500655, + "grad_norm": 1.7100777626037598, + "learning_rate": 4.30611750416167e-05, + "loss": 0.3553, + "num_input_tokens_seen": 3249464, + "step": 12160 + }, + { + "epoch": 3.1887287024901703, + "grad_norm": 2.089423179626465, + "learning_rate": 4.305326511413369e-05, + "loss": 0.5189, + "num_input_tokens_seen": 3250792, + "step": 12165 + }, + { + "epoch": 3.1900393184796854, + "grad_norm": 2.159780263900757, + "learning_rate": 4.304535140813349e-05, + "loss": 0.4789, + "num_input_tokens_seen": 3251944, + "step": 12170 + }, + { + "epoch": 3.1913499344692005, + "grad_norm": 5.428356170654297, + "learning_rate": 4.303743392527242e-05, + "loss": 0.5828, + "num_input_tokens_seen": 3253304, + "step": 12175 + }, + { + "epoch": 3.1926605504587156, + "grad_norm": 1.1363245248794556, + "learning_rate": 4.302951266720759e-05, + "loss": 0.9591, + "num_input_tokens_seen": 3254472, + "step": 12180 + }, + { + "epoch": 3.1939711664482306, + "grad_norm": 2.3537967205047607, + "learning_rate": 4.3021587635596914e-05, + "loss": 0.3648, + "num_input_tokens_seen": 3255960, + "step": 12185 + }, + { + "epoch": 3.1952817824377457, + "grad_norm": 1.8489009141921997, + "learning_rate": 4.301365883209909e-05, + "loss": 0.5833, + "num_input_tokens_seen": 3257416, + "step": 12190 + }, + { + "epoch": 3.196592398427261, + "grad_norm": 1.2163419723510742, + "learning_rate": 4.300572625837359e-05, + "loss": 0.5983, + "num_input_tokens_seen": 3258856, + "step": 12195 + }, + { + "epoch": 3.197903014416776, + "grad_norm": 4.7187275886535645, + "learning_rate": 4.29977899160807e-05, + "loss": 0.5428, + "num_input_tokens_seen": 3259848, + "step": 12200 + }, + { + "epoch": 3.199213630406291, + "grad_norm": 1.8904266357421875, + "learning_rate": 4.298984980688148e-05, + "loss": 0.399, + "num_input_tokens_seen": 3261032, + "step": 12205 + }, + { + "epoch": 3.200524246395806, + "grad_norm": 3.983302354812622, + "learning_rate": 4.298190593243778e-05, + "loss": 0.5177, + "num_input_tokens_seen": 3262392, + "step": 12210 + }, + { + "epoch": 3.2018348623853212, + "grad_norm": 2.9027292728424072, + "learning_rate": 4.297395829441223e-05, + "loss": 0.3641, + "num_input_tokens_seen": 3263832, + "step": 12215 + }, + { + "epoch": 3.2031454783748363, + "grad_norm": 1.9677133560180664, + "learning_rate": 4.2966006894468274e-05, + "loss": 0.4356, + "num_input_tokens_seen": 3265240, + "step": 12220 + }, + { + "epoch": 3.2044560943643514, + "grad_norm": 2.5643513202667236, + "learning_rate": 4.295805173427012e-05, + "loss": 0.4723, + "num_input_tokens_seen": 3267096, + "step": 12225 + }, + { + "epoch": 3.2057667103538665, + "grad_norm": 2.7666642665863037, + "learning_rate": 4.295009281548276e-05, + "loss": 0.3036, + "num_input_tokens_seen": 3268552, + "step": 12230 + }, + { + "epoch": 3.2070773263433816, + "grad_norm": 16.0517520904541, + "learning_rate": 4.294213013977199e-05, + "loss": 0.5925, + "num_input_tokens_seen": 3269928, + "step": 12235 + }, + { + "epoch": 3.2083879423328963, + "grad_norm": 2.0799410343170166, + "learning_rate": 4.2934163708804385e-05, + "loss": 0.5794, + "num_input_tokens_seen": 3271048, + "step": 12240 + }, + { + "epoch": 3.2096985583224114, + "grad_norm": 3.037667989730835, + "learning_rate": 4.29261935242473e-05, + "loss": 0.5427, + "num_input_tokens_seen": 3272216, + "step": 12245 + }, + { + "epoch": 3.2110091743119265, + "grad_norm": 1.7171928882598877, + "learning_rate": 4.291821958776889e-05, + "loss": 0.5438, + "num_input_tokens_seen": 3273528, + "step": 12250 + }, + { + "epoch": 3.2123197903014415, + "grad_norm": 2.917189121246338, + "learning_rate": 4.291024190103808e-05, + "loss": 0.4858, + "num_input_tokens_seen": 3274776, + "step": 12255 + }, + { + "epoch": 3.2136304062909566, + "grad_norm": 2.9725747108459473, + "learning_rate": 4.290226046572459e-05, + "loss": 0.4815, + "num_input_tokens_seen": 3276216, + "step": 12260 + }, + { + "epoch": 3.2149410222804717, + "grad_norm": 2.556379556655884, + "learning_rate": 4.289427528349891e-05, + "loss": 0.4096, + "num_input_tokens_seen": 3277240, + "step": 12265 + }, + { + "epoch": 3.216251638269987, + "grad_norm": 3.4594476222991943, + "learning_rate": 4.2886286356032334e-05, + "loss": 0.3485, + "num_input_tokens_seen": 3278936, + "step": 12270 + }, + { + "epoch": 3.217562254259502, + "grad_norm": 1.8288301229476929, + "learning_rate": 4.287829368499693e-05, + "loss": 0.784, + "num_input_tokens_seen": 3279944, + "step": 12275 + }, + { + "epoch": 3.218872870249017, + "grad_norm": 1.7721713781356812, + "learning_rate": 4.2870297272065546e-05, + "loss": 0.4117, + "num_input_tokens_seen": 3281096, + "step": 12280 + }, + { + "epoch": 3.220183486238532, + "grad_norm": 1.9754234552383423, + "learning_rate": 4.286229711891182e-05, + "loss": 0.6941, + "num_input_tokens_seen": 3282344, + "step": 12285 + }, + { + "epoch": 3.221494102228047, + "grad_norm": 2.108135223388672, + "learning_rate": 4.285429322721017e-05, + "loss": 0.4605, + "num_input_tokens_seen": 3283640, + "step": 12290 + }, + { + "epoch": 3.2228047182175623, + "grad_norm": 5.500300407409668, + "learning_rate": 4.284628559863581e-05, + "loss": 0.4519, + "num_input_tokens_seen": 3284680, + "step": 12295 + }, + { + "epoch": 3.2241153342070774, + "grad_norm": 2.361050605773926, + "learning_rate": 4.2838274234864694e-05, + "loss": 0.3242, + "num_input_tokens_seen": 3286168, + "step": 12300 + }, + { + "epoch": 3.2254259501965925, + "grad_norm": 2.505918264389038, + "learning_rate": 4.283025913757361e-05, + "loss": 0.6391, + "num_input_tokens_seen": 3287912, + "step": 12305 + }, + { + "epoch": 3.2267365661861076, + "grad_norm": 1.7115931510925293, + "learning_rate": 4.2822240308440096e-05, + "loss": 0.4562, + "num_input_tokens_seen": 3289256, + "step": 12310 + }, + { + "epoch": 3.2280471821756227, + "grad_norm": 1.7264857292175293, + "learning_rate": 4.281421774914248e-05, + "loss": 0.4013, + "num_input_tokens_seen": 3290440, + "step": 12315 + }, + { + "epoch": 3.229357798165138, + "grad_norm": 1.38658607006073, + "learning_rate": 4.280619146135988e-05, + "loss": 0.5048, + "num_input_tokens_seen": 3291480, + "step": 12320 + }, + { + "epoch": 3.230668414154653, + "grad_norm": 1.9455790519714355, + "learning_rate": 4.2798161446772156e-05, + "loss": 0.3318, + "num_input_tokens_seen": 3292744, + "step": 12325 + }, + { + "epoch": 3.231979030144168, + "grad_norm": 2.7108874320983887, + "learning_rate": 4.279012770706e-05, + "loss": 0.4833, + "num_input_tokens_seen": 3294264, + "step": 12330 + }, + { + "epoch": 3.2332896461336826, + "grad_norm": 4.427041053771973, + "learning_rate": 4.278209024390486e-05, + "loss": 0.4658, + "num_input_tokens_seen": 3295928, + "step": 12335 + }, + { + "epoch": 3.2346002621231977, + "grad_norm": 1.5041087865829468, + "learning_rate": 4.2774049058988955e-05, + "loss": 0.4233, + "num_input_tokens_seen": 3297384, + "step": 12340 + }, + { + "epoch": 3.235910878112713, + "grad_norm": 2.4027860164642334, + "learning_rate": 4.276600415399529e-05, + "loss": 0.4798, + "num_input_tokens_seen": 3298568, + "step": 12345 + }, + { + "epoch": 3.237221494102228, + "grad_norm": 5.635895252227783, + "learning_rate": 4.275795553060765e-05, + "loss": 0.6808, + "num_input_tokens_seen": 3299976, + "step": 12350 + }, + { + "epoch": 3.238532110091743, + "grad_norm": 4.433472156524658, + "learning_rate": 4.274990319051061e-05, + "loss": 0.6817, + "num_input_tokens_seen": 3300904, + "step": 12355 + }, + { + "epoch": 3.239842726081258, + "grad_norm": 5.908500671386719, + "learning_rate": 4.2741847135389495e-05, + "loss": 0.3713, + "num_input_tokens_seen": 3302296, + "step": 12360 + }, + { + "epoch": 3.241153342070773, + "grad_norm": 1.4044122695922852, + "learning_rate": 4.2733787366930436e-05, + "loss": 0.2342, + "num_input_tokens_seen": 3304168, + "step": 12365 + }, + { + "epoch": 3.2424639580602883, + "grad_norm": 2.990426540374756, + "learning_rate": 4.2725723886820314e-05, + "loss": 0.4675, + "num_input_tokens_seen": 3305384, + "step": 12370 + }, + { + "epoch": 3.2437745740498034, + "grad_norm": 2.654287338256836, + "learning_rate": 4.271765669674682e-05, + "loss": 0.5286, + "num_input_tokens_seen": 3306808, + "step": 12375 + }, + { + "epoch": 3.2450851900393185, + "grad_norm": 2.031214714050293, + "learning_rate": 4.270958579839839e-05, + "loss": 0.5642, + "num_input_tokens_seen": 3308104, + "step": 12380 + }, + { + "epoch": 3.2463958060288336, + "grad_norm": 1.976611852645874, + "learning_rate": 4.270151119346424e-05, + "loss": 0.4599, + "num_input_tokens_seen": 3309416, + "step": 12385 + }, + { + "epoch": 3.2477064220183487, + "grad_norm": 3.6133434772491455, + "learning_rate": 4.26934328836344e-05, + "loss": 0.4547, + "num_input_tokens_seen": 3310936, + "step": 12390 + }, + { + "epoch": 3.249017038007864, + "grad_norm": 1.2464869022369385, + "learning_rate": 4.268535087059962e-05, + "loss": 0.3937, + "num_input_tokens_seen": 3312680, + "step": 12395 + }, + { + "epoch": 3.250327653997379, + "grad_norm": 0.9353690147399902, + "learning_rate": 4.267726515605146e-05, + "loss": 0.5222, + "num_input_tokens_seen": 3314680, + "step": 12400 + }, + { + "epoch": 3.251638269986894, + "grad_norm": 3.4066247940063477, + "learning_rate": 4.2669175741682255e-05, + "loss": 0.6454, + "num_input_tokens_seen": 3316072, + "step": 12405 + }, + { + "epoch": 3.252948885976409, + "grad_norm": 1.573951005935669, + "learning_rate": 4.266108262918509e-05, + "loss": 0.497, + "num_input_tokens_seen": 3317480, + "step": 12410 + }, + { + "epoch": 3.254259501965924, + "grad_norm": 1.7035962343215942, + "learning_rate": 4.265298582025385e-05, + "loss": 0.339, + "num_input_tokens_seen": 3318872, + "step": 12415 + }, + { + "epoch": 3.255570117955439, + "grad_norm": 2.150559663772583, + "learning_rate": 4.264488531658317e-05, + "loss": 0.6626, + "num_input_tokens_seen": 3319848, + "step": 12420 + }, + { + "epoch": 3.2568807339449544, + "grad_norm": 5.128783702850342, + "learning_rate": 4.26367811198685e-05, + "loss": 0.3921, + "num_input_tokens_seen": 3320952, + "step": 12425 + }, + { + "epoch": 3.258191349934469, + "grad_norm": 29.052217483520508, + "learning_rate": 4.2628673231806e-05, + "loss": 0.588, + "num_input_tokens_seen": 3322312, + "step": 12430 + }, + { + "epoch": 3.259501965923984, + "grad_norm": 3.8858935832977295, + "learning_rate": 4.262056165409266e-05, + "loss": 0.3517, + "num_input_tokens_seen": 3323320, + "step": 12435 + }, + { + "epoch": 3.260812581913499, + "grad_norm": 2.803032159805298, + "learning_rate": 4.261244638842621e-05, + "loss": 0.3969, + "num_input_tokens_seen": 3324472, + "step": 12440 + }, + { + "epoch": 3.2621231979030143, + "grad_norm": 2.8432087898254395, + "learning_rate": 4.260432743650516e-05, + "loss": 0.3246, + "num_input_tokens_seen": 3325736, + "step": 12445 + }, + { + "epoch": 3.2634338138925294, + "grad_norm": 4.870303630828857, + "learning_rate": 4.25962048000288e-05, + "loss": 0.4778, + "num_input_tokens_seen": 3326840, + "step": 12450 + }, + { + "epoch": 3.2647444298820445, + "grad_norm": 8.26906681060791, + "learning_rate": 4.258807848069718e-05, + "loss": 0.4603, + "num_input_tokens_seen": 3328952, + "step": 12455 + }, + { + "epoch": 3.2660550458715596, + "grad_norm": 2.820481061935425, + "learning_rate": 4.257994848021112e-05, + "loss": 0.3904, + "num_input_tokens_seen": 3330152, + "step": 12460 + }, + { + "epoch": 3.2673656618610747, + "grad_norm": 1.4537684917449951, + "learning_rate": 4.257181480027222e-05, + "loss": 0.7136, + "num_input_tokens_seen": 3331416, + "step": 12465 + }, + { + "epoch": 3.26867627785059, + "grad_norm": 6.814783573150635, + "learning_rate": 4.256367744258285e-05, + "loss": 0.5403, + "num_input_tokens_seen": 3332536, + "step": 12470 + }, + { + "epoch": 3.269986893840105, + "grad_norm": 6.108974456787109, + "learning_rate": 4.255553640884613e-05, + "loss": 0.6618, + "num_input_tokens_seen": 3333816, + "step": 12475 + }, + { + "epoch": 3.27129750982962, + "grad_norm": 2.595151662826538, + "learning_rate": 4.254739170076597e-05, + "loss": 0.474, + "num_input_tokens_seen": 3335272, + "step": 12480 + }, + { + "epoch": 3.272608125819135, + "grad_norm": 3.0631697177886963, + "learning_rate": 4.253924332004706e-05, + "loss": 0.5543, + "num_input_tokens_seen": 3336344, + "step": 12485 + }, + { + "epoch": 3.27391874180865, + "grad_norm": 16.249351501464844, + "learning_rate": 4.2531091268394816e-05, + "loss": 0.5508, + "num_input_tokens_seen": 3337400, + "step": 12490 + }, + { + "epoch": 3.2752293577981653, + "grad_norm": 5.246446132659912, + "learning_rate": 4.252293554751547e-05, + "loss": 0.5908, + "num_input_tokens_seen": 3338712, + "step": 12495 + }, + { + "epoch": 3.2765399737876804, + "grad_norm": 8.059348106384277, + "learning_rate": 4.251477615911599e-05, + "loss": 0.5235, + "num_input_tokens_seen": 3340040, + "step": 12500 + }, + { + "epoch": 3.2778505897771955, + "grad_norm": 2.3794219493865967, + "learning_rate": 4.250661310490412e-05, + "loss": 0.3124, + "num_input_tokens_seen": 3341048, + "step": 12505 + }, + { + "epoch": 3.2791612057667106, + "grad_norm": 2.4902727603912354, + "learning_rate": 4.249844638658838e-05, + "loss": 0.9295, + "num_input_tokens_seen": 3342520, + "step": 12510 + }, + { + "epoch": 3.280471821756225, + "grad_norm": 2.244110107421875, + "learning_rate": 4.249027600587804e-05, + "loss": 0.4186, + "num_input_tokens_seen": 3343512, + "step": 12515 + }, + { + "epoch": 3.2817824377457407, + "grad_norm": 1.440434455871582, + "learning_rate": 4.248210196448316e-05, + "loss": 0.3366, + "num_input_tokens_seen": 3345096, + "step": 12520 + }, + { + "epoch": 3.2830930537352554, + "grad_norm": 3.7415523529052734, + "learning_rate": 4.247392426411454e-05, + "loss": 0.5031, + "num_input_tokens_seen": 3346200, + "step": 12525 + }, + { + "epoch": 3.2844036697247705, + "grad_norm": 2.366713285446167, + "learning_rate": 4.246574290648376e-05, + "loss": 0.4758, + "num_input_tokens_seen": 3347320, + "step": 12530 + }, + { + "epoch": 3.2857142857142856, + "grad_norm": 1.5144613981246948, + "learning_rate": 4.245755789330318e-05, + "loss": 0.4296, + "num_input_tokens_seen": 3348888, + "step": 12535 + }, + { + "epoch": 3.2870249017038007, + "grad_norm": 1.2099273204803467, + "learning_rate": 4.2449369226285884e-05, + "loss": 0.5342, + "num_input_tokens_seen": 3350152, + "step": 12540 + }, + { + "epoch": 3.288335517693316, + "grad_norm": 5.620386600494385, + "learning_rate": 4.2441176907145775e-05, + "loss": 0.6591, + "num_input_tokens_seen": 3351160, + "step": 12545 + }, + { + "epoch": 3.289646133682831, + "grad_norm": 1.95139479637146, + "learning_rate": 4.243298093759746e-05, + "loss": 0.5418, + "num_input_tokens_seen": 3352872, + "step": 12550 + }, + { + "epoch": 3.290956749672346, + "grad_norm": 6.284833908081055, + "learning_rate": 4.242478131935637e-05, + "loss": 0.5189, + "num_input_tokens_seen": 3353944, + "step": 12555 + }, + { + "epoch": 3.292267365661861, + "grad_norm": 1.7228201627731323, + "learning_rate": 4.241657805413865e-05, + "loss": 0.6146, + "num_input_tokens_seen": 3355224, + "step": 12560 + }, + { + "epoch": 3.293577981651376, + "grad_norm": 1.3854132890701294, + "learning_rate": 4.240837114366123e-05, + "loss": 0.5157, + "num_input_tokens_seen": 3356632, + "step": 12565 + }, + { + "epoch": 3.2948885976408913, + "grad_norm": 6.354742050170898, + "learning_rate": 4.2400160589641825e-05, + "loss": 0.5064, + "num_input_tokens_seen": 3357928, + "step": 12570 + }, + { + "epoch": 3.2961992136304064, + "grad_norm": 7.807467460632324, + "learning_rate": 4.239194639379887e-05, + "loss": 0.4943, + "num_input_tokens_seen": 3359432, + "step": 12575 + }, + { + "epoch": 3.2975098296199215, + "grad_norm": 3.039174795150757, + "learning_rate": 4.238372855785158e-05, + "loss": 0.6699, + "num_input_tokens_seen": 3360648, + "step": 12580 + }, + { + "epoch": 3.2988204456094365, + "grad_norm": 1.8414734601974487, + "learning_rate": 4.2375507083519936e-05, + "loss": 0.4739, + "num_input_tokens_seen": 3361800, + "step": 12585 + }, + { + "epoch": 3.3001310615989516, + "grad_norm": 3.4811172485351562, + "learning_rate": 4.2367281972524685e-05, + "loss": 0.3059, + "num_input_tokens_seen": 3363320, + "step": 12590 + }, + { + "epoch": 3.3014416775884667, + "grad_norm": 2.9992239475250244, + "learning_rate": 4.235905322658733e-05, + "loss": 0.4388, + "num_input_tokens_seen": 3364376, + "step": 12595 + }, + { + "epoch": 3.302752293577982, + "grad_norm": 2.868952751159668, + "learning_rate": 4.235082084743013e-05, + "loss": 0.3264, + "num_input_tokens_seen": 3365432, + "step": 12600 + }, + { + "epoch": 3.304062909567497, + "grad_norm": 2.325162410736084, + "learning_rate": 4.23425848367761e-05, + "loss": 0.4002, + "num_input_tokens_seen": 3366408, + "step": 12605 + }, + { + "epoch": 3.3053735255570116, + "grad_norm": 4.64836311340332, + "learning_rate": 4.2334345196349036e-05, + "loss": 0.4864, + "num_input_tokens_seen": 3367496, + "step": 12610 + }, + { + "epoch": 3.3066841415465267, + "grad_norm": 2.8417904376983643, + "learning_rate": 4.2326101927873464e-05, + "loss": 0.3372, + "num_input_tokens_seen": 3368504, + "step": 12615 + }, + { + "epoch": 3.3079947575360418, + "grad_norm": 4.7999067306518555, + "learning_rate": 4.2317855033074706e-05, + "loss": 0.3994, + "num_input_tokens_seen": 3369736, + "step": 12620 + }, + { + "epoch": 3.309305373525557, + "grad_norm": 2.806180238723755, + "learning_rate": 4.2309604513678804e-05, + "loss": 0.3429, + "num_input_tokens_seen": 3371016, + "step": 12625 + }, + { + "epoch": 3.310615989515072, + "grad_norm": 2.9966325759887695, + "learning_rate": 4.230135037141259e-05, + "loss": 0.4156, + "num_input_tokens_seen": 3372392, + "step": 12630 + }, + { + "epoch": 3.311926605504587, + "grad_norm": 3.0565922260284424, + "learning_rate": 4.229309260800364e-05, + "loss": 0.4005, + "num_input_tokens_seen": 3373672, + "step": 12635 + }, + { + "epoch": 3.313237221494102, + "grad_norm": 5.826170444488525, + "learning_rate": 4.228483122518029e-05, + "loss": 0.3552, + "num_input_tokens_seen": 3375080, + "step": 12640 + }, + { + "epoch": 3.3145478374836173, + "grad_norm": 3.3033506870269775, + "learning_rate": 4.227656622467162e-05, + "loss": 0.7858, + "num_input_tokens_seen": 3376488, + "step": 12645 + }, + { + "epoch": 3.3158584534731324, + "grad_norm": 3.1191799640655518, + "learning_rate": 4.2268297608207496e-05, + "loss": 0.4235, + "num_input_tokens_seen": 3377704, + "step": 12650 + }, + { + "epoch": 3.3171690694626474, + "grad_norm": 2.745704412460327, + "learning_rate": 4.226002537751852e-05, + "loss": 0.6129, + "num_input_tokens_seen": 3378920, + "step": 12655 + }, + { + "epoch": 3.3184796854521625, + "grad_norm": 3.3554434776306152, + "learning_rate": 4.225174953433605e-05, + "loss": 0.4691, + "num_input_tokens_seen": 3380152, + "step": 12660 + }, + { + "epoch": 3.3197903014416776, + "grad_norm": 2.1005871295928955, + "learning_rate": 4.224347008039222e-05, + "loss": 0.535, + "num_input_tokens_seen": 3381992, + "step": 12665 + }, + { + "epoch": 3.3211009174311927, + "grad_norm": 2.647270441055298, + "learning_rate": 4.22351870174199e-05, + "loss": 0.4485, + "num_input_tokens_seen": 3383336, + "step": 12670 + }, + { + "epoch": 3.322411533420708, + "grad_norm": 1.799717903137207, + "learning_rate": 4.222690034715271e-05, + "loss": 0.6078, + "num_input_tokens_seen": 3384680, + "step": 12675 + }, + { + "epoch": 3.323722149410223, + "grad_norm": 2.2633821964263916, + "learning_rate": 4.2218610071325036e-05, + "loss": 0.4494, + "num_input_tokens_seen": 3385864, + "step": 12680 + }, + { + "epoch": 3.325032765399738, + "grad_norm": 1.9332189559936523, + "learning_rate": 4.221031619167204e-05, + "loss": 0.5961, + "num_input_tokens_seen": 3387512, + "step": 12685 + }, + { + "epoch": 3.326343381389253, + "grad_norm": 1.1154454946517944, + "learning_rate": 4.2202018709929595e-05, + "loss": 0.4902, + "num_input_tokens_seen": 3390584, + "step": 12690 + }, + { + "epoch": 3.327653997378768, + "grad_norm": 2.511505126953125, + "learning_rate": 4.219371762783436e-05, + "loss": 0.4105, + "num_input_tokens_seen": 3391416, + "step": 12695 + }, + { + "epoch": 3.3289646133682833, + "grad_norm": 2.278118133544922, + "learning_rate": 4.218541294712373e-05, + "loss": 0.6367, + "num_input_tokens_seen": 3392664, + "step": 12700 + }, + { + "epoch": 3.330275229357798, + "grad_norm": 1.6414189338684082, + "learning_rate": 4.217710466953586e-05, + "loss": 0.8636, + "num_input_tokens_seen": 3394024, + "step": 12705 + }, + { + "epoch": 3.331585845347313, + "grad_norm": 2.9214630126953125, + "learning_rate": 4.2168792796809674e-05, + "loss": 0.3262, + "num_input_tokens_seen": 3395896, + "step": 12710 + }, + { + "epoch": 3.332896461336828, + "grad_norm": 1.7939388751983643, + "learning_rate": 4.216047733068481e-05, + "loss": 0.468, + "num_input_tokens_seen": 3397144, + "step": 12715 + }, + { + "epoch": 3.3342070773263432, + "grad_norm": 2.120528221130371, + "learning_rate": 4.21521582729017e-05, + "loss": 0.3724, + "num_input_tokens_seen": 3398584, + "step": 12720 + }, + { + "epoch": 3.3355176933158583, + "grad_norm": 1.5784807205200195, + "learning_rate": 4.21438356252015e-05, + "loss": 0.486, + "num_input_tokens_seen": 3399688, + "step": 12725 + }, + { + "epoch": 3.3368283093053734, + "grad_norm": 2.3456342220306396, + "learning_rate": 4.213550938932612e-05, + "loss": 0.3677, + "num_input_tokens_seen": 3400936, + "step": 12730 + }, + { + "epoch": 3.3381389252948885, + "grad_norm": 4.203124046325684, + "learning_rate": 4.2127179567018235e-05, + "loss": 0.4046, + "num_input_tokens_seen": 3402248, + "step": 12735 + }, + { + "epoch": 3.3394495412844036, + "grad_norm": 3.7841711044311523, + "learning_rate": 4.211884616002126e-05, + "loss": 0.4734, + "num_input_tokens_seen": 3403416, + "step": 12740 + }, + { + "epoch": 3.3407601572739187, + "grad_norm": 2.05029559135437, + "learning_rate": 4.211050917007937e-05, + "loss": 0.4471, + "num_input_tokens_seen": 3404760, + "step": 12745 + }, + { + "epoch": 3.342070773263434, + "grad_norm": 2.8199820518493652, + "learning_rate": 4.210216859893747e-05, + "loss": 0.4204, + "num_input_tokens_seen": 3405992, + "step": 12750 + }, + { + "epoch": 3.343381389252949, + "grad_norm": 3.632310152053833, + "learning_rate": 4.2093824448341226e-05, + "loss": 0.3954, + "num_input_tokens_seen": 3407176, + "step": 12755 + }, + { + "epoch": 3.344692005242464, + "grad_norm": 2.1368587017059326, + "learning_rate": 4.208547672003707e-05, + "loss": 0.4061, + "num_input_tokens_seen": 3408456, + "step": 12760 + }, + { + "epoch": 3.346002621231979, + "grad_norm": 1.092263102531433, + "learning_rate": 4.207712541577215e-05, + "loss": 0.524, + "num_input_tokens_seen": 3409848, + "step": 12765 + }, + { + "epoch": 3.347313237221494, + "grad_norm": 2.876154661178589, + "learning_rate": 4.2068770537294395e-05, + "loss": 0.4586, + "num_input_tokens_seen": 3411432, + "step": 12770 + }, + { + "epoch": 3.3486238532110093, + "grad_norm": 2.995006561279297, + "learning_rate": 4.2060412086352455e-05, + "loss": 0.4284, + "num_input_tokens_seen": 3412696, + "step": 12775 + }, + { + "epoch": 3.3499344692005244, + "grad_norm": 3.1759350299835205, + "learning_rate": 4.2052050064695746e-05, + "loss": 0.5144, + "num_input_tokens_seen": 3413736, + "step": 12780 + }, + { + "epoch": 3.3512450851900395, + "grad_norm": 1.57246994972229, + "learning_rate": 4.204368447407442e-05, + "loss": 0.4302, + "num_input_tokens_seen": 3415256, + "step": 12785 + }, + { + "epoch": 3.352555701179554, + "grad_norm": 1.9064555168151855, + "learning_rate": 4.203531531623938e-05, + "loss": 0.5097, + "num_input_tokens_seen": 3416456, + "step": 12790 + }, + { + "epoch": 3.3538663171690697, + "grad_norm": 1.7141088247299194, + "learning_rate": 4.2026942592942285e-05, + "loss": 0.4874, + "num_input_tokens_seen": 3418152, + "step": 12795 + }, + { + "epoch": 3.3551769331585843, + "grad_norm": 2.7995433807373047, + "learning_rate": 4.2018566305935525e-05, + "loss": 0.3545, + "num_input_tokens_seen": 3419256, + "step": 12800 + }, + { + "epoch": 3.3564875491480994, + "grad_norm": 2.058701515197754, + "learning_rate": 4.201018645697224e-05, + "loss": 0.3118, + "num_input_tokens_seen": 3420408, + "step": 12805 + }, + { + "epoch": 3.3577981651376145, + "grad_norm": 1.7748310565948486, + "learning_rate": 4.200180304780632e-05, + "loss": 0.4736, + "num_input_tokens_seen": 3421592, + "step": 12810 + }, + { + "epoch": 3.3591087811271296, + "grad_norm": 3.151360034942627, + "learning_rate": 4.1993416080192404e-05, + "loss": 0.3758, + "num_input_tokens_seen": 3422888, + "step": 12815 + }, + { + "epoch": 3.3604193971166447, + "grad_norm": 2.2410647869110107, + "learning_rate": 4.198502555588586e-05, + "loss": 0.3582, + "num_input_tokens_seen": 3424296, + "step": 12820 + }, + { + "epoch": 3.36173001310616, + "grad_norm": 8.241905212402344, + "learning_rate": 4.197663147664281e-05, + "loss": 0.3632, + "num_input_tokens_seen": 3425272, + "step": 12825 + }, + { + "epoch": 3.363040629095675, + "grad_norm": 2.9513638019561768, + "learning_rate": 4.196823384422014e-05, + "loss": 0.417, + "num_input_tokens_seen": 3426328, + "step": 12830 + }, + { + "epoch": 3.36435124508519, + "grad_norm": 4.403711795806885, + "learning_rate": 4.195983266037543e-05, + "loss": 0.3359, + "num_input_tokens_seen": 3427528, + "step": 12835 + }, + { + "epoch": 3.365661861074705, + "grad_norm": 5.335869789123535, + "learning_rate": 4.1951427926867056e-05, + "loss": 0.6155, + "num_input_tokens_seen": 3428824, + "step": 12840 + }, + { + "epoch": 3.36697247706422, + "grad_norm": 3.525550365447998, + "learning_rate": 4.1943019645454115e-05, + "loss": 0.4741, + "num_input_tokens_seen": 3429832, + "step": 12845 + }, + { + "epoch": 3.3682830930537353, + "grad_norm": 1.6100692749023438, + "learning_rate": 4.193460781789642e-05, + "loss": 0.4617, + "num_input_tokens_seen": 3431208, + "step": 12850 + }, + { + "epoch": 3.3695937090432504, + "grad_norm": 2.872535467147827, + "learning_rate": 4.192619244595458e-05, + "loss": 0.3622, + "num_input_tokens_seen": 3432344, + "step": 12855 + }, + { + "epoch": 3.3709043250327655, + "grad_norm": 1.8976712226867676, + "learning_rate": 4.191777353138991e-05, + "loss": 0.4931, + "num_input_tokens_seen": 3433832, + "step": 12860 + }, + { + "epoch": 3.3722149410222806, + "grad_norm": 1.7923122644424438, + "learning_rate": 4.1909351075964464e-05, + "loss": 0.7435, + "num_input_tokens_seen": 3435432, + "step": 12865 + }, + { + "epoch": 3.3735255570117957, + "grad_norm": 1.8181023597717285, + "learning_rate": 4.190092508144106e-05, + "loss": 0.4595, + "num_input_tokens_seen": 3436632, + "step": 12870 + }, + { + "epoch": 3.374836173001311, + "grad_norm": 2.140110731124878, + "learning_rate": 4.189249554958324e-05, + "loss": 0.3351, + "num_input_tokens_seen": 3438168, + "step": 12875 + }, + { + "epoch": 3.376146788990826, + "grad_norm": 3.238630533218384, + "learning_rate": 4.188406248215528e-05, + "loss": 0.38, + "num_input_tokens_seen": 3439320, + "step": 12880 + }, + { + "epoch": 3.3774574049803405, + "grad_norm": 2.516829490661621, + "learning_rate": 4.1875625880922225e-05, + "loss": 0.4544, + "num_input_tokens_seen": 3440744, + "step": 12885 + }, + { + "epoch": 3.378768020969856, + "grad_norm": 2.0201003551483154, + "learning_rate": 4.186718574764983e-05, + "loss": 0.5166, + "num_input_tokens_seen": 3441992, + "step": 12890 + }, + { + "epoch": 3.3800786369593707, + "grad_norm": 2.6296446323394775, + "learning_rate": 4.18587420841046e-05, + "loss": 0.5105, + "num_input_tokens_seen": 3443128, + "step": 12895 + }, + { + "epoch": 3.381389252948886, + "grad_norm": 2.265410900115967, + "learning_rate": 4.185029489205379e-05, + "loss": 0.4655, + "num_input_tokens_seen": 3444328, + "step": 12900 + }, + { + "epoch": 3.382699868938401, + "grad_norm": 1.287039041519165, + "learning_rate": 4.1841844173265357e-05, + "loss": 0.3996, + "num_input_tokens_seen": 3445768, + "step": 12905 + }, + { + "epoch": 3.384010484927916, + "grad_norm": 1.6157633066177368, + "learning_rate": 4.183338992950805e-05, + "loss": 0.4644, + "num_input_tokens_seen": 3447288, + "step": 12910 + }, + { + "epoch": 3.385321100917431, + "grad_norm": 3.8836543560028076, + "learning_rate": 4.182493216255132e-05, + "loss": 0.4437, + "num_input_tokens_seen": 3448360, + "step": 12915 + }, + { + "epoch": 3.386631716906946, + "grad_norm": 3.6440956592559814, + "learning_rate": 4.181647087416536e-05, + "loss": 0.4374, + "num_input_tokens_seen": 3449912, + "step": 12920 + }, + { + "epoch": 3.3879423328964613, + "grad_norm": 2.2043614387512207, + "learning_rate": 4.18080060661211e-05, + "loss": 0.4795, + "num_input_tokens_seen": 3451304, + "step": 12925 + }, + { + "epoch": 3.3892529488859764, + "grad_norm": 2.713613510131836, + "learning_rate": 4.179953774019021e-05, + "loss": 0.3426, + "num_input_tokens_seen": 3452216, + "step": 12930 + }, + { + "epoch": 3.3905635648754915, + "grad_norm": 2.939441680908203, + "learning_rate": 4.1791065898145095e-05, + "loss": 0.4562, + "num_input_tokens_seen": 3453400, + "step": 12935 + }, + { + "epoch": 3.3918741808650066, + "grad_norm": 1.940915822982788, + "learning_rate": 4.178259054175891e-05, + "loss": 0.5988, + "num_input_tokens_seen": 3454712, + "step": 12940 + }, + { + "epoch": 3.3931847968545217, + "grad_norm": 2.35465145111084, + "learning_rate": 4.177411167280551e-05, + "loss": 0.3393, + "num_input_tokens_seen": 3456184, + "step": 12945 + }, + { + "epoch": 3.3944954128440368, + "grad_norm": 1.2920045852661133, + "learning_rate": 4.176562929305953e-05, + "loss": 0.3524, + "num_input_tokens_seen": 3457640, + "step": 12950 + }, + { + "epoch": 3.395806028833552, + "grad_norm": 3.08408522605896, + "learning_rate": 4.1757143404296305e-05, + "loss": 0.413, + "num_input_tokens_seen": 3458984, + "step": 12955 + }, + { + "epoch": 3.397116644823067, + "grad_norm": 3.559230327606201, + "learning_rate": 4.1748654008291914e-05, + "loss": 0.5846, + "num_input_tokens_seen": 3459944, + "step": 12960 + }, + { + "epoch": 3.398427260812582, + "grad_norm": 1.8004214763641357, + "learning_rate": 4.1740161106823185e-05, + "loss": 0.4733, + "num_input_tokens_seen": 3461064, + "step": 12965 + }, + { + "epoch": 3.399737876802097, + "grad_norm": 1.986181378364563, + "learning_rate": 4.173166470166766e-05, + "loss": 0.7226, + "num_input_tokens_seen": 3462344, + "step": 12970 + }, + { + "epoch": 3.4010484927916123, + "grad_norm": 6.018317699432373, + "learning_rate": 4.172316479460362e-05, + "loss": 0.5017, + "num_input_tokens_seen": 3463704, + "step": 12975 + }, + { + "epoch": 3.402359108781127, + "grad_norm": 6.70018196105957, + "learning_rate": 4.171466138741009e-05, + "loss": 0.567, + "num_input_tokens_seen": 3464552, + "step": 12980 + }, + { + "epoch": 3.4036697247706424, + "grad_norm": 1.6213704347610474, + "learning_rate": 4.170615448186681e-05, + "loss": 0.3453, + "num_input_tokens_seen": 3466104, + "step": 12985 + }, + { + "epoch": 3.404980340760157, + "grad_norm": 4.0602264404296875, + "learning_rate": 4.169764407975426e-05, + "loss": 0.3872, + "num_input_tokens_seen": 3467128, + "step": 12990 + }, + { + "epoch": 3.406290956749672, + "grad_norm": 2.344557523727417, + "learning_rate": 4.168913018285366e-05, + "loss": 0.4174, + "num_input_tokens_seen": 3468392, + "step": 12995 + }, + { + "epoch": 3.4076015727391873, + "grad_norm": 2.165178060531616, + "learning_rate": 4.168061279294694e-05, + "loss": 0.4552, + "num_input_tokens_seen": 3469496, + "step": 13000 + }, + { + "epoch": 3.4089121887287024, + "grad_norm": 3.544482469558716, + "learning_rate": 4.16720919118168e-05, + "loss": 0.4995, + "num_input_tokens_seen": 3470760, + "step": 13005 + }, + { + "epoch": 3.4102228047182175, + "grad_norm": 1.6858679056167603, + "learning_rate": 4.1663567541246615e-05, + "loss": 0.7049, + "num_input_tokens_seen": 3471752, + "step": 13010 + }, + { + "epoch": 3.4115334207077326, + "grad_norm": 6.0725908279418945, + "learning_rate": 4.165503968302055e-05, + "loss": 0.4612, + "num_input_tokens_seen": 3472616, + "step": 13015 + }, + { + "epoch": 3.4128440366972477, + "grad_norm": 1.5176907777786255, + "learning_rate": 4.1646508338923446e-05, + "loss": 0.463, + "num_input_tokens_seen": 3473768, + "step": 13020 + }, + { + "epoch": 3.4141546526867628, + "grad_norm": 1.620079517364502, + "learning_rate": 4.1637973510740914e-05, + "loss": 0.4089, + "num_input_tokens_seen": 3475064, + "step": 13025 + }, + { + "epoch": 3.415465268676278, + "grad_norm": 1.9413129091262817, + "learning_rate": 4.162943520025927e-05, + "loss": 0.6522, + "num_input_tokens_seen": 3477000, + "step": 13030 + }, + { + "epoch": 3.416775884665793, + "grad_norm": 2.996131420135498, + "learning_rate": 4.162089340926558e-05, + "loss": 0.5194, + "num_input_tokens_seen": 3478520, + "step": 13035 + }, + { + "epoch": 3.418086500655308, + "grad_norm": 2.7355291843414307, + "learning_rate": 4.161234813954761e-05, + "loss": 0.314, + "num_input_tokens_seen": 3479976, + "step": 13040 + }, + { + "epoch": 3.419397116644823, + "grad_norm": 7.578691482543945, + "learning_rate": 4.160379939289388e-05, + "loss": 0.515, + "num_input_tokens_seen": 3481352, + "step": 13045 + }, + { + "epoch": 3.4207077326343382, + "grad_norm": 3.7907400131225586, + "learning_rate": 4.1595247171093635e-05, + "loss": 0.4145, + "num_input_tokens_seen": 3482760, + "step": 13050 + }, + { + "epoch": 3.4220183486238533, + "grad_norm": 1.8568543195724487, + "learning_rate": 4.158669147593682e-05, + "loss": 0.6624, + "num_input_tokens_seen": 3484232, + "step": 13055 + }, + { + "epoch": 3.4233289646133684, + "grad_norm": 1.048789620399475, + "learning_rate": 4.157813230921414e-05, + "loss": 0.35, + "num_input_tokens_seen": 3485960, + "step": 13060 + }, + { + "epoch": 3.4246395806028835, + "grad_norm": 1.1594876050949097, + "learning_rate": 4.156956967271702e-05, + "loss": 0.2719, + "num_input_tokens_seen": 3488296, + "step": 13065 + }, + { + "epoch": 3.4259501965923986, + "grad_norm": 2.125053644180298, + "learning_rate": 4.156100356823759e-05, + "loss": 0.4971, + "num_input_tokens_seen": 3489640, + "step": 13070 + }, + { + "epoch": 3.4272608125819133, + "grad_norm": 2.288902997970581, + "learning_rate": 4.1552433997568736e-05, + "loss": 0.4923, + "num_input_tokens_seen": 3490856, + "step": 13075 + }, + { + "epoch": 3.4285714285714284, + "grad_norm": 4.509305953979492, + "learning_rate": 4.154386096250404e-05, + "loss": 0.9064, + "num_input_tokens_seen": 3492248, + "step": 13080 + }, + { + "epoch": 3.4298820445609435, + "grad_norm": 1.5998525619506836, + "learning_rate": 4.153528446483783e-05, + "loss": 0.505, + "num_input_tokens_seen": 3493176, + "step": 13085 + }, + { + "epoch": 3.4311926605504586, + "grad_norm": 1.3858137130737305, + "learning_rate": 4.1526704506365153e-05, + "loss": 0.3711, + "num_input_tokens_seen": 3494696, + "step": 13090 + }, + { + "epoch": 3.4325032765399737, + "grad_norm": 1.9501672983169556, + "learning_rate": 4.1518121088881785e-05, + "loss": 0.4154, + "num_input_tokens_seen": 3495544, + "step": 13095 + }, + { + "epoch": 3.4338138925294888, + "grad_norm": 6.545881271362305, + "learning_rate": 4.1509534214184206e-05, + "loss": 0.8645, + "num_input_tokens_seen": 3496632, + "step": 13100 + }, + { + "epoch": 3.435124508519004, + "grad_norm": 4.522979259490967, + "learning_rate": 4.1500943884069644e-05, + "loss": 0.3789, + "num_input_tokens_seen": 3498056, + "step": 13105 + }, + { + "epoch": 3.436435124508519, + "grad_norm": 3.5239102840423584, + "learning_rate": 4.149235010033604e-05, + "loss": 0.4884, + "num_input_tokens_seen": 3499512, + "step": 13110 + }, + { + "epoch": 3.437745740498034, + "grad_norm": 1.4546486139297485, + "learning_rate": 4.148375286478206e-05, + "loss": 0.524, + "num_input_tokens_seen": 3501032, + "step": 13115 + }, + { + "epoch": 3.439056356487549, + "grad_norm": 2.0702409744262695, + "learning_rate": 4.147515217920709e-05, + "loss": 0.4982, + "num_input_tokens_seen": 3503048, + "step": 13120 + }, + { + "epoch": 3.4403669724770642, + "grad_norm": 1.1830066442489624, + "learning_rate": 4.146654804541122e-05, + "loss": 0.6106, + "num_input_tokens_seen": 3504792, + "step": 13125 + }, + { + "epoch": 3.4416775884665793, + "grad_norm": 2.084977865219116, + "learning_rate": 4.145794046519531e-05, + "loss": 0.4584, + "num_input_tokens_seen": 3505880, + "step": 13130 + }, + { + "epoch": 3.4429882044560944, + "grad_norm": 3.418858528137207, + "learning_rate": 4.14493294403609e-05, + "loss": 0.3677, + "num_input_tokens_seen": 3507080, + "step": 13135 + }, + { + "epoch": 3.4442988204456095, + "grad_norm": 0.9783729910850525, + "learning_rate": 4.144071497271025e-05, + "loss": 0.5723, + "num_input_tokens_seen": 3508872, + "step": 13140 + }, + { + "epoch": 3.4456094364351246, + "grad_norm": 1.2326632738113403, + "learning_rate": 4.143209706404637e-05, + "loss": 0.8224, + "num_input_tokens_seen": 3509928, + "step": 13145 + }, + { + "epoch": 3.4469200524246397, + "grad_norm": 3.3913331031799316, + "learning_rate": 4.142347571617296e-05, + "loss": 0.4118, + "num_input_tokens_seen": 3511288, + "step": 13150 + }, + { + "epoch": 3.448230668414155, + "grad_norm": 3.1125478744506836, + "learning_rate": 4.1414850930894466e-05, + "loss": 0.4083, + "num_input_tokens_seen": 3512776, + "step": 13155 + }, + { + "epoch": 3.44954128440367, + "grad_norm": 1.8433901071548462, + "learning_rate": 4.1406222710016034e-05, + "loss": 0.404, + "num_input_tokens_seen": 3514072, + "step": 13160 + }, + { + "epoch": 3.450851900393185, + "grad_norm": 2.0399091243743896, + "learning_rate": 4.1397591055343534e-05, + "loss": 0.4966, + "num_input_tokens_seen": 3515752, + "step": 13165 + }, + { + "epoch": 3.4521625163826997, + "grad_norm": 3.08992075920105, + "learning_rate": 4.1388955968683565e-05, + "loss": 0.646, + "num_input_tokens_seen": 3516888, + "step": 13170 + }, + { + "epoch": 3.4534731323722148, + "grad_norm": 4.398414611816406, + "learning_rate": 4.1380317451843425e-05, + "loss": 0.4677, + "num_input_tokens_seen": 3518264, + "step": 13175 + }, + { + "epoch": 3.45478374836173, + "grad_norm": 2.04459810256958, + "learning_rate": 4.137167550663115e-05, + "loss": 0.453, + "num_input_tokens_seen": 3520392, + "step": 13180 + }, + { + "epoch": 3.456094364351245, + "grad_norm": 2.121272325515747, + "learning_rate": 4.136303013485547e-05, + "loss": 0.6606, + "num_input_tokens_seen": 3521560, + "step": 13185 + }, + { + "epoch": 3.45740498034076, + "grad_norm": 8.099733352661133, + "learning_rate": 4.1354381338325864e-05, + "loss": 0.7117, + "num_input_tokens_seen": 3522424, + "step": 13190 + }, + { + "epoch": 3.458715596330275, + "grad_norm": 10.986327171325684, + "learning_rate": 4.13457291188525e-05, + "loss": 0.5953, + "num_input_tokens_seen": 3523576, + "step": 13195 + }, + { + "epoch": 3.4600262123197902, + "grad_norm": 1.443988561630249, + "learning_rate": 4.1337073478246275e-05, + "loss": 0.3072, + "num_input_tokens_seen": 3525176, + "step": 13200 + }, + { + "epoch": 3.4613368283093053, + "grad_norm": 3.2431869506835938, + "learning_rate": 4.1328414418318795e-05, + "loss": 0.5293, + "num_input_tokens_seen": 3526328, + "step": 13205 + }, + { + "epoch": 3.4626474442988204, + "grad_norm": 5.940613269805908, + "learning_rate": 4.13197519408824e-05, + "loss": 0.4586, + "num_input_tokens_seen": 3527704, + "step": 13210 + }, + { + "epoch": 3.4639580602883355, + "grad_norm": 2.2511653900146484, + "learning_rate": 4.1311086047750114e-05, + "loss": 0.4397, + "num_input_tokens_seen": 3528840, + "step": 13215 + }, + { + "epoch": 3.4652686762778506, + "grad_norm": 1.8816132545471191, + "learning_rate": 4.13024167407357e-05, + "loss": 0.3593, + "num_input_tokens_seen": 3530120, + "step": 13220 + }, + { + "epoch": 3.4665792922673657, + "grad_norm": 1.8069645166397095, + "learning_rate": 4.129374402165363e-05, + "loss": 0.6704, + "num_input_tokens_seen": 3531320, + "step": 13225 + }, + { + "epoch": 3.467889908256881, + "grad_norm": 2.7613346576690674, + "learning_rate": 4.128506789231911e-05, + "loss": 0.4797, + "num_input_tokens_seen": 3532232, + "step": 13230 + }, + { + "epoch": 3.469200524246396, + "grad_norm": 6.310164928436279, + "learning_rate": 4.1276388354548e-05, + "loss": 0.4848, + "num_input_tokens_seen": 3533320, + "step": 13235 + }, + { + "epoch": 3.470511140235911, + "grad_norm": 1.8221491575241089, + "learning_rate": 4.126770541015693e-05, + "loss": 0.5086, + "num_input_tokens_seen": 3534584, + "step": 13240 + }, + { + "epoch": 3.471821756225426, + "grad_norm": 2.1953184604644775, + "learning_rate": 4.1259019060963236e-05, + "loss": 0.3151, + "num_input_tokens_seen": 3535928, + "step": 13245 + }, + { + "epoch": 3.473132372214941, + "grad_norm": 1.9899753332138062, + "learning_rate": 4.125032930878493e-05, + "loss": 0.4331, + "num_input_tokens_seen": 3537480, + "step": 13250 + }, + { + "epoch": 3.474442988204456, + "grad_norm": 10.936208724975586, + "learning_rate": 4.1241636155440785e-05, + "loss": 0.7124, + "num_input_tokens_seen": 3538520, + "step": 13255 + }, + { + "epoch": 3.4757536041939714, + "grad_norm": 2.078204870223999, + "learning_rate": 4.123293960275027e-05, + "loss": 0.5549, + "num_input_tokens_seen": 3539704, + "step": 13260 + }, + { + "epoch": 3.477064220183486, + "grad_norm": 1.585879921913147, + "learning_rate": 4.1224239652533526e-05, + "loss": 0.5551, + "num_input_tokens_seen": 3540952, + "step": 13265 + }, + { + "epoch": 3.478374836173001, + "grad_norm": 2.495692491531372, + "learning_rate": 4.121553630661147e-05, + "loss": 0.4417, + "num_input_tokens_seen": 3542392, + "step": 13270 + }, + { + "epoch": 3.4796854521625162, + "grad_norm": 4.365414619445801, + "learning_rate": 4.120682956680567e-05, + "loss": 0.3399, + "num_input_tokens_seen": 3543944, + "step": 13275 + }, + { + "epoch": 3.4809960681520313, + "grad_norm": 1.776075005531311, + "learning_rate": 4.1198119434938455e-05, + "loss": 0.5078, + "num_input_tokens_seen": 3545288, + "step": 13280 + }, + { + "epoch": 3.4823066841415464, + "grad_norm": 2.0115528106689453, + "learning_rate": 4.118940591283282e-05, + "loss": 0.5481, + "num_input_tokens_seen": 3547016, + "step": 13285 + }, + { + "epoch": 3.4836173001310615, + "grad_norm": 9.290218353271484, + "learning_rate": 4.118068900231251e-05, + "loss": 0.5944, + "num_input_tokens_seen": 3548200, + "step": 13290 + }, + { + "epoch": 3.4849279161205766, + "grad_norm": 4.6097822189331055, + "learning_rate": 4.117196870520195e-05, + "loss": 1.0694, + "num_input_tokens_seen": 3549400, + "step": 13295 + }, + { + "epoch": 3.4862385321100917, + "grad_norm": 1.7512990236282349, + "learning_rate": 4.116324502332628e-05, + "loss": 0.5082, + "num_input_tokens_seen": 3551016, + "step": 13300 + }, + { + "epoch": 3.487549148099607, + "grad_norm": 5.0749616622924805, + "learning_rate": 4.115451795851136e-05, + "loss": 0.5046, + "num_input_tokens_seen": 3552808, + "step": 13305 + }, + { + "epoch": 3.488859764089122, + "grad_norm": 2.315213203430176, + "learning_rate": 4.1145787512583736e-05, + "loss": 0.5445, + "num_input_tokens_seen": 3554184, + "step": 13310 + }, + { + "epoch": 3.490170380078637, + "grad_norm": 1.7278529405593872, + "learning_rate": 4.113705368737069e-05, + "loss": 0.4985, + "num_input_tokens_seen": 3555256, + "step": 13315 + }, + { + "epoch": 3.491480996068152, + "grad_norm": 1.1377629041671753, + "learning_rate": 4.112831648470018e-05, + "loss": 0.4616, + "num_input_tokens_seen": 3556632, + "step": 13320 + }, + { + "epoch": 3.492791612057667, + "grad_norm": 5.788106441497803, + "learning_rate": 4.111957590640091e-05, + "loss": 0.6903, + "num_input_tokens_seen": 3557688, + "step": 13325 + }, + { + "epoch": 3.4941022280471823, + "grad_norm": 1.4898769855499268, + "learning_rate": 4.111083195430225e-05, + "loss": 0.3416, + "num_input_tokens_seen": 3559032, + "step": 13330 + }, + { + "epoch": 3.4954128440366974, + "grad_norm": 6.8366312980651855, + "learning_rate": 4.110208463023429e-05, + "loss": 0.4018, + "num_input_tokens_seen": 3560008, + "step": 13335 + }, + { + "epoch": 3.4967234600262125, + "grad_norm": 1.980196475982666, + "learning_rate": 4.109333393602786e-05, + "loss": 0.4608, + "num_input_tokens_seen": 3561304, + "step": 13340 + }, + { + "epoch": 3.4980340760157276, + "grad_norm": 0.7747876048088074, + "learning_rate": 4.108457987351443e-05, + "loss": 0.5388, + "num_input_tokens_seen": 3563064, + "step": 13345 + }, + { + "epoch": 3.4993446920052422, + "grad_norm": 2.2415213584899902, + "learning_rate": 4.107582244452622e-05, + "loss": 0.4789, + "num_input_tokens_seen": 3564408, + "step": 13350 + }, + { + "epoch": 3.5006553079947578, + "grad_norm": 2.5532515048980713, + "learning_rate": 4.106706165089616e-05, + "loss": 0.5541, + "num_input_tokens_seen": 3565816, + "step": 13355 + }, + { + "epoch": 3.5009174311926605, + "eval_loss": 0.46881935000419617, + "eval_runtime": 18.199, + "eval_samples_per_second": 46.596, + "eval_steps_per_second": 23.298, + "num_input_tokens_seen": 3566184, + "step": 13356 + }, + { + "epoch": 3.5019659239842724, + "grad_norm": 3.1459269523620605, + "learning_rate": 4.105829749445785e-05, + "loss": 0.5593, + "num_input_tokens_seen": 3567432, + "step": 13360 + }, + { + "epoch": 3.5032765399737875, + "grad_norm": 4.3888421058654785, + "learning_rate": 4.104952997704563e-05, + "loss": 0.5182, + "num_input_tokens_seen": 3568408, + "step": 13365 + }, + { + "epoch": 3.5045871559633026, + "grad_norm": 2.9086921215057373, + "learning_rate": 4.1040759100494506e-05, + "loss": 0.4414, + "num_input_tokens_seen": 3569976, + "step": 13370 + }, + { + "epoch": 3.5058977719528177, + "grad_norm": 1.965527057647705, + "learning_rate": 4.103198486664023e-05, + "loss": 0.4833, + "num_input_tokens_seen": 3571320, + "step": 13375 + }, + { + "epoch": 3.507208387942333, + "grad_norm": 4.443036079406738, + "learning_rate": 4.1023207277319213e-05, + "loss": 0.5513, + "num_input_tokens_seen": 3572808, + "step": 13380 + }, + { + "epoch": 3.508519003931848, + "grad_norm": 1.1340224742889404, + "learning_rate": 4.10144263343686e-05, + "loss": 0.2917, + "num_input_tokens_seen": 3574136, + "step": 13385 + }, + { + "epoch": 3.509829619921363, + "grad_norm": 6.067286491394043, + "learning_rate": 4.1005642039626235e-05, + "loss": 0.4186, + "num_input_tokens_seen": 3575336, + "step": 13390 + }, + { + "epoch": 3.511140235910878, + "grad_norm": 2.619685649871826, + "learning_rate": 4.099685439493064e-05, + "loss": 0.4097, + "num_input_tokens_seen": 3576552, + "step": 13395 + }, + { + "epoch": 3.512450851900393, + "grad_norm": 5.43024206161499, + "learning_rate": 4.098806340212106e-05, + "loss": 0.5589, + "num_input_tokens_seen": 3578040, + "step": 13400 + }, + { + "epoch": 3.5137614678899083, + "grad_norm": 2.840414524078369, + "learning_rate": 4.097926906303744e-05, + "loss": 0.3698, + "num_input_tokens_seen": 3578968, + "step": 13405 + }, + { + "epoch": 3.5150720838794234, + "grad_norm": 1.0551886558532715, + "learning_rate": 4.0970471379520414e-05, + "loss": 0.5384, + "num_input_tokens_seen": 3580088, + "step": 13410 + }, + { + "epoch": 3.5163826998689385, + "grad_norm": 3.2319321632385254, + "learning_rate": 4.096167035341132e-05, + "loss": 0.5268, + "num_input_tokens_seen": 3581624, + "step": 13415 + }, + { + "epoch": 3.5176933158584536, + "grad_norm": 2.9333784580230713, + "learning_rate": 4.09528659865522e-05, + "loss": 0.4145, + "num_input_tokens_seen": 3583864, + "step": 13420 + }, + { + "epoch": 3.5190039318479687, + "grad_norm": 1.8383086919784546, + "learning_rate": 4.094405828078579e-05, + "loss": 0.5893, + "num_input_tokens_seen": 3585160, + "step": 13425 + }, + { + "epoch": 3.5203145478374838, + "grad_norm": 3.0145204067230225, + "learning_rate": 4.093524723795554e-05, + "loss": 0.5814, + "num_input_tokens_seen": 3586872, + "step": 13430 + }, + { + "epoch": 3.5216251638269984, + "grad_norm": 1.9875003099441528, + "learning_rate": 4.0926432859905574e-05, + "loss": 0.4882, + "num_input_tokens_seen": 3588456, + "step": 13435 + }, + { + "epoch": 3.522935779816514, + "grad_norm": 6.664308547973633, + "learning_rate": 4.0917615148480726e-05, + "loss": 0.4466, + "num_input_tokens_seen": 3589960, + "step": 13440 + }, + { + "epoch": 3.5242463958060286, + "grad_norm": 1.7891480922698975, + "learning_rate": 4.0908794105526537e-05, + "loss": 0.4161, + "num_input_tokens_seen": 3591096, + "step": 13445 + }, + { + "epoch": 3.525557011795544, + "grad_norm": 3.3077855110168457, + "learning_rate": 4.089996973288922e-05, + "loss": 0.4804, + "num_input_tokens_seen": 3592168, + "step": 13450 + }, + { + "epoch": 3.526867627785059, + "grad_norm": 3.4087491035461426, + "learning_rate": 4.089114203241572e-05, + "loss": 0.4744, + "num_input_tokens_seen": 3593256, + "step": 13455 + }, + { + "epoch": 3.528178243774574, + "grad_norm": 3.900193452835083, + "learning_rate": 4.088231100595364e-05, + "loss": 0.7331, + "num_input_tokens_seen": 3594504, + "step": 13460 + }, + { + "epoch": 3.529488859764089, + "grad_norm": 15.55175495147705, + "learning_rate": 4.087347665535133e-05, + "loss": 0.6021, + "num_input_tokens_seen": 3595640, + "step": 13465 + }, + { + "epoch": 3.530799475753604, + "grad_norm": 2.510932207107544, + "learning_rate": 4.0864638982457766e-05, + "loss": 0.5564, + "num_input_tokens_seen": 3597864, + "step": 13470 + }, + { + "epoch": 3.532110091743119, + "grad_norm": 1.7295775413513184, + "learning_rate": 4.085579798912268e-05, + "loss": 0.7006, + "num_input_tokens_seen": 3599160, + "step": 13475 + }, + { + "epoch": 3.5334207077326343, + "grad_norm": 3.146660804748535, + "learning_rate": 4.084695367719647e-05, + "loss": 0.3913, + "num_input_tokens_seen": 3600136, + "step": 13480 + }, + { + "epoch": 3.5347313237221494, + "grad_norm": 1.7503807544708252, + "learning_rate": 4.083810604853023e-05, + "loss": 0.4662, + "num_input_tokens_seen": 3601528, + "step": 13485 + }, + { + "epoch": 3.5360419397116645, + "grad_norm": 4.245983123779297, + "learning_rate": 4.082925510497577e-05, + "loss": 0.5221, + "num_input_tokens_seen": 3602520, + "step": 13490 + }, + { + "epoch": 3.5373525557011796, + "grad_norm": 10.548373222351074, + "learning_rate": 4.0820400848385564e-05, + "loss": 0.3883, + "num_input_tokens_seen": 3603512, + "step": 13495 + }, + { + "epoch": 3.5386631716906947, + "grad_norm": 3.127352237701416, + "learning_rate": 4.0811543280612805e-05, + "loss": 0.7009, + "num_input_tokens_seen": 3604536, + "step": 13500 + }, + { + "epoch": 3.5399737876802098, + "grad_norm": 3.789565324783325, + "learning_rate": 4.080268240351135e-05, + "loss": 0.455, + "num_input_tokens_seen": 3605976, + "step": 13505 + }, + { + "epoch": 3.541284403669725, + "grad_norm": 1.3125404119491577, + "learning_rate": 4.079381821893576e-05, + "loss": 0.4773, + "num_input_tokens_seen": 3607512, + "step": 13510 + }, + { + "epoch": 3.54259501965924, + "grad_norm": 3.2661337852478027, + "learning_rate": 4.078495072874132e-05, + "loss": 0.4679, + "num_input_tokens_seen": 3608536, + "step": 13515 + }, + { + "epoch": 3.543905635648755, + "grad_norm": 2.5279791355133057, + "learning_rate": 4.0776079934783975e-05, + "loss": 0.485, + "num_input_tokens_seen": 3609976, + "step": 13520 + }, + { + "epoch": 3.54521625163827, + "grad_norm": 2.6940689086914062, + "learning_rate": 4.076720583892034e-05, + "loss": 0.4688, + "num_input_tokens_seen": 3611112, + "step": 13525 + }, + { + "epoch": 3.546526867627785, + "grad_norm": 3.199435234069824, + "learning_rate": 4.075832844300778e-05, + "loss": 0.393, + "num_input_tokens_seen": 3612584, + "step": 13530 + }, + { + "epoch": 3.5478374836173003, + "grad_norm": 3.499659776687622, + "learning_rate": 4.0749447748904304e-05, + "loss": 0.5101, + "num_input_tokens_seen": 3613960, + "step": 13535 + }, + { + "epoch": 3.549148099606815, + "grad_norm": 5.353103160858154, + "learning_rate": 4.074056375846862e-05, + "loss": 0.7699, + "num_input_tokens_seen": 3615336, + "step": 13540 + }, + { + "epoch": 3.5504587155963305, + "grad_norm": 2.4762210845947266, + "learning_rate": 4.0731676473560145e-05, + "loss": 0.4711, + "num_input_tokens_seen": 3616632, + "step": 13545 + }, + { + "epoch": 3.551769331585845, + "grad_norm": 5.157802104949951, + "learning_rate": 4.072278589603896e-05, + "loss": 0.5983, + "num_input_tokens_seen": 3617592, + "step": 13550 + }, + { + "epoch": 3.5530799475753603, + "grad_norm": 2.753689765930176, + "learning_rate": 4.0713892027765863e-05, + "loss": 0.4485, + "num_input_tokens_seen": 3618488, + "step": 13555 + }, + { + "epoch": 3.5543905635648754, + "grad_norm": 3.1220757961273193, + "learning_rate": 4.070499487060231e-05, + "loss": 0.4438, + "num_input_tokens_seen": 3619624, + "step": 13560 + }, + { + "epoch": 3.5557011795543905, + "grad_norm": 2.209648370742798, + "learning_rate": 4.069609442641046e-05, + "loss": 0.3397, + "num_input_tokens_seen": 3620904, + "step": 13565 + }, + { + "epoch": 3.5570117955439056, + "grad_norm": 2.3399295806884766, + "learning_rate": 4.068719069705318e-05, + "loss": 0.5125, + "num_input_tokens_seen": 3622312, + "step": 13570 + }, + { + "epoch": 3.5583224115334207, + "grad_norm": 2.9018936157226562, + "learning_rate": 4.0678283684393995e-05, + "loss": 0.5101, + "num_input_tokens_seen": 3623576, + "step": 13575 + }, + { + "epoch": 3.5596330275229358, + "grad_norm": 2.2590503692626953, + "learning_rate": 4.0669373390297115e-05, + "loss": 0.4488, + "num_input_tokens_seen": 3625144, + "step": 13580 + }, + { + "epoch": 3.560943643512451, + "grad_norm": 3.3129122257232666, + "learning_rate": 4.066045981662746e-05, + "loss": 0.4719, + "num_input_tokens_seen": 3626568, + "step": 13585 + }, + { + "epoch": 3.562254259501966, + "grad_norm": 10.220577239990234, + "learning_rate": 4.065154296525063e-05, + "loss": 0.4909, + "num_input_tokens_seen": 3627768, + "step": 13590 + }, + { + "epoch": 3.563564875491481, + "grad_norm": 8.145442962646484, + "learning_rate": 4.064262283803289e-05, + "loss": 0.6984, + "num_input_tokens_seen": 3628856, + "step": 13595 + }, + { + "epoch": 3.564875491480996, + "grad_norm": 2.0642266273498535, + "learning_rate": 4.0633699436841224e-05, + "loss": 0.5056, + "num_input_tokens_seen": 3630456, + "step": 13600 + }, + { + "epoch": 3.5661861074705112, + "grad_norm": 3.599928140640259, + "learning_rate": 4.062477276354328e-05, + "loss": 0.4391, + "num_input_tokens_seen": 3631928, + "step": 13605 + }, + { + "epoch": 3.5674967234600263, + "grad_norm": 3.771709680557251, + "learning_rate": 4.06158428200074e-05, + "loss": 0.629, + "num_input_tokens_seen": 3633112, + "step": 13610 + }, + { + "epoch": 3.5688073394495414, + "grad_norm": 3.1735124588012695, + "learning_rate": 4.0606909608102595e-05, + "loss": 0.5125, + "num_input_tokens_seen": 3634152, + "step": 13615 + }, + { + "epoch": 3.5701179554390565, + "grad_norm": 2.466002941131592, + "learning_rate": 4.059797312969857e-05, + "loss": 0.4484, + "num_input_tokens_seen": 3635208, + "step": 13620 + }, + { + "epoch": 3.571428571428571, + "grad_norm": 1.1987226009368896, + "learning_rate": 4.058903338666573e-05, + "loss": 0.419, + "num_input_tokens_seen": 3636408, + "step": 13625 + }, + { + "epoch": 3.5727391874180867, + "grad_norm": 30.796852111816406, + "learning_rate": 4.058009038087513e-05, + "loss": 0.6926, + "num_input_tokens_seen": 3639416, + "step": 13630 + }, + { + "epoch": 3.5740498034076014, + "grad_norm": 9.461061477661133, + "learning_rate": 4.0571144114198536e-05, + "loss": 0.3876, + "num_input_tokens_seen": 3640440, + "step": 13635 + }, + { + "epoch": 3.575360419397117, + "grad_norm": 1.4735254049301147, + "learning_rate": 4.056219458850838e-05, + "loss": 0.4148, + "num_input_tokens_seen": 3641864, + "step": 13640 + }, + { + "epoch": 3.5766710353866316, + "grad_norm": 7.556875705718994, + "learning_rate": 4.05532418056778e-05, + "loss": 0.5615, + "num_input_tokens_seen": 3642840, + "step": 13645 + }, + { + "epoch": 3.5779816513761467, + "grad_norm": 3.1799583435058594, + "learning_rate": 4.054428576758057e-05, + "loss": 0.5329, + "num_input_tokens_seen": 3644040, + "step": 13650 + }, + { + "epoch": 3.5792922673656618, + "grad_norm": 1.0262467861175537, + "learning_rate": 4.053532647609119e-05, + "loss": 0.5011, + "num_input_tokens_seen": 3645640, + "step": 13655 + }, + { + "epoch": 3.580602883355177, + "grad_norm": 2.655125141143799, + "learning_rate": 4.0526363933084824e-05, + "loss": 0.4001, + "num_input_tokens_seen": 3647096, + "step": 13660 + }, + { + "epoch": 3.581913499344692, + "grad_norm": 5.8514814376831055, + "learning_rate": 4.051739814043731e-05, + "loss": 0.5138, + "num_input_tokens_seen": 3648280, + "step": 13665 + }, + { + "epoch": 3.583224115334207, + "grad_norm": 1.6915122270584106, + "learning_rate": 4.0508429100025185e-05, + "loss": 0.5059, + "num_input_tokens_seen": 3649720, + "step": 13670 + }, + { + "epoch": 3.584534731323722, + "grad_norm": 2.7762668132781982, + "learning_rate": 4.049945681372565e-05, + "loss": 0.361, + "num_input_tokens_seen": 3650968, + "step": 13675 + }, + { + "epoch": 3.5858453473132372, + "grad_norm": 2.1223440170288086, + "learning_rate": 4.049048128341657e-05, + "loss": 0.4422, + "num_input_tokens_seen": 3652280, + "step": 13680 + }, + { + "epoch": 3.5871559633027523, + "grad_norm": 2.4381487369537354, + "learning_rate": 4.048150251097653e-05, + "loss": 0.6255, + "num_input_tokens_seen": 3653448, + "step": 13685 + }, + { + "epoch": 3.5884665792922674, + "grad_norm": 7.637173652648926, + "learning_rate": 4.047252049828476e-05, + "loss": 0.5647, + "num_input_tokens_seen": 3654648, + "step": 13690 + }, + { + "epoch": 3.5897771952817825, + "grad_norm": 1.0855579376220703, + "learning_rate": 4.046353524722119e-05, + "loss": 0.4874, + "num_input_tokens_seen": 3655992, + "step": 13695 + }, + { + "epoch": 3.5910878112712976, + "grad_norm": 1.8355480432510376, + "learning_rate": 4.0454546759666414e-05, + "loss": 0.5817, + "num_input_tokens_seen": 3657256, + "step": 13700 + }, + { + "epoch": 3.5923984272608127, + "grad_norm": 1.2910668849945068, + "learning_rate": 4.04455550375017e-05, + "loss": 0.5888, + "num_input_tokens_seen": 3658920, + "step": 13705 + }, + { + "epoch": 3.593709043250328, + "grad_norm": 2.4650518894195557, + "learning_rate": 4.0436560082609e-05, + "loss": 0.3803, + "num_input_tokens_seen": 3660120, + "step": 13710 + }, + { + "epoch": 3.595019659239843, + "grad_norm": 3.7926864624023438, + "learning_rate": 4.0427561896870955e-05, + "loss": 0.3611, + "num_input_tokens_seen": 3661768, + "step": 13715 + }, + { + "epoch": 3.5963302752293576, + "grad_norm": 6.157313823699951, + "learning_rate": 4.041856048217085e-05, + "loss": 0.398, + "num_input_tokens_seen": 3663032, + "step": 13720 + }, + { + "epoch": 3.597640891218873, + "grad_norm": 2.044774055480957, + "learning_rate": 4.040955584039269e-05, + "loss": 0.5778, + "num_input_tokens_seen": 3664296, + "step": 13725 + }, + { + "epoch": 3.5989515072083877, + "grad_norm": 1.133025884628296, + "learning_rate": 4.040054797342112e-05, + "loss": 0.4907, + "num_input_tokens_seen": 3665736, + "step": 13730 + }, + { + "epoch": 3.6002621231979033, + "grad_norm": 3.8211941719055176, + "learning_rate": 4.039153688314145e-05, + "loss": 0.5379, + "num_input_tokens_seen": 3667672, + "step": 13735 + }, + { + "epoch": 3.601572739187418, + "grad_norm": 2.347043752670288, + "learning_rate": 4.038252257143973e-05, + "loss": 0.3844, + "num_input_tokens_seen": 3668840, + "step": 13740 + }, + { + "epoch": 3.602883355176933, + "grad_norm": 2.650506019592285, + "learning_rate": 4.03735050402026e-05, + "loss": 0.4018, + "num_input_tokens_seen": 3669704, + "step": 13745 + }, + { + "epoch": 3.604193971166448, + "grad_norm": 4.699620246887207, + "learning_rate": 4.036448429131743e-05, + "loss": 0.4007, + "num_input_tokens_seen": 3670760, + "step": 13750 + }, + { + "epoch": 3.6055045871559632, + "grad_norm": 64.23672485351562, + "learning_rate": 4.035546032667225e-05, + "loss": 0.5713, + "num_input_tokens_seen": 3671544, + "step": 13755 + }, + { + "epoch": 3.6068152031454783, + "grad_norm": 1.0087836980819702, + "learning_rate": 4.034643314815575e-05, + "loss": 0.3929, + "num_input_tokens_seen": 3672840, + "step": 13760 + }, + { + "epoch": 3.6081258191349934, + "grad_norm": 32.10564422607422, + "learning_rate": 4.0337402757657314e-05, + "loss": 0.3833, + "num_input_tokens_seen": 3673960, + "step": 13765 + }, + { + "epoch": 3.6094364351245085, + "grad_norm": 4.769331932067871, + "learning_rate": 4.032836915706698e-05, + "loss": 0.4869, + "num_input_tokens_seen": 3675224, + "step": 13770 + }, + { + "epoch": 3.6107470511140236, + "grad_norm": 2.9801950454711914, + "learning_rate": 4.0319332348275465e-05, + "loss": 0.4306, + "num_input_tokens_seen": 3676568, + "step": 13775 + }, + { + "epoch": 3.6120576671035387, + "grad_norm": 5.40316915512085, + "learning_rate": 4.031029233317416e-05, + "loss": 0.532, + "num_input_tokens_seen": 3677832, + "step": 13780 + }, + { + "epoch": 3.613368283093054, + "grad_norm": 2.607699155807495, + "learning_rate": 4.0301249113655125e-05, + "loss": 0.3537, + "num_input_tokens_seen": 3678856, + "step": 13785 + }, + { + "epoch": 3.614678899082569, + "grad_norm": 3.7833073139190674, + "learning_rate": 4.0292202691611094e-05, + "loss": 0.5905, + "num_input_tokens_seen": 3679912, + "step": 13790 + }, + { + "epoch": 3.615989515072084, + "grad_norm": 9.679998397827148, + "learning_rate": 4.028315306893545e-05, + "loss": 0.5704, + "num_input_tokens_seen": 3681608, + "step": 13795 + }, + { + "epoch": 3.617300131061599, + "grad_norm": 3.217388391494751, + "learning_rate": 4.0274100247522287e-05, + "loss": 0.4203, + "num_input_tokens_seen": 3682888, + "step": 13800 + }, + { + "epoch": 3.618610747051114, + "grad_norm": 2.355254650115967, + "learning_rate": 4.026504422926632e-05, + "loss": 0.441, + "num_input_tokens_seen": 3684520, + "step": 13805 + }, + { + "epoch": 3.6199213630406293, + "grad_norm": 3.3436567783355713, + "learning_rate": 4.025598501606299e-05, + "loss": 0.4548, + "num_input_tokens_seen": 3685688, + "step": 13810 + }, + { + "epoch": 3.621231979030144, + "grad_norm": 1.9286513328552246, + "learning_rate": 4.024692260980835e-05, + "loss": 0.4442, + "num_input_tokens_seen": 3687256, + "step": 13815 + }, + { + "epoch": 3.6225425950196595, + "grad_norm": 2.475893974304199, + "learning_rate": 4.023785701239915e-05, + "loss": 0.5115, + "num_input_tokens_seen": 3688856, + "step": 13820 + }, + { + "epoch": 3.623853211009174, + "grad_norm": 2.571798801422119, + "learning_rate": 4.022878822573281e-05, + "loss": 0.4121, + "num_input_tokens_seen": 3690008, + "step": 13825 + }, + { + "epoch": 3.625163826998689, + "grad_norm": 2.385157823562622, + "learning_rate": 4.0219716251707395e-05, + "loss": 0.3467, + "num_input_tokens_seen": 3691032, + "step": 13830 + }, + { + "epoch": 3.6264744429882043, + "grad_norm": 2.5173251628875732, + "learning_rate": 4.021064109222168e-05, + "loss": 0.4041, + "num_input_tokens_seen": 3692248, + "step": 13835 + }, + { + "epoch": 3.6277850589777194, + "grad_norm": 18.89099884033203, + "learning_rate": 4.020156274917506e-05, + "loss": 0.535, + "num_input_tokens_seen": 3693560, + "step": 13840 + }, + { + "epoch": 3.6290956749672345, + "grad_norm": 4.17224645614624, + "learning_rate": 4.019248122446763e-05, + "loss": 0.6388, + "num_input_tokens_seen": 3694696, + "step": 13845 + }, + { + "epoch": 3.6304062909567496, + "grad_norm": 2.2351949214935303, + "learning_rate": 4.0183396520000114e-05, + "loss": 0.5532, + "num_input_tokens_seen": 3696008, + "step": 13850 + }, + { + "epoch": 3.6317169069462647, + "grad_norm": 7.390750885009766, + "learning_rate": 4.0174308637673956e-05, + "loss": 0.5436, + "num_input_tokens_seen": 3697112, + "step": 13855 + }, + { + "epoch": 3.63302752293578, + "grad_norm": 2.0511152744293213, + "learning_rate": 4.016521757939121e-05, + "loss": 0.382, + "num_input_tokens_seen": 3698264, + "step": 13860 + }, + { + "epoch": 3.634338138925295, + "grad_norm": 6.8107476234436035, + "learning_rate": 4.015612334705464e-05, + "loss": 0.3748, + "num_input_tokens_seen": 3699320, + "step": 13865 + }, + { + "epoch": 3.63564875491481, + "grad_norm": 1.707677960395813, + "learning_rate": 4.014702594256764e-05, + "loss": 0.3291, + "num_input_tokens_seen": 3700392, + "step": 13870 + }, + { + "epoch": 3.636959370904325, + "grad_norm": 4.166021347045898, + "learning_rate": 4.013792536783427e-05, + "loss": 0.3776, + "num_input_tokens_seen": 3701560, + "step": 13875 + }, + { + "epoch": 3.63826998689384, + "grad_norm": 3.1908867359161377, + "learning_rate": 4.012882162475928e-05, + "loss": 0.511, + "num_input_tokens_seen": 3702856, + "step": 13880 + }, + { + "epoch": 3.6395806028833553, + "grad_norm": 1.343970537185669, + "learning_rate": 4.011971471524807e-05, + "loss": 0.5949, + "num_input_tokens_seen": 3704136, + "step": 13885 + }, + { + "epoch": 3.6408912188728704, + "grad_norm": 1.301363229751587, + "learning_rate": 4.011060464120669e-05, + "loss": 0.4384, + "num_input_tokens_seen": 3705480, + "step": 13890 + }, + { + "epoch": 3.6422018348623855, + "grad_norm": 1.7653529644012451, + "learning_rate": 4.010149140454188e-05, + "loss": 0.4414, + "num_input_tokens_seen": 3706840, + "step": 13895 + }, + { + "epoch": 3.6435124508519, + "grad_norm": 4.367697715759277, + "learning_rate": 4.009237500716101e-05, + "loss": 0.4607, + "num_input_tokens_seen": 3708312, + "step": 13900 + }, + { + "epoch": 3.6448230668414157, + "grad_norm": 1.6357060670852661, + "learning_rate": 4.008325545097212e-05, + "loss": 0.8923, + "num_input_tokens_seen": 3709752, + "step": 13905 + }, + { + "epoch": 3.6461336828309303, + "grad_norm": 1.2585744857788086, + "learning_rate": 4.0074132737883927e-05, + "loss": 0.567, + "num_input_tokens_seen": 3711000, + "step": 13910 + }, + { + "epoch": 3.647444298820446, + "grad_norm": 2.129730463027954, + "learning_rate": 4.00650068698058e-05, + "loss": 0.5145, + "num_input_tokens_seen": 3712648, + "step": 13915 + }, + { + "epoch": 3.6487549148099605, + "grad_norm": 2.3030240535736084, + "learning_rate": 4.005587784864776e-05, + "loss": 0.4864, + "num_input_tokens_seen": 3713992, + "step": 13920 + }, + { + "epoch": 3.6500655307994756, + "grad_norm": 1.3827183246612549, + "learning_rate": 4.0046745676320504e-05, + "loss": 0.4401, + "num_input_tokens_seen": 3715592, + "step": 13925 + }, + { + "epoch": 3.6513761467889907, + "grad_norm": 5.255937099456787, + "learning_rate": 4.003761035473538e-05, + "loss": 0.6342, + "num_input_tokens_seen": 3716584, + "step": 13930 + }, + { + "epoch": 3.652686762778506, + "grad_norm": 2.0847415924072266, + "learning_rate": 4.002847188580439e-05, + "loss": 0.4549, + "num_input_tokens_seen": 3717944, + "step": 13935 + }, + { + "epoch": 3.653997378768021, + "grad_norm": 4.272425174713135, + "learning_rate": 4.0019330271440206e-05, + "loss": 0.5684, + "num_input_tokens_seen": 3719112, + "step": 13940 + }, + { + "epoch": 3.655307994757536, + "grad_norm": 1.7819417715072632, + "learning_rate": 4.001018551355614e-05, + "loss": 0.5288, + "num_input_tokens_seen": 3720360, + "step": 13945 + }, + { + "epoch": 3.656618610747051, + "grad_norm": 2.6649534702301025, + "learning_rate": 4.0001037614066184e-05, + "loss": 0.3874, + "num_input_tokens_seen": 3721816, + "step": 13950 + }, + { + "epoch": 3.657929226736566, + "grad_norm": 1.8117824792861938, + "learning_rate": 3.9991886574884975e-05, + "loss": 0.3477, + "num_input_tokens_seen": 3723224, + "step": 13955 + }, + { + "epoch": 3.6592398427260813, + "grad_norm": 2.487265110015869, + "learning_rate": 3.998273239792781e-05, + "loss": 0.4667, + "num_input_tokens_seen": 3724584, + "step": 13960 + }, + { + "epoch": 3.6605504587155964, + "grad_norm": 1.555302619934082, + "learning_rate": 3.997357508511064e-05, + "loss": 0.3415, + "num_input_tokens_seen": 3728152, + "step": 13965 + }, + { + "epoch": 3.6618610747051115, + "grad_norm": 6.064899921417236, + "learning_rate": 3.996441463835008e-05, + "loss": 0.5004, + "num_input_tokens_seen": 3729288, + "step": 13970 + }, + { + "epoch": 3.6631716906946266, + "grad_norm": 2.8111555576324463, + "learning_rate": 3.995525105956339e-05, + "loss": 0.342, + "num_input_tokens_seen": 3731160, + "step": 13975 + }, + { + "epoch": 3.6644823066841417, + "grad_norm": 2.468923330307007, + "learning_rate": 3.9946084350668506e-05, + "loss": 0.7121, + "num_input_tokens_seen": 3732296, + "step": 13980 + }, + { + "epoch": 3.6657929226736568, + "grad_norm": 6.218713283538818, + "learning_rate": 3.993691451358398e-05, + "loss": 0.745, + "num_input_tokens_seen": 3733720, + "step": 13985 + }, + { + "epoch": 3.667103538663172, + "grad_norm": 2.1852235794067383, + "learning_rate": 3.992774155022906e-05, + "loss": 0.6475, + "num_input_tokens_seen": 3735224, + "step": 13990 + }, + { + "epoch": 3.6684141546526865, + "grad_norm": 22.504674911499023, + "learning_rate": 3.991856546252362e-05, + "loss": 0.3687, + "num_input_tokens_seen": 3736392, + "step": 13995 + }, + { + "epoch": 3.669724770642202, + "grad_norm": 1.3865940570831299, + "learning_rate": 3.9909386252388215e-05, + "loss": 0.3584, + "num_input_tokens_seen": 3738136, + "step": 14000 + }, + { + "epoch": 3.6710353866317167, + "grad_norm": 4.247875213623047, + "learning_rate": 3.990020392174402e-05, + "loss": 0.3761, + "num_input_tokens_seen": 3739016, + "step": 14005 + }, + { + "epoch": 3.6723460026212322, + "grad_norm": 2.9622802734375, + "learning_rate": 3.9891018472512895e-05, + "loss": 0.5249, + "num_input_tokens_seen": 3740536, + "step": 14010 + }, + { + "epoch": 3.673656618610747, + "grad_norm": 2.78124737739563, + "learning_rate": 3.988182990661734e-05, + "loss": 0.4374, + "num_input_tokens_seen": 3742744, + "step": 14015 + }, + { + "epoch": 3.674967234600262, + "grad_norm": 4.675894737243652, + "learning_rate": 3.987263822598049e-05, + "loss": 0.5258, + "num_input_tokens_seen": 3743944, + "step": 14020 + }, + { + "epoch": 3.676277850589777, + "grad_norm": 3.8755502700805664, + "learning_rate": 3.986344343252615e-05, + "loss": 0.5426, + "num_input_tokens_seen": 3745272, + "step": 14025 + }, + { + "epoch": 3.677588466579292, + "grad_norm": 3.3685593605041504, + "learning_rate": 3.9854245528178804e-05, + "loss": 0.449, + "num_input_tokens_seen": 3746376, + "step": 14030 + }, + { + "epoch": 3.6788990825688073, + "grad_norm": 2.907137632369995, + "learning_rate": 3.984504451486352e-05, + "loss": 0.5476, + "num_input_tokens_seen": 3747640, + "step": 14035 + }, + { + "epoch": 3.6802096985583224, + "grad_norm": 2.887878179550171, + "learning_rate": 3.983584039450607e-05, + "loss": 0.5948, + "num_input_tokens_seen": 3749000, + "step": 14040 + }, + { + "epoch": 3.6815203145478375, + "grad_norm": 10.49286937713623, + "learning_rate": 3.982663316903286e-05, + "loss": 0.4769, + "num_input_tokens_seen": 3750168, + "step": 14045 + }, + { + "epoch": 3.6828309305373526, + "grad_norm": 4.428913116455078, + "learning_rate": 3.981742284037095e-05, + "loss": 0.3674, + "num_input_tokens_seen": 3751272, + "step": 14050 + }, + { + "epoch": 3.6841415465268676, + "grad_norm": 1.9116560220718384, + "learning_rate": 3.980820941044803e-05, + "loss": 0.6927, + "num_input_tokens_seen": 3752488, + "step": 14055 + }, + { + "epoch": 3.6854521625163827, + "grad_norm": 6.580589771270752, + "learning_rate": 3.979899288119248e-05, + "loss": 0.5517, + "num_input_tokens_seen": 3753816, + "step": 14060 + }, + { + "epoch": 3.686762778505898, + "grad_norm": 4.812602996826172, + "learning_rate": 3.978977325453329e-05, + "loss": 0.3536, + "num_input_tokens_seen": 3755000, + "step": 14065 + }, + { + "epoch": 3.688073394495413, + "grad_norm": 5.216959476470947, + "learning_rate": 3.9780550532400106e-05, + "loss": 0.5482, + "num_input_tokens_seen": 3756168, + "step": 14070 + }, + { + "epoch": 3.689384010484928, + "grad_norm": 5.710012435913086, + "learning_rate": 3.9771324716723246e-05, + "loss": 0.786, + "num_input_tokens_seen": 3757512, + "step": 14075 + }, + { + "epoch": 3.690694626474443, + "grad_norm": 5.139296531677246, + "learning_rate": 3.976209580943363e-05, + "loss": 0.4478, + "num_input_tokens_seen": 3758568, + "step": 14080 + }, + { + "epoch": 3.6920052424639582, + "grad_norm": 1.0508058071136475, + "learning_rate": 3.975286381246288e-05, + "loss": 0.3971, + "num_input_tokens_seen": 3759816, + "step": 14085 + }, + { + "epoch": 3.693315858453473, + "grad_norm": 1.8148722648620605, + "learning_rate": 3.9743628727743224e-05, + "loss": 0.3874, + "num_input_tokens_seen": 3760936, + "step": 14090 + }, + { + "epoch": 3.6946264744429884, + "grad_norm": 2.4455082416534424, + "learning_rate": 3.973439055720755e-05, + "loss": 0.4616, + "num_input_tokens_seen": 3762408, + "step": 14095 + }, + { + "epoch": 3.695937090432503, + "grad_norm": 2.086087226867676, + "learning_rate": 3.972514930278938e-05, + "loss": 0.4003, + "num_input_tokens_seen": 3763864, + "step": 14100 + }, + { + "epoch": 3.6972477064220186, + "grad_norm": 2.908172845840454, + "learning_rate": 3.971590496642291e-05, + "loss": 0.4732, + "num_input_tokens_seen": 3765400, + "step": 14105 + }, + { + "epoch": 3.6985583224115333, + "grad_norm": 14.561662673950195, + "learning_rate": 3.970665755004296e-05, + "loss": 0.5362, + "num_input_tokens_seen": 3766296, + "step": 14110 + }, + { + "epoch": 3.6998689384010484, + "grad_norm": 3.4615447521209717, + "learning_rate": 3.9697407055585e-05, + "loss": 0.3989, + "num_input_tokens_seen": 3767720, + "step": 14115 + }, + { + "epoch": 3.7011795543905635, + "grad_norm": 1.2126444578170776, + "learning_rate": 3.9688153484985135e-05, + "loss": 0.4397, + "num_input_tokens_seen": 3768872, + "step": 14120 + }, + { + "epoch": 3.7024901703800785, + "grad_norm": 1.2793993949890137, + "learning_rate": 3.9678896840180125e-05, + "loss": 0.4488, + "num_input_tokens_seen": 3770808, + "step": 14125 + }, + { + "epoch": 3.7038007863695936, + "grad_norm": 3.7611045837402344, + "learning_rate": 3.9669637123107374e-05, + "loss": 0.3698, + "num_input_tokens_seen": 3772264, + "step": 14130 + }, + { + "epoch": 3.7051114023591087, + "grad_norm": 1.422258973121643, + "learning_rate": 3.9660374335704906e-05, + "loss": 0.6189, + "num_input_tokens_seen": 3773976, + "step": 14135 + }, + { + "epoch": 3.706422018348624, + "grad_norm": 2.464543104171753, + "learning_rate": 3.965110847991143e-05, + "loss": 0.5462, + "num_input_tokens_seen": 3775368, + "step": 14140 + }, + { + "epoch": 3.707732634338139, + "grad_norm": 2.6136574745178223, + "learning_rate": 3.964183955766627e-05, + "loss": 0.4852, + "num_input_tokens_seen": 3776536, + "step": 14145 + }, + { + "epoch": 3.709043250327654, + "grad_norm": 1.1063271760940552, + "learning_rate": 3.963256757090938e-05, + "loss": 0.3138, + "num_input_tokens_seen": 3778632, + "step": 14150 + }, + { + "epoch": 3.710353866317169, + "grad_norm": 1.8330732583999634, + "learning_rate": 3.962329252158139e-05, + "loss": 0.3972, + "num_input_tokens_seen": 3779816, + "step": 14155 + }, + { + "epoch": 3.711664482306684, + "grad_norm": 4.002154350280762, + "learning_rate": 3.961401441162354e-05, + "loss": 0.4605, + "num_input_tokens_seen": 3781048, + "step": 14160 + }, + { + "epoch": 3.7129750982961993, + "grad_norm": 1.5970847606658936, + "learning_rate": 3.960473324297772e-05, + "loss": 0.2946, + "num_input_tokens_seen": 3782152, + "step": 14165 + }, + { + "epoch": 3.7142857142857144, + "grad_norm": 1.1162989139556885, + "learning_rate": 3.959544901758646e-05, + "loss": 0.5323, + "num_input_tokens_seen": 3783736, + "step": 14170 + }, + { + "epoch": 3.7155963302752295, + "grad_norm": 3.2276127338409424, + "learning_rate": 3.958616173739295e-05, + "loss": 0.5641, + "num_input_tokens_seen": 3785192, + "step": 14175 + }, + { + "epoch": 3.7169069462647446, + "grad_norm": 4.573520660400391, + "learning_rate": 3.9576871404340994e-05, + "loss": 0.4663, + "num_input_tokens_seen": 3786872, + "step": 14180 + }, + { + "epoch": 3.7182175622542593, + "grad_norm": 1.8491054773330688, + "learning_rate": 3.9567578020375036e-05, + "loss": 0.5274, + "num_input_tokens_seen": 3788520, + "step": 14185 + }, + { + "epoch": 3.719528178243775, + "grad_norm": 3.9951796531677246, + "learning_rate": 3.955828158744017e-05, + "loss": 0.4255, + "num_input_tokens_seen": 3789704, + "step": 14190 + }, + { + "epoch": 3.7208387942332894, + "grad_norm": 1.5546778440475464, + "learning_rate": 3.954898210748211e-05, + "loss": 0.3998, + "num_input_tokens_seen": 3790728, + "step": 14195 + }, + { + "epoch": 3.722149410222805, + "grad_norm": 1.4710763692855835, + "learning_rate": 3.953967958244725e-05, + "loss": 0.5495, + "num_input_tokens_seen": 3791864, + "step": 14200 + }, + { + "epoch": 3.7234600262123196, + "grad_norm": 2.8310389518737793, + "learning_rate": 3.9530374014282574e-05, + "loss": 0.4532, + "num_input_tokens_seen": 3793000, + "step": 14205 + }, + { + "epoch": 3.7247706422018347, + "grad_norm": 2.6914477348327637, + "learning_rate": 3.952106540493572e-05, + "loss": 0.4507, + "num_input_tokens_seen": 3794504, + "step": 14210 + }, + { + "epoch": 3.72608125819135, + "grad_norm": 4.706211090087891, + "learning_rate": 3.951175375635497e-05, + "loss": 0.436, + "num_input_tokens_seen": 3795976, + "step": 14215 + }, + { + "epoch": 3.727391874180865, + "grad_norm": 3.8700428009033203, + "learning_rate": 3.9502439070489224e-05, + "loss": 0.4037, + "num_input_tokens_seen": 3797304, + "step": 14220 + }, + { + "epoch": 3.72870249017038, + "grad_norm": 41.4946174621582, + "learning_rate": 3.949312134928805e-05, + "loss": 0.5811, + "num_input_tokens_seen": 3798488, + "step": 14225 + }, + { + "epoch": 3.730013106159895, + "grad_norm": 15.087957382202148, + "learning_rate": 3.948380059470161e-05, + "loss": 0.5879, + "num_input_tokens_seen": 3799672, + "step": 14230 + }, + { + "epoch": 3.73132372214941, + "grad_norm": 3.440800905227661, + "learning_rate": 3.947447680868074e-05, + "loss": 0.5409, + "num_input_tokens_seen": 3800808, + "step": 14235 + }, + { + "epoch": 3.7326343381389253, + "grad_norm": 3.2137863636016846, + "learning_rate": 3.9465149993176885e-05, + "loss": 0.4728, + "num_input_tokens_seen": 3802184, + "step": 14240 + }, + { + "epoch": 3.7339449541284404, + "grad_norm": 2.925546169281006, + "learning_rate": 3.945582015014212e-05, + "loss": 0.3572, + "num_input_tokens_seen": 3803544, + "step": 14245 + }, + { + "epoch": 3.7352555701179555, + "grad_norm": 2.2378275394439697, + "learning_rate": 3.944648728152919e-05, + "loss": 0.4166, + "num_input_tokens_seen": 3804696, + "step": 14250 + }, + { + "epoch": 3.7365661861074706, + "grad_norm": 1.430721402168274, + "learning_rate": 3.943715138929142e-05, + "loss": 0.4754, + "num_input_tokens_seen": 3805912, + "step": 14255 + }, + { + "epoch": 3.7378768020969857, + "grad_norm": 3.6426053047180176, + "learning_rate": 3.942781247538282e-05, + "loss": 0.4735, + "num_input_tokens_seen": 3807288, + "step": 14260 + }, + { + "epoch": 3.739187418086501, + "grad_norm": 1.1731704473495483, + "learning_rate": 3.941847054175799e-05, + "loss": 0.7072, + "num_input_tokens_seen": 3808824, + "step": 14265 + }, + { + "epoch": 3.740498034076016, + "grad_norm": 1.376876950263977, + "learning_rate": 3.940912559037219e-05, + "loss": 0.3818, + "num_input_tokens_seen": 3810040, + "step": 14270 + }, + { + "epoch": 3.741808650065531, + "grad_norm": 2.005920648574829, + "learning_rate": 3.939977762318131e-05, + "loss": 0.4301, + "num_input_tokens_seen": 3811672, + "step": 14275 + }, + { + "epoch": 3.7431192660550456, + "grad_norm": 1.0279498100280762, + "learning_rate": 3.939042664214184e-05, + "loss": 0.3473, + "num_input_tokens_seen": 3813176, + "step": 14280 + }, + { + "epoch": 3.744429882044561, + "grad_norm": 3.1663594245910645, + "learning_rate": 3.938107264921095e-05, + "loss": 0.5135, + "num_input_tokens_seen": 3814280, + "step": 14285 + }, + { + "epoch": 3.745740498034076, + "grad_norm": 1.690097451210022, + "learning_rate": 3.9371715646346396e-05, + "loss": 0.4746, + "num_input_tokens_seen": 3815800, + "step": 14290 + }, + { + "epoch": 3.747051114023591, + "grad_norm": 2.334838390350342, + "learning_rate": 3.936235563550659e-05, + "loss": 0.433, + "num_input_tokens_seen": 3817112, + "step": 14295 + }, + { + "epoch": 3.748361730013106, + "grad_norm": 1.2480294704437256, + "learning_rate": 3.935299261865057e-05, + "loss": 0.452, + "num_input_tokens_seen": 3818648, + "step": 14300 + }, + { + "epoch": 3.749672346002621, + "grad_norm": 2.1920876502990723, + "learning_rate": 3.934362659773799e-05, + "loss": 0.3395, + "num_input_tokens_seen": 3819816, + "step": 14305 + }, + { + "epoch": 3.750982961992136, + "grad_norm": 0.9876725673675537, + "learning_rate": 3.933425757472915e-05, + "loss": 0.6937, + "num_input_tokens_seen": 3821320, + "step": 14310 + }, + { + "epoch": 3.7522935779816513, + "grad_norm": 3.779960870742798, + "learning_rate": 3.932488555158497e-05, + "loss": 0.4174, + "num_input_tokens_seen": 3822568, + "step": 14315 + }, + { + "epoch": 3.7536041939711664, + "grad_norm": 3.181766986846924, + "learning_rate": 3.931551053026699e-05, + "loss": 0.5215, + "num_input_tokens_seen": 3823912, + "step": 14320 + }, + { + "epoch": 3.7549148099606815, + "grad_norm": 2.0033302307128906, + "learning_rate": 3.9306132512737395e-05, + "loss": 0.3761, + "num_input_tokens_seen": 3825752, + "step": 14325 + }, + { + "epoch": 3.7562254259501966, + "grad_norm": 6.257669448852539, + "learning_rate": 3.929675150095898e-05, + "loss": 0.5173, + "num_input_tokens_seen": 3826744, + "step": 14330 + }, + { + "epoch": 3.7575360419397117, + "grad_norm": 2.698423147201538, + "learning_rate": 3.928736749689519e-05, + "loss": 0.2858, + "num_input_tokens_seen": 3827928, + "step": 14335 + }, + { + "epoch": 3.758846657929227, + "grad_norm": 3.5550239086151123, + "learning_rate": 3.927798050251006e-05, + "loss": 0.4526, + "num_input_tokens_seen": 3829352, + "step": 14340 + }, + { + "epoch": 3.760157273918742, + "grad_norm": 1.4032396078109741, + "learning_rate": 3.926859051976828e-05, + "loss": 0.4181, + "num_input_tokens_seen": 3830424, + "step": 14345 + }, + { + "epoch": 3.761467889908257, + "grad_norm": 1.099145531654358, + "learning_rate": 3.9259197550635164e-05, + "loss": 0.6185, + "num_input_tokens_seen": 3832072, + "step": 14350 + }, + { + "epoch": 3.762778505897772, + "grad_norm": 1.4192395210266113, + "learning_rate": 3.924980159707664e-05, + "loss": 0.4237, + "num_input_tokens_seen": 3834424, + "step": 14355 + }, + { + "epoch": 3.764089121887287, + "grad_norm": 1.5488269329071045, + "learning_rate": 3.924040266105926e-05, + "loss": 0.6098, + "num_input_tokens_seen": 3835928, + "step": 14360 + }, + { + "epoch": 3.765399737876802, + "grad_norm": 2.0519392490386963, + "learning_rate": 3.9231000744550205e-05, + "loss": 0.7748, + "num_input_tokens_seen": 3837064, + "step": 14365 + }, + { + "epoch": 3.7667103538663174, + "grad_norm": 2.8242743015289307, + "learning_rate": 3.922159584951729e-05, + "loss": 0.5393, + "num_input_tokens_seen": 3838248, + "step": 14370 + }, + { + "epoch": 3.768020969855832, + "grad_norm": 1.407334804534912, + "learning_rate": 3.921218797792893e-05, + "loss": 0.383, + "num_input_tokens_seen": 3839448, + "step": 14375 + }, + { + "epoch": 3.7693315858453476, + "grad_norm": 2.4042797088623047, + "learning_rate": 3.9202777131754187e-05, + "loss": 0.4501, + "num_input_tokens_seen": 3841032, + "step": 14380 + }, + { + "epoch": 3.770642201834862, + "grad_norm": 10.006510734558105, + "learning_rate": 3.9193363312962725e-05, + "loss": 0.3805, + "num_input_tokens_seen": 3842168, + "step": 14385 + }, + { + "epoch": 3.7719528178243773, + "grad_norm": 1.8870376348495483, + "learning_rate": 3.9183946523524856e-05, + "loss": 0.3958, + "num_input_tokens_seen": 3843800, + "step": 14390 + }, + { + "epoch": 3.7732634338138924, + "grad_norm": 2.562368154525757, + "learning_rate": 3.917452676541148e-05, + "loss": 0.4062, + "num_input_tokens_seen": 3845048, + "step": 14395 + }, + { + "epoch": 3.7745740498034075, + "grad_norm": 3.965449810028076, + "learning_rate": 3.916510404059415e-05, + "loss": 0.4473, + "num_input_tokens_seen": 3846744, + "step": 14400 + }, + { + "epoch": 3.7758846657929226, + "grad_norm": 1.1760503053665161, + "learning_rate": 3.9155678351045014e-05, + "loss": 0.5834, + "num_input_tokens_seen": 3848808, + "step": 14405 + }, + { + "epoch": 3.7771952817824377, + "grad_norm": 1.5886043310165405, + "learning_rate": 3.914624969873686e-05, + "loss": 0.433, + "num_input_tokens_seen": 3850040, + "step": 14410 + }, + { + "epoch": 3.778505897771953, + "grad_norm": 1.6368463039398193, + "learning_rate": 3.913681808564309e-05, + "loss": 0.6166, + "num_input_tokens_seen": 3851160, + "step": 14415 + }, + { + "epoch": 3.779816513761468, + "grad_norm": 2.9934613704681396, + "learning_rate": 3.912738351373772e-05, + "loss": 0.7362, + "num_input_tokens_seen": 3852360, + "step": 14420 + }, + { + "epoch": 3.781127129750983, + "grad_norm": 1.6012389659881592, + "learning_rate": 3.911794598499539e-05, + "loss": 0.4791, + "num_input_tokens_seen": 3853992, + "step": 14425 + }, + { + "epoch": 3.782437745740498, + "grad_norm": 3.0935115814208984, + "learning_rate": 3.9108505501391355e-05, + "loss": 0.4823, + "num_input_tokens_seen": 3855544, + "step": 14430 + }, + { + "epoch": 3.783748361730013, + "grad_norm": 2.07999587059021, + "learning_rate": 3.9099062064901497e-05, + "loss": 0.492, + "num_input_tokens_seen": 3856760, + "step": 14435 + }, + { + "epoch": 3.7850589777195283, + "grad_norm": 3.2491254806518555, + "learning_rate": 3.908961567750231e-05, + "loss": 0.5959, + "num_input_tokens_seen": 3857928, + "step": 14440 + }, + { + "epoch": 3.7863695937090434, + "grad_norm": 3.7545487880706787, + "learning_rate": 3.908016634117092e-05, + "loss": 0.9103, + "num_input_tokens_seen": 3859256, + "step": 14445 + }, + { + "epoch": 3.7876802096985585, + "grad_norm": 43.01646423339844, + "learning_rate": 3.907071405788503e-05, + "loss": 0.4025, + "num_input_tokens_seen": 3860328, + "step": 14450 + }, + { + "epoch": 3.7889908256880735, + "grad_norm": 1.7759708166122437, + "learning_rate": 3.9061258829622995e-05, + "loss": 0.4176, + "num_input_tokens_seen": 3861784, + "step": 14455 + }, + { + "epoch": 3.790301441677588, + "grad_norm": 1.0899381637573242, + "learning_rate": 3.905180065836379e-05, + "loss": 0.4404, + "num_input_tokens_seen": 3863256, + "step": 14460 + }, + { + "epoch": 3.7916120576671037, + "grad_norm": 1.733238697052002, + "learning_rate": 3.904233954608699e-05, + "loss": 0.3616, + "num_input_tokens_seen": 3864504, + "step": 14465 + }, + { + "epoch": 3.7929226736566184, + "grad_norm": 2.5000007152557373, + "learning_rate": 3.903287549477279e-05, + "loss": 0.3969, + "num_input_tokens_seen": 3866456, + "step": 14470 + }, + { + "epoch": 3.794233289646134, + "grad_norm": 31.248905181884766, + "learning_rate": 3.9023408506401987e-05, + "loss": 0.4636, + "num_input_tokens_seen": 3867480, + "step": 14475 + }, + { + "epoch": 3.7955439056356486, + "grad_norm": 3.264150381088257, + "learning_rate": 3.901393858295602e-05, + "loss": 0.3789, + "num_input_tokens_seen": 3869112, + "step": 14480 + }, + { + "epoch": 3.7968545216251637, + "grad_norm": 3.0803568363189697, + "learning_rate": 3.900446572641692e-05, + "loss": 1.3147, + "num_input_tokens_seen": 3870296, + "step": 14485 + }, + { + "epoch": 3.7981651376146788, + "grad_norm": 4.480445861816406, + "learning_rate": 3.899498993876733e-05, + "loss": 0.4911, + "num_input_tokens_seen": 3871256, + "step": 14490 + }, + { + "epoch": 3.799475753604194, + "grad_norm": 5.673820972442627, + "learning_rate": 3.898551122199054e-05, + "loss": 0.4338, + "num_input_tokens_seen": 3872360, + "step": 14495 + }, + { + "epoch": 3.800786369593709, + "grad_norm": 3.279346227645874, + "learning_rate": 3.8976029578070415e-05, + "loss": 0.4586, + "num_input_tokens_seen": 3873896, + "step": 14500 + }, + { + "epoch": 3.802096985583224, + "grad_norm": 3.2328579425811768, + "learning_rate": 3.896654500899145e-05, + "loss": 0.3502, + "num_input_tokens_seen": 3875352, + "step": 14505 + }, + { + "epoch": 3.803407601572739, + "grad_norm": 2.490330219268799, + "learning_rate": 3.895705751673874e-05, + "loss": 0.4227, + "num_input_tokens_seen": 3876552, + "step": 14510 + }, + { + "epoch": 3.8047182175622543, + "grad_norm": 2.8207788467407227, + "learning_rate": 3.894756710329801e-05, + "loss": 0.7348, + "num_input_tokens_seen": 3877976, + "step": 14515 + }, + { + "epoch": 3.8060288335517694, + "grad_norm": 1.6658613681793213, + "learning_rate": 3.893807377065559e-05, + "loss": 0.4804, + "num_input_tokens_seen": 3879544, + "step": 14520 + }, + { + "epoch": 3.8073394495412844, + "grad_norm": 6.764564514160156, + "learning_rate": 3.8928577520798407e-05, + "loss": 0.4923, + "num_input_tokens_seen": 3880808, + "step": 14525 + }, + { + "epoch": 3.8086500655307995, + "grad_norm": 2.174452543258667, + "learning_rate": 3.8919078355714025e-05, + "loss": 0.5267, + "num_input_tokens_seen": 3881832, + "step": 14530 + }, + { + "epoch": 3.8099606815203146, + "grad_norm": 1.6234499216079712, + "learning_rate": 3.890957627739058e-05, + "loss": 0.3415, + "num_input_tokens_seen": 3882968, + "step": 14535 + }, + { + "epoch": 3.8112712975098297, + "grad_norm": 3.5793025493621826, + "learning_rate": 3.890007128781686e-05, + "loss": 0.6017, + "num_input_tokens_seen": 3884408, + "step": 14540 + }, + { + "epoch": 3.812581913499345, + "grad_norm": 2.2047410011291504, + "learning_rate": 3.889056338898224e-05, + "loss": 0.3653, + "num_input_tokens_seen": 3885816, + "step": 14545 + }, + { + "epoch": 3.81389252948886, + "grad_norm": 2.1666297912597656, + "learning_rate": 3.8881052582876695e-05, + "loss": 0.442, + "num_input_tokens_seen": 3887672, + "step": 14550 + }, + { + "epoch": 3.8152031454783746, + "grad_norm": 1.7982780933380127, + "learning_rate": 3.887153887149084e-05, + "loss": 0.4661, + "num_input_tokens_seen": 3888936, + "step": 14555 + }, + { + "epoch": 3.81651376146789, + "grad_norm": 2.1546738147735596, + "learning_rate": 3.8862022256815865e-05, + "loss": 0.2923, + "num_input_tokens_seen": 3890296, + "step": 14560 + }, + { + "epoch": 3.8178243774574048, + "grad_norm": 5.0605387687683105, + "learning_rate": 3.885250274084358e-05, + "loss": 0.5567, + "num_input_tokens_seen": 3891352, + "step": 14565 + }, + { + "epoch": 3.8191349934469203, + "grad_norm": 7.3904128074646, + "learning_rate": 3.88429803255664e-05, + "loss": 0.443, + "num_input_tokens_seen": 3892472, + "step": 14570 + }, + { + "epoch": 3.820445609436435, + "grad_norm": 1.3599354028701782, + "learning_rate": 3.883345501297737e-05, + "loss": 0.4444, + "num_input_tokens_seen": 3894168, + "step": 14575 + }, + { + "epoch": 3.82175622542595, + "grad_norm": 39.873016357421875, + "learning_rate": 3.88239268050701e-05, + "loss": 0.7898, + "num_input_tokens_seen": 3895112, + "step": 14580 + }, + { + "epoch": 3.823066841415465, + "grad_norm": 3.3452610969543457, + "learning_rate": 3.881439570383884e-05, + "loss": 0.5801, + "num_input_tokens_seen": 3896488, + "step": 14585 + }, + { + "epoch": 3.8243774574049803, + "grad_norm": 2.2320356369018555, + "learning_rate": 3.8804861711278426e-05, + "loss": 0.4687, + "num_input_tokens_seen": 3898088, + "step": 14590 + }, + { + "epoch": 3.8256880733944953, + "grad_norm": 3.080181121826172, + "learning_rate": 3.879532482938431e-05, + "loss": 0.4305, + "num_input_tokens_seen": 3899256, + "step": 14595 + }, + { + "epoch": 3.8269986893840104, + "grad_norm": 6.734894752502441, + "learning_rate": 3.878578506015254e-05, + "loss": 0.3454, + "num_input_tokens_seen": 3900392, + "step": 14600 + }, + { + "epoch": 3.8283093053735255, + "grad_norm": 13.527750015258789, + "learning_rate": 3.877624240557978e-05, + "loss": 0.4847, + "num_input_tokens_seen": 3901976, + "step": 14605 + }, + { + "epoch": 3.8296199213630406, + "grad_norm": 2.4553375244140625, + "learning_rate": 3.876669686766329e-05, + "loss": 0.4757, + "num_input_tokens_seen": 3903272, + "step": 14610 + }, + { + "epoch": 3.8309305373525557, + "grad_norm": 1.783753514289856, + "learning_rate": 3.875714844840093e-05, + "loss": 0.6759, + "num_input_tokens_seen": 3904296, + "step": 14615 + }, + { + "epoch": 3.832241153342071, + "grad_norm": 2.033705472946167, + "learning_rate": 3.874759714979117e-05, + "loss": 0.4885, + "num_input_tokens_seen": 3905944, + "step": 14620 + }, + { + "epoch": 3.833551769331586, + "grad_norm": 9.079862594604492, + "learning_rate": 3.873804297383308e-05, + "loss": 0.3626, + "num_input_tokens_seen": 3907192, + "step": 14625 + }, + { + "epoch": 3.834862385321101, + "grad_norm": 6.866241931915283, + "learning_rate": 3.872848592252634e-05, + "loss": 0.4271, + "num_input_tokens_seen": 3908584, + "step": 14630 + }, + { + "epoch": 3.836173001310616, + "grad_norm": 2.8191475868225098, + "learning_rate": 3.871892599787121e-05, + "loss": 0.431, + "num_input_tokens_seen": 3909832, + "step": 14635 + }, + { + "epoch": 3.837483617300131, + "grad_norm": 3.6427812576293945, + "learning_rate": 3.8709363201868576e-05, + "loss": 0.4933, + "num_input_tokens_seen": 3911224, + "step": 14640 + }, + { + "epoch": 3.8387942332896463, + "grad_norm": 2.3370773792266846, + "learning_rate": 3.869979753651991e-05, + "loss": 0.344, + "num_input_tokens_seen": 3912232, + "step": 14645 + }, + { + "epoch": 3.840104849279161, + "grad_norm": 3.341705322265625, + "learning_rate": 3.869022900382729e-05, + "loss": 0.6055, + "num_input_tokens_seen": 3913288, + "step": 14650 + }, + { + "epoch": 3.8414154652686765, + "grad_norm": 1.0885815620422363, + "learning_rate": 3.868065760579339e-05, + "loss": 0.3403, + "num_input_tokens_seen": 3915080, + "step": 14655 + }, + { + "epoch": 3.842726081258191, + "grad_norm": 1.6909152269363403, + "learning_rate": 3.86710833444215e-05, + "loss": 0.4003, + "num_input_tokens_seen": 3916472, + "step": 14660 + }, + { + "epoch": 3.8440366972477067, + "grad_norm": 3.0159010887145996, + "learning_rate": 3.866150622171549e-05, + "loss": 0.3841, + "num_input_tokens_seen": 3917688, + "step": 14665 + }, + { + "epoch": 3.8453473132372213, + "grad_norm": 1.2041049003601074, + "learning_rate": 3.8651926239679824e-05, + "loss": 0.4868, + "num_input_tokens_seen": 3919256, + "step": 14670 + }, + { + "epoch": 3.8466579292267364, + "grad_norm": 1.4596984386444092, + "learning_rate": 3.8642343400319594e-05, + "loss": 0.3674, + "num_input_tokens_seen": 3920632, + "step": 14675 + }, + { + "epoch": 3.8479685452162515, + "grad_norm": 4.956817150115967, + "learning_rate": 3.863275770564046e-05, + "loss": 0.2918, + "num_input_tokens_seen": 3922120, + "step": 14680 + }, + { + "epoch": 3.8492791612057666, + "grad_norm": 2.225128412246704, + "learning_rate": 3.862316915764869e-05, + "loss": 0.4883, + "num_input_tokens_seen": 3923608, + "step": 14685 + }, + { + "epoch": 3.8505897771952817, + "grad_norm": 3.1850082874298096, + "learning_rate": 3.861357775835115e-05, + "loss": 0.5943, + "num_input_tokens_seen": 3924776, + "step": 14690 + }, + { + "epoch": 3.851900393184797, + "grad_norm": 1.56737220287323, + "learning_rate": 3.8603983509755316e-05, + "loss": 0.3931, + "num_input_tokens_seen": 3925992, + "step": 14695 + }, + { + "epoch": 3.853211009174312, + "grad_norm": 2.071767568588257, + "learning_rate": 3.8594386413869235e-05, + "loss": 0.3046, + "num_input_tokens_seen": 3927080, + "step": 14700 + }, + { + "epoch": 3.854521625163827, + "grad_norm": 1.2753599882125854, + "learning_rate": 3.8584786472701575e-05, + "loss": 0.3649, + "num_input_tokens_seen": 3928472, + "step": 14705 + }, + { + "epoch": 3.855832241153342, + "grad_norm": 1.6901326179504395, + "learning_rate": 3.857518368826157e-05, + "loss": 0.4541, + "num_input_tokens_seen": 3929656, + "step": 14710 + }, + { + "epoch": 3.857142857142857, + "grad_norm": 1.5120787620544434, + "learning_rate": 3.856557806255908e-05, + "loss": 0.367, + "num_input_tokens_seen": 3931096, + "step": 14715 + }, + { + "epoch": 3.8584534731323723, + "grad_norm": 1.9224971532821655, + "learning_rate": 3.8555969597604536e-05, + "loss": 0.2887, + "num_input_tokens_seen": 3932856, + "step": 14720 + }, + { + "epoch": 3.8597640891218874, + "grad_norm": 1.9600108861923218, + "learning_rate": 3.854635829540898e-05, + "loss": 0.4091, + "num_input_tokens_seen": 3934264, + "step": 14725 + }, + { + "epoch": 3.8610747051114025, + "grad_norm": 3.217226505279541, + "learning_rate": 3.853674415798404e-05, + "loss": 0.4488, + "num_input_tokens_seen": 3935512, + "step": 14730 + }, + { + "epoch": 3.8623853211009176, + "grad_norm": 2.325516939163208, + "learning_rate": 3.8527127187341936e-05, + "loss": 0.379, + "num_input_tokens_seen": 3937000, + "step": 14735 + }, + { + "epoch": 3.8636959370904327, + "grad_norm": 2.6497137546539307, + "learning_rate": 3.8517507385495486e-05, + "loss": 0.6886, + "num_input_tokens_seen": 3938696, + "step": 14740 + }, + { + "epoch": 3.8650065530799473, + "grad_norm": 3.002516269683838, + "learning_rate": 3.85078847544581e-05, + "loss": 0.6239, + "num_input_tokens_seen": 3939960, + "step": 14745 + }, + { + "epoch": 3.866317169069463, + "grad_norm": 2.3647286891937256, + "learning_rate": 3.849825929624377e-05, + "loss": 0.8633, + "num_input_tokens_seen": 3941160, + "step": 14750 + }, + { + "epoch": 3.8676277850589775, + "grad_norm": 2.5714242458343506, + "learning_rate": 3.8488631012867095e-05, + "loss": 0.4481, + "num_input_tokens_seen": 3942200, + "step": 14755 + }, + { + "epoch": 3.8689384010484926, + "grad_norm": 3.4588100910186768, + "learning_rate": 3.847899990634326e-05, + "loss": 0.4537, + "num_input_tokens_seen": 3943480, + "step": 14760 + }, + { + "epoch": 3.8702490170380077, + "grad_norm": 1.7194007635116577, + "learning_rate": 3.846936597868802e-05, + "loss": 0.4775, + "num_input_tokens_seen": 3944632, + "step": 14765 + }, + { + "epoch": 3.871559633027523, + "grad_norm": 3.6419944763183594, + "learning_rate": 3.845972923191776e-05, + "loss": 0.3574, + "num_input_tokens_seen": 3945960, + "step": 14770 + }, + { + "epoch": 3.872870249017038, + "grad_norm": 1.6626487970352173, + "learning_rate": 3.845008966804944e-05, + "loss": 0.4182, + "num_input_tokens_seen": 3947528, + "step": 14775 + }, + { + "epoch": 3.874180865006553, + "grad_norm": 3.15082049369812, + "learning_rate": 3.844044728910058e-05, + "loss": 0.4687, + "num_input_tokens_seen": 3948792, + "step": 14780 + }, + { + "epoch": 3.875491480996068, + "grad_norm": 5.845637321472168, + "learning_rate": 3.843080209708933e-05, + "loss": 0.5409, + "num_input_tokens_seen": 3950200, + "step": 14785 + }, + { + "epoch": 3.876802096985583, + "grad_norm": 2.3210365772247314, + "learning_rate": 3.8421154094034404e-05, + "loss": 0.4631, + "num_input_tokens_seen": 3951288, + "step": 14790 + }, + { + "epoch": 3.8781127129750983, + "grad_norm": 1.8460580110549927, + "learning_rate": 3.841150328195512e-05, + "loss": 0.4299, + "num_input_tokens_seen": 3952264, + "step": 14795 + }, + { + "epoch": 3.8794233289646134, + "grad_norm": 2.114474058151245, + "learning_rate": 3.840184966287137e-05, + "loss": 0.6892, + "num_input_tokens_seen": 3953768, + "step": 14800 + }, + { + "epoch": 3.8807339449541285, + "grad_norm": 1.3647130727767944, + "learning_rate": 3.839219323880365e-05, + "loss": 0.6565, + "num_input_tokens_seen": 3955128, + "step": 14805 + }, + { + "epoch": 3.8820445609436436, + "grad_norm": 1.7195045948028564, + "learning_rate": 3.838253401177302e-05, + "loss": 0.3283, + "num_input_tokens_seen": 3956456, + "step": 14810 + }, + { + "epoch": 3.8833551769331587, + "grad_norm": 2.570110559463501, + "learning_rate": 3.837287198380114e-05, + "loss": 0.5253, + "num_input_tokens_seen": 3957720, + "step": 14815 + }, + { + "epoch": 3.8846657929226738, + "grad_norm": 5.993634223937988, + "learning_rate": 3.836320715691027e-05, + "loss": 0.5266, + "num_input_tokens_seen": 3958952, + "step": 14820 + }, + { + "epoch": 3.885976408912189, + "grad_norm": 1.0756607055664062, + "learning_rate": 3.835353953312322e-05, + "loss": 0.3837, + "num_input_tokens_seen": 3960632, + "step": 14825 + }, + { + "epoch": 3.8872870249017035, + "grad_norm": 1.1352704763412476, + "learning_rate": 3.8343869114463424e-05, + "loss": 0.2903, + "num_input_tokens_seen": 3962008, + "step": 14830 + }, + { + "epoch": 3.888597640891219, + "grad_norm": 1.9956796169281006, + "learning_rate": 3.8334195902954885e-05, + "loss": 0.5813, + "num_input_tokens_seen": 3963592, + "step": 14835 + }, + { + "epoch": 3.8899082568807337, + "grad_norm": 3.0366525650024414, + "learning_rate": 3.8324519900622175e-05, + "loss": 0.3578, + "num_input_tokens_seen": 3965064, + "step": 14840 + }, + { + "epoch": 3.8912188728702493, + "grad_norm": 1.948647141456604, + "learning_rate": 3.8314841109490474e-05, + "loss": 0.4115, + "num_input_tokens_seen": 3966712, + "step": 14845 + }, + { + "epoch": 3.892529488859764, + "grad_norm": 1.4991111755371094, + "learning_rate": 3.830515953158553e-05, + "loss": 0.5817, + "num_input_tokens_seen": 3967880, + "step": 14850 + }, + { + "epoch": 3.893840104849279, + "grad_norm": 3.6995856761932373, + "learning_rate": 3.82954751689337e-05, + "loss": 0.4704, + "num_input_tokens_seen": 3969384, + "step": 14855 + }, + { + "epoch": 3.895150720838794, + "grad_norm": 7.330676555633545, + "learning_rate": 3.828578802356188e-05, + "loss": 0.4865, + "num_input_tokens_seen": 3972536, + "step": 14860 + }, + { + "epoch": 3.896461336828309, + "grad_norm": 2.555142641067505, + "learning_rate": 3.8276098097497584e-05, + "loss": 0.6213, + "num_input_tokens_seen": 3974008, + "step": 14865 + }, + { + "epoch": 3.8977719528178243, + "grad_norm": 2.2730395793914795, + "learning_rate": 3.8266405392768904e-05, + "loss": 0.5588, + "num_input_tokens_seen": 3975592, + "step": 14870 + }, + { + "epoch": 3.8990825688073394, + "grad_norm": 2.1858084201812744, + "learning_rate": 3.8256709911404484e-05, + "loss": 0.6087, + "num_input_tokens_seen": 3976712, + "step": 14875 + }, + { + "epoch": 3.9003931847968545, + "grad_norm": 2.590245485305786, + "learning_rate": 3.82470116554336e-05, + "loss": 0.3941, + "num_input_tokens_seen": 3977992, + "step": 14880 + }, + { + "epoch": 3.9017038007863696, + "grad_norm": 5.436695098876953, + "learning_rate": 3.823731062688605e-05, + "loss": 0.332, + "num_input_tokens_seen": 3979240, + "step": 14885 + }, + { + "epoch": 3.9030144167758847, + "grad_norm": 3.0820112228393555, + "learning_rate": 3.8227606827792265e-05, + "loss": 0.4686, + "num_input_tokens_seen": 3980712, + "step": 14890 + }, + { + "epoch": 3.9043250327653998, + "grad_norm": 2.473599672317505, + "learning_rate": 3.8217900260183224e-05, + "loss": 0.5552, + "num_input_tokens_seen": 3981816, + "step": 14895 + }, + { + "epoch": 3.905635648754915, + "grad_norm": 4.4141693115234375, + "learning_rate": 3.82081909260905e-05, + "loss": 0.5249, + "num_input_tokens_seen": 3983096, + "step": 14900 + }, + { + "epoch": 3.90694626474443, + "grad_norm": 2.201582670211792, + "learning_rate": 3.819847882754623e-05, + "loss": 0.4286, + "num_input_tokens_seen": 3984424, + "step": 14905 + }, + { + "epoch": 3.908256880733945, + "grad_norm": 4.637107849121094, + "learning_rate": 3.8188763966583165e-05, + "loss": 0.5084, + "num_input_tokens_seen": 3985928, + "step": 14910 + }, + { + "epoch": 3.90956749672346, + "grad_norm": 1.875683307647705, + "learning_rate": 3.817904634523458e-05, + "loss": 0.4982, + "num_input_tokens_seen": 3987240, + "step": 14915 + }, + { + "epoch": 3.9108781127129753, + "grad_norm": 3.2756028175354004, + "learning_rate": 3.816932596553436e-05, + "loss": 0.4623, + "num_input_tokens_seen": 3988488, + "step": 14920 + }, + { + "epoch": 3.91218872870249, + "grad_norm": 2.0011444091796875, + "learning_rate": 3.8159602829516975e-05, + "loss": 0.534, + "num_input_tokens_seen": 3990072, + "step": 14925 + }, + { + "epoch": 3.9134993446920054, + "grad_norm": 1.6455705165863037, + "learning_rate": 3.8149876939217454e-05, + "loss": 0.5764, + "num_input_tokens_seen": 3991656, + "step": 14930 + }, + { + "epoch": 3.91480996068152, + "grad_norm": 1.6178518533706665, + "learning_rate": 3.814014829667142e-05, + "loss": 0.5178, + "num_input_tokens_seen": 3993480, + "step": 14935 + }, + { + "epoch": 3.9161205766710356, + "grad_norm": 3.194270610809326, + "learning_rate": 3.8130416903915045e-05, + "loss": 0.5014, + "num_input_tokens_seen": 3994504, + "step": 14940 + }, + { + "epoch": 3.9174311926605503, + "grad_norm": 2.7118728160858154, + "learning_rate": 3.812068276298509e-05, + "loss": 0.7459, + "num_input_tokens_seen": 3996152, + "step": 14945 + }, + { + "epoch": 3.9187418086500654, + "grad_norm": 8.479361534118652, + "learning_rate": 3.811094587591892e-05, + "loss": 0.5686, + "num_input_tokens_seen": 3997576, + "step": 14950 + }, + { + "epoch": 3.9200524246395805, + "grad_norm": 4.244515419006348, + "learning_rate": 3.810120624475443e-05, + "loss": 0.6343, + "num_input_tokens_seen": 3998616, + "step": 14955 + }, + { + "epoch": 3.9213630406290956, + "grad_norm": 2.431015968322754, + "learning_rate": 3.809146387153011e-05, + "loss": 0.4385, + "num_input_tokens_seen": 4000200, + "step": 14960 + }, + { + "epoch": 3.9226736566186107, + "grad_norm": 3.2173304557800293, + "learning_rate": 3.808171875828501e-05, + "loss": 0.5235, + "num_input_tokens_seen": 4001160, + "step": 14965 + }, + { + "epoch": 3.9239842726081258, + "grad_norm": 1.6746269464492798, + "learning_rate": 3.8071970907058786e-05, + "loss": 0.4111, + "num_input_tokens_seen": 4002648, + "step": 14970 + }, + { + "epoch": 3.925294888597641, + "grad_norm": 1.9016671180725098, + "learning_rate": 3.806222031989164e-05, + "loss": 0.5144, + "num_input_tokens_seen": 4004328, + "step": 14975 + }, + { + "epoch": 3.926605504587156, + "grad_norm": 1.5896390676498413, + "learning_rate": 3.8052466998824344e-05, + "loss": 0.4827, + "num_input_tokens_seen": 4005512, + "step": 14980 + }, + { + "epoch": 3.927916120576671, + "grad_norm": 1.471923589706421, + "learning_rate": 3.804271094589826e-05, + "loss": 0.3941, + "num_input_tokens_seen": 4006728, + "step": 14985 + }, + { + "epoch": 3.929226736566186, + "grad_norm": 3.4368817806243896, + "learning_rate": 3.803295216315532e-05, + "loss": 0.7038, + "num_input_tokens_seen": 4007672, + "step": 14990 + }, + { + "epoch": 3.9305373525557012, + "grad_norm": 3.4997801780700684, + "learning_rate": 3.802319065263801e-05, + "loss": 0.5818, + "num_input_tokens_seen": 4009048, + "step": 14995 + }, + { + "epoch": 3.9318479685452163, + "grad_norm": 4.41640567779541, + "learning_rate": 3.8013426416389385e-05, + "loss": 0.4103, + "num_input_tokens_seen": 4009976, + "step": 15000 + }, + { + "epoch": 3.9331585845347314, + "grad_norm": 2.8841545581817627, + "learning_rate": 3.80036594564531e-05, + "loss": 0.5064, + "num_input_tokens_seen": 4011000, + "step": 15005 + }, + { + "epoch": 3.9344692005242465, + "grad_norm": 3.0345921516418457, + "learning_rate": 3.799388977487337e-05, + "loss": 0.5554, + "num_input_tokens_seen": 4012024, + "step": 15010 + }, + { + "epoch": 3.9357798165137616, + "grad_norm": 1.281989336013794, + "learning_rate": 3.798411737369495e-05, + "loss": 0.4897, + "num_input_tokens_seen": 4013320, + "step": 15015 + }, + { + "epoch": 3.9370904325032763, + "grad_norm": 2.65049409866333, + "learning_rate": 3.7974342254963214e-05, + "loss": 0.5413, + "num_input_tokens_seen": 4014600, + "step": 15020 + }, + { + "epoch": 3.938401048492792, + "grad_norm": 2.627937078475952, + "learning_rate": 3.7964564420724044e-05, + "loss": 0.5872, + "num_input_tokens_seen": 4016104, + "step": 15025 + }, + { + "epoch": 3.9397116644823065, + "grad_norm": 6.630642890930176, + "learning_rate": 3.7954783873023955e-05, + "loss": 0.5351, + "num_input_tokens_seen": 4017480, + "step": 15030 + }, + { + "epoch": 3.941022280471822, + "grad_norm": 1.2103965282440186, + "learning_rate": 3.7945000613909975e-05, + "loss": 0.5215, + "num_input_tokens_seen": 4019096, + "step": 15035 + }, + { + "epoch": 3.9423328964613367, + "grad_norm": 2.0134778022766113, + "learning_rate": 3.793521464542974e-05, + "loss": 1.1085, + "num_input_tokens_seen": 4020280, + "step": 15040 + }, + { + "epoch": 3.9436435124508518, + "grad_norm": 2.719555139541626, + "learning_rate": 3.7925425969631435e-05, + "loss": 0.725, + "num_input_tokens_seen": 4021656, + "step": 15045 + }, + { + "epoch": 3.944954128440367, + "grad_norm": 2.0568394660949707, + "learning_rate": 3.7915634588563804e-05, + "loss": 0.367, + "num_input_tokens_seen": 4023032, + "step": 15050 + }, + { + "epoch": 3.946264744429882, + "grad_norm": 2.3581085205078125, + "learning_rate": 3.790584050427616e-05, + "loss": 0.53, + "num_input_tokens_seen": 4024472, + "step": 15055 + }, + { + "epoch": 3.947575360419397, + "grad_norm": 7.581479072570801, + "learning_rate": 3.7896043718818406e-05, + "loss": 0.4214, + "num_input_tokens_seen": 4025592, + "step": 15060 + }, + { + "epoch": 3.948885976408912, + "grad_norm": 2.5242695808410645, + "learning_rate": 3.788624423424099e-05, + "loss": 0.4232, + "num_input_tokens_seen": 4026648, + "step": 15065 + }, + { + "epoch": 3.9501965923984272, + "grad_norm": 3.6118743419647217, + "learning_rate": 3.7876442052594915e-05, + "loss": 0.4865, + "num_input_tokens_seen": 4028024, + "step": 15070 + }, + { + "epoch": 3.9515072083879423, + "grad_norm": 1.3180865049362183, + "learning_rate": 3.7866637175931774e-05, + "loss": 0.288, + "num_input_tokens_seen": 4029272, + "step": 15075 + }, + { + "epoch": 3.9528178243774574, + "grad_norm": 4.394166469573975, + "learning_rate": 3.7856829606303696e-05, + "loss": 0.4239, + "num_input_tokens_seen": 4031096, + "step": 15080 + }, + { + "epoch": 3.9541284403669725, + "grad_norm": 2.513155221939087, + "learning_rate": 3.784701934576339e-05, + "loss": 0.428, + "num_input_tokens_seen": 4032744, + "step": 15085 + }, + { + "epoch": 3.9554390563564876, + "grad_norm": 3.6229686737060547, + "learning_rate": 3.783720639636415e-05, + "loss": 0.4102, + "num_input_tokens_seen": 4034392, + "step": 15090 + }, + { + "epoch": 3.9567496723460027, + "grad_norm": 2.5211691856384277, + "learning_rate": 3.782739076015978e-05, + "loss": 0.38, + "num_input_tokens_seen": 4036088, + "step": 15095 + }, + { + "epoch": 3.958060288335518, + "grad_norm": 6.166409492492676, + "learning_rate": 3.7817572439204695e-05, + "loss": 0.3946, + "num_input_tokens_seen": 4037304, + "step": 15100 + }, + { + "epoch": 3.959370904325033, + "grad_norm": 3.6396615505218506, + "learning_rate": 3.780775143555384e-05, + "loss": 0.4233, + "num_input_tokens_seen": 4039096, + "step": 15105 + }, + { + "epoch": 3.960681520314548, + "grad_norm": 3.7218761444091797, + "learning_rate": 3.779792775126275e-05, + "loss": 0.623, + "num_input_tokens_seen": 4040104, + "step": 15110 + }, + { + "epoch": 3.9619921363040627, + "grad_norm": 2.3138298988342285, + "learning_rate": 3.778810138838748e-05, + "loss": 0.6378, + "num_input_tokens_seen": 4041544, + "step": 15115 + }, + { + "epoch": 3.963302752293578, + "grad_norm": 4.813689708709717, + "learning_rate": 3.777827234898469e-05, + "loss": 0.4001, + "num_input_tokens_seen": 4042600, + "step": 15120 + }, + { + "epoch": 3.964613368283093, + "grad_norm": 1.4512275457382202, + "learning_rate": 3.776844063511158e-05, + "loss": 0.6492, + "num_input_tokens_seen": 4044024, + "step": 15125 + }, + { + "epoch": 3.9659239842726084, + "grad_norm": 3.2423393726348877, + "learning_rate": 3.7758606248825914e-05, + "loss": 0.3437, + "num_input_tokens_seen": 4045736, + "step": 15130 + }, + { + "epoch": 3.967234600262123, + "grad_norm": 1.9841337203979492, + "learning_rate": 3.774876919218599e-05, + "loss": 0.4288, + "num_input_tokens_seen": 4047208, + "step": 15135 + }, + { + "epoch": 3.968545216251638, + "grad_norm": 4.038857936859131, + "learning_rate": 3.773892946725071e-05, + "loss": 0.5271, + "num_input_tokens_seen": 4048408, + "step": 15140 + }, + { + "epoch": 3.9698558322411532, + "grad_norm": 2.3324954509735107, + "learning_rate": 3.7729087076079505e-05, + "loss": 0.563, + "num_input_tokens_seen": 4049736, + "step": 15145 + }, + { + "epoch": 3.9711664482306683, + "grad_norm": 7.02081298828125, + "learning_rate": 3.771924202073236e-05, + "loss": 0.4342, + "num_input_tokens_seen": 4051032, + "step": 15150 + }, + { + "epoch": 3.9724770642201834, + "grad_norm": 1.6143847703933716, + "learning_rate": 3.770939430326985e-05, + "loss": 0.3564, + "num_input_tokens_seen": 4052424, + "step": 15155 + }, + { + "epoch": 3.9737876802096985, + "grad_norm": 1.9381102323532104, + "learning_rate": 3.7699543925753064e-05, + "loss": 0.2922, + "num_input_tokens_seen": 4053400, + "step": 15160 + }, + { + "epoch": 3.9750982961992136, + "grad_norm": 1.9858111143112183, + "learning_rate": 3.768969089024368e-05, + "loss": 0.6504, + "num_input_tokens_seen": 4054632, + "step": 15165 + }, + { + "epoch": 3.9764089121887287, + "grad_norm": 1.6514595746994019, + "learning_rate": 3.767983519880392e-05, + "loss": 0.5046, + "num_input_tokens_seen": 4056344, + "step": 15170 + }, + { + "epoch": 3.977719528178244, + "grad_norm": 2.277308702468872, + "learning_rate": 3.766997685349656e-05, + "loss": 0.4014, + "num_input_tokens_seen": 4057672, + "step": 15175 + }, + { + "epoch": 3.979030144167759, + "grad_norm": 2.2496681213378906, + "learning_rate": 3.766011585638494e-05, + "loss": 0.4518, + "num_input_tokens_seen": 4058792, + "step": 15180 + }, + { + "epoch": 3.980340760157274, + "grad_norm": 2.4180891513824463, + "learning_rate": 3.765025220953294e-05, + "loss": 0.6543, + "num_input_tokens_seen": 4060376, + "step": 15185 + }, + { + "epoch": 3.981651376146789, + "grad_norm": 2.43955659866333, + "learning_rate": 3.764038591500502e-05, + "loss": 0.3677, + "num_input_tokens_seen": 4061816, + "step": 15190 + }, + { + "epoch": 3.982961992136304, + "grad_norm": 1.0037041902542114, + "learning_rate": 3.7630516974866166e-05, + "loss": 0.4095, + "num_input_tokens_seen": 4063544, + "step": 15195 + }, + { + "epoch": 3.9842726081258193, + "grad_norm": 2.125861167907715, + "learning_rate": 3.7620645391181926e-05, + "loss": 0.1716, + "num_input_tokens_seen": 4064536, + "step": 15200 + }, + { + "epoch": 3.9855832241153344, + "grad_norm": 7.0011186599731445, + "learning_rate": 3.761077116601842e-05, + "loss": 0.6114, + "num_input_tokens_seen": 4066024, + "step": 15205 + }, + { + "epoch": 3.986893840104849, + "grad_norm": 2.578354835510254, + "learning_rate": 3.760089430144229e-05, + "loss": 0.3046, + "num_input_tokens_seen": 4066968, + "step": 15210 + }, + { + "epoch": 3.9882044560943646, + "grad_norm": 3.2422478199005127, + "learning_rate": 3.759101479952076e-05, + "loss": 0.5016, + "num_input_tokens_seen": 4068552, + "step": 15215 + }, + { + "epoch": 3.9895150720838792, + "grad_norm": 2.363964319229126, + "learning_rate": 3.7581132662321597e-05, + "loss": 0.3693, + "num_input_tokens_seen": 4069576, + "step": 15220 + }, + { + "epoch": 3.9908256880733948, + "grad_norm": 2.3254127502441406, + "learning_rate": 3.7571247891913095e-05, + "loss": 0.3734, + "num_input_tokens_seen": 4071448, + "step": 15225 + }, + { + "epoch": 3.9921363040629094, + "grad_norm": 2.131592273712158, + "learning_rate": 3.756136049036413e-05, + "loss": 0.5101, + "num_input_tokens_seen": 4072712, + "step": 15230 + }, + { + "epoch": 3.9934469200524245, + "grad_norm": 8.363170623779297, + "learning_rate": 3.7551470459744126e-05, + "loss": 0.5138, + "num_input_tokens_seen": 4074024, + "step": 15235 + }, + { + "epoch": 3.9947575360419396, + "grad_norm": 1.2068195343017578, + "learning_rate": 3.754157780212304e-05, + "loss": 0.4799, + "num_input_tokens_seen": 4075768, + "step": 15240 + }, + { + "epoch": 3.9960681520314547, + "grad_norm": 0.9891070127487183, + "learning_rate": 3.753168251957139e-05, + "loss": 0.4705, + "num_input_tokens_seen": 4077176, + "step": 15245 + }, + { + "epoch": 3.99737876802097, + "grad_norm": 10.138951301574707, + "learning_rate": 3.752178461416024e-05, + "loss": 0.5857, + "num_input_tokens_seen": 4078664, + "step": 15250 + }, + { + "epoch": 3.998689384010485, + "grad_norm": 1.7257682085037231, + "learning_rate": 3.75118840879612e-05, + "loss": 0.4029, + "num_input_tokens_seen": 4079944, + "step": 15255 + }, + { + "epoch": 4.0, + "grad_norm": 3.229863166809082, + "learning_rate": 3.750198094304644e-05, + "loss": 0.5049, + "num_input_tokens_seen": 4081024, + "step": 15260 + }, + { + "epoch": 4.001048492791612, + "eval_loss": 0.4691040515899658, + "eval_runtime": 18.2025, + "eval_samples_per_second": 46.587, + "eval_steps_per_second": 23.294, + "num_input_tokens_seen": 4082080, + "step": 15264 + }, + { + "epoch": 4.001310615989515, + "grad_norm": 1.343145489692688, + "learning_rate": 3.749207518148869e-05, + "loss": 0.5042, + "num_input_tokens_seen": 4082384, + "step": 15265 + }, + { + "epoch": 4.00262123197903, + "grad_norm": 2.088621139526367, + "learning_rate": 3.7482166805361175e-05, + "loss": 0.2973, + "num_input_tokens_seen": 4083360, + "step": 15270 + }, + { + "epoch": 4.003931847968545, + "grad_norm": 3.672483205795288, + "learning_rate": 3.747225581673771e-05, + "loss": 0.4349, + "num_input_tokens_seen": 4085120, + "step": 15275 + }, + { + "epoch": 4.00524246395806, + "grad_norm": 1.5813684463500977, + "learning_rate": 3.746234221769266e-05, + "loss": 0.445, + "num_input_tokens_seen": 4087008, + "step": 15280 + }, + { + "epoch": 4.006553079947575, + "grad_norm": 3.545092821121216, + "learning_rate": 3.7452426010300915e-05, + "loss": 0.4283, + "num_input_tokens_seen": 4088256, + "step": 15285 + }, + { + "epoch": 4.007863695937091, + "grad_norm": 1.9758586883544922, + "learning_rate": 3.7442507196637925e-05, + "loss": 0.4343, + "num_input_tokens_seen": 4089648, + "step": 15290 + }, + { + "epoch": 4.009174311926605, + "grad_norm": 1.6570073366165161, + "learning_rate": 3.743258577877968e-05, + "loss": 0.4544, + "num_input_tokens_seen": 4091680, + "step": 15295 + }, + { + "epoch": 4.010484927916121, + "grad_norm": 2.550278425216675, + "learning_rate": 3.74226617588027e-05, + "loss": 0.7913, + "num_input_tokens_seen": 4093216, + "step": 15300 + }, + { + "epoch": 4.011795543905635, + "grad_norm": 1.8612686395645142, + "learning_rate": 3.7412735138784096e-05, + "loss": 0.5387, + "num_input_tokens_seen": 4094416, + "step": 15305 + }, + { + "epoch": 4.013106159895151, + "grad_norm": 29.70441246032715, + "learning_rate": 3.740280592080147e-05, + "loss": 0.5062, + "num_input_tokens_seen": 4095984, + "step": 15310 + }, + { + "epoch": 4.014416775884666, + "grad_norm": 1.4332515001296997, + "learning_rate": 3.7392874106932985e-05, + "loss": 0.5727, + "num_input_tokens_seen": 4097440, + "step": 15315 + }, + { + "epoch": 4.015727391874181, + "grad_norm": 11.380539894104004, + "learning_rate": 3.738293969925737e-05, + "loss": 0.5761, + "num_input_tokens_seen": 4098688, + "step": 15320 + }, + { + "epoch": 4.017038007863696, + "grad_norm": 1.627843976020813, + "learning_rate": 3.737300269985388e-05, + "loss": 0.567, + "num_input_tokens_seen": 4100336, + "step": 15325 + }, + { + "epoch": 4.018348623853211, + "grad_norm": 2.0381674766540527, + "learning_rate": 3.736306311080229e-05, + "loss": 0.6303, + "num_input_tokens_seen": 4101840, + "step": 15330 + }, + { + "epoch": 4.019659239842726, + "grad_norm": 6.714572429656982, + "learning_rate": 3.7353120934182956e-05, + "loss": 0.5101, + "num_input_tokens_seen": 4102736, + "step": 15335 + }, + { + "epoch": 4.0209698558322415, + "grad_norm": 1.6465650796890259, + "learning_rate": 3.7343176172076755e-05, + "loss": 0.3893, + "num_input_tokens_seen": 4103984, + "step": 15340 + }, + { + "epoch": 4.022280471821756, + "grad_norm": 1.5043619871139526, + "learning_rate": 3.7333228826565115e-05, + "loss": 0.4273, + "num_input_tokens_seen": 4105488, + "step": 15345 + }, + { + "epoch": 4.023591087811272, + "grad_norm": 3.211212635040283, + "learning_rate": 3.7323278899729986e-05, + "loss": 0.4428, + "num_input_tokens_seen": 4106976, + "step": 15350 + }, + { + "epoch": 4.024901703800786, + "grad_norm": 5.945802211761475, + "learning_rate": 3.731332639365387e-05, + "loss": 0.3459, + "num_input_tokens_seen": 4107984, + "step": 15355 + }, + { + "epoch": 4.026212319790301, + "grad_norm": 2.6059892177581787, + "learning_rate": 3.7303371310419835e-05, + "loss": 0.3759, + "num_input_tokens_seen": 4109904, + "step": 15360 + }, + { + "epoch": 4.027522935779817, + "grad_norm": 1.455033540725708, + "learning_rate": 3.729341365211143e-05, + "loss": 0.4489, + "num_input_tokens_seen": 4111360, + "step": 15365 + }, + { + "epoch": 4.028833551769331, + "grad_norm": 7.169518947601318, + "learning_rate": 3.7283453420812786e-05, + "loss": 0.5168, + "num_input_tokens_seen": 4112624, + "step": 15370 + }, + { + "epoch": 4.030144167758847, + "grad_norm": 3.6221201419830322, + "learning_rate": 3.727349061860857e-05, + "loss": 0.4231, + "num_input_tokens_seen": 4113952, + "step": 15375 + }, + { + "epoch": 4.031454783748361, + "grad_norm": 2.9367282390594482, + "learning_rate": 3.726352524758397e-05, + "loss": 0.4536, + "num_input_tokens_seen": 4115296, + "step": 15380 + }, + { + "epoch": 4.032765399737877, + "grad_norm": 2.311264753341675, + "learning_rate": 3.725355730982474e-05, + "loss": 0.5227, + "num_input_tokens_seen": 4116624, + "step": 15385 + }, + { + "epoch": 4.034076015727392, + "grad_norm": 1.7857524156570435, + "learning_rate": 3.724358680741713e-05, + "loss": 0.5446, + "num_input_tokens_seen": 4117856, + "step": 15390 + }, + { + "epoch": 4.035386631716907, + "grad_norm": 1.9854406118392944, + "learning_rate": 3.723361374244795e-05, + "loss": 0.4312, + "num_input_tokens_seen": 4119040, + "step": 15395 + }, + { + "epoch": 4.036697247706422, + "grad_norm": 2.0599193572998047, + "learning_rate": 3.722363811700455e-05, + "loss": 0.4238, + "num_input_tokens_seen": 4120304, + "step": 15400 + }, + { + "epoch": 4.038007863695937, + "grad_norm": 1.4996411800384521, + "learning_rate": 3.721365993317482e-05, + "loss": 0.3315, + "num_input_tokens_seen": 4122048, + "step": 15405 + }, + { + "epoch": 4.039318479685452, + "grad_norm": 4.504283905029297, + "learning_rate": 3.720367919304717e-05, + "loss": 0.5094, + "num_input_tokens_seen": 4122832, + "step": 15410 + }, + { + "epoch": 4.0406290956749675, + "grad_norm": 3.2616195678710938, + "learning_rate": 3.7193695898710545e-05, + "loss": 0.7085, + "num_input_tokens_seen": 4124560, + "step": 15415 + }, + { + "epoch": 4.041939711664482, + "grad_norm": 0.8014583587646484, + "learning_rate": 3.718371005225445e-05, + "loss": 0.4134, + "num_input_tokens_seen": 4126240, + "step": 15420 + }, + { + "epoch": 4.043250327653998, + "grad_norm": 4.469857692718506, + "learning_rate": 3.717372165576888e-05, + "loss": 0.3779, + "num_input_tokens_seen": 4127328, + "step": 15425 + }, + { + "epoch": 4.044560943643512, + "grad_norm": 2.926894187927246, + "learning_rate": 3.71637307113444e-05, + "loss": 0.3567, + "num_input_tokens_seen": 4128480, + "step": 15430 + }, + { + "epoch": 4.045871559633028, + "grad_norm": 3.400387763977051, + "learning_rate": 3.715373722107211e-05, + "loss": 0.2053, + "num_input_tokens_seen": 4129712, + "step": 15435 + }, + { + "epoch": 4.047182175622543, + "grad_norm": 4.193431854248047, + "learning_rate": 3.714374118704362e-05, + "loss": 0.4311, + "num_input_tokens_seen": 4130896, + "step": 15440 + }, + { + "epoch": 4.048492791612058, + "grad_norm": 2.791665554046631, + "learning_rate": 3.713374261135107e-05, + "loss": 0.6096, + "num_input_tokens_seen": 4132448, + "step": 15445 + }, + { + "epoch": 4.049803407601573, + "grad_norm": 1.2912461757659912, + "learning_rate": 3.712374149608717e-05, + "loss": 0.4886, + "num_input_tokens_seen": 4134272, + "step": 15450 + }, + { + "epoch": 4.051114023591087, + "grad_norm": 2.2011513710021973, + "learning_rate": 3.711373784334511e-05, + "loss": 0.3896, + "num_input_tokens_seen": 4135792, + "step": 15455 + }, + { + "epoch": 4.052424639580603, + "grad_norm": 2.1477715969085693, + "learning_rate": 3.7103731655218664e-05, + "loss": 0.4366, + "num_input_tokens_seen": 4137248, + "step": 15460 + }, + { + "epoch": 4.053735255570118, + "grad_norm": 2.24470591545105, + "learning_rate": 3.7093722933802095e-05, + "loss": 0.4089, + "num_input_tokens_seen": 4138672, + "step": 15465 + }, + { + "epoch": 4.055045871559633, + "grad_norm": 1.4550317525863647, + "learning_rate": 3.70837116811902e-05, + "loss": 0.5104, + "num_input_tokens_seen": 4140512, + "step": 15470 + }, + { + "epoch": 4.056356487549148, + "grad_norm": 1.6570086479187012, + "learning_rate": 3.707369789947834e-05, + "loss": 0.4951, + "num_input_tokens_seen": 4141664, + "step": 15475 + }, + { + "epoch": 4.057667103538663, + "grad_norm": 3.085469961166382, + "learning_rate": 3.706368159076238e-05, + "loss": 0.4056, + "num_input_tokens_seen": 4143488, + "step": 15480 + }, + { + "epoch": 4.058977719528178, + "grad_norm": 3.6058623790740967, + "learning_rate": 3.7053662757138686e-05, + "loss": 0.4496, + "num_input_tokens_seen": 4145024, + "step": 15485 + }, + { + "epoch": 4.0602883355176935, + "grad_norm": 2.4534623622894287, + "learning_rate": 3.704364140070421e-05, + "loss": 0.6045, + "num_input_tokens_seen": 4146448, + "step": 15490 + }, + { + "epoch": 4.061598951507208, + "grad_norm": 9.008027076721191, + "learning_rate": 3.703361752355641e-05, + "loss": 0.5245, + "num_input_tokens_seen": 4147488, + "step": 15495 + }, + { + "epoch": 4.062909567496724, + "grad_norm": 1.0443986654281616, + "learning_rate": 3.702359112779325e-05, + "loss": 0.2356, + "num_input_tokens_seen": 4148560, + "step": 15500 + }, + { + "epoch": 4.064220183486238, + "grad_norm": 5.145569801330566, + "learning_rate": 3.701356221551324e-05, + "loss": 0.4698, + "num_input_tokens_seen": 4149600, + "step": 15505 + }, + { + "epoch": 4.065530799475754, + "grad_norm": 2.2418787479400635, + "learning_rate": 3.700353078881541e-05, + "loss": 0.4111, + "num_input_tokens_seen": 4151104, + "step": 15510 + }, + { + "epoch": 4.066841415465269, + "grad_norm": 1.8241424560546875, + "learning_rate": 3.699349684979932e-05, + "loss": 0.6809, + "num_input_tokens_seen": 4152816, + "step": 15515 + }, + { + "epoch": 4.068152031454784, + "grad_norm": 1.787584900856018, + "learning_rate": 3.6983460400565086e-05, + "loss": 0.4438, + "num_input_tokens_seen": 4154224, + "step": 15520 + }, + { + "epoch": 4.069462647444299, + "grad_norm": 2.7952516078948975, + "learning_rate": 3.697342144321329e-05, + "loss": 0.4627, + "num_input_tokens_seen": 4155776, + "step": 15525 + }, + { + "epoch": 4.070773263433814, + "grad_norm": 4.244103908538818, + "learning_rate": 3.6963379979845066e-05, + "loss": 0.4471, + "num_input_tokens_seen": 4157088, + "step": 15530 + }, + { + "epoch": 4.072083879423329, + "grad_norm": 1.9214767217636108, + "learning_rate": 3.6953336012562084e-05, + "loss": 0.5609, + "num_input_tokens_seen": 4158384, + "step": 15535 + }, + { + "epoch": 4.073394495412844, + "grad_norm": 2.5030553340911865, + "learning_rate": 3.6943289543466536e-05, + "loss": 0.483, + "num_input_tokens_seen": 4159920, + "step": 15540 + }, + { + "epoch": 4.074705111402359, + "grad_norm": 2.3187808990478516, + "learning_rate": 3.693324057466111e-05, + "loss": 0.5275, + "num_input_tokens_seen": 4161296, + "step": 15545 + }, + { + "epoch": 4.076015727391874, + "grad_norm": 1.3238924741744995, + "learning_rate": 3.692318910824906e-05, + "loss": 0.4372, + "num_input_tokens_seen": 4162688, + "step": 15550 + }, + { + "epoch": 4.077326343381389, + "grad_norm": 4.43329381942749, + "learning_rate": 3.691313514633413e-05, + "loss": 0.808, + "num_input_tokens_seen": 4163584, + "step": 15555 + }, + { + "epoch": 4.078636959370904, + "grad_norm": 3.0197932720184326, + "learning_rate": 3.69030786910206e-05, + "loss": 0.5329, + "num_input_tokens_seen": 4164688, + "step": 15560 + }, + { + "epoch": 4.0799475753604195, + "grad_norm": 2.222385883331299, + "learning_rate": 3.6893019744413265e-05, + "loss": 0.5666, + "num_input_tokens_seen": 4165840, + "step": 15565 + }, + { + "epoch": 4.081258191349934, + "grad_norm": 3.3548696041107178, + "learning_rate": 3.688295830861744e-05, + "loss": 0.7209, + "num_input_tokens_seen": 4166816, + "step": 15570 + }, + { + "epoch": 4.08256880733945, + "grad_norm": 16.19577407836914, + "learning_rate": 3.687289438573899e-05, + "loss": 0.5595, + "num_input_tokens_seen": 4168016, + "step": 15575 + }, + { + "epoch": 4.083879423328964, + "grad_norm": 2.1581358909606934, + "learning_rate": 3.6862827977884246e-05, + "loss": 0.395, + "num_input_tokens_seen": 4169376, + "step": 15580 + }, + { + "epoch": 4.08519003931848, + "grad_norm": 3.3085596561431885, + "learning_rate": 3.6852759087160105e-05, + "loss": 0.3118, + "num_input_tokens_seen": 4170944, + "step": 15585 + }, + { + "epoch": 4.086500655307995, + "grad_norm": 1.429914951324463, + "learning_rate": 3.6842687715673977e-05, + "loss": 0.4301, + "num_input_tokens_seen": 4172560, + "step": 15590 + }, + { + "epoch": 4.08781127129751, + "grad_norm": 3.2864320278167725, + "learning_rate": 3.683261386553376e-05, + "loss": 0.3343, + "num_input_tokens_seen": 4174064, + "step": 15595 + }, + { + "epoch": 4.089121887287025, + "grad_norm": 3.420828104019165, + "learning_rate": 3.6822537538847914e-05, + "loss": 0.409, + "num_input_tokens_seen": 4175344, + "step": 15600 + }, + { + "epoch": 4.09043250327654, + "grad_norm": 2.878153085708618, + "learning_rate": 3.6812458737725384e-05, + "loss": 0.5038, + "num_input_tokens_seen": 4176448, + "step": 15605 + }, + { + "epoch": 4.091743119266055, + "grad_norm": 3.676396131515503, + "learning_rate": 3.680237746427565e-05, + "loss": 0.5763, + "num_input_tokens_seen": 4177584, + "step": 15610 + }, + { + "epoch": 4.0930537352555705, + "grad_norm": 1.8787773847579956, + "learning_rate": 3.679229372060871e-05, + "loss": 0.3703, + "num_input_tokens_seen": 4178544, + "step": 15615 + }, + { + "epoch": 4.094364351245085, + "grad_norm": 1.5370283126831055, + "learning_rate": 3.678220750883507e-05, + "loss": 0.4701, + "num_input_tokens_seen": 4179680, + "step": 15620 + }, + { + "epoch": 4.095674967234601, + "grad_norm": 2.347665309906006, + "learning_rate": 3.6772118831065755e-05, + "loss": 0.3238, + "num_input_tokens_seen": 4180720, + "step": 15625 + }, + { + "epoch": 4.096985583224115, + "grad_norm": 2.547318935394287, + "learning_rate": 3.6762027689412305e-05, + "loss": 0.6113, + "num_input_tokens_seen": 4181824, + "step": 15630 + }, + { + "epoch": 4.09829619921363, + "grad_norm": 4.646096229553223, + "learning_rate": 3.675193408598679e-05, + "loss": 0.4364, + "num_input_tokens_seen": 4182848, + "step": 15635 + }, + { + "epoch": 4.0996068152031455, + "grad_norm": 2.020940065383911, + "learning_rate": 3.674183802290178e-05, + "loss": 0.4786, + "num_input_tokens_seen": 4184240, + "step": 15640 + }, + { + "epoch": 4.10091743119266, + "grad_norm": 1.9229618310928345, + "learning_rate": 3.6731739502270344e-05, + "loss": 0.5757, + "num_input_tokens_seen": 4185664, + "step": 15645 + }, + { + "epoch": 4.102228047182176, + "grad_norm": 1.4257078170776367, + "learning_rate": 3.6721638526206115e-05, + "loss": 0.4251, + "num_input_tokens_seen": 4186608, + "step": 15650 + }, + { + "epoch": 4.10353866317169, + "grad_norm": 3.152033805847168, + "learning_rate": 3.671153509682319e-05, + "loss": 0.4413, + "num_input_tokens_seen": 4187632, + "step": 15655 + }, + { + "epoch": 4.104849279161206, + "grad_norm": 1.4027118682861328, + "learning_rate": 3.67014292162362e-05, + "loss": 0.6228, + "num_input_tokens_seen": 4188960, + "step": 15660 + }, + { + "epoch": 4.1061598951507206, + "grad_norm": 4.54227352142334, + "learning_rate": 3.66913208865603e-05, + "loss": 0.6085, + "num_input_tokens_seen": 4190320, + "step": 15665 + }, + { + "epoch": 4.107470511140236, + "grad_norm": 5.15736198425293, + "learning_rate": 3.668121010991115e-05, + "loss": 0.4459, + "num_input_tokens_seen": 4191536, + "step": 15670 + }, + { + "epoch": 4.108781127129751, + "grad_norm": 35.104522705078125, + "learning_rate": 3.66710968884049e-05, + "loss": 1.1433, + "num_input_tokens_seen": 4192720, + "step": 15675 + }, + { + "epoch": 4.110091743119266, + "grad_norm": 2.7772200107574463, + "learning_rate": 3.666098122415823e-05, + "loss": 0.6105, + "num_input_tokens_seen": 4193872, + "step": 15680 + }, + { + "epoch": 4.111402359108781, + "grad_norm": 7.223027229309082, + "learning_rate": 3.665086311928834e-05, + "loss": 0.5594, + "num_input_tokens_seen": 4195152, + "step": 15685 + }, + { + "epoch": 4.1127129750982965, + "grad_norm": 2.1038026809692383, + "learning_rate": 3.664074257591293e-05, + "loss": 0.5599, + "num_input_tokens_seen": 4196416, + "step": 15690 + }, + { + "epoch": 4.114023591087811, + "grad_norm": 2.9854440689086914, + "learning_rate": 3.6630619596150225e-05, + "loss": 0.4231, + "num_input_tokens_seen": 4197616, + "step": 15695 + }, + { + "epoch": 4.115334207077327, + "grad_norm": 2.53328537940979, + "learning_rate": 3.662049418211892e-05, + "loss": 0.4935, + "num_input_tokens_seen": 4198896, + "step": 15700 + }, + { + "epoch": 4.116644823066841, + "grad_norm": 1.7196141481399536, + "learning_rate": 3.6610366335938266e-05, + "loss": 0.49, + "num_input_tokens_seen": 4200464, + "step": 15705 + }, + { + "epoch": 4.117955439056357, + "grad_norm": 13.487162590026855, + "learning_rate": 3.6600236059728e-05, + "loss": 0.8976, + "num_input_tokens_seen": 4201840, + "step": 15710 + }, + { + "epoch": 4.1192660550458715, + "grad_norm": 2.561091661453247, + "learning_rate": 3.6590103355608374e-05, + "loss": 1.6076, + "num_input_tokens_seen": 4202768, + "step": 15715 + }, + { + "epoch": 4.120576671035387, + "grad_norm": 5.765556335449219, + "learning_rate": 3.657996822570013e-05, + "loss": 0.487, + "num_input_tokens_seen": 4203920, + "step": 15720 + }, + { + "epoch": 4.121887287024902, + "grad_norm": 5.817470073699951, + "learning_rate": 3.656983067212456e-05, + "loss": 0.372, + "num_input_tokens_seen": 4204976, + "step": 15725 + }, + { + "epoch": 4.123197903014416, + "grad_norm": 3.6427230834960938, + "learning_rate": 3.655969069700341e-05, + "loss": 0.7472, + "num_input_tokens_seen": 4206416, + "step": 15730 + }, + { + "epoch": 4.124508519003932, + "grad_norm": 3.0017874240875244, + "learning_rate": 3.654954830245898e-05, + "loss": 0.4616, + "num_input_tokens_seen": 4207712, + "step": 15735 + }, + { + "epoch": 4.1258191349934465, + "grad_norm": 3.0994765758514404, + "learning_rate": 3.6539403490614046e-05, + "loss": 0.2899, + "num_input_tokens_seen": 4208928, + "step": 15740 + }, + { + "epoch": 4.127129750982962, + "grad_norm": 2.6691837310791016, + "learning_rate": 3.6529256263591885e-05, + "loss": 0.6157, + "num_input_tokens_seen": 4210624, + "step": 15745 + }, + { + "epoch": 4.128440366972477, + "grad_norm": 1.983094573020935, + "learning_rate": 3.651910662351632e-05, + "loss": 0.5482, + "num_input_tokens_seen": 4212080, + "step": 15750 + }, + { + "epoch": 4.129750982961992, + "grad_norm": 3.7980170249938965, + "learning_rate": 3.650895457251165e-05, + "loss": 0.653, + "num_input_tokens_seen": 4215152, + "step": 15755 + }, + { + "epoch": 4.131061598951507, + "grad_norm": 2.6737661361694336, + "learning_rate": 3.6498800112702664e-05, + "loss": 0.5801, + "num_input_tokens_seen": 4216240, + "step": 15760 + }, + { + "epoch": 4.1323722149410225, + "grad_norm": 2.515955686569214, + "learning_rate": 3.648864324621468e-05, + "loss": 0.4602, + "num_input_tokens_seen": 4217328, + "step": 15765 + }, + { + "epoch": 4.133682830930537, + "grad_norm": 2.1234130859375, + "learning_rate": 3.647848397517353e-05, + "loss": 0.4776, + "num_input_tokens_seen": 4218800, + "step": 15770 + }, + { + "epoch": 4.134993446920053, + "grad_norm": 1.7984298467636108, + "learning_rate": 3.6468322301705504e-05, + "loss": 0.4536, + "num_input_tokens_seen": 4219904, + "step": 15775 + }, + { + "epoch": 4.136304062909567, + "grad_norm": 1.7779369354248047, + "learning_rate": 3.6458158227937435e-05, + "loss": 0.4017, + "num_input_tokens_seen": 4221200, + "step": 15780 + }, + { + "epoch": 4.137614678899083, + "grad_norm": 1.7457247972488403, + "learning_rate": 3.6447991755996654e-05, + "loss": 0.5142, + "num_input_tokens_seen": 4222176, + "step": 15785 + }, + { + "epoch": 4.1389252948885975, + "grad_norm": 1.5520920753479004, + "learning_rate": 3.643782288801098e-05, + "loss": 0.5476, + "num_input_tokens_seen": 4223504, + "step": 15790 + }, + { + "epoch": 4.140235910878113, + "grad_norm": 2.1548359394073486, + "learning_rate": 3.642765162610873e-05, + "loss": 0.5078, + "num_input_tokens_seen": 4225344, + "step": 15795 + }, + { + "epoch": 4.141546526867628, + "grad_norm": 42.42886734008789, + "learning_rate": 3.6417477972418745e-05, + "loss": 0.6137, + "num_input_tokens_seen": 4226544, + "step": 15800 + }, + { + "epoch": 4.142857142857143, + "grad_norm": 0.8606060743331909, + "learning_rate": 3.640730192907035e-05, + "loss": 0.5493, + "num_input_tokens_seen": 4228224, + "step": 15805 + }, + { + "epoch": 4.144167758846658, + "grad_norm": 1.134299874305725, + "learning_rate": 3.639712349819337e-05, + "loss": 0.3136, + "num_input_tokens_seen": 4229824, + "step": 15810 + }, + { + "epoch": 4.145478374836173, + "grad_norm": 1.9213074445724487, + "learning_rate": 3.6386942681918135e-05, + "loss": 0.4309, + "num_input_tokens_seen": 4231088, + "step": 15815 + }, + { + "epoch": 4.146788990825688, + "grad_norm": 1.3449063301086426, + "learning_rate": 3.637675948237547e-05, + "loss": 0.6003, + "num_input_tokens_seen": 4232352, + "step": 15820 + }, + { + "epoch": 4.148099606815203, + "grad_norm": 1.5821645259857178, + "learning_rate": 3.6366573901696704e-05, + "loss": 0.3566, + "num_input_tokens_seen": 4233888, + "step": 15825 + }, + { + "epoch": 4.149410222804718, + "grad_norm": 1.3144258260726929, + "learning_rate": 3.635638594201366e-05, + "loss": 0.3934, + "num_input_tokens_seen": 4235664, + "step": 15830 + }, + { + "epoch": 4.150720838794233, + "grad_norm": 1.8963282108306885, + "learning_rate": 3.6346195605458664e-05, + "loss": 0.5001, + "num_input_tokens_seen": 4237488, + "step": 15835 + }, + { + "epoch": 4.1520314547837485, + "grad_norm": 2.2284927368164062, + "learning_rate": 3.633600289416453e-05, + "loss": 0.3308, + "num_input_tokens_seen": 4238784, + "step": 15840 + }, + { + "epoch": 4.153342070773263, + "grad_norm": 3.112727403640747, + "learning_rate": 3.632580781026457e-05, + "loss": 0.3976, + "num_input_tokens_seen": 4240048, + "step": 15845 + }, + { + "epoch": 4.154652686762779, + "grad_norm": 1.6586867570877075, + "learning_rate": 3.631561035589262e-05, + "loss": 0.6037, + "num_input_tokens_seen": 4241328, + "step": 15850 + }, + { + "epoch": 4.155963302752293, + "grad_norm": 3.3088395595550537, + "learning_rate": 3.630541053318297e-05, + "loss": 0.4804, + "num_input_tokens_seen": 4242416, + "step": 15855 + }, + { + "epoch": 4.157273918741809, + "grad_norm": 2.312814235687256, + "learning_rate": 3.6295208344270415e-05, + "loss": 0.4893, + "num_input_tokens_seen": 4243600, + "step": 15860 + }, + { + "epoch": 4.1585845347313235, + "grad_norm": 1.4521777629852295, + "learning_rate": 3.628500379129028e-05, + "loss": 0.3529, + "num_input_tokens_seen": 4245136, + "step": 15865 + }, + { + "epoch": 4.159895150720839, + "grad_norm": 2.5303385257720947, + "learning_rate": 3.627479687637836e-05, + "loss": 0.3978, + "num_input_tokens_seen": 4247536, + "step": 15870 + }, + { + "epoch": 4.161205766710354, + "grad_norm": 1.8936010599136353, + "learning_rate": 3.6264587601670924e-05, + "loss": 0.4729, + "num_input_tokens_seen": 4248720, + "step": 15875 + }, + { + "epoch": 4.162516382699869, + "grad_norm": 1.0021127462387085, + "learning_rate": 3.625437596930476e-05, + "loss": 0.3418, + "num_input_tokens_seen": 4249952, + "step": 15880 + }, + { + "epoch": 4.163826998689384, + "grad_norm": 8.283982276916504, + "learning_rate": 3.6244161981417165e-05, + "loss": 0.4392, + "num_input_tokens_seen": 4251248, + "step": 15885 + }, + { + "epoch": 4.165137614678899, + "grad_norm": 4.720279216766357, + "learning_rate": 3.62339456401459e-05, + "loss": 0.4705, + "num_input_tokens_seen": 4252336, + "step": 15890 + }, + { + "epoch": 4.166448230668414, + "grad_norm": 1.389763593673706, + "learning_rate": 3.622372694762921e-05, + "loss": 0.2797, + "num_input_tokens_seen": 4253904, + "step": 15895 + }, + { + "epoch": 4.16775884665793, + "grad_norm": 2.179610252380371, + "learning_rate": 3.621350590600587e-05, + "loss": 0.6013, + "num_input_tokens_seen": 4254912, + "step": 15900 + }, + { + "epoch": 4.169069462647444, + "grad_norm": 6.381736755371094, + "learning_rate": 3.6203282517415114e-05, + "loss": 0.7324, + "num_input_tokens_seen": 4256032, + "step": 15905 + }, + { + "epoch": 4.17038007863696, + "grad_norm": 6.406212329864502, + "learning_rate": 3.619305678399669e-05, + "loss": 0.7766, + "num_input_tokens_seen": 4257456, + "step": 15910 + }, + { + "epoch": 4.1716906946264745, + "grad_norm": 3.3507606983184814, + "learning_rate": 3.6182828707890816e-05, + "loss": 0.3294, + "num_input_tokens_seen": 4258384, + "step": 15915 + }, + { + "epoch": 4.173001310615989, + "grad_norm": 2.0809414386749268, + "learning_rate": 3.617259829123822e-05, + "loss": 0.4758, + "num_input_tokens_seen": 4259728, + "step": 15920 + }, + { + "epoch": 4.174311926605505, + "grad_norm": 3.6960134506225586, + "learning_rate": 3.6162365536180105e-05, + "loss": 0.4222, + "num_input_tokens_seen": 4261296, + "step": 15925 + }, + { + "epoch": 4.175622542595019, + "grad_norm": 1.706628680229187, + "learning_rate": 3.615213044485817e-05, + "loss": 0.5564, + "num_input_tokens_seen": 4263168, + "step": 15930 + }, + { + "epoch": 4.176933158584535, + "grad_norm": 2.936422824859619, + "learning_rate": 3.6141893019414606e-05, + "loss": 0.3208, + "num_input_tokens_seen": 4264096, + "step": 15935 + }, + { + "epoch": 4.1782437745740495, + "grad_norm": 2.657756805419922, + "learning_rate": 3.6131653261992085e-05, + "loss": 0.3865, + "num_input_tokens_seen": 4265696, + "step": 15940 + }, + { + "epoch": 4.179554390563565, + "grad_norm": 1.871038556098938, + "learning_rate": 3.612141117473377e-05, + "loss": 0.3531, + "num_input_tokens_seen": 4266736, + "step": 15945 + }, + { + "epoch": 4.18086500655308, + "grad_norm": 8.643420219421387, + "learning_rate": 3.61111667597833e-05, + "loss": 0.768, + "num_input_tokens_seen": 4267856, + "step": 15950 + }, + { + "epoch": 4.182175622542595, + "grad_norm": 2.633197546005249, + "learning_rate": 3.610092001928484e-05, + "loss": 0.4842, + "num_input_tokens_seen": 4268928, + "step": 15955 + }, + { + "epoch": 4.18348623853211, + "grad_norm": 2.587554693222046, + "learning_rate": 3.6090670955383e-05, + "loss": 0.4106, + "num_input_tokens_seen": 4270032, + "step": 15960 + }, + { + "epoch": 4.184796854521625, + "grad_norm": 1.5853036642074585, + "learning_rate": 3.6080419570222885e-05, + "loss": 0.4169, + "num_input_tokens_seen": 4271024, + "step": 15965 + }, + { + "epoch": 4.18610747051114, + "grad_norm": 2.40661358833313, + "learning_rate": 3.60701658659501e-05, + "loss": 0.4731, + "num_input_tokens_seen": 4272000, + "step": 15970 + }, + { + "epoch": 4.187418086500656, + "grad_norm": 5.990711212158203, + "learning_rate": 3.605990984471073e-05, + "loss": 0.4407, + "num_input_tokens_seen": 4273248, + "step": 15975 + }, + { + "epoch": 4.18872870249017, + "grad_norm": 1.9335557222366333, + "learning_rate": 3.6049651508651336e-05, + "loss": 0.5056, + "num_input_tokens_seen": 4275152, + "step": 15980 + }, + { + "epoch": 4.190039318479686, + "grad_norm": 2.0539870262145996, + "learning_rate": 3.603939085991898e-05, + "loss": 0.5977, + "num_input_tokens_seen": 4276496, + "step": 15985 + }, + { + "epoch": 4.1913499344692005, + "grad_norm": 1.0690386295318604, + "learning_rate": 3.6029127900661196e-05, + "loss": 0.3204, + "num_input_tokens_seen": 4278432, + "step": 15990 + }, + { + "epoch": 4.192660550458716, + "grad_norm": 2.8285789489746094, + "learning_rate": 3.601886263302599e-05, + "loss": 0.5306, + "num_input_tokens_seen": 4279632, + "step": 15995 + }, + { + "epoch": 4.193971166448231, + "grad_norm": 3.9504554271698, + "learning_rate": 3.600859505916187e-05, + "loss": 0.782, + "num_input_tokens_seen": 4281024, + "step": 16000 + }, + { + "epoch": 4.195281782437746, + "grad_norm": 2.1852455139160156, + "learning_rate": 3.599832518121784e-05, + "loss": 0.392, + "num_input_tokens_seen": 4282912, + "step": 16005 + }, + { + "epoch": 4.196592398427261, + "grad_norm": 1.942561149597168, + "learning_rate": 3.598805300134334e-05, + "loss": 0.4104, + "num_input_tokens_seen": 4284080, + "step": 16010 + }, + { + "epoch": 4.1979030144167755, + "grad_norm": 2.2186455726623535, + "learning_rate": 3.597777852168834e-05, + "loss": 0.6384, + "num_input_tokens_seen": 4285472, + "step": 16015 + }, + { + "epoch": 4.199213630406291, + "grad_norm": 5.792991638183594, + "learning_rate": 3.596750174440326e-05, + "loss": 0.5401, + "num_input_tokens_seen": 4286576, + "step": 16020 + }, + { + "epoch": 4.200524246395806, + "grad_norm": 4.150318145751953, + "learning_rate": 3.595722267163901e-05, + "loss": 0.4363, + "num_input_tokens_seen": 4288096, + "step": 16025 + }, + { + "epoch": 4.201834862385321, + "grad_norm": 4.536371231079102, + "learning_rate": 3.5946941305547e-05, + "loss": 0.4988, + "num_input_tokens_seen": 4289440, + "step": 16030 + }, + { + "epoch": 4.203145478374836, + "grad_norm": 6.626453876495361, + "learning_rate": 3.593665764827908e-05, + "loss": 0.5826, + "num_input_tokens_seen": 4290448, + "step": 16035 + }, + { + "epoch": 4.204456094364351, + "grad_norm": 1.8477915525436401, + "learning_rate": 3.592637170198762e-05, + "loss": 0.5341, + "num_input_tokens_seen": 4291440, + "step": 16040 + }, + { + "epoch": 4.205766710353866, + "grad_norm": 3.754084825515747, + "learning_rate": 3.591608346882543e-05, + "loss": 0.5369, + "num_input_tokens_seen": 4292976, + "step": 16045 + }, + { + "epoch": 4.207077326343382, + "grad_norm": 1.001760482788086, + "learning_rate": 3.590579295094584e-05, + "loss": 0.4516, + "num_input_tokens_seen": 4294672, + "step": 16050 + }, + { + "epoch": 4.208387942332896, + "grad_norm": 1.2694658041000366, + "learning_rate": 3.589550015050263e-05, + "loss": 0.6638, + "num_input_tokens_seen": 4296224, + "step": 16055 + }, + { + "epoch": 4.209698558322412, + "grad_norm": 2.2111825942993164, + "learning_rate": 3.5885205069650056e-05, + "loss": 0.6471, + "num_input_tokens_seen": 4297360, + "step": 16060 + }, + { + "epoch": 4.2110091743119265, + "grad_norm": 3.42858624458313, + "learning_rate": 3.587490771054288e-05, + "loss": 0.4198, + "num_input_tokens_seen": 4298768, + "step": 16065 + }, + { + "epoch": 4.212319790301442, + "grad_norm": 2.71714448928833, + "learning_rate": 3.586460807533631e-05, + "loss": 0.3475, + "num_input_tokens_seen": 4299984, + "step": 16070 + }, + { + "epoch": 4.213630406290957, + "grad_norm": 2.484699010848999, + "learning_rate": 3.585430616618604e-05, + "loss": 0.3566, + "num_input_tokens_seen": 4301472, + "step": 16075 + }, + { + "epoch": 4.214941022280472, + "grad_norm": 3.3567962646484375, + "learning_rate": 3.584400198524824e-05, + "loss": 0.2617, + "num_input_tokens_seen": 4302416, + "step": 16080 + }, + { + "epoch": 4.216251638269987, + "grad_norm": 2.342733860015869, + "learning_rate": 3.583369553467956e-05, + "loss": 0.7415, + "num_input_tokens_seen": 4303648, + "step": 16085 + }, + { + "epoch": 4.217562254259502, + "grad_norm": 4.4239959716796875, + "learning_rate": 3.582338681663713e-05, + "loss": 0.4119, + "num_input_tokens_seen": 4304752, + "step": 16090 + }, + { + "epoch": 4.218872870249017, + "grad_norm": 2.7481942176818848, + "learning_rate": 3.581307583327854e-05, + "loss": 0.4823, + "num_input_tokens_seen": 4306256, + "step": 16095 + }, + { + "epoch": 4.220183486238533, + "grad_norm": 3.232060670852661, + "learning_rate": 3.580276258676186e-05, + "loss": 0.5305, + "num_input_tokens_seen": 4308080, + "step": 16100 + }, + { + "epoch": 4.221494102228047, + "grad_norm": 2.0854945182800293, + "learning_rate": 3.579244707924564e-05, + "loss": 0.5738, + "num_input_tokens_seen": 4309840, + "step": 16105 + }, + { + "epoch": 4.222804718217562, + "grad_norm": 134.27244567871094, + "learning_rate": 3.5782129312888885e-05, + "loss": 0.6806, + "num_input_tokens_seen": 4311632, + "step": 16110 + }, + { + "epoch": 4.224115334207077, + "grad_norm": 2.1521825790405273, + "learning_rate": 3.5771809289851097e-05, + "loss": 0.618, + "num_input_tokens_seen": 4313104, + "step": 16115 + }, + { + "epoch": 4.225425950196592, + "grad_norm": 1.253336787223816, + "learning_rate": 3.576148701229224e-05, + "loss": 0.4307, + "num_input_tokens_seen": 4314656, + "step": 16120 + }, + { + "epoch": 4.226736566186108, + "grad_norm": 4.9405364990234375, + "learning_rate": 3.575116248237275e-05, + "loss": 0.387, + "num_input_tokens_seen": 4316064, + "step": 16125 + }, + { + "epoch": 4.228047182175622, + "grad_norm": 3.3537116050720215, + "learning_rate": 3.574083570225352e-05, + "loss": 0.5236, + "num_input_tokens_seen": 4317136, + "step": 16130 + }, + { + "epoch": 4.229357798165138, + "grad_norm": 4.5580034255981445, + "learning_rate": 3.573050667409594e-05, + "loss": 0.4343, + "num_input_tokens_seen": 4318480, + "step": 16135 + }, + { + "epoch": 4.2306684141546524, + "grad_norm": 2.9832513332366943, + "learning_rate": 3.572017540006186e-05, + "loss": 0.4753, + "num_input_tokens_seen": 4320032, + "step": 16140 + }, + { + "epoch": 4.231979030144168, + "grad_norm": 3.668379068374634, + "learning_rate": 3.570984188231359e-05, + "loss": 0.4706, + "num_input_tokens_seen": 4321088, + "step": 16145 + }, + { + "epoch": 4.233289646133683, + "grad_norm": 1.3515676259994507, + "learning_rate": 3.5699506123013906e-05, + "loss": 0.3659, + "num_input_tokens_seen": 4322336, + "step": 16150 + }, + { + "epoch": 4.234600262123198, + "grad_norm": 2.4164938926696777, + "learning_rate": 3.568916812432609e-05, + "loss": 0.5978, + "num_input_tokens_seen": 4323632, + "step": 16155 + }, + { + "epoch": 4.235910878112713, + "grad_norm": 3.0876951217651367, + "learning_rate": 3.5678827888413856e-05, + "loss": 0.4219, + "num_input_tokens_seen": 4324960, + "step": 16160 + }, + { + "epoch": 4.237221494102228, + "grad_norm": 3.0084593296051025, + "learning_rate": 3.56684854174414e-05, + "loss": 0.3598, + "num_input_tokens_seen": 4326144, + "step": 16165 + }, + { + "epoch": 4.238532110091743, + "grad_norm": 3.802082061767578, + "learning_rate": 3.5658140713573365e-05, + "loss": 0.4665, + "num_input_tokens_seen": 4327808, + "step": 16170 + }, + { + "epoch": 4.239842726081259, + "grad_norm": 3.396256923675537, + "learning_rate": 3.56477937789749e-05, + "loss": 0.4474, + "num_input_tokens_seen": 4328880, + "step": 16175 + }, + { + "epoch": 4.241153342070773, + "grad_norm": 1.652116298675537, + "learning_rate": 3.56374446158116e-05, + "loss": 0.5379, + "num_input_tokens_seen": 4330336, + "step": 16180 + }, + { + "epoch": 4.242463958060289, + "grad_norm": 6.661996364593506, + "learning_rate": 3.5627093226249524e-05, + "loss": 0.7322, + "num_input_tokens_seen": 4331440, + "step": 16185 + }, + { + "epoch": 4.243774574049803, + "grad_norm": 1.6538710594177246, + "learning_rate": 3.5616739612455196e-05, + "loss": 0.5282, + "num_input_tokens_seen": 4334112, + "step": 16190 + }, + { + "epoch": 4.245085190039318, + "grad_norm": 1.9271225929260254, + "learning_rate": 3.560638377659561e-05, + "loss": 0.4511, + "num_input_tokens_seen": 4335024, + "step": 16195 + }, + { + "epoch": 4.246395806028834, + "grad_norm": 2.918299436569214, + "learning_rate": 3.559602572083822e-05, + "loss": 0.3548, + "num_input_tokens_seen": 4336160, + "step": 16200 + }, + { + "epoch": 4.247706422018348, + "grad_norm": 3.6079540252685547, + "learning_rate": 3.558566544735096e-05, + "loss": 0.5233, + "num_input_tokens_seen": 4338032, + "step": 16205 + }, + { + "epoch": 4.249017038007864, + "grad_norm": 2.5969271659851074, + "learning_rate": 3.5575302958302214e-05, + "loss": 0.4693, + "num_input_tokens_seen": 4339520, + "step": 16210 + }, + { + "epoch": 4.250327653997378, + "grad_norm": 1.9179362058639526, + "learning_rate": 3.556493825586083e-05, + "loss": 0.3094, + "num_input_tokens_seen": 4340848, + "step": 16215 + }, + { + "epoch": 4.251638269986894, + "grad_norm": 2.6818158626556396, + "learning_rate": 3.555457134219613e-05, + "loss": 0.5697, + "num_input_tokens_seen": 4342176, + "step": 16220 + }, + { + "epoch": 4.252948885976409, + "grad_norm": 3.3952200412750244, + "learning_rate": 3.554420221947788e-05, + "loss": 0.5122, + "num_input_tokens_seen": 4343360, + "step": 16225 + }, + { + "epoch": 4.254259501965924, + "grad_norm": 2.91845440864563, + "learning_rate": 3.553383088987632e-05, + "loss": 0.4394, + "num_input_tokens_seen": 4344608, + "step": 16230 + }, + { + "epoch": 4.255570117955439, + "grad_norm": 1.5595431327819824, + "learning_rate": 3.552345735556216e-05, + "loss": 0.4298, + "num_input_tokens_seen": 4346096, + "step": 16235 + }, + { + "epoch": 4.256880733944954, + "grad_norm": 1.548219084739685, + "learning_rate": 3.5513081618706555e-05, + "loss": 0.5138, + "num_input_tokens_seen": 4347408, + "step": 16240 + }, + { + "epoch": 4.258191349934469, + "grad_norm": 1.9269428253173828, + "learning_rate": 3.550270368148113e-05, + "loss": 0.5675, + "num_input_tokens_seen": 4348848, + "step": 16245 + }, + { + "epoch": 4.259501965923985, + "grad_norm": 2.8023531436920166, + "learning_rate": 3.549232354605798e-05, + "loss": 0.5741, + "num_input_tokens_seen": 4350176, + "step": 16250 + }, + { + "epoch": 4.260812581913499, + "grad_norm": 1.6540749073028564, + "learning_rate": 3.548194121460964e-05, + "loss": 0.5096, + "num_input_tokens_seen": 4351344, + "step": 16255 + }, + { + "epoch": 4.262123197903015, + "grad_norm": 1.7167341709136963, + "learning_rate": 3.547155668930912e-05, + "loss": 0.4158, + "num_input_tokens_seen": 4352672, + "step": 16260 + }, + { + "epoch": 4.263433813892529, + "grad_norm": 2.0690364837646484, + "learning_rate": 3.546116997232987e-05, + "loss": 0.4033, + "num_input_tokens_seen": 4353904, + "step": 16265 + }, + { + "epoch": 4.264744429882045, + "grad_norm": 4.724527359008789, + "learning_rate": 3.545078106584582e-05, + "loss": 0.6072, + "num_input_tokens_seen": 4355296, + "step": 16270 + }, + { + "epoch": 4.26605504587156, + "grad_norm": 2.4165449142456055, + "learning_rate": 3.544038997203135e-05, + "loss": 0.6096, + "num_input_tokens_seen": 4356608, + "step": 16275 + }, + { + "epoch": 4.267365661861074, + "grad_norm": 4.555400371551514, + "learning_rate": 3.5429996693061304e-05, + "loss": 0.3585, + "num_input_tokens_seen": 4357776, + "step": 16280 + }, + { + "epoch": 4.26867627785059, + "grad_norm": 2.064983606338501, + "learning_rate": 3.541960123111097e-05, + "loss": 0.4416, + "num_input_tokens_seen": 4359024, + "step": 16285 + }, + { + "epoch": 4.269986893840104, + "grad_norm": 6.6770172119140625, + "learning_rate": 3.54092035883561e-05, + "loss": 0.3735, + "num_input_tokens_seen": 4360240, + "step": 16290 + }, + { + "epoch": 4.27129750982962, + "grad_norm": 4.430813789367676, + "learning_rate": 3.53988037669729e-05, + "loss": 0.4897, + "num_input_tokens_seen": 4361184, + "step": 16295 + }, + { + "epoch": 4.272608125819135, + "grad_norm": 1.4800920486450195, + "learning_rate": 3.5388401769138053e-05, + "loss": 0.5675, + "num_input_tokens_seen": 4362688, + "step": 16300 + }, + { + "epoch": 4.27391874180865, + "grad_norm": 2.071070671081543, + "learning_rate": 3.5377997597028656e-05, + "loss": 0.4176, + "num_input_tokens_seen": 4363808, + "step": 16305 + }, + { + "epoch": 4.275229357798165, + "grad_norm": 1.6590381860733032, + "learning_rate": 3.53675912528223e-05, + "loss": 0.3599, + "num_input_tokens_seen": 4365488, + "step": 16310 + }, + { + "epoch": 4.27653997378768, + "grad_norm": 1.6458157300949097, + "learning_rate": 3.535718273869699e-05, + "loss": 0.4949, + "num_input_tokens_seen": 4367152, + "step": 16315 + }, + { + "epoch": 4.277850589777195, + "grad_norm": 2.611449718475342, + "learning_rate": 3.534677205683125e-05, + "loss": 0.5983, + "num_input_tokens_seen": 4368208, + "step": 16320 + }, + { + "epoch": 4.2791612057667106, + "grad_norm": 7.6197509765625, + "learning_rate": 3.533635920940399e-05, + "loss": 0.5691, + "num_input_tokens_seen": 4369888, + "step": 16325 + }, + { + "epoch": 4.280471821756225, + "grad_norm": 2.614469051361084, + "learning_rate": 3.5325944198594607e-05, + "loss": 0.56, + "num_input_tokens_seen": 4371408, + "step": 16330 + }, + { + "epoch": 4.281782437745741, + "grad_norm": 3.314027786254883, + "learning_rate": 3.531552702658294e-05, + "loss": 0.3404, + "num_input_tokens_seen": 4372832, + "step": 16335 + }, + { + "epoch": 4.283093053735255, + "grad_norm": 1.0475798845291138, + "learning_rate": 3.53051076955493e-05, + "loss": 0.504, + "num_input_tokens_seen": 4374240, + "step": 16340 + }, + { + "epoch": 4.284403669724771, + "grad_norm": 1.189955472946167, + "learning_rate": 3.5294686207674414e-05, + "loss": 0.5809, + "num_input_tokens_seen": 4375424, + "step": 16345 + }, + { + "epoch": 4.285714285714286, + "grad_norm": 1.5220086574554443, + "learning_rate": 3.5284262565139494e-05, + "loss": 0.2599, + "num_input_tokens_seen": 4376768, + "step": 16350 + }, + { + "epoch": 4.287024901703801, + "grad_norm": 1.2763906717300415, + "learning_rate": 3.5273836770126204e-05, + "loss": 0.3954, + "num_input_tokens_seen": 4378224, + "step": 16355 + }, + { + "epoch": 4.288335517693316, + "grad_norm": 3.927760601043701, + "learning_rate": 3.5263408824816615e-05, + "loss": 0.5207, + "num_input_tokens_seen": 4379504, + "step": 16360 + }, + { + "epoch": 4.289646133682831, + "grad_norm": 1.776286244392395, + "learning_rate": 3.52529787313933e-05, + "loss": 0.48, + "num_input_tokens_seen": 4380832, + "step": 16365 + }, + { + "epoch": 4.290956749672346, + "grad_norm": 2.467876672744751, + "learning_rate": 3.524254649203925e-05, + "loss": 0.5214, + "num_input_tokens_seen": 4382064, + "step": 16370 + }, + { + "epoch": 4.292267365661861, + "grad_norm": 1.962683916091919, + "learning_rate": 3.5232112108937914e-05, + "loss": 0.5641, + "num_input_tokens_seen": 4383648, + "step": 16375 + }, + { + "epoch": 4.293577981651376, + "grad_norm": 3.849789619445801, + "learning_rate": 3.52216755842732e-05, + "loss": 0.4894, + "num_input_tokens_seen": 4384704, + "step": 16380 + }, + { + "epoch": 4.294888597640891, + "grad_norm": 2.2731032371520996, + "learning_rate": 3.521123692022944e-05, + "loss": 0.6146, + "num_input_tokens_seen": 4386240, + "step": 16385 + }, + { + "epoch": 4.296199213630406, + "grad_norm": 2.2451364994049072, + "learning_rate": 3.520079611899144e-05, + "loss": 0.4586, + "num_input_tokens_seen": 4387472, + "step": 16390 + }, + { + "epoch": 4.297509829619921, + "grad_norm": 2.488159656524658, + "learning_rate": 3.519035318274444e-05, + "loss": 0.4139, + "num_input_tokens_seen": 4388784, + "step": 16395 + }, + { + "epoch": 4.2988204456094365, + "grad_norm": 2.3639039993286133, + "learning_rate": 3.517990811367412e-05, + "loss": 0.386, + "num_input_tokens_seen": 4390016, + "step": 16400 + }, + { + "epoch": 4.300131061598951, + "grad_norm": 2.787994623184204, + "learning_rate": 3.516946091396662e-05, + "loss": 0.7442, + "num_input_tokens_seen": 4391216, + "step": 16405 + }, + { + "epoch": 4.301441677588467, + "grad_norm": 1.3398683071136475, + "learning_rate": 3.5159011585808525e-05, + "loss": 0.4606, + "num_input_tokens_seen": 4392736, + "step": 16410 + }, + { + "epoch": 4.302752293577981, + "grad_norm": 4.915399074554443, + "learning_rate": 3.5148560131386867e-05, + "loss": 0.3933, + "num_input_tokens_seen": 4393792, + "step": 16415 + }, + { + "epoch": 4.304062909567497, + "grad_norm": 1.6949917078018188, + "learning_rate": 3.5138106552889106e-05, + "loss": 0.4619, + "num_input_tokens_seen": 4395136, + "step": 16420 + }, + { + "epoch": 4.305373525557012, + "grad_norm": 3.0986194610595703, + "learning_rate": 3.5127650852503156e-05, + "loss": 0.4062, + "num_input_tokens_seen": 4396144, + "step": 16425 + }, + { + "epoch": 4.306684141546527, + "grad_norm": 1.6095994710922241, + "learning_rate": 3.511719303241738e-05, + "loss": 0.4795, + "num_input_tokens_seen": 4397536, + "step": 16430 + }, + { + "epoch": 4.307994757536042, + "grad_norm": 1.2216920852661133, + "learning_rate": 3.5106733094820586e-05, + "loss": 0.3409, + "num_input_tokens_seen": 4399088, + "step": 16435 + }, + { + "epoch": 4.309305373525557, + "grad_norm": 1.1357476711273193, + "learning_rate": 3.509627104190202e-05, + "loss": 0.5566, + "num_input_tokens_seen": 4400528, + "step": 16440 + }, + { + "epoch": 4.310615989515072, + "grad_norm": 3.9740777015686035, + "learning_rate": 3.5085806875851365e-05, + "loss": 0.6621, + "num_input_tokens_seen": 4401840, + "step": 16445 + }, + { + "epoch": 4.3119266055045875, + "grad_norm": 8.063874244689941, + "learning_rate": 3.507534059885876e-05, + "loss": 0.5067, + "num_input_tokens_seen": 4403024, + "step": 16450 + }, + { + "epoch": 4.313237221494102, + "grad_norm": 1.601525902748108, + "learning_rate": 3.5064872213114784e-05, + "loss": 0.597, + "num_input_tokens_seen": 4404432, + "step": 16455 + }, + { + "epoch": 4.314547837483618, + "grad_norm": 2.4602115154266357, + "learning_rate": 3.505440172081044e-05, + "loss": 0.4848, + "num_input_tokens_seen": 4405920, + "step": 16460 + }, + { + "epoch": 4.315858453473132, + "grad_norm": 2.1808009147644043, + "learning_rate": 3.504392912413719e-05, + "loss": 0.441, + "num_input_tokens_seen": 4406976, + "step": 16465 + }, + { + "epoch": 4.317169069462647, + "grad_norm": 3.812412738800049, + "learning_rate": 3.503345442528692e-05, + "loss": 0.5286, + "num_input_tokens_seen": 4408336, + "step": 16470 + }, + { + "epoch": 4.3184796854521625, + "grad_norm": 4.352725505828857, + "learning_rate": 3.5022977626451985e-05, + "loss": 0.6259, + "num_input_tokens_seen": 4409904, + "step": 16475 + }, + { + "epoch": 4.319790301441677, + "grad_norm": 2.6824862957000732, + "learning_rate": 3.501249872982515e-05, + "loss": 0.6388, + "num_input_tokens_seen": 4411296, + "step": 16480 + }, + { + "epoch": 4.321100917431193, + "grad_norm": 1.4099664688110352, + "learning_rate": 3.500201773759963e-05, + "loss": 0.5123, + "num_input_tokens_seen": 4412768, + "step": 16485 + }, + { + "epoch": 4.322411533420707, + "grad_norm": 1.7036930322647095, + "learning_rate": 3.499153465196907e-05, + "loss": 0.4641, + "num_input_tokens_seen": 4414064, + "step": 16490 + }, + { + "epoch": 4.323722149410223, + "grad_norm": 5.111568450927734, + "learning_rate": 3.4981049475127584e-05, + "loss": 0.7575, + "num_input_tokens_seen": 4415136, + "step": 16495 + }, + { + "epoch": 4.325032765399738, + "grad_norm": 5.747180938720703, + "learning_rate": 3.4970562209269684e-05, + "loss": 0.5078, + "num_input_tokens_seen": 4416304, + "step": 16500 + }, + { + "epoch": 4.326343381389253, + "grad_norm": 1.556764841079712, + "learning_rate": 3.496007285659034e-05, + "loss": 0.6647, + "num_input_tokens_seen": 4417888, + "step": 16505 + }, + { + "epoch": 4.327653997378768, + "grad_norm": 2.674917459487915, + "learning_rate": 3.494958141928496e-05, + "loss": 0.4827, + "num_input_tokens_seen": 4418784, + "step": 16510 + }, + { + "epoch": 4.328964613368283, + "grad_norm": 3.2310590744018555, + "learning_rate": 3.4939087899549386e-05, + "loss": 0.7218, + "num_input_tokens_seen": 4419968, + "step": 16515 + }, + { + "epoch": 4.330275229357798, + "grad_norm": 1.2034496068954468, + "learning_rate": 3.4928592299579874e-05, + "loss": 0.4162, + "num_input_tokens_seen": 4421264, + "step": 16520 + }, + { + "epoch": 4.3315858453473135, + "grad_norm": 3.506653308868408, + "learning_rate": 3.4918094621573164e-05, + "loss": 0.3667, + "num_input_tokens_seen": 4422576, + "step": 16525 + }, + { + "epoch": 4.332896461336828, + "grad_norm": 1.2423869371414185, + "learning_rate": 3.490759486772637e-05, + "loss": 0.4298, + "num_input_tokens_seen": 4425536, + "step": 16530 + }, + { + "epoch": 4.334207077326344, + "grad_norm": 2.3462605476379395, + "learning_rate": 3.48970930402371e-05, + "loss": 0.366, + "num_input_tokens_seen": 4426576, + "step": 16535 + }, + { + "epoch": 4.335517693315858, + "grad_norm": 2.8157870769500732, + "learning_rate": 3.488658914130336e-05, + "loss": 0.5644, + "num_input_tokens_seen": 4428112, + "step": 16540 + }, + { + "epoch": 4.336828309305374, + "grad_norm": 2.314225673675537, + "learning_rate": 3.487608317312358e-05, + "loss": 0.4688, + "num_input_tokens_seen": 4429184, + "step": 16545 + }, + { + "epoch": 4.3381389252948885, + "grad_norm": 2.3864314556121826, + "learning_rate": 3.486557513789668e-05, + "loss": 0.4813, + "num_input_tokens_seen": 4430640, + "step": 16550 + }, + { + "epoch": 4.339449541284404, + "grad_norm": 1.9328793287277222, + "learning_rate": 3.485506503782194e-05, + "loss": 0.3307, + "num_input_tokens_seen": 4431904, + "step": 16555 + }, + { + "epoch": 4.340760157273919, + "grad_norm": 3.4048452377319336, + "learning_rate": 3.484455287509911e-05, + "loss": 0.487, + "num_input_tokens_seen": 4433424, + "step": 16560 + }, + { + "epoch": 4.342070773263433, + "grad_norm": 3.955986976623535, + "learning_rate": 3.483403865192838e-05, + "loss": 0.3909, + "num_input_tokens_seen": 4434320, + "step": 16565 + }, + { + "epoch": 4.343381389252949, + "grad_norm": 2.597184419631958, + "learning_rate": 3.482352237051035e-05, + "loss": 0.4798, + "num_input_tokens_seen": 4435856, + "step": 16570 + }, + { + "epoch": 4.344692005242464, + "grad_norm": 4.6065592765808105, + "learning_rate": 3.481300403304606e-05, + "loss": 0.5761, + "num_input_tokens_seen": 4436640, + "step": 16575 + }, + { + "epoch": 4.346002621231979, + "grad_norm": 3.1853601932525635, + "learning_rate": 3.480248364173698e-05, + "loss": 0.4322, + "num_input_tokens_seen": 4437776, + "step": 16580 + }, + { + "epoch": 4.347313237221494, + "grad_norm": 2.595991373062134, + "learning_rate": 3.4791961198785014e-05, + "loss": 0.3216, + "num_input_tokens_seen": 4438896, + "step": 16585 + }, + { + "epoch": 4.348623853211009, + "grad_norm": 3.43558406829834, + "learning_rate": 3.478143670639248e-05, + "loss": 0.4429, + "num_input_tokens_seen": 4440160, + "step": 16590 + }, + { + "epoch": 4.349934469200524, + "grad_norm": 1.2716434001922607, + "learning_rate": 3.477091016676216e-05, + "loss": 0.548, + "num_input_tokens_seen": 4442000, + "step": 16595 + }, + { + "epoch": 4.3512450851900395, + "grad_norm": 1.936081886291504, + "learning_rate": 3.476038158209722e-05, + "loss": 0.3941, + "num_input_tokens_seen": 4443632, + "step": 16600 + }, + { + "epoch": 4.352555701179554, + "grad_norm": 1.3810384273529053, + "learning_rate": 3.4749850954601274e-05, + "loss": 0.3473, + "num_input_tokens_seen": 4445024, + "step": 16605 + }, + { + "epoch": 4.35386631716907, + "grad_norm": 2.1513264179229736, + "learning_rate": 3.473931828647837e-05, + "loss": 0.416, + "num_input_tokens_seen": 4446192, + "step": 16610 + }, + { + "epoch": 4.355176933158584, + "grad_norm": 38.54049301147461, + "learning_rate": 3.472878357993298e-05, + "loss": 0.5147, + "num_input_tokens_seen": 4447296, + "step": 16615 + }, + { + "epoch": 4.3564875491481, + "grad_norm": 4.355993270874023, + "learning_rate": 3.4718246837169996e-05, + "loss": 0.4068, + "num_input_tokens_seen": 4448528, + "step": 16620 + }, + { + "epoch": 4.3577981651376145, + "grad_norm": 2.821674108505249, + "learning_rate": 3.470770806039474e-05, + "loss": 0.5068, + "num_input_tokens_seen": 4449696, + "step": 16625 + }, + { + "epoch": 4.35910878112713, + "grad_norm": 2.82965087890625, + "learning_rate": 3.469716725181296e-05, + "loss": 0.4036, + "num_input_tokens_seen": 4451168, + "step": 16630 + }, + { + "epoch": 4.360419397116645, + "grad_norm": 14.417654991149902, + "learning_rate": 3.468662441363083e-05, + "loss": 0.6765, + "num_input_tokens_seen": 4452464, + "step": 16635 + }, + { + "epoch": 4.36173001310616, + "grad_norm": 3.9766716957092285, + "learning_rate": 3.467607954805495e-05, + "loss": 0.5073, + "num_input_tokens_seen": 4453712, + "step": 16640 + }, + { + "epoch": 4.363040629095675, + "grad_norm": 2.419063091278076, + "learning_rate": 3.466553265729233e-05, + "loss": 0.3577, + "num_input_tokens_seen": 4455200, + "step": 16645 + }, + { + "epoch": 4.3643512450851905, + "grad_norm": 2.560380458831787, + "learning_rate": 3.465498374355044e-05, + "loss": 0.3841, + "num_input_tokens_seen": 4456416, + "step": 16650 + }, + { + "epoch": 4.365661861074705, + "grad_norm": 2.2476236820220947, + "learning_rate": 3.4644432809037134e-05, + "loss": 0.6097, + "num_input_tokens_seen": 4458032, + "step": 16655 + }, + { + "epoch": 4.36697247706422, + "grad_norm": 2.219202995300293, + "learning_rate": 3.463387985596069e-05, + "loss": 0.4502, + "num_input_tokens_seen": 4459552, + "step": 16660 + }, + { + "epoch": 4.368283093053735, + "grad_norm": 1.808117389678955, + "learning_rate": 3.462332488652985e-05, + "loss": 0.3643, + "num_input_tokens_seen": 4460800, + "step": 16665 + }, + { + "epoch": 4.36959370904325, + "grad_norm": 1.6166713237762451, + "learning_rate": 3.461276790295374e-05, + "loss": 0.4692, + "num_input_tokens_seen": 4462592, + "step": 16670 + }, + { + "epoch": 4.3709043250327655, + "grad_norm": 2.837146759033203, + "learning_rate": 3.4602208907441915e-05, + "loss": 0.6136, + "num_input_tokens_seen": 4463712, + "step": 16675 + }, + { + "epoch": 4.37221494102228, + "grad_norm": 2.754598617553711, + "learning_rate": 3.4591647902204355e-05, + "loss": 0.5439, + "num_input_tokens_seen": 4465008, + "step": 16680 + }, + { + "epoch": 4.373525557011796, + "grad_norm": 3.2312023639678955, + "learning_rate": 3.4581084889451466e-05, + "loss": 0.3907, + "num_input_tokens_seen": 4467200, + "step": 16685 + }, + { + "epoch": 4.37483617300131, + "grad_norm": 5.294649600982666, + "learning_rate": 3.457051987139406e-05, + "loss": 0.4633, + "num_input_tokens_seen": 4468464, + "step": 16690 + }, + { + "epoch": 4.376146788990826, + "grad_norm": 3.4172565937042236, + "learning_rate": 3.455995285024338e-05, + "loss": 0.6192, + "num_input_tokens_seen": 4469968, + "step": 16695 + }, + { + "epoch": 4.3774574049803405, + "grad_norm": 2.3195106983184814, + "learning_rate": 3.454938382821109e-05, + "loss": 0.5501, + "num_input_tokens_seen": 4471520, + "step": 16700 + }, + { + "epoch": 4.378768020969856, + "grad_norm": 2.2411632537841797, + "learning_rate": 3.453881280750927e-05, + "loss": 0.4459, + "num_input_tokens_seen": 4473088, + "step": 16705 + }, + { + "epoch": 4.380078636959371, + "grad_norm": 1.721932291984558, + "learning_rate": 3.452823979035041e-05, + "loss": 0.7405, + "num_input_tokens_seen": 4474432, + "step": 16710 + }, + { + "epoch": 4.381389252948886, + "grad_norm": 3.9116289615631104, + "learning_rate": 3.451766477894742e-05, + "loss": 0.532, + "num_input_tokens_seen": 4475696, + "step": 16715 + }, + { + "epoch": 4.382699868938401, + "grad_norm": 6.157450199127197, + "learning_rate": 3.450708777551364e-05, + "loss": 0.3941, + "num_input_tokens_seen": 4477008, + "step": 16720 + }, + { + "epoch": 4.3840104849279165, + "grad_norm": 8.803646087646484, + "learning_rate": 3.4496508782262804e-05, + "loss": 0.523, + "num_input_tokens_seen": 4478256, + "step": 16725 + }, + { + "epoch": 4.385321100917431, + "grad_norm": 1.355545997619629, + "learning_rate": 3.4485927801409105e-05, + "loss": 0.3921, + "num_input_tokens_seen": 4479888, + "step": 16730 + }, + { + "epoch": 4.386631716906947, + "grad_norm": 2.7418065071105957, + "learning_rate": 3.44753448351671e-05, + "loss": 0.3832, + "num_input_tokens_seen": 4481456, + "step": 16735 + }, + { + "epoch": 4.387942332896461, + "grad_norm": 3.4968414306640625, + "learning_rate": 3.4464759885751796e-05, + "loss": 0.3581, + "num_input_tokens_seen": 4482544, + "step": 16740 + }, + { + "epoch": 4.389252948885977, + "grad_norm": 3.4774959087371826, + "learning_rate": 3.4454172955378595e-05, + "loss": 0.3804, + "num_input_tokens_seen": 4483856, + "step": 16745 + }, + { + "epoch": 4.3905635648754915, + "grad_norm": 1.2533535957336426, + "learning_rate": 3.444358404626335e-05, + "loss": 0.5582, + "num_input_tokens_seen": 4485536, + "step": 16750 + }, + { + "epoch": 4.391874180865006, + "grad_norm": 3.224433660507202, + "learning_rate": 3.4432993160622256e-05, + "loss": 0.5267, + "num_input_tokens_seen": 4486832, + "step": 16755 + }, + { + "epoch": 4.393184796854522, + "grad_norm": 2.00844144821167, + "learning_rate": 3.4422400300672e-05, + "loss": 0.5746, + "num_input_tokens_seen": 4488272, + "step": 16760 + }, + { + "epoch": 4.394495412844036, + "grad_norm": 1.1582804918289185, + "learning_rate": 3.441180546862965e-05, + "loss": 0.2507, + "num_input_tokens_seen": 4489376, + "step": 16765 + }, + { + "epoch": 4.395806028833552, + "grad_norm": 4.162428855895996, + "learning_rate": 3.440120866671268e-05, + "loss": 0.3968, + "num_input_tokens_seen": 4490688, + "step": 16770 + }, + { + "epoch": 4.3971166448230665, + "grad_norm": 4.785272121429443, + "learning_rate": 3.4390609897138965e-05, + "loss": 0.4277, + "num_input_tokens_seen": 4491616, + "step": 16775 + }, + { + "epoch": 4.398427260812582, + "grad_norm": 3.083266496658325, + "learning_rate": 3.438000916212683e-05, + "loss": 0.4526, + "num_input_tokens_seen": 4492816, + "step": 16780 + }, + { + "epoch": 4.399737876802097, + "grad_norm": 2.727847099304199, + "learning_rate": 3.436940646389499e-05, + "loss": 0.3859, + "num_input_tokens_seen": 4493936, + "step": 16785 + }, + { + "epoch": 4.401048492791612, + "grad_norm": 1.6183972358703613, + "learning_rate": 3.435880180466257e-05, + "loss": 0.3404, + "num_input_tokens_seen": 4495264, + "step": 16790 + }, + { + "epoch": 4.402359108781127, + "grad_norm": 2.390446186065674, + "learning_rate": 3.434819518664909e-05, + "loss": 0.367, + "num_input_tokens_seen": 4496816, + "step": 16795 + }, + { + "epoch": 4.4036697247706424, + "grad_norm": 3.7074503898620605, + "learning_rate": 3.4337586612074515e-05, + "loss": 0.36, + "num_input_tokens_seen": 4497808, + "step": 16800 + }, + { + "epoch": 4.404980340760157, + "grad_norm": 2.5245697498321533, + "learning_rate": 3.4326976083159194e-05, + "loss": 0.4763, + "num_input_tokens_seen": 4498800, + "step": 16805 + }, + { + "epoch": 4.406290956749673, + "grad_norm": 1.8170832395553589, + "learning_rate": 3.43163636021239e-05, + "loss": 0.6162, + "num_input_tokens_seen": 4499984, + "step": 16810 + }, + { + "epoch": 4.407601572739187, + "grad_norm": 2.2587928771972656, + "learning_rate": 3.43057491711898e-05, + "loss": 0.5741, + "num_input_tokens_seen": 4501440, + "step": 16815 + }, + { + "epoch": 4.408912188728703, + "grad_norm": 2.5120420455932617, + "learning_rate": 3.429513279257848e-05, + "loss": 0.5087, + "num_input_tokens_seen": 4502736, + "step": 16820 + }, + { + "epoch": 4.4102228047182175, + "grad_norm": 3.1494481563568115, + "learning_rate": 3.428451446851192e-05, + "loss": 0.5077, + "num_input_tokens_seen": 4503904, + "step": 16825 + }, + { + "epoch": 4.411533420707733, + "grad_norm": 3.949517250061035, + "learning_rate": 3.427389420121253e-05, + "loss": 0.4131, + "num_input_tokens_seen": 4505040, + "step": 16830 + }, + { + "epoch": 4.412844036697248, + "grad_norm": 1.3941199779510498, + "learning_rate": 3.42632719929031e-05, + "loss": 0.4691, + "num_input_tokens_seen": 4507024, + "step": 16835 + }, + { + "epoch": 4.414154652686763, + "grad_norm": 2.3322806358337402, + "learning_rate": 3.425264784580686e-05, + "loss": 0.4439, + "num_input_tokens_seen": 4508592, + "step": 16840 + }, + { + "epoch": 4.415465268676278, + "grad_norm": 1.5422812700271606, + "learning_rate": 3.424202176214741e-05, + "loss": 0.3251, + "num_input_tokens_seen": 4509984, + "step": 16845 + }, + { + "epoch": 4.4167758846657925, + "grad_norm": 1.756988286972046, + "learning_rate": 3.423139374414877e-05, + "loss": 0.3563, + "num_input_tokens_seen": 4511728, + "step": 16850 + }, + { + "epoch": 4.418086500655308, + "grad_norm": 7.454423427581787, + "learning_rate": 3.422076379403537e-05, + "loss": 0.6471, + "num_input_tokens_seen": 4513104, + "step": 16855 + }, + { + "epoch": 4.419397116644823, + "grad_norm": 2.2167534828186035, + "learning_rate": 3.421013191403203e-05, + "loss": 0.4277, + "num_input_tokens_seen": 4515152, + "step": 16860 + }, + { + "epoch": 4.420707732634338, + "grad_norm": 1.541232943534851, + "learning_rate": 3.419949810636401e-05, + "loss": 0.5097, + "num_input_tokens_seen": 4516528, + "step": 16865 + }, + { + "epoch": 4.422018348623853, + "grad_norm": 1.7279467582702637, + "learning_rate": 3.418886237325693e-05, + "loss": 0.4766, + "num_input_tokens_seen": 4517600, + "step": 16870 + }, + { + "epoch": 4.423328964613368, + "grad_norm": 2.261981248855591, + "learning_rate": 3.417822471693683e-05, + "loss": 0.578, + "num_input_tokens_seen": 4518736, + "step": 16875 + }, + { + "epoch": 4.424639580602883, + "grad_norm": 4.264520168304443, + "learning_rate": 3.4167585139630156e-05, + "loss": 0.5037, + "num_input_tokens_seen": 4519792, + "step": 16880 + }, + { + "epoch": 4.425950196592399, + "grad_norm": 1.5458613634109497, + "learning_rate": 3.415694364356375e-05, + "loss": 0.2184, + "num_input_tokens_seen": 4521120, + "step": 16885 + }, + { + "epoch": 4.427260812581913, + "grad_norm": 3.3095955848693848, + "learning_rate": 3.414630023096486e-05, + "loss": 0.4678, + "num_input_tokens_seen": 4522208, + "step": 16890 + }, + { + "epoch": 4.428571428571429, + "grad_norm": 3.662733793258667, + "learning_rate": 3.413565490406114e-05, + "loss": 0.6215, + "num_input_tokens_seen": 4523488, + "step": 16895 + }, + { + "epoch": 4.4298820445609435, + "grad_norm": 0.7257294654846191, + "learning_rate": 3.412500766508063e-05, + "loss": 0.5127, + "num_input_tokens_seen": 4525072, + "step": 16900 + }, + { + "epoch": 4.431192660550459, + "grad_norm": 2.7536981105804443, + "learning_rate": 3.411435851625179e-05, + "loss": 0.526, + "num_input_tokens_seen": 4526224, + "step": 16905 + }, + { + "epoch": 4.432503276539974, + "grad_norm": 2.3716776371002197, + "learning_rate": 3.4103707459803456e-05, + "loss": 0.3604, + "num_input_tokens_seen": 4527408, + "step": 16910 + }, + { + "epoch": 4.433813892529489, + "grad_norm": 3.3310422897338867, + "learning_rate": 3.4093054497964884e-05, + "loss": 0.6164, + "num_input_tokens_seen": 4528400, + "step": 16915 + }, + { + "epoch": 4.435124508519004, + "grad_norm": 2.0843346118927, + "learning_rate": 3.40823996329657e-05, + "loss": 0.4309, + "num_input_tokens_seen": 4529872, + "step": 16920 + }, + { + "epoch": 4.436435124508519, + "grad_norm": 2.0595784187316895, + "learning_rate": 3.4071742867035985e-05, + "loss": 0.4708, + "num_input_tokens_seen": 4531152, + "step": 16925 + }, + { + "epoch": 4.437745740498034, + "grad_norm": 1.9158910512924194, + "learning_rate": 3.4061084202406154e-05, + "loss": 0.6228, + "num_input_tokens_seen": 4532256, + "step": 16930 + }, + { + "epoch": 4.43905635648755, + "grad_norm": 4.335877895355225, + "learning_rate": 3.405042364130706e-05, + "loss": 0.4233, + "num_input_tokens_seen": 4533488, + "step": 16935 + }, + { + "epoch": 4.440366972477064, + "grad_norm": 1.9142975807189941, + "learning_rate": 3.403976118596994e-05, + "loss": 0.426, + "num_input_tokens_seen": 4535024, + "step": 16940 + }, + { + "epoch": 4.441677588466579, + "grad_norm": 9.663172721862793, + "learning_rate": 3.4029096838626415e-05, + "loss": 0.5178, + "num_input_tokens_seen": 4535936, + "step": 16945 + }, + { + "epoch": 4.442988204456094, + "grad_norm": 1.9128750562667847, + "learning_rate": 3.401843060150853e-05, + "loss": 0.6921, + "num_input_tokens_seen": 4537488, + "step": 16950 + }, + { + "epoch": 4.444298820445609, + "grad_norm": 1.814889669418335, + "learning_rate": 3.400776247684869e-05, + "loss": 0.5054, + "num_input_tokens_seen": 4538880, + "step": 16955 + }, + { + "epoch": 4.445609436435125, + "grad_norm": 2.8250651359558105, + "learning_rate": 3.399709246687974e-05, + "loss": 0.3066, + "num_input_tokens_seen": 4539872, + "step": 16960 + }, + { + "epoch": 4.446920052424639, + "grad_norm": 3.174685478210449, + "learning_rate": 3.398642057383488e-05, + "loss": 0.6363, + "num_input_tokens_seen": 4540928, + "step": 16965 + }, + { + "epoch": 4.448230668414155, + "grad_norm": 2.4295294284820557, + "learning_rate": 3.397574679994772e-05, + "loss": 0.3417, + "num_input_tokens_seen": 4542096, + "step": 16970 + }, + { + "epoch": 4.4495412844036695, + "grad_norm": 2.8905029296875, + "learning_rate": 3.396507114745226e-05, + "loss": 0.4773, + "num_input_tokens_seen": 4543584, + "step": 16975 + }, + { + "epoch": 4.450851900393185, + "grad_norm": 7.234783172607422, + "learning_rate": 3.3954393618582905e-05, + "loss": 0.4128, + "num_input_tokens_seen": 4544848, + "step": 16980 + }, + { + "epoch": 4.4521625163827, + "grad_norm": 5.905637264251709, + "learning_rate": 3.394371421557444e-05, + "loss": 0.4608, + "num_input_tokens_seen": 4546032, + "step": 16985 + }, + { + "epoch": 4.453473132372215, + "grad_norm": 6.2284040451049805, + "learning_rate": 3.3933032940662034e-05, + "loss": 0.8297, + "num_input_tokens_seen": 4547104, + "step": 16990 + }, + { + "epoch": 4.45478374836173, + "grad_norm": 3.6868677139282227, + "learning_rate": 3.392234979608128e-05, + "loss": 0.5418, + "num_input_tokens_seen": 4548224, + "step": 16995 + }, + { + "epoch": 4.456094364351245, + "grad_norm": 2.7326245307922363, + "learning_rate": 3.391166478406812e-05, + "loss": 0.3853, + "num_input_tokens_seen": 4549792, + "step": 17000 + }, + { + "epoch": 4.45740498034076, + "grad_norm": 1.8734828233718872, + "learning_rate": 3.390097790685892e-05, + "loss": 0.353, + "num_input_tokens_seen": 4551696, + "step": 17005 + }, + { + "epoch": 4.458715596330276, + "grad_norm": 2.400662660598755, + "learning_rate": 3.3890289166690425e-05, + "loss": 0.4881, + "num_input_tokens_seen": 4553248, + "step": 17010 + }, + { + "epoch": 4.46002621231979, + "grad_norm": 2.910292863845825, + "learning_rate": 3.387959856579977e-05, + "loss": 0.6604, + "num_input_tokens_seen": 4554272, + "step": 17015 + }, + { + "epoch": 4.461336828309306, + "grad_norm": 1.597212553024292, + "learning_rate": 3.386890610642448e-05, + "loss": 0.4096, + "num_input_tokens_seen": 4555488, + "step": 17020 + }, + { + "epoch": 4.46264744429882, + "grad_norm": 2.2232861518859863, + "learning_rate": 3.3858211790802466e-05, + "loss": 0.3459, + "num_input_tokens_seen": 4556608, + "step": 17025 + }, + { + "epoch": 4.463958060288336, + "grad_norm": 6.853398323059082, + "learning_rate": 3.384751562117203e-05, + "loss": 0.3774, + "num_input_tokens_seen": 4558288, + "step": 17030 + }, + { + "epoch": 4.465268676277851, + "grad_norm": 2.5960965156555176, + "learning_rate": 3.3836817599771854e-05, + "loss": 0.4349, + "num_input_tokens_seen": 4559504, + "step": 17035 + }, + { + "epoch": 4.466579292267365, + "grad_norm": 2.461695909500122, + "learning_rate": 3.382611772884103e-05, + "loss": 0.4824, + "num_input_tokens_seen": 4561136, + "step": 17040 + }, + { + "epoch": 4.467889908256881, + "grad_norm": 1.9760228395462036, + "learning_rate": 3.3815416010619013e-05, + "loss": 0.5074, + "num_input_tokens_seen": 4562496, + "step": 17045 + }, + { + "epoch": 4.4692005242463955, + "grad_norm": 2.267754316329956, + "learning_rate": 3.380471244734566e-05, + "loss": 0.8607, + "num_input_tokens_seen": 4563904, + "step": 17050 + }, + { + "epoch": 4.470511140235911, + "grad_norm": 1.949327826499939, + "learning_rate": 3.3794007041261205e-05, + "loss": 0.4998, + "num_input_tokens_seen": 4565120, + "step": 17055 + }, + { + "epoch": 4.471821756225426, + "grad_norm": 2.271176338195801, + "learning_rate": 3.378329979460627e-05, + "loss": 0.4997, + "num_input_tokens_seen": 4566320, + "step": 17060 + }, + { + "epoch": 4.473132372214941, + "grad_norm": 1.433983564376831, + "learning_rate": 3.3772590709621864e-05, + "loss": 0.3896, + "num_input_tokens_seen": 4567504, + "step": 17065 + }, + { + "epoch": 4.474442988204456, + "grad_norm": 1.1043493747711182, + "learning_rate": 3.376187978854938e-05, + "loss": 0.6516, + "num_input_tokens_seen": 4569184, + "step": 17070 + }, + { + "epoch": 4.475753604193971, + "grad_norm": 1.9033021926879883, + "learning_rate": 3.3751167033630596e-05, + "loss": 0.6261, + "num_input_tokens_seen": 4570176, + "step": 17075 + }, + { + "epoch": 4.477064220183486, + "grad_norm": 1.369865894317627, + "learning_rate": 3.374045244710767e-05, + "loss": 0.4629, + "num_input_tokens_seen": 4571408, + "step": 17080 + }, + { + "epoch": 4.478374836173002, + "grad_norm": 1.8318809270858765, + "learning_rate": 3.372973603122316e-05, + "loss": 0.4442, + "num_input_tokens_seen": 4572576, + "step": 17085 + }, + { + "epoch": 4.479685452162516, + "grad_norm": 2.3251092433929443, + "learning_rate": 3.371901778821997e-05, + "loss": 0.6382, + "num_input_tokens_seen": 4574096, + "step": 17090 + }, + { + "epoch": 4.480996068152032, + "grad_norm": 2.953695297241211, + "learning_rate": 3.3708297720341425e-05, + "loss": 0.7072, + "num_input_tokens_seen": 4575120, + "step": 17095 + }, + { + "epoch": 4.482306684141546, + "grad_norm": 2.2733957767486572, + "learning_rate": 3.369757582983122e-05, + "loss": 0.3168, + "num_input_tokens_seen": 4576640, + "step": 17100 + }, + { + "epoch": 4.483617300131062, + "grad_norm": 1.863789439201355, + "learning_rate": 3.368685211893342e-05, + "loss": 0.5052, + "num_input_tokens_seen": 4578160, + "step": 17105 + }, + { + "epoch": 4.484927916120577, + "grad_norm": 2.4104387760162354, + "learning_rate": 3.367612658989247e-05, + "loss": 0.5788, + "num_input_tokens_seen": 4580576, + "step": 17110 + }, + { + "epoch": 4.486238532110092, + "grad_norm": 3.4703712463378906, + "learning_rate": 3.3665399244953236e-05, + "loss": 0.5892, + "num_input_tokens_seen": 4581840, + "step": 17115 + }, + { + "epoch": 4.487549148099607, + "grad_norm": 3.2491343021392822, + "learning_rate": 3.36546700863609e-05, + "loss": 0.4638, + "num_input_tokens_seen": 4582880, + "step": 17120 + }, + { + "epoch": 4.488859764089122, + "grad_norm": 3.210892677307129, + "learning_rate": 3.364393911636107e-05, + "loss": 0.5195, + "num_input_tokens_seen": 4583920, + "step": 17125 + }, + { + "epoch": 4.490170380078637, + "grad_norm": 1.6289547681808472, + "learning_rate": 3.363320633719972e-05, + "loss": 0.4251, + "num_input_tokens_seen": 4584992, + "step": 17130 + }, + { + "epoch": 4.491480996068152, + "grad_norm": 10.696221351623535, + "learning_rate": 3.3622471751123196e-05, + "loss": 0.2774, + "num_input_tokens_seen": 4586368, + "step": 17135 + }, + { + "epoch": 4.492791612057667, + "grad_norm": 2.6117777824401855, + "learning_rate": 3.361173536037824e-05, + "loss": 0.4605, + "num_input_tokens_seen": 4587264, + "step": 17140 + }, + { + "epoch": 4.494102228047182, + "grad_norm": 3.2012007236480713, + "learning_rate": 3.360099716721195e-05, + "loss": 0.3976, + "num_input_tokens_seen": 4588416, + "step": 17145 + }, + { + "epoch": 4.495412844036697, + "grad_norm": 1.7735042572021484, + "learning_rate": 3.3590257173871807e-05, + "loss": 0.4861, + "num_input_tokens_seen": 4589536, + "step": 17150 + }, + { + "epoch": 4.496723460026212, + "grad_norm": 3.0888617038726807, + "learning_rate": 3.357951538260569e-05, + "loss": 0.5394, + "num_input_tokens_seen": 4590736, + "step": 17155 + }, + { + "epoch": 4.498034076015728, + "grad_norm": 3.8832976818084717, + "learning_rate": 3.356877179566183e-05, + "loss": 0.7871, + "num_input_tokens_seen": 4592176, + "step": 17160 + }, + { + "epoch": 4.499344692005242, + "grad_norm": 1.2261083126068115, + "learning_rate": 3.355802641528883e-05, + "loss": 0.4946, + "num_input_tokens_seen": 4593616, + "step": 17165 + }, + { + "epoch": 4.500655307994758, + "grad_norm": 3.424750804901123, + "learning_rate": 3.3547279243735695e-05, + "loss": 0.5938, + "num_input_tokens_seen": 4594688, + "step": 17170 + }, + { + "epoch": 4.501179554390563, + "eval_loss": 0.46691471338272095, + "eval_runtime": 18.2025, + "eval_samples_per_second": 46.587, + "eval_steps_per_second": 23.293, + "num_input_tokens_seen": 4595520, + "step": 17172 + }, + { + "epoch": 4.501965923984272, + "grad_norm": 2.7251181602478027, + "learning_rate": 3.3536530283251784e-05, + "loss": 0.3855, + "num_input_tokens_seen": 4596288, + "step": 17175 + }, + { + "epoch": 4.503276539973788, + "grad_norm": 2.350126028060913, + "learning_rate": 3.3525779536086826e-05, + "loss": 0.3994, + "num_input_tokens_seen": 4597552, + "step": 17180 + }, + { + "epoch": 4.504587155963303, + "grad_norm": 1.421417474746704, + "learning_rate": 3.351502700449095e-05, + "loss": 0.7036, + "num_input_tokens_seen": 4598896, + "step": 17185 + }, + { + "epoch": 4.505897771952818, + "grad_norm": 2.566842555999756, + "learning_rate": 3.350427269071463e-05, + "loss": 0.5236, + "num_input_tokens_seen": 4600208, + "step": 17190 + }, + { + "epoch": 4.507208387942333, + "grad_norm": 3.0784575939178467, + "learning_rate": 3.3493516597008734e-05, + "loss": 0.5577, + "num_input_tokens_seen": 4601440, + "step": 17195 + }, + { + "epoch": 4.508519003931848, + "grad_norm": 1.8006294965744019, + "learning_rate": 3.348275872562449e-05, + "loss": 0.4525, + "num_input_tokens_seen": 4603264, + "step": 17200 + }, + { + "epoch": 4.509829619921363, + "grad_norm": 10.365370750427246, + "learning_rate": 3.34719990788135e-05, + "loss": 0.4061, + "num_input_tokens_seen": 4604144, + "step": 17205 + }, + { + "epoch": 4.511140235910878, + "grad_norm": 8.045939445495605, + "learning_rate": 3.346123765882774e-05, + "loss": 0.4113, + "num_input_tokens_seen": 4605456, + "step": 17210 + }, + { + "epoch": 4.512450851900393, + "grad_norm": 4.731143951416016, + "learning_rate": 3.345047446791956e-05, + "loss": 0.5313, + "num_input_tokens_seen": 4606544, + "step": 17215 + }, + { + "epoch": 4.513761467889909, + "grad_norm": 1.3282408714294434, + "learning_rate": 3.3439709508341676e-05, + "loss": 0.415, + "num_input_tokens_seen": 4607632, + "step": 17220 + }, + { + "epoch": 4.515072083879423, + "grad_norm": 1.6482821702957153, + "learning_rate": 3.342894278234717e-05, + "loss": 0.5305, + "num_input_tokens_seen": 4609232, + "step": 17225 + }, + { + "epoch": 4.516382699868938, + "grad_norm": 3.6561174392700195, + "learning_rate": 3.341817429218951e-05, + "loss": 0.3088, + "num_input_tokens_seen": 4610736, + "step": 17230 + }, + { + "epoch": 4.517693315858454, + "grad_norm": 2.9261274337768555, + "learning_rate": 3.340740404012251e-05, + "loss": 0.4658, + "num_input_tokens_seen": 4612800, + "step": 17235 + }, + { + "epoch": 4.519003931847968, + "grad_norm": 2.236300468444824, + "learning_rate": 3.339663202840038e-05, + "loss": 0.4698, + "num_input_tokens_seen": 4614176, + "step": 17240 + }, + { + "epoch": 4.520314547837484, + "grad_norm": 1.5183296203613281, + "learning_rate": 3.3385858259277666e-05, + "loss": 0.4056, + "num_input_tokens_seen": 4615456, + "step": 17245 + }, + { + "epoch": 4.521625163826998, + "grad_norm": 2.2466790676116943, + "learning_rate": 3.337508273500931e-05, + "loss": 0.3931, + "num_input_tokens_seen": 4616592, + "step": 17250 + }, + { + "epoch": 4.522935779816514, + "grad_norm": 1.6765334606170654, + "learning_rate": 3.336430545785061e-05, + "loss": 0.4983, + "num_input_tokens_seen": 4617968, + "step": 17255 + }, + { + "epoch": 4.524246395806029, + "grad_norm": 1.4058171510696411, + "learning_rate": 3.335352643005723e-05, + "loss": 0.7373, + "num_input_tokens_seen": 4619456, + "step": 17260 + }, + { + "epoch": 4.525557011795544, + "grad_norm": 5.064301013946533, + "learning_rate": 3.33427456538852e-05, + "loss": 0.3955, + "num_input_tokens_seen": 4620784, + "step": 17265 + }, + { + "epoch": 4.526867627785059, + "grad_norm": 1.6071637868881226, + "learning_rate": 3.3331963131590916e-05, + "loss": 0.6254, + "num_input_tokens_seen": 4622560, + "step": 17270 + }, + { + "epoch": 4.528178243774574, + "grad_norm": 3.023360013961792, + "learning_rate": 3.3321178865431155e-05, + "loss": 0.4128, + "num_input_tokens_seen": 4623856, + "step": 17275 + }, + { + "epoch": 4.529488859764089, + "grad_norm": 3.1054587364196777, + "learning_rate": 3.331039285766303e-05, + "loss": 0.4297, + "num_input_tokens_seen": 4625568, + "step": 17280 + }, + { + "epoch": 4.5307994757536045, + "grad_norm": 1.3704229593276978, + "learning_rate": 3.329960511054404e-05, + "loss": 0.4457, + "num_input_tokens_seen": 4626816, + "step": 17285 + }, + { + "epoch": 4.532110091743119, + "grad_norm": 2.0058882236480713, + "learning_rate": 3.3288815626332035e-05, + "loss": 0.351, + "num_input_tokens_seen": 4629072, + "step": 17290 + }, + { + "epoch": 4.533420707732635, + "grad_norm": 5.72261905670166, + "learning_rate": 3.327802440728526e-05, + "loss": 0.5619, + "num_input_tokens_seen": 4629936, + "step": 17295 + }, + { + "epoch": 4.534731323722149, + "grad_norm": 4.2594099044799805, + "learning_rate": 3.3267231455662265e-05, + "loss": 0.4508, + "num_input_tokens_seen": 4630928, + "step": 17300 + }, + { + "epoch": 4.536041939711664, + "grad_norm": 1.644005298614502, + "learning_rate": 3.325643677372201e-05, + "loss": 0.3868, + "num_input_tokens_seen": 4632080, + "step": 17305 + }, + { + "epoch": 4.53735255570118, + "grad_norm": 25.250812530517578, + "learning_rate": 3.324564036372381e-05, + "loss": 0.6287, + "num_input_tokens_seen": 4633280, + "step": 17310 + }, + { + "epoch": 4.538663171690695, + "grad_norm": 1.6353189945220947, + "learning_rate": 3.323484222792732e-05, + "loss": 0.4764, + "num_input_tokens_seen": 4634576, + "step": 17315 + }, + { + "epoch": 4.53997378768021, + "grad_norm": 2.8982455730438232, + "learning_rate": 3.322404236859259e-05, + "loss": 0.5643, + "num_input_tokens_seen": 4635792, + "step": 17320 + }, + { + "epoch": 4.541284403669724, + "grad_norm": 5.1313018798828125, + "learning_rate": 3.321324078798e-05, + "loss": 0.4148, + "num_input_tokens_seen": 4637200, + "step": 17325 + }, + { + "epoch": 4.54259501965924, + "grad_norm": 1.5295064449310303, + "learning_rate": 3.32024374883503e-05, + "loss": 0.3701, + "num_input_tokens_seen": 4638688, + "step": 17330 + }, + { + "epoch": 4.543905635648755, + "grad_norm": 1.4826273918151855, + "learning_rate": 3.319163247196462e-05, + "loss": 0.6025, + "num_input_tokens_seen": 4639920, + "step": 17335 + }, + { + "epoch": 4.54521625163827, + "grad_norm": 3.676624298095703, + "learning_rate": 3.318082574108441e-05, + "loss": 0.6065, + "num_input_tokens_seen": 4641504, + "step": 17340 + }, + { + "epoch": 4.546526867627785, + "grad_norm": 3.1345505714416504, + "learning_rate": 3.31700172979715e-05, + "loss": 0.4894, + "num_input_tokens_seen": 4643344, + "step": 17345 + }, + { + "epoch": 4.5478374836173, + "grad_norm": 2.180575132369995, + "learning_rate": 3.315920714488808e-05, + "loss": 0.4556, + "num_input_tokens_seen": 4644528, + "step": 17350 + }, + { + "epoch": 4.549148099606815, + "grad_norm": 2.3013648986816406, + "learning_rate": 3.314839528409672e-05, + "loss": 0.3898, + "num_input_tokens_seen": 4646032, + "step": 17355 + }, + { + "epoch": 4.5504587155963305, + "grad_norm": 2.2402281761169434, + "learning_rate": 3.313758171786031e-05, + "loss": 0.6418, + "num_input_tokens_seen": 4647120, + "step": 17360 + }, + { + "epoch": 4.551769331585845, + "grad_norm": 5.057740211486816, + "learning_rate": 3.3126766448442095e-05, + "loss": 0.4866, + "num_input_tokens_seen": 4648544, + "step": 17365 + }, + { + "epoch": 4.553079947575361, + "grad_norm": 7.093415260314941, + "learning_rate": 3.311594947810571e-05, + "loss": 0.6653, + "num_input_tokens_seen": 4649488, + "step": 17370 + }, + { + "epoch": 4.554390563564875, + "grad_norm": 2.2693498134613037, + "learning_rate": 3.310513080911512e-05, + "loss": 0.4693, + "num_input_tokens_seen": 4650864, + "step": 17375 + }, + { + "epoch": 4.555701179554391, + "grad_norm": 1.7825818061828613, + "learning_rate": 3.309431044373465e-05, + "loss": 0.4793, + "num_input_tokens_seen": 4651984, + "step": 17380 + }, + { + "epoch": 4.557011795543906, + "grad_norm": 3.0080482959747314, + "learning_rate": 3.3083488384228995e-05, + "loss": 0.3764, + "num_input_tokens_seen": 4653168, + "step": 17385 + }, + { + "epoch": 4.558322411533421, + "grad_norm": 3.2361698150634766, + "learning_rate": 3.3072664632863195e-05, + "loss": 0.4224, + "num_input_tokens_seen": 4654336, + "step": 17390 + }, + { + "epoch": 4.559633027522936, + "grad_norm": 3.2331552505493164, + "learning_rate": 3.3061839191902635e-05, + "loss": 0.5285, + "num_input_tokens_seen": 4655520, + "step": 17395 + }, + { + "epoch": 4.56094364351245, + "grad_norm": 5.112798690795898, + "learning_rate": 3.3051012063613054e-05, + "loss": 0.4269, + "num_input_tokens_seen": 4656896, + "step": 17400 + }, + { + "epoch": 4.562254259501966, + "grad_norm": 1.0553150177001953, + "learning_rate": 3.3040183250260556e-05, + "loss": 0.389, + "num_input_tokens_seen": 4658240, + "step": 17405 + }, + { + "epoch": 4.5635648754914815, + "grad_norm": 1.1043448448181152, + "learning_rate": 3.30293527541116e-05, + "loss": 0.2561, + "num_input_tokens_seen": 4659456, + "step": 17410 + }, + { + "epoch": 4.564875491480996, + "grad_norm": 1.9619731903076172, + "learning_rate": 3.3018520577432996e-05, + "loss": 0.5751, + "num_input_tokens_seen": 4660800, + "step": 17415 + }, + { + "epoch": 4.566186107470511, + "grad_norm": 2.7361810207366943, + "learning_rate": 3.3007686722491885e-05, + "loss": 0.6323, + "num_input_tokens_seen": 4662224, + "step": 17420 + }, + { + "epoch": 4.567496723460026, + "grad_norm": 1.0894091129302979, + "learning_rate": 3.299685119155577e-05, + "loss": 0.5412, + "num_input_tokens_seen": 4663728, + "step": 17425 + }, + { + "epoch": 4.568807339449541, + "grad_norm": 18.678668975830078, + "learning_rate": 3.298601398689253e-05, + "loss": 0.7792, + "num_input_tokens_seen": 4664832, + "step": 17430 + }, + { + "epoch": 4.5701179554390565, + "grad_norm": 2.0495989322662354, + "learning_rate": 3.297517511077035e-05, + "loss": 0.2679, + "num_input_tokens_seen": 4665712, + "step": 17435 + }, + { + "epoch": 4.571428571428571, + "grad_norm": 2.8541007041931152, + "learning_rate": 3.296433456545781e-05, + "loss": 0.4842, + "num_input_tokens_seen": 4666896, + "step": 17440 + }, + { + "epoch": 4.572739187418087, + "grad_norm": 1.6484688520431519, + "learning_rate": 3.2953492353223805e-05, + "loss": 0.3567, + "num_input_tokens_seen": 4668240, + "step": 17445 + }, + { + "epoch": 4.574049803407601, + "grad_norm": 2.4309234619140625, + "learning_rate": 3.294264847633759e-05, + "loss": 0.5986, + "num_input_tokens_seen": 4670096, + "step": 17450 + }, + { + "epoch": 4.575360419397117, + "grad_norm": 1.7956318855285645, + "learning_rate": 3.2931802937068776e-05, + "loss": 0.5918, + "num_input_tokens_seen": 4671264, + "step": 17455 + }, + { + "epoch": 4.576671035386632, + "grad_norm": 2.734501838684082, + "learning_rate": 3.292095573768731e-05, + "loss": 0.3282, + "num_input_tokens_seen": 4672272, + "step": 17460 + }, + { + "epoch": 4.577981651376147, + "grad_norm": 1.5101007223129272, + "learning_rate": 3.2910106880463486e-05, + "loss": 0.3787, + "num_input_tokens_seen": 4673600, + "step": 17465 + }, + { + "epoch": 4.579292267365662, + "grad_norm": 3.3340041637420654, + "learning_rate": 3.289925636766797e-05, + "loss": 0.3926, + "num_input_tokens_seen": 4674848, + "step": 17470 + }, + { + "epoch": 4.580602883355177, + "grad_norm": 1.026029109954834, + "learning_rate": 3.2888404201571743e-05, + "loss": 0.32, + "num_input_tokens_seen": 4676416, + "step": 17475 + }, + { + "epoch": 4.581913499344692, + "grad_norm": 2.3871243000030518, + "learning_rate": 3.2877550384446146e-05, + "loss": 0.5053, + "num_input_tokens_seen": 4677712, + "step": 17480 + }, + { + "epoch": 4.5832241153342075, + "grad_norm": 1.5367634296417236, + "learning_rate": 3.2866694918562866e-05, + "loss": 0.4177, + "num_input_tokens_seen": 4678912, + "step": 17485 + }, + { + "epoch": 4.584534731323722, + "grad_norm": 1.9796251058578491, + "learning_rate": 3.285583780619394e-05, + "loss": 0.5325, + "num_input_tokens_seen": 4680336, + "step": 17490 + }, + { + "epoch": 4.585845347313237, + "grad_norm": 1.681483268737793, + "learning_rate": 3.284497904961172e-05, + "loss": 0.4754, + "num_input_tokens_seen": 4681664, + "step": 17495 + }, + { + "epoch": 4.587155963302752, + "grad_norm": 3.8389108180999756, + "learning_rate": 3.283411865108894e-05, + "loss": 0.3866, + "num_input_tokens_seen": 4682896, + "step": 17500 + }, + { + "epoch": 4.588466579292267, + "grad_norm": 2.6651620864868164, + "learning_rate": 3.2823256612898676e-05, + "loss": 0.5634, + "num_input_tokens_seen": 4684064, + "step": 17505 + }, + { + "epoch": 4.5897771952817825, + "grad_norm": 3.9023823738098145, + "learning_rate": 3.2812392937314316e-05, + "loss": 0.386, + "num_input_tokens_seen": 4685248, + "step": 17510 + }, + { + "epoch": 4.591087811271297, + "grad_norm": 4.681086540222168, + "learning_rate": 3.2801527626609606e-05, + "loss": 0.7921, + "num_input_tokens_seen": 4686992, + "step": 17515 + }, + { + "epoch": 4.592398427260813, + "grad_norm": 2.801811933517456, + "learning_rate": 3.2790660683058646e-05, + "loss": 0.9588, + "num_input_tokens_seen": 4688112, + "step": 17520 + }, + { + "epoch": 4.593709043250327, + "grad_norm": 1.78270423412323, + "learning_rate": 3.277979210893587e-05, + "loss": 0.388, + "num_input_tokens_seen": 4689808, + "step": 17525 + }, + { + "epoch": 4.595019659239843, + "grad_norm": 1.327563762664795, + "learning_rate": 3.276892190651605e-05, + "loss": 0.4819, + "num_input_tokens_seen": 4690864, + "step": 17530 + }, + { + "epoch": 4.5963302752293576, + "grad_norm": 3.5798747539520264, + "learning_rate": 3.275805007807429e-05, + "loss": 0.4182, + "num_input_tokens_seen": 4692736, + "step": 17535 + }, + { + "epoch": 4.597640891218873, + "grad_norm": 2.062272071838379, + "learning_rate": 3.274717662588606e-05, + "loss": 0.4585, + "num_input_tokens_seen": 4693648, + "step": 17540 + }, + { + "epoch": 4.598951507208388, + "grad_norm": 4.855706691741943, + "learning_rate": 3.273630155222715e-05, + "loss": 0.8025, + "num_input_tokens_seen": 4694736, + "step": 17545 + }, + { + "epoch": 4.600262123197903, + "grad_norm": 2.7177839279174805, + "learning_rate": 3.272542485937369e-05, + "loss": 0.5377, + "num_input_tokens_seen": 4696240, + "step": 17550 + }, + { + "epoch": 4.601572739187418, + "grad_norm": 2.634727954864502, + "learning_rate": 3.271454654960215e-05, + "loss": 0.6041, + "num_input_tokens_seen": 4697360, + "step": 17555 + }, + { + "epoch": 4.6028833551769335, + "grad_norm": 1.9012656211853027, + "learning_rate": 3.270366662518936e-05, + "loss": 0.5466, + "num_input_tokens_seen": 4698880, + "step": 17560 + }, + { + "epoch": 4.604193971166448, + "grad_norm": 2.8706576824188232, + "learning_rate": 3.2692785088412444e-05, + "loss": 0.5221, + "num_input_tokens_seen": 4700368, + "step": 17565 + }, + { + "epoch": 4.605504587155964, + "grad_norm": 3.999077081680298, + "learning_rate": 3.268190194154891e-05, + "loss": 0.4421, + "num_input_tokens_seen": 4701824, + "step": 17570 + }, + { + "epoch": 4.606815203145478, + "grad_norm": 5.124202728271484, + "learning_rate": 3.267101718687657e-05, + "loss": 0.4632, + "num_input_tokens_seen": 4702832, + "step": 17575 + }, + { + "epoch": 4.608125819134994, + "grad_norm": 3.5937020778656006, + "learning_rate": 3.266013082667357e-05, + "loss": 0.3917, + "num_input_tokens_seen": 4704480, + "step": 17580 + }, + { + "epoch": 4.6094364351245085, + "grad_norm": 4.432927131652832, + "learning_rate": 3.2649242863218446e-05, + "loss": 0.3324, + "num_input_tokens_seen": 4706496, + "step": 17585 + }, + { + "epoch": 4.610747051114023, + "grad_norm": 3.726027011871338, + "learning_rate": 3.2638353298790005e-05, + "loss": 0.4215, + "num_input_tokens_seen": 4707536, + "step": 17590 + }, + { + "epoch": 4.612057667103539, + "grad_norm": 4.507038593292236, + "learning_rate": 3.2627462135667405e-05, + "loss": 0.3464, + "num_input_tokens_seen": 4708672, + "step": 17595 + }, + { + "epoch": 4.613368283093053, + "grad_norm": 2.431084156036377, + "learning_rate": 3.2616569376130175e-05, + "loss": 0.3598, + "num_input_tokens_seen": 4709872, + "step": 17600 + }, + { + "epoch": 4.614678899082569, + "grad_norm": 2.8054795265197754, + "learning_rate": 3.260567502245813e-05, + "loss": 0.4001, + "num_input_tokens_seen": 4711040, + "step": 17605 + }, + { + "epoch": 4.6159895150720835, + "grad_norm": 1.1433181762695312, + "learning_rate": 3.259477907693144e-05, + "loss": 0.6594, + "num_input_tokens_seen": 4712640, + "step": 17610 + }, + { + "epoch": 4.617300131061599, + "grad_norm": 2.722831964492798, + "learning_rate": 3.2583881541830604e-05, + "loss": 0.5287, + "num_input_tokens_seen": 4713728, + "step": 17615 + }, + { + "epoch": 4.618610747051114, + "grad_norm": 3.475226402282715, + "learning_rate": 3.257298241943648e-05, + "loss": 0.5554, + "num_input_tokens_seen": 4715456, + "step": 17620 + }, + { + "epoch": 4.619921363040629, + "grad_norm": 2.1030170917510986, + "learning_rate": 3.256208171203021e-05, + "loss": 0.5557, + "num_input_tokens_seen": 4716560, + "step": 17625 + }, + { + "epoch": 4.621231979030144, + "grad_norm": 1.8575470447540283, + "learning_rate": 3.255117942189331e-05, + "loss": 0.537, + "num_input_tokens_seen": 4717968, + "step": 17630 + }, + { + "epoch": 4.6225425950196595, + "grad_norm": 1.7957600355148315, + "learning_rate": 3.2540275551307605e-05, + "loss": 0.3715, + "num_input_tokens_seen": 4718992, + "step": 17635 + }, + { + "epoch": 4.623853211009174, + "grad_norm": 12.158308029174805, + "learning_rate": 3.252937010255525e-05, + "loss": 0.2873, + "num_input_tokens_seen": 4720320, + "step": 17640 + }, + { + "epoch": 4.62516382699869, + "grad_norm": 2.005871057510376, + "learning_rate": 3.2518463077918746e-05, + "loss": 0.4199, + "num_input_tokens_seen": 4721440, + "step": 17645 + }, + { + "epoch": 4.626474442988204, + "grad_norm": 5.940285682678223, + "learning_rate": 3.250755447968091e-05, + "loss": 0.5445, + "num_input_tokens_seen": 4722576, + "step": 17650 + }, + { + "epoch": 4.62778505897772, + "grad_norm": 3.8042099475860596, + "learning_rate": 3.24966443101249e-05, + "loss": 0.5005, + "num_input_tokens_seen": 4723760, + "step": 17655 + }, + { + "epoch": 4.6290956749672345, + "grad_norm": 5.112757682800293, + "learning_rate": 3.248573257153418e-05, + "loss": 0.6432, + "num_input_tokens_seen": 4725120, + "step": 17660 + }, + { + "epoch": 4.63040629095675, + "grad_norm": 2.5805611610412598, + "learning_rate": 3.2474819266192574e-05, + "loss": 0.5459, + "num_input_tokens_seen": 4726288, + "step": 17665 + }, + { + "epoch": 4.631716906946265, + "grad_norm": 8.552793502807617, + "learning_rate": 3.24639043963842e-05, + "loss": 0.5045, + "num_input_tokens_seen": 4727632, + "step": 17670 + }, + { + "epoch": 4.63302752293578, + "grad_norm": 2.173290491104126, + "learning_rate": 3.245298796439354e-05, + "loss": 0.4785, + "num_input_tokens_seen": 4728944, + "step": 17675 + }, + { + "epoch": 4.634338138925295, + "grad_norm": 2.9382171630859375, + "learning_rate": 3.244206997250538e-05, + "loss": 0.636, + "num_input_tokens_seen": 4729888, + "step": 17680 + }, + { + "epoch": 4.6356487549148095, + "grad_norm": 4.591835021972656, + "learning_rate": 3.243115042300483e-05, + "loss": 0.4552, + "num_input_tokens_seen": 4731360, + "step": 17685 + }, + { + "epoch": 4.636959370904325, + "grad_norm": 2.164822578430176, + "learning_rate": 3.242022931817734e-05, + "loss": 0.3711, + "num_input_tokens_seen": 4732880, + "step": 17690 + }, + { + "epoch": 4.63826998689384, + "grad_norm": 18.37268829345703, + "learning_rate": 3.240930666030867e-05, + "loss": 0.5058, + "num_input_tokens_seen": 4733824, + "step": 17695 + }, + { + "epoch": 4.639580602883355, + "grad_norm": 1.5830745697021484, + "learning_rate": 3.239838245168492e-05, + "loss": 0.4151, + "num_input_tokens_seen": 4735232, + "step": 17700 + }, + { + "epoch": 4.64089121887287, + "grad_norm": 2.404184341430664, + "learning_rate": 3.238745669459251e-05, + "loss": 0.4145, + "num_input_tokens_seen": 4736768, + "step": 17705 + }, + { + "epoch": 4.6422018348623855, + "grad_norm": 2.2028932571411133, + "learning_rate": 3.237652939131818e-05, + "loss": 0.5325, + "num_input_tokens_seen": 4738112, + "step": 17710 + }, + { + "epoch": 4.6435124508519, + "grad_norm": 6.196743965148926, + "learning_rate": 3.236560054414899e-05, + "loss": 0.5558, + "num_input_tokens_seen": 4739296, + "step": 17715 + }, + { + "epoch": 4.644823066841416, + "grad_norm": 1.374284267425537, + "learning_rate": 3.2354670155372326e-05, + "loss": 0.4241, + "num_input_tokens_seen": 4740528, + "step": 17720 + }, + { + "epoch": 4.64613368283093, + "grad_norm": 4.472676753997803, + "learning_rate": 3.234373822727591e-05, + "loss": 0.6118, + "num_input_tokens_seen": 4742016, + "step": 17725 + }, + { + "epoch": 4.647444298820446, + "grad_norm": 1.9020124673843384, + "learning_rate": 3.233280476214778e-05, + "loss": 0.4785, + "num_input_tokens_seen": 4743248, + "step": 17730 + }, + { + "epoch": 4.6487549148099605, + "grad_norm": 2.166980743408203, + "learning_rate": 3.232186976227628e-05, + "loss": 0.5611, + "num_input_tokens_seen": 4744848, + "step": 17735 + }, + { + "epoch": 4.650065530799476, + "grad_norm": 3.4261679649353027, + "learning_rate": 3.231093322995008e-05, + "loss": 0.5113, + "num_input_tokens_seen": 4746080, + "step": 17740 + }, + { + "epoch": 4.651376146788991, + "grad_norm": 1.790570855140686, + "learning_rate": 3.2299995167458196e-05, + "loss": 0.4093, + "num_input_tokens_seen": 4747616, + "step": 17745 + }, + { + "epoch": 4.652686762778506, + "grad_norm": 6.9264373779296875, + "learning_rate": 3.2289055577089936e-05, + "loss": 0.4516, + "num_input_tokens_seen": 4749040, + "step": 17750 + }, + { + "epoch": 4.653997378768021, + "grad_norm": 1.7117482423782349, + "learning_rate": 3.227811446113494e-05, + "loss": 0.7473, + "num_input_tokens_seen": 4750464, + "step": 17755 + }, + { + "epoch": 4.655307994757536, + "grad_norm": 1.66000497341156, + "learning_rate": 3.226717182188316e-05, + "loss": 0.4043, + "num_input_tokens_seen": 4751552, + "step": 17760 + }, + { + "epoch": 4.656618610747051, + "grad_norm": 9.196859359741211, + "learning_rate": 3.2256227661624874e-05, + "loss": 0.4606, + "num_input_tokens_seen": 4753200, + "step": 17765 + }, + { + "epoch": 4.657929226736567, + "grad_norm": 2.0656769275665283, + "learning_rate": 3.2245281982650683e-05, + "loss": 1.3815, + "num_input_tokens_seen": 4754688, + "step": 17770 + }, + { + "epoch": 4.659239842726081, + "grad_norm": 5.080794334411621, + "learning_rate": 3.2234334787251487e-05, + "loss": 0.4918, + "num_input_tokens_seen": 4755680, + "step": 17775 + }, + { + "epoch": 4.660550458715596, + "grad_norm": 9.40517807006836, + "learning_rate": 3.222338607771851e-05, + "loss": 0.5247, + "num_input_tokens_seen": 4756560, + "step": 17780 + }, + { + "epoch": 4.6618610747051115, + "grad_norm": 2.244342803955078, + "learning_rate": 3.221243585634333e-05, + "loss": 0.5044, + "num_input_tokens_seen": 4757632, + "step": 17785 + }, + { + "epoch": 4.663171690694626, + "grad_norm": 1.2117762565612793, + "learning_rate": 3.2201484125417766e-05, + "loss": 0.2525, + "num_input_tokens_seen": 4758752, + "step": 17790 + }, + { + "epoch": 4.664482306684142, + "grad_norm": 3.4011497497558594, + "learning_rate": 3.219053088723403e-05, + "loss": 0.3304, + "num_input_tokens_seen": 4759872, + "step": 17795 + }, + { + "epoch": 4.665792922673656, + "grad_norm": 1.187624216079712, + "learning_rate": 3.21795761440846e-05, + "loss": 0.3987, + "num_input_tokens_seen": 4761344, + "step": 17800 + }, + { + "epoch": 4.667103538663172, + "grad_norm": 2.1258342266082764, + "learning_rate": 3.2168619898262294e-05, + "loss": 0.4213, + "num_input_tokens_seen": 4762880, + "step": 17805 + }, + { + "epoch": 4.6684141546526865, + "grad_norm": 6.72589111328125, + "learning_rate": 3.215766215206022e-05, + "loss": 0.6239, + "num_input_tokens_seen": 4764144, + "step": 17810 + }, + { + "epoch": 4.669724770642202, + "grad_norm": 5.021882057189941, + "learning_rate": 3.2146702907771824e-05, + "loss": 0.5535, + "num_input_tokens_seen": 4765312, + "step": 17815 + }, + { + "epoch": 4.671035386631717, + "grad_norm": 1.5776777267456055, + "learning_rate": 3.2135742167690866e-05, + "loss": 0.4041, + "num_input_tokens_seen": 4766704, + "step": 17820 + }, + { + "epoch": 4.672346002621232, + "grad_norm": 4.051356792449951, + "learning_rate": 3.212477993411141e-05, + "loss": 0.5381, + "num_input_tokens_seen": 4768080, + "step": 17825 + }, + { + "epoch": 4.673656618610747, + "grad_norm": 1.36646568775177, + "learning_rate": 3.211381620932781e-05, + "loss": 0.4373, + "num_input_tokens_seen": 4769312, + "step": 17830 + }, + { + "epoch": 4.674967234600262, + "grad_norm": 3.53678560256958, + "learning_rate": 3.2102850995634765e-05, + "loss": 0.3469, + "num_input_tokens_seen": 4770352, + "step": 17835 + }, + { + "epoch": 4.676277850589777, + "grad_norm": 3.0833725929260254, + "learning_rate": 3.209188429532729e-05, + "loss": 0.542, + "num_input_tokens_seen": 4771520, + "step": 17840 + }, + { + "epoch": 4.677588466579293, + "grad_norm": 2.292370080947876, + "learning_rate": 3.2080916110700684e-05, + "loss": 0.6009, + "num_input_tokens_seen": 4772736, + "step": 17845 + }, + { + "epoch": 4.678899082568807, + "grad_norm": 2.7332420349121094, + "learning_rate": 3.206994644405057e-05, + "loss": 0.5134, + "num_input_tokens_seen": 4774576, + "step": 17850 + }, + { + "epoch": 4.680209698558322, + "grad_norm": 2.1062815189361572, + "learning_rate": 3.2058975297672874e-05, + "loss": 0.5563, + "num_input_tokens_seen": 4775968, + "step": 17855 + }, + { + "epoch": 4.6815203145478375, + "grad_norm": 6.494403839111328, + "learning_rate": 3.204800267386385e-05, + "loss": 0.3522, + "num_input_tokens_seen": 4777280, + "step": 17860 + }, + { + "epoch": 4.682830930537353, + "grad_norm": 0.998525857925415, + "learning_rate": 3.203702857492005e-05, + "loss": 0.8323, + "num_input_tokens_seen": 4779040, + "step": 17865 + }, + { + "epoch": 4.684141546526868, + "grad_norm": 3.732203245162964, + "learning_rate": 3.202605300313833e-05, + "loss": 0.4353, + "num_input_tokens_seen": 4780384, + "step": 17870 + }, + { + "epoch": 4.685452162516382, + "grad_norm": 1.3991804122924805, + "learning_rate": 3.201507596081585e-05, + "loss": 0.3461, + "num_input_tokens_seen": 4781520, + "step": 17875 + }, + { + "epoch": 4.686762778505898, + "grad_norm": 2.1104249954223633, + "learning_rate": 3.200409745025009e-05, + "loss": 0.418, + "num_input_tokens_seen": 4782976, + "step": 17880 + }, + { + "epoch": 4.6880733944954125, + "grad_norm": 1.6505874395370483, + "learning_rate": 3.199311747373885e-05, + "loss": 0.4423, + "num_input_tokens_seen": 4784608, + "step": 17885 + }, + { + "epoch": 4.689384010484928, + "grad_norm": 4.343260288238525, + "learning_rate": 3.198213603358019e-05, + "loss": 0.4768, + "num_input_tokens_seen": 4785968, + "step": 17890 + }, + { + "epoch": 4.690694626474443, + "grad_norm": 2.230698823928833, + "learning_rate": 3.197115313207252e-05, + "loss": 0.5487, + "num_input_tokens_seen": 4787424, + "step": 17895 + }, + { + "epoch": 4.692005242463958, + "grad_norm": 2.734635353088379, + "learning_rate": 3.1960168771514565e-05, + "loss": 0.462, + "num_input_tokens_seen": 4788512, + "step": 17900 + }, + { + "epoch": 4.693315858453473, + "grad_norm": 4.304066181182861, + "learning_rate": 3.1949182954205295e-05, + "loss": 0.4891, + "num_input_tokens_seen": 4789888, + "step": 17905 + }, + { + "epoch": 4.694626474442988, + "grad_norm": 13.496975898742676, + "learning_rate": 3.193819568244405e-05, + "loss": 0.7618, + "num_input_tokens_seen": 4791344, + "step": 17910 + }, + { + "epoch": 4.695937090432503, + "grad_norm": 2.75004506111145, + "learning_rate": 3.192720695853043e-05, + "loss": 0.3865, + "num_input_tokens_seen": 4792992, + "step": 17915 + }, + { + "epoch": 4.697247706422019, + "grad_norm": 1.9608319997787476, + "learning_rate": 3.1916216784764365e-05, + "loss": 0.4629, + "num_input_tokens_seen": 4794240, + "step": 17920 + }, + { + "epoch": 4.698558322411533, + "grad_norm": 1.8596380949020386, + "learning_rate": 3.1905225163446075e-05, + "loss": 0.4064, + "num_input_tokens_seen": 4795952, + "step": 17925 + }, + { + "epoch": 4.699868938401049, + "grad_norm": 2.1937453746795654, + "learning_rate": 3.189423209687609e-05, + "loss": 0.5159, + "num_input_tokens_seen": 4797504, + "step": 17930 + }, + { + "epoch": 4.7011795543905635, + "grad_norm": 3.2861270904541016, + "learning_rate": 3.188323758735524e-05, + "loss": 0.4758, + "num_input_tokens_seen": 4798928, + "step": 17935 + }, + { + "epoch": 4.702490170380079, + "grad_norm": 1.8191465139389038, + "learning_rate": 3.187224163718466e-05, + "loss": 0.438, + "num_input_tokens_seen": 4800032, + "step": 17940 + }, + { + "epoch": 4.703800786369594, + "grad_norm": 1.8043999671936035, + "learning_rate": 3.186124424866578e-05, + "loss": 0.403, + "num_input_tokens_seen": 4801264, + "step": 17945 + }, + { + "epoch": 4.705111402359108, + "grad_norm": 5.853120803833008, + "learning_rate": 3.185024542410033e-05, + "loss": 0.5945, + "num_input_tokens_seen": 4802544, + "step": 17950 + }, + { + "epoch": 4.706422018348624, + "grad_norm": 2.486701726913452, + "learning_rate": 3.183924516579034e-05, + "loss": 0.5941, + "num_input_tokens_seen": 4803856, + "step": 17955 + }, + { + "epoch": 4.707732634338139, + "grad_norm": 7.802852153778076, + "learning_rate": 3.182824347603818e-05, + "loss": 0.5645, + "num_input_tokens_seen": 4805072, + "step": 17960 + }, + { + "epoch": 4.709043250327654, + "grad_norm": 1.2409121990203857, + "learning_rate": 3.181724035714644e-05, + "loss": 0.5291, + "num_input_tokens_seen": 4806304, + "step": 17965 + }, + { + "epoch": 4.710353866317169, + "grad_norm": 1.554979681968689, + "learning_rate": 3.1806235811418074e-05, + "loss": 0.4762, + "num_input_tokens_seen": 4807920, + "step": 17970 + }, + { + "epoch": 4.711664482306684, + "grad_norm": 4.558489799499512, + "learning_rate": 3.179522984115632e-05, + "loss": 0.4584, + "num_input_tokens_seen": 4809216, + "step": 17975 + }, + { + "epoch": 4.712975098296199, + "grad_norm": 2.335991382598877, + "learning_rate": 3.178422244866469e-05, + "loss": 0.4946, + "num_input_tokens_seen": 4810304, + "step": 17980 + }, + { + "epoch": 4.714285714285714, + "grad_norm": 2.0695362091064453, + "learning_rate": 3.177321363624703e-05, + "loss": 0.4533, + "num_input_tokens_seen": 4811440, + "step": 17985 + }, + { + "epoch": 4.715596330275229, + "grad_norm": 0.7020010948181152, + "learning_rate": 3.176220340620746e-05, + "loss": 0.5122, + "num_input_tokens_seen": 4812928, + "step": 17990 + }, + { + "epoch": 4.716906946264745, + "grad_norm": 1.4462666511535645, + "learning_rate": 3.17511917608504e-05, + "loss": 0.3997, + "num_input_tokens_seen": 4814096, + "step": 17995 + }, + { + "epoch": 4.718217562254259, + "grad_norm": 2.164576530456543, + "learning_rate": 3.174017870248057e-05, + "loss": 0.3521, + "num_input_tokens_seen": 4815264, + "step": 18000 + }, + { + "epoch": 4.719528178243775, + "grad_norm": 1.9689006805419922, + "learning_rate": 3.172916423340299e-05, + "loss": 0.3615, + "num_input_tokens_seen": 4816320, + "step": 18005 + }, + { + "epoch": 4.7208387942332894, + "grad_norm": 2.999441146850586, + "learning_rate": 3.171814835592294e-05, + "loss": 0.5858, + "num_input_tokens_seen": 4817280, + "step": 18010 + }, + { + "epoch": 4.722149410222805, + "grad_norm": 2.438143491744995, + "learning_rate": 3.170713107234607e-05, + "loss": 0.4118, + "num_input_tokens_seen": 4818656, + "step": 18015 + }, + { + "epoch": 4.72346002621232, + "grad_norm": 1.477035641670227, + "learning_rate": 3.169611238497825e-05, + "loss": 0.3127, + "num_input_tokens_seen": 4819888, + "step": 18020 + }, + { + "epoch": 4.724770642201835, + "grad_norm": 1.1376197338104248, + "learning_rate": 3.168509229612568e-05, + "loss": 0.5247, + "num_input_tokens_seen": 4821152, + "step": 18025 + }, + { + "epoch": 4.72608125819135, + "grad_norm": 1.2881567478179932, + "learning_rate": 3.167407080809484e-05, + "loss": 0.3869, + "num_input_tokens_seen": 4822752, + "step": 18030 + }, + { + "epoch": 4.727391874180865, + "grad_norm": 4.837313175201416, + "learning_rate": 3.166304792319251e-05, + "loss": 0.5156, + "num_input_tokens_seen": 4824048, + "step": 18035 + }, + { + "epoch": 4.72870249017038, + "grad_norm": 0.9220665097236633, + "learning_rate": 3.165202364372577e-05, + "loss": 0.463, + "num_input_tokens_seen": 4825616, + "step": 18040 + }, + { + "epoch": 4.730013106159895, + "grad_norm": 7.445042133331299, + "learning_rate": 3.164099797200197e-05, + "loss": 0.4168, + "num_input_tokens_seen": 4826352, + "step": 18045 + }, + { + "epoch": 4.73132372214941, + "grad_norm": 2.2587971687316895, + "learning_rate": 3.162997091032876e-05, + "loss": 0.5282, + "num_input_tokens_seen": 4827552, + "step": 18050 + }, + { + "epoch": 4.732634338138926, + "grad_norm": 3.592818260192871, + "learning_rate": 3.161894246101411e-05, + "loss": 0.4202, + "num_input_tokens_seen": 4828768, + "step": 18055 + }, + { + "epoch": 4.73394495412844, + "grad_norm": 2.394233465194702, + "learning_rate": 3.160791262636624e-05, + "loss": 0.5879, + "num_input_tokens_seen": 4829760, + "step": 18060 + }, + { + "epoch": 4.735255570117955, + "grad_norm": 0.6624240875244141, + "learning_rate": 3.159688140869366e-05, + "loss": 0.4254, + "num_input_tokens_seen": 4830864, + "step": 18065 + }, + { + "epoch": 4.736566186107471, + "grad_norm": 1.993308424949646, + "learning_rate": 3.158584881030521e-05, + "loss": 0.4451, + "num_input_tokens_seen": 4832080, + "step": 18070 + }, + { + "epoch": 4.737876802096985, + "grad_norm": 1.9830083847045898, + "learning_rate": 3.157481483350998e-05, + "loss": 0.5598, + "num_input_tokens_seen": 4833440, + "step": 18075 + }, + { + "epoch": 4.739187418086501, + "grad_norm": 2.400496006011963, + "learning_rate": 3.156377948061737e-05, + "loss": 0.4138, + "num_input_tokens_seen": 4834640, + "step": 18080 + }, + { + "epoch": 4.740498034076015, + "grad_norm": 1.1562221050262451, + "learning_rate": 3.155274275393705e-05, + "loss": 0.479, + "num_input_tokens_seen": 4835968, + "step": 18085 + }, + { + "epoch": 4.741808650065531, + "grad_norm": 2.5397560596466064, + "learning_rate": 3.1541704655779e-05, + "loss": 0.3874, + "num_input_tokens_seen": 4837296, + "step": 18090 + }, + { + "epoch": 4.743119266055046, + "grad_norm": 1.2436325550079346, + "learning_rate": 3.1530665188453464e-05, + "loss": 0.9107, + "num_input_tokens_seen": 4838592, + "step": 18095 + }, + { + "epoch": 4.744429882044561, + "grad_norm": 2.146879196166992, + "learning_rate": 3.151962435427099e-05, + "loss": 0.524, + "num_input_tokens_seen": 4840000, + "step": 18100 + }, + { + "epoch": 4.745740498034076, + "grad_norm": 3.399367094039917, + "learning_rate": 3.15085821555424e-05, + "loss": 0.3418, + "num_input_tokens_seen": 4841504, + "step": 18105 + }, + { + "epoch": 4.747051114023591, + "grad_norm": 6.685293674468994, + "learning_rate": 3.1497538594578814e-05, + "loss": 0.7318, + "num_input_tokens_seen": 4842624, + "step": 18110 + }, + { + "epoch": 4.748361730013106, + "grad_norm": 4.027723789215088, + "learning_rate": 3.1486493673691624e-05, + "loss": 0.3652, + "num_input_tokens_seen": 4843792, + "step": 18115 + }, + { + "epoch": 4.749672346002622, + "grad_norm": 5.113737106323242, + "learning_rate": 3.147544739519253e-05, + "loss": 0.4008, + "num_input_tokens_seen": 4845008, + "step": 18120 + }, + { + "epoch": 4.750982961992136, + "grad_norm": 1.7697408199310303, + "learning_rate": 3.1464399761393465e-05, + "loss": 0.6111, + "num_input_tokens_seen": 4846352, + "step": 18125 + }, + { + "epoch": 4.752293577981652, + "grad_norm": 1.1000897884368896, + "learning_rate": 3.145335077460671e-05, + "loss": 0.6113, + "num_input_tokens_seen": 4847952, + "step": 18130 + }, + { + "epoch": 4.753604193971166, + "grad_norm": 2.1061670780181885, + "learning_rate": 3.144230043714479e-05, + "loss": 0.3775, + "num_input_tokens_seen": 4848848, + "step": 18135 + }, + { + "epoch": 4.754914809960681, + "grad_norm": 1.721928358078003, + "learning_rate": 3.143124875132052e-05, + "loss": 0.4128, + "num_input_tokens_seen": 4850288, + "step": 18140 + }, + { + "epoch": 4.756225425950197, + "grad_norm": 4.867207050323486, + "learning_rate": 3.142019571944699e-05, + "loss": 0.4395, + "num_input_tokens_seen": 4851392, + "step": 18145 + }, + { + "epoch": 4.757536041939712, + "grad_norm": 6.980640411376953, + "learning_rate": 3.140914134383759e-05, + "loss": 0.3694, + "num_input_tokens_seen": 4852496, + "step": 18150 + }, + { + "epoch": 4.758846657929227, + "grad_norm": 1.4591834545135498, + "learning_rate": 3.139808562680599e-05, + "loss": 0.4962, + "num_input_tokens_seen": 4854144, + "step": 18155 + }, + { + "epoch": 4.760157273918741, + "grad_norm": 1.7309808731079102, + "learning_rate": 3.138702857066612e-05, + "loss": 0.6133, + "num_input_tokens_seen": 4855312, + "step": 18160 + }, + { + "epoch": 4.761467889908257, + "grad_norm": 1.1738097667694092, + "learning_rate": 3.137597017773221e-05, + "loss": 0.4283, + "num_input_tokens_seen": 4856448, + "step": 18165 + }, + { + "epoch": 4.762778505897772, + "grad_norm": 2.651726245880127, + "learning_rate": 3.136491045031875e-05, + "loss": 0.457, + "num_input_tokens_seen": 4858032, + "step": 18170 + }, + { + "epoch": 4.764089121887287, + "grad_norm": 3.1344406604766846, + "learning_rate": 3.135384939074054e-05, + "loss": 0.4842, + "num_input_tokens_seen": 4859136, + "step": 18175 + }, + { + "epoch": 4.765399737876802, + "grad_norm": 2.2755699157714844, + "learning_rate": 3.1342787001312615e-05, + "loss": 0.4492, + "num_input_tokens_seen": 4860976, + "step": 18180 + }, + { + "epoch": 4.766710353866317, + "grad_norm": 2.7684381008148193, + "learning_rate": 3.133172328435034e-05, + "loss": 0.4949, + "num_input_tokens_seen": 4862208, + "step": 18185 + }, + { + "epoch": 4.768020969855832, + "grad_norm": 2.5454330444335938, + "learning_rate": 3.132065824216932e-05, + "loss": 0.5243, + "num_input_tokens_seen": 4863408, + "step": 18190 + }, + { + "epoch": 4.7693315858453476, + "grad_norm": 1.2542108297348022, + "learning_rate": 3.1309591877085464e-05, + "loss": 0.382, + "num_input_tokens_seen": 4864896, + "step": 18195 + }, + { + "epoch": 4.770642201834862, + "grad_norm": 5.233691692352295, + "learning_rate": 3.1298524191414915e-05, + "loss": 0.5749, + "num_input_tokens_seen": 4866336, + "step": 18200 + }, + { + "epoch": 4.771952817824378, + "grad_norm": 13.78478717803955, + "learning_rate": 3.128745518747414e-05, + "loss": 0.4016, + "num_input_tokens_seen": 4867472, + "step": 18205 + }, + { + "epoch": 4.773263433813892, + "grad_norm": 3.7025203704833984, + "learning_rate": 3.127638486757985e-05, + "loss": 0.3539, + "num_input_tokens_seen": 4869232, + "step": 18210 + }, + { + "epoch": 4.774574049803408, + "grad_norm": 2.387498140335083, + "learning_rate": 3.126531323404907e-05, + "loss": 0.3708, + "num_input_tokens_seen": 4870672, + "step": 18215 + }, + { + "epoch": 4.775884665792923, + "grad_norm": 1.4205150604248047, + "learning_rate": 3.125424028919904e-05, + "loss": 0.5938, + "num_input_tokens_seen": 4872672, + "step": 18220 + }, + { + "epoch": 4.777195281782438, + "grad_norm": 4.92432165145874, + "learning_rate": 3.1243166035347335e-05, + "loss": 0.547, + "num_input_tokens_seen": 4874064, + "step": 18225 + }, + { + "epoch": 4.778505897771953, + "grad_norm": 1.2624009847640991, + "learning_rate": 3.123209047481177e-05, + "loss": 0.392, + "num_input_tokens_seen": 4875760, + "step": 18230 + }, + { + "epoch": 4.779816513761467, + "grad_norm": 2.366976261138916, + "learning_rate": 3.1221013609910424e-05, + "loss": 0.3653, + "num_input_tokens_seen": 4877104, + "step": 18235 + }, + { + "epoch": 4.781127129750983, + "grad_norm": 3.0500781536102295, + "learning_rate": 3.120993544296168e-05, + "loss": 0.336, + "num_input_tokens_seen": 4878256, + "step": 18240 + }, + { + "epoch": 4.7824377457404985, + "grad_norm": 4.711564540863037, + "learning_rate": 3.119885597628418e-05, + "loss": 0.4932, + "num_input_tokens_seen": 4880016, + "step": 18245 + }, + { + "epoch": 4.783748361730013, + "grad_norm": 2.090930700302124, + "learning_rate": 3.1187775212196835e-05, + "loss": 0.6609, + "num_input_tokens_seen": 4881584, + "step": 18250 + }, + { + "epoch": 4.785058977719528, + "grad_norm": 2.968168258666992, + "learning_rate": 3.117669315301883e-05, + "loss": 0.2971, + "num_input_tokens_seen": 4882960, + "step": 18255 + }, + { + "epoch": 4.786369593709043, + "grad_norm": 2.3732237815856934, + "learning_rate": 3.116560980106962e-05, + "loss": 0.5204, + "num_input_tokens_seen": 4884384, + "step": 18260 + }, + { + "epoch": 4.787680209698558, + "grad_norm": 4.974424362182617, + "learning_rate": 3.115452515866893e-05, + "loss": 0.5578, + "num_input_tokens_seen": 4885408, + "step": 18265 + }, + { + "epoch": 4.7889908256880735, + "grad_norm": 0.9032499194145203, + "learning_rate": 3.114343922813677e-05, + "loss": 0.4135, + "num_input_tokens_seen": 4886656, + "step": 18270 + }, + { + "epoch": 4.790301441677588, + "grad_norm": 2.673173427581787, + "learning_rate": 3.1132352011793384e-05, + "loss": 0.4392, + "num_input_tokens_seen": 4887984, + "step": 18275 + }, + { + "epoch": 4.791612057667104, + "grad_norm": 1.9515858888626099, + "learning_rate": 3.112126351195932e-05, + "loss": 0.4475, + "num_input_tokens_seen": 4889200, + "step": 18280 + }, + { + "epoch": 4.792922673656618, + "grad_norm": 1.1176223754882812, + "learning_rate": 3.1110173730955374e-05, + "loss": 0.3634, + "num_input_tokens_seen": 4890272, + "step": 18285 + }, + { + "epoch": 4.794233289646134, + "grad_norm": 3.6766529083251953, + "learning_rate": 3.109908267110263e-05, + "loss": 0.4903, + "num_input_tokens_seen": 4891504, + "step": 18290 + }, + { + "epoch": 4.795543905635649, + "grad_norm": 1.4469822645187378, + "learning_rate": 3.108799033472242e-05, + "loss": 0.3594, + "num_input_tokens_seen": 4892608, + "step": 18295 + }, + { + "epoch": 4.796854521625164, + "grad_norm": 2.3743324279785156, + "learning_rate": 3.107689672413636e-05, + "loss": 0.4252, + "num_input_tokens_seen": 4894032, + "step": 18300 + }, + { + "epoch": 4.798165137614679, + "grad_norm": 2.026834726333618, + "learning_rate": 3.10658018416663e-05, + "loss": 0.3162, + "num_input_tokens_seen": 4895424, + "step": 18305 + }, + { + "epoch": 4.799475753604194, + "grad_norm": 2.447049140930176, + "learning_rate": 3.1054705689634405e-05, + "loss": 0.3754, + "num_input_tokens_seen": 4896896, + "step": 18310 + }, + { + "epoch": 4.800786369593709, + "grad_norm": 1.3351237773895264, + "learning_rate": 3.1043608270363075e-05, + "loss": 0.3725, + "num_input_tokens_seen": 4898208, + "step": 18315 + }, + { + "epoch": 4.8020969855832245, + "grad_norm": 1.9994040727615356, + "learning_rate": 3.103250958617497e-05, + "loss": 0.3856, + "num_input_tokens_seen": 4899648, + "step": 18320 + }, + { + "epoch": 4.803407601572739, + "grad_norm": 2.8677210807800293, + "learning_rate": 3.102140963939303e-05, + "loss": 0.4948, + "num_input_tokens_seen": 4900720, + "step": 18325 + }, + { + "epoch": 4.804718217562254, + "grad_norm": 2.1619343757629395, + "learning_rate": 3.101030843234046e-05, + "loss": 0.4993, + "num_input_tokens_seen": 4901808, + "step": 18330 + }, + { + "epoch": 4.806028833551769, + "grad_norm": 2.630629062652588, + "learning_rate": 3.099920596734073e-05, + "loss": 0.5405, + "num_input_tokens_seen": 4903184, + "step": 18335 + }, + { + "epoch": 4.807339449541285, + "grad_norm": 2.2677996158599854, + "learning_rate": 3.0988102246717544e-05, + "loss": 0.5527, + "num_input_tokens_seen": 4904640, + "step": 18340 + }, + { + "epoch": 4.8086500655307995, + "grad_norm": 3.963609457015991, + "learning_rate": 3.0976997272794906e-05, + "loss": 0.5067, + "num_input_tokens_seen": 4905920, + "step": 18345 + }, + { + "epoch": 4.809960681520314, + "grad_norm": 2.6287522315979004, + "learning_rate": 3.096589104789707e-05, + "loss": 0.5196, + "num_input_tokens_seen": 4907440, + "step": 18350 + }, + { + "epoch": 4.81127129750983, + "grad_norm": 2.2707996368408203, + "learning_rate": 3.095478357434854e-05, + "loss": 0.5236, + "num_input_tokens_seen": 4908768, + "step": 18355 + }, + { + "epoch": 4.812581913499344, + "grad_norm": 3.9868414402008057, + "learning_rate": 3.09436748544741e-05, + "loss": 0.6278, + "num_input_tokens_seen": 4910208, + "step": 18360 + }, + { + "epoch": 4.81389252948886, + "grad_norm": 5.936359405517578, + "learning_rate": 3.093256489059878e-05, + "loss": 0.428, + "num_input_tokens_seen": 4911392, + "step": 18365 + }, + { + "epoch": 4.815203145478375, + "grad_norm": 2.5122721195220947, + "learning_rate": 3.092145368504789e-05, + "loss": 0.5514, + "num_input_tokens_seen": 4912400, + "step": 18370 + }, + { + "epoch": 4.81651376146789, + "grad_norm": 4.04823112487793, + "learning_rate": 3.0910341240146966e-05, + "loss": 0.8554, + "num_input_tokens_seen": 4913488, + "step": 18375 + }, + { + "epoch": 4.817824377457405, + "grad_norm": 5.226048469543457, + "learning_rate": 3.089922755822183e-05, + "loss": 0.5753, + "num_input_tokens_seen": 4914480, + "step": 18380 + }, + { + "epoch": 4.81913499344692, + "grad_norm": 0.5933981537818909, + "learning_rate": 3.088811264159856e-05, + "loss": 0.4335, + "num_input_tokens_seen": 4916192, + "step": 18385 + }, + { + "epoch": 4.820445609436435, + "grad_norm": 4.65023136138916, + "learning_rate": 3.08769964926035e-05, + "loss": 0.4909, + "num_input_tokens_seen": 4917312, + "step": 18390 + }, + { + "epoch": 4.8217562254259505, + "grad_norm": 2.0258800983428955, + "learning_rate": 3.086587911356322e-05, + "loss": 0.4614, + "num_input_tokens_seen": 4918592, + "step": 18395 + }, + { + "epoch": 4.823066841415465, + "grad_norm": 1.913126826286316, + "learning_rate": 3.085476050680458e-05, + "loss": 0.6989, + "num_input_tokens_seen": 4920096, + "step": 18400 + }, + { + "epoch": 4.824377457404981, + "grad_norm": 1.7195206880569458, + "learning_rate": 3.0843640674654675e-05, + "loss": 0.48, + "num_input_tokens_seen": 4921696, + "step": 18405 + }, + { + "epoch": 4.825688073394495, + "grad_norm": 5.335377216339111, + "learning_rate": 3.083251961944088e-05, + "loss": 0.3473, + "num_input_tokens_seen": 4922640, + "step": 18410 + }, + { + "epoch": 4.826998689384011, + "grad_norm": 9.443229675292969, + "learning_rate": 3.08213973434908e-05, + "loss": 0.5861, + "num_input_tokens_seen": 4923568, + "step": 18415 + }, + { + "epoch": 4.8283093053735255, + "grad_norm": 2.5144150257110596, + "learning_rate": 3.081027384913232e-05, + "loss": 0.2803, + "num_input_tokens_seen": 4925104, + "step": 18420 + }, + { + "epoch": 4.82961992136304, + "grad_norm": 6.857880115509033, + "learning_rate": 3.079914913869356e-05, + "loss": 0.7377, + "num_input_tokens_seen": 4926224, + "step": 18425 + }, + { + "epoch": 4.830930537352556, + "grad_norm": 2.759376049041748, + "learning_rate": 3.0788023214502915e-05, + "loss": 0.5243, + "num_input_tokens_seen": 4928288, + "step": 18430 + }, + { + "epoch": 4.832241153342071, + "grad_norm": 3.1941118240356445, + "learning_rate": 3.0776896078889e-05, + "loss": 0.595, + "num_input_tokens_seen": 4930080, + "step": 18435 + }, + { + "epoch": 4.833551769331586, + "grad_norm": 2.5912983417510986, + "learning_rate": 3.0765767734180714e-05, + "loss": 0.4903, + "num_input_tokens_seen": 4931440, + "step": 18440 + }, + { + "epoch": 4.834862385321101, + "grad_norm": 2.080821990966797, + "learning_rate": 3.075463818270722e-05, + "loss": 0.4136, + "num_input_tokens_seen": 4932752, + "step": 18445 + }, + { + "epoch": 4.836173001310616, + "grad_norm": 7.338587760925293, + "learning_rate": 3.074350742679789e-05, + "loss": 0.5028, + "num_input_tokens_seen": 4934000, + "step": 18450 + }, + { + "epoch": 4.837483617300131, + "grad_norm": 1.6168640851974487, + "learning_rate": 3.0732375468782384e-05, + "loss": 0.2915, + "num_input_tokens_seen": 4935200, + "step": 18455 + }, + { + "epoch": 4.838794233289646, + "grad_norm": 2.887720823287964, + "learning_rate": 3.072124231099058e-05, + "loss": 0.4816, + "num_input_tokens_seen": 4936224, + "step": 18460 + }, + { + "epoch": 4.840104849279161, + "grad_norm": 6.639423370361328, + "learning_rate": 3.0710107955752664e-05, + "loss": 0.4618, + "num_input_tokens_seen": 4937216, + "step": 18465 + }, + { + "epoch": 4.8414154652686765, + "grad_norm": 10.01048755645752, + "learning_rate": 3.069897240539901e-05, + "loss": 0.3754, + "num_input_tokens_seen": 4938752, + "step": 18470 + }, + { + "epoch": 4.842726081258191, + "grad_norm": 2.3196849822998047, + "learning_rate": 3.068783566226028e-05, + "loss": 0.4088, + "num_input_tokens_seen": 4940272, + "step": 18475 + }, + { + "epoch": 4.844036697247707, + "grad_norm": 2.746737241744995, + "learning_rate": 3.067669772866738e-05, + "loss": 0.3632, + "num_input_tokens_seen": 4941472, + "step": 18480 + }, + { + "epoch": 4.845347313237221, + "grad_norm": 0.9500427842140198, + "learning_rate": 3.066555860695145e-05, + "loss": 0.5045, + "num_input_tokens_seen": 4943088, + "step": 18485 + }, + { + "epoch": 4.846657929226737, + "grad_norm": 3.751929998397827, + "learning_rate": 3.065441829944389e-05, + "loss": 0.4029, + "num_input_tokens_seen": 4944224, + "step": 18490 + }, + { + "epoch": 4.8479685452162515, + "grad_norm": 2.851041078567505, + "learning_rate": 3.064327680847635e-05, + "loss": 0.4161, + "num_input_tokens_seen": 4945808, + "step": 18495 + }, + { + "epoch": 4.849279161205767, + "grad_norm": 1.0600801706314087, + "learning_rate": 3.063213413638073e-05, + "loss": 0.5238, + "num_input_tokens_seen": 4947776, + "step": 18500 + }, + { + "epoch": 4.850589777195282, + "grad_norm": 1.7351902723312378, + "learning_rate": 3.0620990285489165e-05, + "loss": 0.6102, + "num_input_tokens_seen": 4948976, + "step": 18505 + }, + { + "epoch": 4.851900393184797, + "grad_norm": 2.657625436782837, + "learning_rate": 3.0609845258134046e-05, + "loss": 0.3083, + "num_input_tokens_seen": 4950368, + "step": 18510 + }, + { + "epoch": 4.853211009174312, + "grad_norm": 2.4788260459899902, + "learning_rate": 3.0598699056648005e-05, + "loss": 0.347, + "num_input_tokens_seen": 4951408, + "step": 18515 + }, + { + "epoch": 4.854521625163827, + "grad_norm": 2.208868980407715, + "learning_rate": 3.0587551683363926e-05, + "loss": 0.4896, + "num_input_tokens_seen": 4952672, + "step": 18520 + }, + { + "epoch": 4.855832241153342, + "grad_norm": 2.5552546977996826, + "learning_rate": 3.057640314061493e-05, + "loss": 0.5447, + "num_input_tokens_seen": 4953920, + "step": 18525 + }, + { + "epoch": 4.857142857142857, + "grad_norm": 1.257943868637085, + "learning_rate": 3.05652534307344e-05, + "loss": 0.4253, + "num_input_tokens_seen": 4955552, + "step": 18530 + }, + { + "epoch": 4.858453473132372, + "grad_norm": 2.1040618419647217, + "learning_rate": 3.055410255605594e-05, + "loss": 0.4216, + "num_input_tokens_seen": 4956688, + "step": 18535 + }, + { + "epoch": 4.859764089121887, + "grad_norm": 3.671144485473633, + "learning_rate": 3.054295051891341e-05, + "loss": 0.5498, + "num_input_tokens_seen": 4957728, + "step": 18540 + }, + { + "epoch": 4.8610747051114025, + "grad_norm": 1.8425853252410889, + "learning_rate": 3.053179732164092e-05, + "loss": 0.3439, + "num_input_tokens_seen": 4959136, + "step": 18545 + }, + { + "epoch": 4.862385321100917, + "grad_norm": 2.0176939964294434, + "learning_rate": 3.0520642966572805e-05, + "loss": 0.7849, + "num_input_tokens_seen": 4960896, + "step": 18550 + }, + { + "epoch": 4.863695937090433, + "grad_norm": 61.54301834106445, + "learning_rate": 3.050948745604365e-05, + "loss": 0.4619, + "num_input_tokens_seen": 4962224, + "step": 18555 + }, + { + "epoch": 4.865006553079947, + "grad_norm": 6.943523406982422, + "learning_rate": 3.049833079238829e-05, + "loss": 0.4874, + "num_input_tokens_seen": 4963440, + "step": 18560 + }, + { + "epoch": 4.866317169069463, + "grad_norm": 1.766796588897705, + "learning_rate": 3.0487172977941806e-05, + "loss": 0.3803, + "num_input_tokens_seen": 4964768, + "step": 18565 + }, + { + "epoch": 4.8676277850589775, + "grad_norm": 2.3223366737365723, + "learning_rate": 3.0476014015039496e-05, + "loss": 0.4418, + "num_input_tokens_seen": 4966464, + "step": 18570 + }, + { + "epoch": 4.868938401048493, + "grad_norm": 4.317482948303223, + "learning_rate": 3.046485390601691e-05, + "loss": 0.4654, + "num_input_tokens_seen": 4967552, + "step": 18575 + }, + { + "epoch": 4.870249017038008, + "grad_norm": 3.103905200958252, + "learning_rate": 3.045369265320984e-05, + "loss": 0.5239, + "num_input_tokens_seen": 4968528, + "step": 18580 + }, + { + "epoch": 4.871559633027523, + "grad_norm": 1.3325939178466797, + "learning_rate": 3.0442530258954334e-05, + "loss": 0.452, + "num_input_tokens_seen": 4969632, + "step": 18585 + }, + { + "epoch": 4.872870249017038, + "grad_norm": 2.9839699268341064, + "learning_rate": 3.0431366725586642e-05, + "loss": 0.6207, + "num_input_tokens_seen": 4971168, + "step": 18590 + }, + { + "epoch": 4.8741808650065535, + "grad_norm": 1.4727551937103271, + "learning_rate": 3.0420202055443276e-05, + "loss": 0.5325, + "num_input_tokens_seen": 4972592, + "step": 18595 + }, + { + "epoch": 4.875491480996068, + "grad_norm": 4.155409336090088, + "learning_rate": 3.040903625086099e-05, + "loss": 0.4225, + "num_input_tokens_seen": 4973776, + "step": 18600 + }, + { + "epoch": 4.876802096985584, + "grad_norm": 3.9574520587921143, + "learning_rate": 3.0397869314176762e-05, + "loss": 0.3594, + "num_input_tokens_seen": 4974864, + "step": 18605 + }, + { + "epoch": 4.878112712975098, + "grad_norm": 4.6312408447265625, + "learning_rate": 3.0386701247727807e-05, + "loss": 0.3421, + "num_input_tokens_seen": 4976016, + "step": 18610 + }, + { + "epoch": 4.879423328964613, + "grad_norm": 1.9570711851119995, + "learning_rate": 3.0375532053851592e-05, + "loss": 0.3779, + "num_input_tokens_seen": 4977312, + "step": 18615 + }, + { + "epoch": 4.8807339449541285, + "grad_norm": 5.470591068267822, + "learning_rate": 3.036436173488581e-05, + "loss": 0.3796, + "num_input_tokens_seen": 4978544, + "step": 18620 + }, + { + "epoch": 4.882044560943643, + "grad_norm": 2.2933311462402344, + "learning_rate": 3.0353190293168388e-05, + "loss": 0.5141, + "num_input_tokens_seen": 4980080, + "step": 18625 + }, + { + "epoch": 4.883355176933159, + "grad_norm": 1.5288269519805908, + "learning_rate": 3.0342017731037482e-05, + "loss": 0.4923, + "num_input_tokens_seen": 4981360, + "step": 18630 + }, + { + "epoch": 4.884665792922673, + "grad_norm": 2.7610294818878174, + "learning_rate": 3.0330844050831492e-05, + "loss": 0.5564, + "num_input_tokens_seen": 4983120, + "step": 18635 + }, + { + "epoch": 4.885976408912189, + "grad_norm": 1.9887282848358154, + "learning_rate": 3.0319669254889055e-05, + "loss": 0.4307, + "num_input_tokens_seen": 4984784, + "step": 18640 + }, + { + "epoch": 4.8872870249017035, + "grad_norm": 2.0129449367523193, + "learning_rate": 3.030849334554904e-05, + "loss": 0.6138, + "num_input_tokens_seen": 4986032, + "step": 18645 + }, + { + "epoch": 4.888597640891219, + "grad_norm": 4.492800235748291, + "learning_rate": 3.0297316325150537e-05, + "loss": 0.4196, + "num_input_tokens_seen": 4987632, + "step": 18650 + }, + { + "epoch": 4.889908256880734, + "grad_norm": 3.196037769317627, + "learning_rate": 3.0286138196032876e-05, + "loss": 0.2859, + "num_input_tokens_seen": 4989488, + "step": 18655 + }, + { + "epoch": 4.891218872870249, + "grad_norm": 2.076340436935425, + "learning_rate": 3.0274958960535632e-05, + "loss": 0.2938, + "num_input_tokens_seen": 4991008, + "step": 18660 + }, + { + "epoch": 4.892529488859764, + "grad_norm": 1.537253737449646, + "learning_rate": 3.0263778620998583e-05, + "loss": 0.3761, + "num_input_tokens_seen": 4995968, + "step": 18665 + }, + { + "epoch": 4.8938401048492794, + "grad_norm": 3.5074970722198486, + "learning_rate": 3.025259717976177e-05, + "loss": 0.4856, + "num_input_tokens_seen": 4997200, + "step": 18670 + }, + { + "epoch": 4.895150720838794, + "grad_norm": 30.916494369506836, + "learning_rate": 3.0241414639165444e-05, + "loss": 0.6017, + "num_input_tokens_seen": 4998432, + "step": 18675 + }, + { + "epoch": 4.89646133682831, + "grad_norm": 2.3685808181762695, + "learning_rate": 3.0230231001550086e-05, + "loss": 0.5487, + "num_input_tokens_seen": 4999792, + "step": 18680 + }, + { + "epoch": 4.897771952817824, + "grad_norm": 1.9285871982574463, + "learning_rate": 3.0219046269256425e-05, + "loss": 0.6427, + "num_input_tokens_seen": 5000992, + "step": 18685 + }, + { + "epoch": 4.89908256880734, + "grad_norm": 1.3632872104644775, + "learning_rate": 3.020786044462539e-05, + "loss": 0.4437, + "num_input_tokens_seen": 5002560, + "step": 18690 + }, + { + "epoch": 4.9003931847968545, + "grad_norm": 3.008117437362671, + "learning_rate": 3.019667352999816e-05, + "loss": 0.4239, + "num_input_tokens_seen": 5003872, + "step": 18695 + }, + { + "epoch": 4.90170380078637, + "grad_norm": 2.198913812637329, + "learning_rate": 3.018548552771615e-05, + "loss": 0.2913, + "num_input_tokens_seen": 5004992, + "step": 18700 + }, + { + "epoch": 4.903014416775885, + "grad_norm": 1.606119990348816, + "learning_rate": 3.0174296440120968e-05, + "loss": 0.3537, + "num_input_tokens_seen": 5006048, + "step": 18705 + }, + { + "epoch": 4.904325032765399, + "grad_norm": 1.0346734523773193, + "learning_rate": 3.0163106269554485e-05, + "loss": 0.5863, + "num_input_tokens_seen": 5007424, + "step": 18710 + }, + { + "epoch": 4.905635648754915, + "grad_norm": 8.908926010131836, + "learning_rate": 3.0151915018358785e-05, + "loss": 0.5637, + "num_input_tokens_seen": 5008496, + "step": 18715 + }, + { + "epoch": 4.9069462647444295, + "grad_norm": 1.6600687503814697, + "learning_rate": 3.0140722688876166e-05, + "loss": 0.5808, + "num_input_tokens_seen": 5009520, + "step": 18720 + }, + { + "epoch": 4.908256880733945, + "grad_norm": 2.6971771717071533, + "learning_rate": 3.0129529283449177e-05, + "loss": 0.563, + "num_input_tokens_seen": 5010672, + "step": 18725 + }, + { + "epoch": 4.90956749672346, + "grad_norm": 3.056373357772827, + "learning_rate": 3.011833480442057e-05, + "loss": 0.4569, + "num_input_tokens_seen": 5011728, + "step": 18730 + }, + { + "epoch": 4.910878112712975, + "grad_norm": 2.4454472064971924, + "learning_rate": 3.0107139254133332e-05, + "loss": 0.6098, + "num_input_tokens_seen": 5013120, + "step": 18735 + }, + { + "epoch": 4.91218872870249, + "grad_norm": 2.15289044380188, + "learning_rate": 3.009594263493068e-05, + "loss": 0.3783, + "num_input_tokens_seen": 5014384, + "step": 18740 + }, + { + "epoch": 4.913499344692005, + "grad_norm": 2.1788454055786133, + "learning_rate": 3.008474494915603e-05, + "loss": 0.4547, + "num_input_tokens_seen": 5015488, + "step": 18745 + }, + { + "epoch": 4.91480996068152, + "grad_norm": 1.751869797706604, + "learning_rate": 3.0073546199153053e-05, + "loss": 0.8974, + "num_input_tokens_seen": 5016848, + "step": 18750 + }, + { + "epoch": 4.916120576671036, + "grad_norm": 0.9209473729133606, + "learning_rate": 3.006234638726561e-05, + "loss": 0.593, + "num_input_tokens_seen": 5018320, + "step": 18755 + }, + { + "epoch": 4.91743119266055, + "grad_norm": 3.381697416305542, + "learning_rate": 3.005114551583783e-05, + "loss": 0.4092, + "num_input_tokens_seen": 5019760, + "step": 18760 + }, + { + "epoch": 4.918741808650066, + "grad_norm": 1.8087921142578125, + "learning_rate": 3.003994358721402e-05, + "loss": 0.442, + "num_input_tokens_seen": 5021248, + "step": 18765 + }, + { + "epoch": 4.9200524246395805, + "grad_norm": 2.008765697479248, + "learning_rate": 3.002874060373872e-05, + "loss": 0.8121, + "num_input_tokens_seen": 5022528, + "step": 18770 + }, + { + "epoch": 4.921363040629096, + "grad_norm": 1.382112741470337, + "learning_rate": 3.0017536567756705e-05, + "loss": 0.5302, + "num_input_tokens_seen": 5023552, + "step": 18775 + }, + { + "epoch": 4.922673656618611, + "grad_norm": 1.670391321182251, + "learning_rate": 3.0006331481612955e-05, + "loss": 0.3787, + "num_input_tokens_seen": 5024672, + "step": 18780 + }, + { + "epoch": 4.923984272608125, + "grad_norm": 2.0525875091552734, + "learning_rate": 2.999512534765267e-05, + "loss": 0.3707, + "num_input_tokens_seen": 5025936, + "step": 18785 + }, + { + "epoch": 4.925294888597641, + "grad_norm": 2.545175790786743, + "learning_rate": 2.9983918168221287e-05, + "loss": 0.4841, + "num_input_tokens_seen": 5027264, + "step": 18790 + }, + { + "epoch": 4.926605504587156, + "grad_norm": 2.985321044921875, + "learning_rate": 2.997270994566444e-05, + "loss": 0.4104, + "num_input_tokens_seen": 5028240, + "step": 18795 + }, + { + "epoch": 4.927916120576671, + "grad_norm": 1.4874539375305176, + "learning_rate": 2.9961500682328e-05, + "loss": 0.3181, + "num_input_tokens_seen": 5029856, + "step": 18800 + }, + { + "epoch": 4.929226736566186, + "grad_norm": 3.3543014526367188, + "learning_rate": 2.995029038055804e-05, + "loss": 0.3281, + "num_input_tokens_seen": 5030688, + "step": 18805 + }, + { + "epoch": 4.930537352555701, + "grad_norm": 2.734015941619873, + "learning_rate": 2.9939079042700847e-05, + "loss": 0.4582, + "num_input_tokens_seen": 5031952, + "step": 18810 + }, + { + "epoch": 4.931847968545216, + "grad_norm": 1.7224853038787842, + "learning_rate": 2.9927866671102946e-05, + "loss": 0.5013, + "num_input_tokens_seen": 5033424, + "step": 18815 + }, + { + "epoch": 4.933158584534731, + "grad_norm": 2.8737690448760986, + "learning_rate": 2.9916653268111074e-05, + "loss": 0.597, + "num_input_tokens_seen": 5034784, + "step": 18820 + }, + { + "epoch": 4.934469200524246, + "grad_norm": 3.8727080821990967, + "learning_rate": 2.9905438836072165e-05, + "loss": 0.3732, + "num_input_tokens_seen": 5036192, + "step": 18825 + }, + { + "epoch": 4.935779816513762, + "grad_norm": 1.6449865102767944, + "learning_rate": 2.989422337733338e-05, + "loss": 0.4456, + "num_input_tokens_seen": 5037360, + "step": 18830 + }, + { + "epoch": 4.937090432503276, + "grad_norm": 2.887085199356079, + "learning_rate": 2.98830068942421e-05, + "loss": 0.331, + "num_input_tokens_seen": 5038608, + "step": 18835 + }, + { + "epoch": 4.938401048492792, + "grad_norm": 1.9858777523040771, + "learning_rate": 2.987178938914591e-05, + "loss": 0.5556, + "num_input_tokens_seen": 5039888, + "step": 18840 + }, + { + "epoch": 4.9397116644823065, + "grad_norm": 1.812888741493225, + "learning_rate": 2.9860570864392625e-05, + "loss": 0.3578, + "num_input_tokens_seen": 5041200, + "step": 18845 + }, + { + "epoch": 4.941022280471822, + "grad_norm": 1.6431493759155273, + "learning_rate": 2.9849351322330255e-05, + "loss": 0.4552, + "num_input_tokens_seen": 5043200, + "step": 18850 + }, + { + "epoch": 4.942332896461337, + "grad_norm": 1.7750922441482544, + "learning_rate": 2.983813076530703e-05, + "loss": 0.3845, + "num_input_tokens_seen": 5044576, + "step": 18855 + }, + { + "epoch": 4.943643512450852, + "grad_norm": 2.6524572372436523, + "learning_rate": 2.9826909195671403e-05, + "loss": 0.4657, + "num_input_tokens_seen": 5045840, + "step": 18860 + }, + { + "epoch": 4.944954128440367, + "grad_norm": 3.0590202808380127, + "learning_rate": 2.981568661577202e-05, + "loss": 0.4663, + "num_input_tokens_seen": 5046960, + "step": 18865 + }, + { + "epoch": 4.946264744429882, + "grad_norm": 1.8052760362625122, + "learning_rate": 2.9804463027957736e-05, + "loss": 0.5253, + "num_input_tokens_seen": 5047984, + "step": 18870 + }, + { + "epoch": 4.947575360419397, + "grad_norm": 1.1140354871749878, + "learning_rate": 2.9793238434577652e-05, + "loss": 0.4123, + "num_input_tokens_seen": 5049424, + "step": 18875 + }, + { + "epoch": 4.948885976408912, + "grad_norm": 4.244843482971191, + "learning_rate": 2.978201283798105e-05, + "loss": 0.2962, + "num_input_tokens_seen": 5050848, + "step": 18880 + }, + { + "epoch": 4.950196592398427, + "grad_norm": 1.7901426553726196, + "learning_rate": 2.977078624051742e-05, + "loss": 0.3585, + "num_input_tokens_seen": 5052240, + "step": 18885 + }, + { + "epoch": 4.951507208387943, + "grad_norm": 4.367071628570557, + "learning_rate": 2.9759558644536467e-05, + "loss": 0.4905, + "num_input_tokens_seen": 5053488, + "step": 18890 + }, + { + "epoch": 4.952817824377457, + "grad_norm": 4.262113571166992, + "learning_rate": 2.9748330052388117e-05, + "loss": 0.3781, + "num_input_tokens_seen": 5054704, + "step": 18895 + }, + { + "epoch": 4.954128440366972, + "grad_norm": 2.32106351852417, + "learning_rate": 2.9737100466422498e-05, + "loss": 0.5197, + "num_input_tokens_seen": 5056112, + "step": 18900 + }, + { + "epoch": 4.955439056356488, + "grad_norm": 2.1062211990356445, + "learning_rate": 2.9725869888989926e-05, + "loss": 0.4055, + "num_input_tokens_seen": 5057264, + "step": 18905 + }, + { + "epoch": 4.956749672346002, + "grad_norm": 1.6809900999069214, + "learning_rate": 2.9714638322440956e-05, + "loss": 0.285, + "num_input_tokens_seen": 5058816, + "step": 18910 + }, + { + "epoch": 4.958060288335518, + "grad_norm": 2.6406378746032715, + "learning_rate": 2.9703405769126334e-05, + "loss": 0.4036, + "num_input_tokens_seen": 5060304, + "step": 18915 + }, + { + "epoch": 4.9593709043250325, + "grad_norm": 4.840281009674072, + "learning_rate": 2.969217223139702e-05, + "loss": 0.3265, + "num_input_tokens_seen": 5061296, + "step": 18920 + }, + { + "epoch": 4.960681520314548, + "grad_norm": 7.154180526733398, + "learning_rate": 2.9680937711604146e-05, + "loss": 0.45, + "num_input_tokens_seen": 5062912, + "step": 18925 + }, + { + "epoch": 4.961992136304063, + "grad_norm": 2.53654408454895, + "learning_rate": 2.9669702212099104e-05, + "loss": 0.4151, + "num_input_tokens_seen": 5064160, + "step": 18930 + }, + { + "epoch": 4.963302752293578, + "grad_norm": 16.946250915527344, + "learning_rate": 2.9658465735233467e-05, + "loss": 0.5303, + "num_input_tokens_seen": 5065712, + "step": 18935 + }, + { + "epoch": 4.964613368283093, + "grad_norm": 2.4422738552093506, + "learning_rate": 2.9647228283358995e-05, + "loss": 0.5256, + "num_input_tokens_seen": 5067232, + "step": 18940 + }, + { + "epoch": 4.965923984272608, + "grad_norm": 1.4563850164413452, + "learning_rate": 2.9635989858827673e-05, + "loss": 0.5145, + "num_input_tokens_seen": 5068832, + "step": 18945 + }, + { + "epoch": 4.967234600262123, + "grad_norm": 2.478813648223877, + "learning_rate": 2.962475046399168e-05, + "loss": 0.4992, + "num_input_tokens_seen": 5069952, + "step": 18950 + }, + { + "epoch": 4.968545216251639, + "grad_norm": 1.4412962198257446, + "learning_rate": 2.9613510101203406e-05, + "loss": 0.6768, + "num_input_tokens_seen": 5071056, + "step": 18955 + }, + { + "epoch": 4.969855832241153, + "grad_norm": 2.043811082839966, + "learning_rate": 2.9602268772815445e-05, + "loss": 0.4229, + "num_input_tokens_seen": 5072512, + "step": 18960 + }, + { + "epoch": 4.971166448230669, + "grad_norm": 1.8419848680496216, + "learning_rate": 2.959102648118058e-05, + "loss": 0.4373, + "num_input_tokens_seen": 5073584, + "step": 18965 + }, + { + "epoch": 4.972477064220183, + "grad_norm": 2.7383224964141846, + "learning_rate": 2.9579783228651802e-05, + "loss": 0.489, + "num_input_tokens_seen": 5074960, + "step": 18970 + }, + { + "epoch": 4.973787680209698, + "grad_norm": 3.5042288303375244, + "learning_rate": 2.956853901758231e-05, + "loss": 0.5882, + "num_input_tokens_seen": 5076288, + "step": 18975 + }, + { + "epoch": 4.975098296199214, + "grad_norm": 4.036025047302246, + "learning_rate": 2.9557293850325496e-05, + "loss": 0.2061, + "num_input_tokens_seen": 5077536, + "step": 18980 + }, + { + "epoch": 4.976408912188729, + "grad_norm": 2.2453930377960205, + "learning_rate": 2.954604772923495e-05, + "loss": 0.4883, + "num_input_tokens_seen": 5078784, + "step": 18985 + }, + { + "epoch": 4.977719528178244, + "grad_norm": 1.9153876304626465, + "learning_rate": 2.9534800656664468e-05, + "loss": 0.5051, + "num_input_tokens_seen": 5080352, + "step": 18990 + }, + { + "epoch": 4.9790301441677585, + "grad_norm": 2.718548536300659, + "learning_rate": 2.9523552634968056e-05, + "loss": 0.5056, + "num_input_tokens_seen": 5081376, + "step": 18995 + }, + { + "epoch": 4.980340760157274, + "grad_norm": 3.3170969486236572, + "learning_rate": 2.9512303666499886e-05, + "loss": 0.3316, + "num_input_tokens_seen": 5082816, + "step": 19000 + }, + { + "epoch": 4.981651376146789, + "grad_norm": 2.0762360095977783, + "learning_rate": 2.9501053753614354e-05, + "loss": 0.4263, + "num_input_tokens_seen": 5084176, + "step": 19005 + }, + { + "epoch": 4.982961992136304, + "grad_norm": 1.8580797910690308, + "learning_rate": 2.9489802898666042e-05, + "loss": 0.3627, + "num_input_tokens_seen": 5085600, + "step": 19010 + }, + { + "epoch": 4.984272608125819, + "grad_norm": 4.668809413909912, + "learning_rate": 2.9478551104009756e-05, + "loss": 0.3726, + "num_input_tokens_seen": 5086880, + "step": 19015 + }, + { + "epoch": 4.985583224115334, + "grad_norm": 2.5424275398254395, + "learning_rate": 2.946729837200045e-05, + "loss": 0.3707, + "num_input_tokens_seen": 5088384, + "step": 19020 + }, + { + "epoch": 4.986893840104849, + "grad_norm": 17.75820541381836, + "learning_rate": 2.945604470499332e-05, + "loss": 0.4255, + "num_input_tokens_seen": 5090240, + "step": 19025 + }, + { + "epoch": 4.988204456094365, + "grad_norm": 1.7815076112747192, + "learning_rate": 2.9444790105343735e-05, + "loss": 0.4917, + "num_input_tokens_seen": 5091504, + "step": 19030 + }, + { + "epoch": 4.989515072083879, + "grad_norm": 2.0754153728485107, + "learning_rate": 2.9433534575407257e-05, + "loss": 0.6475, + "num_input_tokens_seen": 5092832, + "step": 19035 + }, + { + "epoch": 4.990825688073395, + "grad_norm": 2.787902593612671, + "learning_rate": 2.942227811753965e-05, + "loss": 0.4546, + "num_input_tokens_seen": 5094128, + "step": 19040 + }, + { + "epoch": 4.992136304062909, + "grad_norm": 1.3611977100372314, + "learning_rate": 2.9411020734096874e-05, + "loss": 0.4006, + "num_input_tokens_seen": 5095760, + "step": 19045 + }, + { + "epoch": 4.993446920052425, + "grad_norm": 1.3350262641906738, + "learning_rate": 2.9399762427435084e-05, + "loss": 0.391, + "num_input_tokens_seen": 5097008, + "step": 19050 + }, + { + "epoch": 4.99475753604194, + "grad_norm": 2.151550054550171, + "learning_rate": 2.938850319991062e-05, + "loss": 0.5033, + "num_input_tokens_seen": 5098240, + "step": 19055 + }, + { + "epoch": 4.996068152031455, + "grad_norm": 3.515399217605591, + "learning_rate": 2.937724305388002e-05, + "loss": 0.3661, + "num_input_tokens_seen": 5099520, + "step": 19060 + }, + { + "epoch": 4.99737876802097, + "grad_norm": 2.473829507827759, + "learning_rate": 2.936598199170001e-05, + "loss": 0.5209, + "num_input_tokens_seen": 5101360, + "step": 19065 + }, + { + "epoch": 4.9986893840104845, + "grad_norm": 1.4249680042266846, + "learning_rate": 2.9354720015727515e-05, + "loss": 0.3938, + "num_input_tokens_seen": 5102640, + "step": 19070 + }, + { + "epoch": 5.0, + "grad_norm": 5.515725135803223, + "learning_rate": 2.934345712831964e-05, + "loss": 0.6439, + "num_input_tokens_seen": 5104040, + "step": 19075 + }, + { + "epoch": 5.001310615989515, + "grad_norm": 1.9852383136749268, + "learning_rate": 2.9332193331833696e-05, + "loss": 0.4097, + "num_input_tokens_seen": 5105064, + "step": 19080 + }, + { + "epoch": 5.001310615989515, + "eval_loss": 0.46647951006889343, + "eval_runtime": 18.2122, + "eval_samples_per_second": 46.562, + "eval_steps_per_second": 23.281, + "num_input_tokens_seen": 5105064, + "step": 19080 + }, + { + "epoch": 5.00262123197903, + "grad_norm": 3.6007299423217773, + "learning_rate": 2.9320928628627175e-05, + "loss": 0.5129, + "num_input_tokens_seen": 5106232, + "step": 19085 + }, + { + "epoch": 5.003931847968545, + "grad_norm": 2.224386215209961, + "learning_rate": 2.9309663021057754e-05, + "loss": 0.5685, + "num_input_tokens_seen": 5107448, + "step": 19090 + }, + { + "epoch": 5.00524246395806, + "grad_norm": 1.2648564577102661, + "learning_rate": 2.929839651148331e-05, + "loss": 0.4419, + "num_input_tokens_seen": 5108632, + "step": 19095 + }, + { + "epoch": 5.006553079947575, + "grad_norm": 1.7410507202148438, + "learning_rate": 2.9287129102261907e-05, + "loss": 0.6007, + "num_input_tokens_seen": 5109656, + "step": 19100 + }, + { + "epoch": 5.007863695937091, + "grad_norm": 2.440098285675049, + "learning_rate": 2.9275860795751786e-05, + "loss": 0.3623, + "num_input_tokens_seen": 5111240, + "step": 19105 + }, + { + "epoch": 5.009174311926605, + "grad_norm": 2.3105268478393555, + "learning_rate": 2.926459159431139e-05, + "loss": 0.7059, + "num_input_tokens_seen": 5112248, + "step": 19110 + }, + { + "epoch": 5.010484927916121, + "grad_norm": 2.7875938415527344, + "learning_rate": 2.9253321500299354e-05, + "loss": 0.4026, + "num_input_tokens_seen": 5113672, + "step": 19115 + }, + { + "epoch": 5.011795543905635, + "grad_norm": 4.063239574432373, + "learning_rate": 2.9242050516074466e-05, + "loss": 0.4381, + "num_input_tokens_seen": 5114904, + "step": 19120 + }, + { + "epoch": 5.013106159895151, + "grad_norm": 2.6400461196899414, + "learning_rate": 2.9230778643995726e-05, + "loss": 0.5699, + "num_input_tokens_seen": 5116200, + "step": 19125 + }, + { + "epoch": 5.014416775884666, + "grad_norm": 1.3220754861831665, + "learning_rate": 2.9219505886422338e-05, + "loss": 0.2772, + "num_input_tokens_seen": 5117432, + "step": 19130 + }, + { + "epoch": 5.015727391874181, + "grad_norm": 2.411905288696289, + "learning_rate": 2.9208232245713657e-05, + "loss": 0.4745, + "num_input_tokens_seen": 5119192, + "step": 19135 + }, + { + "epoch": 5.017038007863696, + "grad_norm": 5.915231704711914, + "learning_rate": 2.9196957724229234e-05, + "loss": 0.6727, + "num_input_tokens_seen": 5120232, + "step": 19140 + }, + { + "epoch": 5.018348623853211, + "grad_norm": 2.1763219833374023, + "learning_rate": 2.9185682324328815e-05, + "loss": 0.5324, + "num_input_tokens_seen": 5121880, + "step": 19145 + }, + { + "epoch": 5.019659239842726, + "grad_norm": 1.701239824295044, + "learning_rate": 2.9174406048372316e-05, + "loss": 0.4641, + "num_input_tokens_seen": 5123176, + "step": 19150 + }, + { + "epoch": 5.0209698558322415, + "grad_norm": 3.9974772930145264, + "learning_rate": 2.9163128898719845e-05, + "loss": 0.5158, + "num_input_tokens_seen": 5124408, + "step": 19155 + }, + { + "epoch": 5.022280471821756, + "grad_norm": 3.6044297218322754, + "learning_rate": 2.915185087773169e-05, + "loss": 0.6168, + "num_input_tokens_seen": 5125544, + "step": 19160 + }, + { + "epoch": 5.023591087811272, + "grad_norm": 2.121950626373291, + "learning_rate": 2.9140571987768316e-05, + "loss": 0.4914, + "num_input_tokens_seen": 5126968, + "step": 19165 + }, + { + "epoch": 5.024901703800786, + "grad_norm": 1.3781023025512695, + "learning_rate": 2.9129292231190386e-05, + "loss": 0.4677, + "num_input_tokens_seen": 5128200, + "step": 19170 + }, + { + "epoch": 5.026212319790301, + "grad_norm": 1.4310328960418701, + "learning_rate": 2.9118011610358726e-05, + "loss": 0.4152, + "num_input_tokens_seen": 5129560, + "step": 19175 + }, + { + "epoch": 5.027522935779817, + "grad_norm": 4.322203159332275, + "learning_rate": 2.9106730127634346e-05, + "loss": 0.5666, + "num_input_tokens_seen": 5130744, + "step": 19180 + }, + { + "epoch": 5.028833551769331, + "grad_norm": 1.8754621744155884, + "learning_rate": 2.9095447785378443e-05, + "loss": 0.4224, + "num_input_tokens_seen": 5132264, + "step": 19185 + }, + { + "epoch": 5.030144167758847, + "grad_norm": 1.8207802772521973, + "learning_rate": 2.9084164585952405e-05, + "loss": 0.4276, + "num_input_tokens_seen": 5134040, + "step": 19190 + }, + { + "epoch": 5.031454783748361, + "grad_norm": 2.0588266849517822, + "learning_rate": 2.9072880531717773e-05, + "loss": 0.485, + "num_input_tokens_seen": 5135096, + "step": 19195 + }, + { + "epoch": 5.032765399737877, + "grad_norm": 2.0504963397979736, + "learning_rate": 2.9061595625036277e-05, + "loss": 0.5169, + "num_input_tokens_seen": 5136296, + "step": 19200 + }, + { + "epoch": 5.034076015727392, + "grad_norm": 4.919119358062744, + "learning_rate": 2.905030986826984e-05, + "loss": 0.3634, + "num_input_tokens_seen": 5137272, + "step": 19205 + }, + { + "epoch": 5.035386631716907, + "grad_norm": 2.1194071769714355, + "learning_rate": 2.903902326378054e-05, + "loss": 0.5852, + "num_input_tokens_seen": 5138824, + "step": 19210 + }, + { + "epoch": 5.036697247706422, + "grad_norm": 2.1254186630249023, + "learning_rate": 2.9027735813930658e-05, + "loss": 0.3077, + "num_input_tokens_seen": 5139896, + "step": 19215 + }, + { + "epoch": 5.038007863695937, + "grad_norm": 2.2118659019470215, + "learning_rate": 2.9016447521082625e-05, + "loss": 0.3007, + "num_input_tokens_seen": 5142744, + "step": 19220 + }, + { + "epoch": 5.039318479685452, + "grad_norm": 1.9199473857879639, + "learning_rate": 2.9005158387599068e-05, + "loss": 0.298, + "num_input_tokens_seen": 5144152, + "step": 19225 + }, + { + "epoch": 5.0406290956749675, + "grad_norm": 4.106074333190918, + "learning_rate": 2.8993868415842785e-05, + "loss": 0.4792, + "num_input_tokens_seen": 5145528, + "step": 19230 + }, + { + "epoch": 5.041939711664482, + "grad_norm": 2.4114365577697754, + "learning_rate": 2.8982577608176737e-05, + "loss": 0.4134, + "num_input_tokens_seen": 5146872, + "step": 19235 + }, + { + "epoch": 5.043250327653998, + "grad_norm": 2.5306992530822754, + "learning_rate": 2.8971285966964075e-05, + "loss": 0.583, + "num_input_tokens_seen": 5148168, + "step": 19240 + }, + { + "epoch": 5.044560943643512, + "grad_norm": 2.3425004482269287, + "learning_rate": 2.8959993494568123e-05, + "loss": 0.4699, + "num_input_tokens_seen": 5149432, + "step": 19245 + }, + { + "epoch": 5.045871559633028, + "grad_norm": 3.187239170074463, + "learning_rate": 2.894870019335238e-05, + "loss": 0.5717, + "num_input_tokens_seen": 5150568, + "step": 19250 + }, + { + "epoch": 5.047182175622543, + "grad_norm": 3.6869242191314697, + "learning_rate": 2.8937406065680505e-05, + "loss": 0.3812, + "num_input_tokens_seen": 5151512, + "step": 19255 + }, + { + "epoch": 5.048492791612058, + "grad_norm": 1.5287961959838867, + "learning_rate": 2.8926111113916344e-05, + "loss": 0.8756, + "num_input_tokens_seen": 5152472, + "step": 19260 + }, + { + "epoch": 5.049803407601573, + "grad_norm": 1.6605561971664429, + "learning_rate": 2.891481534042391e-05, + "loss": 1.1274, + "num_input_tokens_seen": 5153688, + "step": 19265 + }, + { + "epoch": 5.051114023591087, + "grad_norm": 3.0079691410064697, + "learning_rate": 2.8903518747567394e-05, + "loss": 0.5274, + "num_input_tokens_seen": 5155128, + "step": 19270 + }, + { + "epoch": 5.052424639580603, + "grad_norm": 2.707760810852051, + "learning_rate": 2.8892221337711143e-05, + "loss": 0.8701, + "num_input_tokens_seen": 5156056, + "step": 19275 + }, + { + "epoch": 5.053735255570118, + "grad_norm": 1.1302632093429565, + "learning_rate": 2.88809231132197e-05, + "loss": 0.5409, + "num_input_tokens_seen": 5157320, + "step": 19280 + }, + { + "epoch": 5.055045871559633, + "grad_norm": 3.8365538120269775, + "learning_rate": 2.8869624076457745e-05, + "loss": 0.617, + "num_input_tokens_seen": 5158664, + "step": 19285 + }, + { + "epoch": 5.056356487549148, + "grad_norm": 2.801910638809204, + "learning_rate": 2.885832422979017e-05, + "loss": 0.4424, + "num_input_tokens_seen": 5159768, + "step": 19290 + }, + { + "epoch": 5.057667103538663, + "grad_norm": 2.0091283321380615, + "learning_rate": 2.8847023575581988e-05, + "loss": 0.4472, + "num_input_tokens_seen": 5161320, + "step": 19295 + }, + { + "epoch": 5.058977719528178, + "grad_norm": 1.8762141466140747, + "learning_rate": 2.8835722116198415e-05, + "loss": 0.6921, + "num_input_tokens_seen": 5162776, + "step": 19300 + }, + { + "epoch": 5.0602883355176935, + "grad_norm": 1.7761049270629883, + "learning_rate": 2.882441985400484e-05, + "loss": 0.3121, + "num_input_tokens_seen": 5164200, + "step": 19305 + }, + { + "epoch": 5.061598951507208, + "grad_norm": 3.869995594024658, + "learning_rate": 2.88131167913668e-05, + "loss": 0.6208, + "num_input_tokens_seen": 5166136, + "step": 19310 + }, + { + "epoch": 5.062909567496724, + "grad_norm": 4.908606052398682, + "learning_rate": 2.8801812930650002e-05, + "loss": 0.5182, + "num_input_tokens_seen": 5167560, + "step": 19315 + }, + { + "epoch": 5.064220183486238, + "grad_norm": 3.5312702655792236, + "learning_rate": 2.8790508274220323e-05, + "loss": 0.3826, + "num_input_tokens_seen": 5168536, + "step": 19320 + }, + { + "epoch": 5.065530799475754, + "grad_norm": 6.177748680114746, + "learning_rate": 2.877920282444381e-05, + "loss": 0.4605, + "num_input_tokens_seen": 5169720, + "step": 19325 + }, + { + "epoch": 5.066841415465269, + "grad_norm": 3.0100457668304443, + "learning_rate": 2.8767896583686682e-05, + "loss": 0.4279, + "num_input_tokens_seen": 5171496, + "step": 19330 + }, + { + "epoch": 5.068152031454784, + "grad_norm": 1.0781532526016235, + "learning_rate": 2.875658955431531e-05, + "loss": 0.378, + "num_input_tokens_seen": 5173048, + "step": 19335 + }, + { + "epoch": 5.069462647444299, + "grad_norm": 2.3744702339172363, + "learning_rate": 2.874528173869623e-05, + "loss": 0.7436, + "num_input_tokens_seen": 5174408, + "step": 19340 + }, + { + "epoch": 5.070773263433814, + "grad_norm": 1.472615122795105, + "learning_rate": 2.873397313919616e-05, + "loss": 0.3429, + "num_input_tokens_seen": 5175480, + "step": 19345 + }, + { + "epoch": 5.072083879423329, + "grad_norm": 3.625675678253174, + "learning_rate": 2.872266375818196e-05, + "loss": 0.5733, + "num_input_tokens_seen": 5176792, + "step": 19350 + }, + { + "epoch": 5.073394495412844, + "grad_norm": 4.988677978515625, + "learning_rate": 2.8711353598020662e-05, + "loss": 0.5766, + "num_input_tokens_seen": 5178056, + "step": 19355 + }, + { + "epoch": 5.074705111402359, + "grad_norm": 2.863325834274292, + "learning_rate": 2.8700042661079472e-05, + "loss": 0.441, + "num_input_tokens_seen": 5179000, + "step": 19360 + }, + { + "epoch": 5.076015727391874, + "grad_norm": 2.061185836791992, + "learning_rate": 2.8688730949725752e-05, + "loss": 0.5646, + "num_input_tokens_seen": 5180264, + "step": 19365 + }, + { + "epoch": 5.077326343381389, + "grad_norm": 5.491246700286865, + "learning_rate": 2.867741846632701e-05, + "loss": 0.5857, + "num_input_tokens_seen": 5181704, + "step": 19370 + }, + { + "epoch": 5.078636959370904, + "grad_norm": 1.6012024879455566, + "learning_rate": 2.8666105213250943e-05, + "loss": 0.4583, + "num_input_tokens_seen": 5183320, + "step": 19375 + }, + { + "epoch": 5.0799475753604195, + "grad_norm": 1.6951913833618164, + "learning_rate": 2.8654791192865388e-05, + "loss": 0.3436, + "num_input_tokens_seen": 5184568, + "step": 19380 + }, + { + "epoch": 5.081258191349934, + "grad_norm": 1.767833948135376, + "learning_rate": 2.864347640753835e-05, + "loss": 0.4875, + "num_input_tokens_seen": 5185960, + "step": 19385 + }, + { + "epoch": 5.08256880733945, + "grad_norm": 2.521277666091919, + "learning_rate": 2.8632160859637998e-05, + "loss": 0.4898, + "num_input_tokens_seen": 5187240, + "step": 19390 + }, + { + "epoch": 5.083879423328964, + "grad_norm": 5.2876386642456055, + "learning_rate": 2.8620844551532648e-05, + "loss": 0.5545, + "num_input_tokens_seen": 5188616, + "step": 19395 + }, + { + "epoch": 5.08519003931848, + "grad_norm": 4.711249351501465, + "learning_rate": 2.8609527485590794e-05, + "loss": 0.3465, + "num_input_tokens_seen": 5189640, + "step": 19400 + }, + { + "epoch": 5.086500655307995, + "grad_norm": 1.7217260599136353, + "learning_rate": 2.8598209664181073e-05, + "loss": 0.5604, + "num_input_tokens_seen": 5190968, + "step": 19405 + }, + { + "epoch": 5.08781127129751, + "grad_norm": 2.879106044769287, + "learning_rate": 2.858689108967229e-05, + "loss": 0.3669, + "num_input_tokens_seen": 5192856, + "step": 19410 + }, + { + "epoch": 5.089121887287025, + "grad_norm": 2.5376503467559814, + "learning_rate": 2.8575571764433396e-05, + "loss": 0.448, + "num_input_tokens_seen": 5194488, + "step": 19415 + }, + { + "epoch": 5.09043250327654, + "grad_norm": 3.251253366470337, + "learning_rate": 2.8564251690833505e-05, + "loss": 0.3194, + "num_input_tokens_seen": 5196456, + "step": 19420 + }, + { + "epoch": 5.091743119266055, + "grad_norm": 2.9524083137512207, + "learning_rate": 2.8552930871241906e-05, + "loss": 0.4216, + "num_input_tokens_seen": 5197608, + "step": 19425 + }, + { + "epoch": 5.0930537352555705, + "grad_norm": 6.009607791900635, + "learning_rate": 2.8541609308028006e-05, + "loss": 0.4133, + "num_input_tokens_seen": 5198760, + "step": 19430 + }, + { + "epoch": 5.094364351245085, + "grad_norm": 15.411477088928223, + "learning_rate": 2.8530287003561402e-05, + "loss": 0.3833, + "num_input_tokens_seen": 5199976, + "step": 19435 + }, + { + "epoch": 5.095674967234601, + "grad_norm": 1.1571029424667358, + "learning_rate": 2.8518963960211814e-05, + "loss": 0.4629, + "num_input_tokens_seen": 5201720, + "step": 19440 + }, + { + "epoch": 5.096985583224115, + "grad_norm": 7.154335975646973, + "learning_rate": 2.850764018034917e-05, + "loss": 0.5047, + "num_input_tokens_seen": 5202872, + "step": 19445 + }, + { + "epoch": 5.09829619921363, + "grad_norm": 7.906871795654297, + "learning_rate": 2.8496315666343487e-05, + "loss": 0.4168, + "num_input_tokens_seen": 5203848, + "step": 19450 + }, + { + "epoch": 5.0996068152031455, + "grad_norm": 1.2264738082885742, + "learning_rate": 2.8484990420564972e-05, + "loss": 0.5308, + "num_input_tokens_seen": 5205448, + "step": 19455 + }, + { + "epoch": 5.10091743119266, + "grad_norm": 1.7019833326339722, + "learning_rate": 2.847366444538399e-05, + "loss": 0.5919, + "num_input_tokens_seen": 5207992, + "step": 19460 + }, + { + "epoch": 5.102228047182176, + "grad_norm": 2.271395206451416, + "learning_rate": 2.8462337743171046e-05, + "loss": 0.5472, + "num_input_tokens_seen": 5209976, + "step": 19465 + }, + { + "epoch": 5.10353866317169, + "grad_norm": 6.759373664855957, + "learning_rate": 2.8451010316296783e-05, + "loss": 0.4342, + "num_input_tokens_seen": 5210984, + "step": 19470 + }, + { + "epoch": 5.104849279161206, + "grad_norm": 1.8478056192398071, + "learning_rate": 2.8439682167132027e-05, + "loss": 0.3613, + "num_input_tokens_seen": 5212136, + "step": 19475 + }, + { + "epoch": 5.1061598951507206, + "grad_norm": 2.2838943004608154, + "learning_rate": 2.8428353298047738e-05, + "loss": 0.5633, + "num_input_tokens_seen": 5213704, + "step": 19480 + }, + { + "epoch": 5.107470511140236, + "grad_norm": 2.544140338897705, + "learning_rate": 2.8417023711415036e-05, + "loss": 0.5797, + "num_input_tokens_seen": 5215112, + "step": 19485 + }, + { + "epoch": 5.108781127129751, + "grad_norm": 2.867446184158325, + "learning_rate": 2.8405693409605166e-05, + "loss": 0.5042, + "num_input_tokens_seen": 5216264, + "step": 19490 + }, + { + "epoch": 5.110091743119266, + "grad_norm": 1.9133296012878418, + "learning_rate": 2.839436239498955e-05, + "loss": 0.506, + "num_input_tokens_seen": 5217688, + "step": 19495 + }, + { + "epoch": 5.111402359108781, + "grad_norm": 3.396519184112549, + "learning_rate": 2.8383030669939752e-05, + "loss": 0.5245, + "num_input_tokens_seen": 5218776, + "step": 19500 + }, + { + "epoch": 5.1127129750982965, + "grad_norm": 14.286208152770996, + "learning_rate": 2.837169823682748e-05, + "loss": 0.4698, + "num_input_tokens_seen": 5219736, + "step": 19505 + }, + { + "epoch": 5.114023591087811, + "grad_norm": 1.280198097229004, + "learning_rate": 2.8360365098024593e-05, + "loss": 0.5214, + "num_input_tokens_seen": 5221160, + "step": 19510 + }, + { + "epoch": 5.115334207077327, + "grad_norm": 1.9190493822097778, + "learning_rate": 2.8349031255903098e-05, + "loss": 0.6636, + "num_input_tokens_seen": 5222760, + "step": 19515 + }, + { + "epoch": 5.116644823066841, + "grad_norm": 3.3021626472473145, + "learning_rate": 2.8337696712835143e-05, + "loss": 0.4836, + "num_input_tokens_seen": 5224040, + "step": 19520 + }, + { + "epoch": 5.117955439056357, + "grad_norm": 1.691683053970337, + "learning_rate": 2.8326361471193035e-05, + "loss": 0.3723, + "num_input_tokens_seen": 5225336, + "step": 19525 + }, + { + "epoch": 5.1192660550458715, + "grad_norm": 2.0493807792663574, + "learning_rate": 2.831502553334922e-05, + "loss": 0.3629, + "num_input_tokens_seen": 5226744, + "step": 19530 + }, + { + "epoch": 5.120576671035387, + "grad_norm": 3.9267871379852295, + "learning_rate": 2.8303688901676288e-05, + "loss": 0.4867, + "num_input_tokens_seen": 5227800, + "step": 19535 + }, + { + "epoch": 5.121887287024902, + "grad_norm": 2.895934581756592, + "learning_rate": 2.8292351578546982e-05, + "loss": 0.4024, + "num_input_tokens_seen": 5228984, + "step": 19540 + }, + { + "epoch": 5.123197903014416, + "grad_norm": 1.7605431079864502, + "learning_rate": 2.8281013566334182e-05, + "loss": 0.3681, + "num_input_tokens_seen": 5230456, + "step": 19545 + }, + { + "epoch": 5.124508519003932, + "grad_norm": 2.5793263912200928, + "learning_rate": 2.8269674867410907e-05, + "loss": 0.5256, + "num_input_tokens_seen": 5232040, + "step": 19550 + }, + { + "epoch": 5.1258191349934465, + "grad_norm": 2.5026464462280273, + "learning_rate": 2.8258335484150323e-05, + "loss": 0.3891, + "num_input_tokens_seen": 5233384, + "step": 19555 + }, + { + "epoch": 5.127129750982962, + "grad_norm": 1.7879647016525269, + "learning_rate": 2.8246995418925764e-05, + "loss": 0.5704, + "num_input_tokens_seen": 5234776, + "step": 19560 + }, + { + "epoch": 5.128440366972477, + "grad_norm": 2.1462457180023193, + "learning_rate": 2.823565467411067e-05, + "loss": 0.432, + "num_input_tokens_seen": 5235976, + "step": 19565 + }, + { + "epoch": 5.129750982961992, + "grad_norm": 5.292938709259033, + "learning_rate": 2.8224313252078637e-05, + "loss": 0.4706, + "num_input_tokens_seen": 5237496, + "step": 19570 + }, + { + "epoch": 5.131061598951507, + "grad_norm": 3.801158905029297, + "learning_rate": 2.8212971155203422e-05, + "loss": 0.5493, + "num_input_tokens_seen": 5239176, + "step": 19575 + }, + { + "epoch": 5.1323722149410225, + "grad_norm": 1.6130075454711914, + "learning_rate": 2.820162838585889e-05, + "loss": 0.4054, + "num_input_tokens_seen": 5240504, + "step": 19580 + }, + { + "epoch": 5.133682830930537, + "grad_norm": 2.1338889598846436, + "learning_rate": 2.819028494641907e-05, + "loss": 0.5012, + "num_input_tokens_seen": 5241496, + "step": 19585 + }, + { + "epoch": 5.134993446920053, + "grad_norm": 2.5872161388397217, + "learning_rate": 2.817894083925812e-05, + "loss": 0.6456, + "num_input_tokens_seen": 5242696, + "step": 19590 + }, + { + "epoch": 5.136304062909567, + "grad_norm": 1.8010903596878052, + "learning_rate": 2.816759606675035e-05, + "loss": 0.564, + "num_input_tokens_seen": 5244200, + "step": 19595 + }, + { + "epoch": 5.137614678899083, + "grad_norm": 4.578303337097168, + "learning_rate": 2.81562506312702e-05, + "loss": 0.4799, + "num_input_tokens_seen": 5245288, + "step": 19600 + }, + { + "epoch": 5.1389252948885975, + "grad_norm": 1.4108350276947021, + "learning_rate": 2.814490453519224e-05, + "loss": 0.4861, + "num_input_tokens_seen": 5246808, + "step": 19605 + }, + { + "epoch": 5.140235910878113, + "grad_norm": 2.4420878887176514, + "learning_rate": 2.8133557780891194e-05, + "loss": 0.3796, + "num_input_tokens_seen": 5248072, + "step": 19610 + }, + { + "epoch": 5.141546526867628, + "grad_norm": 1.289031982421875, + "learning_rate": 2.812221037074192e-05, + "loss": 0.3776, + "num_input_tokens_seen": 5249464, + "step": 19615 + }, + { + "epoch": 5.142857142857143, + "grad_norm": 16.051057815551758, + "learning_rate": 2.8110862307119413e-05, + "loss": 0.6924, + "num_input_tokens_seen": 5250808, + "step": 19620 + }, + { + "epoch": 5.144167758846658, + "grad_norm": 2.9325921535491943, + "learning_rate": 2.80995135923988e-05, + "loss": 0.6149, + "num_input_tokens_seen": 5252120, + "step": 19625 + }, + { + "epoch": 5.145478374836173, + "grad_norm": 1.2896299362182617, + "learning_rate": 2.8088164228955345e-05, + "loss": 0.3841, + "num_input_tokens_seen": 5253832, + "step": 19630 + }, + { + "epoch": 5.146788990825688, + "grad_norm": 1.7314554452896118, + "learning_rate": 2.8076814219164456e-05, + "loss": 0.3639, + "num_input_tokens_seen": 5255128, + "step": 19635 + }, + { + "epoch": 5.148099606815203, + "grad_norm": 3.356966018676758, + "learning_rate": 2.806546356540166e-05, + "loss": 0.5508, + "num_input_tokens_seen": 5256248, + "step": 19640 + }, + { + "epoch": 5.149410222804718, + "grad_norm": 7.965069770812988, + "learning_rate": 2.8054112270042644e-05, + "loss": 0.4024, + "num_input_tokens_seen": 5257288, + "step": 19645 + }, + { + "epoch": 5.150720838794233, + "grad_norm": 2.4845330715179443, + "learning_rate": 2.8042760335463203e-05, + "loss": 0.5113, + "num_input_tokens_seen": 5258984, + "step": 19650 + }, + { + "epoch": 5.1520314547837485, + "grad_norm": 1.3158965110778809, + "learning_rate": 2.8031407764039287e-05, + "loss": 0.4172, + "num_input_tokens_seen": 5260360, + "step": 19655 + }, + { + "epoch": 5.153342070773263, + "grad_norm": 1.7388975620269775, + "learning_rate": 2.8020054558146964e-05, + "loss": 0.4045, + "num_input_tokens_seen": 5262056, + "step": 19660 + }, + { + "epoch": 5.154652686762779, + "grad_norm": 5.9295759201049805, + "learning_rate": 2.8008700720162433e-05, + "loss": 0.4943, + "num_input_tokens_seen": 5263432, + "step": 19665 + }, + { + "epoch": 5.155963302752293, + "grad_norm": 2.4096739292144775, + "learning_rate": 2.7997346252462035e-05, + "loss": 0.4704, + "num_input_tokens_seen": 5264760, + "step": 19670 + }, + { + "epoch": 5.157273918741809, + "grad_norm": 2.490806818008423, + "learning_rate": 2.7985991157422254e-05, + "loss": 0.7916, + "num_input_tokens_seen": 5266104, + "step": 19675 + }, + { + "epoch": 5.1585845347313235, + "grad_norm": 1.7353445291519165, + "learning_rate": 2.797463543741969e-05, + "loss": 0.5961, + "num_input_tokens_seen": 5267144, + "step": 19680 + }, + { + "epoch": 5.159895150720839, + "grad_norm": 1.0430394411087036, + "learning_rate": 2.7963279094831057e-05, + "loss": 0.5028, + "num_input_tokens_seen": 5268984, + "step": 19685 + }, + { + "epoch": 5.161205766710354, + "grad_norm": 1.7457596063613892, + "learning_rate": 2.7951922132033235e-05, + "loss": 0.413, + "num_input_tokens_seen": 5270456, + "step": 19690 + }, + { + "epoch": 5.162516382699869, + "grad_norm": 1.4718965291976929, + "learning_rate": 2.794056455140321e-05, + "loss": 0.3024, + "num_input_tokens_seen": 5271480, + "step": 19695 + }, + { + "epoch": 5.163826998689384, + "grad_norm": 7.08998441696167, + "learning_rate": 2.7929206355318104e-05, + "loss": 0.4153, + "num_input_tokens_seen": 5272792, + "step": 19700 + }, + { + "epoch": 5.165137614678899, + "grad_norm": 4.009179592132568, + "learning_rate": 2.791784754615517e-05, + "loss": 0.5242, + "num_input_tokens_seen": 5274104, + "step": 19705 + }, + { + "epoch": 5.166448230668414, + "grad_norm": 3.288926124572754, + "learning_rate": 2.7906488126291784e-05, + "loss": 0.4353, + "num_input_tokens_seen": 5275384, + "step": 19710 + }, + { + "epoch": 5.16775884665793, + "grad_norm": 4.517029762268066, + "learning_rate": 2.789512809810546e-05, + "loss": 0.3692, + "num_input_tokens_seen": 5276664, + "step": 19715 + }, + { + "epoch": 5.169069462647444, + "grad_norm": 1.1343053579330444, + "learning_rate": 2.7883767463973832e-05, + "loss": 0.4982, + "num_input_tokens_seen": 5278136, + "step": 19720 + }, + { + "epoch": 5.17038007863696, + "grad_norm": 2.5580997467041016, + "learning_rate": 2.787240622627465e-05, + "loss": 0.501, + "num_input_tokens_seen": 5279736, + "step": 19725 + }, + { + "epoch": 5.1716906946264745, + "grad_norm": 37.70621109008789, + "learning_rate": 2.78610443873858e-05, + "loss": 0.6304, + "num_input_tokens_seen": 5280952, + "step": 19730 + }, + { + "epoch": 5.173001310615989, + "grad_norm": 3.357898473739624, + "learning_rate": 2.7849681949685314e-05, + "loss": 0.6035, + "num_input_tokens_seen": 5282392, + "step": 19735 + }, + { + "epoch": 5.174311926605505, + "grad_norm": 14.822545051574707, + "learning_rate": 2.783831891555132e-05, + "loss": 0.6463, + "num_input_tokens_seen": 5284136, + "step": 19740 + }, + { + "epoch": 5.175622542595019, + "grad_norm": 1.5693423748016357, + "learning_rate": 2.7826955287362077e-05, + "loss": 0.4749, + "num_input_tokens_seen": 5285304, + "step": 19745 + }, + { + "epoch": 5.176933158584535, + "grad_norm": 2.7036166191101074, + "learning_rate": 2.7815591067495977e-05, + "loss": 0.7409, + "num_input_tokens_seen": 5286808, + "step": 19750 + }, + { + "epoch": 5.1782437745740495, + "grad_norm": 1.6109745502471924, + "learning_rate": 2.7804226258331535e-05, + "loss": 0.3407, + "num_input_tokens_seen": 5288200, + "step": 19755 + }, + { + "epoch": 5.179554390563565, + "grad_norm": 2.121126413345337, + "learning_rate": 2.7792860862247377e-05, + "loss": 0.5099, + "num_input_tokens_seen": 5289448, + "step": 19760 + }, + { + "epoch": 5.18086500655308, + "grad_norm": 2.623462438583374, + "learning_rate": 2.7781494881622268e-05, + "loss": 0.4533, + "num_input_tokens_seen": 5290888, + "step": 19765 + }, + { + "epoch": 5.182175622542595, + "grad_norm": 5.659484386444092, + "learning_rate": 2.7770128318835087e-05, + "loss": 0.4783, + "num_input_tokens_seen": 5292168, + "step": 19770 + }, + { + "epoch": 5.18348623853211, + "grad_norm": 6.555447101593018, + "learning_rate": 2.7758761176264835e-05, + "loss": 0.4399, + "num_input_tokens_seen": 5293368, + "step": 19775 + }, + { + "epoch": 5.184796854521625, + "grad_norm": 2.597019910812378, + "learning_rate": 2.7747393456290644e-05, + "loss": 0.4266, + "num_input_tokens_seen": 5294584, + "step": 19780 + }, + { + "epoch": 5.18610747051114, + "grad_norm": 2.7053232192993164, + "learning_rate": 2.7736025161291734e-05, + "loss": 0.3919, + "num_input_tokens_seen": 5295832, + "step": 19785 + }, + { + "epoch": 5.187418086500656, + "grad_norm": 2.7043638229370117, + "learning_rate": 2.7724656293647494e-05, + "loss": 0.3655, + "num_input_tokens_seen": 5296920, + "step": 19790 + }, + { + "epoch": 5.18872870249017, + "grad_norm": 3.4037115573883057, + "learning_rate": 2.7713286855737403e-05, + "loss": 0.4869, + "num_input_tokens_seen": 5298328, + "step": 19795 + }, + { + "epoch": 5.190039318479686, + "grad_norm": 2.1539926528930664, + "learning_rate": 2.7701916849941056e-05, + "loss": 0.5196, + "num_input_tokens_seen": 5299608, + "step": 19800 + }, + { + "epoch": 5.1913499344692005, + "grad_norm": 1.84861421585083, + "learning_rate": 2.7690546278638184e-05, + "loss": 0.3467, + "num_input_tokens_seen": 5300792, + "step": 19805 + }, + { + "epoch": 5.192660550458716, + "grad_norm": 1.9525718688964844, + "learning_rate": 2.7679175144208626e-05, + "loss": 0.4412, + "num_input_tokens_seen": 5302696, + "step": 19810 + }, + { + "epoch": 5.193971166448231, + "grad_norm": 3.139885902404785, + "learning_rate": 2.7667803449032333e-05, + "loss": 0.5886, + "num_input_tokens_seen": 5304104, + "step": 19815 + }, + { + "epoch": 5.195281782437746, + "grad_norm": 4.294738292694092, + "learning_rate": 2.765643119548939e-05, + "loss": 0.3714, + "num_input_tokens_seen": 5305352, + "step": 19820 + }, + { + "epoch": 5.196592398427261, + "grad_norm": 2.9013073444366455, + "learning_rate": 2.7645058385959987e-05, + "loss": 0.5025, + "num_input_tokens_seen": 5306776, + "step": 19825 + }, + { + "epoch": 5.1979030144167755, + "grad_norm": 2.4881339073181152, + "learning_rate": 2.7633685022824436e-05, + "loss": 0.328, + "num_input_tokens_seen": 5307848, + "step": 19830 + }, + { + "epoch": 5.199213630406291, + "grad_norm": 2.1865665912628174, + "learning_rate": 2.7622311108463167e-05, + "loss": 0.5893, + "num_input_tokens_seen": 5309128, + "step": 19835 + }, + { + "epoch": 5.200524246395806, + "grad_norm": 7.392697811126709, + "learning_rate": 2.76109366452567e-05, + "loss": 0.3656, + "num_input_tokens_seen": 5310280, + "step": 19840 + }, + { + "epoch": 5.201834862385321, + "grad_norm": 5.173124313354492, + "learning_rate": 2.7599561635585713e-05, + "loss": 0.8381, + "num_input_tokens_seen": 5311640, + "step": 19845 + }, + { + "epoch": 5.203145478374836, + "grad_norm": 5.286405086517334, + "learning_rate": 2.7588186081830965e-05, + "loss": 0.602, + "num_input_tokens_seen": 5313064, + "step": 19850 + }, + { + "epoch": 5.204456094364351, + "grad_norm": 1.7886662483215332, + "learning_rate": 2.7576809986373348e-05, + "loss": 0.4038, + "num_input_tokens_seen": 5314616, + "step": 19855 + }, + { + "epoch": 5.205766710353866, + "grad_norm": 1.9908447265625, + "learning_rate": 2.756543335159385e-05, + "loss": 0.6212, + "num_input_tokens_seen": 5315672, + "step": 19860 + }, + { + "epoch": 5.207077326343382, + "grad_norm": 6.02347469329834, + "learning_rate": 2.7554056179873583e-05, + "loss": 0.3513, + "num_input_tokens_seen": 5317352, + "step": 19865 + }, + { + "epoch": 5.208387942332896, + "grad_norm": 4.212154865264893, + "learning_rate": 2.7542678473593762e-05, + "loss": 0.3254, + "num_input_tokens_seen": 5318872, + "step": 19870 + }, + { + "epoch": 5.209698558322412, + "grad_norm": 2.8479158878326416, + "learning_rate": 2.753130023513574e-05, + "loss": 0.4618, + "num_input_tokens_seen": 5319880, + "step": 19875 + }, + { + "epoch": 5.2110091743119265, + "grad_norm": 2.4524242877960205, + "learning_rate": 2.7519921466880954e-05, + "loss": 0.4962, + "num_input_tokens_seen": 5320968, + "step": 19880 + }, + { + "epoch": 5.212319790301442, + "grad_norm": 3.212958335876465, + "learning_rate": 2.7508542171210953e-05, + "loss": 0.4933, + "num_input_tokens_seen": 5321976, + "step": 19885 + }, + { + "epoch": 5.213630406290957, + "grad_norm": 2.145601987838745, + "learning_rate": 2.7497162350507415e-05, + "loss": 0.4494, + "num_input_tokens_seen": 5323336, + "step": 19890 + }, + { + "epoch": 5.214941022280472, + "grad_norm": 1.9865138530731201, + "learning_rate": 2.7485782007152117e-05, + "loss": 0.5127, + "num_input_tokens_seen": 5324920, + "step": 19895 + }, + { + "epoch": 5.216251638269987, + "grad_norm": 9.532285690307617, + "learning_rate": 2.747440114352693e-05, + "loss": 0.3988, + "num_input_tokens_seen": 5326328, + "step": 19900 + }, + { + "epoch": 5.217562254259502, + "grad_norm": 3.1837916374206543, + "learning_rate": 2.7463019762013858e-05, + "loss": 0.4753, + "num_input_tokens_seen": 5327608, + "step": 19905 + }, + { + "epoch": 5.218872870249017, + "grad_norm": 2.9698474407196045, + "learning_rate": 2.745163786499501e-05, + "loss": 0.4799, + "num_input_tokens_seen": 5328616, + "step": 19910 + }, + { + "epoch": 5.220183486238533, + "grad_norm": 2.140730142593384, + "learning_rate": 2.74402554548526e-05, + "loss": 0.486, + "num_input_tokens_seen": 5329864, + "step": 19915 + }, + { + "epoch": 5.221494102228047, + "grad_norm": 5.720542907714844, + "learning_rate": 2.742887253396893e-05, + "loss": 0.4955, + "num_input_tokens_seen": 5330968, + "step": 19920 + }, + { + "epoch": 5.222804718217562, + "grad_norm": 1.9971528053283691, + "learning_rate": 2.7417489104726435e-05, + "loss": 0.4379, + "num_input_tokens_seen": 5332200, + "step": 19925 + }, + { + "epoch": 5.224115334207077, + "grad_norm": 1.0482745170593262, + "learning_rate": 2.740610516950765e-05, + "loss": 0.5409, + "num_input_tokens_seen": 5334024, + "step": 19930 + }, + { + "epoch": 5.225425950196592, + "grad_norm": 5.290614604949951, + "learning_rate": 2.739472073069521e-05, + "loss": 0.3549, + "num_input_tokens_seen": 5335064, + "step": 19935 + }, + { + "epoch": 5.226736566186108, + "grad_norm": 1.7220996618270874, + "learning_rate": 2.738333579067186e-05, + "loss": 0.4179, + "num_input_tokens_seen": 5336424, + "step": 19940 + }, + { + "epoch": 5.228047182175622, + "grad_norm": 2.044811248779297, + "learning_rate": 2.7371950351820445e-05, + "loss": 0.2891, + "num_input_tokens_seen": 5337736, + "step": 19945 + }, + { + "epoch": 5.229357798165138, + "grad_norm": 3.0841870307922363, + "learning_rate": 2.736056441652392e-05, + "loss": 0.3696, + "num_input_tokens_seen": 5338936, + "step": 19950 + }, + { + "epoch": 5.2306684141546524, + "grad_norm": 4.920641899108887, + "learning_rate": 2.7349177987165342e-05, + "loss": 0.4038, + "num_input_tokens_seen": 5340600, + "step": 19955 + }, + { + "epoch": 5.231979030144168, + "grad_norm": 1.4615036249160767, + "learning_rate": 2.7337791066127872e-05, + "loss": 0.3502, + "num_input_tokens_seen": 5341624, + "step": 19960 + }, + { + "epoch": 5.233289646133683, + "grad_norm": 3.1438074111938477, + "learning_rate": 2.732640365579477e-05, + "loss": 0.559, + "num_input_tokens_seen": 5342952, + "step": 19965 + }, + { + "epoch": 5.234600262123198, + "grad_norm": 9.085844993591309, + "learning_rate": 2.73150157585494e-05, + "loss": 0.731, + "num_input_tokens_seen": 5344472, + "step": 19970 + }, + { + "epoch": 5.235910878112713, + "grad_norm": 2.3312671184539795, + "learning_rate": 2.7303627376775242e-05, + "loss": 0.4532, + "num_input_tokens_seen": 5345544, + "step": 19975 + }, + { + "epoch": 5.237221494102228, + "grad_norm": 5.815787315368652, + "learning_rate": 2.7292238512855844e-05, + "loss": 0.4077, + "num_input_tokens_seen": 5346520, + "step": 19980 + }, + { + "epoch": 5.238532110091743, + "grad_norm": 1.5695937871932983, + "learning_rate": 2.7280849169174887e-05, + "loss": 0.854, + "num_input_tokens_seen": 5347720, + "step": 19985 + }, + { + "epoch": 5.239842726081259, + "grad_norm": 1.6477127075195312, + "learning_rate": 2.7269459348116143e-05, + "loss": 0.3601, + "num_input_tokens_seen": 5349192, + "step": 19990 + }, + { + "epoch": 5.241153342070773, + "grad_norm": 2.1316113471984863, + "learning_rate": 2.7258069052063477e-05, + "loss": 0.4288, + "num_input_tokens_seen": 5350616, + "step": 19995 + }, + { + "epoch": 5.242463958060289, + "grad_norm": 0.7099505066871643, + "learning_rate": 2.724667828340086e-05, + "loss": 0.28, + "num_input_tokens_seen": 5352296, + "step": 20000 + }, + { + "epoch": 5.243774574049803, + "grad_norm": 2.2469849586486816, + "learning_rate": 2.7235287044512365e-05, + "loss": 0.2318, + "num_input_tokens_seen": 5353704, + "step": 20005 + }, + { + "epoch": 5.245085190039318, + "grad_norm": 8.039796829223633, + "learning_rate": 2.7223895337782145e-05, + "loss": 0.3966, + "num_input_tokens_seen": 5354568, + "step": 20010 + }, + { + "epoch": 5.246395806028834, + "grad_norm": 4.465226650238037, + "learning_rate": 2.7212503165594483e-05, + "loss": 0.3416, + "num_input_tokens_seen": 5355848, + "step": 20015 + }, + { + "epoch": 5.247706422018348, + "grad_norm": 1.6047344207763672, + "learning_rate": 2.7201110530333722e-05, + "loss": 0.3954, + "num_input_tokens_seen": 5357256, + "step": 20020 + }, + { + "epoch": 5.249017038007864, + "grad_norm": 2.626105546951294, + "learning_rate": 2.7189717434384338e-05, + "loss": 0.5955, + "num_input_tokens_seen": 5358408, + "step": 20025 + }, + { + "epoch": 5.250327653997378, + "grad_norm": 1.9031754732131958, + "learning_rate": 2.7178323880130878e-05, + "loss": 0.5383, + "num_input_tokens_seen": 5359624, + "step": 20030 + }, + { + "epoch": 5.251638269986894, + "grad_norm": 6.015999794006348, + "learning_rate": 2.716692986995799e-05, + "loss": 0.5224, + "num_input_tokens_seen": 5360744, + "step": 20035 + }, + { + "epoch": 5.252948885976409, + "grad_norm": 4.3567352294921875, + "learning_rate": 2.7155535406250425e-05, + "loss": 0.321, + "num_input_tokens_seen": 5362344, + "step": 20040 + }, + { + "epoch": 5.254259501965924, + "grad_norm": 2.510427713394165, + "learning_rate": 2.7144140491393018e-05, + "loss": 0.506, + "num_input_tokens_seen": 5364168, + "step": 20045 + }, + { + "epoch": 5.255570117955439, + "grad_norm": 1.1367766857147217, + "learning_rate": 2.713274512777072e-05, + "loss": 0.4739, + "num_input_tokens_seen": 5365752, + "step": 20050 + }, + { + "epoch": 5.256880733944954, + "grad_norm": 3.398630142211914, + "learning_rate": 2.7121349317768552e-05, + "loss": 0.5291, + "num_input_tokens_seen": 5366904, + "step": 20055 + }, + { + "epoch": 5.258191349934469, + "grad_norm": 3.432382822036743, + "learning_rate": 2.710995306377163e-05, + "loss": 0.434, + "num_input_tokens_seen": 5368088, + "step": 20060 + }, + { + "epoch": 5.259501965923985, + "grad_norm": 10.818266868591309, + "learning_rate": 2.7098556368165186e-05, + "loss": 0.6076, + "num_input_tokens_seen": 5369176, + "step": 20065 + }, + { + "epoch": 5.260812581913499, + "grad_norm": 2.0465095043182373, + "learning_rate": 2.7087159233334512e-05, + "loss": 0.4909, + "num_input_tokens_seen": 5370456, + "step": 20070 + }, + { + "epoch": 5.262123197903015, + "grad_norm": 2.6542563438415527, + "learning_rate": 2.707576166166502e-05, + "loss": 0.3903, + "num_input_tokens_seen": 5371448, + "step": 20075 + }, + { + "epoch": 5.263433813892529, + "grad_norm": 1.3819283246994019, + "learning_rate": 2.7064363655542203e-05, + "loss": 0.3536, + "num_input_tokens_seen": 5372728, + "step": 20080 + }, + { + "epoch": 5.264744429882045, + "grad_norm": 2.1111507415771484, + "learning_rate": 2.7052965217351633e-05, + "loss": 0.4593, + "num_input_tokens_seen": 5374152, + "step": 20085 + }, + { + "epoch": 5.26605504587156, + "grad_norm": 3.1083672046661377, + "learning_rate": 2.7041566349479003e-05, + "loss": 0.5169, + "num_input_tokens_seen": 5375688, + "step": 20090 + }, + { + "epoch": 5.267365661861074, + "grad_norm": 1.2064719200134277, + "learning_rate": 2.7030167054310052e-05, + "loss": 0.3064, + "num_input_tokens_seen": 5377432, + "step": 20095 + }, + { + "epoch": 5.26867627785059, + "grad_norm": 1.3516491651535034, + "learning_rate": 2.701876733423064e-05, + "loss": 0.4624, + "num_input_tokens_seen": 5379528, + "step": 20100 + }, + { + "epoch": 5.269986893840104, + "grad_norm": 2.2981925010681152, + "learning_rate": 2.700736719162672e-05, + "loss": 0.3057, + "num_input_tokens_seen": 5380632, + "step": 20105 + }, + { + "epoch": 5.27129750982962, + "grad_norm": 2.5443809032440186, + "learning_rate": 2.699596662888432e-05, + "loss": 0.2858, + "num_input_tokens_seen": 5382168, + "step": 20110 + }, + { + "epoch": 5.272608125819135, + "grad_norm": 1.9319891929626465, + "learning_rate": 2.698456564838955e-05, + "loss": 0.5891, + "num_input_tokens_seen": 5386136, + "step": 20115 + }, + { + "epoch": 5.27391874180865, + "grad_norm": 4.413564682006836, + "learning_rate": 2.6973164252528615e-05, + "loss": 0.5884, + "num_input_tokens_seen": 5387624, + "step": 20120 + }, + { + "epoch": 5.275229357798165, + "grad_norm": 3.027351140975952, + "learning_rate": 2.696176244368781e-05, + "loss": 0.4189, + "num_input_tokens_seen": 5388904, + "step": 20125 + }, + { + "epoch": 5.27653997378768, + "grad_norm": 0.8992070555686951, + "learning_rate": 2.6950360224253524e-05, + "loss": 1.1464, + "num_input_tokens_seen": 5390568, + "step": 20130 + }, + { + "epoch": 5.277850589777195, + "grad_norm": 1.3824018239974976, + "learning_rate": 2.6938957596612203e-05, + "loss": 0.4941, + "num_input_tokens_seen": 5391768, + "step": 20135 + }, + { + "epoch": 5.2791612057667106, + "grad_norm": 4.905298709869385, + "learning_rate": 2.6927554563150414e-05, + "loss": 0.3994, + "num_input_tokens_seen": 5393032, + "step": 20140 + }, + { + "epoch": 5.280471821756225, + "grad_norm": 5.209184646606445, + "learning_rate": 2.6916151126254785e-05, + "loss": 0.5664, + "num_input_tokens_seen": 5394632, + "step": 20145 + }, + { + "epoch": 5.281782437745741, + "grad_norm": 4.331415176391602, + "learning_rate": 2.6904747288312036e-05, + "loss": 0.4643, + "num_input_tokens_seen": 5395816, + "step": 20150 + }, + { + "epoch": 5.283093053735255, + "grad_norm": 4.563414573669434, + "learning_rate": 2.689334305170897e-05, + "loss": 0.3469, + "num_input_tokens_seen": 5396888, + "step": 20155 + }, + { + "epoch": 5.284403669724771, + "grad_norm": 3.6273555755615234, + "learning_rate": 2.688193841883247e-05, + "loss": 0.4392, + "num_input_tokens_seen": 5398088, + "step": 20160 + }, + { + "epoch": 5.285714285714286, + "grad_norm": 4.467587947845459, + "learning_rate": 2.6870533392069518e-05, + "loss": 0.4576, + "num_input_tokens_seen": 5399304, + "step": 20165 + }, + { + "epoch": 5.287024901703801, + "grad_norm": 8.936509132385254, + "learning_rate": 2.6859127973807158e-05, + "loss": 0.4213, + "num_input_tokens_seen": 5400424, + "step": 20170 + }, + { + "epoch": 5.288335517693316, + "grad_norm": 2.442530393600464, + "learning_rate": 2.6847722166432522e-05, + "loss": 0.6694, + "num_input_tokens_seen": 5401912, + "step": 20175 + }, + { + "epoch": 5.289646133682831, + "grad_norm": 4.2523722648620605, + "learning_rate": 2.6836315972332827e-05, + "loss": 0.6712, + "num_input_tokens_seen": 5403208, + "step": 20180 + }, + { + "epoch": 5.290956749672346, + "grad_norm": 2.231657028198242, + "learning_rate": 2.6824909393895375e-05, + "loss": 0.3737, + "num_input_tokens_seen": 5404328, + "step": 20185 + }, + { + "epoch": 5.292267365661861, + "grad_norm": 3.22268009185791, + "learning_rate": 2.681350243350754e-05, + "loss": 0.6709, + "num_input_tokens_seen": 5405560, + "step": 20190 + }, + { + "epoch": 5.293577981651376, + "grad_norm": 1.5491092205047607, + "learning_rate": 2.680209509355678e-05, + "loss": 0.4821, + "num_input_tokens_seen": 5407016, + "step": 20195 + }, + { + "epoch": 5.294888597640891, + "grad_norm": 5.1705217361450195, + "learning_rate": 2.679068737643063e-05, + "loss": 0.4868, + "num_input_tokens_seen": 5408232, + "step": 20200 + }, + { + "epoch": 5.296199213630406, + "grad_norm": 1.5876654386520386, + "learning_rate": 2.6779279284516706e-05, + "loss": 0.4551, + "num_input_tokens_seen": 5409496, + "step": 20205 + }, + { + "epoch": 5.297509829619921, + "grad_norm": 1.472095251083374, + "learning_rate": 2.6767870820202707e-05, + "loss": 0.3871, + "num_input_tokens_seen": 5410968, + "step": 20210 + }, + { + "epoch": 5.2988204456094365, + "grad_norm": 3.425081729888916, + "learning_rate": 2.675646198587639e-05, + "loss": 0.2989, + "num_input_tokens_seen": 5412392, + "step": 20215 + }, + { + "epoch": 5.300131061598951, + "grad_norm": 1.5591347217559814, + "learning_rate": 2.6745052783925616e-05, + "loss": 0.4563, + "num_input_tokens_seen": 5413816, + "step": 20220 + }, + { + "epoch": 5.301441677588467, + "grad_norm": 4.302979469299316, + "learning_rate": 2.6733643216738325e-05, + "loss": 0.4241, + "num_input_tokens_seen": 5414888, + "step": 20225 + }, + { + "epoch": 5.302752293577981, + "grad_norm": 2.286207437515259, + "learning_rate": 2.6722233286702492e-05, + "loss": 0.3426, + "num_input_tokens_seen": 5415992, + "step": 20230 + }, + { + "epoch": 5.304062909567497, + "grad_norm": 8.410097122192383, + "learning_rate": 2.671082299620621e-05, + "loss": 0.6032, + "num_input_tokens_seen": 5416888, + "step": 20235 + }, + { + "epoch": 5.305373525557012, + "grad_norm": 1.2563362121582031, + "learning_rate": 2.6699412347637626e-05, + "loss": 0.3813, + "num_input_tokens_seen": 5418360, + "step": 20240 + }, + { + "epoch": 5.306684141546527, + "grad_norm": 1.4723410606384277, + "learning_rate": 2.6688001343384977e-05, + "loss": 0.4755, + "num_input_tokens_seen": 5419784, + "step": 20245 + }, + { + "epoch": 5.307994757536042, + "grad_norm": 5.560487747192383, + "learning_rate": 2.6676589985836563e-05, + "loss": 0.5973, + "num_input_tokens_seen": 5421048, + "step": 20250 + }, + { + "epoch": 5.309305373525557, + "grad_norm": 2.0771560668945312, + "learning_rate": 2.6665178277380766e-05, + "loss": 0.5357, + "num_input_tokens_seen": 5422728, + "step": 20255 + }, + { + "epoch": 5.310615989515072, + "grad_norm": 6.98594856262207, + "learning_rate": 2.6653766220406022e-05, + "loss": 0.5138, + "num_input_tokens_seen": 5424024, + "step": 20260 + }, + { + "epoch": 5.3119266055045875, + "grad_norm": 2.5329458713531494, + "learning_rate": 2.664235381730088e-05, + "loss": 0.3372, + "num_input_tokens_seen": 5425096, + "step": 20265 + }, + { + "epoch": 5.313237221494102, + "grad_norm": 4.535355091094971, + "learning_rate": 2.663094107045391e-05, + "loss": 0.3657, + "num_input_tokens_seen": 5427000, + "step": 20270 + }, + { + "epoch": 5.314547837483618, + "grad_norm": 3.0388221740722656, + "learning_rate": 2.6619527982253794e-05, + "loss": 0.439, + "num_input_tokens_seen": 5428088, + "step": 20275 + }, + { + "epoch": 5.315858453473132, + "grad_norm": 3.3377795219421387, + "learning_rate": 2.6608114555089275e-05, + "loss": 0.7612, + "num_input_tokens_seen": 5429384, + "step": 20280 + }, + { + "epoch": 5.317169069462647, + "grad_norm": 1.285828709602356, + "learning_rate": 2.659670079134916e-05, + "loss": 0.5246, + "num_input_tokens_seen": 5430904, + "step": 20285 + }, + { + "epoch": 5.3184796854521625, + "grad_norm": 4.398950576782227, + "learning_rate": 2.6585286693422322e-05, + "loss": 0.5146, + "num_input_tokens_seen": 5432104, + "step": 20290 + }, + { + "epoch": 5.319790301441677, + "grad_norm": 3.8557729721069336, + "learning_rate": 2.6573872263697724e-05, + "loss": 0.3709, + "num_input_tokens_seen": 5433560, + "step": 20295 + }, + { + "epoch": 5.321100917431193, + "grad_norm": 12.130836486816406, + "learning_rate": 2.6562457504564375e-05, + "loss": 0.4209, + "num_input_tokens_seen": 5435032, + "step": 20300 + }, + { + "epoch": 5.322411533420707, + "grad_norm": 1.6508535146713257, + "learning_rate": 2.655104241841137e-05, + "loss": 0.9514, + "num_input_tokens_seen": 5436168, + "step": 20305 + }, + { + "epoch": 5.323722149410223, + "grad_norm": 3.972233533859253, + "learning_rate": 2.653962700762787e-05, + "loss": 0.4499, + "num_input_tokens_seen": 5437304, + "step": 20310 + }, + { + "epoch": 5.325032765399738, + "grad_norm": 2.5493645668029785, + "learning_rate": 2.6528211274603092e-05, + "loss": 0.5105, + "num_input_tokens_seen": 5438488, + "step": 20315 + }, + { + "epoch": 5.326343381389253, + "grad_norm": 1.6787323951721191, + "learning_rate": 2.6516795221726344e-05, + "loss": 0.4733, + "num_input_tokens_seen": 5439848, + "step": 20320 + }, + { + "epoch": 5.327653997378768, + "grad_norm": 2.3684630393981934, + "learning_rate": 2.6505378851386976e-05, + "loss": 0.4262, + "num_input_tokens_seen": 5441176, + "step": 20325 + }, + { + "epoch": 5.328964613368283, + "grad_norm": 4.962738037109375, + "learning_rate": 2.6493962165974405e-05, + "loss": 0.7867, + "num_input_tokens_seen": 5442456, + "step": 20330 + }, + { + "epoch": 5.330275229357798, + "grad_norm": 2.185181140899658, + "learning_rate": 2.648254516787814e-05, + "loss": 0.398, + "num_input_tokens_seen": 5443688, + "step": 20335 + }, + { + "epoch": 5.3315858453473135, + "grad_norm": 3.6152892112731934, + "learning_rate": 2.6471127859487732e-05, + "loss": 0.4534, + "num_input_tokens_seen": 5444952, + "step": 20340 + }, + { + "epoch": 5.332896461336828, + "grad_norm": 3.485536813735962, + "learning_rate": 2.6459710243192816e-05, + "loss": 0.6006, + "num_input_tokens_seen": 5446824, + "step": 20345 + }, + { + "epoch": 5.334207077326344, + "grad_norm": 3.94844651222229, + "learning_rate": 2.6448292321383057e-05, + "loss": 0.3745, + "num_input_tokens_seen": 5447976, + "step": 20350 + }, + { + "epoch": 5.335517693315858, + "grad_norm": 1.787806749343872, + "learning_rate": 2.6436874096448216e-05, + "loss": 0.5214, + "num_input_tokens_seen": 5449800, + "step": 20355 + }, + { + "epoch": 5.336828309305374, + "grad_norm": 1.7314717769622803, + "learning_rate": 2.642545557077811e-05, + "loss": 0.3888, + "num_input_tokens_seen": 5450872, + "step": 20360 + }, + { + "epoch": 5.3381389252948885, + "grad_norm": 2.4283955097198486, + "learning_rate": 2.641403674676262e-05, + "loss": 0.5939, + "num_input_tokens_seen": 5452168, + "step": 20365 + }, + { + "epoch": 5.339449541284404, + "grad_norm": 2.4746766090393066, + "learning_rate": 2.6402617626791682e-05, + "loss": 0.3953, + "num_input_tokens_seen": 5453624, + "step": 20370 + }, + { + "epoch": 5.340760157273919, + "grad_norm": 1.4470292329788208, + "learning_rate": 2.6391198213255297e-05, + "loss": 0.2493, + "num_input_tokens_seen": 5454776, + "step": 20375 + }, + { + "epoch": 5.342070773263433, + "grad_norm": 1.9413390159606934, + "learning_rate": 2.637977850854353e-05, + "loss": 0.5622, + "num_input_tokens_seen": 5455976, + "step": 20380 + }, + { + "epoch": 5.343381389252949, + "grad_norm": 3.9772591590881348, + "learning_rate": 2.6368358515046503e-05, + "loss": 0.3138, + "num_input_tokens_seen": 5457208, + "step": 20385 + }, + { + "epoch": 5.344692005242464, + "grad_norm": 3.105205774307251, + "learning_rate": 2.6356938235154404e-05, + "loss": 0.543, + "num_input_tokens_seen": 5458792, + "step": 20390 + }, + { + "epoch": 5.346002621231979, + "grad_norm": 4.096020698547363, + "learning_rate": 2.634551767125748e-05, + "loss": 0.3966, + "num_input_tokens_seen": 5459944, + "step": 20395 + }, + { + "epoch": 5.347313237221494, + "grad_norm": 2.1689059734344482, + "learning_rate": 2.6334096825746035e-05, + "loss": 0.4885, + "num_input_tokens_seen": 5461256, + "step": 20400 + }, + { + "epoch": 5.348623853211009, + "grad_norm": 1.8309191465377808, + "learning_rate": 2.6322675701010423e-05, + "loss": 0.5498, + "num_input_tokens_seen": 5462856, + "step": 20405 + }, + { + "epoch": 5.349934469200524, + "grad_norm": 1.4674630165100098, + "learning_rate": 2.6311254299441072e-05, + "loss": 0.5841, + "num_input_tokens_seen": 5464536, + "step": 20410 + }, + { + "epoch": 5.3512450851900395, + "grad_norm": 5.534310817718506, + "learning_rate": 2.629983262342846e-05, + "loss": 0.4307, + "num_input_tokens_seen": 5465992, + "step": 20415 + }, + { + "epoch": 5.352555701179554, + "grad_norm": 0.728522777557373, + "learning_rate": 2.628841067536314e-05, + "loss": 0.4264, + "num_input_tokens_seen": 5467720, + "step": 20420 + }, + { + "epoch": 5.35386631716907, + "grad_norm": 1.0563637018203735, + "learning_rate": 2.6276988457635675e-05, + "loss": 0.3285, + "num_input_tokens_seen": 5468664, + "step": 20425 + }, + { + "epoch": 5.355176933158584, + "grad_norm": 3.977853775024414, + "learning_rate": 2.626556597263674e-05, + "loss": 0.4627, + "num_input_tokens_seen": 5470120, + "step": 20430 + }, + { + "epoch": 5.3564875491481, + "grad_norm": 2.3904850482940674, + "learning_rate": 2.625414322275703e-05, + "loss": 0.4856, + "num_input_tokens_seen": 5471560, + "step": 20435 + }, + { + "epoch": 5.3577981651376145, + "grad_norm": 1.3468854427337646, + "learning_rate": 2.6242720210387317e-05, + "loss": 0.4731, + "num_input_tokens_seen": 5472920, + "step": 20440 + }, + { + "epoch": 5.35910878112713, + "grad_norm": 1.1212303638458252, + "learning_rate": 2.623129693791841e-05, + "loss": 0.3677, + "num_input_tokens_seen": 5474104, + "step": 20445 + }, + { + "epoch": 5.360419397116645, + "grad_norm": 4.526829242706299, + "learning_rate": 2.6219873407741175e-05, + "loss": 0.3963, + "num_input_tokens_seen": 5475624, + "step": 20450 + }, + { + "epoch": 5.36173001310616, + "grad_norm": 12.39188003540039, + "learning_rate": 2.620844962224655e-05, + "loss": 0.5207, + "num_input_tokens_seen": 5476904, + "step": 20455 + }, + { + "epoch": 5.363040629095675, + "grad_norm": 5.011063098907471, + "learning_rate": 2.6197025583825508e-05, + "loss": 0.4695, + "num_input_tokens_seen": 5478344, + "step": 20460 + }, + { + "epoch": 5.3643512450851905, + "grad_norm": 1.1026123762130737, + "learning_rate": 2.6185601294869077e-05, + "loss": 0.4011, + "num_input_tokens_seen": 5479960, + "step": 20465 + }, + { + "epoch": 5.365661861074705, + "grad_norm": 5.528193473815918, + "learning_rate": 2.617417675776834e-05, + "loss": 0.4287, + "num_input_tokens_seen": 5481416, + "step": 20470 + }, + { + "epoch": 5.36697247706422, + "grad_norm": 3.321394920349121, + "learning_rate": 2.6162751974914433e-05, + "loss": 0.4491, + "num_input_tokens_seen": 5482648, + "step": 20475 + }, + { + "epoch": 5.368283093053735, + "grad_norm": 4.2532782554626465, + "learning_rate": 2.6151326948698557e-05, + "loss": 0.4552, + "num_input_tokens_seen": 5483960, + "step": 20480 + }, + { + "epoch": 5.36959370904325, + "grad_norm": 2.993577718734741, + "learning_rate": 2.6139901681511935e-05, + "loss": 0.4275, + "num_input_tokens_seen": 5485176, + "step": 20485 + }, + { + "epoch": 5.3709043250327655, + "grad_norm": 2.0315232276916504, + "learning_rate": 2.6128476175745865e-05, + "loss": 0.4354, + "num_input_tokens_seen": 5486376, + "step": 20490 + }, + { + "epoch": 5.37221494102228, + "grad_norm": 2.6111176013946533, + "learning_rate": 2.6117050433791672e-05, + "loss": 0.6914, + "num_input_tokens_seen": 5487448, + "step": 20495 + }, + { + "epoch": 5.373525557011796, + "grad_norm": 5.6291422843933105, + "learning_rate": 2.6105624458040763e-05, + "loss": 0.7548, + "num_input_tokens_seen": 5488488, + "step": 20500 + }, + { + "epoch": 5.37483617300131, + "grad_norm": 2.263366460800171, + "learning_rate": 2.6094198250884562e-05, + "loss": 0.5972, + "num_input_tokens_seen": 5490328, + "step": 20505 + }, + { + "epoch": 5.376146788990826, + "grad_norm": 1.5063613653182983, + "learning_rate": 2.6082771814714552e-05, + "loss": 0.5316, + "num_input_tokens_seen": 5492392, + "step": 20510 + }, + { + "epoch": 5.3774574049803405, + "grad_norm": 3.474027156829834, + "learning_rate": 2.6071345151922276e-05, + "loss": 0.405, + "num_input_tokens_seen": 5494024, + "step": 20515 + }, + { + "epoch": 5.378768020969856, + "grad_norm": 7.2791314125061035, + "learning_rate": 2.605991826489932e-05, + "loss": 0.483, + "num_input_tokens_seen": 5495288, + "step": 20520 + }, + { + "epoch": 5.380078636959371, + "grad_norm": 1.4587273597717285, + "learning_rate": 2.604849115603729e-05, + "loss": 0.5032, + "num_input_tokens_seen": 5496344, + "step": 20525 + }, + { + "epoch": 5.381389252948886, + "grad_norm": 3.9445066452026367, + "learning_rate": 2.6037063827727874e-05, + "loss": 0.6473, + "num_input_tokens_seen": 5497496, + "step": 20530 + }, + { + "epoch": 5.382699868938401, + "grad_norm": 4.725435733795166, + "learning_rate": 2.6025636282362792e-05, + "loss": 0.3677, + "num_input_tokens_seen": 5498760, + "step": 20535 + }, + { + "epoch": 5.3840104849279165, + "grad_norm": 6.0677714347839355, + "learning_rate": 2.6014208522333815e-05, + "loss": 0.5606, + "num_input_tokens_seen": 5500488, + "step": 20540 + }, + { + "epoch": 5.385321100917431, + "grad_norm": 8.741910934448242, + "learning_rate": 2.6002780550032747e-05, + "loss": 0.389, + "num_input_tokens_seen": 5501528, + "step": 20545 + }, + { + "epoch": 5.386631716906947, + "grad_norm": 2.61211895942688, + "learning_rate": 2.5991352367851435e-05, + "loss": 0.4705, + "num_input_tokens_seen": 5502872, + "step": 20550 + }, + { + "epoch": 5.387942332896461, + "grad_norm": 2.170980215072632, + "learning_rate": 2.5979923978181787e-05, + "loss": 0.4063, + "num_input_tokens_seen": 5504328, + "step": 20555 + }, + { + "epoch": 5.389252948885977, + "grad_norm": 1.3111993074417114, + "learning_rate": 2.5968495383415742e-05, + "loss": 0.3304, + "num_input_tokens_seen": 5506120, + "step": 20560 + }, + { + "epoch": 5.3905635648754915, + "grad_norm": 14.069317817687988, + "learning_rate": 2.595706658594529e-05, + "loss": 0.6812, + "num_input_tokens_seen": 5507304, + "step": 20565 + }, + { + "epoch": 5.391874180865006, + "grad_norm": 4.2196478843688965, + "learning_rate": 2.5945637588162452e-05, + "loss": 0.4868, + "num_input_tokens_seen": 5508376, + "step": 20570 + }, + { + "epoch": 5.393184796854522, + "grad_norm": 7.516977787017822, + "learning_rate": 2.5934208392459304e-05, + "loss": 0.5129, + "num_input_tokens_seen": 5509496, + "step": 20575 + }, + { + "epoch": 5.394495412844036, + "grad_norm": 1.087876796722412, + "learning_rate": 2.592277900122796e-05, + "loss": 0.4454, + "num_input_tokens_seen": 5511112, + "step": 20580 + }, + { + "epoch": 5.395806028833552, + "grad_norm": 2.6800405979156494, + "learning_rate": 2.5911349416860558e-05, + "loss": 0.253, + "num_input_tokens_seen": 5512280, + "step": 20585 + }, + { + "epoch": 5.3971166448230665, + "grad_norm": 1.830464482307434, + "learning_rate": 2.5899919641749286e-05, + "loss": 0.5247, + "num_input_tokens_seen": 5513304, + "step": 20590 + }, + { + "epoch": 5.398427260812582, + "grad_norm": 3.505066156387329, + "learning_rate": 2.588848967828641e-05, + "loss": 0.5651, + "num_input_tokens_seen": 5514232, + "step": 20595 + }, + { + "epoch": 5.399737876802097, + "grad_norm": 3.447442054748535, + "learning_rate": 2.5877059528864168e-05, + "loss": 0.5075, + "num_input_tokens_seen": 5515560, + "step": 20600 + }, + { + "epoch": 5.401048492791612, + "grad_norm": 1.9964892864227295, + "learning_rate": 2.5865629195874886e-05, + "loss": 0.3489, + "num_input_tokens_seen": 5516712, + "step": 20605 + }, + { + "epoch": 5.402359108781127, + "grad_norm": 1.854394793510437, + "learning_rate": 2.58541986817109e-05, + "loss": 0.3987, + "num_input_tokens_seen": 5518136, + "step": 20610 + }, + { + "epoch": 5.4036697247706424, + "grad_norm": 1.484655737876892, + "learning_rate": 2.5842767988764615e-05, + "loss": 0.5072, + "num_input_tokens_seen": 5519832, + "step": 20615 + }, + { + "epoch": 5.404980340760157, + "grad_norm": 2.829101085662842, + "learning_rate": 2.5831337119428446e-05, + "loss": 0.5259, + "num_input_tokens_seen": 5520904, + "step": 20620 + }, + { + "epoch": 5.406290956749673, + "grad_norm": 3.976296901702881, + "learning_rate": 2.5819906076094848e-05, + "loss": 0.6369, + "num_input_tokens_seen": 5522376, + "step": 20625 + }, + { + "epoch": 5.407601572739187, + "grad_norm": 3.966970920562744, + "learning_rate": 2.5808474861156328e-05, + "loss": 0.3206, + "num_input_tokens_seen": 5524088, + "step": 20630 + }, + { + "epoch": 5.408912188728703, + "grad_norm": 5.187856197357178, + "learning_rate": 2.5797043477005423e-05, + "loss": 0.3233, + "num_input_tokens_seen": 5524888, + "step": 20635 + }, + { + "epoch": 5.4102228047182175, + "grad_norm": 2.572458505630493, + "learning_rate": 2.5785611926034686e-05, + "loss": 0.5918, + "num_input_tokens_seen": 5526344, + "step": 20640 + }, + { + "epoch": 5.411533420707733, + "grad_norm": 2.3978147506713867, + "learning_rate": 2.5774180210636728e-05, + "loss": 0.3196, + "num_input_tokens_seen": 5527528, + "step": 20645 + }, + { + "epoch": 5.412844036697248, + "grad_norm": 3.738821268081665, + "learning_rate": 2.5762748333204196e-05, + "loss": 0.4003, + "num_input_tokens_seen": 5528696, + "step": 20650 + }, + { + "epoch": 5.414154652686763, + "grad_norm": 3.0731019973754883, + "learning_rate": 2.575131629612976e-05, + "loss": 0.504, + "num_input_tokens_seen": 5529928, + "step": 20655 + }, + { + "epoch": 5.415465268676278, + "grad_norm": 5.221053123474121, + "learning_rate": 2.5739884101806117e-05, + "loss": 0.4411, + "num_input_tokens_seen": 5531320, + "step": 20660 + }, + { + "epoch": 5.4167758846657925, + "grad_norm": 2.2466657161712646, + "learning_rate": 2.5728451752626008e-05, + "loss": 0.3792, + "num_input_tokens_seen": 5532680, + "step": 20665 + }, + { + "epoch": 5.418086500655308, + "grad_norm": 3.244718313217163, + "learning_rate": 2.571701925098221e-05, + "loss": 0.4966, + "num_input_tokens_seen": 5533672, + "step": 20670 + }, + { + "epoch": 5.419397116644823, + "grad_norm": 1.4007140398025513, + "learning_rate": 2.5705586599267518e-05, + "loss": 0.2591, + "num_input_tokens_seen": 5535112, + "step": 20675 + }, + { + "epoch": 5.420707732634338, + "grad_norm": 2.837409019470215, + "learning_rate": 2.569415379987477e-05, + "loss": 0.3824, + "num_input_tokens_seen": 5536376, + "step": 20680 + }, + { + "epoch": 5.422018348623853, + "grad_norm": 3.4815657138824463, + "learning_rate": 2.5682720855196836e-05, + "loss": 0.5181, + "num_input_tokens_seen": 5537704, + "step": 20685 + }, + { + "epoch": 5.423328964613368, + "grad_norm": 3.8575923442840576, + "learning_rate": 2.56712877676266e-05, + "loss": 0.3573, + "num_input_tokens_seen": 5538952, + "step": 20690 + }, + { + "epoch": 5.424639580602883, + "grad_norm": 5.185723781585693, + "learning_rate": 2.5659854539556998e-05, + "loss": 0.3696, + "num_input_tokens_seen": 5540328, + "step": 20695 + }, + { + "epoch": 5.425950196592399, + "grad_norm": 1.6216552257537842, + "learning_rate": 2.5648421173380977e-05, + "loss": 0.5528, + "num_input_tokens_seen": 5541576, + "step": 20700 + }, + { + "epoch": 5.427260812581913, + "grad_norm": 2.3479526042938232, + "learning_rate": 2.5636987671491525e-05, + "loss": 0.3818, + "num_input_tokens_seen": 5542792, + "step": 20705 + }, + { + "epoch": 5.428571428571429, + "grad_norm": 1.1184707880020142, + "learning_rate": 2.562555403628166e-05, + "loss": 0.4461, + "num_input_tokens_seen": 5544232, + "step": 20710 + }, + { + "epoch": 5.4298820445609435, + "grad_norm": 2.64208984375, + "learning_rate": 2.5614120270144415e-05, + "loss": 0.3467, + "num_input_tokens_seen": 5545416, + "step": 20715 + }, + { + "epoch": 5.431192660550459, + "grad_norm": 2.1641345024108887, + "learning_rate": 2.5602686375472856e-05, + "loss": 0.6568, + "num_input_tokens_seen": 5546728, + "step": 20720 + }, + { + "epoch": 5.432503276539974, + "grad_norm": 1.9829050302505493, + "learning_rate": 2.5591252354660083e-05, + "loss": 0.4898, + "num_input_tokens_seen": 5547944, + "step": 20725 + }, + { + "epoch": 5.433813892529489, + "grad_norm": 2.426543712615967, + "learning_rate": 2.557981821009921e-05, + "loss": 0.4687, + "num_input_tokens_seen": 5549240, + "step": 20730 + }, + { + "epoch": 5.435124508519004, + "grad_norm": 4.300614833831787, + "learning_rate": 2.556838394418339e-05, + "loss": 0.6613, + "num_input_tokens_seen": 5550616, + "step": 20735 + }, + { + "epoch": 5.436435124508519, + "grad_norm": 5.2445831298828125, + "learning_rate": 2.5556949559305786e-05, + "loss": 0.5266, + "num_input_tokens_seen": 5552200, + "step": 20740 + }, + { + "epoch": 5.437745740498034, + "grad_norm": 2.3947339057922363, + "learning_rate": 2.554551505785961e-05, + "loss": 0.4095, + "num_input_tokens_seen": 5553224, + "step": 20745 + }, + { + "epoch": 5.43905635648755, + "grad_norm": 4.496621131896973, + "learning_rate": 2.553408044223807e-05, + "loss": 0.5656, + "num_input_tokens_seen": 5554392, + "step": 20750 + }, + { + "epoch": 5.440366972477064, + "grad_norm": 2.486616611480713, + "learning_rate": 2.5522645714834426e-05, + "loss": 0.4104, + "num_input_tokens_seen": 5555464, + "step": 20755 + }, + { + "epoch": 5.441677588466579, + "grad_norm": 2.5493576526641846, + "learning_rate": 2.551121087804192e-05, + "loss": 0.3775, + "num_input_tokens_seen": 5556888, + "step": 20760 + }, + { + "epoch": 5.442988204456094, + "grad_norm": 0.9273203015327454, + "learning_rate": 2.5499775934253865e-05, + "loss": 0.4402, + "num_input_tokens_seen": 5558312, + "step": 20765 + }, + { + "epoch": 5.444298820445609, + "grad_norm": 1.9582091569900513, + "learning_rate": 2.5488340885863578e-05, + "loss": 0.5042, + "num_input_tokens_seen": 5560072, + "step": 20770 + }, + { + "epoch": 5.445609436435125, + "grad_norm": 1.845465064048767, + "learning_rate": 2.547690573526439e-05, + "loss": 0.3935, + "num_input_tokens_seen": 5561576, + "step": 20775 + }, + { + "epoch": 5.446920052424639, + "grad_norm": 2.8775644302368164, + "learning_rate": 2.546547048484965e-05, + "loss": 0.5438, + "num_input_tokens_seen": 5563016, + "step": 20780 + }, + { + "epoch": 5.448230668414155, + "grad_norm": 1.5237928628921509, + "learning_rate": 2.545403513701274e-05, + "loss": 0.6681, + "num_input_tokens_seen": 5564296, + "step": 20785 + }, + { + "epoch": 5.4495412844036695, + "grad_norm": 1.9765374660491943, + "learning_rate": 2.544259969414706e-05, + "loss": 0.5368, + "num_input_tokens_seen": 5565720, + "step": 20790 + }, + { + "epoch": 5.450851900393185, + "grad_norm": 2.318225145339966, + "learning_rate": 2.543116415864603e-05, + "loss": 0.3987, + "num_input_tokens_seen": 5566920, + "step": 20795 + }, + { + "epoch": 5.4521625163827, + "grad_norm": 1.7832832336425781, + "learning_rate": 2.541972853290309e-05, + "loss": 0.4431, + "num_input_tokens_seen": 5568456, + "step": 20800 + }, + { + "epoch": 5.453473132372215, + "grad_norm": 2.170212507247925, + "learning_rate": 2.540829281931169e-05, + "loss": 0.5644, + "num_input_tokens_seen": 5569944, + "step": 20805 + }, + { + "epoch": 5.45478374836173, + "grad_norm": 2.5700843334198, + "learning_rate": 2.5396857020265307e-05, + "loss": 0.4732, + "num_input_tokens_seen": 5571432, + "step": 20810 + }, + { + "epoch": 5.456094364351245, + "grad_norm": 1.48385751247406, + "learning_rate": 2.538542113815744e-05, + "loss": 0.4395, + "num_input_tokens_seen": 5573240, + "step": 20815 + }, + { + "epoch": 5.45740498034076, + "grad_norm": 2.1752681732177734, + "learning_rate": 2.5373985175381594e-05, + "loss": 0.3942, + "num_input_tokens_seen": 5574600, + "step": 20820 + }, + { + "epoch": 5.458715596330276, + "grad_norm": 2.0996649265289307, + "learning_rate": 2.536254913433129e-05, + "loss": 0.6651, + "num_input_tokens_seen": 5576152, + "step": 20825 + }, + { + "epoch": 5.46002621231979, + "grad_norm": 1.997583270072937, + "learning_rate": 2.535111301740009e-05, + "loss": 0.377, + "num_input_tokens_seen": 5577880, + "step": 20830 + }, + { + "epoch": 5.461336828309306, + "grad_norm": 3.9273083209991455, + "learning_rate": 2.5339676826981534e-05, + "loss": 0.5083, + "num_input_tokens_seen": 5579176, + "step": 20835 + }, + { + "epoch": 5.46264744429882, + "grad_norm": 4.572756767272949, + "learning_rate": 2.532824056546921e-05, + "loss": 0.3386, + "num_input_tokens_seen": 5580344, + "step": 20840 + }, + { + "epoch": 5.463958060288336, + "grad_norm": 4.009039878845215, + "learning_rate": 2.531680423525669e-05, + "loss": 0.4348, + "num_input_tokens_seen": 5582024, + "step": 20845 + }, + { + "epoch": 5.465268676277851, + "grad_norm": 2.9686598777770996, + "learning_rate": 2.5305367838737608e-05, + "loss": 0.3549, + "num_input_tokens_seen": 5583304, + "step": 20850 + }, + { + "epoch": 5.466579292267365, + "grad_norm": 2.1868274211883545, + "learning_rate": 2.5293931378305558e-05, + "loss": 0.4148, + "num_input_tokens_seen": 5584664, + "step": 20855 + }, + { + "epoch": 5.467889908256881, + "grad_norm": 2.259747266769409, + "learning_rate": 2.5282494856354176e-05, + "loss": 0.4406, + "num_input_tokens_seen": 5585944, + "step": 20860 + }, + { + "epoch": 5.4692005242463955, + "grad_norm": 3.8997769355773926, + "learning_rate": 2.5271058275277117e-05, + "loss": 0.5116, + "num_input_tokens_seen": 5587160, + "step": 20865 + }, + { + "epoch": 5.470511140235911, + "grad_norm": 1.662650227546692, + "learning_rate": 2.5259621637468016e-05, + "loss": 0.5889, + "num_input_tokens_seen": 5588664, + "step": 20870 + }, + { + "epoch": 5.471821756225426, + "grad_norm": 4.793196201324463, + "learning_rate": 2.5248184945320563e-05, + "loss": 0.4688, + "num_input_tokens_seen": 5589848, + "step": 20875 + }, + { + "epoch": 5.473132372214941, + "grad_norm": 1.218489170074463, + "learning_rate": 2.5236748201228428e-05, + "loss": 0.5247, + "num_input_tokens_seen": 5591496, + "step": 20880 + }, + { + "epoch": 5.474442988204456, + "grad_norm": 1.5466939210891724, + "learning_rate": 2.5225311407585306e-05, + "loss": 0.3491, + "num_input_tokens_seen": 5593000, + "step": 20885 + }, + { + "epoch": 5.475753604193971, + "grad_norm": 1.7825514078140259, + "learning_rate": 2.5213874566784902e-05, + "loss": 0.2924, + "num_input_tokens_seen": 5594504, + "step": 20890 + }, + { + "epoch": 5.477064220183486, + "grad_norm": 5.273685932159424, + "learning_rate": 2.5202437681220907e-05, + "loss": 0.3877, + "num_input_tokens_seen": 5595880, + "step": 20895 + }, + { + "epoch": 5.478374836173002, + "grad_norm": 2.2809507846832275, + "learning_rate": 2.5191000753287064e-05, + "loss": 0.6095, + "num_input_tokens_seen": 5597128, + "step": 20900 + }, + { + "epoch": 5.479685452162516, + "grad_norm": 1.419559359550476, + "learning_rate": 2.5179563785377088e-05, + "loss": 0.5482, + "num_input_tokens_seen": 5598776, + "step": 20905 + }, + { + "epoch": 5.480996068152032, + "grad_norm": 1.17024564743042, + "learning_rate": 2.5168126779884727e-05, + "loss": 0.4448, + "num_input_tokens_seen": 5600584, + "step": 20910 + }, + { + "epoch": 5.482306684141546, + "grad_norm": 9.674995422363281, + "learning_rate": 2.5156689739203714e-05, + "loss": 0.4024, + "num_input_tokens_seen": 5601704, + "step": 20915 + }, + { + "epoch": 5.483617300131062, + "grad_norm": 11.303888320922852, + "learning_rate": 2.5145252665727815e-05, + "loss": 0.4421, + "num_input_tokens_seen": 5603032, + "step": 20920 + }, + { + "epoch": 5.484927916120577, + "grad_norm": 8.33417797088623, + "learning_rate": 2.5133815561850782e-05, + "loss": 0.3863, + "num_input_tokens_seen": 5604360, + "step": 20925 + }, + { + "epoch": 5.486238532110092, + "grad_norm": 8.083722114562988, + "learning_rate": 2.5122378429966387e-05, + "loss": 0.5447, + "num_input_tokens_seen": 5605656, + "step": 20930 + }, + { + "epoch": 5.487549148099607, + "grad_norm": 3.197309970855713, + "learning_rate": 2.511094127246839e-05, + "loss": 0.6623, + "num_input_tokens_seen": 5606968, + "step": 20935 + }, + { + "epoch": 5.488859764089122, + "grad_norm": 2.0988714694976807, + "learning_rate": 2.5099504091750585e-05, + "loss": 0.4735, + "num_input_tokens_seen": 5608712, + "step": 20940 + }, + { + "epoch": 5.490170380078637, + "grad_norm": 2.473424196243286, + "learning_rate": 2.508806689020674e-05, + "loss": 0.4731, + "num_input_tokens_seen": 5610440, + "step": 20945 + }, + { + "epoch": 5.491480996068152, + "grad_norm": 5.0267438888549805, + "learning_rate": 2.5076629670230655e-05, + "loss": 0.4978, + "num_input_tokens_seen": 5611496, + "step": 20950 + }, + { + "epoch": 5.492791612057667, + "grad_norm": 2.2318503856658936, + "learning_rate": 2.5065192434216106e-05, + "loss": 0.4512, + "num_input_tokens_seen": 5612616, + "step": 20955 + }, + { + "epoch": 5.494102228047182, + "grad_norm": 4.403797626495361, + "learning_rate": 2.5053755184556886e-05, + "loss": 0.4049, + "num_input_tokens_seen": 5614296, + "step": 20960 + }, + { + "epoch": 5.495412844036697, + "grad_norm": 2.9404706954956055, + "learning_rate": 2.5042317923646812e-05, + "loss": 0.354, + "num_input_tokens_seen": 5615336, + "step": 20965 + }, + { + "epoch": 5.496723460026212, + "grad_norm": 4.672333240509033, + "learning_rate": 2.5030880653879667e-05, + "loss": 0.6662, + "num_input_tokens_seen": 5616376, + "step": 20970 + }, + { + "epoch": 5.498034076015728, + "grad_norm": 2.9846808910369873, + "learning_rate": 2.5019443377649247e-05, + "loss": 0.4678, + "num_input_tokens_seen": 5617624, + "step": 20975 + }, + { + "epoch": 5.499344692005242, + "grad_norm": 3.1386284828186035, + "learning_rate": 2.500800609734937e-05, + "loss": 0.3922, + "num_input_tokens_seen": 5619096, + "step": 20980 + }, + { + "epoch": 5.500655307994758, + "grad_norm": 1.4262598752975464, + "learning_rate": 2.4996568815373827e-05, + "loss": 0.3633, + "num_input_tokens_seen": 5620824, + "step": 20985 + }, + { + "epoch": 5.501441677588467, + "eval_loss": 0.4644910991191864, + "eval_runtime": 18.2245, + "eval_samples_per_second": 46.531, + "eval_steps_per_second": 23.265, + "num_input_tokens_seen": 5623240, + "step": 20988 + }, + { + "epoch": 5.501965923984272, + "grad_norm": 6.901313304901123, + "learning_rate": 2.4985131534116422e-05, + "loss": 0.6214, + "num_input_tokens_seen": 5623736, + "step": 20990 + }, + { + "epoch": 5.503276539973788, + "grad_norm": 1.6294701099395752, + "learning_rate": 2.4973694255970966e-05, + "loss": 0.5306, + "num_input_tokens_seen": 5625480, + "step": 20995 + }, + { + "epoch": 5.504587155963303, + "grad_norm": 4.836513519287109, + "learning_rate": 2.4962256983331254e-05, + "loss": 0.42, + "num_input_tokens_seen": 5626792, + "step": 21000 + }, + { + "epoch": 5.505897771952818, + "grad_norm": 2.6839962005615234, + "learning_rate": 2.4950819718591094e-05, + "loss": 0.3835, + "num_input_tokens_seen": 5628136, + "step": 21005 + }, + { + "epoch": 5.507208387942333, + "grad_norm": 2.892720937728882, + "learning_rate": 2.493938246414428e-05, + "loss": 0.4873, + "num_input_tokens_seen": 5629176, + "step": 21010 + }, + { + "epoch": 5.508519003931848, + "grad_norm": 2.5447826385498047, + "learning_rate": 2.4927945222384613e-05, + "loss": 0.3566, + "num_input_tokens_seen": 5630520, + "step": 21015 + }, + { + "epoch": 5.509829619921363, + "grad_norm": 5.950489521026611, + "learning_rate": 2.491650799570588e-05, + "loss": 0.5538, + "num_input_tokens_seen": 5631448, + "step": 21020 + }, + { + "epoch": 5.511140235910878, + "grad_norm": 1.7463992834091187, + "learning_rate": 2.4905070786501887e-05, + "loss": 0.4659, + "num_input_tokens_seen": 5632872, + "step": 21025 + }, + { + "epoch": 5.512450851900393, + "grad_norm": 2.1882195472717285, + "learning_rate": 2.4893633597166405e-05, + "loss": 0.5312, + "num_input_tokens_seen": 5634232, + "step": 21030 + }, + { + "epoch": 5.513761467889909, + "grad_norm": 5.30229377746582, + "learning_rate": 2.4882196430093233e-05, + "loss": 0.4008, + "num_input_tokens_seen": 5635736, + "step": 21035 + }, + { + "epoch": 5.515072083879423, + "grad_norm": 64.21373748779297, + "learning_rate": 2.4870759287676143e-05, + "loss": 0.5628, + "num_input_tokens_seen": 5637144, + "step": 21040 + }, + { + "epoch": 5.516382699868938, + "grad_norm": 1.8401591777801514, + "learning_rate": 2.485932217230892e-05, + "loss": 0.3623, + "num_input_tokens_seen": 5639128, + "step": 21045 + }, + { + "epoch": 5.517693315858454, + "grad_norm": 7.978780746459961, + "learning_rate": 2.4847885086385314e-05, + "loss": 0.5693, + "num_input_tokens_seen": 5640456, + "step": 21050 + }, + { + "epoch": 5.519003931847968, + "grad_norm": 7.3389506340026855, + "learning_rate": 2.4836448032299097e-05, + "loss": 0.6276, + "num_input_tokens_seen": 5641512, + "step": 21055 + }, + { + "epoch": 5.520314547837484, + "grad_norm": 2.6139872074127197, + "learning_rate": 2.4825011012444026e-05, + "loss": 0.4552, + "num_input_tokens_seen": 5643016, + "step": 21060 + }, + { + "epoch": 5.521625163826998, + "grad_norm": 2.8128511905670166, + "learning_rate": 2.4813574029213853e-05, + "loss": 0.6117, + "num_input_tokens_seen": 5644424, + "step": 21065 + }, + { + "epoch": 5.522935779816514, + "grad_norm": 1.7796908617019653, + "learning_rate": 2.4802137085002306e-05, + "loss": 0.4527, + "num_input_tokens_seen": 5645592, + "step": 21070 + }, + { + "epoch": 5.524246395806029, + "grad_norm": 1.7137088775634766, + "learning_rate": 2.479070018220313e-05, + "loss": 0.706, + "num_input_tokens_seen": 5647064, + "step": 21075 + }, + { + "epoch": 5.525557011795544, + "grad_norm": 3.3117759227752686, + "learning_rate": 2.4779263323210036e-05, + "loss": 0.8194, + "num_input_tokens_seen": 5648504, + "step": 21080 + }, + { + "epoch": 5.526867627785059, + "grad_norm": 4.075483322143555, + "learning_rate": 2.4767826510416753e-05, + "loss": 0.5115, + "num_input_tokens_seen": 5649784, + "step": 21085 + }, + { + "epoch": 5.528178243774574, + "grad_norm": 2.963872194290161, + "learning_rate": 2.4756389746216974e-05, + "loss": 0.61, + "num_input_tokens_seen": 5651240, + "step": 21090 + }, + { + "epoch": 5.529488859764089, + "grad_norm": 2.894514560699463, + "learning_rate": 2.4744953033004408e-05, + "loss": 0.4426, + "num_input_tokens_seen": 5652504, + "step": 21095 + }, + { + "epoch": 5.5307994757536045, + "grad_norm": 3.076230049133301, + "learning_rate": 2.4733516373172728e-05, + "loss": 0.4024, + "num_input_tokens_seen": 5653736, + "step": 21100 + }, + { + "epoch": 5.532110091743119, + "grad_norm": 2.0561630725860596, + "learning_rate": 2.4722079769115617e-05, + "loss": 0.4901, + "num_input_tokens_seen": 5654872, + "step": 21105 + }, + { + "epoch": 5.533420707732635, + "grad_norm": 1.3984758853912354, + "learning_rate": 2.4710643223226722e-05, + "loss": 0.4675, + "num_input_tokens_seen": 5656568, + "step": 21110 + }, + { + "epoch": 5.534731323722149, + "grad_norm": 1.2866641283035278, + "learning_rate": 2.4699206737899707e-05, + "loss": 0.4654, + "num_input_tokens_seen": 5658120, + "step": 21115 + }, + { + "epoch": 5.536041939711664, + "grad_norm": 2.991948366165161, + "learning_rate": 2.4687770315528195e-05, + "loss": 0.5438, + "num_input_tokens_seen": 5659096, + "step": 21120 + }, + { + "epoch": 5.53735255570118, + "grad_norm": 2.187286138534546, + "learning_rate": 2.4676333958505827e-05, + "loss": 0.4021, + "num_input_tokens_seen": 5660488, + "step": 21125 + }, + { + "epoch": 5.538663171690695, + "grad_norm": 5.7059712409973145, + "learning_rate": 2.4664897669226193e-05, + "loss": 0.4933, + "num_input_tokens_seen": 5661480, + "step": 21130 + }, + { + "epoch": 5.53997378768021, + "grad_norm": 0.9729784727096558, + "learning_rate": 2.46534614500829e-05, + "loss": 0.5899, + "num_input_tokens_seen": 5663144, + "step": 21135 + }, + { + "epoch": 5.541284403669724, + "grad_norm": 3.4356586933135986, + "learning_rate": 2.4642025303469527e-05, + "loss": 0.4416, + "num_input_tokens_seen": 5664360, + "step": 21140 + }, + { + "epoch": 5.54259501965924, + "grad_norm": 2.0192296504974365, + "learning_rate": 2.4630589231779633e-05, + "loss": 0.4358, + "num_input_tokens_seen": 5665688, + "step": 21145 + }, + { + "epoch": 5.543905635648755, + "grad_norm": 2.1229500770568848, + "learning_rate": 2.4619153237406782e-05, + "loss": 0.438, + "num_input_tokens_seen": 5666872, + "step": 21150 + }, + { + "epoch": 5.54521625163827, + "grad_norm": 3.065804958343506, + "learning_rate": 2.4607717322744498e-05, + "loss": 0.5004, + "num_input_tokens_seen": 5667992, + "step": 21155 + }, + { + "epoch": 5.546526867627785, + "grad_norm": 3.941155195236206, + "learning_rate": 2.4596281490186303e-05, + "loss": 0.5574, + "num_input_tokens_seen": 5669512, + "step": 21160 + }, + { + "epoch": 5.5478374836173, + "grad_norm": 1.8309992551803589, + "learning_rate": 2.4584845742125696e-05, + "loss": 0.4409, + "num_input_tokens_seen": 5671000, + "step": 21165 + }, + { + "epoch": 5.549148099606815, + "grad_norm": 4.105476379394531, + "learning_rate": 2.457341008095615e-05, + "loss": 0.3904, + "num_input_tokens_seen": 5672376, + "step": 21170 + }, + { + "epoch": 5.5504587155963305, + "grad_norm": 3.3862853050231934, + "learning_rate": 2.4561974509071143e-05, + "loss": 0.5474, + "num_input_tokens_seen": 5673672, + "step": 21175 + }, + { + "epoch": 5.551769331585845, + "grad_norm": 16.140005111694336, + "learning_rate": 2.4550539028864113e-05, + "loss": 0.4192, + "num_input_tokens_seen": 5674856, + "step": 21180 + }, + { + "epoch": 5.553079947575361, + "grad_norm": 2.880896806716919, + "learning_rate": 2.4539103642728494e-05, + "loss": 0.5071, + "num_input_tokens_seen": 5676232, + "step": 21185 + }, + { + "epoch": 5.554390563564875, + "grad_norm": 1.121984601020813, + "learning_rate": 2.4527668353057683e-05, + "loss": 0.2985, + "num_input_tokens_seen": 5677432, + "step": 21190 + }, + { + "epoch": 5.555701179554391, + "grad_norm": 4.29901123046875, + "learning_rate": 2.4516233162245063e-05, + "loss": 0.5014, + "num_input_tokens_seen": 5678712, + "step": 21195 + }, + { + "epoch": 5.557011795543906, + "grad_norm": 1.9604051113128662, + "learning_rate": 2.450479807268401e-05, + "loss": 0.3973, + "num_input_tokens_seen": 5680088, + "step": 21200 + }, + { + "epoch": 5.558322411533421, + "grad_norm": 2.1323347091674805, + "learning_rate": 2.449336308676786e-05, + "loss": 0.3399, + "num_input_tokens_seen": 5681192, + "step": 21205 + }, + { + "epoch": 5.559633027522936, + "grad_norm": 2.3578717708587646, + "learning_rate": 2.4481928206889946e-05, + "loss": 0.6478, + "num_input_tokens_seen": 5682280, + "step": 21210 + }, + { + "epoch": 5.56094364351245, + "grad_norm": 1.5566188097000122, + "learning_rate": 2.4470493435443558e-05, + "loss": 0.5305, + "num_input_tokens_seen": 5683512, + "step": 21215 + }, + { + "epoch": 5.562254259501966, + "grad_norm": 1.616485595703125, + "learning_rate": 2.4459058774821992e-05, + "loss": 0.3212, + "num_input_tokens_seen": 5684936, + "step": 21220 + }, + { + "epoch": 5.5635648754914815, + "grad_norm": 3.014721155166626, + "learning_rate": 2.4447624227418482e-05, + "loss": 0.4314, + "num_input_tokens_seen": 5686472, + "step": 21225 + }, + { + "epoch": 5.564875491480996, + "grad_norm": 3.7459945678710938, + "learning_rate": 2.4436189795626262e-05, + "loss": 0.4164, + "num_input_tokens_seen": 5687592, + "step": 21230 + }, + { + "epoch": 5.566186107470511, + "grad_norm": 1.4219378232955933, + "learning_rate": 2.442475548183855e-05, + "loss": 0.4619, + "num_input_tokens_seen": 5688792, + "step": 21235 + }, + { + "epoch": 5.567496723460026, + "grad_norm": 3.194453477859497, + "learning_rate": 2.4413321288448518e-05, + "loss": 0.6203, + "num_input_tokens_seen": 5689864, + "step": 21240 + }, + { + "epoch": 5.568807339449541, + "grad_norm": 7.635506629943848, + "learning_rate": 2.440188721784933e-05, + "loss": 0.3635, + "num_input_tokens_seen": 5690728, + "step": 21245 + }, + { + "epoch": 5.5701179554390565, + "grad_norm": 2.1941275596618652, + "learning_rate": 2.4390453272434115e-05, + "loss": 0.3931, + "num_input_tokens_seen": 5692376, + "step": 21250 + }, + { + "epoch": 5.571428571428571, + "grad_norm": 3.0057778358459473, + "learning_rate": 2.4379019454595973e-05, + "loss": 0.4303, + "num_input_tokens_seen": 5693704, + "step": 21255 + }, + { + "epoch": 5.572739187418087, + "grad_norm": 3.1277804374694824, + "learning_rate": 2.436758576672798e-05, + "loss": 0.3434, + "num_input_tokens_seen": 5694712, + "step": 21260 + }, + { + "epoch": 5.574049803407601, + "grad_norm": 1.8575791120529175, + "learning_rate": 2.4356152211223198e-05, + "loss": 0.8391, + "num_input_tokens_seen": 5695704, + "step": 21265 + }, + { + "epoch": 5.575360419397117, + "grad_norm": 1.4883688688278198, + "learning_rate": 2.434471879047465e-05, + "loss": 0.4358, + "num_input_tokens_seen": 5697176, + "step": 21270 + }, + { + "epoch": 5.576671035386632, + "grad_norm": 1.7822537422180176, + "learning_rate": 2.4333285506875322e-05, + "loss": 0.5177, + "num_input_tokens_seen": 5698712, + "step": 21275 + }, + { + "epoch": 5.577981651376147, + "grad_norm": 3.3753767013549805, + "learning_rate": 2.4321852362818195e-05, + "loss": 0.461, + "num_input_tokens_seen": 5700184, + "step": 21280 + }, + { + "epoch": 5.579292267365662, + "grad_norm": 3.9261958599090576, + "learning_rate": 2.431041936069619e-05, + "loss": 0.3591, + "num_input_tokens_seen": 5701368, + "step": 21285 + }, + { + "epoch": 5.580602883355177, + "grad_norm": 2.6209545135498047, + "learning_rate": 2.4298986502902218e-05, + "loss": 0.4495, + "num_input_tokens_seen": 5702552, + "step": 21290 + }, + { + "epoch": 5.581913499344692, + "grad_norm": 1.9974260330200195, + "learning_rate": 2.4287553791829166e-05, + "loss": 0.4078, + "num_input_tokens_seen": 5703992, + "step": 21295 + }, + { + "epoch": 5.5832241153342075, + "grad_norm": 1.916765809059143, + "learning_rate": 2.4276121229869875e-05, + "loss": 0.5461, + "num_input_tokens_seen": 5705112, + "step": 21300 + }, + { + "epoch": 5.584534731323722, + "grad_norm": 1.3534560203552246, + "learning_rate": 2.4264688819417155e-05, + "loss": 0.7161, + "num_input_tokens_seen": 5706120, + "step": 21305 + }, + { + "epoch": 5.585845347313237, + "grad_norm": 4.291120529174805, + "learning_rate": 2.4253256562863795e-05, + "loss": 0.4996, + "num_input_tokens_seen": 5707528, + "step": 21310 + }, + { + "epoch": 5.587155963302752, + "grad_norm": 1.4822478294372559, + "learning_rate": 2.4241824462602546e-05, + "loss": 0.6219, + "num_input_tokens_seen": 5708584, + "step": 21315 + }, + { + "epoch": 5.588466579292267, + "grad_norm": 7.942789554595947, + "learning_rate": 2.4230392521026123e-05, + "loss": 0.6228, + "num_input_tokens_seen": 5709784, + "step": 21320 + }, + { + "epoch": 5.5897771952817825, + "grad_norm": 2.1993000507354736, + "learning_rate": 2.4218960740527218e-05, + "loss": 0.5732, + "num_input_tokens_seen": 5711256, + "step": 21325 + }, + { + "epoch": 5.591087811271297, + "grad_norm": 11.703696250915527, + "learning_rate": 2.4207529123498485e-05, + "loss": 0.4821, + "num_input_tokens_seen": 5712328, + "step": 21330 + }, + { + "epoch": 5.592398427260813, + "grad_norm": 1.6497057676315308, + "learning_rate": 2.4196097672332533e-05, + "loss": 0.5398, + "num_input_tokens_seen": 5713640, + "step": 21335 + }, + { + "epoch": 5.593709043250327, + "grad_norm": 4.91134786605835, + "learning_rate": 2.4184666389421957e-05, + "loss": 0.4782, + "num_input_tokens_seen": 5714552, + "step": 21340 + }, + { + "epoch": 5.595019659239843, + "grad_norm": 2.011643886566162, + "learning_rate": 2.417323527715929e-05, + "loss": 0.4328, + "num_input_tokens_seen": 5715896, + "step": 21345 + }, + { + "epoch": 5.5963302752293576, + "grad_norm": 2.5836243629455566, + "learning_rate": 2.416180433793705e-05, + "loss": 0.3762, + "num_input_tokens_seen": 5717496, + "step": 21350 + }, + { + "epoch": 5.597640891218873, + "grad_norm": 2.6128151416778564, + "learning_rate": 2.4150373574147715e-05, + "loss": 0.5033, + "num_input_tokens_seen": 5719064, + "step": 21355 + }, + { + "epoch": 5.598951507208388, + "grad_norm": 2.8972017765045166, + "learning_rate": 2.413894298818373e-05, + "loss": 0.5552, + "num_input_tokens_seen": 5720296, + "step": 21360 + }, + { + "epoch": 5.600262123197903, + "grad_norm": 8.1451416015625, + "learning_rate": 2.4127512582437485e-05, + "loss": 0.4302, + "num_input_tokens_seen": 5721624, + "step": 21365 + }, + { + "epoch": 5.601572739187418, + "grad_norm": 4.39388370513916, + "learning_rate": 2.4116082359301346e-05, + "loss": 0.754, + "num_input_tokens_seen": 5723000, + "step": 21370 + }, + { + "epoch": 5.6028833551769335, + "grad_norm": 4.551173210144043, + "learning_rate": 2.4104652321167638e-05, + "loss": 0.53, + "num_input_tokens_seen": 5724168, + "step": 21375 + }, + { + "epoch": 5.604193971166448, + "grad_norm": 1.7179676294326782, + "learning_rate": 2.4093222470428657e-05, + "loss": 0.3494, + "num_input_tokens_seen": 5725400, + "step": 21380 + }, + { + "epoch": 5.605504587155964, + "grad_norm": 3.2913577556610107, + "learning_rate": 2.4081792809476643e-05, + "loss": 0.3654, + "num_input_tokens_seen": 5728184, + "step": 21385 + }, + { + "epoch": 5.606815203145478, + "grad_norm": 1.1844284534454346, + "learning_rate": 2.4070363340703812e-05, + "loss": 0.5154, + "num_input_tokens_seen": 5729544, + "step": 21390 + }, + { + "epoch": 5.608125819134994, + "grad_norm": 3.549842357635498, + "learning_rate": 2.4058934066502333e-05, + "loss": 0.3422, + "num_input_tokens_seen": 5731224, + "step": 21395 + }, + { + "epoch": 5.6094364351245085, + "grad_norm": 1.9593017101287842, + "learning_rate": 2.4047504989264318e-05, + "loss": 0.5777, + "num_input_tokens_seen": 5732584, + "step": 21400 + }, + { + "epoch": 5.610747051114023, + "grad_norm": 2.224504232406616, + "learning_rate": 2.4036076111381867e-05, + "loss": 0.6392, + "num_input_tokens_seen": 5733912, + "step": 21405 + }, + { + "epoch": 5.612057667103539, + "grad_norm": 3.329418182373047, + "learning_rate": 2.4024647435247018e-05, + "loss": 0.3727, + "num_input_tokens_seen": 5734904, + "step": 21410 + }, + { + "epoch": 5.613368283093053, + "grad_norm": 2.540369749069214, + "learning_rate": 2.4013218963251778e-05, + "loss": 0.424, + "num_input_tokens_seen": 5735896, + "step": 21415 + }, + { + "epoch": 5.614678899082569, + "grad_norm": 6.209473133087158, + "learning_rate": 2.4001790697788107e-05, + "loss": 0.4174, + "num_input_tokens_seen": 5736696, + "step": 21420 + }, + { + "epoch": 5.6159895150720835, + "grad_norm": 2.9603707790374756, + "learning_rate": 2.3990362641247915e-05, + "loss": 0.5089, + "num_input_tokens_seen": 5738008, + "step": 21425 + }, + { + "epoch": 5.617300131061599, + "grad_norm": 1.4801722764968872, + "learning_rate": 2.3978934796023076e-05, + "loss": 0.3931, + "num_input_tokens_seen": 5739416, + "step": 21430 + }, + { + "epoch": 5.618610747051114, + "grad_norm": 2.0892841815948486, + "learning_rate": 2.3967507164505416e-05, + "loss": 0.5264, + "num_input_tokens_seen": 5740952, + "step": 21435 + }, + { + "epoch": 5.619921363040629, + "grad_norm": 2.116448163986206, + "learning_rate": 2.3956079749086728e-05, + "loss": 0.441, + "num_input_tokens_seen": 5742376, + "step": 21440 + }, + { + "epoch": 5.621231979030144, + "grad_norm": 3.0835375785827637, + "learning_rate": 2.3944652552158742e-05, + "loss": 0.6469, + "num_input_tokens_seen": 5743800, + "step": 21445 + }, + { + "epoch": 5.6225425950196595, + "grad_norm": 3.6478865146636963, + "learning_rate": 2.3933225576113154e-05, + "loss": 0.4348, + "num_input_tokens_seen": 5744888, + "step": 21450 + }, + { + "epoch": 5.623853211009174, + "grad_norm": 2.6611061096191406, + "learning_rate": 2.392179882334162e-05, + "loss": 0.4482, + "num_input_tokens_seen": 5746088, + "step": 21455 + }, + { + "epoch": 5.62516382699869, + "grad_norm": 7.815239429473877, + "learning_rate": 2.3910372296235717e-05, + "loss": 0.4146, + "num_input_tokens_seen": 5747128, + "step": 21460 + }, + { + "epoch": 5.626474442988204, + "grad_norm": 3.3761816024780273, + "learning_rate": 2.3898945997187012e-05, + "loss": 0.5102, + "num_input_tokens_seen": 5748152, + "step": 21465 + }, + { + "epoch": 5.62778505897772, + "grad_norm": 2.203855276107788, + "learning_rate": 2.3887519928587007e-05, + "loss": 0.4652, + "num_input_tokens_seen": 5749736, + "step": 21470 + }, + { + "epoch": 5.6290956749672345, + "grad_norm": 3.647296905517578, + "learning_rate": 2.387609409282716e-05, + "loss": 0.3639, + "num_input_tokens_seen": 5751272, + "step": 21475 + }, + { + "epoch": 5.63040629095675, + "grad_norm": 2.520648956298828, + "learning_rate": 2.3864668492298884e-05, + "loss": 0.4194, + "num_input_tokens_seen": 5752904, + "step": 21480 + }, + { + "epoch": 5.631716906946265, + "grad_norm": 2.168586492538452, + "learning_rate": 2.3853243129393524e-05, + "loss": 0.5082, + "num_input_tokens_seen": 5754104, + "step": 21485 + }, + { + "epoch": 5.63302752293578, + "grad_norm": 2.4019343852996826, + "learning_rate": 2.384181800650239e-05, + "loss": 0.3347, + "num_input_tokens_seen": 5755544, + "step": 21490 + }, + { + "epoch": 5.634338138925295, + "grad_norm": 1.261440634727478, + "learning_rate": 2.383039312601675e-05, + "loss": 0.4695, + "num_input_tokens_seen": 5757288, + "step": 21495 + }, + { + "epoch": 5.6356487549148095, + "grad_norm": 0.9679906964302063, + "learning_rate": 2.3818968490327813e-05, + "loss": 0.4673, + "num_input_tokens_seen": 5758872, + "step": 21500 + }, + { + "epoch": 5.636959370904325, + "grad_norm": 6.144745349884033, + "learning_rate": 2.3807544101826722e-05, + "loss": 0.5969, + "num_input_tokens_seen": 5760248, + "step": 21505 + }, + { + "epoch": 5.63826998689384, + "grad_norm": 3.1001529693603516, + "learning_rate": 2.3796119962904597e-05, + "loss": 0.7592, + "num_input_tokens_seen": 5761688, + "step": 21510 + }, + { + "epoch": 5.639580602883355, + "grad_norm": 1.922358751296997, + "learning_rate": 2.3784696075952487e-05, + "loss": 0.4981, + "num_input_tokens_seen": 5763176, + "step": 21515 + }, + { + "epoch": 5.64089121887287, + "grad_norm": 2.3582077026367188, + "learning_rate": 2.3773272443361383e-05, + "loss": 0.4417, + "num_input_tokens_seen": 5764344, + "step": 21520 + }, + { + "epoch": 5.6422018348623855, + "grad_norm": 3.496306896209717, + "learning_rate": 2.3761849067522235e-05, + "loss": 0.6331, + "num_input_tokens_seen": 5765464, + "step": 21525 + }, + { + "epoch": 5.6435124508519, + "grad_norm": 2.0832605361938477, + "learning_rate": 2.3750425950825943e-05, + "loss": 0.3903, + "num_input_tokens_seen": 5766616, + "step": 21530 + }, + { + "epoch": 5.644823066841416, + "grad_norm": 2.7526092529296875, + "learning_rate": 2.3739003095663334e-05, + "loss": 0.4984, + "num_input_tokens_seen": 5767976, + "step": 21535 + }, + { + "epoch": 5.64613368283093, + "grad_norm": 6.826116561889648, + "learning_rate": 2.3727580504425208e-05, + "loss": 0.3902, + "num_input_tokens_seen": 5769064, + "step": 21540 + }, + { + "epoch": 5.647444298820446, + "grad_norm": 1.193244457244873, + "learning_rate": 2.3716158179502277e-05, + "loss": 0.5134, + "num_input_tokens_seen": 5770936, + "step": 21545 + }, + { + "epoch": 5.6487549148099605, + "grad_norm": 1.8233027458190918, + "learning_rate": 2.3704736123285214e-05, + "loss": 0.4116, + "num_input_tokens_seen": 5772216, + "step": 21550 + }, + { + "epoch": 5.650065530799476, + "grad_norm": 1.936830997467041, + "learning_rate": 2.369331433816465e-05, + "loss": 0.6083, + "num_input_tokens_seen": 5773624, + "step": 21555 + }, + { + "epoch": 5.651376146788991, + "grad_norm": 2.0360701084136963, + "learning_rate": 2.3681892826531137e-05, + "loss": 0.3873, + "num_input_tokens_seen": 5774744, + "step": 21560 + }, + { + "epoch": 5.652686762778506, + "grad_norm": 2.842482328414917, + "learning_rate": 2.3670471590775174e-05, + "loss": 0.4847, + "num_input_tokens_seen": 5776504, + "step": 21565 + }, + { + "epoch": 5.653997378768021, + "grad_norm": 15.015660285949707, + "learning_rate": 2.3659050633287217e-05, + "loss": 0.5256, + "num_input_tokens_seen": 5777432, + "step": 21570 + }, + { + "epoch": 5.655307994757536, + "grad_norm": 2.577393054962158, + "learning_rate": 2.3647629956457647e-05, + "loss": 0.3256, + "num_input_tokens_seen": 5778456, + "step": 21575 + }, + { + "epoch": 5.656618610747051, + "grad_norm": 1.3553190231323242, + "learning_rate": 2.3636209562676784e-05, + "loss": 0.514, + "num_input_tokens_seen": 5779832, + "step": 21580 + }, + { + "epoch": 5.657929226736567, + "grad_norm": 2.459245443344116, + "learning_rate": 2.3624789454334905e-05, + "loss": 0.3842, + "num_input_tokens_seen": 5781000, + "step": 21585 + }, + { + "epoch": 5.659239842726081, + "grad_norm": 153.4912109375, + "learning_rate": 2.3613369633822217e-05, + "loss": 1.3322, + "num_input_tokens_seen": 5782184, + "step": 21590 + }, + { + "epoch": 5.660550458715596, + "grad_norm": 3.8578438758850098, + "learning_rate": 2.360195010352887e-05, + "loss": 0.6402, + "num_input_tokens_seen": 5783768, + "step": 21595 + }, + { + "epoch": 5.6618610747051115, + "grad_norm": 3.284747838973999, + "learning_rate": 2.3590530865844952e-05, + "loss": 0.5927, + "num_input_tokens_seen": 5784856, + "step": 21600 + }, + { + "epoch": 5.663171690694626, + "grad_norm": 1.9943257570266724, + "learning_rate": 2.357911192316047e-05, + "loss": 0.6117, + "num_input_tokens_seen": 5786680, + "step": 21605 + }, + { + "epoch": 5.664482306684142, + "grad_norm": 2.9537675380706787, + "learning_rate": 2.356769327786543e-05, + "loss": 0.5926, + "num_input_tokens_seen": 5787848, + "step": 21610 + }, + { + "epoch": 5.665792922673656, + "grad_norm": 1.7383939027786255, + "learning_rate": 2.3556274932349703e-05, + "loss": 0.3236, + "num_input_tokens_seen": 5789112, + "step": 21615 + }, + { + "epoch": 5.667103538663172, + "grad_norm": 2.5420212745666504, + "learning_rate": 2.354485688900314e-05, + "loss": 0.3426, + "num_input_tokens_seen": 5790328, + "step": 21620 + }, + { + "epoch": 5.6684141546526865, + "grad_norm": 1.346182942390442, + "learning_rate": 2.3533439150215517e-05, + "loss": 0.4007, + "num_input_tokens_seen": 5791304, + "step": 21625 + }, + { + "epoch": 5.669724770642202, + "grad_norm": 4.020904064178467, + "learning_rate": 2.352202171837655e-05, + "loss": 0.449, + "num_input_tokens_seen": 5792584, + "step": 21630 + }, + { + "epoch": 5.671035386631717, + "grad_norm": 2.9626030921936035, + "learning_rate": 2.3510604595875877e-05, + "loss": 0.486, + "num_input_tokens_seen": 5793768, + "step": 21635 + }, + { + "epoch": 5.672346002621232, + "grad_norm": 1.7929707765579224, + "learning_rate": 2.3499187785103087e-05, + "loss": 0.4339, + "num_input_tokens_seen": 5795000, + "step": 21640 + }, + { + "epoch": 5.673656618610747, + "grad_norm": 1.3210010528564453, + "learning_rate": 2.3487771288447705e-05, + "loss": 0.5308, + "num_input_tokens_seen": 5796264, + "step": 21645 + }, + { + "epoch": 5.674967234600262, + "grad_norm": 22.244911193847656, + "learning_rate": 2.3476355108299174e-05, + "loss": 0.5633, + "num_input_tokens_seen": 5797624, + "step": 21650 + }, + { + "epoch": 5.676277850589777, + "grad_norm": 2.3738272190093994, + "learning_rate": 2.3464939247046896e-05, + "loss": 0.5053, + "num_input_tokens_seen": 5799224, + "step": 21655 + }, + { + "epoch": 5.677588466579293, + "grad_norm": 2.9229719638824463, + "learning_rate": 2.3453523707080165e-05, + "loss": 0.3569, + "num_input_tokens_seen": 5800488, + "step": 21660 + }, + { + "epoch": 5.678899082568807, + "grad_norm": 2.2811472415924072, + "learning_rate": 2.344210849078826e-05, + "loss": 0.601, + "num_input_tokens_seen": 5801880, + "step": 21665 + }, + { + "epoch": 5.680209698558322, + "grad_norm": 1.0830860137939453, + "learning_rate": 2.343069360056036e-05, + "loss": 0.3203, + "num_input_tokens_seen": 5803544, + "step": 21670 + }, + { + "epoch": 5.6815203145478375, + "grad_norm": 1.820557951927185, + "learning_rate": 2.3419279038785575e-05, + "loss": 0.4208, + "num_input_tokens_seen": 5804696, + "step": 21675 + }, + { + "epoch": 5.682830930537353, + "grad_norm": 1.0548561811447144, + "learning_rate": 2.3407864807852952e-05, + "loss": 0.5287, + "num_input_tokens_seen": 5805992, + "step": 21680 + }, + { + "epoch": 5.684141546526868, + "grad_norm": 5.549178600311279, + "learning_rate": 2.3396450910151476e-05, + "loss": 0.6084, + "num_input_tokens_seen": 5807336, + "step": 21685 + }, + { + "epoch": 5.685452162516382, + "grad_norm": 2.1804027557373047, + "learning_rate": 2.338503734807006e-05, + "loss": 0.4393, + "num_input_tokens_seen": 5808728, + "step": 21690 + }, + { + "epoch": 5.686762778505898, + "grad_norm": 4.8537397384643555, + "learning_rate": 2.337362412399753e-05, + "loss": 0.3714, + "num_input_tokens_seen": 5810216, + "step": 21695 + }, + { + "epoch": 5.6880733944954125, + "grad_norm": 2.2261226177215576, + "learning_rate": 2.3362211240322663e-05, + "loss": 0.5124, + "num_input_tokens_seen": 5811336, + "step": 21700 + }, + { + "epoch": 5.689384010484928, + "grad_norm": 3.2926924228668213, + "learning_rate": 2.335079869943415e-05, + "loss": 0.4838, + "num_input_tokens_seen": 5812536, + "step": 21705 + }, + { + "epoch": 5.690694626474443, + "grad_norm": 1.930579423904419, + "learning_rate": 2.333938650372062e-05, + "loss": 0.3668, + "num_input_tokens_seen": 5813896, + "step": 21710 + }, + { + "epoch": 5.692005242463958, + "grad_norm": 2.8191821575164795, + "learning_rate": 2.3327974655570633e-05, + "loss": 0.4479, + "num_input_tokens_seen": 5814952, + "step": 21715 + }, + { + "epoch": 5.693315858453473, + "grad_norm": 2.8498752117156982, + "learning_rate": 2.3316563157372645e-05, + "loss": 0.4743, + "num_input_tokens_seen": 5816152, + "step": 21720 + }, + { + "epoch": 5.694626474442988, + "grad_norm": 6.6694769859313965, + "learning_rate": 2.3305152011515098e-05, + "loss": 0.6352, + "num_input_tokens_seen": 5817224, + "step": 21725 + }, + { + "epoch": 5.695937090432503, + "grad_norm": 8.781671524047852, + "learning_rate": 2.3293741220386296e-05, + "loss": 0.4022, + "num_input_tokens_seen": 5818296, + "step": 21730 + }, + { + "epoch": 5.697247706422019, + "grad_norm": 3.8936784267425537, + "learning_rate": 2.328233078637451e-05, + "loss": 0.5403, + "num_input_tokens_seen": 5819464, + "step": 21735 + }, + { + "epoch": 5.698558322411533, + "grad_norm": 1.8732428550720215, + "learning_rate": 2.327092071186792e-05, + "loss": 0.4527, + "num_input_tokens_seen": 5820648, + "step": 21740 + }, + { + "epoch": 5.699868938401049, + "grad_norm": 2.2572367191314697, + "learning_rate": 2.3259510999254645e-05, + "loss": 0.4074, + "num_input_tokens_seen": 5821912, + "step": 21745 + }, + { + "epoch": 5.7011795543905635, + "grad_norm": 1.4986249208450317, + "learning_rate": 2.324810165092271e-05, + "loss": 0.5152, + "num_input_tokens_seen": 5823432, + "step": 21750 + }, + { + "epoch": 5.702490170380079, + "grad_norm": 2.2403624057769775, + "learning_rate": 2.323669266926007e-05, + "loss": 0.3745, + "num_input_tokens_seen": 5824440, + "step": 21755 + }, + { + "epoch": 5.703800786369594, + "grad_norm": 5.4116597175598145, + "learning_rate": 2.3225284056654607e-05, + "loss": 0.4877, + "num_input_tokens_seen": 5825800, + "step": 21760 + }, + { + "epoch": 5.705111402359108, + "grad_norm": 2.0568583011627197, + "learning_rate": 2.3213875815494122e-05, + "loss": 0.577, + "num_input_tokens_seen": 5827288, + "step": 21765 + }, + { + "epoch": 5.706422018348624, + "grad_norm": 6.40414571762085, + "learning_rate": 2.3202467948166343e-05, + "loss": 0.3997, + "num_input_tokens_seen": 5828680, + "step": 21770 + }, + { + "epoch": 5.707732634338139, + "grad_norm": 2.7680301666259766, + "learning_rate": 2.319106045705891e-05, + "loss": 0.5049, + "num_input_tokens_seen": 5830296, + "step": 21775 + }, + { + "epoch": 5.709043250327654, + "grad_norm": 1.0484168529510498, + "learning_rate": 2.317965334455941e-05, + "loss": 0.4204, + "num_input_tokens_seen": 5831832, + "step": 21780 + }, + { + "epoch": 5.710353866317169, + "grad_norm": 10.941601753234863, + "learning_rate": 2.3168246613055317e-05, + "loss": 0.4915, + "num_input_tokens_seen": 5833064, + "step": 21785 + }, + { + "epoch": 5.711664482306684, + "grad_norm": 2.262772560119629, + "learning_rate": 2.3156840264934045e-05, + "loss": 0.352, + "num_input_tokens_seen": 5834376, + "step": 21790 + }, + { + "epoch": 5.712975098296199, + "grad_norm": 19.03804588317871, + "learning_rate": 2.3145434302582913e-05, + "loss": 0.4182, + "num_input_tokens_seen": 5835512, + "step": 21795 + }, + { + "epoch": 5.714285714285714, + "grad_norm": 2.8741981983184814, + "learning_rate": 2.313402872838918e-05, + "loss": 0.5664, + "num_input_tokens_seen": 5836664, + "step": 21800 + }, + { + "epoch": 5.715596330275229, + "grad_norm": 5.144537448883057, + "learning_rate": 2.312262354474001e-05, + "loss": 0.4506, + "num_input_tokens_seen": 5837816, + "step": 21805 + }, + { + "epoch": 5.716906946264745, + "grad_norm": 2.0727908611297607, + "learning_rate": 2.3111218754022495e-05, + "loss": 0.5388, + "num_input_tokens_seen": 5839128, + "step": 21810 + }, + { + "epoch": 5.718217562254259, + "grad_norm": 2.074847936630249, + "learning_rate": 2.309981435862363e-05, + "loss": 0.3519, + "num_input_tokens_seen": 5840312, + "step": 21815 + }, + { + "epoch": 5.719528178243775, + "grad_norm": 2.0777246952056885, + "learning_rate": 2.308841036093033e-05, + "loss": 0.3633, + "num_input_tokens_seen": 5841800, + "step": 21820 + }, + { + "epoch": 5.7208387942332894, + "grad_norm": 6.8212714195251465, + "learning_rate": 2.307700676332944e-05, + "loss": 0.4317, + "num_input_tokens_seen": 5842984, + "step": 21825 + }, + { + "epoch": 5.722149410222805, + "grad_norm": 2.537108898162842, + "learning_rate": 2.3065603568207715e-05, + "loss": 0.5126, + "num_input_tokens_seen": 5844312, + "step": 21830 + }, + { + "epoch": 5.72346002621232, + "grad_norm": 2.6533751487731934, + "learning_rate": 2.3054200777951807e-05, + "loss": 0.4904, + "num_input_tokens_seen": 5845752, + "step": 21835 + }, + { + "epoch": 5.724770642201835, + "grad_norm": 1.6155344247817993, + "learning_rate": 2.3042798394948326e-05, + "loss": 0.5021, + "num_input_tokens_seen": 5846904, + "step": 21840 + }, + { + "epoch": 5.72608125819135, + "grad_norm": 2.3481805324554443, + "learning_rate": 2.303139642158376e-05, + "loss": 0.3731, + "num_input_tokens_seen": 5848216, + "step": 21845 + }, + { + "epoch": 5.727391874180865, + "grad_norm": 3.2254130840301514, + "learning_rate": 2.3019994860244523e-05, + "loss": 0.3971, + "num_input_tokens_seen": 5849752, + "step": 21850 + }, + { + "epoch": 5.72870249017038, + "grad_norm": 2.074455976486206, + "learning_rate": 2.3008593713316937e-05, + "loss": 0.462, + "num_input_tokens_seen": 5853224, + "step": 21855 + }, + { + "epoch": 5.730013106159895, + "grad_norm": 2.9052646160125732, + "learning_rate": 2.2997192983187245e-05, + "loss": 0.3083, + "num_input_tokens_seen": 5854408, + "step": 21860 + }, + { + "epoch": 5.73132372214941, + "grad_norm": 1.674842357635498, + "learning_rate": 2.2985792672241608e-05, + "loss": 0.4132, + "num_input_tokens_seen": 5855384, + "step": 21865 + }, + { + "epoch": 5.732634338138926, + "grad_norm": 1.2262144088745117, + "learning_rate": 2.2974392782866078e-05, + "loss": 0.572, + "num_input_tokens_seen": 5856728, + "step": 21870 + }, + { + "epoch": 5.73394495412844, + "grad_norm": 3.6379737854003906, + "learning_rate": 2.296299331744664e-05, + "loss": 0.4233, + "num_input_tokens_seen": 5858152, + "step": 21875 + }, + { + "epoch": 5.735255570117955, + "grad_norm": 2.53336501121521, + "learning_rate": 2.2951594278369178e-05, + "loss": 0.4993, + "num_input_tokens_seen": 5859480, + "step": 21880 + }, + { + "epoch": 5.736566186107471, + "grad_norm": 5.545539855957031, + "learning_rate": 2.2940195668019497e-05, + "loss": 0.5594, + "num_input_tokens_seen": 5860680, + "step": 21885 + }, + { + "epoch": 5.737876802096985, + "grad_norm": 2.1896474361419678, + "learning_rate": 2.2928797488783298e-05, + "loss": 0.4458, + "num_input_tokens_seen": 5862568, + "step": 21890 + }, + { + "epoch": 5.739187418086501, + "grad_norm": 1.3542163372039795, + "learning_rate": 2.2917399743046208e-05, + "loss": 0.6775, + "num_input_tokens_seen": 5864040, + "step": 21895 + }, + { + "epoch": 5.740498034076015, + "grad_norm": 1.9192805290222168, + "learning_rate": 2.2906002433193757e-05, + "loss": 0.6436, + "num_input_tokens_seen": 5865320, + "step": 21900 + }, + { + "epoch": 5.741808650065531, + "grad_norm": 2.878291606903076, + "learning_rate": 2.2894605561611382e-05, + "loss": 0.4036, + "num_input_tokens_seen": 5866696, + "step": 21905 + }, + { + "epoch": 5.743119266055046, + "grad_norm": 1.7805453538894653, + "learning_rate": 2.288320913068442e-05, + "loss": 0.4851, + "num_input_tokens_seen": 5868504, + "step": 21910 + }, + { + "epoch": 5.744429882044561, + "grad_norm": 2.24504017829895, + "learning_rate": 2.2871813142798133e-05, + "loss": 0.2739, + "num_input_tokens_seen": 5869624, + "step": 21915 + }, + { + "epoch": 5.745740498034076, + "grad_norm": 2.943147897720337, + "learning_rate": 2.2860417600337673e-05, + "loss": 0.3617, + "num_input_tokens_seen": 5870792, + "step": 21920 + }, + { + "epoch": 5.747051114023591, + "grad_norm": 1.8338762521743774, + "learning_rate": 2.284902250568812e-05, + "loss": 0.4207, + "num_input_tokens_seen": 5872696, + "step": 21925 + }, + { + "epoch": 5.748361730013106, + "grad_norm": 7.394570827484131, + "learning_rate": 2.2837627861234434e-05, + "loss": 0.3705, + "num_input_tokens_seen": 5874024, + "step": 21930 + }, + { + "epoch": 5.749672346002622, + "grad_norm": 2.5811171531677246, + "learning_rate": 2.2826233669361504e-05, + "loss": 0.4422, + "num_input_tokens_seen": 5876280, + "step": 21935 + }, + { + "epoch": 5.750982961992136, + "grad_norm": 8.28326416015625, + "learning_rate": 2.2814839932454103e-05, + "loss": 0.4208, + "num_input_tokens_seen": 5877688, + "step": 21940 + }, + { + "epoch": 5.752293577981652, + "grad_norm": 1.4144134521484375, + "learning_rate": 2.2803446652896935e-05, + "loss": 0.3527, + "num_input_tokens_seen": 5878904, + "step": 21945 + }, + { + "epoch": 5.753604193971166, + "grad_norm": 5.51743221282959, + "learning_rate": 2.279205383307458e-05, + "loss": 0.3744, + "num_input_tokens_seen": 5880184, + "step": 21950 + }, + { + "epoch": 5.754914809960681, + "grad_norm": 4.046903133392334, + "learning_rate": 2.278066147537155e-05, + "loss": 0.6616, + "num_input_tokens_seen": 5881512, + "step": 21955 + }, + { + "epoch": 5.756225425950197, + "grad_norm": 3.2102208137512207, + "learning_rate": 2.276926958217224e-05, + "loss": 0.4883, + "num_input_tokens_seen": 5882952, + "step": 21960 + }, + { + "epoch": 5.757536041939712, + "grad_norm": 8.523531913757324, + "learning_rate": 2.2757878155860948e-05, + "loss": 0.6174, + "num_input_tokens_seen": 5884088, + "step": 21965 + }, + { + "epoch": 5.758846657929227, + "grad_norm": 2.3888707160949707, + "learning_rate": 2.2746487198821883e-05, + "loss": 0.4358, + "num_input_tokens_seen": 5885048, + "step": 21970 + }, + { + "epoch": 5.760157273918741, + "grad_norm": 3.361614942550659, + "learning_rate": 2.2735096713439156e-05, + "loss": 0.5285, + "num_input_tokens_seen": 5885960, + "step": 21975 + }, + { + "epoch": 5.761467889908257, + "grad_norm": 1.6971075534820557, + "learning_rate": 2.2723706702096774e-05, + "loss": 0.5527, + "num_input_tokens_seen": 5887192, + "step": 21980 + }, + { + "epoch": 5.762778505897772, + "grad_norm": 1.3230466842651367, + "learning_rate": 2.2712317167178648e-05, + "loss": 0.3942, + "num_input_tokens_seen": 5888488, + "step": 21985 + }, + { + "epoch": 5.764089121887287, + "grad_norm": 1.1851367950439453, + "learning_rate": 2.2700928111068582e-05, + "loss": 0.4592, + "num_input_tokens_seen": 5890200, + "step": 21990 + }, + { + "epoch": 5.765399737876802, + "grad_norm": 1.9672540426254272, + "learning_rate": 2.2689539536150294e-05, + "loss": 0.5451, + "num_input_tokens_seen": 5891592, + "step": 21995 + }, + { + "epoch": 5.766710353866317, + "grad_norm": 2.241180419921875, + "learning_rate": 2.2678151444807384e-05, + "loss": 0.5579, + "num_input_tokens_seen": 5892888, + "step": 22000 + }, + { + "epoch": 5.768020969855832, + "grad_norm": 1.0076088905334473, + "learning_rate": 2.266676383942336e-05, + "loss": 0.3936, + "num_input_tokens_seen": 5894568, + "step": 22005 + }, + { + "epoch": 5.7693315858453476, + "grad_norm": 3.567314624786377, + "learning_rate": 2.265537672238164e-05, + "loss": 1.0165, + "num_input_tokens_seen": 5895464, + "step": 22010 + }, + { + "epoch": 5.770642201834862, + "grad_norm": 2.141348123550415, + "learning_rate": 2.2643990096065518e-05, + "loss": 0.3023, + "num_input_tokens_seen": 5896856, + "step": 22015 + }, + { + "epoch": 5.771952817824378, + "grad_norm": 3.323404312133789, + "learning_rate": 2.2632603962858207e-05, + "loss": 0.4924, + "num_input_tokens_seen": 5898008, + "step": 22020 + }, + { + "epoch": 5.773263433813892, + "grad_norm": 5.8209638595581055, + "learning_rate": 2.2621218325142788e-05, + "loss": 0.6611, + "num_input_tokens_seen": 5899448, + "step": 22025 + }, + { + "epoch": 5.774574049803408, + "grad_norm": 2.5337157249450684, + "learning_rate": 2.2609833185302262e-05, + "loss": 0.472, + "num_input_tokens_seen": 5900696, + "step": 22030 + }, + { + "epoch": 5.775884665792923, + "grad_norm": 4.188205718994141, + "learning_rate": 2.2598448545719525e-05, + "loss": 0.5438, + "num_input_tokens_seen": 5902008, + "step": 22035 + }, + { + "epoch": 5.777195281782438, + "grad_norm": 0.842322826385498, + "learning_rate": 2.2587064408777352e-05, + "loss": 0.4128, + "num_input_tokens_seen": 5903400, + "step": 22040 + }, + { + "epoch": 5.778505897771953, + "grad_norm": 0.9332048296928406, + "learning_rate": 2.257568077685844e-05, + "loss": 0.5206, + "num_input_tokens_seen": 5905464, + "step": 22045 + }, + { + "epoch": 5.779816513761467, + "grad_norm": 1.7305256128311157, + "learning_rate": 2.2564297652345347e-05, + "loss": 0.5684, + "num_input_tokens_seen": 5906968, + "step": 22050 + }, + { + "epoch": 5.781127129750983, + "grad_norm": 2.684548854827881, + "learning_rate": 2.255291503762054e-05, + "loss": 0.6137, + "num_input_tokens_seen": 5908184, + "step": 22055 + }, + { + "epoch": 5.7824377457404985, + "grad_norm": 3.785219192504883, + "learning_rate": 2.2541532935066394e-05, + "loss": 0.4481, + "num_input_tokens_seen": 5909096, + "step": 22060 + }, + { + "epoch": 5.783748361730013, + "grad_norm": 1.9247146844863892, + "learning_rate": 2.2530151347065153e-05, + "loss": 0.3675, + "num_input_tokens_seen": 5910296, + "step": 22065 + }, + { + "epoch": 5.785058977719528, + "grad_norm": 3.381842613220215, + "learning_rate": 2.2518770275998967e-05, + "loss": 0.4386, + "num_input_tokens_seen": 5911384, + "step": 22070 + }, + { + "epoch": 5.786369593709043, + "grad_norm": 1.3411957025527954, + "learning_rate": 2.2507389724249878e-05, + "loss": 0.4166, + "num_input_tokens_seen": 5912568, + "step": 22075 + }, + { + "epoch": 5.787680209698558, + "grad_norm": 2.0378355979919434, + "learning_rate": 2.2496009694199815e-05, + "loss": 0.9998, + "num_input_tokens_seen": 5913768, + "step": 22080 + }, + { + "epoch": 5.7889908256880735, + "grad_norm": 1.840600609779358, + "learning_rate": 2.2484630188230596e-05, + "loss": 0.5579, + "num_input_tokens_seen": 5915240, + "step": 22085 + }, + { + "epoch": 5.790301441677588, + "grad_norm": 4.852047443389893, + "learning_rate": 2.2473251208723927e-05, + "loss": 0.8095, + "num_input_tokens_seen": 5916376, + "step": 22090 + }, + { + "epoch": 5.791612057667104, + "grad_norm": 3.6883716583251953, + "learning_rate": 2.246187275806142e-05, + "loss": 0.4408, + "num_input_tokens_seen": 5917752, + "step": 22095 + }, + { + "epoch": 5.792922673656618, + "grad_norm": 5.625666618347168, + "learning_rate": 2.2450494838624554e-05, + "loss": 0.5393, + "num_input_tokens_seen": 5919144, + "step": 22100 + }, + { + "epoch": 5.794233289646134, + "grad_norm": 1.7626293897628784, + "learning_rate": 2.2439117452794723e-05, + "loss": 0.4751, + "num_input_tokens_seen": 5920728, + "step": 22105 + }, + { + "epoch": 5.795543905635649, + "grad_norm": 2.1020667552948, + "learning_rate": 2.2427740602953175e-05, + "loss": 0.5123, + "num_input_tokens_seen": 5922232, + "step": 22110 + }, + { + "epoch": 5.796854521625164, + "grad_norm": 4.6131911277771, + "learning_rate": 2.2416364291481075e-05, + "loss": 0.3131, + "num_input_tokens_seen": 5923368, + "step": 22115 + }, + { + "epoch": 5.798165137614679, + "grad_norm": 4.593222141265869, + "learning_rate": 2.240498852075946e-05, + "loss": 0.4032, + "num_input_tokens_seen": 5924536, + "step": 22120 + }, + { + "epoch": 5.799475753604194, + "grad_norm": 2.3518333435058594, + "learning_rate": 2.239361329316927e-05, + "loss": 0.3195, + "num_input_tokens_seen": 5925752, + "step": 22125 + }, + { + "epoch": 5.800786369593709, + "grad_norm": 2.7046048641204834, + "learning_rate": 2.2382238611091318e-05, + "loss": 0.5071, + "num_input_tokens_seen": 5926872, + "step": 22130 + }, + { + "epoch": 5.8020969855832245, + "grad_norm": 3.4741108417510986, + "learning_rate": 2.2370864476906296e-05, + "loss": 0.5296, + "num_input_tokens_seen": 5928056, + "step": 22135 + }, + { + "epoch": 5.803407601572739, + "grad_norm": 7.7444047927856445, + "learning_rate": 2.23594908929948e-05, + "loss": 0.5281, + "num_input_tokens_seen": 5929000, + "step": 22140 + }, + { + "epoch": 5.804718217562254, + "grad_norm": 1.8733488321304321, + "learning_rate": 2.2348117861737297e-05, + "loss": 0.4764, + "num_input_tokens_seen": 5930120, + "step": 22145 + }, + { + "epoch": 5.806028833551769, + "grad_norm": 3.3188607692718506, + "learning_rate": 2.233674538551414e-05, + "loss": 0.3757, + "num_input_tokens_seen": 5931672, + "step": 22150 + }, + { + "epoch": 5.807339449541285, + "grad_norm": 4.2684407234191895, + "learning_rate": 2.2325373466705573e-05, + "loss": 0.3914, + "num_input_tokens_seen": 5933032, + "step": 22155 + }, + { + "epoch": 5.8086500655307995, + "grad_norm": 2.3624632358551025, + "learning_rate": 2.2314002107691727e-05, + "loss": 0.4725, + "num_input_tokens_seen": 5934024, + "step": 22160 + }, + { + "epoch": 5.809960681520314, + "grad_norm": 2.585310935974121, + "learning_rate": 2.230263131085259e-05, + "loss": 0.4977, + "num_input_tokens_seen": 5935624, + "step": 22165 + }, + { + "epoch": 5.81127129750983, + "grad_norm": 5.45912504196167, + "learning_rate": 2.2291261078568057e-05, + "loss": 0.3909, + "num_input_tokens_seen": 5936776, + "step": 22170 + }, + { + "epoch": 5.812581913499344, + "grad_norm": 4.856486797332764, + "learning_rate": 2.22798914132179e-05, + "loss": 0.6314, + "num_input_tokens_seen": 5937752, + "step": 22175 + }, + { + "epoch": 5.81389252948886, + "grad_norm": 2.3336825370788574, + "learning_rate": 2.226852231718177e-05, + "loss": 0.4916, + "num_input_tokens_seen": 5939432, + "step": 22180 + }, + { + "epoch": 5.815203145478375, + "grad_norm": 2.3712806701660156, + "learning_rate": 2.2257153792839202e-05, + "loss": 0.3406, + "num_input_tokens_seen": 5940824, + "step": 22185 + }, + { + "epoch": 5.81651376146789, + "grad_norm": 3.730072021484375, + "learning_rate": 2.2245785842569604e-05, + "loss": 0.4442, + "num_input_tokens_seen": 5941992, + "step": 22190 + }, + { + "epoch": 5.817824377457405, + "grad_norm": 4.320250034332275, + "learning_rate": 2.2234418468752276e-05, + "loss": 0.6794, + "num_input_tokens_seen": 5943112, + "step": 22195 + }, + { + "epoch": 5.81913499344692, + "grad_norm": 1.5568982362747192, + "learning_rate": 2.222305167376638e-05, + "loss": 0.4441, + "num_input_tokens_seen": 5944360, + "step": 22200 + }, + { + "epoch": 5.820445609436435, + "grad_norm": 2.764535427093506, + "learning_rate": 2.221168545999097e-05, + "loss": 0.3293, + "num_input_tokens_seen": 5945528, + "step": 22205 + }, + { + "epoch": 5.8217562254259505, + "grad_norm": 3.344456911087036, + "learning_rate": 2.2200319829804978e-05, + "loss": 0.6511, + "num_input_tokens_seen": 5946648, + "step": 22210 + }, + { + "epoch": 5.823066841415465, + "grad_norm": 2.5733745098114014, + "learning_rate": 2.2188954785587207e-05, + "loss": 0.5402, + "num_input_tokens_seen": 5947992, + "step": 22215 + }, + { + "epoch": 5.824377457404981, + "grad_norm": 4.829158306121826, + "learning_rate": 2.2177590329716345e-05, + "loss": 0.445, + "num_input_tokens_seen": 5949320, + "step": 22220 + }, + { + "epoch": 5.825688073394495, + "grad_norm": 8.674412727355957, + "learning_rate": 2.2166226464570947e-05, + "loss": 0.5804, + "num_input_tokens_seen": 5950440, + "step": 22225 + }, + { + "epoch": 5.826998689384011, + "grad_norm": 4.2057671546936035, + "learning_rate": 2.2154863192529453e-05, + "loss": 0.3927, + "num_input_tokens_seen": 5952120, + "step": 22230 + }, + { + "epoch": 5.8283093053735255, + "grad_norm": 2.0921437740325928, + "learning_rate": 2.2143500515970173e-05, + "loss": 0.436, + "num_input_tokens_seen": 5953464, + "step": 22235 + }, + { + "epoch": 5.82961992136304, + "grad_norm": 3.4538607597351074, + "learning_rate": 2.21321384372713e-05, + "loss": 0.4501, + "num_input_tokens_seen": 5954440, + "step": 22240 + }, + { + "epoch": 5.830930537352556, + "grad_norm": 2.3911476135253906, + "learning_rate": 2.2120776958810902e-05, + "loss": 0.4012, + "num_input_tokens_seen": 5956152, + "step": 22245 + }, + { + "epoch": 5.832241153342071, + "grad_norm": 2.6633853912353516, + "learning_rate": 2.2109416082966907e-05, + "loss": 0.5718, + "num_input_tokens_seen": 5957416, + "step": 22250 + }, + { + "epoch": 5.833551769331586, + "grad_norm": 2.1679089069366455, + "learning_rate": 2.2098055812117143e-05, + "loss": 0.465, + "num_input_tokens_seen": 5959048, + "step": 22255 + }, + { + "epoch": 5.834862385321101, + "grad_norm": 2.5251474380493164, + "learning_rate": 2.208669614863927e-05, + "loss": 0.7751, + "num_input_tokens_seen": 5960328, + "step": 22260 + }, + { + "epoch": 5.836173001310616, + "grad_norm": 2.5136210918426514, + "learning_rate": 2.2075337094910868e-05, + "loss": 0.4313, + "num_input_tokens_seen": 5961608, + "step": 22265 + }, + { + "epoch": 5.837483617300131, + "grad_norm": 1.5877436399459839, + "learning_rate": 2.2063978653309354e-05, + "loss": 0.4479, + "num_input_tokens_seen": 5962856, + "step": 22270 + }, + { + "epoch": 5.838794233289646, + "grad_norm": 1.3315857648849487, + "learning_rate": 2.2052620826212035e-05, + "loss": 0.3944, + "num_input_tokens_seen": 5964232, + "step": 22275 + }, + { + "epoch": 5.840104849279161, + "grad_norm": 2.042814016342163, + "learning_rate": 2.204126361599609e-05, + "loss": 0.4497, + "num_input_tokens_seen": 5965624, + "step": 22280 + }, + { + "epoch": 5.8414154652686765, + "grad_norm": 2.4507384300231934, + "learning_rate": 2.2029907025038548e-05, + "loss": 0.4524, + "num_input_tokens_seen": 5966792, + "step": 22285 + }, + { + "epoch": 5.842726081258191, + "grad_norm": 3.2982139587402344, + "learning_rate": 2.201855105571634e-05, + "loss": 0.5683, + "num_input_tokens_seen": 5967912, + "step": 22290 + }, + { + "epoch": 5.844036697247707, + "grad_norm": 3.140080213546753, + "learning_rate": 2.2007195710406236e-05, + "loss": 0.5196, + "num_input_tokens_seen": 5969272, + "step": 22295 + }, + { + "epoch": 5.845347313237221, + "grad_norm": 5.550419807434082, + "learning_rate": 2.1995840991484906e-05, + "loss": 0.4107, + "num_input_tokens_seen": 5970328, + "step": 22300 + }, + { + "epoch": 5.846657929226737, + "grad_norm": 1.831133246421814, + "learning_rate": 2.198448690132886e-05, + "loss": 0.4779, + "num_input_tokens_seen": 5971768, + "step": 22305 + }, + { + "epoch": 5.8479685452162515, + "grad_norm": 3.0265955924987793, + "learning_rate": 2.19731334423145e-05, + "loss": 0.5751, + "num_input_tokens_seen": 5973112, + "step": 22310 + }, + { + "epoch": 5.849279161205767, + "grad_norm": 3.710340738296509, + "learning_rate": 2.196178061681809e-05, + "loss": 0.3487, + "num_input_tokens_seen": 5974264, + "step": 22315 + }, + { + "epoch": 5.850589777195282, + "grad_norm": 1.479618787765503, + "learning_rate": 2.1950428427215737e-05, + "loss": 0.565, + "num_input_tokens_seen": 5975592, + "step": 22320 + }, + { + "epoch": 5.851900393184797, + "grad_norm": 1.474673867225647, + "learning_rate": 2.1939076875883448e-05, + "loss": 0.4923, + "num_input_tokens_seen": 5976824, + "step": 22325 + }, + { + "epoch": 5.853211009174312, + "grad_norm": 15.887931823730469, + "learning_rate": 2.1927725965197074e-05, + "loss": 0.4459, + "num_input_tokens_seen": 5978296, + "step": 22330 + }, + { + "epoch": 5.854521625163827, + "grad_norm": 2.6616580486297607, + "learning_rate": 2.191637569753235e-05, + "loss": 0.521, + "num_input_tokens_seen": 5979320, + "step": 22335 + }, + { + "epoch": 5.855832241153342, + "grad_norm": 1.059397578239441, + "learning_rate": 2.1905026075264873e-05, + "loss": 0.3463, + "num_input_tokens_seen": 5980744, + "step": 22340 + }, + { + "epoch": 5.857142857142857, + "grad_norm": 5.23103141784668, + "learning_rate": 2.1893677100770084e-05, + "loss": 0.4609, + "num_input_tokens_seen": 5981768, + "step": 22345 + }, + { + "epoch": 5.858453473132372, + "grad_norm": 2.4172191619873047, + "learning_rate": 2.18823287764233e-05, + "loss": 0.3435, + "num_input_tokens_seen": 5984232, + "step": 22350 + }, + { + "epoch": 5.859764089121887, + "grad_norm": 2.609449863433838, + "learning_rate": 2.187098110459973e-05, + "loss": 0.3063, + "num_input_tokens_seen": 5985736, + "step": 22355 + }, + { + "epoch": 5.8610747051114025, + "grad_norm": 3.498201608657837, + "learning_rate": 2.1859634087674398e-05, + "loss": 0.2933, + "num_input_tokens_seen": 5986984, + "step": 22360 + }, + { + "epoch": 5.862385321100917, + "grad_norm": 6.352112770080566, + "learning_rate": 2.1848287728022233e-05, + "loss": 0.5314, + "num_input_tokens_seen": 5987944, + "step": 22365 + }, + { + "epoch": 5.863695937090433, + "grad_norm": 2.2873330116271973, + "learning_rate": 2.1836942028018e-05, + "loss": 0.5042, + "num_input_tokens_seen": 5989064, + "step": 22370 + }, + { + "epoch": 5.865006553079947, + "grad_norm": 2.0461552143096924, + "learning_rate": 2.182559699003634e-05, + "loss": 0.5474, + "num_input_tokens_seen": 5990456, + "step": 22375 + }, + { + "epoch": 5.866317169069463, + "grad_norm": 2.154531717300415, + "learning_rate": 2.1814252616451735e-05, + "loss": 0.282, + "num_input_tokens_seen": 5991992, + "step": 22380 + }, + { + "epoch": 5.8676277850589775, + "grad_norm": 1.3204232454299927, + "learning_rate": 2.1802908909638558e-05, + "loss": 0.3752, + "num_input_tokens_seen": 5993272, + "step": 22385 + }, + { + "epoch": 5.868938401048493, + "grad_norm": 2.3509137630462646, + "learning_rate": 2.1791565871971024e-05, + "loss": 0.4394, + "num_input_tokens_seen": 5994664, + "step": 22390 + }, + { + "epoch": 5.870249017038008, + "grad_norm": 3.40907883644104, + "learning_rate": 2.1780223505823212e-05, + "loss": 0.5083, + "num_input_tokens_seen": 5995928, + "step": 22395 + }, + { + "epoch": 5.871559633027523, + "grad_norm": 3.6717615127563477, + "learning_rate": 2.1768881813569046e-05, + "loss": 0.3898, + "num_input_tokens_seen": 5997128, + "step": 22400 + }, + { + "epoch": 5.872870249017038, + "grad_norm": 3.1110568046569824, + "learning_rate": 2.175754079758234e-05, + "loss": 0.7603, + "num_input_tokens_seen": 5998728, + "step": 22405 + }, + { + "epoch": 5.8741808650065535, + "grad_norm": 3.0555317401885986, + "learning_rate": 2.1746200460236736e-05, + "loss": 0.5328, + "num_input_tokens_seen": 6000328, + "step": 22410 + }, + { + "epoch": 5.875491480996068, + "grad_norm": 5.565809726715088, + "learning_rate": 2.173486080390576e-05, + "loss": 0.4102, + "num_input_tokens_seen": 6001800, + "step": 22415 + }, + { + "epoch": 5.876802096985584, + "grad_norm": 2.2263150215148926, + "learning_rate": 2.1723521830962774e-05, + "loss": 0.5565, + "num_input_tokens_seen": 6003592, + "step": 22420 + }, + { + "epoch": 5.878112712975098, + "grad_norm": 1.0364294052124023, + "learning_rate": 2.1712183543781005e-05, + "loss": 0.4567, + "num_input_tokens_seen": 6005512, + "step": 22425 + }, + { + "epoch": 5.879423328964613, + "grad_norm": 1.9651602506637573, + "learning_rate": 2.1700845944733543e-05, + "loss": 0.4322, + "num_input_tokens_seen": 6006776, + "step": 22430 + }, + { + "epoch": 5.8807339449541285, + "grad_norm": 3.298535108566284, + "learning_rate": 2.1689509036193333e-05, + "loss": 0.8505, + "num_input_tokens_seen": 6008136, + "step": 22435 + }, + { + "epoch": 5.882044560943643, + "grad_norm": 2.1954903602600098, + "learning_rate": 2.167817282053315e-05, + "loss": 0.3424, + "num_input_tokens_seen": 6009272, + "step": 22440 + }, + { + "epoch": 5.883355176933159, + "grad_norm": 3.49891996383667, + "learning_rate": 2.166683730012566e-05, + "loss": 0.6147, + "num_input_tokens_seen": 6010648, + "step": 22445 + }, + { + "epoch": 5.884665792922673, + "grad_norm": 1.4730981588363647, + "learning_rate": 2.1655502477343363e-05, + "loss": 0.4319, + "num_input_tokens_seen": 6012040, + "step": 22450 + }, + { + "epoch": 5.885976408912189, + "grad_norm": 1.307352900505066, + "learning_rate": 2.164416835455862e-05, + "loss": 0.463, + "num_input_tokens_seen": 6013368, + "step": 22455 + }, + { + "epoch": 5.8872870249017035, + "grad_norm": 2.7470548152923584, + "learning_rate": 2.1632834934143642e-05, + "loss": 0.5817, + "num_input_tokens_seen": 6014776, + "step": 22460 + }, + { + "epoch": 5.888597640891219, + "grad_norm": 2.772092342376709, + "learning_rate": 2.162150221847048e-05, + "loss": 0.4184, + "num_input_tokens_seen": 6016056, + "step": 22465 + }, + { + "epoch": 5.889908256880734, + "grad_norm": 2.222832441329956, + "learning_rate": 2.1610170209911085e-05, + "loss": 0.4058, + "num_input_tokens_seen": 6017272, + "step": 22470 + }, + { + "epoch": 5.891218872870249, + "grad_norm": 6.511477947235107, + "learning_rate": 2.15988389108372e-05, + "loss": 0.3973, + "num_input_tokens_seen": 6018424, + "step": 22475 + }, + { + "epoch": 5.892529488859764, + "grad_norm": 2.1483969688415527, + "learning_rate": 2.1587508323620452e-05, + "loss": 0.377, + "num_input_tokens_seen": 6019624, + "step": 22480 + }, + { + "epoch": 5.8938401048492794, + "grad_norm": 5.0956130027771, + "learning_rate": 2.1576178450632324e-05, + "loss": 0.3112, + "num_input_tokens_seen": 6020840, + "step": 22485 + }, + { + "epoch": 5.895150720838794, + "grad_norm": 2.9788811206817627, + "learning_rate": 2.1564849294244126e-05, + "loss": 0.4238, + "num_input_tokens_seen": 6021912, + "step": 22490 + }, + { + "epoch": 5.89646133682831, + "grad_norm": 2.2478604316711426, + "learning_rate": 2.1553520856827038e-05, + "loss": 0.47, + "num_input_tokens_seen": 6023032, + "step": 22495 + }, + { + "epoch": 5.897771952817824, + "grad_norm": 1.7379266023635864, + "learning_rate": 2.1542193140752074e-05, + "loss": 0.6604, + "num_input_tokens_seen": 6024136, + "step": 22500 + }, + { + "epoch": 5.89908256880734, + "grad_norm": 1.937932014465332, + "learning_rate": 2.1530866148390114e-05, + "loss": 0.4523, + "num_input_tokens_seen": 6026168, + "step": 22505 + }, + { + "epoch": 5.9003931847968545, + "grad_norm": 1.4928897619247437, + "learning_rate": 2.151953988211187e-05, + "loss": 0.3813, + "num_input_tokens_seen": 6027464, + "step": 22510 + }, + { + "epoch": 5.90170380078637, + "grad_norm": 1.4877759218215942, + "learning_rate": 2.150821434428793e-05, + "loss": 0.457, + "num_input_tokens_seen": 6028984, + "step": 22515 + }, + { + "epoch": 5.903014416775885, + "grad_norm": 4.612005710601807, + "learning_rate": 2.149688953728867e-05, + "loss": 0.4885, + "num_input_tokens_seen": 6030472, + "step": 22520 + }, + { + "epoch": 5.904325032765399, + "grad_norm": 3.2539641857147217, + "learning_rate": 2.1485565463484395e-05, + "loss": 0.6565, + "num_input_tokens_seen": 6031608, + "step": 22525 + }, + { + "epoch": 5.905635648754915, + "grad_norm": 2.3412251472473145, + "learning_rate": 2.1474242125245192e-05, + "loss": 0.3967, + "num_input_tokens_seen": 6033000, + "step": 22530 + }, + { + "epoch": 5.9069462647444295, + "grad_norm": 5.2409539222717285, + "learning_rate": 2.1462919524941018e-05, + "loss": 0.5687, + "num_input_tokens_seen": 6034568, + "step": 22535 + }, + { + "epoch": 5.908256880733945, + "grad_norm": 1.2713801860809326, + "learning_rate": 2.145159766494168e-05, + "loss": 0.3705, + "num_input_tokens_seen": 6035784, + "step": 22540 + }, + { + "epoch": 5.90956749672346, + "grad_norm": 6.899596214294434, + "learning_rate": 2.1440276547616815e-05, + "loss": 0.3196, + "num_input_tokens_seen": 6036744, + "step": 22545 + }, + { + "epoch": 5.910878112712975, + "grad_norm": 2.4265811443328857, + "learning_rate": 2.1428956175335922e-05, + "loss": 0.4232, + "num_input_tokens_seen": 6038216, + "step": 22550 + }, + { + "epoch": 5.91218872870249, + "grad_norm": 3.990649938583374, + "learning_rate": 2.141763655046833e-05, + "loss": 0.4796, + "num_input_tokens_seen": 6039416, + "step": 22555 + }, + { + "epoch": 5.913499344692005, + "grad_norm": 5.904153347015381, + "learning_rate": 2.140631767538322e-05, + "loss": 0.4099, + "num_input_tokens_seen": 6040520, + "step": 22560 + }, + { + "epoch": 5.91480996068152, + "grad_norm": 1.7246456146240234, + "learning_rate": 2.1394999552449608e-05, + "loss": 0.4798, + "num_input_tokens_seen": 6041544, + "step": 22565 + }, + { + "epoch": 5.916120576671036, + "grad_norm": 1.611495852470398, + "learning_rate": 2.1383682184036357e-05, + "loss": 0.4334, + "num_input_tokens_seen": 6043144, + "step": 22570 + }, + { + "epoch": 5.91743119266055, + "grad_norm": 2.7604293823242188, + "learning_rate": 2.1372365572512188e-05, + "loss": 0.3993, + "num_input_tokens_seen": 6044312, + "step": 22575 + }, + { + "epoch": 5.918741808650066, + "grad_norm": 1.6858426332473755, + "learning_rate": 2.136104972024562e-05, + "loss": 0.4252, + "num_input_tokens_seen": 6045816, + "step": 22580 + }, + { + "epoch": 5.9200524246395805, + "grad_norm": 2.3302114009857178, + "learning_rate": 2.134973462960507e-05, + "loss": 0.3944, + "num_input_tokens_seen": 6046968, + "step": 22585 + }, + { + "epoch": 5.921363040629096, + "grad_norm": 2.7397236824035645, + "learning_rate": 2.133842030295875e-05, + "loss": 0.4256, + "num_input_tokens_seen": 6048264, + "step": 22590 + }, + { + "epoch": 5.922673656618611, + "grad_norm": 4.083434581756592, + "learning_rate": 2.1327106742674735e-05, + "loss": 0.5215, + "num_input_tokens_seen": 6049432, + "step": 22595 + }, + { + "epoch": 5.923984272608125, + "grad_norm": 0.8032022714614868, + "learning_rate": 2.1315793951120934e-05, + "loss": 0.4818, + "num_input_tokens_seen": 6051640, + "step": 22600 + }, + { + "epoch": 5.925294888597641, + "grad_norm": 1.4066333770751953, + "learning_rate": 2.1304481930665095e-05, + "loss": 0.511, + "num_input_tokens_seen": 6053112, + "step": 22605 + }, + { + "epoch": 5.926605504587156, + "grad_norm": 6.0575480461120605, + "learning_rate": 2.1293170683674803e-05, + "loss": 0.3734, + "num_input_tokens_seen": 6054440, + "step": 22610 + }, + { + "epoch": 5.927916120576671, + "grad_norm": 9.632198333740234, + "learning_rate": 2.1281860212517477e-05, + "loss": 0.5071, + "num_input_tokens_seen": 6055448, + "step": 22615 + }, + { + "epoch": 5.929226736566186, + "grad_norm": 5.742950916290283, + "learning_rate": 2.1270550519560387e-05, + "loss": 0.7538, + "num_input_tokens_seen": 6056616, + "step": 22620 + }, + { + "epoch": 5.930537352555701, + "grad_norm": 4.336487770080566, + "learning_rate": 2.125924160717063e-05, + "loss": 0.3967, + "num_input_tokens_seen": 6057800, + "step": 22625 + }, + { + "epoch": 5.931847968545216, + "grad_norm": 1.5091769695281982, + "learning_rate": 2.1247933477715145e-05, + "loss": 0.3749, + "num_input_tokens_seen": 6059032, + "step": 22630 + }, + { + "epoch": 5.933158584534731, + "grad_norm": 4.01979923248291, + "learning_rate": 2.123662613356068e-05, + "loss": 1.0162, + "num_input_tokens_seen": 6059800, + "step": 22635 + }, + { + "epoch": 5.934469200524246, + "grad_norm": 1.6322834491729736, + "learning_rate": 2.1225319577073877e-05, + "loss": 0.4117, + "num_input_tokens_seen": 6061144, + "step": 22640 + }, + { + "epoch": 5.935779816513762, + "grad_norm": 2.726872444152832, + "learning_rate": 2.121401381062117e-05, + "loss": 0.402, + "num_input_tokens_seen": 6062088, + "step": 22645 + }, + { + "epoch": 5.937090432503276, + "grad_norm": 0.9079626798629761, + "learning_rate": 2.1202708836568826e-05, + "loss": 0.5692, + "num_input_tokens_seen": 6063432, + "step": 22650 + }, + { + "epoch": 5.938401048492792, + "grad_norm": 3.5128133296966553, + "learning_rate": 2.1191404657282957e-05, + "loss": 0.7389, + "num_input_tokens_seen": 6064744, + "step": 22655 + }, + { + "epoch": 5.9397116644823065, + "grad_norm": 1.4126572608947754, + "learning_rate": 2.1180101275129515e-05, + "loss": 0.4209, + "num_input_tokens_seen": 6065960, + "step": 22660 + }, + { + "epoch": 5.941022280471822, + "grad_norm": 2.8525900840759277, + "learning_rate": 2.1168798692474277e-05, + "loss": 0.4803, + "num_input_tokens_seen": 6067320, + "step": 22665 + }, + { + "epoch": 5.942332896461337, + "grad_norm": 5.9075446128845215, + "learning_rate": 2.115749691168286e-05, + "loss": 0.3998, + "num_input_tokens_seen": 6069320, + "step": 22670 + }, + { + "epoch": 5.943643512450852, + "grad_norm": 2.9839396476745605, + "learning_rate": 2.114619593512069e-05, + "loss": 0.422, + "num_input_tokens_seen": 6070360, + "step": 22675 + }, + { + "epoch": 5.944954128440367, + "grad_norm": 2.490067958831787, + "learning_rate": 2.1134895765153057e-05, + "loss": 0.5093, + "num_input_tokens_seen": 6071544, + "step": 22680 + }, + { + "epoch": 5.946264744429882, + "grad_norm": 3.1667611598968506, + "learning_rate": 2.112359640414506e-05, + "loss": 0.4499, + "num_input_tokens_seen": 6072936, + "step": 22685 + }, + { + "epoch": 5.947575360419397, + "grad_norm": 1.2846016883850098, + "learning_rate": 2.111229785446164e-05, + "loss": 0.6152, + "num_input_tokens_seen": 6074136, + "step": 22690 + }, + { + "epoch": 5.948885976408912, + "grad_norm": 4.307449817657471, + "learning_rate": 2.110100011846755e-05, + "loss": 0.5657, + "num_input_tokens_seen": 6075192, + "step": 22695 + }, + { + "epoch": 5.950196592398427, + "grad_norm": 3.3416380882263184, + "learning_rate": 2.1089703198527407e-05, + "loss": 0.4833, + "num_input_tokens_seen": 6076568, + "step": 22700 + }, + { + "epoch": 5.951507208387943, + "grad_norm": 2.4592783451080322, + "learning_rate": 2.1078407097005632e-05, + "loss": 0.5564, + "num_input_tokens_seen": 6077864, + "step": 22705 + }, + { + "epoch": 5.952817824377457, + "grad_norm": 4.833880424499512, + "learning_rate": 2.106711181626647e-05, + "loss": 0.5153, + "num_input_tokens_seen": 6079352, + "step": 22710 + }, + { + "epoch": 5.954128440366972, + "grad_norm": 2.8722939491271973, + "learning_rate": 2.105581735867401e-05, + "loss": 0.439, + "num_input_tokens_seen": 6080904, + "step": 22715 + }, + { + "epoch": 5.955439056356488, + "grad_norm": 2.3223912715911865, + "learning_rate": 2.1044523726592165e-05, + "loss": 0.5723, + "num_input_tokens_seen": 6082328, + "step": 22720 + }, + { + "epoch": 5.956749672346002, + "grad_norm": 1.6644021272659302, + "learning_rate": 2.1033230922384672e-05, + "loss": 0.3963, + "num_input_tokens_seen": 6083592, + "step": 22725 + }, + { + "epoch": 5.958060288335518, + "grad_norm": 2.089752435684204, + "learning_rate": 2.1021938948415086e-05, + "loss": 1.045, + "num_input_tokens_seen": 6084760, + "step": 22730 + }, + { + "epoch": 5.9593709043250325, + "grad_norm": 3.4004135131835938, + "learning_rate": 2.1010647807046806e-05, + "loss": 0.4654, + "num_input_tokens_seen": 6086024, + "step": 22735 + }, + { + "epoch": 5.960681520314548, + "grad_norm": 1.248490571975708, + "learning_rate": 2.099935750064305e-05, + "loss": 0.5397, + "num_input_tokens_seen": 6087208, + "step": 22740 + }, + { + "epoch": 5.961992136304063, + "grad_norm": 2.455012559890747, + "learning_rate": 2.0988068031566854e-05, + "loss": 0.3803, + "num_input_tokens_seen": 6088152, + "step": 22745 + }, + { + "epoch": 5.963302752293578, + "grad_norm": 9.703611373901367, + "learning_rate": 2.097677940218108e-05, + "loss": 0.5114, + "num_input_tokens_seen": 6089288, + "step": 22750 + }, + { + "epoch": 5.964613368283093, + "grad_norm": 1.6965969800949097, + "learning_rate": 2.0965491614848433e-05, + "loss": 0.4014, + "num_input_tokens_seen": 6090776, + "step": 22755 + }, + { + "epoch": 5.965923984272608, + "grad_norm": 2.4323959350585938, + "learning_rate": 2.095420467193142e-05, + "loss": 0.4422, + "num_input_tokens_seen": 6091912, + "step": 22760 + }, + { + "epoch": 5.967234600262123, + "grad_norm": 1.27641761302948, + "learning_rate": 2.0942918575792382e-05, + "loss": 0.5325, + "num_input_tokens_seen": 6093528, + "step": 22765 + }, + { + "epoch": 5.968545216251639, + "grad_norm": 2.9984242916107178, + "learning_rate": 2.0931633328793472e-05, + "loss": 0.4022, + "num_input_tokens_seen": 6094840, + "step": 22770 + }, + { + "epoch": 5.969855832241153, + "grad_norm": 5.9579973220825195, + "learning_rate": 2.0920348933296678e-05, + "loss": 0.3443, + "num_input_tokens_seen": 6095864, + "step": 22775 + }, + { + "epoch": 5.971166448230669, + "grad_norm": 9.722179412841797, + "learning_rate": 2.09090653916638e-05, + "loss": 1.0, + "num_input_tokens_seen": 6097080, + "step": 22780 + }, + { + "epoch": 5.972477064220183, + "grad_norm": 4.3190226554870605, + "learning_rate": 2.0897782706256474e-05, + "loss": 0.4557, + "num_input_tokens_seen": 6098104, + "step": 22785 + }, + { + "epoch": 5.973787680209698, + "grad_norm": 4.014211654663086, + "learning_rate": 2.0886500879436136e-05, + "loss": 0.4201, + "num_input_tokens_seen": 6099368, + "step": 22790 + }, + { + "epoch": 5.975098296199214, + "grad_norm": 2.6522319316864014, + "learning_rate": 2.0875219913564053e-05, + "loss": 0.3864, + "num_input_tokens_seen": 6100696, + "step": 22795 + }, + { + "epoch": 5.976408912188729, + "grad_norm": 1.2930279970169067, + "learning_rate": 2.0863939811001316e-05, + "loss": 0.4227, + "num_input_tokens_seen": 6102200, + "step": 22800 + }, + { + "epoch": 5.977719528178244, + "grad_norm": 2.191941022872925, + "learning_rate": 2.0852660574108833e-05, + "loss": 0.4232, + "num_input_tokens_seen": 6103288, + "step": 22805 + }, + { + "epoch": 5.9790301441677585, + "grad_norm": 7.622509479522705, + "learning_rate": 2.084138220524732e-05, + "loss": 0.4136, + "num_input_tokens_seen": 6104568, + "step": 22810 + }, + { + "epoch": 5.980340760157274, + "grad_norm": 3.4985458850860596, + "learning_rate": 2.0830104706777325e-05, + "loss": 0.4171, + "num_input_tokens_seen": 6105720, + "step": 22815 + }, + { + "epoch": 5.981651376146789, + "grad_norm": 1.950383186340332, + "learning_rate": 2.081882808105922e-05, + "loss": 0.4158, + "num_input_tokens_seen": 6106904, + "step": 22820 + }, + { + "epoch": 5.982961992136304, + "grad_norm": 2.4268600940704346, + "learning_rate": 2.0807552330453166e-05, + "loss": 0.3382, + "num_input_tokens_seen": 6108584, + "step": 22825 + }, + { + "epoch": 5.984272608125819, + "grad_norm": 2.7207765579223633, + "learning_rate": 2.079627745731917e-05, + "loss": 0.4874, + "num_input_tokens_seen": 6109784, + "step": 22830 + }, + { + "epoch": 5.985583224115334, + "grad_norm": 2.2109906673431396, + "learning_rate": 2.0785003464017032e-05, + "loss": 0.4645, + "num_input_tokens_seen": 6110888, + "step": 22835 + }, + { + "epoch": 5.986893840104849, + "grad_norm": 2.394625425338745, + "learning_rate": 2.0773730352906394e-05, + "loss": 0.5456, + "num_input_tokens_seen": 6112328, + "step": 22840 + }, + { + "epoch": 5.988204456094365, + "grad_norm": 3.640747308731079, + "learning_rate": 2.076245812634669e-05, + "loss": 0.4201, + "num_input_tokens_seen": 6113816, + "step": 22845 + }, + { + "epoch": 5.989515072083879, + "grad_norm": 1.184476613998413, + "learning_rate": 2.0751186786697175e-05, + "loss": 0.5059, + "num_input_tokens_seen": 6115944, + "step": 22850 + }, + { + "epoch": 5.990825688073395, + "grad_norm": 1.1286286115646362, + "learning_rate": 2.0739916336316924e-05, + "loss": 0.3788, + "num_input_tokens_seen": 6117128, + "step": 22855 + }, + { + "epoch": 5.992136304062909, + "grad_norm": 7.381196975708008, + "learning_rate": 2.072864677756482e-05, + "loss": 0.5681, + "num_input_tokens_seen": 6118344, + "step": 22860 + }, + { + "epoch": 5.993446920052425, + "grad_norm": 1.9844664335250854, + "learning_rate": 2.0717378112799563e-05, + "loss": 0.52, + "num_input_tokens_seen": 6120008, + "step": 22865 + }, + { + "epoch": 5.99475753604194, + "grad_norm": 1.748782992362976, + "learning_rate": 2.070611034437967e-05, + "loss": 0.5385, + "num_input_tokens_seen": 6121160, + "step": 22870 + }, + { + "epoch": 5.996068152031455, + "grad_norm": 1.328748106956482, + "learning_rate": 2.0694843474663465e-05, + "loss": 0.4326, + "num_input_tokens_seen": 6122536, + "step": 22875 + }, + { + "epoch": 5.99737876802097, + "grad_norm": 2.8275997638702393, + "learning_rate": 2.0683577506009087e-05, + "loss": 0.5808, + "num_input_tokens_seen": 6123896, + "step": 22880 + }, + { + "epoch": 5.9986893840104845, + "grad_norm": 3.2569425106048584, + "learning_rate": 2.0672312440774473e-05, + "loss": 0.7497, + "num_input_tokens_seen": 6124904, + "step": 22885 + }, + { + "epoch": 6.0, + "grad_norm": 2.1076738834381104, + "learning_rate": 2.0661048281317388e-05, + "loss": 0.4635, + "num_input_tokens_seen": 6126192, + "step": 22890 + }, + { + "epoch": 6.001310615989515, + "grad_norm": 2.3881216049194336, + "learning_rate": 2.06497850299954e-05, + "loss": 0.5529, + "num_input_tokens_seen": 6127344, + "step": 22895 + }, + { + "epoch": 6.001572739187418, + "eval_loss": 0.46811333298683167, + "eval_runtime": 18.2014, + "eval_samples_per_second": 46.59, + "eval_steps_per_second": 23.295, + "num_input_tokens_seen": 6127632, + "step": 22896 + }, + { + "epoch": 6.00262123197903, + "grad_norm": 1.6255927085876465, + "learning_rate": 2.063852268916589e-05, + "loss": 0.432, + "num_input_tokens_seen": 6128832, + "step": 22900 + }, + { + "epoch": 6.003931847968545, + "grad_norm": 2.62389874458313, + "learning_rate": 2.062726126118605e-05, + "loss": 0.4888, + "num_input_tokens_seen": 6130208, + "step": 22905 + }, + { + "epoch": 6.00524246395806, + "grad_norm": 1.899611473083496, + "learning_rate": 2.0616000748412864e-05, + "loss": 0.3422, + "num_input_tokens_seen": 6131632, + "step": 22910 + }, + { + "epoch": 6.006553079947575, + "grad_norm": 1.6332066059112549, + "learning_rate": 2.060474115320315e-05, + "loss": 0.3453, + "num_input_tokens_seen": 6133056, + "step": 22915 + }, + { + "epoch": 6.007863695937091, + "grad_norm": 1.4072328805923462, + "learning_rate": 2.059348247791352e-05, + "loss": 0.3303, + "num_input_tokens_seen": 6134544, + "step": 22920 + }, + { + "epoch": 6.009174311926605, + "grad_norm": 7.626305103302002, + "learning_rate": 2.058222472490038e-05, + "loss": 0.5149, + "num_input_tokens_seen": 6135680, + "step": 22925 + }, + { + "epoch": 6.010484927916121, + "grad_norm": 4.056648254394531, + "learning_rate": 2.0570967896519983e-05, + "loss": 0.4289, + "num_input_tokens_seen": 6136704, + "step": 22930 + }, + { + "epoch": 6.011795543905635, + "grad_norm": 2.494072914123535, + "learning_rate": 2.0559711995128353e-05, + "loss": 0.3912, + "num_input_tokens_seen": 6138400, + "step": 22935 + }, + { + "epoch": 6.013106159895151, + "grad_norm": 3.760221004486084, + "learning_rate": 2.0548457023081333e-05, + "loss": 0.6426, + "num_input_tokens_seen": 6139632, + "step": 22940 + }, + { + "epoch": 6.014416775884666, + "grad_norm": 3.729058027267456, + "learning_rate": 2.0537202982734564e-05, + "loss": 0.3684, + "num_input_tokens_seen": 6140880, + "step": 22945 + }, + { + "epoch": 6.015727391874181, + "grad_norm": 1.6149930953979492, + "learning_rate": 2.0525949876443497e-05, + "loss": 0.575, + "num_input_tokens_seen": 6142672, + "step": 22950 + }, + { + "epoch": 6.017038007863696, + "grad_norm": 5.560789585113525, + "learning_rate": 2.051469770656339e-05, + "loss": 0.4542, + "num_input_tokens_seen": 6143696, + "step": 22955 + }, + { + "epoch": 6.018348623853211, + "grad_norm": 2.1142165660858154, + "learning_rate": 2.050344647544931e-05, + "loss": 0.3741, + "num_input_tokens_seen": 6145296, + "step": 22960 + }, + { + "epoch": 6.019659239842726, + "grad_norm": 3.4637014865875244, + "learning_rate": 2.0492196185456108e-05, + "loss": 0.2735, + "num_input_tokens_seen": 6146240, + "step": 22965 + }, + { + "epoch": 6.0209698558322415, + "grad_norm": 7.239351749420166, + "learning_rate": 2.0480946838938455e-05, + "loss": 0.4618, + "num_input_tokens_seen": 6147600, + "step": 22970 + }, + { + "epoch": 6.022280471821756, + "grad_norm": 2.99902081489563, + "learning_rate": 2.046969843825082e-05, + "loss": 0.5858, + "num_input_tokens_seen": 6148736, + "step": 22975 + }, + { + "epoch": 6.023591087811272, + "grad_norm": 3.062811851501465, + "learning_rate": 2.0458450985747468e-05, + "loss": 0.4437, + "num_input_tokens_seen": 6149888, + "step": 22980 + }, + { + "epoch": 6.024901703800786, + "grad_norm": 2.9727492332458496, + "learning_rate": 2.0447204483782482e-05, + "loss": 0.446, + "num_input_tokens_seen": 6151168, + "step": 22985 + }, + { + "epoch": 6.026212319790301, + "grad_norm": 2.5744636058807373, + "learning_rate": 2.0435958934709738e-05, + "loss": 0.4895, + "num_input_tokens_seen": 6152176, + "step": 22990 + }, + { + "epoch": 6.027522935779817, + "grad_norm": 11.672140121459961, + "learning_rate": 2.0424714340882893e-05, + "loss": 0.4165, + "num_input_tokens_seen": 6153376, + "step": 22995 + }, + { + "epoch": 6.028833551769331, + "grad_norm": 1.986388921737671, + "learning_rate": 2.0413470704655445e-05, + "loss": 0.4979, + "num_input_tokens_seen": 6154864, + "step": 23000 + }, + { + "epoch": 6.030144167758847, + "grad_norm": 5.874687194824219, + "learning_rate": 2.0402228028380644e-05, + "loss": 0.4845, + "num_input_tokens_seen": 6156016, + "step": 23005 + }, + { + "epoch": 6.031454783748361, + "grad_norm": 3.377333402633667, + "learning_rate": 2.0390986314411573e-05, + "loss": 0.4992, + "num_input_tokens_seen": 6157296, + "step": 23010 + }, + { + "epoch": 6.032765399737877, + "grad_norm": 5.2204508781433105, + "learning_rate": 2.03797455651011e-05, + "loss": 0.684, + "num_input_tokens_seen": 6158656, + "step": 23015 + }, + { + "epoch": 6.034076015727392, + "grad_norm": 2.717100143432617, + "learning_rate": 2.0368505782801904e-05, + "loss": 0.4616, + "num_input_tokens_seen": 6160016, + "step": 23020 + }, + { + "epoch": 6.035386631716907, + "grad_norm": 1.7439979314804077, + "learning_rate": 2.0357266969866443e-05, + "loss": 0.4532, + "num_input_tokens_seen": 6161280, + "step": 23025 + }, + { + "epoch": 6.036697247706422, + "grad_norm": 2.4337661266326904, + "learning_rate": 2.034602912864698e-05, + "loss": 0.5425, + "num_input_tokens_seen": 6162432, + "step": 23030 + }, + { + "epoch": 6.038007863695937, + "grad_norm": 3.5164501667022705, + "learning_rate": 2.0334792261495576e-05, + "loss": 0.5747, + "num_input_tokens_seen": 6163808, + "step": 23035 + }, + { + "epoch": 6.039318479685452, + "grad_norm": 4.325860977172852, + "learning_rate": 2.032355637076409e-05, + "loss": 0.5945, + "num_input_tokens_seen": 6165088, + "step": 23040 + }, + { + "epoch": 6.0406290956749675, + "grad_norm": 2.880913734436035, + "learning_rate": 2.031232145880418e-05, + "loss": 0.7141, + "num_input_tokens_seen": 6166096, + "step": 23045 + }, + { + "epoch": 6.041939711664482, + "grad_norm": 2.046290397644043, + "learning_rate": 2.030108752796729e-05, + "loss": 0.3002, + "num_input_tokens_seen": 6167264, + "step": 23050 + }, + { + "epoch": 6.043250327653998, + "grad_norm": 2.6237545013427734, + "learning_rate": 2.0289854580604667e-05, + "loss": 0.4573, + "num_input_tokens_seen": 6168624, + "step": 23055 + }, + { + "epoch": 6.044560943643512, + "grad_norm": 4.410861492156982, + "learning_rate": 2.0278622619067334e-05, + "loss": 0.4291, + "num_input_tokens_seen": 6169760, + "step": 23060 + }, + { + "epoch": 6.045871559633028, + "grad_norm": 2.354405164718628, + "learning_rate": 2.026739164570613e-05, + "loss": 0.4709, + "num_input_tokens_seen": 6170880, + "step": 23065 + }, + { + "epoch": 6.047182175622543, + "grad_norm": 1.0844545364379883, + "learning_rate": 2.0256161662871675e-05, + "loss": 0.4056, + "num_input_tokens_seen": 6172720, + "step": 23070 + }, + { + "epoch": 6.048492791612058, + "grad_norm": 2.9591686725616455, + "learning_rate": 2.0244932672914392e-05, + "loss": 0.5202, + "num_input_tokens_seen": 6174576, + "step": 23075 + }, + { + "epoch": 6.049803407601573, + "grad_norm": 3.4375667572021484, + "learning_rate": 2.0233704678184485e-05, + "loss": 0.443, + "num_input_tokens_seen": 6175760, + "step": 23080 + }, + { + "epoch": 6.051114023591087, + "grad_norm": 4.353974342346191, + "learning_rate": 2.0222477681031947e-05, + "loss": 0.3149, + "num_input_tokens_seen": 6178592, + "step": 23085 + }, + { + "epoch": 6.052424639580603, + "grad_norm": 4.039008140563965, + "learning_rate": 2.021125168380658e-05, + "loss": 0.4105, + "num_input_tokens_seen": 6180160, + "step": 23090 + }, + { + "epoch": 6.053735255570118, + "grad_norm": 1.1338244676589966, + "learning_rate": 2.0200026688857955e-05, + "loss": 0.5695, + "num_input_tokens_seen": 6181520, + "step": 23095 + }, + { + "epoch": 6.055045871559633, + "grad_norm": 3.5665552616119385, + "learning_rate": 2.0188802698535455e-05, + "loss": 0.369, + "num_input_tokens_seen": 6182752, + "step": 23100 + }, + { + "epoch": 6.056356487549148, + "grad_norm": 2.080087184906006, + "learning_rate": 2.0177579715188234e-05, + "loss": 0.3959, + "num_input_tokens_seen": 6184400, + "step": 23105 + }, + { + "epoch": 6.057667103538663, + "grad_norm": 2.6996867656707764, + "learning_rate": 2.016635774116525e-05, + "loss": 0.7813, + "num_input_tokens_seen": 6185968, + "step": 23110 + }, + { + "epoch": 6.058977719528178, + "grad_norm": 5.407234191894531, + "learning_rate": 2.0155136778815246e-05, + "loss": 0.3679, + "num_input_tokens_seen": 6187104, + "step": 23115 + }, + { + "epoch": 6.0602883355176935, + "grad_norm": 1.2466716766357422, + "learning_rate": 2.0143916830486734e-05, + "loss": 0.3396, + "num_input_tokens_seen": 6189024, + "step": 23120 + }, + { + "epoch": 6.061598951507208, + "grad_norm": 3.579486131668091, + "learning_rate": 2.0132697898528037e-05, + "loss": 0.4559, + "num_input_tokens_seen": 6190256, + "step": 23125 + }, + { + "epoch": 6.062909567496724, + "grad_norm": 2.2311770915985107, + "learning_rate": 2.0121479985287264e-05, + "loss": 0.4223, + "num_input_tokens_seen": 6191808, + "step": 23130 + }, + { + "epoch": 6.064220183486238, + "grad_norm": 2.9449515342712402, + "learning_rate": 2.01102630931123e-05, + "loss": 0.5991, + "num_input_tokens_seen": 6193136, + "step": 23135 + }, + { + "epoch": 6.065530799475754, + "grad_norm": 4.066915512084961, + "learning_rate": 2.0099047224350825e-05, + "loss": 0.4428, + "num_input_tokens_seen": 6194160, + "step": 23140 + }, + { + "epoch": 6.066841415465269, + "grad_norm": 1.5539965629577637, + "learning_rate": 2.0087832381350295e-05, + "loss": 0.3185, + "num_input_tokens_seen": 6195648, + "step": 23145 + }, + { + "epoch": 6.068152031454784, + "grad_norm": 5.51610803604126, + "learning_rate": 2.007661856645796e-05, + "loss": 0.4372, + "num_input_tokens_seen": 6196896, + "step": 23150 + }, + { + "epoch": 6.069462647444299, + "grad_norm": 4.308101177215576, + "learning_rate": 2.0065405782020844e-05, + "loss": 0.3527, + "num_input_tokens_seen": 6197808, + "step": 23155 + }, + { + "epoch": 6.070773263433814, + "grad_norm": 3.1671688556671143, + "learning_rate": 2.0054194030385774e-05, + "loss": 0.8095, + "num_input_tokens_seen": 6199168, + "step": 23160 + }, + { + "epoch": 6.072083879423329, + "grad_norm": 2.676234006881714, + "learning_rate": 2.0042983313899356e-05, + "loss": 0.4706, + "num_input_tokens_seen": 6200768, + "step": 23165 + }, + { + "epoch": 6.073394495412844, + "grad_norm": 2.0591073036193848, + "learning_rate": 2.0031773634907962e-05, + "loss": 0.6156, + "num_input_tokens_seen": 6202080, + "step": 23170 + }, + { + "epoch": 6.074705111402359, + "grad_norm": 5.724255084991455, + "learning_rate": 2.0020564995757765e-05, + "loss": 0.4639, + "num_input_tokens_seen": 6203504, + "step": 23175 + }, + { + "epoch": 6.076015727391874, + "grad_norm": 4.540925979614258, + "learning_rate": 2.0009357398794707e-05, + "loss": 0.5582, + "num_input_tokens_seen": 6204480, + "step": 23180 + }, + { + "epoch": 6.077326343381389, + "grad_norm": 1.182540774345398, + "learning_rate": 1.999815084636452e-05, + "loss": 0.6089, + "num_input_tokens_seen": 6206544, + "step": 23185 + }, + { + "epoch": 6.078636959370904, + "grad_norm": 1.357940435409546, + "learning_rate": 1.9986945340812722e-05, + "loss": 0.5084, + "num_input_tokens_seen": 6207984, + "step": 23190 + }, + { + "epoch": 6.0799475753604195, + "grad_norm": 6.258735656738281, + "learning_rate": 1.9975740884484606e-05, + "loss": 0.5519, + "num_input_tokens_seen": 6209184, + "step": 23195 + }, + { + "epoch": 6.081258191349934, + "grad_norm": 6.18679141998291, + "learning_rate": 1.9964537479725234e-05, + "loss": 0.4639, + "num_input_tokens_seen": 6210576, + "step": 23200 + }, + { + "epoch": 6.08256880733945, + "grad_norm": 2.4107065200805664, + "learning_rate": 1.9953335128879468e-05, + "loss": 0.5432, + "num_input_tokens_seen": 6212144, + "step": 23205 + }, + { + "epoch": 6.083879423328964, + "grad_norm": 5.836817264556885, + "learning_rate": 1.9942133834291936e-05, + "loss": 0.4976, + "num_input_tokens_seen": 6213552, + "step": 23210 + }, + { + "epoch": 6.08519003931848, + "grad_norm": 2.1353814601898193, + "learning_rate": 1.9930933598307058e-05, + "loss": 0.4811, + "num_input_tokens_seen": 6214688, + "step": 23215 + }, + { + "epoch": 6.086500655307995, + "grad_norm": 5.814011096954346, + "learning_rate": 1.991973442326902e-05, + "loss": 0.6874, + "num_input_tokens_seen": 6216032, + "step": 23220 + }, + { + "epoch": 6.08781127129751, + "grad_norm": 9.680994987487793, + "learning_rate": 1.990853631152179e-05, + "loss": 0.5737, + "num_input_tokens_seen": 6217136, + "step": 23225 + }, + { + "epoch": 6.089121887287025, + "grad_norm": 2.036886215209961, + "learning_rate": 1.989733926540911e-05, + "loss": 0.3068, + "num_input_tokens_seen": 6218016, + "step": 23230 + }, + { + "epoch": 6.09043250327654, + "grad_norm": 1.428392767906189, + "learning_rate": 1.988614328727451e-05, + "loss": 0.9453, + "num_input_tokens_seen": 6219184, + "step": 23235 + }, + { + "epoch": 6.091743119266055, + "grad_norm": 1.5728732347488403, + "learning_rate": 1.9874948379461277e-05, + "loss": 0.5097, + "num_input_tokens_seen": 6221360, + "step": 23240 + }, + { + "epoch": 6.0930537352555705, + "grad_norm": 2.6253879070281982, + "learning_rate": 1.9863754544312495e-05, + "loss": 0.4609, + "num_input_tokens_seen": 6222576, + "step": 23245 + }, + { + "epoch": 6.094364351245085, + "grad_norm": 1.6395275592803955, + "learning_rate": 1.9852561784171014e-05, + "loss": 0.4022, + "num_input_tokens_seen": 6223488, + "step": 23250 + }, + { + "epoch": 6.095674967234601, + "grad_norm": 3.4120500087738037, + "learning_rate": 1.9841370101379463e-05, + "loss": 0.4141, + "num_input_tokens_seen": 6224608, + "step": 23255 + }, + { + "epoch": 6.096985583224115, + "grad_norm": 3.487813711166382, + "learning_rate": 1.983017949828023e-05, + "loss": 0.4326, + "num_input_tokens_seen": 6226064, + "step": 23260 + }, + { + "epoch": 6.09829619921363, + "grad_norm": 1.6539324522018433, + "learning_rate": 1.981898997721548e-05, + "loss": 0.3765, + "num_input_tokens_seen": 6227168, + "step": 23265 + }, + { + "epoch": 6.0996068152031455, + "grad_norm": 4.410440921783447, + "learning_rate": 1.98078015405272e-05, + "loss": 0.5437, + "num_input_tokens_seen": 6228240, + "step": 23270 + }, + { + "epoch": 6.10091743119266, + "grad_norm": 2.0960593223571777, + "learning_rate": 1.9796614190557074e-05, + "loss": 0.7972, + "num_input_tokens_seen": 6229680, + "step": 23275 + }, + { + "epoch": 6.102228047182176, + "grad_norm": 7.559669494628906, + "learning_rate": 1.978542792964661e-05, + "loss": 0.6909, + "num_input_tokens_seen": 6230768, + "step": 23280 + }, + { + "epoch": 6.10353866317169, + "grad_norm": 2.3530383110046387, + "learning_rate": 1.977424276013707e-05, + "loss": 0.4821, + "num_input_tokens_seen": 6232448, + "step": 23285 + }, + { + "epoch": 6.104849279161206, + "grad_norm": 3.411506414413452, + "learning_rate": 1.9763058684369495e-05, + "loss": 0.3249, + "num_input_tokens_seen": 6234096, + "step": 23290 + }, + { + "epoch": 6.1061598951507206, + "grad_norm": 2.506471633911133, + "learning_rate": 1.9751875704684684e-05, + "loss": 0.3607, + "num_input_tokens_seen": 6235056, + "step": 23295 + }, + { + "epoch": 6.107470511140236, + "grad_norm": 1.6199965476989746, + "learning_rate": 1.974069382342322e-05, + "loss": 0.3246, + "num_input_tokens_seen": 6236512, + "step": 23300 + }, + { + "epoch": 6.108781127129751, + "grad_norm": 1.023024082183838, + "learning_rate": 1.9729513042925455e-05, + "loss": 0.5823, + "num_input_tokens_seen": 6239872, + "step": 23305 + }, + { + "epoch": 6.110091743119266, + "grad_norm": 2.661891222000122, + "learning_rate": 1.97183333655315e-05, + "loss": 0.3361, + "num_input_tokens_seen": 6241136, + "step": 23310 + }, + { + "epoch": 6.111402359108781, + "grad_norm": 19.159658432006836, + "learning_rate": 1.970715479358125e-05, + "loss": 0.478, + "num_input_tokens_seen": 6242240, + "step": 23315 + }, + { + "epoch": 6.1127129750982965, + "grad_norm": 4.7092108726501465, + "learning_rate": 1.9695977329414352e-05, + "loss": 0.5003, + "num_input_tokens_seen": 6243648, + "step": 23320 + }, + { + "epoch": 6.114023591087811, + "grad_norm": 1.9452379941940308, + "learning_rate": 1.9684800975370224e-05, + "loss": 0.5393, + "num_input_tokens_seen": 6245216, + "step": 23325 + }, + { + "epoch": 6.115334207077327, + "grad_norm": 1.0229536294937134, + "learning_rate": 1.9673625733788084e-05, + "loss": 0.3294, + "num_input_tokens_seen": 6247152, + "step": 23330 + }, + { + "epoch": 6.116644823066841, + "grad_norm": 1.50642991065979, + "learning_rate": 1.966245160700687e-05, + "loss": 0.3817, + "num_input_tokens_seen": 6248608, + "step": 23335 + }, + { + "epoch": 6.117955439056357, + "grad_norm": 2.6571414470672607, + "learning_rate": 1.965127859736531e-05, + "loss": 0.5071, + "num_input_tokens_seen": 6249808, + "step": 23340 + }, + { + "epoch": 6.1192660550458715, + "grad_norm": 3.131559371948242, + "learning_rate": 1.96401067072019e-05, + "loss": 0.5269, + "num_input_tokens_seen": 6250976, + "step": 23345 + }, + { + "epoch": 6.120576671035387, + "grad_norm": 2.328425645828247, + "learning_rate": 1.9628935938854897e-05, + "loss": 0.5745, + "num_input_tokens_seen": 6252544, + "step": 23350 + }, + { + "epoch": 6.121887287024902, + "grad_norm": 5.620305061340332, + "learning_rate": 1.961776629466232e-05, + "loss": 0.3689, + "num_input_tokens_seen": 6253872, + "step": 23355 + }, + { + "epoch": 6.123197903014416, + "grad_norm": 8.457983016967773, + "learning_rate": 1.9606597776961955e-05, + "loss": 0.438, + "num_input_tokens_seen": 6255088, + "step": 23360 + }, + { + "epoch": 6.124508519003932, + "grad_norm": 1.7571306228637695, + "learning_rate": 1.9595430388091358e-05, + "loss": 0.4582, + "num_input_tokens_seen": 6256080, + "step": 23365 + }, + { + "epoch": 6.1258191349934465, + "grad_norm": 1.547071933746338, + "learning_rate": 1.958426413038784e-05, + "loss": 0.3663, + "num_input_tokens_seen": 6257328, + "step": 23370 + }, + { + "epoch": 6.127129750982962, + "grad_norm": 1.5022943019866943, + "learning_rate": 1.957309900618849e-05, + "loss": 0.4201, + "num_input_tokens_seen": 6259296, + "step": 23375 + }, + { + "epoch": 6.128440366972477, + "grad_norm": 5.847530841827393, + "learning_rate": 1.956193501783012e-05, + "loss": 0.5591, + "num_input_tokens_seen": 6260608, + "step": 23380 + }, + { + "epoch": 6.129750982961992, + "grad_norm": 7.9949421882629395, + "learning_rate": 1.9550772167649378e-05, + "loss": 0.579, + "num_input_tokens_seen": 6261600, + "step": 23385 + }, + { + "epoch": 6.131061598951507, + "grad_norm": 2.276953935623169, + "learning_rate": 1.9539610457982596e-05, + "loss": 0.4108, + "num_input_tokens_seen": 6262944, + "step": 23390 + }, + { + "epoch": 6.1323722149410225, + "grad_norm": 6.415992259979248, + "learning_rate": 1.9528449891165913e-05, + "loss": 0.5754, + "num_input_tokens_seen": 6264000, + "step": 23395 + }, + { + "epoch": 6.133682830930537, + "grad_norm": 8.125576972961426, + "learning_rate": 1.951729046953521e-05, + "loss": 0.4878, + "num_input_tokens_seen": 6265552, + "step": 23400 + }, + { + "epoch": 6.134993446920053, + "grad_norm": 5.73450231552124, + "learning_rate": 1.9506132195426138e-05, + "loss": 0.7126, + "num_input_tokens_seen": 6266672, + "step": 23405 + }, + { + "epoch": 6.136304062909567, + "grad_norm": 4.159032821655273, + "learning_rate": 1.9494975071174112e-05, + "loss": 0.5458, + "num_input_tokens_seen": 6269584, + "step": 23410 + }, + { + "epoch": 6.137614678899083, + "grad_norm": 2.564826250076294, + "learning_rate": 1.9483819099114286e-05, + "loss": 0.397, + "num_input_tokens_seen": 6271088, + "step": 23415 + }, + { + "epoch": 6.1389252948885975, + "grad_norm": 3.113300085067749, + "learning_rate": 1.9472664281581594e-05, + "loss": 0.4968, + "num_input_tokens_seen": 6272256, + "step": 23420 + }, + { + "epoch": 6.140235910878113, + "grad_norm": 1.360522747039795, + "learning_rate": 1.9461510620910713e-05, + "loss": 0.4178, + "num_input_tokens_seen": 6273536, + "step": 23425 + }, + { + "epoch": 6.141546526867628, + "grad_norm": 2.9210426807403564, + "learning_rate": 1.94503581194361e-05, + "loss": 0.5166, + "num_input_tokens_seen": 6274544, + "step": 23430 + }, + { + "epoch": 6.142857142857143, + "grad_norm": 3.6499767303466797, + "learning_rate": 1.9439206779491934e-05, + "loss": 0.4846, + "num_input_tokens_seen": 6275424, + "step": 23435 + }, + { + "epoch": 6.144167758846658, + "grad_norm": 6.2548508644104, + "learning_rate": 1.942805660341217e-05, + "loss": 0.7494, + "num_input_tokens_seen": 6276768, + "step": 23440 + }, + { + "epoch": 6.145478374836173, + "grad_norm": 1.7231138944625854, + "learning_rate": 1.9416907593530548e-05, + "loss": 0.3806, + "num_input_tokens_seen": 6278096, + "step": 23445 + }, + { + "epoch": 6.146788990825688, + "grad_norm": 2.186605215072632, + "learning_rate": 1.9405759752180515e-05, + "loss": 0.5746, + "num_input_tokens_seen": 6279216, + "step": 23450 + }, + { + "epoch": 6.148099606815203, + "grad_norm": 2.904074192047119, + "learning_rate": 1.9394613081695302e-05, + "loss": 0.4225, + "num_input_tokens_seen": 6280544, + "step": 23455 + }, + { + "epoch": 6.149410222804718, + "grad_norm": 3.7180984020233154, + "learning_rate": 1.9383467584407883e-05, + "loss": 0.3746, + "num_input_tokens_seen": 6281584, + "step": 23460 + }, + { + "epoch": 6.150720838794233, + "grad_norm": 2.895826816558838, + "learning_rate": 1.9372323262650992e-05, + "loss": 0.42, + "num_input_tokens_seen": 6283312, + "step": 23465 + }, + { + "epoch": 6.1520314547837485, + "grad_norm": 6.988287448883057, + "learning_rate": 1.936118011875713e-05, + "loss": 0.3156, + "num_input_tokens_seen": 6286240, + "step": 23470 + }, + { + "epoch": 6.153342070773263, + "grad_norm": 1.9371130466461182, + "learning_rate": 1.9350038155058515e-05, + "loss": 0.5661, + "num_input_tokens_seen": 6287584, + "step": 23475 + }, + { + "epoch": 6.154652686762779, + "grad_norm": 2.7011842727661133, + "learning_rate": 1.9338897373887154e-05, + "loss": 0.4501, + "num_input_tokens_seen": 6289280, + "step": 23480 + }, + { + "epoch": 6.155963302752293, + "grad_norm": 1.8984954357147217, + "learning_rate": 1.932775777757479e-05, + "loss": 0.4726, + "num_input_tokens_seen": 6290848, + "step": 23485 + }, + { + "epoch": 6.157273918741809, + "grad_norm": 3.1297707557678223, + "learning_rate": 1.9316619368452928e-05, + "loss": 0.5588, + "num_input_tokens_seen": 6291712, + "step": 23490 + }, + { + "epoch": 6.1585845347313235, + "grad_norm": 2.821823835372925, + "learning_rate": 1.9305482148852794e-05, + "loss": 0.561, + "num_input_tokens_seen": 6292992, + "step": 23495 + }, + { + "epoch": 6.159895150720839, + "grad_norm": 1.919467568397522, + "learning_rate": 1.929434612110542e-05, + "loss": 0.3628, + "num_input_tokens_seen": 6294304, + "step": 23500 + }, + { + "epoch": 6.161205766710354, + "grad_norm": 2.1890010833740234, + "learning_rate": 1.9283211287541548e-05, + "loss": 0.3485, + "num_input_tokens_seen": 6295552, + "step": 23505 + }, + { + "epoch": 6.162516382699869, + "grad_norm": 2.7197303771972656, + "learning_rate": 1.9272077650491673e-05, + "loss": 0.5248, + "num_input_tokens_seen": 6296528, + "step": 23510 + }, + { + "epoch": 6.163826998689384, + "grad_norm": 3.1436455249786377, + "learning_rate": 1.9260945212286047e-05, + "loss": 0.7618, + "num_input_tokens_seen": 6297936, + "step": 23515 + }, + { + "epoch": 6.165137614678899, + "grad_norm": 2.985042095184326, + "learning_rate": 1.9249813975254673e-05, + "loss": 0.4544, + "num_input_tokens_seen": 6299104, + "step": 23520 + }, + { + "epoch": 6.166448230668414, + "grad_norm": 2.5996384620666504, + "learning_rate": 1.92386839417273e-05, + "loss": 0.4451, + "num_input_tokens_seen": 6300224, + "step": 23525 + }, + { + "epoch": 6.16775884665793, + "grad_norm": 1.5631873607635498, + "learning_rate": 1.9227555114033422e-05, + "loss": 0.4784, + "num_input_tokens_seen": 6301488, + "step": 23530 + }, + { + "epoch": 6.169069462647444, + "grad_norm": 3.136214017868042, + "learning_rate": 1.921642749450228e-05, + "loss": 0.534, + "num_input_tokens_seen": 6302624, + "step": 23535 + }, + { + "epoch": 6.17038007863696, + "grad_norm": 2.853914737701416, + "learning_rate": 1.9205301085462877e-05, + "loss": 0.3836, + "num_input_tokens_seen": 6304032, + "step": 23540 + }, + { + "epoch": 6.1716906946264745, + "grad_norm": 3.2419586181640625, + "learning_rate": 1.919417588924394e-05, + "loss": 0.4506, + "num_input_tokens_seen": 6305312, + "step": 23545 + }, + { + "epoch": 6.173001310615989, + "grad_norm": 2.537984848022461, + "learning_rate": 1.9183051908173963e-05, + "loss": 0.419, + "num_input_tokens_seen": 6306384, + "step": 23550 + }, + { + "epoch": 6.174311926605505, + "grad_norm": 3.4040913581848145, + "learning_rate": 1.9171929144581156e-05, + "loss": 0.6048, + "num_input_tokens_seen": 6307824, + "step": 23555 + }, + { + "epoch": 6.175622542595019, + "grad_norm": 3.0533089637756348, + "learning_rate": 1.9160807600793517e-05, + "loss": 0.6008, + "num_input_tokens_seen": 6308928, + "step": 23560 + }, + { + "epoch": 6.176933158584535, + "grad_norm": 3.6238412857055664, + "learning_rate": 1.9149687279138768e-05, + "loss": 0.3635, + "num_input_tokens_seen": 6310032, + "step": 23565 + }, + { + "epoch": 6.1782437745740495, + "grad_norm": 7.016526222229004, + "learning_rate": 1.9138568181944357e-05, + "loss": 0.4834, + "num_input_tokens_seen": 6311280, + "step": 23570 + }, + { + "epoch": 6.179554390563565, + "grad_norm": 2.6385250091552734, + "learning_rate": 1.9127450311537494e-05, + "loss": 0.5426, + "num_input_tokens_seen": 6312528, + "step": 23575 + }, + { + "epoch": 6.18086500655308, + "grad_norm": 3.177093505859375, + "learning_rate": 1.9116333670245135e-05, + "loss": 0.5109, + "num_input_tokens_seen": 6313968, + "step": 23580 + }, + { + "epoch": 6.182175622542595, + "grad_norm": 1.8589425086975098, + "learning_rate": 1.9105218260393975e-05, + "loss": 0.4665, + "num_input_tokens_seen": 6315264, + "step": 23585 + }, + { + "epoch": 6.18348623853211, + "grad_norm": 2.0373926162719727, + "learning_rate": 1.9094104084310446e-05, + "loss": 0.4865, + "num_input_tokens_seen": 6316848, + "step": 23590 + }, + { + "epoch": 6.184796854521625, + "grad_norm": 1.998284935951233, + "learning_rate": 1.908299114432073e-05, + "loss": 0.7518, + "num_input_tokens_seen": 6317984, + "step": 23595 + }, + { + "epoch": 6.18610747051114, + "grad_norm": 3.8692495822906494, + "learning_rate": 1.907187944275074e-05, + "loss": 0.5042, + "num_input_tokens_seen": 6319632, + "step": 23600 + }, + { + "epoch": 6.187418086500656, + "grad_norm": 5.436378479003906, + "learning_rate": 1.9060768981926137e-05, + "loss": 0.4906, + "num_input_tokens_seen": 6320704, + "step": 23605 + }, + { + "epoch": 6.18872870249017, + "grad_norm": 2.4815614223480225, + "learning_rate": 1.904965976417232e-05, + "loss": 0.4611, + "num_input_tokens_seen": 6321904, + "step": 23610 + }, + { + "epoch": 6.190039318479686, + "grad_norm": 2.4405438899993896, + "learning_rate": 1.9038551791814446e-05, + "loss": 0.563, + "num_input_tokens_seen": 6323360, + "step": 23615 + }, + { + "epoch": 6.1913499344692005, + "grad_norm": 2.4605977535247803, + "learning_rate": 1.902744506717738e-05, + "loss": 0.4521, + "num_input_tokens_seen": 6324512, + "step": 23620 + }, + { + "epoch": 6.192660550458716, + "grad_norm": 2.4272096157073975, + "learning_rate": 1.9016339592585737e-05, + "loss": 0.5673, + "num_input_tokens_seen": 6325984, + "step": 23625 + }, + { + "epoch": 6.193971166448231, + "grad_norm": 14.277854919433594, + "learning_rate": 1.9005235370363877e-05, + "loss": 0.7011, + "num_input_tokens_seen": 6326896, + "step": 23630 + }, + { + "epoch": 6.195281782437746, + "grad_norm": 6.261676788330078, + "learning_rate": 1.89941324028359e-05, + "loss": 0.4478, + "num_input_tokens_seen": 6327824, + "step": 23635 + }, + { + "epoch": 6.196592398427261, + "grad_norm": 1.8819833993911743, + "learning_rate": 1.8983030692325632e-05, + "loss": 0.5913, + "num_input_tokens_seen": 6329104, + "step": 23640 + }, + { + "epoch": 6.1979030144167755, + "grad_norm": 6.465756893157959, + "learning_rate": 1.8971930241156646e-05, + "loss": 0.3749, + "num_input_tokens_seen": 6330144, + "step": 23645 + }, + { + "epoch": 6.199213630406291, + "grad_norm": 7.42081880569458, + "learning_rate": 1.8960831051652243e-05, + "loss": 0.4117, + "num_input_tokens_seen": 6331360, + "step": 23650 + }, + { + "epoch": 6.200524246395806, + "grad_norm": 1.947418451309204, + "learning_rate": 1.8949733126135464e-05, + "loss": 0.5717, + "num_input_tokens_seen": 6332624, + "step": 23655 + }, + { + "epoch": 6.201834862385321, + "grad_norm": 2.5491037368774414, + "learning_rate": 1.8938636466929088e-05, + "loss": 0.3258, + "num_input_tokens_seen": 6333856, + "step": 23660 + }, + { + "epoch": 6.203145478374836, + "grad_norm": 2.192847728729248, + "learning_rate": 1.8927541076355626e-05, + "loss": 0.4084, + "num_input_tokens_seen": 6335136, + "step": 23665 + }, + { + "epoch": 6.204456094364351, + "grad_norm": 1.0824263095855713, + "learning_rate": 1.8916446956737317e-05, + "loss": 0.4117, + "num_input_tokens_seen": 6336880, + "step": 23670 + }, + { + "epoch": 6.205766710353866, + "grad_norm": 8.531237602233887, + "learning_rate": 1.890535411039616e-05, + "loss": 0.4858, + "num_input_tokens_seen": 6337792, + "step": 23675 + }, + { + "epoch": 6.207077326343382, + "grad_norm": 2.5583841800689697, + "learning_rate": 1.8894262539653858e-05, + "loss": 0.4723, + "num_input_tokens_seen": 6339056, + "step": 23680 + }, + { + "epoch": 6.208387942332896, + "grad_norm": 1.853534460067749, + "learning_rate": 1.888317224683185e-05, + "loss": 0.5307, + "num_input_tokens_seen": 6340544, + "step": 23685 + }, + { + "epoch": 6.209698558322412, + "grad_norm": 1.5476523637771606, + "learning_rate": 1.8872083234251324e-05, + "loss": 0.402, + "num_input_tokens_seen": 6341744, + "step": 23690 + }, + { + "epoch": 6.2110091743119265, + "grad_norm": 2.376169443130493, + "learning_rate": 1.8860995504233188e-05, + "loss": 0.5735, + "num_input_tokens_seen": 6342848, + "step": 23695 + }, + { + "epoch": 6.212319790301442, + "grad_norm": 2.7386109828948975, + "learning_rate": 1.8849909059098087e-05, + "loss": 0.294, + "num_input_tokens_seen": 6344304, + "step": 23700 + }, + { + "epoch": 6.213630406290957, + "grad_norm": 2.0161850452423096, + "learning_rate": 1.8838823901166396e-05, + "loss": 0.4115, + "num_input_tokens_seen": 6345392, + "step": 23705 + }, + { + "epoch": 6.214941022280472, + "grad_norm": 2.4145843982696533, + "learning_rate": 1.8827740032758215e-05, + "loss": 0.435, + "num_input_tokens_seen": 6346592, + "step": 23710 + }, + { + "epoch": 6.216251638269987, + "grad_norm": 3.1168646812438965, + "learning_rate": 1.8816657456193376e-05, + "loss": 0.5345, + "num_input_tokens_seen": 6347936, + "step": 23715 + }, + { + "epoch": 6.217562254259502, + "grad_norm": 2.3466482162475586, + "learning_rate": 1.880557617379145e-05, + "loss": 0.4498, + "num_input_tokens_seen": 6349344, + "step": 23720 + }, + { + "epoch": 6.218872870249017, + "grad_norm": 4.658753395080566, + "learning_rate": 1.8794496187871725e-05, + "loss": 0.4218, + "num_input_tokens_seen": 6350752, + "step": 23725 + }, + { + "epoch": 6.220183486238533, + "grad_norm": 6.900380611419678, + "learning_rate": 1.8783417500753224e-05, + "loss": 0.6081, + "num_input_tokens_seen": 6351680, + "step": 23730 + }, + { + "epoch": 6.221494102228047, + "grad_norm": 4.818958282470703, + "learning_rate": 1.8772340114754705e-05, + "loss": 0.4997, + "num_input_tokens_seen": 6353216, + "step": 23735 + }, + { + "epoch": 6.222804718217562, + "grad_norm": 2.2125136852264404, + "learning_rate": 1.8761264032194642e-05, + "loss": 0.4151, + "num_input_tokens_seen": 6354400, + "step": 23740 + }, + { + "epoch": 6.224115334207077, + "grad_norm": 4.533153057098389, + "learning_rate": 1.875018925539123e-05, + "loss": 0.5993, + "num_input_tokens_seen": 6355376, + "step": 23745 + }, + { + "epoch": 6.225425950196592, + "grad_norm": 1.8433500528335571, + "learning_rate": 1.8739115786662415e-05, + "loss": 0.4298, + "num_input_tokens_seen": 6356464, + "step": 23750 + }, + { + "epoch": 6.226736566186108, + "grad_norm": 7.247882843017578, + "learning_rate": 1.872804362832584e-05, + "loss": 0.4201, + "num_input_tokens_seen": 6357552, + "step": 23755 + }, + { + "epoch": 6.228047182175622, + "grad_norm": 2.873232364654541, + "learning_rate": 1.8716972782698905e-05, + "loss": 0.4332, + "num_input_tokens_seen": 6358768, + "step": 23760 + }, + { + "epoch": 6.229357798165138, + "grad_norm": 1.0776824951171875, + "learning_rate": 1.8705903252098704e-05, + "loss": 0.9152, + "num_input_tokens_seen": 6360336, + "step": 23765 + }, + { + "epoch": 6.2306684141546524, + "grad_norm": 5.71271276473999, + "learning_rate": 1.869483503884208e-05, + "loss": 0.3896, + "num_input_tokens_seen": 6361648, + "step": 23770 + }, + { + "epoch": 6.231979030144168, + "grad_norm": 1.7083042860031128, + "learning_rate": 1.8683768145245586e-05, + "loss": 0.397, + "num_input_tokens_seen": 6363120, + "step": 23775 + }, + { + "epoch": 6.233289646133683, + "grad_norm": 2.007005453109741, + "learning_rate": 1.8672702573625505e-05, + "loss": 0.3986, + "num_input_tokens_seen": 6364800, + "step": 23780 + }, + { + "epoch": 6.234600262123198, + "grad_norm": 3.085984945297241, + "learning_rate": 1.866163832629784e-05, + "loss": 0.5003, + "num_input_tokens_seen": 6366096, + "step": 23785 + }, + { + "epoch": 6.235910878112713, + "grad_norm": 1.2011326551437378, + "learning_rate": 1.8650575405578324e-05, + "loss": 0.2868, + "num_input_tokens_seen": 6367456, + "step": 23790 + }, + { + "epoch": 6.237221494102228, + "grad_norm": 1.7487878799438477, + "learning_rate": 1.8639513813782407e-05, + "loss": 0.4353, + "num_input_tokens_seen": 6368736, + "step": 23795 + }, + { + "epoch": 6.238532110091743, + "grad_norm": 2.8737313747406006, + "learning_rate": 1.8628453553225262e-05, + "loss": 0.5764, + "num_input_tokens_seen": 6369744, + "step": 23800 + }, + { + "epoch": 6.239842726081259, + "grad_norm": 8.622292518615723, + "learning_rate": 1.8617394626221774e-05, + "loss": 0.7085, + "num_input_tokens_seen": 6371232, + "step": 23805 + }, + { + "epoch": 6.241153342070773, + "grad_norm": 6.422327041625977, + "learning_rate": 1.8606337035086564e-05, + "loss": 0.3142, + "num_input_tokens_seen": 6372224, + "step": 23810 + }, + { + "epoch": 6.242463958060289, + "grad_norm": 1.562195062637329, + "learning_rate": 1.859528078213396e-05, + "loss": 0.3539, + "num_input_tokens_seen": 6373600, + "step": 23815 + }, + { + "epoch": 6.243774574049803, + "grad_norm": 9.634483337402344, + "learning_rate": 1.8584225869678033e-05, + "loss": 0.3872, + "num_input_tokens_seen": 6374944, + "step": 23820 + }, + { + "epoch": 6.245085190039318, + "grad_norm": 4.290041923522949, + "learning_rate": 1.8573172300032538e-05, + "loss": 0.3697, + "num_input_tokens_seen": 6376464, + "step": 23825 + }, + { + "epoch": 6.246395806028834, + "grad_norm": 4.111637592315674, + "learning_rate": 1.8562120075510976e-05, + "loss": 0.3561, + "num_input_tokens_seen": 6380096, + "step": 23830 + }, + { + "epoch": 6.247706422018348, + "grad_norm": 2.8830161094665527, + "learning_rate": 1.8551069198426556e-05, + "loss": 0.5024, + "num_input_tokens_seen": 6381328, + "step": 23835 + }, + { + "epoch": 6.249017038007864, + "grad_norm": 3.694854497909546, + "learning_rate": 1.854001967109221e-05, + "loss": 0.4077, + "num_input_tokens_seen": 6382656, + "step": 23840 + }, + { + "epoch": 6.250327653997378, + "grad_norm": 2.4715356826782227, + "learning_rate": 1.8528971495820584e-05, + "loss": 0.508, + "num_input_tokens_seen": 6383872, + "step": 23845 + }, + { + "epoch": 6.251638269986894, + "grad_norm": 1.9161360263824463, + "learning_rate": 1.8517924674924048e-05, + "loss": 0.4222, + "num_input_tokens_seen": 6385568, + "step": 23850 + }, + { + "epoch": 6.252948885976409, + "grad_norm": 2.109057664871216, + "learning_rate": 1.8506879210714685e-05, + "loss": 0.5521, + "num_input_tokens_seen": 6386704, + "step": 23855 + }, + { + "epoch": 6.254259501965924, + "grad_norm": 2.3162877559661865, + "learning_rate": 1.849583510550427e-05, + "loss": 0.426, + "num_input_tokens_seen": 6387712, + "step": 23860 + }, + { + "epoch": 6.255570117955439, + "grad_norm": 5.546384811401367, + "learning_rate": 1.8484792361604336e-05, + "loss": 0.548, + "num_input_tokens_seen": 6388848, + "step": 23865 + }, + { + "epoch": 6.256880733944954, + "grad_norm": 1.5562869310379028, + "learning_rate": 1.84737509813261e-05, + "loss": 0.6005, + "num_input_tokens_seen": 6390272, + "step": 23870 + }, + { + "epoch": 6.258191349934469, + "grad_norm": 2.2355499267578125, + "learning_rate": 1.8462710966980506e-05, + "loss": 0.5581, + "num_input_tokens_seen": 6391792, + "step": 23875 + }, + { + "epoch": 6.259501965923985, + "grad_norm": 6.21510648727417, + "learning_rate": 1.845167232087822e-05, + "loss": 0.5632, + "num_input_tokens_seen": 6393040, + "step": 23880 + }, + { + "epoch": 6.260812581913499, + "grad_norm": 3.6930103302001953, + "learning_rate": 1.8440635045329595e-05, + "loss": 0.6234, + "num_input_tokens_seen": 6394240, + "step": 23885 + }, + { + "epoch": 6.262123197903015, + "grad_norm": 2.3286099433898926, + "learning_rate": 1.8429599142644724e-05, + "loss": 0.3285, + "num_input_tokens_seen": 6395248, + "step": 23890 + }, + { + "epoch": 6.263433813892529, + "grad_norm": 2.335928201675415, + "learning_rate": 1.84185646151334e-05, + "loss": 0.3806, + "num_input_tokens_seen": 6396112, + "step": 23895 + }, + { + "epoch": 6.264744429882045, + "grad_norm": 2.346669912338257, + "learning_rate": 1.840753146510512e-05, + "loss": 0.8045, + "num_input_tokens_seen": 6397744, + "step": 23900 + }, + { + "epoch": 6.26605504587156, + "grad_norm": 2.233994960784912, + "learning_rate": 1.8396499694869123e-05, + "loss": 0.5155, + "num_input_tokens_seen": 6399104, + "step": 23905 + }, + { + "epoch": 6.267365661861074, + "grad_norm": 6.894655227661133, + "learning_rate": 1.8385469306734327e-05, + "loss": 0.4996, + "num_input_tokens_seen": 6400320, + "step": 23910 + }, + { + "epoch": 6.26867627785059, + "grad_norm": 3.246248960494995, + "learning_rate": 1.8374440303009382e-05, + "loss": 0.4302, + "num_input_tokens_seen": 6401616, + "step": 23915 + }, + { + "epoch": 6.269986893840104, + "grad_norm": 4.321455955505371, + "learning_rate": 1.836341268600263e-05, + "loss": 0.4099, + "num_input_tokens_seen": 6403216, + "step": 23920 + }, + { + "epoch": 6.27129750982962, + "grad_norm": 1.4920928478240967, + "learning_rate": 1.8352386458022135e-05, + "loss": 0.545, + "num_input_tokens_seen": 6404688, + "step": 23925 + }, + { + "epoch": 6.272608125819135, + "grad_norm": 1.8319275379180908, + "learning_rate": 1.8341361621375666e-05, + "loss": 0.5177, + "num_input_tokens_seen": 6406336, + "step": 23930 + }, + { + "epoch": 6.27391874180865, + "grad_norm": 7.065600872039795, + "learning_rate": 1.8330338178370706e-05, + "loss": 0.4277, + "num_input_tokens_seen": 6407872, + "step": 23935 + }, + { + "epoch": 6.275229357798165, + "grad_norm": 2.161468744277954, + "learning_rate": 1.8319316131314447e-05, + "loss": 0.7763, + "num_input_tokens_seen": 6409584, + "step": 23940 + }, + { + "epoch": 6.27653997378768, + "grad_norm": 2.2392263412475586, + "learning_rate": 1.830829548251377e-05, + "loss": 0.5141, + "num_input_tokens_seen": 6411088, + "step": 23945 + }, + { + "epoch": 6.277850589777195, + "grad_norm": 9.734177589416504, + "learning_rate": 1.829727623427529e-05, + "loss": 0.5542, + "num_input_tokens_seen": 6412192, + "step": 23950 + }, + { + "epoch": 6.2791612057667106, + "grad_norm": 3.8816473484039307, + "learning_rate": 1.8286258388905308e-05, + "loss": 0.5203, + "num_input_tokens_seen": 6413408, + "step": 23955 + }, + { + "epoch": 6.280471821756225, + "grad_norm": 3.145413875579834, + "learning_rate": 1.8275241948709847e-05, + "loss": 0.3701, + "num_input_tokens_seen": 6414544, + "step": 23960 + }, + { + "epoch": 6.281782437745741, + "grad_norm": 28.22792625427246, + "learning_rate": 1.8264226915994627e-05, + "loss": 0.5438, + "num_input_tokens_seen": 6415952, + "step": 23965 + }, + { + "epoch": 6.283093053735255, + "grad_norm": 6.619207859039307, + "learning_rate": 1.8253213293065077e-05, + "loss": 0.4848, + "num_input_tokens_seen": 6417248, + "step": 23970 + }, + { + "epoch": 6.284403669724771, + "grad_norm": 2.024956464767456, + "learning_rate": 1.8242201082226332e-05, + "loss": 0.4927, + "num_input_tokens_seen": 6418784, + "step": 23975 + }, + { + "epoch": 6.285714285714286, + "grad_norm": 2.156810998916626, + "learning_rate": 1.8231190285783216e-05, + "loss": 0.4052, + "num_input_tokens_seen": 6420416, + "step": 23980 + }, + { + "epoch": 6.287024901703801, + "grad_norm": 1.373143196105957, + "learning_rate": 1.822018090604028e-05, + "loss": 0.3892, + "num_input_tokens_seen": 6421728, + "step": 23985 + }, + { + "epoch": 6.288335517693316, + "grad_norm": 2.1943132877349854, + "learning_rate": 1.8209172945301768e-05, + "loss": 0.4143, + "num_input_tokens_seen": 6422848, + "step": 23990 + }, + { + "epoch": 6.289646133682831, + "grad_norm": 2.7427823543548584, + "learning_rate": 1.819816640587162e-05, + "loss": 0.6059, + "num_input_tokens_seen": 6424176, + "step": 23995 + }, + { + "epoch": 6.290956749672346, + "grad_norm": 4.744399070739746, + "learning_rate": 1.8187161290053498e-05, + "loss": 0.3833, + "num_input_tokens_seen": 6425440, + "step": 24000 + }, + { + "epoch": 6.292267365661861, + "grad_norm": 3.814157247543335, + "learning_rate": 1.817615760015074e-05, + "loss": 0.4204, + "num_input_tokens_seen": 6426736, + "step": 24005 + }, + { + "epoch": 6.293577981651376, + "grad_norm": 2.0026140213012695, + "learning_rate": 1.8165155338466404e-05, + "loss": 0.4943, + "num_input_tokens_seen": 6427744, + "step": 24010 + }, + { + "epoch": 6.294888597640891, + "grad_norm": 2.8567399978637695, + "learning_rate": 1.8154154507303245e-05, + "loss": 0.4111, + "num_input_tokens_seen": 6428880, + "step": 24015 + }, + { + "epoch": 6.296199213630406, + "grad_norm": 1.4260332584381104, + "learning_rate": 1.814315510896372e-05, + "loss": 0.463, + "num_input_tokens_seen": 6430352, + "step": 24020 + }, + { + "epoch": 6.297509829619921, + "grad_norm": 3.7276439666748047, + "learning_rate": 1.813215714574999e-05, + "loss": 0.4329, + "num_input_tokens_seen": 6431440, + "step": 24025 + }, + { + "epoch": 6.2988204456094365, + "grad_norm": 3.4870989322662354, + "learning_rate": 1.8121160619963894e-05, + "loss": 0.4547, + "num_input_tokens_seen": 6433088, + "step": 24030 + }, + { + "epoch": 6.300131061598951, + "grad_norm": 3.9024834632873535, + "learning_rate": 1.8110165533907007e-05, + "loss": 0.4326, + "num_input_tokens_seen": 6434400, + "step": 24035 + }, + { + "epoch": 6.301441677588467, + "grad_norm": 10.411323547363281, + "learning_rate": 1.809917188988056e-05, + "loss": 0.4518, + "num_input_tokens_seen": 6435664, + "step": 24040 + }, + { + "epoch": 6.302752293577981, + "grad_norm": 3.697288990020752, + "learning_rate": 1.8088179690185515e-05, + "loss": 0.379, + "num_input_tokens_seen": 6436832, + "step": 24045 + }, + { + "epoch": 6.304062909567497, + "grad_norm": 6.194120407104492, + "learning_rate": 1.8077188937122515e-05, + "loss": 0.451, + "num_input_tokens_seen": 6438240, + "step": 24050 + }, + { + "epoch": 6.305373525557012, + "grad_norm": 1.8007792234420776, + "learning_rate": 1.8066199632991913e-05, + "loss": 0.3836, + "num_input_tokens_seen": 6439536, + "step": 24055 + }, + { + "epoch": 6.306684141546527, + "grad_norm": 3.085712432861328, + "learning_rate": 1.805521178009375e-05, + "loss": 0.3457, + "num_input_tokens_seen": 6440976, + "step": 24060 + }, + { + "epoch": 6.307994757536042, + "grad_norm": 2.422980546951294, + "learning_rate": 1.804422538072775e-05, + "loss": 0.6211, + "num_input_tokens_seen": 6442096, + "step": 24065 + }, + { + "epoch": 6.309305373525557, + "grad_norm": 1.3518458604812622, + "learning_rate": 1.8033240437193364e-05, + "loss": 0.503, + "num_input_tokens_seen": 6443840, + "step": 24070 + }, + { + "epoch": 6.310615989515072, + "grad_norm": 7.996537208557129, + "learning_rate": 1.8022256951789718e-05, + "loss": 0.4386, + "num_input_tokens_seen": 6445136, + "step": 24075 + }, + { + "epoch": 6.3119266055045875, + "grad_norm": 2.405174493789673, + "learning_rate": 1.8011274926815635e-05, + "loss": 0.3649, + "num_input_tokens_seen": 6446512, + "step": 24080 + }, + { + "epoch": 6.313237221494102, + "grad_norm": 1.8301442861557007, + "learning_rate": 1.8000294364569636e-05, + "loss": 0.4015, + "num_input_tokens_seen": 6447968, + "step": 24085 + }, + { + "epoch": 6.314547837483618, + "grad_norm": 2.898913860321045, + "learning_rate": 1.7989315267349936e-05, + "loss": 0.6067, + "num_input_tokens_seen": 6449008, + "step": 24090 + }, + { + "epoch": 6.315858453473132, + "grad_norm": 1.3546570539474487, + "learning_rate": 1.7978337637454433e-05, + "loss": 0.5723, + "num_input_tokens_seen": 6450448, + "step": 24095 + }, + { + "epoch": 6.317169069462647, + "grad_norm": 2.916247844696045, + "learning_rate": 1.7967361477180727e-05, + "loss": 0.4058, + "num_input_tokens_seen": 6451984, + "step": 24100 + }, + { + "epoch": 6.3184796854521625, + "grad_norm": 4.317192554473877, + "learning_rate": 1.7956386788826112e-05, + "loss": 0.326, + "num_input_tokens_seen": 6452992, + "step": 24105 + }, + { + "epoch": 6.319790301441677, + "grad_norm": 5.851595878601074, + "learning_rate": 1.7945413574687574e-05, + "loss": 0.3735, + "num_input_tokens_seen": 6454304, + "step": 24110 + }, + { + "epoch": 6.321100917431193, + "grad_norm": 4.281556606292725, + "learning_rate": 1.7934441837061795e-05, + "loss": 0.4939, + "num_input_tokens_seen": 6455424, + "step": 24115 + }, + { + "epoch": 6.322411533420707, + "grad_norm": 9.49299430847168, + "learning_rate": 1.792347157824512e-05, + "loss": 0.6182, + "num_input_tokens_seen": 6456544, + "step": 24120 + }, + { + "epoch": 6.323722149410223, + "grad_norm": 5.209400177001953, + "learning_rate": 1.7912502800533604e-05, + "loss": 0.4771, + "num_input_tokens_seen": 6457760, + "step": 24125 + }, + { + "epoch": 6.325032765399738, + "grad_norm": 1.866782784461975, + "learning_rate": 1.7901535506223027e-05, + "loss": 0.4765, + "num_input_tokens_seen": 6459024, + "step": 24130 + }, + { + "epoch": 6.326343381389253, + "grad_norm": 2.577653408050537, + "learning_rate": 1.7890569697608793e-05, + "loss": 0.4021, + "num_input_tokens_seen": 6461056, + "step": 24135 + }, + { + "epoch": 6.327653997378768, + "grad_norm": 2.009073495864868, + "learning_rate": 1.787960537698604e-05, + "loss": 0.328, + "num_input_tokens_seen": 6462160, + "step": 24140 + }, + { + "epoch": 6.328964613368283, + "grad_norm": 2.0231354236602783, + "learning_rate": 1.786864254664958e-05, + "loss": 0.6153, + "num_input_tokens_seen": 6463520, + "step": 24145 + }, + { + "epoch": 6.330275229357798, + "grad_norm": 4.373239517211914, + "learning_rate": 1.785768120889392e-05, + "loss": 0.443, + "num_input_tokens_seen": 6465328, + "step": 24150 + }, + { + "epoch": 6.3315858453473135, + "grad_norm": 3.2270548343658447, + "learning_rate": 1.784672136601323e-05, + "loss": 0.3125, + "num_input_tokens_seen": 6466496, + "step": 24155 + }, + { + "epoch": 6.332896461336828, + "grad_norm": 1.5191535949707031, + "learning_rate": 1.7835763020301403e-05, + "loss": 0.4488, + "num_input_tokens_seen": 6467872, + "step": 24160 + }, + { + "epoch": 6.334207077326344, + "grad_norm": 7.07719612121582, + "learning_rate": 1.7824806174051995e-05, + "loss": 0.5033, + "num_input_tokens_seen": 6469200, + "step": 24165 + }, + { + "epoch": 6.335517693315858, + "grad_norm": 3.1436619758605957, + "learning_rate": 1.7813850829558257e-05, + "loss": 0.555, + "num_input_tokens_seen": 6470272, + "step": 24170 + }, + { + "epoch": 6.336828309305374, + "grad_norm": 3.316307783126831, + "learning_rate": 1.7802896989113127e-05, + "loss": 0.6017, + "num_input_tokens_seen": 6471504, + "step": 24175 + }, + { + "epoch": 6.3381389252948885, + "grad_norm": 1.754928469657898, + "learning_rate": 1.779194465500922e-05, + "loss": 0.3939, + "num_input_tokens_seen": 6473008, + "step": 24180 + }, + { + "epoch": 6.339449541284404, + "grad_norm": 1.243839979171753, + "learning_rate": 1.7780993829538836e-05, + "loss": 0.522, + "num_input_tokens_seen": 6474880, + "step": 24185 + }, + { + "epoch": 6.340760157273919, + "grad_norm": 5.128986835479736, + "learning_rate": 1.7770044514993973e-05, + "loss": 0.4706, + "num_input_tokens_seen": 6476032, + "step": 24190 + }, + { + "epoch": 6.342070773263433, + "grad_norm": 3.2709004878997803, + "learning_rate": 1.7759096713666305e-05, + "loss": 0.4361, + "num_input_tokens_seen": 6477136, + "step": 24195 + }, + { + "epoch": 6.343381389252949, + "grad_norm": 2.068885087966919, + "learning_rate": 1.774815042784718e-05, + "loss": 0.4853, + "num_input_tokens_seen": 6478640, + "step": 24200 + }, + { + "epoch": 6.344692005242464, + "grad_norm": 3.7399845123291016, + "learning_rate": 1.773720565982764e-05, + "loss": 0.4407, + "num_input_tokens_seen": 6479920, + "step": 24205 + }, + { + "epoch": 6.346002621231979, + "grad_norm": 6.034506320953369, + "learning_rate": 1.7726262411898413e-05, + "loss": 0.4865, + "num_input_tokens_seen": 6481216, + "step": 24210 + }, + { + "epoch": 6.347313237221494, + "grad_norm": 3.4343032836914062, + "learning_rate": 1.771532068634989e-05, + "loss": 0.4607, + "num_input_tokens_seen": 6482240, + "step": 24215 + }, + { + "epoch": 6.348623853211009, + "grad_norm": 7.677488803863525, + "learning_rate": 1.770438048547216e-05, + "loss": 0.6394, + "num_input_tokens_seen": 6483200, + "step": 24220 + }, + { + "epoch": 6.349934469200524, + "grad_norm": 3.7322683334350586, + "learning_rate": 1.769344181155499e-05, + "loss": 0.4631, + "num_input_tokens_seen": 6484768, + "step": 24225 + }, + { + "epoch": 6.3512450851900395, + "grad_norm": 3.708528757095337, + "learning_rate": 1.7682504666887823e-05, + "loss": 0.544, + "num_input_tokens_seen": 6486000, + "step": 24230 + }, + { + "epoch": 6.352555701179554, + "grad_norm": 4.756847381591797, + "learning_rate": 1.767156905375979e-05, + "loss": 0.4283, + "num_input_tokens_seen": 6487056, + "step": 24235 + }, + { + "epoch": 6.35386631716907, + "grad_norm": 2.19622802734375, + "learning_rate": 1.7660634974459674e-05, + "loss": 0.5696, + "num_input_tokens_seen": 6488096, + "step": 24240 + }, + { + "epoch": 6.355176933158584, + "grad_norm": 2.1818387508392334, + "learning_rate": 1.7649702431275994e-05, + "loss": 0.5685, + "num_input_tokens_seen": 6489744, + "step": 24245 + }, + { + "epoch": 6.3564875491481, + "grad_norm": 1.3854869604110718, + "learning_rate": 1.7638771426496893e-05, + "loss": 0.5428, + "num_input_tokens_seen": 6491424, + "step": 24250 + }, + { + "epoch": 6.3577981651376145, + "grad_norm": 3.5459413528442383, + "learning_rate": 1.7627841962410208e-05, + "loss": 0.3735, + "num_input_tokens_seen": 6492400, + "step": 24255 + }, + { + "epoch": 6.35910878112713, + "grad_norm": 2.302804470062256, + "learning_rate": 1.7616914041303463e-05, + "loss": 0.8852, + "num_input_tokens_seen": 6493408, + "step": 24260 + }, + { + "epoch": 6.360419397116645, + "grad_norm": 0.8320634365081787, + "learning_rate": 1.7605987665463854e-05, + "loss": 0.5352, + "num_input_tokens_seen": 6495488, + "step": 24265 + }, + { + "epoch": 6.36173001310616, + "grad_norm": 12.013195037841797, + "learning_rate": 1.7595062837178254e-05, + "loss": 0.7196, + "num_input_tokens_seen": 6496720, + "step": 24270 + }, + { + "epoch": 6.363040629095675, + "grad_norm": 2.7360525131225586, + "learning_rate": 1.7584139558733205e-05, + "loss": 0.3953, + "num_input_tokens_seen": 6498896, + "step": 24275 + }, + { + "epoch": 6.3643512450851905, + "grad_norm": 3.0265235900878906, + "learning_rate": 1.757321783241493e-05, + "loss": 0.5068, + "num_input_tokens_seen": 6500160, + "step": 24280 + }, + { + "epoch": 6.365661861074705, + "grad_norm": 3.4128782749176025, + "learning_rate": 1.7562297660509334e-05, + "loss": 0.5044, + "num_input_tokens_seen": 6501808, + "step": 24285 + }, + { + "epoch": 6.36697247706422, + "grad_norm": 3.578380823135376, + "learning_rate": 1.7551379045301987e-05, + "loss": 0.4586, + "num_input_tokens_seen": 6503008, + "step": 24290 + }, + { + "epoch": 6.368283093053735, + "grad_norm": 3.2705094814300537, + "learning_rate": 1.7540461989078132e-05, + "loss": 0.4653, + "num_input_tokens_seen": 6504320, + "step": 24295 + }, + { + "epoch": 6.36959370904325, + "grad_norm": 25.05501937866211, + "learning_rate": 1.7529546494122685e-05, + "loss": 0.5738, + "num_input_tokens_seen": 6505312, + "step": 24300 + }, + { + "epoch": 6.3709043250327655, + "grad_norm": 3.6016364097595215, + "learning_rate": 1.7518632562720263e-05, + "loss": 0.4936, + "num_input_tokens_seen": 6506416, + "step": 24305 + }, + { + "epoch": 6.37221494102228, + "grad_norm": 1.828049659729004, + "learning_rate": 1.7507720197155114e-05, + "loss": 0.3777, + "num_input_tokens_seen": 6507424, + "step": 24310 + }, + { + "epoch": 6.373525557011796, + "grad_norm": 4.007122993469238, + "learning_rate": 1.749680939971118e-05, + "loss": 0.4184, + "num_input_tokens_seen": 6508464, + "step": 24315 + }, + { + "epoch": 6.37483617300131, + "grad_norm": 19.644521713256836, + "learning_rate": 1.7485900172672076e-05, + "loss": 0.5743, + "num_input_tokens_seen": 6509328, + "step": 24320 + }, + { + "epoch": 6.376146788990826, + "grad_norm": 3.1949610710144043, + "learning_rate": 1.7474992518321083e-05, + "loss": 0.4274, + "num_input_tokens_seen": 6510624, + "step": 24325 + }, + { + "epoch": 6.3774574049803405, + "grad_norm": 4.361968994140625, + "learning_rate": 1.7464086438941156e-05, + "loss": 0.3156, + "num_input_tokens_seen": 6511840, + "step": 24330 + }, + { + "epoch": 6.378768020969856, + "grad_norm": 4.641468524932861, + "learning_rate": 1.7453181936814915e-05, + "loss": 0.3086, + "num_input_tokens_seen": 6513024, + "step": 24335 + }, + { + "epoch": 6.380078636959371, + "grad_norm": 5.84036111831665, + "learning_rate": 1.7442279014224654e-05, + "loss": 0.4096, + "num_input_tokens_seen": 6514336, + "step": 24340 + }, + { + "epoch": 6.381389252948886, + "grad_norm": 6.333864212036133, + "learning_rate": 1.7431377673452336e-05, + "loss": 0.3705, + "num_input_tokens_seen": 6515280, + "step": 24345 + }, + { + "epoch": 6.382699868938401, + "grad_norm": 1.716847538948059, + "learning_rate": 1.74204779167796e-05, + "loss": 0.4841, + "num_input_tokens_seen": 6516704, + "step": 24350 + }, + { + "epoch": 6.3840104849279165, + "grad_norm": 2.027841567993164, + "learning_rate": 1.740957974648772e-05, + "loss": 0.4151, + "num_input_tokens_seen": 6518144, + "step": 24355 + }, + { + "epoch": 6.385321100917431, + "grad_norm": 6.474591255187988, + "learning_rate": 1.7398683164857697e-05, + "loss": 0.4136, + "num_input_tokens_seen": 6519104, + "step": 24360 + }, + { + "epoch": 6.386631716906947, + "grad_norm": 0.7675861120223999, + "learning_rate": 1.7387788174170157e-05, + "loss": 0.3175, + "num_input_tokens_seen": 6520688, + "step": 24365 + }, + { + "epoch": 6.387942332896461, + "grad_norm": 4.911305904388428, + "learning_rate": 1.7376894776705397e-05, + "loss": 0.4281, + "num_input_tokens_seen": 6522208, + "step": 24370 + }, + { + "epoch": 6.389252948885977, + "grad_norm": 3.136002540588379, + "learning_rate": 1.7366002974743383e-05, + "loss": 0.4641, + "num_input_tokens_seen": 6523664, + "step": 24375 + }, + { + "epoch": 6.3905635648754915, + "grad_norm": 2.130223512649536, + "learning_rate": 1.7355112770563764e-05, + "loss": 0.4336, + "num_input_tokens_seen": 6525280, + "step": 24380 + }, + { + "epoch": 6.391874180865006, + "grad_norm": 3.456307888031006, + "learning_rate": 1.734422416644583e-05, + "loss": 0.4134, + "num_input_tokens_seen": 6526496, + "step": 24385 + }, + { + "epoch": 6.393184796854522, + "grad_norm": 4.3005242347717285, + "learning_rate": 1.733333716466855e-05, + "loss": 0.5687, + "num_input_tokens_seen": 6527840, + "step": 24390 + }, + { + "epoch": 6.394495412844036, + "grad_norm": 2.808234930038452, + "learning_rate": 1.732245176751055e-05, + "loss": 0.5609, + "num_input_tokens_seen": 6529232, + "step": 24395 + }, + { + "epoch": 6.395806028833552, + "grad_norm": 1.6002638339996338, + "learning_rate": 1.7311567977250133e-05, + "loss": 0.4884, + "num_input_tokens_seen": 6530464, + "step": 24400 + }, + { + "epoch": 6.3971166448230665, + "grad_norm": 3.362884521484375, + "learning_rate": 1.7300685796165255e-05, + "loss": 0.3876, + "num_input_tokens_seen": 6531552, + "step": 24405 + }, + { + "epoch": 6.398427260812582, + "grad_norm": 1.9853761196136475, + "learning_rate": 1.728980522653354e-05, + "loss": 0.3991, + "num_input_tokens_seen": 6532912, + "step": 24410 + }, + { + "epoch": 6.399737876802097, + "grad_norm": 8.47537612915039, + "learning_rate": 1.727892627063225e-05, + "loss": 0.6209, + "num_input_tokens_seen": 6534144, + "step": 24415 + }, + { + "epoch": 6.401048492791612, + "grad_norm": 3.1139142513275146, + "learning_rate": 1.726804893073837e-05, + "loss": 0.3589, + "num_input_tokens_seen": 6535168, + "step": 24420 + }, + { + "epoch": 6.402359108781127, + "grad_norm": 5.042975902557373, + "learning_rate": 1.725717320912848e-05, + "loss": 0.2941, + "num_input_tokens_seen": 6536384, + "step": 24425 + }, + { + "epoch": 6.4036697247706424, + "grad_norm": 4.20203971862793, + "learning_rate": 1.724629910807886e-05, + "loss": 0.3602, + "num_input_tokens_seen": 6538032, + "step": 24430 + }, + { + "epoch": 6.404980340760157, + "grad_norm": 0.9270637035369873, + "learning_rate": 1.7235426629865445e-05, + "loss": 0.4125, + "num_input_tokens_seen": 6539424, + "step": 24435 + }, + { + "epoch": 6.406290956749673, + "grad_norm": 5.448267936706543, + "learning_rate": 1.7224555776763814e-05, + "loss": 0.3315, + "num_input_tokens_seen": 6541152, + "step": 24440 + }, + { + "epoch": 6.407601572739187, + "grad_norm": 4.638136863708496, + "learning_rate": 1.7213686551049233e-05, + "loss": 0.4825, + "num_input_tokens_seen": 6542144, + "step": 24445 + }, + { + "epoch": 6.408912188728703, + "grad_norm": 4.543063640594482, + "learning_rate": 1.7202818954996597e-05, + "loss": 0.4914, + "num_input_tokens_seen": 6543376, + "step": 24450 + }, + { + "epoch": 6.4102228047182175, + "grad_norm": 2.1888484954833984, + "learning_rate": 1.7191952990880476e-05, + "loss": 0.3596, + "num_input_tokens_seen": 6544480, + "step": 24455 + }, + { + "epoch": 6.411533420707733, + "grad_norm": 1.2121407985687256, + "learning_rate": 1.7181088660975107e-05, + "loss": 0.4044, + "num_input_tokens_seen": 6545808, + "step": 24460 + }, + { + "epoch": 6.412844036697248, + "grad_norm": 13.700643539428711, + "learning_rate": 1.7170225967554374e-05, + "loss": 0.6071, + "num_input_tokens_seen": 6547120, + "step": 24465 + }, + { + "epoch": 6.414154652686763, + "grad_norm": 3.3637583255767822, + "learning_rate": 1.715936491289181e-05, + "loss": 0.551, + "num_input_tokens_seen": 6548352, + "step": 24470 + }, + { + "epoch": 6.415465268676278, + "grad_norm": 2.3174057006835938, + "learning_rate": 1.7148505499260627e-05, + "loss": 0.3812, + "num_input_tokens_seen": 6549920, + "step": 24475 + }, + { + "epoch": 6.4167758846657925, + "grad_norm": 1.5164521932601929, + "learning_rate": 1.713764772893368e-05, + "loss": 0.3171, + "num_input_tokens_seen": 6551296, + "step": 24480 + }, + { + "epoch": 6.418086500655308, + "grad_norm": 2.701894760131836, + "learning_rate": 1.7126791604183474e-05, + "loss": 0.5643, + "num_input_tokens_seen": 6552512, + "step": 24485 + }, + { + "epoch": 6.419397116644823, + "grad_norm": 3.2761967182159424, + "learning_rate": 1.7115937127282182e-05, + "loss": 0.5281, + "num_input_tokens_seen": 6553824, + "step": 24490 + }, + { + "epoch": 6.420707732634338, + "grad_norm": 3.7010419368743896, + "learning_rate": 1.7105084300501627e-05, + "loss": 0.4516, + "num_input_tokens_seen": 6555072, + "step": 24495 + }, + { + "epoch": 6.422018348623853, + "grad_norm": 3.6815237998962402, + "learning_rate": 1.7094233126113285e-05, + "loss": 0.4915, + "num_input_tokens_seen": 6556224, + "step": 24500 + }, + { + "epoch": 6.423328964613368, + "grad_norm": 2.199026107788086, + "learning_rate": 1.70833836063883e-05, + "loss": 0.4141, + "num_input_tokens_seen": 6557680, + "step": 24505 + }, + { + "epoch": 6.424639580602883, + "grad_norm": 2.528822898864746, + "learning_rate": 1.7072535743597433e-05, + "loss": 0.5127, + "num_input_tokens_seen": 6558496, + "step": 24510 + }, + { + "epoch": 6.425950196592399, + "grad_norm": 1.2646111249923706, + "learning_rate": 1.7061689540011144e-05, + "loss": 0.4552, + "num_input_tokens_seen": 6559872, + "step": 24515 + }, + { + "epoch": 6.427260812581913, + "grad_norm": 8.066503524780273, + "learning_rate": 1.7050844997899517e-05, + "loss": 0.5532, + "num_input_tokens_seen": 6561328, + "step": 24520 + }, + { + "epoch": 6.428571428571429, + "grad_norm": 8.671228408813477, + "learning_rate": 1.70400021195323e-05, + "loss": 0.34, + "num_input_tokens_seen": 6562848, + "step": 24525 + }, + { + "epoch": 6.4298820445609435, + "grad_norm": 2.1547844409942627, + "learning_rate": 1.702916090717887e-05, + "loss": 0.3603, + "num_input_tokens_seen": 6564656, + "step": 24530 + }, + { + "epoch": 6.431192660550459, + "grad_norm": 2.0320303440093994, + "learning_rate": 1.7018321363108303e-05, + "loss": 0.4334, + "num_input_tokens_seen": 6565712, + "step": 24535 + }, + { + "epoch": 6.432503276539974, + "grad_norm": 8.203434944152832, + "learning_rate": 1.700748348958929e-05, + "loss": 0.3438, + "num_input_tokens_seen": 6566880, + "step": 24540 + }, + { + "epoch": 6.433813892529489, + "grad_norm": 6.259047508239746, + "learning_rate": 1.699664728889016e-05, + "loss": 0.4411, + "num_input_tokens_seen": 6568128, + "step": 24545 + }, + { + "epoch": 6.435124508519004, + "grad_norm": 3.503896474838257, + "learning_rate": 1.698581276327893e-05, + "loss": 0.4253, + "num_input_tokens_seen": 6569600, + "step": 24550 + }, + { + "epoch": 6.436435124508519, + "grad_norm": 2.4775521755218506, + "learning_rate": 1.6974979915023237e-05, + "loss": 0.354, + "num_input_tokens_seen": 6570640, + "step": 24555 + }, + { + "epoch": 6.437745740498034, + "grad_norm": 2.0727531909942627, + "learning_rate": 1.6964148746390383e-05, + "loss": 0.7281, + "num_input_tokens_seen": 6571712, + "step": 24560 + }, + { + "epoch": 6.43905635648755, + "grad_norm": 2.056541681289673, + "learning_rate": 1.6953319259647323e-05, + "loss": 0.5809, + "num_input_tokens_seen": 6572768, + "step": 24565 + }, + { + "epoch": 6.440366972477064, + "grad_norm": 4.808716773986816, + "learning_rate": 1.6942491457060626e-05, + "loss": 0.5349, + "num_input_tokens_seen": 6573792, + "step": 24570 + }, + { + "epoch": 6.441677588466579, + "grad_norm": 2.578075408935547, + "learning_rate": 1.6931665340896545e-05, + "loss": 0.4793, + "num_input_tokens_seen": 6575376, + "step": 24575 + }, + { + "epoch": 6.442988204456094, + "grad_norm": 3.985480308532715, + "learning_rate": 1.692084091342097e-05, + "loss": 0.4533, + "num_input_tokens_seen": 6576480, + "step": 24580 + }, + { + "epoch": 6.444298820445609, + "grad_norm": 11.481932640075684, + "learning_rate": 1.6910018176899424e-05, + "loss": 0.3509, + "num_input_tokens_seen": 6577584, + "step": 24585 + }, + { + "epoch": 6.445609436435125, + "grad_norm": 8.023473739624023, + "learning_rate": 1.68991971335971e-05, + "loss": 0.4803, + "num_input_tokens_seen": 6578768, + "step": 24590 + }, + { + "epoch": 6.446920052424639, + "grad_norm": 2.6987738609313965, + "learning_rate": 1.688837778577882e-05, + "loss": 0.5475, + "num_input_tokens_seen": 6580320, + "step": 24595 + }, + { + "epoch": 6.448230668414155, + "grad_norm": 2.65008544921875, + "learning_rate": 1.687756013570906e-05, + "loss": 0.335, + "num_input_tokens_seen": 6582032, + "step": 24600 + }, + { + "epoch": 6.4495412844036695, + "grad_norm": 2.257495403289795, + "learning_rate": 1.6866744185651922e-05, + "loss": 0.3926, + "num_input_tokens_seen": 6583424, + "step": 24605 + }, + { + "epoch": 6.450851900393185, + "grad_norm": 2.692531108856201, + "learning_rate": 1.6855929937871175e-05, + "loss": 0.2818, + "num_input_tokens_seen": 6584688, + "step": 24610 + }, + { + "epoch": 6.4521625163827, + "grad_norm": 2.0350093841552734, + "learning_rate": 1.6845117394630223e-05, + "loss": 0.4026, + "num_input_tokens_seen": 6586112, + "step": 24615 + }, + { + "epoch": 6.453473132372215, + "grad_norm": 2.3788704872131348, + "learning_rate": 1.683430655819211e-05, + "loss": 0.4317, + "num_input_tokens_seen": 6587664, + "step": 24620 + }, + { + "epoch": 6.45478374836173, + "grad_norm": 1.1029149293899536, + "learning_rate": 1.6823497430819518e-05, + "loss": 0.5164, + "num_input_tokens_seen": 6589216, + "step": 24625 + }, + { + "epoch": 6.456094364351245, + "grad_norm": 4.049336910247803, + "learning_rate": 1.681269001477479e-05, + "loss": 0.4363, + "num_input_tokens_seen": 6590544, + "step": 24630 + }, + { + "epoch": 6.45740498034076, + "grad_norm": 0.964241623878479, + "learning_rate": 1.6801884312319895e-05, + "loss": 0.4585, + "num_input_tokens_seen": 6592336, + "step": 24635 + }, + { + "epoch": 6.458715596330276, + "grad_norm": 3.3048784732818604, + "learning_rate": 1.6791080325716445e-05, + "loss": 0.4788, + "num_input_tokens_seen": 6593488, + "step": 24640 + }, + { + "epoch": 6.46002621231979, + "grad_norm": 2.745117425918579, + "learning_rate": 1.6780278057225697e-05, + "loss": 0.4973, + "num_input_tokens_seen": 6594912, + "step": 24645 + }, + { + "epoch": 6.461336828309306, + "grad_norm": 1.3796725273132324, + "learning_rate": 1.6769477509108556e-05, + "loss": 0.5713, + "num_input_tokens_seen": 6596032, + "step": 24650 + }, + { + "epoch": 6.46264744429882, + "grad_norm": 2.9856154918670654, + "learning_rate": 1.675867868362555e-05, + "loss": 0.3461, + "num_input_tokens_seen": 6597392, + "step": 24655 + }, + { + "epoch": 6.463958060288336, + "grad_norm": 4.295116901397705, + "learning_rate": 1.6747881583036858e-05, + "loss": 0.4434, + "num_input_tokens_seen": 6598592, + "step": 24660 + }, + { + "epoch": 6.465268676277851, + "grad_norm": 3.287236213684082, + "learning_rate": 1.673708620960229e-05, + "loss": 0.535, + "num_input_tokens_seen": 6600192, + "step": 24665 + }, + { + "epoch": 6.466579292267365, + "grad_norm": 11.71035385131836, + "learning_rate": 1.6726292565581298e-05, + "loss": 0.4916, + "num_input_tokens_seen": 6601552, + "step": 24670 + }, + { + "epoch": 6.467889908256881, + "grad_norm": 1.393242597579956, + "learning_rate": 1.6715500653232974e-05, + "loss": 0.4894, + "num_input_tokens_seen": 6603008, + "step": 24675 + }, + { + "epoch": 6.4692005242463955, + "grad_norm": 2.2782132625579834, + "learning_rate": 1.670471047481606e-05, + "loss": 0.476, + "num_input_tokens_seen": 6604368, + "step": 24680 + }, + { + "epoch": 6.470511140235911, + "grad_norm": 1.277086615562439, + "learning_rate": 1.66939220325889e-05, + "loss": 0.5717, + "num_input_tokens_seen": 6605776, + "step": 24685 + }, + { + "epoch": 6.471821756225426, + "grad_norm": 2.143329381942749, + "learning_rate": 1.6683135328809512e-05, + "loss": 0.4118, + "num_input_tokens_seen": 6607072, + "step": 24690 + }, + { + "epoch": 6.473132372214941, + "grad_norm": 3.174471139907837, + "learning_rate": 1.6672350365735528e-05, + "loss": 0.4326, + "num_input_tokens_seen": 6608208, + "step": 24695 + }, + { + "epoch": 6.474442988204456, + "grad_norm": 1.6982004642486572, + "learning_rate": 1.6661567145624215e-05, + "loss": 0.4503, + "num_input_tokens_seen": 6609680, + "step": 24700 + }, + { + "epoch": 6.475753604193971, + "grad_norm": 7.935152530670166, + "learning_rate": 1.6650785670732498e-05, + "loss": 0.4251, + "num_input_tokens_seen": 6611168, + "step": 24705 + }, + { + "epoch": 6.477064220183486, + "grad_norm": 2.0621330738067627, + "learning_rate": 1.6640005943316914e-05, + "loss": 0.4425, + "num_input_tokens_seen": 6612880, + "step": 24710 + }, + { + "epoch": 6.478374836173002, + "grad_norm": 3.2722232341766357, + "learning_rate": 1.6629227965633644e-05, + "loss": 0.5773, + "num_input_tokens_seen": 6614048, + "step": 24715 + }, + { + "epoch": 6.479685452162516, + "grad_norm": 1.8373576402664185, + "learning_rate": 1.6618451739938495e-05, + "loss": 0.4187, + "num_input_tokens_seen": 6615408, + "step": 24720 + }, + { + "epoch": 6.480996068152032, + "grad_norm": 1.842573881149292, + "learning_rate": 1.6607677268486914e-05, + "loss": 0.7408, + "num_input_tokens_seen": 6616384, + "step": 24725 + }, + { + "epoch": 6.482306684141546, + "grad_norm": 1.9698364734649658, + "learning_rate": 1.659690455353398e-05, + "loss": 0.5468, + "num_input_tokens_seen": 6618240, + "step": 24730 + }, + { + "epoch": 6.483617300131062, + "grad_norm": 1.5459396839141846, + "learning_rate": 1.65861335973344e-05, + "loss": 0.556, + "num_input_tokens_seen": 6619936, + "step": 24735 + }, + { + "epoch": 6.484927916120577, + "grad_norm": 2.952272653579712, + "learning_rate": 1.6575364402142528e-05, + "loss": 0.4917, + "num_input_tokens_seen": 6621568, + "step": 24740 + }, + { + "epoch": 6.486238532110092, + "grad_norm": 2.6472818851470947, + "learning_rate": 1.6564596970212327e-05, + "loss": 0.5469, + "num_input_tokens_seen": 6622944, + "step": 24745 + }, + { + "epoch": 6.487549148099607, + "grad_norm": 9.691254615783691, + "learning_rate": 1.6553831303797405e-05, + "loss": 0.5527, + "num_input_tokens_seen": 6623696, + "step": 24750 + }, + { + "epoch": 6.488859764089122, + "grad_norm": 2.1982481479644775, + "learning_rate": 1.6543067405150994e-05, + "loss": 0.9283, + "num_input_tokens_seen": 6624640, + "step": 24755 + }, + { + "epoch": 6.490170380078637, + "grad_norm": 6.748383522033691, + "learning_rate": 1.653230527652596e-05, + "loss": 0.4725, + "num_input_tokens_seen": 6625552, + "step": 24760 + }, + { + "epoch": 6.491480996068152, + "grad_norm": 1.677208423614502, + "learning_rate": 1.6521544920174803e-05, + "loss": 0.3208, + "num_input_tokens_seen": 6626688, + "step": 24765 + }, + { + "epoch": 6.492791612057667, + "grad_norm": 4.271915912628174, + "learning_rate": 1.6510786338349645e-05, + "loss": 0.6869, + "num_input_tokens_seen": 6627648, + "step": 24770 + }, + { + "epoch": 6.494102228047182, + "grad_norm": 1.816726565361023, + "learning_rate": 1.6500029533302247e-05, + "loss": 0.3252, + "num_input_tokens_seen": 6629280, + "step": 24775 + }, + { + "epoch": 6.495412844036697, + "grad_norm": 2.5348873138427734, + "learning_rate": 1.648927450728397e-05, + "loss": 0.6458, + "num_input_tokens_seen": 6630336, + "step": 24780 + }, + { + "epoch": 6.496723460026212, + "grad_norm": 5.215158939361572, + "learning_rate": 1.6478521262545837e-05, + "loss": 0.3923, + "num_input_tokens_seen": 6631424, + "step": 24785 + }, + { + "epoch": 6.498034076015728, + "grad_norm": 4.065864562988281, + "learning_rate": 1.6467769801338478e-05, + "loss": 0.4753, + "num_input_tokens_seen": 6633232, + "step": 24790 + }, + { + "epoch": 6.499344692005242, + "grad_norm": 4.674156188964844, + "learning_rate": 1.645702012591216e-05, + "loss": 0.4164, + "num_input_tokens_seen": 6634320, + "step": 24795 + }, + { + "epoch": 6.500655307994758, + "grad_norm": 1.7277292013168335, + "learning_rate": 1.644627223851677e-05, + "loss": 0.4049, + "num_input_tokens_seen": 6635712, + "step": 24800 + }, + { + "epoch": 6.50170380078637, + "eval_loss": 0.46464020013809204, + "eval_runtime": 18.2082, + "eval_samples_per_second": 46.572, + "eval_steps_per_second": 23.286, + "num_input_tokens_seen": 6636608, + "step": 24804 + }, + { + "epoch": 6.501965923984272, + "grad_norm": 2.2343618869781494, + "learning_rate": 1.643552614140182e-05, + "loss": 0.5338, + "num_input_tokens_seen": 6636928, + "step": 24805 + }, + { + "epoch": 6.503276539973788, + "grad_norm": 2.657496452331543, + "learning_rate": 1.6424781836816443e-05, + "loss": 0.4934, + "num_input_tokens_seen": 6638128, + "step": 24810 + }, + { + "epoch": 6.504587155963303, + "grad_norm": 1.038689374923706, + "learning_rate": 1.641403932700941e-05, + "loss": 0.4551, + "num_input_tokens_seen": 6640016, + "step": 24815 + }, + { + "epoch": 6.505897771952818, + "grad_norm": 1.3459970951080322, + "learning_rate": 1.6403298614229117e-05, + "loss": 0.556, + "num_input_tokens_seen": 6641328, + "step": 24820 + }, + { + "epoch": 6.507208387942333, + "grad_norm": 7.514122486114502, + "learning_rate": 1.639255970072357e-05, + "loss": 0.6258, + "num_input_tokens_seen": 6642448, + "step": 24825 + }, + { + "epoch": 6.508519003931848, + "grad_norm": 5.818509101867676, + "learning_rate": 1.6381822588740404e-05, + "loss": 0.4548, + "num_input_tokens_seen": 6643472, + "step": 24830 + }, + { + "epoch": 6.509829619921363, + "grad_norm": 2.988475799560547, + "learning_rate": 1.6371087280526888e-05, + "loss": 0.4737, + "num_input_tokens_seen": 6644704, + "step": 24835 + }, + { + "epoch": 6.511140235910878, + "grad_norm": 1.6362345218658447, + "learning_rate": 1.6360353778329883e-05, + "loss": 0.4316, + "num_input_tokens_seen": 6646496, + "step": 24840 + }, + { + "epoch": 6.512450851900393, + "grad_norm": 3.511000156402588, + "learning_rate": 1.6349622084395903e-05, + "loss": 0.4397, + "num_input_tokens_seen": 6647616, + "step": 24845 + }, + { + "epoch": 6.513761467889909, + "grad_norm": 7.358226776123047, + "learning_rate": 1.6338892200971078e-05, + "loss": 0.4971, + "num_input_tokens_seen": 6648592, + "step": 24850 + }, + { + "epoch": 6.515072083879423, + "grad_norm": 1.5750823020935059, + "learning_rate": 1.6328164130301157e-05, + "loss": 0.7566, + "num_input_tokens_seen": 6650048, + "step": 24855 + }, + { + "epoch": 6.516382699868938, + "grad_norm": 2.224167823791504, + "learning_rate": 1.631743787463149e-05, + "loss": 0.4082, + "num_input_tokens_seen": 6651264, + "step": 24860 + }, + { + "epoch": 6.517693315858454, + "grad_norm": 2.876500368118286, + "learning_rate": 1.630671343620708e-05, + "loss": 0.4932, + "num_input_tokens_seen": 6652416, + "step": 24865 + }, + { + "epoch": 6.519003931847968, + "grad_norm": 11.684536933898926, + "learning_rate": 1.6295990817272516e-05, + "loss": 0.4535, + "num_input_tokens_seen": 6653568, + "step": 24870 + }, + { + "epoch": 6.520314547837484, + "grad_norm": 2.3204498291015625, + "learning_rate": 1.6285270020072044e-05, + "loss": 0.5963, + "num_input_tokens_seen": 6654960, + "step": 24875 + }, + { + "epoch": 6.521625163826998, + "grad_norm": 2.5026817321777344, + "learning_rate": 1.6274551046849495e-05, + "loss": 0.3364, + "num_input_tokens_seen": 6656368, + "step": 24880 + }, + { + "epoch": 6.522935779816514, + "grad_norm": 2.7011942863464355, + "learning_rate": 1.6263833899848337e-05, + "loss": 0.6654, + "num_input_tokens_seen": 6657680, + "step": 24885 + }, + { + "epoch": 6.524246395806029, + "grad_norm": 3.4574801921844482, + "learning_rate": 1.625311858131165e-05, + "loss": 0.3369, + "num_input_tokens_seen": 6658688, + "step": 24890 + }, + { + "epoch": 6.525557011795544, + "grad_norm": 1.4360239505767822, + "learning_rate": 1.6242405093482133e-05, + "loss": 0.398, + "num_input_tokens_seen": 6659968, + "step": 24895 + }, + { + "epoch": 6.526867627785059, + "grad_norm": 2.3102450370788574, + "learning_rate": 1.6231693438602097e-05, + "loss": 0.3187, + "num_input_tokens_seen": 6661136, + "step": 24900 + }, + { + "epoch": 6.528178243774574, + "grad_norm": 1.8885244131088257, + "learning_rate": 1.6220983618913472e-05, + "loss": 0.3538, + "num_input_tokens_seen": 6662176, + "step": 24905 + }, + { + "epoch": 6.529488859764089, + "grad_norm": 1.2655295133590698, + "learning_rate": 1.6210275636657805e-05, + "loss": 0.3921, + "num_input_tokens_seen": 6663616, + "step": 24910 + }, + { + "epoch": 6.5307994757536045, + "grad_norm": 3.047456979751587, + "learning_rate": 1.6199569494076265e-05, + "loss": 0.4817, + "num_input_tokens_seen": 6664880, + "step": 24915 + }, + { + "epoch": 6.532110091743119, + "grad_norm": 4.696083068847656, + "learning_rate": 1.618886519340962e-05, + "loss": 0.2487, + "num_input_tokens_seen": 6666224, + "step": 24920 + }, + { + "epoch": 6.533420707732635, + "grad_norm": 3.201982021331787, + "learning_rate": 1.6178162736898264e-05, + "loss": 0.3349, + "num_input_tokens_seen": 6667696, + "step": 24925 + }, + { + "epoch": 6.534731323722149, + "grad_norm": 3.257911443710327, + "learning_rate": 1.61674621267822e-05, + "loss": 0.4855, + "num_input_tokens_seen": 6669152, + "step": 24930 + }, + { + "epoch": 6.536041939711664, + "grad_norm": 5.826011657714844, + "learning_rate": 1.6156763365301054e-05, + "loss": 0.5366, + "num_input_tokens_seen": 6670400, + "step": 24935 + }, + { + "epoch": 6.53735255570118, + "grad_norm": 2.9334168434143066, + "learning_rate": 1.614606645469406e-05, + "loss": 0.3024, + "num_input_tokens_seen": 6671552, + "step": 24940 + }, + { + "epoch": 6.538663171690695, + "grad_norm": 3.4170279502868652, + "learning_rate": 1.6135371397200057e-05, + "loss": 0.5803, + "num_input_tokens_seen": 6672960, + "step": 24945 + }, + { + "epoch": 6.53997378768021, + "grad_norm": 1.6614866256713867, + "learning_rate": 1.6124678195057507e-05, + "loss": 0.3338, + "num_input_tokens_seen": 6674112, + "step": 24950 + }, + { + "epoch": 6.541284403669724, + "grad_norm": 1.4210761785507202, + "learning_rate": 1.6113986850504475e-05, + "loss": 0.4469, + "num_input_tokens_seen": 6676160, + "step": 24955 + }, + { + "epoch": 6.54259501965924, + "grad_norm": 1.7675191164016724, + "learning_rate": 1.6103297365778636e-05, + "loss": 0.526, + "num_input_tokens_seen": 6677296, + "step": 24960 + }, + { + "epoch": 6.543905635648755, + "grad_norm": 1.9306446313858032, + "learning_rate": 1.609260974311729e-05, + "loss": 0.2842, + "num_input_tokens_seen": 6678896, + "step": 24965 + }, + { + "epoch": 6.54521625163827, + "grad_norm": 2.5330958366394043, + "learning_rate": 1.6081923984757332e-05, + "loss": 0.4161, + "num_input_tokens_seen": 6680272, + "step": 24970 + }, + { + "epoch": 6.546526867627785, + "grad_norm": 2.5444562435150146, + "learning_rate": 1.607124009293528e-05, + "loss": 0.4415, + "num_input_tokens_seen": 6681344, + "step": 24975 + }, + { + "epoch": 6.5478374836173, + "grad_norm": 2.3476855754852295, + "learning_rate": 1.6060558069887244e-05, + "loss": 0.6916, + "num_input_tokens_seen": 6682592, + "step": 24980 + }, + { + "epoch": 6.549148099606815, + "grad_norm": 1.604485273361206, + "learning_rate": 1.6049877917848948e-05, + "loss": 0.3605, + "num_input_tokens_seen": 6683712, + "step": 24985 + }, + { + "epoch": 6.5504587155963305, + "grad_norm": 5.210792541503906, + "learning_rate": 1.6039199639055755e-05, + "loss": 0.3054, + "num_input_tokens_seen": 6684720, + "step": 24990 + }, + { + "epoch": 6.551769331585845, + "grad_norm": 2.7137176990509033, + "learning_rate": 1.602852323574259e-05, + "loss": 0.4265, + "num_input_tokens_seen": 6685904, + "step": 24995 + }, + { + "epoch": 6.553079947575361, + "grad_norm": 2.0333425998687744, + "learning_rate": 1.6017848710144006e-05, + "loss": 0.4629, + "num_input_tokens_seen": 6687056, + "step": 25000 + }, + { + "epoch": 6.554390563564875, + "grad_norm": 6.121101379394531, + "learning_rate": 1.600717606449417e-05, + "loss": 0.4204, + "num_input_tokens_seen": 6688368, + "step": 25005 + }, + { + "epoch": 6.555701179554391, + "grad_norm": 6.708584308624268, + "learning_rate": 1.5996505301026848e-05, + "loss": 0.4067, + "num_input_tokens_seen": 6689568, + "step": 25010 + }, + { + "epoch": 6.557011795543906, + "grad_norm": 3.126622438430786, + "learning_rate": 1.598583642197541e-05, + "loss": 0.5364, + "num_input_tokens_seen": 6691632, + "step": 25015 + }, + { + "epoch": 6.558322411533421, + "grad_norm": 6.394060134887695, + "learning_rate": 1.5975169429572828e-05, + "loss": 0.491, + "num_input_tokens_seen": 6692688, + "step": 25020 + }, + { + "epoch": 6.559633027522936, + "grad_norm": 2.170161724090576, + "learning_rate": 1.5964504326051693e-05, + "loss": 0.5999, + "num_input_tokens_seen": 6693728, + "step": 25025 + }, + { + "epoch": 6.56094364351245, + "grad_norm": 3.4708993434906006, + "learning_rate": 1.595384111364419e-05, + "loss": 0.4151, + "num_input_tokens_seen": 6695168, + "step": 25030 + }, + { + "epoch": 6.562254259501966, + "grad_norm": 1.9372975826263428, + "learning_rate": 1.594317979458212e-05, + "loss": 0.4807, + "num_input_tokens_seen": 6696720, + "step": 25035 + }, + { + "epoch": 6.5635648754914815, + "grad_norm": 2.458956241607666, + "learning_rate": 1.593252037109686e-05, + "loss": 0.7621, + "num_input_tokens_seen": 6697984, + "step": 25040 + }, + { + "epoch": 6.564875491480996, + "grad_norm": 1.8836945295333862, + "learning_rate": 1.5921862845419417e-05, + "loss": 0.4974, + "num_input_tokens_seen": 6699136, + "step": 25045 + }, + { + "epoch": 6.566186107470511, + "grad_norm": 4.853231906890869, + "learning_rate": 1.5911207219780398e-05, + "loss": 0.5206, + "num_input_tokens_seen": 6700384, + "step": 25050 + }, + { + "epoch": 6.567496723460026, + "grad_norm": 5.576383590698242, + "learning_rate": 1.5900553496410004e-05, + "loss": 0.3979, + "num_input_tokens_seen": 6701648, + "step": 25055 + }, + { + "epoch": 6.568807339449541, + "grad_norm": 1.1193078756332397, + "learning_rate": 1.5889901677538037e-05, + "loss": 0.4846, + "num_input_tokens_seen": 6702992, + "step": 25060 + }, + { + "epoch": 6.5701179554390565, + "grad_norm": 3.1875085830688477, + "learning_rate": 1.5879251765393914e-05, + "loss": 0.5582, + "num_input_tokens_seen": 6704032, + "step": 25065 + }, + { + "epoch": 6.571428571428571, + "grad_norm": 2.8105194568634033, + "learning_rate": 1.5868603762206637e-05, + "loss": 0.4744, + "num_input_tokens_seen": 6706176, + "step": 25070 + }, + { + "epoch": 6.572739187418087, + "grad_norm": 2.6147327423095703, + "learning_rate": 1.5857957670204816e-05, + "loss": 0.5092, + "num_input_tokens_seen": 6707520, + "step": 25075 + }, + { + "epoch": 6.574049803407601, + "grad_norm": 2.154344320297241, + "learning_rate": 1.5847313491616654e-05, + "loss": 0.4254, + "num_input_tokens_seen": 6709072, + "step": 25080 + }, + { + "epoch": 6.575360419397117, + "grad_norm": 1.184433937072754, + "learning_rate": 1.5836671228669962e-05, + "loss": 0.4261, + "num_input_tokens_seen": 6710240, + "step": 25085 + }, + { + "epoch": 6.576671035386632, + "grad_norm": 2.7953286170959473, + "learning_rate": 1.582603088359216e-05, + "loss": 0.5177, + "num_input_tokens_seen": 6711312, + "step": 25090 + }, + { + "epoch": 6.577981651376147, + "grad_norm": 3.8222591876983643, + "learning_rate": 1.581539245861023e-05, + "loss": 0.4144, + "num_input_tokens_seen": 6713136, + "step": 25095 + }, + { + "epoch": 6.579292267365662, + "grad_norm": 5.183117866516113, + "learning_rate": 1.5804755955950785e-05, + "loss": 0.5725, + "num_input_tokens_seen": 6714336, + "step": 25100 + }, + { + "epoch": 6.580602883355177, + "grad_norm": 5.522197723388672, + "learning_rate": 1.579412137784004e-05, + "loss": 0.494, + "num_input_tokens_seen": 6715824, + "step": 25105 + }, + { + "epoch": 6.581913499344692, + "grad_norm": 2.407804012298584, + "learning_rate": 1.578348872650378e-05, + "loss": 0.5026, + "num_input_tokens_seen": 6717040, + "step": 25110 + }, + { + "epoch": 6.5832241153342075, + "grad_norm": 3.9510252475738525, + "learning_rate": 1.5772858004167407e-05, + "loss": 0.5231, + "num_input_tokens_seen": 6718496, + "step": 25115 + }, + { + "epoch": 6.584534731323722, + "grad_norm": 2.2906668186187744, + "learning_rate": 1.5762229213055914e-05, + "loss": 0.2826, + "num_input_tokens_seen": 6719968, + "step": 25120 + }, + { + "epoch": 6.585845347313237, + "grad_norm": 2.3052263259887695, + "learning_rate": 1.5751602355393878e-05, + "loss": 0.4413, + "num_input_tokens_seen": 6721136, + "step": 25125 + }, + { + "epoch": 6.587155963302752, + "grad_norm": 1.991910696029663, + "learning_rate": 1.5740977433405496e-05, + "loss": 0.3645, + "num_input_tokens_seen": 6722496, + "step": 25130 + }, + { + "epoch": 6.588466579292267, + "grad_norm": 2.375887870788574, + "learning_rate": 1.5730354449314534e-05, + "loss": 0.3244, + "num_input_tokens_seen": 6723808, + "step": 25135 + }, + { + "epoch": 6.5897771952817825, + "grad_norm": 6.4174485206604, + "learning_rate": 1.5719733405344373e-05, + "loss": 0.5205, + "num_input_tokens_seen": 6724752, + "step": 25140 + }, + { + "epoch": 6.591087811271297, + "grad_norm": 3.378793716430664, + "learning_rate": 1.570911430371797e-05, + "loss": 0.3576, + "num_input_tokens_seen": 6725888, + "step": 25145 + }, + { + "epoch": 6.592398427260813, + "grad_norm": 2.869586229324341, + "learning_rate": 1.5698497146657897e-05, + "loss": 0.8052, + "num_input_tokens_seen": 6727488, + "step": 25150 + }, + { + "epoch": 6.593709043250327, + "grad_norm": 1.8426172733306885, + "learning_rate": 1.5687881936386294e-05, + "loss": 0.4232, + "num_input_tokens_seen": 6728672, + "step": 25155 + }, + { + "epoch": 6.595019659239843, + "grad_norm": 4.6450629234313965, + "learning_rate": 1.5677268675124902e-05, + "loss": 0.4587, + "num_input_tokens_seen": 6729936, + "step": 25160 + }, + { + "epoch": 6.5963302752293576, + "grad_norm": 0.7964534163475037, + "learning_rate": 1.5666657365095083e-05, + "loss": 0.4386, + "num_input_tokens_seen": 6731648, + "step": 25165 + }, + { + "epoch": 6.597640891218873, + "grad_norm": 2.632723331451416, + "learning_rate": 1.565604800851774e-05, + "loss": 0.5698, + "num_input_tokens_seen": 6734048, + "step": 25170 + }, + { + "epoch": 6.598951507208388, + "grad_norm": 5.17559289932251, + "learning_rate": 1.5645440607613413e-05, + "loss": 0.431, + "num_input_tokens_seen": 6735568, + "step": 25175 + }, + { + "epoch": 6.600262123197903, + "grad_norm": 3.5396697521209717, + "learning_rate": 1.56348351646022e-05, + "loss": 0.4636, + "num_input_tokens_seen": 6736832, + "step": 25180 + }, + { + "epoch": 6.601572739187418, + "grad_norm": 4.654937744140625, + "learning_rate": 1.5624231681703812e-05, + "loss": 0.5227, + "num_input_tokens_seen": 6738288, + "step": 25185 + }, + { + "epoch": 6.6028833551769335, + "grad_norm": 1.9874011278152466, + "learning_rate": 1.5613630161137523e-05, + "loss": 0.5196, + "num_input_tokens_seen": 6739344, + "step": 25190 + }, + { + "epoch": 6.604193971166448, + "grad_norm": 1.137853741645813, + "learning_rate": 1.5603030605122226e-05, + "loss": 0.4756, + "num_input_tokens_seen": 6740656, + "step": 25195 + }, + { + "epoch": 6.605504587155964, + "grad_norm": 3.070787191390991, + "learning_rate": 1.5592433015876383e-05, + "loss": 0.4096, + "num_input_tokens_seen": 6741744, + "step": 25200 + }, + { + "epoch": 6.606815203145478, + "grad_norm": 3.074436902999878, + "learning_rate": 1.558183739561806e-05, + "loss": 0.2199, + "num_input_tokens_seen": 6743056, + "step": 25205 + }, + { + "epoch": 6.608125819134994, + "grad_norm": 4.863202095031738, + "learning_rate": 1.55712437465649e-05, + "loss": 0.341, + "num_input_tokens_seen": 6744272, + "step": 25210 + }, + { + "epoch": 6.6094364351245085, + "grad_norm": 6.182097911834717, + "learning_rate": 1.5560652070934117e-05, + "loss": 0.4822, + "num_input_tokens_seen": 6745712, + "step": 25215 + }, + { + "epoch": 6.610747051114023, + "grad_norm": 7.707995891571045, + "learning_rate": 1.5550062370942556e-05, + "loss": 0.8047, + "num_input_tokens_seen": 6747088, + "step": 25220 + }, + { + "epoch": 6.612057667103539, + "grad_norm": 1.3567112684249878, + "learning_rate": 1.553947464880662e-05, + "loss": 0.3811, + "num_input_tokens_seen": 6748272, + "step": 25225 + }, + { + "epoch": 6.613368283093053, + "grad_norm": 3.011573314666748, + "learning_rate": 1.5528888906742285e-05, + "loss": 0.3777, + "num_input_tokens_seen": 6749312, + "step": 25230 + }, + { + "epoch": 6.614678899082569, + "grad_norm": 2.508965015411377, + "learning_rate": 1.5518305146965133e-05, + "loss": 0.685, + "num_input_tokens_seen": 6750384, + "step": 25235 + }, + { + "epoch": 6.6159895150720835, + "grad_norm": 4.410300254821777, + "learning_rate": 1.5507723371690336e-05, + "loss": 0.389, + "num_input_tokens_seen": 6751520, + "step": 25240 + }, + { + "epoch": 6.617300131061599, + "grad_norm": 3.9045610427856445, + "learning_rate": 1.5497143583132638e-05, + "loss": 0.3269, + "num_input_tokens_seen": 6752784, + "step": 25245 + }, + { + "epoch": 6.618610747051114, + "grad_norm": 8.222853660583496, + "learning_rate": 1.5486565783506367e-05, + "loss": 0.4897, + "num_input_tokens_seen": 6754256, + "step": 25250 + }, + { + "epoch": 6.619921363040629, + "grad_norm": 2.0275957584381104, + "learning_rate": 1.547598997502544e-05, + "loss": 0.6433, + "num_input_tokens_seen": 6755840, + "step": 25255 + }, + { + "epoch": 6.621231979030144, + "grad_norm": 1.958586573600769, + "learning_rate": 1.546541615990335e-05, + "loss": 0.4582, + "num_input_tokens_seen": 6756864, + "step": 25260 + }, + { + "epoch": 6.6225425950196595, + "grad_norm": 2.7957346439361572, + "learning_rate": 1.545484434035319e-05, + "loss": 0.4768, + "num_input_tokens_seen": 6758176, + "step": 25265 + }, + { + "epoch": 6.623853211009174, + "grad_norm": 3.9449267387390137, + "learning_rate": 1.5444274518587617e-05, + "loss": 0.5559, + "num_input_tokens_seen": 6759408, + "step": 25270 + }, + { + "epoch": 6.62516382699869, + "grad_norm": 2.602003335952759, + "learning_rate": 1.543370669681886e-05, + "loss": 0.4599, + "num_input_tokens_seen": 6760592, + "step": 25275 + }, + { + "epoch": 6.626474442988204, + "grad_norm": 3.4282991886138916, + "learning_rate": 1.5423140877258784e-05, + "loss": 0.4102, + "num_input_tokens_seen": 6761920, + "step": 25280 + }, + { + "epoch": 6.62778505897772, + "grad_norm": 10.068952560424805, + "learning_rate": 1.5412577062118763e-05, + "loss": 0.4175, + "num_input_tokens_seen": 6763328, + "step": 25285 + }, + { + "epoch": 6.6290956749672345, + "grad_norm": 4.639484882354736, + "learning_rate": 1.5402015253609796e-05, + "loss": 0.4192, + "num_input_tokens_seen": 6764704, + "step": 25290 + }, + { + "epoch": 6.63040629095675, + "grad_norm": 1.8020859956741333, + "learning_rate": 1.539145545394246e-05, + "loss": 0.5038, + "num_input_tokens_seen": 6766640, + "step": 25295 + }, + { + "epoch": 6.631716906946265, + "grad_norm": 4.896599292755127, + "learning_rate": 1.538089766532689e-05, + "loss": 0.655, + "num_input_tokens_seen": 6768192, + "step": 25300 + }, + { + "epoch": 6.63302752293578, + "grad_norm": 1.966156005859375, + "learning_rate": 1.537034188997282e-05, + "loss": 0.3666, + "num_input_tokens_seen": 6769664, + "step": 25305 + }, + { + "epoch": 6.634338138925295, + "grad_norm": 4.724384784698486, + "learning_rate": 1.5359788130089558e-05, + "loss": 0.3575, + "num_input_tokens_seen": 6770976, + "step": 25310 + }, + { + "epoch": 6.6356487549148095, + "grad_norm": 9.043694496154785, + "learning_rate": 1.5349236387885976e-05, + "loss": 0.6444, + "num_input_tokens_seen": 6772224, + "step": 25315 + }, + { + "epoch": 6.636959370904325, + "grad_norm": 1.292860984802246, + "learning_rate": 1.5338686665570546e-05, + "loss": 0.4075, + "num_input_tokens_seen": 6773872, + "step": 25320 + }, + { + "epoch": 6.63826998689384, + "grad_norm": 2.319762945175171, + "learning_rate": 1.5328138965351307e-05, + "loss": 0.4764, + "num_input_tokens_seen": 6775392, + "step": 25325 + }, + { + "epoch": 6.639580602883355, + "grad_norm": 12.219168663024902, + "learning_rate": 1.5317593289435855e-05, + "loss": 0.3906, + "num_input_tokens_seen": 6776768, + "step": 25330 + }, + { + "epoch": 6.64089121887287, + "grad_norm": 2.3046159744262695, + "learning_rate": 1.530704964003141e-05, + "loss": 0.4689, + "num_input_tokens_seen": 6778048, + "step": 25335 + }, + { + "epoch": 6.6422018348623855, + "grad_norm": 4.766210079193115, + "learning_rate": 1.529650801934473e-05, + "loss": 0.4816, + "num_input_tokens_seen": 6779872, + "step": 25340 + }, + { + "epoch": 6.6435124508519, + "grad_norm": 1.8364298343658447, + "learning_rate": 1.5285968429582153e-05, + "loss": 0.5189, + "num_input_tokens_seen": 6781024, + "step": 25345 + }, + { + "epoch": 6.644823066841416, + "grad_norm": 1.6064540147781372, + "learning_rate": 1.52754308729496e-05, + "loss": 0.4545, + "num_input_tokens_seen": 6782528, + "step": 25350 + }, + { + "epoch": 6.64613368283093, + "grad_norm": 2.091944456100464, + "learning_rate": 1.5264895351652555e-05, + "loss": 0.4102, + "num_input_tokens_seen": 6784144, + "step": 25355 + }, + { + "epoch": 6.647444298820446, + "grad_norm": 2.9940717220306396, + "learning_rate": 1.52543618678961e-05, + "loss": 0.5222, + "num_input_tokens_seen": 6785472, + "step": 25360 + }, + { + "epoch": 6.6487549148099605, + "grad_norm": 1.811781406402588, + "learning_rate": 1.5243830423884867e-05, + "loss": 0.3487, + "num_input_tokens_seen": 6787536, + "step": 25365 + }, + { + "epoch": 6.650065530799476, + "grad_norm": 4.717513084411621, + "learning_rate": 1.5233301021823065e-05, + "loss": 0.6983, + "num_input_tokens_seen": 6788528, + "step": 25370 + }, + { + "epoch": 6.651376146788991, + "grad_norm": 3.864806652069092, + "learning_rate": 1.5222773663914486e-05, + "loss": 0.3343, + "num_input_tokens_seen": 6789872, + "step": 25375 + }, + { + "epoch": 6.652686762778506, + "grad_norm": 2.0607516765594482, + "learning_rate": 1.5212248352362482e-05, + "loss": 0.4334, + "num_input_tokens_seen": 6791168, + "step": 25380 + }, + { + "epoch": 6.653997378768021, + "grad_norm": 3.373782157897949, + "learning_rate": 1.5201725089369991e-05, + "loss": 0.5285, + "num_input_tokens_seen": 6792400, + "step": 25385 + }, + { + "epoch": 6.655307994757536, + "grad_norm": 3.318969249725342, + "learning_rate": 1.5191203877139498e-05, + "loss": 0.438, + "num_input_tokens_seen": 6793712, + "step": 25390 + }, + { + "epoch": 6.656618610747051, + "grad_norm": 6.156524658203125, + "learning_rate": 1.5180684717873089e-05, + "loss": 0.5821, + "num_input_tokens_seen": 6795472, + "step": 25395 + }, + { + "epoch": 6.657929226736567, + "grad_norm": 3.762943983078003, + "learning_rate": 1.5170167613772412e-05, + "loss": 0.5821, + "num_input_tokens_seen": 6796912, + "step": 25400 + }, + { + "epoch": 6.659239842726081, + "grad_norm": 0.9809470772743225, + "learning_rate": 1.5159652567038663e-05, + "loss": 0.4756, + "num_input_tokens_seen": 6797984, + "step": 25405 + }, + { + "epoch": 6.660550458715596, + "grad_norm": 1.7395511865615845, + "learning_rate": 1.5149139579872629e-05, + "loss": 0.5943, + "num_input_tokens_seen": 6799104, + "step": 25410 + }, + { + "epoch": 6.6618610747051115, + "grad_norm": 4.610198497772217, + "learning_rate": 1.5138628654474656e-05, + "loss": 0.3482, + "num_input_tokens_seen": 6800656, + "step": 25415 + }, + { + "epoch": 6.663171690694626, + "grad_norm": 2.850907325744629, + "learning_rate": 1.5128119793044677e-05, + "loss": 0.5272, + "num_input_tokens_seen": 6801696, + "step": 25420 + }, + { + "epoch": 6.664482306684142, + "grad_norm": 2.0158581733703613, + "learning_rate": 1.511761299778216e-05, + "loss": 0.4562, + "num_input_tokens_seen": 6803056, + "step": 25425 + }, + { + "epoch": 6.665792922673656, + "grad_norm": 13.686270713806152, + "learning_rate": 1.5107108270886166e-05, + "loss": 0.4464, + "num_input_tokens_seen": 6804528, + "step": 25430 + }, + { + "epoch": 6.667103538663172, + "grad_norm": 2.215947389602661, + "learning_rate": 1.509660561455532e-05, + "loss": 0.3896, + "num_input_tokens_seen": 6805904, + "step": 25435 + }, + { + "epoch": 6.6684141546526865, + "grad_norm": 2.174964666366577, + "learning_rate": 1.5086105030987807e-05, + "loss": 0.4843, + "num_input_tokens_seen": 6807472, + "step": 25440 + }, + { + "epoch": 6.669724770642202, + "grad_norm": 6.607484340667725, + "learning_rate": 1.5075606522381375e-05, + "loss": 0.2847, + "num_input_tokens_seen": 6808880, + "step": 25445 + }, + { + "epoch": 6.671035386631717, + "grad_norm": 1.4045828580856323, + "learning_rate": 1.5065110090933354e-05, + "loss": 0.6483, + "num_input_tokens_seen": 6809968, + "step": 25450 + }, + { + "epoch": 6.672346002621232, + "grad_norm": 4.4416632652282715, + "learning_rate": 1.5054615738840627e-05, + "loss": 0.5613, + "num_input_tokens_seen": 6811104, + "step": 25455 + }, + { + "epoch": 6.673656618610747, + "grad_norm": 2.1711220741271973, + "learning_rate": 1.5044123468299648e-05, + "loss": 0.3265, + "num_input_tokens_seen": 6812304, + "step": 25460 + }, + { + "epoch": 6.674967234600262, + "grad_norm": 2.0451624393463135, + "learning_rate": 1.5033633281506416e-05, + "loss": 0.3882, + "num_input_tokens_seen": 6813872, + "step": 25465 + }, + { + "epoch": 6.676277850589777, + "grad_norm": 4.252297878265381, + "learning_rate": 1.5023145180656517e-05, + "loss": 0.4759, + "num_input_tokens_seen": 6815168, + "step": 25470 + }, + { + "epoch": 6.677588466579293, + "grad_norm": 2.1362764835357666, + "learning_rate": 1.5012659167945097e-05, + "loss": 0.4803, + "num_input_tokens_seen": 6816448, + "step": 25475 + }, + { + "epoch": 6.678899082568807, + "grad_norm": 1.3165714740753174, + "learning_rate": 1.5002175245566857e-05, + "loss": 0.4618, + "num_input_tokens_seen": 6818384, + "step": 25480 + }, + { + "epoch": 6.680209698558322, + "grad_norm": 9.326155662536621, + "learning_rate": 1.4991693415716062e-05, + "loss": 0.7138, + "num_input_tokens_seen": 6819568, + "step": 25485 + }, + { + "epoch": 6.6815203145478375, + "grad_norm": 1.6631617546081543, + "learning_rate": 1.498121368058654e-05, + "loss": 0.5848, + "num_input_tokens_seen": 6821248, + "step": 25490 + }, + { + "epoch": 6.682830930537353, + "grad_norm": 0.7883427143096924, + "learning_rate": 1.4970736042371685e-05, + "loss": 0.3497, + "num_input_tokens_seen": 6822896, + "step": 25495 + }, + { + "epoch": 6.684141546526868, + "grad_norm": 2.335841178894043, + "learning_rate": 1.4960260503264445e-05, + "loss": 0.4634, + "num_input_tokens_seen": 6824112, + "step": 25500 + }, + { + "epoch": 6.685452162516382, + "grad_norm": 7.433170795440674, + "learning_rate": 1.494978706545733e-05, + "loss": 0.507, + "num_input_tokens_seen": 6825328, + "step": 25505 + }, + { + "epoch": 6.686762778505898, + "grad_norm": 2.2736315727233887, + "learning_rate": 1.4939315731142423e-05, + "loss": 0.5354, + "num_input_tokens_seen": 6826512, + "step": 25510 + }, + { + "epoch": 6.6880733944954125, + "grad_norm": 3.509885311126709, + "learning_rate": 1.4928846502511353e-05, + "loss": 0.4237, + "num_input_tokens_seen": 6827728, + "step": 25515 + }, + { + "epoch": 6.689384010484928, + "grad_norm": 0.920952558517456, + "learning_rate": 1.4918379381755303e-05, + "loss": 0.478, + "num_input_tokens_seen": 6829408, + "step": 25520 + }, + { + "epoch": 6.690694626474443, + "grad_norm": 5.441708564758301, + "learning_rate": 1.4907914371065032e-05, + "loss": 0.3463, + "num_input_tokens_seen": 6830544, + "step": 25525 + }, + { + "epoch": 6.692005242463958, + "grad_norm": 0.876693844795227, + "learning_rate": 1.4897451472630844e-05, + "loss": 0.7231, + "num_input_tokens_seen": 6832160, + "step": 25530 + }, + { + "epoch": 6.693315858453473, + "grad_norm": 4.982016563415527, + "learning_rate": 1.4886990688642605e-05, + "loss": 0.515, + "num_input_tokens_seen": 6833264, + "step": 25535 + }, + { + "epoch": 6.694626474442988, + "grad_norm": 7.908217430114746, + "learning_rate": 1.487653202128975e-05, + "loss": 0.251, + "num_input_tokens_seen": 6835776, + "step": 25540 + }, + { + "epoch": 6.695937090432503, + "grad_norm": 3.1726016998291016, + "learning_rate": 1.4866075472761243e-05, + "loss": 0.4322, + "num_input_tokens_seen": 6837392, + "step": 25545 + }, + { + "epoch": 6.697247706422019, + "grad_norm": 2.0294907093048096, + "learning_rate": 1.485562104524563e-05, + "loss": 0.488, + "num_input_tokens_seen": 6838816, + "step": 25550 + }, + { + "epoch": 6.698558322411533, + "grad_norm": 1.6148020029067993, + "learning_rate": 1.4845168740931004e-05, + "loss": 0.7115, + "num_input_tokens_seen": 6840016, + "step": 25555 + }, + { + "epoch": 6.699868938401049, + "grad_norm": 1.5461727380752563, + "learning_rate": 1.4834718562005007e-05, + "loss": 0.3477, + "num_input_tokens_seen": 6841120, + "step": 25560 + }, + { + "epoch": 6.7011795543905635, + "grad_norm": 8.461490631103516, + "learning_rate": 1.4824270510654859e-05, + "loss": 0.516, + "num_input_tokens_seen": 6842304, + "step": 25565 + }, + { + "epoch": 6.702490170380079, + "grad_norm": 2.3458876609802246, + "learning_rate": 1.4813824589067309e-05, + "loss": 0.3668, + "num_input_tokens_seen": 6843328, + "step": 25570 + }, + { + "epoch": 6.703800786369594, + "grad_norm": 1.2805362939834595, + "learning_rate": 1.4803380799428679e-05, + "loss": 0.4059, + "num_input_tokens_seen": 6844832, + "step": 25575 + }, + { + "epoch": 6.705111402359108, + "grad_norm": 17.123056411743164, + "learning_rate": 1.4792939143924822e-05, + "loss": 0.5016, + "num_input_tokens_seen": 6846320, + "step": 25580 + }, + { + "epoch": 6.706422018348624, + "grad_norm": 2.0815577507019043, + "learning_rate": 1.478249962474117e-05, + "loss": 0.43, + "num_input_tokens_seen": 6847776, + "step": 25585 + }, + { + "epoch": 6.707732634338139, + "grad_norm": 2.0667881965637207, + "learning_rate": 1.4772062244062687e-05, + "loss": 0.4365, + "num_input_tokens_seen": 6849008, + "step": 25590 + }, + { + "epoch": 6.709043250327654, + "grad_norm": 2.9614884853363037, + "learning_rate": 1.4761627004073905e-05, + "loss": 0.3659, + "num_input_tokens_seen": 6850336, + "step": 25595 + }, + { + "epoch": 6.710353866317169, + "grad_norm": 4.976016044616699, + "learning_rate": 1.4751193906958907e-05, + "loss": 0.3774, + "num_input_tokens_seen": 6851696, + "step": 25600 + }, + { + "epoch": 6.711664482306684, + "grad_norm": 3.495351791381836, + "learning_rate": 1.474076295490131e-05, + "loss": 0.6072, + "num_input_tokens_seen": 6853040, + "step": 25605 + }, + { + "epoch": 6.712975098296199, + "grad_norm": 1.6102590560913086, + "learning_rate": 1.4730334150084302e-05, + "loss": 0.3037, + "num_input_tokens_seen": 6854624, + "step": 25610 + }, + { + "epoch": 6.714285714285714, + "grad_norm": 5.3785576820373535, + "learning_rate": 1.4719907494690604e-05, + "loss": 0.5494, + "num_input_tokens_seen": 6856048, + "step": 25615 + }, + { + "epoch": 6.715596330275229, + "grad_norm": 4.775707244873047, + "learning_rate": 1.470948299090251e-05, + "loss": 0.3518, + "num_input_tokens_seen": 6857152, + "step": 25620 + }, + { + "epoch": 6.716906946264745, + "grad_norm": 2.4675681591033936, + "learning_rate": 1.4699060640901845e-05, + "loss": 0.4564, + "num_input_tokens_seen": 6858448, + "step": 25625 + }, + { + "epoch": 6.718217562254259, + "grad_norm": 7.0348052978515625, + "learning_rate": 1.4688640446869988e-05, + "loss": 0.3697, + "num_input_tokens_seen": 6859392, + "step": 25630 + }, + { + "epoch": 6.719528178243775, + "grad_norm": 1.3935301303863525, + "learning_rate": 1.467822241098788e-05, + "loss": 0.4207, + "num_input_tokens_seen": 6860464, + "step": 25635 + }, + { + "epoch": 6.7208387942332894, + "grad_norm": 1.4519438743591309, + "learning_rate": 1.4667806535435974e-05, + "loss": 0.4554, + "num_input_tokens_seen": 6861888, + "step": 25640 + }, + { + "epoch": 6.722149410222805, + "grad_norm": 6.344418525695801, + "learning_rate": 1.4657392822394312e-05, + "loss": 0.5992, + "num_input_tokens_seen": 6863056, + "step": 25645 + }, + { + "epoch": 6.72346002621232, + "grad_norm": 2.0579164028167725, + "learning_rate": 1.464698127404246e-05, + "loss": 0.3194, + "num_input_tokens_seen": 6864224, + "step": 25650 + }, + { + "epoch": 6.724770642201835, + "grad_norm": 1.75296950340271, + "learning_rate": 1.4636571892559547e-05, + "loss": 0.4518, + "num_input_tokens_seen": 6865424, + "step": 25655 + }, + { + "epoch": 6.72608125819135, + "grad_norm": 2.307795763015747, + "learning_rate": 1.4626164680124221e-05, + "loss": 0.3291, + "num_input_tokens_seen": 6866272, + "step": 25660 + }, + { + "epoch": 6.727391874180865, + "grad_norm": 1.1784831285476685, + "learning_rate": 1.4615759638914705e-05, + "loss": 0.4326, + "num_input_tokens_seen": 6867616, + "step": 25665 + }, + { + "epoch": 6.72870249017038, + "grad_norm": 1.6310540437698364, + "learning_rate": 1.4605356771108753e-05, + "loss": 0.6592, + "num_input_tokens_seen": 6869008, + "step": 25670 + }, + { + "epoch": 6.730013106159895, + "grad_norm": 1.1724390983581543, + "learning_rate": 1.4594956078883665e-05, + "loss": 0.4775, + "num_input_tokens_seen": 6870704, + "step": 25675 + }, + { + "epoch": 6.73132372214941, + "grad_norm": 4.338613033294678, + "learning_rate": 1.4584557564416296e-05, + "loss": 0.48, + "num_input_tokens_seen": 6871712, + "step": 25680 + }, + { + "epoch": 6.732634338138926, + "grad_norm": 1.842241883277893, + "learning_rate": 1.457416122988303e-05, + "loss": 0.312, + "num_input_tokens_seen": 6873104, + "step": 25685 + }, + { + "epoch": 6.73394495412844, + "grad_norm": 3.3740766048431396, + "learning_rate": 1.4563767077459805e-05, + "loss": 0.4374, + "num_input_tokens_seen": 6874128, + "step": 25690 + }, + { + "epoch": 6.735255570117955, + "grad_norm": 4.514527797698975, + "learning_rate": 1.4553375109322098e-05, + "loss": 0.3307, + "num_input_tokens_seen": 6875200, + "step": 25695 + }, + { + "epoch": 6.736566186107471, + "grad_norm": 1.751186728477478, + "learning_rate": 1.4542985327644931e-05, + "loss": 0.3715, + "num_input_tokens_seen": 6876704, + "step": 25700 + }, + { + "epoch": 6.737876802096985, + "grad_norm": 2.6104354858398438, + "learning_rate": 1.4532597734602877e-05, + "loss": 0.5109, + "num_input_tokens_seen": 6877904, + "step": 25705 + }, + { + "epoch": 6.739187418086501, + "grad_norm": 1.629697322845459, + "learning_rate": 1.4522212332370019e-05, + "loss": 0.5093, + "num_input_tokens_seen": 6879232, + "step": 25710 + }, + { + "epoch": 6.740498034076015, + "grad_norm": 1.6415278911590576, + "learning_rate": 1.4511829123120018e-05, + "loss": 0.327, + "num_input_tokens_seen": 6880208, + "step": 25715 + }, + { + "epoch": 6.741808650065531, + "grad_norm": 4.238846778869629, + "learning_rate": 1.4501448109026055e-05, + "loss": 0.3339, + "num_input_tokens_seen": 6881776, + "step": 25720 + }, + { + "epoch": 6.743119266055046, + "grad_norm": 1.1236543655395508, + "learning_rate": 1.4491069292260868e-05, + "loss": 0.495, + "num_input_tokens_seen": 6883200, + "step": 25725 + }, + { + "epoch": 6.744429882044561, + "grad_norm": 1.0938338041305542, + "learning_rate": 1.448069267499671e-05, + "loss": 0.5603, + "num_input_tokens_seen": 6884672, + "step": 25730 + }, + { + "epoch": 6.745740498034076, + "grad_norm": 3.7824289798736572, + "learning_rate": 1.4470318259405407e-05, + "loss": 0.7671, + "num_input_tokens_seen": 6885792, + "step": 25735 + }, + { + "epoch": 6.747051114023591, + "grad_norm": 1.3771098852157593, + "learning_rate": 1.4459946047658306e-05, + "loss": 0.5468, + "num_input_tokens_seen": 6887152, + "step": 25740 + }, + { + "epoch": 6.748361730013106, + "grad_norm": 1.6846179962158203, + "learning_rate": 1.444957604192628e-05, + "loss": 0.5874, + "num_input_tokens_seen": 6888864, + "step": 25745 + }, + { + "epoch": 6.749672346002622, + "grad_norm": 5.3832316398620605, + "learning_rate": 1.4439208244379757e-05, + "loss": 0.5608, + "num_input_tokens_seen": 6890240, + "step": 25750 + }, + { + "epoch": 6.750982961992136, + "grad_norm": 2.353959560394287, + "learning_rate": 1.4428842657188701e-05, + "loss": 0.5167, + "num_input_tokens_seen": 6891552, + "step": 25755 + }, + { + "epoch": 6.752293577981652, + "grad_norm": 2.820366144180298, + "learning_rate": 1.441847928252261e-05, + "loss": 0.4136, + "num_input_tokens_seen": 6893056, + "step": 25760 + }, + { + "epoch": 6.753604193971166, + "grad_norm": 2.482203483581543, + "learning_rate": 1.4408118122550524e-05, + "loss": 0.6345, + "num_input_tokens_seen": 6894480, + "step": 25765 + }, + { + "epoch": 6.754914809960681, + "grad_norm": 2.0948026180267334, + "learning_rate": 1.4397759179441012e-05, + "loss": 0.5549, + "num_input_tokens_seen": 6895936, + "step": 25770 + }, + { + "epoch": 6.756225425950197, + "grad_norm": 9.037862777709961, + "learning_rate": 1.4387402455362181e-05, + "loss": 0.4827, + "num_input_tokens_seen": 6896960, + "step": 25775 + }, + { + "epoch": 6.757536041939712, + "grad_norm": 1.8976749181747437, + "learning_rate": 1.4377047952481685e-05, + "loss": 0.4953, + "num_input_tokens_seen": 6898240, + "step": 25780 + }, + { + "epoch": 6.758846657929227, + "grad_norm": 7.834442615509033, + "learning_rate": 1.4366695672966705e-05, + "loss": 0.7174, + "num_input_tokens_seen": 6899200, + "step": 25785 + }, + { + "epoch": 6.760157273918741, + "grad_norm": 4.114680290222168, + "learning_rate": 1.4356345618983927e-05, + "loss": 0.4018, + "num_input_tokens_seen": 6900432, + "step": 25790 + }, + { + "epoch": 6.761467889908257, + "grad_norm": 2.2916934490203857, + "learning_rate": 1.434599779269963e-05, + "loss": 0.7909, + "num_input_tokens_seen": 6902336, + "step": 25795 + }, + { + "epoch": 6.762778505897772, + "grad_norm": 15.553308486938477, + "learning_rate": 1.4335652196279587e-05, + "loss": 0.3413, + "num_input_tokens_seen": 6903488, + "step": 25800 + }, + { + "epoch": 6.764089121887287, + "grad_norm": 5.485769748687744, + "learning_rate": 1.4325308831889115e-05, + "loss": 0.7706, + "num_input_tokens_seen": 6904864, + "step": 25805 + }, + { + "epoch": 6.765399737876802, + "grad_norm": 1.9779337644577026, + "learning_rate": 1.4314967701693055e-05, + "loss": 0.2795, + "num_input_tokens_seen": 6906272, + "step": 25810 + }, + { + "epoch": 6.766710353866317, + "grad_norm": 4.596672534942627, + "learning_rate": 1.4304628807855797e-05, + "loss": 0.4392, + "num_input_tokens_seen": 6907360, + "step": 25815 + }, + { + "epoch": 6.768020969855832, + "grad_norm": 1.7503243684768677, + "learning_rate": 1.4294292152541258e-05, + "loss": 0.2749, + "num_input_tokens_seen": 6908720, + "step": 25820 + }, + { + "epoch": 6.7693315858453476, + "grad_norm": 2.692340135574341, + "learning_rate": 1.4283957737912862e-05, + "loss": 0.43, + "num_input_tokens_seen": 6910224, + "step": 25825 + }, + { + "epoch": 6.770642201834862, + "grad_norm": 2.399967908859253, + "learning_rate": 1.4273625566133596e-05, + "loss": 0.6369, + "num_input_tokens_seen": 6911440, + "step": 25830 + }, + { + "epoch": 6.771952817824378, + "grad_norm": 4.352700233459473, + "learning_rate": 1.426329563936597e-05, + "loss": 0.445, + "num_input_tokens_seen": 6912816, + "step": 25835 + }, + { + "epoch": 6.773263433813892, + "grad_norm": 3.2094295024871826, + "learning_rate": 1.4252967959772013e-05, + "loss": 0.6142, + "num_input_tokens_seen": 6914528, + "step": 25840 + }, + { + "epoch": 6.774574049803408, + "grad_norm": 9.544609069824219, + "learning_rate": 1.4242642529513278e-05, + "loss": 0.5226, + "num_input_tokens_seen": 6915664, + "step": 25845 + }, + { + "epoch": 6.775884665792923, + "grad_norm": 3.1508545875549316, + "learning_rate": 1.4232319350750888e-05, + "loss": 0.5469, + "num_input_tokens_seen": 6916944, + "step": 25850 + }, + { + "epoch": 6.777195281782438, + "grad_norm": 2.6378366947174072, + "learning_rate": 1.422199842564545e-05, + "loss": 0.4372, + "num_input_tokens_seen": 6918560, + "step": 25855 + }, + { + "epoch": 6.778505897771953, + "grad_norm": 3.787736415863037, + "learning_rate": 1.4211679756357132e-05, + "loss": 0.3572, + "num_input_tokens_seen": 6919936, + "step": 25860 + }, + { + "epoch": 6.779816513761467, + "grad_norm": 2.5661723613739014, + "learning_rate": 1.4201363345045588e-05, + "loss": 0.2304, + "num_input_tokens_seen": 6921152, + "step": 25865 + }, + { + "epoch": 6.781127129750983, + "grad_norm": 4.938706874847412, + "learning_rate": 1.4191049193870037e-05, + "loss": 0.5098, + "num_input_tokens_seen": 6922544, + "step": 25870 + }, + { + "epoch": 6.7824377457404985, + "grad_norm": 4.002720832824707, + "learning_rate": 1.4180737304989216e-05, + "loss": 0.5578, + "num_input_tokens_seen": 6924272, + "step": 25875 + }, + { + "epoch": 6.783748361730013, + "grad_norm": 2.9650633335113525, + "learning_rate": 1.4170427680561382e-05, + "loss": 0.3336, + "num_input_tokens_seen": 6925424, + "step": 25880 + }, + { + "epoch": 6.785058977719528, + "grad_norm": 6.482110977172852, + "learning_rate": 1.4160120322744322e-05, + "loss": 0.3778, + "num_input_tokens_seen": 6926576, + "step": 25885 + }, + { + "epoch": 6.786369593709043, + "grad_norm": 11.379250526428223, + "learning_rate": 1.4149815233695346e-05, + "loss": 0.4862, + "num_input_tokens_seen": 6927536, + "step": 25890 + }, + { + "epoch": 6.787680209698558, + "grad_norm": 3.208271026611328, + "learning_rate": 1.4139512415571298e-05, + "loss": 0.5138, + "num_input_tokens_seen": 6929008, + "step": 25895 + }, + { + "epoch": 6.7889908256880735, + "grad_norm": 2.232133388519287, + "learning_rate": 1.4129211870528534e-05, + "loss": 0.4098, + "num_input_tokens_seen": 6930080, + "step": 25900 + }, + { + "epoch": 6.790301441677588, + "grad_norm": 1.5190527439117432, + "learning_rate": 1.411891360072294e-05, + "loss": 0.3526, + "num_input_tokens_seen": 6931472, + "step": 25905 + }, + { + "epoch": 6.791612057667104, + "grad_norm": 6.465576648712158, + "learning_rate": 1.4108617608309932e-05, + "loss": 0.4997, + "num_input_tokens_seen": 6932816, + "step": 25910 + }, + { + "epoch": 6.792922673656618, + "grad_norm": 9.090031623840332, + "learning_rate": 1.4098323895444437e-05, + "loss": 0.4687, + "num_input_tokens_seen": 6933792, + "step": 25915 + }, + { + "epoch": 6.794233289646134, + "grad_norm": 8.809243202209473, + "learning_rate": 1.4088032464280914e-05, + "loss": 0.2709, + "num_input_tokens_seen": 6934944, + "step": 25920 + }, + { + "epoch": 6.795543905635649, + "grad_norm": 2.375333309173584, + "learning_rate": 1.4077743316973346e-05, + "loss": 0.376, + "num_input_tokens_seen": 6936416, + "step": 25925 + }, + { + "epoch": 6.796854521625164, + "grad_norm": 4.008575916290283, + "learning_rate": 1.4067456455675226e-05, + "loss": 0.6265, + "num_input_tokens_seen": 6937824, + "step": 25930 + }, + { + "epoch": 6.798165137614679, + "grad_norm": 1.2827376127243042, + "learning_rate": 1.4057171882539583e-05, + "loss": 0.4702, + "num_input_tokens_seen": 6939344, + "step": 25935 + }, + { + "epoch": 6.799475753604194, + "grad_norm": 2.8602826595306396, + "learning_rate": 1.4046889599718962e-05, + "loss": 0.9396, + "num_input_tokens_seen": 6940592, + "step": 25940 + }, + { + "epoch": 6.800786369593709, + "grad_norm": 2.0785446166992188, + "learning_rate": 1.4036609609365414e-05, + "loss": 0.4706, + "num_input_tokens_seen": 6942032, + "step": 25945 + }, + { + "epoch": 6.8020969855832245, + "grad_norm": 1.924432396888733, + "learning_rate": 1.4026331913630536e-05, + "loss": 0.3706, + "num_input_tokens_seen": 6943360, + "step": 25950 + }, + { + "epoch": 6.803407601572739, + "grad_norm": 2.958191156387329, + "learning_rate": 1.401605651466542e-05, + "loss": 0.397, + "num_input_tokens_seen": 6944944, + "step": 25955 + }, + { + "epoch": 6.804718217562254, + "grad_norm": 2.0153205394744873, + "learning_rate": 1.4005783414620694e-05, + "loss": 0.3487, + "num_input_tokens_seen": 6946480, + "step": 25960 + }, + { + "epoch": 6.806028833551769, + "grad_norm": 13.919561386108398, + "learning_rate": 1.3995512615646509e-05, + "loss": 0.5369, + "num_input_tokens_seen": 6948320, + "step": 25965 + }, + { + "epoch": 6.807339449541285, + "grad_norm": 3.1908981800079346, + "learning_rate": 1.3985244119892518e-05, + "loss": 0.3089, + "num_input_tokens_seen": 6949520, + "step": 25970 + }, + { + "epoch": 6.8086500655307995, + "grad_norm": 2.99393367767334, + "learning_rate": 1.3974977929507913e-05, + "loss": 0.4316, + "num_input_tokens_seen": 6951008, + "step": 25975 + }, + { + "epoch": 6.809960681520314, + "grad_norm": 4.840995788574219, + "learning_rate": 1.3964714046641366e-05, + "loss": 0.6154, + "num_input_tokens_seen": 6952144, + "step": 25980 + }, + { + "epoch": 6.81127129750983, + "grad_norm": 8.995997428894043, + "learning_rate": 1.3954452473441104e-05, + "loss": 1.0026, + "num_input_tokens_seen": 6952960, + "step": 25985 + }, + { + "epoch": 6.812581913499344, + "grad_norm": 2.374066114425659, + "learning_rate": 1.3944193212054851e-05, + "loss": 0.4825, + "num_input_tokens_seen": 6954368, + "step": 25990 + }, + { + "epoch": 6.81389252948886, + "grad_norm": 4.226680755615234, + "learning_rate": 1.393393626462986e-05, + "loss": 0.4232, + "num_input_tokens_seen": 6955648, + "step": 25995 + }, + { + "epoch": 6.815203145478375, + "grad_norm": 3.8108747005462646, + "learning_rate": 1.3923681633312885e-05, + "loss": 0.593, + "num_input_tokens_seen": 6956688, + "step": 26000 + }, + { + "epoch": 6.81651376146789, + "grad_norm": 3.8976831436157227, + "learning_rate": 1.3913429320250203e-05, + "loss": 0.5782, + "num_input_tokens_seen": 6958192, + "step": 26005 + }, + { + "epoch": 6.817824377457405, + "grad_norm": 1.5724998712539673, + "learning_rate": 1.3903179327587612e-05, + "loss": 0.5623, + "num_input_tokens_seen": 6960096, + "step": 26010 + }, + { + "epoch": 6.81913499344692, + "grad_norm": 2.0926692485809326, + "learning_rate": 1.3892931657470415e-05, + "loss": 0.3526, + "num_input_tokens_seen": 6961296, + "step": 26015 + }, + { + "epoch": 6.820445609436435, + "grad_norm": 1.208078145980835, + "learning_rate": 1.3882686312043428e-05, + "loss": 0.3864, + "num_input_tokens_seen": 6962672, + "step": 26020 + }, + { + "epoch": 6.8217562254259505, + "grad_norm": 1.8006623983383179, + "learning_rate": 1.3872443293450987e-05, + "loss": 0.4073, + "num_input_tokens_seen": 6964144, + "step": 26025 + }, + { + "epoch": 6.823066841415465, + "grad_norm": 3.2602405548095703, + "learning_rate": 1.3862202603836938e-05, + "loss": 0.3256, + "num_input_tokens_seen": 6965312, + "step": 26030 + }, + { + "epoch": 6.824377457404981, + "grad_norm": 1.0100961923599243, + "learning_rate": 1.385196424534464e-05, + "loss": 0.6936, + "num_input_tokens_seen": 6967312, + "step": 26035 + }, + { + "epoch": 6.825688073394495, + "grad_norm": 2.4809255599975586, + "learning_rate": 1.3841728220116961e-05, + "loss": 0.593, + "num_input_tokens_seen": 6968928, + "step": 26040 + }, + { + "epoch": 6.826998689384011, + "grad_norm": 2.904125213623047, + "learning_rate": 1.3831494530296291e-05, + "loss": 0.4375, + "num_input_tokens_seen": 6969968, + "step": 26045 + }, + { + "epoch": 6.8283093053735255, + "grad_norm": 15.969979286193848, + "learning_rate": 1.3821263178024512e-05, + "loss": 0.4576, + "num_input_tokens_seen": 6971344, + "step": 26050 + }, + { + "epoch": 6.82961992136304, + "grad_norm": 2.311769723892212, + "learning_rate": 1.3811034165443038e-05, + "loss": 0.6515, + "num_input_tokens_seen": 6972480, + "step": 26055 + }, + { + "epoch": 6.830930537352556, + "grad_norm": 2.762019395828247, + "learning_rate": 1.3800807494692785e-05, + "loss": 0.3592, + "num_input_tokens_seen": 6973744, + "step": 26060 + }, + { + "epoch": 6.832241153342071, + "grad_norm": 27.269508361816406, + "learning_rate": 1.3790583167914168e-05, + "loss": 0.571, + "num_input_tokens_seen": 6974800, + "step": 26065 + }, + { + "epoch": 6.833551769331586, + "grad_norm": 2.8648228645324707, + "learning_rate": 1.3780361187247125e-05, + "loss": 0.3087, + "num_input_tokens_seen": 6976544, + "step": 26070 + }, + { + "epoch": 6.834862385321101, + "grad_norm": 2.8331403732299805, + "learning_rate": 1.377014155483109e-05, + "loss": 0.4085, + "num_input_tokens_seen": 6977968, + "step": 26075 + }, + { + "epoch": 6.836173001310616, + "grad_norm": 3.1464622020721436, + "learning_rate": 1.375992427280503e-05, + "loss": 0.5029, + "num_input_tokens_seen": 6979024, + "step": 26080 + }, + { + "epoch": 6.837483617300131, + "grad_norm": 1.9218237400054932, + "learning_rate": 1.3749709343307401e-05, + "loss": 0.4748, + "num_input_tokens_seen": 6980352, + "step": 26085 + }, + { + "epoch": 6.838794233289646, + "grad_norm": 2.783071517944336, + "learning_rate": 1.3739496768476168e-05, + "loss": 0.3208, + "num_input_tokens_seen": 6981424, + "step": 26090 + }, + { + "epoch": 6.840104849279161, + "grad_norm": 2.1458346843719482, + "learning_rate": 1.3729286550448814e-05, + "loss": 0.4113, + "num_input_tokens_seen": 6982720, + "step": 26095 + }, + { + "epoch": 6.8414154652686765, + "grad_norm": 2.225637912750244, + "learning_rate": 1.3719078691362302e-05, + "loss": 0.3806, + "num_input_tokens_seen": 6984544, + "step": 26100 + }, + { + "epoch": 6.842726081258191, + "grad_norm": 2.1181163787841797, + "learning_rate": 1.3708873193353125e-05, + "loss": 0.3348, + "num_input_tokens_seen": 6985744, + "step": 26105 + }, + { + "epoch": 6.844036697247707, + "grad_norm": 2.6374104022979736, + "learning_rate": 1.3698670058557281e-05, + "loss": 0.5458, + "num_input_tokens_seen": 6987088, + "step": 26110 + }, + { + "epoch": 6.845347313237221, + "grad_norm": 1.522915244102478, + "learning_rate": 1.3688469289110268e-05, + "loss": 0.4131, + "num_input_tokens_seen": 6988368, + "step": 26115 + }, + { + "epoch": 6.846657929226737, + "grad_norm": 3.094176769256592, + "learning_rate": 1.3678270887147085e-05, + "loss": 0.4936, + "num_input_tokens_seen": 6989536, + "step": 26120 + }, + { + "epoch": 6.8479685452162515, + "grad_norm": 2.206068754196167, + "learning_rate": 1.3668074854802243e-05, + "loss": 0.692, + "num_input_tokens_seen": 6990816, + "step": 26125 + }, + { + "epoch": 6.849279161205767, + "grad_norm": 2.2424376010894775, + "learning_rate": 1.3657881194209754e-05, + "loss": 0.541, + "num_input_tokens_seen": 6992272, + "step": 26130 + }, + { + "epoch": 6.850589777195282, + "grad_norm": 2.134683847427368, + "learning_rate": 1.3647689907503136e-05, + "loss": 0.4397, + "num_input_tokens_seen": 6993536, + "step": 26135 + }, + { + "epoch": 6.851900393184797, + "grad_norm": 2.3871567249298096, + "learning_rate": 1.3637500996815402e-05, + "loss": 0.4132, + "num_input_tokens_seen": 6994800, + "step": 26140 + }, + { + "epoch": 6.853211009174312, + "grad_norm": 1.806617259979248, + "learning_rate": 1.3627314464279078e-05, + "loss": 0.3652, + "num_input_tokens_seen": 6995872, + "step": 26145 + }, + { + "epoch": 6.854521625163827, + "grad_norm": 2.0363242626190186, + "learning_rate": 1.3617130312026188e-05, + "loss": 0.4094, + "num_input_tokens_seen": 6997232, + "step": 26150 + }, + { + "epoch": 6.855832241153342, + "grad_norm": 5.4619526863098145, + "learning_rate": 1.3606948542188252e-05, + "loss": 0.4635, + "num_input_tokens_seen": 6998400, + "step": 26155 + }, + { + "epoch": 6.857142857142857, + "grad_norm": 1.8566848039627075, + "learning_rate": 1.3596769156896303e-05, + "loss": 0.4939, + "num_input_tokens_seen": 6999648, + "step": 26160 + }, + { + "epoch": 6.858453473132372, + "grad_norm": 1.4520189762115479, + "learning_rate": 1.3586592158280864e-05, + "loss": 0.427, + "num_input_tokens_seen": 7001200, + "step": 26165 + }, + { + "epoch": 6.859764089121887, + "grad_norm": 2.583343505859375, + "learning_rate": 1.3576417548471967e-05, + "loss": 0.6606, + "num_input_tokens_seen": 7002592, + "step": 26170 + }, + { + "epoch": 6.8610747051114025, + "grad_norm": 5.110483169555664, + "learning_rate": 1.3566245329599148e-05, + "loss": 0.4653, + "num_input_tokens_seen": 7003664, + "step": 26175 + }, + { + "epoch": 6.862385321100917, + "grad_norm": 4.005038738250732, + "learning_rate": 1.3556075503791416e-05, + "loss": 0.4735, + "num_input_tokens_seen": 7004976, + "step": 26180 + }, + { + "epoch": 6.863695937090433, + "grad_norm": 2.0960047245025635, + "learning_rate": 1.3545908073177308e-05, + "loss": 0.4581, + "num_input_tokens_seen": 7006096, + "step": 26185 + }, + { + "epoch": 6.865006553079947, + "grad_norm": 4.653040885925293, + "learning_rate": 1.3535743039884841e-05, + "loss": 0.5762, + "num_input_tokens_seen": 7007728, + "step": 26190 + }, + { + "epoch": 6.866317169069463, + "grad_norm": 5.086483478546143, + "learning_rate": 1.3525580406041552e-05, + "loss": 0.4612, + "num_input_tokens_seen": 7008736, + "step": 26195 + }, + { + "epoch": 6.8676277850589775, + "grad_norm": 6.08259916305542, + "learning_rate": 1.3515420173774463e-05, + "loss": 0.4118, + "num_input_tokens_seen": 7009792, + "step": 26200 + }, + { + "epoch": 6.868938401048493, + "grad_norm": 3.9090237617492676, + "learning_rate": 1.3505262345210084e-05, + "loss": 0.4863, + "num_input_tokens_seen": 7011328, + "step": 26205 + }, + { + "epoch": 6.870249017038008, + "grad_norm": 2.8507280349731445, + "learning_rate": 1.3495106922474444e-05, + "loss": 0.5974, + "num_input_tokens_seen": 7012576, + "step": 26210 + }, + { + "epoch": 6.871559633027523, + "grad_norm": 2.411947250366211, + "learning_rate": 1.3484953907693038e-05, + "loss": 0.3731, + "num_input_tokens_seen": 7014000, + "step": 26215 + }, + { + "epoch": 6.872870249017038, + "grad_norm": 28.007844924926758, + "learning_rate": 1.3474803302990885e-05, + "loss": 0.3174, + "num_input_tokens_seen": 7015472, + "step": 26220 + }, + { + "epoch": 6.8741808650065535, + "grad_norm": 1.1223673820495605, + "learning_rate": 1.3464655110492484e-05, + "loss": 0.4973, + "num_input_tokens_seen": 7016864, + "step": 26225 + }, + { + "epoch": 6.875491480996068, + "grad_norm": 7.774084568023682, + "learning_rate": 1.3454509332321834e-05, + "loss": 0.4139, + "num_input_tokens_seen": 7018304, + "step": 26230 + }, + { + "epoch": 6.876802096985584, + "grad_norm": 2.871079444885254, + "learning_rate": 1.3444365970602436e-05, + "loss": 0.3678, + "num_input_tokens_seen": 7019440, + "step": 26235 + }, + { + "epoch": 6.878112712975098, + "grad_norm": 3.0095343589782715, + "learning_rate": 1.3434225027457274e-05, + "loss": 0.5254, + "num_input_tokens_seen": 7020880, + "step": 26240 + }, + { + "epoch": 6.879423328964613, + "grad_norm": 1.884472370147705, + "learning_rate": 1.3424086505008827e-05, + "loss": 0.4779, + "num_input_tokens_seen": 7022464, + "step": 26245 + }, + { + "epoch": 6.8807339449541285, + "grad_norm": 1.788521647453308, + "learning_rate": 1.3413950405379075e-05, + "loss": 0.4224, + "num_input_tokens_seen": 7023632, + "step": 26250 + }, + { + "epoch": 6.882044560943643, + "grad_norm": 1.0909208059310913, + "learning_rate": 1.3403816730689483e-05, + "loss": 0.4264, + "num_input_tokens_seen": 7025104, + "step": 26255 + }, + { + "epoch": 6.883355176933159, + "grad_norm": 5.364969730377197, + "learning_rate": 1.3393685483061014e-05, + "loss": 0.4601, + "num_input_tokens_seen": 7026272, + "step": 26260 + }, + { + "epoch": 6.884665792922673, + "grad_norm": 1.5499870777130127, + "learning_rate": 1.3383556664614117e-05, + "loss": 0.4515, + "num_input_tokens_seen": 7027984, + "step": 26265 + }, + { + "epoch": 6.885976408912189, + "grad_norm": 2.7105352878570557, + "learning_rate": 1.337343027746874e-05, + "loss": 0.488, + "num_input_tokens_seen": 7029440, + "step": 26270 + }, + { + "epoch": 6.8872870249017035, + "grad_norm": 6.104650020599365, + "learning_rate": 1.3363306323744315e-05, + "loss": 0.3758, + "num_input_tokens_seen": 7030752, + "step": 26275 + }, + { + "epoch": 6.888597640891219, + "grad_norm": 4.623847484588623, + "learning_rate": 1.335318480555977e-05, + "loss": 0.6966, + "num_input_tokens_seen": 7031904, + "step": 26280 + }, + { + "epoch": 6.889908256880734, + "grad_norm": 2.7169837951660156, + "learning_rate": 1.3343065725033522e-05, + "loss": 0.582, + "num_input_tokens_seen": 7032976, + "step": 26285 + }, + { + "epoch": 6.891218872870249, + "grad_norm": 3.7762112617492676, + "learning_rate": 1.3332949084283478e-05, + "loss": 0.5104, + "num_input_tokens_seen": 7034128, + "step": 26290 + }, + { + "epoch": 6.892529488859764, + "grad_norm": 3.7517354488372803, + "learning_rate": 1.3322834885427038e-05, + "loss": 0.4635, + "num_input_tokens_seen": 7035344, + "step": 26295 + }, + { + "epoch": 6.8938401048492794, + "grad_norm": 3.1108431816101074, + "learning_rate": 1.3312723130581073e-05, + "loss": 0.4446, + "num_input_tokens_seen": 7036544, + "step": 26300 + }, + { + "epoch": 6.895150720838794, + "grad_norm": 2.524177074432373, + "learning_rate": 1.3302613821861951e-05, + "loss": 0.5279, + "num_input_tokens_seen": 7038656, + "step": 26305 + }, + { + "epoch": 6.89646133682831, + "grad_norm": 4.392932891845703, + "learning_rate": 1.3292506961385553e-05, + "loss": 0.309, + "num_input_tokens_seen": 7040688, + "step": 26310 + }, + { + "epoch": 6.897771952817824, + "grad_norm": 1.185590386390686, + "learning_rate": 1.3282402551267223e-05, + "loss": 0.495, + "num_input_tokens_seen": 7042496, + "step": 26315 + }, + { + "epoch": 6.89908256880734, + "grad_norm": 4.1335225105285645, + "learning_rate": 1.3272300593621789e-05, + "loss": 0.3844, + "num_input_tokens_seen": 7043856, + "step": 26320 + }, + { + "epoch": 6.9003931847968545, + "grad_norm": 3.413682699203491, + "learning_rate": 1.3262201090563581e-05, + "loss": 0.3093, + "num_input_tokens_seen": 7045792, + "step": 26325 + }, + { + "epoch": 6.90170380078637, + "grad_norm": 3.540663719177246, + "learning_rate": 1.325210404420641e-05, + "loss": 0.4311, + "num_input_tokens_seen": 7047056, + "step": 26330 + }, + { + "epoch": 6.903014416775885, + "grad_norm": 3.143089532852173, + "learning_rate": 1.324200945666355e-05, + "loss": 0.4855, + "num_input_tokens_seen": 7048416, + "step": 26335 + }, + { + "epoch": 6.904325032765399, + "grad_norm": 3.3994667530059814, + "learning_rate": 1.3231917330047802e-05, + "loss": 0.5166, + "num_input_tokens_seen": 7049952, + "step": 26340 + }, + { + "epoch": 6.905635648754915, + "grad_norm": 2.083611011505127, + "learning_rate": 1.322182766647142e-05, + "loss": 0.2829, + "num_input_tokens_seen": 7051376, + "step": 26345 + }, + { + "epoch": 6.9069462647444295, + "grad_norm": 2.2097256183624268, + "learning_rate": 1.3211740468046158e-05, + "loss": 0.3563, + "num_input_tokens_seen": 7053056, + "step": 26350 + }, + { + "epoch": 6.908256880733945, + "grad_norm": 1.6168676614761353, + "learning_rate": 1.3201655736883244e-05, + "loss": 0.3508, + "num_input_tokens_seen": 7054400, + "step": 26355 + }, + { + "epoch": 6.90956749672346, + "grad_norm": 3.6321351528167725, + "learning_rate": 1.31915734750934e-05, + "loss": 0.5021, + "num_input_tokens_seen": 7055392, + "step": 26360 + }, + { + "epoch": 6.910878112712975, + "grad_norm": 2.561715602874756, + "learning_rate": 1.3181493684786825e-05, + "loss": 0.4953, + "num_input_tokens_seen": 7056368, + "step": 26365 + }, + { + "epoch": 6.91218872870249, + "grad_norm": 2.2511467933654785, + "learning_rate": 1.31714163680732e-05, + "loss": 0.6348, + "num_input_tokens_seen": 7057584, + "step": 26370 + }, + { + "epoch": 6.913499344692005, + "grad_norm": 2.800827980041504, + "learning_rate": 1.316134152706169e-05, + "loss": 0.5475, + "num_input_tokens_seen": 7058736, + "step": 26375 + }, + { + "epoch": 6.91480996068152, + "grad_norm": 4.093933582305908, + "learning_rate": 1.3151269163860947e-05, + "loss": 0.5098, + "num_input_tokens_seen": 7060112, + "step": 26380 + }, + { + "epoch": 6.916120576671036, + "grad_norm": 7.7908196449279785, + "learning_rate": 1.3141199280579092e-05, + "loss": 0.5135, + "num_input_tokens_seen": 7061648, + "step": 26385 + }, + { + "epoch": 6.91743119266055, + "grad_norm": 3.49546217918396, + "learning_rate": 1.3131131879323736e-05, + "loss": 0.4855, + "num_input_tokens_seen": 7063344, + "step": 26390 + }, + { + "epoch": 6.918741808650066, + "grad_norm": 4.371997833251953, + "learning_rate": 1.3121066962201971e-05, + "loss": 0.5586, + "num_input_tokens_seen": 7064416, + "step": 26395 + }, + { + "epoch": 6.9200524246395805, + "grad_norm": 1.571158766746521, + "learning_rate": 1.3111004531320364e-05, + "loss": 0.5346, + "num_input_tokens_seen": 7065936, + "step": 26400 + }, + { + "epoch": 6.921363040629096, + "grad_norm": 59.565467834472656, + "learning_rate": 1.3100944588784969e-05, + "loss": 0.6644, + "num_input_tokens_seen": 7067184, + "step": 26405 + }, + { + "epoch": 6.922673656618611, + "grad_norm": 0.9776583909988403, + "learning_rate": 1.3090887136701319e-05, + "loss": 0.3333, + "num_input_tokens_seen": 7068512, + "step": 26410 + }, + { + "epoch": 6.923984272608125, + "grad_norm": 2.291599988937378, + "learning_rate": 1.3080832177174406e-05, + "loss": 0.3575, + "num_input_tokens_seen": 7069728, + "step": 26415 + }, + { + "epoch": 6.925294888597641, + "grad_norm": 3.028275489807129, + "learning_rate": 1.3070779712308711e-05, + "loss": 0.3127, + "num_input_tokens_seen": 7071072, + "step": 26420 + }, + { + "epoch": 6.926605504587156, + "grad_norm": 1.996232509613037, + "learning_rate": 1.3060729744208222e-05, + "loss": 0.4231, + "num_input_tokens_seen": 7072384, + "step": 26425 + }, + { + "epoch": 6.927916120576671, + "grad_norm": 3.003638505935669, + "learning_rate": 1.3050682274976369e-05, + "loss": 0.404, + "num_input_tokens_seen": 7073840, + "step": 26430 + }, + { + "epoch": 6.929226736566186, + "grad_norm": 4.262087345123291, + "learning_rate": 1.304063730671607e-05, + "loss": 0.2608, + "num_input_tokens_seen": 7074816, + "step": 26435 + }, + { + "epoch": 6.930537352555701, + "grad_norm": 2.5129055976867676, + "learning_rate": 1.3030594841529714e-05, + "loss": 0.4559, + "num_input_tokens_seen": 7075872, + "step": 26440 + }, + { + "epoch": 6.931847968545216, + "grad_norm": 4.195748805999756, + "learning_rate": 1.3020554881519187e-05, + "loss": 0.4203, + "num_input_tokens_seen": 7076896, + "step": 26445 + }, + { + "epoch": 6.933158584534731, + "grad_norm": 3.4067466259002686, + "learning_rate": 1.3010517428785818e-05, + "loss": 0.4999, + "num_input_tokens_seen": 7078448, + "step": 26450 + }, + { + "epoch": 6.934469200524246, + "grad_norm": 5.5650458335876465, + "learning_rate": 1.3000482485430433e-05, + "loss": 0.5266, + "num_input_tokens_seen": 7080096, + "step": 26455 + }, + { + "epoch": 6.935779816513762, + "grad_norm": 2.1084625720977783, + "learning_rate": 1.299045005355333e-05, + "loss": 0.4381, + "num_input_tokens_seen": 7081664, + "step": 26460 + }, + { + "epoch": 6.937090432503276, + "grad_norm": 5.188677787780762, + "learning_rate": 1.2980420135254279e-05, + "loss": 0.7054, + "num_input_tokens_seen": 7082800, + "step": 26465 + }, + { + "epoch": 6.938401048492792, + "grad_norm": 13.57404899597168, + "learning_rate": 1.2970392732632527e-05, + "loss": 0.5416, + "num_input_tokens_seen": 7083856, + "step": 26470 + }, + { + "epoch": 6.9397116644823065, + "grad_norm": 4.759698390960693, + "learning_rate": 1.2960367847786781e-05, + "loss": 0.5284, + "num_input_tokens_seen": 7085040, + "step": 26475 + }, + { + "epoch": 6.941022280471822, + "grad_norm": 2.8184003829956055, + "learning_rate": 1.295034548281526e-05, + "loss": 0.5721, + "num_input_tokens_seen": 7085952, + "step": 26480 + }, + { + "epoch": 6.942332896461337, + "grad_norm": 6.051124095916748, + "learning_rate": 1.2940325639815599e-05, + "loss": 0.5444, + "num_input_tokens_seen": 7087024, + "step": 26485 + }, + { + "epoch": 6.943643512450852, + "grad_norm": 1.4126659631729126, + "learning_rate": 1.2930308320884946e-05, + "loss": 0.4435, + "num_input_tokens_seen": 7088448, + "step": 26490 + }, + { + "epoch": 6.944954128440367, + "grad_norm": 7.692073345184326, + "learning_rate": 1.2920293528119903e-05, + "loss": 0.5484, + "num_input_tokens_seen": 7089792, + "step": 26495 + }, + { + "epoch": 6.946264744429882, + "grad_norm": 30.270931243896484, + "learning_rate": 1.2910281263616559e-05, + "loss": 0.4612, + "num_input_tokens_seen": 7090704, + "step": 26500 + }, + { + "epoch": 6.947575360419397, + "grad_norm": 2.3547933101654053, + "learning_rate": 1.2900271529470451e-05, + "loss": 0.4307, + "num_input_tokens_seen": 7092272, + "step": 26505 + }, + { + "epoch": 6.948885976408912, + "grad_norm": 3.425410509109497, + "learning_rate": 1.2890264327776608e-05, + "loss": 0.4667, + "num_input_tokens_seen": 7093872, + "step": 26510 + }, + { + "epoch": 6.950196592398427, + "grad_norm": 2.574453353881836, + "learning_rate": 1.2880259660629516e-05, + "loss": 0.3947, + "num_input_tokens_seen": 7095024, + "step": 26515 + }, + { + "epoch": 6.951507208387943, + "grad_norm": 2.120279550552368, + "learning_rate": 1.2870257530123136e-05, + "loss": 0.41, + "num_input_tokens_seen": 7096080, + "step": 26520 + }, + { + "epoch": 6.952817824377457, + "grad_norm": 3.2184925079345703, + "learning_rate": 1.2860257938350895e-05, + "loss": 0.4136, + "num_input_tokens_seen": 7097712, + "step": 26525 + }, + { + "epoch": 6.954128440366972, + "grad_norm": 2.6699771881103516, + "learning_rate": 1.2850260887405702e-05, + "loss": 0.4269, + "num_input_tokens_seen": 7098832, + "step": 26530 + }, + { + "epoch": 6.955439056356488, + "grad_norm": 3.716053009033203, + "learning_rate": 1.2840266379379896e-05, + "loss": 0.3709, + "num_input_tokens_seen": 7100128, + "step": 26535 + }, + { + "epoch": 6.956749672346002, + "grad_norm": 4.40641450881958, + "learning_rate": 1.2830274416365334e-05, + "loss": 0.4245, + "num_input_tokens_seen": 7101152, + "step": 26540 + }, + { + "epoch": 6.958060288335518, + "grad_norm": 4.137094497680664, + "learning_rate": 1.2820285000453309e-05, + "loss": 0.3471, + "num_input_tokens_seen": 7102400, + "step": 26545 + }, + { + "epoch": 6.9593709043250325, + "grad_norm": 5.578113555908203, + "learning_rate": 1.281029813373459e-05, + "loss": 0.4992, + "num_input_tokens_seen": 7103792, + "step": 26550 + }, + { + "epoch": 6.960681520314548, + "grad_norm": 2.122875928878784, + "learning_rate": 1.2800313818299403e-05, + "loss": 0.359, + "num_input_tokens_seen": 7104944, + "step": 26555 + }, + { + "epoch": 6.961992136304063, + "grad_norm": 1.6667053699493408, + "learning_rate": 1.2790332056237458e-05, + "loss": 0.3704, + "num_input_tokens_seen": 7106144, + "step": 26560 + }, + { + "epoch": 6.963302752293578, + "grad_norm": 3.729678153991699, + "learning_rate": 1.2780352849637928e-05, + "loss": 0.618, + "num_input_tokens_seen": 7107424, + "step": 26565 + }, + { + "epoch": 6.964613368283093, + "grad_norm": 5.311074733734131, + "learning_rate": 1.2770376200589423e-05, + "loss": 0.4488, + "num_input_tokens_seen": 7108992, + "step": 26570 + }, + { + "epoch": 6.965923984272608, + "grad_norm": 2.6659293174743652, + "learning_rate": 1.2760402111180048e-05, + "loss": 0.5123, + "num_input_tokens_seen": 7110640, + "step": 26575 + }, + { + "epoch": 6.967234600262123, + "grad_norm": 7.703089714050293, + "learning_rate": 1.2750430583497363e-05, + "loss": 0.4245, + "num_input_tokens_seen": 7111968, + "step": 26580 + }, + { + "epoch": 6.968545216251639, + "grad_norm": 1.6236867904663086, + "learning_rate": 1.2740461619628392e-05, + "loss": 0.536, + "num_input_tokens_seen": 7113376, + "step": 26585 + }, + { + "epoch": 6.969855832241153, + "grad_norm": 2.233104705810547, + "learning_rate": 1.2730495221659607e-05, + "loss": 0.4949, + "num_input_tokens_seen": 7114576, + "step": 26590 + }, + { + "epoch": 6.971166448230669, + "grad_norm": 1.9903984069824219, + "learning_rate": 1.2720531391676987e-05, + "loss": 0.436, + "num_input_tokens_seen": 7115712, + "step": 26595 + }, + { + "epoch": 6.972477064220183, + "grad_norm": 1.1708776950836182, + "learning_rate": 1.2710570131765937e-05, + "loss": 0.5911, + "num_input_tokens_seen": 7117968, + "step": 26600 + }, + { + "epoch": 6.973787680209698, + "grad_norm": 2.538797616958618, + "learning_rate": 1.2700611444011312e-05, + "loss": 0.4649, + "num_input_tokens_seen": 7119184, + "step": 26605 + }, + { + "epoch": 6.975098296199214, + "grad_norm": 3.920365333557129, + "learning_rate": 1.2690655330497464e-05, + "loss": 0.6984, + "num_input_tokens_seen": 7120416, + "step": 26610 + }, + { + "epoch": 6.976408912188729, + "grad_norm": 4.843588352203369, + "learning_rate": 1.2680701793308187e-05, + "loss": 0.5154, + "num_input_tokens_seen": 7121728, + "step": 26615 + }, + { + "epoch": 6.977719528178244, + "grad_norm": 9.519847869873047, + "learning_rate": 1.2670750834526735e-05, + "loss": 0.4651, + "num_input_tokens_seen": 7122864, + "step": 26620 + }, + { + "epoch": 6.9790301441677585, + "grad_norm": 1.902405858039856, + "learning_rate": 1.266080245623583e-05, + "loss": 0.3904, + "num_input_tokens_seen": 7124608, + "step": 26625 + }, + { + "epoch": 6.980340760157274, + "grad_norm": 16.13047981262207, + "learning_rate": 1.2650856660517652e-05, + "loss": 0.469, + "num_input_tokens_seen": 7125632, + "step": 26630 + }, + { + "epoch": 6.981651376146789, + "grad_norm": 1.990519404411316, + "learning_rate": 1.2640913449453837e-05, + "loss": 0.3029, + "num_input_tokens_seen": 7126752, + "step": 26635 + }, + { + "epoch": 6.982961992136304, + "grad_norm": 3.5578742027282715, + "learning_rate": 1.2630972825125482e-05, + "loss": 0.6306, + "num_input_tokens_seen": 7128064, + "step": 26640 + }, + { + "epoch": 6.984272608125819, + "grad_norm": 8.859781265258789, + "learning_rate": 1.2621034789613151e-05, + "loss": 0.8696, + "num_input_tokens_seen": 7129392, + "step": 26645 + }, + { + "epoch": 6.985583224115334, + "grad_norm": 11.722240447998047, + "learning_rate": 1.2611099344996832e-05, + "loss": 0.4997, + "num_input_tokens_seen": 7130816, + "step": 26650 + }, + { + "epoch": 6.986893840104849, + "grad_norm": 1.2489382028579712, + "learning_rate": 1.2601166493356023e-05, + "loss": 0.5404, + "num_input_tokens_seen": 7132864, + "step": 26655 + }, + { + "epoch": 6.988204456094365, + "grad_norm": 3.874467372894287, + "learning_rate": 1.2591236236769643e-05, + "loss": 0.5432, + "num_input_tokens_seen": 7134080, + "step": 26660 + }, + { + "epoch": 6.989515072083879, + "grad_norm": 8.242684364318848, + "learning_rate": 1.2581308577316079e-05, + "loss": 0.6231, + "num_input_tokens_seen": 7135216, + "step": 26665 + }, + { + "epoch": 6.990825688073395, + "grad_norm": 39.08094024658203, + "learning_rate": 1.2571383517073172e-05, + "loss": 0.4648, + "num_input_tokens_seen": 7136624, + "step": 26670 + }, + { + "epoch": 6.992136304062909, + "grad_norm": 2.437466621398926, + "learning_rate": 1.2561461058118219e-05, + "loss": 0.4785, + "num_input_tokens_seen": 7138208, + "step": 26675 + }, + { + "epoch": 6.993446920052425, + "grad_norm": 8.457503318786621, + "learning_rate": 1.2551541202527984e-05, + "loss": 0.6991, + "num_input_tokens_seen": 7139344, + "step": 26680 + }, + { + "epoch": 6.99475753604194, + "grad_norm": 5.044585704803467, + "learning_rate": 1.2541623952378657e-05, + "loss": 0.6422, + "num_input_tokens_seen": 7140496, + "step": 26685 + }, + { + "epoch": 6.996068152031455, + "grad_norm": 1.0663235187530518, + "learning_rate": 1.2531709309745907e-05, + "loss": 0.3997, + "num_input_tokens_seen": 7142272, + "step": 26690 + }, + { + "epoch": 6.99737876802097, + "grad_norm": 3.8820981979370117, + "learning_rate": 1.2521797276704858e-05, + "loss": 0.5554, + "num_input_tokens_seen": 7143472, + "step": 26695 + }, + { + "epoch": 6.9986893840104845, + "grad_norm": 2.1360738277435303, + "learning_rate": 1.251188785533008e-05, + "loss": 0.4826, + "num_input_tokens_seen": 7144608, + "step": 26700 + }, + { + "epoch": 7.0, + "grad_norm": 6.06490421295166, + "learning_rate": 1.2501981047695583e-05, + "loss": 0.5039, + "num_input_tokens_seen": 7146368, + "step": 26705 + }, + { + "epoch": 7.001310615989515, + "grad_norm": 2.1234076023101807, + "learning_rate": 1.2492076855874868e-05, + "loss": 0.5062, + "num_input_tokens_seen": 7147568, + "step": 26710 + }, + { + "epoch": 7.001834862385321, + "eval_loss": 0.4646163880825043, + "eval_runtime": 18.2137, + "eval_samples_per_second": 46.558, + "eval_steps_per_second": 23.279, + "num_input_tokens_seen": 7148208, + "step": 26712 + }, + { + "epoch": 7.00262123197903, + "grad_norm": 2.4688720703125, + "learning_rate": 1.2482175281940855e-05, + "loss": 0.3144, + "num_input_tokens_seen": 7149424, + "step": 26715 + }, + { + "epoch": 7.003931847968545, + "grad_norm": 14.773076057434082, + "learning_rate": 1.2472276327965938e-05, + "loss": 0.6418, + "num_input_tokens_seen": 7150896, + "step": 26720 + }, + { + "epoch": 7.00524246395806, + "grad_norm": 11.983713150024414, + "learning_rate": 1.2462379996021928e-05, + "loss": 0.4079, + "num_input_tokens_seen": 7152224, + "step": 26725 + }, + { + "epoch": 7.006553079947575, + "grad_norm": 2.308548927307129, + "learning_rate": 1.2452486288180124e-05, + "loss": 0.5627, + "num_input_tokens_seen": 7153824, + "step": 26730 + }, + { + "epoch": 7.007863695937091, + "grad_norm": 1.5499259233474731, + "learning_rate": 1.2442595206511257e-05, + "loss": 0.4361, + "num_input_tokens_seen": 7155008, + "step": 26735 + }, + { + "epoch": 7.009174311926605, + "grad_norm": 2.4212839603424072, + "learning_rate": 1.2432706753085516e-05, + "loss": 0.5003, + "num_input_tokens_seen": 7156256, + "step": 26740 + }, + { + "epoch": 7.010484927916121, + "grad_norm": 2.697370767593384, + "learning_rate": 1.2422820929972537e-05, + "loss": 0.4468, + "num_input_tokens_seen": 7157760, + "step": 26745 + }, + { + "epoch": 7.011795543905635, + "grad_norm": 2.074246644973755, + "learning_rate": 1.2412937739241407e-05, + "loss": 0.3089, + "num_input_tokens_seen": 7159168, + "step": 26750 + }, + { + "epoch": 7.013106159895151, + "grad_norm": 2.3513338565826416, + "learning_rate": 1.2403057182960655e-05, + "loss": 0.4055, + "num_input_tokens_seen": 7160512, + "step": 26755 + }, + { + "epoch": 7.014416775884666, + "grad_norm": 4.135067939758301, + "learning_rate": 1.2393179263198273e-05, + "loss": 0.5295, + "num_input_tokens_seen": 7161680, + "step": 26760 + }, + { + "epoch": 7.015727391874181, + "grad_norm": 3.462540626525879, + "learning_rate": 1.2383303982021688e-05, + "loss": 0.3999, + "num_input_tokens_seen": 7162624, + "step": 26765 + }, + { + "epoch": 7.017038007863696, + "grad_norm": 2.0407326221466064, + "learning_rate": 1.237343134149778e-05, + "loss": 0.5308, + "num_input_tokens_seen": 7164288, + "step": 26770 + }, + { + "epoch": 7.018348623853211, + "grad_norm": 1.9573253393173218, + "learning_rate": 1.2363561343692873e-05, + "loss": 0.4939, + "num_input_tokens_seen": 7165648, + "step": 26775 + }, + { + "epoch": 7.019659239842726, + "grad_norm": 2.233880043029785, + "learning_rate": 1.2353693990672743e-05, + "loss": 0.4429, + "num_input_tokens_seen": 7166976, + "step": 26780 + }, + { + "epoch": 7.0209698558322415, + "grad_norm": 1.3875606060028076, + "learning_rate": 1.2343829284502612e-05, + "loss": 0.5455, + "num_input_tokens_seen": 7168416, + "step": 26785 + }, + { + "epoch": 7.022280471821756, + "grad_norm": 3.6885578632354736, + "learning_rate": 1.2333967227247143e-05, + "loss": 0.3786, + "num_input_tokens_seen": 7169488, + "step": 26790 + }, + { + "epoch": 7.023591087811272, + "grad_norm": 4.718465328216553, + "learning_rate": 1.2324107820970445e-05, + "loss": 0.2715, + "num_input_tokens_seen": 7170528, + "step": 26795 + }, + { + "epoch": 7.024901703800786, + "grad_norm": 2.6939127445220947, + "learning_rate": 1.231425106773609e-05, + "loss": 0.3973, + "num_input_tokens_seen": 7172176, + "step": 26800 + }, + { + "epoch": 7.026212319790301, + "grad_norm": 1.6823240518569946, + "learning_rate": 1.2304396969607057e-05, + "loss": 0.4322, + "num_input_tokens_seen": 7173984, + "step": 26805 + }, + { + "epoch": 7.027522935779817, + "grad_norm": 3.428297758102417, + "learning_rate": 1.2294545528645804e-05, + "loss": 0.4927, + "num_input_tokens_seen": 7175024, + "step": 26810 + }, + { + "epoch": 7.028833551769331, + "grad_norm": 4.644112586975098, + "learning_rate": 1.2284696746914216e-05, + "loss": 0.4936, + "num_input_tokens_seen": 7176304, + "step": 26815 + }, + { + "epoch": 7.030144167758847, + "grad_norm": 1.9844173192977905, + "learning_rate": 1.227485062647362e-05, + "loss": 0.4655, + "num_input_tokens_seen": 7177632, + "step": 26820 + }, + { + "epoch": 7.031454783748361, + "grad_norm": 3.117440700531006, + "learning_rate": 1.226500716938481e-05, + "loss": 0.3916, + "num_input_tokens_seen": 7179072, + "step": 26825 + }, + { + "epoch": 7.032765399737877, + "grad_norm": 2.923340320587158, + "learning_rate": 1.2255166377707995e-05, + "loss": 0.4096, + "num_input_tokens_seen": 7180336, + "step": 26830 + }, + { + "epoch": 7.034076015727392, + "grad_norm": 4.238308429718018, + "learning_rate": 1.2245328253502844e-05, + "loss": 0.4867, + "num_input_tokens_seen": 7181552, + "step": 26835 + }, + { + "epoch": 7.035386631716907, + "grad_norm": 2.6472530364990234, + "learning_rate": 1.2235492798828444e-05, + "loss": 0.4639, + "num_input_tokens_seen": 7182496, + "step": 26840 + }, + { + "epoch": 7.036697247706422, + "grad_norm": 1.6733726263046265, + "learning_rate": 1.2225660015743343e-05, + "loss": 0.4073, + "num_input_tokens_seen": 7183728, + "step": 26845 + }, + { + "epoch": 7.038007863695937, + "grad_norm": 4.104498863220215, + "learning_rate": 1.2215829906305531e-05, + "loss": 0.5093, + "num_input_tokens_seen": 7184880, + "step": 26850 + }, + { + "epoch": 7.039318479685452, + "grad_norm": 2.899144172668457, + "learning_rate": 1.2206002472572428e-05, + "loss": 0.3449, + "num_input_tokens_seen": 7186064, + "step": 26855 + }, + { + "epoch": 7.0406290956749675, + "grad_norm": 1.6914780139923096, + "learning_rate": 1.2196177716600904e-05, + "loss": 0.4742, + "num_input_tokens_seen": 7187440, + "step": 26860 + }, + { + "epoch": 7.041939711664482, + "grad_norm": 3.9383602142333984, + "learning_rate": 1.2186355640447259e-05, + "loss": 0.4897, + "num_input_tokens_seen": 7188688, + "step": 26865 + }, + { + "epoch": 7.043250327653998, + "grad_norm": 2.6294515132904053, + "learning_rate": 1.2176536246167235e-05, + "loss": 0.6042, + "num_input_tokens_seen": 7190272, + "step": 26870 + }, + { + "epoch": 7.044560943643512, + "grad_norm": 4.703229904174805, + "learning_rate": 1.2166719535816024e-05, + "loss": 0.4433, + "num_input_tokens_seen": 7191648, + "step": 26875 + }, + { + "epoch": 7.045871559633028, + "grad_norm": 2.567078113555908, + "learning_rate": 1.2156905511448235e-05, + "loss": 0.3265, + "num_input_tokens_seen": 7194304, + "step": 26880 + }, + { + "epoch": 7.047182175622543, + "grad_norm": 4.075915813446045, + "learning_rate": 1.2147094175117932e-05, + "loss": 0.4803, + "num_input_tokens_seen": 7195392, + "step": 26885 + }, + { + "epoch": 7.048492791612058, + "grad_norm": 1.6603200435638428, + "learning_rate": 1.2137285528878611e-05, + "loss": 0.5623, + "num_input_tokens_seen": 7196848, + "step": 26890 + }, + { + "epoch": 7.049803407601573, + "grad_norm": 3.8483989238739014, + "learning_rate": 1.2127479574783205e-05, + "loss": 0.474, + "num_input_tokens_seen": 7198352, + "step": 26895 + }, + { + "epoch": 7.051114023591087, + "grad_norm": 2.7747154235839844, + "learning_rate": 1.211767631488408e-05, + "loss": 0.5139, + "num_input_tokens_seen": 7199760, + "step": 26900 + }, + { + "epoch": 7.052424639580603, + "grad_norm": 2.521568536758423, + "learning_rate": 1.2107875751233045e-05, + "loss": 0.5043, + "num_input_tokens_seen": 7201440, + "step": 26905 + }, + { + "epoch": 7.053735255570118, + "grad_norm": 1.6076395511627197, + "learning_rate": 1.2098077885881336e-05, + "loss": 0.5214, + "num_input_tokens_seen": 7203200, + "step": 26910 + }, + { + "epoch": 7.055045871559633, + "grad_norm": 1.0670889616012573, + "learning_rate": 1.2088282720879646e-05, + "loss": 0.3494, + "num_input_tokens_seen": 7204656, + "step": 26915 + }, + { + "epoch": 7.056356487549148, + "grad_norm": 2.1101388931274414, + "learning_rate": 1.2078490258278061e-05, + "loss": 0.545, + "num_input_tokens_seen": 7206048, + "step": 26920 + }, + { + "epoch": 7.057667103538663, + "grad_norm": 1.578600525856018, + "learning_rate": 1.206870050012614e-05, + "loss": 0.5636, + "num_input_tokens_seen": 7207408, + "step": 26925 + }, + { + "epoch": 7.058977719528178, + "grad_norm": 1.4198557138442993, + "learning_rate": 1.205891344847286e-05, + "loss": 0.6761, + "num_input_tokens_seen": 7208784, + "step": 26930 + }, + { + "epoch": 7.0602883355176935, + "grad_norm": 1.6933176517486572, + "learning_rate": 1.2049129105366625e-05, + "loss": 0.466, + "num_input_tokens_seen": 7210224, + "step": 26935 + }, + { + "epoch": 7.061598951507208, + "grad_norm": 3.481503486633301, + "learning_rate": 1.20393474728553e-05, + "loss": 0.3404, + "num_input_tokens_seen": 7211632, + "step": 26940 + }, + { + "epoch": 7.062909567496724, + "grad_norm": 2.3746893405914307, + "learning_rate": 1.2029568552986156e-05, + "loss": 0.654, + "num_input_tokens_seen": 7212832, + "step": 26945 + }, + { + "epoch": 7.064220183486238, + "grad_norm": 19.01474952697754, + "learning_rate": 1.2019792347805902e-05, + "loss": 0.5584, + "num_input_tokens_seen": 7214080, + "step": 26950 + }, + { + "epoch": 7.065530799475754, + "grad_norm": 1.6801081895828247, + "learning_rate": 1.2010018859360691e-05, + "loss": 0.4427, + "num_input_tokens_seen": 7215600, + "step": 26955 + }, + { + "epoch": 7.066841415465269, + "grad_norm": 3.0401241779327393, + "learning_rate": 1.2000248089696078e-05, + "loss": 0.3384, + "num_input_tokens_seen": 7216768, + "step": 26960 + }, + { + "epoch": 7.068152031454784, + "grad_norm": 4.916090488433838, + "learning_rate": 1.1990480040857076e-05, + "loss": 0.4936, + "num_input_tokens_seen": 7218704, + "step": 26965 + }, + { + "epoch": 7.069462647444299, + "grad_norm": 2.744964599609375, + "learning_rate": 1.1980714714888127e-05, + "loss": 0.4321, + "num_input_tokens_seen": 7220080, + "step": 26970 + }, + { + "epoch": 7.070773263433814, + "grad_norm": 9.105704307556152, + "learning_rate": 1.197095211383309e-05, + "loss": 0.462, + "num_input_tokens_seen": 7221408, + "step": 26975 + }, + { + "epoch": 7.072083879423329, + "grad_norm": 3.9265928268432617, + "learning_rate": 1.1961192239735264e-05, + "loss": 0.4191, + "num_input_tokens_seen": 7222640, + "step": 26980 + }, + { + "epoch": 7.073394495412844, + "grad_norm": 4.1935224533081055, + "learning_rate": 1.1951435094637373e-05, + "loss": 0.4888, + "num_input_tokens_seen": 7223744, + "step": 26985 + }, + { + "epoch": 7.074705111402359, + "grad_norm": 7.911445617675781, + "learning_rate": 1.194168068058157e-05, + "loss": 0.7966, + "num_input_tokens_seen": 7225072, + "step": 26990 + }, + { + "epoch": 7.076015727391874, + "grad_norm": 2.297353982925415, + "learning_rate": 1.1931928999609438e-05, + "loss": 0.399, + "num_input_tokens_seen": 7226352, + "step": 26995 + }, + { + "epoch": 7.077326343381389, + "grad_norm": 2.889137029647827, + "learning_rate": 1.1922180053761985e-05, + "loss": 0.4453, + "num_input_tokens_seen": 7227440, + "step": 27000 + }, + { + "epoch": 7.078636959370904, + "grad_norm": 6.052199363708496, + "learning_rate": 1.1912433845079652e-05, + "loss": 0.7693, + "num_input_tokens_seen": 7228464, + "step": 27005 + }, + { + "epoch": 7.0799475753604195, + "grad_norm": 3.390277147293091, + "learning_rate": 1.19026903756023e-05, + "loss": 0.6177, + "num_input_tokens_seen": 7229968, + "step": 27010 + }, + { + "epoch": 7.081258191349934, + "grad_norm": 1.76405930519104, + "learning_rate": 1.1892949647369223e-05, + "loss": 0.3901, + "num_input_tokens_seen": 7231424, + "step": 27015 + }, + { + "epoch": 7.08256880733945, + "grad_norm": 3.0881736278533936, + "learning_rate": 1.1883211662419137e-05, + "loss": 0.3692, + "num_input_tokens_seen": 7232976, + "step": 27020 + }, + { + "epoch": 7.083879423328964, + "grad_norm": 2.9709978103637695, + "learning_rate": 1.1873476422790187e-05, + "loss": 0.4947, + "num_input_tokens_seen": 7235408, + "step": 27025 + }, + { + "epoch": 7.08519003931848, + "grad_norm": 15.065858840942383, + "learning_rate": 1.1863743930519939e-05, + "loss": 0.4264, + "num_input_tokens_seen": 7236320, + "step": 27030 + }, + { + "epoch": 7.086500655307995, + "grad_norm": 1.6109249591827393, + "learning_rate": 1.1854014187645398e-05, + "loss": 0.3151, + "num_input_tokens_seen": 7237472, + "step": 27035 + }, + { + "epoch": 7.08781127129751, + "grad_norm": 5.63012170791626, + "learning_rate": 1.1844287196202967e-05, + "loss": 0.5995, + "num_input_tokens_seen": 7238592, + "step": 27040 + }, + { + "epoch": 7.089121887287025, + "grad_norm": 6.784441947937012, + "learning_rate": 1.183456295822849e-05, + "loss": 0.4527, + "num_input_tokens_seen": 7239552, + "step": 27045 + }, + { + "epoch": 7.09043250327654, + "grad_norm": 1.6092556715011597, + "learning_rate": 1.1824841475757232e-05, + "loss": 0.3873, + "num_input_tokens_seen": 7240640, + "step": 27050 + }, + { + "epoch": 7.091743119266055, + "grad_norm": 3.549830675125122, + "learning_rate": 1.1815122750823896e-05, + "loss": 0.4834, + "num_input_tokens_seen": 7241936, + "step": 27055 + }, + { + "epoch": 7.0930537352555705, + "grad_norm": 1.2639409303665161, + "learning_rate": 1.1805406785462587e-05, + "loss": 0.4936, + "num_input_tokens_seen": 7243632, + "step": 27060 + }, + { + "epoch": 7.094364351245085, + "grad_norm": 1.9754445552825928, + "learning_rate": 1.1795693581706838e-05, + "loss": 0.6618, + "num_input_tokens_seen": 7244896, + "step": 27065 + }, + { + "epoch": 7.095674967234601, + "grad_norm": 4.596470832824707, + "learning_rate": 1.1785983141589618e-05, + "loss": 0.4805, + "num_input_tokens_seen": 7245856, + "step": 27070 + }, + { + "epoch": 7.096985583224115, + "grad_norm": 3.5257017612457275, + "learning_rate": 1.1776275467143283e-05, + "loss": 0.3603, + "num_input_tokens_seen": 7247040, + "step": 27075 + }, + { + "epoch": 7.09829619921363, + "grad_norm": 2.374562978744507, + "learning_rate": 1.1766570560399647e-05, + "loss": 0.3219, + "num_input_tokens_seen": 7248192, + "step": 27080 + }, + { + "epoch": 7.0996068152031455, + "grad_norm": 3.3565797805786133, + "learning_rate": 1.1756868423389925e-05, + "loss": 0.727, + "num_input_tokens_seen": 7249472, + "step": 27085 + }, + { + "epoch": 7.10091743119266, + "grad_norm": 3.022841215133667, + "learning_rate": 1.1747169058144761e-05, + "loss": 0.6069, + "num_input_tokens_seen": 7250864, + "step": 27090 + }, + { + "epoch": 7.102228047182176, + "grad_norm": 2.599672555923462, + "learning_rate": 1.1737472466694213e-05, + "loss": 0.3005, + "num_input_tokens_seen": 7252672, + "step": 27095 + }, + { + "epoch": 7.10353866317169, + "grad_norm": 1.6078988313674927, + "learning_rate": 1.1727778651067761e-05, + "loss": 0.4297, + "num_input_tokens_seen": 7254528, + "step": 27100 + }, + { + "epoch": 7.104849279161206, + "grad_norm": 1.1715054512023926, + "learning_rate": 1.1718087613294307e-05, + "loss": 0.3927, + "num_input_tokens_seen": 7255728, + "step": 27105 + }, + { + "epoch": 7.1061598951507206, + "grad_norm": 9.632615089416504, + "learning_rate": 1.1708399355402163e-05, + "loss": 0.344, + "num_input_tokens_seen": 7257120, + "step": 27110 + }, + { + "epoch": 7.107470511140236, + "grad_norm": 8.535378456115723, + "learning_rate": 1.1698713879419072e-05, + "loss": 0.4375, + "num_input_tokens_seen": 7258528, + "step": 27115 + }, + { + "epoch": 7.108781127129751, + "grad_norm": 2.5347414016723633, + "learning_rate": 1.168903118737218e-05, + "loss": 0.3899, + "num_input_tokens_seen": 7259616, + "step": 27120 + }, + { + "epoch": 7.110091743119266, + "grad_norm": 2.6115005016326904, + "learning_rate": 1.1679351281288062e-05, + "loss": 0.3906, + "num_input_tokens_seen": 7260864, + "step": 27125 + }, + { + "epoch": 7.111402359108781, + "grad_norm": 1.0615527629852295, + "learning_rate": 1.1669674163192703e-05, + "loss": 0.327, + "num_input_tokens_seen": 7263072, + "step": 27130 + }, + { + "epoch": 7.1127129750982965, + "grad_norm": 3.532097816467285, + "learning_rate": 1.165999983511151e-05, + "loss": 0.4494, + "num_input_tokens_seen": 7264080, + "step": 27135 + }, + { + "epoch": 7.114023591087811, + "grad_norm": 2.5425875186920166, + "learning_rate": 1.1650328299069301e-05, + "loss": 0.5508, + "num_input_tokens_seen": 7265600, + "step": 27140 + }, + { + "epoch": 7.115334207077327, + "grad_norm": 4.27174186706543, + "learning_rate": 1.1640659557090313e-05, + "loss": 0.397, + "num_input_tokens_seen": 7266656, + "step": 27145 + }, + { + "epoch": 7.116644823066841, + "grad_norm": 3.259796142578125, + "learning_rate": 1.1630993611198207e-05, + "loss": 0.5115, + "num_input_tokens_seen": 7268112, + "step": 27150 + }, + { + "epoch": 7.117955439056357, + "grad_norm": 6.858980178833008, + "learning_rate": 1.1621330463416027e-05, + "loss": 0.462, + "num_input_tokens_seen": 7269360, + "step": 27155 + }, + { + "epoch": 7.1192660550458715, + "grad_norm": 3.9366989135742188, + "learning_rate": 1.1611670115766265e-05, + "loss": 0.3519, + "num_input_tokens_seen": 7270656, + "step": 27160 + }, + { + "epoch": 7.120576671035387, + "grad_norm": 4.135689735412598, + "learning_rate": 1.1602012570270806e-05, + "loss": 0.5585, + "num_input_tokens_seen": 7272064, + "step": 27165 + }, + { + "epoch": 7.121887287024902, + "grad_norm": 15.745954513549805, + "learning_rate": 1.1592357828950976e-05, + "loss": 0.5439, + "num_input_tokens_seen": 7273568, + "step": 27170 + }, + { + "epoch": 7.123197903014416, + "grad_norm": 3.1988751888275146, + "learning_rate": 1.1582705893827483e-05, + "loss": 0.4985, + "num_input_tokens_seen": 7274992, + "step": 27175 + }, + { + "epoch": 7.124508519003932, + "grad_norm": 2.155407428741455, + "learning_rate": 1.1573056766920465e-05, + "loss": 0.3084, + "num_input_tokens_seen": 7276160, + "step": 27180 + }, + { + "epoch": 7.1258191349934465, + "grad_norm": 3.7455227375030518, + "learning_rate": 1.1563410450249468e-05, + "loss": 0.3675, + "num_input_tokens_seen": 7277552, + "step": 27185 + }, + { + "epoch": 7.127129750982962, + "grad_norm": 2.397585391998291, + "learning_rate": 1.1553766945833452e-05, + "loss": 0.3094, + "num_input_tokens_seen": 7278832, + "step": 27190 + }, + { + "epoch": 7.128440366972477, + "grad_norm": 2.5380353927612305, + "learning_rate": 1.1544126255690774e-05, + "loss": 0.5672, + "num_input_tokens_seen": 7280256, + "step": 27195 + }, + { + "epoch": 7.129750982961992, + "grad_norm": 1.4739576578140259, + "learning_rate": 1.153448838183922e-05, + "loss": 0.6281, + "num_input_tokens_seen": 7281712, + "step": 27200 + }, + { + "epoch": 7.131061598951507, + "grad_norm": 1.533361554145813, + "learning_rate": 1.1524853326295982e-05, + "loss": 0.9381, + "num_input_tokens_seen": 7283104, + "step": 27205 + }, + { + "epoch": 7.1323722149410225, + "grad_norm": 4.96901273727417, + "learning_rate": 1.151522109107766e-05, + "loss": 0.3909, + "num_input_tokens_seen": 7284688, + "step": 27210 + }, + { + "epoch": 7.133682830930537, + "grad_norm": 2.210754871368408, + "learning_rate": 1.150559167820026e-05, + "loss": 0.3082, + "num_input_tokens_seen": 7286176, + "step": 27215 + }, + { + "epoch": 7.134993446920053, + "grad_norm": 2.5631227493286133, + "learning_rate": 1.14959650896792e-05, + "loss": 0.4711, + "num_input_tokens_seen": 7287520, + "step": 27220 + }, + { + "epoch": 7.136304062909567, + "grad_norm": 4.038083553314209, + "learning_rate": 1.1486341327529332e-05, + "loss": 0.5606, + "num_input_tokens_seen": 7288432, + "step": 27225 + }, + { + "epoch": 7.137614678899083, + "grad_norm": 3.8529555797576904, + "learning_rate": 1.1476720393764864e-05, + "loss": 0.6323, + "num_input_tokens_seen": 7289536, + "step": 27230 + }, + { + "epoch": 7.1389252948885975, + "grad_norm": 4.938140392303467, + "learning_rate": 1.146710229039945e-05, + "loss": 0.396, + "num_input_tokens_seen": 7291056, + "step": 27235 + }, + { + "epoch": 7.140235910878113, + "grad_norm": 3.2357699871063232, + "learning_rate": 1.1457487019446145e-05, + "loss": 0.4213, + "num_input_tokens_seen": 7292064, + "step": 27240 + }, + { + "epoch": 7.141546526867628, + "grad_norm": 9.534383773803711, + "learning_rate": 1.1447874582917407e-05, + "loss": 0.3771, + "num_input_tokens_seen": 7293664, + "step": 27245 + }, + { + "epoch": 7.142857142857143, + "grad_norm": 1.9958041906356812, + "learning_rate": 1.1438264982825104e-05, + "loss": 0.473, + "num_input_tokens_seen": 7294928, + "step": 27250 + }, + { + "epoch": 7.144167758846658, + "grad_norm": 5.226180076599121, + "learning_rate": 1.1428658221180506e-05, + "loss": 0.4102, + "num_input_tokens_seen": 7296192, + "step": 27255 + }, + { + "epoch": 7.145478374836173, + "grad_norm": 2.1021947860717773, + "learning_rate": 1.1419054299994292e-05, + "loss": 0.4224, + "num_input_tokens_seen": 7297456, + "step": 27260 + }, + { + "epoch": 7.146788990825688, + "grad_norm": 2.28995943069458, + "learning_rate": 1.1409453221276545e-05, + "loss": 0.5144, + "num_input_tokens_seen": 7298832, + "step": 27265 + }, + { + "epoch": 7.148099606815203, + "grad_norm": 8.602524757385254, + "learning_rate": 1.1399854987036768e-05, + "loss": 0.4232, + "num_input_tokens_seen": 7299792, + "step": 27270 + }, + { + "epoch": 7.149410222804718, + "grad_norm": 4.766420364379883, + "learning_rate": 1.1390259599283835e-05, + "loss": 0.4243, + "num_input_tokens_seen": 7301056, + "step": 27275 + }, + { + "epoch": 7.150720838794233, + "grad_norm": 10.650915145874023, + "learning_rate": 1.1380667060026041e-05, + "loss": 1.0512, + "num_input_tokens_seen": 7302768, + "step": 27280 + }, + { + "epoch": 7.1520314547837485, + "grad_norm": 2.509835720062256, + "learning_rate": 1.1371077371271108e-05, + "loss": 0.5667, + "num_input_tokens_seen": 7303952, + "step": 27285 + }, + { + "epoch": 7.153342070773263, + "grad_norm": 1.5202564001083374, + "learning_rate": 1.136149053502613e-05, + "loss": 0.5744, + "num_input_tokens_seen": 7305520, + "step": 27290 + }, + { + "epoch": 7.154652686762779, + "grad_norm": 3.7856528759002686, + "learning_rate": 1.1351906553297622e-05, + "loss": 0.4236, + "num_input_tokens_seen": 7306688, + "step": 27295 + }, + { + "epoch": 7.155963302752293, + "grad_norm": 1.4936925172805786, + "learning_rate": 1.1342325428091488e-05, + "loss": 0.3827, + "num_input_tokens_seen": 7308368, + "step": 27300 + }, + { + "epoch": 7.157273918741809, + "grad_norm": 6.340560436248779, + "learning_rate": 1.1332747161413054e-05, + "loss": 0.4894, + "num_input_tokens_seen": 7309360, + "step": 27305 + }, + { + "epoch": 7.1585845347313235, + "grad_norm": 2.68902325630188, + "learning_rate": 1.1323171755267015e-05, + "loss": 0.4404, + "num_input_tokens_seen": 7310432, + "step": 27310 + }, + { + "epoch": 7.159895150720839, + "grad_norm": 2.0491645336151123, + "learning_rate": 1.1313599211657494e-05, + "loss": 0.4643, + "num_input_tokens_seen": 7311616, + "step": 27315 + }, + { + "epoch": 7.161205766710354, + "grad_norm": 3.169692277908325, + "learning_rate": 1.130402953258801e-05, + "loss": 0.4313, + "num_input_tokens_seen": 7313376, + "step": 27320 + }, + { + "epoch": 7.162516382699869, + "grad_norm": 2.304887294769287, + "learning_rate": 1.129446272006148e-05, + "loss": 0.5594, + "num_input_tokens_seen": 7314624, + "step": 27325 + }, + { + "epoch": 7.163826998689384, + "grad_norm": 1.4523773193359375, + "learning_rate": 1.1284898776080222e-05, + "loss": 0.3615, + "num_input_tokens_seen": 7316224, + "step": 27330 + }, + { + "epoch": 7.165137614678899, + "grad_norm": 4.1719889640808105, + "learning_rate": 1.127533770264594e-05, + "loss": 0.4753, + "num_input_tokens_seen": 7317440, + "step": 27335 + }, + { + "epoch": 7.166448230668414, + "grad_norm": 1.549621343612671, + "learning_rate": 1.1265779501759786e-05, + "loss": 0.4198, + "num_input_tokens_seen": 7318736, + "step": 27340 + }, + { + "epoch": 7.16775884665793, + "grad_norm": 1.4357951879501343, + "learning_rate": 1.125622417542224e-05, + "loss": 0.7348, + "num_input_tokens_seen": 7320112, + "step": 27345 + }, + { + "epoch": 7.169069462647444, + "grad_norm": 1.674551010131836, + "learning_rate": 1.1246671725633227e-05, + "loss": 0.5443, + "num_input_tokens_seen": 7321392, + "step": 27350 + }, + { + "epoch": 7.17038007863696, + "grad_norm": 1.4832404851913452, + "learning_rate": 1.1237122154392058e-05, + "loss": 0.531, + "num_input_tokens_seen": 7322960, + "step": 27355 + }, + { + "epoch": 7.1716906946264745, + "grad_norm": 5.313262939453125, + "learning_rate": 1.122757546369744e-05, + "loss": 0.5176, + "num_input_tokens_seen": 7323984, + "step": 27360 + }, + { + "epoch": 7.173001310615989, + "grad_norm": 4.589288234710693, + "learning_rate": 1.1218031655547482e-05, + "loss": 0.4252, + "num_input_tokens_seen": 7325504, + "step": 27365 + }, + { + "epoch": 7.174311926605505, + "grad_norm": 2.669217824935913, + "learning_rate": 1.1208490731939683e-05, + "loss": 0.6102, + "num_input_tokens_seen": 7326864, + "step": 27370 + }, + { + "epoch": 7.175622542595019, + "grad_norm": 5.871502876281738, + "learning_rate": 1.1198952694870945e-05, + "loss": 0.4306, + "num_input_tokens_seen": 7328000, + "step": 27375 + }, + { + "epoch": 7.176933158584535, + "grad_norm": 2.7191526889801025, + "learning_rate": 1.1189417546337564e-05, + "loss": 0.3602, + "num_input_tokens_seen": 7329456, + "step": 27380 + }, + { + "epoch": 7.1782437745740495, + "grad_norm": 1.2822188138961792, + "learning_rate": 1.1179885288335235e-05, + "loss": 0.3242, + "num_input_tokens_seen": 7330576, + "step": 27385 + }, + { + "epoch": 7.179554390563565, + "grad_norm": 1.3858356475830078, + "learning_rate": 1.1170355922859024e-05, + "loss": 0.2948, + "num_input_tokens_seen": 7331744, + "step": 27390 + }, + { + "epoch": 7.18086500655308, + "grad_norm": 3.74884033203125, + "learning_rate": 1.1160829451903418e-05, + "loss": 0.3892, + "num_input_tokens_seen": 7333568, + "step": 27395 + }, + { + "epoch": 7.182175622542595, + "grad_norm": 4.746233940124512, + "learning_rate": 1.1151305877462304e-05, + "loss": 0.6362, + "num_input_tokens_seen": 7334624, + "step": 27400 + }, + { + "epoch": 7.18348623853211, + "grad_norm": 1.492077112197876, + "learning_rate": 1.1141785201528936e-05, + "loss": 0.4219, + "num_input_tokens_seen": 7336144, + "step": 27405 + }, + { + "epoch": 7.184796854521625, + "grad_norm": 0.8806624412536621, + "learning_rate": 1.1132267426095985e-05, + "loss": 0.3463, + "num_input_tokens_seen": 7337696, + "step": 27410 + }, + { + "epoch": 7.18610747051114, + "grad_norm": 4.505166053771973, + "learning_rate": 1.11227525531555e-05, + "loss": 0.4372, + "num_input_tokens_seen": 7338864, + "step": 27415 + }, + { + "epoch": 7.187418086500656, + "grad_norm": 3.1988070011138916, + "learning_rate": 1.1113240584698927e-05, + "loss": 0.6563, + "num_input_tokens_seen": 7340144, + "step": 27420 + }, + { + "epoch": 7.18872870249017, + "grad_norm": 3.96533465385437, + "learning_rate": 1.1103731522717115e-05, + "loss": 0.3836, + "num_input_tokens_seen": 7341424, + "step": 27425 + }, + { + "epoch": 7.190039318479686, + "grad_norm": 3.1776468753814697, + "learning_rate": 1.1094225369200272e-05, + "loss": 0.3809, + "num_input_tokens_seen": 7342368, + "step": 27430 + }, + { + "epoch": 7.1913499344692005, + "grad_norm": 6.617096900939941, + "learning_rate": 1.1084722126138037e-05, + "loss": 0.6, + "num_input_tokens_seen": 7343568, + "step": 27435 + }, + { + "epoch": 7.192660550458716, + "grad_norm": 3.257808208465576, + "learning_rate": 1.1075221795519413e-05, + "loss": 0.4259, + "num_input_tokens_seen": 7345280, + "step": 27440 + }, + { + "epoch": 7.193971166448231, + "grad_norm": 1.9860756397247314, + "learning_rate": 1.106572437933281e-05, + "loss": 0.3835, + "num_input_tokens_seen": 7346656, + "step": 27445 + }, + { + "epoch": 7.195281782437746, + "grad_norm": 1.364261269569397, + "learning_rate": 1.1056229879566004e-05, + "loss": 0.4121, + "num_input_tokens_seen": 7347856, + "step": 27450 + }, + { + "epoch": 7.196592398427261, + "grad_norm": 1.3823587894439697, + "learning_rate": 1.1046738298206205e-05, + "loss": 0.4329, + "num_input_tokens_seen": 7349648, + "step": 27455 + }, + { + "epoch": 7.1979030144167755, + "grad_norm": 3.562032699584961, + "learning_rate": 1.103724963723998e-05, + "loss": 0.4815, + "num_input_tokens_seen": 7350768, + "step": 27460 + }, + { + "epoch": 7.199213630406291, + "grad_norm": 5.273187637329102, + "learning_rate": 1.1027763898653268e-05, + "loss": 0.7129, + "num_input_tokens_seen": 7352016, + "step": 27465 + }, + { + "epoch": 7.200524246395806, + "grad_norm": 1.7579790353775024, + "learning_rate": 1.1018281084431432e-05, + "loss": 0.478, + "num_input_tokens_seen": 7353456, + "step": 27470 + }, + { + "epoch": 7.201834862385321, + "grad_norm": 3.0089149475097656, + "learning_rate": 1.1008801196559204e-05, + "loss": 0.4582, + "num_input_tokens_seen": 7354848, + "step": 27475 + }, + { + "epoch": 7.203145478374836, + "grad_norm": 3.2578911781311035, + "learning_rate": 1.0999324237020714e-05, + "loss": 0.4581, + "num_input_tokens_seen": 7356592, + "step": 27480 + }, + { + "epoch": 7.204456094364351, + "grad_norm": 4.112592697143555, + "learning_rate": 1.0989850207799469e-05, + "loss": 0.4829, + "num_input_tokens_seen": 7357904, + "step": 27485 + }, + { + "epoch": 7.205766710353866, + "grad_norm": 2.239109516143799, + "learning_rate": 1.0980379110878366e-05, + "loss": 0.5762, + "num_input_tokens_seen": 7359472, + "step": 27490 + }, + { + "epoch": 7.207077326343382, + "grad_norm": 2.865509271621704, + "learning_rate": 1.0970910948239694e-05, + "loss": 0.7727, + "num_input_tokens_seen": 7360784, + "step": 27495 + }, + { + "epoch": 7.208387942332896, + "grad_norm": 3.0187528133392334, + "learning_rate": 1.0961445721865116e-05, + "loss": 0.5031, + "num_input_tokens_seen": 7361856, + "step": 27500 + }, + { + "epoch": 7.209698558322412, + "grad_norm": 2.6490535736083984, + "learning_rate": 1.0951983433735704e-05, + "loss": 0.4915, + "num_input_tokens_seen": 7363184, + "step": 27505 + }, + { + "epoch": 7.2110091743119265, + "grad_norm": 4.370042324066162, + "learning_rate": 1.0942524085831868e-05, + "loss": 0.5609, + "num_input_tokens_seen": 7364608, + "step": 27510 + }, + { + "epoch": 7.212319790301442, + "grad_norm": 2.6505885124206543, + "learning_rate": 1.0933067680133457e-05, + "loss": 0.6183, + "num_input_tokens_seen": 7366016, + "step": 27515 + }, + { + "epoch": 7.213630406290957, + "grad_norm": 3.9169607162475586, + "learning_rate": 1.092361421861968e-05, + "loss": 0.4793, + "num_input_tokens_seen": 7367072, + "step": 27520 + }, + { + "epoch": 7.214941022280472, + "grad_norm": 1.6626514196395874, + "learning_rate": 1.091416370326912e-05, + "loss": 0.414, + "num_input_tokens_seen": 7368288, + "step": 27525 + }, + { + "epoch": 7.216251638269987, + "grad_norm": 5.891592502593994, + "learning_rate": 1.090471613605976e-05, + "loss": 0.5008, + "num_input_tokens_seen": 7369456, + "step": 27530 + }, + { + "epoch": 7.217562254259502, + "grad_norm": 3.773040533065796, + "learning_rate": 1.0895271518968962e-05, + "loss": 0.4072, + "num_input_tokens_seen": 7370448, + "step": 27535 + }, + { + "epoch": 7.218872870249017, + "grad_norm": 2.452003240585327, + "learning_rate": 1.0885829853973467e-05, + "loss": 0.5744, + "num_input_tokens_seen": 7372336, + "step": 27540 + }, + { + "epoch": 7.220183486238533, + "grad_norm": 2.8878960609436035, + "learning_rate": 1.087639114304939e-05, + "loss": 0.4863, + "num_input_tokens_seen": 7373424, + "step": 27545 + }, + { + "epoch": 7.221494102228047, + "grad_norm": 4.259128093719482, + "learning_rate": 1.0866955388172245e-05, + "loss": 0.7284, + "num_input_tokens_seen": 7374768, + "step": 27550 + }, + { + "epoch": 7.222804718217562, + "grad_norm": 3.0682432651519775, + "learning_rate": 1.0857522591316915e-05, + "loss": 0.6486, + "num_input_tokens_seen": 7375936, + "step": 27555 + }, + { + "epoch": 7.224115334207077, + "grad_norm": 1.3136460781097412, + "learning_rate": 1.084809275445767e-05, + "loss": 0.5083, + "num_input_tokens_seen": 7377392, + "step": 27560 + }, + { + "epoch": 7.225425950196592, + "grad_norm": 2.817042112350464, + "learning_rate": 1.083866587956815e-05, + "loss": 0.4898, + "num_input_tokens_seen": 7378480, + "step": 27565 + }, + { + "epoch": 7.226736566186108, + "grad_norm": 2.1925761699676514, + "learning_rate": 1.0829241968621401e-05, + "loss": 0.5104, + "num_input_tokens_seen": 7379888, + "step": 27570 + }, + { + "epoch": 7.228047182175622, + "grad_norm": 2.399871826171875, + "learning_rate": 1.0819821023589832e-05, + "loss": 0.6651, + "num_input_tokens_seen": 7381232, + "step": 27575 + }, + { + "epoch": 7.229357798165138, + "grad_norm": 1.224013328552246, + "learning_rate": 1.0810403046445214e-05, + "loss": 0.5193, + "num_input_tokens_seen": 7383360, + "step": 27580 + }, + { + "epoch": 7.2306684141546524, + "grad_norm": 3.96947979927063, + "learning_rate": 1.0800988039158719e-05, + "loss": 0.3921, + "num_input_tokens_seen": 7384880, + "step": 27585 + }, + { + "epoch": 7.231979030144168, + "grad_norm": 2.446038007736206, + "learning_rate": 1.0791576003700895e-05, + "loss": 0.5402, + "num_input_tokens_seen": 7386192, + "step": 27590 + }, + { + "epoch": 7.233289646133683, + "grad_norm": 8.415780067443848, + "learning_rate": 1.0782166942041663e-05, + "loss": 0.3174, + "num_input_tokens_seen": 7387264, + "step": 27595 + }, + { + "epoch": 7.234600262123198, + "grad_norm": 3.6825978755950928, + "learning_rate": 1.0772760856150322e-05, + "loss": 0.5396, + "num_input_tokens_seen": 7388816, + "step": 27600 + }, + { + "epoch": 7.235910878112713, + "grad_norm": 1.914732813835144, + "learning_rate": 1.0763357747995551e-05, + "loss": 0.479, + "num_input_tokens_seen": 7390624, + "step": 27605 + }, + { + "epoch": 7.237221494102228, + "grad_norm": 1.7497165203094482, + "learning_rate": 1.0753957619545405e-05, + "loss": 0.4387, + "num_input_tokens_seen": 7392480, + "step": 27610 + }, + { + "epoch": 7.238532110091743, + "grad_norm": 1.735542893409729, + "learning_rate": 1.0744560472767315e-05, + "loss": 0.4817, + "num_input_tokens_seen": 7393536, + "step": 27615 + }, + { + "epoch": 7.239842726081259, + "grad_norm": 2.556222677230835, + "learning_rate": 1.0735166309628083e-05, + "loss": 0.3269, + "num_input_tokens_seen": 7395104, + "step": 27620 + }, + { + "epoch": 7.241153342070773, + "grad_norm": 6.256320476531982, + "learning_rate": 1.0725775132093899e-05, + "loss": 0.5775, + "num_input_tokens_seen": 7396192, + "step": 27625 + }, + { + "epoch": 7.242463958060289, + "grad_norm": 1.9446640014648438, + "learning_rate": 1.0716386942130313e-05, + "loss": 0.7244, + "num_input_tokens_seen": 7397728, + "step": 27630 + }, + { + "epoch": 7.243774574049803, + "grad_norm": 1.9290871620178223, + "learning_rate": 1.0707001741702261e-05, + "loss": 0.651, + "num_input_tokens_seen": 7399616, + "step": 27635 + }, + { + "epoch": 7.245085190039318, + "grad_norm": 2.29412841796875, + "learning_rate": 1.0697619532774047e-05, + "loss": 0.4659, + "num_input_tokens_seen": 7401200, + "step": 27640 + }, + { + "epoch": 7.246395806028834, + "grad_norm": 4.206180572509766, + "learning_rate": 1.0688240317309355e-05, + "loss": 0.4184, + "num_input_tokens_seen": 7402624, + "step": 27645 + }, + { + "epoch": 7.247706422018348, + "grad_norm": 4.904531955718994, + "learning_rate": 1.0678864097271234e-05, + "loss": 0.3873, + "num_input_tokens_seen": 7403776, + "step": 27650 + }, + { + "epoch": 7.249017038007864, + "grad_norm": 3.8448479175567627, + "learning_rate": 1.0669490874622112e-05, + "loss": 0.3705, + "num_input_tokens_seen": 7404992, + "step": 27655 + }, + { + "epoch": 7.250327653997378, + "grad_norm": 6.695855617523193, + "learning_rate": 1.0660120651323799e-05, + "loss": 0.4656, + "num_input_tokens_seen": 7406336, + "step": 27660 + }, + { + "epoch": 7.251638269986894, + "grad_norm": 3.966033697128296, + "learning_rate": 1.0650753429337446e-05, + "loss": 0.5571, + "num_input_tokens_seen": 7407392, + "step": 27665 + }, + { + "epoch": 7.252948885976409, + "grad_norm": 11.3234281539917, + "learning_rate": 1.0641389210623607e-05, + "loss": 0.5866, + "num_input_tokens_seen": 7408800, + "step": 27670 + }, + { + "epoch": 7.254259501965924, + "grad_norm": 3.857879638671875, + "learning_rate": 1.0632027997142196e-05, + "loss": 0.3715, + "num_input_tokens_seen": 7410176, + "step": 27675 + }, + { + "epoch": 7.255570117955439, + "grad_norm": 5.490236759185791, + "learning_rate": 1.062266979085249e-05, + "loss": 0.6175, + "num_input_tokens_seen": 7411392, + "step": 27680 + }, + { + "epoch": 7.256880733944954, + "grad_norm": 6.398519992828369, + "learning_rate": 1.0613314593713164e-05, + "loss": 0.4539, + "num_input_tokens_seen": 7412464, + "step": 27685 + }, + { + "epoch": 7.258191349934469, + "grad_norm": 2.7716636657714844, + "learning_rate": 1.0603962407682236e-05, + "loss": 0.3446, + "num_input_tokens_seen": 7413904, + "step": 27690 + }, + { + "epoch": 7.259501965923985, + "grad_norm": 10.231462478637695, + "learning_rate": 1.059461323471711e-05, + "loss": 0.2835, + "num_input_tokens_seen": 7415472, + "step": 27695 + }, + { + "epoch": 7.260812581913499, + "grad_norm": 9.792598724365234, + "learning_rate": 1.0585267076774536e-05, + "loss": 0.4821, + "num_input_tokens_seen": 7416784, + "step": 27700 + }, + { + "epoch": 7.262123197903015, + "grad_norm": 4.362481594085693, + "learning_rate": 1.0575923935810656e-05, + "loss": 0.553, + "num_input_tokens_seen": 7418096, + "step": 27705 + }, + { + "epoch": 7.263433813892529, + "grad_norm": 5.610560894012451, + "learning_rate": 1.056658381378097e-05, + "loss": 0.4808, + "num_input_tokens_seen": 7419296, + "step": 27710 + }, + { + "epoch": 7.264744429882045, + "grad_norm": 3.537689447402954, + "learning_rate": 1.0557246712640356e-05, + "loss": 0.4642, + "num_input_tokens_seen": 7420784, + "step": 27715 + }, + { + "epoch": 7.26605504587156, + "grad_norm": 4.766613960266113, + "learning_rate": 1.054791263434305e-05, + "loss": 0.4691, + "num_input_tokens_seen": 7421968, + "step": 27720 + }, + { + "epoch": 7.267365661861074, + "grad_norm": 2.880091428756714, + "learning_rate": 1.0538581580842655e-05, + "loss": 0.448, + "num_input_tokens_seen": 7423152, + "step": 27725 + }, + { + "epoch": 7.26867627785059, + "grad_norm": 3.45727276802063, + "learning_rate": 1.052925355409215e-05, + "loss": 0.4384, + "num_input_tokens_seen": 7424816, + "step": 27730 + }, + { + "epoch": 7.269986893840104, + "grad_norm": 1.6395896673202515, + "learning_rate": 1.0519928556043873e-05, + "loss": 0.4644, + "num_input_tokens_seen": 7426272, + "step": 27735 + }, + { + "epoch": 7.27129750982962, + "grad_norm": 3.0906753540039062, + "learning_rate": 1.0510606588649526e-05, + "loss": 0.2777, + "num_input_tokens_seen": 7427168, + "step": 27740 + }, + { + "epoch": 7.272608125819135, + "grad_norm": 2.870229959487915, + "learning_rate": 1.0501287653860189e-05, + "loss": 0.3954, + "num_input_tokens_seen": 7428320, + "step": 27745 + }, + { + "epoch": 7.27391874180865, + "grad_norm": 1.9690911769866943, + "learning_rate": 1.0491971753626292e-05, + "loss": 0.3819, + "num_input_tokens_seen": 7429840, + "step": 27750 + }, + { + "epoch": 7.275229357798165, + "grad_norm": 1.541184425354004, + "learning_rate": 1.048265888989764e-05, + "loss": 0.4501, + "num_input_tokens_seen": 7431408, + "step": 27755 + }, + { + "epoch": 7.27653997378768, + "grad_norm": 3.761406421661377, + "learning_rate": 1.0473349064623397e-05, + "loss": 0.5582, + "num_input_tokens_seen": 7432512, + "step": 27760 + }, + { + "epoch": 7.277850589777195, + "grad_norm": 5.875897407531738, + "learning_rate": 1.0464042279752096e-05, + "loss": 0.5126, + "num_input_tokens_seen": 7433792, + "step": 27765 + }, + { + "epoch": 7.2791612057667106, + "grad_norm": 1.875113606452942, + "learning_rate": 1.0454738537231628e-05, + "loss": 0.4677, + "num_input_tokens_seen": 7435696, + "step": 27770 + }, + { + "epoch": 7.280471821756225, + "grad_norm": 2.5415399074554443, + "learning_rate": 1.0445437839009265e-05, + "loss": 0.3686, + "num_input_tokens_seen": 7437184, + "step": 27775 + }, + { + "epoch": 7.281782437745741, + "grad_norm": 6.36135196685791, + "learning_rate": 1.0436140187031601e-05, + "loss": 0.5385, + "num_input_tokens_seen": 7438880, + "step": 27780 + }, + { + "epoch": 7.283093053735255, + "grad_norm": 3.3205368518829346, + "learning_rate": 1.0426845583244632e-05, + "loss": 0.3901, + "num_input_tokens_seen": 7440064, + "step": 27785 + }, + { + "epoch": 7.284403669724771, + "grad_norm": 3.9712324142456055, + "learning_rate": 1.04175540295937e-05, + "loss": 1.3327, + "num_input_tokens_seen": 7441296, + "step": 27790 + }, + { + "epoch": 7.285714285714286, + "grad_norm": 4.421400547027588, + "learning_rate": 1.0408265528023508e-05, + "loss": 0.3592, + "num_input_tokens_seen": 7442400, + "step": 27795 + }, + { + "epoch": 7.287024901703801, + "grad_norm": 2.129439353942871, + "learning_rate": 1.0398980080478139e-05, + "loss": 0.4278, + "num_input_tokens_seen": 7443392, + "step": 27800 + }, + { + "epoch": 7.288335517693316, + "grad_norm": 1.2402920722961426, + "learning_rate": 1.0389697688901003e-05, + "loss": 0.3021, + "num_input_tokens_seen": 7444880, + "step": 27805 + }, + { + "epoch": 7.289646133682831, + "grad_norm": 3.066850423812866, + "learning_rate": 1.038041835523491e-05, + "loss": 0.6237, + "num_input_tokens_seen": 7445952, + "step": 27810 + }, + { + "epoch": 7.290956749672346, + "grad_norm": 13.423921585083008, + "learning_rate": 1.0371142081421987e-05, + "loss": 0.5313, + "num_input_tokens_seen": 7446896, + "step": 27815 + }, + { + "epoch": 7.292267365661861, + "grad_norm": 3.761197328567505, + "learning_rate": 1.0361868869403745e-05, + "loss": 0.5298, + "num_input_tokens_seen": 7448464, + "step": 27820 + }, + { + "epoch": 7.293577981651376, + "grad_norm": 0.8859174251556396, + "learning_rate": 1.035259872112106e-05, + "loss": 0.3288, + "num_input_tokens_seen": 7450080, + "step": 27825 + }, + { + "epoch": 7.294888597640891, + "grad_norm": 2.7830045223236084, + "learning_rate": 1.0343331638514154e-05, + "loss": 0.6255, + "num_input_tokens_seen": 7451264, + "step": 27830 + }, + { + "epoch": 7.296199213630406, + "grad_norm": 2.0214555263519287, + "learning_rate": 1.0334067623522609e-05, + "loss": 0.56, + "num_input_tokens_seen": 7452688, + "step": 27835 + }, + { + "epoch": 7.297509829619921, + "grad_norm": 9.625120162963867, + "learning_rate": 1.032480667808537e-05, + "loss": 0.4066, + "num_input_tokens_seen": 7453792, + "step": 27840 + }, + { + "epoch": 7.2988204456094365, + "grad_norm": 3.347195625305176, + "learning_rate": 1.0315548804140736e-05, + "loss": 0.3544, + "num_input_tokens_seen": 7455184, + "step": 27845 + }, + { + "epoch": 7.300131061598951, + "grad_norm": 13.58961009979248, + "learning_rate": 1.0306294003626365e-05, + "loss": 0.424, + "num_input_tokens_seen": 7456416, + "step": 27850 + }, + { + "epoch": 7.301441677588467, + "grad_norm": 4.317567348480225, + "learning_rate": 1.029704227847927e-05, + "loss": 0.4674, + "num_input_tokens_seen": 7457632, + "step": 27855 + }, + { + "epoch": 7.302752293577981, + "grad_norm": 2.630927324295044, + "learning_rate": 1.0287793630635822e-05, + "loss": 0.4039, + "num_input_tokens_seen": 7459072, + "step": 27860 + }, + { + "epoch": 7.304062909567497, + "grad_norm": 2.6140382289886475, + "learning_rate": 1.0278548062031748e-05, + "loss": 0.4664, + "num_input_tokens_seen": 7460304, + "step": 27865 + }, + { + "epoch": 7.305373525557012, + "grad_norm": 3.7854866981506348, + "learning_rate": 1.0269305574602125e-05, + "loss": 0.4094, + "num_input_tokens_seen": 7461776, + "step": 27870 + }, + { + "epoch": 7.306684141546527, + "grad_norm": 3.8645191192626953, + "learning_rate": 1.0260066170281395e-05, + "loss": 0.3457, + "num_input_tokens_seen": 7463376, + "step": 27875 + }, + { + "epoch": 7.307994757536042, + "grad_norm": 4.678675174713135, + "learning_rate": 1.0250829851003346e-05, + "loss": 0.4768, + "num_input_tokens_seen": 7464464, + "step": 27880 + }, + { + "epoch": 7.309305373525557, + "grad_norm": 2.78240704536438, + "learning_rate": 1.0241596618701124e-05, + "loss": 0.5176, + "num_input_tokens_seen": 7465440, + "step": 27885 + }, + { + "epoch": 7.310615989515072, + "grad_norm": 5.53643274307251, + "learning_rate": 1.0232366475307228e-05, + "loss": 0.514, + "num_input_tokens_seen": 7466688, + "step": 27890 + }, + { + "epoch": 7.3119266055045875, + "grad_norm": 3.843292474746704, + "learning_rate": 1.0223139422753524e-05, + "loss": 0.4933, + "num_input_tokens_seen": 7468576, + "step": 27895 + }, + { + "epoch": 7.313237221494102, + "grad_norm": 4.874563694000244, + "learning_rate": 1.0213915462971197e-05, + "loss": 0.8792, + "num_input_tokens_seen": 7469744, + "step": 27900 + }, + { + "epoch": 7.314547837483618, + "grad_norm": 3.903240919113159, + "learning_rate": 1.0204694597890812e-05, + "loss": 0.7998, + "num_input_tokens_seen": 7471168, + "step": 27905 + }, + { + "epoch": 7.315858453473132, + "grad_norm": 3.2925145626068115, + "learning_rate": 1.0195476829442277e-05, + "loss": 0.6085, + "num_input_tokens_seen": 7472368, + "step": 27910 + }, + { + "epoch": 7.317169069462647, + "grad_norm": 6.016820907592773, + "learning_rate": 1.0186262159554872e-05, + "loss": 0.3717, + "num_input_tokens_seen": 7473552, + "step": 27915 + }, + { + "epoch": 7.3184796854521625, + "grad_norm": 1.3555868864059448, + "learning_rate": 1.0177050590157197e-05, + "loss": 0.6002, + "num_input_tokens_seen": 7474976, + "step": 27920 + }, + { + "epoch": 7.319790301441677, + "grad_norm": 3.4419453144073486, + "learning_rate": 1.0167842123177223e-05, + "loss": 0.3858, + "num_input_tokens_seen": 7476256, + "step": 27925 + }, + { + "epoch": 7.321100917431193, + "grad_norm": 1.9278768301010132, + "learning_rate": 1.0158636760542276e-05, + "loss": 0.3862, + "num_input_tokens_seen": 7477456, + "step": 27930 + }, + { + "epoch": 7.322411533420707, + "grad_norm": 2.380173683166504, + "learning_rate": 1.0149434504178997e-05, + "loss": 0.4634, + "num_input_tokens_seen": 7478768, + "step": 27935 + }, + { + "epoch": 7.323722149410223, + "grad_norm": 17.654815673828125, + "learning_rate": 1.0140235356013422e-05, + "loss": 0.5875, + "num_input_tokens_seen": 7480368, + "step": 27940 + }, + { + "epoch": 7.325032765399738, + "grad_norm": 4.323280334472656, + "learning_rate": 1.0131039317970908e-05, + "loss": 0.5695, + "num_input_tokens_seen": 7482000, + "step": 27945 + }, + { + "epoch": 7.326343381389253, + "grad_norm": 1.3434196710586548, + "learning_rate": 1.0121846391976173e-05, + "loss": 0.3195, + "num_input_tokens_seen": 7483616, + "step": 27950 + }, + { + "epoch": 7.327653997378768, + "grad_norm": 2.5448951721191406, + "learning_rate": 1.0112656579953283e-05, + "loss": 0.5462, + "num_input_tokens_seen": 7485104, + "step": 27955 + }, + { + "epoch": 7.328964613368283, + "grad_norm": 2.06251859664917, + "learning_rate": 1.0103469883825645e-05, + "loss": 0.36, + "num_input_tokens_seen": 7486352, + "step": 27960 + }, + { + "epoch": 7.330275229357798, + "grad_norm": 5.8678131103515625, + "learning_rate": 1.0094286305516023e-05, + "loss": 0.4465, + "num_input_tokens_seen": 7487984, + "step": 27965 + }, + { + "epoch": 7.3315858453473135, + "grad_norm": 1.9760240316390991, + "learning_rate": 1.008510584694652e-05, + "loss": 0.5108, + "num_input_tokens_seen": 7489232, + "step": 27970 + }, + { + "epoch": 7.332896461336828, + "grad_norm": 5.903847694396973, + "learning_rate": 1.0075928510038593e-05, + "loss": 0.4118, + "num_input_tokens_seen": 7490560, + "step": 27975 + }, + { + "epoch": 7.334207077326344, + "grad_norm": 2.415942907333374, + "learning_rate": 1.0066754296713043e-05, + "loss": 0.5463, + "num_input_tokens_seen": 7491600, + "step": 27980 + }, + { + "epoch": 7.335517693315858, + "grad_norm": 2.008681058883667, + "learning_rate": 1.0057583208890017e-05, + "loss": 0.5266, + "num_input_tokens_seen": 7493120, + "step": 27985 + }, + { + "epoch": 7.336828309305374, + "grad_norm": 2.341080904006958, + "learning_rate": 1.0048415248489004e-05, + "loss": 0.3666, + "num_input_tokens_seen": 7494880, + "step": 27990 + }, + { + "epoch": 7.3381389252948885, + "grad_norm": 1.7961355447769165, + "learning_rate": 1.0039250417428847e-05, + "loss": 0.5284, + "num_input_tokens_seen": 7496032, + "step": 27995 + }, + { + "epoch": 7.339449541284404, + "grad_norm": 2.6747982501983643, + "learning_rate": 1.0030088717627729e-05, + "loss": 0.4158, + "num_input_tokens_seen": 7497328, + "step": 28000 + }, + { + "epoch": 7.340760157273919, + "grad_norm": 10.893765449523926, + "learning_rate": 1.0020930151003174e-05, + "loss": 0.3657, + "num_input_tokens_seen": 7498640, + "step": 28005 + }, + { + "epoch": 7.342070773263433, + "grad_norm": 3.3657398223876953, + "learning_rate": 1.0011774719472064e-05, + "loss": 0.4634, + "num_input_tokens_seen": 7500368, + "step": 28010 + }, + { + "epoch": 7.343381389252949, + "grad_norm": 4.713408946990967, + "learning_rate": 1.0002622424950604e-05, + "loss": 0.4452, + "num_input_tokens_seen": 7501664, + "step": 28015 + }, + { + "epoch": 7.344692005242464, + "grad_norm": 2.9350292682647705, + "learning_rate": 9.99347326935435e-06, + "loss": 0.4945, + "num_input_tokens_seen": 7502992, + "step": 28020 + }, + { + "epoch": 7.346002621231979, + "grad_norm": 6.455787658691406, + "learning_rate": 9.984327254598208e-06, + "loss": 0.5081, + "num_input_tokens_seen": 7504192, + "step": 28025 + }, + { + "epoch": 7.347313237221494, + "grad_norm": 2.9115231037139893, + "learning_rate": 9.975184382596434e-06, + "loss": 0.4089, + "num_input_tokens_seen": 7505296, + "step": 28030 + }, + { + "epoch": 7.348623853211009, + "grad_norm": 8.19745922088623, + "learning_rate": 9.966044655262607e-06, + "loss": 0.3845, + "num_input_tokens_seen": 7506448, + "step": 28035 + }, + { + "epoch": 7.349934469200524, + "grad_norm": 5.541910648345947, + "learning_rate": 9.956908074509658e-06, + "loss": 0.5149, + "num_input_tokens_seen": 7508064, + "step": 28040 + }, + { + "epoch": 7.3512450851900395, + "grad_norm": 2.930173873901367, + "learning_rate": 9.947774642249863e-06, + "loss": 0.4362, + "num_input_tokens_seen": 7509088, + "step": 28045 + }, + { + "epoch": 7.352555701179554, + "grad_norm": 1.1979721784591675, + "learning_rate": 9.938644360394822e-06, + "loss": 0.3694, + "num_input_tokens_seen": 7510448, + "step": 28050 + }, + { + "epoch": 7.35386631716907, + "grad_norm": 1.3200510740280151, + "learning_rate": 9.92951723085549e-06, + "loss": 0.3259, + "num_input_tokens_seen": 7511728, + "step": 28055 + }, + { + "epoch": 7.355176933158584, + "grad_norm": 1.4924215078353882, + "learning_rate": 9.920393255542163e-06, + "loss": 0.4864, + "num_input_tokens_seen": 7513152, + "step": 28060 + }, + { + "epoch": 7.3564875491481, + "grad_norm": 2.802443504333496, + "learning_rate": 9.911272436364469e-06, + "loss": 0.2547, + "num_input_tokens_seen": 7514304, + "step": 28065 + }, + { + "epoch": 7.3577981651376145, + "grad_norm": 2.573765277862549, + "learning_rate": 9.902154775231386e-06, + "loss": 0.418, + "num_input_tokens_seen": 7515552, + "step": 28070 + }, + { + "epoch": 7.35910878112713, + "grad_norm": 3.3579795360565186, + "learning_rate": 9.893040274051221e-06, + "loss": 0.5021, + "num_input_tokens_seen": 7516864, + "step": 28075 + }, + { + "epoch": 7.360419397116645, + "grad_norm": 2.781102180480957, + "learning_rate": 9.883928934731615e-06, + "loss": 0.4273, + "num_input_tokens_seen": 7518672, + "step": 28080 + }, + { + "epoch": 7.36173001310616, + "grad_norm": 2.7142281532287598, + "learning_rate": 9.874820759179581e-06, + "loss": 0.66, + "num_input_tokens_seen": 7520000, + "step": 28085 + }, + { + "epoch": 7.363040629095675, + "grad_norm": 2.502506971359253, + "learning_rate": 9.865715749301419e-06, + "loss": 0.5295, + "num_input_tokens_seen": 7521584, + "step": 28090 + }, + { + "epoch": 7.3643512450851905, + "grad_norm": 1.9284080266952515, + "learning_rate": 9.8566139070028e-06, + "loss": 0.4051, + "num_input_tokens_seen": 7522592, + "step": 28095 + }, + { + "epoch": 7.365661861074705, + "grad_norm": 2.2454068660736084, + "learning_rate": 9.847515234188726e-06, + "loss": 0.3876, + "num_input_tokens_seen": 7523648, + "step": 28100 + }, + { + "epoch": 7.36697247706422, + "grad_norm": 4.282873630523682, + "learning_rate": 9.838419732763531e-06, + "loss": 0.4444, + "num_input_tokens_seen": 7524768, + "step": 28105 + }, + { + "epoch": 7.368283093053735, + "grad_norm": 1.973257303237915, + "learning_rate": 9.829327404630891e-06, + "loss": 0.4833, + "num_input_tokens_seen": 7525904, + "step": 28110 + }, + { + "epoch": 7.36959370904325, + "grad_norm": 3.0383150577545166, + "learning_rate": 9.820238251693811e-06, + "loss": 0.6167, + "num_input_tokens_seen": 7527184, + "step": 28115 + }, + { + "epoch": 7.3709043250327655, + "grad_norm": 5.653012275695801, + "learning_rate": 9.811152275854635e-06, + "loss": 0.3693, + "num_input_tokens_seen": 7528752, + "step": 28120 + }, + { + "epoch": 7.37221494102228, + "grad_norm": 1.8801065683364868, + "learning_rate": 9.802069479015044e-06, + "loss": 0.3311, + "num_input_tokens_seen": 7530240, + "step": 28125 + }, + { + "epoch": 7.373525557011796, + "grad_norm": 2.020434856414795, + "learning_rate": 9.792989863076064e-06, + "loss": 0.5435, + "num_input_tokens_seen": 7531296, + "step": 28130 + }, + { + "epoch": 7.37483617300131, + "grad_norm": 2.877758502960205, + "learning_rate": 9.783913429938015e-06, + "loss": 0.5731, + "num_input_tokens_seen": 7532464, + "step": 28135 + }, + { + "epoch": 7.376146788990826, + "grad_norm": 3.1399052143096924, + "learning_rate": 9.774840181500588e-06, + "loss": 0.541, + "num_input_tokens_seen": 7533888, + "step": 28140 + }, + { + "epoch": 7.3774574049803405, + "grad_norm": 3.575221538543701, + "learning_rate": 9.765770119662813e-06, + "loss": 0.3778, + "num_input_tokens_seen": 7535152, + "step": 28145 + }, + { + "epoch": 7.378768020969856, + "grad_norm": 7.567577362060547, + "learning_rate": 9.756703246323026e-06, + "loss": 0.391, + "num_input_tokens_seen": 7536304, + "step": 28150 + }, + { + "epoch": 7.380078636959371, + "grad_norm": 3.021777629852295, + "learning_rate": 9.747639563378916e-06, + "loss": 0.3591, + "num_input_tokens_seen": 7537744, + "step": 28155 + }, + { + "epoch": 7.381389252948886, + "grad_norm": 5.460872173309326, + "learning_rate": 9.738579072727488e-06, + "loss": 0.4409, + "num_input_tokens_seen": 7538864, + "step": 28160 + }, + { + "epoch": 7.382699868938401, + "grad_norm": 0.9456722140312195, + "learning_rate": 9.7295217762651e-06, + "loss": 0.3677, + "num_input_tokens_seen": 7540720, + "step": 28165 + }, + { + "epoch": 7.3840104849279165, + "grad_norm": 2.4899380207061768, + "learning_rate": 9.72046767588741e-06, + "loss": 0.4558, + "num_input_tokens_seen": 7541728, + "step": 28170 + }, + { + "epoch": 7.385321100917431, + "grad_norm": 4.219250202178955, + "learning_rate": 9.711416773489434e-06, + "loss": 0.4545, + "num_input_tokens_seen": 7542928, + "step": 28175 + }, + { + "epoch": 7.386631716906947, + "grad_norm": 1.9928386211395264, + "learning_rate": 9.702369070965512e-06, + "loss": 0.3848, + "num_input_tokens_seen": 7544128, + "step": 28180 + }, + { + "epoch": 7.387942332896461, + "grad_norm": 3.4575929641723633, + "learning_rate": 9.693324570209312e-06, + "loss": 0.3795, + "num_input_tokens_seen": 7545232, + "step": 28185 + }, + { + "epoch": 7.389252948885977, + "grad_norm": 7.54035758972168, + "learning_rate": 9.684283273113829e-06, + "loss": 0.4638, + "num_input_tokens_seen": 7546624, + "step": 28190 + }, + { + "epoch": 7.3905635648754915, + "grad_norm": 4.076102256774902, + "learning_rate": 9.675245181571385e-06, + "loss": 0.5183, + "num_input_tokens_seen": 7548592, + "step": 28195 + }, + { + "epoch": 7.391874180865006, + "grad_norm": 5.190038204193115, + "learning_rate": 9.666210297473663e-06, + "loss": 0.4925, + "num_input_tokens_seen": 7550160, + "step": 28200 + }, + { + "epoch": 7.393184796854522, + "grad_norm": 6.1346354484558105, + "learning_rate": 9.657178622711624e-06, + "loss": 0.5233, + "num_input_tokens_seen": 7551360, + "step": 28205 + }, + { + "epoch": 7.394495412844036, + "grad_norm": 4.002377986907959, + "learning_rate": 9.648150159175587e-06, + "loss": 0.4033, + "num_input_tokens_seen": 7552592, + "step": 28210 + }, + { + "epoch": 7.395806028833552, + "grad_norm": 1.5914322137832642, + "learning_rate": 9.639124908755196e-06, + "loss": 0.3451, + "num_input_tokens_seen": 7553648, + "step": 28215 + }, + { + "epoch": 7.3971166448230665, + "grad_norm": 2.464599847793579, + "learning_rate": 9.630102873339419e-06, + "loss": 0.5022, + "num_input_tokens_seen": 7554912, + "step": 28220 + }, + { + "epoch": 7.398427260812582, + "grad_norm": 2.4604244232177734, + "learning_rate": 9.621084054816557e-06, + "loss": 0.3844, + "num_input_tokens_seen": 7556480, + "step": 28225 + }, + { + "epoch": 7.399737876802097, + "grad_norm": 8.36047649383545, + "learning_rate": 9.612068455074227e-06, + "loss": 0.5099, + "num_input_tokens_seen": 7557856, + "step": 28230 + }, + { + "epoch": 7.401048492791612, + "grad_norm": 2.960749864578247, + "learning_rate": 9.603056075999382e-06, + "loss": 0.343, + "num_input_tokens_seen": 7558880, + "step": 28235 + }, + { + "epoch": 7.402359108781127, + "grad_norm": 2.772076368331909, + "learning_rate": 9.594046919478294e-06, + "loss": 0.5816, + "num_input_tokens_seen": 7559872, + "step": 28240 + }, + { + "epoch": 7.4036697247706424, + "grad_norm": 2.933476448059082, + "learning_rate": 9.58504098739658e-06, + "loss": 0.4829, + "num_input_tokens_seen": 7561104, + "step": 28245 + }, + { + "epoch": 7.404980340760157, + "grad_norm": 2.5902698040008545, + "learning_rate": 9.576038281639138e-06, + "loss": 0.873, + "num_input_tokens_seen": 7562160, + "step": 28250 + }, + { + "epoch": 7.406290956749673, + "grad_norm": 3.161325693130493, + "learning_rate": 9.567038804090227e-06, + "loss": 0.489, + "num_input_tokens_seen": 7563808, + "step": 28255 + }, + { + "epoch": 7.407601572739187, + "grad_norm": 4.429383277893066, + "learning_rate": 9.55804255663344e-06, + "loss": 0.499, + "num_input_tokens_seen": 7565296, + "step": 28260 + }, + { + "epoch": 7.408912188728703, + "grad_norm": 4.365433692932129, + "learning_rate": 9.549049541151659e-06, + "loss": 0.4865, + "num_input_tokens_seen": 7566384, + "step": 28265 + }, + { + "epoch": 7.4102228047182175, + "grad_norm": 4.330374717712402, + "learning_rate": 9.540059759527115e-06, + "loss": 0.5129, + "num_input_tokens_seen": 7567536, + "step": 28270 + }, + { + "epoch": 7.411533420707733, + "grad_norm": 5.197701930999756, + "learning_rate": 9.53107321364135e-06, + "loss": 0.4595, + "num_input_tokens_seen": 7568640, + "step": 28275 + }, + { + "epoch": 7.412844036697248, + "grad_norm": 2.340690851211548, + "learning_rate": 9.522089905375231e-06, + "loss": 0.5136, + "num_input_tokens_seen": 7570272, + "step": 28280 + }, + { + "epoch": 7.414154652686763, + "grad_norm": 9.483389854431152, + "learning_rate": 9.513109836608963e-06, + "loss": 0.353, + "num_input_tokens_seen": 7571424, + "step": 28285 + }, + { + "epoch": 7.415465268676278, + "grad_norm": 3.4824488162994385, + "learning_rate": 9.504133009222035e-06, + "loss": 0.6633, + "num_input_tokens_seen": 7573056, + "step": 28290 + }, + { + "epoch": 7.4167758846657925, + "grad_norm": 3.138538122177124, + "learning_rate": 9.495159425093298e-06, + "loss": 0.4224, + "num_input_tokens_seen": 7574400, + "step": 28295 + }, + { + "epoch": 7.418086500655308, + "grad_norm": 2.2021195888519287, + "learning_rate": 9.4861890861009e-06, + "loss": 0.2566, + "num_input_tokens_seen": 7577312, + "step": 28300 + }, + { + "epoch": 7.419397116644823, + "grad_norm": 3.2003209590911865, + "learning_rate": 9.477221994122321e-06, + "loss": 0.6581, + "num_input_tokens_seen": 7578352, + "step": 28305 + }, + { + "epoch": 7.420707732634338, + "grad_norm": 4.6367292404174805, + "learning_rate": 9.468258151034349e-06, + "loss": 0.4432, + "num_input_tokens_seen": 7579456, + "step": 28310 + }, + { + "epoch": 7.422018348623853, + "grad_norm": 5.431668758392334, + "learning_rate": 9.459297558713115e-06, + "loss": 0.6743, + "num_input_tokens_seen": 7581040, + "step": 28315 + }, + { + "epoch": 7.423328964613368, + "grad_norm": 3.0968542098999023, + "learning_rate": 9.45034021903406e-06, + "loss": 0.3589, + "num_input_tokens_seen": 7582288, + "step": 28320 + }, + { + "epoch": 7.424639580602883, + "grad_norm": 2.4890077114105225, + "learning_rate": 9.441386133871922e-06, + "loss": 0.5535, + "num_input_tokens_seen": 7583728, + "step": 28325 + }, + { + "epoch": 7.425950196592399, + "grad_norm": 1.0060148239135742, + "learning_rate": 9.432435305100782e-06, + "loss": 0.3418, + "num_input_tokens_seen": 7585184, + "step": 28330 + }, + { + "epoch": 7.427260812581913, + "grad_norm": 1.2469213008880615, + "learning_rate": 9.423487734594036e-06, + "loss": 0.4139, + "num_input_tokens_seen": 7586272, + "step": 28335 + }, + { + "epoch": 7.428571428571429, + "grad_norm": 1.6515814065933228, + "learning_rate": 9.41454342422439e-06, + "loss": 0.6514, + "num_input_tokens_seen": 7587712, + "step": 28340 + }, + { + "epoch": 7.4298820445609435, + "grad_norm": 1.8970959186553955, + "learning_rate": 9.405602375863876e-06, + "loss": 0.4382, + "num_input_tokens_seen": 7589120, + "step": 28345 + }, + { + "epoch": 7.431192660550459, + "grad_norm": 2.9997682571411133, + "learning_rate": 9.39666459138384e-06, + "loss": 0.6491, + "num_input_tokens_seen": 7590608, + "step": 28350 + }, + { + "epoch": 7.432503276539974, + "grad_norm": 2.4363489151000977, + "learning_rate": 9.387730072654946e-06, + "loss": 0.4865, + "num_input_tokens_seen": 7591968, + "step": 28355 + }, + { + "epoch": 7.433813892529489, + "grad_norm": 4.6696367263793945, + "learning_rate": 9.378798821547171e-06, + "loss": 0.4616, + "num_input_tokens_seen": 7593456, + "step": 28360 + }, + { + "epoch": 7.435124508519004, + "grad_norm": 2.215759515762329, + "learning_rate": 9.36987083992982e-06, + "loss": 0.3156, + "num_input_tokens_seen": 7594544, + "step": 28365 + }, + { + "epoch": 7.436435124508519, + "grad_norm": 1.1673287153244019, + "learning_rate": 9.36094612967148e-06, + "loss": 0.3916, + "num_input_tokens_seen": 7595760, + "step": 28370 + }, + { + "epoch": 7.437745740498034, + "grad_norm": 1.7086424827575684, + "learning_rate": 9.352024692640098e-06, + "loss": 0.4588, + "num_input_tokens_seen": 7597152, + "step": 28375 + }, + { + "epoch": 7.43905635648755, + "grad_norm": 2.9863598346710205, + "learning_rate": 9.343106530702913e-06, + "loss": 0.4146, + "num_input_tokens_seen": 7598192, + "step": 28380 + }, + { + "epoch": 7.440366972477064, + "grad_norm": 4.187107563018799, + "learning_rate": 9.334191645726475e-06, + "loss": 0.4827, + "num_input_tokens_seen": 7599184, + "step": 28385 + }, + { + "epoch": 7.441677588466579, + "grad_norm": 4.922520637512207, + "learning_rate": 9.325280039576661e-06, + "loss": 0.4759, + "num_input_tokens_seen": 7600256, + "step": 28390 + }, + { + "epoch": 7.442988204456094, + "grad_norm": 2.447265625, + "learning_rate": 9.31637171411865e-06, + "loss": 0.4294, + "num_input_tokens_seen": 7602064, + "step": 28395 + }, + { + "epoch": 7.444298820445609, + "grad_norm": 1.2292448282241821, + "learning_rate": 9.307466671216946e-06, + "loss": 0.4292, + "num_input_tokens_seen": 7603744, + "step": 28400 + }, + { + "epoch": 7.445609436435125, + "grad_norm": 3.040088653564453, + "learning_rate": 9.298564912735347e-06, + "loss": 0.4818, + "num_input_tokens_seen": 7604864, + "step": 28405 + }, + { + "epoch": 7.446920052424639, + "grad_norm": 2.9230294227600098, + "learning_rate": 9.289666440536981e-06, + "loss": 0.3941, + "num_input_tokens_seen": 7606576, + "step": 28410 + }, + { + "epoch": 7.448230668414155, + "grad_norm": 1.9360953569412231, + "learning_rate": 9.280771256484284e-06, + "loss": 0.5587, + "num_input_tokens_seen": 7607536, + "step": 28415 + }, + { + "epoch": 7.4495412844036695, + "grad_norm": 5.680616855621338, + "learning_rate": 9.271879362439e-06, + "loss": 0.5247, + "num_input_tokens_seen": 7608576, + "step": 28420 + }, + { + "epoch": 7.450851900393185, + "grad_norm": 2.339953899383545, + "learning_rate": 9.262990760262183e-06, + "loss": 0.3933, + "num_input_tokens_seen": 7609792, + "step": 28425 + }, + { + "epoch": 7.4521625163827, + "grad_norm": 2.0463814735412598, + "learning_rate": 9.254105451814215e-06, + "loss": 0.3111, + "num_input_tokens_seen": 7610896, + "step": 28430 + }, + { + "epoch": 7.453473132372215, + "grad_norm": 3.740114688873291, + "learning_rate": 9.245223438954781e-06, + "loss": 0.455, + "num_input_tokens_seen": 7612064, + "step": 28435 + }, + { + "epoch": 7.45478374836173, + "grad_norm": 2.703627347946167, + "learning_rate": 9.236344723542847e-06, + "loss": 0.5974, + "num_input_tokens_seen": 7613568, + "step": 28440 + }, + { + "epoch": 7.456094364351245, + "grad_norm": 5.872389316558838, + "learning_rate": 9.227469307436728e-06, + "loss": 0.3241, + "num_input_tokens_seen": 7614784, + "step": 28445 + }, + { + "epoch": 7.45740498034076, + "grad_norm": 2.5291597843170166, + "learning_rate": 9.21859719249403e-06, + "loss": 0.4649, + "num_input_tokens_seen": 7615920, + "step": 28450 + }, + { + "epoch": 7.458715596330276, + "grad_norm": 2.4412312507629395, + "learning_rate": 9.209728380571669e-06, + "loss": 0.7053, + "num_input_tokens_seen": 7617216, + "step": 28455 + }, + { + "epoch": 7.46002621231979, + "grad_norm": 3.8588876724243164, + "learning_rate": 9.200862873525876e-06, + "loss": 0.4298, + "num_input_tokens_seen": 7619168, + "step": 28460 + }, + { + "epoch": 7.461336828309306, + "grad_norm": 7.095777988433838, + "learning_rate": 9.192000673212184e-06, + "loss": 0.3636, + "num_input_tokens_seen": 7620320, + "step": 28465 + }, + { + "epoch": 7.46264744429882, + "grad_norm": 2.3516335487365723, + "learning_rate": 9.183141781485439e-06, + "loss": 0.4904, + "num_input_tokens_seen": 7621456, + "step": 28470 + }, + { + "epoch": 7.463958060288336, + "grad_norm": 3.5134999752044678, + "learning_rate": 9.174286200199786e-06, + "loss": 0.4169, + "num_input_tokens_seen": 7622416, + "step": 28475 + }, + { + "epoch": 7.465268676277851, + "grad_norm": 28.388185501098633, + "learning_rate": 9.165433931208697e-06, + "loss": 0.4726, + "num_input_tokens_seen": 7623792, + "step": 28480 + }, + { + "epoch": 7.466579292267365, + "grad_norm": 2.598766326904297, + "learning_rate": 9.15658497636491e-06, + "loss": 0.682, + "num_input_tokens_seen": 7625440, + "step": 28485 + }, + { + "epoch": 7.467889908256881, + "grad_norm": 2.1648974418640137, + "learning_rate": 9.147739337520519e-06, + "loss": 0.5223, + "num_input_tokens_seen": 7626656, + "step": 28490 + }, + { + "epoch": 7.4692005242463955, + "grad_norm": 6.209239482879639, + "learning_rate": 9.138897016526895e-06, + "loss": 0.3821, + "num_input_tokens_seen": 7627680, + "step": 28495 + }, + { + "epoch": 7.470511140235911, + "grad_norm": 1.7165422439575195, + "learning_rate": 9.130058015234722e-06, + "loss": 0.5894, + "num_input_tokens_seen": 7629152, + "step": 28500 + }, + { + "epoch": 7.471821756225426, + "grad_norm": 4.2501220703125, + "learning_rate": 9.12122233549398e-06, + "loss": 0.4185, + "num_input_tokens_seen": 7630672, + "step": 28505 + }, + { + "epoch": 7.473132372214941, + "grad_norm": 2.5746829509735107, + "learning_rate": 9.11238997915397e-06, + "loss": 0.3823, + "num_input_tokens_seen": 7632208, + "step": 28510 + }, + { + "epoch": 7.474442988204456, + "grad_norm": 4.320060729980469, + "learning_rate": 9.103560948063286e-06, + "loss": 0.4809, + "num_input_tokens_seen": 7633120, + "step": 28515 + }, + { + "epoch": 7.475753604193971, + "grad_norm": 1.4788517951965332, + "learning_rate": 9.094735244069839e-06, + "loss": 0.5656, + "num_input_tokens_seen": 7634544, + "step": 28520 + }, + { + "epoch": 7.477064220183486, + "grad_norm": 3.2148075103759766, + "learning_rate": 9.08591286902081e-06, + "loss": 0.4633, + "num_input_tokens_seen": 7635808, + "step": 28525 + }, + { + "epoch": 7.478374836173002, + "grad_norm": 1.8995238542556763, + "learning_rate": 9.077093824762722e-06, + "loss": 0.3726, + "num_input_tokens_seen": 7637072, + "step": 28530 + }, + { + "epoch": 7.479685452162516, + "grad_norm": 6.532533168792725, + "learning_rate": 9.068278113141385e-06, + "loss": 0.6612, + "num_input_tokens_seen": 7638464, + "step": 28535 + }, + { + "epoch": 7.480996068152032, + "grad_norm": 10.828871726989746, + "learning_rate": 9.059465736001901e-06, + "loss": 0.5291, + "num_input_tokens_seen": 7640416, + "step": 28540 + }, + { + "epoch": 7.482306684141546, + "grad_norm": 2.5511319637298584, + "learning_rate": 9.050656695188706e-06, + "loss": 0.3932, + "num_input_tokens_seen": 7641648, + "step": 28545 + }, + { + "epoch": 7.483617300131062, + "grad_norm": 1.2870523929595947, + "learning_rate": 9.041850992545503e-06, + "loss": 0.3888, + "num_input_tokens_seen": 7643328, + "step": 28550 + }, + { + "epoch": 7.484927916120577, + "grad_norm": 16.011343002319336, + "learning_rate": 9.033048629915326e-06, + "loss": 0.6253, + "num_input_tokens_seen": 7644256, + "step": 28555 + }, + { + "epoch": 7.486238532110092, + "grad_norm": 4.688170909881592, + "learning_rate": 9.02424960914047e-06, + "loss": 0.4712, + "num_input_tokens_seen": 7645280, + "step": 28560 + }, + { + "epoch": 7.487549148099607, + "grad_norm": 3.351672649383545, + "learning_rate": 9.015453932062568e-06, + "loss": 0.5193, + "num_input_tokens_seen": 7648576, + "step": 28565 + }, + { + "epoch": 7.488859764089122, + "grad_norm": 3.6209092140197754, + "learning_rate": 9.006661600522539e-06, + "loss": 0.3494, + "num_input_tokens_seen": 7649952, + "step": 28570 + }, + { + "epoch": 7.490170380078637, + "grad_norm": 13.723076820373535, + "learning_rate": 8.997872616360603e-06, + "loss": 0.3833, + "num_input_tokens_seen": 7651264, + "step": 28575 + }, + { + "epoch": 7.491480996068152, + "grad_norm": 2.7123262882232666, + "learning_rate": 8.989086981416278e-06, + "loss": 0.538, + "num_input_tokens_seen": 7652368, + "step": 28580 + }, + { + "epoch": 7.492791612057667, + "grad_norm": 1.4308421611785889, + "learning_rate": 8.980304697528385e-06, + "loss": 0.33, + "num_input_tokens_seen": 7654000, + "step": 28585 + }, + { + "epoch": 7.494102228047182, + "grad_norm": 3.278214931488037, + "learning_rate": 8.971525766535038e-06, + "loss": 0.4115, + "num_input_tokens_seen": 7655216, + "step": 28590 + }, + { + "epoch": 7.495412844036697, + "grad_norm": 2.2701337337493896, + "learning_rate": 8.962750190273653e-06, + "loss": 0.5219, + "num_input_tokens_seen": 7656384, + "step": 28595 + }, + { + "epoch": 7.496723460026212, + "grad_norm": 10.432271003723145, + "learning_rate": 8.953977970580942e-06, + "loss": 0.4468, + "num_input_tokens_seen": 7657488, + "step": 28600 + }, + { + "epoch": 7.498034076015728, + "grad_norm": 2.888063669204712, + "learning_rate": 8.945209109292919e-06, + "loss": 0.6191, + "num_input_tokens_seen": 7658816, + "step": 28605 + }, + { + "epoch": 7.499344692005242, + "grad_norm": 5.560892581939697, + "learning_rate": 8.936443608244885e-06, + "loss": 0.548, + "num_input_tokens_seen": 7659872, + "step": 28610 + }, + { + "epoch": 7.500655307994758, + "grad_norm": 4.841058731079102, + "learning_rate": 8.927681469271451e-06, + "loss": 0.4807, + "num_input_tokens_seen": 7661040, + "step": 28615 + }, + { + "epoch": 7.501965923984272, + "grad_norm": 2.015261650085449, + "learning_rate": 8.918922694206513e-06, + "loss": 0.5667, + "num_input_tokens_seen": 7662240, + "step": 28620 + }, + { + "epoch": 7.501965923984272, + "eval_loss": 0.46405282616615295, + "eval_runtime": 18.2128, + "eval_samples_per_second": 46.561, + "eval_steps_per_second": 23.28, + "num_input_tokens_seen": 7662240, + "step": 28620 + }, + { + "epoch": 7.503276539973788, + "grad_norm": 2.7298827171325684, + "learning_rate": 8.91016728488327e-06, + "loss": 0.4938, + "num_input_tokens_seen": 7663632, + "step": 28625 + }, + { + "epoch": 7.504587155963303, + "grad_norm": 1.9593621492385864, + "learning_rate": 8.901415243134211e-06, + "loss": 0.5526, + "num_input_tokens_seen": 7665072, + "step": 28630 + }, + { + "epoch": 7.505897771952818, + "grad_norm": 3.6143805980682373, + "learning_rate": 8.892666570791136e-06, + "loss": 0.5314, + "num_input_tokens_seen": 7666416, + "step": 28635 + }, + { + "epoch": 7.507208387942333, + "grad_norm": 4.347865581512451, + "learning_rate": 8.883921269685108e-06, + "loss": 0.3882, + "num_input_tokens_seen": 7668272, + "step": 28640 + }, + { + "epoch": 7.508519003931848, + "grad_norm": 1.4749680757522583, + "learning_rate": 8.875179341646511e-06, + "loss": 0.4554, + "num_input_tokens_seen": 7669200, + "step": 28645 + }, + { + "epoch": 7.509829619921363, + "grad_norm": 9.049057960510254, + "learning_rate": 8.866440788505018e-06, + "loss": 0.4916, + "num_input_tokens_seen": 7671088, + "step": 28650 + }, + { + "epoch": 7.511140235910878, + "grad_norm": 2.767794370651245, + "learning_rate": 8.857705612089581e-06, + "loss": 0.4609, + "num_input_tokens_seen": 7672336, + "step": 28655 + }, + { + "epoch": 7.512450851900393, + "grad_norm": 1.6307090520858765, + "learning_rate": 8.848973814228478e-06, + "loss": 0.3336, + "num_input_tokens_seen": 7673552, + "step": 28660 + }, + { + "epoch": 7.513761467889909, + "grad_norm": 14.659150123596191, + "learning_rate": 8.840245396749247e-06, + "loss": 0.4786, + "num_input_tokens_seen": 7674976, + "step": 28665 + }, + { + "epoch": 7.515072083879423, + "grad_norm": 10.347916603088379, + "learning_rate": 8.831520361478744e-06, + "loss": 0.4763, + "num_input_tokens_seen": 7676192, + "step": 28670 + }, + { + "epoch": 7.516382699868938, + "grad_norm": 10.646052360534668, + "learning_rate": 8.82279871024308e-06, + "loss": 0.5498, + "num_input_tokens_seen": 7677440, + "step": 28675 + }, + { + "epoch": 7.517693315858454, + "grad_norm": 3.135681629180908, + "learning_rate": 8.814080444867695e-06, + "loss": 0.4232, + "num_input_tokens_seen": 7678688, + "step": 28680 + }, + { + "epoch": 7.519003931847968, + "grad_norm": 2.2771377563476562, + "learning_rate": 8.805365567177307e-06, + "loss": 0.4612, + "num_input_tokens_seen": 7680544, + "step": 28685 + }, + { + "epoch": 7.520314547837484, + "grad_norm": 2.554137706756592, + "learning_rate": 8.796654078995921e-06, + "loss": 0.4046, + "num_input_tokens_seen": 7681920, + "step": 28690 + }, + { + "epoch": 7.521625163826998, + "grad_norm": 39.35218048095703, + "learning_rate": 8.787945982146841e-06, + "loss": 1.0551, + "num_input_tokens_seen": 7683216, + "step": 28695 + }, + { + "epoch": 7.522935779816514, + "grad_norm": 3.436063051223755, + "learning_rate": 8.779241278452657e-06, + "loss": 0.4002, + "num_input_tokens_seen": 7684400, + "step": 28700 + }, + { + "epoch": 7.524246395806029, + "grad_norm": 3.952352285385132, + "learning_rate": 8.770539969735241e-06, + "loss": 0.3334, + "num_input_tokens_seen": 7686304, + "step": 28705 + }, + { + "epoch": 7.525557011795544, + "grad_norm": 1.8979518413543701, + "learning_rate": 8.761842057815772e-06, + "loss": 0.4726, + "num_input_tokens_seen": 7687488, + "step": 28710 + }, + { + "epoch": 7.526867627785059, + "grad_norm": 2.4892492294311523, + "learning_rate": 8.753147544514701e-06, + "loss": 0.4149, + "num_input_tokens_seen": 7688688, + "step": 28715 + }, + { + "epoch": 7.528178243774574, + "grad_norm": 2.582181215286255, + "learning_rate": 8.744456431651777e-06, + "loss": 0.475, + "num_input_tokens_seen": 7690176, + "step": 28720 + }, + { + "epoch": 7.529488859764089, + "grad_norm": 1.743487000465393, + "learning_rate": 8.73576872104603e-06, + "loss": 0.4128, + "num_input_tokens_seen": 7691568, + "step": 28725 + }, + { + "epoch": 7.5307994757536045, + "grad_norm": 6.564072132110596, + "learning_rate": 8.727084414515793e-06, + "loss": 0.4507, + "num_input_tokens_seen": 7692624, + "step": 28730 + }, + { + "epoch": 7.532110091743119, + "grad_norm": 1.0844546556472778, + "learning_rate": 8.718403513878668e-06, + "loss": 0.4502, + "num_input_tokens_seen": 7694352, + "step": 28735 + }, + { + "epoch": 7.533420707732635, + "grad_norm": 2.14273738861084, + "learning_rate": 8.709726020951557e-06, + "loss": 0.3036, + "num_input_tokens_seen": 7696144, + "step": 28740 + }, + { + "epoch": 7.534731323722149, + "grad_norm": 2.111417055130005, + "learning_rate": 8.701051937550638e-06, + "loss": 0.5031, + "num_input_tokens_seen": 7697552, + "step": 28745 + }, + { + "epoch": 7.536041939711664, + "grad_norm": 13.264822959899902, + "learning_rate": 8.69238126549139e-06, + "loss": 0.5847, + "num_input_tokens_seen": 7698512, + "step": 28750 + }, + { + "epoch": 7.53735255570118, + "grad_norm": 2.9830965995788574, + "learning_rate": 8.683714006588576e-06, + "loss": 0.4087, + "num_input_tokens_seen": 7699664, + "step": 28755 + }, + { + "epoch": 7.538663171690695, + "grad_norm": 5.019702911376953, + "learning_rate": 8.675050162656215e-06, + "loss": 0.5131, + "num_input_tokens_seen": 7701056, + "step": 28760 + }, + { + "epoch": 7.53997378768021, + "grad_norm": 4.803126335144043, + "learning_rate": 8.666389735507648e-06, + "loss": 0.5884, + "num_input_tokens_seen": 7702256, + "step": 28765 + }, + { + "epoch": 7.541284403669724, + "grad_norm": 2.8477463722229004, + "learning_rate": 8.65773272695548e-06, + "loss": 0.4717, + "num_input_tokens_seen": 7703376, + "step": 28770 + }, + { + "epoch": 7.54259501965924, + "grad_norm": 4.739023685455322, + "learning_rate": 8.649079138811625e-06, + "loss": 0.3925, + "num_input_tokens_seen": 7704896, + "step": 28775 + }, + { + "epoch": 7.543905635648755, + "grad_norm": 2.35905385017395, + "learning_rate": 8.640428972887252e-06, + "loss": 0.5597, + "num_input_tokens_seen": 7706144, + "step": 28780 + }, + { + "epoch": 7.54521625163827, + "grad_norm": 3.3850483894348145, + "learning_rate": 8.63178223099283e-06, + "loss": 0.5111, + "num_input_tokens_seen": 7707552, + "step": 28785 + }, + { + "epoch": 7.546526867627785, + "grad_norm": 6.161430835723877, + "learning_rate": 8.623138914938111e-06, + "loss": 0.4346, + "num_input_tokens_seen": 7708768, + "step": 28790 + }, + { + "epoch": 7.5478374836173, + "grad_norm": 1.9766544103622437, + "learning_rate": 8.61449902653211e-06, + "loss": 0.5582, + "num_input_tokens_seen": 7710208, + "step": 28795 + }, + { + "epoch": 7.549148099606815, + "grad_norm": 6.238855361938477, + "learning_rate": 8.605862567583155e-06, + "loss": 0.3695, + "num_input_tokens_seen": 7711504, + "step": 28800 + }, + { + "epoch": 7.5504587155963305, + "grad_norm": 2.293480634689331, + "learning_rate": 8.597229539898835e-06, + "loss": 0.4275, + "num_input_tokens_seen": 7713008, + "step": 28805 + }, + { + "epoch": 7.551769331585845, + "grad_norm": 2.2571542263031006, + "learning_rate": 8.588599945286033e-06, + "loss": 0.5221, + "num_input_tokens_seen": 7714640, + "step": 28810 + }, + { + "epoch": 7.553079947575361, + "grad_norm": 19.659011840820312, + "learning_rate": 8.579973785550904e-06, + "loss": 0.7159, + "num_input_tokens_seen": 7716352, + "step": 28815 + }, + { + "epoch": 7.554390563564875, + "grad_norm": 3.409639596939087, + "learning_rate": 8.571351062498888e-06, + "loss": 0.5108, + "num_input_tokens_seen": 7717728, + "step": 28820 + }, + { + "epoch": 7.555701179554391, + "grad_norm": 6.983000755310059, + "learning_rate": 8.562731777934713e-06, + "loss": 0.5331, + "num_input_tokens_seen": 7718800, + "step": 28825 + }, + { + "epoch": 7.557011795543906, + "grad_norm": 1.7936627864837646, + "learning_rate": 8.55411593366237e-06, + "loss": 0.3399, + "num_input_tokens_seen": 7720112, + "step": 28830 + }, + { + "epoch": 7.558322411533421, + "grad_norm": 2.657160758972168, + "learning_rate": 8.545503531485149e-06, + "loss": 0.513, + "num_input_tokens_seen": 7721600, + "step": 28835 + }, + { + "epoch": 7.559633027522936, + "grad_norm": 3.5952181816101074, + "learning_rate": 8.536894573205608e-06, + "loss": 0.4096, + "num_input_tokens_seen": 7722624, + "step": 28840 + }, + { + "epoch": 7.56094364351245, + "grad_norm": 1.8023078441619873, + "learning_rate": 8.528289060625586e-06, + "loss": 0.4692, + "num_input_tokens_seen": 7724064, + "step": 28845 + }, + { + "epoch": 7.562254259501966, + "grad_norm": 3.3666951656341553, + "learning_rate": 8.519686995546203e-06, + "loss": 0.3778, + "num_input_tokens_seen": 7725488, + "step": 28850 + }, + { + "epoch": 7.5635648754914815, + "grad_norm": 6.596519470214844, + "learning_rate": 8.511088379767857e-06, + "loss": 0.4831, + "num_input_tokens_seen": 7726864, + "step": 28855 + }, + { + "epoch": 7.564875491480996, + "grad_norm": 2.4717235565185547, + "learning_rate": 8.502493215090218e-06, + "loss": 0.5071, + "num_input_tokens_seen": 7728048, + "step": 28860 + }, + { + "epoch": 7.566186107470511, + "grad_norm": 10.156692504882812, + "learning_rate": 8.493901503312249e-06, + "loss": 0.3634, + "num_input_tokens_seen": 7729264, + "step": 28865 + }, + { + "epoch": 7.567496723460026, + "grad_norm": 5.498894691467285, + "learning_rate": 8.48531324623218e-06, + "loss": 0.6405, + "num_input_tokens_seen": 7730144, + "step": 28870 + }, + { + "epoch": 7.568807339449541, + "grad_norm": 3.8619446754455566, + "learning_rate": 8.476728445647506e-06, + "loss": 0.4117, + "num_input_tokens_seen": 7731392, + "step": 28875 + }, + { + "epoch": 7.5701179554390565, + "grad_norm": 2.4156734943389893, + "learning_rate": 8.468147103355018e-06, + "loss": 0.4411, + "num_input_tokens_seen": 7732640, + "step": 28880 + }, + { + "epoch": 7.571428571428571, + "grad_norm": 5.761648654937744, + "learning_rate": 8.45956922115077e-06, + "loss": 0.5602, + "num_input_tokens_seen": 7733952, + "step": 28885 + }, + { + "epoch": 7.572739187418087, + "grad_norm": 0.8301056027412415, + "learning_rate": 8.450994800830112e-06, + "loss": 0.5658, + "num_input_tokens_seen": 7735488, + "step": 28890 + }, + { + "epoch": 7.574049803407601, + "grad_norm": 5.172355651855469, + "learning_rate": 8.442423844187647e-06, + "loss": 0.708, + "num_input_tokens_seen": 7736720, + "step": 28895 + }, + { + "epoch": 7.575360419397117, + "grad_norm": 2.760969877243042, + "learning_rate": 8.433856353017261e-06, + "loss": 0.4987, + "num_input_tokens_seen": 7737664, + "step": 28900 + }, + { + "epoch": 7.576671035386632, + "grad_norm": 7.665335178375244, + "learning_rate": 8.425292329112124e-06, + "loss": 0.5856, + "num_input_tokens_seen": 7738784, + "step": 28905 + }, + { + "epoch": 7.577981651376147, + "grad_norm": 6.557796955108643, + "learning_rate": 8.416731774264658e-06, + "loss": 0.6102, + "num_input_tokens_seen": 7740240, + "step": 28910 + }, + { + "epoch": 7.579292267365662, + "grad_norm": 2.774282693862915, + "learning_rate": 8.408174690266577e-06, + "loss": 0.4901, + "num_input_tokens_seen": 7741360, + "step": 28915 + }, + { + "epoch": 7.580602883355177, + "grad_norm": 3.6304562091827393, + "learning_rate": 8.399621078908865e-06, + "loss": 0.5044, + "num_input_tokens_seen": 7742400, + "step": 28920 + }, + { + "epoch": 7.581913499344692, + "grad_norm": 3.895494222640991, + "learning_rate": 8.391070941981778e-06, + "loss": 0.2596, + "num_input_tokens_seen": 7743664, + "step": 28925 + }, + { + "epoch": 7.5832241153342075, + "grad_norm": 3.6553285121917725, + "learning_rate": 8.382524281274844e-06, + "loss": 0.2782, + "num_input_tokens_seen": 7744640, + "step": 28930 + }, + { + "epoch": 7.584534731323722, + "grad_norm": 14.076478004455566, + "learning_rate": 8.373981098576865e-06, + "loss": 0.5012, + "num_input_tokens_seen": 7746480, + "step": 28935 + }, + { + "epoch": 7.585845347313237, + "grad_norm": 2.741234064102173, + "learning_rate": 8.365441395675907e-06, + "loss": 0.5644, + "num_input_tokens_seen": 7747552, + "step": 28940 + }, + { + "epoch": 7.587155963302752, + "grad_norm": 2.3303656578063965, + "learning_rate": 8.356905174359338e-06, + "loss": 0.4839, + "num_input_tokens_seen": 7748592, + "step": 28945 + }, + { + "epoch": 7.588466579292267, + "grad_norm": 4.91783332824707, + "learning_rate": 8.348372436413754e-06, + "loss": 0.485, + "num_input_tokens_seen": 7750096, + "step": 28950 + }, + { + "epoch": 7.5897771952817825, + "grad_norm": 2.2409474849700928, + "learning_rate": 8.339843183625046e-06, + "loss": 0.6956, + "num_input_tokens_seen": 7751648, + "step": 28955 + }, + { + "epoch": 7.591087811271297, + "grad_norm": 1.9546600580215454, + "learning_rate": 8.33131741777837e-06, + "loss": 0.5228, + "num_input_tokens_seen": 7752752, + "step": 28960 + }, + { + "epoch": 7.592398427260813, + "grad_norm": 1.4495973587036133, + "learning_rate": 8.32279514065816e-06, + "loss": 0.4222, + "num_input_tokens_seen": 7754000, + "step": 28965 + }, + { + "epoch": 7.593709043250327, + "grad_norm": 2.4985735416412354, + "learning_rate": 8.314276354048109e-06, + "loss": 0.3509, + "num_input_tokens_seen": 7755280, + "step": 28970 + }, + { + "epoch": 7.595019659239843, + "grad_norm": 3.3617706298828125, + "learning_rate": 8.30576105973119e-06, + "loss": 0.4879, + "num_input_tokens_seen": 7756416, + "step": 28975 + }, + { + "epoch": 7.5963302752293576, + "grad_norm": 3.4285342693328857, + "learning_rate": 8.297249259489634e-06, + "loss": 0.5656, + "num_input_tokens_seen": 7757872, + "step": 28980 + }, + { + "epoch": 7.597640891218873, + "grad_norm": 2.926694393157959, + "learning_rate": 8.288740955104946e-06, + "loss": 0.5118, + "num_input_tokens_seen": 7759040, + "step": 28985 + }, + { + "epoch": 7.598951507208388, + "grad_norm": 2.018120527267456, + "learning_rate": 8.280236148357911e-06, + "loss": 0.3319, + "num_input_tokens_seen": 7760304, + "step": 28990 + }, + { + "epoch": 7.600262123197903, + "grad_norm": 2.0697405338287354, + "learning_rate": 8.271734841028553e-06, + "loss": 0.3837, + "num_input_tokens_seen": 7763360, + "step": 28995 + }, + { + "epoch": 7.601572739187418, + "grad_norm": 2.5061872005462646, + "learning_rate": 8.263237034896179e-06, + "loss": 0.3967, + "num_input_tokens_seen": 7764624, + "step": 29000 + }, + { + "epoch": 7.6028833551769335, + "grad_norm": 4.346799373626709, + "learning_rate": 8.254742731739384e-06, + "loss": 0.3747, + "num_input_tokens_seen": 7765984, + "step": 29005 + }, + { + "epoch": 7.604193971166448, + "grad_norm": 2.963078498840332, + "learning_rate": 8.246251933336e-06, + "loss": 0.7989, + "num_input_tokens_seen": 7767040, + "step": 29010 + }, + { + "epoch": 7.605504587155964, + "grad_norm": 8.558320999145508, + "learning_rate": 8.237764641463139e-06, + "loss": 0.8265, + "num_input_tokens_seen": 7767952, + "step": 29015 + }, + { + "epoch": 7.606815203145478, + "grad_norm": 2.3432745933532715, + "learning_rate": 8.229280857897176e-06, + "loss": 0.4679, + "num_input_tokens_seen": 7769680, + "step": 29020 + }, + { + "epoch": 7.608125819134994, + "grad_norm": 3.42234206199646, + "learning_rate": 8.22080058441376e-06, + "loss": 0.6525, + "num_input_tokens_seen": 7771328, + "step": 29025 + }, + { + "epoch": 7.6094364351245085, + "grad_norm": 4.574453353881836, + "learning_rate": 8.212323822787782e-06, + "loss": 0.4572, + "num_input_tokens_seen": 7772688, + "step": 29030 + }, + { + "epoch": 7.610747051114023, + "grad_norm": 5.885719299316406, + "learning_rate": 8.20385057479342e-06, + "loss": 0.4215, + "num_input_tokens_seen": 7774448, + "step": 29035 + }, + { + "epoch": 7.612057667103539, + "grad_norm": 5.380017280578613, + "learning_rate": 8.195380842204112e-06, + "loss": 0.4319, + "num_input_tokens_seen": 7775680, + "step": 29040 + }, + { + "epoch": 7.613368283093053, + "grad_norm": 4.387514114379883, + "learning_rate": 8.186914626792558e-06, + "loss": 0.4778, + "num_input_tokens_seen": 7776928, + "step": 29045 + }, + { + "epoch": 7.614678899082569, + "grad_norm": 3.6007020473480225, + "learning_rate": 8.178451930330719e-06, + "loss": 0.2439, + "num_input_tokens_seen": 7777856, + "step": 29050 + }, + { + "epoch": 7.6159895150720835, + "grad_norm": 2.2460899353027344, + "learning_rate": 8.169992754589819e-06, + "loss": 0.4071, + "num_input_tokens_seen": 7779104, + "step": 29055 + }, + { + "epoch": 7.617300131061599, + "grad_norm": 3.6535167694091797, + "learning_rate": 8.161537101340375e-06, + "loss": 0.4752, + "num_input_tokens_seen": 7780256, + "step": 29060 + }, + { + "epoch": 7.618610747051114, + "grad_norm": 5.23991584777832, + "learning_rate": 8.15308497235211e-06, + "loss": 0.5597, + "num_input_tokens_seen": 7781328, + "step": 29065 + }, + { + "epoch": 7.619921363040629, + "grad_norm": 2.198667526245117, + "learning_rate": 8.144636369394048e-06, + "loss": 0.5706, + "num_input_tokens_seen": 7782416, + "step": 29070 + }, + { + "epoch": 7.621231979030144, + "grad_norm": 2.3739469051361084, + "learning_rate": 8.136191294234474e-06, + "loss": 0.738, + "num_input_tokens_seen": 7783744, + "step": 29075 + }, + { + "epoch": 7.6225425950196595, + "grad_norm": 2.4246115684509277, + "learning_rate": 8.127749748640923e-06, + "loss": 0.4082, + "num_input_tokens_seen": 7785104, + "step": 29080 + }, + { + "epoch": 7.623853211009174, + "grad_norm": 4.890942573547363, + "learning_rate": 8.119311734380197e-06, + "loss": 0.3796, + "num_input_tokens_seen": 7786368, + "step": 29085 + }, + { + "epoch": 7.62516382699869, + "grad_norm": 3.3024725914001465, + "learning_rate": 8.110877253218358e-06, + "loss": 0.359, + "num_input_tokens_seen": 7787520, + "step": 29090 + }, + { + "epoch": 7.626474442988204, + "grad_norm": 1.4706089496612549, + "learning_rate": 8.102446306920725e-06, + "loss": 0.4088, + "num_input_tokens_seen": 7788992, + "step": 29095 + }, + { + "epoch": 7.62778505897772, + "grad_norm": 0.6612501740455627, + "learning_rate": 8.094018897251887e-06, + "loss": 0.3051, + "num_input_tokens_seen": 7790304, + "step": 29100 + }, + { + "epoch": 7.6290956749672345, + "grad_norm": 1.6375762224197388, + "learning_rate": 8.085595025975687e-06, + "loss": 0.3857, + "num_input_tokens_seen": 7791744, + "step": 29105 + }, + { + "epoch": 7.63040629095675, + "grad_norm": 5.777542591094971, + "learning_rate": 8.077174694855216e-06, + "loss": 0.4729, + "num_input_tokens_seen": 7792864, + "step": 29110 + }, + { + "epoch": 7.631716906946265, + "grad_norm": 5.9064130783081055, + "learning_rate": 8.068757905652832e-06, + "loss": 0.4639, + "num_input_tokens_seen": 7794656, + "step": 29115 + }, + { + "epoch": 7.63302752293578, + "grad_norm": 5.972199440002441, + "learning_rate": 8.060344660130172e-06, + "loss": 0.5531, + "num_input_tokens_seen": 7795904, + "step": 29120 + }, + { + "epoch": 7.634338138925295, + "grad_norm": 4.48258113861084, + "learning_rate": 8.051934960048104e-06, + "loss": 0.5527, + "num_input_tokens_seen": 7797008, + "step": 29125 + }, + { + "epoch": 7.6356487549148095, + "grad_norm": 6.621611595153809, + "learning_rate": 8.043528807166767e-06, + "loss": 0.3419, + "num_input_tokens_seen": 7798112, + "step": 29130 + }, + { + "epoch": 7.636959370904325, + "grad_norm": 2.0612635612487793, + "learning_rate": 8.035126203245552e-06, + "loss": 0.4852, + "num_input_tokens_seen": 7799408, + "step": 29135 + }, + { + "epoch": 7.63826998689384, + "grad_norm": 2.4101390838623047, + "learning_rate": 8.026727150043118e-06, + "loss": 0.393, + "num_input_tokens_seen": 7800544, + "step": 29140 + }, + { + "epoch": 7.639580602883355, + "grad_norm": 1.2123245000839233, + "learning_rate": 8.018331649317353e-06, + "loss": 0.3248, + "num_input_tokens_seen": 7801744, + "step": 29145 + }, + { + "epoch": 7.64089121887287, + "grad_norm": 3.3965566158294678, + "learning_rate": 8.009939702825433e-06, + "loss": 0.5883, + "num_input_tokens_seen": 7802992, + "step": 29150 + }, + { + "epoch": 7.6422018348623855, + "grad_norm": 2.149488925933838, + "learning_rate": 8.001551312323776e-06, + "loss": 0.4864, + "num_input_tokens_seen": 7804272, + "step": 29155 + }, + { + "epoch": 7.6435124508519, + "grad_norm": 2.5961508750915527, + "learning_rate": 7.993166479568063e-06, + "loss": 0.436, + "num_input_tokens_seen": 7805696, + "step": 29160 + }, + { + "epoch": 7.644823066841416, + "grad_norm": 3.9217052459716797, + "learning_rate": 7.984785206313213e-06, + "loss": 0.4056, + "num_input_tokens_seen": 7806736, + "step": 29165 + }, + { + "epoch": 7.64613368283093, + "grad_norm": 2.4739041328430176, + "learning_rate": 7.976407494313418e-06, + "loss": 0.495, + "num_input_tokens_seen": 7807888, + "step": 29170 + }, + { + "epoch": 7.647444298820446, + "grad_norm": 5.169532775878906, + "learning_rate": 7.968033345322123e-06, + "loss": 0.3269, + "num_input_tokens_seen": 7808896, + "step": 29175 + }, + { + "epoch": 7.6487549148099605, + "grad_norm": 10.813941955566406, + "learning_rate": 7.95966276109203e-06, + "loss": 0.436, + "num_input_tokens_seen": 7809888, + "step": 29180 + }, + { + "epoch": 7.650065530799476, + "grad_norm": 3.1217904090881348, + "learning_rate": 7.95129574337507e-06, + "loss": 0.5291, + "num_input_tokens_seen": 7811296, + "step": 29185 + }, + { + "epoch": 7.651376146788991, + "grad_norm": 3.1848390102386475, + "learning_rate": 7.942932293922453e-06, + "loss": 0.6967, + "num_input_tokens_seen": 7812496, + "step": 29190 + }, + { + "epoch": 7.652686762778506, + "grad_norm": 6.478772163391113, + "learning_rate": 7.934572414484634e-06, + "loss": 0.5261, + "num_input_tokens_seen": 7813904, + "step": 29195 + }, + { + "epoch": 7.653997378768021, + "grad_norm": 2.271134376525879, + "learning_rate": 7.926216106811321e-06, + "loss": 0.3671, + "num_input_tokens_seen": 7815280, + "step": 29200 + }, + { + "epoch": 7.655307994757536, + "grad_norm": 3.4848246574401855, + "learning_rate": 7.917863372651477e-06, + "loss": 0.4733, + "num_input_tokens_seen": 7816496, + "step": 29205 + }, + { + "epoch": 7.656618610747051, + "grad_norm": 2.4469261169433594, + "learning_rate": 7.909514213753314e-06, + "loss": 0.471, + "num_input_tokens_seen": 7817856, + "step": 29210 + }, + { + "epoch": 7.657929226736567, + "grad_norm": 2.967446804046631, + "learning_rate": 7.901168631864294e-06, + "loss": 0.3991, + "num_input_tokens_seen": 7819312, + "step": 29215 + }, + { + "epoch": 7.659239842726081, + "grad_norm": 1.9235938787460327, + "learning_rate": 7.892826628731135e-06, + "loss": 0.4287, + "num_input_tokens_seen": 7820912, + "step": 29220 + }, + { + "epoch": 7.660550458715596, + "grad_norm": 9.639278411865234, + "learning_rate": 7.884488206099814e-06, + "loss": 0.5854, + "num_input_tokens_seen": 7822080, + "step": 29225 + }, + { + "epoch": 7.6618610747051115, + "grad_norm": 2.6375925540924072, + "learning_rate": 7.876153365715519e-06, + "loss": 0.4996, + "num_input_tokens_seen": 7823376, + "step": 29230 + }, + { + "epoch": 7.663171690694626, + "grad_norm": 2.4145538806915283, + "learning_rate": 7.867822109322747e-06, + "loss": 0.4254, + "num_input_tokens_seen": 7824832, + "step": 29235 + }, + { + "epoch": 7.664482306684142, + "grad_norm": 1.7911522388458252, + "learning_rate": 7.859494438665204e-06, + "loss": 0.4841, + "num_input_tokens_seen": 7826288, + "step": 29240 + }, + { + "epoch": 7.665792922673656, + "grad_norm": 7.844268321990967, + "learning_rate": 7.85117035548586e-06, + "loss": 0.5813, + "num_input_tokens_seen": 7827664, + "step": 29245 + }, + { + "epoch": 7.667103538663172, + "grad_norm": 6.371004581451416, + "learning_rate": 7.842849861526929e-06, + "loss": 0.4562, + "num_input_tokens_seen": 7829392, + "step": 29250 + }, + { + "epoch": 7.6684141546526865, + "grad_norm": 2.167337656021118, + "learning_rate": 7.834532958529878e-06, + "loss": 0.3393, + "num_input_tokens_seen": 7830624, + "step": 29255 + }, + { + "epoch": 7.669724770642202, + "grad_norm": 3.104269504547119, + "learning_rate": 7.826219648235424e-06, + "loss": 0.5137, + "num_input_tokens_seen": 7831712, + "step": 29260 + }, + { + "epoch": 7.671035386631717, + "grad_norm": 6.761919975280762, + "learning_rate": 7.817909932383519e-06, + "loss": 0.3255, + "num_input_tokens_seen": 7832880, + "step": 29265 + }, + { + "epoch": 7.672346002621232, + "grad_norm": 10.962610244750977, + "learning_rate": 7.809603812713373e-06, + "loss": 0.5488, + "num_input_tokens_seen": 7833888, + "step": 29270 + }, + { + "epoch": 7.673656618610747, + "grad_norm": 3.4182353019714355, + "learning_rate": 7.80130129096345e-06, + "loss": 0.5232, + "num_input_tokens_seen": 7835376, + "step": 29275 + }, + { + "epoch": 7.674967234600262, + "grad_norm": 4.720802307128906, + "learning_rate": 7.793002368871446e-06, + "loss": 0.3927, + "num_input_tokens_seen": 7836560, + "step": 29280 + }, + { + "epoch": 7.676277850589777, + "grad_norm": 3.377849578857422, + "learning_rate": 7.78470704817431e-06, + "loss": 0.5317, + "num_input_tokens_seen": 7838000, + "step": 29285 + }, + { + "epoch": 7.677588466579293, + "grad_norm": 3.6014342308044434, + "learning_rate": 7.77641533060825e-06, + "loss": 0.3818, + "num_input_tokens_seen": 7839376, + "step": 29290 + }, + { + "epoch": 7.678899082568807, + "grad_norm": 2.1975622177124023, + "learning_rate": 7.76812721790871e-06, + "loss": 0.4562, + "num_input_tokens_seen": 7840608, + "step": 29295 + }, + { + "epoch": 7.680209698558322, + "grad_norm": 6.184089183807373, + "learning_rate": 7.759842711810361e-06, + "loss": 0.5827, + "num_input_tokens_seen": 7841952, + "step": 29300 + }, + { + "epoch": 7.6815203145478375, + "grad_norm": 2.18989634513855, + "learning_rate": 7.751561814047142e-06, + "loss": 0.46, + "num_input_tokens_seen": 7843424, + "step": 29305 + }, + { + "epoch": 7.682830930537353, + "grad_norm": 2.1911332607269287, + "learning_rate": 7.743284526352235e-06, + "loss": 0.4715, + "num_input_tokens_seen": 7844496, + "step": 29310 + }, + { + "epoch": 7.684141546526868, + "grad_norm": 5.079965591430664, + "learning_rate": 7.735010850458057e-06, + "loss": 0.3093, + "num_input_tokens_seen": 7845616, + "step": 29315 + }, + { + "epoch": 7.685452162516382, + "grad_norm": 2.6317429542541504, + "learning_rate": 7.726740788096276e-06, + "loss": 0.5594, + "num_input_tokens_seen": 7847104, + "step": 29320 + }, + { + "epoch": 7.686762778505898, + "grad_norm": 3.8152754306793213, + "learning_rate": 7.718474340997804e-06, + "loss": 0.5325, + "num_input_tokens_seen": 7848432, + "step": 29325 + }, + { + "epoch": 7.6880733944954125, + "grad_norm": 1.54348886013031, + "learning_rate": 7.710211510892792e-06, + "loss": 0.3529, + "num_input_tokens_seen": 7849888, + "step": 29330 + }, + { + "epoch": 7.689384010484928, + "grad_norm": 8.189632415771484, + "learning_rate": 7.701952299510634e-06, + "loss": 0.4616, + "num_input_tokens_seen": 7850928, + "step": 29335 + }, + { + "epoch": 7.690694626474443, + "grad_norm": 5.154665946960449, + "learning_rate": 7.69369670857998e-06, + "loss": 0.4901, + "num_input_tokens_seen": 7852176, + "step": 29340 + }, + { + "epoch": 7.692005242463958, + "grad_norm": 2.023052930831909, + "learning_rate": 7.685444739828688e-06, + "loss": 0.4014, + "num_input_tokens_seen": 7853328, + "step": 29345 + }, + { + "epoch": 7.693315858453473, + "grad_norm": 1.4958281517028809, + "learning_rate": 7.677196394983902e-06, + "loss": 0.466, + "num_input_tokens_seen": 7854416, + "step": 29350 + }, + { + "epoch": 7.694626474442988, + "grad_norm": 5.3421502113342285, + "learning_rate": 7.668951675771979e-06, + "loss": 0.4404, + "num_input_tokens_seen": 7856176, + "step": 29355 + }, + { + "epoch": 7.695937090432503, + "grad_norm": 18.301515579223633, + "learning_rate": 7.660710583918524e-06, + "loss": 0.417, + "num_input_tokens_seen": 7857200, + "step": 29360 + }, + { + "epoch": 7.697247706422019, + "grad_norm": 2.525092601776123, + "learning_rate": 7.652473121148383e-06, + "loss": 0.6537, + "num_input_tokens_seen": 7858416, + "step": 29365 + }, + { + "epoch": 7.698558322411533, + "grad_norm": 5.472678184509277, + "learning_rate": 7.644239289185644e-06, + "loss": 0.51, + "num_input_tokens_seen": 7859904, + "step": 29370 + }, + { + "epoch": 7.699868938401049, + "grad_norm": 6.782070636749268, + "learning_rate": 7.636009089753643e-06, + "loss": 0.5557, + "num_input_tokens_seen": 7861472, + "step": 29375 + }, + { + "epoch": 7.7011795543905635, + "grad_norm": 0.6332657933235168, + "learning_rate": 7.627782524574925e-06, + "loss": 0.5642, + "num_input_tokens_seen": 7864400, + "step": 29380 + }, + { + "epoch": 7.702490170380079, + "grad_norm": 3.8451685905456543, + "learning_rate": 7.619559595371312e-06, + "loss": 0.4547, + "num_input_tokens_seen": 7865840, + "step": 29385 + }, + { + "epoch": 7.703800786369594, + "grad_norm": 6.253728866577148, + "learning_rate": 7.611340303863843e-06, + "loss": 0.5625, + "num_input_tokens_seen": 7866944, + "step": 29390 + }, + { + "epoch": 7.705111402359108, + "grad_norm": 3.7546098232269287, + "learning_rate": 7.603124651772803e-06, + "loss": 0.4397, + "num_input_tokens_seen": 7868384, + "step": 29395 + }, + { + "epoch": 7.706422018348624, + "grad_norm": 2.631671667098999, + "learning_rate": 7.594912640817705e-06, + "loss": 0.4537, + "num_input_tokens_seen": 7869952, + "step": 29400 + }, + { + "epoch": 7.707732634338139, + "grad_norm": 2.981036901473999, + "learning_rate": 7.586704272717327e-06, + "loss": 0.4061, + "num_input_tokens_seen": 7871312, + "step": 29405 + }, + { + "epoch": 7.709043250327654, + "grad_norm": 5.20271110534668, + "learning_rate": 7.578499549189655e-06, + "loss": 0.5352, + "num_input_tokens_seen": 7872704, + "step": 29410 + }, + { + "epoch": 7.710353866317169, + "grad_norm": 4.173034191131592, + "learning_rate": 7.570298471951934e-06, + "loss": 0.6548, + "num_input_tokens_seen": 7873680, + "step": 29415 + }, + { + "epoch": 7.711664482306684, + "grad_norm": 1.0877888202667236, + "learning_rate": 7.562101042720621e-06, + "loss": 0.3557, + "num_input_tokens_seen": 7875552, + "step": 29420 + }, + { + "epoch": 7.712975098296199, + "grad_norm": 10.276432991027832, + "learning_rate": 7.553907263211427e-06, + "loss": 0.7647, + "num_input_tokens_seen": 7876976, + "step": 29425 + }, + { + "epoch": 7.714285714285714, + "grad_norm": 3.809394598007202, + "learning_rate": 7.545717135139299e-06, + "loss": 0.4809, + "num_input_tokens_seen": 7878128, + "step": 29430 + }, + { + "epoch": 7.715596330275229, + "grad_norm": 5.723487854003906, + "learning_rate": 7.537530660218414e-06, + "loss": 0.4273, + "num_input_tokens_seen": 7879728, + "step": 29435 + }, + { + "epoch": 7.716906946264745, + "grad_norm": 5.035701751708984, + "learning_rate": 7.529347840162193e-06, + "loss": 0.6173, + "num_input_tokens_seen": 7881152, + "step": 29440 + }, + { + "epoch": 7.718217562254259, + "grad_norm": 2.063490629196167, + "learning_rate": 7.52116867668328e-06, + "loss": 0.2947, + "num_input_tokens_seen": 7882416, + "step": 29445 + }, + { + "epoch": 7.719528178243775, + "grad_norm": 3.424621105194092, + "learning_rate": 7.51299317149356e-06, + "loss": 0.5285, + "num_input_tokens_seen": 7883712, + "step": 29450 + }, + { + "epoch": 7.7208387942332894, + "grad_norm": 1.7821861505508423, + "learning_rate": 7.504821326304157e-06, + "loss": 0.4705, + "num_input_tokens_seen": 7885056, + "step": 29455 + }, + { + "epoch": 7.722149410222805, + "grad_norm": 3.249748945236206, + "learning_rate": 7.496653142825419e-06, + "loss": 0.3701, + "num_input_tokens_seen": 7886688, + "step": 29460 + }, + { + "epoch": 7.72346002621232, + "grad_norm": 2.6365244388580322, + "learning_rate": 7.4884886227669375e-06, + "loss": 0.5464, + "num_input_tokens_seen": 7887776, + "step": 29465 + }, + { + "epoch": 7.724770642201835, + "grad_norm": 4.543306827545166, + "learning_rate": 7.480327767837525e-06, + "loss": 0.3404, + "num_input_tokens_seen": 7889408, + "step": 29470 + }, + { + "epoch": 7.72608125819135, + "grad_norm": 3.873034954071045, + "learning_rate": 7.472170579745244e-06, + "loss": 0.513, + "num_input_tokens_seen": 7890368, + "step": 29475 + }, + { + "epoch": 7.727391874180865, + "grad_norm": 5.893413066864014, + "learning_rate": 7.464017060197373e-06, + "loss": 0.5008, + "num_input_tokens_seen": 7893680, + "step": 29480 + }, + { + "epoch": 7.72870249017038, + "grad_norm": 2.4016520977020264, + "learning_rate": 7.45586721090043e-06, + "loss": 0.4861, + "num_input_tokens_seen": 7895008, + "step": 29485 + }, + { + "epoch": 7.730013106159895, + "grad_norm": 1.7071523666381836, + "learning_rate": 7.447721033560168e-06, + "loss": 0.5605, + "num_input_tokens_seen": 7896624, + "step": 29490 + }, + { + "epoch": 7.73132372214941, + "grad_norm": 4.641177177429199, + "learning_rate": 7.439578529881572e-06, + "loss": 0.4731, + "num_input_tokens_seen": 7897904, + "step": 29495 + }, + { + "epoch": 7.732634338138926, + "grad_norm": 1.5100449323654175, + "learning_rate": 7.431439701568841e-06, + "loss": 0.3197, + "num_input_tokens_seen": 7899200, + "step": 29500 + }, + { + "epoch": 7.73394495412844, + "grad_norm": 2.587829828262329, + "learning_rate": 7.423304550325422e-06, + "loss": 0.3892, + "num_input_tokens_seen": 7900560, + "step": 29505 + }, + { + "epoch": 7.735255570117955, + "grad_norm": 2.6369354724884033, + "learning_rate": 7.415173077853993e-06, + "loss": 0.4088, + "num_input_tokens_seen": 7902000, + "step": 29510 + }, + { + "epoch": 7.736566186107471, + "grad_norm": 1.3513611555099487, + "learning_rate": 7.407045285856445e-06, + "loss": 0.4604, + "num_input_tokens_seen": 7903904, + "step": 29515 + }, + { + "epoch": 7.737876802096985, + "grad_norm": 2.3384287357330322, + "learning_rate": 7.39892117603393e-06, + "loss": 0.4941, + "num_input_tokens_seen": 7905648, + "step": 29520 + }, + { + "epoch": 7.739187418086501, + "grad_norm": 4.198498249053955, + "learning_rate": 7.390800750086799e-06, + "loss": 0.5202, + "num_input_tokens_seen": 7906880, + "step": 29525 + }, + { + "epoch": 7.740498034076015, + "grad_norm": 2.9462344646453857, + "learning_rate": 7.382684009714652e-06, + "loss": 0.3564, + "num_input_tokens_seen": 7908336, + "step": 29530 + }, + { + "epoch": 7.741808650065531, + "grad_norm": 2.7424120903015137, + "learning_rate": 7.374570956616295e-06, + "loss": 0.5099, + "num_input_tokens_seen": 7909888, + "step": 29535 + }, + { + "epoch": 7.743119266055046, + "grad_norm": 5.171253681182861, + "learning_rate": 7.366461592489782e-06, + "loss": 0.4667, + "num_input_tokens_seen": 7910992, + "step": 29540 + }, + { + "epoch": 7.744429882044561, + "grad_norm": 3.1083755493164062, + "learning_rate": 7.358355919032389e-06, + "loss": 0.266, + "num_input_tokens_seen": 7912176, + "step": 29545 + }, + { + "epoch": 7.745740498034076, + "grad_norm": 1.9074985980987549, + "learning_rate": 7.350253937940621e-06, + "loss": 0.4424, + "num_input_tokens_seen": 7913776, + "step": 29550 + }, + { + "epoch": 7.747051114023591, + "grad_norm": 2.822066307067871, + "learning_rate": 7.342155650910207e-06, + "loss": 0.4461, + "num_input_tokens_seen": 7915680, + "step": 29555 + }, + { + "epoch": 7.748361730013106, + "grad_norm": 1.5945488214492798, + "learning_rate": 7.334061059636102e-06, + "loss": 0.3004, + "num_input_tokens_seen": 7916944, + "step": 29560 + }, + { + "epoch": 7.749672346002622, + "grad_norm": 2.981483221054077, + "learning_rate": 7.325970165812496e-06, + "loss": 0.4636, + "num_input_tokens_seen": 7917968, + "step": 29565 + }, + { + "epoch": 7.750982961992136, + "grad_norm": 1.2773809432983398, + "learning_rate": 7.3178829711327945e-06, + "loss": 0.3352, + "num_input_tokens_seen": 7919520, + "step": 29570 + }, + { + "epoch": 7.752293577981652, + "grad_norm": 4.6406683921813965, + "learning_rate": 7.309799477289636e-06, + "loss": 0.5834, + "num_input_tokens_seen": 7920704, + "step": 29575 + }, + { + "epoch": 7.753604193971166, + "grad_norm": 5.359013080596924, + "learning_rate": 7.301719685974881e-06, + "loss": 0.4198, + "num_input_tokens_seen": 7921984, + "step": 29580 + }, + { + "epoch": 7.754914809960681, + "grad_norm": 1.518312692642212, + "learning_rate": 7.293643598879612e-06, + "loss": 0.3348, + "num_input_tokens_seen": 7923504, + "step": 29585 + }, + { + "epoch": 7.756225425950197, + "grad_norm": 2.371727466583252, + "learning_rate": 7.285571217694148e-06, + "loss": 0.3519, + "num_input_tokens_seen": 7924592, + "step": 29590 + }, + { + "epoch": 7.757536041939712, + "grad_norm": 4.385661602020264, + "learning_rate": 7.277502544108017e-06, + "loss": 0.407, + "num_input_tokens_seen": 7925872, + "step": 29595 + }, + { + "epoch": 7.758846657929227, + "grad_norm": 2.2127346992492676, + "learning_rate": 7.2694375798099815e-06, + "loss": 0.3386, + "num_input_tokens_seen": 7927344, + "step": 29600 + }, + { + "epoch": 7.760157273918741, + "grad_norm": 2.0371389389038086, + "learning_rate": 7.261376326488026e-06, + "loss": 0.4193, + "num_input_tokens_seen": 7928928, + "step": 29605 + }, + { + "epoch": 7.761467889908257, + "grad_norm": 0.8981499671936035, + "learning_rate": 7.253318785829363e-06, + "loss": 0.5443, + "num_input_tokens_seen": 7930544, + "step": 29610 + }, + { + "epoch": 7.762778505897772, + "grad_norm": 1.7751870155334473, + "learning_rate": 7.245264959520406e-06, + "loss": 0.3182, + "num_input_tokens_seen": 7931680, + "step": 29615 + }, + { + "epoch": 7.764089121887287, + "grad_norm": 3.7137303352355957, + "learning_rate": 7.237214849246815e-06, + "loss": 0.4919, + "num_input_tokens_seen": 7932960, + "step": 29620 + }, + { + "epoch": 7.765399737876802, + "grad_norm": 2.4290077686309814, + "learning_rate": 7.2291684566934634e-06, + "loss": 0.2705, + "num_input_tokens_seen": 7935040, + "step": 29625 + }, + { + "epoch": 7.766710353866317, + "grad_norm": 1.9749490022659302, + "learning_rate": 7.221125783544441e-06, + "loss": 0.4967, + "num_input_tokens_seen": 7936288, + "step": 29630 + }, + { + "epoch": 7.768020969855832, + "grad_norm": 1.3277404308319092, + "learning_rate": 7.213086831483082e-06, + "loss": 0.5024, + "num_input_tokens_seen": 7937584, + "step": 29635 + }, + { + "epoch": 7.7693315858453476, + "grad_norm": 1.413864016532898, + "learning_rate": 7.205051602191914e-06, + "loss": 0.3177, + "num_input_tokens_seen": 7938960, + "step": 29640 + }, + { + "epoch": 7.770642201834862, + "grad_norm": 5.023472309112549, + "learning_rate": 7.1970200973527e-06, + "loss": 0.4742, + "num_input_tokens_seen": 7940320, + "step": 29645 + }, + { + "epoch": 7.771952817824378, + "grad_norm": 2.0058815479278564, + "learning_rate": 7.188992318646423e-06, + "loss": 0.5378, + "num_input_tokens_seen": 7941584, + "step": 29650 + }, + { + "epoch": 7.773263433813892, + "grad_norm": 4.572968482971191, + "learning_rate": 7.180968267753274e-06, + "loss": 0.6415, + "num_input_tokens_seen": 7942608, + "step": 29655 + }, + { + "epoch": 7.774574049803408, + "grad_norm": 3.2345330715179443, + "learning_rate": 7.172947946352676e-06, + "loss": 0.6418, + "num_input_tokens_seen": 7943952, + "step": 29660 + }, + { + "epoch": 7.775884665792923, + "grad_norm": 3.105570077896118, + "learning_rate": 7.164931356123269e-06, + "loss": 0.5224, + "num_input_tokens_seen": 7944992, + "step": 29665 + }, + { + "epoch": 7.777195281782438, + "grad_norm": 2.902604818344116, + "learning_rate": 7.156918498742915e-06, + "loss": 0.3713, + "num_input_tokens_seen": 7946160, + "step": 29670 + }, + { + "epoch": 7.778505897771953, + "grad_norm": 1.8661247491836548, + "learning_rate": 7.148909375888688e-06, + "loss": 0.456, + "num_input_tokens_seen": 7947504, + "step": 29675 + }, + { + "epoch": 7.779816513761467, + "grad_norm": 2.090345859527588, + "learning_rate": 7.140903989236886e-06, + "loss": 0.5406, + "num_input_tokens_seen": 7948688, + "step": 29680 + }, + { + "epoch": 7.781127129750983, + "grad_norm": 4.320595741271973, + "learning_rate": 7.132902340463018e-06, + "loss": 0.5629, + "num_input_tokens_seen": 7950000, + "step": 29685 + }, + { + "epoch": 7.7824377457404985, + "grad_norm": 5.849262237548828, + "learning_rate": 7.124904431241819e-06, + "loss": 0.3777, + "num_input_tokens_seen": 7951312, + "step": 29690 + }, + { + "epoch": 7.783748361730013, + "grad_norm": 2.2469565868377686, + "learning_rate": 7.116910263247237e-06, + "loss": 0.2912, + "num_input_tokens_seen": 7952752, + "step": 29695 + }, + { + "epoch": 7.785058977719528, + "grad_norm": 6.771210193634033, + "learning_rate": 7.108919838152436e-06, + "loss": 0.3343, + "num_input_tokens_seen": 7953792, + "step": 29700 + }, + { + "epoch": 7.786369593709043, + "grad_norm": 8.794093132019043, + "learning_rate": 7.100933157629802e-06, + "loss": 0.4947, + "num_input_tokens_seen": 7955312, + "step": 29705 + }, + { + "epoch": 7.787680209698558, + "grad_norm": 4.784348964691162, + "learning_rate": 7.09295022335093e-06, + "loss": 0.6148, + "num_input_tokens_seen": 7956432, + "step": 29710 + }, + { + "epoch": 7.7889908256880735, + "grad_norm": 2.918259382247925, + "learning_rate": 7.0849710369866345e-06, + "loss": 0.4313, + "num_input_tokens_seen": 7957600, + "step": 29715 + }, + { + "epoch": 7.790301441677588, + "grad_norm": 5.733238220214844, + "learning_rate": 7.076995600206948e-06, + "loss": 0.6284, + "num_input_tokens_seen": 7958720, + "step": 29720 + }, + { + "epoch": 7.791612057667104, + "grad_norm": 2.8009581565856934, + "learning_rate": 7.069023914681111e-06, + "loss": 0.4814, + "num_input_tokens_seen": 7959888, + "step": 29725 + }, + { + "epoch": 7.792922673656618, + "grad_norm": 3.7674014568328857, + "learning_rate": 7.0610559820775965e-06, + "loss": 0.2686, + "num_input_tokens_seen": 7961280, + "step": 29730 + }, + { + "epoch": 7.794233289646134, + "grad_norm": 1.4231586456298828, + "learning_rate": 7.05309180406406e-06, + "loss": 0.3728, + "num_input_tokens_seen": 7962624, + "step": 29735 + }, + { + "epoch": 7.795543905635649, + "grad_norm": 5.490416526794434, + "learning_rate": 7.045131382307399e-06, + "loss": 0.6355, + "num_input_tokens_seen": 7964048, + "step": 29740 + }, + { + "epoch": 7.796854521625164, + "grad_norm": 5.40217399597168, + "learning_rate": 7.0371747184737075e-06, + "loss": 0.3569, + "num_input_tokens_seen": 7965216, + "step": 29745 + }, + { + "epoch": 7.798165137614679, + "grad_norm": 1.7175720930099487, + "learning_rate": 7.02922181422832e-06, + "loss": 0.6162, + "num_input_tokens_seen": 7966384, + "step": 29750 + }, + { + "epoch": 7.799475753604194, + "grad_norm": 2.8676958084106445, + "learning_rate": 7.021272671235754e-06, + "loss": 0.4284, + "num_input_tokens_seen": 7967888, + "step": 29755 + }, + { + "epoch": 7.800786369593709, + "grad_norm": 3.211277961730957, + "learning_rate": 7.013327291159755e-06, + "loss": 0.3066, + "num_input_tokens_seen": 7969008, + "step": 29760 + }, + { + "epoch": 7.8020969855832245, + "grad_norm": 6.053005218505859, + "learning_rate": 7.0053856756632815e-06, + "loss": 0.4387, + "num_input_tokens_seen": 7970448, + "step": 29765 + }, + { + "epoch": 7.803407601572739, + "grad_norm": 13.46610164642334, + "learning_rate": 6.997447826408485e-06, + "loss": 0.3961, + "num_input_tokens_seen": 7972064, + "step": 29770 + }, + { + "epoch": 7.804718217562254, + "grad_norm": 2.664911985397339, + "learning_rate": 6.989513745056753e-06, + "loss": 0.6539, + "num_input_tokens_seen": 7973440, + "step": 29775 + }, + { + "epoch": 7.806028833551769, + "grad_norm": 3.3076815605163574, + "learning_rate": 6.981583433268676e-06, + "loss": 0.4514, + "num_input_tokens_seen": 7974944, + "step": 29780 + }, + { + "epoch": 7.807339449541285, + "grad_norm": 2.801750421524048, + "learning_rate": 6.97365689270405e-06, + "loss": 0.4296, + "num_input_tokens_seen": 7976240, + "step": 29785 + }, + { + "epoch": 7.8086500655307995, + "grad_norm": 1.8777276277542114, + "learning_rate": 6.965734125021892e-06, + "loss": 0.3463, + "num_input_tokens_seen": 7977616, + "step": 29790 + }, + { + "epoch": 7.809960681520314, + "grad_norm": 3.556595802307129, + "learning_rate": 6.957815131880421e-06, + "loss": 0.4385, + "num_input_tokens_seen": 7979056, + "step": 29795 + }, + { + "epoch": 7.81127129750983, + "grad_norm": 1.2670748233795166, + "learning_rate": 6.949899914937066e-06, + "loss": 0.544, + "num_input_tokens_seen": 7980784, + "step": 29800 + }, + { + "epoch": 7.812581913499344, + "grad_norm": 7.757167816162109, + "learning_rate": 6.941988475848474e-06, + "loss": 0.4296, + "num_input_tokens_seen": 7981952, + "step": 29805 + }, + { + "epoch": 7.81389252948886, + "grad_norm": 3.599581003189087, + "learning_rate": 6.934080816270488e-06, + "loss": 0.6521, + "num_input_tokens_seen": 7983792, + "step": 29810 + }, + { + "epoch": 7.815203145478375, + "grad_norm": 7.453869342803955, + "learning_rate": 6.926176937858175e-06, + "loss": 0.3656, + "num_input_tokens_seen": 7984784, + "step": 29815 + }, + { + "epoch": 7.81651376146789, + "grad_norm": 1.9235111474990845, + "learning_rate": 6.9182768422658e-06, + "loss": 0.274, + "num_input_tokens_seen": 7986000, + "step": 29820 + }, + { + "epoch": 7.817824377457405, + "grad_norm": 1.441921591758728, + "learning_rate": 6.910380531146837e-06, + "loss": 0.5189, + "num_input_tokens_seen": 7987680, + "step": 29825 + }, + { + "epoch": 7.81913499344692, + "grad_norm": 17.88463592529297, + "learning_rate": 6.902488006153973e-06, + "loss": 0.5047, + "num_input_tokens_seen": 7988832, + "step": 29830 + }, + { + "epoch": 7.820445609436435, + "grad_norm": 1.4093828201293945, + "learning_rate": 6.894599268939101e-06, + "loss": 0.356, + "num_input_tokens_seen": 7990016, + "step": 29835 + }, + { + "epoch": 7.8217562254259505, + "grad_norm": 4.259943962097168, + "learning_rate": 6.886714321153315e-06, + "loss": 0.3992, + "num_input_tokens_seen": 7991312, + "step": 29840 + }, + { + "epoch": 7.823066841415465, + "grad_norm": 2.3754265308380127, + "learning_rate": 6.8788331644469265e-06, + "loss": 0.3449, + "num_input_tokens_seen": 7992848, + "step": 29845 + }, + { + "epoch": 7.824377457404981, + "grad_norm": 9.903606414794922, + "learning_rate": 6.870955800469453e-06, + "loss": 0.4862, + "num_input_tokens_seen": 7993856, + "step": 29850 + }, + { + "epoch": 7.825688073394495, + "grad_norm": 5.480794429779053, + "learning_rate": 6.863082230869597e-06, + "loss": 0.5046, + "num_input_tokens_seen": 7995040, + "step": 29855 + }, + { + "epoch": 7.826998689384011, + "grad_norm": 5.291403770446777, + "learning_rate": 6.855212457295285e-06, + "loss": 0.3872, + "num_input_tokens_seen": 7996112, + "step": 29860 + }, + { + "epoch": 7.8283093053735255, + "grad_norm": 4.089907169342041, + "learning_rate": 6.847346481393663e-06, + "loss": 0.8223, + "num_input_tokens_seen": 7997296, + "step": 29865 + }, + { + "epoch": 7.82961992136304, + "grad_norm": 5.174185752868652, + "learning_rate": 6.839484304811055e-06, + "loss": 0.8338, + "num_input_tokens_seen": 7998208, + "step": 29870 + }, + { + "epoch": 7.830930537352556, + "grad_norm": 2.953801393508911, + "learning_rate": 6.831625929193006e-06, + "loss": 0.4078, + "num_input_tokens_seen": 7999536, + "step": 29875 + }, + { + "epoch": 7.832241153342071, + "grad_norm": 5.220383167266846, + "learning_rate": 6.8237713561842556e-06, + "loss": 0.4071, + "num_input_tokens_seen": 8000912, + "step": 29880 + }, + { + "epoch": 7.833551769331586, + "grad_norm": 1.200162410736084, + "learning_rate": 6.815920587428759e-06, + "loss": 0.536, + "num_input_tokens_seen": 8002304, + "step": 29885 + }, + { + "epoch": 7.834862385321101, + "grad_norm": 3.2317395210266113, + "learning_rate": 6.808073624569661e-06, + "loss": 0.5091, + "num_input_tokens_seen": 8003568, + "step": 29890 + }, + { + "epoch": 7.836173001310616, + "grad_norm": 2.197267532348633, + "learning_rate": 6.8002304692493165e-06, + "loss": 0.4653, + "num_input_tokens_seen": 8005152, + "step": 29895 + }, + { + "epoch": 7.837483617300131, + "grad_norm": 6.214138984680176, + "learning_rate": 6.79239112310929e-06, + "loss": 0.4552, + "num_input_tokens_seen": 8006176, + "step": 29900 + }, + { + "epoch": 7.838794233289646, + "grad_norm": 5.304162502288818, + "learning_rate": 6.784555587790342e-06, + "loss": 0.5677, + "num_input_tokens_seen": 8007472, + "step": 29905 + }, + { + "epoch": 7.840104849279161, + "grad_norm": 3.1504595279693604, + "learning_rate": 6.7767238649324325e-06, + "loss": 0.5041, + "num_input_tokens_seen": 8008800, + "step": 29910 + }, + { + "epoch": 7.8414154652686765, + "grad_norm": 3.0759899616241455, + "learning_rate": 6.768895956174726e-06, + "loss": 0.6408, + "num_input_tokens_seen": 8010048, + "step": 29915 + }, + { + "epoch": 7.842726081258191, + "grad_norm": 5.115838527679443, + "learning_rate": 6.761071863155607e-06, + "loss": 1.0057, + "num_input_tokens_seen": 8011040, + "step": 29920 + }, + { + "epoch": 7.844036697247707, + "grad_norm": 2.6806325912475586, + "learning_rate": 6.7532515875126295e-06, + "loss": 0.5563, + "num_input_tokens_seen": 8012480, + "step": 29925 + }, + { + "epoch": 7.845347313237221, + "grad_norm": 3.3001184463500977, + "learning_rate": 6.7454351308825634e-06, + "loss": 0.4604, + "num_input_tokens_seen": 8014048, + "step": 29930 + }, + { + "epoch": 7.846657929226737, + "grad_norm": 4.289942264556885, + "learning_rate": 6.737622494901383e-06, + "loss": 0.3265, + "num_input_tokens_seen": 8014896, + "step": 29935 + }, + { + "epoch": 7.8479685452162515, + "grad_norm": 5.8844804763793945, + "learning_rate": 6.7298136812042615e-06, + "loss": 0.4928, + "num_input_tokens_seen": 8016208, + "step": 29940 + }, + { + "epoch": 7.849279161205767, + "grad_norm": 2.0596017837524414, + "learning_rate": 6.722008691425566e-06, + "loss": 0.5326, + "num_input_tokens_seen": 8017376, + "step": 29945 + }, + { + "epoch": 7.850589777195282, + "grad_norm": 2.117124319076538, + "learning_rate": 6.7142075271988705e-06, + "loss": 0.4752, + "num_input_tokens_seen": 8018816, + "step": 29950 + }, + { + "epoch": 7.851900393184797, + "grad_norm": 6.641891002655029, + "learning_rate": 6.706410190156945e-06, + "loss": 0.3827, + "num_input_tokens_seen": 8019872, + "step": 29955 + }, + { + "epoch": 7.853211009174312, + "grad_norm": 3.505631923675537, + "learning_rate": 6.698616681931754e-06, + "loss": 0.6308, + "num_input_tokens_seen": 8020832, + "step": 29960 + }, + { + "epoch": 7.854521625163827, + "grad_norm": 2.175035238265991, + "learning_rate": 6.6908270041544815e-06, + "loss": 0.3352, + "num_input_tokens_seen": 8022176, + "step": 29965 + }, + { + "epoch": 7.855832241153342, + "grad_norm": 2.1063616275787354, + "learning_rate": 6.683041158455472e-06, + "loss": 0.4882, + "num_input_tokens_seen": 8023312, + "step": 29970 + }, + { + "epoch": 7.857142857142857, + "grad_norm": 2.959904670715332, + "learning_rate": 6.675259146464291e-06, + "loss": 0.6022, + "num_input_tokens_seen": 8024704, + "step": 29975 + }, + { + "epoch": 7.858453473132372, + "grad_norm": 1.5350273847579956, + "learning_rate": 6.667480969809714e-06, + "loss": 0.4806, + "num_input_tokens_seen": 8026656, + "step": 29980 + }, + { + "epoch": 7.859764089121887, + "grad_norm": 3.581462860107422, + "learning_rate": 6.659706630119697e-06, + "loss": 0.5454, + "num_input_tokens_seen": 8027920, + "step": 29985 + }, + { + "epoch": 7.8610747051114025, + "grad_norm": 2.5484704971313477, + "learning_rate": 6.651936129021391e-06, + "loss": 0.5508, + "num_input_tokens_seen": 8029504, + "step": 29990 + }, + { + "epoch": 7.862385321100917, + "grad_norm": 2.9509220123291016, + "learning_rate": 6.644169468141148e-06, + "loss": 0.6256, + "num_input_tokens_seen": 8030720, + "step": 29995 + }, + { + "epoch": 7.863695937090433, + "grad_norm": 3.3730547428131104, + "learning_rate": 6.636406649104529e-06, + "loss": 0.5692, + "num_input_tokens_seen": 8032144, + "step": 30000 + }, + { + "epoch": 7.865006553079947, + "grad_norm": 3.4790568351745605, + "learning_rate": 6.628647673536262e-06, + "loss": 0.4067, + "num_input_tokens_seen": 8033664, + "step": 30005 + }, + { + "epoch": 7.866317169069463, + "grad_norm": 1.4505493640899658, + "learning_rate": 6.620892543060292e-06, + "loss": 0.5884, + "num_input_tokens_seen": 8035040, + "step": 30010 + }, + { + "epoch": 7.8676277850589775, + "grad_norm": 2.1228702068328857, + "learning_rate": 6.613141259299754e-06, + "loss": 0.4099, + "num_input_tokens_seen": 8036048, + "step": 30015 + }, + { + "epoch": 7.868938401048493, + "grad_norm": 3.2959115505218506, + "learning_rate": 6.6053938238769845e-06, + "loss": 0.4847, + "num_input_tokens_seen": 8037984, + "step": 30020 + }, + { + "epoch": 7.870249017038008, + "grad_norm": 3.6973485946655273, + "learning_rate": 6.5976502384135025e-06, + "loss": 0.3896, + "num_input_tokens_seen": 8038976, + "step": 30025 + }, + { + "epoch": 7.871559633027523, + "grad_norm": 10.481034278869629, + "learning_rate": 6.58991050453002e-06, + "loss": 0.4194, + "num_input_tokens_seen": 8040128, + "step": 30030 + }, + { + "epoch": 7.872870249017038, + "grad_norm": 2.8255176544189453, + "learning_rate": 6.582174623846477e-06, + "loss": 0.4857, + "num_input_tokens_seen": 8041712, + "step": 30035 + }, + { + "epoch": 7.8741808650065535, + "grad_norm": 2.525700807571411, + "learning_rate": 6.57444259798195e-06, + "loss": 0.4911, + "num_input_tokens_seen": 8043104, + "step": 30040 + }, + { + "epoch": 7.875491480996068, + "grad_norm": 3.2095425128936768, + "learning_rate": 6.566714428554752e-06, + "loss": 0.6238, + "num_input_tokens_seen": 8044416, + "step": 30045 + }, + { + "epoch": 7.876802096985584, + "grad_norm": 3.371885061264038, + "learning_rate": 6.558990117182376e-06, + "loss": 0.3159, + "num_input_tokens_seen": 8045824, + "step": 30050 + }, + { + "epoch": 7.878112712975098, + "grad_norm": 2.069716453552246, + "learning_rate": 6.551269665481502e-06, + "loss": 0.4202, + "num_input_tokens_seen": 8046768, + "step": 30055 + }, + { + "epoch": 7.879423328964613, + "grad_norm": 3.980436325073242, + "learning_rate": 6.543553075068012e-06, + "loss": 0.5937, + "num_input_tokens_seen": 8047856, + "step": 30060 + }, + { + "epoch": 7.8807339449541285, + "grad_norm": 8.383040428161621, + "learning_rate": 6.535840347556971e-06, + "loss": 0.7089, + "num_input_tokens_seen": 8048832, + "step": 30065 + }, + { + "epoch": 7.882044560943643, + "grad_norm": 3.8432817459106445, + "learning_rate": 6.528131484562644e-06, + "loss": 0.6044, + "num_input_tokens_seen": 8049952, + "step": 30070 + }, + { + "epoch": 7.883355176933159, + "grad_norm": 3.319859743118286, + "learning_rate": 6.520426487698478e-06, + "loss": 0.411, + "num_input_tokens_seen": 8051392, + "step": 30075 + }, + { + "epoch": 7.884665792922673, + "grad_norm": 4.481537818908691, + "learning_rate": 6.512725358577121e-06, + "loss": 0.5137, + "num_input_tokens_seen": 8052720, + "step": 30080 + }, + { + "epoch": 7.885976408912189, + "grad_norm": 2.448218822479248, + "learning_rate": 6.505028098810406e-06, + "loss": 0.4428, + "num_input_tokens_seen": 8053936, + "step": 30085 + }, + { + "epoch": 7.8872870249017035, + "grad_norm": 3.7051517963409424, + "learning_rate": 6.497334710009342e-06, + "loss": 0.2883, + "num_input_tokens_seen": 8055072, + "step": 30090 + }, + { + "epoch": 7.888597640891219, + "grad_norm": 2.9773666858673096, + "learning_rate": 6.489645193784158e-06, + "loss": 0.424, + "num_input_tokens_seen": 8056624, + "step": 30095 + }, + { + "epoch": 7.889908256880734, + "grad_norm": 2.620925188064575, + "learning_rate": 6.481959551744254e-06, + "loss": 0.5726, + "num_input_tokens_seen": 8058192, + "step": 30100 + }, + { + "epoch": 7.891218872870249, + "grad_norm": 1.7986420392990112, + "learning_rate": 6.474277785498217e-06, + "loss": 0.326, + "num_input_tokens_seen": 8059520, + "step": 30105 + }, + { + "epoch": 7.892529488859764, + "grad_norm": 1.838405728340149, + "learning_rate": 6.46659989665383e-06, + "loss": 0.4864, + "num_input_tokens_seen": 8061008, + "step": 30110 + }, + { + "epoch": 7.8938401048492794, + "grad_norm": 5.679198741912842, + "learning_rate": 6.458925886818062e-06, + "loss": 0.6119, + "num_input_tokens_seen": 8062336, + "step": 30115 + }, + { + "epoch": 7.895150720838794, + "grad_norm": 4.894464015960693, + "learning_rate": 6.451255757597078e-06, + "loss": 0.6883, + "num_input_tokens_seen": 8063632, + "step": 30120 + }, + { + "epoch": 7.89646133682831, + "grad_norm": 2.196251630783081, + "learning_rate": 6.443589510596204e-06, + "loss": 0.4268, + "num_input_tokens_seen": 8064832, + "step": 30125 + }, + { + "epoch": 7.897771952817824, + "grad_norm": 1.7235143184661865, + "learning_rate": 6.435927147419984e-06, + "loss": 0.4884, + "num_input_tokens_seen": 8065888, + "step": 30130 + }, + { + "epoch": 7.89908256880734, + "grad_norm": 2.1491587162017822, + "learning_rate": 6.4282686696721356e-06, + "loss": 0.4901, + "num_input_tokens_seen": 8067024, + "step": 30135 + }, + { + "epoch": 7.9003931847968545, + "grad_norm": 2.0169894695281982, + "learning_rate": 6.420614078955564e-06, + "loss": 0.3951, + "num_input_tokens_seen": 8068096, + "step": 30140 + }, + { + "epoch": 7.90170380078637, + "grad_norm": 2.24297833442688, + "learning_rate": 6.412963376872358e-06, + "loss": 0.5301, + "num_input_tokens_seen": 8069680, + "step": 30145 + }, + { + "epoch": 7.903014416775885, + "grad_norm": 3.009547710418701, + "learning_rate": 6.405316565023806e-06, + "loss": 0.3877, + "num_input_tokens_seen": 8071056, + "step": 30150 + }, + { + "epoch": 7.904325032765399, + "grad_norm": 1.3628367185592651, + "learning_rate": 6.397673645010377e-06, + "loss": 0.5065, + "num_input_tokens_seen": 8072432, + "step": 30155 + }, + { + "epoch": 7.905635648754915, + "grad_norm": 1.6497844457626343, + "learning_rate": 6.390034618431704e-06, + "loss": 0.4744, + "num_input_tokens_seen": 8074016, + "step": 30160 + }, + { + "epoch": 7.9069462647444295, + "grad_norm": 1.504987120628357, + "learning_rate": 6.382399486886631e-06, + "loss": 0.2919, + "num_input_tokens_seen": 8075616, + "step": 30165 + }, + { + "epoch": 7.908256880733945, + "grad_norm": 7.382762908935547, + "learning_rate": 6.374768251973174e-06, + "loss": 0.3302, + "num_input_tokens_seen": 8076656, + "step": 30170 + }, + { + "epoch": 7.90956749672346, + "grad_norm": 2.6663875579833984, + "learning_rate": 6.367140915288542e-06, + "loss": 0.5004, + "num_input_tokens_seen": 8078096, + "step": 30175 + }, + { + "epoch": 7.910878112712975, + "grad_norm": 2.4434497356414795, + "learning_rate": 6.3595174784291195e-06, + "loss": 0.5501, + "num_input_tokens_seen": 8079344, + "step": 30180 + }, + { + "epoch": 7.91218872870249, + "grad_norm": 1.365761160850525, + "learning_rate": 6.351897942990481e-06, + "loss": 0.55, + "num_input_tokens_seen": 8080624, + "step": 30185 + }, + { + "epoch": 7.913499344692005, + "grad_norm": 3.057002305984497, + "learning_rate": 6.344282310567384e-06, + "loss": 0.4934, + "num_input_tokens_seen": 8081808, + "step": 30190 + }, + { + "epoch": 7.91480996068152, + "grad_norm": 4.605463981628418, + "learning_rate": 6.336670582753762e-06, + "loss": 0.4923, + "num_input_tokens_seen": 8082800, + "step": 30195 + }, + { + "epoch": 7.916120576671036, + "grad_norm": 4.792738914489746, + "learning_rate": 6.329062761142748e-06, + "loss": 0.4909, + "num_input_tokens_seen": 8084256, + "step": 30200 + }, + { + "epoch": 7.91743119266055, + "grad_norm": 1.4866057634353638, + "learning_rate": 6.3214588473266225e-06, + "loss": 0.3935, + "num_input_tokens_seen": 8085520, + "step": 30205 + }, + { + "epoch": 7.918741808650066, + "grad_norm": 2.9980242252349854, + "learning_rate": 6.313858842896895e-06, + "loss": 0.5483, + "num_input_tokens_seen": 8086976, + "step": 30210 + }, + { + "epoch": 7.9200524246395805, + "grad_norm": 2.887082576751709, + "learning_rate": 6.306262749444219e-06, + "loss": 0.5321, + "num_input_tokens_seen": 8088256, + "step": 30215 + }, + { + "epoch": 7.921363040629096, + "grad_norm": 3.9723567962646484, + "learning_rate": 6.298670568558454e-06, + "loss": 0.3077, + "num_input_tokens_seen": 8089616, + "step": 30220 + }, + { + "epoch": 7.922673656618611, + "grad_norm": 10.422383308410645, + "learning_rate": 6.291082301828624e-06, + "loss": 0.6563, + "num_input_tokens_seen": 8090720, + "step": 30225 + }, + { + "epoch": 7.923984272608125, + "grad_norm": 2.858647346496582, + "learning_rate": 6.283497950842942e-06, + "loss": 0.3799, + "num_input_tokens_seen": 8091888, + "step": 30230 + }, + { + "epoch": 7.925294888597641, + "grad_norm": 4.0317769050598145, + "learning_rate": 6.275917517188809e-06, + "loss": 0.4153, + "num_input_tokens_seen": 8093008, + "step": 30235 + }, + { + "epoch": 7.926605504587156, + "grad_norm": 3.6798555850982666, + "learning_rate": 6.2683410024527764e-06, + "loss": 0.3818, + "num_input_tokens_seen": 8094000, + "step": 30240 + }, + { + "epoch": 7.927916120576671, + "grad_norm": 1.0582311153411865, + "learning_rate": 6.260768408220607e-06, + "loss": 0.6266, + "num_input_tokens_seen": 8095680, + "step": 30245 + }, + { + "epoch": 7.929226736566186, + "grad_norm": 6.129692077636719, + "learning_rate": 6.253199736077231e-06, + "loss": 0.6074, + "num_input_tokens_seen": 8096720, + "step": 30250 + }, + { + "epoch": 7.930537352555701, + "grad_norm": 2.350327253341675, + "learning_rate": 6.24563498760676e-06, + "loss": 0.3161, + "num_input_tokens_seen": 8097728, + "step": 30255 + }, + { + "epoch": 7.931847968545216, + "grad_norm": 1.485217809677124, + "learning_rate": 6.238074164392471e-06, + "loss": 0.3896, + "num_input_tokens_seen": 8099184, + "step": 30260 + }, + { + "epoch": 7.933158584534731, + "grad_norm": 4.989798545837402, + "learning_rate": 6.23051726801685e-06, + "loss": 0.5691, + "num_input_tokens_seen": 8100256, + "step": 30265 + }, + { + "epoch": 7.934469200524246, + "grad_norm": 5.933127403259277, + "learning_rate": 6.22296430006154e-06, + "loss": 0.7852, + "num_input_tokens_seen": 8101536, + "step": 30270 + }, + { + "epoch": 7.935779816513762, + "grad_norm": 2.505340099334717, + "learning_rate": 6.215415262107352e-06, + "loss": 0.5138, + "num_input_tokens_seen": 8102816, + "step": 30275 + }, + { + "epoch": 7.937090432503276, + "grad_norm": 1.4888299703598022, + "learning_rate": 6.207870155734291e-06, + "loss": 0.486, + "num_input_tokens_seen": 8104864, + "step": 30280 + }, + { + "epoch": 7.938401048492792, + "grad_norm": 1.461180329322815, + "learning_rate": 6.200328982521536e-06, + "loss": 0.449, + "num_input_tokens_seen": 8106416, + "step": 30285 + }, + { + "epoch": 7.9397116644823065, + "grad_norm": 3.3282485008239746, + "learning_rate": 6.192791744047444e-06, + "loss": 0.5096, + "num_input_tokens_seen": 8107696, + "step": 30290 + }, + { + "epoch": 7.941022280471822, + "grad_norm": 3.1533865928649902, + "learning_rate": 6.185258441889541e-06, + "loss": 0.5615, + "num_input_tokens_seen": 8109440, + "step": 30295 + }, + { + "epoch": 7.942332896461337, + "grad_norm": 1.7448891401290894, + "learning_rate": 6.177729077624539e-06, + "loss": 0.5168, + "num_input_tokens_seen": 8110592, + "step": 30300 + }, + { + "epoch": 7.943643512450852, + "grad_norm": 2.1619222164154053, + "learning_rate": 6.170203652828316e-06, + "loss": 0.9062, + "num_input_tokens_seen": 8111872, + "step": 30305 + }, + { + "epoch": 7.944954128440367, + "grad_norm": 2.2828917503356934, + "learning_rate": 6.162682169075934e-06, + "loss": 0.4342, + "num_input_tokens_seen": 8113920, + "step": 30310 + }, + { + "epoch": 7.946264744429882, + "grad_norm": 6.429940700531006, + "learning_rate": 6.155164627941626e-06, + "loss": 0.4946, + "num_input_tokens_seen": 8115152, + "step": 30315 + }, + { + "epoch": 7.947575360419397, + "grad_norm": 1.5713310241699219, + "learning_rate": 6.147651030998799e-06, + "loss": 0.3352, + "num_input_tokens_seen": 8116448, + "step": 30320 + }, + { + "epoch": 7.948885976408912, + "grad_norm": 3.168391704559326, + "learning_rate": 6.140141379820038e-06, + "loss": 0.5074, + "num_input_tokens_seen": 8117600, + "step": 30325 + }, + { + "epoch": 7.950196592398427, + "grad_norm": 5.927618026733398, + "learning_rate": 6.132635675977099e-06, + "loss": 0.5305, + "num_input_tokens_seen": 8118704, + "step": 30330 + }, + { + "epoch": 7.951507208387943, + "grad_norm": 1.2092620134353638, + "learning_rate": 6.125133921040912e-06, + "loss": 0.3028, + "num_input_tokens_seen": 8120960, + "step": 30335 + }, + { + "epoch": 7.952817824377457, + "grad_norm": 2.581958055496216, + "learning_rate": 6.1176361165815845e-06, + "loss": 0.3222, + "num_input_tokens_seen": 8122288, + "step": 30340 + }, + { + "epoch": 7.954128440366972, + "grad_norm": 1.8403809070587158, + "learning_rate": 6.1101422641683895e-06, + "loss": 0.3991, + "num_input_tokens_seen": 8123440, + "step": 30345 + }, + { + "epoch": 7.955439056356488, + "grad_norm": 2.8065671920776367, + "learning_rate": 6.102652365369779e-06, + "loss": 0.3675, + "num_input_tokens_seen": 8124880, + "step": 30350 + }, + { + "epoch": 7.956749672346002, + "grad_norm": 34.093936920166016, + "learning_rate": 6.095166421753385e-06, + "loss": 0.7255, + "num_input_tokens_seen": 8125904, + "step": 30355 + }, + { + "epoch": 7.958060288335518, + "grad_norm": 1.3845579624176025, + "learning_rate": 6.087684434885985e-06, + "loss": 0.3849, + "num_input_tokens_seen": 8127472, + "step": 30360 + }, + { + "epoch": 7.9593709043250325, + "grad_norm": 4.472171306610107, + "learning_rate": 6.080206406333555e-06, + "loss": 0.5733, + "num_input_tokens_seen": 8128576, + "step": 30365 + }, + { + "epoch": 7.960681520314548, + "grad_norm": 1.8823914527893066, + "learning_rate": 6.072732337661236e-06, + "loss": 0.4657, + "num_input_tokens_seen": 8129824, + "step": 30370 + }, + { + "epoch": 7.961992136304063, + "grad_norm": 1.6879463195800781, + "learning_rate": 6.065262230433325e-06, + "loss": 0.4296, + "num_input_tokens_seen": 8131056, + "step": 30375 + }, + { + "epoch": 7.963302752293578, + "grad_norm": 4.008354663848877, + "learning_rate": 6.057796086213319e-06, + "loss": 0.5916, + "num_input_tokens_seen": 8132192, + "step": 30380 + }, + { + "epoch": 7.964613368283093, + "grad_norm": 2.1399126052856445, + "learning_rate": 6.050333906563865e-06, + "loss": 0.4972, + "num_input_tokens_seen": 8133520, + "step": 30385 + }, + { + "epoch": 7.965923984272608, + "grad_norm": 1.743483304977417, + "learning_rate": 6.0428756930467876e-06, + "loss": 0.3584, + "num_input_tokens_seen": 8134832, + "step": 30390 + }, + { + "epoch": 7.967234600262123, + "grad_norm": 3.4852728843688965, + "learning_rate": 6.035421447223064e-06, + "loss": 0.4051, + "num_input_tokens_seen": 8136688, + "step": 30395 + }, + { + "epoch": 7.968545216251639, + "grad_norm": 6.695756912231445, + "learning_rate": 6.027971170652866e-06, + "loss": 0.8112, + "num_input_tokens_seen": 8137776, + "step": 30400 + }, + { + "epoch": 7.969855832241153, + "grad_norm": 4.496495246887207, + "learning_rate": 6.020524864895519e-06, + "loss": 0.4683, + "num_input_tokens_seen": 8138912, + "step": 30405 + }, + { + "epoch": 7.971166448230669, + "grad_norm": 1.1162606477737427, + "learning_rate": 6.013082531509523e-06, + "loss": 0.4737, + "num_input_tokens_seen": 8140256, + "step": 30410 + }, + { + "epoch": 7.972477064220183, + "grad_norm": 2.878350257873535, + "learning_rate": 6.005644172052549e-06, + "loss": 0.384, + "num_input_tokens_seen": 8141520, + "step": 30415 + }, + { + "epoch": 7.973787680209698, + "grad_norm": 3.064523220062256, + "learning_rate": 5.998209788081427e-06, + "loss": 0.3725, + "num_input_tokens_seen": 8142704, + "step": 30420 + }, + { + "epoch": 7.975098296199214, + "grad_norm": 2.818798065185547, + "learning_rate": 5.990779381152168e-06, + "loss": 0.3959, + "num_input_tokens_seen": 8143968, + "step": 30425 + }, + { + "epoch": 7.976408912188729, + "grad_norm": 6.404308319091797, + "learning_rate": 5.9833529528199344e-06, + "loss": 0.4705, + "num_input_tokens_seen": 8145216, + "step": 30430 + }, + { + "epoch": 7.977719528178244, + "grad_norm": 4.185575485229492, + "learning_rate": 5.9759305046390715e-06, + "loss": 0.4306, + "num_input_tokens_seen": 8146352, + "step": 30435 + }, + { + "epoch": 7.9790301441677585, + "grad_norm": 1.893722653388977, + "learning_rate": 5.968512038163085e-06, + "loss": 0.4273, + "num_input_tokens_seen": 8148464, + "step": 30440 + }, + { + "epoch": 7.980340760157274, + "grad_norm": 3.4606776237487793, + "learning_rate": 5.9610975549446414e-06, + "loss": 0.5394, + "num_input_tokens_seen": 8149616, + "step": 30445 + }, + { + "epoch": 7.981651376146789, + "grad_norm": 2.146358013153076, + "learning_rate": 5.953687056535584e-06, + "loss": 0.3754, + "num_input_tokens_seen": 8150864, + "step": 30450 + }, + { + "epoch": 7.982961992136304, + "grad_norm": 4.662749290466309, + "learning_rate": 5.946280544486918e-06, + "loss": 0.5441, + "num_input_tokens_seen": 8152000, + "step": 30455 + }, + { + "epoch": 7.984272608125819, + "grad_norm": 3.1112279891967773, + "learning_rate": 5.938878020348809e-06, + "loss": 0.3719, + "num_input_tokens_seen": 8153440, + "step": 30460 + }, + { + "epoch": 7.985583224115334, + "grad_norm": 3.9445934295654297, + "learning_rate": 5.931479485670599e-06, + "loss": 0.6596, + "num_input_tokens_seen": 8154864, + "step": 30465 + }, + { + "epoch": 7.986893840104849, + "grad_norm": 2.5020015239715576, + "learning_rate": 5.924084942000787e-06, + "loss": 0.5172, + "num_input_tokens_seen": 8156448, + "step": 30470 + }, + { + "epoch": 7.988204456094365, + "grad_norm": 4.7255096435546875, + "learning_rate": 5.916694390887034e-06, + "loss": 0.3276, + "num_input_tokens_seen": 8157392, + "step": 30475 + }, + { + "epoch": 7.989515072083879, + "grad_norm": 2.659611701965332, + "learning_rate": 5.909307833876168e-06, + "loss": 0.498, + "num_input_tokens_seen": 8158560, + "step": 30480 + }, + { + "epoch": 7.990825688073395, + "grad_norm": 5.841415882110596, + "learning_rate": 5.901925272514186e-06, + "loss": 0.629, + "num_input_tokens_seen": 8159776, + "step": 30485 + }, + { + "epoch": 7.992136304062909, + "grad_norm": 2.7284045219421387, + "learning_rate": 5.8945467083462405e-06, + "loss": 0.5406, + "num_input_tokens_seen": 8161184, + "step": 30490 + }, + { + "epoch": 7.993446920052425, + "grad_norm": 1.9305483102798462, + "learning_rate": 5.887172142916663e-06, + "loss": 0.4916, + "num_input_tokens_seen": 8162464, + "step": 30495 + }, + { + "epoch": 7.99475753604194, + "grad_norm": 3.851289987564087, + "learning_rate": 5.879801577768934e-06, + "loss": 0.3509, + "num_input_tokens_seen": 8163984, + "step": 30500 + }, + { + "epoch": 7.996068152031455, + "grad_norm": 5.176931381225586, + "learning_rate": 5.872435014445696e-06, + "loss": 0.4492, + "num_input_tokens_seen": 8165120, + "step": 30505 + }, + { + "epoch": 7.99737876802097, + "grad_norm": 3.383728504180908, + "learning_rate": 5.865072454488765e-06, + "loss": 0.4442, + "num_input_tokens_seen": 8166656, + "step": 30510 + }, + { + "epoch": 7.9986893840104845, + "grad_norm": 3.47408127784729, + "learning_rate": 5.857713899439101e-06, + "loss": 0.635, + "num_input_tokens_seen": 8167568, + "step": 30515 + }, + { + "epoch": 8.0, + "grad_norm": 9.116125106811523, + "learning_rate": 5.850359350836842e-06, + "loss": 0.4758, + "num_input_tokens_seen": 8168664, + "step": 30520 + }, + { + "epoch": 8.001310615989516, + "grad_norm": 2.636604070663452, + "learning_rate": 5.843008810221284e-06, + "loss": 0.3807, + "num_input_tokens_seen": 8169944, + "step": 30525 + }, + { + "epoch": 8.002096985583224, + "eval_loss": 0.4648008644580841, + "eval_runtime": 18.2169, + "eval_samples_per_second": 46.55, + "eval_steps_per_second": 23.275, + "num_input_tokens_seen": 8170632, + "step": 30528 + }, + { + "epoch": 8.00262123197903, + "grad_norm": 2.2837581634521484, + "learning_rate": 5.835662279130879e-06, + "loss": 0.2536, + "num_input_tokens_seen": 8171080, + "step": 30530 + }, + { + "epoch": 8.003931847968545, + "grad_norm": 3.998188018798828, + "learning_rate": 5.828319759103248e-06, + "loss": 0.5725, + "num_input_tokens_seen": 8172472, + "step": 30535 + }, + { + "epoch": 8.00524246395806, + "grad_norm": 1.6730095148086548, + "learning_rate": 5.8209812516751646e-06, + "loss": 0.4785, + "num_input_tokens_seen": 8173672, + "step": 30540 + }, + { + "epoch": 8.006553079947576, + "grad_norm": 3.2477028369903564, + "learning_rate": 5.8136467583825685e-06, + "loss": 0.4539, + "num_input_tokens_seen": 8174744, + "step": 30545 + }, + { + "epoch": 8.00786369593709, + "grad_norm": 3.0398781299591064, + "learning_rate": 5.806316280760551e-06, + "loss": 0.5378, + "num_input_tokens_seen": 8175848, + "step": 30550 + }, + { + "epoch": 8.009174311926605, + "grad_norm": 1.7706221342086792, + "learning_rate": 5.7989898203433725e-06, + "loss": 0.3497, + "num_input_tokens_seen": 8177080, + "step": 30555 + }, + { + "epoch": 8.01048492791612, + "grad_norm": 5.194352149963379, + "learning_rate": 5.791667378664451e-06, + "loss": 0.4234, + "num_input_tokens_seen": 8177944, + "step": 30560 + }, + { + "epoch": 8.011795543905636, + "grad_norm": 2.8427391052246094, + "learning_rate": 5.784348957256352e-06, + "loss": 0.3688, + "num_input_tokens_seen": 8179176, + "step": 30565 + }, + { + "epoch": 8.01310615989515, + "grad_norm": 3.552497625350952, + "learning_rate": 5.7770345576508176e-06, + "loss": 0.3999, + "num_input_tokens_seen": 8180568, + "step": 30570 + }, + { + "epoch": 8.014416775884666, + "grad_norm": 3.0117075443267822, + "learning_rate": 5.769724181378736e-06, + "loss": 0.4856, + "num_input_tokens_seen": 8182040, + "step": 30575 + }, + { + "epoch": 8.015727391874181, + "grad_norm": 3.4113757610321045, + "learning_rate": 5.7624178299701546e-06, + "loss": 0.4082, + "num_input_tokens_seen": 8183400, + "step": 30580 + }, + { + "epoch": 8.017038007863697, + "grad_norm": 5.798975944519043, + "learning_rate": 5.755115504954284e-06, + "loss": 0.6067, + "num_input_tokens_seen": 8184648, + "step": 30585 + }, + { + "epoch": 8.01834862385321, + "grad_norm": 2.125502586364746, + "learning_rate": 5.747817207859491e-06, + "loss": 0.5336, + "num_input_tokens_seen": 8185880, + "step": 30590 + }, + { + "epoch": 8.019659239842726, + "grad_norm": 2.517913818359375, + "learning_rate": 5.740522940213283e-06, + "loss": 0.3242, + "num_input_tokens_seen": 8187576, + "step": 30595 + }, + { + "epoch": 8.020969855832242, + "grad_norm": 2.9778120517730713, + "learning_rate": 5.733232703542346e-06, + "loss": 0.6011, + "num_input_tokens_seen": 8188952, + "step": 30600 + }, + { + "epoch": 8.022280471821757, + "grad_norm": 3.321803092956543, + "learning_rate": 5.725946499372506e-06, + "loss": 0.3292, + "num_input_tokens_seen": 8190072, + "step": 30605 + }, + { + "epoch": 8.02359108781127, + "grad_norm": 1.5840461254119873, + "learning_rate": 5.7186643292287705e-06, + "loss": 0.3275, + "num_input_tokens_seen": 8191208, + "step": 30610 + }, + { + "epoch": 8.024901703800786, + "grad_norm": 2.003227949142456, + "learning_rate": 5.711386194635274e-06, + "loss": 0.4572, + "num_input_tokens_seen": 8192312, + "step": 30615 + }, + { + "epoch": 8.026212319790302, + "grad_norm": 2.8645546436309814, + "learning_rate": 5.704112097115316e-06, + "loss": 0.3833, + "num_input_tokens_seen": 8194040, + "step": 30620 + }, + { + "epoch": 8.027522935779816, + "grad_norm": 4.348957061767578, + "learning_rate": 5.696842038191366e-06, + "loss": 0.3438, + "num_input_tokens_seen": 8195336, + "step": 30625 + }, + { + "epoch": 8.028833551769331, + "grad_norm": 2.126875162124634, + "learning_rate": 5.689576019385015e-06, + "loss": 0.3168, + "num_input_tokens_seen": 8196344, + "step": 30630 + }, + { + "epoch": 8.030144167758847, + "grad_norm": 8.028481483459473, + "learning_rate": 5.682314042217041e-06, + "loss": 0.4851, + "num_input_tokens_seen": 8197512, + "step": 30635 + }, + { + "epoch": 8.031454783748362, + "grad_norm": 1.6729463338851929, + "learning_rate": 5.675056108207355e-06, + "loss": 0.4454, + "num_input_tokens_seen": 8198936, + "step": 30640 + }, + { + "epoch": 8.032765399737876, + "grad_norm": 0.9819478988647461, + "learning_rate": 5.667802218875037e-06, + "loss": 0.3549, + "num_input_tokens_seen": 8200120, + "step": 30645 + }, + { + "epoch": 8.034076015727392, + "grad_norm": 1.6757971048355103, + "learning_rate": 5.66055237573831e-06, + "loss": 0.333, + "num_input_tokens_seen": 8201176, + "step": 30650 + }, + { + "epoch": 8.035386631716907, + "grad_norm": 3.363227128982544, + "learning_rate": 5.6533065803145545e-06, + "loss": 0.6272, + "num_input_tokens_seen": 8202360, + "step": 30655 + }, + { + "epoch": 8.036697247706423, + "grad_norm": 1.3841958045959473, + "learning_rate": 5.646064834120304e-06, + "loss": 0.5215, + "num_input_tokens_seen": 8203640, + "step": 30660 + }, + { + "epoch": 8.038007863695936, + "grad_norm": 3.4793477058410645, + "learning_rate": 5.638827138671243e-06, + "loss": 0.4406, + "num_input_tokens_seen": 8205048, + "step": 30665 + }, + { + "epoch": 8.039318479685452, + "grad_norm": 4.929070949554443, + "learning_rate": 5.631593495482209e-06, + "loss": 0.4806, + "num_input_tokens_seen": 8206216, + "step": 30670 + }, + { + "epoch": 8.040629095674968, + "grad_norm": 2.932630777359009, + "learning_rate": 5.62436390606719e-06, + "loss": 0.5072, + "num_input_tokens_seen": 8207496, + "step": 30675 + }, + { + "epoch": 8.041939711664483, + "grad_norm": 3.0231950283050537, + "learning_rate": 5.617138371939326e-06, + "loss": 0.3739, + "num_input_tokens_seen": 8208840, + "step": 30680 + }, + { + "epoch": 8.043250327653997, + "grad_norm": 2.337599515914917, + "learning_rate": 5.6099168946109114e-06, + "loss": 0.6185, + "num_input_tokens_seen": 8210200, + "step": 30685 + }, + { + "epoch": 8.044560943643512, + "grad_norm": 4.593106269836426, + "learning_rate": 5.602699475593387e-06, + "loss": 0.4603, + "num_input_tokens_seen": 8211192, + "step": 30690 + }, + { + "epoch": 8.045871559633028, + "grad_norm": 4.436142444610596, + "learning_rate": 5.595486116397344e-06, + "loss": 0.5337, + "num_input_tokens_seen": 8212840, + "step": 30695 + }, + { + "epoch": 8.047182175622543, + "grad_norm": 1.391870379447937, + "learning_rate": 5.588276818532531e-06, + "loss": 0.4443, + "num_input_tokens_seen": 8214568, + "step": 30700 + }, + { + "epoch": 8.048492791612057, + "grad_norm": 3.5050768852233887, + "learning_rate": 5.5810715835078434e-06, + "loss": 0.5428, + "num_input_tokens_seen": 8215608, + "step": 30705 + }, + { + "epoch": 8.049803407601573, + "grad_norm": 3.9091384410858154, + "learning_rate": 5.573870412831314e-06, + "loss": 0.3187, + "num_input_tokens_seen": 8216824, + "step": 30710 + }, + { + "epoch": 8.051114023591088, + "grad_norm": 13.957024574279785, + "learning_rate": 5.566673308010142e-06, + "loss": 0.5018, + "num_input_tokens_seen": 8217816, + "step": 30715 + }, + { + "epoch": 8.052424639580602, + "grad_norm": 6.8842387199401855, + "learning_rate": 5.5594802705506626e-06, + "loss": 0.6505, + "num_input_tokens_seen": 8219080, + "step": 30720 + }, + { + "epoch": 8.053735255570118, + "grad_norm": 4.683077335357666, + "learning_rate": 5.552291301958379e-06, + "loss": 0.4434, + "num_input_tokens_seen": 8220360, + "step": 30725 + }, + { + "epoch": 8.055045871559633, + "grad_norm": 8.104729652404785, + "learning_rate": 5.545106403737921e-06, + "loss": 0.5496, + "num_input_tokens_seen": 8221288, + "step": 30730 + }, + { + "epoch": 8.056356487549149, + "grad_norm": 3.529205560684204, + "learning_rate": 5.537925577393077e-06, + "loss": 0.5015, + "num_input_tokens_seen": 8222760, + "step": 30735 + }, + { + "epoch": 8.057667103538662, + "grad_norm": 2.032961845397949, + "learning_rate": 5.530748824426782e-06, + "loss": 0.3423, + "num_input_tokens_seen": 8224200, + "step": 30740 + }, + { + "epoch": 8.058977719528178, + "grad_norm": 5.110462188720703, + "learning_rate": 5.523576146341128e-06, + "loss": 0.5305, + "num_input_tokens_seen": 8225240, + "step": 30745 + }, + { + "epoch": 8.060288335517694, + "grad_norm": 1.9216947555541992, + "learning_rate": 5.5164075446373225e-06, + "loss": 0.5384, + "num_input_tokens_seen": 8227400, + "step": 30750 + }, + { + "epoch": 8.061598951507209, + "grad_norm": 4.710636138916016, + "learning_rate": 5.509243020815755e-06, + "loss": 0.4618, + "num_input_tokens_seen": 8228632, + "step": 30755 + }, + { + "epoch": 8.062909567496723, + "grad_norm": 2.5602707862854004, + "learning_rate": 5.502082576375947e-06, + "loss": 0.6523, + "num_input_tokens_seen": 8230104, + "step": 30760 + }, + { + "epoch": 8.064220183486238, + "grad_norm": 1.5625455379486084, + "learning_rate": 5.494926212816562e-06, + "loss": 0.4873, + "num_input_tokens_seen": 8231400, + "step": 30765 + }, + { + "epoch": 8.065530799475754, + "grad_norm": 1.6356621980667114, + "learning_rate": 5.4877739316354236e-06, + "loss": 0.3935, + "num_input_tokens_seen": 8233016, + "step": 30770 + }, + { + "epoch": 8.06684141546527, + "grad_norm": 3.728158712387085, + "learning_rate": 5.4806257343294755e-06, + "loss": 0.5457, + "num_input_tokens_seen": 8234088, + "step": 30775 + }, + { + "epoch": 8.068152031454783, + "grad_norm": 5.124449729919434, + "learning_rate": 5.473481622394849e-06, + "loss": 0.2862, + "num_input_tokens_seen": 8235320, + "step": 30780 + }, + { + "epoch": 8.069462647444299, + "grad_norm": 2.344275951385498, + "learning_rate": 5.466341597326774e-06, + "loss": 0.4313, + "num_input_tokens_seen": 8236632, + "step": 30785 + }, + { + "epoch": 8.070773263433814, + "grad_norm": 2.856093168258667, + "learning_rate": 5.459205660619651e-06, + "loss": 0.4201, + "num_input_tokens_seen": 8238104, + "step": 30790 + }, + { + "epoch": 8.07208387942333, + "grad_norm": 2.8277461528778076, + "learning_rate": 5.452073813767019e-06, + "loss": 0.3999, + "num_input_tokens_seen": 8239832, + "step": 30795 + }, + { + "epoch": 8.073394495412844, + "grad_norm": 4.825343608856201, + "learning_rate": 5.444946058261563e-06, + "loss": 0.4901, + "num_input_tokens_seen": 8241128, + "step": 30800 + }, + { + "epoch": 8.07470511140236, + "grad_norm": 2.284234046936035, + "learning_rate": 5.4378223955951074e-06, + "loss": 0.4335, + "num_input_tokens_seen": 8242392, + "step": 30805 + }, + { + "epoch": 8.076015727391875, + "grad_norm": 2.7665488719940186, + "learning_rate": 5.430702827258622e-06, + "loss": 0.3554, + "num_input_tokens_seen": 8243544, + "step": 30810 + }, + { + "epoch": 8.077326343381388, + "grad_norm": 2.512882709503174, + "learning_rate": 5.423587354742224e-06, + "loss": 0.3291, + "num_input_tokens_seen": 8244824, + "step": 30815 + }, + { + "epoch": 8.078636959370904, + "grad_norm": 4.386537075042725, + "learning_rate": 5.4164759795351655e-06, + "loss": 0.6542, + "num_input_tokens_seen": 8245928, + "step": 30820 + }, + { + "epoch": 8.07994757536042, + "grad_norm": 5.254810333251953, + "learning_rate": 5.4093687031258554e-06, + "loss": 0.5989, + "num_input_tokens_seen": 8247096, + "step": 30825 + }, + { + "epoch": 8.081258191349935, + "grad_norm": 2.125267744064331, + "learning_rate": 5.402265527001818e-06, + "loss": 0.4574, + "num_input_tokens_seen": 8248600, + "step": 30830 + }, + { + "epoch": 8.082568807339449, + "grad_norm": 2.655221462249756, + "learning_rate": 5.3951664526497356e-06, + "loss": 0.3944, + "num_input_tokens_seen": 8249752, + "step": 30835 + }, + { + "epoch": 8.083879423328964, + "grad_norm": 1.7365909814834595, + "learning_rate": 5.388071481555448e-06, + "loss": 0.4433, + "num_input_tokens_seen": 8251384, + "step": 30840 + }, + { + "epoch": 8.08519003931848, + "grad_norm": 1.459088683128357, + "learning_rate": 5.3809806152039125e-06, + "loss": 0.7731, + "num_input_tokens_seen": 8252760, + "step": 30845 + }, + { + "epoch": 8.086500655307995, + "grad_norm": 4.8604655265808105, + "learning_rate": 5.373893855079235e-06, + "loss": 0.453, + "num_input_tokens_seen": 8253736, + "step": 30850 + }, + { + "epoch": 8.08781127129751, + "grad_norm": 4.759625434875488, + "learning_rate": 5.366811202664662e-06, + "loss": 0.4202, + "num_input_tokens_seen": 8255048, + "step": 30855 + }, + { + "epoch": 8.089121887287025, + "grad_norm": 2.7206339836120605, + "learning_rate": 5.35973265944259e-06, + "loss": 0.4141, + "num_input_tokens_seen": 8256216, + "step": 30860 + }, + { + "epoch": 8.09043250327654, + "grad_norm": 2.903944253921509, + "learning_rate": 5.352658226894527e-06, + "loss": 0.4624, + "num_input_tokens_seen": 8257768, + "step": 30865 + }, + { + "epoch": 8.091743119266056, + "grad_norm": 16.430402755737305, + "learning_rate": 5.345587906501148e-06, + "loss": 0.4064, + "num_input_tokens_seen": 8258792, + "step": 30870 + }, + { + "epoch": 8.09305373525557, + "grad_norm": 6.74241304397583, + "learning_rate": 5.338521699742263e-06, + "loss": 0.5499, + "num_input_tokens_seen": 8259912, + "step": 30875 + }, + { + "epoch": 8.094364351245085, + "grad_norm": 4.214361190795898, + "learning_rate": 5.331459608096815e-06, + "loss": 0.3789, + "num_input_tokens_seen": 8261240, + "step": 30880 + }, + { + "epoch": 8.0956749672346, + "grad_norm": 1.4251759052276611, + "learning_rate": 5.324401633042883e-06, + "loss": 0.3878, + "num_input_tokens_seen": 8263208, + "step": 30885 + }, + { + "epoch": 8.096985583224116, + "grad_norm": 2.486323118209839, + "learning_rate": 5.317347776057685e-06, + "loss": 0.3658, + "num_input_tokens_seen": 8264184, + "step": 30890 + }, + { + "epoch": 8.09829619921363, + "grad_norm": 2.6036183834075928, + "learning_rate": 5.310298038617606e-06, + "loss": 0.3926, + "num_input_tokens_seen": 8265672, + "step": 30895 + }, + { + "epoch": 8.099606815203146, + "grad_norm": 3.224684953689575, + "learning_rate": 5.303252422198115e-06, + "loss": 0.3699, + "num_input_tokens_seen": 8267160, + "step": 30900 + }, + { + "epoch": 8.100917431192661, + "grad_norm": 3.302833318710327, + "learning_rate": 5.296210928273862e-06, + "loss": 0.5558, + "num_input_tokens_seen": 8268856, + "step": 30905 + }, + { + "epoch": 8.102228047182175, + "grad_norm": 22.578248977661133, + "learning_rate": 5.289173558318617e-06, + "loss": 0.5581, + "num_input_tokens_seen": 8269928, + "step": 30910 + }, + { + "epoch": 8.10353866317169, + "grad_norm": 2.5167198181152344, + "learning_rate": 5.282140313805284e-06, + "loss": 0.417, + "num_input_tokens_seen": 8272056, + "step": 30915 + }, + { + "epoch": 8.104849279161206, + "grad_norm": 9.832356452941895, + "learning_rate": 5.275111196205917e-06, + "loss": 0.4265, + "num_input_tokens_seen": 8273272, + "step": 30920 + }, + { + "epoch": 8.106159895150721, + "grad_norm": 1.7284303903579712, + "learning_rate": 5.268086206991693e-06, + "loss": 0.46, + "num_input_tokens_seen": 8274840, + "step": 30925 + }, + { + "epoch": 8.107470511140235, + "grad_norm": 4.746047496795654, + "learning_rate": 5.2610653476329345e-06, + "loss": 0.5415, + "num_input_tokens_seen": 8276616, + "step": 30930 + }, + { + "epoch": 8.10878112712975, + "grad_norm": 2.752080202102661, + "learning_rate": 5.254048619599089e-06, + "loss": 0.469, + "num_input_tokens_seen": 8277752, + "step": 30935 + }, + { + "epoch": 8.110091743119266, + "grad_norm": 4.531562328338623, + "learning_rate": 5.247036024358759e-06, + "loss": 0.5971, + "num_input_tokens_seen": 8278936, + "step": 30940 + }, + { + "epoch": 8.111402359108782, + "grad_norm": 2.8487532138824463, + "learning_rate": 5.240027563379649e-06, + "loss": 0.5304, + "num_input_tokens_seen": 8279960, + "step": 30945 + }, + { + "epoch": 8.112712975098296, + "grad_norm": 8.91334342956543, + "learning_rate": 5.233023238128623e-06, + "loss": 0.4605, + "num_input_tokens_seen": 8281288, + "step": 30950 + }, + { + "epoch": 8.114023591087811, + "grad_norm": 3.2177188396453857, + "learning_rate": 5.226023050071682e-06, + "loss": 0.7151, + "num_input_tokens_seen": 8282504, + "step": 30955 + }, + { + "epoch": 8.115334207077327, + "grad_norm": 5.06614351272583, + "learning_rate": 5.219027000673954e-06, + "loss": 0.4241, + "num_input_tokens_seen": 8285368, + "step": 30960 + }, + { + "epoch": 8.116644823066842, + "grad_norm": 6.196769714355469, + "learning_rate": 5.212035091399694e-06, + "loss": 0.6148, + "num_input_tokens_seen": 8286616, + "step": 30965 + }, + { + "epoch": 8.117955439056356, + "grad_norm": 3.197248935699463, + "learning_rate": 5.205047323712298e-06, + "loss": 0.4575, + "num_input_tokens_seen": 8287848, + "step": 30970 + }, + { + "epoch": 8.119266055045872, + "grad_norm": 5.691487789154053, + "learning_rate": 5.198063699074293e-06, + "loss": 0.5061, + "num_input_tokens_seen": 8289704, + "step": 30975 + }, + { + "epoch": 8.120576671035387, + "grad_norm": 2.415482997894287, + "learning_rate": 5.191084218947351e-06, + "loss": 0.5388, + "num_input_tokens_seen": 8290872, + "step": 30980 + }, + { + "epoch": 8.1218872870249, + "grad_norm": 3.908787250518799, + "learning_rate": 5.184108884792244e-06, + "loss": 0.4034, + "num_input_tokens_seen": 8292248, + "step": 30985 + }, + { + "epoch": 8.123197903014416, + "grad_norm": 3.5224769115448, + "learning_rate": 5.177137698068912e-06, + "loss": 0.381, + "num_input_tokens_seen": 8293608, + "step": 30990 + }, + { + "epoch": 8.124508519003932, + "grad_norm": 12.329084396362305, + "learning_rate": 5.170170660236406e-06, + "loss": 0.4466, + "num_input_tokens_seen": 8294920, + "step": 30995 + }, + { + "epoch": 8.125819134993447, + "grad_norm": 3.243396759033203, + "learning_rate": 5.163207772752918e-06, + "loss": 0.5639, + "num_input_tokens_seen": 8296024, + "step": 31000 + }, + { + "epoch": 8.127129750982961, + "grad_norm": 5.594143867492676, + "learning_rate": 5.156249037075761e-06, + "loss": 0.3887, + "num_input_tokens_seen": 8297688, + "step": 31005 + }, + { + "epoch": 8.128440366972477, + "grad_norm": 3.566286563873291, + "learning_rate": 5.149294454661399e-06, + "loss": 0.442, + "num_input_tokens_seen": 8298920, + "step": 31010 + }, + { + "epoch": 8.129750982961992, + "grad_norm": 1.2951115369796753, + "learning_rate": 5.142344026965418e-06, + "loss": 0.4554, + "num_input_tokens_seen": 8300616, + "step": 31015 + }, + { + "epoch": 8.131061598951508, + "grad_norm": 3.689288377761841, + "learning_rate": 5.135397755442512e-06, + "loss": 0.412, + "num_input_tokens_seen": 8301720, + "step": 31020 + }, + { + "epoch": 8.132372214941022, + "grad_norm": 2.5456995964050293, + "learning_rate": 5.1284556415465315e-06, + "loss": 0.4611, + "num_input_tokens_seen": 8302872, + "step": 31025 + }, + { + "epoch": 8.133682830930537, + "grad_norm": 3.3529257774353027, + "learning_rate": 5.121517686730451e-06, + "loss": 0.5086, + "num_input_tokens_seen": 8304248, + "step": 31030 + }, + { + "epoch": 8.134993446920053, + "grad_norm": 4.180994510650635, + "learning_rate": 5.1145838924463725e-06, + "loss": 0.4393, + "num_input_tokens_seen": 8305560, + "step": 31035 + }, + { + "epoch": 8.136304062909568, + "grad_norm": 2.6065187454223633, + "learning_rate": 5.107654260145523e-06, + "loss": 0.4425, + "num_input_tokens_seen": 8306888, + "step": 31040 + }, + { + "epoch": 8.137614678899082, + "grad_norm": 2.565548896789551, + "learning_rate": 5.100728791278267e-06, + "loss": 0.3708, + "num_input_tokens_seen": 8308328, + "step": 31045 + }, + { + "epoch": 8.138925294888598, + "grad_norm": 1.267237901687622, + "learning_rate": 5.0938074872940915e-06, + "loss": 0.3078, + "num_input_tokens_seen": 8309464, + "step": 31050 + }, + { + "epoch": 8.140235910878113, + "grad_norm": 4.366708278656006, + "learning_rate": 5.086890349641616e-06, + "loss": 0.3167, + "num_input_tokens_seen": 8310728, + "step": 31055 + }, + { + "epoch": 8.141546526867629, + "grad_norm": 3.109163284301758, + "learning_rate": 5.079977379768588e-06, + "loss": 0.392, + "num_input_tokens_seen": 8312216, + "step": 31060 + }, + { + "epoch": 8.142857142857142, + "grad_norm": 2.7430171966552734, + "learning_rate": 5.073068579121862e-06, + "loss": 0.6054, + "num_input_tokens_seen": 8314120, + "step": 31065 + }, + { + "epoch": 8.144167758846658, + "grad_norm": 2.1253626346588135, + "learning_rate": 5.066163949147457e-06, + "loss": 0.6202, + "num_input_tokens_seen": 8315208, + "step": 31070 + }, + { + "epoch": 8.145478374836173, + "grad_norm": 3.6869919300079346, + "learning_rate": 5.059263491290495e-06, + "loss": 0.4268, + "num_input_tokens_seen": 8316280, + "step": 31075 + }, + { + "epoch": 8.146788990825687, + "grad_norm": 1.5461149215698242, + "learning_rate": 5.052367206995229e-06, + "loss": 0.4941, + "num_input_tokens_seen": 8317736, + "step": 31080 + }, + { + "epoch": 8.148099606815203, + "grad_norm": 1.6060377359390259, + "learning_rate": 5.0454750977050406e-06, + "loss": 0.5083, + "num_input_tokens_seen": 8318904, + "step": 31085 + }, + { + "epoch": 8.149410222804718, + "grad_norm": 4.238748073577881, + "learning_rate": 5.038587164862435e-06, + "loss": 0.5395, + "num_input_tokens_seen": 8320376, + "step": 31090 + }, + { + "epoch": 8.150720838794234, + "grad_norm": 2.9884321689605713, + "learning_rate": 5.031703409909052e-06, + "loss": 0.6657, + "num_input_tokens_seen": 8322056, + "step": 31095 + }, + { + "epoch": 8.152031454783748, + "grad_norm": 4.813848972320557, + "learning_rate": 5.024823834285636e-06, + "loss": 0.4596, + "num_input_tokens_seen": 8323688, + "step": 31100 + }, + { + "epoch": 8.153342070773263, + "grad_norm": 2.527625799179077, + "learning_rate": 5.017948439432077e-06, + "loss": 0.6436, + "num_input_tokens_seen": 8325000, + "step": 31105 + }, + { + "epoch": 8.154652686762779, + "grad_norm": 1.475007176399231, + "learning_rate": 5.0110772267873836e-06, + "loss": 0.3445, + "num_input_tokens_seen": 8325928, + "step": 31110 + }, + { + "epoch": 8.155963302752294, + "grad_norm": 3.146958112716675, + "learning_rate": 5.004210197789688e-06, + "loss": 0.4062, + "num_input_tokens_seen": 8327656, + "step": 31115 + }, + { + "epoch": 8.157273918741808, + "grad_norm": 3.9819226264953613, + "learning_rate": 4.997347353876242e-06, + "loss": 0.377, + "num_input_tokens_seen": 8328760, + "step": 31120 + }, + { + "epoch": 8.158584534731324, + "grad_norm": 6.710964202880859, + "learning_rate": 4.990488696483439e-06, + "loss": 0.4565, + "num_input_tokens_seen": 8330392, + "step": 31125 + }, + { + "epoch": 8.159895150720839, + "grad_norm": 0.7801979780197144, + "learning_rate": 4.983634227046785e-06, + "loss": 0.4148, + "num_input_tokens_seen": 8331976, + "step": 31130 + }, + { + "epoch": 8.161205766710355, + "grad_norm": 1.9496744871139526, + "learning_rate": 4.9767839470008935e-06, + "loss": 0.3603, + "num_input_tokens_seen": 8333192, + "step": 31135 + }, + { + "epoch": 8.162516382699868, + "grad_norm": 2.0289716720581055, + "learning_rate": 4.969937857779528e-06, + "loss": 0.4321, + "num_input_tokens_seen": 8334664, + "step": 31140 + }, + { + "epoch": 8.163826998689384, + "grad_norm": 5.142909049987793, + "learning_rate": 4.963095960815556e-06, + "loss": 0.5282, + "num_input_tokens_seen": 8335784, + "step": 31145 + }, + { + "epoch": 8.1651376146789, + "grad_norm": 2.4776711463928223, + "learning_rate": 4.95625825754098e-06, + "loss": 0.3648, + "num_input_tokens_seen": 8336888, + "step": 31150 + }, + { + "epoch": 8.166448230668415, + "grad_norm": 2.5538926124572754, + "learning_rate": 4.9494247493869165e-06, + "loss": 0.5543, + "num_input_tokens_seen": 8338024, + "step": 31155 + }, + { + "epoch": 8.167758846657929, + "grad_norm": 4.101945400238037, + "learning_rate": 4.942595437783609e-06, + "loss": 0.2688, + "num_input_tokens_seen": 8339368, + "step": 31160 + }, + { + "epoch": 8.169069462647444, + "grad_norm": 6.098082065582275, + "learning_rate": 4.9357703241604185e-06, + "loss": 0.3765, + "num_input_tokens_seen": 8340536, + "step": 31165 + }, + { + "epoch": 8.17038007863696, + "grad_norm": 3.303030252456665, + "learning_rate": 4.928949409945832e-06, + "loss": 0.3722, + "num_input_tokens_seen": 8341880, + "step": 31170 + }, + { + "epoch": 8.171690694626474, + "grad_norm": 1.9899882078170776, + "learning_rate": 4.922132696567464e-06, + "loss": 0.3046, + "num_input_tokens_seen": 8342904, + "step": 31175 + }, + { + "epoch": 8.17300131061599, + "grad_norm": 2.066087007522583, + "learning_rate": 4.915320185452013e-06, + "loss": 0.4685, + "num_input_tokens_seen": 8344328, + "step": 31180 + }, + { + "epoch": 8.174311926605505, + "grad_norm": 2.7956252098083496, + "learning_rate": 4.9085118780253515e-06, + "loss": 0.4296, + "num_input_tokens_seen": 8345864, + "step": 31185 + }, + { + "epoch": 8.17562254259502, + "grad_norm": 4.960803508758545, + "learning_rate": 4.9017077757124385e-06, + "loss": 0.5188, + "num_input_tokens_seen": 8347304, + "step": 31190 + }, + { + "epoch": 8.176933158584534, + "grad_norm": 5.772447109222412, + "learning_rate": 4.89490787993736e-06, + "loss": 0.3095, + "num_input_tokens_seen": 8348824, + "step": 31195 + }, + { + "epoch": 8.17824377457405, + "grad_norm": 1.6324362754821777, + "learning_rate": 4.888112192123323e-06, + "loss": 0.46, + "num_input_tokens_seen": 8350760, + "step": 31200 + }, + { + "epoch": 8.179554390563565, + "grad_norm": 5.312380790710449, + "learning_rate": 4.8813207136926555e-06, + "loss": 0.4933, + "num_input_tokens_seen": 8351928, + "step": 31205 + }, + { + "epoch": 8.18086500655308, + "grad_norm": 6.889612674713135, + "learning_rate": 4.8745334460668005e-06, + "loss": 0.5967, + "num_input_tokens_seen": 8353352, + "step": 31210 + }, + { + "epoch": 8.182175622542594, + "grad_norm": 8.272258758544922, + "learning_rate": 4.867750390666326e-06, + "loss": 0.4264, + "num_input_tokens_seen": 8354536, + "step": 31215 + }, + { + "epoch": 8.18348623853211, + "grad_norm": 4.3578925132751465, + "learning_rate": 4.860971548910903e-06, + "loss": 0.3908, + "num_input_tokens_seen": 8355736, + "step": 31220 + }, + { + "epoch": 8.184796854521625, + "grad_norm": 2.82590913772583, + "learning_rate": 4.854196922219337e-06, + "loss": 0.605, + "num_input_tokens_seen": 8357224, + "step": 31225 + }, + { + "epoch": 8.186107470511141, + "grad_norm": 3.047529458999634, + "learning_rate": 4.847426512009545e-06, + "loss": 0.5108, + "num_input_tokens_seen": 8358696, + "step": 31230 + }, + { + "epoch": 8.187418086500655, + "grad_norm": 2.1943624019622803, + "learning_rate": 4.840660319698559e-06, + "loss": 0.4396, + "num_input_tokens_seen": 8360040, + "step": 31235 + }, + { + "epoch": 8.18872870249017, + "grad_norm": 2.5206313133239746, + "learning_rate": 4.8338983467025394e-06, + "loss": 0.3772, + "num_input_tokens_seen": 8361032, + "step": 31240 + }, + { + "epoch": 8.190039318479686, + "grad_norm": 6.659091949462891, + "learning_rate": 4.827140594436752e-06, + "loss": 0.4349, + "num_input_tokens_seen": 8362296, + "step": 31245 + }, + { + "epoch": 8.191349934469201, + "grad_norm": 2.225346803665161, + "learning_rate": 4.82038706431559e-06, + "loss": 0.4159, + "num_input_tokens_seen": 8363560, + "step": 31250 + }, + { + "epoch": 8.192660550458715, + "grad_norm": 2.1384024620056152, + "learning_rate": 4.81363775775254e-06, + "loss": 0.456, + "num_input_tokens_seen": 8365064, + "step": 31255 + }, + { + "epoch": 8.19397116644823, + "grad_norm": 2.4348480701446533, + "learning_rate": 4.80689267616023e-06, + "loss": 0.5532, + "num_input_tokens_seen": 8366712, + "step": 31260 + }, + { + "epoch": 8.195281782437746, + "grad_norm": 2.843230724334717, + "learning_rate": 4.800151820950391e-06, + "loss": 0.624, + "num_input_tokens_seen": 8368024, + "step": 31265 + }, + { + "epoch": 8.19659239842726, + "grad_norm": 3.857541561126709, + "learning_rate": 4.7934151935338706e-06, + "loss": 0.7462, + "num_input_tokens_seen": 8369640, + "step": 31270 + }, + { + "epoch": 8.197903014416775, + "grad_norm": 2.46854829788208, + "learning_rate": 4.786682795320638e-06, + "loss": 0.3301, + "num_input_tokens_seen": 8371096, + "step": 31275 + }, + { + "epoch": 8.199213630406291, + "grad_norm": 3.4151766300201416, + "learning_rate": 4.779954627719771e-06, + "loss": 0.4703, + "num_input_tokens_seen": 8372440, + "step": 31280 + }, + { + "epoch": 8.200524246395807, + "grad_norm": 4.499403953552246, + "learning_rate": 4.773230692139463e-06, + "loss": 0.37, + "num_input_tokens_seen": 8373464, + "step": 31285 + }, + { + "epoch": 8.20183486238532, + "grad_norm": 3.475187301635742, + "learning_rate": 4.766510989987022e-06, + "loss": 0.5716, + "num_input_tokens_seen": 8374952, + "step": 31290 + }, + { + "epoch": 8.203145478374836, + "grad_norm": 1.9425688982009888, + "learning_rate": 4.7597955226688666e-06, + "loss": 0.5578, + "num_input_tokens_seen": 8377976, + "step": 31295 + }, + { + "epoch": 8.204456094364351, + "grad_norm": 2.8150219917297363, + "learning_rate": 4.75308429159054e-06, + "loss": 0.3447, + "num_input_tokens_seen": 8379192, + "step": 31300 + }, + { + "epoch": 8.205766710353867, + "grad_norm": 2.9993274211883545, + "learning_rate": 4.746377298156685e-06, + "loss": 0.5962, + "num_input_tokens_seen": 8380424, + "step": 31305 + }, + { + "epoch": 8.20707732634338, + "grad_norm": 2.9506542682647705, + "learning_rate": 4.739674543771066e-06, + "loss": 0.4438, + "num_input_tokens_seen": 8381656, + "step": 31310 + }, + { + "epoch": 8.208387942332896, + "grad_norm": 2.2332539558410645, + "learning_rate": 4.732976029836561e-06, + "loss": 0.4318, + "num_input_tokens_seen": 8382648, + "step": 31315 + }, + { + "epoch": 8.209698558322412, + "grad_norm": 14.19393539428711, + "learning_rate": 4.726281757755149e-06, + "loss": 0.3716, + "num_input_tokens_seen": 8385736, + "step": 31320 + }, + { + "epoch": 8.211009174311927, + "grad_norm": 2.793254852294922, + "learning_rate": 4.719591728927938e-06, + "loss": 0.6616, + "num_input_tokens_seen": 8386904, + "step": 31325 + }, + { + "epoch": 8.212319790301441, + "grad_norm": 3.7740345001220703, + "learning_rate": 4.712905944755144e-06, + "loss": 0.3484, + "num_input_tokens_seen": 8387944, + "step": 31330 + }, + { + "epoch": 8.213630406290957, + "grad_norm": 4.21055793762207, + "learning_rate": 4.706224406636072e-06, + "loss": 0.4483, + "num_input_tokens_seen": 8389288, + "step": 31335 + }, + { + "epoch": 8.214941022280472, + "grad_norm": 3.6894991397857666, + "learning_rate": 4.699547115969172e-06, + "loss": 0.5163, + "num_input_tokens_seen": 8390664, + "step": 31340 + }, + { + "epoch": 8.216251638269988, + "grad_norm": 0.9001460075378418, + "learning_rate": 4.692874074151979e-06, + "loss": 0.801, + "num_input_tokens_seen": 8391928, + "step": 31345 + }, + { + "epoch": 8.217562254259501, + "grad_norm": 1.8720333576202393, + "learning_rate": 4.686205282581152e-06, + "loss": 0.5682, + "num_input_tokens_seen": 8393128, + "step": 31350 + }, + { + "epoch": 8.218872870249017, + "grad_norm": 2.9097046852111816, + "learning_rate": 4.679540742652463e-06, + "loss": 0.5224, + "num_input_tokens_seen": 8394040, + "step": 31355 + }, + { + "epoch": 8.220183486238533, + "grad_norm": 1.5918338298797607, + "learning_rate": 4.672880455760786e-06, + "loss": 0.4984, + "num_input_tokens_seen": 8395688, + "step": 31360 + }, + { + "epoch": 8.221494102228046, + "grad_norm": 6.3405656814575195, + "learning_rate": 4.666224423300114e-06, + "loss": 0.5538, + "num_input_tokens_seen": 8397592, + "step": 31365 + }, + { + "epoch": 8.222804718217562, + "grad_norm": 4.859597682952881, + "learning_rate": 4.659572646663532e-06, + "loss": 0.4397, + "num_input_tokens_seen": 8398712, + "step": 31370 + }, + { + "epoch": 8.224115334207077, + "grad_norm": 6.089107990264893, + "learning_rate": 4.652925127243246e-06, + "loss": 0.5105, + "num_input_tokens_seen": 8399816, + "step": 31375 + }, + { + "epoch": 8.225425950196593, + "grad_norm": 2.814331531524658, + "learning_rate": 4.646281866430574e-06, + "loss": 0.4128, + "num_input_tokens_seen": 8401064, + "step": 31380 + }, + { + "epoch": 8.226736566186107, + "grad_norm": 4.599028587341309, + "learning_rate": 4.6396428656159424e-06, + "loss": 0.4534, + "num_input_tokens_seen": 8402088, + "step": 31385 + }, + { + "epoch": 8.228047182175622, + "grad_norm": 3.755464553833008, + "learning_rate": 4.633008126188876e-06, + "loss": 0.3399, + "num_input_tokens_seen": 8403032, + "step": 31390 + }, + { + "epoch": 8.229357798165138, + "grad_norm": 3.8630495071411133, + "learning_rate": 4.626377649538019e-06, + "loss": 0.69, + "num_input_tokens_seen": 8404600, + "step": 31395 + }, + { + "epoch": 8.230668414154653, + "grad_norm": 3.408644199371338, + "learning_rate": 4.619751437051114e-06, + "loss": 0.5296, + "num_input_tokens_seen": 8405896, + "step": 31400 + }, + { + "epoch": 8.231979030144167, + "grad_norm": 1.2820957899093628, + "learning_rate": 4.613129490115023e-06, + "loss": 0.4122, + "num_input_tokens_seen": 8407608, + "step": 31405 + }, + { + "epoch": 8.233289646133683, + "grad_norm": 14.836932182312012, + "learning_rate": 4.6065118101157014e-06, + "loss": 0.5448, + "num_input_tokens_seen": 8409000, + "step": 31410 + }, + { + "epoch": 8.234600262123198, + "grad_norm": 5.640592098236084, + "learning_rate": 4.599898398438221e-06, + "loss": 0.478, + "num_input_tokens_seen": 8409896, + "step": 31415 + }, + { + "epoch": 8.235910878112714, + "grad_norm": 2.176564931869507, + "learning_rate": 4.593289256466757e-06, + "loss": 0.3882, + "num_input_tokens_seen": 8411288, + "step": 31420 + }, + { + "epoch": 8.237221494102227, + "grad_norm": 16.35874366760254, + "learning_rate": 4.586684385584592e-06, + "loss": 0.4878, + "num_input_tokens_seen": 8412440, + "step": 31425 + }, + { + "epoch": 8.238532110091743, + "grad_norm": 2.0917375087738037, + "learning_rate": 4.580083787174114e-06, + "loss": 0.3931, + "num_input_tokens_seen": 8413752, + "step": 31430 + }, + { + "epoch": 8.239842726081259, + "grad_norm": 5.110379219055176, + "learning_rate": 4.573487462616815e-06, + "loss": 0.7097, + "num_input_tokens_seen": 8415144, + "step": 31435 + }, + { + "epoch": 8.241153342070774, + "grad_norm": 4.915318012237549, + "learning_rate": 4.566895413293298e-06, + "loss": 0.4047, + "num_input_tokens_seen": 8416952, + "step": 31440 + }, + { + "epoch": 8.242463958060288, + "grad_norm": 3.4541513919830322, + "learning_rate": 4.560307640583264e-06, + "loss": 0.6292, + "num_input_tokens_seen": 8418232, + "step": 31445 + }, + { + "epoch": 8.243774574049803, + "grad_norm": 3.4159302711486816, + "learning_rate": 4.553724145865529e-06, + "loss": 0.4606, + "num_input_tokens_seen": 8419672, + "step": 31450 + }, + { + "epoch": 8.245085190039319, + "grad_norm": 1.815692663192749, + "learning_rate": 4.547144930517996e-06, + "loss": 0.5568, + "num_input_tokens_seen": 8421080, + "step": 31455 + }, + { + "epoch": 8.246395806028833, + "grad_norm": 1.7038661241531372, + "learning_rate": 4.54056999591769e-06, + "loss": 0.7762, + "num_input_tokens_seen": 8422216, + "step": 31460 + }, + { + "epoch": 8.247706422018348, + "grad_norm": 3.931739091873169, + "learning_rate": 4.533999343440728e-06, + "loss": 0.6007, + "num_input_tokens_seen": 8424072, + "step": 31465 + }, + { + "epoch": 8.249017038007864, + "grad_norm": 5.089694976806641, + "learning_rate": 4.527432974462345e-06, + "loss": 0.3754, + "num_input_tokens_seen": 8425400, + "step": 31470 + }, + { + "epoch": 8.25032765399738, + "grad_norm": 7.945149898529053, + "learning_rate": 4.520870890356868e-06, + "loss": 0.4558, + "num_input_tokens_seen": 8426312, + "step": 31475 + }, + { + "epoch": 8.251638269986893, + "grad_norm": 6.199655055999756, + "learning_rate": 4.5143130924977265e-06, + "loss": 0.3774, + "num_input_tokens_seen": 8428248, + "step": 31480 + }, + { + "epoch": 8.252948885976409, + "grad_norm": 3.6786983013153076, + "learning_rate": 4.5077595822574645e-06, + "loss": 0.4348, + "num_input_tokens_seen": 8430056, + "step": 31485 + }, + { + "epoch": 8.254259501965924, + "grad_norm": 1.7846136093139648, + "learning_rate": 4.50121036100771e-06, + "loss": 0.4599, + "num_input_tokens_seen": 8431416, + "step": 31490 + }, + { + "epoch": 8.25557011795544, + "grad_norm": 4.295989513397217, + "learning_rate": 4.494665430119208e-06, + "loss": 0.5112, + "num_input_tokens_seen": 8432632, + "step": 31495 + }, + { + "epoch": 8.256880733944953, + "grad_norm": 3.371737241744995, + "learning_rate": 4.488124790961798e-06, + "loss": 0.5065, + "num_input_tokens_seen": 8433992, + "step": 31500 + }, + { + "epoch": 8.258191349934469, + "grad_norm": 10.546168327331543, + "learning_rate": 4.4815884449044275e-06, + "loss": 0.2818, + "num_input_tokens_seen": 8435528, + "step": 31505 + }, + { + "epoch": 8.259501965923985, + "grad_norm": 3.7903289794921875, + "learning_rate": 4.475056393315144e-06, + "loss": 0.5449, + "num_input_tokens_seen": 8436904, + "step": 31510 + }, + { + "epoch": 8.2608125819135, + "grad_norm": 3.903442859649658, + "learning_rate": 4.468528637561095e-06, + "loss": 0.4958, + "num_input_tokens_seen": 8437976, + "step": 31515 + }, + { + "epoch": 8.262123197903014, + "grad_norm": 2.0650715827941895, + "learning_rate": 4.462005179008525e-06, + "loss": 0.4457, + "num_input_tokens_seen": 8439128, + "step": 31520 + }, + { + "epoch": 8.26343381389253, + "grad_norm": 2.3499932289123535, + "learning_rate": 4.455486019022781e-06, + "loss": 0.5198, + "num_input_tokens_seen": 8440936, + "step": 31525 + }, + { + "epoch": 8.264744429882045, + "grad_norm": 4.089449882507324, + "learning_rate": 4.448971158968318e-06, + "loss": 0.3814, + "num_input_tokens_seen": 8442312, + "step": 31530 + }, + { + "epoch": 8.26605504587156, + "grad_norm": 7.0092573165893555, + "learning_rate": 4.4424606002086814e-06, + "loss": 0.5041, + "num_input_tokens_seen": 8443640, + "step": 31535 + }, + { + "epoch": 8.267365661861074, + "grad_norm": 2.0962655544281006, + "learning_rate": 4.435954344106522e-06, + "loss": 0.4645, + "num_input_tokens_seen": 8445336, + "step": 31540 + }, + { + "epoch": 8.26867627785059, + "grad_norm": 5.4282073974609375, + "learning_rate": 4.429452392023584e-06, + "loss": 0.4948, + "num_input_tokens_seen": 8446616, + "step": 31545 + }, + { + "epoch": 8.269986893840105, + "grad_norm": 2.9598047733306885, + "learning_rate": 4.4229547453207175e-06, + "loss": 0.5403, + "num_input_tokens_seen": 8447880, + "step": 31550 + }, + { + "epoch": 8.271297509829619, + "grad_norm": 3.3853328227996826, + "learning_rate": 4.416461405357869e-06, + "loss": 0.4832, + "num_input_tokens_seen": 8449144, + "step": 31555 + }, + { + "epoch": 8.272608125819135, + "grad_norm": 12.027291297912598, + "learning_rate": 4.409972373494084e-06, + "loss": 0.3377, + "num_input_tokens_seen": 8450232, + "step": 31560 + }, + { + "epoch": 8.27391874180865, + "grad_norm": 3.4063258171081543, + "learning_rate": 4.403487651087509e-06, + "loss": 0.5036, + "num_input_tokens_seen": 8451320, + "step": 31565 + }, + { + "epoch": 8.275229357798166, + "grad_norm": 2.8224592208862305, + "learning_rate": 4.397007239495376e-06, + "loss": 0.3631, + "num_input_tokens_seen": 8452712, + "step": 31570 + }, + { + "epoch": 8.27653997378768, + "grad_norm": 3.1220455169677734, + "learning_rate": 4.390531140074028e-06, + "loss": 0.5314, + "num_input_tokens_seen": 8454040, + "step": 31575 + }, + { + "epoch": 8.277850589777195, + "grad_norm": 4.146581649780273, + "learning_rate": 4.384059354178893e-06, + "loss": 0.3451, + "num_input_tokens_seen": 8455016, + "step": 31580 + }, + { + "epoch": 8.27916120576671, + "grad_norm": 14.519503593444824, + "learning_rate": 4.377591883164522e-06, + "loss": 0.3154, + "num_input_tokens_seen": 8457064, + "step": 31585 + }, + { + "epoch": 8.280471821756226, + "grad_norm": 6.366556644439697, + "learning_rate": 4.371128728384538e-06, + "loss": 0.5731, + "num_input_tokens_seen": 8458536, + "step": 31590 + }, + { + "epoch": 8.28178243774574, + "grad_norm": 1.361627459526062, + "learning_rate": 4.364669891191667e-06, + "loss": 0.3367, + "num_input_tokens_seen": 8460024, + "step": 31595 + }, + { + "epoch": 8.283093053735255, + "grad_norm": 1.4951213598251343, + "learning_rate": 4.358215372937738e-06, + "loss": 0.4112, + "num_input_tokens_seen": 8461224, + "step": 31600 + }, + { + "epoch": 8.284403669724771, + "grad_norm": 2.1103100776672363, + "learning_rate": 4.35176517497366e-06, + "loss": 0.3963, + "num_input_tokens_seen": 8462456, + "step": 31605 + }, + { + "epoch": 8.285714285714286, + "grad_norm": 4.670657634735107, + "learning_rate": 4.34531929864945e-06, + "loss": 0.3779, + "num_input_tokens_seen": 8463400, + "step": 31610 + }, + { + "epoch": 8.2870249017038, + "grad_norm": 2.3284993171691895, + "learning_rate": 4.3388777453142265e-06, + "loss": 0.4359, + "num_input_tokens_seen": 8464584, + "step": 31615 + }, + { + "epoch": 8.288335517693316, + "grad_norm": 4.38856840133667, + "learning_rate": 4.332440516316188e-06, + "loss": 0.5717, + "num_input_tokens_seen": 8466072, + "step": 31620 + }, + { + "epoch": 8.289646133682831, + "grad_norm": 5.673497200012207, + "learning_rate": 4.326007613002642e-06, + "loss": 0.4423, + "num_input_tokens_seen": 8467512, + "step": 31625 + }, + { + "epoch": 8.290956749672347, + "grad_norm": 4.113978385925293, + "learning_rate": 4.319579036719979e-06, + "loss": 0.4057, + "num_input_tokens_seen": 8468584, + "step": 31630 + }, + { + "epoch": 8.29226736566186, + "grad_norm": 1.5557422637939453, + "learning_rate": 4.313154788813684e-06, + "loss": 0.2806, + "num_input_tokens_seen": 8470376, + "step": 31635 + }, + { + "epoch": 8.293577981651376, + "grad_norm": 2.9459972381591797, + "learning_rate": 4.30673487062836e-06, + "loss": 0.3146, + "num_input_tokens_seen": 8471720, + "step": 31640 + }, + { + "epoch": 8.294888597640892, + "grad_norm": 10.468439102172852, + "learning_rate": 4.300319283507667e-06, + "loss": 0.3236, + "num_input_tokens_seen": 8472616, + "step": 31645 + }, + { + "epoch": 8.296199213630405, + "grad_norm": 5.118487358093262, + "learning_rate": 4.29390802879438e-06, + "loss": 0.6389, + "num_input_tokens_seen": 8473528, + "step": 31650 + }, + { + "epoch": 8.297509829619921, + "grad_norm": 1.9927204847335815, + "learning_rate": 4.287501107830366e-06, + "loss": 0.5753, + "num_input_tokens_seen": 8474648, + "step": 31655 + }, + { + "epoch": 8.298820445609437, + "grad_norm": 6.383204936981201, + "learning_rate": 4.2810985219565835e-06, + "loss": 0.2783, + "num_input_tokens_seen": 8475688, + "step": 31660 + }, + { + "epoch": 8.300131061598952, + "grad_norm": 3.9344098567962646, + "learning_rate": 4.27470027251308e-06, + "loss": 0.588, + "num_input_tokens_seen": 8476952, + "step": 31665 + }, + { + "epoch": 8.301441677588466, + "grad_norm": 9.445164680480957, + "learning_rate": 4.268306360838998e-06, + "loss": 0.4759, + "num_input_tokens_seen": 8477992, + "step": 31670 + }, + { + "epoch": 8.302752293577981, + "grad_norm": 2.804105281829834, + "learning_rate": 4.261916788272571e-06, + "loss": 0.5937, + "num_input_tokens_seen": 8479304, + "step": 31675 + }, + { + "epoch": 8.304062909567497, + "grad_norm": 3.799840211868286, + "learning_rate": 4.2555315561511326e-06, + "loss": 0.3303, + "num_input_tokens_seen": 8480936, + "step": 31680 + }, + { + "epoch": 8.305373525557012, + "grad_norm": 1.3610645532608032, + "learning_rate": 4.249150665811097e-06, + "loss": 0.477, + "num_input_tokens_seen": 8482280, + "step": 31685 + }, + { + "epoch": 8.306684141546526, + "grad_norm": 2.124208927154541, + "learning_rate": 4.242774118587969e-06, + "loss": 0.4786, + "num_input_tokens_seen": 8483528, + "step": 31690 + }, + { + "epoch": 8.307994757536042, + "grad_norm": 2.15779447555542, + "learning_rate": 4.236401915816346e-06, + "loss": 0.4569, + "num_input_tokens_seen": 8485112, + "step": 31695 + }, + { + "epoch": 8.309305373525557, + "grad_norm": 4.728836536407471, + "learning_rate": 4.230034058829932e-06, + "loss": 0.3881, + "num_input_tokens_seen": 8486568, + "step": 31700 + }, + { + "epoch": 8.310615989515073, + "grad_norm": 2.166679859161377, + "learning_rate": 4.2236705489615e-06, + "loss": 0.4305, + "num_input_tokens_seen": 8487976, + "step": 31705 + }, + { + "epoch": 8.311926605504587, + "grad_norm": 2.501723527908325, + "learning_rate": 4.2173113875429246e-06, + "loss": 0.4164, + "num_input_tokens_seen": 8489304, + "step": 31710 + }, + { + "epoch": 8.313237221494102, + "grad_norm": 2.444606304168701, + "learning_rate": 4.2109565759051645e-06, + "loss": 0.4514, + "num_input_tokens_seen": 8490392, + "step": 31715 + }, + { + "epoch": 8.314547837483618, + "grad_norm": 3.425015449523926, + "learning_rate": 4.204606115378282e-06, + "loss": 0.482, + "num_input_tokens_seen": 8491448, + "step": 31720 + }, + { + "epoch": 8.315858453473133, + "grad_norm": 2.320739269256592, + "learning_rate": 4.198260007291399e-06, + "loss": 0.7761, + "num_input_tokens_seen": 8492616, + "step": 31725 + }, + { + "epoch": 8.317169069462647, + "grad_norm": 5.742697238922119, + "learning_rate": 4.191918252972754e-06, + "loss": 1.4282, + "num_input_tokens_seen": 8493624, + "step": 31730 + }, + { + "epoch": 8.318479685452163, + "grad_norm": 3.0274643898010254, + "learning_rate": 4.185580853749666e-06, + "loss": 0.3041, + "num_input_tokens_seen": 8494792, + "step": 31735 + }, + { + "epoch": 8.319790301441678, + "grad_norm": 3.264674186706543, + "learning_rate": 4.179247810948539e-06, + "loss": 0.4629, + "num_input_tokens_seen": 8495864, + "step": 31740 + }, + { + "epoch": 8.321100917431192, + "grad_norm": 1.3096957206726074, + "learning_rate": 4.172919125894869e-06, + "loss": 0.4571, + "num_input_tokens_seen": 8498136, + "step": 31745 + }, + { + "epoch": 8.322411533420707, + "grad_norm": 4.423628330230713, + "learning_rate": 4.166594799913237e-06, + "loss": 0.59, + "num_input_tokens_seen": 8499448, + "step": 31750 + }, + { + "epoch": 8.323722149410223, + "grad_norm": 1.9373342990875244, + "learning_rate": 4.160274834327327e-06, + "loss": 0.5461, + "num_input_tokens_seen": 8500968, + "step": 31755 + }, + { + "epoch": 8.325032765399738, + "grad_norm": 2.7524242401123047, + "learning_rate": 4.153959230459878e-06, + "loss": 0.5095, + "num_input_tokens_seen": 8502584, + "step": 31760 + }, + { + "epoch": 8.326343381389252, + "grad_norm": 2.366374969482422, + "learning_rate": 4.147647989632744e-06, + "loss": 0.4408, + "num_input_tokens_seen": 8503608, + "step": 31765 + }, + { + "epoch": 8.327653997378768, + "grad_norm": 1.4142696857452393, + "learning_rate": 4.141341113166855e-06, + "loss": 0.5951, + "num_input_tokens_seen": 8505032, + "step": 31770 + }, + { + "epoch": 8.328964613368283, + "grad_norm": 1.98081374168396, + "learning_rate": 4.135038602382232e-06, + "loss": 0.433, + "num_input_tokens_seen": 8506552, + "step": 31775 + }, + { + "epoch": 8.330275229357799, + "grad_norm": 1.9638701677322388, + "learning_rate": 4.1287404585979764e-06, + "loss": 0.3264, + "num_input_tokens_seen": 8507912, + "step": 31780 + }, + { + "epoch": 8.331585845347313, + "grad_norm": 6.416074275970459, + "learning_rate": 4.122446683132281e-06, + "loss": 0.4834, + "num_input_tokens_seen": 8509416, + "step": 31785 + }, + { + "epoch": 8.332896461336828, + "grad_norm": 5.318577766418457, + "learning_rate": 4.116157277302418e-06, + "loss": 0.405, + "num_input_tokens_seen": 8511064, + "step": 31790 + }, + { + "epoch": 8.334207077326344, + "grad_norm": 3.1984665393829346, + "learning_rate": 4.109872242424756e-06, + "loss": 0.4091, + "num_input_tokens_seen": 8514520, + "step": 31795 + }, + { + "epoch": 8.33551769331586, + "grad_norm": 1.4141638278961182, + "learning_rate": 4.1035915798147415e-06, + "loss": 0.6046, + "num_input_tokens_seen": 8516152, + "step": 31800 + }, + { + "epoch": 8.336828309305373, + "grad_norm": 2.4779815673828125, + "learning_rate": 4.097315290786899e-06, + "loss": 0.3518, + "num_input_tokens_seen": 8517416, + "step": 31805 + }, + { + "epoch": 8.338138925294889, + "grad_norm": 6.931008338928223, + "learning_rate": 4.091043376654843e-06, + "loss": 0.4175, + "num_input_tokens_seen": 8518392, + "step": 31810 + }, + { + "epoch": 8.339449541284404, + "grad_norm": 2.411073684692383, + "learning_rate": 4.084775838731286e-06, + "loss": 0.884, + "num_input_tokens_seen": 8519688, + "step": 31815 + }, + { + "epoch": 8.34076015727392, + "grad_norm": 5.362305641174316, + "learning_rate": 4.078512678328009e-06, + "loss": 0.309, + "num_input_tokens_seen": 8520968, + "step": 31820 + }, + { + "epoch": 8.342070773263433, + "grad_norm": 3.2805020809173584, + "learning_rate": 4.072253896755876e-06, + "loss": 0.494, + "num_input_tokens_seen": 8522344, + "step": 31825 + }, + { + "epoch": 8.343381389252949, + "grad_norm": 1.5567680597305298, + "learning_rate": 4.065999495324843e-06, + "loss": 0.3931, + "num_input_tokens_seen": 8523864, + "step": 31830 + }, + { + "epoch": 8.344692005242464, + "grad_norm": 1.9394618272781372, + "learning_rate": 4.059749475343949e-06, + "loss": 0.4157, + "num_input_tokens_seen": 8525224, + "step": 31835 + }, + { + "epoch": 8.346002621231978, + "grad_norm": 4.230229377746582, + "learning_rate": 4.053503838121306e-06, + "loss": 0.2941, + "num_input_tokens_seen": 8526904, + "step": 31840 + }, + { + "epoch": 8.347313237221494, + "grad_norm": 2.393202781677246, + "learning_rate": 4.047262584964112e-06, + "loss": 0.6056, + "num_input_tokens_seen": 8528392, + "step": 31845 + }, + { + "epoch": 8.34862385321101, + "grad_norm": 2.045015335083008, + "learning_rate": 4.0410257171786595e-06, + "loss": 0.4783, + "num_input_tokens_seen": 8529752, + "step": 31850 + }, + { + "epoch": 8.349934469200525, + "grad_norm": 1.471076250076294, + "learning_rate": 4.034793236070308e-06, + "loss": 0.5095, + "num_input_tokens_seen": 8531192, + "step": 31855 + }, + { + "epoch": 8.351245085190039, + "grad_norm": 3.8090901374816895, + "learning_rate": 4.028565142943508e-06, + "loss": 0.4232, + "num_input_tokens_seen": 8532264, + "step": 31860 + }, + { + "epoch": 8.352555701179554, + "grad_norm": 2.917130708694458, + "learning_rate": 4.022341439101782e-06, + "loss": 0.3452, + "num_input_tokens_seen": 8533416, + "step": 31865 + }, + { + "epoch": 8.35386631716907, + "grad_norm": 6.850009441375732, + "learning_rate": 4.016122125847751e-06, + "loss": 0.4739, + "num_input_tokens_seen": 8534552, + "step": 31870 + }, + { + "epoch": 8.355176933158585, + "grad_norm": 4.643163204193115, + "learning_rate": 4.009907204483107e-06, + "loss": 0.5271, + "num_input_tokens_seen": 8535608, + "step": 31875 + }, + { + "epoch": 8.356487549148099, + "grad_norm": 1.7884303331375122, + "learning_rate": 4.003696676308613e-06, + "loss": 0.4667, + "num_input_tokens_seen": 8537160, + "step": 31880 + }, + { + "epoch": 8.357798165137615, + "grad_norm": 4.989171028137207, + "learning_rate": 3.997490542624124e-06, + "loss": 0.422, + "num_input_tokens_seen": 8538728, + "step": 31885 + }, + { + "epoch": 8.35910878112713, + "grad_norm": 2.2009999752044678, + "learning_rate": 3.991288804728577e-06, + "loss": 0.497, + "num_input_tokens_seen": 8539752, + "step": 31890 + }, + { + "epoch": 8.360419397116646, + "grad_norm": 1.3282700777053833, + "learning_rate": 3.985091463919982e-06, + "loss": 0.4696, + "num_input_tokens_seen": 8540952, + "step": 31895 + }, + { + "epoch": 8.36173001310616, + "grad_norm": 7.789153575897217, + "learning_rate": 3.978898521495433e-06, + "loss": 0.5362, + "num_input_tokens_seen": 8542040, + "step": 31900 + }, + { + "epoch": 8.363040629095675, + "grad_norm": 7.54129695892334, + "learning_rate": 3.972709978751099e-06, + "loss": 0.3541, + "num_input_tokens_seen": 8543208, + "step": 31905 + }, + { + "epoch": 8.36435124508519, + "grad_norm": 5.6150593757629395, + "learning_rate": 3.966525836982237e-06, + "loss": 0.3732, + "num_input_tokens_seen": 8544104, + "step": 31910 + }, + { + "epoch": 8.365661861074706, + "grad_norm": 2.431480884552002, + "learning_rate": 3.960346097483175e-06, + "loss": 0.4031, + "num_input_tokens_seen": 8545256, + "step": 31915 + }, + { + "epoch": 8.36697247706422, + "grad_norm": 3.4237143993377686, + "learning_rate": 3.954170761547324e-06, + "loss": 0.2065, + "num_input_tokens_seen": 8546696, + "step": 31920 + }, + { + "epoch": 8.368283093053735, + "grad_norm": 2.673034191131592, + "learning_rate": 3.947999830467159e-06, + "loss": 0.4112, + "num_input_tokens_seen": 8547944, + "step": 31925 + }, + { + "epoch": 8.36959370904325, + "grad_norm": 3.61987042427063, + "learning_rate": 3.941833305534259e-06, + "loss": 0.7182, + "num_input_tokens_seen": 8549096, + "step": 31930 + }, + { + "epoch": 8.370904325032765, + "grad_norm": 6.776269912719727, + "learning_rate": 3.935671188039264e-06, + "loss": 0.6403, + "num_input_tokens_seen": 8550152, + "step": 31935 + }, + { + "epoch": 8.37221494102228, + "grad_norm": 3.3360769748687744, + "learning_rate": 3.9295134792718905e-06, + "loss": 0.295, + "num_input_tokens_seen": 8551704, + "step": 31940 + }, + { + "epoch": 8.373525557011796, + "grad_norm": 3.085723638534546, + "learning_rate": 3.923360180520941e-06, + "loss": 0.5037, + "num_input_tokens_seen": 8553720, + "step": 31945 + }, + { + "epoch": 8.374836173001311, + "grad_norm": 2.863191843032837, + "learning_rate": 3.917211293074285e-06, + "loss": 0.3593, + "num_input_tokens_seen": 8555000, + "step": 31950 + }, + { + "epoch": 8.376146788990825, + "grad_norm": 1.1201870441436768, + "learning_rate": 3.911066818218886e-06, + "loss": 0.3898, + "num_input_tokens_seen": 8556408, + "step": 31955 + }, + { + "epoch": 8.37745740498034, + "grad_norm": 4.6715617179870605, + "learning_rate": 3.9049267572407565e-06, + "loss": 0.5561, + "num_input_tokens_seen": 8557720, + "step": 31960 + }, + { + "epoch": 8.378768020969856, + "grad_norm": 6.724741458892822, + "learning_rate": 3.898791111425007e-06, + "loss": 0.4773, + "num_input_tokens_seen": 8559192, + "step": 31965 + }, + { + "epoch": 8.380078636959372, + "grad_norm": 3.398831844329834, + "learning_rate": 3.892659882055819e-06, + "loss": 0.5357, + "num_input_tokens_seen": 8560408, + "step": 31970 + }, + { + "epoch": 8.381389252948885, + "grad_norm": 4.6850905418396, + "learning_rate": 3.886533070416445e-06, + "loss": 0.5672, + "num_input_tokens_seen": 8561560, + "step": 31975 + }, + { + "epoch": 8.382699868938401, + "grad_norm": 2.250779867172241, + "learning_rate": 3.880410677789214e-06, + "loss": 0.5018, + "num_input_tokens_seen": 8562712, + "step": 31980 + }, + { + "epoch": 8.384010484927916, + "grad_norm": 0.9032430052757263, + "learning_rate": 3.87429270545554e-06, + "loss": 0.3483, + "num_input_tokens_seen": 8564072, + "step": 31985 + }, + { + "epoch": 8.385321100917432, + "grad_norm": 8.734784126281738, + "learning_rate": 3.868179154695909e-06, + "loss": 0.4269, + "num_input_tokens_seen": 8565144, + "step": 31990 + }, + { + "epoch": 8.386631716906946, + "grad_norm": 3.6495237350463867, + "learning_rate": 3.86207002678986e-06, + "loss": 0.6328, + "num_input_tokens_seen": 8566520, + "step": 31995 + }, + { + "epoch": 8.387942332896461, + "grad_norm": 3.5813755989074707, + "learning_rate": 3.855965323016031e-06, + "loss": 0.5696, + "num_input_tokens_seen": 8567560, + "step": 32000 + }, + { + "epoch": 8.389252948885977, + "grad_norm": 3.2447965145111084, + "learning_rate": 3.8498650446521264e-06, + "loss": 0.429, + "num_input_tokens_seen": 8569544, + "step": 32005 + }, + { + "epoch": 8.390563564875492, + "grad_norm": 6.511360168457031, + "learning_rate": 3.843769192974925e-06, + "loss": 0.3898, + "num_input_tokens_seen": 8570872, + "step": 32010 + }, + { + "epoch": 8.391874180865006, + "grad_norm": 1.6601344347000122, + "learning_rate": 3.837677769260273e-06, + "loss": 0.5533, + "num_input_tokens_seen": 8572472, + "step": 32015 + }, + { + "epoch": 8.393184796854522, + "grad_norm": 2.6398699283599854, + "learning_rate": 3.8315907747831e-06, + "loss": 0.5998, + "num_input_tokens_seen": 8573512, + "step": 32020 + }, + { + "epoch": 8.394495412844037, + "grad_norm": 7.236937522888184, + "learning_rate": 3.825508210817402e-06, + "loss": 0.5103, + "num_input_tokens_seen": 8574776, + "step": 32025 + }, + { + "epoch": 8.395806028833551, + "grad_norm": 10.434103965759277, + "learning_rate": 3.8194300786362465e-06, + "loss": 0.5385, + "num_input_tokens_seen": 8576264, + "step": 32030 + }, + { + "epoch": 8.397116644823067, + "grad_norm": 5.054757595062256, + "learning_rate": 3.8133563795117816e-06, + "loss": 0.6463, + "num_input_tokens_seen": 8577608, + "step": 32035 + }, + { + "epoch": 8.398427260812582, + "grad_norm": 5.379391670227051, + "learning_rate": 3.807287114715216e-06, + "loss": 0.5772, + "num_input_tokens_seen": 8578616, + "step": 32040 + }, + { + "epoch": 8.399737876802098, + "grad_norm": 3.9136598110198975, + "learning_rate": 3.801222285516842e-06, + "loss": 0.3251, + "num_input_tokens_seen": 8579752, + "step": 32045 + }, + { + "epoch": 8.401048492791611, + "grad_norm": 1.9813429117202759, + "learning_rate": 3.7951618931860112e-06, + "loss": 0.376, + "num_input_tokens_seen": 8580904, + "step": 32050 + }, + { + "epoch": 8.402359108781127, + "grad_norm": 1.2691731452941895, + "learning_rate": 3.789105938991161e-06, + "loss": 0.499, + "num_input_tokens_seen": 8582504, + "step": 32055 + }, + { + "epoch": 8.403669724770642, + "grad_norm": 4.047717094421387, + "learning_rate": 3.7830544241997864e-06, + "loss": 0.3648, + "num_input_tokens_seen": 8583736, + "step": 32060 + }, + { + "epoch": 8.404980340760158, + "grad_norm": 8.619447708129883, + "learning_rate": 3.7770073500784596e-06, + "loss": 0.4895, + "num_input_tokens_seen": 8584744, + "step": 32065 + }, + { + "epoch": 8.406290956749672, + "grad_norm": 3.915311574935913, + "learning_rate": 3.7709647178928246e-06, + "loss": 0.3901, + "num_input_tokens_seen": 8585672, + "step": 32070 + }, + { + "epoch": 8.407601572739187, + "grad_norm": 9.890520095825195, + "learning_rate": 3.7649265289076003e-06, + "loss": 0.6066, + "num_input_tokens_seen": 8586648, + "step": 32075 + }, + { + "epoch": 8.408912188728703, + "grad_norm": 1.5971131324768066, + "learning_rate": 3.7588927843865556e-06, + "loss": 0.494, + "num_input_tokens_seen": 8588376, + "step": 32080 + }, + { + "epoch": 8.410222804718218, + "grad_norm": 8.246161460876465, + "learning_rate": 3.7528634855925475e-06, + "loss": 0.3748, + "num_input_tokens_seen": 8589512, + "step": 32085 + }, + { + "epoch": 8.411533420707732, + "grad_norm": 2.5616254806518555, + "learning_rate": 3.7468386337875028e-06, + "loss": 0.3379, + "num_input_tokens_seen": 8590504, + "step": 32090 + }, + { + "epoch": 8.412844036697248, + "grad_norm": 1.9691693782806396, + "learning_rate": 3.7408182302324014e-06, + "loss": 0.7108, + "num_input_tokens_seen": 8591880, + "step": 32095 + }, + { + "epoch": 8.414154652686763, + "grad_norm": 1.6043977737426758, + "learning_rate": 3.734802276187319e-06, + "loss": 0.4191, + "num_input_tokens_seen": 8593336, + "step": 32100 + }, + { + "epoch": 8.415465268676279, + "grad_norm": 8.206808090209961, + "learning_rate": 3.7287907729113754e-06, + "loss": 0.5241, + "num_input_tokens_seen": 8594584, + "step": 32105 + }, + { + "epoch": 8.416775884665793, + "grad_norm": 1.5711439847946167, + "learning_rate": 3.722783721662773e-06, + "loss": 0.5383, + "num_input_tokens_seen": 8595800, + "step": 32110 + }, + { + "epoch": 8.418086500655308, + "grad_norm": 3.1900620460510254, + "learning_rate": 3.7167811236987704e-06, + "loss": 0.4788, + "num_input_tokens_seen": 8597032, + "step": 32115 + }, + { + "epoch": 8.419397116644824, + "grad_norm": 2.747201919555664, + "learning_rate": 3.710782980275704e-06, + "loss": 0.4004, + "num_input_tokens_seen": 8598248, + "step": 32120 + }, + { + "epoch": 8.420707732634337, + "grad_norm": 1.4808269739151, + "learning_rate": 3.704789292648975e-06, + "loss": 0.5302, + "num_input_tokens_seen": 8599320, + "step": 32125 + }, + { + "epoch": 8.422018348623853, + "grad_norm": 5.4524126052856445, + "learning_rate": 3.6988000620730516e-06, + "loss": 0.5898, + "num_input_tokens_seen": 8600888, + "step": 32130 + }, + { + "epoch": 8.423328964613368, + "grad_norm": 1.6624374389648438, + "learning_rate": 3.6928152898014684e-06, + "loss": 0.5246, + "num_input_tokens_seen": 8602328, + "step": 32135 + }, + { + "epoch": 8.424639580602884, + "grad_norm": 7.438071250915527, + "learning_rate": 3.686834977086831e-06, + "loss": 0.3198, + "num_input_tokens_seen": 8603672, + "step": 32140 + }, + { + "epoch": 8.425950196592398, + "grad_norm": 27.837188720703125, + "learning_rate": 3.680859125180805e-06, + "loss": 0.3817, + "num_input_tokens_seen": 8604744, + "step": 32145 + }, + { + "epoch": 8.427260812581913, + "grad_norm": 1.7200762033462524, + "learning_rate": 3.674887735334126e-06, + "loss": 0.3206, + "num_input_tokens_seen": 8605944, + "step": 32150 + }, + { + "epoch": 8.428571428571429, + "grad_norm": 1.8121353387832642, + "learning_rate": 3.6689208087966e-06, + "loss": 0.4291, + "num_input_tokens_seen": 8607352, + "step": 32155 + }, + { + "epoch": 8.429882044560944, + "grad_norm": 1.614431619644165, + "learning_rate": 3.6629583468170914e-06, + "loss": 0.6009, + "num_input_tokens_seen": 8608840, + "step": 32160 + }, + { + "epoch": 8.431192660550458, + "grad_norm": 4.3146653175354, + "learning_rate": 3.65700035064353e-06, + "loss": 0.6478, + "num_input_tokens_seen": 8609720, + "step": 32165 + }, + { + "epoch": 8.432503276539974, + "grad_norm": 4.65165901184082, + "learning_rate": 3.6510468215229215e-06, + "loss": 0.452, + "num_input_tokens_seen": 8611368, + "step": 32170 + }, + { + "epoch": 8.43381389252949, + "grad_norm": 2.6685991287231445, + "learning_rate": 3.6450977607013217e-06, + "loss": 0.5476, + "num_input_tokens_seen": 8612584, + "step": 32175 + }, + { + "epoch": 8.435124508519005, + "grad_norm": 2.6238298416137695, + "learning_rate": 3.6391531694238647e-06, + "loss": 0.6149, + "num_input_tokens_seen": 8613864, + "step": 32180 + }, + { + "epoch": 8.436435124508519, + "grad_norm": 4.314866542816162, + "learning_rate": 3.6332130489347387e-06, + "loss": 0.3965, + "num_input_tokens_seen": 8615048, + "step": 32185 + }, + { + "epoch": 8.437745740498034, + "grad_norm": 4.135809421539307, + "learning_rate": 3.627277400477211e-06, + "loss": 0.4241, + "num_input_tokens_seen": 8616456, + "step": 32190 + }, + { + "epoch": 8.43905635648755, + "grad_norm": 2.0216329097747803, + "learning_rate": 3.6213462252935844e-06, + "loss": 0.427, + "num_input_tokens_seen": 8617720, + "step": 32195 + }, + { + "epoch": 8.440366972477065, + "grad_norm": 4.240630149841309, + "learning_rate": 3.6154195246252575e-06, + "loss": 0.5861, + "num_input_tokens_seen": 8618712, + "step": 32200 + }, + { + "epoch": 8.441677588466579, + "grad_norm": 3.5650432109832764, + "learning_rate": 3.609497299712672e-06, + "loss": 0.5328, + "num_input_tokens_seen": 8620136, + "step": 32205 + }, + { + "epoch": 8.442988204456094, + "grad_norm": 20.145597457885742, + "learning_rate": 3.6035795517953357e-06, + "loss": 0.6897, + "num_input_tokens_seen": 8621288, + "step": 32210 + }, + { + "epoch": 8.44429882044561, + "grad_norm": 2.854807138442993, + "learning_rate": 3.597666282111836e-06, + "loss": 0.4329, + "num_input_tokens_seen": 8622504, + "step": 32215 + }, + { + "epoch": 8.445609436435124, + "grad_norm": 3.471259355545044, + "learning_rate": 3.5917574918998024e-06, + "loss": 0.4014, + "num_input_tokens_seen": 8623496, + "step": 32220 + }, + { + "epoch": 8.44692005242464, + "grad_norm": 1.1333229541778564, + "learning_rate": 3.58585318239594e-06, + "loss": 0.3585, + "num_input_tokens_seen": 8624888, + "step": 32225 + }, + { + "epoch": 8.448230668414155, + "grad_norm": 2.0813004970550537, + "learning_rate": 3.579953354836002e-06, + "loss": 0.6668, + "num_input_tokens_seen": 8625896, + "step": 32230 + }, + { + "epoch": 8.44954128440367, + "grad_norm": 2.0227620601654053, + "learning_rate": 3.5740580104548145e-06, + "loss": 0.4565, + "num_input_tokens_seen": 8627192, + "step": 32235 + }, + { + "epoch": 8.450851900393184, + "grad_norm": 5.561039924621582, + "learning_rate": 3.568167150486265e-06, + "loss": 0.4664, + "num_input_tokens_seen": 8628232, + "step": 32240 + }, + { + "epoch": 8.4521625163827, + "grad_norm": 3.0528812408447266, + "learning_rate": 3.5622807761632985e-06, + "loss": 0.4093, + "num_input_tokens_seen": 8629336, + "step": 32245 + }, + { + "epoch": 8.453473132372215, + "grad_norm": 5.213983058929443, + "learning_rate": 3.5563988887179266e-06, + "loss": 0.6543, + "num_input_tokens_seen": 8630440, + "step": 32250 + }, + { + "epoch": 8.45478374836173, + "grad_norm": 5.019857883453369, + "learning_rate": 3.5505214893812115e-06, + "loss": 0.5999, + "num_input_tokens_seen": 8631960, + "step": 32255 + }, + { + "epoch": 8.456094364351245, + "grad_norm": 2.7999796867370605, + "learning_rate": 3.5446485793832918e-06, + "loss": 0.3581, + "num_input_tokens_seen": 8633288, + "step": 32260 + }, + { + "epoch": 8.45740498034076, + "grad_norm": 4.073810577392578, + "learning_rate": 3.5387801599533475e-06, + "loss": 0.3653, + "num_input_tokens_seen": 8634376, + "step": 32265 + }, + { + "epoch": 8.458715596330276, + "grad_norm": 6.48486852645874, + "learning_rate": 3.5329162323196356e-06, + "loss": 0.4226, + "num_input_tokens_seen": 8635304, + "step": 32270 + }, + { + "epoch": 8.460026212319791, + "grad_norm": 10.460210800170898, + "learning_rate": 3.527056797709466e-06, + "loss": 0.4772, + "num_input_tokens_seen": 8636488, + "step": 32275 + }, + { + "epoch": 8.461336828309305, + "grad_norm": 3.342729091644287, + "learning_rate": 3.5212018573492047e-06, + "loss": 0.3531, + "num_input_tokens_seen": 8637800, + "step": 32280 + }, + { + "epoch": 8.46264744429882, + "grad_norm": 5.530272483825684, + "learning_rate": 3.5153514124642822e-06, + "loss": 0.3745, + "num_input_tokens_seen": 8639464, + "step": 32285 + }, + { + "epoch": 8.463958060288336, + "grad_norm": 2.201104164123535, + "learning_rate": 3.5095054642791862e-06, + "loss": 0.4778, + "num_input_tokens_seen": 8640520, + "step": 32290 + }, + { + "epoch": 8.46526867627785, + "grad_norm": 3.974435329437256, + "learning_rate": 3.503664014017463e-06, + "loss": 0.5102, + "num_input_tokens_seen": 8641640, + "step": 32295 + }, + { + "epoch": 8.466579292267365, + "grad_norm": 3.11423921585083, + "learning_rate": 3.4978270629017207e-06, + "loss": 0.4044, + "num_input_tokens_seen": 8643160, + "step": 32300 + }, + { + "epoch": 8.46788990825688, + "grad_norm": 3.9198338985443115, + "learning_rate": 3.491994612153618e-06, + "loss": 0.597, + "num_input_tokens_seen": 8644312, + "step": 32305 + }, + { + "epoch": 8.469200524246396, + "grad_norm": 6.188119411468506, + "learning_rate": 3.4861666629938896e-06, + "loss": 0.4304, + "num_input_tokens_seen": 8645672, + "step": 32310 + }, + { + "epoch": 8.47051114023591, + "grad_norm": 3.29258394241333, + "learning_rate": 3.4803432166422983e-06, + "loss": 0.4463, + "num_input_tokens_seen": 8647080, + "step": 32315 + }, + { + "epoch": 8.471821756225426, + "grad_norm": 4.158838748931885, + "learning_rate": 3.474524274317689e-06, + "loss": 0.4324, + "num_input_tokens_seen": 8648120, + "step": 32320 + }, + { + "epoch": 8.473132372214941, + "grad_norm": 4.331899166107178, + "learning_rate": 3.468709837237949e-06, + "loss": 0.4913, + "num_input_tokens_seen": 8649336, + "step": 32325 + }, + { + "epoch": 8.474442988204457, + "grad_norm": 2.4167182445526123, + "learning_rate": 3.4628999066200427e-06, + "loss": 0.7259, + "num_input_tokens_seen": 8650376, + "step": 32330 + }, + { + "epoch": 8.47575360419397, + "grad_norm": 4.608827114105225, + "learning_rate": 3.4570944836799734e-06, + "loss": 0.4286, + "num_input_tokens_seen": 8651480, + "step": 32335 + }, + { + "epoch": 8.477064220183486, + "grad_norm": 4.429877758026123, + "learning_rate": 3.4512935696328045e-06, + "loss": 0.4259, + "num_input_tokens_seen": 8653096, + "step": 32340 + }, + { + "epoch": 8.478374836173002, + "grad_norm": 5.826183319091797, + "learning_rate": 3.4454971656926655e-06, + "loss": 0.564, + "num_input_tokens_seen": 8654264, + "step": 32345 + }, + { + "epoch": 8.479685452162517, + "grad_norm": 1.8714585304260254, + "learning_rate": 3.4397052730727214e-06, + "loss": 0.4353, + "num_input_tokens_seen": 8655544, + "step": 32350 + }, + { + "epoch": 8.48099606815203, + "grad_norm": 6.717642307281494, + "learning_rate": 3.4339178929852085e-06, + "loss": 0.5821, + "num_input_tokens_seen": 8656904, + "step": 32355 + }, + { + "epoch": 8.482306684141546, + "grad_norm": 2.7946255207061768, + "learning_rate": 3.4281350266414158e-06, + "loss": 0.4861, + "num_input_tokens_seen": 8658088, + "step": 32360 + }, + { + "epoch": 8.483617300131062, + "grad_norm": 3.3576102256774902, + "learning_rate": 3.422356675251692e-06, + "loss": 0.5721, + "num_input_tokens_seen": 8659256, + "step": 32365 + }, + { + "epoch": 8.484927916120578, + "grad_norm": 31.37765121459961, + "learning_rate": 3.4165828400254314e-06, + "loss": 0.5492, + "num_input_tokens_seen": 8660536, + "step": 32370 + }, + { + "epoch": 8.486238532110091, + "grad_norm": 2.955782175064087, + "learning_rate": 3.4108135221710885e-06, + "loss": 0.8138, + "num_input_tokens_seen": 8661816, + "step": 32375 + }, + { + "epoch": 8.487549148099607, + "grad_norm": 8.668697357177734, + "learning_rate": 3.405048722896173e-06, + "loss": 0.5985, + "num_input_tokens_seen": 8662856, + "step": 32380 + }, + { + "epoch": 8.488859764089122, + "grad_norm": 3.8321568965911865, + "learning_rate": 3.3992884434072474e-06, + "loss": 0.5503, + "num_input_tokens_seen": 8664056, + "step": 32385 + }, + { + "epoch": 8.490170380078636, + "grad_norm": 1.9365930557250977, + "learning_rate": 3.3935326849099253e-06, + "loss": 0.4184, + "num_input_tokens_seen": 8665304, + "step": 32390 + }, + { + "epoch": 8.491480996068152, + "grad_norm": 7.315883159637451, + "learning_rate": 3.3877814486088816e-06, + "loss": 0.4656, + "num_input_tokens_seen": 8666296, + "step": 32395 + }, + { + "epoch": 8.492791612057667, + "grad_norm": 4.292545318603516, + "learning_rate": 3.382034735707834e-06, + "loss": 0.5637, + "num_input_tokens_seen": 8667608, + "step": 32400 + }, + { + "epoch": 8.494102228047183, + "grad_norm": 4.730447292327881, + "learning_rate": 3.376292547409568e-06, + "loss": 0.4749, + "num_input_tokens_seen": 8668632, + "step": 32405 + }, + { + "epoch": 8.495412844036696, + "grad_norm": 10.76811408996582, + "learning_rate": 3.3705548849159054e-06, + "loss": 0.3281, + "num_input_tokens_seen": 8669784, + "step": 32410 + }, + { + "epoch": 8.496723460026212, + "grad_norm": 4.195574760437012, + "learning_rate": 3.364821749427732e-06, + "loss": 0.4657, + "num_input_tokens_seen": 8670840, + "step": 32415 + }, + { + "epoch": 8.498034076015728, + "grad_norm": 1.3009300231933594, + "learning_rate": 3.3590931421449867e-06, + "loss": 0.4662, + "num_input_tokens_seen": 8672312, + "step": 32420 + }, + { + "epoch": 8.499344692005243, + "grad_norm": 2.559251546859741, + "learning_rate": 3.3533690642666595e-06, + "loss": 0.5078, + "num_input_tokens_seen": 8673816, + "step": 32425 + }, + { + "epoch": 8.500655307994757, + "grad_norm": 4.092755317687988, + "learning_rate": 3.347649516990778e-06, + "loss": 0.5677, + "num_input_tokens_seen": 8674936, + "step": 32430 + }, + { + "epoch": 8.501965923984272, + "grad_norm": 0.8494312763214111, + "learning_rate": 3.3419345015144406e-06, + "loss": 0.4253, + "num_input_tokens_seen": 8676536, + "step": 32435 + }, + { + "epoch": 8.502228047182175, + "eval_loss": 0.46438896656036377, + "eval_runtime": 18.218, + "eval_samples_per_second": 46.547, + "eval_steps_per_second": 23.274, + "num_input_tokens_seen": 8676728, + "step": 32436 + }, + { + "epoch": 8.503276539973788, + "grad_norm": 2.293449878692627, + "learning_rate": 3.3362240190337817e-06, + "loss": 0.3373, + "num_input_tokens_seen": 8677672, + "step": 32440 + }, + { + "epoch": 8.504587155963304, + "grad_norm": 2.9440088272094727, + "learning_rate": 3.3305180707440124e-06, + "loss": 0.427, + "num_input_tokens_seen": 8679048, + "step": 32445 + }, + { + "epoch": 8.505897771952817, + "grad_norm": 2.830986261367798, + "learning_rate": 3.3248166578393665e-06, + "loss": 0.5735, + "num_input_tokens_seen": 8680232, + "step": 32450 + }, + { + "epoch": 8.507208387942333, + "grad_norm": 2.151811361312866, + "learning_rate": 3.319119781513144e-06, + "loss": 0.3788, + "num_input_tokens_seen": 8681624, + "step": 32455 + }, + { + "epoch": 8.508519003931848, + "grad_norm": 3.633958101272583, + "learning_rate": 3.313427442957695e-06, + "loss": 0.6219, + "num_input_tokens_seen": 8682952, + "step": 32460 + }, + { + "epoch": 8.509829619921362, + "grad_norm": 3.6655266284942627, + "learning_rate": 3.3077396433644046e-06, + "loss": 0.3212, + "num_input_tokens_seen": 8684296, + "step": 32465 + }, + { + "epoch": 8.511140235910878, + "grad_norm": 0.9536294341087341, + "learning_rate": 3.3020563839237267e-06, + "loss": 0.4395, + "num_input_tokens_seen": 8686312, + "step": 32470 + }, + { + "epoch": 8.512450851900393, + "grad_norm": 2.92100191116333, + "learning_rate": 3.2963776658251537e-06, + "loss": 0.4182, + "num_input_tokens_seen": 8687960, + "step": 32475 + }, + { + "epoch": 8.513761467889909, + "grad_norm": 0.29565152525901794, + "learning_rate": 3.2907034902572382e-06, + "loss": 0.5988, + "num_input_tokens_seen": 8689176, + "step": 32480 + }, + { + "epoch": 8.515072083879424, + "grad_norm": 2.2594661712646484, + "learning_rate": 3.2850338584075686e-06, + "loss": 0.5206, + "num_input_tokens_seen": 8690616, + "step": 32485 + }, + { + "epoch": 8.516382699868938, + "grad_norm": 1.8286584615707397, + "learning_rate": 3.279368771462793e-06, + "loss": 0.3121, + "num_input_tokens_seen": 8691896, + "step": 32490 + }, + { + "epoch": 8.517693315858454, + "grad_norm": 3.437333822250366, + "learning_rate": 3.273708230608602e-06, + "loss": 0.3954, + "num_input_tokens_seen": 8693432, + "step": 32495 + }, + { + "epoch": 8.51900393184797, + "grad_norm": 3.516164541244507, + "learning_rate": 3.2680522370297397e-06, + "loss": 0.5962, + "num_input_tokens_seen": 8694520, + "step": 32500 + }, + { + "epoch": 8.520314547837483, + "grad_norm": 5.8210978507995605, + "learning_rate": 3.262400791909992e-06, + "loss": 0.5781, + "num_input_tokens_seen": 8695768, + "step": 32505 + }, + { + "epoch": 8.521625163826998, + "grad_norm": 9.838863372802734, + "learning_rate": 3.256753896432202e-06, + "loss": 0.4166, + "num_input_tokens_seen": 8696632, + "step": 32510 + }, + { + "epoch": 8.522935779816514, + "grad_norm": 2.314159393310547, + "learning_rate": 3.2511115517782494e-06, + "loss": 0.4496, + "num_input_tokens_seen": 8698200, + "step": 32515 + }, + { + "epoch": 8.52424639580603, + "grad_norm": 2.2862162590026855, + "learning_rate": 3.2454737591290695e-06, + "loss": 0.434, + "num_input_tokens_seen": 8699336, + "step": 32520 + }, + { + "epoch": 8.525557011795543, + "grad_norm": 3.70784592628479, + "learning_rate": 3.2398405196646443e-06, + "loss": 0.4606, + "num_input_tokens_seen": 8700616, + "step": 32525 + }, + { + "epoch": 8.526867627785059, + "grad_norm": 9.277105331420898, + "learning_rate": 3.2342118345639993e-06, + "loss": 0.521, + "num_input_tokens_seen": 8701512, + "step": 32530 + }, + { + "epoch": 8.528178243774574, + "grad_norm": 1.638107419013977, + "learning_rate": 3.2285877050052093e-06, + "loss": 0.4043, + "num_input_tokens_seen": 8702840, + "step": 32535 + }, + { + "epoch": 8.52948885976409, + "grad_norm": 0.4943486750125885, + "learning_rate": 3.2229681321653966e-06, + "loss": 0.3809, + "num_input_tokens_seen": 8705352, + "step": 32540 + }, + { + "epoch": 8.530799475753604, + "grad_norm": 1.4706332683563232, + "learning_rate": 3.2173531172207283e-06, + "loss": 0.5957, + "num_input_tokens_seen": 8706824, + "step": 32545 + }, + { + "epoch": 8.53211009174312, + "grad_norm": 2.893143892288208, + "learning_rate": 3.211742661346409e-06, + "loss": 0.4996, + "num_input_tokens_seen": 8708392, + "step": 32550 + }, + { + "epoch": 8.533420707732635, + "grad_norm": 3.333706855773926, + "learning_rate": 3.2061367657167025e-06, + "loss": 0.4921, + "num_input_tokens_seen": 8709576, + "step": 32555 + }, + { + "epoch": 8.534731323722148, + "grad_norm": 9.364167213439941, + "learning_rate": 3.200535431504914e-06, + "loss": 0.3944, + "num_input_tokens_seen": 8710472, + "step": 32560 + }, + { + "epoch": 8.536041939711664, + "grad_norm": 7.044729232788086, + "learning_rate": 3.194938659883398e-06, + "loss": 0.613, + "num_input_tokens_seen": 8711560, + "step": 32565 + }, + { + "epoch": 8.53735255570118, + "grad_norm": 1.8173372745513916, + "learning_rate": 3.189346452023542e-06, + "loss": 0.4403, + "num_input_tokens_seen": 8712632, + "step": 32570 + }, + { + "epoch": 8.538663171690695, + "grad_norm": 2.7125847339630127, + "learning_rate": 3.1837588090957853e-06, + "loss": 0.3396, + "num_input_tokens_seen": 8714056, + "step": 32575 + }, + { + "epoch": 8.539973787680209, + "grad_norm": 2.8744070529937744, + "learning_rate": 3.1781757322696224e-06, + "loss": 0.6608, + "num_input_tokens_seen": 8715320, + "step": 32580 + }, + { + "epoch": 8.541284403669724, + "grad_norm": 11.620566368103027, + "learning_rate": 3.172597222713569e-06, + "loss": 0.3625, + "num_input_tokens_seen": 8716648, + "step": 32585 + }, + { + "epoch": 8.54259501965924, + "grad_norm": 2.152395725250244, + "learning_rate": 3.1670232815951994e-06, + "loss": 0.3595, + "num_input_tokens_seen": 8718152, + "step": 32590 + }, + { + "epoch": 8.543905635648755, + "grad_norm": 4.081813812255859, + "learning_rate": 3.1614539100811356e-06, + "loss": 0.8338, + "num_input_tokens_seen": 8719432, + "step": 32595 + }, + { + "epoch": 8.54521625163827, + "grad_norm": 1.7068098783493042, + "learning_rate": 3.1558891093370318e-06, + "loss": 0.5033, + "num_input_tokens_seen": 8721240, + "step": 32600 + }, + { + "epoch": 8.546526867627785, + "grad_norm": 2.7729084491729736, + "learning_rate": 3.150328880527595e-06, + "loss": 0.521, + "num_input_tokens_seen": 8722712, + "step": 32605 + }, + { + "epoch": 8.5478374836173, + "grad_norm": 2.744551420211792, + "learning_rate": 3.1447732248165633e-06, + "loss": 0.3556, + "num_input_tokens_seen": 8724136, + "step": 32610 + }, + { + "epoch": 8.549148099606816, + "grad_norm": 3.5610547065734863, + "learning_rate": 3.139222143366746e-06, + "loss": 0.2523, + "num_input_tokens_seen": 8725208, + "step": 32615 + }, + { + "epoch": 8.55045871559633, + "grad_norm": 8.909448623657227, + "learning_rate": 3.1336756373399556e-06, + "loss": 0.4059, + "num_input_tokens_seen": 8726520, + "step": 32620 + }, + { + "epoch": 8.551769331585845, + "grad_norm": 3.1390879154205322, + "learning_rate": 3.128133707897071e-06, + "loss": 0.4408, + "num_input_tokens_seen": 8727576, + "step": 32625 + }, + { + "epoch": 8.55307994757536, + "grad_norm": 1.6571964025497437, + "learning_rate": 3.12259635619801e-06, + "loss": 0.4575, + "num_input_tokens_seen": 8728952, + "step": 32630 + }, + { + "epoch": 8.554390563564876, + "grad_norm": 2.9722142219543457, + "learning_rate": 3.1170635834017303e-06, + "loss": 0.414, + "num_input_tokens_seen": 8730120, + "step": 32635 + }, + { + "epoch": 8.55570117955439, + "grad_norm": 2.9672513008117676, + "learning_rate": 3.1115353906662313e-06, + "loss": 0.4861, + "num_input_tokens_seen": 8731288, + "step": 32640 + }, + { + "epoch": 8.557011795543906, + "grad_norm": 15.047938346862793, + "learning_rate": 3.1060117791485544e-06, + "loss": 0.2616, + "num_input_tokens_seen": 8732104, + "step": 32645 + }, + { + "epoch": 8.558322411533421, + "grad_norm": 1.7605503797531128, + "learning_rate": 3.1004927500047827e-06, + "loss": 0.5204, + "num_input_tokens_seen": 8733656, + "step": 32650 + }, + { + "epoch": 8.559633027522935, + "grad_norm": 2.7712955474853516, + "learning_rate": 3.0949783043900393e-06, + "loss": 0.4932, + "num_input_tokens_seen": 8735096, + "step": 32655 + }, + { + "epoch": 8.56094364351245, + "grad_norm": 0.9411232471466064, + "learning_rate": 3.0894684434584958e-06, + "loss": 0.3868, + "num_input_tokens_seen": 8736728, + "step": 32660 + }, + { + "epoch": 8.562254259501966, + "grad_norm": 2.638190746307373, + "learning_rate": 3.083963168363341e-06, + "loss": 0.3636, + "num_input_tokens_seen": 8738120, + "step": 32665 + }, + { + "epoch": 8.563564875491481, + "grad_norm": 8.791812896728516, + "learning_rate": 3.0784624802568192e-06, + "loss": 0.5498, + "num_input_tokens_seen": 8739480, + "step": 32670 + }, + { + "epoch": 8.564875491480995, + "grad_norm": 2.434105634689331, + "learning_rate": 3.072966380290232e-06, + "loss": 0.459, + "num_input_tokens_seen": 8740584, + "step": 32675 + }, + { + "epoch": 8.56618610747051, + "grad_norm": 1.7149865627288818, + "learning_rate": 3.0674748696138927e-06, + "loss": 0.4243, + "num_input_tokens_seen": 8742664, + "step": 32680 + }, + { + "epoch": 8.567496723460026, + "grad_norm": 3.917341947555542, + "learning_rate": 3.0619879493771654e-06, + "loss": 0.3918, + "num_input_tokens_seen": 8744616, + "step": 32685 + }, + { + "epoch": 8.568807339449542, + "grad_norm": 2.859779119491577, + "learning_rate": 3.056505620728456e-06, + "loss": 0.5779, + "num_input_tokens_seen": 8746152, + "step": 32690 + }, + { + "epoch": 8.570117955439056, + "grad_norm": 4.306607246398926, + "learning_rate": 3.051027884815211e-06, + "loss": 0.3236, + "num_input_tokens_seen": 8747256, + "step": 32695 + }, + { + "epoch": 8.571428571428571, + "grad_norm": 3.641507148742676, + "learning_rate": 3.0455547427838986e-06, + "loss": 0.4347, + "num_input_tokens_seen": 8748392, + "step": 32700 + }, + { + "epoch": 8.572739187418087, + "grad_norm": 3.269625425338745, + "learning_rate": 3.0400861957800423e-06, + "loss": 0.515, + "num_input_tokens_seen": 8749480, + "step": 32705 + }, + { + "epoch": 8.574049803407602, + "grad_norm": 1.811838984489441, + "learning_rate": 3.034622244948204e-06, + "loss": 0.4546, + "num_input_tokens_seen": 8751224, + "step": 32710 + }, + { + "epoch": 8.575360419397116, + "grad_norm": 2.457167148590088, + "learning_rate": 3.029162891431975e-06, + "loss": 0.4474, + "num_input_tokens_seen": 8752472, + "step": 32715 + }, + { + "epoch": 8.576671035386632, + "grad_norm": 106.37034606933594, + "learning_rate": 3.02370813637399e-06, + "loss": 0.9954, + "num_input_tokens_seen": 8753704, + "step": 32720 + }, + { + "epoch": 8.577981651376147, + "grad_norm": 2.2641046047210693, + "learning_rate": 3.018257980915917e-06, + "loss": 0.345, + "num_input_tokens_seen": 8755672, + "step": 32725 + }, + { + "epoch": 8.579292267365663, + "grad_norm": 1.970775842666626, + "learning_rate": 3.0128124261984693e-06, + "loss": 0.3597, + "num_input_tokens_seen": 8757480, + "step": 32730 + }, + { + "epoch": 8.580602883355176, + "grad_norm": 5.622256755828857, + "learning_rate": 3.007371473361398e-06, + "loss": 0.2593, + "num_input_tokens_seen": 8758440, + "step": 32735 + }, + { + "epoch": 8.581913499344692, + "grad_norm": 3.55991268157959, + "learning_rate": 3.00193512354347e-06, + "loss": 0.4177, + "num_input_tokens_seen": 8759608, + "step": 32740 + }, + { + "epoch": 8.583224115334207, + "grad_norm": 1.955259919166565, + "learning_rate": 2.996503377882512e-06, + "loss": 0.4329, + "num_input_tokens_seen": 8761208, + "step": 32745 + }, + { + "epoch": 8.584534731323721, + "grad_norm": 5.524657249450684, + "learning_rate": 2.9910762375153752e-06, + "loss": 0.4009, + "num_input_tokens_seen": 8762376, + "step": 32750 + }, + { + "epoch": 8.585845347313237, + "grad_norm": 7.414572715759277, + "learning_rate": 2.9856537035779564e-06, + "loss": 0.6027, + "num_input_tokens_seen": 8763400, + "step": 32755 + }, + { + "epoch": 8.587155963302752, + "grad_norm": 8.909586906433105, + "learning_rate": 2.980235777205179e-06, + "loss": 0.4813, + "num_input_tokens_seen": 8765224, + "step": 32760 + }, + { + "epoch": 8.588466579292268, + "grad_norm": 1.9553418159484863, + "learning_rate": 2.9748224595310042e-06, + "loss": 0.3798, + "num_input_tokens_seen": 8766792, + "step": 32765 + }, + { + "epoch": 8.589777195281782, + "grad_norm": 1.678082823753357, + "learning_rate": 2.9694137516884353e-06, + "loss": 0.4565, + "num_input_tokens_seen": 8768248, + "step": 32770 + }, + { + "epoch": 8.591087811271297, + "grad_norm": 1.4927386045455933, + "learning_rate": 2.9640096548094992e-06, + "loss": 0.5348, + "num_input_tokens_seen": 8769816, + "step": 32775 + }, + { + "epoch": 8.592398427260813, + "grad_norm": 3.2452995777130127, + "learning_rate": 2.958610170025275e-06, + "loss": 0.4581, + "num_input_tokens_seen": 8771192, + "step": 32780 + }, + { + "epoch": 8.593709043250328, + "grad_norm": 1.7157657146453857, + "learning_rate": 2.9532152984658446e-06, + "loss": 0.3252, + "num_input_tokens_seen": 8772984, + "step": 32785 + }, + { + "epoch": 8.595019659239842, + "grad_norm": 1.598798155784607, + "learning_rate": 2.9478250412603657e-06, + "loss": 0.3292, + "num_input_tokens_seen": 8774600, + "step": 32790 + }, + { + "epoch": 8.596330275229358, + "grad_norm": 3.2343223094940186, + "learning_rate": 2.942439399537e-06, + "loss": 0.776, + "num_input_tokens_seen": 8775816, + "step": 32795 + }, + { + "epoch": 8.597640891218873, + "grad_norm": 4.4449005126953125, + "learning_rate": 2.9370583744229572e-06, + "loss": 0.5657, + "num_input_tokens_seen": 8777096, + "step": 32800 + }, + { + "epoch": 8.598951507208389, + "grad_norm": 1.582850456237793, + "learning_rate": 2.931681967044475e-06, + "loss": 0.4004, + "num_input_tokens_seen": 8779304, + "step": 32805 + }, + { + "epoch": 8.600262123197902, + "grad_norm": 1.6600313186645508, + "learning_rate": 2.9263101785268254e-06, + "loss": 0.4894, + "num_input_tokens_seen": 8780312, + "step": 32810 + }, + { + "epoch": 8.601572739187418, + "grad_norm": 4.6506781578063965, + "learning_rate": 2.9209430099943204e-06, + "loss": 0.4394, + "num_input_tokens_seen": 8781144, + "step": 32815 + }, + { + "epoch": 8.602883355176933, + "grad_norm": 2.4605133533477783, + "learning_rate": 2.915580462570289e-06, + "loss": 0.4962, + "num_input_tokens_seen": 8782616, + "step": 32820 + }, + { + "epoch": 8.604193971166449, + "grad_norm": 3.5529799461364746, + "learning_rate": 2.910222537377108e-06, + "loss": 0.5816, + "num_input_tokens_seen": 8784440, + "step": 32825 + }, + { + "epoch": 8.605504587155963, + "grad_norm": 10.374228477478027, + "learning_rate": 2.904869235536181e-06, + "loss": 0.2821, + "num_input_tokens_seen": 8785608, + "step": 32830 + }, + { + "epoch": 8.606815203145478, + "grad_norm": 4.0001702308654785, + "learning_rate": 2.8995205581679452e-06, + "loss": 0.4196, + "num_input_tokens_seen": 8786984, + "step": 32835 + }, + { + "epoch": 8.608125819134994, + "grad_norm": 3.994563341140747, + "learning_rate": 2.8941765063918657e-06, + "loss": 0.5625, + "num_input_tokens_seen": 8788312, + "step": 32840 + }, + { + "epoch": 8.609436435124508, + "grad_norm": 3.1236324310302734, + "learning_rate": 2.888837081326451e-06, + "loss": 0.3732, + "num_input_tokens_seen": 8789736, + "step": 32845 + }, + { + "epoch": 8.610747051114023, + "grad_norm": 1.8138487339019775, + "learning_rate": 2.8835022840892373e-06, + "loss": 0.6167, + "num_input_tokens_seen": 8791400, + "step": 32850 + }, + { + "epoch": 8.612057667103539, + "grad_norm": 1.9874194860458374, + "learning_rate": 2.8781721157967734e-06, + "loss": 0.5712, + "num_input_tokens_seen": 8792680, + "step": 32855 + }, + { + "epoch": 8.613368283093054, + "grad_norm": 7.680117130279541, + "learning_rate": 2.872846577564664e-06, + "loss": 0.3982, + "num_input_tokens_seen": 8793784, + "step": 32860 + }, + { + "epoch": 8.614678899082568, + "grad_norm": 2.267057418823242, + "learning_rate": 2.8675256705075343e-06, + "loss": 0.5545, + "num_input_tokens_seen": 8795160, + "step": 32865 + }, + { + "epoch": 8.615989515072084, + "grad_norm": 1.9560284614562988, + "learning_rate": 2.8622093957390377e-06, + "loss": 0.4327, + "num_input_tokens_seen": 8796360, + "step": 32870 + }, + { + "epoch": 8.617300131061599, + "grad_norm": 11.333175659179688, + "learning_rate": 2.856897754371865e-06, + "loss": 0.4154, + "num_input_tokens_seen": 8797640, + "step": 32875 + }, + { + "epoch": 8.618610747051115, + "grad_norm": 2.5761027336120605, + "learning_rate": 2.8515907475177293e-06, + "loss": 0.3341, + "num_input_tokens_seen": 8799032, + "step": 32880 + }, + { + "epoch": 8.619921363040628, + "grad_norm": 1.3244388103485107, + "learning_rate": 2.8462883762873842e-06, + "loss": 0.5719, + "num_input_tokens_seen": 8800424, + "step": 32885 + }, + { + "epoch": 8.621231979030144, + "grad_norm": 1.7521897554397583, + "learning_rate": 2.8409906417906e-06, + "loss": 0.2829, + "num_input_tokens_seen": 8801704, + "step": 32890 + }, + { + "epoch": 8.62254259501966, + "grad_norm": 4.282098293304443, + "learning_rate": 2.8356975451361873e-06, + "loss": 0.4553, + "num_input_tokens_seen": 8802968, + "step": 32895 + }, + { + "epoch": 8.623853211009175, + "grad_norm": 4.178332328796387, + "learning_rate": 2.8304090874319843e-06, + "loss": 0.4305, + "num_input_tokens_seen": 8804344, + "step": 32900 + }, + { + "epoch": 8.625163826998689, + "grad_norm": 9.019598007202148, + "learning_rate": 2.8251252697848505e-06, + "loss": 0.3914, + "num_input_tokens_seen": 8805816, + "step": 32905 + }, + { + "epoch": 8.626474442988204, + "grad_norm": 79.05650329589844, + "learning_rate": 2.819846093300682e-06, + "loss": 0.4362, + "num_input_tokens_seen": 8806792, + "step": 32910 + }, + { + "epoch": 8.62778505897772, + "grad_norm": 2.140080451965332, + "learning_rate": 2.814571559084403e-06, + "loss": 0.313, + "num_input_tokens_seen": 8808232, + "step": 32915 + }, + { + "epoch": 8.629095674967235, + "grad_norm": 2.2190053462982178, + "learning_rate": 2.8093016682399636e-06, + "loss": 0.4481, + "num_input_tokens_seen": 8809624, + "step": 32920 + }, + { + "epoch": 8.63040629095675, + "grad_norm": 3.6241111755371094, + "learning_rate": 2.804036421870343e-06, + "loss": 0.5302, + "num_input_tokens_seen": 8810776, + "step": 32925 + }, + { + "epoch": 8.631716906946265, + "grad_norm": 1.759587287902832, + "learning_rate": 2.798775821077551e-06, + "loss": 0.3403, + "num_input_tokens_seen": 8811928, + "step": 32930 + }, + { + "epoch": 8.63302752293578, + "grad_norm": 1.263648271560669, + "learning_rate": 2.7935198669626155e-06, + "loss": 0.3426, + "num_input_tokens_seen": 8813320, + "step": 32935 + }, + { + "epoch": 8.634338138925294, + "grad_norm": 2.1444625854492188, + "learning_rate": 2.788268560625601e-06, + "loss": 0.4898, + "num_input_tokens_seen": 8815528, + "step": 32940 + }, + { + "epoch": 8.63564875491481, + "grad_norm": 3.145885705947876, + "learning_rate": 2.7830219031655973e-06, + "loss": 0.4358, + "num_input_tokens_seen": 8816872, + "step": 32945 + }, + { + "epoch": 8.636959370904325, + "grad_norm": 4.559032440185547, + "learning_rate": 2.777779895680721e-06, + "loss": 0.452, + "num_input_tokens_seen": 8818184, + "step": 32950 + }, + { + "epoch": 8.63826998689384, + "grad_norm": 3.093364953994751, + "learning_rate": 2.772542539268114e-06, + "loss": 0.3982, + "num_input_tokens_seen": 8819112, + "step": 32955 + }, + { + "epoch": 8.639580602883354, + "grad_norm": 2.56544828414917, + "learning_rate": 2.7673098350239485e-06, + "loss": 0.8146, + "num_input_tokens_seen": 8820376, + "step": 32960 + }, + { + "epoch": 8.64089121887287, + "grad_norm": 3.4281864166259766, + "learning_rate": 2.7620817840434216e-06, + "loss": 0.5036, + "num_input_tokens_seen": 8821400, + "step": 32965 + }, + { + "epoch": 8.642201834862385, + "grad_norm": 1.9862706661224365, + "learning_rate": 2.756858387420758e-06, + "loss": 0.4125, + "num_input_tokens_seen": 8822952, + "step": 32970 + }, + { + "epoch": 8.643512450851901, + "grad_norm": 2.805009603500366, + "learning_rate": 2.7516396462491934e-06, + "loss": 0.3426, + "num_input_tokens_seen": 8824424, + "step": 32975 + }, + { + "epoch": 8.644823066841415, + "grad_norm": 3.7867629528045654, + "learning_rate": 2.7464255616210106e-06, + "loss": 0.3601, + "num_input_tokens_seen": 8825960, + "step": 32980 + }, + { + "epoch": 8.64613368283093, + "grad_norm": 2.2258658409118652, + "learning_rate": 2.7412161346275057e-06, + "loss": 0.6314, + "num_input_tokens_seen": 8827720, + "step": 32985 + }, + { + "epoch": 8.647444298820446, + "grad_norm": 4.45399284362793, + "learning_rate": 2.7360113663590036e-06, + "loss": 0.5816, + "num_input_tokens_seen": 8829016, + "step": 32990 + }, + { + "epoch": 8.648754914809961, + "grad_norm": 5.464005470275879, + "learning_rate": 2.7308112579048555e-06, + "loss": 0.5104, + "num_input_tokens_seen": 8830008, + "step": 32995 + }, + { + "epoch": 8.650065530799475, + "grad_norm": 4.789607524871826, + "learning_rate": 2.7256158103534298e-06, + "loss": 0.532, + "num_input_tokens_seen": 8831368, + "step": 33000 + }, + { + "epoch": 8.65137614678899, + "grad_norm": 8.318199157714844, + "learning_rate": 2.7204250247921308e-06, + "loss": 0.361, + "num_input_tokens_seen": 8832408, + "step": 33005 + }, + { + "epoch": 8.652686762778506, + "grad_norm": 2.3564138412475586, + "learning_rate": 2.7152389023073766e-06, + "loss": 0.3398, + "num_input_tokens_seen": 8833464, + "step": 33010 + }, + { + "epoch": 8.653997378768022, + "grad_norm": 4.640978813171387, + "learning_rate": 2.710057443984615e-06, + "loss": 0.4288, + "num_input_tokens_seen": 8834840, + "step": 33015 + }, + { + "epoch": 8.655307994757536, + "grad_norm": 3.162649631500244, + "learning_rate": 2.7048806509083146e-06, + "loss": 0.5069, + "num_input_tokens_seen": 8838024, + "step": 33020 + }, + { + "epoch": 8.656618610747051, + "grad_norm": 2.6991031169891357, + "learning_rate": 2.6997085241619728e-06, + "loss": 0.4128, + "num_input_tokens_seen": 8839320, + "step": 33025 + }, + { + "epoch": 8.657929226736567, + "grad_norm": 2.1275076866149902, + "learning_rate": 2.6945410648281044e-06, + "loss": 0.5858, + "num_input_tokens_seen": 8840344, + "step": 33030 + }, + { + "epoch": 8.65923984272608, + "grad_norm": 2.9502217769622803, + "learning_rate": 2.6893782739882523e-06, + "loss": 0.3544, + "num_input_tokens_seen": 8841320, + "step": 33035 + }, + { + "epoch": 8.660550458715596, + "grad_norm": 4.29616117477417, + "learning_rate": 2.6842201527229743e-06, + "loss": 0.3613, + "num_input_tokens_seen": 8842424, + "step": 33040 + }, + { + "epoch": 8.661861074705111, + "grad_norm": 2.735761880874634, + "learning_rate": 2.6790667021118626e-06, + "loss": 0.3548, + "num_input_tokens_seen": 8844168, + "step": 33045 + }, + { + "epoch": 8.663171690694627, + "grad_norm": 1.836193561553955, + "learning_rate": 2.6739179232335264e-06, + "loss": 0.4866, + "num_input_tokens_seen": 8845848, + "step": 33050 + }, + { + "epoch": 8.66448230668414, + "grad_norm": 3.3240838050842285, + "learning_rate": 2.6687738171655873e-06, + "loss": 0.7218, + "num_input_tokens_seen": 8846920, + "step": 33055 + }, + { + "epoch": 8.665792922673656, + "grad_norm": 2.850552558898926, + "learning_rate": 2.6636343849847033e-06, + "loss": 0.3199, + "num_input_tokens_seen": 8848184, + "step": 33060 + }, + { + "epoch": 8.667103538663172, + "grad_norm": 3.335305690765381, + "learning_rate": 2.6584996277665475e-06, + "loss": 0.4394, + "num_input_tokens_seen": 8849304, + "step": 33065 + }, + { + "epoch": 8.668414154652687, + "grad_norm": 1.4569040536880493, + "learning_rate": 2.653369546585813e-06, + "loss": 0.4108, + "num_input_tokens_seen": 8850760, + "step": 33070 + }, + { + "epoch": 8.669724770642201, + "grad_norm": 5.452258586883545, + "learning_rate": 2.6482441425162235e-06, + "loss": 0.458, + "num_input_tokens_seen": 8851912, + "step": 33075 + }, + { + "epoch": 8.671035386631717, + "grad_norm": 2.6017894744873047, + "learning_rate": 2.6431234166305135e-06, + "loss": 0.5273, + "num_input_tokens_seen": 8853128, + "step": 33080 + }, + { + "epoch": 8.672346002621232, + "grad_norm": 7.24777889251709, + "learning_rate": 2.6380073700004504e-06, + "loss": 0.4414, + "num_input_tokens_seen": 8854088, + "step": 33085 + }, + { + "epoch": 8.673656618610748, + "grad_norm": 1.104638695716858, + "learning_rate": 2.6328960036967996e-06, + "loss": 0.4856, + "num_input_tokens_seen": 8855640, + "step": 33090 + }, + { + "epoch": 8.674967234600262, + "grad_norm": 3.7577993869781494, + "learning_rate": 2.627789318789367e-06, + "loss": 0.5229, + "num_input_tokens_seen": 8856632, + "step": 33095 + }, + { + "epoch": 8.676277850589777, + "grad_norm": 6.48061990737915, + "learning_rate": 2.6226873163469752e-06, + "loss": 0.4006, + "num_input_tokens_seen": 8857736, + "step": 33100 + }, + { + "epoch": 8.677588466579293, + "grad_norm": 1.7304445505142212, + "learning_rate": 2.6175899974374614e-06, + "loss": 0.4459, + "num_input_tokens_seen": 8859128, + "step": 33105 + }, + { + "epoch": 8.678899082568808, + "grad_norm": 2.285322904586792, + "learning_rate": 2.6124973631276888e-06, + "loss": 0.5461, + "num_input_tokens_seen": 8860312, + "step": 33110 + }, + { + "epoch": 8.680209698558322, + "grad_norm": 6.062207221984863, + "learning_rate": 2.607409414483536e-06, + "loss": 0.5046, + "num_input_tokens_seen": 8862072, + "step": 33115 + }, + { + "epoch": 8.681520314547837, + "grad_norm": 1.648390293121338, + "learning_rate": 2.6023261525699003e-06, + "loss": 0.3727, + "num_input_tokens_seen": 8863288, + "step": 33120 + }, + { + "epoch": 8.682830930537353, + "grad_norm": 1.7778027057647705, + "learning_rate": 2.597247578450701e-06, + "loss": 0.4611, + "num_input_tokens_seen": 8864888, + "step": 33125 + }, + { + "epoch": 8.684141546526867, + "grad_norm": 2.8090765476226807, + "learning_rate": 2.592173693188876e-06, + "loss": 0.6012, + "num_input_tokens_seen": 8866376, + "step": 33130 + }, + { + "epoch": 8.685452162516382, + "grad_norm": 2.6667962074279785, + "learning_rate": 2.5871044978463764e-06, + "loss": 0.4775, + "num_input_tokens_seen": 8867784, + "step": 33135 + }, + { + "epoch": 8.686762778505898, + "grad_norm": 2.6963798999786377, + "learning_rate": 2.582039993484181e-06, + "loss": 0.5043, + "num_input_tokens_seen": 8869464, + "step": 33140 + }, + { + "epoch": 8.688073394495413, + "grad_norm": 4.08290958404541, + "learning_rate": 2.576980181162278e-06, + "loss": 0.4234, + "num_input_tokens_seen": 8870552, + "step": 33145 + }, + { + "epoch": 8.689384010484927, + "grad_norm": 8.107934951782227, + "learning_rate": 2.571925061939681e-06, + "loss": 0.4533, + "num_input_tokens_seen": 8871592, + "step": 33150 + }, + { + "epoch": 8.690694626474443, + "grad_norm": 3.420680046081543, + "learning_rate": 2.566874636874414e-06, + "loss": 0.3674, + "num_input_tokens_seen": 8872584, + "step": 33155 + }, + { + "epoch": 8.692005242463958, + "grad_norm": 5.14133882522583, + "learning_rate": 2.561828907023525e-06, + "loss": 0.6398, + "num_input_tokens_seen": 8873752, + "step": 33160 + }, + { + "epoch": 8.693315858453474, + "grad_norm": 1.691115379333496, + "learning_rate": 2.556787873443081e-06, + "loss": 0.4725, + "num_input_tokens_seen": 8875528, + "step": 33165 + }, + { + "epoch": 8.694626474442988, + "grad_norm": 3.679473400115967, + "learning_rate": 2.5517515371881516e-06, + "loss": 0.4062, + "num_input_tokens_seen": 8876504, + "step": 33170 + }, + { + "epoch": 8.695937090432503, + "grad_norm": 4.43670129776001, + "learning_rate": 2.5467198993128355e-06, + "loss": 0.2972, + "num_input_tokens_seen": 8877528, + "step": 33175 + }, + { + "epoch": 8.697247706422019, + "grad_norm": 1.8841911554336548, + "learning_rate": 2.541692960870251e-06, + "loss": 0.4079, + "num_input_tokens_seen": 8879128, + "step": 33180 + }, + { + "epoch": 8.698558322411534, + "grad_norm": 40.916805267333984, + "learning_rate": 2.536670722912518e-06, + "loss": 0.7319, + "num_input_tokens_seen": 8880616, + "step": 33185 + }, + { + "epoch": 8.699868938401048, + "grad_norm": 5.04597806930542, + "learning_rate": 2.5316531864907955e-06, + "loss": 0.4745, + "num_input_tokens_seen": 8881960, + "step": 33190 + }, + { + "epoch": 8.701179554390563, + "grad_norm": 1.9333078861236572, + "learning_rate": 2.526640352655238e-06, + "loss": 0.5412, + "num_input_tokens_seen": 8883208, + "step": 33195 + }, + { + "epoch": 8.702490170380079, + "grad_norm": 2.082979917526245, + "learning_rate": 2.5216322224550226e-06, + "loss": 0.4542, + "num_input_tokens_seen": 8885048, + "step": 33200 + }, + { + "epoch": 8.703800786369595, + "grad_norm": 2.7125144004821777, + "learning_rate": 2.51662879693835e-06, + "loss": 0.6205, + "num_input_tokens_seen": 8886232, + "step": 33205 + }, + { + "epoch": 8.705111402359108, + "grad_norm": 2.475069761276245, + "learning_rate": 2.511630077152416e-06, + "loss": 0.3464, + "num_input_tokens_seen": 8887400, + "step": 33210 + }, + { + "epoch": 8.706422018348624, + "grad_norm": 1.7703763246536255, + "learning_rate": 2.50663606414345e-06, + "loss": 0.4737, + "num_input_tokens_seen": 8889144, + "step": 33215 + }, + { + "epoch": 8.70773263433814, + "grad_norm": 3.289094924926758, + "learning_rate": 2.501646758956691e-06, + "loss": 0.4741, + "num_input_tokens_seen": 8890264, + "step": 33220 + }, + { + "epoch": 8.709043250327653, + "grad_norm": 3.1482160091400146, + "learning_rate": 2.49666216263639e-06, + "loss": 0.3833, + "num_input_tokens_seen": 8891256, + "step": 33225 + }, + { + "epoch": 8.710353866317169, + "grad_norm": 1.1169230937957764, + "learning_rate": 2.491682276225818e-06, + "loss": 0.445, + "num_input_tokens_seen": 8892904, + "step": 33230 + }, + { + "epoch": 8.711664482306684, + "grad_norm": 1.7618119716644287, + "learning_rate": 2.486707100767252e-06, + "loss": 0.3106, + "num_input_tokens_seen": 8894264, + "step": 33235 + }, + { + "epoch": 8.7129750982962, + "grad_norm": 4.543776512145996, + "learning_rate": 2.4817366373019902e-06, + "loss": 0.4379, + "num_input_tokens_seen": 8895624, + "step": 33240 + }, + { + "epoch": 8.714285714285714, + "grad_norm": 4.329852104187012, + "learning_rate": 2.4767708868703414e-06, + "loss": 0.497, + "num_input_tokens_seen": 8896616, + "step": 33245 + }, + { + "epoch": 8.715596330275229, + "grad_norm": 4.016679286956787, + "learning_rate": 2.4718098505116305e-06, + "loss": 0.3727, + "num_input_tokens_seen": 8897944, + "step": 33250 + }, + { + "epoch": 8.716906946264745, + "grad_norm": 2.8815107345581055, + "learning_rate": 2.4668535292641904e-06, + "loss": 0.5503, + "num_input_tokens_seen": 8899112, + "step": 33255 + }, + { + "epoch": 8.71821756225426, + "grad_norm": 2.4064245223999023, + "learning_rate": 2.4619019241653716e-06, + "loss": 0.3396, + "num_input_tokens_seen": 8900424, + "step": 33260 + }, + { + "epoch": 8.719528178243774, + "grad_norm": 3.5999813079833984, + "learning_rate": 2.4569550362515403e-06, + "loss": 0.3063, + "num_input_tokens_seen": 8901528, + "step": 33265 + }, + { + "epoch": 8.72083879423329, + "grad_norm": 3.3880631923675537, + "learning_rate": 2.452012866558065e-06, + "loss": 0.499, + "num_input_tokens_seen": 8903080, + "step": 33270 + }, + { + "epoch": 8.722149410222805, + "grad_norm": 2.251687526702881, + "learning_rate": 2.4470754161193406e-06, + "loss": 0.3775, + "num_input_tokens_seen": 8904648, + "step": 33275 + }, + { + "epoch": 8.72346002621232, + "grad_norm": 8.021501541137695, + "learning_rate": 2.4421426859687597e-06, + "loss": 0.6156, + "num_input_tokens_seen": 8905496, + "step": 33280 + }, + { + "epoch": 8.724770642201834, + "grad_norm": 1.3290027379989624, + "learning_rate": 2.437214677138744e-06, + "loss": 0.5518, + "num_input_tokens_seen": 8906968, + "step": 33285 + }, + { + "epoch": 8.72608125819135, + "grad_norm": 2.297287940979004, + "learning_rate": 2.432291390660707e-06, + "loss": 0.4519, + "num_input_tokens_seen": 8908344, + "step": 33290 + }, + { + "epoch": 8.727391874180865, + "grad_norm": 4.419931888580322, + "learning_rate": 2.427372827565086e-06, + "loss": 0.3985, + "num_input_tokens_seen": 8909784, + "step": 33295 + }, + { + "epoch": 8.728702490170381, + "grad_norm": 2.7365236282348633, + "learning_rate": 2.4224589888813263e-06, + "loss": 0.5757, + "num_input_tokens_seen": 8911256, + "step": 33300 + }, + { + "epoch": 8.730013106159895, + "grad_norm": 6.37371301651001, + "learning_rate": 2.4175498756378924e-06, + "loss": 0.5242, + "num_input_tokens_seen": 8913448, + "step": 33305 + }, + { + "epoch": 8.73132372214941, + "grad_norm": 5.806975841522217, + "learning_rate": 2.41264548886225e-06, + "loss": 0.4844, + "num_input_tokens_seen": 8914488, + "step": 33310 + }, + { + "epoch": 8.732634338138926, + "grad_norm": 2.1004199981689453, + "learning_rate": 2.407745829580882e-06, + "loss": 0.4659, + "num_input_tokens_seen": 8915816, + "step": 33315 + }, + { + "epoch": 8.73394495412844, + "grad_norm": 1.8991992473602295, + "learning_rate": 2.402850898819278e-06, + "loss": 0.518, + "num_input_tokens_seen": 8917400, + "step": 33320 + }, + { + "epoch": 8.735255570117955, + "grad_norm": 5.516261577606201, + "learning_rate": 2.397960697601931e-06, + "loss": 0.4477, + "num_input_tokens_seen": 8918856, + "step": 33325 + }, + { + "epoch": 8.73656618610747, + "grad_norm": 2.4840030670166016, + "learning_rate": 2.393075226952357e-06, + "loss": 0.5363, + "num_input_tokens_seen": 8920360, + "step": 33330 + }, + { + "epoch": 8.737876802096986, + "grad_norm": 1.437688946723938, + "learning_rate": 2.388194487893078e-06, + "loss": 0.6769, + "num_input_tokens_seen": 8921800, + "step": 33335 + }, + { + "epoch": 8.7391874180865, + "grad_norm": 1.5880341529846191, + "learning_rate": 2.3833184814456204e-06, + "loss": 0.4026, + "num_input_tokens_seen": 8923752, + "step": 33340 + }, + { + "epoch": 8.740498034076015, + "grad_norm": 3.8024842739105225, + "learning_rate": 2.378447208630527e-06, + "loss": 0.3894, + "num_input_tokens_seen": 8925064, + "step": 33345 + }, + { + "epoch": 8.741808650065531, + "grad_norm": 6.604581356048584, + "learning_rate": 2.373580670467343e-06, + "loss": 0.5615, + "num_input_tokens_seen": 8926216, + "step": 33350 + }, + { + "epoch": 8.743119266055047, + "grad_norm": 3.437502384185791, + "learning_rate": 2.3687188679746315e-06, + "loss": 0.6661, + "num_input_tokens_seen": 8927576, + "step": 33355 + }, + { + "epoch": 8.74442988204456, + "grad_norm": 1.8482294082641602, + "learning_rate": 2.363861802169959e-06, + "loss": 0.6698, + "num_input_tokens_seen": 8928888, + "step": 33360 + }, + { + "epoch": 8.745740498034076, + "grad_norm": 7.063354015350342, + "learning_rate": 2.3590094740698955e-06, + "loss": 0.472, + "num_input_tokens_seen": 8930376, + "step": 33365 + }, + { + "epoch": 8.747051114023591, + "grad_norm": 2.600224494934082, + "learning_rate": 2.3541618846900316e-06, + "loss": 0.2961, + "num_input_tokens_seen": 8931704, + "step": 33370 + }, + { + "epoch": 8.748361730013107, + "grad_norm": 4.842316150665283, + "learning_rate": 2.3493190350449557e-06, + "loss": 0.4972, + "num_input_tokens_seen": 8933080, + "step": 33375 + }, + { + "epoch": 8.74967234600262, + "grad_norm": 4.8623857498168945, + "learning_rate": 2.3444809261482653e-06, + "loss": 0.5387, + "num_input_tokens_seen": 8934408, + "step": 33380 + }, + { + "epoch": 8.750982961992136, + "grad_norm": 3.651210069656372, + "learning_rate": 2.3396475590125753e-06, + "loss": 0.4202, + "num_input_tokens_seen": 8935704, + "step": 33385 + }, + { + "epoch": 8.752293577981652, + "grad_norm": 3.1210944652557373, + "learning_rate": 2.334818934649494e-06, + "loss": 0.3851, + "num_input_tokens_seen": 8936824, + "step": 33390 + }, + { + "epoch": 8.753604193971167, + "grad_norm": 1.3954740762710571, + "learning_rate": 2.3299950540696504e-06, + "loss": 0.4933, + "num_input_tokens_seen": 8938264, + "step": 33395 + }, + { + "epoch": 8.754914809960681, + "grad_norm": 2.1065051555633545, + "learning_rate": 2.325175918282674e-06, + "loss": 0.2781, + "num_input_tokens_seen": 8939688, + "step": 33400 + }, + { + "epoch": 8.756225425950197, + "grad_norm": 3.3589208126068115, + "learning_rate": 2.3203615282971934e-06, + "loss": 0.5585, + "num_input_tokens_seen": 8940904, + "step": 33405 + }, + { + "epoch": 8.757536041939712, + "grad_norm": 2.9602582454681396, + "learning_rate": 2.315551885120859e-06, + "loss": 0.4764, + "num_input_tokens_seen": 8942072, + "step": 33410 + }, + { + "epoch": 8.758846657929226, + "grad_norm": 4.666058540344238, + "learning_rate": 2.3107469897603152e-06, + "loss": 0.5996, + "num_input_tokens_seen": 8943272, + "step": 33415 + }, + { + "epoch": 8.760157273918741, + "grad_norm": 1.777022361755371, + "learning_rate": 2.305946843221224e-06, + "loss": 0.5042, + "num_input_tokens_seen": 8945000, + "step": 33420 + }, + { + "epoch": 8.761467889908257, + "grad_norm": 2.7306079864501953, + "learning_rate": 2.3011514465082485e-06, + "loss": 0.5294, + "num_input_tokens_seen": 8946520, + "step": 33425 + }, + { + "epoch": 8.762778505897773, + "grad_norm": 2.775650978088379, + "learning_rate": 2.2963608006250555e-06, + "loss": 0.3958, + "num_input_tokens_seen": 8947896, + "step": 33430 + }, + { + "epoch": 8.764089121887286, + "grad_norm": 7.108036518096924, + "learning_rate": 2.2915749065743176e-06, + "loss": 0.5553, + "num_input_tokens_seen": 8949160, + "step": 33435 + }, + { + "epoch": 8.765399737876802, + "grad_norm": 5.475337505340576, + "learning_rate": 2.2867937653577173e-06, + "loss": 0.3711, + "num_input_tokens_seen": 8950568, + "step": 33440 + }, + { + "epoch": 8.766710353866317, + "grad_norm": 2.0587687492370605, + "learning_rate": 2.282017377975934e-06, + "loss": 0.5761, + "num_input_tokens_seen": 8951848, + "step": 33445 + }, + { + "epoch": 8.768020969855833, + "grad_norm": 12.92982292175293, + "learning_rate": 2.2772457454286578e-06, + "loss": 0.6671, + "num_input_tokens_seen": 8953320, + "step": 33450 + }, + { + "epoch": 8.769331585845347, + "grad_norm": 9.14317798614502, + "learning_rate": 2.2724788687145865e-06, + "loss": 0.5085, + "num_input_tokens_seen": 8954472, + "step": 33455 + }, + { + "epoch": 8.770642201834862, + "grad_norm": 48.261966705322266, + "learning_rate": 2.2677167488314193e-06, + "loss": 0.5275, + "num_input_tokens_seen": 8956264, + "step": 33460 + }, + { + "epoch": 8.771952817824378, + "grad_norm": 9.658742904663086, + "learning_rate": 2.2629593867758564e-06, + "loss": 0.4495, + "num_input_tokens_seen": 8957544, + "step": 33465 + }, + { + "epoch": 8.773263433813893, + "grad_norm": 1.7345727682113647, + "learning_rate": 2.258206783543601e-06, + "loss": 0.5004, + "num_input_tokens_seen": 8959160, + "step": 33470 + }, + { + "epoch": 8.774574049803407, + "grad_norm": 5.981035232543945, + "learning_rate": 2.253458940129383e-06, + "loss": 0.4483, + "num_input_tokens_seen": 8960248, + "step": 33475 + }, + { + "epoch": 8.775884665792923, + "grad_norm": 3.3594677448272705, + "learning_rate": 2.2487158575269006e-06, + "loss": 0.5316, + "num_input_tokens_seen": 8961384, + "step": 33480 + }, + { + "epoch": 8.777195281782438, + "grad_norm": 6.618640422821045, + "learning_rate": 2.243977536728878e-06, + "loss": 0.3614, + "num_input_tokens_seen": 8962728, + "step": 33485 + }, + { + "epoch": 8.778505897771954, + "grad_norm": 2.674121141433716, + "learning_rate": 2.2392439787270373e-06, + "loss": 0.5009, + "num_input_tokens_seen": 8963864, + "step": 33490 + }, + { + "epoch": 8.779816513761467, + "grad_norm": 2.62898588180542, + "learning_rate": 2.234515184512104e-06, + "loss": 0.4049, + "num_input_tokens_seen": 8965176, + "step": 33495 + }, + { + "epoch": 8.781127129750983, + "grad_norm": 8.386787414550781, + "learning_rate": 2.229791155073807e-06, + "loss": 0.5228, + "num_input_tokens_seen": 8967016, + "step": 33500 + }, + { + "epoch": 8.782437745740499, + "grad_norm": 2.6192471981048584, + "learning_rate": 2.225071891400879e-06, + "loss": 0.6346, + "num_input_tokens_seen": 8968856, + "step": 33505 + }, + { + "epoch": 8.783748361730012, + "grad_norm": 4.039888858795166, + "learning_rate": 2.2203573944810514e-06, + "loss": 0.5276, + "num_input_tokens_seen": 8970072, + "step": 33510 + }, + { + "epoch": 8.785058977719528, + "grad_norm": 2.8244500160217285, + "learning_rate": 2.21564766530106e-06, + "loss": 0.523, + "num_input_tokens_seen": 8971368, + "step": 33515 + }, + { + "epoch": 8.786369593709043, + "grad_norm": 16.10043716430664, + "learning_rate": 2.210942704846647e-06, + "loss": 0.6224, + "num_input_tokens_seen": 8972584, + "step": 33520 + }, + { + "epoch": 8.787680209698559, + "grad_norm": 1.2095730304718018, + "learning_rate": 2.2062425141025465e-06, + "loss": 0.5778, + "num_input_tokens_seen": 8974168, + "step": 33525 + }, + { + "epoch": 8.788990825688073, + "grad_norm": 2.4199042320251465, + "learning_rate": 2.2015470940525003e-06, + "loss": 0.362, + "num_input_tokens_seen": 8975512, + "step": 33530 + }, + { + "epoch": 8.790301441677588, + "grad_norm": 5.3335676193237305, + "learning_rate": 2.1968564456792573e-06, + "loss": 0.504, + "num_input_tokens_seen": 8976808, + "step": 33535 + }, + { + "epoch": 8.791612057667104, + "grad_norm": 1.5172346830368042, + "learning_rate": 2.1921705699645607e-06, + "loss": 0.4798, + "num_input_tokens_seen": 8978520, + "step": 33540 + }, + { + "epoch": 8.79292267365662, + "grad_norm": 3.9290497303009033, + "learning_rate": 2.187489467889153e-06, + "loss": 0.4384, + "num_input_tokens_seen": 8979432, + "step": 33545 + }, + { + "epoch": 8.794233289646133, + "grad_norm": 1.988110899925232, + "learning_rate": 2.182813140432785e-06, + "loss": 0.4022, + "num_input_tokens_seen": 8981160, + "step": 33550 + }, + { + "epoch": 8.795543905635649, + "grad_norm": 2.152194023132324, + "learning_rate": 2.1781415885742037e-06, + "loss": 0.4027, + "num_input_tokens_seen": 8982664, + "step": 33555 + }, + { + "epoch": 8.796854521625164, + "grad_norm": 5.268988609313965, + "learning_rate": 2.1734748132911515e-06, + "loss": 0.4293, + "num_input_tokens_seen": 8984424, + "step": 33560 + }, + { + "epoch": 8.79816513761468, + "grad_norm": 8.183021545410156, + "learning_rate": 2.1688128155603783e-06, + "loss": 0.5073, + "num_input_tokens_seen": 8985432, + "step": 33565 + }, + { + "epoch": 8.799475753604193, + "grad_norm": 34.60999298095703, + "learning_rate": 2.1641555963576366e-06, + "loss": 0.4872, + "num_input_tokens_seen": 8986808, + "step": 33570 + }, + { + "epoch": 8.800786369593709, + "grad_norm": 1.4440162181854248, + "learning_rate": 2.159503156657669e-06, + "loss": 0.4029, + "num_input_tokens_seen": 8988168, + "step": 33575 + }, + { + "epoch": 8.802096985583225, + "grad_norm": 2.459449291229248, + "learning_rate": 2.1548554974342244e-06, + "loss": 0.4139, + "num_input_tokens_seen": 8989992, + "step": 33580 + }, + { + "epoch": 8.80340760157274, + "grad_norm": 2.505298137664795, + "learning_rate": 2.15021261966005e-06, + "loss": 0.5626, + "num_input_tokens_seen": 8991368, + "step": 33585 + }, + { + "epoch": 8.804718217562254, + "grad_norm": 3.779865026473999, + "learning_rate": 2.145574524306898e-06, + "loss": 0.4463, + "num_input_tokens_seen": 8992248, + "step": 33590 + }, + { + "epoch": 8.80602883355177, + "grad_norm": 2.4362196922302246, + "learning_rate": 2.140941212345507e-06, + "loss": 0.4835, + "num_input_tokens_seen": 8993352, + "step": 33595 + }, + { + "epoch": 8.807339449541285, + "grad_norm": 3.921614408493042, + "learning_rate": 2.136312684745623e-06, + "loss": 0.3868, + "num_input_tokens_seen": 8994472, + "step": 33600 + }, + { + "epoch": 8.808650065530799, + "grad_norm": 1.2257968187332153, + "learning_rate": 2.1316889424759874e-06, + "loss": 0.5191, + "num_input_tokens_seen": 8996712, + "step": 33605 + }, + { + "epoch": 8.809960681520314, + "grad_norm": 9.01860237121582, + "learning_rate": 2.1270699865043405e-06, + "loss": 0.5513, + "num_input_tokens_seen": 8997928, + "step": 33610 + }, + { + "epoch": 8.81127129750983, + "grad_norm": 5.699176788330078, + "learning_rate": 2.122455817797428e-06, + "loss": 0.4132, + "num_input_tokens_seen": 8998968, + "step": 33615 + }, + { + "epoch": 8.812581913499345, + "grad_norm": 4.605726718902588, + "learning_rate": 2.1178464373209807e-06, + "loss": 0.7334, + "num_input_tokens_seen": 8999960, + "step": 33620 + }, + { + "epoch": 8.813892529488859, + "grad_norm": 3.7643935680389404, + "learning_rate": 2.1132418460397376e-06, + "loss": 0.4464, + "num_input_tokens_seen": 9001528, + "step": 33625 + }, + { + "epoch": 8.815203145478375, + "grad_norm": 1.2382242679595947, + "learning_rate": 2.1086420449174303e-06, + "loss": 0.443, + "num_input_tokens_seen": 9002968, + "step": 33630 + }, + { + "epoch": 8.81651376146789, + "grad_norm": 12.399721145629883, + "learning_rate": 2.1040470349167923e-06, + "loss": 0.4128, + "num_input_tokens_seen": 9004136, + "step": 33635 + }, + { + "epoch": 8.817824377457406, + "grad_norm": 2.833815813064575, + "learning_rate": 2.0994568169995455e-06, + "loss": 0.4511, + "num_input_tokens_seen": 9005192, + "step": 33640 + }, + { + "epoch": 8.81913499344692, + "grad_norm": 2.9633731842041016, + "learning_rate": 2.0948713921264214e-06, + "loss": 0.342, + "num_input_tokens_seen": 9006584, + "step": 33645 + }, + { + "epoch": 8.820445609436435, + "grad_norm": 2.9900448322296143, + "learning_rate": 2.0902907612571355e-06, + "loss": 0.3605, + "num_input_tokens_seen": 9007576, + "step": 33650 + }, + { + "epoch": 8.82175622542595, + "grad_norm": 1.9147443771362305, + "learning_rate": 2.085714925350407e-06, + "loss": 0.5767, + "num_input_tokens_seen": 9008904, + "step": 33655 + }, + { + "epoch": 8.823066841415466, + "grad_norm": 1.5306757688522339, + "learning_rate": 2.0811438853639537e-06, + "loss": 0.4199, + "num_input_tokens_seen": 9010248, + "step": 33660 + }, + { + "epoch": 8.82437745740498, + "grad_norm": 1.4540656805038452, + "learning_rate": 2.076577642254485e-06, + "loss": 0.3927, + "num_input_tokens_seen": 9011928, + "step": 33665 + }, + { + "epoch": 8.825688073394495, + "grad_norm": 4.811779022216797, + "learning_rate": 2.072016196977708e-06, + "loss": 0.5648, + "num_input_tokens_seen": 9013160, + "step": 33670 + }, + { + "epoch": 8.82699868938401, + "grad_norm": 1.6005785465240479, + "learning_rate": 2.0674595504883265e-06, + "loss": 0.3091, + "num_input_tokens_seen": 9014312, + "step": 33675 + }, + { + "epoch": 8.828309305373526, + "grad_norm": 2.219839334487915, + "learning_rate": 2.062907703740033e-06, + "loss": 0.2842, + "num_input_tokens_seen": 9015400, + "step": 33680 + }, + { + "epoch": 8.82961992136304, + "grad_norm": 1.567028522491455, + "learning_rate": 2.0583606576855236e-06, + "loss": 0.4834, + "num_input_tokens_seen": 9016600, + "step": 33685 + }, + { + "epoch": 8.830930537352556, + "grad_norm": 1.7921339273452759, + "learning_rate": 2.0538184132764897e-06, + "loss": 0.4221, + "num_input_tokens_seen": 9018120, + "step": 33690 + }, + { + "epoch": 8.832241153342071, + "grad_norm": 10.41969108581543, + "learning_rate": 2.0492809714636156e-06, + "loss": 0.6093, + "num_input_tokens_seen": 9019400, + "step": 33695 + }, + { + "epoch": 8.833551769331585, + "grad_norm": 9.32044792175293, + "learning_rate": 2.044748333196572e-06, + "loss": 0.4159, + "num_input_tokens_seen": 9020568, + "step": 33700 + }, + { + "epoch": 8.8348623853211, + "grad_norm": 2.860307455062866, + "learning_rate": 2.040220499424042e-06, + "loss": 0.4457, + "num_input_tokens_seen": 9022088, + "step": 33705 + }, + { + "epoch": 8.836173001310616, + "grad_norm": 3.2502620220184326, + "learning_rate": 2.0356974710936976e-06, + "loss": 0.4659, + "num_input_tokens_seen": 9023480, + "step": 33710 + }, + { + "epoch": 8.837483617300132, + "grad_norm": 3.221660852432251, + "learning_rate": 2.0311792491521874e-06, + "loss": 0.2736, + "num_input_tokens_seen": 9025224, + "step": 33715 + }, + { + "epoch": 8.838794233289645, + "grad_norm": 1.7659180164337158, + "learning_rate": 2.0266658345451715e-06, + "loss": 0.3881, + "num_input_tokens_seen": 9026888, + "step": 33720 + }, + { + "epoch": 8.840104849279161, + "grad_norm": 3.806874990463257, + "learning_rate": 2.022157228217303e-06, + "loss": 0.3448, + "num_input_tokens_seen": 9028168, + "step": 33725 + }, + { + "epoch": 8.841415465268676, + "grad_norm": 1.8443055152893066, + "learning_rate": 2.0176534311122263e-06, + "loss": 0.3227, + "num_input_tokens_seen": 9029784, + "step": 33730 + }, + { + "epoch": 8.842726081258192, + "grad_norm": 1.360216736793518, + "learning_rate": 2.0131544441725715e-06, + "loss": 0.581, + "num_input_tokens_seen": 9031256, + "step": 33735 + }, + { + "epoch": 8.844036697247706, + "grad_norm": 5.986873149871826, + "learning_rate": 2.0086602683399764e-06, + "loss": 0.5346, + "num_input_tokens_seen": 9032232, + "step": 33740 + }, + { + "epoch": 8.845347313237221, + "grad_norm": 3.8245675563812256, + "learning_rate": 2.0041709045550612e-06, + "loss": 0.5025, + "num_input_tokens_seen": 9033496, + "step": 33745 + }, + { + "epoch": 8.846657929226737, + "grad_norm": 4.687044143676758, + "learning_rate": 1.9996863537574406e-06, + "loss": 0.4728, + "num_input_tokens_seen": 9034888, + "step": 33750 + }, + { + "epoch": 8.847968545216252, + "grad_norm": 2.3191468715667725, + "learning_rate": 1.9952066168857252e-06, + "loss": 0.3801, + "num_input_tokens_seen": 9036856, + "step": 33755 + }, + { + "epoch": 8.849279161205766, + "grad_norm": 1.4690594673156738, + "learning_rate": 1.990731694877515e-06, + "loss": 0.4199, + "num_input_tokens_seen": 9038616, + "step": 33760 + }, + { + "epoch": 8.850589777195282, + "grad_norm": 2.832284450531006, + "learning_rate": 1.9862615886694054e-06, + "loss": 0.5985, + "num_input_tokens_seen": 9040264, + "step": 33765 + }, + { + "epoch": 8.851900393184797, + "grad_norm": 8.48769474029541, + "learning_rate": 1.9817962991969805e-06, + "loss": 0.4317, + "num_input_tokens_seen": 9041352, + "step": 33770 + }, + { + "epoch": 8.853211009174313, + "grad_norm": 2.964555025100708, + "learning_rate": 1.977335827394816e-06, + "loss": 0.5828, + "num_input_tokens_seen": 9042936, + "step": 33775 + }, + { + "epoch": 8.854521625163827, + "grad_norm": 2.08186674118042, + "learning_rate": 1.972880174196484e-06, + "loss": 0.5251, + "num_input_tokens_seen": 9044152, + "step": 33780 + }, + { + "epoch": 8.855832241153342, + "grad_norm": 5.766287803649902, + "learning_rate": 1.9684293405345445e-06, + "loss": 0.4485, + "num_input_tokens_seen": 9045384, + "step": 33785 + }, + { + "epoch": 8.857142857142858, + "grad_norm": 5.327592849731445, + "learning_rate": 1.9639833273405516e-06, + "loss": 0.6975, + "num_input_tokens_seen": 9046840, + "step": 33790 + }, + { + "epoch": 8.858453473132371, + "grad_norm": 4.5617547035217285, + "learning_rate": 1.959542135545042e-06, + "loss": 0.5416, + "num_input_tokens_seen": 9047896, + "step": 33795 + }, + { + "epoch": 8.859764089121887, + "grad_norm": 2.5326175689697266, + "learning_rate": 1.9551057660775527e-06, + "loss": 0.3364, + "num_input_tokens_seen": 9049112, + "step": 33800 + }, + { + "epoch": 8.861074705111402, + "grad_norm": 2.537968873977661, + "learning_rate": 1.950674219866608e-06, + "loss": 0.4281, + "num_input_tokens_seen": 9050360, + "step": 33805 + }, + { + "epoch": 8.862385321100918, + "grad_norm": 3.8017852306365967, + "learning_rate": 1.9462474978397204e-06, + "loss": 0.4687, + "num_input_tokens_seen": 9051576, + "step": 33810 + }, + { + "epoch": 8.863695937090432, + "grad_norm": 4.4486589431762695, + "learning_rate": 1.941825600923397e-06, + "loss": 0.3554, + "num_input_tokens_seen": 9052792, + "step": 33815 + }, + { + "epoch": 8.865006553079947, + "grad_norm": 1.684401512145996, + "learning_rate": 1.937408530043136e-06, + "loss": 0.3112, + "num_input_tokens_seen": 9054360, + "step": 33820 + }, + { + "epoch": 8.866317169069463, + "grad_norm": 1.706894040107727, + "learning_rate": 1.9329962861234257e-06, + "loss": 0.2992, + "num_input_tokens_seen": 9056184, + "step": 33825 + }, + { + "epoch": 8.867627785058978, + "grad_norm": 5.165828227996826, + "learning_rate": 1.9285888700877303e-06, + "loss": 0.6855, + "num_input_tokens_seen": 9058152, + "step": 33830 + }, + { + "epoch": 8.868938401048492, + "grad_norm": 4.6107707023620605, + "learning_rate": 1.92418628285852e-06, + "loss": 0.5341, + "num_input_tokens_seen": 9059272, + "step": 33835 + }, + { + "epoch": 8.870249017038008, + "grad_norm": 9.35816478729248, + "learning_rate": 1.9197885253572497e-06, + "loss": 0.4765, + "num_input_tokens_seen": 9060744, + "step": 33840 + }, + { + "epoch": 8.871559633027523, + "grad_norm": 2.181687116622925, + "learning_rate": 1.915395598504363e-06, + "loss": 0.4205, + "num_input_tokens_seen": 9062232, + "step": 33845 + }, + { + "epoch": 8.872870249017039, + "grad_norm": 3.3293752670288086, + "learning_rate": 1.9110075032192894e-06, + "loss": 0.5669, + "num_input_tokens_seen": 9063304, + "step": 33850 + }, + { + "epoch": 8.874180865006553, + "grad_norm": 2.727841377258301, + "learning_rate": 1.9066242404204487e-06, + "loss": 0.4871, + "num_input_tokens_seen": 9064568, + "step": 33855 + }, + { + "epoch": 8.875491480996068, + "grad_norm": 1.427282452583313, + "learning_rate": 1.902245811025255e-06, + "loss": 0.3331, + "num_input_tokens_seen": 9066296, + "step": 33860 + }, + { + "epoch": 8.876802096985584, + "grad_norm": 1.738394856452942, + "learning_rate": 1.8978722159501029e-06, + "loss": 0.4727, + "num_input_tokens_seen": 9067592, + "step": 33865 + }, + { + "epoch": 8.8781127129751, + "grad_norm": 1.1835179328918457, + "learning_rate": 1.8935034561103764e-06, + "loss": 0.6009, + "num_input_tokens_seen": 9069848, + "step": 33870 + }, + { + "epoch": 8.879423328964613, + "grad_norm": 5.9063591957092285, + "learning_rate": 1.889139532420453e-06, + "loss": 0.4376, + "num_input_tokens_seen": 9070840, + "step": 33875 + }, + { + "epoch": 8.880733944954128, + "grad_norm": 3.296322822570801, + "learning_rate": 1.8847804457936934e-06, + "loss": 0.4135, + "num_input_tokens_seen": 9071912, + "step": 33880 + }, + { + "epoch": 8.882044560943644, + "grad_norm": 3.213876962661743, + "learning_rate": 1.8804261971424458e-06, + "loss": 0.6219, + "num_input_tokens_seen": 9072952, + "step": 33885 + }, + { + "epoch": 8.883355176933158, + "grad_norm": 2.6267991065979004, + "learning_rate": 1.8760767873780476e-06, + "loss": 0.4277, + "num_input_tokens_seen": 9074360, + "step": 33890 + }, + { + "epoch": 8.884665792922673, + "grad_norm": 3.436613082885742, + "learning_rate": 1.871732217410821e-06, + "loss": 0.4917, + "num_input_tokens_seen": 9075720, + "step": 33895 + }, + { + "epoch": 8.885976408912189, + "grad_norm": 1.9252835512161255, + "learning_rate": 1.8673924881500826e-06, + "loss": 0.4295, + "num_input_tokens_seen": 9077256, + "step": 33900 + }, + { + "epoch": 8.887287024901704, + "grad_norm": 2.137075185775757, + "learning_rate": 1.8630576005041228e-06, + "loss": 0.5963, + "num_input_tokens_seen": 9078520, + "step": 33905 + }, + { + "epoch": 8.888597640891218, + "grad_norm": 2.929325580596924, + "learning_rate": 1.8587275553802352e-06, + "loss": 0.4219, + "num_input_tokens_seen": 9079816, + "step": 33910 + }, + { + "epoch": 8.889908256880734, + "grad_norm": 2.6635985374450684, + "learning_rate": 1.854402353684681e-06, + "loss": 0.3748, + "num_input_tokens_seen": 9081016, + "step": 33915 + }, + { + "epoch": 8.89121887287025, + "grad_norm": 2.1117630004882812, + "learning_rate": 1.8500819963227223e-06, + "loss": 0.3823, + "num_input_tokens_seen": 9082584, + "step": 33920 + }, + { + "epoch": 8.892529488859765, + "grad_norm": 2.9809279441833496, + "learning_rate": 1.8457664841986023e-06, + "loss": 0.4101, + "num_input_tokens_seen": 9083960, + "step": 33925 + }, + { + "epoch": 8.893840104849279, + "grad_norm": 6.060624599456787, + "learning_rate": 1.8414558182155456e-06, + "loss": 0.5076, + "num_input_tokens_seen": 9085528, + "step": 33930 + }, + { + "epoch": 8.895150720838794, + "grad_norm": 4.1738457679748535, + "learning_rate": 1.8371499992757723e-06, + "loss": 0.7728, + "num_input_tokens_seen": 9086728, + "step": 33935 + }, + { + "epoch": 8.89646133682831, + "grad_norm": 2.757796049118042, + "learning_rate": 1.832849028280484e-06, + "loss": 0.4357, + "num_input_tokens_seen": 9088008, + "step": 33940 + }, + { + "epoch": 8.897771952817825, + "grad_norm": 2.746246576309204, + "learning_rate": 1.8285529061298684e-06, + "loss": 0.5622, + "num_input_tokens_seen": 9089224, + "step": 33945 + }, + { + "epoch": 8.899082568807339, + "grad_norm": 3.3590786457061768, + "learning_rate": 1.82426163372309e-06, + "loss": 0.7397, + "num_input_tokens_seen": 9090456, + "step": 33950 + }, + { + "epoch": 8.900393184796854, + "grad_norm": 3.2175633907318115, + "learning_rate": 1.8199752119583052e-06, + "loss": 0.2947, + "num_input_tokens_seen": 9091560, + "step": 33955 + }, + { + "epoch": 8.90170380078637, + "grad_norm": 4.901510238647461, + "learning_rate": 1.8156936417326576e-06, + "loss": 0.3955, + "num_input_tokens_seen": 9092648, + "step": 33960 + }, + { + "epoch": 8.903014416775886, + "grad_norm": 1.2736613750457764, + "learning_rate": 1.8114169239422691e-06, + "loss": 0.3282, + "num_input_tokens_seen": 9094072, + "step": 33965 + }, + { + "epoch": 8.9043250327654, + "grad_norm": 3.1754443645477295, + "learning_rate": 1.8071450594822543e-06, + "loss": 0.4758, + "num_input_tokens_seen": 9095064, + "step": 33970 + }, + { + "epoch": 8.905635648754915, + "grad_norm": 3.3542957305908203, + "learning_rate": 1.8028780492467007e-06, + "loss": 0.3477, + "num_input_tokens_seen": 9096312, + "step": 33975 + }, + { + "epoch": 8.90694626474443, + "grad_norm": 3.6109750270843506, + "learning_rate": 1.798615894128694e-06, + "loss": 0.6141, + "num_input_tokens_seen": 9097736, + "step": 33980 + }, + { + "epoch": 8.908256880733944, + "grad_norm": 3.9093832969665527, + "learning_rate": 1.7943585950202902e-06, + "loss": 0.4446, + "num_input_tokens_seen": 9098888, + "step": 33985 + }, + { + "epoch": 8.90956749672346, + "grad_norm": 2.0063109397888184, + "learning_rate": 1.7901061528125347e-06, + "loss": 0.9155, + "num_input_tokens_seen": 9100024, + "step": 33990 + }, + { + "epoch": 8.910878112712975, + "grad_norm": 5.136262893676758, + "learning_rate": 1.7858585683954597e-06, + "loss": 0.5052, + "num_input_tokens_seen": 9101240, + "step": 33995 + }, + { + "epoch": 8.91218872870249, + "grad_norm": 2.3743555545806885, + "learning_rate": 1.7816158426580742e-06, + "loss": 0.6099, + "num_input_tokens_seen": 9102696, + "step": 34000 + }, + { + "epoch": 8.913499344692005, + "grad_norm": 4.5567097663879395, + "learning_rate": 1.7773779764883758e-06, + "loss": 0.5154, + "num_input_tokens_seen": 9103848, + "step": 34005 + }, + { + "epoch": 8.91480996068152, + "grad_norm": 2.1555423736572266, + "learning_rate": 1.7731449707733412e-06, + "loss": 0.4899, + "num_input_tokens_seen": 9105320, + "step": 34010 + }, + { + "epoch": 8.916120576671036, + "grad_norm": 1.5502132177352905, + "learning_rate": 1.7689168263989309e-06, + "loss": 0.2184, + "num_input_tokens_seen": 9106696, + "step": 34015 + }, + { + "epoch": 8.917431192660551, + "grad_norm": 2.292630434036255, + "learning_rate": 1.7646935442500872e-06, + "loss": 0.2538, + "num_input_tokens_seen": 9107880, + "step": 34020 + }, + { + "epoch": 8.918741808650065, + "grad_norm": 2.1289403438568115, + "learning_rate": 1.760475125210745e-06, + "loss": 0.5112, + "num_input_tokens_seen": 9108920, + "step": 34025 + }, + { + "epoch": 8.92005242463958, + "grad_norm": 2.4942471981048584, + "learning_rate": 1.7562615701637975e-06, + "loss": 0.66, + "num_input_tokens_seen": 9110024, + "step": 34030 + }, + { + "epoch": 8.921363040629096, + "grad_norm": 1.8633724451065063, + "learning_rate": 1.7520528799911395e-06, + "loss": 0.6023, + "num_input_tokens_seen": 9111848, + "step": 34035 + }, + { + "epoch": 8.922673656618612, + "grad_norm": 2.4766881465911865, + "learning_rate": 1.7478490555736471e-06, + "loss": 0.4649, + "num_input_tokens_seen": 9113224, + "step": 34040 + }, + { + "epoch": 8.923984272608125, + "grad_norm": 3.127255439758301, + "learning_rate": 1.7436500977911663e-06, + "loss": 0.5188, + "num_input_tokens_seen": 9114632, + "step": 34045 + }, + { + "epoch": 8.92529488859764, + "grad_norm": 5.793189525604248, + "learning_rate": 1.7394560075225414e-06, + "loss": 0.3205, + "num_input_tokens_seen": 9116440, + "step": 34050 + }, + { + "epoch": 8.926605504587156, + "grad_norm": 3.690842866897583, + "learning_rate": 1.7352667856455812e-06, + "loss": 0.4703, + "num_input_tokens_seen": 9117768, + "step": 34055 + }, + { + "epoch": 8.927916120576672, + "grad_norm": 8.084329605102539, + "learning_rate": 1.7310824330370873e-06, + "loss": 0.5892, + "num_input_tokens_seen": 9119304, + "step": 34060 + }, + { + "epoch": 8.929226736566186, + "grad_norm": 7.646589279174805, + "learning_rate": 1.7269029505728313e-06, + "loss": 0.4547, + "num_input_tokens_seen": 9120408, + "step": 34065 + }, + { + "epoch": 8.930537352555701, + "grad_norm": 0.9721210598945618, + "learning_rate": 1.7227283391275772e-06, + "loss": 0.4847, + "num_input_tokens_seen": 9121736, + "step": 34070 + }, + { + "epoch": 8.931847968545217, + "grad_norm": 2.746924877166748, + "learning_rate": 1.7185585995750598e-06, + "loss": 0.5119, + "num_input_tokens_seen": 9122936, + "step": 34075 + }, + { + "epoch": 8.93315858453473, + "grad_norm": 4.703282356262207, + "learning_rate": 1.714393732788e-06, + "loss": 0.6126, + "num_input_tokens_seen": 9124312, + "step": 34080 + }, + { + "epoch": 8.934469200524246, + "grad_norm": 2.4867053031921387, + "learning_rate": 1.710233739638098e-06, + "loss": 0.4168, + "num_input_tokens_seen": 9125672, + "step": 34085 + }, + { + "epoch": 8.935779816513762, + "grad_norm": 2.2165141105651855, + "learning_rate": 1.706078620996035e-06, + "loss": 0.5701, + "num_input_tokens_seen": 9126664, + "step": 34090 + }, + { + "epoch": 8.937090432503277, + "grad_norm": 5.555600643157959, + "learning_rate": 1.7019283777314653e-06, + "loss": 0.4117, + "num_input_tokens_seen": 9127976, + "step": 34095 + }, + { + "epoch": 8.938401048492791, + "grad_norm": 2.7149131298065186, + "learning_rate": 1.6977830107130305e-06, + "loss": 0.4228, + "num_input_tokens_seen": 9129192, + "step": 34100 + }, + { + "epoch": 8.939711664482306, + "grad_norm": 2.1792173385620117, + "learning_rate": 1.6936425208083473e-06, + "loss": 0.4276, + "num_input_tokens_seen": 9130696, + "step": 34105 + }, + { + "epoch": 8.941022280471822, + "grad_norm": 5.914462566375732, + "learning_rate": 1.6895069088840144e-06, + "loss": 0.4139, + "num_input_tokens_seen": 9132280, + "step": 34110 + }, + { + "epoch": 8.942332896461338, + "grad_norm": 6.189234733581543, + "learning_rate": 1.6853761758056114e-06, + "loss": 0.5057, + "num_input_tokens_seen": 9133768, + "step": 34115 + }, + { + "epoch": 8.943643512450851, + "grad_norm": 2.3017280101776123, + "learning_rate": 1.6812503224376857e-06, + "loss": 0.2915, + "num_input_tokens_seen": 9135272, + "step": 34120 + }, + { + "epoch": 8.944954128440367, + "grad_norm": 3.4770607948303223, + "learning_rate": 1.67712934964378e-06, + "loss": 0.3285, + "num_input_tokens_seen": 9136792, + "step": 34125 + }, + { + "epoch": 8.946264744429882, + "grad_norm": 2.507164239883423, + "learning_rate": 1.673013258286399e-06, + "loss": 0.4125, + "num_input_tokens_seen": 9138152, + "step": 34130 + }, + { + "epoch": 8.947575360419398, + "grad_norm": 2.737752676010132, + "learning_rate": 1.668902049227039e-06, + "loss": 0.4581, + "num_input_tokens_seen": 9139416, + "step": 34135 + }, + { + "epoch": 8.948885976408912, + "grad_norm": 7.390409469604492, + "learning_rate": 1.664795723326168e-06, + "loss": 0.8912, + "num_input_tokens_seen": 9140616, + "step": 34140 + }, + { + "epoch": 8.950196592398427, + "grad_norm": 3.842625617980957, + "learning_rate": 1.660694281443234e-06, + "loss": 0.2998, + "num_input_tokens_seen": 9141992, + "step": 34145 + }, + { + "epoch": 8.951507208387943, + "grad_norm": 10.006539344787598, + "learning_rate": 1.6565977244366564e-06, + "loss": 0.4537, + "num_input_tokens_seen": 9143128, + "step": 34150 + }, + { + "epoch": 8.952817824377458, + "grad_norm": 6.279438495635986, + "learning_rate": 1.6525060531638408e-06, + "loss": 0.5105, + "num_input_tokens_seen": 9144344, + "step": 34155 + }, + { + "epoch": 8.954128440366972, + "grad_norm": 3.5621497631073, + "learning_rate": 1.6484192684811634e-06, + "loss": 0.4829, + "num_input_tokens_seen": 9145688, + "step": 34160 + }, + { + "epoch": 8.955439056356488, + "grad_norm": 3.9021241664886475, + "learning_rate": 1.644337371243987e-06, + "loss": 0.3815, + "num_input_tokens_seen": 9146856, + "step": 34165 + }, + { + "epoch": 8.956749672346003, + "grad_norm": 2.288755178451538, + "learning_rate": 1.6402603623066448e-06, + "loss": 0.6212, + "num_input_tokens_seen": 9148200, + "step": 34170 + }, + { + "epoch": 8.958060288335517, + "grad_norm": 4.2530837059021, + "learning_rate": 1.6361882425224461e-06, + "loss": 0.4008, + "num_input_tokens_seen": 9149432, + "step": 34175 + }, + { + "epoch": 8.959370904325032, + "grad_norm": 1.954293966293335, + "learning_rate": 1.6321210127436814e-06, + "loss": 0.3558, + "num_input_tokens_seen": 9150648, + "step": 34180 + }, + { + "epoch": 8.960681520314548, + "grad_norm": 4.0449347496032715, + "learning_rate": 1.6280586738216114e-06, + "loss": 0.5355, + "num_input_tokens_seen": 9152088, + "step": 34185 + }, + { + "epoch": 8.961992136304064, + "grad_norm": 2.2327234745025635, + "learning_rate": 1.6240012266064752e-06, + "loss": 0.5388, + "num_input_tokens_seen": 9153320, + "step": 34190 + }, + { + "epoch": 8.963302752293577, + "grad_norm": 2.4310929775238037, + "learning_rate": 1.619948671947491e-06, + "loss": 0.4076, + "num_input_tokens_seen": 9154424, + "step": 34195 + }, + { + "epoch": 8.964613368283093, + "grad_norm": 2.5348637104034424, + "learning_rate": 1.6159010106928523e-06, + "loss": 0.8102, + "num_input_tokens_seen": 9156184, + "step": 34200 + }, + { + "epoch": 8.965923984272608, + "grad_norm": 4.1426615715026855, + "learning_rate": 1.611858243689729e-06, + "loss": 0.6706, + "num_input_tokens_seen": 9157480, + "step": 34205 + }, + { + "epoch": 8.967234600262124, + "grad_norm": 2.8301243782043457, + "learning_rate": 1.6078203717842633e-06, + "loss": 0.6299, + "num_input_tokens_seen": 9158824, + "step": 34210 + }, + { + "epoch": 8.968545216251638, + "grad_norm": 2.8504724502563477, + "learning_rate": 1.6037873958215738e-06, + "loss": 0.5307, + "num_input_tokens_seen": 9160568, + "step": 34215 + }, + { + "epoch": 8.969855832241153, + "grad_norm": 1.8505300283432007, + "learning_rate": 1.5997593166457575e-06, + "loss": 0.339, + "num_input_tokens_seen": 9161832, + "step": 34220 + }, + { + "epoch": 8.971166448230669, + "grad_norm": 12.569279670715332, + "learning_rate": 1.5957361350998846e-06, + "loss": 0.5763, + "num_input_tokens_seen": 9163048, + "step": 34225 + }, + { + "epoch": 8.972477064220184, + "grad_norm": 2.4533979892730713, + "learning_rate": 1.5917178520259979e-06, + "loss": 0.49, + "num_input_tokens_seen": 9164312, + "step": 34230 + }, + { + "epoch": 8.973787680209698, + "grad_norm": 2.8253366947174072, + "learning_rate": 1.5877044682651194e-06, + "loss": 0.6327, + "num_input_tokens_seen": 9165224, + "step": 34235 + }, + { + "epoch": 8.975098296199214, + "grad_norm": 4.034358978271484, + "learning_rate": 1.5836959846572408e-06, + "loss": 0.5334, + "num_input_tokens_seen": 9166696, + "step": 34240 + }, + { + "epoch": 8.97640891218873, + "grad_norm": 1.5629727840423584, + "learning_rate": 1.579692402041333e-06, + "loss": 0.5164, + "num_input_tokens_seen": 9168024, + "step": 34245 + }, + { + "epoch": 8.977719528178245, + "grad_norm": 3.227451801300049, + "learning_rate": 1.5756937212553391e-06, + "loss": 0.5231, + "num_input_tokens_seen": 9169336, + "step": 34250 + }, + { + "epoch": 8.979030144167758, + "grad_norm": 3.1810214519500732, + "learning_rate": 1.5716999431361762e-06, + "loss": 0.3804, + "num_input_tokens_seen": 9170296, + "step": 34255 + }, + { + "epoch": 8.980340760157274, + "grad_norm": 2.077378749847412, + "learning_rate": 1.5677110685197366e-06, + "loss": 0.5277, + "num_input_tokens_seen": 9171736, + "step": 34260 + }, + { + "epoch": 8.98165137614679, + "grad_norm": 2.745291233062744, + "learning_rate": 1.5637270982408775e-06, + "loss": 0.5066, + "num_input_tokens_seen": 9173096, + "step": 34265 + }, + { + "epoch": 8.982961992136303, + "grad_norm": 3.42781925201416, + "learning_rate": 1.559748033133443e-06, + "loss": 0.3607, + "num_input_tokens_seen": 9174296, + "step": 34270 + }, + { + "epoch": 8.984272608125819, + "grad_norm": 10.47968864440918, + "learning_rate": 1.555773874030242e-06, + "loss": 0.48, + "num_input_tokens_seen": 9175560, + "step": 34275 + }, + { + "epoch": 8.985583224115334, + "grad_norm": 1.2068703174591064, + "learning_rate": 1.5518046217630615e-06, + "loss": 0.4069, + "num_input_tokens_seen": 9177304, + "step": 34280 + }, + { + "epoch": 8.98689384010485, + "grad_norm": 1.9044114351272583, + "learning_rate": 1.5478402771626599e-06, + "loss": 0.3103, + "num_input_tokens_seen": 9178760, + "step": 34285 + }, + { + "epoch": 8.988204456094364, + "grad_norm": 3.6748032569885254, + "learning_rate": 1.5438808410587646e-06, + "loss": 0.4299, + "num_input_tokens_seen": 9179848, + "step": 34290 + }, + { + "epoch": 8.98951507208388, + "grad_norm": 2.9791340827941895, + "learning_rate": 1.5399263142800825e-06, + "loss": 0.5104, + "num_input_tokens_seen": 9181336, + "step": 34295 + }, + { + "epoch": 8.990825688073395, + "grad_norm": 2.281458616256714, + "learning_rate": 1.5359766976542873e-06, + "loss": 0.4715, + "num_input_tokens_seen": 9182488, + "step": 34300 + }, + { + "epoch": 8.99213630406291, + "grad_norm": 2.8175835609436035, + "learning_rate": 1.5320319920080262e-06, + "loss": 0.5954, + "num_input_tokens_seen": 9183912, + "step": 34305 + }, + { + "epoch": 8.993446920052424, + "grad_norm": 3.459115505218506, + "learning_rate": 1.5280921981669166e-06, + "loss": 0.4986, + "num_input_tokens_seen": 9184968, + "step": 34310 + }, + { + "epoch": 8.99475753604194, + "grad_norm": 2.0297317504882812, + "learning_rate": 1.5241573169555573e-06, + "loss": 0.2702, + "num_input_tokens_seen": 9186440, + "step": 34315 + }, + { + "epoch": 8.996068152031455, + "grad_norm": 15.479453086853027, + "learning_rate": 1.5202273491975061e-06, + "loss": 0.5479, + "num_input_tokens_seen": 9187544, + "step": 34320 + }, + { + "epoch": 8.997378768020969, + "grad_norm": 2.059860944747925, + "learning_rate": 1.5163022957152994e-06, + "loss": 0.6873, + "num_input_tokens_seen": 9189080, + "step": 34325 + }, + { + "epoch": 8.998689384010484, + "grad_norm": 3.8344576358795166, + "learning_rate": 1.5123821573304442e-06, + "loss": 0.3671, + "num_input_tokens_seen": 9190856, + "step": 34330 + }, + { + "epoch": 9.0, + "grad_norm": 4.219173908233643, + "learning_rate": 1.5084669348634255e-06, + "loss": 0.6636, + "num_input_tokens_seen": 9192352, + "step": 34335 + }, + { + "epoch": 9.001310615989516, + "grad_norm": 2.2672839164733887, + "learning_rate": 1.5045566291336854e-06, + "loss": 0.4878, + "num_input_tokens_seen": 9193840, + "step": 34340 + }, + { + "epoch": 9.002359108781127, + "eval_loss": 0.4643946588039398, + "eval_runtime": 18.2173, + "eval_samples_per_second": 46.549, + "eval_steps_per_second": 23.275, + "num_input_tokens_seen": 9194752, + "step": 34344 + }, + { + "epoch": 9.00262123197903, + "grad_norm": 1.4564013481140137, + "learning_rate": 1.500651240959644e-06, + "loss": 0.5075, + "num_input_tokens_seen": 9195520, + "step": 34345 + }, + { + "epoch": 9.003931847968545, + "grad_norm": 1.4877833127975464, + "learning_rate": 1.4967507711586976e-06, + "loss": 0.4974, + "num_input_tokens_seen": 9197056, + "step": 34350 + }, + { + "epoch": 9.00524246395806, + "grad_norm": 1.5413388013839722, + "learning_rate": 1.4928552205472012e-06, + "loss": 0.4649, + "num_input_tokens_seen": 9198336, + "step": 34355 + }, + { + "epoch": 9.006553079947576, + "grad_norm": 4.959609508514404, + "learning_rate": 1.488964589940489e-06, + "loss": 0.3868, + "num_input_tokens_seen": 9199488, + "step": 34360 + }, + { + "epoch": 9.00786369593709, + "grad_norm": 1.9724352359771729, + "learning_rate": 1.4850788801528653e-06, + "loss": 0.3322, + "num_input_tokens_seen": 9200752, + "step": 34365 + }, + { + "epoch": 9.009174311926605, + "grad_norm": 1.8944079875946045, + "learning_rate": 1.4811980919976043e-06, + "loss": 0.5696, + "num_input_tokens_seen": 9202208, + "step": 34370 + }, + { + "epoch": 9.01048492791612, + "grad_norm": 2.4743244647979736, + "learning_rate": 1.4773222262869423e-06, + "loss": 0.4976, + "num_input_tokens_seen": 9203824, + "step": 34375 + }, + { + "epoch": 9.011795543905636, + "grad_norm": 3.5267536640167236, + "learning_rate": 1.4734512838320974e-06, + "loss": 0.442, + "num_input_tokens_seen": 9205392, + "step": 34380 + }, + { + "epoch": 9.01310615989515, + "grad_norm": 3.8050713539123535, + "learning_rate": 1.4695852654432463e-06, + "loss": 0.4075, + "num_input_tokens_seen": 9206528, + "step": 34385 + }, + { + "epoch": 9.014416775884666, + "grad_norm": 5.014498710632324, + "learning_rate": 1.4657241719295362e-06, + "loss": 0.4543, + "num_input_tokens_seen": 9207552, + "step": 34390 + }, + { + "epoch": 9.015727391874181, + "grad_norm": 2.88602352142334, + "learning_rate": 1.4618680040990983e-06, + "loss": 0.3212, + "num_input_tokens_seen": 9209104, + "step": 34395 + }, + { + "epoch": 9.017038007863697, + "grad_norm": 2.3894753456115723, + "learning_rate": 1.4580167627590153e-06, + "loss": 0.4634, + "num_input_tokens_seen": 9210656, + "step": 34400 + }, + { + "epoch": 9.01834862385321, + "grad_norm": 2.5297818183898926, + "learning_rate": 1.4541704487153474e-06, + "loss": 0.4159, + "num_input_tokens_seen": 9211968, + "step": 34405 + }, + { + "epoch": 9.019659239842726, + "grad_norm": 8.938386917114258, + "learning_rate": 1.4503290627731175e-06, + "loss": 0.5498, + "num_input_tokens_seen": 9212960, + "step": 34410 + }, + { + "epoch": 9.020969855832242, + "grad_norm": 3.892439365386963, + "learning_rate": 1.4464926057363298e-06, + "loss": 0.5469, + "num_input_tokens_seen": 9214368, + "step": 34415 + }, + { + "epoch": 9.022280471821757, + "grad_norm": 2.3504483699798584, + "learning_rate": 1.4426610784079391e-06, + "loss": 0.3886, + "num_input_tokens_seen": 9215600, + "step": 34420 + }, + { + "epoch": 9.02359108781127, + "grad_norm": 2.390437602996826, + "learning_rate": 1.438834481589882e-06, + "loss": 0.5124, + "num_input_tokens_seen": 9216960, + "step": 34425 + }, + { + "epoch": 9.024901703800786, + "grad_norm": 4.026634693145752, + "learning_rate": 1.4350128160830562e-06, + "loss": 0.4849, + "num_input_tokens_seen": 9218208, + "step": 34430 + }, + { + "epoch": 9.026212319790302, + "grad_norm": 7.818960189819336, + "learning_rate": 1.4311960826873305e-06, + "loss": 0.3701, + "num_input_tokens_seen": 9219088, + "step": 34435 + }, + { + "epoch": 9.027522935779816, + "grad_norm": 3.080639600753784, + "learning_rate": 1.427384282201541e-06, + "loss": 0.4442, + "num_input_tokens_seen": 9220128, + "step": 34440 + }, + { + "epoch": 9.028833551769331, + "grad_norm": 1.4973219633102417, + "learning_rate": 1.4235774154234855e-06, + "loss": 0.743, + "num_input_tokens_seen": 9221504, + "step": 34445 + }, + { + "epoch": 9.030144167758847, + "grad_norm": 1.7595537900924683, + "learning_rate": 1.4197754831499488e-06, + "loss": 0.3761, + "num_input_tokens_seen": 9222832, + "step": 34450 + }, + { + "epoch": 9.031454783748362, + "grad_norm": 1.3296682834625244, + "learning_rate": 1.4159784861766584e-06, + "loss": 0.4225, + "num_input_tokens_seen": 9224752, + "step": 34455 + }, + { + "epoch": 9.032765399737876, + "grad_norm": 3.883890151977539, + "learning_rate": 1.4121864252983174e-06, + "loss": 0.4957, + "num_input_tokens_seen": 9226064, + "step": 34460 + }, + { + "epoch": 9.034076015727392, + "grad_norm": 1.4291505813598633, + "learning_rate": 1.408399301308605e-06, + "loss": 0.624, + "num_input_tokens_seen": 9227216, + "step": 34465 + }, + { + "epoch": 9.035386631716907, + "grad_norm": 4.208733558654785, + "learning_rate": 1.4046171150001508e-06, + "loss": 0.407, + "num_input_tokens_seen": 9228736, + "step": 34470 + }, + { + "epoch": 9.036697247706423, + "grad_norm": 3.129603147506714, + "learning_rate": 1.4008398671645688e-06, + "loss": 0.4758, + "num_input_tokens_seen": 9230320, + "step": 34475 + }, + { + "epoch": 9.038007863695936, + "grad_norm": 6.518104076385498, + "learning_rate": 1.3970675585924241e-06, + "loss": 0.5921, + "num_input_tokens_seen": 9231360, + "step": 34480 + }, + { + "epoch": 9.039318479685452, + "grad_norm": 3.386561870574951, + "learning_rate": 1.3933001900732572e-06, + "loss": 0.4905, + "num_input_tokens_seen": 9234640, + "step": 34485 + }, + { + "epoch": 9.040629095674968, + "grad_norm": 3.3229198455810547, + "learning_rate": 1.3895377623955707e-06, + "loss": 0.2699, + "num_input_tokens_seen": 9235712, + "step": 34490 + }, + { + "epoch": 9.041939711664483, + "grad_norm": 2.983163356781006, + "learning_rate": 1.3857802763468403e-06, + "loss": 0.4796, + "num_input_tokens_seen": 9236736, + "step": 34495 + }, + { + "epoch": 9.043250327653997, + "grad_norm": 4.615990161895752, + "learning_rate": 1.3820277327134866e-06, + "loss": 0.5384, + "num_input_tokens_seen": 9237840, + "step": 34500 + }, + { + "epoch": 9.044560943643512, + "grad_norm": 2.005601644515991, + "learning_rate": 1.3782801322809262e-06, + "loss": 0.3978, + "num_input_tokens_seen": 9239520, + "step": 34505 + }, + { + "epoch": 9.045871559633028, + "grad_norm": 5.616878032684326, + "learning_rate": 1.374537475833515e-06, + "loss": 0.4691, + "num_input_tokens_seen": 9240928, + "step": 34510 + }, + { + "epoch": 9.047182175622543, + "grad_norm": 3.2023415565490723, + "learning_rate": 1.37079976415459e-06, + "loss": 0.4512, + "num_input_tokens_seen": 9242208, + "step": 34515 + }, + { + "epoch": 9.048492791612057, + "grad_norm": 2.3044052124023438, + "learning_rate": 1.3670669980264477e-06, + "loss": 0.5627, + "num_input_tokens_seen": 9244160, + "step": 34520 + }, + { + "epoch": 9.049803407601573, + "grad_norm": 4.823723793029785, + "learning_rate": 1.3633391782303468e-06, + "loss": 0.6994, + "num_input_tokens_seen": 9245392, + "step": 34525 + }, + { + "epoch": 9.051114023591088, + "grad_norm": 1.9743802547454834, + "learning_rate": 1.3596163055465154e-06, + "loss": 0.5626, + "num_input_tokens_seen": 9246704, + "step": 34530 + }, + { + "epoch": 9.052424639580602, + "grad_norm": 2.917560577392578, + "learning_rate": 1.3558983807541476e-06, + "loss": 0.5343, + "num_input_tokens_seen": 9247840, + "step": 34535 + }, + { + "epoch": 9.053735255570118, + "grad_norm": 5.18890380859375, + "learning_rate": 1.3521854046313897e-06, + "loss": 0.3531, + "num_input_tokens_seen": 9248848, + "step": 34540 + }, + { + "epoch": 9.055045871559633, + "grad_norm": 4.025966167449951, + "learning_rate": 1.3484773779553677e-06, + "loss": 0.5978, + "num_input_tokens_seen": 9249984, + "step": 34545 + }, + { + "epoch": 9.056356487549149, + "grad_norm": 2.129265069961548, + "learning_rate": 1.3447743015021636e-06, + "loss": 0.5851, + "num_input_tokens_seen": 9251040, + "step": 34550 + }, + { + "epoch": 9.057667103538662, + "grad_norm": 1.588087558746338, + "learning_rate": 1.3410761760468265e-06, + "loss": 0.4817, + "num_input_tokens_seen": 9253056, + "step": 34555 + }, + { + "epoch": 9.058977719528178, + "grad_norm": 2.2193617820739746, + "learning_rate": 1.3373830023633598e-06, + "loss": 0.4039, + "num_input_tokens_seen": 9254464, + "step": 34560 + }, + { + "epoch": 9.060288335517694, + "grad_norm": 3.2164793014526367, + "learning_rate": 1.3336947812247507e-06, + "loss": 0.5888, + "num_input_tokens_seen": 9255552, + "step": 34565 + }, + { + "epoch": 9.061598951507209, + "grad_norm": 3.3096370697021484, + "learning_rate": 1.3300115134029345e-06, + "loss": 0.4475, + "num_input_tokens_seen": 9256656, + "step": 34570 + }, + { + "epoch": 9.062909567496723, + "grad_norm": 2.1192827224731445, + "learning_rate": 1.3263331996688055e-06, + "loss": 0.2927, + "num_input_tokens_seen": 9258256, + "step": 34575 + }, + { + "epoch": 9.064220183486238, + "grad_norm": 1.6465007066726685, + "learning_rate": 1.3226598407922342e-06, + "loss": 0.4291, + "num_input_tokens_seen": 9260240, + "step": 34580 + }, + { + "epoch": 9.065530799475754, + "grad_norm": 1.6611133813858032, + "learning_rate": 1.318991437542047e-06, + "loss": 0.4694, + "num_input_tokens_seen": 9262064, + "step": 34585 + }, + { + "epoch": 9.06684141546527, + "grad_norm": 1.2353923320770264, + "learning_rate": 1.3153279906860355e-06, + "loss": 0.383, + "num_input_tokens_seen": 9263328, + "step": 34590 + }, + { + "epoch": 9.068152031454783, + "grad_norm": 1.575118899345398, + "learning_rate": 1.3116695009909474e-06, + "loss": 0.4576, + "num_input_tokens_seen": 9264528, + "step": 34595 + }, + { + "epoch": 9.069462647444299, + "grad_norm": 6.84708309173584, + "learning_rate": 1.3080159692225063e-06, + "loss": 0.4345, + "num_input_tokens_seen": 9265552, + "step": 34600 + }, + { + "epoch": 9.070773263433814, + "grad_norm": 3.114293098449707, + "learning_rate": 1.304367396145384e-06, + "loss": 0.4688, + "num_input_tokens_seen": 9267104, + "step": 34605 + }, + { + "epoch": 9.07208387942333, + "grad_norm": 8.277276992797852, + "learning_rate": 1.300723782523225e-06, + "loss": 0.367, + "num_input_tokens_seen": 9268256, + "step": 34610 + }, + { + "epoch": 9.073394495412844, + "grad_norm": 4.358921051025391, + "learning_rate": 1.2970851291186276e-06, + "loss": 0.6046, + "num_input_tokens_seen": 9269728, + "step": 34615 + }, + { + "epoch": 9.07470511140236, + "grad_norm": 2.0237207412719727, + "learning_rate": 1.2934514366931578e-06, + "loss": 0.6172, + "num_input_tokens_seen": 9271504, + "step": 34620 + }, + { + "epoch": 9.076015727391875, + "grad_norm": 2.152257204055786, + "learning_rate": 1.2898227060073403e-06, + "loss": 0.3211, + "num_input_tokens_seen": 9272816, + "step": 34625 + }, + { + "epoch": 9.077326343381388, + "grad_norm": 2.474926710128784, + "learning_rate": 1.2861989378206624e-06, + "loss": 0.4226, + "num_input_tokens_seen": 9274000, + "step": 34630 + }, + { + "epoch": 9.078636959370904, + "grad_norm": 1.913926362991333, + "learning_rate": 1.2825801328915699e-06, + "loss": 0.6946, + "num_input_tokens_seen": 9275280, + "step": 34635 + }, + { + "epoch": 9.07994757536042, + "grad_norm": 9.139344215393066, + "learning_rate": 1.2789662919774792e-06, + "loss": 0.4923, + "num_input_tokens_seen": 9276448, + "step": 34640 + }, + { + "epoch": 9.081258191349935, + "grad_norm": 2.069775104522705, + "learning_rate": 1.2753574158347547e-06, + "loss": 0.4388, + "num_input_tokens_seen": 9277808, + "step": 34645 + }, + { + "epoch": 9.082568807339449, + "grad_norm": 3.9888558387756348, + "learning_rate": 1.271753505218734e-06, + "loss": 0.6755, + "num_input_tokens_seen": 9279392, + "step": 34650 + }, + { + "epoch": 9.083879423328964, + "grad_norm": 3.131077527999878, + "learning_rate": 1.2681545608837026e-06, + "loss": 0.3159, + "num_input_tokens_seen": 9281056, + "step": 34655 + }, + { + "epoch": 9.08519003931848, + "grad_norm": 2.007174015045166, + "learning_rate": 1.2645605835829132e-06, + "loss": 0.3868, + "num_input_tokens_seen": 9282400, + "step": 34660 + }, + { + "epoch": 9.086500655307995, + "grad_norm": 4.68031120300293, + "learning_rate": 1.2609715740685869e-06, + "loss": 0.4076, + "num_input_tokens_seen": 9283568, + "step": 34665 + }, + { + "epoch": 9.08781127129751, + "grad_norm": 0.5522331595420837, + "learning_rate": 1.257387533091889e-06, + "loss": 0.315, + "num_input_tokens_seen": 9286448, + "step": 34670 + }, + { + "epoch": 9.089121887287025, + "grad_norm": 1.0572540760040283, + "learning_rate": 1.2538084614029527e-06, + "loss": 0.3557, + "num_input_tokens_seen": 9288048, + "step": 34675 + }, + { + "epoch": 9.09043250327654, + "grad_norm": 3.2085604667663574, + "learning_rate": 1.2502343597508792e-06, + "loss": 0.5195, + "num_input_tokens_seen": 9289408, + "step": 34680 + }, + { + "epoch": 9.091743119266056, + "grad_norm": 1.5650553703308105, + "learning_rate": 1.2466652288837227e-06, + "loss": 0.5, + "num_input_tokens_seen": 9290912, + "step": 34685 + }, + { + "epoch": 9.09305373525557, + "grad_norm": 2.993685007095337, + "learning_rate": 1.2431010695484858e-06, + "loss": 0.4481, + "num_input_tokens_seen": 9292352, + "step": 34690 + }, + { + "epoch": 9.094364351245085, + "grad_norm": 2.4522979259490967, + "learning_rate": 1.2395418824911464e-06, + "loss": 0.3599, + "num_input_tokens_seen": 9293376, + "step": 34695 + }, + { + "epoch": 9.0956749672346, + "grad_norm": 5.595747470855713, + "learning_rate": 1.2359876684566368e-06, + "loss": 0.4849, + "num_input_tokens_seen": 9294800, + "step": 34700 + }, + { + "epoch": 9.096985583224116, + "grad_norm": 15.209753036499023, + "learning_rate": 1.232438428188848e-06, + "loss": 0.7993, + "num_input_tokens_seen": 9295872, + "step": 34705 + }, + { + "epoch": 9.09829619921363, + "grad_norm": 2.178891897201538, + "learning_rate": 1.228894162430627e-06, + "loss": 0.3643, + "num_input_tokens_seen": 9297456, + "step": 34710 + }, + { + "epoch": 9.099606815203146, + "grad_norm": 3.0013649463653564, + "learning_rate": 1.225354871923784e-06, + "loss": 1.16, + "num_input_tokens_seen": 9298496, + "step": 34715 + }, + { + "epoch": 9.100917431192661, + "grad_norm": 3.4097914695739746, + "learning_rate": 1.2218205574090896e-06, + "loss": 0.4819, + "num_input_tokens_seen": 9299360, + "step": 34720 + }, + { + "epoch": 9.102228047182175, + "grad_norm": 3.8071508407592773, + "learning_rate": 1.2182912196262664e-06, + "loss": 0.6873, + "num_input_tokens_seen": 9300640, + "step": 34725 + }, + { + "epoch": 9.10353866317169, + "grad_norm": 2.3823466300964355, + "learning_rate": 1.2147668593139982e-06, + "loss": 0.6026, + "num_input_tokens_seen": 9302096, + "step": 34730 + }, + { + "epoch": 9.104849279161206, + "grad_norm": 5.189424514770508, + "learning_rate": 1.2112474772099285e-06, + "loss": 0.431, + "num_input_tokens_seen": 9303424, + "step": 34735 + }, + { + "epoch": 9.106159895150721, + "grad_norm": 7.304616928100586, + "learning_rate": 1.2077330740506598e-06, + "loss": 0.5773, + "num_input_tokens_seen": 9304384, + "step": 34740 + }, + { + "epoch": 9.107470511140235, + "grad_norm": 13.202887535095215, + "learning_rate": 1.2042236505717452e-06, + "loss": 0.3985, + "num_input_tokens_seen": 9305344, + "step": 34745 + }, + { + "epoch": 9.10878112712975, + "grad_norm": 4.353958606719971, + "learning_rate": 1.2007192075077085e-06, + "loss": 0.4578, + "num_input_tokens_seen": 9306624, + "step": 34750 + }, + { + "epoch": 9.110091743119266, + "grad_norm": 3.2523036003112793, + "learning_rate": 1.1972197455920181e-06, + "loss": 0.4673, + "num_input_tokens_seen": 9307888, + "step": 34755 + }, + { + "epoch": 9.111402359108782, + "grad_norm": 1.8359781503677368, + "learning_rate": 1.193725265557108e-06, + "loss": 0.5125, + "num_input_tokens_seen": 9309600, + "step": 34760 + }, + { + "epoch": 9.112712975098296, + "grad_norm": 2.422805070877075, + "learning_rate": 1.1902357681343623e-06, + "loss": 0.5306, + "num_input_tokens_seen": 9311312, + "step": 34765 + }, + { + "epoch": 9.114023591087811, + "grad_norm": 2.7963271141052246, + "learning_rate": 1.1867512540541388e-06, + "loss": 0.3944, + "num_input_tokens_seen": 9312544, + "step": 34770 + }, + { + "epoch": 9.115334207077327, + "grad_norm": 1.8000760078430176, + "learning_rate": 1.1832717240457231e-06, + "loss": 0.3793, + "num_input_tokens_seen": 9313600, + "step": 34775 + }, + { + "epoch": 9.116644823066842, + "grad_norm": 2.9343082904815674, + "learning_rate": 1.1797971788373857e-06, + "loss": 0.3983, + "num_input_tokens_seen": 9314752, + "step": 34780 + }, + { + "epoch": 9.117955439056356, + "grad_norm": 2.12395977973938, + "learning_rate": 1.1763276191563422e-06, + "loss": 0.6114, + "num_input_tokens_seen": 9317488, + "step": 34785 + }, + { + "epoch": 9.119266055045872, + "grad_norm": 6.9714508056640625, + "learning_rate": 1.1728630457287587e-06, + "loss": 0.5379, + "num_input_tokens_seen": 9318944, + "step": 34790 + }, + { + "epoch": 9.120576671035387, + "grad_norm": 3.4018490314483643, + "learning_rate": 1.1694034592797747e-06, + "loss": 0.3859, + "num_input_tokens_seen": 9320320, + "step": 34795 + }, + { + "epoch": 9.1218872870249, + "grad_norm": 3.6750497817993164, + "learning_rate": 1.1659488605334695e-06, + "loss": 0.5164, + "num_input_tokens_seen": 9321968, + "step": 34800 + }, + { + "epoch": 9.123197903014416, + "grad_norm": 14.863167762756348, + "learning_rate": 1.1624992502128868e-06, + "loss": 0.576, + "num_input_tokens_seen": 9323728, + "step": 34805 + }, + { + "epoch": 9.124508519003932, + "grad_norm": 2.22867488861084, + "learning_rate": 1.159054629040024e-06, + "loss": 0.522, + "num_input_tokens_seen": 9324864, + "step": 34810 + }, + { + "epoch": 9.125819134993447, + "grad_norm": 5.285652160644531, + "learning_rate": 1.1556149977358295e-06, + "loss": 0.6934, + "num_input_tokens_seen": 9325984, + "step": 34815 + }, + { + "epoch": 9.127129750982961, + "grad_norm": 2.2512192726135254, + "learning_rate": 1.1521803570202188e-06, + "loss": 0.6786, + "num_input_tokens_seen": 9327248, + "step": 34820 + }, + { + "epoch": 9.128440366972477, + "grad_norm": 2.9351489543914795, + "learning_rate": 1.1487507076120507e-06, + "loss": 0.5698, + "num_input_tokens_seen": 9328576, + "step": 34825 + }, + { + "epoch": 9.129750982961992, + "grad_norm": 2.47727108001709, + "learning_rate": 1.1453260502291507e-06, + "loss": 0.455, + "num_input_tokens_seen": 9329680, + "step": 34830 + }, + { + "epoch": 9.131061598951508, + "grad_norm": 2.2022204399108887, + "learning_rate": 1.14190638558829e-06, + "loss": 0.4729, + "num_input_tokens_seen": 9331120, + "step": 34835 + }, + { + "epoch": 9.132372214941022, + "grad_norm": 2.3094606399536133, + "learning_rate": 1.1384917144051987e-06, + "loss": 0.4314, + "num_input_tokens_seen": 9333008, + "step": 34840 + }, + { + "epoch": 9.133682830930537, + "grad_norm": 3.6016948223114014, + "learning_rate": 1.135082037394561e-06, + "loss": 0.4656, + "num_input_tokens_seen": 9334080, + "step": 34845 + }, + { + "epoch": 9.134993446920053, + "grad_norm": 3.745727300643921, + "learning_rate": 1.1316773552700193e-06, + "loss": 0.6025, + "num_input_tokens_seen": 9335344, + "step": 34850 + }, + { + "epoch": 9.136304062909568, + "grad_norm": 4.2073211669921875, + "learning_rate": 1.1282776687441626e-06, + "loss": 0.4081, + "num_input_tokens_seen": 9336448, + "step": 34855 + }, + { + "epoch": 9.137614678899082, + "grad_norm": 2.326551675796509, + "learning_rate": 1.1248829785285463e-06, + "loss": 0.29, + "num_input_tokens_seen": 9337520, + "step": 34860 + }, + { + "epoch": 9.138925294888598, + "grad_norm": 4.019786834716797, + "learning_rate": 1.121493285333669e-06, + "loss": 0.5047, + "num_input_tokens_seen": 9338736, + "step": 34865 + }, + { + "epoch": 9.140235910878113, + "grad_norm": 2.5481112003326416, + "learning_rate": 1.1181085898689881e-06, + "loss": 0.5045, + "num_input_tokens_seen": 9340608, + "step": 34870 + }, + { + "epoch": 9.141546526867629, + "grad_norm": 3.0751211643218994, + "learning_rate": 1.1147288928429118e-06, + "loss": 0.4552, + "num_input_tokens_seen": 9342480, + "step": 34875 + }, + { + "epoch": 9.142857142857142, + "grad_norm": 2.762625217437744, + "learning_rate": 1.1113541949628104e-06, + "loss": 0.3465, + "num_input_tokens_seen": 9344128, + "step": 34880 + }, + { + "epoch": 9.144167758846658, + "grad_norm": 2.2380712032318115, + "learning_rate": 1.1079844969350023e-06, + "loss": 0.3542, + "num_input_tokens_seen": 9345616, + "step": 34885 + }, + { + "epoch": 9.145478374836173, + "grad_norm": 6.116968154907227, + "learning_rate": 1.104619799464754e-06, + "loss": 0.3422, + "num_input_tokens_seen": 9346816, + "step": 34890 + }, + { + "epoch": 9.146788990825687, + "grad_norm": 4.02264404296875, + "learning_rate": 1.1012601032562935e-06, + "loss": 0.4683, + "num_input_tokens_seen": 9348224, + "step": 34895 + }, + { + "epoch": 9.148099606815203, + "grad_norm": 2.0465543270111084, + "learning_rate": 1.0979054090127978e-06, + "loss": 0.3794, + "num_input_tokens_seen": 9349632, + "step": 34900 + }, + { + "epoch": 9.149410222804718, + "grad_norm": 1.4238834381103516, + "learning_rate": 1.094555717436399e-06, + "loss": 0.4558, + "num_input_tokens_seen": 9350960, + "step": 34905 + }, + { + "epoch": 9.150720838794234, + "grad_norm": 2.612501621246338, + "learning_rate": 1.0912110292281868e-06, + "loss": 0.4668, + "num_input_tokens_seen": 9352048, + "step": 34910 + }, + { + "epoch": 9.152031454783748, + "grad_norm": 2.006617784500122, + "learning_rate": 1.0878713450881928e-06, + "loss": 0.6889, + "num_input_tokens_seen": 9353280, + "step": 34915 + }, + { + "epoch": 9.153342070773263, + "grad_norm": 2.419492244720459, + "learning_rate": 1.084536665715416e-06, + "loss": 0.295, + "num_input_tokens_seen": 9354656, + "step": 34920 + }, + { + "epoch": 9.154652686762779, + "grad_norm": 3.511359453201294, + "learning_rate": 1.0812069918077844e-06, + "loss": 0.4016, + "num_input_tokens_seen": 9355760, + "step": 34925 + }, + { + "epoch": 9.155963302752294, + "grad_norm": 1.415355920791626, + "learning_rate": 1.0778823240622043e-06, + "loss": 0.4797, + "num_input_tokens_seen": 9357040, + "step": 34930 + }, + { + "epoch": 9.157273918741808, + "grad_norm": 3.5567686557769775, + "learning_rate": 1.0745626631745193e-06, + "loss": 0.3613, + "num_input_tokens_seen": 9358128, + "step": 34935 + }, + { + "epoch": 9.158584534731324, + "grad_norm": 4.026987552642822, + "learning_rate": 1.0712480098395289e-06, + "loss": 0.4732, + "num_input_tokens_seen": 9359216, + "step": 34940 + }, + { + "epoch": 9.159895150720839, + "grad_norm": 10.25666332244873, + "learning_rate": 1.0679383647509839e-06, + "loss": 0.396, + "num_input_tokens_seen": 9360896, + "step": 34945 + }, + { + "epoch": 9.161205766710355, + "grad_norm": 4.345870494842529, + "learning_rate": 1.0646337286015856e-06, + "loss": 0.3902, + "num_input_tokens_seen": 9362144, + "step": 34950 + }, + { + "epoch": 9.162516382699868, + "grad_norm": 2.993691921234131, + "learning_rate": 1.0613341020829948e-06, + "loss": 0.5793, + "num_input_tokens_seen": 9363504, + "step": 34955 + }, + { + "epoch": 9.163826998689384, + "grad_norm": 2.111318349838257, + "learning_rate": 1.0580394858858117e-06, + "loss": 0.6026, + "num_input_tokens_seen": 9364544, + "step": 34960 + }, + { + "epoch": 9.1651376146789, + "grad_norm": 3.6088876724243164, + "learning_rate": 1.054749880699593e-06, + "loss": 0.5899, + "num_input_tokens_seen": 9365680, + "step": 34965 + }, + { + "epoch": 9.166448230668415, + "grad_norm": 5.313335418701172, + "learning_rate": 1.051465287212855e-06, + "loss": 0.5265, + "num_input_tokens_seen": 9366864, + "step": 34970 + }, + { + "epoch": 9.167758846657929, + "grad_norm": 6.877971172332764, + "learning_rate": 1.04818570611305e-06, + "loss": 0.526, + "num_input_tokens_seen": 9368144, + "step": 34975 + }, + { + "epoch": 9.169069462647444, + "grad_norm": 1.456050992012024, + "learning_rate": 1.0449111380865906e-06, + "loss": 0.4952, + "num_input_tokens_seen": 9369648, + "step": 34980 + }, + { + "epoch": 9.17038007863696, + "grad_norm": 3.751650094985962, + "learning_rate": 1.0416415838188393e-06, + "loss": 0.4264, + "num_input_tokens_seen": 9370608, + "step": 34985 + }, + { + "epoch": 9.171690694626474, + "grad_norm": 1.3114784955978394, + "learning_rate": 1.038377043994107e-06, + "loss": 0.306, + "num_input_tokens_seen": 9372288, + "step": 34990 + }, + { + "epoch": 9.17300131061599, + "grad_norm": 1.8512636423110962, + "learning_rate": 1.0351175192956585e-06, + "loss": 0.3164, + "num_input_tokens_seen": 9374016, + "step": 34995 + }, + { + "epoch": 9.174311926605505, + "grad_norm": 3.2444019317626953, + "learning_rate": 1.0318630104057037e-06, + "loss": 0.3977, + "num_input_tokens_seen": 9375184, + "step": 35000 + }, + { + "epoch": 9.17562254259502, + "grad_norm": 6.43087911605835, + "learning_rate": 1.0286135180054114e-06, + "loss": 0.4978, + "num_input_tokens_seen": 9376416, + "step": 35005 + }, + { + "epoch": 9.176933158584534, + "grad_norm": 7.541406154632568, + "learning_rate": 1.0253690427748874e-06, + "loss": 0.4864, + "num_input_tokens_seen": 9377472, + "step": 35010 + }, + { + "epoch": 9.17824377457405, + "grad_norm": 2.2140164375305176, + "learning_rate": 1.0221295853931973e-06, + "loss": 0.3964, + "num_input_tokens_seen": 9378912, + "step": 35015 + }, + { + "epoch": 9.179554390563565, + "grad_norm": 1.5528959035873413, + "learning_rate": 1.0188951465383511e-06, + "loss": 0.3525, + "num_input_tokens_seen": 9380432, + "step": 35020 + }, + { + "epoch": 9.18086500655308, + "grad_norm": 2.2917251586914062, + "learning_rate": 1.015665726887316e-06, + "loss": 0.5102, + "num_input_tokens_seen": 9382112, + "step": 35025 + }, + { + "epoch": 9.182175622542594, + "grad_norm": 2.5486671924591064, + "learning_rate": 1.0124413271160038e-06, + "loss": 0.4162, + "num_input_tokens_seen": 9383312, + "step": 35030 + }, + { + "epoch": 9.18348623853211, + "grad_norm": 14.282694816589355, + "learning_rate": 1.0092219478992776e-06, + "loss": 0.6404, + "num_input_tokens_seen": 9384288, + "step": 35035 + }, + { + "epoch": 9.184796854521625, + "grad_norm": 5.690136432647705, + "learning_rate": 1.0060075899109427e-06, + "loss": 0.6157, + "num_input_tokens_seen": 9385408, + "step": 35040 + }, + { + "epoch": 9.186107470511141, + "grad_norm": 1.5584126710891724, + "learning_rate": 1.0027982538237612e-06, + "loss": 0.5384, + "num_input_tokens_seen": 9386672, + "step": 35045 + }, + { + "epoch": 9.187418086500655, + "grad_norm": 2.1214911937713623, + "learning_rate": 9.995939403094402e-07, + "loss": 0.8441, + "num_input_tokens_seen": 9388192, + "step": 35050 + }, + { + "epoch": 9.18872870249017, + "grad_norm": 3.5522539615631104, + "learning_rate": 9.963946500386374e-07, + "loss": 0.5656, + "num_input_tokens_seen": 9389584, + "step": 35055 + }, + { + "epoch": 9.190039318479686, + "grad_norm": 2.5918936729431152, + "learning_rate": 9.932003836809616e-07, + "loss": 0.3924, + "num_input_tokens_seen": 9392768, + "step": 35060 + }, + { + "epoch": 9.191349934469201, + "grad_norm": 3.544966220855713, + "learning_rate": 9.900111419049619e-07, + "loss": 0.566, + "num_input_tokens_seen": 9394560, + "step": 35065 + }, + { + "epoch": 9.192660550458715, + "grad_norm": 3.1727099418640137, + "learning_rate": 9.868269253781453e-07, + "loss": 0.5304, + "num_input_tokens_seen": 9395824, + "step": 35070 + }, + { + "epoch": 9.19397116644823, + "grad_norm": 4.294496536254883, + "learning_rate": 9.836477347669626e-07, + "loss": 0.3923, + "num_input_tokens_seen": 9397248, + "step": 35075 + }, + { + "epoch": 9.195281782437746, + "grad_norm": 3.546719551086426, + "learning_rate": 9.804735707368118e-07, + "loss": 0.42, + "num_input_tokens_seen": 9398288, + "step": 35080 + }, + { + "epoch": 9.19659239842726, + "grad_norm": 2.6862306594848633, + "learning_rate": 9.773044339520392e-07, + "loss": 0.4939, + "num_input_tokens_seen": 9399840, + "step": 35085 + }, + { + "epoch": 9.197903014416775, + "grad_norm": 4.887294769287109, + "learning_rate": 9.741403250759424e-07, + "loss": 0.5339, + "num_input_tokens_seen": 9401008, + "step": 35090 + }, + { + "epoch": 9.199213630406291, + "grad_norm": 3.8725202083587646, + "learning_rate": 9.709812447707606e-07, + "loss": 0.462, + "num_input_tokens_seen": 9402192, + "step": 35095 + }, + { + "epoch": 9.200524246395807, + "grad_norm": 7.991739749908447, + "learning_rate": 9.678271936976847e-07, + "loss": 0.4786, + "num_input_tokens_seen": 9403328, + "step": 35100 + }, + { + "epoch": 9.20183486238532, + "grad_norm": 2.2972774505615234, + "learning_rate": 9.64678172516853e-07, + "loss": 0.3697, + "num_input_tokens_seen": 9406976, + "step": 35105 + }, + { + "epoch": 9.203145478374836, + "grad_norm": 2.5907769203186035, + "learning_rate": 9.615341818873496e-07, + "loss": 0.4445, + "num_input_tokens_seen": 9408384, + "step": 35110 + }, + { + "epoch": 9.204456094364351, + "grad_norm": 6.6918864250183105, + "learning_rate": 9.583952224672094e-07, + "loss": 0.5002, + "num_input_tokens_seen": 9409376, + "step": 35115 + }, + { + "epoch": 9.205766710353867, + "grad_norm": 3.92692232131958, + "learning_rate": 9.552612949134094e-07, + "loss": 0.4172, + "num_input_tokens_seen": 9410640, + "step": 35120 + }, + { + "epoch": 9.20707732634338, + "grad_norm": 2.4216325283050537, + "learning_rate": 9.521323998818694e-07, + "loss": 0.6242, + "num_input_tokens_seen": 9411776, + "step": 35125 + }, + { + "epoch": 9.208387942332896, + "grad_norm": 3.3646392822265625, + "learning_rate": 9.490085380274682e-07, + "loss": 0.6537, + "num_input_tokens_seen": 9413456, + "step": 35130 + }, + { + "epoch": 9.209698558322412, + "grad_norm": 1.5261340141296387, + "learning_rate": 9.45889710004022e-07, + "loss": 0.4017, + "num_input_tokens_seen": 9414736, + "step": 35135 + }, + { + "epoch": 9.211009174311927, + "grad_norm": 2.9677622318267822, + "learning_rate": 9.427759164642974e-07, + "loss": 0.3683, + "num_input_tokens_seen": 9416112, + "step": 35140 + }, + { + "epoch": 9.212319790301441, + "grad_norm": 1.7511866092681885, + "learning_rate": 9.396671580600064e-07, + "loss": 0.5064, + "num_input_tokens_seen": 9417744, + "step": 35145 + }, + { + "epoch": 9.213630406290957, + "grad_norm": 3.205085515975952, + "learning_rate": 9.36563435441809e-07, + "loss": 0.4885, + "num_input_tokens_seen": 9418720, + "step": 35150 + }, + { + "epoch": 9.214941022280472, + "grad_norm": 11.002187728881836, + "learning_rate": 9.334647492593051e-07, + "loss": 0.669, + "num_input_tokens_seen": 9419936, + "step": 35155 + }, + { + "epoch": 9.216251638269988, + "grad_norm": 27.945463180541992, + "learning_rate": 9.303711001610454e-07, + "loss": 0.487, + "num_input_tokens_seen": 9421184, + "step": 35160 + }, + { + "epoch": 9.217562254259501, + "grad_norm": 21.658267974853516, + "learning_rate": 9.27282488794523e-07, + "loss": 0.415, + "num_input_tokens_seen": 9422176, + "step": 35165 + }, + { + "epoch": 9.218872870249017, + "grad_norm": 5.738841533660889, + "learning_rate": 9.24198915806182e-07, + "loss": 0.4049, + "num_input_tokens_seen": 9423280, + "step": 35170 + }, + { + "epoch": 9.220183486238533, + "grad_norm": 8.644062995910645, + "learning_rate": 9.211203818414088e-07, + "loss": 0.5129, + "num_input_tokens_seen": 9424304, + "step": 35175 + }, + { + "epoch": 9.221494102228046, + "grad_norm": 8.929327011108398, + "learning_rate": 9.180468875445352e-07, + "loss": 0.3913, + "num_input_tokens_seen": 9425712, + "step": 35180 + }, + { + "epoch": 9.222804718217562, + "grad_norm": 3.359752893447876, + "learning_rate": 9.149784335588357e-07, + "loss": 0.5395, + "num_input_tokens_seen": 9426784, + "step": 35185 + }, + { + "epoch": 9.224115334207077, + "grad_norm": 4.203570365905762, + "learning_rate": 9.119150205265326e-07, + "loss": 0.4816, + "num_input_tokens_seen": 9428416, + "step": 35190 + }, + { + "epoch": 9.225425950196593, + "grad_norm": 1.9474326372146606, + "learning_rate": 9.088566490888017e-07, + "loss": 0.4394, + "num_input_tokens_seen": 9429744, + "step": 35195 + }, + { + "epoch": 9.226736566186107, + "grad_norm": 2.1557109355926514, + "learning_rate": 9.058033198857424e-07, + "loss": 0.3867, + "num_input_tokens_seen": 9430944, + "step": 35200 + }, + { + "epoch": 9.228047182175622, + "grad_norm": 16.941068649291992, + "learning_rate": 9.027550335564184e-07, + "loss": 0.597, + "num_input_tokens_seen": 9432320, + "step": 35205 + }, + { + "epoch": 9.229357798165138, + "grad_norm": 3.0376346111297607, + "learning_rate": 8.997117907388275e-07, + "loss": 0.3235, + "num_input_tokens_seen": 9433712, + "step": 35210 + }, + { + "epoch": 9.230668414154653, + "grad_norm": 5.669178485870361, + "learning_rate": 8.966735920699187e-07, + "loss": 0.7923, + "num_input_tokens_seen": 9434848, + "step": 35215 + }, + { + "epoch": 9.231979030144167, + "grad_norm": 3.45344877243042, + "learning_rate": 8.936404381855806e-07, + "loss": 0.2755, + "num_input_tokens_seen": 9435952, + "step": 35220 + }, + { + "epoch": 9.233289646133683, + "grad_norm": 2.912449359893799, + "learning_rate": 8.906123297206442e-07, + "loss": 0.4929, + "num_input_tokens_seen": 9437296, + "step": 35225 + }, + { + "epoch": 9.234600262123198, + "grad_norm": 3.545485019683838, + "learning_rate": 8.875892673088887e-07, + "loss": 0.3005, + "num_input_tokens_seen": 9438224, + "step": 35230 + }, + { + "epoch": 9.235910878112714, + "grad_norm": 2.0517096519470215, + "learning_rate": 8.845712515830385e-07, + "loss": 0.4757, + "num_input_tokens_seen": 9439280, + "step": 35235 + }, + { + "epoch": 9.237221494102227, + "grad_norm": 13.446227073669434, + "learning_rate": 8.815582831747576e-07, + "loss": 0.5462, + "num_input_tokens_seen": 9440352, + "step": 35240 + }, + { + "epoch": 9.238532110091743, + "grad_norm": 4.3488664627075195, + "learning_rate": 8.785503627146501e-07, + "loss": 0.3748, + "num_input_tokens_seen": 9441680, + "step": 35245 + }, + { + "epoch": 9.239842726081259, + "grad_norm": 8.42081069946289, + "learning_rate": 8.755474908322764e-07, + "loss": 0.4894, + "num_input_tokens_seen": 9442912, + "step": 35250 + }, + { + "epoch": 9.241153342070774, + "grad_norm": 3.362757444381714, + "learning_rate": 8.725496681561279e-07, + "loss": 0.3982, + "num_input_tokens_seen": 9444080, + "step": 35255 + }, + { + "epoch": 9.242463958060288, + "grad_norm": 2.7450828552246094, + "learning_rate": 8.695568953136446e-07, + "loss": 0.4247, + "num_input_tokens_seen": 9445568, + "step": 35260 + }, + { + "epoch": 9.243774574049803, + "grad_norm": 3.481066942214966, + "learning_rate": 8.665691729312115e-07, + "loss": 0.4618, + "num_input_tokens_seen": 9447424, + "step": 35265 + }, + { + "epoch": 9.245085190039319, + "grad_norm": 7.6015944480896, + "learning_rate": 8.635865016341477e-07, + "loss": 0.6035, + "num_input_tokens_seen": 9449232, + "step": 35270 + }, + { + "epoch": 9.246395806028833, + "grad_norm": 9.819400787353516, + "learning_rate": 8.606088820467318e-07, + "loss": 0.5253, + "num_input_tokens_seen": 9450800, + "step": 35275 + }, + { + "epoch": 9.247706422018348, + "grad_norm": 2.22330904006958, + "learning_rate": 8.576363147921596e-07, + "loss": 0.371, + "num_input_tokens_seen": 9451904, + "step": 35280 + }, + { + "epoch": 9.249017038007864, + "grad_norm": 2.609114170074463, + "learning_rate": 8.546688004925946e-07, + "loss": 0.5565, + "num_input_tokens_seen": 9452960, + "step": 35285 + }, + { + "epoch": 9.25032765399738, + "grad_norm": 6.393649101257324, + "learning_rate": 8.517063397691288e-07, + "loss": 0.4251, + "num_input_tokens_seen": 9454352, + "step": 35290 + }, + { + "epoch": 9.251638269986893, + "grad_norm": 1.9047974348068237, + "learning_rate": 8.487489332418025e-07, + "loss": 0.3926, + "num_input_tokens_seen": 9455824, + "step": 35295 + }, + { + "epoch": 9.252948885976409, + "grad_norm": 1.6647944450378418, + "learning_rate": 8.457965815295926e-07, + "loss": 0.5801, + "num_input_tokens_seen": 9457712, + "step": 35300 + }, + { + "epoch": 9.254259501965924, + "grad_norm": 1.4253937005996704, + "learning_rate": 8.428492852504216e-07, + "loss": 0.3826, + "num_input_tokens_seen": 9458832, + "step": 35305 + }, + { + "epoch": 9.25557011795544, + "grad_norm": 1.4716923236846924, + "learning_rate": 8.399070450211599e-07, + "loss": 0.4913, + "num_input_tokens_seen": 9460160, + "step": 35310 + }, + { + "epoch": 9.256880733944953, + "grad_norm": 2.26806378364563, + "learning_rate": 8.369698614576038e-07, + "loss": 0.4198, + "num_input_tokens_seen": 9461472, + "step": 35315 + }, + { + "epoch": 9.258191349934469, + "grad_norm": 2.691068172454834, + "learning_rate": 8.340377351745088e-07, + "loss": 0.5688, + "num_input_tokens_seen": 9462976, + "step": 35320 + }, + { + "epoch": 9.259501965923985, + "grad_norm": 6.833580493927002, + "learning_rate": 8.311106667855562e-07, + "loss": 0.5284, + "num_input_tokens_seen": 9464416, + "step": 35325 + }, + { + "epoch": 9.2608125819135, + "grad_norm": 6.21604585647583, + "learning_rate": 8.281886569033836e-07, + "loss": 0.383, + "num_input_tokens_seen": 9465488, + "step": 35330 + }, + { + "epoch": 9.262123197903014, + "grad_norm": 3.2177023887634277, + "learning_rate": 8.252717061395576e-07, + "loss": 0.4911, + "num_input_tokens_seen": 9466624, + "step": 35335 + }, + { + "epoch": 9.26343381389253, + "grad_norm": 3.6092751026153564, + "learning_rate": 8.223598151045952e-07, + "loss": 0.6643, + "num_input_tokens_seen": 9468016, + "step": 35340 + }, + { + "epoch": 9.264744429882045, + "grad_norm": 1.0986213684082031, + "learning_rate": 8.194529844079451e-07, + "loss": 0.3338, + "num_input_tokens_seen": 9469776, + "step": 35345 + }, + { + "epoch": 9.26605504587156, + "grad_norm": 5.623603820800781, + "learning_rate": 8.16551214658004e-07, + "loss": 0.6079, + "num_input_tokens_seen": 9470880, + "step": 35350 + }, + { + "epoch": 9.267365661861074, + "grad_norm": 3.8224456310272217, + "learning_rate": 8.136545064621137e-07, + "loss": 0.5638, + "num_input_tokens_seen": 9472304, + "step": 35355 + }, + { + "epoch": 9.26867627785059, + "grad_norm": 6.686797618865967, + "learning_rate": 8.107628604265366e-07, + "loss": 0.4307, + "num_input_tokens_seen": 9473728, + "step": 35360 + }, + { + "epoch": 9.269986893840105, + "grad_norm": 5.736859321594238, + "learning_rate": 8.078762771564996e-07, + "loss": 0.5548, + "num_input_tokens_seen": 9474800, + "step": 35365 + }, + { + "epoch": 9.271297509829619, + "grad_norm": 1.1811884641647339, + "learning_rate": 8.049947572561584e-07, + "loss": 0.3002, + "num_input_tokens_seen": 9476128, + "step": 35370 + }, + { + "epoch": 9.272608125819135, + "grad_norm": 4.530514717102051, + "learning_rate": 8.021183013286082e-07, + "loss": 0.3901, + "num_input_tokens_seen": 9477632, + "step": 35375 + }, + { + "epoch": 9.27391874180865, + "grad_norm": 1.9962141513824463, + "learning_rate": 7.99246909975887e-07, + "loss": 0.7959, + "num_input_tokens_seen": 9478800, + "step": 35380 + }, + { + "epoch": 9.275229357798166, + "grad_norm": 2.3386595249176025, + "learning_rate": 7.963805837989724e-07, + "loss": 0.3329, + "num_input_tokens_seen": 9480416, + "step": 35385 + }, + { + "epoch": 9.27653997378768, + "grad_norm": 2.870232582092285, + "learning_rate": 7.935193233977845e-07, + "loss": 0.4208, + "num_input_tokens_seen": 9481504, + "step": 35390 + }, + { + "epoch": 9.277850589777195, + "grad_norm": 3.3847713470458984, + "learning_rate": 7.906631293711719e-07, + "loss": 0.444, + "num_input_tokens_seen": 9482864, + "step": 35395 + }, + { + "epoch": 9.27916120576671, + "grad_norm": 2.014010429382324, + "learning_rate": 7.878120023169344e-07, + "loss": 0.4982, + "num_input_tokens_seen": 9484320, + "step": 35400 + }, + { + "epoch": 9.280471821756226, + "grad_norm": 1.6614536046981812, + "learning_rate": 7.849659428318113e-07, + "loss": 0.6436, + "num_input_tokens_seen": 9485968, + "step": 35405 + }, + { + "epoch": 9.28178243774574, + "grad_norm": 1.6443352699279785, + "learning_rate": 7.821249515114787e-07, + "loss": 0.3662, + "num_input_tokens_seen": 9487680, + "step": 35410 + }, + { + "epoch": 9.283093053735255, + "grad_norm": 3.0807738304138184, + "learning_rate": 7.792890289505444e-07, + "loss": 0.5149, + "num_input_tokens_seen": 9489120, + "step": 35415 + }, + { + "epoch": 9.284403669724771, + "grad_norm": 2.675462245941162, + "learning_rate": 7.764581757425642e-07, + "loss": 0.4281, + "num_input_tokens_seen": 9490976, + "step": 35420 + }, + { + "epoch": 9.285714285714286, + "grad_norm": 3.357726573944092, + "learning_rate": 7.736323924800365e-07, + "loss": 0.4544, + "num_input_tokens_seen": 9492544, + "step": 35425 + }, + { + "epoch": 9.2870249017038, + "grad_norm": 4.6731157302856445, + "learning_rate": 7.708116797543907e-07, + "loss": 0.6249, + "num_input_tokens_seen": 9493712, + "step": 35430 + }, + { + "epoch": 9.288335517693316, + "grad_norm": 3.041100263595581, + "learning_rate": 7.679960381559936e-07, + "loss": 0.4906, + "num_input_tokens_seen": 9495344, + "step": 35435 + }, + { + "epoch": 9.289646133682831, + "grad_norm": 1.5485293865203857, + "learning_rate": 7.651854682741572e-07, + "loss": 0.5812, + "num_input_tokens_seen": 9496768, + "step": 35440 + }, + { + "epoch": 9.290956749672347, + "grad_norm": 8.279304504394531, + "learning_rate": 7.623799706971274e-07, + "loss": 0.207, + "num_input_tokens_seen": 9497840, + "step": 35445 + }, + { + "epoch": 9.29226736566186, + "grad_norm": 2.6507296562194824, + "learning_rate": 7.595795460120903e-07, + "loss": 0.4319, + "num_input_tokens_seen": 9499200, + "step": 35450 + }, + { + "epoch": 9.293577981651376, + "grad_norm": 7.340606212615967, + "learning_rate": 7.567841948051685e-07, + "loss": 0.5874, + "num_input_tokens_seen": 9500096, + "step": 35455 + }, + { + "epoch": 9.294888597640892, + "grad_norm": 1.057113528251648, + "learning_rate": 7.539939176614247e-07, + "loss": 0.5056, + "num_input_tokens_seen": 9501584, + "step": 35460 + }, + { + "epoch": 9.296199213630405, + "grad_norm": 3.8686647415161133, + "learning_rate": 7.512087151648639e-07, + "loss": 0.4502, + "num_input_tokens_seen": 9502640, + "step": 35465 + }, + { + "epoch": 9.297509829619921, + "grad_norm": 3.3599069118499756, + "learning_rate": 7.484285878984171e-07, + "loss": 0.6409, + "num_input_tokens_seen": 9504208, + "step": 35470 + }, + { + "epoch": 9.298820445609437, + "grad_norm": 3.6627285480499268, + "learning_rate": 7.45653536443966e-07, + "loss": 0.3658, + "num_input_tokens_seen": 9505392, + "step": 35475 + }, + { + "epoch": 9.300131061598952, + "grad_norm": 14.464285850524902, + "learning_rate": 7.428835613823182e-07, + "loss": 0.5416, + "num_input_tokens_seen": 9506944, + "step": 35480 + }, + { + "epoch": 9.301441677588466, + "grad_norm": 2.5077621936798096, + "learning_rate": 7.401186632932294e-07, + "loss": 0.7647, + "num_input_tokens_seen": 9508336, + "step": 35485 + }, + { + "epoch": 9.302752293577981, + "grad_norm": 7.809906482696533, + "learning_rate": 7.373588427553869e-07, + "loss": 0.4743, + "num_input_tokens_seen": 9509616, + "step": 35490 + }, + { + "epoch": 9.304062909567497, + "grad_norm": 3.426680088043213, + "learning_rate": 7.346041003464172e-07, + "loss": 0.2327, + "num_input_tokens_seen": 9510784, + "step": 35495 + }, + { + "epoch": 9.305373525557012, + "grad_norm": 3.3497557640075684, + "learning_rate": 7.318544366428814e-07, + "loss": 0.482, + "num_input_tokens_seen": 9512256, + "step": 35500 + }, + { + "epoch": 9.306684141546526, + "grad_norm": 1.677761197090149, + "learning_rate": 7.291098522202777e-07, + "loss": 0.4367, + "num_input_tokens_seen": 9514096, + "step": 35505 + }, + { + "epoch": 9.307994757536042, + "grad_norm": 2.2305545806884766, + "learning_rate": 7.263703476530492e-07, + "loss": 0.2595, + "num_input_tokens_seen": 9515296, + "step": 35510 + }, + { + "epoch": 9.309305373525557, + "grad_norm": 1.7186979055404663, + "learning_rate": 7.236359235145624e-07, + "loss": 0.5828, + "num_input_tokens_seen": 9516512, + "step": 35515 + }, + { + "epoch": 9.310615989515073, + "grad_norm": 1.8311760425567627, + "learning_rate": 7.209065803771315e-07, + "loss": 0.4883, + "num_input_tokens_seen": 9518144, + "step": 35520 + }, + { + "epoch": 9.311926605504587, + "grad_norm": 2.7368593215942383, + "learning_rate": 7.181823188120024e-07, + "loss": 0.5533, + "num_input_tokens_seen": 9519456, + "step": 35525 + }, + { + "epoch": 9.313237221494102, + "grad_norm": 2.85176157951355, + "learning_rate": 7.154631393893552e-07, + "loss": 0.4252, + "num_input_tokens_seen": 9520752, + "step": 35530 + }, + { + "epoch": 9.314547837483618, + "grad_norm": 3.244013547897339, + "learning_rate": 7.127490426783123e-07, + "loss": 0.358, + "num_input_tokens_seen": 9522128, + "step": 35535 + }, + { + "epoch": 9.315858453473133, + "grad_norm": 4.479596138000488, + "learning_rate": 7.100400292469333e-07, + "loss": 0.3969, + "num_input_tokens_seen": 9523536, + "step": 35540 + }, + { + "epoch": 9.317169069462647, + "grad_norm": 2.8483195304870605, + "learning_rate": 7.073360996622064e-07, + "loss": 0.3553, + "num_input_tokens_seen": 9524736, + "step": 35545 + }, + { + "epoch": 9.318479685452163, + "grad_norm": 1.6064356565475464, + "learning_rate": 7.046372544900592e-07, + "loss": 0.3419, + "num_input_tokens_seen": 9526288, + "step": 35550 + }, + { + "epoch": 9.319790301441678, + "grad_norm": 2.7354543209075928, + "learning_rate": 7.01943494295354e-07, + "loss": 0.4732, + "num_input_tokens_seen": 9527520, + "step": 35555 + }, + { + "epoch": 9.321100917431192, + "grad_norm": 4.259365081787109, + "learning_rate": 6.992548196418924e-07, + "loss": 0.4535, + "num_input_tokens_seen": 9528720, + "step": 35560 + }, + { + "epoch": 9.322411533420707, + "grad_norm": 2.4613196849823, + "learning_rate": 6.965712310924078e-07, + "loss": 0.4602, + "num_input_tokens_seen": 9529936, + "step": 35565 + }, + { + "epoch": 9.323722149410223, + "grad_norm": 3.3839659690856934, + "learning_rate": 6.9389272920857e-07, + "loss": 0.3315, + "num_input_tokens_seen": 9531264, + "step": 35570 + }, + { + "epoch": 9.325032765399738, + "grad_norm": 1.883752703666687, + "learning_rate": 6.912193145509893e-07, + "loss": 0.6194, + "num_input_tokens_seen": 9532640, + "step": 35575 + }, + { + "epoch": 9.326343381389252, + "grad_norm": 2.831390380859375, + "learning_rate": 6.885509876792012e-07, + "loss": 0.5353, + "num_input_tokens_seen": 9533728, + "step": 35580 + }, + { + "epoch": 9.327653997378768, + "grad_norm": 1.9143990278244019, + "learning_rate": 6.858877491516869e-07, + "loss": 0.5665, + "num_input_tokens_seen": 9535040, + "step": 35585 + }, + { + "epoch": 9.328964613368283, + "grad_norm": 3.348304033279419, + "learning_rate": 6.832295995258531e-07, + "loss": 0.8273, + "num_input_tokens_seen": 9536064, + "step": 35590 + }, + { + "epoch": 9.330275229357799, + "grad_norm": 2.3617217540740967, + "learning_rate": 6.805765393580493e-07, + "loss": 0.3641, + "num_input_tokens_seen": 9537344, + "step": 35595 + }, + { + "epoch": 9.331585845347313, + "grad_norm": 2.1379098892211914, + "learning_rate": 6.779285692035536e-07, + "loss": 0.3648, + "num_input_tokens_seen": 9538672, + "step": 35600 + }, + { + "epoch": 9.332896461336828, + "grad_norm": 9.426433563232422, + "learning_rate": 6.752856896165866e-07, + "loss": 0.599, + "num_input_tokens_seen": 9539616, + "step": 35605 + }, + { + "epoch": 9.334207077326344, + "grad_norm": 13.843708038330078, + "learning_rate": 6.726479011502917e-07, + "loss": 0.491, + "num_input_tokens_seen": 9541328, + "step": 35610 + }, + { + "epoch": 9.33551769331586, + "grad_norm": 3.368746042251587, + "learning_rate": 6.700152043567554e-07, + "loss": 0.4305, + "num_input_tokens_seen": 9542704, + "step": 35615 + }, + { + "epoch": 9.336828309305373, + "grad_norm": 3.0319056510925293, + "learning_rate": 6.673875997870005e-07, + "loss": 0.3891, + "num_input_tokens_seen": 9544288, + "step": 35620 + }, + { + "epoch": 9.338138925294889, + "grad_norm": 6.642776966094971, + "learning_rate": 6.64765087990979e-07, + "loss": 0.4751, + "num_input_tokens_seen": 9545392, + "step": 35625 + }, + { + "epoch": 9.339449541284404, + "grad_norm": 2.5887467861175537, + "learning_rate": 6.621476695175738e-07, + "loss": 0.3795, + "num_input_tokens_seen": 9547088, + "step": 35630 + }, + { + "epoch": 9.34076015727392, + "grad_norm": 2.023913621902466, + "learning_rate": 6.595353449146108e-07, + "loss": 0.5107, + "num_input_tokens_seen": 9548576, + "step": 35635 + }, + { + "epoch": 9.342070773263433, + "grad_norm": 1.5753891468048096, + "learning_rate": 6.569281147288414e-07, + "loss": 0.4545, + "num_input_tokens_seen": 9550032, + "step": 35640 + }, + { + "epoch": 9.343381389252949, + "grad_norm": 4.623026371002197, + "learning_rate": 6.54325979505957e-07, + "loss": 0.5102, + "num_input_tokens_seen": 9551216, + "step": 35645 + }, + { + "epoch": 9.344692005242464, + "grad_norm": 2.9154560565948486, + "learning_rate": 6.517289397905746e-07, + "loss": 0.3213, + "num_input_tokens_seen": 9552320, + "step": 35650 + }, + { + "epoch": 9.346002621231978, + "grad_norm": 4.704843521118164, + "learning_rate": 6.491369961262566e-07, + "loss": 0.548, + "num_input_tokens_seen": 9553376, + "step": 35655 + }, + { + "epoch": 9.347313237221494, + "grad_norm": 3.654385566711426, + "learning_rate": 6.465501490554914e-07, + "loss": 0.3906, + "num_input_tokens_seen": 9554384, + "step": 35660 + }, + { + "epoch": 9.34862385321101, + "grad_norm": 3.093122720718384, + "learning_rate": 6.439683991197043e-07, + "loss": 0.5217, + "num_input_tokens_seen": 9555648, + "step": 35665 + }, + { + "epoch": 9.349934469200525, + "grad_norm": 2.702704668045044, + "learning_rate": 6.413917468592434e-07, + "loss": 0.508, + "num_input_tokens_seen": 9556864, + "step": 35670 + }, + { + "epoch": 9.351245085190039, + "grad_norm": 2.1530051231384277, + "learning_rate": 6.388201928134025e-07, + "loss": 0.4748, + "num_input_tokens_seen": 9558144, + "step": 35675 + }, + { + "epoch": 9.352555701179554, + "grad_norm": 4.881879806518555, + "learning_rate": 6.362537375204009e-07, + "loss": 0.5009, + "num_input_tokens_seen": 9559664, + "step": 35680 + }, + { + "epoch": 9.35386631716907, + "grad_norm": 1.4735699892044067, + "learning_rate": 6.336923815173951e-07, + "loss": 0.6738, + "num_input_tokens_seen": 9561200, + "step": 35685 + }, + { + "epoch": 9.355176933158585, + "grad_norm": 2.1862802505493164, + "learning_rate": 6.31136125340473e-07, + "loss": 0.4135, + "num_input_tokens_seen": 9562672, + "step": 35690 + }, + { + "epoch": 9.356487549148099, + "grad_norm": 1.4744738340377808, + "learning_rate": 6.285849695246537e-07, + "loss": 0.4772, + "num_input_tokens_seen": 9564528, + "step": 35695 + }, + { + "epoch": 9.357798165137615, + "grad_norm": 4.066432952880859, + "learning_rate": 6.260389146038881e-07, + "loss": 0.6234, + "num_input_tokens_seen": 9566032, + "step": 35700 + }, + { + "epoch": 9.35910878112713, + "grad_norm": 6.81235933303833, + "learning_rate": 6.234979611110637e-07, + "loss": 0.3902, + "num_input_tokens_seen": 9567024, + "step": 35705 + }, + { + "epoch": 9.360419397116646, + "grad_norm": 1.6074835062026978, + "learning_rate": 6.209621095779993e-07, + "loss": 0.4173, + "num_input_tokens_seen": 9568448, + "step": 35710 + }, + { + "epoch": 9.36173001310616, + "grad_norm": 2.3601043224334717, + "learning_rate": 6.184313605354375e-07, + "loss": 0.5058, + "num_input_tokens_seen": 9569776, + "step": 35715 + }, + { + "epoch": 9.363040629095675, + "grad_norm": 8.244030952453613, + "learning_rate": 6.159057145130653e-07, + "loss": 0.5728, + "num_input_tokens_seen": 9570944, + "step": 35720 + }, + { + "epoch": 9.36435124508519, + "grad_norm": 3.513824701309204, + "learning_rate": 6.133851720394962e-07, + "loss": 0.4502, + "num_input_tokens_seen": 9572208, + "step": 35725 + }, + { + "epoch": 9.365661861074706, + "grad_norm": 4.065837383270264, + "learning_rate": 6.10869733642272e-07, + "loss": 0.5791, + "num_input_tokens_seen": 9573392, + "step": 35730 + }, + { + "epoch": 9.36697247706422, + "grad_norm": 4.627383708953857, + "learning_rate": 6.083593998478715e-07, + "loss": 0.6777, + "num_input_tokens_seen": 9574480, + "step": 35735 + }, + { + "epoch": 9.368283093053735, + "grad_norm": 2.8408589363098145, + "learning_rate": 6.058541711817023e-07, + "loss": 0.539, + "num_input_tokens_seen": 9575584, + "step": 35740 + }, + { + "epoch": 9.36959370904325, + "grad_norm": 1.297999382019043, + "learning_rate": 6.033540481681088e-07, + "loss": 0.3919, + "num_input_tokens_seen": 9576800, + "step": 35745 + }, + { + "epoch": 9.370904325032765, + "grad_norm": 2.2314677238464355, + "learning_rate": 6.008590313303586e-07, + "loss": 0.4946, + "num_input_tokens_seen": 9578256, + "step": 35750 + }, + { + "epoch": 9.37221494102228, + "grad_norm": 5.6034979820251465, + "learning_rate": 5.983691211906506e-07, + "loss": 0.3439, + "num_input_tokens_seen": 9580816, + "step": 35755 + }, + { + "epoch": 9.373525557011796, + "grad_norm": 2.22308349609375, + "learning_rate": 5.958843182701263e-07, + "loss": 0.4927, + "num_input_tokens_seen": 9582416, + "step": 35760 + }, + { + "epoch": 9.374836173001311, + "grad_norm": 3.109503746032715, + "learning_rate": 5.934046230888418e-07, + "loss": 0.6583, + "num_input_tokens_seen": 9584048, + "step": 35765 + }, + { + "epoch": 9.376146788990825, + "grad_norm": 2.7540245056152344, + "learning_rate": 5.909300361658015e-07, + "loss": 0.3591, + "num_input_tokens_seen": 9585136, + "step": 35770 + }, + { + "epoch": 9.37745740498034, + "grad_norm": 2.725062131881714, + "learning_rate": 5.8846055801893e-07, + "loss": 0.569, + "num_input_tokens_seen": 9586224, + "step": 35775 + }, + { + "epoch": 9.378768020969856, + "grad_norm": 1.3777809143066406, + "learning_rate": 5.859961891650861e-07, + "loss": 0.3623, + "num_input_tokens_seen": 9587648, + "step": 35780 + }, + { + "epoch": 9.380078636959372, + "grad_norm": 2.4590084552764893, + "learning_rate": 5.835369301200516e-07, + "loss": 0.3217, + "num_input_tokens_seen": 9589072, + "step": 35785 + }, + { + "epoch": 9.381389252948885, + "grad_norm": 3.5999860763549805, + "learning_rate": 5.810827813985509e-07, + "loss": 0.3637, + "num_input_tokens_seen": 9590336, + "step": 35790 + }, + { + "epoch": 9.382699868938401, + "grad_norm": 4.719509601593018, + "learning_rate": 5.786337435142314e-07, + "loss": 0.472, + "num_input_tokens_seen": 9591728, + "step": 35795 + }, + { + "epoch": 9.384010484927916, + "grad_norm": 11.526660919189453, + "learning_rate": 5.761898169796721e-07, + "loss": 0.3578, + "num_input_tokens_seen": 9592832, + "step": 35800 + }, + { + "epoch": 9.385321100917432, + "grad_norm": 3.820835828781128, + "learning_rate": 5.73751002306383e-07, + "loss": 0.4605, + "num_input_tokens_seen": 9594112, + "step": 35805 + }, + { + "epoch": 9.386631716906946, + "grad_norm": 3.5377235412597656, + "learning_rate": 5.713173000048033e-07, + "loss": 0.4657, + "num_input_tokens_seen": 9595296, + "step": 35810 + }, + { + "epoch": 9.387942332896461, + "grad_norm": 1.0267821550369263, + "learning_rate": 5.688887105843032e-07, + "loss": 0.3229, + "num_input_tokens_seen": 9596784, + "step": 35815 + }, + { + "epoch": 9.389252948885977, + "grad_norm": 3.187756061553955, + "learning_rate": 5.664652345531846e-07, + "loss": 0.5576, + "num_input_tokens_seen": 9598240, + "step": 35820 + }, + { + "epoch": 9.390563564875492, + "grad_norm": 2.8306872844696045, + "learning_rate": 5.640468724186721e-07, + "loss": 0.4824, + "num_input_tokens_seen": 9599376, + "step": 35825 + }, + { + "epoch": 9.391874180865006, + "grad_norm": 2.3861544132232666, + "learning_rate": 5.616336246869275e-07, + "loss": 0.3368, + "num_input_tokens_seen": 9600656, + "step": 35830 + }, + { + "epoch": 9.393184796854522, + "grad_norm": 3.3852732181549072, + "learning_rate": 5.592254918630385e-07, + "loss": 0.4997, + "num_input_tokens_seen": 9602256, + "step": 35835 + }, + { + "epoch": 9.394495412844037, + "grad_norm": 5.011051177978516, + "learning_rate": 5.568224744510242e-07, + "loss": 0.3361, + "num_input_tokens_seen": 9603552, + "step": 35840 + }, + { + "epoch": 9.395806028833551, + "grad_norm": 4.976907253265381, + "learning_rate": 5.544245729538322e-07, + "loss": 0.2179, + "num_input_tokens_seen": 9604816, + "step": 35845 + }, + { + "epoch": 9.397116644823067, + "grad_norm": 3.119114875793457, + "learning_rate": 5.520317878733389e-07, + "loss": 0.5035, + "num_input_tokens_seen": 9605840, + "step": 35850 + }, + { + "epoch": 9.398427260812582, + "grad_norm": 4.343940734863281, + "learning_rate": 5.496441197103463e-07, + "loss": 0.4253, + "num_input_tokens_seen": 9607040, + "step": 35855 + }, + { + "epoch": 9.399737876802098, + "grad_norm": 10.998224258422852, + "learning_rate": 5.472615689645965e-07, + "loss": 0.4932, + "num_input_tokens_seen": 9608112, + "step": 35860 + }, + { + "epoch": 9.401048492791611, + "grad_norm": 10.299112319946289, + "learning_rate": 5.44884136134749e-07, + "loss": 0.6399, + "num_input_tokens_seen": 9609536, + "step": 35865 + }, + { + "epoch": 9.402359108781127, + "grad_norm": 5.018993854522705, + "learning_rate": 5.425118217183945e-07, + "loss": 0.374, + "num_input_tokens_seen": 9610976, + "step": 35870 + }, + { + "epoch": 9.403669724770642, + "grad_norm": 3.9586081504821777, + "learning_rate": 5.40144626212058e-07, + "loss": 0.3847, + "num_input_tokens_seen": 9612304, + "step": 35875 + }, + { + "epoch": 9.404980340760158, + "grad_norm": 2.474353075027466, + "learning_rate": 5.377825501111849e-07, + "loss": 0.3949, + "num_input_tokens_seen": 9614000, + "step": 35880 + }, + { + "epoch": 9.406290956749672, + "grad_norm": 1.1504909992218018, + "learning_rate": 5.354255939101576e-07, + "loss": 0.4152, + "num_input_tokens_seen": 9615440, + "step": 35885 + }, + { + "epoch": 9.407601572739187, + "grad_norm": 4.791656970977783, + "learning_rate": 5.330737581022843e-07, + "loss": 0.4191, + "num_input_tokens_seen": 9616544, + "step": 35890 + }, + { + "epoch": 9.408912188728703, + "grad_norm": 2.079939842224121, + "learning_rate": 5.307270431797962e-07, + "loss": 0.35, + "num_input_tokens_seen": 9617680, + "step": 35895 + }, + { + "epoch": 9.410222804718218, + "grad_norm": 2.751727342605591, + "learning_rate": 5.283854496338586e-07, + "loss": 0.628, + "num_input_tokens_seen": 9619008, + "step": 35900 + }, + { + "epoch": 9.411533420707732, + "grad_norm": 1.922353744506836, + "learning_rate": 5.26048977954563e-07, + "loss": 0.6099, + "num_input_tokens_seen": 9620448, + "step": 35905 + }, + { + "epoch": 9.412844036697248, + "grad_norm": 17.79767608642578, + "learning_rate": 5.237176286309292e-07, + "loss": 0.7651, + "num_input_tokens_seen": 9621424, + "step": 35910 + }, + { + "epoch": 9.414154652686763, + "grad_norm": 3.3822343349456787, + "learning_rate": 5.213914021509031e-07, + "loss": 0.5369, + "num_input_tokens_seen": 9622432, + "step": 35915 + }, + { + "epoch": 9.415465268676279, + "grad_norm": 4.392385959625244, + "learning_rate": 5.190702990013618e-07, + "loss": 0.4307, + "num_input_tokens_seen": 9623520, + "step": 35920 + }, + { + "epoch": 9.416775884665793, + "grad_norm": 2.5616180896759033, + "learning_rate": 5.167543196681058e-07, + "loss": 0.3317, + "num_input_tokens_seen": 9624672, + "step": 35925 + }, + { + "epoch": 9.418086500655308, + "grad_norm": 1.4229775667190552, + "learning_rate": 5.144434646358665e-07, + "loss": 0.451, + "num_input_tokens_seen": 9625952, + "step": 35930 + }, + { + "epoch": 9.419397116644824, + "grad_norm": 2.449812173843384, + "learning_rate": 5.121377343883016e-07, + "loss": 0.2767, + "num_input_tokens_seen": 9627584, + "step": 35935 + }, + { + "epoch": 9.420707732634337, + "grad_norm": 3.0756711959838867, + "learning_rate": 5.098371294080001e-07, + "loss": 0.4567, + "num_input_tokens_seen": 9628992, + "step": 35940 + }, + { + "epoch": 9.422018348623853, + "grad_norm": 2.04005765914917, + "learning_rate": 5.075416501764685e-07, + "loss": 0.3661, + "num_input_tokens_seen": 9630352, + "step": 35945 + }, + { + "epoch": 9.423328964613368, + "grad_norm": 4.815462589263916, + "learning_rate": 5.052512971741529e-07, + "loss": 0.4214, + "num_input_tokens_seen": 9631856, + "step": 35950 + }, + { + "epoch": 9.424639580602884, + "grad_norm": 6.943310737609863, + "learning_rate": 5.029660708804146e-07, + "loss": 0.3973, + "num_input_tokens_seen": 9633008, + "step": 35955 + }, + { + "epoch": 9.425950196592398, + "grad_norm": 2.155665874481201, + "learning_rate": 5.006859717735512e-07, + "loss": 0.4729, + "num_input_tokens_seen": 9634080, + "step": 35960 + }, + { + "epoch": 9.427260812581913, + "grad_norm": 4.689642429351807, + "learning_rate": 4.984110003307812e-07, + "loss": 0.3675, + "num_input_tokens_seen": 9635616, + "step": 35965 + }, + { + "epoch": 9.428571428571429, + "grad_norm": 3.7618327140808105, + "learning_rate": 4.961411570282543e-07, + "loss": 0.5221, + "num_input_tokens_seen": 9637072, + "step": 35970 + }, + { + "epoch": 9.429882044560944, + "grad_norm": 3.7443740367889404, + "learning_rate": 4.938764423410458e-07, + "loss": 0.4546, + "num_input_tokens_seen": 9638320, + "step": 35975 + }, + { + "epoch": 9.431192660550458, + "grad_norm": 1.8458434343338013, + "learning_rate": 4.916168567431544e-07, + "loss": 0.455, + "num_input_tokens_seen": 9639600, + "step": 35980 + }, + { + "epoch": 9.432503276539974, + "grad_norm": 5.426109313964844, + "learning_rate": 4.893624007075048e-07, + "loss": 0.4237, + "num_input_tokens_seen": 9641104, + "step": 35985 + }, + { + "epoch": 9.43381389252949, + "grad_norm": 3.337475061416626, + "learning_rate": 4.871130747059554e-07, + "loss": 0.411, + "num_input_tokens_seen": 9642432, + "step": 35990 + }, + { + "epoch": 9.435124508519005, + "grad_norm": 2.9626352787017822, + "learning_rate": 4.848688792092798e-07, + "loss": 0.4748, + "num_input_tokens_seen": 9644320, + "step": 35995 + }, + { + "epoch": 9.436435124508519, + "grad_norm": 5.873270511627197, + "learning_rate": 4.826298146871938e-07, + "loss": 0.4393, + "num_input_tokens_seen": 9645248, + "step": 36000 + }, + { + "epoch": 9.437745740498034, + "grad_norm": 1.4383858442306519, + "learning_rate": 4.803958816083254e-07, + "loss": 0.3376, + "num_input_tokens_seen": 9646496, + "step": 36005 + }, + { + "epoch": 9.43905635648755, + "grad_norm": 2.559854030609131, + "learning_rate": 4.781670804402283e-07, + "loss": 0.4649, + "num_input_tokens_seen": 9647776, + "step": 36010 + }, + { + "epoch": 9.440366972477065, + "grad_norm": 1.8274197578430176, + "learning_rate": 4.75943411649396e-07, + "loss": 0.5064, + "num_input_tokens_seen": 9648912, + "step": 36015 + }, + { + "epoch": 9.441677588466579, + "grad_norm": 3.5599184036254883, + "learning_rate": 4.7372487570123135e-07, + "loss": 0.3658, + "num_input_tokens_seen": 9650032, + "step": 36020 + }, + { + "epoch": 9.442988204456094, + "grad_norm": 1.25680410861969, + "learning_rate": 4.7151147306007117e-07, + "loss": 0.5041, + "num_input_tokens_seen": 9651744, + "step": 36025 + }, + { + "epoch": 9.44429882044561, + "grad_norm": 4.548203468322754, + "learning_rate": 4.6930320418917817e-07, + "loss": 0.2879, + "num_input_tokens_seen": 9652816, + "step": 36030 + }, + { + "epoch": 9.445609436435124, + "grad_norm": 3.4165594577789307, + "learning_rate": 4.6710006955073826e-07, + "loss": 0.393, + "num_input_tokens_seen": 9654112, + "step": 36035 + }, + { + "epoch": 9.44692005242464, + "grad_norm": 3.944941759109497, + "learning_rate": 4.649020696058659e-07, + "loss": 0.433, + "num_input_tokens_seen": 9655136, + "step": 36040 + }, + { + "epoch": 9.448230668414155, + "grad_norm": 2.7573835849761963, + "learning_rate": 4.627092048145959e-07, + "loss": 0.338, + "num_input_tokens_seen": 9656720, + "step": 36045 + }, + { + "epoch": 9.44954128440367, + "grad_norm": 1.8367669582366943, + "learning_rate": 4.6052147563589167e-07, + "loss": 0.4506, + "num_input_tokens_seen": 9657824, + "step": 36050 + }, + { + "epoch": 9.450851900393184, + "grad_norm": 3.6252927780151367, + "learning_rate": 4.583388825276397e-07, + "loss": 0.3894, + "num_input_tokens_seen": 9659072, + "step": 36055 + }, + { + "epoch": 9.4521625163827, + "grad_norm": 2.4598841667175293, + "learning_rate": 4.561614259466579e-07, + "loss": 0.4935, + "num_input_tokens_seen": 9660160, + "step": 36060 + }, + { + "epoch": 9.453473132372215, + "grad_norm": 3.3087098598480225, + "learning_rate": 4.53989106348679e-07, + "loss": 0.3651, + "num_input_tokens_seen": 9661632, + "step": 36065 + }, + { + "epoch": 9.45478374836173, + "grad_norm": 3.7555136680603027, + "learning_rate": 4.518219241883698e-07, + "loss": 0.3474, + "num_input_tokens_seen": 9662720, + "step": 36070 + }, + { + "epoch": 9.456094364351245, + "grad_norm": 2.779111862182617, + "learning_rate": 4.4965987991931466e-07, + "loss": 0.288, + "num_input_tokens_seen": 9664064, + "step": 36075 + }, + { + "epoch": 9.45740498034076, + "grad_norm": 2.6549084186553955, + "learning_rate": 4.475029739940295e-07, + "loss": 0.4276, + "num_input_tokens_seen": 9665712, + "step": 36080 + }, + { + "epoch": 9.458715596330276, + "grad_norm": 1.3344117403030396, + "learning_rate": 4.4535120686394484e-07, + "loss": 0.4038, + "num_input_tokens_seen": 9666896, + "step": 36085 + }, + { + "epoch": 9.460026212319791, + "grad_norm": 2.5019407272338867, + "learning_rate": 4.4320457897942815e-07, + "loss": 0.4067, + "num_input_tokens_seen": 9668320, + "step": 36090 + }, + { + "epoch": 9.461336828309305, + "grad_norm": 6.317018032073975, + "learning_rate": 4.410630907897617e-07, + "loss": 0.6427, + "num_input_tokens_seen": 9669472, + "step": 36095 + }, + { + "epoch": 9.46264744429882, + "grad_norm": 3.1661155223846436, + "learning_rate": 4.389267427431565e-07, + "loss": 0.4879, + "num_input_tokens_seen": 9670528, + "step": 36100 + }, + { + "epoch": 9.463958060288336, + "grad_norm": 4.179480075836182, + "learning_rate": 4.367955352867437e-07, + "loss": 0.4428, + "num_input_tokens_seen": 9671664, + "step": 36105 + }, + { + "epoch": 9.46526867627785, + "grad_norm": 2.538316488265991, + "learning_rate": 4.346694688665831e-07, + "loss": 0.5307, + "num_input_tokens_seen": 9672784, + "step": 36110 + }, + { + "epoch": 9.466579292267365, + "grad_norm": 2.073091506958008, + "learning_rate": 4.325485439276578e-07, + "loss": 0.494, + "num_input_tokens_seen": 9674016, + "step": 36115 + }, + { + "epoch": 9.46788990825688, + "grad_norm": 8.627588272094727, + "learning_rate": 4.304327609138736e-07, + "loss": 0.3522, + "num_input_tokens_seen": 9674944, + "step": 36120 + }, + { + "epoch": 9.469200524246396, + "grad_norm": 1.541882038116455, + "learning_rate": 4.2832212026805986e-07, + "loss": 0.3885, + "num_input_tokens_seen": 9677888, + "step": 36125 + }, + { + "epoch": 9.47051114023591, + "grad_norm": 2.275390863418579, + "learning_rate": 4.262166224319686e-07, + "loss": 0.8751, + "num_input_tokens_seen": 9678960, + "step": 36130 + }, + { + "epoch": 9.471821756225426, + "grad_norm": 4.690054416656494, + "learning_rate": 4.2411626784628065e-07, + "loss": 0.4298, + "num_input_tokens_seen": 9680320, + "step": 36135 + }, + { + "epoch": 9.473132372214941, + "grad_norm": 4.4723944664001465, + "learning_rate": 4.220210569505917e-07, + "loss": 0.5216, + "num_input_tokens_seen": 9681456, + "step": 36140 + }, + { + "epoch": 9.474442988204457, + "grad_norm": 3.5025391578674316, + "learning_rate": 4.199309901834286e-07, + "loss": 0.4054, + "num_input_tokens_seen": 9682672, + "step": 36145 + }, + { + "epoch": 9.47575360419397, + "grad_norm": 3.2226197719573975, + "learning_rate": 4.178460679822388e-07, + "loss": 0.319, + "num_input_tokens_seen": 9684144, + "step": 36150 + }, + { + "epoch": 9.477064220183486, + "grad_norm": 2.2666749954223633, + "learning_rate": 4.157662907833898e-07, + "loss": 0.405, + "num_input_tokens_seen": 9685200, + "step": 36155 + }, + { + "epoch": 9.478374836173002, + "grad_norm": 3.222724199295044, + "learning_rate": 4.1369165902218065e-07, + "loss": 0.3468, + "num_input_tokens_seen": 9686352, + "step": 36160 + }, + { + "epoch": 9.479685452162517, + "grad_norm": 5.775931358337402, + "learning_rate": 4.1162217313282236e-07, + "loss": 0.3806, + "num_input_tokens_seen": 9687408, + "step": 36165 + }, + { + "epoch": 9.48099606815203, + "grad_norm": 2.9195659160614014, + "learning_rate": 4.0955783354846013e-07, + "loss": 0.3435, + "num_input_tokens_seen": 9688544, + "step": 36170 + }, + { + "epoch": 9.482306684141546, + "grad_norm": 1.9731709957122803, + "learning_rate": 4.0749864070115387e-07, + "loss": 0.6431, + "num_input_tokens_seen": 9690016, + "step": 36175 + }, + { + "epoch": 9.483617300131062, + "grad_norm": 4.225332260131836, + "learning_rate": 4.0544459502188934e-07, + "loss": 0.3818, + "num_input_tokens_seen": 9691392, + "step": 36180 + }, + { + "epoch": 9.484927916120578, + "grad_norm": 3.0387792587280273, + "learning_rate": 4.0339569694057554e-07, + "loss": 0.3287, + "num_input_tokens_seen": 9692576, + "step": 36185 + }, + { + "epoch": 9.486238532110091, + "grad_norm": 5.3085761070251465, + "learning_rate": 4.013519468860416e-07, + "loss": 0.5651, + "num_input_tokens_seen": 9693952, + "step": 36190 + }, + { + "epoch": 9.487549148099607, + "grad_norm": 13.108534812927246, + "learning_rate": 3.993133452860426e-07, + "loss": 0.4706, + "num_input_tokens_seen": 9695568, + "step": 36195 + }, + { + "epoch": 9.488859764089122, + "grad_norm": 4.290036678314209, + "learning_rate": 3.9727989256725384e-07, + "loss": 0.4982, + "num_input_tokens_seen": 9696640, + "step": 36200 + }, + { + "epoch": 9.490170380078636, + "grad_norm": 2.8155996799468994, + "learning_rate": 3.9525158915527383e-07, + "loss": 0.6203, + "num_input_tokens_seen": 9697952, + "step": 36205 + }, + { + "epoch": 9.491480996068152, + "grad_norm": 4.795003890991211, + "learning_rate": 3.932284354746213e-07, + "loss": 0.5621, + "num_input_tokens_seen": 9699536, + "step": 36210 + }, + { + "epoch": 9.492791612057667, + "grad_norm": 3.0965168476104736, + "learning_rate": 3.9121043194874365e-07, + "loss": 0.7582, + "num_input_tokens_seen": 9701008, + "step": 36215 + }, + { + "epoch": 9.494102228047183, + "grad_norm": 2.3816747665405273, + "learning_rate": 3.8919757899999743e-07, + "loss": 0.3695, + "num_input_tokens_seen": 9702096, + "step": 36220 + }, + { + "epoch": 9.495412844036696, + "grad_norm": 1.2161275148391724, + "learning_rate": 3.8718987704967345e-07, + "loss": 0.3997, + "num_input_tokens_seen": 9703616, + "step": 36225 + }, + { + "epoch": 9.496723460026212, + "grad_norm": 4.2974982261657715, + "learning_rate": 3.8518732651798284e-07, + "loss": 0.3122, + "num_input_tokens_seen": 9704848, + "step": 36230 + }, + { + "epoch": 9.498034076015728, + "grad_norm": 1.9398059844970703, + "learning_rate": 3.8318992782405414e-07, + "loss": 0.6037, + "num_input_tokens_seen": 9706480, + "step": 36235 + }, + { + "epoch": 9.499344692005243, + "grad_norm": 2.8328356742858887, + "learning_rate": 3.811976813859364e-07, + "loss": 0.4715, + "num_input_tokens_seen": 9707568, + "step": 36240 + }, + { + "epoch": 9.500655307994757, + "grad_norm": 2.520714521408081, + "learning_rate": 3.792105876206098e-07, + "loss": 0.5604, + "num_input_tokens_seen": 9708768, + "step": 36245 + }, + { + "epoch": 9.501965923984272, + "grad_norm": 2.3650498390197754, + "learning_rate": 3.7722864694396397e-07, + "loss": 0.6231, + "num_input_tokens_seen": 9710384, + "step": 36250 + }, + { + "epoch": 9.502490170380078, + "eval_loss": 0.4644148349761963, + "eval_runtime": 18.2234, + "eval_samples_per_second": 46.534, + "eval_steps_per_second": 23.267, + "num_input_tokens_seen": 9710784, + "step": 36252 + }, + { + "epoch": 9.503276539973788, + "grad_norm": 1.0170708894729614, + "learning_rate": 3.7525185977081713e-07, + "loss": 0.2843, + "num_input_tokens_seen": 9711936, + "step": 36255 + }, + { + "epoch": 9.504587155963304, + "grad_norm": 3.977498769760132, + "learning_rate": 3.7328022651490767e-07, + "loss": 0.382, + "num_input_tokens_seen": 9713136, + "step": 36260 + }, + { + "epoch": 9.505897771952817, + "grad_norm": 4.7744317054748535, + "learning_rate": 3.7131374758889446e-07, + "loss": 0.7423, + "num_input_tokens_seen": 9714352, + "step": 36265 + }, + { + "epoch": 9.507208387942333, + "grad_norm": 8.834290504455566, + "learning_rate": 3.693524234043594e-07, + "loss": 0.4664, + "num_input_tokens_seen": 9715424, + "step": 36270 + }, + { + "epoch": 9.508519003931848, + "grad_norm": 13.774864196777344, + "learning_rate": 3.6739625437180457e-07, + "loss": 0.501, + "num_input_tokens_seen": 9716592, + "step": 36275 + }, + { + "epoch": 9.509829619921362, + "grad_norm": 1.6659451723098755, + "learning_rate": 3.654452409006498e-07, + "loss": 0.4567, + "num_input_tokens_seen": 9717936, + "step": 36280 + }, + { + "epoch": 9.511140235910878, + "grad_norm": 3.0819942951202393, + "learning_rate": 3.634993833992434e-07, + "loss": 0.5936, + "num_input_tokens_seen": 9719840, + "step": 36285 + }, + { + "epoch": 9.512450851900393, + "grad_norm": 4.088659286499023, + "learning_rate": 3.615586822748457e-07, + "loss": 0.4132, + "num_input_tokens_seen": 9720944, + "step": 36290 + }, + { + "epoch": 9.513761467889909, + "grad_norm": 2.5036792755126953, + "learning_rate": 3.596231379336429e-07, + "loss": 0.5189, + "num_input_tokens_seen": 9722288, + "step": 36295 + }, + { + "epoch": 9.515072083879424, + "grad_norm": 3.682349920272827, + "learning_rate": 3.5769275078074425e-07, + "loss": 0.4806, + "num_input_tokens_seen": 9723696, + "step": 36300 + }, + { + "epoch": 9.516382699868938, + "grad_norm": 5.805666446685791, + "learning_rate": 3.5576752122017385e-07, + "loss": 0.2844, + "num_input_tokens_seen": 9724720, + "step": 36305 + }, + { + "epoch": 9.517693315858454, + "grad_norm": 3.654193639755249, + "learning_rate": 3.5384744965487883e-07, + "loss": 0.5908, + "num_input_tokens_seen": 9726032, + "step": 36310 + }, + { + "epoch": 9.51900393184797, + "grad_norm": 5.153464317321777, + "learning_rate": 3.519325364867265e-07, + "loss": 0.3595, + "num_input_tokens_seen": 9726928, + "step": 36315 + }, + { + "epoch": 9.520314547837483, + "grad_norm": 3.1041676998138428, + "learning_rate": 3.5002278211650476e-07, + "loss": 0.3025, + "num_input_tokens_seen": 9728528, + "step": 36320 + }, + { + "epoch": 9.521625163826998, + "grad_norm": 3.192943572998047, + "learning_rate": 3.4811818694392706e-07, + "loss": 0.3353, + "num_input_tokens_seen": 9729808, + "step": 36325 + }, + { + "epoch": 9.522935779816514, + "grad_norm": 2.953446865081787, + "learning_rate": 3.462187513676135e-07, + "loss": 0.3755, + "num_input_tokens_seen": 9730800, + "step": 36330 + }, + { + "epoch": 9.52424639580603, + "grad_norm": 1.5842770338058472, + "learning_rate": 3.4432447578511826e-07, + "loss": 0.327, + "num_input_tokens_seen": 9732144, + "step": 36335 + }, + { + "epoch": 9.525557011795543, + "grad_norm": 3.710353136062622, + "learning_rate": 3.4243536059291035e-07, + "loss": 0.5464, + "num_input_tokens_seen": 9733424, + "step": 36340 + }, + { + "epoch": 9.526867627785059, + "grad_norm": 2.8160712718963623, + "learning_rate": 3.405514061863735e-07, + "loss": 0.4453, + "num_input_tokens_seen": 9734800, + "step": 36345 + }, + { + "epoch": 9.528178243774574, + "grad_norm": 2.8967597484588623, + "learning_rate": 3.3867261295982e-07, + "loss": 0.5759, + "num_input_tokens_seen": 9736272, + "step": 36350 + }, + { + "epoch": 9.52948885976409, + "grad_norm": 8.149152755737305, + "learning_rate": 3.36798981306477e-07, + "loss": 0.4823, + "num_input_tokens_seen": 9737664, + "step": 36355 + }, + { + "epoch": 9.530799475753604, + "grad_norm": 2.889207601547241, + "learning_rate": 3.3493051161849477e-07, + "loss": 0.6085, + "num_input_tokens_seen": 9738896, + "step": 36360 + }, + { + "epoch": 9.53211009174312, + "grad_norm": 1.1550350189208984, + "learning_rate": 3.330672042869354e-07, + "loss": 0.3231, + "num_input_tokens_seen": 9740288, + "step": 36365 + }, + { + "epoch": 9.533420707732635, + "grad_norm": 3.4501683712005615, + "learning_rate": 3.312090597017925e-07, + "loss": 0.4255, + "num_input_tokens_seen": 9741424, + "step": 36370 + }, + { + "epoch": 9.534731323722148, + "grad_norm": 7.659945487976074, + "learning_rate": 3.2935607825196614e-07, + "loss": 0.3459, + "num_input_tokens_seen": 9742656, + "step": 36375 + }, + { + "epoch": 9.536041939711664, + "grad_norm": 3.0688464641571045, + "learning_rate": 3.275082603252877e-07, + "loss": 0.4675, + "num_input_tokens_seen": 9743952, + "step": 36380 + }, + { + "epoch": 9.53735255570118, + "grad_norm": 1.4146919250488281, + "learning_rate": 3.2566560630849784e-07, + "loss": 0.4096, + "num_input_tokens_seen": 9745344, + "step": 36385 + }, + { + "epoch": 9.538663171690695, + "grad_norm": 1.9991121292114258, + "learning_rate": 3.2382811658726033e-07, + "loss": 0.361, + "num_input_tokens_seen": 9747088, + "step": 36390 + }, + { + "epoch": 9.539973787680209, + "grad_norm": 8.91376781463623, + "learning_rate": 3.219957915461619e-07, + "loss": 0.4035, + "num_input_tokens_seen": 9748416, + "step": 36395 + }, + { + "epoch": 9.541284403669724, + "grad_norm": 2.2569775581359863, + "learning_rate": 3.201686315687041e-07, + "loss": 0.5251, + "num_input_tokens_seen": 9749552, + "step": 36400 + }, + { + "epoch": 9.54259501965924, + "grad_norm": 2.6948537826538086, + "learning_rate": 3.183466370373117e-07, + "loss": 0.2695, + "num_input_tokens_seen": 9750624, + "step": 36405 + }, + { + "epoch": 9.543905635648755, + "grad_norm": 2.006354570388794, + "learning_rate": 3.1652980833331846e-07, + "loss": 0.387, + "num_input_tokens_seen": 9752272, + "step": 36410 + }, + { + "epoch": 9.54521625163827, + "grad_norm": 4.830735206604004, + "learning_rate": 3.147181458369897e-07, + "loss": 0.5501, + "num_input_tokens_seen": 9753552, + "step": 36415 + }, + { + "epoch": 9.546526867627785, + "grad_norm": 4.030601501464844, + "learning_rate": 3.12911649927497e-07, + "loss": 0.638, + "num_input_tokens_seen": 9755072, + "step": 36420 + }, + { + "epoch": 9.5478374836173, + "grad_norm": 2.8008391857147217, + "learning_rate": 3.111103209829408e-07, + "loss": 0.4726, + "num_input_tokens_seen": 9756448, + "step": 36425 + }, + { + "epoch": 9.549148099606816, + "grad_norm": 2.6798222064971924, + "learning_rate": 3.0931415938033336e-07, + "loss": 0.4905, + "num_input_tokens_seen": 9757920, + "step": 36430 + }, + { + "epoch": 9.55045871559633, + "grad_norm": 6.7065510749816895, + "learning_rate": 3.075231654956129e-07, + "loss": 0.4567, + "num_input_tokens_seen": 9759008, + "step": 36435 + }, + { + "epoch": 9.551769331585845, + "grad_norm": 3.5546653270721436, + "learning_rate": 3.0573733970362674e-07, + "loss": 0.3707, + "num_input_tokens_seen": 9759904, + "step": 36440 + }, + { + "epoch": 9.55307994757536, + "grad_norm": 4.359969139099121, + "learning_rate": 3.0395668237814813e-07, + "loss": 0.4859, + "num_input_tokens_seen": 9761520, + "step": 36445 + }, + { + "epoch": 9.554390563564876, + "grad_norm": 1.772533893585205, + "learning_rate": 3.0218119389186503e-07, + "loss": 0.5199, + "num_input_tokens_seen": 9763184, + "step": 36450 + }, + { + "epoch": 9.55570117955439, + "grad_norm": 2.009676456451416, + "learning_rate": 3.0041087461638304e-07, + "loss": 0.4943, + "num_input_tokens_seen": 9764224, + "step": 36455 + }, + { + "epoch": 9.557011795543906, + "grad_norm": 3.8506200313568115, + "learning_rate": 2.986457249222252e-07, + "loss": 0.5861, + "num_input_tokens_seen": 9765696, + "step": 36460 + }, + { + "epoch": 9.558322411533421, + "grad_norm": 2.7284085750579834, + "learning_rate": 2.968857451788404e-07, + "loss": 0.459, + "num_input_tokens_seen": 9767072, + "step": 36465 + }, + { + "epoch": 9.559633027522935, + "grad_norm": 1.722091794013977, + "learning_rate": 2.951309357545812e-07, + "loss": 0.3379, + "num_input_tokens_seen": 9768400, + "step": 36470 + }, + { + "epoch": 9.56094364351245, + "grad_norm": 2.646179437637329, + "learning_rate": 2.9338129701673434e-07, + "loss": 0.268, + "num_input_tokens_seen": 9769696, + "step": 36475 + }, + { + "epoch": 9.562254259501966, + "grad_norm": 1.331217646598816, + "learning_rate": 2.916368293314903e-07, + "loss": 0.4903, + "num_input_tokens_seen": 9771040, + "step": 36480 + }, + { + "epoch": 9.563564875491481, + "grad_norm": 3.229957342147827, + "learning_rate": 2.898975330639653e-07, + "loss": 0.3791, + "num_input_tokens_seen": 9772352, + "step": 36485 + }, + { + "epoch": 9.564875491480995, + "grad_norm": 2.7599101066589355, + "learning_rate": 2.8816340857819034e-07, + "loss": 0.5536, + "num_input_tokens_seen": 9773712, + "step": 36490 + }, + { + "epoch": 9.56618610747051, + "grad_norm": 2.7193872928619385, + "learning_rate": 2.8643445623711684e-07, + "loss": 0.4325, + "num_input_tokens_seen": 9774896, + "step": 36495 + }, + { + "epoch": 9.567496723460026, + "grad_norm": 2.3896429538726807, + "learning_rate": 2.847106764026081e-07, + "loss": 0.4535, + "num_input_tokens_seen": 9776672, + "step": 36500 + }, + { + "epoch": 9.568807339449542, + "grad_norm": 3.0143637657165527, + "learning_rate": 2.8299206943545044e-07, + "loss": 0.4658, + "num_input_tokens_seen": 9777616, + "step": 36505 + }, + { + "epoch": 9.570117955439056, + "grad_norm": 2.3099029064178467, + "learning_rate": 2.812786356953451e-07, + "loss": 0.4222, + "num_input_tokens_seen": 9778720, + "step": 36510 + }, + { + "epoch": 9.571428571428571, + "grad_norm": 2.8446991443634033, + "learning_rate": 2.795703755409107e-07, + "loss": 0.4351, + "num_input_tokens_seen": 9780304, + "step": 36515 + }, + { + "epoch": 9.572739187418087, + "grad_norm": 13.12182903289795, + "learning_rate": 2.7786728932968354e-07, + "loss": 0.5777, + "num_input_tokens_seen": 9781856, + "step": 36520 + }, + { + "epoch": 9.574049803407602, + "grad_norm": 2.6488234996795654, + "learning_rate": 2.761693774181173e-07, + "loss": 0.3893, + "num_input_tokens_seen": 9783312, + "step": 36525 + }, + { + "epoch": 9.575360419397116, + "grad_norm": 4.952175617218018, + "learning_rate": 2.7447664016157783e-07, + "loss": 0.3799, + "num_input_tokens_seen": 9784832, + "step": 36530 + }, + { + "epoch": 9.576671035386632, + "grad_norm": 4.461991310119629, + "learning_rate": 2.727890779143566e-07, + "loss": 0.3913, + "num_input_tokens_seen": 9786016, + "step": 36535 + }, + { + "epoch": 9.577981651376147, + "grad_norm": 10.575848579406738, + "learning_rate": 2.7110669102965723e-07, + "loss": 0.4441, + "num_input_tokens_seen": 9787312, + "step": 36540 + }, + { + "epoch": 9.579292267365663, + "grad_norm": 2.5164225101470947, + "learning_rate": 2.694294798595981e-07, + "loss": 0.5725, + "num_input_tokens_seen": 9788368, + "step": 36545 + }, + { + "epoch": 9.580602883355176, + "grad_norm": 1.951934576034546, + "learning_rate": 2.677574447552178e-07, + "loss": 0.6227, + "num_input_tokens_seen": 9789632, + "step": 36550 + }, + { + "epoch": 9.581913499344692, + "grad_norm": 1.8228892087936401, + "learning_rate": 2.660905860664697e-07, + "loss": 0.497, + "num_input_tokens_seen": 9791040, + "step": 36555 + }, + { + "epoch": 9.583224115334207, + "grad_norm": 3.567514657974243, + "learning_rate": 2.6442890414222487e-07, + "loss": 0.2953, + "num_input_tokens_seen": 9792320, + "step": 36560 + }, + { + "epoch": 9.584534731323721, + "grad_norm": 5.487175941467285, + "learning_rate": 2.6277239933027163e-07, + "loss": 0.524, + "num_input_tokens_seen": 9793552, + "step": 36565 + }, + { + "epoch": 9.585845347313237, + "grad_norm": 3.148735761642456, + "learning_rate": 2.611210719773133e-07, + "loss": 0.3568, + "num_input_tokens_seen": 9795136, + "step": 36570 + }, + { + "epoch": 9.587155963302752, + "grad_norm": 2.67765736579895, + "learning_rate": 2.5947492242896776e-07, + "loss": 0.5105, + "num_input_tokens_seen": 9796848, + "step": 36575 + }, + { + "epoch": 9.588466579292268, + "grad_norm": 5.837456703186035, + "learning_rate": 2.5783395102977057e-07, + "loss": 0.2868, + "num_input_tokens_seen": 9797904, + "step": 36580 + }, + { + "epoch": 9.589777195281782, + "grad_norm": 7.266870498657227, + "learning_rate": 2.5619815812318037e-07, + "loss": 0.503, + "num_input_tokens_seen": 9799248, + "step": 36585 + }, + { + "epoch": 9.591087811271297, + "grad_norm": 3.6531901359558105, + "learning_rate": 2.5456754405155934e-07, + "loss": 0.3767, + "num_input_tokens_seen": 9800304, + "step": 36590 + }, + { + "epoch": 9.592398427260813, + "grad_norm": 15.189360618591309, + "learning_rate": 2.5294210915619564e-07, + "loss": 0.3537, + "num_input_tokens_seen": 9801520, + "step": 36595 + }, + { + "epoch": 9.593709043250328, + "grad_norm": 5.411869049072266, + "learning_rate": 2.5132185377728656e-07, + "loss": 0.5526, + "num_input_tokens_seen": 9802720, + "step": 36600 + }, + { + "epoch": 9.595019659239842, + "grad_norm": 5.23284912109375, + "learning_rate": 2.497067782539553e-07, + "loss": 0.5554, + "num_input_tokens_seen": 9804384, + "step": 36605 + }, + { + "epoch": 9.596330275229358, + "grad_norm": 1.0142531394958496, + "learning_rate": 2.480968829242286e-07, + "loss": 0.4207, + "num_input_tokens_seen": 9805936, + "step": 36610 + }, + { + "epoch": 9.597640891218873, + "grad_norm": 1.9848331212997437, + "learning_rate": 2.4649216812505373e-07, + "loss": 0.3299, + "num_input_tokens_seen": 9807392, + "step": 36615 + }, + { + "epoch": 9.598951507208389, + "grad_norm": 1.9304910898208618, + "learning_rate": 2.448926341923008e-07, + "loss": 0.5301, + "num_input_tokens_seen": 9808624, + "step": 36620 + }, + { + "epoch": 9.600262123197902, + "grad_norm": 2.0615248680114746, + "learning_rate": 2.4329828146074095e-07, + "loss": 0.4366, + "num_input_tokens_seen": 9809808, + "step": 36625 + }, + { + "epoch": 9.601572739187418, + "grad_norm": 4.534021854400635, + "learning_rate": 2.417091102640795e-07, + "loss": 0.3898, + "num_input_tokens_seen": 9810992, + "step": 36630 + }, + { + "epoch": 9.602883355176933, + "grad_norm": 2.0459139347076416, + "learning_rate": 2.401251209349198e-07, + "loss": 0.5057, + "num_input_tokens_seen": 9812416, + "step": 36635 + }, + { + "epoch": 9.604193971166449, + "grad_norm": 3.952566385269165, + "learning_rate": 2.385463138047911e-07, + "loss": 0.4565, + "num_input_tokens_seen": 9813520, + "step": 36640 + }, + { + "epoch": 9.605504587155963, + "grad_norm": 5.043124198913574, + "learning_rate": 2.3697268920413462e-07, + "loss": 0.5264, + "num_input_tokens_seen": 9814960, + "step": 36645 + }, + { + "epoch": 9.606815203145478, + "grad_norm": 3.696488857269287, + "learning_rate": 2.3540424746230916e-07, + "loss": 0.5389, + "num_input_tokens_seen": 9816000, + "step": 36650 + }, + { + "epoch": 9.608125819134994, + "grad_norm": 2.3491358757019043, + "learning_rate": 2.3384098890758267e-07, + "loss": 0.8096, + "num_input_tokens_seen": 9817200, + "step": 36655 + }, + { + "epoch": 9.609436435124508, + "grad_norm": 2.5469813346862793, + "learning_rate": 2.322829138671434e-07, + "loss": 0.5564, + "num_input_tokens_seen": 9818336, + "step": 36660 + }, + { + "epoch": 9.610747051114023, + "grad_norm": 3.0777313709259033, + "learning_rate": 2.3073002266709443e-07, + "loss": 0.6724, + "num_input_tokens_seen": 9820016, + "step": 36665 + }, + { + "epoch": 9.612057667103539, + "grad_norm": 2.745176315307617, + "learning_rate": 2.2918231563245629e-07, + "loss": 0.5387, + "num_input_tokens_seen": 9821072, + "step": 36670 + }, + { + "epoch": 9.613368283093054, + "grad_norm": 4.7656378746032715, + "learning_rate": 2.2763979308715877e-07, + "loss": 0.282, + "num_input_tokens_seen": 9822144, + "step": 36675 + }, + { + "epoch": 9.614678899082568, + "grad_norm": 4.626561641693115, + "learning_rate": 2.2610245535404638e-07, + "loss": 0.4783, + "num_input_tokens_seen": 9823104, + "step": 36680 + }, + { + "epoch": 9.615989515072084, + "grad_norm": 5.703025817871094, + "learning_rate": 2.2457030275488676e-07, + "loss": 0.6589, + "num_input_tokens_seen": 9824736, + "step": 36685 + }, + { + "epoch": 9.617300131061599, + "grad_norm": 7.195417404174805, + "learning_rate": 2.2304333561035396e-07, + "loss": 0.3905, + "num_input_tokens_seen": 9825872, + "step": 36690 + }, + { + "epoch": 9.618610747051115, + "grad_norm": 7.020646095275879, + "learning_rate": 2.2152155424003952e-07, + "loss": 0.3989, + "num_input_tokens_seen": 9827168, + "step": 36695 + }, + { + "epoch": 9.619921363040628, + "grad_norm": 3.34566330909729, + "learning_rate": 2.2000495896244976e-07, + "loss": 0.5941, + "num_input_tokens_seen": 9828448, + "step": 36700 + }, + { + "epoch": 9.621231979030144, + "grad_norm": 2.0993969440460205, + "learning_rate": 2.1849355009500582e-07, + "loss": 0.5156, + "num_input_tokens_seen": 9829648, + "step": 36705 + }, + { + "epoch": 9.62254259501966, + "grad_norm": 1.654142141342163, + "learning_rate": 2.1698732795404074e-07, + "loss": 0.4565, + "num_input_tokens_seen": 9831216, + "step": 36710 + }, + { + "epoch": 9.623853211009175, + "grad_norm": 1.196428656578064, + "learning_rate": 2.1548629285480792e-07, + "loss": 0.3016, + "num_input_tokens_seen": 9832976, + "step": 36715 + }, + { + "epoch": 9.625163826998689, + "grad_norm": 3.7811927795410156, + "learning_rate": 2.1399044511146993e-07, + "loss": 0.5339, + "num_input_tokens_seen": 9833904, + "step": 36720 + }, + { + "epoch": 9.626474442988204, + "grad_norm": 4.578232765197754, + "learning_rate": 2.124997850371041e-07, + "loss": 0.4297, + "num_input_tokens_seen": 9834944, + "step": 36725 + }, + { + "epoch": 9.62778505897772, + "grad_norm": 3.2911782264709473, + "learning_rate": 2.1101431294370533e-07, + "loss": 0.2881, + "num_input_tokens_seen": 9835920, + "step": 36730 + }, + { + "epoch": 9.629095674967235, + "grad_norm": 3.6512258052825928, + "learning_rate": 2.0953402914217768e-07, + "loss": 0.5307, + "num_input_tokens_seen": 9837264, + "step": 36735 + }, + { + "epoch": 9.63040629095675, + "grad_norm": 2.1630778312683105, + "learning_rate": 2.0805893394234e-07, + "loss": 0.7567, + "num_input_tokens_seen": 9838336, + "step": 36740 + }, + { + "epoch": 9.631716906946265, + "grad_norm": 9.633103370666504, + "learning_rate": 2.065890276529342e-07, + "loss": 0.5949, + "num_input_tokens_seen": 9839424, + "step": 36745 + }, + { + "epoch": 9.63302752293578, + "grad_norm": 3.7974154949188232, + "learning_rate": 2.051243105816031e-07, + "loss": 0.4292, + "num_input_tokens_seen": 9840656, + "step": 36750 + }, + { + "epoch": 9.634338138925294, + "grad_norm": 1.7192325592041016, + "learning_rate": 2.0366478303491255e-07, + "loss": 0.3512, + "num_input_tokens_seen": 9842096, + "step": 36755 + }, + { + "epoch": 9.63564875491481, + "grad_norm": 4.755732536315918, + "learning_rate": 2.0221044531834043e-07, + "loss": 0.4569, + "num_input_tokens_seen": 9843392, + "step": 36760 + }, + { + "epoch": 9.636959370904325, + "grad_norm": 2.3121190071105957, + "learning_rate": 2.0076129773627105e-07, + "loss": 0.5385, + "num_input_tokens_seen": 9844896, + "step": 36765 + }, + { + "epoch": 9.63826998689384, + "grad_norm": 8.314573287963867, + "learning_rate": 1.9931734059201457e-07, + "loss": 0.645, + "num_input_tokens_seen": 9846048, + "step": 36770 + }, + { + "epoch": 9.639580602883354, + "grad_norm": 3.426619291305542, + "learning_rate": 1.9787857418778478e-07, + "loss": 0.5354, + "num_input_tokens_seen": 9847360, + "step": 36775 + }, + { + "epoch": 9.64089121887287, + "grad_norm": 6.9236040115356445, + "learning_rate": 1.9644499882471578e-07, + "loss": 0.6062, + "num_input_tokens_seen": 9848672, + "step": 36780 + }, + { + "epoch": 9.642201834862385, + "grad_norm": 4.573215484619141, + "learning_rate": 1.9501661480285095e-07, + "loss": 0.5071, + "num_input_tokens_seen": 9849776, + "step": 36785 + }, + { + "epoch": 9.643512450851901, + "grad_norm": 5.549885272979736, + "learning_rate": 1.935934224211483e-07, + "loss": 0.4328, + "num_input_tokens_seen": 9851200, + "step": 36790 + }, + { + "epoch": 9.644823066841415, + "grad_norm": 7.138340950012207, + "learning_rate": 1.9217542197748074e-07, + "loss": 0.7537, + "num_input_tokens_seen": 9852560, + "step": 36795 + }, + { + "epoch": 9.64613368283093, + "grad_norm": 4.535216808319092, + "learning_rate": 1.9076261376863303e-07, + "loss": 0.5039, + "num_input_tokens_seen": 9853872, + "step": 36800 + }, + { + "epoch": 9.647444298820446, + "grad_norm": 2.997465133666992, + "learning_rate": 1.8935499809030476e-07, + "loss": 0.3572, + "num_input_tokens_seen": 9855584, + "step": 36805 + }, + { + "epoch": 9.648754914809961, + "grad_norm": 3.639089822769165, + "learning_rate": 1.8795257523710197e-07, + "loss": 0.4215, + "num_input_tokens_seen": 9856624, + "step": 36810 + }, + { + "epoch": 9.650065530799475, + "grad_norm": 6.013915538787842, + "learning_rate": 1.8655534550255648e-07, + "loss": 0.4103, + "num_input_tokens_seen": 9858384, + "step": 36815 + }, + { + "epoch": 9.65137614678899, + "grad_norm": 2.4636402130126953, + "learning_rate": 1.8516330917910108e-07, + "loss": 0.8813, + "num_input_tokens_seen": 9859328, + "step": 36820 + }, + { + "epoch": 9.652686762778506, + "grad_norm": 2.7072949409484863, + "learning_rate": 1.8377646655808877e-07, + "loss": 0.5863, + "num_input_tokens_seen": 9861040, + "step": 36825 + }, + { + "epoch": 9.653997378768022, + "grad_norm": 2.8962762355804443, + "learning_rate": 1.8239481792978464e-07, + "loss": 0.5199, + "num_input_tokens_seen": 9862128, + "step": 36830 + }, + { + "epoch": 9.655307994757536, + "grad_norm": 2.197871446609497, + "learning_rate": 1.8101836358336287e-07, + "loss": 0.5259, + "num_input_tokens_seen": 9863664, + "step": 36835 + }, + { + "epoch": 9.656618610747051, + "grad_norm": 4.020021915435791, + "learning_rate": 1.7964710380691251e-07, + "loss": 0.5546, + "num_input_tokens_seen": 9864592, + "step": 36840 + }, + { + "epoch": 9.657929226736567, + "grad_norm": 1.5576765537261963, + "learning_rate": 1.782810388874373e-07, + "loss": 0.5533, + "num_input_tokens_seen": 9866448, + "step": 36845 + }, + { + "epoch": 9.65923984272608, + "grad_norm": 4.660694122314453, + "learning_rate": 1.7692016911085295e-07, + "loss": 0.4996, + "num_input_tokens_seen": 9868144, + "step": 36850 + }, + { + "epoch": 9.660550458715596, + "grad_norm": 4.276114463806152, + "learning_rate": 1.7556449476198445e-07, + "loss": 0.5635, + "num_input_tokens_seen": 9869312, + "step": 36855 + }, + { + "epoch": 9.661861074705111, + "grad_norm": 3.0154879093170166, + "learning_rate": 1.7421401612457423e-07, + "loss": 0.3785, + "num_input_tokens_seen": 9870368, + "step": 36860 + }, + { + "epoch": 9.663171690694627, + "grad_norm": 1.682773232460022, + "learning_rate": 1.7286873348127676e-07, + "loss": 0.3791, + "num_input_tokens_seen": 9871520, + "step": 36865 + }, + { + "epoch": 9.66448230668414, + "grad_norm": 2.8315536975860596, + "learning_rate": 1.715286471136529e-07, + "loss": 0.6127, + "num_input_tokens_seen": 9872928, + "step": 36870 + }, + { + "epoch": 9.665792922673656, + "grad_norm": 1.9034093618392944, + "learning_rate": 1.7019375730218667e-07, + "loss": 0.3197, + "num_input_tokens_seen": 9874160, + "step": 36875 + }, + { + "epoch": 9.667103538663172, + "grad_norm": 2.348507881164551, + "learning_rate": 1.688640643262629e-07, + "loss": 0.3932, + "num_input_tokens_seen": 9875328, + "step": 36880 + }, + { + "epoch": 9.668414154652687, + "grad_norm": 2.4731128215789795, + "learning_rate": 1.6753956846418395e-07, + "loss": 0.3459, + "num_input_tokens_seen": 9876624, + "step": 36885 + }, + { + "epoch": 9.669724770642201, + "grad_norm": 5.995449066162109, + "learning_rate": 1.6622026999316697e-07, + "loss": 0.6014, + "num_input_tokens_seen": 9877984, + "step": 36890 + }, + { + "epoch": 9.671035386631717, + "grad_norm": 2.2575790882110596, + "learning_rate": 1.6490616918933554e-07, + "loss": 0.333, + "num_input_tokens_seen": 9879152, + "step": 36895 + }, + { + "epoch": 9.672346002621232, + "grad_norm": 3.72969651222229, + "learning_rate": 1.6359726632773353e-07, + "loss": 0.3574, + "num_input_tokens_seen": 9880192, + "step": 36900 + }, + { + "epoch": 9.673656618610748, + "grad_norm": 1.2998420000076294, + "learning_rate": 1.6229356168231125e-07, + "loss": 0.4911, + "num_input_tokens_seen": 9881408, + "step": 36905 + }, + { + "epoch": 9.674967234600262, + "grad_norm": 1.942789077758789, + "learning_rate": 1.609950555259282e-07, + "loss": 0.6371, + "num_input_tokens_seen": 9883088, + "step": 36910 + }, + { + "epoch": 9.676277850589777, + "grad_norm": 4.619419097900391, + "learning_rate": 1.5970174813036144e-07, + "loss": 0.338, + "num_input_tokens_seen": 9884688, + "step": 36915 + }, + { + "epoch": 9.677588466579293, + "grad_norm": 3.7484405040740967, + "learning_rate": 1.5841363976629998e-07, + "loss": 0.5053, + "num_input_tokens_seen": 9886176, + "step": 36920 + }, + { + "epoch": 9.678899082568808, + "grad_norm": 3.6500370502471924, + "learning_rate": 1.5713073070333927e-07, + "loss": 0.3, + "num_input_tokens_seen": 9887232, + "step": 36925 + }, + { + "epoch": 9.680209698558322, + "grad_norm": 2.2415902614593506, + "learning_rate": 1.5585302120998956e-07, + "loss": 0.4015, + "num_input_tokens_seen": 9888784, + "step": 36930 + }, + { + "epoch": 9.681520314547837, + "grad_norm": 5.596822738647461, + "learning_rate": 1.545805115536786e-07, + "loss": 0.3544, + "num_input_tokens_seen": 9889888, + "step": 36935 + }, + { + "epoch": 9.682830930537353, + "grad_norm": 2.780852794647217, + "learning_rate": 1.5331320200073497e-07, + "loss": 0.4375, + "num_input_tokens_seen": 9891424, + "step": 36940 + }, + { + "epoch": 9.684141546526867, + "grad_norm": 3.539799213409424, + "learning_rate": 1.520510928164076e-07, + "loss": 0.3923, + "num_input_tokens_seen": 9892464, + "step": 36945 + }, + { + "epoch": 9.685452162516382, + "grad_norm": 3.722968339920044, + "learning_rate": 1.5079418426485193e-07, + "loss": 0.4986, + "num_input_tokens_seen": 9893632, + "step": 36950 + }, + { + "epoch": 9.686762778505898, + "grad_norm": 3.3041367530822754, + "learning_rate": 1.49542476609138e-07, + "loss": 0.3674, + "num_input_tokens_seen": 9895056, + "step": 36955 + }, + { + "epoch": 9.688073394495413, + "grad_norm": 4.116991996765137, + "learning_rate": 1.482959701112452e-07, + "loss": 0.432, + "num_input_tokens_seen": 9896400, + "step": 36960 + }, + { + "epoch": 9.689384010484927, + "grad_norm": 7.76773738861084, + "learning_rate": 1.47054665032062e-07, + "loss": 0.4488, + "num_input_tokens_seen": 9897920, + "step": 36965 + }, + { + "epoch": 9.690694626474443, + "grad_norm": 4.103569030761719, + "learning_rate": 1.4581856163140008e-07, + "loss": 0.5548, + "num_input_tokens_seen": 9899696, + "step": 36970 + }, + { + "epoch": 9.692005242463958, + "grad_norm": 1.9561855792999268, + "learning_rate": 1.4458766016796632e-07, + "loss": 0.4107, + "num_input_tokens_seen": 9901232, + "step": 36975 + }, + { + "epoch": 9.693315858453474, + "grad_norm": 2.660496950149536, + "learning_rate": 1.4336196089938802e-07, + "loss": 0.3285, + "num_input_tokens_seen": 9902736, + "step": 36980 + }, + { + "epoch": 9.694626474442988, + "grad_norm": 2.093114137649536, + "learning_rate": 1.4214146408220163e-07, + "loss": 0.4428, + "num_input_tokens_seen": 9904272, + "step": 36985 + }, + { + "epoch": 9.695937090432503, + "grad_norm": 5.724133491516113, + "learning_rate": 1.4092616997185837e-07, + "loss": 0.2677, + "num_input_tokens_seen": 9905376, + "step": 36990 + }, + { + "epoch": 9.697247706422019, + "grad_norm": 6.902961730957031, + "learning_rate": 1.397160788227131e-07, + "loss": 0.4453, + "num_input_tokens_seen": 9906368, + "step": 36995 + }, + { + "epoch": 9.698558322411534, + "grad_norm": 2.591120719909668, + "learning_rate": 1.385111908880382e-07, + "loss": 0.4303, + "num_input_tokens_seen": 9907760, + "step": 37000 + }, + { + "epoch": 9.699868938401048, + "grad_norm": 1.3520300388336182, + "learning_rate": 1.3731150642001255e-07, + "loss": 0.613, + "num_input_tokens_seen": 9909280, + "step": 37005 + }, + { + "epoch": 9.701179554390563, + "grad_norm": 4.899772644042969, + "learning_rate": 1.3611702566972694e-07, + "loss": 0.517, + "num_input_tokens_seen": 9910608, + "step": 37010 + }, + { + "epoch": 9.702490170380079, + "grad_norm": 11.537737846374512, + "learning_rate": 1.3492774888718974e-07, + "loss": 0.9993, + "num_input_tokens_seen": 9911600, + "step": 37015 + }, + { + "epoch": 9.703800786369595, + "grad_norm": 4.372610569000244, + "learning_rate": 1.3374367632131014e-07, + "loss": 0.5074, + "num_input_tokens_seen": 9912944, + "step": 37020 + }, + { + "epoch": 9.705111402359108, + "grad_norm": 9.523377418518066, + "learning_rate": 1.3256480821991213e-07, + "loss": 0.6065, + "num_input_tokens_seen": 9914080, + "step": 37025 + }, + { + "epoch": 9.706422018348624, + "grad_norm": 8.278286933898926, + "learning_rate": 1.3139114482973448e-07, + "loss": 0.4257, + "num_input_tokens_seen": 9915088, + "step": 37030 + }, + { + "epoch": 9.70773263433814, + "grad_norm": 3.0970189571380615, + "learning_rate": 1.3022268639641956e-07, + "loss": 0.3233, + "num_input_tokens_seen": 9916208, + "step": 37035 + }, + { + "epoch": 9.709043250327653, + "grad_norm": 3.2616777420043945, + "learning_rate": 1.290594331645245e-07, + "loss": 0.443, + "num_input_tokens_seen": 9917360, + "step": 37040 + }, + { + "epoch": 9.710353866317169, + "grad_norm": 1.9186309576034546, + "learning_rate": 1.279013853775185e-07, + "loss": 0.5076, + "num_input_tokens_seen": 9918784, + "step": 37045 + }, + { + "epoch": 9.711664482306684, + "grad_norm": 2.487856388092041, + "learning_rate": 1.267485432777743e-07, + "loss": 0.2904, + "num_input_tokens_seen": 9920272, + "step": 37050 + }, + { + "epoch": 9.7129750982962, + "grad_norm": 8.509283065795898, + "learning_rate": 1.2560090710658223e-07, + "loss": 0.4639, + "num_input_tokens_seen": 9921792, + "step": 37055 + }, + { + "epoch": 9.714285714285714, + "grad_norm": 32.62371063232422, + "learning_rate": 1.2445847710414183e-07, + "loss": 0.5242, + "num_input_tokens_seen": 9922992, + "step": 37060 + }, + { + "epoch": 9.715596330275229, + "grad_norm": 2.700806140899658, + "learning_rate": 1.233212535095618e-07, + "loss": 0.4581, + "num_input_tokens_seen": 9924256, + "step": 37065 + }, + { + "epoch": 9.716906946264745, + "grad_norm": 5.2650017738342285, + "learning_rate": 1.2218923656085735e-07, + "loss": 0.5968, + "num_input_tokens_seen": 9925184, + "step": 37070 + }, + { + "epoch": 9.71821756225426, + "grad_norm": 4.049737930297852, + "learning_rate": 1.2106242649496112e-07, + "loss": 0.5842, + "num_input_tokens_seen": 9926528, + "step": 37075 + }, + { + "epoch": 9.719528178243774, + "grad_norm": 2.5973260402679443, + "learning_rate": 1.199408235477123e-07, + "loss": 0.3363, + "num_input_tokens_seen": 9927920, + "step": 37080 + }, + { + "epoch": 9.72083879423329, + "grad_norm": 4.856617450714111, + "learning_rate": 1.188244279538564e-07, + "loss": 0.4191, + "num_input_tokens_seen": 9929344, + "step": 37085 + }, + { + "epoch": 9.722149410222805, + "grad_norm": 0.8349355459213257, + "learning_rate": 1.1771323994705929e-07, + "loss": 0.3482, + "num_input_tokens_seen": 9930560, + "step": 37090 + }, + { + "epoch": 9.72346002621232, + "grad_norm": 5.063260078430176, + "learning_rate": 1.1660725975988773e-07, + "loss": 0.3133, + "num_input_tokens_seen": 9931680, + "step": 37095 + }, + { + "epoch": 9.724770642201834, + "grad_norm": 3.770831823348999, + "learning_rate": 1.1550648762382044e-07, + "loss": 0.467, + "num_input_tokens_seen": 9932704, + "step": 37100 + }, + { + "epoch": 9.72608125819135, + "grad_norm": 4.847357273101807, + "learning_rate": 1.144109237692509e-07, + "loss": 0.5936, + "num_input_tokens_seen": 9933776, + "step": 37105 + }, + { + "epoch": 9.727391874180865, + "grad_norm": 4.397853851318359, + "learning_rate": 1.1332056842547344e-07, + "loss": 0.3316, + "num_input_tokens_seen": 9935376, + "step": 37110 + }, + { + "epoch": 9.728702490170381, + "grad_norm": 3.8477017879486084, + "learning_rate": 1.1223542182070002e-07, + "loss": 0.3898, + "num_input_tokens_seen": 9936624, + "step": 37115 + }, + { + "epoch": 9.730013106159895, + "grad_norm": 3.8804571628570557, + "learning_rate": 1.1115548418205168e-07, + "loss": 0.4055, + "num_input_tokens_seen": 9937744, + "step": 37120 + }, + { + "epoch": 9.73132372214941, + "grad_norm": 1.7989721298217773, + "learning_rate": 1.1008075573555599e-07, + "loss": 0.4577, + "num_input_tokens_seen": 9939040, + "step": 37125 + }, + { + "epoch": 9.732634338138926, + "grad_norm": 9.669116020202637, + "learning_rate": 1.0901123670614965e-07, + "loss": 0.3461, + "num_input_tokens_seen": 9940352, + "step": 37130 + }, + { + "epoch": 9.73394495412844, + "grad_norm": 9.339088439941406, + "learning_rate": 1.0794692731768419e-07, + "loss": 0.461, + "num_input_tokens_seen": 9941552, + "step": 37135 + }, + { + "epoch": 9.735255570117955, + "grad_norm": 1.350570797920227, + "learning_rate": 1.0688782779291473e-07, + "loss": 0.4573, + "num_input_tokens_seen": 9943088, + "step": 37140 + }, + { + "epoch": 9.73656618610747, + "grad_norm": 4.3970947265625, + "learning_rate": 1.0583393835351396e-07, + "loss": 0.4006, + "num_input_tokens_seen": 9944192, + "step": 37145 + }, + { + "epoch": 9.737876802096986, + "grad_norm": 1.7985260486602783, + "learning_rate": 1.0478525922005545e-07, + "loss": 0.3065, + "num_input_tokens_seen": 9945936, + "step": 37150 + }, + { + "epoch": 9.7391874180865, + "grad_norm": 4.74815559387207, + "learning_rate": 1.0374179061202749e-07, + "loss": 0.4585, + "num_input_tokens_seen": 9947504, + "step": 37155 + }, + { + "epoch": 9.740498034076015, + "grad_norm": 2.026550054550171, + "learning_rate": 1.0270353274782207e-07, + "loss": 0.2372, + "num_input_tokens_seen": 9948656, + "step": 37160 + }, + { + "epoch": 9.741808650065531, + "grad_norm": 2.2053470611572266, + "learning_rate": 1.0167048584475147e-07, + "loss": 0.3198, + "num_input_tokens_seen": 9950832, + "step": 37165 + }, + { + "epoch": 9.743119266055047, + "grad_norm": 4.119060039520264, + "learning_rate": 1.006426501190233e-07, + "loss": 0.419, + "num_input_tokens_seen": 9952656, + "step": 37170 + }, + { + "epoch": 9.74442988204456, + "grad_norm": 1.5536768436431885, + "learning_rate": 9.962002578576823e-08, + "loss": 0.3943, + "num_input_tokens_seen": 9953648, + "step": 37175 + }, + { + "epoch": 9.745740498034076, + "grad_norm": 3.428016185760498, + "learning_rate": 9.860261305901785e-08, + "loss": 0.6126, + "num_input_tokens_seen": 9954720, + "step": 37180 + }, + { + "epoch": 9.747051114023591, + "grad_norm": 10.715883255004883, + "learning_rate": 9.759041215171295e-08, + "loss": 0.6923, + "num_input_tokens_seen": 9955776, + "step": 37185 + }, + { + "epoch": 9.748361730013107, + "grad_norm": 3.034113883972168, + "learning_rate": 9.658342327570902e-08, + "loss": 0.5265, + "num_input_tokens_seen": 9957072, + "step": 37190 + }, + { + "epoch": 9.74967234600262, + "grad_norm": 2.625889301300049, + "learning_rate": 9.55816466417625e-08, + "loss": 0.3366, + "num_input_tokens_seen": 9958464, + "step": 37195 + }, + { + "epoch": 9.750982961992136, + "grad_norm": 7.12210750579834, + "learning_rate": 9.458508245954456e-08, + "loss": 0.5896, + "num_input_tokens_seen": 9959824, + "step": 37200 + }, + { + "epoch": 9.752293577981652, + "grad_norm": 2.597959280014038, + "learning_rate": 9.359373093763835e-08, + "loss": 0.5354, + "num_input_tokens_seen": 9961232, + "step": 37205 + }, + { + "epoch": 9.753604193971167, + "grad_norm": 4.032848358154297, + "learning_rate": 9.260759228352789e-08, + "loss": 0.5544, + "num_input_tokens_seen": 9963040, + "step": 37210 + }, + { + "epoch": 9.754914809960681, + "grad_norm": 3.5044031143188477, + "learning_rate": 9.162666670361198e-08, + "loss": 0.4281, + "num_input_tokens_seen": 9964192, + "step": 37215 + }, + { + "epoch": 9.756225425950197, + "grad_norm": 4.879634380340576, + "learning_rate": 9.065095440319582e-08, + "loss": 0.7711, + "num_input_tokens_seen": 9965648, + "step": 37220 + }, + { + "epoch": 9.757536041939712, + "grad_norm": 1.4151219129562378, + "learning_rate": 8.968045558649663e-08, + "loss": 0.373, + "num_input_tokens_seen": 9967216, + "step": 37225 + }, + { + "epoch": 9.758846657929226, + "grad_norm": 4.699089527130127, + "learning_rate": 8.871517045663524e-08, + "loss": 0.4866, + "num_input_tokens_seen": 9968896, + "step": 37230 + }, + { + "epoch": 9.760157273918741, + "grad_norm": 3.172208786010742, + "learning_rate": 8.775509921564728e-08, + "loss": 0.4039, + "num_input_tokens_seen": 9970576, + "step": 37235 + }, + { + "epoch": 9.761467889908257, + "grad_norm": 1.5838569402694702, + "learning_rate": 8.680024206446924e-08, + "loss": 0.6578, + "num_input_tokens_seen": 9972080, + "step": 37240 + }, + { + "epoch": 9.762778505897773, + "grad_norm": 1.4297672510147095, + "learning_rate": 8.585059920295512e-08, + "loss": 0.4219, + "num_input_tokens_seen": 9973536, + "step": 37245 + }, + { + "epoch": 9.764089121887286, + "grad_norm": 8.92773723602295, + "learning_rate": 8.490617082985986e-08, + "loss": 0.3138, + "num_input_tokens_seen": 9974592, + "step": 37250 + }, + { + "epoch": 9.765399737876802, + "grad_norm": 1.6688332557678223, + "learning_rate": 8.396695714285585e-08, + "loss": 0.418, + "num_input_tokens_seen": 9975760, + "step": 37255 + }, + { + "epoch": 9.766710353866317, + "grad_norm": 12.398496627807617, + "learning_rate": 8.303295833851365e-08, + "loss": 0.3359, + "num_input_tokens_seen": 9976832, + "step": 37260 + }, + { + "epoch": 9.768020969855833, + "grad_norm": 1.6135272979736328, + "learning_rate": 8.210417461232412e-08, + "loss": 0.8076, + "num_input_tokens_seen": 9978224, + "step": 37265 + }, + { + "epoch": 9.769331585845347, + "grad_norm": 2.8665318489074707, + "learning_rate": 8.11806061586734e-08, + "loss": 0.4236, + "num_input_tokens_seen": 9979600, + "step": 37270 + }, + { + "epoch": 9.770642201834862, + "grad_norm": 2.3882784843444824, + "learning_rate": 8.026225317086522e-08, + "loss": 0.7952, + "num_input_tokens_seen": 9981040, + "step": 37275 + }, + { + "epoch": 9.771952817824378, + "grad_norm": 3.480428695678711, + "learning_rate": 7.934911584110971e-08, + "loss": 0.3827, + "num_input_tokens_seen": 9982288, + "step": 37280 + }, + { + "epoch": 9.773263433813893, + "grad_norm": 0.9266604781150818, + "learning_rate": 7.844119436052622e-08, + "loss": 0.4753, + "num_input_tokens_seen": 9983424, + "step": 37285 + }, + { + "epoch": 9.774574049803407, + "grad_norm": 2.5613365173339844, + "learning_rate": 7.753848891913772e-08, + "loss": 0.4159, + "num_input_tokens_seen": 9984528, + "step": 37290 + }, + { + "epoch": 9.775884665792923, + "grad_norm": 1.5828609466552734, + "learning_rate": 7.664099970588201e-08, + "loss": 0.3387, + "num_input_tokens_seen": 9985840, + "step": 37295 + }, + { + "epoch": 9.777195281782438, + "grad_norm": 2.566232204437256, + "learning_rate": 7.57487269085977e-08, + "loss": 0.3063, + "num_input_tokens_seen": 9987248, + "step": 37300 + }, + { + "epoch": 9.778505897771954, + "grad_norm": 6.689155101776123, + "learning_rate": 7.486167071404371e-08, + "loss": 0.3989, + "num_input_tokens_seen": 9988240, + "step": 37305 + }, + { + "epoch": 9.779816513761467, + "grad_norm": 9.079622268676758, + "learning_rate": 7.397983130787156e-08, + "loss": 0.3123, + "num_input_tokens_seen": 9989536, + "step": 37310 + }, + { + "epoch": 9.781127129750983, + "grad_norm": 4.548454761505127, + "learning_rate": 7.310320887465305e-08, + "loss": 0.4459, + "num_input_tokens_seen": 9990608, + "step": 37315 + }, + { + "epoch": 9.782437745740499, + "grad_norm": 4.055171012878418, + "learning_rate": 7.223180359786086e-08, + "loss": 0.3141, + "num_input_tokens_seen": 9992048, + "step": 37320 + }, + { + "epoch": 9.783748361730012, + "grad_norm": 3.2844672203063965, + "learning_rate": 7.136561565988242e-08, + "loss": 0.5881, + "num_input_tokens_seen": 9993328, + "step": 37325 + }, + { + "epoch": 9.785058977719528, + "grad_norm": 3.006943702697754, + "learning_rate": 7.050464524200607e-08, + "loss": 0.5805, + "num_input_tokens_seen": 9994560, + "step": 37330 + }, + { + "epoch": 9.786369593709043, + "grad_norm": 1.6528805494308472, + "learning_rate": 6.964889252443485e-08, + "loss": 0.5419, + "num_input_tokens_seen": 9996320, + "step": 37335 + }, + { + "epoch": 9.787680209698559, + "grad_norm": 5.51164436340332, + "learning_rate": 6.879835768627274e-08, + "loss": 0.278, + "num_input_tokens_seen": 9997712, + "step": 37340 + }, + { + "epoch": 9.788990825688073, + "grad_norm": 3.4132301807403564, + "learning_rate": 6.795304090553567e-08, + "loss": 0.4306, + "num_input_tokens_seen": 9998960, + "step": 37345 + }, + { + "epoch": 9.790301441677588, + "grad_norm": 20.38096809387207, + "learning_rate": 6.711294235914877e-08, + "loss": 0.4672, + "num_input_tokens_seen": 10000592, + "step": 37350 + }, + { + "epoch": 9.791612057667104, + "grad_norm": 1.7401520013809204, + "learning_rate": 6.627806222294086e-08, + "loss": 0.4097, + "num_input_tokens_seen": 10002080, + "step": 37355 + }, + { + "epoch": 9.79292267365662, + "grad_norm": 4.043654918670654, + "learning_rate": 6.544840067165548e-08, + "loss": 0.4745, + "num_input_tokens_seen": 10003392, + "step": 37360 + }, + { + "epoch": 9.794233289646133, + "grad_norm": 1.974805235862732, + "learning_rate": 6.462395787893427e-08, + "loss": 0.4986, + "num_input_tokens_seen": 10004928, + "step": 37365 + }, + { + "epoch": 9.795543905635649, + "grad_norm": 8.8792724609375, + "learning_rate": 6.380473401733366e-08, + "loss": 0.4327, + "num_input_tokens_seen": 10005904, + "step": 37370 + }, + { + "epoch": 9.796854521625164, + "grad_norm": 2.3082032203674316, + "learning_rate": 6.299072925831373e-08, + "loss": 0.6154, + "num_input_tokens_seen": 10006944, + "step": 37375 + }, + { + "epoch": 9.79816513761468, + "grad_norm": 1.6091580390930176, + "learning_rate": 6.218194377224928e-08, + "loss": 0.5157, + "num_input_tokens_seen": 10008320, + "step": 37380 + }, + { + "epoch": 9.799475753604193, + "grad_norm": 2.8473055362701416, + "learning_rate": 6.137837772841326e-08, + "loss": 0.4487, + "num_input_tokens_seen": 10009504, + "step": 37385 + }, + { + "epoch": 9.800786369593709, + "grad_norm": 3.319814682006836, + "learning_rate": 6.058003129499334e-08, + "loss": 0.3696, + "num_input_tokens_seen": 10010896, + "step": 37390 + }, + { + "epoch": 9.802096985583225, + "grad_norm": 15.69229507446289, + "learning_rate": 5.978690463908088e-08, + "loss": 0.3567, + "num_input_tokens_seen": 10012048, + "step": 37395 + }, + { + "epoch": 9.80340760157274, + "grad_norm": 5.8064680099487305, + "learning_rate": 5.8998997926676405e-08, + "loss": 0.5638, + "num_input_tokens_seen": 10013200, + "step": 37400 + }, + { + "epoch": 9.804718217562254, + "grad_norm": 7.521440029144287, + "learning_rate": 5.821631132268412e-08, + "loss": 0.6094, + "num_input_tokens_seen": 10014320, + "step": 37405 + }, + { + "epoch": 9.80602883355177, + "grad_norm": 0.8078501224517822, + "learning_rate": 5.743884499092578e-08, + "loss": 0.4961, + "num_input_tokens_seen": 10016336, + "step": 37410 + }, + { + "epoch": 9.807339449541285, + "grad_norm": 1.9062607288360596, + "learning_rate": 5.6666599094115646e-08, + "loss": 0.2503, + "num_input_tokens_seen": 10017728, + "step": 37415 + }, + { + "epoch": 9.808650065530799, + "grad_norm": 2.212507486343384, + "learning_rate": 5.58995737938911e-08, + "loss": 0.5375, + "num_input_tokens_seen": 10018992, + "step": 37420 + }, + { + "epoch": 9.809960681520314, + "grad_norm": 1.812024712562561, + "learning_rate": 5.513776925078207e-08, + "loss": 0.5454, + "num_input_tokens_seen": 10020400, + "step": 37425 + }, + { + "epoch": 9.81127129750983, + "grad_norm": 3.8293848037719727, + "learning_rate": 5.4381185624238776e-08, + "loss": 0.4164, + "num_input_tokens_seen": 10021392, + "step": 37430 + }, + { + "epoch": 9.812581913499345, + "grad_norm": 2.1468758583068848, + "learning_rate": 5.362982307261233e-08, + "loss": 0.4264, + "num_input_tokens_seen": 10022784, + "step": 37435 + }, + { + "epoch": 9.813892529488859, + "grad_norm": 2.339840888977051, + "learning_rate": 5.2883681753157497e-08, + "loss": 0.8296, + "num_input_tokens_seen": 10023888, + "step": 37440 + }, + { + "epoch": 9.815203145478375, + "grad_norm": 1.8636722564697266, + "learning_rate": 5.214276182204381e-08, + "loss": 0.3616, + "num_input_tokens_seen": 10025376, + "step": 37445 + }, + { + "epoch": 9.81651376146789, + "grad_norm": 1.8993425369262695, + "learning_rate": 5.140706343434165e-08, + "loss": 0.5374, + "num_input_tokens_seen": 10027248, + "step": 37450 + }, + { + "epoch": 9.817824377457406, + "grad_norm": 2.1093709468841553, + "learning_rate": 5.0676586744036194e-08, + "loss": 0.2843, + "num_input_tokens_seen": 10028704, + "step": 37455 + }, + { + "epoch": 9.81913499344692, + "grad_norm": 3.045285701751709, + "learning_rate": 4.9951331904007915e-08, + "loss": 0.5301, + "num_input_tokens_seen": 10030080, + "step": 37460 + }, + { + "epoch": 9.820445609436435, + "grad_norm": 2.584074020385742, + "learning_rate": 4.923129906606039e-08, + "loss": 0.3898, + "num_input_tokens_seen": 10031936, + "step": 37465 + }, + { + "epoch": 9.82175622542595, + "grad_norm": 12.879185676574707, + "learning_rate": 4.851648838088696e-08, + "loss": 0.4562, + "num_input_tokens_seen": 10032944, + "step": 37470 + }, + { + "epoch": 9.823066841415466, + "grad_norm": 4.808753967285156, + "learning_rate": 4.7806899998101283e-08, + "loss": 0.4719, + "num_input_tokens_seen": 10034128, + "step": 37475 + }, + { + "epoch": 9.82437745740498, + "grad_norm": 2.1943981647491455, + "learning_rate": 4.71025340662179e-08, + "loss": 0.4303, + "num_input_tokens_seen": 10035216, + "step": 37480 + }, + { + "epoch": 9.825688073394495, + "grad_norm": 3.718449592590332, + "learning_rate": 4.6403390732654985e-08, + "loss": 0.4351, + "num_input_tokens_seen": 10036736, + "step": 37485 + }, + { + "epoch": 9.82699868938401, + "grad_norm": 0.8467246294021606, + "learning_rate": 4.570947014374827e-08, + "loss": 0.3419, + "num_input_tokens_seen": 10038496, + "step": 37490 + }, + { + "epoch": 9.828309305373526, + "grad_norm": 3.3202102184295654, + "learning_rate": 4.502077244473435e-08, + "loss": 0.6545, + "num_input_tokens_seen": 10039744, + "step": 37495 + }, + { + "epoch": 9.82961992136304, + "grad_norm": 1.8360629081726074, + "learning_rate": 4.4337297779750705e-08, + "loss": 0.4536, + "num_input_tokens_seen": 10041888, + "step": 37500 + }, + { + "epoch": 9.830930537352556, + "grad_norm": 2.6758902072906494, + "learning_rate": 4.365904629185236e-08, + "loss": 0.2949, + "num_input_tokens_seen": 10043856, + "step": 37505 + }, + { + "epoch": 9.832241153342071, + "grad_norm": 2.46090030670166, + "learning_rate": 4.298601812299241e-08, + "loss": 0.2457, + "num_input_tokens_seen": 10045184, + "step": 37510 + }, + { + "epoch": 9.833551769331585, + "grad_norm": 3.3170642852783203, + "learning_rate": 4.2318213414038745e-08, + "loss": 0.5211, + "num_input_tokens_seen": 10046416, + "step": 37515 + }, + { + "epoch": 9.8348623853211, + "grad_norm": 10.51761245727539, + "learning_rate": 4.1655632304757334e-08, + "loss": 0.4201, + "num_input_tokens_seen": 10047808, + "step": 37520 + }, + { + "epoch": 9.836173001310616, + "grad_norm": 3.861800193786621, + "learning_rate": 4.0998274933828905e-08, + "loss": 0.348, + "num_input_tokens_seen": 10049104, + "step": 37525 + }, + { + "epoch": 9.837483617300132, + "grad_norm": 12.530860900878906, + "learning_rate": 4.034614143883508e-08, + "loss": 0.5667, + "num_input_tokens_seen": 10050144, + "step": 37530 + }, + { + "epoch": 9.838794233289645, + "grad_norm": 1.049141764640808, + "learning_rate": 3.969923195626668e-08, + "loss": 0.6578, + "num_input_tokens_seen": 10051488, + "step": 37535 + }, + { + "epoch": 9.840104849279161, + "grad_norm": 6.50029182434082, + "learning_rate": 3.9057546621520946e-08, + "loss": 0.3663, + "num_input_tokens_seen": 10052336, + "step": 37540 + }, + { + "epoch": 9.841415465268676, + "grad_norm": 7.642570495605469, + "learning_rate": 3.842108556890156e-08, + "loss": 0.5609, + "num_input_tokens_seen": 10053600, + "step": 37545 + }, + { + "epoch": 9.842726081258192, + "grad_norm": 4.271821022033691, + "learning_rate": 3.778984893161863e-08, + "loss": 0.394, + "num_input_tokens_seen": 10054832, + "step": 37550 + }, + { + "epoch": 9.844036697247706, + "grad_norm": 2.262721538543701, + "learning_rate": 3.7163836841791476e-08, + "loss": 0.4783, + "num_input_tokens_seen": 10056080, + "step": 37555 + }, + { + "epoch": 9.845347313237221, + "grad_norm": 3.2165677547454834, + "learning_rate": 3.654304943043752e-08, + "loss": 0.4629, + "num_input_tokens_seen": 10057248, + "step": 37560 + }, + { + "epoch": 9.846657929226737, + "grad_norm": 3.3012335300445557, + "learning_rate": 3.5927486827491696e-08, + "loss": 0.3988, + "num_input_tokens_seen": 10058528, + "step": 37565 + }, + { + "epoch": 9.847968545216252, + "grad_norm": 2.4823319911956787, + "learning_rate": 3.531714916178708e-08, + "loss": 0.8589, + "num_input_tokens_seen": 10059792, + "step": 37570 + }, + { + "epoch": 9.849279161205766, + "grad_norm": 5.932900428771973, + "learning_rate": 3.4712036561068693e-08, + "loss": 0.5301, + "num_input_tokens_seen": 10061472, + "step": 37575 + }, + { + "epoch": 9.850589777195282, + "grad_norm": 2.1746878623962402, + "learning_rate": 3.4112149151982466e-08, + "loss": 0.3067, + "num_input_tokens_seen": 10062976, + "step": 37580 + }, + { + "epoch": 9.851900393184797, + "grad_norm": 3.1900694370269775, + "learning_rate": 3.351748706008628e-08, + "loss": 0.4163, + "num_input_tokens_seen": 10064272, + "step": 37585 + }, + { + "epoch": 9.853211009174313, + "grad_norm": 5.902387619018555, + "learning_rate": 3.292805040984171e-08, + "loss": 0.4554, + "num_input_tokens_seen": 10065696, + "step": 37590 + }, + { + "epoch": 9.854521625163827, + "grad_norm": 2.535313606262207, + "learning_rate": 3.234383932461671e-08, + "loss": 0.4883, + "num_input_tokens_seen": 10066656, + "step": 37595 + }, + { + "epoch": 9.855832241153342, + "grad_norm": 2.1270313262939453, + "learning_rate": 3.176485392668571e-08, + "loss": 0.3162, + "num_input_tokens_seen": 10067632, + "step": 37600 + }, + { + "epoch": 9.857142857142858, + "grad_norm": 4.2351250648498535, + "learning_rate": 3.119109433722955e-08, + "loss": 0.3895, + "num_input_tokens_seen": 10068832, + "step": 37605 + }, + { + "epoch": 9.858453473132371, + "grad_norm": 1.552340030670166, + "learning_rate": 3.0622560676332734e-08, + "loss": 0.5132, + "num_input_tokens_seen": 10070224, + "step": 37610 + }, + { + "epoch": 9.859764089121887, + "grad_norm": 4.165393352508545, + "learning_rate": 3.005925306299173e-08, + "loss": 0.4392, + "num_input_tokens_seen": 10071952, + "step": 37615 + }, + { + "epoch": 9.861074705111402, + "grad_norm": 9.926936149597168, + "learning_rate": 2.9501171615103907e-08, + "loss": 0.5632, + "num_input_tokens_seen": 10073200, + "step": 37620 + }, + { + "epoch": 9.862385321100918, + "grad_norm": 1.9410256147384644, + "learning_rate": 2.8948316449473044e-08, + "loss": 0.4667, + "num_input_tokens_seen": 10074672, + "step": 37625 + }, + { + "epoch": 9.863695937090432, + "grad_norm": 9.928619384765625, + "learning_rate": 2.840068768181492e-08, + "loss": 0.4994, + "num_input_tokens_seen": 10075952, + "step": 37630 + }, + { + "epoch": 9.865006553079947, + "grad_norm": 4.1241984367370605, + "learning_rate": 2.785828542674618e-08, + "loss": 0.4288, + "num_input_tokens_seen": 10077008, + "step": 37635 + }, + { + "epoch": 9.866317169069463, + "grad_norm": 2.1396706104278564, + "learning_rate": 2.7321109797787125e-08, + "loss": 0.4825, + "num_input_tokens_seen": 10078224, + "step": 37640 + }, + { + "epoch": 9.867627785058978, + "grad_norm": 1.6534976959228516, + "learning_rate": 2.6789160907372822e-08, + "loss": 0.5178, + "num_input_tokens_seen": 10079392, + "step": 37645 + }, + { + "epoch": 9.868938401048492, + "grad_norm": 2.521904706954956, + "learning_rate": 2.626243886683366e-08, + "loss": 0.605, + "num_input_tokens_seen": 10080752, + "step": 37650 + }, + { + "epoch": 9.870249017038008, + "grad_norm": 3.6226933002471924, + "learning_rate": 2.574094378641756e-08, + "loss": 0.4248, + "num_input_tokens_seen": 10081968, + "step": 37655 + }, + { + "epoch": 9.871559633027523, + "grad_norm": 2.3589096069335938, + "learning_rate": 2.5224675775270544e-08, + "loss": 0.4038, + "num_input_tokens_seen": 10083488, + "step": 37660 + }, + { + "epoch": 9.872870249017039, + "grad_norm": 1.2116461992263794, + "learning_rate": 2.4713634941442298e-08, + "loss": 0.2978, + "num_input_tokens_seen": 10085296, + "step": 37665 + }, + { + "epoch": 9.874180865006553, + "grad_norm": 1.684102177619934, + "learning_rate": 2.4207821391900032e-08, + "loss": 0.4928, + "num_input_tokens_seen": 10086544, + "step": 37670 + }, + { + "epoch": 9.875491480996068, + "grad_norm": 1.3849166631698608, + "learning_rate": 2.370723523250351e-08, + "loss": 0.4888, + "num_input_tokens_seen": 10087824, + "step": 37675 + }, + { + "epoch": 9.876802096985584, + "grad_norm": 10.857104301452637, + "learning_rate": 2.321187656802726e-08, + "loss": 0.5324, + "num_input_tokens_seen": 10089248, + "step": 37680 + }, + { + "epoch": 9.8781127129751, + "grad_norm": 6.787441253662109, + "learning_rate": 2.272174550214945e-08, + "loss": 0.7052, + "num_input_tokens_seen": 10090400, + "step": 37685 + }, + { + "epoch": 9.879423328964613, + "grad_norm": 2.811203956604004, + "learning_rate": 2.223684213745192e-08, + "loss": 0.5775, + "num_input_tokens_seen": 10091440, + "step": 37690 + }, + { + "epoch": 9.880733944954128, + "grad_norm": 4.943232536315918, + "learning_rate": 2.1757166575425702e-08, + "loss": 0.4851, + "num_input_tokens_seen": 10092784, + "step": 37695 + }, + { + "epoch": 9.882044560943644, + "grad_norm": 5.154699325561523, + "learning_rate": 2.1282718916465494e-08, + "loss": 0.4529, + "num_input_tokens_seen": 10094352, + "step": 37700 + }, + { + "epoch": 9.883355176933158, + "grad_norm": 3.255122423171997, + "learning_rate": 2.0813499259872414e-08, + "loss": 0.4666, + "num_input_tokens_seen": 10095488, + "step": 37705 + }, + { + "epoch": 9.884665792922673, + "grad_norm": 4.345831871032715, + "learning_rate": 2.0349507703851244e-08, + "loss": 0.5032, + "num_input_tokens_seen": 10096496, + "step": 37710 + }, + { + "epoch": 9.885976408912189, + "grad_norm": 3.997427463531494, + "learning_rate": 1.9890744345518742e-08, + "loss": 0.4291, + "num_input_tokens_seen": 10098080, + "step": 37715 + }, + { + "epoch": 9.887287024901704, + "grad_norm": 2.39752459526062, + "learning_rate": 1.9437209280889768e-08, + "loss": 0.3265, + "num_input_tokens_seen": 10099808, + "step": 37720 + }, + { + "epoch": 9.888597640891218, + "grad_norm": 8.76745891571045, + "learning_rate": 1.8988902604891166e-08, + "loss": 0.5593, + "num_input_tokens_seen": 10100736, + "step": 37725 + }, + { + "epoch": 9.889908256880734, + "grad_norm": 2.592496395111084, + "learning_rate": 1.8545824411350665e-08, + "loss": 0.4665, + "num_input_tokens_seen": 10101856, + "step": 37730 + }, + { + "epoch": 9.89121887287025, + "grad_norm": 4.498986721038818, + "learning_rate": 1.810797479300519e-08, + "loss": 0.5775, + "num_input_tokens_seen": 10102944, + "step": 37735 + }, + { + "epoch": 9.892529488859765, + "grad_norm": 1.3209362030029297, + "learning_rate": 1.7675353841495325e-08, + "loss": 0.4107, + "num_input_tokens_seen": 10105040, + "step": 37740 + }, + { + "epoch": 9.893840104849279, + "grad_norm": 4.647012710571289, + "learning_rate": 1.7247961647368082e-08, + "loss": 0.5416, + "num_input_tokens_seen": 10106576, + "step": 37745 + }, + { + "epoch": 9.895150720838794, + "grad_norm": 3.4594969749450684, + "learning_rate": 1.6825798300074137e-08, + "loss": 0.4093, + "num_input_tokens_seen": 10108000, + "step": 37750 + }, + { + "epoch": 9.89646133682831, + "grad_norm": 2.9057064056396484, + "learning_rate": 1.640886388797336e-08, + "loss": 0.3691, + "num_input_tokens_seen": 10109776, + "step": 37755 + }, + { + "epoch": 9.897771952817825, + "grad_norm": 3.291598081588745, + "learning_rate": 1.5997158498329277e-08, + "loss": 0.5016, + "num_input_tokens_seen": 10111344, + "step": 37760 + }, + { + "epoch": 9.899082568807339, + "grad_norm": 3.8150367736816406, + "learning_rate": 1.559068221731186e-08, + "loss": 0.4656, + "num_input_tokens_seen": 10112816, + "step": 37765 + }, + { + "epoch": 9.900393184796854, + "grad_norm": 2.182119369506836, + "learning_rate": 1.518943512999471e-08, + "loss": 0.4262, + "num_input_tokens_seen": 10114112, + "step": 37770 + }, + { + "epoch": 9.90170380078637, + "grad_norm": 2.9042575359344482, + "learning_rate": 1.4793417320357882e-08, + "loss": 0.5565, + "num_input_tokens_seen": 10115584, + "step": 37775 + }, + { + "epoch": 9.903014416775886, + "grad_norm": 2.5614371299743652, + "learning_rate": 1.4402628871285073e-08, + "loss": 0.4864, + "num_input_tokens_seen": 10117760, + "step": 37780 + }, + { + "epoch": 9.9043250327654, + "grad_norm": 2.831359624862671, + "learning_rate": 1.401706986457474e-08, + "loss": 0.5071, + "num_input_tokens_seen": 10118720, + "step": 37785 + }, + { + "epoch": 9.905635648754915, + "grad_norm": 2.773726463317871, + "learning_rate": 1.3636740380915113e-08, + "loss": 0.6881, + "num_input_tokens_seen": 10120128, + "step": 37790 + }, + { + "epoch": 9.90694626474443, + "grad_norm": 2.061652421951294, + "learning_rate": 1.326164049991474e-08, + "loss": 0.4767, + "num_input_tokens_seen": 10122112, + "step": 37795 + }, + { + "epoch": 9.908256880733944, + "grad_norm": 2.2342231273651123, + "learning_rate": 1.2891770300080263e-08, + "loss": 0.5301, + "num_input_tokens_seen": 10123536, + "step": 37800 + }, + { + "epoch": 9.90956749672346, + "grad_norm": 1.7479207515716553, + "learning_rate": 1.2527129858821985e-08, + "loss": 0.4071, + "num_input_tokens_seen": 10125088, + "step": 37805 + }, + { + "epoch": 9.910878112712975, + "grad_norm": 2.7545838356018066, + "learning_rate": 1.2167719252462183e-08, + "loss": 0.5652, + "num_input_tokens_seen": 10126512, + "step": 37810 + }, + { + "epoch": 9.91218872870249, + "grad_norm": 2.3677830696105957, + "learning_rate": 1.1813538556221249e-08, + "loss": 0.6393, + "num_input_tokens_seen": 10127488, + "step": 37815 + }, + { + "epoch": 9.913499344692005, + "grad_norm": 2.7720677852630615, + "learning_rate": 1.1464587844231544e-08, + "loss": 0.4564, + "num_input_tokens_seen": 10128704, + "step": 37820 + }, + { + "epoch": 9.91480996068152, + "grad_norm": 6.138557434082031, + "learning_rate": 1.1120867189526318e-08, + "loss": 0.5635, + "num_input_tokens_seen": 10129904, + "step": 37825 + }, + { + "epoch": 9.916120576671036, + "grad_norm": 2.9976770877838135, + "learning_rate": 1.0782376664045246e-08, + "loss": 0.4425, + "num_input_tokens_seen": 10131232, + "step": 37830 + }, + { + "epoch": 9.917431192660551, + "grad_norm": 1.7294683456420898, + "learning_rate": 1.044911633863721e-08, + "loss": 0.3406, + "num_input_tokens_seen": 10132640, + "step": 37835 + }, + { + "epoch": 9.918741808650065, + "grad_norm": 6.635354518890381, + "learning_rate": 1.0121086283049198e-08, + "loss": 0.5424, + "num_input_tokens_seen": 10133728, + "step": 37840 + }, + { + "epoch": 9.92005242463958, + "grad_norm": 3.1084799766540527, + "learning_rate": 9.798286565937399e-09, + "loss": 0.4494, + "num_input_tokens_seen": 10134992, + "step": 37845 + }, + { + "epoch": 9.921363040629096, + "grad_norm": 12.739140510559082, + "learning_rate": 9.48071725486721e-09, + "loss": 0.3853, + "num_input_tokens_seen": 10136000, + "step": 37850 + }, + { + "epoch": 9.922673656618612, + "grad_norm": 4.972506523132324, + "learning_rate": 9.168378416299362e-09, + "loss": 0.5436, + "num_input_tokens_seen": 10137008, + "step": 37855 + }, + { + "epoch": 9.923984272608125, + "grad_norm": 2.0442073345184326, + "learning_rate": 8.861270115612108e-09, + "loss": 0.3707, + "num_input_tokens_seen": 10138576, + "step": 37860 + }, + { + "epoch": 9.92529488859764, + "grad_norm": 4.373222827911377, + "learning_rate": 8.559392417079038e-09, + "loss": 0.3039, + "num_input_tokens_seen": 10139728, + "step": 37865 + }, + { + "epoch": 9.926605504587156, + "grad_norm": 9.884424209594727, + "learning_rate": 8.262745383880166e-09, + "loss": 0.62, + "num_input_tokens_seen": 10141328, + "step": 37870 + }, + { + "epoch": 9.927916120576672, + "grad_norm": 4.578514575958252, + "learning_rate": 7.971329078110268e-09, + "loss": 0.4175, + "num_input_tokens_seen": 10142432, + "step": 37875 + }, + { + "epoch": 9.929226736566186, + "grad_norm": 2.044703722000122, + "learning_rate": 7.68514356075667e-09, + "loss": 0.5269, + "num_input_tokens_seen": 10143936, + "step": 37880 + }, + { + "epoch": 9.930537352555701, + "grad_norm": 24.877199172973633, + "learning_rate": 7.4041888917186775e-09, + "loss": 0.3476, + "num_input_tokens_seen": 10145152, + "step": 37885 + }, + { + "epoch": 9.931847968545217, + "grad_norm": 2.2890727519989014, + "learning_rate": 7.1284651297992556e-09, + "loss": 0.522, + "num_input_tokens_seen": 10146720, + "step": 37890 + }, + { + "epoch": 9.93315858453473, + "grad_norm": 5.274936199188232, + "learning_rate": 6.8579723327105715e-09, + "loss": 0.4965, + "num_input_tokens_seen": 10147776, + "step": 37895 + }, + { + "epoch": 9.934469200524246, + "grad_norm": 4.156195640563965, + "learning_rate": 6.5927105570601224e-09, + "loss": 0.5095, + "num_input_tokens_seen": 10148848, + "step": 37900 + }, + { + "epoch": 9.935779816513762, + "grad_norm": 5.230598449707031, + "learning_rate": 6.3326798583729364e-09, + "loss": 0.5331, + "num_input_tokens_seen": 10149872, + "step": 37905 + }, + { + "epoch": 9.937090432503277, + "grad_norm": 8.94764232635498, + "learning_rate": 6.077880291069371e-09, + "loss": 0.5014, + "num_input_tokens_seen": 10151088, + "step": 37910 + }, + { + "epoch": 9.938401048492791, + "grad_norm": 4.458880424499512, + "learning_rate": 5.8283119084789895e-09, + "loss": 0.5018, + "num_input_tokens_seen": 10152384, + "step": 37915 + }, + { + "epoch": 9.939711664482306, + "grad_norm": 4.048871040344238, + "learning_rate": 5.583974762835009e-09, + "loss": 0.5274, + "num_input_tokens_seen": 10153728, + "step": 37920 + }, + { + "epoch": 9.941022280471822, + "grad_norm": 4.441789627075195, + "learning_rate": 5.344868905279854e-09, + "loss": 0.4422, + "num_input_tokens_seen": 10155152, + "step": 37925 + }, + { + "epoch": 9.942332896461338, + "grad_norm": 2.974168062210083, + "learning_rate": 5.110994385856826e-09, + "loss": 0.5795, + "num_input_tokens_seen": 10156560, + "step": 37930 + }, + { + "epoch": 9.943643512450851, + "grad_norm": 1.6276882886886597, + "learning_rate": 4.8823512535128845e-09, + "loss": 0.3342, + "num_input_tokens_seen": 10158288, + "step": 37935 + }, + { + "epoch": 9.944954128440367, + "grad_norm": 3.7455270290374756, + "learning_rate": 4.658939556104191e-09, + "loss": 0.3313, + "num_input_tokens_seen": 10159664, + "step": 37940 + }, + { + "epoch": 9.946264744429882, + "grad_norm": 5.146298885345459, + "learning_rate": 4.440759340393341e-09, + "loss": 0.5568, + "num_input_tokens_seen": 10160720, + "step": 37945 + }, + { + "epoch": 9.947575360419398, + "grad_norm": 2.7797248363494873, + "learning_rate": 4.227810652041031e-09, + "loss": 0.5958, + "num_input_tokens_seen": 10161984, + "step": 37950 + }, + { + "epoch": 9.948885976408912, + "grad_norm": 4.525611877441406, + "learning_rate": 4.0200935356171645e-09, + "loss": 0.4652, + "num_input_tokens_seen": 10163232, + "step": 37955 + }, + { + "epoch": 9.950196592398427, + "grad_norm": 6.228055477142334, + "learning_rate": 3.817608034600851e-09, + "loss": 0.5117, + "num_input_tokens_seen": 10164272, + "step": 37960 + }, + { + "epoch": 9.951507208387943, + "grad_norm": 2.7895443439483643, + "learning_rate": 3.620354191366526e-09, + "loss": 0.4646, + "num_input_tokens_seen": 10165328, + "step": 37965 + }, + { + "epoch": 9.952817824377458, + "grad_norm": 40.875118255615234, + "learning_rate": 3.4283320472033864e-09, + "loss": 0.5442, + "num_input_tokens_seen": 10166352, + "step": 37970 + }, + { + "epoch": 9.954128440366972, + "grad_norm": 1.690145492553711, + "learning_rate": 3.241541642298729e-09, + "loss": 0.6032, + "num_input_tokens_seen": 10168000, + "step": 37975 + }, + { + "epoch": 9.955439056356488, + "grad_norm": 2.9472155570983887, + "learning_rate": 3.059983015749057e-09, + "loss": 0.3812, + "num_input_tokens_seen": 10169088, + "step": 37980 + }, + { + "epoch": 9.956749672346003, + "grad_norm": 1.0865963697433472, + "learning_rate": 2.88365620555453e-09, + "loss": 0.4265, + "num_input_tokens_seen": 10171456, + "step": 37985 + }, + { + "epoch": 9.958060288335517, + "grad_norm": 5.26079797744751, + "learning_rate": 2.712561248618961e-09, + "loss": 0.5511, + "num_input_tokens_seen": 10172672, + "step": 37990 + }, + { + "epoch": 9.959370904325032, + "grad_norm": 6.731012344360352, + "learning_rate": 2.546698180749818e-09, + "loss": 0.6493, + "num_input_tokens_seen": 10173888, + "step": 37995 + }, + { + "epoch": 9.960681520314548, + "grad_norm": 3.268022060394287, + "learning_rate": 2.3860670366665505e-09, + "loss": 0.4543, + "num_input_tokens_seen": 10175392, + "step": 38000 + }, + { + "epoch": 9.961992136304064, + "grad_norm": 5.012324810028076, + "learning_rate": 2.2306678499867116e-09, + "loss": 0.4227, + "num_input_tokens_seen": 10176576, + "step": 38005 + }, + { + "epoch": 9.963302752293577, + "grad_norm": 4.575169563293457, + "learning_rate": 2.080500653234285e-09, + "loss": 0.366, + "num_input_tokens_seen": 10177664, + "step": 38010 + }, + { + "epoch": 9.964613368283093, + "grad_norm": 2.6589903831481934, + "learning_rate": 1.935565477839685e-09, + "loss": 0.3758, + "num_input_tokens_seen": 10179152, + "step": 38015 + }, + { + "epoch": 9.965923984272608, + "grad_norm": 6.698283672332764, + "learning_rate": 1.7958623541397546e-09, + "loss": 0.5286, + "num_input_tokens_seen": 10180096, + "step": 38020 + }, + { + "epoch": 9.967234600262124, + "grad_norm": 1.4664866924285889, + "learning_rate": 1.6613913113694424e-09, + "loss": 0.5744, + "num_input_tokens_seen": 10181456, + "step": 38025 + }, + { + "epoch": 9.968545216251638, + "grad_norm": 2.5204086303710938, + "learning_rate": 1.532152377678453e-09, + "loss": 0.6039, + "num_input_tokens_seen": 10182752, + "step": 38030 + }, + { + "epoch": 9.969855832241153, + "grad_norm": 6.685367107391357, + "learning_rate": 1.4081455801145948e-09, + "loss": 0.4691, + "num_input_tokens_seen": 10184032, + "step": 38035 + }, + { + "epoch": 9.971166448230669, + "grad_norm": 2.386237621307373, + "learning_rate": 1.289370944629331e-09, + "loss": 0.4383, + "num_input_tokens_seen": 10186016, + "step": 38040 + }, + { + "epoch": 9.972477064220184, + "grad_norm": 1.2985855340957642, + "learning_rate": 1.1758284960861066e-09, + "loss": 0.5702, + "num_input_tokens_seen": 10187632, + "step": 38045 + }, + { + "epoch": 9.973787680209698, + "grad_norm": 5.091379165649414, + "learning_rate": 1.0675182582464693e-09, + "loss": 0.4261, + "num_input_tokens_seen": 10188736, + "step": 38050 + }, + { + "epoch": 9.975098296199214, + "grad_norm": 5.778327941894531, + "learning_rate": 9.644402537811736e-10, + "loss": 0.4982, + "num_input_tokens_seen": 10190368, + "step": 38055 + }, + { + "epoch": 9.97640891218873, + "grad_norm": 2.153388023376465, + "learning_rate": 8.665945042618529e-10, + "loss": 0.4371, + "num_input_tokens_seen": 10191504, + "step": 38060 + }, + { + "epoch": 9.977719528178245, + "grad_norm": 6.889730453491211, + "learning_rate": 7.739810301693462e-10, + "loss": 0.471, + "num_input_tokens_seen": 10192816, + "step": 38065 + }, + { + "epoch": 9.979030144167758, + "grad_norm": 31.94826316833496, + "learning_rate": 6.865998508881477e-10, + "loss": 0.5922, + "num_input_tokens_seen": 10193808, + "step": 38070 + }, + { + "epoch": 9.980340760157274, + "grad_norm": 5.713347434997559, + "learning_rate": 6.044509847064062e-10, + "loss": 0.666, + "num_input_tokens_seen": 10194784, + "step": 38075 + }, + { + "epoch": 9.98165137614679, + "grad_norm": 4.347081661224365, + "learning_rate": 5.275344488187006e-10, + "loss": 0.5533, + "num_input_tokens_seen": 10195936, + "step": 38080 + }, + { + "epoch": 9.982961992136303, + "grad_norm": 3.640455484390259, + "learning_rate": 4.5585025932048943e-10, + "loss": 0.4056, + "num_input_tokens_seen": 10197312, + "step": 38085 + }, + { + "epoch": 9.984272608125819, + "grad_norm": 5.773710250854492, + "learning_rate": 3.893984312164367e-10, + "loss": 0.5213, + "num_input_tokens_seen": 10198608, + "step": 38090 + }, + { + "epoch": 9.985583224115334, + "grad_norm": 4.293521404266357, + "learning_rate": 3.28178978417637e-10, + "loss": 0.6512, + "num_input_tokens_seen": 10200032, + "step": 38095 + }, + { + "epoch": 9.98689384010485, + "grad_norm": 3.6116082668304443, + "learning_rate": 2.7219191373328846e-10, + "loss": 0.4037, + "num_input_tokens_seen": 10201680, + "step": 38100 + }, + { + "epoch": 9.988204456094364, + "grad_norm": 2.0918591022491455, + "learning_rate": 2.2143724888179506e-10, + "loss": 0.4281, + "num_input_tokens_seen": 10203056, + "step": 38105 + }, + { + "epoch": 9.98951507208388, + "grad_norm": 6.9108123779296875, + "learning_rate": 1.759149944879912e-10, + "loss": 0.4263, + "num_input_tokens_seen": 10204032, + "step": 38110 + }, + { + "epoch": 9.990825688073395, + "grad_norm": 2.158986806869507, + "learning_rate": 1.3562516008036597e-10, + "loss": 0.3833, + "num_input_tokens_seen": 10205488, + "step": 38115 + }, + { + "epoch": 9.99213630406291, + "grad_norm": 13.370691299438477, + "learning_rate": 1.0056775408828767e-10, + "loss": 0.4283, + "num_input_tokens_seen": 10206608, + "step": 38120 + }, + { + "epoch": 9.993446920052424, + "grad_norm": 13.357122421264648, + "learning_rate": 7.07427838503305e-11, + "loss": 0.6102, + "num_input_tokens_seen": 10207808, + "step": 38125 + }, + { + "epoch": 9.99475753604194, + "grad_norm": 2.988194227218628, + "learning_rate": 4.6150255611498996e-11, + "loss": 0.4566, + "num_input_tokens_seen": 10208976, + "step": 38130 + }, + { + "epoch": 9.996068152031455, + "grad_norm": 4.48499059677124, + "learning_rate": 2.679017451490129e-11, + "loss": 0.4748, + "num_input_tokens_seen": 10210128, + "step": 38135 + }, + { + "epoch": 9.997378768020969, + "grad_norm": 2.8613994121551514, + "learning_rate": 1.2662544615626993e-11, + "loss": 0.4218, + "num_input_tokens_seen": 10211488, + "step": 38140 + }, + { + "epoch": 9.998689384010484, + "grad_norm": 2.747965097427368, + "learning_rate": 3.767368869644905e-12, + "loss": 0.517, + "num_input_tokens_seen": 10213040, + "step": 38145 + }, + { + "epoch": 10.0, + "grad_norm": 5.274493217468262, + "learning_rate": 1.0464913657859399e-13, + "loss": 0.4409, + "num_input_tokens_seen": 10214104, + "step": 38150 + }, + { + "epoch": 10.0, + "num_input_tokens_seen": 10214104, + "step": 38150, + "total_flos": 5.0107394251771085e+17, + "train_loss": 0.5066599989749814, + "train_runtime": 5951.357, + "train_samples_per_second": 12.819, + "train_steps_per_second": 6.41 + } + ], + "logging_steps": 5, + "max_steps": 38150, + "num_input_tokens_seen": 10214104, + "num_train_epochs": 10, + "save_steps": 1908, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 5.0107394251771085e+17, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}