{ "best_global_step": 7632, "best_metric": 0.49409183859825134, "best_model_checkpoint": "saves/prefix-tuning/llama-3-8b-instruct/train_codealpacapy_1756735779/checkpoint-7632", "epoch": 10.0, "eval_steps": 1908, "global_step": 38150, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.001310615989515072, "grad_norm": 78.85181427001953, "learning_rate": 5.2424639580602885e-08, "loss": 6.7988, "num_input_tokens_seen": 1184, "step": 5 }, { "epoch": 0.002621231979030144, "grad_norm": 41.020729064941406, "learning_rate": 1.1795543905635651e-07, "loss": 6.4596, "num_input_tokens_seen": 2448, "step": 10 }, { "epoch": 0.003931847968545216, "grad_norm": 41.45704650878906, "learning_rate": 1.8348623853211012e-07, "loss": 6.3352, "num_input_tokens_seen": 4128, "step": 15 }, { "epoch": 0.005242463958060288, "grad_norm": 81.63996887207031, "learning_rate": 2.490170380078637e-07, "loss": 6.3743, "num_input_tokens_seen": 5664, "step": 20 }, { "epoch": 0.00655307994757536, "grad_norm": 64.59480285644531, "learning_rate": 3.1454783748361734e-07, "loss": 6.9953, "num_input_tokens_seen": 7040, "step": 25 }, { "epoch": 0.007863695937090432, "grad_norm": 66.04780578613281, "learning_rate": 3.8007863695937093e-07, "loss": 6.254, "num_input_tokens_seen": 8352, "step": 30 }, { "epoch": 0.009174311926605505, "grad_norm": 44.92106628417969, "learning_rate": 4.4560943643512453e-07, "loss": 6.3418, "num_input_tokens_seen": 9680, "step": 35 }, { "epoch": 0.010484927916120577, "grad_norm": 39.61148452758789, "learning_rate": 5.111402359108782e-07, "loss": 6.0337, "num_input_tokens_seen": 11424, "step": 40 }, { "epoch": 0.011795543905635648, "grad_norm": 69.95035552978516, "learning_rate": 5.766710353866317e-07, "loss": 6.127, "num_input_tokens_seen": 12544, "step": 45 }, { "epoch": 0.01310615989515072, "grad_norm": 59.79043960571289, "learning_rate": 6.422018348623854e-07, "loss": 5.9314, "num_input_tokens_seen": 14112, "step": 50 }, { "epoch": 0.014416775884665793, "grad_norm": 41.60962677001953, "learning_rate": 7.07732634338139e-07, "loss": 6.1952, "num_input_tokens_seen": 15232, "step": 55 }, { "epoch": 0.015727391874180863, "grad_norm": 53.85437774658203, "learning_rate": 7.732634338138926e-07, "loss": 5.5476, "num_input_tokens_seen": 16816, "step": 60 }, { "epoch": 0.01703800786369594, "grad_norm": 52.799495697021484, "learning_rate": 8.387942332896462e-07, "loss": 5.1698, "num_input_tokens_seen": 17760, "step": 65 }, { "epoch": 0.01834862385321101, "grad_norm": 47.800113677978516, "learning_rate": 9.043250327653998e-07, "loss": 4.9766, "num_input_tokens_seen": 18976, "step": 70 }, { "epoch": 0.019659239842726082, "grad_norm": 42.85136795043945, "learning_rate": 9.698558322411533e-07, "loss": 4.9762, "num_input_tokens_seen": 20368, "step": 75 }, { "epoch": 0.020969855832241154, "grad_norm": 43.31123352050781, "learning_rate": 1.035386631716907e-06, "loss": 4.6494, "num_input_tokens_seen": 21456, "step": 80 }, { "epoch": 0.022280471821756225, "grad_norm": 67.13323974609375, "learning_rate": 1.1009174311926608e-06, "loss": 4.7273, "num_input_tokens_seen": 22592, "step": 85 }, { "epoch": 0.023591087811271297, "grad_norm": 35.138458251953125, "learning_rate": 1.1664482306684142e-06, "loss": 4.7757, "num_input_tokens_seen": 23728, "step": 90 }, { "epoch": 0.02490170380078637, "grad_norm": 39.420166015625, "learning_rate": 1.2319790301441677e-06, "loss": 4.1652, "num_input_tokens_seen": 24864, "step": 95 }, { "epoch": 0.02621231979030144, "grad_norm": 32.41032409667969, "learning_rate": 1.2975098296199214e-06, "loss": 4.319, "num_input_tokens_seen": 26272, "step": 100 }, { "epoch": 0.027522935779816515, "grad_norm": 31.285568237304688, "learning_rate": 1.363040629095675e-06, "loss": 4.355, "num_input_tokens_seen": 27248, "step": 105 }, { "epoch": 0.028833551769331587, "grad_norm": 29.255451202392578, "learning_rate": 1.4285714285714286e-06, "loss": 3.3366, "num_input_tokens_seen": 28688, "step": 110 }, { "epoch": 0.03014416775884666, "grad_norm": 25.10015106201172, "learning_rate": 1.4941022280471821e-06, "loss": 4.3782, "num_input_tokens_seen": 29792, "step": 115 }, { "epoch": 0.03145478374836173, "grad_norm": 24.30801010131836, "learning_rate": 1.559633027522936e-06, "loss": 3.5134, "num_input_tokens_seen": 31424, "step": 120 }, { "epoch": 0.0327653997378768, "grad_norm": 36.23066329956055, "learning_rate": 1.6251638269986893e-06, "loss": 3.8634, "num_input_tokens_seen": 32384, "step": 125 }, { "epoch": 0.03407601572739188, "grad_norm": 34.126651763916016, "learning_rate": 1.690694626474443e-06, "loss": 3.898, "num_input_tokens_seen": 33808, "step": 130 }, { "epoch": 0.035386631716906945, "grad_norm": 56.42647933959961, "learning_rate": 1.7562254259501965e-06, "loss": 3.6393, "num_input_tokens_seen": 34912, "step": 135 }, { "epoch": 0.03669724770642202, "grad_norm": 20.71720314025879, "learning_rate": 1.8217562254259502e-06, "loss": 3.5168, "num_input_tokens_seen": 36576, "step": 140 }, { "epoch": 0.03800786369593709, "grad_norm": 34.7415771484375, "learning_rate": 1.8872870249017041e-06, "loss": 3.6464, "num_input_tokens_seen": 37664, "step": 145 }, { "epoch": 0.039318479685452164, "grad_norm": 45.20863342285156, "learning_rate": 1.9528178243774574e-06, "loss": 3.6348, "num_input_tokens_seen": 38816, "step": 150 }, { "epoch": 0.04062909567496723, "grad_norm": 32.540260314941406, "learning_rate": 2.0183486238532113e-06, "loss": 3.2512, "num_input_tokens_seen": 40272, "step": 155 }, { "epoch": 0.04193971166448231, "grad_norm": 54.34537124633789, "learning_rate": 2.083879423328965e-06, "loss": 3.3999, "num_input_tokens_seen": 41392, "step": 160 }, { "epoch": 0.04325032765399738, "grad_norm": 22.871353149414062, "learning_rate": 2.1494102228047183e-06, "loss": 3.6457, "num_input_tokens_seen": 43152, "step": 165 }, { "epoch": 0.04456094364351245, "grad_norm": 18.423627853393555, "learning_rate": 2.2149410222804718e-06, "loss": 3.3635, "num_input_tokens_seen": 44992, "step": 170 }, { "epoch": 0.045871559633027525, "grad_norm": 33.693145751953125, "learning_rate": 2.2804718217562257e-06, "loss": 3.0724, "num_input_tokens_seen": 46192, "step": 175 }, { "epoch": 0.047182175622542594, "grad_norm": 52.8044319152832, "learning_rate": 2.346002621231979e-06, "loss": 3.4663, "num_input_tokens_seen": 47216, "step": 180 }, { "epoch": 0.04849279161205767, "grad_norm": 26.974430084228516, "learning_rate": 2.4115334207077327e-06, "loss": 2.649, "num_input_tokens_seen": 48944, "step": 185 }, { "epoch": 0.04980340760157274, "grad_norm": 30.683292388916016, "learning_rate": 2.4770642201834866e-06, "loss": 2.7592, "num_input_tokens_seen": 50288, "step": 190 }, { "epoch": 0.05111402359108781, "grad_norm": 24.398881912231445, "learning_rate": 2.54259501965924e-06, "loss": 2.583, "num_input_tokens_seen": 51552, "step": 195 }, { "epoch": 0.05242463958060288, "grad_norm": 26.586990356445312, "learning_rate": 2.6081258191349936e-06, "loss": 3.2712, "num_input_tokens_seen": 53328, "step": 200 }, { "epoch": 0.053735255570117955, "grad_norm": 32.74496841430664, "learning_rate": 2.673656618610747e-06, "loss": 2.5977, "num_input_tokens_seen": 54640, "step": 205 }, { "epoch": 0.05504587155963303, "grad_norm": 32.16419982910156, "learning_rate": 2.739187418086501e-06, "loss": 2.747, "num_input_tokens_seen": 55776, "step": 210 }, { "epoch": 0.0563564875491481, "grad_norm": 18.71208953857422, "learning_rate": 2.8047182175622545e-06, "loss": 2.5956, "num_input_tokens_seen": 57216, "step": 215 }, { "epoch": 0.057667103538663174, "grad_norm": 12.590314865112305, "learning_rate": 2.870249017038008e-06, "loss": 2.7148, "num_input_tokens_seen": 59360, "step": 220 }, { "epoch": 0.05897771952817824, "grad_norm": 26.79210662841797, "learning_rate": 2.935779816513762e-06, "loss": 2.8768, "num_input_tokens_seen": 60384, "step": 225 }, { "epoch": 0.06028833551769332, "grad_norm": 31.221885681152344, "learning_rate": 3.0013106159895154e-06, "loss": 2.4627, "num_input_tokens_seen": 61888, "step": 230 }, { "epoch": 0.061598951507208385, "grad_norm": 40.7152214050293, "learning_rate": 3.066841415465269e-06, "loss": 2.2571, "num_input_tokens_seen": 62800, "step": 235 }, { "epoch": 0.06290956749672345, "grad_norm": 23.824317932128906, "learning_rate": 3.1323722149410228e-06, "loss": 2.3205, "num_input_tokens_seen": 64256, "step": 240 }, { "epoch": 0.06422018348623854, "grad_norm": 34.902645111083984, "learning_rate": 3.1979030144167763e-06, "loss": 2.4592, "num_input_tokens_seen": 65728, "step": 245 }, { "epoch": 0.0655307994757536, "grad_norm": 18.403425216674805, "learning_rate": 3.2634338138925293e-06, "loss": 2.6116, "num_input_tokens_seen": 66912, "step": 250 }, { "epoch": 0.06684141546526867, "grad_norm": 26.418415069580078, "learning_rate": 3.328964613368283e-06, "loss": 2.2267, "num_input_tokens_seen": 67904, "step": 255 }, { "epoch": 0.06815203145478375, "grad_norm": 22.388019561767578, "learning_rate": 3.394495412844037e-06, "loss": 2.3389, "num_input_tokens_seen": 69456, "step": 260 }, { "epoch": 0.06946264744429882, "grad_norm": 22.1889705657959, "learning_rate": 3.4600262123197906e-06, "loss": 2.8775, "num_input_tokens_seen": 70832, "step": 265 }, { "epoch": 0.07077326343381389, "grad_norm": 19.02516746520996, "learning_rate": 3.5255570117955437e-06, "loss": 2.7468, "num_input_tokens_seen": 72448, "step": 270 }, { "epoch": 0.07208387942332896, "grad_norm": 23.846927642822266, "learning_rate": 3.591087811271298e-06, "loss": 2.5026, "num_input_tokens_seen": 73920, "step": 275 }, { "epoch": 0.07339449541284404, "grad_norm": 53.351097106933594, "learning_rate": 3.6566186107470515e-06, "loss": 2.5371, "num_input_tokens_seen": 75472, "step": 280 }, { "epoch": 0.07470511140235911, "grad_norm": 15.377866744995117, "learning_rate": 3.7221494102228046e-06, "loss": 2.404, "num_input_tokens_seen": 76976, "step": 285 }, { "epoch": 0.07601572739187418, "grad_norm": 21.78667640686035, "learning_rate": 3.787680209698558e-06, "loss": 2.3197, "num_input_tokens_seen": 78496, "step": 290 }, { "epoch": 0.07732634338138926, "grad_norm": 26.924175262451172, "learning_rate": 3.853211009174312e-06, "loss": 1.9647, "num_input_tokens_seen": 80032, "step": 295 }, { "epoch": 0.07863695937090433, "grad_norm": 23.564693450927734, "learning_rate": 3.918741808650066e-06, "loss": 1.6902, "num_input_tokens_seen": 81104, "step": 300 }, { "epoch": 0.0799475753604194, "grad_norm": 40.69663619995117, "learning_rate": 3.984272608125819e-06, "loss": 1.713, "num_input_tokens_seen": 82224, "step": 305 }, { "epoch": 0.08125819134993446, "grad_norm": 27.559953689575195, "learning_rate": 4.049803407601573e-06, "loss": 1.7123, "num_input_tokens_seen": 84160, "step": 310 }, { "epoch": 0.08256880733944955, "grad_norm": 22.61702537536621, "learning_rate": 4.115334207077327e-06, "loss": 2.0062, "num_input_tokens_seen": 85488, "step": 315 }, { "epoch": 0.08387942332896461, "grad_norm": 19.799457550048828, "learning_rate": 4.18086500655308e-06, "loss": 2.202, "num_input_tokens_seen": 86768, "step": 320 }, { "epoch": 0.08519003931847968, "grad_norm": 19.331878662109375, "learning_rate": 4.246395806028834e-06, "loss": 1.6681, "num_input_tokens_seen": 87920, "step": 325 }, { "epoch": 0.08650065530799476, "grad_norm": 22.45851707458496, "learning_rate": 4.311926605504588e-06, "loss": 2.5665, "num_input_tokens_seen": 89136, "step": 330 }, { "epoch": 0.08781127129750983, "grad_norm": 18.617366790771484, "learning_rate": 4.377457404980341e-06, "loss": 1.9125, "num_input_tokens_seen": 90464, "step": 335 }, { "epoch": 0.0891218872870249, "grad_norm": 17.070053100585938, "learning_rate": 4.442988204456095e-06, "loss": 1.8467, "num_input_tokens_seen": 91984, "step": 340 }, { "epoch": 0.09043250327653997, "grad_norm": 25.171110153198242, "learning_rate": 4.508519003931848e-06, "loss": 1.3421, "num_input_tokens_seen": 93024, "step": 345 }, { "epoch": 0.09174311926605505, "grad_norm": 47.312095642089844, "learning_rate": 4.574049803407602e-06, "loss": 1.5021, "num_input_tokens_seen": 94464, "step": 350 }, { "epoch": 0.09305373525557012, "grad_norm": 19.13548469543457, "learning_rate": 4.639580602883356e-06, "loss": 1.774, "num_input_tokens_seen": 95664, "step": 355 }, { "epoch": 0.09436435124508519, "grad_norm": 48.80869674682617, "learning_rate": 4.705111402359109e-06, "loss": 1.4325, "num_input_tokens_seen": 96608, "step": 360 }, { "epoch": 0.09567496723460026, "grad_norm": 12.515359878540039, "learning_rate": 4.7706422018348626e-06, "loss": 1.333, "num_input_tokens_seen": 97968, "step": 365 }, { "epoch": 0.09698558322411534, "grad_norm": 20.451143264770508, "learning_rate": 4.8361730013106165e-06, "loss": 1.1737, "num_input_tokens_seen": 99152, "step": 370 }, { "epoch": 0.0982961992136304, "grad_norm": 18.291444778442383, "learning_rate": 4.9017038007863695e-06, "loss": 1.253, "num_input_tokens_seen": 100592, "step": 375 }, { "epoch": 0.09960681520314547, "grad_norm": 18.427112579345703, "learning_rate": 4.9672346002621235e-06, "loss": 1.047, "num_input_tokens_seen": 101856, "step": 380 }, { "epoch": 0.10091743119266056, "grad_norm": 9.26123046875, "learning_rate": 5.032765399737877e-06, "loss": 1.0784, "num_input_tokens_seen": 103024, "step": 385 }, { "epoch": 0.10222804718217562, "grad_norm": 20.02321434020996, "learning_rate": 5.0982961992136304e-06, "loss": 0.7362, "num_input_tokens_seen": 104512, "step": 390 }, { "epoch": 0.10353866317169069, "grad_norm": 12.981529235839844, "learning_rate": 5.163826998689384e-06, "loss": 0.7474, "num_input_tokens_seen": 105904, "step": 395 }, { "epoch": 0.10484927916120576, "grad_norm": 12.749576568603516, "learning_rate": 5.229357798165138e-06, "loss": 0.8223, "num_input_tokens_seen": 107088, "step": 400 }, { "epoch": 0.10615989515072084, "grad_norm": 11.017773628234863, "learning_rate": 5.294888597640891e-06, "loss": 1.0916, "num_input_tokens_seen": 108240, "step": 405 }, { "epoch": 0.10747051114023591, "grad_norm": 6.939527988433838, "learning_rate": 5.360419397116645e-06, "loss": 0.814, "num_input_tokens_seen": 109472, "step": 410 }, { "epoch": 0.10878112712975098, "grad_norm": 11.323343276977539, "learning_rate": 5.425950196592398e-06, "loss": 0.861, "num_input_tokens_seen": 111056, "step": 415 }, { "epoch": 0.11009174311926606, "grad_norm": 10.066669464111328, "learning_rate": 5.491480996068152e-06, "loss": 0.696, "num_input_tokens_seen": 112448, "step": 420 }, { "epoch": 0.11140235910878113, "grad_norm": 10.207267761230469, "learning_rate": 5.557011795543906e-06, "loss": 0.8794, "num_input_tokens_seen": 113648, "step": 425 }, { "epoch": 0.1127129750982962, "grad_norm": 6.772650241851807, "learning_rate": 5.622542595019659e-06, "loss": 1.2266, "num_input_tokens_seen": 114864, "step": 430 }, { "epoch": 0.11402359108781127, "grad_norm": 24.065044403076172, "learning_rate": 5.688073394495413e-06, "loss": 0.8133, "num_input_tokens_seen": 115968, "step": 435 }, { "epoch": 0.11533420707732635, "grad_norm": 8.515714645385742, "learning_rate": 5.753604193971167e-06, "loss": 0.8083, "num_input_tokens_seen": 117680, "step": 440 }, { "epoch": 0.11664482306684142, "grad_norm": 9.364624977111816, "learning_rate": 5.81913499344692e-06, "loss": 0.7652, "num_input_tokens_seen": 118720, "step": 445 }, { "epoch": 0.11795543905635648, "grad_norm": 5.285806179046631, "learning_rate": 5.884665792922674e-06, "loss": 0.5732, "num_input_tokens_seen": 120288, "step": 450 }, { "epoch": 0.11926605504587157, "grad_norm": 53.94522476196289, "learning_rate": 5.950196592398428e-06, "loss": 0.6981, "num_input_tokens_seen": 121728, "step": 455 }, { "epoch": 0.12057667103538663, "grad_norm": 8.656790733337402, "learning_rate": 6.015727391874181e-06, "loss": 0.6557, "num_input_tokens_seen": 122880, "step": 460 }, { "epoch": 0.1218872870249017, "grad_norm": 11.836997985839844, "learning_rate": 6.081258191349935e-06, "loss": 0.7686, "num_input_tokens_seen": 124112, "step": 465 }, { "epoch": 0.12319790301441677, "grad_norm": 12.190650939941406, "learning_rate": 6.146788990825689e-06, "loss": 0.939, "num_input_tokens_seen": 125568, "step": 470 }, { "epoch": 0.12450851900393185, "grad_norm": 5.932990550994873, "learning_rate": 6.212319790301442e-06, "loss": 0.6379, "num_input_tokens_seen": 127216, "step": 475 }, { "epoch": 0.1258191349934469, "grad_norm": 9.715404510498047, "learning_rate": 6.277850589777196e-06, "loss": 0.9168, "num_input_tokens_seen": 128560, "step": 480 }, { "epoch": 0.127129750982962, "grad_norm": 9.483189582824707, "learning_rate": 6.343381389252949e-06, "loss": 0.6431, "num_input_tokens_seen": 129952, "step": 485 }, { "epoch": 0.12844036697247707, "grad_norm": 6.418022632598877, "learning_rate": 6.408912188728703e-06, "loss": 0.8309, "num_input_tokens_seen": 130976, "step": 490 }, { "epoch": 0.12975098296199214, "grad_norm": 8.721658706665039, "learning_rate": 6.474442988204456e-06, "loss": 0.4063, "num_input_tokens_seen": 132256, "step": 495 }, { "epoch": 0.1310615989515072, "grad_norm": 6.9230055809021, "learning_rate": 6.539973787680211e-06, "loss": 0.837, "num_input_tokens_seen": 134256, "step": 500 }, { "epoch": 0.13237221494102228, "grad_norm": 5.154301166534424, "learning_rate": 6.605504587155964e-06, "loss": 0.5043, "num_input_tokens_seen": 135584, "step": 505 }, { "epoch": 0.13368283093053734, "grad_norm": 5.893485069274902, "learning_rate": 6.671035386631718e-06, "loss": 0.6325, "num_input_tokens_seen": 136576, "step": 510 }, { "epoch": 0.1349934469200524, "grad_norm": 11.127357482910156, "learning_rate": 6.736566186107471e-06, "loss": 0.5974, "num_input_tokens_seen": 137952, "step": 515 }, { "epoch": 0.1363040629095675, "grad_norm": 6.57765531539917, "learning_rate": 6.8020969855832246e-06, "loss": 0.5231, "num_input_tokens_seen": 139024, "step": 520 }, { "epoch": 0.13761467889908258, "grad_norm": 11.020800590515137, "learning_rate": 6.867627785058978e-06, "loss": 0.5588, "num_input_tokens_seen": 140464, "step": 525 }, { "epoch": 0.13892529488859764, "grad_norm": 28.254995346069336, "learning_rate": 6.933158584534731e-06, "loss": 0.5278, "num_input_tokens_seen": 141824, "step": 530 }, { "epoch": 0.1402359108781127, "grad_norm": 10.407012939453125, "learning_rate": 6.9986893840104855e-06, "loss": 0.601, "num_input_tokens_seen": 142864, "step": 535 }, { "epoch": 0.14154652686762778, "grad_norm": 8.060755729675293, "learning_rate": 7.064220183486239e-06, "loss": 0.8334, "num_input_tokens_seen": 143856, "step": 540 }, { "epoch": 0.14285714285714285, "grad_norm": 6.0524091720581055, "learning_rate": 7.1297509829619924e-06, "loss": 0.5285, "num_input_tokens_seen": 145248, "step": 545 }, { "epoch": 0.14416775884665792, "grad_norm": 3.398391008377075, "learning_rate": 7.195281782437746e-06, "loss": 0.5365, "num_input_tokens_seen": 146912, "step": 550 }, { "epoch": 0.145478374836173, "grad_norm": 7.9956440925598145, "learning_rate": 7.260812581913499e-06, "loss": 0.5778, "num_input_tokens_seen": 148224, "step": 555 }, { "epoch": 0.14678899082568808, "grad_norm": 6.1303911209106445, "learning_rate": 7.326343381389253e-06, "loss": 0.5417, "num_input_tokens_seen": 149568, "step": 560 }, { "epoch": 0.14809960681520315, "grad_norm": 10.488219261169434, "learning_rate": 7.391874180865006e-06, "loss": 0.6856, "num_input_tokens_seen": 150912, "step": 565 }, { "epoch": 0.14941022280471822, "grad_norm": 4.999405860900879, "learning_rate": 7.457404980340761e-06, "loss": 0.6477, "num_input_tokens_seen": 152352, "step": 570 }, { "epoch": 0.15072083879423329, "grad_norm": 4.705912113189697, "learning_rate": 7.522935779816514e-06, "loss": 0.4554, "num_input_tokens_seen": 153728, "step": 575 }, { "epoch": 0.15203145478374835, "grad_norm": 33.82322692871094, "learning_rate": 7.588466579292268e-06, "loss": 1.0347, "num_input_tokens_seen": 155200, "step": 580 }, { "epoch": 0.15334207077326342, "grad_norm": 6.281925201416016, "learning_rate": 7.653997378768021e-06, "loss": 0.6114, "num_input_tokens_seen": 156288, "step": 585 }, { "epoch": 0.15465268676277852, "grad_norm": 23.61570930480957, "learning_rate": 7.719528178243775e-06, "loss": 0.6107, "num_input_tokens_seen": 157664, "step": 590 }, { "epoch": 0.1559633027522936, "grad_norm": 9.322854042053223, "learning_rate": 7.785058977719529e-06, "loss": 0.3819, "num_input_tokens_seen": 158832, "step": 595 }, { "epoch": 0.15727391874180865, "grad_norm": 11.59537410736084, "learning_rate": 7.850589777195281e-06, "loss": 0.4905, "num_input_tokens_seen": 159984, "step": 600 }, { "epoch": 0.15858453473132372, "grad_norm": 8.02344036102295, "learning_rate": 7.916120576671037e-06, "loss": 0.494, "num_input_tokens_seen": 161248, "step": 605 }, { "epoch": 0.1598951507208388, "grad_norm": 5.351619243621826, "learning_rate": 7.981651376146789e-06, "loss": 0.6348, "num_input_tokens_seen": 162608, "step": 610 }, { "epoch": 0.16120576671035386, "grad_norm": 6.902489185333252, "learning_rate": 8.047182175622543e-06, "loss": 0.4705, "num_input_tokens_seen": 163696, "step": 615 }, { "epoch": 0.16251638269986893, "grad_norm": 5.372810363769531, "learning_rate": 8.112712975098297e-06, "loss": 0.6646, "num_input_tokens_seen": 165056, "step": 620 }, { "epoch": 0.16382699868938402, "grad_norm": 12.597624778747559, "learning_rate": 8.17824377457405e-06, "loss": 0.4328, "num_input_tokens_seen": 166848, "step": 625 }, { "epoch": 0.1651376146788991, "grad_norm": 14.66284465789795, "learning_rate": 8.243774574049803e-06, "loss": 0.6292, "num_input_tokens_seen": 168240, "step": 630 }, { "epoch": 0.16644823066841416, "grad_norm": 7.103930950164795, "learning_rate": 8.309305373525557e-06, "loss": 0.7707, "num_input_tokens_seen": 169840, "step": 635 }, { "epoch": 0.16775884665792923, "grad_norm": 7.039085865020752, "learning_rate": 8.374836173001311e-06, "loss": 0.6284, "num_input_tokens_seen": 171152, "step": 640 }, { "epoch": 0.1690694626474443, "grad_norm": 7.620847702026367, "learning_rate": 8.440366972477065e-06, "loss": 0.7534, "num_input_tokens_seen": 172144, "step": 645 }, { "epoch": 0.17038007863695936, "grad_norm": 7.485767364501953, "learning_rate": 8.505897771952819e-06, "loss": 0.6067, "num_input_tokens_seen": 173440, "step": 650 }, { "epoch": 0.17169069462647443, "grad_norm": 20.503398895263672, "learning_rate": 8.571428571428573e-06, "loss": 0.6322, "num_input_tokens_seen": 174928, "step": 655 }, { "epoch": 0.17300131061598953, "grad_norm": 3.1943178176879883, "learning_rate": 8.636959370904325e-06, "loss": 0.5868, "num_input_tokens_seen": 176672, "step": 660 }, { "epoch": 0.1743119266055046, "grad_norm": 50.336544036865234, "learning_rate": 8.702490170380079e-06, "loss": 0.8532, "num_input_tokens_seen": 177712, "step": 665 }, { "epoch": 0.17562254259501967, "grad_norm": 7.113663673400879, "learning_rate": 8.768020969855833e-06, "loss": 0.7438, "num_input_tokens_seen": 179232, "step": 670 }, { "epoch": 0.17693315858453473, "grad_norm": 9.251459121704102, "learning_rate": 8.833551769331587e-06, "loss": 0.5196, "num_input_tokens_seen": 180688, "step": 675 }, { "epoch": 0.1782437745740498, "grad_norm": 3.8759820461273193, "learning_rate": 8.89908256880734e-06, "loss": 0.589, "num_input_tokens_seen": 181824, "step": 680 }, { "epoch": 0.17955439056356487, "grad_norm": 6.008286952972412, "learning_rate": 8.964613368283094e-06, "loss": 0.7088, "num_input_tokens_seen": 182768, "step": 685 }, { "epoch": 0.18086500655307994, "grad_norm": 4.4908671379089355, "learning_rate": 9.030144167758847e-06, "loss": 0.7538, "num_input_tokens_seen": 184240, "step": 690 }, { "epoch": 0.182175622542595, "grad_norm": 13.397222518920898, "learning_rate": 9.0956749672346e-06, "loss": 0.5687, "num_input_tokens_seen": 185200, "step": 695 }, { "epoch": 0.1834862385321101, "grad_norm": 5.166349411010742, "learning_rate": 9.161205766710354e-06, "loss": 0.43, "num_input_tokens_seen": 186464, "step": 700 }, { "epoch": 0.18479685452162517, "grad_norm": 3.4410459995269775, "learning_rate": 9.226736566186107e-06, "loss": 0.4524, "num_input_tokens_seen": 187728, "step": 705 }, { "epoch": 0.18610747051114024, "grad_norm": 9.359235763549805, "learning_rate": 9.29226736566186e-06, "loss": 0.5504, "num_input_tokens_seen": 188944, "step": 710 }, { "epoch": 0.1874180865006553, "grad_norm": 3.447437047958374, "learning_rate": 9.357798165137616e-06, "loss": 0.6473, "num_input_tokens_seen": 190288, "step": 715 }, { "epoch": 0.18872870249017037, "grad_norm": 8.461204528808594, "learning_rate": 9.423328964613368e-06, "loss": 0.4042, "num_input_tokens_seen": 191264, "step": 720 }, { "epoch": 0.19003931847968544, "grad_norm": 28.27265167236328, "learning_rate": 9.488859764089122e-06, "loss": 0.6551, "num_input_tokens_seen": 192336, "step": 725 }, { "epoch": 0.1913499344692005, "grad_norm": 4.53128719329834, "learning_rate": 9.554390563564876e-06, "loss": 0.5857, "num_input_tokens_seen": 193584, "step": 730 }, { "epoch": 0.1926605504587156, "grad_norm": 3.848435163497925, "learning_rate": 9.619921363040628e-06, "loss": 0.5511, "num_input_tokens_seen": 194800, "step": 735 }, { "epoch": 0.19397116644823068, "grad_norm": 16.699588775634766, "learning_rate": 9.685452162516382e-06, "loss": 0.5202, "num_input_tokens_seen": 196192, "step": 740 }, { "epoch": 0.19528178243774574, "grad_norm": 19.730575561523438, "learning_rate": 9.750982961992136e-06, "loss": 0.5277, "num_input_tokens_seen": 197200, "step": 745 }, { "epoch": 0.1965923984272608, "grad_norm": 7.053399562835693, "learning_rate": 9.81651376146789e-06, "loss": 0.7124, "num_input_tokens_seen": 198384, "step": 750 }, { "epoch": 0.19790301441677588, "grad_norm": 3.579075336456299, "learning_rate": 9.882044560943644e-06, "loss": 0.649, "num_input_tokens_seen": 199680, "step": 755 }, { "epoch": 0.19921363040629095, "grad_norm": 3.78265118598938, "learning_rate": 9.947575360419398e-06, "loss": 0.533, "num_input_tokens_seen": 200960, "step": 760 }, { "epoch": 0.20052424639580602, "grad_norm": 5.192812442779541, "learning_rate": 1.0013106159895152e-05, "loss": 0.6791, "num_input_tokens_seen": 203120, "step": 765 }, { "epoch": 0.2018348623853211, "grad_norm": 6.076825141906738, "learning_rate": 1.0078636959370904e-05, "loss": 0.7958, "num_input_tokens_seen": 204848, "step": 770 }, { "epoch": 0.20314547837483618, "grad_norm": 8.492985725402832, "learning_rate": 1.0144167758846658e-05, "loss": 0.746, "num_input_tokens_seen": 206080, "step": 775 }, { "epoch": 0.20445609436435125, "grad_norm": 2.90004301071167, "learning_rate": 1.0209698558322412e-05, "loss": 0.6857, "num_input_tokens_seen": 207920, "step": 780 }, { "epoch": 0.20576671035386632, "grad_norm": 3.7898874282836914, "learning_rate": 1.0275229357798166e-05, "loss": 0.675, "num_input_tokens_seen": 209040, "step": 785 }, { "epoch": 0.20707732634338138, "grad_norm": 2.094329357147217, "learning_rate": 1.034076015727392e-05, "loss": 0.5556, "num_input_tokens_seen": 210384, "step": 790 }, { "epoch": 0.20838794233289645, "grad_norm": 4.728037357330322, "learning_rate": 1.0406290956749674e-05, "loss": 0.6568, "num_input_tokens_seen": 212032, "step": 795 }, { "epoch": 0.20969855832241152, "grad_norm": 40.5416374206543, "learning_rate": 1.0471821756225426e-05, "loss": 0.6257, "num_input_tokens_seen": 212896, "step": 800 }, { "epoch": 0.21100917431192662, "grad_norm": 21.48481559753418, "learning_rate": 1.053735255570118e-05, "loss": 0.8714, "num_input_tokens_seen": 214352, "step": 805 }, { "epoch": 0.21231979030144169, "grad_norm": 8.870784759521484, "learning_rate": 1.0602883355176934e-05, "loss": 0.7145, "num_input_tokens_seen": 215664, "step": 810 }, { "epoch": 0.21363040629095675, "grad_norm": 5.430676460266113, "learning_rate": 1.0668414154652686e-05, "loss": 0.5441, "num_input_tokens_seen": 217104, "step": 815 }, { "epoch": 0.21494102228047182, "grad_norm": 8.41822338104248, "learning_rate": 1.0733944954128442e-05, "loss": 0.6288, "num_input_tokens_seen": 218224, "step": 820 }, { "epoch": 0.2162516382699869, "grad_norm": 6.499901294708252, "learning_rate": 1.0799475753604196e-05, "loss": 0.5544, "num_input_tokens_seen": 219584, "step": 825 }, { "epoch": 0.21756225425950196, "grad_norm": 2.438255548477173, "learning_rate": 1.0865006553079948e-05, "loss": 0.6436, "num_input_tokens_seen": 220912, "step": 830 }, { "epoch": 0.21887287024901703, "grad_norm": 14.680463790893555, "learning_rate": 1.0930537352555702e-05, "loss": 0.4972, "num_input_tokens_seen": 222272, "step": 835 }, { "epoch": 0.22018348623853212, "grad_norm": 2.9180948734283447, "learning_rate": 1.0996068152031456e-05, "loss": 0.7088, "num_input_tokens_seen": 223600, "step": 840 }, { "epoch": 0.2214941022280472, "grad_norm": 3.4285786151885986, "learning_rate": 1.1061598951507208e-05, "loss": 0.8075, "num_input_tokens_seen": 225232, "step": 845 }, { "epoch": 0.22280471821756226, "grad_norm": 8.85015869140625, "learning_rate": 1.1127129750982962e-05, "loss": 0.5162, "num_input_tokens_seen": 226512, "step": 850 }, { "epoch": 0.22411533420707733, "grad_norm": 8.431280136108398, "learning_rate": 1.1192660550458717e-05, "loss": 0.7515, "num_input_tokens_seen": 227552, "step": 855 }, { "epoch": 0.2254259501965924, "grad_norm": 6.236743927001953, "learning_rate": 1.125819134993447e-05, "loss": 0.6401, "num_input_tokens_seen": 228608, "step": 860 }, { "epoch": 0.22673656618610746, "grad_norm": 0.9077622294425964, "learning_rate": 1.1323722149410223e-05, "loss": 0.3691, "num_input_tokens_seen": 230192, "step": 865 }, { "epoch": 0.22804718217562253, "grad_norm": 2.3008577823638916, "learning_rate": 1.1389252948885977e-05, "loss": 0.5598, "num_input_tokens_seen": 231776, "step": 870 }, { "epoch": 0.22935779816513763, "grad_norm": 5.630849838256836, "learning_rate": 1.145478374836173e-05, "loss": 0.7994, "num_input_tokens_seen": 233104, "step": 875 }, { "epoch": 0.2306684141546527, "grad_norm": 14.09296703338623, "learning_rate": 1.1520314547837483e-05, "loss": 0.5982, "num_input_tokens_seen": 234128, "step": 880 }, { "epoch": 0.23197903014416776, "grad_norm": 5.670069694519043, "learning_rate": 1.1585845347313237e-05, "loss": 0.641, "num_input_tokens_seen": 235312, "step": 885 }, { "epoch": 0.23328964613368283, "grad_norm": 8.877077102661133, "learning_rate": 1.1651376146788991e-05, "loss": 0.5376, "num_input_tokens_seen": 236896, "step": 890 }, { "epoch": 0.2346002621231979, "grad_norm": 2.224792718887329, "learning_rate": 1.1716906946264745e-05, "loss": 0.6308, "num_input_tokens_seen": 237936, "step": 895 }, { "epoch": 0.23591087811271297, "grad_norm": 3.6192586421966553, "learning_rate": 1.1782437745740499e-05, "loss": 0.684, "num_input_tokens_seen": 239552, "step": 900 }, { "epoch": 0.23722149410222804, "grad_norm": 5.991044521331787, "learning_rate": 1.1847968545216253e-05, "loss": 0.5091, "num_input_tokens_seen": 240832, "step": 905 }, { "epoch": 0.23853211009174313, "grad_norm": 4.697566509246826, "learning_rate": 1.1913499344692005e-05, "loss": 0.6124, "num_input_tokens_seen": 242192, "step": 910 }, { "epoch": 0.2398427260812582, "grad_norm": 5.229562759399414, "learning_rate": 1.197903014416776e-05, "loss": 0.628, "num_input_tokens_seen": 243456, "step": 915 }, { "epoch": 0.24115334207077327, "grad_norm": 2.5291748046875, "learning_rate": 1.2044560943643513e-05, "loss": 0.7448, "num_input_tokens_seen": 245424, "step": 920 }, { "epoch": 0.24246395806028834, "grad_norm": 2.4825387001037598, "learning_rate": 1.2110091743119267e-05, "loss": 0.5062, "num_input_tokens_seen": 246752, "step": 925 }, { "epoch": 0.2437745740498034, "grad_norm": 2.8981821537017822, "learning_rate": 1.2175622542595021e-05, "loss": 0.5031, "num_input_tokens_seen": 248256, "step": 930 }, { "epoch": 0.24508519003931847, "grad_norm": 7.413928985595703, "learning_rate": 1.2241153342070775e-05, "loss": 0.5012, "num_input_tokens_seen": 249648, "step": 935 }, { "epoch": 0.24639580602883354, "grad_norm": 2.647071361541748, "learning_rate": 1.2306684141546527e-05, "loss": 0.6731, "num_input_tokens_seen": 251600, "step": 940 }, { "epoch": 0.24770642201834864, "grad_norm": 2.396348476409912, "learning_rate": 1.2372214941022281e-05, "loss": 0.4135, "num_input_tokens_seen": 252544, "step": 945 }, { "epoch": 0.2490170380078637, "grad_norm": 2.595533609390259, "learning_rate": 1.2437745740498035e-05, "loss": 0.7714, "num_input_tokens_seen": 253776, "step": 950 }, { "epoch": 0.2503276539973788, "grad_norm": 5.056262016296387, "learning_rate": 1.2503276539973787e-05, "loss": 0.4239, "num_input_tokens_seen": 255232, "step": 955 }, { "epoch": 0.2516382699868938, "grad_norm": 5.719480514526367, "learning_rate": 1.2568807339449543e-05, "loss": 0.6549, "num_input_tokens_seen": 256640, "step": 960 }, { "epoch": 0.2529488859764089, "grad_norm": 9.946965217590332, "learning_rate": 1.2634338138925295e-05, "loss": 0.5892, "num_input_tokens_seen": 257760, "step": 965 }, { "epoch": 0.254259501965924, "grad_norm": 3.3111963272094727, "learning_rate": 1.2699868938401049e-05, "loss": 0.5002, "num_input_tokens_seen": 258960, "step": 970 }, { "epoch": 0.25557011795543905, "grad_norm": 5.54409122467041, "learning_rate": 1.2765399737876801e-05, "loss": 0.471, "num_input_tokens_seen": 260528, "step": 975 }, { "epoch": 0.25688073394495414, "grad_norm": 2.8781471252441406, "learning_rate": 1.2830930537352557e-05, "loss": 0.5201, "num_input_tokens_seen": 261744, "step": 980 }, { "epoch": 0.2581913499344692, "grad_norm": 3.819031238555908, "learning_rate": 1.289646133682831e-05, "loss": 0.7073, "num_input_tokens_seen": 262928, "step": 985 }, { "epoch": 0.2595019659239843, "grad_norm": 1.9269148111343384, "learning_rate": 1.2961992136304063e-05, "loss": 0.5288, "num_input_tokens_seen": 264384, "step": 990 }, { "epoch": 0.2608125819134993, "grad_norm": 32.780799865722656, "learning_rate": 1.3027522935779818e-05, "loss": 0.8788, "num_input_tokens_seen": 265392, "step": 995 }, { "epoch": 0.2621231979030144, "grad_norm": 11.850534439086914, "learning_rate": 1.309305373525557e-05, "loss": 0.9137, "num_input_tokens_seen": 266608, "step": 1000 }, { "epoch": 0.2634338138925295, "grad_norm": 2.8229050636291504, "learning_rate": 1.3158584534731325e-05, "loss": 0.5584, "num_input_tokens_seen": 267984, "step": 1005 }, { "epoch": 0.26474442988204455, "grad_norm": 4.865471839904785, "learning_rate": 1.3224115334207077e-05, "loss": 0.7446, "num_input_tokens_seen": 268944, "step": 1010 }, { "epoch": 0.26605504587155965, "grad_norm": 3.247163772583008, "learning_rate": 1.328964613368283e-05, "loss": 0.5295, "num_input_tokens_seen": 270672, "step": 1015 }, { "epoch": 0.2673656618610747, "grad_norm": 3.8408443927764893, "learning_rate": 1.3355176933158586e-05, "loss": 0.4811, "num_input_tokens_seen": 272208, "step": 1020 }, { "epoch": 0.2686762778505898, "grad_norm": 2.835538387298584, "learning_rate": 1.3420707732634339e-05, "loss": 0.605, "num_input_tokens_seen": 274720, "step": 1025 }, { "epoch": 0.2699868938401048, "grad_norm": 118.93071746826172, "learning_rate": 1.3486238532110092e-05, "loss": 2.3075, "num_input_tokens_seen": 275664, "step": 1030 }, { "epoch": 0.2712975098296199, "grad_norm": 3.809112071990967, "learning_rate": 1.3551769331585845e-05, "loss": 0.779, "num_input_tokens_seen": 277568, "step": 1035 }, { "epoch": 0.272608125819135, "grad_norm": 2.967628240585327, "learning_rate": 1.36173001310616e-05, "loss": 0.4939, "num_input_tokens_seen": 278848, "step": 1040 }, { "epoch": 0.27391874180865006, "grad_norm": 1.6521815061569214, "learning_rate": 1.3682830930537352e-05, "loss": 0.4899, "num_input_tokens_seen": 279968, "step": 1045 }, { "epoch": 0.27522935779816515, "grad_norm": 2.938727378845215, "learning_rate": 1.3748361730013106e-05, "loss": 0.8051, "num_input_tokens_seen": 281184, "step": 1050 }, { "epoch": 0.2765399737876802, "grad_norm": 2.6933324337005615, "learning_rate": 1.3813892529488862e-05, "loss": 0.4287, "num_input_tokens_seen": 282576, "step": 1055 }, { "epoch": 0.2778505897771953, "grad_norm": 2.861161231994629, "learning_rate": 1.3879423328964614e-05, "loss": 0.5296, "num_input_tokens_seen": 284064, "step": 1060 }, { "epoch": 0.27916120576671033, "grad_norm": 14.395566940307617, "learning_rate": 1.3944954128440368e-05, "loss": 0.5659, "num_input_tokens_seen": 285200, "step": 1065 }, { "epoch": 0.2804718217562254, "grad_norm": 4.832371711730957, "learning_rate": 1.401048492791612e-05, "loss": 0.5686, "num_input_tokens_seen": 286432, "step": 1070 }, { "epoch": 0.2817824377457405, "grad_norm": 3.439277410507202, "learning_rate": 1.4076015727391876e-05, "loss": 0.5534, "num_input_tokens_seen": 287888, "step": 1075 }, { "epoch": 0.28309305373525556, "grad_norm": 3.0723893642425537, "learning_rate": 1.4141546526867626e-05, "loss": 0.606, "num_input_tokens_seen": 289120, "step": 1080 }, { "epoch": 0.28440366972477066, "grad_norm": 5.0017313957214355, "learning_rate": 1.4207077326343382e-05, "loss": 0.7526, "num_input_tokens_seen": 290032, "step": 1085 }, { "epoch": 0.2857142857142857, "grad_norm": 2.289461612701416, "learning_rate": 1.4272608125819138e-05, "loss": 0.49, "num_input_tokens_seen": 291328, "step": 1090 }, { "epoch": 0.2870249017038008, "grad_norm": 2.8156521320343018, "learning_rate": 1.4338138925294888e-05, "loss": 0.4005, "num_input_tokens_seen": 293632, "step": 1095 }, { "epoch": 0.28833551769331583, "grad_norm": 3.548586368560791, "learning_rate": 1.4403669724770644e-05, "loss": 0.7492, "num_input_tokens_seen": 294880, "step": 1100 }, { "epoch": 0.28964613368283093, "grad_norm": 3.3822519779205322, "learning_rate": 1.4469200524246396e-05, "loss": 0.523, "num_input_tokens_seen": 296208, "step": 1105 }, { "epoch": 0.290956749672346, "grad_norm": 1.7299913167953491, "learning_rate": 1.453473132372215e-05, "loss": 0.5571, "num_input_tokens_seen": 297760, "step": 1110 }, { "epoch": 0.29226736566186107, "grad_norm": 4.685419082641602, "learning_rate": 1.4600262123197902e-05, "loss": 0.5401, "num_input_tokens_seen": 298832, "step": 1115 }, { "epoch": 0.29357798165137616, "grad_norm": 2.366166353225708, "learning_rate": 1.4665792922673658e-05, "loss": 0.5632, "num_input_tokens_seen": 300112, "step": 1120 }, { "epoch": 0.2948885976408912, "grad_norm": 3.1106951236724854, "learning_rate": 1.4731323722149412e-05, "loss": 0.6186, "num_input_tokens_seen": 301248, "step": 1125 }, { "epoch": 0.2961992136304063, "grad_norm": 6.185274124145508, "learning_rate": 1.4796854521625164e-05, "loss": 0.5048, "num_input_tokens_seen": 302656, "step": 1130 }, { "epoch": 0.29750982961992134, "grad_norm": 2.4686310291290283, "learning_rate": 1.486238532110092e-05, "loss": 0.5719, "num_input_tokens_seen": 304000, "step": 1135 }, { "epoch": 0.29882044560943644, "grad_norm": 4.812318325042725, "learning_rate": 1.4927916120576672e-05, "loss": 0.6592, "num_input_tokens_seen": 305472, "step": 1140 }, { "epoch": 0.30013106159895153, "grad_norm": 5.288152694702148, "learning_rate": 1.4993446920052426e-05, "loss": 0.3469, "num_input_tokens_seen": 306448, "step": 1145 }, { "epoch": 0.30144167758846657, "grad_norm": 25.66158103942871, "learning_rate": 1.5058977719528178e-05, "loss": 0.8703, "num_input_tokens_seen": 307632, "step": 1150 }, { "epoch": 0.30275229357798167, "grad_norm": 2.2138991355895996, "learning_rate": 1.5124508519003932e-05, "loss": 0.3953, "num_input_tokens_seen": 308784, "step": 1155 }, { "epoch": 0.3040629095674967, "grad_norm": 6.181709289550781, "learning_rate": 1.5190039318479687e-05, "loss": 0.511, "num_input_tokens_seen": 310096, "step": 1160 }, { "epoch": 0.3053735255570118, "grad_norm": 3.6387598514556885, "learning_rate": 1.525557011795544e-05, "loss": 1.234, "num_input_tokens_seen": 311472, "step": 1165 }, { "epoch": 0.30668414154652685, "grad_norm": 1.4975757598876953, "learning_rate": 1.5321100917431195e-05, "loss": 0.5748, "num_input_tokens_seen": 312704, "step": 1170 }, { "epoch": 0.30799475753604194, "grad_norm": 4.869255065917969, "learning_rate": 1.5386631716906946e-05, "loss": 0.4746, "num_input_tokens_seen": 313888, "step": 1175 }, { "epoch": 0.30930537352555704, "grad_norm": 2.780200481414795, "learning_rate": 1.54521625163827e-05, "loss": 0.5761, "num_input_tokens_seen": 315040, "step": 1180 }, { "epoch": 0.3106159895150721, "grad_norm": 3.6357228755950928, "learning_rate": 1.5517693315858454e-05, "loss": 0.9553, "num_input_tokens_seen": 316352, "step": 1185 }, { "epoch": 0.3119266055045872, "grad_norm": 4.174716949462891, "learning_rate": 1.5583224115334208e-05, "loss": 0.6184, "num_input_tokens_seen": 317616, "step": 1190 }, { "epoch": 0.3132372214941022, "grad_norm": 3.681637763977051, "learning_rate": 1.564875491480996e-05, "loss": 0.6654, "num_input_tokens_seen": 319184, "step": 1195 }, { "epoch": 0.3145478374836173, "grad_norm": 4.555323600769043, "learning_rate": 1.5714285714285715e-05, "loss": 0.5972, "num_input_tokens_seen": 320752, "step": 1200 }, { "epoch": 0.31585845347313235, "grad_norm": 5.134693622589111, "learning_rate": 1.577981651376147e-05, "loss": 0.4852, "num_input_tokens_seen": 322032, "step": 1205 }, { "epoch": 0.31716906946264745, "grad_norm": 4.212268352508545, "learning_rate": 1.5845347313237223e-05, "loss": 0.507, "num_input_tokens_seen": 323360, "step": 1210 }, { "epoch": 0.31847968545216254, "grad_norm": 8.982877731323242, "learning_rate": 1.5910878112712977e-05, "loss": 0.4859, "num_input_tokens_seen": 324528, "step": 1215 }, { "epoch": 0.3197903014416776, "grad_norm": 7.022196292877197, "learning_rate": 1.5976408912188728e-05, "loss": 0.5087, "num_input_tokens_seen": 326928, "step": 1220 }, { "epoch": 0.3211009174311927, "grad_norm": 3.4443554878234863, "learning_rate": 1.604193971166448e-05, "loss": 0.432, "num_input_tokens_seen": 328432, "step": 1225 }, { "epoch": 0.3224115334207077, "grad_norm": 12.509011268615723, "learning_rate": 1.610747051114024e-05, "loss": 0.6027, "num_input_tokens_seen": 329440, "step": 1230 }, { "epoch": 0.3237221494102228, "grad_norm": 2.5812745094299316, "learning_rate": 1.617300131061599e-05, "loss": 0.4072, "num_input_tokens_seen": 330912, "step": 1235 }, { "epoch": 0.32503276539973786, "grad_norm": 2.3451120853424072, "learning_rate": 1.6238532110091743e-05, "loss": 0.5806, "num_input_tokens_seen": 332288, "step": 1240 }, { "epoch": 0.32634338138925295, "grad_norm": 1.840508222579956, "learning_rate": 1.6304062909567497e-05, "loss": 0.6111, "num_input_tokens_seen": 333584, "step": 1245 }, { "epoch": 0.32765399737876805, "grad_norm": 2.1190602779388428, "learning_rate": 1.636959370904325e-05, "loss": 0.5152, "num_input_tokens_seen": 335392, "step": 1250 }, { "epoch": 0.3289646133682831, "grad_norm": 3.62638258934021, "learning_rate": 1.6435124508519005e-05, "loss": 0.6256, "num_input_tokens_seen": 336560, "step": 1255 }, { "epoch": 0.3302752293577982, "grad_norm": 3.258556365966797, "learning_rate": 1.650065530799476e-05, "loss": 0.6303, "num_input_tokens_seen": 338192, "step": 1260 }, { "epoch": 0.3315858453473132, "grad_norm": 3.5412490367889404, "learning_rate": 1.6566186107470513e-05, "loss": 0.6393, "num_input_tokens_seen": 339568, "step": 1265 }, { "epoch": 0.3328964613368283, "grad_norm": 7.9454193115234375, "learning_rate": 1.6631716906946267e-05, "loss": 0.4624, "num_input_tokens_seen": 340736, "step": 1270 }, { "epoch": 0.33420707732634336, "grad_norm": 6.095376968383789, "learning_rate": 1.669724770642202e-05, "loss": 1.2262, "num_input_tokens_seen": 341680, "step": 1275 }, { "epoch": 0.33551769331585846, "grad_norm": 2.614762306213379, "learning_rate": 1.676277850589777e-05, "loss": 0.6406, "num_input_tokens_seen": 342800, "step": 1280 }, { "epoch": 0.33682830930537355, "grad_norm": 3.5753233432769775, "learning_rate": 1.682830930537353e-05, "loss": 0.64, "num_input_tokens_seen": 344000, "step": 1285 }, { "epoch": 0.3381389252948886, "grad_norm": 4.9823899269104, "learning_rate": 1.689384010484928e-05, "loss": 0.5243, "num_input_tokens_seen": 345968, "step": 1290 }, { "epoch": 0.3394495412844037, "grad_norm": 5.558934688568115, "learning_rate": 1.6959370904325033e-05, "loss": 0.4954, "num_input_tokens_seen": 347264, "step": 1295 }, { "epoch": 0.34076015727391873, "grad_norm": 27.250900268554688, "learning_rate": 1.702490170380079e-05, "loss": 0.5076, "num_input_tokens_seen": 348720, "step": 1300 }, { "epoch": 0.3420707732634338, "grad_norm": 3.5045039653778076, "learning_rate": 1.709043250327654e-05, "loss": 0.4386, "num_input_tokens_seen": 350032, "step": 1305 }, { "epoch": 0.34338138925294887, "grad_norm": 3.0216636657714844, "learning_rate": 1.7155963302752295e-05, "loss": 0.5246, "num_input_tokens_seen": 351408, "step": 1310 }, { "epoch": 0.34469200524246396, "grad_norm": 1.7246674299240112, "learning_rate": 1.722149410222805e-05, "loss": 0.4417, "num_input_tokens_seen": 352560, "step": 1315 }, { "epoch": 0.34600262123197906, "grad_norm": 4.161496162414551, "learning_rate": 1.7287024901703802e-05, "loss": 0.6305, "num_input_tokens_seen": 353824, "step": 1320 }, { "epoch": 0.3473132372214941, "grad_norm": 3.190852165222168, "learning_rate": 1.7352555701179553e-05, "loss": 0.4237, "num_input_tokens_seen": 355456, "step": 1325 }, { "epoch": 0.3486238532110092, "grad_norm": 2.0025367736816406, "learning_rate": 1.741808650065531e-05, "loss": 0.4906, "num_input_tokens_seen": 356528, "step": 1330 }, { "epoch": 0.34993446920052423, "grad_norm": 2.531231164932251, "learning_rate": 1.7483617300131064e-05, "loss": 0.3692, "num_input_tokens_seen": 357664, "step": 1335 }, { "epoch": 0.35124508519003933, "grad_norm": 1.7449383735656738, "learning_rate": 1.7549148099606815e-05, "loss": 0.6657, "num_input_tokens_seen": 359136, "step": 1340 }, { "epoch": 0.35255570117955437, "grad_norm": 2.5152766704559326, "learning_rate": 1.7614678899082572e-05, "loss": 0.3857, "num_input_tokens_seen": 360560, "step": 1345 }, { "epoch": 0.35386631716906947, "grad_norm": 2.7248895168304443, "learning_rate": 1.7680209698558323e-05, "loss": 0.6477, "num_input_tokens_seen": 361952, "step": 1350 }, { "epoch": 0.35517693315858456, "grad_norm": 4.4747796058654785, "learning_rate": 1.7745740498034076e-05, "loss": 0.5307, "num_input_tokens_seen": 363360, "step": 1355 }, { "epoch": 0.3564875491480996, "grad_norm": 3.6998848915100098, "learning_rate": 1.781127129750983e-05, "loss": 0.5986, "num_input_tokens_seen": 364480, "step": 1360 }, { "epoch": 0.3577981651376147, "grad_norm": 16.500486373901367, "learning_rate": 1.7876802096985584e-05, "loss": 0.7447, "num_input_tokens_seen": 365584, "step": 1365 }, { "epoch": 0.35910878112712974, "grad_norm": 3.0840508937835693, "learning_rate": 1.7942332896461335e-05, "loss": 0.7333, "num_input_tokens_seen": 366704, "step": 1370 }, { "epoch": 0.36041939711664484, "grad_norm": 4.8327836990356445, "learning_rate": 1.8007863695937092e-05, "loss": 0.8173, "num_input_tokens_seen": 368000, "step": 1375 }, { "epoch": 0.3617300131061599, "grad_norm": 4.091284275054932, "learning_rate": 1.8073394495412846e-05, "loss": 0.4185, "num_input_tokens_seen": 369360, "step": 1380 }, { "epoch": 0.36304062909567497, "grad_norm": 1.5369758605957031, "learning_rate": 1.8138925294888597e-05, "loss": 0.5342, "num_input_tokens_seen": 370864, "step": 1385 }, { "epoch": 0.36435124508519, "grad_norm": 3.236823558807373, "learning_rate": 1.8204456094364354e-05, "loss": 0.365, "num_input_tokens_seen": 372000, "step": 1390 }, { "epoch": 0.3656618610747051, "grad_norm": 55.99557876586914, "learning_rate": 1.8269986893840104e-05, "loss": 0.9095, "num_input_tokens_seen": 373168, "step": 1395 }, { "epoch": 0.3669724770642202, "grad_norm": 5.798702239990234, "learning_rate": 1.833551769331586e-05, "loss": 0.651, "num_input_tokens_seen": 374304, "step": 1400 }, { "epoch": 0.36828309305373524, "grad_norm": 3.69032621383667, "learning_rate": 1.8401048492791612e-05, "loss": 0.3197, "num_input_tokens_seen": 375888, "step": 1405 }, { "epoch": 0.36959370904325034, "grad_norm": 5.794120788574219, "learning_rate": 1.8466579292267366e-05, "loss": 0.5491, "num_input_tokens_seen": 377296, "step": 1410 }, { "epoch": 0.3709043250327654, "grad_norm": 5.924728870391846, "learning_rate": 1.853211009174312e-05, "loss": 0.4044, "num_input_tokens_seen": 378608, "step": 1415 }, { "epoch": 0.3722149410222805, "grad_norm": 1.8516135215759277, "learning_rate": 1.8597640891218874e-05, "loss": 0.5017, "num_input_tokens_seen": 380192, "step": 1420 }, { "epoch": 0.3735255570117955, "grad_norm": 3.4478466510772705, "learning_rate": 1.8663171690694628e-05, "loss": 0.583, "num_input_tokens_seen": 381616, "step": 1425 }, { "epoch": 0.3748361730013106, "grad_norm": 3.99037766456604, "learning_rate": 1.872870249017038e-05, "loss": 0.5024, "num_input_tokens_seen": 382672, "step": 1430 }, { "epoch": 0.3761467889908257, "grad_norm": 4.1149396896362305, "learning_rate": 1.8794233289646136e-05, "loss": 0.5785, "num_input_tokens_seen": 384176, "step": 1435 }, { "epoch": 0.37745740498034075, "grad_norm": 2.0330164432525635, "learning_rate": 1.8859764089121886e-05, "loss": 0.7361, "num_input_tokens_seen": 385584, "step": 1440 }, { "epoch": 0.37876802096985585, "grad_norm": 2.4837565422058105, "learning_rate": 1.892529488859764e-05, "loss": 0.4607, "num_input_tokens_seen": 386832, "step": 1445 }, { "epoch": 0.3800786369593709, "grad_norm": 2.0028533935546875, "learning_rate": 1.8990825688073397e-05, "loss": 0.466, "num_input_tokens_seen": 388416, "step": 1450 }, { "epoch": 0.381389252948886, "grad_norm": 3.968302011489868, "learning_rate": 1.9056356487549148e-05, "loss": 0.4032, "num_input_tokens_seen": 389648, "step": 1455 }, { "epoch": 0.382699868938401, "grad_norm": 2.1822702884674072, "learning_rate": 1.9121887287024902e-05, "loss": 0.3536, "num_input_tokens_seen": 390864, "step": 1460 }, { "epoch": 0.3840104849279161, "grad_norm": 2.3429641723632812, "learning_rate": 1.9187418086500656e-05, "loss": 0.4989, "num_input_tokens_seen": 392128, "step": 1465 }, { "epoch": 0.3853211009174312, "grad_norm": 2.4910130500793457, "learning_rate": 1.925294888597641e-05, "loss": 0.609, "num_input_tokens_seen": 394864, "step": 1470 }, { "epoch": 0.38663171690694625, "grad_norm": 8.46894359588623, "learning_rate": 1.9318479685452164e-05, "loss": 0.4123, "num_input_tokens_seen": 395776, "step": 1475 }, { "epoch": 0.38794233289646135, "grad_norm": 2.6442813873291016, "learning_rate": 1.9384010484927918e-05, "loss": 0.6988, "num_input_tokens_seen": 396880, "step": 1480 }, { "epoch": 0.3892529488859764, "grad_norm": 3.278961181640625, "learning_rate": 1.944954128440367e-05, "loss": 0.4707, "num_input_tokens_seen": 398512, "step": 1485 }, { "epoch": 0.3905635648754915, "grad_norm": 4.856677532196045, "learning_rate": 1.9515072083879425e-05, "loss": 0.756, "num_input_tokens_seen": 399664, "step": 1490 }, { "epoch": 0.3918741808650065, "grad_norm": 5.141520977020264, "learning_rate": 1.958060288335518e-05, "loss": 0.3666, "num_input_tokens_seen": 400864, "step": 1495 }, { "epoch": 0.3931847968545216, "grad_norm": 1.8487346172332764, "learning_rate": 1.964613368283093e-05, "loss": 0.5834, "num_input_tokens_seen": 402352, "step": 1500 }, { "epoch": 0.3944954128440367, "grad_norm": 1.9902998208999634, "learning_rate": 1.9711664482306684e-05, "loss": 0.549, "num_input_tokens_seen": 403744, "step": 1505 }, { "epoch": 0.39580602883355176, "grad_norm": 4.140905857086182, "learning_rate": 1.9777195281782438e-05, "loss": 0.5027, "num_input_tokens_seen": 405104, "step": 1510 }, { "epoch": 0.39711664482306686, "grad_norm": 6.926522254943848, "learning_rate": 1.984272608125819e-05, "loss": 0.7384, "num_input_tokens_seen": 406208, "step": 1515 }, { "epoch": 0.3984272608125819, "grad_norm": 4.066760063171387, "learning_rate": 1.9908256880733945e-05, "loss": 0.5337, "num_input_tokens_seen": 407120, "step": 1520 }, { "epoch": 0.399737876802097, "grad_norm": 2.480729579925537, "learning_rate": 1.99737876802097e-05, "loss": 0.6012, "num_input_tokens_seen": 408080, "step": 1525 }, { "epoch": 0.40104849279161203, "grad_norm": 2.9193475246429443, "learning_rate": 2.0039318479685453e-05, "loss": 0.5894, "num_input_tokens_seen": 409360, "step": 1530 }, { "epoch": 0.40235910878112713, "grad_norm": 2.885507583618164, "learning_rate": 2.0104849279161207e-05, "loss": 0.4954, "num_input_tokens_seen": 410720, "step": 1535 }, { "epoch": 0.4036697247706422, "grad_norm": 3.043856382369995, "learning_rate": 2.017038007863696e-05, "loss": 0.5587, "num_input_tokens_seen": 411904, "step": 1540 }, { "epoch": 0.40498034076015726, "grad_norm": 2.6498143672943115, "learning_rate": 2.023591087811271e-05, "loss": 0.5367, "num_input_tokens_seen": 412960, "step": 1545 }, { "epoch": 0.40629095674967236, "grad_norm": 1.8738378286361694, "learning_rate": 2.030144167758847e-05, "loss": 0.5362, "num_input_tokens_seen": 414784, "step": 1550 }, { "epoch": 0.4076015727391874, "grad_norm": 1.6748307943344116, "learning_rate": 2.0366972477064223e-05, "loss": 0.5578, "num_input_tokens_seen": 416208, "step": 1555 }, { "epoch": 0.4089121887287025, "grad_norm": 3.6126832962036133, "learning_rate": 2.0432503276539973e-05, "loss": 0.4681, "num_input_tokens_seen": 417440, "step": 1560 }, { "epoch": 0.41022280471821754, "grad_norm": 1.3035978078842163, "learning_rate": 2.049803407601573e-05, "loss": 0.578, "num_input_tokens_seen": 419008, "step": 1565 }, { "epoch": 0.41153342070773263, "grad_norm": 3.3946073055267334, "learning_rate": 2.056356487549148e-05, "loss": 0.5003, "num_input_tokens_seen": 420384, "step": 1570 }, { "epoch": 0.41284403669724773, "grad_norm": 4.49365234375, "learning_rate": 2.0629095674967235e-05, "loss": 0.6083, "num_input_tokens_seen": 421440, "step": 1575 }, { "epoch": 0.41415465268676277, "grad_norm": 2.4396347999572754, "learning_rate": 2.069462647444299e-05, "loss": 0.528, "num_input_tokens_seen": 422608, "step": 1580 }, { "epoch": 0.41546526867627787, "grad_norm": 2.640653371810913, "learning_rate": 2.0760157273918743e-05, "loss": 0.5247, "num_input_tokens_seen": 423952, "step": 1585 }, { "epoch": 0.4167758846657929, "grad_norm": 2.1404130458831787, "learning_rate": 2.0825688073394497e-05, "loss": 0.5102, "num_input_tokens_seen": 425552, "step": 1590 }, { "epoch": 0.418086500655308, "grad_norm": 3.697516441345215, "learning_rate": 2.089121887287025e-05, "loss": 0.6127, "num_input_tokens_seen": 426656, "step": 1595 }, { "epoch": 0.41939711664482304, "grad_norm": 1.5919452905654907, "learning_rate": 2.0956749672346005e-05, "loss": 0.7057, "num_input_tokens_seen": 427696, "step": 1600 }, { "epoch": 0.42070773263433814, "grad_norm": 3.9976584911346436, "learning_rate": 2.1022280471821755e-05, "loss": 0.5847, "num_input_tokens_seen": 428656, "step": 1605 }, { "epoch": 0.42201834862385323, "grad_norm": 2.3983733654022217, "learning_rate": 2.1087811271297513e-05, "loss": 0.696, "num_input_tokens_seen": 430128, "step": 1610 }, { "epoch": 0.4233289646133683, "grad_norm": 1.463672399520874, "learning_rate": 2.1153342070773263e-05, "loss": 0.391, "num_input_tokens_seen": 431792, "step": 1615 }, { "epoch": 0.42463958060288337, "grad_norm": 2.2722384929656982, "learning_rate": 2.1218872870249017e-05, "loss": 0.4164, "num_input_tokens_seen": 433200, "step": 1620 }, { "epoch": 0.4259501965923984, "grad_norm": 1.6852360963821411, "learning_rate": 2.1284403669724774e-05, "loss": 0.4642, "num_input_tokens_seen": 434688, "step": 1625 }, { "epoch": 0.4272608125819135, "grad_norm": 2.5211703777313232, "learning_rate": 2.1349934469200525e-05, "loss": 0.7214, "num_input_tokens_seen": 435920, "step": 1630 }, { "epoch": 0.42857142857142855, "grad_norm": 2.2542707920074463, "learning_rate": 2.141546526867628e-05, "loss": 0.6131, "num_input_tokens_seen": 438000, "step": 1635 }, { "epoch": 0.42988204456094364, "grad_norm": 2.988687038421631, "learning_rate": 2.1480996068152033e-05, "loss": 0.543, "num_input_tokens_seen": 439104, "step": 1640 }, { "epoch": 0.43119266055045874, "grad_norm": 3.019631862640381, "learning_rate": 2.1546526867627787e-05, "loss": 0.5606, "num_input_tokens_seen": 440400, "step": 1645 }, { "epoch": 0.4325032765399738, "grad_norm": 4.057520866394043, "learning_rate": 2.1612057667103537e-05, "loss": 0.5824, "num_input_tokens_seen": 442592, "step": 1650 }, { "epoch": 0.4338138925294889, "grad_norm": 3.4875128269195557, "learning_rate": 2.1677588466579294e-05, "loss": 0.6625, "num_input_tokens_seen": 444016, "step": 1655 }, { "epoch": 0.4351245085190039, "grad_norm": 3.0716159343719482, "learning_rate": 2.1743119266055048e-05, "loss": 0.5156, "num_input_tokens_seen": 445424, "step": 1660 }, { "epoch": 0.436435124508519, "grad_norm": 1.4717578887939453, "learning_rate": 2.18086500655308e-05, "loss": 0.4807, "num_input_tokens_seen": 446768, "step": 1665 }, { "epoch": 0.43774574049803405, "grad_norm": 2.251211643218994, "learning_rate": 2.1874180865006556e-05, "loss": 0.7684, "num_input_tokens_seen": 448288, "step": 1670 }, { "epoch": 0.43905635648754915, "grad_norm": 2.8987057209014893, "learning_rate": 2.1939711664482307e-05, "loss": 0.7574, "num_input_tokens_seen": 449440, "step": 1675 }, { "epoch": 0.44036697247706424, "grad_norm": 4.709256172180176, "learning_rate": 2.200524246395806e-05, "loss": 0.4843, "num_input_tokens_seen": 450832, "step": 1680 }, { "epoch": 0.4416775884665793, "grad_norm": 2.2053301334381104, "learning_rate": 2.2070773263433814e-05, "loss": 0.6426, "num_input_tokens_seen": 451984, "step": 1685 }, { "epoch": 0.4429882044560944, "grad_norm": 3.258676290512085, "learning_rate": 2.213630406290957e-05, "loss": 0.5525, "num_input_tokens_seen": 453376, "step": 1690 }, { "epoch": 0.4442988204456094, "grad_norm": 2.4151039123535156, "learning_rate": 2.2201834862385322e-05, "loss": 0.4433, "num_input_tokens_seen": 454672, "step": 1695 }, { "epoch": 0.4456094364351245, "grad_norm": 1.9659720659255981, "learning_rate": 2.2267365661861076e-05, "loss": 0.4438, "num_input_tokens_seen": 456240, "step": 1700 }, { "epoch": 0.44692005242463956, "grad_norm": 10.494597434997559, "learning_rate": 2.233289646133683e-05, "loss": 0.4939, "num_input_tokens_seen": 457632, "step": 1705 }, { "epoch": 0.44823066841415465, "grad_norm": 6.122323036193848, "learning_rate": 2.239842726081258e-05, "loss": 0.384, "num_input_tokens_seen": 459200, "step": 1710 }, { "epoch": 0.44954128440366975, "grad_norm": 2.1752569675445557, "learning_rate": 2.2463958060288338e-05, "loss": 0.6029, "num_input_tokens_seen": 460384, "step": 1715 }, { "epoch": 0.4508519003931848, "grad_norm": 2.5353310108184814, "learning_rate": 2.252948885976409e-05, "loss": 0.5166, "num_input_tokens_seen": 461712, "step": 1720 }, { "epoch": 0.4521625163826999, "grad_norm": 2.3990066051483154, "learning_rate": 2.2595019659239842e-05, "loss": 0.3839, "num_input_tokens_seen": 462832, "step": 1725 }, { "epoch": 0.4534731323722149, "grad_norm": 1.6242684125900269, "learning_rate": 2.26605504587156e-05, "loss": 0.4182, "num_input_tokens_seen": 464224, "step": 1730 }, { "epoch": 0.45478374836173, "grad_norm": 1.0428502559661865, "learning_rate": 2.272608125819135e-05, "loss": 0.4975, "num_input_tokens_seen": 465696, "step": 1735 }, { "epoch": 0.45609436435124506, "grad_norm": 2.4753592014312744, "learning_rate": 2.2791612057667104e-05, "loss": 0.5524, "num_input_tokens_seen": 466944, "step": 1740 }, { "epoch": 0.45740498034076016, "grad_norm": 1.976385474205017, "learning_rate": 2.2857142857142858e-05, "loss": 0.579, "num_input_tokens_seen": 468256, "step": 1745 }, { "epoch": 0.45871559633027525, "grad_norm": 3.355137586593628, "learning_rate": 2.2922673656618612e-05, "loss": 0.5612, "num_input_tokens_seen": 469376, "step": 1750 }, { "epoch": 0.4600262123197903, "grad_norm": 2.097940683364868, "learning_rate": 2.2988204456094366e-05, "loss": 0.6303, "num_input_tokens_seen": 470336, "step": 1755 }, { "epoch": 0.4613368283093054, "grad_norm": 2.124408006668091, "learning_rate": 2.305373525557012e-05, "loss": 0.611, "num_input_tokens_seen": 471360, "step": 1760 }, { "epoch": 0.46264744429882043, "grad_norm": 3.308434247970581, "learning_rate": 2.3119266055045874e-05, "loss": 0.6561, "num_input_tokens_seen": 472576, "step": 1765 }, { "epoch": 0.4639580602883355, "grad_norm": 3.734894037246704, "learning_rate": 2.3184796854521628e-05, "loss": 0.6158, "num_input_tokens_seen": 474224, "step": 1770 }, { "epoch": 0.46526867627785057, "grad_norm": 2.685161828994751, "learning_rate": 2.325032765399738e-05, "loss": 0.5072, "num_input_tokens_seen": 475376, "step": 1775 }, { "epoch": 0.46657929226736566, "grad_norm": 2.611184597015381, "learning_rate": 2.3315858453473132e-05, "loss": 0.6766, "num_input_tokens_seen": 476592, "step": 1780 }, { "epoch": 0.46788990825688076, "grad_norm": 8.008881568908691, "learning_rate": 2.3381389252948886e-05, "loss": 0.6705, "num_input_tokens_seen": 477504, "step": 1785 }, { "epoch": 0.4692005242463958, "grad_norm": 4.252854347229004, "learning_rate": 2.344692005242464e-05, "loss": 0.4777, "num_input_tokens_seen": 478560, "step": 1790 }, { "epoch": 0.4705111402359109, "grad_norm": 1.3049945831298828, "learning_rate": 2.3512450851900394e-05, "loss": 0.4486, "num_input_tokens_seen": 479920, "step": 1795 }, { "epoch": 0.47182175622542594, "grad_norm": 2.172233819961548, "learning_rate": 2.3577981651376148e-05, "loss": 0.3998, "num_input_tokens_seen": 481264, "step": 1800 }, { "epoch": 0.47313237221494103, "grad_norm": 52.008480072021484, "learning_rate": 2.36435124508519e-05, "loss": 0.8518, "num_input_tokens_seen": 482608, "step": 1805 }, { "epoch": 0.4744429882044561, "grad_norm": 3.9156692028045654, "learning_rate": 2.3709043250327656e-05, "loss": 0.5054, "num_input_tokens_seen": 484096, "step": 1810 }, { "epoch": 0.47575360419397117, "grad_norm": 4.9985761642456055, "learning_rate": 2.377457404980341e-05, "loss": 0.5315, "num_input_tokens_seen": 485120, "step": 1815 }, { "epoch": 0.47706422018348627, "grad_norm": 6.736708164215088, "learning_rate": 2.3840104849279163e-05, "loss": 0.3992, "num_input_tokens_seen": 486080, "step": 1820 }, { "epoch": 0.4783748361730013, "grad_norm": 3.089327335357666, "learning_rate": 2.3905635648754914e-05, "loss": 0.5554, "num_input_tokens_seen": 487344, "step": 1825 }, { "epoch": 0.4796854521625164, "grad_norm": 9.291513442993164, "learning_rate": 2.397116644823067e-05, "loss": 0.4807, "num_input_tokens_seen": 488608, "step": 1830 }, { "epoch": 0.48099606815203144, "grad_norm": 2.017587900161743, "learning_rate": 2.4036697247706425e-05, "loss": 0.519, "num_input_tokens_seen": 489664, "step": 1835 }, { "epoch": 0.48230668414154654, "grad_norm": 30.55312156677246, "learning_rate": 2.4102228047182176e-05, "loss": 0.5205, "num_input_tokens_seen": 491120, "step": 1840 }, { "epoch": 0.4836173001310616, "grad_norm": 3.0239756107330322, "learning_rate": 2.4167758846657933e-05, "loss": 0.4896, "num_input_tokens_seen": 492464, "step": 1845 }, { "epoch": 0.4849279161205767, "grad_norm": 2.2593679428100586, "learning_rate": 2.4233289646133683e-05, "loss": 0.5451, "num_input_tokens_seen": 493712, "step": 1850 }, { "epoch": 0.48623853211009177, "grad_norm": 2.5638322830200195, "learning_rate": 2.4298820445609437e-05, "loss": 0.3908, "num_input_tokens_seen": 494816, "step": 1855 }, { "epoch": 0.4875491480996068, "grad_norm": 1.986190676689148, "learning_rate": 2.436435124508519e-05, "loss": 0.4548, "num_input_tokens_seen": 496256, "step": 1860 }, { "epoch": 0.4888597640891219, "grad_norm": 2.8630595207214355, "learning_rate": 2.4429882044560945e-05, "loss": 0.5502, "num_input_tokens_seen": 497280, "step": 1865 }, { "epoch": 0.49017038007863695, "grad_norm": 0.8331786394119263, "learning_rate": 2.44954128440367e-05, "loss": 0.6226, "num_input_tokens_seen": 498144, "step": 1870 }, { "epoch": 0.49148099606815204, "grad_norm": 3.678101062774658, "learning_rate": 2.4560943643512453e-05, "loss": 0.5476, "num_input_tokens_seen": 499648, "step": 1875 }, { "epoch": 0.4927916120576671, "grad_norm": 1.555010199546814, "learning_rate": 2.4626474442988207e-05, "loss": 0.4044, "num_input_tokens_seen": 501488, "step": 1880 }, { "epoch": 0.4941022280471822, "grad_norm": 1.9225693941116333, "learning_rate": 2.4692005242463957e-05, "loss": 0.7042, "num_input_tokens_seen": 503456, "step": 1885 }, { "epoch": 0.4954128440366973, "grad_norm": 1.87907874584198, "learning_rate": 2.4757536041939715e-05, "loss": 0.3805, "num_input_tokens_seen": 504544, "step": 1890 }, { "epoch": 0.4967234600262123, "grad_norm": 3.818004846572876, "learning_rate": 2.4823066841415465e-05, "loss": 0.5373, "num_input_tokens_seen": 505728, "step": 1895 }, { "epoch": 0.4980340760157274, "grad_norm": 38.935203552246094, "learning_rate": 2.488859764089122e-05, "loss": 0.6764, "num_input_tokens_seen": 506640, "step": 1900 }, { "epoch": 0.49934469200524245, "grad_norm": 3.069854259490967, "learning_rate": 2.4954128440366977e-05, "loss": 0.4264, "num_input_tokens_seen": 507904, "step": 1905 }, { "epoch": 0.5001310615989515, "eval_loss": 0.5402384996414185, "eval_runtime": 16.66, "eval_samples_per_second": 50.9, "eval_steps_per_second": 25.45, "num_input_tokens_seen": 508608, "step": 1908 }, { "epoch": 0.5006553079947575, "grad_norm": 3.434157371520996, "learning_rate": 2.501965923984273e-05, "loss": 0.556, "num_input_tokens_seen": 508992, "step": 1910 }, { "epoch": 0.5019659239842726, "grad_norm": 2.275768280029297, "learning_rate": 2.5085190039318478e-05, "loss": 0.3117, "num_input_tokens_seen": 510256, "step": 1915 }, { "epoch": 0.5032765399737876, "grad_norm": 4.502960205078125, "learning_rate": 2.5150720838794235e-05, "loss": 0.5645, "num_input_tokens_seen": 511536, "step": 1920 }, { "epoch": 0.5045871559633027, "grad_norm": 2.9038889408111572, "learning_rate": 2.521625163826999e-05, "loss": 0.6014, "num_input_tokens_seen": 512992, "step": 1925 }, { "epoch": 0.5058977719528178, "grad_norm": 2.7696759700775146, "learning_rate": 2.5281782437745743e-05, "loss": 0.4367, "num_input_tokens_seen": 514208, "step": 1930 }, { "epoch": 0.5072083879423329, "grad_norm": 4.398025035858154, "learning_rate": 2.5347313237221493e-05, "loss": 0.719, "num_input_tokens_seen": 515536, "step": 1935 }, { "epoch": 0.508519003931848, "grad_norm": 1.7317771911621094, "learning_rate": 2.5412844036697247e-05, "loss": 0.6151, "num_input_tokens_seen": 516720, "step": 1940 }, { "epoch": 0.509829619921363, "grad_norm": 2.8493947982788086, "learning_rate": 2.5478374836173e-05, "loss": 0.4086, "num_input_tokens_seen": 517808, "step": 1945 }, { "epoch": 0.5111402359108781, "grad_norm": 3.238162040710449, "learning_rate": 2.554390563564876e-05, "loss": 0.4957, "num_input_tokens_seen": 519312, "step": 1950 }, { "epoch": 0.5124508519003932, "grad_norm": 1.648484706878662, "learning_rate": 2.5609436435124512e-05, "loss": 0.501, "num_input_tokens_seen": 520528, "step": 1955 }, { "epoch": 0.5137614678899083, "grad_norm": 3.6903152465820312, "learning_rate": 2.5674967234600263e-05, "loss": 0.4812, "num_input_tokens_seen": 521744, "step": 1960 }, { "epoch": 0.5150720838794234, "grad_norm": 10.216867446899414, "learning_rate": 2.5740498034076017e-05, "loss": 0.4916, "num_input_tokens_seen": 522976, "step": 1965 }, { "epoch": 0.5163826998689384, "grad_norm": 7.164393901824951, "learning_rate": 2.580602883355177e-05, "loss": 0.5115, "num_input_tokens_seen": 524560, "step": 1970 }, { "epoch": 0.5176933158584535, "grad_norm": 1.1070894002914429, "learning_rate": 2.5871559633027525e-05, "loss": 0.3358, "num_input_tokens_seen": 525968, "step": 1975 }, { "epoch": 0.5190039318479686, "grad_norm": 0.9421691298484802, "learning_rate": 2.5937090432503282e-05, "loss": 0.5695, "num_input_tokens_seen": 527808, "step": 1980 }, { "epoch": 0.5203145478374837, "grad_norm": 1.8897169828414917, "learning_rate": 2.600262123197903e-05, "loss": 0.4778, "num_input_tokens_seen": 529472, "step": 1985 }, { "epoch": 0.5216251638269986, "grad_norm": 1.7084450721740723, "learning_rate": 2.6068152031454783e-05, "loss": 0.4683, "num_input_tokens_seen": 530864, "step": 1990 }, { "epoch": 0.5229357798165137, "grad_norm": 3.298898458480835, "learning_rate": 2.613368283093054e-05, "loss": 0.4999, "num_input_tokens_seen": 532448, "step": 1995 }, { "epoch": 0.5242463958060288, "grad_norm": 1.2518061399459839, "learning_rate": 2.6199213630406294e-05, "loss": 0.5012, "num_input_tokens_seen": 534064, "step": 2000 }, { "epoch": 0.5255570117955439, "grad_norm": 2.773449420928955, "learning_rate": 2.6264744429882045e-05, "loss": 0.6247, "num_input_tokens_seen": 535328, "step": 2005 }, { "epoch": 0.526867627785059, "grad_norm": 6.637743949890137, "learning_rate": 2.63302752293578e-05, "loss": 0.7652, "num_input_tokens_seen": 536576, "step": 2010 }, { "epoch": 0.528178243774574, "grad_norm": 2.03828763961792, "learning_rate": 2.6395806028833552e-05, "loss": 0.3845, "num_input_tokens_seen": 537792, "step": 2015 }, { "epoch": 0.5294888597640891, "grad_norm": 1.9719680547714233, "learning_rate": 2.6461336828309306e-05, "loss": 0.3912, "num_input_tokens_seen": 539120, "step": 2020 }, { "epoch": 0.5307994757536042, "grad_norm": 2.903881072998047, "learning_rate": 2.6526867627785064e-05, "loss": 0.5121, "num_input_tokens_seen": 540320, "step": 2025 }, { "epoch": 0.5321100917431193, "grad_norm": 2.1104884147644043, "learning_rate": 2.659239842726081e-05, "loss": 0.569, "num_input_tokens_seen": 541760, "step": 2030 }, { "epoch": 0.5334207077326344, "grad_norm": 13.863626480102539, "learning_rate": 2.6657929226736568e-05, "loss": 0.6838, "num_input_tokens_seen": 542784, "step": 2035 }, { "epoch": 0.5347313237221494, "grad_norm": 2.4781527519226074, "learning_rate": 2.6723460026212322e-05, "loss": 0.5885, "num_input_tokens_seen": 543968, "step": 2040 }, { "epoch": 0.5360419397116645, "grad_norm": 1.1079294681549072, "learning_rate": 2.6788990825688076e-05, "loss": 0.4155, "num_input_tokens_seen": 545248, "step": 2045 }, { "epoch": 0.5373525557011796, "grad_norm": 3.948120594024658, "learning_rate": 2.685452162516383e-05, "loss": 0.699, "num_input_tokens_seen": 546400, "step": 2050 }, { "epoch": 0.5386631716906947, "grad_norm": 1.774869441986084, "learning_rate": 2.692005242463958e-05, "loss": 0.4568, "num_input_tokens_seen": 547856, "step": 2055 }, { "epoch": 0.5399737876802096, "grad_norm": 1.9233063459396362, "learning_rate": 2.6985583224115334e-05, "loss": 0.4005, "num_input_tokens_seen": 549152, "step": 2060 }, { "epoch": 0.5412844036697247, "grad_norm": 8.367643356323242, "learning_rate": 2.7051114023591088e-05, "loss": 0.4726, "num_input_tokens_seen": 550656, "step": 2065 }, { "epoch": 0.5425950196592398, "grad_norm": 5.480030059814453, "learning_rate": 2.7116644823066845e-05, "loss": 0.4615, "num_input_tokens_seen": 551680, "step": 2070 }, { "epoch": 0.5439056356487549, "grad_norm": 1.500766396522522, "learning_rate": 2.7182175622542593e-05, "loss": 0.4512, "num_input_tokens_seen": 553088, "step": 2075 }, { "epoch": 0.54521625163827, "grad_norm": 4.978320598602295, "learning_rate": 2.724770642201835e-05, "loss": 0.48, "num_input_tokens_seen": 554304, "step": 2080 }, { "epoch": 0.546526867627785, "grad_norm": 2.5358550548553467, "learning_rate": 2.7313237221494104e-05, "loss": 0.6518, "num_input_tokens_seen": 555584, "step": 2085 }, { "epoch": 0.5478374836173001, "grad_norm": 2.632516860961914, "learning_rate": 2.7378768020969858e-05, "loss": 0.4942, "num_input_tokens_seen": 556768, "step": 2090 }, { "epoch": 0.5491480996068152, "grad_norm": 2.1802945137023926, "learning_rate": 2.744429882044561e-05, "loss": 0.365, "num_input_tokens_seen": 558208, "step": 2095 }, { "epoch": 0.5504587155963303, "grad_norm": 2.2905917167663574, "learning_rate": 2.7509829619921362e-05, "loss": 0.5263, "num_input_tokens_seen": 559664, "step": 2100 }, { "epoch": 0.5517693315858454, "grad_norm": 3.1913740634918213, "learning_rate": 2.7575360419397116e-05, "loss": 0.4994, "num_input_tokens_seen": 560800, "step": 2105 }, { "epoch": 0.5530799475753604, "grad_norm": 3.8853983879089355, "learning_rate": 2.7640891218872873e-05, "loss": 0.5338, "num_input_tokens_seen": 562640, "step": 2110 }, { "epoch": 0.5543905635648755, "grad_norm": 1.902052879333496, "learning_rate": 2.7706422018348627e-05, "loss": 0.5973, "num_input_tokens_seen": 563760, "step": 2115 }, { "epoch": 0.5557011795543906, "grad_norm": 4.302713394165039, "learning_rate": 2.777195281782438e-05, "loss": 0.4523, "num_input_tokens_seen": 565280, "step": 2120 }, { "epoch": 0.5570117955439057, "grad_norm": 1.8023966550827026, "learning_rate": 2.7837483617300132e-05, "loss": 0.4923, "num_input_tokens_seen": 566560, "step": 2125 }, { "epoch": 0.5583224115334207, "grad_norm": 3.718881368637085, "learning_rate": 2.7903014416775886e-05, "loss": 0.3822, "num_input_tokens_seen": 567776, "step": 2130 }, { "epoch": 0.5596330275229358, "grad_norm": 2.2356045246124268, "learning_rate": 2.796854521625164e-05, "loss": 0.587, "num_input_tokens_seen": 568960, "step": 2135 }, { "epoch": 0.5609436435124509, "grad_norm": 1.3925820589065552, "learning_rate": 2.8034076015727394e-05, "loss": 0.5336, "num_input_tokens_seen": 570432, "step": 2140 }, { "epoch": 0.562254259501966, "grad_norm": 1.8763049840927124, "learning_rate": 2.8099606815203144e-05, "loss": 0.6016, "num_input_tokens_seen": 571408, "step": 2145 }, { "epoch": 0.563564875491481, "grad_norm": 3.100895404815674, "learning_rate": 2.8165137614678898e-05, "loss": 0.5158, "num_input_tokens_seen": 572592, "step": 2150 }, { "epoch": 0.564875491480996, "grad_norm": 1.417417049407959, "learning_rate": 2.8230668414154655e-05, "loss": 0.5736, "num_input_tokens_seen": 573968, "step": 2155 }, { "epoch": 0.5661861074705111, "grad_norm": 2.9745612144470215, "learning_rate": 2.829619921363041e-05, "loss": 0.5089, "num_input_tokens_seen": 575088, "step": 2160 }, { "epoch": 0.5674967234600262, "grad_norm": 1.502740740776062, "learning_rate": 2.8361730013106163e-05, "loss": 0.5181, "num_input_tokens_seen": 576368, "step": 2165 }, { "epoch": 0.5688073394495413, "grad_norm": 3.3592724800109863, "learning_rate": 2.8427260812581914e-05, "loss": 0.4945, "num_input_tokens_seen": 577520, "step": 2170 }, { "epoch": 0.5701179554390564, "grad_norm": 2.2729785442352295, "learning_rate": 2.8492791612057668e-05, "loss": 0.7469, "num_input_tokens_seen": 579152, "step": 2175 }, { "epoch": 0.5714285714285714, "grad_norm": 2.079936981201172, "learning_rate": 2.855832241153342e-05, "loss": 0.3893, "num_input_tokens_seen": 580720, "step": 2180 }, { "epoch": 0.5727391874180865, "grad_norm": 4.161647319793701, "learning_rate": 2.862385321100918e-05, "loss": 0.4183, "num_input_tokens_seen": 582128, "step": 2185 }, { "epoch": 0.5740498034076016, "grad_norm": 2.6937177181243896, "learning_rate": 2.8689384010484926e-05, "loss": 0.4846, "num_input_tokens_seen": 584000, "step": 2190 }, { "epoch": 0.5753604193971167, "grad_norm": 6.311154365539551, "learning_rate": 2.875491480996068e-05, "loss": 0.7678, "num_input_tokens_seen": 585104, "step": 2195 }, { "epoch": 0.5766710353866317, "grad_norm": 2.0084595680236816, "learning_rate": 2.8820445609436437e-05, "loss": 0.4618, "num_input_tokens_seen": 586432, "step": 2200 }, { "epoch": 0.5779816513761468, "grad_norm": 1.4200323820114136, "learning_rate": 2.888597640891219e-05, "loss": 0.7214, "num_input_tokens_seen": 587648, "step": 2205 }, { "epoch": 0.5792922673656619, "grad_norm": 5.574283599853516, "learning_rate": 2.8951507208387945e-05, "loss": 0.5714, "num_input_tokens_seen": 588672, "step": 2210 }, { "epoch": 0.580602883355177, "grad_norm": 2.7227280139923096, "learning_rate": 2.9017038007863695e-05, "loss": 0.4278, "num_input_tokens_seen": 590192, "step": 2215 }, { "epoch": 0.581913499344692, "grad_norm": 6.194467067718506, "learning_rate": 2.908256880733945e-05, "loss": 0.5471, "num_input_tokens_seen": 591072, "step": 2220 }, { "epoch": 0.583224115334207, "grad_norm": 1.7511792182922363, "learning_rate": 2.9148099606815203e-05, "loss": 0.4078, "num_input_tokens_seen": 592192, "step": 2225 }, { "epoch": 0.5845347313237221, "grad_norm": 1.6587177515029907, "learning_rate": 2.921363040629096e-05, "loss": 0.5351, "num_input_tokens_seen": 593520, "step": 2230 }, { "epoch": 0.5858453473132372, "grad_norm": 3.308610200881958, "learning_rate": 2.9279161205766714e-05, "loss": 0.4078, "num_input_tokens_seen": 594528, "step": 2235 }, { "epoch": 0.5871559633027523, "grad_norm": 1.3610674142837524, "learning_rate": 2.9344692005242465e-05, "loss": 0.381, "num_input_tokens_seen": 595824, "step": 2240 }, { "epoch": 0.5884665792922673, "grad_norm": 2.024686098098755, "learning_rate": 2.941022280471822e-05, "loss": 0.3295, "num_input_tokens_seen": 597136, "step": 2245 }, { "epoch": 0.5897771952817824, "grad_norm": 11.501989364624023, "learning_rate": 2.9475753604193973e-05, "loss": 0.6205, "num_input_tokens_seen": 598320, "step": 2250 }, { "epoch": 0.5910878112712975, "grad_norm": 3.0324478149414062, "learning_rate": 2.9541284403669727e-05, "loss": 0.4186, "num_input_tokens_seen": 599792, "step": 2255 }, { "epoch": 0.5923984272608126, "grad_norm": 1.9314618110656738, "learning_rate": 2.9606815203145477e-05, "loss": 0.3461, "num_input_tokens_seen": 601120, "step": 2260 }, { "epoch": 0.5937090432503277, "grad_norm": 1.372018814086914, "learning_rate": 2.967234600262123e-05, "loss": 0.4597, "num_input_tokens_seen": 602368, "step": 2265 }, { "epoch": 0.5950196592398427, "grad_norm": 1.2908564805984497, "learning_rate": 2.9737876802096985e-05, "loss": 0.3602, "num_input_tokens_seen": 603824, "step": 2270 }, { "epoch": 0.5963302752293578, "grad_norm": 2.111158847808838, "learning_rate": 2.9803407601572742e-05, "loss": 0.5429, "num_input_tokens_seen": 605376, "step": 2275 }, { "epoch": 0.5976408912188729, "grad_norm": 6.602642059326172, "learning_rate": 2.9868938401048496e-05, "loss": 0.5188, "num_input_tokens_seen": 606672, "step": 2280 }, { "epoch": 0.598951507208388, "grad_norm": 1.3216384649276733, "learning_rate": 2.9934469200524247e-05, "loss": 0.3367, "num_input_tokens_seen": 608096, "step": 2285 }, { "epoch": 0.6002621231979031, "grad_norm": 1.3463813066482544, "learning_rate": 3e-05, "loss": 0.3935, "num_input_tokens_seen": 609408, "step": 2290 }, { "epoch": 0.601572739187418, "grad_norm": 1.8525469303131104, "learning_rate": 3.0065530799475755e-05, "loss": 0.4949, "num_input_tokens_seen": 610720, "step": 2295 }, { "epoch": 0.6028833551769331, "grad_norm": 2.7590696811676025, "learning_rate": 3.013106159895151e-05, "loss": 0.6532, "num_input_tokens_seen": 611888, "step": 2300 }, { "epoch": 0.6041939711664482, "grad_norm": 1.3104685544967651, "learning_rate": 3.0196592398427266e-05, "loss": 0.6514, "num_input_tokens_seen": 613008, "step": 2305 }, { "epoch": 0.6055045871559633, "grad_norm": 2.669842004776001, "learning_rate": 3.0262123197903013e-05, "loss": 0.4294, "num_input_tokens_seen": 614096, "step": 2310 }, { "epoch": 0.6068152031454783, "grad_norm": 2.141486644744873, "learning_rate": 3.032765399737877e-05, "loss": 0.5692, "num_input_tokens_seen": 615088, "step": 2315 }, { "epoch": 0.6081258191349934, "grad_norm": 4.448790550231934, "learning_rate": 3.0393184796854524e-05, "loss": 0.5173, "num_input_tokens_seen": 616400, "step": 2320 }, { "epoch": 0.6094364351245085, "grad_norm": 1.543714165687561, "learning_rate": 3.0458715596330278e-05, "loss": 0.6402, "num_input_tokens_seen": 618032, "step": 2325 }, { "epoch": 0.6107470511140236, "grad_norm": 2.097818613052368, "learning_rate": 3.052424639580603e-05, "loss": 0.6164, "num_input_tokens_seen": 619296, "step": 2330 }, { "epoch": 0.6120576671035387, "grad_norm": 3.7025816440582275, "learning_rate": 3.058977719528178e-05, "loss": 0.4386, "num_input_tokens_seen": 620736, "step": 2335 }, { "epoch": 0.6133682830930537, "grad_norm": 2.9317495822906494, "learning_rate": 3.0655307994757537e-05, "loss": 0.6197, "num_input_tokens_seen": 621888, "step": 2340 }, { "epoch": 0.6146788990825688, "grad_norm": 4.181792736053467, "learning_rate": 3.0720838794233294e-05, "loss": 0.4142, "num_input_tokens_seen": 622992, "step": 2345 }, { "epoch": 0.6159895150720839, "grad_norm": 1.0318039655685425, "learning_rate": 3.0786369593709044e-05, "loss": 0.6539, "num_input_tokens_seen": 624304, "step": 2350 }, { "epoch": 0.617300131061599, "grad_norm": 1.2228163480758667, "learning_rate": 3.0851900393184795e-05, "loss": 0.5192, "num_input_tokens_seen": 625744, "step": 2355 }, { "epoch": 0.6186107470511141, "grad_norm": 2.3236467838287354, "learning_rate": 3.091743119266055e-05, "loss": 0.55, "num_input_tokens_seen": 627120, "step": 2360 }, { "epoch": 0.6199213630406291, "grad_norm": 4.4309306144714355, "learning_rate": 3.09829619921363e-05, "loss": 0.6149, "num_input_tokens_seen": 628144, "step": 2365 }, { "epoch": 0.6212319790301442, "grad_norm": 1.3920481204986572, "learning_rate": 3.104849279161206e-05, "loss": 0.3572, "num_input_tokens_seen": 629344, "step": 2370 }, { "epoch": 0.6225425950196593, "grad_norm": 5.893730640411377, "learning_rate": 3.111402359108782e-05, "loss": 0.782, "num_input_tokens_seen": 630928, "step": 2375 }, { "epoch": 0.6238532110091743, "grad_norm": 2.0787031650543213, "learning_rate": 3.117955439056357e-05, "loss": 0.5117, "num_input_tokens_seen": 632272, "step": 2380 }, { "epoch": 0.6251638269986893, "grad_norm": 1.1112409830093384, "learning_rate": 3.124508519003932e-05, "loss": 0.5447, "num_input_tokens_seen": 633440, "step": 2385 }, { "epoch": 0.6264744429882044, "grad_norm": 2.87857723236084, "learning_rate": 3.1310615989515076e-05, "loss": 0.4829, "num_input_tokens_seen": 634624, "step": 2390 }, { "epoch": 0.6277850589777195, "grad_norm": 1.4129356145858765, "learning_rate": 3.1376146788990826e-05, "loss": 0.8482, "num_input_tokens_seen": 635664, "step": 2395 }, { "epoch": 0.6290956749672346, "grad_norm": 2.0380897521972656, "learning_rate": 3.144167758846658e-05, "loss": 0.4431, "num_input_tokens_seen": 637104, "step": 2400 }, { "epoch": 0.6304062909567497, "grad_norm": 1.0678879022598267, "learning_rate": 3.1507208387942334e-05, "loss": 0.3554, "num_input_tokens_seen": 638448, "step": 2405 }, { "epoch": 0.6317169069462647, "grad_norm": 2.1511006355285645, "learning_rate": 3.1572739187418085e-05, "loss": 0.3563, "num_input_tokens_seen": 641536, "step": 2410 }, { "epoch": 0.6330275229357798, "grad_norm": 1.1208857297897339, "learning_rate": 3.163826998689384e-05, "loss": 0.4532, "num_input_tokens_seen": 642960, "step": 2415 }, { "epoch": 0.6343381389252949, "grad_norm": 2.422799825668335, "learning_rate": 3.17038007863696e-05, "loss": 0.5056, "num_input_tokens_seen": 643952, "step": 2420 }, { "epoch": 0.63564875491481, "grad_norm": 3.3149569034576416, "learning_rate": 3.176933158584535e-05, "loss": 0.4983, "num_input_tokens_seen": 645184, "step": 2425 }, { "epoch": 0.6369593709043251, "grad_norm": 1.712993860244751, "learning_rate": 3.18348623853211e-05, "loss": 0.5997, "num_input_tokens_seen": 646880, "step": 2430 }, { "epoch": 0.6382699868938401, "grad_norm": 1.857866644859314, "learning_rate": 3.190039318479686e-05, "loss": 0.4265, "num_input_tokens_seen": 648064, "step": 2435 }, { "epoch": 0.6395806028833552, "grad_norm": 2.042656183242798, "learning_rate": 3.196592398427261e-05, "loss": 0.7177, "num_input_tokens_seen": 649552, "step": 2440 }, { "epoch": 0.6408912188728703, "grad_norm": 1.9189003705978394, "learning_rate": 3.2031454783748365e-05, "loss": 0.604, "num_input_tokens_seen": 650848, "step": 2445 }, { "epoch": 0.6422018348623854, "grad_norm": 3.6986427307128906, "learning_rate": 3.2096985583224116e-05, "loss": 0.7344, "num_input_tokens_seen": 652240, "step": 2450 }, { "epoch": 0.6435124508519003, "grad_norm": 2.424818992614746, "learning_rate": 3.216251638269987e-05, "loss": 0.4734, "num_input_tokens_seen": 653616, "step": 2455 }, { "epoch": 0.6448230668414154, "grad_norm": 4.151462554931641, "learning_rate": 3.2228047182175624e-05, "loss": 0.7304, "num_input_tokens_seen": 654864, "step": 2460 }, { "epoch": 0.6461336828309305, "grad_norm": 3.571312427520752, "learning_rate": 3.229357798165138e-05, "loss": 0.4208, "num_input_tokens_seen": 656288, "step": 2465 }, { "epoch": 0.6474442988204456, "grad_norm": 5.857045650482178, "learning_rate": 3.235910878112713e-05, "loss": 0.7095, "num_input_tokens_seen": 657360, "step": 2470 }, { "epoch": 0.6487549148099607, "grad_norm": 1.8801190853118896, "learning_rate": 3.242463958060288e-05, "loss": 0.63, "num_input_tokens_seen": 658704, "step": 2475 }, { "epoch": 0.6500655307994757, "grad_norm": 6.484066486358643, "learning_rate": 3.249017038007864e-05, "loss": 0.4915, "num_input_tokens_seen": 659792, "step": 2480 }, { "epoch": 0.6513761467889908, "grad_norm": 3.4781076908111572, "learning_rate": 3.255570117955439e-05, "loss": 0.4013, "num_input_tokens_seen": 661152, "step": 2485 }, { "epoch": 0.6526867627785059, "grad_norm": 1.7792664766311646, "learning_rate": 3.262123197903015e-05, "loss": 0.3262, "num_input_tokens_seen": 662288, "step": 2490 }, { "epoch": 0.653997378768021, "grad_norm": 1.9830656051635742, "learning_rate": 3.26867627785059e-05, "loss": 0.5866, "num_input_tokens_seen": 663568, "step": 2495 }, { "epoch": 0.6553079947575361, "grad_norm": 4.25563907623291, "learning_rate": 3.2752293577981655e-05, "loss": 0.5814, "num_input_tokens_seen": 664960, "step": 2500 }, { "epoch": 0.6566186107470511, "grad_norm": 1.682287335395813, "learning_rate": 3.2817824377457405e-05, "loss": 0.4626, "num_input_tokens_seen": 666416, "step": 2505 }, { "epoch": 0.6579292267365662, "grad_norm": 6.45681095123291, "learning_rate": 3.288335517693316e-05, "loss": 0.6387, "num_input_tokens_seen": 667488, "step": 2510 }, { "epoch": 0.6592398427260813, "grad_norm": 1.3837732076644897, "learning_rate": 3.294888597640891e-05, "loss": 0.2612, "num_input_tokens_seen": 669072, "step": 2515 }, { "epoch": 0.6605504587155964, "grad_norm": 1.876477599143982, "learning_rate": 3.3014416775884664e-05, "loss": 0.6158, "num_input_tokens_seen": 670256, "step": 2520 }, { "epoch": 0.6618610747051114, "grad_norm": 3.196418046951294, "learning_rate": 3.307994757536042e-05, "loss": 0.3864, "num_input_tokens_seen": 671520, "step": 2525 }, { "epoch": 0.6631716906946264, "grad_norm": 4.16778039932251, "learning_rate": 3.314547837483618e-05, "loss": 0.6017, "num_input_tokens_seen": 673056, "step": 2530 }, { "epoch": 0.6644823066841415, "grad_norm": 1.5347477197647095, "learning_rate": 3.321100917431193e-05, "loss": 0.4167, "num_input_tokens_seen": 674096, "step": 2535 }, { "epoch": 0.6657929226736566, "grad_norm": 1.4946584701538086, "learning_rate": 3.327653997378768e-05, "loss": 0.3559, "num_input_tokens_seen": 675264, "step": 2540 }, { "epoch": 0.6671035386631717, "grad_norm": 2.2691457271575928, "learning_rate": 3.334207077326344e-05, "loss": 0.4499, "num_input_tokens_seen": 676784, "step": 2545 }, { "epoch": 0.6684141546526867, "grad_norm": 2.9327380657196045, "learning_rate": 3.340760157273919e-05, "loss": 0.5844, "num_input_tokens_seen": 678064, "step": 2550 }, { "epoch": 0.6697247706422018, "grad_norm": 1.7220855951309204, "learning_rate": 3.3473132372214945e-05, "loss": 0.4998, "num_input_tokens_seen": 679728, "step": 2555 }, { "epoch": 0.6710353866317169, "grad_norm": 0.8584604263305664, "learning_rate": 3.3538663171690695e-05, "loss": 0.4667, "num_input_tokens_seen": 680976, "step": 2560 }, { "epoch": 0.672346002621232, "grad_norm": 1.0375261306762695, "learning_rate": 3.3604193971166446e-05, "loss": 0.4154, "num_input_tokens_seen": 682512, "step": 2565 }, { "epoch": 0.6736566186107471, "grad_norm": 1.6503781080245972, "learning_rate": 3.36697247706422e-05, "loss": 0.5485, "num_input_tokens_seen": 684080, "step": 2570 }, { "epoch": 0.6749672346002621, "grad_norm": 2.068755865097046, "learning_rate": 3.373525557011796e-05, "loss": 0.3939, "num_input_tokens_seen": 685248, "step": 2575 }, { "epoch": 0.6762778505897772, "grad_norm": 1.2142843008041382, "learning_rate": 3.380078636959371e-05, "loss": 0.5037, "num_input_tokens_seen": 686480, "step": 2580 }, { "epoch": 0.6775884665792923, "grad_norm": 4.525426864624023, "learning_rate": 3.386631716906947e-05, "loss": 0.5082, "num_input_tokens_seen": 687760, "step": 2585 }, { "epoch": 0.6788990825688074, "grad_norm": 2.4733333587646484, "learning_rate": 3.393184796854522e-05, "loss": 0.5672, "num_input_tokens_seen": 688896, "step": 2590 }, { "epoch": 0.6802096985583224, "grad_norm": 1.8542877435684204, "learning_rate": 3.399737876802097e-05, "loss": 0.5368, "num_input_tokens_seen": 689968, "step": 2595 }, { "epoch": 0.6815203145478375, "grad_norm": 1.457539439201355, "learning_rate": 3.4062909567496726e-05, "loss": 0.5633, "num_input_tokens_seen": 691232, "step": 2600 }, { "epoch": 0.6828309305373526, "grad_norm": 1.7383458614349365, "learning_rate": 3.4128440366972484e-05, "loss": 0.83, "num_input_tokens_seen": 692304, "step": 2605 }, { "epoch": 0.6841415465268676, "grad_norm": 0.8868491649627686, "learning_rate": 3.419397116644823e-05, "loss": 0.4118, "num_input_tokens_seen": 693744, "step": 2610 }, { "epoch": 0.6854521625163827, "grad_norm": 2.504481315612793, "learning_rate": 3.4259501965923985e-05, "loss": 0.6219, "num_input_tokens_seen": 695184, "step": 2615 }, { "epoch": 0.6867627785058977, "grad_norm": 1.7288988828659058, "learning_rate": 3.432503276539974e-05, "loss": 0.6611, "num_input_tokens_seen": 696416, "step": 2620 }, { "epoch": 0.6880733944954128, "grad_norm": 1.7229682207107544, "learning_rate": 3.439056356487549e-05, "loss": 0.4841, "num_input_tokens_seen": 697824, "step": 2625 }, { "epoch": 0.6893840104849279, "grad_norm": 0.9151841402053833, "learning_rate": 3.445609436435125e-05, "loss": 0.574, "num_input_tokens_seen": 699328, "step": 2630 }, { "epoch": 0.690694626474443, "grad_norm": 1.8206734657287598, "learning_rate": 3.4521625163827e-05, "loss": 0.5874, "num_input_tokens_seen": 700320, "step": 2635 }, { "epoch": 0.6920052424639581, "grad_norm": 1.3129353523254395, "learning_rate": 3.458715596330275e-05, "loss": 0.5497, "num_input_tokens_seen": 701712, "step": 2640 }, { "epoch": 0.6933158584534731, "grad_norm": 2.0036935806274414, "learning_rate": 3.465268676277851e-05, "loss": 0.6065, "num_input_tokens_seen": 703952, "step": 2645 }, { "epoch": 0.6946264744429882, "grad_norm": 0.8819103240966797, "learning_rate": 3.4718217562254266e-05, "loss": 0.4823, "num_input_tokens_seen": 705168, "step": 2650 }, { "epoch": 0.6959370904325033, "grad_norm": 0.980019748210907, "learning_rate": 3.4783748361730016e-05, "loss": 0.2963, "num_input_tokens_seen": 706272, "step": 2655 }, { "epoch": 0.6972477064220184, "grad_norm": 1.0561647415161133, "learning_rate": 3.484927916120577e-05, "loss": 0.357, "num_input_tokens_seen": 707680, "step": 2660 }, { "epoch": 0.6985583224115334, "grad_norm": 2.012563705444336, "learning_rate": 3.4914809960681524e-05, "loss": 0.5062, "num_input_tokens_seen": 709024, "step": 2665 }, { "epoch": 0.6998689384010485, "grad_norm": 3.8637020587921143, "learning_rate": 3.4980340760157274e-05, "loss": 0.5083, "num_input_tokens_seen": 710576, "step": 2670 }, { "epoch": 0.7011795543905636, "grad_norm": 1.6531529426574707, "learning_rate": 3.504587155963303e-05, "loss": 0.7233, "num_input_tokens_seen": 711984, "step": 2675 }, { "epoch": 0.7024901703800787, "grad_norm": 1.2874553203582764, "learning_rate": 3.511140235910878e-05, "loss": 0.4035, "num_input_tokens_seen": 713344, "step": 2680 }, { "epoch": 0.7038007863695938, "grad_norm": 1.120280385017395, "learning_rate": 3.517693315858453e-05, "loss": 0.4057, "num_input_tokens_seen": 714704, "step": 2685 }, { "epoch": 0.7051114023591087, "grad_norm": 2.632725238800049, "learning_rate": 3.524246395806029e-05, "loss": 0.6582, "num_input_tokens_seen": 716288, "step": 2690 }, { "epoch": 0.7064220183486238, "grad_norm": 3.4262731075286865, "learning_rate": 3.530799475753605e-05, "loss": 0.4644, "num_input_tokens_seen": 717392, "step": 2695 }, { "epoch": 0.7077326343381389, "grad_norm": 7.881598472595215, "learning_rate": 3.53735255570118e-05, "loss": 0.6434, "num_input_tokens_seen": 718784, "step": 2700 }, { "epoch": 0.709043250327654, "grad_norm": 2.3157427310943604, "learning_rate": 3.543905635648755e-05, "loss": 0.5766, "num_input_tokens_seen": 719824, "step": 2705 }, { "epoch": 0.7103538663171691, "grad_norm": 1.1855077743530273, "learning_rate": 3.5504587155963306e-05, "loss": 0.4848, "num_input_tokens_seen": 721424, "step": 2710 }, { "epoch": 0.7116644823066841, "grad_norm": 1.510707974433899, "learning_rate": 3.5570117955439056e-05, "loss": 0.5339, "num_input_tokens_seen": 722784, "step": 2715 }, { "epoch": 0.7129750982961992, "grad_norm": 1.3519115447998047, "learning_rate": 3.5635648754914814e-05, "loss": 0.569, "num_input_tokens_seen": 724160, "step": 2720 }, { "epoch": 0.7142857142857143, "grad_norm": 2.6192209720611572, "learning_rate": 3.570117955439057e-05, "loss": 0.4979, "num_input_tokens_seen": 725360, "step": 2725 }, { "epoch": 0.7155963302752294, "grad_norm": 2.862760305404663, "learning_rate": 3.5766710353866315e-05, "loss": 0.509, "num_input_tokens_seen": 726704, "step": 2730 }, { "epoch": 0.7169069462647444, "grad_norm": 1.060274362564087, "learning_rate": 3.583224115334207e-05, "loss": 0.5332, "num_input_tokens_seen": 728528, "step": 2735 }, { "epoch": 0.7182175622542595, "grad_norm": 1.7978585958480835, "learning_rate": 3.589777195281783e-05, "loss": 0.4857, "num_input_tokens_seen": 730096, "step": 2740 }, { "epoch": 0.7195281782437746, "grad_norm": 1.7255842685699463, "learning_rate": 3.596330275229358e-05, "loss": 0.5598, "num_input_tokens_seen": 731168, "step": 2745 }, { "epoch": 0.7208387942332897, "grad_norm": 1.333540439605713, "learning_rate": 3.602883355176933e-05, "loss": 0.4465, "num_input_tokens_seen": 733136, "step": 2750 }, { "epoch": 0.7221494102228048, "grad_norm": 1.5121113061904907, "learning_rate": 3.609436435124509e-05, "loss": 0.7126, "num_input_tokens_seen": 734448, "step": 2755 }, { "epoch": 0.7234600262123198, "grad_norm": 2.147348403930664, "learning_rate": 3.615989515072084e-05, "loss": 0.3628, "num_input_tokens_seen": 735952, "step": 2760 }, { "epoch": 0.7247706422018348, "grad_norm": 4.228831768035889, "learning_rate": 3.6225425950196595e-05, "loss": 0.7112, "num_input_tokens_seen": 737392, "step": 2765 }, { "epoch": 0.7260812581913499, "grad_norm": 2.906744956970215, "learning_rate": 3.629095674967235e-05, "loss": 0.4919, "num_input_tokens_seen": 738816, "step": 2770 }, { "epoch": 0.727391874180865, "grad_norm": 1.3440158367156982, "learning_rate": 3.6356487549148097e-05, "loss": 0.5913, "num_input_tokens_seen": 740336, "step": 2775 }, { "epoch": 0.72870249017038, "grad_norm": 3.6847968101501465, "learning_rate": 3.6422018348623854e-05, "loss": 0.4144, "num_input_tokens_seen": 741536, "step": 2780 }, { "epoch": 0.7300131061598951, "grad_norm": 3.969026565551758, "learning_rate": 3.648754914809961e-05, "loss": 0.4087, "num_input_tokens_seen": 743392, "step": 2785 }, { "epoch": 0.7313237221494102, "grad_norm": 4.676729202270508, "learning_rate": 3.655307994757536e-05, "loss": 0.4788, "num_input_tokens_seen": 744608, "step": 2790 }, { "epoch": 0.7326343381389253, "grad_norm": 8.177637100219727, "learning_rate": 3.661861074705111e-05, "loss": 0.4582, "num_input_tokens_seen": 745840, "step": 2795 }, { "epoch": 0.7339449541284404, "grad_norm": 2.78497576713562, "learning_rate": 3.668414154652687e-05, "loss": 0.6163, "num_input_tokens_seen": 746896, "step": 2800 }, { "epoch": 0.7352555701179554, "grad_norm": 2.0066730976104736, "learning_rate": 3.674967234600262e-05, "loss": 0.6236, "num_input_tokens_seen": 748176, "step": 2805 }, { "epoch": 0.7365661861074705, "grad_norm": 2.087784767150879, "learning_rate": 3.681520314547838e-05, "loss": 0.6458, "num_input_tokens_seen": 749312, "step": 2810 }, { "epoch": 0.7378768020969856, "grad_norm": 3.1604738235473633, "learning_rate": 3.6880733944954135e-05, "loss": 0.634, "num_input_tokens_seen": 750864, "step": 2815 }, { "epoch": 0.7391874180865007, "grad_norm": 3.8251841068267822, "learning_rate": 3.694626474442988e-05, "loss": 0.6352, "num_input_tokens_seen": 751952, "step": 2820 }, { "epoch": 0.7404980340760158, "grad_norm": 2.117314100265503, "learning_rate": 3.7011795543905636e-05, "loss": 0.5339, "num_input_tokens_seen": 752928, "step": 2825 }, { "epoch": 0.7418086500655308, "grad_norm": 1.470259666442871, "learning_rate": 3.707732634338139e-05, "loss": 0.7916, "num_input_tokens_seen": 754400, "step": 2830 }, { "epoch": 0.7431192660550459, "grad_norm": 1.931189775466919, "learning_rate": 3.7142857142857143e-05, "loss": 0.4198, "num_input_tokens_seen": 755712, "step": 2835 }, { "epoch": 0.744429882044561, "grad_norm": 3.133629560470581, "learning_rate": 3.72083879423329e-05, "loss": 0.6553, "num_input_tokens_seen": 758464, "step": 2840 }, { "epoch": 0.745740498034076, "grad_norm": 2.0653603076934814, "learning_rate": 3.727391874180865e-05, "loss": 0.5338, "num_input_tokens_seen": 759904, "step": 2845 }, { "epoch": 0.747051114023591, "grad_norm": 1.8256139755249023, "learning_rate": 3.73394495412844e-05, "loss": 0.3967, "num_input_tokens_seen": 761328, "step": 2850 }, { "epoch": 0.7483617300131061, "grad_norm": 2.75216007232666, "learning_rate": 3.740498034076016e-05, "loss": 0.6587, "num_input_tokens_seen": 762688, "step": 2855 }, { "epoch": 0.7496723460026212, "grad_norm": 3.2348203659057617, "learning_rate": 3.7470511140235916e-05, "loss": 0.4167, "num_input_tokens_seen": 763888, "step": 2860 }, { "epoch": 0.7509829619921363, "grad_norm": 2.1470789909362793, "learning_rate": 3.753604193971167e-05, "loss": 0.5065, "num_input_tokens_seen": 765296, "step": 2865 }, { "epoch": 0.7522935779816514, "grad_norm": 1.7849810123443604, "learning_rate": 3.760157273918742e-05, "loss": 0.5666, "num_input_tokens_seen": 766928, "step": 2870 }, { "epoch": 0.7536041939711664, "grad_norm": 3.3253672122955322, "learning_rate": 3.7667103538663175e-05, "loss": 0.5495, "num_input_tokens_seen": 768480, "step": 2875 }, { "epoch": 0.7549148099606815, "grad_norm": 3.5438053607940674, "learning_rate": 3.7732634338138925e-05, "loss": 0.4594, "num_input_tokens_seen": 769584, "step": 2880 }, { "epoch": 0.7562254259501966, "grad_norm": 0.9622169733047485, "learning_rate": 3.779816513761468e-05, "loss": 0.5531, "num_input_tokens_seen": 770960, "step": 2885 }, { "epoch": 0.7575360419397117, "grad_norm": 2.3652477264404297, "learning_rate": 3.786369593709043e-05, "loss": 0.6364, "num_input_tokens_seen": 772272, "step": 2890 }, { "epoch": 0.7588466579292268, "grad_norm": 2.092395305633545, "learning_rate": 3.7929226736566184e-05, "loss": 0.4002, "num_input_tokens_seen": 775536, "step": 2895 }, { "epoch": 0.7601572739187418, "grad_norm": 4.773683071136475, "learning_rate": 3.799475753604194e-05, "loss": 0.6081, "num_input_tokens_seen": 777216, "step": 2900 }, { "epoch": 0.7614678899082569, "grad_norm": 1.8409453630447388, "learning_rate": 3.80602883355177e-05, "loss": 0.4526, "num_input_tokens_seen": 778320, "step": 2905 }, { "epoch": 0.762778505897772, "grad_norm": 1.4166276454925537, "learning_rate": 3.812581913499345e-05, "loss": 0.5602, "num_input_tokens_seen": 779712, "step": 2910 }, { "epoch": 0.7640891218872871, "grad_norm": 3.0207102298736572, "learning_rate": 3.81913499344692e-05, "loss": 0.5046, "num_input_tokens_seen": 780928, "step": 2915 }, { "epoch": 0.765399737876802, "grad_norm": 1.3461695909500122, "learning_rate": 3.825688073394496e-05, "loss": 0.3556, "num_input_tokens_seen": 782064, "step": 2920 }, { "epoch": 0.7667103538663171, "grad_norm": 1.5074098110198975, "learning_rate": 3.832241153342071e-05, "loss": 0.5295, "num_input_tokens_seen": 783680, "step": 2925 }, { "epoch": 0.7680209698558322, "grad_norm": 1.952462911605835, "learning_rate": 3.8387942332896464e-05, "loss": 0.5014, "num_input_tokens_seen": 784912, "step": 2930 }, { "epoch": 0.7693315858453473, "grad_norm": 2.6567676067352295, "learning_rate": 3.8453473132372215e-05, "loss": 0.5363, "num_input_tokens_seen": 786368, "step": 2935 }, { "epoch": 0.7706422018348624, "grad_norm": 3.920560598373413, "learning_rate": 3.851900393184797e-05, "loss": 0.6843, "num_input_tokens_seen": 787600, "step": 2940 }, { "epoch": 0.7719528178243774, "grad_norm": 7.3719892501831055, "learning_rate": 3.858453473132372e-05, "loss": 0.6794, "num_input_tokens_seen": 788800, "step": 2945 }, { "epoch": 0.7732634338138925, "grad_norm": 1.2593191862106323, "learning_rate": 3.865006553079948e-05, "loss": 0.6653, "num_input_tokens_seen": 790464, "step": 2950 }, { "epoch": 0.7745740498034076, "grad_norm": 2.8490493297576904, "learning_rate": 3.871559633027523e-05, "loss": 1.092, "num_input_tokens_seen": 791392, "step": 2955 }, { "epoch": 0.7758846657929227, "grad_norm": 3.2696659564971924, "learning_rate": 3.878112712975098e-05, "loss": 0.5082, "num_input_tokens_seen": 792432, "step": 2960 }, { "epoch": 0.7771952817824378, "grad_norm": 1.282567024230957, "learning_rate": 3.884665792922674e-05, "loss": 0.5551, "num_input_tokens_seen": 794096, "step": 2965 }, { "epoch": 0.7785058977719528, "grad_norm": 2.949777603149414, "learning_rate": 3.891218872870249e-05, "loss": 0.4289, "num_input_tokens_seen": 795664, "step": 2970 }, { "epoch": 0.7798165137614679, "grad_norm": 2.812333583831787, "learning_rate": 3.8977719528178246e-05, "loss": 0.5806, "num_input_tokens_seen": 796992, "step": 2975 }, { "epoch": 0.781127129750983, "grad_norm": 1.6913937330245972, "learning_rate": 3.9043250327654004e-05, "loss": 0.3893, "num_input_tokens_seen": 798144, "step": 2980 }, { "epoch": 0.7824377457404981, "grad_norm": 2.1659700870513916, "learning_rate": 3.9108781127129754e-05, "loss": 0.5062, "num_input_tokens_seen": 800224, "step": 2985 }, { "epoch": 0.783748361730013, "grad_norm": 1.6434013843536377, "learning_rate": 3.9174311926605505e-05, "loss": 0.3938, "num_input_tokens_seen": 801632, "step": 2990 }, { "epoch": 0.7850589777195282, "grad_norm": 19.47600555419922, "learning_rate": 3.923984272608126e-05, "loss": 0.7312, "num_input_tokens_seen": 803088, "step": 2995 }, { "epoch": 0.7863695937090432, "grad_norm": 2.395024538040161, "learning_rate": 3.930537352555701e-05, "loss": 0.6317, "num_input_tokens_seen": 804240, "step": 3000 }, { "epoch": 0.7876802096985583, "grad_norm": 8.009413719177246, "learning_rate": 3.937090432503276e-05, "loss": 0.7781, "num_input_tokens_seen": 805312, "step": 3005 }, { "epoch": 0.7889908256880734, "grad_norm": 1.95327889919281, "learning_rate": 3.943643512450852e-05, "loss": 0.2945, "num_input_tokens_seen": 807184, "step": 3010 }, { "epoch": 0.7903014416775884, "grad_norm": 7.327889919281006, "learning_rate": 3.950196592398428e-05, "loss": 0.622, "num_input_tokens_seen": 808400, "step": 3015 }, { "epoch": 0.7916120576671035, "grad_norm": 2.3555102348327637, "learning_rate": 3.956749672346003e-05, "loss": 0.5144, "num_input_tokens_seen": 809936, "step": 3020 }, { "epoch": 0.7929226736566186, "grad_norm": 1.780224084854126, "learning_rate": 3.9633027522935785e-05, "loss": 0.5322, "num_input_tokens_seen": 811216, "step": 3025 }, { "epoch": 0.7942332896461337, "grad_norm": 2.32491397857666, "learning_rate": 3.9698558322411536e-05, "loss": 0.4713, "num_input_tokens_seen": 812288, "step": 3030 }, { "epoch": 0.7955439056356488, "grad_norm": 1.7406319379806519, "learning_rate": 3.9764089121887286e-05, "loss": 0.4999, "num_input_tokens_seen": 813536, "step": 3035 }, { "epoch": 0.7968545216251638, "grad_norm": 3.3187549114227295, "learning_rate": 3.9829619921363044e-05, "loss": 0.5852, "num_input_tokens_seen": 814768, "step": 3040 }, { "epoch": 0.7981651376146789, "grad_norm": 1.3625065088272095, "learning_rate": 3.9895150720838794e-05, "loss": 0.3661, "num_input_tokens_seen": 815984, "step": 3045 }, { "epoch": 0.799475753604194, "grad_norm": 1.6920275688171387, "learning_rate": 3.996068152031455e-05, "loss": 0.4267, "num_input_tokens_seen": 817744, "step": 3050 }, { "epoch": 0.8007863695937091, "grad_norm": 2.3205044269561768, "learning_rate": 4.00262123197903e-05, "loss": 0.4964, "num_input_tokens_seen": 818944, "step": 3055 }, { "epoch": 0.8020969855832241, "grad_norm": 1.7545166015625, "learning_rate": 4.009174311926606e-05, "loss": 0.4525, "num_input_tokens_seen": 820240, "step": 3060 }, { "epoch": 0.8034076015727392, "grad_norm": 1.3729503154754639, "learning_rate": 4.015727391874181e-05, "loss": 0.4565, "num_input_tokens_seen": 822032, "step": 3065 }, { "epoch": 0.8047182175622543, "grad_norm": 2.1463189125061035, "learning_rate": 4.022280471821757e-05, "loss": 0.4064, "num_input_tokens_seen": 823536, "step": 3070 }, { "epoch": 0.8060288335517694, "grad_norm": 0.6248984336853027, "learning_rate": 4.028833551769332e-05, "loss": 0.5668, "num_input_tokens_seen": 824864, "step": 3075 }, { "epoch": 0.8073394495412844, "grad_norm": 2.2870969772338867, "learning_rate": 4.035386631716907e-05, "loss": 0.5505, "num_input_tokens_seen": 825936, "step": 3080 }, { "epoch": 0.8086500655307994, "grad_norm": 1.2926721572875977, "learning_rate": 4.0419397116644826e-05, "loss": 0.5973, "num_input_tokens_seen": 826960, "step": 3085 }, { "epoch": 0.8099606815203145, "grad_norm": 7.615579605102539, "learning_rate": 4.048492791612058e-05, "loss": 0.2533, "num_input_tokens_seen": 827840, "step": 3090 }, { "epoch": 0.8112712975098296, "grad_norm": 2.2990803718566895, "learning_rate": 4.0550458715596333e-05, "loss": 0.5718, "num_input_tokens_seen": 829056, "step": 3095 }, { "epoch": 0.8125819134993447, "grad_norm": 1.1908137798309326, "learning_rate": 4.0615989515072084e-05, "loss": 0.343, "num_input_tokens_seen": 830528, "step": 3100 }, { "epoch": 0.8138925294888598, "grad_norm": 4.403345584869385, "learning_rate": 4.068152031454784e-05, "loss": 0.508, "num_input_tokens_seen": 831920, "step": 3105 }, { "epoch": 0.8152031454783748, "grad_norm": 3.675224781036377, "learning_rate": 4.074705111402359e-05, "loss": 0.552, "num_input_tokens_seen": 835104, "step": 3110 }, { "epoch": 0.8165137614678899, "grad_norm": 1.2305803298950195, "learning_rate": 4.081258191349935e-05, "loss": 0.3931, "num_input_tokens_seen": 837472, "step": 3115 }, { "epoch": 0.817824377457405, "grad_norm": 2.3848931789398193, "learning_rate": 4.08781127129751e-05, "loss": 0.4064, "num_input_tokens_seen": 838688, "step": 3120 }, { "epoch": 0.8191349934469201, "grad_norm": 2.7347137928009033, "learning_rate": 4.094364351245085e-05, "loss": 0.6062, "num_input_tokens_seen": 840448, "step": 3125 }, { "epoch": 0.8204456094364351, "grad_norm": 1.1643943786621094, "learning_rate": 4.100917431192661e-05, "loss": 0.3752, "num_input_tokens_seen": 841712, "step": 3130 }, { "epoch": 0.8217562254259502, "grad_norm": 1.6261266469955444, "learning_rate": 4.1074705111402365e-05, "loss": 0.8746, "num_input_tokens_seen": 843024, "step": 3135 }, { "epoch": 0.8230668414154653, "grad_norm": 1.3165112733840942, "learning_rate": 4.1140235910878115e-05, "loss": 0.728, "num_input_tokens_seen": 844736, "step": 3140 }, { "epoch": 0.8243774574049804, "grad_norm": 1.1527777910232544, "learning_rate": 4.1205766710353866e-05, "loss": 0.4727, "num_input_tokens_seen": 846432, "step": 3145 }, { "epoch": 0.8256880733944955, "grad_norm": 1.2509970664978027, "learning_rate": 4.127129750982962e-05, "loss": 0.5708, "num_input_tokens_seen": 847984, "step": 3150 }, { "epoch": 0.8269986893840104, "grad_norm": 1.5048619508743286, "learning_rate": 4.1336828309305374e-05, "loss": 0.5581, "num_input_tokens_seen": 849072, "step": 3155 }, { "epoch": 0.8283093053735255, "grad_norm": 1.7178633213043213, "learning_rate": 4.140235910878113e-05, "loss": 0.5581, "num_input_tokens_seen": 850496, "step": 3160 }, { "epoch": 0.8296199213630406, "grad_norm": 2.705397605895996, "learning_rate": 4.146788990825689e-05, "loss": 0.2745, "num_input_tokens_seen": 851904, "step": 3165 }, { "epoch": 0.8309305373525557, "grad_norm": 2.6228818893432617, "learning_rate": 4.153342070773263e-05, "loss": 0.4752, "num_input_tokens_seen": 852864, "step": 3170 }, { "epoch": 0.8322411533420708, "grad_norm": 1.890653371810913, "learning_rate": 4.159895150720839e-05, "loss": 0.5925, "num_input_tokens_seen": 854032, "step": 3175 }, { "epoch": 0.8335517693315858, "grad_norm": 2.0723142623901367, "learning_rate": 4.1664482306684147e-05, "loss": 0.4814, "num_input_tokens_seen": 855120, "step": 3180 }, { "epoch": 0.8348623853211009, "grad_norm": 1.0690852403640747, "learning_rate": 4.17300131061599e-05, "loss": 0.4767, "num_input_tokens_seen": 856320, "step": 3185 }, { "epoch": 0.836173001310616, "grad_norm": 1.492992877960205, "learning_rate": 4.1795543905635654e-05, "loss": 0.3494, "num_input_tokens_seen": 857520, "step": 3190 }, { "epoch": 0.8374836173001311, "grad_norm": 0.9519442319869995, "learning_rate": 4.1861074705111405e-05, "loss": 0.3567, "num_input_tokens_seen": 858864, "step": 3195 }, { "epoch": 0.8387942332896461, "grad_norm": 1.8586342334747314, "learning_rate": 4.1926605504587155e-05, "loss": 0.5791, "num_input_tokens_seen": 861424, "step": 3200 }, { "epoch": 0.8401048492791612, "grad_norm": 3.863041639328003, "learning_rate": 4.199213630406291e-05, "loss": 0.8194, "num_input_tokens_seen": 862448, "step": 3205 }, { "epoch": 0.8414154652686763, "grad_norm": 2.906264543533325, "learning_rate": 4.205766710353867e-05, "loss": 0.4977, "num_input_tokens_seen": 863408, "step": 3210 }, { "epoch": 0.8427260812581914, "grad_norm": 3.8230233192443848, "learning_rate": 4.2123197903014414e-05, "loss": 0.5157, "num_input_tokens_seen": 864432, "step": 3215 }, { "epoch": 0.8440366972477065, "grad_norm": 2.590656280517578, "learning_rate": 4.218872870249017e-05, "loss": 0.4625, "num_input_tokens_seen": 865648, "step": 3220 }, { "epoch": 0.8453473132372215, "grad_norm": 1.7023906707763672, "learning_rate": 4.225425950196593e-05, "loss": 0.4343, "num_input_tokens_seen": 867536, "step": 3225 }, { "epoch": 0.8466579292267365, "grad_norm": 1.7146395444869995, "learning_rate": 4.231979030144168e-05, "loss": 0.6107, "num_input_tokens_seen": 868864, "step": 3230 }, { "epoch": 0.8479685452162516, "grad_norm": 1.2369645833969116, "learning_rate": 4.2385321100917436e-05, "loss": 0.6765, "num_input_tokens_seen": 870272, "step": 3235 }, { "epoch": 0.8492791612057667, "grad_norm": 4.90701961517334, "learning_rate": 4.245085190039319e-05, "loss": 0.6877, "num_input_tokens_seen": 871344, "step": 3240 }, { "epoch": 0.8505897771952818, "grad_norm": 2.687236785888672, "learning_rate": 4.251638269986894e-05, "loss": 0.5565, "num_input_tokens_seen": 872848, "step": 3245 }, { "epoch": 0.8519003931847968, "grad_norm": 2.4967379570007324, "learning_rate": 4.2581913499344695e-05, "loss": 0.6355, "num_input_tokens_seen": 874160, "step": 3250 }, { "epoch": 0.8532110091743119, "grad_norm": 1.2244840860366821, "learning_rate": 4.264744429882045e-05, "loss": 0.3577, "num_input_tokens_seen": 875424, "step": 3255 }, { "epoch": 0.854521625163827, "grad_norm": 5.210635185241699, "learning_rate": 4.27129750982962e-05, "loss": 0.6817, "num_input_tokens_seen": 876544, "step": 3260 }, { "epoch": 0.8558322411533421, "grad_norm": 1.3335932493209839, "learning_rate": 4.277850589777195e-05, "loss": 0.4583, "num_input_tokens_seen": 878032, "step": 3265 }, { "epoch": 0.8571428571428571, "grad_norm": 1.4052375555038452, "learning_rate": 4.284403669724771e-05, "loss": 0.6311, "num_input_tokens_seen": 880112, "step": 3270 }, { "epoch": 0.8584534731323722, "grad_norm": 1.1433418989181519, "learning_rate": 4.290956749672346e-05, "loss": 0.4091, "num_input_tokens_seen": 881520, "step": 3275 }, { "epoch": 0.8597640891218873, "grad_norm": 1.6060913801193237, "learning_rate": 4.297509829619922e-05, "loss": 0.4139, "num_input_tokens_seen": 882592, "step": 3280 }, { "epoch": 0.8610747051114024, "grad_norm": 1.9087657928466797, "learning_rate": 4.304062909567497e-05, "loss": 0.4337, "num_input_tokens_seen": 883792, "step": 3285 }, { "epoch": 0.8623853211009175, "grad_norm": 1.7102701663970947, "learning_rate": 4.310615989515072e-05, "loss": 0.5552, "num_input_tokens_seen": 885200, "step": 3290 }, { "epoch": 0.8636959370904325, "grad_norm": 1.2906825542449951, "learning_rate": 4.3171690694626476e-05, "loss": 0.4763, "num_input_tokens_seen": 886752, "step": 3295 }, { "epoch": 0.8650065530799476, "grad_norm": 2.1032416820526123, "learning_rate": 4.3237221494102234e-05, "loss": 0.5778, "num_input_tokens_seen": 888080, "step": 3300 }, { "epoch": 0.8663171690694627, "grad_norm": 1.5788081884384155, "learning_rate": 4.3302752293577984e-05, "loss": 0.6777, "num_input_tokens_seen": 889248, "step": 3305 }, { "epoch": 0.8676277850589778, "grad_norm": 3.457411050796509, "learning_rate": 4.3368283093053735e-05, "loss": 0.5583, "num_input_tokens_seen": 890368, "step": 3310 }, { "epoch": 0.8689384010484927, "grad_norm": 1.8271673917770386, "learning_rate": 4.343381389252949e-05, "loss": 0.6281, "num_input_tokens_seen": 891472, "step": 3315 }, { "epoch": 0.8702490170380078, "grad_norm": 2.049283504486084, "learning_rate": 4.349934469200524e-05, "loss": 0.5094, "num_input_tokens_seen": 892688, "step": 3320 }, { "epoch": 0.8715596330275229, "grad_norm": 1.2658488750457764, "learning_rate": 4.3564875491481e-05, "loss": 0.5827, "num_input_tokens_seen": 893920, "step": 3325 }, { "epoch": 0.872870249017038, "grad_norm": 8.27939510345459, "learning_rate": 4.363040629095675e-05, "loss": 0.4716, "num_input_tokens_seen": 895072, "step": 3330 }, { "epoch": 0.8741808650065531, "grad_norm": 1.1694151163101196, "learning_rate": 4.36959370904325e-05, "loss": 0.56, "num_input_tokens_seen": 896272, "step": 3335 }, { "epoch": 0.8754914809960681, "grad_norm": 0.9440947771072388, "learning_rate": 4.376146788990826e-05, "loss": 0.6519, "num_input_tokens_seen": 897648, "step": 3340 }, { "epoch": 0.8768020969855832, "grad_norm": 2.3456943035125732, "learning_rate": 4.3826998689384016e-05, "loss": 0.6922, "num_input_tokens_seen": 898768, "step": 3345 }, { "epoch": 0.8781127129750983, "grad_norm": 1.061538815498352, "learning_rate": 4.3892529488859766e-05, "loss": 0.5273, "num_input_tokens_seen": 900240, "step": 3350 }, { "epoch": 0.8794233289646134, "grad_norm": 0.8263496160507202, "learning_rate": 4.395806028833552e-05, "loss": 0.5387, "num_input_tokens_seen": 902000, "step": 3355 }, { "epoch": 0.8807339449541285, "grad_norm": 2.1540143489837646, "learning_rate": 4.4023591087811274e-05, "loss": 0.5575, "num_input_tokens_seen": 903488, "step": 3360 }, { "epoch": 0.8820445609436435, "grad_norm": 1.0837451219558716, "learning_rate": 4.4089121887287024e-05, "loss": 0.5084, "num_input_tokens_seen": 904736, "step": 3365 }, { "epoch": 0.8833551769331586, "grad_norm": 1.2159174680709839, "learning_rate": 4.415465268676278e-05, "loss": 0.4491, "num_input_tokens_seen": 907040, "step": 3370 }, { "epoch": 0.8846657929226737, "grad_norm": 3.576995611190796, "learning_rate": 4.422018348623854e-05, "loss": 0.6555, "num_input_tokens_seen": 908128, "step": 3375 }, { "epoch": 0.8859764089121888, "grad_norm": 1.5551244020462036, "learning_rate": 4.428571428571428e-05, "loss": 0.5988, "num_input_tokens_seen": 909712, "step": 3380 }, { "epoch": 0.8872870249017037, "grad_norm": 1.3243414163589478, "learning_rate": 4.435124508519004e-05, "loss": 0.6543, "num_input_tokens_seen": 910944, "step": 3385 }, { "epoch": 0.8885976408912188, "grad_norm": 1.3442959785461426, "learning_rate": 4.44167758846658e-05, "loss": 0.496, "num_input_tokens_seen": 912096, "step": 3390 }, { "epoch": 0.8899082568807339, "grad_norm": 2.0901970863342285, "learning_rate": 4.448230668414155e-05, "loss": 0.6832, "num_input_tokens_seen": 913248, "step": 3395 }, { "epoch": 0.891218872870249, "grad_norm": 3.396979570388794, "learning_rate": 4.45478374836173e-05, "loss": 0.47, "num_input_tokens_seen": 914288, "step": 3400 }, { "epoch": 0.8925294888597641, "grad_norm": 1.4748756885528564, "learning_rate": 4.4613368283093056e-05, "loss": 0.4848, "num_input_tokens_seen": 915312, "step": 3405 }, { "epoch": 0.8938401048492791, "grad_norm": 1.0362719297409058, "learning_rate": 4.4678899082568806e-05, "loss": 0.3799, "num_input_tokens_seen": 916624, "step": 3410 }, { "epoch": 0.8951507208387942, "grad_norm": 1.6466543674468994, "learning_rate": 4.4744429882044564e-05, "loss": 0.4894, "num_input_tokens_seen": 918048, "step": 3415 }, { "epoch": 0.8964613368283093, "grad_norm": 1.196942925453186, "learning_rate": 4.480996068152032e-05, "loss": 0.7305, "num_input_tokens_seen": 919472, "step": 3420 }, { "epoch": 0.8977719528178244, "grad_norm": 1.6224817037582397, "learning_rate": 4.487549148099607e-05, "loss": 0.5118, "num_input_tokens_seen": 920688, "step": 3425 }, { "epoch": 0.8990825688073395, "grad_norm": 4.809433937072754, "learning_rate": 4.494102228047182e-05, "loss": 0.6354, "num_input_tokens_seen": 921456, "step": 3430 }, { "epoch": 0.9003931847968545, "grad_norm": 2.6695423126220703, "learning_rate": 4.500655307994758e-05, "loss": 0.4735, "num_input_tokens_seen": 922816, "step": 3435 }, { "epoch": 0.9017038007863696, "grad_norm": 1.15375554561615, "learning_rate": 4.507208387942333e-05, "loss": 0.8554, "num_input_tokens_seen": 924304, "step": 3440 }, { "epoch": 0.9030144167758847, "grad_norm": 2.2427282333374023, "learning_rate": 4.513761467889909e-05, "loss": 0.4468, "num_input_tokens_seen": 925696, "step": 3445 }, { "epoch": 0.9043250327653998, "grad_norm": 1.9167665243148804, "learning_rate": 4.520314547837484e-05, "loss": 0.762, "num_input_tokens_seen": 926688, "step": 3450 }, { "epoch": 0.9056356487549148, "grad_norm": 1.6116359233856201, "learning_rate": 4.526867627785059e-05, "loss": 0.68, "num_input_tokens_seen": 928336, "step": 3455 }, { "epoch": 0.9069462647444299, "grad_norm": 0.755664587020874, "learning_rate": 4.5334207077326345e-05, "loss": 0.3359, "num_input_tokens_seen": 929664, "step": 3460 }, { "epoch": 0.908256880733945, "grad_norm": 1.622778058052063, "learning_rate": 4.53997378768021e-05, "loss": 0.6306, "num_input_tokens_seen": 931216, "step": 3465 }, { "epoch": 0.90956749672346, "grad_norm": 1.1285862922668457, "learning_rate": 4.546526867627785e-05, "loss": 0.7095, "num_input_tokens_seen": 932512, "step": 3470 }, { "epoch": 0.9108781127129751, "grad_norm": 3.926293134689331, "learning_rate": 4.5530799475753604e-05, "loss": 0.6392, "num_input_tokens_seen": 933440, "step": 3475 }, { "epoch": 0.9121887287024901, "grad_norm": 1.2938542366027832, "learning_rate": 4.559633027522936e-05, "loss": 0.5058, "num_input_tokens_seen": 934736, "step": 3480 }, { "epoch": 0.9134993446920052, "grad_norm": 1.5012218952178955, "learning_rate": 4.566186107470511e-05, "loss": 0.3591, "num_input_tokens_seen": 935920, "step": 3485 }, { "epoch": 0.9148099606815203, "grad_norm": 2.722320795059204, "learning_rate": 4.572739187418087e-05, "loss": 0.5897, "num_input_tokens_seen": 937392, "step": 3490 }, { "epoch": 0.9161205766710354, "grad_norm": 1.2485910654067993, "learning_rate": 4.579292267365662e-05, "loss": 0.4681, "num_input_tokens_seen": 938560, "step": 3495 }, { "epoch": 0.9174311926605505, "grad_norm": 2.3634727001190186, "learning_rate": 4.585845347313238e-05, "loss": 0.6132, "num_input_tokens_seen": 939776, "step": 3500 }, { "epoch": 0.9187418086500655, "grad_norm": 6.562047958374023, "learning_rate": 4.592398427260813e-05, "loss": 0.7277, "num_input_tokens_seen": 940960, "step": 3505 }, { "epoch": 0.9200524246395806, "grad_norm": 0.9076332449913025, "learning_rate": 4.5989515072083885e-05, "loss": 0.4567, "num_input_tokens_seen": 942384, "step": 3510 }, { "epoch": 0.9213630406290957, "grad_norm": 2.4651002883911133, "learning_rate": 4.6055045871559635e-05, "loss": 0.552, "num_input_tokens_seen": 943872, "step": 3515 }, { "epoch": 0.9226736566186108, "grad_norm": 2.7189385890960693, "learning_rate": 4.6120576671035386e-05, "loss": 0.5655, "num_input_tokens_seen": 945360, "step": 3520 }, { "epoch": 0.9239842726081258, "grad_norm": 4.4070658683776855, "learning_rate": 4.618610747051114e-05, "loss": 0.4636, "num_input_tokens_seen": 946480, "step": 3525 }, { "epoch": 0.9252948885976409, "grad_norm": 2.004542827606201, "learning_rate": 4.6251638269986893e-05, "loss": 0.5247, "num_input_tokens_seen": 947744, "step": 3530 }, { "epoch": 0.926605504587156, "grad_norm": 1.8918020725250244, "learning_rate": 4.631716906946265e-05, "loss": 0.6477, "num_input_tokens_seen": 948976, "step": 3535 }, { "epoch": 0.927916120576671, "grad_norm": 1.1594223976135254, "learning_rate": 4.63826998689384e-05, "loss": 0.5908, "num_input_tokens_seen": 950384, "step": 3540 }, { "epoch": 0.9292267365661862, "grad_norm": 1.7376312017440796, "learning_rate": 4.644823066841416e-05, "loss": 0.5307, "num_input_tokens_seen": 951584, "step": 3545 }, { "epoch": 0.9305373525557011, "grad_norm": 1.504585862159729, "learning_rate": 4.651376146788991e-05, "loss": 0.6542, "num_input_tokens_seen": 952912, "step": 3550 }, { "epoch": 0.9318479685452162, "grad_norm": 1.3518586158752441, "learning_rate": 4.6579292267365666e-05, "loss": 0.7257, "num_input_tokens_seen": 954368, "step": 3555 }, { "epoch": 0.9331585845347313, "grad_norm": 1.3862948417663574, "learning_rate": 4.664482306684142e-05, "loss": 0.479, "num_input_tokens_seen": 956000, "step": 3560 }, { "epoch": 0.9344692005242464, "grad_norm": 1.299380898475647, "learning_rate": 4.671035386631717e-05, "loss": 0.526, "num_input_tokens_seen": 957328, "step": 3565 }, { "epoch": 0.9357798165137615, "grad_norm": 1.379024624824524, "learning_rate": 4.6775884665792925e-05, "loss": 0.5299, "num_input_tokens_seen": 958528, "step": 3570 }, { "epoch": 0.9370904325032765, "grad_norm": 4.457588195800781, "learning_rate": 4.684141546526868e-05, "loss": 0.8826, "num_input_tokens_seen": 959824, "step": 3575 }, { "epoch": 0.9384010484927916, "grad_norm": 1.5250426530838013, "learning_rate": 4.690694626474443e-05, "loss": 0.4874, "num_input_tokens_seen": 961232, "step": 3580 }, { "epoch": 0.9397116644823067, "grad_norm": 1.5440568923950195, "learning_rate": 4.697247706422019e-05, "loss": 0.6796, "num_input_tokens_seen": 962480, "step": 3585 }, { "epoch": 0.9410222804718218, "grad_norm": 0.6969276070594788, "learning_rate": 4.703800786369594e-05, "loss": 0.2759, "num_input_tokens_seen": 964128, "step": 3590 }, { "epoch": 0.9423328964613368, "grad_norm": 2.3999459743499756, "learning_rate": 4.710353866317169e-05, "loss": 0.5991, "num_input_tokens_seen": 965584, "step": 3595 }, { "epoch": 0.9436435124508519, "grad_norm": 2.7910265922546387, "learning_rate": 4.716906946264745e-05, "loss": 0.5176, "num_input_tokens_seen": 967136, "step": 3600 }, { "epoch": 0.944954128440367, "grad_norm": 2.6156835556030273, "learning_rate": 4.72346002621232e-05, "loss": 0.6051, "num_input_tokens_seen": 968512, "step": 3605 }, { "epoch": 0.9462647444298821, "grad_norm": 2.3243534564971924, "learning_rate": 4.730013106159895e-05, "loss": 0.7481, "num_input_tokens_seen": 969744, "step": 3610 }, { "epoch": 0.9475753604193972, "grad_norm": 1.1220234632492065, "learning_rate": 4.7365661861074707e-05, "loss": 0.3956, "num_input_tokens_seen": 970976, "step": 3615 }, { "epoch": 0.9488859764089121, "grad_norm": 2.2485506534576416, "learning_rate": 4.7431192660550464e-05, "loss": 0.4518, "num_input_tokens_seen": 972240, "step": 3620 }, { "epoch": 0.9501965923984272, "grad_norm": 3.04068922996521, "learning_rate": 4.7496723460026214e-05, "loss": 0.5815, "num_input_tokens_seen": 973904, "step": 3625 }, { "epoch": 0.9515072083879423, "grad_norm": 1.1449778079986572, "learning_rate": 4.756225425950197e-05, "loss": 0.4036, "num_input_tokens_seen": 975392, "step": 3630 }, { "epoch": 0.9528178243774574, "grad_norm": 2.6147878170013428, "learning_rate": 4.762778505897772e-05, "loss": 0.4668, "num_input_tokens_seen": 976560, "step": 3635 }, { "epoch": 0.9541284403669725, "grad_norm": 1.8183071613311768, "learning_rate": 4.769331585845347e-05, "loss": 0.7507, "num_input_tokens_seen": 977600, "step": 3640 }, { "epoch": 0.9554390563564875, "grad_norm": 6.0100884437561035, "learning_rate": 4.775884665792923e-05, "loss": 0.6251, "num_input_tokens_seen": 978896, "step": 3645 }, { "epoch": 0.9567496723460026, "grad_norm": 1.6118465662002563, "learning_rate": 4.782437745740499e-05, "loss": 0.5008, "num_input_tokens_seen": 980320, "step": 3650 }, { "epoch": 0.9580602883355177, "grad_norm": 1.6128661632537842, "learning_rate": 4.788990825688074e-05, "loss": 0.7189, "num_input_tokens_seen": 981744, "step": 3655 }, { "epoch": 0.9593709043250328, "grad_norm": 2.042003631591797, "learning_rate": 4.795543905635649e-05, "loss": 0.4507, "num_input_tokens_seen": 983088, "step": 3660 }, { "epoch": 0.9606815203145478, "grad_norm": 0.9308379888534546, "learning_rate": 4.8020969855832246e-05, "loss": 0.4173, "num_input_tokens_seen": 984320, "step": 3665 }, { "epoch": 0.9619921363040629, "grad_norm": 0.8330432176589966, "learning_rate": 4.8086500655307996e-05, "loss": 0.3997, "num_input_tokens_seen": 985520, "step": 3670 }, { "epoch": 0.963302752293578, "grad_norm": 1.0502641201019287, "learning_rate": 4.8152031454783754e-05, "loss": 0.4385, "num_input_tokens_seen": 986464, "step": 3675 }, { "epoch": 0.9646133682830931, "grad_norm": 0.896272599697113, "learning_rate": 4.8217562254259504e-05, "loss": 0.3703, "num_input_tokens_seen": 987680, "step": 3680 }, { "epoch": 0.9659239842726082, "grad_norm": 1.4144014120101929, "learning_rate": 4.8283093053735255e-05, "loss": 0.43, "num_input_tokens_seen": 989408, "step": 3685 }, { "epoch": 0.9672346002621232, "grad_norm": 2.630642890930176, "learning_rate": 4.834862385321101e-05, "loss": 0.3541, "num_input_tokens_seen": 990368, "step": 3690 }, { "epoch": 0.9685452162516383, "grad_norm": 1.2460017204284668, "learning_rate": 4.841415465268677e-05, "loss": 0.4756, "num_input_tokens_seen": 992112, "step": 3695 }, { "epoch": 0.9698558322411533, "grad_norm": 1.431044101715088, "learning_rate": 4.847968545216252e-05, "loss": 0.5093, "num_input_tokens_seen": 993488, "step": 3700 }, { "epoch": 0.9711664482306684, "grad_norm": 2.2571959495544434, "learning_rate": 4.854521625163827e-05, "loss": 0.5544, "num_input_tokens_seen": 994800, "step": 3705 }, { "epoch": 0.9724770642201835, "grad_norm": 0.9467846751213074, "learning_rate": 4.861074705111403e-05, "loss": 0.4846, "num_input_tokens_seen": 996192, "step": 3710 }, { "epoch": 0.9737876802096985, "grad_norm": 1.5490169525146484, "learning_rate": 4.867627785058978e-05, "loss": 0.4247, "num_input_tokens_seen": 997872, "step": 3715 }, { "epoch": 0.9750982961992136, "grad_norm": 3.141561269760132, "learning_rate": 4.8741808650065535e-05, "loss": 0.594, "num_input_tokens_seen": 999520, "step": 3720 }, { "epoch": 0.9764089121887287, "grad_norm": 1.1119529008865356, "learning_rate": 4.880733944954129e-05, "loss": 0.5427, "num_input_tokens_seen": 1000496, "step": 3725 }, { "epoch": 0.9777195281782438, "grad_norm": 1.5083746910095215, "learning_rate": 4.8872870249017036e-05, "loss": 0.485, "num_input_tokens_seen": 1001472, "step": 3730 }, { "epoch": 0.9790301441677588, "grad_norm": 1.5696074962615967, "learning_rate": 4.8938401048492794e-05, "loss": 0.5095, "num_input_tokens_seen": 1002688, "step": 3735 }, { "epoch": 0.9803407601572739, "grad_norm": 1.723433494567871, "learning_rate": 4.900393184796855e-05, "loss": 0.5889, "num_input_tokens_seen": 1004000, "step": 3740 }, { "epoch": 0.981651376146789, "grad_norm": 4.878201961517334, "learning_rate": 4.90694626474443e-05, "loss": 0.5998, "num_input_tokens_seen": 1005744, "step": 3745 }, { "epoch": 0.9829619921363041, "grad_norm": 0.9303051233291626, "learning_rate": 4.913499344692005e-05, "loss": 0.54, "num_input_tokens_seen": 1007344, "step": 3750 }, { "epoch": 0.9842726081258192, "grad_norm": 2.7708616256713867, "learning_rate": 4.920052424639581e-05, "loss": 0.3986, "num_input_tokens_seen": 1008736, "step": 3755 }, { "epoch": 0.9855832241153342, "grad_norm": 1.441434621810913, "learning_rate": 4.926605504587156e-05, "loss": 0.3851, "num_input_tokens_seen": 1009808, "step": 3760 }, { "epoch": 0.9868938401048493, "grad_norm": 1.2710583209991455, "learning_rate": 4.933158584534732e-05, "loss": 0.4003, "num_input_tokens_seen": 1010688, "step": 3765 }, { "epoch": 0.9882044560943644, "grad_norm": 1.6222196817398071, "learning_rate": 4.9397116644823075e-05, "loss": 0.5061, "num_input_tokens_seen": 1012320, "step": 3770 }, { "epoch": 0.9895150720838795, "grad_norm": 1.2657767534255981, "learning_rate": 4.946264744429882e-05, "loss": 0.3955, "num_input_tokens_seen": 1013328, "step": 3775 }, { "epoch": 0.9908256880733946, "grad_norm": 2.4563724994659424, "learning_rate": 4.9528178243774576e-05, "loss": 0.4538, "num_input_tokens_seen": 1014912, "step": 3780 }, { "epoch": 0.9921363040629095, "grad_norm": 1.3591992855072021, "learning_rate": 4.959370904325033e-05, "loss": 0.4016, "num_input_tokens_seen": 1016336, "step": 3785 }, { "epoch": 0.9934469200524246, "grad_norm": 0.47212669253349304, "learning_rate": 4.9659239842726083e-05, "loss": 0.606, "num_input_tokens_seen": 1017568, "step": 3790 }, { "epoch": 0.9947575360419397, "grad_norm": 1.1116055250167847, "learning_rate": 4.972477064220184e-05, "loss": 0.401, "num_input_tokens_seen": 1018736, "step": 3795 }, { "epoch": 0.9960681520314548, "grad_norm": 4.614126205444336, "learning_rate": 4.979030144167759e-05, "loss": 0.383, "num_input_tokens_seen": 1019744, "step": 3800 }, { "epoch": 0.9973787680209698, "grad_norm": 1.6825178861618042, "learning_rate": 4.985583224115334e-05, "loss": 0.5684, "num_input_tokens_seen": 1020880, "step": 3805 }, { "epoch": 0.9986893840104849, "grad_norm": 1.2593885660171509, "learning_rate": 4.99213630406291e-05, "loss": 0.7315, "num_input_tokens_seen": 1022400, "step": 3810 }, { "epoch": 1.0, "grad_norm": 6.468271255493164, "learning_rate": 4.9986893840104856e-05, "loss": 0.5745, "num_input_tokens_seen": 1023568, "step": 3815 }, { "epoch": 1.000262123197903, "eval_loss": 0.5149140357971191, "eval_runtime": 16.6568, "eval_samples_per_second": 50.91, "eval_steps_per_second": 25.455, "num_input_tokens_seen": 1023840, "step": 3816 }, { "epoch": 1.001310615989515, "grad_norm": 1.8686693906784058, "learning_rate": 4.9999998325613814e-05, "loss": 0.3699, "num_input_tokens_seen": 1024656, "step": 3820 }, { "epoch": 1.0026212319790302, "grad_norm": 1.280236005783081, "learning_rate": 4.999999152342031e-05, "loss": 0.5139, "num_input_tokens_seen": 1025904, "step": 3825 }, { "epoch": 1.0039318479685453, "grad_norm": 1.7583696842193604, "learning_rate": 4.999997948877178e-05, "loss": 0.4651, "num_input_tokens_seen": 1026816, "step": 3830 }, { "epoch": 1.0052424639580604, "grad_norm": 1.6051653623580933, "learning_rate": 4.9999962221670737e-05, "loss": 0.594, "num_input_tokens_seen": 1028176, "step": 3835 }, { "epoch": 1.0065530799475753, "grad_norm": 6.371858596801758, "learning_rate": 4.9999939722120804e-05, "loss": 0.4304, "num_input_tokens_seen": 1029232, "step": 3840 }, { "epoch": 1.0078636959370904, "grad_norm": 1.2764511108398438, "learning_rate": 4.9999911990126674e-05, "loss": 0.3921, "num_input_tokens_seen": 1030512, "step": 3845 }, { "epoch": 1.0091743119266054, "grad_norm": 2.1170363426208496, "learning_rate": 4.999987902569416e-05, "loss": 0.3623, "num_input_tokens_seen": 1031488, "step": 3850 }, { "epoch": 1.0104849279161205, "grad_norm": 0.9283241629600525, "learning_rate": 4.9999840828830167e-05, "loss": 0.3856, "num_input_tokens_seen": 1033152, "step": 3855 }, { "epoch": 1.0117955439056356, "grad_norm": 1.9830763339996338, "learning_rate": 4.999979739954269e-05, "loss": 0.4549, "num_input_tokens_seen": 1034624, "step": 3860 }, { "epoch": 1.0131061598951507, "grad_norm": 0.8171852827072144, "learning_rate": 4.999974873784081e-05, "loss": 0.4254, "num_input_tokens_seen": 1035584, "step": 3865 }, { "epoch": 1.0144167758846658, "grad_norm": 1.5159668922424316, "learning_rate": 4.9999694843734715e-05, "loss": 0.4543, "num_input_tokens_seen": 1036960, "step": 3870 }, { "epoch": 1.015727391874181, "grad_norm": 0.7735707759857178, "learning_rate": 4.999963571723568e-05, "loss": 0.5468, "num_input_tokens_seen": 1038720, "step": 3875 }, { "epoch": 1.017038007863696, "grad_norm": 0.7263953685760498, "learning_rate": 4.9999571358356095e-05, "loss": 0.414, "num_input_tokens_seen": 1040400, "step": 3880 }, { "epoch": 1.018348623853211, "grad_norm": 0.9627657532691956, "learning_rate": 4.9999501767109416e-05, "loss": 0.3922, "num_input_tokens_seen": 1042304, "step": 3885 }, { "epoch": 1.019659239842726, "grad_norm": 1.1743758916854858, "learning_rate": 4.999942694351021e-05, "loss": 0.5602, "num_input_tokens_seen": 1043728, "step": 3890 }, { "epoch": 1.020969855832241, "grad_norm": 1.268827199935913, "learning_rate": 4.9999346887574146e-05, "loss": 0.4492, "num_input_tokens_seen": 1044752, "step": 3895 }, { "epoch": 1.0222804718217562, "grad_norm": 1.292625069618225, "learning_rate": 4.999926159931797e-05, "loss": 0.5645, "num_input_tokens_seen": 1045968, "step": 3900 }, { "epoch": 1.0235910878112713, "grad_norm": 1.550226092338562, "learning_rate": 4.999917107875954e-05, "loss": 0.519, "num_input_tokens_seen": 1046848, "step": 3905 }, { "epoch": 1.0249017038007864, "grad_norm": 2.6793899536132812, "learning_rate": 4.999907532591779e-05, "loss": 0.5615, "num_input_tokens_seen": 1048208, "step": 3910 }, { "epoch": 1.0262123197903015, "grad_norm": 0.9728404879570007, "learning_rate": 4.9998974340812775e-05, "loss": 0.3562, "num_input_tokens_seen": 1049744, "step": 3915 }, { "epoch": 1.0275229357798166, "grad_norm": 1.959608793258667, "learning_rate": 4.999886812346563e-05, "loss": 0.5275, "num_input_tokens_seen": 1050816, "step": 3920 }, { "epoch": 1.0288335517693317, "grad_norm": 5.384112358093262, "learning_rate": 4.999875667389858e-05, "loss": 0.4476, "num_input_tokens_seen": 1051888, "step": 3925 }, { "epoch": 1.0301441677588468, "grad_norm": 2.3228116035461426, "learning_rate": 4.999863999213495e-05, "loss": 0.2859, "num_input_tokens_seen": 1052960, "step": 3930 }, { "epoch": 1.0314547837483616, "grad_norm": 1.7633055448532104, "learning_rate": 4.999851807819916e-05, "loss": 0.4581, "num_input_tokens_seen": 1054592, "step": 3935 }, { "epoch": 1.0327653997378767, "grad_norm": 0.7197217345237732, "learning_rate": 4.9998390932116734e-05, "loss": 0.4138, "num_input_tokens_seen": 1056080, "step": 3940 }, { "epoch": 1.0340760157273918, "grad_norm": 1.2878060340881348, "learning_rate": 4.9998258553914286e-05, "loss": 0.4525, "num_input_tokens_seen": 1057520, "step": 3945 }, { "epoch": 1.035386631716907, "grad_norm": 0.935342013835907, "learning_rate": 4.999812094361951e-05, "loss": 0.6388, "num_input_tokens_seen": 1058928, "step": 3950 }, { "epoch": 1.036697247706422, "grad_norm": 1.1796636581420898, "learning_rate": 4.9997978101261215e-05, "loss": 0.247, "num_input_tokens_seen": 1060400, "step": 3955 }, { "epoch": 1.0380078636959371, "grad_norm": 1.0787718296051025, "learning_rate": 4.9997830026869294e-05, "loss": 0.5269, "num_input_tokens_seen": 1061824, "step": 3960 }, { "epoch": 1.0393184796854522, "grad_norm": 1.0790520906448364, "learning_rate": 4.999767672047474e-05, "loss": 0.7039, "num_input_tokens_seen": 1063584, "step": 3965 }, { "epoch": 1.0406290956749673, "grad_norm": 2.2454566955566406, "learning_rate": 4.999751818210965e-05, "loss": 0.5759, "num_input_tokens_seen": 1064736, "step": 3970 }, { "epoch": 1.0419397116644824, "grad_norm": 1.4464811086654663, "learning_rate": 4.999735441180718e-05, "loss": 0.4956, "num_input_tokens_seen": 1065744, "step": 3975 }, { "epoch": 1.0432503276539973, "grad_norm": 1.7107837200164795, "learning_rate": 4.999718540960164e-05, "loss": 0.4995, "num_input_tokens_seen": 1068432, "step": 3980 }, { "epoch": 1.0445609436435124, "grad_norm": 9.214110374450684, "learning_rate": 4.999701117552839e-05, "loss": 0.5019, "num_input_tokens_seen": 1069712, "step": 3985 }, { "epoch": 1.0458715596330275, "grad_norm": 2.445868730545044, "learning_rate": 4.999683170962388e-05, "loss": 0.3592, "num_input_tokens_seen": 1071328, "step": 3990 }, { "epoch": 1.0471821756225426, "grad_norm": 1.8769757747650146, "learning_rate": 4.9996647011925685e-05, "loss": 0.6953, "num_input_tokens_seen": 1072288, "step": 3995 }, { "epoch": 1.0484927916120577, "grad_norm": 1.2512292861938477, "learning_rate": 4.9996457082472465e-05, "loss": 0.8264, "num_input_tokens_seen": 1073872, "step": 4000 }, { "epoch": 1.0498034076015728, "grad_norm": 2.5100560188293457, "learning_rate": 4.999626192130397e-05, "loss": 0.3389, "num_input_tokens_seen": 1075728, "step": 4005 }, { "epoch": 1.0511140235910879, "grad_norm": 1.2133710384368896, "learning_rate": 4.999606152846104e-05, "loss": 0.3751, "num_input_tokens_seen": 1076864, "step": 4010 }, { "epoch": 1.052424639580603, "grad_norm": 1.849005103111267, "learning_rate": 4.9995855903985624e-05, "loss": 0.3799, "num_input_tokens_seen": 1078192, "step": 4015 }, { "epoch": 1.053735255570118, "grad_norm": 1.935469627380371, "learning_rate": 4.9995645047920756e-05, "loss": 0.5599, "num_input_tokens_seen": 1079280, "step": 4020 }, { "epoch": 1.0550458715596331, "grad_norm": 1.4641064405441284, "learning_rate": 4.999542896031056e-05, "loss": 0.4837, "num_input_tokens_seen": 1080672, "step": 4025 }, { "epoch": 1.056356487549148, "grad_norm": 0.7548826336860657, "learning_rate": 4.999520764120029e-05, "loss": 0.4684, "num_input_tokens_seen": 1082208, "step": 4030 }, { "epoch": 1.0576671035386631, "grad_norm": 1.841504454612732, "learning_rate": 4.999498109063624e-05, "loss": 0.4286, "num_input_tokens_seen": 1083584, "step": 4035 }, { "epoch": 1.0589777195281782, "grad_norm": 1.560070276260376, "learning_rate": 4.999474930866583e-05, "loss": 0.452, "num_input_tokens_seen": 1084864, "step": 4040 }, { "epoch": 1.0602883355176933, "grad_norm": 1.1004029512405396, "learning_rate": 4.999451229533759e-05, "loss": 0.5035, "num_input_tokens_seen": 1085872, "step": 4045 }, { "epoch": 1.0615989515072084, "grad_norm": 0.6239252090454102, "learning_rate": 4.99942700507011e-05, "loss": 0.5486, "num_input_tokens_seen": 1087408, "step": 4050 }, { "epoch": 1.0629095674967235, "grad_norm": 1.4874398708343506, "learning_rate": 4.9994022574807085e-05, "loss": 0.5492, "num_input_tokens_seen": 1088576, "step": 4055 }, { "epoch": 1.0642201834862386, "grad_norm": 1.005002737045288, "learning_rate": 4.999376986770733e-05, "loss": 0.3757, "num_input_tokens_seen": 1090144, "step": 4060 }, { "epoch": 1.0655307994757537, "grad_norm": 9.359614372253418, "learning_rate": 4.999351192945473e-05, "loss": 0.6331, "num_input_tokens_seen": 1091296, "step": 4065 }, { "epoch": 1.0668414154652686, "grad_norm": 1.2244986295700073, "learning_rate": 4.999324876010326e-05, "loss": 0.3924, "num_input_tokens_seen": 1092448, "step": 4070 }, { "epoch": 1.0681520314547837, "grad_norm": 0.6149687170982361, "learning_rate": 4.999298035970801e-05, "loss": 0.3613, "num_input_tokens_seen": 1093552, "step": 4075 }, { "epoch": 1.0694626474442988, "grad_norm": 0.6168872117996216, "learning_rate": 4.999270672832516e-05, "loss": 0.4642, "num_input_tokens_seen": 1094640, "step": 4080 }, { "epoch": 1.0707732634338138, "grad_norm": 2.4991228580474854, "learning_rate": 4.9992427866011974e-05, "loss": 0.5494, "num_input_tokens_seen": 1096288, "step": 4085 }, { "epoch": 1.072083879423329, "grad_norm": 0.6309290528297424, "learning_rate": 4.9992143772826826e-05, "loss": 0.4818, "num_input_tokens_seen": 1097888, "step": 4090 }, { "epoch": 1.073394495412844, "grad_norm": 4.188943386077881, "learning_rate": 4.9991854448829165e-05, "loss": 0.4657, "num_input_tokens_seen": 1099248, "step": 4095 }, { "epoch": 1.0747051114023591, "grad_norm": 0.5446180105209351, "learning_rate": 4.999155989407954e-05, "loss": 0.771, "num_input_tokens_seen": 1100736, "step": 4100 }, { "epoch": 1.0760157273918742, "grad_norm": 2.995225191116333, "learning_rate": 4.999126010863963e-05, "loss": 0.3963, "num_input_tokens_seen": 1102208, "step": 4105 }, { "epoch": 1.0773263433813893, "grad_norm": 1.8120501041412354, "learning_rate": 4.999095509257214e-05, "loss": 0.4242, "num_input_tokens_seen": 1103216, "step": 4110 }, { "epoch": 1.0786369593709044, "grad_norm": 1.1895047426223755, "learning_rate": 4.999064484594095e-05, "loss": 0.3426, "num_input_tokens_seen": 1104400, "step": 4115 }, { "epoch": 1.0799475753604193, "grad_norm": 2.0030670166015625, "learning_rate": 4.999032936881096e-05, "loss": 0.7343, "num_input_tokens_seen": 1105552, "step": 4120 }, { "epoch": 1.0812581913499344, "grad_norm": 3.6586716175079346, "learning_rate": 4.999000866124822e-05, "loss": 0.7544, "num_input_tokens_seen": 1107616, "step": 4125 }, { "epoch": 1.0825688073394495, "grad_norm": 1.3761804103851318, "learning_rate": 4.9989682723319846e-05, "loss": 0.5269, "num_input_tokens_seen": 1109056, "step": 4130 }, { "epoch": 1.0838794233289646, "grad_norm": 0.767091691493988, "learning_rate": 4.9989351555094055e-05, "loss": 0.4381, "num_input_tokens_seen": 1110400, "step": 4135 }, { "epoch": 1.0851900393184797, "grad_norm": 1.8386061191558838, "learning_rate": 4.998901515664017e-05, "loss": 0.4653, "num_input_tokens_seen": 1111536, "step": 4140 }, { "epoch": 1.0865006553079948, "grad_norm": 0.7687604427337646, "learning_rate": 4.998867352802858e-05, "loss": 0.443, "num_input_tokens_seen": 1113936, "step": 4145 }, { "epoch": 1.0878112712975099, "grad_norm": 2.409327507019043, "learning_rate": 4.9988326669330797e-05, "loss": 0.4846, "num_input_tokens_seen": 1115312, "step": 4150 }, { "epoch": 1.089121887287025, "grad_norm": 1.5002739429473877, "learning_rate": 4.998797458061942e-05, "loss": 0.44, "num_input_tokens_seen": 1116384, "step": 4155 }, { "epoch": 1.09043250327654, "grad_norm": 3.055368423461914, "learning_rate": 4.998761726196814e-05, "loss": 0.6437, "num_input_tokens_seen": 1117360, "step": 4160 }, { "epoch": 1.091743119266055, "grad_norm": 1.2930845022201538, "learning_rate": 4.9987254713451754e-05, "loss": 0.3888, "num_input_tokens_seen": 1118544, "step": 4165 }, { "epoch": 1.09305373525557, "grad_norm": 1.3530598878860474, "learning_rate": 4.998688693514612e-05, "loss": 0.5856, "num_input_tokens_seen": 1119824, "step": 4170 }, { "epoch": 1.0943643512450851, "grad_norm": 0.5773869752883911, "learning_rate": 4.998651392712823e-05, "loss": 0.5513, "num_input_tokens_seen": 1121264, "step": 4175 }, { "epoch": 1.0956749672346002, "grad_norm": 1.3707647323608398, "learning_rate": 4.998613568947614e-05, "loss": 0.3421, "num_input_tokens_seen": 1122560, "step": 4180 }, { "epoch": 1.0969855832241153, "grad_norm": 1.9001654386520386, "learning_rate": 4.998575222226903e-05, "loss": 0.4724, "num_input_tokens_seen": 1123920, "step": 4185 }, { "epoch": 1.0982961992136304, "grad_norm": 0.909902811050415, "learning_rate": 4.9985363525587145e-05, "loss": 0.4059, "num_input_tokens_seen": 1125280, "step": 4190 }, { "epoch": 1.0996068152031455, "grad_norm": 1.4051512479782104, "learning_rate": 4.998496959951185e-05, "loss": 0.3734, "num_input_tokens_seen": 1127056, "step": 4195 }, { "epoch": 1.1009174311926606, "grad_norm": 1.505509376525879, "learning_rate": 4.998457044412559e-05, "loss": 0.6037, "num_input_tokens_seen": 1128720, "step": 4200 }, { "epoch": 1.1022280471821757, "grad_norm": 0.6877798438072205, "learning_rate": 4.998416605951191e-05, "loss": 0.3224, "num_input_tokens_seen": 1129968, "step": 4205 }, { "epoch": 1.1035386631716908, "grad_norm": 1.1755434274673462, "learning_rate": 4.998375644575543e-05, "loss": 0.4786, "num_input_tokens_seen": 1131408, "step": 4210 }, { "epoch": 1.1048492791612057, "grad_norm": 1.1008023023605347, "learning_rate": 4.99833416029419e-05, "loss": 0.4297, "num_input_tokens_seen": 1132400, "step": 4215 }, { "epoch": 1.1061598951507208, "grad_norm": 1.015079140663147, "learning_rate": 4.998292153115814e-05, "loss": 0.5618, "num_input_tokens_seen": 1133712, "step": 4220 }, { "epoch": 1.1074705111402359, "grad_norm": 0.7426710724830627, "learning_rate": 4.9982496230492085e-05, "loss": 0.4925, "num_input_tokens_seen": 1135232, "step": 4225 }, { "epoch": 1.108781127129751, "grad_norm": 1.6841890811920166, "learning_rate": 4.998206570103271e-05, "loss": 0.6249, "num_input_tokens_seen": 1136288, "step": 4230 }, { "epoch": 1.110091743119266, "grad_norm": 1.5601511001586914, "learning_rate": 4.998162994287017e-05, "loss": 0.4481, "num_input_tokens_seen": 1137584, "step": 4235 }, { "epoch": 1.1114023591087812, "grad_norm": 0.7838784456253052, "learning_rate": 4.998118895609564e-05, "loss": 0.4442, "num_input_tokens_seen": 1139408, "step": 4240 }, { "epoch": 1.1127129750982963, "grad_norm": 9.483351707458496, "learning_rate": 4.9980742740801425e-05, "loss": 0.6414, "num_input_tokens_seen": 1140448, "step": 4245 }, { "epoch": 1.1140235910878113, "grad_norm": 1.7240267992019653, "learning_rate": 4.998029129708092e-05, "loss": 0.4227, "num_input_tokens_seen": 1142000, "step": 4250 }, { "epoch": 1.1153342070773264, "grad_norm": 0.7027513384819031, "learning_rate": 4.997983462502861e-05, "loss": 0.4781, "num_input_tokens_seen": 1143568, "step": 4255 }, { "epoch": 1.1166448230668413, "grad_norm": 2.139810562133789, "learning_rate": 4.997937272474007e-05, "loss": 0.389, "num_input_tokens_seen": 1144800, "step": 4260 }, { "epoch": 1.1179554390563564, "grad_norm": 9.06335163116455, "learning_rate": 4.9978905596311985e-05, "loss": 0.6555, "num_input_tokens_seen": 1146080, "step": 4265 }, { "epoch": 1.1192660550458715, "grad_norm": 28.752614974975586, "learning_rate": 4.9978433239842115e-05, "loss": 0.4973, "num_input_tokens_seen": 1147488, "step": 4270 }, { "epoch": 1.1205766710353866, "grad_norm": 1.3653117418289185, "learning_rate": 4.997795565542933e-05, "loss": 1.0447, "num_input_tokens_seen": 1148368, "step": 4275 }, { "epoch": 1.1218872870249017, "grad_norm": 1.0317771434783936, "learning_rate": 4.997747284317358e-05, "loss": 0.5197, "num_input_tokens_seen": 1149808, "step": 4280 }, { "epoch": 1.1231979030144168, "grad_norm": 4.802574634552002, "learning_rate": 4.997698480317593e-05, "loss": 0.6284, "num_input_tokens_seen": 1151184, "step": 4285 }, { "epoch": 1.124508519003932, "grad_norm": 35.22478485107422, "learning_rate": 4.997649153553851e-05, "loss": 0.57, "num_input_tokens_seen": 1152288, "step": 4290 }, { "epoch": 1.125819134993447, "grad_norm": 4.537015914916992, "learning_rate": 4.9975993040364574e-05, "loss": 0.6402, "num_input_tokens_seen": 1153648, "step": 4295 }, { "epoch": 1.127129750982962, "grad_norm": 0.8938350081443787, "learning_rate": 4.9975489317758446e-05, "loss": 0.6017, "num_input_tokens_seen": 1154864, "step": 4300 }, { "epoch": 1.1284403669724772, "grad_norm": 0.9685462713241577, "learning_rate": 4.9974980367825555e-05, "loss": 0.5579, "num_input_tokens_seen": 1156608, "step": 4305 }, { "epoch": 1.129750982961992, "grad_norm": 1.9766652584075928, "learning_rate": 4.997446619067243e-05, "loss": 0.4255, "num_input_tokens_seen": 1157968, "step": 4310 }, { "epoch": 1.1310615989515072, "grad_norm": 3.73512864112854, "learning_rate": 4.997394678640669e-05, "loss": 0.7881, "num_input_tokens_seen": 1159152, "step": 4315 }, { "epoch": 1.1323722149410222, "grad_norm": 1.2350808382034302, "learning_rate": 4.997342215513703e-05, "loss": 0.5077, "num_input_tokens_seen": 1160144, "step": 4320 }, { "epoch": 1.1336828309305373, "grad_norm": 1.9417966604232788, "learning_rate": 4.9972892296973263e-05, "loss": 0.484, "num_input_tokens_seen": 1161136, "step": 4325 }, { "epoch": 1.1349934469200524, "grad_norm": 7.728499412536621, "learning_rate": 4.997235721202629e-05, "loss": 0.5822, "num_input_tokens_seen": 1162384, "step": 4330 }, { "epoch": 1.1363040629095675, "grad_norm": 0.9209542870521545, "learning_rate": 4.997181690040811e-05, "loss": 0.4863, "num_input_tokens_seen": 1163696, "step": 4335 }, { "epoch": 1.1376146788990826, "grad_norm": 3.846818447113037, "learning_rate": 4.997127136223179e-05, "loss": 0.5426, "num_input_tokens_seen": 1164864, "step": 4340 }, { "epoch": 1.1389252948885977, "grad_norm": 1.1326251029968262, "learning_rate": 4.997072059761153e-05, "loss": 0.3894, "num_input_tokens_seen": 1165968, "step": 4345 }, { "epoch": 1.1402359108781126, "grad_norm": 1.616620659828186, "learning_rate": 4.997016460666258e-05, "loss": 0.5472, "num_input_tokens_seen": 1167488, "step": 4350 }, { "epoch": 1.1415465268676277, "grad_norm": 1.1172770261764526, "learning_rate": 4.996960338950134e-05, "loss": 0.7841, "num_input_tokens_seen": 1169024, "step": 4355 }, { "epoch": 1.1428571428571428, "grad_norm": 5.275417327880859, "learning_rate": 4.996903694624525e-05, "loss": 0.3959, "num_input_tokens_seen": 1170352, "step": 4360 }, { "epoch": 1.144167758846658, "grad_norm": 1.165823221206665, "learning_rate": 4.996846527701287e-05, "loss": 0.5444, "num_input_tokens_seen": 1171808, "step": 4365 }, { "epoch": 1.145478374836173, "grad_norm": 1.1215718984603882, "learning_rate": 4.996788838192384e-05, "loss": 0.5191, "num_input_tokens_seen": 1173104, "step": 4370 }, { "epoch": 1.146788990825688, "grad_norm": 1.1373196840286255, "learning_rate": 4.9967306261098924e-05, "loss": 0.5298, "num_input_tokens_seen": 1174208, "step": 4375 }, { "epoch": 1.1480996068152032, "grad_norm": 1.1910570859909058, "learning_rate": 4.996671891465994e-05, "loss": 0.7357, "num_input_tokens_seen": 1176528, "step": 4380 }, { "epoch": 1.1494102228047183, "grad_norm": 2.3721392154693604, "learning_rate": 4.996612634272983e-05, "loss": 0.6776, "num_input_tokens_seen": 1177936, "step": 4385 }, { "epoch": 1.1507208387942334, "grad_norm": 1.0523141622543335, "learning_rate": 4.996552854543262e-05, "loss": 0.3928, "num_input_tokens_seen": 1178896, "step": 4390 }, { "epoch": 1.1520314547837485, "grad_norm": 2.108607292175293, "learning_rate": 4.9964925522893416e-05, "loss": 0.6931, "num_input_tokens_seen": 1180160, "step": 4395 }, { "epoch": 1.1533420707732633, "grad_norm": 1.2326674461364746, "learning_rate": 4.9964317275238435e-05, "loss": 0.3487, "num_input_tokens_seen": 1181648, "step": 4400 }, { "epoch": 1.1546526867627784, "grad_norm": 0.9787944555282593, "learning_rate": 4.996370380259499e-05, "loss": 0.9172, "num_input_tokens_seen": 1182896, "step": 4405 }, { "epoch": 1.1559633027522935, "grad_norm": 1.3900665044784546, "learning_rate": 4.996308510509147e-05, "loss": 0.6454, "num_input_tokens_seen": 1184112, "step": 4410 }, { "epoch": 1.1572739187418086, "grad_norm": 1.1860154867172241, "learning_rate": 4.9962461182857366e-05, "loss": 0.3236, "num_input_tokens_seen": 1185312, "step": 4415 }, { "epoch": 1.1585845347313237, "grad_norm": 4.458273410797119, "learning_rate": 4.9961832036023275e-05, "loss": 0.3794, "num_input_tokens_seen": 1186336, "step": 4420 }, { "epoch": 1.1598951507208388, "grad_norm": 0.991655170917511, "learning_rate": 4.996119766472087e-05, "loss": 0.6543, "num_input_tokens_seen": 1187776, "step": 4425 }, { "epoch": 1.161205766710354, "grad_norm": 3.4048075675964355, "learning_rate": 4.996055806908292e-05, "loss": 0.4693, "num_input_tokens_seen": 1189040, "step": 4430 }, { "epoch": 1.162516382699869, "grad_norm": 1.815104365348816, "learning_rate": 4.99599132492433e-05, "loss": 0.4407, "num_input_tokens_seen": 1190272, "step": 4435 }, { "epoch": 1.163826998689384, "grad_norm": 1.1027878522872925, "learning_rate": 4.995926320533695e-05, "loss": 0.5154, "num_input_tokens_seen": 1191472, "step": 4440 }, { "epoch": 1.165137614678899, "grad_norm": 0.6620709300041199, "learning_rate": 4.995860793749995e-05, "loss": 0.4608, "num_input_tokens_seen": 1192832, "step": 4445 }, { "epoch": 1.166448230668414, "grad_norm": 0.7143833041191101, "learning_rate": 4.995794744586942e-05, "loss": 0.5844, "num_input_tokens_seen": 1194288, "step": 4450 }, { "epoch": 1.1677588466579292, "grad_norm": 0.6483161449432373, "learning_rate": 4.9957281730583635e-05, "loss": 0.5588, "num_input_tokens_seen": 1195392, "step": 4455 }, { "epoch": 1.1690694626474443, "grad_norm": 1.4145514965057373, "learning_rate": 4.9956610791781885e-05, "loss": 0.4973, "num_input_tokens_seen": 1196880, "step": 4460 }, { "epoch": 1.1703800786369594, "grad_norm": 1.945336103439331, "learning_rate": 4.995593462960464e-05, "loss": 0.5428, "num_input_tokens_seen": 1198064, "step": 4465 }, { "epoch": 1.1716906946264745, "grad_norm": 1.4219011068344116, "learning_rate": 4.9955253244193375e-05, "loss": 0.5483, "num_input_tokens_seen": 1199152, "step": 4470 }, { "epoch": 1.1730013106159896, "grad_norm": 1.4455604553222656, "learning_rate": 4.9954566635690744e-05, "loss": 0.5169, "num_input_tokens_seen": 1200256, "step": 4475 }, { "epoch": 1.1743119266055047, "grad_norm": 0.8753124475479126, "learning_rate": 4.995387480424043e-05, "loss": 0.5995, "num_input_tokens_seen": 1201936, "step": 4480 }, { "epoch": 1.1756225425950197, "grad_norm": 0.9920641183853149, "learning_rate": 4.995317774998723e-05, "loss": 0.4241, "num_input_tokens_seen": 1205088, "step": 4485 }, { "epoch": 1.1769331585845348, "grad_norm": 1.253334879875183, "learning_rate": 4.995247547307704e-05, "loss": 0.4149, "num_input_tokens_seen": 1206176, "step": 4490 }, { "epoch": 1.1782437745740497, "grad_norm": 0.44011181592941284, "learning_rate": 4.995176797365686e-05, "loss": 0.4627, "num_input_tokens_seen": 1207568, "step": 4495 }, { "epoch": 1.1795543905635648, "grad_norm": 1.9314247369766235, "learning_rate": 4.995105525187475e-05, "loss": 0.3372, "num_input_tokens_seen": 1208816, "step": 4500 }, { "epoch": 1.18086500655308, "grad_norm": 1.2268073558807373, "learning_rate": 4.9950337307879884e-05, "loss": 0.5801, "num_input_tokens_seen": 1209952, "step": 4505 }, { "epoch": 1.182175622542595, "grad_norm": 0.6963161826133728, "learning_rate": 4.994961414182254e-05, "loss": 0.3691, "num_input_tokens_seen": 1211408, "step": 4510 }, { "epoch": 1.18348623853211, "grad_norm": 1.1661295890808105, "learning_rate": 4.9948885753854056e-05, "loss": 0.3369, "num_input_tokens_seen": 1212816, "step": 4515 }, { "epoch": 1.1847968545216252, "grad_norm": 1.0930291414260864, "learning_rate": 4.99481521441269e-05, "loss": 0.4933, "num_input_tokens_seen": 1214048, "step": 4520 }, { "epoch": 1.1861074705111403, "grad_norm": 1.3197475671768188, "learning_rate": 4.99474133127946e-05, "loss": 0.5292, "num_input_tokens_seen": 1215360, "step": 4525 }, { "epoch": 1.1874180865006554, "grad_norm": 1.1987184286117554, "learning_rate": 4.99466692600118e-05, "loss": 0.7782, "num_input_tokens_seen": 1216352, "step": 4530 }, { "epoch": 1.1887287024901703, "grad_norm": 0.7648068070411682, "learning_rate": 4.994591998593423e-05, "loss": 0.5001, "num_input_tokens_seen": 1217680, "step": 4535 }, { "epoch": 1.1900393184796854, "grad_norm": 2.009330987930298, "learning_rate": 4.9945165490718714e-05, "loss": 0.4417, "num_input_tokens_seen": 1219472, "step": 4540 }, { "epoch": 1.1913499344692005, "grad_norm": 2.9883131980895996, "learning_rate": 4.994440577452316e-05, "loss": 0.447, "num_input_tokens_seen": 1221008, "step": 4545 }, { "epoch": 1.1926605504587156, "grad_norm": 2.44889497756958, "learning_rate": 4.9943640837506574e-05, "loss": 0.6817, "num_input_tokens_seen": 1222016, "step": 4550 }, { "epoch": 1.1939711664482306, "grad_norm": 1.2081036567687988, "learning_rate": 4.994287067982907e-05, "loss": 0.5502, "num_input_tokens_seen": 1223136, "step": 4555 }, { "epoch": 1.1952817824377457, "grad_norm": 1.665197730064392, "learning_rate": 4.994209530165183e-05, "loss": 0.9779, "num_input_tokens_seen": 1224368, "step": 4560 }, { "epoch": 1.1965923984272608, "grad_norm": 1.050563931465149, "learning_rate": 4.9941314703137134e-05, "loss": 0.4841, "num_input_tokens_seen": 1226448, "step": 4565 }, { "epoch": 1.197903014416776, "grad_norm": 0.910430908203125, "learning_rate": 4.994052888444837e-05, "loss": 0.4677, "num_input_tokens_seen": 1227792, "step": 4570 }, { "epoch": 1.199213630406291, "grad_norm": 4.4951558113098145, "learning_rate": 4.993973784575e-05, "loss": 0.3899, "num_input_tokens_seen": 1229056, "step": 4575 }, { "epoch": 1.2005242463958061, "grad_norm": 1.38774836063385, "learning_rate": 4.9938941587207586e-05, "loss": 0.4989, "num_input_tokens_seen": 1230272, "step": 4580 }, { "epoch": 1.2018348623853212, "grad_norm": 0.605683445930481, "learning_rate": 4.9938140108987795e-05, "loss": 0.7254, "num_input_tokens_seen": 1231648, "step": 4585 }, { "epoch": 1.203145478374836, "grad_norm": 0.9556915760040283, "learning_rate": 4.9937333411258366e-05, "loss": 0.4665, "num_input_tokens_seen": 1232736, "step": 4590 }, { "epoch": 1.2044560943643512, "grad_norm": 1.7923600673675537, "learning_rate": 4.993652149418815e-05, "loss": 0.4013, "num_input_tokens_seen": 1233904, "step": 4595 }, { "epoch": 1.2057667103538663, "grad_norm": 0.7912031412124634, "learning_rate": 4.9935704357947065e-05, "loss": 0.609, "num_input_tokens_seen": 1235232, "step": 4600 }, { "epoch": 1.2070773263433814, "grad_norm": 1.342315673828125, "learning_rate": 4.9934882002706154e-05, "loss": 0.4589, "num_input_tokens_seen": 1236608, "step": 4605 }, { "epoch": 1.2083879423328965, "grad_norm": 0.8676767945289612, "learning_rate": 4.9934054428637514e-05, "loss": 0.6659, "num_input_tokens_seen": 1237712, "step": 4610 }, { "epoch": 1.2096985583224116, "grad_norm": 4.962038516998291, "learning_rate": 4.9933221635914365e-05, "loss": 0.5159, "num_input_tokens_seen": 1239152, "step": 4615 }, { "epoch": 1.2110091743119267, "grad_norm": 1.9116156101226807, "learning_rate": 4.9932383624711014e-05, "loss": 0.3265, "num_input_tokens_seen": 1240784, "step": 4620 }, { "epoch": 1.2123197903014418, "grad_norm": 3.999201536178589, "learning_rate": 4.993154039520285e-05, "loss": 0.6078, "num_input_tokens_seen": 1242752, "step": 4625 }, { "epoch": 1.2136304062909566, "grad_norm": 1.3416532278060913, "learning_rate": 4.9930691947566354e-05, "loss": 0.3944, "num_input_tokens_seen": 1244176, "step": 4630 }, { "epoch": 1.2149410222804717, "grad_norm": 1.2804819345474243, "learning_rate": 4.992983828197911e-05, "loss": 0.3481, "num_input_tokens_seen": 1245376, "step": 4635 }, { "epoch": 1.2162516382699868, "grad_norm": 2.228559970855713, "learning_rate": 4.99289793986198e-05, "loss": 0.4792, "num_input_tokens_seen": 1246544, "step": 4640 }, { "epoch": 1.217562254259502, "grad_norm": 0.8631333112716675, "learning_rate": 4.992811529766816e-05, "loss": 0.4523, "num_input_tokens_seen": 1248016, "step": 4645 }, { "epoch": 1.218872870249017, "grad_norm": 1.8554017543792725, "learning_rate": 4.992724597930508e-05, "loss": 0.5155, "num_input_tokens_seen": 1249392, "step": 4650 }, { "epoch": 1.2201834862385321, "grad_norm": 0.6502385139465332, "learning_rate": 4.9926371443712474e-05, "loss": 0.4617, "num_input_tokens_seen": 1250960, "step": 4655 }, { "epoch": 1.2214941022280472, "grad_norm": 2.4360740184783936, "learning_rate": 4.9925491691073396e-05, "loss": 0.4248, "num_input_tokens_seen": 1252320, "step": 4660 }, { "epoch": 1.2228047182175623, "grad_norm": 1.355713963508606, "learning_rate": 4.9924606721571975e-05, "loss": 0.4757, "num_input_tokens_seen": 1254080, "step": 4665 }, { "epoch": 1.2241153342070774, "grad_norm": 0.6427122950553894, "learning_rate": 4.9923716535393434e-05, "loss": 0.5979, "num_input_tokens_seen": 1255168, "step": 4670 }, { "epoch": 1.2254259501965925, "grad_norm": 0.9243772625923157, "learning_rate": 4.9922821132724086e-05, "loss": 0.5548, "num_input_tokens_seen": 1256528, "step": 4675 }, { "epoch": 1.2267365661861074, "grad_norm": 1.4700371026992798, "learning_rate": 4.992192051375134e-05, "loss": 0.4897, "num_input_tokens_seen": 1257808, "step": 4680 }, { "epoch": 1.2280471821756225, "grad_norm": 2.9411213397979736, "learning_rate": 4.992101467866369e-05, "loss": 0.5646, "num_input_tokens_seen": 1259040, "step": 4685 }, { "epoch": 1.2293577981651376, "grad_norm": 1.642642617225647, "learning_rate": 4.9920103627650734e-05, "loss": 0.5418, "num_input_tokens_seen": 1260592, "step": 4690 }, { "epoch": 1.2306684141546527, "grad_norm": 3.3916871547698975, "learning_rate": 4.991918736090314e-05, "loss": 0.6362, "num_input_tokens_seen": 1261680, "step": 4695 }, { "epoch": 1.2319790301441678, "grad_norm": 4.231936931610107, "learning_rate": 4.991826587861269e-05, "loss": 0.635, "num_input_tokens_seen": 1262816, "step": 4700 }, { "epoch": 1.2332896461336829, "grad_norm": 0.7245665788650513, "learning_rate": 4.991733918097225e-05, "loss": 0.5849, "num_input_tokens_seen": 1264464, "step": 4705 }, { "epoch": 1.234600262123198, "grad_norm": 1.6168075799942017, "learning_rate": 4.9916407268175775e-05, "loss": 0.4652, "num_input_tokens_seen": 1265536, "step": 4710 }, { "epoch": 1.235910878112713, "grad_norm": 0.7353128790855408, "learning_rate": 4.991547014041831e-05, "loss": 0.4704, "num_input_tokens_seen": 1267104, "step": 4715 }, { "epoch": 1.237221494102228, "grad_norm": 1.3222131729125977, "learning_rate": 4.9914527797895995e-05, "loss": 0.4375, "num_input_tokens_seen": 1268256, "step": 4720 }, { "epoch": 1.238532110091743, "grad_norm": 4.194340705871582, "learning_rate": 4.9913580240806054e-05, "loss": 0.619, "num_input_tokens_seen": 1269184, "step": 4725 }, { "epoch": 1.2398427260812581, "grad_norm": 0.9990293979644775, "learning_rate": 4.991262746934682e-05, "loss": 0.4748, "num_input_tokens_seen": 1270768, "step": 4730 }, { "epoch": 1.2411533420707732, "grad_norm": 1.3202980756759644, "learning_rate": 4.991166948371771e-05, "loss": 0.413, "num_input_tokens_seen": 1272336, "step": 4735 }, { "epoch": 1.2424639580602883, "grad_norm": 0.8801778554916382, "learning_rate": 4.991070628411921e-05, "loss": 0.3629, "num_input_tokens_seen": 1273520, "step": 4740 }, { "epoch": 1.2437745740498034, "grad_norm": 1.2366838455200195, "learning_rate": 4.9909737870752935e-05, "loss": 0.3915, "num_input_tokens_seen": 1274976, "step": 4745 }, { "epoch": 1.2450851900393185, "grad_norm": 1.0465301275253296, "learning_rate": 4.990876424382156e-05, "loss": 0.5218, "num_input_tokens_seen": 1276096, "step": 4750 }, { "epoch": 1.2463958060288336, "grad_norm": 1.276018500328064, "learning_rate": 4.9907785403528863e-05, "loss": 0.3155, "num_input_tokens_seen": 1278192, "step": 4755 }, { "epoch": 1.2477064220183487, "grad_norm": 0.5353769659996033, "learning_rate": 4.9906801350079724e-05, "loss": 0.3998, "num_input_tokens_seen": 1279520, "step": 4760 }, { "epoch": 1.2490170380078638, "grad_norm": 2.3472845554351807, "learning_rate": 4.9905812083680105e-05, "loss": 0.5643, "num_input_tokens_seen": 1280608, "step": 4765 }, { "epoch": 1.2503276539973789, "grad_norm": 2.4411816596984863, "learning_rate": 4.990481760453704e-05, "loss": 0.4226, "num_input_tokens_seen": 1282416, "step": 4770 }, { "epoch": 1.2516382699868938, "grad_norm": 3.2240447998046875, "learning_rate": 4.9903817912858684e-05, "loss": 0.6281, "num_input_tokens_seen": 1283456, "step": 4775 }, { "epoch": 1.2529488859764089, "grad_norm": 1.4853649139404297, "learning_rate": 4.9902813008854274e-05, "loss": 0.5356, "num_input_tokens_seen": 1284624, "step": 4780 }, { "epoch": 1.254259501965924, "grad_norm": 1.4789369106292725, "learning_rate": 4.990180289273413e-05, "loss": 0.5191, "num_input_tokens_seen": 1285888, "step": 4785 }, { "epoch": 1.255570117955439, "grad_norm": 1.4205297231674194, "learning_rate": 4.990078756470966e-05, "loss": 0.3981, "num_input_tokens_seen": 1287184, "step": 4790 }, { "epoch": 1.2568807339449541, "grad_norm": 1.8537791967391968, "learning_rate": 4.989976702499339e-05, "loss": 0.4957, "num_input_tokens_seen": 1288592, "step": 4795 }, { "epoch": 1.2581913499344692, "grad_norm": 1.4361358880996704, "learning_rate": 4.98987412737989e-05, "loss": 0.7347, "num_input_tokens_seen": 1289744, "step": 4800 }, { "epoch": 1.2595019659239843, "grad_norm": 1.6070441007614136, "learning_rate": 4.989771031134088e-05, "loss": 0.5614, "num_input_tokens_seen": 1290960, "step": 4805 }, { "epoch": 1.2608125819134992, "grad_norm": 2.210671901702881, "learning_rate": 4.989667413783511e-05, "loss": 0.4958, "num_input_tokens_seen": 1292208, "step": 4810 }, { "epoch": 1.2621231979030143, "grad_norm": 0.783543586730957, "learning_rate": 4.9895632753498465e-05, "loss": 0.5419, "num_input_tokens_seen": 1293472, "step": 4815 }, { "epoch": 1.2634338138925294, "grad_norm": 1.0579088926315308, "learning_rate": 4.98945861585489e-05, "loss": 0.3846, "num_input_tokens_seen": 1294640, "step": 4820 }, { "epoch": 1.2647444298820445, "grad_norm": 2.7245538234710693, "learning_rate": 4.989353435320547e-05, "loss": 0.5074, "num_input_tokens_seen": 1296096, "step": 4825 }, { "epoch": 1.2660550458715596, "grad_norm": 2.875556468963623, "learning_rate": 4.9892477337688306e-05, "loss": 0.5357, "num_input_tokens_seen": 1297200, "step": 4830 }, { "epoch": 1.2673656618610747, "grad_norm": 0.7214925289154053, "learning_rate": 4.9891415112218654e-05, "loss": 0.3462, "num_input_tokens_seen": 1298512, "step": 4835 }, { "epoch": 1.2686762778505898, "grad_norm": 2.092400312423706, "learning_rate": 4.9890347677018826e-05, "loss": 0.5125, "num_input_tokens_seen": 1299616, "step": 4840 }, { "epoch": 1.2699868938401049, "grad_norm": 0.6186462640762329, "learning_rate": 4.9889275032312235e-05, "loss": 0.4071, "num_input_tokens_seen": 1300784, "step": 4845 }, { "epoch": 1.27129750982962, "grad_norm": 0.780430793762207, "learning_rate": 4.988819717832338e-05, "loss": 0.4788, "num_input_tokens_seen": 1302304, "step": 4850 }, { "epoch": 1.272608125819135, "grad_norm": 1.0290547609329224, "learning_rate": 4.988711411527786e-05, "loss": 0.4595, "num_input_tokens_seen": 1304224, "step": 4855 }, { "epoch": 1.2739187418086502, "grad_norm": 0.7073336839675903, "learning_rate": 4.988602584340236e-05, "loss": 0.614, "num_input_tokens_seen": 1305584, "step": 4860 }, { "epoch": 1.2752293577981653, "grad_norm": 0.8128321170806885, "learning_rate": 4.988493236292465e-05, "loss": 0.4867, "num_input_tokens_seen": 1306848, "step": 4865 }, { "epoch": 1.2765399737876801, "grad_norm": 1.5303391218185425, "learning_rate": 4.98838336740736e-05, "loss": 0.5443, "num_input_tokens_seen": 1308064, "step": 4870 }, { "epoch": 1.2778505897771952, "grad_norm": 0.7519951462745667, "learning_rate": 4.9882729777079154e-05, "loss": 0.3898, "num_input_tokens_seen": 1309152, "step": 4875 }, { "epoch": 1.2791612057667103, "grad_norm": 1.6224100589752197, "learning_rate": 4.9881620672172355e-05, "loss": 0.5063, "num_input_tokens_seen": 1310352, "step": 4880 }, { "epoch": 1.2804718217562254, "grad_norm": 1.0204194784164429, "learning_rate": 4.988050635958535e-05, "loss": 0.4255, "num_input_tokens_seen": 1311408, "step": 4885 }, { "epoch": 1.2817824377457405, "grad_norm": 1.4285943508148193, "learning_rate": 4.987938683955135e-05, "loss": 0.5127, "num_input_tokens_seen": 1312656, "step": 4890 }, { "epoch": 1.2830930537352556, "grad_norm": 1.8321727514266968, "learning_rate": 4.987826211230467e-05, "loss": 0.5074, "num_input_tokens_seen": 1314096, "step": 4895 }, { "epoch": 1.2844036697247707, "grad_norm": 0.7773303985595703, "learning_rate": 4.987713217808072e-05, "loss": 0.3186, "num_input_tokens_seen": 1315168, "step": 4900 }, { "epoch": 1.2857142857142856, "grad_norm": 0.5115456581115723, "learning_rate": 4.987599703711599e-05, "loss": 0.6232, "num_input_tokens_seen": 1316592, "step": 4905 }, { "epoch": 1.2870249017038007, "grad_norm": 3.0620837211608887, "learning_rate": 4.9874856689648065e-05, "loss": 0.4972, "num_input_tokens_seen": 1317488, "step": 4910 }, { "epoch": 1.2883355176933158, "grad_norm": 1.1962873935699463, "learning_rate": 4.9873711135915605e-05, "loss": 0.3169, "num_input_tokens_seen": 1319216, "step": 4915 }, { "epoch": 1.2896461336828309, "grad_norm": 1.2171231508255005, "learning_rate": 4.987256037615839e-05, "loss": 0.5332, "num_input_tokens_seen": 1320320, "step": 4920 }, { "epoch": 1.290956749672346, "grad_norm": 1.1007181406021118, "learning_rate": 4.987140441061726e-05, "loss": 0.3922, "num_input_tokens_seen": 1321664, "step": 4925 }, { "epoch": 1.292267365661861, "grad_norm": 1.397943377494812, "learning_rate": 4.987024323953417e-05, "loss": 0.3127, "num_input_tokens_seen": 1323040, "step": 4930 }, { "epoch": 1.2935779816513762, "grad_norm": 0.9013969898223877, "learning_rate": 4.9869076863152134e-05, "loss": 0.3669, "num_input_tokens_seen": 1324224, "step": 4935 }, { "epoch": 1.2948885976408913, "grad_norm": 0.8513376712799072, "learning_rate": 4.986790528171529e-05, "loss": 0.3132, "num_input_tokens_seen": 1325744, "step": 4940 }, { "epoch": 1.2961992136304064, "grad_norm": 0.654806911945343, "learning_rate": 4.986672849546883e-05, "loss": 0.463, "num_input_tokens_seen": 1327136, "step": 4945 }, { "epoch": 1.2975098296199215, "grad_norm": 1.3261380195617676, "learning_rate": 4.9865546504659063e-05, "loss": 0.5526, "num_input_tokens_seen": 1329040, "step": 4950 }, { "epoch": 1.2988204456094365, "grad_norm": 0.7423044443130493, "learning_rate": 4.986435930953338e-05, "loss": 0.5139, "num_input_tokens_seen": 1330512, "step": 4955 }, { "epoch": 1.3001310615989516, "grad_norm": 1.2502152919769287, "learning_rate": 4.986316691034026e-05, "loss": 0.3546, "num_input_tokens_seen": 1331408, "step": 4960 }, { "epoch": 1.3014416775884665, "grad_norm": 1.5479899644851685, "learning_rate": 4.986196930732926e-05, "loss": 0.6419, "num_input_tokens_seen": 1332512, "step": 4965 }, { "epoch": 1.3027522935779816, "grad_norm": 0.822586178779602, "learning_rate": 4.986076650075104e-05, "loss": 0.301, "num_input_tokens_seen": 1334048, "step": 4970 }, { "epoch": 1.3040629095674967, "grad_norm": 1.588680624961853, "learning_rate": 4.985955849085735e-05, "loss": 0.6515, "num_input_tokens_seen": 1335168, "step": 4975 }, { "epoch": 1.3053735255570118, "grad_norm": 0.6667677760124207, "learning_rate": 4.9858345277901034e-05, "loss": 0.3537, "num_input_tokens_seen": 1336624, "step": 4980 }, { "epoch": 1.306684141546527, "grad_norm": 1.9958804845809937, "learning_rate": 4.9857126862135984e-05, "loss": 0.4691, "num_input_tokens_seen": 1337728, "step": 4985 }, { "epoch": 1.307994757536042, "grad_norm": 1.3039921522140503, "learning_rate": 4.985590324381724e-05, "loss": 0.4551, "num_input_tokens_seen": 1339136, "step": 4990 }, { "epoch": 1.309305373525557, "grad_norm": 1.505985140800476, "learning_rate": 4.98546744232009e-05, "loss": 0.6262, "num_input_tokens_seen": 1340784, "step": 4995 }, { "epoch": 1.310615989515072, "grad_norm": 1.7048349380493164, "learning_rate": 4.9853440400544136e-05, "loss": 0.3113, "num_input_tokens_seen": 1342384, "step": 5000 }, { "epoch": 1.311926605504587, "grad_norm": 1.1584246158599854, "learning_rate": 4.985220117610525e-05, "loss": 0.425, "num_input_tokens_seen": 1343952, "step": 5005 }, { "epoch": 1.3132372214941022, "grad_norm": 0.7541625499725342, "learning_rate": 4.985095675014359e-05, "loss": 0.3113, "num_input_tokens_seen": 1345600, "step": 5010 }, { "epoch": 1.3145478374836173, "grad_norm": 0.8177048563957214, "learning_rate": 4.984970712291963e-05, "loss": 0.5669, "num_input_tokens_seen": 1347216, "step": 5015 }, { "epoch": 1.3158584534731324, "grad_norm": 0.571933925151825, "learning_rate": 4.98484522946949e-05, "loss": 0.439, "num_input_tokens_seen": 1348368, "step": 5020 }, { "epoch": 1.3171690694626474, "grad_norm": 1.2217748165130615, "learning_rate": 4.984719226573205e-05, "loss": 0.4751, "num_input_tokens_seen": 1349504, "step": 5025 }, { "epoch": 1.3184796854521625, "grad_norm": 1.5772972106933594, "learning_rate": 4.984592703629478e-05, "loss": 0.5976, "num_input_tokens_seen": 1350688, "step": 5030 }, { "epoch": 1.3197903014416776, "grad_norm": 0.8667944073677063, "learning_rate": 4.9844656606647924e-05, "loss": 0.5997, "num_input_tokens_seen": 1352160, "step": 5035 }, { "epoch": 1.3211009174311927, "grad_norm": 0.8807387948036194, "learning_rate": 4.984338097705736e-05, "loss": 0.38, "num_input_tokens_seen": 1353504, "step": 5040 }, { "epoch": 1.3224115334207078, "grad_norm": 0.5245699286460876, "learning_rate": 4.984210014779008e-05, "loss": 0.4686, "num_input_tokens_seen": 1354992, "step": 5045 }, { "epoch": 1.323722149410223, "grad_norm": 5.562580108642578, "learning_rate": 4.984081411911417e-05, "loss": 0.5442, "num_input_tokens_seen": 1355984, "step": 5050 }, { "epoch": 1.3250327653997378, "grad_norm": 0.7142163515090942, "learning_rate": 4.9839522891298784e-05, "loss": 0.4396, "num_input_tokens_seen": 1357360, "step": 5055 }, { "epoch": 1.326343381389253, "grad_norm": 1.821625828742981, "learning_rate": 4.983822646461417e-05, "loss": 0.54, "num_input_tokens_seen": 1358416, "step": 5060 }, { "epoch": 1.327653997378768, "grad_norm": 0.6053028106689453, "learning_rate": 4.9836924839331686e-05, "loss": 0.632, "num_input_tokens_seen": 1359824, "step": 5065 }, { "epoch": 1.328964613368283, "grad_norm": 1.7480442523956299, "learning_rate": 4.983561801572374e-05, "loss": 0.5001, "num_input_tokens_seen": 1361056, "step": 5070 }, { "epoch": 1.3302752293577982, "grad_norm": 1.2072457075119019, "learning_rate": 4.983430599406386e-05, "loss": 0.5386, "num_input_tokens_seen": 1362048, "step": 5075 }, { "epoch": 1.3315858453473133, "grad_norm": 0.8014470934867859, "learning_rate": 4.983298877462664e-05, "loss": 0.3528, "num_input_tokens_seen": 1362960, "step": 5080 }, { "epoch": 1.3328964613368284, "grad_norm": 0.5434121489524841, "learning_rate": 4.983166635768778e-05, "loss": 0.3358, "num_input_tokens_seen": 1364192, "step": 5085 }, { "epoch": 1.3342070773263432, "grad_norm": 0.6086885333061218, "learning_rate": 4.9830338743524064e-05, "loss": 0.4566, "num_input_tokens_seen": 1366240, "step": 5090 }, { "epoch": 1.3355176933158583, "grad_norm": 1.173873782157898, "learning_rate": 4.982900593241334e-05, "loss": 0.4429, "num_input_tokens_seen": 1367584, "step": 5095 }, { "epoch": 1.3368283093053734, "grad_norm": 1.2639012336730957, "learning_rate": 4.982766792463458e-05, "loss": 0.4435, "num_input_tokens_seen": 1369152, "step": 5100 }, { "epoch": 1.3381389252948885, "grad_norm": 2.4236795902252197, "learning_rate": 4.9826324720467834e-05, "loss": 0.4057, "num_input_tokens_seen": 1370256, "step": 5105 }, { "epoch": 1.3394495412844036, "grad_norm": 3.274639129638672, "learning_rate": 4.982497632019421e-05, "loss": 1.0208, "num_input_tokens_seen": 1373184, "step": 5110 }, { "epoch": 1.3407601572739187, "grad_norm": 1.9863746166229248, "learning_rate": 4.9823622724095936e-05, "loss": 0.6557, "num_input_tokens_seen": 1374544, "step": 5115 }, { "epoch": 1.3420707732634338, "grad_norm": 1.7269021272659302, "learning_rate": 4.982226393245632e-05, "loss": 0.5711, "num_input_tokens_seen": 1375808, "step": 5120 }, { "epoch": 1.343381389252949, "grad_norm": 0.8907382488250732, "learning_rate": 4.982089994555975e-05, "loss": 0.5467, "num_input_tokens_seen": 1377216, "step": 5125 }, { "epoch": 1.344692005242464, "grad_norm": 1.6904405355453491, "learning_rate": 4.981953076369171e-05, "loss": 0.5458, "num_input_tokens_seen": 1378448, "step": 5130 }, { "epoch": 1.3460026212319791, "grad_norm": 0.8438498973846436, "learning_rate": 4.981815638713877e-05, "loss": 0.6452, "num_input_tokens_seen": 1379728, "step": 5135 }, { "epoch": 1.3473132372214942, "grad_norm": 2.2139389514923096, "learning_rate": 4.981677681618858e-05, "loss": 0.6321, "num_input_tokens_seen": 1380880, "step": 5140 }, { "epoch": 1.3486238532110093, "grad_norm": 1.7017005681991577, "learning_rate": 4.981539205112988e-05, "loss": 0.7648, "num_input_tokens_seen": 1382672, "step": 5145 }, { "epoch": 1.3499344692005242, "grad_norm": 1.7518823146820068, "learning_rate": 4.98140020922525e-05, "loss": 0.4952, "num_input_tokens_seen": 1384176, "step": 5150 }, { "epoch": 1.3512450851900393, "grad_norm": 1.240859031677246, "learning_rate": 4.9812606939847356e-05, "loss": 0.5321, "num_input_tokens_seen": 1385440, "step": 5155 }, { "epoch": 1.3525557011795544, "grad_norm": 1.0636147260665894, "learning_rate": 4.981120659420646e-05, "loss": 0.5374, "num_input_tokens_seen": 1387040, "step": 5160 }, { "epoch": 1.3538663171690695, "grad_norm": 0.6412643790245056, "learning_rate": 4.9809801055622887e-05, "loss": 0.5066, "num_input_tokens_seen": 1388048, "step": 5165 }, { "epoch": 1.3551769331585846, "grad_norm": 1.2036141157150269, "learning_rate": 4.980839032439082e-05, "loss": 0.6217, "num_input_tokens_seen": 1389072, "step": 5170 }, { "epoch": 1.3564875491480997, "grad_norm": 1.324585199356079, "learning_rate": 4.980697440080553e-05, "loss": 0.4046, "num_input_tokens_seen": 1390720, "step": 5175 }, { "epoch": 1.3577981651376148, "grad_norm": 0.78847736120224, "learning_rate": 4.980555328516335e-05, "loss": 0.2337, "num_input_tokens_seen": 1392096, "step": 5180 }, { "epoch": 1.3591087811271296, "grad_norm": 0.9229063987731934, "learning_rate": 4.9804126977761747e-05, "loss": 0.3974, "num_input_tokens_seen": 1393376, "step": 5185 }, { "epoch": 1.3604193971166447, "grad_norm": 0.8572607636451721, "learning_rate": 4.980269547889921e-05, "loss": 0.5202, "num_input_tokens_seen": 1394704, "step": 5190 }, { "epoch": 1.3617300131061598, "grad_norm": 1.2383371591567993, "learning_rate": 4.980125878887537e-05, "loss": 0.6242, "num_input_tokens_seen": 1395888, "step": 5195 }, { "epoch": 1.363040629095675, "grad_norm": 0.8132635354995728, "learning_rate": 4.979981690799092e-05, "loss": 0.6619, "num_input_tokens_seen": 1397280, "step": 5200 }, { "epoch": 1.36435124508519, "grad_norm": 2.1658332347869873, "learning_rate": 4.9798369836547644e-05, "loss": 0.4406, "num_input_tokens_seen": 1398320, "step": 5205 }, { "epoch": 1.365661861074705, "grad_norm": 1.0454928874969482, "learning_rate": 4.97969175748484e-05, "loss": 0.4474, "num_input_tokens_seen": 1399632, "step": 5210 }, { "epoch": 1.3669724770642202, "grad_norm": 1.5086307525634766, "learning_rate": 4.9795460123197166e-05, "loss": 0.5241, "num_input_tokens_seen": 1400768, "step": 5215 }, { "epoch": 1.3682830930537353, "grad_norm": 2.041823148727417, "learning_rate": 4.979399748189896e-05, "loss": 0.4059, "num_input_tokens_seen": 1402352, "step": 5220 }, { "epoch": 1.3695937090432504, "grad_norm": 0.506029486656189, "learning_rate": 4.979252965125993e-05, "loss": 0.3485, "num_input_tokens_seen": 1403328, "step": 5225 }, { "epoch": 1.3709043250327655, "grad_norm": 1.0456629991531372, "learning_rate": 4.9791056631587276e-05, "loss": 0.577, "num_input_tokens_seen": 1404592, "step": 5230 }, { "epoch": 1.3722149410222806, "grad_norm": 0.6734380722045898, "learning_rate": 4.978957842318931e-05, "loss": 0.4829, "num_input_tokens_seen": 1405840, "step": 5235 }, { "epoch": 1.3735255570117955, "grad_norm": 1.3565478324890137, "learning_rate": 4.978809502637541e-05, "loss": 0.5562, "num_input_tokens_seen": 1407248, "step": 5240 }, { "epoch": 1.3748361730013106, "grad_norm": 1.0673725605010986, "learning_rate": 4.9786606441456054e-05, "loss": 0.4734, "num_input_tokens_seen": 1408528, "step": 5245 }, { "epoch": 1.3761467889908257, "grad_norm": 0.9056631326675415, "learning_rate": 4.9785112668742796e-05, "loss": 0.3551, "num_input_tokens_seen": 1409792, "step": 5250 }, { "epoch": 1.3774574049803407, "grad_norm": 1.7862955331802368, "learning_rate": 4.978361370854828e-05, "loss": 0.4334, "num_input_tokens_seen": 1411168, "step": 5255 }, { "epoch": 1.3787680209698558, "grad_norm": 0.7105101346969604, "learning_rate": 4.978210956118624e-05, "loss": 0.506, "num_input_tokens_seen": 1412512, "step": 5260 }, { "epoch": 1.380078636959371, "grad_norm": 0.7658453583717346, "learning_rate": 4.9780600226971486e-05, "loss": 0.564, "num_input_tokens_seen": 1413584, "step": 5265 }, { "epoch": 1.381389252948886, "grad_norm": 0.6066240072250366, "learning_rate": 4.977908570621993e-05, "loss": 0.6005, "num_input_tokens_seen": 1415520, "step": 5270 }, { "epoch": 1.382699868938401, "grad_norm": 0.9593902230262756, "learning_rate": 4.977756599924854e-05, "loss": 0.5842, "num_input_tokens_seen": 1416576, "step": 5275 }, { "epoch": 1.384010484927916, "grad_norm": 0.8831272125244141, "learning_rate": 4.97760411063754e-05, "loss": 0.3497, "num_input_tokens_seen": 1418032, "step": 5280 }, { "epoch": 1.385321100917431, "grad_norm": 0.7092305421829224, "learning_rate": 4.977451102791968e-05, "loss": 0.5371, "num_input_tokens_seen": 1419264, "step": 5285 }, { "epoch": 1.3866317169069462, "grad_norm": 2.44988751411438, "learning_rate": 4.977297576420159e-05, "loss": 0.5691, "num_input_tokens_seen": 1420224, "step": 5290 }, { "epoch": 1.3879423328964613, "grad_norm": 2.40803599357605, "learning_rate": 4.977143531554249e-05, "loss": 0.5393, "num_input_tokens_seen": 1421552, "step": 5295 }, { "epoch": 1.3892529488859764, "grad_norm": 0.5504088401794434, "learning_rate": 4.9769889682264774e-05, "loss": 0.5455, "num_input_tokens_seen": 1422832, "step": 5300 }, { "epoch": 1.3905635648754915, "grad_norm": 1.0337762832641602, "learning_rate": 4.976833886469196e-05, "loss": 0.4894, "num_input_tokens_seen": 1423952, "step": 5305 }, { "epoch": 1.3918741808650066, "grad_norm": 1.0061286687850952, "learning_rate": 4.97667828631486e-05, "loss": 0.36, "num_input_tokens_seen": 1425168, "step": 5310 }, { "epoch": 1.3931847968545217, "grad_norm": 1.2127015590667725, "learning_rate": 4.976522167796038e-05, "loss": 0.5481, "num_input_tokens_seen": 1426160, "step": 5315 }, { "epoch": 1.3944954128440368, "grad_norm": 0.79617840051651, "learning_rate": 4.9763655309454064e-05, "loss": 0.5663, "num_input_tokens_seen": 1427440, "step": 5320 }, { "epoch": 1.3958060288335519, "grad_norm": 0.5120932459831238, "learning_rate": 4.976208375795748e-05, "loss": 0.473, "num_input_tokens_seen": 1428496, "step": 5325 }, { "epoch": 1.397116644823067, "grad_norm": 2.243412971496582, "learning_rate": 4.9760507023799544e-05, "loss": 0.4347, "num_input_tokens_seen": 1429568, "step": 5330 }, { "epoch": 1.3984272608125818, "grad_norm": 1.0802245140075684, "learning_rate": 4.9758925107310276e-05, "loss": 0.5096, "num_input_tokens_seen": 1430800, "step": 5335 }, { "epoch": 1.399737876802097, "grad_norm": 1.9309486150741577, "learning_rate": 4.975733800882077e-05, "loss": 0.4978, "num_input_tokens_seen": 1431968, "step": 5340 }, { "epoch": 1.401048492791612, "grad_norm": 3.206530809402466, "learning_rate": 4.975574572866318e-05, "loss": 0.6304, "num_input_tokens_seen": 1433440, "step": 5345 }, { "epoch": 1.4023591087811271, "grad_norm": 0.8427588939666748, "learning_rate": 4.97541482671708e-05, "loss": 0.3991, "num_input_tokens_seen": 1434640, "step": 5350 }, { "epoch": 1.4036697247706422, "grad_norm": 0.6707924008369446, "learning_rate": 4.975254562467794e-05, "loss": 0.5023, "num_input_tokens_seen": 1436224, "step": 5355 }, { "epoch": 1.4049803407601573, "grad_norm": 0.4408380687236786, "learning_rate": 4.9750937801520064e-05, "loss": 0.4383, "num_input_tokens_seen": 1437136, "step": 5360 }, { "epoch": 1.4062909567496724, "grad_norm": 1.282902479171753, "learning_rate": 4.974932479803367e-05, "loss": 0.5683, "num_input_tokens_seen": 1438384, "step": 5365 }, { "epoch": 1.4076015727391873, "grad_norm": 1.1191471815109253, "learning_rate": 4.974770661455636e-05, "loss": 0.6337, "num_input_tokens_seen": 1439600, "step": 5370 }, { "epoch": 1.4089121887287024, "grad_norm": 0.7335794568061829, "learning_rate": 4.974608325142681e-05, "loss": 0.5366, "num_input_tokens_seen": 1440768, "step": 5375 }, { "epoch": 1.4102228047182175, "grad_norm": 0.7241779565811157, "learning_rate": 4.97444547089848e-05, "loss": 0.447, "num_input_tokens_seen": 1442320, "step": 5380 }, { "epoch": 1.4115334207077326, "grad_norm": 4.521408557891846, "learning_rate": 4.974282098757118e-05, "loss": 0.4804, "num_input_tokens_seen": 1443456, "step": 5385 }, { "epoch": 1.4128440366972477, "grad_norm": 0.8154172301292419, "learning_rate": 4.974118208752787e-05, "loss": 0.4294, "num_input_tokens_seen": 1444864, "step": 5390 }, { "epoch": 1.4141546526867628, "grad_norm": 0.775244414806366, "learning_rate": 4.97395380091979e-05, "loss": 0.4417, "num_input_tokens_seen": 1446176, "step": 5395 }, { "epoch": 1.4154652686762779, "grad_norm": 0.9156679511070251, "learning_rate": 4.9737888752925375e-05, "loss": 0.4792, "num_input_tokens_seen": 1447376, "step": 5400 }, { "epoch": 1.416775884665793, "grad_norm": 1.1594759225845337, "learning_rate": 4.973623431905548e-05, "loss": 0.6091, "num_input_tokens_seen": 1448368, "step": 5405 }, { "epoch": 1.418086500655308, "grad_norm": 0.9071454405784607, "learning_rate": 4.973457470793447e-05, "loss": 0.5432, "num_input_tokens_seen": 1449456, "step": 5410 }, { "epoch": 1.4193971166448232, "grad_norm": 0.6250770688056946, "learning_rate": 4.973290991990972e-05, "loss": 0.4781, "num_input_tokens_seen": 1451088, "step": 5415 }, { "epoch": 1.4207077326343382, "grad_norm": 1.4288370609283447, "learning_rate": 4.973123995532966e-05, "loss": 0.5407, "num_input_tokens_seen": 1452240, "step": 5420 }, { "epoch": 1.4220183486238533, "grad_norm": 1.0550423860549927, "learning_rate": 4.972956481454381e-05, "loss": 0.4772, "num_input_tokens_seen": 1453648, "step": 5425 }, { "epoch": 1.4233289646133682, "grad_norm": 1.0372899770736694, "learning_rate": 4.972788449790277e-05, "loss": 0.5638, "num_input_tokens_seen": 1455584, "step": 5430 }, { "epoch": 1.4246395806028833, "grad_norm": 0.7931000590324402, "learning_rate": 4.9726199005758234e-05, "loss": 0.3917, "num_input_tokens_seen": 1457408, "step": 5435 }, { "epoch": 1.4259501965923984, "grad_norm": 0.8892136216163635, "learning_rate": 4.972450833846297e-05, "loss": 0.4885, "num_input_tokens_seen": 1458704, "step": 5440 }, { "epoch": 1.4272608125819135, "grad_norm": 1.1158099174499512, "learning_rate": 4.972281249637083e-05, "loss": 0.3522, "num_input_tokens_seen": 1461232, "step": 5445 }, { "epoch": 1.4285714285714286, "grad_norm": 0.7592016458511353, "learning_rate": 4.9721111479836756e-05, "loss": 0.5044, "num_input_tokens_seen": 1462768, "step": 5450 }, { "epoch": 1.4298820445609437, "grad_norm": 0.6013795137405396, "learning_rate": 4.971940528921676e-05, "loss": 0.4616, "num_input_tokens_seen": 1464160, "step": 5455 }, { "epoch": 1.4311926605504588, "grad_norm": 1.2013453245162964, "learning_rate": 4.9717693924867944e-05, "loss": 0.4215, "num_input_tokens_seen": 1465264, "step": 5460 }, { "epoch": 1.4325032765399737, "grad_norm": 0.8987486958503723, "learning_rate": 4.9715977387148504e-05, "loss": 0.408, "num_input_tokens_seen": 1466592, "step": 5465 }, { "epoch": 1.4338138925294888, "grad_norm": 1.1830995082855225, "learning_rate": 4.971425567641771e-05, "loss": 0.5673, "num_input_tokens_seen": 1467744, "step": 5470 }, { "epoch": 1.4351245085190039, "grad_norm": 0.9222856760025024, "learning_rate": 4.97125287930359e-05, "loss": 0.4117, "num_input_tokens_seen": 1468768, "step": 5475 }, { "epoch": 1.436435124508519, "grad_norm": 1.9534437656402588, "learning_rate": 4.971079673736451e-05, "loss": 0.4894, "num_input_tokens_seen": 1470144, "step": 5480 }, { "epoch": 1.437745740498034, "grad_norm": 0.5576744079589844, "learning_rate": 4.970905950976606e-05, "loss": 0.6436, "num_input_tokens_seen": 1471488, "step": 5485 }, { "epoch": 1.4390563564875491, "grad_norm": 3.0003979206085205, "learning_rate": 4.970731711060415e-05, "loss": 0.6993, "num_input_tokens_seen": 1472784, "step": 5490 }, { "epoch": 1.4403669724770642, "grad_norm": 0.8651063442230225, "learning_rate": 4.970556954024346e-05, "loss": 0.4219, "num_input_tokens_seen": 1474112, "step": 5495 }, { "epoch": 1.4416775884665793, "grad_norm": 2.3586463928222656, "learning_rate": 4.970381679904975e-05, "loss": 0.3723, "num_input_tokens_seen": 1475120, "step": 5500 }, { "epoch": 1.4429882044560944, "grad_norm": 0.8201201558113098, "learning_rate": 4.970205888738988e-05, "loss": 0.6312, "num_input_tokens_seen": 1476368, "step": 5505 }, { "epoch": 1.4442988204456095, "grad_norm": 2.547727584838867, "learning_rate": 4.970029580563176e-05, "loss": 0.4603, "num_input_tokens_seen": 1477760, "step": 5510 }, { "epoch": 1.4456094364351246, "grad_norm": 1.162937879562378, "learning_rate": 4.96985275541444e-05, "loss": 0.3972, "num_input_tokens_seen": 1479136, "step": 5515 }, { "epoch": 1.4469200524246395, "grad_norm": 1.1882344484329224, "learning_rate": 4.969675413329791e-05, "loss": 0.3346, "num_input_tokens_seen": 1480048, "step": 5520 }, { "epoch": 1.4482306684141546, "grad_norm": 0.5743182897567749, "learning_rate": 4.969497554346344e-05, "loss": 0.4001, "num_input_tokens_seen": 1481360, "step": 5525 }, { "epoch": 1.4495412844036697, "grad_norm": 1.1041983366012573, "learning_rate": 4.969319178501327e-05, "loss": 0.6215, "num_input_tokens_seen": 1482864, "step": 5530 }, { "epoch": 1.4508519003931848, "grad_norm": 1.1661235094070435, "learning_rate": 4.969140285832072e-05, "loss": 0.3463, "num_input_tokens_seen": 1484336, "step": 5535 }, { "epoch": 1.4521625163826999, "grad_norm": 0.852237343788147, "learning_rate": 4.968960876376022e-05, "loss": 0.4374, "num_input_tokens_seen": 1485472, "step": 5540 }, { "epoch": 1.453473132372215, "grad_norm": 0.7179476022720337, "learning_rate": 4.9687809501707264e-05, "loss": 0.4204, "num_input_tokens_seen": 1487264, "step": 5545 }, { "epoch": 1.45478374836173, "grad_norm": 0.815875232219696, "learning_rate": 4.968600507253843e-05, "loss": 0.4793, "num_input_tokens_seen": 1488400, "step": 5550 }, { "epoch": 1.456094364351245, "grad_norm": 0.9726129174232483, "learning_rate": 4.968419547663139e-05, "loss": 0.3934, "num_input_tokens_seen": 1489424, "step": 5555 }, { "epoch": 1.45740498034076, "grad_norm": 1.5403904914855957, "learning_rate": 4.9682380714364897e-05, "loss": 0.6029, "num_input_tokens_seen": 1490320, "step": 5560 }, { "epoch": 1.4587155963302751, "grad_norm": 0.6364911794662476, "learning_rate": 4.968056078611876e-05, "loss": 0.5997, "num_input_tokens_seen": 1492640, "step": 5565 }, { "epoch": 1.4600262123197902, "grad_norm": 1.2287840843200684, "learning_rate": 4.96787356922739e-05, "loss": 0.5832, "num_input_tokens_seen": 1493872, "step": 5570 }, { "epoch": 1.4613368283093053, "grad_norm": 1.1933811902999878, "learning_rate": 4.9676905433212296e-05, "loss": 0.506, "num_input_tokens_seen": 1495216, "step": 5575 }, { "epoch": 1.4626474442988204, "grad_norm": 1.396740198135376, "learning_rate": 4.967507000931702e-05, "loss": 0.4736, "num_input_tokens_seen": 1496480, "step": 5580 }, { "epoch": 1.4639580602883355, "grad_norm": 0.6638656854629517, "learning_rate": 4.967322942097225e-05, "loss": 0.4606, "num_input_tokens_seen": 1497936, "step": 5585 }, { "epoch": 1.4652686762778506, "grad_norm": 1.5556180477142334, "learning_rate": 4.9671383668563167e-05, "loss": 0.4677, "num_input_tokens_seen": 1499280, "step": 5590 }, { "epoch": 1.4665792922673657, "grad_norm": 0.7043840885162354, "learning_rate": 4.966953275247612e-05, "loss": 0.5463, "num_input_tokens_seen": 1500464, "step": 5595 }, { "epoch": 1.4678899082568808, "grad_norm": 0.5942767858505249, "learning_rate": 4.966767667309849e-05, "loss": 0.3518, "num_input_tokens_seen": 1501824, "step": 5600 }, { "epoch": 1.469200524246396, "grad_norm": 0.6405045986175537, "learning_rate": 4.966581543081876e-05, "loss": 0.4883, "num_input_tokens_seen": 1503024, "step": 5605 }, { "epoch": 1.470511140235911, "grad_norm": 3.227539539337158, "learning_rate": 4.966394902602649e-05, "loss": 0.6247, "num_input_tokens_seen": 1504800, "step": 5610 }, { "epoch": 1.4718217562254259, "grad_norm": 2.670102119445801, "learning_rate": 4.966207745911229e-05, "loss": 0.3964, "num_input_tokens_seen": 1506016, "step": 5615 }, { "epoch": 1.473132372214941, "grad_norm": 1.097899317741394, "learning_rate": 4.9660200730467894e-05, "loss": 0.6105, "num_input_tokens_seen": 1507360, "step": 5620 }, { "epoch": 1.474442988204456, "grad_norm": 1.0324790477752686, "learning_rate": 4.9658318840486095e-05, "loss": 0.6996, "num_input_tokens_seen": 1508528, "step": 5625 }, { "epoch": 1.4757536041939712, "grad_norm": 2.1002299785614014, "learning_rate": 4.9656431789560765e-05, "loss": 0.5123, "num_input_tokens_seen": 1509584, "step": 5630 }, { "epoch": 1.4770642201834863, "grad_norm": 1.5997265577316284, "learning_rate": 4.9654539578086865e-05, "loss": 0.8493, "num_input_tokens_seen": 1511008, "step": 5635 }, { "epoch": 1.4783748361730014, "grad_norm": 0.8657549023628235, "learning_rate": 4.965264220646044e-05, "loss": 0.547, "num_input_tokens_seen": 1512448, "step": 5640 }, { "epoch": 1.4796854521625165, "grad_norm": 1.7330175638198853, "learning_rate": 4.965073967507859e-05, "loss": 0.5323, "num_input_tokens_seen": 1513712, "step": 5645 }, { "epoch": 1.4809960681520313, "grad_norm": 1.246508240699768, "learning_rate": 4.9648831984339525e-05, "loss": 0.6146, "num_input_tokens_seen": 1515632, "step": 5650 }, { "epoch": 1.4823066841415464, "grad_norm": 1.1842312812805176, "learning_rate": 4.9646919134642513e-05, "loss": 0.3592, "num_input_tokens_seen": 1516592, "step": 5655 }, { "epoch": 1.4836173001310615, "grad_norm": 2.6348187923431396, "learning_rate": 4.964500112638791e-05, "loss": 0.4562, "num_input_tokens_seen": 1518080, "step": 5660 }, { "epoch": 1.4849279161205766, "grad_norm": 1.2573635578155518, "learning_rate": 4.964307795997716e-05, "loss": 0.3608, "num_input_tokens_seen": 1519344, "step": 5665 }, { "epoch": 1.4862385321100917, "grad_norm": 2.731027841567993, "learning_rate": 4.964114963581277e-05, "loss": 0.486, "num_input_tokens_seen": 1520384, "step": 5670 }, { "epoch": 1.4875491480996068, "grad_norm": 1.9607807397842407, "learning_rate": 4.963921615429833e-05, "loss": 0.5224, "num_input_tokens_seen": 1521840, "step": 5675 }, { "epoch": 1.488859764089122, "grad_norm": 0.9827050566673279, "learning_rate": 4.963727751583853e-05, "loss": 0.5583, "num_input_tokens_seen": 1522960, "step": 5680 }, { "epoch": 1.490170380078637, "grad_norm": 0.6259966492652893, "learning_rate": 4.963533372083912e-05, "loss": 0.4622, "num_input_tokens_seen": 1524416, "step": 5685 }, { "epoch": 1.491480996068152, "grad_norm": 1.9256447553634644, "learning_rate": 4.963338476970692e-05, "loss": 0.5388, "num_input_tokens_seen": 1525808, "step": 5690 }, { "epoch": 1.4927916120576672, "grad_norm": 2.1677982807159424, "learning_rate": 4.963143066284986e-05, "loss": 0.3986, "num_input_tokens_seen": 1527088, "step": 5695 }, { "epoch": 1.4941022280471823, "grad_norm": 1.024311900138855, "learning_rate": 4.962947140067691e-05, "loss": 0.5411, "num_input_tokens_seen": 1528432, "step": 5700 }, { "epoch": 1.4954128440366974, "grad_norm": 0.7777459621429443, "learning_rate": 4.962750698359816e-05, "loss": 0.6251, "num_input_tokens_seen": 1529520, "step": 5705 }, { "epoch": 1.4967234600262123, "grad_norm": 0.8918207287788391, "learning_rate": 4.962553741202475e-05, "loss": 0.4923, "num_input_tokens_seen": 1530720, "step": 5710 }, { "epoch": 1.4980340760157274, "grad_norm": 1.448420524597168, "learning_rate": 4.962356268636891e-05, "loss": 0.5652, "num_input_tokens_seen": 1531952, "step": 5715 }, { "epoch": 1.4993446920052425, "grad_norm": 0.759056031703949, "learning_rate": 4.962158280704395e-05, "loss": 0.4672, "num_input_tokens_seen": 1533456, "step": 5720 }, { "epoch": 1.5003931847968546, "eval_loss": 0.5127978920936584, "eval_runtime": 16.6598, "eval_samples_per_second": 50.901, "eval_steps_per_second": 25.451, "num_input_tokens_seen": 1534400, "step": 5724 }, { "epoch": 1.5006553079947575, "grad_norm": 0.7319673895835876, "learning_rate": 4.961959777446425e-05, "loss": 0.5325, "num_input_tokens_seen": 1534720, "step": 5725 }, { "epoch": 1.5019659239842726, "grad_norm": 0.7141506671905518, "learning_rate": 4.961760758904527e-05, "loss": 0.3283, "num_input_tokens_seen": 1536992, "step": 5730 }, { "epoch": 1.5032765399737875, "grad_norm": 1.4572545289993286, "learning_rate": 4.961561225120356e-05, "loss": 0.5976, "num_input_tokens_seen": 1538288, "step": 5735 }, { "epoch": 1.5045871559633026, "grad_norm": 0.7808509469032288, "learning_rate": 4.9613611761356736e-05, "loss": 0.4768, "num_input_tokens_seen": 1539296, "step": 5740 }, { "epoch": 1.5058977719528177, "grad_norm": 0.585465133190155, "learning_rate": 4.96116061199235e-05, "loss": 0.4092, "num_input_tokens_seen": 1540896, "step": 5745 }, { "epoch": 1.5072083879423328, "grad_norm": 1.0976871252059937, "learning_rate": 4.960959532732363e-05, "loss": 0.4265, "num_input_tokens_seen": 1542000, "step": 5750 }, { "epoch": 1.508519003931848, "grad_norm": 0.8123722672462463, "learning_rate": 4.9607579383977984e-05, "loss": 0.4397, "num_input_tokens_seen": 1543536, "step": 5755 }, { "epoch": 1.509829619921363, "grad_norm": 1.592822790145874, "learning_rate": 4.9605558290308484e-05, "loss": 0.5714, "num_input_tokens_seen": 1545056, "step": 5760 }, { "epoch": 1.511140235910878, "grad_norm": 2.625445604324341, "learning_rate": 4.9603532046738154e-05, "loss": 0.4011, "num_input_tokens_seen": 1546128, "step": 5765 }, { "epoch": 1.5124508519003932, "grad_norm": 0.9269005656242371, "learning_rate": 4.960150065369108e-05, "loss": 0.3011, "num_input_tokens_seen": 1547536, "step": 5770 }, { "epoch": 1.5137614678899083, "grad_norm": 0.5243197083473206, "learning_rate": 4.9599464111592417e-05, "loss": 0.478, "num_input_tokens_seen": 1549152, "step": 5775 }, { "epoch": 1.5150720838794234, "grad_norm": 0.6148406267166138, "learning_rate": 4.959742242086843e-05, "loss": 0.4777, "num_input_tokens_seen": 1550464, "step": 5780 }, { "epoch": 1.5163826998689385, "grad_norm": 0.7037737965583801, "learning_rate": 4.959537558194643e-05, "loss": 0.4542, "num_input_tokens_seen": 1551808, "step": 5785 }, { "epoch": 1.5176933158584536, "grad_norm": 1.0745958089828491, "learning_rate": 4.959332359525482e-05, "loss": 0.4703, "num_input_tokens_seen": 1553056, "step": 5790 }, { "epoch": 1.5190039318479687, "grad_norm": 0.9227848052978516, "learning_rate": 4.9591266461223075e-05, "loss": 0.4574, "num_input_tokens_seen": 1554032, "step": 5795 }, { "epoch": 1.5203145478374838, "grad_norm": 0.8383831977844238, "learning_rate": 4.958920418028175e-05, "loss": 0.358, "num_input_tokens_seen": 1555360, "step": 5800 }, { "epoch": 1.5216251638269986, "grad_norm": 0.8909528255462646, "learning_rate": 4.958713675286247e-05, "loss": 0.5044, "num_input_tokens_seen": 1556384, "step": 5805 }, { "epoch": 1.5229357798165137, "grad_norm": 0.947526752948761, "learning_rate": 4.9585064179397955e-05, "loss": 0.471, "num_input_tokens_seen": 1557968, "step": 5810 }, { "epoch": 1.5242463958060288, "grad_norm": 1.4732797145843506, "learning_rate": 4.958298646032199e-05, "loss": 0.5812, "num_input_tokens_seen": 1559056, "step": 5815 }, { "epoch": 1.525557011795544, "grad_norm": 1.3241379261016846, "learning_rate": 4.958090359606943e-05, "loss": 0.4554, "num_input_tokens_seen": 1560336, "step": 5820 }, { "epoch": 1.526867627785059, "grad_norm": 1.2322802543640137, "learning_rate": 4.957881558707622e-05, "loss": 0.3966, "num_input_tokens_seen": 1561344, "step": 5825 }, { "epoch": 1.528178243774574, "grad_norm": 0.7963037490844727, "learning_rate": 4.957672243377938e-05, "loss": 0.5405, "num_input_tokens_seen": 1562960, "step": 5830 }, { "epoch": 1.529488859764089, "grad_norm": 1.0853794813156128, "learning_rate": 4.9574624136617e-05, "loss": 0.4587, "num_input_tokens_seen": 1564736, "step": 5835 }, { "epoch": 1.530799475753604, "grad_norm": 1.6213120222091675, "learning_rate": 4.9572520696028245e-05, "loss": 0.4091, "num_input_tokens_seen": 1566624, "step": 5840 }, { "epoch": 1.5321100917431192, "grad_norm": 0.6520669460296631, "learning_rate": 4.9570412112453377e-05, "loss": 0.4358, "num_input_tokens_seen": 1568336, "step": 5845 }, { "epoch": 1.5334207077326343, "grad_norm": 0.5852640271186829, "learning_rate": 4.95682983863337e-05, "loss": 0.4986, "num_input_tokens_seen": 1570208, "step": 5850 }, { "epoch": 1.5347313237221494, "grad_norm": 1.4085580110549927, "learning_rate": 4.9566179518111625e-05, "loss": 0.4717, "num_input_tokens_seen": 1571504, "step": 5855 }, { "epoch": 1.5360419397116645, "grad_norm": 2.09163498878479, "learning_rate": 4.9564055508230616e-05, "loss": 0.5031, "num_input_tokens_seen": 1572400, "step": 5860 }, { "epoch": 1.5373525557011796, "grad_norm": 6.488420486450195, "learning_rate": 4.9561926357135235e-05, "loss": 0.5164, "num_input_tokens_seen": 1573856, "step": 5865 }, { "epoch": 1.5386631716906947, "grad_norm": 0.8113301992416382, "learning_rate": 4.9559792065271116e-05, "loss": 0.3819, "num_input_tokens_seen": 1574976, "step": 5870 }, { "epoch": 1.5399737876802098, "grad_norm": 0.9420294761657715, "learning_rate": 4.955765263308495e-05, "loss": 0.6455, "num_input_tokens_seen": 1576016, "step": 5875 }, { "epoch": 1.5412844036697249, "grad_norm": 1.6135375499725342, "learning_rate": 4.9555508061024514e-05, "loss": 0.4014, "num_input_tokens_seen": 1577552, "step": 5880 }, { "epoch": 1.54259501965924, "grad_norm": 1.0153688192367554, "learning_rate": 4.955335834953867e-05, "loss": 0.3147, "num_input_tokens_seen": 1578944, "step": 5885 }, { "epoch": 1.543905635648755, "grad_norm": 7.119976043701172, "learning_rate": 4.955120349907736e-05, "loss": 0.3183, "num_input_tokens_seen": 1580336, "step": 5890 }, { "epoch": 1.5452162516382701, "grad_norm": 0.9584914445877075, "learning_rate": 4.954904351009157e-05, "loss": 0.8244, "num_input_tokens_seen": 1581456, "step": 5895 }, { "epoch": 1.546526867627785, "grad_norm": 1.0080095529556274, "learning_rate": 4.954687838303339e-05, "loss": 0.484, "num_input_tokens_seen": 1582976, "step": 5900 }, { "epoch": 1.5478374836173001, "grad_norm": 0.6299540400505066, "learning_rate": 4.9544708118355984e-05, "loss": 0.3864, "num_input_tokens_seen": 1584544, "step": 5905 }, { "epoch": 1.5491480996068152, "grad_norm": 0.9310789108276367, "learning_rate": 4.9542532716513576e-05, "loss": 0.6542, "num_input_tokens_seen": 1585936, "step": 5910 }, { "epoch": 1.5504587155963303, "grad_norm": 0.8613308668136597, "learning_rate": 4.9540352177961466e-05, "loss": 0.4892, "num_input_tokens_seen": 1587056, "step": 5915 }, { "epoch": 1.5517693315858454, "grad_norm": 1.0555955171585083, "learning_rate": 4.953816650315606e-05, "loss": 0.4552, "num_input_tokens_seen": 1588704, "step": 5920 }, { "epoch": 1.5530799475753603, "grad_norm": 1.136110544204712, "learning_rate": 4.9535975692554805e-05, "loss": 0.3534, "num_input_tokens_seen": 1590016, "step": 5925 }, { "epoch": 1.5543905635648754, "grad_norm": 2.647056818008423, "learning_rate": 4.953377974661623e-05, "loss": 0.5305, "num_input_tokens_seen": 1591200, "step": 5930 }, { "epoch": 1.5557011795543905, "grad_norm": 1.0518839359283447, "learning_rate": 4.953157866579994e-05, "loss": 0.5669, "num_input_tokens_seen": 1592576, "step": 5935 }, { "epoch": 1.5570117955439056, "grad_norm": 1.6293509006500244, "learning_rate": 4.9529372450566624e-05, "loss": 0.2821, "num_input_tokens_seen": 1593680, "step": 5940 }, { "epoch": 1.5583224115334207, "grad_norm": 0.5052666068077087, "learning_rate": 4.952716110137804e-05, "loss": 0.3905, "num_input_tokens_seen": 1594800, "step": 5945 }, { "epoch": 1.5596330275229358, "grad_norm": 0.9673407077789307, "learning_rate": 4.952494461869702e-05, "loss": 0.4517, "num_input_tokens_seen": 1596096, "step": 5950 }, { "epoch": 1.5609436435124509, "grad_norm": 0.8610644936561584, "learning_rate": 4.952272300298747e-05, "loss": 0.542, "num_input_tokens_seen": 1597344, "step": 5955 }, { "epoch": 1.562254259501966, "grad_norm": 1.1957364082336426, "learning_rate": 4.952049625471436e-05, "loss": 0.5057, "num_input_tokens_seen": 1598864, "step": 5960 }, { "epoch": 1.563564875491481, "grad_norm": 0.5641737580299377, "learning_rate": 4.951826437434375e-05, "loss": 0.4298, "num_input_tokens_seen": 1600144, "step": 5965 }, { "epoch": 1.5648754914809961, "grad_norm": 0.33124494552612305, "learning_rate": 4.951602736234278e-05, "loss": 0.4165, "num_input_tokens_seen": 1602080, "step": 5970 }, { "epoch": 1.5661861074705112, "grad_norm": 2.2246875762939453, "learning_rate": 4.9513785219179634e-05, "loss": 0.4644, "num_input_tokens_seen": 1603664, "step": 5975 }, { "epoch": 1.5674967234600263, "grad_norm": 1.4759539365768433, "learning_rate": 4.95115379453236e-05, "loss": 0.545, "num_input_tokens_seen": 1605616, "step": 5980 }, { "epoch": 1.5688073394495414, "grad_norm": 0.42818430066108704, "learning_rate": 4.950928554124503e-05, "loss": 0.4556, "num_input_tokens_seen": 1606720, "step": 5985 }, { "epoch": 1.5701179554390565, "grad_norm": 0.7715561985969543, "learning_rate": 4.950702800741534e-05, "loss": 0.8327, "num_input_tokens_seen": 1608240, "step": 5990 }, { "epoch": 1.5714285714285714, "grad_norm": 1.6144659519195557, "learning_rate": 4.950476534430703e-05, "loss": 0.6651, "num_input_tokens_seen": 1609504, "step": 5995 }, { "epoch": 1.5727391874180865, "grad_norm": 0.7934569716453552, "learning_rate": 4.950249755239369e-05, "loss": 0.4726, "num_input_tokens_seen": 1610608, "step": 6000 }, { "epoch": 1.5740498034076016, "grad_norm": 2.551246404647827, "learning_rate": 4.950022463214994e-05, "loss": 0.5899, "num_input_tokens_seen": 1611968, "step": 6005 }, { "epoch": 1.5753604193971167, "grad_norm": 0.5054857730865479, "learning_rate": 4.949794658405151e-05, "loss": 0.2993, "num_input_tokens_seen": 1613712, "step": 6010 }, { "epoch": 1.5766710353866316, "grad_norm": 0.578861653804779, "learning_rate": 4.9495663408575185e-05, "loss": 0.4274, "num_input_tokens_seen": 1615280, "step": 6015 }, { "epoch": 1.5779816513761467, "grad_norm": 0.47164011001586914, "learning_rate": 4.949337510619884e-05, "loss": 0.4214, "num_input_tokens_seen": 1616992, "step": 6020 }, { "epoch": 1.5792922673656618, "grad_norm": 0.7292202711105347, "learning_rate": 4.94910816774014e-05, "loss": 0.5024, "num_input_tokens_seen": 1618592, "step": 6025 }, { "epoch": 1.5806028833551768, "grad_norm": 1.0429317951202393, "learning_rate": 4.9488783122662886e-05, "loss": 0.4706, "num_input_tokens_seen": 1619824, "step": 6030 }, { "epoch": 1.581913499344692, "grad_norm": 0.8012551665306091, "learning_rate": 4.9486479442464374e-05, "loss": 0.5033, "num_input_tokens_seen": 1620800, "step": 6035 }, { "epoch": 1.583224115334207, "grad_norm": 1.6855294704437256, "learning_rate": 4.948417063728803e-05, "loss": 0.5061, "num_input_tokens_seen": 1622000, "step": 6040 }, { "epoch": 1.5845347313237221, "grad_norm": 0.9201597571372986, "learning_rate": 4.948185670761707e-05, "loss": 0.3619, "num_input_tokens_seen": 1623328, "step": 6045 }, { "epoch": 1.5858453473132372, "grad_norm": 2.1506130695343018, "learning_rate": 4.94795376539358e-05, "loss": 0.4016, "num_input_tokens_seen": 1624432, "step": 6050 }, { "epoch": 1.5871559633027523, "grad_norm": 2.056123733520508, "learning_rate": 4.947721347672961e-05, "loss": 0.459, "num_input_tokens_seen": 1626304, "step": 6055 }, { "epoch": 1.5884665792922674, "grad_norm": 0.9874494671821594, "learning_rate": 4.9474884176484924e-05, "loss": 0.6411, "num_input_tokens_seen": 1628480, "step": 6060 }, { "epoch": 1.5897771952817825, "grad_norm": 1.8352646827697754, "learning_rate": 4.947254975368927e-05, "loss": 0.6251, "num_input_tokens_seen": 1630048, "step": 6065 }, { "epoch": 1.5910878112712976, "grad_norm": 0.38193973898887634, "learning_rate": 4.947021020883124e-05, "loss": 0.4789, "num_input_tokens_seen": 1631600, "step": 6070 }, { "epoch": 1.5923984272608127, "grad_norm": 0.754138708114624, "learning_rate": 4.9467865542400494e-05, "loss": 0.442, "num_input_tokens_seen": 1633232, "step": 6075 }, { "epoch": 1.5937090432503278, "grad_norm": 0.6343916654586792, "learning_rate": 4.946551575488776e-05, "loss": 0.4548, "num_input_tokens_seen": 1634592, "step": 6080 }, { "epoch": 1.5950196592398427, "grad_norm": 0.7448921203613281, "learning_rate": 4.946316084678486e-05, "loss": 0.4246, "num_input_tokens_seen": 1635760, "step": 6085 }, { "epoch": 1.5963302752293578, "grad_norm": 0.9134271740913391, "learning_rate": 4.946080081858466e-05, "loss": 0.6168, "num_input_tokens_seen": 1637040, "step": 6090 }, { "epoch": 1.5976408912188729, "grad_norm": 0.9658303260803223, "learning_rate": 4.945843567078111e-05, "loss": 0.4377, "num_input_tokens_seen": 1638464, "step": 6095 }, { "epoch": 1.598951507208388, "grad_norm": 0.635071337223053, "learning_rate": 4.9456065403869246e-05, "loss": 0.615, "num_input_tokens_seen": 1639792, "step": 6100 }, { "epoch": 1.600262123197903, "grad_norm": 0.6069138646125793, "learning_rate": 4.9453690018345144e-05, "loss": 0.3219, "num_input_tokens_seen": 1641056, "step": 6105 }, { "epoch": 1.601572739187418, "grad_norm": 1.2640292644500732, "learning_rate": 4.945130951470597e-05, "loss": 0.7561, "num_input_tokens_seen": 1642208, "step": 6110 }, { "epoch": 1.602883355176933, "grad_norm": 4.838681221008301, "learning_rate": 4.944892389344997e-05, "loss": 0.3168, "num_input_tokens_seen": 1643664, "step": 6115 }, { "epoch": 1.6041939711664481, "grad_norm": 0.7814904451370239, "learning_rate": 4.9446533155076444e-05, "loss": 0.4053, "num_input_tokens_seen": 1644976, "step": 6120 }, { "epoch": 1.6055045871559632, "grad_norm": 1.5928938388824463, "learning_rate": 4.9444137300085766e-05, "loss": 0.7602, "num_input_tokens_seen": 1646352, "step": 6125 }, { "epoch": 1.6068152031454783, "grad_norm": 1.1783299446105957, "learning_rate": 4.944173632897938e-05, "loss": 0.438, "num_input_tokens_seen": 1647296, "step": 6130 }, { "epoch": 1.6081258191349934, "grad_norm": 0.7145092487335205, "learning_rate": 4.943933024225983e-05, "loss": 0.6084, "num_input_tokens_seen": 1649152, "step": 6135 }, { "epoch": 1.6094364351245085, "grad_norm": 0.7572266459465027, "learning_rate": 4.943691904043067e-05, "loss": 0.5478, "num_input_tokens_seen": 1650208, "step": 6140 }, { "epoch": 1.6107470511140236, "grad_norm": 0.6965382695198059, "learning_rate": 4.943450272399659e-05, "loss": 0.5426, "num_input_tokens_seen": 1650960, "step": 6145 }, { "epoch": 1.6120576671035387, "grad_norm": 2.955700159072876, "learning_rate": 4.94320812934633e-05, "loss": 0.407, "num_input_tokens_seen": 1652240, "step": 6150 }, { "epoch": 1.6133682830930538, "grad_norm": 0.9901660084724426, "learning_rate": 4.942965474933762e-05, "loss": 0.5964, "num_input_tokens_seen": 1653568, "step": 6155 }, { "epoch": 1.614678899082569, "grad_norm": 0.9327901601791382, "learning_rate": 4.9427223092127406e-05, "loss": 0.6486, "num_input_tokens_seen": 1654944, "step": 6160 }, { "epoch": 1.615989515072084, "grad_norm": 0.48634281754493713, "learning_rate": 4.942478632234161e-05, "loss": 0.6433, "num_input_tokens_seen": 1656352, "step": 6165 }, { "epoch": 1.617300131061599, "grad_norm": 1.0848687887191772, "learning_rate": 4.942234444049024e-05, "loss": 0.4469, "num_input_tokens_seen": 1657792, "step": 6170 }, { "epoch": 1.6186107470511142, "grad_norm": 1.0310423374176025, "learning_rate": 4.9419897447084375e-05, "loss": 0.463, "num_input_tokens_seen": 1659072, "step": 6175 }, { "epoch": 1.619921363040629, "grad_norm": 0.47011858224868774, "learning_rate": 4.9417445342636166e-05, "loss": 0.2199, "num_input_tokens_seen": 1660688, "step": 6180 }, { "epoch": 1.6212319790301442, "grad_norm": 1.5750653743743896, "learning_rate": 4.941498812765884e-05, "loss": 0.4368, "num_input_tokens_seen": 1662352, "step": 6185 }, { "epoch": 1.6225425950196593, "grad_norm": 5.238009452819824, "learning_rate": 4.941252580266668e-05, "loss": 0.6413, "num_input_tokens_seen": 1663168, "step": 6190 }, { "epoch": 1.6238532110091743, "grad_norm": 1.0451902151107788, "learning_rate": 4.941005836817506e-05, "loss": 0.537, "num_input_tokens_seen": 1664896, "step": 6195 }, { "epoch": 1.6251638269986892, "grad_norm": 1.254346489906311, "learning_rate": 4.940758582470039e-05, "loss": 0.4581, "num_input_tokens_seen": 1666048, "step": 6200 }, { "epoch": 1.6264744429882043, "grad_norm": 0.8843708038330078, "learning_rate": 4.940510817276019e-05, "loss": 0.6308, "num_input_tokens_seen": 1667184, "step": 6205 }, { "epoch": 1.6277850589777194, "grad_norm": 0.552084743976593, "learning_rate": 4.940262541287302e-05, "loss": 0.2956, "num_input_tokens_seen": 1668496, "step": 6210 }, { "epoch": 1.6290956749672345, "grad_norm": 1.0052968263626099, "learning_rate": 4.94001375455585e-05, "loss": 0.8038, "num_input_tokens_seen": 1669776, "step": 6215 }, { "epoch": 1.6304062909567496, "grad_norm": 0.40440645813941956, "learning_rate": 4.939764457133737e-05, "loss": 0.5252, "num_input_tokens_seen": 1671296, "step": 6220 }, { "epoch": 1.6317169069462647, "grad_norm": 1.4839963912963867, "learning_rate": 4.939514649073137e-05, "loss": 0.44, "num_input_tokens_seen": 1672672, "step": 6225 }, { "epoch": 1.6330275229357798, "grad_norm": 1.1047978401184082, "learning_rate": 4.939264330426337e-05, "loss": 0.4103, "num_input_tokens_seen": 1673904, "step": 6230 }, { "epoch": 1.634338138925295, "grad_norm": 1.312233805656433, "learning_rate": 4.9390135012457275e-05, "loss": 0.4145, "num_input_tokens_seen": 1675296, "step": 6235 }, { "epoch": 1.63564875491481, "grad_norm": 0.8910013437271118, "learning_rate": 4.938762161583807e-05, "loss": 0.7432, "num_input_tokens_seen": 1676400, "step": 6240 }, { "epoch": 1.636959370904325, "grad_norm": 0.6921913623809814, "learning_rate": 4.9385103114931784e-05, "loss": 0.4417, "num_input_tokens_seen": 1678064, "step": 6245 }, { "epoch": 1.6382699868938402, "grad_norm": 1.019027590751648, "learning_rate": 4.9382579510265556e-05, "loss": 0.5192, "num_input_tokens_seen": 1679824, "step": 6250 }, { "epoch": 1.6395806028833553, "grad_norm": 0.1895265132188797, "learning_rate": 4.9380050802367573e-05, "loss": 0.4091, "num_input_tokens_seen": 1680960, "step": 6255 }, { "epoch": 1.6408912188728704, "grad_norm": 0.8579305410385132, "learning_rate": 4.937751699176708e-05, "loss": 0.5073, "num_input_tokens_seen": 1682144, "step": 6260 }, { "epoch": 1.6422018348623855, "grad_norm": 0.9244486689567566, "learning_rate": 4.93749780789944e-05, "loss": 0.5087, "num_input_tokens_seen": 1683744, "step": 6265 }, { "epoch": 1.6435124508519003, "grad_norm": 1.008666753768921, "learning_rate": 4.937243406458093e-05, "loss": 0.3179, "num_input_tokens_seen": 1685136, "step": 6270 }, { "epoch": 1.6448230668414154, "grad_norm": 4.169241905212402, "learning_rate": 4.9369884949059115e-05, "loss": 0.6063, "num_input_tokens_seen": 1686400, "step": 6275 }, { "epoch": 1.6461336828309305, "grad_norm": 0.8493562340736389, "learning_rate": 4.936733073296249e-05, "loss": 0.4029, "num_input_tokens_seen": 1687424, "step": 6280 }, { "epoch": 1.6474442988204456, "grad_norm": 1.5128765106201172, "learning_rate": 4.936477141682565e-05, "loss": 0.455, "num_input_tokens_seen": 1688976, "step": 6285 }, { "epoch": 1.6487549148099607, "grad_norm": 0.41897645592689514, "learning_rate": 4.9362207001184246e-05, "loss": 0.327, "num_input_tokens_seen": 1690608, "step": 6290 }, { "epoch": 1.6500655307994756, "grad_norm": 1.7220394611358643, "learning_rate": 4.9359637486575016e-05, "loss": 0.5294, "num_input_tokens_seen": 1691760, "step": 6295 }, { "epoch": 1.6513761467889907, "grad_norm": 1.6031759977340698, "learning_rate": 4.9357062873535746e-05, "loss": 0.5092, "num_input_tokens_seen": 1693280, "step": 6300 }, { "epoch": 1.6526867627785058, "grad_norm": 0.6525447368621826, "learning_rate": 4.9354483162605305e-05, "loss": 0.437, "num_input_tokens_seen": 1694272, "step": 6305 }, { "epoch": 1.6539973787680209, "grad_norm": 0.7283455729484558, "learning_rate": 4.9351898354323625e-05, "loss": 0.4836, "num_input_tokens_seen": 1695696, "step": 6310 }, { "epoch": 1.655307994757536, "grad_norm": 1.326164722442627, "learning_rate": 4.934930844923168e-05, "loss": 0.4579, "num_input_tokens_seen": 1696704, "step": 6315 }, { "epoch": 1.656618610747051, "grad_norm": 1.1957917213439941, "learning_rate": 4.934671344787156e-05, "loss": 0.4546, "num_input_tokens_seen": 1697936, "step": 6320 }, { "epoch": 1.6579292267365662, "grad_norm": 0.734409511089325, "learning_rate": 4.9344113350786386e-05, "loss": 0.4213, "num_input_tokens_seen": 1699456, "step": 6325 }, { "epoch": 1.6592398427260813, "grad_norm": 1.6726239919662476, "learning_rate": 4.934150815852034e-05, "loss": 0.428, "num_input_tokens_seen": 1700512, "step": 6330 }, { "epoch": 1.6605504587155964, "grad_norm": 1.3881809711456299, "learning_rate": 4.933889787161871e-05, "loss": 0.4616, "num_input_tokens_seen": 1701760, "step": 6335 }, { "epoch": 1.6618610747051115, "grad_norm": 0.8431746959686279, "learning_rate": 4.9336282490627803e-05, "loss": 0.6746, "num_input_tokens_seen": 1702880, "step": 6340 }, { "epoch": 1.6631716906946266, "grad_norm": 1.0772448778152466, "learning_rate": 4.9333662016095016e-05, "loss": 0.6467, "num_input_tokens_seen": 1704064, "step": 6345 }, { "epoch": 1.6644823066841417, "grad_norm": 0.6958007216453552, "learning_rate": 4.9331036448568826e-05, "loss": 0.4696, "num_input_tokens_seen": 1705264, "step": 6350 }, { "epoch": 1.6657929226736568, "grad_norm": 0.6979279518127441, "learning_rate": 4.932840578859873e-05, "loss": 0.3906, "num_input_tokens_seen": 1706736, "step": 6355 }, { "epoch": 1.6671035386631718, "grad_norm": 1.2867227792739868, "learning_rate": 4.932577003673535e-05, "loss": 0.4826, "num_input_tokens_seen": 1708000, "step": 6360 }, { "epoch": 1.6684141546526867, "grad_norm": 1.018479824066162, "learning_rate": 4.932312919353034e-05, "loss": 0.4898, "num_input_tokens_seen": 1709536, "step": 6365 }, { "epoch": 1.6697247706422018, "grad_norm": 0.6861900687217712, "learning_rate": 4.932048325953641e-05, "loss": 0.416, "num_input_tokens_seen": 1710816, "step": 6370 }, { "epoch": 1.671035386631717, "grad_norm": 1.1885331869125366, "learning_rate": 4.931783223530736e-05, "loss": 0.4295, "num_input_tokens_seen": 1712288, "step": 6375 }, { "epoch": 1.672346002621232, "grad_norm": 1.1245275735855103, "learning_rate": 4.9315176121398035e-05, "loss": 0.4488, "num_input_tokens_seen": 1714192, "step": 6380 }, { "epoch": 1.673656618610747, "grad_norm": 0.610616147518158, "learning_rate": 4.931251491836436e-05, "loss": 0.3399, "num_input_tokens_seen": 1715280, "step": 6385 }, { "epoch": 1.674967234600262, "grad_norm": 0.8407655358314514, "learning_rate": 4.930984862676332e-05, "loss": 0.3599, "num_input_tokens_seen": 1716864, "step": 6390 }, { "epoch": 1.676277850589777, "grad_norm": 4.456713676452637, "learning_rate": 4.930717724715298e-05, "loss": 0.4963, "num_input_tokens_seen": 1717984, "step": 6395 }, { "epoch": 1.6775884665792922, "grad_norm": 1.228685736656189, "learning_rate": 4.930450078009243e-05, "loss": 0.7987, "num_input_tokens_seen": 1719120, "step": 6400 }, { "epoch": 1.6788990825688073, "grad_norm": 1.7197134494781494, "learning_rate": 4.930181922614186e-05, "loss": 0.7762, "num_input_tokens_seen": 1721136, "step": 6405 }, { "epoch": 1.6802096985583224, "grad_norm": 0.7076876759529114, "learning_rate": 4.929913258586252e-05, "loss": 0.5135, "num_input_tokens_seen": 1722528, "step": 6410 }, { "epoch": 1.6815203145478375, "grad_norm": 0.6559166312217712, "learning_rate": 4.9296440859816715e-05, "loss": 0.4964, "num_input_tokens_seen": 1723856, "step": 6415 }, { "epoch": 1.6828309305373526, "grad_norm": 1.0205610990524292, "learning_rate": 4.9293744048567825e-05, "loss": 0.4738, "num_input_tokens_seen": 1724976, "step": 6420 }, { "epoch": 1.6841415465268676, "grad_norm": 1.621030330657959, "learning_rate": 4.929104215268027e-05, "loss": 0.7173, "num_input_tokens_seen": 1726656, "step": 6425 }, { "epoch": 1.6854521625163827, "grad_norm": 0.8691216707229614, "learning_rate": 4.9288335172719575e-05, "loss": 0.6423, "num_input_tokens_seen": 1727920, "step": 6430 }, { "epoch": 1.6867627785058978, "grad_norm": 0.9194053411483765, "learning_rate": 4.9285623109252286e-05, "loss": 0.5724, "num_input_tokens_seen": 1729200, "step": 6435 }, { "epoch": 1.688073394495413, "grad_norm": 0.8431970477104187, "learning_rate": 4.928290596284606e-05, "loss": 0.297, "num_input_tokens_seen": 1730352, "step": 6440 }, { "epoch": 1.689384010484928, "grad_norm": 0.5960290431976318, "learning_rate": 4.928018373406955e-05, "loss": 0.3877, "num_input_tokens_seen": 1731616, "step": 6445 }, { "epoch": 1.6906946264744431, "grad_norm": 3.1750872135162354, "learning_rate": 4.927745642349255e-05, "loss": 0.6975, "num_input_tokens_seen": 1732960, "step": 6450 }, { "epoch": 1.6920052424639582, "grad_norm": 1.2113443613052368, "learning_rate": 4.927472403168587e-05, "loss": 0.6525, "num_input_tokens_seen": 1734160, "step": 6455 }, { "epoch": 1.693315858453473, "grad_norm": 1.0301929712295532, "learning_rate": 4.92719865592214e-05, "loss": 0.5038, "num_input_tokens_seen": 1735392, "step": 6460 }, { "epoch": 1.6946264744429882, "grad_norm": 1.0386505126953125, "learning_rate": 4.9269244006672065e-05, "loss": 0.533, "num_input_tokens_seen": 1736576, "step": 6465 }, { "epoch": 1.6959370904325033, "grad_norm": 7.552898406982422, "learning_rate": 4.9266496374611904e-05, "loss": 0.5008, "num_input_tokens_seen": 1737792, "step": 6470 }, { "epoch": 1.6972477064220184, "grad_norm": 1.0811614990234375, "learning_rate": 4.926374366361598e-05, "loss": 0.6107, "num_input_tokens_seen": 1739184, "step": 6475 }, { "epoch": 1.6985583224115333, "grad_norm": 0.9648706316947937, "learning_rate": 4.926098587426043e-05, "loss": 0.4872, "num_input_tokens_seen": 1740624, "step": 6480 }, { "epoch": 1.6998689384010484, "grad_norm": 0.7501538991928101, "learning_rate": 4.9258223007122454e-05, "loss": 0.3768, "num_input_tokens_seen": 1742480, "step": 6485 }, { "epoch": 1.7011795543905635, "grad_norm": 0.4591003954410553, "learning_rate": 4.925545506278032e-05, "loss": 0.4322, "num_input_tokens_seen": 1743808, "step": 6490 }, { "epoch": 1.7024901703800785, "grad_norm": 0.8012029528617859, "learning_rate": 4.9252682041813356e-05, "loss": 0.7028, "num_input_tokens_seen": 1744704, "step": 6495 }, { "epoch": 1.7038007863695936, "grad_norm": 0.6217153072357178, "learning_rate": 4.924990394480195e-05, "loss": 0.3545, "num_input_tokens_seen": 1745872, "step": 6500 }, { "epoch": 1.7051114023591087, "grad_norm": 0.6636559367179871, "learning_rate": 4.9247120772327534e-05, "loss": 0.4908, "num_input_tokens_seen": 1747184, "step": 6505 }, { "epoch": 1.7064220183486238, "grad_norm": 1.1624473333358765, "learning_rate": 4.924433252497264e-05, "loss": 0.639, "num_input_tokens_seen": 1748880, "step": 6510 }, { "epoch": 1.707732634338139, "grad_norm": 1.9010640382766724, "learning_rate": 4.924153920332084e-05, "loss": 0.6765, "num_input_tokens_seen": 1750400, "step": 6515 }, { "epoch": 1.709043250327654, "grad_norm": 0.7727935314178467, "learning_rate": 4.923874080795677e-05, "loss": 0.3932, "num_input_tokens_seen": 1752000, "step": 6520 }, { "epoch": 1.7103538663171691, "grad_norm": 0.8389325737953186, "learning_rate": 4.923593733946614e-05, "loss": 0.6646, "num_input_tokens_seen": 1753520, "step": 6525 }, { "epoch": 1.7116644823066842, "grad_norm": 1.351881980895996, "learning_rate": 4.92331287984357e-05, "loss": 0.5579, "num_input_tokens_seen": 1754704, "step": 6530 }, { "epoch": 1.7129750982961993, "grad_norm": 1.2248071432113647, "learning_rate": 4.9230315185453266e-05, "loss": 0.4575, "num_input_tokens_seen": 1755488, "step": 6535 }, { "epoch": 1.7142857142857144, "grad_norm": 0.7606407403945923, "learning_rate": 4.922749650110773e-05, "loss": 0.3642, "num_input_tokens_seen": 1756688, "step": 6540 }, { "epoch": 1.7155963302752295, "grad_norm": 0.5733204483985901, "learning_rate": 4.9224672745989045e-05, "loss": 0.5313, "num_input_tokens_seen": 1758256, "step": 6545 }, { "epoch": 1.7169069462647444, "grad_norm": 2.7744271755218506, "learning_rate": 4.922184392068821e-05, "loss": 0.6103, "num_input_tokens_seen": 1759392, "step": 6550 }, { "epoch": 1.7182175622542595, "grad_norm": 1.4623180627822876, "learning_rate": 4.921901002579729e-05, "loss": 0.4935, "num_input_tokens_seen": 1760800, "step": 6555 }, { "epoch": 1.7195281782437746, "grad_norm": 0.9519086480140686, "learning_rate": 4.921617106190942e-05, "loss": 0.4041, "num_input_tokens_seen": 1761872, "step": 6560 }, { "epoch": 1.7208387942332897, "grad_norm": 2.8568873405456543, "learning_rate": 4.9213327029618796e-05, "loss": 0.507, "num_input_tokens_seen": 1763392, "step": 6565 }, { "epoch": 1.7221494102228048, "grad_norm": 0.682806134223938, "learning_rate": 4.921047792952066e-05, "loss": 0.4878, "num_input_tokens_seen": 1764832, "step": 6570 }, { "epoch": 1.7234600262123196, "grad_norm": 0.9472496509552002, "learning_rate": 4.920762376221132e-05, "loss": 0.3727, "num_input_tokens_seen": 1766304, "step": 6575 }, { "epoch": 1.7247706422018347, "grad_norm": 0.49992188811302185, "learning_rate": 4.920476452828815e-05, "loss": 0.3027, "num_input_tokens_seen": 1767632, "step": 6580 }, { "epoch": 1.7260812581913498, "grad_norm": 1.5605952739715576, "learning_rate": 4.9201900228349594e-05, "loss": 0.3547, "num_input_tokens_seen": 1768944, "step": 6585 }, { "epoch": 1.727391874180865, "grad_norm": 1.0718650817871094, "learning_rate": 4.9199030862995144e-05, "loss": 0.5509, "num_input_tokens_seen": 1770672, "step": 6590 }, { "epoch": 1.72870249017038, "grad_norm": 1.126039743423462, "learning_rate": 4.919615643282533e-05, "loss": 0.5559, "num_input_tokens_seen": 1771808, "step": 6595 }, { "epoch": 1.7300131061598951, "grad_norm": 0.9381118416786194, "learning_rate": 4.91932769384418e-05, "loss": 0.5746, "num_input_tokens_seen": 1773056, "step": 6600 }, { "epoch": 1.7313237221494102, "grad_norm": 0.861495316028595, "learning_rate": 4.91903923804472e-05, "loss": 0.6368, "num_input_tokens_seen": 1774528, "step": 6605 }, { "epoch": 1.7326343381389253, "grad_norm": 1.1521104574203491, "learning_rate": 4.918750275944528e-05, "loss": 0.7543, "num_input_tokens_seen": 1775680, "step": 6610 }, { "epoch": 1.7339449541284404, "grad_norm": 0.5581307411193848, "learning_rate": 4.9184608076040816e-05, "loss": 0.4046, "num_input_tokens_seen": 1776944, "step": 6615 }, { "epoch": 1.7352555701179555, "grad_norm": 0.5516573190689087, "learning_rate": 4.918170833083967e-05, "loss": 0.438, "num_input_tokens_seen": 1778432, "step": 6620 }, { "epoch": 1.7365661861074706, "grad_norm": 0.4821214973926544, "learning_rate": 4.9178803524448756e-05, "loss": 0.5342, "num_input_tokens_seen": 1779520, "step": 6625 }, { "epoch": 1.7378768020969857, "grad_norm": 0.96257084608078, "learning_rate": 4.9175893657476046e-05, "loss": 0.7148, "num_input_tokens_seen": 1780864, "step": 6630 }, { "epoch": 1.7391874180865008, "grad_norm": 1.058562994003296, "learning_rate": 4.917297873053056e-05, "loss": 0.4534, "num_input_tokens_seen": 1782608, "step": 6635 }, { "epoch": 1.7404980340760159, "grad_norm": 1.2307953834533691, "learning_rate": 4.91700587442224e-05, "loss": 0.4737, "num_input_tokens_seen": 1783680, "step": 6640 }, { "epoch": 1.7418086500655308, "grad_norm": 4.781888484954834, "learning_rate": 4.91671336991627e-05, "loss": 0.5315, "num_input_tokens_seen": 1784720, "step": 6645 }, { "epoch": 1.7431192660550459, "grad_norm": 2.5541961193084717, "learning_rate": 4.916420359596368e-05, "loss": 0.6538, "num_input_tokens_seen": 1785792, "step": 6650 }, { "epoch": 1.744429882044561, "grad_norm": 0.29191887378692627, "learning_rate": 4.91612684352386e-05, "loss": 0.3535, "num_input_tokens_seen": 1787008, "step": 6655 }, { "epoch": 1.745740498034076, "grad_norm": 0.8759594559669495, "learning_rate": 4.915832821760178e-05, "loss": 0.3954, "num_input_tokens_seen": 1788384, "step": 6660 }, { "epoch": 1.747051114023591, "grad_norm": 0.6974801421165466, "learning_rate": 4.915538294366861e-05, "loss": 0.7385, "num_input_tokens_seen": 1789888, "step": 6665 }, { "epoch": 1.748361730013106, "grad_norm": 0.9153414368629456, "learning_rate": 4.915243261405553e-05, "loss": 0.4532, "num_input_tokens_seen": 1791088, "step": 6670 }, { "epoch": 1.7496723460026211, "grad_norm": 2.4410016536712646, "learning_rate": 4.9149477229380023e-05, "loss": 0.4921, "num_input_tokens_seen": 1792384, "step": 6675 }, { "epoch": 1.7509829619921362, "grad_norm": 1.0986213684082031, "learning_rate": 4.914651679026068e-05, "loss": 0.551, "num_input_tokens_seen": 1793776, "step": 6680 }, { "epoch": 1.7522935779816513, "grad_norm": 0.6800899505615234, "learning_rate": 4.9143551297317076e-05, "loss": 0.3604, "num_input_tokens_seen": 1794880, "step": 6685 }, { "epoch": 1.7536041939711664, "grad_norm": 0.6268681287765503, "learning_rate": 4.9140580751169906e-05, "loss": 0.3901, "num_input_tokens_seen": 1796016, "step": 6690 }, { "epoch": 1.7549148099606815, "grad_norm": 0.9493885636329651, "learning_rate": 4.91376051524409e-05, "loss": 0.5346, "num_input_tokens_seen": 1797264, "step": 6695 }, { "epoch": 1.7562254259501966, "grad_norm": 0.4828123450279236, "learning_rate": 4.913462450175285e-05, "loss": 0.3536, "num_input_tokens_seen": 1798368, "step": 6700 }, { "epoch": 1.7575360419397117, "grad_norm": 0.5010225772857666, "learning_rate": 4.913163879972957e-05, "loss": 0.3929, "num_input_tokens_seen": 1799744, "step": 6705 }, { "epoch": 1.7588466579292268, "grad_norm": 1.5448187589645386, "learning_rate": 4.9128648046996004e-05, "loss": 0.5041, "num_input_tokens_seen": 1800992, "step": 6710 }, { "epoch": 1.7601572739187419, "grad_norm": 0.9373982548713684, "learning_rate": 4.912565224417809e-05, "loss": 0.5375, "num_input_tokens_seen": 1802640, "step": 6715 }, { "epoch": 1.761467889908257, "grad_norm": 0.6767665147781372, "learning_rate": 4.9122651391902855e-05, "loss": 0.3754, "num_input_tokens_seen": 1803920, "step": 6720 }, { "epoch": 1.762778505897772, "grad_norm": 1.1261706352233887, "learning_rate": 4.9119645490798347e-05, "loss": 0.5246, "num_input_tokens_seen": 1805088, "step": 6725 }, { "epoch": 1.7640891218872872, "grad_norm": 0.5417119860649109, "learning_rate": 4.911663454149372e-05, "loss": 0.6203, "num_input_tokens_seen": 1806512, "step": 6730 }, { "epoch": 1.765399737876802, "grad_norm": 0.8019611835479736, "learning_rate": 4.911361854461916e-05, "loss": 0.6543, "num_input_tokens_seen": 1808064, "step": 6735 }, { "epoch": 1.7667103538663171, "grad_norm": 0.5308572053909302, "learning_rate": 4.91105975008059e-05, "loss": 0.2587, "num_input_tokens_seen": 1809216, "step": 6740 }, { "epoch": 1.7680209698558322, "grad_norm": 1.723603367805481, "learning_rate": 4.910757141068624e-05, "loss": 0.4724, "num_input_tokens_seen": 1810368, "step": 6745 }, { "epoch": 1.7693315858453473, "grad_norm": 0.6646420955657959, "learning_rate": 4.910454027489354e-05, "loss": 0.3655, "num_input_tokens_seen": 1812000, "step": 6750 }, { "epoch": 1.7706422018348624, "grad_norm": 1.0058733224868774, "learning_rate": 4.910150409406221e-05, "loss": 0.5391, "num_input_tokens_seen": 1813232, "step": 6755 }, { "epoch": 1.7719528178243773, "grad_norm": 2.5067317485809326, "learning_rate": 4.909846286882772e-05, "loss": 0.4088, "num_input_tokens_seen": 1814832, "step": 6760 }, { "epoch": 1.7732634338138924, "grad_norm": 1.7278116941452026, "learning_rate": 4.909541659982659e-05, "loss": 0.3842, "num_input_tokens_seen": 1815808, "step": 6765 }, { "epoch": 1.7745740498034075, "grad_norm": 0.925725519657135, "learning_rate": 4.9092365287696396e-05, "loss": 0.4042, "num_input_tokens_seen": 1817296, "step": 6770 }, { "epoch": 1.7758846657929226, "grad_norm": 1.2817479372024536, "learning_rate": 4.908930893307578e-05, "loss": 0.3336, "num_input_tokens_seen": 1818528, "step": 6775 }, { "epoch": 1.7771952817824377, "grad_norm": 1.064948558807373, "learning_rate": 4.9086247536604425e-05, "loss": 0.4931, "num_input_tokens_seen": 1819888, "step": 6780 }, { "epoch": 1.7785058977719528, "grad_norm": 0.9128989577293396, "learning_rate": 4.908318109892308e-05, "loss": 0.4024, "num_input_tokens_seen": 1821232, "step": 6785 }, { "epoch": 1.7798165137614679, "grad_norm": 0.7392311096191406, "learning_rate": 4.9080109620673546e-05, "loss": 0.3975, "num_input_tokens_seen": 1822544, "step": 6790 }, { "epoch": 1.781127129750983, "grad_norm": 1.134143590927124, "learning_rate": 4.9077033102498666e-05, "loss": 0.5811, "num_input_tokens_seen": 1823712, "step": 6795 }, { "epoch": 1.782437745740498, "grad_norm": 0.9781844019889832, "learning_rate": 4.907395154504236e-05, "loss": 0.4371, "num_input_tokens_seen": 1824736, "step": 6800 }, { "epoch": 1.7837483617300132, "grad_norm": 0.9062517881393433, "learning_rate": 4.9070864948949604e-05, "loss": 0.3096, "num_input_tokens_seen": 1825920, "step": 6805 }, { "epoch": 1.7850589777195283, "grad_norm": 0.444137305021286, "learning_rate": 4.9067773314866396e-05, "loss": 0.5162, "num_input_tokens_seen": 1827104, "step": 6810 }, { "epoch": 1.7863695937090434, "grad_norm": 1.1803456544876099, "learning_rate": 4.906467664343982e-05, "loss": 0.363, "num_input_tokens_seen": 1829024, "step": 6815 }, { "epoch": 1.7876802096985585, "grad_norm": 1.1048377752304077, "learning_rate": 4.9061574935318006e-05, "loss": 0.3901, "num_input_tokens_seen": 1830336, "step": 6820 }, { "epoch": 1.7889908256880735, "grad_norm": 1.7686864137649536, "learning_rate": 4.905846819115013e-05, "loss": 0.4432, "num_input_tokens_seen": 1831360, "step": 6825 }, { "epoch": 1.7903014416775884, "grad_norm": 0.7421029806137085, "learning_rate": 4.905535641158644e-05, "loss": 0.2388, "num_input_tokens_seen": 1832800, "step": 6830 }, { "epoch": 1.7916120576671035, "grad_norm": 1.8699222803115845, "learning_rate": 4.9052239597278206e-05, "loss": 0.3833, "num_input_tokens_seen": 1833792, "step": 6835 }, { "epoch": 1.7929226736566186, "grad_norm": 2.3744802474975586, "learning_rate": 4.904911774887779e-05, "loss": 0.4249, "num_input_tokens_seen": 1835008, "step": 6840 }, { "epoch": 1.7942332896461337, "grad_norm": 0.8089054822921753, "learning_rate": 4.904599086703858e-05, "loss": 0.4375, "num_input_tokens_seen": 1836256, "step": 6845 }, { "epoch": 1.7955439056356488, "grad_norm": 2.1441421508789062, "learning_rate": 4.904285895241503e-05, "loss": 0.4768, "num_input_tokens_seen": 1837408, "step": 6850 }, { "epoch": 1.7968545216251637, "grad_norm": 1.1263066530227661, "learning_rate": 4.903972200566265e-05, "loss": 0.5525, "num_input_tokens_seen": 1838736, "step": 6855 }, { "epoch": 1.7981651376146788, "grad_norm": 0.699854850769043, "learning_rate": 4.903658002743798e-05, "loss": 0.4391, "num_input_tokens_seen": 1840016, "step": 6860 }, { "epoch": 1.7994757536041939, "grad_norm": 0.7241958975791931, "learning_rate": 4.9033433018398654e-05, "loss": 0.4952, "num_input_tokens_seen": 1841040, "step": 6865 }, { "epoch": 1.800786369593709, "grad_norm": 0.4314689040184021, "learning_rate": 4.9030280979203314e-05, "loss": 0.3754, "num_input_tokens_seen": 1842128, "step": 6870 }, { "epoch": 1.802096985583224, "grad_norm": 0.9407453536987305, "learning_rate": 4.90271239105117e-05, "loss": 0.2286, "num_input_tokens_seen": 1843632, "step": 6875 }, { "epoch": 1.8034076015727392, "grad_norm": 1.7983976602554321, "learning_rate": 4.902396181298455e-05, "loss": 0.4954, "num_input_tokens_seen": 1844704, "step": 6880 }, { "epoch": 1.8047182175622543, "grad_norm": 0.5178771018981934, "learning_rate": 4.902079468728371e-05, "loss": 0.519, "num_input_tokens_seen": 1846080, "step": 6885 }, { "epoch": 1.8060288335517694, "grad_norm": 1.0932344198226929, "learning_rate": 4.901762253407205e-05, "loss": 0.4288, "num_input_tokens_seen": 1847024, "step": 6890 }, { "epoch": 1.8073394495412844, "grad_norm": 1.2140244245529175, "learning_rate": 4.9014445354013496e-05, "loss": 0.4732, "num_input_tokens_seen": 1848352, "step": 6895 }, { "epoch": 1.8086500655307995, "grad_norm": 0.8260429501533508, "learning_rate": 4.901126314777301e-05, "loss": 0.6025, "num_input_tokens_seen": 1849360, "step": 6900 }, { "epoch": 1.8099606815203146, "grad_norm": 0.45188823342323303, "learning_rate": 4.9008075916016646e-05, "loss": 0.4676, "num_input_tokens_seen": 1850848, "step": 6905 }, { "epoch": 1.8112712975098297, "grad_norm": 0.8287671208381653, "learning_rate": 4.900488365941147e-05, "loss": 0.4678, "num_input_tokens_seen": 1851824, "step": 6910 }, { "epoch": 1.8125819134993448, "grad_norm": 0.6101911664009094, "learning_rate": 4.900168637862563e-05, "loss": 0.4341, "num_input_tokens_seen": 1853056, "step": 6915 }, { "epoch": 1.81389252948886, "grad_norm": 2.0748720169067383, "learning_rate": 4.899848407432829e-05, "loss": 0.442, "num_input_tokens_seen": 1854480, "step": 6920 }, { "epoch": 1.8152031454783748, "grad_norm": 0.716832160949707, "learning_rate": 4.8995276747189714e-05, "loss": 0.423, "num_input_tokens_seen": 1856384, "step": 6925 }, { "epoch": 1.81651376146789, "grad_norm": 0.4405304491519928, "learning_rate": 4.899206439788117e-05, "loss": 0.2919, "num_input_tokens_seen": 1857584, "step": 6930 }, { "epoch": 1.817824377457405, "grad_norm": 1.4279087781906128, "learning_rate": 4.898884702707499e-05, "loss": 0.7611, "num_input_tokens_seen": 1858960, "step": 6935 }, { "epoch": 1.81913499344692, "grad_norm": 1.0433109998703003, "learning_rate": 4.898562463544459e-05, "loss": 0.5676, "num_input_tokens_seen": 1859920, "step": 6940 }, { "epoch": 1.820445609436435, "grad_norm": 1.0660706758499146, "learning_rate": 4.89823972236644e-05, "loss": 0.5601, "num_input_tokens_seen": 1861280, "step": 6945 }, { "epoch": 1.82175622542595, "grad_norm": 1.1305385828018188, "learning_rate": 4.89791647924099e-05, "loss": 0.5975, "num_input_tokens_seen": 1862576, "step": 6950 }, { "epoch": 1.8230668414154652, "grad_norm": 0.4609510898590088, "learning_rate": 4.8975927342357646e-05, "loss": 0.5637, "num_input_tokens_seen": 1864192, "step": 6955 }, { "epoch": 1.8243774574049803, "grad_norm": 0.6815983057022095, "learning_rate": 4.897268487418523e-05, "loss": 0.4284, "num_input_tokens_seen": 1865664, "step": 6960 }, { "epoch": 1.8256880733944953, "grad_norm": 0.9924338459968567, "learning_rate": 4.896943738857128e-05, "loss": 0.4882, "num_input_tokens_seen": 1867072, "step": 6965 }, { "epoch": 1.8269986893840104, "grad_norm": 0.937540590763092, "learning_rate": 4.896618488619551e-05, "loss": 0.4059, "num_input_tokens_seen": 1868160, "step": 6970 }, { "epoch": 1.8283093053735255, "grad_norm": 1.491365671157837, "learning_rate": 4.8962927367738655e-05, "loss": 0.4655, "num_input_tokens_seen": 1869104, "step": 6975 }, { "epoch": 1.8296199213630406, "grad_norm": 0.7606911659240723, "learning_rate": 4.89596648338825e-05, "loss": 0.4131, "num_input_tokens_seen": 1870448, "step": 6980 }, { "epoch": 1.8309305373525557, "grad_norm": 0.8070969581604004, "learning_rate": 4.89563972853099e-05, "loss": 0.6022, "num_input_tokens_seen": 1871904, "step": 6985 }, { "epoch": 1.8322411533420708, "grad_norm": 0.7106375694274902, "learning_rate": 4.895312472270475e-05, "loss": 0.298, "num_input_tokens_seen": 1873216, "step": 6990 }, { "epoch": 1.833551769331586, "grad_norm": 0.6376100182533264, "learning_rate": 4.894984714675197e-05, "loss": 0.579, "num_input_tokens_seen": 1874480, "step": 6995 }, { "epoch": 1.834862385321101, "grad_norm": 0.6005480289459229, "learning_rate": 4.8946564558137566e-05, "loss": 0.5231, "num_input_tokens_seen": 1876064, "step": 7000 }, { "epoch": 1.8361730013106161, "grad_norm": 1.0351533889770508, "learning_rate": 4.894327695754858e-05, "loss": 0.2947, "num_input_tokens_seen": 1877488, "step": 7005 }, { "epoch": 1.8374836173001312, "grad_norm": 1.0640324354171753, "learning_rate": 4.893998434567309e-05, "loss": 0.4503, "num_input_tokens_seen": 1878880, "step": 7010 }, { "epoch": 1.838794233289646, "grad_norm": 1.579492449760437, "learning_rate": 4.893668672320025e-05, "loss": 0.4441, "num_input_tokens_seen": 1880416, "step": 7015 }, { "epoch": 1.8401048492791612, "grad_norm": 0.6291418671607971, "learning_rate": 4.893338409082023e-05, "loss": 0.4893, "num_input_tokens_seen": 1881936, "step": 7020 }, { "epoch": 1.8414154652686763, "grad_norm": 1.2711575031280518, "learning_rate": 4.8930076449224285e-05, "loss": 0.4489, "num_input_tokens_seen": 1883056, "step": 7025 }, { "epoch": 1.8427260812581914, "grad_norm": 0.8683908581733704, "learning_rate": 4.892676379910468e-05, "loss": 0.4462, "num_input_tokens_seen": 1884192, "step": 7030 }, { "epoch": 1.8440366972477065, "grad_norm": 0.7845191955566406, "learning_rate": 4.8923446141154746e-05, "loss": 0.4124, "num_input_tokens_seen": 1885968, "step": 7035 }, { "epoch": 1.8453473132372213, "grad_norm": 1.34452486038208, "learning_rate": 4.8920123476068894e-05, "loss": 0.5432, "num_input_tokens_seen": 1887024, "step": 7040 }, { "epoch": 1.8466579292267364, "grad_norm": 0.9982672929763794, "learning_rate": 4.891679580454251e-05, "loss": 0.5151, "num_input_tokens_seen": 1888288, "step": 7045 }, { "epoch": 1.8479685452162515, "grad_norm": 0.8999399542808533, "learning_rate": 4.89134631272721e-05, "loss": 0.5425, "num_input_tokens_seen": 1890128, "step": 7050 }, { "epoch": 1.8492791612057666, "grad_norm": 0.6143273711204529, "learning_rate": 4.891012544495517e-05, "loss": 0.5913, "num_input_tokens_seen": 1891584, "step": 7055 }, { "epoch": 1.8505897771952817, "grad_norm": 1.5042873620986938, "learning_rate": 4.890678275829029e-05, "loss": 0.4084, "num_input_tokens_seen": 1892944, "step": 7060 }, { "epoch": 1.8519003931847968, "grad_norm": 1.0172028541564941, "learning_rate": 4.89034350679771e-05, "loss": 0.4063, "num_input_tokens_seen": 1894144, "step": 7065 }, { "epoch": 1.853211009174312, "grad_norm": 1.2649682760238647, "learning_rate": 4.890008237471625e-05, "loss": 0.6676, "num_input_tokens_seen": 1895056, "step": 7070 }, { "epoch": 1.854521625163827, "grad_norm": 0.7674832940101624, "learning_rate": 4.889672467920945e-05, "loss": 0.4497, "num_input_tokens_seen": 1896416, "step": 7075 }, { "epoch": 1.855832241153342, "grad_norm": 1.0371665954589844, "learning_rate": 4.889336198215947e-05, "loss": 0.5831, "num_input_tokens_seen": 1897456, "step": 7080 }, { "epoch": 1.8571428571428572, "grad_norm": 0.5730865001678467, "learning_rate": 4.888999428427011e-05, "loss": 0.4698, "num_input_tokens_seen": 1898816, "step": 7085 }, { "epoch": 1.8584534731323723, "grad_norm": 1.045027732849121, "learning_rate": 4.8886621586246226e-05, "loss": 0.5691, "num_input_tokens_seen": 1900176, "step": 7090 }, { "epoch": 1.8597640891218874, "grad_norm": 0.4577600955963135, "learning_rate": 4.8883243888793715e-05, "loss": 0.3212, "num_input_tokens_seen": 1901424, "step": 7095 }, { "epoch": 1.8610747051114025, "grad_norm": 0.944434642791748, "learning_rate": 4.887986119261953e-05, "loss": 0.3285, "num_input_tokens_seen": 1903616, "step": 7100 }, { "epoch": 1.8623853211009176, "grad_norm": 0.7478685975074768, "learning_rate": 4.887647349843165e-05, "loss": 0.3257, "num_input_tokens_seen": 1904928, "step": 7105 }, { "epoch": 1.8636959370904325, "grad_norm": 1.6358598470687866, "learning_rate": 4.887308080693913e-05, "loss": 0.4497, "num_input_tokens_seen": 1907632, "step": 7110 }, { "epoch": 1.8650065530799476, "grad_norm": 1.0343245267868042, "learning_rate": 4.886968311885204e-05, "loss": 0.3965, "num_input_tokens_seen": 1908672, "step": 7115 }, { "epoch": 1.8663171690694627, "grad_norm": 0.6615481376647949, "learning_rate": 4.8866280434881516e-05, "loss": 1.0075, "num_input_tokens_seen": 1909808, "step": 7120 }, { "epoch": 1.8676277850589778, "grad_norm": 0.9559385180473328, "learning_rate": 4.886287275573974e-05, "loss": 0.4652, "num_input_tokens_seen": 1911248, "step": 7125 }, { "epoch": 1.8689384010484926, "grad_norm": 1.5528619289398193, "learning_rate": 4.885946008213993e-05, "loss": 0.5468, "num_input_tokens_seen": 1912688, "step": 7130 }, { "epoch": 1.8702490170380077, "grad_norm": 1.2007777690887451, "learning_rate": 4.885604241479635e-05, "loss": 0.5446, "num_input_tokens_seen": 1914384, "step": 7135 }, { "epoch": 1.8715596330275228, "grad_norm": 0.5862559080123901, "learning_rate": 4.885261975442431e-05, "loss": 0.3864, "num_input_tokens_seen": 1915696, "step": 7140 }, { "epoch": 1.872870249017038, "grad_norm": 0.6858235597610474, "learning_rate": 4.884919210174017e-05, "loss": 0.6468, "num_input_tokens_seen": 1917424, "step": 7145 }, { "epoch": 1.874180865006553, "grad_norm": 0.8065508604049683, "learning_rate": 4.8845759457461326e-05, "loss": 0.3608, "num_input_tokens_seen": 1918800, "step": 7150 }, { "epoch": 1.875491480996068, "grad_norm": 0.6845648884773254, "learning_rate": 4.8842321822306235e-05, "loss": 0.3639, "num_input_tokens_seen": 1920240, "step": 7155 }, { "epoch": 1.8768020969855832, "grad_norm": 0.5481168031692505, "learning_rate": 4.8838879196994385e-05, "loss": 0.5869, "num_input_tokens_seen": 1921696, "step": 7160 }, { "epoch": 1.8781127129750983, "grad_norm": 1.4231078624725342, "learning_rate": 4.88354315822463e-05, "loss": 0.6143, "num_input_tokens_seen": 1922848, "step": 7165 }, { "epoch": 1.8794233289646134, "grad_norm": 0.6936765313148499, "learning_rate": 4.8831978978783585e-05, "loss": 0.4319, "num_input_tokens_seen": 1924160, "step": 7170 }, { "epoch": 1.8807339449541285, "grad_norm": 0.4884100556373596, "learning_rate": 4.882852138732884e-05, "loss": 0.4315, "num_input_tokens_seen": 1925792, "step": 7175 }, { "epoch": 1.8820445609436436, "grad_norm": 0.6424300670623779, "learning_rate": 4.882505880860574e-05, "loss": 0.3243, "num_input_tokens_seen": 1927312, "step": 7180 }, { "epoch": 1.8833551769331587, "grad_norm": 1.1730873584747314, "learning_rate": 4.8821591243339e-05, "loss": 0.4104, "num_input_tokens_seen": 1929248, "step": 7185 }, { "epoch": 1.8846657929226738, "grad_norm": 0.8907444477081299, "learning_rate": 4.8818118692254375e-05, "loss": 0.3467, "num_input_tokens_seen": 1930464, "step": 7190 }, { "epoch": 1.8859764089121889, "grad_norm": 0.8938016295433044, "learning_rate": 4.881464115607865e-05, "loss": 0.4919, "num_input_tokens_seen": 1931552, "step": 7195 }, { "epoch": 1.8872870249017037, "grad_norm": 0.3641728460788727, "learning_rate": 4.88111586355397e-05, "loss": 0.2714, "num_input_tokens_seen": 1932768, "step": 7200 }, { "epoch": 1.8885976408912188, "grad_norm": 1.0881500244140625, "learning_rate": 4.880767113136637e-05, "loss": 0.3929, "num_input_tokens_seen": 1934064, "step": 7205 }, { "epoch": 1.889908256880734, "grad_norm": 0.8638085126876831, "learning_rate": 4.880417864428862e-05, "loss": 0.4931, "num_input_tokens_seen": 1935744, "step": 7210 }, { "epoch": 1.891218872870249, "grad_norm": 0.7231525182723999, "learning_rate": 4.880068117503741e-05, "loss": 0.5083, "num_input_tokens_seen": 1937232, "step": 7215 }, { "epoch": 1.8925294888597641, "grad_norm": 1.1456974744796753, "learning_rate": 4.879717872434475e-05, "loss": 0.516, "num_input_tokens_seen": 1938704, "step": 7220 }, { "epoch": 1.893840104849279, "grad_norm": 1.0615862607955933, "learning_rate": 4.879367129294371e-05, "loss": 0.6729, "num_input_tokens_seen": 1939696, "step": 7225 }, { "epoch": 1.895150720838794, "grad_norm": 0.7003543972969055, "learning_rate": 4.879015888156837e-05, "loss": 0.553, "num_input_tokens_seen": 1940960, "step": 7230 }, { "epoch": 1.8964613368283092, "grad_norm": 2.048478364944458, "learning_rate": 4.8786641490953886e-05, "loss": 0.5512, "num_input_tokens_seen": 1942160, "step": 7235 }, { "epoch": 1.8977719528178243, "grad_norm": 0.9915602803230286, "learning_rate": 4.878311912183644e-05, "loss": 0.4997, "num_input_tokens_seen": 1943840, "step": 7240 }, { "epoch": 1.8990825688073394, "grad_norm": 0.5905684232711792, "learning_rate": 4.877959177495326e-05, "loss": 0.309, "num_input_tokens_seen": 1944896, "step": 7245 }, { "epoch": 1.9003931847968545, "grad_norm": 0.4719136953353882, "learning_rate": 4.87760594510426e-05, "loss": 0.5092, "num_input_tokens_seen": 1945952, "step": 7250 }, { "epoch": 1.9017038007863696, "grad_norm": 1.1023880243301392, "learning_rate": 4.877252215084379e-05, "loss": 0.4376, "num_input_tokens_seen": 1947072, "step": 7255 }, { "epoch": 1.9030144167758847, "grad_norm": 0.5127874612808228, "learning_rate": 4.876897987509716e-05, "loss": 0.3793, "num_input_tokens_seen": 1948128, "step": 7260 }, { "epoch": 1.9043250327653998, "grad_norm": 0.7499842643737793, "learning_rate": 4.876543262454412e-05, "loss": 0.5058, "num_input_tokens_seen": 1949344, "step": 7265 }, { "epoch": 1.9056356487549149, "grad_norm": 0.7125519514083862, "learning_rate": 4.8761880399927084e-05, "loss": 0.4692, "num_input_tokens_seen": 1950848, "step": 7270 }, { "epoch": 1.90694626474443, "grad_norm": 1.0635372400283813, "learning_rate": 4.875832320198954e-05, "loss": 0.487, "num_input_tokens_seen": 1951904, "step": 7275 }, { "epoch": 1.908256880733945, "grad_norm": 0.6712915301322937, "learning_rate": 4.8754761031476005e-05, "loss": 0.5505, "num_input_tokens_seen": 1953664, "step": 7280 }, { "epoch": 1.9095674967234602, "grad_norm": 1.7918682098388672, "learning_rate": 4.875119388913203e-05, "loss": 0.4727, "num_input_tokens_seen": 1955328, "step": 7285 }, { "epoch": 1.9108781127129753, "grad_norm": 0.8559989929199219, "learning_rate": 4.874762177570422e-05, "loss": 0.4517, "num_input_tokens_seen": 1956624, "step": 7290 }, { "epoch": 1.9121887287024901, "grad_norm": 0.9230644106864929, "learning_rate": 4.874404469194019e-05, "loss": 0.4726, "num_input_tokens_seen": 1957872, "step": 7295 }, { "epoch": 1.9134993446920052, "grad_norm": 0.7732226848602295, "learning_rate": 4.8740462638588644e-05, "loss": 0.3246, "num_input_tokens_seen": 1958864, "step": 7300 }, { "epoch": 1.9148099606815203, "grad_norm": 0.7017455697059631, "learning_rate": 4.873687561639929e-05, "loss": 0.5356, "num_input_tokens_seen": 1960160, "step": 7305 }, { "epoch": 1.9161205766710354, "grad_norm": 1.560852289199829, "learning_rate": 4.8733283626122875e-05, "loss": 0.6258, "num_input_tokens_seen": 1961248, "step": 7310 }, { "epoch": 1.9174311926605505, "grad_norm": 8.44855785369873, "learning_rate": 4.872968666851121e-05, "loss": 0.3841, "num_input_tokens_seen": 1962400, "step": 7315 }, { "epoch": 1.9187418086500654, "grad_norm": 0.8995791673660278, "learning_rate": 4.872608474431712e-05, "loss": 0.5505, "num_input_tokens_seen": 1964352, "step": 7320 }, { "epoch": 1.9200524246395805, "grad_norm": 0.8487437963485718, "learning_rate": 4.87224778542945e-05, "loss": 0.3056, "num_input_tokens_seen": 1965376, "step": 7325 }, { "epoch": 1.9213630406290956, "grad_norm": 0.6935005187988281, "learning_rate": 4.871886599919825e-05, "loss": 0.4364, "num_input_tokens_seen": 1966512, "step": 7330 }, { "epoch": 1.9226736566186107, "grad_norm": 1.236052393913269, "learning_rate": 4.871524917978433e-05, "loss": 0.5592, "num_input_tokens_seen": 1968112, "step": 7335 }, { "epoch": 1.9239842726081258, "grad_norm": 0.34944120049476624, "learning_rate": 4.8711627396809725e-05, "loss": 0.3916, "num_input_tokens_seen": 1969472, "step": 7340 }, { "epoch": 1.9252948885976409, "grad_norm": 1.9920048713684082, "learning_rate": 4.870800065103248e-05, "loss": 0.5303, "num_input_tokens_seen": 1970768, "step": 7345 }, { "epoch": 1.926605504587156, "grad_norm": 2.245887517929077, "learning_rate": 4.870436894321167e-05, "loss": 0.6357, "num_input_tokens_seen": 1972432, "step": 7350 }, { "epoch": 1.927916120576671, "grad_norm": 3.51375412940979, "learning_rate": 4.87007322741074e-05, "loss": 0.7518, "num_input_tokens_seen": 1974128, "step": 7355 }, { "epoch": 1.9292267365661862, "grad_norm": 1.0011354684829712, "learning_rate": 4.8697090644480804e-05, "loss": 0.3668, "num_input_tokens_seen": 1975616, "step": 7360 }, { "epoch": 1.9305373525557012, "grad_norm": 0.5385328531265259, "learning_rate": 4.86934440550941e-05, "loss": 0.6013, "num_input_tokens_seen": 1976944, "step": 7365 }, { "epoch": 1.9318479685452163, "grad_norm": 2.8192293643951416, "learning_rate": 4.868979250671047e-05, "loss": 0.5366, "num_input_tokens_seen": 1977856, "step": 7370 }, { "epoch": 1.9331585845347314, "grad_norm": 1.9690088033676147, "learning_rate": 4.868613600009422e-05, "loss": 0.4332, "num_input_tokens_seen": 1979184, "step": 7375 }, { "epoch": 1.9344692005242465, "grad_norm": 0.5745998024940491, "learning_rate": 4.8682474536010624e-05, "loss": 0.4004, "num_input_tokens_seen": 1980512, "step": 7380 }, { "epoch": 1.9357798165137616, "grad_norm": 0.6280592679977417, "learning_rate": 4.8678808115226026e-05, "loss": 0.3132, "num_input_tokens_seen": 1981936, "step": 7385 }, { "epoch": 1.9370904325032765, "grad_norm": 0.8607454895973206, "learning_rate": 4.867513673850781e-05, "loss": 0.5072, "num_input_tokens_seen": 1983376, "step": 7390 }, { "epoch": 1.9384010484927916, "grad_norm": 0.41156861186027527, "learning_rate": 4.867146040662438e-05, "loss": 0.5803, "num_input_tokens_seen": 1984912, "step": 7395 }, { "epoch": 1.9397116644823067, "grad_norm": 1.9264333248138428, "learning_rate": 4.8667779120345185e-05, "loss": 0.4393, "num_input_tokens_seen": 1986112, "step": 7400 }, { "epoch": 1.9410222804718218, "grad_norm": 0.4985097050666809, "learning_rate": 4.866409288044072e-05, "loss": 0.4696, "num_input_tokens_seen": 1987824, "step": 7405 }, { "epoch": 1.9423328964613367, "grad_norm": 0.6533564925193787, "learning_rate": 4.86604016876825e-05, "loss": 0.4379, "num_input_tokens_seen": 1988752, "step": 7410 }, { "epoch": 1.9436435124508518, "grad_norm": 1.2371602058410645, "learning_rate": 4.8656705542843085e-05, "loss": 0.5543, "num_input_tokens_seen": 1990384, "step": 7415 }, { "epoch": 1.9449541284403669, "grad_norm": 0.5798689723014832, "learning_rate": 4.865300444669608e-05, "loss": 0.7042, "num_input_tokens_seen": 1991504, "step": 7420 }, { "epoch": 1.946264744429882, "grad_norm": 0.8448935151100159, "learning_rate": 4.864929840001612e-05, "loss": 0.3148, "num_input_tokens_seen": 1993040, "step": 7425 }, { "epoch": 1.947575360419397, "grad_norm": 0.6315438151359558, "learning_rate": 4.8645587403578855e-05, "loss": 0.3478, "num_input_tokens_seen": 1994176, "step": 7430 }, { "epoch": 1.9488859764089121, "grad_norm": 0.6832272410392761, "learning_rate": 4.864187145816101e-05, "loss": 0.3681, "num_input_tokens_seen": 1995232, "step": 7435 }, { "epoch": 1.9501965923984272, "grad_norm": 4.150203227996826, "learning_rate": 4.863815056454032e-05, "loss": 0.627, "num_input_tokens_seen": 1996464, "step": 7440 }, { "epoch": 1.9515072083879423, "grad_norm": 0.2572241425514221, "learning_rate": 4.863442472349555e-05, "loss": 0.3803, "num_input_tokens_seen": 1997488, "step": 7445 }, { "epoch": 1.9528178243774574, "grad_norm": 0.820608913898468, "learning_rate": 4.8630693935806536e-05, "loss": 0.4416, "num_input_tokens_seen": 1998608, "step": 7450 }, { "epoch": 1.9541284403669725, "grad_norm": 1.626372218132019, "learning_rate": 4.8626958202254106e-05, "loss": 0.6401, "num_input_tokens_seen": 1999728, "step": 7455 }, { "epoch": 1.9554390563564876, "grad_norm": 0.5626112222671509, "learning_rate": 4.862321752362015e-05, "loss": 0.6034, "num_input_tokens_seen": 2000848, "step": 7460 }, { "epoch": 1.9567496723460027, "grad_norm": 1.8096959590911865, "learning_rate": 4.8619471900687583e-05, "loss": 0.7379, "num_input_tokens_seen": 2001840, "step": 7465 }, { "epoch": 1.9580602883355178, "grad_norm": 3.913302183151245, "learning_rate": 4.861572133424036e-05, "loss": 0.4158, "num_input_tokens_seen": 2003072, "step": 7470 }, { "epoch": 1.959370904325033, "grad_norm": 0.5786474347114563, "learning_rate": 4.8611965825063466e-05, "loss": 0.4933, "num_input_tokens_seen": 2004448, "step": 7475 }, { "epoch": 1.9606815203145478, "grad_norm": 0.8555788993835449, "learning_rate": 4.8608205373942915e-05, "loss": 0.4657, "num_input_tokens_seen": 2005472, "step": 7480 }, { "epoch": 1.9619921363040629, "grad_norm": 0.7093372941017151, "learning_rate": 4.860443998166578e-05, "loss": 0.3521, "num_input_tokens_seen": 2006752, "step": 7485 }, { "epoch": 1.963302752293578, "grad_norm": 1.2195024490356445, "learning_rate": 4.860066964902014e-05, "loss": 0.5095, "num_input_tokens_seen": 2008048, "step": 7490 }, { "epoch": 1.964613368283093, "grad_norm": 0.422389417886734, "learning_rate": 4.859689437679512e-05, "loss": 0.5144, "num_input_tokens_seen": 2009520, "step": 7495 }, { "epoch": 1.9659239842726082, "grad_norm": 1.601709008216858, "learning_rate": 4.859311416578088e-05, "loss": 0.4333, "num_input_tokens_seen": 2010848, "step": 7500 }, { "epoch": 1.967234600262123, "grad_norm": 0.5003732442855835, "learning_rate": 4.8589329016768615e-05, "loss": 0.7015, "num_input_tokens_seen": 2012336, "step": 7505 }, { "epoch": 1.9685452162516381, "grad_norm": 0.6085672378540039, "learning_rate": 4.8585538930550547e-05, "loss": 0.5533, "num_input_tokens_seen": 2013872, "step": 7510 }, { "epoch": 1.9698558322411532, "grad_norm": 1.4773823022842407, "learning_rate": 4.858174390791993e-05, "loss": 0.4016, "num_input_tokens_seen": 2015040, "step": 7515 }, { "epoch": 1.9711664482306683, "grad_norm": 0.6098823547363281, "learning_rate": 4.857794394967106e-05, "loss": 0.3361, "num_input_tokens_seen": 2016256, "step": 7520 }, { "epoch": 1.9724770642201834, "grad_norm": 0.5034121870994568, "learning_rate": 4.8574139056599256e-05, "loss": 0.3653, "num_input_tokens_seen": 2017984, "step": 7525 }, { "epoch": 1.9737876802096985, "grad_norm": 0.9562009572982788, "learning_rate": 4.8570329229500884e-05, "loss": 0.5499, "num_input_tokens_seen": 2019488, "step": 7530 }, { "epoch": 1.9750982961992136, "grad_norm": 2.1058411598205566, "learning_rate": 4.8566514469173326e-05, "loss": 0.3764, "num_input_tokens_seen": 2020736, "step": 7535 }, { "epoch": 1.9764089121887287, "grad_norm": 0.35773107409477234, "learning_rate": 4.8562694776415016e-05, "loss": 0.4331, "num_input_tokens_seen": 2022240, "step": 7540 }, { "epoch": 1.9777195281782438, "grad_norm": 5.913983345031738, "learning_rate": 4.8558870152025404e-05, "loss": 0.7079, "num_input_tokens_seen": 2023936, "step": 7545 }, { "epoch": 1.979030144167759, "grad_norm": 0.3639586269855499, "learning_rate": 4.855504059680497e-05, "loss": 0.3996, "num_input_tokens_seen": 2025248, "step": 7550 }, { "epoch": 1.980340760157274, "grad_norm": 0.440568208694458, "learning_rate": 4.855120611155524e-05, "loss": 0.3253, "num_input_tokens_seen": 2026960, "step": 7555 }, { "epoch": 1.981651376146789, "grad_norm": 0.7400256991386414, "learning_rate": 4.854736669707876e-05, "loss": 0.4233, "num_input_tokens_seen": 2028384, "step": 7560 }, { "epoch": 1.9829619921363042, "grad_norm": 1.0712593793869019, "learning_rate": 4.854352235417913e-05, "loss": 0.519, "num_input_tokens_seen": 2029856, "step": 7565 }, { "epoch": 1.9842726081258193, "grad_norm": 1.200717568397522, "learning_rate": 4.853967308366094e-05, "loss": 0.3763, "num_input_tokens_seen": 2031104, "step": 7570 }, { "epoch": 1.9855832241153342, "grad_norm": 1.162473201751709, "learning_rate": 4.8535818886329855e-05, "loss": 0.5295, "num_input_tokens_seen": 2032368, "step": 7575 }, { "epoch": 1.9868938401048493, "grad_norm": 0.5971775650978088, "learning_rate": 4.853195976299254e-05, "loss": 0.4139, "num_input_tokens_seen": 2033536, "step": 7580 }, { "epoch": 1.9882044560943644, "grad_norm": 0.6405415534973145, "learning_rate": 4.85280957144567e-05, "loss": 0.4604, "num_input_tokens_seen": 2035600, "step": 7585 }, { "epoch": 1.9895150720838795, "grad_norm": 1.1007750034332275, "learning_rate": 4.852422674153109e-05, "loss": 0.5578, "num_input_tokens_seen": 2036928, "step": 7590 }, { "epoch": 1.9908256880733946, "grad_norm": 1.224938988685608, "learning_rate": 4.852035284502547e-05, "loss": 0.3217, "num_input_tokens_seen": 2038144, "step": 7595 }, { "epoch": 1.9921363040629094, "grad_norm": 0.591335654258728, "learning_rate": 4.851647402575063e-05, "loss": 0.5144, "num_input_tokens_seen": 2039328, "step": 7600 }, { "epoch": 1.9934469200524245, "grad_norm": 0.7974773645401001, "learning_rate": 4.851259028451842e-05, "loss": 0.5338, "num_input_tokens_seen": 2040496, "step": 7605 }, { "epoch": 1.9947575360419396, "grad_norm": 0.8852341175079346, "learning_rate": 4.850870162214168e-05, "loss": 0.4701, "num_input_tokens_seen": 2041904, "step": 7610 }, { "epoch": 1.9960681520314547, "grad_norm": 0.19125144183635712, "learning_rate": 4.850480803943432e-05, "loss": 0.623, "num_input_tokens_seen": 2043376, "step": 7615 }, { "epoch": 1.9973787680209698, "grad_norm": 1.3045613765716553, "learning_rate": 4.8500909537211246e-05, "loss": 0.4689, "num_input_tokens_seen": 2044736, "step": 7620 }, { "epoch": 1.998689384010485, "grad_norm": 0.5382546782493591, "learning_rate": 4.849700611628841e-05, "loss": 0.3406, "num_input_tokens_seen": 2046016, "step": 7625 }, { "epoch": 2.0, "grad_norm": 0.6879667639732361, "learning_rate": 4.84930977774828e-05, "loss": 0.4912, "num_input_tokens_seen": 2047128, "step": 7630 }, { "epoch": 2.000524246395806, "eval_loss": 0.49409183859825134, "eval_runtime": 16.678, "eval_samples_per_second": 50.846, "eval_steps_per_second": 25.423, "num_input_tokens_seen": 2047464, "step": 7632 }, { "epoch": 2.001310615989515, "grad_norm": 0.5073279142379761, "learning_rate": 4.8489184521612416e-05, "loss": 0.4047, "num_input_tokens_seen": 2048456, "step": 7635 }, { "epoch": 2.00262123197903, "grad_norm": 1.5758949518203735, "learning_rate": 4.848526634949629e-05, "loss": 0.487, "num_input_tokens_seen": 2049368, "step": 7640 }, { "epoch": 2.0039318479685453, "grad_norm": 1.2738746404647827, "learning_rate": 4.848134326195451e-05, "loss": 0.3526, "num_input_tokens_seen": 2050728, "step": 7645 }, { "epoch": 2.0052424639580604, "grad_norm": 1.254355788230896, "learning_rate": 4.847741525980815e-05, "loss": 0.3587, "num_input_tokens_seen": 2051864, "step": 7650 }, { "epoch": 2.0065530799475755, "grad_norm": 0.7049255967140198, "learning_rate": 4.8473482343879344e-05, "loss": 0.6309, "num_input_tokens_seen": 2053240, "step": 7655 }, { "epoch": 2.0078636959370906, "grad_norm": 1.1705043315887451, "learning_rate": 4.8469544514991235e-05, "loss": 0.4558, "num_input_tokens_seen": 2054280, "step": 7660 }, { "epoch": 2.0091743119266057, "grad_norm": 0.7467496395111084, "learning_rate": 4.846560177396802e-05, "loss": 0.307, "num_input_tokens_seen": 2055400, "step": 7665 }, { "epoch": 2.0104849279161208, "grad_norm": 0.5276650190353394, "learning_rate": 4.8461654121634905e-05, "loss": 0.4832, "num_input_tokens_seen": 2056952, "step": 7670 }, { "epoch": 2.011795543905636, "grad_norm": 1.2656160593032837, "learning_rate": 4.8457701558818106e-05, "loss": 0.3828, "num_input_tokens_seen": 2058104, "step": 7675 }, { "epoch": 2.0131061598951505, "grad_norm": 0.9983326196670532, "learning_rate": 4.845374408634491e-05, "loss": 0.3452, "num_input_tokens_seen": 2059224, "step": 7680 }, { "epoch": 2.0144167758846656, "grad_norm": 1.0417670011520386, "learning_rate": 4.84497817050436e-05, "loss": 0.4057, "num_input_tokens_seen": 2060776, "step": 7685 }, { "epoch": 2.0157273918741807, "grad_norm": 2.8644213676452637, "learning_rate": 4.8445814415743504e-05, "loss": 0.5588, "num_input_tokens_seen": 2062120, "step": 7690 }, { "epoch": 2.017038007863696, "grad_norm": 0.7203768491744995, "learning_rate": 4.844184221927495e-05, "loss": 0.399, "num_input_tokens_seen": 2063208, "step": 7695 }, { "epoch": 2.018348623853211, "grad_norm": 0.7051616311073303, "learning_rate": 4.843786511646933e-05, "loss": 0.6413, "num_input_tokens_seen": 2065096, "step": 7700 }, { "epoch": 2.019659239842726, "grad_norm": 1.589298129081726, "learning_rate": 4.843388310815904e-05, "loss": 0.5262, "num_input_tokens_seen": 2067304, "step": 7705 }, { "epoch": 2.020969855832241, "grad_norm": 0.768596887588501, "learning_rate": 4.8429896195177504e-05, "loss": 0.4201, "num_input_tokens_seen": 2068728, "step": 7710 }, { "epoch": 2.022280471821756, "grad_norm": 1.0093637704849243, "learning_rate": 4.842590437835918e-05, "loss": 0.4967, "num_input_tokens_seen": 2069880, "step": 7715 }, { "epoch": 2.0235910878112713, "grad_norm": 1.017867088317871, "learning_rate": 4.8421907658539545e-05, "loss": 0.372, "num_input_tokens_seen": 2071096, "step": 7720 }, { "epoch": 2.0249017038007864, "grad_norm": 0.7268002033233643, "learning_rate": 4.84179060365551e-05, "loss": 0.433, "num_input_tokens_seen": 2074312, "step": 7725 }, { "epoch": 2.0262123197903015, "grad_norm": 3.607614278793335, "learning_rate": 4.84138995132434e-05, "loss": 0.5673, "num_input_tokens_seen": 2075688, "step": 7730 }, { "epoch": 2.0275229357798166, "grad_norm": 0.5425214767456055, "learning_rate": 4.840988808944298e-05, "loss": 0.388, "num_input_tokens_seen": 2076808, "step": 7735 }, { "epoch": 2.0288335517693317, "grad_norm": 0.4413887858390808, "learning_rate": 4.8405871765993433e-05, "loss": 0.3792, "num_input_tokens_seen": 2078440, "step": 7740 }, { "epoch": 2.0301441677588468, "grad_norm": 0.4695689082145691, "learning_rate": 4.840185054373536e-05, "loss": 0.4815, "num_input_tokens_seen": 2079656, "step": 7745 }, { "epoch": 2.031454783748362, "grad_norm": 3.1990952491760254, "learning_rate": 4.839782442351042e-05, "loss": 0.5919, "num_input_tokens_seen": 2080712, "step": 7750 }, { "epoch": 2.032765399737877, "grad_norm": 0.7755283117294312, "learning_rate": 4.839379340616124e-05, "loss": 0.4443, "num_input_tokens_seen": 2082088, "step": 7755 }, { "epoch": 2.034076015727392, "grad_norm": 1.1045335531234741, "learning_rate": 4.838975749253153e-05, "loss": 0.4636, "num_input_tokens_seen": 2083208, "step": 7760 }, { "epoch": 2.035386631716907, "grad_norm": 1.396257758140564, "learning_rate": 4.838571668346599e-05, "loss": 0.4171, "num_input_tokens_seen": 2084312, "step": 7765 }, { "epoch": 2.036697247706422, "grad_norm": 0.49973440170288086, "learning_rate": 4.838167097981036e-05, "loss": 0.3818, "num_input_tokens_seen": 2085880, "step": 7770 }, { "epoch": 2.038007863695937, "grad_norm": 0.6683505177497864, "learning_rate": 4.8377620382411385e-05, "loss": 0.4323, "num_input_tokens_seen": 2087080, "step": 7775 }, { "epoch": 2.039318479685452, "grad_norm": 0.902432918548584, "learning_rate": 4.837356489211686e-05, "loss": 0.3997, "num_input_tokens_seen": 2088440, "step": 7780 }, { "epoch": 2.040629095674967, "grad_norm": 0.8663788437843323, "learning_rate": 4.836950450977559e-05, "loss": 0.421, "num_input_tokens_seen": 2089432, "step": 7785 }, { "epoch": 2.041939711664482, "grad_norm": 0.6652033925056458, "learning_rate": 4.836543923623741e-05, "loss": 0.3026, "num_input_tokens_seen": 2090952, "step": 7790 }, { "epoch": 2.0432503276539973, "grad_norm": 0.8631477952003479, "learning_rate": 4.8361369072353164e-05, "loss": 0.6333, "num_input_tokens_seen": 2092648, "step": 7795 }, { "epoch": 2.0445609436435124, "grad_norm": 1.0817385911941528, "learning_rate": 4.835729401897474e-05, "loss": 0.4751, "num_input_tokens_seen": 2093896, "step": 7800 }, { "epoch": 2.0458715596330275, "grad_norm": 0.989139974117279, "learning_rate": 4.835321407695503e-05, "loss": 0.4335, "num_input_tokens_seen": 2095320, "step": 7805 }, { "epoch": 2.0471821756225426, "grad_norm": 0.6938352584838867, "learning_rate": 4.834912924714797e-05, "loss": 0.3745, "num_input_tokens_seen": 2096488, "step": 7810 }, { "epoch": 2.0484927916120577, "grad_norm": 1.587478756904602, "learning_rate": 4.834503953040851e-05, "loss": 0.5356, "num_input_tokens_seen": 2097880, "step": 7815 }, { "epoch": 2.0498034076015728, "grad_norm": 2.077134847640991, "learning_rate": 4.83409449275926e-05, "loss": 0.4777, "num_input_tokens_seen": 2099160, "step": 7820 }, { "epoch": 2.051114023591088, "grad_norm": 1.869240164756775, "learning_rate": 4.8336845439557256e-05, "loss": 0.3739, "num_input_tokens_seen": 2100632, "step": 7825 }, { "epoch": 2.052424639580603, "grad_norm": 1.6039412021636963, "learning_rate": 4.833274106716048e-05, "loss": 0.2713, "num_input_tokens_seen": 2101688, "step": 7830 }, { "epoch": 2.053735255570118, "grad_norm": 0.46313178539276123, "learning_rate": 4.832863181126132e-05, "loss": 0.3649, "num_input_tokens_seen": 2103176, "step": 7835 }, { "epoch": 2.055045871559633, "grad_norm": 0.5038869976997375, "learning_rate": 4.832451767271983e-05, "loss": 0.7591, "num_input_tokens_seen": 2104360, "step": 7840 }, { "epoch": 2.0563564875491482, "grad_norm": 2.4724645614624023, "learning_rate": 4.832039865239708e-05, "loss": 0.3673, "num_input_tokens_seen": 2106072, "step": 7845 }, { "epoch": 2.0576671035386633, "grad_norm": 0.6493978500366211, "learning_rate": 4.83162747511552e-05, "loss": 0.3394, "num_input_tokens_seen": 2107288, "step": 7850 }, { "epoch": 2.0589777195281784, "grad_norm": 1.3475422859191895, "learning_rate": 4.83121459698573e-05, "loss": 0.4041, "num_input_tokens_seen": 2108456, "step": 7855 }, { "epoch": 2.0602883355176935, "grad_norm": 0.9086787700653076, "learning_rate": 4.830801230936752e-05, "loss": 0.5448, "num_input_tokens_seen": 2109896, "step": 7860 }, { "epoch": 2.061598951507208, "grad_norm": 1.6861252784729004, "learning_rate": 4.830387377055105e-05, "loss": 0.4116, "num_input_tokens_seen": 2111560, "step": 7865 }, { "epoch": 2.0629095674967233, "grad_norm": 1.8205486536026, "learning_rate": 4.829973035427406e-05, "loss": 0.5013, "num_input_tokens_seen": 2112584, "step": 7870 }, { "epoch": 2.0642201834862384, "grad_norm": 1.1589665412902832, "learning_rate": 4.8295582061403766e-05, "loss": 0.3259, "num_input_tokens_seen": 2114424, "step": 7875 }, { "epoch": 2.0655307994757535, "grad_norm": 0.7561389207839966, "learning_rate": 4.8291428892808396e-05, "loss": 0.4723, "num_input_tokens_seen": 2115672, "step": 7880 }, { "epoch": 2.0668414154652686, "grad_norm": 0.9455597400665283, "learning_rate": 4.8287270849357205e-05, "loss": 0.5828, "num_input_tokens_seen": 2117336, "step": 7885 }, { "epoch": 2.0681520314547837, "grad_norm": 0.5780449509620667, "learning_rate": 4.828310793192046e-05, "loss": 0.3274, "num_input_tokens_seen": 2118792, "step": 7890 }, { "epoch": 2.0694626474442988, "grad_norm": 1.1007542610168457, "learning_rate": 4.827894014136945e-05, "loss": 0.4806, "num_input_tokens_seen": 2120072, "step": 7895 }, { "epoch": 2.070773263433814, "grad_norm": 0.9203476905822754, "learning_rate": 4.827476747857651e-05, "loss": 0.3971, "num_input_tokens_seen": 2120952, "step": 7900 }, { "epoch": 2.072083879423329, "grad_norm": 1.5245367288589478, "learning_rate": 4.827058994441494e-05, "loss": 0.3878, "num_input_tokens_seen": 2121880, "step": 7905 }, { "epoch": 2.073394495412844, "grad_norm": 0.7637484669685364, "learning_rate": 4.82664075397591e-05, "loss": 0.5218, "num_input_tokens_seen": 2123128, "step": 7910 }, { "epoch": 2.074705111402359, "grad_norm": 0.661803662776947, "learning_rate": 4.826222026548437e-05, "loss": 0.4309, "num_input_tokens_seen": 2125016, "step": 7915 }, { "epoch": 2.0760157273918742, "grad_norm": 2.600520372390747, "learning_rate": 4.825802812246713e-05, "loss": 1.1946, "num_input_tokens_seen": 2126088, "step": 7920 }, { "epoch": 2.0773263433813893, "grad_norm": 0.7656107544898987, "learning_rate": 4.8253831111584794e-05, "loss": 0.3936, "num_input_tokens_seen": 2127576, "step": 7925 }, { "epoch": 2.0786369593709044, "grad_norm": 0.7555944323539734, "learning_rate": 4.8249629233715776e-05, "loss": 0.3559, "num_input_tokens_seen": 2128776, "step": 7930 }, { "epoch": 2.0799475753604195, "grad_norm": 1.0195581912994385, "learning_rate": 4.824542248973954e-05, "loss": 0.4379, "num_input_tokens_seen": 2130056, "step": 7935 }, { "epoch": 2.0812581913499346, "grad_norm": 0.5601898431777954, "learning_rate": 4.824121088053654e-05, "loss": 0.3939, "num_input_tokens_seen": 2131432, "step": 7940 }, { "epoch": 2.0825688073394497, "grad_norm": 1.3408119678497314, "learning_rate": 4.8236994406988255e-05, "loss": 0.3493, "num_input_tokens_seen": 2132584, "step": 7945 }, { "epoch": 2.083879423328965, "grad_norm": 0.5840147137641907, "learning_rate": 4.82327730699772e-05, "loss": 0.3543, "num_input_tokens_seen": 2134184, "step": 7950 }, { "epoch": 2.08519003931848, "grad_norm": 0.8780110478401184, "learning_rate": 4.822854687038688e-05, "loss": 0.493, "num_input_tokens_seen": 2135624, "step": 7955 }, { "epoch": 2.0865006553079946, "grad_norm": 1.0718653202056885, "learning_rate": 4.8224315809101836e-05, "loss": 0.4465, "num_input_tokens_seen": 2136712, "step": 7960 }, { "epoch": 2.0878112712975097, "grad_norm": 1.8258345127105713, "learning_rate": 4.822007988700762e-05, "loss": 0.6523, "num_input_tokens_seen": 2138264, "step": 7965 }, { "epoch": 2.0891218872870247, "grad_norm": 1.4603102207183838, "learning_rate": 4.821583910499081e-05, "loss": 0.4828, "num_input_tokens_seen": 2139368, "step": 7970 }, { "epoch": 2.09043250327654, "grad_norm": 1.5973820686340332, "learning_rate": 4.8211593463938985e-05, "loss": 0.5833, "num_input_tokens_seen": 2140456, "step": 7975 }, { "epoch": 2.091743119266055, "grad_norm": 0.688015878200531, "learning_rate": 4.8207342964740755e-05, "loss": 0.4934, "num_input_tokens_seen": 2142024, "step": 7980 }, { "epoch": 2.09305373525557, "grad_norm": 1.1461091041564941, "learning_rate": 4.8203087608285744e-05, "loss": 0.5381, "num_input_tokens_seen": 2143208, "step": 7985 }, { "epoch": 2.094364351245085, "grad_norm": 0.5470507144927979, "learning_rate": 4.819882739546458e-05, "loss": 0.2767, "num_input_tokens_seen": 2144712, "step": 7990 }, { "epoch": 2.0956749672346002, "grad_norm": 0.5757023096084595, "learning_rate": 4.8194562327168946e-05, "loss": 0.4888, "num_input_tokens_seen": 2146392, "step": 7995 }, { "epoch": 2.0969855832241153, "grad_norm": 0.5811610221862793, "learning_rate": 4.819029240429148e-05, "loss": 0.5002, "num_input_tokens_seen": 2148136, "step": 8000 }, { "epoch": 2.0982961992136304, "grad_norm": 0.6180412769317627, "learning_rate": 4.818601762772588e-05, "loss": 0.4633, "num_input_tokens_seen": 2149480, "step": 8005 }, { "epoch": 2.0996068152031455, "grad_norm": 0.7798029184341431, "learning_rate": 4.818173799836686e-05, "loss": 0.6001, "num_input_tokens_seen": 2150904, "step": 8010 }, { "epoch": 2.1009174311926606, "grad_norm": 0.6852022409439087, "learning_rate": 4.817745351711013e-05, "loss": 0.3398, "num_input_tokens_seen": 2152056, "step": 8015 }, { "epoch": 2.1022280471821757, "grad_norm": 3.575195074081421, "learning_rate": 4.817316418485243e-05, "loss": 0.5211, "num_input_tokens_seen": 2153080, "step": 8020 }, { "epoch": 2.103538663171691, "grad_norm": 1.1600300073623657, "learning_rate": 4.81688700024915e-05, "loss": 0.234, "num_input_tokens_seen": 2154312, "step": 8025 }, { "epoch": 2.104849279161206, "grad_norm": 0.6803120970726013, "learning_rate": 4.816457097092611e-05, "loss": 0.4559, "num_input_tokens_seen": 2155528, "step": 8030 }, { "epoch": 2.106159895150721, "grad_norm": 2.970127582550049, "learning_rate": 4.816026709105605e-05, "loss": 0.3233, "num_input_tokens_seen": 2156568, "step": 8035 }, { "epoch": 2.107470511140236, "grad_norm": 1.4367763996124268, "learning_rate": 4.81559583637821e-05, "loss": 0.5408, "num_input_tokens_seen": 2158056, "step": 8040 }, { "epoch": 2.108781127129751, "grad_norm": 0.8749766945838928, "learning_rate": 4.8151644790006077e-05, "loss": 0.3634, "num_input_tokens_seen": 2159160, "step": 8045 }, { "epoch": 2.1100917431192663, "grad_norm": 0.9572346806526184, "learning_rate": 4.8147326370630796e-05, "loss": 0.4301, "num_input_tokens_seen": 2160632, "step": 8050 }, { "epoch": 2.111402359108781, "grad_norm": 1.2107853889465332, "learning_rate": 4.814300310656011e-05, "loss": 0.4383, "num_input_tokens_seen": 2162152, "step": 8055 }, { "epoch": 2.112712975098296, "grad_norm": 1.0117788314819336, "learning_rate": 4.8138674998698854e-05, "loss": 0.5468, "num_input_tokens_seen": 2163896, "step": 8060 }, { "epoch": 2.114023591087811, "grad_norm": 0.8269990086555481, "learning_rate": 4.8134342047952906e-05, "loss": 0.3243, "num_input_tokens_seen": 2165240, "step": 8065 }, { "epoch": 2.1153342070773262, "grad_norm": 0.3805731236934662, "learning_rate": 4.8130004255229146e-05, "loss": 0.3082, "num_input_tokens_seen": 2166600, "step": 8070 }, { "epoch": 2.1166448230668413, "grad_norm": 0.7177361249923706, "learning_rate": 4.8125661621435445e-05, "loss": 0.4608, "num_input_tokens_seen": 2168056, "step": 8075 }, { "epoch": 2.1179554390563564, "grad_norm": 1.139682412147522, "learning_rate": 4.8121314147480736e-05, "loss": 0.3557, "num_input_tokens_seen": 2169480, "step": 8080 }, { "epoch": 2.1192660550458715, "grad_norm": 0.6454281806945801, "learning_rate": 4.811696183427493e-05, "loss": 0.5347, "num_input_tokens_seen": 2170632, "step": 8085 }, { "epoch": 2.1205766710353866, "grad_norm": 0.9357003569602966, "learning_rate": 4.811260468272895e-05, "loss": 0.4523, "num_input_tokens_seen": 2172072, "step": 8090 }, { "epoch": 2.1218872870249017, "grad_norm": 0.8719678521156311, "learning_rate": 4.810824269375475e-05, "loss": 0.6985, "num_input_tokens_seen": 2173416, "step": 8095 }, { "epoch": 2.123197903014417, "grad_norm": 2.73244309425354, "learning_rate": 4.810387586826527e-05, "loss": 0.4998, "num_input_tokens_seen": 2174536, "step": 8100 }, { "epoch": 2.124508519003932, "grad_norm": 0.8113266229629517, "learning_rate": 4.80995042071745e-05, "loss": 0.5552, "num_input_tokens_seen": 2176280, "step": 8105 }, { "epoch": 2.125819134993447, "grad_norm": 0.8399121165275574, "learning_rate": 4.809512771139741e-05, "loss": 0.3435, "num_input_tokens_seen": 2177928, "step": 8110 }, { "epoch": 2.127129750982962, "grad_norm": 1.3674434423446655, "learning_rate": 4.809074638184999e-05, "loss": 0.3492, "num_input_tokens_seen": 2179208, "step": 8115 }, { "epoch": 2.128440366972477, "grad_norm": 1.3159587383270264, "learning_rate": 4.808636021944927e-05, "loss": 0.3619, "num_input_tokens_seen": 2180520, "step": 8120 }, { "epoch": 2.1297509829619923, "grad_norm": 1.361639142036438, "learning_rate": 4.808196922511323e-05, "loss": 0.5426, "num_input_tokens_seen": 2181800, "step": 8125 }, { "epoch": 2.1310615989515074, "grad_norm": 1.5382716655731201, "learning_rate": 4.8077573399760914e-05, "loss": 0.4392, "num_input_tokens_seen": 2183016, "step": 8130 }, { "epoch": 2.1323722149410225, "grad_norm": 1.134087085723877, "learning_rate": 4.807317274431236e-05, "loss": 0.4409, "num_input_tokens_seen": 2184504, "step": 8135 }, { "epoch": 2.133682830930537, "grad_norm": 1.5371627807617188, "learning_rate": 4.806876725968862e-05, "loss": 0.3591, "num_input_tokens_seen": 2185544, "step": 8140 }, { "epoch": 2.134993446920052, "grad_norm": 0.3377857506275177, "learning_rate": 4.806435694681175e-05, "loss": 0.2765, "num_input_tokens_seen": 2187192, "step": 8145 }, { "epoch": 2.1363040629095673, "grad_norm": 1.5562925338745117, "learning_rate": 4.805994180660483e-05, "loss": 0.5136, "num_input_tokens_seen": 2188504, "step": 8150 }, { "epoch": 2.1376146788990824, "grad_norm": 1.6988061666488647, "learning_rate": 4.805552183999193e-05, "loss": 0.435, "num_input_tokens_seen": 2189928, "step": 8155 }, { "epoch": 2.1389252948885975, "grad_norm": 1.6491096019744873, "learning_rate": 4.8051097047898144e-05, "loss": 0.5862, "num_input_tokens_seen": 2191768, "step": 8160 }, { "epoch": 2.1402359108781126, "grad_norm": 1.2294342517852783, "learning_rate": 4.8046667431249585e-05, "loss": 0.3979, "num_input_tokens_seen": 2193080, "step": 8165 }, { "epoch": 2.1415465268676277, "grad_norm": 0.5613675117492676, "learning_rate": 4.804223299097334e-05, "loss": 0.6868, "num_input_tokens_seen": 2194104, "step": 8170 }, { "epoch": 2.142857142857143, "grad_norm": 1.2831707000732422, "learning_rate": 4.803779372799755e-05, "loss": 0.4601, "num_input_tokens_seen": 2197176, "step": 8175 }, { "epoch": 2.144167758846658, "grad_norm": 2.0360093116760254, "learning_rate": 4.803334964325134e-05, "loss": 0.3973, "num_input_tokens_seen": 2198232, "step": 8180 }, { "epoch": 2.145478374836173, "grad_norm": 0.620308518409729, "learning_rate": 4.802890073766485e-05, "loss": 0.478, "num_input_tokens_seen": 2199576, "step": 8185 }, { "epoch": 2.146788990825688, "grad_norm": 1.0775619745254517, "learning_rate": 4.802444701216923e-05, "loss": 0.7292, "num_input_tokens_seen": 2200728, "step": 8190 }, { "epoch": 2.148099606815203, "grad_norm": 0.6699600219726562, "learning_rate": 4.801998846769662e-05, "loss": 0.4275, "num_input_tokens_seen": 2202296, "step": 8195 }, { "epoch": 2.1494102228047183, "grad_norm": 1.1954923868179321, "learning_rate": 4.801552510518021e-05, "loss": 0.4994, "num_input_tokens_seen": 2203480, "step": 8200 }, { "epoch": 2.1507208387942334, "grad_norm": 0.7678935527801514, "learning_rate": 4.801105692555415e-05, "loss": 0.4089, "num_input_tokens_seen": 2204584, "step": 8205 }, { "epoch": 2.1520314547837485, "grad_norm": 3.2889564037323, "learning_rate": 4.800658392975366e-05, "loss": 0.4274, "num_input_tokens_seen": 2205464, "step": 8210 }, { "epoch": 2.1533420707732636, "grad_norm": 1.191123127937317, "learning_rate": 4.800210611871488e-05, "loss": 0.6732, "num_input_tokens_seen": 2206936, "step": 8215 }, { "epoch": 2.1546526867627787, "grad_norm": 0.7594806551933289, "learning_rate": 4.7997623493375054e-05, "loss": 0.5149, "num_input_tokens_seen": 2208120, "step": 8220 }, { "epoch": 2.1559633027522938, "grad_norm": 0.7626320719718933, "learning_rate": 4.7993136054672356e-05, "loss": 0.4776, "num_input_tokens_seen": 2209608, "step": 8225 }, { "epoch": 2.157273918741809, "grad_norm": 0.7398156523704529, "learning_rate": 4.798864380354601e-05, "loss": 0.2754, "num_input_tokens_seen": 2211080, "step": 8230 }, { "epoch": 2.1585845347313235, "grad_norm": 0.3356604278087616, "learning_rate": 4.798414674093624e-05, "loss": 0.5766, "num_input_tokens_seen": 2212984, "step": 8235 }, { "epoch": 2.1598951507208386, "grad_norm": 1.5446648597717285, "learning_rate": 4.7979644867784276e-05, "loss": 0.4363, "num_input_tokens_seen": 2214136, "step": 8240 }, { "epoch": 2.1612057667103537, "grad_norm": 4.8551812171936035, "learning_rate": 4.797513818503234e-05, "loss": 0.523, "num_input_tokens_seen": 2215080, "step": 8245 }, { "epoch": 2.162516382699869, "grad_norm": 1.1935569047927856, "learning_rate": 4.797062669362369e-05, "loss": 0.3114, "num_input_tokens_seen": 2216328, "step": 8250 }, { "epoch": 2.163826998689384, "grad_norm": 0.9565759897232056, "learning_rate": 4.796611039450255e-05, "loss": 0.5585, "num_input_tokens_seen": 2217608, "step": 8255 }, { "epoch": 2.165137614678899, "grad_norm": 0.7555602788925171, "learning_rate": 4.796158928861419e-05, "loss": 0.447, "num_input_tokens_seen": 2218712, "step": 8260 }, { "epoch": 2.166448230668414, "grad_norm": 1.4092086553573608, "learning_rate": 4.7957063376904875e-05, "loss": 0.3964, "num_input_tokens_seen": 2219880, "step": 8265 }, { "epoch": 2.167758846657929, "grad_norm": 1.2574660778045654, "learning_rate": 4.7952532660321854e-05, "loss": 0.543, "num_input_tokens_seen": 2221064, "step": 8270 }, { "epoch": 2.1690694626474443, "grad_norm": 2.992576837539673, "learning_rate": 4.794799713981341e-05, "loss": 0.4219, "num_input_tokens_seen": 2222440, "step": 8275 }, { "epoch": 2.1703800786369594, "grad_norm": 0.5478504300117493, "learning_rate": 4.7943456816328815e-05, "loss": 0.4101, "num_input_tokens_seen": 2223768, "step": 8280 }, { "epoch": 2.1716906946264745, "grad_norm": 0.8603192567825317, "learning_rate": 4.7938911690818347e-05, "loss": 0.4088, "num_input_tokens_seen": 2225320, "step": 8285 }, { "epoch": 2.1730013106159896, "grad_norm": 2.220376491546631, "learning_rate": 4.79343617642333e-05, "loss": 0.392, "num_input_tokens_seen": 2226792, "step": 8290 }, { "epoch": 2.1743119266055047, "grad_norm": 1.1760821342468262, "learning_rate": 4.792980703752597e-05, "loss": 0.4362, "num_input_tokens_seen": 2228248, "step": 8295 }, { "epoch": 2.1756225425950197, "grad_norm": 0.9497908353805542, "learning_rate": 4.792524751164964e-05, "loss": 0.3739, "num_input_tokens_seen": 2229640, "step": 8300 }, { "epoch": 2.176933158584535, "grad_norm": 0.9091820120811462, "learning_rate": 4.792068318755861e-05, "loss": 0.31, "num_input_tokens_seen": 2230936, "step": 8305 }, { "epoch": 2.17824377457405, "grad_norm": 0.5964632630348206, "learning_rate": 4.791611406620821e-05, "loss": 0.4359, "num_input_tokens_seen": 2232264, "step": 8310 }, { "epoch": 2.179554390563565, "grad_norm": 0.7622513771057129, "learning_rate": 4.7911540148554725e-05, "loss": 0.3685, "num_input_tokens_seen": 2233416, "step": 8315 }, { "epoch": 2.18086500655308, "grad_norm": 0.8429498672485352, "learning_rate": 4.790696143555546e-05, "loss": 0.3969, "num_input_tokens_seen": 2235032, "step": 8320 }, { "epoch": 2.1821756225425952, "grad_norm": 1.1077492237091064, "learning_rate": 4.790237792816877e-05, "loss": 0.7751, "num_input_tokens_seen": 2236440, "step": 8325 }, { "epoch": 2.18348623853211, "grad_norm": 0.5450628399848938, "learning_rate": 4.789778962735394e-05, "loss": 0.305, "num_input_tokens_seen": 2237912, "step": 8330 }, { "epoch": 2.184796854521625, "grad_norm": 4.004177093505859, "learning_rate": 4.78931965340713e-05, "loss": 0.4473, "num_input_tokens_seen": 2239144, "step": 8335 }, { "epoch": 2.18610747051114, "grad_norm": 1.566059947013855, "learning_rate": 4.78885986492822e-05, "loss": 0.5083, "num_input_tokens_seen": 2240840, "step": 8340 }, { "epoch": 2.187418086500655, "grad_norm": 1.0312576293945312, "learning_rate": 4.788399597394894e-05, "loss": 0.5069, "num_input_tokens_seen": 2242296, "step": 8345 }, { "epoch": 2.1887287024901703, "grad_norm": 0.8091779947280884, "learning_rate": 4.787938850903486e-05, "loss": 0.4492, "num_input_tokens_seen": 2243608, "step": 8350 }, { "epoch": 2.1900393184796854, "grad_norm": 1.9120004177093506, "learning_rate": 4.78747762555043e-05, "loss": 0.5428, "num_input_tokens_seen": 2244888, "step": 8355 }, { "epoch": 2.1913499344692005, "grad_norm": 0.9732306003570557, "learning_rate": 4.7870159214322594e-05, "loss": 0.4036, "num_input_tokens_seen": 2246136, "step": 8360 }, { "epoch": 2.1926605504587156, "grad_norm": 2.4127728939056396, "learning_rate": 4.7865537386456084e-05, "loss": 0.4833, "num_input_tokens_seen": 2247480, "step": 8365 }, { "epoch": 2.1939711664482306, "grad_norm": 0.7270041108131409, "learning_rate": 4.7860910772872104e-05, "loss": 0.3178, "num_input_tokens_seen": 2248456, "step": 8370 }, { "epoch": 2.1952817824377457, "grad_norm": 1.3355820178985596, "learning_rate": 4.7856279374539e-05, "loss": 0.7852, "num_input_tokens_seen": 2249608, "step": 8375 }, { "epoch": 2.196592398427261, "grad_norm": 3.3613951206207275, "learning_rate": 4.785164319242612e-05, "loss": 0.4276, "num_input_tokens_seen": 2250648, "step": 8380 }, { "epoch": 2.197903014416776, "grad_norm": 0.6110643744468689, "learning_rate": 4.78470022275038e-05, "loss": 0.6113, "num_input_tokens_seen": 2252168, "step": 8385 }, { "epoch": 2.199213630406291, "grad_norm": 1.1338820457458496, "learning_rate": 4.784235648074339e-05, "loss": 0.2813, "num_input_tokens_seen": 2253464, "step": 8390 }, { "epoch": 2.200524246395806, "grad_norm": 1.0611249208450317, "learning_rate": 4.783770595311724e-05, "loss": 0.3557, "num_input_tokens_seen": 2254712, "step": 8395 }, { "epoch": 2.2018348623853212, "grad_norm": 0.4728388488292694, "learning_rate": 4.78330506455987e-05, "loss": 0.2774, "num_input_tokens_seen": 2255800, "step": 8400 }, { "epoch": 2.2031454783748363, "grad_norm": 0.42215877771377563, "learning_rate": 4.78283905591621e-05, "loss": 0.4672, "num_input_tokens_seen": 2257208, "step": 8405 }, { "epoch": 2.2044560943643514, "grad_norm": 1.1588200330734253, "learning_rate": 4.78237256947828e-05, "loss": 0.4243, "num_input_tokens_seen": 2258600, "step": 8410 }, { "epoch": 2.2057667103538665, "grad_norm": 0.7151453495025635, "learning_rate": 4.7819056053437165e-05, "loss": 0.5211, "num_input_tokens_seen": 2259960, "step": 8415 }, { "epoch": 2.2070773263433816, "grad_norm": 2.1966800689697266, "learning_rate": 4.7814381636102515e-05, "loss": 0.5132, "num_input_tokens_seen": 2260888, "step": 8420 }, { "epoch": 2.2083879423328963, "grad_norm": 0.8171327114105225, "learning_rate": 4.780970244375721e-05, "loss": 0.5039, "num_input_tokens_seen": 2262376, "step": 8425 }, { "epoch": 2.2096985583224114, "grad_norm": 0.6097788214683533, "learning_rate": 4.780501847738059e-05, "loss": 0.4211, "num_input_tokens_seen": 2263432, "step": 8430 }, { "epoch": 2.2110091743119265, "grad_norm": 0.7413409948348999, "learning_rate": 4.7800329737953016e-05, "loss": 0.419, "num_input_tokens_seen": 2264808, "step": 8435 }, { "epoch": 2.2123197903014415, "grad_norm": 0.8850277662277222, "learning_rate": 4.779563622645582e-05, "loss": 0.4193, "num_input_tokens_seen": 2266248, "step": 8440 }, { "epoch": 2.2136304062909566, "grad_norm": 1.0269036293029785, "learning_rate": 4.779093794387135e-05, "loss": 0.4207, "num_input_tokens_seen": 2267416, "step": 8445 }, { "epoch": 2.2149410222804717, "grad_norm": 0.8551937937736511, "learning_rate": 4.778623489118295e-05, "loss": 0.4217, "num_input_tokens_seen": 2268440, "step": 8450 }, { "epoch": 2.216251638269987, "grad_norm": 3.1878063678741455, "learning_rate": 4.7781527069374957e-05, "loss": 0.4652, "num_input_tokens_seen": 2269800, "step": 8455 }, { "epoch": 2.217562254259502, "grad_norm": 0.7930054068565369, "learning_rate": 4.777681447943271e-05, "loss": 0.4561, "num_input_tokens_seen": 2271256, "step": 8460 }, { "epoch": 2.218872870249017, "grad_norm": 0.8992612957954407, "learning_rate": 4.777209712234255e-05, "loss": 0.3378, "num_input_tokens_seen": 2272728, "step": 8465 }, { "epoch": 2.220183486238532, "grad_norm": 0.7872147560119629, "learning_rate": 4.776737499909181e-05, "loss": 0.7519, "num_input_tokens_seen": 2274072, "step": 8470 }, { "epoch": 2.221494102228047, "grad_norm": 0.6284301280975342, "learning_rate": 4.776264811066882e-05, "loss": 0.4812, "num_input_tokens_seen": 2275560, "step": 8475 }, { "epoch": 2.2228047182175623, "grad_norm": 0.3827400505542755, "learning_rate": 4.7757916458062905e-05, "loss": 0.3472, "num_input_tokens_seen": 2276840, "step": 8480 }, { "epoch": 2.2241153342070774, "grad_norm": 1.0331089496612549, "learning_rate": 4.7753180042264404e-05, "loss": 0.592, "num_input_tokens_seen": 2277944, "step": 8485 }, { "epoch": 2.2254259501965925, "grad_norm": 2.4483041763305664, "learning_rate": 4.7748438864264636e-05, "loss": 0.334, "num_input_tokens_seen": 2278792, "step": 8490 }, { "epoch": 2.2267365661861076, "grad_norm": 1.6473430395126343, "learning_rate": 4.774369292505592e-05, "loss": 0.4016, "num_input_tokens_seen": 2280488, "step": 8495 }, { "epoch": 2.2280471821756227, "grad_norm": 0.6965255737304688, "learning_rate": 4.773894222563157e-05, "loss": 0.401, "num_input_tokens_seen": 2281640, "step": 8500 }, { "epoch": 2.229357798165138, "grad_norm": 0.897974967956543, "learning_rate": 4.7734186766985904e-05, "loss": 0.3659, "num_input_tokens_seen": 2282744, "step": 8505 }, { "epoch": 2.230668414154653, "grad_norm": 0.8043052554130554, "learning_rate": 4.772942655011423e-05, "loss": 0.2331, "num_input_tokens_seen": 2283848, "step": 8510 }, { "epoch": 2.231979030144168, "grad_norm": 0.8080447316169739, "learning_rate": 4.772466157601285e-05, "loss": 0.4527, "num_input_tokens_seen": 2285000, "step": 8515 }, { "epoch": 2.2332896461336826, "grad_norm": 0.9964724779129028, "learning_rate": 4.771989184567907e-05, "loss": 0.3433, "num_input_tokens_seen": 2286056, "step": 8520 }, { "epoch": 2.2346002621231977, "grad_norm": 1.5948034524917603, "learning_rate": 4.7715117360111185e-05, "loss": 0.6894, "num_input_tokens_seen": 2287320, "step": 8525 }, { "epoch": 2.235910878112713, "grad_norm": 1.2823892831802368, "learning_rate": 4.771033812030849e-05, "loss": 0.4944, "num_input_tokens_seen": 2288280, "step": 8530 }, { "epoch": 2.237221494102228, "grad_norm": 7.546942710876465, "learning_rate": 4.7705554127271254e-05, "loss": 0.4757, "num_input_tokens_seen": 2289576, "step": 8535 }, { "epoch": 2.238532110091743, "grad_norm": 0.7780389785766602, "learning_rate": 4.770076538200078e-05, "loss": 0.318, "num_input_tokens_seen": 2291032, "step": 8540 }, { "epoch": 2.239842726081258, "grad_norm": 0.5902000665664673, "learning_rate": 4.769597188549932e-05, "loss": 0.441, "num_input_tokens_seen": 2292824, "step": 8545 }, { "epoch": 2.241153342070773, "grad_norm": 0.6218693256378174, "learning_rate": 4.769117363877018e-05, "loss": 0.417, "num_input_tokens_seen": 2294264, "step": 8550 }, { "epoch": 2.2424639580602883, "grad_norm": 0.7651249766349792, "learning_rate": 4.76863706428176e-05, "loss": 0.528, "num_input_tokens_seen": 2295960, "step": 8555 }, { "epoch": 2.2437745740498034, "grad_norm": 1.1986885070800781, "learning_rate": 4.7681562898646836e-05, "loss": 0.4909, "num_input_tokens_seen": 2297336, "step": 8560 }, { "epoch": 2.2450851900393185, "grad_norm": 1.1260629892349243, "learning_rate": 4.767675040726416e-05, "loss": 0.3623, "num_input_tokens_seen": 2298568, "step": 8565 }, { "epoch": 2.2463958060288336, "grad_norm": 0.4500885605812073, "learning_rate": 4.76719331696768e-05, "loss": 0.4489, "num_input_tokens_seen": 2299576, "step": 8570 }, { "epoch": 2.2477064220183487, "grad_norm": 1.2212257385253906, "learning_rate": 4.7667111186893e-05, "loss": 0.533, "num_input_tokens_seen": 2301288, "step": 8575 }, { "epoch": 2.249017038007864, "grad_norm": 0.3955844044685364, "learning_rate": 4.766228445992199e-05, "loss": 0.3693, "num_input_tokens_seen": 2302664, "step": 8580 }, { "epoch": 2.250327653997379, "grad_norm": 0.8682078123092651, "learning_rate": 4.7657452989774e-05, "loss": 0.3706, "num_input_tokens_seen": 2303960, "step": 8585 }, { "epoch": 2.251638269986894, "grad_norm": 0.7314596772193909, "learning_rate": 4.765261677746026e-05, "loss": 0.3902, "num_input_tokens_seen": 2305112, "step": 8590 }, { "epoch": 2.252948885976409, "grad_norm": 0.6795572638511658, "learning_rate": 4.764777582399296e-05, "loss": 0.3545, "num_input_tokens_seen": 2306600, "step": 8595 }, { "epoch": 2.254259501965924, "grad_norm": 2.291071653366089, "learning_rate": 4.764293013038531e-05, "loss": 0.4562, "num_input_tokens_seen": 2307928, "step": 8600 }, { "epoch": 2.255570117955439, "grad_norm": 0.6718518733978271, "learning_rate": 4.7638079697651514e-05, "loss": 0.5339, "num_input_tokens_seen": 2309144, "step": 8605 }, { "epoch": 2.2568807339449544, "grad_norm": 0.8995252847671509, "learning_rate": 4.763322452680675e-05, "loss": 0.3773, "num_input_tokens_seen": 2310296, "step": 8610 }, { "epoch": 2.258191349934469, "grad_norm": 0.5060803890228271, "learning_rate": 4.76283646188672e-05, "loss": 0.3368, "num_input_tokens_seen": 2311816, "step": 8615 }, { "epoch": 2.259501965923984, "grad_norm": 0.8340950608253479, "learning_rate": 4.7623499974850036e-05, "loss": 0.4195, "num_input_tokens_seen": 2313112, "step": 8620 }, { "epoch": 2.260812581913499, "grad_norm": 0.9068667888641357, "learning_rate": 4.7618630595773424e-05, "loss": 0.4356, "num_input_tokens_seen": 2314312, "step": 8625 }, { "epoch": 2.2621231979030143, "grad_norm": 2.7974045276641846, "learning_rate": 4.7613756482656505e-05, "loss": 0.6862, "num_input_tokens_seen": 2315432, "step": 8630 }, { "epoch": 2.2634338138925294, "grad_norm": 0.5369064211845398, "learning_rate": 4.760887763651942e-05, "loss": 0.3856, "num_input_tokens_seen": 2317128, "step": 8635 }, { "epoch": 2.2647444298820445, "grad_norm": 0.9934021234512329, "learning_rate": 4.7603994058383315e-05, "loss": 0.4458, "num_input_tokens_seen": 2318904, "step": 8640 }, { "epoch": 2.2660550458715596, "grad_norm": 0.7386358976364136, "learning_rate": 4.759910574927032e-05, "loss": 0.5311, "num_input_tokens_seen": 2321384, "step": 8645 }, { "epoch": 2.2673656618610747, "grad_norm": 0.9608674645423889, "learning_rate": 4.759421271020353e-05, "loss": 0.3724, "num_input_tokens_seen": 2322776, "step": 8650 }, { "epoch": 2.26867627785059, "grad_norm": 0.7007020711898804, "learning_rate": 4.758931494220707e-05, "loss": 0.3432, "num_input_tokens_seen": 2324376, "step": 8655 }, { "epoch": 2.269986893840105, "grad_norm": 0.6575281023979187, "learning_rate": 4.7584412446306e-05, "loss": 0.4732, "num_input_tokens_seen": 2325688, "step": 8660 }, { "epoch": 2.27129750982962, "grad_norm": 0.7603439092636108, "learning_rate": 4.7579505223526446e-05, "loss": 0.6317, "num_input_tokens_seen": 2326808, "step": 8665 }, { "epoch": 2.272608125819135, "grad_norm": 0.48440414667129517, "learning_rate": 4.757459327489546e-05, "loss": 0.2489, "num_input_tokens_seen": 2328344, "step": 8670 }, { "epoch": 2.27391874180865, "grad_norm": 1.1126253604888916, "learning_rate": 4.756967660144111e-05, "loss": 0.3033, "num_input_tokens_seen": 2329560, "step": 8675 }, { "epoch": 2.2752293577981653, "grad_norm": 0.5945998430252075, "learning_rate": 4.756475520419244e-05, "loss": 0.5526, "num_input_tokens_seen": 2330888, "step": 8680 }, { "epoch": 2.2765399737876804, "grad_norm": 2.250704050064087, "learning_rate": 4.7559829084179494e-05, "loss": 0.6434, "num_input_tokens_seen": 2332232, "step": 8685 }, { "epoch": 2.2778505897771955, "grad_norm": 0.6627398133277893, "learning_rate": 4.755489824243331e-05, "loss": 0.7811, "num_input_tokens_seen": 2333512, "step": 8690 }, { "epoch": 2.2791612057667106, "grad_norm": 0.4201726019382477, "learning_rate": 4.754996267998589e-05, "loss": 0.4466, "num_input_tokens_seen": 2334792, "step": 8695 }, { "epoch": 2.280471821756225, "grad_norm": 0.9147955775260925, "learning_rate": 4.754502239787024e-05, "loss": 0.3722, "num_input_tokens_seen": 2336040, "step": 8700 }, { "epoch": 2.2817824377457407, "grad_norm": 2.3934547901153564, "learning_rate": 4.7540077397120365e-05, "loss": 0.5347, "num_input_tokens_seen": 2336936, "step": 8705 }, { "epoch": 2.2830930537352554, "grad_norm": 0.6833648085594177, "learning_rate": 4.7535127678771233e-05, "loss": 0.526, "num_input_tokens_seen": 2338664, "step": 8710 }, { "epoch": 2.2844036697247705, "grad_norm": 0.9030707478523254, "learning_rate": 4.7530173243858814e-05, "loss": 0.4645, "num_input_tokens_seen": 2339816, "step": 8715 }, { "epoch": 2.2857142857142856, "grad_norm": 1.0262186527252197, "learning_rate": 4.7525214093420065e-05, "loss": 0.4219, "num_input_tokens_seen": 2341128, "step": 8720 }, { "epoch": 2.2870249017038007, "grad_norm": 0.9317236542701721, "learning_rate": 4.7520250228492934e-05, "loss": 0.3463, "num_input_tokens_seen": 2342184, "step": 8725 }, { "epoch": 2.288335517693316, "grad_norm": 0.7503220438957214, "learning_rate": 4.751528165011633e-05, "loss": 0.3402, "num_input_tokens_seen": 2343960, "step": 8730 }, { "epoch": 2.289646133682831, "grad_norm": 0.8217684030532837, "learning_rate": 4.7510308359330194e-05, "loss": 0.3896, "num_input_tokens_seen": 2345208, "step": 8735 }, { "epoch": 2.290956749672346, "grad_norm": 0.9560125470161438, "learning_rate": 4.7505330357175406e-05, "loss": 0.5047, "num_input_tokens_seen": 2346568, "step": 8740 }, { "epoch": 2.292267365661861, "grad_norm": 1.372713565826416, "learning_rate": 4.750034764469386e-05, "loss": 0.4926, "num_input_tokens_seen": 2347656, "step": 8745 }, { "epoch": 2.293577981651376, "grad_norm": 0.5214519500732422, "learning_rate": 4.749536022292842e-05, "loss": 0.3984, "num_input_tokens_seen": 2349112, "step": 8750 }, { "epoch": 2.2948885976408913, "grad_norm": 0.6465591192245483, "learning_rate": 4.7490368092922964e-05, "loss": 0.3111, "num_input_tokens_seen": 2350536, "step": 8755 }, { "epoch": 2.2961992136304064, "grad_norm": 0.7180967926979065, "learning_rate": 4.748537125572233e-05, "loss": 0.7997, "num_input_tokens_seen": 2352472, "step": 8760 }, { "epoch": 2.2975098296199215, "grad_norm": 2.98630428314209, "learning_rate": 4.748036971237234e-05, "loss": 0.6602, "num_input_tokens_seen": 2353784, "step": 8765 }, { "epoch": 2.2988204456094365, "grad_norm": 2.8553547859191895, "learning_rate": 4.747536346391981e-05, "loss": 0.3767, "num_input_tokens_seen": 2355240, "step": 8770 }, { "epoch": 2.3001310615989516, "grad_norm": 0.7701948881149292, "learning_rate": 4.747035251141254e-05, "loss": 0.365, "num_input_tokens_seen": 2356312, "step": 8775 }, { "epoch": 2.3014416775884667, "grad_norm": 1.5245234966278076, "learning_rate": 4.746533685589932e-05, "loss": 0.3594, "num_input_tokens_seen": 2357624, "step": 8780 }, { "epoch": 2.302752293577982, "grad_norm": 0.9659894704818726, "learning_rate": 4.746031649842991e-05, "loss": 0.4074, "num_input_tokens_seen": 2358984, "step": 8785 }, { "epoch": 2.304062909567497, "grad_norm": 0.8815451860427856, "learning_rate": 4.745529144005507e-05, "loss": 0.3368, "num_input_tokens_seen": 2360056, "step": 8790 }, { "epoch": 2.3053735255570116, "grad_norm": 1.05913245677948, "learning_rate": 4.745026168182654e-05, "loss": 0.6964, "num_input_tokens_seen": 2363800, "step": 8795 }, { "epoch": 2.3066841415465267, "grad_norm": 0.5870732665061951, "learning_rate": 4.744522722479701e-05, "loss": 0.4086, "num_input_tokens_seen": 2365224, "step": 8800 }, { "epoch": 2.3079947575360418, "grad_norm": 0.7687622904777527, "learning_rate": 4.7440188070020217e-05, "loss": 0.6208, "num_input_tokens_seen": 2366840, "step": 8805 }, { "epoch": 2.309305373525557, "grad_norm": 0.7129191756248474, "learning_rate": 4.7435144218550834e-05, "loss": 0.3667, "num_input_tokens_seen": 2368424, "step": 8810 }, { "epoch": 2.310615989515072, "grad_norm": 1.1346328258514404, "learning_rate": 4.7430095671444526e-05, "loss": 0.9835, "num_input_tokens_seen": 2370024, "step": 8815 }, { "epoch": 2.311926605504587, "grad_norm": 1.8663386106491089, "learning_rate": 4.742504242975795e-05, "loss": 0.3051, "num_input_tokens_seen": 2371448, "step": 8820 }, { "epoch": 2.313237221494102, "grad_norm": 1.108597993850708, "learning_rate": 4.7419984494548744e-05, "loss": 0.5556, "num_input_tokens_seen": 2373080, "step": 8825 }, { "epoch": 2.3145478374836173, "grad_norm": 2.8733224868774414, "learning_rate": 4.7414921866875524e-05, "loss": 0.5539, "num_input_tokens_seen": 2374088, "step": 8830 }, { "epoch": 2.3158584534731324, "grad_norm": 0.7134377360343933, "learning_rate": 4.740985454779788e-05, "loss": 0.3995, "num_input_tokens_seen": 2375848, "step": 8835 }, { "epoch": 2.3171690694626474, "grad_norm": 1.0640044212341309, "learning_rate": 4.740478253837641e-05, "loss": 0.3596, "num_input_tokens_seen": 2376920, "step": 8840 }, { "epoch": 2.3184796854521625, "grad_norm": 1.6816201210021973, "learning_rate": 4.739970583967266e-05, "loss": 0.5796, "num_input_tokens_seen": 2378584, "step": 8845 }, { "epoch": 2.3197903014416776, "grad_norm": 1.0498420000076294, "learning_rate": 4.739462445274919e-05, "loss": 0.6549, "num_input_tokens_seen": 2379848, "step": 8850 }, { "epoch": 2.3211009174311927, "grad_norm": 0.8621310591697693, "learning_rate": 4.738953837866951e-05, "loss": 0.4265, "num_input_tokens_seen": 2381048, "step": 8855 }, { "epoch": 2.322411533420708, "grad_norm": 0.8602033853530884, "learning_rate": 4.738444761849813e-05, "loss": 0.364, "num_input_tokens_seen": 2382232, "step": 8860 }, { "epoch": 2.323722149410223, "grad_norm": 1.025470495223999, "learning_rate": 4.7379352173300553e-05, "loss": 0.542, "num_input_tokens_seen": 2383368, "step": 8865 }, { "epoch": 2.325032765399738, "grad_norm": 2.3524584770202637, "learning_rate": 4.737425204414322e-05, "loss": 0.4018, "num_input_tokens_seen": 2385032, "step": 8870 }, { "epoch": 2.326343381389253, "grad_norm": 3.662275791168213, "learning_rate": 4.73691472320936e-05, "loss": 0.2713, "num_input_tokens_seen": 2388264, "step": 8875 }, { "epoch": 2.327653997378768, "grad_norm": 1.5980836153030396, "learning_rate": 4.736403773822012e-05, "loss": 0.481, "num_input_tokens_seen": 2389464, "step": 8880 }, { "epoch": 2.3289646133682833, "grad_norm": 1.2280826568603516, "learning_rate": 4.735892356359217e-05, "loss": 0.4141, "num_input_tokens_seen": 2390456, "step": 8885 }, { "epoch": 2.330275229357798, "grad_norm": 1.4143821001052856, "learning_rate": 4.735380470928017e-05, "loss": 0.5926, "num_input_tokens_seen": 2392168, "step": 8890 }, { "epoch": 2.331585845347313, "grad_norm": 0.538390576839447, "learning_rate": 4.7348681176355456e-05, "loss": 0.3358, "num_input_tokens_seen": 2393464, "step": 8895 }, { "epoch": 2.332896461336828, "grad_norm": 0.6208410859107971, "learning_rate": 4.734355296589039e-05, "loss": 0.6081, "num_input_tokens_seen": 2395064, "step": 8900 }, { "epoch": 2.3342070773263432, "grad_norm": 0.45204707980155945, "learning_rate": 4.73384200789583e-05, "loss": 0.4603, "num_input_tokens_seen": 2396232, "step": 8905 }, { "epoch": 2.3355176933158583, "grad_norm": 1.6654459238052368, "learning_rate": 4.733328251663348e-05, "loss": 0.4144, "num_input_tokens_seen": 2397544, "step": 8910 }, { "epoch": 2.3368283093053734, "grad_norm": 0.8880354166030884, "learning_rate": 4.732814027999122e-05, "loss": 0.3295, "num_input_tokens_seen": 2398824, "step": 8915 }, { "epoch": 2.3381389252948885, "grad_norm": 1.0009287595748901, "learning_rate": 4.7322993370107775e-05, "loss": 0.688, "num_input_tokens_seen": 2399816, "step": 8920 }, { "epoch": 2.3394495412844036, "grad_norm": 1.0262188911437988, "learning_rate": 4.7317841788060394e-05, "loss": 0.4514, "num_input_tokens_seen": 2400952, "step": 8925 }, { "epoch": 2.3407601572739187, "grad_norm": 0.9422820806503296, "learning_rate": 4.731268553492729e-05, "loss": 0.3584, "num_input_tokens_seen": 2402120, "step": 8930 }, { "epoch": 2.342070773263434, "grad_norm": 0.45888906717300415, "learning_rate": 4.730752461178766e-05, "loss": 0.3618, "num_input_tokens_seen": 2403432, "step": 8935 }, { "epoch": 2.343381389252949, "grad_norm": 0.5397550463676453, "learning_rate": 4.730235901972166e-05, "loss": 0.42, "num_input_tokens_seen": 2405160, "step": 8940 }, { "epoch": 2.344692005242464, "grad_norm": 0.4947550892829895, "learning_rate": 4.729718875981046e-05, "loss": 0.3122, "num_input_tokens_seen": 2406536, "step": 8945 }, { "epoch": 2.346002621231979, "grad_norm": 1.1918772459030151, "learning_rate": 4.729201383313618e-05, "loss": 0.5108, "num_input_tokens_seen": 2407992, "step": 8950 }, { "epoch": 2.347313237221494, "grad_norm": 1.1888861656188965, "learning_rate": 4.7286834240781925e-05, "loss": 0.2159, "num_input_tokens_seen": 2409016, "step": 8955 }, { "epoch": 2.3486238532110093, "grad_norm": 2.496642827987671, "learning_rate": 4.728164998383177e-05, "loss": 0.4938, "num_input_tokens_seen": 2410472, "step": 8960 }, { "epoch": 2.3499344692005244, "grad_norm": 0.9024525880813599, "learning_rate": 4.727646106337077e-05, "loss": 0.3576, "num_input_tokens_seen": 2411496, "step": 8965 }, { "epoch": 2.3512450851900395, "grad_norm": 1.2637792825698853, "learning_rate": 4.727126748048496e-05, "loss": 0.3967, "num_input_tokens_seen": 2412696, "step": 8970 }, { "epoch": 2.352555701179554, "grad_norm": 0.8905666470527649, "learning_rate": 4.726606923626135e-05, "loss": 0.3474, "num_input_tokens_seen": 2414008, "step": 8975 }, { "epoch": 2.3538663171690697, "grad_norm": 1.0399484634399414, "learning_rate": 4.726086633178792e-05, "loss": 0.3403, "num_input_tokens_seen": 2415272, "step": 8980 }, { "epoch": 2.3551769331585843, "grad_norm": 1.4865294694900513, "learning_rate": 4.7255658768153635e-05, "loss": 0.5076, "num_input_tokens_seen": 2417240, "step": 8985 }, { "epoch": 2.3564875491480994, "grad_norm": 2.6090729236602783, "learning_rate": 4.725044654644842e-05, "loss": 0.5528, "num_input_tokens_seen": 2418376, "step": 8990 }, { "epoch": 2.3577981651376145, "grad_norm": 0.6258254647254944, "learning_rate": 4.724522966776319e-05, "loss": 0.4037, "num_input_tokens_seen": 2420120, "step": 8995 }, { "epoch": 2.3591087811271296, "grad_norm": 1.9731671810150146, "learning_rate": 4.7240008133189836e-05, "loss": 0.4713, "num_input_tokens_seen": 2421160, "step": 9000 }, { "epoch": 2.3604193971166447, "grad_norm": 0.8087565898895264, "learning_rate": 4.723478194382119e-05, "loss": 0.3417, "num_input_tokens_seen": 2423048, "step": 9005 }, { "epoch": 2.36173001310616, "grad_norm": 1.2754124402999878, "learning_rate": 4.722955110075112e-05, "loss": 0.3438, "num_input_tokens_seen": 2424504, "step": 9010 }, { "epoch": 2.363040629095675, "grad_norm": 2.1395795345306396, "learning_rate": 4.7224315605074403e-05, "loss": 0.3637, "num_input_tokens_seen": 2425416, "step": 9015 }, { "epoch": 2.36435124508519, "grad_norm": 1.2118288278579712, "learning_rate": 4.7219075457886836e-05, "loss": 0.391, "num_input_tokens_seen": 2427176, "step": 9020 }, { "epoch": 2.365661861074705, "grad_norm": 0.9681674242019653, "learning_rate": 4.721383066028516e-05, "loss": 0.5139, "num_input_tokens_seen": 2428360, "step": 9025 }, { "epoch": 2.36697247706422, "grad_norm": 0.7735799551010132, "learning_rate": 4.720858121336712e-05, "loss": 0.3412, "num_input_tokens_seen": 2429688, "step": 9030 }, { "epoch": 2.3682830930537353, "grad_norm": 1.5500423908233643, "learning_rate": 4.72033271182314e-05, "loss": 0.4646, "num_input_tokens_seen": 2430712, "step": 9035 }, { "epoch": 2.3695937090432504, "grad_norm": 1.296911597251892, "learning_rate": 4.7198068375977675e-05, "loss": 0.4652, "num_input_tokens_seen": 2432072, "step": 9040 }, { "epoch": 2.3709043250327655, "grad_norm": 0.5930824279785156, "learning_rate": 4.71928049877066e-05, "loss": 0.3483, "num_input_tokens_seen": 2433288, "step": 9045 }, { "epoch": 2.3722149410222806, "grad_norm": 1.7538156509399414, "learning_rate": 4.718753695451978e-05, "loss": 0.3713, "num_input_tokens_seen": 2434616, "step": 9050 }, { "epoch": 2.3735255570117957, "grad_norm": 0.966044545173645, "learning_rate": 4.718226427751982e-05, "loss": 0.5425, "num_input_tokens_seen": 2435688, "step": 9055 }, { "epoch": 2.374836173001311, "grad_norm": 0.8035081028938293, "learning_rate": 4.7176986957810275e-05, "loss": 0.4783, "num_input_tokens_seen": 2437416, "step": 9060 }, { "epoch": 2.376146788990826, "grad_norm": 0.9015527963638306, "learning_rate": 4.717170499649568e-05, "loss": 0.4786, "num_input_tokens_seen": 2438280, "step": 9065 }, { "epoch": 2.3774574049803405, "grad_norm": 1.5886253118515015, "learning_rate": 4.7166418394681526e-05, "loss": 0.5231, "num_input_tokens_seen": 2439784, "step": 9070 }, { "epoch": 2.378768020969856, "grad_norm": 1.2012112140655518, "learning_rate": 4.716112715347432e-05, "loss": 0.3702, "num_input_tokens_seen": 2441160, "step": 9075 }, { "epoch": 2.3800786369593707, "grad_norm": 0.6584828495979309, "learning_rate": 4.715583127398148e-05, "loss": 0.4083, "num_input_tokens_seen": 2442328, "step": 9080 }, { "epoch": 2.381389252948886, "grad_norm": 1.211094856262207, "learning_rate": 4.715053075731144e-05, "loss": 0.2532, "num_input_tokens_seen": 2443432, "step": 9085 }, { "epoch": 2.382699868938401, "grad_norm": 0.9386574029922485, "learning_rate": 4.714522560457359e-05, "loss": 0.4425, "num_input_tokens_seen": 2444488, "step": 9090 }, { "epoch": 2.384010484927916, "grad_norm": 1.012618899345398, "learning_rate": 4.713991581687827e-05, "loss": 0.5882, "num_input_tokens_seen": 2445592, "step": 9095 }, { "epoch": 2.385321100917431, "grad_norm": 0.7707961201667786, "learning_rate": 4.7134601395336836e-05, "loss": 0.3934, "num_input_tokens_seen": 2447256, "step": 9100 }, { "epoch": 2.386631716906946, "grad_norm": 0.8171420693397522, "learning_rate": 4.712928234106158e-05, "loss": 0.3388, "num_input_tokens_seen": 2448472, "step": 9105 }, { "epoch": 2.3879423328964613, "grad_norm": 0.6093966960906982, "learning_rate": 4.712395865516576e-05, "loss": 0.4308, "num_input_tokens_seen": 2449880, "step": 9110 }, { "epoch": 2.3892529488859764, "grad_norm": 1.0854917764663696, "learning_rate": 4.711863033876362e-05, "loss": 0.3791, "num_input_tokens_seen": 2450936, "step": 9115 }, { "epoch": 2.3905635648754915, "grad_norm": 0.8276032209396362, "learning_rate": 4.7113297392970365e-05, "loss": 0.4649, "num_input_tokens_seen": 2452488, "step": 9120 }, { "epoch": 2.3918741808650066, "grad_norm": 1.132495641708374, "learning_rate": 4.710795981890218e-05, "loss": 0.4529, "num_input_tokens_seen": 2453960, "step": 9125 }, { "epoch": 2.3931847968545217, "grad_norm": 0.4544866681098938, "learning_rate": 4.7102617617676203e-05, "loss": 0.4767, "num_input_tokens_seen": 2455208, "step": 9130 }, { "epoch": 2.3944954128440368, "grad_norm": 1.2117393016815186, "learning_rate": 4.7097270790410555e-05, "loss": 0.3005, "num_input_tokens_seen": 2456152, "step": 9135 }, { "epoch": 2.395806028833552, "grad_norm": 0.7712363600730896, "learning_rate": 4.709191933822431e-05, "loss": 0.4895, "num_input_tokens_seen": 2457528, "step": 9140 }, { "epoch": 2.397116644823067, "grad_norm": 0.3548509180545807, "learning_rate": 4.7086563262237514e-05, "loss": 0.2617, "num_input_tokens_seen": 2458792, "step": 9145 }, { "epoch": 2.398427260812582, "grad_norm": 0.7140108942985535, "learning_rate": 4.7081202563571194e-05, "loss": 0.3779, "num_input_tokens_seen": 2459928, "step": 9150 }, { "epoch": 2.399737876802097, "grad_norm": 1.5850672721862793, "learning_rate": 4.707583724334732e-05, "loss": 0.4875, "num_input_tokens_seen": 2461304, "step": 9155 }, { "epoch": 2.4010484927916123, "grad_norm": 0.951117217540741, "learning_rate": 4.707046730268887e-05, "loss": 0.557, "num_input_tokens_seen": 2462440, "step": 9160 }, { "epoch": 2.402359108781127, "grad_norm": 0.5609908699989319, "learning_rate": 4.706509274271974e-05, "loss": 0.5576, "num_input_tokens_seen": 2463864, "step": 9165 }, { "epoch": 2.4036697247706424, "grad_norm": 2.169494152069092, "learning_rate": 4.705971356456483e-05, "loss": 0.6575, "num_input_tokens_seen": 2465688, "step": 9170 }, { "epoch": 2.404980340760157, "grad_norm": 1.0646811723709106, "learning_rate": 4.7054329769349984e-05, "loss": 0.4454, "num_input_tokens_seen": 2467320, "step": 9175 }, { "epoch": 2.406290956749672, "grad_norm": 0.38475462794303894, "learning_rate": 4.704894135820202e-05, "loss": 0.314, "num_input_tokens_seen": 2468680, "step": 9180 }, { "epoch": 2.4076015727391873, "grad_norm": 0.7083476781845093, "learning_rate": 4.704354833224873e-05, "loss": 0.6147, "num_input_tokens_seen": 2470088, "step": 9185 }, { "epoch": 2.4089121887287024, "grad_norm": 1.688554048538208, "learning_rate": 4.7038150692618864e-05, "loss": 0.4943, "num_input_tokens_seen": 2471416, "step": 9190 }, { "epoch": 2.4102228047182175, "grad_norm": 0.7750020623207092, "learning_rate": 4.703274844044214e-05, "loss": 0.4638, "num_input_tokens_seen": 2472456, "step": 9195 }, { "epoch": 2.4115334207077326, "grad_norm": 1.2817070484161377, "learning_rate": 4.702734157684924e-05, "loss": 0.5072, "num_input_tokens_seen": 2473400, "step": 9200 }, { "epoch": 2.4128440366972477, "grad_norm": 0.5082091093063354, "learning_rate": 4.7021930102971796e-05, "loss": 0.3212, "num_input_tokens_seen": 2474616, "step": 9205 }, { "epoch": 2.4141546526867628, "grad_norm": 0.8567540049552917, "learning_rate": 4.701651401994244e-05, "loss": 0.4179, "num_input_tokens_seen": 2475752, "step": 9210 }, { "epoch": 2.415465268676278, "grad_norm": 0.749630868434906, "learning_rate": 4.701109332889474e-05, "loss": 0.3746, "num_input_tokens_seen": 2477064, "step": 9215 }, { "epoch": 2.416775884665793, "grad_norm": 0.9330957531929016, "learning_rate": 4.700566803096324e-05, "loss": 0.4322, "num_input_tokens_seen": 2478472, "step": 9220 }, { "epoch": 2.418086500655308, "grad_norm": 0.4857800304889679, "learning_rate": 4.700023812728344e-05, "loss": 0.2794, "num_input_tokens_seen": 2480408, "step": 9225 }, { "epoch": 2.419397116644823, "grad_norm": 0.9840391874313354, "learning_rate": 4.699480361899181e-05, "loss": 0.5259, "num_input_tokens_seen": 2481704, "step": 9230 }, { "epoch": 2.4207077326343382, "grad_norm": 0.4369877576828003, "learning_rate": 4.6989364507225795e-05, "loss": 0.2434, "num_input_tokens_seen": 2482776, "step": 9235 }, { "epoch": 2.4220183486238533, "grad_norm": 1.642298698425293, "learning_rate": 4.698392079312378e-05, "loss": 0.4032, "num_input_tokens_seen": 2484376, "step": 9240 }, { "epoch": 2.4233289646133684, "grad_norm": 0.6500970125198364, "learning_rate": 4.697847247782513e-05, "loss": 0.4674, "num_input_tokens_seen": 2486088, "step": 9245 }, { "epoch": 2.4246395806028835, "grad_norm": 0.5750691294670105, "learning_rate": 4.6973019562470166e-05, "loss": 0.5645, "num_input_tokens_seen": 2487320, "step": 9250 }, { "epoch": 2.4259501965923986, "grad_norm": 1.202391505241394, "learning_rate": 4.6967562048200176e-05, "loss": 0.4771, "num_input_tokens_seen": 2488424, "step": 9255 }, { "epoch": 2.4272608125819133, "grad_norm": 1.067195177078247, "learning_rate": 4.69620999361574e-05, "loss": 0.3815, "num_input_tokens_seen": 2489480, "step": 9260 }, { "epoch": 2.4285714285714284, "grad_norm": 1.0230413675308228, "learning_rate": 4.695663322748506e-05, "loss": 0.5785, "num_input_tokens_seen": 2490872, "step": 9265 }, { "epoch": 2.4298820445609435, "grad_norm": 2.216106414794922, "learning_rate": 4.6951161923327324e-05, "loss": 0.3873, "num_input_tokens_seen": 2492152, "step": 9270 }, { "epoch": 2.4311926605504586, "grad_norm": 1.486968755722046, "learning_rate": 4.694568602482933e-05, "loss": 0.3834, "num_input_tokens_seen": 2493640, "step": 9275 }, { "epoch": 2.4325032765399737, "grad_norm": 0.48706895112991333, "learning_rate": 4.694020553313716e-05, "loss": 0.32, "num_input_tokens_seen": 2494792, "step": 9280 }, { "epoch": 2.4338138925294888, "grad_norm": 0.715540885925293, "learning_rate": 4.693472044939788e-05, "loss": 0.3531, "num_input_tokens_seen": 2496200, "step": 9285 }, { "epoch": 2.435124508519004, "grad_norm": 0.695114254951477, "learning_rate": 4.692923077475952e-05, "loss": 0.4364, "num_input_tokens_seen": 2497544, "step": 9290 }, { "epoch": 2.436435124508519, "grad_norm": 0.828644871711731, "learning_rate": 4.6923736510371054e-05, "loss": 0.5125, "num_input_tokens_seen": 2498856, "step": 9295 }, { "epoch": 2.437745740498034, "grad_norm": 0.7951806783676147, "learning_rate": 4.69182376573824e-05, "loss": 0.4684, "num_input_tokens_seen": 2500136, "step": 9300 }, { "epoch": 2.439056356487549, "grad_norm": 1.5468088388442993, "learning_rate": 4.6912734216944486e-05, "loss": 0.4265, "num_input_tokens_seen": 2501304, "step": 9305 }, { "epoch": 2.4403669724770642, "grad_norm": 1.0518008470535278, "learning_rate": 4.6907226190209165e-05, "loss": 0.3987, "num_input_tokens_seen": 2502792, "step": 9310 }, { "epoch": 2.4416775884665793, "grad_norm": 0.9397377967834473, "learning_rate": 4.6901713578329255e-05, "loss": 0.3068, "num_input_tokens_seen": 2503944, "step": 9315 }, { "epoch": 2.4429882044560944, "grad_norm": 0.677692711353302, "learning_rate": 4.689619638245853e-05, "loss": 0.5133, "num_input_tokens_seen": 2505960, "step": 9320 }, { "epoch": 2.4442988204456095, "grad_norm": 1.2240961790084839, "learning_rate": 4.6890674603751746e-05, "loss": 0.4371, "num_input_tokens_seen": 2507480, "step": 9325 }, { "epoch": 2.4456094364351246, "grad_norm": 1.4485275745391846, "learning_rate": 4.688514824336459e-05, "loss": 0.5481, "num_input_tokens_seen": 2509928, "step": 9330 }, { "epoch": 2.4469200524246397, "grad_norm": 1.3446344137191772, "learning_rate": 4.687961730245371e-05, "loss": 0.5494, "num_input_tokens_seen": 2511496, "step": 9335 }, { "epoch": 2.448230668414155, "grad_norm": 0.8576650619506836, "learning_rate": 4.687408178217674e-05, "loss": 0.3625, "num_input_tokens_seen": 2512984, "step": 9340 }, { "epoch": 2.44954128440367, "grad_norm": 0.634669303894043, "learning_rate": 4.6868541683692244e-05, "loss": 0.4191, "num_input_tokens_seen": 2514536, "step": 9345 }, { "epoch": 2.450851900393185, "grad_norm": 0.6707493662834167, "learning_rate": 4.6862997008159774e-05, "loss": 0.3635, "num_input_tokens_seen": 2515656, "step": 9350 }, { "epoch": 2.4521625163826997, "grad_norm": 0.9634779095649719, "learning_rate": 4.685744775673979e-05, "loss": 0.5123, "num_input_tokens_seen": 2517096, "step": 9355 }, { "epoch": 2.4534731323722148, "grad_norm": 0.938165009021759, "learning_rate": 4.685189393059377e-05, "loss": 0.5252, "num_input_tokens_seen": 2518408, "step": 9360 }, { "epoch": 2.45478374836173, "grad_norm": 1.8038935661315918, "learning_rate": 4.68463355308841e-05, "loss": 0.3548, "num_input_tokens_seen": 2519512, "step": 9365 }, { "epoch": 2.456094364351245, "grad_norm": 0.8191643953323364, "learning_rate": 4.684077255877415e-05, "loss": 0.3831, "num_input_tokens_seen": 2520808, "step": 9370 }, { "epoch": 2.45740498034076, "grad_norm": 1.0681109428405762, "learning_rate": 4.6835205015428246e-05, "loss": 0.5973, "num_input_tokens_seen": 2522120, "step": 9375 }, { "epoch": 2.458715596330275, "grad_norm": 0.6765396595001221, "learning_rate": 4.682963290201166e-05, "loss": 0.4643, "num_input_tokens_seen": 2523384, "step": 9380 }, { "epoch": 2.4600262123197902, "grad_norm": 0.6007382869720459, "learning_rate": 4.682405621969062e-05, "loss": 0.4616, "num_input_tokens_seen": 2525144, "step": 9385 }, { "epoch": 2.4613368283093053, "grad_norm": 1.266541838645935, "learning_rate": 4.681847496963233e-05, "loss": 0.4311, "num_input_tokens_seen": 2526168, "step": 9390 }, { "epoch": 2.4626474442988204, "grad_norm": 0.9054426550865173, "learning_rate": 4.6812889153004925e-05, "loss": 0.5995, "num_input_tokens_seen": 2527560, "step": 9395 }, { "epoch": 2.4639580602883355, "grad_norm": 2.0182793140411377, "learning_rate": 4.6807298770977516e-05, "loss": 0.4142, "num_input_tokens_seen": 2529176, "step": 9400 }, { "epoch": 2.4652686762778506, "grad_norm": 1.7828328609466553, "learning_rate": 4.6801703824720146e-05, "loss": 0.5115, "num_input_tokens_seen": 2530088, "step": 9405 }, { "epoch": 2.4665792922673657, "grad_norm": 0.9613820314407349, "learning_rate": 4.679610431540384e-05, "loss": 0.4067, "num_input_tokens_seen": 2531272, "step": 9410 }, { "epoch": 2.467889908256881, "grad_norm": 0.40499863028526306, "learning_rate": 4.679050024420056e-05, "loss": 0.5001, "num_input_tokens_seen": 2532696, "step": 9415 }, { "epoch": 2.469200524246396, "grad_norm": 1.6695805788040161, "learning_rate": 4.678489161228323e-05, "loss": 0.4322, "num_input_tokens_seen": 2533640, "step": 9420 }, { "epoch": 2.470511140235911, "grad_norm": 0.9405520558357239, "learning_rate": 4.677927842082572e-05, "loss": 0.3752, "num_input_tokens_seen": 2534856, "step": 9425 }, { "epoch": 2.471821756225426, "grad_norm": 0.9273393750190735, "learning_rate": 4.6773660671002885e-05, "loss": 0.3508, "num_input_tokens_seen": 2536152, "step": 9430 }, { "epoch": 2.473132372214941, "grad_norm": 0.5438724756240845, "learning_rate": 4.676803836399048e-05, "loss": 0.5753, "num_input_tokens_seen": 2537560, "step": 9435 }, { "epoch": 2.474442988204456, "grad_norm": 0.7576953172683716, "learning_rate": 4.6762411500965265e-05, "loss": 0.3363, "num_input_tokens_seen": 2538792, "step": 9440 }, { "epoch": 2.4757536041939714, "grad_norm": 0.8391208648681641, "learning_rate": 4.675678008310492e-05, "loss": 0.5878, "num_input_tokens_seen": 2539896, "step": 9445 }, { "epoch": 2.477064220183486, "grad_norm": 1.0493543148040771, "learning_rate": 4.6751144111588094e-05, "loss": 0.4452, "num_input_tokens_seen": 2541032, "step": 9450 }, { "epoch": 2.478374836173001, "grad_norm": 0.8029890656471252, "learning_rate": 4.6745503587594396e-05, "loss": 0.3479, "num_input_tokens_seen": 2542120, "step": 9455 }, { "epoch": 2.4796854521625162, "grad_norm": 0.348682165145874, "learning_rate": 4.6739858512304366e-05, "loss": 0.5795, "num_input_tokens_seen": 2543096, "step": 9460 }, { "epoch": 2.4809960681520313, "grad_norm": 1.0933228731155396, "learning_rate": 4.673420888689952e-05, "loss": 0.3413, "num_input_tokens_seen": 2544328, "step": 9465 }, { "epoch": 2.4823066841415464, "grad_norm": 2.1066360473632812, "learning_rate": 4.67285547125623e-05, "loss": 0.462, "num_input_tokens_seen": 2545448, "step": 9470 }, { "epoch": 2.4836173001310615, "grad_norm": 1.2098709344863892, "learning_rate": 4.6722895990476126e-05, "loss": 0.4009, "num_input_tokens_seen": 2546584, "step": 9475 }, { "epoch": 2.4849279161205766, "grad_norm": 1.1483012437820435, "learning_rate": 4.671723272182535e-05, "loss": 0.3655, "num_input_tokens_seen": 2547832, "step": 9480 }, { "epoch": 2.4862385321100917, "grad_norm": 1.3184795379638672, "learning_rate": 4.67115649077953e-05, "loss": 0.5863, "num_input_tokens_seen": 2549016, "step": 9485 }, { "epoch": 2.487549148099607, "grad_norm": 1.0963811874389648, "learning_rate": 4.670589254957222e-05, "loss": 0.3346, "num_input_tokens_seen": 2550392, "step": 9490 }, { "epoch": 2.488859764089122, "grad_norm": 0.7686243057250977, "learning_rate": 4.6700215648343336e-05, "loss": 0.3656, "num_input_tokens_seen": 2551992, "step": 9495 }, { "epoch": 2.490170380078637, "grad_norm": 0.47060176730155945, "learning_rate": 4.669453420529682e-05, "loss": 0.3441, "num_input_tokens_seen": 2553736, "step": 9500 }, { "epoch": 2.491480996068152, "grad_norm": 2.373819589614868, "learning_rate": 4.668884822162177e-05, "loss": 0.5035, "num_input_tokens_seen": 2554968, "step": 9505 }, { "epoch": 2.492791612057667, "grad_norm": 0.8240037560462952, "learning_rate": 4.668315769850827e-05, "loss": 0.3261, "num_input_tokens_seen": 2556120, "step": 9510 }, { "epoch": 2.4941022280471823, "grad_norm": 0.687270998954773, "learning_rate": 4.667746263714733e-05, "loss": 0.4259, "num_input_tokens_seen": 2557560, "step": 9515 }, { "epoch": 2.4954128440366974, "grad_norm": 0.9331863522529602, "learning_rate": 4.667176303873091e-05, "loss": 0.4729, "num_input_tokens_seen": 2558888, "step": 9520 }, { "epoch": 2.4967234600262125, "grad_norm": 0.7603746056556702, "learning_rate": 4.666605890445193e-05, "loss": 0.3144, "num_input_tokens_seen": 2560024, "step": 9525 }, { "epoch": 2.4980340760157276, "grad_norm": 2.0291335582733154, "learning_rate": 4.6660350235504265e-05, "loss": 0.4811, "num_input_tokens_seen": 2561192, "step": 9530 }, { "epoch": 2.4993446920052422, "grad_norm": 1.0198439359664917, "learning_rate": 4.6654637033082715e-05, "loss": 0.4181, "num_input_tokens_seen": 2562648, "step": 9535 }, { "epoch": 2.5006553079947578, "grad_norm": 0.8620941042900085, "learning_rate": 4.6648919298383055e-05, "loss": 0.687, "num_input_tokens_seen": 2563624, "step": 9540 }, { "epoch": 2.5006553079947578, "eval_loss": 0.5038492679595947, "eval_runtime": 16.6572, "eval_samples_per_second": 50.909, "eval_steps_per_second": 25.455, "num_input_tokens_seen": 2563624, "step": 9540 }, { "epoch": 2.5019659239842724, "grad_norm": 0.7289140820503235, "learning_rate": 4.664319703260199e-05, "loss": 0.5718, "num_input_tokens_seen": 2565160, "step": 9545 }, { "epoch": 2.5032765399737875, "grad_norm": 0.804179310798645, "learning_rate": 4.6637470236937186e-05, "loss": 0.5495, "num_input_tokens_seen": 2566232, "step": 9550 }, { "epoch": 2.5045871559633026, "grad_norm": 0.9942927956581116, "learning_rate": 4.663173891258724e-05, "loss": 0.2937, "num_input_tokens_seen": 2567400, "step": 9555 }, { "epoch": 2.5058977719528177, "grad_norm": 0.47411879897117615, "learning_rate": 4.662600306075172e-05, "loss": 0.3025, "num_input_tokens_seen": 2568840, "step": 9560 }, { "epoch": 2.507208387942333, "grad_norm": 0.6688489317893982, "learning_rate": 4.662026268263113e-05, "loss": 0.6666, "num_input_tokens_seen": 2570392, "step": 9565 }, { "epoch": 2.508519003931848, "grad_norm": 0.7938612103462219, "learning_rate": 4.6614517779426917e-05, "loss": 0.5651, "num_input_tokens_seen": 2572072, "step": 9570 }, { "epoch": 2.509829619921363, "grad_norm": 1.4165377616882324, "learning_rate": 4.660876835234148e-05, "loss": 0.6555, "num_input_tokens_seen": 2573096, "step": 9575 }, { "epoch": 2.511140235910878, "grad_norm": 2.599578857421875, "learning_rate": 4.660301440257816e-05, "loss": 0.5819, "num_input_tokens_seen": 2574424, "step": 9580 }, { "epoch": 2.512450851900393, "grad_norm": 1.4428513050079346, "learning_rate": 4.659725593134125e-05, "loss": 0.3222, "num_input_tokens_seen": 2575688, "step": 9585 }, { "epoch": 2.5137614678899083, "grad_norm": 1.0225791931152344, "learning_rate": 4.6591492939836e-05, "loss": 0.4992, "num_input_tokens_seen": 2577240, "step": 9590 }, { "epoch": 2.5150720838794234, "grad_norm": 1.1921937465667725, "learning_rate": 4.6585725429268585e-05, "loss": 0.5198, "num_input_tokens_seen": 2578760, "step": 9595 }, { "epoch": 2.5163826998689385, "grad_norm": 0.5060738921165466, "learning_rate": 4.657995340084613e-05, "loss": 0.4577, "num_input_tokens_seen": 2580328, "step": 9600 }, { "epoch": 2.5176933158584536, "grad_norm": 1.380427360534668, "learning_rate": 4.657417685577671e-05, "loss": 0.5546, "num_input_tokens_seen": 2581400, "step": 9605 }, { "epoch": 2.5190039318479687, "grad_norm": 0.7122948169708252, "learning_rate": 4.656839579526936e-05, "loss": 0.4363, "num_input_tokens_seen": 2582568, "step": 9610 }, { "epoch": 2.5203145478374838, "grad_norm": 0.9503388404846191, "learning_rate": 4.6562610220534035e-05, "loss": 0.4143, "num_input_tokens_seen": 2583816, "step": 9615 }, { "epoch": 2.5216251638269984, "grad_norm": 1.0455378293991089, "learning_rate": 4.6556820132781646e-05, "loss": 0.3581, "num_input_tokens_seen": 2585128, "step": 9620 }, { "epoch": 2.522935779816514, "grad_norm": 5.219253063201904, "learning_rate": 4.6551025533224056e-05, "loss": 0.3222, "num_input_tokens_seen": 2586120, "step": 9625 }, { "epoch": 2.5242463958060286, "grad_norm": 1.254984736442566, "learning_rate": 4.654522642307405e-05, "loss": 0.4654, "num_input_tokens_seen": 2587352, "step": 9630 }, { "epoch": 2.525557011795544, "grad_norm": 0.5088933110237122, "learning_rate": 4.653942280354539e-05, "loss": 0.3947, "num_input_tokens_seen": 2588840, "step": 9635 }, { "epoch": 2.526867627785059, "grad_norm": 1.079953670501709, "learning_rate": 4.6533614675852755e-05, "loss": 0.3552, "num_input_tokens_seen": 2590232, "step": 9640 }, { "epoch": 2.528178243774574, "grad_norm": 0.9030967950820923, "learning_rate": 4.652780204121177e-05, "loss": 0.5103, "num_input_tokens_seen": 2591400, "step": 9645 }, { "epoch": 2.529488859764089, "grad_norm": 0.8732820153236389, "learning_rate": 4.652198490083901e-05, "loss": 0.3792, "num_input_tokens_seen": 2592488, "step": 9650 }, { "epoch": 2.530799475753604, "grad_norm": 0.6662008762359619, "learning_rate": 4.6516163255952004e-05, "loss": 0.3581, "num_input_tokens_seen": 2593544, "step": 9655 }, { "epoch": 2.532110091743119, "grad_norm": 1.432569146156311, "learning_rate": 4.651033710776921e-05, "loss": 0.3684, "num_input_tokens_seen": 2594696, "step": 9660 }, { "epoch": 2.5334207077326343, "grad_norm": 3.9627373218536377, "learning_rate": 4.6504506457510025e-05, "loss": 0.54, "num_input_tokens_seen": 2595864, "step": 9665 }, { "epoch": 2.5347313237221494, "grad_norm": 1.3510109186172485, "learning_rate": 4.649867130639479e-05, "loss": 0.6789, "num_input_tokens_seen": 2597080, "step": 9670 }, { "epoch": 2.5360419397116645, "grad_norm": 0.9601966738700867, "learning_rate": 4.64928316556448e-05, "loss": 0.3675, "num_input_tokens_seen": 2598200, "step": 9675 }, { "epoch": 2.5373525557011796, "grad_norm": 1.0932148694992065, "learning_rate": 4.648698750648228e-05, "loss": 0.4732, "num_input_tokens_seen": 2599512, "step": 9680 }, { "epoch": 2.5386631716906947, "grad_norm": 0.6759830713272095, "learning_rate": 4.64811388601304e-05, "loss": 0.448, "num_input_tokens_seen": 2600984, "step": 9685 }, { "epoch": 2.5399737876802098, "grad_norm": 1.896186351776123, "learning_rate": 4.647528571781328e-05, "loss": 0.4674, "num_input_tokens_seen": 2602024, "step": 9690 }, { "epoch": 2.541284403669725, "grad_norm": 0.9210941791534424, "learning_rate": 4.646942808075596e-05, "loss": 0.6476, "num_input_tokens_seen": 2603848, "step": 9695 }, { "epoch": 2.54259501965924, "grad_norm": 0.6519945859909058, "learning_rate": 4.646356595018444e-05, "loss": 0.301, "num_input_tokens_seen": 2605384, "step": 9700 }, { "epoch": 2.543905635648755, "grad_norm": 1.1131277084350586, "learning_rate": 4.6457699327325655e-05, "loss": 0.5167, "num_input_tokens_seen": 2606600, "step": 9705 }, { "epoch": 2.54521625163827, "grad_norm": 0.9151872992515564, "learning_rate": 4.645182821340747e-05, "loss": 0.4387, "num_input_tokens_seen": 2608072, "step": 9710 }, { "epoch": 2.546526867627785, "grad_norm": 0.7752858996391296, "learning_rate": 4.644595260965872e-05, "loss": 0.3511, "num_input_tokens_seen": 2609400, "step": 9715 }, { "epoch": 2.5478374836173003, "grad_norm": 0.9080662727355957, "learning_rate": 4.644007251730913e-05, "loss": 0.5342, "num_input_tokens_seen": 2610648, "step": 9720 }, { "epoch": 2.549148099606815, "grad_norm": 0.5518311858177185, "learning_rate": 4.643418793758941e-05, "loss": 0.4696, "num_input_tokens_seen": 2612392, "step": 9725 }, { "epoch": 2.5504587155963305, "grad_norm": 0.7866399884223938, "learning_rate": 4.642829887173119e-05, "loss": 0.4153, "num_input_tokens_seen": 2614088, "step": 9730 }, { "epoch": 2.551769331585845, "grad_norm": 0.8948689699172974, "learning_rate": 4.642240532096705e-05, "loss": 0.4009, "num_input_tokens_seen": 2615256, "step": 9735 }, { "epoch": 2.5530799475753603, "grad_norm": 2.7729647159576416, "learning_rate": 4.641650728653049e-05, "loss": 0.5348, "num_input_tokens_seen": 2616568, "step": 9740 }, { "epoch": 2.5543905635648754, "grad_norm": 0.4956773817539215, "learning_rate": 4.641060476965595e-05, "loss": 0.2747, "num_input_tokens_seen": 2617896, "step": 9745 }, { "epoch": 2.5557011795543905, "grad_norm": 0.6980568766593933, "learning_rate": 4.640469777157883e-05, "loss": 0.438, "num_input_tokens_seen": 2618968, "step": 9750 }, { "epoch": 2.5570117955439056, "grad_norm": 1.477591633796692, "learning_rate": 4.6398786293535455e-05, "loss": 0.656, "num_input_tokens_seen": 2619928, "step": 9755 }, { "epoch": 2.5583224115334207, "grad_norm": 0.8293538689613342, "learning_rate": 4.639287033676308e-05, "loss": 0.4139, "num_input_tokens_seen": 2621048, "step": 9760 }, { "epoch": 2.5596330275229358, "grad_norm": 0.8800724148750305, "learning_rate": 4.638694990249991e-05, "loss": 0.3919, "num_input_tokens_seen": 2622152, "step": 9765 }, { "epoch": 2.560943643512451, "grad_norm": 0.614510715007782, "learning_rate": 4.638102499198508e-05, "loss": 0.4369, "num_input_tokens_seen": 2624104, "step": 9770 }, { "epoch": 2.562254259501966, "grad_norm": 0.5661119222640991, "learning_rate": 4.637509560645865e-05, "loss": 0.488, "num_input_tokens_seen": 2625416, "step": 9775 }, { "epoch": 2.563564875491481, "grad_norm": 0.6907255053520203, "learning_rate": 4.636916174716166e-05, "loss": 0.7861, "num_input_tokens_seen": 2626600, "step": 9780 }, { "epoch": 2.564875491480996, "grad_norm": 0.8253029584884644, "learning_rate": 4.636322341533603e-05, "loss": 0.5441, "num_input_tokens_seen": 2627864, "step": 9785 }, { "epoch": 2.5661861074705112, "grad_norm": 0.8723334074020386, "learning_rate": 4.6357280612224656e-05, "loss": 0.8065, "num_input_tokens_seen": 2628984, "step": 9790 }, { "epoch": 2.5674967234600263, "grad_norm": 0.8949697613716125, "learning_rate": 4.635133333907135e-05, "loss": 0.4043, "num_input_tokens_seen": 2630376, "step": 9795 }, { "epoch": 2.5688073394495414, "grad_norm": 2.929939031600952, "learning_rate": 4.634538159712088e-05, "loss": 0.2764, "num_input_tokens_seen": 2631448, "step": 9800 }, { "epoch": 2.5701179554390565, "grad_norm": 0.8427737355232239, "learning_rate": 4.6339425387618906e-05, "loss": 0.3164, "num_input_tokens_seen": 2632664, "step": 9805 }, { "epoch": 2.571428571428571, "grad_norm": 0.7121486067771912, "learning_rate": 4.633346471181207e-05, "loss": 0.4077, "num_input_tokens_seen": 2633896, "step": 9810 }, { "epoch": 2.5727391874180867, "grad_norm": 0.967987596988678, "learning_rate": 4.632749957094794e-05, "loss": 0.418, "num_input_tokens_seen": 2635416, "step": 9815 }, { "epoch": 2.5740498034076014, "grad_norm": 1.3039953708648682, "learning_rate": 4.6321529966275e-05, "loss": 0.399, "num_input_tokens_seen": 2636632, "step": 9820 }, { "epoch": 2.575360419397117, "grad_norm": 1.1095337867736816, "learning_rate": 4.6315555899042686e-05, "loss": 0.6166, "num_input_tokens_seen": 2637672, "step": 9825 }, { "epoch": 2.5766710353866316, "grad_norm": 0.9945749640464783, "learning_rate": 4.630957737050134e-05, "loss": 0.5278, "num_input_tokens_seen": 2639064, "step": 9830 }, { "epoch": 2.5779816513761467, "grad_norm": 1.319035530090332, "learning_rate": 4.630359438190227e-05, "loss": 0.4315, "num_input_tokens_seen": 2640184, "step": 9835 }, { "epoch": 2.5792922673656618, "grad_norm": 1.128167748451233, "learning_rate": 4.629760693449772e-05, "loss": 0.4984, "num_input_tokens_seen": 2641400, "step": 9840 }, { "epoch": 2.580602883355177, "grad_norm": 0.594484806060791, "learning_rate": 4.6291615029540826e-05, "loss": 0.4323, "num_input_tokens_seen": 2642792, "step": 9845 }, { "epoch": 2.581913499344692, "grad_norm": 0.8767303228378296, "learning_rate": 4.62856186682857e-05, "loss": 0.3967, "num_input_tokens_seen": 2643992, "step": 9850 }, { "epoch": 2.583224115334207, "grad_norm": 1.8526558876037598, "learning_rate": 4.627961785198736e-05, "loss": 0.4231, "num_input_tokens_seen": 2644808, "step": 9855 }, { "epoch": 2.584534731323722, "grad_norm": 1.7484591007232666, "learning_rate": 4.627361258190178e-05, "loss": 0.4033, "num_input_tokens_seen": 2646168, "step": 9860 }, { "epoch": 2.5858453473132372, "grad_norm": 1.0366355180740356, "learning_rate": 4.626760285928584e-05, "loss": 0.4481, "num_input_tokens_seen": 2647512, "step": 9865 }, { "epoch": 2.5871559633027523, "grad_norm": 1.4488345384597778, "learning_rate": 4.626158868539737e-05, "loss": 0.2824, "num_input_tokens_seen": 2648584, "step": 9870 }, { "epoch": 2.5884665792922674, "grad_norm": 3.6443517208099365, "learning_rate": 4.625557006149512e-05, "loss": 0.4849, "num_input_tokens_seen": 2650120, "step": 9875 }, { "epoch": 2.5897771952817825, "grad_norm": 0.6085895299911499, "learning_rate": 4.624954698883879e-05, "loss": 0.5821, "num_input_tokens_seen": 2651624, "step": 9880 }, { "epoch": 2.5910878112712976, "grad_norm": 1.0562587976455688, "learning_rate": 4.624351946868899e-05, "loss": 0.5351, "num_input_tokens_seen": 2652888, "step": 9885 }, { "epoch": 2.5923984272608127, "grad_norm": 1.5085564851760864, "learning_rate": 4.6237487502307265e-05, "loss": 0.4201, "num_input_tokens_seen": 2654184, "step": 9890 }, { "epoch": 2.593709043250328, "grad_norm": 1.4866795539855957, "learning_rate": 4.623145109095611e-05, "loss": 0.3689, "num_input_tokens_seen": 2655832, "step": 9895 }, { "epoch": 2.595019659239843, "grad_norm": 1.8416800498962402, "learning_rate": 4.622541023589893e-05, "loss": 0.6369, "num_input_tokens_seen": 2657288, "step": 9900 }, { "epoch": 2.5963302752293576, "grad_norm": 0.9054094552993774, "learning_rate": 4.621936493840005e-05, "loss": 0.6522, "num_input_tokens_seen": 2658760, "step": 9905 }, { "epoch": 2.597640891218873, "grad_norm": 5.882595062255859, "learning_rate": 4.621331519972476e-05, "loss": 0.794, "num_input_tokens_seen": 2659928, "step": 9910 }, { "epoch": 2.5989515072083877, "grad_norm": 0.671150267124176, "learning_rate": 4.6207261021139246e-05, "loss": 0.3716, "num_input_tokens_seen": 2661160, "step": 9915 }, { "epoch": 2.6002621231979033, "grad_norm": 1.147092342376709, "learning_rate": 4.620120240391065e-05, "loss": 0.3892, "num_input_tokens_seen": 2662328, "step": 9920 }, { "epoch": 2.601572739187418, "grad_norm": 0.8344385027885437, "learning_rate": 4.6195139349307024e-05, "loss": 0.3152, "num_input_tokens_seen": 2663672, "step": 9925 }, { "epoch": 2.602883355176933, "grad_norm": 0.9184563755989075, "learning_rate": 4.6189071858597355e-05, "loss": 0.4307, "num_input_tokens_seen": 2665000, "step": 9930 }, { "epoch": 2.604193971166448, "grad_norm": 0.6267595887184143, "learning_rate": 4.6182999933051554e-05, "loss": 0.3951, "num_input_tokens_seen": 2666632, "step": 9935 }, { "epoch": 2.6055045871559632, "grad_norm": 0.8355779051780701, "learning_rate": 4.617692357394047e-05, "loss": 0.3654, "num_input_tokens_seen": 2668392, "step": 9940 }, { "epoch": 2.6068152031454783, "grad_norm": 1.2300312519073486, "learning_rate": 4.617084278253587e-05, "loss": 0.3819, "num_input_tokens_seen": 2669736, "step": 9945 }, { "epoch": 2.6081258191349934, "grad_norm": 0.66218101978302, "learning_rate": 4.616475756011046e-05, "loss": 0.5105, "num_input_tokens_seen": 2671112, "step": 9950 }, { "epoch": 2.6094364351245085, "grad_norm": 0.5859202146530151, "learning_rate": 4.615866790793786e-05, "loss": 0.4392, "num_input_tokens_seen": 2672856, "step": 9955 }, { "epoch": 2.6107470511140236, "grad_norm": 0.7324390411376953, "learning_rate": 4.6152573827292636e-05, "loss": 0.5356, "num_input_tokens_seen": 2674008, "step": 9960 }, { "epoch": 2.6120576671035387, "grad_norm": 0.6828048229217529, "learning_rate": 4.6146475319450244e-05, "loss": 0.7617, "num_input_tokens_seen": 2675176, "step": 9965 }, { "epoch": 2.613368283093054, "grad_norm": 1.2162121534347534, "learning_rate": 4.6140372385687114e-05, "loss": 0.4436, "num_input_tokens_seen": 2676664, "step": 9970 }, { "epoch": 2.614678899082569, "grad_norm": 0.8908351063728333, "learning_rate": 4.6134265027280574e-05, "loss": 0.5452, "num_input_tokens_seen": 2678104, "step": 9975 }, { "epoch": 2.615989515072084, "grad_norm": 1.7291003465652466, "learning_rate": 4.6128153245508876e-05, "loss": 0.4717, "num_input_tokens_seen": 2679112, "step": 9980 }, { "epoch": 2.617300131061599, "grad_norm": 0.9272906184196472, "learning_rate": 4.612203704165121e-05, "loss": 0.3378, "num_input_tokens_seen": 2680600, "step": 9985 }, { "epoch": 2.618610747051114, "grad_norm": 1.4246816635131836, "learning_rate": 4.6115916416987686e-05, "loss": 0.6308, "num_input_tokens_seen": 2681672, "step": 9990 }, { "epoch": 2.6199213630406293, "grad_norm": 1.8993912935256958, "learning_rate": 4.610979137279935e-05, "loss": 0.3405, "num_input_tokens_seen": 2682792, "step": 9995 }, { "epoch": 2.621231979030144, "grad_norm": 2.5771448612213135, "learning_rate": 4.6103661910368146e-05, "loss": 0.4526, "num_input_tokens_seen": 2684088, "step": 10000 }, { "epoch": 2.6225425950196595, "grad_norm": 0.6719486117362976, "learning_rate": 4.6097528030976964e-05, "loss": 0.2707, "num_input_tokens_seen": 2685624, "step": 10005 }, { "epoch": 2.623853211009174, "grad_norm": 0.5062150955200195, "learning_rate": 4.609138973590963e-05, "loss": 0.512, "num_input_tokens_seen": 2686808, "step": 10010 }, { "epoch": 2.625163826998689, "grad_norm": 0.4827198088169098, "learning_rate": 4.6085247026450866e-05, "loss": 0.4329, "num_input_tokens_seen": 2687976, "step": 10015 }, { "epoch": 2.6264744429882043, "grad_norm": 1.2518573999404907, "learning_rate": 4.607909990388632e-05, "loss": 0.5016, "num_input_tokens_seen": 2689400, "step": 10020 }, { "epoch": 2.6277850589777194, "grad_norm": 0.801075279712677, "learning_rate": 4.607294836950261e-05, "loss": 0.4405, "num_input_tokens_seen": 2690664, "step": 10025 }, { "epoch": 2.6290956749672345, "grad_norm": 0.6064937710762024, "learning_rate": 4.606679242458719e-05, "loss": 0.4034, "num_input_tokens_seen": 2691848, "step": 10030 }, { "epoch": 2.6304062909567496, "grad_norm": 0.6487647294998169, "learning_rate": 4.606063207042853e-05, "loss": 0.3672, "num_input_tokens_seen": 2693128, "step": 10035 }, { "epoch": 2.6317169069462647, "grad_norm": 0.9830914735794067, "learning_rate": 4.6054467308315975e-05, "loss": 0.4426, "num_input_tokens_seen": 2694392, "step": 10040 }, { "epoch": 2.63302752293578, "grad_norm": 0.9056069850921631, "learning_rate": 4.604829813953978e-05, "loss": 0.5656, "num_input_tokens_seen": 2695560, "step": 10045 }, { "epoch": 2.634338138925295, "grad_norm": 1.1045368909835815, "learning_rate": 4.6042124565391155e-05, "loss": 0.4186, "num_input_tokens_seen": 2696872, "step": 10050 }, { "epoch": 2.63564875491481, "grad_norm": 3.0564255714416504, "learning_rate": 4.603594658716221e-05, "loss": 0.8341, "num_input_tokens_seen": 2698104, "step": 10055 }, { "epoch": 2.636959370904325, "grad_norm": 0.9066421389579773, "learning_rate": 4.6029764206146006e-05, "loss": 0.2625, "num_input_tokens_seen": 2699800, "step": 10060 }, { "epoch": 2.63826998689384, "grad_norm": 0.9739185571670532, "learning_rate": 4.602357742363649e-05, "loss": 0.2511, "num_input_tokens_seen": 2701400, "step": 10065 }, { "epoch": 2.6395806028833553, "grad_norm": 0.5248993039131165, "learning_rate": 4.601738624092853e-05, "loss": 0.4096, "num_input_tokens_seen": 2702744, "step": 10070 }, { "epoch": 2.6408912188728704, "grad_norm": 0.8596011400222778, "learning_rate": 4.601119065931796e-05, "loss": 0.2716, "num_input_tokens_seen": 2704296, "step": 10075 }, { "epoch": 2.6422018348623855, "grad_norm": 1.2339235544204712, "learning_rate": 4.600499068010148e-05, "loss": 0.4772, "num_input_tokens_seen": 2705992, "step": 10080 }, { "epoch": 2.6435124508519, "grad_norm": 0.8585870265960693, "learning_rate": 4.599878630457674e-05, "loss": 0.4907, "num_input_tokens_seen": 2707816, "step": 10085 }, { "epoch": 2.6448230668414157, "grad_norm": 1.2911244630813599, "learning_rate": 4.599257753404233e-05, "loss": 0.5018, "num_input_tokens_seen": 2709144, "step": 10090 }, { "epoch": 2.6461336828309303, "grad_norm": 1.2550843954086304, "learning_rate": 4.59863643697977e-05, "loss": 0.5308, "num_input_tokens_seen": 2710152, "step": 10095 }, { "epoch": 2.647444298820446, "grad_norm": 1.5561838150024414, "learning_rate": 4.5980146813143264e-05, "loss": 0.6531, "num_input_tokens_seen": 2711304, "step": 10100 }, { "epoch": 2.6487549148099605, "grad_norm": 2.1558480262756348, "learning_rate": 4.597392486538036e-05, "loss": 0.5709, "num_input_tokens_seen": 2712328, "step": 10105 }, { "epoch": 2.6500655307994756, "grad_norm": 0.6759009957313538, "learning_rate": 4.596769852781122e-05, "loss": 0.4862, "num_input_tokens_seen": 2714120, "step": 10110 }, { "epoch": 2.6513761467889907, "grad_norm": 0.8388283252716064, "learning_rate": 4.596146780173901e-05, "loss": 0.5375, "num_input_tokens_seen": 2715800, "step": 10115 }, { "epoch": 2.652686762778506, "grad_norm": 0.9842453598976135, "learning_rate": 4.59552326884678e-05, "loss": 0.544, "num_input_tokens_seen": 2717016, "step": 10120 }, { "epoch": 2.653997378768021, "grad_norm": 0.09416304528713226, "learning_rate": 4.5948993189302605e-05, "loss": 0.6001, "num_input_tokens_seen": 2719176, "step": 10125 }, { "epoch": 2.655307994757536, "grad_norm": 0.5972254276275635, "learning_rate": 4.5942749305549325e-05, "loss": 0.3859, "num_input_tokens_seen": 2720456, "step": 10130 }, { "epoch": 2.656618610747051, "grad_norm": 0.6657077670097351, "learning_rate": 4.593650103851481e-05, "loss": 0.3071, "num_input_tokens_seen": 2721576, "step": 10135 }, { "epoch": 2.657929226736566, "grad_norm": 2.6367228031158447, "learning_rate": 4.5930248389506794e-05, "loss": 0.7646, "num_input_tokens_seen": 2722680, "step": 10140 }, { "epoch": 2.6592398427260813, "grad_norm": 0.7363440990447998, "learning_rate": 4.592399135983396e-05, "loss": 0.3826, "num_input_tokens_seen": 2724104, "step": 10145 }, { "epoch": 2.6605504587155964, "grad_norm": 1.1682610511779785, "learning_rate": 4.591772995080589e-05, "loss": 0.3659, "num_input_tokens_seen": 2725432, "step": 10150 }, { "epoch": 2.6618610747051115, "grad_norm": 0.5655229687690735, "learning_rate": 4.591146416373308e-05, "loss": 0.5388, "num_input_tokens_seen": 2726712, "step": 10155 }, { "epoch": 2.6631716906946266, "grad_norm": 0.6535264849662781, "learning_rate": 4.590519399992695e-05, "loss": 0.4906, "num_input_tokens_seen": 2727880, "step": 10160 }, { "epoch": 2.6644823066841417, "grad_norm": 1.3587167263031006, "learning_rate": 4.5898919460699845e-05, "loss": 0.4649, "num_input_tokens_seen": 2729096, "step": 10165 }, { "epoch": 2.6657929226736568, "grad_norm": 0.9230599403381348, "learning_rate": 4.589264054736501e-05, "loss": 0.4747, "num_input_tokens_seen": 2730264, "step": 10170 }, { "epoch": 2.667103538663172, "grad_norm": 0.7680661082267761, "learning_rate": 4.58863572612366e-05, "loss": 0.3868, "num_input_tokens_seen": 2731352, "step": 10175 }, { "epoch": 2.6684141546526865, "grad_norm": 1.0163977146148682, "learning_rate": 4.5880069603629705e-05, "loss": 0.4923, "num_input_tokens_seen": 2732552, "step": 10180 }, { "epoch": 2.669724770642202, "grad_norm": 1.9734443426132202, "learning_rate": 4.587377757586032e-05, "loss": 0.539, "num_input_tokens_seen": 2734232, "step": 10185 }, { "epoch": 2.6710353866317167, "grad_norm": 0.7722833156585693, "learning_rate": 4.586748117924536e-05, "loss": 0.4539, "num_input_tokens_seen": 2735480, "step": 10190 }, { "epoch": 2.6723460026212322, "grad_norm": 0.9596455693244934, "learning_rate": 4.586118041510264e-05, "loss": 0.5484, "num_input_tokens_seen": 2736824, "step": 10195 }, { "epoch": 2.673656618610747, "grad_norm": 2.6396987438201904, "learning_rate": 4.585487528475091e-05, "loss": 0.3364, "num_input_tokens_seen": 2737992, "step": 10200 }, { "epoch": 2.674967234600262, "grad_norm": 1.508176326751709, "learning_rate": 4.584856578950981e-05, "loss": 0.444, "num_input_tokens_seen": 2739272, "step": 10205 }, { "epoch": 2.676277850589777, "grad_norm": 0.6327779293060303, "learning_rate": 4.584225193069992e-05, "loss": 0.3925, "num_input_tokens_seen": 2740840, "step": 10210 }, { "epoch": 2.677588466579292, "grad_norm": 0.7989467978477478, "learning_rate": 4.583593370964271e-05, "loss": 0.4571, "num_input_tokens_seen": 2742632, "step": 10215 }, { "epoch": 2.6788990825688073, "grad_norm": 1.288407325744629, "learning_rate": 4.582961112766058e-05, "loss": 0.4566, "num_input_tokens_seen": 2743944, "step": 10220 }, { "epoch": 2.6802096985583224, "grad_norm": 1.0544451475143433, "learning_rate": 4.582328418607682e-05, "loss": 0.4192, "num_input_tokens_seen": 2745064, "step": 10225 }, { "epoch": 2.6815203145478375, "grad_norm": 1.1804770231246948, "learning_rate": 4.581695288621568e-05, "loss": 0.5688, "num_input_tokens_seen": 2746312, "step": 10230 }, { "epoch": 2.6828309305373526, "grad_norm": 0.7538242936134338, "learning_rate": 4.581061722940225e-05, "loss": 0.4829, "num_input_tokens_seen": 2747672, "step": 10235 }, { "epoch": 2.6841415465268676, "grad_norm": 1.2226948738098145, "learning_rate": 4.580427721696261e-05, "loss": 0.5425, "num_input_tokens_seen": 2748680, "step": 10240 }, { "epoch": 2.6854521625163827, "grad_norm": 1.5607408285140991, "learning_rate": 4.579793285022368e-05, "loss": 0.379, "num_input_tokens_seen": 2749944, "step": 10245 }, { "epoch": 2.686762778505898, "grad_norm": 0.55815589427948, "learning_rate": 4.579158413051335e-05, "loss": 0.4989, "num_input_tokens_seen": 2751320, "step": 10250 }, { "epoch": 2.688073394495413, "grad_norm": 0.688313364982605, "learning_rate": 4.578523105916038e-05, "loss": 0.3492, "num_input_tokens_seen": 2753064, "step": 10255 }, { "epoch": 2.689384010484928, "grad_norm": 0.8243019580841064, "learning_rate": 4.577887363749447e-05, "loss": 0.6171, "num_input_tokens_seen": 2754024, "step": 10260 }, { "epoch": 2.690694626474443, "grad_norm": 0.5729010105133057, "learning_rate": 4.57725118668462e-05, "loss": 0.652, "num_input_tokens_seen": 2755800, "step": 10265 }, { "epoch": 2.6920052424639582, "grad_norm": 0.6267798542976379, "learning_rate": 4.5766145748547104e-05, "loss": 0.3711, "num_input_tokens_seen": 2756872, "step": 10270 }, { "epoch": 2.693315858453473, "grad_norm": 0.7688126564025879, "learning_rate": 4.575977528392957e-05, "loss": 0.4103, "num_input_tokens_seen": 2758440, "step": 10275 }, { "epoch": 2.6946264744429884, "grad_norm": 0.6295357942581177, "learning_rate": 4.5753400474326946e-05, "loss": 0.3087, "num_input_tokens_seen": 2759816, "step": 10280 }, { "epoch": 2.695937090432503, "grad_norm": 0.7819052934646606, "learning_rate": 4.574702132107346e-05, "loss": 0.4929, "num_input_tokens_seen": 2761400, "step": 10285 }, { "epoch": 2.6972477064220186, "grad_norm": 0.5963521003723145, "learning_rate": 4.5740637825504265e-05, "loss": 0.4569, "num_input_tokens_seen": 2762392, "step": 10290 }, { "epoch": 2.6985583224115333, "grad_norm": 1.2496589422225952, "learning_rate": 4.5734249988955405e-05, "loss": 0.3955, "num_input_tokens_seen": 2763864, "step": 10295 }, { "epoch": 2.6998689384010484, "grad_norm": 1.2228084802627563, "learning_rate": 4.572785781276385e-05, "loss": 0.4471, "num_input_tokens_seen": 2764856, "step": 10300 }, { "epoch": 2.7011795543905635, "grad_norm": 3.2185683250427246, "learning_rate": 4.5721461298267466e-05, "loss": 0.5549, "num_input_tokens_seen": 2765944, "step": 10305 }, { "epoch": 2.7024901703800785, "grad_norm": 1.7451478242874146, "learning_rate": 4.5715060446805044e-05, "loss": 0.3082, "num_input_tokens_seen": 2766888, "step": 10310 }, { "epoch": 2.7038007863695936, "grad_norm": 0.6927734613418579, "learning_rate": 4.570865525971626e-05, "loss": 0.363, "num_input_tokens_seen": 2768328, "step": 10315 }, { "epoch": 2.7051114023591087, "grad_norm": 1.17095947265625, "learning_rate": 4.570224573834171e-05, "loss": 0.3487, "num_input_tokens_seen": 2769256, "step": 10320 }, { "epoch": 2.706422018348624, "grad_norm": 1.6055806875228882, "learning_rate": 4.569583188402289e-05, "loss": 0.422, "num_input_tokens_seen": 2770744, "step": 10325 }, { "epoch": 2.707732634338139, "grad_norm": 1.1072933673858643, "learning_rate": 4.5689413698102226e-05, "loss": 0.4756, "num_input_tokens_seen": 2772200, "step": 10330 }, { "epoch": 2.709043250327654, "grad_norm": 0.5367483496665955, "learning_rate": 4.5682991181923024e-05, "loss": 0.3413, "num_input_tokens_seen": 2773752, "step": 10335 }, { "epoch": 2.710353866317169, "grad_norm": 2.0468027591705322, "learning_rate": 4.56765643368295e-05, "loss": 0.5611, "num_input_tokens_seen": 2774952, "step": 10340 }, { "epoch": 2.711664482306684, "grad_norm": 2.0456156730651855, "learning_rate": 4.5670133164166795e-05, "loss": 0.397, "num_input_tokens_seen": 2776072, "step": 10345 }, { "epoch": 2.7129750982961993, "grad_norm": 0.5866532325744629, "learning_rate": 4.566369766528093e-05, "loss": 0.3609, "num_input_tokens_seen": 2777560, "step": 10350 }, { "epoch": 2.7142857142857144, "grad_norm": 1.4686737060546875, "learning_rate": 4.565725784151884e-05, "loss": 0.5654, "num_input_tokens_seen": 2778456, "step": 10355 }, { "epoch": 2.7155963302752295, "grad_norm": 1.2935333251953125, "learning_rate": 4.565081369422839e-05, "loss": 0.4263, "num_input_tokens_seen": 2779672, "step": 10360 }, { "epoch": 2.7169069462647446, "grad_norm": 3.3202288150787354, "learning_rate": 4.56443652247583e-05, "loss": 0.3742, "num_input_tokens_seen": 2780824, "step": 10365 }, { "epoch": 2.7182175622542593, "grad_norm": 1.9295376539230347, "learning_rate": 4.5637912434458254e-05, "loss": 0.9734, "num_input_tokens_seen": 2781960, "step": 10370 }, { "epoch": 2.719528178243775, "grad_norm": 1.0904392004013062, "learning_rate": 4.563145532467878e-05, "loss": 0.4838, "num_input_tokens_seen": 2783064, "step": 10375 }, { "epoch": 2.7208387942332894, "grad_norm": 3.1570825576782227, "learning_rate": 4.562499389677137e-05, "loss": 0.5964, "num_input_tokens_seen": 2784280, "step": 10380 }, { "epoch": 2.722149410222805, "grad_norm": 1.3850469589233398, "learning_rate": 4.561852815208837e-05, "loss": 0.727, "num_input_tokens_seen": 2785736, "step": 10385 }, { "epoch": 2.7234600262123196, "grad_norm": 0.7361521124839783, "learning_rate": 4.561205809198305e-05, "loss": 0.4115, "num_input_tokens_seen": 2787064, "step": 10390 }, { "epoch": 2.7247706422018347, "grad_norm": 1.097764015197754, "learning_rate": 4.560558371780959e-05, "loss": 0.4688, "num_input_tokens_seen": 2788120, "step": 10395 }, { "epoch": 2.72608125819135, "grad_norm": 3.0858652591705322, "learning_rate": 4.559910503092306e-05, "loss": 0.4196, "num_input_tokens_seen": 2789288, "step": 10400 }, { "epoch": 2.727391874180865, "grad_norm": 0.9343308210372925, "learning_rate": 4.559262203267945e-05, "loss": 0.5226, "num_input_tokens_seen": 2790280, "step": 10405 }, { "epoch": 2.72870249017038, "grad_norm": 0.8517870306968689, "learning_rate": 4.5586134724435614e-05, "loss": 0.2696, "num_input_tokens_seen": 2791320, "step": 10410 }, { "epoch": 2.730013106159895, "grad_norm": 0.9281659722328186, "learning_rate": 4.557964310754935e-05, "loss": 0.4455, "num_input_tokens_seen": 2792856, "step": 10415 }, { "epoch": 2.73132372214941, "grad_norm": 0.7465404272079468, "learning_rate": 4.557314718337935e-05, "loss": 0.4904, "num_input_tokens_seen": 2794200, "step": 10420 }, { "epoch": 2.7326343381389253, "grad_norm": 0.9770485758781433, "learning_rate": 4.556664695328519e-05, "loss": 0.3813, "num_input_tokens_seen": 2795192, "step": 10425 }, { "epoch": 2.7339449541284404, "grad_norm": 0.8306324481964111, "learning_rate": 4.556014241862736e-05, "loss": 0.3235, "num_input_tokens_seen": 2796296, "step": 10430 }, { "epoch": 2.7352555701179555, "grad_norm": 0.7465631365776062, "learning_rate": 4.555363358076725e-05, "loss": 0.3988, "num_input_tokens_seen": 2797512, "step": 10435 }, { "epoch": 2.7365661861074706, "grad_norm": 4.302236080169678, "learning_rate": 4.5547120441067136e-05, "loss": 0.5127, "num_input_tokens_seen": 2798504, "step": 10440 }, { "epoch": 2.7378768020969857, "grad_norm": 0.5179964303970337, "learning_rate": 4.554060300089022e-05, "loss": 0.3942, "num_input_tokens_seen": 2799832, "step": 10445 }, { "epoch": 2.739187418086501, "grad_norm": 0.9240724444389343, "learning_rate": 4.553408126160059e-05, "loss": 0.3834, "num_input_tokens_seen": 2801048, "step": 10450 }, { "epoch": 2.740498034076016, "grad_norm": 0.7508686780929565, "learning_rate": 4.552755522456324e-05, "loss": 0.4673, "num_input_tokens_seen": 2802568, "step": 10455 }, { "epoch": 2.741808650065531, "grad_norm": 1.389841079711914, "learning_rate": 4.552102489114404e-05, "loss": 0.507, "num_input_tokens_seen": 2804184, "step": 10460 }, { "epoch": 2.7431192660550456, "grad_norm": 0.7893614768981934, "learning_rate": 4.551449026270979e-05, "loss": 0.3741, "num_input_tokens_seen": 2805512, "step": 10465 }, { "epoch": 2.744429882044561, "grad_norm": 1.6113132238388062, "learning_rate": 4.550795134062817e-05, "loss": 0.5046, "num_input_tokens_seen": 2806664, "step": 10470 }, { "epoch": 2.745740498034076, "grad_norm": 0.8689745664596558, "learning_rate": 4.5501408126267784e-05, "loss": 0.4424, "num_input_tokens_seen": 2808200, "step": 10475 }, { "epoch": 2.747051114023591, "grad_norm": 0.8488057255744934, "learning_rate": 4.5494860620998086e-05, "loss": 0.3842, "num_input_tokens_seen": 2809528, "step": 10480 }, { "epoch": 2.748361730013106, "grad_norm": 2.5881004333496094, "learning_rate": 4.548830882618948e-05, "loss": 0.5063, "num_input_tokens_seen": 2810984, "step": 10485 }, { "epoch": 2.749672346002621, "grad_norm": 0.9705827832221985, "learning_rate": 4.5481752743213234e-05, "loss": 0.3944, "num_input_tokens_seen": 2812600, "step": 10490 }, { "epoch": 2.750982961992136, "grad_norm": 1.4332165718078613, "learning_rate": 4.5475192373441525e-05, "loss": 0.4307, "num_input_tokens_seen": 2814568, "step": 10495 }, { "epoch": 2.7522935779816513, "grad_norm": 0.7154757380485535, "learning_rate": 4.5468627718247436e-05, "loss": 0.6888, "num_input_tokens_seen": 2816216, "step": 10500 }, { "epoch": 2.7536041939711664, "grad_norm": 1.3058146238327026, "learning_rate": 4.546205877900494e-05, "loss": 0.5149, "num_input_tokens_seen": 2817560, "step": 10505 }, { "epoch": 2.7549148099606815, "grad_norm": 1.8384207487106323, "learning_rate": 4.5455485557088885e-05, "loss": 0.5955, "num_input_tokens_seen": 2818616, "step": 10510 }, { "epoch": 2.7562254259501966, "grad_norm": 0.9141144752502441, "learning_rate": 4.544890805387505e-05, "loss": 0.396, "num_input_tokens_seen": 2819976, "step": 10515 }, { "epoch": 2.7575360419397117, "grad_norm": 0.4181059002876282, "learning_rate": 4.544232627074009e-05, "loss": 0.4918, "num_input_tokens_seen": 2821464, "step": 10520 }, { "epoch": 2.758846657929227, "grad_norm": 0.7489643692970276, "learning_rate": 4.5435740209061574e-05, "loss": 0.4272, "num_input_tokens_seen": 2822744, "step": 10525 }, { "epoch": 2.760157273918742, "grad_norm": 0.9428383708000183, "learning_rate": 4.5429149870217936e-05, "loss": 0.7109, "num_input_tokens_seen": 2823992, "step": 10530 }, { "epoch": 2.761467889908257, "grad_norm": 1.2240785360336304, "learning_rate": 4.5422555255588524e-05, "loss": 0.3629, "num_input_tokens_seen": 2825256, "step": 10535 }, { "epoch": 2.762778505897772, "grad_norm": 1.1704312562942505, "learning_rate": 4.541595636655359e-05, "loss": 0.4293, "num_input_tokens_seen": 2826312, "step": 10540 }, { "epoch": 2.764089121887287, "grad_norm": 1.2741886377334595, "learning_rate": 4.540935320449426e-05, "loss": 0.3665, "num_input_tokens_seen": 2827816, "step": 10545 }, { "epoch": 2.765399737876802, "grad_norm": 1.2153253555297852, "learning_rate": 4.540274577079258e-05, "loss": 0.558, "num_input_tokens_seen": 2829272, "step": 10550 }, { "epoch": 2.7667103538663174, "grad_norm": 0.9573599696159363, "learning_rate": 4.539613406683145e-05, "loss": 0.5754, "num_input_tokens_seen": 2830648, "step": 10555 }, { "epoch": 2.768020969855832, "grad_norm": 1.5360604524612427, "learning_rate": 4.538951809399471e-05, "loss": 0.637, "num_input_tokens_seen": 2832088, "step": 10560 }, { "epoch": 2.7693315858453476, "grad_norm": 0.5979472398757935, "learning_rate": 4.538289785366706e-05, "loss": 0.32, "num_input_tokens_seen": 2833720, "step": 10565 }, { "epoch": 2.770642201834862, "grad_norm": 1.2098339796066284, "learning_rate": 4.53762733472341e-05, "loss": 0.5764, "num_input_tokens_seen": 2834888, "step": 10570 }, { "epoch": 2.7719528178243773, "grad_norm": 1.147695541381836, "learning_rate": 4.536964457608235e-05, "loss": 0.3799, "num_input_tokens_seen": 2836232, "step": 10575 }, { "epoch": 2.7732634338138924, "grad_norm": 1.6820716857910156, "learning_rate": 4.5363011541599185e-05, "loss": 0.5439, "num_input_tokens_seen": 2837720, "step": 10580 }, { "epoch": 2.7745740498034075, "grad_norm": 1.0538495779037476, "learning_rate": 4.535637424517288e-05, "loss": 0.6059, "num_input_tokens_seen": 2839000, "step": 10585 }, { "epoch": 2.7758846657929226, "grad_norm": 1.2445604801177979, "learning_rate": 4.5349732688192624e-05, "loss": 0.4402, "num_input_tokens_seen": 2840360, "step": 10590 }, { "epoch": 2.7771952817824377, "grad_norm": 1.1281843185424805, "learning_rate": 4.534308687204848e-05, "loss": 0.452, "num_input_tokens_seen": 2841576, "step": 10595 }, { "epoch": 2.778505897771953, "grad_norm": 1.5193393230438232, "learning_rate": 4.53364367981314e-05, "loss": 0.5864, "num_input_tokens_seen": 2842888, "step": 10600 }, { "epoch": 2.779816513761468, "grad_norm": 2.047736883163452, "learning_rate": 4.5329782467833236e-05, "loss": 0.3395, "num_input_tokens_seen": 2844040, "step": 10605 }, { "epoch": 2.781127129750983, "grad_norm": 2.434643030166626, "learning_rate": 4.532312388254674e-05, "loss": 0.4111, "num_input_tokens_seen": 2845208, "step": 10610 }, { "epoch": 2.782437745740498, "grad_norm": 0.2900049686431885, "learning_rate": 4.5316461043665516e-05, "loss": 0.5442, "num_input_tokens_seen": 2846792, "step": 10615 }, { "epoch": 2.783748361730013, "grad_norm": 0.7413959503173828, "learning_rate": 4.53097939525841e-05, "loss": 0.3262, "num_input_tokens_seen": 2848120, "step": 10620 }, { "epoch": 2.7850589777195283, "grad_norm": 0.5136945843696594, "learning_rate": 4.530312261069791e-05, "loss": 0.4699, "num_input_tokens_seen": 2849544, "step": 10625 }, { "epoch": 2.7863695937090434, "grad_norm": 1.1655598878860474, "learning_rate": 4.5296447019403245e-05, "loss": 0.5193, "num_input_tokens_seen": 2850648, "step": 10630 }, { "epoch": 2.7876802096985585, "grad_norm": 1.7492032051086426, "learning_rate": 4.5289767180097275e-05, "loss": 0.5641, "num_input_tokens_seen": 2851864, "step": 10635 }, { "epoch": 2.7889908256880735, "grad_norm": 1.3174082040786743, "learning_rate": 4.528308309417809e-05, "loss": 0.3605, "num_input_tokens_seen": 2852984, "step": 10640 }, { "epoch": 2.790301441677588, "grad_norm": 0.6743907928466797, "learning_rate": 4.527639476304467e-05, "loss": 0.4487, "num_input_tokens_seen": 2854152, "step": 10645 }, { "epoch": 2.7916120576671037, "grad_norm": 0.8679163455963135, "learning_rate": 4.526970218809686e-05, "loss": 0.3784, "num_input_tokens_seen": 2856104, "step": 10650 }, { "epoch": 2.7929226736566184, "grad_norm": 1.5290807485580444, "learning_rate": 4.52630053707354e-05, "loss": 0.5406, "num_input_tokens_seen": 2857176, "step": 10655 }, { "epoch": 2.794233289646134, "grad_norm": 0.9494523406028748, "learning_rate": 4.525630431236193e-05, "loss": 0.3565, "num_input_tokens_seen": 2858632, "step": 10660 }, { "epoch": 2.7955439056356486, "grad_norm": 0.7960340976715088, "learning_rate": 4.524959901437897e-05, "loss": 0.4718, "num_input_tokens_seen": 2860312, "step": 10665 }, { "epoch": 2.7968545216251637, "grad_norm": 4.222696304321289, "learning_rate": 4.5242889478189926e-05, "loss": 0.4597, "num_input_tokens_seen": 2861512, "step": 10670 }, { "epoch": 2.7981651376146788, "grad_norm": 1.2262110710144043, "learning_rate": 4.5236175705199094e-05, "loss": 0.4796, "num_input_tokens_seen": 2862904, "step": 10675 }, { "epoch": 2.799475753604194, "grad_norm": 1.1220972537994385, "learning_rate": 4.522945769681164e-05, "loss": 0.4581, "num_input_tokens_seen": 2864232, "step": 10680 }, { "epoch": 2.800786369593709, "grad_norm": 9.467529296875, "learning_rate": 4.522273545443365e-05, "loss": 0.2812, "num_input_tokens_seen": 2865608, "step": 10685 }, { "epoch": 2.802096985583224, "grad_norm": 1.138298749923706, "learning_rate": 4.521600897947208e-05, "loss": 0.4895, "num_input_tokens_seen": 2867224, "step": 10690 }, { "epoch": 2.803407601572739, "grad_norm": 0.6650664806365967, "learning_rate": 4.520927827333475e-05, "loss": 0.4683, "num_input_tokens_seen": 2868344, "step": 10695 }, { "epoch": 2.8047182175622543, "grad_norm": 1.0977435111999512, "learning_rate": 4.5202543337430405e-05, "loss": 0.3804, "num_input_tokens_seen": 2869528, "step": 10700 }, { "epoch": 2.8060288335517694, "grad_norm": 0.48201167583465576, "learning_rate": 4.519580417316863e-05, "loss": 0.2403, "num_input_tokens_seen": 2871112, "step": 10705 }, { "epoch": 2.8073394495412844, "grad_norm": 1.0160572528839111, "learning_rate": 4.518906078195996e-05, "loss": 0.4257, "num_input_tokens_seen": 2872184, "step": 10710 }, { "epoch": 2.8086500655307995, "grad_norm": 1.2010172605514526, "learning_rate": 4.518231316521574e-05, "loss": 0.4347, "num_input_tokens_seen": 2873448, "step": 10715 }, { "epoch": 2.8099606815203146, "grad_norm": 0.8558980226516724, "learning_rate": 4.517556132434825e-05, "loss": 0.3564, "num_input_tokens_seen": 2874664, "step": 10720 }, { "epoch": 2.8112712975098297, "grad_norm": 1.0365791320800781, "learning_rate": 4.5168805260770633e-05, "loss": 0.5023, "num_input_tokens_seen": 2876072, "step": 10725 }, { "epoch": 2.812581913499345, "grad_norm": 1.1080867052078247, "learning_rate": 4.516204497589692e-05, "loss": 0.3922, "num_input_tokens_seen": 2877144, "step": 10730 }, { "epoch": 2.81389252948886, "grad_norm": 0.7588904500007629, "learning_rate": 4.5155280471142036e-05, "loss": 0.4154, "num_input_tokens_seen": 2878680, "step": 10735 }, { "epoch": 2.8152031454783746, "grad_norm": 0.5754498243331909, "learning_rate": 4.514851174792177e-05, "loss": 0.5063, "num_input_tokens_seen": 2879640, "step": 10740 }, { "epoch": 2.81651376146789, "grad_norm": 1.0377075672149658, "learning_rate": 4.514173880765282e-05, "loss": 0.5452, "num_input_tokens_seen": 2880904, "step": 10745 }, { "epoch": 2.8178243774574048, "grad_norm": 1.35277259349823, "learning_rate": 4.513496165175273e-05, "loss": 0.4965, "num_input_tokens_seen": 2882312, "step": 10750 }, { "epoch": 2.8191349934469203, "grad_norm": 1.0376431941986084, "learning_rate": 4.512818028163996e-05, "loss": 0.3541, "num_input_tokens_seen": 2883624, "step": 10755 }, { "epoch": 2.820445609436435, "grad_norm": 0.7090166211128235, "learning_rate": 4.512139469873384e-05, "loss": 0.4004, "num_input_tokens_seen": 2884680, "step": 10760 }, { "epoch": 2.82175622542595, "grad_norm": 0.5703844428062439, "learning_rate": 4.511460490445457e-05, "loss": 0.3121, "num_input_tokens_seen": 2885800, "step": 10765 }, { "epoch": 2.823066841415465, "grad_norm": 0.30476489663124084, "learning_rate": 4.510781090022325e-05, "loss": 0.3702, "num_input_tokens_seen": 2887352, "step": 10770 }, { "epoch": 2.8243774574049803, "grad_norm": 2.1522512435913086, "learning_rate": 4.5101012687461854e-05, "loss": 0.4287, "num_input_tokens_seen": 2888632, "step": 10775 }, { "epoch": 2.8256880733944953, "grad_norm": 1.1662527322769165, "learning_rate": 4.509421026759323e-05, "loss": 0.3922, "num_input_tokens_seen": 2889576, "step": 10780 }, { "epoch": 2.8269986893840104, "grad_norm": 1.3883858919143677, "learning_rate": 4.508740364204113e-05, "loss": 0.367, "num_input_tokens_seen": 2891320, "step": 10785 }, { "epoch": 2.8283093053735255, "grad_norm": 0.7772650122642517, "learning_rate": 4.508059281223015e-05, "loss": 0.4425, "num_input_tokens_seen": 2892984, "step": 10790 }, { "epoch": 2.8296199213630406, "grad_norm": 2.251997232437134, "learning_rate": 4.507377777958578e-05, "loss": 0.482, "num_input_tokens_seen": 2894232, "step": 10795 }, { "epoch": 2.8309305373525557, "grad_norm": 1.760118007659912, "learning_rate": 4.5066958545534415e-05, "loss": 0.3804, "num_input_tokens_seen": 2895224, "step": 10800 }, { "epoch": 2.832241153342071, "grad_norm": 0.5552123188972473, "learning_rate": 4.50601351115033e-05, "loss": 0.4384, "num_input_tokens_seen": 2896536, "step": 10805 }, { "epoch": 2.833551769331586, "grad_norm": 0.5738458037376404, "learning_rate": 4.5053307478920564e-05, "loss": 0.4721, "num_input_tokens_seen": 2897992, "step": 10810 }, { "epoch": 2.834862385321101, "grad_norm": 0.6763349175453186, "learning_rate": 4.504647564921523e-05, "loss": 0.412, "num_input_tokens_seen": 2899496, "step": 10815 }, { "epoch": 2.836173001310616, "grad_norm": 7.47363805770874, "learning_rate": 4.503963962381717e-05, "loss": 0.4963, "num_input_tokens_seen": 2900696, "step": 10820 }, { "epoch": 2.837483617300131, "grad_norm": 0.946419358253479, "learning_rate": 4.503279940415717e-05, "loss": 0.3425, "num_input_tokens_seen": 2902072, "step": 10825 }, { "epoch": 2.8387942332896463, "grad_norm": 1.7990496158599854, "learning_rate": 4.5025954991666865e-05, "loss": 0.4314, "num_input_tokens_seen": 2903256, "step": 10830 }, { "epoch": 2.840104849279161, "grad_norm": 0.7960519194602966, "learning_rate": 4.501910638777878e-05, "loss": 0.4663, "num_input_tokens_seen": 2904632, "step": 10835 }, { "epoch": 2.8414154652686765, "grad_norm": 0.45873865485191345, "learning_rate": 4.5012253593926324e-05, "loss": 0.3424, "num_input_tokens_seen": 2905848, "step": 10840 }, { "epoch": 2.842726081258191, "grad_norm": 1.241364598274231, "learning_rate": 4.5005396611543765e-05, "loss": 0.3627, "num_input_tokens_seen": 2907208, "step": 10845 }, { "epoch": 2.8440366972477067, "grad_norm": 0.5873698592185974, "learning_rate": 4.499853544206626e-05, "loss": 0.4389, "num_input_tokens_seen": 2908424, "step": 10850 }, { "epoch": 2.8453473132372213, "grad_norm": 0.8283502459526062, "learning_rate": 4.499167008692985e-05, "loss": 0.5166, "num_input_tokens_seen": 2909496, "step": 10855 }, { "epoch": 2.8466579292267364, "grad_norm": 0.5990482568740845, "learning_rate": 4.4984800547571425e-05, "loss": 0.3246, "num_input_tokens_seen": 2911144, "step": 10860 }, { "epoch": 2.8479685452162515, "grad_norm": 0.5440280437469482, "learning_rate": 4.497792682542878e-05, "loss": 0.3505, "num_input_tokens_seen": 2912536, "step": 10865 }, { "epoch": 2.8492791612057666, "grad_norm": 1.4577785730361938, "learning_rate": 4.4971048921940575e-05, "loss": 0.4069, "num_input_tokens_seen": 2913672, "step": 10870 }, { "epoch": 2.8505897771952817, "grad_norm": 2.596254348754883, "learning_rate": 4.496416683854633e-05, "loss": 0.6002, "num_input_tokens_seen": 2914616, "step": 10875 }, { "epoch": 2.851900393184797, "grad_norm": 0.902684211730957, "learning_rate": 4.495728057668647e-05, "loss": 0.5686, "num_input_tokens_seen": 2915912, "step": 10880 }, { "epoch": 2.853211009174312, "grad_norm": 1.004073977470398, "learning_rate": 4.4950390137802254e-05, "loss": 0.5428, "num_input_tokens_seen": 2917336, "step": 10885 }, { "epoch": 2.854521625163827, "grad_norm": 1.4365122318267822, "learning_rate": 4.4943495523335865e-05, "loss": 0.6272, "num_input_tokens_seen": 2918776, "step": 10890 }, { "epoch": 2.855832241153342, "grad_norm": 1.0838050842285156, "learning_rate": 4.493659673473032e-05, "loss": 0.4413, "num_input_tokens_seen": 2921144, "step": 10895 }, { "epoch": 2.857142857142857, "grad_norm": 0.6855168342590332, "learning_rate": 4.492969377342952e-05, "loss": 0.2722, "num_input_tokens_seen": 2922776, "step": 10900 }, { "epoch": 2.8584534731323723, "grad_norm": 1.0021765232086182, "learning_rate": 4.4922786640878265e-05, "loss": 0.4067, "num_input_tokens_seen": 2923992, "step": 10905 }, { "epoch": 2.8597640891218874, "grad_norm": 0.9853978753089905, "learning_rate": 4.491587533852218e-05, "loss": 0.3402, "num_input_tokens_seen": 2925432, "step": 10910 }, { "epoch": 2.8610747051114025, "grad_norm": 0.9148572683334351, "learning_rate": 4.49089598678078e-05, "loss": 0.4418, "num_input_tokens_seen": 2926776, "step": 10915 }, { "epoch": 2.8623853211009176, "grad_norm": 1.3103586435317993, "learning_rate": 4.490204023018252e-05, "loss": 0.4102, "num_input_tokens_seen": 2927656, "step": 10920 }, { "epoch": 2.8636959370904327, "grad_norm": 0.6972234845161438, "learning_rate": 4.4895116427094605e-05, "loss": 0.5084, "num_input_tokens_seen": 2928808, "step": 10925 }, { "epoch": 2.8650065530799473, "grad_norm": 1.471651315689087, "learning_rate": 4.48881884599932e-05, "loss": 0.3954, "num_input_tokens_seen": 2930472, "step": 10930 }, { "epoch": 2.866317169069463, "grad_norm": 1.4324820041656494, "learning_rate": 4.4881256330328314e-05, "loss": 0.3925, "num_input_tokens_seen": 2931976, "step": 10935 }, { "epoch": 2.8676277850589775, "grad_norm": 0.9355971217155457, "learning_rate": 4.4874320039550834e-05, "loss": 0.4232, "num_input_tokens_seen": 2932968, "step": 10940 }, { "epoch": 2.8689384010484926, "grad_norm": 1.3156145811080933, "learning_rate": 4.48673795891125e-05, "loss": 0.4492, "num_input_tokens_seen": 2934072, "step": 10945 }, { "epoch": 2.8702490170380077, "grad_norm": 3.297308921813965, "learning_rate": 4.4860434980465954e-05, "loss": 0.5857, "num_input_tokens_seen": 2935560, "step": 10950 }, { "epoch": 2.871559633027523, "grad_norm": 0.7950226664543152, "learning_rate": 4.485348621506468e-05, "loss": 0.3486, "num_input_tokens_seen": 2937112, "step": 10955 }, { "epoch": 2.872870249017038, "grad_norm": 1.7278326749801636, "learning_rate": 4.484653329436305e-05, "loss": 0.471, "num_input_tokens_seen": 2938360, "step": 10960 }, { "epoch": 2.874180865006553, "grad_norm": 1.054707646369934, "learning_rate": 4.483957621981629e-05, "loss": 0.5951, "num_input_tokens_seen": 2939624, "step": 10965 }, { "epoch": 2.875491480996068, "grad_norm": 0.7023601531982422, "learning_rate": 4.483261499288051e-05, "loss": 0.3073, "num_input_tokens_seen": 2941032, "step": 10970 }, { "epoch": 2.876802096985583, "grad_norm": 1.9351731538772583, "learning_rate": 4.482564961501268e-05, "loss": 0.732, "num_input_tokens_seen": 2942200, "step": 10975 }, { "epoch": 2.8781127129750983, "grad_norm": 3.4908292293548584, "learning_rate": 4.481868008767063e-05, "loss": 0.4321, "num_input_tokens_seen": 2945208, "step": 10980 }, { "epoch": 2.8794233289646134, "grad_norm": 1.8144824504852295, "learning_rate": 4.4811706412313094e-05, "loss": 0.4034, "num_input_tokens_seen": 2946248, "step": 10985 }, { "epoch": 2.8807339449541285, "grad_norm": 0.619545042514801, "learning_rate": 4.4804728590399635e-05, "loss": 0.4189, "num_input_tokens_seen": 2948088, "step": 10990 }, { "epoch": 2.8820445609436436, "grad_norm": 1.1654947996139526, "learning_rate": 4.47977466233907e-05, "loss": 0.3029, "num_input_tokens_seen": 2949240, "step": 10995 }, { "epoch": 2.8833551769331587, "grad_norm": 3.2465660572052, "learning_rate": 4.479076051274761e-05, "loss": 0.5459, "num_input_tokens_seen": 2950328, "step": 11000 }, { "epoch": 2.8846657929226738, "grad_norm": 0.4212243854999542, "learning_rate": 4.478377025993254e-05, "loss": 0.3602, "num_input_tokens_seen": 2951848, "step": 11005 }, { "epoch": 2.885976408912189, "grad_norm": 1.5197100639343262, "learning_rate": 4.477677586640854e-05, "loss": 0.4934, "num_input_tokens_seen": 2952888, "step": 11010 }, { "epoch": 2.8872870249017035, "grad_norm": 1.285616159439087, "learning_rate": 4.476977733363951e-05, "loss": 0.5045, "num_input_tokens_seen": 2953912, "step": 11015 }, { "epoch": 2.888597640891219, "grad_norm": 2.498528003692627, "learning_rate": 4.4762774663090256e-05, "loss": 0.483, "num_input_tokens_seen": 2955288, "step": 11020 }, { "epoch": 2.8899082568807337, "grad_norm": 0.8270884156227112, "learning_rate": 4.4755767856226405e-05, "loss": 0.3404, "num_input_tokens_seen": 2956536, "step": 11025 }, { "epoch": 2.8912188728702493, "grad_norm": 0.8422662019729614, "learning_rate": 4.474875691451448e-05, "loss": 0.3724, "num_input_tokens_seen": 2957528, "step": 11030 }, { "epoch": 2.892529488859764, "grad_norm": 0.9516127705574036, "learning_rate": 4.474174183942186e-05, "loss": 0.3889, "num_input_tokens_seen": 2958520, "step": 11035 }, { "epoch": 2.893840104849279, "grad_norm": 0.7700480222702026, "learning_rate": 4.473472263241678e-05, "loss": 0.6615, "num_input_tokens_seen": 2959752, "step": 11040 }, { "epoch": 2.895150720838794, "grad_norm": 1.117974877357483, "learning_rate": 4.472769929496835e-05, "loss": 0.4299, "num_input_tokens_seen": 2961352, "step": 11045 }, { "epoch": 2.896461336828309, "grad_norm": 1.165185570716858, "learning_rate": 4.472067182854654e-05, "loss": 0.5679, "num_input_tokens_seen": 2962568, "step": 11050 }, { "epoch": 2.8977719528178243, "grad_norm": 2.393207311630249, "learning_rate": 4.4713640234622205e-05, "loss": 0.478, "num_input_tokens_seen": 2963816, "step": 11055 }, { "epoch": 2.8990825688073394, "grad_norm": 0.42401066422462463, "learning_rate": 4.4706604514667025e-05, "loss": 0.4333, "num_input_tokens_seen": 2965096, "step": 11060 }, { "epoch": 2.9003931847968545, "grad_norm": 0.7206100225448608, "learning_rate": 4.469956467015357e-05, "loss": 0.6074, "num_input_tokens_seen": 2966744, "step": 11065 }, { "epoch": 2.9017038007863696, "grad_norm": 1.0507038831710815, "learning_rate": 4.469252070255527e-05, "loss": 0.5076, "num_input_tokens_seen": 2967768, "step": 11070 }, { "epoch": 2.9030144167758847, "grad_norm": 0.5395247936248779, "learning_rate": 4.4685472613346404e-05, "loss": 0.5177, "num_input_tokens_seen": 2969240, "step": 11075 }, { "epoch": 2.9043250327653998, "grad_norm": 1.1053310632705688, "learning_rate": 4.467842040400214e-05, "loss": 0.4897, "num_input_tokens_seen": 2970872, "step": 11080 }, { "epoch": 2.905635648754915, "grad_norm": 0.3597072958946228, "learning_rate": 4.467136407599849e-05, "loss": 0.3903, "num_input_tokens_seen": 2972296, "step": 11085 }, { "epoch": 2.90694626474443, "grad_norm": 0.4368862509727478, "learning_rate": 4.466430363081233e-05, "loss": 0.3655, "num_input_tokens_seen": 2974408, "step": 11090 }, { "epoch": 2.908256880733945, "grad_norm": 0.8946187496185303, "learning_rate": 4.465723906992139e-05, "loss": 0.522, "num_input_tokens_seen": 2975448, "step": 11095 }, { "epoch": 2.90956749672346, "grad_norm": 2.109252452850342, "learning_rate": 4.465017039480428e-05, "loss": 0.5057, "num_input_tokens_seen": 2976648, "step": 11100 }, { "epoch": 2.9108781127129753, "grad_norm": 0.5017095804214478, "learning_rate": 4.464309760694047e-05, "loss": 0.4409, "num_input_tokens_seen": 2978216, "step": 11105 }, { "epoch": 2.91218872870249, "grad_norm": 0.8688420653343201, "learning_rate": 4.4636020707810254e-05, "loss": 0.4293, "num_input_tokens_seen": 2979368, "step": 11110 }, { "epoch": 2.9134993446920054, "grad_norm": 0.6314570903778076, "learning_rate": 4.462893969889484e-05, "loss": 0.3623, "num_input_tokens_seen": 2980760, "step": 11115 }, { "epoch": 2.91480996068152, "grad_norm": 1.359541893005371, "learning_rate": 4.462185458167626e-05, "loss": 0.409, "num_input_tokens_seen": 2982280, "step": 11120 }, { "epoch": 2.9161205766710356, "grad_norm": 1.0684316158294678, "learning_rate": 4.4614765357637425e-05, "loss": 0.3929, "num_input_tokens_seen": 2983624, "step": 11125 }, { "epoch": 2.9174311926605503, "grad_norm": 0.6183621287345886, "learning_rate": 4.4607672028262094e-05, "loss": 0.4359, "num_input_tokens_seen": 2984936, "step": 11130 }, { "epoch": 2.9187418086500654, "grad_norm": 1.7711127996444702, "learning_rate": 4.460057459503488e-05, "loss": 0.5407, "num_input_tokens_seen": 2985864, "step": 11135 }, { "epoch": 2.9200524246395805, "grad_norm": 0.529178261756897, "learning_rate": 4.4593473059441274e-05, "loss": 0.6462, "num_input_tokens_seen": 2987336, "step": 11140 }, { "epoch": 2.9213630406290956, "grad_norm": 1.637194037437439, "learning_rate": 4.458636742296761e-05, "loss": 0.4046, "num_input_tokens_seen": 2988712, "step": 11145 }, { "epoch": 2.9226736566186107, "grad_norm": 1.8990720510482788, "learning_rate": 4.4579257687101097e-05, "loss": 0.4545, "num_input_tokens_seen": 2990184, "step": 11150 }, { "epoch": 2.9239842726081258, "grad_norm": 0.3569866716861725, "learning_rate": 4.457214385332977e-05, "loss": 0.3959, "num_input_tokens_seen": 2991720, "step": 11155 }, { "epoch": 2.925294888597641, "grad_norm": 1.0908446311950684, "learning_rate": 4.456502592314256e-05, "loss": 0.4136, "num_input_tokens_seen": 2993352, "step": 11160 }, { "epoch": 2.926605504587156, "grad_norm": 0.5755401253700256, "learning_rate": 4.455790389802922e-05, "loss": 0.429, "num_input_tokens_seen": 2995048, "step": 11165 }, { "epoch": 2.927916120576671, "grad_norm": 0.7114539742469788, "learning_rate": 4.4550777779480393e-05, "loss": 0.3059, "num_input_tokens_seen": 2996296, "step": 11170 }, { "epoch": 2.929226736566186, "grad_norm": 0.5872527956962585, "learning_rate": 4.454364756898756e-05, "loss": 0.3617, "num_input_tokens_seen": 2997240, "step": 11175 }, { "epoch": 2.9305373525557012, "grad_norm": 0.6484159827232361, "learning_rate": 4.453651326804305e-05, "loss": 0.4203, "num_input_tokens_seen": 2998328, "step": 11180 }, { "epoch": 2.9318479685452163, "grad_norm": 1.1291297674179077, "learning_rate": 4.452937487814007e-05, "loss": 0.3571, "num_input_tokens_seen": 2999656, "step": 11185 }, { "epoch": 2.9331585845347314, "grad_norm": 1.186318278312683, "learning_rate": 4.452223240077269e-05, "loss": 0.2981, "num_input_tokens_seen": 3000968, "step": 11190 }, { "epoch": 2.9344692005242465, "grad_norm": 0.8699777126312256, "learning_rate": 4.451508583743578e-05, "loss": 0.5768, "num_input_tokens_seen": 3002696, "step": 11195 }, { "epoch": 2.9357798165137616, "grad_norm": 0.9419727921485901, "learning_rate": 4.450793518962514e-05, "loss": 0.6667, "num_input_tokens_seen": 3003832, "step": 11200 }, { "epoch": 2.9370904325032763, "grad_norm": 1.4011867046356201, "learning_rate": 4.450078045883736e-05, "loss": 0.4377, "num_input_tokens_seen": 3004984, "step": 11205 }, { "epoch": 2.938401048492792, "grad_norm": 1.9926025867462158, "learning_rate": 4.449362164656993e-05, "loss": 0.5448, "num_input_tokens_seen": 3006040, "step": 11210 }, { "epoch": 2.9397116644823065, "grad_norm": 0.8680263161659241, "learning_rate": 4.448645875432117e-05, "loss": 0.4584, "num_input_tokens_seen": 3007448, "step": 11215 }, { "epoch": 2.941022280471822, "grad_norm": 0.48358863592147827, "learning_rate": 4.447929178359026e-05, "loss": 0.4483, "num_input_tokens_seen": 3008776, "step": 11220 }, { "epoch": 2.9423328964613367, "grad_norm": 0.7852267622947693, "learning_rate": 4.4472120735877246e-05, "loss": 0.4643, "num_input_tokens_seen": 3009864, "step": 11225 }, { "epoch": 2.9436435124508518, "grad_norm": 0.4182153344154358, "learning_rate": 4.4464945612683005e-05, "loss": 0.5691, "num_input_tokens_seen": 3011176, "step": 11230 }, { "epoch": 2.944954128440367, "grad_norm": 0.7670403122901917, "learning_rate": 4.4457766415509275e-05, "loss": 0.3691, "num_input_tokens_seen": 3012664, "step": 11235 }, { "epoch": 2.946264744429882, "grad_norm": 1.0395978689193726, "learning_rate": 4.445058314585866e-05, "loss": 0.4718, "num_input_tokens_seen": 3013704, "step": 11240 }, { "epoch": 2.947575360419397, "grad_norm": 1.0511772632598877, "learning_rate": 4.44433958052346e-05, "loss": 0.482, "num_input_tokens_seen": 3015224, "step": 11245 }, { "epoch": 2.948885976408912, "grad_norm": 1.1080737113952637, "learning_rate": 4.443620439514138e-05, "loss": 0.5043, "num_input_tokens_seen": 3016776, "step": 11250 }, { "epoch": 2.9501965923984272, "grad_norm": 0.5182813405990601, "learning_rate": 4.442900891708418e-05, "loss": 0.5361, "num_input_tokens_seen": 3017960, "step": 11255 }, { "epoch": 2.9515072083879423, "grad_norm": 1.1424126625061035, "learning_rate": 4.442180937256898e-05, "loss": 0.4362, "num_input_tokens_seen": 3019176, "step": 11260 }, { "epoch": 2.9528178243774574, "grad_norm": 1.0460125207901, "learning_rate": 4.4414605763102635e-05, "loss": 0.4871, "num_input_tokens_seen": 3020392, "step": 11265 }, { "epoch": 2.9541284403669725, "grad_norm": 0.6798220276832581, "learning_rate": 4.4407398090192846e-05, "loss": 0.3267, "num_input_tokens_seen": 3021496, "step": 11270 }, { "epoch": 2.9554390563564876, "grad_norm": 1.6447490453720093, "learning_rate": 4.440018635534818e-05, "loss": 0.5667, "num_input_tokens_seen": 3022536, "step": 11275 }, { "epoch": 2.9567496723460027, "grad_norm": 2.436394214630127, "learning_rate": 4.439297056007802e-05, "loss": 0.6269, "num_input_tokens_seen": 3024024, "step": 11280 }, { "epoch": 2.958060288335518, "grad_norm": 1.1841866970062256, "learning_rate": 4.4385750705892634e-05, "loss": 0.4075, "num_input_tokens_seen": 3025256, "step": 11285 }, { "epoch": 2.959370904325033, "grad_norm": 1.1634842157363892, "learning_rate": 4.437852679430313e-05, "loss": 0.3865, "num_input_tokens_seen": 3026792, "step": 11290 }, { "epoch": 2.960681520314548, "grad_norm": 0.9909886717796326, "learning_rate": 4.437129882682144e-05, "loss": 0.468, "num_input_tokens_seen": 3028424, "step": 11295 }, { "epoch": 2.9619921363040627, "grad_norm": 0.9761015772819519, "learning_rate": 4.436406680496038e-05, "loss": 0.3984, "num_input_tokens_seen": 3029416, "step": 11300 }, { "epoch": 2.963302752293578, "grad_norm": 0.7200692892074585, "learning_rate": 4.43568307302336e-05, "loss": 0.4267, "num_input_tokens_seen": 3030776, "step": 11305 }, { "epoch": 2.964613368283093, "grad_norm": 1.5596954822540283, "learning_rate": 4.4349590604155586e-05, "loss": 0.4276, "num_input_tokens_seen": 3031896, "step": 11310 }, { "epoch": 2.9659239842726084, "grad_norm": 1.767038106918335, "learning_rate": 4.43423464282417e-05, "loss": 0.5792, "num_input_tokens_seen": 3032952, "step": 11315 }, { "epoch": 2.967234600262123, "grad_norm": 1.0322219133377075, "learning_rate": 4.4335098204008126e-05, "loss": 0.5295, "num_input_tokens_seen": 3034760, "step": 11320 }, { "epoch": 2.968545216251638, "grad_norm": 0.7681983709335327, "learning_rate": 4.43278459329719e-05, "loss": 0.5438, "num_input_tokens_seen": 3035976, "step": 11325 }, { "epoch": 2.9698558322411532, "grad_norm": 0.529216468334198, "learning_rate": 4.432058961665092e-05, "loss": 0.6107, "num_input_tokens_seen": 3037304, "step": 11330 }, { "epoch": 2.9711664482306683, "grad_norm": 0.7971901893615723, "learning_rate": 4.4313329256563916e-05, "loss": 0.3243, "num_input_tokens_seen": 3038664, "step": 11335 }, { "epoch": 2.9724770642201834, "grad_norm": 0.730625331401825, "learning_rate": 4.430606485423048e-05, "loss": 0.39, "num_input_tokens_seen": 3040024, "step": 11340 }, { "epoch": 2.9737876802096985, "grad_norm": 0.5416414737701416, "learning_rate": 4.4298796411171015e-05, "loss": 0.5508, "num_input_tokens_seen": 3041320, "step": 11345 }, { "epoch": 2.9750982961992136, "grad_norm": 0.5781485438346863, "learning_rate": 4.429152392890681e-05, "loss": 0.4039, "num_input_tokens_seen": 3042408, "step": 11350 }, { "epoch": 2.9764089121887287, "grad_norm": 0.8013659119606018, "learning_rate": 4.428424740895998e-05, "loss": 0.4687, "num_input_tokens_seen": 3043640, "step": 11355 }, { "epoch": 2.977719528178244, "grad_norm": 0.8434219360351562, "learning_rate": 4.427696685285349e-05, "loss": 0.4738, "num_input_tokens_seen": 3045176, "step": 11360 }, { "epoch": 2.979030144167759, "grad_norm": 0.6642666459083557, "learning_rate": 4.4269682262111145e-05, "loss": 0.3777, "num_input_tokens_seen": 3046504, "step": 11365 }, { "epoch": 2.980340760157274, "grad_norm": 2.9187700748443604, "learning_rate": 4.42623936382576e-05, "loss": 0.3743, "num_input_tokens_seen": 3047752, "step": 11370 }, { "epoch": 2.981651376146789, "grad_norm": 1.021810531616211, "learning_rate": 4.425510098281835e-05, "loss": 0.6287, "num_input_tokens_seen": 3049368, "step": 11375 }, { "epoch": 2.982961992136304, "grad_norm": 0.8792248964309692, "learning_rate": 4.424780429731973e-05, "loss": 0.4583, "num_input_tokens_seen": 3050712, "step": 11380 }, { "epoch": 2.9842726081258193, "grad_norm": 0.8930190801620483, "learning_rate": 4.424050358328893e-05, "loss": 0.504, "num_input_tokens_seen": 3052488, "step": 11385 }, { "epoch": 2.9855832241153344, "grad_norm": 1.8062995672225952, "learning_rate": 4.423319884225398e-05, "loss": 0.4805, "num_input_tokens_seen": 3053736, "step": 11390 }, { "epoch": 2.986893840104849, "grad_norm": 1.2413045167922974, "learning_rate": 4.422589007574374e-05, "loss": 0.3228, "num_input_tokens_seen": 3054856, "step": 11395 }, { "epoch": 2.9882044560943646, "grad_norm": 1.0496180057525635, "learning_rate": 4.421857728528792e-05, "loss": 0.3876, "num_input_tokens_seen": 3055848, "step": 11400 }, { "epoch": 2.9895150720838792, "grad_norm": 0.5235907435417175, "learning_rate": 4.421126047241709e-05, "loss": 0.3885, "num_input_tokens_seen": 3057112, "step": 11405 }, { "epoch": 2.9908256880733948, "grad_norm": 0.747406542301178, "learning_rate": 4.420393963866263e-05, "loss": 0.5226, "num_input_tokens_seen": 3058312, "step": 11410 }, { "epoch": 2.9921363040629094, "grad_norm": 0.8095989227294922, "learning_rate": 4.4196614785556795e-05, "loss": 0.4006, "num_input_tokens_seen": 3060008, "step": 11415 }, { "epoch": 2.9934469200524245, "grad_norm": 4.557903289794922, "learning_rate": 4.418928591463265e-05, "loss": 0.4781, "num_input_tokens_seen": 3061016, "step": 11420 }, { "epoch": 2.9947575360419396, "grad_norm": 0.5235884785652161, "learning_rate": 4.418195302742412e-05, "loss": 0.2844, "num_input_tokens_seen": 3062632, "step": 11425 }, { "epoch": 2.9960681520314547, "grad_norm": 0.5899786949157715, "learning_rate": 4.417461612546596e-05, "loss": 0.4836, "num_input_tokens_seen": 3064008, "step": 11430 }, { "epoch": 2.99737876802097, "grad_norm": 2.45700740814209, "learning_rate": 4.416727521029379e-05, "loss": 0.4982, "num_input_tokens_seen": 3065256, "step": 11435 }, { "epoch": 2.998689384010485, "grad_norm": 1.019124150276184, "learning_rate": 4.415993028344403e-05, "loss": 0.3908, "num_input_tokens_seen": 3066456, "step": 11440 }, { "epoch": 3.0, "grad_norm": 5.559087753295898, "learning_rate": 4.415258134645396e-05, "loss": 0.4155, "num_input_tokens_seen": 3067768, "step": 11445 }, { "epoch": 3.000786369593709, "eval_loss": 0.49623367190361023, "eval_runtime": 16.685, "eval_samples_per_second": 50.824, "eval_steps_per_second": 25.412, "num_input_tokens_seen": 3068424, "step": 11448 }, { "epoch": 3.001310615989515, "grad_norm": 0.8290225267410278, "learning_rate": 4.414522840086172e-05, "loss": 0.596, "num_input_tokens_seen": 3068920, "step": 11450 }, { "epoch": 3.00262123197903, "grad_norm": 0.8040452003479004, "learning_rate": 4.413787144820625e-05, "loss": 0.4509, "num_input_tokens_seen": 3070328, "step": 11455 }, { "epoch": 3.0039318479685453, "grad_norm": 2.1803174018859863, "learning_rate": 4.413051049002735e-05, "loss": 0.3851, "num_input_tokens_seen": 3071368, "step": 11460 }, { "epoch": 3.0052424639580604, "grad_norm": 1.0317872762680054, "learning_rate": 4.412314552786566e-05, "loss": 0.446, "num_input_tokens_seen": 3072840, "step": 11465 }, { "epoch": 3.0065530799475755, "grad_norm": 0.916805624961853, "learning_rate": 4.4115776563262656e-05, "loss": 0.3448, "num_input_tokens_seen": 3074440, "step": 11470 }, { "epoch": 3.0078636959370906, "grad_norm": 2.381649971008301, "learning_rate": 4.410840359776065e-05, "loss": 0.3528, "num_input_tokens_seen": 3075592, "step": 11475 }, { "epoch": 3.0091743119266057, "grad_norm": 1.7830843925476074, "learning_rate": 4.410102663290278e-05, "loss": 0.421, "num_input_tokens_seen": 3076616, "step": 11480 }, { "epoch": 3.0104849279161208, "grad_norm": 0.5519474148750305, "learning_rate": 4.4093645670233045e-05, "loss": 0.3752, "num_input_tokens_seen": 3078088, "step": 11485 }, { "epoch": 3.011795543905636, "grad_norm": 1.2938812971115112, "learning_rate": 4.4086260711296265e-05, "loss": 0.4526, "num_input_tokens_seen": 3079064, "step": 11490 }, { "epoch": 3.0131061598951505, "grad_norm": 1.1425122022628784, "learning_rate": 4.407887175763809e-05, "loss": 0.3522, "num_input_tokens_seen": 3080184, "step": 11495 }, { "epoch": 3.0144167758846656, "grad_norm": 0.47444406151771545, "learning_rate": 4.407147881080503e-05, "loss": 0.3721, "num_input_tokens_seen": 3082008, "step": 11500 }, { "epoch": 3.0157273918741807, "grad_norm": 0.9128749966621399, "learning_rate": 4.40640818723444e-05, "loss": 0.4344, "num_input_tokens_seen": 3083352, "step": 11505 }, { "epoch": 3.017038007863696, "grad_norm": 1.7618298530578613, "learning_rate": 4.4056680943804374e-05, "loss": 0.3651, "num_input_tokens_seen": 3084248, "step": 11510 }, { "epoch": 3.018348623853211, "grad_norm": 0.7683572173118591, "learning_rate": 4.404927602673396e-05, "loss": 0.424, "num_input_tokens_seen": 3085864, "step": 11515 }, { "epoch": 3.019659239842726, "grad_norm": 1.32582426071167, "learning_rate": 4.404186712268299e-05, "loss": 0.483, "num_input_tokens_seen": 3087112, "step": 11520 }, { "epoch": 3.020969855832241, "grad_norm": 1.0521025657653809, "learning_rate": 4.4034454233202124e-05, "loss": 0.2853, "num_input_tokens_seen": 3088632, "step": 11525 }, { "epoch": 3.022280471821756, "grad_norm": 0.6893375515937805, "learning_rate": 4.4027037359842884e-05, "loss": 0.3726, "num_input_tokens_seen": 3089560, "step": 11530 }, { "epoch": 3.0235910878112713, "grad_norm": 0.9906198382377625, "learning_rate": 4.401961650415759e-05, "loss": 0.3251, "num_input_tokens_seen": 3090600, "step": 11535 }, { "epoch": 3.0249017038007864, "grad_norm": 0.8933149576187134, "learning_rate": 4.401219166769943e-05, "loss": 0.3926, "num_input_tokens_seen": 3091704, "step": 11540 }, { "epoch": 3.0262123197903015, "grad_norm": 2.0617618560791016, "learning_rate": 4.400476285202241e-05, "loss": 0.4219, "num_input_tokens_seen": 3092744, "step": 11545 }, { "epoch": 3.0275229357798166, "grad_norm": 0.7230196595191956, "learning_rate": 4.399733005868136e-05, "loss": 0.5048, "num_input_tokens_seen": 3094408, "step": 11550 }, { "epoch": 3.0288335517693317, "grad_norm": 1.3118135929107666, "learning_rate": 4.3989893289231954e-05, "loss": 0.3346, "num_input_tokens_seen": 3095800, "step": 11555 }, { "epoch": 3.0301441677588468, "grad_norm": 0.9546906352043152, "learning_rate": 4.398245254523069e-05, "loss": 0.5449, "num_input_tokens_seen": 3096888, "step": 11560 }, { "epoch": 3.031454783748362, "grad_norm": 1.027496337890625, "learning_rate": 4.397500782823492e-05, "loss": 0.6476, "num_input_tokens_seen": 3098856, "step": 11565 }, { "epoch": 3.032765399737877, "grad_norm": 1.1368443965911865, "learning_rate": 4.3967559139802785e-05, "loss": 0.3666, "num_input_tokens_seen": 3100184, "step": 11570 }, { "epoch": 3.034076015727392, "grad_norm": 0.3887004256248474, "learning_rate": 4.39601064814933e-05, "loss": 0.4111, "num_input_tokens_seen": 3101688, "step": 11575 }, { "epoch": 3.035386631716907, "grad_norm": 0.6131666898727417, "learning_rate": 4.395264985486629e-05, "loss": 0.3848, "num_input_tokens_seen": 3103048, "step": 11580 }, { "epoch": 3.036697247706422, "grad_norm": 0.5541452765464783, "learning_rate": 4.394518926148241e-05, "loss": 0.3139, "num_input_tokens_seen": 3104376, "step": 11585 }, { "epoch": 3.038007863695937, "grad_norm": 0.7444257140159607, "learning_rate": 4.393772470290315e-05, "loss": 0.4225, "num_input_tokens_seen": 3105720, "step": 11590 }, { "epoch": 3.039318479685452, "grad_norm": 1.3128618001937866, "learning_rate": 4.393025618069084e-05, "loss": 0.4866, "num_input_tokens_seen": 3107080, "step": 11595 }, { "epoch": 3.040629095674967, "grad_norm": 1.34038507938385, "learning_rate": 4.3922783696408615e-05, "loss": 0.3524, "num_input_tokens_seen": 3108648, "step": 11600 }, { "epoch": 3.041939711664482, "grad_norm": 0.5209506154060364, "learning_rate": 4.3915307251620456e-05, "loss": 0.4254, "num_input_tokens_seen": 3110040, "step": 11605 }, { "epoch": 3.0432503276539973, "grad_norm": 2.999112606048584, "learning_rate": 4.390782684789118e-05, "loss": 0.4557, "num_input_tokens_seen": 3111672, "step": 11610 }, { "epoch": 3.0445609436435124, "grad_norm": 1.3946667909622192, "learning_rate": 4.3900342486786405e-05, "loss": 0.4702, "num_input_tokens_seen": 3112936, "step": 11615 }, { "epoch": 3.0458715596330275, "grad_norm": 2.2576050758361816, "learning_rate": 4.389285416987261e-05, "loss": 0.4388, "num_input_tokens_seen": 3113928, "step": 11620 }, { "epoch": 3.0471821756225426, "grad_norm": 1.5710160732269287, "learning_rate": 4.388536189871708e-05, "loss": 0.3888, "num_input_tokens_seen": 3115384, "step": 11625 }, { "epoch": 3.0484927916120577, "grad_norm": 0.8613280057907104, "learning_rate": 4.387786567488794e-05, "loss": 0.9137, "num_input_tokens_seen": 3116616, "step": 11630 }, { "epoch": 3.0498034076015728, "grad_norm": 1.191860556602478, "learning_rate": 4.387036549995413e-05, "loss": 0.3724, "num_input_tokens_seen": 3117768, "step": 11635 }, { "epoch": 3.051114023591088, "grad_norm": 1.1306427717208862, "learning_rate": 4.386286137548543e-05, "loss": 0.3889, "num_input_tokens_seen": 3119064, "step": 11640 }, { "epoch": 3.052424639580603, "grad_norm": 1.464879035949707, "learning_rate": 4.3855353303052436e-05, "loss": 0.502, "num_input_tokens_seen": 3120520, "step": 11645 }, { "epoch": 3.053735255570118, "grad_norm": 1.7546124458312988, "learning_rate": 4.384784128422657e-05, "loss": 0.4218, "num_input_tokens_seen": 3121416, "step": 11650 }, { "epoch": 3.055045871559633, "grad_norm": 0.9712050557136536, "learning_rate": 4.3840325320580086e-05, "loss": 0.5726, "num_input_tokens_seen": 3122744, "step": 11655 }, { "epoch": 3.0563564875491482, "grad_norm": 0.6443157196044922, "learning_rate": 4.383280541368608e-05, "loss": 0.2981, "num_input_tokens_seen": 3125080, "step": 11660 }, { "epoch": 3.0576671035386633, "grad_norm": 0.4667041003704071, "learning_rate": 4.382528156511843e-05, "loss": 0.3537, "num_input_tokens_seen": 3126104, "step": 11665 }, { "epoch": 3.0589777195281784, "grad_norm": 1.286192536354065, "learning_rate": 4.381775377645188e-05, "loss": 0.4198, "num_input_tokens_seen": 3127304, "step": 11670 }, { "epoch": 3.0602883355176935, "grad_norm": 1.2646572589874268, "learning_rate": 4.381022204926198e-05, "loss": 0.3719, "num_input_tokens_seen": 3128712, "step": 11675 }, { "epoch": 3.061598951507208, "grad_norm": 0.6576411724090576, "learning_rate": 4.38026863851251e-05, "loss": 0.6207, "num_input_tokens_seen": 3129720, "step": 11680 }, { "epoch": 3.0629095674967233, "grad_norm": 0.5319696068763733, "learning_rate": 4.379514678561846e-05, "loss": 0.4519, "num_input_tokens_seen": 3131224, "step": 11685 }, { "epoch": 3.0642201834862384, "grad_norm": 0.5498784780502319, "learning_rate": 4.3787603252320065e-05, "loss": 0.2679, "num_input_tokens_seen": 3132888, "step": 11690 }, { "epoch": 3.0655307994757535, "grad_norm": 0.8331080675125122, "learning_rate": 4.378005578680877e-05, "loss": 0.3908, "num_input_tokens_seen": 3134472, "step": 11695 }, { "epoch": 3.0668414154652686, "grad_norm": 1.8894786834716797, "learning_rate": 4.377250439066426e-05, "loss": 0.5317, "num_input_tokens_seen": 3135416, "step": 11700 }, { "epoch": 3.0681520314547837, "grad_norm": 1.0051054954528809, "learning_rate": 4.3764949065467006e-05, "loss": 0.7469, "num_input_tokens_seen": 3136616, "step": 11705 }, { "epoch": 3.0694626474442988, "grad_norm": 1.7835105657577515, "learning_rate": 4.375738981279834e-05, "loss": 0.4443, "num_input_tokens_seen": 3137720, "step": 11710 }, { "epoch": 3.070773263433814, "grad_norm": 0.8046889305114746, "learning_rate": 4.37498266342404e-05, "loss": 0.4423, "num_input_tokens_seen": 3138680, "step": 11715 }, { "epoch": 3.072083879423329, "grad_norm": 0.7626618146896362, "learning_rate": 4.3742259531376135e-05, "loss": 0.2513, "num_input_tokens_seen": 3139672, "step": 11720 }, { "epoch": 3.073394495412844, "grad_norm": 2.222651481628418, "learning_rate": 4.373468850578933e-05, "loss": 0.2509, "num_input_tokens_seen": 3140824, "step": 11725 }, { "epoch": 3.074705111402359, "grad_norm": 1.9595853090286255, "learning_rate": 4.37271135590646e-05, "loss": 0.4485, "num_input_tokens_seen": 3142040, "step": 11730 }, { "epoch": 3.0760157273918742, "grad_norm": 0.46594345569610596, "learning_rate": 4.371953469278737e-05, "loss": 0.1889, "num_input_tokens_seen": 3143080, "step": 11735 }, { "epoch": 3.0773263433813893, "grad_norm": 0.7599170804023743, "learning_rate": 4.371195190854386e-05, "loss": 0.3818, "num_input_tokens_seen": 3144504, "step": 11740 }, { "epoch": 3.0786369593709044, "grad_norm": 0.9280793070793152, "learning_rate": 4.370436520792115e-05, "loss": 0.4638, "num_input_tokens_seen": 3145816, "step": 11745 }, { "epoch": 3.0799475753604195, "grad_norm": 3.7168097496032715, "learning_rate": 4.369677459250712e-05, "loss": 0.5184, "num_input_tokens_seen": 3147528, "step": 11750 }, { "epoch": 3.0812581913499346, "grad_norm": 1.5560232400894165, "learning_rate": 4.3689180063890476e-05, "loss": 0.5212, "num_input_tokens_seen": 3148648, "step": 11755 }, { "epoch": 3.0825688073394497, "grad_norm": 0.845289945602417, "learning_rate": 4.368158162366073e-05, "loss": 0.3628, "num_input_tokens_seen": 3149976, "step": 11760 }, { "epoch": 3.083879423328965, "grad_norm": 2.9016313552856445, "learning_rate": 4.367397927340824e-05, "loss": 0.5579, "num_input_tokens_seen": 3151624, "step": 11765 }, { "epoch": 3.08519003931848, "grad_norm": 1.983708143234253, "learning_rate": 4.366637301472415e-05, "loss": 0.4522, "num_input_tokens_seen": 3153160, "step": 11770 }, { "epoch": 3.0865006553079946, "grad_norm": 0.8780083656311035, "learning_rate": 4.365876284920044e-05, "loss": 0.4055, "num_input_tokens_seen": 3154600, "step": 11775 }, { "epoch": 3.0878112712975097, "grad_norm": 1.2206320762634277, "learning_rate": 4.365114877842991e-05, "loss": 0.297, "num_input_tokens_seen": 3155912, "step": 11780 }, { "epoch": 3.0891218872870247, "grad_norm": 1.8751294612884521, "learning_rate": 4.364353080400616e-05, "loss": 0.4721, "num_input_tokens_seen": 3156936, "step": 11785 }, { "epoch": 3.09043250327654, "grad_norm": 1.1428658962249756, "learning_rate": 4.363590892752363e-05, "loss": 0.39, "num_input_tokens_seen": 3158216, "step": 11790 }, { "epoch": 3.091743119266055, "grad_norm": 0.6433303952217102, "learning_rate": 4.362828315057756e-05, "loss": 0.4473, "num_input_tokens_seen": 3159544, "step": 11795 }, { "epoch": 3.09305373525557, "grad_norm": 0.9364408850669861, "learning_rate": 4.362065347476401e-05, "loss": 0.3607, "num_input_tokens_seen": 3161384, "step": 11800 }, { "epoch": 3.094364351245085, "grad_norm": 0.9295558929443359, "learning_rate": 4.361301990167988e-05, "loss": 0.2538, "num_input_tokens_seen": 3162360, "step": 11805 }, { "epoch": 3.0956749672346002, "grad_norm": 1.3614798784255981, "learning_rate": 4.360538243292282e-05, "loss": 0.4715, "num_input_tokens_seen": 3163352, "step": 11810 }, { "epoch": 3.0969855832241153, "grad_norm": 0.6146789193153381, "learning_rate": 4.359774107009138e-05, "loss": 0.3396, "num_input_tokens_seen": 3164648, "step": 11815 }, { "epoch": 3.0982961992136304, "grad_norm": 1.424957036972046, "learning_rate": 4.3590095814784866e-05, "loss": 0.3853, "num_input_tokens_seen": 3165752, "step": 11820 }, { "epoch": 3.0996068152031455, "grad_norm": 1.0626883506774902, "learning_rate": 4.3582446668603414e-05, "loss": 0.2994, "num_input_tokens_seen": 3167384, "step": 11825 }, { "epoch": 3.1009174311926606, "grad_norm": 0.9283444285392761, "learning_rate": 4.357479363314799e-05, "loss": 0.4065, "num_input_tokens_seen": 3169288, "step": 11830 }, { "epoch": 3.1022280471821757, "grad_norm": 1.2955172061920166, "learning_rate": 4.356713671002035e-05, "loss": 0.312, "num_input_tokens_seen": 3170264, "step": 11835 }, { "epoch": 3.103538663171691, "grad_norm": 1.2753205299377441, "learning_rate": 4.355947590082308e-05, "loss": 0.3732, "num_input_tokens_seen": 3171448, "step": 11840 }, { "epoch": 3.104849279161206, "grad_norm": 0.9723863005638123, "learning_rate": 4.355181120715957e-05, "loss": 0.4351, "num_input_tokens_seen": 3173016, "step": 11845 }, { "epoch": 3.106159895150721, "grad_norm": 0.6143611073493958, "learning_rate": 4.3544142630634024e-05, "loss": 0.3103, "num_input_tokens_seen": 3174872, "step": 11850 }, { "epoch": 3.107470511140236, "grad_norm": 0.7917157411575317, "learning_rate": 4.3536470172851476e-05, "loss": 0.3615, "num_input_tokens_seen": 3176840, "step": 11855 }, { "epoch": 3.108781127129751, "grad_norm": 1.1499496698379517, "learning_rate": 4.352879383541775e-05, "loss": 0.4216, "num_input_tokens_seen": 3178328, "step": 11860 }, { "epoch": 3.1100917431192663, "grad_norm": 0.6670563220977783, "learning_rate": 4.352111361993948e-05, "loss": 0.4607, "num_input_tokens_seen": 3179672, "step": 11865 }, { "epoch": 3.111402359108781, "grad_norm": 2.2965030670166016, "learning_rate": 4.351342952802414e-05, "loss": 0.4236, "num_input_tokens_seen": 3181128, "step": 11870 }, { "epoch": 3.112712975098296, "grad_norm": 1.0013840198516846, "learning_rate": 4.3505741561279984e-05, "loss": 0.2298, "num_input_tokens_seen": 3182648, "step": 11875 }, { "epoch": 3.114023591087811, "grad_norm": 2.408662796020508, "learning_rate": 4.349804972131609e-05, "loss": 0.3832, "num_input_tokens_seen": 3184312, "step": 11880 }, { "epoch": 3.1153342070773262, "grad_norm": 2.1826703548431396, "learning_rate": 4.349035400974236e-05, "loss": 0.4891, "num_input_tokens_seen": 3185448, "step": 11885 }, { "epoch": 3.1166448230668413, "grad_norm": 1.1475571393966675, "learning_rate": 4.348265442816949e-05, "loss": 0.3178, "num_input_tokens_seen": 3186920, "step": 11890 }, { "epoch": 3.1179554390563564, "grad_norm": 0.8346600532531738, "learning_rate": 4.347495097820897e-05, "loss": 0.4886, "num_input_tokens_seen": 3188568, "step": 11895 }, { "epoch": 3.1192660550458715, "grad_norm": 0.8939752578735352, "learning_rate": 4.346724366147314e-05, "loss": 0.377, "num_input_tokens_seen": 3189800, "step": 11900 }, { "epoch": 3.1205766710353866, "grad_norm": 2.2865500450134277, "learning_rate": 4.345953247957512e-05, "loss": 0.3152, "num_input_tokens_seen": 3190856, "step": 11905 }, { "epoch": 3.1218872870249017, "grad_norm": 0.7817928194999695, "learning_rate": 4.345181743412885e-05, "loss": 0.3502, "num_input_tokens_seen": 3192552, "step": 11910 }, { "epoch": 3.123197903014417, "grad_norm": 0.9188166260719299, "learning_rate": 4.344409852674907e-05, "loss": 0.4716, "num_input_tokens_seen": 3193672, "step": 11915 }, { "epoch": 3.124508519003932, "grad_norm": 0.949137270450592, "learning_rate": 4.343637575905133e-05, "loss": 0.3606, "num_input_tokens_seen": 3194760, "step": 11920 }, { "epoch": 3.125819134993447, "grad_norm": 1.2605834007263184, "learning_rate": 4.342864913265201e-05, "loss": 0.2471, "num_input_tokens_seen": 3195768, "step": 11925 }, { "epoch": 3.127129750982962, "grad_norm": 5.929020881652832, "learning_rate": 4.3420918649168265e-05, "loss": 0.4573, "num_input_tokens_seen": 3196936, "step": 11930 }, { "epoch": 3.128440366972477, "grad_norm": 8.03878116607666, "learning_rate": 4.341318431021808e-05, "loss": 0.4404, "num_input_tokens_seen": 3198360, "step": 11935 }, { "epoch": 3.1297509829619923, "grad_norm": 1.7889974117279053, "learning_rate": 4.340544611742023e-05, "loss": 0.4376, "num_input_tokens_seen": 3199816, "step": 11940 }, { "epoch": 3.1310615989515074, "grad_norm": 0.7878460884094238, "learning_rate": 4.339770407239431e-05, "loss": 0.4257, "num_input_tokens_seen": 3200808, "step": 11945 }, { "epoch": 3.1323722149410225, "grad_norm": 0.5995755195617676, "learning_rate": 4.3389958176760714e-05, "loss": 0.535, "num_input_tokens_seen": 3202984, "step": 11950 }, { "epoch": 3.133682830930537, "grad_norm": 0.6746127009391785, "learning_rate": 4.3382208432140646e-05, "loss": 0.2994, "num_input_tokens_seen": 3204296, "step": 11955 }, { "epoch": 3.134993446920052, "grad_norm": 1.6235499382019043, "learning_rate": 4.337445484015613e-05, "loss": 0.3724, "num_input_tokens_seen": 3205704, "step": 11960 }, { "epoch": 3.1363040629095673, "grad_norm": 2.882927179336548, "learning_rate": 4.3366697402429945e-05, "loss": 0.449, "num_input_tokens_seen": 3207240, "step": 11965 }, { "epoch": 3.1376146788990824, "grad_norm": 1.1278483867645264, "learning_rate": 4.335893612058574e-05, "loss": 0.3877, "num_input_tokens_seen": 3208264, "step": 11970 }, { "epoch": 3.1389252948885975, "grad_norm": 2.7744648456573486, "learning_rate": 4.335117099624792e-05, "loss": 0.4427, "num_input_tokens_seen": 3209656, "step": 11975 }, { "epoch": 3.1402359108781126, "grad_norm": 0.939441442489624, "learning_rate": 4.3343402031041726e-05, "loss": 0.3482, "num_input_tokens_seen": 3210616, "step": 11980 }, { "epoch": 3.1415465268676277, "grad_norm": 2.0664098262786865, "learning_rate": 4.333562922659318e-05, "loss": 0.4225, "num_input_tokens_seen": 3211640, "step": 11985 }, { "epoch": 3.142857142857143, "grad_norm": 0.9291212558746338, "learning_rate": 4.332785258452911e-05, "loss": 0.3735, "num_input_tokens_seen": 3213096, "step": 11990 }, { "epoch": 3.144167758846658, "grad_norm": 1.3390967845916748, "learning_rate": 4.332007210647717e-05, "loss": 0.3963, "num_input_tokens_seen": 3214072, "step": 11995 }, { "epoch": 3.145478374836173, "grad_norm": 1.1319376230239868, "learning_rate": 4.3312287794065795e-05, "loss": 0.3891, "num_input_tokens_seen": 3215416, "step": 12000 }, { "epoch": 3.146788990825688, "grad_norm": 0.8100825548171997, "learning_rate": 4.330449964892422e-05, "loss": 0.2425, "num_input_tokens_seen": 3216680, "step": 12005 }, { "epoch": 3.148099606815203, "grad_norm": 1.24164879322052, "learning_rate": 4.32967076726825e-05, "loss": 0.4162, "num_input_tokens_seen": 3217704, "step": 12010 }, { "epoch": 3.1494102228047183, "grad_norm": 2.112908124923706, "learning_rate": 4.328891186697148e-05, "loss": 0.5105, "num_input_tokens_seen": 3218840, "step": 12015 }, { "epoch": 3.1507208387942334, "grad_norm": 1.8827714920043945, "learning_rate": 4.328111223342281e-05, "loss": 0.2529, "num_input_tokens_seen": 3220264, "step": 12020 }, { "epoch": 3.1520314547837485, "grad_norm": 0.8639865517616272, "learning_rate": 4.327330877366893e-05, "loss": 0.8794, "num_input_tokens_seen": 3221416, "step": 12025 }, { "epoch": 3.1533420707732636, "grad_norm": 0.9369497895240784, "learning_rate": 4.3265501489343094e-05, "loss": 0.5436, "num_input_tokens_seen": 3222616, "step": 12030 }, { "epoch": 3.1546526867627787, "grad_norm": 1.0737946033477783, "learning_rate": 4.325769038207936e-05, "loss": 0.3308, "num_input_tokens_seen": 3223624, "step": 12035 }, { "epoch": 3.1559633027522938, "grad_norm": 0.6535418629646301, "learning_rate": 4.3249875453512586e-05, "loss": 0.3119, "num_input_tokens_seen": 3225256, "step": 12040 }, { "epoch": 3.157273918741809, "grad_norm": 1.87607741355896, "learning_rate": 4.32420567052784e-05, "loss": 0.4687, "num_input_tokens_seen": 3226632, "step": 12045 }, { "epoch": 3.1585845347313235, "grad_norm": 1.4278680086135864, "learning_rate": 4.323423413901327e-05, "loss": 0.3421, "num_input_tokens_seen": 3227768, "step": 12050 }, { "epoch": 3.1598951507208386, "grad_norm": 0.5801369547843933, "learning_rate": 4.322640775635445e-05, "loss": 0.3814, "num_input_tokens_seen": 3229128, "step": 12055 }, { "epoch": 3.1612057667103537, "grad_norm": 1.3262988328933716, "learning_rate": 4.321857755893996e-05, "loss": 0.4206, "num_input_tokens_seen": 3230216, "step": 12060 }, { "epoch": 3.162516382699869, "grad_norm": 1.5633172988891602, "learning_rate": 4.3210743548408684e-05, "loss": 0.3565, "num_input_tokens_seen": 3231256, "step": 12065 }, { "epoch": 3.163826998689384, "grad_norm": 1.2642590999603271, "learning_rate": 4.3202905726400236e-05, "loss": 0.5635, "num_input_tokens_seen": 3232808, "step": 12070 }, { "epoch": 3.165137614678899, "grad_norm": 0.9782804846763611, "learning_rate": 4.3195064094555085e-05, "loss": 0.3861, "num_input_tokens_seen": 3234776, "step": 12075 }, { "epoch": 3.166448230668414, "grad_norm": 9.464893341064453, "learning_rate": 4.318721865451445e-05, "loss": 0.4862, "num_input_tokens_seen": 3236088, "step": 12080 }, { "epoch": 3.167758846657929, "grad_norm": 1.1808239221572876, "learning_rate": 4.317936940792037e-05, "loss": 0.4195, "num_input_tokens_seen": 3237448, "step": 12085 }, { "epoch": 3.1690694626474443, "grad_norm": 0.9277758598327637, "learning_rate": 4.317151635641569e-05, "loss": 0.3095, "num_input_tokens_seen": 3238552, "step": 12090 }, { "epoch": 3.1703800786369594, "grad_norm": 2.3976047039031982, "learning_rate": 4.316365950164404e-05, "loss": 0.4139, "num_input_tokens_seen": 3239704, "step": 12095 }, { "epoch": 3.1716906946264745, "grad_norm": 0.5812339782714844, "learning_rate": 4.3155798845249827e-05, "loss": 0.4927, "num_input_tokens_seen": 3241256, "step": 12100 }, { "epoch": 3.1730013106159896, "grad_norm": 1.0861685276031494, "learning_rate": 4.3147934388878295e-05, "loss": 0.3851, "num_input_tokens_seen": 3242696, "step": 12105 }, { "epoch": 3.1743119266055047, "grad_norm": 1.5172971487045288, "learning_rate": 4.314006613417546e-05, "loss": 0.413, "num_input_tokens_seen": 3244280, "step": 12110 }, { "epoch": 3.1756225425950197, "grad_norm": 0.9100217819213867, "learning_rate": 4.3132194082788116e-05, "loss": 0.2828, "num_input_tokens_seen": 3245608, "step": 12115 }, { "epoch": 3.176933158584535, "grad_norm": 1.455655813217163, "learning_rate": 4.312431823636388e-05, "loss": 0.3254, "num_input_tokens_seen": 3246904, "step": 12120 }, { "epoch": 3.17824377457405, "grad_norm": 0.979512095451355, "learning_rate": 4.311643859655116e-05, "loss": 0.3588, "num_input_tokens_seen": 3248360, "step": 12125 }, { "epoch": 3.179554390563565, "grad_norm": 0.9720458388328552, "learning_rate": 4.3108555164999144e-05, "loss": 0.4271, "num_input_tokens_seen": 3249560, "step": 12130 }, { "epoch": 3.18086500655308, "grad_norm": 2.3408942222595215, "learning_rate": 4.310066794335782e-05, "loss": 0.4238, "num_input_tokens_seen": 3250792, "step": 12135 }, { "epoch": 3.1821756225425952, "grad_norm": 0.8741540312767029, "learning_rate": 4.309277693327797e-05, "loss": 0.3045, "num_input_tokens_seen": 3252120, "step": 12140 }, { "epoch": 3.18348623853211, "grad_norm": 1.034574270248413, "learning_rate": 4.308488213641116e-05, "loss": 0.2958, "num_input_tokens_seen": 3253432, "step": 12145 }, { "epoch": 3.184796854521625, "grad_norm": 1.4452708959579468, "learning_rate": 4.3076983554409784e-05, "loss": 0.3451, "num_input_tokens_seen": 3254600, "step": 12150 }, { "epoch": 3.18610747051114, "grad_norm": 1.2498011589050293, "learning_rate": 4.306908118892698e-05, "loss": 0.4181, "num_input_tokens_seen": 3255704, "step": 12155 }, { "epoch": 3.187418086500655, "grad_norm": 1.2280375957489014, "learning_rate": 4.30611750416167e-05, "loss": 0.3387, "num_input_tokens_seen": 3256904, "step": 12160 }, { "epoch": 3.1887287024901703, "grad_norm": 0.9341961741447449, "learning_rate": 4.305326511413369e-05, "loss": 0.2807, "num_input_tokens_seen": 3258184, "step": 12165 }, { "epoch": 3.1900393184796854, "grad_norm": 0.4049975872039795, "learning_rate": 4.304535140813349e-05, "loss": 0.4166, "num_input_tokens_seen": 3259432, "step": 12170 }, { "epoch": 3.1913499344692005, "grad_norm": 1.5861231088638306, "learning_rate": 4.303743392527242e-05, "loss": 0.4476, "num_input_tokens_seen": 3260616, "step": 12175 }, { "epoch": 3.1926605504587156, "grad_norm": 0.9539099335670471, "learning_rate": 4.302951266720759e-05, "loss": 0.5141, "num_input_tokens_seen": 3262200, "step": 12180 }, { "epoch": 3.1939711664482306, "grad_norm": 0.8138977885246277, "learning_rate": 4.3021587635596914e-05, "loss": 0.353, "num_input_tokens_seen": 3264136, "step": 12185 }, { "epoch": 3.1952817824377457, "grad_norm": 2.56648588180542, "learning_rate": 4.301365883209909e-05, "loss": 0.569, "num_input_tokens_seen": 3265112, "step": 12190 }, { "epoch": 3.196592398427261, "grad_norm": 1.6595916748046875, "learning_rate": 4.300572625837359e-05, "loss": 0.4812, "num_input_tokens_seen": 3268472, "step": 12195 }, { "epoch": 3.197903014416776, "grad_norm": 0.8588946461677551, "learning_rate": 4.29977899160807e-05, "loss": 0.4994, "num_input_tokens_seen": 3270168, "step": 12200 }, { "epoch": 3.199213630406291, "grad_norm": 0.6541647911071777, "learning_rate": 4.298984980688148e-05, "loss": 0.2946, "num_input_tokens_seen": 3271752, "step": 12205 }, { "epoch": 3.200524246395806, "grad_norm": 2.810349702835083, "learning_rate": 4.298190593243778e-05, "loss": 0.3259, "num_input_tokens_seen": 3272888, "step": 12210 }, { "epoch": 3.2018348623853212, "grad_norm": 0.8164893388748169, "learning_rate": 4.297395829441223e-05, "loss": 0.5364, "num_input_tokens_seen": 3273960, "step": 12215 }, { "epoch": 3.2031454783748363, "grad_norm": 1.1217353343963623, "learning_rate": 4.2966006894468274e-05, "loss": 0.4714, "num_input_tokens_seen": 3275240, "step": 12220 }, { "epoch": 3.2044560943643514, "grad_norm": 0.7755135297775269, "learning_rate": 4.295805173427012e-05, "loss": 0.3437, "num_input_tokens_seen": 3277096, "step": 12225 }, { "epoch": 3.2057667103538665, "grad_norm": 1.0968739986419678, "learning_rate": 4.295009281548276e-05, "loss": 0.3869, "num_input_tokens_seen": 3278744, "step": 12230 }, { "epoch": 3.2070773263433816, "grad_norm": 0.9663660526275635, "learning_rate": 4.294213013977199e-05, "loss": 0.3217, "num_input_tokens_seen": 3280616, "step": 12235 }, { "epoch": 3.2083879423328963, "grad_norm": 1.5205885171890259, "learning_rate": 4.2934163708804385e-05, "loss": 0.4601, "num_input_tokens_seen": 3282008, "step": 12240 }, { "epoch": 3.2096985583224114, "grad_norm": 1.838966727256775, "learning_rate": 4.29261935242473e-05, "loss": 0.419, "num_input_tokens_seen": 3283288, "step": 12245 }, { "epoch": 3.2110091743119265, "grad_norm": 0.9143408536911011, "learning_rate": 4.291821958776889e-05, "loss": 0.3811, "num_input_tokens_seen": 3284696, "step": 12250 }, { "epoch": 3.2123197903014415, "grad_norm": 0.8312178254127502, "learning_rate": 4.291024190103808e-05, "loss": 0.4152, "num_input_tokens_seen": 3285880, "step": 12255 }, { "epoch": 3.2136304062909566, "grad_norm": 1.99077570438385, "learning_rate": 4.290226046572459e-05, "loss": 0.5058, "num_input_tokens_seen": 3287112, "step": 12260 }, { "epoch": 3.2149410222804717, "grad_norm": 1.4216557741165161, "learning_rate": 4.289427528349891e-05, "loss": 0.4346, "num_input_tokens_seen": 3288792, "step": 12265 }, { "epoch": 3.216251638269987, "grad_norm": 0.9937379956245422, "learning_rate": 4.2886286356032334e-05, "loss": 0.3394, "num_input_tokens_seen": 3290104, "step": 12270 }, { "epoch": 3.217562254259502, "grad_norm": 1.0400047302246094, "learning_rate": 4.287829368499693e-05, "loss": 0.4039, "num_input_tokens_seen": 3291304, "step": 12275 }, { "epoch": 3.218872870249017, "grad_norm": 1.8716009855270386, "learning_rate": 4.2870297272065546e-05, "loss": 0.5122, "num_input_tokens_seen": 3292600, "step": 12280 }, { "epoch": 3.220183486238532, "grad_norm": 1.2077854871749878, "learning_rate": 4.286229711891182e-05, "loss": 0.5003, "num_input_tokens_seen": 3293672, "step": 12285 }, { "epoch": 3.221494102228047, "grad_norm": 1.1941382884979248, "learning_rate": 4.285429322721017e-05, "loss": 0.4837, "num_input_tokens_seen": 3295192, "step": 12290 }, { "epoch": 3.2228047182175623, "grad_norm": 0.600077748298645, "learning_rate": 4.284628559863581e-05, "loss": 0.3891, "num_input_tokens_seen": 3296216, "step": 12295 }, { "epoch": 3.2241153342070774, "grad_norm": 0.9203482270240784, "learning_rate": 4.2838274234864694e-05, "loss": 0.4773, "num_input_tokens_seen": 3297464, "step": 12300 }, { "epoch": 3.2254259501965925, "grad_norm": 1.1899926662445068, "learning_rate": 4.283025913757361e-05, "loss": 0.5577, "num_input_tokens_seen": 3298744, "step": 12305 }, { "epoch": 3.2267365661861076, "grad_norm": 0.7191978096961975, "learning_rate": 4.2822240308440096e-05, "loss": 0.2028, "num_input_tokens_seen": 3299848, "step": 12310 }, { "epoch": 3.2280471821756227, "grad_norm": 0.8685644865036011, "learning_rate": 4.281421774914248e-05, "loss": 0.2475, "num_input_tokens_seen": 3301432, "step": 12315 }, { "epoch": 3.229357798165138, "grad_norm": 1.607553482055664, "learning_rate": 4.280619146135988e-05, "loss": 0.4822, "num_input_tokens_seen": 3303176, "step": 12320 }, { "epoch": 3.230668414154653, "grad_norm": 1.1230361461639404, "learning_rate": 4.2798161446772156e-05, "loss": 0.4388, "num_input_tokens_seen": 3304824, "step": 12325 }, { "epoch": 3.231979030144168, "grad_norm": 0.8110120296478271, "learning_rate": 4.279012770706e-05, "loss": 0.2912, "num_input_tokens_seen": 3306312, "step": 12330 }, { "epoch": 3.2332896461336826, "grad_norm": 1.0608501434326172, "learning_rate": 4.278209024390486e-05, "loss": 0.4093, "num_input_tokens_seen": 3307864, "step": 12335 }, { "epoch": 3.2346002621231977, "grad_norm": 1.9760822057724, "learning_rate": 4.2774049058988955e-05, "loss": 0.4859, "num_input_tokens_seen": 3309304, "step": 12340 }, { "epoch": 3.235910878112713, "grad_norm": 1.066863775253296, "learning_rate": 4.276600415399529e-05, "loss": 0.328, "num_input_tokens_seen": 3310568, "step": 12345 }, { "epoch": 3.237221494102228, "grad_norm": 1.9405947923660278, "learning_rate": 4.275795553060765e-05, "loss": 0.3713, "num_input_tokens_seen": 3312024, "step": 12350 }, { "epoch": 3.238532110091743, "grad_norm": 0.9786284565925598, "learning_rate": 4.274990319051061e-05, "loss": 0.4072, "num_input_tokens_seen": 3313336, "step": 12355 }, { "epoch": 3.239842726081258, "grad_norm": 1.6610158681869507, "learning_rate": 4.2741847135389495e-05, "loss": 0.5071, "num_input_tokens_seen": 3314616, "step": 12360 }, { "epoch": 3.241153342070773, "grad_norm": 1.233563780784607, "learning_rate": 4.2733787366930436e-05, "loss": 0.4027, "num_input_tokens_seen": 3315864, "step": 12365 }, { "epoch": 3.2424639580602883, "grad_norm": 1.2495805025100708, "learning_rate": 4.2725723886820314e-05, "loss": 0.556, "num_input_tokens_seen": 3317144, "step": 12370 }, { "epoch": 3.2437745740498034, "grad_norm": 0.5736082792282104, "learning_rate": 4.271765669674682e-05, "loss": 0.3369, "num_input_tokens_seen": 3318728, "step": 12375 }, { "epoch": 3.2450851900393185, "grad_norm": 0.6864911913871765, "learning_rate": 4.270958579839839e-05, "loss": 0.3178, "num_input_tokens_seen": 3320008, "step": 12380 }, { "epoch": 3.2463958060288336, "grad_norm": 0.7428423166275024, "learning_rate": 4.270151119346424e-05, "loss": 0.4648, "num_input_tokens_seen": 3321496, "step": 12385 }, { "epoch": 3.2477064220183487, "grad_norm": 1.3102647066116333, "learning_rate": 4.26934328836344e-05, "loss": 0.4677, "num_input_tokens_seen": 3322632, "step": 12390 }, { "epoch": 3.249017038007864, "grad_norm": 0.7702975273132324, "learning_rate": 4.268535087059962e-05, "loss": 0.2702, "num_input_tokens_seen": 3324168, "step": 12395 }, { "epoch": 3.250327653997379, "grad_norm": 1.164160132408142, "learning_rate": 4.267726515605146e-05, "loss": 0.406, "num_input_tokens_seen": 3325592, "step": 12400 }, { "epoch": 3.251638269986894, "grad_norm": 1.917345643043518, "learning_rate": 4.2669175741682255e-05, "loss": 0.3787, "num_input_tokens_seen": 3326920, "step": 12405 }, { "epoch": 3.252948885976409, "grad_norm": 1.3492202758789062, "learning_rate": 4.266108262918509e-05, "loss": 0.362, "num_input_tokens_seen": 3327912, "step": 12410 }, { "epoch": 3.254259501965924, "grad_norm": 1.5545824766159058, "learning_rate": 4.265298582025385e-05, "loss": 0.6706, "num_input_tokens_seen": 3329352, "step": 12415 }, { "epoch": 3.255570117955439, "grad_norm": 1.6622095108032227, "learning_rate": 4.264488531658317e-05, "loss": 0.5163, "num_input_tokens_seen": 3330456, "step": 12420 }, { "epoch": 3.2568807339449544, "grad_norm": 1.1697379350662231, "learning_rate": 4.26367811198685e-05, "loss": 0.3548, "num_input_tokens_seen": 3331752, "step": 12425 }, { "epoch": 3.258191349934469, "grad_norm": 1.0058196783065796, "learning_rate": 4.2628673231806e-05, "loss": 0.4441, "num_input_tokens_seen": 3333128, "step": 12430 }, { "epoch": 3.259501965923984, "grad_norm": 0.9957892298698425, "learning_rate": 4.262056165409266e-05, "loss": 0.4199, "num_input_tokens_seen": 3334200, "step": 12435 }, { "epoch": 3.260812581913499, "grad_norm": 1.053330659866333, "learning_rate": 4.261244638842621e-05, "loss": 0.216, "num_input_tokens_seen": 3335496, "step": 12440 }, { "epoch": 3.2621231979030143, "grad_norm": 3.3475470542907715, "learning_rate": 4.260432743650516e-05, "loss": 0.6577, "num_input_tokens_seen": 3336488, "step": 12445 }, { "epoch": 3.2634338138925294, "grad_norm": 0.84101402759552, "learning_rate": 4.25962048000288e-05, "loss": 0.4633, "num_input_tokens_seen": 3337832, "step": 12450 }, { "epoch": 3.2647444298820445, "grad_norm": 3.7470293045043945, "learning_rate": 4.258807848069718e-05, "loss": 0.5391, "num_input_tokens_seen": 3339384, "step": 12455 }, { "epoch": 3.2660550458715596, "grad_norm": 1.4123915433883667, "learning_rate": 4.257994848021112e-05, "loss": 0.4882, "num_input_tokens_seen": 3340904, "step": 12460 }, { "epoch": 3.2673656618610747, "grad_norm": 1.1883420944213867, "learning_rate": 4.257181480027222e-05, "loss": 0.4778, "num_input_tokens_seen": 3342296, "step": 12465 }, { "epoch": 3.26867627785059, "grad_norm": 1.3919060230255127, "learning_rate": 4.256367744258285e-05, "loss": 0.6924, "num_input_tokens_seen": 3343336, "step": 12470 }, { "epoch": 3.269986893840105, "grad_norm": 1.2090609073638916, "learning_rate": 4.255553640884613e-05, "loss": 0.28, "num_input_tokens_seen": 3344472, "step": 12475 }, { "epoch": 3.27129750982962, "grad_norm": 1.5007638931274414, "learning_rate": 4.254739170076597e-05, "loss": 0.3965, "num_input_tokens_seen": 3345720, "step": 12480 }, { "epoch": 3.272608125819135, "grad_norm": 2.2019338607788086, "learning_rate": 4.253924332004706e-05, "loss": 0.4089, "num_input_tokens_seen": 3346920, "step": 12485 }, { "epoch": 3.27391874180865, "grad_norm": 1.1110869646072388, "learning_rate": 4.2531091268394816e-05, "loss": 0.3657, "num_input_tokens_seen": 3348232, "step": 12490 }, { "epoch": 3.2752293577981653, "grad_norm": 1.8245795965194702, "learning_rate": 4.252293554751547e-05, "loss": 0.369, "num_input_tokens_seen": 3349400, "step": 12495 }, { "epoch": 3.2765399737876804, "grad_norm": 1.610334038734436, "learning_rate": 4.251477615911599e-05, "loss": 0.593, "num_input_tokens_seen": 3350552, "step": 12500 }, { "epoch": 3.2778505897771955, "grad_norm": 4.5193681716918945, "learning_rate": 4.250661310490412e-05, "loss": 0.3633, "num_input_tokens_seen": 3351832, "step": 12505 }, { "epoch": 3.2791612057667106, "grad_norm": 0.3863404393196106, "learning_rate": 4.249844638658838e-05, "loss": 0.2605, "num_input_tokens_seen": 3353512, "step": 12510 }, { "epoch": 3.280471821756225, "grad_norm": 1.1126062870025635, "learning_rate": 4.249027600587804e-05, "loss": 0.3516, "num_input_tokens_seen": 3354744, "step": 12515 }, { "epoch": 3.2817824377457407, "grad_norm": 1.2054901123046875, "learning_rate": 4.248210196448316e-05, "loss": 0.2952, "num_input_tokens_seen": 3355960, "step": 12520 }, { "epoch": 3.2830930537352554, "grad_norm": 2.457953691482544, "learning_rate": 4.247392426411454e-05, "loss": 0.4209, "num_input_tokens_seen": 3356952, "step": 12525 }, { "epoch": 3.2844036697247705, "grad_norm": 0.6931677460670471, "learning_rate": 4.246574290648376e-05, "loss": 0.4611, "num_input_tokens_seen": 3358312, "step": 12530 }, { "epoch": 3.2857142857142856, "grad_norm": 0.824364185333252, "learning_rate": 4.245755789330318e-05, "loss": 0.4348, "num_input_tokens_seen": 3359912, "step": 12535 }, { "epoch": 3.2870249017038007, "grad_norm": 1.5108625888824463, "learning_rate": 4.2449369226285884e-05, "loss": 0.4414, "num_input_tokens_seen": 3360920, "step": 12540 }, { "epoch": 3.288335517693316, "grad_norm": 1.1532230377197266, "learning_rate": 4.2441176907145775e-05, "loss": 0.3522, "num_input_tokens_seen": 3362104, "step": 12545 }, { "epoch": 3.289646133682831, "grad_norm": 4.002058029174805, "learning_rate": 4.243298093759746e-05, "loss": 0.3267, "num_input_tokens_seen": 3363288, "step": 12550 }, { "epoch": 3.290956749672346, "grad_norm": 1.4848051071166992, "learning_rate": 4.242478131935637e-05, "loss": 0.3871, "num_input_tokens_seen": 3364424, "step": 12555 }, { "epoch": 3.292267365661861, "grad_norm": 1.2580984830856323, "learning_rate": 4.241657805413865e-05, "loss": 0.467, "num_input_tokens_seen": 3365736, "step": 12560 }, { "epoch": 3.293577981651376, "grad_norm": 1.3203423023223877, "learning_rate": 4.240837114366123e-05, "loss": 0.3476, "num_input_tokens_seen": 3366968, "step": 12565 }, { "epoch": 3.2948885976408913, "grad_norm": 0.7901044487953186, "learning_rate": 4.2400160589641825e-05, "loss": 0.3167, "num_input_tokens_seen": 3368216, "step": 12570 }, { "epoch": 3.2961992136304064, "grad_norm": 1.066307544708252, "learning_rate": 4.239194639379887e-05, "loss": 0.4032, "num_input_tokens_seen": 3369464, "step": 12575 }, { "epoch": 3.2975098296199215, "grad_norm": 1.2655882835388184, "learning_rate": 4.238372855785158e-05, "loss": 0.3968, "num_input_tokens_seen": 3370584, "step": 12580 }, { "epoch": 3.2988204456094365, "grad_norm": 0.7398244142532349, "learning_rate": 4.2375507083519936e-05, "loss": 0.4244, "num_input_tokens_seen": 3371848, "step": 12585 }, { "epoch": 3.3001310615989516, "grad_norm": 1.4887526035308838, "learning_rate": 4.2367281972524685e-05, "loss": 0.411, "num_input_tokens_seen": 3373128, "step": 12590 }, { "epoch": 3.3014416775884667, "grad_norm": 1.5324937105178833, "learning_rate": 4.235905322658733e-05, "loss": 0.4422, "num_input_tokens_seen": 3374296, "step": 12595 }, { "epoch": 3.302752293577982, "grad_norm": 0.5556575059890747, "learning_rate": 4.235082084743013e-05, "loss": 0.2818, "num_input_tokens_seen": 3375592, "step": 12600 }, { "epoch": 3.304062909567497, "grad_norm": 0.8782303929328918, "learning_rate": 4.23425848367761e-05, "loss": 0.3691, "num_input_tokens_seen": 3377336, "step": 12605 }, { "epoch": 3.3053735255570116, "grad_norm": 1.842785358428955, "learning_rate": 4.2334345196349036e-05, "loss": 0.4464, "num_input_tokens_seen": 3378440, "step": 12610 }, { "epoch": 3.3066841415465267, "grad_norm": 1.0687837600708008, "learning_rate": 4.2326101927873464e-05, "loss": 0.4332, "num_input_tokens_seen": 3380088, "step": 12615 }, { "epoch": 3.3079947575360418, "grad_norm": 1.7325464487075806, "learning_rate": 4.2317855033074706e-05, "loss": 0.5012, "num_input_tokens_seen": 3381208, "step": 12620 }, { "epoch": 3.309305373525557, "grad_norm": 1.4151911735534668, "learning_rate": 4.2309604513678804e-05, "loss": 0.5606, "num_input_tokens_seen": 3382568, "step": 12625 }, { "epoch": 3.310615989515072, "grad_norm": 1.2849715948104858, "learning_rate": 4.230135037141259e-05, "loss": 0.7137, "num_input_tokens_seen": 3383768, "step": 12630 }, { "epoch": 3.311926605504587, "grad_norm": 2.4282007217407227, "learning_rate": 4.229309260800364e-05, "loss": 0.2959, "num_input_tokens_seen": 3384712, "step": 12635 }, { "epoch": 3.313237221494102, "grad_norm": 1.135986328125, "learning_rate": 4.228483122518029e-05, "loss": 0.475, "num_input_tokens_seen": 3386040, "step": 12640 }, { "epoch": 3.3145478374836173, "grad_norm": 0.8256007432937622, "learning_rate": 4.227656622467162e-05, "loss": 0.3718, "num_input_tokens_seen": 3387208, "step": 12645 }, { "epoch": 3.3158584534731324, "grad_norm": 1.0377116203308105, "learning_rate": 4.2268297608207496e-05, "loss": 0.4565, "num_input_tokens_seen": 3388408, "step": 12650 }, { "epoch": 3.3171690694626474, "grad_norm": 0.8304136991500854, "learning_rate": 4.226002537751852e-05, "loss": 0.4693, "num_input_tokens_seen": 3389448, "step": 12655 }, { "epoch": 3.3184796854521625, "grad_norm": 1.292287826538086, "learning_rate": 4.225174953433605e-05, "loss": 0.4085, "num_input_tokens_seen": 3390568, "step": 12660 }, { "epoch": 3.3197903014416776, "grad_norm": 2.097417116165161, "learning_rate": 4.224347008039222e-05, "loss": 0.3276, "num_input_tokens_seen": 3391768, "step": 12665 }, { "epoch": 3.3211009174311927, "grad_norm": 1.4430811405181885, "learning_rate": 4.22351870174199e-05, "loss": 0.4063, "num_input_tokens_seen": 3393208, "step": 12670 }, { "epoch": 3.322411533420708, "grad_norm": 0.9457472562789917, "learning_rate": 4.222690034715271e-05, "loss": 0.2526, "num_input_tokens_seen": 3394248, "step": 12675 }, { "epoch": 3.323722149410223, "grad_norm": 1.4896221160888672, "learning_rate": 4.2218610071325036e-05, "loss": 0.4486, "num_input_tokens_seen": 3396232, "step": 12680 }, { "epoch": 3.325032765399738, "grad_norm": 0.681951105594635, "learning_rate": 4.221031619167204e-05, "loss": 0.3529, "num_input_tokens_seen": 3397432, "step": 12685 }, { "epoch": 3.326343381389253, "grad_norm": 0.891793429851532, "learning_rate": 4.2202018709929595e-05, "loss": 0.3266, "num_input_tokens_seen": 3398760, "step": 12690 }, { "epoch": 3.327653997378768, "grad_norm": 1.287151575088501, "learning_rate": 4.219371762783436e-05, "loss": 0.51, "num_input_tokens_seen": 3399848, "step": 12695 }, { "epoch": 3.3289646133682833, "grad_norm": 0.9530618190765381, "learning_rate": 4.218541294712373e-05, "loss": 0.4549, "num_input_tokens_seen": 3401240, "step": 12700 }, { "epoch": 3.330275229357798, "grad_norm": 2.0875966548919678, "learning_rate": 4.217710466953586e-05, "loss": 1.0398, "num_input_tokens_seen": 3402920, "step": 12705 }, { "epoch": 3.331585845347313, "grad_norm": 0.3703330159187317, "learning_rate": 4.2168792796809674e-05, "loss": 0.3626, "num_input_tokens_seen": 3404184, "step": 12710 }, { "epoch": 3.332896461336828, "grad_norm": 0.622894287109375, "learning_rate": 4.216047733068481e-05, "loss": 0.3415, "num_input_tokens_seen": 3405576, "step": 12715 }, { "epoch": 3.3342070773263432, "grad_norm": 1.8503421545028687, "learning_rate": 4.21521582729017e-05, "loss": 0.4928, "num_input_tokens_seen": 3406712, "step": 12720 }, { "epoch": 3.3355176933158583, "grad_norm": 2.2909958362579346, "learning_rate": 4.21438356252015e-05, "loss": 0.4823, "num_input_tokens_seen": 3407928, "step": 12725 }, { "epoch": 3.3368283093053734, "grad_norm": 1.2955501079559326, "learning_rate": 4.213550938932612e-05, "loss": 0.518, "num_input_tokens_seen": 3409272, "step": 12730 }, { "epoch": 3.3381389252948885, "grad_norm": 0.8777226805686951, "learning_rate": 4.2127179567018235e-05, "loss": 0.3369, "num_input_tokens_seen": 3410568, "step": 12735 }, { "epoch": 3.3394495412844036, "grad_norm": 1.5804678201675415, "learning_rate": 4.211884616002126e-05, "loss": 0.3476, "num_input_tokens_seen": 3412248, "step": 12740 }, { "epoch": 3.3407601572739187, "grad_norm": 0.6606892347335815, "learning_rate": 4.211050917007937e-05, "loss": 0.4204, "num_input_tokens_seen": 3414104, "step": 12745 }, { "epoch": 3.342070773263434, "grad_norm": 1.223755121231079, "learning_rate": 4.210216859893747e-05, "loss": 0.3167, "num_input_tokens_seen": 3415256, "step": 12750 }, { "epoch": 3.343381389252949, "grad_norm": 3.7373077869415283, "learning_rate": 4.2093824448341226e-05, "loss": 0.4368, "num_input_tokens_seen": 3416264, "step": 12755 }, { "epoch": 3.344692005242464, "grad_norm": 1.769940972328186, "learning_rate": 4.208547672003707e-05, "loss": 0.3894, "num_input_tokens_seen": 3417592, "step": 12760 }, { "epoch": 3.346002621231979, "grad_norm": 0.7917900085449219, "learning_rate": 4.207712541577215e-05, "loss": 0.4253, "num_input_tokens_seen": 3419032, "step": 12765 }, { "epoch": 3.347313237221494, "grad_norm": 1.3272076845169067, "learning_rate": 4.2068770537294395e-05, "loss": 0.3894, "num_input_tokens_seen": 3420104, "step": 12770 }, { "epoch": 3.3486238532110093, "grad_norm": 1.2209746837615967, "learning_rate": 4.2060412086352455e-05, "loss": 0.3206, "num_input_tokens_seen": 3421704, "step": 12775 }, { "epoch": 3.3499344692005244, "grad_norm": 1.0302109718322754, "learning_rate": 4.2052050064695746e-05, "loss": 0.3482, "num_input_tokens_seen": 3422904, "step": 12780 }, { "epoch": 3.3512450851900395, "grad_norm": 0.5879092216491699, "learning_rate": 4.204368447407442e-05, "loss": 0.2897, "num_input_tokens_seen": 3424456, "step": 12785 }, { "epoch": 3.352555701179554, "grad_norm": 0.5557191371917725, "learning_rate": 4.203531531623938e-05, "loss": 0.2629, "num_input_tokens_seen": 3425576, "step": 12790 }, { "epoch": 3.3538663171690697, "grad_norm": 0.7454561591148376, "learning_rate": 4.2026942592942285e-05, "loss": 0.3628, "num_input_tokens_seen": 3426920, "step": 12795 }, { "epoch": 3.3551769331585843, "grad_norm": 0.8825490474700928, "learning_rate": 4.2018566305935525e-05, "loss": 0.422, "num_input_tokens_seen": 3428504, "step": 12800 }, { "epoch": 3.3564875491480994, "grad_norm": 0.6989990472793579, "learning_rate": 4.201018645697224e-05, "loss": 0.2942, "num_input_tokens_seen": 3429416, "step": 12805 }, { "epoch": 3.3577981651376145, "grad_norm": 1.4563369750976562, "learning_rate": 4.200180304780632e-05, "loss": 0.3333, "num_input_tokens_seen": 3430776, "step": 12810 }, { "epoch": 3.3591087811271296, "grad_norm": 1.6099334955215454, "learning_rate": 4.1993416080192404e-05, "loss": 0.5345, "num_input_tokens_seen": 3431832, "step": 12815 }, { "epoch": 3.3604193971166447, "grad_norm": 0.92412930727005, "learning_rate": 4.198502555588586e-05, "loss": 0.5276, "num_input_tokens_seen": 3432904, "step": 12820 }, { "epoch": 3.36173001310616, "grad_norm": 1.3629952669143677, "learning_rate": 4.197663147664281e-05, "loss": 0.3318, "num_input_tokens_seen": 3433976, "step": 12825 }, { "epoch": 3.363040629095675, "grad_norm": 1.2072886228561401, "learning_rate": 4.196823384422014e-05, "loss": 0.3641, "num_input_tokens_seen": 3435320, "step": 12830 }, { "epoch": 3.36435124508519, "grad_norm": 0.7115566730499268, "learning_rate": 4.195983266037543e-05, "loss": 0.3732, "num_input_tokens_seen": 3436552, "step": 12835 }, { "epoch": 3.365661861074705, "grad_norm": 1.3484517335891724, "learning_rate": 4.1951427926867056e-05, "loss": 0.3424, "num_input_tokens_seen": 3437656, "step": 12840 }, { "epoch": 3.36697247706422, "grad_norm": 3.7108097076416016, "learning_rate": 4.1943019645454115e-05, "loss": 0.4483, "num_input_tokens_seen": 3438616, "step": 12845 }, { "epoch": 3.3682830930537353, "grad_norm": 1.053591251373291, "learning_rate": 4.193460781789642e-05, "loss": 0.4466, "num_input_tokens_seen": 3439848, "step": 12850 }, { "epoch": 3.3695937090432504, "grad_norm": 0.564730703830719, "learning_rate": 4.192619244595458e-05, "loss": 0.4094, "num_input_tokens_seen": 3441176, "step": 12855 }, { "epoch": 3.3709043250327655, "grad_norm": 0.9811622500419617, "learning_rate": 4.191777353138991e-05, "loss": 0.4596, "num_input_tokens_seen": 3442696, "step": 12860 }, { "epoch": 3.3722149410222806, "grad_norm": 1.0767258405685425, "learning_rate": 4.1909351075964464e-05, "loss": 0.3724, "num_input_tokens_seen": 3443832, "step": 12865 }, { "epoch": 3.3735255570117957, "grad_norm": 0.6086859107017517, "learning_rate": 4.190092508144106e-05, "loss": 0.3895, "num_input_tokens_seen": 3444952, "step": 12870 }, { "epoch": 3.374836173001311, "grad_norm": 0.81880784034729, "learning_rate": 4.189249554958324e-05, "loss": 0.3898, "num_input_tokens_seen": 3446232, "step": 12875 }, { "epoch": 3.376146788990826, "grad_norm": 0.7379269003868103, "learning_rate": 4.188406248215528e-05, "loss": 0.3492, "num_input_tokens_seen": 3447528, "step": 12880 }, { "epoch": 3.3774574049803405, "grad_norm": 1.0684304237365723, "learning_rate": 4.1875625880922225e-05, "loss": 0.7531, "num_input_tokens_seen": 3449352, "step": 12885 }, { "epoch": 3.378768020969856, "grad_norm": 0.8594215512275696, "learning_rate": 4.186718574764983e-05, "loss": 0.5399, "num_input_tokens_seen": 3450536, "step": 12890 }, { "epoch": 3.3800786369593707, "grad_norm": 1.0020599365234375, "learning_rate": 4.18587420841046e-05, "loss": 0.3774, "num_input_tokens_seen": 3451688, "step": 12895 }, { "epoch": 3.381389252948886, "grad_norm": 0.9474236369132996, "learning_rate": 4.185029489205379e-05, "loss": 0.4063, "num_input_tokens_seen": 3452760, "step": 12900 }, { "epoch": 3.382699868938401, "grad_norm": 0.6056210994720459, "learning_rate": 4.1841844173265357e-05, "loss": 0.3203, "num_input_tokens_seen": 3454712, "step": 12905 }, { "epoch": 3.384010484927916, "grad_norm": 1.2588502168655396, "learning_rate": 4.183338992950805e-05, "loss": 0.46, "num_input_tokens_seen": 3456632, "step": 12910 }, { "epoch": 3.385321100917431, "grad_norm": 1.4912420511245728, "learning_rate": 4.182493216255132e-05, "loss": 0.3916, "num_input_tokens_seen": 3457992, "step": 12915 }, { "epoch": 3.386631716906946, "grad_norm": 0.9583026170730591, "learning_rate": 4.181647087416536e-05, "loss": 0.2606, "num_input_tokens_seen": 3459224, "step": 12920 }, { "epoch": 3.3879423328964613, "grad_norm": 1.0645703077316284, "learning_rate": 4.18080060661211e-05, "loss": 0.3607, "num_input_tokens_seen": 3460680, "step": 12925 }, { "epoch": 3.3892529488859764, "grad_norm": 1.192298412322998, "learning_rate": 4.179953774019021e-05, "loss": 0.4129, "num_input_tokens_seen": 3462520, "step": 12930 }, { "epoch": 3.3905635648754915, "grad_norm": 0.8138737082481384, "learning_rate": 4.1791065898145095e-05, "loss": 0.3745, "num_input_tokens_seen": 3463768, "step": 12935 }, { "epoch": 3.3918741808650066, "grad_norm": 0.8071034550666809, "learning_rate": 4.178259054175891e-05, "loss": 0.3563, "num_input_tokens_seen": 3465032, "step": 12940 }, { "epoch": 3.3931847968545217, "grad_norm": 1.259445309638977, "learning_rate": 4.177411167280551e-05, "loss": 0.3545, "num_input_tokens_seen": 3466520, "step": 12945 }, { "epoch": 3.3944954128440368, "grad_norm": 1.1913692951202393, "learning_rate": 4.176562929305953e-05, "loss": 0.5198, "num_input_tokens_seen": 3467800, "step": 12950 }, { "epoch": 3.395806028833552, "grad_norm": 1.3259100914001465, "learning_rate": 4.1757143404296305e-05, "loss": 0.3592, "num_input_tokens_seen": 3468840, "step": 12955 }, { "epoch": 3.397116644823067, "grad_norm": 0.6978281140327454, "learning_rate": 4.1748654008291914e-05, "loss": 0.5228, "num_input_tokens_seen": 3469912, "step": 12960 }, { "epoch": 3.398427260812582, "grad_norm": 0.989033043384552, "learning_rate": 4.1740161106823185e-05, "loss": 0.3669, "num_input_tokens_seen": 3471736, "step": 12965 }, { "epoch": 3.399737876802097, "grad_norm": 0.6922523975372314, "learning_rate": 4.173166470166766e-05, "loss": 0.3496, "num_input_tokens_seen": 3473304, "step": 12970 }, { "epoch": 3.4010484927916123, "grad_norm": 1.8260563611984253, "learning_rate": 4.172316479460362e-05, "loss": 0.4037, "num_input_tokens_seen": 3474552, "step": 12975 }, { "epoch": 3.402359108781127, "grad_norm": 0.8751673698425293, "learning_rate": 4.171466138741009e-05, "loss": 0.4263, "num_input_tokens_seen": 3475864, "step": 12980 }, { "epoch": 3.4036697247706424, "grad_norm": 0.48081159591674805, "learning_rate": 4.170615448186681e-05, "loss": 0.324, "num_input_tokens_seen": 3477288, "step": 12985 }, { "epoch": 3.404980340760157, "grad_norm": 0.8565744757652283, "learning_rate": 4.169764407975426e-05, "loss": 0.3417, "num_input_tokens_seen": 3478712, "step": 12990 }, { "epoch": 3.406290956749672, "grad_norm": 1.2342891693115234, "learning_rate": 4.168913018285366e-05, "loss": 0.6233, "num_input_tokens_seen": 3480040, "step": 12995 }, { "epoch": 3.4076015727391873, "grad_norm": 0.9038825035095215, "learning_rate": 4.168061279294694e-05, "loss": 0.6298, "num_input_tokens_seen": 3481864, "step": 13000 }, { "epoch": 3.4089121887287024, "grad_norm": 0.8759608864784241, "learning_rate": 4.16720919118168e-05, "loss": 0.3869, "num_input_tokens_seen": 3483224, "step": 13005 }, { "epoch": 3.4102228047182175, "grad_norm": 0.34537312388420105, "learning_rate": 4.1663567541246615e-05, "loss": 0.3488, "num_input_tokens_seen": 3484728, "step": 13010 }, { "epoch": 3.4115334207077326, "grad_norm": 0.6590739488601685, "learning_rate": 4.165503968302055e-05, "loss": 0.4874, "num_input_tokens_seen": 3486376, "step": 13015 }, { "epoch": 3.4128440366972477, "grad_norm": 1.0835388898849487, "learning_rate": 4.1646508338923446e-05, "loss": 0.5474, "num_input_tokens_seen": 3487640, "step": 13020 }, { "epoch": 3.4141546526867628, "grad_norm": 0.7188493609428406, "learning_rate": 4.1637973510740914e-05, "loss": 0.3907, "num_input_tokens_seen": 3489160, "step": 13025 }, { "epoch": 3.415465268676278, "grad_norm": 0.6147576570510864, "learning_rate": 4.162943520025927e-05, "loss": 0.3889, "num_input_tokens_seen": 3490888, "step": 13030 }, { "epoch": 3.416775884665793, "grad_norm": 1.0244979858398438, "learning_rate": 4.162089340926558e-05, "loss": 0.519, "num_input_tokens_seen": 3492344, "step": 13035 }, { "epoch": 3.418086500655308, "grad_norm": 0.5332514047622681, "learning_rate": 4.161234813954761e-05, "loss": 0.4786, "num_input_tokens_seen": 3493384, "step": 13040 }, { "epoch": 3.419397116644823, "grad_norm": 0.6890641450881958, "learning_rate": 4.160379939289388e-05, "loss": 0.3131, "num_input_tokens_seen": 3494760, "step": 13045 }, { "epoch": 3.4207077326343382, "grad_norm": 1.2409075498580933, "learning_rate": 4.1595247171093635e-05, "loss": 0.3724, "num_input_tokens_seen": 3495896, "step": 13050 }, { "epoch": 3.4220183486238533, "grad_norm": 0.6885039806365967, "learning_rate": 4.158669147593682e-05, "loss": 0.5024, "num_input_tokens_seen": 3497160, "step": 13055 }, { "epoch": 3.4233289646133684, "grad_norm": 2.4803006649017334, "learning_rate": 4.157813230921414e-05, "loss": 0.2734, "num_input_tokens_seen": 3498472, "step": 13060 }, { "epoch": 3.4246395806028835, "grad_norm": 1.5909579992294312, "learning_rate": 4.156956967271702e-05, "loss": 0.4503, "num_input_tokens_seen": 3499688, "step": 13065 }, { "epoch": 3.4259501965923986, "grad_norm": 1.294309377670288, "learning_rate": 4.156100356823759e-05, "loss": 0.2499, "num_input_tokens_seen": 3500984, "step": 13070 }, { "epoch": 3.4272608125819133, "grad_norm": 1.244260549545288, "learning_rate": 4.1552433997568736e-05, "loss": 0.8027, "num_input_tokens_seen": 3502184, "step": 13075 }, { "epoch": 3.4285714285714284, "grad_norm": 0.2697485089302063, "learning_rate": 4.154386096250404e-05, "loss": 0.3594, "num_input_tokens_seen": 3503640, "step": 13080 }, { "epoch": 3.4298820445609435, "grad_norm": 1.7621493339538574, "learning_rate": 4.153528446483783e-05, "loss": 0.4077, "num_input_tokens_seen": 3505032, "step": 13085 }, { "epoch": 3.4311926605504586, "grad_norm": 1.0419808626174927, "learning_rate": 4.1526704506365153e-05, "loss": 0.4146, "num_input_tokens_seen": 3506088, "step": 13090 }, { "epoch": 3.4325032765399737, "grad_norm": 0.8649885654449463, "learning_rate": 4.1518121088881785e-05, "loss": 0.5812, "num_input_tokens_seen": 3507048, "step": 13095 }, { "epoch": 3.4338138925294888, "grad_norm": 0.6419506669044495, "learning_rate": 4.1509534214184206e-05, "loss": 0.3323, "num_input_tokens_seen": 3508296, "step": 13100 }, { "epoch": 3.435124508519004, "grad_norm": 2.3104469776153564, "learning_rate": 4.1500943884069644e-05, "loss": 0.7419, "num_input_tokens_seen": 3509496, "step": 13105 }, { "epoch": 3.436435124508519, "grad_norm": 0.5615848302841187, "learning_rate": 4.149235010033604e-05, "loss": 0.3819, "num_input_tokens_seen": 3511080, "step": 13110 }, { "epoch": 3.437745740498034, "grad_norm": 0.864291250705719, "learning_rate": 4.148375286478206e-05, "loss": 0.388, "num_input_tokens_seen": 3512440, "step": 13115 }, { "epoch": 3.439056356487549, "grad_norm": 1.046037197113037, "learning_rate": 4.147515217920709e-05, "loss": 0.3499, "num_input_tokens_seen": 3513720, "step": 13120 }, { "epoch": 3.4403669724770642, "grad_norm": 1.8774871826171875, "learning_rate": 4.146654804541122e-05, "loss": 0.4064, "num_input_tokens_seen": 3515240, "step": 13125 }, { "epoch": 3.4416775884665793, "grad_norm": 0.6962029337882996, "learning_rate": 4.145794046519531e-05, "loss": 0.3346, "num_input_tokens_seen": 3516424, "step": 13130 }, { "epoch": 3.4429882044560944, "grad_norm": 1.2493752241134644, "learning_rate": 4.14493294403609e-05, "loss": 0.5694, "num_input_tokens_seen": 3517592, "step": 13135 }, { "epoch": 3.4442988204456095, "grad_norm": 1.64742112159729, "learning_rate": 4.144071497271025e-05, "loss": 0.336, "num_input_tokens_seen": 3518728, "step": 13140 }, { "epoch": 3.4456094364351246, "grad_norm": 1.0433791875839233, "learning_rate": 4.143209706404637e-05, "loss": 0.3681, "num_input_tokens_seen": 3519960, "step": 13145 }, { "epoch": 3.4469200524246397, "grad_norm": 0.9383648037910461, "learning_rate": 4.142347571617296e-05, "loss": 0.3203, "num_input_tokens_seen": 3521096, "step": 13150 }, { "epoch": 3.448230668414155, "grad_norm": 0.5950590372085571, "learning_rate": 4.1414850930894466e-05, "loss": 0.4866, "num_input_tokens_seen": 3522360, "step": 13155 }, { "epoch": 3.44954128440367, "grad_norm": 1.3085261583328247, "learning_rate": 4.1406222710016034e-05, "loss": 0.299, "num_input_tokens_seen": 3523768, "step": 13160 }, { "epoch": 3.450851900393185, "grad_norm": 1.8174036741256714, "learning_rate": 4.1397591055343534e-05, "loss": 0.4029, "num_input_tokens_seen": 3525368, "step": 13165 }, { "epoch": 3.4521625163826997, "grad_norm": 7.219829559326172, "learning_rate": 4.1388955968683565e-05, "loss": 0.7779, "num_input_tokens_seen": 3526664, "step": 13170 }, { "epoch": 3.4534731323722148, "grad_norm": 0.9889565110206604, "learning_rate": 4.1380317451843425e-05, "loss": 0.4595, "num_input_tokens_seen": 3528088, "step": 13175 }, { "epoch": 3.45478374836173, "grad_norm": 0.8553174734115601, "learning_rate": 4.137167550663115e-05, "loss": 0.2997, "num_input_tokens_seen": 3529496, "step": 13180 }, { "epoch": 3.456094364351245, "grad_norm": 1.2988044023513794, "learning_rate": 4.136303013485547e-05, "loss": 0.3917, "num_input_tokens_seen": 3531720, "step": 13185 }, { "epoch": 3.45740498034076, "grad_norm": 0.5477215051651001, "learning_rate": 4.1354381338325864e-05, "loss": 0.4582, "num_input_tokens_seen": 3533032, "step": 13190 }, { "epoch": 3.458715596330275, "grad_norm": 1.9718542098999023, "learning_rate": 4.13457291188525e-05, "loss": 0.5103, "num_input_tokens_seen": 3534136, "step": 13195 }, { "epoch": 3.4600262123197902, "grad_norm": 0.590114176273346, "learning_rate": 4.1337073478246275e-05, "loss": 0.6707, "num_input_tokens_seen": 3535272, "step": 13200 }, { "epoch": 3.4613368283093053, "grad_norm": 0.6338545083999634, "learning_rate": 4.1328414418318795e-05, "loss": 0.3337, "num_input_tokens_seen": 3536536, "step": 13205 }, { "epoch": 3.4626474442988204, "grad_norm": 2.4920654296875, "learning_rate": 4.13197519408824e-05, "loss": 0.3584, "num_input_tokens_seen": 3537496, "step": 13210 }, { "epoch": 3.4639580602883355, "grad_norm": 1.3035608530044556, "learning_rate": 4.1311086047750114e-05, "loss": 0.4631, "num_input_tokens_seen": 3538696, "step": 13215 }, { "epoch": 3.4652686762778506, "grad_norm": 1.0671148300170898, "learning_rate": 4.13024167407357e-05, "loss": 0.7007, "num_input_tokens_seen": 3540184, "step": 13220 }, { "epoch": 3.4665792922673657, "grad_norm": 1.6458625793457031, "learning_rate": 4.129374402165363e-05, "loss": 0.3234, "num_input_tokens_seen": 3541576, "step": 13225 }, { "epoch": 3.467889908256881, "grad_norm": 0.8900466561317444, "learning_rate": 4.128506789231911e-05, "loss": 0.3375, "num_input_tokens_seen": 3544776, "step": 13230 }, { "epoch": 3.469200524246396, "grad_norm": 1.481719970703125, "learning_rate": 4.1276388354548e-05, "loss": 0.3088, "num_input_tokens_seen": 3545976, "step": 13235 }, { "epoch": 3.470511140235911, "grad_norm": 0.9532942771911621, "learning_rate": 4.126770541015693e-05, "loss": 0.3123, "num_input_tokens_seen": 3547160, "step": 13240 }, { "epoch": 3.471821756225426, "grad_norm": 1.3175382614135742, "learning_rate": 4.1259019060963236e-05, "loss": 0.3148, "num_input_tokens_seen": 3548296, "step": 13245 }, { "epoch": 3.473132372214941, "grad_norm": 0.9113063812255859, "learning_rate": 4.125032930878493e-05, "loss": 0.3252, "num_input_tokens_seen": 3549624, "step": 13250 }, { "epoch": 3.474442988204456, "grad_norm": 2.001901388168335, "learning_rate": 4.1241636155440785e-05, "loss": 0.3777, "num_input_tokens_seen": 3550824, "step": 13255 }, { "epoch": 3.4757536041939714, "grad_norm": 0.6200873851776123, "learning_rate": 4.123293960275027e-05, "loss": 0.3523, "num_input_tokens_seen": 3552232, "step": 13260 }, { "epoch": 3.477064220183486, "grad_norm": 1.5529171228408813, "learning_rate": 4.1224239652533526e-05, "loss": 0.4646, "num_input_tokens_seen": 3553512, "step": 13265 }, { "epoch": 3.478374836173001, "grad_norm": 0.7355450391769409, "learning_rate": 4.121553630661147e-05, "loss": 0.32, "num_input_tokens_seen": 3554728, "step": 13270 }, { "epoch": 3.4796854521625162, "grad_norm": 1.6189631223678589, "learning_rate": 4.120682956680567e-05, "loss": 0.5929, "num_input_tokens_seen": 3555800, "step": 13275 }, { "epoch": 3.4809960681520313, "grad_norm": 0.4972996115684509, "learning_rate": 4.1198119434938455e-05, "loss": 0.4182, "num_input_tokens_seen": 3557400, "step": 13280 }, { "epoch": 3.4823066841415464, "grad_norm": 0.7944049835205078, "learning_rate": 4.118940591283282e-05, "loss": 0.3584, "num_input_tokens_seen": 3558504, "step": 13285 }, { "epoch": 3.4836173001310615, "grad_norm": 0.7036214470863342, "learning_rate": 4.118068900231251e-05, "loss": 0.3923, "num_input_tokens_seen": 3559928, "step": 13290 }, { "epoch": 3.4849279161205766, "grad_norm": 2.1731207370758057, "learning_rate": 4.117196870520195e-05, "loss": 0.8308, "num_input_tokens_seen": 3561320, "step": 13295 }, { "epoch": 3.4862385321100917, "grad_norm": 1.1253196001052856, "learning_rate": 4.116324502332628e-05, "loss": 0.5671, "num_input_tokens_seen": 3562200, "step": 13300 }, { "epoch": 3.487549148099607, "grad_norm": 0.49132072925567627, "learning_rate": 4.115451795851136e-05, "loss": 0.5522, "num_input_tokens_seen": 3563544, "step": 13305 }, { "epoch": 3.488859764089122, "grad_norm": 0.5061679482460022, "learning_rate": 4.1145787512583736e-05, "loss": 0.4547, "num_input_tokens_seen": 3565320, "step": 13310 }, { "epoch": 3.490170380078637, "grad_norm": 0.9248661994934082, "learning_rate": 4.113705368737069e-05, "loss": 0.4635, "num_input_tokens_seen": 3567288, "step": 13315 }, { "epoch": 3.491480996068152, "grad_norm": 1.5652366876602173, "learning_rate": 4.112831648470018e-05, "loss": 0.4676, "num_input_tokens_seen": 3568616, "step": 13320 }, { "epoch": 3.492791612057667, "grad_norm": 0.5612421631813049, "learning_rate": 4.111957590640091e-05, "loss": 0.4263, "num_input_tokens_seen": 3570184, "step": 13325 }, { "epoch": 3.4941022280471823, "grad_norm": 4.39119815826416, "learning_rate": 4.111083195430225e-05, "loss": 0.5029, "num_input_tokens_seen": 3572040, "step": 13330 }, { "epoch": 3.4954128440366974, "grad_norm": 0.8048673272132874, "learning_rate": 4.110208463023429e-05, "loss": 0.4419, "num_input_tokens_seen": 3573560, "step": 13335 }, { "epoch": 3.4967234600262125, "grad_norm": 0.31344175338745117, "learning_rate": 4.109333393602786e-05, "loss": 0.4881, "num_input_tokens_seen": 3575048, "step": 13340 }, { "epoch": 3.4980340760157276, "grad_norm": 0.6461431384086609, "learning_rate": 4.108457987351443e-05, "loss": 0.409, "num_input_tokens_seen": 3576408, "step": 13345 }, { "epoch": 3.4993446920052422, "grad_norm": 0.9406044483184814, "learning_rate": 4.107582244452622e-05, "loss": 0.2777, "num_input_tokens_seen": 3577448, "step": 13350 }, { "epoch": 3.5006553079947578, "grad_norm": 1.0301584005355835, "learning_rate": 4.106706165089616e-05, "loss": 0.38, "num_input_tokens_seen": 3579256, "step": 13355 }, { "epoch": 3.5009174311926605, "eval_loss": 0.5108230710029602, "eval_runtime": 16.669, "eval_samples_per_second": 50.873, "eval_steps_per_second": 25.436, "num_input_tokens_seen": 3579432, "step": 13356 }, { "epoch": 3.5019659239842724, "grad_norm": 1.4594475030899048, "learning_rate": 4.105829749445785e-05, "loss": 0.4126, "num_input_tokens_seen": 3580584, "step": 13360 }, { "epoch": 3.5032765399737875, "grad_norm": 0.5254836082458496, "learning_rate": 4.104952997704563e-05, "loss": 0.3263, "num_input_tokens_seen": 3582440, "step": 13365 }, { "epoch": 3.5045871559633026, "grad_norm": 1.0862256288528442, "learning_rate": 4.1040759100494506e-05, "loss": 0.3293, "num_input_tokens_seen": 3583496, "step": 13370 }, { "epoch": 3.5058977719528177, "grad_norm": 0.8625341653823853, "learning_rate": 4.103198486664023e-05, "loss": 0.218, "num_input_tokens_seen": 3584856, "step": 13375 }, { "epoch": 3.507208387942333, "grad_norm": 1.3016166687011719, "learning_rate": 4.1023207277319213e-05, "loss": 0.371, "num_input_tokens_seen": 3586120, "step": 13380 }, { "epoch": 3.508519003931848, "grad_norm": 0.9517269730567932, "learning_rate": 4.10144263343686e-05, "loss": 0.2049, "num_input_tokens_seen": 3587688, "step": 13385 }, { "epoch": 3.509829619921363, "grad_norm": 1.6520429849624634, "learning_rate": 4.1005642039626235e-05, "loss": 0.5946, "num_input_tokens_seen": 3588888, "step": 13390 }, { "epoch": 3.511140235910878, "grad_norm": 2.0031018257141113, "learning_rate": 4.099685439493064e-05, "loss": 0.3185, "num_input_tokens_seen": 3590488, "step": 13395 }, { "epoch": 3.512450851900393, "grad_norm": 1.0867513418197632, "learning_rate": 4.098806340212106e-05, "loss": 0.3237, "num_input_tokens_seen": 3591640, "step": 13400 }, { "epoch": 3.5137614678899083, "grad_norm": 0.43772566318511963, "learning_rate": 4.097926906303744e-05, "loss": 0.3404, "num_input_tokens_seen": 3592904, "step": 13405 }, { "epoch": 3.5150720838794234, "grad_norm": 0.9377123713493347, "learning_rate": 4.0970471379520414e-05, "loss": 0.4717, "num_input_tokens_seen": 3594712, "step": 13410 }, { "epoch": 3.5163826998689385, "grad_norm": 0.45314404368400574, "learning_rate": 4.096167035341132e-05, "loss": 0.3442, "num_input_tokens_seen": 3595896, "step": 13415 }, { "epoch": 3.5176933158584536, "grad_norm": 1.191202163696289, "learning_rate": 4.09528659865522e-05, "loss": 0.446, "num_input_tokens_seen": 3597416, "step": 13420 }, { "epoch": 3.5190039318479687, "grad_norm": 1.2033077478408813, "learning_rate": 4.094405828078579e-05, "loss": 0.4291, "num_input_tokens_seen": 3598664, "step": 13425 }, { "epoch": 3.5203145478374838, "grad_norm": 0.9075400233268738, "learning_rate": 4.093524723795554e-05, "loss": 0.4886, "num_input_tokens_seen": 3599784, "step": 13430 }, { "epoch": 3.5216251638269984, "grad_norm": 0.8704401254653931, "learning_rate": 4.0926432859905574e-05, "loss": 0.4603, "num_input_tokens_seen": 3601336, "step": 13435 }, { "epoch": 3.522935779816514, "grad_norm": 1.4615336656570435, "learning_rate": 4.0917615148480726e-05, "loss": 0.3569, "num_input_tokens_seen": 3602968, "step": 13440 }, { "epoch": 3.5242463958060286, "grad_norm": 0.9119053483009338, "learning_rate": 4.0908794105526537e-05, "loss": 0.4037, "num_input_tokens_seen": 3604312, "step": 13445 }, { "epoch": 3.525557011795544, "grad_norm": 1.1939709186553955, "learning_rate": 4.089996973288922e-05, "loss": 0.4382, "num_input_tokens_seen": 3605608, "step": 13450 }, { "epoch": 3.526867627785059, "grad_norm": 0.7271466255187988, "learning_rate": 4.089114203241572e-05, "loss": 0.4564, "num_input_tokens_seen": 3607448, "step": 13455 }, { "epoch": 3.528178243774574, "grad_norm": 0.539291501045227, "learning_rate": 4.088231100595364e-05, "loss": 0.4396, "num_input_tokens_seen": 3608728, "step": 13460 }, { "epoch": 3.529488859764089, "grad_norm": 0.7806124687194824, "learning_rate": 4.087347665535133e-05, "loss": 0.3366, "num_input_tokens_seen": 3609896, "step": 13465 }, { "epoch": 3.530799475753604, "grad_norm": 0.8101630210876465, "learning_rate": 4.0864638982457766e-05, "loss": 0.3975, "num_input_tokens_seen": 3611064, "step": 13470 }, { "epoch": 3.532110091743119, "grad_norm": 1.5209051370620728, "learning_rate": 4.085579798912268e-05, "loss": 0.317, "num_input_tokens_seen": 3612280, "step": 13475 }, { "epoch": 3.5334207077326343, "grad_norm": 0.9467312097549438, "learning_rate": 4.084695367719647e-05, "loss": 0.4488, "num_input_tokens_seen": 3613464, "step": 13480 }, { "epoch": 3.5347313237221494, "grad_norm": 1.4486536979675293, "learning_rate": 4.083810604853023e-05, "loss": 0.3844, "num_input_tokens_seen": 3614936, "step": 13485 }, { "epoch": 3.5360419397116645, "grad_norm": 1.1961058378219604, "learning_rate": 4.082925510497577e-05, "loss": 0.3308, "num_input_tokens_seen": 3616232, "step": 13490 }, { "epoch": 3.5373525557011796, "grad_norm": 1.0371572971343994, "learning_rate": 4.0820400848385564e-05, "loss": 0.2373, "num_input_tokens_seen": 3617048, "step": 13495 }, { "epoch": 3.5386631716906947, "grad_norm": 0.9827103614807129, "learning_rate": 4.0811543280612805e-05, "loss": 0.4417, "num_input_tokens_seen": 3618440, "step": 13500 }, { "epoch": 3.5399737876802098, "grad_norm": 0.9018757343292236, "learning_rate": 4.080268240351135e-05, "loss": 0.5202, "num_input_tokens_seen": 3619800, "step": 13505 }, { "epoch": 3.541284403669725, "grad_norm": 0.8623421788215637, "learning_rate": 4.079381821893576e-05, "loss": 0.334, "num_input_tokens_seen": 3620984, "step": 13510 }, { "epoch": 3.54259501965924, "grad_norm": 1.3889050483703613, "learning_rate": 4.078495072874132e-05, "loss": 0.319, "num_input_tokens_seen": 3621960, "step": 13515 }, { "epoch": 3.543905635648755, "grad_norm": 0.9492108821868896, "learning_rate": 4.0776079934783975e-05, "loss": 0.2588, "num_input_tokens_seen": 3623016, "step": 13520 }, { "epoch": 3.54521625163827, "grad_norm": 0.6232650279998779, "learning_rate": 4.076720583892034e-05, "loss": 0.4128, "num_input_tokens_seen": 3624408, "step": 13525 }, { "epoch": 3.546526867627785, "grad_norm": 3.5490918159484863, "learning_rate": 4.075832844300778e-05, "loss": 0.3895, "num_input_tokens_seen": 3625624, "step": 13530 }, { "epoch": 3.5478374836173003, "grad_norm": 0.9016141891479492, "learning_rate": 4.0749447748904304e-05, "loss": 0.3467, "num_input_tokens_seen": 3626808, "step": 13535 }, { "epoch": 3.549148099606815, "grad_norm": 2.6728174686431885, "learning_rate": 4.074056375846862e-05, "loss": 0.5352, "num_input_tokens_seen": 3628104, "step": 13540 }, { "epoch": 3.5504587155963305, "grad_norm": 2.3157660961151123, "learning_rate": 4.0731676473560145e-05, "loss": 0.3621, "num_input_tokens_seen": 3629048, "step": 13545 }, { "epoch": 3.551769331585845, "grad_norm": 2.7368876934051514, "learning_rate": 4.072278589603896e-05, "loss": 0.5034, "num_input_tokens_seen": 3630536, "step": 13550 }, { "epoch": 3.5530799475753603, "grad_norm": 0.7409206032752991, "learning_rate": 4.0713892027765863e-05, "loss": 0.2809, "num_input_tokens_seen": 3631560, "step": 13555 }, { "epoch": 3.5543905635648754, "grad_norm": 0.7424606084823608, "learning_rate": 4.070499487060231e-05, "loss": 0.3377, "num_input_tokens_seen": 3632712, "step": 13560 }, { "epoch": 3.5557011795543905, "grad_norm": 0.9267617464065552, "learning_rate": 4.069609442641046e-05, "loss": 0.4547, "num_input_tokens_seen": 3634088, "step": 13565 }, { "epoch": 3.5570117955439056, "grad_norm": 4.359254837036133, "learning_rate": 4.068719069705318e-05, "loss": 0.5236, "num_input_tokens_seen": 3635480, "step": 13570 }, { "epoch": 3.5583224115334207, "grad_norm": 0.7831059098243713, "learning_rate": 4.0678283684393995e-05, "loss": 0.5101, "num_input_tokens_seen": 3637080, "step": 13575 }, { "epoch": 3.5596330275229358, "grad_norm": 0.9109848141670227, "learning_rate": 4.0669373390297115e-05, "loss": 0.4407, "num_input_tokens_seen": 3638248, "step": 13580 }, { "epoch": 3.560943643512451, "grad_norm": 1.2190130949020386, "learning_rate": 4.066045981662746e-05, "loss": 0.3451, "num_input_tokens_seen": 3639448, "step": 13585 }, { "epoch": 3.562254259501966, "grad_norm": 1.6526154279708862, "learning_rate": 4.065154296525063e-05, "loss": 0.3384, "num_input_tokens_seen": 3640648, "step": 13590 }, { "epoch": 3.563564875491481, "grad_norm": 0.3741651475429535, "learning_rate": 4.064262283803289e-05, "loss": 0.3489, "num_input_tokens_seen": 3642120, "step": 13595 }, { "epoch": 3.564875491480996, "grad_norm": 1.0218112468719482, "learning_rate": 4.0633699436841224e-05, "loss": 0.466, "num_input_tokens_seen": 3643208, "step": 13600 }, { "epoch": 3.5661861074705112, "grad_norm": 0.5050973296165466, "learning_rate": 4.062477276354328e-05, "loss": 0.3592, "num_input_tokens_seen": 3644776, "step": 13605 }, { "epoch": 3.5674967234600263, "grad_norm": 1.2577061653137207, "learning_rate": 4.06158428200074e-05, "loss": 0.3721, "num_input_tokens_seen": 3646056, "step": 13610 }, { "epoch": 3.5688073394495414, "grad_norm": 1.2881280183792114, "learning_rate": 4.0606909608102595e-05, "loss": 0.3933, "num_input_tokens_seen": 3647800, "step": 13615 }, { "epoch": 3.5701179554390565, "grad_norm": 0.8214377760887146, "learning_rate": 4.059797312969857e-05, "loss": 0.4087, "num_input_tokens_seen": 3649240, "step": 13620 }, { "epoch": 3.571428571428571, "grad_norm": 0.7521225214004517, "learning_rate": 4.058903338666573e-05, "loss": 0.4699, "num_input_tokens_seen": 3650568, "step": 13625 }, { "epoch": 3.5727391874180867, "grad_norm": 0.5228440761566162, "learning_rate": 4.058009038087513e-05, "loss": 0.4888, "num_input_tokens_seen": 3652056, "step": 13630 }, { "epoch": 3.5740498034076014, "grad_norm": 1.623928189277649, "learning_rate": 4.0571144114198536e-05, "loss": 0.3267, "num_input_tokens_seen": 3653128, "step": 13635 }, { "epoch": 3.575360419397117, "grad_norm": 0.919354259967804, "learning_rate": 4.056219458850838e-05, "loss": 0.4832, "num_input_tokens_seen": 3654648, "step": 13640 }, { "epoch": 3.5766710353866316, "grad_norm": 1.887241244316101, "learning_rate": 4.05532418056778e-05, "loss": 0.3322, "num_input_tokens_seen": 3655896, "step": 13645 }, { "epoch": 3.5779816513761467, "grad_norm": 1.727588415145874, "learning_rate": 4.054428576758057e-05, "loss": 0.3777, "num_input_tokens_seen": 3657624, "step": 13650 }, { "epoch": 3.5792922673656618, "grad_norm": 1.9364272356033325, "learning_rate": 4.053532647609119e-05, "loss": 0.3021, "num_input_tokens_seen": 3658808, "step": 13655 }, { "epoch": 3.580602883355177, "grad_norm": 2.069234848022461, "learning_rate": 4.0526363933084824e-05, "loss": 0.587, "num_input_tokens_seen": 3660088, "step": 13660 }, { "epoch": 3.581913499344692, "grad_norm": 0.13682712614536285, "learning_rate": 4.051739814043731e-05, "loss": 0.2003, "num_input_tokens_seen": 3661160, "step": 13665 }, { "epoch": 3.583224115334207, "grad_norm": 1.5421077013015747, "learning_rate": 4.0508429100025185e-05, "loss": 0.441, "num_input_tokens_seen": 3662424, "step": 13670 }, { "epoch": 3.584534731323722, "grad_norm": 1.3167057037353516, "learning_rate": 4.049945681372565e-05, "loss": 0.3205, "num_input_tokens_seen": 3663656, "step": 13675 }, { "epoch": 3.5858453473132372, "grad_norm": 0.9032373428344727, "learning_rate": 4.049048128341657e-05, "loss": 0.2599, "num_input_tokens_seen": 3664712, "step": 13680 }, { "epoch": 3.5871559633027523, "grad_norm": 1.8343915939331055, "learning_rate": 4.048150251097653e-05, "loss": 0.4007, "num_input_tokens_seen": 3666312, "step": 13685 }, { "epoch": 3.5884665792922674, "grad_norm": 1.5592379570007324, "learning_rate": 4.047252049828476e-05, "loss": 0.402, "num_input_tokens_seen": 3667304, "step": 13690 }, { "epoch": 3.5897771952817825, "grad_norm": 1.020726203918457, "learning_rate": 4.046353524722119e-05, "loss": 0.3913, "num_input_tokens_seen": 3668328, "step": 13695 }, { "epoch": 3.5910878112712976, "grad_norm": 1.1044061183929443, "learning_rate": 4.0454546759666414e-05, "loss": 0.5269, "num_input_tokens_seen": 3669464, "step": 13700 }, { "epoch": 3.5923984272608127, "grad_norm": 1.0969526767730713, "learning_rate": 4.04455550375017e-05, "loss": 0.3128, "num_input_tokens_seen": 3670584, "step": 13705 }, { "epoch": 3.593709043250328, "grad_norm": 0.9509369134902954, "learning_rate": 4.0436560082609e-05, "loss": 0.4118, "num_input_tokens_seen": 3671592, "step": 13710 }, { "epoch": 3.595019659239843, "grad_norm": 1.969576120376587, "learning_rate": 4.0427561896870955e-05, "loss": 0.4354, "num_input_tokens_seen": 3673048, "step": 13715 }, { "epoch": 3.5963302752293576, "grad_norm": 0.6164811253547668, "learning_rate": 4.041856048217085e-05, "loss": 0.5657, "num_input_tokens_seen": 3674568, "step": 13720 }, { "epoch": 3.597640891218873, "grad_norm": 3.7255730628967285, "learning_rate": 4.040955584039269e-05, "loss": 0.3794, "num_input_tokens_seen": 3675608, "step": 13725 }, { "epoch": 3.5989515072083877, "grad_norm": 0.40278875827789307, "learning_rate": 4.040054797342112e-05, "loss": 0.3506, "num_input_tokens_seen": 3676872, "step": 13730 }, { "epoch": 3.6002621231979033, "grad_norm": 0.954251766204834, "learning_rate": 4.039153688314145e-05, "loss": 0.2952, "num_input_tokens_seen": 3678360, "step": 13735 }, { "epoch": 3.601572739187418, "grad_norm": 1.0304450988769531, "learning_rate": 4.038252257143973e-05, "loss": 0.3738, "num_input_tokens_seen": 3679992, "step": 13740 }, { "epoch": 3.602883355176933, "grad_norm": 2.307598829269409, "learning_rate": 4.03735050402026e-05, "loss": 0.4494, "num_input_tokens_seen": 3682120, "step": 13745 }, { "epoch": 3.604193971166448, "grad_norm": 0.8704211711883545, "learning_rate": 4.036448429131743e-05, "loss": 0.3632, "num_input_tokens_seen": 3683480, "step": 13750 }, { "epoch": 3.6055045871559632, "grad_norm": 0.7679812908172607, "learning_rate": 4.035546032667225e-05, "loss": 0.3659, "num_input_tokens_seen": 3685368, "step": 13755 }, { "epoch": 3.6068152031454783, "grad_norm": 2.096409559249878, "learning_rate": 4.034643314815575e-05, "loss": 0.5755, "num_input_tokens_seen": 3686552, "step": 13760 }, { "epoch": 3.6081258191349934, "grad_norm": 1.7113841772079468, "learning_rate": 4.0337402757657314e-05, "loss": 0.3643, "num_input_tokens_seen": 3687816, "step": 13765 }, { "epoch": 3.6094364351245085, "grad_norm": 1.2056057453155518, "learning_rate": 4.032836915706698e-05, "loss": 0.3433, "num_input_tokens_seen": 3689176, "step": 13770 }, { "epoch": 3.6107470511140236, "grad_norm": 0.6311746835708618, "learning_rate": 4.0319332348275465e-05, "loss": 0.336, "num_input_tokens_seen": 3690232, "step": 13775 }, { "epoch": 3.6120576671035387, "grad_norm": 1.1757292747497559, "learning_rate": 4.031029233317416e-05, "loss": 0.3272, "num_input_tokens_seen": 3691480, "step": 13780 }, { "epoch": 3.613368283093054, "grad_norm": 2.1751608848571777, "learning_rate": 4.0301249113655125e-05, "loss": 0.4596, "num_input_tokens_seen": 3693016, "step": 13785 }, { "epoch": 3.614678899082569, "grad_norm": 1.3399169445037842, "learning_rate": 4.0292202691611094e-05, "loss": 0.3883, "num_input_tokens_seen": 3694136, "step": 13790 }, { "epoch": 3.615989515072084, "grad_norm": 0.7792652249336243, "learning_rate": 4.028315306893545e-05, "loss": 0.5513, "num_input_tokens_seen": 3695672, "step": 13795 }, { "epoch": 3.617300131061599, "grad_norm": 0.4776395857334137, "learning_rate": 4.0274100247522287e-05, "loss": 0.3083, "num_input_tokens_seen": 3697032, "step": 13800 }, { "epoch": 3.618610747051114, "grad_norm": 3.075652599334717, "learning_rate": 4.026504422926632e-05, "loss": 0.3492, "num_input_tokens_seen": 3698088, "step": 13805 }, { "epoch": 3.6199213630406293, "grad_norm": 0.9932785630226135, "learning_rate": 4.025598501606299e-05, "loss": 0.3779, "num_input_tokens_seen": 3699624, "step": 13810 }, { "epoch": 3.621231979030144, "grad_norm": 0.6084864139556885, "learning_rate": 4.024692260980835e-05, "loss": 0.3841, "num_input_tokens_seen": 3701080, "step": 13815 }, { "epoch": 3.6225425950196595, "grad_norm": 3.7317771911621094, "learning_rate": 4.023785701239915e-05, "loss": 0.4062, "num_input_tokens_seen": 3702088, "step": 13820 }, { "epoch": 3.623853211009174, "grad_norm": 1.2050566673278809, "learning_rate": 4.022878822573281e-05, "loss": 0.3261, "num_input_tokens_seen": 3703224, "step": 13825 }, { "epoch": 3.625163826998689, "grad_norm": 2.253376007080078, "learning_rate": 4.0219716251707395e-05, "loss": 0.4479, "num_input_tokens_seen": 3704520, "step": 13830 }, { "epoch": 3.6264744429882043, "grad_norm": 1.1646244525909424, "learning_rate": 4.021064109222168e-05, "loss": 0.2553, "num_input_tokens_seen": 3706584, "step": 13835 }, { "epoch": 3.6277850589777194, "grad_norm": 1.8463507890701294, "learning_rate": 4.020156274917506e-05, "loss": 0.5173, "num_input_tokens_seen": 3708200, "step": 13840 }, { "epoch": 3.6290956749672345, "grad_norm": 0.6389721035957336, "learning_rate": 4.019248122446763e-05, "loss": 0.3495, "num_input_tokens_seen": 3709560, "step": 13845 }, { "epoch": 3.6304062909567496, "grad_norm": 1.2110682725906372, "learning_rate": 4.0183396520000114e-05, "loss": 0.5514, "num_input_tokens_seen": 3710952, "step": 13850 }, { "epoch": 3.6317169069462647, "grad_norm": 0.8643091320991516, "learning_rate": 4.0174308637673956e-05, "loss": 0.5366, "num_input_tokens_seen": 3712392, "step": 13855 }, { "epoch": 3.63302752293578, "grad_norm": 1.1190720796585083, "learning_rate": 4.016521757939121e-05, "loss": 0.4754, "num_input_tokens_seen": 3713464, "step": 13860 }, { "epoch": 3.634338138925295, "grad_norm": 1.7050834894180298, "learning_rate": 4.015612334705464e-05, "loss": 0.3435, "num_input_tokens_seen": 3714968, "step": 13865 }, { "epoch": 3.63564875491481, "grad_norm": 0.613486111164093, "learning_rate": 4.014702594256764e-05, "loss": 0.571, "num_input_tokens_seen": 3716248, "step": 13870 }, { "epoch": 3.636959370904325, "grad_norm": 2.762303113937378, "learning_rate": 4.013792536783427e-05, "loss": 0.4097, "num_input_tokens_seen": 3717736, "step": 13875 }, { "epoch": 3.63826998689384, "grad_norm": 0.9963790774345398, "learning_rate": 4.012882162475928e-05, "loss": 0.419, "num_input_tokens_seen": 3718904, "step": 13880 }, { "epoch": 3.6395806028833553, "grad_norm": 0.5557476282119751, "learning_rate": 4.011971471524807e-05, "loss": 0.2966, "num_input_tokens_seen": 3720104, "step": 13885 }, { "epoch": 3.6408912188728704, "grad_norm": 2.5372731685638428, "learning_rate": 4.011060464120669e-05, "loss": 0.3712, "num_input_tokens_seen": 3721288, "step": 13890 }, { "epoch": 3.6422018348623855, "grad_norm": 1.1839570999145508, "learning_rate": 4.010149140454188e-05, "loss": 0.4028, "num_input_tokens_seen": 3722856, "step": 13895 }, { "epoch": 3.6435124508519, "grad_norm": 0.9247267842292786, "learning_rate": 4.009237500716101e-05, "loss": 0.3486, "num_input_tokens_seen": 3724264, "step": 13900 }, { "epoch": 3.6448230668414157, "grad_norm": 0.3485855758190155, "learning_rate": 4.008325545097212e-05, "loss": 0.1769, "num_input_tokens_seen": 3725528, "step": 13905 }, { "epoch": 3.6461336828309303, "grad_norm": 1.8404299020767212, "learning_rate": 4.0074132737883927e-05, "loss": 0.4182, "num_input_tokens_seen": 3726888, "step": 13910 }, { "epoch": 3.647444298820446, "grad_norm": 2.9511680603027344, "learning_rate": 4.00650068698058e-05, "loss": 0.4994, "num_input_tokens_seen": 3728472, "step": 13915 }, { "epoch": 3.6487549148099605, "grad_norm": 1.084266185760498, "learning_rate": 4.005587784864776e-05, "loss": 0.479, "num_input_tokens_seen": 3729704, "step": 13920 }, { "epoch": 3.6500655307994756, "grad_norm": 1.1485730409622192, "learning_rate": 4.0046745676320504e-05, "loss": 0.3496, "num_input_tokens_seen": 3731144, "step": 13925 }, { "epoch": 3.6513761467889907, "grad_norm": 1.985418438911438, "learning_rate": 4.003761035473538e-05, "loss": 0.4705, "num_input_tokens_seen": 3732456, "step": 13930 }, { "epoch": 3.652686762778506, "grad_norm": 0.8610768914222717, "learning_rate": 4.002847188580439e-05, "loss": 0.467, "num_input_tokens_seen": 3734312, "step": 13935 }, { "epoch": 3.653997378768021, "grad_norm": 1.3236161470413208, "learning_rate": 4.0019330271440206e-05, "loss": 0.3966, "num_input_tokens_seen": 3735944, "step": 13940 }, { "epoch": 3.655307994757536, "grad_norm": 0.652395486831665, "learning_rate": 4.001018551355614e-05, "loss": 0.347, "num_input_tokens_seen": 3737304, "step": 13945 }, { "epoch": 3.656618610747051, "grad_norm": 0.47080251574516296, "learning_rate": 4.0001037614066184e-05, "loss": 0.3291, "num_input_tokens_seen": 3738696, "step": 13950 }, { "epoch": 3.657929226736566, "grad_norm": 1.0724682807922363, "learning_rate": 3.9991886574884975e-05, "loss": 0.4242, "num_input_tokens_seen": 3739752, "step": 13955 }, { "epoch": 3.6592398427260813, "grad_norm": 1.325995922088623, "learning_rate": 3.998273239792781e-05, "loss": 0.4187, "num_input_tokens_seen": 3741176, "step": 13960 }, { "epoch": 3.6605504587155964, "grad_norm": 1.3198798894882202, "learning_rate": 3.997357508511064e-05, "loss": 0.3776, "num_input_tokens_seen": 3742472, "step": 13965 }, { "epoch": 3.6618610747051115, "grad_norm": 0.7613834738731384, "learning_rate": 3.996441463835008e-05, "loss": 0.4141, "num_input_tokens_seen": 3743592, "step": 13970 }, { "epoch": 3.6631716906946266, "grad_norm": 0.7715174555778503, "learning_rate": 3.995525105956339e-05, "loss": 0.3229, "num_input_tokens_seen": 3744888, "step": 13975 }, { "epoch": 3.6644823066841417, "grad_norm": 1.1214805841445923, "learning_rate": 3.9946084350668506e-05, "loss": 0.4123, "num_input_tokens_seen": 3745928, "step": 13980 }, { "epoch": 3.6657929226736568, "grad_norm": 1.5684713125228882, "learning_rate": 3.993691451358398e-05, "loss": 0.4973, "num_input_tokens_seen": 3746936, "step": 13985 }, { "epoch": 3.667103538663172, "grad_norm": 0.7104909420013428, "learning_rate": 3.992774155022906e-05, "loss": 0.354, "num_input_tokens_seen": 3748600, "step": 13990 }, { "epoch": 3.6684141546526865, "grad_norm": 1.4452846050262451, "learning_rate": 3.991856546252362e-05, "loss": 0.4909, "num_input_tokens_seen": 3750104, "step": 13995 }, { "epoch": 3.669724770642202, "grad_norm": 1.1462008953094482, "learning_rate": 3.9909386252388215e-05, "loss": 0.3469, "num_input_tokens_seen": 3751752, "step": 14000 }, { "epoch": 3.6710353866317167, "grad_norm": 2.1872518062591553, "learning_rate": 3.990020392174402e-05, "loss": 0.3211, "num_input_tokens_seen": 3753224, "step": 14005 }, { "epoch": 3.6723460026212322, "grad_norm": 2.04902720451355, "learning_rate": 3.9891018472512895e-05, "loss": 0.4769, "num_input_tokens_seen": 3754280, "step": 14010 }, { "epoch": 3.673656618610747, "grad_norm": 0.7173255681991577, "learning_rate": 3.988182990661734e-05, "loss": 0.3491, "num_input_tokens_seen": 3755640, "step": 14015 }, { "epoch": 3.674967234600262, "grad_norm": 0.7126530408859253, "learning_rate": 3.987263822598049e-05, "loss": 0.3734, "num_input_tokens_seen": 3757064, "step": 14020 }, { "epoch": 3.676277850589777, "grad_norm": 0.8680738806724548, "learning_rate": 3.986344343252615e-05, "loss": 0.3347, "num_input_tokens_seen": 3758296, "step": 14025 }, { "epoch": 3.677588466579292, "grad_norm": 0.9646985530853271, "learning_rate": 3.9854245528178804e-05, "loss": 0.3324, "num_input_tokens_seen": 3759496, "step": 14030 }, { "epoch": 3.6788990825688073, "grad_norm": 3.0601248741149902, "learning_rate": 3.984504451486352e-05, "loss": 0.6255, "num_input_tokens_seen": 3760712, "step": 14035 }, { "epoch": 3.6802096985583224, "grad_norm": 1.1124377250671387, "learning_rate": 3.983584039450607e-05, "loss": 0.4012, "num_input_tokens_seen": 3761800, "step": 14040 }, { "epoch": 3.6815203145478375, "grad_norm": 1.4636439085006714, "learning_rate": 3.982663316903286e-05, "loss": 0.3005, "num_input_tokens_seen": 3763368, "step": 14045 }, { "epoch": 3.6828309305373526, "grad_norm": 1.1983970403671265, "learning_rate": 3.981742284037095e-05, "loss": 0.4141, "num_input_tokens_seen": 3764632, "step": 14050 }, { "epoch": 3.6841415465268676, "grad_norm": 0.9334726333618164, "learning_rate": 3.980820941044803e-05, "loss": 0.4321, "num_input_tokens_seen": 3765896, "step": 14055 }, { "epoch": 3.6854521625163827, "grad_norm": 1.2987070083618164, "learning_rate": 3.979899288119248e-05, "loss": 0.312, "num_input_tokens_seen": 3767128, "step": 14060 }, { "epoch": 3.686762778505898, "grad_norm": 1.35708749294281, "learning_rate": 3.978977325453329e-05, "loss": 0.3117, "num_input_tokens_seen": 3768136, "step": 14065 }, { "epoch": 3.688073394495413, "grad_norm": 1.5041320323944092, "learning_rate": 3.9780550532400106e-05, "loss": 0.352, "num_input_tokens_seen": 3769048, "step": 14070 }, { "epoch": 3.689384010484928, "grad_norm": 0.5406659245491028, "learning_rate": 3.9771324716723246e-05, "loss": 0.3709, "num_input_tokens_seen": 3770312, "step": 14075 }, { "epoch": 3.690694626474443, "grad_norm": 2.902353286743164, "learning_rate": 3.976209580943363e-05, "loss": 0.4562, "num_input_tokens_seen": 3771304, "step": 14080 }, { "epoch": 3.6920052424639582, "grad_norm": 0.5223312377929688, "learning_rate": 3.975286381246288e-05, "loss": 0.4706, "num_input_tokens_seen": 3772616, "step": 14085 }, { "epoch": 3.693315858453473, "grad_norm": 1.0552000999450684, "learning_rate": 3.9743628727743224e-05, "loss": 0.4446, "num_input_tokens_seen": 3773704, "step": 14090 }, { "epoch": 3.6946264744429884, "grad_norm": 2.142364501953125, "learning_rate": 3.973439055720755e-05, "loss": 0.3386, "num_input_tokens_seen": 3774728, "step": 14095 }, { "epoch": 3.695937090432503, "grad_norm": 1.8685067892074585, "learning_rate": 3.972514930278938e-05, "loss": 0.7705, "num_input_tokens_seen": 3776440, "step": 14100 }, { "epoch": 3.6972477064220186, "grad_norm": 0.7835041284561157, "learning_rate": 3.971590496642291e-05, "loss": 0.3457, "num_input_tokens_seen": 3777768, "step": 14105 }, { "epoch": 3.6985583224115333, "grad_norm": 1.3196215629577637, "learning_rate": 3.970665755004296e-05, "loss": 0.3576, "num_input_tokens_seen": 3778824, "step": 14110 }, { "epoch": 3.6998689384010484, "grad_norm": 0.5475981831550598, "learning_rate": 3.9697407055585e-05, "loss": 0.3088, "num_input_tokens_seen": 3780344, "step": 14115 }, { "epoch": 3.7011795543905635, "grad_norm": 1.9128377437591553, "learning_rate": 3.9688153484985135e-05, "loss": 0.4003, "num_input_tokens_seen": 3781384, "step": 14120 }, { "epoch": 3.7024901703800785, "grad_norm": 1.2888175249099731, "learning_rate": 3.9678896840180125e-05, "loss": 1.8516, "num_input_tokens_seen": 3782488, "step": 14125 }, { "epoch": 3.7038007863695936, "grad_norm": 0.6049098968505859, "learning_rate": 3.9669637123107374e-05, "loss": 0.4699, "num_input_tokens_seen": 3784408, "step": 14130 }, { "epoch": 3.7051114023591087, "grad_norm": 1.114485502243042, "learning_rate": 3.9660374335704906e-05, "loss": 0.4304, "num_input_tokens_seen": 3785608, "step": 14135 }, { "epoch": 3.706422018348624, "grad_norm": 0.9368941783905029, "learning_rate": 3.965110847991143e-05, "loss": 0.4, "num_input_tokens_seen": 3786712, "step": 14140 }, { "epoch": 3.707732634338139, "grad_norm": 0.6259164810180664, "learning_rate": 3.964183955766627e-05, "loss": 0.4291, "num_input_tokens_seen": 3787928, "step": 14145 }, { "epoch": 3.709043250327654, "grad_norm": 0.744727611541748, "learning_rate": 3.963256757090938e-05, "loss": 0.4275, "num_input_tokens_seen": 3789480, "step": 14150 }, { "epoch": 3.710353866317169, "grad_norm": 1.066974401473999, "learning_rate": 3.962329252158139e-05, "loss": 0.2174, "num_input_tokens_seen": 3790696, "step": 14155 }, { "epoch": 3.711664482306684, "grad_norm": 1.116836667060852, "learning_rate": 3.961401441162354e-05, "loss": 0.3769, "num_input_tokens_seen": 3791832, "step": 14160 }, { "epoch": 3.7129750982961993, "grad_norm": 1.109150767326355, "learning_rate": 3.960473324297772e-05, "loss": 0.495, "num_input_tokens_seen": 3792888, "step": 14165 }, { "epoch": 3.7142857142857144, "grad_norm": 1.0132731199264526, "learning_rate": 3.959544901758646e-05, "loss": 0.43, "num_input_tokens_seen": 3794792, "step": 14170 }, { "epoch": 3.7155963302752295, "grad_norm": 1.925199031829834, "learning_rate": 3.958616173739295e-05, "loss": 0.6454, "num_input_tokens_seen": 3796712, "step": 14175 }, { "epoch": 3.7169069462647446, "grad_norm": 0.8587391376495361, "learning_rate": 3.9576871404340994e-05, "loss": 0.4579, "num_input_tokens_seen": 3797896, "step": 14180 }, { "epoch": 3.7182175622542593, "grad_norm": 0.774343729019165, "learning_rate": 3.9567578020375036e-05, "loss": 0.7153, "num_input_tokens_seen": 3798808, "step": 14185 }, { "epoch": 3.719528178243775, "grad_norm": 1.1390503644943237, "learning_rate": 3.955828158744017e-05, "loss": 0.3847, "num_input_tokens_seen": 3800264, "step": 14190 }, { "epoch": 3.7208387942332894, "grad_norm": 0.6294752955436707, "learning_rate": 3.954898210748211e-05, "loss": 0.314, "num_input_tokens_seen": 3802008, "step": 14195 }, { "epoch": 3.722149410222805, "grad_norm": 1.314095139503479, "learning_rate": 3.953967958244725e-05, "loss": 0.3284, "num_input_tokens_seen": 3803256, "step": 14200 }, { "epoch": 3.7234600262123196, "grad_norm": 1.4656463861465454, "learning_rate": 3.9530374014282574e-05, "loss": 0.6075, "num_input_tokens_seen": 3804504, "step": 14205 }, { "epoch": 3.7247706422018347, "grad_norm": 0.829846203327179, "learning_rate": 3.952106540493572e-05, "loss": 0.4408, "num_input_tokens_seen": 3806424, "step": 14210 }, { "epoch": 3.72608125819135, "grad_norm": 1.1798505783081055, "learning_rate": 3.951175375635497e-05, "loss": 0.2921, "num_input_tokens_seen": 3807752, "step": 14215 }, { "epoch": 3.727391874180865, "grad_norm": 0.7585403323173523, "learning_rate": 3.9502439070489224e-05, "loss": 0.3919, "num_input_tokens_seen": 3808856, "step": 14220 }, { "epoch": 3.72870249017038, "grad_norm": 0.7328468561172485, "learning_rate": 3.949312134928805e-05, "loss": 0.5031, "num_input_tokens_seen": 3810040, "step": 14225 }, { "epoch": 3.730013106159895, "grad_norm": 0.7913419604301453, "learning_rate": 3.948380059470161e-05, "loss": 0.3749, "num_input_tokens_seen": 3811224, "step": 14230 }, { "epoch": 3.73132372214941, "grad_norm": 1.578640341758728, "learning_rate": 3.947447680868074e-05, "loss": 0.6335, "num_input_tokens_seen": 3812424, "step": 14235 }, { "epoch": 3.7326343381389253, "grad_norm": 1.4512447118759155, "learning_rate": 3.9465149993176885e-05, "loss": 0.4298, "num_input_tokens_seen": 3813576, "step": 14240 }, { "epoch": 3.7339449541284404, "grad_norm": 0.9129408597946167, "learning_rate": 3.945582015014212e-05, "loss": 0.3605, "num_input_tokens_seen": 3814536, "step": 14245 }, { "epoch": 3.7352555701179555, "grad_norm": 1.439340353012085, "learning_rate": 3.944648728152919e-05, "loss": 0.378, "num_input_tokens_seen": 3816504, "step": 14250 }, { "epoch": 3.7365661861074706, "grad_norm": 0.9618875980377197, "learning_rate": 3.943715138929142e-05, "loss": 0.403, "num_input_tokens_seen": 3817640, "step": 14255 }, { "epoch": 3.7378768020969857, "grad_norm": 2.4991753101348877, "learning_rate": 3.942781247538282e-05, "loss": 0.4744, "num_input_tokens_seen": 3818696, "step": 14260 }, { "epoch": 3.739187418086501, "grad_norm": 2.9364407062530518, "learning_rate": 3.941847054175799e-05, "loss": 0.2322, "num_input_tokens_seen": 3819736, "step": 14265 }, { "epoch": 3.740498034076016, "grad_norm": 0.9145066738128662, "learning_rate": 3.940912559037219e-05, "loss": 0.6669, "num_input_tokens_seen": 3821256, "step": 14270 }, { "epoch": 3.741808650065531, "grad_norm": 1.243040680885315, "learning_rate": 3.939977762318131e-05, "loss": 0.4392, "num_input_tokens_seen": 3822472, "step": 14275 }, { "epoch": 3.7431192660550456, "grad_norm": 2.484257221221924, "learning_rate": 3.939042664214184e-05, "loss": 0.4072, "num_input_tokens_seen": 3823400, "step": 14280 }, { "epoch": 3.744429882044561, "grad_norm": 1.2355892658233643, "learning_rate": 3.938107264921095e-05, "loss": 0.5192, "num_input_tokens_seen": 3824472, "step": 14285 }, { "epoch": 3.745740498034076, "grad_norm": 0.9257277250289917, "learning_rate": 3.9371715646346396e-05, "loss": 0.2397, "num_input_tokens_seen": 3826408, "step": 14290 }, { "epoch": 3.747051114023591, "grad_norm": 0.8580982685089111, "learning_rate": 3.936235563550659e-05, "loss": 0.4495, "num_input_tokens_seen": 3828456, "step": 14295 }, { "epoch": 3.748361730013106, "grad_norm": 0.9841142892837524, "learning_rate": 3.935299261865057e-05, "loss": 0.3603, "num_input_tokens_seen": 3829640, "step": 14300 }, { "epoch": 3.749672346002621, "grad_norm": 0.9687109589576721, "learning_rate": 3.934362659773799e-05, "loss": 0.3514, "num_input_tokens_seen": 3830680, "step": 14305 }, { "epoch": 3.750982961992136, "grad_norm": 2.873581647872925, "learning_rate": 3.933425757472915e-05, "loss": 0.4315, "num_input_tokens_seen": 3831768, "step": 14310 }, { "epoch": 3.7522935779816513, "grad_norm": 1.1622958183288574, "learning_rate": 3.932488555158497e-05, "loss": 0.4915, "num_input_tokens_seen": 3833272, "step": 14315 }, { "epoch": 3.7536041939711664, "grad_norm": 1.1393346786499023, "learning_rate": 3.931551053026699e-05, "loss": 0.2634, "num_input_tokens_seen": 3834408, "step": 14320 }, { "epoch": 3.7549148099606815, "grad_norm": 1.104943871498108, "learning_rate": 3.9306132512737395e-05, "loss": 0.35, "num_input_tokens_seen": 3835432, "step": 14325 }, { "epoch": 3.7562254259501966, "grad_norm": 3.2872087955474854, "learning_rate": 3.929675150095898e-05, "loss": 0.2977, "num_input_tokens_seen": 3836808, "step": 14330 }, { "epoch": 3.7575360419397117, "grad_norm": 0.7097954154014587, "learning_rate": 3.928736749689519e-05, "loss": 0.382, "num_input_tokens_seen": 3837976, "step": 14335 }, { "epoch": 3.758846657929227, "grad_norm": 0.7637161016464233, "learning_rate": 3.927798050251006e-05, "loss": 0.3297, "num_input_tokens_seen": 3839288, "step": 14340 }, { "epoch": 3.760157273918742, "grad_norm": 1.2677721977233887, "learning_rate": 3.926859051976828e-05, "loss": 0.4502, "num_input_tokens_seen": 3840600, "step": 14345 }, { "epoch": 3.761467889908257, "grad_norm": 1.1503814458847046, "learning_rate": 3.9259197550635164e-05, "loss": 0.4055, "num_input_tokens_seen": 3841704, "step": 14350 }, { "epoch": 3.762778505897772, "grad_norm": 3.2775981426239014, "learning_rate": 3.924980159707664e-05, "loss": 0.5567, "num_input_tokens_seen": 3842776, "step": 14355 }, { "epoch": 3.764089121887287, "grad_norm": 1.3013134002685547, "learning_rate": 3.924040266105926e-05, "loss": 0.2586, "num_input_tokens_seen": 3844072, "step": 14360 }, { "epoch": 3.765399737876802, "grad_norm": 1.1480563879013062, "learning_rate": 3.9231000744550205e-05, "loss": 0.4391, "num_input_tokens_seen": 3845208, "step": 14365 }, { "epoch": 3.7667103538663174, "grad_norm": 0.462841659784317, "learning_rate": 3.922159584951729e-05, "loss": 0.4593, "num_input_tokens_seen": 3846584, "step": 14370 }, { "epoch": 3.768020969855832, "grad_norm": 0.747713029384613, "learning_rate": 3.921218797792893e-05, "loss": 0.4338, "num_input_tokens_seen": 3848040, "step": 14375 }, { "epoch": 3.7693315858453476, "grad_norm": 1.3139632940292358, "learning_rate": 3.9202777131754187e-05, "loss": 0.4574, "num_input_tokens_seen": 3849224, "step": 14380 }, { "epoch": 3.770642201834862, "grad_norm": 1.0444220304489136, "learning_rate": 3.9193363312962725e-05, "loss": 0.403, "num_input_tokens_seen": 3850424, "step": 14385 }, { "epoch": 3.7719528178243773, "grad_norm": 0.8396021127700806, "learning_rate": 3.9183946523524856e-05, "loss": 0.3492, "num_input_tokens_seen": 3852008, "step": 14390 }, { "epoch": 3.7732634338138924, "grad_norm": 0.5594427585601807, "learning_rate": 3.917452676541148e-05, "loss": 0.4584, "num_input_tokens_seen": 3853160, "step": 14395 }, { "epoch": 3.7745740498034075, "grad_norm": 1.3309781551361084, "learning_rate": 3.916510404059415e-05, "loss": 0.4198, "num_input_tokens_seen": 3854568, "step": 14400 }, { "epoch": 3.7758846657929226, "grad_norm": 1.176584005355835, "learning_rate": 3.9155678351045014e-05, "loss": 0.4768, "num_input_tokens_seen": 3856280, "step": 14405 }, { "epoch": 3.7771952817824377, "grad_norm": 1.245524287223816, "learning_rate": 3.914624969873686e-05, "loss": 0.3329, "num_input_tokens_seen": 3857672, "step": 14410 }, { "epoch": 3.778505897771953, "grad_norm": 3.493666887283325, "learning_rate": 3.913681808564309e-05, "loss": 0.4928, "num_input_tokens_seen": 3858904, "step": 14415 }, { "epoch": 3.779816513761468, "grad_norm": 0.9795220494270325, "learning_rate": 3.912738351373772e-05, "loss": 0.464, "num_input_tokens_seen": 3860344, "step": 14420 }, { "epoch": 3.781127129750983, "grad_norm": 2.3869247436523438, "learning_rate": 3.911794598499539e-05, "loss": 0.4469, "num_input_tokens_seen": 3861368, "step": 14425 }, { "epoch": 3.782437745740498, "grad_norm": 1.1637309789657593, "learning_rate": 3.9108505501391355e-05, "loss": 0.4457, "num_input_tokens_seen": 3862504, "step": 14430 }, { "epoch": 3.783748361730013, "grad_norm": 0.8809982538223267, "learning_rate": 3.9099062064901497e-05, "loss": 0.3577, "num_input_tokens_seen": 3863752, "step": 14435 }, { "epoch": 3.7850589777195283, "grad_norm": 1.0453778505325317, "learning_rate": 3.908961567750231e-05, "loss": 0.3144, "num_input_tokens_seen": 3865176, "step": 14440 }, { "epoch": 3.7863695937090434, "grad_norm": 1.4418493509292603, "learning_rate": 3.908016634117092e-05, "loss": 0.4395, "num_input_tokens_seen": 3866392, "step": 14445 }, { "epoch": 3.7876802096985585, "grad_norm": 0.6362911462783813, "learning_rate": 3.907071405788503e-05, "loss": 0.4157, "num_input_tokens_seen": 3867416, "step": 14450 }, { "epoch": 3.7889908256880735, "grad_norm": 1.8845957517623901, "learning_rate": 3.9061258829622995e-05, "loss": 0.4655, "num_input_tokens_seen": 3868712, "step": 14455 }, { "epoch": 3.790301441677588, "grad_norm": 0.6175870299339294, "learning_rate": 3.905180065836379e-05, "loss": 0.2663, "num_input_tokens_seen": 3870152, "step": 14460 }, { "epoch": 3.7916120576671037, "grad_norm": 2.6970109939575195, "learning_rate": 3.904233954608699e-05, "loss": 0.4459, "num_input_tokens_seen": 3871480, "step": 14465 }, { "epoch": 3.7929226736566184, "grad_norm": 1.5881115198135376, "learning_rate": 3.903287549477279e-05, "loss": 0.291, "num_input_tokens_seen": 3872792, "step": 14470 }, { "epoch": 3.794233289646134, "grad_norm": 1.2107913494110107, "learning_rate": 3.9023408506401987e-05, "loss": 0.5015, "num_input_tokens_seen": 3873944, "step": 14475 }, { "epoch": 3.7955439056356486, "grad_norm": 1.3371944427490234, "learning_rate": 3.901393858295602e-05, "loss": 0.5388, "num_input_tokens_seen": 3876200, "step": 14480 }, { "epoch": 3.7968545216251637, "grad_norm": 1.0547019243240356, "learning_rate": 3.900446572641692e-05, "loss": 0.4268, "num_input_tokens_seen": 3877928, "step": 14485 }, { "epoch": 3.7981651376146788, "grad_norm": 0.8453646302223206, "learning_rate": 3.899498993876733e-05, "loss": 0.4591, "num_input_tokens_seen": 3878904, "step": 14490 }, { "epoch": 3.799475753604194, "grad_norm": 0.79286128282547, "learning_rate": 3.898551122199054e-05, "loss": 0.3265, "num_input_tokens_seen": 3880136, "step": 14495 }, { "epoch": 3.800786369593709, "grad_norm": 0.6775321960449219, "learning_rate": 3.8976029578070415e-05, "loss": 0.4722, "num_input_tokens_seen": 3881272, "step": 14500 }, { "epoch": 3.802096985583224, "grad_norm": 0.8248684406280518, "learning_rate": 3.896654500899145e-05, "loss": 0.4279, "num_input_tokens_seen": 3882504, "step": 14505 }, { "epoch": 3.803407601572739, "grad_norm": 0.9849101305007935, "learning_rate": 3.895705751673874e-05, "loss": 0.495, "num_input_tokens_seen": 3885368, "step": 14510 }, { "epoch": 3.8047182175622543, "grad_norm": 0.9346420168876648, "learning_rate": 3.894756710329801e-05, "loss": 0.5482, "num_input_tokens_seen": 3886904, "step": 14515 }, { "epoch": 3.8060288335517694, "grad_norm": 1.1741716861724854, "learning_rate": 3.893807377065559e-05, "loss": 0.4162, "num_input_tokens_seen": 3887976, "step": 14520 }, { "epoch": 3.8073394495412844, "grad_norm": 1.1521551609039307, "learning_rate": 3.8928577520798407e-05, "loss": 0.3655, "num_input_tokens_seen": 3889064, "step": 14525 }, { "epoch": 3.8086500655307995, "grad_norm": 0.796657145023346, "learning_rate": 3.8919078355714025e-05, "loss": 0.2186, "num_input_tokens_seen": 3890344, "step": 14530 }, { "epoch": 3.8099606815203146, "grad_norm": 1.0217984914779663, "learning_rate": 3.890957627739058e-05, "loss": 0.4583, "num_input_tokens_seen": 3892056, "step": 14535 }, { "epoch": 3.8112712975098297, "grad_norm": 0.8225256204605103, "learning_rate": 3.890007128781686e-05, "loss": 0.2696, "num_input_tokens_seen": 3893304, "step": 14540 }, { "epoch": 3.812581913499345, "grad_norm": 1.5093042850494385, "learning_rate": 3.889056338898224e-05, "loss": 0.2586, "num_input_tokens_seen": 3894872, "step": 14545 }, { "epoch": 3.81389252948886, "grad_norm": 1.3072141408920288, "learning_rate": 3.8881052582876695e-05, "loss": 0.5862, "num_input_tokens_seen": 3896504, "step": 14550 }, { "epoch": 3.8152031454783746, "grad_norm": 1.5119061470031738, "learning_rate": 3.887153887149084e-05, "loss": 0.3624, "num_input_tokens_seen": 3897768, "step": 14555 }, { "epoch": 3.81651376146789, "grad_norm": 0.9465123414993286, "learning_rate": 3.8862022256815865e-05, "loss": 0.3292, "num_input_tokens_seen": 3898872, "step": 14560 }, { "epoch": 3.8178243774574048, "grad_norm": 1.5128434896469116, "learning_rate": 3.885250274084358e-05, "loss": 0.4106, "num_input_tokens_seen": 3900008, "step": 14565 }, { "epoch": 3.8191349934469203, "grad_norm": 0.9806982278823853, "learning_rate": 3.88429803255664e-05, "loss": 0.3589, "num_input_tokens_seen": 3901064, "step": 14570 }, { "epoch": 3.820445609436435, "grad_norm": 1.508307933807373, "learning_rate": 3.883345501297737e-05, "loss": 0.3424, "num_input_tokens_seen": 3902568, "step": 14575 }, { "epoch": 3.82175622542595, "grad_norm": 1.3347301483154297, "learning_rate": 3.88239268050701e-05, "loss": 0.6653, "num_input_tokens_seen": 3904088, "step": 14580 }, { "epoch": 3.823066841415465, "grad_norm": 1.5408132076263428, "learning_rate": 3.881439570383884e-05, "loss": 0.6049, "num_input_tokens_seen": 3905464, "step": 14585 }, { "epoch": 3.8243774574049803, "grad_norm": 0.5267147421836853, "learning_rate": 3.8804861711278426e-05, "loss": 0.3739, "num_input_tokens_seen": 3906952, "step": 14590 }, { "epoch": 3.8256880733944953, "grad_norm": 2.8164420127868652, "learning_rate": 3.879532482938431e-05, "loss": 0.4786, "num_input_tokens_seen": 3908056, "step": 14595 }, { "epoch": 3.8269986893840104, "grad_norm": 0.8307332992553711, "learning_rate": 3.878578506015254e-05, "loss": 0.3639, "num_input_tokens_seen": 3909256, "step": 14600 }, { "epoch": 3.8283093053735255, "grad_norm": 1.388845443725586, "learning_rate": 3.877624240557978e-05, "loss": 0.5493, "num_input_tokens_seen": 3910424, "step": 14605 }, { "epoch": 3.8296199213630406, "grad_norm": 1.159963607788086, "learning_rate": 3.876669686766329e-05, "loss": 0.5182, "num_input_tokens_seen": 3912536, "step": 14610 }, { "epoch": 3.8309305373525557, "grad_norm": 1.9478963613510132, "learning_rate": 3.875714844840093e-05, "loss": 0.3871, "num_input_tokens_seen": 3913688, "step": 14615 }, { "epoch": 3.832241153342071, "grad_norm": 1.4422844648361206, "learning_rate": 3.874759714979117e-05, "loss": 0.351, "num_input_tokens_seen": 3914616, "step": 14620 }, { "epoch": 3.833551769331586, "grad_norm": 4.4598259925842285, "learning_rate": 3.873804297383308e-05, "loss": 0.4598, "num_input_tokens_seen": 3916120, "step": 14625 }, { "epoch": 3.834862385321101, "grad_norm": 2.870821475982666, "learning_rate": 3.872848592252634e-05, "loss": 0.3932, "num_input_tokens_seen": 3919400, "step": 14630 }, { "epoch": 3.836173001310616, "grad_norm": 1.3837535381317139, "learning_rate": 3.871892599787121e-05, "loss": 0.4162, "num_input_tokens_seen": 3920408, "step": 14635 }, { "epoch": 3.837483617300131, "grad_norm": 2.0211470127105713, "learning_rate": 3.8709363201868576e-05, "loss": 0.4557, "num_input_tokens_seen": 3921640, "step": 14640 }, { "epoch": 3.8387942332896463, "grad_norm": 1.5732158422470093, "learning_rate": 3.869979753651991e-05, "loss": 0.4548, "num_input_tokens_seen": 3922744, "step": 14645 }, { "epoch": 3.840104849279161, "grad_norm": 0.8678003549575806, "learning_rate": 3.869022900382729e-05, "loss": 0.2257, "num_input_tokens_seen": 3924152, "step": 14650 }, { "epoch": 3.8414154652686765, "grad_norm": 1.0522682666778564, "learning_rate": 3.868065760579339e-05, "loss": 0.3203, "num_input_tokens_seen": 3925848, "step": 14655 }, { "epoch": 3.842726081258191, "grad_norm": 0.9274669289588928, "learning_rate": 3.86710833444215e-05, "loss": 0.3986, "num_input_tokens_seen": 3927400, "step": 14660 }, { "epoch": 3.8440366972477067, "grad_norm": 1.61357581615448, "learning_rate": 3.866150622171549e-05, "loss": 0.4733, "num_input_tokens_seen": 3928536, "step": 14665 }, { "epoch": 3.8453473132372213, "grad_norm": 1.8953113555908203, "learning_rate": 3.8651926239679824e-05, "loss": 0.4988, "num_input_tokens_seen": 3929544, "step": 14670 }, { "epoch": 3.8466579292267364, "grad_norm": 0.7717346549034119, "learning_rate": 3.8642343400319594e-05, "loss": 0.4336, "num_input_tokens_seen": 3931096, "step": 14675 }, { "epoch": 3.8479685452162515, "grad_norm": 1.3774197101593018, "learning_rate": 3.863275770564046e-05, "loss": 0.359, "num_input_tokens_seen": 3932568, "step": 14680 }, { "epoch": 3.8492791612057666, "grad_norm": 0.6123676896095276, "learning_rate": 3.862316915764869e-05, "loss": 0.2505, "num_input_tokens_seen": 3933752, "step": 14685 }, { "epoch": 3.8505897771952817, "grad_norm": 0.7028438448905945, "learning_rate": 3.861357775835115e-05, "loss": 0.4151, "num_input_tokens_seen": 3935000, "step": 14690 }, { "epoch": 3.851900393184797, "grad_norm": 1.3766824007034302, "learning_rate": 3.8603983509755316e-05, "loss": 0.3201, "num_input_tokens_seen": 3936280, "step": 14695 }, { "epoch": 3.853211009174312, "grad_norm": 0.738145112991333, "learning_rate": 3.8594386413869235e-05, "loss": 0.3283, "num_input_tokens_seen": 3937912, "step": 14700 }, { "epoch": 3.854521625163827, "grad_norm": 0.9239733219146729, "learning_rate": 3.8584786472701575e-05, "loss": 0.4125, "num_input_tokens_seen": 3939528, "step": 14705 }, { "epoch": 3.855832241153342, "grad_norm": 0.7961690425872803, "learning_rate": 3.857518368826157e-05, "loss": 0.4321, "num_input_tokens_seen": 3940888, "step": 14710 }, { "epoch": 3.857142857142857, "grad_norm": 0.6966882944107056, "learning_rate": 3.856557806255908e-05, "loss": 0.3863, "num_input_tokens_seen": 3942296, "step": 14715 }, { "epoch": 3.8584534731323723, "grad_norm": 1.7900174856185913, "learning_rate": 3.8555969597604536e-05, "loss": 0.3552, "num_input_tokens_seen": 3943432, "step": 14720 }, { "epoch": 3.8597640891218874, "grad_norm": 1.0908517837524414, "learning_rate": 3.854635829540898e-05, "loss": 0.2647, "num_input_tokens_seen": 3944760, "step": 14725 }, { "epoch": 3.8610747051114025, "grad_norm": 0.8533638119697571, "learning_rate": 3.853674415798404e-05, "loss": 0.2777, "num_input_tokens_seen": 3947544, "step": 14730 }, { "epoch": 3.8623853211009176, "grad_norm": 0.8029381036758423, "learning_rate": 3.8527127187341936e-05, "loss": 0.3075, "num_input_tokens_seen": 3949016, "step": 14735 }, { "epoch": 3.8636959370904327, "grad_norm": 1.4114874601364136, "learning_rate": 3.8517507385495486e-05, "loss": 0.435, "num_input_tokens_seen": 3950584, "step": 14740 }, { "epoch": 3.8650065530799473, "grad_norm": 0.7523203492164612, "learning_rate": 3.85078847544581e-05, "loss": 0.582, "num_input_tokens_seen": 3951720, "step": 14745 }, { "epoch": 3.866317169069463, "grad_norm": 0.7905870676040649, "learning_rate": 3.849825929624377e-05, "loss": 0.303, "num_input_tokens_seen": 3952984, "step": 14750 }, { "epoch": 3.8676277850589775, "grad_norm": 0.9459598660469055, "learning_rate": 3.8488631012867095e-05, "loss": 0.4331, "num_input_tokens_seen": 3954680, "step": 14755 }, { "epoch": 3.8689384010484926, "grad_norm": 0.8635359406471252, "learning_rate": 3.847899990634326e-05, "loss": 0.3355, "num_input_tokens_seen": 3956024, "step": 14760 }, { "epoch": 3.8702490170380077, "grad_norm": 0.7497474551200867, "learning_rate": 3.846936597868802e-05, "loss": 0.3968, "num_input_tokens_seen": 3957320, "step": 14765 }, { "epoch": 3.871559633027523, "grad_norm": 1.3741674423217773, "learning_rate": 3.845972923191776e-05, "loss": 0.5432, "num_input_tokens_seen": 3958376, "step": 14770 }, { "epoch": 3.872870249017038, "grad_norm": 0.9989532828330994, "learning_rate": 3.845008966804944e-05, "loss": 0.3567, "num_input_tokens_seen": 3959624, "step": 14775 }, { "epoch": 3.874180865006553, "grad_norm": 1.1938927173614502, "learning_rate": 3.844044728910058e-05, "loss": 0.37, "num_input_tokens_seen": 3960760, "step": 14780 }, { "epoch": 3.875491480996068, "grad_norm": 1.734110713005066, "learning_rate": 3.843080209708933e-05, "loss": 0.3604, "num_input_tokens_seen": 3962504, "step": 14785 }, { "epoch": 3.876802096985583, "grad_norm": 0.6229527592658997, "learning_rate": 3.8421154094034404e-05, "loss": 0.3337, "num_input_tokens_seen": 3964200, "step": 14790 }, { "epoch": 3.8781127129750983, "grad_norm": 0.9114784002304077, "learning_rate": 3.841150328195512e-05, "loss": 0.3229, "num_input_tokens_seen": 3965192, "step": 14795 }, { "epoch": 3.8794233289646134, "grad_norm": 1.4485070705413818, "learning_rate": 3.840184966287137e-05, "loss": 0.4557, "num_input_tokens_seen": 3966744, "step": 14800 }, { "epoch": 3.8807339449541285, "grad_norm": 0.7846187353134155, "learning_rate": 3.839219323880365e-05, "loss": 0.3656, "num_input_tokens_seen": 3968440, "step": 14805 }, { "epoch": 3.8820445609436436, "grad_norm": 0.8970635533332825, "learning_rate": 3.838253401177302e-05, "loss": 0.4274, "num_input_tokens_seen": 3970072, "step": 14810 }, { "epoch": 3.8833551769331587, "grad_norm": 0.5233820080757141, "learning_rate": 3.837287198380114e-05, "loss": 0.3428, "num_input_tokens_seen": 3971112, "step": 14815 }, { "epoch": 3.8846657929226738, "grad_norm": 9.287395477294922, "learning_rate": 3.836320715691027e-05, "loss": 0.4686, "num_input_tokens_seen": 3972120, "step": 14820 }, { "epoch": 3.885976408912189, "grad_norm": 1.6292757987976074, "learning_rate": 3.835353953312322e-05, "loss": 0.3802, "num_input_tokens_seen": 3973448, "step": 14825 }, { "epoch": 3.8872870249017035, "grad_norm": 0.5460254549980164, "learning_rate": 3.8343869114463424e-05, "loss": 0.5609, "num_input_tokens_seen": 3975192, "step": 14830 }, { "epoch": 3.888597640891219, "grad_norm": 1.455540657043457, "learning_rate": 3.8334195902954885e-05, "loss": 0.4602, "num_input_tokens_seen": 3976344, "step": 14835 }, { "epoch": 3.8899082568807337, "grad_norm": 0.7341175675392151, "learning_rate": 3.8324519900622175e-05, "loss": 0.4039, "num_input_tokens_seen": 3977832, "step": 14840 }, { "epoch": 3.8912188728702493, "grad_norm": 0.9172179102897644, "learning_rate": 3.8314841109490474e-05, "loss": 0.3801, "num_input_tokens_seen": 3979128, "step": 14845 }, { "epoch": 3.892529488859764, "grad_norm": 1.2292873859405518, "learning_rate": 3.830515953158553e-05, "loss": 0.453, "num_input_tokens_seen": 3980248, "step": 14850 }, { "epoch": 3.893840104849279, "grad_norm": 1.3497289419174194, "learning_rate": 3.82954751689337e-05, "loss": 0.3097, "num_input_tokens_seen": 3981320, "step": 14855 }, { "epoch": 3.895150720838794, "grad_norm": 1.0957204103469849, "learning_rate": 3.828578802356188e-05, "loss": 0.5522, "num_input_tokens_seen": 3982584, "step": 14860 }, { "epoch": 3.896461336828309, "grad_norm": 1.262648582458496, "learning_rate": 3.8276098097497584e-05, "loss": 0.2894, "num_input_tokens_seen": 3983928, "step": 14865 }, { "epoch": 3.8977719528178243, "grad_norm": 0.6311514973640442, "learning_rate": 3.8266405392768904e-05, "loss": 0.3871, "num_input_tokens_seen": 3985816, "step": 14870 }, { "epoch": 3.8990825688073394, "grad_norm": 1.0741389989852905, "learning_rate": 3.8256709911404484e-05, "loss": 0.389, "num_input_tokens_seen": 3987208, "step": 14875 }, { "epoch": 3.9003931847968545, "grad_norm": 0.9446879625320435, "learning_rate": 3.82470116554336e-05, "loss": 0.4234, "num_input_tokens_seen": 3988440, "step": 14880 }, { "epoch": 3.9017038007863696, "grad_norm": 1.8258028030395508, "learning_rate": 3.823731062688605e-05, "loss": 0.3502, "num_input_tokens_seen": 3989464, "step": 14885 }, { "epoch": 3.9030144167758847, "grad_norm": 1.2829755544662476, "learning_rate": 3.8227606827792265e-05, "loss": 0.3685, "num_input_tokens_seen": 3990872, "step": 14890 }, { "epoch": 3.9043250327653998, "grad_norm": 1.3106721639633179, "learning_rate": 3.8217900260183224e-05, "loss": 0.3787, "num_input_tokens_seen": 3992008, "step": 14895 }, { "epoch": 3.905635648754915, "grad_norm": 0.6422778367996216, "learning_rate": 3.82081909260905e-05, "loss": 0.2709, "num_input_tokens_seen": 3993288, "step": 14900 }, { "epoch": 3.90694626474443, "grad_norm": 0.6372278928756714, "learning_rate": 3.819847882754623e-05, "loss": 0.391, "num_input_tokens_seen": 3994728, "step": 14905 }, { "epoch": 3.908256880733945, "grad_norm": 0.657509446144104, "learning_rate": 3.8188763966583165e-05, "loss": 0.3405, "num_input_tokens_seen": 3995928, "step": 14910 }, { "epoch": 3.90956749672346, "grad_norm": 0.9248924255371094, "learning_rate": 3.817904634523458e-05, "loss": 0.4356, "num_input_tokens_seen": 3997208, "step": 14915 }, { "epoch": 3.9108781127129753, "grad_norm": 1.8870102167129517, "learning_rate": 3.816932596553436e-05, "loss": 0.3586, "num_input_tokens_seen": 3998808, "step": 14920 }, { "epoch": 3.91218872870249, "grad_norm": 0.7947030663490295, "learning_rate": 3.8159602829516975e-05, "loss": 0.4333, "num_input_tokens_seen": 4000232, "step": 14925 }, { "epoch": 3.9134993446920054, "grad_norm": 1.5478394031524658, "learning_rate": 3.8149876939217454e-05, "loss": 0.3612, "num_input_tokens_seen": 4001256, "step": 14930 }, { "epoch": 3.91480996068152, "grad_norm": 1.3280823230743408, "learning_rate": 3.814014829667142e-05, "loss": 0.3201, "num_input_tokens_seen": 4002568, "step": 14935 }, { "epoch": 3.9161205766710356, "grad_norm": 0.8468579649925232, "learning_rate": 3.8130416903915045e-05, "loss": 0.3906, "num_input_tokens_seen": 4003976, "step": 14940 }, { "epoch": 3.9174311926605503, "grad_norm": 0.8908202052116394, "learning_rate": 3.812068276298509e-05, "loss": 0.353, "num_input_tokens_seen": 4005848, "step": 14945 }, { "epoch": 3.9187418086500654, "grad_norm": 4.807361125946045, "learning_rate": 3.811094587591892e-05, "loss": 0.3844, "num_input_tokens_seen": 4006808, "step": 14950 }, { "epoch": 3.9200524246395805, "grad_norm": 1.0162426233291626, "learning_rate": 3.810120624475443e-05, "loss": 0.3416, "num_input_tokens_seen": 4008184, "step": 14955 }, { "epoch": 3.9213630406290956, "grad_norm": 0.6243327856063843, "learning_rate": 3.809146387153011e-05, "loss": 0.384, "num_input_tokens_seen": 4009656, "step": 14960 }, { "epoch": 3.9226736566186107, "grad_norm": 0.8414437174797058, "learning_rate": 3.808171875828501e-05, "loss": 0.518, "num_input_tokens_seen": 4010760, "step": 14965 }, { "epoch": 3.9239842726081258, "grad_norm": 2.6277472972869873, "learning_rate": 3.8071970907058786e-05, "loss": 0.6566, "num_input_tokens_seen": 4011752, "step": 14970 }, { "epoch": 3.925294888597641, "grad_norm": 0.9146928191184998, "learning_rate": 3.806222031989164e-05, "loss": 0.4656, "num_input_tokens_seen": 4012872, "step": 14975 }, { "epoch": 3.926605504587156, "grad_norm": 0.7698646187782288, "learning_rate": 3.8052466998824344e-05, "loss": 0.2866, "num_input_tokens_seen": 4013896, "step": 14980 }, { "epoch": 3.927916120576671, "grad_norm": 0.6607807874679565, "learning_rate": 3.804271094589826e-05, "loss": 0.4339, "num_input_tokens_seen": 4015560, "step": 14985 }, { "epoch": 3.929226736566186, "grad_norm": 2.0358328819274902, "learning_rate": 3.803295216315532e-05, "loss": 0.3704, "num_input_tokens_seen": 4016696, "step": 14990 }, { "epoch": 3.9305373525557012, "grad_norm": 1.2310099601745605, "learning_rate": 3.802319065263801e-05, "loss": 0.3665, "num_input_tokens_seen": 4017800, "step": 14995 }, { "epoch": 3.9318479685452163, "grad_norm": 1.9436471462249756, "learning_rate": 3.8013426416389385e-05, "loss": 0.388, "num_input_tokens_seen": 4019224, "step": 15000 }, { "epoch": 3.9331585845347314, "grad_norm": 1.6127839088439941, "learning_rate": 3.80036594564531e-05, "loss": 0.3805, "num_input_tokens_seen": 4020488, "step": 15005 }, { "epoch": 3.9344692005242465, "grad_norm": 0.9893333911895752, "learning_rate": 3.799388977487337e-05, "loss": 0.3402, "num_input_tokens_seen": 4021592, "step": 15010 }, { "epoch": 3.9357798165137616, "grad_norm": 0.9352397322654724, "learning_rate": 3.798411737369495e-05, "loss": 0.3509, "num_input_tokens_seen": 4022728, "step": 15015 }, { "epoch": 3.9370904325032763, "grad_norm": 1.0618655681610107, "learning_rate": 3.7974342254963214e-05, "loss": 0.7432, "num_input_tokens_seen": 4024328, "step": 15020 }, { "epoch": 3.938401048492792, "grad_norm": 0.6937621831893921, "learning_rate": 3.7964564420724044e-05, "loss": 0.4464, "num_input_tokens_seen": 4025880, "step": 15025 }, { "epoch": 3.9397116644823065, "grad_norm": 0.7405502200126648, "learning_rate": 3.7954783873023955e-05, "loss": 0.7156, "num_input_tokens_seen": 4027192, "step": 15030 }, { "epoch": 3.941022280471822, "grad_norm": 2.1510064601898193, "learning_rate": 3.7945000613909975e-05, "loss": 0.3292, "num_input_tokens_seen": 4028536, "step": 15035 }, { "epoch": 3.9423328964613367, "grad_norm": 1.9360071420669556, "learning_rate": 3.793521464542974e-05, "loss": 0.3899, "num_input_tokens_seen": 4029944, "step": 15040 }, { "epoch": 3.9436435124508518, "grad_norm": 1.3576067686080933, "learning_rate": 3.7925425969631435e-05, "loss": 0.4549, "num_input_tokens_seen": 4031080, "step": 15045 }, { "epoch": 3.944954128440367, "grad_norm": 0.7620101571083069, "learning_rate": 3.7915634588563804e-05, "loss": 0.2826, "num_input_tokens_seen": 4033128, "step": 15050 }, { "epoch": 3.946264744429882, "grad_norm": 1.4963370561599731, "learning_rate": 3.790584050427616e-05, "loss": 0.4468, "num_input_tokens_seen": 4034472, "step": 15055 }, { "epoch": 3.947575360419397, "grad_norm": 0.649645209312439, "learning_rate": 3.7896043718818406e-05, "loss": 0.4449, "num_input_tokens_seen": 4035768, "step": 15060 }, { "epoch": 3.948885976408912, "grad_norm": 1.2839727401733398, "learning_rate": 3.788624423424099e-05, "loss": 0.4876, "num_input_tokens_seen": 4036984, "step": 15065 }, { "epoch": 3.9501965923984272, "grad_norm": 1.5318702459335327, "learning_rate": 3.7876442052594915e-05, "loss": 0.6981, "num_input_tokens_seen": 4038184, "step": 15070 }, { "epoch": 3.9515072083879423, "grad_norm": 0.9313368201255798, "learning_rate": 3.7866637175931774e-05, "loss": 0.3878, "num_input_tokens_seen": 4039448, "step": 15075 }, { "epoch": 3.9528178243774574, "grad_norm": 0.784034252166748, "learning_rate": 3.7856829606303696e-05, "loss": 0.4549, "num_input_tokens_seen": 4040792, "step": 15080 }, { "epoch": 3.9541284403669725, "grad_norm": 1.075270175933838, "learning_rate": 3.784701934576339e-05, "loss": 0.3993, "num_input_tokens_seen": 4042008, "step": 15085 }, { "epoch": 3.9554390563564876, "grad_norm": 0.3716960549354553, "learning_rate": 3.783720639636415e-05, "loss": 0.3188, "num_input_tokens_seen": 4043368, "step": 15090 }, { "epoch": 3.9567496723460027, "grad_norm": 0.5645270943641663, "learning_rate": 3.782739076015978e-05, "loss": 0.5041, "num_input_tokens_seen": 4044472, "step": 15095 }, { "epoch": 3.958060288335518, "grad_norm": 1.789473056793213, "learning_rate": 3.7817572439204695e-05, "loss": 0.2873, "num_input_tokens_seen": 4045400, "step": 15100 }, { "epoch": 3.959370904325033, "grad_norm": 1.310235857963562, "learning_rate": 3.780775143555384e-05, "loss": 0.3193, "num_input_tokens_seen": 4046760, "step": 15105 }, { "epoch": 3.960681520314548, "grad_norm": 0.3856951892375946, "learning_rate": 3.779792775126275e-05, "loss": 0.4755, "num_input_tokens_seen": 4048360, "step": 15110 }, { "epoch": 3.9619921363040627, "grad_norm": 2.0929763317108154, "learning_rate": 3.778810138838748e-05, "loss": 0.2977, "num_input_tokens_seen": 4049528, "step": 15115 }, { "epoch": 3.963302752293578, "grad_norm": 1.458464503288269, "learning_rate": 3.777827234898469e-05, "loss": 0.5211, "num_input_tokens_seen": 4050664, "step": 15120 }, { "epoch": 3.964613368283093, "grad_norm": 0.7064275145530701, "learning_rate": 3.776844063511158e-05, "loss": 0.4641, "num_input_tokens_seen": 4052216, "step": 15125 }, { "epoch": 3.9659239842726084, "grad_norm": 0.6829923987388611, "learning_rate": 3.7758606248825914e-05, "loss": 0.4749, "num_input_tokens_seen": 4053720, "step": 15130 }, { "epoch": 3.967234600262123, "grad_norm": 1.2913763523101807, "learning_rate": 3.774876919218599e-05, "loss": 0.5882, "num_input_tokens_seen": 4055080, "step": 15135 }, { "epoch": 3.968545216251638, "grad_norm": 0.577420175075531, "learning_rate": 3.773892946725071e-05, "loss": 0.3344, "num_input_tokens_seen": 4056328, "step": 15140 }, { "epoch": 3.9698558322411532, "grad_norm": 1.1378531455993652, "learning_rate": 3.7729087076079505e-05, "loss": 0.4703, "num_input_tokens_seen": 4057944, "step": 15145 }, { "epoch": 3.9711664482306683, "grad_norm": 0.9726043343544006, "learning_rate": 3.771924202073236e-05, "loss": 0.3981, "num_input_tokens_seen": 4059128, "step": 15150 }, { "epoch": 3.9724770642201834, "grad_norm": 1.0263450145721436, "learning_rate": 3.770939430326985e-05, "loss": 0.3403, "num_input_tokens_seen": 4060472, "step": 15155 }, { "epoch": 3.9737876802096985, "grad_norm": 0.7833340167999268, "learning_rate": 3.7699543925753064e-05, "loss": 0.3425, "num_input_tokens_seen": 4061608, "step": 15160 }, { "epoch": 3.9750982961992136, "grad_norm": 0.7336474061012268, "learning_rate": 3.768969089024368e-05, "loss": 0.4062, "num_input_tokens_seen": 4063624, "step": 15165 }, { "epoch": 3.9764089121887287, "grad_norm": 1.0704506635665894, "learning_rate": 3.767983519880392e-05, "loss": 0.3281, "num_input_tokens_seen": 4066152, "step": 15170 }, { "epoch": 3.977719528178244, "grad_norm": 3.6577582359313965, "learning_rate": 3.766997685349656e-05, "loss": 0.4033, "num_input_tokens_seen": 4067304, "step": 15175 }, { "epoch": 3.979030144167759, "grad_norm": 0.7972487807273865, "learning_rate": 3.766011585638494e-05, "loss": 0.3086, "num_input_tokens_seen": 4068456, "step": 15180 }, { "epoch": 3.980340760157274, "grad_norm": 0.7862528562545776, "learning_rate": 3.765025220953294e-05, "loss": 0.2291, "num_input_tokens_seen": 4069864, "step": 15185 }, { "epoch": 3.981651376146789, "grad_norm": 0.8994476795196533, "learning_rate": 3.764038591500502e-05, "loss": 0.3518, "num_input_tokens_seen": 4071096, "step": 15190 }, { "epoch": 3.982961992136304, "grad_norm": 1.3877050876617432, "learning_rate": 3.7630516974866166e-05, "loss": 0.621, "num_input_tokens_seen": 4072504, "step": 15195 }, { "epoch": 3.9842726081258193, "grad_norm": 1.0842753648757935, "learning_rate": 3.7620645391181926e-05, "loss": 0.3174, "num_input_tokens_seen": 4073880, "step": 15200 }, { "epoch": 3.9855832241153344, "grad_norm": 0.3286502957344055, "learning_rate": 3.761077116601842e-05, "loss": 0.3758, "num_input_tokens_seen": 4075096, "step": 15205 }, { "epoch": 3.986893840104849, "grad_norm": 0.9006760716438293, "learning_rate": 3.760089430144229e-05, "loss": 0.6088, "num_input_tokens_seen": 4076648, "step": 15210 }, { "epoch": 3.9882044560943646, "grad_norm": 1.3272546529769897, "learning_rate": 3.759101479952076e-05, "loss": 0.4566, "num_input_tokens_seen": 4078088, "step": 15215 }, { "epoch": 3.9895150720838792, "grad_norm": 0.4521055221557617, "learning_rate": 3.7581132662321597e-05, "loss": 0.4513, "num_input_tokens_seen": 4079416, "step": 15220 }, { "epoch": 3.9908256880733948, "grad_norm": 2.0390663146972656, "learning_rate": 3.7571247891913095e-05, "loss": 0.374, "num_input_tokens_seen": 4081288, "step": 15225 }, { "epoch": 3.9921363040629094, "grad_norm": 1.4461923837661743, "learning_rate": 3.756136049036413e-05, "loss": 0.591, "num_input_tokens_seen": 4082584, "step": 15230 }, { "epoch": 3.9934469200524245, "grad_norm": 0.7820754647254944, "learning_rate": 3.7551470459744126e-05, "loss": 0.3588, "num_input_tokens_seen": 4083848, "step": 15235 }, { "epoch": 3.9947575360419396, "grad_norm": 2.2004661560058594, "learning_rate": 3.754157780212304e-05, "loss": 0.4292, "num_input_tokens_seen": 4085160, "step": 15240 }, { "epoch": 3.9960681520314547, "grad_norm": 0.369859516620636, "learning_rate": 3.753168251957139e-05, "loss": 0.2214, "num_input_tokens_seen": 4086600, "step": 15245 }, { "epoch": 3.99737876802097, "grad_norm": 0.7584067583084106, "learning_rate": 3.752178461416024e-05, "loss": 0.3093, "num_input_tokens_seen": 4087736, "step": 15250 }, { "epoch": 3.998689384010485, "grad_norm": 2.6543655395507812, "learning_rate": 3.75118840879612e-05, "loss": 0.4824, "num_input_tokens_seen": 4089000, "step": 15255 }, { "epoch": 4.0, "grad_norm": 0.9112398624420166, "learning_rate": 3.750198094304644e-05, "loss": 0.4852, "num_input_tokens_seen": 4089872, "step": 15260 }, { "epoch": 4.001048492791612, "eval_loss": 0.5102021098136902, "eval_runtime": 16.7075, "eval_samples_per_second": 50.756, "eval_steps_per_second": 25.378, "num_input_tokens_seen": 4090736, "step": 15264 }, { "epoch": 4.001310615989515, "grad_norm": 0.5710874199867249, "learning_rate": 3.749207518148869e-05, "loss": 0.4124, "num_input_tokens_seen": 4091024, "step": 15265 }, { "epoch": 4.00262123197903, "grad_norm": 1.6767256259918213, "learning_rate": 3.7482166805361175e-05, "loss": 0.2868, "num_input_tokens_seen": 4092256, "step": 15270 }, { "epoch": 4.003931847968545, "grad_norm": 2.018289089202881, "learning_rate": 3.747225581673771e-05, "loss": 0.2686, "num_input_tokens_seen": 4093168, "step": 15275 }, { "epoch": 4.00524246395806, "grad_norm": 0.3900015652179718, "learning_rate": 3.746234221769266e-05, "loss": 0.3011, "num_input_tokens_seen": 4094480, "step": 15280 }, { "epoch": 4.006553079947575, "grad_norm": 1.014138102531433, "learning_rate": 3.7452426010300915e-05, "loss": 1.0277, "num_input_tokens_seen": 4095472, "step": 15285 }, { "epoch": 4.007863695937091, "grad_norm": 1.6088920831680298, "learning_rate": 3.7442507196637925e-05, "loss": 0.395, "num_input_tokens_seen": 4096656, "step": 15290 }, { "epoch": 4.009174311926605, "grad_norm": 1.3494174480438232, "learning_rate": 3.743258577877968e-05, "loss": 0.4907, "num_input_tokens_seen": 4097760, "step": 15295 }, { "epoch": 4.010484927916121, "grad_norm": 1.1706165075302124, "learning_rate": 3.74226617588027e-05, "loss": 0.2615, "num_input_tokens_seen": 4099168, "step": 15300 }, { "epoch": 4.011795543905635, "grad_norm": 0.6922947764396667, "learning_rate": 3.7412735138784096e-05, "loss": 0.3265, "num_input_tokens_seen": 4100576, "step": 15305 }, { "epoch": 4.013106159895151, "grad_norm": 1.1416230201721191, "learning_rate": 3.740280592080147e-05, "loss": 0.35, "num_input_tokens_seen": 4101904, "step": 15310 }, { "epoch": 4.014416775884666, "grad_norm": 1.0017750263214111, "learning_rate": 3.7392874106932985e-05, "loss": 0.3358, "num_input_tokens_seen": 4103584, "step": 15315 }, { "epoch": 4.015727391874181, "grad_norm": 1.882419228553772, "learning_rate": 3.738293969925737e-05, "loss": 0.3393, "num_input_tokens_seen": 4104640, "step": 15320 }, { "epoch": 4.017038007863696, "grad_norm": 0.7629048228263855, "learning_rate": 3.737300269985388e-05, "loss": 0.328, "num_input_tokens_seen": 4106272, "step": 15325 }, { "epoch": 4.018348623853211, "grad_norm": 0.9349744915962219, "learning_rate": 3.736306311080229e-05, "loss": 0.382, "num_input_tokens_seen": 4107808, "step": 15330 }, { "epoch": 4.019659239842726, "grad_norm": 2.474996566772461, "learning_rate": 3.7353120934182956e-05, "loss": 0.5427, "num_input_tokens_seen": 4109152, "step": 15335 }, { "epoch": 4.0209698558322415, "grad_norm": 0.7123661041259766, "learning_rate": 3.7343176172076755e-05, "loss": 0.2391, "num_input_tokens_seen": 4110496, "step": 15340 }, { "epoch": 4.022280471821756, "grad_norm": 0.500874400138855, "learning_rate": 3.7333228826565115e-05, "loss": 0.2977, "num_input_tokens_seen": 4111824, "step": 15345 }, { "epoch": 4.023591087811272, "grad_norm": 2.4373834133148193, "learning_rate": 3.7323278899729986e-05, "loss": 0.3753, "num_input_tokens_seen": 4112880, "step": 15350 }, { "epoch": 4.024901703800786, "grad_norm": 3.1125118732452393, "learning_rate": 3.731332639365387e-05, "loss": 0.3627, "num_input_tokens_seen": 4113840, "step": 15355 }, { "epoch": 4.026212319790301, "grad_norm": 0.9370334148406982, "learning_rate": 3.7303371310419835e-05, "loss": 0.3914, "num_input_tokens_seen": 4114800, "step": 15360 }, { "epoch": 4.027522935779817, "grad_norm": 1.220414638519287, "learning_rate": 3.729341365211143e-05, "loss": 0.3767, "num_input_tokens_seen": 4116416, "step": 15365 }, { "epoch": 4.028833551769331, "grad_norm": 1.3489309549331665, "learning_rate": 3.7283453420812786e-05, "loss": 0.4277, "num_input_tokens_seen": 4117584, "step": 15370 }, { "epoch": 4.030144167758847, "grad_norm": 1.3635720014572144, "learning_rate": 3.727349061860857e-05, "loss": 0.4517, "num_input_tokens_seen": 4118848, "step": 15375 }, { "epoch": 4.031454783748361, "grad_norm": 1.2767181396484375, "learning_rate": 3.726352524758397e-05, "loss": 0.3096, "num_input_tokens_seen": 4120032, "step": 15380 }, { "epoch": 4.032765399737877, "grad_norm": 0.9051564335823059, "learning_rate": 3.725355730982474e-05, "loss": 0.3104, "num_input_tokens_seen": 4121680, "step": 15385 }, { "epoch": 4.034076015727392, "grad_norm": 1.529902458190918, "learning_rate": 3.724358680741713e-05, "loss": 0.2001, "num_input_tokens_seen": 4122800, "step": 15390 }, { "epoch": 4.035386631716907, "grad_norm": 0.8181842565536499, "learning_rate": 3.723361374244795e-05, "loss": 0.3889, "num_input_tokens_seen": 4124336, "step": 15395 }, { "epoch": 4.036697247706422, "grad_norm": 0.6840779781341553, "learning_rate": 3.722363811700455e-05, "loss": 0.431, "num_input_tokens_seen": 4125904, "step": 15400 }, { "epoch": 4.038007863695937, "grad_norm": 2.819944381713867, "learning_rate": 3.721365993317482e-05, "loss": 0.3311, "num_input_tokens_seen": 4127376, "step": 15405 }, { "epoch": 4.039318479685452, "grad_norm": 1.1946598291397095, "learning_rate": 3.720367919304717e-05, "loss": 0.3918, "num_input_tokens_seen": 4128944, "step": 15410 }, { "epoch": 4.0406290956749675, "grad_norm": 0.767782986164093, "learning_rate": 3.7193695898710545e-05, "loss": 0.2845, "num_input_tokens_seen": 4130176, "step": 15415 }, { "epoch": 4.041939711664482, "grad_norm": 1.2381718158721924, "learning_rate": 3.718371005225445e-05, "loss": 0.2966, "num_input_tokens_seen": 4131680, "step": 15420 }, { "epoch": 4.043250327653998, "grad_norm": 0.7335060238838196, "learning_rate": 3.717372165576888e-05, "loss": 0.3541, "num_input_tokens_seen": 4132736, "step": 15425 }, { "epoch": 4.044560943643512, "grad_norm": 1.9448546171188354, "learning_rate": 3.71637307113444e-05, "loss": 0.2301, "num_input_tokens_seen": 4134240, "step": 15430 }, { "epoch": 4.045871559633028, "grad_norm": 1.4325217008590698, "learning_rate": 3.715373722107211e-05, "loss": 0.3728, "num_input_tokens_seen": 4135616, "step": 15435 }, { "epoch": 4.047182175622543, "grad_norm": 0.9633174538612366, "learning_rate": 3.714374118704362e-05, "loss": 0.3703, "num_input_tokens_seen": 4137376, "step": 15440 }, { "epoch": 4.048492791612058, "grad_norm": 1.3096901178359985, "learning_rate": 3.713374261135107e-05, "loss": 0.5855, "num_input_tokens_seen": 4138640, "step": 15445 }, { "epoch": 4.049803407601573, "grad_norm": 1.610138177871704, "learning_rate": 3.712374149608717e-05, "loss": 0.3106, "num_input_tokens_seen": 4140032, "step": 15450 }, { "epoch": 4.051114023591087, "grad_norm": 1.589325189590454, "learning_rate": 3.711373784334511e-05, "loss": 0.563, "num_input_tokens_seen": 4141488, "step": 15455 }, { "epoch": 4.052424639580603, "grad_norm": 1.1939072608947754, "learning_rate": 3.7103731655218664e-05, "loss": 0.5036, "num_input_tokens_seen": 4142864, "step": 15460 }, { "epoch": 4.053735255570118, "grad_norm": 1.4312946796417236, "learning_rate": 3.7093722933802095e-05, "loss": 0.2945, "num_input_tokens_seen": 4145888, "step": 15465 }, { "epoch": 4.055045871559633, "grad_norm": 1.0872986316680908, "learning_rate": 3.70837116811902e-05, "loss": 0.322, "num_input_tokens_seen": 4147264, "step": 15470 }, { "epoch": 4.056356487549148, "grad_norm": 0.3884176015853882, "learning_rate": 3.707369789947834e-05, "loss": 0.3428, "num_input_tokens_seen": 4149184, "step": 15475 }, { "epoch": 4.057667103538663, "grad_norm": 2.0219218730926514, "learning_rate": 3.706368159076238e-05, "loss": 0.4422, "num_input_tokens_seen": 4150624, "step": 15480 }, { "epoch": 4.058977719528178, "grad_norm": 5.671409606933594, "learning_rate": 3.7053662757138686e-05, "loss": 0.378, "num_input_tokens_seen": 4152256, "step": 15485 }, { "epoch": 4.0602883355176935, "grad_norm": 0.765018105506897, "learning_rate": 3.704364140070421e-05, "loss": 0.3426, "num_input_tokens_seen": 4153552, "step": 15490 }, { "epoch": 4.061598951507208, "grad_norm": 1.3567652702331543, "learning_rate": 3.703361752355641e-05, "loss": 0.3171, "num_input_tokens_seen": 4154800, "step": 15495 }, { "epoch": 4.062909567496724, "grad_norm": 1.2356395721435547, "learning_rate": 3.702359112779325e-05, "loss": 0.227, "num_input_tokens_seen": 4155872, "step": 15500 }, { "epoch": 4.064220183486238, "grad_norm": 1.230805516242981, "learning_rate": 3.701356221551324e-05, "loss": 0.3697, "num_input_tokens_seen": 4157200, "step": 15505 }, { "epoch": 4.065530799475754, "grad_norm": 1.0596007108688354, "learning_rate": 3.700353078881541e-05, "loss": 0.2447, "num_input_tokens_seen": 4158416, "step": 15510 }, { "epoch": 4.066841415465269, "grad_norm": 0.5798215270042419, "learning_rate": 3.699349684979932e-05, "loss": 0.3802, "num_input_tokens_seen": 4159888, "step": 15515 }, { "epoch": 4.068152031454784, "grad_norm": 0.8329735398292542, "learning_rate": 3.6983460400565086e-05, "loss": 0.1836, "num_input_tokens_seen": 4161120, "step": 15520 }, { "epoch": 4.069462647444299, "grad_norm": 1.9830948114395142, "learning_rate": 3.697342144321329e-05, "loss": 0.6108, "num_input_tokens_seen": 4162288, "step": 15525 }, { "epoch": 4.070773263433814, "grad_norm": 1.2912129163742065, "learning_rate": 3.6963379979845066e-05, "loss": 0.4895, "num_input_tokens_seen": 4165056, "step": 15530 }, { "epoch": 4.072083879423329, "grad_norm": 1.4555463790893555, "learning_rate": 3.6953336012562084e-05, "loss": 0.3561, "num_input_tokens_seen": 4166656, "step": 15535 }, { "epoch": 4.073394495412844, "grad_norm": 1.276933193206787, "learning_rate": 3.6943289543466536e-05, "loss": 0.255, "num_input_tokens_seen": 4167728, "step": 15540 }, { "epoch": 4.074705111402359, "grad_norm": 1.2360544204711914, "learning_rate": 3.693324057466111e-05, "loss": 0.4359, "num_input_tokens_seen": 4169120, "step": 15545 }, { "epoch": 4.076015727391874, "grad_norm": 0.9674568176269531, "learning_rate": 3.692318910824906e-05, "loss": 0.2721, "num_input_tokens_seen": 4170864, "step": 15550 }, { "epoch": 4.077326343381389, "grad_norm": 1.194679856300354, "learning_rate": 3.691313514633413e-05, "loss": 0.4647, "num_input_tokens_seen": 4171824, "step": 15555 }, { "epoch": 4.078636959370904, "grad_norm": 1.1207356452941895, "learning_rate": 3.69030786910206e-05, "loss": 0.359, "num_input_tokens_seen": 4173056, "step": 15560 }, { "epoch": 4.0799475753604195, "grad_norm": 0.9049304127693176, "learning_rate": 3.6893019744413265e-05, "loss": 0.2858, "num_input_tokens_seen": 4174080, "step": 15565 }, { "epoch": 4.081258191349934, "grad_norm": 1.6314843893051147, "learning_rate": 3.688295830861744e-05, "loss": 0.3283, "num_input_tokens_seen": 4175136, "step": 15570 }, { "epoch": 4.08256880733945, "grad_norm": 0.960339367389679, "learning_rate": 3.687289438573899e-05, "loss": 0.2939, "num_input_tokens_seen": 4176208, "step": 15575 }, { "epoch": 4.083879423328964, "grad_norm": 0.5315578579902649, "learning_rate": 3.6862827977884246e-05, "loss": 0.3777, "num_input_tokens_seen": 4177680, "step": 15580 }, { "epoch": 4.08519003931848, "grad_norm": 1.7825731039047241, "learning_rate": 3.6852759087160105e-05, "loss": 0.3047, "num_input_tokens_seen": 4179632, "step": 15585 }, { "epoch": 4.086500655307995, "grad_norm": 1.3463691473007202, "learning_rate": 3.6842687715673977e-05, "loss": 0.3382, "num_input_tokens_seen": 4180992, "step": 15590 }, { "epoch": 4.08781127129751, "grad_norm": 1.9474496841430664, "learning_rate": 3.683261386553376e-05, "loss": 0.3723, "num_input_tokens_seen": 4182608, "step": 15595 }, { "epoch": 4.089121887287025, "grad_norm": 0.9942028522491455, "learning_rate": 3.6822537538847914e-05, "loss": 0.4365, "num_input_tokens_seen": 4183952, "step": 15600 }, { "epoch": 4.09043250327654, "grad_norm": 1.529155969619751, "learning_rate": 3.6812458737725384e-05, "loss": 0.3529, "num_input_tokens_seen": 4185920, "step": 15605 }, { "epoch": 4.091743119266055, "grad_norm": 2.3804378509521484, "learning_rate": 3.680237746427565e-05, "loss": 0.3824, "num_input_tokens_seen": 4186896, "step": 15610 }, { "epoch": 4.0930537352555705, "grad_norm": 1.0819133520126343, "learning_rate": 3.679229372060871e-05, "loss": 0.3386, "num_input_tokens_seen": 4188032, "step": 15615 }, { "epoch": 4.094364351245085, "grad_norm": 0.5615158081054688, "learning_rate": 3.678220750883507e-05, "loss": 0.2913, "num_input_tokens_seen": 4189728, "step": 15620 }, { "epoch": 4.095674967234601, "grad_norm": 1.6393334865570068, "learning_rate": 3.6772118831065755e-05, "loss": 0.4371, "num_input_tokens_seen": 4191024, "step": 15625 }, { "epoch": 4.096985583224115, "grad_norm": 1.589302897453308, "learning_rate": 3.6762027689412305e-05, "loss": 0.5434, "num_input_tokens_seen": 4192672, "step": 15630 }, { "epoch": 4.09829619921363, "grad_norm": 1.0135445594787598, "learning_rate": 3.675193408598679e-05, "loss": 0.4229, "num_input_tokens_seen": 4194144, "step": 15635 }, { "epoch": 4.0996068152031455, "grad_norm": 0.9298484325408936, "learning_rate": 3.674183802290178e-05, "loss": 0.3345, "num_input_tokens_seen": 4196128, "step": 15640 }, { "epoch": 4.10091743119266, "grad_norm": 1.246108889579773, "learning_rate": 3.6731739502270344e-05, "loss": 0.3567, "num_input_tokens_seen": 4197856, "step": 15645 }, { "epoch": 4.102228047182176, "grad_norm": 4.304207801818848, "learning_rate": 3.6721638526206115e-05, "loss": 0.3401, "num_input_tokens_seen": 4199008, "step": 15650 }, { "epoch": 4.10353866317169, "grad_norm": 0.5303115844726562, "learning_rate": 3.671153509682319e-05, "loss": 0.3384, "num_input_tokens_seen": 4200752, "step": 15655 }, { "epoch": 4.104849279161206, "grad_norm": 1.8351447582244873, "learning_rate": 3.67014292162362e-05, "loss": 0.4393, "num_input_tokens_seen": 4201888, "step": 15660 }, { "epoch": 4.1061598951507206, "grad_norm": 1.0766528844833374, "learning_rate": 3.66913208865603e-05, "loss": 0.2955, "num_input_tokens_seen": 4202928, "step": 15665 }, { "epoch": 4.107470511140236, "grad_norm": 0.9429026246070862, "learning_rate": 3.668121010991115e-05, "loss": 0.2973, "num_input_tokens_seen": 4204304, "step": 15670 }, { "epoch": 4.108781127129751, "grad_norm": 1.6416596174240112, "learning_rate": 3.66710968884049e-05, "loss": 0.3376, "num_input_tokens_seen": 4205888, "step": 15675 }, { "epoch": 4.110091743119266, "grad_norm": 0.7739332318305969, "learning_rate": 3.666098122415823e-05, "loss": 0.3229, "num_input_tokens_seen": 4207728, "step": 15680 }, { "epoch": 4.111402359108781, "grad_norm": 0.9910470247268677, "learning_rate": 3.665086311928834e-05, "loss": 0.3362, "num_input_tokens_seen": 4208848, "step": 15685 }, { "epoch": 4.1127129750982965, "grad_norm": 0.8342692852020264, "learning_rate": 3.664074257591293e-05, "loss": 0.2706, "num_input_tokens_seen": 4210144, "step": 15690 }, { "epoch": 4.114023591087811, "grad_norm": 1.092233419418335, "learning_rate": 3.6630619596150225e-05, "loss": 0.3589, "num_input_tokens_seen": 4211488, "step": 15695 }, { "epoch": 4.115334207077327, "grad_norm": 0.7290071845054626, "learning_rate": 3.662049418211892e-05, "loss": 0.3831, "num_input_tokens_seen": 4213184, "step": 15700 }, { "epoch": 4.116644823066841, "grad_norm": 6.214095592498779, "learning_rate": 3.6610366335938266e-05, "loss": 0.5436, "num_input_tokens_seen": 4214624, "step": 15705 }, { "epoch": 4.117955439056357, "grad_norm": 1.3182779550552368, "learning_rate": 3.6600236059728e-05, "loss": 0.3427, "num_input_tokens_seen": 4215536, "step": 15710 }, { "epoch": 4.1192660550458715, "grad_norm": 1.3761115074157715, "learning_rate": 3.6590103355608374e-05, "loss": 0.4309, "num_input_tokens_seen": 4216624, "step": 15715 }, { "epoch": 4.120576671035387, "grad_norm": 1.6875160932540894, "learning_rate": 3.657996822570013e-05, "loss": 0.4814, "num_input_tokens_seen": 4218048, "step": 15720 }, { "epoch": 4.121887287024902, "grad_norm": 0.678435742855072, "learning_rate": 3.656983067212456e-05, "loss": 0.2969, "num_input_tokens_seen": 4219552, "step": 15725 }, { "epoch": 4.123197903014416, "grad_norm": 1.3292630910873413, "learning_rate": 3.655969069700341e-05, "loss": 0.3081, "num_input_tokens_seen": 4220528, "step": 15730 }, { "epoch": 4.124508519003932, "grad_norm": 1.2618353366851807, "learning_rate": 3.654954830245898e-05, "loss": 0.3813, "num_input_tokens_seen": 4221552, "step": 15735 }, { "epoch": 4.1258191349934465, "grad_norm": 1.1498795747756958, "learning_rate": 3.6539403490614046e-05, "loss": 0.3413, "num_input_tokens_seen": 4222784, "step": 15740 }, { "epoch": 4.127129750982962, "grad_norm": 1.9231582880020142, "learning_rate": 3.6529256263591885e-05, "loss": 0.3259, "num_input_tokens_seen": 4223840, "step": 15745 }, { "epoch": 4.128440366972477, "grad_norm": 1.5483767986297607, "learning_rate": 3.651910662351632e-05, "loss": 0.5441, "num_input_tokens_seen": 4225728, "step": 15750 }, { "epoch": 4.129750982961992, "grad_norm": 0.9939459562301636, "learning_rate": 3.650895457251165e-05, "loss": 0.3045, "num_input_tokens_seen": 4227120, "step": 15755 }, { "epoch": 4.131061598951507, "grad_norm": 3.5622735023498535, "learning_rate": 3.6498800112702664e-05, "loss": 0.3836, "num_input_tokens_seen": 4228240, "step": 15760 }, { "epoch": 4.1323722149410225, "grad_norm": 1.8145718574523926, "learning_rate": 3.648864324621468e-05, "loss": 0.4037, "num_input_tokens_seen": 4229696, "step": 15765 }, { "epoch": 4.133682830930537, "grad_norm": 1.7264305353164673, "learning_rate": 3.647848397517353e-05, "loss": 0.3065, "num_input_tokens_seen": 4231024, "step": 15770 }, { "epoch": 4.134993446920053, "grad_norm": 1.3556705713272095, "learning_rate": 3.6468322301705504e-05, "loss": 0.3288, "num_input_tokens_seen": 4232800, "step": 15775 }, { "epoch": 4.136304062909567, "grad_norm": 1.6293288469314575, "learning_rate": 3.6458158227937435e-05, "loss": 0.3455, "num_input_tokens_seen": 4234192, "step": 15780 }, { "epoch": 4.137614678899083, "grad_norm": 1.1197826862335205, "learning_rate": 3.6447991755996654e-05, "loss": 0.3075, "num_input_tokens_seen": 4235296, "step": 15785 }, { "epoch": 4.1389252948885975, "grad_norm": 0.9535305500030518, "learning_rate": 3.643782288801098e-05, "loss": 0.5036, "num_input_tokens_seen": 4236752, "step": 15790 }, { "epoch": 4.140235910878113, "grad_norm": 1.12507963180542, "learning_rate": 3.642765162610873e-05, "loss": 0.1849, "num_input_tokens_seen": 4237856, "step": 15795 }, { "epoch": 4.141546526867628, "grad_norm": 2.390890121459961, "learning_rate": 3.6417477972418745e-05, "loss": 0.4214, "num_input_tokens_seen": 4239280, "step": 15800 }, { "epoch": 4.142857142857143, "grad_norm": 0.6109883785247803, "learning_rate": 3.640730192907035e-05, "loss": 0.3941, "num_input_tokens_seen": 4240912, "step": 15805 }, { "epoch": 4.144167758846658, "grad_norm": 1.4457385540008545, "learning_rate": 3.639712349819337e-05, "loss": 0.1803, "num_input_tokens_seen": 4241808, "step": 15810 }, { "epoch": 4.145478374836173, "grad_norm": 1.3595077991485596, "learning_rate": 3.6386942681918135e-05, "loss": 0.3502, "num_input_tokens_seen": 4242928, "step": 15815 }, { "epoch": 4.146788990825688, "grad_norm": 0.9665805101394653, "learning_rate": 3.637675948237547e-05, "loss": 0.2129, "num_input_tokens_seen": 4244048, "step": 15820 }, { "epoch": 4.148099606815203, "grad_norm": 1.2214876413345337, "learning_rate": 3.6366573901696704e-05, "loss": 0.1809, "num_input_tokens_seen": 4245104, "step": 15825 }, { "epoch": 4.149410222804718, "grad_norm": 2.343961477279663, "learning_rate": 3.635638594201366e-05, "loss": 0.3709, "num_input_tokens_seen": 4246192, "step": 15830 }, { "epoch": 4.150720838794233, "grad_norm": 1.2524042129516602, "learning_rate": 3.6346195605458664e-05, "loss": 0.3049, "num_input_tokens_seen": 4247312, "step": 15835 }, { "epoch": 4.1520314547837485, "grad_norm": 1.58927583694458, "learning_rate": 3.633600289416453e-05, "loss": 0.2187, "num_input_tokens_seen": 4248368, "step": 15840 }, { "epoch": 4.153342070773263, "grad_norm": 1.746366024017334, "learning_rate": 3.632580781026457e-05, "loss": 0.3579, "num_input_tokens_seen": 4249872, "step": 15845 }, { "epoch": 4.154652686762779, "grad_norm": 1.1776533126831055, "learning_rate": 3.631561035589262e-05, "loss": 0.3023, "num_input_tokens_seen": 4252336, "step": 15850 }, { "epoch": 4.155963302752293, "grad_norm": 0.7475170493125916, "learning_rate": 3.630541053318297e-05, "loss": 0.246, "num_input_tokens_seen": 4253920, "step": 15855 }, { "epoch": 4.157273918741809, "grad_norm": 1.2155375480651855, "learning_rate": 3.6295208344270415e-05, "loss": 0.4795, "num_input_tokens_seen": 4255312, "step": 15860 }, { "epoch": 4.1585845347313235, "grad_norm": 1.5461000204086304, "learning_rate": 3.628500379129028e-05, "loss": 0.3592, "num_input_tokens_seen": 4256400, "step": 15865 }, { "epoch": 4.159895150720839, "grad_norm": 0.9961380958557129, "learning_rate": 3.627479687637836e-05, "loss": 0.2707, "num_input_tokens_seen": 4258896, "step": 15870 }, { "epoch": 4.161205766710354, "grad_norm": 1.8197455406188965, "learning_rate": 3.6264587601670924e-05, "loss": 0.3718, "num_input_tokens_seen": 4260192, "step": 15875 }, { "epoch": 4.162516382699869, "grad_norm": 1.1484287977218628, "learning_rate": 3.625437596930476e-05, "loss": 0.3821, "num_input_tokens_seen": 4261280, "step": 15880 }, { "epoch": 4.163826998689384, "grad_norm": 0.9106109142303467, "learning_rate": 3.6244161981417165e-05, "loss": 0.265, "num_input_tokens_seen": 4262192, "step": 15885 }, { "epoch": 4.165137614678899, "grad_norm": 1.74653160572052, "learning_rate": 3.62339456401459e-05, "loss": 0.31, "num_input_tokens_seen": 4263392, "step": 15890 }, { "epoch": 4.166448230668414, "grad_norm": 13.547574996948242, "learning_rate": 3.622372694762921e-05, "loss": 0.6072, "num_input_tokens_seen": 4264896, "step": 15895 }, { "epoch": 4.16775884665793, "grad_norm": 2.075763702392578, "learning_rate": 3.621350590600587e-05, "loss": 0.3897, "num_input_tokens_seen": 4265984, "step": 15900 }, { "epoch": 4.169069462647444, "grad_norm": 1.3882046937942505, "learning_rate": 3.6203282517415114e-05, "loss": 0.3443, "num_input_tokens_seen": 4266864, "step": 15905 }, { "epoch": 4.17038007863696, "grad_norm": 0.9018324017524719, "learning_rate": 3.619305678399669e-05, "loss": 0.3708, "num_input_tokens_seen": 4268208, "step": 15910 }, { "epoch": 4.1716906946264745, "grad_norm": 1.149835228919983, "learning_rate": 3.6182828707890816e-05, "loss": 0.4141, "num_input_tokens_seen": 4269392, "step": 15915 }, { "epoch": 4.173001310615989, "grad_norm": 1.3747724294662476, "learning_rate": 3.617259829123822e-05, "loss": 0.4881, "num_input_tokens_seen": 4270800, "step": 15920 }, { "epoch": 4.174311926605505, "grad_norm": 4.301419734954834, "learning_rate": 3.6162365536180105e-05, "loss": 0.4194, "num_input_tokens_seen": 4271840, "step": 15925 }, { "epoch": 4.175622542595019, "grad_norm": 0.8534736633300781, "learning_rate": 3.615213044485817e-05, "loss": 0.265, "num_input_tokens_seen": 4273056, "step": 15930 }, { "epoch": 4.176933158584535, "grad_norm": 1.7066736221313477, "learning_rate": 3.6141893019414606e-05, "loss": 0.2772, "num_input_tokens_seen": 4274112, "step": 15935 }, { "epoch": 4.1782437745740495, "grad_norm": 1.969143271446228, "learning_rate": 3.6131653261992085e-05, "loss": 0.4007, "num_input_tokens_seen": 4275344, "step": 15940 }, { "epoch": 4.179554390563565, "grad_norm": 1.5072801113128662, "learning_rate": 3.612141117473377e-05, "loss": 0.3625, "num_input_tokens_seen": 4278688, "step": 15945 }, { "epoch": 4.18086500655308, "grad_norm": 2.970231294631958, "learning_rate": 3.61111667597833e-05, "loss": 0.4097, "num_input_tokens_seen": 4279920, "step": 15950 }, { "epoch": 4.182175622542595, "grad_norm": 1.9330261945724487, "learning_rate": 3.610092001928484e-05, "loss": 0.4554, "num_input_tokens_seen": 4280960, "step": 15955 }, { "epoch": 4.18348623853211, "grad_norm": 0.914412260055542, "learning_rate": 3.6090670955383e-05, "loss": 0.3449, "num_input_tokens_seen": 4282416, "step": 15960 }, { "epoch": 4.184796854521625, "grad_norm": 1.9971624612808228, "learning_rate": 3.6080419570222885e-05, "loss": 0.3635, "num_input_tokens_seen": 4283584, "step": 15965 }, { "epoch": 4.18610747051114, "grad_norm": 2.0902841091156006, "learning_rate": 3.60701658659501e-05, "loss": 0.3383, "num_input_tokens_seen": 4284864, "step": 15970 }, { "epoch": 4.187418086500656, "grad_norm": 1.0101394653320312, "learning_rate": 3.605990984471073e-05, "loss": 0.4034, "num_input_tokens_seen": 4286320, "step": 15975 }, { "epoch": 4.18872870249017, "grad_norm": 2.2743349075317383, "learning_rate": 3.6049651508651336e-05, "loss": 0.2098, "num_input_tokens_seen": 4287200, "step": 15980 }, { "epoch": 4.190039318479686, "grad_norm": 1.575887680053711, "learning_rate": 3.603939085991898e-05, "loss": 0.3518, "num_input_tokens_seen": 4288368, "step": 15985 }, { "epoch": 4.1913499344692005, "grad_norm": 10.255946159362793, "learning_rate": 3.6029127900661196e-05, "loss": 0.2729, "num_input_tokens_seen": 4289344, "step": 15990 }, { "epoch": 4.192660550458716, "grad_norm": 1.2610193490982056, "learning_rate": 3.601886263302599e-05, "loss": 0.3895, "num_input_tokens_seen": 4291568, "step": 15995 }, { "epoch": 4.193971166448231, "grad_norm": 5.306450366973877, "learning_rate": 3.600859505916187e-05, "loss": 0.287, "num_input_tokens_seen": 4292688, "step": 16000 }, { "epoch": 4.195281782437746, "grad_norm": 1.8762067556381226, "learning_rate": 3.599832518121784e-05, "loss": 0.3689, "num_input_tokens_seen": 4294352, "step": 16005 }, { "epoch": 4.196592398427261, "grad_norm": 4.361361980438232, "learning_rate": 3.598805300134334e-05, "loss": 0.3909, "num_input_tokens_seen": 4295520, "step": 16010 }, { "epoch": 4.1979030144167755, "grad_norm": 3.3090779781341553, "learning_rate": 3.597777852168834e-05, "loss": 0.4455, "num_input_tokens_seen": 4296528, "step": 16015 }, { "epoch": 4.199213630406291, "grad_norm": 1.4984967708587646, "learning_rate": 3.596750174440326e-05, "loss": 0.5652, "num_input_tokens_seen": 4297824, "step": 16020 }, { "epoch": 4.200524246395806, "grad_norm": 2.067063570022583, "learning_rate": 3.595722267163901e-05, "loss": 0.2359, "num_input_tokens_seen": 4299056, "step": 16025 }, { "epoch": 4.201834862385321, "grad_norm": 1.564071774482727, "learning_rate": 3.5946941305547e-05, "loss": 0.4556, "num_input_tokens_seen": 4300400, "step": 16030 }, { "epoch": 4.203145478374836, "grad_norm": 1.073004961013794, "learning_rate": 3.593665764827908e-05, "loss": 0.3946, "num_input_tokens_seen": 4301552, "step": 16035 }, { "epoch": 4.204456094364351, "grad_norm": 0.9177672266960144, "learning_rate": 3.592637170198762e-05, "loss": 0.3203, "num_input_tokens_seen": 4302736, "step": 16040 }, { "epoch": 4.205766710353866, "grad_norm": 2.708549737930298, "learning_rate": 3.591608346882543e-05, "loss": 0.5164, "num_input_tokens_seen": 4304048, "step": 16045 }, { "epoch": 4.207077326343382, "grad_norm": 1.2947360277175903, "learning_rate": 3.590579295094584e-05, "loss": 0.5208, "num_input_tokens_seen": 4305968, "step": 16050 }, { "epoch": 4.208387942332896, "grad_norm": 1.0046409368515015, "learning_rate": 3.589550015050263e-05, "loss": 0.5046, "num_input_tokens_seen": 4307216, "step": 16055 }, { "epoch": 4.209698558322412, "grad_norm": 1.0387824773788452, "learning_rate": 3.5885205069650056e-05, "loss": 0.2353, "num_input_tokens_seen": 4308336, "step": 16060 }, { "epoch": 4.2110091743119265, "grad_norm": 1.1554970741271973, "learning_rate": 3.587490771054288e-05, "loss": 0.348, "num_input_tokens_seen": 4310016, "step": 16065 }, { "epoch": 4.212319790301442, "grad_norm": 1.2760416269302368, "learning_rate": 3.586460807533631e-05, "loss": 0.8231, "num_input_tokens_seen": 4311136, "step": 16070 }, { "epoch": 4.213630406290957, "grad_norm": 2.262627124786377, "learning_rate": 3.585430616618604e-05, "loss": 0.3043, "num_input_tokens_seen": 4312384, "step": 16075 }, { "epoch": 4.214941022280472, "grad_norm": 1.1456384658813477, "learning_rate": 3.584400198524824e-05, "loss": 0.3, "num_input_tokens_seen": 4314544, "step": 16080 }, { "epoch": 4.216251638269987, "grad_norm": 1.0621742010116577, "learning_rate": 3.583369553467956e-05, "loss": 0.2841, "num_input_tokens_seen": 4316176, "step": 16085 }, { "epoch": 4.217562254259502, "grad_norm": 0.6280931830406189, "learning_rate": 3.582338681663713e-05, "loss": 0.2525, "num_input_tokens_seen": 4317312, "step": 16090 }, { "epoch": 4.218872870249017, "grad_norm": 1.1733187437057495, "learning_rate": 3.581307583327854e-05, "loss": 0.5015, "num_input_tokens_seen": 4318768, "step": 16095 }, { "epoch": 4.220183486238533, "grad_norm": 0.9018048048019409, "learning_rate": 3.580276258676186e-05, "loss": 0.4547, "num_input_tokens_seen": 4319968, "step": 16100 }, { "epoch": 4.221494102228047, "grad_norm": 0.9558042287826538, "learning_rate": 3.579244707924564e-05, "loss": 0.4162, "num_input_tokens_seen": 4321072, "step": 16105 }, { "epoch": 4.222804718217562, "grad_norm": 0.9775887727737427, "learning_rate": 3.5782129312888885e-05, "loss": 0.2823, "num_input_tokens_seen": 4322192, "step": 16110 }, { "epoch": 4.224115334207077, "grad_norm": 2.886800765991211, "learning_rate": 3.5771809289851097e-05, "loss": 0.4208, "num_input_tokens_seen": 4323376, "step": 16115 }, { "epoch": 4.225425950196592, "grad_norm": 0.6653011441230774, "learning_rate": 3.576148701229224e-05, "loss": 0.3812, "num_input_tokens_seen": 4324656, "step": 16120 }, { "epoch": 4.226736566186108, "grad_norm": 1.396384358406067, "learning_rate": 3.575116248237275e-05, "loss": 0.3543, "num_input_tokens_seen": 4326208, "step": 16125 }, { "epoch": 4.228047182175622, "grad_norm": 1.1437876224517822, "learning_rate": 3.574083570225352e-05, "loss": 0.3394, "num_input_tokens_seen": 4327392, "step": 16130 }, { "epoch": 4.229357798165138, "grad_norm": 1.3182317018508911, "learning_rate": 3.573050667409594e-05, "loss": 0.337, "num_input_tokens_seen": 4329024, "step": 16135 }, { "epoch": 4.2306684141546524, "grad_norm": 1.4825668334960938, "learning_rate": 3.572017540006186e-05, "loss": 0.3961, "num_input_tokens_seen": 4330352, "step": 16140 }, { "epoch": 4.231979030144168, "grad_norm": 1.2205891609191895, "learning_rate": 3.570984188231359e-05, "loss": 0.2234, "num_input_tokens_seen": 4331488, "step": 16145 }, { "epoch": 4.233289646133683, "grad_norm": 0.6287804245948792, "learning_rate": 3.5699506123013906e-05, "loss": 0.3995, "num_input_tokens_seen": 4332608, "step": 16150 }, { "epoch": 4.234600262123198, "grad_norm": 1.5979969501495361, "learning_rate": 3.568916812432609e-05, "loss": 0.2918, "num_input_tokens_seen": 4334208, "step": 16155 }, { "epoch": 4.235910878112713, "grad_norm": 1.0276682376861572, "learning_rate": 3.5678827888413856e-05, "loss": 0.312, "num_input_tokens_seen": 4335424, "step": 16160 }, { "epoch": 4.237221494102228, "grad_norm": 6.303406238555908, "learning_rate": 3.56684854174414e-05, "loss": 0.3136, "num_input_tokens_seen": 4336512, "step": 16165 }, { "epoch": 4.238532110091743, "grad_norm": 0.2096768468618393, "learning_rate": 3.5658140713573365e-05, "loss": 0.1615, "num_input_tokens_seen": 4337920, "step": 16170 }, { "epoch": 4.239842726081259, "grad_norm": 1.1213366985321045, "learning_rate": 3.56477937789749e-05, "loss": 0.4802, "num_input_tokens_seen": 4339264, "step": 16175 }, { "epoch": 4.241153342070773, "grad_norm": 0.6339240670204163, "learning_rate": 3.56374446158116e-05, "loss": 0.2379, "num_input_tokens_seen": 4340560, "step": 16180 }, { "epoch": 4.242463958060289, "grad_norm": 0.8244946002960205, "learning_rate": 3.5627093226249524e-05, "loss": 0.2298, "num_input_tokens_seen": 4342000, "step": 16185 }, { "epoch": 4.243774574049803, "grad_norm": 1.2266769409179688, "learning_rate": 3.5616739612455196e-05, "loss": 0.4415, "num_input_tokens_seen": 4343568, "step": 16190 }, { "epoch": 4.245085190039318, "grad_norm": 0.5980857610702515, "learning_rate": 3.560638377659561e-05, "loss": 0.3038, "num_input_tokens_seen": 4345136, "step": 16195 }, { "epoch": 4.246395806028834, "grad_norm": 1.0081841945648193, "learning_rate": 3.559602572083822e-05, "loss": 0.2971, "num_input_tokens_seen": 4346240, "step": 16200 }, { "epoch": 4.247706422018348, "grad_norm": 0.8107646107673645, "learning_rate": 3.558566544735096e-05, "loss": 0.358, "num_input_tokens_seen": 4347728, "step": 16205 }, { "epoch": 4.249017038007864, "grad_norm": 1.0586708784103394, "learning_rate": 3.5575302958302214e-05, "loss": 0.4366, "num_input_tokens_seen": 4349168, "step": 16210 }, { "epoch": 4.250327653997378, "grad_norm": 1.7920902967453003, "learning_rate": 3.556493825586083e-05, "loss": 0.2486, "num_input_tokens_seen": 4350432, "step": 16215 }, { "epoch": 4.251638269986894, "grad_norm": 2.054274797439575, "learning_rate": 3.555457134219613e-05, "loss": 0.4853, "num_input_tokens_seen": 4351600, "step": 16220 }, { "epoch": 4.252948885976409, "grad_norm": 1.2635972499847412, "learning_rate": 3.554420221947788e-05, "loss": 0.6594, "num_input_tokens_seen": 4352784, "step": 16225 }, { "epoch": 4.254259501965924, "grad_norm": 2.9394259452819824, "learning_rate": 3.553383088987632e-05, "loss": 0.5482, "num_input_tokens_seen": 4353952, "step": 16230 }, { "epoch": 4.255570117955439, "grad_norm": 1.994411826133728, "learning_rate": 3.552345735556216e-05, "loss": 0.3256, "num_input_tokens_seen": 4355328, "step": 16235 }, { "epoch": 4.256880733944954, "grad_norm": 1.3151346445083618, "learning_rate": 3.5513081618706555e-05, "loss": 0.3568, "num_input_tokens_seen": 4356672, "step": 16240 }, { "epoch": 4.258191349934469, "grad_norm": 1.317007064819336, "learning_rate": 3.550270368148113e-05, "loss": 0.3799, "num_input_tokens_seen": 4357840, "step": 16245 }, { "epoch": 4.259501965923985, "grad_norm": 0.991162896156311, "learning_rate": 3.549232354605798e-05, "loss": 0.3672, "num_input_tokens_seen": 4358816, "step": 16250 }, { "epoch": 4.260812581913499, "grad_norm": 1.8905587196350098, "learning_rate": 3.548194121460964e-05, "loss": 0.3656, "num_input_tokens_seen": 4359920, "step": 16255 }, { "epoch": 4.262123197903015, "grad_norm": 3.317809820175171, "learning_rate": 3.547155668930912e-05, "loss": 0.4317, "num_input_tokens_seen": 4360944, "step": 16260 }, { "epoch": 4.263433813892529, "grad_norm": 1.051470398902893, "learning_rate": 3.546116997232987e-05, "loss": 0.2992, "num_input_tokens_seen": 4362112, "step": 16265 }, { "epoch": 4.264744429882045, "grad_norm": 0.8268030285835266, "learning_rate": 3.545078106584582e-05, "loss": 0.3248, "num_input_tokens_seen": 4363232, "step": 16270 }, { "epoch": 4.26605504587156, "grad_norm": 0.6040062308311462, "learning_rate": 3.544038997203135e-05, "loss": 0.2344, "num_input_tokens_seen": 4364800, "step": 16275 }, { "epoch": 4.267365661861074, "grad_norm": 1.763629674911499, "learning_rate": 3.5429996693061304e-05, "loss": 0.4793, "num_input_tokens_seen": 4366096, "step": 16280 }, { "epoch": 4.26867627785059, "grad_norm": 1.150772213935852, "learning_rate": 3.541960123111097e-05, "loss": 0.2455, "num_input_tokens_seen": 4367744, "step": 16285 }, { "epoch": 4.269986893840104, "grad_norm": 1.7110486030578613, "learning_rate": 3.54092035883561e-05, "loss": 0.3275, "num_input_tokens_seen": 4369040, "step": 16290 }, { "epoch": 4.27129750982962, "grad_norm": 3.0629565715789795, "learning_rate": 3.53988037669729e-05, "loss": 0.5025, "num_input_tokens_seen": 4370784, "step": 16295 }, { "epoch": 4.272608125819135, "grad_norm": 0.895214319229126, "learning_rate": 3.5388401769138053e-05, "loss": 0.2916, "num_input_tokens_seen": 4372880, "step": 16300 }, { "epoch": 4.27391874180865, "grad_norm": 2.4177229404449463, "learning_rate": 3.5377997597028656e-05, "loss": 0.2725, "num_input_tokens_seen": 4373840, "step": 16305 }, { "epoch": 4.275229357798165, "grad_norm": 1.0316458940505981, "learning_rate": 3.53675912528223e-05, "loss": 0.342, "num_input_tokens_seen": 4375536, "step": 16310 }, { "epoch": 4.27653997378768, "grad_norm": 1.2866771221160889, "learning_rate": 3.535718273869699e-05, "loss": 0.2755, "num_input_tokens_seen": 4376464, "step": 16315 }, { "epoch": 4.277850589777195, "grad_norm": 1.59391188621521, "learning_rate": 3.534677205683125e-05, "loss": 0.3561, "num_input_tokens_seen": 4377776, "step": 16320 }, { "epoch": 4.2791612057667106, "grad_norm": 1.022371768951416, "learning_rate": 3.533635920940399e-05, "loss": 0.3833, "num_input_tokens_seen": 4379120, "step": 16325 }, { "epoch": 4.280471821756225, "grad_norm": 0.8015239238739014, "learning_rate": 3.5325944198594607e-05, "loss": 0.3234, "num_input_tokens_seen": 4380352, "step": 16330 }, { "epoch": 4.281782437745741, "grad_norm": 0.809281587600708, "learning_rate": 3.531552702658294e-05, "loss": 0.2581, "num_input_tokens_seen": 4381488, "step": 16335 }, { "epoch": 4.283093053735255, "grad_norm": 1.4925849437713623, "learning_rate": 3.53051076955493e-05, "loss": 0.3053, "num_input_tokens_seen": 4382672, "step": 16340 }, { "epoch": 4.284403669724771, "grad_norm": 2.578418493270874, "learning_rate": 3.5294686207674414e-05, "loss": 0.4985, "num_input_tokens_seen": 4384000, "step": 16345 }, { "epoch": 4.285714285714286, "grad_norm": 1.068687081336975, "learning_rate": 3.5284262565139494e-05, "loss": 0.3343, "num_input_tokens_seen": 4385168, "step": 16350 }, { "epoch": 4.287024901703801, "grad_norm": 3.333160400390625, "learning_rate": 3.5273836770126204e-05, "loss": 0.3793, "num_input_tokens_seen": 4386400, "step": 16355 }, { "epoch": 4.288335517693316, "grad_norm": 0.5527365803718567, "learning_rate": 3.5263408824816615e-05, "loss": 0.3258, "num_input_tokens_seen": 4387840, "step": 16360 }, { "epoch": 4.289646133682831, "grad_norm": 0.6313728094100952, "learning_rate": 3.52529787313933e-05, "loss": 0.4548, "num_input_tokens_seen": 4389440, "step": 16365 }, { "epoch": 4.290956749672346, "grad_norm": 1.5900036096572876, "learning_rate": 3.524254649203925e-05, "loss": 0.2701, "num_input_tokens_seen": 4390848, "step": 16370 }, { "epoch": 4.292267365661861, "grad_norm": 0.9884350299835205, "learning_rate": 3.5232112108937914e-05, "loss": 0.3911, "num_input_tokens_seen": 4392208, "step": 16375 }, { "epoch": 4.293577981651376, "grad_norm": 1.0434958934783936, "learning_rate": 3.52216755842732e-05, "loss": 0.5693, "num_input_tokens_seen": 4393968, "step": 16380 }, { "epoch": 4.294888597640891, "grad_norm": 1.639180064201355, "learning_rate": 3.521123692022944e-05, "loss": 0.4382, "num_input_tokens_seen": 4394944, "step": 16385 }, { "epoch": 4.296199213630406, "grad_norm": 1.252429485321045, "learning_rate": 3.520079611899144e-05, "loss": 0.3224, "num_input_tokens_seen": 4396112, "step": 16390 }, { "epoch": 4.297509829619921, "grad_norm": 0.6594598293304443, "learning_rate": 3.519035318274444e-05, "loss": 0.3178, "num_input_tokens_seen": 4397376, "step": 16395 }, { "epoch": 4.2988204456094365, "grad_norm": 1.5361226797103882, "learning_rate": 3.517990811367412e-05, "loss": 0.4448, "num_input_tokens_seen": 4398624, "step": 16400 }, { "epoch": 4.300131061598951, "grad_norm": 2.520146608352661, "learning_rate": 3.516946091396662e-05, "loss": 0.3936, "num_input_tokens_seen": 4399744, "step": 16405 }, { "epoch": 4.301441677588467, "grad_norm": 3.5149660110473633, "learning_rate": 3.5159011585808525e-05, "loss": 0.5309, "num_input_tokens_seen": 4400720, "step": 16410 }, { "epoch": 4.302752293577981, "grad_norm": 1.225192666053772, "learning_rate": 3.5148560131386867e-05, "loss": 0.449, "num_input_tokens_seen": 4402464, "step": 16415 }, { "epoch": 4.304062909567497, "grad_norm": 0.7681854963302612, "learning_rate": 3.5138106552889106e-05, "loss": 0.3347, "num_input_tokens_seen": 4403968, "step": 16420 }, { "epoch": 4.305373525557012, "grad_norm": 1.3416621685028076, "learning_rate": 3.5127650852503156e-05, "loss": 0.2256, "num_input_tokens_seen": 4405200, "step": 16425 }, { "epoch": 4.306684141546527, "grad_norm": 2.6530556678771973, "learning_rate": 3.511719303241738e-05, "loss": 0.2018, "num_input_tokens_seen": 4406432, "step": 16430 }, { "epoch": 4.307994757536042, "grad_norm": 3.2312471866607666, "learning_rate": 3.5106733094820586e-05, "loss": 0.4583, "num_input_tokens_seen": 4407648, "step": 16435 }, { "epoch": 4.309305373525557, "grad_norm": 2.6420023441314697, "learning_rate": 3.509627104190202e-05, "loss": 0.3414, "num_input_tokens_seen": 4408736, "step": 16440 }, { "epoch": 4.310615989515072, "grad_norm": 3.8113491535186768, "learning_rate": 3.5085806875851365e-05, "loss": 0.3714, "num_input_tokens_seen": 4409984, "step": 16445 }, { "epoch": 4.3119266055045875, "grad_norm": 2.2139394283294678, "learning_rate": 3.507534059885876e-05, "loss": 0.3142, "num_input_tokens_seen": 4410960, "step": 16450 }, { "epoch": 4.313237221494102, "grad_norm": 1.902824878692627, "learning_rate": 3.5064872213114784e-05, "loss": 0.4213, "num_input_tokens_seen": 4412016, "step": 16455 }, { "epoch": 4.314547837483618, "grad_norm": 0.8734674453735352, "learning_rate": 3.505440172081044e-05, "loss": 0.3321, "num_input_tokens_seen": 4413776, "step": 16460 }, { "epoch": 4.315858453473132, "grad_norm": 1.607588768005371, "learning_rate": 3.504392912413719e-05, "loss": 0.2576, "num_input_tokens_seen": 4414720, "step": 16465 }, { "epoch": 4.317169069462647, "grad_norm": 1.1587462425231934, "learning_rate": 3.503345442528692e-05, "loss": 0.3217, "num_input_tokens_seen": 4415824, "step": 16470 }, { "epoch": 4.3184796854521625, "grad_norm": 1.415827751159668, "learning_rate": 3.5022977626451985e-05, "loss": 0.2062, "num_input_tokens_seen": 4416832, "step": 16475 }, { "epoch": 4.319790301441677, "grad_norm": 1.7727583646774292, "learning_rate": 3.501249872982515e-05, "loss": 0.4124, "num_input_tokens_seen": 4418176, "step": 16480 }, { "epoch": 4.321100917431193, "grad_norm": 1.199699878692627, "learning_rate": 3.500201773759963e-05, "loss": 0.3792, "num_input_tokens_seen": 4419504, "step": 16485 }, { "epoch": 4.322411533420707, "grad_norm": 1.7805516719818115, "learning_rate": 3.499153465196907e-05, "loss": 0.6166, "num_input_tokens_seen": 4420784, "step": 16490 }, { "epoch": 4.323722149410223, "grad_norm": 2.110133409500122, "learning_rate": 3.4981049475127584e-05, "loss": 0.4666, "num_input_tokens_seen": 4421968, "step": 16495 }, { "epoch": 4.325032765399738, "grad_norm": 0.9011319279670715, "learning_rate": 3.4970562209269684e-05, "loss": 0.2533, "num_input_tokens_seen": 4423088, "step": 16500 }, { "epoch": 4.326343381389253, "grad_norm": 1.2642605304718018, "learning_rate": 3.496007285659034e-05, "loss": 0.5291, "num_input_tokens_seen": 4424672, "step": 16505 }, { "epoch": 4.327653997378768, "grad_norm": 1.0062477588653564, "learning_rate": 3.494958141928496e-05, "loss": 0.2335, "num_input_tokens_seen": 4425632, "step": 16510 }, { "epoch": 4.328964613368283, "grad_norm": 0.9766978025436401, "learning_rate": 3.4939087899549386e-05, "loss": 0.3556, "num_input_tokens_seen": 4426960, "step": 16515 }, { "epoch": 4.330275229357798, "grad_norm": 1.6188976764678955, "learning_rate": 3.4928592299579874e-05, "loss": 0.3204, "num_input_tokens_seen": 4428816, "step": 16520 }, { "epoch": 4.3315858453473135, "grad_norm": 1.3712010383605957, "learning_rate": 3.4918094621573164e-05, "loss": 0.3584, "num_input_tokens_seen": 4429888, "step": 16525 }, { "epoch": 4.332896461336828, "grad_norm": 0.6879979968070984, "learning_rate": 3.490759486772637e-05, "loss": 0.369, "num_input_tokens_seen": 4431424, "step": 16530 }, { "epoch": 4.334207077326344, "grad_norm": 0.7778570055961609, "learning_rate": 3.48970930402371e-05, "loss": 0.2689, "num_input_tokens_seen": 4432880, "step": 16535 }, { "epoch": 4.335517693315858, "grad_norm": 0.91136234998703, "learning_rate": 3.488658914130336e-05, "loss": 0.2782, "num_input_tokens_seen": 4433920, "step": 16540 }, { "epoch": 4.336828309305374, "grad_norm": 1.3941502571105957, "learning_rate": 3.487608317312358e-05, "loss": 0.3457, "num_input_tokens_seen": 4435104, "step": 16545 }, { "epoch": 4.3381389252948885, "grad_norm": 1.9551451206207275, "learning_rate": 3.486557513789668e-05, "loss": 0.3596, "num_input_tokens_seen": 4436416, "step": 16550 }, { "epoch": 4.339449541284404, "grad_norm": 2.354416847229004, "learning_rate": 3.485506503782194e-05, "loss": 0.3976, "num_input_tokens_seen": 4437600, "step": 16555 }, { "epoch": 4.340760157273919, "grad_norm": 0.5950000286102295, "learning_rate": 3.484455287509911e-05, "loss": 0.327, "num_input_tokens_seen": 4438928, "step": 16560 }, { "epoch": 4.342070773263433, "grad_norm": 0.8557286858558655, "learning_rate": 3.483403865192838e-05, "loss": 0.4345, "num_input_tokens_seen": 4440400, "step": 16565 }, { "epoch": 4.343381389252949, "grad_norm": 1.566717267036438, "learning_rate": 3.482352237051035e-05, "loss": 0.4212, "num_input_tokens_seen": 4441728, "step": 16570 }, { "epoch": 4.344692005242464, "grad_norm": 2.0889768600463867, "learning_rate": 3.481300403304606e-05, "loss": 0.3577, "num_input_tokens_seen": 4442768, "step": 16575 }, { "epoch": 4.346002621231979, "grad_norm": 1.0581681728363037, "learning_rate": 3.480248364173698e-05, "loss": 0.3112, "num_input_tokens_seen": 4444448, "step": 16580 }, { "epoch": 4.347313237221494, "grad_norm": 1.0606915950775146, "learning_rate": 3.4791961198785014e-05, "loss": 0.2649, "num_input_tokens_seen": 4445568, "step": 16585 }, { "epoch": 4.348623853211009, "grad_norm": 2.3250155448913574, "learning_rate": 3.478143670639248e-05, "loss": 0.3948, "num_input_tokens_seen": 4446896, "step": 16590 }, { "epoch": 4.349934469200524, "grad_norm": 1.1872657537460327, "learning_rate": 3.477091016676216e-05, "loss": 0.3318, "num_input_tokens_seen": 4448192, "step": 16595 }, { "epoch": 4.3512450851900395, "grad_norm": 1.1743543148040771, "learning_rate": 3.476038158209722e-05, "loss": 0.4071, "num_input_tokens_seen": 4449296, "step": 16600 }, { "epoch": 4.352555701179554, "grad_norm": 1.9296534061431885, "learning_rate": 3.4749850954601274e-05, "loss": 0.3168, "num_input_tokens_seen": 4451136, "step": 16605 }, { "epoch": 4.35386631716907, "grad_norm": 0.6387118101119995, "learning_rate": 3.473931828647837e-05, "loss": 0.2533, "num_input_tokens_seen": 4452480, "step": 16610 }, { "epoch": 4.355176933158584, "grad_norm": 2.7919907569885254, "learning_rate": 3.472878357993298e-05, "loss": 0.3694, "num_input_tokens_seen": 4453536, "step": 16615 }, { "epoch": 4.3564875491481, "grad_norm": 1.1446566581726074, "learning_rate": 3.4718246837169996e-05, "loss": 0.539, "num_input_tokens_seen": 4455712, "step": 16620 }, { "epoch": 4.3577981651376145, "grad_norm": 1.03729248046875, "learning_rate": 3.470770806039474e-05, "loss": 0.3759, "num_input_tokens_seen": 4457056, "step": 16625 }, { "epoch": 4.35910878112713, "grad_norm": 1.1601433753967285, "learning_rate": 3.469716725181296e-05, "loss": 0.3228, "num_input_tokens_seen": 4458352, "step": 16630 }, { "epoch": 4.360419397116645, "grad_norm": 2.867678165435791, "learning_rate": 3.468662441363083e-05, "loss": 0.323, "num_input_tokens_seen": 4459456, "step": 16635 }, { "epoch": 4.36173001310616, "grad_norm": 1.0764873027801514, "learning_rate": 3.467607954805495e-05, "loss": 0.5093, "num_input_tokens_seen": 4460704, "step": 16640 }, { "epoch": 4.363040629095675, "grad_norm": 0.9911305904388428, "learning_rate": 3.466553265729233e-05, "loss": 0.2859, "num_input_tokens_seen": 4462192, "step": 16645 }, { "epoch": 4.3643512450851905, "grad_norm": 2.039288282394409, "learning_rate": 3.465498374355044e-05, "loss": 0.3979, "num_input_tokens_seen": 4463328, "step": 16650 }, { "epoch": 4.365661861074705, "grad_norm": 1.4290627241134644, "learning_rate": 3.4644432809037134e-05, "loss": 0.2669, "num_input_tokens_seen": 4464384, "step": 16655 }, { "epoch": 4.36697247706422, "grad_norm": 2.5127010345458984, "learning_rate": 3.463387985596069e-05, "loss": 0.4103, "num_input_tokens_seen": 4465808, "step": 16660 }, { "epoch": 4.368283093053735, "grad_norm": 1.2354819774627686, "learning_rate": 3.462332488652985e-05, "loss": 0.4157, "num_input_tokens_seen": 4467200, "step": 16665 }, { "epoch": 4.36959370904325, "grad_norm": 1.197962999343872, "learning_rate": 3.461276790295374e-05, "loss": 0.3094, "num_input_tokens_seen": 4468608, "step": 16670 }, { "epoch": 4.3709043250327655, "grad_norm": 1.0911953449249268, "learning_rate": 3.4602208907441915e-05, "loss": 0.2754, "num_input_tokens_seen": 4469920, "step": 16675 }, { "epoch": 4.37221494102228, "grad_norm": 1.2658835649490356, "learning_rate": 3.4591647902204355e-05, "loss": 0.4854, "num_input_tokens_seen": 4470896, "step": 16680 }, { "epoch": 4.373525557011796, "grad_norm": 1.761495590209961, "learning_rate": 3.4581084889451466e-05, "loss": 0.4462, "num_input_tokens_seen": 4472864, "step": 16685 }, { "epoch": 4.37483617300131, "grad_norm": 1.2119874954223633, "learning_rate": 3.457051987139406e-05, "loss": 0.367, "num_input_tokens_seen": 4474032, "step": 16690 }, { "epoch": 4.376146788990826, "grad_norm": 3.6500518321990967, "learning_rate": 3.455995285024338e-05, "loss": 0.4494, "num_input_tokens_seen": 4475728, "step": 16695 }, { "epoch": 4.3774574049803405, "grad_norm": 5.542580604553223, "learning_rate": 3.454938382821109e-05, "loss": 0.5579, "num_input_tokens_seen": 4477344, "step": 16700 }, { "epoch": 4.378768020969856, "grad_norm": 1.1905821561813354, "learning_rate": 3.453881280750927e-05, "loss": 0.3403, "num_input_tokens_seen": 4478480, "step": 16705 }, { "epoch": 4.380078636959371, "grad_norm": 1.3696770668029785, "learning_rate": 3.452823979035041e-05, "loss": 0.3075, "num_input_tokens_seen": 4479552, "step": 16710 }, { "epoch": 4.381389252948886, "grad_norm": 1.6426670551300049, "learning_rate": 3.451766477894742e-05, "loss": 0.406, "num_input_tokens_seen": 4480704, "step": 16715 }, { "epoch": 4.382699868938401, "grad_norm": 2.4242665767669678, "learning_rate": 3.450708777551364e-05, "loss": 0.3936, "num_input_tokens_seen": 4482176, "step": 16720 }, { "epoch": 4.3840104849279165, "grad_norm": 0.6132290363311768, "learning_rate": 3.4496508782262804e-05, "loss": 0.3009, "num_input_tokens_seen": 4483376, "step": 16725 }, { "epoch": 4.385321100917431, "grad_norm": 2.7995011806488037, "learning_rate": 3.4485927801409105e-05, "loss": 0.34, "num_input_tokens_seen": 4484656, "step": 16730 }, { "epoch": 4.386631716906947, "grad_norm": 0.8609116077423096, "learning_rate": 3.44753448351671e-05, "loss": 0.3136, "num_input_tokens_seen": 4485968, "step": 16735 }, { "epoch": 4.387942332896461, "grad_norm": 3.3436646461486816, "learning_rate": 3.4464759885751796e-05, "loss": 0.361, "num_input_tokens_seen": 4487888, "step": 16740 }, { "epoch": 4.389252948885977, "grad_norm": 0.8006501197814941, "learning_rate": 3.4454172955378595e-05, "loss": 0.3321, "num_input_tokens_seen": 4488944, "step": 16745 }, { "epoch": 4.3905635648754915, "grad_norm": 0.8130530714988708, "learning_rate": 3.444358404626335e-05, "loss": 0.319, "num_input_tokens_seen": 4490336, "step": 16750 }, { "epoch": 4.391874180865006, "grad_norm": 0.6712907552719116, "learning_rate": 3.4432993160622256e-05, "loss": 0.2834, "num_input_tokens_seen": 4491504, "step": 16755 }, { "epoch": 4.393184796854522, "grad_norm": 1.1884721517562866, "learning_rate": 3.4422400300672e-05, "loss": 0.4298, "num_input_tokens_seen": 4492976, "step": 16760 }, { "epoch": 4.394495412844036, "grad_norm": 1.7421600818634033, "learning_rate": 3.441180546862965e-05, "loss": 0.5182, "num_input_tokens_seen": 4493968, "step": 16765 }, { "epoch": 4.395806028833552, "grad_norm": 1.6899566650390625, "learning_rate": 3.440120866671268e-05, "loss": 0.5206, "num_input_tokens_seen": 4495168, "step": 16770 }, { "epoch": 4.3971166448230665, "grad_norm": 1.7369353771209717, "learning_rate": 3.4390609897138965e-05, "loss": 0.3397, "num_input_tokens_seen": 4496256, "step": 16775 }, { "epoch": 4.398427260812582, "grad_norm": 1.074692964553833, "learning_rate": 3.438000916212683e-05, "loss": 0.295, "num_input_tokens_seen": 4497216, "step": 16780 }, { "epoch": 4.399737876802097, "grad_norm": 1.793020248413086, "learning_rate": 3.436940646389499e-05, "loss": 0.2403, "num_input_tokens_seen": 4498448, "step": 16785 }, { "epoch": 4.401048492791612, "grad_norm": 1.421364665031433, "learning_rate": 3.435880180466257e-05, "loss": 0.3302, "num_input_tokens_seen": 4499776, "step": 16790 }, { "epoch": 4.402359108781127, "grad_norm": 1.4720966815948486, "learning_rate": 3.434819518664909e-05, "loss": 0.3782, "num_input_tokens_seen": 4501104, "step": 16795 }, { "epoch": 4.4036697247706424, "grad_norm": 1.1266885995864868, "learning_rate": 3.4337586612074515e-05, "loss": 0.3918, "num_input_tokens_seen": 4504208, "step": 16800 }, { "epoch": 4.404980340760157, "grad_norm": 1.1365020275115967, "learning_rate": 3.4326976083159194e-05, "loss": 0.3684, "num_input_tokens_seen": 4505920, "step": 16805 }, { "epoch": 4.406290956749673, "grad_norm": 1.6604540348052979, "learning_rate": 3.43163636021239e-05, "loss": 0.4053, "num_input_tokens_seen": 4507536, "step": 16810 }, { "epoch": 4.407601572739187, "grad_norm": 1.8749252557754517, "learning_rate": 3.43057491711898e-05, "loss": 0.2948, "num_input_tokens_seen": 4508768, "step": 16815 }, { "epoch": 4.408912188728703, "grad_norm": 0.6499437689781189, "learning_rate": 3.429513279257848e-05, "loss": 0.3321, "num_input_tokens_seen": 4510336, "step": 16820 }, { "epoch": 4.4102228047182175, "grad_norm": 1.4025542736053467, "learning_rate": 3.428451446851192e-05, "loss": 0.3788, "num_input_tokens_seen": 4511840, "step": 16825 }, { "epoch": 4.411533420707733, "grad_norm": 1.3784167766571045, "learning_rate": 3.427389420121253e-05, "loss": 0.3447, "num_input_tokens_seen": 4513008, "step": 16830 }, { "epoch": 4.412844036697248, "grad_norm": 1.5165982246398926, "learning_rate": 3.42632719929031e-05, "loss": 0.2561, "num_input_tokens_seen": 4514352, "step": 16835 }, { "epoch": 4.414154652686763, "grad_norm": 1.8574429750442505, "learning_rate": 3.425264784580686e-05, "loss": 0.3785, "num_input_tokens_seen": 4515792, "step": 16840 }, { "epoch": 4.415465268676278, "grad_norm": 1.3686634302139282, "learning_rate": 3.424202176214741e-05, "loss": 0.2529, "num_input_tokens_seen": 4516752, "step": 16845 }, { "epoch": 4.4167758846657925, "grad_norm": 1.0044394731521606, "learning_rate": 3.423139374414877e-05, "loss": 0.2947, "num_input_tokens_seen": 4517808, "step": 16850 }, { "epoch": 4.418086500655308, "grad_norm": 1.4048179388046265, "learning_rate": 3.422076379403537e-05, "loss": 0.4275, "num_input_tokens_seen": 4518944, "step": 16855 }, { "epoch": 4.419397116644823, "grad_norm": 3.9554102420806885, "learning_rate": 3.421013191403203e-05, "loss": 0.2809, "num_input_tokens_seen": 4520160, "step": 16860 }, { "epoch": 4.420707732634338, "grad_norm": 1.2668232917785645, "learning_rate": 3.419949810636401e-05, "loss": 0.31, "num_input_tokens_seen": 4521264, "step": 16865 }, { "epoch": 4.422018348623853, "grad_norm": 1.4060118198394775, "learning_rate": 3.418886237325693e-05, "loss": 0.404, "num_input_tokens_seen": 4522272, "step": 16870 }, { "epoch": 4.423328964613368, "grad_norm": 1.0694692134857178, "learning_rate": 3.417822471693683e-05, "loss": 0.3969, "num_input_tokens_seen": 4523344, "step": 16875 }, { "epoch": 4.424639580602883, "grad_norm": 3.5909481048583984, "learning_rate": 3.4167585139630156e-05, "loss": 0.3626, "num_input_tokens_seen": 4524944, "step": 16880 }, { "epoch": 4.425950196592399, "grad_norm": 0.8935874104499817, "learning_rate": 3.415694364356375e-05, "loss": 0.3864, "num_input_tokens_seen": 4526112, "step": 16885 }, { "epoch": 4.427260812581913, "grad_norm": 1.4901190996170044, "learning_rate": 3.414630023096486e-05, "loss": 0.3721, "num_input_tokens_seen": 4527920, "step": 16890 }, { "epoch": 4.428571428571429, "grad_norm": 1.0702438354492188, "learning_rate": 3.413565490406114e-05, "loss": 0.2791, "num_input_tokens_seen": 4529280, "step": 16895 }, { "epoch": 4.4298820445609435, "grad_norm": 1.625253677368164, "learning_rate": 3.412500766508063e-05, "loss": 0.3281, "num_input_tokens_seen": 4530736, "step": 16900 }, { "epoch": 4.431192660550459, "grad_norm": 0.5635849833488464, "learning_rate": 3.411435851625179e-05, "loss": 0.4035, "num_input_tokens_seen": 4532336, "step": 16905 }, { "epoch": 4.432503276539974, "grad_norm": 1.6652357578277588, "learning_rate": 3.4103707459803456e-05, "loss": 0.3491, "num_input_tokens_seen": 4533728, "step": 16910 }, { "epoch": 4.433813892529489, "grad_norm": 0.9639471173286438, "learning_rate": 3.4093054497964884e-05, "loss": 0.3709, "num_input_tokens_seen": 4534992, "step": 16915 }, { "epoch": 4.435124508519004, "grad_norm": 1.473474383354187, "learning_rate": 3.40823996329657e-05, "loss": 0.2874, "num_input_tokens_seen": 4536512, "step": 16920 }, { "epoch": 4.436435124508519, "grad_norm": 0.7343546152114868, "learning_rate": 3.4071742867035985e-05, "loss": 0.4692, "num_input_tokens_seen": 4537984, "step": 16925 }, { "epoch": 4.437745740498034, "grad_norm": 1.3514927625656128, "learning_rate": 3.4061084202406154e-05, "loss": 0.2605, "num_input_tokens_seen": 4540032, "step": 16930 }, { "epoch": 4.43905635648755, "grad_norm": 0.6499561667442322, "learning_rate": 3.405042364130706e-05, "loss": 0.2796, "num_input_tokens_seen": 4541792, "step": 16935 }, { "epoch": 4.440366972477064, "grad_norm": 1.0656449794769287, "learning_rate": 3.403976118596994e-05, "loss": 0.3713, "num_input_tokens_seen": 4542976, "step": 16940 }, { "epoch": 4.441677588466579, "grad_norm": 0.9863684177398682, "learning_rate": 3.4029096838626415e-05, "loss": 0.346, "num_input_tokens_seen": 4544368, "step": 16945 }, { "epoch": 4.442988204456094, "grad_norm": 1.57991361618042, "learning_rate": 3.401843060150853e-05, "loss": 0.2989, "num_input_tokens_seen": 4545648, "step": 16950 }, { "epoch": 4.444298820445609, "grad_norm": 1.1065069437026978, "learning_rate": 3.400776247684869e-05, "loss": 0.2979, "num_input_tokens_seen": 4546624, "step": 16955 }, { "epoch": 4.445609436435125, "grad_norm": 1.4316527843475342, "learning_rate": 3.399709246687974e-05, "loss": 0.2594, "num_input_tokens_seen": 4547968, "step": 16960 }, { "epoch": 4.446920052424639, "grad_norm": 1.0148862600326538, "learning_rate": 3.398642057383488e-05, "loss": 0.3572, "num_input_tokens_seen": 4549872, "step": 16965 }, { "epoch": 4.448230668414155, "grad_norm": 1.2101045846939087, "learning_rate": 3.397574679994772e-05, "loss": 0.2768, "num_input_tokens_seen": 4551216, "step": 16970 }, { "epoch": 4.4495412844036695, "grad_norm": 2.0488626956939697, "learning_rate": 3.396507114745226e-05, "loss": 0.3767, "num_input_tokens_seen": 4552288, "step": 16975 }, { "epoch": 4.450851900393185, "grad_norm": 0.9082562327384949, "learning_rate": 3.3954393618582905e-05, "loss": 0.338, "num_input_tokens_seen": 4553664, "step": 16980 }, { "epoch": 4.4521625163827, "grad_norm": 1.433767318725586, "learning_rate": 3.394371421557444e-05, "loss": 0.4724, "num_input_tokens_seen": 4555104, "step": 16985 }, { "epoch": 4.453473132372215, "grad_norm": 1.3095331192016602, "learning_rate": 3.3933032940662034e-05, "loss": 0.3763, "num_input_tokens_seen": 4556752, "step": 16990 }, { "epoch": 4.45478374836173, "grad_norm": 1.5582737922668457, "learning_rate": 3.392234979608128e-05, "loss": 0.3414, "num_input_tokens_seen": 4557728, "step": 16995 }, { "epoch": 4.456094364351245, "grad_norm": 1.0662201642990112, "learning_rate": 3.391166478406812e-05, "loss": 0.2461, "num_input_tokens_seen": 4559088, "step": 17000 }, { "epoch": 4.45740498034076, "grad_norm": 1.3722349405288696, "learning_rate": 3.390097790685892e-05, "loss": 0.3667, "num_input_tokens_seen": 4560304, "step": 17005 }, { "epoch": 4.458715596330276, "grad_norm": 1.003305435180664, "learning_rate": 3.3890289166690425e-05, "loss": 0.3081, "num_input_tokens_seen": 4561808, "step": 17010 }, { "epoch": 4.46002621231979, "grad_norm": 1.030430793762207, "learning_rate": 3.387959856579977e-05, "loss": 0.3805, "num_input_tokens_seen": 4563984, "step": 17015 }, { "epoch": 4.461336828309306, "grad_norm": 0.9437968730926514, "learning_rate": 3.386890610642448e-05, "loss": 0.5515, "num_input_tokens_seen": 4565744, "step": 17020 }, { "epoch": 4.46264744429882, "grad_norm": 1.39519464969635, "learning_rate": 3.3858211790802466e-05, "loss": 0.3094, "num_input_tokens_seen": 4566768, "step": 17025 }, { "epoch": 4.463958060288336, "grad_norm": 2.2872087955474854, "learning_rate": 3.384751562117203e-05, "loss": 0.3496, "num_input_tokens_seen": 4567952, "step": 17030 }, { "epoch": 4.465268676277851, "grad_norm": 1.081591248512268, "learning_rate": 3.3836817599771854e-05, "loss": 0.3436, "num_input_tokens_seen": 4569120, "step": 17035 }, { "epoch": 4.466579292267365, "grad_norm": 0.7517439126968384, "learning_rate": 3.382611772884103e-05, "loss": 0.456, "num_input_tokens_seen": 4570448, "step": 17040 }, { "epoch": 4.467889908256881, "grad_norm": 7.374482154846191, "learning_rate": 3.3815416010619013e-05, "loss": 0.3058, "num_input_tokens_seen": 4572144, "step": 17045 }, { "epoch": 4.4692005242463955, "grad_norm": 0.9888402223587036, "learning_rate": 3.380471244734566e-05, "loss": 0.247, "num_input_tokens_seen": 4573296, "step": 17050 }, { "epoch": 4.470511140235911, "grad_norm": 1.8144021034240723, "learning_rate": 3.3794007041261205e-05, "loss": 0.2479, "num_input_tokens_seen": 4574336, "step": 17055 }, { "epoch": 4.471821756225426, "grad_norm": 1.1229172945022583, "learning_rate": 3.378329979460627e-05, "loss": 0.3499, "num_input_tokens_seen": 4575840, "step": 17060 }, { "epoch": 4.473132372214941, "grad_norm": 1.162177562713623, "learning_rate": 3.3772590709621864e-05, "loss": 0.2641, "num_input_tokens_seen": 4577072, "step": 17065 }, { "epoch": 4.474442988204456, "grad_norm": 2.0642282962799072, "learning_rate": 3.376187978854938e-05, "loss": 0.3126, "num_input_tokens_seen": 4578384, "step": 17070 }, { "epoch": 4.475753604193971, "grad_norm": 1.0217961072921753, "learning_rate": 3.3751167033630596e-05, "loss": 0.2442, "num_input_tokens_seen": 4579808, "step": 17075 }, { "epoch": 4.477064220183486, "grad_norm": 10.422452926635742, "learning_rate": 3.374045244710767e-05, "loss": 0.3627, "num_input_tokens_seen": 4580736, "step": 17080 }, { "epoch": 4.478374836173002, "grad_norm": 1.1199649572372437, "learning_rate": 3.372973603122316e-05, "loss": 0.2876, "num_input_tokens_seen": 4581792, "step": 17085 }, { "epoch": 4.479685452162516, "grad_norm": 3.1055965423583984, "learning_rate": 3.371901778821997e-05, "loss": 0.3201, "num_input_tokens_seen": 4583040, "step": 17090 }, { "epoch": 4.480996068152032, "grad_norm": 0.802042543888092, "learning_rate": 3.3708297720341425e-05, "loss": 0.3873, "num_input_tokens_seen": 4584320, "step": 17095 }, { "epoch": 4.482306684141546, "grad_norm": 1.0711039304733276, "learning_rate": 3.369757582983122e-05, "loss": 0.4405, "num_input_tokens_seen": 4585584, "step": 17100 }, { "epoch": 4.483617300131062, "grad_norm": 0.8210697174072266, "learning_rate": 3.368685211893342e-05, "loss": 0.4625, "num_input_tokens_seen": 4587136, "step": 17105 }, { "epoch": 4.484927916120577, "grad_norm": 1.2251169681549072, "learning_rate": 3.367612658989247e-05, "loss": 0.3832, "num_input_tokens_seen": 4588624, "step": 17110 }, { "epoch": 4.486238532110092, "grad_norm": 2.747007369995117, "learning_rate": 3.3665399244953236e-05, "loss": 0.3045, "num_input_tokens_seen": 4589632, "step": 17115 }, { "epoch": 4.487549148099607, "grad_norm": 3.6267905235290527, "learning_rate": 3.36546700863609e-05, "loss": 0.3078, "num_input_tokens_seen": 4590512, "step": 17120 }, { "epoch": 4.488859764089122, "grad_norm": 3.0889782905578613, "learning_rate": 3.364393911636107e-05, "loss": 0.3421, "num_input_tokens_seen": 4591600, "step": 17125 }, { "epoch": 4.490170380078637, "grad_norm": 1.175399661064148, "learning_rate": 3.363320633719972e-05, "loss": 0.3793, "num_input_tokens_seen": 4592640, "step": 17130 }, { "epoch": 4.491480996068152, "grad_norm": 2.898271083831787, "learning_rate": 3.3622471751123196e-05, "loss": 0.5354, "num_input_tokens_seen": 4594224, "step": 17135 }, { "epoch": 4.492791612057667, "grad_norm": 0.9100838303565979, "learning_rate": 3.361173536037824e-05, "loss": 0.2927, "num_input_tokens_seen": 4595808, "step": 17140 }, { "epoch": 4.494102228047182, "grad_norm": 1.603742003440857, "learning_rate": 3.360099716721195e-05, "loss": 0.294, "num_input_tokens_seen": 4597264, "step": 17145 }, { "epoch": 4.495412844036697, "grad_norm": 1.0675230026245117, "learning_rate": 3.3590257173871807e-05, "loss": 0.363, "num_input_tokens_seen": 4598544, "step": 17150 }, { "epoch": 4.496723460026212, "grad_norm": 0.9479562640190125, "learning_rate": 3.357951538260569e-05, "loss": 0.4405, "num_input_tokens_seen": 4599728, "step": 17155 }, { "epoch": 4.498034076015728, "grad_norm": 0.8173244595527649, "learning_rate": 3.356877179566183e-05, "loss": 0.2826, "num_input_tokens_seen": 4601216, "step": 17160 }, { "epoch": 4.499344692005242, "grad_norm": 1.262125849723816, "learning_rate": 3.355802641528883e-05, "loss": 0.4, "num_input_tokens_seen": 4602576, "step": 17165 }, { "epoch": 4.500655307994758, "grad_norm": 1.073205828666687, "learning_rate": 3.3547279243735695e-05, "loss": 0.3171, "num_input_tokens_seen": 4603984, "step": 17170 }, { "epoch": 4.501179554390563, "eval_loss": 0.5320618748664856, "eval_runtime": 16.6868, "eval_samples_per_second": 50.819, "eval_steps_per_second": 25.409, "num_input_tokens_seen": 4604416, "step": 17172 }, { "epoch": 4.501965923984272, "grad_norm": 0.36322739720344543, "learning_rate": 3.3536530283251784e-05, "loss": 0.3659, "num_input_tokens_seen": 4605232, "step": 17175 }, { "epoch": 4.503276539973788, "grad_norm": 0.5253325700759888, "learning_rate": 3.3525779536086826e-05, "loss": 0.2808, "num_input_tokens_seen": 4606368, "step": 17180 }, { "epoch": 4.504587155963303, "grad_norm": 1.2849162817001343, "learning_rate": 3.351502700449095e-05, "loss": 0.4726, "num_input_tokens_seen": 4607600, "step": 17185 }, { "epoch": 4.505897771952818, "grad_norm": 2.0533149242401123, "learning_rate": 3.350427269071463e-05, "loss": 0.3117, "num_input_tokens_seen": 4608800, "step": 17190 }, { "epoch": 4.507208387942333, "grad_norm": 2.169210910797119, "learning_rate": 3.3493516597008734e-05, "loss": 0.3696, "num_input_tokens_seen": 4609696, "step": 17195 }, { "epoch": 4.508519003931848, "grad_norm": 0.7058953046798706, "learning_rate": 3.348275872562449e-05, "loss": 0.2604, "num_input_tokens_seen": 4610944, "step": 17200 }, { "epoch": 4.509829619921363, "grad_norm": 1.0049231052398682, "learning_rate": 3.34719990788135e-05, "loss": 0.3271, "num_input_tokens_seen": 4612368, "step": 17205 }, { "epoch": 4.511140235910878, "grad_norm": 1.7192293405532837, "learning_rate": 3.346123765882774e-05, "loss": 0.2845, "num_input_tokens_seen": 4613472, "step": 17210 }, { "epoch": 4.512450851900393, "grad_norm": 1.1062496900558472, "learning_rate": 3.345047446791956e-05, "loss": 0.4879, "num_input_tokens_seen": 4614688, "step": 17215 }, { "epoch": 4.513761467889909, "grad_norm": 0.5679647922515869, "learning_rate": 3.3439709508341676e-05, "loss": 0.303, "num_input_tokens_seen": 4616160, "step": 17220 }, { "epoch": 4.515072083879423, "grad_norm": 0.3867059051990509, "learning_rate": 3.342894278234717e-05, "loss": 0.2622, "num_input_tokens_seen": 4617264, "step": 17225 }, { "epoch": 4.516382699868938, "grad_norm": 0.9726400971412659, "learning_rate": 3.341817429218951e-05, "loss": 0.3732, "num_input_tokens_seen": 4619456, "step": 17230 }, { "epoch": 4.517693315858454, "grad_norm": 0.7944853901863098, "learning_rate": 3.340740404012251e-05, "loss": 0.2826, "num_input_tokens_seen": 4620960, "step": 17235 }, { "epoch": 4.519003931847968, "grad_norm": 0.8770360350608826, "learning_rate": 3.339663202840038e-05, "loss": 0.2577, "num_input_tokens_seen": 4622176, "step": 17240 }, { "epoch": 4.520314547837484, "grad_norm": 1.7073373794555664, "learning_rate": 3.3385858259277666e-05, "loss": 0.3638, "num_input_tokens_seen": 4623808, "step": 17245 }, { "epoch": 4.521625163826998, "grad_norm": 2.3150758743286133, "learning_rate": 3.337508273500931e-05, "loss": 0.3949, "num_input_tokens_seen": 4626640, "step": 17250 }, { "epoch": 4.522935779816514, "grad_norm": 1.2864753007888794, "learning_rate": 3.336430545785061e-05, "loss": 0.242, "num_input_tokens_seen": 4627824, "step": 17255 }, { "epoch": 4.524246395806029, "grad_norm": 1.044870376586914, "learning_rate": 3.335352643005723e-05, "loss": 0.2886, "num_input_tokens_seen": 4629232, "step": 17260 }, { "epoch": 4.525557011795544, "grad_norm": 1.346872091293335, "learning_rate": 3.33427456538852e-05, "loss": 0.4758, "num_input_tokens_seen": 4630416, "step": 17265 }, { "epoch": 4.526867627785059, "grad_norm": 0.9082647562026978, "learning_rate": 3.3331963131590916e-05, "loss": 0.444, "num_input_tokens_seen": 4631984, "step": 17270 }, { "epoch": 4.528178243774574, "grad_norm": 1.0945342779159546, "learning_rate": 3.3321178865431155e-05, "loss": 0.2368, "num_input_tokens_seen": 4633104, "step": 17275 }, { "epoch": 4.529488859764089, "grad_norm": 1.7758946418762207, "learning_rate": 3.331039285766303e-05, "loss": 0.4326, "num_input_tokens_seen": 4634656, "step": 17280 }, { "epoch": 4.5307994757536045, "grad_norm": 1.6179851293563843, "learning_rate": 3.329960511054404e-05, "loss": 0.4674, "num_input_tokens_seen": 4635616, "step": 17285 }, { "epoch": 4.532110091743119, "grad_norm": 1.5101810693740845, "learning_rate": 3.3288815626332035e-05, "loss": 0.2064, "num_input_tokens_seen": 4636720, "step": 17290 }, { "epoch": 4.533420707732635, "grad_norm": 3.891439199447632, "learning_rate": 3.327802440728526e-05, "loss": 0.41, "num_input_tokens_seen": 4637904, "step": 17295 }, { "epoch": 4.534731323722149, "grad_norm": 2.6917383670806885, "learning_rate": 3.3267231455662265e-05, "loss": 0.3501, "num_input_tokens_seen": 4639264, "step": 17300 }, { "epoch": 4.536041939711664, "grad_norm": 1.428631067276001, "learning_rate": 3.325643677372201e-05, "loss": 0.2624, "num_input_tokens_seen": 4640176, "step": 17305 }, { "epoch": 4.53735255570118, "grad_norm": 1.684189796447754, "learning_rate": 3.324564036372381e-05, "loss": 0.4363, "num_input_tokens_seen": 4641440, "step": 17310 }, { "epoch": 4.538663171690695, "grad_norm": 1.313414454460144, "learning_rate": 3.323484222792732e-05, "loss": 0.3804, "num_input_tokens_seen": 4642496, "step": 17315 }, { "epoch": 4.53997378768021, "grad_norm": 0.9854404926300049, "learning_rate": 3.322404236859259e-05, "loss": 0.4433, "num_input_tokens_seen": 4643744, "step": 17320 }, { "epoch": 4.541284403669724, "grad_norm": 3.494051218032837, "learning_rate": 3.321324078798e-05, "loss": 0.2731, "num_input_tokens_seen": 4645024, "step": 17325 }, { "epoch": 4.54259501965924, "grad_norm": 0.9226874709129333, "learning_rate": 3.32024374883503e-05, "loss": 0.2775, "num_input_tokens_seen": 4645984, "step": 17330 }, { "epoch": 4.543905635648755, "grad_norm": 2.325460910797119, "learning_rate": 3.319163247196462e-05, "loss": 0.4362, "num_input_tokens_seen": 4647072, "step": 17335 }, { "epoch": 4.54521625163827, "grad_norm": 0.8318047523498535, "learning_rate": 3.318082574108441e-05, "loss": 0.3123, "num_input_tokens_seen": 4648304, "step": 17340 }, { "epoch": 4.546526867627785, "grad_norm": 1.1312474012374878, "learning_rate": 3.31700172979715e-05, "loss": 0.2998, "num_input_tokens_seen": 4649728, "step": 17345 }, { "epoch": 4.5478374836173, "grad_norm": 0.747988760471344, "learning_rate": 3.315920714488808e-05, "loss": 0.2583, "num_input_tokens_seen": 4650816, "step": 17350 }, { "epoch": 4.549148099606815, "grad_norm": 2.8147192001342773, "learning_rate": 3.314839528409672e-05, "loss": 0.4058, "num_input_tokens_seen": 4651888, "step": 17355 }, { "epoch": 4.5504587155963305, "grad_norm": 1.8526959419250488, "learning_rate": 3.313758171786031e-05, "loss": 0.4154, "num_input_tokens_seen": 4653008, "step": 17360 }, { "epoch": 4.551769331585845, "grad_norm": 1.591301679611206, "learning_rate": 3.3126766448442095e-05, "loss": 0.4453, "num_input_tokens_seen": 4654560, "step": 17365 }, { "epoch": 4.553079947575361, "grad_norm": 0.9847530722618103, "learning_rate": 3.311594947810571e-05, "loss": 0.3392, "num_input_tokens_seen": 4655552, "step": 17370 }, { "epoch": 4.554390563564875, "grad_norm": 1.7157139778137207, "learning_rate": 3.310513080911512e-05, "loss": 0.4899, "num_input_tokens_seen": 4656800, "step": 17375 }, { "epoch": 4.555701179554391, "grad_norm": 1.0904216766357422, "learning_rate": 3.309431044373465e-05, "loss": 0.4066, "num_input_tokens_seen": 4658208, "step": 17380 }, { "epoch": 4.557011795543906, "grad_norm": 1.8107309341430664, "learning_rate": 3.3083488384228995e-05, "loss": 0.2069, "num_input_tokens_seen": 4659248, "step": 17385 }, { "epoch": 4.558322411533421, "grad_norm": 1.075240969657898, "learning_rate": 3.3072664632863195e-05, "loss": 0.2821, "num_input_tokens_seen": 4660480, "step": 17390 }, { "epoch": 4.559633027522936, "grad_norm": 2.0191657543182373, "learning_rate": 3.3061839191902635e-05, "loss": 0.353, "num_input_tokens_seen": 4661408, "step": 17395 }, { "epoch": 4.56094364351245, "grad_norm": 0.717889666557312, "learning_rate": 3.3051012063613054e-05, "loss": 0.2316, "num_input_tokens_seen": 4662464, "step": 17400 }, { "epoch": 4.562254259501966, "grad_norm": 2.242738962173462, "learning_rate": 3.3040183250260556e-05, "loss": 0.4916, "num_input_tokens_seen": 4663744, "step": 17405 }, { "epoch": 4.5635648754914815, "grad_norm": 1.2516344785690308, "learning_rate": 3.30293527541116e-05, "loss": 0.3597, "num_input_tokens_seen": 4665008, "step": 17410 }, { "epoch": 4.564875491480996, "grad_norm": 0.3255641460418701, "learning_rate": 3.3018520577432996e-05, "loss": 0.3276, "num_input_tokens_seen": 4666672, "step": 17415 }, { "epoch": 4.566186107470511, "grad_norm": 1.2726229429244995, "learning_rate": 3.3007686722491885e-05, "loss": 0.293, "num_input_tokens_seen": 4667552, "step": 17420 }, { "epoch": 4.567496723460026, "grad_norm": 0.6791831851005554, "learning_rate": 3.299685119155577e-05, "loss": 0.4108, "num_input_tokens_seen": 4668592, "step": 17425 }, { "epoch": 4.568807339449541, "grad_norm": 1.4055804014205933, "learning_rate": 3.298601398689253e-05, "loss": 0.2346, "num_input_tokens_seen": 4669984, "step": 17430 }, { "epoch": 4.5701179554390565, "grad_norm": 1.9092109203338623, "learning_rate": 3.297517511077035e-05, "loss": 0.5064, "num_input_tokens_seen": 4671168, "step": 17435 }, { "epoch": 4.571428571428571, "grad_norm": 1.6873809099197388, "learning_rate": 3.296433456545781e-05, "loss": 0.3495, "num_input_tokens_seen": 4672768, "step": 17440 }, { "epoch": 4.572739187418087, "grad_norm": 0.9169571399688721, "learning_rate": 3.2953492353223805e-05, "loss": 0.3292, "num_input_tokens_seen": 4674544, "step": 17445 }, { "epoch": 4.574049803407601, "grad_norm": 1.746522307395935, "learning_rate": 3.294264847633759e-05, "loss": 0.4412, "num_input_tokens_seen": 4675584, "step": 17450 }, { "epoch": 4.575360419397117, "grad_norm": 1.987412929534912, "learning_rate": 3.2931802937068776e-05, "loss": 0.5927, "num_input_tokens_seen": 4676784, "step": 17455 }, { "epoch": 4.576671035386632, "grad_norm": 0.9479457139968872, "learning_rate": 3.292095573768731e-05, "loss": 0.3293, "num_input_tokens_seen": 4677856, "step": 17460 }, { "epoch": 4.577981651376147, "grad_norm": 1.1216373443603516, "learning_rate": 3.2910106880463486e-05, "loss": 0.3113, "num_input_tokens_seen": 4679040, "step": 17465 }, { "epoch": 4.579292267365662, "grad_norm": 2.6949872970581055, "learning_rate": 3.289925636766797e-05, "loss": 0.3, "num_input_tokens_seen": 4680400, "step": 17470 }, { "epoch": 4.580602883355177, "grad_norm": 3.088322401046753, "learning_rate": 3.2888404201571743e-05, "loss": 0.3508, "num_input_tokens_seen": 4681888, "step": 17475 }, { "epoch": 4.581913499344692, "grad_norm": 1.0131120681762695, "learning_rate": 3.2877550384446146e-05, "loss": 0.3698, "num_input_tokens_seen": 4682848, "step": 17480 }, { "epoch": 4.5832241153342075, "grad_norm": 1.817142367362976, "learning_rate": 3.2866694918562866e-05, "loss": 0.3866, "num_input_tokens_seen": 4684096, "step": 17485 }, { "epoch": 4.584534731323722, "grad_norm": 0.5655882954597473, "learning_rate": 3.285583780619394e-05, "loss": 0.2799, "num_input_tokens_seen": 4685344, "step": 17490 }, { "epoch": 4.585845347313237, "grad_norm": 0.7644572854042053, "learning_rate": 3.284497904961172e-05, "loss": 0.3547, "num_input_tokens_seen": 4686768, "step": 17495 }, { "epoch": 4.587155963302752, "grad_norm": 2.415073871612549, "learning_rate": 3.283411865108894e-05, "loss": 0.4096, "num_input_tokens_seen": 4687776, "step": 17500 }, { "epoch": 4.588466579292267, "grad_norm": 1.522179365158081, "learning_rate": 3.2823256612898676e-05, "loss": 0.3954, "num_input_tokens_seen": 4688752, "step": 17505 }, { "epoch": 4.5897771952817825, "grad_norm": 1.5998930931091309, "learning_rate": 3.2812392937314316e-05, "loss": 0.3053, "num_input_tokens_seen": 4690128, "step": 17510 }, { "epoch": 4.591087811271297, "grad_norm": 2.816850423812866, "learning_rate": 3.2801527626609606e-05, "loss": 0.3357, "num_input_tokens_seen": 4691312, "step": 17515 }, { "epoch": 4.592398427260813, "grad_norm": 1.335915446281433, "learning_rate": 3.2790660683058646e-05, "loss": 0.3626, "num_input_tokens_seen": 4693120, "step": 17520 }, { "epoch": 4.593709043250327, "grad_norm": 0.5113058090209961, "learning_rate": 3.277979210893587e-05, "loss": 0.3339, "num_input_tokens_seen": 4694640, "step": 17525 }, { "epoch": 4.595019659239843, "grad_norm": 0.5241709351539612, "learning_rate": 3.276892190651605e-05, "loss": 0.2774, "num_input_tokens_seen": 4695888, "step": 17530 }, { "epoch": 4.5963302752293576, "grad_norm": 1.9735461473464966, "learning_rate": 3.275805007807429e-05, "loss": 0.2859, "num_input_tokens_seen": 4696928, "step": 17535 }, { "epoch": 4.597640891218873, "grad_norm": 0.9554915428161621, "learning_rate": 3.274717662588606e-05, "loss": 0.4601, "num_input_tokens_seen": 4698608, "step": 17540 }, { "epoch": 4.598951507208388, "grad_norm": 0.8704994916915894, "learning_rate": 3.273630155222715e-05, "loss": 0.2135, "num_input_tokens_seen": 4699856, "step": 17545 }, { "epoch": 4.600262123197903, "grad_norm": 1.5572388172149658, "learning_rate": 3.272542485937369e-05, "loss": 0.3247, "num_input_tokens_seen": 4701488, "step": 17550 }, { "epoch": 4.601572739187418, "grad_norm": 0.9712796211242676, "learning_rate": 3.271454654960215e-05, "loss": 0.4417, "num_input_tokens_seen": 4702960, "step": 17555 }, { "epoch": 4.6028833551769335, "grad_norm": 2.1970973014831543, "learning_rate": 3.270366662518936e-05, "loss": 0.2791, "num_input_tokens_seen": 4704208, "step": 17560 }, { "epoch": 4.604193971166448, "grad_norm": 1.1810392141342163, "learning_rate": 3.2692785088412444e-05, "loss": 0.1648, "num_input_tokens_seen": 4705440, "step": 17565 }, { "epoch": 4.605504587155964, "grad_norm": 3.432814121246338, "learning_rate": 3.268190194154891e-05, "loss": 0.3876, "num_input_tokens_seen": 4706944, "step": 17570 }, { "epoch": 4.606815203145478, "grad_norm": 2.923427104949951, "learning_rate": 3.267101718687657e-05, "loss": 0.3255, "num_input_tokens_seen": 4708304, "step": 17575 }, { "epoch": 4.608125819134994, "grad_norm": 0.7126815915107727, "learning_rate": 3.266013082667357e-05, "loss": 0.1875, "num_input_tokens_seen": 4709424, "step": 17580 }, { "epoch": 4.6094364351245085, "grad_norm": 2.2169580459594727, "learning_rate": 3.2649242863218446e-05, "loss": 0.3599, "num_input_tokens_seen": 4710832, "step": 17585 }, { "epoch": 4.610747051114023, "grad_norm": 1.2263528108596802, "learning_rate": 3.2638353298790005e-05, "loss": 0.3269, "num_input_tokens_seen": 4712144, "step": 17590 }, { "epoch": 4.612057667103539, "grad_norm": 1.188721776008606, "learning_rate": 3.2627462135667405e-05, "loss": 0.4781, "num_input_tokens_seen": 4713648, "step": 17595 }, { "epoch": 4.613368283093053, "grad_norm": 1.3668652772903442, "learning_rate": 3.2616569376130175e-05, "loss": 0.2287, "num_input_tokens_seen": 4715168, "step": 17600 }, { "epoch": 4.614678899082569, "grad_norm": 1.8802216053009033, "learning_rate": 3.260567502245813e-05, "loss": 0.5434, "num_input_tokens_seen": 4716272, "step": 17605 }, { "epoch": 4.6159895150720835, "grad_norm": 1.0616182088851929, "learning_rate": 3.259477907693144e-05, "loss": 0.3086, "num_input_tokens_seen": 4717504, "step": 17610 }, { "epoch": 4.617300131061599, "grad_norm": 1.990250587463379, "learning_rate": 3.2583881541830604e-05, "loss": 0.3713, "num_input_tokens_seen": 4718768, "step": 17615 }, { "epoch": 4.618610747051114, "grad_norm": 3.712127208709717, "learning_rate": 3.257298241943648e-05, "loss": 0.3486, "num_input_tokens_seen": 4720160, "step": 17620 }, { "epoch": 4.619921363040629, "grad_norm": 1.8061048984527588, "learning_rate": 3.256208171203021e-05, "loss": 0.3632, "num_input_tokens_seen": 4721168, "step": 17625 }, { "epoch": 4.621231979030144, "grad_norm": 1.995659351348877, "learning_rate": 3.255117942189331e-05, "loss": 0.3081, "num_input_tokens_seen": 4722304, "step": 17630 }, { "epoch": 4.6225425950196595, "grad_norm": 6.85903263092041, "learning_rate": 3.2540275551307605e-05, "loss": 0.2295, "num_input_tokens_seen": 4723792, "step": 17635 }, { "epoch": 4.623853211009174, "grad_norm": 1.9550557136535645, "learning_rate": 3.252937010255525e-05, "loss": 0.3249, "num_input_tokens_seen": 4725312, "step": 17640 }, { "epoch": 4.62516382699869, "grad_norm": 2.12274169921875, "learning_rate": 3.2518463077918746e-05, "loss": 0.2497, "num_input_tokens_seen": 4726400, "step": 17645 }, { "epoch": 4.626474442988204, "grad_norm": 0.9498224258422852, "learning_rate": 3.250755447968091e-05, "loss": 0.3796, "num_input_tokens_seen": 4728416, "step": 17650 }, { "epoch": 4.62778505897772, "grad_norm": 2.5365023612976074, "learning_rate": 3.24966443101249e-05, "loss": 0.4414, "num_input_tokens_seen": 4729472, "step": 17655 }, { "epoch": 4.6290956749672345, "grad_norm": 0.8436103463172913, "learning_rate": 3.248573257153418e-05, "loss": 0.4139, "num_input_tokens_seen": 4730912, "step": 17660 }, { "epoch": 4.63040629095675, "grad_norm": 2.161898612976074, "learning_rate": 3.2474819266192574e-05, "loss": 0.22, "num_input_tokens_seen": 4732064, "step": 17665 }, { "epoch": 4.631716906946265, "grad_norm": 0.6871098279953003, "learning_rate": 3.24639043963842e-05, "loss": 0.3745, "num_input_tokens_seen": 4733536, "step": 17670 }, { "epoch": 4.63302752293578, "grad_norm": 0.9018607139587402, "learning_rate": 3.245298796439354e-05, "loss": 0.2067, "num_input_tokens_seen": 4734992, "step": 17675 }, { "epoch": 4.634338138925295, "grad_norm": 1.1304051876068115, "learning_rate": 3.244206997250538e-05, "loss": 0.3415, "num_input_tokens_seen": 4736480, "step": 17680 }, { "epoch": 4.6356487549148095, "grad_norm": 4.1544880867004395, "learning_rate": 3.243115042300483e-05, "loss": 0.3395, "num_input_tokens_seen": 4737936, "step": 17685 }, { "epoch": 4.636959370904325, "grad_norm": 1.3328922986984253, "learning_rate": 3.242022931817734e-05, "loss": 0.3379, "num_input_tokens_seen": 4739088, "step": 17690 }, { "epoch": 4.63826998689384, "grad_norm": 1.6705633401870728, "learning_rate": 3.240930666030867e-05, "loss": 0.3323, "num_input_tokens_seen": 4740352, "step": 17695 }, { "epoch": 4.639580602883355, "grad_norm": 1.674945592880249, "learning_rate": 3.239838245168492e-05, "loss": 0.3359, "num_input_tokens_seen": 4741520, "step": 17700 }, { "epoch": 4.64089121887287, "grad_norm": 1.9718936681747437, "learning_rate": 3.238745669459251e-05, "loss": 0.332, "num_input_tokens_seen": 4742672, "step": 17705 }, { "epoch": 4.6422018348623855, "grad_norm": 1.108044147491455, "learning_rate": 3.237652939131818e-05, "loss": 0.4053, "num_input_tokens_seen": 4743920, "step": 17710 }, { "epoch": 4.6435124508519, "grad_norm": 0.8206292390823364, "learning_rate": 3.236560054414899e-05, "loss": 0.3556, "num_input_tokens_seen": 4745472, "step": 17715 }, { "epoch": 4.644823066841416, "grad_norm": 1.983821153640747, "learning_rate": 3.2354670155372326e-05, "loss": 0.5141, "num_input_tokens_seen": 4746528, "step": 17720 }, { "epoch": 4.64613368283093, "grad_norm": 1.3358550071716309, "learning_rate": 3.234373822727591e-05, "loss": 0.5522, "num_input_tokens_seen": 4747744, "step": 17725 }, { "epoch": 4.647444298820446, "grad_norm": 3.604257583618164, "learning_rate": 3.233280476214778e-05, "loss": 0.3078, "num_input_tokens_seen": 4748992, "step": 17730 }, { "epoch": 4.6487549148099605, "grad_norm": 1.502956509590149, "learning_rate": 3.232186976227628e-05, "loss": 0.4461, "num_input_tokens_seen": 4750240, "step": 17735 }, { "epoch": 4.650065530799476, "grad_norm": 1.0101648569107056, "learning_rate": 3.231093322995008e-05, "loss": 0.3458, "num_input_tokens_seen": 4751376, "step": 17740 }, { "epoch": 4.651376146788991, "grad_norm": 1.0317356586456299, "learning_rate": 3.2299995167458196e-05, "loss": 0.25, "num_input_tokens_seen": 4752720, "step": 17745 }, { "epoch": 4.652686762778506, "grad_norm": 1.1127538681030273, "learning_rate": 3.2289055577089936e-05, "loss": 0.4173, "num_input_tokens_seen": 4754144, "step": 17750 }, { "epoch": 4.653997378768021, "grad_norm": 1.3433735370635986, "learning_rate": 3.227811446113494e-05, "loss": 0.4126, "num_input_tokens_seen": 4755568, "step": 17755 }, { "epoch": 4.655307994757536, "grad_norm": 0.4701380729675293, "learning_rate": 3.226717182188316e-05, "loss": 0.4213, "num_input_tokens_seen": 4756816, "step": 17760 }, { "epoch": 4.656618610747051, "grad_norm": 1.7572624683380127, "learning_rate": 3.2256227661624874e-05, "loss": 0.584, "num_input_tokens_seen": 4758080, "step": 17765 }, { "epoch": 4.657929226736567, "grad_norm": 5.391920566558838, "learning_rate": 3.2245281982650683e-05, "loss": 0.5601, "num_input_tokens_seen": 4759248, "step": 17770 }, { "epoch": 4.659239842726081, "grad_norm": 1.0308719873428345, "learning_rate": 3.2234334787251487e-05, "loss": 0.3898, "num_input_tokens_seen": 4760720, "step": 17775 }, { "epoch": 4.660550458715596, "grad_norm": 1.203073263168335, "learning_rate": 3.222338607771851e-05, "loss": 0.3116, "num_input_tokens_seen": 4762256, "step": 17780 }, { "epoch": 4.6618610747051115, "grad_norm": 1.4271950721740723, "learning_rate": 3.221243585634333e-05, "loss": 0.3876, "num_input_tokens_seen": 4763632, "step": 17785 }, { "epoch": 4.663171690694626, "grad_norm": 0.9300002455711365, "learning_rate": 3.2201484125417766e-05, "loss": 0.2265, "num_input_tokens_seen": 4765456, "step": 17790 }, { "epoch": 4.664482306684142, "grad_norm": 3.7457685470581055, "learning_rate": 3.219053088723403e-05, "loss": 0.3518, "num_input_tokens_seen": 4766400, "step": 17795 }, { "epoch": 4.665792922673656, "grad_norm": 1.2115554809570312, "learning_rate": 3.21795761440846e-05, "loss": 0.4383, "num_input_tokens_seen": 4767552, "step": 17800 }, { "epoch": 4.667103538663172, "grad_norm": 1.9685455560684204, "learning_rate": 3.2168619898262294e-05, "loss": 0.3638, "num_input_tokens_seen": 4768656, "step": 17805 }, { "epoch": 4.6684141546526865, "grad_norm": 1.6084606647491455, "learning_rate": 3.215766215206022e-05, "loss": 0.5377, "num_input_tokens_seen": 4770288, "step": 17810 }, { "epoch": 4.669724770642202, "grad_norm": 1.787102222442627, "learning_rate": 3.2146702907771824e-05, "loss": 0.411, "num_input_tokens_seen": 4771744, "step": 17815 }, { "epoch": 4.671035386631717, "grad_norm": 1.0314890146255493, "learning_rate": 3.2135742167690866e-05, "loss": 0.3805, "num_input_tokens_seen": 4773232, "step": 17820 }, { "epoch": 4.672346002621232, "grad_norm": 0.8983182311058044, "learning_rate": 3.212477993411141e-05, "loss": 0.6358, "num_input_tokens_seen": 4774832, "step": 17825 }, { "epoch": 4.673656618610747, "grad_norm": 1.5946640968322754, "learning_rate": 3.211381620932781e-05, "loss": 0.4067, "num_input_tokens_seen": 4776352, "step": 17830 }, { "epoch": 4.674967234600262, "grad_norm": 0.6667571067810059, "learning_rate": 3.2102850995634765e-05, "loss": 0.3606, "num_input_tokens_seen": 4777968, "step": 17835 }, { "epoch": 4.676277850589777, "grad_norm": 0.876111626625061, "learning_rate": 3.209188429532729e-05, "loss": 0.3446, "num_input_tokens_seen": 4779248, "step": 17840 }, { "epoch": 4.677588466579293, "grad_norm": 1.1644935607910156, "learning_rate": 3.2080916110700684e-05, "loss": 0.2256, "num_input_tokens_seen": 4780368, "step": 17845 }, { "epoch": 4.678899082568807, "grad_norm": 1.033519983291626, "learning_rate": 3.206994644405057e-05, "loss": 0.3845, "num_input_tokens_seen": 4782304, "step": 17850 }, { "epoch": 4.680209698558322, "grad_norm": 2.6828579902648926, "learning_rate": 3.2058975297672874e-05, "loss": 0.2836, "num_input_tokens_seen": 4783232, "step": 17855 }, { "epoch": 4.6815203145478375, "grad_norm": 0.6725556254386902, "learning_rate": 3.204800267386385e-05, "loss": 0.3069, "num_input_tokens_seen": 4784448, "step": 17860 }, { "epoch": 4.682830930537353, "grad_norm": 0.9342696070671082, "learning_rate": 3.203702857492005e-05, "loss": 0.259, "num_input_tokens_seen": 4785376, "step": 17865 }, { "epoch": 4.684141546526868, "grad_norm": 1.5700206756591797, "learning_rate": 3.202605300313833e-05, "loss": 0.4303, "num_input_tokens_seen": 4786992, "step": 17870 }, { "epoch": 4.685452162516382, "grad_norm": 0.6048070192337036, "learning_rate": 3.201507596081585e-05, "loss": 0.2729, "num_input_tokens_seen": 4787920, "step": 17875 }, { "epoch": 4.686762778505898, "grad_norm": 2.0076916217803955, "learning_rate": 3.200409745025009e-05, "loss": 0.2826, "num_input_tokens_seen": 4789392, "step": 17880 }, { "epoch": 4.6880733944954125, "grad_norm": 0.9271293878555298, "learning_rate": 3.199311747373885e-05, "loss": 0.4296, "num_input_tokens_seen": 4791888, "step": 17885 }, { "epoch": 4.689384010484928, "grad_norm": 1.4584506750106812, "learning_rate": 3.198213603358019e-05, "loss": 0.288, "num_input_tokens_seen": 4793008, "step": 17890 }, { "epoch": 4.690694626474443, "grad_norm": 1.1095099449157715, "learning_rate": 3.197115313207252e-05, "loss": 0.3103, "num_input_tokens_seen": 4794272, "step": 17895 }, { "epoch": 4.692005242463958, "grad_norm": 1.132169485092163, "learning_rate": 3.1960168771514565e-05, "loss": 0.3938, "num_input_tokens_seen": 4795552, "step": 17900 }, { "epoch": 4.693315858453473, "grad_norm": 3.169623374938965, "learning_rate": 3.1949182954205295e-05, "loss": 0.2179, "num_input_tokens_seen": 4796832, "step": 17905 }, { "epoch": 4.694626474442988, "grad_norm": 1.1213635206222534, "learning_rate": 3.193819568244405e-05, "loss": 0.4199, "num_input_tokens_seen": 4798432, "step": 17910 }, { "epoch": 4.695937090432503, "grad_norm": 1.2064170837402344, "learning_rate": 3.192720695853043e-05, "loss": 0.3233, "num_input_tokens_seen": 4799792, "step": 17915 }, { "epoch": 4.697247706422019, "grad_norm": 1.729030966758728, "learning_rate": 3.1916216784764365e-05, "loss": 0.3865, "num_input_tokens_seen": 4800864, "step": 17920 }, { "epoch": 4.698558322411533, "grad_norm": 1.3022931814193726, "learning_rate": 3.1905225163446075e-05, "loss": 0.3095, "num_input_tokens_seen": 4802176, "step": 17925 }, { "epoch": 4.699868938401049, "grad_norm": 1.6783095598220825, "learning_rate": 3.189423209687609e-05, "loss": 0.3926, "num_input_tokens_seen": 4803568, "step": 17930 }, { "epoch": 4.7011795543905635, "grad_norm": 1.0355173349380493, "learning_rate": 3.188323758735524e-05, "loss": 0.323, "num_input_tokens_seen": 4804784, "step": 17935 }, { "epoch": 4.702490170380079, "grad_norm": 1.3551372289657593, "learning_rate": 3.187224163718466e-05, "loss": 0.447, "num_input_tokens_seen": 4805920, "step": 17940 }, { "epoch": 4.703800786369594, "grad_norm": 0.980518639087677, "learning_rate": 3.186124424866578e-05, "loss": 0.3112, "num_input_tokens_seen": 4807472, "step": 17945 }, { "epoch": 4.705111402359108, "grad_norm": 1.6676244735717773, "learning_rate": 3.185024542410033e-05, "loss": 0.3737, "num_input_tokens_seen": 4808912, "step": 17950 }, { "epoch": 4.706422018348624, "grad_norm": 1.2091588973999023, "learning_rate": 3.183924516579034e-05, "loss": 0.3116, "num_input_tokens_seen": 4810016, "step": 17955 }, { "epoch": 4.707732634338139, "grad_norm": 1.729274868965149, "learning_rate": 3.182824347603818e-05, "loss": 0.2639, "num_input_tokens_seen": 4811392, "step": 17960 }, { "epoch": 4.709043250327654, "grad_norm": 1.0474050045013428, "learning_rate": 3.181724035714644e-05, "loss": 0.4108, "num_input_tokens_seen": 4812944, "step": 17965 }, { "epoch": 4.710353866317169, "grad_norm": 1.0662624835968018, "learning_rate": 3.1806235811418074e-05, "loss": 0.4221, "num_input_tokens_seen": 4814448, "step": 17970 }, { "epoch": 4.711664482306684, "grad_norm": 1.2666679620742798, "learning_rate": 3.179522984115632e-05, "loss": 0.3535, "num_input_tokens_seen": 4815632, "step": 17975 }, { "epoch": 4.712975098296199, "grad_norm": 1.378671646118164, "learning_rate": 3.178422244866469e-05, "loss": 0.4003, "num_input_tokens_seen": 4817120, "step": 17980 }, { "epoch": 4.714285714285714, "grad_norm": 0.6710944771766663, "learning_rate": 3.177321363624703e-05, "loss": 0.2418, "num_input_tokens_seen": 4818208, "step": 17985 }, { "epoch": 4.715596330275229, "grad_norm": 1.5209535360336304, "learning_rate": 3.176220340620746e-05, "loss": 0.2561, "num_input_tokens_seen": 4819600, "step": 17990 }, { "epoch": 4.716906946264745, "grad_norm": 1.413262963294983, "learning_rate": 3.17511917608504e-05, "loss": 0.283, "num_input_tokens_seen": 4820672, "step": 17995 }, { "epoch": 4.718217562254259, "grad_norm": 1.416476845741272, "learning_rate": 3.174017870248057e-05, "loss": 0.3128, "num_input_tokens_seen": 4821888, "step": 18000 }, { "epoch": 4.719528178243775, "grad_norm": 0.80423504114151, "learning_rate": 3.172916423340299e-05, "loss": 0.3161, "num_input_tokens_seen": 4823504, "step": 18005 }, { "epoch": 4.7208387942332894, "grad_norm": 1.2630188465118408, "learning_rate": 3.171814835592294e-05, "loss": 0.3354, "num_input_tokens_seen": 4825136, "step": 18010 }, { "epoch": 4.722149410222805, "grad_norm": 1.1828079223632812, "learning_rate": 3.170713107234607e-05, "loss": 0.5675, "num_input_tokens_seen": 4826672, "step": 18015 }, { "epoch": 4.72346002621232, "grad_norm": 7.2230329513549805, "learning_rate": 3.169611238497825e-05, "loss": 0.3204, "num_input_tokens_seen": 4828496, "step": 18020 }, { "epoch": 4.724770642201835, "grad_norm": 1.462994933128357, "learning_rate": 3.168509229612568e-05, "loss": 0.5146, "num_input_tokens_seen": 4829792, "step": 18025 }, { "epoch": 4.72608125819135, "grad_norm": 0.8194270133972168, "learning_rate": 3.167407080809484e-05, "loss": 0.3845, "num_input_tokens_seen": 4831216, "step": 18030 }, { "epoch": 4.727391874180865, "grad_norm": 2.0587422847747803, "learning_rate": 3.166304792319251e-05, "loss": 0.2311, "num_input_tokens_seen": 4832208, "step": 18035 }, { "epoch": 4.72870249017038, "grad_norm": 16.584733963012695, "learning_rate": 3.165202364372577e-05, "loss": 0.4004, "num_input_tokens_seen": 4833616, "step": 18040 }, { "epoch": 4.730013106159895, "grad_norm": 0.48976120352745056, "learning_rate": 3.164099797200197e-05, "loss": 0.3454, "num_input_tokens_seen": 4834928, "step": 18045 }, { "epoch": 4.73132372214941, "grad_norm": 1.289254903793335, "learning_rate": 3.162997091032876e-05, "loss": 0.5094, "num_input_tokens_seen": 4836128, "step": 18050 }, { "epoch": 4.732634338138926, "grad_norm": 0.9039122462272644, "learning_rate": 3.161894246101411e-05, "loss": 0.2948, "num_input_tokens_seen": 4837504, "step": 18055 }, { "epoch": 4.73394495412844, "grad_norm": 1.0824521780014038, "learning_rate": 3.160791262636624e-05, "loss": 0.3722, "num_input_tokens_seen": 4838720, "step": 18060 }, { "epoch": 4.735255570117955, "grad_norm": 1.4971970319747925, "learning_rate": 3.159688140869366e-05, "loss": 0.3766, "num_input_tokens_seen": 4840336, "step": 18065 }, { "epoch": 4.736566186107471, "grad_norm": 1.7019201517105103, "learning_rate": 3.158584881030521e-05, "loss": 0.3315, "num_input_tokens_seen": 4841424, "step": 18070 }, { "epoch": 4.737876802096985, "grad_norm": 1.1136701107025146, "learning_rate": 3.157481483350998e-05, "loss": 0.4879, "num_input_tokens_seen": 4842704, "step": 18075 }, { "epoch": 4.739187418086501, "grad_norm": 1.3943884372711182, "learning_rate": 3.156377948061737e-05, "loss": 0.4764, "num_input_tokens_seen": 4843888, "step": 18080 }, { "epoch": 4.740498034076015, "grad_norm": 1.56150484085083, "learning_rate": 3.155274275393705e-05, "loss": 0.2416, "num_input_tokens_seen": 4844960, "step": 18085 }, { "epoch": 4.741808650065531, "grad_norm": 0.7801145911216736, "learning_rate": 3.1541704655779e-05, "loss": 0.3787, "num_input_tokens_seen": 4846272, "step": 18090 }, { "epoch": 4.743119266055046, "grad_norm": 1.0426219701766968, "learning_rate": 3.1530665188453464e-05, "loss": 0.4857, "num_input_tokens_seen": 4847456, "step": 18095 }, { "epoch": 4.744429882044561, "grad_norm": 1.3886425495147705, "learning_rate": 3.151962435427099e-05, "loss": 0.3385, "num_input_tokens_seen": 4848560, "step": 18100 }, { "epoch": 4.745740498034076, "grad_norm": 0.8269982933998108, "learning_rate": 3.15085821555424e-05, "loss": 0.3029, "num_input_tokens_seen": 4850160, "step": 18105 }, { "epoch": 4.747051114023591, "grad_norm": 1.1903682947158813, "learning_rate": 3.1497538594578814e-05, "loss": 0.3418, "num_input_tokens_seen": 4851696, "step": 18110 }, { "epoch": 4.748361730013106, "grad_norm": 1.0869154930114746, "learning_rate": 3.1486493673691624e-05, "loss": 0.3486, "num_input_tokens_seen": 4853008, "step": 18115 }, { "epoch": 4.749672346002622, "grad_norm": 0.919447124004364, "learning_rate": 3.147544739519253e-05, "loss": 0.4115, "num_input_tokens_seen": 4854352, "step": 18120 }, { "epoch": 4.750982961992136, "grad_norm": 0.7950528264045715, "learning_rate": 3.1464399761393465e-05, "loss": 0.2497, "num_input_tokens_seen": 4855904, "step": 18125 }, { "epoch": 4.752293577981652, "grad_norm": 0.9616645574569702, "learning_rate": 3.145335077460671e-05, "loss": 0.3306, "num_input_tokens_seen": 4857536, "step": 18130 }, { "epoch": 4.753604193971166, "grad_norm": 1.4580256938934326, "learning_rate": 3.144230043714479e-05, "loss": 0.2525, "num_input_tokens_seen": 4858480, "step": 18135 }, { "epoch": 4.754914809960681, "grad_norm": 0.5954152941703796, "learning_rate": 3.143124875132052e-05, "loss": 0.246, "num_input_tokens_seen": 4859664, "step": 18140 }, { "epoch": 4.756225425950197, "grad_norm": 0.7615203261375427, "learning_rate": 3.142019571944699e-05, "loss": 0.2922, "num_input_tokens_seen": 4861472, "step": 18145 }, { "epoch": 4.757536041939712, "grad_norm": 3.659101724624634, "learning_rate": 3.140914134383759e-05, "loss": 0.4382, "num_input_tokens_seen": 4862592, "step": 18150 }, { "epoch": 4.758846657929227, "grad_norm": 1.2411603927612305, "learning_rate": 3.139808562680599e-05, "loss": 0.5929, "num_input_tokens_seen": 4863696, "step": 18155 }, { "epoch": 4.760157273918741, "grad_norm": 1.2910197973251343, "learning_rate": 3.138702857066612e-05, "loss": 0.2909, "num_input_tokens_seen": 4864768, "step": 18160 }, { "epoch": 4.761467889908257, "grad_norm": 0.7136368751525879, "learning_rate": 3.137597017773221e-05, "loss": 0.3293, "num_input_tokens_seen": 4866112, "step": 18165 }, { "epoch": 4.762778505897772, "grad_norm": 0.9191770553588867, "learning_rate": 3.136491045031875e-05, "loss": 0.2606, "num_input_tokens_seen": 4867456, "step": 18170 }, { "epoch": 4.764089121887287, "grad_norm": 13.249166488647461, "learning_rate": 3.135384939074054e-05, "loss": 0.4066, "num_input_tokens_seen": 4868592, "step": 18175 }, { "epoch": 4.765399737876802, "grad_norm": 6.395598888397217, "learning_rate": 3.1342787001312615e-05, "loss": 0.4004, "num_input_tokens_seen": 4869824, "step": 18180 }, { "epoch": 4.766710353866317, "grad_norm": 1.153091549873352, "learning_rate": 3.133172328435034e-05, "loss": 0.3174, "num_input_tokens_seen": 4871200, "step": 18185 }, { "epoch": 4.768020969855832, "grad_norm": 2.4372758865356445, "learning_rate": 3.132065824216932e-05, "loss": 0.3388, "num_input_tokens_seen": 4872224, "step": 18190 }, { "epoch": 4.7693315858453476, "grad_norm": 1.2384003400802612, "learning_rate": 3.1309591877085464e-05, "loss": 0.5409, "num_input_tokens_seen": 4873520, "step": 18195 }, { "epoch": 4.770642201834862, "grad_norm": 1.1120833158493042, "learning_rate": 3.1298524191414915e-05, "loss": 0.3464, "num_input_tokens_seen": 4874992, "step": 18200 }, { "epoch": 4.771952817824378, "grad_norm": 1.0420832633972168, "learning_rate": 3.128745518747414e-05, "loss": 0.2414, "num_input_tokens_seen": 4876112, "step": 18205 }, { "epoch": 4.773263433813892, "grad_norm": 1.5847417116165161, "learning_rate": 3.127638486757985e-05, "loss": 0.1936, "num_input_tokens_seen": 4877136, "step": 18210 }, { "epoch": 4.774574049803408, "grad_norm": 0.7616410255432129, "learning_rate": 3.126531323404907e-05, "loss": 0.2324, "num_input_tokens_seen": 4878848, "step": 18215 }, { "epoch": 4.775884665792923, "grad_norm": 0.8076962232589722, "learning_rate": 3.125424028919904e-05, "loss": 0.2518, "num_input_tokens_seen": 4880912, "step": 18220 }, { "epoch": 4.777195281782438, "grad_norm": 1.1223700046539307, "learning_rate": 3.1243166035347335e-05, "loss": 0.3917, "num_input_tokens_seen": 4882448, "step": 18225 }, { "epoch": 4.778505897771953, "grad_norm": 1.132076382637024, "learning_rate": 3.123209047481177e-05, "loss": 0.5161, "num_input_tokens_seen": 4883792, "step": 18230 }, { "epoch": 4.779816513761467, "grad_norm": 2.7057735919952393, "learning_rate": 3.1221013609910424e-05, "loss": 0.4325, "num_input_tokens_seen": 4884928, "step": 18235 }, { "epoch": 4.781127129750983, "grad_norm": 1.535140872001648, "learning_rate": 3.120993544296168e-05, "loss": 0.4393, "num_input_tokens_seen": 4886288, "step": 18240 }, { "epoch": 4.7824377457404985, "grad_norm": 0.9387999176979065, "learning_rate": 3.119885597628418e-05, "loss": 0.3582, "num_input_tokens_seen": 4887712, "step": 18245 }, { "epoch": 4.783748361730013, "grad_norm": 1.059517741203308, "learning_rate": 3.1187775212196835e-05, "loss": 0.2656, "num_input_tokens_seen": 4888880, "step": 18250 }, { "epoch": 4.785058977719528, "grad_norm": 0.8509910106658936, "learning_rate": 3.117669315301883e-05, "loss": 0.4074, "num_input_tokens_seen": 4890208, "step": 18255 }, { "epoch": 4.786369593709043, "grad_norm": 1.3930435180664062, "learning_rate": 3.116560980106962e-05, "loss": 0.2763, "num_input_tokens_seen": 4891440, "step": 18260 }, { "epoch": 4.787680209698558, "grad_norm": 1.0894118547439575, "learning_rate": 3.115452515866893e-05, "loss": 0.3058, "num_input_tokens_seen": 4892592, "step": 18265 }, { "epoch": 4.7889908256880735, "grad_norm": 0.7526819705963135, "learning_rate": 3.114343922813677e-05, "loss": 0.279, "num_input_tokens_seen": 4893856, "step": 18270 }, { "epoch": 4.790301441677588, "grad_norm": 0.8162720203399658, "learning_rate": 3.1132352011793384e-05, "loss": 0.3646, "num_input_tokens_seen": 4895120, "step": 18275 }, { "epoch": 4.791612057667104, "grad_norm": 1.6686255931854248, "learning_rate": 3.112126351195932e-05, "loss": 0.3036, "num_input_tokens_seen": 4896512, "step": 18280 }, { "epoch": 4.792922673656618, "grad_norm": 1.15106201171875, "learning_rate": 3.1110173730955374e-05, "loss": 0.4465, "num_input_tokens_seen": 4897824, "step": 18285 }, { "epoch": 4.794233289646134, "grad_norm": 1.2193071842193604, "learning_rate": 3.109908267110263e-05, "loss": 0.4173, "num_input_tokens_seen": 4899184, "step": 18290 }, { "epoch": 4.795543905635649, "grad_norm": 0.9079327583312988, "learning_rate": 3.108799033472242e-05, "loss": 0.3963, "num_input_tokens_seen": 4900976, "step": 18295 }, { "epoch": 4.796854521625164, "grad_norm": 1.109818696975708, "learning_rate": 3.107689672413636e-05, "loss": 0.2772, "num_input_tokens_seen": 4902480, "step": 18300 }, { "epoch": 4.798165137614679, "grad_norm": 0.7707464098930359, "learning_rate": 3.10658018416663e-05, "loss": 0.2277, "num_input_tokens_seen": 4903744, "step": 18305 }, { "epoch": 4.799475753604194, "grad_norm": 0.7357063293457031, "learning_rate": 3.1054705689634405e-05, "loss": 0.2634, "num_input_tokens_seen": 4905216, "step": 18310 }, { "epoch": 4.800786369593709, "grad_norm": 1.0819969177246094, "learning_rate": 3.1043608270363075e-05, "loss": 0.3033, "num_input_tokens_seen": 4906512, "step": 18315 }, { "epoch": 4.8020969855832245, "grad_norm": 0.977950930595398, "learning_rate": 3.103250958617497e-05, "loss": 0.291, "num_input_tokens_seen": 4907808, "step": 18320 }, { "epoch": 4.803407601572739, "grad_norm": 0.7060366868972778, "learning_rate": 3.102140963939303e-05, "loss": 0.4619, "num_input_tokens_seen": 4909728, "step": 18325 }, { "epoch": 4.804718217562254, "grad_norm": 0.7873033881187439, "learning_rate": 3.101030843234046e-05, "loss": 0.3318, "num_input_tokens_seen": 4910896, "step": 18330 }, { "epoch": 4.806028833551769, "grad_norm": 1.555127739906311, "learning_rate": 3.099920596734073e-05, "loss": 0.5063, "num_input_tokens_seen": 4912288, "step": 18335 }, { "epoch": 4.807339449541285, "grad_norm": 1.037789225578308, "learning_rate": 3.0988102246717544e-05, "loss": 0.2073, "num_input_tokens_seen": 4913312, "step": 18340 }, { "epoch": 4.8086500655307995, "grad_norm": 2.594956874847412, "learning_rate": 3.0976997272794906e-05, "loss": 0.2653, "num_input_tokens_seen": 4914544, "step": 18345 }, { "epoch": 4.809960681520314, "grad_norm": 1.6264424324035645, "learning_rate": 3.096589104789707e-05, "loss": 0.3135, "num_input_tokens_seen": 4915776, "step": 18350 }, { "epoch": 4.81127129750983, "grad_norm": 1.638982892036438, "learning_rate": 3.095478357434854e-05, "loss": 0.4779, "num_input_tokens_seen": 4917008, "step": 18355 }, { "epoch": 4.812581913499344, "grad_norm": 2.915095567703247, "learning_rate": 3.09436748544741e-05, "loss": 0.2487, "num_input_tokens_seen": 4918032, "step": 18360 }, { "epoch": 4.81389252948886, "grad_norm": 1.0382609367370605, "learning_rate": 3.093256489059878e-05, "loss": 0.3763, "num_input_tokens_seen": 4919568, "step": 18365 }, { "epoch": 4.815203145478375, "grad_norm": 2.647488594055176, "learning_rate": 3.092145368504789e-05, "loss": 0.3847, "num_input_tokens_seen": 4920912, "step": 18370 }, { "epoch": 4.81651376146789, "grad_norm": 1.4380463361740112, "learning_rate": 3.0910341240146966e-05, "loss": 0.3544, "num_input_tokens_seen": 4922160, "step": 18375 }, { "epoch": 4.817824377457405, "grad_norm": 1.6443504095077515, "learning_rate": 3.089922755822183e-05, "loss": 0.393, "num_input_tokens_seen": 4923472, "step": 18380 }, { "epoch": 4.81913499344692, "grad_norm": 0.8551878929138184, "learning_rate": 3.088811264159856e-05, "loss": 0.4524, "num_input_tokens_seen": 4924528, "step": 18385 }, { "epoch": 4.820445609436435, "grad_norm": 1.3523719310760498, "learning_rate": 3.08769964926035e-05, "loss": 0.3017, "num_input_tokens_seen": 4925936, "step": 18390 }, { "epoch": 4.8217562254259505, "grad_norm": 1.783098578453064, "learning_rate": 3.086587911356322e-05, "loss": 0.4879, "num_input_tokens_seen": 4927504, "step": 18395 }, { "epoch": 4.823066841415465, "grad_norm": 0.8783156275749207, "learning_rate": 3.085476050680458e-05, "loss": 0.3742, "num_input_tokens_seen": 4928800, "step": 18400 }, { "epoch": 4.824377457404981, "grad_norm": 1.648777723312378, "learning_rate": 3.0843640674654675e-05, "loss": 0.311, "num_input_tokens_seen": 4930480, "step": 18405 }, { "epoch": 4.825688073394495, "grad_norm": 0.5037420988082886, "learning_rate": 3.083251961944088e-05, "loss": 0.3366, "num_input_tokens_seen": 4931904, "step": 18410 }, { "epoch": 4.826998689384011, "grad_norm": 1.3155970573425293, "learning_rate": 3.08213973434908e-05, "loss": 0.3018, "num_input_tokens_seen": 4933728, "step": 18415 }, { "epoch": 4.8283093053735255, "grad_norm": 1.3327407836914062, "learning_rate": 3.081027384913232e-05, "loss": 0.5017, "num_input_tokens_seen": 4935264, "step": 18420 }, { "epoch": 4.82961992136304, "grad_norm": 2.257939100265503, "learning_rate": 3.079914913869356e-05, "loss": 0.4036, "num_input_tokens_seen": 4936736, "step": 18425 }, { "epoch": 4.830930537352556, "grad_norm": 1.6298880577087402, "learning_rate": 3.0788023214502915e-05, "loss": 0.3532, "num_input_tokens_seen": 4938448, "step": 18430 }, { "epoch": 4.832241153342071, "grad_norm": 2.0351030826568604, "learning_rate": 3.0776896078889e-05, "loss": 0.2807, "num_input_tokens_seen": 4939632, "step": 18435 }, { "epoch": 4.833551769331586, "grad_norm": 3.210517406463623, "learning_rate": 3.0765767734180714e-05, "loss": 0.312, "num_input_tokens_seen": 4940800, "step": 18440 }, { "epoch": 4.834862385321101, "grad_norm": 0.846977949142456, "learning_rate": 3.075463818270722e-05, "loss": 0.3374, "num_input_tokens_seen": 4942032, "step": 18445 }, { "epoch": 4.836173001310616, "grad_norm": 0.4473035931587219, "learning_rate": 3.074350742679789e-05, "loss": 0.2436, "num_input_tokens_seen": 4943312, "step": 18450 }, { "epoch": 4.837483617300131, "grad_norm": 2.3329074382781982, "learning_rate": 3.0732375468782384e-05, "loss": 0.452, "num_input_tokens_seen": 4944448, "step": 18455 }, { "epoch": 4.838794233289646, "grad_norm": 0.8379557728767395, "learning_rate": 3.072124231099058e-05, "loss": 0.2853, "num_input_tokens_seen": 4946000, "step": 18460 }, { "epoch": 4.840104849279161, "grad_norm": 0.9778307676315308, "learning_rate": 3.0710107955752664e-05, "loss": 0.327, "num_input_tokens_seen": 4947152, "step": 18465 }, { "epoch": 4.8414154652686765, "grad_norm": 0.8185500502586365, "learning_rate": 3.069897240539901e-05, "loss": 0.3097, "num_input_tokens_seen": 4948704, "step": 18470 }, { "epoch": 4.842726081258191, "grad_norm": 0.8154501914978027, "learning_rate": 3.068783566226028e-05, "loss": 0.3632, "num_input_tokens_seen": 4950096, "step": 18475 }, { "epoch": 4.844036697247707, "grad_norm": 1.0165022611618042, "learning_rate": 3.067669772866738e-05, "loss": 0.4068, "num_input_tokens_seen": 4951408, "step": 18480 }, { "epoch": 4.845347313237221, "grad_norm": 3.6908469200134277, "learning_rate": 3.066555860695145e-05, "loss": 0.5501, "num_input_tokens_seen": 4952352, "step": 18485 }, { "epoch": 4.846657929226737, "grad_norm": 2.464735984802246, "learning_rate": 3.065441829944389e-05, "loss": 0.4836, "num_input_tokens_seen": 4953696, "step": 18490 }, { "epoch": 4.8479685452162515, "grad_norm": 0.7682361006736755, "learning_rate": 3.064327680847635e-05, "loss": 0.2775, "num_input_tokens_seen": 4955376, "step": 18495 }, { "epoch": 4.849279161205767, "grad_norm": 1.329329013824463, "learning_rate": 3.063213413638073e-05, "loss": 0.3886, "num_input_tokens_seen": 4957024, "step": 18500 }, { "epoch": 4.850589777195282, "grad_norm": 1.020242691040039, "learning_rate": 3.0620990285489165e-05, "loss": 0.4394, "num_input_tokens_seen": 4958272, "step": 18505 }, { "epoch": 4.851900393184797, "grad_norm": 1.2942956686019897, "learning_rate": 3.0609845258134046e-05, "loss": 0.4015, "num_input_tokens_seen": 4959328, "step": 18510 }, { "epoch": 4.853211009174312, "grad_norm": 1.6885311603546143, "learning_rate": 3.0598699056648005e-05, "loss": 0.3374, "num_input_tokens_seen": 4960352, "step": 18515 }, { "epoch": 4.854521625163827, "grad_norm": 1.2706032991409302, "learning_rate": 3.0587551683363926e-05, "loss": 0.3046, "num_input_tokens_seen": 4961488, "step": 18520 }, { "epoch": 4.855832241153342, "grad_norm": 1.5603673458099365, "learning_rate": 3.057640314061493e-05, "loss": 0.3383, "num_input_tokens_seen": 4962816, "step": 18525 }, { "epoch": 4.857142857142857, "grad_norm": 0.7254924774169922, "learning_rate": 3.05652534307344e-05, "loss": 0.2499, "num_input_tokens_seen": 4964256, "step": 18530 }, { "epoch": 4.858453473132372, "grad_norm": 0.13568423688411713, "learning_rate": 3.055410255605594e-05, "loss": 0.3222, "num_input_tokens_seen": 4965680, "step": 18535 }, { "epoch": 4.859764089121887, "grad_norm": 4.321174621582031, "learning_rate": 3.054295051891341e-05, "loss": 0.2943, "num_input_tokens_seen": 4967216, "step": 18540 }, { "epoch": 4.8610747051114025, "grad_norm": 1.8483421802520752, "learning_rate": 3.053179732164092e-05, "loss": 0.2351, "num_input_tokens_seen": 4968496, "step": 18545 }, { "epoch": 4.862385321100917, "grad_norm": 4.474368572235107, "learning_rate": 3.0520642966572805e-05, "loss": 0.4113, "num_input_tokens_seen": 4970112, "step": 18550 }, { "epoch": 4.863695937090433, "grad_norm": 1.6773356199264526, "learning_rate": 3.050948745604365e-05, "loss": 0.3722, "num_input_tokens_seen": 4971568, "step": 18555 }, { "epoch": 4.865006553079947, "grad_norm": 2.754406690597534, "learning_rate": 3.049833079238829e-05, "loss": 0.6832, "num_input_tokens_seen": 4972704, "step": 18560 }, { "epoch": 4.866317169069463, "grad_norm": 1.4428521394729614, "learning_rate": 3.0487172977941806e-05, "loss": 0.2968, "num_input_tokens_seen": 4973936, "step": 18565 }, { "epoch": 4.8676277850589775, "grad_norm": 1.5865527391433716, "learning_rate": 3.0476014015039496e-05, "loss": 0.3564, "num_input_tokens_seen": 4975184, "step": 18570 }, { "epoch": 4.868938401048493, "grad_norm": 1.2789565324783325, "learning_rate": 3.046485390601691e-05, "loss": 0.2926, "num_input_tokens_seen": 4976304, "step": 18575 }, { "epoch": 4.870249017038008, "grad_norm": 0.7097507119178772, "learning_rate": 3.045369265320984e-05, "loss": 0.3781, "num_input_tokens_seen": 4977696, "step": 18580 }, { "epoch": 4.871559633027523, "grad_norm": 0.8739421963691711, "learning_rate": 3.0442530258954334e-05, "loss": 0.5125, "num_input_tokens_seen": 4979360, "step": 18585 }, { "epoch": 4.872870249017038, "grad_norm": 0.9244402647018433, "learning_rate": 3.0431366725586642e-05, "loss": 0.3162, "num_input_tokens_seen": 4980656, "step": 18590 }, { "epoch": 4.8741808650065535, "grad_norm": 0.9057483077049255, "learning_rate": 3.0420202055443276e-05, "loss": 0.2564, "num_input_tokens_seen": 4981920, "step": 18595 }, { "epoch": 4.875491480996068, "grad_norm": 2.1877009868621826, "learning_rate": 3.040903625086099e-05, "loss": 0.3418, "num_input_tokens_seen": 4983136, "step": 18600 }, { "epoch": 4.876802096985584, "grad_norm": 0.9826922416687012, "learning_rate": 3.0397869314176762e-05, "loss": 0.4861, "num_input_tokens_seen": 4984304, "step": 18605 }, { "epoch": 4.878112712975098, "grad_norm": 0.8205198049545288, "learning_rate": 3.0386701247727807e-05, "loss": 0.3493, "num_input_tokens_seen": 4985824, "step": 18610 }, { "epoch": 4.879423328964613, "grad_norm": 1.5266989469528198, "learning_rate": 3.0375532053851592e-05, "loss": 0.418, "num_input_tokens_seen": 4987280, "step": 18615 }, { "epoch": 4.8807339449541285, "grad_norm": 0.8170895576477051, "learning_rate": 3.036436173488581e-05, "loss": 0.2104, "num_input_tokens_seen": 4988272, "step": 18620 }, { "epoch": 4.882044560943643, "grad_norm": 1.8604276180267334, "learning_rate": 3.0353190293168388e-05, "loss": 0.2137, "num_input_tokens_seen": 4989392, "step": 18625 }, { "epoch": 4.883355176933159, "grad_norm": 0.7580130696296692, "learning_rate": 3.0342017731037482e-05, "loss": 0.3943, "num_input_tokens_seen": 4991296, "step": 18630 }, { "epoch": 4.884665792922673, "grad_norm": 0.8518902659416199, "learning_rate": 3.0330844050831492e-05, "loss": 0.3944, "num_input_tokens_seen": 4992720, "step": 18635 }, { "epoch": 4.885976408912189, "grad_norm": 1.3241097927093506, "learning_rate": 3.0319669254889055e-05, "loss": 0.3985, "num_input_tokens_seen": 4993856, "step": 18640 }, { "epoch": 4.8872870249017035, "grad_norm": 1.1530187129974365, "learning_rate": 3.030849334554904e-05, "loss": 0.3012, "num_input_tokens_seen": 4994864, "step": 18645 }, { "epoch": 4.888597640891219, "grad_norm": 1.2482197284698486, "learning_rate": 3.0297316325150537e-05, "loss": 0.2983, "num_input_tokens_seen": 4996048, "step": 18650 }, { "epoch": 4.889908256880734, "grad_norm": 1.563889503479004, "learning_rate": 3.0286138196032876e-05, "loss": 0.3628, "num_input_tokens_seen": 4997568, "step": 18655 }, { "epoch": 4.891218872870249, "grad_norm": 1.3659075498580933, "learning_rate": 3.0274958960535632e-05, "loss": 0.3845, "num_input_tokens_seen": 4998640, "step": 18660 }, { "epoch": 4.892529488859764, "grad_norm": 1.6598716974258423, "learning_rate": 3.0263778620998583e-05, "loss": 0.4795, "num_input_tokens_seen": 5000064, "step": 18665 }, { "epoch": 4.8938401048492794, "grad_norm": 1.13080894947052, "learning_rate": 3.025259717976177e-05, "loss": 0.1773, "num_input_tokens_seen": 5001152, "step": 18670 }, { "epoch": 4.895150720838794, "grad_norm": 1.1034263372421265, "learning_rate": 3.0241414639165444e-05, "loss": 0.3077, "num_input_tokens_seen": 5002480, "step": 18675 }, { "epoch": 4.89646133682831, "grad_norm": 2.067324638366699, "learning_rate": 3.0230231001550086e-05, "loss": 0.3177, "num_input_tokens_seen": 5003696, "step": 18680 }, { "epoch": 4.897771952817824, "grad_norm": 0.7730072736740112, "learning_rate": 3.0219046269256425e-05, "loss": 0.2628, "num_input_tokens_seen": 5005312, "step": 18685 }, { "epoch": 4.89908256880734, "grad_norm": 4.629087448120117, "learning_rate": 3.020786044462539e-05, "loss": 0.4097, "num_input_tokens_seen": 5006688, "step": 18690 }, { "epoch": 4.9003931847968545, "grad_norm": 1.0289722681045532, "learning_rate": 3.019667352999816e-05, "loss": 0.3898, "num_input_tokens_seen": 5008160, "step": 18695 }, { "epoch": 4.90170380078637, "grad_norm": 1.2344999313354492, "learning_rate": 3.018548552771615e-05, "loss": 0.4269, "num_input_tokens_seen": 5009744, "step": 18700 }, { "epoch": 4.903014416775885, "grad_norm": 0.6255630850791931, "learning_rate": 3.0174296440120968e-05, "loss": 0.267, "num_input_tokens_seen": 5010944, "step": 18705 }, { "epoch": 4.904325032765399, "grad_norm": 0.9877576231956482, "learning_rate": 3.0163106269554485e-05, "loss": 0.4268, "num_input_tokens_seen": 5012816, "step": 18710 }, { "epoch": 4.905635648754915, "grad_norm": 1.1958277225494385, "learning_rate": 3.0151915018358785e-05, "loss": 0.3724, "num_input_tokens_seen": 5013936, "step": 18715 }, { "epoch": 4.9069462647444295, "grad_norm": 1.1001087427139282, "learning_rate": 3.0140722688876166e-05, "loss": 0.3662, "num_input_tokens_seen": 5015344, "step": 18720 }, { "epoch": 4.908256880733945, "grad_norm": 0.7078176736831665, "learning_rate": 3.0129529283449177e-05, "loss": 0.2663, "num_input_tokens_seen": 5016544, "step": 18725 }, { "epoch": 4.90956749672346, "grad_norm": 1.1125634908676147, "learning_rate": 3.011833480442057e-05, "loss": 0.3739, "num_input_tokens_seen": 5018128, "step": 18730 }, { "epoch": 4.910878112712975, "grad_norm": 1.1210917234420776, "learning_rate": 3.0107139254133332e-05, "loss": 0.3561, "num_input_tokens_seen": 5019504, "step": 18735 }, { "epoch": 4.91218872870249, "grad_norm": 1.5429974794387817, "learning_rate": 3.009594263493068e-05, "loss": 0.2899, "num_input_tokens_seen": 5021040, "step": 18740 }, { "epoch": 4.913499344692005, "grad_norm": 1.8637839555740356, "learning_rate": 3.008474494915603e-05, "loss": 0.3204, "num_input_tokens_seen": 5022304, "step": 18745 }, { "epoch": 4.91480996068152, "grad_norm": 1.628784418106079, "learning_rate": 3.0073546199153053e-05, "loss": 0.3444, "num_input_tokens_seen": 5023808, "step": 18750 }, { "epoch": 4.916120576671036, "grad_norm": 1.5521663427352905, "learning_rate": 3.006234638726561e-05, "loss": 0.4593, "num_input_tokens_seen": 5025200, "step": 18755 }, { "epoch": 4.91743119266055, "grad_norm": 1.4065786600112915, "learning_rate": 3.005114551583783e-05, "loss": 0.3162, "num_input_tokens_seen": 5026656, "step": 18760 }, { "epoch": 4.918741808650066, "grad_norm": 1.0386676788330078, "learning_rate": 3.003994358721402e-05, "loss": 0.2834, "num_input_tokens_seen": 5027616, "step": 18765 }, { "epoch": 4.9200524246395805, "grad_norm": 8.241569519042969, "learning_rate": 3.002874060373872e-05, "loss": 0.6287, "num_input_tokens_seen": 5029584, "step": 18770 }, { "epoch": 4.921363040629096, "grad_norm": 0.8415012955665588, "learning_rate": 3.0017536567756705e-05, "loss": 0.2896, "num_input_tokens_seen": 5030672, "step": 18775 }, { "epoch": 4.922673656618611, "grad_norm": 0.6514761447906494, "learning_rate": 3.0006331481612955e-05, "loss": 0.2805, "num_input_tokens_seen": 5032352, "step": 18780 }, { "epoch": 4.923984272608125, "grad_norm": 0.9651486277580261, "learning_rate": 2.999512534765267e-05, "loss": 0.1847, "num_input_tokens_seen": 5033744, "step": 18785 }, { "epoch": 4.925294888597641, "grad_norm": 1.2363812923431396, "learning_rate": 2.9983918168221287e-05, "loss": 0.2825, "num_input_tokens_seen": 5034800, "step": 18790 }, { "epoch": 4.926605504587156, "grad_norm": 1.1375688314437866, "learning_rate": 2.997270994566444e-05, "loss": 0.3688, "num_input_tokens_seen": 5035952, "step": 18795 }, { "epoch": 4.927916120576671, "grad_norm": 0.6241050958633423, "learning_rate": 2.9961500682328e-05, "loss": 0.2477, "num_input_tokens_seen": 5037088, "step": 18800 }, { "epoch": 4.929226736566186, "grad_norm": 1.1478254795074463, "learning_rate": 2.995029038055804e-05, "loss": 0.6002, "num_input_tokens_seen": 5038624, "step": 18805 }, { "epoch": 4.930537352555701, "grad_norm": 0.9090135097503662, "learning_rate": 2.9939079042700847e-05, "loss": 0.3814, "num_input_tokens_seen": 5040416, "step": 18810 }, { "epoch": 4.931847968545216, "grad_norm": 1.2981040477752686, "learning_rate": 2.9927866671102946e-05, "loss": 0.3325, "num_input_tokens_seen": 5041696, "step": 18815 }, { "epoch": 4.933158584534731, "grad_norm": 2.3526346683502197, "learning_rate": 2.9916653268111074e-05, "loss": 0.2711, "num_input_tokens_seen": 5043168, "step": 18820 }, { "epoch": 4.934469200524246, "grad_norm": 0.7767801880836487, "learning_rate": 2.9905438836072165e-05, "loss": 0.4408, "num_input_tokens_seen": 5044736, "step": 18825 }, { "epoch": 4.935779816513762, "grad_norm": 1.204754114151001, "learning_rate": 2.989422337733338e-05, "loss": 0.4491, "num_input_tokens_seen": 5046256, "step": 18830 }, { "epoch": 4.937090432503276, "grad_norm": 1.6337769031524658, "learning_rate": 2.98830068942421e-05, "loss": 0.4885, "num_input_tokens_seen": 5047536, "step": 18835 }, { "epoch": 4.938401048492792, "grad_norm": 1.252376675605774, "learning_rate": 2.987178938914591e-05, "loss": 0.5233, "num_input_tokens_seen": 5048768, "step": 18840 }, { "epoch": 4.9397116644823065, "grad_norm": 1.7673293352127075, "learning_rate": 2.9860570864392625e-05, "loss": 0.3501, "num_input_tokens_seen": 5049760, "step": 18845 }, { "epoch": 4.941022280471822, "grad_norm": 3.090947151184082, "learning_rate": 2.9849351322330255e-05, "loss": 0.3596, "num_input_tokens_seen": 5050720, "step": 18850 }, { "epoch": 4.942332896461337, "grad_norm": 0.7884257435798645, "learning_rate": 2.983813076530703e-05, "loss": 0.316, "num_input_tokens_seen": 5051808, "step": 18855 }, { "epoch": 4.943643512450852, "grad_norm": 1.3672308921813965, "learning_rate": 2.9826909195671403e-05, "loss": 0.1575, "num_input_tokens_seen": 5053168, "step": 18860 }, { "epoch": 4.944954128440367, "grad_norm": 1.2516543865203857, "learning_rate": 2.981568661577202e-05, "loss": 0.4313, "num_input_tokens_seen": 5054368, "step": 18865 }, { "epoch": 4.946264744429882, "grad_norm": 1.7434855699539185, "learning_rate": 2.9804463027957736e-05, "loss": 0.5559, "num_input_tokens_seen": 5055600, "step": 18870 }, { "epoch": 4.947575360419397, "grad_norm": 0.9941691160202026, "learning_rate": 2.9793238434577652e-05, "loss": 0.2945, "num_input_tokens_seen": 5057040, "step": 18875 }, { "epoch": 4.948885976408912, "grad_norm": 1.2285586595535278, "learning_rate": 2.978201283798105e-05, "loss": 0.2298, "num_input_tokens_seen": 5059248, "step": 18880 }, { "epoch": 4.950196592398427, "grad_norm": 0.7933909893035889, "learning_rate": 2.977078624051742e-05, "loss": 0.3411, "num_input_tokens_seen": 5060640, "step": 18885 }, { "epoch": 4.951507208387943, "grad_norm": 1.724945306777954, "learning_rate": 2.9759558644536467e-05, "loss": 0.392, "num_input_tokens_seen": 5062032, "step": 18890 }, { "epoch": 4.952817824377457, "grad_norm": 1.678563117980957, "learning_rate": 2.9748330052388117e-05, "loss": 0.3462, "num_input_tokens_seen": 5063424, "step": 18895 }, { "epoch": 4.954128440366972, "grad_norm": 1.6706761121749878, "learning_rate": 2.9737100466422498e-05, "loss": 0.3279, "num_input_tokens_seen": 5064384, "step": 18900 }, { "epoch": 4.955439056356488, "grad_norm": 1.2831873893737793, "learning_rate": 2.9725869888989926e-05, "loss": 0.291, "num_input_tokens_seen": 5065728, "step": 18905 }, { "epoch": 4.956749672346002, "grad_norm": 1.65487802028656, "learning_rate": 2.9714638322440956e-05, "loss": 0.263, "num_input_tokens_seen": 5067408, "step": 18910 }, { "epoch": 4.958060288335518, "grad_norm": 1.6282583475112915, "learning_rate": 2.9703405769126334e-05, "loss": 0.3975, "num_input_tokens_seen": 5068592, "step": 18915 }, { "epoch": 4.9593709043250325, "grad_norm": 1.502694010734558, "learning_rate": 2.969217223139702e-05, "loss": 0.4318, "num_input_tokens_seen": 5070224, "step": 18920 }, { "epoch": 4.960681520314548, "grad_norm": 1.3327168226242065, "learning_rate": 2.9680937711604146e-05, "loss": 0.3383, "num_input_tokens_seen": 5071424, "step": 18925 }, { "epoch": 4.961992136304063, "grad_norm": 0.9034335613250732, "learning_rate": 2.9669702212099104e-05, "loss": 0.2528, "num_input_tokens_seen": 5072880, "step": 18930 }, { "epoch": 4.963302752293578, "grad_norm": 0.9951410889625549, "learning_rate": 2.9658465735233467e-05, "loss": 0.2958, "num_input_tokens_seen": 5074128, "step": 18935 }, { "epoch": 4.964613368283093, "grad_norm": 2.2347302436828613, "learning_rate": 2.9647228283358995e-05, "loss": 0.4595, "num_input_tokens_seen": 5075216, "step": 18940 }, { "epoch": 4.965923984272608, "grad_norm": 4.000086307525635, "learning_rate": 2.9635989858827673e-05, "loss": 0.4868, "num_input_tokens_seen": 5076368, "step": 18945 }, { "epoch": 4.967234600262123, "grad_norm": 2.796105146408081, "learning_rate": 2.962475046399168e-05, "loss": 0.5165, "num_input_tokens_seen": 5077760, "step": 18950 }, { "epoch": 4.968545216251639, "grad_norm": 1.0637303590774536, "learning_rate": 2.9613510101203406e-05, "loss": 0.3726, "num_input_tokens_seen": 5079728, "step": 18955 }, { "epoch": 4.969855832241153, "grad_norm": 3.0130796432495117, "learning_rate": 2.9602268772815445e-05, "loss": 0.3254, "num_input_tokens_seen": 5081232, "step": 18960 }, { "epoch": 4.971166448230669, "grad_norm": 1.439610481262207, "learning_rate": 2.959102648118058e-05, "loss": 0.5101, "num_input_tokens_seen": 5082672, "step": 18965 }, { "epoch": 4.972477064220183, "grad_norm": 1.4358810186386108, "learning_rate": 2.9579783228651802e-05, "loss": 0.3985, "num_input_tokens_seen": 5083824, "step": 18970 }, { "epoch": 4.973787680209698, "grad_norm": 1.8635120391845703, "learning_rate": 2.956853901758231e-05, "loss": 0.2942, "num_input_tokens_seen": 5084912, "step": 18975 }, { "epoch": 4.975098296199214, "grad_norm": 0.6217031478881836, "learning_rate": 2.9557293850325496e-05, "loss": 0.3482, "num_input_tokens_seen": 5086528, "step": 18980 }, { "epoch": 4.976408912188729, "grad_norm": 1.608855128288269, "learning_rate": 2.954604772923495e-05, "loss": 0.265, "num_input_tokens_seen": 5087856, "step": 18985 }, { "epoch": 4.977719528178244, "grad_norm": 1.336466908454895, "learning_rate": 2.9534800656664468e-05, "loss": 0.3828, "num_input_tokens_seen": 5089328, "step": 18990 }, { "epoch": 4.9790301441677585, "grad_norm": 0.9116246700286865, "learning_rate": 2.9523552634968056e-05, "loss": 0.2678, "num_input_tokens_seen": 5090400, "step": 18995 }, { "epoch": 4.980340760157274, "grad_norm": 1.1097005605697632, "learning_rate": 2.9512303666499886e-05, "loss": 0.275, "num_input_tokens_seen": 5091984, "step": 19000 }, { "epoch": 4.981651376146789, "grad_norm": 1.1491345167160034, "learning_rate": 2.9501053753614354e-05, "loss": 0.2939, "num_input_tokens_seen": 5093424, "step": 19005 }, { "epoch": 4.982961992136304, "grad_norm": 5.6867289543151855, "learning_rate": 2.9489802898666042e-05, "loss": 0.2823, "num_input_tokens_seen": 5094752, "step": 19010 }, { "epoch": 4.984272608125819, "grad_norm": 1.4166263341903687, "learning_rate": 2.9478551104009756e-05, "loss": 0.3132, "num_input_tokens_seen": 5096240, "step": 19015 }, { "epoch": 4.985583224115334, "grad_norm": 1.795531988143921, "learning_rate": 2.946729837200045e-05, "loss": 0.2549, "num_input_tokens_seen": 5098176, "step": 19020 }, { "epoch": 4.986893840104849, "grad_norm": 1.607467532157898, "learning_rate": 2.945604470499332e-05, "loss": 0.3538, "num_input_tokens_seen": 5099520, "step": 19025 }, { "epoch": 4.988204456094365, "grad_norm": 2.686338424682617, "learning_rate": 2.9444790105343735e-05, "loss": 0.3795, "num_input_tokens_seen": 5100688, "step": 19030 }, { "epoch": 4.989515072083879, "grad_norm": 1.9512938261032104, "learning_rate": 2.9433534575407257e-05, "loss": 0.323, "num_input_tokens_seen": 5101888, "step": 19035 }, { "epoch": 4.990825688073395, "grad_norm": 1.5468538999557495, "learning_rate": 2.942227811753965e-05, "loss": 0.3505, "num_input_tokens_seen": 5103264, "step": 19040 }, { "epoch": 4.992136304062909, "grad_norm": 0.7480335235595703, "learning_rate": 2.9411020734096874e-05, "loss": 0.3736, "num_input_tokens_seen": 5104976, "step": 19045 }, { "epoch": 4.993446920052425, "grad_norm": 1.02678644657135, "learning_rate": 2.9399762427435084e-05, "loss": 0.3018, "num_input_tokens_seen": 5106320, "step": 19050 }, { "epoch": 4.99475753604194, "grad_norm": 1.0783634185791016, "learning_rate": 2.938850319991062e-05, "loss": 0.3959, "num_input_tokens_seen": 5107744, "step": 19055 }, { "epoch": 4.996068152031455, "grad_norm": 0.9274812936782837, "learning_rate": 2.937724305388002e-05, "loss": 0.2759, "num_input_tokens_seen": 5109136, "step": 19060 }, { "epoch": 4.99737876802097, "grad_norm": 1.0421197414398193, "learning_rate": 2.936598199170001e-05, "loss": 0.2816, "num_input_tokens_seen": 5110544, "step": 19065 }, { "epoch": 4.9986893840104845, "grad_norm": 1.8542288541793823, "learning_rate": 2.9354720015727515e-05, "loss": 0.5024, "num_input_tokens_seen": 5111680, "step": 19070 }, { "epoch": 5.0, "grad_norm": 1.4108448028564453, "learning_rate": 2.934345712831964e-05, "loss": 0.2881, "num_input_tokens_seen": 5113312, "step": 19075 }, { "epoch": 5.001310615989515, "grad_norm": 1.403334617614746, "learning_rate": 2.9332193331833696e-05, "loss": 0.2889, "num_input_tokens_seen": 5114800, "step": 19080 }, { "epoch": 5.001310615989515, "eval_loss": 0.5386597514152527, "eval_runtime": 16.6954, "eval_samples_per_second": 50.792, "eval_steps_per_second": 25.396, "num_input_tokens_seen": 5114800, "step": 19080 }, { "epoch": 5.00262123197903, "grad_norm": 1.3693547248840332, "learning_rate": 2.9320928628627175e-05, "loss": 0.2316, "num_input_tokens_seen": 5116176, "step": 19085 }, { "epoch": 5.003931847968545, "grad_norm": 1.3085978031158447, "learning_rate": 2.9309663021057754e-05, "loss": 0.2678, "num_input_tokens_seen": 5117904, "step": 19090 }, { "epoch": 5.00524246395806, "grad_norm": 0.9885488748550415, "learning_rate": 2.929839651148331e-05, "loss": 0.2921, "num_input_tokens_seen": 5119520, "step": 19095 }, { "epoch": 5.006553079947575, "grad_norm": 2.1272263526916504, "learning_rate": 2.9287129102261907e-05, "loss": 0.2953, "num_input_tokens_seen": 5120688, "step": 19100 }, { "epoch": 5.007863695937091, "grad_norm": 1.5486340522766113, "learning_rate": 2.9275860795751786e-05, "loss": 0.4206, "num_input_tokens_seen": 5121936, "step": 19105 }, { "epoch": 5.009174311926605, "grad_norm": 3.501323938369751, "learning_rate": 2.926459159431139e-05, "loss": 0.2989, "num_input_tokens_seen": 5122896, "step": 19110 }, { "epoch": 5.010484927916121, "grad_norm": 0.8315783739089966, "learning_rate": 2.9253321500299354e-05, "loss": 0.3971, "num_input_tokens_seen": 5124256, "step": 19115 }, { "epoch": 5.011795543905635, "grad_norm": 2.1914496421813965, "learning_rate": 2.9242050516074466e-05, "loss": 0.3643, "num_input_tokens_seen": 5125696, "step": 19120 }, { "epoch": 5.013106159895151, "grad_norm": 1.0710703134536743, "learning_rate": 2.9230778643995726e-05, "loss": 0.3661, "num_input_tokens_seen": 5126960, "step": 19125 }, { "epoch": 5.014416775884666, "grad_norm": 1.393040418624878, "learning_rate": 2.9219505886422338e-05, "loss": 0.3219, "num_input_tokens_seen": 5128336, "step": 19130 }, { "epoch": 5.015727391874181, "grad_norm": 2.047931671142578, "learning_rate": 2.9208232245713657e-05, "loss": 0.4305, "num_input_tokens_seen": 5129984, "step": 19135 }, { "epoch": 5.017038007863696, "grad_norm": 1.3057986497879028, "learning_rate": 2.9196957724229234e-05, "loss": 0.2253, "num_input_tokens_seen": 5131120, "step": 19140 }, { "epoch": 5.018348623853211, "grad_norm": 0.9187046885490417, "learning_rate": 2.9185682324328815e-05, "loss": 0.2473, "num_input_tokens_seen": 5132352, "step": 19145 }, { "epoch": 5.019659239842726, "grad_norm": 0.8415477871894836, "learning_rate": 2.9174406048372316e-05, "loss": 0.6009, "num_input_tokens_seen": 5134048, "step": 19150 }, { "epoch": 5.0209698558322415, "grad_norm": 1.4099576473236084, "learning_rate": 2.9163128898719845e-05, "loss": 0.2826, "num_input_tokens_seen": 5135600, "step": 19155 }, { "epoch": 5.022280471821756, "grad_norm": 1.2961570024490356, "learning_rate": 2.915185087773169e-05, "loss": 0.2787, "num_input_tokens_seen": 5136784, "step": 19160 }, { "epoch": 5.023591087811272, "grad_norm": 1.4560167789459229, "learning_rate": 2.9140571987768316e-05, "loss": 0.3424, "num_input_tokens_seen": 5138656, "step": 19165 }, { "epoch": 5.024901703800786, "grad_norm": 1.43406081199646, "learning_rate": 2.9129292231190386e-05, "loss": 0.2634, "num_input_tokens_seen": 5140256, "step": 19170 }, { "epoch": 5.026212319790301, "grad_norm": 0.7970129251480103, "learning_rate": 2.9118011610358726e-05, "loss": 0.3177, "num_input_tokens_seen": 5141648, "step": 19175 }, { "epoch": 5.027522935779817, "grad_norm": 2.116795539855957, "learning_rate": 2.9106730127634346e-05, "loss": 0.386, "num_input_tokens_seen": 5142848, "step": 19180 }, { "epoch": 5.028833551769331, "grad_norm": 0.9179461002349854, "learning_rate": 2.9095447785378443e-05, "loss": 0.242, "num_input_tokens_seen": 5144128, "step": 19185 }, { "epoch": 5.030144167758847, "grad_norm": 1.317246437072754, "learning_rate": 2.9084164585952405e-05, "loss": 0.1398, "num_input_tokens_seen": 5145808, "step": 19190 }, { "epoch": 5.031454783748361, "grad_norm": 0.7756890058517456, "learning_rate": 2.9072880531717773e-05, "loss": 0.1921, "num_input_tokens_seen": 5147728, "step": 19195 }, { "epoch": 5.032765399737877, "grad_norm": 1.0747019052505493, "learning_rate": 2.9061595625036277e-05, "loss": 0.289, "num_input_tokens_seen": 5149200, "step": 19200 }, { "epoch": 5.034076015727392, "grad_norm": 2.401092529296875, "learning_rate": 2.905030986826984e-05, "loss": 0.3385, "num_input_tokens_seen": 5150336, "step": 19205 }, { "epoch": 5.035386631716907, "grad_norm": 0.7224874496459961, "learning_rate": 2.903902326378054e-05, "loss": 0.1796, "num_input_tokens_seen": 5151344, "step": 19210 }, { "epoch": 5.036697247706422, "grad_norm": 1.368513822555542, "learning_rate": 2.9027735813930658e-05, "loss": 0.2426, "num_input_tokens_seen": 5152944, "step": 19215 }, { "epoch": 5.038007863695937, "grad_norm": 1.5318354368209839, "learning_rate": 2.9016447521082625e-05, "loss": 0.4687, "num_input_tokens_seen": 5154384, "step": 19220 }, { "epoch": 5.039318479685452, "grad_norm": 1.5467764139175415, "learning_rate": 2.9005158387599068e-05, "loss": 0.2264, "num_input_tokens_seen": 5155632, "step": 19225 }, { "epoch": 5.0406290956749675, "grad_norm": 6.953861713409424, "learning_rate": 2.8993868415842785e-05, "loss": 0.5888, "num_input_tokens_seen": 5156992, "step": 19230 }, { "epoch": 5.041939711664482, "grad_norm": 1.502655267715454, "learning_rate": 2.8982577608176737e-05, "loss": 0.2733, "num_input_tokens_seen": 5158528, "step": 19235 }, { "epoch": 5.043250327653998, "grad_norm": 1.510900616645813, "learning_rate": 2.8971285966964075e-05, "loss": 0.3761, "num_input_tokens_seen": 5160208, "step": 19240 }, { "epoch": 5.044560943643512, "grad_norm": 1.2023422718048096, "learning_rate": 2.8959993494568123e-05, "loss": 0.3178, "num_input_tokens_seen": 5161536, "step": 19245 }, { "epoch": 5.045871559633028, "grad_norm": 1.5850929021835327, "learning_rate": 2.894870019335238e-05, "loss": 0.3348, "num_input_tokens_seen": 5164688, "step": 19250 }, { "epoch": 5.047182175622543, "grad_norm": 1.575480580329895, "learning_rate": 2.8937406065680505e-05, "loss": 0.312, "num_input_tokens_seen": 5166160, "step": 19255 }, { "epoch": 5.048492791612058, "grad_norm": 1.410156488418579, "learning_rate": 2.8926111113916344e-05, "loss": 0.2625, "num_input_tokens_seen": 5167232, "step": 19260 }, { "epoch": 5.049803407601573, "grad_norm": 1.4302335977554321, "learning_rate": 2.891481534042391e-05, "loss": 0.3221, "num_input_tokens_seen": 5168288, "step": 19265 }, { "epoch": 5.051114023591087, "grad_norm": 1.4506804943084717, "learning_rate": 2.8903518747567394e-05, "loss": 0.2394, "num_input_tokens_seen": 5169344, "step": 19270 }, { "epoch": 5.052424639580603, "grad_norm": 1.7831027507781982, "learning_rate": 2.8892221337711143e-05, "loss": 0.2414, "num_input_tokens_seen": 5170288, "step": 19275 }, { "epoch": 5.053735255570118, "grad_norm": 0.9683497548103333, "learning_rate": 2.88809231132197e-05, "loss": 0.277, "num_input_tokens_seen": 5171568, "step": 19280 }, { "epoch": 5.055045871559633, "grad_norm": 2.71570086479187, "learning_rate": 2.8869624076457745e-05, "loss": 0.2944, "num_input_tokens_seen": 5172832, "step": 19285 }, { "epoch": 5.056356487549148, "grad_norm": 1.4018663167953491, "learning_rate": 2.885832422979017e-05, "loss": 0.3159, "num_input_tokens_seen": 5174400, "step": 19290 }, { "epoch": 5.057667103538663, "grad_norm": 0.9125452041625977, "learning_rate": 2.8847023575581988e-05, "loss": 0.2308, "num_input_tokens_seen": 5175456, "step": 19295 }, { "epoch": 5.058977719528178, "grad_norm": 1.6285388469696045, "learning_rate": 2.8835722116198415e-05, "loss": 0.2561, "num_input_tokens_seen": 5176672, "step": 19300 }, { "epoch": 5.0602883355176935, "grad_norm": 0.8585681319236755, "learning_rate": 2.882441985400484e-05, "loss": 0.2202, "num_input_tokens_seen": 5177728, "step": 19305 }, { "epoch": 5.061598951507208, "grad_norm": 1.123124122619629, "learning_rate": 2.88131167913668e-05, "loss": 0.2919, "num_input_tokens_seen": 5178992, "step": 19310 }, { "epoch": 5.062909567496724, "grad_norm": 1.1969578266143799, "learning_rate": 2.8801812930650002e-05, "loss": 0.2116, "num_input_tokens_seen": 5180832, "step": 19315 }, { "epoch": 5.064220183486238, "grad_norm": 1.7442402839660645, "learning_rate": 2.8790508274220323e-05, "loss": 0.5522, "num_input_tokens_seen": 5182176, "step": 19320 }, { "epoch": 5.065530799475754, "grad_norm": 1.1408292055130005, "learning_rate": 2.877920282444381e-05, "loss": 0.3332, "num_input_tokens_seen": 5183424, "step": 19325 }, { "epoch": 5.066841415465269, "grad_norm": 2.6194612979888916, "learning_rate": 2.8767896583686682e-05, "loss": 0.5339, "num_input_tokens_seen": 5184912, "step": 19330 }, { "epoch": 5.068152031454784, "grad_norm": 1.0523053407669067, "learning_rate": 2.875658955431531e-05, "loss": 0.2548, "num_input_tokens_seen": 5186512, "step": 19335 }, { "epoch": 5.069462647444299, "grad_norm": 0.7993674278259277, "learning_rate": 2.874528173869623e-05, "loss": 0.2014, "num_input_tokens_seen": 5187984, "step": 19340 }, { "epoch": 5.070773263433814, "grad_norm": 0.7222397327423096, "learning_rate": 2.873397313919616e-05, "loss": 0.1999, "num_input_tokens_seen": 5189296, "step": 19345 }, { "epoch": 5.072083879423329, "grad_norm": 1.3827570676803589, "learning_rate": 2.872266375818196e-05, "loss": 0.2489, "num_input_tokens_seen": 5190400, "step": 19350 }, { "epoch": 5.073394495412844, "grad_norm": 1.0401521921157837, "learning_rate": 2.8711353598020662e-05, "loss": 0.2259, "num_input_tokens_seen": 5191408, "step": 19355 }, { "epoch": 5.074705111402359, "grad_norm": 2.300774574279785, "learning_rate": 2.8700042661079472e-05, "loss": 0.1964, "num_input_tokens_seen": 5192672, "step": 19360 }, { "epoch": 5.076015727391874, "grad_norm": 1.0454461574554443, "learning_rate": 2.8688730949725752e-05, "loss": 0.2951, "num_input_tokens_seen": 5194112, "step": 19365 }, { "epoch": 5.077326343381389, "grad_norm": 1.2965301275253296, "learning_rate": 2.867741846632701e-05, "loss": 0.2152, "num_input_tokens_seen": 5195264, "step": 19370 }, { "epoch": 5.078636959370904, "grad_norm": 1.523871898651123, "learning_rate": 2.8666105213250943e-05, "loss": 0.2743, "num_input_tokens_seen": 5197040, "step": 19375 }, { "epoch": 5.0799475753604195, "grad_norm": 2.868211269378662, "learning_rate": 2.8654791192865388e-05, "loss": 0.2824, "num_input_tokens_seen": 5198160, "step": 19380 }, { "epoch": 5.081258191349934, "grad_norm": 2.2424676418304443, "learning_rate": 2.864347640753835e-05, "loss": 0.3037, "num_input_tokens_seen": 5199120, "step": 19385 }, { "epoch": 5.08256880733945, "grad_norm": 1.8733497858047485, "learning_rate": 2.8632160859637998e-05, "loss": 0.2769, "num_input_tokens_seen": 5200544, "step": 19390 }, { "epoch": 5.083879423328964, "grad_norm": 2.090456247329712, "learning_rate": 2.8620844551532648e-05, "loss": 0.2882, "num_input_tokens_seen": 5201840, "step": 19395 }, { "epoch": 5.08519003931848, "grad_norm": 0.9077737927436829, "learning_rate": 2.8609527485590794e-05, "loss": 0.2956, "num_input_tokens_seen": 5203104, "step": 19400 }, { "epoch": 5.086500655307995, "grad_norm": 1.3665276765823364, "learning_rate": 2.8598209664181073e-05, "loss": 0.3672, "num_input_tokens_seen": 5204352, "step": 19405 }, { "epoch": 5.08781127129751, "grad_norm": 1.7192860841751099, "learning_rate": 2.858689108967229e-05, "loss": 0.2081, "num_input_tokens_seen": 5205920, "step": 19410 }, { "epoch": 5.089121887287025, "grad_norm": 1.2511093616485596, "learning_rate": 2.8575571764433396e-05, "loss": 0.3415, "num_input_tokens_seen": 5206992, "step": 19415 }, { "epoch": 5.09043250327654, "grad_norm": 2.3576228618621826, "learning_rate": 2.8564251690833505e-05, "loss": 0.2999, "num_input_tokens_seen": 5208400, "step": 19420 }, { "epoch": 5.091743119266055, "grad_norm": 1.8969295024871826, "learning_rate": 2.8552930871241906e-05, "loss": 0.4349, "num_input_tokens_seen": 5209680, "step": 19425 }, { "epoch": 5.0930537352555705, "grad_norm": 1.6432344913482666, "learning_rate": 2.8541609308028006e-05, "loss": 0.375, "num_input_tokens_seen": 5210832, "step": 19430 }, { "epoch": 5.094364351245085, "grad_norm": 1.9410762786865234, "learning_rate": 2.8530287003561402e-05, "loss": 0.2228, "num_input_tokens_seen": 5212160, "step": 19435 }, { "epoch": 5.095674967234601, "grad_norm": 1.2009270191192627, "learning_rate": 2.8518963960211814e-05, "loss": 0.3144, "num_input_tokens_seen": 5213728, "step": 19440 }, { "epoch": 5.096985583224115, "grad_norm": 0.7226784229278564, "learning_rate": 2.850764018034917e-05, "loss": 0.3046, "num_input_tokens_seen": 5215536, "step": 19445 }, { "epoch": 5.09829619921363, "grad_norm": 1.2037876844406128, "learning_rate": 2.8496315666343487e-05, "loss": 0.2775, "num_input_tokens_seen": 5216704, "step": 19450 }, { "epoch": 5.0996068152031455, "grad_norm": 4.002134799957275, "learning_rate": 2.8484990420564972e-05, "loss": 0.2479, "num_input_tokens_seen": 5218016, "step": 19455 }, { "epoch": 5.10091743119266, "grad_norm": 1.8861137628555298, "learning_rate": 2.847366444538399e-05, "loss": 0.2163, "num_input_tokens_seen": 5218960, "step": 19460 }, { "epoch": 5.102228047182176, "grad_norm": 5.494444847106934, "learning_rate": 2.8462337743171046e-05, "loss": 0.334, "num_input_tokens_seen": 5220352, "step": 19465 }, { "epoch": 5.10353866317169, "grad_norm": 1.0145267248153687, "learning_rate": 2.8451010316296783e-05, "loss": 0.2206, "num_input_tokens_seen": 5221776, "step": 19470 }, { "epoch": 5.104849279161206, "grad_norm": 3.1173887252807617, "learning_rate": 2.8439682167132027e-05, "loss": 0.5212, "num_input_tokens_seen": 5222944, "step": 19475 }, { "epoch": 5.1061598951507206, "grad_norm": 3.206892967224121, "learning_rate": 2.8428353298047738e-05, "loss": 0.2464, "num_input_tokens_seen": 5223904, "step": 19480 }, { "epoch": 5.107470511140236, "grad_norm": 2.0012872219085693, "learning_rate": 2.8417023711415036e-05, "loss": 0.2432, "num_input_tokens_seen": 5225408, "step": 19485 }, { "epoch": 5.108781127129751, "grad_norm": 0.7162761688232422, "learning_rate": 2.8405693409605166e-05, "loss": 0.2865, "num_input_tokens_seen": 5226368, "step": 19490 }, { "epoch": 5.110091743119266, "grad_norm": 1.6161377429962158, "learning_rate": 2.839436239498955e-05, "loss": 0.4766, "num_input_tokens_seen": 5227696, "step": 19495 }, { "epoch": 5.111402359108781, "grad_norm": 0.7055450081825256, "learning_rate": 2.8383030669939752e-05, "loss": 0.2679, "num_input_tokens_seen": 5229264, "step": 19500 }, { "epoch": 5.1127129750982965, "grad_norm": 1.9027858972549438, "learning_rate": 2.837169823682748e-05, "loss": 0.2702, "num_input_tokens_seen": 5230528, "step": 19505 }, { "epoch": 5.114023591087811, "grad_norm": 2.0162174701690674, "learning_rate": 2.8360365098024593e-05, "loss": 0.3002, "num_input_tokens_seen": 5231840, "step": 19510 }, { "epoch": 5.115334207077327, "grad_norm": 1.3012276887893677, "learning_rate": 2.8349031255903098e-05, "loss": 0.2366, "num_input_tokens_seen": 5232912, "step": 19515 }, { "epoch": 5.116644823066841, "grad_norm": 1.2676416635513306, "learning_rate": 2.8337696712835143e-05, "loss": 0.2951, "num_input_tokens_seen": 5234752, "step": 19520 }, { "epoch": 5.117955439056357, "grad_norm": 1.5724029541015625, "learning_rate": 2.8326361471193035e-05, "loss": 0.3952, "num_input_tokens_seen": 5235968, "step": 19525 }, { "epoch": 5.1192660550458715, "grad_norm": 1.5424597263336182, "learning_rate": 2.831502553334922e-05, "loss": 0.2139, "num_input_tokens_seen": 5237168, "step": 19530 }, { "epoch": 5.120576671035387, "grad_norm": 1.8203363418579102, "learning_rate": 2.8303688901676288e-05, "loss": 0.2457, "num_input_tokens_seen": 5238304, "step": 19535 }, { "epoch": 5.121887287024902, "grad_norm": 1.735131025314331, "learning_rate": 2.8292351578546982e-05, "loss": 0.3116, "num_input_tokens_seen": 5239840, "step": 19540 }, { "epoch": 5.123197903014416, "grad_norm": 1.0465484857559204, "learning_rate": 2.8281013566334182e-05, "loss": 0.2702, "num_input_tokens_seen": 5241088, "step": 19545 }, { "epoch": 5.124508519003932, "grad_norm": 1.5649826526641846, "learning_rate": 2.8269674867410907e-05, "loss": 0.2572, "num_input_tokens_seen": 5242208, "step": 19550 }, { "epoch": 5.1258191349934465, "grad_norm": 1.06915283203125, "learning_rate": 2.8258335484150323e-05, "loss": 0.189, "num_input_tokens_seen": 5243552, "step": 19555 }, { "epoch": 5.127129750982962, "grad_norm": 1.1181913614273071, "learning_rate": 2.8246995418925764e-05, "loss": 0.2728, "num_input_tokens_seen": 5244800, "step": 19560 }, { "epoch": 5.128440366972477, "grad_norm": 1.6085069179534912, "learning_rate": 2.823565467411067e-05, "loss": 0.3277, "num_input_tokens_seen": 5246400, "step": 19565 }, { "epoch": 5.129750982961992, "grad_norm": 1.4694204330444336, "learning_rate": 2.8224313252078637e-05, "loss": 0.2565, "num_input_tokens_seen": 5247600, "step": 19570 }, { "epoch": 5.131061598951507, "grad_norm": 1.069425344467163, "learning_rate": 2.8212971155203422e-05, "loss": 0.3036, "num_input_tokens_seen": 5248976, "step": 19575 }, { "epoch": 5.1323722149410225, "grad_norm": 2.603379249572754, "learning_rate": 2.820162838585889e-05, "loss": 0.3316, "num_input_tokens_seen": 5249872, "step": 19580 }, { "epoch": 5.133682830930537, "grad_norm": 1.5016220808029175, "learning_rate": 2.819028494641907e-05, "loss": 0.3685, "num_input_tokens_seen": 5251376, "step": 19585 }, { "epoch": 5.134993446920053, "grad_norm": 1.488887071609497, "learning_rate": 2.817894083925812e-05, "loss": 0.36, "num_input_tokens_seen": 5252672, "step": 19590 }, { "epoch": 5.136304062909567, "grad_norm": 0.9939960241317749, "learning_rate": 2.816759606675035e-05, "loss": 0.2547, "num_input_tokens_seen": 5253984, "step": 19595 }, { "epoch": 5.137614678899083, "grad_norm": 1.7712420225143433, "learning_rate": 2.81562506312702e-05, "loss": 0.379, "num_input_tokens_seen": 5255552, "step": 19600 }, { "epoch": 5.1389252948885975, "grad_norm": 1.641875982284546, "learning_rate": 2.814490453519224e-05, "loss": 0.2699, "num_input_tokens_seen": 5256592, "step": 19605 }, { "epoch": 5.140235910878113, "grad_norm": 1.673984169960022, "learning_rate": 2.8133557780891194e-05, "loss": 0.3391, "num_input_tokens_seen": 5257728, "step": 19610 }, { "epoch": 5.141546526867628, "grad_norm": 1.4041446447372437, "learning_rate": 2.812221037074192e-05, "loss": 0.2034, "num_input_tokens_seen": 5258816, "step": 19615 }, { "epoch": 5.142857142857143, "grad_norm": 2.2137796878814697, "learning_rate": 2.8110862307119413e-05, "loss": 0.2093, "num_input_tokens_seen": 5259904, "step": 19620 }, { "epoch": 5.144167758846658, "grad_norm": 0.8885729908943176, "learning_rate": 2.80995135923988e-05, "loss": 0.2522, "num_input_tokens_seen": 5261216, "step": 19625 }, { "epoch": 5.145478374836173, "grad_norm": 1.8608922958374023, "learning_rate": 2.8088164228955345e-05, "loss": 0.1998, "num_input_tokens_seen": 5262400, "step": 19630 }, { "epoch": 5.146788990825688, "grad_norm": 1.5705660581588745, "learning_rate": 2.8076814219164456e-05, "loss": 0.2542, "num_input_tokens_seen": 5263280, "step": 19635 }, { "epoch": 5.148099606815203, "grad_norm": 1.1709566116333008, "learning_rate": 2.806546356540166e-05, "loss": 0.4867, "num_input_tokens_seen": 5264688, "step": 19640 }, { "epoch": 5.149410222804718, "grad_norm": 1.173229694366455, "learning_rate": 2.8054112270042644e-05, "loss": 0.3102, "num_input_tokens_seen": 5265952, "step": 19645 }, { "epoch": 5.150720838794233, "grad_norm": 0.9713665246963501, "learning_rate": 2.8042760335463203e-05, "loss": 0.1873, "num_input_tokens_seen": 5267680, "step": 19650 }, { "epoch": 5.1520314547837485, "grad_norm": 1.5309196710586548, "learning_rate": 2.8031407764039287e-05, "loss": 0.2885, "num_input_tokens_seen": 5268944, "step": 19655 }, { "epoch": 5.153342070773263, "grad_norm": 1.0021322965621948, "learning_rate": 2.8020054558146964e-05, "loss": 0.2362, "num_input_tokens_seen": 5270096, "step": 19660 }, { "epoch": 5.154652686762779, "grad_norm": 1.1441389322280884, "learning_rate": 2.8008700720162433e-05, "loss": 0.5799, "num_input_tokens_seen": 5271024, "step": 19665 }, { "epoch": 5.155963302752293, "grad_norm": 1.4081135988235474, "learning_rate": 2.7997346252462035e-05, "loss": 0.3302, "num_input_tokens_seen": 5271984, "step": 19670 }, { "epoch": 5.157273918741809, "grad_norm": 1.7042102813720703, "learning_rate": 2.7985991157422254e-05, "loss": 0.2302, "num_input_tokens_seen": 5273456, "step": 19675 }, { "epoch": 5.1585845347313235, "grad_norm": 0.7274054884910583, "learning_rate": 2.797463543741969e-05, "loss": 0.2054, "num_input_tokens_seen": 5274720, "step": 19680 }, { "epoch": 5.159895150720839, "grad_norm": 1.6302541494369507, "learning_rate": 2.7963279094831057e-05, "loss": 0.3043, "num_input_tokens_seen": 5275840, "step": 19685 }, { "epoch": 5.161205766710354, "grad_norm": 2.8614542484283447, "learning_rate": 2.7951922132033235e-05, "loss": 0.2413, "num_input_tokens_seen": 5277344, "step": 19690 }, { "epoch": 5.162516382699869, "grad_norm": 1.5475577116012573, "learning_rate": 2.794056455140321e-05, "loss": 0.3186, "num_input_tokens_seen": 5279568, "step": 19695 }, { "epoch": 5.163826998689384, "grad_norm": 2.225407123565674, "learning_rate": 2.7929206355318104e-05, "loss": 0.3158, "num_input_tokens_seen": 5281120, "step": 19700 }, { "epoch": 5.165137614678899, "grad_norm": 1.6106297969818115, "learning_rate": 2.791784754615517e-05, "loss": 0.4631, "num_input_tokens_seen": 5282816, "step": 19705 }, { "epoch": 5.166448230668414, "grad_norm": 0.8583155274391174, "learning_rate": 2.7906488126291784e-05, "loss": 0.2211, "num_input_tokens_seen": 5284144, "step": 19710 }, { "epoch": 5.16775884665793, "grad_norm": 1.016308307647705, "learning_rate": 2.789512809810546e-05, "loss": 0.2545, "num_input_tokens_seen": 5285616, "step": 19715 }, { "epoch": 5.169069462647444, "grad_norm": 2.645247459411621, "learning_rate": 2.7883767463973832e-05, "loss": 0.3091, "num_input_tokens_seen": 5286720, "step": 19720 }, { "epoch": 5.17038007863696, "grad_norm": 1.5523228645324707, "learning_rate": 2.787240622627465e-05, "loss": 0.4464, "num_input_tokens_seen": 5287936, "step": 19725 }, { "epoch": 5.1716906946264745, "grad_norm": 0.9644277691841125, "learning_rate": 2.78610443873858e-05, "loss": 0.2992, "num_input_tokens_seen": 5289152, "step": 19730 }, { "epoch": 5.173001310615989, "grad_norm": 1.0235252380371094, "learning_rate": 2.7849681949685314e-05, "loss": 0.198, "num_input_tokens_seen": 5290160, "step": 19735 }, { "epoch": 5.174311926605505, "grad_norm": 0.5527048707008362, "learning_rate": 2.783831891555132e-05, "loss": 0.2164, "num_input_tokens_seen": 5291760, "step": 19740 }, { "epoch": 5.175622542595019, "grad_norm": 13.102550506591797, "learning_rate": 2.7826955287362077e-05, "loss": 0.4592, "num_input_tokens_seen": 5293024, "step": 19745 }, { "epoch": 5.176933158584535, "grad_norm": 0.6751660704612732, "learning_rate": 2.7815591067495977e-05, "loss": 0.1647, "num_input_tokens_seen": 5293920, "step": 19750 }, { "epoch": 5.1782437745740495, "grad_norm": 1.5503387451171875, "learning_rate": 2.7804226258331535e-05, "loss": 0.2918, "num_input_tokens_seen": 5295200, "step": 19755 }, { "epoch": 5.179554390563565, "grad_norm": 0.912816047668457, "learning_rate": 2.7792860862247377e-05, "loss": 0.4091, "num_input_tokens_seen": 5297088, "step": 19760 }, { "epoch": 5.18086500655308, "grad_norm": 1.6752779483795166, "learning_rate": 2.7781494881622268e-05, "loss": 0.2992, "num_input_tokens_seen": 5298480, "step": 19765 }, { "epoch": 5.182175622542595, "grad_norm": 1.8707168102264404, "learning_rate": 2.7770128318835087e-05, "loss": 0.3015, "num_input_tokens_seen": 5299648, "step": 19770 }, { "epoch": 5.18348623853211, "grad_norm": 1.0149427652359009, "learning_rate": 2.7758761176264835e-05, "loss": 0.339, "num_input_tokens_seen": 5300752, "step": 19775 }, { "epoch": 5.184796854521625, "grad_norm": 4.564137935638428, "learning_rate": 2.7747393456290644e-05, "loss": 0.3643, "num_input_tokens_seen": 5302384, "step": 19780 }, { "epoch": 5.18610747051114, "grad_norm": 2.428157329559326, "learning_rate": 2.7736025161291734e-05, "loss": 0.2658, "num_input_tokens_seen": 5303584, "step": 19785 }, { "epoch": 5.187418086500656, "grad_norm": 1.7014962434768677, "learning_rate": 2.7724656293647494e-05, "loss": 0.2433, "num_input_tokens_seen": 5305072, "step": 19790 }, { "epoch": 5.18872870249017, "grad_norm": 0.7357364296913147, "learning_rate": 2.7713286855737403e-05, "loss": 0.27, "num_input_tokens_seen": 5306368, "step": 19795 }, { "epoch": 5.190039318479686, "grad_norm": 1.0679035186767578, "learning_rate": 2.7701916849941056e-05, "loss": 0.1978, "num_input_tokens_seen": 5307600, "step": 19800 }, { "epoch": 5.1913499344692005, "grad_norm": 0.7673362493515015, "learning_rate": 2.7690546278638184e-05, "loss": 0.3684, "num_input_tokens_seen": 5308688, "step": 19805 }, { "epoch": 5.192660550458716, "grad_norm": 4.448248386383057, "learning_rate": 2.7679175144208626e-05, "loss": 0.4351, "num_input_tokens_seen": 5310192, "step": 19810 }, { "epoch": 5.193971166448231, "grad_norm": 1.7519475221633911, "learning_rate": 2.7667803449032333e-05, "loss": 0.3129, "num_input_tokens_seen": 5311392, "step": 19815 }, { "epoch": 5.195281782437746, "grad_norm": 1.7683005332946777, "learning_rate": 2.765643119548939e-05, "loss": 0.2276, "num_input_tokens_seen": 5312848, "step": 19820 }, { "epoch": 5.196592398427261, "grad_norm": 1.5368481874465942, "learning_rate": 2.7645058385959987e-05, "loss": 0.2608, "num_input_tokens_seen": 5314528, "step": 19825 }, { "epoch": 5.1979030144167755, "grad_norm": 1.0681610107421875, "learning_rate": 2.7633685022824436e-05, "loss": 0.2573, "num_input_tokens_seen": 5316000, "step": 19830 }, { "epoch": 5.199213630406291, "grad_norm": 1.5405285358428955, "learning_rate": 2.7622311108463167e-05, "loss": 0.2868, "num_input_tokens_seen": 5317328, "step": 19835 }, { "epoch": 5.200524246395806, "grad_norm": 0.8212360739707947, "learning_rate": 2.76109366452567e-05, "loss": 0.1739, "num_input_tokens_seen": 5318336, "step": 19840 }, { "epoch": 5.201834862385321, "grad_norm": 12.742072105407715, "learning_rate": 2.7599561635585713e-05, "loss": 0.9583, "num_input_tokens_seen": 5319488, "step": 19845 }, { "epoch": 5.203145478374836, "grad_norm": 0.6607145667076111, "learning_rate": 2.7588186081830965e-05, "loss": 0.303, "num_input_tokens_seen": 5320560, "step": 19850 }, { "epoch": 5.204456094364351, "grad_norm": 0.6623199582099915, "learning_rate": 2.7576809986373348e-05, "loss": 0.2371, "num_input_tokens_seen": 5321920, "step": 19855 }, { "epoch": 5.205766710353866, "grad_norm": 0.7815482020378113, "learning_rate": 2.756543335159385e-05, "loss": 0.3306, "num_input_tokens_seen": 5323664, "step": 19860 }, { "epoch": 5.207077326343382, "grad_norm": 1.8894668817520142, "learning_rate": 2.7554056179873583e-05, "loss": 0.2297, "num_input_tokens_seen": 5324896, "step": 19865 }, { "epoch": 5.208387942332896, "grad_norm": 1.5889298915863037, "learning_rate": 2.7542678473593762e-05, "loss": 0.4544, "num_input_tokens_seen": 5326288, "step": 19870 }, { "epoch": 5.209698558322412, "grad_norm": 1.0176523923873901, "learning_rate": 2.753130023513574e-05, "loss": 0.1528, "num_input_tokens_seen": 5327472, "step": 19875 }, { "epoch": 5.2110091743119265, "grad_norm": 1.2905219793319702, "learning_rate": 2.7519921466880954e-05, "loss": 0.2972, "num_input_tokens_seen": 5328832, "step": 19880 }, { "epoch": 5.212319790301442, "grad_norm": 1.209609031677246, "learning_rate": 2.7508542171210953e-05, "loss": 0.2559, "num_input_tokens_seen": 5329936, "step": 19885 }, { "epoch": 5.213630406290957, "grad_norm": 1.0681642293930054, "learning_rate": 2.7497162350507415e-05, "loss": 0.2821, "num_input_tokens_seen": 5331248, "step": 19890 }, { "epoch": 5.214941022280472, "grad_norm": 1.8289755582809448, "learning_rate": 2.7485782007152117e-05, "loss": 0.2757, "num_input_tokens_seen": 5332720, "step": 19895 }, { "epoch": 5.216251638269987, "grad_norm": 0.9987429976463318, "learning_rate": 2.747440114352693e-05, "loss": 0.3286, "num_input_tokens_seen": 5333760, "step": 19900 }, { "epoch": 5.217562254259502, "grad_norm": 2.0314629077911377, "learning_rate": 2.7463019762013858e-05, "loss": 0.5005, "num_input_tokens_seen": 5335696, "step": 19905 }, { "epoch": 5.218872870249017, "grad_norm": 1.5342919826507568, "learning_rate": 2.745163786499501e-05, "loss": 0.2609, "num_input_tokens_seen": 5337088, "step": 19910 }, { "epoch": 5.220183486238533, "grad_norm": 1.5729012489318848, "learning_rate": 2.74402554548526e-05, "loss": 0.2773, "num_input_tokens_seen": 5338480, "step": 19915 }, { "epoch": 5.221494102228047, "grad_norm": 1.0755575895309448, "learning_rate": 2.742887253396893e-05, "loss": 0.3158, "num_input_tokens_seen": 5339680, "step": 19920 }, { "epoch": 5.222804718217562, "grad_norm": 2.6883938312530518, "learning_rate": 2.7417489104726435e-05, "loss": 0.3392, "num_input_tokens_seen": 5340880, "step": 19925 }, { "epoch": 5.224115334207077, "grad_norm": 2.204216480255127, "learning_rate": 2.740610516950765e-05, "loss": 0.3542, "num_input_tokens_seen": 5341888, "step": 19930 }, { "epoch": 5.225425950196592, "grad_norm": 1.322090983390808, "learning_rate": 2.739472073069521e-05, "loss": 0.2755, "num_input_tokens_seen": 5343424, "step": 19935 }, { "epoch": 5.226736566186108, "grad_norm": 0.8490749597549438, "learning_rate": 2.738333579067186e-05, "loss": 0.2534, "num_input_tokens_seen": 5345024, "step": 19940 }, { "epoch": 5.228047182175622, "grad_norm": 1.9409488439559937, "learning_rate": 2.7371950351820445e-05, "loss": 0.2588, "num_input_tokens_seen": 5346144, "step": 19945 }, { "epoch": 5.229357798165138, "grad_norm": 1.3417514562606812, "learning_rate": 2.736056441652392e-05, "loss": 0.1539, "num_input_tokens_seen": 5346960, "step": 19950 }, { "epoch": 5.2306684141546524, "grad_norm": 2.3494489192962646, "learning_rate": 2.7349177987165342e-05, "loss": 0.3004, "num_input_tokens_seen": 5348496, "step": 19955 }, { "epoch": 5.231979030144168, "grad_norm": 0.8034600019454956, "learning_rate": 2.7337791066127872e-05, "loss": 0.329, "num_input_tokens_seen": 5349632, "step": 19960 }, { "epoch": 5.233289646133683, "grad_norm": 1.9654561281204224, "learning_rate": 2.732640365579477e-05, "loss": 0.2741, "num_input_tokens_seen": 5350672, "step": 19965 }, { "epoch": 5.234600262123198, "grad_norm": 1.9072750806808472, "learning_rate": 2.73150157585494e-05, "loss": 0.1904, "num_input_tokens_seen": 5351952, "step": 19970 }, { "epoch": 5.235910878112713, "grad_norm": 1.200706958770752, "learning_rate": 2.7303627376775242e-05, "loss": 0.2096, "num_input_tokens_seen": 5353552, "step": 19975 }, { "epoch": 5.237221494102228, "grad_norm": 0.9895368814468384, "learning_rate": 2.7292238512855844e-05, "loss": 0.4788, "num_input_tokens_seen": 5354800, "step": 19980 }, { "epoch": 5.238532110091743, "grad_norm": 2.5578420162200928, "learning_rate": 2.7280849169174887e-05, "loss": 0.2921, "num_input_tokens_seen": 5355696, "step": 19985 }, { "epoch": 5.239842726081259, "grad_norm": 1.158872127532959, "learning_rate": 2.7269459348116143e-05, "loss": 0.3159, "num_input_tokens_seen": 5357232, "step": 19990 }, { "epoch": 5.241153342070773, "grad_norm": 3.8600497245788574, "learning_rate": 2.7258069052063477e-05, "loss": 0.3031, "num_input_tokens_seen": 5358528, "step": 19995 }, { "epoch": 5.242463958060289, "grad_norm": 1.5000160932540894, "learning_rate": 2.724667828340086e-05, "loss": 0.2315, "num_input_tokens_seen": 5359840, "step": 20000 }, { "epoch": 5.243774574049803, "grad_norm": 6.247227191925049, "learning_rate": 2.7235287044512365e-05, "loss": 0.3191, "num_input_tokens_seen": 5360944, "step": 20005 }, { "epoch": 5.245085190039318, "grad_norm": 2.3619022369384766, "learning_rate": 2.7223895337782145e-05, "loss": 0.3552, "num_input_tokens_seen": 5362704, "step": 20010 }, { "epoch": 5.246395806028834, "grad_norm": 2.7200069427490234, "learning_rate": 2.7212503165594483e-05, "loss": 0.2122, "num_input_tokens_seen": 5363584, "step": 20015 }, { "epoch": 5.247706422018348, "grad_norm": 1.8875439167022705, "learning_rate": 2.7201110530333722e-05, "loss": 0.258, "num_input_tokens_seen": 5364544, "step": 20020 }, { "epoch": 5.249017038007864, "grad_norm": 2.037452459335327, "learning_rate": 2.7189717434384338e-05, "loss": 0.402, "num_input_tokens_seen": 5366144, "step": 20025 }, { "epoch": 5.250327653997378, "grad_norm": 5.9480438232421875, "learning_rate": 2.7178323880130878e-05, "loss": 0.3799, "num_input_tokens_seen": 5367408, "step": 20030 }, { "epoch": 5.251638269986894, "grad_norm": 4.135739803314209, "learning_rate": 2.716692986995799e-05, "loss": 0.2272, "num_input_tokens_seen": 5368480, "step": 20035 }, { "epoch": 5.252948885976409, "grad_norm": 4.2548651695251465, "learning_rate": 2.7155535406250425e-05, "loss": 0.4359, "num_input_tokens_seen": 5369744, "step": 20040 }, { "epoch": 5.254259501965924, "grad_norm": 1.2298641204833984, "learning_rate": 2.7144140491393018e-05, "loss": 0.2897, "num_input_tokens_seen": 5370736, "step": 20045 }, { "epoch": 5.255570117955439, "grad_norm": 1.334626317024231, "learning_rate": 2.713274512777072e-05, "loss": 0.1521, "num_input_tokens_seen": 5371744, "step": 20050 }, { "epoch": 5.256880733944954, "grad_norm": 1.2442331314086914, "learning_rate": 2.7121349317768552e-05, "loss": 0.4472, "num_input_tokens_seen": 5373200, "step": 20055 }, { "epoch": 5.258191349934469, "grad_norm": 2.3081703186035156, "learning_rate": 2.710995306377163e-05, "loss": 0.2277, "num_input_tokens_seen": 5374400, "step": 20060 }, { "epoch": 5.259501965923985, "grad_norm": 1.7875142097473145, "learning_rate": 2.7098556368165186e-05, "loss": 0.3954, "num_input_tokens_seen": 5375776, "step": 20065 }, { "epoch": 5.260812581913499, "grad_norm": 1.6765859127044678, "learning_rate": 2.7087159233334512e-05, "loss": 0.3681, "num_input_tokens_seen": 5376848, "step": 20070 }, { "epoch": 5.262123197903015, "grad_norm": 3.9064972400665283, "learning_rate": 2.707576166166502e-05, "loss": 0.267, "num_input_tokens_seen": 5378144, "step": 20075 }, { "epoch": 5.263433813892529, "grad_norm": 0.942844808101654, "learning_rate": 2.7064363655542203e-05, "loss": 0.1041, "num_input_tokens_seen": 5379392, "step": 20080 }, { "epoch": 5.264744429882045, "grad_norm": 0.32746246457099915, "learning_rate": 2.7052965217351633e-05, "loss": 0.4226, "num_input_tokens_seen": 5380576, "step": 20085 }, { "epoch": 5.26605504587156, "grad_norm": 1.1362756490707397, "learning_rate": 2.7041566349479003e-05, "loss": 0.1781, "num_input_tokens_seen": 5381680, "step": 20090 }, { "epoch": 5.267365661861074, "grad_norm": 1.3776780366897583, "learning_rate": 2.7030167054310052e-05, "loss": 0.3451, "num_input_tokens_seen": 5382848, "step": 20095 }, { "epoch": 5.26867627785059, "grad_norm": 1.1900873184204102, "learning_rate": 2.701876733423064e-05, "loss": 0.1758, "num_input_tokens_seen": 5384080, "step": 20100 }, { "epoch": 5.269986893840104, "grad_norm": 1.116032361984253, "learning_rate": 2.700736719162672e-05, "loss": 0.3642, "num_input_tokens_seen": 5385920, "step": 20105 }, { "epoch": 5.27129750982962, "grad_norm": 1.1854000091552734, "learning_rate": 2.699596662888432e-05, "loss": 0.3132, "num_input_tokens_seen": 5387264, "step": 20110 }, { "epoch": 5.272608125819135, "grad_norm": 0.7284985780715942, "learning_rate": 2.698456564838955e-05, "loss": 0.3227, "num_input_tokens_seen": 5388576, "step": 20115 }, { "epoch": 5.27391874180865, "grad_norm": 1.3735376596450806, "learning_rate": 2.6973164252528615e-05, "loss": 0.2389, "num_input_tokens_seen": 5389872, "step": 20120 }, { "epoch": 5.275229357798165, "grad_norm": 1.8929708003997803, "learning_rate": 2.696176244368781e-05, "loss": 0.2513, "num_input_tokens_seen": 5391376, "step": 20125 }, { "epoch": 5.27653997378768, "grad_norm": 1.1634336709976196, "learning_rate": 2.6950360224253524e-05, "loss": 0.3205, "num_input_tokens_seen": 5392736, "step": 20130 }, { "epoch": 5.277850589777195, "grad_norm": 2.6988260746002197, "learning_rate": 2.6938957596612203e-05, "loss": 0.2809, "num_input_tokens_seen": 5393888, "step": 20135 }, { "epoch": 5.2791612057667106, "grad_norm": 1.573703408241272, "learning_rate": 2.6927554563150414e-05, "loss": 0.2586, "num_input_tokens_seen": 5394848, "step": 20140 }, { "epoch": 5.280471821756225, "grad_norm": 1.8991388082504272, "learning_rate": 2.6916151126254785e-05, "loss": 0.4941, "num_input_tokens_seen": 5395872, "step": 20145 }, { "epoch": 5.281782437745741, "grad_norm": 2.7661232948303223, "learning_rate": 2.6904747288312036e-05, "loss": 0.2938, "num_input_tokens_seen": 5397056, "step": 20150 }, { "epoch": 5.283093053735255, "grad_norm": 1.959211826324463, "learning_rate": 2.689334305170897e-05, "loss": 0.2484, "num_input_tokens_seen": 5398368, "step": 20155 }, { "epoch": 5.284403669724771, "grad_norm": 1.4602481126785278, "learning_rate": 2.688193841883247e-05, "loss": 0.2847, "num_input_tokens_seen": 5399664, "step": 20160 }, { "epoch": 5.285714285714286, "grad_norm": 0.8975512385368347, "learning_rate": 2.6870533392069518e-05, "loss": 0.2614, "num_input_tokens_seen": 5400816, "step": 20165 }, { "epoch": 5.287024901703801, "grad_norm": 1.0219534635543823, "learning_rate": 2.6859127973807158e-05, "loss": 0.2689, "num_input_tokens_seen": 5401952, "step": 20170 }, { "epoch": 5.288335517693316, "grad_norm": 2.0100860595703125, "learning_rate": 2.6847722166432522e-05, "loss": 0.324, "num_input_tokens_seen": 5403648, "step": 20175 }, { "epoch": 5.289646133682831, "grad_norm": 1.0379687547683716, "learning_rate": 2.6836315972332827e-05, "loss": 0.3182, "num_input_tokens_seen": 5405728, "step": 20180 }, { "epoch": 5.290956749672346, "grad_norm": 2.606717348098755, "learning_rate": 2.6824909393895375e-05, "loss": 0.258, "num_input_tokens_seen": 5407088, "step": 20185 }, { "epoch": 5.292267365661861, "grad_norm": 1.1896939277648926, "learning_rate": 2.681350243350754e-05, "loss": 0.2743, "num_input_tokens_seen": 5408592, "step": 20190 }, { "epoch": 5.293577981651376, "grad_norm": 1.5315154790878296, "learning_rate": 2.680209509355678e-05, "loss": 0.2967, "num_input_tokens_seen": 5409808, "step": 20195 }, { "epoch": 5.294888597640891, "grad_norm": 2.2698254585266113, "learning_rate": 2.679068737643063e-05, "loss": 0.3313, "num_input_tokens_seen": 5410784, "step": 20200 }, { "epoch": 5.296199213630406, "grad_norm": 1.7014349699020386, "learning_rate": 2.6779279284516706e-05, "loss": 0.2809, "num_input_tokens_seen": 5411984, "step": 20205 }, { "epoch": 5.297509829619921, "grad_norm": 2.3103461265563965, "learning_rate": 2.6767870820202707e-05, "loss": 0.3464, "num_input_tokens_seen": 5413232, "step": 20210 }, { "epoch": 5.2988204456094365, "grad_norm": 1.4863743782043457, "learning_rate": 2.675646198587639e-05, "loss": 0.1916, "num_input_tokens_seen": 5414400, "step": 20215 }, { "epoch": 5.300131061598951, "grad_norm": 1.5253169536590576, "learning_rate": 2.6745052783925616e-05, "loss": 0.3047, "num_input_tokens_seen": 5415872, "step": 20220 }, { "epoch": 5.301441677588467, "grad_norm": 4.949212551116943, "learning_rate": 2.6733643216738325e-05, "loss": 0.3324, "num_input_tokens_seen": 5417152, "step": 20225 }, { "epoch": 5.302752293577981, "grad_norm": 1.3558292388916016, "learning_rate": 2.6722233286702492e-05, "loss": 0.2709, "num_input_tokens_seen": 5418368, "step": 20230 }, { "epoch": 5.304062909567497, "grad_norm": 1.1113015413284302, "learning_rate": 2.671082299620621e-05, "loss": 0.415, "num_input_tokens_seen": 5419728, "step": 20235 }, { "epoch": 5.305373525557012, "grad_norm": 2.733553409576416, "learning_rate": 2.6699412347637626e-05, "loss": 0.3643, "num_input_tokens_seen": 5421408, "step": 20240 }, { "epoch": 5.306684141546527, "grad_norm": 0.8561173677444458, "learning_rate": 2.6688001343384977e-05, "loss": 0.164, "num_input_tokens_seen": 5422768, "step": 20245 }, { "epoch": 5.307994757536042, "grad_norm": 1.2245137691497803, "learning_rate": 2.6676589985836563e-05, "loss": 0.2395, "num_input_tokens_seen": 5424592, "step": 20250 }, { "epoch": 5.309305373525557, "grad_norm": 0.8195958137512207, "learning_rate": 2.6665178277380766e-05, "loss": 0.2071, "num_input_tokens_seen": 5426416, "step": 20255 }, { "epoch": 5.310615989515072, "grad_norm": 2.6474156379699707, "learning_rate": 2.6653766220406022e-05, "loss": 0.4562, "num_input_tokens_seen": 5427536, "step": 20260 }, { "epoch": 5.3119266055045875, "grad_norm": 1.0737898349761963, "learning_rate": 2.664235381730088e-05, "loss": 0.1645, "num_input_tokens_seen": 5428752, "step": 20265 }, { "epoch": 5.313237221494102, "grad_norm": 2.5196242332458496, "learning_rate": 2.663094107045391e-05, "loss": 0.2132, "num_input_tokens_seen": 5429808, "step": 20270 }, { "epoch": 5.314547837483618, "grad_norm": 1.422742247581482, "learning_rate": 2.6619527982253794e-05, "loss": 0.2061, "num_input_tokens_seen": 5431264, "step": 20275 }, { "epoch": 5.315858453473132, "grad_norm": 1.0161882638931274, "learning_rate": 2.6608114555089275e-05, "loss": 0.369, "num_input_tokens_seen": 5432672, "step": 20280 }, { "epoch": 5.317169069462647, "grad_norm": 1.7167760133743286, "learning_rate": 2.659670079134916e-05, "loss": 0.2806, "num_input_tokens_seen": 5433696, "step": 20285 }, { "epoch": 5.3184796854521625, "grad_norm": 0.905925452709198, "learning_rate": 2.6585286693422322e-05, "loss": 0.4688, "num_input_tokens_seen": 5435280, "step": 20290 }, { "epoch": 5.319790301441677, "grad_norm": 4.2790350914001465, "learning_rate": 2.6573872263697724e-05, "loss": 0.6193, "num_input_tokens_seen": 5436416, "step": 20295 }, { "epoch": 5.321100917431193, "grad_norm": 0.8695805072784424, "learning_rate": 2.6562457504564375e-05, "loss": 0.1746, "num_input_tokens_seen": 5438016, "step": 20300 }, { "epoch": 5.322411533420707, "grad_norm": 16.33668327331543, "learning_rate": 2.655104241841137e-05, "loss": 0.4408, "num_input_tokens_seen": 5439392, "step": 20305 }, { "epoch": 5.323722149410223, "grad_norm": 3.4023513793945312, "learning_rate": 2.653962700762787e-05, "loss": 0.2997, "num_input_tokens_seen": 5440960, "step": 20310 }, { "epoch": 5.325032765399738, "grad_norm": 2.118263006210327, "learning_rate": 2.6528211274603092e-05, "loss": 0.3726, "num_input_tokens_seen": 5442816, "step": 20315 }, { "epoch": 5.326343381389253, "grad_norm": 1.6713093519210815, "learning_rate": 2.6516795221726344e-05, "loss": 0.261, "num_input_tokens_seen": 5443792, "step": 20320 }, { "epoch": 5.327653997378768, "grad_norm": 2.6569406986236572, "learning_rate": 2.6505378851386976e-05, "loss": 0.4706, "num_input_tokens_seen": 5445312, "step": 20325 }, { "epoch": 5.328964613368283, "grad_norm": 0.9943259358406067, "learning_rate": 2.6493962165974405e-05, "loss": 0.2595, "num_input_tokens_seen": 5446224, "step": 20330 }, { "epoch": 5.330275229357798, "grad_norm": 2.0713932514190674, "learning_rate": 2.648254516787814e-05, "loss": 0.1562, "num_input_tokens_seen": 5447344, "step": 20335 }, { "epoch": 5.3315858453473135, "grad_norm": 1.8026095628738403, "learning_rate": 2.6471127859487732e-05, "loss": 0.2198, "num_input_tokens_seen": 5448752, "step": 20340 }, { "epoch": 5.332896461336828, "grad_norm": 2.423302173614502, "learning_rate": 2.6459710243192816e-05, "loss": 0.413, "num_input_tokens_seen": 5450000, "step": 20345 }, { "epoch": 5.334207077326344, "grad_norm": 1.1896154880523682, "learning_rate": 2.6448292321383057e-05, "loss": 0.252, "num_input_tokens_seen": 5451104, "step": 20350 }, { "epoch": 5.335517693315858, "grad_norm": 1.8926136493682861, "learning_rate": 2.6436874096448216e-05, "loss": 0.3246, "num_input_tokens_seen": 5452336, "step": 20355 }, { "epoch": 5.336828309305374, "grad_norm": 2.7079062461853027, "learning_rate": 2.642545557077811e-05, "loss": 0.3347, "num_input_tokens_seen": 5453904, "step": 20360 }, { "epoch": 5.3381389252948885, "grad_norm": 2.675870180130005, "learning_rate": 2.641403674676262e-05, "loss": 0.3798, "num_input_tokens_seen": 5455136, "step": 20365 }, { "epoch": 5.339449541284404, "grad_norm": 0.8692746162414551, "learning_rate": 2.6402617626791682e-05, "loss": 0.4158, "num_input_tokens_seen": 5456352, "step": 20370 }, { "epoch": 5.340760157273919, "grad_norm": 1.7851464748382568, "learning_rate": 2.6391198213255297e-05, "loss": 0.375, "num_input_tokens_seen": 5457392, "step": 20375 }, { "epoch": 5.342070773263433, "grad_norm": 2.1244008541107178, "learning_rate": 2.637977850854353e-05, "loss": 0.3473, "num_input_tokens_seen": 5459056, "step": 20380 }, { "epoch": 5.343381389252949, "grad_norm": 1.5165082216262817, "learning_rate": 2.6368358515046503e-05, "loss": 0.2467, "num_input_tokens_seen": 5460080, "step": 20385 }, { "epoch": 5.344692005242464, "grad_norm": 1.5453307628631592, "learning_rate": 2.6356938235154404e-05, "loss": 0.2718, "num_input_tokens_seen": 5461536, "step": 20390 }, { "epoch": 5.346002621231979, "grad_norm": 0.8672682642936707, "learning_rate": 2.634551767125748e-05, "loss": 0.2399, "num_input_tokens_seen": 5462960, "step": 20395 }, { "epoch": 5.347313237221494, "grad_norm": 1.747979760169983, "learning_rate": 2.6334096825746035e-05, "loss": 0.4611, "num_input_tokens_seen": 5464192, "step": 20400 }, { "epoch": 5.348623853211009, "grad_norm": 1.8611595630645752, "learning_rate": 2.6322675701010423e-05, "loss": 0.251, "num_input_tokens_seen": 5465344, "step": 20405 }, { "epoch": 5.349934469200524, "grad_norm": 1.7000659704208374, "learning_rate": 2.6311254299441072e-05, "loss": 0.3546, "num_input_tokens_seen": 5466240, "step": 20410 }, { "epoch": 5.3512450851900395, "grad_norm": 1.4286887645721436, "learning_rate": 2.629983262342846e-05, "loss": 0.2218, "num_input_tokens_seen": 5467360, "step": 20415 }, { "epoch": 5.352555701179554, "grad_norm": 2.9303815364837646, "learning_rate": 2.628841067536314e-05, "loss": 0.2829, "num_input_tokens_seen": 5468512, "step": 20420 }, { "epoch": 5.35386631716907, "grad_norm": 4.476258754730225, "learning_rate": 2.6276988457635675e-05, "loss": 0.3332, "num_input_tokens_seen": 5469552, "step": 20425 }, { "epoch": 5.355176933158584, "grad_norm": 1.4072932004928589, "learning_rate": 2.626556597263674e-05, "loss": 0.3451, "num_input_tokens_seen": 5470864, "step": 20430 }, { "epoch": 5.3564875491481, "grad_norm": 1.3130810260772705, "learning_rate": 2.625414322275703e-05, "loss": 0.3923, "num_input_tokens_seen": 5472080, "step": 20435 }, { "epoch": 5.3577981651376145, "grad_norm": 1.283618450164795, "learning_rate": 2.6242720210387317e-05, "loss": 0.2902, "num_input_tokens_seen": 5473488, "step": 20440 }, { "epoch": 5.35910878112713, "grad_norm": 0.8583328127861023, "learning_rate": 2.623129693791841e-05, "loss": 0.1996, "num_input_tokens_seen": 5474640, "step": 20445 }, { "epoch": 5.360419397116645, "grad_norm": 0.7487677931785583, "learning_rate": 2.6219873407741175e-05, "loss": 0.219, "num_input_tokens_seen": 5475808, "step": 20450 }, { "epoch": 5.36173001310616, "grad_norm": 2.135082721710205, "learning_rate": 2.620844962224655e-05, "loss": 0.2762, "num_input_tokens_seen": 5477072, "step": 20455 }, { "epoch": 5.363040629095675, "grad_norm": 7.247494697570801, "learning_rate": 2.6197025583825508e-05, "loss": 0.2983, "num_input_tokens_seen": 5478352, "step": 20460 }, { "epoch": 5.3643512450851905, "grad_norm": 6.937630653381348, "learning_rate": 2.6185601294869077e-05, "loss": 0.1809, "num_input_tokens_seen": 5479808, "step": 20465 }, { "epoch": 5.365661861074705, "grad_norm": 1.54380464553833, "learning_rate": 2.617417675776834e-05, "loss": 0.2933, "num_input_tokens_seen": 5481104, "step": 20470 }, { "epoch": 5.36697247706422, "grad_norm": 1.9410635232925415, "learning_rate": 2.6162751974914433e-05, "loss": 0.3181, "num_input_tokens_seen": 5482432, "step": 20475 }, { "epoch": 5.368283093053735, "grad_norm": 1.2099945545196533, "learning_rate": 2.6151326948698557e-05, "loss": 0.3733, "num_input_tokens_seen": 5483792, "step": 20480 }, { "epoch": 5.36959370904325, "grad_norm": 1.5608065128326416, "learning_rate": 2.6139901681511935e-05, "loss": 0.3869, "num_input_tokens_seen": 5486112, "step": 20485 }, { "epoch": 5.3709043250327655, "grad_norm": 1.6017704010009766, "learning_rate": 2.6128476175745865e-05, "loss": 0.2028, "num_input_tokens_seen": 5487408, "step": 20490 }, { "epoch": 5.37221494102228, "grad_norm": 0.7544400691986084, "learning_rate": 2.6117050433791672e-05, "loss": 0.3094, "num_input_tokens_seen": 5488656, "step": 20495 }, { "epoch": 5.373525557011796, "grad_norm": 0.7731003165245056, "learning_rate": 2.6105624458040763e-05, "loss": 0.336, "num_input_tokens_seen": 5489984, "step": 20500 }, { "epoch": 5.37483617300131, "grad_norm": 1.050256609916687, "learning_rate": 2.6094198250884562e-05, "loss": 0.3222, "num_input_tokens_seen": 5491664, "step": 20505 }, { "epoch": 5.376146788990826, "grad_norm": 1.3824807405471802, "learning_rate": 2.6082771814714552e-05, "loss": 0.3791, "num_input_tokens_seen": 5492816, "step": 20510 }, { "epoch": 5.3774574049803405, "grad_norm": 20.119606018066406, "learning_rate": 2.6071345151922276e-05, "loss": 0.3367, "num_input_tokens_seen": 5494304, "step": 20515 }, { "epoch": 5.378768020969856, "grad_norm": 2.7069501876831055, "learning_rate": 2.605991826489932e-05, "loss": 0.297, "num_input_tokens_seen": 5495264, "step": 20520 }, { "epoch": 5.380078636959371, "grad_norm": 1.2244843244552612, "learning_rate": 2.604849115603729e-05, "loss": 0.2911, "num_input_tokens_seen": 5496736, "step": 20525 }, { "epoch": 5.381389252948886, "grad_norm": 1.7061306238174438, "learning_rate": 2.6037063827727874e-05, "loss": 0.3438, "num_input_tokens_seen": 5497840, "step": 20530 }, { "epoch": 5.382699868938401, "grad_norm": 0.708448588848114, "learning_rate": 2.6025636282362792e-05, "loss": 0.2685, "num_input_tokens_seen": 5498992, "step": 20535 }, { "epoch": 5.3840104849279165, "grad_norm": 1.7129814624786377, "learning_rate": 2.6014208522333815e-05, "loss": 0.29, "num_input_tokens_seen": 5500256, "step": 20540 }, { "epoch": 5.385321100917431, "grad_norm": 1.3680624961853027, "learning_rate": 2.6002780550032747e-05, "loss": 0.2666, "num_input_tokens_seen": 5501600, "step": 20545 }, { "epoch": 5.386631716906947, "grad_norm": 1.7456402778625488, "learning_rate": 2.5991352367851435e-05, "loss": 0.1669, "num_input_tokens_seen": 5503024, "step": 20550 }, { "epoch": 5.387942332896461, "grad_norm": 2.5520708560943604, "learning_rate": 2.5979923978181787e-05, "loss": 0.42, "num_input_tokens_seen": 5504352, "step": 20555 }, { "epoch": 5.389252948885977, "grad_norm": 1.5238032341003418, "learning_rate": 2.5968495383415742e-05, "loss": 0.2425, "num_input_tokens_seen": 5505808, "step": 20560 }, { "epoch": 5.3905635648754915, "grad_norm": 1.180689811706543, "learning_rate": 2.595706658594529e-05, "loss": 0.4574, "num_input_tokens_seen": 5506944, "step": 20565 }, { "epoch": 5.391874180865006, "grad_norm": 0.884657084941864, "learning_rate": 2.5945637588162452e-05, "loss": 0.2876, "num_input_tokens_seen": 5507888, "step": 20570 }, { "epoch": 5.393184796854522, "grad_norm": 1.7741113901138306, "learning_rate": 2.5934208392459304e-05, "loss": 0.2433, "num_input_tokens_seen": 5509088, "step": 20575 }, { "epoch": 5.394495412844036, "grad_norm": 1.048124074935913, "learning_rate": 2.592277900122796e-05, "loss": 0.2683, "num_input_tokens_seen": 5510448, "step": 20580 }, { "epoch": 5.395806028833552, "grad_norm": 2.1404151916503906, "learning_rate": 2.5911349416860558e-05, "loss": 0.2993, "num_input_tokens_seen": 5511728, "step": 20585 }, { "epoch": 5.3971166448230665, "grad_norm": 1.5012961626052856, "learning_rate": 2.5899919641749286e-05, "loss": 0.3153, "num_input_tokens_seen": 5512880, "step": 20590 }, { "epoch": 5.398427260812582, "grad_norm": 2.654632568359375, "learning_rate": 2.588848967828641e-05, "loss": 0.3555, "num_input_tokens_seen": 5514032, "step": 20595 }, { "epoch": 5.399737876802097, "grad_norm": 1.3884079456329346, "learning_rate": 2.5877059528864168e-05, "loss": 0.2412, "num_input_tokens_seen": 5515504, "step": 20600 }, { "epoch": 5.401048492791612, "grad_norm": 1.3531696796417236, "learning_rate": 2.5865629195874886e-05, "loss": 0.2736, "num_input_tokens_seen": 5516864, "step": 20605 }, { "epoch": 5.402359108781127, "grad_norm": 1.1527453660964966, "learning_rate": 2.58541986817109e-05, "loss": 0.1759, "num_input_tokens_seen": 5518384, "step": 20610 }, { "epoch": 5.4036697247706424, "grad_norm": 1.413683295249939, "learning_rate": 2.5842767988764615e-05, "loss": 0.2983, "num_input_tokens_seen": 5519696, "step": 20615 }, { "epoch": 5.404980340760157, "grad_norm": 2.022336006164551, "learning_rate": 2.5831337119428446e-05, "loss": 0.3255, "num_input_tokens_seen": 5520800, "step": 20620 }, { "epoch": 5.406290956749673, "grad_norm": 1.513763427734375, "learning_rate": 2.5819906076094848e-05, "loss": 0.3272, "num_input_tokens_seen": 5521968, "step": 20625 }, { "epoch": 5.407601572739187, "grad_norm": 1.584681749343872, "learning_rate": 2.5808474861156328e-05, "loss": 0.364, "num_input_tokens_seen": 5523024, "step": 20630 }, { "epoch": 5.408912188728703, "grad_norm": 1.7371647357940674, "learning_rate": 2.5797043477005423e-05, "loss": 0.3659, "num_input_tokens_seen": 5524224, "step": 20635 }, { "epoch": 5.4102228047182175, "grad_norm": 0.9815015196800232, "learning_rate": 2.5785611926034686e-05, "loss": 0.3396, "num_input_tokens_seen": 5525360, "step": 20640 }, { "epoch": 5.411533420707733, "grad_norm": 0.21297472715377808, "learning_rate": 2.5774180210636728e-05, "loss": 0.2336, "num_input_tokens_seen": 5527984, "step": 20645 }, { "epoch": 5.412844036697248, "grad_norm": 2.800964593887329, "learning_rate": 2.5762748333204196e-05, "loss": 0.3711, "num_input_tokens_seen": 5529744, "step": 20650 }, { "epoch": 5.414154652686763, "grad_norm": 1.1514092683792114, "learning_rate": 2.575131629612976e-05, "loss": 0.2109, "num_input_tokens_seen": 5531152, "step": 20655 }, { "epoch": 5.415465268676278, "grad_norm": 1.3627357482910156, "learning_rate": 2.5739884101806117e-05, "loss": 0.2543, "num_input_tokens_seen": 5532592, "step": 20660 }, { "epoch": 5.4167758846657925, "grad_norm": 1.0112524032592773, "learning_rate": 2.5728451752626008e-05, "loss": 0.2894, "num_input_tokens_seen": 5533824, "step": 20665 }, { "epoch": 5.418086500655308, "grad_norm": 1.5858067274093628, "learning_rate": 2.571701925098221e-05, "loss": 0.4912, "num_input_tokens_seen": 5535792, "step": 20670 }, { "epoch": 5.419397116644823, "grad_norm": 1.162253975868225, "learning_rate": 2.5705586599267518e-05, "loss": 0.291, "num_input_tokens_seen": 5537456, "step": 20675 }, { "epoch": 5.420707732634338, "grad_norm": 0.9977052807807922, "learning_rate": 2.569415379987477e-05, "loss": 0.2946, "num_input_tokens_seen": 5538688, "step": 20680 }, { "epoch": 5.422018348623853, "grad_norm": 1.1870025396347046, "learning_rate": 2.5682720855196836e-05, "loss": 0.279, "num_input_tokens_seen": 5540464, "step": 20685 }, { "epoch": 5.423328964613368, "grad_norm": 1.2885215282440186, "learning_rate": 2.56712877676266e-05, "loss": 0.3367, "num_input_tokens_seen": 5541440, "step": 20690 }, { "epoch": 5.424639580602883, "grad_norm": 1.4357815980911255, "learning_rate": 2.5659854539556998e-05, "loss": 0.3329, "num_input_tokens_seen": 5542640, "step": 20695 }, { "epoch": 5.425950196592399, "grad_norm": 1.0864776372909546, "learning_rate": 2.5648421173380977e-05, "loss": 0.2962, "num_input_tokens_seen": 5543920, "step": 20700 }, { "epoch": 5.427260812581913, "grad_norm": 1.9996299743652344, "learning_rate": 2.5636987671491525e-05, "loss": 0.2806, "num_input_tokens_seen": 5545152, "step": 20705 }, { "epoch": 5.428571428571429, "grad_norm": 0.9394932389259338, "learning_rate": 2.562555403628166e-05, "loss": 0.2583, "num_input_tokens_seen": 5546336, "step": 20710 }, { "epoch": 5.4298820445609435, "grad_norm": 3.2119619846343994, "learning_rate": 2.5614120270144415e-05, "loss": 0.2258, "num_input_tokens_seen": 5547824, "step": 20715 }, { "epoch": 5.431192660550459, "grad_norm": 3.7539024353027344, "learning_rate": 2.5602686375472856e-05, "loss": 0.3969, "num_input_tokens_seen": 5548976, "step": 20720 }, { "epoch": 5.432503276539974, "grad_norm": 2.563953161239624, "learning_rate": 2.5591252354660083e-05, "loss": 0.5446, "num_input_tokens_seen": 5550160, "step": 20725 }, { "epoch": 5.433813892529489, "grad_norm": 1.4749304056167603, "learning_rate": 2.557981821009921e-05, "loss": 0.339, "num_input_tokens_seen": 5551280, "step": 20730 }, { "epoch": 5.435124508519004, "grad_norm": 0.9765479564666748, "learning_rate": 2.556838394418339e-05, "loss": 0.1791, "num_input_tokens_seen": 5552480, "step": 20735 }, { "epoch": 5.436435124508519, "grad_norm": 1.5518637895584106, "learning_rate": 2.5556949559305786e-05, "loss": 0.1883, "num_input_tokens_seen": 5553840, "step": 20740 }, { "epoch": 5.437745740498034, "grad_norm": 1.2186598777770996, "learning_rate": 2.554551505785961e-05, "loss": 0.2157, "num_input_tokens_seen": 5555248, "step": 20745 }, { "epoch": 5.43905635648755, "grad_norm": 1.6952106952667236, "learning_rate": 2.553408044223807e-05, "loss": 0.2875, "num_input_tokens_seen": 5556592, "step": 20750 }, { "epoch": 5.440366972477064, "grad_norm": 1.6791247129440308, "learning_rate": 2.5522645714834426e-05, "loss": 0.3326, "num_input_tokens_seen": 5557888, "step": 20755 }, { "epoch": 5.441677588466579, "grad_norm": 2.484325408935547, "learning_rate": 2.551121087804192e-05, "loss": 0.2604, "num_input_tokens_seen": 5559024, "step": 20760 }, { "epoch": 5.442988204456094, "grad_norm": 4.417399883270264, "learning_rate": 2.5499775934253865e-05, "loss": 0.2301, "num_input_tokens_seen": 5560048, "step": 20765 }, { "epoch": 5.444298820445609, "grad_norm": 0.7939776182174683, "learning_rate": 2.5488340885863578e-05, "loss": 0.2686, "num_input_tokens_seen": 5561296, "step": 20770 }, { "epoch": 5.445609436435125, "grad_norm": 2.8306686878204346, "learning_rate": 2.547690573526439e-05, "loss": 0.3478, "num_input_tokens_seen": 5563136, "step": 20775 }, { "epoch": 5.446920052424639, "grad_norm": 2.4455811977386475, "learning_rate": 2.546547048484965e-05, "loss": 0.3137, "num_input_tokens_seen": 5564560, "step": 20780 }, { "epoch": 5.448230668414155, "grad_norm": 1.1897295713424683, "learning_rate": 2.545403513701274e-05, "loss": 0.304, "num_input_tokens_seen": 5565696, "step": 20785 }, { "epoch": 5.4495412844036695, "grad_norm": 2.277740716934204, "learning_rate": 2.544259969414706e-05, "loss": 0.2977, "num_input_tokens_seen": 5567296, "step": 20790 }, { "epoch": 5.450851900393185, "grad_norm": 1.6389942169189453, "learning_rate": 2.543116415864603e-05, "loss": 0.3333, "num_input_tokens_seen": 5568752, "step": 20795 }, { "epoch": 5.4521625163827, "grad_norm": 1.0025200843811035, "learning_rate": 2.541972853290309e-05, "loss": 0.2009, "num_input_tokens_seen": 5569904, "step": 20800 }, { "epoch": 5.453473132372215, "grad_norm": 1.382649302482605, "learning_rate": 2.540829281931169e-05, "loss": 0.3782, "num_input_tokens_seen": 5571568, "step": 20805 }, { "epoch": 5.45478374836173, "grad_norm": 1.69512140750885, "learning_rate": 2.5396857020265307e-05, "loss": 0.3203, "num_input_tokens_seen": 5572704, "step": 20810 }, { "epoch": 5.456094364351245, "grad_norm": 2.9091579914093018, "learning_rate": 2.538542113815744e-05, "loss": 0.3792, "num_input_tokens_seen": 5574576, "step": 20815 }, { "epoch": 5.45740498034076, "grad_norm": 0.5492436289787292, "learning_rate": 2.5373985175381594e-05, "loss": 0.2348, "num_input_tokens_seen": 5575648, "step": 20820 }, { "epoch": 5.458715596330276, "grad_norm": 8.184869766235352, "learning_rate": 2.536254913433129e-05, "loss": 0.2149, "num_input_tokens_seen": 5576784, "step": 20825 }, { "epoch": 5.46002621231979, "grad_norm": 2.2438817024230957, "learning_rate": 2.535111301740009e-05, "loss": 0.3304, "num_input_tokens_seen": 5577920, "step": 20830 }, { "epoch": 5.461336828309306, "grad_norm": 1.8269575834274292, "learning_rate": 2.5339676826981534e-05, "loss": 0.3678, "num_input_tokens_seen": 5579280, "step": 20835 }, { "epoch": 5.46264744429882, "grad_norm": 1.0219815969467163, "learning_rate": 2.532824056546921e-05, "loss": 0.3079, "num_input_tokens_seen": 5582000, "step": 20840 }, { "epoch": 5.463958060288336, "grad_norm": 1.5388704538345337, "learning_rate": 2.531680423525669e-05, "loss": 0.2908, "num_input_tokens_seen": 5583264, "step": 20845 }, { "epoch": 5.465268676277851, "grad_norm": 2.0788426399230957, "learning_rate": 2.5305367838737608e-05, "loss": 0.2487, "num_input_tokens_seen": 5584304, "step": 20850 }, { "epoch": 5.466579292267365, "grad_norm": 1.4323455095291138, "learning_rate": 2.5293931378305558e-05, "loss": 0.3516, "num_input_tokens_seen": 5585824, "step": 20855 }, { "epoch": 5.467889908256881, "grad_norm": 1.723847508430481, "learning_rate": 2.5282494856354176e-05, "loss": 0.3489, "num_input_tokens_seen": 5587152, "step": 20860 }, { "epoch": 5.4692005242463955, "grad_norm": 1.5980154275894165, "learning_rate": 2.5271058275277117e-05, "loss": 0.2728, "num_input_tokens_seen": 5588368, "step": 20865 }, { "epoch": 5.470511140235911, "grad_norm": 1.4042749404907227, "learning_rate": 2.5259621637468016e-05, "loss": 0.2885, "num_input_tokens_seen": 5590128, "step": 20870 }, { "epoch": 5.471821756225426, "grad_norm": 1.6241838932037354, "learning_rate": 2.5248184945320563e-05, "loss": 0.3696, "num_input_tokens_seen": 5591712, "step": 20875 }, { "epoch": 5.473132372214941, "grad_norm": 2.165687084197998, "learning_rate": 2.5236748201228428e-05, "loss": 0.3222, "num_input_tokens_seen": 5592880, "step": 20880 }, { "epoch": 5.474442988204456, "grad_norm": 0.7275102138519287, "learning_rate": 2.5225311407585306e-05, "loss": 0.1993, "num_input_tokens_seen": 5594064, "step": 20885 }, { "epoch": 5.475753604193971, "grad_norm": 1.611797571182251, "learning_rate": 2.5213874566784902e-05, "loss": 0.2959, "num_input_tokens_seen": 5595344, "step": 20890 }, { "epoch": 5.477064220183486, "grad_norm": 1.914402723312378, "learning_rate": 2.5202437681220907e-05, "loss": 0.3371, "num_input_tokens_seen": 5596576, "step": 20895 }, { "epoch": 5.478374836173002, "grad_norm": 1.1047813892364502, "learning_rate": 2.5191000753287064e-05, "loss": 0.2875, "num_input_tokens_seen": 5597712, "step": 20900 }, { "epoch": 5.479685452162516, "grad_norm": 1.2318493127822876, "learning_rate": 2.5179563785377088e-05, "loss": 0.3148, "num_input_tokens_seen": 5598848, "step": 20905 }, { "epoch": 5.480996068152032, "grad_norm": 2.5819766521453857, "learning_rate": 2.5168126779884727e-05, "loss": 0.3682, "num_input_tokens_seen": 5600336, "step": 20910 }, { "epoch": 5.482306684141546, "grad_norm": 0.6741869449615479, "learning_rate": 2.5156689739203714e-05, "loss": 0.3345, "num_input_tokens_seen": 5601392, "step": 20915 }, { "epoch": 5.483617300131062, "grad_norm": 1.4873913526535034, "learning_rate": 2.5145252665727815e-05, "loss": 0.315, "num_input_tokens_seen": 5602864, "step": 20920 }, { "epoch": 5.484927916120577, "grad_norm": 0.8145521879196167, "learning_rate": 2.5133815561850782e-05, "loss": 0.2728, "num_input_tokens_seen": 5604016, "step": 20925 }, { "epoch": 5.486238532110092, "grad_norm": 1.2537059783935547, "learning_rate": 2.5122378429966387e-05, "loss": 0.2468, "num_input_tokens_seen": 5605120, "step": 20930 }, { "epoch": 5.487549148099607, "grad_norm": 2.2570900917053223, "learning_rate": 2.511094127246839e-05, "loss": 0.3595, "num_input_tokens_seen": 5606256, "step": 20935 }, { "epoch": 5.488859764089122, "grad_norm": 1.0669488906860352, "learning_rate": 2.5099504091750585e-05, "loss": 0.1629, "num_input_tokens_seen": 5607424, "step": 20940 }, { "epoch": 5.490170380078637, "grad_norm": 1.6709719896316528, "learning_rate": 2.508806689020674e-05, "loss": 0.5017, "num_input_tokens_seen": 5608528, "step": 20945 }, { "epoch": 5.491480996068152, "grad_norm": 1.6123181581497192, "learning_rate": 2.5076629670230655e-05, "loss": 0.3244, "num_input_tokens_seen": 5609824, "step": 20950 }, { "epoch": 5.492791612057667, "grad_norm": 1.021705150604248, "learning_rate": 2.5065192434216106e-05, "loss": 0.2765, "num_input_tokens_seen": 5610944, "step": 20955 }, { "epoch": 5.494102228047182, "grad_norm": 1.1845554113388062, "learning_rate": 2.5053755184556886e-05, "loss": 0.136, "num_input_tokens_seen": 5612016, "step": 20960 }, { "epoch": 5.495412844036697, "grad_norm": 1.287732481956482, "learning_rate": 2.5042317923646812e-05, "loss": 0.3382, "num_input_tokens_seen": 5613632, "step": 20965 }, { "epoch": 5.496723460026212, "grad_norm": 2.263669967651367, "learning_rate": 2.5030880653879667e-05, "loss": 0.3395, "num_input_tokens_seen": 5615056, "step": 20970 }, { "epoch": 5.498034076015728, "grad_norm": 0.5658301711082458, "learning_rate": 2.5019443377649247e-05, "loss": 0.165, "num_input_tokens_seen": 5616192, "step": 20975 }, { "epoch": 5.499344692005242, "grad_norm": 2.307352304458618, "learning_rate": 2.500800609734937e-05, "loss": 0.2846, "num_input_tokens_seen": 5617296, "step": 20980 }, { "epoch": 5.500655307994758, "grad_norm": 1.2929041385650635, "learning_rate": 2.4996568815373827e-05, "loss": 0.4, "num_input_tokens_seen": 5618656, "step": 20985 }, { "epoch": 5.501441677588467, "eval_loss": 0.588385283946991, "eval_runtime": 16.6994, "eval_samples_per_second": 50.78, "eval_steps_per_second": 25.39, "num_input_tokens_seen": 5619904, "step": 20988 }, { "epoch": 5.501965923984272, "grad_norm": 3.7175960540771484, "learning_rate": 2.4985131534116422e-05, "loss": 0.4433, "num_input_tokens_seen": 5620480, "step": 20990 }, { "epoch": 5.503276539973788, "grad_norm": 1.70937979221344, "learning_rate": 2.4973694255970966e-05, "loss": 0.4726, "num_input_tokens_seen": 5621648, "step": 20995 }, { "epoch": 5.504587155963303, "grad_norm": 0.9829655885696411, "learning_rate": 2.4962256983331254e-05, "loss": 0.3227, "num_input_tokens_seen": 5622768, "step": 21000 }, { "epoch": 5.505897771952818, "grad_norm": 1.3478831052780151, "learning_rate": 2.4950819718591094e-05, "loss": 0.2562, "num_input_tokens_seen": 5624048, "step": 21005 }, { "epoch": 5.507208387942333, "grad_norm": 1.143445372581482, "learning_rate": 2.493938246414428e-05, "loss": 0.2132, "num_input_tokens_seen": 5625344, "step": 21010 }, { "epoch": 5.508519003931848, "grad_norm": 1.9640520811080933, "learning_rate": 2.4927945222384613e-05, "loss": 0.3369, "num_input_tokens_seen": 5626480, "step": 21015 }, { "epoch": 5.509829619921363, "grad_norm": 1.9779683351516724, "learning_rate": 2.491650799570588e-05, "loss": 0.2278, "num_input_tokens_seen": 5627744, "step": 21020 }, { "epoch": 5.511140235910878, "grad_norm": 2.0416245460510254, "learning_rate": 2.4905070786501887e-05, "loss": 0.2751, "num_input_tokens_seen": 5629008, "step": 21025 }, { "epoch": 5.512450851900393, "grad_norm": 1.7028287649154663, "learning_rate": 2.4893633597166405e-05, "loss": 0.4069, "num_input_tokens_seen": 5630144, "step": 21030 }, { "epoch": 5.513761467889909, "grad_norm": 1.7972054481506348, "learning_rate": 2.4882196430093233e-05, "loss": 0.3453, "num_input_tokens_seen": 5632112, "step": 21035 }, { "epoch": 5.515072083879423, "grad_norm": 1.053876280784607, "learning_rate": 2.4870759287676143e-05, "loss": 0.3267, "num_input_tokens_seen": 5633664, "step": 21040 }, { "epoch": 5.516382699868938, "grad_norm": 1.9761183261871338, "learning_rate": 2.485932217230892e-05, "loss": 0.3045, "num_input_tokens_seen": 5634992, "step": 21045 }, { "epoch": 5.517693315858454, "grad_norm": 1.666211485862732, "learning_rate": 2.4847885086385314e-05, "loss": 0.3452, "num_input_tokens_seen": 5636000, "step": 21050 }, { "epoch": 5.519003931847968, "grad_norm": 1.8493047952651978, "learning_rate": 2.4836448032299097e-05, "loss": 0.2105, "num_input_tokens_seen": 5637088, "step": 21055 }, { "epoch": 5.520314547837484, "grad_norm": 1.2289950847625732, "learning_rate": 2.4825011012444026e-05, "loss": 0.3078, "num_input_tokens_seen": 5638240, "step": 21060 }, { "epoch": 5.521625163826998, "grad_norm": 0.9469043016433716, "learning_rate": 2.4813574029213853e-05, "loss": 0.3565, "num_input_tokens_seen": 5639952, "step": 21065 }, { "epoch": 5.522935779816514, "grad_norm": 1.2690155506134033, "learning_rate": 2.4802137085002306e-05, "loss": 0.345, "num_input_tokens_seen": 5641504, "step": 21070 }, { "epoch": 5.524246395806029, "grad_norm": 3.9917945861816406, "learning_rate": 2.479070018220313e-05, "loss": 0.6214, "num_input_tokens_seen": 5642512, "step": 21075 }, { "epoch": 5.525557011795544, "grad_norm": 0.5862970352172852, "learning_rate": 2.4779263323210036e-05, "loss": 0.239, "num_input_tokens_seen": 5643776, "step": 21080 }, { "epoch": 5.526867627785059, "grad_norm": 0.7906697392463684, "learning_rate": 2.4767826510416753e-05, "loss": 0.2769, "num_input_tokens_seen": 5645536, "step": 21085 }, { "epoch": 5.528178243774574, "grad_norm": 1.6603772640228271, "learning_rate": 2.4756389746216974e-05, "loss": 0.2021, "num_input_tokens_seen": 5646976, "step": 21090 }, { "epoch": 5.529488859764089, "grad_norm": 0.9273515939712524, "learning_rate": 2.4744953033004408e-05, "loss": 0.2322, "num_input_tokens_seen": 5648368, "step": 21095 }, { "epoch": 5.5307994757536045, "grad_norm": 1.0944797992706299, "learning_rate": 2.4733516373172728e-05, "loss": 0.2908, "num_input_tokens_seen": 5649824, "step": 21100 }, { "epoch": 5.532110091743119, "grad_norm": 1.1516096591949463, "learning_rate": 2.4722079769115617e-05, "loss": 0.2987, "num_input_tokens_seen": 5650976, "step": 21105 }, { "epoch": 5.533420707732635, "grad_norm": 2.5784096717834473, "learning_rate": 2.4710643223226722e-05, "loss": 0.352, "num_input_tokens_seen": 5652608, "step": 21110 }, { "epoch": 5.534731323722149, "grad_norm": 2.21742844581604, "learning_rate": 2.4699206737899707e-05, "loss": 0.3183, "num_input_tokens_seen": 5653808, "step": 21115 }, { "epoch": 5.536041939711664, "grad_norm": 1.1114078760147095, "learning_rate": 2.4687770315528195e-05, "loss": 0.3645, "num_input_tokens_seen": 5655200, "step": 21120 }, { "epoch": 5.53735255570118, "grad_norm": 1.525415062904358, "learning_rate": 2.4676333958505827e-05, "loss": 0.3466, "num_input_tokens_seen": 5656192, "step": 21125 }, { "epoch": 5.538663171690695, "grad_norm": 1.5467272996902466, "learning_rate": 2.4664897669226193e-05, "loss": 0.3025, "num_input_tokens_seen": 5657472, "step": 21130 }, { "epoch": 5.53997378768021, "grad_norm": 0.9436496496200562, "learning_rate": 2.46534614500829e-05, "loss": 0.2957, "num_input_tokens_seen": 5658416, "step": 21135 }, { "epoch": 5.541284403669724, "grad_norm": 1.7344677448272705, "learning_rate": 2.4642025303469527e-05, "loss": 0.2797, "num_input_tokens_seen": 5659776, "step": 21140 }, { "epoch": 5.54259501965924, "grad_norm": 1.1087876558303833, "learning_rate": 2.4630589231779633e-05, "loss": 0.2542, "num_input_tokens_seen": 5660912, "step": 21145 }, { "epoch": 5.543905635648755, "grad_norm": 3.9150989055633545, "learning_rate": 2.4619153237406782e-05, "loss": 0.3407, "num_input_tokens_seen": 5663440, "step": 21150 }, { "epoch": 5.54521625163827, "grad_norm": 0.9305122494697571, "learning_rate": 2.4607717322744498e-05, "loss": 0.2451, "num_input_tokens_seen": 5664960, "step": 21155 }, { "epoch": 5.546526867627785, "grad_norm": 1.0053982734680176, "learning_rate": 2.4596281490186303e-05, "loss": 0.4239, "num_input_tokens_seen": 5666048, "step": 21160 }, { "epoch": 5.5478374836173, "grad_norm": 1.5032737255096436, "learning_rate": 2.4584845742125696e-05, "loss": 0.1552, "num_input_tokens_seen": 5667616, "step": 21165 }, { "epoch": 5.549148099606815, "grad_norm": 2.3495945930480957, "learning_rate": 2.457341008095615e-05, "loss": 0.258, "num_input_tokens_seen": 5668880, "step": 21170 }, { "epoch": 5.5504587155963305, "grad_norm": 0.898339569568634, "learning_rate": 2.4561974509071143e-05, "loss": 0.3103, "num_input_tokens_seen": 5670032, "step": 21175 }, { "epoch": 5.551769331585845, "grad_norm": 2.4878220558166504, "learning_rate": 2.4550539028864113e-05, "loss": 0.3258, "num_input_tokens_seen": 5671280, "step": 21180 }, { "epoch": 5.553079947575361, "grad_norm": 1.9366796016693115, "learning_rate": 2.4539103642728494e-05, "loss": 0.3288, "num_input_tokens_seen": 5672544, "step": 21185 }, { "epoch": 5.554390563564875, "grad_norm": 2.3049144744873047, "learning_rate": 2.4527668353057683e-05, "loss": 0.231, "num_input_tokens_seen": 5673792, "step": 21190 }, { "epoch": 5.555701179554391, "grad_norm": 1.4197998046875, "learning_rate": 2.4516233162245063e-05, "loss": 0.4142, "num_input_tokens_seen": 5674848, "step": 21195 }, { "epoch": 5.557011795543906, "grad_norm": 1.3624705076217651, "learning_rate": 2.450479807268401e-05, "loss": 0.2793, "num_input_tokens_seen": 5676064, "step": 21200 }, { "epoch": 5.558322411533421, "grad_norm": 1.5463988780975342, "learning_rate": 2.449336308676786e-05, "loss": 0.3116, "num_input_tokens_seen": 5677472, "step": 21205 }, { "epoch": 5.559633027522936, "grad_norm": 1.4991949796676636, "learning_rate": 2.4481928206889946e-05, "loss": 0.3948, "num_input_tokens_seen": 5678752, "step": 21210 }, { "epoch": 5.56094364351245, "grad_norm": 1.7323695421218872, "learning_rate": 2.4470493435443558e-05, "loss": 0.46, "num_input_tokens_seen": 5680288, "step": 21215 }, { "epoch": 5.562254259501966, "grad_norm": 1.855953335762024, "learning_rate": 2.4459058774821992e-05, "loss": 0.2762, "num_input_tokens_seen": 5681360, "step": 21220 }, { "epoch": 5.5635648754914815, "grad_norm": 1.4537312984466553, "learning_rate": 2.4447624227418482e-05, "loss": 0.3172, "num_input_tokens_seen": 5682432, "step": 21225 }, { "epoch": 5.564875491480996, "grad_norm": 0.7472761273384094, "learning_rate": 2.4436189795626262e-05, "loss": 0.2273, "num_input_tokens_seen": 5683808, "step": 21230 }, { "epoch": 5.566186107470511, "grad_norm": 2.4340600967407227, "learning_rate": 2.442475548183855e-05, "loss": 0.223, "num_input_tokens_seen": 5684928, "step": 21235 }, { "epoch": 5.567496723460026, "grad_norm": 1.6921464204788208, "learning_rate": 2.4413321288448518e-05, "loss": 0.2339, "num_input_tokens_seen": 5685840, "step": 21240 }, { "epoch": 5.568807339449541, "grad_norm": 1.7607451677322388, "learning_rate": 2.440188721784933e-05, "loss": 0.3988, "num_input_tokens_seen": 5687008, "step": 21245 }, { "epoch": 5.5701179554390565, "grad_norm": 1.1313989162445068, "learning_rate": 2.4390453272434115e-05, "loss": 0.2422, "num_input_tokens_seen": 5688400, "step": 21250 }, { "epoch": 5.571428571428571, "grad_norm": 1.9669772386550903, "learning_rate": 2.4379019454595973e-05, "loss": 0.321, "num_input_tokens_seen": 5689712, "step": 21255 }, { "epoch": 5.572739187418087, "grad_norm": 1.265230417251587, "learning_rate": 2.436758576672798e-05, "loss": 0.3192, "num_input_tokens_seen": 5690816, "step": 21260 }, { "epoch": 5.574049803407601, "grad_norm": 1.0909309387207031, "learning_rate": 2.4356152211223198e-05, "loss": 0.2045, "num_input_tokens_seen": 5692032, "step": 21265 }, { "epoch": 5.575360419397117, "grad_norm": 3.621068239212036, "learning_rate": 2.434471879047465e-05, "loss": 0.4213, "num_input_tokens_seen": 5693568, "step": 21270 }, { "epoch": 5.576671035386632, "grad_norm": 3.982790470123291, "learning_rate": 2.4333285506875322e-05, "loss": 0.3118, "num_input_tokens_seen": 5695056, "step": 21275 }, { "epoch": 5.577981651376147, "grad_norm": 3.5290331840515137, "learning_rate": 2.4321852362818195e-05, "loss": 0.3163, "num_input_tokens_seen": 5696480, "step": 21280 }, { "epoch": 5.579292267365662, "grad_norm": 0.8566789627075195, "learning_rate": 2.431041936069619e-05, "loss": 0.2661, "num_input_tokens_seen": 5697632, "step": 21285 }, { "epoch": 5.580602883355177, "grad_norm": 7.724262714385986, "learning_rate": 2.4298986502902218e-05, "loss": 0.4323, "num_input_tokens_seen": 5698752, "step": 21290 }, { "epoch": 5.581913499344692, "grad_norm": 4.05936336517334, "learning_rate": 2.4287553791829166e-05, "loss": 0.3856, "num_input_tokens_seen": 5700192, "step": 21295 }, { "epoch": 5.5832241153342075, "grad_norm": 0.8279931545257568, "learning_rate": 2.4276121229869875e-05, "loss": 0.1939, "num_input_tokens_seen": 5701520, "step": 21300 }, { "epoch": 5.584534731323722, "grad_norm": 0.8708128333091736, "learning_rate": 2.4264688819417155e-05, "loss": 0.2643, "num_input_tokens_seen": 5703280, "step": 21305 }, { "epoch": 5.585845347313237, "grad_norm": 1.5754224061965942, "learning_rate": 2.4253256562863795e-05, "loss": 0.1929, "num_input_tokens_seen": 5704848, "step": 21310 }, { "epoch": 5.587155963302752, "grad_norm": 2.8901901245117188, "learning_rate": 2.4241824462602546e-05, "loss": 0.3436, "num_input_tokens_seen": 5706608, "step": 21315 }, { "epoch": 5.588466579292267, "grad_norm": 1.213430643081665, "learning_rate": 2.4230392521026123e-05, "loss": 0.3552, "num_input_tokens_seen": 5707936, "step": 21320 }, { "epoch": 5.5897771952817825, "grad_norm": 2.218508720397949, "learning_rate": 2.4218960740527218e-05, "loss": 0.3128, "num_input_tokens_seen": 5709200, "step": 21325 }, { "epoch": 5.591087811271297, "grad_norm": 1.8137261867523193, "learning_rate": 2.4207529123498485e-05, "loss": 0.253, "num_input_tokens_seen": 5710320, "step": 21330 }, { "epoch": 5.592398427260813, "grad_norm": 1.181134581565857, "learning_rate": 2.4196097672332533e-05, "loss": 0.2703, "num_input_tokens_seen": 5711888, "step": 21335 }, { "epoch": 5.593709043250327, "grad_norm": 2.5147953033447266, "learning_rate": 2.4184666389421957e-05, "loss": 0.4109, "num_input_tokens_seen": 5713456, "step": 21340 }, { "epoch": 5.595019659239843, "grad_norm": 1.8545939922332764, "learning_rate": 2.417323527715929e-05, "loss": 0.3213, "num_input_tokens_seen": 5714720, "step": 21345 }, { "epoch": 5.5963302752293576, "grad_norm": 1.2307093143463135, "learning_rate": 2.416180433793705e-05, "loss": 0.3684, "num_input_tokens_seen": 5716048, "step": 21350 }, { "epoch": 5.597640891218873, "grad_norm": 1.4052627086639404, "learning_rate": 2.4150373574147715e-05, "loss": 0.3938, "num_input_tokens_seen": 5717840, "step": 21355 }, { "epoch": 5.598951507208388, "grad_norm": 0.7330812215805054, "learning_rate": 2.413894298818373e-05, "loss": 0.2441, "num_input_tokens_seen": 5719184, "step": 21360 }, { "epoch": 5.600262123197903, "grad_norm": 0.7220891118049622, "learning_rate": 2.4127512582437485e-05, "loss": 0.2832, "num_input_tokens_seen": 5720704, "step": 21365 }, { "epoch": 5.601572739187418, "grad_norm": 2.07869291305542, "learning_rate": 2.4116082359301346e-05, "loss": 0.1549, "num_input_tokens_seen": 5722160, "step": 21370 }, { "epoch": 5.6028833551769335, "grad_norm": 1.154119849205017, "learning_rate": 2.4104652321167638e-05, "loss": 0.2727, "num_input_tokens_seen": 5723392, "step": 21375 }, { "epoch": 5.604193971166448, "grad_norm": 0.5035355091094971, "learning_rate": 2.4093222470428657e-05, "loss": 0.2357, "num_input_tokens_seen": 5724528, "step": 21380 }, { "epoch": 5.605504587155964, "grad_norm": 0.7432346940040588, "learning_rate": 2.4081792809476643e-05, "loss": 0.2509, "num_input_tokens_seen": 5725984, "step": 21385 }, { "epoch": 5.606815203145478, "grad_norm": 2.001821994781494, "learning_rate": 2.4070363340703812e-05, "loss": 0.403, "num_input_tokens_seen": 5727280, "step": 21390 }, { "epoch": 5.608125819134994, "grad_norm": 1.8197418451309204, "learning_rate": 2.4058934066502333e-05, "loss": 0.4806, "num_input_tokens_seen": 5728336, "step": 21395 }, { "epoch": 5.6094364351245085, "grad_norm": 2.6696746349334717, "learning_rate": 2.4047504989264318e-05, "loss": 0.2929, "num_input_tokens_seen": 5729600, "step": 21400 }, { "epoch": 5.610747051114023, "grad_norm": 1.5209650993347168, "learning_rate": 2.4036076111381867e-05, "loss": 0.3213, "num_input_tokens_seen": 5731008, "step": 21405 }, { "epoch": 5.612057667103539, "grad_norm": 1.8287805318832397, "learning_rate": 2.4024647435247018e-05, "loss": 0.2996, "num_input_tokens_seen": 5732224, "step": 21410 }, { "epoch": 5.613368283093053, "grad_norm": 1.2796417474746704, "learning_rate": 2.4013218963251778e-05, "loss": 0.3652, "num_input_tokens_seen": 5733760, "step": 21415 }, { "epoch": 5.614678899082569, "grad_norm": 0.693709135055542, "learning_rate": 2.4001790697788107e-05, "loss": 0.2087, "num_input_tokens_seen": 5735264, "step": 21420 }, { "epoch": 5.6159895150720835, "grad_norm": 1.2516529560089111, "learning_rate": 2.3990362641247915e-05, "loss": 0.2172, "num_input_tokens_seen": 5737152, "step": 21425 }, { "epoch": 5.617300131061599, "grad_norm": 1.2723875045776367, "learning_rate": 2.3978934796023076e-05, "loss": 0.3589, "num_input_tokens_seen": 5738400, "step": 21430 }, { "epoch": 5.618610747051114, "grad_norm": 0.8897961974143982, "learning_rate": 2.3967507164505416e-05, "loss": 0.1944, "num_input_tokens_seen": 5739632, "step": 21435 }, { "epoch": 5.619921363040629, "grad_norm": 2.855067729949951, "learning_rate": 2.3956079749086728e-05, "loss": 0.2959, "num_input_tokens_seen": 5740976, "step": 21440 }, { "epoch": 5.621231979030144, "grad_norm": 0.6343970894813538, "learning_rate": 2.3944652552158742e-05, "loss": 0.3127, "num_input_tokens_seen": 5742320, "step": 21445 }, { "epoch": 5.6225425950196595, "grad_norm": 1.4756700992584229, "learning_rate": 2.3933225576113154e-05, "loss": 0.3194, "num_input_tokens_seen": 5743728, "step": 21450 }, { "epoch": 5.623853211009174, "grad_norm": 3.2923998832702637, "learning_rate": 2.392179882334162e-05, "loss": 0.2601, "num_input_tokens_seen": 5744944, "step": 21455 }, { "epoch": 5.62516382699869, "grad_norm": 1.8805774450302124, "learning_rate": 2.3910372296235717e-05, "loss": 0.2283, "num_input_tokens_seen": 5745856, "step": 21460 }, { "epoch": 5.626474442988204, "grad_norm": 3.68729829788208, "learning_rate": 2.3898945997187012e-05, "loss": 0.3543, "num_input_tokens_seen": 5747328, "step": 21465 }, { "epoch": 5.62778505897772, "grad_norm": 1.590311884880066, "learning_rate": 2.3887519928587007e-05, "loss": 0.5428, "num_input_tokens_seen": 5748336, "step": 21470 }, { "epoch": 5.6290956749672345, "grad_norm": 0.9351285696029663, "learning_rate": 2.387609409282716e-05, "loss": 0.299, "num_input_tokens_seen": 5749712, "step": 21475 }, { "epoch": 5.63040629095675, "grad_norm": 1.3518164157867432, "learning_rate": 2.3864668492298884e-05, "loss": 0.2638, "num_input_tokens_seen": 5751328, "step": 21480 }, { "epoch": 5.631716906946265, "grad_norm": 0.6519835591316223, "learning_rate": 2.3853243129393524e-05, "loss": 0.2251, "num_input_tokens_seen": 5752352, "step": 21485 }, { "epoch": 5.63302752293578, "grad_norm": 1.2250027656555176, "learning_rate": 2.384181800650239e-05, "loss": 0.244, "num_input_tokens_seen": 5753856, "step": 21490 }, { "epoch": 5.634338138925295, "grad_norm": 0.8785263895988464, "learning_rate": 2.383039312601675e-05, "loss": 0.2352, "num_input_tokens_seen": 5755216, "step": 21495 }, { "epoch": 5.6356487549148095, "grad_norm": 1.1938610076904297, "learning_rate": 2.3818968490327813e-05, "loss": 0.3087, "num_input_tokens_seen": 5756512, "step": 21500 }, { "epoch": 5.636959370904325, "grad_norm": 1.4809653759002686, "learning_rate": 2.3807544101826722e-05, "loss": 0.2489, "num_input_tokens_seen": 5758160, "step": 21505 }, { "epoch": 5.63826998689384, "grad_norm": 1.9685050249099731, "learning_rate": 2.3796119962904597e-05, "loss": 0.2564, "num_input_tokens_seen": 5759776, "step": 21510 }, { "epoch": 5.639580602883355, "grad_norm": 1.06106698513031, "learning_rate": 2.3784696075952487e-05, "loss": 0.1806, "num_input_tokens_seen": 5760944, "step": 21515 }, { "epoch": 5.64089121887287, "grad_norm": 1.5365785360336304, "learning_rate": 2.3773272443361383e-05, "loss": 0.3025, "num_input_tokens_seen": 5762288, "step": 21520 }, { "epoch": 5.6422018348623855, "grad_norm": 1.437898874282837, "learning_rate": 2.3761849067522235e-05, "loss": 0.2692, "num_input_tokens_seen": 5763488, "step": 21525 }, { "epoch": 5.6435124508519, "grad_norm": 1.7692862749099731, "learning_rate": 2.3750425950825943e-05, "loss": 0.3686, "num_input_tokens_seen": 5764896, "step": 21530 }, { "epoch": 5.644823066841416, "grad_norm": 2.4920334815979004, "learning_rate": 2.3739003095663334e-05, "loss": 0.294, "num_input_tokens_seen": 5766112, "step": 21535 }, { "epoch": 5.64613368283093, "grad_norm": 2.4492099285125732, "learning_rate": 2.3727580504425208e-05, "loss": 0.2418, "num_input_tokens_seen": 5767472, "step": 21540 }, { "epoch": 5.647444298820446, "grad_norm": 1.4949698448181152, "learning_rate": 2.3716158179502277e-05, "loss": 0.3367, "num_input_tokens_seen": 5769056, "step": 21545 }, { "epoch": 5.6487549148099605, "grad_norm": 1.9633846282958984, "learning_rate": 2.3704736123285214e-05, "loss": 0.3503, "num_input_tokens_seen": 5770448, "step": 21550 }, { "epoch": 5.650065530799476, "grad_norm": 0.9257538914680481, "learning_rate": 2.369331433816465e-05, "loss": 0.5642, "num_input_tokens_seen": 5771808, "step": 21555 }, { "epoch": 5.651376146788991, "grad_norm": 1.960965871810913, "learning_rate": 2.3681892826531137e-05, "loss": 0.2769, "num_input_tokens_seen": 5773760, "step": 21560 }, { "epoch": 5.652686762778506, "grad_norm": 1.0345983505249023, "learning_rate": 2.3670471590775174e-05, "loss": 0.2382, "num_input_tokens_seen": 5775584, "step": 21565 }, { "epoch": 5.653997378768021, "grad_norm": 1.335160732269287, "learning_rate": 2.3659050633287217e-05, "loss": 0.2351, "num_input_tokens_seen": 5776768, "step": 21570 }, { "epoch": 5.655307994757536, "grad_norm": 1.4161320924758911, "learning_rate": 2.3647629956457647e-05, "loss": 0.358, "num_input_tokens_seen": 5778336, "step": 21575 }, { "epoch": 5.656618610747051, "grad_norm": 1.373909592628479, "learning_rate": 2.3636209562676784e-05, "loss": 0.3078, "num_input_tokens_seen": 5779952, "step": 21580 }, { "epoch": 5.657929226736567, "grad_norm": 1.1404807567596436, "learning_rate": 2.3624789454334905e-05, "loss": 0.3829, "num_input_tokens_seen": 5781376, "step": 21585 }, { "epoch": 5.659239842726081, "grad_norm": 1.1231058835983276, "learning_rate": 2.3613369633822217e-05, "loss": 0.2929, "num_input_tokens_seen": 5783424, "step": 21590 }, { "epoch": 5.660550458715596, "grad_norm": 1.2758110761642456, "learning_rate": 2.360195010352887e-05, "loss": 0.2929, "num_input_tokens_seen": 5784848, "step": 21595 }, { "epoch": 5.6618610747051115, "grad_norm": 2.629783868789673, "learning_rate": 2.3590530865844952e-05, "loss": 0.3318, "num_input_tokens_seen": 5786128, "step": 21600 }, { "epoch": 5.663171690694626, "grad_norm": 1.6000722646713257, "learning_rate": 2.357911192316047e-05, "loss": 0.2885, "num_input_tokens_seen": 5787440, "step": 21605 }, { "epoch": 5.664482306684142, "grad_norm": 1.3771421909332275, "learning_rate": 2.356769327786543e-05, "loss": 0.2979, "num_input_tokens_seen": 5789056, "step": 21610 }, { "epoch": 5.665792922673656, "grad_norm": 1.175499439239502, "learning_rate": 2.3556274932349703e-05, "loss": 0.3234, "num_input_tokens_seen": 5790416, "step": 21615 }, { "epoch": 5.667103538663172, "grad_norm": 0.7201309204101562, "learning_rate": 2.354485688900314e-05, "loss": 0.3215, "num_input_tokens_seen": 5792000, "step": 21620 }, { "epoch": 5.6684141546526865, "grad_norm": 1.0354862213134766, "learning_rate": 2.3533439150215517e-05, "loss": 0.2315, "num_input_tokens_seen": 5793792, "step": 21625 }, { "epoch": 5.669724770642202, "grad_norm": 1.2057524919509888, "learning_rate": 2.352202171837655e-05, "loss": 0.2833, "num_input_tokens_seen": 5795232, "step": 21630 }, { "epoch": 5.671035386631717, "grad_norm": 0.828131914138794, "learning_rate": 2.3510604595875877e-05, "loss": 0.239, "num_input_tokens_seen": 5796560, "step": 21635 }, { "epoch": 5.672346002621232, "grad_norm": 1.404248833656311, "learning_rate": 2.3499187785103087e-05, "loss": 0.2602, "num_input_tokens_seen": 5797984, "step": 21640 }, { "epoch": 5.673656618610747, "grad_norm": 1.2220226526260376, "learning_rate": 2.3487771288447705e-05, "loss": 0.268, "num_input_tokens_seen": 5799088, "step": 21645 }, { "epoch": 5.674967234600262, "grad_norm": 2.0054972171783447, "learning_rate": 2.3476355108299174e-05, "loss": 0.3393, "num_input_tokens_seen": 5800464, "step": 21650 }, { "epoch": 5.676277850589777, "grad_norm": 1.507490873336792, "learning_rate": 2.3464939247046896e-05, "loss": 0.3583, "num_input_tokens_seen": 5802032, "step": 21655 }, { "epoch": 5.677588466579293, "grad_norm": 0.8536332249641418, "learning_rate": 2.3453523707080165e-05, "loss": 0.3889, "num_input_tokens_seen": 5803424, "step": 21660 }, { "epoch": 5.678899082568807, "grad_norm": 2.325211524963379, "learning_rate": 2.344210849078826e-05, "loss": 0.2468, "num_input_tokens_seen": 5804896, "step": 21665 }, { "epoch": 5.680209698558322, "grad_norm": 8.457039833068848, "learning_rate": 2.343069360056036e-05, "loss": 0.2642, "num_input_tokens_seen": 5806096, "step": 21670 }, { "epoch": 5.6815203145478375, "grad_norm": 2.2145028114318848, "learning_rate": 2.3419279038785575e-05, "loss": 0.4233, "num_input_tokens_seen": 5807616, "step": 21675 }, { "epoch": 5.682830930537353, "grad_norm": 1.1756761074066162, "learning_rate": 2.3407864807852952e-05, "loss": 0.1831, "num_input_tokens_seen": 5808768, "step": 21680 }, { "epoch": 5.684141546526868, "grad_norm": 2.0283219814300537, "learning_rate": 2.3396450910151476e-05, "loss": 0.2109, "num_input_tokens_seen": 5809792, "step": 21685 }, { "epoch": 5.685452162516382, "grad_norm": 0.9125142097473145, "learning_rate": 2.338503734807006e-05, "loss": 0.2758, "num_input_tokens_seen": 5811600, "step": 21690 }, { "epoch": 5.686762778505898, "grad_norm": 1.9588528871536255, "learning_rate": 2.337362412399753e-05, "loss": 0.3285, "num_input_tokens_seen": 5812736, "step": 21695 }, { "epoch": 5.6880733944954125, "grad_norm": 1.3307849168777466, "learning_rate": 2.3362211240322663e-05, "loss": 0.2105, "num_input_tokens_seen": 5813600, "step": 21700 }, { "epoch": 5.689384010484928, "grad_norm": 1.3807462453842163, "learning_rate": 2.335079869943415e-05, "loss": 0.3675, "num_input_tokens_seen": 5814736, "step": 21705 }, { "epoch": 5.690694626474443, "grad_norm": 1.2740672826766968, "learning_rate": 2.333938650372062e-05, "loss": 0.3367, "num_input_tokens_seen": 5816032, "step": 21710 }, { "epoch": 5.692005242463958, "grad_norm": 0.9653410315513611, "learning_rate": 2.3327974655570633e-05, "loss": 0.2066, "num_input_tokens_seen": 5816992, "step": 21715 }, { "epoch": 5.693315858453473, "grad_norm": 3.1443514823913574, "learning_rate": 2.3316563157372645e-05, "loss": 0.3443, "num_input_tokens_seen": 5818032, "step": 21720 }, { "epoch": 5.694626474442988, "grad_norm": 1.7097524404525757, "learning_rate": 2.3305152011515098e-05, "loss": 0.1725, "num_input_tokens_seen": 5819232, "step": 21725 }, { "epoch": 5.695937090432503, "grad_norm": 3.660918712615967, "learning_rate": 2.3293741220386296e-05, "loss": 0.325, "num_input_tokens_seen": 5820608, "step": 21730 }, { "epoch": 5.697247706422019, "grad_norm": 1.9249787330627441, "learning_rate": 2.328233078637451e-05, "loss": 0.2076, "num_input_tokens_seen": 5821952, "step": 21735 }, { "epoch": 5.698558322411533, "grad_norm": 1.2461059093475342, "learning_rate": 2.327092071186792e-05, "loss": 0.389, "num_input_tokens_seen": 5823296, "step": 21740 }, { "epoch": 5.699868938401049, "grad_norm": 1.4009236097335815, "learning_rate": 2.3259510999254645e-05, "loss": 0.232, "num_input_tokens_seen": 5824624, "step": 21745 }, { "epoch": 5.7011795543905635, "grad_norm": 0.9734497666358948, "learning_rate": 2.324810165092271e-05, "loss": 0.246, "num_input_tokens_seen": 5825632, "step": 21750 }, { "epoch": 5.702490170380079, "grad_norm": 1.3825798034667969, "learning_rate": 2.323669266926007e-05, "loss": 0.2503, "num_input_tokens_seen": 5827136, "step": 21755 }, { "epoch": 5.703800786369594, "grad_norm": 2.785872459411621, "learning_rate": 2.3225284056654607e-05, "loss": 0.2228, "num_input_tokens_seen": 5828224, "step": 21760 }, { "epoch": 5.705111402359108, "grad_norm": 0.9600896835327148, "learning_rate": 2.3213875815494122e-05, "loss": 0.2052, "num_input_tokens_seen": 5829360, "step": 21765 }, { "epoch": 5.706422018348624, "grad_norm": 1.126656413078308, "learning_rate": 2.3202467948166343e-05, "loss": 0.3428, "num_input_tokens_seen": 5830640, "step": 21770 }, { "epoch": 5.707732634338139, "grad_norm": 0.8512653112411499, "learning_rate": 2.319106045705891e-05, "loss": 0.1747, "num_input_tokens_seen": 5832224, "step": 21775 }, { "epoch": 5.709043250327654, "grad_norm": 0.9374132752418518, "learning_rate": 2.317965334455941e-05, "loss": 0.2123, "num_input_tokens_seen": 5833712, "step": 21780 }, { "epoch": 5.710353866317169, "grad_norm": 1.788504719734192, "learning_rate": 2.3168246613055317e-05, "loss": 0.3758, "num_input_tokens_seen": 5834880, "step": 21785 }, { "epoch": 5.711664482306684, "grad_norm": 1.5274293422698975, "learning_rate": 2.3156840264934045e-05, "loss": 0.3037, "num_input_tokens_seen": 5836304, "step": 21790 }, { "epoch": 5.712975098296199, "grad_norm": 1.559735894203186, "learning_rate": 2.3145434302582913e-05, "loss": 0.3654, "num_input_tokens_seen": 5837920, "step": 21795 }, { "epoch": 5.714285714285714, "grad_norm": 1.2425915002822876, "learning_rate": 2.313402872838918e-05, "loss": 0.2959, "num_input_tokens_seen": 5841344, "step": 21800 }, { "epoch": 5.715596330275229, "grad_norm": 1.7801851034164429, "learning_rate": 2.312262354474001e-05, "loss": 0.4949, "num_input_tokens_seen": 5842592, "step": 21805 }, { "epoch": 5.716906946264745, "grad_norm": 2.3260529041290283, "learning_rate": 2.3111218754022495e-05, "loss": 0.2188, "num_input_tokens_seen": 5843712, "step": 21810 }, { "epoch": 5.718217562254259, "grad_norm": 1.0423305034637451, "learning_rate": 2.309981435862363e-05, "loss": 0.3124, "num_input_tokens_seen": 5845376, "step": 21815 }, { "epoch": 5.719528178243775, "grad_norm": 2.1461598873138428, "learning_rate": 2.308841036093033e-05, "loss": 0.3019, "num_input_tokens_seen": 5846896, "step": 21820 }, { "epoch": 5.7208387942332894, "grad_norm": 2.0777587890625, "learning_rate": 2.307700676332944e-05, "loss": 0.4765, "num_input_tokens_seen": 5848240, "step": 21825 }, { "epoch": 5.722149410222805, "grad_norm": 0.8475099205970764, "learning_rate": 2.3065603568207715e-05, "loss": 0.2137, "num_input_tokens_seen": 5849456, "step": 21830 }, { "epoch": 5.72346002621232, "grad_norm": 1.8642553091049194, "learning_rate": 2.3054200777951807e-05, "loss": 0.2216, "num_input_tokens_seen": 5850928, "step": 21835 }, { "epoch": 5.724770642201835, "grad_norm": 1.2695940732955933, "learning_rate": 2.3042798394948326e-05, "loss": 0.2526, "num_input_tokens_seen": 5852320, "step": 21840 }, { "epoch": 5.72608125819135, "grad_norm": 0.9926271438598633, "learning_rate": 2.303139642158376e-05, "loss": 0.3429, "num_input_tokens_seen": 5853552, "step": 21845 }, { "epoch": 5.727391874180865, "grad_norm": 1.4325923919677734, "learning_rate": 2.3019994860244523e-05, "loss": 0.5065, "num_input_tokens_seen": 5855040, "step": 21850 }, { "epoch": 5.72870249017038, "grad_norm": 0.9082145690917969, "learning_rate": 2.3008593713316937e-05, "loss": 0.315, "num_input_tokens_seen": 5857152, "step": 21855 }, { "epoch": 5.730013106159895, "grad_norm": 2.904888153076172, "learning_rate": 2.2997192983187245e-05, "loss": 0.4065, "num_input_tokens_seen": 5858272, "step": 21860 }, { "epoch": 5.73132372214941, "grad_norm": 0.9179049134254456, "learning_rate": 2.2985792672241608e-05, "loss": 0.2133, "num_input_tokens_seen": 5859504, "step": 21865 }, { "epoch": 5.732634338138926, "grad_norm": 1.4290448427200317, "learning_rate": 2.2974392782866078e-05, "loss": 0.3413, "num_input_tokens_seen": 5861072, "step": 21870 }, { "epoch": 5.73394495412844, "grad_norm": 2.638882875442505, "learning_rate": 2.296299331744664e-05, "loss": 0.2061, "num_input_tokens_seen": 5862496, "step": 21875 }, { "epoch": 5.735255570117955, "grad_norm": 2.2977142333984375, "learning_rate": 2.2951594278369178e-05, "loss": 0.3492, "num_input_tokens_seen": 5863872, "step": 21880 }, { "epoch": 5.736566186107471, "grad_norm": 1.7513105869293213, "learning_rate": 2.2940195668019497e-05, "loss": 0.4606, "num_input_tokens_seen": 5865216, "step": 21885 }, { "epoch": 5.737876802096985, "grad_norm": 0.8326264023780823, "learning_rate": 2.2928797488783298e-05, "loss": 0.2933, "num_input_tokens_seen": 5866656, "step": 21890 }, { "epoch": 5.739187418086501, "grad_norm": 0.9437413811683655, "learning_rate": 2.2917399743046208e-05, "loss": 0.2456, "num_input_tokens_seen": 5867920, "step": 21895 }, { "epoch": 5.740498034076015, "grad_norm": 1.0345286130905151, "learning_rate": 2.2906002433193757e-05, "loss": 0.2863, "num_input_tokens_seen": 5869472, "step": 21900 }, { "epoch": 5.741808650065531, "grad_norm": 1.2589845657348633, "learning_rate": 2.2894605561611382e-05, "loss": 0.1737, "num_input_tokens_seen": 5870480, "step": 21905 }, { "epoch": 5.743119266055046, "grad_norm": 1.8203274011611938, "learning_rate": 2.288320913068442e-05, "loss": 0.4025, "num_input_tokens_seen": 5871888, "step": 21910 }, { "epoch": 5.744429882044561, "grad_norm": 1.0309844017028809, "learning_rate": 2.2871813142798133e-05, "loss": 0.2712, "num_input_tokens_seen": 5873216, "step": 21915 }, { "epoch": 5.745740498034076, "grad_norm": 1.476834774017334, "learning_rate": 2.2860417600337673e-05, "loss": 0.2576, "num_input_tokens_seen": 5874320, "step": 21920 }, { "epoch": 5.747051114023591, "grad_norm": 1.3081468343734741, "learning_rate": 2.284902250568812e-05, "loss": 0.4306, "num_input_tokens_seen": 5875648, "step": 21925 }, { "epoch": 5.748361730013106, "grad_norm": 2.35640811920166, "learning_rate": 2.2837627861234434e-05, "loss": 0.2571, "num_input_tokens_seen": 5876624, "step": 21930 }, { "epoch": 5.749672346002622, "grad_norm": 1.0627663135528564, "learning_rate": 2.2826233669361504e-05, "loss": 0.1936, "num_input_tokens_seen": 5877760, "step": 21935 }, { "epoch": 5.750982961992136, "grad_norm": 1.498871088027954, "learning_rate": 2.2814839932454103e-05, "loss": 0.4571, "num_input_tokens_seen": 5879152, "step": 21940 }, { "epoch": 5.752293577981652, "grad_norm": 5.5438971519470215, "learning_rate": 2.2803446652896935e-05, "loss": 0.3548, "num_input_tokens_seen": 5880640, "step": 21945 }, { "epoch": 5.753604193971166, "grad_norm": 2.4764668941497803, "learning_rate": 2.279205383307458e-05, "loss": 0.5427, "num_input_tokens_seen": 5883504, "step": 21950 }, { "epoch": 5.754914809960681, "grad_norm": 2.005509853363037, "learning_rate": 2.278066147537155e-05, "loss": 0.309, "num_input_tokens_seen": 5884768, "step": 21955 }, { "epoch": 5.756225425950197, "grad_norm": 1.8143160343170166, "learning_rate": 2.276926958217224e-05, "loss": 0.3372, "num_input_tokens_seen": 5885920, "step": 21960 }, { "epoch": 5.757536041939712, "grad_norm": 1.1276295185089111, "learning_rate": 2.2757878155860948e-05, "loss": 0.3462, "num_input_tokens_seen": 5887952, "step": 21965 }, { "epoch": 5.758846657929227, "grad_norm": 2.6148571968078613, "learning_rate": 2.2746487198821883e-05, "loss": 0.4306, "num_input_tokens_seen": 5889872, "step": 21970 }, { "epoch": 5.760157273918741, "grad_norm": 1.3954148292541504, "learning_rate": 2.2735096713439156e-05, "loss": 0.3039, "num_input_tokens_seen": 5891408, "step": 21975 }, { "epoch": 5.761467889908257, "grad_norm": 2.1414175033569336, "learning_rate": 2.2723706702096774e-05, "loss": 0.4072, "num_input_tokens_seen": 5892448, "step": 21980 }, { "epoch": 5.762778505897772, "grad_norm": 0.9492318630218506, "learning_rate": 2.2712317167178648e-05, "loss": 0.2957, "num_input_tokens_seen": 5894208, "step": 21985 }, { "epoch": 5.764089121887287, "grad_norm": 3.791926383972168, "learning_rate": 2.2700928111068582e-05, "loss": 0.2626, "num_input_tokens_seen": 5895184, "step": 21990 }, { "epoch": 5.765399737876802, "grad_norm": 1.0927481651306152, "learning_rate": 2.2689539536150294e-05, "loss": 0.2704, "num_input_tokens_seen": 5896240, "step": 21995 }, { "epoch": 5.766710353866317, "grad_norm": 2.1720242500305176, "learning_rate": 2.2678151444807384e-05, "loss": 0.4148, "num_input_tokens_seen": 5897584, "step": 22000 }, { "epoch": 5.768020969855832, "grad_norm": 1.3342280387878418, "learning_rate": 2.266676383942336e-05, "loss": 0.3582, "num_input_tokens_seen": 5898800, "step": 22005 }, { "epoch": 5.7693315858453476, "grad_norm": 4.533392906188965, "learning_rate": 2.265537672238164e-05, "loss": 0.3817, "num_input_tokens_seen": 5900080, "step": 22010 }, { "epoch": 5.770642201834862, "grad_norm": 2.7910821437835693, "learning_rate": 2.2643990096065518e-05, "loss": 0.3193, "num_input_tokens_seen": 5901200, "step": 22015 }, { "epoch": 5.771952817824378, "grad_norm": 0.8897469639778137, "learning_rate": 2.2632603962858207e-05, "loss": 0.1486, "num_input_tokens_seen": 5902416, "step": 22020 }, { "epoch": 5.773263433813892, "grad_norm": 0.7089112401008606, "learning_rate": 2.2621218325142788e-05, "loss": 0.412, "num_input_tokens_seen": 5903712, "step": 22025 }, { "epoch": 5.774574049803408, "grad_norm": 1.923475980758667, "learning_rate": 2.2609833185302262e-05, "loss": 0.5405, "num_input_tokens_seen": 5905216, "step": 22030 }, { "epoch": 5.775884665792923, "grad_norm": 1.7022171020507812, "learning_rate": 2.2598448545719525e-05, "loss": 0.3429, "num_input_tokens_seen": 5906576, "step": 22035 }, { "epoch": 5.777195281782438, "grad_norm": 1.102648138999939, "learning_rate": 2.2587064408777352e-05, "loss": 0.2446, "num_input_tokens_seen": 5907648, "step": 22040 }, { "epoch": 5.778505897771953, "grad_norm": 0.6918660402297974, "learning_rate": 2.257568077685844e-05, "loss": 0.3648, "num_input_tokens_seen": 5909440, "step": 22045 }, { "epoch": 5.779816513761467, "grad_norm": 4.849746227264404, "learning_rate": 2.2564297652345347e-05, "loss": 0.3794, "num_input_tokens_seen": 5910432, "step": 22050 }, { "epoch": 5.781127129750983, "grad_norm": 2.188580274581909, "learning_rate": 2.255291503762054e-05, "loss": 0.2458, "num_input_tokens_seen": 5911648, "step": 22055 }, { "epoch": 5.7824377457404985, "grad_norm": 2.578535795211792, "learning_rate": 2.2541532935066394e-05, "loss": 0.3635, "num_input_tokens_seen": 5913168, "step": 22060 }, { "epoch": 5.783748361730013, "grad_norm": 1.7971712350845337, "learning_rate": 2.2530151347065153e-05, "loss": 0.2126, "num_input_tokens_seen": 5914528, "step": 22065 }, { "epoch": 5.785058977719528, "grad_norm": 1.0288854837417603, "learning_rate": 2.2518770275998967e-05, "loss": 0.253, "num_input_tokens_seen": 5915968, "step": 22070 }, { "epoch": 5.786369593709043, "grad_norm": 0.7776690721511841, "learning_rate": 2.2507389724249878e-05, "loss": 0.2678, "num_input_tokens_seen": 5917184, "step": 22075 }, { "epoch": 5.787680209698558, "grad_norm": 1.3917999267578125, "learning_rate": 2.2496009694199815e-05, "loss": 0.3038, "num_input_tokens_seen": 5918496, "step": 22080 }, { "epoch": 5.7889908256880735, "grad_norm": 1.4752827882766724, "learning_rate": 2.2484630188230596e-05, "loss": 0.2365, "num_input_tokens_seen": 5920096, "step": 22085 }, { "epoch": 5.790301441677588, "grad_norm": 1.210536241531372, "learning_rate": 2.2473251208723927e-05, "loss": 0.4099, "num_input_tokens_seen": 5922176, "step": 22090 }, { "epoch": 5.791612057667104, "grad_norm": 3.3471391201019287, "learning_rate": 2.246187275806142e-05, "loss": 0.3993, "num_input_tokens_seen": 5923328, "step": 22095 }, { "epoch": 5.792922673656618, "grad_norm": 1.1658010482788086, "learning_rate": 2.2450494838624554e-05, "loss": 0.3027, "num_input_tokens_seen": 5924912, "step": 22100 }, { "epoch": 5.794233289646134, "grad_norm": 1.9675785303115845, "learning_rate": 2.2439117452794723e-05, "loss": 0.3191, "num_input_tokens_seen": 5926512, "step": 22105 }, { "epoch": 5.795543905635649, "grad_norm": 0.5384573936462402, "learning_rate": 2.2427740602953175e-05, "loss": 0.2594, "num_input_tokens_seen": 5927584, "step": 22110 }, { "epoch": 5.796854521625164, "grad_norm": 2.7668895721435547, "learning_rate": 2.2416364291481075e-05, "loss": 0.3709, "num_input_tokens_seen": 5929040, "step": 22115 }, { "epoch": 5.798165137614679, "grad_norm": 3.646857500076294, "learning_rate": 2.240498852075946e-05, "loss": 0.3461, "num_input_tokens_seen": 5931664, "step": 22120 }, { "epoch": 5.799475753604194, "grad_norm": 1.4164015054702759, "learning_rate": 2.239361329316927e-05, "loss": 0.3156, "num_input_tokens_seen": 5932704, "step": 22125 }, { "epoch": 5.800786369593709, "grad_norm": 2.3589723110198975, "learning_rate": 2.2382238611091318e-05, "loss": 0.2556, "num_input_tokens_seen": 5933936, "step": 22130 }, { "epoch": 5.8020969855832245, "grad_norm": 0.8365362286567688, "learning_rate": 2.2370864476906296e-05, "loss": 0.2864, "num_input_tokens_seen": 5935232, "step": 22135 }, { "epoch": 5.803407601572739, "grad_norm": 3.423438310623169, "learning_rate": 2.23594908929948e-05, "loss": 0.296, "num_input_tokens_seen": 5936400, "step": 22140 }, { "epoch": 5.804718217562254, "grad_norm": 2.8964574337005615, "learning_rate": 2.2348117861737297e-05, "loss": 0.4936, "num_input_tokens_seen": 5937648, "step": 22145 }, { "epoch": 5.806028833551769, "grad_norm": 1.2780601978302002, "learning_rate": 2.233674538551414e-05, "loss": 0.2992, "num_input_tokens_seen": 5939120, "step": 22150 }, { "epoch": 5.807339449541285, "grad_norm": 1.409487247467041, "learning_rate": 2.2325373466705573e-05, "loss": 0.3705, "num_input_tokens_seen": 5940720, "step": 22155 }, { "epoch": 5.8086500655307995, "grad_norm": 1.0259147882461548, "learning_rate": 2.2314002107691727e-05, "loss": 0.3119, "num_input_tokens_seen": 5942320, "step": 22160 }, { "epoch": 5.809960681520314, "grad_norm": 0.9329002499580383, "learning_rate": 2.230263131085259e-05, "loss": 0.256, "num_input_tokens_seen": 5943280, "step": 22165 }, { "epoch": 5.81127129750983, "grad_norm": 0.9876506328582764, "learning_rate": 2.2291261078568057e-05, "loss": 0.4365, "num_input_tokens_seen": 5944672, "step": 22170 }, { "epoch": 5.812581913499344, "grad_norm": 2.0058271884918213, "learning_rate": 2.22798914132179e-05, "loss": 0.267, "num_input_tokens_seen": 5946208, "step": 22175 }, { "epoch": 5.81389252948886, "grad_norm": 1.2021589279174805, "learning_rate": 2.226852231718177e-05, "loss": 0.3189, "num_input_tokens_seen": 5947456, "step": 22180 }, { "epoch": 5.815203145478375, "grad_norm": 1.5750606060028076, "learning_rate": 2.2257153792839202e-05, "loss": 0.3432, "num_input_tokens_seen": 5948560, "step": 22185 }, { "epoch": 5.81651376146789, "grad_norm": 3.837115526199341, "learning_rate": 2.2245785842569604e-05, "loss": 0.3771, "num_input_tokens_seen": 5949504, "step": 22190 }, { "epoch": 5.817824377457405, "grad_norm": 1.0370965003967285, "learning_rate": 2.2234418468752276e-05, "loss": 0.4691, "num_input_tokens_seen": 5950720, "step": 22195 }, { "epoch": 5.81913499344692, "grad_norm": 2.4086124897003174, "learning_rate": 2.222305167376638e-05, "loss": 0.2461, "num_input_tokens_seen": 5951856, "step": 22200 }, { "epoch": 5.820445609436435, "grad_norm": 6.362067222595215, "learning_rate": 2.221168545999097e-05, "loss": 0.2637, "num_input_tokens_seen": 5953184, "step": 22205 }, { "epoch": 5.8217562254259505, "grad_norm": 1.761861801147461, "learning_rate": 2.2200319829804978e-05, "loss": 0.2702, "num_input_tokens_seen": 5954208, "step": 22210 }, { "epoch": 5.823066841415465, "grad_norm": 1.0159358978271484, "learning_rate": 2.2188954785587207e-05, "loss": 0.202, "num_input_tokens_seen": 5955376, "step": 22215 }, { "epoch": 5.824377457404981, "grad_norm": 2.5740532875061035, "learning_rate": 2.2177590329716345e-05, "loss": 0.2851, "num_input_tokens_seen": 5956752, "step": 22220 }, { "epoch": 5.825688073394495, "grad_norm": 1.1980196237564087, "learning_rate": 2.2166226464570947e-05, "loss": 0.1988, "num_input_tokens_seen": 5957760, "step": 22225 }, { "epoch": 5.826998689384011, "grad_norm": 0.9109039306640625, "learning_rate": 2.2154863192529453e-05, "loss": 0.3713, "num_input_tokens_seen": 5959152, "step": 22230 }, { "epoch": 5.8283093053735255, "grad_norm": 2.1873435974121094, "learning_rate": 2.2143500515970173e-05, "loss": 0.23, "num_input_tokens_seen": 5960368, "step": 22235 }, { "epoch": 5.82961992136304, "grad_norm": 1.134504795074463, "learning_rate": 2.21321384372713e-05, "loss": 0.2981, "num_input_tokens_seen": 5962016, "step": 22240 }, { "epoch": 5.830930537352556, "grad_norm": 0.04522417485713959, "learning_rate": 2.2120776958810902e-05, "loss": 0.1759, "num_input_tokens_seen": 5963040, "step": 22245 }, { "epoch": 5.832241153342071, "grad_norm": 1.025832176208496, "learning_rate": 2.2109416082966907e-05, "loss": 0.2561, "num_input_tokens_seen": 5964256, "step": 22250 }, { "epoch": 5.833551769331586, "grad_norm": 2.0656614303588867, "learning_rate": 2.2098055812117143e-05, "loss": 0.2544, "num_input_tokens_seen": 5965200, "step": 22255 }, { "epoch": 5.834862385321101, "grad_norm": 1.3696671724319458, "learning_rate": 2.208669614863927e-05, "loss": 0.3458, "num_input_tokens_seen": 5966784, "step": 22260 }, { "epoch": 5.836173001310616, "grad_norm": 2.5834848880767822, "learning_rate": 2.2075337094910868e-05, "loss": 0.4065, "num_input_tokens_seen": 5967872, "step": 22265 }, { "epoch": 5.837483617300131, "grad_norm": 2.770618200302124, "learning_rate": 2.2063978653309354e-05, "loss": 0.3708, "num_input_tokens_seen": 5969648, "step": 22270 }, { "epoch": 5.838794233289646, "grad_norm": 1.2828247547149658, "learning_rate": 2.2052620826212035e-05, "loss": 0.4079, "num_input_tokens_seen": 5971168, "step": 22275 }, { "epoch": 5.840104849279161, "grad_norm": 1.3308048248291016, "learning_rate": 2.204126361599609e-05, "loss": 0.3766, "num_input_tokens_seen": 5973200, "step": 22280 }, { "epoch": 5.8414154652686765, "grad_norm": 2.389864683151245, "learning_rate": 2.2029907025038548e-05, "loss": 0.2067, "num_input_tokens_seen": 5974224, "step": 22285 }, { "epoch": 5.842726081258191, "grad_norm": 1.2932283878326416, "learning_rate": 2.201855105571634e-05, "loss": 0.3728, "num_input_tokens_seen": 5975504, "step": 22290 }, { "epoch": 5.844036697247707, "grad_norm": 1.613929033279419, "learning_rate": 2.2007195710406236e-05, "loss": 0.2506, "num_input_tokens_seen": 5976608, "step": 22295 }, { "epoch": 5.845347313237221, "grad_norm": 1.6917530298233032, "learning_rate": 2.1995840991484906e-05, "loss": 0.3463, "num_input_tokens_seen": 5977760, "step": 22300 }, { "epoch": 5.846657929226737, "grad_norm": 1.568916916847229, "learning_rate": 2.198448690132886e-05, "loss": 0.256, "num_input_tokens_seen": 5978896, "step": 22305 }, { "epoch": 5.8479685452162515, "grad_norm": 1.167933464050293, "learning_rate": 2.19731334423145e-05, "loss": 0.328, "num_input_tokens_seen": 5980576, "step": 22310 }, { "epoch": 5.849279161205767, "grad_norm": 1.317266583442688, "learning_rate": 2.196178061681809e-05, "loss": 0.3732, "num_input_tokens_seen": 5981632, "step": 22315 }, { "epoch": 5.850589777195282, "grad_norm": 1.0510672330856323, "learning_rate": 2.1950428427215737e-05, "loss": 0.2336, "num_input_tokens_seen": 5983392, "step": 22320 }, { "epoch": 5.851900393184797, "grad_norm": 0.48780885338783264, "learning_rate": 2.1939076875883448e-05, "loss": 0.3157, "num_input_tokens_seen": 5985072, "step": 22325 }, { "epoch": 5.853211009174312, "grad_norm": 1.7657959461212158, "learning_rate": 2.1927725965197074e-05, "loss": 0.4695, "num_input_tokens_seen": 5986368, "step": 22330 }, { "epoch": 5.854521625163827, "grad_norm": 1.4005889892578125, "learning_rate": 2.191637569753235e-05, "loss": 0.303, "num_input_tokens_seen": 5987264, "step": 22335 }, { "epoch": 5.855832241153342, "grad_norm": 2.2351579666137695, "learning_rate": 2.1905026075264873e-05, "loss": 0.25, "num_input_tokens_seen": 5988320, "step": 22340 }, { "epoch": 5.857142857142857, "grad_norm": 1.8753050565719604, "learning_rate": 2.1893677100770084e-05, "loss": 0.2923, "num_input_tokens_seen": 5989600, "step": 22345 }, { "epoch": 5.858453473132372, "grad_norm": 1.5387030839920044, "learning_rate": 2.18823287764233e-05, "loss": 0.4191, "num_input_tokens_seen": 5991280, "step": 22350 }, { "epoch": 5.859764089121887, "grad_norm": 1.8516439199447632, "learning_rate": 2.187098110459973e-05, "loss": 0.2562, "num_input_tokens_seen": 5992416, "step": 22355 }, { "epoch": 5.8610747051114025, "grad_norm": 1.5588432550430298, "learning_rate": 2.1859634087674398e-05, "loss": 0.3085, "num_input_tokens_seen": 5993680, "step": 22360 }, { "epoch": 5.862385321100917, "grad_norm": 0.843021035194397, "learning_rate": 2.1848287728022233e-05, "loss": 0.268, "num_input_tokens_seen": 5994912, "step": 22365 }, { "epoch": 5.863695937090433, "grad_norm": 2.896810531616211, "learning_rate": 2.1836942028018e-05, "loss": 0.3305, "num_input_tokens_seen": 5996336, "step": 22370 }, { "epoch": 5.865006553079947, "grad_norm": 1.6956467628479004, "learning_rate": 2.182559699003634e-05, "loss": 0.2926, "num_input_tokens_seen": 5997504, "step": 22375 }, { "epoch": 5.866317169069463, "grad_norm": 0.7977932691574097, "learning_rate": 2.1814252616451735e-05, "loss": 0.3571, "num_input_tokens_seen": 5999040, "step": 22380 }, { "epoch": 5.8676277850589775, "grad_norm": 1.9836019277572632, "learning_rate": 2.1802908909638558e-05, "loss": 0.2428, "num_input_tokens_seen": 6000336, "step": 22385 }, { "epoch": 5.868938401048493, "grad_norm": 1.2209959030151367, "learning_rate": 2.1791565871971024e-05, "loss": 0.2504, "num_input_tokens_seen": 6001584, "step": 22390 }, { "epoch": 5.870249017038008, "grad_norm": 1.718729019165039, "learning_rate": 2.1780223505823212e-05, "loss": 0.3046, "num_input_tokens_seen": 6002688, "step": 22395 }, { "epoch": 5.871559633027523, "grad_norm": 2.3819398880004883, "learning_rate": 2.1768881813569046e-05, "loss": 0.3958, "num_input_tokens_seen": 6003952, "step": 22400 }, { "epoch": 5.872870249017038, "grad_norm": 1.845976710319519, "learning_rate": 2.175754079758234e-05, "loss": 0.3342, "num_input_tokens_seen": 6005184, "step": 22405 }, { "epoch": 5.8741808650065535, "grad_norm": 1.176844596862793, "learning_rate": 2.1746200460236736e-05, "loss": 0.3204, "num_input_tokens_seen": 6006416, "step": 22410 }, { "epoch": 5.875491480996068, "grad_norm": 1.1258748769760132, "learning_rate": 2.173486080390576e-05, "loss": 0.3183, "num_input_tokens_seen": 6007616, "step": 22415 }, { "epoch": 5.876802096985584, "grad_norm": 2.0158984661102295, "learning_rate": 2.1723521830962774e-05, "loss": 0.3067, "num_input_tokens_seen": 6009040, "step": 22420 }, { "epoch": 5.878112712975098, "grad_norm": 0.740014910697937, "learning_rate": 2.1712183543781005e-05, "loss": 0.2464, "num_input_tokens_seen": 6010144, "step": 22425 }, { "epoch": 5.879423328964613, "grad_norm": 1.7233095169067383, "learning_rate": 2.1700845944733543e-05, "loss": 0.2545, "num_input_tokens_seen": 6011232, "step": 22430 }, { "epoch": 5.8807339449541285, "grad_norm": 2.0287222862243652, "learning_rate": 2.1689509036193333e-05, "loss": 0.2425, "num_input_tokens_seen": 6012096, "step": 22435 }, { "epoch": 5.882044560943643, "grad_norm": 1.4063918590545654, "learning_rate": 2.167817282053315e-05, "loss": 0.398, "num_input_tokens_seen": 6013632, "step": 22440 }, { "epoch": 5.883355176933159, "grad_norm": 3.7948033809661865, "learning_rate": 2.166683730012566e-05, "loss": 0.3393, "num_input_tokens_seen": 6014608, "step": 22445 }, { "epoch": 5.884665792922673, "grad_norm": 0.9378402233123779, "learning_rate": 2.1655502477343363e-05, "loss": 0.2882, "num_input_tokens_seen": 6015824, "step": 22450 }, { "epoch": 5.885976408912189, "grad_norm": 1.9097117185592651, "learning_rate": 2.164416835455862e-05, "loss": 0.3315, "num_input_tokens_seen": 6017440, "step": 22455 }, { "epoch": 5.8872870249017035, "grad_norm": 1.7598260641098022, "learning_rate": 2.1632834934143642e-05, "loss": 0.2467, "num_input_tokens_seen": 6018592, "step": 22460 }, { "epoch": 5.888597640891219, "grad_norm": 1.1772183179855347, "learning_rate": 2.162150221847048e-05, "loss": 0.2465, "num_input_tokens_seen": 6019776, "step": 22465 }, { "epoch": 5.889908256880734, "grad_norm": 0.9620100855827332, "learning_rate": 2.1610170209911085e-05, "loss": 0.4344, "num_input_tokens_seen": 6021056, "step": 22470 }, { "epoch": 5.891218872870249, "grad_norm": 1.6631426811218262, "learning_rate": 2.15988389108372e-05, "loss": 0.2871, "num_input_tokens_seen": 6022528, "step": 22475 }, { "epoch": 5.892529488859764, "grad_norm": 2.165530204772949, "learning_rate": 2.1587508323620452e-05, "loss": 0.3523, "num_input_tokens_seen": 6023872, "step": 22480 }, { "epoch": 5.8938401048492794, "grad_norm": 0.9419733881950378, "learning_rate": 2.1576178450632324e-05, "loss": 0.2838, "num_input_tokens_seen": 6025600, "step": 22485 }, { "epoch": 5.895150720838794, "grad_norm": 1.275312900543213, "learning_rate": 2.1564849294244126e-05, "loss": 0.2497, "num_input_tokens_seen": 6026864, "step": 22490 }, { "epoch": 5.89646133682831, "grad_norm": 0.6835410594940186, "learning_rate": 2.1553520856827038e-05, "loss": 0.3269, "num_input_tokens_seen": 6028016, "step": 22495 }, { "epoch": 5.897771952817824, "grad_norm": 1.4181642532348633, "learning_rate": 2.1542193140752074e-05, "loss": 0.2998, "num_input_tokens_seen": 6029584, "step": 22500 }, { "epoch": 5.89908256880734, "grad_norm": 1.8006480932235718, "learning_rate": 2.1530866148390114e-05, "loss": 0.3353, "num_input_tokens_seen": 6030640, "step": 22505 }, { "epoch": 5.9003931847968545, "grad_norm": 1.9057719707489014, "learning_rate": 2.151953988211187e-05, "loss": 0.3233, "num_input_tokens_seen": 6031904, "step": 22510 }, { "epoch": 5.90170380078637, "grad_norm": 0.9592381715774536, "learning_rate": 2.150821434428793e-05, "loss": 0.2936, "num_input_tokens_seen": 6033584, "step": 22515 }, { "epoch": 5.903014416775885, "grad_norm": 2.070441484451294, "learning_rate": 2.149688953728867e-05, "loss": 0.2408, "num_input_tokens_seen": 6034704, "step": 22520 }, { "epoch": 5.904325032765399, "grad_norm": 2.406811475753784, "learning_rate": 2.1485565463484395e-05, "loss": 0.5053, "num_input_tokens_seen": 6036624, "step": 22525 }, { "epoch": 5.905635648754915, "grad_norm": 1.6500517129898071, "learning_rate": 2.1474242125245192e-05, "loss": 0.3563, "num_input_tokens_seen": 6037840, "step": 22530 }, { "epoch": 5.9069462647444295, "grad_norm": 2.0058183670043945, "learning_rate": 2.1462919524941018e-05, "loss": 0.3628, "num_input_tokens_seen": 6039344, "step": 22535 }, { "epoch": 5.908256880733945, "grad_norm": 1.258126974105835, "learning_rate": 2.145159766494168e-05, "loss": 0.3621, "num_input_tokens_seen": 6040912, "step": 22540 }, { "epoch": 5.90956749672346, "grad_norm": 1.7804046869277954, "learning_rate": 2.1440276547616815e-05, "loss": 0.1955, "num_input_tokens_seen": 6042048, "step": 22545 }, { "epoch": 5.910878112712975, "grad_norm": 1.2644410133361816, "learning_rate": 2.1428956175335922e-05, "loss": 0.3131, "num_input_tokens_seen": 6043328, "step": 22550 }, { "epoch": 5.91218872870249, "grad_norm": 1.767244577407837, "learning_rate": 2.141763655046833e-05, "loss": 0.2326, "num_input_tokens_seen": 6044288, "step": 22555 }, { "epoch": 5.913499344692005, "grad_norm": 1.6923768520355225, "learning_rate": 2.140631767538322e-05, "loss": 0.2345, "num_input_tokens_seen": 6045392, "step": 22560 }, { "epoch": 5.91480996068152, "grad_norm": 0.9197333455085754, "learning_rate": 2.1394999552449608e-05, "loss": 0.2942, "num_input_tokens_seen": 6046688, "step": 22565 }, { "epoch": 5.916120576671036, "grad_norm": 2.0731358528137207, "learning_rate": 2.1383682184036357e-05, "loss": 0.4185, "num_input_tokens_seen": 6047584, "step": 22570 }, { "epoch": 5.91743119266055, "grad_norm": 3.7055201530456543, "learning_rate": 2.1372365572512188e-05, "loss": 0.3632, "num_input_tokens_seen": 6048912, "step": 22575 }, { "epoch": 5.918741808650066, "grad_norm": 2.876371383666992, "learning_rate": 2.136104972024562e-05, "loss": 0.2758, "num_input_tokens_seen": 6049856, "step": 22580 }, { "epoch": 5.9200524246395805, "grad_norm": 4.0817437171936035, "learning_rate": 2.134973462960507e-05, "loss": 0.2697, "num_input_tokens_seen": 6051152, "step": 22585 }, { "epoch": 5.921363040629096, "grad_norm": 1.1927992105484009, "learning_rate": 2.133842030295875e-05, "loss": 0.1963, "num_input_tokens_seen": 6052672, "step": 22590 }, { "epoch": 5.922673656618611, "grad_norm": 1.3904850482940674, "learning_rate": 2.1327106742674735e-05, "loss": 0.3338, "num_input_tokens_seen": 6053840, "step": 22595 }, { "epoch": 5.923984272608125, "grad_norm": 1.6878644227981567, "learning_rate": 2.1315793951120934e-05, "loss": 0.3408, "num_input_tokens_seen": 6055696, "step": 22600 }, { "epoch": 5.925294888597641, "grad_norm": 1.466389775276184, "learning_rate": 2.1304481930665095e-05, "loss": 0.3122, "num_input_tokens_seen": 6056720, "step": 22605 }, { "epoch": 5.926605504587156, "grad_norm": 2.001713514328003, "learning_rate": 2.1293170683674803e-05, "loss": 0.3074, "num_input_tokens_seen": 6058080, "step": 22610 }, { "epoch": 5.927916120576671, "grad_norm": 1.5633223056793213, "learning_rate": 2.1281860212517477e-05, "loss": 0.37, "num_input_tokens_seen": 6059616, "step": 22615 }, { "epoch": 5.929226736566186, "grad_norm": 0.7426407337188721, "learning_rate": 2.1270550519560387e-05, "loss": 0.2933, "num_input_tokens_seen": 6061408, "step": 22620 }, { "epoch": 5.930537352555701, "grad_norm": 1.0303096771240234, "learning_rate": 2.125924160717063e-05, "loss": 0.2178, "num_input_tokens_seen": 6062864, "step": 22625 }, { "epoch": 5.931847968545216, "grad_norm": 1.4672282934188843, "learning_rate": 2.1247933477715145e-05, "loss": 0.3086, "num_input_tokens_seen": 6064784, "step": 22630 }, { "epoch": 5.933158584534731, "grad_norm": 0.9438732862472534, "learning_rate": 2.123662613356068e-05, "loss": 0.2457, "num_input_tokens_seen": 6066464, "step": 22635 }, { "epoch": 5.934469200524246, "grad_norm": 1.900701642036438, "learning_rate": 2.1225319577073877e-05, "loss": 0.2601, "num_input_tokens_seen": 6067584, "step": 22640 }, { "epoch": 5.935779816513762, "grad_norm": 0.6984930038452148, "learning_rate": 2.121401381062117e-05, "loss": 0.3741, "num_input_tokens_seen": 6068672, "step": 22645 }, { "epoch": 5.937090432503276, "grad_norm": 0.6236879229545593, "learning_rate": 2.1202708836568826e-05, "loss": 0.2219, "num_input_tokens_seen": 6070352, "step": 22650 }, { "epoch": 5.938401048492792, "grad_norm": 0.9309603571891785, "learning_rate": 2.1191404657282957e-05, "loss": 0.342, "num_input_tokens_seen": 6071648, "step": 22655 }, { "epoch": 5.9397116644823065, "grad_norm": 2.26261830329895, "learning_rate": 2.1180101275129515e-05, "loss": 0.2182, "num_input_tokens_seen": 6072528, "step": 22660 }, { "epoch": 5.941022280471822, "grad_norm": 1.0180822610855103, "learning_rate": 2.1168798692474277e-05, "loss": 0.2417, "num_input_tokens_seen": 6073744, "step": 22665 }, { "epoch": 5.942332896461337, "grad_norm": 1.552895426750183, "learning_rate": 2.115749691168286e-05, "loss": 0.2568, "num_input_tokens_seen": 6074848, "step": 22670 }, { "epoch": 5.943643512450852, "grad_norm": 1.1321470737457275, "learning_rate": 2.114619593512069e-05, "loss": 0.2596, "num_input_tokens_seen": 6076128, "step": 22675 }, { "epoch": 5.944954128440367, "grad_norm": 1.5146031379699707, "learning_rate": 2.1134895765153057e-05, "loss": 0.2274, "num_input_tokens_seen": 6077280, "step": 22680 }, { "epoch": 5.946264744429882, "grad_norm": 1.946197748184204, "learning_rate": 2.112359640414506e-05, "loss": 0.3827, "num_input_tokens_seen": 6078768, "step": 22685 }, { "epoch": 5.947575360419397, "grad_norm": 1.0857242345809937, "learning_rate": 2.111229785446164e-05, "loss": 0.3341, "num_input_tokens_seen": 6080528, "step": 22690 }, { "epoch": 5.948885976408912, "grad_norm": 3.3897130489349365, "learning_rate": 2.110100011846755e-05, "loss": 0.3438, "num_input_tokens_seen": 6081792, "step": 22695 }, { "epoch": 5.950196592398427, "grad_norm": 1.5815824270248413, "learning_rate": 2.1089703198527407e-05, "loss": 0.2112, "num_input_tokens_seen": 6083136, "step": 22700 }, { "epoch": 5.951507208387943, "grad_norm": 2.5152413845062256, "learning_rate": 2.1078407097005632e-05, "loss": 0.2751, "num_input_tokens_seen": 6084240, "step": 22705 }, { "epoch": 5.952817824377457, "grad_norm": 0.6482627391815186, "learning_rate": 2.106711181626647e-05, "loss": 0.3395, "num_input_tokens_seen": 6086384, "step": 22710 }, { "epoch": 5.954128440366972, "grad_norm": 1.3603384494781494, "learning_rate": 2.105581735867401e-05, "loss": 0.309, "num_input_tokens_seen": 6087456, "step": 22715 }, { "epoch": 5.955439056356488, "grad_norm": 1.4782334566116333, "learning_rate": 2.1044523726592165e-05, "loss": 0.2283, "num_input_tokens_seen": 6088464, "step": 22720 }, { "epoch": 5.956749672346002, "grad_norm": 1.7213712930679321, "learning_rate": 2.1033230922384672e-05, "loss": 0.2712, "num_input_tokens_seen": 6090464, "step": 22725 }, { "epoch": 5.958060288335518, "grad_norm": 3.4148929119110107, "learning_rate": 2.1021938948415086e-05, "loss": 0.1888, "num_input_tokens_seen": 6091360, "step": 22730 }, { "epoch": 5.9593709043250325, "grad_norm": 2.704771041870117, "learning_rate": 2.1010647807046806e-05, "loss": 0.2344, "num_input_tokens_seen": 6092912, "step": 22735 }, { "epoch": 5.960681520314548, "grad_norm": 2.8111116886138916, "learning_rate": 2.099935750064305e-05, "loss": 0.3027, "num_input_tokens_seen": 6094384, "step": 22740 }, { "epoch": 5.961992136304063, "grad_norm": 1.0825709104537964, "learning_rate": 2.0988068031566854e-05, "loss": 0.389, "num_input_tokens_seen": 6095616, "step": 22745 }, { "epoch": 5.963302752293578, "grad_norm": 2.937549352645874, "learning_rate": 2.097677940218108e-05, "loss": 0.2791, "num_input_tokens_seen": 6096720, "step": 22750 }, { "epoch": 5.964613368283093, "grad_norm": 1.186156988143921, "learning_rate": 2.0965491614848433e-05, "loss": 0.4278, "num_input_tokens_seen": 6098592, "step": 22755 }, { "epoch": 5.965923984272608, "grad_norm": 1.5962685346603394, "learning_rate": 2.095420467193142e-05, "loss": 0.3612, "num_input_tokens_seen": 6099920, "step": 22760 }, { "epoch": 5.967234600262123, "grad_norm": 1.2310409545898438, "learning_rate": 2.0942918575792382e-05, "loss": 0.3882, "num_input_tokens_seen": 6101232, "step": 22765 }, { "epoch": 5.968545216251639, "grad_norm": 1.4049311876296997, "learning_rate": 2.0931633328793472e-05, "loss": 0.2604, "num_input_tokens_seen": 6102608, "step": 22770 }, { "epoch": 5.969855832241153, "grad_norm": 1.3499363660812378, "learning_rate": 2.0920348933296678e-05, "loss": 0.2941, "num_input_tokens_seen": 6104928, "step": 22775 }, { "epoch": 5.971166448230669, "grad_norm": 1.7695424556732178, "learning_rate": 2.09090653916638e-05, "loss": 0.3903, "num_input_tokens_seen": 6106096, "step": 22780 }, { "epoch": 5.972477064220183, "grad_norm": 1.169029951095581, "learning_rate": 2.0897782706256474e-05, "loss": 0.3081, "num_input_tokens_seen": 6107584, "step": 22785 }, { "epoch": 5.973787680209698, "grad_norm": 1.2921189069747925, "learning_rate": 2.0886500879436136e-05, "loss": 0.391, "num_input_tokens_seen": 6109232, "step": 22790 }, { "epoch": 5.975098296199214, "grad_norm": 1.1840955018997192, "learning_rate": 2.0875219913564053e-05, "loss": 0.3208, "num_input_tokens_seen": 6110384, "step": 22795 }, { "epoch": 5.976408912188729, "grad_norm": 0.9171605706214905, "learning_rate": 2.0863939811001316e-05, "loss": 0.2381, "num_input_tokens_seen": 6111840, "step": 22800 }, { "epoch": 5.977719528178244, "grad_norm": 1.1586036682128906, "learning_rate": 2.0852660574108833e-05, "loss": 0.2877, "num_input_tokens_seen": 6113504, "step": 22805 }, { "epoch": 5.9790301441677585, "grad_norm": 1.5134283304214478, "learning_rate": 2.084138220524732e-05, "loss": 0.2366, "num_input_tokens_seen": 6114576, "step": 22810 }, { "epoch": 5.980340760157274, "grad_norm": 1.9248579740524292, "learning_rate": 2.0830104706777325e-05, "loss": 0.2853, "num_input_tokens_seen": 6115632, "step": 22815 }, { "epoch": 5.981651376146789, "grad_norm": 1.0289345979690552, "learning_rate": 2.081882808105922e-05, "loss": 0.2607, "num_input_tokens_seen": 6117200, "step": 22820 }, { "epoch": 5.982961992136304, "grad_norm": 3.4398691654205322, "learning_rate": 2.0807552330453166e-05, "loss": 0.2401, "num_input_tokens_seen": 6118256, "step": 22825 }, { "epoch": 5.984272608125819, "grad_norm": 1.0764249563217163, "learning_rate": 2.079627745731917e-05, "loss": 0.3379, "num_input_tokens_seen": 6119840, "step": 22830 }, { "epoch": 5.985583224115334, "grad_norm": 0.9589481949806213, "learning_rate": 2.0785003464017032e-05, "loss": 0.2777, "num_input_tokens_seen": 6121280, "step": 22835 }, { "epoch": 5.986893840104849, "grad_norm": 1.390449047088623, "learning_rate": 2.0773730352906394e-05, "loss": 0.2289, "num_input_tokens_seen": 6122624, "step": 22840 }, { "epoch": 5.988204456094365, "grad_norm": 1.2812846899032593, "learning_rate": 2.076245812634669e-05, "loss": 0.2842, "num_input_tokens_seen": 6123504, "step": 22845 }, { "epoch": 5.989515072083879, "grad_norm": 0.3759801387786865, "learning_rate": 2.0751186786697175e-05, "loss": 0.2683, "num_input_tokens_seen": 6124496, "step": 22850 }, { "epoch": 5.990825688073395, "grad_norm": 2.350432872772217, "learning_rate": 2.0739916336316924e-05, "loss": 0.3591, "num_input_tokens_seen": 6125888, "step": 22855 }, { "epoch": 5.992136304062909, "grad_norm": 1.044498085975647, "learning_rate": 2.072864677756482e-05, "loss": 0.3921, "num_input_tokens_seen": 6127520, "step": 22860 }, { "epoch": 5.993446920052425, "grad_norm": 1.5146089792251587, "learning_rate": 2.0717378112799563e-05, "loss": 0.2755, "num_input_tokens_seen": 6128992, "step": 22865 }, { "epoch": 5.99475753604194, "grad_norm": 3.8892273902893066, "learning_rate": 2.070611034437967e-05, "loss": 0.2858, "num_input_tokens_seen": 6130128, "step": 22870 }, { "epoch": 5.996068152031455, "grad_norm": 2.0816168785095215, "learning_rate": 2.0694843474663465e-05, "loss": 0.226, "num_input_tokens_seen": 6131616, "step": 22875 }, { "epoch": 5.99737876802097, "grad_norm": 1.158579707145691, "learning_rate": 2.0683577506009087e-05, "loss": 0.3461, "num_input_tokens_seen": 6133120, "step": 22880 }, { "epoch": 5.9986893840104845, "grad_norm": 1.7067307233810425, "learning_rate": 2.0672312440774473e-05, "loss": 0.1675, "num_input_tokens_seen": 6134224, "step": 22885 }, { "epoch": 6.0, "grad_norm": 3.0496950149536133, "learning_rate": 2.0661048281317388e-05, "loss": 0.3055, "num_input_tokens_seen": 6135704, "step": 22890 }, { "epoch": 6.001310615989515, "grad_norm": 0.9247754812240601, "learning_rate": 2.06497850299954e-05, "loss": 0.2025, "num_input_tokens_seen": 6137080, "step": 22895 }, { "epoch": 6.001572739187418, "eval_loss": 0.5994988083839417, "eval_runtime": 16.7025, "eval_samples_per_second": 50.771, "eval_steps_per_second": 25.385, "num_input_tokens_seen": 6137320, "step": 22896 }, { "epoch": 6.00262123197903, "grad_norm": 1.7780745029449463, "learning_rate": 2.063852268916589e-05, "loss": 0.2058, "num_input_tokens_seen": 6138248, "step": 22900 }, { "epoch": 6.003931847968545, "grad_norm": 0.9727507829666138, "learning_rate": 2.062726126118605e-05, "loss": 0.206, "num_input_tokens_seen": 6139960, "step": 22905 }, { "epoch": 6.00524246395806, "grad_norm": 0.7458198070526123, "learning_rate": 2.0616000748412864e-05, "loss": 0.2473, "num_input_tokens_seen": 6141400, "step": 22910 }, { "epoch": 6.006553079947575, "grad_norm": 2.4211833477020264, "learning_rate": 2.060474115320315e-05, "loss": 0.1879, "num_input_tokens_seen": 6142536, "step": 22915 }, { "epoch": 6.007863695937091, "grad_norm": 3.107539176940918, "learning_rate": 2.059348247791352e-05, "loss": 0.3079, "num_input_tokens_seen": 6143800, "step": 22920 }, { "epoch": 6.009174311926605, "grad_norm": 2.536679983139038, "learning_rate": 2.058222472490038e-05, "loss": 0.2565, "num_input_tokens_seen": 6144936, "step": 22925 }, { "epoch": 6.010484927916121, "grad_norm": 1.2277942895889282, "learning_rate": 2.0570967896519983e-05, "loss": 0.2468, "num_input_tokens_seen": 6146504, "step": 22930 }, { "epoch": 6.011795543905635, "grad_norm": 1.191809058189392, "learning_rate": 2.0559711995128353e-05, "loss": 0.2623, "num_input_tokens_seen": 6147464, "step": 22935 }, { "epoch": 6.013106159895151, "grad_norm": 1.4322420358657837, "learning_rate": 2.0548457023081333e-05, "loss": 0.1588, "num_input_tokens_seen": 6148952, "step": 22940 }, { "epoch": 6.014416775884666, "grad_norm": 1.0610461235046387, "learning_rate": 2.0537202982734564e-05, "loss": 0.171, "num_input_tokens_seen": 6150152, "step": 22945 }, { "epoch": 6.015727391874181, "grad_norm": 1.9310803413391113, "learning_rate": 2.0525949876443497e-05, "loss": 0.2592, "num_input_tokens_seen": 6151592, "step": 22950 }, { "epoch": 6.017038007863696, "grad_norm": 2.2980880737304688, "learning_rate": 2.051469770656339e-05, "loss": 0.2369, "num_input_tokens_seen": 6152920, "step": 22955 }, { "epoch": 6.018348623853211, "grad_norm": 4.21765661239624, "learning_rate": 2.050344647544931e-05, "loss": 0.3105, "num_input_tokens_seen": 6154696, "step": 22960 }, { "epoch": 6.019659239842726, "grad_norm": 2.085530996322632, "learning_rate": 2.0492196185456108e-05, "loss": 0.2922, "num_input_tokens_seen": 6156200, "step": 22965 }, { "epoch": 6.0209698558322415, "grad_norm": 2.1921024322509766, "learning_rate": 2.0480946838938455e-05, "loss": 0.1975, "num_input_tokens_seen": 6157496, "step": 22970 }, { "epoch": 6.022280471821756, "grad_norm": 4.3539137840271, "learning_rate": 2.046969843825082e-05, "loss": 0.2123, "num_input_tokens_seen": 6158584, "step": 22975 }, { "epoch": 6.023591087811272, "grad_norm": 1.8683215379714966, "learning_rate": 2.0458450985747468e-05, "loss": 0.2158, "num_input_tokens_seen": 6160120, "step": 22980 }, { "epoch": 6.024901703800786, "grad_norm": 1.4988722801208496, "learning_rate": 2.0447204483782482e-05, "loss": 0.2339, "num_input_tokens_seen": 6161240, "step": 22985 }, { "epoch": 6.026212319790301, "grad_norm": 1.7570724487304688, "learning_rate": 2.0435958934709738e-05, "loss": 0.1755, "num_input_tokens_seen": 6162408, "step": 22990 }, { "epoch": 6.027522935779817, "grad_norm": 0.9486292004585266, "learning_rate": 2.0424714340882893e-05, "loss": 0.135, "num_input_tokens_seen": 6163704, "step": 22995 }, { "epoch": 6.028833551769331, "grad_norm": 0.9947705268859863, "learning_rate": 2.0413470704655445e-05, "loss": 0.2499, "num_input_tokens_seen": 6164968, "step": 23000 }, { "epoch": 6.030144167758847, "grad_norm": 0.5718144178390503, "learning_rate": 2.0402228028380644e-05, "loss": 0.2636, "num_input_tokens_seen": 6168184, "step": 23005 }, { "epoch": 6.031454783748361, "grad_norm": 1.4893466234207153, "learning_rate": 2.0390986314411573e-05, "loss": 0.0855, "num_input_tokens_seen": 6169304, "step": 23010 }, { "epoch": 6.032765399737877, "grad_norm": 0.1827617883682251, "learning_rate": 2.03797455651011e-05, "loss": 0.1915, "num_input_tokens_seen": 6170328, "step": 23015 }, { "epoch": 6.034076015727392, "grad_norm": 2.4341421127319336, "learning_rate": 2.0368505782801904e-05, "loss": 0.2008, "num_input_tokens_seen": 6171112, "step": 23020 }, { "epoch": 6.035386631716907, "grad_norm": 5.257660865783691, "learning_rate": 2.0357266969866443e-05, "loss": 0.2447, "num_input_tokens_seen": 6172376, "step": 23025 }, { "epoch": 6.036697247706422, "grad_norm": 2.645627975463867, "learning_rate": 2.034602912864698e-05, "loss": 0.265, "num_input_tokens_seen": 6173800, "step": 23030 }, { "epoch": 6.038007863695937, "grad_norm": 1.6898442506790161, "learning_rate": 2.0334792261495576e-05, "loss": 0.2602, "num_input_tokens_seen": 6175176, "step": 23035 }, { "epoch": 6.039318479685452, "grad_norm": 2.11633038520813, "learning_rate": 2.032355637076409e-05, "loss": 0.2342, "num_input_tokens_seen": 6176856, "step": 23040 }, { "epoch": 6.0406290956749675, "grad_norm": 1.8042981624603271, "learning_rate": 2.031232145880418e-05, "loss": 0.277, "num_input_tokens_seen": 6178072, "step": 23045 }, { "epoch": 6.041939711664482, "grad_norm": 2.7152342796325684, "learning_rate": 2.030108752796729e-05, "loss": 0.2194, "num_input_tokens_seen": 6179224, "step": 23050 }, { "epoch": 6.043250327653998, "grad_norm": 1.6915545463562012, "learning_rate": 2.0289854580604667e-05, "loss": 0.2066, "num_input_tokens_seen": 6180312, "step": 23055 }, { "epoch": 6.044560943643512, "grad_norm": 1.2682957649230957, "learning_rate": 2.0278622619067334e-05, "loss": 0.2352, "num_input_tokens_seen": 6181896, "step": 23060 }, { "epoch": 6.045871559633028, "grad_norm": 1.2331777811050415, "learning_rate": 2.026739164570613e-05, "loss": 0.2356, "num_input_tokens_seen": 6183272, "step": 23065 }, { "epoch": 6.047182175622543, "grad_norm": 1.8118054866790771, "learning_rate": 2.0256161662871675e-05, "loss": 0.208, "num_input_tokens_seen": 6184648, "step": 23070 }, { "epoch": 6.048492791612058, "grad_norm": 2.3742597103118896, "learning_rate": 2.0244932672914392e-05, "loss": 0.2136, "num_input_tokens_seen": 6185656, "step": 23075 }, { "epoch": 6.049803407601573, "grad_norm": 1.9109985828399658, "learning_rate": 2.0233704678184485e-05, "loss": 0.132, "num_input_tokens_seen": 6186760, "step": 23080 }, { "epoch": 6.051114023591087, "grad_norm": 1.2164453268051147, "learning_rate": 2.0222477681031947e-05, "loss": 0.2454, "num_input_tokens_seen": 6188008, "step": 23085 }, { "epoch": 6.052424639580603, "grad_norm": 2.5910959243774414, "learning_rate": 2.021125168380658e-05, "loss": 0.276, "num_input_tokens_seen": 6189944, "step": 23090 }, { "epoch": 6.053735255570118, "grad_norm": 0.5451599359512329, "learning_rate": 2.0200026688857955e-05, "loss": 0.1362, "num_input_tokens_seen": 6190952, "step": 23095 }, { "epoch": 6.055045871559633, "grad_norm": 1.030251145362854, "learning_rate": 2.0188802698535455e-05, "loss": 0.2833, "num_input_tokens_seen": 6192840, "step": 23100 }, { "epoch": 6.056356487549148, "grad_norm": 1.6892142295837402, "learning_rate": 2.0177579715188234e-05, "loss": 0.1571, "num_input_tokens_seen": 6193944, "step": 23105 }, { "epoch": 6.057667103538663, "grad_norm": 1.5088753700256348, "learning_rate": 2.016635774116525e-05, "loss": 0.2789, "num_input_tokens_seen": 6195368, "step": 23110 }, { "epoch": 6.058977719528178, "grad_norm": 1.4492833614349365, "learning_rate": 2.0155136778815246e-05, "loss": 0.2343, "num_input_tokens_seen": 6196536, "step": 23115 }, { "epoch": 6.0602883355176935, "grad_norm": 2.089726448059082, "learning_rate": 2.0143916830486734e-05, "loss": 0.2268, "num_input_tokens_seen": 6197896, "step": 23120 }, { "epoch": 6.061598951507208, "grad_norm": 1.703696370124817, "learning_rate": 2.0132697898528037e-05, "loss": 0.186, "num_input_tokens_seen": 6199112, "step": 23125 }, { "epoch": 6.062909567496724, "grad_norm": 2.4885382652282715, "learning_rate": 2.0121479985287264e-05, "loss": 0.249, "num_input_tokens_seen": 6200760, "step": 23130 }, { "epoch": 6.064220183486238, "grad_norm": 1.887976884841919, "learning_rate": 2.01102630931123e-05, "loss": 0.1913, "num_input_tokens_seen": 6201816, "step": 23135 }, { "epoch": 6.065530799475754, "grad_norm": 1.0896891355514526, "learning_rate": 2.0099047224350825e-05, "loss": 0.3996, "num_input_tokens_seen": 6203512, "step": 23140 }, { "epoch": 6.066841415465269, "grad_norm": 1.7686148881912231, "learning_rate": 2.0087832381350295e-05, "loss": 0.2305, "num_input_tokens_seen": 6204712, "step": 23145 }, { "epoch": 6.068152031454784, "grad_norm": 4.3917927742004395, "learning_rate": 2.007661856645796e-05, "loss": 0.2308, "num_input_tokens_seen": 6205640, "step": 23150 }, { "epoch": 6.069462647444299, "grad_norm": 1.3477911949157715, "learning_rate": 2.0065405782020844e-05, "loss": 0.1734, "num_input_tokens_seen": 6206520, "step": 23155 }, { "epoch": 6.070773263433814, "grad_norm": 3.116642951965332, "learning_rate": 2.0054194030385774e-05, "loss": 0.2447, "num_input_tokens_seen": 6207656, "step": 23160 }, { "epoch": 6.072083879423329, "grad_norm": 2.210923671722412, "learning_rate": 2.0042983313899356e-05, "loss": 0.23, "num_input_tokens_seen": 6208792, "step": 23165 }, { "epoch": 6.073394495412844, "grad_norm": 3.383868455886841, "learning_rate": 2.0031773634907962e-05, "loss": 0.2856, "num_input_tokens_seen": 6210456, "step": 23170 }, { "epoch": 6.074705111402359, "grad_norm": 0.7445526719093323, "learning_rate": 2.0020564995757765e-05, "loss": 0.1314, "num_input_tokens_seen": 6211512, "step": 23175 }, { "epoch": 6.076015727391874, "grad_norm": 1.2462213039398193, "learning_rate": 2.0009357398794707e-05, "loss": 0.2272, "num_input_tokens_seen": 6213032, "step": 23180 }, { "epoch": 6.077326343381389, "grad_norm": 1.0494025945663452, "learning_rate": 1.999815084636452e-05, "loss": 0.1803, "num_input_tokens_seen": 6214152, "step": 23185 }, { "epoch": 6.078636959370904, "grad_norm": 1.2873364686965942, "learning_rate": 1.9986945340812722e-05, "loss": 0.2104, "num_input_tokens_seen": 6215656, "step": 23190 }, { "epoch": 6.0799475753604195, "grad_norm": 2.0015969276428223, "learning_rate": 1.9975740884484606e-05, "loss": 0.3842, "num_input_tokens_seen": 6216840, "step": 23195 }, { "epoch": 6.081258191349934, "grad_norm": 1.1611076593399048, "learning_rate": 1.9964537479725234e-05, "loss": 0.2107, "num_input_tokens_seen": 6218024, "step": 23200 }, { "epoch": 6.08256880733945, "grad_norm": 3.083024263381958, "learning_rate": 1.9953335128879468e-05, "loss": 0.2062, "num_input_tokens_seen": 6219256, "step": 23205 }, { "epoch": 6.083879423328964, "grad_norm": 1.055400013923645, "learning_rate": 1.9942133834291936e-05, "loss": 0.2704, "num_input_tokens_seen": 6220872, "step": 23210 }, { "epoch": 6.08519003931848, "grad_norm": 2.977884531021118, "learning_rate": 1.9930933598307058e-05, "loss": 0.1882, "num_input_tokens_seen": 6222216, "step": 23215 }, { "epoch": 6.086500655307995, "grad_norm": 2.821376323699951, "learning_rate": 1.991973442326902e-05, "loss": 0.1782, "num_input_tokens_seen": 6223368, "step": 23220 }, { "epoch": 6.08781127129751, "grad_norm": 1.5630364418029785, "learning_rate": 1.990853631152179e-05, "loss": 0.2155, "num_input_tokens_seen": 6224696, "step": 23225 }, { "epoch": 6.089121887287025, "grad_norm": 3.335427761077881, "learning_rate": 1.989733926540911e-05, "loss": 0.2108, "num_input_tokens_seen": 6225768, "step": 23230 }, { "epoch": 6.09043250327654, "grad_norm": 1.6145172119140625, "learning_rate": 1.988614328727451e-05, "loss": 0.1673, "num_input_tokens_seen": 6226808, "step": 23235 }, { "epoch": 6.091743119266055, "grad_norm": 1.454304814338684, "learning_rate": 1.9874948379461277e-05, "loss": 0.3434, "num_input_tokens_seen": 6228248, "step": 23240 }, { "epoch": 6.0930537352555705, "grad_norm": 0.7989770174026489, "learning_rate": 1.9863754544312495e-05, "loss": 0.1561, "num_input_tokens_seen": 6230104, "step": 23245 }, { "epoch": 6.094364351245085, "grad_norm": 8.595321655273438, "learning_rate": 1.9852561784171014e-05, "loss": 0.1987, "num_input_tokens_seen": 6231448, "step": 23250 }, { "epoch": 6.095674967234601, "grad_norm": 3.7822201251983643, "learning_rate": 1.9841370101379463e-05, "loss": 0.291, "num_input_tokens_seen": 6232664, "step": 23255 }, { "epoch": 6.096985583224115, "grad_norm": 0.8293286561965942, "learning_rate": 1.983017949828023e-05, "loss": 0.2692, "num_input_tokens_seen": 6234280, "step": 23260 }, { "epoch": 6.09829619921363, "grad_norm": 2.4832561016082764, "learning_rate": 1.981898997721548e-05, "loss": 0.4301, "num_input_tokens_seen": 6235528, "step": 23265 }, { "epoch": 6.0996068152031455, "grad_norm": 2.1517810821533203, "learning_rate": 1.98078015405272e-05, "loss": 0.1877, "num_input_tokens_seen": 6236536, "step": 23270 }, { "epoch": 6.10091743119266, "grad_norm": 1.4630705118179321, "learning_rate": 1.9796614190557074e-05, "loss": 0.2889, "num_input_tokens_seen": 6238088, "step": 23275 }, { "epoch": 6.102228047182176, "grad_norm": 1.3763991594314575, "learning_rate": 1.978542792964661e-05, "loss": 0.216, "num_input_tokens_seen": 6239336, "step": 23280 }, { "epoch": 6.10353866317169, "grad_norm": 1.923537254333496, "learning_rate": 1.977424276013707e-05, "loss": 0.2814, "num_input_tokens_seen": 6240680, "step": 23285 }, { "epoch": 6.104849279161206, "grad_norm": 1.8193836212158203, "learning_rate": 1.9763058684369495e-05, "loss": 0.2051, "num_input_tokens_seen": 6242296, "step": 23290 }, { "epoch": 6.1061598951507206, "grad_norm": 1.8255839347839355, "learning_rate": 1.9751875704684684e-05, "loss": 0.1154, "num_input_tokens_seen": 6243880, "step": 23295 }, { "epoch": 6.107470511140236, "grad_norm": 3.022209644317627, "learning_rate": 1.974069382342322e-05, "loss": 0.3403, "num_input_tokens_seen": 6245160, "step": 23300 }, { "epoch": 6.108781127129751, "grad_norm": 1.2186864614486694, "learning_rate": 1.9729513042925455e-05, "loss": 0.2323, "num_input_tokens_seen": 6246280, "step": 23305 }, { "epoch": 6.110091743119266, "grad_norm": 5.980576038360596, "learning_rate": 1.97183333655315e-05, "loss": 0.1299, "num_input_tokens_seen": 6247416, "step": 23310 }, { "epoch": 6.111402359108781, "grad_norm": 1.4596282243728638, "learning_rate": 1.970715479358125e-05, "loss": 0.3576, "num_input_tokens_seen": 6248440, "step": 23315 }, { "epoch": 6.1127129750982965, "grad_norm": 1.2243062257766724, "learning_rate": 1.9695977329414352e-05, "loss": 0.1373, "num_input_tokens_seen": 6249800, "step": 23320 }, { "epoch": 6.114023591087811, "grad_norm": 1.28800630569458, "learning_rate": 1.9684800975370224e-05, "loss": 0.1346, "num_input_tokens_seen": 6250856, "step": 23325 }, { "epoch": 6.115334207077327, "grad_norm": 1.208818793296814, "learning_rate": 1.9673625733788084e-05, "loss": 0.2537, "num_input_tokens_seen": 6252376, "step": 23330 }, { "epoch": 6.116644823066841, "grad_norm": 2.617971897125244, "learning_rate": 1.966245160700687e-05, "loss": 0.1299, "num_input_tokens_seen": 6253592, "step": 23335 }, { "epoch": 6.117955439056357, "grad_norm": 2.444141149520874, "learning_rate": 1.965127859736531e-05, "loss": 0.1529, "num_input_tokens_seen": 6255208, "step": 23340 }, { "epoch": 6.1192660550458715, "grad_norm": 3.099005937576294, "learning_rate": 1.96401067072019e-05, "loss": 0.2242, "num_input_tokens_seen": 6256216, "step": 23345 }, { "epoch": 6.120576671035387, "grad_norm": 1.7820981740951538, "learning_rate": 1.9628935938854897e-05, "loss": 0.1682, "num_input_tokens_seen": 6257752, "step": 23350 }, { "epoch": 6.121887287024902, "grad_norm": 1.6129531860351562, "learning_rate": 1.961776629466232e-05, "loss": 0.1855, "num_input_tokens_seen": 6258904, "step": 23355 }, { "epoch": 6.123197903014416, "grad_norm": 1.8081200122833252, "learning_rate": 1.9606597776961955e-05, "loss": 0.2506, "num_input_tokens_seen": 6260120, "step": 23360 }, { "epoch": 6.124508519003932, "grad_norm": 2.0522286891937256, "learning_rate": 1.9595430388091358e-05, "loss": 0.1594, "num_input_tokens_seen": 6261448, "step": 23365 }, { "epoch": 6.1258191349934465, "grad_norm": 2.265688896179199, "learning_rate": 1.958426413038784e-05, "loss": 0.2237, "num_input_tokens_seen": 6262728, "step": 23370 }, { "epoch": 6.127129750982962, "grad_norm": 1.9122334718704224, "learning_rate": 1.957309900618849e-05, "loss": 0.2599, "num_input_tokens_seen": 6264216, "step": 23375 }, { "epoch": 6.128440366972477, "grad_norm": 2.087097644805908, "learning_rate": 1.956193501783012e-05, "loss": 0.2199, "num_input_tokens_seen": 6265464, "step": 23380 }, { "epoch": 6.129750982961992, "grad_norm": 1.6881234645843506, "learning_rate": 1.9550772167649378e-05, "loss": 0.153, "num_input_tokens_seen": 6266680, "step": 23385 }, { "epoch": 6.131061598951507, "grad_norm": 4.945633411407471, "learning_rate": 1.9539610457982596e-05, "loss": 0.4317, "num_input_tokens_seen": 6267640, "step": 23390 }, { "epoch": 6.1323722149410225, "grad_norm": 1.478943109512329, "learning_rate": 1.9528449891165913e-05, "loss": 0.2513, "num_input_tokens_seen": 6269080, "step": 23395 }, { "epoch": 6.133682830930537, "grad_norm": 2.3020598888397217, "learning_rate": 1.951729046953521e-05, "loss": 0.2022, "num_input_tokens_seen": 6270360, "step": 23400 }, { "epoch": 6.134993446920053, "grad_norm": 2.844237804412842, "learning_rate": 1.9506132195426138e-05, "loss": 0.2662, "num_input_tokens_seen": 6271720, "step": 23405 }, { "epoch": 6.136304062909567, "grad_norm": 2.1211001873016357, "learning_rate": 1.9494975071174112e-05, "loss": 0.2509, "num_input_tokens_seen": 6273352, "step": 23410 }, { "epoch": 6.137614678899083, "grad_norm": 2.111917734146118, "learning_rate": 1.9483819099114286e-05, "loss": 0.2668, "num_input_tokens_seen": 6274808, "step": 23415 }, { "epoch": 6.1389252948885975, "grad_norm": 2.052447557449341, "learning_rate": 1.9472664281581594e-05, "loss": 0.2719, "num_input_tokens_seen": 6275704, "step": 23420 }, { "epoch": 6.140235910878113, "grad_norm": 1.6535288095474243, "learning_rate": 1.9461510620910713e-05, "loss": 0.2197, "num_input_tokens_seen": 6277400, "step": 23425 }, { "epoch": 6.141546526867628, "grad_norm": 0.9617841839790344, "learning_rate": 1.94503581194361e-05, "loss": 0.2557, "num_input_tokens_seen": 6278696, "step": 23430 }, { "epoch": 6.142857142857143, "grad_norm": 1.5443767309188843, "learning_rate": 1.9439206779491934e-05, "loss": 0.2073, "num_input_tokens_seen": 6280024, "step": 23435 }, { "epoch": 6.144167758846658, "grad_norm": 2.6362287998199463, "learning_rate": 1.942805660341217e-05, "loss": 0.3252, "num_input_tokens_seen": 6281048, "step": 23440 }, { "epoch": 6.145478374836173, "grad_norm": 0.7500015497207642, "learning_rate": 1.9416907593530548e-05, "loss": 0.1613, "num_input_tokens_seen": 6282360, "step": 23445 }, { "epoch": 6.146788990825688, "grad_norm": 2.1943881511688232, "learning_rate": 1.9405759752180515e-05, "loss": 0.091, "num_input_tokens_seen": 6283080, "step": 23450 }, { "epoch": 6.148099606815203, "grad_norm": 1.647273302078247, "learning_rate": 1.9394613081695302e-05, "loss": 0.1625, "num_input_tokens_seen": 6284824, "step": 23455 }, { "epoch": 6.149410222804718, "grad_norm": 5.003935813903809, "learning_rate": 1.9383467584407883e-05, "loss": 0.2155, "num_input_tokens_seen": 6285896, "step": 23460 }, { "epoch": 6.150720838794233, "grad_norm": 3.6027138233184814, "learning_rate": 1.9372323262650992e-05, "loss": 0.2955, "num_input_tokens_seen": 6287080, "step": 23465 }, { "epoch": 6.1520314547837485, "grad_norm": 2.2102253437042236, "learning_rate": 1.936118011875713e-05, "loss": 0.2181, "num_input_tokens_seen": 6288088, "step": 23470 }, { "epoch": 6.153342070773263, "grad_norm": 2.3659448623657227, "learning_rate": 1.9350038155058515e-05, "loss": 0.2508, "num_input_tokens_seen": 6289176, "step": 23475 }, { "epoch": 6.154652686762779, "grad_norm": 1.5780726671218872, "learning_rate": 1.9338897373887154e-05, "loss": 0.2357, "num_input_tokens_seen": 6290376, "step": 23480 }, { "epoch": 6.155963302752293, "grad_norm": 3.026690721511841, "learning_rate": 1.932775777757479e-05, "loss": 0.2889, "num_input_tokens_seen": 6291688, "step": 23485 }, { "epoch": 6.157273918741809, "grad_norm": 1.0133432149887085, "learning_rate": 1.9316619368452928e-05, "loss": 0.1581, "num_input_tokens_seen": 6292792, "step": 23490 }, { "epoch": 6.1585845347313235, "grad_norm": 2.0317025184631348, "learning_rate": 1.9305482148852794e-05, "loss": 0.1968, "num_input_tokens_seen": 6294184, "step": 23495 }, { "epoch": 6.159895150720839, "grad_norm": 3.672166347503662, "learning_rate": 1.929434612110542e-05, "loss": 0.209, "num_input_tokens_seen": 6295384, "step": 23500 }, { "epoch": 6.161205766710354, "grad_norm": 2.8412928581237793, "learning_rate": 1.9283211287541548e-05, "loss": 0.2424, "num_input_tokens_seen": 6296568, "step": 23505 }, { "epoch": 6.162516382699869, "grad_norm": 1.803467035293579, "learning_rate": 1.9272077650491673e-05, "loss": 0.2384, "num_input_tokens_seen": 6297880, "step": 23510 }, { "epoch": 6.163826998689384, "grad_norm": 2.763880729675293, "learning_rate": 1.9260945212286047e-05, "loss": 0.3544, "num_input_tokens_seen": 6298824, "step": 23515 }, { "epoch": 6.165137614678899, "grad_norm": 1.5195196866989136, "learning_rate": 1.9249813975254673e-05, "loss": 0.1445, "num_input_tokens_seen": 6299960, "step": 23520 }, { "epoch": 6.166448230668414, "grad_norm": 2.1381888389587402, "learning_rate": 1.92386839417273e-05, "loss": 0.216, "num_input_tokens_seen": 6301272, "step": 23525 }, { "epoch": 6.16775884665793, "grad_norm": 2.1650636196136475, "learning_rate": 1.9227555114033422e-05, "loss": 0.2841, "num_input_tokens_seen": 6302504, "step": 23530 }, { "epoch": 6.169069462647444, "grad_norm": 2.2650644779205322, "learning_rate": 1.921642749450228e-05, "loss": 0.2343, "num_input_tokens_seen": 6303832, "step": 23535 }, { "epoch": 6.17038007863696, "grad_norm": 1.05382239818573, "learning_rate": 1.9205301085462877e-05, "loss": 0.2295, "num_input_tokens_seen": 6305336, "step": 23540 }, { "epoch": 6.1716906946264745, "grad_norm": 2.6359760761260986, "learning_rate": 1.919417588924394e-05, "loss": 0.3011, "num_input_tokens_seen": 6306568, "step": 23545 }, { "epoch": 6.173001310615989, "grad_norm": 1.5257558822631836, "learning_rate": 1.9183051908173963e-05, "loss": 0.2259, "num_input_tokens_seen": 6307880, "step": 23550 }, { "epoch": 6.174311926605505, "grad_norm": 0.6757317781448364, "learning_rate": 1.9171929144581156e-05, "loss": 0.1982, "num_input_tokens_seen": 6309448, "step": 23555 }, { "epoch": 6.175622542595019, "grad_norm": 1.2013474702835083, "learning_rate": 1.9160807600793517e-05, "loss": 0.2092, "num_input_tokens_seen": 6310888, "step": 23560 }, { "epoch": 6.176933158584535, "grad_norm": 2.396584987640381, "learning_rate": 1.9149687279138768e-05, "loss": 0.3557, "num_input_tokens_seen": 6312344, "step": 23565 }, { "epoch": 6.1782437745740495, "grad_norm": 2.802947521209717, "learning_rate": 1.9138568181944357e-05, "loss": 0.3335, "num_input_tokens_seen": 6313704, "step": 23570 }, { "epoch": 6.179554390563565, "grad_norm": 2.6237246990203857, "learning_rate": 1.9127450311537494e-05, "loss": 0.229, "num_input_tokens_seen": 6315016, "step": 23575 }, { "epoch": 6.18086500655308, "grad_norm": 2.025775909423828, "learning_rate": 1.9116333670245135e-05, "loss": 0.1708, "num_input_tokens_seen": 6315992, "step": 23580 }, { "epoch": 6.182175622542595, "grad_norm": 3.538409471511841, "learning_rate": 1.9105218260393975e-05, "loss": 0.2383, "num_input_tokens_seen": 6317320, "step": 23585 }, { "epoch": 6.18348623853211, "grad_norm": 2.400808095932007, "learning_rate": 1.9094104084310446e-05, "loss": 0.4849, "num_input_tokens_seen": 6318424, "step": 23590 }, { "epoch": 6.184796854521625, "grad_norm": 4.256579399108887, "learning_rate": 1.908299114432073e-05, "loss": 0.2519, "num_input_tokens_seen": 6319736, "step": 23595 }, { "epoch": 6.18610747051114, "grad_norm": 1.8736729621887207, "learning_rate": 1.907187944275074e-05, "loss": 0.1795, "num_input_tokens_seen": 6320904, "step": 23600 }, { "epoch": 6.187418086500656, "grad_norm": 12.827842712402344, "learning_rate": 1.9060768981926137e-05, "loss": 0.365, "num_input_tokens_seen": 6321896, "step": 23605 }, { "epoch": 6.18872870249017, "grad_norm": 2.5936026573181152, "learning_rate": 1.904965976417232e-05, "loss": 0.1849, "num_input_tokens_seen": 6323176, "step": 23610 }, { "epoch": 6.190039318479686, "grad_norm": 2.3901100158691406, "learning_rate": 1.9038551791814446e-05, "loss": 0.2217, "num_input_tokens_seen": 6324776, "step": 23615 }, { "epoch": 6.1913499344692005, "grad_norm": 1.2902779579162598, "learning_rate": 1.902744506717738e-05, "loss": 0.2537, "num_input_tokens_seen": 6326152, "step": 23620 }, { "epoch": 6.192660550458716, "grad_norm": 1.5970227718353271, "learning_rate": 1.9016339592585737e-05, "loss": 0.1392, "num_input_tokens_seen": 6327112, "step": 23625 }, { "epoch": 6.193971166448231, "grad_norm": 1.4757474660873413, "learning_rate": 1.9005235370363877e-05, "loss": 0.3227, "num_input_tokens_seen": 6328616, "step": 23630 }, { "epoch": 6.195281782437746, "grad_norm": 2.9789626598358154, "learning_rate": 1.89941324028359e-05, "loss": 0.2195, "num_input_tokens_seen": 6329736, "step": 23635 }, { "epoch": 6.196592398427261, "grad_norm": 1.4470077753067017, "learning_rate": 1.8983030692325632e-05, "loss": 0.1977, "num_input_tokens_seen": 6330984, "step": 23640 }, { "epoch": 6.1979030144167755, "grad_norm": 1.6329295635223389, "learning_rate": 1.8971930241156646e-05, "loss": 0.205, "num_input_tokens_seen": 6332168, "step": 23645 }, { "epoch": 6.199213630406291, "grad_norm": 0.8929493427276611, "learning_rate": 1.8960831051652243e-05, "loss": 0.1782, "num_input_tokens_seen": 6333320, "step": 23650 }, { "epoch": 6.200524246395806, "grad_norm": 1.2304115295410156, "learning_rate": 1.8949733126135464e-05, "loss": 0.2395, "num_input_tokens_seen": 6334712, "step": 23655 }, { "epoch": 6.201834862385321, "grad_norm": 1.569131851196289, "learning_rate": 1.8938636466929088e-05, "loss": 0.2766, "num_input_tokens_seen": 6335752, "step": 23660 }, { "epoch": 6.203145478374836, "grad_norm": 1.0415278673171997, "learning_rate": 1.8927541076355626e-05, "loss": 0.1808, "num_input_tokens_seen": 6336872, "step": 23665 }, { "epoch": 6.204456094364351, "grad_norm": 2.3485546112060547, "learning_rate": 1.8916446956737317e-05, "loss": 0.4657, "num_input_tokens_seen": 6338296, "step": 23670 }, { "epoch": 6.205766710353866, "grad_norm": 1.6441831588745117, "learning_rate": 1.890535411039616e-05, "loss": 0.2532, "num_input_tokens_seen": 6339560, "step": 23675 }, { "epoch": 6.207077326343382, "grad_norm": 5.327146053314209, "learning_rate": 1.8894262539653858e-05, "loss": 0.2059, "num_input_tokens_seen": 6340744, "step": 23680 }, { "epoch": 6.208387942332896, "grad_norm": 1.4604032039642334, "learning_rate": 1.888317224683185e-05, "loss": 0.2012, "num_input_tokens_seen": 6342040, "step": 23685 }, { "epoch": 6.209698558322412, "grad_norm": 2.1814229488372803, "learning_rate": 1.8872083234251324e-05, "loss": 0.2137, "num_input_tokens_seen": 6343000, "step": 23690 }, { "epoch": 6.2110091743119265, "grad_norm": 1.6021459102630615, "learning_rate": 1.8860995504233188e-05, "loss": 0.2318, "num_input_tokens_seen": 6343992, "step": 23695 }, { "epoch": 6.212319790301442, "grad_norm": 1.9387896060943604, "learning_rate": 1.8849909059098087e-05, "loss": 0.3788, "num_input_tokens_seen": 6345640, "step": 23700 }, { "epoch": 6.213630406290957, "grad_norm": 1.1846922636032104, "learning_rate": 1.8838823901166396e-05, "loss": 0.208, "num_input_tokens_seen": 6347128, "step": 23705 }, { "epoch": 6.214941022280472, "grad_norm": 1.43584144115448, "learning_rate": 1.8827740032758215e-05, "loss": 0.2698, "num_input_tokens_seen": 6348872, "step": 23710 }, { "epoch": 6.216251638269987, "grad_norm": 1.2124840021133423, "learning_rate": 1.8816657456193376e-05, "loss": 0.2233, "num_input_tokens_seen": 6350536, "step": 23715 }, { "epoch": 6.217562254259502, "grad_norm": 0.9596245884895325, "learning_rate": 1.880557617379145e-05, "loss": 0.1429, "num_input_tokens_seen": 6352200, "step": 23720 }, { "epoch": 6.218872870249017, "grad_norm": 3.5399649143218994, "learning_rate": 1.8794496187871725e-05, "loss": 0.2496, "num_input_tokens_seen": 6353528, "step": 23725 }, { "epoch": 6.220183486238533, "grad_norm": 5.11917781829834, "learning_rate": 1.8783417500753224e-05, "loss": 0.2827, "num_input_tokens_seen": 6355016, "step": 23730 }, { "epoch": 6.221494102228047, "grad_norm": 1.2735427618026733, "learning_rate": 1.8772340114754705e-05, "loss": 0.1533, "num_input_tokens_seen": 6356536, "step": 23735 }, { "epoch": 6.222804718217562, "grad_norm": 1.745728850364685, "learning_rate": 1.8761264032194642e-05, "loss": 0.1955, "num_input_tokens_seen": 6358200, "step": 23740 }, { "epoch": 6.224115334207077, "grad_norm": 1.2870686054229736, "learning_rate": 1.875018925539123e-05, "loss": 0.1441, "num_input_tokens_seen": 6359240, "step": 23745 }, { "epoch": 6.225425950196592, "grad_norm": 3.7927165031433105, "learning_rate": 1.8739115786662415e-05, "loss": 0.2233, "num_input_tokens_seen": 6361096, "step": 23750 }, { "epoch": 6.226736566186108, "grad_norm": 3.089327812194824, "learning_rate": 1.872804362832584e-05, "loss": 0.1827, "num_input_tokens_seen": 6362200, "step": 23755 }, { "epoch": 6.228047182175622, "grad_norm": 0.9229252934455872, "learning_rate": 1.8716972782698905e-05, "loss": 0.1166, "num_input_tokens_seen": 6363448, "step": 23760 }, { "epoch": 6.229357798165138, "grad_norm": 1.842820644378662, "learning_rate": 1.8705903252098704e-05, "loss": 0.1709, "num_input_tokens_seen": 6364376, "step": 23765 }, { "epoch": 6.2306684141546524, "grad_norm": 1.2118839025497437, "learning_rate": 1.869483503884208e-05, "loss": 0.273, "num_input_tokens_seen": 6365688, "step": 23770 }, { "epoch": 6.231979030144168, "grad_norm": 1.4971134662628174, "learning_rate": 1.8683768145245586e-05, "loss": 0.1719, "num_input_tokens_seen": 6366808, "step": 23775 }, { "epoch": 6.233289646133683, "grad_norm": 1.8931725025177002, "learning_rate": 1.8672702573625505e-05, "loss": 0.3098, "num_input_tokens_seen": 6368296, "step": 23780 }, { "epoch": 6.234600262123198, "grad_norm": 1.7305455207824707, "learning_rate": 1.866163832629784e-05, "loss": 0.2679, "num_input_tokens_seen": 6369224, "step": 23785 }, { "epoch": 6.235910878112713, "grad_norm": 2.173326015472412, "learning_rate": 1.8650575405578324e-05, "loss": 0.2092, "num_input_tokens_seen": 6370232, "step": 23790 }, { "epoch": 6.237221494102228, "grad_norm": 1.552901029586792, "learning_rate": 1.8639513813782407e-05, "loss": 0.2247, "num_input_tokens_seen": 6371336, "step": 23795 }, { "epoch": 6.238532110091743, "grad_norm": 2.2920312881469727, "learning_rate": 1.8628453553225262e-05, "loss": 0.2598, "num_input_tokens_seen": 6372824, "step": 23800 }, { "epoch": 6.239842726081259, "grad_norm": 2.1994969844818115, "learning_rate": 1.8617394626221774e-05, "loss": 0.2398, "num_input_tokens_seen": 6374168, "step": 23805 }, { "epoch": 6.241153342070773, "grad_norm": 1.2005432844161987, "learning_rate": 1.8606337035086564e-05, "loss": 0.2483, "num_input_tokens_seen": 6375576, "step": 23810 }, { "epoch": 6.242463958060289, "grad_norm": 1.2242989540100098, "learning_rate": 1.859528078213396e-05, "loss": 0.3228, "num_input_tokens_seen": 6376584, "step": 23815 }, { "epoch": 6.243774574049803, "grad_norm": 1.4676826000213623, "learning_rate": 1.8584225869678033e-05, "loss": 0.2179, "num_input_tokens_seen": 6377752, "step": 23820 }, { "epoch": 6.245085190039318, "grad_norm": 1.7195942401885986, "learning_rate": 1.8573172300032538e-05, "loss": 0.2809, "num_input_tokens_seen": 6378776, "step": 23825 }, { "epoch": 6.246395806028834, "grad_norm": 1.2856121063232422, "learning_rate": 1.8562120075510976e-05, "loss": 0.148, "num_input_tokens_seen": 6379976, "step": 23830 }, { "epoch": 6.247706422018348, "grad_norm": 3.007185220718384, "learning_rate": 1.8551069198426556e-05, "loss": 0.1612, "num_input_tokens_seen": 6381768, "step": 23835 }, { "epoch": 6.249017038007864, "grad_norm": 1.9614125490188599, "learning_rate": 1.854001967109221e-05, "loss": 0.4135, "num_input_tokens_seen": 6383256, "step": 23840 }, { "epoch": 6.250327653997378, "grad_norm": 1.312628984451294, "learning_rate": 1.8528971495820584e-05, "loss": 0.2067, "num_input_tokens_seen": 6384744, "step": 23845 }, { "epoch": 6.251638269986894, "grad_norm": 2.2830889225006104, "learning_rate": 1.8517924674924048e-05, "loss": 0.2475, "num_input_tokens_seen": 6386136, "step": 23850 }, { "epoch": 6.252948885976409, "grad_norm": 1.4618113040924072, "learning_rate": 1.8506879210714685e-05, "loss": 0.1425, "num_input_tokens_seen": 6387160, "step": 23855 }, { "epoch": 6.254259501965924, "grad_norm": 0.9106518626213074, "learning_rate": 1.849583510550427e-05, "loss": 0.1391, "num_input_tokens_seen": 6389128, "step": 23860 }, { "epoch": 6.255570117955439, "grad_norm": 4.664388656616211, "learning_rate": 1.8484792361604336e-05, "loss": 0.2443, "num_input_tokens_seen": 6390120, "step": 23865 }, { "epoch": 6.256880733944954, "grad_norm": 1.5369030237197876, "learning_rate": 1.84737509813261e-05, "loss": 0.1984, "num_input_tokens_seen": 6391544, "step": 23870 }, { "epoch": 6.258191349934469, "grad_norm": 1.3597623109817505, "learning_rate": 1.8462710966980506e-05, "loss": 0.2124, "num_input_tokens_seen": 6392792, "step": 23875 }, { "epoch": 6.259501965923985, "grad_norm": 2.2725863456726074, "learning_rate": 1.845167232087822e-05, "loss": 0.2372, "num_input_tokens_seen": 6394248, "step": 23880 }, { "epoch": 6.260812581913499, "grad_norm": 1.1381410360336304, "learning_rate": 1.8440635045329595e-05, "loss": 0.1613, "num_input_tokens_seen": 6395416, "step": 23885 }, { "epoch": 6.262123197903015, "grad_norm": 3.111097812652588, "learning_rate": 1.8429599142644724e-05, "loss": 0.2057, "num_input_tokens_seen": 6396536, "step": 23890 }, { "epoch": 6.263433813892529, "grad_norm": 1.9054431915283203, "learning_rate": 1.84185646151334e-05, "loss": 0.2182, "num_input_tokens_seen": 6397704, "step": 23895 }, { "epoch": 6.264744429882045, "grad_norm": 4.20647668838501, "learning_rate": 1.840753146510512e-05, "loss": 0.2586, "num_input_tokens_seen": 6398840, "step": 23900 }, { "epoch": 6.26605504587156, "grad_norm": 1.529327154159546, "learning_rate": 1.8396499694869123e-05, "loss": 0.2393, "num_input_tokens_seen": 6400264, "step": 23905 }, { "epoch": 6.267365661861074, "grad_norm": 2.19577956199646, "learning_rate": 1.8385469306734327e-05, "loss": 0.2158, "num_input_tokens_seen": 6401224, "step": 23910 }, { "epoch": 6.26867627785059, "grad_norm": 2.0123841762542725, "learning_rate": 1.8374440303009382e-05, "loss": 0.2783, "num_input_tokens_seen": 6403080, "step": 23915 }, { "epoch": 6.269986893840104, "grad_norm": 2.0985045433044434, "learning_rate": 1.836341268600263e-05, "loss": 0.1981, "num_input_tokens_seen": 6404280, "step": 23920 }, { "epoch": 6.27129750982962, "grad_norm": 2.1559488773345947, "learning_rate": 1.8352386458022135e-05, "loss": 0.245, "num_input_tokens_seen": 6405656, "step": 23925 }, { "epoch": 6.272608125819135, "grad_norm": 1.9318557977676392, "learning_rate": 1.8341361621375666e-05, "loss": 0.2551, "num_input_tokens_seen": 6406936, "step": 23930 }, { "epoch": 6.27391874180865, "grad_norm": 1.376910924911499, "learning_rate": 1.8330338178370706e-05, "loss": 0.2232, "num_input_tokens_seen": 6408216, "step": 23935 }, { "epoch": 6.275229357798165, "grad_norm": 1.663450837135315, "learning_rate": 1.8319316131314447e-05, "loss": 0.1907, "num_input_tokens_seen": 6409208, "step": 23940 }, { "epoch": 6.27653997378768, "grad_norm": 0.9535619020462036, "learning_rate": 1.830829548251377e-05, "loss": 0.2292, "num_input_tokens_seen": 6411704, "step": 23945 }, { "epoch": 6.277850589777195, "grad_norm": 1.3526182174682617, "learning_rate": 1.829727623427529e-05, "loss": 0.2587, "num_input_tokens_seen": 6413256, "step": 23950 }, { "epoch": 6.2791612057667106, "grad_norm": 0.8010938167572021, "learning_rate": 1.8286258388905308e-05, "loss": 0.1567, "num_input_tokens_seen": 6414440, "step": 23955 }, { "epoch": 6.280471821756225, "grad_norm": 1.9810678958892822, "learning_rate": 1.8275241948709847e-05, "loss": 0.2176, "num_input_tokens_seen": 6415336, "step": 23960 }, { "epoch": 6.281782437745741, "grad_norm": 1.6247448921203613, "learning_rate": 1.8264226915994627e-05, "loss": 0.1535, "num_input_tokens_seen": 6417224, "step": 23965 }, { "epoch": 6.283093053735255, "grad_norm": 1.4778156280517578, "learning_rate": 1.8253213293065077e-05, "loss": 0.2508, "num_input_tokens_seen": 6418456, "step": 23970 }, { "epoch": 6.284403669724771, "grad_norm": 2.283013105392456, "learning_rate": 1.8242201082226332e-05, "loss": 0.3832, "num_input_tokens_seen": 6419768, "step": 23975 }, { "epoch": 6.285714285714286, "grad_norm": 1.7543927431106567, "learning_rate": 1.8231190285783216e-05, "loss": 0.4498, "num_input_tokens_seen": 6421128, "step": 23980 }, { "epoch": 6.287024901703801, "grad_norm": 1.9676792621612549, "learning_rate": 1.822018090604028e-05, "loss": 0.2873, "num_input_tokens_seen": 6422712, "step": 23985 }, { "epoch": 6.288335517693316, "grad_norm": 1.2209227085113525, "learning_rate": 1.8209172945301768e-05, "loss": 0.2635, "num_input_tokens_seen": 6423608, "step": 23990 }, { "epoch": 6.289646133682831, "grad_norm": 1.4692747592926025, "learning_rate": 1.819816640587162e-05, "loss": 0.2124, "num_input_tokens_seen": 6424888, "step": 23995 }, { "epoch": 6.290956749672346, "grad_norm": 4.582939624786377, "learning_rate": 1.8187161290053498e-05, "loss": 0.2573, "num_input_tokens_seen": 6425896, "step": 24000 }, { "epoch": 6.292267365661861, "grad_norm": 1.3160076141357422, "learning_rate": 1.817615760015074e-05, "loss": 0.1953, "num_input_tokens_seen": 6427144, "step": 24005 }, { "epoch": 6.293577981651376, "grad_norm": 1.120308518409729, "learning_rate": 1.8165155338466404e-05, "loss": 0.1492, "num_input_tokens_seen": 6428424, "step": 24010 }, { "epoch": 6.294888597640891, "grad_norm": 1.7292041778564453, "learning_rate": 1.8154154507303245e-05, "loss": 0.1656, "num_input_tokens_seen": 6429848, "step": 24015 }, { "epoch": 6.296199213630406, "grad_norm": 1.0670430660247803, "learning_rate": 1.814315510896372e-05, "loss": 0.2464, "num_input_tokens_seen": 6431096, "step": 24020 }, { "epoch": 6.297509829619921, "grad_norm": 1.9515669345855713, "learning_rate": 1.813215714574999e-05, "loss": 0.26, "num_input_tokens_seen": 6432232, "step": 24025 }, { "epoch": 6.2988204456094365, "grad_norm": 1.814413070678711, "learning_rate": 1.8121160619963894e-05, "loss": 0.1739, "num_input_tokens_seen": 6433624, "step": 24030 }, { "epoch": 6.300131061598951, "grad_norm": 1.2957247495651245, "learning_rate": 1.8110165533907007e-05, "loss": 0.2497, "num_input_tokens_seen": 6434696, "step": 24035 }, { "epoch": 6.301441677588467, "grad_norm": 1.1105080842971802, "learning_rate": 1.809917188988056e-05, "loss": 0.1908, "num_input_tokens_seen": 6436104, "step": 24040 }, { "epoch": 6.302752293577981, "grad_norm": 2.7681214809417725, "learning_rate": 1.8088179690185515e-05, "loss": 0.2876, "num_input_tokens_seen": 6437336, "step": 24045 }, { "epoch": 6.304062909567497, "grad_norm": 1.9363659620285034, "learning_rate": 1.8077188937122515e-05, "loss": 0.1665, "num_input_tokens_seen": 6438504, "step": 24050 }, { "epoch": 6.305373525557012, "grad_norm": 1.0231547355651855, "learning_rate": 1.8066199632991913e-05, "loss": 0.2186, "num_input_tokens_seen": 6439592, "step": 24055 }, { "epoch": 6.306684141546527, "grad_norm": 2.126044511795044, "learning_rate": 1.805521178009375e-05, "loss": 0.3632, "num_input_tokens_seen": 6441000, "step": 24060 }, { "epoch": 6.307994757536042, "grad_norm": 1.974825143814087, "learning_rate": 1.804422538072775e-05, "loss": 0.2062, "num_input_tokens_seen": 6442376, "step": 24065 }, { "epoch": 6.309305373525557, "grad_norm": 1.9851739406585693, "learning_rate": 1.8033240437193364e-05, "loss": 0.1818, "num_input_tokens_seen": 6443688, "step": 24070 }, { "epoch": 6.310615989515072, "grad_norm": 1.3586838245391846, "learning_rate": 1.8022256951789718e-05, "loss": 0.2833, "num_input_tokens_seen": 6445176, "step": 24075 }, { "epoch": 6.3119266055045875, "grad_norm": 1.2991379499435425, "learning_rate": 1.8011274926815635e-05, "loss": 0.1588, "num_input_tokens_seen": 6446360, "step": 24080 }, { "epoch": 6.313237221494102, "grad_norm": 3.302372455596924, "learning_rate": 1.8000294364569636e-05, "loss": 0.492, "num_input_tokens_seen": 6447560, "step": 24085 }, { "epoch": 6.314547837483618, "grad_norm": 1.3871076107025146, "learning_rate": 1.7989315267349936e-05, "loss": 0.2062, "num_input_tokens_seen": 6448792, "step": 24090 }, { "epoch": 6.315858453473132, "grad_norm": 1.3708348274230957, "learning_rate": 1.7978337637454433e-05, "loss": 0.1867, "num_input_tokens_seen": 6450232, "step": 24095 }, { "epoch": 6.317169069462647, "grad_norm": 1.1732125282287598, "learning_rate": 1.7967361477180727e-05, "loss": 0.2806, "num_input_tokens_seen": 6451736, "step": 24100 }, { "epoch": 6.3184796854521625, "grad_norm": 2.742499351501465, "learning_rate": 1.7956386788826112e-05, "loss": 0.2655, "num_input_tokens_seen": 6452696, "step": 24105 }, { "epoch": 6.319790301441677, "grad_norm": 2.1856327056884766, "learning_rate": 1.7945413574687574e-05, "loss": 0.2818, "num_input_tokens_seen": 6454168, "step": 24110 }, { "epoch": 6.321100917431193, "grad_norm": 3.8306632041931152, "learning_rate": 1.7934441837061795e-05, "loss": 0.3588, "num_input_tokens_seen": 6455480, "step": 24115 }, { "epoch": 6.322411533420707, "grad_norm": 3.0346128940582275, "learning_rate": 1.792347157824512e-05, "loss": 0.197, "num_input_tokens_seen": 6456728, "step": 24120 }, { "epoch": 6.323722149410223, "grad_norm": 2.122591018676758, "learning_rate": 1.7912502800533604e-05, "loss": 0.2554, "num_input_tokens_seen": 6458152, "step": 24125 }, { "epoch": 6.325032765399738, "grad_norm": 2.8716728687286377, "learning_rate": 1.7901535506223027e-05, "loss": 0.1752, "num_input_tokens_seen": 6459304, "step": 24130 }, { "epoch": 6.326343381389253, "grad_norm": 1.6440620422363281, "learning_rate": 1.7890569697608793e-05, "loss": 0.2979, "num_input_tokens_seen": 6460376, "step": 24135 }, { "epoch": 6.327653997378768, "grad_norm": 1.8461905717849731, "learning_rate": 1.787960537698604e-05, "loss": 0.2139, "num_input_tokens_seen": 6461528, "step": 24140 }, { "epoch": 6.328964613368283, "grad_norm": 1.3707926273345947, "learning_rate": 1.786864254664958e-05, "loss": 0.1536, "num_input_tokens_seen": 6463512, "step": 24145 }, { "epoch": 6.330275229357798, "grad_norm": 1.3260747194290161, "learning_rate": 1.785768120889392e-05, "loss": 0.1607, "num_input_tokens_seen": 6464568, "step": 24150 }, { "epoch": 6.3315858453473135, "grad_norm": 0.5661066770553589, "learning_rate": 1.784672136601323e-05, "loss": 0.2634, "num_input_tokens_seen": 6465688, "step": 24155 }, { "epoch": 6.332896461336828, "grad_norm": 0.9794564843177795, "learning_rate": 1.7835763020301403e-05, "loss": 0.1665, "num_input_tokens_seen": 6467288, "step": 24160 }, { "epoch": 6.334207077326344, "grad_norm": 0.31028950214385986, "learning_rate": 1.7824806174051995e-05, "loss": 0.1541, "num_input_tokens_seen": 6468232, "step": 24165 }, { "epoch": 6.335517693315858, "grad_norm": 3.757484197616577, "learning_rate": 1.7813850829558257e-05, "loss": 0.2435, "num_input_tokens_seen": 6469512, "step": 24170 }, { "epoch": 6.336828309305374, "grad_norm": 5.451851844787598, "learning_rate": 1.7802896989113127e-05, "loss": 0.2644, "num_input_tokens_seen": 6470744, "step": 24175 }, { "epoch": 6.3381389252948885, "grad_norm": 2.0616185665130615, "learning_rate": 1.779194465500922e-05, "loss": 0.2389, "num_input_tokens_seen": 6471864, "step": 24180 }, { "epoch": 6.339449541284404, "grad_norm": 5.970306873321533, "learning_rate": 1.7780993829538836e-05, "loss": 0.2508, "num_input_tokens_seen": 6472904, "step": 24185 }, { "epoch": 6.340760157273919, "grad_norm": 1.243808388710022, "learning_rate": 1.7770044514993973e-05, "loss": 0.2898, "num_input_tokens_seen": 6474808, "step": 24190 }, { "epoch": 6.342070773263433, "grad_norm": 1.9525961875915527, "learning_rate": 1.7759096713666305e-05, "loss": 0.385, "num_input_tokens_seen": 6476168, "step": 24195 }, { "epoch": 6.343381389252949, "grad_norm": 0.8247703909873962, "learning_rate": 1.774815042784718e-05, "loss": 0.1178, "num_input_tokens_seen": 6477160, "step": 24200 }, { "epoch": 6.344692005242464, "grad_norm": 2.8676912784576416, "learning_rate": 1.773720565982764e-05, "loss": 0.2506, "num_input_tokens_seen": 6478600, "step": 24205 }, { "epoch": 6.346002621231979, "grad_norm": 1.8138171434402466, "learning_rate": 1.7726262411898413e-05, "loss": 0.2852, "num_input_tokens_seen": 6480408, "step": 24210 }, { "epoch": 6.347313237221494, "grad_norm": 3.5433120727539062, "learning_rate": 1.771532068634989e-05, "loss": 0.3277, "num_input_tokens_seen": 6481416, "step": 24215 }, { "epoch": 6.348623853211009, "grad_norm": 2.1543407440185547, "learning_rate": 1.770438048547216e-05, "loss": 0.4252, "num_input_tokens_seen": 6482840, "step": 24220 }, { "epoch": 6.349934469200524, "grad_norm": 1.9296263456344604, "learning_rate": 1.769344181155499e-05, "loss": 0.3148, "num_input_tokens_seen": 6484024, "step": 24225 }, { "epoch": 6.3512450851900395, "grad_norm": 1.7218178510665894, "learning_rate": 1.7682504666887823e-05, "loss": 0.2217, "num_input_tokens_seen": 6485192, "step": 24230 }, { "epoch": 6.352555701179554, "grad_norm": 1.2423213720321655, "learning_rate": 1.767156905375979e-05, "loss": 0.1762, "num_input_tokens_seen": 6486760, "step": 24235 }, { "epoch": 6.35386631716907, "grad_norm": 3.1959035396575928, "learning_rate": 1.7660634974459674e-05, "loss": 0.2699, "num_input_tokens_seen": 6487960, "step": 24240 }, { "epoch": 6.355176933158584, "grad_norm": 4.614531517028809, "learning_rate": 1.7649702431275994e-05, "loss": 0.4633, "num_input_tokens_seen": 6489096, "step": 24245 }, { "epoch": 6.3564875491481, "grad_norm": 1.091967225074768, "learning_rate": 1.7638771426496893e-05, "loss": 0.2818, "num_input_tokens_seen": 6490344, "step": 24250 }, { "epoch": 6.3577981651376145, "grad_norm": 2.8079586029052734, "learning_rate": 1.7627841962410208e-05, "loss": 0.3132, "num_input_tokens_seen": 6491832, "step": 24255 }, { "epoch": 6.35910878112713, "grad_norm": 4.734869956970215, "learning_rate": 1.7616914041303463e-05, "loss": 0.2968, "num_input_tokens_seen": 6492952, "step": 24260 }, { "epoch": 6.360419397116645, "grad_norm": 6.4378461837768555, "learning_rate": 1.7605987665463854e-05, "loss": 0.1951, "num_input_tokens_seen": 6494248, "step": 24265 }, { "epoch": 6.36173001310616, "grad_norm": 0.7111827731132507, "learning_rate": 1.7595062837178254e-05, "loss": 0.1308, "num_input_tokens_seen": 6495336, "step": 24270 }, { "epoch": 6.363040629095675, "grad_norm": 1.5657480955123901, "learning_rate": 1.7584139558733205e-05, "loss": 0.1464, "num_input_tokens_seen": 6496312, "step": 24275 }, { "epoch": 6.3643512450851905, "grad_norm": 1.4484822750091553, "learning_rate": 1.757321783241493e-05, "loss": 0.2213, "num_input_tokens_seen": 6497528, "step": 24280 }, { "epoch": 6.365661861074705, "grad_norm": 1.1638484001159668, "learning_rate": 1.7562297660509334e-05, "loss": 0.3419, "num_input_tokens_seen": 6498936, "step": 24285 }, { "epoch": 6.36697247706422, "grad_norm": 1.5099220275878906, "learning_rate": 1.7551379045301987e-05, "loss": 0.2302, "num_input_tokens_seen": 6500024, "step": 24290 }, { "epoch": 6.368283093053735, "grad_norm": 0.9374719858169556, "learning_rate": 1.7540461989078132e-05, "loss": 0.3396, "num_input_tokens_seen": 6501640, "step": 24295 }, { "epoch": 6.36959370904325, "grad_norm": 1.4567286968231201, "learning_rate": 1.7529546494122685e-05, "loss": 0.2593, "num_input_tokens_seen": 6503144, "step": 24300 }, { "epoch": 6.3709043250327655, "grad_norm": 1.7515714168548584, "learning_rate": 1.7518632562720263e-05, "loss": 0.2798, "num_input_tokens_seen": 6504360, "step": 24305 }, { "epoch": 6.37221494102228, "grad_norm": 1.2562603950500488, "learning_rate": 1.7507720197155114e-05, "loss": 0.2141, "num_input_tokens_seen": 6505992, "step": 24310 }, { "epoch": 6.373525557011796, "grad_norm": 0.8368186354637146, "learning_rate": 1.749680939971118e-05, "loss": 0.446, "num_input_tokens_seen": 6507608, "step": 24315 }, { "epoch": 6.37483617300131, "grad_norm": 3.0467283725738525, "learning_rate": 1.7485900172672076e-05, "loss": 0.2283, "num_input_tokens_seen": 6509112, "step": 24320 }, { "epoch": 6.376146788990826, "grad_norm": 1.2039673328399658, "learning_rate": 1.7474992518321083e-05, "loss": 0.1477, "num_input_tokens_seen": 6510408, "step": 24325 }, { "epoch": 6.3774574049803405, "grad_norm": 4.031517028808594, "learning_rate": 1.7464086438941156e-05, "loss": 0.4144, "num_input_tokens_seen": 6511992, "step": 24330 }, { "epoch": 6.378768020969856, "grad_norm": 2.6966354846954346, "learning_rate": 1.7453181936814915e-05, "loss": 0.19, "num_input_tokens_seen": 6513288, "step": 24335 }, { "epoch": 6.380078636959371, "grad_norm": 2.8070106506347656, "learning_rate": 1.7442279014224654e-05, "loss": 0.2922, "num_input_tokens_seen": 6514456, "step": 24340 }, { "epoch": 6.381389252948886, "grad_norm": 2.240384101867676, "learning_rate": 1.7431377673452336e-05, "loss": 0.2088, "num_input_tokens_seen": 6515832, "step": 24345 }, { "epoch": 6.382699868938401, "grad_norm": 1.1353641748428345, "learning_rate": 1.74204779167796e-05, "loss": 0.2806, "num_input_tokens_seen": 6517000, "step": 24350 }, { "epoch": 6.3840104849279165, "grad_norm": 1.934184193611145, "learning_rate": 1.740957974648772e-05, "loss": 0.2987, "num_input_tokens_seen": 6519128, "step": 24355 }, { "epoch": 6.385321100917431, "grad_norm": 1.3806270360946655, "learning_rate": 1.7398683164857697e-05, "loss": 0.1483, "num_input_tokens_seen": 6520104, "step": 24360 }, { "epoch": 6.386631716906947, "grad_norm": 1.5703027248382568, "learning_rate": 1.7387788174170157e-05, "loss": 0.2454, "num_input_tokens_seen": 6521720, "step": 24365 }, { "epoch": 6.387942332896461, "grad_norm": 1.35746169090271, "learning_rate": 1.7376894776705397e-05, "loss": 0.1972, "num_input_tokens_seen": 6523304, "step": 24370 }, { "epoch": 6.389252948885977, "grad_norm": 2.4454495906829834, "learning_rate": 1.7366002974743383e-05, "loss": 0.2109, "num_input_tokens_seen": 6524536, "step": 24375 }, { "epoch": 6.3905635648754915, "grad_norm": 1.3558001518249512, "learning_rate": 1.7355112770563764e-05, "loss": 0.3106, "num_input_tokens_seen": 6526024, "step": 24380 }, { "epoch": 6.391874180865006, "grad_norm": 3.6879026889801025, "learning_rate": 1.734422416644583e-05, "loss": 0.3029, "num_input_tokens_seen": 6527112, "step": 24385 }, { "epoch": 6.393184796854522, "grad_norm": 2.200411319732666, "learning_rate": 1.733333716466855e-05, "loss": 0.358, "num_input_tokens_seen": 6528552, "step": 24390 }, { "epoch": 6.394495412844036, "grad_norm": 4.020582675933838, "learning_rate": 1.732245176751055e-05, "loss": 0.2539, "num_input_tokens_seen": 6529688, "step": 24395 }, { "epoch": 6.395806028833552, "grad_norm": 2.6692473888397217, "learning_rate": 1.7311567977250133e-05, "loss": 0.2862, "num_input_tokens_seen": 6531032, "step": 24400 }, { "epoch": 6.3971166448230665, "grad_norm": 3.4636194705963135, "learning_rate": 1.7300685796165255e-05, "loss": 0.1661, "num_input_tokens_seen": 6531944, "step": 24405 }, { "epoch": 6.398427260812582, "grad_norm": 0.5234965085983276, "learning_rate": 1.728980522653354e-05, "loss": 0.1646, "num_input_tokens_seen": 6532968, "step": 24410 }, { "epoch": 6.399737876802097, "grad_norm": 3.850360870361328, "learning_rate": 1.727892627063225e-05, "loss": 0.2221, "num_input_tokens_seen": 6534152, "step": 24415 }, { "epoch": 6.401048492791612, "grad_norm": 1.358048915863037, "learning_rate": 1.726804893073837e-05, "loss": 0.293, "num_input_tokens_seen": 6535608, "step": 24420 }, { "epoch": 6.402359108781127, "grad_norm": 1.8971186876296997, "learning_rate": 1.725717320912848e-05, "loss": 0.4633, "num_input_tokens_seen": 6536760, "step": 24425 }, { "epoch": 6.4036697247706424, "grad_norm": 2.2974343299865723, "learning_rate": 1.724629910807886e-05, "loss": 0.3403, "num_input_tokens_seen": 6538232, "step": 24430 }, { "epoch": 6.404980340760157, "grad_norm": 2.935946226119995, "learning_rate": 1.7235426629865445e-05, "loss": 0.1319, "num_input_tokens_seen": 6539464, "step": 24435 }, { "epoch": 6.406290956749673, "grad_norm": 1.0858100652694702, "learning_rate": 1.7224555776763814e-05, "loss": 0.1665, "num_input_tokens_seen": 6540568, "step": 24440 }, { "epoch": 6.407601572739187, "grad_norm": 2.809264659881592, "learning_rate": 1.7213686551049233e-05, "loss": 0.2833, "num_input_tokens_seen": 6543944, "step": 24445 }, { "epoch": 6.408912188728703, "grad_norm": 1.587375283241272, "learning_rate": 1.7202818954996597e-05, "loss": 0.1652, "num_input_tokens_seen": 6544920, "step": 24450 }, { "epoch": 6.4102228047182175, "grad_norm": 1.4607826471328735, "learning_rate": 1.7191952990880476e-05, "loss": 0.2651, "num_input_tokens_seen": 6546136, "step": 24455 }, { "epoch": 6.411533420707733, "grad_norm": 2.187437057495117, "learning_rate": 1.7181088660975107e-05, "loss": 0.2627, "num_input_tokens_seen": 6547272, "step": 24460 }, { "epoch": 6.412844036697248, "grad_norm": 1.7777844667434692, "learning_rate": 1.7170225967554374e-05, "loss": 0.178, "num_input_tokens_seen": 6548376, "step": 24465 }, { "epoch": 6.414154652686763, "grad_norm": 1.3393548727035522, "learning_rate": 1.715936491289181e-05, "loss": 0.2437, "num_input_tokens_seen": 6550072, "step": 24470 }, { "epoch": 6.415465268676278, "grad_norm": 1.1529039144515991, "learning_rate": 1.7148505499260627e-05, "loss": 0.2845, "num_input_tokens_seen": 6551736, "step": 24475 }, { "epoch": 6.4167758846657925, "grad_norm": 0.8895043134689331, "learning_rate": 1.713764772893368e-05, "loss": 0.2417, "num_input_tokens_seen": 6553144, "step": 24480 }, { "epoch": 6.418086500655308, "grad_norm": 3.3019392490386963, "learning_rate": 1.7126791604183474e-05, "loss": 0.203, "num_input_tokens_seen": 6554152, "step": 24485 }, { "epoch": 6.419397116644823, "grad_norm": 1.556369662284851, "learning_rate": 1.7115937127282182e-05, "loss": 0.1811, "num_input_tokens_seen": 6555160, "step": 24490 }, { "epoch": 6.420707732634338, "grad_norm": 2.5180537700653076, "learning_rate": 1.7105084300501627e-05, "loss": 0.2229, "num_input_tokens_seen": 6556232, "step": 24495 }, { "epoch": 6.422018348623853, "grad_norm": 1.6980621814727783, "learning_rate": 1.7094233126113285e-05, "loss": 0.4364, "num_input_tokens_seen": 6557448, "step": 24500 }, { "epoch": 6.423328964613368, "grad_norm": 6.672554969787598, "learning_rate": 1.70833836063883e-05, "loss": 0.2398, "num_input_tokens_seen": 6558680, "step": 24505 }, { "epoch": 6.424639580602883, "grad_norm": 13.667848587036133, "learning_rate": 1.7072535743597433e-05, "loss": 0.3029, "num_input_tokens_seen": 6559848, "step": 24510 }, { "epoch": 6.425950196592399, "grad_norm": 0.37781015038490295, "learning_rate": 1.7061689540011144e-05, "loss": 0.1892, "num_input_tokens_seen": 6560904, "step": 24515 }, { "epoch": 6.427260812581913, "grad_norm": 2.5187230110168457, "learning_rate": 1.7050844997899517e-05, "loss": 0.2522, "num_input_tokens_seen": 6562184, "step": 24520 }, { "epoch": 6.428571428571429, "grad_norm": 1.2539806365966797, "learning_rate": 1.70400021195323e-05, "loss": 0.1797, "num_input_tokens_seen": 6563656, "step": 24525 }, { "epoch": 6.4298820445609435, "grad_norm": 2.206974744796753, "learning_rate": 1.702916090717887e-05, "loss": 0.2553, "num_input_tokens_seen": 6564808, "step": 24530 }, { "epoch": 6.431192660550459, "grad_norm": 1.8368867635726929, "learning_rate": 1.7018321363108303e-05, "loss": 0.2261, "num_input_tokens_seen": 6565656, "step": 24535 }, { "epoch": 6.432503276539974, "grad_norm": 1.2990002632141113, "learning_rate": 1.700748348958929e-05, "loss": 0.2734, "num_input_tokens_seen": 6567176, "step": 24540 }, { "epoch": 6.433813892529489, "grad_norm": 2.1917500495910645, "learning_rate": 1.699664728889016e-05, "loss": 0.21, "num_input_tokens_seen": 6568520, "step": 24545 }, { "epoch": 6.435124508519004, "grad_norm": 2.620882272720337, "learning_rate": 1.698581276327893e-05, "loss": 0.1849, "num_input_tokens_seen": 6569592, "step": 24550 }, { "epoch": 6.436435124508519, "grad_norm": 2.6678614616394043, "learning_rate": 1.6974979915023237e-05, "loss": 0.2106, "num_input_tokens_seen": 6570744, "step": 24555 }, { "epoch": 6.437745740498034, "grad_norm": 2.767241954803467, "learning_rate": 1.6964148746390383e-05, "loss": 0.1347, "num_input_tokens_seen": 6571704, "step": 24560 }, { "epoch": 6.43905635648755, "grad_norm": 1.8563843965530396, "learning_rate": 1.6953319259647323e-05, "loss": 0.3417, "num_input_tokens_seen": 6572904, "step": 24565 }, { "epoch": 6.440366972477064, "grad_norm": 2.656893491744995, "learning_rate": 1.6942491457060626e-05, "loss": 0.4795, "num_input_tokens_seen": 6574520, "step": 24570 }, { "epoch": 6.441677588466579, "grad_norm": 3.9321134090423584, "learning_rate": 1.6931665340896545e-05, "loss": 0.2995, "num_input_tokens_seen": 6576152, "step": 24575 }, { "epoch": 6.442988204456094, "grad_norm": 1.7351505756378174, "learning_rate": 1.692084091342097e-05, "loss": 0.2669, "num_input_tokens_seen": 6577512, "step": 24580 }, { "epoch": 6.444298820445609, "grad_norm": 0.7733346223831177, "learning_rate": 1.6910018176899424e-05, "loss": 0.1343, "num_input_tokens_seen": 6578648, "step": 24585 }, { "epoch": 6.445609436435125, "grad_norm": 1.3747119903564453, "learning_rate": 1.68991971335971e-05, "loss": 0.2241, "num_input_tokens_seen": 6579992, "step": 24590 }, { "epoch": 6.446920052424639, "grad_norm": 1.2436082363128662, "learning_rate": 1.688837778577882e-05, "loss": 0.1643, "num_input_tokens_seen": 6581096, "step": 24595 }, { "epoch": 6.448230668414155, "grad_norm": 2.7274105548858643, "learning_rate": 1.687756013570906e-05, "loss": 0.2414, "num_input_tokens_seen": 6582344, "step": 24600 }, { "epoch": 6.4495412844036695, "grad_norm": 3.041038751602173, "learning_rate": 1.6866744185651922e-05, "loss": 0.2619, "num_input_tokens_seen": 6583800, "step": 24605 }, { "epoch": 6.450851900393185, "grad_norm": 1.2803078889846802, "learning_rate": 1.6855929937871175e-05, "loss": 0.258, "num_input_tokens_seen": 6585896, "step": 24610 }, { "epoch": 6.4521625163827, "grad_norm": 2.4919912815093994, "learning_rate": 1.6845117394630223e-05, "loss": 0.2434, "num_input_tokens_seen": 6587224, "step": 24615 }, { "epoch": 6.453473132372215, "grad_norm": 1.425834059715271, "learning_rate": 1.683430655819211e-05, "loss": 0.2575, "num_input_tokens_seen": 6588792, "step": 24620 }, { "epoch": 6.45478374836173, "grad_norm": 1.415590524673462, "learning_rate": 1.6823497430819518e-05, "loss": 0.2061, "num_input_tokens_seen": 6590104, "step": 24625 }, { "epoch": 6.456094364351245, "grad_norm": 1.690026879310608, "learning_rate": 1.681269001477479e-05, "loss": 0.24, "num_input_tokens_seen": 6591032, "step": 24630 }, { "epoch": 6.45740498034076, "grad_norm": 1.6734195947647095, "learning_rate": 1.6801884312319895e-05, "loss": 0.2864, "num_input_tokens_seen": 6592664, "step": 24635 }, { "epoch": 6.458715596330276, "grad_norm": 1.2938700914382935, "learning_rate": 1.6791080325716445e-05, "loss": 0.1757, "num_input_tokens_seen": 6594184, "step": 24640 }, { "epoch": 6.46002621231979, "grad_norm": 1.770066738128662, "learning_rate": 1.6780278057225697e-05, "loss": 0.1703, "num_input_tokens_seen": 6595320, "step": 24645 }, { "epoch": 6.461336828309306, "grad_norm": 1.9267300367355347, "learning_rate": 1.6769477509108556e-05, "loss": 0.3198, "num_input_tokens_seen": 6596680, "step": 24650 }, { "epoch": 6.46264744429882, "grad_norm": 1.8813124895095825, "learning_rate": 1.675867868362555e-05, "loss": 0.267, "num_input_tokens_seen": 6598840, "step": 24655 }, { "epoch": 6.463958060288336, "grad_norm": 2.0711350440979004, "learning_rate": 1.6747881583036858e-05, "loss": 0.2557, "num_input_tokens_seen": 6600088, "step": 24660 }, { "epoch": 6.465268676277851, "grad_norm": 2.85628080368042, "learning_rate": 1.673708620960229e-05, "loss": 0.2549, "num_input_tokens_seen": 6601320, "step": 24665 }, { "epoch": 6.466579292267365, "grad_norm": 4.5888237953186035, "learning_rate": 1.6726292565581298e-05, "loss": 0.1583, "num_input_tokens_seen": 6602376, "step": 24670 }, { "epoch": 6.467889908256881, "grad_norm": 3.051175355911255, "learning_rate": 1.6715500653232974e-05, "loss": 0.2226, "num_input_tokens_seen": 6604344, "step": 24675 }, { "epoch": 6.4692005242463955, "grad_norm": 1.45399808883667, "learning_rate": 1.670471047481606e-05, "loss": 0.2887, "num_input_tokens_seen": 6606088, "step": 24680 }, { "epoch": 6.470511140235911, "grad_norm": 1.6758090257644653, "learning_rate": 1.66939220325889e-05, "loss": 0.1844, "num_input_tokens_seen": 6607560, "step": 24685 }, { "epoch": 6.471821756225426, "grad_norm": 1.7879748344421387, "learning_rate": 1.6683135328809512e-05, "loss": 0.2619, "num_input_tokens_seen": 6608904, "step": 24690 }, { "epoch": 6.473132372214941, "grad_norm": 1.4551335573196411, "learning_rate": 1.6672350365735528e-05, "loss": 0.2737, "num_input_tokens_seen": 6610088, "step": 24695 }, { "epoch": 6.474442988204456, "grad_norm": 1.6529558897018433, "learning_rate": 1.6661567145624215e-05, "loss": 0.3121, "num_input_tokens_seen": 6611640, "step": 24700 }, { "epoch": 6.475753604193971, "grad_norm": 2.583927869796753, "learning_rate": 1.6650785670732498e-05, "loss": 0.2092, "num_input_tokens_seen": 6612968, "step": 24705 }, { "epoch": 6.477064220183486, "grad_norm": 1.995395541191101, "learning_rate": 1.6640005943316914e-05, "loss": 0.1554, "num_input_tokens_seen": 6614072, "step": 24710 }, { "epoch": 6.478374836173002, "grad_norm": 1.0252865552902222, "learning_rate": 1.6629227965633644e-05, "loss": 0.1853, "num_input_tokens_seen": 6615384, "step": 24715 }, { "epoch": 6.479685452162516, "grad_norm": 2.0460526943206787, "learning_rate": 1.6618451739938495e-05, "loss": 0.3221, "num_input_tokens_seen": 6616904, "step": 24720 }, { "epoch": 6.480996068152032, "grad_norm": 1.7532262802124023, "learning_rate": 1.6607677268486914e-05, "loss": 0.5414, "num_input_tokens_seen": 6618152, "step": 24725 }, { "epoch": 6.482306684141546, "grad_norm": 3.0901601314544678, "learning_rate": 1.659690455353398e-05, "loss": 0.2941, "num_input_tokens_seen": 6619432, "step": 24730 }, { "epoch": 6.483617300131062, "grad_norm": 3.3154759407043457, "learning_rate": 1.65861335973344e-05, "loss": 0.2108, "num_input_tokens_seen": 6620760, "step": 24735 }, { "epoch": 6.484927916120577, "grad_norm": 3.2769339084625244, "learning_rate": 1.6575364402142528e-05, "loss": 0.369, "num_input_tokens_seen": 6622104, "step": 24740 }, { "epoch": 6.486238532110092, "grad_norm": 0.9127000570297241, "learning_rate": 1.6564596970212327e-05, "loss": 0.2271, "num_input_tokens_seen": 6623224, "step": 24745 }, { "epoch": 6.487549148099607, "grad_norm": 0.17395110428333282, "learning_rate": 1.6553831303797405e-05, "loss": 0.252, "num_input_tokens_seen": 6624264, "step": 24750 }, { "epoch": 6.488859764089122, "grad_norm": 3.018648147583008, "learning_rate": 1.6543067405150994e-05, "loss": 0.3049, "num_input_tokens_seen": 6625240, "step": 24755 }, { "epoch": 6.490170380078637, "grad_norm": 2.1631357669830322, "learning_rate": 1.653230527652596e-05, "loss": 0.215, "num_input_tokens_seen": 6626552, "step": 24760 }, { "epoch": 6.491480996068152, "grad_norm": 1.5574642419815063, "learning_rate": 1.6521544920174803e-05, "loss": 0.324, "num_input_tokens_seen": 6627768, "step": 24765 }, { "epoch": 6.492791612057667, "grad_norm": 0.773709237575531, "learning_rate": 1.6510786338349645e-05, "loss": 0.2154, "num_input_tokens_seen": 6629288, "step": 24770 }, { "epoch": 6.494102228047182, "grad_norm": 1.6064163446426392, "learning_rate": 1.6500029533302247e-05, "loss": 0.3645, "num_input_tokens_seen": 6630296, "step": 24775 }, { "epoch": 6.495412844036697, "grad_norm": 1.5949465036392212, "learning_rate": 1.648927450728397e-05, "loss": 0.2799, "num_input_tokens_seen": 6631400, "step": 24780 }, { "epoch": 6.496723460026212, "grad_norm": 3.810789108276367, "learning_rate": 1.6478521262545837e-05, "loss": 0.2002, "num_input_tokens_seen": 6632424, "step": 24785 }, { "epoch": 6.498034076015728, "grad_norm": 1.1819249391555786, "learning_rate": 1.6467769801338478e-05, "loss": 0.1522, "num_input_tokens_seen": 6634120, "step": 24790 }, { "epoch": 6.499344692005242, "grad_norm": 2.8783295154571533, "learning_rate": 1.645702012591216e-05, "loss": 0.2108, "num_input_tokens_seen": 6635256, "step": 24795 }, { "epoch": 6.500655307994758, "grad_norm": 3.8708746433258057, "learning_rate": 1.644627223851677e-05, "loss": 0.2342, "num_input_tokens_seen": 6636552, "step": 24800 }, { "epoch": 6.50170380078637, "eval_loss": 0.6671392321586609, "eval_runtime": 16.681, "eval_samples_per_second": 50.836, "eval_steps_per_second": 25.418, "num_input_tokens_seen": 6637864, "step": 24804 }, { "epoch": 6.501965923984272, "grad_norm": 2.7949609756469727, "learning_rate": 1.643552614140182e-05, "loss": 0.2578, "num_input_tokens_seen": 6638008, "step": 24805 }, { "epoch": 6.503276539973788, "grad_norm": 2.883180618286133, "learning_rate": 1.6424781836816443e-05, "loss": 0.2418, "num_input_tokens_seen": 6639304, "step": 24810 }, { "epoch": 6.504587155963303, "grad_norm": 1.5648415088653564, "learning_rate": 1.641403932700941e-05, "loss": 0.1757, "num_input_tokens_seen": 6640776, "step": 24815 }, { "epoch": 6.505897771952818, "grad_norm": 3.1461751461029053, "learning_rate": 1.6403298614229117e-05, "loss": 0.2295, "num_input_tokens_seen": 6642024, "step": 24820 }, { "epoch": 6.507208387942333, "grad_norm": 3.328606128692627, "learning_rate": 1.639255970072357e-05, "loss": 0.2644, "num_input_tokens_seen": 6643288, "step": 24825 }, { "epoch": 6.508519003931848, "grad_norm": 3.244825839996338, "learning_rate": 1.6381822588740404e-05, "loss": 0.2818, "num_input_tokens_seen": 6644232, "step": 24830 }, { "epoch": 6.509829619921363, "grad_norm": 1.2884443998336792, "learning_rate": 1.6371087280526888e-05, "loss": 0.1702, "num_input_tokens_seen": 6645496, "step": 24835 }, { "epoch": 6.511140235910878, "grad_norm": 2.4371886253356934, "learning_rate": 1.6360353778329883e-05, "loss": 0.2781, "num_input_tokens_seen": 6647128, "step": 24840 }, { "epoch": 6.512450851900393, "grad_norm": 2.497892379760742, "learning_rate": 1.6349622084395903e-05, "loss": 0.2564, "num_input_tokens_seen": 6648056, "step": 24845 }, { "epoch": 6.513761467889909, "grad_norm": 1.6126786470413208, "learning_rate": 1.6338892200971078e-05, "loss": 0.3495, "num_input_tokens_seen": 6649176, "step": 24850 }, { "epoch": 6.515072083879423, "grad_norm": 0.9111617207527161, "learning_rate": 1.6328164130301157e-05, "loss": 0.2248, "num_input_tokens_seen": 6650536, "step": 24855 }, { "epoch": 6.516382699868938, "grad_norm": 1.239813208580017, "learning_rate": 1.631743787463149e-05, "loss": 0.2619, "num_input_tokens_seen": 6652072, "step": 24860 }, { "epoch": 6.517693315858454, "grad_norm": 0.9693565368652344, "learning_rate": 1.630671343620708e-05, "loss": 0.1365, "num_input_tokens_seen": 6653240, "step": 24865 }, { "epoch": 6.519003931847968, "grad_norm": 1.936277985572815, "learning_rate": 1.6295990817272516e-05, "loss": 0.2627, "num_input_tokens_seen": 6654440, "step": 24870 }, { "epoch": 6.520314547837484, "grad_norm": 2.100653648376465, "learning_rate": 1.6285270020072044e-05, "loss": 0.204, "num_input_tokens_seen": 6655608, "step": 24875 }, { "epoch": 6.521625163826998, "grad_norm": 1.016603946685791, "learning_rate": 1.6274551046849495e-05, "loss": 0.1295, "num_input_tokens_seen": 6656712, "step": 24880 }, { "epoch": 6.522935779816514, "grad_norm": 2.34100079536438, "learning_rate": 1.6263833899848337e-05, "loss": 0.2731, "num_input_tokens_seen": 6658936, "step": 24885 }, { "epoch": 6.524246395806029, "grad_norm": 2.968716621398926, "learning_rate": 1.625311858131165e-05, "loss": 0.3245, "num_input_tokens_seen": 6660328, "step": 24890 }, { "epoch": 6.525557011795544, "grad_norm": 2.2013230323791504, "learning_rate": 1.6242405093482133e-05, "loss": 0.277, "num_input_tokens_seen": 6661352, "step": 24895 }, { "epoch": 6.526867627785059, "grad_norm": 0.8769590854644775, "learning_rate": 1.6231693438602097e-05, "loss": 0.3063, "num_input_tokens_seen": 6662600, "step": 24900 }, { "epoch": 6.528178243774574, "grad_norm": 1.2925572395324707, "learning_rate": 1.6220983618913472e-05, "loss": 0.2764, "num_input_tokens_seen": 6663656, "step": 24905 }, { "epoch": 6.529488859764089, "grad_norm": 4.478983402252197, "learning_rate": 1.6210275636657805e-05, "loss": 0.2323, "num_input_tokens_seen": 6664744, "step": 24910 }, { "epoch": 6.5307994757536045, "grad_norm": 1.6121459007263184, "learning_rate": 1.6199569494076265e-05, "loss": 0.2712, "num_input_tokens_seen": 6666104, "step": 24915 }, { "epoch": 6.532110091743119, "grad_norm": 2.219001054763794, "learning_rate": 1.618886519340962e-05, "loss": 0.2638, "num_input_tokens_seen": 6667352, "step": 24920 }, { "epoch": 6.533420707732635, "grad_norm": 1.651220679283142, "learning_rate": 1.6178162736898264e-05, "loss": 0.3062, "num_input_tokens_seen": 6669080, "step": 24925 }, { "epoch": 6.534731323722149, "grad_norm": 3.8128371238708496, "learning_rate": 1.61674621267822e-05, "loss": 0.3361, "num_input_tokens_seen": 6670712, "step": 24930 }, { "epoch": 6.536041939711664, "grad_norm": 1.4987390041351318, "learning_rate": 1.6156763365301054e-05, "loss": 0.2239, "num_input_tokens_seen": 6671944, "step": 24935 }, { "epoch": 6.53735255570118, "grad_norm": 1.9172990322113037, "learning_rate": 1.614606645469406e-05, "loss": 0.2674, "num_input_tokens_seen": 6673080, "step": 24940 }, { "epoch": 6.538663171690695, "grad_norm": 1.605698823928833, "learning_rate": 1.6135371397200057e-05, "loss": 0.3047, "num_input_tokens_seen": 6674056, "step": 24945 }, { "epoch": 6.53997378768021, "grad_norm": 1.5918400287628174, "learning_rate": 1.6124678195057507e-05, "loss": 0.2122, "num_input_tokens_seen": 6675368, "step": 24950 }, { "epoch": 6.541284403669724, "grad_norm": 2.3243768215179443, "learning_rate": 1.6113986850504475e-05, "loss": 0.3778, "num_input_tokens_seen": 6676696, "step": 24955 }, { "epoch": 6.54259501965924, "grad_norm": 1.2824896574020386, "learning_rate": 1.6103297365778636e-05, "loss": 0.3004, "num_input_tokens_seen": 6678136, "step": 24960 }, { "epoch": 6.543905635648755, "grad_norm": 2.2578928470611572, "learning_rate": 1.609260974311729e-05, "loss": 0.2633, "num_input_tokens_seen": 6679656, "step": 24965 }, { "epoch": 6.54521625163827, "grad_norm": 1.3307678699493408, "learning_rate": 1.6081923984757332e-05, "loss": 0.2084, "num_input_tokens_seen": 6680856, "step": 24970 }, { "epoch": 6.546526867627785, "grad_norm": 3.1362671852111816, "learning_rate": 1.607124009293528e-05, "loss": 0.2155, "num_input_tokens_seen": 6682296, "step": 24975 }, { "epoch": 6.5478374836173, "grad_norm": 1.7210544347763062, "learning_rate": 1.6060558069887244e-05, "loss": 0.2131, "num_input_tokens_seen": 6683624, "step": 24980 }, { "epoch": 6.549148099606815, "grad_norm": 2.2995076179504395, "learning_rate": 1.6049877917848948e-05, "loss": 0.2682, "num_input_tokens_seen": 6685160, "step": 24985 }, { "epoch": 6.5504587155963305, "grad_norm": 1.3495991230010986, "learning_rate": 1.6039199639055755e-05, "loss": 0.32, "num_input_tokens_seen": 6687512, "step": 24990 }, { "epoch": 6.551769331585845, "grad_norm": 1.1311200857162476, "learning_rate": 1.602852323574259e-05, "loss": 0.3237, "num_input_tokens_seen": 6689320, "step": 24995 }, { "epoch": 6.553079947575361, "grad_norm": 1.3546226024627686, "learning_rate": 1.6017848710144006e-05, "loss": 0.1585, "num_input_tokens_seen": 6690920, "step": 25000 }, { "epoch": 6.554390563564875, "grad_norm": 2.165797472000122, "learning_rate": 1.600717606449417e-05, "loss": 0.2468, "num_input_tokens_seen": 6692088, "step": 25005 }, { "epoch": 6.555701179554391, "grad_norm": 1.3406788110733032, "learning_rate": 1.5996505301026848e-05, "loss": 0.2406, "num_input_tokens_seen": 6693288, "step": 25010 }, { "epoch": 6.557011795543906, "grad_norm": 2.139265298843384, "learning_rate": 1.598583642197541e-05, "loss": 0.2278, "num_input_tokens_seen": 6694632, "step": 25015 }, { "epoch": 6.558322411533421, "grad_norm": 2.4070146083831787, "learning_rate": 1.5975169429572828e-05, "loss": 0.0703, "num_input_tokens_seen": 6695576, "step": 25020 }, { "epoch": 6.559633027522936, "grad_norm": 3.156869888305664, "learning_rate": 1.5964504326051693e-05, "loss": 0.2404, "num_input_tokens_seen": 6696536, "step": 25025 }, { "epoch": 6.56094364351245, "grad_norm": 1.286544680595398, "learning_rate": 1.595384111364419e-05, "loss": 0.1738, "num_input_tokens_seen": 6697960, "step": 25030 }, { "epoch": 6.562254259501966, "grad_norm": 1.7316354513168335, "learning_rate": 1.594317979458212e-05, "loss": 0.2579, "num_input_tokens_seen": 6699560, "step": 25035 }, { "epoch": 6.5635648754914815, "grad_norm": 1.2571359872817993, "learning_rate": 1.593252037109686e-05, "loss": 0.1402, "num_input_tokens_seen": 6700632, "step": 25040 }, { "epoch": 6.564875491480996, "grad_norm": 2.194887161254883, "learning_rate": 1.5921862845419417e-05, "loss": 0.2476, "num_input_tokens_seen": 6702312, "step": 25045 }, { "epoch": 6.566186107470511, "grad_norm": 1.8740257024765015, "learning_rate": 1.5911207219780398e-05, "loss": 0.3064, "num_input_tokens_seen": 6703816, "step": 25050 }, { "epoch": 6.567496723460026, "grad_norm": 1.1894400119781494, "learning_rate": 1.5900553496410004e-05, "loss": 0.1935, "num_input_tokens_seen": 6705256, "step": 25055 }, { "epoch": 6.568807339449541, "grad_norm": 2.7665188312530518, "learning_rate": 1.5889901677538037e-05, "loss": 0.2583, "num_input_tokens_seen": 6706408, "step": 25060 }, { "epoch": 6.5701179554390565, "grad_norm": 1.4620097875595093, "learning_rate": 1.5879251765393914e-05, "loss": 0.2915, "num_input_tokens_seen": 6707448, "step": 25065 }, { "epoch": 6.571428571428571, "grad_norm": 1.4056648015975952, "learning_rate": 1.5868603762206637e-05, "loss": 0.2786, "num_input_tokens_seen": 6708648, "step": 25070 }, { "epoch": 6.572739187418087, "grad_norm": 1.7485697269439697, "learning_rate": 1.5857957670204816e-05, "loss": 0.2219, "num_input_tokens_seen": 6709928, "step": 25075 }, { "epoch": 6.574049803407601, "grad_norm": 1.7296308279037476, "learning_rate": 1.5847313491616654e-05, "loss": 0.2274, "num_input_tokens_seen": 6711816, "step": 25080 }, { "epoch": 6.575360419397117, "grad_norm": 1.3984544277191162, "learning_rate": 1.5836671228669962e-05, "loss": 0.2435, "num_input_tokens_seen": 6713128, "step": 25085 }, { "epoch": 6.576671035386632, "grad_norm": 3.3322594165802, "learning_rate": 1.582603088359216e-05, "loss": 0.1436, "num_input_tokens_seen": 6714776, "step": 25090 }, { "epoch": 6.577981651376147, "grad_norm": 3.3710360527038574, "learning_rate": 1.581539245861023e-05, "loss": 0.2726, "num_input_tokens_seen": 6715672, "step": 25095 }, { "epoch": 6.579292267365662, "grad_norm": 1.5992205142974854, "learning_rate": 1.5804755955950785e-05, "loss": 0.2324, "num_input_tokens_seen": 6717016, "step": 25100 }, { "epoch": 6.580602883355177, "grad_norm": 1.8252872228622437, "learning_rate": 1.579412137784004e-05, "loss": 0.1518, "num_input_tokens_seen": 6718104, "step": 25105 }, { "epoch": 6.581913499344692, "grad_norm": 1.5925283432006836, "learning_rate": 1.578348872650378e-05, "loss": 0.1654, "num_input_tokens_seen": 6719064, "step": 25110 }, { "epoch": 6.5832241153342075, "grad_norm": 3.352980613708496, "learning_rate": 1.5772858004167407e-05, "loss": 0.3567, "num_input_tokens_seen": 6720376, "step": 25115 }, { "epoch": 6.584534731323722, "grad_norm": 1.1687238216400146, "learning_rate": 1.5762229213055914e-05, "loss": 0.1791, "num_input_tokens_seen": 6721784, "step": 25120 }, { "epoch": 6.585845347313237, "grad_norm": 1.6434283256530762, "learning_rate": 1.5751602355393878e-05, "loss": 0.2025, "num_input_tokens_seen": 6723192, "step": 25125 }, { "epoch": 6.587155963302752, "grad_norm": 5.324244022369385, "learning_rate": 1.5740977433405496e-05, "loss": 0.1928, "num_input_tokens_seen": 6724392, "step": 25130 }, { "epoch": 6.588466579292267, "grad_norm": 1.3904563188552856, "learning_rate": 1.5730354449314534e-05, "loss": 0.2529, "num_input_tokens_seen": 6725672, "step": 25135 }, { "epoch": 6.5897771952817825, "grad_norm": 1.1530088186264038, "learning_rate": 1.5719733405344373e-05, "loss": 0.191, "num_input_tokens_seen": 6727288, "step": 25140 }, { "epoch": 6.591087811271297, "grad_norm": 2.9965555667877197, "learning_rate": 1.570911430371797e-05, "loss": 0.289, "num_input_tokens_seen": 6728744, "step": 25145 }, { "epoch": 6.592398427260813, "grad_norm": 1.6946923732757568, "learning_rate": 1.5698497146657897e-05, "loss": 0.3248, "num_input_tokens_seen": 6730040, "step": 25150 }, { "epoch": 6.593709043250327, "grad_norm": 1.2059102058410645, "learning_rate": 1.5687881936386294e-05, "loss": 0.2924, "num_input_tokens_seen": 6731592, "step": 25155 }, { "epoch": 6.595019659239843, "grad_norm": 3.4804656505584717, "learning_rate": 1.5677268675124902e-05, "loss": 0.2983, "num_input_tokens_seen": 6733064, "step": 25160 }, { "epoch": 6.5963302752293576, "grad_norm": 1.9182997941970825, "learning_rate": 1.5666657365095083e-05, "loss": 0.1329, "num_input_tokens_seen": 6734216, "step": 25165 }, { "epoch": 6.597640891218873, "grad_norm": 1.7446786165237427, "learning_rate": 1.565604800851774e-05, "loss": 0.1303, "num_input_tokens_seen": 6735224, "step": 25170 }, { "epoch": 6.598951507208388, "grad_norm": 1.6113282442092896, "learning_rate": 1.5645440607613413e-05, "loss": 0.2162, "num_input_tokens_seen": 6736728, "step": 25175 }, { "epoch": 6.600262123197903, "grad_norm": 0.8543061017990112, "learning_rate": 1.56348351646022e-05, "loss": 0.2452, "num_input_tokens_seen": 6737720, "step": 25180 }, { "epoch": 6.601572739187418, "grad_norm": 1.8244683742523193, "learning_rate": 1.5624231681703812e-05, "loss": 0.2665, "num_input_tokens_seen": 6739016, "step": 25185 }, { "epoch": 6.6028833551769335, "grad_norm": 1.0774368047714233, "learning_rate": 1.5613630161137523e-05, "loss": 0.2015, "num_input_tokens_seen": 6740328, "step": 25190 }, { "epoch": 6.604193971166448, "grad_norm": 2.6089954376220703, "learning_rate": 1.5603030605122226e-05, "loss": 0.2073, "num_input_tokens_seen": 6741384, "step": 25195 }, { "epoch": 6.605504587155964, "grad_norm": 2.103705883026123, "learning_rate": 1.5592433015876383e-05, "loss": 0.2834, "num_input_tokens_seen": 6742744, "step": 25200 }, { "epoch": 6.606815203145478, "grad_norm": 2.467336654663086, "learning_rate": 1.558183739561806e-05, "loss": 0.4044, "num_input_tokens_seen": 6746088, "step": 25205 }, { "epoch": 6.608125819134994, "grad_norm": 2.427124261856079, "learning_rate": 1.55712437465649e-05, "loss": 0.1885, "num_input_tokens_seen": 6747784, "step": 25210 }, { "epoch": 6.6094364351245085, "grad_norm": 1.9293110370635986, "learning_rate": 1.5560652070934117e-05, "loss": 0.3641, "num_input_tokens_seen": 6748920, "step": 25215 }, { "epoch": 6.610747051114023, "grad_norm": 2.7111783027648926, "learning_rate": 1.5550062370942556e-05, "loss": 0.2517, "num_input_tokens_seen": 6750584, "step": 25220 }, { "epoch": 6.612057667103539, "grad_norm": 1.3960341215133667, "learning_rate": 1.553947464880662e-05, "loss": 0.228, "num_input_tokens_seen": 6752184, "step": 25225 }, { "epoch": 6.613368283093053, "grad_norm": 2.3221819400787354, "learning_rate": 1.5528888906742285e-05, "loss": 0.4528, "num_input_tokens_seen": 6753480, "step": 25230 }, { "epoch": 6.614678899082569, "grad_norm": 1.7363542318344116, "learning_rate": 1.5518305146965133e-05, "loss": 0.2712, "num_input_tokens_seen": 6755032, "step": 25235 }, { "epoch": 6.6159895150720835, "grad_norm": 1.3967534303665161, "learning_rate": 1.5507723371690336e-05, "loss": 0.1471, "num_input_tokens_seen": 6756680, "step": 25240 }, { "epoch": 6.617300131061599, "grad_norm": 4.570379257202148, "learning_rate": 1.5497143583132638e-05, "loss": 0.2519, "num_input_tokens_seen": 6757896, "step": 25245 }, { "epoch": 6.618610747051114, "grad_norm": 1.3877174854278564, "learning_rate": 1.5486565783506367e-05, "loss": 0.3071, "num_input_tokens_seen": 6759800, "step": 25250 }, { "epoch": 6.619921363040629, "grad_norm": 1.97965669631958, "learning_rate": 1.547598997502544e-05, "loss": 0.4131, "num_input_tokens_seen": 6761096, "step": 25255 }, { "epoch": 6.621231979030144, "grad_norm": 0.9430743455886841, "learning_rate": 1.546541615990335e-05, "loss": 0.2172, "num_input_tokens_seen": 6762328, "step": 25260 }, { "epoch": 6.6225425950196595, "grad_norm": 2.7556965351104736, "learning_rate": 1.545484434035319e-05, "loss": 0.1888, "num_input_tokens_seen": 6763496, "step": 25265 }, { "epoch": 6.623853211009174, "grad_norm": 2.4072751998901367, "learning_rate": 1.5444274518587617e-05, "loss": 0.3152, "num_input_tokens_seen": 6764904, "step": 25270 }, { "epoch": 6.62516382699869, "grad_norm": 1.9261829853057861, "learning_rate": 1.543370669681886e-05, "loss": 0.1961, "num_input_tokens_seen": 6766168, "step": 25275 }, { "epoch": 6.626474442988204, "grad_norm": 1.963349461555481, "learning_rate": 1.5423140877258784e-05, "loss": 0.2579, "num_input_tokens_seen": 6767272, "step": 25280 }, { "epoch": 6.62778505897772, "grad_norm": 2.020838499069214, "learning_rate": 1.5412577062118763e-05, "loss": 0.3646, "num_input_tokens_seen": 6768456, "step": 25285 }, { "epoch": 6.6290956749672345, "grad_norm": 1.533765435218811, "learning_rate": 1.5402015253609796e-05, "loss": 0.2195, "num_input_tokens_seen": 6769800, "step": 25290 }, { "epoch": 6.63040629095675, "grad_norm": 1.5476419925689697, "learning_rate": 1.539145545394246e-05, "loss": 0.1937, "num_input_tokens_seen": 6771112, "step": 25295 }, { "epoch": 6.631716906946265, "grad_norm": 2.9777655601501465, "learning_rate": 1.538089766532689e-05, "loss": 0.3079, "num_input_tokens_seen": 6772840, "step": 25300 }, { "epoch": 6.63302752293578, "grad_norm": 1.7996022701263428, "learning_rate": 1.537034188997282e-05, "loss": 0.1929, "num_input_tokens_seen": 6774008, "step": 25305 }, { "epoch": 6.634338138925295, "grad_norm": 2.939380168914795, "learning_rate": 1.5359788130089558e-05, "loss": 0.3174, "num_input_tokens_seen": 6775336, "step": 25310 }, { "epoch": 6.6356487549148095, "grad_norm": 0.9413760304450989, "learning_rate": 1.5349236387885976e-05, "loss": 0.2304, "num_input_tokens_seen": 6776872, "step": 25315 }, { "epoch": 6.636959370904325, "grad_norm": 1.4571613073349, "learning_rate": 1.5338686665570546e-05, "loss": 0.2539, "num_input_tokens_seen": 6778040, "step": 25320 }, { "epoch": 6.63826998689384, "grad_norm": 1.408513069152832, "learning_rate": 1.5328138965351307e-05, "loss": 0.2494, "num_input_tokens_seen": 6779864, "step": 25325 }, { "epoch": 6.639580602883355, "grad_norm": 1.786520004272461, "learning_rate": 1.5317593289435855e-05, "loss": 0.2977, "num_input_tokens_seen": 6781800, "step": 25330 }, { "epoch": 6.64089121887287, "grad_norm": 2.118067979812622, "learning_rate": 1.530704964003141e-05, "loss": 0.2728, "num_input_tokens_seen": 6783208, "step": 25335 }, { "epoch": 6.6422018348623855, "grad_norm": 1.7073715925216675, "learning_rate": 1.529650801934473e-05, "loss": 0.2535, "num_input_tokens_seen": 6784664, "step": 25340 }, { "epoch": 6.6435124508519, "grad_norm": 1.6824665069580078, "learning_rate": 1.5285968429582153e-05, "loss": 0.2014, "num_input_tokens_seen": 6786056, "step": 25345 }, { "epoch": 6.644823066841416, "grad_norm": 1.7510491609573364, "learning_rate": 1.52754308729496e-05, "loss": 0.32, "num_input_tokens_seen": 6787656, "step": 25350 }, { "epoch": 6.64613368283093, "grad_norm": 4.108894348144531, "learning_rate": 1.5264895351652555e-05, "loss": 0.3461, "num_input_tokens_seen": 6788728, "step": 25355 }, { "epoch": 6.647444298820446, "grad_norm": 1.5117878913879395, "learning_rate": 1.52543618678961e-05, "loss": 0.1509, "num_input_tokens_seen": 6789832, "step": 25360 }, { "epoch": 6.6487549148099605, "grad_norm": 3.6214053630828857, "learning_rate": 1.5243830423884867e-05, "loss": 0.1556, "num_input_tokens_seen": 6791368, "step": 25365 }, { "epoch": 6.650065530799476, "grad_norm": 3.0171380043029785, "learning_rate": 1.5233301021823065e-05, "loss": 0.2983, "num_input_tokens_seen": 6792648, "step": 25370 }, { "epoch": 6.651376146788991, "grad_norm": 10.055960655212402, "learning_rate": 1.5222773663914486e-05, "loss": 0.1774, "num_input_tokens_seen": 6794120, "step": 25375 }, { "epoch": 6.652686762778506, "grad_norm": 1.7527207136154175, "learning_rate": 1.5212248352362482e-05, "loss": 0.331, "num_input_tokens_seen": 6795240, "step": 25380 }, { "epoch": 6.653997378768021, "grad_norm": 1.4931011199951172, "learning_rate": 1.5201725089369991e-05, "loss": 0.1863, "num_input_tokens_seen": 6796632, "step": 25385 }, { "epoch": 6.655307994757536, "grad_norm": 2.894707202911377, "learning_rate": 1.5191203877139498e-05, "loss": 0.1812, "num_input_tokens_seen": 6798120, "step": 25390 }, { "epoch": 6.656618610747051, "grad_norm": 1.440462350845337, "learning_rate": 1.5180684717873089e-05, "loss": 0.3453, "num_input_tokens_seen": 6799784, "step": 25395 }, { "epoch": 6.657929226736567, "grad_norm": 1.7111172676086426, "learning_rate": 1.5170167613772412e-05, "loss": 0.2123, "num_input_tokens_seen": 6800952, "step": 25400 }, { "epoch": 6.659239842726081, "grad_norm": 3.0455307960510254, "learning_rate": 1.5159652567038663e-05, "loss": 0.3462, "num_input_tokens_seen": 6802504, "step": 25405 }, { "epoch": 6.660550458715596, "grad_norm": 1.4798980951309204, "learning_rate": 1.5149139579872629e-05, "loss": 0.3113, "num_input_tokens_seen": 6803880, "step": 25410 }, { "epoch": 6.6618610747051115, "grad_norm": 1.4796780347824097, "learning_rate": 1.5138628654474656e-05, "loss": 0.2711, "num_input_tokens_seen": 6805176, "step": 25415 }, { "epoch": 6.663171690694626, "grad_norm": 1.1806204319000244, "learning_rate": 1.5128119793044677e-05, "loss": 0.1911, "num_input_tokens_seen": 6806456, "step": 25420 }, { "epoch": 6.664482306684142, "grad_norm": 3.813420057296753, "learning_rate": 1.511761299778216e-05, "loss": 0.2264, "num_input_tokens_seen": 6807592, "step": 25425 }, { "epoch": 6.665792922673656, "grad_norm": 1.7209250926971436, "learning_rate": 1.5107108270886166e-05, "loss": 0.3168, "num_input_tokens_seen": 6809448, "step": 25430 }, { "epoch": 6.667103538663172, "grad_norm": 1.6921199560165405, "learning_rate": 1.509660561455532e-05, "loss": 0.2794, "num_input_tokens_seen": 6810520, "step": 25435 }, { "epoch": 6.6684141546526865, "grad_norm": 3.5738561153411865, "learning_rate": 1.5086105030987807e-05, "loss": 0.2782, "num_input_tokens_seen": 6811656, "step": 25440 }, { "epoch": 6.669724770642202, "grad_norm": 1.2525302171707153, "learning_rate": 1.5075606522381375e-05, "loss": 0.1662, "num_input_tokens_seen": 6812600, "step": 25445 }, { "epoch": 6.671035386631717, "grad_norm": 1.8258386850357056, "learning_rate": 1.5065110090933354e-05, "loss": 0.2902, "num_input_tokens_seen": 6813912, "step": 25450 }, { "epoch": 6.672346002621232, "grad_norm": 1.2999995946884155, "learning_rate": 1.5054615738840627e-05, "loss": 0.2712, "num_input_tokens_seen": 6814904, "step": 25455 }, { "epoch": 6.673656618610747, "grad_norm": 1.318206787109375, "learning_rate": 1.5044123468299648e-05, "loss": 0.331, "num_input_tokens_seen": 6815992, "step": 25460 }, { "epoch": 6.674967234600262, "grad_norm": 1.5522310733795166, "learning_rate": 1.5033633281506416e-05, "loss": 0.3218, "num_input_tokens_seen": 6817640, "step": 25465 }, { "epoch": 6.676277850589777, "grad_norm": 1.7117424011230469, "learning_rate": 1.5023145180656517e-05, "loss": 0.3076, "num_input_tokens_seen": 6819288, "step": 25470 }, { "epoch": 6.677588466579293, "grad_norm": 3.5350029468536377, "learning_rate": 1.5012659167945097e-05, "loss": 0.2594, "num_input_tokens_seen": 6820440, "step": 25475 }, { "epoch": 6.678899082568807, "grad_norm": 1.0524544715881348, "learning_rate": 1.5002175245566857e-05, "loss": 0.1815, "num_input_tokens_seen": 6822104, "step": 25480 }, { "epoch": 6.680209698558322, "grad_norm": 3.37886643409729, "learning_rate": 1.4991693415716062e-05, "loss": 0.3935, "num_input_tokens_seen": 6823208, "step": 25485 }, { "epoch": 6.6815203145478375, "grad_norm": 2.1885814666748047, "learning_rate": 1.498121368058654e-05, "loss": 0.2755, "num_input_tokens_seen": 6824872, "step": 25490 }, { "epoch": 6.682830930537353, "grad_norm": 1.7825169563293457, "learning_rate": 1.4970736042371685e-05, "loss": 0.2035, "num_input_tokens_seen": 6826040, "step": 25495 }, { "epoch": 6.684141546526868, "grad_norm": 1.8996835947036743, "learning_rate": 1.4960260503264445e-05, "loss": 0.2564, "num_input_tokens_seen": 6827400, "step": 25500 }, { "epoch": 6.685452162516382, "grad_norm": 1.5034120082855225, "learning_rate": 1.494978706545733e-05, "loss": 0.1704, "num_input_tokens_seen": 6828760, "step": 25505 }, { "epoch": 6.686762778505898, "grad_norm": 2.636913776397705, "learning_rate": 1.4939315731142423e-05, "loss": 0.1768, "num_input_tokens_seen": 6830296, "step": 25510 }, { "epoch": 6.6880733944954125, "grad_norm": 1.7624200582504272, "learning_rate": 1.4928846502511353e-05, "loss": 0.3062, "num_input_tokens_seen": 6832152, "step": 25515 }, { "epoch": 6.689384010484928, "grad_norm": 1.910718560218811, "learning_rate": 1.4918379381755303e-05, "loss": 0.2084, "num_input_tokens_seen": 6833288, "step": 25520 }, { "epoch": 6.690694626474443, "grad_norm": 0.6148360967636108, "learning_rate": 1.4907914371065032e-05, "loss": 0.1004, "num_input_tokens_seen": 6834792, "step": 25525 }, { "epoch": 6.692005242463958, "grad_norm": 3.1673388481140137, "learning_rate": 1.4897451472630844e-05, "loss": 0.3622, "num_input_tokens_seen": 6836360, "step": 25530 }, { "epoch": 6.693315858453473, "grad_norm": 0.5678116679191589, "learning_rate": 1.4886990688642605e-05, "loss": 0.1906, "num_input_tokens_seen": 6837448, "step": 25535 }, { "epoch": 6.694626474442988, "grad_norm": 2.5160632133483887, "learning_rate": 1.487653202128975e-05, "loss": 0.3146, "num_input_tokens_seen": 6838632, "step": 25540 }, { "epoch": 6.695937090432503, "grad_norm": 3.314711093902588, "learning_rate": 1.4866075472761243e-05, "loss": 0.2421, "num_input_tokens_seen": 6839720, "step": 25545 }, { "epoch": 6.697247706422019, "grad_norm": 1.7912801504135132, "learning_rate": 1.485562104524563e-05, "loss": 0.2414, "num_input_tokens_seen": 6840856, "step": 25550 }, { "epoch": 6.698558322411533, "grad_norm": 1.4185876846313477, "learning_rate": 1.4845168740931004e-05, "loss": 0.2118, "num_input_tokens_seen": 6842088, "step": 25555 }, { "epoch": 6.699868938401049, "grad_norm": 1.10417640209198, "learning_rate": 1.4834718562005007e-05, "loss": 0.1703, "num_input_tokens_seen": 6843528, "step": 25560 }, { "epoch": 6.7011795543905635, "grad_norm": 2.4899046421051025, "learning_rate": 1.4824270510654859e-05, "loss": 0.2247, "num_input_tokens_seen": 6844776, "step": 25565 }, { "epoch": 6.702490170380079, "grad_norm": 0.8302702307701111, "learning_rate": 1.4813824589067309e-05, "loss": 0.1519, "num_input_tokens_seen": 6846216, "step": 25570 }, { "epoch": 6.703800786369594, "grad_norm": 3.6547799110412598, "learning_rate": 1.4803380799428679e-05, "loss": 0.2134, "num_input_tokens_seen": 6847224, "step": 25575 }, { "epoch": 6.705111402359108, "grad_norm": 1.7030041217803955, "learning_rate": 1.4792939143924822e-05, "loss": 0.2652, "num_input_tokens_seen": 6848536, "step": 25580 }, { "epoch": 6.706422018348624, "grad_norm": 1.7795188426971436, "learning_rate": 1.478249962474117e-05, "loss": 0.3419, "num_input_tokens_seen": 6849672, "step": 25585 }, { "epoch": 6.707732634338139, "grad_norm": 2.0287067890167236, "learning_rate": 1.4772062244062687e-05, "loss": 0.2788, "num_input_tokens_seen": 6850824, "step": 25590 }, { "epoch": 6.709043250327654, "grad_norm": 2.7299952507019043, "learning_rate": 1.4761627004073905e-05, "loss": 0.2568, "num_input_tokens_seen": 6852072, "step": 25595 }, { "epoch": 6.710353866317169, "grad_norm": 3.708306312561035, "learning_rate": 1.4751193906958907e-05, "loss": 0.265, "num_input_tokens_seen": 6853240, "step": 25600 }, { "epoch": 6.711664482306684, "grad_norm": 1.6317461729049683, "learning_rate": 1.474076295490131e-05, "loss": 0.1852, "num_input_tokens_seen": 6854840, "step": 25605 }, { "epoch": 6.712975098296199, "grad_norm": 2.065112829208374, "learning_rate": 1.4730334150084302e-05, "loss": 0.2036, "num_input_tokens_seen": 6855848, "step": 25610 }, { "epoch": 6.714285714285714, "grad_norm": 1.043625831604004, "learning_rate": 1.4719907494690604e-05, "loss": 0.2406, "num_input_tokens_seen": 6857320, "step": 25615 }, { "epoch": 6.715596330275229, "grad_norm": 2.6936655044555664, "learning_rate": 1.470948299090251e-05, "loss": 0.1984, "num_input_tokens_seen": 6858568, "step": 25620 }, { "epoch": 6.716906946264745, "grad_norm": 3.9324398040771484, "learning_rate": 1.4699060640901845e-05, "loss": 0.125, "num_input_tokens_seen": 6859912, "step": 25625 }, { "epoch": 6.718217562254259, "grad_norm": 1.3162046670913696, "learning_rate": 1.4688640446869988e-05, "loss": 0.1588, "num_input_tokens_seen": 6860760, "step": 25630 }, { "epoch": 6.719528178243775, "grad_norm": 5.153762340545654, "learning_rate": 1.467822241098788e-05, "loss": 0.2161, "num_input_tokens_seen": 6861928, "step": 25635 }, { "epoch": 6.7208387942332894, "grad_norm": 2.9122025966644287, "learning_rate": 1.4667806535435974e-05, "loss": 0.1585, "num_input_tokens_seen": 6864136, "step": 25640 }, { "epoch": 6.722149410222805, "grad_norm": 0.9297611117362976, "learning_rate": 1.4657392822394312e-05, "loss": 0.1579, "num_input_tokens_seen": 6865480, "step": 25645 }, { "epoch": 6.72346002621232, "grad_norm": 1.878532886505127, "learning_rate": 1.464698127404246e-05, "loss": 0.2002, "num_input_tokens_seen": 6866888, "step": 25650 }, { "epoch": 6.724770642201835, "grad_norm": 1.8530993461608887, "learning_rate": 1.4636571892559547e-05, "loss": 0.2826, "num_input_tokens_seen": 6868248, "step": 25655 }, { "epoch": 6.72608125819135, "grad_norm": 3.919110059738159, "learning_rate": 1.4626164680124221e-05, "loss": 0.2247, "num_input_tokens_seen": 6869352, "step": 25660 }, { "epoch": 6.727391874180865, "grad_norm": 1.5562161207199097, "learning_rate": 1.4615759638914705e-05, "loss": 0.2249, "num_input_tokens_seen": 6870440, "step": 25665 }, { "epoch": 6.72870249017038, "grad_norm": 2.7133195400238037, "learning_rate": 1.4605356771108753e-05, "loss": 0.2774, "num_input_tokens_seen": 6871704, "step": 25670 }, { "epoch": 6.730013106159895, "grad_norm": 1.3756768703460693, "learning_rate": 1.4594956078883665e-05, "loss": 0.2221, "num_input_tokens_seen": 6873208, "step": 25675 }, { "epoch": 6.73132372214941, "grad_norm": 2.4878644943237305, "learning_rate": 1.4584557564416296e-05, "loss": 0.2629, "num_input_tokens_seen": 6874616, "step": 25680 }, { "epoch": 6.732634338138926, "grad_norm": 2.0295469760894775, "learning_rate": 1.457416122988303e-05, "loss": 0.1926, "num_input_tokens_seen": 6875672, "step": 25685 }, { "epoch": 6.73394495412844, "grad_norm": 1.4900530576705933, "learning_rate": 1.4563767077459805e-05, "loss": 0.2106, "num_input_tokens_seen": 6877064, "step": 25690 }, { "epoch": 6.735255570117955, "grad_norm": 2.872729778289795, "learning_rate": 1.4553375109322098e-05, "loss": 0.2883, "num_input_tokens_seen": 6878328, "step": 25695 }, { "epoch": 6.736566186107471, "grad_norm": 1.8286409378051758, "learning_rate": 1.4542985327644931e-05, "loss": 0.2175, "num_input_tokens_seen": 6879528, "step": 25700 }, { "epoch": 6.737876802096985, "grad_norm": 1.1876640319824219, "learning_rate": 1.4532597734602877e-05, "loss": 0.2987, "num_input_tokens_seen": 6880920, "step": 25705 }, { "epoch": 6.739187418086501, "grad_norm": 1.973356008529663, "learning_rate": 1.4522212332370019e-05, "loss": 0.2211, "num_input_tokens_seen": 6882344, "step": 25710 }, { "epoch": 6.740498034076015, "grad_norm": 2.384884834289551, "learning_rate": 1.4511829123120018e-05, "loss": 0.3666, "num_input_tokens_seen": 6883624, "step": 25715 }, { "epoch": 6.741808650065531, "grad_norm": 1.8340263366699219, "learning_rate": 1.4501448109026055e-05, "loss": 0.2765, "num_input_tokens_seen": 6885432, "step": 25720 }, { "epoch": 6.743119266055046, "grad_norm": 0.7784782648086548, "learning_rate": 1.4491069292260868e-05, "loss": 0.187, "num_input_tokens_seen": 6887480, "step": 25725 }, { "epoch": 6.744429882044561, "grad_norm": 0.8483484983444214, "learning_rate": 1.448069267499671e-05, "loss": 0.2161, "num_input_tokens_seen": 6888728, "step": 25730 }, { "epoch": 6.745740498034076, "grad_norm": 2.6021997928619385, "learning_rate": 1.4470318259405407e-05, "loss": 0.5593, "num_input_tokens_seen": 6889848, "step": 25735 }, { "epoch": 6.747051114023591, "grad_norm": 1.415408968925476, "learning_rate": 1.4459946047658306e-05, "loss": 0.2352, "num_input_tokens_seen": 6891016, "step": 25740 }, { "epoch": 6.748361730013106, "grad_norm": 7.352571487426758, "learning_rate": 1.444957604192628e-05, "loss": 0.1295, "num_input_tokens_seen": 6892472, "step": 25745 }, { "epoch": 6.749672346002622, "grad_norm": 0.447363942861557, "learning_rate": 1.4439208244379757e-05, "loss": 0.2879, "num_input_tokens_seen": 6893912, "step": 25750 }, { "epoch": 6.750982961992136, "grad_norm": 1.5927385091781616, "learning_rate": 1.4428842657188701e-05, "loss": 0.2231, "num_input_tokens_seen": 6894920, "step": 25755 }, { "epoch": 6.752293577981652, "grad_norm": 1.5900635719299316, "learning_rate": 1.441847928252261e-05, "loss": 0.2403, "num_input_tokens_seen": 6896344, "step": 25760 }, { "epoch": 6.753604193971166, "grad_norm": 1.4900022745132446, "learning_rate": 1.4408118122550524e-05, "loss": 0.3365, "num_input_tokens_seen": 6898072, "step": 25765 }, { "epoch": 6.754914809960681, "grad_norm": 1.5566685199737549, "learning_rate": 1.4397759179441012e-05, "loss": 0.4765, "num_input_tokens_seen": 6899304, "step": 25770 }, { "epoch": 6.756225425950197, "grad_norm": 2.6553330421447754, "learning_rate": 1.4387402455362181e-05, "loss": 0.1576, "num_input_tokens_seen": 6901000, "step": 25775 }, { "epoch": 6.757536041939712, "grad_norm": 1.4405725002288818, "learning_rate": 1.4377047952481685e-05, "loss": 0.2867, "num_input_tokens_seen": 6902184, "step": 25780 }, { "epoch": 6.758846657929227, "grad_norm": 1.9002352952957153, "learning_rate": 1.4366695672966705e-05, "loss": 0.1984, "num_input_tokens_seen": 6903480, "step": 25785 }, { "epoch": 6.760157273918741, "grad_norm": 1.919329285621643, "learning_rate": 1.4356345618983927e-05, "loss": 0.2275, "num_input_tokens_seen": 6904536, "step": 25790 }, { "epoch": 6.761467889908257, "grad_norm": 2.753903388977051, "learning_rate": 1.434599779269963e-05, "loss": 0.1588, "num_input_tokens_seen": 6905768, "step": 25795 }, { "epoch": 6.762778505897772, "grad_norm": 0.806069016456604, "learning_rate": 1.4335652196279587e-05, "loss": 0.1312, "num_input_tokens_seen": 6907096, "step": 25800 }, { "epoch": 6.764089121887287, "grad_norm": 3.6966445446014404, "learning_rate": 1.4325308831889115e-05, "loss": 0.2414, "num_input_tokens_seen": 6908344, "step": 25805 }, { "epoch": 6.765399737876802, "grad_norm": 2.2133870124816895, "learning_rate": 1.4314967701693055e-05, "loss": 0.2189, "num_input_tokens_seen": 6909592, "step": 25810 }, { "epoch": 6.766710353866317, "grad_norm": 2.013005256652832, "learning_rate": 1.4304628807855797e-05, "loss": 0.184, "num_input_tokens_seen": 6910552, "step": 25815 }, { "epoch": 6.768020969855832, "grad_norm": 2.803690195083618, "learning_rate": 1.4294292152541258e-05, "loss": 0.247, "num_input_tokens_seen": 6911880, "step": 25820 }, { "epoch": 6.7693315858453476, "grad_norm": 1.70244300365448, "learning_rate": 1.4283957737912862e-05, "loss": 0.2292, "num_input_tokens_seen": 6913176, "step": 25825 }, { "epoch": 6.770642201834862, "grad_norm": 2.4592223167419434, "learning_rate": 1.4273625566133596e-05, "loss": 0.3281, "num_input_tokens_seen": 6914424, "step": 25830 }, { "epoch": 6.771952817824378, "grad_norm": 1.7925246953964233, "learning_rate": 1.426329563936597e-05, "loss": 0.2495, "num_input_tokens_seen": 6915896, "step": 25835 }, { "epoch": 6.773263433813892, "grad_norm": 0.827044665813446, "learning_rate": 1.4252967959772013e-05, "loss": 0.2148, "num_input_tokens_seen": 6917672, "step": 25840 }, { "epoch": 6.774574049803408, "grad_norm": 0.8980286121368408, "learning_rate": 1.4242642529513278e-05, "loss": 0.1401, "num_input_tokens_seen": 6918968, "step": 25845 }, { "epoch": 6.775884665792923, "grad_norm": 1.9692636728286743, "learning_rate": 1.4232319350750888e-05, "loss": 0.2377, "num_input_tokens_seen": 6920536, "step": 25850 }, { "epoch": 6.777195281782438, "grad_norm": 1.6602963209152222, "learning_rate": 1.422199842564545e-05, "loss": 0.2144, "num_input_tokens_seen": 6922056, "step": 25855 }, { "epoch": 6.778505897771953, "grad_norm": 1.700974464416504, "learning_rate": 1.4211679756357132e-05, "loss": 0.2522, "num_input_tokens_seen": 6923080, "step": 25860 }, { "epoch": 6.779816513761467, "grad_norm": 1.5624394416809082, "learning_rate": 1.4201363345045588e-05, "loss": 0.2247, "num_input_tokens_seen": 6924776, "step": 25865 }, { "epoch": 6.781127129750983, "grad_norm": 2.5073399543762207, "learning_rate": 1.4191049193870037e-05, "loss": 0.2249, "num_input_tokens_seen": 6926216, "step": 25870 }, { "epoch": 6.7824377457404985, "grad_norm": 2.2330780029296875, "learning_rate": 1.4180737304989216e-05, "loss": 0.2403, "num_input_tokens_seen": 6927432, "step": 25875 }, { "epoch": 6.783748361730013, "grad_norm": 1.8624987602233887, "learning_rate": 1.4170427680561382e-05, "loss": 0.2143, "num_input_tokens_seen": 6929064, "step": 25880 }, { "epoch": 6.785058977719528, "grad_norm": 1.0904632806777954, "learning_rate": 1.4160120322744322e-05, "loss": 0.232, "num_input_tokens_seen": 6930584, "step": 25885 }, { "epoch": 6.786369593709043, "grad_norm": 2.3367044925689697, "learning_rate": 1.4149815233695346e-05, "loss": 0.3502, "num_input_tokens_seen": 6932312, "step": 25890 }, { "epoch": 6.787680209698558, "grad_norm": 1.7990397214889526, "learning_rate": 1.4139512415571298e-05, "loss": 0.3503, "num_input_tokens_seen": 6934152, "step": 25895 }, { "epoch": 6.7889908256880735, "grad_norm": 2.189303398132324, "learning_rate": 1.4129211870528534e-05, "loss": 0.2093, "num_input_tokens_seen": 6935400, "step": 25900 }, { "epoch": 6.790301441677588, "grad_norm": 1.3718849420547485, "learning_rate": 1.411891360072294e-05, "loss": 0.1522, "num_input_tokens_seen": 6936424, "step": 25905 }, { "epoch": 6.791612057667104, "grad_norm": 1.8248573541641235, "learning_rate": 1.4108617608309932e-05, "loss": 0.1647, "num_input_tokens_seen": 6937624, "step": 25910 }, { "epoch": 6.792922673656618, "grad_norm": 1.3750749826431274, "learning_rate": 1.4098323895444437e-05, "loss": 0.23, "num_input_tokens_seen": 6939032, "step": 25915 }, { "epoch": 6.794233289646134, "grad_norm": 1.232538104057312, "learning_rate": 1.4088032464280914e-05, "loss": 0.328, "num_input_tokens_seen": 6940296, "step": 25920 }, { "epoch": 6.795543905635649, "grad_norm": 1.7560913562774658, "learning_rate": 1.4077743316973346e-05, "loss": 0.2337, "num_input_tokens_seen": 6942200, "step": 25925 }, { "epoch": 6.796854521625164, "grad_norm": 4.382491111755371, "learning_rate": 1.4067456455675226e-05, "loss": 0.2533, "num_input_tokens_seen": 6944472, "step": 25930 }, { "epoch": 6.798165137614679, "grad_norm": 1.0725537538528442, "learning_rate": 1.4057171882539583e-05, "loss": 0.2586, "num_input_tokens_seen": 6945944, "step": 25935 }, { "epoch": 6.799475753604194, "grad_norm": 2.434145450592041, "learning_rate": 1.4046889599718962e-05, "loss": 0.2045, "num_input_tokens_seen": 6947096, "step": 25940 }, { "epoch": 6.800786369593709, "grad_norm": 2.4589684009552, "learning_rate": 1.4036609609365414e-05, "loss": 0.2499, "num_input_tokens_seen": 6948360, "step": 25945 }, { "epoch": 6.8020969855832245, "grad_norm": 2.0672683715820312, "learning_rate": 1.4026331913630536e-05, "loss": 0.2343, "num_input_tokens_seen": 6949576, "step": 25950 }, { "epoch": 6.803407601572739, "grad_norm": 3.1219024658203125, "learning_rate": 1.401605651466542e-05, "loss": 0.1471, "num_input_tokens_seen": 6950840, "step": 25955 }, { "epoch": 6.804718217562254, "grad_norm": 1.7720856666564941, "learning_rate": 1.4005783414620694e-05, "loss": 0.1636, "num_input_tokens_seen": 6952296, "step": 25960 }, { "epoch": 6.806028833551769, "grad_norm": 2.261397361755371, "learning_rate": 1.3995512615646509e-05, "loss": 0.3419, "num_input_tokens_seen": 6953512, "step": 25965 }, { "epoch": 6.807339449541285, "grad_norm": 3.3323700428009033, "learning_rate": 1.3985244119892518e-05, "loss": 0.235, "num_input_tokens_seen": 6954744, "step": 25970 }, { "epoch": 6.8086500655307995, "grad_norm": 0.9993792176246643, "learning_rate": 1.3974977929507913e-05, "loss": 0.1682, "num_input_tokens_seen": 6956280, "step": 25975 }, { "epoch": 6.809960681520314, "grad_norm": 3.230445384979248, "learning_rate": 1.3964714046641366e-05, "loss": 0.2114, "num_input_tokens_seen": 6957304, "step": 25980 }, { "epoch": 6.81127129750983, "grad_norm": 1.291414499282837, "learning_rate": 1.3954452473441104e-05, "loss": 0.2362, "num_input_tokens_seen": 6958296, "step": 25985 }, { "epoch": 6.812581913499344, "grad_norm": 1.4077110290527344, "learning_rate": 1.3944193212054851e-05, "loss": 0.2787, "num_input_tokens_seen": 6959400, "step": 25990 }, { "epoch": 6.81389252948886, "grad_norm": 1.868666648864746, "learning_rate": 1.393393626462986e-05, "loss": 0.2104, "num_input_tokens_seen": 6960904, "step": 25995 }, { "epoch": 6.815203145478375, "grad_norm": 4.574467182159424, "learning_rate": 1.3923681633312885e-05, "loss": 0.1402, "num_input_tokens_seen": 6962120, "step": 26000 }, { "epoch": 6.81651376146789, "grad_norm": 1.8807827234268188, "learning_rate": 1.3913429320250203e-05, "loss": 0.2039, "num_input_tokens_seen": 6963160, "step": 26005 }, { "epoch": 6.817824377457405, "grad_norm": 0.5960428714752197, "learning_rate": 1.3903179327587612e-05, "loss": 0.2932, "num_input_tokens_seen": 6964696, "step": 26010 }, { "epoch": 6.81913499344692, "grad_norm": 2.2622146606445312, "learning_rate": 1.3892931657470415e-05, "loss": 0.1925, "num_input_tokens_seen": 6966152, "step": 26015 }, { "epoch": 6.820445609436435, "grad_norm": 4.039807319641113, "learning_rate": 1.3882686312043428e-05, "loss": 0.3117, "num_input_tokens_seen": 6967320, "step": 26020 }, { "epoch": 6.8217562254259505, "grad_norm": 2.528869152069092, "learning_rate": 1.3872443293450987e-05, "loss": 0.2567, "num_input_tokens_seen": 6968440, "step": 26025 }, { "epoch": 6.823066841415465, "grad_norm": 1.997273325920105, "learning_rate": 1.3862202603836938e-05, "loss": 0.2826, "num_input_tokens_seen": 6969800, "step": 26030 }, { "epoch": 6.824377457404981, "grad_norm": 2.280055522918701, "learning_rate": 1.385196424534464e-05, "loss": 0.2333, "num_input_tokens_seen": 6971752, "step": 26035 }, { "epoch": 6.825688073394495, "grad_norm": 1.2910295724868774, "learning_rate": 1.3841728220116961e-05, "loss": 0.1609, "num_input_tokens_seen": 6972968, "step": 26040 }, { "epoch": 6.826998689384011, "grad_norm": 1.22651207447052, "learning_rate": 1.3831494530296291e-05, "loss": 0.1378, "num_input_tokens_seen": 6974664, "step": 26045 }, { "epoch": 6.8283093053735255, "grad_norm": 1.6468349695205688, "learning_rate": 1.3821263178024512e-05, "loss": 0.2113, "num_input_tokens_seen": 6976152, "step": 26050 }, { "epoch": 6.82961992136304, "grad_norm": 1.8722444772720337, "learning_rate": 1.3811034165443038e-05, "loss": 0.1439, "num_input_tokens_seen": 6977400, "step": 26055 }, { "epoch": 6.830930537352556, "grad_norm": 2.382549285888672, "learning_rate": 1.3800807494692785e-05, "loss": 0.3195, "num_input_tokens_seen": 6978808, "step": 26060 }, { "epoch": 6.832241153342071, "grad_norm": 1.9036399126052856, "learning_rate": 1.3790583167914168e-05, "loss": 0.2281, "num_input_tokens_seen": 6980424, "step": 26065 }, { "epoch": 6.833551769331586, "grad_norm": 3.9257349967956543, "learning_rate": 1.3780361187247125e-05, "loss": 0.1843, "num_input_tokens_seen": 6981736, "step": 26070 }, { "epoch": 6.834862385321101, "grad_norm": 1.788182258605957, "learning_rate": 1.377014155483109e-05, "loss": 0.3447, "num_input_tokens_seen": 6983112, "step": 26075 }, { "epoch": 6.836173001310616, "grad_norm": 2.063608169555664, "learning_rate": 1.375992427280503e-05, "loss": 0.317, "num_input_tokens_seen": 6984376, "step": 26080 }, { "epoch": 6.837483617300131, "grad_norm": 1.5282411575317383, "learning_rate": 1.3749709343307401e-05, "loss": 0.278, "num_input_tokens_seen": 6985576, "step": 26085 }, { "epoch": 6.838794233289646, "grad_norm": 1.863983392715454, "learning_rate": 1.3739496768476168e-05, "loss": 0.2248, "num_input_tokens_seen": 6986888, "step": 26090 }, { "epoch": 6.840104849279161, "grad_norm": 1.7301137447357178, "learning_rate": 1.3729286550448814e-05, "loss": 0.3227, "num_input_tokens_seen": 6988648, "step": 26095 }, { "epoch": 6.8414154652686765, "grad_norm": 1.5712090730667114, "learning_rate": 1.3719078691362302e-05, "loss": 0.2056, "num_input_tokens_seen": 6990264, "step": 26100 }, { "epoch": 6.842726081258191, "grad_norm": 0.6412261724472046, "learning_rate": 1.3708873193353125e-05, "loss": 0.2294, "num_input_tokens_seen": 6991256, "step": 26105 }, { "epoch": 6.844036697247707, "grad_norm": 1.5746053457260132, "learning_rate": 1.3698670058557281e-05, "loss": 0.2166, "num_input_tokens_seen": 6992360, "step": 26110 }, { "epoch": 6.845347313237221, "grad_norm": 0.6001418828964233, "learning_rate": 1.3688469289110268e-05, "loss": 0.4309, "num_input_tokens_seen": 6994360, "step": 26115 }, { "epoch": 6.846657929226737, "grad_norm": 1.5436376333236694, "learning_rate": 1.3678270887147085e-05, "loss": 0.3101, "num_input_tokens_seen": 6995464, "step": 26120 }, { "epoch": 6.8479685452162515, "grad_norm": 2.137930154800415, "learning_rate": 1.3668074854802243e-05, "loss": 0.244, "num_input_tokens_seen": 6996872, "step": 26125 }, { "epoch": 6.849279161205767, "grad_norm": 4.883234977722168, "learning_rate": 1.3657881194209754e-05, "loss": 0.4217, "num_input_tokens_seen": 6998248, "step": 26130 }, { "epoch": 6.850589777195282, "grad_norm": 7.8630242347717285, "learning_rate": 1.3647689907503136e-05, "loss": 0.275, "num_input_tokens_seen": 6999352, "step": 26135 }, { "epoch": 6.851900393184797, "grad_norm": 1.6285459995269775, "learning_rate": 1.3637500996815402e-05, "loss": 0.2588, "num_input_tokens_seen": 7000760, "step": 26140 }, { "epoch": 6.853211009174312, "grad_norm": 1.3706039190292358, "learning_rate": 1.3627314464279078e-05, "loss": 0.2047, "num_input_tokens_seen": 7002200, "step": 26145 }, { "epoch": 6.854521625163827, "grad_norm": 3.468414545059204, "learning_rate": 1.3617130312026188e-05, "loss": 0.3497, "num_input_tokens_seen": 7004136, "step": 26150 }, { "epoch": 6.855832241153342, "grad_norm": 1.6736639738082886, "learning_rate": 1.3606948542188252e-05, "loss": 0.2344, "num_input_tokens_seen": 7005288, "step": 26155 }, { "epoch": 6.857142857142857, "grad_norm": 2.3680357933044434, "learning_rate": 1.3596769156896303e-05, "loss": 0.2673, "num_input_tokens_seen": 7006264, "step": 26160 }, { "epoch": 6.858453473132372, "grad_norm": 1.3594814538955688, "learning_rate": 1.3586592158280864e-05, "loss": 0.1941, "num_input_tokens_seen": 7007608, "step": 26165 }, { "epoch": 6.859764089121887, "grad_norm": 1.0338691473007202, "learning_rate": 1.3576417548471967e-05, "loss": 0.2309, "num_input_tokens_seen": 7009224, "step": 26170 }, { "epoch": 6.8610747051114025, "grad_norm": 1.498755693435669, "learning_rate": 1.3566245329599148e-05, "loss": 0.2671, "num_input_tokens_seen": 7010728, "step": 26175 }, { "epoch": 6.862385321100917, "grad_norm": 1.395979881286621, "learning_rate": 1.3556075503791416e-05, "loss": 0.2474, "num_input_tokens_seen": 7011752, "step": 26180 }, { "epoch": 6.863695937090433, "grad_norm": 1.674803376197815, "learning_rate": 1.3545908073177308e-05, "loss": 0.4035, "num_input_tokens_seen": 7013016, "step": 26185 }, { "epoch": 6.865006553079947, "grad_norm": 2.03755521774292, "learning_rate": 1.3535743039884841e-05, "loss": 0.2226, "num_input_tokens_seen": 7014360, "step": 26190 }, { "epoch": 6.866317169069463, "grad_norm": 1.2099854946136475, "learning_rate": 1.3525580406041552e-05, "loss": 0.3512, "num_input_tokens_seen": 7015304, "step": 26195 }, { "epoch": 6.8676277850589775, "grad_norm": 1.7429351806640625, "learning_rate": 1.3515420173774463e-05, "loss": 0.4022, "num_input_tokens_seen": 7017096, "step": 26200 }, { "epoch": 6.868938401048493, "grad_norm": 2.227949380874634, "learning_rate": 1.3505262345210084e-05, "loss": 0.8023, "num_input_tokens_seen": 7018248, "step": 26205 }, { "epoch": 6.870249017038008, "grad_norm": 2.44122314453125, "learning_rate": 1.3495106922474444e-05, "loss": 0.2896, "num_input_tokens_seen": 7019960, "step": 26210 }, { "epoch": 6.871559633027523, "grad_norm": 3.276092767715454, "learning_rate": 1.3484953907693038e-05, "loss": 0.3219, "num_input_tokens_seen": 7021016, "step": 26215 }, { "epoch": 6.872870249017038, "grad_norm": 2.9858343601226807, "learning_rate": 1.3474803302990885e-05, "loss": 0.1974, "num_input_tokens_seen": 7022520, "step": 26220 }, { "epoch": 6.8741808650065535, "grad_norm": 1.8386037349700928, "learning_rate": 1.3464655110492484e-05, "loss": 0.2565, "num_input_tokens_seen": 7024056, "step": 26225 }, { "epoch": 6.875491480996068, "grad_norm": 0.7269723415374756, "learning_rate": 1.3454509332321834e-05, "loss": 0.1889, "num_input_tokens_seen": 7025256, "step": 26230 }, { "epoch": 6.876802096985584, "grad_norm": 1.5698318481445312, "learning_rate": 1.3444365970602436e-05, "loss": 0.1788, "num_input_tokens_seen": 7026424, "step": 26235 }, { "epoch": 6.878112712975098, "grad_norm": 1.0880733728408813, "learning_rate": 1.3434225027457274e-05, "loss": 0.1606, "num_input_tokens_seen": 7027704, "step": 26240 }, { "epoch": 6.879423328964613, "grad_norm": 1.616504192352295, "learning_rate": 1.3424086505008827e-05, "loss": 0.1936, "num_input_tokens_seen": 7029288, "step": 26245 }, { "epoch": 6.8807339449541285, "grad_norm": 1.5635665655136108, "learning_rate": 1.3413950405379075e-05, "loss": 0.2847, "num_input_tokens_seen": 7030648, "step": 26250 }, { "epoch": 6.882044560943643, "grad_norm": 2.4287970066070557, "learning_rate": 1.3403816730689483e-05, "loss": 0.2015, "num_input_tokens_seen": 7031992, "step": 26255 }, { "epoch": 6.883355176933159, "grad_norm": 1.8609051704406738, "learning_rate": 1.3393685483061014e-05, "loss": 0.1801, "num_input_tokens_seen": 7033144, "step": 26260 }, { "epoch": 6.884665792922673, "grad_norm": 2.0625834465026855, "learning_rate": 1.3383556664614117e-05, "loss": 0.2838, "num_input_tokens_seen": 7034824, "step": 26265 }, { "epoch": 6.885976408912189, "grad_norm": 1.0623091459274292, "learning_rate": 1.337343027746874e-05, "loss": 0.2261, "num_input_tokens_seen": 7036008, "step": 26270 }, { "epoch": 6.8872870249017035, "grad_norm": 1.12856125831604, "learning_rate": 1.3363306323744315e-05, "loss": 0.2297, "num_input_tokens_seen": 7037624, "step": 26275 }, { "epoch": 6.888597640891219, "grad_norm": 0.8132712841033936, "learning_rate": 1.335318480555977e-05, "loss": 0.3551, "num_input_tokens_seen": 7039096, "step": 26280 }, { "epoch": 6.889908256880734, "grad_norm": 1.4827752113342285, "learning_rate": 1.3343065725033522e-05, "loss": 0.3137, "num_input_tokens_seen": 7040664, "step": 26285 }, { "epoch": 6.891218872870249, "grad_norm": 1.9584498405456543, "learning_rate": 1.3332949084283478e-05, "loss": 0.4, "num_input_tokens_seen": 7042136, "step": 26290 }, { "epoch": 6.892529488859764, "grad_norm": 3.14552903175354, "learning_rate": 1.3322834885427038e-05, "loss": 0.2798, "num_input_tokens_seen": 7043464, "step": 26295 }, { "epoch": 6.8938401048492794, "grad_norm": 2.7801294326782227, "learning_rate": 1.3312723130581073e-05, "loss": 0.2223, "num_input_tokens_seen": 7044568, "step": 26300 }, { "epoch": 6.895150720838794, "grad_norm": 1.3170039653778076, "learning_rate": 1.3302613821861951e-05, "loss": 0.1442, "num_input_tokens_seen": 7045896, "step": 26305 }, { "epoch": 6.89646133682831, "grad_norm": 1.4198129177093506, "learning_rate": 1.3292506961385553e-05, "loss": 0.2251, "num_input_tokens_seen": 7047288, "step": 26310 }, { "epoch": 6.897771952817824, "grad_norm": 3.700468063354492, "learning_rate": 1.3282402551267223e-05, "loss": 0.3175, "num_input_tokens_seen": 7048472, "step": 26315 }, { "epoch": 6.89908256880734, "grad_norm": 2.815046787261963, "learning_rate": 1.3272300593621789e-05, "loss": 0.2015, "num_input_tokens_seen": 7049832, "step": 26320 }, { "epoch": 6.9003931847968545, "grad_norm": 1.164143443107605, "learning_rate": 1.3262201090563581e-05, "loss": 0.3355, "num_input_tokens_seen": 7051176, "step": 26325 }, { "epoch": 6.90170380078637, "grad_norm": 1.1115081310272217, "learning_rate": 1.325210404420641e-05, "loss": 0.2599, "num_input_tokens_seen": 7052888, "step": 26330 }, { "epoch": 6.903014416775885, "grad_norm": 1.1789785623550415, "learning_rate": 1.324200945666355e-05, "loss": 0.1747, "num_input_tokens_seen": 7053960, "step": 26335 }, { "epoch": 6.904325032765399, "grad_norm": 3.4092981815338135, "learning_rate": 1.3231917330047802e-05, "loss": 0.3223, "num_input_tokens_seen": 7055160, "step": 26340 }, { "epoch": 6.905635648754915, "grad_norm": 0.37722039222717285, "learning_rate": 1.322182766647142e-05, "loss": 0.1895, "num_input_tokens_seen": 7056344, "step": 26345 }, { "epoch": 6.9069462647444295, "grad_norm": 3.601919412612915, "learning_rate": 1.3211740468046158e-05, "loss": 0.234, "num_input_tokens_seen": 7057368, "step": 26350 }, { "epoch": 6.908256880733945, "grad_norm": 4.52142858505249, "learning_rate": 1.3201655736883244e-05, "loss": 0.2523, "num_input_tokens_seen": 7058696, "step": 26355 }, { "epoch": 6.90956749672346, "grad_norm": 2.8521392345428467, "learning_rate": 1.31915734750934e-05, "loss": 0.1913, "num_input_tokens_seen": 7060216, "step": 26360 }, { "epoch": 6.910878112712975, "grad_norm": 1.3078420162200928, "learning_rate": 1.3181493684786825e-05, "loss": 0.3327, "num_input_tokens_seen": 7062168, "step": 26365 }, { "epoch": 6.91218872870249, "grad_norm": 2.7595319747924805, "learning_rate": 1.31714163680732e-05, "loss": 0.2508, "num_input_tokens_seen": 7063416, "step": 26370 }, { "epoch": 6.913499344692005, "grad_norm": 0.7506441473960876, "learning_rate": 1.316134152706169e-05, "loss": 0.1885, "num_input_tokens_seen": 7064728, "step": 26375 }, { "epoch": 6.91480996068152, "grad_norm": 1.9058836698532104, "learning_rate": 1.3151269163860947e-05, "loss": 0.3901, "num_input_tokens_seen": 7067624, "step": 26380 }, { "epoch": 6.916120576671036, "grad_norm": 1.689331293106079, "learning_rate": 1.3141199280579092e-05, "loss": 0.4708, "num_input_tokens_seen": 7068760, "step": 26385 }, { "epoch": 6.91743119266055, "grad_norm": 2.324613094329834, "learning_rate": 1.3131131879323736e-05, "loss": 0.2458, "num_input_tokens_seen": 7070328, "step": 26390 }, { "epoch": 6.918741808650066, "grad_norm": 1.615390658378601, "learning_rate": 1.3121066962201971e-05, "loss": 0.1352, "num_input_tokens_seen": 7071832, "step": 26395 }, { "epoch": 6.9200524246395805, "grad_norm": 3.32611346244812, "learning_rate": 1.3111004531320364e-05, "loss": 0.1711, "num_input_tokens_seen": 7072952, "step": 26400 }, { "epoch": 6.921363040629096, "grad_norm": 1.8740664720535278, "learning_rate": 1.3100944588784969e-05, "loss": 0.2588, "num_input_tokens_seen": 7074488, "step": 26405 }, { "epoch": 6.922673656618611, "grad_norm": 1.8963661193847656, "learning_rate": 1.3090887136701319e-05, "loss": 0.0726, "num_input_tokens_seen": 7077080, "step": 26410 }, { "epoch": 6.923984272608125, "grad_norm": 1.387555480003357, "learning_rate": 1.3080832177174406e-05, "loss": 0.1887, "num_input_tokens_seen": 7078616, "step": 26415 }, { "epoch": 6.925294888597641, "grad_norm": 4.8475236892700195, "learning_rate": 1.3070779712308711e-05, "loss": 0.2875, "num_input_tokens_seen": 7080120, "step": 26420 }, { "epoch": 6.926605504587156, "grad_norm": 2.8362295627593994, "learning_rate": 1.3060729744208222e-05, "loss": 0.2986, "num_input_tokens_seen": 7081784, "step": 26425 }, { "epoch": 6.927916120576671, "grad_norm": 1.2548385858535767, "learning_rate": 1.3050682274976369e-05, "loss": 0.1862, "num_input_tokens_seen": 7083208, "step": 26430 }, { "epoch": 6.929226736566186, "grad_norm": 2.438864231109619, "learning_rate": 1.304063730671607e-05, "loss": 0.2437, "num_input_tokens_seen": 7084456, "step": 26435 }, { "epoch": 6.930537352555701, "grad_norm": 2.4865646362304688, "learning_rate": 1.3030594841529714e-05, "loss": 0.2948, "num_input_tokens_seen": 7085800, "step": 26440 }, { "epoch": 6.931847968545216, "grad_norm": 1.3798091411590576, "learning_rate": 1.3020554881519187e-05, "loss": 0.1686, "num_input_tokens_seen": 7087144, "step": 26445 }, { "epoch": 6.933158584534731, "grad_norm": 0.6929230690002441, "learning_rate": 1.3010517428785818e-05, "loss": 0.2176, "num_input_tokens_seen": 7088632, "step": 26450 }, { "epoch": 6.934469200524246, "grad_norm": 1.5564186573028564, "learning_rate": 1.3000482485430433e-05, "loss": 0.1556, "num_input_tokens_seen": 7089800, "step": 26455 }, { "epoch": 6.935779816513762, "grad_norm": 1.5588442087173462, "learning_rate": 1.299045005355333e-05, "loss": 0.2387, "num_input_tokens_seen": 7091016, "step": 26460 }, { "epoch": 6.937090432503276, "grad_norm": 2.5915346145629883, "learning_rate": 1.2980420135254279e-05, "loss": 0.309, "num_input_tokens_seen": 7092264, "step": 26465 }, { "epoch": 6.938401048492792, "grad_norm": 1.304560899734497, "learning_rate": 1.2970392732632527e-05, "loss": 0.2968, "num_input_tokens_seen": 7093320, "step": 26470 }, { "epoch": 6.9397116644823065, "grad_norm": 1.1663705110549927, "learning_rate": 1.2960367847786781e-05, "loss": 0.2087, "num_input_tokens_seen": 7094712, "step": 26475 }, { "epoch": 6.941022280471822, "grad_norm": 1.971138834953308, "learning_rate": 1.295034548281526e-05, "loss": 0.2038, "num_input_tokens_seen": 7095880, "step": 26480 }, { "epoch": 6.942332896461337, "grad_norm": 2.830897092819214, "learning_rate": 1.2940325639815599e-05, "loss": 0.2565, "num_input_tokens_seen": 7097288, "step": 26485 }, { "epoch": 6.943643512450852, "grad_norm": 1.8097134828567505, "learning_rate": 1.2930308320884946e-05, "loss": 0.1955, "num_input_tokens_seen": 7098424, "step": 26490 }, { "epoch": 6.944954128440367, "grad_norm": 1.9675096273422241, "learning_rate": 1.2920293528119903e-05, "loss": 0.1377, "num_input_tokens_seen": 7099608, "step": 26495 }, { "epoch": 6.946264744429882, "grad_norm": 1.617258906364441, "learning_rate": 1.2910281263616559e-05, "loss": 0.406, "num_input_tokens_seen": 7101208, "step": 26500 }, { "epoch": 6.947575360419397, "grad_norm": 2.957867383956909, "learning_rate": 1.2900271529470451e-05, "loss": 0.244, "num_input_tokens_seen": 7102616, "step": 26505 }, { "epoch": 6.948885976408912, "grad_norm": 2.362659454345703, "learning_rate": 1.2890264327776608e-05, "loss": 0.2144, "num_input_tokens_seen": 7103720, "step": 26510 }, { "epoch": 6.950196592398427, "grad_norm": 1.9929786920547485, "learning_rate": 1.2880259660629516e-05, "loss": 0.2764, "num_input_tokens_seen": 7104888, "step": 26515 }, { "epoch": 6.951507208387943, "grad_norm": 2.9362285137176514, "learning_rate": 1.2870257530123136e-05, "loss": 0.2012, "num_input_tokens_seen": 7106088, "step": 26520 }, { "epoch": 6.952817824377457, "grad_norm": 1.455930471420288, "learning_rate": 1.2860257938350895e-05, "loss": 0.5097, "num_input_tokens_seen": 7107320, "step": 26525 }, { "epoch": 6.954128440366972, "grad_norm": 2.9836373329162598, "learning_rate": 1.2850260887405702e-05, "loss": 0.3653, "num_input_tokens_seen": 7109000, "step": 26530 }, { "epoch": 6.955439056356488, "grad_norm": 0.9004831910133362, "learning_rate": 1.2840266379379896e-05, "loss": 0.1621, "num_input_tokens_seen": 7110504, "step": 26535 }, { "epoch": 6.956749672346002, "grad_norm": 2.9095473289489746, "learning_rate": 1.2830274416365334e-05, "loss": 0.3192, "num_input_tokens_seen": 7111912, "step": 26540 }, { "epoch": 6.958060288335518, "grad_norm": 2.1235785484313965, "learning_rate": 1.2820285000453309e-05, "loss": 0.253, "num_input_tokens_seen": 7114440, "step": 26545 }, { "epoch": 6.9593709043250325, "grad_norm": 2.5533218383789062, "learning_rate": 1.281029813373459e-05, "loss": 0.1503, "num_input_tokens_seen": 7115928, "step": 26550 }, { "epoch": 6.960681520314548, "grad_norm": 1.9069334268569946, "learning_rate": 1.2800313818299403e-05, "loss": 0.3686, "num_input_tokens_seen": 7117144, "step": 26555 }, { "epoch": 6.961992136304063, "grad_norm": 3.437084674835205, "learning_rate": 1.2790332056237458e-05, "loss": 0.439, "num_input_tokens_seen": 7118552, "step": 26560 }, { "epoch": 6.963302752293578, "grad_norm": 3.3195083141326904, "learning_rate": 1.2780352849637928e-05, "loss": 0.2111, "num_input_tokens_seen": 7119640, "step": 26565 }, { "epoch": 6.964613368283093, "grad_norm": 1.3326771259307861, "learning_rate": 1.2770376200589423e-05, "loss": 0.2995, "num_input_tokens_seen": 7121192, "step": 26570 }, { "epoch": 6.965923984272608, "grad_norm": 2.9842259883880615, "learning_rate": 1.2760402111180048e-05, "loss": 0.4483, "num_input_tokens_seen": 7122360, "step": 26575 }, { "epoch": 6.967234600262123, "grad_norm": 1.122489333152771, "learning_rate": 1.2750430583497363e-05, "loss": 0.3772, "num_input_tokens_seen": 7124104, "step": 26580 }, { "epoch": 6.968545216251639, "grad_norm": 5.45490837097168, "learning_rate": 1.2740461619628392e-05, "loss": 0.2276, "num_input_tokens_seen": 7125496, "step": 26585 }, { "epoch": 6.969855832241153, "grad_norm": 1.711745023727417, "learning_rate": 1.2730495221659607e-05, "loss": 0.3266, "num_input_tokens_seen": 7126792, "step": 26590 }, { "epoch": 6.971166448230669, "grad_norm": 3.821594476699829, "learning_rate": 1.2720531391676987e-05, "loss": 0.2482, "num_input_tokens_seen": 7128536, "step": 26595 }, { "epoch": 6.972477064220183, "grad_norm": 5.006137847900391, "learning_rate": 1.2710570131765937e-05, "loss": 0.2095, "num_input_tokens_seen": 7129896, "step": 26600 }, { "epoch": 6.973787680209698, "grad_norm": 3.11627197265625, "learning_rate": 1.2700611444011312e-05, "loss": 0.2841, "num_input_tokens_seen": 7131080, "step": 26605 }, { "epoch": 6.975098296199214, "grad_norm": 1.4280518293380737, "learning_rate": 1.2690655330497464e-05, "loss": 0.2077, "num_input_tokens_seen": 7132504, "step": 26610 }, { "epoch": 6.976408912188729, "grad_norm": 1.788801670074463, "learning_rate": 1.2680701793308187e-05, "loss": 0.1352, "num_input_tokens_seen": 7133720, "step": 26615 }, { "epoch": 6.977719528178244, "grad_norm": 2.7688231468200684, "learning_rate": 1.2670750834526735e-05, "loss": 0.2663, "num_input_tokens_seen": 7135176, "step": 26620 }, { "epoch": 6.9790301441677585, "grad_norm": 2.5871148109436035, "learning_rate": 1.266080245623583e-05, "loss": 0.3114, "num_input_tokens_seen": 7136648, "step": 26625 }, { "epoch": 6.980340760157274, "grad_norm": 1.6878184080123901, "learning_rate": 1.2650856660517652e-05, "loss": 0.4235, "num_input_tokens_seen": 7137640, "step": 26630 }, { "epoch": 6.981651376146789, "grad_norm": 2.7197396755218506, "learning_rate": 1.2640913449453837e-05, "loss": 0.2921, "num_input_tokens_seen": 7138824, "step": 26635 }, { "epoch": 6.982961992136304, "grad_norm": 1.7500115633010864, "learning_rate": 1.2630972825125482e-05, "loss": 0.328, "num_input_tokens_seen": 7140376, "step": 26640 }, { "epoch": 6.984272608125819, "grad_norm": 2.070449113845825, "learning_rate": 1.2621034789613151e-05, "loss": 0.4429, "num_input_tokens_seen": 7141896, "step": 26645 }, { "epoch": 6.985583224115334, "grad_norm": 1.5306406021118164, "learning_rate": 1.2611099344996832e-05, "loss": 0.2321, "num_input_tokens_seen": 7143432, "step": 26650 }, { "epoch": 6.986893840104849, "grad_norm": 2.4949142932891846, "learning_rate": 1.2601166493356023e-05, "loss": 0.2276, "num_input_tokens_seen": 7144520, "step": 26655 }, { "epoch": 6.988204456094365, "grad_norm": 1.7446280717849731, "learning_rate": 1.2591236236769643e-05, "loss": 0.2371, "num_input_tokens_seen": 7145816, "step": 26660 }, { "epoch": 6.989515072083879, "grad_norm": 1.533094882965088, "learning_rate": 1.2581308577316079e-05, "loss": 0.2778, "num_input_tokens_seen": 7147192, "step": 26665 }, { "epoch": 6.990825688073395, "grad_norm": 2.323965549468994, "learning_rate": 1.2571383517073172e-05, "loss": 0.2452, "num_input_tokens_seen": 7148488, "step": 26670 }, { "epoch": 6.992136304062909, "grad_norm": 4.810091495513916, "learning_rate": 1.2561461058118219e-05, "loss": 0.3604, "num_input_tokens_seen": 7149256, "step": 26675 }, { "epoch": 6.993446920052425, "grad_norm": 2.6772878170013428, "learning_rate": 1.2551541202527984e-05, "loss": 0.2237, "num_input_tokens_seen": 7151000, "step": 26680 }, { "epoch": 6.99475753604194, "grad_norm": 1.0656349658966064, "learning_rate": 1.2541623952378657e-05, "loss": 0.2342, "num_input_tokens_seen": 7152008, "step": 26685 }, { "epoch": 6.996068152031455, "grad_norm": 2.455427885055542, "learning_rate": 1.2531709309745907e-05, "loss": 0.202, "num_input_tokens_seen": 7153320, "step": 26690 }, { "epoch": 6.99737876802097, "grad_norm": 3.3356800079345703, "learning_rate": 1.2521797276704858e-05, "loss": 0.404, "num_input_tokens_seen": 7154424, "step": 26695 }, { "epoch": 6.9986893840104845, "grad_norm": 1.469804048538208, "learning_rate": 1.251188785533008e-05, "loss": 0.2865, "num_input_tokens_seen": 7156472, "step": 26700 }, { "epoch": 7.0, "grad_norm": 3.538595199584961, "learning_rate": 1.2501981047695583e-05, "loss": 0.1817, "num_input_tokens_seen": 7157624, "step": 26705 }, { "epoch": 7.001310615989515, "grad_norm": 0.9130023121833801, "learning_rate": 1.2492076855874868e-05, "loss": 0.1245, "num_input_tokens_seen": 7158776, "step": 26710 }, { "epoch": 7.001834862385321, "eval_loss": 0.6657854318618774, "eval_runtime": 16.709, "eval_samples_per_second": 50.751, "eval_steps_per_second": 25.376, "num_input_tokens_seen": 7159688, "step": 26712 }, { "epoch": 7.00262123197903, "grad_norm": 1.0543054342269897, "learning_rate": 1.2482175281940855e-05, "loss": 0.1443, "num_input_tokens_seen": 7160488, "step": 26715 }, { "epoch": 7.003931847968545, "grad_norm": 1.635632038116455, "learning_rate": 1.2472276327965938e-05, "loss": 0.2202, "num_input_tokens_seen": 7162232, "step": 26720 }, { "epoch": 7.00524246395806, "grad_norm": 1.6818921566009521, "learning_rate": 1.2462379996021928e-05, "loss": 0.2708, "num_input_tokens_seen": 7164264, "step": 26725 }, { "epoch": 7.006553079947575, "grad_norm": 2.138751268386841, "learning_rate": 1.2452486288180124e-05, "loss": 0.1202, "num_input_tokens_seen": 7165720, "step": 26730 }, { "epoch": 7.007863695937091, "grad_norm": 1.8140935897827148, "learning_rate": 1.2442595206511257e-05, "loss": 0.2193, "num_input_tokens_seen": 7167048, "step": 26735 }, { "epoch": 7.009174311926605, "grad_norm": 2.391697883605957, "learning_rate": 1.2432706753085516e-05, "loss": 0.1876, "num_input_tokens_seen": 7169224, "step": 26740 }, { "epoch": 7.010484927916121, "grad_norm": 1.9397395849227905, "learning_rate": 1.2422820929972537e-05, "loss": 0.1483, "num_input_tokens_seen": 7170472, "step": 26745 }, { "epoch": 7.011795543905635, "grad_norm": 0.6848123669624329, "learning_rate": 1.2412937739241407e-05, "loss": 0.1439, "num_input_tokens_seen": 7171416, "step": 26750 }, { "epoch": 7.013106159895151, "grad_norm": 1.8236838579177856, "learning_rate": 1.2403057182960655e-05, "loss": 0.1055, "num_input_tokens_seen": 7172856, "step": 26755 }, { "epoch": 7.014416775884666, "grad_norm": 3.4868996143341064, "learning_rate": 1.2393179263198273e-05, "loss": 0.1643, "num_input_tokens_seen": 7174200, "step": 26760 }, { "epoch": 7.015727391874181, "grad_norm": 1.3386534452438354, "learning_rate": 1.2383303982021688e-05, "loss": 0.1721, "num_input_tokens_seen": 7175384, "step": 26765 }, { "epoch": 7.017038007863696, "grad_norm": 2.080150842666626, "learning_rate": 1.237343134149778e-05, "loss": 0.179, "num_input_tokens_seen": 7176824, "step": 26770 }, { "epoch": 7.018348623853211, "grad_norm": 2.6052446365356445, "learning_rate": 1.2363561343692873e-05, "loss": 0.2636, "num_input_tokens_seen": 7178040, "step": 26775 }, { "epoch": 7.019659239842726, "grad_norm": 0.8706895112991333, "learning_rate": 1.2353693990672743e-05, "loss": 0.1491, "num_input_tokens_seen": 7179224, "step": 26780 }, { "epoch": 7.0209698558322415, "grad_norm": 4.230594635009766, "learning_rate": 1.2343829284502612e-05, "loss": 0.1922, "num_input_tokens_seen": 7180280, "step": 26785 }, { "epoch": 7.022280471821756, "grad_norm": 2.5949866771698, "learning_rate": 1.2333967227247143e-05, "loss": 0.1117, "num_input_tokens_seen": 7181624, "step": 26790 }, { "epoch": 7.023591087811272, "grad_norm": 2.10923433303833, "learning_rate": 1.2324107820970445e-05, "loss": 0.3219, "num_input_tokens_seen": 7182872, "step": 26795 }, { "epoch": 7.024901703800786, "grad_norm": 1.057800054550171, "learning_rate": 1.231425106773609e-05, "loss": 0.1584, "num_input_tokens_seen": 7184120, "step": 26800 }, { "epoch": 7.026212319790301, "grad_norm": 2.8530967235565186, "learning_rate": 1.2304396969607057e-05, "loss": 0.1717, "num_input_tokens_seen": 7185000, "step": 26805 }, { "epoch": 7.027522935779817, "grad_norm": 2.1300723552703857, "learning_rate": 1.2294545528645804e-05, "loss": 0.1994, "num_input_tokens_seen": 7186456, "step": 26810 }, { "epoch": 7.028833551769331, "grad_norm": 1.9189119338989258, "learning_rate": 1.2284696746914216e-05, "loss": 0.285, "num_input_tokens_seen": 7187496, "step": 26815 }, { "epoch": 7.030144167758847, "grad_norm": 1.8213449716567993, "learning_rate": 1.227485062647362e-05, "loss": 0.1229, "num_input_tokens_seen": 7189144, "step": 26820 }, { "epoch": 7.031454783748361, "grad_norm": 0.9465539455413818, "learning_rate": 1.226500716938481e-05, "loss": 0.0857, "num_input_tokens_seen": 7190248, "step": 26825 }, { "epoch": 7.032765399737877, "grad_norm": 1.6380267143249512, "learning_rate": 1.2255166377707995e-05, "loss": 0.1263, "num_input_tokens_seen": 7191768, "step": 26830 }, { "epoch": 7.034076015727392, "grad_norm": 3.5983474254608154, "learning_rate": 1.2245328253502844e-05, "loss": 0.1922, "num_input_tokens_seen": 7193272, "step": 26835 }, { "epoch": 7.035386631716907, "grad_norm": 1.4829351902008057, "learning_rate": 1.2235492798828444e-05, "loss": 0.1647, "num_input_tokens_seen": 7194792, "step": 26840 }, { "epoch": 7.036697247706422, "grad_norm": 2.937469720840454, "learning_rate": 1.2225660015743343e-05, "loss": 0.2499, "num_input_tokens_seen": 7195992, "step": 26845 }, { "epoch": 7.038007863695937, "grad_norm": 1.902072548866272, "learning_rate": 1.2215829906305531e-05, "loss": 0.2402, "num_input_tokens_seen": 7197544, "step": 26850 }, { "epoch": 7.039318479685452, "grad_norm": 2.6508755683898926, "learning_rate": 1.2206002472572428e-05, "loss": 0.2414, "num_input_tokens_seen": 7198904, "step": 26855 }, { "epoch": 7.0406290956749675, "grad_norm": 1.4231199026107788, "learning_rate": 1.2196177716600904e-05, "loss": 0.1461, "num_input_tokens_seen": 7200200, "step": 26860 }, { "epoch": 7.041939711664482, "grad_norm": 1.7390528917312622, "learning_rate": 1.2186355640447259e-05, "loss": 0.2164, "num_input_tokens_seen": 7201576, "step": 26865 }, { "epoch": 7.043250327653998, "grad_norm": 1.4358378648757935, "learning_rate": 1.2176536246167235e-05, "loss": 0.0961, "num_input_tokens_seen": 7203272, "step": 26870 }, { "epoch": 7.044560943643512, "grad_norm": 1.3010835647583008, "learning_rate": 1.2166719535816024e-05, "loss": 0.1193, "num_input_tokens_seen": 7204504, "step": 26875 }, { "epoch": 7.045871559633028, "grad_norm": 1.7512954473495483, "learning_rate": 1.2156905511448235e-05, "loss": 0.0979, "num_input_tokens_seen": 7205544, "step": 26880 }, { "epoch": 7.047182175622543, "grad_norm": 2.924560070037842, "learning_rate": 1.2147094175117932e-05, "loss": 0.2667, "num_input_tokens_seen": 7206760, "step": 26885 }, { "epoch": 7.048492791612058, "grad_norm": 4.151130676269531, "learning_rate": 1.2137285528878611e-05, "loss": 0.2045, "num_input_tokens_seen": 7208264, "step": 26890 }, { "epoch": 7.049803407601573, "grad_norm": 4.626003265380859, "learning_rate": 1.2127479574783205e-05, "loss": 0.2859, "num_input_tokens_seen": 7210936, "step": 26895 }, { "epoch": 7.051114023591087, "grad_norm": 1.5223242044448853, "learning_rate": 1.211767631488408e-05, "loss": 0.1505, "num_input_tokens_seen": 7212200, "step": 26900 }, { "epoch": 7.052424639580603, "grad_norm": 0.4196576476097107, "learning_rate": 1.2107875751233045e-05, "loss": 0.2422, "num_input_tokens_seen": 7213272, "step": 26905 }, { "epoch": 7.053735255570118, "grad_norm": 1.4536633491516113, "learning_rate": 1.2098077885881336e-05, "loss": 0.1253, "num_input_tokens_seen": 7214472, "step": 26910 }, { "epoch": 7.055045871559633, "grad_norm": 1.8766010999679565, "learning_rate": 1.2088282720879646e-05, "loss": 0.2062, "num_input_tokens_seen": 7215608, "step": 26915 }, { "epoch": 7.056356487549148, "grad_norm": 3.637805461883545, "learning_rate": 1.2078490258278061e-05, "loss": 0.2422, "num_input_tokens_seen": 7216952, "step": 26920 }, { "epoch": 7.057667103538663, "grad_norm": 2.759347915649414, "learning_rate": 1.206870050012614e-05, "loss": 0.1905, "num_input_tokens_seen": 7218824, "step": 26925 }, { "epoch": 7.058977719528178, "grad_norm": 2.678624153137207, "learning_rate": 1.205891344847286e-05, "loss": 0.1506, "num_input_tokens_seen": 7220136, "step": 26930 }, { "epoch": 7.0602883355176935, "grad_norm": 5.715234756469727, "learning_rate": 1.2049129105366625e-05, "loss": 0.238, "num_input_tokens_seen": 7221112, "step": 26935 }, { "epoch": 7.061598951507208, "grad_norm": 1.0641841888427734, "learning_rate": 1.20393474728553e-05, "loss": 0.2042, "num_input_tokens_seen": 7222696, "step": 26940 }, { "epoch": 7.062909567496724, "grad_norm": 1.5144569873809814, "learning_rate": 1.2029568552986156e-05, "loss": 0.1535, "num_input_tokens_seen": 7223928, "step": 26945 }, { "epoch": 7.064220183486238, "grad_norm": 5.744327068328857, "learning_rate": 1.2019792347805902e-05, "loss": 0.211, "num_input_tokens_seen": 7225160, "step": 26950 }, { "epoch": 7.065530799475754, "grad_norm": 1.6774028539657593, "learning_rate": 1.2010018859360691e-05, "loss": 0.2735, "num_input_tokens_seen": 7226728, "step": 26955 }, { "epoch": 7.066841415465269, "grad_norm": 1.5821352005004883, "learning_rate": 1.2000248089696078e-05, "loss": 0.1436, "num_input_tokens_seen": 7227944, "step": 26960 }, { "epoch": 7.068152031454784, "grad_norm": 2.211000919342041, "learning_rate": 1.1990480040857076e-05, "loss": 0.1762, "num_input_tokens_seen": 7229208, "step": 26965 }, { "epoch": 7.069462647444299, "grad_norm": 1.4113107919692993, "learning_rate": 1.1980714714888127e-05, "loss": 0.1023, "num_input_tokens_seen": 7230616, "step": 26970 }, { "epoch": 7.070773263433814, "grad_norm": 3.231330156326294, "learning_rate": 1.197095211383309e-05, "loss": 0.2856, "num_input_tokens_seen": 7231896, "step": 26975 }, { "epoch": 7.072083879423329, "grad_norm": 2.9505186080932617, "learning_rate": 1.1961192239735264e-05, "loss": 0.2797, "num_input_tokens_seen": 7233496, "step": 26980 }, { "epoch": 7.073394495412844, "grad_norm": 1.309305191040039, "learning_rate": 1.1951435094637373e-05, "loss": 0.1461, "num_input_tokens_seen": 7234936, "step": 26985 }, { "epoch": 7.074705111402359, "grad_norm": 1.943398356437683, "learning_rate": 1.194168068058157e-05, "loss": 0.1845, "num_input_tokens_seen": 7236200, "step": 26990 }, { "epoch": 7.076015727391874, "grad_norm": 1.6783092021942139, "learning_rate": 1.1931928999609438e-05, "loss": 0.1219, "num_input_tokens_seen": 7237672, "step": 26995 }, { "epoch": 7.077326343381389, "grad_norm": 0.91379714012146, "learning_rate": 1.1922180053761985e-05, "loss": 0.1887, "num_input_tokens_seen": 7238952, "step": 27000 }, { "epoch": 7.078636959370904, "grad_norm": 1.2796794176101685, "learning_rate": 1.1912433845079652e-05, "loss": 0.3314, "num_input_tokens_seen": 7240296, "step": 27005 }, { "epoch": 7.0799475753604195, "grad_norm": 2.708179235458374, "learning_rate": 1.19026903756023e-05, "loss": 0.1782, "num_input_tokens_seen": 7241448, "step": 27010 }, { "epoch": 7.081258191349934, "grad_norm": 2.2169454097747803, "learning_rate": 1.1892949647369223e-05, "loss": 0.1797, "num_input_tokens_seen": 7242920, "step": 27015 }, { "epoch": 7.08256880733945, "grad_norm": 2.6228957176208496, "learning_rate": 1.1883211662419137e-05, "loss": 0.1374, "num_input_tokens_seen": 7244488, "step": 27020 }, { "epoch": 7.083879423328964, "grad_norm": 2.8414690494537354, "learning_rate": 1.1873476422790187e-05, "loss": 0.2168, "num_input_tokens_seen": 7245928, "step": 27025 }, { "epoch": 7.08519003931848, "grad_norm": 2.381960868835449, "learning_rate": 1.1863743930519939e-05, "loss": 0.257, "num_input_tokens_seen": 7247208, "step": 27030 }, { "epoch": 7.086500655307995, "grad_norm": 2.605909824371338, "learning_rate": 1.1854014187645398e-05, "loss": 0.2074, "num_input_tokens_seen": 7248360, "step": 27035 }, { "epoch": 7.08781127129751, "grad_norm": 1.01252281665802, "learning_rate": 1.1844287196202967e-05, "loss": 0.1261, "num_input_tokens_seen": 7249800, "step": 27040 }, { "epoch": 7.089121887287025, "grad_norm": 2.1162760257720947, "learning_rate": 1.183456295822849e-05, "loss": 0.1663, "num_input_tokens_seen": 7251448, "step": 27045 }, { "epoch": 7.09043250327654, "grad_norm": 2.6340761184692383, "learning_rate": 1.1824841475757232e-05, "loss": 0.3197, "num_input_tokens_seen": 7252712, "step": 27050 }, { "epoch": 7.091743119266055, "grad_norm": 2.138136386871338, "learning_rate": 1.1815122750823896e-05, "loss": 0.2635, "num_input_tokens_seen": 7254968, "step": 27055 }, { "epoch": 7.0930537352555705, "grad_norm": 3.509516954421997, "learning_rate": 1.1805406785462587e-05, "loss": 0.1393, "num_input_tokens_seen": 7256312, "step": 27060 }, { "epoch": 7.094364351245085, "grad_norm": 2.93847393989563, "learning_rate": 1.1795693581706838e-05, "loss": 0.1685, "num_input_tokens_seen": 7257528, "step": 27065 }, { "epoch": 7.095674967234601, "grad_norm": 1.7351616621017456, "learning_rate": 1.1785983141589618e-05, "loss": 0.1473, "num_input_tokens_seen": 7258808, "step": 27070 }, { "epoch": 7.096985583224115, "grad_norm": 1.7774673700332642, "learning_rate": 1.1776275467143283e-05, "loss": 0.1971, "num_input_tokens_seen": 7260200, "step": 27075 }, { "epoch": 7.09829619921363, "grad_norm": 4.406271934509277, "learning_rate": 1.1766570560399647e-05, "loss": 0.1615, "num_input_tokens_seen": 7261320, "step": 27080 }, { "epoch": 7.0996068152031455, "grad_norm": 2.125261068344116, "learning_rate": 1.1756868423389925e-05, "loss": 0.2132, "num_input_tokens_seen": 7262696, "step": 27085 }, { "epoch": 7.10091743119266, "grad_norm": 2.072974681854248, "learning_rate": 1.1747169058144761e-05, "loss": 0.2605, "num_input_tokens_seen": 7263944, "step": 27090 }, { "epoch": 7.102228047182176, "grad_norm": 1.9412260055541992, "learning_rate": 1.1737472466694213e-05, "loss": 0.2586, "num_input_tokens_seen": 7265112, "step": 27095 }, { "epoch": 7.10353866317169, "grad_norm": 3.6648168563842773, "learning_rate": 1.1727778651067761e-05, "loss": 0.1419, "num_input_tokens_seen": 7266424, "step": 27100 }, { "epoch": 7.104849279161206, "grad_norm": 2.2473604679107666, "learning_rate": 1.1718087613294307e-05, "loss": 0.225, "num_input_tokens_seen": 7267848, "step": 27105 }, { "epoch": 7.1061598951507206, "grad_norm": 1.3061103820800781, "learning_rate": 1.1708399355402163e-05, "loss": 0.173, "num_input_tokens_seen": 7269416, "step": 27110 }, { "epoch": 7.107470511140236, "grad_norm": 1.7051465511322021, "learning_rate": 1.1698713879419072e-05, "loss": 0.2101, "num_input_tokens_seen": 7271032, "step": 27115 }, { "epoch": 7.108781127129751, "grad_norm": 3.0182318687438965, "learning_rate": 1.168903118737218e-05, "loss": 0.1755, "num_input_tokens_seen": 7272216, "step": 27120 }, { "epoch": 7.110091743119266, "grad_norm": 1.533500075340271, "learning_rate": 1.1679351281288062e-05, "loss": 0.2155, "num_input_tokens_seen": 7273688, "step": 27125 }, { "epoch": 7.111402359108781, "grad_norm": 2.62107253074646, "learning_rate": 1.1669674163192703e-05, "loss": 0.2981, "num_input_tokens_seen": 7274808, "step": 27130 }, { "epoch": 7.1127129750982965, "grad_norm": 1.8501136302947998, "learning_rate": 1.165999983511151e-05, "loss": 0.2386, "num_input_tokens_seen": 7276664, "step": 27135 }, { "epoch": 7.114023591087811, "grad_norm": 2.0646440982818604, "learning_rate": 1.1650328299069301e-05, "loss": 0.1542, "num_input_tokens_seen": 7277912, "step": 27140 }, { "epoch": 7.115334207077327, "grad_norm": 3.5200936794281006, "learning_rate": 1.1640659557090313e-05, "loss": 0.1714, "num_input_tokens_seen": 7278920, "step": 27145 }, { "epoch": 7.116644823066841, "grad_norm": 7.323924541473389, "learning_rate": 1.1630993611198207e-05, "loss": 0.254, "num_input_tokens_seen": 7280184, "step": 27150 }, { "epoch": 7.117955439056357, "grad_norm": 0.8671079874038696, "learning_rate": 1.1621330463416027e-05, "loss": 0.2269, "num_input_tokens_seen": 7281736, "step": 27155 }, { "epoch": 7.1192660550458715, "grad_norm": 3.4464738368988037, "learning_rate": 1.1611670115766265e-05, "loss": 0.1948, "num_input_tokens_seen": 7283112, "step": 27160 }, { "epoch": 7.120576671035387, "grad_norm": 4.470269680023193, "learning_rate": 1.1602012570270806e-05, "loss": 0.1242, "num_input_tokens_seen": 7284136, "step": 27165 }, { "epoch": 7.121887287024902, "grad_norm": 1.4973258972167969, "learning_rate": 1.1592357828950976e-05, "loss": 0.2809, "num_input_tokens_seen": 7285576, "step": 27170 }, { "epoch": 7.123197903014416, "grad_norm": 1.0281286239624023, "learning_rate": 1.1582705893827483e-05, "loss": 0.1127, "num_input_tokens_seen": 7286856, "step": 27175 }, { "epoch": 7.124508519003932, "grad_norm": 1.745289921760559, "learning_rate": 1.1573056766920465e-05, "loss": 0.1378, "num_input_tokens_seen": 7288168, "step": 27180 }, { "epoch": 7.1258191349934465, "grad_norm": 2.1499338150024414, "learning_rate": 1.1563410450249468e-05, "loss": 0.1471, "num_input_tokens_seen": 7289448, "step": 27185 }, { "epoch": 7.127129750982962, "grad_norm": 2.499143123626709, "learning_rate": 1.1553766945833452e-05, "loss": 0.1826, "num_input_tokens_seen": 7290744, "step": 27190 }, { "epoch": 7.128440366972477, "grad_norm": 1.2673308849334717, "learning_rate": 1.1544126255690774e-05, "loss": 0.1559, "num_input_tokens_seen": 7292296, "step": 27195 }, { "epoch": 7.129750982961992, "grad_norm": 6.271773338317871, "learning_rate": 1.153448838183922e-05, "loss": 0.1489, "num_input_tokens_seen": 7293336, "step": 27200 }, { "epoch": 7.131061598951507, "grad_norm": 1.9893853664398193, "learning_rate": 1.1524853326295982e-05, "loss": 0.1325, "num_input_tokens_seen": 7294392, "step": 27205 }, { "epoch": 7.1323722149410225, "grad_norm": 0.9682709574699402, "learning_rate": 1.151522109107766e-05, "loss": 0.1818, "num_input_tokens_seen": 7295816, "step": 27210 }, { "epoch": 7.133682830930537, "grad_norm": 1.6595070362091064, "learning_rate": 1.150559167820026e-05, "loss": 0.3032, "num_input_tokens_seen": 7297320, "step": 27215 }, { "epoch": 7.134993446920053, "grad_norm": 3.3437912464141846, "learning_rate": 1.14959650896792e-05, "loss": 0.1686, "num_input_tokens_seen": 7298440, "step": 27220 }, { "epoch": 7.136304062909567, "grad_norm": 1.1297025680541992, "learning_rate": 1.1486341327529332e-05, "loss": 0.1449, "num_input_tokens_seen": 7300104, "step": 27225 }, { "epoch": 7.137614678899083, "grad_norm": 1.687880039215088, "learning_rate": 1.1476720393764864e-05, "loss": 0.1509, "num_input_tokens_seen": 7301224, "step": 27230 }, { "epoch": 7.1389252948885975, "grad_norm": 0.7995802164077759, "learning_rate": 1.146710229039945e-05, "loss": 0.1552, "num_input_tokens_seen": 7302616, "step": 27235 }, { "epoch": 7.140235910878113, "grad_norm": 1.4614832401275635, "learning_rate": 1.1457487019446145e-05, "loss": 0.2349, "num_input_tokens_seen": 7303752, "step": 27240 }, { "epoch": 7.141546526867628, "grad_norm": 0.9592230319976807, "learning_rate": 1.1447874582917407e-05, "loss": 0.1422, "num_input_tokens_seen": 7304968, "step": 27245 }, { "epoch": 7.142857142857143, "grad_norm": 2.163356065750122, "learning_rate": 1.1438264982825104e-05, "loss": 0.1866, "num_input_tokens_seen": 7306248, "step": 27250 }, { "epoch": 7.144167758846658, "grad_norm": 5.752673625946045, "learning_rate": 1.1428658221180506e-05, "loss": 0.1653, "num_input_tokens_seen": 7308120, "step": 27255 }, { "epoch": 7.145478374836173, "grad_norm": 0.038059089332818985, "learning_rate": 1.1419054299994292e-05, "loss": 0.175, "num_input_tokens_seen": 7309144, "step": 27260 }, { "epoch": 7.146788990825688, "grad_norm": 2.4828855991363525, "learning_rate": 1.1409453221276545e-05, "loss": 0.2448, "num_input_tokens_seen": 7310632, "step": 27265 }, { "epoch": 7.148099606815203, "grad_norm": 1.4939231872558594, "learning_rate": 1.1399854987036768e-05, "loss": 0.1731, "num_input_tokens_seen": 7311848, "step": 27270 }, { "epoch": 7.149410222804718, "grad_norm": 6.534052848815918, "learning_rate": 1.1390259599283835e-05, "loss": 0.2228, "num_input_tokens_seen": 7313048, "step": 27275 }, { "epoch": 7.150720838794233, "grad_norm": 18.89065933227539, "learning_rate": 1.1380667060026041e-05, "loss": 0.1775, "num_input_tokens_seen": 7313992, "step": 27280 }, { "epoch": 7.1520314547837485, "grad_norm": 1.7153112888336182, "learning_rate": 1.1371077371271108e-05, "loss": 0.1645, "num_input_tokens_seen": 7315464, "step": 27285 }, { "epoch": 7.153342070773263, "grad_norm": 1.6619694232940674, "learning_rate": 1.136149053502613e-05, "loss": 0.2283, "num_input_tokens_seen": 7316840, "step": 27290 }, { "epoch": 7.154652686762779, "grad_norm": 1.5162651538848877, "learning_rate": 1.1351906553297622e-05, "loss": 0.2032, "num_input_tokens_seen": 7318232, "step": 27295 }, { "epoch": 7.155963302752293, "grad_norm": 2.170682668685913, "learning_rate": 1.1342325428091488e-05, "loss": 0.1119, "num_input_tokens_seen": 7319320, "step": 27300 }, { "epoch": 7.157273918741809, "grad_norm": 2.9140563011169434, "learning_rate": 1.1332747161413054e-05, "loss": 0.1332, "num_input_tokens_seen": 7320600, "step": 27305 }, { "epoch": 7.1585845347313235, "grad_norm": 1.1790186166763306, "learning_rate": 1.1323171755267015e-05, "loss": 0.1564, "num_input_tokens_seen": 7322072, "step": 27310 }, { "epoch": 7.159895150720839, "grad_norm": 2.021036386489868, "learning_rate": 1.1313599211657494e-05, "loss": 0.1369, "num_input_tokens_seen": 7323032, "step": 27315 }, { "epoch": 7.161205766710354, "grad_norm": 1.2740497589111328, "learning_rate": 1.130402953258801e-05, "loss": 0.2224, "num_input_tokens_seen": 7324472, "step": 27320 }, { "epoch": 7.162516382699869, "grad_norm": 3.4810287952423096, "learning_rate": 1.129446272006148e-05, "loss": 0.3237, "num_input_tokens_seen": 7325704, "step": 27325 }, { "epoch": 7.163826998689384, "grad_norm": 1.9162752628326416, "learning_rate": 1.1284898776080222e-05, "loss": 0.1917, "num_input_tokens_seen": 7327368, "step": 27330 }, { "epoch": 7.165137614678899, "grad_norm": 5.22629451751709, "learning_rate": 1.127533770264594e-05, "loss": 0.2437, "num_input_tokens_seen": 7328664, "step": 27335 }, { "epoch": 7.166448230668414, "grad_norm": 1.4933265447616577, "learning_rate": 1.1265779501759786e-05, "loss": 0.2249, "num_input_tokens_seen": 7329912, "step": 27340 }, { "epoch": 7.16775884665793, "grad_norm": 1.9103487730026245, "learning_rate": 1.125622417542224e-05, "loss": 0.1981, "num_input_tokens_seen": 7331208, "step": 27345 }, { "epoch": 7.169069462647444, "grad_norm": 1.4217923879623413, "learning_rate": 1.1246671725633227e-05, "loss": 0.1337, "num_input_tokens_seen": 7332216, "step": 27350 }, { "epoch": 7.17038007863696, "grad_norm": 1.0020629167556763, "learning_rate": 1.1237122154392058e-05, "loss": 0.0758, "num_input_tokens_seen": 7333784, "step": 27355 }, { "epoch": 7.1716906946264745, "grad_norm": 2.866149663925171, "learning_rate": 1.122757546369744e-05, "loss": 0.2502, "num_input_tokens_seen": 7335304, "step": 27360 }, { "epoch": 7.173001310615989, "grad_norm": 1.9756108522415161, "learning_rate": 1.1218031655547482e-05, "loss": 0.1642, "num_input_tokens_seen": 7336680, "step": 27365 }, { "epoch": 7.174311926605505, "grad_norm": 1.7211952209472656, "learning_rate": 1.1208490731939683e-05, "loss": 0.233, "num_input_tokens_seen": 7338088, "step": 27370 }, { "epoch": 7.175622542595019, "grad_norm": 1.6190333366394043, "learning_rate": 1.1198952694870945e-05, "loss": 0.2352, "num_input_tokens_seen": 7339384, "step": 27375 }, { "epoch": 7.176933158584535, "grad_norm": 5.766364097595215, "learning_rate": 1.1189417546337564e-05, "loss": 0.2687, "num_input_tokens_seen": 7340648, "step": 27380 }, { "epoch": 7.1782437745740495, "grad_norm": 2.1443047523498535, "learning_rate": 1.1179885288335235e-05, "loss": 0.1412, "num_input_tokens_seen": 7341912, "step": 27385 }, { "epoch": 7.179554390563565, "grad_norm": 4.179052829742432, "learning_rate": 1.1170355922859024e-05, "loss": 0.198, "num_input_tokens_seen": 7343272, "step": 27390 }, { "epoch": 7.18086500655308, "grad_norm": 4.266551494598389, "learning_rate": 1.1160829451903418e-05, "loss": 0.1973, "num_input_tokens_seen": 7344696, "step": 27395 }, { "epoch": 7.182175622542595, "grad_norm": 2.6195709705352783, "learning_rate": 1.1151305877462304e-05, "loss": 0.1852, "num_input_tokens_seen": 7346008, "step": 27400 }, { "epoch": 7.18348623853211, "grad_norm": 1.663454532623291, "learning_rate": 1.1141785201528936e-05, "loss": 0.1757, "num_input_tokens_seen": 7347512, "step": 27405 }, { "epoch": 7.184796854521625, "grad_norm": 0.5521488785743713, "learning_rate": 1.1132267426095985e-05, "loss": 0.1822, "num_input_tokens_seen": 7348936, "step": 27410 }, { "epoch": 7.18610747051114, "grad_norm": 1.5403611660003662, "learning_rate": 1.11227525531555e-05, "loss": 0.114, "num_input_tokens_seen": 7350184, "step": 27415 }, { "epoch": 7.187418086500656, "grad_norm": 2.7626755237579346, "learning_rate": 1.1113240584698927e-05, "loss": 0.1567, "num_input_tokens_seen": 7351432, "step": 27420 }, { "epoch": 7.18872870249017, "grad_norm": 1.266979455947876, "learning_rate": 1.1103731522717115e-05, "loss": 0.1336, "num_input_tokens_seen": 7352504, "step": 27425 }, { "epoch": 7.190039318479686, "grad_norm": 1.4258825778961182, "learning_rate": 1.1094225369200272e-05, "loss": 0.1637, "num_input_tokens_seen": 7354024, "step": 27430 }, { "epoch": 7.1913499344692005, "grad_norm": 3.816251516342163, "learning_rate": 1.1084722126138037e-05, "loss": 0.3055, "num_input_tokens_seen": 7355336, "step": 27435 }, { "epoch": 7.192660550458716, "grad_norm": 1.7219024896621704, "learning_rate": 1.1075221795519413e-05, "loss": 0.232, "num_input_tokens_seen": 7356696, "step": 27440 }, { "epoch": 7.193971166448231, "grad_norm": 1.7668417692184448, "learning_rate": 1.106572437933281e-05, "loss": 0.1942, "num_input_tokens_seen": 7358312, "step": 27445 }, { "epoch": 7.195281782437746, "grad_norm": 3.61645770072937, "learning_rate": 1.1056229879566004e-05, "loss": 0.1821, "num_input_tokens_seen": 7359512, "step": 27450 }, { "epoch": 7.196592398427261, "grad_norm": 1.1250476837158203, "learning_rate": 1.1046738298206205e-05, "loss": 0.2535, "num_input_tokens_seen": 7360808, "step": 27455 }, { "epoch": 7.1979030144167755, "grad_norm": 1.9767892360687256, "learning_rate": 1.103724963723998e-05, "loss": 0.2577, "num_input_tokens_seen": 7362408, "step": 27460 }, { "epoch": 7.199213630406291, "grad_norm": 1.7213493585586548, "learning_rate": 1.1027763898653268e-05, "loss": 0.1674, "num_input_tokens_seen": 7364184, "step": 27465 }, { "epoch": 7.200524246395806, "grad_norm": 0.4514981806278229, "learning_rate": 1.1018281084431432e-05, "loss": 0.1378, "num_input_tokens_seen": 7365208, "step": 27470 }, { "epoch": 7.201834862385321, "grad_norm": 1.4117326736450195, "learning_rate": 1.1008801196559204e-05, "loss": 0.1319, "num_input_tokens_seen": 7366264, "step": 27475 }, { "epoch": 7.203145478374836, "grad_norm": 2.4062962532043457, "learning_rate": 1.0999324237020714e-05, "loss": 0.2734, "num_input_tokens_seen": 7368328, "step": 27480 }, { "epoch": 7.204456094364351, "grad_norm": 3.518488883972168, "learning_rate": 1.0989850207799469e-05, "loss": 0.2063, "num_input_tokens_seen": 7370296, "step": 27485 }, { "epoch": 7.205766710353866, "grad_norm": 7.163586616516113, "learning_rate": 1.0980379110878366e-05, "loss": 0.2537, "num_input_tokens_seen": 7371176, "step": 27490 }, { "epoch": 7.207077326343382, "grad_norm": 2.988058090209961, "learning_rate": 1.0970910948239694e-05, "loss": 0.2141, "num_input_tokens_seen": 7372312, "step": 27495 }, { "epoch": 7.208387942332896, "grad_norm": 1.9175490140914917, "learning_rate": 1.0961445721865116e-05, "loss": 0.1911, "num_input_tokens_seen": 7373544, "step": 27500 }, { "epoch": 7.209698558322412, "grad_norm": 1.3215522766113281, "learning_rate": 1.0951983433735704e-05, "loss": 0.1335, "num_input_tokens_seen": 7374888, "step": 27505 }, { "epoch": 7.2110091743119265, "grad_norm": 2.6241660118103027, "learning_rate": 1.0942524085831868e-05, "loss": 0.2107, "num_input_tokens_seen": 7376504, "step": 27510 }, { "epoch": 7.212319790301442, "grad_norm": 1.7905809879302979, "learning_rate": 1.0933067680133457e-05, "loss": 0.1174, "num_input_tokens_seen": 7377752, "step": 27515 }, { "epoch": 7.213630406290957, "grad_norm": 1.7978090047836304, "learning_rate": 1.092361421861968e-05, "loss": 0.0848, "num_input_tokens_seen": 7378952, "step": 27520 }, { "epoch": 7.214941022280472, "grad_norm": 1.4546412229537964, "learning_rate": 1.091416370326912e-05, "loss": 0.1592, "num_input_tokens_seen": 7380520, "step": 27525 }, { "epoch": 7.216251638269987, "grad_norm": 2.664584159851074, "learning_rate": 1.090471613605976e-05, "loss": 0.2495, "num_input_tokens_seen": 7382008, "step": 27530 }, { "epoch": 7.217562254259502, "grad_norm": 1.7767168283462524, "learning_rate": 1.0895271518968962e-05, "loss": 0.0884, "num_input_tokens_seen": 7383592, "step": 27535 }, { "epoch": 7.218872870249017, "grad_norm": 1.788527250289917, "learning_rate": 1.0885829853973467e-05, "loss": 0.1348, "num_input_tokens_seen": 7384728, "step": 27540 }, { "epoch": 7.220183486238533, "grad_norm": 1.2370933294296265, "learning_rate": 1.087639114304939e-05, "loss": 0.2586, "num_input_tokens_seen": 7385880, "step": 27545 }, { "epoch": 7.221494102228047, "grad_norm": 1.6648838520050049, "learning_rate": 1.0866955388172245e-05, "loss": 0.1991, "num_input_tokens_seen": 7387000, "step": 27550 }, { "epoch": 7.222804718217562, "grad_norm": 1.0647857189178467, "learning_rate": 1.0857522591316915e-05, "loss": 0.0974, "num_input_tokens_seen": 7388520, "step": 27555 }, { "epoch": 7.224115334207077, "grad_norm": 1.2700934410095215, "learning_rate": 1.084809275445767e-05, "loss": 0.1713, "num_input_tokens_seen": 7389832, "step": 27560 }, { "epoch": 7.225425950196592, "grad_norm": 2.46150541305542, "learning_rate": 1.083866587956815e-05, "loss": 0.3029, "num_input_tokens_seen": 7390712, "step": 27565 }, { "epoch": 7.226736566186108, "grad_norm": 4.790217876434326, "learning_rate": 1.0829241968621401e-05, "loss": 0.2651, "num_input_tokens_seen": 7392344, "step": 27570 }, { "epoch": 7.228047182175622, "grad_norm": 1.5074634552001953, "learning_rate": 1.0819821023589832e-05, "loss": 0.182, "num_input_tokens_seen": 7393736, "step": 27575 }, { "epoch": 7.229357798165138, "grad_norm": 1.628108263015747, "learning_rate": 1.0810403046445214e-05, "loss": 0.1678, "num_input_tokens_seen": 7395256, "step": 27580 }, { "epoch": 7.2306684141546524, "grad_norm": 0.544040322303772, "learning_rate": 1.0800988039158719e-05, "loss": 0.0664, "num_input_tokens_seen": 7396712, "step": 27585 }, { "epoch": 7.231979030144168, "grad_norm": 2.2694551944732666, "learning_rate": 1.0791576003700895e-05, "loss": 0.1628, "num_input_tokens_seen": 7397832, "step": 27590 }, { "epoch": 7.233289646133683, "grad_norm": 1.8549150228500366, "learning_rate": 1.0782166942041663e-05, "loss": 0.121, "num_input_tokens_seen": 7398904, "step": 27595 }, { "epoch": 7.234600262123198, "grad_norm": 2.4082891941070557, "learning_rate": 1.0772760856150322e-05, "loss": 0.1989, "num_input_tokens_seen": 7399896, "step": 27600 }, { "epoch": 7.235910878112713, "grad_norm": 2.2112984657287598, "learning_rate": 1.0763357747995551e-05, "loss": 0.1948, "num_input_tokens_seen": 7401128, "step": 27605 }, { "epoch": 7.237221494102228, "grad_norm": 1.9851064682006836, "learning_rate": 1.0753957619545405e-05, "loss": 0.3834, "num_input_tokens_seen": 7402632, "step": 27610 }, { "epoch": 7.238532110091743, "grad_norm": 1.3332091569900513, "learning_rate": 1.0744560472767315e-05, "loss": 0.1612, "num_input_tokens_seen": 7403960, "step": 27615 }, { "epoch": 7.239842726081259, "grad_norm": 1.3950926065444946, "learning_rate": 1.0735166309628083e-05, "loss": 0.105, "num_input_tokens_seen": 7405064, "step": 27620 }, { "epoch": 7.241153342070773, "grad_norm": 1.3448046445846558, "learning_rate": 1.0725775132093899e-05, "loss": 0.2346, "num_input_tokens_seen": 7406504, "step": 27625 }, { "epoch": 7.242463958060289, "grad_norm": 1.5374263525009155, "learning_rate": 1.0716386942130313e-05, "loss": 0.1475, "num_input_tokens_seen": 7407928, "step": 27630 }, { "epoch": 7.243774574049803, "grad_norm": 1.8022342920303345, "learning_rate": 1.0707001741702261e-05, "loss": 0.2104, "num_input_tokens_seen": 7409432, "step": 27635 }, { "epoch": 7.245085190039318, "grad_norm": 2.2327442169189453, "learning_rate": 1.0697619532774047e-05, "loss": 0.2491, "num_input_tokens_seen": 7410600, "step": 27640 }, { "epoch": 7.246395806028834, "grad_norm": 1.4418843984603882, "learning_rate": 1.0688240317309355e-05, "loss": 0.214, "num_input_tokens_seen": 7411736, "step": 27645 }, { "epoch": 7.247706422018348, "grad_norm": 1.7069255113601685, "learning_rate": 1.0678864097271234e-05, "loss": 0.2929, "num_input_tokens_seen": 7413304, "step": 27650 }, { "epoch": 7.249017038007864, "grad_norm": 1.0489062070846558, "learning_rate": 1.0669490874622112e-05, "loss": 0.1177, "num_input_tokens_seen": 7414744, "step": 27655 }, { "epoch": 7.250327653997378, "grad_norm": 2.227433919906616, "learning_rate": 1.0660120651323799e-05, "loss": 0.1731, "num_input_tokens_seen": 7415720, "step": 27660 }, { "epoch": 7.251638269986894, "grad_norm": 0.9364787340164185, "learning_rate": 1.0650753429337446e-05, "loss": 0.2049, "num_input_tokens_seen": 7417160, "step": 27665 }, { "epoch": 7.252948885976409, "grad_norm": 1.8661956787109375, "learning_rate": 1.0641389210623607e-05, "loss": 0.1943, "num_input_tokens_seen": 7418696, "step": 27670 }, { "epoch": 7.254259501965924, "grad_norm": 1.0694975852966309, "learning_rate": 1.0632027997142196e-05, "loss": 0.1544, "num_input_tokens_seen": 7420264, "step": 27675 }, { "epoch": 7.255570117955439, "grad_norm": 1.7257393598556519, "learning_rate": 1.062266979085249e-05, "loss": 0.2005, "num_input_tokens_seen": 7421160, "step": 27680 }, { "epoch": 7.256880733944954, "grad_norm": 0.7332062125205994, "learning_rate": 1.0613314593713164e-05, "loss": 0.115, "num_input_tokens_seen": 7422360, "step": 27685 }, { "epoch": 7.258191349934469, "grad_norm": 7.01608419418335, "learning_rate": 1.0603962407682236e-05, "loss": 0.302, "num_input_tokens_seen": 7423720, "step": 27690 }, { "epoch": 7.259501965923985, "grad_norm": 1.8371080160140991, "learning_rate": 1.059461323471711e-05, "loss": 0.1772, "num_input_tokens_seen": 7425336, "step": 27695 }, { "epoch": 7.260812581913499, "grad_norm": 2.0186429023742676, "learning_rate": 1.0585267076774536e-05, "loss": 0.3062, "num_input_tokens_seen": 7426616, "step": 27700 }, { "epoch": 7.262123197903015, "grad_norm": 2.7032642364501953, "learning_rate": 1.0575923935810656e-05, "loss": 0.1824, "num_input_tokens_seen": 7428424, "step": 27705 }, { "epoch": 7.263433813892529, "grad_norm": 4.683781147003174, "learning_rate": 1.056658381378097e-05, "loss": 0.4344, "num_input_tokens_seen": 7429672, "step": 27710 }, { "epoch": 7.264744429882045, "grad_norm": 3.5058889389038086, "learning_rate": 1.0557246712640356e-05, "loss": 0.1149, "num_input_tokens_seen": 7430616, "step": 27715 }, { "epoch": 7.26605504587156, "grad_norm": 2.5342862606048584, "learning_rate": 1.054791263434305e-05, "loss": 0.1959, "num_input_tokens_seen": 7432104, "step": 27720 }, { "epoch": 7.267365661861074, "grad_norm": 2.238186836242676, "learning_rate": 1.0538581580842655e-05, "loss": 0.187, "num_input_tokens_seen": 7433144, "step": 27725 }, { "epoch": 7.26867627785059, "grad_norm": 0.3219383656978607, "learning_rate": 1.052925355409215e-05, "loss": 0.3116, "num_input_tokens_seen": 7434664, "step": 27730 }, { "epoch": 7.269986893840104, "grad_norm": 0.849750280380249, "learning_rate": 1.0519928556043873e-05, "loss": 0.2362, "num_input_tokens_seen": 7435624, "step": 27735 }, { "epoch": 7.27129750982962, "grad_norm": 1.1273976564407349, "learning_rate": 1.0510606588649526e-05, "loss": 0.2224, "num_input_tokens_seen": 7436840, "step": 27740 }, { "epoch": 7.272608125819135, "grad_norm": 1.9184279441833496, "learning_rate": 1.0501287653860189e-05, "loss": 0.3683, "num_input_tokens_seen": 7438232, "step": 27745 }, { "epoch": 7.27391874180865, "grad_norm": 1.8899272680282593, "learning_rate": 1.0491971753626292e-05, "loss": 0.2662, "num_input_tokens_seen": 7441576, "step": 27750 }, { "epoch": 7.275229357798165, "grad_norm": 2.076481342315674, "learning_rate": 1.048265888989764e-05, "loss": 0.2307, "num_input_tokens_seen": 7443000, "step": 27755 }, { "epoch": 7.27653997378768, "grad_norm": 1.3641387224197388, "learning_rate": 1.0473349064623397e-05, "loss": 0.1181, "num_input_tokens_seen": 7444488, "step": 27760 }, { "epoch": 7.277850589777195, "grad_norm": 1.2328306436538696, "learning_rate": 1.0464042279752096e-05, "loss": 0.2529, "num_input_tokens_seen": 7446088, "step": 27765 }, { "epoch": 7.2791612057667106, "grad_norm": 2.0276992321014404, "learning_rate": 1.0454738537231628e-05, "loss": 0.1888, "num_input_tokens_seen": 7447464, "step": 27770 }, { "epoch": 7.280471821756225, "grad_norm": 1.819591760635376, "learning_rate": 1.0445437839009265e-05, "loss": 0.2129, "num_input_tokens_seen": 7448664, "step": 27775 }, { "epoch": 7.281782437745741, "grad_norm": 2.848031997680664, "learning_rate": 1.0436140187031601e-05, "loss": 0.2708, "num_input_tokens_seen": 7449640, "step": 27780 }, { "epoch": 7.283093053735255, "grad_norm": 2.1113288402557373, "learning_rate": 1.0426845583244632e-05, "loss": 0.2167, "num_input_tokens_seen": 7450984, "step": 27785 }, { "epoch": 7.284403669724771, "grad_norm": 3.1720919609069824, "learning_rate": 1.04175540295937e-05, "loss": 0.1914, "num_input_tokens_seen": 7452200, "step": 27790 }, { "epoch": 7.285714285714286, "grad_norm": 1.7106475830078125, "learning_rate": 1.0408265528023508e-05, "loss": 0.1862, "num_input_tokens_seen": 7453160, "step": 27795 }, { "epoch": 7.287024901703801, "grad_norm": 3.0476367473602295, "learning_rate": 1.0398980080478139e-05, "loss": 0.1916, "num_input_tokens_seen": 7454280, "step": 27800 }, { "epoch": 7.288335517693316, "grad_norm": 3.945732593536377, "learning_rate": 1.0389697688901003e-05, "loss": 0.261, "num_input_tokens_seen": 7456184, "step": 27805 }, { "epoch": 7.289646133682831, "grad_norm": 3.7536046504974365, "learning_rate": 1.038041835523491e-05, "loss": 0.1423, "num_input_tokens_seen": 7457496, "step": 27810 }, { "epoch": 7.290956749672346, "grad_norm": 2.760735034942627, "learning_rate": 1.0371142081421987e-05, "loss": 0.2528, "num_input_tokens_seen": 7458824, "step": 27815 }, { "epoch": 7.292267365661861, "grad_norm": 10.3235502243042, "learning_rate": 1.0361868869403745e-05, "loss": 0.2027, "num_input_tokens_seen": 7459960, "step": 27820 }, { "epoch": 7.293577981651376, "grad_norm": 2.3040404319763184, "learning_rate": 1.035259872112106e-05, "loss": 0.3172, "num_input_tokens_seen": 7461256, "step": 27825 }, { "epoch": 7.294888597640891, "grad_norm": 1.948953628540039, "learning_rate": 1.0343331638514154e-05, "loss": 0.1474, "num_input_tokens_seen": 7462440, "step": 27830 }, { "epoch": 7.296199213630406, "grad_norm": 0.9308282136917114, "learning_rate": 1.0334067623522609e-05, "loss": 0.0839, "num_input_tokens_seen": 7463688, "step": 27835 }, { "epoch": 7.297509829619921, "grad_norm": 3.607354164123535, "learning_rate": 1.032480667808537e-05, "loss": 0.2226, "num_input_tokens_seen": 7465096, "step": 27840 }, { "epoch": 7.2988204456094365, "grad_norm": 1.1608482599258423, "learning_rate": 1.0315548804140736e-05, "loss": 0.1574, "num_input_tokens_seen": 7466680, "step": 27845 }, { "epoch": 7.300131061598951, "grad_norm": 2.0135879516601562, "learning_rate": 1.0306294003626365e-05, "loss": 0.1323, "num_input_tokens_seen": 7467768, "step": 27850 }, { "epoch": 7.301441677588467, "grad_norm": 2.560546398162842, "learning_rate": 1.029704227847927e-05, "loss": 0.1194, "num_input_tokens_seen": 7468984, "step": 27855 }, { "epoch": 7.302752293577981, "grad_norm": 1.896600365638733, "learning_rate": 1.0287793630635822e-05, "loss": 0.1532, "num_input_tokens_seen": 7470584, "step": 27860 }, { "epoch": 7.304062909567497, "grad_norm": 10.417558670043945, "learning_rate": 1.0278548062031748e-05, "loss": 0.2016, "num_input_tokens_seen": 7472056, "step": 27865 }, { "epoch": 7.305373525557012, "grad_norm": 2.9850852489471436, "learning_rate": 1.0269305574602125e-05, "loss": 0.1634, "num_input_tokens_seen": 7473576, "step": 27870 }, { "epoch": 7.306684141546527, "grad_norm": 3.3900671005249023, "learning_rate": 1.0260066170281395e-05, "loss": 0.1373, "num_input_tokens_seen": 7474856, "step": 27875 }, { "epoch": 7.307994757536042, "grad_norm": 1.3034130334854126, "learning_rate": 1.0250829851003346e-05, "loss": 0.1646, "num_input_tokens_seen": 7476392, "step": 27880 }, { "epoch": 7.309305373525557, "grad_norm": 3.0527570247650146, "learning_rate": 1.0241596618701124e-05, "loss": 0.2546, "num_input_tokens_seen": 7477624, "step": 27885 }, { "epoch": 7.310615989515072, "grad_norm": 4.24608039855957, "learning_rate": 1.0232366475307228e-05, "loss": 0.1679, "num_input_tokens_seen": 7478584, "step": 27890 }, { "epoch": 7.3119266055045875, "grad_norm": 0.11830302327871323, "learning_rate": 1.0223139422753524e-05, "loss": 0.1549, "num_input_tokens_seen": 7479848, "step": 27895 }, { "epoch": 7.313237221494102, "grad_norm": 4.535591125488281, "learning_rate": 1.0213915462971197e-05, "loss": 0.207, "num_input_tokens_seen": 7481512, "step": 27900 }, { "epoch": 7.314547837483618, "grad_norm": 0.36843323707580566, "learning_rate": 1.0204694597890812e-05, "loss": 0.2167, "num_input_tokens_seen": 7482728, "step": 27905 }, { "epoch": 7.315858453473132, "grad_norm": 2.625762462615967, "learning_rate": 1.0195476829442277e-05, "loss": 0.2128, "num_input_tokens_seen": 7483944, "step": 27910 }, { "epoch": 7.317169069462647, "grad_norm": 1.7747896909713745, "learning_rate": 1.0186262159554872e-05, "loss": 0.3077, "num_input_tokens_seen": 7485608, "step": 27915 }, { "epoch": 7.3184796854521625, "grad_norm": 3.3718678951263428, "learning_rate": 1.0177050590157197e-05, "loss": 0.1616, "num_input_tokens_seen": 7486728, "step": 27920 }, { "epoch": 7.319790301441677, "grad_norm": 2.6970653533935547, "learning_rate": 1.0167842123177223e-05, "loss": 0.1806, "num_input_tokens_seen": 7487912, "step": 27925 }, { "epoch": 7.321100917431193, "grad_norm": 1.3486049175262451, "learning_rate": 1.0158636760542276e-05, "loss": 0.1317, "num_input_tokens_seen": 7489048, "step": 27930 }, { "epoch": 7.322411533420707, "grad_norm": 78.71602630615234, "learning_rate": 1.0149434504178997e-05, "loss": 0.3209, "num_input_tokens_seen": 7490312, "step": 27935 }, { "epoch": 7.323722149410223, "grad_norm": 3.381246566772461, "learning_rate": 1.0140235356013422e-05, "loss": 0.1104, "num_input_tokens_seen": 7491352, "step": 27940 }, { "epoch": 7.325032765399738, "grad_norm": 1.7056745290756226, "learning_rate": 1.0131039317970908e-05, "loss": 0.1231, "num_input_tokens_seen": 7492280, "step": 27945 }, { "epoch": 7.326343381389253, "grad_norm": 2.4176712036132812, "learning_rate": 1.0121846391976173e-05, "loss": 0.3461, "num_input_tokens_seen": 7493800, "step": 27950 }, { "epoch": 7.327653997378768, "grad_norm": 2.0547444820404053, "learning_rate": 1.0112656579953283e-05, "loss": 0.1707, "num_input_tokens_seen": 7495176, "step": 27955 }, { "epoch": 7.328964613368283, "grad_norm": 8.265409469604492, "learning_rate": 1.0103469883825645e-05, "loss": 0.3737, "num_input_tokens_seen": 7496472, "step": 27960 }, { "epoch": 7.330275229357798, "grad_norm": 0.9855308532714844, "learning_rate": 1.0094286305516023e-05, "loss": 0.1475, "num_input_tokens_seen": 7497752, "step": 27965 }, { "epoch": 7.3315858453473135, "grad_norm": 1.1224299669265747, "learning_rate": 1.008510584694652e-05, "loss": 0.1889, "num_input_tokens_seen": 7499000, "step": 27970 }, { "epoch": 7.332896461336828, "grad_norm": 1.8485409021377563, "learning_rate": 1.0075928510038593e-05, "loss": 0.1029, "num_input_tokens_seen": 7499992, "step": 27975 }, { "epoch": 7.334207077326344, "grad_norm": 0.8038206696510315, "learning_rate": 1.0066754296713043e-05, "loss": 0.118, "num_input_tokens_seen": 7501192, "step": 27980 }, { "epoch": 7.335517693315858, "grad_norm": 1.7217367887496948, "learning_rate": 1.0057583208890017e-05, "loss": 0.1469, "num_input_tokens_seen": 7502504, "step": 27985 }, { "epoch": 7.336828309305374, "grad_norm": 1.9840595722198486, "learning_rate": 1.0048415248489004e-05, "loss": 0.1755, "num_input_tokens_seen": 7503944, "step": 27990 }, { "epoch": 7.3381389252948885, "grad_norm": 3.9843857288360596, "learning_rate": 1.0039250417428847e-05, "loss": 0.1524, "num_input_tokens_seen": 7505240, "step": 27995 }, { "epoch": 7.339449541284404, "grad_norm": 1.7382800579071045, "learning_rate": 1.0030088717627729e-05, "loss": 0.1736, "num_input_tokens_seen": 7506344, "step": 28000 }, { "epoch": 7.340760157273919, "grad_norm": 2.3440744876861572, "learning_rate": 1.0020930151003174e-05, "loss": 0.1207, "num_input_tokens_seen": 7507688, "step": 28005 }, { "epoch": 7.342070773263433, "grad_norm": 5.0250749588012695, "learning_rate": 1.0011774719472064e-05, "loss": 0.2548, "num_input_tokens_seen": 7508760, "step": 28010 }, { "epoch": 7.343381389252949, "grad_norm": 1.3591216802597046, "learning_rate": 1.0002622424950604e-05, "loss": 0.1486, "num_input_tokens_seen": 7510136, "step": 28015 }, { "epoch": 7.344692005242464, "grad_norm": 1.1868897676467896, "learning_rate": 9.99347326935435e-06, "loss": 0.1604, "num_input_tokens_seen": 7511864, "step": 28020 }, { "epoch": 7.346002621231979, "grad_norm": 4.367127895355225, "learning_rate": 9.984327254598208e-06, "loss": 0.1362, "num_input_tokens_seen": 7513096, "step": 28025 }, { "epoch": 7.347313237221494, "grad_norm": 1.9809437990188599, "learning_rate": 9.975184382596434e-06, "loss": 0.1472, "num_input_tokens_seen": 7514136, "step": 28030 }, { "epoch": 7.348623853211009, "grad_norm": 0.9011715650558472, "learning_rate": 9.966044655262607e-06, "loss": 0.0965, "num_input_tokens_seen": 7515544, "step": 28035 }, { "epoch": 7.349934469200524, "grad_norm": 1.2488704919815063, "learning_rate": 9.956908074509658e-06, "loss": 0.1594, "num_input_tokens_seen": 7516888, "step": 28040 }, { "epoch": 7.3512450851900395, "grad_norm": 1.6671940088272095, "learning_rate": 9.947774642249863e-06, "loss": 0.135, "num_input_tokens_seen": 7517944, "step": 28045 }, { "epoch": 7.352555701179554, "grad_norm": 1.6078159809112549, "learning_rate": 9.938644360394822e-06, "loss": 0.1677, "num_input_tokens_seen": 7519560, "step": 28050 }, { "epoch": 7.35386631716907, "grad_norm": 1.8897641897201538, "learning_rate": 9.92951723085549e-06, "loss": 0.1378, "num_input_tokens_seen": 7520808, "step": 28055 }, { "epoch": 7.355176933158584, "grad_norm": 6.701071262359619, "learning_rate": 9.920393255542163e-06, "loss": 0.1605, "num_input_tokens_seen": 7521944, "step": 28060 }, { "epoch": 7.3564875491481, "grad_norm": 0.9722893238067627, "learning_rate": 9.911272436364469e-06, "loss": 0.2091, "num_input_tokens_seen": 7523464, "step": 28065 }, { "epoch": 7.3577981651376145, "grad_norm": 3.2754719257354736, "learning_rate": 9.902154775231386e-06, "loss": 0.2021, "num_input_tokens_seen": 7525304, "step": 28070 }, { "epoch": 7.35910878112713, "grad_norm": 1.54898202419281, "learning_rate": 9.893040274051221e-06, "loss": 0.1706, "num_input_tokens_seen": 7526792, "step": 28075 }, { "epoch": 7.360419397116645, "grad_norm": 1.9235692024230957, "learning_rate": 9.883928934731615e-06, "loss": 0.2795, "num_input_tokens_seen": 7527864, "step": 28080 }, { "epoch": 7.36173001310616, "grad_norm": 4.265389919281006, "learning_rate": 9.874820759179581e-06, "loss": 0.196, "num_input_tokens_seen": 7529208, "step": 28085 }, { "epoch": 7.363040629095675, "grad_norm": 1.8284046649932861, "learning_rate": 9.865715749301419e-06, "loss": 0.1929, "num_input_tokens_seen": 7530696, "step": 28090 }, { "epoch": 7.3643512450851905, "grad_norm": 2.109560251235962, "learning_rate": 9.8566139070028e-06, "loss": 0.1292, "num_input_tokens_seen": 7531800, "step": 28095 }, { "epoch": 7.365661861074705, "grad_norm": 1.453358769416809, "learning_rate": 9.847515234188726e-06, "loss": 0.1986, "num_input_tokens_seen": 7532968, "step": 28100 }, { "epoch": 7.36697247706422, "grad_norm": 1.8742204904556274, "learning_rate": 9.838419732763531e-06, "loss": 0.1093, "num_input_tokens_seen": 7534264, "step": 28105 }, { "epoch": 7.368283093053735, "grad_norm": 1.5008366107940674, "learning_rate": 9.829327404630891e-06, "loss": 0.1732, "num_input_tokens_seen": 7535720, "step": 28110 }, { "epoch": 7.36959370904325, "grad_norm": 1.394046664237976, "learning_rate": 9.820238251693811e-06, "loss": 0.2317, "num_input_tokens_seen": 7536792, "step": 28115 }, { "epoch": 7.3709043250327655, "grad_norm": 1.4865810871124268, "learning_rate": 9.811152275854635e-06, "loss": 0.1395, "num_input_tokens_seen": 7537864, "step": 28120 }, { "epoch": 7.37221494102228, "grad_norm": 1.3845146894454956, "learning_rate": 9.802069479015044e-06, "loss": 0.238, "num_input_tokens_seen": 7539272, "step": 28125 }, { "epoch": 7.373525557011796, "grad_norm": 4.966456413269043, "learning_rate": 9.792989863076064e-06, "loss": 0.2933, "num_input_tokens_seen": 7540296, "step": 28130 }, { "epoch": 7.37483617300131, "grad_norm": 1.763920545578003, "learning_rate": 9.783913429938015e-06, "loss": 0.2707, "num_input_tokens_seen": 7541768, "step": 28135 }, { "epoch": 7.376146788990826, "grad_norm": 1.9199695587158203, "learning_rate": 9.774840181500588e-06, "loss": 0.181, "num_input_tokens_seen": 7542952, "step": 28140 }, { "epoch": 7.3774574049803405, "grad_norm": 5.605641841888428, "learning_rate": 9.765770119662813e-06, "loss": 0.1553, "num_input_tokens_seen": 7543912, "step": 28145 }, { "epoch": 7.378768020969856, "grad_norm": 1.695617914199829, "learning_rate": 9.756703246323026e-06, "loss": 0.2064, "num_input_tokens_seen": 7544856, "step": 28150 }, { "epoch": 7.380078636959371, "grad_norm": 1.4334255456924438, "learning_rate": 9.747639563378916e-06, "loss": 0.1802, "num_input_tokens_seen": 7545992, "step": 28155 }, { "epoch": 7.381389252948886, "grad_norm": 1.3057397603988647, "learning_rate": 9.738579072727488e-06, "loss": 0.1512, "num_input_tokens_seen": 7547496, "step": 28160 }, { "epoch": 7.382699868938401, "grad_norm": 2.1998324394226074, "learning_rate": 9.7295217762651e-06, "loss": 0.2058, "num_input_tokens_seen": 7548552, "step": 28165 }, { "epoch": 7.3840104849279165, "grad_norm": 0.8075292110443115, "learning_rate": 9.72046767588741e-06, "loss": 0.2678, "num_input_tokens_seen": 7550280, "step": 28170 }, { "epoch": 7.385321100917431, "grad_norm": 2.437783718109131, "learning_rate": 9.711416773489434e-06, "loss": 0.1623, "num_input_tokens_seen": 7551368, "step": 28175 }, { "epoch": 7.386631716906947, "grad_norm": 14.478818893432617, "learning_rate": 9.702369070965512e-06, "loss": 0.2397, "num_input_tokens_seen": 7552808, "step": 28180 }, { "epoch": 7.387942332896461, "grad_norm": 5.623617649078369, "learning_rate": 9.693324570209312e-06, "loss": 0.1274, "num_input_tokens_seen": 7554072, "step": 28185 }, { "epoch": 7.389252948885977, "grad_norm": 3.4988040924072266, "learning_rate": 9.684283273113829e-06, "loss": 0.3526, "num_input_tokens_seen": 7555336, "step": 28190 }, { "epoch": 7.3905635648754915, "grad_norm": 2.2545993328094482, "learning_rate": 9.675245181571385e-06, "loss": 0.0963, "num_input_tokens_seen": 7556536, "step": 28195 }, { "epoch": 7.391874180865006, "grad_norm": 3.2896738052368164, "learning_rate": 9.666210297473663e-06, "loss": 0.1846, "num_input_tokens_seen": 7557592, "step": 28200 }, { "epoch": 7.393184796854522, "grad_norm": 0.9924479126930237, "learning_rate": 9.657178622711624e-06, "loss": 0.1256, "num_input_tokens_seen": 7558808, "step": 28205 }, { "epoch": 7.394495412844036, "grad_norm": 1.8737738132476807, "learning_rate": 9.648150159175587e-06, "loss": 0.1596, "num_input_tokens_seen": 7559992, "step": 28210 }, { "epoch": 7.395806028833552, "grad_norm": 1.0624275207519531, "learning_rate": 9.639124908755196e-06, "loss": 0.1701, "num_input_tokens_seen": 7561560, "step": 28215 }, { "epoch": 7.3971166448230665, "grad_norm": 1.378556251525879, "learning_rate": 9.630102873339419e-06, "loss": 0.133, "num_input_tokens_seen": 7563000, "step": 28220 }, { "epoch": 7.398427260812582, "grad_norm": 2.282855272293091, "learning_rate": 9.621084054816557e-06, "loss": 0.1772, "num_input_tokens_seen": 7564664, "step": 28225 }, { "epoch": 7.399737876802097, "grad_norm": 1.7512069940567017, "learning_rate": 9.612068455074227e-06, "loss": 0.1611, "num_input_tokens_seen": 7566056, "step": 28230 }, { "epoch": 7.401048492791612, "grad_norm": 1.7702323198318481, "learning_rate": 9.603056075999382e-06, "loss": 0.2816, "num_input_tokens_seen": 7567288, "step": 28235 }, { "epoch": 7.402359108781127, "grad_norm": 4.4307403564453125, "learning_rate": 9.594046919478294e-06, "loss": 0.2917, "num_input_tokens_seen": 7568824, "step": 28240 }, { "epoch": 7.4036697247706424, "grad_norm": 1.101159691810608, "learning_rate": 9.58504098739658e-06, "loss": 0.2539, "num_input_tokens_seen": 7570504, "step": 28245 }, { "epoch": 7.404980340760157, "grad_norm": 1.4986345767974854, "learning_rate": 9.576038281639138e-06, "loss": 0.2213, "num_input_tokens_seen": 7571784, "step": 28250 }, { "epoch": 7.406290956749673, "grad_norm": 4.070072650909424, "learning_rate": 9.567038804090227e-06, "loss": 0.2449, "num_input_tokens_seen": 7573112, "step": 28255 }, { "epoch": 7.407601572739187, "grad_norm": 2.0965218544006348, "learning_rate": 9.55804255663344e-06, "loss": 0.2501, "num_input_tokens_seen": 7574536, "step": 28260 }, { "epoch": 7.408912188728703, "grad_norm": 1.8412283658981323, "learning_rate": 9.549049541151659e-06, "loss": 0.2169, "num_input_tokens_seen": 7575848, "step": 28265 }, { "epoch": 7.4102228047182175, "grad_norm": 19.713834762573242, "learning_rate": 9.540059759527115e-06, "loss": 0.1944, "num_input_tokens_seen": 7577016, "step": 28270 }, { "epoch": 7.411533420707733, "grad_norm": 1.5408648252487183, "learning_rate": 9.53107321364135e-06, "loss": 0.1749, "num_input_tokens_seen": 7578488, "step": 28275 }, { "epoch": 7.412844036697248, "grad_norm": 1.8886696100234985, "learning_rate": 9.522089905375231e-06, "loss": 0.2284, "num_input_tokens_seen": 7579960, "step": 28280 }, { "epoch": 7.414154652686763, "grad_norm": 1.435829520225525, "learning_rate": 9.513109836608963e-06, "loss": 0.1157, "num_input_tokens_seen": 7581432, "step": 28285 }, { "epoch": 7.415465268676278, "grad_norm": 0.4558162987232208, "learning_rate": 9.504133009222035e-06, "loss": 0.1804, "num_input_tokens_seen": 7582552, "step": 28290 }, { "epoch": 7.4167758846657925, "grad_norm": 1.881651520729065, "learning_rate": 9.495159425093298e-06, "loss": 0.1301, "num_input_tokens_seen": 7583944, "step": 28295 }, { "epoch": 7.418086500655308, "grad_norm": 1.3408735990524292, "learning_rate": 9.4861890861009e-06, "loss": 0.213, "num_input_tokens_seen": 7585320, "step": 28300 }, { "epoch": 7.419397116644823, "grad_norm": 1.4939266443252563, "learning_rate": 9.477221994122321e-06, "loss": 0.2343, "num_input_tokens_seen": 7586456, "step": 28305 }, { "epoch": 7.420707732634338, "grad_norm": 0.15036293864250183, "learning_rate": 9.468258151034349e-06, "loss": 0.2093, "num_input_tokens_seen": 7587560, "step": 28310 }, { "epoch": 7.422018348623853, "grad_norm": 1.905910849571228, "learning_rate": 9.459297558713115e-06, "loss": 0.2806, "num_input_tokens_seen": 7588664, "step": 28315 }, { "epoch": 7.423328964613368, "grad_norm": 2.459052085876465, "learning_rate": 9.45034021903406e-06, "loss": 0.3014, "num_input_tokens_seen": 7589848, "step": 28320 }, { "epoch": 7.424639580602883, "grad_norm": 2.335564613342285, "learning_rate": 9.441386133871922e-06, "loss": 0.3239, "num_input_tokens_seen": 7590920, "step": 28325 }, { "epoch": 7.425950196592399, "grad_norm": 1.9642318487167358, "learning_rate": 9.432435305100782e-06, "loss": 0.2263, "num_input_tokens_seen": 7592232, "step": 28330 }, { "epoch": 7.427260812581913, "grad_norm": 1.6650627851486206, "learning_rate": 9.423487734594036e-06, "loss": 0.1737, "num_input_tokens_seen": 7593624, "step": 28335 }, { "epoch": 7.428571428571429, "grad_norm": 1.0627650022506714, "learning_rate": 9.41454342422439e-06, "loss": 0.2504, "num_input_tokens_seen": 7595096, "step": 28340 }, { "epoch": 7.4298820445609435, "grad_norm": 2.2029800415039062, "learning_rate": 9.405602375863876e-06, "loss": 0.1768, "num_input_tokens_seen": 7596664, "step": 28345 }, { "epoch": 7.431192660550459, "grad_norm": 1.794832468032837, "learning_rate": 9.39666459138384e-06, "loss": 0.1535, "num_input_tokens_seen": 7597624, "step": 28350 }, { "epoch": 7.432503276539974, "grad_norm": 5.518884658813477, "learning_rate": 9.387730072654946e-06, "loss": 0.2136, "num_input_tokens_seen": 7598984, "step": 28355 }, { "epoch": 7.433813892529489, "grad_norm": 2.509178876876831, "learning_rate": 9.378798821547171e-06, "loss": 0.3643, "num_input_tokens_seen": 7600216, "step": 28360 }, { "epoch": 7.435124508519004, "grad_norm": 3.0277278423309326, "learning_rate": 9.36987083992982e-06, "loss": 0.19, "num_input_tokens_seen": 7601496, "step": 28365 }, { "epoch": 7.436435124508519, "grad_norm": 2.294452667236328, "learning_rate": 9.36094612967148e-06, "loss": 0.3393, "num_input_tokens_seen": 7603400, "step": 28370 }, { "epoch": 7.437745740498034, "grad_norm": 2.6917128562927246, "learning_rate": 9.352024692640098e-06, "loss": 0.1572, "num_input_tokens_seen": 7604632, "step": 28375 }, { "epoch": 7.43905635648755, "grad_norm": 1.9771366119384766, "learning_rate": 9.343106530702913e-06, "loss": 0.1545, "num_input_tokens_seen": 7605672, "step": 28380 }, { "epoch": 7.440366972477064, "grad_norm": 1.8532308340072632, "learning_rate": 9.334191645726475e-06, "loss": 0.1519, "num_input_tokens_seen": 7606856, "step": 28385 }, { "epoch": 7.441677588466579, "grad_norm": 2.0310122966766357, "learning_rate": 9.325280039576661e-06, "loss": 0.2008, "num_input_tokens_seen": 7608120, "step": 28390 }, { "epoch": 7.442988204456094, "grad_norm": 1.6848902702331543, "learning_rate": 9.31637171411865e-06, "loss": 0.1287, "num_input_tokens_seen": 7609224, "step": 28395 }, { "epoch": 7.444298820445609, "grad_norm": 2.168607711791992, "learning_rate": 9.307466671216946e-06, "loss": 0.2753, "num_input_tokens_seen": 7610520, "step": 28400 }, { "epoch": 7.445609436435125, "grad_norm": 4.580872058868408, "learning_rate": 9.298564912735347e-06, "loss": 0.2044, "num_input_tokens_seen": 7612328, "step": 28405 }, { "epoch": 7.446920052424639, "grad_norm": 1.7408071756362915, "learning_rate": 9.289666440536981e-06, "loss": 0.1924, "num_input_tokens_seen": 7613784, "step": 28410 }, { "epoch": 7.448230668414155, "grad_norm": 2.2302873134613037, "learning_rate": 9.280771256484284e-06, "loss": 0.2891, "num_input_tokens_seen": 7615224, "step": 28415 }, { "epoch": 7.4495412844036695, "grad_norm": 4.373729228973389, "learning_rate": 9.271879362439e-06, "loss": 0.21, "num_input_tokens_seen": 7616600, "step": 28420 }, { "epoch": 7.450851900393185, "grad_norm": 0.5463648438453674, "learning_rate": 9.262990760262183e-06, "loss": 0.0889, "num_input_tokens_seen": 7617720, "step": 28425 }, { "epoch": 7.4521625163827, "grad_norm": 1.904239296913147, "learning_rate": 9.254105451814215e-06, "loss": 0.2138, "num_input_tokens_seen": 7619432, "step": 28430 }, { "epoch": 7.453473132372215, "grad_norm": 2.785254716873169, "learning_rate": 9.245223438954781e-06, "loss": 0.1614, "num_input_tokens_seen": 7620408, "step": 28435 }, { "epoch": 7.45478374836173, "grad_norm": 1.2245383262634277, "learning_rate": 9.236344723542847e-06, "loss": 0.1247, "num_input_tokens_seen": 7621624, "step": 28440 }, { "epoch": 7.456094364351245, "grad_norm": 2.205933094024658, "learning_rate": 9.227469307436728e-06, "loss": 0.1545, "num_input_tokens_seen": 7623080, "step": 28445 }, { "epoch": 7.45740498034076, "grad_norm": 3.1248552799224854, "learning_rate": 9.21859719249403e-06, "loss": 0.1783, "num_input_tokens_seen": 7624120, "step": 28450 }, { "epoch": 7.458715596330276, "grad_norm": 1.3409837484359741, "learning_rate": 9.209728380571669e-06, "loss": 0.1861, "num_input_tokens_seen": 7625496, "step": 28455 }, { "epoch": 7.46002621231979, "grad_norm": 0.7358991503715515, "learning_rate": 9.200862873525876e-06, "loss": 0.2277, "num_input_tokens_seen": 7627656, "step": 28460 }, { "epoch": 7.461336828309306, "grad_norm": 1.5814789533615112, "learning_rate": 9.192000673212184e-06, "loss": 0.1787, "num_input_tokens_seen": 7629000, "step": 28465 }, { "epoch": 7.46264744429882, "grad_norm": 2.4328222274780273, "learning_rate": 9.183141781485439e-06, "loss": 0.2115, "num_input_tokens_seen": 7630200, "step": 28470 }, { "epoch": 7.463958060288336, "grad_norm": 1.9274786710739136, "learning_rate": 9.174286200199786e-06, "loss": 0.2173, "num_input_tokens_seen": 7631208, "step": 28475 }, { "epoch": 7.465268676277851, "grad_norm": 2.968440294265747, "learning_rate": 9.165433931208697e-06, "loss": 0.213, "num_input_tokens_seen": 7632360, "step": 28480 }, { "epoch": 7.466579292267365, "grad_norm": 1.9451289176940918, "learning_rate": 9.15658497636491e-06, "loss": 0.2448, "num_input_tokens_seen": 7633704, "step": 28485 }, { "epoch": 7.467889908256881, "grad_norm": 2.081357717514038, "learning_rate": 9.147739337520519e-06, "loss": 0.1451, "num_input_tokens_seen": 7634776, "step": 28490 }, { "epoch": 7.4692005242463955, "grad_norm": 4.654325485229492, "learning_rate": 9.138897016526895e-06, "loss": 0.3149, "num_input_tokens_seen": 7636456, "step": 28495 }, { "epoch": 7.470511140235911, "grad_norm": 1.8092787265777588, "learning_rate": 9.130058015234722e-06, "loss": 0.242, "num_input_tokens_seen": 7637704, "step": 28500 }, { "epoch": 7.471821756225426, "grad_norm": 1.052457571029663, "learning_rate": 9.12122233549398e-06, "loss": 0.1091, "num_input_tokens_seen": 7639000, "step": 28505 }, { "epoch": 7.473132372214941, "grad_norm": 2.7924840450286865, "learning_rate": 9.11238997915397e-06, "loss": 0.1658, "num_input_tokens_seen": 7640200, "step": 28510 }, { "epoch": 7.474442988204456, "grad_norm": 3.1336400508880615, "learning_rate": 9.103560948063286e-06, "loss": 0.2505, "num_input_tokens_seen": 7641640, "step": 28515 }, { "epoch": 7.475753604193971, "grad_norm": 1.5838547945022583, "learning_rate": 9.094735244069839e-06, "loss": 0.1171, "num_input_tokens_seen": 7642872, "step": 28520 }, { "epoch": 7.477064220183486, "grad_norm": 1.351757287979126, "learning_rate": 9.08591286902081e-06, "loss": 0.2601, "num_input_tokens_seen": 7644456, "step": 28525 }, { "epoch": 7.478374836173002, "grad_norm": 3.032625198364258, "learning_rate": 9.077093824762722e-06, "loss": 0.2074, "num_input_tokens_seen": 7645592, "step": 28530 }, { "epoch": 7.479685452162516, "grad_norm": 3.3576908111572266, "learning_rate": 9.068278113141385e-06, "loss": 0.1077, "num_input_tokens_seen": 7646744, "step": 28535 }, { "epoch": 7.480996068152032, "grad_norm": 1.5651350021362305, "learning_rate": 9.059465736001901e-06, "loss": 0.2834, "num_input_tokens_seen": 7648392, "step": 28540 }, { "epoch": 7.482306684141546, "grad_norm": 2.2848308086395264, "learning_rate": 9.050656695188706e-06, "loss": 0.2211, "num_input_tokens_seen": 7650344, "step": 28545 }, { "epoch": 7.483617300131062, "grad_norm": 2.003840923309326, "learning_rate": 9.041850992545503e-06, "loss": 0.1665, "num_input_tokens_seen": 7651544, "step": 28550 }, { "epoch": 7.484927916120577, "grad_norm": 3.1595118045806885, "learning_rate": 9.033048629915326e-06, "loss": 0.1667, "num_input_tokens_seen": 7653160, "step": 28555 }, { "epoch": 7.486238532110092, "grad_norm": 2.8601856231689453, "learning_rate": 9.02424960914047e-06, "loss": 0.11, "num_input_tokens_seen": 7654152, "step": 28560 }, { "epoch": 7.487549148099607, "grad_norm": 2.254798650741577, "learning_rate": 9.015453932062568e-06, "loss": 0.12, "num_input_tokens_seen": 7655560, "step": 28565 }, { "epoch": 7.488859764089122, "grad_norm": 3.405729055404663, "learning_rate": 9.006661600522539e-06, "loss": 0.3265, "num_input_tokens_seen": 7656744, "step": 28570 }, { "epoch": 7.490170380078637, "grad_norm": 3.3442063331604004, "learning_rate": 8.997872616360603e-06, "loss": 0.296, "num_input_tokens_seen": 7658216, "step": 28575 }, { "epoch": 7.491480996068152, "grad_norm": 1.5085724592208862, "learning_rate": 8.989086981416278e-06, "loss": 0.217, "num_input_tokens_seen": 7660504, "step": 28580 }, { "epoch": 7.492791612057667, "grad_norm": 3.413182020187378, "learning_rate": 8.980304697528385e-06, "loss": 0.2475, "num_input_tokens_seen": 7661720, "step": 28585 }, { "epoch": 7.494102228047182, "grad_norm": 1.9973623752593994, "learning_rate": 8.971525766535038e-06, "loss": 0.2577, "num_input_tokens_seen": 7663448, "step": 28590 }, { "epoch": 7.495412844036697, "grad_norm": 2.568145513534546, "learning_rate": 8.962750190273653e-06, "loss": 0.2848, "num_input_tokens_seen": 7665048, "step": 28595 }, { "epoch": 7.496723460026212, "grad_norm": 1.9960359334945679, "learning_rate": 8.953977970580942e-06, "loss": 0.1572, "num_input_tokens_seen": 7666600, "step": 28600 }, { "epoch": 7.498034076015728, "grad_norm": 1.0535340309143066, "learning_rate": 8.945209109292919e-06, "loss": 0.2154, "num_input_tokens_seen": 7667960, "step": 28605 }, { "epoch": 7.499344692005242, "grad_norm": 0.8445720076560974, "learning_rate": 8.936443608244885e-06, "loss": 0.1806, "num_input_tokens_seen": 7669336, "step": 28610 }, { "epoch": 7.500655307994758, "grad_norm": 2.9602291584014893, "learning_rate": 8.927681469271451e-06, "loss": 0.2156, "num_input_tokens_seen": 7670600, "step": 28615 }, { "epoch": 7.501965923984272, "grad_norm": 1.6550451517105103, "learning_rate": 8.918922694206513e-06, "loss": 0.2067, "num_input_tokens_seen": 7672088, "step": 28620 }, { "epoch": 7.501965923984272, "eval_loss": 0.7494282126426697, "eval_runtime": 16.6763, "eval_samples_per_second": 50.851, "eval_steps_per_second": 25.425, "num_input_tokens_seen": 7672088, "step": 28620 }, { "epoch": 7.503276539973788, "grad_norm": 1.59080970287323, "learning_rate": 8.91016728488327e-06, "loss": 0.1926, "num_input_tokens_seen": 7673528, "step": 28625 }, { "epoch": 7.504587155963303, "grad_norm": 8.985434532165527, "learning_rate": 8.901415243134211e-06, "loss": 0.1384, "num_input_tokens_seen": 7674808, "step": 28630 }, { "epoch": 7.505897771952818, "grad_norm": 1.549377679824829, "learning_rate": 8.892666570791136e-06, "loss": 0.126, "num_input_tokens_seen": 7676008, "step": 28635 }, { "epoch": 7.507208387942333, "grad_norm": 4.406373500823975, "learning_rate": 8.883921269685108e-06, "loss": 0.3204, "num_input_tokens_seen": 7677736, "step": 28640 }, { "epoch": 7.508519003931848, "grad_norm": 1.8129905462265015, "learning_rate": 8.875179341646511e-06, "loss": 0.2673, "num_input_tokens_seen": 7679192, "step": 28645 }, { "epoch": 7.509829619921363, "grad_norm": 2.0973448753356934, "learning_rate": 8.866440788505018e-06, "loss": 0.1244, "num_input_tokens_seen": 7680600, "step": 28650 }, { "epoch": 7.511140235910878, "grad_norm": 1.8935050964355469, "learning_rate": 8.857705612089581e-06, "loss": 0.2358, "num_input_tokens_seen": 7681992, "step": 28655 }, { "epoch": 7.512450851900393, "grad_norm": 1.9875574111938477, "learning_rate": 8.848973814228478e-06, "loss": 0.1693, "num_input_tokens_seen": 7683112, "step": 28660 }, { "epoch": 7.513761467889909, "grad_norm": 2.016632080078125, "learning_rate": 8.840245396749247e-06, "loss": 0.1805, "num_input_tokens_seen": 7684440, "step": 28665 }, { "epoch": 7.515072083879423, "grad_norm": 1.5294437408447266, "learning_rate": 8.831520361478744e-06, "loss": 0.1931, "num_input_tokens_seen": 7686264, "step": 28670 }, { "epoch": 7.516382699868938, "grad_norm": 1.013635516166687, "learning_rate": 8.82279871024308e-06, "loss": 0.1261, "num_input_tokens_seen": 7687688, "step": 28675 }, { "epoch": 7.517693315858454, "grad_norm": 3.9798641204833984, "learning_rate": 8.814080444867695e-06, "loss": 0.367, "num_input_tokens_seen": 7690104, "step": 28680 }, { "epoch": 7.519003931847968, "grad_norm": 1.0838299989700317, "learning_rate": 8.805365567177307e-06, "loss": 0.1417, "num_input_tokens_seen": 7691544, "step": 28685 }, { "epoch": 7.520314547837484, "grad_norm": 1.3281464576721191, "learning_rate": 8.796654078995921e-06, "loss": 0.102, "num_input_tokens_seen": 7692712, "step": 28690 }, { "epoch": 7.521625163826998, "grad_norm": 1.404063105583191, "learning_rate": 8.787945982146841e-06, "loss": 0.149, "num_input_tokens_seen": 7693736, "step": 28695 }, { "epoch": 7.522935779816514, "grad_norm": 1.0460654497146606, "learning_rate": 8.779241278452657e-06, "loss": 0.1286, "num_input_tokens_seen": 7695032, "step": 28700 }, { "epoch": 7.524246395806029, "grad_norm": 3.510505199432373, "learning_rate": 8.770539969735241e-06, "loss": 0.1614, "num_input_tokens_seen": 7696216, "step": 28705 }, { "epoch": 7.525557011795544, "grad_norm": 2.4026217460632324, "learning_rate": 8.761842057815772e-06, "loss": 0.2605, "num_input_tokens_seen": 7697720, "step": 28710 }, { "epoch": 7.526867627785059, "grad_norm": 1.4059795141220093, "learning_rate": 8.753147544514701e-06, "loss": 0.1067, "num_input_tokens_seen": 7699144, "step": 28715 }, { "epoch": 7.528178243774574, "grad_norm": 1.4109972715377808, "learning_rate": 8.744456431651777e-06, "loss": 0.1643, "num_input_tokens_seen": 7700440, "step": 28720 }, { "epoch": 7.529488859764089, "grad_norm": 2.4339401721954346, "learning_rate": 8.73576872104603e-06, "loss": 0.1793, "num_input_tokens_seen": 7701624, "step": 28725 }, { "epoch": 7.5307994757536045, "grad_norm": 3.85253643989563, "learning_rate": 8.727084414515793e-06, "loss": 0.1767, "num_input_tokens_seen": 7702600, "step": 28730 }, { "epoch": 7.532110091743119, "grad_norm": 2.234530210494995, "learning_rate": 8.718403513878668e-06, "loss": 0.1122, "num_input_tokens_seen": 7703672, "step": 28735 }, { "epoch": 7.533420707732635, "grad_norm": 2.1552131175994873, "learning_rate": 8.709726020951557e-06, "loss": 0.201, "num_input_tokens_seen": 7704776, "step": 28740 }, { "epoch": 7.534731323722149, "grad_norm": 2.973658323287964, "learning_rate": 8.701051937550638e-06, "loss": 0.2574, "num_input_tokens_seen": 7706328, "step": 28745 }, { "epoch": 7.536041939711664, "grad_norm": 2.4981024265289307, "learning_rate": 8.69238126549139e-06, "loss": 0.1187, "num_input_tokens_seen": 7707416, "step": 28750 }, { "epoch": 7.53735255570118, "grad_norm": 2.0851104259490967, "learning_rate": 8.683714006588576e-06, "loss": 0.2487, "num_input_tokens_seen": 7708696, "step": 28755 }, { "epoch": 7.538663171690695, "grad_norm": 0.8576210141181946, "learning_rate": 8.675050162656215e-06, "loss": 0.1956, "num_input_tokens_seen": 7709992, "step": 28760 }, { "epoch": 7.53997378768021, "grad_norm": 2.4442081451416016, "learning_rate": 8.666389735507648e-06, "loss": 0.1601, "num_input_tokens_seen": 7710888, "step": 28765 }, { "epoch": 7.541284403669724, "grad_norm": 1.6326568126678467, "learning_rate": 8.65773272695548e-06, "loss": 0.1826, "num_input_tokens_seen": 7711960, "step": 28770 }, { "epoch": 7.54259501965924, "grad_norm": 1.3839585781097412, "learning_rate": 8.649079138811625e-06, "loss": 0.2707, "num_input_tokens_seen": 7713160, "step": 28775 }, { "epoch": 7.543905635648755, "grad_norm": 1.3466675281524658, "learning_rate": 8.640428972887252e-06, "loss": 0.2103, "num_input_tokens_seen": 7714792, "step": 28780 }, { "epoch": 7.54521625163827, "grad_norm": 1.271134614944458, "learning_rate": 8.63178223099283e-06, "loss": 0.1828, "num_input_tokens_seen": 7715896, "step": 28785 }, { "epoch": 7.546526867627785, "grad_norm": 2.8732717037200928, "learning_rate": 8.623138914938111e-06, "loss": 0.1873, "num_input_tokens_seen": 7717144, "step": 28790 }, { "epoch": 7.5478374836173, "grad_norm": 1.5817687511444092, "learning_rate": 8.61449902653211e-06, "loss": 0.2026, "num_input_tokens_seen": 7718488, "step": 28795 }, { "epoch": 7.549148099606815, "grad_norm": 3.3642845153808594, "learning_rate": 8.605862567583155e-06, "loss": 0.1501, "num_input_tokens_seen": 7720312, "step": 28800 }, { "epoch": 7.5504587155963305, "grad_norm": 2.4649782180786133, "learning_rate": 8.597229539898835e-06, "loss": 0.1879, "num_input_tokens_seen": 7721432, "step": 28805 }, { "epoch": 7.551769331585845, "grad_norm": 1.7521203756332397, "learning_rate": 8.588599945286033e-06, "loss": 0.1264, "num_input_tokens_seen": 7722968, "step": 28810 }, { "epoch": 7.553079947575361, "grad_norm": 1.3534550666809082, "learning_rate": 8.579973785550904e-06, "loss": 0.2081, "num_input_tokens_seen": 7724280, "step": 28815 }, { "epoch": 7.554390563564875, "grad_norm": 1.8319790363311768, "learning_rate": 8.571351062498888e-06, "loss": 0.1317, "num_input_tokens_seen": 7725256, "step": 28820 }, { "epoch": 7.555701179554391, "grad_norm": 3.3233842849731445, "learning_rate": 8.562731777934713e-06, "loss": 0.206, "num_input_tokens_seen": 7726168, "step": 28825 }, { "epoch": 7.557011795543906, "grad_norm": 2.918853282928467, "learning_rate": 8.55411593366237e-06, "loss": 0.1555, "num_input_tokens_seen": 7727416, "step": 28830 }, { "epoch": 7.558322411533421, "grad_norm": 0.3891765773296356, "learning_rate": 8.545503531485149e-06, "loss": 0.0852, "num_input_tokens_seen": 7728600, "step": 28835 }, { "epoch": 7.559633027522936, "grad_norm": 3.803697347640991, "learning_rate": 8.536894573205608e-06, "loss": 0.3599, "num_input_tokens_seen": 7730008, "step": 28840 }, { "epoch": 7.56094364351245, "grad_norm": 2.69832444190979, "learning_rate": 8.528289060625586e-06, "loss": 0.237, "num_input_tokens_seen": 7731112, "step": 28845 }, { "epoch": 7.562254259501966, "grad_norm": 2.5260846614837646, "learning_rate": 8.519686995546203e-06, "loss": 0.2121, "num_input_tokens_seen": 7732376, "step": 28850 }, { "epoch": 7.5635648754914815, "grad_norm": 2.2859537601470947, "learning_rate": 8.511088379767857e-06, "loss": 0.1086, "num_input_tokens_seen": 7733640, "step": 28855 }, { "epoch": 7.564875491480996, "grad_norm": 1.232692003250122, "learning_rate": 8.502493215090218e-06, "loss": 0.2722, "num_input_tokens_seen": 7735064, "step": 28860 }, { "epoch": 7.566186107470511, "grad_norm": 1.9170564413070679, "learning_rate": 8.493901503312249e-06, "loss": 0.2319, "num_input_tokens_seen": 7736328, "step": 28865 }, { "epoch": 7.567496723460026, "grad_norm": 1.1926628351211548, "learning_rate": 8.48531324623218e-06, "loss": 0.1181, "num_input_tokens_seen": 7737672, "step": 28870 }, { "epoch": 7.568807339449541, "grad_norm": 1.4438986778259277, "learning_rate": 8.476728445647506e-06, "loss": 0.1623, "num_input_tokens_seen": 7739272, "step": 28875 }, { "epoch": 7.5701179554390565, "grad_norm": 4.3004913330078125, "learning_rate": 8.468147103355018e-06, "loss": 0.2433, "num_input_tokens_seen": 7740376, "step": 28880 }, { "epoch": 7.571428571428571, "grad_norm": 6.9417243003845215, "learning_rate": 8.45956922115077e-06, "loss": 0.2302, "num_input_tokens_seen": 7741816, "step": 28885 }, { "epoch": 7.572739187418087, "grad_norm": 1.5844171047210693, "learning_rate": 8.450994800830112e-06, "loss": 0.2304, "num_input_tokens_seen": 7743016, "step": 28890 }, { "epoch": 7.574049803407601, "grad_norm": 1.2840310335159302, "learning_rate": 8.442423844187647e-06, "loss": 0.1378, "num_input_tokens_seen": 7743928, "step": 28895 }, { "epoch": 7.575360419397117, "grad_norm": 1.586673617362976, "learning_rate": 8.433856353017261e-06, "loss": 0.1187, "num_input_tokens_seen": 7745320, "step": 28900 }, { "epoch": 7.576671035386632, "grad_norm": 1.408018946647644, "learning_rate": 8.425292329112124e-06, "loss": 0.1097, "num_input_tokens_seen": 7746584, "step": 28905 }, { "epoch": 7.577981651376147, "grad_norm": 3.2036426067352295, "learning_rate": 8.416731774264658e-06, "loss": 0.2223, "num_input_tokens_seen": 7748168, "step": 28910 }, { "epoch": 7.579292267365662, "grad_norm": 1.081339716911316, "learning_rate": 8.408174690266577e-06, "loss": 0.1827, "num_input_tokens_seen": 7749400, "step": 28915 }, { "epoch": 7.580602883355177, "grad_norm": 1.5191494226455688, "learning_rate": 8.399621078908865e-06, "loss": 0.1142, "num_input_tokens_seen": 7750488, "step": 28920 }, { "epoch": 7.581913499344692, "grad_norm": 3.897160291671753, "learning_rate": 8.391070941981778e-06, "loss": 0.2733, "num_input_tokens_seen": 7751512, "step": 28925 }, { "epoch": 7.5832241153342075, "grad_norm": 2.105682849884033, "learning_rate": 8.382524281274844e-06, "loss": 0.2124, "num_input_tokens_seen": 7752504, "step": 28930 }, { "epoch": 7.584534731323722, "grad_norm": 1.326270341873169, "learning_rate": 8.373981098576865e-06, "loss": 0.2066, "num_input_tokens_seen": 7754120, "step": 28935 }, { "epoch": 7.585845347313237, "grad_norm": 2.511960506439209, "learning_rate": 8.365441395675907e-06, "loss": 0.223, "num_input_tokens_seen": 7755640, "step": 28940 }, { "epoch": 7.587155963302752, "grad_norm": 3.892486333847046, "learning_rate": 8.356905174359338e-06, "loss": 0.1811, "num_input_tokens_seen": 7757048, "step": 28945 }, { "epoch": 7.588466579292267, "grad_norm": 2.3623170852661133, "learning_rate": 8.348372436413754e-06, "loss": 0.1175, "num_input_tokens_seen": 7758232, "step": 28950 }, { "epoch": 7.5897771952817825, "grad_norm": 2.500748634338379, "learning_rate": 8.339843183625046e-06, "loss": 0.1558, "num_input_tokens_seen": 7759576, "step": 28955 }, { "epoch": 7.591087811271297, "grad_norm": 0.9827706813812256, "learning_rate": 8.33131741777837e-06, "loss": 0.1674, "num_input_tokens_seen": 7761032, "step": 28960 }, { "epoch": 7.592398427260813, "grad_norm": 3.2247097492218018, "learning_rate": 8.32279514065816e-06, "loss": 0.167, "num_input_tokens_seen": 7762392, "step": 28965 }, { "epoch": 7.593709043250327, "grad_norm": 4.066261291503906, "learning_rate": 8.314276354048109e-06, "loss": 0.268, "num_input_tokens_seen": 7763992, "step": 28970 }, { "epoch": 7.595019659239843, "grad_norm": 1.3196282386779785, "learning_rate": 8.30576105973119e-06, "loss": 0.2555, "num_input_tokens_seen": 7764936, "step": 28975 }, { "epoch": 7.5963302752293576, "grad_norm": 2.209340810775757, "learning_rate": 8.297249259489634e-06, "loss": 0.2442, "num_input_tokens_seen": 7766728, "step": 28980 }, { "epoch": 7.597640891218873, "grad_norm": 1.798044204711914, "learning_rate": 8.288740955104946e-06, "loss": 0.1334, "num_input_tokens_seen": 7768088, "step": 28985 }, { "epoch": 7.598951507208388, "grad_norm": 1.0667638778686523, "learning_rate": 8.280236148357911e-06, "loss": 0.1638, "num_input_tokens_seen": 7769416, "step": 28990 }, { "epoch": 7.600262123197903, "grad_norm": 0.7721909880638123, "learning_rate": 8.271734841028553e-06, "loss": 0.1434, "num_input_tokens_seen": 7770552, "step": 28995 }, { "epoch": 7.601572739187418, "grad_norm": 1.9729691743850708, "learning_rate": 8.263237034896179e-06, "loss": 0.1827, "num_input_tokens_seen": 7772232, "step": 29000 }, { "epoch": 7.6028833551769335, "grad_norm": 1.71646249294281, "learning_rate": 8.254742731739384e-06, "loss": 0.2087, "num_input_tokens_seen": 7773432, "step": 29005 }, { "epoch": 7.604193971166448, "grad_norm": 4.3569016456604, "learning_rate": 8.246251933336e-06, "loss": 0.1804, "num_input_tokens_seen": 7774472, "step": 29010 }, { "epoch": 7.605504587155964, "grad_norm": 0.6382154226303101, "learning_rate": 8.237764641463139e-06, "loss": 0.093, "num_input_tokens_seen": 7775448, "step": 29015 }, { "epoch": 7.606815203145478, "grad_norm": 1.0668261051177979, "learning_rate": 8.229280857897176e-06, "loss": 0.2216, "num_input_tokens_seen": 7776520, "step": 29020 }, { "epoch": 7.608125819134994, "grad_norm": 2.2683770656585693, "learning_rate": 8.22080058441376e-06, "loss": 0.2794, "num_input_tokens_seen": 7778200, "step": 29025 }, { "epoch": 7.6094364351245085, "grad_norm": 3.2737202644348145, "learning_rate": 8.212323822787782e-06, "loss": 0.1789, "num_input_tokens_seen": 7779432, "step": 29030 }, { "epoch": 7.610747051114023, "grad_norm": 1.9173610210418701, "learning_rate": 8.20385057479342e-06, "loss": 0.1944, "num_input_tokens_seen": 7781032, "step": 29035 }, { "epoch": 7.612057667103539, "grad_norm": 1.9204694032669067, "learning_rate": 8.195380842204112e-06, "loss": 0.2357, "num_input_tokens_seen": 7782936, "step": 29040 }, { "epoch": 7.613368283093053, "grad_norm": 1.6688029766082764, "learning_rate": 8.186914626792558e-06, "loss": 0.1597, "num_input_tokens_seen": 7784280, "step": 29045 }, { "epoch": 7.614678899082569, "grad_norm": 3.8828799724578857, "learning_rate": 8.178451930330719e-06, "loss": 0.2192, "num_input_tokens_seen": 7785688, "step": 29050 }, { "epoch": 7.6159895150720835, "grad_norm": 3.0062191486358643, "learning_rate": 8.169992754589819e-06, "loss": 0.3193, "num_input_tokens_seen": 7786952, "step": 29055 }, { "epoch": 7.617300131061599, "grad_norm": 7.582383155822754, "learning_rate": 8.161537101340375e-06, "loss": 0.3109, "num_input_tokens_seen": 7787864, "step": 29060 }, { "epoch": 7.618610747051114, "grad_norm": 1.6805453300476074, "learning_rate": 8.15308497235211e-06, "loss": 0.2087, "num_input_tokens_seen": 7789320, "step": 29065 }, { "epoch": 7.619921363040629, "grad_norm": 6.444908618927002, "learning_rate": 8.144636369394048e-06, "loss": 0.2251, "num_input_tokens_seen": 7790680, "step": 29070 }, { "epoch": 7.621231979030144, "grad_norm": 3.3632211685180664, "learning_rate": 8.136191294234474e-06, "loss": 0.1508, "num_input_tokens_seen": 7791976, "step": 29075 }, { "epoch": 7.6225425950196595, "grad_norm": 1.925675630569458, "learning_rate": 8.127749748640923e-06, "loss": 0.176, "num_input_tokens_seen": 7793256, "step": 29080 }, { "epoch": 7.623853211009174, "grad_norm": 1.566051959991455, "learning_rate": 8.119311734380197e-06, "loss": 0.1257, "num_input_tokens_seen": 7794408, "step": 29085 }, { "epoch": 7.62516382699869, "grad_norm": 2.787306070327759, "learning_rate": 8.110877253218358e-06, "loss": 0.201, "num_input_tokens_seen": 7795608, "step": 29090 }, { "epoch": 7.626474442988204, "grad_norm": 6.20051908493042, "learning_rate": 8.102446306920725e-06, "loss": 0.2787, "num_input_tokens_seen": 7796968, "step": 29095 }, { "epoch": 7.62778505897772, "grad_norm": 2.1953516006469727, "learning_rate": 8.094018897251887e-06, "loss": 0.1811, "num_input_tokens_seen": 7798088, "step": 29100 }, { "epoch": 7.6290956749672345, "grad_norm": 1.9840667247772217, "learning_rate": 8.085595025975687e-06, "loss": 0.2001, "num_input_tokens_seen": 7799272, "step": 29105 }, { "epoch": 7.63040629095675, "grad_norm": 2.1938209533691406, "learning_rate": 8.077174694855216e-06, "loss": 0.2171, "num_input_tokens_seen": 7800648, "step": 29110 }, { "epoch": 7.631716906946265, "grad_norm": 2.921952247619629, "learning_rate": 8.068757905652832e-06, "loss": 0.1839, "num_input_tokens_seen": 7802200, "step": 29115 }, { "epoch": 7.63302752293578, "grad_norm": 1.9603427648544312, "learning_rate": 8.060344660130172e-06, "loss": 0.1985, "num_input_tokens_seen": 7803304, "step": 29120 }, { "epoch": 7.634338138925295, "grad_norm": 1.8044602870941162, "learning_rate": 8.051934960048104e-06, "loss": 0.1603, "num_input_tokens_seen": 7804136, "step": 29125 }, { "epoch": 7.6356487549148095, "grad_norm": 2.9604384899139404, "learning_rate": 8.043528807166767e-06, "loss": 0.2999, "num_input_tokens_seen": 7805608, "step": 29130 }, { "epoch": 7.636959370904325, "grad_norm": 1.4245058298110962, "learning_rate": 8.035126203245552e-06, "loss": 0.1376, "num_input_tokens_seen": 7806808, "step": 29135 }, { "epoch": 7.63826998689384, "grad_norm": 3.4748053550720215, "learning_rate": 8.026727150043118e-06, "loss": 0.2425, "num_input_tokens_seen": 7808312, "step": 29140 }, { "epoch": 7.639580602883355, "grad_norm": 2.6761677265167236, "learning_rate": 8.018331649317353e-06, "loss": 0.124, "num_input_tokens_seen": 7809496, "step": 29145 }, { "epoch": 7.64089121887287, "grad_norm": 2.3228468894958496, "learning_rate": 8.009939702825433e-06, "loss": 0.2296, "num_input_tokens_seen": 7810792, "step": 29150 }, { "epoch": 7.6422018348623855, "grad_norm": 1.2219737768173218, "learning_rate": 8.001551312323776e-06, "loss": 0.1611, "num_input_tokens_seen": 7811832, "step": 29155 }, { "epoch": 7.6435124508519, "grad_norm": 9.670119285583496, "learning_rate": 7.993166479568063e-06, "loss": 0.1721, "num_input_tokens_seen": 7813032, "step": 29160 }, { "epoch": 7.644823066841416, "grad_norm": 2.545875310897827, "learning_rate": 7.984785206313213e-06, "loss": 0.2351, "num_input_tokens_seen": 7814472, "step": 29165 }, { "epoch": 7.64613368283093, "grad_norm": 3.1729085445404053, "learning_rate": 7.976407494313418e-06, "loss": 0.1788, "num_input_tokens_seen": 7816312, "step": 29170 }, { "epoch": 7.647444298820446, "grad_norm": 1.9806894063949585, "learning_rate": 7.968033345322123e-06, "loss": 0.1435, "num_input_tokens_seen": 7817736, "step": 29175 }, { "epoch": 7.6487549148099605, "grad_norm": 1.618987798690796, "learning_rate": 7.95966276109203e-06, "loss": 0.1515, "num_input_tokens_seen": 7818984, "step": 29180 }, { "epoch": 7.650065530799476, "grad_norm": 1.1352745294570923, "learning_rate": 7.95129574337507e-06, "loss": 0.2581, "num_input_tokens_seen": 7820280, "step": 29185 }, { "epoch": 7.651376146788991, "grad_norm": 1.7937326431274414, "learning_rate": 7.942932293922453e-06, "loss": 0.1879, "num_input_tokens_seen": 7821544, "step": 29190 }, { "epoch": 7.652686762778506, "grad_norm": 1.1247214078903198, "learning_rate": 7.934572414484634e-06, "loss": 0.1406, "num_input_tokens_seen": 7822776, "step": 29195 }, { "epoch": 7.653997378768021, "grad_norm": 2.1527011394500732, "learning_rate": 7.926216106811321e-06, "loss": 0.1356, "num_input_tokens_seen": 7824008, "step": 29200 }, { "epoch": 7.655307994757536, "grad_norm": 1.3674523830413818, "learning_rate": 7.917863372651477e-06, "loss": 0.2011, "num_input_tokens_seen": 7825528, "step": 29205 }, { "epoch": 7.656618610747051, "grad_norm": 1.9090681076049805, "learning_rate": 7.909514213753314e-06, "loss": 0.1193, "num_input_tokens_seen": 7826952, "step": 29210 }, { "epoch": 7.657929226736567, "grad_norm": 1.2273377180099487, "learning_rate": 7.901168631864294e-06, "loss": 0.1735, "num_input_tokens_seen": 7828152, "step": 29215 }, { "epoch": 7.659239842726081, "grad_norm": 1.6594409942626953, "learning_rate": 7.892826628731135e-06, "loss": 0.3031, "num_input_tokens_seen": 7829864, "step": 29220 }, { "epoch": 7.660550458715596, "grad_norm": 3.7301974296569824, "learning_rate": 7.884488206099814e-06, "loss": 0.1624, "num_input_tokens_seen": 7831096, "step": 29225 }, { "epoch": 7.6618610747051115, "grad_norm": 2.4569056034088135, "learning_rate": 7.876153365715519e-06, "loss": 0.2319, "num_input_tokens_seen": 7832392, "step": 29230 }, { "epoch": 7.663171690694626, "grad_norm": 1.8636467456817627, "learning_rate": 7.867822109322747e-06, "loss": 0.131, "num_input_tokens_seen": 7833528, "step": 29235 }, { "epoch": 7.664482306684142, "grad_norm": 1.6061841249465942, "learning_rate": 7.859494438665204e-06, "loss": 0.2013, "num_input_tokens_seen": 7835064, "step": 29240 }, { "epoch": 7.665792922673656, "grad_norm": 2.7394275665283203, "learning_rate": 7.85117035548586e-06, "loss": 0.1194, "num_input_tokens_seen": 7836440, "step": 29245 }, { "epoch": 7.667103538663172, "grad_norm": 2.0500848293304443, "learning_rate": 7.842849861526929e-06, "loss": 0.251, "num_input_tokens_seen": 7837736, "step": 29250 }, { "epoch": 7.6684141546526865, "grad_norm": 2.9958276748657227, "learning_rate": 7.834532958529878e-06, "loss": 0.1375, "num_input_tokens_seen": 7838920, "step": 29255 }, { "epoch": 7.669724770642202, "grad_norm": 1.4921129941940308, "learning_rate": 7.826219648235424e-06, "loss": 0.1501, "num_input_tokens_seen": 7840296, "step": 29260 }, { "epoch": 7.671035386631717, "grad_norm": 3.820136070251465, "learning_rate": 7.817909932383519e-06, "loss": 0.3308, "num_input_tokens_seen": 7841928, "step": 29265 }, { "epoch": 7.672346002621232, "grad_norm": 4.031387805938721, "learning_rate": 7.809603812713373e-06, "loss": 0.1656, "num_input_tokens_seen": 7843000, "step": 29270 }, { "epoch": 7.673656618610747, "grad_norm": 1.515427589416504, "learning_rate": 7.80130129096345e-06, "loss": 0.2638, "num_input_tokens_seen": 7844616, "step": 29275 }, { "epoch": 7.674967234600262, "grad_norm": 2.0983238220214844, "learning_rate": 7.793002368871446e-06, "loss": 0.1756, "num_input_tokens_seen": 7845752, "step": 29280 }, { "epoch": 7.676277850589777, "grad_norm": 2.1764044761657715, "learning_rate": 7.78470704817431e-06, "loss": 0.158, "num_input_tokens_seen": 7847112, "step": 29285 }, { "epoch": 7.677588466579293, "grad_norm": 3.7191600799560547, "learning_rate": 7.77641533060825e-06, "loss": 0.1241, "num_input_tokens_seen": 7848264, "step": 29290 }, { "epoch": 7.678899082568807, "grad_norm": 2.655216932296753, "learning_rate": 7.76812721790871e-06, "loss": 0.2458, "num_input_tokens_seen": 7849672, "step": 29295 }, { "epoch": 7.680209698558322, "grad_norm": 2.5639235973358154, "learning_rate": 7.759842711810361e-06, "loss": 0.236, "num_input_tokens_seen": 7850952, "step": 29300 }, { "epoch": 7.6815203145478375, "grad_norm": 3.002049446105957, "learning_rate": 7.751561814047142e-06, "loss": 0.2989, "num_input_tokens_seen": 7852216, "step": 29305 }, { "epoch": 7.682830930537353, "grad_norm": 6.726833820343018, "learning_rate": 7.743284526352235e-06, "loss": 0.1406, "num_input_tokens_seen": 7853368, "step": 29310 }, { "epoch": 7.684141546526868, "grad_norm": 3.2967474460601807, "learning_rate": 7.735010850458057e-06, "loss": 0.2383, "num_input_tokens_seen": 7855080, "step": 29315 }, { "epoch": 7.685452162516382, "grad_norm": 2.9733755588531494, "learning_rate": 7.726740788096276e-06, "loss": 0.2531, "num_input_tokens_seen": 7856536, "step": 29320 }, { "epoch": 7.686762778505898, "grad_norm": 3.0954322814941406, "learning_rate": 7.718474340997804e-06, "loss": 0.291, "num_input_tokens_seen": 7858104, "step": 29325 }, { "epoch": 7.6880733944954125, "grad_norm": 2.795128345489502, "learning_rate": 7.710211510892792e-06, "loss": 0.3156, "num_input_tokens_seen": 7859752, "step": 29330 }, { "epoch": 7.689384010484928, "grad_norm": 2.0531797409057617, "learning_rate": 7.701952299510634e-06, "loss": 0.1816, "num_input_tokens_seen": 7862760, "step": 29335 }, { "epoch": 7.690694626474443, "grad_norm": 1.9203252792358398, "learning_rate": 7.69369670857998e-06, "loss": 0.1427, "num_input_tokens_seen": 7864632, "step": 29340 }, { "epoch": 7.692005242463958, "grad_norm": 2.175119161605835, "learning_rate": 7.685444739828688e-06, "loss": 0.1424, "num_input_tokens_seen": 7866152, "step": 29345 }, { "epoch": 7.693315858453473, "grad_norm": 3.931161880493164, "learning_rate": 7.677196394983902e-06, "loss": 0.1478, "num_input_tokens_seen": 7867336, "step": 29350 }, { "epoch": 7.694626474442988, "grad_norm": 5.036560535430908, "learning_rate": 7.668951675771979e-06, "loss": 0.1875, "num_input_tokens_seen": 7868568, "step": 29355 }, { "epoch": 7.695937090432503, "grad_norm": 1.8412014245986938, "learning_rate": 7.660710583918524e-06, "loss": 0.1579, "num_input_tokens_seen": 7869784, "step": 29360 }, { "epoch": 7.697247706422019, "grad_norm": 2.3758106231689453, "learning_rate": 7.652473121148383e-06, "loss": 0.3253, "num_input_tokens_seen": 7871112, "step": 29365 }, { "epoch": 7.698558322411533, "grad_norm": 2.7661209106445312, "learning_rate": 7.644239289185644e-06, "loss": 0.2472, "num_input_tokens_seen": 7872776, "step": 29370 }, { "epoch": 7.699868938401049, "grad_norm": 2.3216047286987305, "learning_rate": 7.636009089753643e-06, "loss": 0.1388, "num_input_tokens_seen": 7874056, "step": 29375 }, { "epoch": 7.7011795543905635, "grad_norm": 2.8059959411621094, "learning_rate": 7.627782524574925e-06, "loss": 0.2462, "num_input_tokens_seen": 7875224, "step": 29380 }, { "epoch": 7.702490170380079, "grad_norm": 0.9595551490783691, "learning_rate": 7.619559595371312e-06, "loss": 0.1864, "num_input_tokens_seen": 7876536, "step": 29385 }, { "epoch": 7.703800786369594, "grad_norm": 2.1924500465393066, "learning_rate": 7.611340303863843e-06, "loss": 0.1804, "num_input_tokens_seen": 7877688, "step": 29390 }, { "epoch": 7.705111402359108, "grad_norm": 3.7278709411621094, "learning_rate": 7.603124651772803e-06, "loss": 0.1976, "num_input_tokens_seen": 7878968, "step": 29395 }, { "epoch": 7.706422018348624, "grad_norm": 1.3390843868255615, "learning_rate": 7.594912640817705e-06, "loss": 0.08, "num_input_tokens_seen": 7880312, "step": 29400 }, { "epoch": 7.707732634338139, "grad_norm": 0.9267050623893738, "learning_rate": 7.586704272717327e-06, "loss": 0.178, "num_input_tokens_seen": 7881832, "step": 29405 }, { "epoch": 7.709043250327654, "grad_norm": 2.234985113143921, "learning_rate": 7.578499549189655e-06, "loss": 0.1502, "num_input_tokens_seen": 7882984, "step": 29410 }, { "epoch": 7.710353866317169, "grad_norm": 1.1909781694412231, "learning_rate": 7.570298471951934e-06, "loss": 0.1201, "num_input_tokens_seen": 7884440, "step": 29415 }, { "epoch": 7.711664482306684, "grad_norm": 1.282825231552124, "learning_rate": 7.562101042720621e-06, "loss": 0.3729, "num_input_tokens_seen": 7885800, "step": 29420 }, { "epoch": 7.712975098296199, "grad_norm": 3.5022592544555664, "learning_rate": 7.553907263211427e-06, "loss": 0.2392, "num_input_tokens_seen": 7886968, "step": 29425 }, { "epoch": 7.714285714285714, "grad_norm": 1.2224808931350708, "learning_rate": 7.545717135139299e-06, "loss": 0.1699, "num_input_tokens_seen": 7888904, "step": 29430 }, { "epoch": 7.715596330275229, "grad_norm": 2.16192889213562, "learning_rate": 7.537530660218414e-06, "loss": 0.1197, "num_input_tokens_seen": 7889944, "step": 29435 }, { "epoch": 7.716906946264745, "grad_norm": 1.8140374422073364, "learning_rate": 7.529347840162193e-06, "loss": 0.194, "num_input_tokens_seen": 7891048, "step": 29440 }, { "epoch": 7.718217562254259, "grad_norm": 5.705775260925293, "learning_rate": 7.52116867668328e-06, "loss": 0.2047, "num_input_tokens_seen": 7892504, "step": 29445 }, { "epoch": 7.719528178243775, "grad_norm": 0.6446337699890137, "learning_rate": 7.51299317149356e-06, "loss": 0.2273, "num_input_tokens_seen": 7893848, "step": 29450 }, { "epoch": 7.7208387942332894, "grad_norm": 1.800856590270996, "learning_rate": 7.504821326304157e-06, "loss": 0.1774, "num_input_tokens_seen": 7895192, "step": 29455 }, { "epoch": 7.722149410222805, "grad_norm": 3.0176384449005127, "learning_rate": 7.496653142825419e-06, "loss": 0.1504, "num_input_tokens_seen": 7896424, "step": 29460 }, { "epoch": 7.72346002621232, "grad_norm": 4.854153156280518, "learning_rate": 7.4884886227669375e-06, "loss": 0.1351, "num_input_tokens_seen": 7897992, "step": 29465 }, { "epoch": 7.724770642201835, "grad_norm": 4.444602966308594, "learning_rate": 7.480327767837525e-06, "loss": 0.1935, "num_input_tokens_seen": 7899128, "step": 29470 }, { "epoch": 7.72608125819135, "grad_norm": 2.020966053009033, "learning_rate": 7.472170579745244e-06, "loss": 0.287, "num_input_tokens_seen": 7900568, "step": 29475 }, { "epoch": 7.727391874180865, "grad_norm": 2.8598949909210205, "learning_rate": 7.464017060197373e-06, "loss": 0.1954, "num_input_tokens_seen": 7902024, "step": 29480 }, { "epoch": 7.72870249017038, "grad_norm": 0.9913416504859924, "learning_rate": 7.45586721090043e-06, "loss": 0.081, "num_input_tokens_seen": 7903416, "step": 29485 }, { "epoch": 7.730013106159895, "grad_norm": 1.2812479734420776, "learning_rate": 7.447721033560168e-06, "loss": 0.2254, "num_input_tokens_seen": 7904552, "step": 29490 }, { "epoch": 7.73132372214941, "grad_norm": 1.2024487257003784, "learning_rate": 7.439578529881572e-06, "loss": 0.1446, "num_input_tokens_seen": 7905752, "step": 29495 }, { "epoch": 7.732634338138926, "grad_norm": 1.937235951423645, "learning_rate": 7.431439701568841e-06, "loss": 0.1452, "num_input_tokens_seen": 7907240, "step": 29500 }, { "epoch": 7.73394495412844, "grad_norm": 2.481937885284424, "learning_rate": 7.423304550325422e-06, "loss": 0.2102, "num_input_tokens_seen": 7908296, "step": 29505 }, { "epoch": 7.735255570117955, "grad_norm": 2.4278318881988525, "learning_rate": 7.415173077853993e-06, "loss": 0.1129, "num_input_tokens_seen": 7909384, "step": 29510 }, { "epoch": 7.736566186107471, "grad_norm": 0.32164981961250305, "learning_rate": 7.407045285856445e-06, "loss": 0.2298, "num_input_tokens_seen": 7911080, "step": 29515 }, { "epoch": 7.737876802096985, "grad_norm": 1.8104740381240845, "learning_rate": 7.39892117603393e-06, "loss": 0.1974, "num_input_tokens_seen": 7911976, "step": 29520 }, { "epoch": 7.739187418086501, "grad_norm": 1.4889023303985596, "learning_rate": 7.390800750086799e-06, "loss": 0.244, "num_input_tokens_seen": 7913416, "step": 29525 }, { "epoch": 7.740498034076015, "grad_norm": 3.6638402938842773, "learning_rate": 7.382684009714652e-06, "loss": 0.2173, "num_input_tokens_seen": 7914936, "step": 29530 }, { "epoch": 7.741808650065531, "grad_norm": 4.843504905700684, "learning_rate": 7.374570956616295e-06, "loss": 0.3823, "num_input_tokens_seen": 7918200, "step": 29535 }, { "epoch": 7.743119266055046, "grad_norm": 0.031204311177134514, "learning_rate": 7.366461592489782e-06, "loss": 0.2331, "num_input_tokens_seen": 7919208, "step": 29540 }, { "epoch": 7.744429882044561, "grad_norm": 1.0501195192337036, "learning_rate": 7.358355919032389e-06, "loss": 0.1633, "num_input_tokens_seen": 7920648, "step": 29545 }, { "epoch": 7.745740498034076, "grad_norm": 2.8215997219085693, "learning_rate": 7.350253937940621e-06, "loss": 0.1129, "num_input_tokens_seen": 7921544, "step": 29550 }, { "epoch": 7.747051114023591, "grad_norm": 3.842176675796509, "learning_rate": 7.342155650910207e-06, "loss": 0.1307, "num_input_tokens_seen": 7922648, "step": 29555 }, { "epoch": 7.748361730013106, "grad_norm": 3.7963175773620605, "learning_rate": 7.334061059636102e-06, "loss": 0.1686, "num_input_tokens_seen": 7923880, "step": 29560 }, { "epoch": 7.749672346002622, "grad_norm": 1.5301496982574463, "learning_rate": 7.325970165812496e-06, "loss": 0.1527, "num_input_tokens_seen": 7925592, "step": 29565 }, { "epoch": 7.750982961992136, "grad_norm": 1.720096468925476, "learning_rate": 7.3178829711327945e-06, "loss": 0.1648, "num_input_tokens_seen": 7926856, "step": 29570 }, { "epoch": 7.752293577981652, "grad_norm": 3.946504831314087, "learning_rate": 7.309799477289636e-06, "loss": 0.1371, "num_input_tokens_seen": 7928056, "step": 29575 }, { "epoch": 7.753604193971166, "grad_norm": 3.6019465923309326, "learning_rate": 7.301719685974881e-06, "loss": 0.2065, "num_input_tokens_seen": 7929032, "step": 29580 }, { "epoch": 7.754914809960681, "grad_norm": 1.707804560661316, "learning_rate": 7.293643598879612e-06, "loss": 0.1725, "num_input_tokens_seen": 7930056, "step": 29585 }, { "epoch": 7.756225425950197, "grad_norm": 1.3160977363586426, "learning_rate": 7.285571217694148e-06, "loss": 0.223, "num_input_tokens_seen": 7931288, "step": 29590 }, { "epoch": 7.757536041939712, "grad_norm": 19.73737335205078, "learning_rate": 7.277502544108017e-06, "loss": 0.2156, "num_input_tokens_seen": 7932424, "step": 29595 }, { "epoch": 7.758846657929227, "grad_norm": 2.2085700035095215, "learning_rate": 7.2694375798099815e-06, "loss": 0.1734, "num_input_tokens_seen": 7933880, "step": 29600 }, { "epoch": 7.760157273918741, "grad_norm": 3.8119821548461914, "learning_rate": 7.261376326488026e-06, "loss": 0.3843, "num_input_tokens_seen": 7935080, "step": 29605 }, { "epoch": 7.761467889908257, "grad_norm": 2.709763765335083, "learning_rate": 7.253318785829363e-06, "loss": 0.2429, "num_input_tokens_seen": 7936168, "step": 29610 }, { "epoch": 7.762778505897772, "grad_norm": 1.8229174613952637, "learning_rate": 7.245264959520406e-06, "loss": 0.1396, "num_input_tokens_seen": 7937352, "step": 29615 }, { "epoch": 7.764089121887287, "grad_norm": 0.3101944625377655, "learning_rate": 7.237214849246815e-06, "loss": 0.1663, "num_input_tokens_seen": 7938712, "step": 29620 }, { "epoch": 7.765399737876802, "grad_norm": 2.445969820022583, "learning_rate": 7.2291684566934634e-06, "loss": 0.3214, "num_input_tokens_seen": 7940664, "step": 29625 }, { "epoch": 7.766710353866317, "grad_norm": 2.290248394012451, "learning_rate": 7.221125783544441e-06, "loss": 0.1691, "num_input_tokens_seen": 7942088, "step": 29630 }, { "epoch": 7.768020969855832, "grad_norm": 1.0802117586135864, "learning_rate": 7.213086831483082e-06, "loss": 0.1154, "num_input_tokens_seen": 7943352, "step": 29635 }, { "epoch": 7.7693315858453476, "grad_norm": 2.825993299484253, "learning_rate": 7.205051602191914e-06, "loss": 0.1718, "num_input_tokens_seen": 7944728, "step": 29640 }, { "epoch": 7.770642201834862, "grad_norm": 2.5481808185577393, "learning_rate": 7.1970200973527e-06, "loss": 0.2578, "num_input_tokens_seen": 7946072, "step": 29645 }, { "epoch": 7.771952817824378, "grad_norm": 1.4962975978851318, "learning_rate": 7.188992318646423e-06, "loss": 0.2032, "num_input_tokens_seen": 7946920, "step": 29650 }, { "epoch": 7.773263433813892, "grad_norm": 2.858924150466919, "learning_rate": 7.180968267753274e-06, "loss": 0.2766, "num_input_tokens_seen": 7948616, "step": 29655 }, { "epoch": 7.774574049803408, "grad_norm": 1.4160724878311157, "learning_rate": 7.172947946352676e-06, "loss": 0.1781, "num_input_tokens_seen": 7950600, "step": 29660 }, { "epoch": 7.775884665792923, "grad_norm": 2.8741469383239746, "learning_rate": 7.164931356123269e-06, "loss": 0.2267, "num_input_tokens_seen": 7951752, "step": 29665 }, { "epoch": 7.777195281782438, "grad_norm": 1.1550499200820923, "learning_rate": 7.156918498742915e-06, "loss": 0.2354, "num_input_tokens_seen": 7953256, "step": 29670 }, { "epoch": 7.778505897771953, "grad_norm": 2.281770706176758, "learning_rate": 7.148909375888688e-06, "loss": 0.0977, "num_input_tokens_seen": 7954488, "step": 29675 }, { "epoch": 7.779816513761467, "grad_norm": 1.8449552059173584, "learning_rate": 7.140903989236886e-06, "loss": 0.2661, "num_input_tokens_seen": 7955816, "step": 29680 }, { "epoch": 7.781127129750983, "grad_norm": 2.130551338195801, "learning_rate": 7.132902340463018e-06, "loss": 0.2287, "num_input_tokens_seen": 7957304, "step": 29685 }, { "epoch": 7.7824377457404985, "grad_norm": 2.1811423301696777, "learning_rate": 7.124904431241819e-06, "loss": 0.3023, "num_input_tokens_seen": 7958616, "step": 29690 }, { "epoch": 7.783748361730013, "grad_norm": 1.4459139108657837, "learning_rate": 7.116910263247237e-06, "loss": 0.1192, "num_input_tokens_seen": 7959848, "step": 29695 }, { "epoch": 7.785058977719528, "grad_norm": 2.329094171524048, "learning_rate": 7.108919838152436e-06, "loss": 0.179, "num_input_tokens_seen": 7961048, "step": 29700 }, { "epoch": 7.786369593709043, "grad_norm": 2.7055559158325195, "learning_rate": 7.100933157629802e-06, "loss": 0.2742, "num_input_tokens_seen": 7962648, "step": 29705 }, { "epoch": 7.787680209698558, "grad_norm": 8.016684532165527, "learning_rate": 7.09295022335093e-06, "loss": 0.273, "num_input_tokens_seen": 7964008, "step": 29710 }, { "epoch": 7.7889908256880735, "grad_norm": 1.7855243682861328, "learning_rate": 7.0849710369866345e-06, "loss": 0.2312, "num_input_tokens_seen": 7965176, "step": 29715 }, { "epoch": 7.790301441677588, "grad_norm": 3.796322822570801, "learning_rate": 7.076995600206948e-06, "loss": 0.1937, "num_input_tokens_seen": 7966568, "step": 29720 }, { "epoch": 7.791612057667104, "grad_norm": 1.5199096202850342, "learning_rate": 7.069023914681111e-06, "loss": 0.2119, "num_input_tokens_seen": 7967880, "step": 29725 }, { "epoch": 7.792922673656618, "grad_norm": 3.5923783779144287, "learning_rate": 7.0610559820775965e-06, "loss": 0.2679, "num_input_tokens_seen": 7969448, "step": 29730 }, { "epoch": 7.794233289646134, "grad_norm": 3.911315441131592, "learning_rate": 7.05309180406406e-06, "loss": 0.2451, "num_input_tokens_seen": 7970488, "step": 29735 }, { "epoch": 7.795543905635649, "grad_norm": 4.666811943054199, "learning_rate": 7.045131382307399e-06, "loss": 0.164, "num_input_tokens_seen": 7971560, "step": 29740 }, { "epoch": 7.796854521625164, "grad_norm": 3.1697065830230713, "learning_rate": 7.0371747184737075e-06, "loss": 0.2406, "num_input_tokens_seen": 7973032, "step": 29745 }, { "epoch": 7.798165137614679, "grad_norm": 1.8610944747924805, "learning_rate": 7.02922181422832e-06, "loss": 0.2919, "num_input_tokens_seen": 7975176, "step": 29750 }, { "epoch": 7.799475753604194, "grad_norm": 2.5031158924102783, "learning_rate": 7.021272671235754e-06, "loss": 0.3754, "num_input_tokens_seen": 7976392, "step": 29755 }, { "epoch": 7.800786369593709, "grad_norm": 1.8906100988388062, "learning_rate": 7.013327291159755e-06, "loss": 0.1985, "num_input_tokens_seen": 7977848, "step": 29760 }, { "epoch": 7.8020969855832245, "grad_norm": 1.5056548118591309, "learning_rate": 7.0053856756632815e-06, "loss": 0.1854, "num_input_tokens_seen": 7979368, "step": 29765 }, { "epoch": 7.803407601572739, "grad_norm": 4.925436019897461, "learning_rate": 6.997447826408485e-06, "loss": 0.3581, "num_input_tokens_seen": 7980616, "step": 29770 }, { "epoch": 7.804718217562254, "grad_norm": 6.2863078117370605, "learning_rate": 6.989513745056753e-06, "loss": 0.172, "num_input_tokens_seen": 7981688, "step": 29775 }, { "epoch": 7.806028833551769, "grad_norm": 3.2342824935913086, "learning_rate": 6.981583433268676e-06, "loss": 0.215, "num_input_tokens_seen": 7983112, "step": 29780 }, { "epoch": 7.807339449541285, "grad_norm": 0.43990254402160645, "learning_rate": 6.97365689270405e-06, "loss": 0.1579, "num_input_tokens_seen": 7984088, "step": 29785 }, { "epoch": 7.8086500655307995, "grad_norm": 2.8296267986297607, "learning_rate": 6.965734125021892e-06, "loss": 0.1622, "num_input_tokens_seen": 7985512, "step": 29790 }, { "epoch": 7.809960681520314, "grad_norm": 2.210918426513672, "learning_rate": 6.957815131880421e-06, "loss": 0.2358, "num_input_tokens_seen": 7987096, "step": 29795 }, { "epoch": 7.81127129750983, "grad_norm": 2.524568557739258, "learning_rate": 6.949899914937066e-06, "loss": 0.1911, "num_input_tokens_seen": 7988824, "step": 29800 }, { "epoch": 7.812581913499344, "grad_norm": 1.6364614963531494, "learning_rate": 6.941988475848474e-06, "loss": 0.1443, "num_input_tokens_seen": 7990408, "step": 29805 }, { "epoch": 7.81389252948886, "grad_norm": 1.4469033479690552, "learning_rate": 6.934080816270488e-06, "loss": 0.2344, "num_input_tokens_seen": 7991544, "step": 29810 }, { "epoch": 7.815203145478375, "grad_norm": 1.699363350868225, "learning_rate": 6.926176937858175e-06, "loss": 0.2919, "num_input_tokens_seen": 7992808, "step": 29815 }, { "epoch": 7.81651376146789, "grad_norm": 0.31893137097358704, "learning_rate": 6.9182768422658e-06, "loss": 0.1223, "num_input_tokens_seen": 7994088, "step": 29820 }, { "epoch": 7.817824377457405, "grad_norm": 1.5001541376113892, "learning_rate": 6.910380531146837e-06, "loss": 0.4079, "num_input_tokens_seen": 7995496, "step": 29825 }, { "epoch": 7.81913499344692, "grad_norm": 1.124242901802063, "learning_rate": 6.902488006153973e-06, "loss": 0.2907, "num_input_tokens_seen": 7996856, "step": 29830 }, { "epoch": 7.820445609436435, "grad_norm": 3.586063861846924, "learning_rate": 6.894599268939101e-06, "loss": 0.2525, "num_input_tokens_seen": 7998008, "step": 29835 }, { "epoch": 7.8217562254259505, "grad_norm": 1.0227556228637695, "learning_rate": 6.886714321153315e-06, "loss": 0.262, "num_input_tokens_seen": 7999432, "step": 29840 }, { "epoch": 7.823066841415465, "grad_norm": 1.9156169891357422, "learning_rate": 6.8788331644469265e-06, "loss": 0.1222, "num_input_tokens_seen": 8000680, "step": 29845 }, { "epoch": 7.824377457404981, "grad_norm": 0.4412085711956024, "learning_rate": 6.870955800469453e-06, "loss": 0.0901, "num_input_tokens_seen": 8001880, "step": 29850 }, { "epoch": 7.825688073394495, "grad_norm": 5.774625301361084, "learning_rate": 6.863082230869597e-06, "loss": 0.2098, "num_input_tokens_seen": 8003288, "step": 29855 }, { "epoch": 7.826998689384011, "grad_norm": 1.99695885181427, "learning_rate": 6.855212457295285e-06, "loss": 0.2565, "num_input_tokens_seen": 8004888, "step": 29860 }, { "epoch": 7.8283093053735255, "grad_norm": 2.2804343700408936, "learning_rate": 6.847346481393663e-06, "loss": 0.1663, "num_input_tokens_seen": 8005944, "step": 29865 }, { "epoch": 7.82961992136304, "grad_norm": 6.022218227386475, "learning_rate": 6.839484304811055e-06, "loss": 0.2069, "num_input_tokens_seen": 8007240, "step": 29870 }, { "epoch": 7.830930537352556, "grad_norm": 1.3433631658554077, "learning_rate": 6.831625929193006e-06, "loss": 0.1757, "num_input_tokens_seen": 8008440, "step": 29875 }, { "epoch": 7.832241153342071, "grad_norm": 2.137457847595215, "learning_rate": 6.8237713561842556e-06, "loss": 0.1841, "num_input_tokens_seen": 8009576, "step": 29880 }, { "epoch": 7.833551769331586, "grad_norm": 2.850203037261963, "learning_rate": 6.815920587428759e-06, "loss": 0.1938, "num_input_tokens_seen": 8011288, "step": 29885 }, { "epoch": 7.834862385321101, "grad_norm": 3.6986172199249268, "learning_rate": 6.808073624569661e-06, "loss": 0.2306, "num_input_tokens_seen": 8014760, "step": 29890 }, { "epoch": 7.836173001310616, "grad_norm": 5.24623966217041, "learning_rate": 6.8002304692493165e-06, "loss": 0.3203, "num_input_tokens_seen": 8015736, "step": 29895 }, { "epoch": 7.837483617300131, "grad_norm": 1.5555472373962402, "learning_rate": 6.79239112310929e-06, "loss": 0.179, "num_input_tokens_seen": 8017032, "step": 29900 }, { "epoch": 7.838794233289646, "grad_norm": 0.95973140001297, "learning_rate": 6.784555587790342e-06, "loss": 0.1965, "num_input_tokens_seen": 8018600, "step": 29905 }, { "epoch": 7.840104849279161, "grad_norm": 1.5080071687698364, "learning_rate": 6.7767238649324325e-06, "loss": 0.2461, "num_input_tokens_seen": 8019944, "step": 29910 }, { "epoch": 7.8414154652686765, "grad_norm": 2.1287448406219482, "learning_rate": 6.768895956174726e-06, "loss": 0.2694, "num_input_tokens_seen": 8021784, "step": 29915 }, { "epoch": 7.842726081258191, "grad_norm": 1.9344820976257324, "learning_rate": 6.761071863155607e-06, "loss": 0.3051, "num_input_tokens_seen": 8022920, "step": 29920 }, { "epoch": 7.844036697247707, "grad_norm": 2.0828402042388916, "learning_rate": 6.7532515875126295e-06, "loss": 0.2666, "num_input_tokens_seen": 8024200, "step": 29925 }, { "epoch": 7.845347313237221, "grad_norm": 3.249284029006958, "learning_rate": 6.7454351308825634e-06, "loss": 0.2683, "num_input_tokens_seen": 8025640, "step": 29930 }, { "epoch": 7.846657929226737, "grad_norm": 2.4616219997406006, "learning_rate": 6.737622494901383e-06, "loss": 0.2895, "num_input_tokens_seen": 8026696, "step": 29935 }, { "epoch": 7.8479685452162515, "grad_norm": 5.0297088623046875, "learning_rate": 6.7298136812042615e-06, "loss": 0.1666, "num_input_tokens_seen": 8027768, "step": 29940 }, { "epoch": 7.849279161205767, "grad_norm": 3.200448989868164, "learning_rate": 6.722008691425566e-06, "loss": 0.2042, "num_input_tokens_seen": 8029384, "step": 29945 }, { "epoch": 7.850589777195282, "grad_norm": 6.174816608428955, "learning_rate": 6.7142075271988705e-06, "loss": 0.1857, "num_input_tokens_seen": 8031144, "step": 29950 }, { "epoch": 7.851900393184797, "grad_norm": 3.522519588470459, "learning_rate": 6.706410190156945e-06, "loss": 0.1748, "num_input_tokens_seen": 8032536, "step": 29955 }, { "epoch": 7.853211009174312, "grad_norm": 2.10976243019104, "learning_rate": 6.698616681931754e-06, "loss": 0.2316, "num_input_tokens_seen": 8033672, "step": 29960 }, { "epoch": 7.854521625163827, "grad_norm": 4.571808815002441, "learning_rate": 6.6908270041544815e-06, "loss": 0.2631, "num_input_tokens_seen": 8035112, "step": 29965 }, { "epoch": 7.855832241153342, "grad_norm": 1.981835961341858, "learning_rate": 6.683041158455472e-06, "loss": 0.2685, "num_input_tokens_seen": 8036264, "step": 29970 }, { "epoch": 7.857142857142857, "grad_norm": 2.4248616695404053, "learning_rate": 6.675259146464291e-06, "loss": 0.1297, "num_input_tokens_seen": 8037768, "step": 29975 }, { "epoch": 7.858453473132372, "grad_norm": 2.4111685752868652, "learning_rate": 6.667480969809714e-06, "loss": 0.2072, "num_input_tokens_seen": 8038840, "step": 29980 }, { "epoch": 7.859764089121887, "grad_norm": 0.9993346929550171, "learning_rate": 6.659706630119697e-06, "loss": 0.1401, "num_input_tokens_seen": 8040184, "step": 29985 }, { "epoch": 7.8610747051114025, "grad_norm": 1.551667332649231, "learning_rate": 6.651936129021391e-06, "loss": 0.1327, "num_input_tokens_seen": 8041576, "step": 29990 }, { "epoch": 7.862385321100917, "grad_norm": 1.1068382263183594, "learning_rate": 6.644169468141148e-06, "loss": 0.126, "num_input_tokens_seen": 8043064, "step": 29995 }, { "epoch": 7.863695937090433, "grad_norm": 3.7594659328460693, "learning_rate": 6.636406649104529e-06, "loss": 0.3573, "num_input_tokens_seen": 8044536, "step": 30000 }, { "epoch": 7.865006553079947, "grad_norm": 1.4383833408355713, "learning_rate": 6.628647673536262e-06, "loss": 0.1804, "num_input_tokens_seen": 8045944, "step": 30005 }, { "epoch": 7.866317169069463, "grad_norm": 4.195449352264404, "learning_rate": 6.620892543060292e-06, "loss": 0.1959, "num_input_tokens_seen": 8047096, "step": 30010 }, { "epoch": 7.8676277850589775, "grad_norm": 0.27907395362854004, "learning_rate": 6.613141259299754e-06, "loss": 0.1142, "num_input_tokens_seen": 8048344, "step": 30015 }, { "epoch": 7.868938401048493, "grad_norm": 1.5397826433181763, "learning_rate": 6.6053938238769845e-06, "loss": 0.1593, "num_input_tokens_seen": 8049320, "step": 30020 }, { "epoch": 7.870249017038008, "grad_norm": 1.8218250274658203, "learning_rate": 6.5976502384135025e-06, "loss": 0.1657, "num_input_tokens_seen": 8050392, "step": 30025 }, { "epoch": 7.871559633027523, "grad_norm": 1.7467329502105713, "learning_rate": 6.58991050453002e-06, "loss": 0.2378, "num_input_tokens_seen": 8052744, "step": 30030 }, { "epoch": 7.872870249017038, "grad_norm": 2.2533209323883057, "learning_rate": 6.582174623846477e-06, "loss": 0.1591, "num_input_tokens_seen": 8054056, "step": 30035 }, { "epoch": 7.8741808650065535, "grad_norm": 1.6588726043701172, "learning_rate": 6.57444259798195e-06, "loss": 0.1523, "num_input_tokens_seen": 8055208, "step": 30040 }, { "epoch": 7.875491480996068, "grad_norm": 11.784451484680176, "learning_rate": 6.566714428554752e-06, "loss": 0.1274, "num_input_tokens_seen": 8056392, "step": 30045 }, { "epoch": 7.876802096985584, "grad_norm": 2.0112650394439697, "learning_rate": 6.558990117182376e-06, "loss": 0.2327, "num_input_tokens_seen": 8057752, "step": 30050 }, { "epoch": 7.878112712975098, "grad_norm": 1.1781842708587646, "learning_rate": 6.551269665481502e-06, "loss": 0.1627, "num_input_tokens_seen": 8059176, "step": 30055 }, { "epoch": 7.879423328964613, "grad_norm": 0.5790190696716309, "learning_rate": 6.543553075068012e-06, "loss": 0.098, "num_input_tokens_seen": 8060296, "step": 30060 }, { "epoch": 7.8807339449541285, "grad_norm": 3.077826976776123, "learning_rate": 6.535840347556971e-06, "loss": 0.217, "num_input_tokens_seen": 8061640, "step": 30065 }, { "epoch": 7.882044560943643, "grad_norm": 3.7948899269104004, "learning_rate": 6.528131484562644e-06, "loss": 0.3479, "num_input_tokens_seen": 8063192, "step": 30070 }, { "epoch": 7.883355176933159, "grad_norm": 1.5566157102584839, "learning_rate": 6.520426487698478e-06, "loss": 0.1381, "num_input_tokens_seen": 8064472, "step": 30075 }, { "epoch": 7.884665792922673, "grad_norm": 2.6323399543762207, "learning_rate": 6.512725358577121e-06, "loss": 0.2343, "num_input_tokens_seen": 8065896, "step": 30080 }, { "epoch": 7.885976408912189, "grad_norm": 6.558634281158447, "learning_rate": 6.505028098810406e-06, "loss": 0.1795, "num_input_tokens_seen": 8066920, "step": 30085 }, { "epoch": 7.8872870249017035, "grad_norm": 4.3448357582092285, "learning_rate": 6.497334710009342e-06, "loss": 0.148, "num_input_tokens_seen": 8068248, "step": 30090 }, { "epoch": 7.888597640891219, "grad_norm": 6.619223117828369, "learning_rate": 6.489645193784158e-06, "loss": 0.2811, "num_input_tokens_seen": 8069352, "step": 30095 }, { "epoch": 7.889908256880734, "grad_norm": 2.9748435020446777, "learning_rate": 6.481959551744254e-06, "loss": 0.3195, "num_input_tokens_seen": 8070696, "step": 30100 }, { "epoch": 7.891218872870249, "grad_norm": 2.306018829345703, "learning_rate": 6.474277785498217e-06, "loss": 0.1769, "num_input_tokens_seen": 8072024, "step": 30105 }, { "epoch": 7.892529488859764, "grad_norm": 2.014460563659668, "learning_rate": 6.46659989665383e-06, "loss": 0.2396, "num_input_tokens_seen": 8073384, "step": 30110 }, { "epoch": 7.8938401048492794, "grad_norm": 2.789626359939575, "learning_rate": 6.458925886818062e-06, "loss": 0.2876, "num_input_tokens_seen": 8074888, "step": 30115 }, { "epoch": 7.895150720838794, "grad_norm": 1.718971610069275, "learning_rate": 6.451255757597078e-06, "loss": 0.1873, "num_input_tokens_seen": 8076456, "step": 30120 }, { "epoch": 7.89646133682831, "grad_norm": 1.9652369022369385, "learning_rate": 6.443589510596204e-06, "loss": 0.1884, "num_input_tokens_seen": 8077512, "step": 30125 }, { "epoch": 7.897771952817824, "grad_norm": 11.596253395080566, "learning_rate": 6.435927147419984e-06, "loss": 0.1457, "num_input_tokens_seen": 8078392, "step": 30130 }, { "epoch": 7.89908256880734, "grad_norm": 1.5882755517959595, "learning_rate": 6.4282686696721356e-06, "loss": 0.2361, "num_input_tokens_seen": 8080120, "step": 30135 }, { "epoch": 7.9003931847968545, "grad_norm": 1.2109239101409912, "learning_rate": 6.420614078955564e-06, "loss": 0.1406, "num_input_tokens_seen": 8081512, "step": 30140 }, { "epoch": 7.90170380078637, "grad_norm": 1.1287133693695068, "learning_rate": 6.412963376872358e-06, "loss": 0.158, "num_input_tokens_seen": 8083320, "step": 30145 }, { "epoch": 7.903014416775885, "grad_norm": 1.7634916305541992, "learning_rate": 6.405316565023806e-06, "loss": 0.1551, "num_input_tokens_seen": 8084584, "step": 30150 }, { "epoch": 7.904325032765399, "grad_norm": 1.8294787406921387, "learning_rate": 6.397673645010377e-06, "loss": 0.1068, "num_input_tokens_seen": 8085800, "step": 30155 }, { "epoch": 7.905635648754915, "grad_norm": 3.783543348312378, "learning_rate": 6.390034618431704e-06, "loss": 0.1988, "num_input_tokens_seen": 8086952, "step": 30160 }, { "epoch": 7.9069462647444295, "grad_norm": 2.381500482559204, "learning_rate": 6.382399486886631e-06, "loss": 0.2057, "num_input_tokens_seen": 8088296, "step": 30165 }, { "epoch": 7.908256880733945, "grad_norm": 3.2405409812927246, "learning_rate": 6.374768251973174e-06, "loss": 0.0951, "num_input_tokens_seen": 8089528, "step": 30170 }, { "epoch": 7.90956749672346, "grad_norm": 1.7107802629470825, "learning_rate": 6.367140915288542e-06, "loss": 0.29, "num_input_tokens_seen": 8091192, "step": 30175 }, { "epoch": 7.910878112712975, "grad_norm": 3.010535717010498, "learning_rate": 6.3595174784291195e-06, "loss": 0.2613, "num_input_tokens_seen": 8093176, "step": 30180 }, { "epoch": 7.91218872870249, "grad_norm": 9.497936248779297, "learning_rate": 6.351897942990481e-06, "loss": 0.2411, "num_input_tokens_seen": 8094408, "step": 30185 }, { "epoch": 7.913499344692005, "grad_norm": 3.0408775806427, "learning_rate": 6.344282310567384e-06, "loss": 0.2323, "num_input_tokens_seen": 8096088, "step": 30190 }, { "epoch": 7.91480996068152, "grad_norm": 2.163917064666748, "learning_rate": 6.336670582753762e-06, "loss": 0.2145, "num_input_tokens_seen": 8097224, "step": 30195 }, { "epoch": 7.916120576671036, "grad_norm": 1.007785439491272, "learning_rate": 6.329062761142748e-06, "loss": 0.1122, "num_input_tokens_seen": 8098216, "step": 30200 }, { "epoch": 7.91743119266055, "grad_norm": 4.1211700439453125, "learning_rate": 6.3214588473266225e-06, "loss": 0.1274, "num_input_tokens_seen": 8099400, "step": 30205 }, { "epoch": 7.918741808650066, "grad_norm": 2.301992416381836, "learning_rate": 6.313858842896895e-06, "loss": 0.1618, "num_input_tokens_seen": 8100312, "step": 30210 }, { "epoch": 7.9200524246395805, "grad_norm": 3.9563770294189453, "learning_rate": 6.306262749444219e-06, "loss": 0.2127, "num_input_tokens_seen": 8101672, "step": 30215 }, { "epoch": 7.921363040629096, "grad_norm": 2.593441963195801, "learning_rate": 6.298670568558454e-06, "loss": 0.1436, "num_input_tokens_seen": 8103000, "step": 30220 }, { "epoch": 7.922673656618611, "grad_norm": 0.6539732813835144, "learning_rate": 6.291082301828624e-06, "loss": 0.1355, "num_input_tokens_seen": 8104104, "step": 30225 }, { "epoch": 7.923984272608125, "grad_norm": 1.7247099876403809, "learning_rate": 6.283497950842942e-06, "loss": 0.1685, "num_input_tokens_seen": 8105432, "step": 30230 }, { "epoch": 7.925294888597641, "grad_norm": 2.910801887512207, "learning_rate": 6.275917517188809e-06, "loss": 0.2323, "num_input_tokens_seen": 8106424, "step": 30235 }, { "epoch": 7.926605504587156, "grad_norm": 1.5982897281646729, "learning_rate": 6.2683410024527764e-06, "loss": 0.1393, "num_input_tokens_seen": 8107928, "step": 30240 }, { "epoch": 7.927916120576671, "grad_norm": 3.5625219345092773, "learning_rate": 6.260768408220607e-06, "loss": 0.2053, "num_input_tokens_seen": 8109208, "step": 30245 }, { "epoch": 7.929226736566186, "grad_norm": 1.5568106174468994, "learning_rate": 6.253199736077231e-06, "loss": 0.1122, "num_input_tokens_seen": 8110472, "step": 30250 }, { "epoch": 7.930537352555701, "grad_norm": 0.36800476908683777, "learning_rate": 6.24563498760676e-06, "loss": 0.2159, "num_input_tokens_seen": 8111688, "step": 30255 }, { "epoch": 7.931847968545216, "grad_norm": 1.2643896341323853, "learning_rate": 6.238074164392471e-06, "loss": 0.1312, "num_input_tokens_seen": 8112872, "step": 30260 }, { "epoch": 7.933158584534731, "grad_norm": 1.4573373794555664, "learning_rate": 6.23051726801685e-06, "loss": 0.1404, "num_input_tokens_seen": 8113944, "step": 30265 }, { "epoch": 7.934469200524246, "grad_norm": 3.1149520874023438, "learning_rate": 6.22296430006154e-06, "loss": 0.1236, "num_input_tokens_seen": 8115384, "step": 30270 }, { "epoch": 7.935779816513762, "grad_norm": 3.471351146697998, "learning_rate": 6.215415262107352e-06, "loss": 0.3267, "num_input_tokens_seen": 8118120, "step": 30275 }, { "epoch": 7.937090432503276, "grad_norm": 3.521735906600952, "learning_rate": 6.207870155734291e-06, "loss": 0.2003, "num_input_tokens_seen": 8119176, "step": 30280 }, { "epoch": 7.938401048492792, "grad_norm": 1.6701315641403198, "learning_rate": 6.200328982521536e-06, "loss": 0.1799, "num_input_tokens_seen": 8120360, "step": 30285 }, { "epoch": 7.9397116644823065, "grad_norm": 2.4202640056610107, "learning_rate": 6.192791744047444e-06, "loss": 0.2253, "num_input_tokens_seen": 8121528, "step": 30290 }, { "epoch": 7.941022280471822, "grad_norm": 0.8272736072540283, "learning_rate": 6.185258441889541e-06, "loss": 0.1897, "num_input_tokens_seen": 8122856, "step": 30295 }, { "epoch": 7.942332896461337, "grad_norm": 2.5084939002990723, "learning_rate": 6.177729077624539e-06, "loss": 0.2758, "num_input_tokens_seen": 8124072, "step": 30300 }, { "epoch": 7.943643512450852, "grad_norm": 2.403850793838501, "learning_rate": 6.170203652828316e-06, "loss": 0.2289, "num_input_tokens_seen": 8125752, "step": 30305 }, { "epoch": 7.944954128440367, "grad_norm": 2.8539230823516846, "learning_rate": 6.162682169075934e-06, "loss": 0.2855, "num_input_tokens_seen": 8126904, "step": 30310 }, { "epoch": 7.946264744429882, "grad_norm": 7.973361492156982, "learning_rate": 6.155164627941626e-06, "loss": 0.21, "num_input_tokens_seen": 8128168, "step": 30315 }, { "epoch": 7.947575360419397, "grad_norm": 2.1792891025543213, "learning_rate": 6.147651030998799e-06, "loss": 0.2679, "num_input_tokens_seen": 8129544, "step": 30320 }, { "epoch": 7.948885976408912, "grad_norm": 1.9918161630630493, "learning_rate": 6.140141379820038e-06, "loss": 0.2015, "num_input_tokens_seen": 8130616, "step": 30325 }, { "epoch": 7.950196592398427, "grad_norm": 4.841213226318359, "learning_rate": 6.132635675977099e-06, "loss": 0.2053, "num_input_tokens_seen": 8131432, "step": 30330 }, { "epoch": 7.951507208387943, "grad_norm": 2.748112201690674, "learning_rate": 6.125133921040912e-06, "loss": 0.0971, "num_input_tokens_seen": 8132520, "step": 30335 }, { "epoch": 7.952817824377457, "grad_norm": 1.9933620691299438, "learning_rate": 6.1176361165815845e-06, "loss": 0.1421, "num_input_tokens_seen": 8133720, "step": 30340 }, { "epoch": 7.954128440366972, "grad_norm": 1.3210337162017822, "learning_rate": 6.1101422641683895e-06, "loss": 0.1611, "num_input_tokens_seen": 8135528, "step": 30345 }, { "epoch": 7.955439056356488, "grad_norm": 1.7511035203933716, "learning_rate": 6.102652365369779e-06, "loss": 0.1735, "num_input_tokens_seen": 8137240, "step": 30350 }, { "epoch": 7.956749672346002, "grad_norm": 1.0502355098724365, "learning_rate": 6.095166421753385e-06, "loss": 0.301, "num_input_tokens_seen": 8138296, "step": 30355 }, { "epoch": 7.958060288335518, "grad_norm": 0.7960491180419922, "learning_rate": 6.087684434885985e-06, "loss": 0.2056, "num_input_tokens_seen": 8139816, "step": 30360 }, { "epoch": 7.9593709043250325, "grad_norm": 4.019136428833008, "learning_rate": 6.080206406333555e-06, "loss": 0.2342, "num_input_tokens_seen": 8140840, "step": 30365 }, { "epoch": 7.960681520314548, "grad_norm": 1.7378650903701782, "learning_rate": 6.072732337661236e-06, "loss": 0.1037, "num_input_tokens_seen": 8141896, "step": 30370 }, { "epoch": 7.961992136304063, "grad_norm": 2.1300888061523438, "learning_rate": 6.065262230433325e-06, "loss": 0.2109, "num_input_tokens_seen": 8143528, "step": 30375 }, { "epoch": 7.963302752293578, "grad_norm": 2.375610589981079, "learning_rate": 6.057796086213319e-06, "loss": 0.256, "num_input_tokens_seen": 8144920, "step": 30380 }, { "epoch": 7.964613368283093, "grad_norm": 2.1505491733551025, "learning_rate": 6.050333906563865e-06, "loss": 0.1768, "num_input_tokens_seen": 8146392, "step": 30385 }, { "epoch": 7.965923984272608, "grad_norm": 2.8870596885681152, "learning_rate": 6.0428756930467876e-06, "loss": 0.2958, "num_input_tokens_seen": 8147768, "step": 30390 }, { "epoch": 7.967234600262123, "grad_norm": 0.9866217970848083, "learning_rate": 6.035421447223064e-06, "loss": 0.2311, "num_input_tokens_seen": 8148840, "step": 30395 }, { "epoch": 7.968545216251639, "grad_norm": 3.2876081466674805, "learning_rate": 6.027971170652866e-06, "loss": 0.1589, "num_input_tokens_seen": 8150408, "step": 30400 }, { "epoch": 7.969855832241153, "grad_norm": 3.1666975021362305, "learning_rate": 6.020524864895519e-06, "loss": 0.2062, "num_input_tokens_seen": 8151448, "step": 30405 }, { "epoch": 7.971166448230669, "grad_norm": 2.1412882804870605, "learning_rate": 6.013082531509523e-06, "loss": 0.2051, "num_input_tokens_seen": 8152936, "step": 30410 }, { "epoch": 7.972477064220183, "grad_norm": 2.5209736824035645, "learning_rate": 6.005644172052549e-06, "loss": 0.1341, "num_input_tokens_seen": 8153864, "step": 30415 }, { "epoch": 7.973787680209698, "grad_norm": 2.210026979446411, "learning_rate": 5.998209788081427e-06, "loss": 0.2068, "num_input_tokens_seen": 8155192, "step": 30420 }, { "epoch": 7.975098296199214, "grad_norm": 8.584059715270996, "learning_rate": 5.990779381152168e-06, "loss": 0.2602, "num_input_tokens_seen": 8156264, "step": 30425 }, { "epoch": 7.976408912188729, "grad_norm": 2.003688335418701, "learning_rate": 5.9833529528199344e-06, "loss": 0.2116, "num_input_tokens_seen": 8158104, "step": 30430 }, { "epoch": 7.977719528178244, "grad_norm": 1.5950992107391357, "learning_rate": 5.9759305046390715e-06, "loss": 0.346, "num_input_tokens_seen": 8159464, "step": 30435 }, { "epoch": 7.9790301441677585, "grad_norm": 1.0614957809448242, "learning_rate": 5.968512038163085e-06, "loss": 0.1418, "num_input_tokens_seen": 8160520, "step": 30440 }, { "epoch": 7.980340760157274, "grad_norm": 2.6570358276367188, "learning_rate": 5.9610975549446414e-06, "loss": 0.0984, "num_input_tokens_seen": 8161640, "step": 30445 }, { "epoch": 7.981651376146789, "grad_norm": 2.792041778564453, "learning_rate": 5.953687056535584e-06, "loss": 0.1707, "num_input_tokens_seen": 8163064, "step": 30450 }, { "epoch": 7.982961992136304, "grad_norm": 1.831498622894287, "learning_rate": 5.946280544486918e-06, "loss": 0.1852, "num_input_tokens_seen": 8164568, "step": 30455 }, { "epoch": 7.984272608125819, "grad_norm": 1.8450294733047485, "learning_rate": 5.938878020348809e-06, "loss": 0.2851, "num_input_tokens_seen": 8165896, "step": 30460 }, { "epoch": 7.985583224115334, "grad_norm": 1.6289806365966797, "learning_rate": 5.931479485670599e-06, "loss": 0.2657, "num_input_tokens_seen": 8167128, "step": 30465 }, { "epoch": 7.986893840104849, "grad_norm": 1.309530258178711, "learning_rate": 5.924084942000787e-06, "loss": 0.1316, "num_input_tokens_seen": 8168696, "step": 30470 }, { "epoch": 7.988204456094365, "grad_norm": 1.8401983976364136, "learning_rate": 5.916694390887034e-06, "loss": 0.1997, "num_input_tokens_seen": 8170024, "step": 30475 }, { "epoch": 7.989515072083879, "grad_norm": 2.283066987991333, "learning_rate": 5.909307833876168e-06, "loss": 0.1241, "num_input_tokens_seen": 8171496, "step": 30480 }, { "epoch": 7.990825688073395, "grad_norm": 3.209890127182007, "learning_rate": 5.901925272514186e-06, "loss": 0.2448, "num_input_tokens_seen": 8172824, "step": 30485 }, { "epoch": 7.992136304062909, "grad_norm": 1.1189991235733032, "learning_rate": 5.8945467083462405e-06, "loss": 0.3415, "num_input_tokens_seen": 8174408, "step": 30490 }, { "epoch": 7.993446920052425, "grad_norm": 1.8638041019439697, "learning_rate": 5.887172142916663e-06, "loss": 0.2371, "num_input_tokens_seen": 8175688, "step": 30495 }, { "epoch": 7.99475753604194, "grad_norm": 2.0132930278778076, "learning_rate": 5.879801577768934e-06, "loss": 0.2907, "num_input_tokens_seen": 8178136, "step": 30500 }, { "epoch": 7.996068152031455, "grad_norm": 3.634066581726074, "learning_rate": 5.872435014445696e-06, "loss": 0.1914, "num_input_tokens_seen": 8179800, "step": 30505 }, { "epoch": 7.99737876802097, "grad_norm": 2.336609363555908, "learning_rate": 5.865072454488765e-06, "loss": 0.7126, "num_input_tokens_seen": 8181512, "step": 30510 }, { "epoch": 7.9986893840104845, "grad_norm": 5.8480730056762695, "learning_rate": 5.857713899439101e-06, "loss": 0.1246, "num_input_tokens_seen": 8182424, "step": 30515 }, { "epoch": 8.0, "grad_norm": 5.315600395202637, "learning_rate": 5.850359350836842e-06, "loss": 0.2185, "num_input_tokens_seen": 8183584, "step": 30520 }, { "epoch": 8.001310615989516, "grad_norm": 2.633748769760132, "learning_rate": 5.843008810221284e-06, "loss": 0.2948, "num_input_tokens_seen": 8185008, "step": 30525 }, { "epoch": 8.002096985583224, "eval_loss": 0.7577313184738159, "eval_runtime": 16.6489, "eval_samples_per_second": 50.934, "eval_steps_per_second": 25.467, "num_input_tokens_seen": 8185712, "step": 30528 }, { "epoch": 8.00262123197903, "grad_norm": 1.1302164793014526, "learning_rate": 5.835662279130879e-06, "loss": 0.3423, "num_input_tokens_seen": 8186432, "step": 30530 }, { "epoch": 8.003931847968545, "grad_norm": 2.644151210784912, "learning_rate": 5.828319759103248e-06, "loss": 0.1196, "num_input_tokens_seen": 8187968, "step": 30535 }, { "epoch": 8.00524246395806, "grad_norm": 2.8875491619110107, "learning_rate": 5.8209812516751646e-06, "loss": 0.1669, "num_input_tokens_seen": 8189072, "step": 30540 }, { "epoch": 8.006553079947576, "grad_norm": 3.493304491043091, "learning_rate": 5.8136467583825685e-06, "loss": 0.1478, "num_input_tokens_seen": 8190256, "step": 30545 }, { "epoch": 8.00786369593709, "grad_norm": 1.506314992904663, "learning_rate": 5.806316280760551e-06, "loss": 0.1167, "num_input_tokens_seen": 8191440, "step": 30550 }, { "epoch": 8.009174311926605, "grad_norm": 1.1972153186798096, "learning_rate": 5.7989898203433725e-06, "loss": 0.1686, "num_input_tokens_seen": 8192976, "step": 30555 }, { "epoch": 8.01048492791612, "grad_norm": 2.2439374923706055, "learning_rate": 5.791667378664451e-06, "loss": 0.1005, "num_input_tokens_seen": 8194544, "step": 30560 }, { "epoch": 8.011795543905636, "grad_norm": 5.123591899871826, "learning_rate": 5.784348957256352e-06, "loss": 0.1504, "num_input_tokens_seen": 8195728, "step": 30565 }, { "epoch": 8.01310615989515, "grad_norm": 0.6124131083488464, "learning_rate": 5.7770345576508176e-06, "loss": 0.154, "num_input_tokens_seen": 8196992, "step": 30570 }, { "epoch": 8.014416775884666, "grad_norm": 1.8984131813049316, "learning_rate": 5.769724181378736e-06, "loss": 0.1266, "num_input_tokens_seen": 8198544, "step": 30575 }, { "epoch": 8.015727391874181, "grad_norm": 3.63521146774292, "learning_rate": 5.7624178299701546e-06, "loss": 0.1359, "num_input_tokens_seen": 8199504, "step": 30580 }, { "epoch": 8.017038007863697, "grad_norm": 3.3590927124023438, "learning_rate": 5.755115504954284e-06, "loss": 0.168, "num_input_tokens_seen": 8200864, "step": 30585 }, { "epoch": 8.01834862385321, "grad_norm": 1.6536164283752441, "learning_rate": 5.747817207859491e-06, "loss": 0.2716, "num_input_tokens_seen": 8202000, "step": 30590 }, { "epoch": 8.019659239842726, "grad_norm": 2.054706335067749, "learning_rate": 5.740522940213283e-06, "loss": 0.1556, "num_input_tokens_seen": 8203200, "step": 30595 }, { "epoch": 8.020969855832242, "grad_norm": 4.208477020263672, "learning_rate": 5.733232703542346e-06, "loss": 0.2044, "num_input_tokens_seen": 8204480, "step": 30600 }, { "epoch": 8.022280471821757, "grad_norm": 2.8196229934692383, "learning_rate": 5.725946499372506e-06, "loss": 0.1213, "num_input_tokens_seen": 8205616, "step": 30605 }, { "epoch": 8.02359108781127, "grad_norm": 1.5885910987854004, "learning_rate": 5.7186643292287705e-06, "loss": 0.1804, "num_input_tokens_seen": 8208176, "step": 30610 }, { "epoch": 8.024901703800786, "grad_norm": 1.6756213903427124, "learning_rate": 5.711386194635274e-06, "loss": 0.1389, "num_input_tokens_seen": 8209168, "step": 30615 }, { "epoch": 8.026212319790302, "grad_norm": 2.0745909214019775, "learning_rate": 5.704112097115316e-06, "loss": 0.188, "num_input_tokens_seen": 8210464, "step": 30620 }, { "epoch": 8.027522935779816, "grad_norm": 0.13347433507442474, "learning_rate": 5.696842038191366e-06, "loss": 0.1101, "num_input_tokens_seen": 8211616, "step": 30625 }, { "epoch": 8.028833551769331, "grad_norm": 1.5255661010742188, "learning_rate": 5.689576019385015e-06, "loss": 0.1572, "num_input_tokens_seen": 8212832, "step": 30630 }, { "epoch": 8.030144167758847, "grad_norm": 3.474591016769409, "learning_rate": 5.682314042217041e-06, "loss": 0.2468, "num_input_tokens_seen": 8214048, "step": 30635 }, { "epoch": 8.031454783748362, "grad_norm": 1.7913269996643066, "learning_rate": 5.675056108207355e-06, "loss": 0.1044, "num_input_tokens_seen": 8215328, "step": 30640 }, { "epoch": 8.032765399737876, "grad_norm": 2.0716774463653564, "learning_rate": 5.667802218875037e-06, "loss": 0.2197, "num_input_tokens_seen": 8216720, "step": 30645 }, { "epoch": 8.034076015727392, "grad_norm": 2.5264344215393066, "learning_rate": 5.66055237573831e-06, "loss": 0.1888, "num_input_tokens_seen": 8218112, "step": 30650 }, { "epoch": 8.035386631716907, "grad_norm": 1.6517751216888428, "learning_rate": 5.6533065803145545e-06, "loss": 0.1013, "num_input_tokens_seen": 8219552, "step": 30655 }, { "epoch": 8.036697247706423, "grad_norm": 1.1132080554962158, "learning_rate": 5.646064834120304e-06, "loss": 0.1647, "num_input_tokens_seen": 8221136, "step": 30660 }, { "epoch": 8.038007863695936, "grad_norm": 1.844874620437622, "learning_rate": 5.638827138671243e-06, "loss": 0.1978, "num_input_tokens_seen": 8222496, "step": 30665 }, { "epoch": 8.039318479685452, "grad_norm": 3.235424757003784, "learning_rate": 5.631593495482209e-06, "loss": 0.0835, "num_input_tokens_seen": 8224272, "step": 30670 }, { "epoch": 8.040629095674968, "grad_norm": 0.7762516736984253, "learning_rate": 5.62436390606719e-06, "loss": 0.0944, "num_input_tokens_seen": 8225632, "step": 30675 }, { "epoch": 8.041939711664483, "grad_norm": 2.1491050720214844, "learning_rate": 5.617138371939326e-06, "loss": 0.1299, "num_input_tokens_seen": 8226928, "step": 30680 }, { "epoch": 8.043250327653997, "grad_norm": 3.4223151206970215, "learning_rate": 5.6099168946109114e-06, "loss": 0.1827, "num_input_tokens_seen": 8228064, "step": 30685 }, { "epoch": 8.044560943643512, "grad_norm": 3.7910313606262207, "learning_rate": 5.602699475593387e-06, "loss": 0.1365, "num_input_tokens_seen": 8229168, "step": 30690 }, { "epoch": 8.045871559633028, "grad_norm": 1.0028245449066162, "learning_rate": 5.595486116397344e-06, "loss": 0.1354, "num_input_tokens_seen": 8230336, "step": 30695 }, { "epoch": 8.047182175622543, "grad_norm": 2.5103068351745605, "learning_rate": 5.588276818532531e-06, "loss": 0.1573, "num_input_tokens_seen": 8231408, "step": 30700 }, { "epoch": 8.048492791612057, "grad_norm": 2.694333791732788, "learning_rate": 5.5810715835078434e-06, "loss": 0.1367, "num_input_tokens_seen": 8232784, "step": 30705 }, { "epoch": 8.049803407601573, "grad_norm": 1.1198112964630127, "learning_rate": 5.573870412831314e-06, "loss": 0.1698, "num_input_tokens_seen": 8234336, "step": 30710 }, { "epoch": 8.051114023591088, "grad_norm": 1.933044672012329, "learning_rate": 5.566673308010142e-06, "loss": 0.129, "num_input_tokens_seen": 8235392, "step": 30715 }, { "epoch": 8.052424639580602, "grad_norm": 2.1622700691223145, "learning_rate": 5.5594802705506626e-06, "loss": 0.187, "num_input_tokens_seen": 8236608, "step": 30720 }, { "epoch": 8.053735255570118, "grad_norm": 1.6946215629577637, "learning_rate": 5.552291301958379e-06, "loss": 0.1827, "num_input_tokens_seen": 8238096, "step": 30725 }, { "epoch": 8.055045871559633, "grad_norm": 1.770411729812622, "learning_rate": 5.545106403737921e-06, "loss": 0.1404, "num_input_tokens_seen": 8239280, "step": 30730 }, { "epoch": 8.056356487549149, "grad_norm": 1.4108421802520752, "learning_rate": 5.537925577393077e-06, "loss": 0.1781, "num_input_tokens_seen": 8240448, "step": 30735 }, { "epoch": 8.057667103538662, "grad_norm": 3.034613847732544, "learning_rate": 5.530748824426782e-06, "loss": 0.1619, "num_input_tokens_seen": 8241504, "step": 30740 }, { "epoch": 8.058977719528178, "grad_norm": 1.1789047718048096, "learning_rate": 5.523576146341128e-06, "loss": 0.0968, "num_input_tokens_seen": 8242784, "step": 30745 }, { "epoch": 8.060288335517694, "grad_norm": 1.9321460723876953, "learning_rate": 5.5164075446373225e-06, "loss": 0.1786, "num_input_tokens_seen": 8243760, "step": 30750 }, { "epoch": 8.061598951507209, "grad_norm": 1.6260814666748047, "learning_rate": 5.509243020815755e-06, "loss": 0.2466, "num_input_tokens_seen": 8245888, "step": 30755 }, { "epoch": 8.062909567496723, "grad_norm": 2.1546552181243896, "learning_rate": 5.502082576375947e-06, "loss": 0.1369, "num_input_tokens_seen": 8246896, "step": 30760 }, { "epoch": 8.064220183486238, "grad_norm": 1.3225247859954834, "learning_rate": 5.494926212816562e-06, "loss": 0.182, "num_input_tokens_seen": 8248704, "step": 30765 }, { "epoch": 8.065530799475754, "grad_norm": 2.5336594581604004, "learning_rate": 5.4877739316354236e-06, "loss": 0.1829, "num_input_tokens_seen": 8250368, "step": 30770 }, { "epoch": 8.06684141546527, "grad_norm": 0.9671874046325684, "learning_rate": 5.4806257343294755e-06, "loss": 0.1571, "num_input_tokens_seen": 8252000, "step": 30775 }, { "epoch": 8.068152031454783, "grad_norm": 0.5707082152366638, "learning_rate": 5.473481622394849e-06, "loss": 0.1158, "num_input_tokens_seen": 8253248, "step": 30780 }, { "epoch": 8.069462647444299, "grad_norm": 2.709831714630127, "learning_rate": 5.466341597326774e-06, "loss": 0.145, "num_input_tokens_seen": 8254400, "step": 30785 }, { "epoch": 8.070773263433814, "grad_norm": 3.215623140335083, "learning_rate": 5.459205660619651e-06, "loss": 0.1718, "num_input_tokens_seen": 8255648, "step": 30790 }, { "epoch": 8.07208387942333, "grad_norm": 1.2389049530029297, "learning_rate": 5.452073813767019e-06, "loss": 0.1564, "num_input_tokens_seen": 8256928, "step": 30795 }, { "epoch": 8.073394495412844, "grad_norm": 1.8191207647323608, "learning_rate": 5.444946058261563e-06, "loss": 0.1408, "num_input_tokens_seen": 8258368, "step": 30800 }, { "epoch": 8.07470511140236, "grad_norm": 68.4483413696289, "learning_rate": 5.4378223955951074e-06, "loss": 0.2378, "num_input_tokens_seen": 8260864, "step": 30805 }, { "epoch": 8.076015727391875, "grad_norm": 2.732006072998047, "learning_rate": 5.430702827258622e-06, "loss": 0.1214, "num_input_tokens_seen": 8261984, "step": 30810 }, { "epoch": 8.077326343381388, "grad_norm": 0.7009357213973999, "learning_rate": 5.423587354742224e-06, "loss": 0.0946, "num_input_tokens_seen": 8263424, "step": 30815 }, { "epoch": 8.078636959370904, "grad_norm": 1.2298024892807007, "learning_rate": 5.4164759795351655e-06, "loss": 0.0879, "num_input_tokens_seen": 8264672, "step": 30820 }, { "epoch": 8.07994757536042, "grad_norm": 2.5948269367218018, "learning_rate": 5.4093687031258554e-06, "loss": 0.1413, "num_input_tokens_seen": 8266272, "step": 30825 }, { "epoch": 8.081258191349935, "grad_norm": 3.985487699508667, "learning_rate": 5.402265527001818e-06, "loss": 0.149, "num_input_tokens_seen": 8267296, "step": 30830 }, { "epoch": 8.082568807339449, "grad_norm": 2.029663562774658, "learning_rate": 5.3951664526497356e-06, "loss": 0.2399, "num_input_tokens_seen": 8268656, "step": 30835 }, { "epoch": 8.083879423328964, "grad_norm": 2.348057508468628, "learning_rate": 5.388071481555448e-06, "loss": 0.1154, "num_input_tokens_seen": 8269904, "step": 30840 }, { "epoch": 8.08519003931848, "grad_norm": 4.310263633728027, "learning_rate": 5.3809806152039125e-06, "loss": 0.1591, "num_input_tokens_seen": 8270992, "step": 30845 }, { "epoch": 8.086500655307995, "grad_norm": 3.139551877975464, "learning_rate": 5.373893855079235e-06, "loss": 0.1791, "num_input_tokens_seen": 8272352, "step": 30850 }, { "epoch": 8.08781127129751, "grad_norm": 19.397871017456055, "learning_rate": 5.366811202664662e-06, "loss": 0.1697, "num_input_tokens_seen": 8273424, "step": 30855 }, { "epoch": 8.089121887287025, "grad_norm": 2.2742090225219727, "learning_rate": 5.35973265944259e-06, "loss": 0.1959, "num_input_tokens_seen": 8274976, "step": 30860 }, { "epoch": 8.09043250327654, "grad_norm": 3.2187418937683105, "learning_rate": 5.352658226894527e-06, "loss": 0.1756, "num_input_tokens_seen": 8276384, "step": 30865 }, { "epoch": 8.091743119266056, "grad_norm": 2.0189666748046875, "learning_rate": 5.345587906501148e-06, "loss": 0.1193, "num_input_tokens_seen": 8277920, "step": 30870 }, { "epoch": 8.09305373525557, "grad_norm": 1.9398599863052368, "learning_rate": 5.338521699742263e-06, "loss": 0.1921, "num_input_tokens_seen": 8279488, "step": 30875 }, { "epoch": 8.094364351245085, "grad_norm": 2.183544635772705, "learning_rate": 5.331459608096815e-06, "loss": 0.1919, "num_input_tokens_seen": 8280912, "step": 30880 }, { "epoch": 8.0956749672346, "grad_norm": 1.5519286394119263, "learning_rate": 5.324401633042883e-06, "loss": 0.2094, "num_input_tokens_seen": 8282016, "step": 30885 }, { "epoch": 8.096985583224116, "grad_norm": 1.74008047580719, "learning_rate": 5.317347776057685e-06, "loss": 0.068, "num_input_tokens_seen": 8283120, "step": 30890 }, { "epoch": 8.09829619921363, "grad_norm": 3.914597988128662, "learning_rate": 5.310298038617606e-06, "loss": 0.0998, "num_input_tokens_seen": 8284496, "step": 30895 }, { "epoch": 8.099606815203146, "grad_norm": 4.15565299987793, "learning_rate": 5.303252422198115e-06, "loss": 0.1794, "num_input_tokens_seen": 8285840, "step": 30900 }, { "epoch": 8.100917431192661, "grad_norm": 0.8132625222206116, "learning_rate": 5.296210928273862e-06, "loss": 0.2137, "num_input_tokens_seen": 8287008, "step": 30905 }, { "epoch": 8.102228047182175, "grad_norm": 5.474616050720215, "learning_rate": 5.289173558318617e-06, "loss": 0.179, "num_input_tokens_seen": 8288416, "step": 30910 }, { "epoch": 8.10353866317169, "grad_norm": 1.4350123405456543, "learning_rate": 5.282140313805284e-06, "loss": 0.0812, "num_input_tokens_seen": 8289808, "step": 30915 }, { "epoch": 8.104849279161206, "grad_norm": 1.1181435585021973, "learning_rate": 5.275111196205917e-06, "loss": 0.1078, "num_input_tokens_seen": 8291184, "step": 30920 }, { "epoch": 8.106159895150721, "grad_norm": 2.3327889442443848, "learning_rate": 5.268086206991693e-06, "loss": 0.1717, "num_input_tokens_seen": 8292784, "step": 30925 }, { "epoch": 8.107470511140235, "grad_norm": 2.597627878189087, "learning_rate": 5.2610653476329345e-06, "loss": 0.1445, "num_input_tokens_seen": 8294304, "step": 30930 }, { "epoch": 8.10878112712975, "grad_norm": 1.9816697835922241, "learning_rate": 5.254048619599089e-06, "loss": 0.1072, "num_input_tokens_seen": 8295488, "step": 30935 }, { "epoch": 8.110091743119266, "grad_norm": 2.3611483573913574, "learning_rate": 5.247036024358759e-06, "loss": 0.1179, "num_input_tokens_seen": 8296544, "step": 30940 }, { "epoch": 8.111402359108782, "grad_norm": 1.396254539489746, "learning_rate": 5.240027563379649e-06, "loss": 0.237, "num_input_tokens_seen": 8299472, "step": 30945 }, { "epoch": 8.112712975098296, "grad_norm": 2.6721408367156982, "learning_rate": 5.233023238128623e-06, "loss": 0.1362, "num_input_tokens_seen": 8302656, "step": 30950 }, { "epoch": 8.114023591087811, "grad_norm": 1.73038649559021, "learning_rate": 5.226023050071682e-06, "loss": 0.1332, "num_input_tokens_seen": 8304304, "step": 30955 }, { "epoch": 8.115334207077327, "grad_norm": 0.17992624640464783, "learning_rate": 5.219027000673954e-06, "loss": 0.0718, "num_input_tokens_seen": 8305344, "step": 30960 }, { "epoch": 8.116644823066842, "grad_norm": 1.3407448530197144, "learning_rate": 5.212035091399694e-06, "loss": 0.1392, "num_input_tokens_seen": 8306560, "step": 30965 }, { "epoch": 8.117955439056356, "grad_norm": 3.4515957832336426, "learning_rate": 5.205047323712298e-06, "loss": 0.1469, "num_input_tokens_seen": 8307888, "step": 30970 }, { "epoch": 8.119266055045872, "grad_norm": 2.0831336975097656, "learning_rate": 5.198063699074293e-06, "loss": 0.127, "num_input_tokens_seen": 8309424, "step": 30975 }, { "epoch": 8.120576671035387, "grad_norm": 2.712031126022339, "learning_rate": 5.191084218947351e-06, "loss": 0.117, "num_input_tokens_seen": 8310752, "step": 30980 }, { "epoch": 8.1218872870249, "grad_norm": 1.0548748970031738, "learning_rate": 5.184108884792244e-06, "loss": 0.0813, "num_input_tokens_seen": 8311888, "step": 30985 }, { "epoch": 8.123197903014416, "grad_norm": 4.120144367218018, "learning_rate": 5.177137698068912e-06, "loss": 0.1972, "num_input_tokens_seen": 8312896, "step": 30990 }, { "epoch": 8.124508519003932, "grad_norm": 2.2447493076324463, "learning_rate": 5.170170660236406e-06, "loss": 0.1856, "num_input_tokens_seen": 8314144, "step": 30995 }, { "epoch": 8.125819134993447, "grad_norm": 2.4121620655059814, "learning_rate": 5.163207772752918e-06, "loss": 0.1081, "num_input_tokens_seen": 8315168, "step": 31000 }, { "epoch": 8.127129750982961, "grad_norm": 2.4007742404937744, "learning_rate": 5.156249037075761e-06, "loss": 0.2146, "num_input_tokens_seen": 8316560, "step": 31005 }, { "epoch": 8.128440366972477, "grad_norm": 2.1110806465148926, "learning_rate": 5.149294454661399e-06, "loss": 0.154, "num_input_tokens_seen": 8317904, "step": 31010 }, { "epoch": 8.129750982961992, "grad_norm": 2.3873767852783203, "learning_rate": 5.142344026965418e-06, "loss": 0.0472, "num_input_tokens_seen": 8318960, "step": 31015 }, { "epoch": 8.131061598951508, "grad_norm": 2.760911226272583, "learning_rate": 5.135397755442512e-06, "loss": 0.1025, "num_input_tokens_seen": 8320352, "step": 31020 }, { "epoch": 8.132372214941022, "grad_norm": 2.2207260131835938, "learning_rate": 5.1284556415465315e-06, "loss": 0.2823, "num_input_tokens_seen": 8322048, "step": 31025 }, { "epoch": 8.133682830930537, "grad_norm": 1.1331685781478882, "learning_rate": 5.121517686730451e-06, "loss": 0.208, "num_input_tokens_seen": 8323584, "step": 31030 }, { "epoch": 8.134993446920053, "grad_norm": 3.241791009902954, "learning_rate": 5.1145838924463725e-06, "loss": 0.1153, "num_input_tokens_seen": 8324944, "step": 31035 }, { "epoch": 8.136304062909568, "grad_norm": 1.448380947113037, "learning_rate": 5.107654260145523e-06, "loss": 0.125, "num_input_tokens_seen": 8326352, "step": 31040 }, { "epoch": 8.137614678899082, "grad_norm": 0.48027729988098145, "learning_rate": 5.100728791278267e-06, "loss": 0.1457, "num_input_tokens_seen": 8327664, "step": 31045 }, { "epoch": 8.138925294888598, "grad_norm": 2.519641876220703, "learning_rate": 5.0938074872940915e-06, "loss": 0.1348, "num_input_tokens_seen": 8328848, "step": 31050 }, { "epoch": 8.140235910878113, "grad_norm": 0.9398940801620483, "learning_rate": 5.086890349641616e-06, "loss": 0.1378, "num_input_tokens_seen": 8329984, "step": 31055 }, { "epoch": 8.141546526867629, "grad_norm": 1.7639001607894897, "learning_rate": 5.079977379768588e-06, "loss": 0.1435, "num_input_tokens_seen": 8331168, "step": 31060 }, { "epoch": 8.142857142857142, "grad_norm": 3.6395583152770996, "learning_rate": 5.073068579121862e-06, "loss": 0.1557, "num_input_tokens_seen": 8332496, "step": 31065 }, { "epoch": 8.144167758846658, "grad_norm": 4.139388084411621, "learning_rate": 5.066163949147457e-06, "loss": 0.207, "num_input_tokens_seen": 8333616, "step": 31070 }, { "epoch": 8.145478374836173, "grad_norm": 3.101496458053589, "learning_rate": 5.059263491290495e-06, "loss": 0.1421, "num_input_tokens_seen": 8334928, "step": 31075 }, { "epoch": 8.146788990825687, "grad_norm": 1.8716093301773071, "learning_rate": 5.052367206995229e-06, "loss": 0.1558, "num_input_tokens_seen": 8336432, "step": 31080 }, { "epoch": 8.148099606815203, "grad_norm": 1.3927209377288818, "learning_rate": 5.0454750977050406e-06, "loss": 0.1112, "num_input_tokens_seen": 8337776, "step": 31085 }, { "epoch": 8.149410222804718, "grad_norm": 2.1844303607940674, "learning_rate": 5.038587164862435e-06, "loss": 0.1499, "num_input_tokens_seen": 8339264, "step": 31090 }, { "epoch": 8.150720838794234, "grad_norm": 4.912164211273193, "learning_rate": 5.031703409909052e-06, "loss": 0.0684, "num_input_tokens_seen": 8340592, "step": 31095 }, { "epoch": 8.152031454783748, "grad_norm": 1.0280675888061523, "learning_rate": 5.024823834285636e-06, "loss": 0.1416, "num_input_tokens_seen": 8341712, "step": 31100 }, { "epoch": 8.153342070773263, "grad_norm": 1.845108985900879, "learning_rate": 5.017948439432077e-06, "loss": 0.1238, "num_input_tokens_seen": 8343312, "step": 31105 }, { "epoch": 8.154652686762779, "grad_norm": 4.057570457458496, "learning_rate": 5.0110772267873836e-06, "loss": 0.1455, "num_input_tokens_seen": 8344384, "step": 31110 }, { "epoch": 8.155963302752294, "grad_norm": 3.176048755645752, "learning_rate": 5.004210197789688e-06, "loss": 0.1211, "num_input_tokens_seen": 8345536, "step": 31115 }, { "epoch": 8.157273918741808, "grad_norm": 5.800510406494141, "learning_rate": 4.997347353876242e-06, "loss": 0.1906, "num_input_tokens_seen": 8346592, "step": 31120 }, { "epoch": 8.158584534731324, "grad_norm": 2.350559711456299, "learning_rate": 4.990488696483439e-06, "loss": 0.1141, "num_input_tokens_seen": 8347616, "step": 31125 }, { "epoch": 8.159895150720839, "grad_norm": 2.558410882949829, "learning_rate": 4.983634227046785e-06, "loss": 0.2347, "num_input_tokens_seen": 8349168, "step": 31130 }, { "epoch": 8.161205766710355, "grad_norm": 3.724531888961792, "learning_rate": 4.9767839470008935e-06, "loss": 0.1345, "num_input_tokens_seen": 8350720, "step": 31135 }, { "epoch": 8.162516382699868, "grad_norm": 2.4305965900421143, "learning_rate": 4.969937857779528e-06, "loss": 0.1697, "num_input_tokens_seen": 8352208, "step": 31140 }, { "epoch": 8.163826998689384, "grad_norm": 3.5015430450439453, "learning_rate": 4.963095960815556e-06, "loss": 0.1472, "num_input_tokens_seen": 8353504, "step": 31145 }, { "epoch": 8.1651376146789, "grad_norm": 1.4030815362930298, "learning_rate": 4.95625825754098e-06, "loss": 0.1708, "num_input_tokens_seen": 8354560, "step": 31150 }, { "epoch": 8.166448230668415, "grad_norm": 5.451794147491455, "learning_rate": 4.9494247493869165e-06, "loss": 0.1414, "num_input_tokens_seen": 8355808, "step": 31155 }, { "epoch": 8.167758846657929, "grad_norm": 1.4228428602218628, "learning_rate": 4.942595437783609e-06, "loss": 0.1695, "num_input_tokens_seen": 8357360, "step": 31160 }, { "epoch": 8.169069462647444, "grad_norm": 1.402602195739746, "learning_rate": 4.9357703241604185e-06, "loss": 0.1385, "num_input_tokens_seen": 8358880, "step": 31165 }, { "epoch": 8.17038007863696, "grad_norm": 2.060976505279541, "learning_rate": 4.928949409945832e-06, "loss": 0.1699, "num_input_tokens_seen": 8360496, "step": 31170 }, { "epoch": 8.171690694626474, "grad_norm": 2.172844648361206, "learning_rate": 4.922132696567464e-06, "loss": 0.2173, "num_input_tokens_seen": 8361776, "step": 31175 }, { "epoch": 8.17300131061599, "grad_norm": 2.2601447105407715, "learning_rate": 4.915320185452013e-06, "loss": 0.1353, "num_input_tokens_seen": 8362928, "step": 31180 }, { "epoch": 8.174311926605505, "grad_norm": 0.7300294637680054, "learning_rate": 4.9085118780253515e-06, "loss": 0.0875, "num_input_tokens_seen": 8364176, "step": 31185 }, { "epoch": 8.17562254259502, "grad_norm": 2.4860851764678955, "learning_rate": 4.9017077757124385e-06, "loss": 0.1098, "num_input_tokens_seen": 8365376, "step": 31190 }, { "epoch": 8.176933158584534, "grad_norm": 3.1913328170776367, "learning_rate": 4.89490787993736e-06, "loss": 0.1468, "num_input_tokens_seen": 8366400, "step": 31195 }, { "epoch": 8.17824377457405, "grad_norm": 2.489621639251709, "learning_rate": 4.888112192123323e-06, "loss": 0.2877, "num_input_tokens_seen": 8367792, "step": 31200 }, { "epoch": 8.179554390563565, "grad_norm": 1.802823781967163, "learning_rate": 4.8813207136926555e-06, "loss": 0.2169, "num_input_tokens_seen": 8369232, "step": 31205 }, { "epoch": 8.18086500655308, "grad_norm": 2.5584869384765625, "learning_rate": 4.8745334460668005e-06, "loss": 0.197, "num_input_tokens_seen": 8370496, "step": 31210 }, { "epoch": 8.182175622542594, "grad_norm": 1.5136901140213013, "learning_rate": 4.867750390666326e-06, "loss": 0.1037, "num_input_tokens_seen": 8371488, "step": 31215 }, { "epoch": 8.18348623853211, "grad_norm": 3.1295762062072754, "learning_rate": 4.860971548910903e-06, "loss": 0.1411, "num_input_tokens_seen": 8372752, "step": 31220 }, { "epoch": 8.184796854521625, "grad_norm": 2.3752167224884033, "learning_rate": 4.854196922219337e-06, "loss": 0.1504, "num_input_tokens_seen": 8374160, "step": 31225 }, { "epoch": 8.186107470511141, "grad_norm": 2.9598827362060547, "learning_rate": 4.847426512009545e-06, "loss": 0.0502, "num_input_tokens_seen": 8375024, "step": 31230 }, { "epoch": 8.187418086500655, "grad_norm": 2.7239291667938232, "learning_rate": 4.840660319698559e-06, "loss": 0.261, "num_input_tokens_seen": 8376448, "step": 31235 }, { "epoch": 8.18872870249017, "grad_norm": 2.2485837936401367, "learning_rate": 4.8338983467025394e-06, "loss": 0.1473, "num_input_tokens_seen": 8378016, "step": 31240 }, { "epoch": 8.190039318479686, "grad_norm": 1.6665523052215576, "learning_rate": 4.827140594436752e-06, "loss": 0.1194, "num_input_tokens_seen": 8379136, "step": 31245 }, { "epoch": 8.191349934469201, "grad_norm": 1.3868532180786133, "learning_rate": 4.82038706431559e-06, "loss": 0.194, "num_input_tokens_seen": 8380688, "step": 31250 }, { "epoch": 8.192660550458715, "grad_norm": 2.732062339782715, "learning_rate": 4.81363775775254e-06, "loss": 0.0906, "num_input_tokens_seen": 8382016, "step": 31255 }, { "epoch": 8.19397116644823, "grad_norm": 2.832245111465454, "learning_rate": 4.80689267616023e-06, "loss": 0.0897, "num_input_tokens_seen": 8383056, "step": 31260 }, { "epoch": 8.195281782437746, "grad_norm": 1.6450941562652588, "learning_rate": 4.800151820950391e-06, "loss": 0.1827, "num_input_tokens_seen": 8384336, "step": 31265 }, { "epoch": 8.19659239842726, "grad_norm": 3.8361005783081055, "learning_rate": 4.7934151935338706e-06, "loss": 0.1608, "num_input_tokens_seen": 8386160, "step": 31270 }, { "epoch": 8.197903014416775, "grad_norm": 2.206979990005493, "learning_rate": 4.786682795320638e-06, "loss": 0.1553, "num_input_tokens_seen": 8387600, "step": 31275 }, { "epoch": 8.199213630406291, "grad_norm": 1.5817803144454956, "learning_rate": 4.779954627719771e-06, "loss": 0.1696, "num_input_tokens_seen": 8389072, "step": 31280 }, { "epoch": 8.200524246395807, "grad_norm": 1.9868839979171753, "learning_rate": 4.773230692139463e-06, "loss": 0.1912, "num_input_tokens_seen": 8390336, "step": 31285 }, { "epoch": 8.20183486238532, "grad_norm": 0.9980146884918213, "learning_rate": 4.766510989987022e-06, "loss": 0.0793, "num_input_tokens_seen": 8391680, "step": 31290 }, { "epoch": 8.203145478374836, "grad_norm": 3.342956304550171, "learning_rate": 4.7597955226688666e-06, "loss": 0.1805, "num_input_tokens_seen": 8392784, "step": 31295 }, { "epoch": 8.204456094364351, "grad_norm": 1.0932917594909668, "learning_rate": 4.75308429159054e-06, "loss": 0.1504, "num_input_tokens_seen": 8394288, "step": 31300 }, { "epoch": 8.205766710353867, "grad_norm": 1.7859035730361938, "learning_rate": 4.746377298156685e-06, "loss": 0.1255, "num_input_tokens_seen": 8395632, "step": 31305 }, { "epoch": 8.20707732634338, "grad_norm": 1.5183217525482178, "learning_rate": 4.739674543771066e-06, "loss": 0.0812, "num_input_tokens_seen": 8396816, "step": 31310 }, { "epoch": 8.208387942332896, "grad_norm": 3.794706106185913, "learning_rate": 4.732976029836561e-06, "loss": 0.1734, "num_input_tokens_seen": 8398048, "step": 31315 }, { "epoch": 8.209698558322412, "grad_norm": 4.520301342010498, "learning_rate": 4.726281757755149e-06, "loss": 0.1486, "num_input_tokens_seen": 8399408, "step": 31320 }, { "epoch": 8.211009174311927, "grad_norm": 2.45900559425354, "learning_rate": 4.719591728927938e-06, "loss": 0.1496, "num_input_tokens_seen": 8400992, "step": 31325 }, { "epoch": 8.212319790301441, "grad_norm": 0.22196795046329498, "learning_rate": 4.712905944755144e-06, "loss": 0.1623, "num_input_tokens_seen": 8402208, "step": 31330 }, { "epoch": 8.213630406290957, "grad_norm": 1.7437082529067993, "learning_rate": 4.706224406636072e-06, "loss": 0.0966, "num_input_tokens_seen": 8403520, "step": 31335 }, { "epoch": 8.214941022280472, "grad_norm": 1.721388816833496, "learning_rate": 4.699547115969172e-06, "loss": 0.1328, "num_input_tokens_seen": 8405040, "step": 31340 }, { "epoch": 8.216251638269988, "grad_norm": 2.1358633041381836, "learning_rate": 4.692874074151979e-06, "loss": 0.1847, "num_input_tokens_seen": 8406512, "step": 31345 }, { "epoch": 8.217562254259501, "grad_norm": 4.1007184982299805, "learning_rate": 4.686205282581152e-06, "loss": 0.1573, "num_input_tokens_seen": 8407584, "step": 31350 }, { "epoch": 8.218872870249017, "grad_norm": 2.5179927349090576, "learning_rate": 4.679540742652463e-06, "loss": 0.1538, "num_input_tokens_seen": 8408672, "step": 31355 }, { "epoch": 8.220183486238533, "grad_norm": 1.5490877628326416, "learning_rate": 4.672880455760786e-06, "loss": 0.0985, "num_input_tokens_seen": 8409712, "step": 31360 }, { "epoch": 8.221494102228046, "grad_norm": 3.1821653842926025, "learning_rate": 4.666224423300114e-06, "loss": 0.2616, "num_input_tokens_seen": 8411040, "step": 31365 }, { "epoch": 8.222804718217562, "grad_norm": 2.4958887100219727, "learning_rate": 4.659572646663532e-06, "loss": 0.0784, "num_input_tokens_seen": 8412448, "step": 31370 }, { "epoch": 8.224115334207077, "grad_norm": 2.9835755825042725, "learning_rate": 4.652925127243246e-06, "loss": 0.1778, "num_input_tokens_seen": 8413568, "step": 31375 }, { "epoch": 8.225425950196593, "grad_norm": 1.1628973484039307, "learning_rate": 4.646281866430574e-06, "loss": 0.3368, "num_input_tokens_seen": 8414688, "step": 31380 }, { "epoch": 8.226736566186107, "grad_norm": 2.0180652141571045, "learning_rate": 4.6396428656159424e-06, "loss": 0.2056, "num_input_tokens_seen": 8416160, "step": 31385 }, { "epoch": 8.228047182175622, "grad_norm": 1.9305076599121094, "learning_rate": 4.633008126188876e-06, "loss": 0.2016, "num_input_tokens_seen": 8417344, "step": 31390 }, { "epoch": 8.229357798165138, "grad_norm": 2.117133140563965, "learning_rate": 4.626377649538019e-06, "loss": 0.1814, "num_input_tokens_seen": 8418528, "step": 31395 }, { "epoch": 8.230668414154653, "grad_norm": 1.847413420677185, "learning_rate": 4.619751437051114e-06, "loss": 0.0944, "num_input_tokens_seen": 8419536, "step": 31400 }, { "epoch": 8.231979030144167, "grad_norm": 0.8468684554100037, "learning_rate": 4.613129490115023e-06, "loss": 0.1188, "num_input_tokens_seen": 8420800, "step": 31405 }, { "epoch": 8.233289646133683, "grad_norm": 1.126416563987732, "learning_rate": 4.6065118101157014e-06, "loss": 0.1026, "num_input_tokens_seen": 8422016, "step": 31410 }, { "epoch": 8.234600262123198, "grad_norm": 1.8972846269607544, "learning_rate": 4.599898398438221e-06, "loss": 0.1737, "num_input_tokens_seen": 8423552, "step": 31415 }, { "epoch": 8.235910878112714, "grad_norm": 1.1720812320709229, "learning_rate": 4.593289256466757e-06, "loss": 0.1903, "num_input_tokens_seen": 8425008, "step": 31420 }, { "epoch": 8.237221494102227, "grad_norm": 0.5745282769203186, "learning_rate": 4.586684385584592e-06, "loss": 0.0876, "num_input_tokens_seen": 8426256, "step": 31425 }, { "epoch": 8.238532110091743, "grad_norm": 1.3409236669540405, "learning_rate": 4.580083787174114e-06, "loss": 0.0831, "num_input_tokens_seen": 8427952, "step": 31430 }, { "epoch": 8.239842726081259, "grad_norm": 2.722959518432617, "learning_rate": 4.573487462616815e-06, "loss": 0.1704, "num_input_tokens_seen": 8429120, "step": 31435 }, { "epoch": 8.241153342070774, "grad_norm": 3.0025475025177, "learning_rate": 4.566895413293298e-06, "loss": 0.0408, "num_input_tokens_seen": 8430032, "step": 31440 }, { "epoch": 8.242463958060288, "grad_norm": 1.8471914529800415, "learning_rate": 4.560307640583264e-06, "loss": 0.1548, "num_input_tokens_seen": 8431024, "step": 31445 }, { "epoch": 8.243774574049803, "grad_norm": 4.789692401885986, "learning_rate": 4.553724145865529e-06, "loss": 0.1543, "num_input_tokens_seen": 8432160, "step": 31450 }, { "epoch": 8.245085190039319, "grad_norm": 4.922976016998291, "learning_rate": 4.547144930517996e-06, "loss": 0.1042, "num_input_tokens_seen": 8433328, "step": 31455 }, { "epoch": 8.246395806028833, "grad_norm": 2.311441659927368, "learning_rate": 4.54056999591769e-06, "loss": 0.206, "num_input_tokens_seen": 8435280, "step": 31460 }, { "epoch": 8.247706422018348, "grad_norm": 0.8724219799041748, "learning_rate": 4.533999343440728e-06, "loss": 0.1354, "num_input_tokens_seen": 8436688, "step": 31465 }, { "epoch": 8.249017038007864, "grad_norm": 5.271705627441406, "learning_rate": 4.527432974462345e-06, "loss": 0.0998, "num_input_tokens_seen": 8437680, "step": 31470 }, { "epoch": 8.25032765399738, "grad_norm": 2.600062847137451, "learning_rate": 4.520870890356868e-06, "loss": 0.2157, "num_input_tokens_seen": 8438864, "step": 31475 }, { "epoch": 8.251638269986893, "grad_norm": 2.005492925643921, "learning_rate": 4.5143130924977265e-06, "loss": 0.18, "num_input_tokens_seen": 8440272, "step": 31480 }, { "epoch": 8.252948885976409, "grad_norm": 5.501739978790283, "learning_rate": 4.5077595822574645e-06, "loss": 0.1905, "num_input_tokens_seen": 8441424, "step": 31485 }, { "epoch": 8.254259501965924, "grad_norm": 0.736234724521637, "learning_rate": 4.50121036100771e-06, "loss": 0.0991, "num_input_tokens_seen": 8442704, "step": 31490 }, { "epoch": 8.25557011795544, "grad_norm": 3.9724924564361572, "learning_rate": 4.494665430119208e-06, "loss": 0.1337, "num_input_tokens_seen": 8443744, "step": 31495 }, { "epoch": 8.256880733944953, "grad_norm": 2.0274856090545654, "learning_rate": 4.488124790961798e-06, "loss": 0.2403, "num_input_tokens_seen": 8445648, "step": 31500 }, { "epoch": 8.258191349934469, "grad_norm": 2.425563097000122, "learning_rate": 4.4815884449044275e-06, "loss": 0.1003, "num_input_tokens_seen": 8447072, "step": 31505 }, { "epoch": 8.259501965923985, "grad_norm": 2.008056163787842, "learning_rate": 4.475056393315144e-06, "loss": 0.1404, "num_input_tokens_seen": 8448496, "step": 31510 }, { "epoch": 8.2608125819135, "grad_norm": 1.3014004230499268, "learning_rate": 4.468528637561095e-06, "loss": 0.1585, "num_input_tokens_seen": 8450336, "step": 31515 }, { "epoch": 8.262123197903014, "grad_norm": 1.772028923034668, "learning_rate": 4.462005179008525e-06, "loss": 0.1902, "num_input_tokens_seen": 8452368, "step": 31520 }, { "epoch": 8.26343381389253, "grad_norm": 3.755340814590454, "learning_rate": 4.455486019022781e-06, "loss": 0.2145, "num_input_tokens_seen": 8453760, "step": 31525 }, { "epoch": 8.264744429882045, "grad_norm": 1.9160677194595337, "learning_rate": 4.448971158968318e-06, "loss": 0.1406, "num_input_tokens_seen": 8455024, "step": 31530 }, { "epoch": 8.26605504587156, "grad_norm": 1.3130797147750854, "learning_rate": 4.4424606002086814e-06, "loss": 0.2506, "num_input_tokens_seen": 8456640, "step": 31535 }, { "epoch": 8.267365661861074, "grad_norm": 1.4824415445327759, "learning_rate": 4.435954344106522e-06, "loss": 0.1753, "num_input_tokens_seen": 8457920, "step": 31540 }, { "epoch": 8.26867627785059, "grad_norm": 6.538980007171631, "learning_rate": 4.429452392023584e-06, "loss": 0.0969, "num_input_tokens_seen": 8459584, "step": 31545 }, { "epoch": 8.269986893840105, "grad_norm": 1.8694268465042114, "learning_rate": 4.4229547453207175e-06, "loss": 0.1775, "num_input_tokens_seen": 8460656, "step": 31550 }, { "epoch": 8.271297509829619, "grad_norm": 2.002279281616211, "learning_rate": 4.416461405357869e-06, "loss": 0.1707, "num_input_tokens_seen": 8462064, "step": 31555 }, { "epoch": 8.272608125819135, "grad_norm": 4.386385440826416, "learning_rate": 4.409972373494084e-06, "loss": 0.1768, "num_input_tokens_seen": 8463264, "step": 31560 }, { "epoch": 8.27391874180865, "grad_norm": 3.9640135765075684, "learning_rate": 4.403487651087509e-06, "loss": 0.2432, "num_input_tokens_seen": 8464560, "step": 31565 }, { "epoch": 8.275229357798166, "grad_norm": 0.8518090844154358, "learning_rate": 4.397007239495376e-06, "loss": 0.113, "num_input_tokens_seen": 8465744, "step": 31570 }, { "epoch": 8.27653997378768, "grad_norm": 1.894340991973877, "learning_rate": 4.390531140074028e-06, "loss": 0.1573, "num_input_tokens_seen": 8467376, "step": 31575 }, { "epoch": 8.277850589777195, "grad_norm": 3.240571975708008, "learning_rate": 4.384059354178893e-06, "loss": 0.1509, "num_input_tokens_seen": 8468432, "step": 31580 }, { "epoch": 8.27916120576671, "grad_norm": 2.1804678440093994, "learning_rate": 4.377591883164522e-06, "loss": 0.1246, "num_input_tokens_seen": 8469680, "step": 31585 }, { "epoch": 8.280471821756226, "grad_norm": 3.271198272705078, "learning_rate": 4.371128728384538e-06, "loss": 0.2396, "num_input_tokens_seen": 8471008, "step": 31590 }, { "epoch": 8.28178243774574, "grad_norm": 1.5681599378585815, "learning_rate": 4.364669891191667e-06, "loss": 0.1392, "num_input_tokens_seen": 8472400, "step": 31595 }, { "epoch": 8.283093053735255, "grad_norm": 1.9818757772445679, "learning_rate": 4.358215372937738e-06, "loss": 0.1142, "num_input_tokens_seen": 8474368, "step": 31600 }, { "epoch": 8.284403669724771, "grad_norm": 1.354896068572998, "learning_rate": 4.35176517497366e-06, "loss": 0.0853, "num_input_tokens_seen": 8475424, "step": 31605 }, { "epoch": 8.285714285714286, "grad_norm": 2.673637628555298, "learning_rate": 4.34531929864945e-06, "loss": 0.2617, "num_input_tokens_seen": 8477120, "step": 31610 }, { "epoch": 8.2870249017038, "grad_norm": 2.5588395595550537, "learning_rate": 4.3388777453142265e-06, "loss": 0.1942, "num_input_tokens_seen": 8478528, "step": 31615 }, { "epoch": 8.288335517693316, "grad_norm": 1.2733838558197021, "learning_rate": 4.332440516316188e-06, "loss": 0.1098, "num_input_tokens_seen": 8480032, "step": 31620 }, { "epoch": 8.289646133682831, "grad_norm": 3.0522730350494385, "learning_rate": 4.326007613002642e-06, "loss": 0.1647, "num_input_tokens_seen": 8481280, "step": 31625 }, { "epoch": 8.290956749672347, "grad_norm": 3.168649196624756, "learning_rate": 4.319579036719979e-06, "loss": 0.1876, "num_input_tokens_seen": 8483040, "step": 31630 }, { "epoch": 8.29226736566186, "grad_norm": 1.8875434398651123, "learning_rate": 4.313154788813684e-06, "loss": 0.1577, "num_input_tokens_seen": 8484016, "step": 31635 }, { "epoch": 8.293577981651376, "grad_norm": 3.0666463375091553, "learning_rate": 4.30673487062836e-06, "loss": 0.1815, "num_input_tokens_seen": 8484960, "step": 31640 }, { "epoch": 8.294888597640892, "grad_norm": 3.0789499282836914, "learning_rate": 4.300319283507667e-06, "loss": 0.1609, "num_input_tokens_seen": 8486112, "step": 31645 }, { "epoch": 8.296199213630405, "grad_norm": 3.9198899269104004, "learning_rate": 4.29390802879438e-06, "loss": 0.1508, "num_input_tokens_seen": 8487072, "step": 31650 }, { "epoch": 8.297509829619921, "grad_norm": 1.8161669969558716, "learning_rate": 4.287501107830366e-06, "loss": 0.2161, "num_input_tokens_seen": 8488592, "step": 31655 }, { "epoch": 8.298820445609437, "grad_norm": 2.5465846061706543, "learning_rate": 4.2810985219565835e-06, "loss": 0.2042, "num_input_tokens_seen": 8490064, "step": 31660 }, { "epoch": 8.300131061598952, "grad_norm": 5.683577537536621, "learning_rate": 4.27470027251308e-06, "loss": 0.1326, "num_input_tokens_seen": 8491648, "step": 31665 }, { "epoch": 8.301441677588466, "grad_norm": 0.7285233736038208, "learning_rate": 4.268306360838998e-06, "loss": 0.1629, "num_input_tokens_seen": 8492672, "step": 31670 }, { "epoch": 8.302752293577981, "grad_norm": 3.030991315841675, "learning_rate": 4.261916788272571e-06, "loss": 0.3437, "num_input_tokens_seen": 8494080, "step": 31675 }, { "epoch": 8.304062909567497, "grad_norm": 2.300032377243042, "learning_rate": 4.2555315561511326e-06, "loss": 0.221, "num_input_tokens_seen": 8495568, "step": 31680 }, { "epoch": 8.305373525557012, "grad_norm": 0.8721568584442139, "learning_rate": 4.249150665811097e-06, "loss": 0.0972, "num_input_tokens_seen": 8496768, "step": 31685 }, { "epoch": 8.306684141546526, "grad_norm": 4.04206657409668, "learning_rate": 4.242774118587969e-06, "loss": 0.1291, "num_input_tokens_seen": 8498384, "step": 31690 }, { "epoch": 8.307994757536042, "grad_norm": 1.5122958421707153, "learning_rate": 4.236401915816346e-06, "loss": 0.2561, "num_input_tokens_seen": 8499872, "step": 31695 }, { "epoch": 8.309305373525557, "grad_norm": 2.4223930835723877, "learning_rate": 4.230034058829932e-06, "loss": 0.0785, "num_input_tokens_seen": 8500944, "step": 31700 }, { "epoch": 8.310615989515073, "grad_norm": 3.0036087036132812, "learning_rate": 4.2236705489615e-06, "loss": 0.1187, "num_input_tokens_seen": 8502224, "step": 31705 }, { "epoch": 8.311926605504587, "grad_norm": 2.924406051635742, "learning_rate": 4.2173113875429246e-06, "loss": 0.0898, "num_input_tokens_seen": 8503312, "step": 31710 }, { "epoch": 8.313237221494102, "grad_norm": 2.0523476600646973, "learning_rate": 4.2109565759051645e-06, "loss": 0.1378, "num_input_tokens_seen": 8504752, "step": 31715 }, { "epoch": 8.314547837483618, "grad_norm": 2.740370988845825, "learning_rate": 4.204606115378282e-06, "loss": 0.1512, "num_input_tokens_seen": 8506160, "step": 31720 }, { "epoch": 8.315858453473133, "grad_norm": 5.084738731384277, "learning_rate": 4.198260007291399e-06, "loss": 0.3664, "num_input_tokens_seen": 8507616, "step": 31725 }, { "epoch": 8.317169069462647, "grad_norm": 1.1096748113632202, "learning_rate": 4.191918252972754e-06, "loss": 0.0975, "num_input_tokens_seen": 8508784, "step": 31730 }, { "epoch": 8.318479685452163, "grad_norm": 1.1122493743896484, "learning_rate": 4.185580853749666e-06, "loss": 0.1464, "num_input_tokens_seen": 8510080, "step": 31735 }, { "epoch": 8.319790301441678, "grad_norm": 1.6478314399719238, "learning_rate": 4.179247810948539e-06, "loss": 0.1281, "num_input_tokens_seen": 8511488, "step": 31740 }, { "epoch": 8.321100917431192, "grad_norm": 2.3799242973327637, "learning_rate": 4.172919125894869e-06, "loss": 0.2368, "num_input_tokens_seen": 8512944, "step": 31745 }, { "epoch": 8.322411533420707, "grad_norm": 4.592325210571289, "learning_rate": 4.166594799913237e-06, "loss": 0.1268, "num_input_tokens_seen": 8514128, "step": 31750 }, { "epoch": 8.323722149410223, "grad_norm": 0.768199622631073, "learning_rate": 4.160274834327327e-06, "loss": 0.1375, "num_input_tokens_seen": 8515760, "step": 31755 }, { "epoch": 8.325032765399738, "grad_norm": 2.1917781829833984, "learning_rate": 4.153959230459878e-06, "loss": 0.2065, "num_input_tokens_seen": 8517104, "step": 31760 }, { "epoch": 8.326343381389252, "grad_norm": 2.9709179401397705, "learning_rate": 4.147647989632744e-06, "loss": 0.1783, "num_input_tokens_seen": 8518256, "step": 31765 }, { "epoch": 8.327653997378768, "grad_norm": 1.7295552492141724, "learning_rate": 4.141341113166855e-06, "loss": 0.2465, "num_input_tokens_seen": 8519600, "step": 31770 }, { "epoch": 8.328964613368283, "grad_norm": 1.2724056243896484, "learning_rate": 4.135038602382232e-06, "loss": 0.1463, "num_input_tokens_seen": 8520944, "step": 31775 }, { "epoch": 8.330275229357799, "grad_norm": 1.824720859527588, "learning_rate": 4.1287404585979764e-06, "loss": 0.1057, "num_input_tokens_seen": 8522160, "step": 31780 }, { "epoch": 8.331585845347313, "grad_norm": 3.7952351570129395, "learning_rate": 4.122446683132281e-06, "loss": 0.1734, "num_input_tokens_seen": 8523808, "step": 31785 }, { "epoch": 8.332896461336828, "grad_norm": 2.3570501804351807, "learning_rate": 4.116157277302418e-06, "loss": 0.4625, "num_input_tokens_seen": 8525616, "step": 31790 }, { "epoch": 8.334207077326344, "grad_norm": 1.4295494556427002, "learning_rate": 4.109872242424756e-06, "loss": 0.181, "num_input_tokens_seen": 8527200, "step": 31795 }, { "epoch": 8.33551769331586, "grad_norm": 0.7609356641769409, "learning_rate": 4.1035915798147415e-06, "loss": 0.1953, "num_input_tokens_seen": 8528720, "step": 31800 }, { "epoch": 8.336828309305373, "grad_norm": 2.2350425720214844, "learning_rate": 4.097315290786899e-06, "loss": 0.1241, "num_input_tokens_seen": 8529888, "step": 31805 }, { "epoch": 8.338138925294889, "grad_norm": 3.4192943572998047, "learning_rate": 4.091043376654843e-06, "loss": 0.1166, "num_input_tokens_seen": 8531360, "step": 31810 }, { "epoch": 8.339449541284404, "grad_norm": 1.2067307233810425, "learning_rate": 4.084775838731286e-06, "loss": 0.0758, "num_input_tokens_seen": 8532656, "step": 31815 }, { "epoch": 8.34076015727392, "grad_norm": 0.16964922845363617, "learning_rate": 4.078512678328009e-06, "loss": 0.1562, "num_input_tokens_seen": 8533712, "step": 31820 }, { "epoch": 8.342070773263433, "grad_norm": 3.9337000846862793, "learning_rate": 4.072253896755876e-06, "loss": 0.116, "num_input_tokens_seen": 8535008, "step": 31825 }, { "epoch": 8.343381389252949, "grad_norm": 2.1106936931610107, "learning_rate": 4.065999495324843e-06, "loss": 0.2383, "num_input_tokens_seen": 8536256, "step": 31830 }, { "epoch": 8.344692005242464, "grad_norm": 1.506483554840088, "learning_rate": 4.059749475343949e-06, "loss": 0.1285, "num_input_tokens_seen": 8537680, "step": 31835 }, { "epoch": 8.346002621231978, "grad_norm": 1.3120815753936768, "learning_rate": 4.053503838121306e-06, "loss": 0.1286, "num_input_tokens_seen": 8539072, "step": 31840 }, { "epoch": 8.347313237221494, "grad_norm": 2.1895041465759277, "learning_rate": 4.047262584964112e-06, "loss": 0.2023, "num_input_tokens_seen": 8540576, "step": 31845 }, { "epoch": 8.34862385321101, "grad_norm": 1.8390376567840576, "learning_rate": 4.0410257171786595e-06, "loss": 0.2138, "num_input_tokens_seen": 8542176, "step": 31850 }, { "epoch": 8.349934469200525, "grad_norm": 2.2440831661224365, "learning_rate": 4.034793236070308e-06, "loss": 0.2396, "num_input_tokens_seen": 8543600, "step": 31855 }, { "epoch": 8.351245085190039, "grad_norm": 1.30652916431427, "learning_rate": 4.028565142943508e-06, "loss": 0.1505, "num_input_tokens_seen": 8545072, "step": 31860 }, { "epoch": 8.352555701179554, "grad_norm": 1.7601975202560425, "learning_rate": 4.022341439101782e-06, "loss": 0.1033, "num_input_tokens_seen": 8545952, "step": 31865 }, { "epoch": 8.35386631716907, "grad_norm": 1.3176350593566895, "learning_rate": 4.016122125847751e-06, "loss": 0.1265, "num_input_tokens_seen": 8547168, "step": 31870 }, { "epoch": 8.355176933158585, "grad_norm": 0.9615864157676697, "learning_rate": 4.009907204483107e-06, "loss": 0.1137, "num_input_tokens_seen": 8548480, "step": 31875 }, { "epoch": 8.356487549148099, "grad_norm": 1.4435642957687378, "learning_rate": 4.003696676308613e-06, "loss": 0.1482, "num_input_tokens_seen": 8549904, "step": 31880 }, { "epoch": 8.357798165137615, "grad_norm": 3.1327528953552246, "learning_rate": 3.997490542624124e-06, "loss": 0.0936, "num_input_tokens_seen": 8550896, "step": 31885 }, { "epoch": 8.35910878112713, "grad_norm": 1.225024700164795, "learning_rate": 3.991288804728577e-06, "loss": 0.1611, "num_input_tokens_seen": 8552096, "step": 31890 }, { "epoch": 8.360419397116646, "grad_norm": 2.8014001846313477, "learning_rate": 3.985091463919982e-06, "loss": 0.1769, "num_input_tokens_seen": 8553680, "step": 31895 }, { "epoch": 8.36173001310616, "grad_norm": 2.0747172832489014, "learning_rate": 3.978898521495433e-06, "loss": 0.2379, "num_input_tokens_seen": 8556080, "step": 31900 }, { "epoch": 8.363040629095675, "grad_norm": 1.571591854095459, "learning_rate": 3.972709978751099e-06, "loss": 0.1595, "num_input_tokens_seen": 8557280, "step": 31905 }, { "epoch": 8.36435124508519, "grad_norm": 0.4797604978084564, "learning_rate": 3.966525836982237e-06, "loss": 0.206, "num_input_tokens_seen": 8558496, "step": 31910 }, { "epoch": 8.365661861074706, "grad_norm": 2.4699833393096924, "learning_rate": 3.960346097483175e-06, "loss": 0.1829, "num_input_tokens_seen": 8560000, "step": 31915 }, { "epoch": 8.36697247706422, "grad_norm": 1.6365010738372803, "learning_rate": 3.954170761547324e-06, "loss": 0.1046, "num_input_tokens_seen": 8561120, "step": 31920 }, { "epoch": 8.368283093053735, "grad_norm": 2.9906349182128906, "learning_rate": 3.947999830467159e-06, "loss": 0.1494, "num_input_tokens_seen": 8562048, "step": 31925 }, { "epoch": 8.36959370904325, "grad_norm": 3.5076727867126465, "learning_rate": 3.941833305534259e-06, "loss": 0.0888, "num_input_tokens_seen": 8562992, "step": 31930 }, { "epoch": 8.370904325032765, "grad_norm": 1.76182222366333, "learning_rate": 3.935671188039264e-06, "loss": 0.1643, "num_input_tokens_seen": 8564816, "step": 31935 }, { "epoch": 8.37221494102228, "grad_norm": 1.2688654661178589, "learning_rate": 3.9295134792718905e-06, "loss": 0.1783, "num_input_tokens_seen": 8566464, "step": 31940 }, { "epoch": 8.373525557011796, "grad_norm": 2.9223783016204834, "learning_rate": 3.923360180520941e-06, "loss": 0.1587, "num_input_tokens_seen": 8567744, "step": 31945 }, { "epoch": 8.374836173001311, "grad_norm": 1.8914504051208496, "learning_rate": 3.917211293074285e-06, "loss": 0.3205, "num_input_tokens_seen": 8569280, "step": 31950 }, { "epoch": 8.376146788990825, "grad_norm": 3.534674644470215, "learning_rate": 3.911066818218886e-06, "loss": 0.2084, "num_input_tokens_seen": 8570256, "step": 31955 }, { "epoch": 8.37745740498034, "grad_norm": 0.20867833495140076, "learning_rate": 3.9049267572407565e-06, "loss": 0.1865, "num_input_tokens_seen": 8572704, "step": 31960 }, { "epoch": 8.378768020969856, "grad_norm": 1.8273518085479736, "learning_rate": 3.898791111425007e-06, "loss": 0.1527, "num_input_tokens_seen": 8574016, "step": 31965 }, { "epoch": 8.380078636959372, "grad_norm": 1.4930132627487183, "learning_rate": 3.892659882055819e-06, "loss": 0.1975, "num_input_tokens_seen": 8575696, "step": 31970 }, { "epoch": 8.381389252948885, "grad_norm": 0.5152679681777954, "learning_rate": 3.886533070416445e-06, "loss": 0.1775, "num_input_tokens_seen": 8577248, "step": 31975 }, { "epoch": 8.382699868938401, "grad_norm": 1.5007617473602295, "learning_rate": 3.880410677789214e-06, "loss": 0.1455, "num_input_tokens_seen": 8578192, "step": 31980 }, { "epoch": 8.384010484927916, "grad_norm": 1.391951322555542, "learning_rate": 3.87429270545554e-06, "loss": 0.0697, "num_input_tokens_seen": 8579488, "step": 31985 }, { "epoch": 8.385321100917432, "grad_norm": 1.4000465869903564, "learning_rate": 3.868179154695909e-06, "loss": 0.0805, "num_input_tokens_seen": 8580432, "step": 31990 }, { "epoch": 8.386631716906946, "grad_norm": 1.6005799770355225, "learning_rate": 3.86207002678986e-06, "loss": 0.2514, "num_input_tokens_seen": 8582064, "step": 31995 }, { "epoch": 8.387942332896461, "grad_norm": 4.3821916580200195, "learning_rate": 3.855965323016031e-06, "loss": 0.2064, "num_input_tokens_seen": 8583424, "step": 32000 }, { "epoch": 8.389252948885977, "grad_norm": 3.9475901126861572, "learning_rate": 3.8498650446521264e-06, "loss": 0.2078, "num_input_tokens_seen": 8584992, "step": 32005 }, { "epoch": 8.390563564875492, "grad_norm": 1.6456671953201294, "learning_rate": 3.843769192974925e-06, "loss": 0.116, "num_input_tokens_seen": 8586080, "step": 32010 }, { "epoch": 8.391874180865006, "grad_norm": 0.833198070526123, "learning_rate": 3.837677769260273e-06, "loss": 0.1017, "num_input_tokens_seen": 8587424, "step": 32015 }, { "epoch": 8.393184796854522, "grad_norm": 2.677875280380249, "learning_rate": 3.8315907747831e-06, "loss": 0.0929, "num_input_tokens_seen": 8588720, "step": 32020 }, { "epoch": 8.394495412844037, "grad_norm": 4.643295764923096, "learning_rate": 3.825508210817402e-06, "loss": 0.1565, "num_input_tokens_seen": 8590144, "step": 32025 }, { "epoch": 8.395806028833551, "grad_norm": 2.493925094604492, "learning_rate": 3.8194300786362465e-06, "loss": 0.1308, "num_input_tokens_seen": 8591520, "step": 32030 }, { "epoch": 8.397116644823067, "grad_norm": 1.5131068229675293, "learning_rate": 3.8133563795117816e-06, "loss": 0.1685, "num_input_tokens_seen": 8592816, "step": 32035 }, { "epoch": 8.398427260812582, "grad_norm": 3.180326461791992, "learning_rate": 3.807287114715216e-06, "loss": 0.1645, "num_input_tokens_seen": 8593824, "step": 32040 }, { "epoch": 8.399737876802098, "grad_norm": 4.09185791015625, "learning_rate": 3.801222285516842e-06, "loss": 0.2185, "num_input_tokens_seen": 8595040, "step": 32045 }, { "epoch": 8.401048492791611, "grad_norm": 5.682479381561279, "learning_rate": 3.7951618931860112e-06, "loss": 0.1715, "num_input_tokens_seen": 8596240, "step": 32050 }, { "epoch": 8.402359108781127, "grad_norm": 1.8056479692459106, "learning_rate": 3.789105938991161e-06, "loss": 0.1847, "num_input_tokens_seen": 8597424, "step": 32055 }, { "epoch": 8.403669724770642, "grad_norm": 0.29889363050460815, "learning_rate": 3.7830544241997864e-06, "loss": 0.1153, "num_input_tokens_seen": 8598512, "step": 32060 }, { "epoch": 8.404980340760158, "grad_norm": 3.7288358211517334, "learning_rate": 3.7770073500784596e-06, "loss": 0.119, "num_input_tokens_seen": 8599568, "step": 32065 }, { "epoch": 8.406290956749672, "grad_norm": 2.924590826034546, "learning_rate": 3.7709647178928246e-06, "loss": 0.1293, "num_input_tokens_seen": 8600752, "step": 32070 }, { "epoch": 8.407601572739187, "grad_norm": 2.131976842880249, "learning_rate": 3.7649265289076003e-06, "loss": 0.0827, "num_input_tokens_seen": 8601632, "step": 32075 }, { "epoch": 8.408912188728703, "grad_norm": 0.6439327597618103, "learning_rate": 3.7588927843865556e-06, "loss": 0.2435, "num_input_tokens_seen": 8603216, "step": 32080 }, { "epoch": 8.410222804718218, "grad_norm": 3.738055944442749, "learning_rate": 3.7528634855925475e-06, "loss": 0.1845, "num_input_tokens_seen": 8604672, "step": 32085 }, { "epoch": 8.411533420707732, "grad_norm": 2.645174980163574, "learning_rate": 3.7468386337875028e-06, "loss": 0.0866, "num_input_tokens_seen": 8605856, "step": 32090 }, { "epoch": 8.412844036697248, "grad_norm": 2.918095827102661, "learning_rate": 3.7408182302324014e-06, "loss": 0.1453, "num_input_tokens_seen": 8606912, "step": 32095 }, { "epoch": 8.414154652686763, "grad_norm": 0.6752572655677795, "learning_rate": 3.734802276187319e-06, "loss": 0.2066, "num_input_tokens_seen": 8608640, "step": 32100 }, { "epoch": 8.415465268676279, "grad_norm": 1.7469940185546875, "learning_rate": 3.7287907729113754e-06, "loss": 0.134, "num_input_tokens_seen": 8609920, "step": 32105 }, { "epoch": 8.416775884665793, "grad_norm": 2.694465398788452, "learning_rate": 3.722783721662773e-06, "loss": 0.1552, "num_input_tokens_seen": 8611328, "step": 32110 }, { "epoch": 8.418086500655308, "grad_norm": 0.7006999254226685, "learning_rate": 3.7167811236987704e-06, "loss": 0.098, "num_input_tokens_seen": 8612608, "step": 32115 }, { "epoch": 8.419397116644824, "grad_norm": 1.363386631011963, "learning_rate": 3.710782980275704e-06, "loss": 0.1691, "num_input_tokens_seen": 8614016, "step": 32120 }, { "epoch": 8.420707732634337, "grad_norm": 2.6073107719421387, "learning_rate": 3.704789292648975e-06, "loss": 0.2358, "num_input_tokens_seen": 8615616, "step": 32125 }, { "epoch": 8.422018348623853, "grad_norm": 4.458653926849365, "learning_rate": 3.6988000620730516e-06, "loss": 0.0962, "num_input_tokens_seen": 8616832, "step": 32130 }, { "epoch": 8.423328964613368, "grad_norm": 1.9747713804244995, "learning_rate": 3.6928152898014684e-06, "loss": 0.1602, "num_input_tokens_seen": 8618560, "step": 32135 }, { "epoch": 8.424639580602884, "grad_norm": 5.556384563446045, "learning_rate": 3.686834977086831e-06, "loss": 0.1698, "num_input_tokens_seen": 8619776, "step": 32140 }, { "epoch": 8.425950196592398, "grad_norm": 1.710625171661377, "learning_rate": 3.680859125180805e-06, "loss": 0.1564, "num_input_tokens_seen": 8621008, "step": 32145 }, { "epoch": 8.427260812581913, "grad_norm": 1.4220054149627686, "learning_rate": 3.674887735334126e-06, "loss": 0.2266, "num_input_tokens_seen": 8622928, "step": 32150 }, { "epoch": 8.428571428571429, "grad_norm": 1.3852684497833252, "learning_rate": 3.6689208087966e-06, "loss": 0.1436, "num_input_tokens_seen": 8624320, "step": 32155 }, { "epoch": 8.429882044560944, "grad_norm": 1.4911799430847168, "learning_rate": 3.6629583468170914e-06, "loss": 0.2022, "num_input_tokens_seen": 8625648, "step": 32160 }, { "epoch": 8.431192660550458, "grad_norm": 2.310631275177002, "learning_rate": 3.65700035064353e-06, "loss": 0.1165, "num_input_tokens_seen": 8627152, "step": 32165 }, { "epoch": 8.432503276539974, "grad_norm": 3.536576271057129, "learning_rate": 3.6510468215229215e-06, "loss": 0.1118, "num_input_tokens_seen": 8628464, "step": 32170 }, { "epoch": 8.43381389252949, "grad_norm": 2.440371036529541, "learning_rate": 3.6450977607013217e-06, "loss": 0.0751, "num_input_tokens_seen": 8629408, "step": 32175 }, { "epoch": 8.435124508519005, "grad_norm": 1.4068293571472168, "learning_rate": 3.6391531694238647e-06, "loss": 0.1817, "num_input_tokens_seen": 8630656, "step": 32180 }, { "epoch": 8.436435124508519, "grad_norm": 2.126176357269287, "learning_rate": 3.6332130489347387e-06, "loss": 0.1237, "num_input_tokens_seen": 8631936, "step": 32185 }, { "epoch": 8.437745740498034, "grad_norm": 1.7091032266616821, "learning_rate": 3.627277400477211e-06, "loss": 0.2248, "num_input_tokens_seen": 8633584, "step": 32190 }, { "epoch": 8.43905635648755, "grad_norm": 1.0085374116897583, "learning_rate": 3.6213462252935844e-06, "loss": 0.11, "num_input_tokens_seen": 8635584, "step": 32195 }, { "epoch": 8.440366972477065, "grad_norm": 2.5083539485931396, "learning_rate": 3.6154195246252575e-06, "loss": 0.2602, "num_input_tokens_seen": 8636912, "step": 32200 }, { "epoch": 8.441677588466579, "grad_norm": 2.651350736618042, "learning_rate": 3.609497299712672e-06, "loss": 0.1935, "num_input_tokens_seen": 8638112, "step": 32205 }, { "epoch": 8.442988204456094, "grad_norm": 1.493116021156311, "learning_rate": 3.6035795517953357e-06, "loss": 0.1508, "num_input_tokens_seen": 8639568, "step": 32210 }, { "epoch": 8.44429882044561, "grad_norm": 2.601992130279541, "learning_rate": 3.597666282111836e-06, "loss": 0.1197, "num_input_tokens_seen": 8640832, "step": 32215 }, { "epoch": 8.445609436435124, "grad_norm": 3.642317295074463, "learning_rate": 3.5917574918998024e-06, "loss": 0.1654, "num_input_tokens_seen": 8642192, "step": 32220 }, { "epoch": 8.44692005242464, "grad_norm": 2.684211015701294, "learning_rate": 3.58585318239594e-06, "loss": 0.1354, "num_input_tokens_seen": 8643328, "step": 32225 }, { "epoch": 8.448230668414155, "grad_norm": 0.0928802564740181, "learning_rate": 3.579953354836002e-06, "loss": 0.1592, "num_input_tokens_seen": 8644400, "step": 32230 }, { "epoch": 8.44954128440367, "grad_norm": 1.7544782161712646, "learning_rate": 3.5740580104548145e-06, "loss": 0.11, "num_input_tokens_seen": 8645920, "step": 32235 }, { "epoch": 8.450851900393184, "grad_norm": 3.2696590423583984, "learning_rate": 3.568167150486265e-06, "loss": 0.1729, "num_input_tokens_seen": 8647152, "step": 32240 }, { "epoch": 8.4521625163827, "grad_norm": 0.7049514055252075, "learning_rate": 3.5622807761632985e-06, "loss": 0.0616, "num_input_tokens_seen": 8648320, "step": 32245 }, { "epoch": 8.453473132372215, "grad_norm": 1.6731959581375122, "learning_rate": 3.5563988887179266e-06, "loss": 0.0597, "num_input_tokens_seen": 8649376, "step": 32250 }, { "epoch": 8.45478374836173, "grad_norm": 2.5219759941101074, "learning_rate": 3.5505214893812115e-06, "loss": 0.1387, "num_input_tokens_seen": 8651136, "step": 32255 }, { "epoch": 8.456094364351245, "grad_norm": 2.152275800704956, "learning_rate": 3.5446485793832918e-06, "loss": 0.1442, "num_input_tokens_seen": 8652704, "step": 32260 }, { "epoch": 8.45740498034076, "grad_norm": 1.6359614133834839, "learning_rate": 3.5387801599533475e-06, "loss": 0.2039, "num_input_tokens_seen": 8654976, "step": 32265 }, { "epoch": 8.458715596330276, "grad_norm": 2.8269002437591553, "learning_rate": 3.5329162323196356e-06, "loss": 0.3272, "num_input_tokens_seen": 8656640, "step": 32270 }, { "epoch": 8.460026212319791, "grad_norm": 3.1835763454437256, "learning_rate": 3.527056797709466e-06, "loss": 0.2214, "num_input_tokens_seen": 8657968, "step": 32275 }, { "epoch": 8.461336828309305, "grad_norm": 1.925493836402893, "learning_rate": 3.5212018573492047e-06, "loss": 0.2049, "num_input_tokens_seen": 8660048, "step": 32280 }, { "epoch": 8.46264744429882, "grad_norm": 0.5592754483222961, "learning_rate": 3.5153514124642822e-06, "loss": 0.0792, "num_input_tokens_seen": 8661280, "step": 32285 }, { "epoch": 8.463958060288336, "grad_norm": 3.818089008331299, "learning_rate": 3.5095054642791862e-06, "loss": 0.2166, "num_input_tokens_seen": 8663200, "step": 32290 }, { "epoch": 8.46526867627785, "grad_norm": 2.3887617588043213, "learning_rate": 3.503664014017463e-06, "loss": 0.1047, "num_input_tokens_seen": 8664768, "step": 32295 }, { "epoch": 8.466579292267365, "grad_norm": 2.704501152038574, "learning_rate": 3.4978270629017207e-06, "loss": 0.1775, "num_input_tokens_seen": 8665824, "step": 32300 }, { "epoch": 8.46788990825688, "grad_norm": 3.9108684062957764, "learning_rate": 3.491994612153618e-06, "loss": 0.1868, "num_input_tokens_seen": 8667200, "step": 32305 }, { "epoch": 8.469200524246396, "grad_norm": 2.366373062133789, "learning_rate": 3.4861666629938896e-06, "loss": 0.1313, "num_input_tokens_seen": 8668640, "step": 32310 }, { "epoch": 8.47051114023591, "grad_norm": 3.8481667041778564, "learning_rate": 3.4803432166422983e-06, "loss": 0.1158, "num_input_tokens_seen": 8669696, "step": 32315 }, { "epoch": 8.471821756225426, "grad_norm": 3.716684341430664, "learning_rate": 3.474524274317689e-06, "loss": 0.1713, "num_input_tokens_seen": 8670848, "step": 32320 }, { "epoch": 8.473132372214941, "grad_norm": 3.0893921852111816, "learning_rate": 3.468709837237949e-06, "loss": 0.1415, "num_input_tokens_seen": 8671968, "step": 32325 }, { "epoch": 8.474442988204457, "grad_norm": 3.0528862476348877, "learning_rate": 3.4628999066200427e-06, "loss": 0.1746, "num_input_tokens_seen": 8672768, "step": 32330 }, { "epoch": 8.47575360419397, "grad_norm": 1.053995132446289, "learning_rate": 3.4570944836799734e-06, "loss": 0.0925, "num_input_tokens_seen": 8674176, "step": 32335 }, { "epoch": 8.477064220183486, "grad_norm": 2.157010078430176, "learning_rate": 3.4512935696328045e-06, "loss": 0.1788, "num_input_tokens_seen": 8675696, "step": 32340 }, { "epoch": 8.478374836173002, "grad_norm": 1.559977412223816, "learning_rate": 3.4454971656926655e-06, "loss": 0.1365, "num_input_tokens_seen": 8677280, "step": 32345 }, { "epoch": 8.479685452162517, "grad_norm": 1.437272548675537, "learning_rate": 3.4397052730727214e-06, "loss": 0.1117, "num_input_tokens_seen": 8678400, "step": 32350 }, { "epoch": 8.48099606815203, "grad_norm": 3.8195137977600098, "learning_rate": 3.4339178929852085e-06, "loss": 0.1695, "num_input_tokens_seen": 8679696, "step": 32355 }, { "epoch": 8.482306684141546, "grad_norm": 2.4894001483917236, "learning_rate": 3.4281350266414158e-06, "loss": 0.1876, "num_input_tokens_seen": 8680720, "step": 32360 }, { "epoch": 8.483617300131062, "grad_norm": 2.4528534412384033, "learning_rate": 3.422356675251692e-06, "loss": 0.1651, "num_input_tokens_seen": 8682160, "step": 32365 }, { "epoch": 8.484927916120578, "grad_norm": 4.329845428466797, "learning_rate": 3.4165828400254314e-06, "loss": 0.179, "num_input_tokens_seen": 8683728, "step": 32370 }, { "epoch": 8.486238532110091, "grad_norm": 6.925877094268799, "learning_rate": 3.4108135221710885e-06, "loss": 0.1244, "num_input_tokens_seen": 8684704, "step": 32375 }, { "epoch": 8.487549148099607, "grad_norm": 1.2462657690048218, "learning_rate": 3.405048722896173e-06, "loss": 0.1629, "num_input_tokens_seen": 8686288, "step": 32380 }, { "epoch": 8.488859764089122, "grad_norm": 2.1521213054656982, "learning_rate": 3.3992884434072474e-06, "loss": 0.1607, "num_input_tokens_seen": 8687408, "step": 32385 }, { "epoch": 8.490170380078636, "grad_norm": 8.843960762023926, "learning_rate": 3.3935326849099253e-06, "loss": 0.1268, "num_input_tokens_seen": 8688592, "step": 32390 }, { "epoch": 8.491480996068152, "grad_norm": 2.0603604316711426, "learning_rate": 3.3877814486088816e-06, "loss": 0.1484, "num_input_tokens_seen": 8689792, "step": 32395 }, { "epoch": 8.492791612057667, "grad_norm": 1.985763669013977, "learning_rate": 3.382034735707834e-06, "loss": 0.124, "num_input_tokens_seen": 8690896, "step": 32400 }, { "epoch": 8.494102228047183, "grad_norm": 1.560104250907898, "learning_rate": 3.376292547409568e-06, "loss": 0.1029, "num_input_tokens_seen": 8691968, "step": 32405 }, { "epoch": 8.495412844036696, "grad_norm": 1.5956625938415527, "learning_rate": 3.3705548849159054e-06, "loss": 0.2085, "num_input_tokens_seen": 8693648, "step": 32410 }, { "epoch": 8.496723460026212, "grad_norm": 0.9066974520683289, "learning_rate": 3.364821749427732e-06, "loss": 0.1569, "num_input_tokens_seen": 8695056, "step": 32415 }, { "epoch": 8.498034076015728, "grad_norm": 1.0181338787078857, "learning_rate": 3.3590931421449867e-06, "loss": 0.0791, "num_input_tokens_seen": 8696368, "step": 32420 }, { "epoch": 8.499344692005243, "grad_norm": 2.331864595413208, "learning_rate": 3.3533690642666595e-06, "loss": 0.0686, "num_input_tokens_seen": 8697472, "step": 32425 }, { "epoch": 8.500655307994757, "grad_norm": 3.0621864795684814, "learning_rate": 3.347649516990778e-06, "loss": 0.1706, "num_input_tokens_seen": 8698880, "step": 32430 }, { "epoch": 8.501965923984272, "grad_norm": 1.4552348852157593, "learning_rate": 3.3419345015144406e-06, "loss": 0.0856, "num_input_tokens_seen": 8700096, "step": 32435 }, { "epoch": 8.502228047182175, "eval_loss": 0.8556064963340759, "eval_runtime": 16.6953, "eval_samples_per_second": 50.793, "eval_steps_per_second": 25.396, "num_input_tokens_seen": 8700416, "step": 32436 }, { "epoch": 8.503276539973788, "grad_norm": 3.864074945449829, "learning_rate": 3.3362240190337817e-06, "loss": 0.3147, "num_input_tokens_seen": 8701568, "step": 32440 }, { "epoch": 8.504587155963304, "grad_norm": 1.7225043773651123, "learning_rate": 3.3305180707440124e-06, "loss": 0.1517, "num_input_tokens_seen": 8702976, "step": 32445 }, { "epoch": 8.505897771952817, "grad_norm": 4.709097385406494, "learning_rate": 3.3248166578393665e-06, "loss": 0.2401, "num_input_tokens_seen": 8704208, "step": 32450 }, { "epoch": 8.507208387942333, "grad_norm": 1.3491195440292358, "learning_rate": 3.319119781513144e-06, "loss": 0.2879, "num_input_tokens_seen": 8705440, "step": 32455 }, { "epoch": 8.508519003931848, "grad_norm": 1.3280109167099, "learning_rate": 3.313427442957695e-06, "loss": 0.0348, "num_input_tokens_seen": 8706368, "step": 32460 }, { "epoch": 8.509829619921362, "grad_norm": 3.083831548690796, "learning_rate": 3.3077396433644046e-06, "loss": 0.1652, "num_input_tokens_seen": 8708048, "step": 32465 }, { "epoch": 8.511140235910878, "grad_norm": 2.7942538261413574, "learning_rate": 3.3020563839237267e-06, "loss": 0.1616, "num_input_tokens_seen": 8709200, "step": 32470 }, { "epoch": 8.512450851900393, "grad_norm": 2.731616735458374, "learning_rate": 3.2963776658251537e-06, "loss": 0.159, "num_input_tokens_seen": 8710304, "step": 32475 }, { "epoch": 8.513761467889909, "grad_norm": 2.678412914276123, "learning_rate": 3.2907034902572382e-06, "loss": 0.3113, "num_input_tokens_seen": 8711632, "step": 32480 }, { "epoch": 8.515072083879424, "grad_norm": 1.3500713109970093, "learning_rate": 3.2850338584075686e-06, "loss": 0.0631, "num_input_tokens_seen": 8712464, "step": 32485 }, { "epoch": 8.516382699868938, "grad_norm": 5.809161186218262, "learning_rate": 3.279368771462793e-06, "loss": 0.2275, "num_input_tokens_seen": 8713568, "step": 32490 }, { "epoch": 8.517693315858454, "grad_norm": 4.065499305725098, "learning_rate": 3.273708230608602e-06, "loss": 0.1845, "num_input_tokens_seen": 8715136, "step": 32495 }, { "epoch": 8.51900393184797, "grad_norm": 2.2365481853485107, "learning_rate": 3.2680522370297397e-06, "loss": 0.109, "num_input_tokens_seen": 8717136, "step": 32500 }, { "epoch": 8.520314547837483, "grad_norm": 1.7314362525939941, "learning_rate": 3.262400791909992e-06, "loss": 0.1344, "num_input_tokens_seen": 8718544, "step": 32505 }, { "epoch": 8.521625163826998, "grad_norm": 1.1906815767288208, "learning_rate": 3.256753896432202e-06, "loss": 0.1414, "num_input_tokens_seen": 8719680, "step": 32510 }, { "epoch": 8.522935779816514, "grad_norm": 2.0105655193328857, "learning_rate": 3.2511115517782494e-06, "loss": 0.0855, "num_input_tokens_seen": 8721344, "step": 32515 }, { "epoch": 8.52424639580603, "grad_norm": 1.5882692337036133, "learning_rate": 3.2454737591290695e-06, "loss": 0.246, "num_input_tokens_seen": 8722752, "step": 32520 }, { "epoch": 8.525557011795543, "grad_norm": 2.018402338027954, "learning_rate": 3.2398405196646443e-06, "loss": 0.0454, "num_input_tokens_seen": 8723984, "step": 32525 }, { "epoch": 8.526867627785059, "grad_norm": 2.7601943016052246, "learning_rate": 3.2342118345639993e-06, "loss": 0.2172, "num_input_tokens_seen": 8725184, "step": 32530 }, { "epoch": 8.528178243774574, "grad_norm": 2.6665139198303223, "learning_rate": 3.2285877050052093e-06, "loss": 0.127, "num_input_tokens_seen": 8726592, "step": 32535 }, { "epoch": 8.52948885976409, "grad_norm": 1.3262053728103638, "learning_rate": 3.2229681321653966e-06, "loss": 0.0901, "num_input_tokens_seen": 8727664, "step": 32540 }, { "epoch": 8.530799475753604, "grad_norm": 2.02828311920166, "learning_rate": 3.2173531172207283e-06, "loss": 0.1748, "num_input_tokens_seen": 8728928, "step": 32545 }, { "epoch": 8.53211009174312, "grad_norm": 3.1216752529144287, "learning_rate": 3.211742661346409e-06, "loss": 0.1292, "num_input_tokens_seen": 8730256, "step": 32550 }, { "epoch": 8.533420707732635, "grad_norm": 0.529518723487854, "learning_rate": 3.2061367657167025e-06, "loss": 0.1356, "num_input_tokens_seen": 8731248, "step": 32555 }, { "epoch": 8.534731323722148, "grad_norm": 2.158207893371582, "learning_rate": 3.200535431504914e-06, "loss": 0.1955, "num_input_tokens_seen": 8732624, "step": 32560 }, { "epoch": 8.536041939711664, "grad_norm": 2.776562452316284, "learning_rate": 3.194938659883398e-06, "loss": 0.1211, "num_input_tokens_seen": 8733728, "step": 32565 }, { "epoch": 8.53735255570118, "grad_norm": 2.57124924659729, "learning_rate": 3.189346452023542e-06, "loss": 0.1766, "num_input_tokens_seen": 8734784, "step": 32570 }, { "epoch": 8.538663171690695, "grad_norm": 2.1726396083831787, "learning_rate": 3.1837588090957853e-06, "loss": 0.1027, "num_input_tokens_seen": 8736320, "step": 32575 }, { "epoch": 8.539973787680209, "grad_norm": 2.9530465602874756, "learning_rate": 3.1781757322696224e-06, "loss": 0.1754, "num_input_tokens_seen": 8737424, "step": 32580 }, { "epoch": 8.541284403669724, "grad_norm": 2.6150407791137695, "learning_rate": 3.172597222713569e-06, "loss": 0.1416, "num_input_tokens_seen": 8738592, "step": 32585 }, { "epoch": 8.54259501965924, "grad_norm": 4.033787727355957, "learning_rate": 3.1670232815951994e-06, "loss": 0.1005, "num_input_tokens_seen": 8739936, "step": 32590 }, { "epoch": 8.543905635648755, "grad_norm": 2.8096976280212402, "learning_rate": 3.1614539100811356e-06, "loss": 0.2203, "num_input_tokens_seen": 8741440, "step": 32595 }, { "epoch": 8.54521625163827, "grad_norm": 1.8388835191726685, "learning_rate": 3.1558891093370318e-06, "loss": 0.2072, "num_input_tokens_seen": 8743024, "step": 32600 }, { "epoch": 8.546526867627785, "grad_norm": 0.049210771918296814, "learning_rate": 3.150328880527595e-06, "loss": 0.087, "num_input_tokens_seen": 8744160, "step": 32605 }, { "epoch": 8.5478374836173, "grad_norm": 2.136540412902832, "learning_rate": 3.1447732248165633e-06, "loss": 0.1094, "num_input_tokens_seen": 8745344, "step": 32610 }, { "epoch": 8.549148099606816, "grad_norm": 1.6268221139907837, "learning_rate": 3.139222143366746e-06, "loss": 0.1611, "num_input_tokens_seen": 8746800, "step": 32615 }, { "epoch": 8.55045871559633, "grad_norm": 2.467240810394287, "learning_rate": 3.1336756373399556e-06, "loss": 0.1201, "num_input_tokens_seen": 8748416, "step": 32620 }, { "epoch": 8.551769331585845, "grad_norm": 1.2280819416046143, "learning_rate": 3.128133707897071e-06, "loss": 0.1595, "num_input_tokens_seen": 8749808, "step": 32625 }, { "epoch": 8.55307994757536, "grad_norm": 0.23791538178920746, "learning_rate": 3.12259635619801e-06, "loss": 0.2437, "num_input_tokens_seen": 8751008, "step": 32630 }, { "epoch": 8.554390563564876, "grad_norm": 1.1963038444519043, "learning_rate": 3.1170635834017303e-06, "loss": 0.1381, "num_input_tokens_seen": 8752416, "step": 32635 }, { "epoch": 8.55570117955439, "grad_norm": 2.925530433654785, "learning_rate": 3.1115353906662313e-06, "loss": 0.1119, "num_input_tokens_seen": 8753456, "step": 32640 }, { "epoch": 8.557011795543906, "grad_norm": 4.614274024963379, "learning_rate": 3.1060117791485544e-06, "loss": 0.1541, "num_input_tokens_seen": 8754768, "step": 32645 }, { "epoch": 8.558322411533421, "grad_norm": 1.5584431886672974, "learning_rate": 3.1004927500047827e-06, "loss": 0.1482, "num_input_tokens_seen": 8756160, "step": 32650 }, { "epoch": 8.559633027522935, "grad_norm": 3.275758981704712, "learning_rate": 3.0949783043900393e-06, "loss": 0.1219, "num_input_tokens_seen": 8757088, "step": 32655 }, { "epoch": 8.56094364351245, "grad_norm": 1.7539896965026855, "learning_rate": 3.0894684434584958e-06, "loss": 0.2306, "num_input_tokens_seen": 8758464, "step": 32660 }, { "epoch": 8.562254259501966, "grad_norm": 3.2358736991882324, "learning_rate": 3.083963168363341e-06, "loss": 0.1423, "num_input_tokens_seen": 8759792, "step": 32665 }, { "epoch": 8.563564875491481, "grad_norm": 2.207368850708008, "learning_rate": 3.0784624802568192e-06, "loss": 0.1669, "num_input_tokens_seen": 8761248, "step": 32670 }, { "epoch": 8.564875491480995, "grad_norm": 3.149803400039673, "learning_rate": 3.072966380290232e-06, "loss": 0.1365, "num_input_tokens_seen": 8762480, "step": 32675 }, { "epoch": 8.56618610747051, "grad_norm": 2.338050603866577, "learning_rate": 3.0674748696138927e-06, "loss": 0.2416, "num_input_tokens_seen": 8764176, "step": 32680 }, { "epoch": 8.567496723460026, "grad_norm": 0.9279637932777405, "learning_rate": 3.0619879493771654e-06, "loss": 0.2441, "num_input_tokens_seen": 8765344, "step": 32685 }, { "epoch": 8.568807339449542, "grad_norm": 2.046204090118408, "learning_rate": 3.056505620728456e-06, "loss": 0.0591, "num_input_tokens_seen": 8766608, "step": 32690 }, { "epoch": 8.570117955439056, "grad_norm": 1.6334306001663208, "learning_rate": 3.051027884815211e-06, "loss": 0.2646, "num_input_tokens_seen": 8768576, "step": 32695 }, { "epoch": 8.571428571428571, "grad_norm": 2.3286564350128174, "learning_rate": 3.0455547427838986e-06, "loss": 0.1051, "num_input_tokens_seen": 8769712, "step": 32700 }, { "epoch": 8.572739187418087, "grad_norm": 0.694810152053833, "learning_rate": 3.0400861957800423e-06, "loss": 0.2622, "num_input_tokens_seen": 8771792, "step": 32705 }, { "epoch": 8.574049803407602, "grad_norm": 1.5022097826004028, "learning_rate": 3.034622244948204e-06, "loss": 0.1677, "num_input_tokens_seen": 8772912, "step": 32710 }, { "epoch": 8.575360419397116, "grad_norm": 0.9320288300514221, "learning_rate": 3.029162891431975e-06, "loss": 0.1415, "num_input_tokens_seen": 8774656, "step": 32715 }, { "epoch": 8.576671035386632, "grad_norm": 1.5586111545562744, "learning_rate": 3.02370813637399e-06, "loss": 0.1123, "num_input_tokens_seen": 8776688, "step": 32720 }, { "epoch": 8.577981651376147, "grad_norm": 1.66806161403656, "learning_rate": 3.018257980915917e-06, "loss": 0.0681, "num_input_tokens_seen": 8777680, "step": 32725 }, { "epoch": 8.579292267365663, "grad_norm": 2.9209375381469727, "learning_rate": 3.0128124261984693e-06, "loss": 0.2023, "num_input_tokens_seen": 8778656, "step": 32730 }, { "epoch": 8.580602883355176, "grad_norm": 3.3488409519195557, "learning_rate": 3.007371473361398e-06, "loss": 0.1234, "num_input_tokens_seen": 8779968, "step": 32735 }, { "epoch": 8.581913499344692, "grad_norm": 1.739315152168274, "learning_rate": 3.00193512354347e-06, "loss": 0.2098, "num_input_tokens_seen": 8781888, "step": 32740 }, { "epoch": 8.583224115334207, "grad_norm": 4.11474609375, "learning_rate": 2.996503377882512e-06, "loss": 0.1873, "num_input_tokens_seen": 8783424, "step": 32745 }, { "epoch": 8.584534731323721, "grad_norm": 2.1893210411071777, "learning_rate": 2.9910762375153752e-06, "loss": 0.0549, "num_input_tokens_seen": 8784496, "step": 32750 }, { "epoch": 8.585845347313237, "grad_norm": 3.7401320934295654, "learning_rate": 2.9856537035779564e-06, "loss": 0.2593, "num_input_tokens_seen": 8785840, "step": 32755 }, { "epoch": 8.587155963302752, "grad_norm": 3.8563175201416016, "learning_rate": 2.980235777205179e-06, "loss": 0.1327, "num_input_tokens_seen": 8787392, "step": 32760 }, { "epoch": 8.588466579292268, "grad_norm": 3.014867067337036, "learning_rate": 2.9748224595310042e-06, "loss": 0.1416, "num_input_tokens_seen": 8788704, "step": 32765 }, { "epoch": 8.589777195281782, "grad_norm": 2.587286949157715, "learning_rate": 2.9694137516884353e-06, "loss": 0.1558, "num_input_tokens_seen": 8789984, "step": 32770 }, { "epoch": 8.591087811271297, "grad_norm": 3.249298572540283, "learning_rate": 2.9640096548094992e-06, "loss": 0.1444, "num_input_tokens_seen": 8791792, "step": 32775 }, { "epoch": 8.592398427260813, "grad_norm": 0.7792115211486816, "learning_rate": 2.958610170025275e-06, "loss": 0.1477, "num_input_tokens_seen": 8793152, "step": 32780 }, { "epoch": 8.593709043250328, "grad_norm": 2.791329860687256, "learning_rate": 2.9532152984658446e-06, "loss": 0.1643, "num_input_tokens_seen": 8794432, "step": 32785 }, { "epoch": 8.595019659239842, "grad_norm": 4.029712677001953, "learning_rate": 2.9478250412603657e-06, "loss": 0.1005, "num_input_tokens_seen": 8795648, "step": 32790 }, { "epoch": 8.596330275229358, "grad_norm": 1.2140578031539917, "learning_rate": 2.942439399537e-06, "loss": 0.091, "num_input_tokens_seen": 8796640, "step": 32795 }, { "epoch": 8.597640891218873, "grad_norm": 2.948127269744873, "learning_rate": 2.9370583744229572e-06, "loss": 0.3083, "num_input_tokens_seen": 8797744, "step": 32800 }, { "epoch": 8.598951507208389, "grad_norm": 2.0531041622161865, "learning_rate": 2.931681967044475e-06, "loss": 0.0995, "num_input_tokens_seen": 8798800, "step": 32805 }, { "epoch": 8.600262123197902, "grad_norm": 2.842202663421631, "learning_rate": 2.9263101785268254e-06, "loss": 0.1758, "num_input_tokens_seen": 8800016, "step": 32810 }, { "epoch": 8.601572739187418, "grad_norm": 3.042788028717041, "learning_rate": 2.9209430099943204e-06, "loss": 0.1535, "num_input_tokens_seen": 8801488, "step": 32815 }, { "epoch": 8.602883355176933, "grad_norm": 2.5491902828216553, "learning_rate": 2.915580462570289e-06, "loss": 0.1218, "num_input_tokens_seen": 8802592, "step": 32820 }, { "epoch": 8.604193971166449, "grad_norm": 3.728121280670166, "learning_rate": 2.910222537377108e-06, "loss": 0.1859, "num_input_tokens_seen": 8803872, "step": 32825 }, { "epoch": 8.605504587155963, "grad_norm": 1.1810035705566406, "learning_rate": 2.904869235536181e-06, "loss": 0.1368, "num_input_tokens_seen": 8805008, "step": 32830 }, { "epoch": 8.606815203145478, "grad_norm": 1.5672125816345215, "learning_rate": 2.8995205581679452e-06, "loss": 0.045, "num_input_tokens_seen": 8806000, "step": 32835 }, { "epoch": 8.608125819134994, "grad_norm": 3.4066720008850098, "learning_rate": 2.8941765063918657e-06, "loss": 0.1114, "num_input_tokens_seen": 8807248, "step": 32840 }, { "epoch": 8.609436435124508, "grad_norm": 1.5718902349472046, "learning_rate": 2.888837081326451e-06, "loss": 0.0965, "num_input_tokens_seen": 8808752, "step": 32845 }, { "epoch": 8.610747051114023, "grad_norm": 2.249486207962036, "learning_rate": 2.8835022840892373e-06, "loss": 0.0898, "num_input_tokens_seen": 8810096, "step": 32850 }, { "epoch": 8.612057667103539, "grad_norm": 1.8250452280044556, "learning_rate": 2.8781721157967734e-06, "loss": 0.2868, "num_input_tokens_seen": 8812144, "step": 32855 }, { "epoch": 8.613368283093054, "grad_norm": 3.054180383682251, "learning_rate": 2.872846577564664e-06, "loss": 0.2291, "num_input_tokens_seen": 8813200, "step": 32860 }, { "epoch": 8.614678899082568, "grad_norm": 1.6001074314117432, "learning_rate": 2.8675256705075343e-06, "loss": 0.1565, "num_input_tokens_seen": 8814496, "step": 32865 }, { "epoch": 8.615989515072084, "grad_norm": 4.04578161239624, "learning_rate": 2.8622093957390377e-06, "loss": 0.2031, "num_input_tokens_seen": 8815600, "step": 32870 }, { "epoch": 8.617300131061599, "grad_norm": 5.4042205810546875, "learning_rate": 2.856897754371865e-06, "loss": 0.2346, "num_input_tokens_seen": 8816672, "step": 32875 }, { "epoch": 8.618610747051115, "grad_norm": 7.422481060028076, "learning_rate": 2.8515907475177293e-06, "loss": 0.1166, "num_input_tokens_seen": 8817824, "step": 32880 }, { "epoch": 8.619921363040628, "grad_norm": 2.221155881881714, "learning_rate": 2.8462883762873842e-06, "loss": 0.0957, "num_input_tokens_seen": 8819072, "step": 32885 }, { "epoch": 8.621231979030144, "grad_norm": 1.7494885921478271, "learning_rate": 2.8409906417906e-06, "loss": 0.0968, "num_input_tokens_seen": 8820272, "step": 32890 }, { "epoch": 8.62254259501966, "grad_norm": 1.7626723051071167, "learning_rate": 2.8356975451361873e-06, "loss": 0.1195, "num_input_tokens_seen": 8821520, "step": 32895 }, { "epoch": 8.623853211009175, "grad_norm": 1.3333401679992676, "learning_rate": 2.8304090874319843e-06, "loss": 0.2078, "num_input_tokens_seen": 8822768, "step": 32900 }, { "epoch": 8.625163826998689, "grad_norm": 5.016307353973389, "learning_rate": 2.8251252697848505e-06, "loss": 0.1781, "num_input_tokens_seen": 8824384, "step": 32905 }, { "epoch": 8.626474442988204, "grad_norm": 1.9804579019546509, "learning_rate": 2.819846093300682e-06, "loss": 0.2354, "num_input_tokens_seen": 8825712, "step": 32910 }, { "epoch": 8.62778505897772, "grad_norm": 1.353192925453186, "learning_rate": 2.814571559084403e-06, "loss": 0.1287, "num_input_tokens_seen": 8826688, "step": 32915 }, { "epoch": 8.629095674967235, "grad_norm": 1.0631028413772583, "learning_rate": 2.8093016682399636e-06, "loss": 0.1804, "num_input_tokens_seen": 8828144, "step": 32920 }, { "epoch": 8.63040629095675, "grad_norm": 0.8730196952819824, "learning_rate": 2.804036421870343e-06, "loss": 0.156, "num_input_tokens_seen": 8829856, "step": 32925 }, { "epoch": 8.631716906946265, "grad_norm": 3.77496600151062, "learning_rate": 2.798775821077551e-06, "loss": 0.0942, "num_input_tokens_seen": 8831280, "step": 32930 }, { "epoch": 8.63302752293578, "grad_norm": 0.43558135628700256, "learning_rate": 2.7935198669626155e-06, "loss": 0.1952, "num_input_tokens_seen": 8832688, "step": 32935 }, { "epoch": 8.634338138925294, "grad_norm": 2.49953293800354, "learning_rate": 2.788268560625601e-06, "loss": 0.117, "num_input_tokens_seen": 8833952, "step": 32940 }, { "epoch": 8.63564875491481, "grad_norm": 5.34187126159668, "learning_rate": 2.7830219031655973e-06, "loss": 0.2005, "num_input_tokens_seen": 8835504, "step": 32945 }, { "epoch": 8.636959370904325, "grad_norm": 6.7802734375, "learning_rate": 2.777779895680721e-06, "loss": 0.1893, "num_input_tokens_seen": 8837216, "step": 32950 }, { "epoch": 8.63826998689384, "grad_norm": 2.123671293258667, "learning_rate": 2.772542539268114e-06, "loss": 0.1651, "num_input_tokens_seen": 8838176, "step": 32955 }, { "epoch": 8.639580602883354, "grad_norm": 4.013540744781494, "learning_rate": 2.7673098350239485e-06, "loss": 0.1492, "num_input_tokens_seen": 8839440, "step": 32960 }, { "epoch": 8.64089121887287, "grad_norm": 2.090508222579956, "learning_rate": 2.7620817840434216e-06, "loss": 0.1173, "num_input_tokens_seen": 8840688, "step": 32965 }, { "epoch": 8.642201834862385, "grad_norm": 0.48181796073913574, "learning_rate": 2.756858387420758e-06, "loss": 0.1627, "num_input_tokens_seen": 8841680, "step": 32970 }, { "epoch": 8.643512450851901, "grad_norm": 1.6519737243652344, "learning_rate": 2.7516396462491934e-06, "loss": 0.1922, "num_input_tokens_seen": 8843120, "step": 32975 }, { "epoch": 8.644823066841415, "grad_norm": 1.808053731918335, "learning_rate": 2.7464255616210106e-06, "loss": 0.0974, "num_input_tokens_seen": 8844752, "step": 32980 }, { "epoch": 8.64613368283093, "grad_norm": 1.4628220796585083, "learning_rate": 2.7412161346275057e-06, "loss": 0.2323, "num_input_tokens_seen": 8845856, "step": 32985 }, { "epoch": 8.647444298820446, "grad_norm": 1.8642383813858032, "learning_rate": 2.7360113663590036e-06, "loss": 0.2219, "num_input_tokens_seen": 8846976, "step": 32990 }, { "epoch": 8.648754914809961, "grad_norm": 1.8206160068511963, "learning_rate": 2.7308112579048555e-06, "loss": 0.2262, "num_input_tokens_seen": 8848832, "step": 32995 }, { "epoch": 8.650065530799475, "grad_norm": 1.581413984298706, "learning_rate": 2.7256158103534298e-06, "loss": 0.1237, "num_input_tokens_seen": 8850080, "step": 33000 }, { "epoch": 8.65137614678899, "grad_norm": 1.3178975582122803, "learning_rate": 2.7204250247921308e-06, "loss": 0.1413, "num_input_tokens_seen": 8851552, "step": 33005 }, { "epoch": 8.652686762778506, "grad_norm": 1.7222307920455933, "learning_rate": 2.7152389023073766e-06, "loss": 0.1868, "num_input_tokens_seen": 8852688, "step": 33010 }, { "epoch": 8.653997378768022, "grad_norm": 1.8122174739837646, "learning_rate": 2.710057443984615e-06, "loss": 0.112, "num_input_tokens_seen": 8854176, "step": 33015 }, { "epoch": 8.655307994757536, "grad_norm": 3.9530932903289795, "learning_rate": 2.7048806509083146e-06, "loss": 0.1547, "num_input_tokens_seen": 8855584, "step": 33020 }, { "epoch": 8.656618610747051, "grad_norm": 3.351153612136841, "learning_rate": 2.6997085241619728e-06, "loss": 0.2194, "num_input_tokens_seen": 8857056, "step": 33025 }, { "epoch": 8.657929226736567, "grad_norm": 2.6135737895965576, "learning_rate": 2.6945410648281044e-06, "loss": 0.21, "num_input_tokens_seen": 8858304, "step": 33030 }, { "epoch": 8.65923984272608, "grad_norm": 1.2511171102523804, "learning_rate": 2.6893782739882523e-06, "loss": 0.1345, "num_input_tokens_seen": 8859936, "step": 33035 }, { "epoch": 8.660550458715596, "grad_norm": 7.370398044586182, "learning_rate": 2.6842201527229743e-06, "loss": 0.1374, "num_input_tokens_seen": 8860976, "step": 33040 }, { "epoch": 8.661861074705111, "grad_norm": 0.6731420755386353, "learning_rate": 2.6790667021118626e-06, "loss": 0.0713, "num_input_tokens_seen": 8861792, "step": 33045 }, { "epoch": 8.663171690694627, "grad_norm": 3.407555341720581, "learning_rate": 2.6739179232335264e-06, "loss": 0.2539, "num_input_tokens_seen": 8862880, "step": 33050 }, { "epoch": 8.66448230668414, "grad_norm": 3.6794941425323486, "learning_rate": 2.6687738171655873e-06, "loss": 0.1916, "num_input_tokens_seen": 8863952, "step": 33055 }, { "epoch": 8.665792922673656, "grad_norm": 1.9179753065109253, "learning_rate": 2.6636343849847033e-06, "loss": 0.219, "num_input_tokens_seen": 8865632, "step": 33060 }, { "epoch": 8.667103538663172, "grad_norm": 2.5183801651000977, "learning_rate": 2.6584996277665475e-06, "loss": 0.1, "num_input_tokens_seen": 8867136, "step": 33065 }, { "epoch": 8.668414154652687, "grad_norm": 1.5444164276123047, "learning_rate": 2.653369546585813e-06, "loss": 0.1466, "num_input_tokens_seen": 8868288, "step": 33070 }, { "epoch": 8.669724770642201, "grad_norm": 1.0248584747314453, "learning_rate": 2.6482441425162235e-06, "loss": 0.0815, "num_input_tokens_seen": 8869584, "step": 33075 }, { "epoch": 8.671035386631717, "grad_norm": 2.6091647148132324, "learning_rate": 2.6431234166305135e-06, "loss": 0.1538, "num_input_tokens_seen": 8871216, "step": 33080 }, { "epoch": 8.672346002621232, "grad_norm": 0.48970434069633484, "learning_rate": 2.6380073700004504e-06, "loss": 0.1021, "num_input_tokens_seen": 8872160, "step": 33085 }, { "epoch": 8.673656618610748, "grad_norm": 2.11242413520813, "learning_rate": 2.6328960036967996e-06, "loss": 0.1287, "num_input_tokens_seen": 8873792, "step": 33090 }, { "epoch": 8.674967234600262, "grad_norm": 5.83452844619751, "learning_rate": 2.627789318789367e-06, "loss": 0.2082, "num_input_tokens_seen": 8874960, "step": 33095 }, { "epoch": 8.676277850589777, "grad_norm": 3.1024281978607178, "learning_rate": 2.6226873163469752e-06, "loss": 0.1551, "num_input_tokens_seen": 8876304, "step": 33100 }, { "epoch": 8.677588466579293, "grad_norm": 2.3377091884613037, "learning_rate": 2.6175899974374614e-06, "loss": 0.0746, "num_input_tokens_seen": 8877312, "step": 33105 }, { "epoch": 8.678899082568808, "grad_norm": 1.8792909383773804, "learning_rate": 2.6124973631276888e-06, "loss": 0.1296, "num_input_tokens_seen": 8878544, "step": 33110 }, { "epoch": 8.680209698558322, "grad_norm": 5.364103317260742, "learning_rate": 2.607409414483536e-06, "loss": 0.2654, "num_input_tokens_seen": 8879424, "step": 33115 }, { "epoch": 8.681520314547837, "grad_norm": 2.3201098442077637, "learning_rate": 2.6023261525699003e-06, "loss": 0.0906, "num_input_tokens_seen": 8880592, "step": 33120 }, { "epoch": 8.682830930537353, "grad_norm": 1.3958053588867188, "learning_rate": 2.597247578450701e-06, "loss": 0.3102, "num_input_tokens_seen": 8882272, "step": 33125 }, { "epoch": 8.684141546526867, "grad_norm": 3.333148717880249, "learning_rate": 2.592173693188876e-06, "loss": 0.1092, "num_input_tokens_seen": 8883248, "step": 33130 }, { "epoch": 8.685452162516382, "grad_norm": 1.4580737352371216, "learning_rate": 2.5871044978463764e-06, "loss": 0.1374, "num_input_tokens_seen": 8884784, "step": 33135 }, { "epoch": 8.686762778505898, "grad_norm": 4.0434346199035645, "learning_rate": 2.582039993484181e-06, "loss": 0.2433, "num_input_tokens_seen": 8886576, "step": 33140 }, { "epoch": 8.688073394495413, "grad_norm": 2.132150888442993, "learning_rate": 2.576980181162278e-06, "loss": 0.1174, "num_input_tokens_seen": 8887616, "step": 33145 }, { "epoch": 8.689384010484927, "grad_norm": 1.9467140436172485, "learning_rate": 2.571925061939681e-06, "loss": 0.1354, "num_input_tokens_seen": 8889232, "step": 33150 }, { "epoch": 8.690694626474443, "grad_norm": 1.369160532951355, "learning_rate": 2.566874636874414e-06, "loss": 0.1777, "num_input_tokens_seen": 8890432, "step": 33155 }, { "epoch": 8.692005242463958, "grad_norm": 6.004194259643555, "learning_rate": 2.561828907023525e-06, "loss": 0.2335, "num_input_tokens_seen": 8892160, "step": 33160 }, { "epoch": 8.693315858453474, "grad_norm": 2.28267240524292, "learning_rate": 2.556787873443081e-06, "loss": 0.1453, "num_input_tokens_seen": 8893360, "step": 33165 }, { "epoch": 8.694626474442988, "grad_norm": 2.92558217048645, "learning_rate": 2.5517515371881516e-06, "loss": 0.108, "num_input_tokens_seen": 8895104, "step": 33170 }, { "epoch": 8.695937090432503, "grad_norm": 1.318305492401123, "learning_rate": 2.5467198993128355e-06, "loss": 0.0837, "num_input_tokens_seen": 8896304, "step": 33175 }, { "epoch": 8.697247706422019, "grad_norm": 2.576427459716797, "learning_rate": 2.541692960870251e-06, "loss": 0.2027, "num_input_tokens_seen": 8897440, "step": 33180 }, { "epoch": 8.698558322411534, "grad_norm": 1.7493723630905151, "learning_rate": 2.536670722912518e-06, "loss": 0.13, "num_input_tokens_seen": 8898736, "step": 33185 }, { "epoch": 8.699868938401048, "grad_norm": 1.268141269683838, "learning_rate": 2.5316531864907955e-06, "loss": 0.0772, "num_input_tokens_seen": 8899632, "step": 33190 }, { "epoch": 8.701179554390563, "grad_norm": 2.2447190284729004, "learning_rate": 2.526640352655238e-06, "loss": 0.1346, "num_input_tokens_seen": 8900848, "step": 33195 }, { "epoch": 8.702490170380079, "grad_norm": 9.667533874511719, "learning_rate": 2.5216322224550226e-06, "loss": 0.074, "num_input_tokens_seen": 8901808, "step": 33200 }, { "epoch": 8.703800786369595, "grad_norm": 1.738473653793335, "learning_rate": 2.51662879693835e-06, "loss": 0.135, "num_input_tokens_seen": 8903008, "step": 33205 }, { "epoch": 8.705111402359108, "grad_norm": 1.6895551681518555, "learning_rate": 2.511630077152416e-06, "loss": 0.1774, "num_input_tokens_seen": 8904192, "step": 33210 }, { "epoch": 8.706422018348624, "grad_norm": 1.7895909547805786, "learning_rate": 2.50663606414345e-06, "loss": 0.2049, "num_input_tokens_seen": 8905696, "step": 33215 }, { "epoch": 8.70773263433814, "grad_norm": 0.7613011598587036, "learning_rate": 2.501646758956691e-06, "loss": 0.2952, "num_input_tokens_seen": 8907440, "step": 33220 }, { "epoch": 8.709043250327653, "grad_norm": 3.909841775894165, "learning_rate": 2.49666216263639e-06, "loss": 0.1935, "num_input_tokens_seen": 8908544, "step": 33225 }, { "epoch": 8.710353866317169, "grad_norm": 4.866018772125244, "learning_rate": 2.491682276225818e-06, "loss": 0.0888, "num_input_tokens_seen": 8909744, "step": 33230 }, { "epoch": 8.711664482306684, "grad_norm": 3.0249626636505127, "learning_rate": 2.486707100767252e-06, "loss": 0.1929, "num_input_tokens_seen": 8911040, "step": 33235 }, { "epoch": 8.7129750982962, "grad_norm": 4.256736755371094, "learning_rate": 2.4817366373019902e-06, "loss": 0.1167, "num_input_tokens_seen": 8912144, "step": 33240 }, { "epoch": 8.714285714285714, "grad_norm": 4.8992438316345215, "learning_rate": 2.4767708868703414e-06, "loss": 0.1588, "num_input_tokens_seen": 8913296, "step": 33245 }, { "epoch": 8.715596330275229, "grad_norm": 1.74736750125885, "learning_rate": 2.4718098505116305e-06, "loss": 0.1208, "num_input_tokens_seen": 8914464, "step": 33250 }, { "epoch": 8.716906946264745, "grad_norm": 2.677234411239624, "learning_rate": 2.4668535292641904e-06, "loss": 0.2307, "num_input_tokens_seen": 8916016, "step": 33255 }, { "epoch": 8.71821756225426, "grad_norm": 2.177039384841919, "learning_rate": 2.4619019241653716e-06, "loss": 0.1432, "num_input_tokens_seen": 8917328, "step": 33260 }, { "epoch": 8.719528178243774, "grad_norm": 1.0738686323165894, "learning_rate": 2.4569550362515403e-06, "loss": 0.1024, "num_input_tokens_seen": 8919232, "step": 33265 }, { "epoch": 8.72083879423329, "grad_norm": 3.043980360031128, "learning_rate": 2.452012866558065e-06, "loss": 0.1271, "num_input_tokens_seen": 8920512, "step": 33270 }, { "epoch": 8.722149410222805, "grad_norm": 3.396092414855957, "learning_rate": 2.4470754161193406e-06, "loss": 0.0904, "num_input_tokens_seen": 8921680, "step": 33275 }, { "epoch": 8.72346002621232, "grad_norm": 3.9442873001098633, "learning_rate": 2.4421426859687597e-06, "loss": 0.2847, "num_input_tokens_seen": 8923056, "step": 33280 }, { "epoch": 8.724770642201834, "grad_norm": 1.6951699256896973, "learning_rate": 2.437214677138744e-06, "loss": 0.0744, "num_input_tokens_seen": 8923968, "step": 33285 }, { "epoch": 8.72608125819135, "grad_norm": 1.7726953029632568, "learning_rate": 2.432291390660707e-06, "loss": 0.1474, "num_input_tokens_seen": 8925248, "step": 33290 }, { "epoch": 8.727391874180865, "grad_norm": 0.5677197575569153, "learning_rate": 2.427372827565086e-06, "loss": 0.1371, "num_input_tokens_seen": 8926320, "step": 33295 }, { "epoch": 8.728702490170381, "grad_norm": 0.7235933542251587, "learning_rate": 2.4224589888813263e-06, "loss": 0.1517, "num_input_tokens_seen": 8927840, "step": 33300 }, { "epoch": 8.730013106159895, "grad_norm": 1.863829255104065, "learning_rate": 2.4175498756378924e-06, "loss": 0.2137, "num_input_tokens_seen": 8929024, "step": 33305 }, { "epoch": 8.73132372214941, "grad_norm": 2.153381109237671, "learning_rate": 2.41264548886225e-06, "loss": 0.1931, "num_input_tokens_seen": 8930240, "step": 33310 }, { "epoch": 8.732634338138926, "grad_norm": 1.0598951578140259, "learning_rate": 2.407745829580882e-06, "loss": 0.0469, "num_input_tokens_seen": 8931280, "step": 33315 }, { "epoch": 8.73394495412844, "grad_norm": 8.718171119689941, "learning_rate": 2.402850898819278e-06, "loss": 0.1634, "num_input_tokens_seen": 8932480, "step": 33320 }, { "epoch": 8.735255570117955, "grad_norm": 3.2657482624053955, "learning_rate": 2.397960697601931e-06, "loss": 0.2779, "num_input_tokens_seen": 8934048, "step": 33325 }, { "epoch": 8.73656618610747, "grad_norm": 2.6694142818450928, "learning_rate": 2.393075226952357e-06, "loss": 0.2419, "num_input_tokens_seen": 8935520, "step": 33330 }, { "epoch": 8.737876802096986, "grad_norm": 3.240748405456543, "learning_rate": 2.388194487893078e-06, "loss": 0.1892, "num_input_tokens_seen": 8937248, "step": 33335 }, { "epoch": 8.7391874180865, "grad_norm": 3.938525676727295, "learning_rate": 2.3833184814456204e-06, "loss": 0.1633, "num_input_tokens_seen": 8938688, "step": 33340 }, { "epoch": 8.740498034076015, "grad_norm": 4.32220458984375, "learning_rate": 2.378447208630527e-06, "loss": 0.2077, "num_input_tokens_seen": 8940176, "step": 33345 }, { "epoch": 8.741808650065531, "grad_norm": 1.3062759637832642, "learning_rate": 2.373580670467343e-06, "loss": 0.1985, "num_input_tokens_seen": 8941568, "step": 33350 }, { "epoch": 8.743119266055047, "grad_norm": 1.9020127058029175, "learning_rate": 2.3687188679746315e-06, "loss": 0.2417, "num_input_tokens_seen": 8943376, "step": 33355 }, { "epoch": 8.74442988204456, "grad_norm": 0.5505346059799194, "learning_rate": 2.363861802169959e-06, "loss": 0.1149, "num_input_tokens_seen": 8944832, "step": 33360 }, { "epoch": 8.745740498034076, "grad_norm": 1.880000352859497, "learning_rate": 2.3590094740698955e-06, "loss": 0.204, "num_input_tokens_seen": 8946368, "step": 33365 }, { "epoch": 8.747051114023591, "grad_norm": 1.4167176485061646, "learning_rate": 2.3541618846900316e-06, "loss": 0.1251, "num_input_tokens_seen": 8947904, "step": 33370 }, { "epoch": 8.748361730013107, "grad_norm": 3.305227279663086, "learning_rate": 2.3493190350449557e-06, "loss": 0.2063, "num_input_tokens_seen": 8949168, "step": 33375 }, { "epoch": 8.74967234600262, "grad_norm": 1.401639461517334, "learning_rate": 2.3444809261482653e-06, "loss": 0.2173, "num_input_tokens_seen": 8950880, "step": 33380 }, { "epoch": 8.750982961992136, "grad_norm": 1.2679133415222168, "learning_rate": 2.3396475590125753e-06, "loss": 0.2304, "num_input_tokens_seen": 8952688, "step": 33385 }, { "epoch": 8.752293577981652, "grad_norm": 3.415750741958618, "learning_rate": 2.334818934649494e-06, "loss": 0.2136, "num_input_tokens_seen": 8954448, "step": 33390 }, { "epoch": 8.753604193971167, "grad_norm": 3.421959638595581, "learning_rate": 2.3299950540696504e-06, "loss": 0.1493, "num_input_tokens_seen": 8955872, "step": 33395 }, { "epoch": 8.754914809960681, "grad_norm": 2.8792037963867188, "learning_rate": 2.325175918282674e-06, "loss": 0.2021, "num_input_tokens_seen": 8957840, "step": 33400 }, { "epoch": 8.756225425950197, "grad_norm": 4.693619251251221, "learning_rate": 2.3203615282971934e-06, "loss": 0.1984, "num_input_tokens_seen": 8959216, "step": 33405 }, { "epoch": 8.757536041939712, "grad_norm": 3.7640466690063477, "learning_rate": 2.315551885120859e-06, "loss": 0.1406, "num_input_tokens_seen": 8960464, "step": 33410 }, { "epoch": 8.758846657929226, "grad_norm": 2.341484308242798, "learning_rate": 2.3107469897603152e-06, "loss": 0.1965, "num_input_tokens_seen": 8961536, "step": 33415 }, { "epoch": 8.760157273918741, "grad_norm": 2.3274986743927, "learning_rate": 2.305946843221224e-06, "loss": 0.0799, "num_input_tokens_seen": 8962720, "step": 33420 }, { "epoch": 8.761467889908257, "grad_norm": 2.3670616149902344, "learning_rate": 2.3011514465082485e-06, "loss": 0.0935, "num_input_tokens_seen": 8963888, "step": 33425 }, { "epoch": 8.762778505897773, "grad_norm": 1.7574084997177124, "learning_rate": 2.2963608006250555e-06, "loss": 0.064, "num_input_tokens_seen": 8965008, "step": 33430 }, { "epoch": 8.764089121887286, "grad_norm": 2.546762704849243, "learning_rate": 2.2915749065743176e-06, "loss": 0.1761, "num_input_tokens_seen": 8966432, "step": 33435 }, { "epoch": 8.765399737876802, "grad_norm": 3.343644142150879, "learning_rate": 2.2867937653577173e-06, "loss": 0.2096, "num_input_tokens_seen": 8968304, "step": 33440 }, { "epoch": 8.766710353866317, "grad_norm": 2.2713146209716797, "learning_rate": 2.282017377975934e-06, "loss": 0.1881, "num_input_tokens_seen": 8969600, "step": 33445 }, { "epoch": 8.768020969855833, "grad_norm": 3.142483949661255, "learning_rate": 2.2772457454286578e-06, "loss": 0.1641, "num_input_tokens_seen": 8970640, "step": 33450 }, { "epoch": 8.769331585845347, "grad_norm": 1.3896517753601074, "learning_rate": 2.2724788687145865e-06, "loss": 0.1593, "num_input_tokens_seen": 8971776, "step": 33455 }, { "epoch": 8.770642201834862, "grad_norm": 1.6070173978805542, "learning_rate": 2.2677167488314193e-06, "loss": 0.1491, "num_input_tokens_seen": 8973136, "step": 33460 }, { "epoch": 8.771952817824378, "grad_norm": 1.7603377103805542, "learning_rate": 2.2629593867758564e-06, "loss": 0.1991, "num_input_tokens_seen": 8974544, "step": 33465 }, { "epoch": 8.773263433813893, "grad_norm": 3.1123054027557373, "learning_rate": 2.258206783543601e-06, "loss": 0.1342, "num_input_tokens_seen": 8975632, "step": 33470 }, { "epoch": 8.774574049803407, "grad_norm": 1.799720048904419, "learning_rate": 2.253458940129383e-06, "loss": 0.1571, "num_input_tokens_seen": 8976800, "step": 33475 }, { "epoch": 8.775884665792923, "grad_norm": 1.34115731716156, "learning_rate": 2.2487158575269006e-06, "loss": 0.0628, "num_input_tokens_seen": 8977952, "step": 33480 }, { "epoch": 8.777195281782438, "grad_norm": 2.4367027282714844, "learning_rate": 2.243977536728878e-06, "loss": 0.1207, "num_input_tokens_seen": 8979392, "step": 33485 }, { "epoch": 8.778505897771954, "grad_norm": 5.024295330047607, "learning_rate": 2.2392439787270373e-06, "loss": 0.1058, "num_input_tokens_seen": 8980416, "step": 33490 }, { "epoch": 8.779816513761467, "grad_norm": 2.0952324867248535, "learning_rate": 2.234515184512104e-06, "loss": 0.1357, "num_input_tokens_seen": 8981904, "step": 33495 }, { "epoch": 8.781127129750983, "grad_norm": 2.708621025085449, "learning_rate": 2.229791155073807e-06, "loss": 0.1357, "num_input_tokens_seen": 8983696, "step": 33500 }, { "epoch": 8.782437745740499, "grad_norm": 0.44706907868385315, "learning_rate": 2.225071891400879e-06, "loss": 0.1329, "num_input_tokens_seen": 8984752, "step": 33505 }, { "epoch": 8.783748361730012, "grad_norm": 4.129914283752441, "learning_rate": 2.2203573944810514e-06, "loss": 0.1029, "num_input_tokens_seen": 8985984, "step": 33510 }, { "epoch": 8.785058977719528, "grad_norm": 3.34647274017334, "learning_rate": 2.21564766530106e-06, "loss": 0.1477, "num_input_tokens_seen": 8987232, "step": 33515 }, { "epoch": 8.786369593709043, "grad_norm": 3.570307970046997, "learning_rate": 2.210942704846647e-06, "loss": 0.1521, "num_input_tokens_seen": 8988912, "step": 33520 }, { "epoch": 8.787680209698559, "grad_norm": 1.5756107568740845, "learning_rate": 2.2062425141025465e-06, "loss": 0.1217, "num_input_tokens_seen": 8990144, "step": 33525 }, { "epoch": 8.788990825688073, "grad_norm": 2.2680797576904297, "learning_rate": 2.2015470940525003e-06, "loss": 0.2287, "num_input_tokens_seen": 8991568, "step": 33530 }, { "epoch": 8.790301441677588, "grad_norm": 2.6886374950408936, "learning_rate": 2.1968564456792573e-06, "loss": 0.2095, "num_input_tokens_seen": 8993072, "step": 33535 }, { "epoch": 8.791612057667104, "grad_norm": 4.2486138343811035, "learning_rate": 2.1921705699645607e-06, "loss": 0.1485, "num_input_tokens_seen": 8994336, "step": 33540 }, { "epoch": 8.79292267365662, "grad_norm": 2.8244404792785645, "learning_rate": 2.187489467889153e-06, "loss": 0.1305, "num_input_tokens_seen": 8995632, "step": 33545 }, { "epoch": 8.794233289646133, "grad_norm": 1.8317714929580688, "learning_rate": 2.182813140432785e-06, "loss": 0.0891, "num_input_tokens_seen": 8996704, "step": 33550 }, { "epoch": 8.795543905635649, "grad_norm": 1.8470107316970825, "learning_rate": 2.1781415885742037e-06, "loss": 0.146, "num_input_tokens_seen": 8998608, "step": 33555 }, { "epoch": 8.796854521625164, "grad_norm": 1.7955381870269775, "learning_rate": 2.1734748132911515e-06, "loss": 0.1566, "num_input_tokens_seen": 8999952, "step": 33560 }, { "epoch": 8.79816513761468, "grad_norm": 1.772293210029602, "learning_rate": 2.1688128155603783e-06, "loss": 0.1084, "num_input_tokens_seen": 9001312, "step": 33565 }, { "epoch": 8.799475753604193, "grad_norm": 1.7174484729766846, "learning_rate": 2.1641555963576366e-06, "loss": 0.0928, "num_input_tokens_seen": 9002528, "step": 33570 }, { "epoch": 8.800786369593709, "grad_norm": 0.7377776503562927, "learning_rate": 2.159503156657669e-06, "loss": 0.1089, "num_input_tokens_seen": 9003776, "step": 33575 }, { "epoch": 8.802096985583225, "grad_norm": 1.9113919734954834, "learning_rate": 2.1548554974342244e-06, "loss": 0.1011, "num_input_tokens_seen": 9004768, "step": 33580 }, { "epoch": 8.80340760157274, "grad_norm": 1.8387210369110107, "learning_rate": 2.15021261966005e-06, "loss": 0.212, "num_input_tokens_seen": 9005984, "step": 33585 }, { "epoch": 8.804718217562254, "grad_norm": 3.6245615482330322, "learning_rate": 2.145574524306898e-06, "loss": 0.1514, "num_input_tokens_seen": 9007200, "step": 33590 }, { "epoch": 8.80602883355177, "grad_norm": 1.7994086742401123, "learning_rate": 2.140941212345507e-06, "loss": 0.1634, "num_input_tokens_seen": 9009152, "step": 33595 }, { "epoch": 8.807339449541285, "grad_norm": 0.34349197149276733, "learning_rate": 2.136312684745623e-06, "loss": 0.0768, "num_input_tokens_seen": 9011744, "step": 33600 }, { "epoch": 8.808650065530799, "grad_norm": 1.1388682126998901, "learning_rate": 2.1316889424759874e-06, "loss": 0.1727, "num_input_tokens_seen": 9013424, "step": 33605 }, { "epoch": 8.809960681520314, "grad_norm": 3.5155210494995117, "learning_rate": 2.1270699865043405e-06, "loss": 0.1564, "num_input_tokens_seen": 9014384, "step": 33610 }, { "epoch": 8.81127129750983, "grad_norm": 3.82850980758667, "learning_rate": 2.122455817797428e-06, "loss": 0.2011, "num_input_tokens_seen": 9015936, "step": 33615 }, { "epoch": 8.812581913499345, "grad_norm": 1.2568384408950806, "learning_rate": 2.1178464373209807e-06, "loss": 0.1366, "num_input_tokens_seen": 9017040, "step": 33620 }, { "epoch": 8.813892529488859, "grad_norm": 3.6976318359375, "learning_rate": 2.1132418460397376e-06, "loss": 0.3266, "num_input_tokens_seen": 9018464, "step": 33625 }, { "epoch": 8.815203145478375, "grad_norm": 2.273711681365967, "learning_rate": 2.1086420449174303e-06, "loss": 0.1096, "num_input_tokens_seen": 9019648, "step": 33630 }, { "epoch": 8.81651376146789, "grad_norm": 1.9782689809799194, "learning_rate": 2.1040470349167923e-06, "loss": 0.1754, "num_input_tokens_seen": 9021360, "step": 33635 }, { "epoch": 8.817824377457406, "grad_norm": 1.931650996208191, "learning_rate": 2.0994568169995455e-06, "loss": 0.1223, "num_input_tokens_seen": 9022496, "step": 33640 }, { "epoch": 8.81913499344692, "grad_norm": 1.8885554075241089, "learning_rate": 2.0948713921264214e-06, "loss": 0.1329, "num_input_tokens_seen": 9023792, "step": 33645 }, { "epoch": 8.820445609436435, "grad_norm": 1.1074391603469849, "learning_rate": 2.0902907612571355e-06, "loss": 0.1012, "num_input_tokens_seen": 9024928, "step": 33650 }, { "epoch": 8.82175622542595, "grad_norm": 2.8282480239868164, "learning_rate": 2.085714925350407e-06, "loss": 0.2293, "num_input_tokens_seen": 9026256, "step": 33655 }, { "epoch": 8.823066841415466, "grad_norm": 2.252067804336548, "learning_rate": 2.0811438853639537e-06, "loss": 0.2507, "num_input_tokens_seen": 9027904, "step": 33660 }, { "epoch": 8.82437745740498, "grad_norm": 2.3114166259765625, "learning_rate": 2.076577642254485e-06, "loss": 0.203, "num_input_tokens_seen": 9029168, "step": 33665 }, { "epoch": 8.825688073394495, "grad_norm": 2.4931528568267822, "learning_rate": 2.072016196977708e-06, "loss": 0.1707, "num_input_tokens_seen": 9030336, "step": 33670 }, { "epoch": 8.82699868938401, "grad_norm": 4.065838813781738, "learning_rate": 2.0674595504883265e-06, "loss": 0.1469, "num_input_tokens_seen": 9031488, "step": 33675 }, { "epoch": 8.828309305373526, "grad_norm": 2.52976655960083, "learning_rate": 2.062907703740033e-06, "loss": 0.1293, "num_input_tokens_seen": 9033152, "step": 33680 }, { "epoch": 8.82961992136304, "grad_norm": 1.8691022396087646, "learning_rate": 2.0583606576855236e-06, "loss": 0.1275, "num_input_tokens_seen": 9034352, "step": 33685 }, { "epoch": 8.830930537352556, "grad_norm": 1.4459967613220215, "learning_rate": 2.0538184132764897e-06, "loss": 0.1684, "num_input_tokens_seen": 9035712, "step": 33690 }, { "epoch": 8.832241153342071, "grad_norm": 2.1590425968170166, "learning_rate": 2.0492809714636156e-06, "loss": 0.1374, "num_input_tokens_seen": 9036784, "step": 33695 }, { "epoch": 8.833551769331585, "grad_norm": 1.410834789276123, "learning_rate": 2.044748333196572e-06, "loss": 0.1189, "num_input_tokens_seen": 9038256, "step": 33700 }, { "epoch": 8.8348623853211, "grad_norm": 1.1854991912841797, "learning_rate": 2.040220499424042e-06, "loss": 0.1156, "num_input_tokens_seen": 9039712, "step": 33705 }, { "epoch": 8.836173001310616, "grad_norm": 2.3122527599334717, "learning_rate": 2.0356974710936976e-06, "loss": 0.1614, "num_input_tokens_seen": 9041440, "step": 33710 }, { "epoch": 8.837483617300132, "grad_norm": 1.489254117012024, "learning_rate": 2.0311792491521874e-06, "loss": 0.0998, "num_input_tokens_seen": 9042448, "step": 33715 }, { "epoch": 8.838794233289645, "grad_norm": 2.2802231311798096, "learning_rate": 2.0266658345451715e-06, "loss": 0.1645, "num_input_tokens_seen": 9043824, "step": 33720 }, { "epoch": 8.840104849279161, "grad_norm": 2.838434934616089, "learning_rate": 2.022157228217303e-06, "loss": 0.1664, "num_input_tokens_seen": 9045168, "step": 33725 }, { "epoch": 8.841415465268676, "grad_norm": 2.903109312057495, "learning_rate": 2.0176534311122263e-06, "loss": 0.128, "num_input_tokens_seen": 9046368, "step": 33730 }, { "epoch": 8.842726081258192, "grad_norm": 4.0327043533325195, "learning_rate": 2.0131544441725715e-06, "loss": 0.2387, "num_input_tokens_seen": 9047520, "step": 33735 }, { "epoch": 8.844036697247706, "grad_norm": 5.717251777648926, "learning_rate": 2.0086602683399764e-06, "loss": 0.1452, "num_input_tokens_seen": 9048960, "step": 33740 }, { "epoch": 8.845347313237221, "grad_norm": 2.111473798751831, "learning_rate": 2.0041709045550612e-06, "loss": 0.0623, "num_input_tokens_seen": 9049744, "step": 33745 }, { "epoch": 8.846657929226737, "grad_norm": 3.2994959354400635, "learning_rate": 1.9996863537574406e-06, "loss": 0.2936, "num_input_tokens_seen": 9051040, "step": 33750 }, { "epoch": 8.847968545216252, "grad_norm": 6.426774501800537, "learning_rate": 1.9952066168857252e-06, "loss": 0.0848, "num_input_tokens_seen": 9051920, "step": 33755 }, { "epoch": 8.849279161205766, "grad_norm": 7.828963279724121, "learning_rate": 1.990731694877515e-06, "loss": 0.4975, "num_input_tokens_seen": 9053600, "step": 33760 }, { "epoch": 8.850589777195282, "grad_norm": 3.547431468963623, "learning_rate": 1.9862615886694054e-06, "loss": 0.1857, "num_input_tokens_seen": 9054848, "step": 33765 }, { "epoch": 8.851900393184797, "grad_norm": 6.166888236999512, "learning_rate": 1.9817962991969805e-06, "loss": 0.185, "num_input_tokens_seen": 9056032, "step": 33770 }, { "epoch": 8.853211009174313, "grad_norm": 3.0907669067382812, "learning_rate": 1.977335827394816e-06, "loss": 0.1935, "num_input_tokens_seen": 9057216, "step": 33775 }, { "epoch": 8.854521625163827, "grad_norm": 2.4439680576324463, "learning_rate": 1.972880174196484e-06, "loss": 0.1577, "num_input_tokens_seen": 9058528, "step": 33780 }, { "epoch": 8.855832241153342, "grad_norm": 3.0838606357574463, "learning_rate": 1.9684293405345445e-06, "loss": 0.1363, "num_input_tokens_seen": 9059568, "step": 33785 }, { "epoch": 8.857142857142858, "grad_norm": 2.1840996742248535, "learning_rate": 1.9639833273405516e-06, "loss": 0.2098, "num_input_tokens_seen": 9060960, "step": 33790 }, { "epoch": 8.858453473132371, "grad_norm": 2.6528968811035156, "learning_rate": 1.959542135545042e-06, "loss": 0.1187, "num_input_tokens_seen": 9062576, "step": 33795 }, { "epoch": 8.859764089121887, "grad_norm": 0.28066977858543396, "learning_rate": 1.9551057660775527e-06, "loss": 0.1352, "num_input_tokens_seen": 9064096, "step": 33800 }, { "epoch": 8.861074705111402, "grad_norm": 1.792215347290039, "learning_rate": 1.950674219866608e-06, "loss": 0.2148, "num_input_tokens_seen": 9065552, "step": 33805 }, { "epoch": 8.862385321100918, "grad_norm": 1.9395787715911865, "learning_rate": 1.9462474978397204e-06, "loss": 0.1384, "num_input_tokens_seen": 9066992, "step": 33810 }, { "epoch": 8.863695937090432, "grad_norm": 2.769362688064575, "learning_rate": 1.941825600923397e-06, "loss": 0.1814, "num_input_tokens_seen": 9068368, "step": 33815 }, { "epoch": 8.865006553079947, "grad_norm": 5.051325798034668, "learning_rate": 1.937408530043136e-06, "loss": 0.2899, "num_input_tokens_seen": 9069840, "step": 33820 }, { "epoch": 8.866317169069463, "grad_norm": 2.2162363529205322, "learning_rate": 1.9329962861234257e-06, "loss": 0.2376, "num_input_tokens_seen": 9071248, "step": 33825 }, { "epoch": 8.867627785058978, "grad_norm": 0.13054540753364563, "learning_rate": 1.9285888700877303e-06, "loss": 0.1063, "num_input_tokens_seen": 9072800, "step": 33830 }, { "epoch": 8.868938401048492, "grad_norm": 2.6157419681549072, "learning_rate": 1.92418628285852e-06, "loss": 0.1449, "num_input_tokens_seen": 9073904, "step": 33835 }, { "epoch": 8.870249017038008, "grad_norm": 1.0545536279678345, "learning_rate": 1.9197885253572497e-06, "loss": 0.0604, "num_input_tokens_seen": 9075184, "step": 33840 }, { "epoch": 8.871559633027523, "grad_norm": 1.8270957469940186, "learning_rate": 1.915395598504363e-06, "loss": 0.1262, "num_input_tokens_seen": 9076464, "step": 33845 }, { "epoch": 8.872870249017039, "grad_norm": 2.6735780239105225, "learning_rate": 1.9110075032192894e-06, "loss": 0.2219, "num_input_tokens_seen": 9077680, "step": 33850 }, { "epoch": 8.874180865006553, "grad_norm": 0.8053278923034668, "learning_rate": 1.9066242404204487e-06, "loss": 0.0638, "num_input_tokens_seen": 9078976, "step": 33855 }, { "epoch": 8.875491480996068, "grad_norm": 1.4670414924621582, "learning_rate": 1.902245811025255e-06, "loss": 0.1245, "num_input_tokens_seen": 9080352, "step": 33860 }, { "epoch": 8.876802096985584, "grad_norm": 1.453296184539795, "learning_rate": 1.8978722159501029e-06, "loss": 0.1482, "num_input_tokens_seen": 9081552, "step": 33865 }, { "epoch": 8.8781127129751, "grad_norm": 1.466187596321106, "learning_rate": 1.8935034561103764e-06, "loss": 0.1298, "num_input_tokens_seen": 9082752, "step": 33870 }, { "epoch": 8.879423328964613, "grad_norm": 1.8410165309906006, "learning_rate": 1.889139532420453e-06, "loss": 0.0865, "num_input_tokens_seen": 9083936, "step": 33875 }, { "epoch": 8.880733944954128, "grad_norm": 3.1124188899993896, "learning_rate": 1.8847804457936934e-06, "loss": 0.0753, "num_input_tokens_seen": 9085264, "step": 33880 }, { "epoch": 8.882044560943644, "grad_norm": 2.319841146469116, "learning_rate": 1.8804261971424458e-06, "loss": 0.047, "num_input_tokens_seen": 9086512, "step": 33885 }, { "epoch": 8.883355176933158, "grad_norm": 2.5108749866485596, "learning_rate": 1.8760767873780476e-06, "loss": 0.0705, "num_input_tokens_seen": 9087648, "step": 33890 }, { "epoch": 8.884665792922673, "grad_norm": 1.517497181892395, "learning_rate": 1.871732217410821e-06, "loss": 0.2254, "num_input_tokens_seen": 9089120, "step": 33895 }, { "epoch": 8.885976408912189, "grad_norm": 0.7717084288597107, "learning_rate": 1.8673924881500826e-06, "loss": 0.072, "num_input_tokens_seen": 9090368, "step": 33900 }, { "epoch": 8.887287024901704, "grad_norm": 1.9316141605377197, "learning_rate": 1.8630576005041228e-06, "loss": 0.1165, "num_input_tokens_seen": 9091552, "step": 33905 }, { "epoch": 8.888597640891218, "grad_norm": 2.9178314208984375, "learning_rate": 1.8587275553802352e-06, "loss": 0.0996, "num_input_tokens_seen": 9092800, "step": 33910 }, { "epoch": 8.889908256880734, "grad_norm": 3.2532646656036377, "learning_rate": 1.854402353684681e-06, "loss": 0.2382, "num_input_tokens_seen": 9094512, "step": 33915 }, { "epoch": 8.89121887287025, "grad_norm": 0.1537981778383255, "learning_rate": 1.8500819963227223e-06, "loss": 0.1238, "num_input_tokens_seen": 9095776, "step": 33920 }, { "epoch": 8.892529488859765, "grad_norm": 1.9477297067642212, "learning_rate": 1.8457664841986023e-06, "loss": 0.1459, "num_input_tokens_seen": 9096992, "step": 33925 }, { "epoch": 8.893840104849279, "grad_norm": 9.729101181030273, "learning_rate": 1.8414558182155456e-06, "loss": 0.0888, "num_input_tokens_seen": 9098160, "step": 33930 }, { "epoch": 8.895150720838794, "grad_norm": 2.5673060417175293, "learning_rate": 1.8371499992757723e-06, "loss": 0.1295, "num_input_tokens_seen": 9100016, "step": 33935 }, { "epoch": 8.89646133682831, "grad_norm": 1.69512140750885, "learning_rate": 1.832849028280484e-06, "loss": 0.0795, "num_input_tokens_seen": 9101248, "step": 33940 }, { "epoch": 8.897771952817825, "grad_norm": 1.1334962844848633, "learning_rate": 1.8285529061298684e-06, "loss": 0.037, "num_input_tokens_seen": 9102448, "step": 33945 }, { "epoch": 8.899082568807339, "grad_norm": 2.4449028968811035, "learning_rate": 1.82426163372309e-06, "loss": 0.0648, "num_input_tokens_seen": 9103568, "step": 33950 }, { "epoch": 8.900393184796854, "grad_norm": 1.808228611946106, "learning_rate": 1.8199752119583052e-06, "loss": 0.0825, "num_input_tokens_seen": 9104656, "step": 33955 }, { "epoch": 8.90170380078637, "grad_norm": 1.9001350402832031, "learning_rate": 1.8156936417326576e-06, "loss": 0.2582, "num_input_tokens_seen": 9106288, "step": 33960 }, { "epoch": 8.903014416775886, "grad_norm": 1.3463313579559326, "learning_rate": 1.8114169239422691e-06, "loss": 0.1101, "num_input_tokens_seen": 9107728, "step": 33965 }, { "epoch": 8.9043250327654, "grad_norm": 2.172478437423706, "learning_rate": 1.8071450594822543e-06, "loss": 0.1721, "num_input_tokens_seen": 9108912, "step": 33970 }, { "epoch": 8.905635648754915, "grad_norm": 0.6818680763244629, "learning_rate": 1.8028780492467007e-06, "loss": 0.091, "num_input_tokens_seen": 9110784, "step": 33975 }, { "epoch": 8.90694626474443, "grad_norm": 4.404728889465332, "learning_rate": 1.798615894128694e-06, "loss": 0.1645, "num_input_tokens_seen": 9112016, "step": 33980 }, { "epoch": 8.908256880733944, "grad_norm": 4.06793212890625, "learning_rate": 1.7943585950202902e-06, "loss": 0.0913, "num_input_tokens_seen": 9113184, "step": 33985 }, { "epoch": 8.90956749672346, "grad_norm": 2.4263956546783447, "learning_rate": 1.7901061528125347e-06, "loss": 0.161, "num_input_tokens_seen": 9114832, "step": 33990 }, { "epoch": 8.910878112712975, "grad_norm": 1.2119075059890747, "learning_rate": 1.7858585683954597e-06, "loss": 0.1357, "num_input_tokens_seen": 9116128, "step": 33995 }, { "epoch": 8.91218872870249, "grad_norm": 1.3376243114471436, "learning_rate": 1.7816158426580742e-06, "loss": 0.1117, "num_input_tokens_seen": 9117408, "step": 34000 }, { "epoch": 8.913499344692005, "grad_norm": 2.282101631164551, "learning_rate": 1.7773779764883758e-06, "loss": 0.0972, "num_input_tokens_seen": 9120192, "step": 34005 }, { "epoch": 8.91480996068152, "grad_norm": 3.554633617401123, "learning_rate": 1.7731449707733412e-06, "loss": 0.1415, "num_input_tokens_seen": 9121616, "step": 34010 }, { "epoch": 8.916120576671036, "grad_norm": 3.4549710750579834, "learning_rate": 1.7689168263989309e-06, "loss": 0.1127, "num_input_tokens_seen": 9122912, "step": 34015 }, { "epoch": 8.917431192660551, "grad_norm": 2.603228807449341, "learning_rate": 1.7646935442500872e-06, "loss": 0.1985, "num_input_tokens_seen": 9124336, "step": 34020 }, { "epoch": 8.918741808650065, "grad_norm": 2.9487340450286865, "learning_rate": 1.760475125210745e-06, "loss": 0.1273, "num_input_tokens_seen": 9125472, "step": 34025 }, { "epoch": 8.92005242463958, "grad_norm": 0.6003196239471436, "learning_rate": 1.7562615701637975e-06, "loss": 0.2448, "num_input_tokens_seen": 9126560, "step": 34030 }, { "epoch": 8.921363040629096, "grad_norm": 3.6549577713012695, "learning_rate": 1.7520528799911395e-06, "loss": 0.1394, "num_input_tokens_seen": 9127776, "step": 34035 }, { "epoch": 8.922673656618612, "grad_norm": 2.2137320041656494, "learning_rate": 1.7478490555736471e-06, "loss": 0.1162, "num_input_tokens_seen": 9129248, "step": 34040 }, { "epoch": 8.923984272608125, "grad_norm": 2.2354276180267334, "learning_rate": 1.7436500977911663e-06, "loss": 0.2044, "num_input_tokens_seen": 9130896, "step": 34045 }, { "epoch": 8.92529488859764, "grad_norm": 4.162858963012695, "learning_rate": 1.7394560075225414e-06, "loss": 0.1418, "num_input_tokens_seen": 9131936, "step": 34050 }, { "epoch": 8.926605504587156, "grad_norm": 1.5876582860946655, "learning_rate": 1.7352667856455812e-06, "loss": 0.1266, "num_input_tokens_seen": 9133168, "step": 34055 }, { "epoch": 8.927916120576672, "grad_norm": 2.5676779747009277, "learning_rate": 1.7310824330370873e-06, "loss": 0.123, "num_input_tokens_seen": 9134272, "step": 34060 }, { "epoch": 8.929226736566186, "grad_norm": 2.190635919570923, "learning_rate": 1.7269029505728313e-06, "loss": 0.1439, "num_input_tokens_seen": 9135520, "step": 34065 }, { "epoch": 8.930537352555701, "grad_norm": 1.7521297931671143, "learning_rate": 1.7227283391275772e-06, "loss": 0.0731, "num_input_tokens_seen": 9136752, "step": 34070 }, { "epoch": 8.931847968545217, "grad_norm": 1.7199889421463013, "learning_rate": 1.7185585995750598e-06, "loss": 0.0518, "num_input_tokens_seen": 9137568, "step": 34075 }, { "epoch": 8.93315858453473, "grad_norm": 2.358344554901123, "learning_rate": 1.714393732788e-06, "loss": 0.0844, "num_input_tokens_seen": 9138688, "step": 34080 }, { "epoch": 8.934469200524246, "grad_norm": 3.3934922218322754, "learning_rate": 1.710233739638098e-06, "loss": 0.0774, "num_input_tokens_seen": 9139872, "step": 34085 }, { "epoch": 8.935779816513762, "grad_norm": 2.196794033050537, "learning_rate": 1.706078620996035e-06, "loss": 0.2076, "num_input_tokens_seen": 9142192, "step": 34090 }, { "epoch": 8.937090432503277, "grad_norm": 1.7987394332885742, "learning_rate": 1.7019283777314653e-06, "loss": 0.1905, "num_input_tokens_seen": 9143664, "step": 34095 }, { "epoch": 8.938401048492791, "grad_norm": 2.48622989654541, "learning_rate": 1.6977830107130305e-06, "loss": 0.1712, "num_input_tokens_seen": 9145136, "step": 34100 }, { "epoch": 8.939711664482306, "grad_norm": 2.2850534915924072, "learning_rate": 1.6936425208083473e-06, "loss": 0.1427, "num_input_tokens_seen": 9146544, "step": 34105 }, { "epoch": 8.941022280471822, "grad_norm": 1.3995418548583984, "learning_rate": 1.6895069088840144e-06, "loss": 0.0989, "num_input_tokens_seen": 9148336, "step": 34110 }, { "epoch": 8.942332896461338, "grad_norm": 2.6262705326080322, "learning_rate": 1.6853761758056114e-06, "loss": 0.3596, "num_input_tokens_seen": 9151376, "step": 34115 }, { "epoch": 8.943643512450851, "grad_norm": 1.8379396200180054, "learning_rate": 1.6812503224376857e-06, "loss": 0.085, "num_input_tokens_seen": 9152304, "step": 34120 }, { "epoch": 8.944954128440367, "grad_norm": 2.5288331508636475, "learning_rate": 1.67712934964378e-06, "loss": 0.1676, "num_input_tokens_seen": 9153632, "step": 34125 }, { "epoch": 8.946264744429882, "grad_norm": 2.6267004013061523, "learning_rate": 1.673013258286399e-06, "loss": 0.1836, "num_input_tokens_seen": 9154928, "step": 34130 }, { "epoch": 8.947575360419398, "grad_norm": 1.6990704536437988, "learning_rate": 1.668902049227039e-06, "loss": 0.0753, "num_input_tokens_seen": 9156096, "step": 34135 }, { "epoch": 8.948885976408912, "grad_norm": 6.255670070648193, "learning_rate": 1.664795723326168e-06, "loss": 0.1673, "num_input_tokens_seen": 9157472, "step": 34140 }, { "epoch": 8.950196592398427, "grad_norm": 2.774883985519409, "learning_rate": 1.660694281443234e-06, "loss": 0.2231, "num_input_tokens_seen": 9159088, "step": 34145 }, { "epoch": 8.951507208387943, "grad_norm": 3.3818347454071045, "learning_rate": 1.6565977244366564e-06, "loss": 0.2205, "num_input_tokens_seen": 9160304, "step": 34150 }, { "epoch": 8.952817824377458, "grad_norm": 1.8262863159179688, "learning_rate": 1.6525060531638408e-06, "loss": 0.1281, "num_input_tokens_seen": 9161824, "step": 34155 }, { "epoch": 8.954128440366972, "grad_norm": 1.1410380601882935, "learning_rate": 1.6484192684811634e-06, "loss": 0.2108, "num_input_tokens_seen": 9163040, "step": 34160 }, { "epoch": 8.955439056356488, "grad_norm": 0.7980388402938843, "learning_rate": 1.644337371243987e-06, "loss": 0.1466, "num_input_tokens_seen": 9164352, "step": 34165 }, { "epoch": 8.956749672346003, "grad_norm": 0.29689130187034607, "learning_rate": 1.6402603623066448e-06, "loss": 0.1232, "num_input_tokens_seen": 9165488, "step": 34170 }, { "epoch": 8.958060288335517, "grad_norm": 1.745961308479309, "learning_rate": 1.6361882425224461e-06, "loss": 0.1799, "num_input_tokens_seen": 9167280, "step": 34175 }, { "epoch": 8.959370904325032, "grad_norm": 2.344991683959961, "learning_rate": 1.6321210127436814e-06, "loss": 0.1357, "num_input_tokens_seen": 9168736, "step": 34180 }, { "epoch": 8.960681520314548, "grad_norm": 1.6383717060089111, "learning_rate": 1.6280586738216114e-06, "loss": 0.1513, "num_input_tokens_seen": 9169776, "step": 34185 }, { "epoch": 8.961992136304064, "grad_norm": 4.012542724609375, "learning_rate": 1.6240012266064752e-06, "loss": 0.2785, "num_input_tokens_seen": 9170752, "step": 34190 }, { "epoch": 8.963302752293577, "grad_norm": 3.202653408050537, "learning_rate": 1.619948671947491e-06, "loss": 0.1045, "num_input_tokens_seen": 9171984, "step": 34195 }, { "epoch": 8.964613368283093, "grad_norm": 1.831483244895935, "learning_rate": 1.6159010106928523e-06, "loss": 0.1617, "num_input_tokens_seen": 9173440, "step": 34200 }, { "epoch": 8.965923984272608, "grad_norm": 3.594496250152588, "learning_rate": 1.611858243689729e-06, "loss": 0.1225, "num_input_tokens_seen": 9174464, "step": 34205 }, { "epoch": 8.967234600262124, "grad_norm": 1.4803361892700195, "learning_rate": 1.6078203717842633e-06, "loss": 0.232, "num_input_tokens_seen": 9176080, "step": 34210 }, { "epoch": 8.968545216251638, "grad_norm": 1.4351037740707397, "learning_rate": 1.6037873958215738e-06, "loss": 0.1267, "num_input_tokens_seen": 9178208, "step": 34215 }, { "epoch": 8.969855832241153, "grad_norm": 1.7286677360534668, "learning_rate": 1.5997593166457575e-06, "loss": 0.1376, "num_input_tokens_seen": 9179456, "step": 34220 }, { "epoch": 8.971166448230669, "grad_norm": 1.6115787029266357, "learning_rate": 1.5957361350998846e-06, "loss": 0.1391, "num_input_tokens_seen": 9180736, "step": 34225 }, { "epoch": 8.972477064220184, "grad_norm": 1.7532846927642822, "learning_rate": 1.5917178520259979e-06, "loss": 0.1413, "num_input_tokens_seen": 9182416, "step": 34230 }, { "epoch": 8.973787680209698, "grad_norm": 1.655179500579834, "learning_rate": 1.5877044682651194e-06, "loss": 0.1695, "num_input_tokens_seen": 9183904, "step": 34235 }, { "epoch": 8.975098296199214, "grad_norm": 2.630847930908203, "learning_rate": 1.5836959846572408e-06, "loss": 0.2963, "num_input_tokens_seen": 9185232, "step": 34240 }, { "epoch": 8.97640891218873, "grad_norm": 1.8809977769851685, "learning_rate": 1.579692402041333e-06, "loss": 0.1125, "num_input_tokens_seen": 9186480, "step": 34245 }, { "epoch": 8.977719528178245, "grad_norm": 2.6355605125427246, "learning_rate": 1.5756937212553391e-06, "loss": 0.1772, "num_input_tokens_seen": 9187696, "step": 34250 }, { "epoch": 8.979030144167758, "grad_norm": 1.9827450513839722, "learning_rate": 1.5716999431361762e-06, "loss": 0.1345, "num_input_tokens_seen": 9189168, "step": 34255 }, { "epoch": 8.980340760157274, "grad_norm": 1.0879855155944824, "learning_rate": 1.5677110685197366e-06, "loss": 0.067, "num_input_tokens_seen": 9190192, "step": 34260 }, { "epoch": 8.98165137614679, "grad_norm": 2.3053832054138184, "learning_rate": 1.5637270982408775e-06, "loss": 0.2328, "num_input_tokens_seen": 9191712, "step": 34265 }, { "epoch": 8.982961992136303, "grad_norm": 2.9295852184295654, "learning_rate": 1.559748033133443e-06, "loss": 0.1825, "num_input_tokens_seen": 9192944, "step": 34270 }, { "epoch": 8.984272608125819, "grad_norm": 2.8638405799865723, "learning_rate": 1.555773874030242e-06, "loss": 0.0795, "num_input_tokens_seen": 9193888, "step": 34275 }, { "epoch": 8.985583224115334, "grad_norm": 2.416872024536133, "learning_rate": 1.5518046217630615e-06, "loss": 0.1457, "num_input_tokens_seen": 9195280, "step": 34280 }, { "epoch": 8.98689384010485, "grad_norm": 0.3113875091075897, "learning_rate": 1.5478402771626599e-06, "loss": 0.157, "num_input_tokens_seen": 9196336, "step": 34285 }, { "epoch": 8.988204456094364, "grad_norm": 2.924851894378662, "learning_rate": 1.5438808410587646e-06, "loss": 0.1633, "num_input_tokens_seen": 9197920, "step": 34290 }, { "epoch": 8.98951507208388, "grad_norm": 1.807958722114563, "learning_rate": 1.5399263142800825e-06, "loss": 0.1744, "num_input_tokens_seen": 9199168, "step": 34295 }, { "epoch": 8.990825688073395, "grad_norm": 4.558920860290527, "learning_rate": 1.5359766976542873e-06, "loss": 0.138, "num_input_tokens_seen": 9200176, "step": 34300 }, { "epoch": 8.99213630406291, "grad_norm": 3.0840256214141846, "learning_rate": 1.5320319920080262e-06, "loss": 0.169, "num_input_tokens_seen": 9201248, "step": 34305 }, { "epoch": 8.993446920052424, "grad_norm": 1.3957526683807373, "learning_rate": 1.5280921981669166e-06, "loss": 0.0929, "num_input_tokens_seen": 9202080, "step": 34310 }, { "epoch": 8.99475753604194, "grad_norm": 1.9096814393997192, "learning_rate": 1.5241573169555573e-06, "loss": 0.2718, "num_input_tokens_seen": 9203952, "step": 34315 }, { "epoch": 8.996068152031455, "grad_norm": 0.6404217481613159, "learning_rate": 1.5202273491975061e-06, "loss": 0.3196, "num_input_tokens_seen": 9204992, "step": 34320 }, { "epoch": 8.997378768020969, "grad_norm": 2.550006151199341, "learning_rate": 1.5163022957152994e-06, "loss": 0.115, "num_input_tokens_seen": 9206160, "step": 34325 }, { "epoch": 8.998689384010484, "grad_norm": 1.3037729263305664, "learning_rate": 1.5123821573304442e-06, "loss": 0.1735, "num_input_tokens_seen": 9207520, "step": 34330 }, { "epoch": 9.0, "grad_norm": 6.823376178741455, "learning_rate": 1.5084669348634255e-06, "loss": 0.1172, "num_input_tokens_seen": 9208488, "step": 34335 }, { "epoch": 9.001310615989516, "grad_norm": 4.594335556030273, "learning_rate": 1.5045566291336854e-06, "loss": 0.0914, "num_input_tokens_seen": 9209608, "step": 34340 }, { "epoch": 9.002359108781127, "eval_loss": 0.85427325963974, "eval_runtime": 16.661, "eval_samples_per_second": 50.897, "eval_steps_per_second": 25.449, "num_input_tokens_seen": 9210648, "step": 34344 }, { "epoch": 9.00262123197903, "grad_norm": 1.5309717655181885, "learning_rate": 1.500651240959644e-06, "loss": 0.2131, "num_input_tokens_seen": 9210792, "step": 34345 }, { "epoch": 9.003931847968545, "grad_norm": 0.9570596218109131, "learning_rate": 1.4967507711586976e-06, "loss": 0.0764, "num_input_tokens_seen": 9211864, "step": 34350 }, { "epoch": 9.00524246395806, "grad_norm": 2.454737424850464, "learning_rate": 1.4928552205472012e-06, "loss": 0.1871, "num_input_tokens_seen": 9213288, "step": 34355 }, { "epoch": 9.006553079947576, "grad_norm": 0.8880041241645813, "learning_rate": 1.488964589940489e-06, "loss": 0.1973, "num_input_tokens_seen": 9215224, "step": 34360 }, { "epoch": 9.00786369593709, "grad_norm": 1.498643159866333, "learning_rate": 1.4850788801528653e-06, "loss": 0.0915, "num_input_tokens_seen": 9216280, "step": 34365 }, { "epoch": 9.009174311926605, "grad_norm": 0.4653771221637726, "learning_rate": 1.4811980919976043e-06, "loss": 0.0635, "num_input_tokens_seen": 9217608, "step": 34370 }, { "epoch": 9.01048492791612, "grad_norm": 2.1605546474456787, "learning_rate": 1.4773222262869423e-06, "loss": 0.1578, "num_input_tokens_seen": 9219000, "step": 34375 }, { "epoch": 9.011795543905636, "grad_norm": 1.7289191484451294, "learning_rate": 1.4734512838320974e-06, "loss": 0.0768, "num_input_tokens_seen": 9220488, "step": 34380 }, { "epoch": 9.01310615989515, "grad_norm": 2.0138726234436035, "learning_rate": 1.4695852654432463e-06, "loss": 0.1303, "num_input_tokens_seen": 9221640, "step": 34385 }, { "epoch": 9.014416775884666, "grad_norm": 0.11174427717924118, "learning_rate": 1.4657241719295362e-06, "loss": 0.1586, "num_input_tokens_seen": 9223064, "step": 34390 }, { "epoch": 9.015727391874181, "grad_norm": 2.3413426876068115, "learning_rate": 1.4618680040990983e-06, "loss": 0.1394, "num_input_tokens_seen": 9224184, "step": 34395 }, { "epoch": 9.017038007863697, "grad_norm": 2.560129404067993, "learning_rate": 1.4580167627590153e-06, "loss": 0.1608, "num_input_tokens_seen": 9225816, "step": 34400 }, { "epoch": 9.01834862385321, "grad_norm": 0.6859720349311829, "learning_rate": 1.4541704487153474e-06, "loss": 0.079, "num_input_tokens_seen": 9226888, "step": 34405 }, { "epoch": 9.019659239842726, "grad_norm": 1.2460691928863525, "learning_rate": 1.4503290627731175e-06, "loss": 0.18, "num_input_tokens_seen": 9229096, "step": 34410 }, { "epoch": 9.020969855832242, "grad_norm": 3.022205352783203, "learning_rate": 1.4464926057363298e-06, "loss": 0.0974, "num_input_tokens_seen": 9229896, "step": 34415 }, { "epoch": 9.022280471821757, "grad_norm": 2.4086785316467285, "learning_rate": 1.4426610784079391e-06, "loss": 0.1185, "num_input_tokens_seen": 9230872, "step": 34420 }, { "epoch": 9.02359108781127, "grad_norm": 1.089856743812561, "learning_rate": 1.438834481589882e-06, "loss": 0.1647, "num_input_tokens_seen": 9232296, "step": 34425 }, { "epoch": 9.024901703800786, "grad_norm": 1.7450950145721436, "learning_rate": 1.4350128160830562e-06, "loss": 0.1677, "num_input_tokens_seen": 9233880, "step": 34430 }, { "epoch": 9.026212319790302, "grad_norm": 1.4695261716842651, "learning_rate": 1.4311960826873305e-06, "loss": 0.1668, "num_input_tokens_seen": 9235000, "step": 34435 }, { "epoch": 9.027522935779816, "grad_norm": 4.13171911239624, "learning_rate": 1.427384282201541e-06, "loss": 0.1543, "num_input_tokens_seen": 9236232, "step": 34440 }, { "epoch": 9.028833551769331, "grad_norm": 1.0815494060516357, "learning_rate": 1.4235774154234855e-06, "loss": 0.1273, "num_input_tokens_seen": 9237656, "step": 34445 }, { "epoch": 9.030144167758847, "grad_norm": 1.933716058731079, "learning_rate": 1.4197754831499488e-06, "loss": 0.1291, "num_input_tokens_seen": 9238984, "step": 34450 }, { "epoch": 9.031454783748362, "grad_norm": 0.7016099691390991, "learning_rate": 1.4159784861766584e-06, "loss": 0.1061, "num_input_tokens_seen": 9240024, "step": 34455 }, { "epoch": 9.032765399737876, "grad_norm": 1.6600512266159058, "learning_rate": 1.4121864252983174e-06, "loss": 0.1514, "num_input_tokens_seen": 9241416, "step": 34460 }, { "epoch": 9.034076015727392, "grad_norm": 1.4121421575546265, "learning_rate": 1.408399301308605e-06, "loss": 0.0981, "num_input_tokens_seen": 9242712, "step": 34465 }, { "epoch": 9.035386631716907, "grad_norm": 9.471601486206055, "learning_rate": 1.4046171150001508e-06, "loss": 0.213, "num_input_tokens_seen": 9244584, "step": 34470 }, { "epoch": 9.036697247706423, "grad_norm": 0.7122043371200562, "learning_rate": 1.4008398671645688e-06, "loss": 0.1191, "num_input_tokens_seen": 9246040, "step": 34475 }, { "epoch": 9.038007863695936, "grad_norm": 2.188317060470581, "learning_rate": 1.3970675585924241e-06, "loss": 0.1186, "num_input_tokens_seen": 9246936, "step": 34480 }, { "epoch": 9.039318479685452, "grad_norm": 1.719199299812317, "learning_rate": 1.3933001900732572e-06, "loss": 0.1932, "num_input_tokens_seen": 9248104, "step": 34485 }, { "epoch": 9.040629095674968, "grad_norm": 1.9917184114456177, "learning_rate": 1.3895377623955707e-06, "loss": 0.101, "num_input_tokens_seen": 9249848, "step": 34490 }, { "epoch": 9.041939711664483, "grad_norm": 1.402761697769165, "learning_rate": 1.3857802763468403e-06, "loss": 0.1304, "num_input_tokens_seen": 9251144, "step": 34495 }, { "epoch": 9.043250327653997, "grad_norm": 1.93567955493927, "learning_rate": 1.3820277327134866e-06, "loss": 0.1539, "num_input_tokens_seen": 9252472, "step": 34500 }, { "epoch": 9.044560943643512, "grad_norm": 1.3774219751358032, "learning_rate": 1.3782801322809262e-06, "loss": 0.0882, "num_input_tokens_seen": 9253576, "step": 34505 }, { "epoch": 9.045871559633028, "grad_norm": 3.62141489982605, "learning_rate": 1.374537475833515e-06, "loss": 0.0872, "num_input_tokens_seen": 9254968, "step": 34510 }, { "epoch": 9.047182175622543, "grad_norm": 2.71498441696167, "learning_rate": 1.37079976415459e-06, "loss": 0.2852, "num_input_tokens_seen": 9256440, "step": 34515 }, { "epoch": 9.048492791612057, "grad_norm": 1.2565834522247314, "learning_rate": 1.3670669980264477e-06, "loss": 0.1108, "num_input_tokens_seen": 9257688, "step": 34520 }, { "epoch": 9.049803407601573, "grad_norm": 1.6010178327560425, "learning_rate": 1.3633391782303468e-06, "loss": 0.0785, "num_input_tokens_seen": 9258664, "step": 34525 }, { "epoch": 9.051114023591088, "grad_norm": 2.956028461456299, "learning_rate": 1.3596163055465154e-06, "loss": 0.1338, "num_input_tokens_seen": 9260344, "step": 34530 }, { "epoch": 9.052424639580602, "grad_norm": 2.680701971054077, "learning_rate": 1.3558983807541476e-06, "loss": 0.2299, "num_input_tokens_seen": 9263144, "step": 34535 }, { "epoch": 9.053735255570118, "grad_norm": 3.646207571029663, "learning_rate": 1.3521854046313897e-06, "loss": 0.1693, "num_input_tokens_seen": 9264136, "step": 34540 }, { "epoch": 9.055045871559633, "grad_norm": 1.6179636716842651, "learning_rate": 1.3484773779553677e-06, "loss": 0.2379, "num_input_tokens_seen": 9266360, "step": 34545 }, { "epoch": 9.056356487549149, "grad_norm": 2.4722697734832764, "learning_rate": 1.3447743015021636e-06, "loss": 0.1891, "num_input_tokens_seen": 9267624, "step": 34550 }, { "epoch": 9.057667103538662, "grad_norm": 2.4089765548706055, "learning_rate": 1.3410761760468265e-06, "loss": 0.1979, "num_input_tokens_seen": 9269112, "step": 34555 }, { "epoch": 9.058977719528178, "grad_norm": 1.8056161403656006, "learning_rate": 1.3373830023633598e-06, "loss": 0.1239, "num_input_tokens_seen": 9270440, "step": 34560 }, { "epoch": 9.060288335517694, "grad_norm": 2.271890878677368, "learning_rate": 1.3336947812247507e-06, "loss": 0.0569, "num_input_tokens_seen": 9271704, "step": 34565 }, { "epoch": 9.061598951507209, "grad_norm": 2.064295768737793, "learning_rate": 1.3300115134029345e-06, "loss": 0.0781, "num_input_tokens_seen": 9272984, "step": 34570 }, { "epoch": 9.062909567496723, "grad_norm": 2.7265565395355225, "learning_rate": 1.3263331996688055e-06, "loss": 0.0586, "num_input_tokens_seen": 9274488, "step": 34575 }, { "epoch": 9.064220183486238, "grad_norm": 1.963314414024353, "learning_rate": 1.3226598407922342e-06, "loss": 0.1368, "num_input_tokens_seen": 9276504, "step": 34580 }, { "epoch": 9.065530799475754, "grad_norm": 2.8815369606018066, "learning_rate": 1.318991437542047e-06, "loss": 0.1066, "num_input_tokens_seen": 9278200, "step": 34585 }, { "epoch": 9.06684141546527, "grad_norm": 0.1186383068561554, "learning_rate": 1.3153279906860355e-06, "loss": 0.1826, "num_input_tokens_seen": 9279880, "step": 34590 }, { "epoch": 9.068152031454783, "grad_norm": 2.6999433040618896, "learning_rate": 1.3116695009909474e-06, "loss": 0.0608, "num_input_tokens_seen": 9280952, "step": 34595 }, { "epoch": 9.069462647444299, "grad_norm": 2.691282272338867, "learning_rate": 1.3080159692225063e-06, "loss": 0.1406, "num_input_tokens_seen": 9282008, "step": 34600 }, { "epoch": 9.070773263433814, "grad_norm": 2.5740036964416504, "learning_rate": 1.304367396145384e-06, "loss": 0.1662, "num_input_tokens_seen": 9283304, "step": 34605 }, { "epoch": 9.07208387942333, "grad_norm": 3.728196859359741, "learning_rate": 1.300723782523225e-06, "loss": 0.1283, "num_input_tokens_seen": 9284696, "step": 34610 }, { "epoch": 9.073394495412844, "grad_norm": 2.9445133209228516, "learning_rate": 1.2970851291186276e-06, "loss": 0.0908, "num_input_tokens_seen": 9285880, "step": 34615 }, { "epoch": 9.07470511140236, "grad_norm": 2.070155620574951, "learning_rate": 1.2934514366931578e-06, "loss": 0.0626, "num_input_tokens_seen": 9287112, "step": 34620 }, { "epoch": 9.076015727391875, "grad_norm": 1.8689771890640259, "learning_rate": 1.2898227060073403e-06, "loss": 0.1082, "num_input_tokens_seen": 9288488, "step": 34625 }, { "epoch": 9.077326343381388, "grad_norm": 1.7486282587051392, "learning_rate": 1.2861989378206624e-06, "loss": 0.1496, "num_input_tokens_seen": 9290072, "step": 34630 }, { "epoch": 9.078636959370904, "grad_norm": 1.0517776012420654, "learning_rate": 1.2825801328915699e-06, "loss": 0.1152, "num_input_tokens_seen": 9291160, "step": 34635 }, { "epoch": 9.07994757536042, "grad_norm": 2.8314735889434814, "learning_rate": 1.2789662919774792e-06, "loss": 0.1418, "num_input_tokens_seen": 9292408, "step": 34640 }, { "epoch": 9.081258191349935, "grad_norm": 3.742419481277466, "learning_rate": 1.2753574158347547e-06, "loss": 0.2247, "num_input_tokens_seen": 9294296, "step": 34645 }, { "epoch": 9.082568807339449, "grad_norm": 1.9452733993530273, "learning_rate": 1.271753505218734e-06, "loss": 0.1494, "num_input_tokens_seen": 9295256, "step": 34650 }, { "epoch": 9.083879423328964, "grad_norm": 2.4086384773254395, "learning_rate": 1.2681545608837026e-06, "loss": 0.1617, "num_input_tokens_seen": 9296696, "step": 34655 }, { "epoch": 9.08519003931848, "grad_norm": 0.46922361850738525, "learning_rate": 1.2645605835829132e-06, "loss": 0.1594, "num_input_tokens_seen": 9298792, "step": 34660 }, { "epoch": 9.086500655307995, "grad_norm": 2.7618086338043213, "learning_rate": 1.2609715740685869e-06, "loss": 0.1629, "num_input_tokens_seen": 9300456, "step": 34665 }, { "epoch": 9.08781127129751, "grad_norm": 1.307860016822815, "learning_rate": 1.257387533091889e-06, "loss": 0.1262, "num_input_tokens_seen": 9302104, "step": 34670 }, { "epoch": 9.089121887287025, "grad_norm": 0.6675728559494019, "learning_rate": 1.2538084614029527e-06, "loss": 0.1103, "num_input_tokens_seen": 9303256, "step": 34675 }, { "epoch": 9.09043250327654, "grad_norm": 2.438037633895874, "learning_rate": 1.2502343597508792e-06, "loss": 0.1129, "num_input_tokens_seen": 9304776, "step": 34680 }, { "epoch": 9.091743119266056, "grad_norm": 1.9229469299316406, "learning_rate": 1.2466652288837227e-06, "loss": 0.3951, "num_input_tokens_seen": 9306296, "step": 34685 }, { "epoch": 9.09305373525557, "grad_norm": 1.5156524181365967, "learning_rate": 1.2431010695484858e-06, "loss": 0.1972, "num_input_tokens_seen": 9307800, "step": 34690 }, { "epoch": 9.094364351245085, "grad_norm": 1.5939979553222656, "learning_rate": 1.2395418824911464e-06, "loss": 0.1351, "num_input_tokens_seen": 9309128, "step": 34695 }, { "epoch": 9.0956749672346, "grad_norm": 2.7233681678771973, "learning_rate": 1.2359876684566368e-06, "loss": 0.1244, "num_input_tokens_seen": 9310520, "step": 34700 }, { "epoch": 9.096985583224116, "grad_norm": 1.6752822399139404, "learning_rate": 1.232438428188848e-06, "loss": 0.1674, "num_input_tokens_seen": 9311640, "step": 34705 }, { "epoch": 9.09829619921363, "grad_norm": 8.946744918823242, "learning_rate": 1.228894162430627e-06, "loss": 0.1018, "num_input_tokens_seen": 9312568, "step": 34710 }, { "epoch": 9.099606815203146, "grad_norm": 1.2082995176315308, "learning_rate": 1.225354871923784e-06, "loss": 0.1521, "num_input_tokens_seen": 9314424, "step": 34715 }, { "epoch": 9.100917431192661, "grad_norm": 1.67416512966156, "learning_rate": 1.2218205574090896e-06, "loss": 0.1366, "num_input_tokens_seen": 9316008, "step": 34720 }, { "epoch": 9.102228047182175, "grad_norm": 1.6673219203948975, "learning_rate": 1.2182912196262664e-06, "loss": 0.096, "num_input_tokens_seen": 9317128, "step": 34725 }, { "epoch": 9.10353866317169, "grad_norm": 1.171129822731018, "learning_rate": 1.2147668593139982e-06, "loss": 0.1461, "num_input_tokens_seen": 9318216, "step": 34730 }, { "epoch": 9.104849279161206, "grad_norm": 3.5045292377471924, "learning_rate": 1.2112474772099285e-06, "loss": 0.1207, "num_input_tokens_seen": 9319208, "step": 34735 }, { "epoch": 9.106159895150721, "grad_norm": 1.789284110069275, "learning_rate": 1.2077330740506598e-06, "loss": 0.1102, "num_input_tokens_seen": 9320248, "step": 34740 }, { "epoch": 9.107470511140235, "grad_norm": 2.0601553916931152, "learning_rate": 1.2042236505717452e-06, "loss": 0.1088, "num_input_tokens_seen": 9321912, "step": 34745 }, { "epoch": 9.10878112712975, "grad_norm": 4.1603193283081055, "learning_rate": 1.2007192075077085e-06, "loss": 0.0969, "num_input_tokens_seen": 9323080, "step": 34750 }, { "epoch": 9.110091743119266, "grad_norm": 2.2902276515960693, "learning_rate": 1.1972197455920181e-06, "loss": 0.1627, "num_input_tokens_seen": 9324328, "step": 34755 }, { "epoch": 9.111402359108782, "grad_norm": 3.9465174674987793, "learning_rate": 1.193725265557108e-06, "loss": 0.0799, "num_input_tokens_seen": 9325320, "step": 34760 }, { "epoch": 9.112712975098296, "grad_norm": 3.2967140674591064, "learning_rate": 1.1902357681343623e-06, "loss": 0.1456, "num_input_tokens_seen": 9326632, "step": 34765 }, { "epoch": 9.114023591087811, "grad_norm": 2.274397850036621, "learning_rate": 1.1867512540541388e-06, "loss": 0.1347, "num_input_tokens_seen": 9327768, "step": 34770 }, { "epoch": 9.115334207077327, "grad_norm": 1.6245286464691162, "learning_rate": 1.1832717240457231e-06, "loss": 0.15, "num_input_tokens_seen": 9329176, "step": 34775 }, { "epoch": 9.116644823066842, "grad_norm": 2.2859268188476562, "learning_rate": 1.1797971788373857e-06, "loss": 0.0989, "num_input_tokens_seen": 9330504, "step": 34780 }, { "epoch": 9.117955439056356, "grad_norm": 2.3509724140167236, "learning_rate": 1.1763276191563422e-06, "loss": 0.1233, "num_input_tokens_seen": 9331768, "step": 34785 }, { "epoch": 9.119266055045872, "grad_norm": 2.18066143989563, "learning_rate": 1.1728630457287587e-06, "loss": 0.1091, "num_input_tokens_seen": 9332824, "step": 34790 }, { "epoch": 9.120576671035387, "grad_norm": 2.5536625385284424, "learning_rate": 1.1694034592797747e-06, "loss": 0.1607, "num_input_tokens_seen": 9334616, "step": 34795 }, { "epoch": 9.1218872870249, "grad_norm": 1.921411395072937, "learning_rate": 1.1659488605334695e-06, "loss": 0.1656, "num_input_tokens_seen": 9335864, "step": 34800 }, { "epoch": 9.123197903014416, "grad_norm": 6.100725173950195, "learning_rate": 1.1624992502128868e-06, "loss": 0.2345, "num_input_tokens_seen": 9337336, "step": 34805 }, { "epoch": 9.124508519003932, "grad_norm": 0.8808035254478455, "learning_rate": 1.159054629040024e-06, "loss": 0.0748, "num_input_tokens_seen": 9338600, "step": 34810 }, { "epoch": 9.125819134993447, "grad_norm": 2.3600265979766846, "learning_rate": 1.1556149977358295e-06, "loss": 0.0783, "num_input_tokens_seen": 9339704, "step": 34815 }, { "epoch": 9.127129750982961, "grad_norm": 0.5366150140762329, "learning_rate": 1.1521803570202188e-06, "loss": 0.0724, "num_input_tokens_seen": 9340712, "step": 34820 }, { "epoch": 9.128440366972477, "grad_norm": 2.6191341876983643, "learning_rate": 1.1487507076120507e-06, "loss": 0.1074, "num_input_tokens_seen": 9341640, "step": 34825 }, { "epoch": 9.129750982961992, "grad_norm": 2.537083148956299, "learning_rate": 1.1453260502291507e-06, "loss": 0.2548, "num_input_tokens_seen": 9343096, "step": 34830 }, { "epoch": 9.131061598951508, "grad_norm": 2.1622273921966553, "learning_rate": 1.14190638558829e-06, "loss": 0.136, "num_input_tokens_seen": 9344184, "step": 34835 }, { "epoch": 9.132372214941022, "grad_norm": 1.4451380968093872, "learning_rate": 1.1384917144051987e-06, "loss": 0.1355, "num_input_tokens_seen": 9345512, "step": 34840 }, { "epoch": 9.133682830930537, "grad_norm": 0.7543509602546692, "learning_rate": 1.135082037394561e-06, "loss": 0.1417, "num_input_tokens_seen": 9346888, "step": 34845 }, { "epoch": 9.134993446920053, "grad_norm": 5.718748569488525, "learning_rate": 1.1316773552700193e-06, "loss": 0.1402, "num_input_tokens_seen": 9348104, "step": 34850 }, { "epoch": 9.136304062909568, "grad_norm": 2.1646957397460938, "learning_rate": 1.1282776687441626e-06, "loss": 0.2211, "num_input_tokens_seen": 9349640, "step": 34855 }, { "epoch": 9.137614678899082, "grad_norm": 0.86445152759552, "learning_rate": 1.1248829785285463e-06, "loss": 0.1368, "num_input_tokens_seen": 9350984, "step": 34860 }, { "epoch": 9.138925294888598, "grad_norm": 2.061021327972412, "learning_rate": 1.121493285333669e-06, "loss": 0.1729, "num_input_tokens_seen": 9352200, "step": 34865 }, { "epoch": 9.140235910878113, "grad_norm": 2.719322919845581, "learning_rate": 1.1181085898689881e-06, "loss": 0.1574, "num_input_tokens_seen": 9353656, "step": 34870 }, { "epoch": 9.141546526867629, "grad_norm": 0.5635553598403931, "learning_rate": 1.1147288928429118e-06, "loss": 0.0941, "num_input_tokens_seen": 9355032, "step": 34875 }, { "epoch": 9.142857142857142, "grad_norm": 1.8518296480178833, "learning_rate": 1.1113541949628104e-06, "loss": 0.1459, "num_input_tokens_seen": 9356248, "step": 34880 }, { "epoch": 9.144167758846658, "grad_norm": 1.8367491960525513, "learning_rate": 1.1079844969350023e-06, "loss": 0.0994, "num_input_tokens_seen": 9357480, "step": 34885 }, { "epoch": 9.145478374836173, "grad_norm": 2.6822004318237305, "learning_rate": 1.104619799464754e-06, "loss": 0.11, "num_input_tokens_seen": 9358472, "step": 34890 }, { "epoch": 9.146788990825687, "grad_norm": 4.489284992218018, "learning_rate": 1.1012601032562935e-06, "loss": 0.1664, "num_input_tokens_seen": 9360072, "step": 34895 }, { "epoch": 9.148099606815203, "grad_norm": 0.6879579424858093, "learning_rate": 1.0979054090127978e-06, "loss": 0.1578, "num_input_tokens_seen": 9361432, "step": 34900 }, { "epoch": 9.149410222804718, "grad_norm": 1.3711817264556885, "learning_rate": 1.094555717436399e-06, "loss": 0.1322, "num_input_tokens_seen": 9362600, "step": 34905 }, { "epoch": 9.150720838794234, "grad_norm": 2.126391649246216, "learning_rate": 1.0912110292281868e-06, "loss": 0.125, "num_input_tokens_seen": 9363832, "step": 34910 }, { "epoch": 9.152031454783748, "grad_norm": 2.460510015487671, "learning_rate": 1.0878713450881928e-06, "loss": 0.1194, "num_input_tokens_seen": 9365160, "step": 34915 }, { "epoch": 9.153342070773263, "grad_norm": 1.3154326677322388, "learning_rate": 1.084536665715416e-06, "loss": 0.4938, "num_input_tokens_seen": 9366504, "step": 34920 }, { "epoch": 9.154652686762779, "grad_norm": 1.7411339282989502, "learning_rate": 1.0812069918077844e-06, "loss": 0.1065, "num_input_tokens_seen": 9367768, "step": 34925 }, { "epoch": 9.155963302752294, "grad_norm": 2.9401347637176514, "learning_rate": 1.0778823240622043e-06, "loss": 0.0997, "num_input_tokens_seen": 9368872, "step": 34930 }, { "epoch": 9.157273918741808, "grad_norm": 2.312685966491699, "learning_rate": 1.0745626631745193e-06, "loss": 0.2265, "num_input_tokens_seen": 9371912, "step": 34935 }, { "epoch": 9.158584534731324, "grad_norm": 3.383019208908081, "learning_rate": 1.0712480098395289e-06, "loss": 0.1502, "num_input_tokens_seen": 9373624, "step": 34940 }, { "epoch": 9.159895150720839, "grad_norm": 0.6052565574645996, "learning_rate": 1.0679383647509839e-06, "loss": 0.0645, "num_input_tokens_seen": 9374984, "step": 34945 }, { "epoch": 9.161205766710355, "grad_norm": 1.555254340171814, "learning_rate": 1.0646337286015856e-06, "loss": 0.1084, "num_input_tokens_seen": 9375880, "step": 34950 }, { "epoch": 9.162516382699868, "grad_norm": 3.212184429168701, "learning_rate": 1.0613341020829948e-06, "loss": 0.1666, "num_input_tokens_seen": 9377112, "step": 34955 }, { "epoch": 9.163826998689384, "grad_norm": 2.1742162704467773, "learning_rate": 1.0580394858858117e-06, "loss": 0.1074, "num_input_tokens_seen": 9378120, "step": 34960 }, { "epoch": 9.1651376146789, "grad_norm": 1.1215310096740723, "learning_rate": 1.054749880699593e-06, "loss": 0.0691, "num_input_tokens_seen": 9379192, "step": 34965 }, { "epoch": 9.166448230668415, "grad_norm": 2.540432929992676, "learning_rate": 1.051465287212855e-06, "loss": 0.1478, "num_input_tokens_seen": 9380312, "step": 34970 }, { "epoch": 9.167758846657929, "grad_norm": 0.8753987550735474, "learning_rate": 1.04818570611305e-06, "loss": 0.1261, "num_input_tokens_seen": 9381976, "step": 34975 }, { "epoch": 9.169069462647444, "grad_norm": 3.6233606338500977, "learning_rate": 1.0449111380865906e-06, "loss": 0.0601, "num_input_tokens_seen": 9383000, "step": 34980 }, { "epoch": 9.17038007863696, "grad_norm": 1.9084573984146118, "learning_rate": 1.0416415838188393e-06, "loss": 0.2446, "num_input_tokens_seen": 9384264, "step": 34985 }, { "epoch": 9.171690694626474, "grad_norm": 0.997616171836853, "learning_rate": 1.038377043994107e-06, "loss": 0.0356, "num_input_tokens_seen": 9385640, "step": 34990 }, { "epoch": 9.17300131061599, "grad_norm": 2.1284561157226562, "learning_rate": 1.0351175192956585e-06, "loss": 0.167, "num_input_tokens_seen": 9387160, "step": 34995 }, { "epoch": 9.174311926605505, "grad_norm": 4.750766754150391, "learning_rate": 1.0318630104057037e-06, "loss": 0.1946, "num_input_tokens_seen": 9388248, "step": 35000 }, { "epoch": 9.17562254259502, "grad_norm": 0.3163827359676361, "learning_rate": 1.0286135180054114e-06, "loss": 0.0751, "num_input_tokens_seen": 9389480, "step": 35005 }, { "epoch": 9.176933158584534, "grad_norm": 3.6484549045562744, "learning_rate": 1.0253690427748874e-06, "loss": 0.2073, "num_input_tokens_seen": 9390856, "step": 35010 }, { "epoch": 9.17824377457405, "grad_norm": 3.369020700454712, "learning_rate": 1.0221295853931973e-06, "loss": 0.0787, "num_input_tokens_seen": 9392488, "step": 35015 }, { "epoch": 9.179554390563565, "grad_norm": 2.2646327018737793, "learning_rate": 1.0188951465383511e-06, "loss": 0.1462, "num_input_tokens_seen": 9394072, "step": 35020 }, { "epoch": 9.18086500655308, "grad_norm": 1.1075454950332642, "learning_rate": 1.015665726887316e-06, "loss": 0.0629, "num_input_tokens_seen": 9395288, "step": 35025 }, { "epoch": 9.182175622542594, "grad_norm": 3.652587652206421, "learning_rate": 1.0124413271160038e-06, "loss": 0.1367, "num_input_tokens_seen": 9396472, "step": 35030 }, { "epoch": 9.18348623853211, "grad_norm": 7.045922756195068, "learning_rate": 1.0092219478992776e-06, "loss": 0.1133, "num_input_tokens_seen": 9397832, "step": 35035 }, { "epoch": 9.184796854521625, "grad_norm": 2.722654104232788, "learning_rate": 1.0060075899109427e-06, "loss": 0.1684, "num_input_tokens_seen": 9399256, "step": 35040 }, { "epoch": 9.186107470511141, "grad_norm": 1.2703063488006592, "learning_rate": 1.0027982538237612e-06, "loss": 0.1562, "num_input_tokens_seen": 9401176, "step": 35045 }, { "epoch": 9.187418086500655, "grad_norm": 3.0519397258758545, "learning_rate": 9.995939403094402e-07, "loss": 0.1126, "num_input_tokens_seen": 9402280, "step": 35050 }, { "epoch": 9.18872870249017, "grad_norm": 2.1225850582122803, "learning_rate": 9.963946500386374e-07, "loss": 0.128, "num_input_tokens_seen": 9403352, "step": 35055 }, { "epoch": 9.190039318479686, "grad_norm": 3.214909315109253, "learning_rate": 9.932003836809616e-07, "loss": 0.2393, "num_input_tokens_seen": 9404568, "step": 35060 }, { "epoch": 9.191349934469201, "grad_norm": 1.4211022853851318, "learning_rate": 9.900111419049619e-07, "loss": 0.0635, "num_input_tokens_seen": 9405496, "step": 35065 }, { "epoch": 9.192660550458715, "grad_norm": 3.099222421646118, "learning_rate": 9.868269253781453e-07, "loss": 0.1395, "num_input_tokens_seen": 9406904, "step": 35070 }, { "epoch": 9.19397116644823, "grad_norm": 2.859208583831787, "learning_rate": 9.836477347669626e-07, "loss": 0.1554, "num_input_tokens_seen": 9408200, "step": 35075 }, { "epoch": 9.195281782437746, "grad_norm": 2.1937294006347656, "learning_rate": 9.804735707368118e-07, "loss": 0.1369, "num_input_tokens_seen": 9409448, "step": 35080 }, { "epoch": 9.19659239842726, "grad_norm": 3.606907606124878, "learning_rate": 9.773044339520392e-07, "loss": 0.1549, "num_input_tokens_seen": 9410888, "step": 35085 }, { "epoch": 9.197903014416775, "grad_norm": 2.8208272457122803, "learning_rate": 9.741403250759424e-07, "loss": 0.0932, "num_input_tokens_seen": 9412008, "step": 35090 }, { "epoch": 9.199213630406291, "grad_norm": 2.0508244037628174, "learning_rate": 9.709812447707606e-07, "loss": 0.1668, "num_input_tokens_seen": 9413304, "step": 35095 }, { "epoch": 9.200524246395807, "grad_norm": 2.9439380168914795, "learning_rate": 9.678271936976847e-07, "loss": 0.1223, "num_input_tokens_seen": 9414440, "step": 35100 }, { "epoch": 9.20183486238532, "grad_norm": 2.307053327560425, "learning_rate": 9.64678172516853e-07, "loss": 0.1319, "num_input_tokens_seen": 9415672, "step": 35105 }, { "epoch": 9.203145478374836, "grad_norm": 1.9681482315063477, "learning_rate": 9.615341818873496e-07, "loss": 0.0595, "num_input_tokens_seen": 9416872, "step": 35110 }, { "epoch": 9.204456094364351, "grad_norm": 4.067656517028809, "learning_rate": 9.583952224672094e-07, "loss": 0.1512, "num_input_tokens_seen": 9417976, "step": 35115 }, { "epoch": 9.205766710353867, "grad_norm": 1.2458912134170532, "learning_rate": 9.552612949134094e-07, "loss": 0.1381, "num_input_tokens_seen": 9419496, "step": 35120 }, { "epoch": 9.20707732634338, "grad_norm": 1.761694073677063, "learning_rate": 9.521323998818694e-07, "loss": 0.1018, "num_input_tokens_seen": 9420536, "step": 35125 }, { "epoch": 9.208387942332896, "grad_norm": 3.221498727798462, "learning_rate": 9.490085380274682e-07, "loss": 0.1439, "num_input_tokens_seen": 9422248, "step": 35130 }, { "epoch": 9.209698558322412, "grad_norm": 1.9054224491119385, "learning_rate": 9.45889710004022e-07, "loss": 0.0923, "num_input_tokens_seen": 9423176, "step": 35135 }, { "epoch": 9.211009174311927, "grad_norm": 2.0755515098571777, "learning_rate": 9.427759164642974e-07, "loss": 0.13, "num_input_tokens_seen": 9424360, "step": 35140 }, { "epoch": 9.212319790301441, "grad_norm": 1.3477365970611572, "learning_rate": 9.396671580600064e-07, "loss": 0.0571, "num_input_tokens_seen": 9425384, "step": 35145 }, { "epoch": 9.213630406290957, "grad_norm": 1.3481425046920776, "learning_rate": 9.36563435441809e-07, "loss": 0.087, "num_input_tokens_seen": 9427304, "step": 35150 }, { "epoch": 9.214941022280472, "grad_norm": 2.320028781890869, "learning_rate": 9.334647492593051e-07, "loss": 0.1045, "num_input_tokens_seen": 9428472, "step": 35155 }, { "epoch": 9.216251638269988, "grad_norm": 2.445328950881958, "learning_rate": 9.303711001610454e-07, "loss": 0.2124, "num_input_tokens_seen": 9429576, "step": 35160 }, { "epoch": 9.217562254259501, "grad_norm": 0.9546011090278625, "learning_rate": 9.27282488794523e-07, "loss": 0.0549, "num_input_tokens_seen": 9430504, "step": 35165 }, { "epoch": 9.218872870249017, "grad_norm": 2.033189058303833, "learning_rate": 9.24198915806182e-07, "loss": 0.1368, "num_input_tokens_seen": 9432408, "step": 35170 }, { "epoch": 9.220183486238533, "grad_norm": 1.7312573194503784, "learning_rate": 9.211203818414088e-07, "loss": 0.103, "num_input_tokens_seen": 9433592, "step": 35175 }, { "epoch": 9.221494102228046, "grad_norm": 3.0195870399475098, "learning_rate": 9.180468875445352e-07, "loss": 0.1064, "num_input_tokens_seen": 9434648, "step": 35180 }, { "epoch": 9.222804718217562, "grad_norm": 1.9414982795715332, "learning_rate": 9.149784335588357e-07, "loss": 0.1038, "num_input_tokens_seen": 9436152, "step": 35185 }, { "epoch": 9.224115334207077, "grad_norm": 4.783400535583496, "learning_rate": 9.119150205265326e-07, "loss": 0.097, "num_input_tokens_seen": 9437336, "step": 35190 }, { "epoch": 9.225425950196593, "grad_norm": 1.3561285734176636, "learning_rate": 9.088566490888017e-07, "loss": 0.1104, "num_input_tokens_seen": 9438728, "step": 35195 }, { "epoch": 9.226736566186107, "grad_norm": 1.245880126953125, "learning_rate": 9.058033198857424e-07, "loss": 0.1539, "num_input_tokens_seen": 9440216, "step": 35200 }, { "epoch": 9.228047182175622, "grad_norm": 3.6632115840911865, "learning_rate": 9.027550335564184e-07, "loss": 0.1086, "num_input_tokens_seen": 9441208, "step": 35205 }, { "epoch": 9.229357798165138, "grad_norm": 1.8943607807159424, "learning_rate": 8.997117907388275e-07, "loss": 0.1238, "num_input_tokens_seen": 9442152, "step": 35210 }, { "epoch": 9.230668414154653, "grad_norm": 0.5480180382728577, "learning_rate": 8.966735920699187e-07, "loss": 0.1082, "num_input_tokens_seen": 9443528, "step": 35215 }, { "epoch": 9.231979030144167, "grad_norm": 0.4066224694252014, "learning_rate": 8.936404381855806e-07, "loss": 0.0832, "num_input_tokens_seen": 9444488, "step": 35220 }, { "epoch": 9.233289646133683, "grad_norm": 3.1085777282714844, "learning_rate": 8.906123297206442e-07, "loss": 0.1057, "num_input_tokens_seen": 9445720, "step": 35225 }, { "epoch": 9.234600262123198, "grad_norm": 1.7457513809204102, "learning_rate": 8.875892673088887e-07, "loss": 0.1465, "num_input_tokens_seen": 9447064, "step": 35230 }, { "epoch": 9.235910878112714, "grad_norm": 2.272001266479492, "learning_rate": 8.845712515830385e-07, "loss": 0.1657, "num_input_tokens_seen": 9448392, "step": 35235 }, { "epoch": 9.237221494102227, "grad_norm": 3.0765938758850098, "learning_rate": 8.815582831747576e-07, "loss": 0.1745, "num_input_tokens_seen": 9449912, "step": 35240 }, { "epoch": 9.238532110091743, "grad_norm": 1.5210216045379639, "learning_rate": 8.785503627146501e-07, "loss": 0.2851, "num_input_tokens_seen": 9451560, "step": 35245 }, { "epoch": 9.239842726081259, "grad_norm": 1.3927600383758545, "learning_rate": 8.755474908322764e-07, "loss": 0.1834, "num_input_tokens_seen": 9452984, "step": 35250 }, { "epoch": 9.241153342070774, "grad_norm": 3.1837568283081055, "learning_rate": 8.725496681561279e-07, "loss": 0.218, "num_input_tokens_seen": 9454568, "step": 35255 }, { "epoch": 9.242463958060288, "grad_norm": 3.7545087337493896, "learning_rate": 8.695568953136446e-07, "loss": 0.1905, "num_input_tokens_seen": 9455880, "step": 35260 }, { "epoch": 9.243774574049803, "grad_norm": 2.79842472076416, "learning_rate": 8.665691729312115e-07, "loss": 0.1849, "num_input_tokens_seen": 9457624, "step": 35265 }, { "epoch": 9.245085190039319, "grad_norm": 1.729865550994873, "learning_rate": 8.635865016341477e-07, "loss": 0.0897, "num_input_tokens_seen": 9459352, "step": 35270 }, { "epoch": 9.246395806028833, "grad_norm": 2.923044443130493, "learning_rate": 8.606088820467318e-07, "loss": 0.0769, "num_input_tokens_seen": 9460328, "step": 35275 }, { "epoch": 9.247706422018348, "grad_norm": 1.3544517755508423, "learning_rate": 8.576363147921596e-07, "loss": 0.2808, "num_input_tokens_seen": 9461880, "step": 35280 }, { "epoch": 9.249017038007864, "grad_norm": 0.9139845371246338, "learning_rate": 8.546688004925946e-07, "loss": 0.0622, "num_input_tokens_seen": 9463336, "step": 35285 }, { "epoch": 9.25032765399738, "grad_norm": 2.3416926860809326, "learning_rate": 8.517063397691288e-07, "loss": 0.1228, "num_input_tokens_seen": 9465400, "step": 35290 }, { "epoch": 9.251638269986893, "grad_norm": 3.6561808586120605, "learning_rate": 8.487489332418025e-07, "loss": 0.1586, "num_input_tokens_seen": 9467048, "step": 35295 }, { "epoch": 9.252948885976409, "grad_norm": 1.9566045999526978, "learning_rate": 8.457965815295926e-07, "loss": 0.1023, "num_input_tokens_seen": 9468104, "step": 35300 }, { "epoch": 9.254259501965924, "grad_norm": 0.9611411094665527, "learning_rate": 8.428492852504216e-07, "loss": 0.1474, "num_input_tokens_seen": 9469128, "step": 35305 }, { "epoch": 9.25557011795544, "grad_norm": 3.363982915878296, "learning_rate": 8.399070450211599e-07, "loss": 0.0923, "num_input_tokens_seen": 9470088, "step": 35310 }, { "epoch": 9.256880733944953, "grad_norm": 2.0684008598327637, "learning_rate": 8.369698614576038e-07, "loss": 0.1324, "num_input_tokens_seen": 9471208, "step": 35315 }, { "epoch": 9.258191349934469, "grad_norm": 1.3858416080474854, "learning_rate": 8.340377351745088e-07, "loss": 0.3219, "num_input_tokens_seen": 9472728, "step": 35320 }, { "epoch": 9.259501965923985, "grad_norm": 0.07065407186746597, "learning_rate": 8.311106667855562e-07, "loss": 0.0927, "num_input_tokens_seen": 9474008, "step": 35325 }, { "epoch": 9.2608125819135, "grad_norm": 1.6445037126541138, "learning_rate": 8.281886569033836e-07, "loss": 0.1399, "num_input_tokens_seen": 9475528, "step": 35330 }, { "epoch": 9.262123197903014, "grad_norm": 3.551419258117676, "learning_rate": 8.252717061395576e-07, "loss": 0.4019, "num_input_tokens_seen": 9476856, "step": 35335 }, { "epoch": 9.26343381389253, "grad_norm": 2.354519844055176, "learning_rate": 8.223598151045952e-07, "loss": 0.0696, "num_input_tokens_seen": 9477880, "step": 35340 }, { "epoch": 9.264744429882045, "grad_norm": 4.283883571624756, "learning_rate": 8.194529844079451e-07, "loss": 0.201, "num_input_tokens_seen": 9479304, "step": 35345 }, { "epoch": 9.26605504587156, "grad_norm": 1.6445306539535522, "learning_rate": 8.16551214658004e-07, "loss": 0.1267, "num_input_tokens_seen": 9480280, "step": 35350 }, { "epoch": 9.267365661861074, "grad_norm": 0.9883849620819092, "learning_rate": 8.136545064621137e-07, "loss": 0.1633, "num_input_tokens_seen": 9481496, "step": 35355 }, { "epoch": 9.26867627785059, "grad_norm": 3.9097368717193604, "learning_rate": 8.107628604265366e-07, "loss": 0.0994, "num_input_tokens_seen": 9483144, "step": 35360 }, { "epoch": 9.269986893840105, "grad_norm": 3.521143674850464, "learning_rate": 8.078762771564996e-07, "loss": 0.1237, "num_input_tokens_seen": 9484616, "step": 35365 }, { "epoch": 9.271297509829619, "grad_norm": 2.7868709564208984, "learning_rate": 8.049947572561584e-07, "loss": 0.0909, "num_input_tokens_seen": 9485800, "step": 35370 }, { "epoch": 9.272608125819135, "grad_norm": 1.9478776454925537, "learning_rate": 8.021183013286082e-07, "loss": 0.0871, "num_input_tokens_seen": 9486984, "step": 35375 }, { "epoch": 9.27391874180865, "grad_norm": 2.9339468479156494, "learning_rate": 7.99246909975887e-07, "loss": 0.2082, "num_input_tokens_seen": 9488632, "step": 35380 }, { "epoch": 9.275229357798166, "grad_norm": 1.570267677307129, "learning_rate": 7.963805837989724e-07, "loss": 0.2168, "num_input_tokens_seen": 9489880, "step": 35385 }, { "epoch": 9.27653997378768, "grad_norm": 1.7526637315750122, "learning_rate": 7.935193233977845e-07, "loss": 0.1017, "num_input_tokens_seen": 9491032, "step": 35390 }, { "epoch": 9.277850589777195, "grad_norm": 2.6688168048858643, "learning_rate": 7.906631293711719e-07, "loss": 0.1099, "num_input_tokens_seen": 9492520, "step": 35395 }, { "epoch": 9.27916120576671, "grad_norm": 0.7037972211837769, "learning_rate": 7.878120023169344e-07, "loss": 0.0886, "num_input_tokens_seen": 9494040, "step": 35400 }, { "epoch": 9.280471821756226, "grad_norm": 2.6732869148254395, "learning_rate": 7.849659428318113e-07, "loss": 0.129, "num_input_tokens_seen": 9495048, "step": 35405 }, { "epoch": 9.28178243774574, "grad_norm": 1.6994290351867676, "learning_rate": 7.821249515114787e-07, "loss": 0.1485, "num_input_tokens_seen": 9496776, "step": 35410 }, { "epoch": 9.283093053735255, "grad_norm": 2.587265729904175, "learning_rate": 7.792890289505444e-07, "loss": 0.2719, "num_input_tokens_seen": 9498040, "step": 35415 }, { "epoch": 9.284403669724771, "grad_norm": 1.334954023361206, "learning_rate": 7.764581757425642e-07, "loss": 0.0649, "num_input_tokens_seen": 9499608, "step": 35420 }, { "epoch": 9.285714285714286, "grad_norm": 3.607517719268799, "learning_rate": 7.736323924800365e-07, "loss": 0.2314, "num_input_tokens_seen": 9500984, "step": 35425 }, { "epoch": 9.2870249017038, "grad_norm": 1.7252888679504395, "learning_rate": 7.708116797543907e-07, "loss": 0.0834, "num_input_tokens_seen": 9502184, "step": 35430 }, { "epoch": 9.288335517693316, "grad_norm": 1.9360830783843994, "learning_rate": 7.679960381559936e-07, "loss": 0.2034, "num_input_tokens_seen": 9503624, "step": 35435 }, { "epoch": 9.289646133682831, "grad_norm": 1.6521998643875122, "learning_rate": 7.651854682741572e-07, "loss": 0.1565, "num_input_tokens_seen": 9505176, "step": 35440 }, { "epoch": 9.290956749672347, "grad_norm": 0.7134284377098083, "learning_rate": 7.623799706971274e-07, "loss": 0.1834, "num_input_tokens_seen": 9506504, "step": 35445 }, { "epoch": 9.29226736566186, "grad_norm": 2.4747676849365234, "learning_rate": 7.595795460120903e-07, "loss": 0.0962, "num_input_tokens_seen": 9507448, "step": 35450 }, { "epoch": 9.293577981651376, "grad_norm": 1.8124415874481201, "learning_rate": 7.567841948051685e-07, "loss": 0.1529, "num_input_tokens_seen": 9508760, "step": 35455 }, { "epoch": 9.294888597640892, "grad_norm": 2.9932544231414795, "learning_rate": 7.539939176614247e-07, "loss": 0.1505, "num_input_tokens_seen": 9510072, "step": 35460 }, { "epoch": 9.296199213630405, "grad_norm": 1.4923834800720215, "learning_rate": 7.512087151648639e-07, "loss": 0.0983, "num_input_tokens_seen": 9511288, "step": 35465 }, { "epoch": 9.297509829619921, "grad_norm": 2.061221122741699, "learning_rate": 7.484285878984171e-07, "loss": 0.1978, "num_input_tokens_seen": 9512408, "step": 35470 }, { "epoch": 9.298820445609437, "grad_norm": 2.112312078475952, "learning_rate": 7.45653536443966e-07, "loss": 0.1101, "num_input_tokens_seen": 9513752, "step": 35475 }, { "epoch": 9.300131061598952, "grad_norm": 1.8693066835403442, "learning_rate": 7.428835613823182e-07, "loss": 0.2371, "num_input_tokens_seen": 9515000, "step": 35480 }, { "epoch": 9.301441677588466, "grad_norm": 1.6978106498718262, "learning_rate": 7.401186632932294e-07, "loss": 0.1024, "num_input_tokens_seen": 9516136, "step": 35485 }, { "epoch": 9.302752293577981, "grad_norm": 1.9708462953567505, "learning_rate": 7.373588427553869e-07, "loss": 0.1386, "num_input_tokens_seen": 9517288, "step": 35490 }, { "epoch": 9.304062909567497, "grad_norm": 2.46830415725708, "learning_rate": 7.346041003464172e-07, "loss": 0.1331, "num_input_tokens_seen": 9518536, "step": 35495 }, { "epoch": 9.305373525557012, "grad_norm": 2.280346393585205, "learning_rate": 7.318544366428814e-07, "loss": 0.1108, "num_input_tokens_seen": 9519608, "step": 35500 }, { "epoch": 9.306684141546526, "grad_norm": 2.170708656311035, "learning_rate": 7.291098522202777e-07, "loss": 0.1212, "num_input_tokens_seen": 9520808, "step": 35505 }, { "epoch": 9.307994757536042, "grad_norm": 1.5391058921813965, "learning_rate": 7.263703476530492e-07, "loss": 0.133, "num_input_tokens_seen": 9522008, "step": 35510 }, { "epoch": 9.309305373525557, "grad_norm": 3.188790798187256, "learning_rate": 7.236359235145624e-07, "loss": 0.1994, "num_input_tokens_seen": 9523288, "step": 35515 }, { "epoch": 9.310615989515073, "grad_norm": 2.3026649951934814, "learning_rate": 7.209065803771315e-07, "loss": 0.1438, "num_input_tokens_seen": 9525064, "step": 35520 }, { "epoch": 9.311926605504587, "grad_norm": 3.5376415252685547, "learning_rate": 7.181823188120024e-07, "loss": 0.1555, "num_input_tokens_seen": 9526120, "step": 35525 }, { "epoch": 9.313237221494102, "grad_norm": 2.5073935985565186, "learning_rate": 7.154631393893552e-07, "loss": 0.1646, "num_input_tokens_seen": 9527512, "step": 35530 }, { "epoch": 9.314547837483618, "grad_norm": 1.8736308813095093, "learning_rate": 7.127490426783123e-07, "loss": 0.2143, "num_input_tokens_seen": 9528600, "step": 35535 }, { "epoch": 9.315858453473133, "grad_norm": 2.0287833213806152, "learning_rate": 7.100400292469333e-07, "loss": 0.2043, "num_input_tokens_seen": 9529800, "step": 35540 }, { "epoch": 9.317169069462647, "grad_norm": 1.8963514566421509, "learning_rate": 7.073360996622064e-07, "loss": 0.1104, "num_input_tokens_seen": 9530968, "step": 35545 }, { "epoch": 9.318479685452163, "grad_norm": 1.4159948825836182, "learning_rate": 7.046372544900592e-07, "loss": 0.0827, "num_input_tokens_seen": 9532168, "step": 35550 }, { "epoch": 9.319790301441678, "grad_norm": 1.1262009143829346, "learning_rate": 7.01943494295354e-07, "loss": 0.155, "num_input_tokens_seen": 9533640, "step": 35555 }, { "epoch": 9.321100917431192, "grad_norm": 3.022366762161255, "learning_rate": 6.992548196418924e-07, "loss": 0.0823, "num_input_tokens_seen": 9534632, "step": 35560 }, { "epoch": 9.322411533420707, "grad_norm": 1.4424632787704468, "learning_rate": 6.965712310924078e-07, "loss": 0.1137, "num_input_tokens_seen": 9536040, "step": 35565 }, { "epoch": 9.323722149410223, "grad_norm": 1.6763579845428467, "learning_rate": 6.9389272920857e-07, "loss": 0.0873, "num_input_tokens_seen": 9537192, "step": 35570 }, { "epoch": 9.325032765399738, "grad_norm": 1.4567729234695435, "learning_rate": 6.912193145509893e-07, "loss": 0.0737, "num_input_tokens_seen": 9538680, "step": 35575 }, { "epoch": 9.326343381389252, "grad_norm": 1.8525739908218384, "learning_rate": 6.885509876792012e-07, "loss": 0.0689, "num_input_tokens_seen": 9539656, "step": 35580 }, { "epoch": 9.327653997378768, "grad_norm": 6.650676727294922, "learning_rate": 6.858877491516869e-07, "loss": 0.1105, "num_input_tokens_seen": 9540968, "step": 35585 }, { "epoch": 9.328964613368283, "grad_norm": 2.2462942600250244, "learning_rate": 6.832295995258531e-07, "loss": 0.2372, "num_input_tokens_seen": 9542248, "step": 35590 }, { "epoch": 9.330275229357799, "grad_norm": 2.4293832778930664, "learning_rate": 6.805765393580493e-07, "loss": 0.1463, "num_input_tokens_seen": 9543416, "step": 35595 }, { "epoch": 9.331585845347313, "grad_norm": 1.7707767486572266, "learning_rate": 6.779285692035536e-07, "loss": 0.165, "num_input_tokens_seen": 9545272, "step": 35600 }, { "epoch": 9.332896461336828, "grad_norm": 1.721814513206482, "learning_rate": 6.752856896165866e-07, "loss": 0.1597, "num_input_tokens_seen": 9546488, "step": 35605 }, { "epoch": 9.334207077326344, "grad_norm": 2.441236972808838, "learning_rate": 6.726479011502917e-07, "loss": 0.0795, "num_input_tokens_seen": 9547496, "step": 35610 }, { "epoch": 9.33551769331586, "grad_norm": 3.3494908809661865, "learning_rate": 6.700152043567554e-07, "loss": 0.114, "num_input_tokens_seen": 9548856, "step": 35615 }, { "epoch": 9.336828309305373, "grad_norm": 3.7816050052642822, "learning_rate": 6.673875997870005e-07, "loss": 0.1159, "num_input_tokens_seen": 9550280, "step": 35620 }, { "epoch": 9.338138925294889, "grad_norm": 1.359512448310852, "learning_rate": 6.64765087990979e-07, "loss": 0.1715, "num_input_tokens_seen": 9551864, "step": 35625 }, { "epoch": 9.339449541284404, "grad_norm": 3.2490220069885254, "learning_rate": 6.621476695175738e-07, "loss": 0.1171, "num_input_tokens_seen": 9553032, "step": 35630 }, { "epoch": 9.34076015727392, "grad_norm": 1.1405823230743408, "learning_rate": 6.595353449146108e-07, "loss": 0.2346, "num_input_tokens_seen": 9554520, "step": 35635 }, { "epoch": 9.342070773263433, "grad_norm": 2.121864080429077, "learning_rate": 6.569281147288414e-07, "loss": 0.2527, "num_input_tokens_seen": 9556424, "step": 35640 }, { "epoch": 9.343381389252949, "grad_norm": 1.98110032081604, "learning_rate": 6.54325979505957e-07, "loss": 0.1813, "num_input_tokens_seen": 9558392, "step": 35645 }, { "epoch": 9.344692005242464, "grad_norm": 6.446577548980713, "learning_rate": 6.517289397905746e-07, "loss": 0.1414, "num_input_tokens_seen": 9559704, "step": 35650 }, { "epoch": 9.346002621231978, "grad_norm": 2.6486828327178955, "learning_rate": 6.491369961262566e-07, "loss": 0.2222, "num_input_tokens_seen": 9561288, "step": 35655 }, { "epoch": 9.347313237221494, "grad_norm": 4.59937047958374, "learning_rate": 6.465501490554914e-07, "loss": 0.1929, "num_input_tokens_seen": 9562440, "step": 35660 }, { "epoch": 9.34862385321101, "grad_norm": 2.4526331424713135, "learning_rate": 6.439683991197043e-07, "loss": 0.1088, "num_input_tokens_seen": 9563880, "step": 35665 }, { "epoch": 9.349934469200525, "grad_norm": 1.8858774900436401, "learning_rate": 6.413917468592434e-07, "loss": 0.1752, "num_input_tokens_seen": 9565032, "step": 35670 }, { "epoch": 9.351245085190039, "grad_norm": 1.8659309148788452, "learning_rate": 6.388201928134025e-07, "loss": 0.0719, "num_input_tokens_seen": 9566008, "step": 35675 }, { "epoch": 9.352555701179554, "grad_norm": 1.5086474418640137, "learning_rate": 6.362537375204009e-07, "loss": 0.2382, "num_input_tokens_seen": 9567336, "step": 35680 }, { "epoch": 9.35386631716907, "grad_norm": 2.8451857566833496, "learning_rate": 6.336923815173951e-07, "loss": 0.0771, "num_input_tokens_seen": 9568312, "step": 35685 }, { "epoch": 9.355176933158585, "grad_norm": 2.5037941932678223, "learning_rate": 6.31136125340473e-07, "loss": 0.1928, "num_input_tokens_seen": 9569816, "step": 35690 }, { "epoch": 9.356487549148099, "grad_norm": 3.3445065021514893, "learning_rate": 6.285849695246537e-07, "loss": 0.1062, "num_input_tokens_seen": 9571224, "step": 35695 }, { "epoch": 9.357798165137615, "grad_norm": 0.9625251889228821, "learning_rate": 6.260389146038881e-07, "loss": 0.1598, "num_input_tokens_seen": 9572360, "step": 35700 }, { "epoch": 9.35910878112713, "grad_norm": 0.31549370288848877, "learning_rate": 6.234979611110637e-07, "loss": 0.1306, "num_input_tokens_seen": 9573656, "step": 35705 }, { "epoch": 9.360419397116646, "grad_norm": 2.778585195541382, "learning_rate": 6.209621095779993e-07, "loss": 0.0632, "num_input_tokens_seen": 9574776, "step": 35710 }, { "epoch": 9.36173001310616, "grad_norm": 2.861632823944092, "learning_rate": 6.184313605354375e-07, "loss": 0.1472, "num_input_tokens_seen": 9576248, "step": 35715 }, { "epoch": 9.363040629095675, "grad_norm": 1.9534103870391846, "learning_rate": 6.159057145130653e-07, "loss": 0.1474, "num_input_tokens_seen": 9577432, "step": 35720 }, { "epoch": 9.36435124508519, "grad_norm": 0.8811678886413574, "learning_rate": 6.133851720394962e-07, "loss": 0.0471, "num_input_tokens_seen": 9578552, "step": 35725 }, { "epoch": 9.365661861074706, "grad_norm": 2.499403715133667, "learning_rate": 6.10869733642272e-07, "loss": 0.2357, "num_input_tokens_seen": 9580024, "step": 35730 }, { "epoch": 9.36697247706422, "grad_norm": 0.6047438979148865, "learning_rate": 6.083593998478715e-07, "loss": 0.1009, "num_input_tokens_seen": 9581496, "step": 35735 }, { "epoch": 9.368283093053735, "grad_norm": 2.5307822227478027, "learning_rate": 6.058541711817023e-07, "loss": 0.1216, "num_input_tokens_seen": 9582440, "step": 35740 }, { "epoch": 9.36959370904325, "grad_norm": 1.2048028707504272, "learning_rate": 6.033540481681088e-07, "loss": 0.101, "num_input_tokens_seen": 9583896, "step": 35745 }, { "epoch": 9.370904325032765, "grad_norm": 3.0305063724517822, "learning_rate": 6.008590313303586e-07, "loss": 0.1489, "num_input_tokens_seen": 9585304, "step": 35750 }, { "epoch": 9.37221494102228, "grad_norm": 2.428302049636841, "learning_rate": 5.983691211906506e-07, "loss": 0.1538, "num_input_tokens_seen": 9586488, "step": 35755 }, { "epoch": 9.373525557011796, "grad_norm": 4.620051860809326, "learning_rate": 5.958843182701263e-07, "loss": 0.1763, "num_input_tokens_seen": 9587880, "step": 35760 }, { "epoch": 9.374836173001311, "grad_norm": 4.881377220153809, "learning_rate": 5.934046230888418e-07, "loss": 0.1175, "num_input_tokens_seen": 9588920, "step": 35765 }, { "epoch": 9.376146788990825, "grad_norm": 5.710839748382568, "learning_rate": 5.909300361658015e-07, "loss": 0.1124, "num_input_tokens_seen": 9589976, "step": 35770 }, { "epoch": 9.37745740498034, "grad_norm": 1.2393168210983276, "learning_rate": 5.8846055801893e-07, "loss": 0.0959, "num_input_tokens_seen": 9591704, "step": 35775 }, { "epoch": 9.378768020969856, "grad_norm": 3.9050543308258057, "learning_rate": 5.859961891650861e-07, "loss": 0.2196, "num_input_tokens_seen": 9593000, "step": 35780 }, { "epoch": 9.380078636959372, "grad_norm": 4.0221757888793945, "learning_rate": 5.835369301200516e-07, "loss": 0.1422, "num_input_tokens_seen": 9594440, "step": 35785 }, { "epoch": 9.381389252948885, "grad_norm": 2.1574180126190186, "learning_rate": 5.810827813985509e-07, "loss": 0.0781, "num_input_tokens_seen": 9595672, "step": 35790 }, { "epoch": 9.382699868938401, "grad_norm": 2.0370004177093506, "learning_rate": 5.786337435142314e-07, "loss": 0.1189, "num_input_tokens_seen": 9597112, "step": 35795 }, { "epoch": 9.384010484927916, "grad_norm": 0.020874088630080223, "learning_rate": 5.761898169796721e-07, "loss": 0.0633, "num_input_tokens_seen": 9598456, "step": 35800 }, { "epoch": 9.385321100917432, "grad_norm": 2.5135414600372314, "learning_rate": 5.73751002306383e-07, "loss": 0.2168, "num_input_tokens_seen": 9599608, "step": 35805 }, { "epoch": 9.386631716906946, "grad_norm": 3.5723726749420166, "learning_rate": 5.713173000048033e-07, "loss": 0.1245, "num_input_tokens_seen": 9601000, "step": 35810 }, { "epoch": 9.387942332896461, "grad_norm": 1.6967244148254395, "learning_rate": 5.688887105843032e-07, "loss": 0.1228, "num_input_tokens_seen": 9602232, "step": 35815 }, { "epoch": 9.389252948885977, "grad_norm": 2.9806692600250244, "learning_rate": 5.664652345531846e-07, "loss": 0.1257, "num_input_tokens_seen": 9603400, "step": 35820 }, { "epoch": 9.390563564875492, "grad_norm": 2.567641496658325, "learning_rate": 5.640468724186721e-07, "loss": 0.1649, "num_input_tokens_seen": 9604648, "step": 35825 }, { "epoch": 9.391874180865006, "grad_norm": 2.18001389503479, "learning_rate": 5.616336246869275e-07, "loss": 0.0858, "num_input_tokens_seen": 9606008, "step": 35830 }, { "epoch": 9.393184796854522, "grad_norm": 3.3440420627593994, "learning_rate": 5.592254918630385e-07, "loss": 0.1968, "num_input_tokens_seen": 9607304, "step": 35835 }, { "epoch": 9.394495412844037, "grad_norm": 4.126683712005615, "learning_rate": 5.568224744510242e-07, "loss": 0.1554, "num_input_tokens_seen": 9608360, "step": 35840 }, { "epoch": 9.395806028833551, "grad_norm": 1.623806118965149, "learning_rate": 5.544245729538322e-07, "loss": 0.1332, "num_input_tokens_seen": 9609720, "step": 35845 }, { "epoch": 9.397116644823067, "grad_norm": 5.862619876861572, "learning_rate": 5.520317878733389e-07, "loss": 0.1039, "num_input_tokens_seen": 9610536, "step": 35850 }, { "epoch": 9.398427260812582, "grad_norm": 1.7068638801574707, "learning_rate": 5.496441197103463e-07, "loss": 0.1462, "num_input_tokens_seen": 9611832, "step": 35855 }, { "epoch": 9.399737876802098, "grad_norm": 1.2046786546707153, "learning_rate": 5.472615689645965e-07, "loss": 0.1491, "num_input_tokens_seen": 9613416, "step": 35860 }, { "epoch": 9.401048492791611, "grad_norm": 1.4225417375564575, "learning_rate": 5.44884136134749e-07, "loss": 0.1124, "num_input_tokens_seen": 9614296, "step": 35865 }, { "epoch": 9.402359108781127, "grad_norm": 2.509650707244873, "learning_rate": 5.425118217183945e-07, "loss": 0.0312, "num_input_tokens_seen": 9615176, "step": 35870 }, { "epoch": 9.403669724770642, "grad_norm": 2.629420042037964, "learning_rate": 5.40144626212058e-07, "loss": 0.1195, "num_input_tokens_seen": 9616472, "step": 35875 }, { "epoch": 9.404980340760158, "grad_norm": 2.5064516067504883, "learning_rate": 5.377825501111849e-07, "loss": 0.1108, "num_input_tokens_seen": 9617784, "step": 35880 }, { "epoch": 9.406290956749672, "grad_norm": 1.5455715656280518, "learning_rate": 5.354255939101576e-07, "loss": 0.1249, "num_input_tokens_seen": 9619272, "step": 35885 }, { "epoch": 9.407601572739187, "grad_norm": 2.3652472496032715, "learning_rate": 5.330737581022843e-07, "loss": 0.13, "num_input_tokens_seen": 9620472, "step": 35890 }, { "epoch": 9.408912188728703, "grad_norm": 0.3465063273906708, "learning_rate": 5.307270431797962e-07, "loss": 0.1398, "num_input_tokens_seen": 9622040, "step": 35895 }, { "epoch": 9.410222804718218, "grad_norm": 2.2696917057037354, "learning_rate": 5.283854496338586e-07, "loss": 0.1704, "num_input_tokens_seen": 9623304, "step": 35900 }, { "epoch": 9.411533420707732, "grad_norm": 1.5396939516067505, "learning_rate": 5.26048977954563e-07, "loss": 0.1038, "num_input_tokens_seen": 9624936, "step": 35905 }, { "epoch": 9.412844036697248, "grad_norm": 2.0652403831481934, "learning_rate": 5.237176286309292e-07, "loss": 0.1058, "num_input_tokens_seen": 9626952, "step": 35910 }, { "epoch": 9.414154652686763, "grad_norm": 1.9160079956054688, "learning_rate": 5.213914021509031e-07, "loss": 0.2019, "num_input_tokens_seen": 9628200, "step": 35915 }, { "epoch": 9.415465268676279, "grad_norm": 1.6642727851867676, "learning_rate": 5.190702990013618e-07, "loss": 0.0869, "num_input_tokens_seen": 9629432, "step": 35920 }, { "epoch": 9.416775884665793, "grad_norm": 1.6829988956451416, "learning_rate": 5.167543196681058e-07, "loss": 0.0715, "num_input_tokens_seen": 9630824, "step": 35925 }, { "epoch": 9.418086500655308, "grad_norm": 3.4000954627990723, "learning_rate": 5.144434646358665e-07, "loss": 0.1621, "num_input_tokens_seen": 9632104, "step": 35930 }, { "epoch": 9.419397116644824, "grad_norm": 1.4094080924987793, "learning_rate": 5.121377343883016e-07, "loss": 0.1022, "num_input_tokens_seen": 9633704, "step": 35935 }, { "epoch": 9.420707732634337, "grad_norm": 1.2390856742858887, "learning_rate": 5.098371294080001e-07, "loss": 0.0729, "num_input_tokens_seen": 9634968, "step": 35940 }, { "epoch": 9.422018348623853, "grad_norm": 2.701275110244751, "learning_rate": 5.075416501764685e-07, "loss": 0.2971, "num_input_tokens_seen": 9636392, "step": 35945 }, { "epoch": 9.423328964613368, "grad_norm": 2.41206431388855, "learning_rate": 5.052512971741529e-07, "loss": 0.1308, "num_input_tokens_seen": 9637400, "step": 35950 }, { "epoch": 9.424639580602884, "grad_norm": 2.0459072589874268, "learning_rate": 5.029660708804146e-07, "loss": 0.1734, "num_input_tokens_seen": 9638952, "step": 35955 }, { "epoch": 9.425950196592398, "grad_norm": 3.0552730560302734, "learning_rate": 5.006859717735512e-07, "loss": 0.1645, "num_input_tokens_seen": 9640232, "step": 35960 }, { "epoch": 9.427260812581913, "grad_norm": 1.4751050472259521, "learning_rate": 4.984110003307812e-07, "loss": 0.1239, "num_input_tokens_seen": 9641384, "step": 35965 }, { "epoch": 9.428571428571429, "grad_norm": 1.7416868209838867, "learning_rate": 4.961411570282543e-07, "loss": 0.0858, "num_input_tokens_seen": 9643128, "step": 35970 }, { "epoch": 9.429882044560944, "grad_norm": 2.5727455615997314, "learning_rate": 4.938764423410458e-07, "loss": 0.0895, "num_input_tokens_seen": 9644312, "step": 35975 }, { "epoch": 9.431192660550458, "grad_norm": 3.6709234714508057, "learning_rate": 4.916168567431544e-07, "loss": 0.1189, "num_input_tokens_seen": 9645544, "step": 35980 }, { "epoch": 9.432503276539974, "grad_norm": 2.089369297027588, "learning_rate": 4.893624007075048e-07, "loss": 0.1423, "num_input_tokens_seen": 9646888, "step": 35985 }, { "epoch": 9.43381389252949, "grad_norm": 2.3688485622406006, "learning_rate": 4.871130747059554e-07, "loss": 0.1427, "num_input_tokens_seen": 9648360, "step": 35990 }, { "epoch": 9.435124508519005, "grad_norm": 1.7121763229370117, "learning_rate": 4.848688792092798e-07, "loss": 0.1499, "num_input_tokens_seen": 9649608, "step": 35995 }, { "epoch": 9.436435124508519, "grad_norm": 1.697890281677246, "learning_rate": 4.826298146871938e-07, "loss": 0.2261, "num_input_tokens_seen": 9650840, "step": 36000 }, { "epoch": 9.437745740498034, "grad_norm": 0.8887071013450623, "learning_rate": 4.803958816083254e-07, "loss": 0.1197, "num_input_tokens_seen": 9651928, "step": 36005 }, { "epoch": 9.43905635648755, "grad_norm": 2.838531494140625, "learning_rate": 4.781670804402283e-07, "loss": 0.1398, "num_input_tokens_seen": 9653192, "step": 36010 }, { "epoch": 9.440366972477065, "grad_norm": 1.8847013711929321, "learning_rate": 4.75943411649396e-07, "loss": 0.1074, "num_input_tokens_seen": 9654152, "step": 36015 }, { "epoch": 9.441677588466579, "grad_norm": 1.6599732637405396, "learning_rate": 4.7372487570123135e-07, "loss": 0.1729, "num_input_tokens_seen": 9655656, "step": 36020 }, { "epoch": 9.442988204456094, "grad_norm": 1.0275921821594238, "learning_rate": 4.7151147306007117e-07, "loss": 0.1046, "num_input_tokens_seen": 9656792, "step": 36025 }, { "epoch": 9.44429882044561, "grad_norm": 4.234565734863281, "learning_rate": 4.6930320418917817e-07, "loss": 0.1735, "num_input_tokens_seen": 9657928, "step": 36030 }, { "epoch": 9.445609436435124, "grad_norm": 2.609384059906006, "learning_rate": 4.6710006955073826e-07, "loss": 0.1178, "num_input_tokens_seen": 9659032, "step": 36035 }, { "epoch": 9.44692005242464, "grad_norm": 2.7509255409240723, "learning_rate": 4.649020696058659e-07, "loss": 0.1347, "num_input_tokens_seen": 9660200, "step": 36040 }, { "epoch": 9.448230668414155, "grad_norm": 1.2598332166671753, "learning_rate": 4.627092048145959e-07, "loss": 0.0538, "num_input_tokens_seen": 9661432, "step": 36045 }, { "epoch": 9.44954128440367, "grad_norm": 3.801981210708618, "learning_rate": 4.6052147563589167e-07, "loss": 0.2051, "num_input_tokens_seen": 9662776, "step": 36050 }, { "epoch": 9.450851900393184, "grad_norm": 0.45869338512420654, "learning_rate": 4.583388825276397e-07, "loss": 0.1153, "num_input_tokens_seen": 9663896, "step": 36055 }, { "epoch": 9.4521625163827, "grad_norm": 1.1521239280700684, "learning_rate": 4.561614259466579e-07, "loss": 0.0696, "num_input_tokens_seen": 9665144, "step": 36060 }, { "epoch": 9.453473132372215, "grad_norm": 3.5778768062591553, "learning_rate": 4.53989106348679e-07, "loss": 0.3006, "num_input_tokens_seen": 9666616, "step": 36065 }, { "epoch": 9.45478374836173, "grad_norm": 2.1212124824523926, "learning_rate": 4.518219241883698e-07, "loss": 0.082, "num_input_tokens_seen": 9667768, "step": 36070 }, { "epoch": 9.456094364351245, "grad_norm": 3.1630325317382812, "learning_rate": 4.4965987991931466e-07, "loss": 0.0883, "num_input_tokens_seen": 9669192, "step": 36075 }, { "epoch": 9.45740498034076, "grad_norm": 1.1580039262771606, "learning_rate": 4.475029739940295e-07, "loss": 0.1166, "num_input_tokens_seen": 9670392, "step": 36080 }, { "epoch": 9.458715596330276, "grad_norm": 1.4649966955184937, "learning_rate": 4.4535120686394484e-07, "loss": 0.1345, "num_input_tokens_seen": 9671656, "step": 36085 }, { "epoch": 9.460026212319791, "grad_norm": 1.1170204877853394, "learning_rate": 4.4320457897942815e-07, "loss": 0.0631, "num_input_tokens_seen": 9672696, "step": 36090 }, { "epoch": 9.461336828309305, "grad_norm": 0.9193148612976074, "learning_rate": 4.410630907897617e-07, "loss": 0.1272, "num_input_tokens_seen": 9674152, "step": 36095 }, { "epoch": 9.46264744429882, "grad_norm": 1.8702424764633179, "learning_rate": 4.389267427431565e-07, "loss": 0.0821, "num_input_tokens_seen": 9675272, "step": 36100 }, { "epoch": 9.463958060288336, "grad_norm": 1.7747126817703247, "learning_rate": 4.367955352867437e-07, "loss": 0.1227, "num_input_tokens_seen": 9676600, "step": 36105 }, { "epoch": 9.46526867627785, "grad_norm": 2.470520257949829, "learning_rate": 4.346694688665831e-07, "loss": 0.096, "num_input_tokens_seen": 9677656, "step": 36110 }, { "epoch": 9.466579292267365, "grad_norm": 2.441338062286377, "learning_rate": 4.325485439276578e-07, "loss": 0.0905, "num_input_tokens_seen": 9678696, "step": 36115 }, { "epoch": 9.46788990825688, "grad_norm": 1.511999487876892, "learning_rate": 4.304327609138736e-07, "loss": 0.1167, "num_input_tokens_seen": 9680280, "step": 36120 }, { "epoch": 9.469200524246396, "grad_norm": 2.111384153366089, "learning_rate": 4.2832212026805986e-07, "loss": 0.1346, "num_input_tokens_seen": 9681432, "step": 36125 }, { "epoch": 9.47051114023591, "grad_norm": 4.047342777252197, "learning_rate": 4.262166224319686e-07, "loss": 0.1547, "num_input_tokens_seen": 9682744, "step": 36130 }, { "epoch": 9.471821756225426, "grad_norm": 1.1931722164154053, "learning_rate": 4.2411626784628065e-07, "loss": 0.2189, "num_input_tokens_seen": 9686088, "step": 36135 }, { "epoch": 9.473132372214941, "grad_norm": 2.033994674682617, "learning_rate": 4.220210569505917e-07, "loss": 0.1486, "num_input_tokens_seen": 9687240, "step": 36140 }, { "epoch": 9.474442988204457, "grad_norm": 0.19807054102420807, "learning_rate": 4.199309901834286e-07, "loss": 0.1304, "num_input_tokens_seen": 9688440, "step": 36145 }, { "epoch": 9.47575360419397, "grad_norm": 1.2443702220916748, "learning_rate": 4.178460679822388e-07, "loss": 0.1937, "num_input_tokens_seen": 9690136, "step": 36150 }, { "epoch": 9.477064220183486, "grad_norm": 2.0889065265655518, "learning_rate": 4.157662907833898e-07, "loss": 0.0815, "num_input_tokens_seen": 9691352, "step": 36155 }, { "epoch": 9.478374836173002, "grad_norm": 1.4408690929412842, "learning_rate": 4.1369165902218065e-07, "loss": 0.1974, "num_input_tokens_seen": 9692760, "step": 36160 }, { "epoch": 9.479685452162517, "grad_norm": 0.30867692828178406, "learning_rate": 4.1162217313282236e-07, "loss": 0.1761, "num_input_tokens_seen": 9694024, "step": 36165 }, { "epoch": 9.48099606815203, "grad_norm": 1.3498482704162598, "learning_rate": 4.0955783354846013e-07, "loss": 0.1141, "num_input_tokens_seen": 9695192, "step": 36170 }, { "epoch": 9.482306684141546, "grad_norm": 1.8935470581054688, "learning_rate": 4.0749864070115387e-07, "loss": 0.161, "num_input_tokens_seen": 9696536, "step": 36175 }, { "epoch": 9.483617300131062, "grad_norm": 1.1048920154571533, "learning_rate": 4.0544459502188934e-07, "loss": 0.0644, "num_input_tokens_seen": 9697928, "step": 36180 }, { "epoch": 9.484927916120578, "grad_norm": 2.04659366607666, "learning_rate": 4.0339569694057554e-07, "loss": 0.187, "num_input_tokens_seen": 9699272, "step": 36185 }, { "epoch": 9.486238532110091, "grad_norm": 3.355172634124756, "learning_rate": 4.013519468860416e-07, "loss": 0.223, "num_input_tokens_seen": 9700968, "step": 36190 }, { "epoch": 9.487549148099607, "grad_norm": 1.689030647277832, "learning_rate": 3.993133452860426e-07, "loss": 0.1992, "num_input_tokens_seen": 9702344, "step": 36195 }, { "epoch": 9.488859764089122, "grad_norm": 3.774273633956909, "learning_rate": 3.9727989256725384e-07, "loss": 0.1482, "num_input_tokens_seen": 9703352, "step": 36200 }, { "epoch": 9.490170380078636, "grad_norm": 2.124027967453003, "learning_rate": 3.9525158915527383e-07, "loss": 0.1234, "num_input_tokens_seen": 9704840, "step": 36205 }, { "epoch": 9.491480996068152, "grad_norm": 1.9508074522018433, "learning_rate": 3.932284354746213e-07, "loss": 0.1153, "num_input_tokens_seen": 9706248, "step": 36210 }, { "epoch": 9.492791612057667, "grad_norm": 4.935507774353027, "learning_rate": 3.9121043194874365e-07, "loss": 0.0911, "num_input_tokens_seen": 9707384, "step": 36215 }, { "epoch": 9.494102228047183, "grad_norm": 0.7054265737533569, "learning_rate": 3.8919757899999743e-07, "loss": 0.0947, "num_input_tokens_seen": 9709208, "step": 36220 }, { "epoch": 9.495412844036696, "grad_norm": 1.367629885673523, "learning_rate": 3.8718987704967345e-07, "loss": 0.1031, "num_input_tokens_seen": 9710392, "step": 36225 }, { "epoch": 9.496723460026212, "grad_norm": 3.5340161323547363, "learning_rate": 3.8518732651798284e-07, "loss": 0.1849, "num_input_tokens_seen": 9711624, "step": 36230 }, { "epoch": 9.498034076015728, "grad_norm": 0.9756973385810852, "learning_rate": 3.8318992782405414e-07, "loss": 0.108, "num_input_tokens_seen": 9712680, "step": 36235 }, { "epoch": 9.499344692005243, "grad_norm": 1.4220720529556274, "learning_rate": 3.811976813859364e-07, "loss": 0.0862, "num_input_tokens_seen": 9714008, "step": 36240 }, { "epoch": 9.500655307994757, "grad_norm": 1.6711879968643188, "learning_rate": 3.792105876206098e-07, "loss": 0.1308, "num_input_tokens_seen": 9715176, "step": 36245 }, { "epoch": 9.501965923984272, "grad_norm": 3.987288475036621, "learning_rate": 3.7722864694396397e-07, "loss": 0.1008, "num_input_tokens_seen": 9716296, "step": 36250 }, { "epoch": 9.502490170380078, "eval_loss": 0.9084105491638184, "eval_runtime": 16.7043, "eval_samples_per_second": 50.765, "eval_steps_per_second": 25.383, "num_input_tokens_seen": 9716920, "step": 36252 }, { "epoch": 9.503276539973788, "grad_norm": 0.4776995778083801, "learning_rate": 3.7525185977081713e-07, "loss": 0.1074, "num_input_tokens_seen": 9719064, "step": 36255 }, { "epoch": 9.504587155963304, "grad_norm": 1.3234055042266846, "learning_rate": 3.7328022651490767e-07, "loss": 0.136, "num_input_tokens_seen": 9720472, "step": 36260 }, { "epoch": 9.505897771952817, "grad_norm": 2.616325855255127, "learning_rate": 3.7131374758889446e-07, "loss": 0.1351, "num_input_tokens_seen": 9721896, "step": 36265 }, { "epoch": 9.507208387942333, "grad_norm": 0.6744889616966248, "learning_rate": 3.693524234043594e-07, "loss": 0.152, "num_input_tokens_seen": 9723480, "step": 36270 }, { "epoch": 9.508519003931848, "grad_norm": 0.7145595550537109, "learning_rate": 3.6739625437180457e-07, "loss": 0.1421, "num_input_tokens_seen": 9725224, "step": 36275 }, { "epoch": 9.509829619921362, "grad_norm": 4.848711967468262, "learning_rate": 3.654452409006498e-07, "loss": 0.3595, "num_input_tokens_seen": 9726520, "step": 36280 }, { "epoch": 9.511140235910878, "grad_norm": 5.102114677429199, "learning_rate": 3.634993833992434e-07, "loss": 0.0853, "num_input_tokens_seen": 9727352, "step": 36285 }, { "epoch": 9.512450851900393, "grad_norm": 3.483776569366455, "learning_rate": 3.615586822748457e-07, "loss": 0.1123, "num_input_tokens_seen": 9728760, "step": 36290 }, { "epoch": 9.513761467889909, "grad_norm": 0.9029681086540222, "learning_rate": 3.596231379336429e-07, "loss": 0.0724, "num_input_tokens_seen": 9730360, "step": 36295 }, { "epoch": 9.515072083879424, "grad_norm": 3.4429242610931396, "learning_rate": 3.5769275078074425e-07, "loss": 0.1198, "num_input_tokens_seen": 9731688, "step": 36300 }, { "epoch": 9.516382699868938, "grad_norm": 1.155398964881897, "learning_rate": 3.5576752122017385e-07, "loss": 0.1929, "num_input_tokens_seen": 9733160, "step": 36305 }, { "epoch": 9.517693315858454, "grad_norm": 0.2384902536869049, "learning_rate": 3.5384744965487883e-07, "loss": 0.093, "num_input_tokens_seen": 9734760, "step": 36310 }, { "epoch": 9.51900393184797, "grad_norm": 1.8751970529556274, "learning_rate": 3.519325364867265e-07, "loss": 0.0743, "num_input_tokens_seen": 9735960, "step": 36315 }, { "epoch": 9.520314547837483, "grad_norm": 0.5909475684165955, "learning_rate": 3.5002278211650476e-07, "loss": 0.1604, "num_input_tokens_seen": 9737416, "step": 36320 }, { "epoch": 9.521625163826998, "grad_norm": 1.1755883693695068, "learning_rate": 3.4811818694392706e-07, "loss": 0.1655, "num_input_tokens_seen": 9738888, "step": 36325 }, { "epoch": 9.522935779816514, "grad_norm": 5.10992956161499, "learning_rate": 3.462187513676135e-07, "loss": 0.0637, "num_input_tokens_seen": 9739880, "step": 36330 }, { "epoch": 9.52424639580603, "grad_norm": 2.771373748779297, "learning_rate": 3.4432447578511826e-07, "loss": 0.1003, "num_input_tokens_seen": 9741336, "step": 36335 }, { "epoch": 9.525557011795543, "grad_norm": 1.0946205854415894, "learning_rate": 3.4243536059291035e-07, "loss": 0.184, "num_input_tokens_seen": 9742776, "step": 36340 }, { "epoch": 9.526867627785059, "grad_norm": 2.1333858966827393, "learning_rate": 3.405514061863735e-07, "loss": 0.0883, "num_input_tokens_seen": 9744072, "step": 36345 }, { "epoch": 9.528178243774574, "grad_norm": 3.8916029930114746, "learning_rate": 3.3867261295982e-07, "loss": 0.2157, "num_input_tokens_seen": 9745384, "step": 36350 }, { "epoch": 9.52948885976409, "grad_norm": 2.114206075668335, "learning_rate": 3.36798981306477e-07, "loss": 0.2285, "num_input_tokens_seen": 9747400, "step": 36355 }, { "epoch": 9.530799475753604, "grad_norm": 2.0251927375793457, "learning_rate": 3.3493051161849477e-07, "loss": 0.1603, "num_input_tokens_seen": 9749032, "step": 36360 }, { "epoch": 9.53211009174312, "grad_norm": 2.34676456451416, "learning_rate": 3.330672042869354e-07, "loss": 0.1984, "num_input_tokens_seen": 9750648, "step": 36365 }, { "epoch": 9.533420707732635, "grad_norm": 2.8135104179382324, "learning_rate": 3.312090597017925e-07, "loss": 0.0962, "num_input_tokens_seen": 9752392, "step": 36370 }, { "epoch": 9.534731323722148, "grad_norm": 3.0202014446258545, "learning_rate": 3.2935607825196614e-07, "loss": 0.12, "num_input_tokens_seen": 9754152, "step": 36375 }, { "epoch": 9.536041939711664, "grad_norm": 1.570130705833435, "learning_rate": 3.275082603252877e-07, "loss": 0.0695, "num_input_tokens_seen": 9755128, "step": 36380 }, { "epoch": 9.53735255570118, "grad_norm": 1.776708960533142, "learning_rate": 3.2566560630849784e-07, "loss": 0.0703, "num_input_tokens_seen": 9756408, "step": 36385 }, { "epoch": 9.538663171690695, "grad_norm": 10.882648468017578, "learning_rate": 3.2382811658726033e-07, "loss": 0.1618, "num_input_tokens_seen": 9757432, "step": 36390 }, { "epoch": 9.539973787680209, "grad_norm": 1.1040400266647339, "learning_rate": 3.219957915461619e-07, "loss": 0.061, "num_input_tokens_seen": 9758856, "step": 36395 }, { "epoch": 9.541284403669724, "grad_norm": 0.7601689696311951, "learning_rate": 3.201686315687041e-07, "loss": 0.13, "num_input_tokens_seen": 9761496, "step": 36400 }, { "epoch": 9.54259501965924, "grad_norm": 2.286191940307617, "learning_rate": 3.183466370373117e-07, "loss": 0.1309, "num_input_tokens_seen": 9762616, "step": 36405 }, { "epoch": 9.543905635648755, "grad_norm": 2.842372417449951, "learning_rate": 3.1652980833331846e-07, "loss": 0.1386, "num_input_tokens_seen": 9763864, "step": 36410 }, { "epoch": 9.54521625163827, "grad_norm": 2.4993197917938232, "learning_rate": 3.147181458369897e-07, "loss": 0.1269, "num_input_tokens_seen": 9764968, "step": 36415 }, { "epoch": 9.546526867627785, "grad_norm": 4.689060211181641, "learning_rate": 3.12911649927497e-07, "loss": 0.169, "num_input_tokens_seen": 9766424, "step": 36420 }, { "epoch": 9.5478374836173, "grad_norm": 1.9543167352676392, "learning_rate": 3.111103209829408e-07, "loss": 0.0868, "num_input_tokens_seen": 9767912, "step": 36425 }, { "epoch": 9.549148099606816, "grad_norm": 3.5764920711517334, "learning_rate": 3.0931415938033336e-07, "loss": 0.0841, "num_input_tokens_seen": 9768904, "step": 36430 }, { "epoch": 9.55045871559633, "grad_norm": 1.6719218492507935, "learning_rate": 3.075231654956129e-07, "loss": 0.1449, "num_input_tokens_seen": 9770552, "step": 36435 }, { "epoch": 9.551769331585845, "grad_norm": 2.4749739170074463, "learning_rate": 3.0573733970362674e-07, "loss": 0.2064, "num_input_tokens_seen": 9771704, "step": 36440 }, { "epoch": 9.55307994757536, "grad_norm": 1.9904348850250244, "learning_rate": 3.0395668237814813e-07, "loss": 0.1631, "num_input_tokens_seen": 9773320, "step": 36445 }, { "epoch": 9.554390563564876, "grad_norm": 0.24971002340316772, "learning_rate": 3.0218119389186503e-07, "loss": 0.1425, "num_input_tokens_seen": 9774424, "step": 36450 }, { "epoch": 9.55570117955439, "grad_norm": 1.9118331670761108, "learning_rate": 3.0041087461638304e-07, "loss": 0.0587, "num_input_tokens_seen": 9775336, "step": 36455 }, { "epoch": 9.557011795543906, "grad_norm": 1.2461457252502441, "learning_rate": 2.986457249222252e-07, "loss": 0.1666, "num_input_tokens_seen": 9776680, "step": 36460 }, { "epoch": 9.558322411533421, "grad_norm": 0.7922095656394958, "learning_rate": 2.968857451788404e-07, "loss": 0.1577, "num_input_tokens_seen": 9778088, "step": 36465 }, { "epoch": 9.559633027522935, "grad_norm": 1.658831238746643, "learning_rate": 2.951309357545812e-07, "loss": 0.1041, "num_input_tokens_seen": 9779752, "step": 36470 }, { "epoch": 9.56094364351245, "grad_norm": 1.220672369003296, "learning_rate": 2.9338129701673434e-07, "loss": 0.0939, "num_input_tokens_seen": 9781368, "step": 36475 }, { "epoch": 9.562254259501966, "grad_norm": 1.4991236925125122, "learning_rate": 2.916368293314903e-07, "loss": 0.0557, "num_input_tokens_seen": 9782600, "step": 36480 }, { "epoch": 9.563564875491481, "grad_norm": 4.0293192863464355, "learning_rate": 2.898975330639653e-07, "loss": 0.155, "num_input_tokens_seen": 9784184, "step": 36485 }, { "epoch": 9.564875491480995, "grad_norm": 2.1732590198516846, "learning_rate": 2.8816340857819034e-07, "loss": 0.0989, "num_input_tokens_seen": 9785672, "step": 36490 }, { "epoch": 9.56618610747051, "grad_norm": 1.8036049604415894, "learning_rate": 2.8643445623711684e-07, "loss": 0.1609, "num_input_tokens_seen": 9786952, "step": 36495 }, { "epoch": 9.567496723460026, "grad_norm": 1.8042634725570679, "learning_rate": 2.847106764026081e-07, "loss": 0.0849, "num_input_tokens_seen": 9788040, "step": 36500 }, { "epoch": 9.568807339449542, "grad_norm": 4.53346061706543, "learning_rate": 2.8299206943545044e-07, "loss": 0.1354, "num_input_tokens_seen": 9789144, "step": 36505 }, { "epoch": 9.570117955439056, "grad_norm": 2.7721800804138184, "learning_rate": 2.812786356953451e-07, "loss": 0.0999, "num_input_tokens_seen": 9790168, "step": 36510 }, { "epoch": 9.571428571428571, "grad_norm": 1.750311017036438, "learning_rate": 2.795703755409107e-07, "loss": 0.159, "num_input_tokens_seen": 9791672, "step": 36515 }, { "epoch": 9.572739187418087, "grad_norm": 1.2728358507156372, "learning_rate": 2.7786728932968354e-07, "loss": 0.1438, "num_input_tokens_seen": 9793048, "step": 36520 }, { "epoch": 9.574049803407602, "grad_norm": 2.2153425216674805, "learning_rate": 2.761693774181173e-07, "loss": 0.1704, "num_input_tokens_seen": 9794648, "step": 36525 }, { "epoch": 9.575360419397116, "grad_norm": 0.47317641973495483, "learning_rate": 2.7447664016157783e-07, "loss": 0.0708, "num_input_tokens_seen": 9795928, "step": 36530 }, { "epoch": 9.576671035386632, "grad_norm": 3.1851441860198975, "learning_rate": 2.727890779143566e-07, "loss": 0.1685, "num_input_tokens_seen": 9797064, "step": 36535 }, { "epoch": 9.577981651376147, "grad_norm": 1.194657325744629, "learning_rate": 2.7110669102965723e-07, "loss": 0.0851, "num_input_tokens_seen": 9798120, "step": 36540 }, { "epoch": 9.579292267365663, "grad_norm": 1.8612163066864014, "learning_rate": 2.694294798595981e-07, "loss": 0.1578, "num_input_tokens_seen": 9799336, "step": 36545 }, { "epoch": 9.580602883355176, "grad_norm": 0.6016544103622437, "learning_rate": 2.677574447552178e-07, "loss": 0.124, "num_input_tokens_seen": 9800664, "step": 36550 }, { "epoch": 9.581913499344692, "grad_norm": 4.270508766174316, "learning_rate": 2.660905860664697e-07, "loss": 0.0636, "num_input_tokens_seen": 9801656, "step": 36555 }, { "epoch": 9.583224115334207, "grad_norm": 0.9184279441833496, "learning_rate": 2.6442890414222487e-07, "loss": 0.1255, "num_input_tokens_seen": 9803416, "step": 36560 }, { "epoch": 9.584534731323721, "grad_norm": 5.1336894035339355, "learning_rate": 2.6277239933027163e-07, "loss": 0.1381, "num_input_tokens_seen": 9804568, "step": 36565 }, { "epoch": 9.585845347313237, "grad_norm": 1.381519079208374, "learning_rate": 2.611210719773133e-07, "loss": 0.0579, "num_input_tokens_seen": 9805528, "step": 36570 }, { "epoch": 9.587155963302752, "grad_norm": 1.753217101097107, "learning_rate": 2.5947492242896776e-07, "loss": 0.0623, "num_input_tokens_seen": 9806616, "step": 36575 }, { "epoch": 9.588466579292268, "grad_norm": 2.4809489250183105, "learning_rate": 2.5783395102977057e-07, "loss": 0.1509, "num_input_tokens_seen": 9807944, "step": 36580 }, { "epoch": 9.589777195281782, "grad_norm": 1.5839974880218506, "learning_rate": 2.5619815812318037e-07, "loss": 0.1805, "num_input_tokens_seen": 9809272, "step": 36585 }, { "epoch": 9.591087811271297, "grad_norm": 2.5769765377044678, "learning_rate": 2.5456754405155934e-07, "loss": 0.1697, "num_input_tokens_seen": 9811080, "step": 36590 }, { "epoch": 9.592398427260813, "grad_norm": 3.166325569152832, "learning_rate": 2.5294210915619564e-07, "loss": 0.1012, "num_input_tokens_seen": 9811960, "step": 36595 }, { "epoch": 9.593709043250328, "grad_norm": 1.0715429782867432, "learning_rate": 2.5132185377728656e-07, "loss": 0.0926, "num_input_tokens_seen": 9813048, "step": 36600 }, { "epoch": 9.595019659239842, "grad_norm": 4.133466720581055, "learning_rate": 2.497067782539553e-07, "loss": 0.1015, "num_input_tokens_seen": 9814168, "step": 36605 }, { "epoch": 9.596330275229358, "grad_norm": 2.97493839263916, "learning_rate": 2.480968829242286e-07, "loss": 0.0663, "num_input_tokens_seen": 9815416, "step": 36610 }, { "epoch": 9.597640891218873, "grad_norm": 2.0710439682006836, "learning_rate": 2.4649216812505373e-07, "loss": 0.1387, "num_input_tokens_seen": 9816632, "step": 36615 }, { "epoch": 9.598951507208389, "grad_norm": 3.1332733631134033, "learning_rate": 2.448926341923008e-07, "loss": 0.0636, "num_input_tokens_seen": 9817736, "step": 36620 }, { "epoch": 9.600262123197902, "grad_norm": 1.6929426193237305, "learning_rate": 2.4329828146074095e-07, "loss": 0.1344, "num_input_tokens_seen": 9819432, "step": 36625 }, { "epoch": 9.601572739187418, "grad_norm": 0.8790777921676636, "learning_rate": 2.417091102640795e-07, "loss": 0.0971, "num_input_tokens_seen": 9821096, "step": 36630 }, { "epoch": 9.602883355176933, "grad_norm": 1.8196675777435303, "learning_rate": 2.401251209349198e-07, "loss": 0.2046, "num_input_tokens_seen": 9822584, "step": 36635 }, { "epoch": 9.604193971166449, "grad_norm": 4.824203968048096, "learning_rate": 2.385463138047911e-07, "loss": 0.1284, "num_input_tokens_seen": 9823720, "step": 36640 }, { "epoch": 9.605504587155963, "grad_norm": 2.326470136642456, "learning_rate": 2.3697268920413462e-07, "loss": 0.1133, "num_input_tokens_seen": 9824984, "step": 36645 }, { "epoch": 9.606815203145478, "grad_norm": 1.7848526239395142, "learning_rate": 2.3540424746230916e-07, "loss": 0.2869, "num_input_tokens_seen": 9826152, "step": 36650 }, { "epoch": 9.608125819134994, "grad_norm": 2.548877716064453, "learning_rate": 2.3384098890758267e-07, "loss": 0.0779, "num_input_tokens_seen": 9827224, "step": 36655 }, { "epoch": 9.609436435124508, "grad_norm": 0.7645861506462097, "learning_rate": 2.322829138671434e-07, "loss": 0.0766, "num_input_tokens_seen": 9828472, "step": 36660 }, { "epoch": 9.610747051114023, "grad_norm": 3.29500150680542, "learning_rate": 2.3073002266709443e-07, "loss": 0.1068, "num_input_tokens_seen": 9829704, "step": 36665 }, { "epoch": 9.612057667103539, "grad_norm": 2.39357328414917, "learning_rate": 2.2918231563245629e-07, "loss": 0.0725, "num_input_tokens_seen": 9831224, "step": 36670 }, { "epoch": 9.613368283093054, "grad_norm": 1.8379261493682861, "learning_rate": 2.2763979308715877e-07, "loss": 0.1218, "num_input_tokens_seen": 9832744, "step": 36675 }, { "epoch": 9.614678899082568, "grad_norm": 2.2624876499176025, "learning_rate": 2.2610245535404638e-07, "loss": 0.1073, "num_input_tokens_seen": 9833960, "step": 36680 }, { "epoch": 9.615989515072084, "grad_norm": 3.6846795082092285, "learning_rate": 2.2457030275488676e-07, "loss": 0.1357, "num_input_tokens_seen": 9835960, "step": 36685 }, { "epoch": 9.617300131061599, "grad_norm": 3.683079242706299, "learning_rate": 2.2304333561035396e-07, "loss": 0.1159, "num_input_tokens_seen": 9836936, "step": 36690 }, { "epoch": 9.618610747051115, "grad_norm": 3.534013509750366, "learning_rate": 2.2152155424003952e-07, "loss": 0.0842, "num_input_tokens_seen": 9838296, "step": 36695 }, { "epoch": 9.619921363040628, "grad_norm": 1.22043776512146, "learning_rate": 2.2000495896244976e-07, "loss": 0.2818, "num_input_tokens_seen": 9839912, "step": 36700 }, { "epoch": 9.621231979030144, "grad_norm": 0.505605936050415, "learning_rate": 2.1849355009500582e-07, "loss": 0.1131, "num_input_tokens_seen": 9841096, "step": 36705 }, { "epoch": 9.62254259501966, "grad_norm": 0.8757616877555847, "learning_rate": 2.1698732795404074e-07, "loss": 0.0888, "num_input_tokens_seen": 9842600, "step": 36710 }, { "epoch": 9.623853211009175, "grad_norm": 0.31006497144699097, "learning_rate": 2.1548629285480792e-07, "loss": 0.0844, "num_input_tokens_seen": 9844456, "step": 36715 }, { "epoch": 9.625163826998689, "grad_norm": 0.7865310907363892, "learning_rate": 2.1399044511146993e-07, "loss": 0.0828, "num_input_tokens_seen": 9845944, "step": 36720 }, { "epoch": 9.626474442988204, "grad_norm": 2.0097899436950684, "learning_rate": 2.124997850371041e-07, "loss": 0.133, "num_input_tokens_seen": 9847400, "step": 36725 }, { "epoch": 9.62778505897772, "grad_norm": 1.4149818420410156, "learning_rate": 2.1101431294370533e-07, "loss": 0.1506, "num_input_tokens_seen": 9848440, "step": 36730 }, { "epoch": 9.629095674967235, "grad_norm": 0.5441896319389343, "learning_rate": 2.0953402914217768e-07, "loss": 0.141, "num_input_tokens_seen": 9849816, "step": 36735 }, { "epoch": 9.63040629095675, "grad_norm": 2.0472006797790527, "learning_rate": 2.0805893394234e-07, "loss": 0.1377, "num_input_tokens_seen": 9851304, "step": 36740 }, { "epoch": 9.631716906946265, "grad_norm": 1.787344217300415, "learning_rate": 2.065890276529342e-07, "loss": 0.0741, "num_input_tokens_seen": 9852568, "step": 36745 }, { "epoch": 9.63302752293578, "grad_norm": 4.289626121520996, "learning_rate": 2.051243105816031e-07, "loss": 0.1489, "num_input_tokens_seen": 9854056, "step": 36750 }, { "epoch": 9.634338138925294, "grad_norm": 0.684329628944397, "learning_rate": 2.0366478303491255e-07, "loss": 0.0996, "num_input_tokens_seen": 9855240, "step": 36755 }, { "epoch": 9.63564875491481, "grad_norm": 0.5943271517753601, "learning_rate": 2.0221044531834043e-07, "loss": 0.1319, "num_input_tokens_seen": 9856760, "step": 36760 }, { "epoch": 9.636959370904325, "grad_norm": 1.0990097522735596, "learning_rate": 2.0076129773627105e-07, "loss": 0.1103, "num_input_tokens_seen": 9858360, "step": 36765 }, { "epoch": 9.63826998689384, "grad_norm": 1.7482212781906128, "learning_rate": 1.9931734059201457e-07, "loss": 0.1375, "num_input_tokens_seen": 9859624, "step": 36770 }, { "epoch": 9.639580602883354, "grad_norm": 1.3678339719772339, "learning_rate": 1.9787857418778478e-07, "loss": 0.0601, "num_input_tokens_seen": 9860856, "step": 36775 }, { "epoch": 9.64089121887287, "grad_norm": 1.4158073663711548, "learning_rate": 1.9644499882471578e-07, "loss": 0.1071, "num_input_tokens_seen": 9862296, "step": 36780 }, { "epoch": 9.642201834862385, "grad_norm": 1.970729112625122, "learning_rate": 1.9501661480285095e-07, "loss": 0.1217, "num_input_tokens_seen": 9863784, "step": 36785 }, { "epoch": 9.643512450851901, "grad_norm": 1.4291642904281616, "learning_rate": 1.935934224211483e-07, "loss": 0.1642, "num_input_tokens_seen": 9865288, "step": 36790 }, { "epoch": 9.644823066841415, "grad_norm": 1.1784981489181519, "learning_rate": 1.9217542197748074e-07, "loss": 0.0517, "num_input_tokens_seen": 9866504, "step": 36795 }, { "epoch": 9.64613368283093, "grad_norm": 1.6323217153549194, "learning_rate": 1.9076261376863303e-07, "loss": 0.1616, "num_input_tokens_seen": 9867480, "step": 36800 }, { "epoch": 9.647444298820446, "grad_norm": 2.3853163719177246, "learning_rate": 1.8935499809030476e-07, "loss": 0.1558, "num_input_tokens_seen": 9869256, "step": 36805 }, { "epoch": 9.648754914809961, "grad_norm": 0.6107277870178223, "learning_rate": 1.8795257523710197e-07, "loss": 0.0872, "num_input_tokens_seen": 9871160, "step": 36810 }, { "epoch": 9.650065530799475, "grad_norm": 1.5324268341064453, "learning_rate": 1.8655534550255648e-07, "loss": 0.0954, "num_input_tokens_seen": 9872504, "step": 36815 }, { "epoch": 9.65137614678899, "grad_norm": 2.663087844848633, "learning_rate": 1.8516330917910108e-07, "loss": 0.0781, "num_input_tokens_seen": 9873928, "step": 36820 }, { "epoch": 9.652686762778506, "grad_norm": 4.937634468078613, "learning_rate": 1.8377646655808877e-07, "loss": 0.1718, "num_input_tokens_seen": 9875368, "step": 36825 }, { "epoch": 9.653997378768022, "grad_norm": 1.4597142934799194, "learning_rate": 1.8239481792978464e-07, "loss": 0.104, "num_input_tokens_seen": 9876584, "step": 36830 }, { "epoch": 9.655307994757536, "grad_norm": 1.669605016708374, "learning_rate": 1.8101836358336287e-07, "loss": 0.0959, "num_input_tokens_seen": 9878584, "step": 36835 }, { "epoch": 9.656618610747051, "grad_norm": 3.6788768768310547, "learning_rate": 1.7964710380691251e-07, "loss": 0.1316, "num_input_tokens_seen": 9879432, "step": 36840 }, { "epoch": 9.657929226736567, "grad_norm": 1.5390880107879639, "learning_rate": 1.782810388874373e-07, "loss": 0.1001, "num_input_tokens_seen": 9880744, "step": 36845 }, { "epoch": 9.65923984272608, "grad_norm": 3.357316493988037, "learning_rate": 1.7692016911085295e-07, "loss": 0.1628, "num_input_tokens_seen": 9882280, "step": 36850 }, { "epoch": 9.660550458715596, "grad_norm": 1.357601523399353, "learning_rate": 1.7556449476198445e-07, "loss": 0.192, "num_input_tokens_seen": 9883816, "step": 36855 }, { "epoch": 9.661861074705111, "grad_norm": 4.80232048034668, "learning_rate": 1.7421401612457423e-07, "loss": 0.1603, "num_input_tokens_seen": 9885512, "step": 36860 }, { "epoch": 9.663171690694627, "grad_norm": 1.840120553970337, "learning_rate": 1.7286873348127676e-07, "loss": 0.0897, "num_input_tokens_seen": 9886600, "step": 36865 }, { "epoch": 9.66448230668414, "grad_norm": 2.2172892093658447, "learning_rate": 1.715286471136529e-07, "loss": 0.0996, "num_input_tokens_seen": 9887592, "step": 36870 }, { "epoch": 9.665792922673656, "grad_norm": 2.0426366329193115, "learning_rate": 1.7019375730218667e-07, "loss": 0.0913, "num_input_tokens_seen": 9888504, "step": 36875 }, { "epoch": 9.667103538663172, "grad_norm": 2.699409246444702, "learning_rate": 1.688640643262629e-07, "loss": 0.1361, "num_input_tokens_seen": 9889688, "step": 36880 }, { "epoch": 9.668414154652687, "grad_norm": 1.3750494718551636, "learning_rate": 1.6753956846418395e-07, "loss": 0.2183, "num_input_tokens_seen": 9891304, "step": 36885 }, { "epoch": 9.669724770642201, "grad_norm": 3.332573652267456, "learning_rate": 1.6622026999316697e-07, "loss": 0.1017, "num_input_tokens_seen": 9893064, "step": 36890 }, { "epoch": 9.671035386631717, "grad_norm": 2.1378109455108643, "learning_rate": 1.6490616918933554e-07, "loss": 0.1402, "num_input_tokens_seen": 9894680, "step": 36895 }, { "epoch": 9.672346002621232, "grad_norm": 1.403340458869934, "learning_rate": 1.6359726632773353e-07, "loss": 0.129, "num_input_tokens_seen": 9896056, "step": 36900 }, { "epoch": 9.673656618610748, "grad_norm": 2.767080545425415, "learning_rate": 1.6229356168231125e-07, "loss": 0.1671, "num_input_tokens_seen": 9897448, "step": 36905 }, { "epoch": 9.674967234600262, "grad_norm": 1.720409870147705, "learning_rate": 1.609950555259282e-07, "loss": 0.1519, "num_input_tokens_seen": 9899128, "step": 36910 }, { "epoch": 9.676277850589777, "grad_norm": 1.6067314147949219, "learning_rate": 1.5970174813036144e-07, "loss": 0.1519, "num_input_tokens_seen": 9900552, "step": 36915 }, { "epoch": 9.677588466579293, "grad_norm": 1.3373714685440063, "learning_rate": 1.5841363976629998e-07, "loss": 0.1652, "num_input_tokens_seen": 9901784, "step": 36920 }, { "epoch": 9.678899082568808, "grad_norm": 0.76328045129776, "learning_rate": 1.5713073070333927e-07, "loss": 0.1005, "num_input_tokens_seen": 9902968, "step": 36925 }, { "epoch": 9.680209698558322, "grad_norm": 3.5606350898742676, "learning_rate": 1.5585302120998956e-07, "loss": 0.14, "num_input_tokens_seen": 9904472, "step": 36930 }, { "epoch": 9.681520314547837, "grad_norm": 1.6349234580993652, "learning_rate": 1.545805115536786e-07, "loss": 0.1206, "num_input_tokens_seen": 9905768, "step": 36935 }, { "epoch": 9.682830930537353, "grad_norm": 1.1323826313018799, "learning_rate": 1.5331320200073497e-07, "loss": 0.11, "num_input_tokens_seen": 9907912, "step": 36940 }, { "epoch": 9.684141546526867, "grad_norm": 2.2093653678894043, "learning_rate": 1.520510928164076e-07, "loss": 0.0552, "num_input_tokens_seen": 9908984, "step": 36945 }, { "epoch": 9.685452162516382, "grad_norm": 6.978725910186768, "learning_rate": 1.5079418426485193e-07, "loss": 0.1837, "num_input_tokens_seen": 9910216, "step": 36950 }, { "epoch": 9.686762778505898, "grad_norm": 2.018296957015991, "learning_rate": 1.49542476609138e-07, "loss": 0.257, "num_input_tokens_seen": 9911544, "step": 36955 }, { "epoch": 9.688073394495413, "grad_norm": 1.9706544876098633, "learning_rate": 1.482959701112452e-07, "loss": 0.1452, "num_input_tokens_seen": 9913352, "step": 36960 }, { "epoch": 9.689384010484927, "grad_norm": 0.7904735803604126, "learning_rate": 1.47054665032062e-07, "loss": 0.091, "num_input_tokens_seen": 9914376, "step": 36965 }, { "epoch": 9.690694626474443, "grad_norm": 1.6907106637954712, "learning_rate": 1.4581856163140008e-07, "loss": 0.111, "num_input_tokens_seen": 9915448, "step": 36970 }, { "epoch": 9.692005242463958, "grad_norm": 1.7372337579727173, "learning_rate": 1.4458766016796632e-07, "loss": 0.1627, "num_input_tokens_seen": 9916936, "step": 36975 }, { "epoch": 9.693315858453474, "grad_norm": 3.988450527191162, "learning_rate": 1.4336196089938802e-07, "loss": 0.1168, "num_input_tokens_seen": 9918200, "step": 36980 }, { "epoch": 9.694626474442988, "grad_norm": 1.2609578371047974, "learning_rate": 1.4214146408220163e-07, "loss": 0.1793, "num_input_tokens_seen": 9920120, "step": 36985 }, { "epoch": 9.695937090432503, "grad_norm": 2.3939707279205322, "learning_rate": 1.4092616997185837e-07, "loss": 0.1296, "num_input_tokens_seen": 9921464, "step": 36990 }, { "epoch": 9.697247706422019, "grad_norm": 1.7909765243530273, "learning_rate": 1.397160788227131e-07, "loss": 0.2593, "num_input_tokens_seen": 9923128, "step": 36995 }, { "epoch": 9.698558322411534, "grad_norm": 4.0987067222595215, "learning_rate": 1.385111908880382e-07, "loss": 0.1757, "num_input_tokens_seen": 9924408, "step": 37000 }, { "epoch": 9.699868938401048, "grad_norm": 1.4610382318496704, "learning_rate": 1.3731150642001255e-07, "loss": 0.1676, "num_input_tokens_seen": 9925912, "step": 37005 }, { "epoch": 9.701179554390563, "grad_norm": 1.681179404258728, "learning_rate": 1.3611702566972694e-07, "loss": 0.1013, "num_input_tokens_seen": 9927000, "step": 37010 }, { "epoch": 9.702490170380079, "grad_norm": 2.8168962001800537, "learning_rate": 1.3492774888718974e-07, "loss": 0.1303, "num_input_tokens_seen": 9927992, "step": 37015 }, { "epoch": 9.703800786369595, "grad_norm": 1.349563479423523, "learning_rate": 1.3374367632131014e-07, "loss": 0.2199, "num_input_tokens_seen": 9929448, "step": 37020 }, { "epoch": 9.705111402359108, "grad_norm": 6.248359203338623, "learning_rate": 1.3256480821991213e-07, "loss": 0.1321, "num_input_tokens_seen": 9930808, "step": 37025 }, { "epoch": 9.706422018348624, "grad_norm": 1.8784503936767578, "learning_rate": 1.3139114482973448e-07, "loss": 0.1686, "num_input_tokens_seen": 9932232, "step": 37030 }, { "epoch": 9.70773263433814, "grad_norm": 0.3291603922843933, "learning_rate": 1.3022268639641956e-07, "loss": 0.1825, "num_input_tokens_seen": 9933656, "step": 37035 }, { "epoch": 9.709043250327653, "grad_norm": 2.7501637935638428, "learning_rate": 1.290594331645245e-07, "loss": 0.1237, "num_input_tokens_seen": 9935368, "step": 37040 }, { "epoch": 9.710353866317169, "grad_norm": 3.576693534851074, "learning_rate": 1.279013853775185e-07, "loss": 0.0451, "num_input_tokens_seen": 9936376, "step": 37045 }, { "epoch": 9.711664482306684, "grad_norm": 1.8335912227630615, "learning_rate": 1.267485432777743e-07, "loss": 0.1491, "num_input_tokens_seen": 9937464, "step": 37050 }, { "epoch": 9.7129750982962, "grad_norm": 0.307525098323822, "learning_rate": 1.2560090710658223e-07, "loss": 0.0893, "num_input_tokens_seen": 9938696, "step": 37055 }, { "epoch": 9.714285714285714, "grad_norm": 0.6838637590408325, "learning_rate": 1.2445847710414183e-07, "loss": 0.018, "num_input_tokens_seen": 9939656, "step": 37060 }, { "epoch": 9.715596330275229, "grad_norm": 2.034111738204956, "learning_rate": 1.233212535095618e-07, "loss": 0.1037, "num_input_tokens_seen": 9940888, "step": 37065 }, { "epoch": 9.716906946264745, "grad_norm": 2.1111350059509277, "learning_rate": 1.2218923656085735e-07, "loss": 0.0728, "num_input_tokens_seen": 9942024, "step": 37070 }, { "epoch": 9.71821756225426, "grad_norm": 0.832830548286438, "learning_rate": 1.2106242649496112e-07, "loss": 0.1347, "num_input_tokens_seen": 9943336, "step": 37075 }, { "epoch": 9.719528178243774, "grad_norm": 1.2508392333984375, "learning_rate": 1.199408235477123e-07, "loss": 0.3417, "num_input_tokens_seen": 9945096, "step": 37080 }, { "epoch": 9.72083879423329, "grad_norm": 2.393211603164673, "learning_rate": 1.188244279538564e-07, "loss": 0.1436, "num_input_tokens_seen": 9946344, "step": 37085 }, { "epoch": 9.722149410222805, "grad_norm": 3.256845235824585, "learning_rate": 1.1771323994705929e-07, "loss": 0.0855, "num_input_tokens_seen": 9947688, "step": 37090 }, { "epoch": 9.72346002621232, "grad_norm": 1.5802679061889648, "learning_rate": 1.1660725975988773e-07, "loss": 0.1388, "num_input_tokens_seen": 9949816, "step": 37095 }, { "epoch": 9.724770642201834, "grad_norm": 2.569258689880371, "learning_rate": 1.1550648762382044e-07, "loss": 0.2345, "num_input_tokens_seen": 9951112, "step": 37100 }, { "epoch": 9.72608125819135, "grad_norm": 3.1870357990264893, "learning_rate": 1.144109237692509e-07, "loss": 0.1289, "num_input_tokens_seen": 9952312, "step": 37105 }, { "epoch": 9.727391874180865, "grad_norm": 2.903853178024292, "learning_rate": 1.1332056842547344e-07, "loss": 0.1475, "num_input_tokens_seen": 9953448, "step": 37110 }, { "epoch": 9.728702490170381, "grad_norm": 2.4577670097351074, "learning_rate": 1.1223542182070002e-07, "loss": 0.1814, "num_input_tokens_seen": 9954936, "step": 37115 }, { "epoch": 9.730013106159895, "grad_norm": 1.6861149072647095, "learning_rate": 1.1115548418205168e-07, "loss": 0.2168, "num_input_tokens_seen": 9956216, "step": 37120 }, { "epoch": 9.73132372214941, "grad_norm": 3.909684658050537, "learning_rate": 1.1008075573555599e-07, "loss": 0.1583, "num_input_tokens_seen": 9957416, "step": 37125 }, { "epoch": 9.732634338138926, "grad_norm": 3.797518730163574, "learning_rate": 1.0901123670614965e-07, "loss": 0.0764, "num_input_tokens_seen": 9958504, "step": 37130 }, { "epoch": 9.73394495412844, "grad_norm": 2.368239641189575, "learning_rate": 1.0794692731768419e-07, "loss": 0.1582, "num_input_tokens_seen": 9959784, "step": 37135 }, { "epoch": 9.735255570117955, "grad_norm": 2.1198716163635254, "learning_rate": 1.0688782779291473e-07, "loss": 0.2522, "num_input_tokens_seen": 9961576, "step": 37140 }, { "epoch": 9.73656618610747, "grad_norm": 2.430509090423584, "learning_rate": 1.0583393835351396e-07, "loss": 0.0973, "num_input_tokens_seen": 9963032, "step": 37145 }, { "epoch": 9.737876802096986, "grad_norm": 0.4448793828487396, "learning_rate": 1.0478525922005545e-07, "loss": 0.076, "num_input_tokens_seen": 9964056, "step": 37150 }, { "epoch": 9.7391874180865, "grad_norm": 1.6928132772445679, "learning_rate": 1.0374179061202749e-07, "loss": 0.0456, "num_input_tokens_seen": 9965368, "step": 37155 }, { "epoch": 9.740498034076015, "grad_norm": 2.217388153076172, "learning_rate": 1.0270353274782207e-07, "loss": 0.1874, "num_input_tokens_seen": 9966680, "step": 37160 }, { "epoch": 9.741808650065531, "grad_norm": 2.5340447425842285, "learning_rate": 1.0167048584475147e-07, "loss": 0.1056, "num_input_tokens_seen": 9967864, "step": 37165 }, { "epoch": 9.743119266055047, "grad_norm": 1.7367072105407715, "learning_rate": 1.006426501190233e-07, "loss": 0.1259, "num_input_tokens_seen": 9969448, "step": 37170 }, { "epoch": 9.74442988204456, "grad_norm": 2.861330509185791, "learning_rate": 9.962002578576823e-08, "loss": 0.0991, "num_input_tokens_seen": 9970712, "step": 37175 }, { "epoch": 9.745740498034076, "grad_norm": 1.166388988494873, "learning_rate": 9.860261305901785e-08, "loss": 0.1242, "num_input_tokens_seen": 9971736, "step": 37180 }, { "epoch": 9.747051114023591, "grad_norm": 6.280514240264893, "learning_rate": 9.759041215171295e-08, "loss": 0.1056, "num_input_tokens_seen": 9972872, "step": 37185 }, { "epoch": 9.748361730013107, "grad_norm": 3.0783865451812744, "learning_rate": 9.658342327570902e-08, "loss": 0.2933, "num_input_tokens_seen": 9973784, "step": 37190 }, { "epoch": 9.74967234600262, "grad_norm": 1.9520740509033203, "learning_rate": 9.55816466417625e-08, "loss": 0.0825, "num_input_tokens_seen": 9974840, "step": 37195 }, { "epoch": 9.750982961992136, "grad_norm": 2.8310792446136475, "learning_rate": 9.458508245954456e-08, "loss": 0.1526, "num_input_tokens_seen": 9976088, "step": 37200 }, { "epoch": 9.752293577981652, "grad_norm": 2.688903331756592, "learning_rate": 9.359373093763835e-08, "loss": 0.1379, "num_input_tokens_seen": 9977432, "step": 37205 }, { "epoch": 9.753604193971167, "grad_norm": 3.0710608959198, "learning_rate": 9.260759228352789e-08, "loss": 0.189, "num_input_tokens_seen": 9978392, "step": 37210 }, { "epoch": 9.754914809960681, "grad_norm": 2.8627703189849854, "learning_rate": 9.162666670361198e-08, "loss": 0.2042, "num_input_tokens_seen": 9980024, "step": 37215 }, { "epoch": 9.756225425950197, "grad_norm": 3.8566431999206543, "learning_rate": 9.065095440319582e-08, "loss": 0.1817, "num_input_tokens_seen": 9981416, "step": 37220 }, { "epoch": 9.757536041939712, "grad_norm": 2.5656239986419678, "learning_rate": 8.968045558649663e-08, "loss": 0.0804, "num_input_tokens_seen": 9982664, "step": 37225 }, { "epoch": 9.758846657929226, "grad_norm": 2.57595157623291, "learning_rate": 8.871517045663524e-08, "loss": 0.081, "num_input_tokens_seen": 9984040, "step": 37230 }, { "epoch": 9.760157273918741, "grad_norm": 3.2677345275878906, "learning_rate": 8.775509921564728e-08, "loss": 0.1813, "num_input_tokens_seen": 9985400, "step": 37235 }, { "epoch": 9.761467889908257, "grad_norm": 5.813194751739502, "learning_rate": 8.680024206446924e-08, "loss": 0.2269, "num_input_tokens_seen": 9986584, "step": 37240 }, { "epoch": 9.762778505897773, "grad_norm": 2.710388660430908, "learning_rate": 8.585059920295512e-08, "loss": 0.1774, "num_input_tokens_seen": 9988088, "step": 37245 }, { "epoch": 9.764089121887286, "grad_norm": 3.797132968902588, "learning_rate": 8.490617082985986e-08, "loss": 0.1763, "num_input_tokens_seen": 9989192, "step": 37250 }, { "epoch": 9.765399737876802, "grad_norm": 1.649535059928894, "learning_rate": 8.396695714285585e-08, "loss": 0.1137, "num_input_tokens_seen": 9990392, "step": 37255 }, { "epoch": 9.766710353866317, "grad_norm": 1.6171553134918213, "learning_rate": 8.303295833851365e-08, "loss": 0.1553, "num_input_tokens_seen": 9991608, "step": 37260 }, { "epoch": 9.768020969855833, "grad_norm": 1.1751186847686768, "learning_rate": 8.210417461232412e-08, "loss": 0.1568, "num_input_tokens_seen": 9992968, "step": 37265 }, { "epoch": 9.769331585845347, "grad_norm": 1.437835931777954, "learning_rate": 8.11806061586734e-08, "loss": 0.1167, "num_input_tokens_seen": 9994280, "step": 37270 }, { "epoch": 9.770642201834862, "grad_norm": 1.988434910774231, "learning_rate": 8.026225317086522e-08, "loss": 0.1431, "num_input_tokens_seen": 9995736, "step": 37275 }, { "epoch": 9.771952817824378, "grad_norm": 2.6981070041656494, "learning_rate": 7.934911584110971e-08, "loss": 0.1556, "num_input_tokens_seen": 9997064, "step": 37280 }, { "epoch": 9.773263433813893, "grad_norm": 1.5637860298156738, "learning_rate": 7.844119436052622e-08, "loss": 0.0886, "num_input_tokens_seen": 9998072, "step": 37285 }, { "epoch": 9.774574049803407, "grad_norm": 2.566011428833008, "learning_rate": 7.753848891913772e-08, "loss": 0.1376, "num_input_tokens_seen": 9999416, "step": 37290 }, { "epoch": 9.775884665792923, "grad_norm": 3.188633680343628, "learning_rate": 7.664099970588201e-08, "loss": 0.1149, "num_input_tokens_seen": 10000648, "step": 37295 }, { "epoch": 9.777195281782438, "grad_norm": 0.4483897089958191, "learning_rate": 7.57487269085977e-08, "loss": 0.0411, "num_input_tokens_seen": 10001544, "step": 37300 }, { "epoch": 9.778505897771954, "grad_norm": 3.2428598403930664, "learning_rate": 7.486167071404371e-08, "loss": 0.0709, "num_input_tokens_seen": 10002648, "step": 37305 }, { "epoch": 9.779816513761467, "grad_norm": 3.9077413082122803, "learning_rate": 7.397983130787156e-08, "loss": 0.1286, "num_input_tokens_seen": 10004024, "step": 37310 }, { "epoch": 9.781127129750983, "grad_norm": 3.4870150089263916, "learning_rate": 7.310320887465305e-08, "loss": 0.1637, "num_input_tokens_seen": 10005288, "step": 37315 }, { "epoch": 9.782437745740499, "grad_norm": 1.4254059791564941, "learning_rate": 7.223180359786086e-08, "loss": 0.1608, "num_input_tokens_seen": 10006600, "step": 37320 }, { "epoch": 9.783748361730012, "grad_norm": 2.309114456176758, "learning_rate": 7.136561565988242e-08, "loss": 0.156, "num_input_tokens_seen": 10008328, "step": 37325 }, { "epoch": 9.785058977719528, "grad_norm": 1.8770674467086792, "learning_rate": 7.050464524200607e-08, "loss": 0.1212, "num_input_tokens_seen": 10009528, "step": 37330 }, { "epoch": 9.786369593709043, "grad_norm": 1.0252941846847534, "learning_rate": 6.964889252443485e-08, "loss": 0.0843, "num_input_tokens_seen": 10010920, "step": 37335 }, { "epoch": 9.787680209698559, "grad_norm": 3.4392759799957275, "learning_rate": 6.879835768627274e-08, "loss": 0.2713, "num_input_tokens_seen": 10012584, "step": 37340 }, { "epoch": 9.788990825688073, "grad_norm": 2.5554087162017822, "learning_rate": 6.795304090553567e-08, "loss": 0.1341, "num_input_tokens_seen": 10013864, "step": 37345 }, { "epoch": 9.790301441677588, "grad_norm": 3.4553442001342773, "learning_rate": 6.711294235914877e-08, "loss": 0.1896, "num_input_tokens_seen": 10016008, "step": 37350 }, { "epoch": 9.791612057667104, "grad_norm": 1.7363635301589966, "learning_rate": 6.627806222294086e-08, "loss": 0.0877, "num_input_tokens_seen": 10017176, "step": 37355 }, { "epoch": 9.79292267365662, "grad_norm": 1.68206787109375, "learning_rate": 6.544840067165548e-08, "loss": 0.0864, "num_input_tokens_seen": 10018584, "step": 37360 }, { "epoch": 9.794233289646133, "grad_norm": 0.9928643107414246, "learning_rate": 6.462395787893427e-08, "loss": 0.1402, "num_input_tokens_seen": 10020392, "step": 37365 }, { "epoch": 9.795543905635649, "grad_norm": 1.1028051376342773, "learning_rate": 6.380473401733366e-08, "loss": 0.1691, "num_input_tokens_seen": 10022424, "step": 37370 }, { "epoch": 9.796854521625164, "grad_norm": 1.5838152170181274, "learning_rate": 6.299072925831373e-08, "loss": 0.1082, "num_input_tokens_seen": 10024296, "step": 37375 }, { "epoch": 9.79816513761468, "grad_norm": 5.441504955291748, "learning_rate": 6.218194377224928e-08, "loss": 0.1706, "num_input_tokens_seen": 10025304, "step": 37380 }, { "epoch": 9.799475753604193, "grad_norm": 0.5730090141296387, "learning_rate": 6.137837772841326e-08, "loss": 0.0489, "num_input_tokens_seen": 10026344, "step": 37385 }, { "epoch": 9.800786369593709, "grad_norm": 2.7586419582366943, "learning_rate": 6.058003129499334e-08, "loss": 0.1299, "num_input_tokens_seen": 10027464, "step": 37390 }, { "epoch": 9.802096985583225, "grad_norm": 3.53641414642334, "learning_rate": 5.978690463908088e-08, "loss": 0.1219, "num_input_tokens_seen": 10028680, "step": 37395 }, { "epoch": 9.80340760157274, "grad_norm": 3.8107943534851074, "learning_rate": 5.8998997926676405e-08, "loss": 0.1695, "num_input_tokens_seen": 10029928, "step": 37400 }, { "epoch": 9.804718217562254, "grad_norm": 0.1493091583251953, "learning_rate": 5.821631132268412e-08, "loss": 0.125, "num_input_tokens_seen": 10031208, "step": 37405 }, { "epoch": 9.80602883355177, "grad_norm": 2.125173330307007, "learning_rate": 5.743884499092578e-08, "loss": 0.0834, "num_input_tokens_seen": 10032136, "step": 37410 }, { "epoch": 9.807339449541285, "grad_norm": 2.6003835201263428, "learning_rate": 5.6666599094115646e-08, "loss": 0.1099, "num_input_tokens_seen": 10033176, "step": 37415 }, { "epoch": 9.808650065530799, "grad_norm": 0.2881726324558258, "learning_rate": 5.58995737938911e-08, "loss": 0.1168, "num_input_tokens_seen": 10034280, "step": 37420 }, { "epoch": 9.809960681520314, "grad_norm": 0.7880100011825562, "learning_rate": 5.513776925078207e-08, "loss": 0.1163, "num_input_tokens_seen": 10035960, "step": 37425 }, { "epoch": 9.81127129750983, "grad_norm": 2.7018661499023438, "learning_rate": 5.4381185624238776e-08, "loss": 0.0485, "num_input_tokens_seen": 10036984, "step": 37430 }, { "epoch": 9.812581913499345, "grad_norm": 2.949230432510376, "learning_rate": 5.362982307261233e-08, "loss": 0.1255, "num_input_tokens_seen": 10038008, "step": 37435 }, { "epoch": 9.813892529488859, "grad_norm": 1.4297879934310913, "learning_rate": 5.2883681753157497e-08, "loss": 0.1374, "num_input_tokens_seen": 10039112, "step": 37440 }, { "epoch": 9.815203145478375, "grad_norm": 1.8527635335922241, "learning_rate": 5.214276182204381e-08, "loss": 0.1028, "num_input_tokens_seen": 10040216, "step": 37445 }, { "epoch": 9.81651376146789, "grad_norm": 2.4659769535064697, "learning_rate": 5.140706343434165e-08, "loss": 0.1403, "num_input_tokens_seen": 10041528, "step": 37450 }, { "epoch": 9.817824377457406, "grad_norm": 5.71954870223999, "learning_rate": 5.0676586744036194e-08, "loss": 0.1066, "num_input_tokens_seen": 10042600, "step": 37455 }, { "epoch": 9.81913499344692, "grad_norm": 1.2018866539001465, "learning_rate": 4.9951331904007915e-08, "loss": 0.158, "num_input_tokens_seen": 10043960, "step": 37460 }, { "epoch": 9.820445609436435, "grad_norm": 3.1432132720947266, "learning_rate": 4.923129906606039e-08, "loss": 0.1343, "num_input_tokens_seen": 10045080, "step": 37465 }, { "epoch": 9.82175622542595, "grad_norm": 0.36548951268196106, "learning_rate": 4.851648838088696e-08, "loss": 0.1046, "num_input_tokens_seen": 10046440, "step": 37470 }, { "epoch": 9.823066841415466, "grad_norm": 2.9251182079315186, "learning_rate": 4.7806899998101283e-08, "loss": 0.1237, "num_input_tokens_seen": 10047416, "step": 37475 }, { "epoch": 9.82437745740498, "grad_norm": 2.289513349533081, "learning_rate": 4.71025340662179e-08, "loss": 0.1087, "num_input_tokens_seen": 10048648, "step": 37480 }, { "epoch": 9.825688073394495, "grad_norm": 2.3381717205047607, "learning_rate": 4.6403390732654985e-08, "loss": 0.1663, "num_input_tokens_seen": 10050008, "step": 37485 }, { "epoch": 9.82699868938401, "grad_norm": 3.4792258739471436, "learning_rate": 4.570947014374827e-08, "loss": 0.1559, "num_input_tokens_seen": 10051752, "step": 37490 }, { "epoch": 9.828309305373526, "grad_norm": 3.820192337036133, "learning_rate": 4.502077244473435e-08, "loss": 0.184, "num_input_tokens_seen": 10053032, "step": 37495 }, { "epoch": 9.82961992136304, "grad_norm": 1.3954282999038696, "learning_rate": 4.4337297779750705e-08, "loss": 0.0667, "num_input_tokens_seen": 10054104, "step": 37500 }, { "epoch": 9.830930537352556, "grad_norm": 5.460801601409912, "learning_rate": 4.365904629185236e-08, "loss": 0.0723, "num_input_tokens_seen": 10055288, "step": 37505 }, { "epoch": 9.832241153342071, "grad_norm": 1.4932693243026733, "learning_rate": 4.298601812299241e-08, "loss": 0.1856, "num_input_tokens_seen": 10056808, "step": 37510 }, { "epoch": 9.833551769331585, "grad_norm": 2.9909884929656982, "learning_rate": 4.2318213414038745e-08, "loss": 0.1168, "num_input_tokens_seen": 10057960, "step": 37515 }, { "epoch": 9.8348623853211, "grad_norm": 1.8662792444229126, "learning_rate": 4.1655632304757334e-08, "loss": 0.1311, "num_input_tokens_seen": 10059544, "step": 37520 }, { "epoch": 9.836173001310616, "grad_norm": 2.1714837551116943, "learning_rate": 4.0998274933828905e-08, "loss": 0.1012, "num_input_tokens_seen": 10060680, "step": 37525 }, { "epoch": 9.837483617300132, "grad_norm": 4.408315181732178, "learning_rate": 4.034614143883508e-08, "loss": 0.1466, "num_input_tokens_seen": 10061848, "step": 37530 }, { "epoch": 9.838794233289645, "grad_norm": 2.3431711196899414, "learning_rate": 3.969923195626668e-08, "loss": 0.1198, "num_input_tokens_seen": 10063224, "step": 37535 }, { "epoch": 9.840104849279161, "grad_norm": 3.9649112224578857, "learning_rate": 3.9057546621520946e-08, "loss": 0.0972, "num_input_tokens_seen": 10064728, "step": 37540 }, { "epoch": 9.841415465268676, "grad_norm": 1.7435944080352783, "learning_rate": 3.842108556890156e-08, "loss": 0.0511, "num_input_tokens_seen": 10065832, "step": 37545 }, { "epoch": 9.842726081258192, "grad_norm": 1.4996514320373535, "learning_rate": 3.778984893161863e-08, "loss": 0.0977, "num_input_tokens_seen": 10067112, "step": 37550 }, { "epoch": 9.844036697247706, "grad_norm": 2.6279664039611816, "learning_rate": 3.7163836841791476e-08, "loss": 0.1898, "num_input_tokens_seen": 10068264, "step": 37555 }, { "epoch": 9.845347313237221, "grad_norm": 3.2953739166259766, "learning_rate": 3.654304943043752e-08, "loss": 0.1518, "num_input_tokens_seen": 10069448, "step": 37560 }, { "epoch": 9.846657929226737, "grad_norm": 9.48287296295166, "learning_rate": 3.5927486827491696e-08, "loss": 0.1902, "num_input_tokens_seen": 10070776, "step": 37565 }, { "epoch": 9.847968545216252, "grad_norm": 0.9757031798362732, "learning_rate": 3.531714916178708e-08, "loss": 0.0726, "num_input_tokens_seen": 10071928, "step": 37570 }, { "epoch": 9.849279161205766, "grad_norm": 2.338756561279297, "learning_rate": 3.4712036561068693e-08, "loss": 0.1119, "num_input_tokens_seen": 10072888, "step": 37575 }, { "epoch": 9.850589777195282, "grad_norm": 0.6509219408035278, "learning_rate": 3.4112149151982466e-08, "loss": 0.0553, "num_input_tokens_seen": 10074008, "step": 37580 }, { "epoch": 9.851900393184797, "grad_norm": 2.6385464668273926, "learning_rate": 3.351748706008628e-08, "loss": 0.1489, "num_input_tokens_seen": 10075544, "step": 37585 }, { "epoch": 9.853211009174313, "grad_norm": 3.355116367340088, "learning_rate": 3.292805040984171e-08, "loss": 0.1933, "num_input_tokens_seen": 10076920, "step": 37590 }, { "epoch": 9.854521625163827, "grad_norm": 1.8833667039871216, "learning_rate": 3.234383932461671e-08, "loss": 0.1413, "num_input_tokens_seen": 10078456, "step": 37595 }, { "epoch": 9.855832241153342, "grad_norm": 0.5719460844993591, "learning_rate": 3.176485392668571e-08, "loss": 0.1613, "num_input_tokens_seen": 10079896, "step": 37600 }, { "epoch": 9.857142857142858, "grad_norm": 0.07313024252653122, "learning_rate": 3.119109433722955e-08, "loss": 0.1368, "num_input_tokens_seen": 10081512, "step": 37605 }, { "epoch": 9.858453473132371, "grad_norm": 3.6871070861816406, "learning_rate": 3.0622560676332734e-08, "loss": 0.306, "num_input_tokens_seen": 10083096, "step": 37610 }, { "epoch": 9.859764089121887, "grad_norm": 1.9405354261398315, "learning_rate": 3.005925306299173e-08, "loss": 0.1717, "num_input_tokens_seen": 10084632, "step": 37615 }, { "epoch": 9.861074705111402, "grad_norm": 2.053358554840088, "learning_rate": 2.9501171615103907e-08, "loss": 0.1314, "num_input_tokens_seen": 10086152, "step": 37620 }, { "epoch": 9.862385321100918, "grad_norm": 2.3347136974334717, "learning_rate": 2.8948316449473044e-08, "loss": 0.1858, "num_input_tokens_seen": 10087864, "step": 37625 }, { "epoch": 9.863695937090432, "grad_norm": 2.5222694873809814, "learning_rate": 2.840068768181492e-08, "loss": 0.1576, "num_input_tokens_seen": 10089288, "step": 37630 }, { "epoch": 9.865006553079947, "grad_norm": 1.12095046043396, "learning_rate": 2.785828542674618e-08, "loss": 0.1762, "num_input_tokens_seen": 10090552, "step": 37635 }, { "epoch": 9.866317169069463, "grad_norm": 0.45096349716186523, "learning_rate": 2.7321109797787125e-08, "loss": 0.146, "num_input_tokens_seen": 10091944, "step": 37640 }, { "epoch": 9.867627785058978, "grad_norm": 2.267075777053833, "learning_rate": 2.6789160907372822e-08, "loss": 0.1375, "num_input_tokens_seen": 10093400, "step": 37645 }, { "epoch": 9.868938401048492, "grad_norm": 0.7796200513839722, "learning_rate": 2.626243886683366e-08, "loss": 0.1358, "num_input_tokens_seen": 10094536, "step": 37650 }, { "epoch": 9.870249017038008, "grad_norm": 2.194305419921875, "learning_rate": 2.574094378641756e-08, "loss": 0.1291, "num_input_tokens_seen": 10095736, "step": 37655 }, { "epoch": 9.871559633027523, "grad_norm": 2.9655609130859375, "learning_rate": 2.5224675775270544e-08, "loss": 0.2504, "num_input_tokens_seen": 10096888, "step": 37660 }, { "epoch": 9.872870249017039, "grad_norm": 1.9312101602554321, "learning_rate": 2.4713634941442298e-08, "loss": 0.0533, "num_input_tokens_seen": 10098072, "step": 37665 }, { "epoch": 9.874180865006553, "grad_norm": 2.1217033863067627, "learning_rate": 2.4207821391900032e-08, "loss": 0.2349, "num_input_tokens_seen": 10099832, "step": 37670 }, { "epoch": 9.875491480996068, "grad_norm": 2.4736502170562744, "learning_rate": 2.370723523250351e-08, "loss": 0.059, "num_input_tokens_seen": 10101352, "step": 37675 }, { "epoch": 9.876802096985584, "grad_norm": 1.9046415090560913, "learning_rate": 2.321187656802726e-08, "loss": 0.1163, "num_input_tokens_seen": 10102632, "step": 37680 }, { "epoch": 9.8781127129751, "grad_norm": 0.7752765417098999, "learning_rate": 2.272174550214945e-08, "loss": 0.1317, "num_input_tokens_seen": 10103688, "step": 37685 }, { "epoch": 9.879423328964613, "grad_norm": 0.2764919102191925, "learning_rate": 2.223684213745192e-08, "loss": 0.0558, "num_input_tokens_seen": 10104904, "step": 37690 }, { "epoch": 9.880733944954128, "grad_norm": 2.463197708129883, "learning_rate": 2.1757166575425702e-08, "loss": 0.1807, "num_input_tokens_seen": 10106328, "step": 37695 }, { "epoch": 9.882044560943644, "grad_norm": 2.486464262008667, "learning_rate": 2.1282718916465494e-08, "loss": 0.1013, "num_input_tokens_seen": 10107496, "step": 37700 }, { "epoch": 9.883355176933158, "grad_norm": 2.110335111618042, "learning_rate": 2.0813499259872414e-08, "loss": 0.1287, "num_input_tokens_seen": 10108632, "step": 37705 }, { "epoch": 9.884665792922673, "grad_norm": 2.203706741333008, "learning_rate": 2.0349507703851244e-08, "loss": 0.1425, "num_input_tokens_seen": 10110008, "step": 37710 }, { "epoch": 9.885976408912189, "grad_norm": 2.1745455265045166, "learning_rate": 1.9890744345518742e-08, "loss": 0.1604, "num_input_tokens_seen": 10111272, "step": 37715 }, { "epoch": 9.887287024901704, "grad_norm": 1.922242522239685, "learning_rate": 1.9437209280889768e-08, "loss": 0.1352, "num_input_tokens_seen": 10112920, "step": 37720 }, { "epoch": 9.888597640891218, "grad_norm": 1.8295544385910034, "learning_rate": 1.8988902604891166e-08, "loss": 0.2724, "num_input_tokens_seen": 10114760, "step": 37725 }, { "epoch": 9.889908256880734, "grad_norm": 2.1269726753234863, "learning_rate": 1.8545824411350665e-08, "loss": 0.267, "num_input_tokens_seen": 10116120, "step": 37730 }, { "epoch": 9.89121887287025, "grad_norm": 2.070903778076172, "learning_rate": 1.810797479300519e-08, "loss": 0.1483, "num_input_tokens_seen": 10117784, "step": 37735 }, { "epoch": 9.892529488859765, "grad_norm": 1.5993167161941528, "learning_rate": 1.7675353841495325e-08, "loss": 0.1647, "num_input_tokens_seen": 10119112, "step": 37740 }, { "epoch": 9.893840104849279, "grad_norm": 4.565347194671631, "learning_rate": 1.7247961647368082e-08, "loss": 0.2018, "num_input_tokens_seen": 10120424, "step": 37745 }, { "epoch": 9.895150720838794, "grad_norm": 3.20430326461792, "learning_rate": 1.6825798300074137e-08, "loss": 0.1114, "num_input_tokens_seen": 10121592, "step": 37750 }, { "epoch": 9.89646133682831, "grad_norm": 0.4800178110599518, "learning_rate": 1.640886388797336e-08, "loss": 0.07, "num_input_tokens_seen": 10122600, "step": 37755 }, { "epoch": 9.897771952817825, "grad_norm": 1.5342737436294556, "learning_rate": 1.5997158498329277e-08, "loss": 0.1285, "num_input_tokens_seen": 10123880, "step": 37760 }, { "epoch": 9.899082568807339, "grad_norm": 1.9078214168548584, "learning_rate": 1.559068221731186e-08, "loss": 0.0753, "num_input_tokens_seen": 10125112, "step": 37765 }, { "epoch": 9.900393184796854, "grad_norm": 2.1263208389282227, "learning_rate": 1.518943512999471e-08, "loss": 0.1168, "num_input_tokens_seen": 10126280, "step": 37770 }, { "epoch": 9.90170380078637, "grad_norm": 2.0671064853668213, "learning_rate": 1.4793417320357882e-08, "loss": 0.1289, "num_input_tokens_seen": 10128024, "step": 37775 }, { "epoch": 9.903014416775886, "grad_norm": 0.9553855657577515, "learning_rate": 1.4402628871285073e-08, "loss": 0.1358, "num_input_tokens_seen": 10129528, "step": 37780 }, { "epoch": 9.9043250327654, "grad_norm": 2.6904337406158447, "learning_rate": 1.401706986457474e-08, "loss": 0.1429, "num_input_tokens_seen": 10131192, "step": 37785 }, { "epoch": 9.905635648754915, "grad_norm": 4.935583114624023, "learning_rate": 1.3636740380915113e-08, "loss": 0.2331, "num_input_tokens_seen": 10132424, "step": 37790 }, { "epoch": 9.90694626474443, "grad_norm": 3.9408562183380127, "learning_rate": 1.326164049991474e-08, "loss": 0.1162, "num_input_tokens_seen": 10133816, "step": 37795 }, { "epoch": 9.908256880733944, "grad_norm": 1.3529729843139648, "learning_rate": 1.2891770300080263e-08, "loss": 0.194, "num_input_tokens_seen": 10135176, "step": 37800 }, { "epoch": 9.90956749672346, "grad_norm": 1.075871467590332, "learning_rate": 1.2527129858821985e-08, "loss": 0.154, "num_input_tokens_seen": 10136376, "step": 37805 }, { "epoch": 9.910878112712975, "grad_norm": 1.760664463043213, "learning_rate": 1.2167719252462183e-08, "loss": 0.2252, "num_input_tokens_seen": 10138408, "step": 37810 }, { "epoch": 9.91218872870249, "grad_norm": 2.9542906284332275, "learning_rate": 1.1813538556221249e-08, "loss": 0.1162, "num_input_tokens_seen": 10139672, "step": 37815 }, { "epoch": 9.913499344692005, "grad_norm": 2.2700345516204834, "learning_rate": 1.1464587844231544e-08, "loss": 0.2525, "num_input_tokens_seen": 10141624, "step": 37820 }, { "epoch": 9.91480996068152, "grad_norm": 2.8707475662231445, "learning_rate": 1.1120867189526318e-08, "loss": 0.1496, "num_input_tokens_seen": 10142920, "step": 37825 }, { "epoch": 9.916120576671036, "grad_norm": 1.12582266330719, "learning_rate": 1.0782376664045246e-08, "loss": 0.1283, "num_input_tokens_seen": 10144664, "step": 37830 }, { "epoch": 9.917431192660551, "grad_norm": 2.1944265365600586, "learning_rate": 1.044911633863721e-08, "loss": 0.0768, "num_input_tokens_seen": 10146008, "step": 37835 }, { "epoch": 9.918741808650065, "grad_norm": 2.808959484100342, "learning_rate": 1.0121086283049198e-08, "loss": 0.1012, "num_input_tokens_seen": 10147608, "step": 37840 }, { "epoch": 9.92005242463958, "grad_norm": 2.5309932231903076, "learning_rate": 9.798286565937399e-09, "loss": 0.184, "num_input_tokens_seen": 10148712, "step": 37845 }, { "epoch": 9.921363040629096, "grad_norm": 1.0472512245178223, "learning_rate": 9.48071725486721e-09, "loss": 0.1798, "num_input_tokens_seen": 10150040, "step": 37850 }, { "epoch": 9.922673656618612, "grad_norm": 0.5333071947097778, "learning_rate": 9.168378416299362e-09, "loss": 0.1222, "num_input_tokens_seen": 10152024, "step": 37855 }, { "epoch": 9.923984272608125, "grad_norm": 9.200300216674805, "learning_rate": 8.861270115612108e-09, "loss": 0.2142, "num_input_tokens_seen": 10153480, "step": 37860 }, { "epoch": 9.92529488859764, "grad_norm": 1.5046266317367554, "learning_rate": 8.559392417079038e-09, "loss": 0.1423, "num_input_tokens_seen": 10154936, "step": 37865 }, { "epoch": 9.926605504587156, "grad_norm": 1.7145956754684448, "learning_rate": 8.262745383880166e-09, "loss": 0.0476, "num_input_tokens_seen": 10155960, "step": 37870 }, { "epoch": 9.927916120576672, "grad_norm": 2.589975118637085, "learning_rate": 7.971329078110268e-09, "loss": 0.2281, "num_input_tokens_seen": 10157160, "step": 37875 }, { "epoch": 9.929226736566186, "grad_norm": 1.8345918655395508, "learning_rate": 7.68514356075667e-09, "loss": 0.1047, "num_input_tokens_seen": 10158472, "step": 37880 }, { "epoch": 9.930537352555701, "grad_norm": 2.3517587184906006, "learning_rate": 7.4041888917186775e-09, "loss": 0.2136, "num_input_tokens_seen": 10159832, "step": 37885 }, { "epoch": 9.931847968545217, "grad_norm": 1.4413022994995117, "learning_rate": 7.1284651297992556e-09, "loss": 0.0821, "num_input_tokens_seen": 10161096, "step": 37890 }, { "epoch": 9.93315858453473, "grad_norm": 1.7480137348175049, "learning_rate": 6.8579723327105715e-09, "loss": 0.0965, "num_input_tokens_seen": 10162616, "step": 37895 }, { "epoch": 9.934469200524246, "grad_norm": 2.351471424102783, "learning_rate": 6.5927105570601224e-09, "loss": 0.1408, "num_input_tokens_seen": 10163912, "step": 37900 }, { "epoch": 9.935779816513762, "grad_norm": 2.0824148654937744, "learning_rate": 6.3326798583729364e-09, "loss": 0.1257, "num_input_tokens_seen": 10165224, "step": 37905 }, { "epoch": 9.937090432503277, "grad_norm": 2.686082601547241, "learning_rate": 6.077880291069371e-09, "loss": 0.1584, "num_input_tokens_seen": 10166904, "step": 37910 }, { "epoch": 9.938401048492791, "grad_norm": 2.6181628704071045, "learning_rate": 5.8283119084789895e-09, "loss": 0.1462, "num_input_tokens_seen": 10168280, "step": 37915 }, { "epoch": 9.939711664482306, "grad_norm": 1.0779887437820435, "learning_rate": 5.583974762835009e-09, "loss": 0.1491, "num_input_tokens_seen": 10169624, "step": 37920 }, { "epoch": 9.941022280471822, "grad_norm": 3.420417308807373, "learning_rate": 5.344868905279854e-09, "loss": 0.0979, "num_input_tokens_seen": 10170840, "step": 37925 }, { "epoch": 9.942332896461338, "grad_norm": 0.5921236872673035, "learning_rate": 5.110994385856826e-09, "loss": 0.092, "num_input_tokens_seen": 10172008, "step": 37930 }, { "epoch": 9.943643512450851, "grad_norm": 1.9504531621932983, "learning_rate": 4.8823512535128845e-09, "loss": 0.0853, "num_input_tokens_seen": 10174936, "step": 37935 }, { "epoch": 9.944954128440367, "grad_norm": 2.1762890815734863, "learning_rate": 4.658939556104191e-09, "loss": 0.2499, "num_input_tokens_seen": 10176232, "step": 37940 }, { "epoch": 9.946264744429882, "grad_norm": 3.8463587760925293, "learning_rate": 4.440759340393341e-09, "loss": 0.1635, "num_input_tokens_seen": 10177416, "step": 37945 }, { "epoch": 9.947575360419398, "grad_norm": 2.193530559539795, "learning_rate": 4.227810652041031e-09, "loss": 0.0945, "num_input_tokens_seen": 10178792, "step": 37950 }, { "epoch": 9.948885976408912, "grad_norm": 1.7719398736953735, "learning_rate": 4.0200935356171645e-09, "loss": 0.0634, "num_input_tokens_seen": 10179912, "step": 37955 }, { "epoch": 9.950196592398427, "grad_norm": 1.6331030130386353, "learning_rate": 3.817608034600851e-09, "loss": 0.0964, "num_input_tokens_seen": 10181160, "step": 37960 }, { "epoch": 9.951507208387943, "grad_norm": 3.1762661933898926, "learning_rate": 3.620354191366526e-09, "loss": 0.2594, "num_input_tokens_seen": 10182376, "step": 37965 }, { "epoch": 9.952817824377458, "grad_norm": 2.8444414138793945, "learning_rate": 3.4283320472033864e-09, "loss": 0.2145, "num_input_tokens_seen": 10183816, "step": 37970 }, { "epoch": 9.954128440366972, "grad_norm": 3.3300461769104004, "learning_rate": 3.241541642298729e-09, "loss": 0.2383, "num_input_tokens_seen": 10185016, "step": 37975 }, { "epoch": 9.955439056356488, "grad_norm": 0.5807785391807556, "learning_rate": 3.059983015749057e-09, "loss": 0.124, "num_input_tokens_seen": 10186248, "step": 37980 }, { "epoch": 9.956749672346003, "grad_norm": 1.3740290403366089, "learning_rate": 2.88365620555453e-09, "loss": 0.0986, "num_input_tokens_seen": 10187944, "step": 37985 }, { "epoch": 9.958060288335517, "grad_norm": 3.979473352432251, "learning_rate": 2.712561248618961e-09, "loss": 0.1467, "num_input_tokens_seen": 10189064, "step": 37990 }, { "epoch": 9.959370904325032, "grad_norm": 1.2816872596740723, "learning_rate": 2.546698180749818e-09, "loss": 0.1451, "num_input_tokens_seen": 10190616, "step": 37995 }, { "epoch": 9.960681520314548, "grad_norm": 2.3440730571746826, "learning_rate": 2.3860670366665505e-09, "loss": 0.1541, "num_input_tokens_seen": 10192200, "step": 38000 }, { "epoch": 9.961992136304064, "grad_norm": 3.1041648387908936, "learning_rate": 2.2306678499867116e-09, "loss": 0.1491, "num_input_tokens_seen": 10193928, "step": 38005 }, { "epoch": 9.963302752293577, "grad_norm": 2.5349721908569336, "learning_rate": 2.080500653234285e-09, "loss": 0.1615, "num_input_tokens_seen": 10195240, "step": 38010 }, { "epoch": 9.964613368283093, "grad_norm": 0.633821427822113, "learning_rate": 1.935565477839685e-09, "loss": 0.1564, "num_input_tokens_seen": 10197112, "step": 38015 }, { "epoch": 9.965923984272608, "grad_norm": 2.493147850036621, "learning_rate": 1.7958623541397546e-09, "loss": 0.1546, "num_input_tokens_seen": 10198408, "step": 38020 }, { "epoch": 9.967234600262124, "grad_norm": 1.8063607215881348, "learning_rate": 1.6613913113694424e-09, "loss": 0.0943, "num_input_tokens_seen": 10199784, "step": 38025 }, { "epoch": 9.968545216251638, "grad_norm": 0.5968029499053955, "learning_rate": 1.532152377678453e-09, "loss": 0.0758, "num_input_tokens_seen": 10200824, "step": 38030 }, { "epoch": 9.969855832241153, "grad_norm": 1.5907138586044312, "learning_rate": 1.4081455801145948e-09, "loss": 0.1086, "num_input_tokens_seen": 10201912, "step": 38035 }, { "epoch": 9.971166448230669, "grad_norm": 1.1253728866577148, "learning_rate": 1.289370944629331e-09, "loss": 0.0755, "num_input_tokens_seen": 10202856, "step": 38040 }, { "epoch": 9.972477064220184, "grad_norm": 1.747207522392273, "learning_rate": 1.1758284960861066e-09, "loss": 0.1101, "num_input_tokens_seen": 10203896, "step": 38045 }, { "epoch": 9.973787680209698, "grad_norm": 0.48657846450805664, "learning_rate": 1.0675182582464693e-09, "loss": 0.1004, "num_input_tokens_seen": 10205272, "step": 38050 }, { "epoch": 9.975098296199214, "grad_norm": 2.388117551803589, "learning_rate": 9.644402537811736e-10, "loss": 0.1659, "num_input_tokens_seen": 10206936, "step": 38055 }, { "epoch": 9.97640891218873, "grad_norm": 3.7141404151916504, "learning_rate": 8.665945042618529e-10, "loss": 0.0795, "num_input_tokens_seen": 10207912, "step": 38060 }, { "epoch": 9.977719528178245, "grad_norm": 1.771722435951233, "learning_rate": 7.739810301693462e-10, "loss": 0.1725, "num_input_tokens_seen": 10208904, "step": 38065 }, { "epoch": 9.979030144167758, "grad_norm": 2.431589365005493, "learning_rate": 6.865998508881477e-10, "loss": 0.2182, "num_input_tokens_seen": 10209928, "step": 38070 }, { "epoch": 9.980340760157274, "grad_norm": 1.579350233078003, "learning_rate": 6.044509847064062e-10, "loss": 0.0792, "num_input_tokens_seen": 10211368, "step": 38075 }, { "epoch": 9.98165137614679, "grad_norm": 2.4286813735961914, "learning_rate": 5.275344488187006e-10, "loss": 0.1433, "num_input_tokens_seen": 10212600, "step": 38080 }, { "epoch": 9.982961992136303, "grad_norm": 1.233528733253479, "learning_rate": 4.5585025932048943e-10, "loss": 0.1122, "num_input_tokens_seen": 10213656, "step": 38085 }, { "epoch": 9.984272608125819, "grad_norm": 2.5066542625427246, "learning_rate": 3.893984312164367e-10, "loss": 0.2141, "num_input_tokens_seen": 10215272, "step": 38090 }, { "epoch": 9.985583224115334, "grad_norm": 1.821300745010376, "learning_rate": 3.28178978417637e-10, "loss": 0.0881, "num_input_tokens_seen": 10216776, "step": 38095 }, { "epoch": 9.98689384010485, "grad_norm": 0.9451442956924438, "learning_rate": 2.7219191373328846e-10, "loss": 0.0894, "num_input_tokens_seen": 10217960, "step": 38100 }, { "epoch": 9.988204456094364, "grad_norm": 0.9172111749649048, "learning_rate": 2.2143724888179506e-10, "loss": 0.0501, "num_input_tokens_seen": 10219192, "step": 38105 }, { "epoch": 9.98951507208388, "grad_norm": 2.6162350177764893, "learning_rate": 1.759149944879912e-10, "loss": 0.2575, "num_input_tokens_seen": 10220904, "step": 38110 }, { "epoch": 9.990825688073395, "grad_norm": 2.8424627780914307, "learning_rate": 1.3562516008036597e-10, "loss": 0.2622, "num_input_tokens_seen": 10222152, "step": 38115 }, { "epoch": 9.99213630406291, "grad_norm": 2.0516107082366943, "learning_rate": 1.0056775408828767e-10, "loss": 0.1456, "num_input_tokens_seen": 10223672, "step": 38120 }, { "epoch": 9.993446920052424, "grad_norm": 2.7187840938568115, "learning_rate": 7.07427838503305e-11, "loss": 0.2053, "num_input_tokens_seen": 10225736, "step": 38125 }, { "epoch": 9.99475753604194, "grad_norm": 1.7900035381317139, "learning_rate": 4.6150255611498996e-11, "loss": 0.1426, "num_input_tokens_seen": 10227320, "step": 38130 }, { "epoch": 9.996068152031455, "grad_norm": 1.4247218370437622, "learning_rate": 2.679017451490129e-11, "loss": 0.1232, "num_input_tokens_seen": 10228504, "step": 38135 }, { "epoch": 9.997378768020969, "grad_norm": 2.9804153442382812, "learning_rate": 1.2662544615626993e-11, "loss": 0.2702, "num_input_tokens_seen": 10229624, "step": 38140 }, { "epoch": 9.998689384010484, "grad_norm": 0.5244209170341492, "learning_rate": 3.767368869644905e-12, "loss": 0.2777, "num_input_tokens_seen": 10231048, "step": 38145 }, { "epoch": 10.0, "grad_norm": 2.7214128971099854, "learning_rate": 1.0464913657859399e-13, "loss": 0.256, "num_input_tokens_seen": 10232192, "step": 38150 }, { "epoch": 10.0, "num_input_tokens_seen": 10232192, "step": 38150, "total_flos": 4.6075098113664614e+17, "train_loss": 0.36039121026702975, "train_runtime": 3784.8796, "train_samples_per_second": 20.157, "train_steps_per_second": 10.08 } ], "logging_steps": 5, "max_steps": 38150, "num_input_tokens_seen": 10232192, "num_train_epochs": 10, "save_steps": 1908, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 4.6075098113664614e+17, "train_batch_size": 2, "trial_name": null, "trial_params": null }