{ "best_global_step": null, "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 500, "global_step": 4674, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0021399529210357373, "grad_norm": 2.7134010791778564, "learning_rate": 3.5460992907801423e-06, "loss": 12.9166, "step": 5 }, { "epoch": 0.004279905842071475, "grad_norm": 2.857581853866577, "learning_rate": 7.092198581560285e-06, "loss": 12.8988, "step": 10 }, { "epoch": 0.006419858763107212, "grad_norm": 3.6460659503936768, "learning_rate": 1.0638297872340426e-05, "loss": 12.596, "step": 15 }, { "epoch": 0.00855981168414295, "grad_norm": 4.4102678298950195, "learning_rate": 1.418439716312057e-05, "loss": 12.0487, "step": 20 }, { "epoch": 0.010699764605178685, "grad_norm": 4.709506034851074, "learning_rate": 1.773049645390071e-05, "loss": 11.2542, "step": 25 }, { "epoch": 0.012839717526214423, "grad_norm": 3.634777069091797, "learning_rate": 2.1276595744680852e-05, "loss": 10.3125, "step": 30 }, { "epoch": 0.014979670447250161, "grad_norm": 3.215787649154663, "learning_rate": 2.4822695035460995e-05, "loss": 9.4876, "step": 35 }, { "epoch": 0.0171196233682859, "grad_norm": 2.2986671924591064, "learning_rate": 2.836879432624114e-05, "loss": 8.8462, "step": 40 }, { "epoch": 0.019259576289321637, "grad_norm": 3.012507677078247, "learning_rate": 3.191489361702128e-05, "loss": 8.3983, "step": 45 }, { "epoch": 0.02139952921035737, "grad_norm": 1.9252734184265137, "learning_rate": 3.546099290780142e-05, "loss": 7.9777, "step": 50 }, { "epoch": 0.02353948213139311, "grad_norm": 4.199497699737549, "learning_rate": 3.900709219858156e-05, "loss": 7.6574, "step": 55 }, { "epoch": 0.025679435052428846, "grad_norm": 6.262833118438721, "learning_rate": 4.2553191489361704e-05, "loss": 7.4121, "step": 60 }, { "epoch": 0.027819387973464584, "grad_norm": 6.515402317047119, "learning_rate": 4.609929078014185e-05, "loss": 7.1827, "step": 65 }, { "epoch": 0.029959340894500322, "grad_norm": 2.35896372795105, "learning_rate": 4.964539007092199e-05, "loss": 6.9596, "step": 70 }, { "epoch": 0.032099293815536056, "grad_norm": 3.057620048522949, "learning_rate": 5.319148936170213e-05, "loss": 6.7407, "step": 75 }, { "epoch": 0.0342392467365718, "grad_norm": 2.1466054916381836, "learning_rate": 5.673758865248228e-05, "loss": 6.4509, "step": 80 }, { "epoch": 0.03637919965760753, "grad_norm": 4.194606781005859, "learning_rate": 6.0283687943262414e-05, "loss": 6.3802, "step": 85 }, { "epoch": 0.03851915257864327, "grad_norm": 2.6796622276306152, "learning_rate": 6.382978723404256e-05, "loss": 6.1995, "step": 90 }, { "epoch": 0.04065910549967901, "grad_norm": 4.852718353271484, "learning_rate": 6.737588652482269e-05, "loss": 5.9464, "step": 95 }, { "epoch": 0.04279905842071474, "grad_norm": 4.550829887390137, "learning_rate": 7.092198581560284e-05, "loss": 5.7867, "step": 100 }, { "epoch": 0.04493901134175048, "grad_norm": 5.130778789520264, "learning_rate": 7.446808510638298e-05, "loss": 5.6702, "step": 105 }, { "epoch": 0.04707896426278622, "grad_norm": 4.426118850708008, "learning_rate": 7.801418439716312e-05, "loss": 5.5532, "step": 110 }, { "epoch": 0.04921891718382196, "grad_norm": 3.7589707374572754, "learning_rate": 8.156028368794327e-05, "loss": 5.2532, "step": 115 }, { "epoch": 0.05135887010485769, "grad_norm": 4.644890785217285, "learning_rate": 8.510638297872341e-05, "loss": 5.1286, "step": 120 }, { "epoch": 0.05349882302589343, "grad_norm": 3.4871771335601807, "learning_rate": 8.865248226950354e-05, "loss": 4.9816, "step": 125 }, { "epoch": 0.05563877594692917, "grad_norm": 6.377041339874268, "learning_rate": 9.21985815602837e-05, "loss": 4.855, "step": 130 }, { "epoch": 0.0577787288679649, "grad_norm": 4.984015941619873, "learning_rate": 9.574468085106384e-05, "loss": 4.6588, "step": 135 }, { "epoch": 0.059918681789000644, "grad_norm": 5.524806976318359, "learning_rate": 9.929078014184398e-05, "loss": 4.5006, "step": 140 }, { "epoch": 0.06205863471003638, "grad_norm": 6.326210021972656, "learning_rate": 9.999980787316854e-05, "loss": 4.5593, "step": 145 }, { "epoch": 0.06419858763107211, "grad_norm": 6.40717887878418, "learning_rate": 9.999902736044627e-05, "loss": 4.2919, "step": 150 }, { "epoch": 0.06633854055210785, "grad_norm": 5.881182670593262, "learning_rate": 9.999764646327135e-05, "loss": 4.2351, "step": 155 }, { "epoch": 0.0684784934731436, "grad_norm": 7.092776298522949, "learning_rate": 9.99956651982255e-05, "loss": 4.1024, "step": 160 }, { "epoch": 0.07061844639417933, "grad_norm": 6.636219501495361, "learning_rate": 9.999308358909955e-05, "loss": 4.0457, "step": 165 }, { "epoch": 0.07275839931521506, "grad_norm": 6.4725141525268555, "learning_rate": 9.998990166689332e-05, "loss": 3.9062, "step": 170 }, { "epoch": 0.0748983522362508, "grad_norm": 9.732144355773926, "learning_rate": 9.998611946981506e-05, "loss": 3.9227, "step": 175 }, { "epoch": 0.07703830515728655, "grad_norm": 8.540387153625488, "learning_rate": 9.998173704328112e-05, "loss": 3.7503, "step": 180 }, { "epoch": 0.07917825807832228, "grad_norm": 7.108048915863037, "learning_rate": 9.99767544399153e-05, "loss": 3.6129, "step": 185 }, { "epoch": 0.08131821099935801, "grad_norm": 7.507834434509277, "learning_rate": 9.997117171954835e-05, "loss": 3.6377, "step": 190 }, { "epoch": 0.08345816392039375, "grad_norm": 7.859340667724609, "learning_rate": 9.996498894921713e-05, "loss": 3.6035, "step": 195 }, { "epoch": 0.08559811684142948, "grad_norm": 7.093273162841797, "learning_rate": 9.995820620316386e-05, "loss": 3.5484, "step": 200 }, { "epoch": 0.08773806976246523, "grad_norm": 8.989423751831055, "learning_rate": 9.995082356283525e-05, "loss": 3.5132, "step": 205 }, { "epoch": 0.08987802268350097, "grad_norm": 8.387621879577637, "learning_rate": 9.994284111688145e-05, "loss": 3.3861, "step": 210 }, { "epoch": 0.0920179756045367, "grad_norm": 6.809689044952393, "learning_rate": 9.993425896115509e-05, "loss": 3.37, "step": 215 }, { "epoch": 0.09415792852557243, "grad_norm": 7.266931056976318, "learning_rate": 9.992507719870998e-05, "loss": 3.4136, "step": 220 }, { "epoch": 0.09629788144660817, "grad_norm": 6.893076419830322, "learning_rate": 9.991529593980006e-05, "loss": 3.3817, "step": 225 }, { "epoch": 0.09843783436764392, "grad_norm": 6.13765287399292, "learning_rate": 9.990491530187791e-05, "loss": 3.3894, "step": 230 }, { "epoch": 0.10057778728867965, "grad_norm": 8.638238906860352, "learning_rate": 9.989393540959343e-05, "loss": 3.2748, "step": 235 }, { "epoch": 0.10271774020971539, "grad_norm": 7.530318737030029, "learning_rate": 9.98823563947923e-05, "loss": 3.1735, "step": 240 }, { "epoch": 0.10485769313075112, "grad_norm": 7.243378639221191, "learning_rate": 9.987017839651447e-05, "loss": 3.186, "step": 245 }, { "epoch": 0.10699764605178685, "grad_norm": 8.053439140319824, "learning_rate": 9.985740156099239e-05, "loss": 3.2066, "step": 250 }, { "epoch": 0.1091375989728226, "grad_norm": 7.392351150512695, "learning_rate": 9.984402604164928e-05, "loss": 3.132, "step": 255 }, { "epoch": 0.11127755189385834, "grad_norm": 6.198112487792969, "learning_rate": 9.983005199909738e-05, "loss": 3.1682, "step": 260 }, { "epoch": 0.11341750481489407, "grad_norm": 7.847325801849365, "learning_rate": 9.981547960113591e-05, "loss": 2.9939, "step": 265 }, { "epoch": 0.1155574577359298, "grad_norm": 6.414429664611816, "learning_rate": 9.980030902274907e-05, "loss": 2.9657, "step": 270 }, { "epoch": 0.11769741065696554, "grad_norm": 7.576863765716553, "learning_rate": 9.9784540446104e-05, "loss": 3.0261, "step": 275 }, { "epoch": 0.11983736357800129, "grad_norm": 7.394157886505127, "learning_rate": 9.976817406054856e-05, "loss": 3.0425, "step": 280 }, { "epoch": 0.12197731649903702, "grad_norm": 6.541194915771484, "learning_rate": 9.975121006260905e-05, "loss": 2.9808, "step": 285 }, { "epoch": 0.12411726942007276, "grad_norm": 7.00990104675293, "learning_rate": 9.973364865598783e-05, "loss": 3.0134, "step": 290 }, { "epoch": 0.1262572223411085, "grad_norm": 8.408513069152832, "learning_rate": 9.97154900515609e-05, "loss": 2.9945, "step": 295 }, { "epoch": 0.12839717526214423, "grad_norm": 6.816072463989258, "learning_rate": 9.96967344673754e-05, "loss": 2.968, "step": 300 }, { "epoch": 0.13053712818317997, "grad_norm": 7.750121593475342, "learning_rate": 9.967738212864692e-05, "loss": 2.9725, "step": 305 }, { "epoch": 0.1326770811042157, "grad_norm": 7.111683368682861, "learning_rate": 9.965743326775686e-05, "loss": 2.9475, "step": 310 }, { "epoch": 0.13481703402525144, "grad_norm": 9.491382598876953, "learning_rate": 9.963688812424958e-05, "loss": 2.9236, "step": 315 }, { "epoch": 0.1369569869462872, "grad_norm": 9.146512031555176, "learning_rate": 9.96157469448296e-05, "loss": 2.8553, "step": 320 }, { "epoch": 0.1390969398673229, "grad_norm": 10.622199058532715, "learning_rate": 9.959400998335855e-05, "loss": 2.7986, "step": 325 }, { "epoch": 0.14123689278835866, "grad_norm": 9.17403793334961, "learning_rate": 9.957167750085217e-05, "loss": 2.8332, "step": 330 }, { "epoch": 0.1433768457093944, "grad_norm": 8.492594718933105, "learning_rate": 9.95487497654772e-05, "loss": 2.8036, "step": 335 }, { "epoch": 0.14551679863043013, "grad_norm": 7.058006763458252, "learning_rate": 9.95252270525481e-05, "loss": 2.8065, "step": 340 }, { "epoch": 0.14765675155146588, "grad_norm": 8.662881851196289, "learning_rate": 9.950110964452382e-05, "loss": 2.829, "step": 345 }, { "epoch": 0.1497967044725016, "grad_norm": 8.287664413452148, "learning_rate": 9.947639783100429e-05, "loss": 2.8611, "step": 350 }, { "epoch": 0.15193665739353734, "grad_norm": 6.383536338806152, "learning_rate": 9.945109190872706e-05, "loss": 2.719, "step": 355 }, { "epoch": 0.1540766103145731, "grad_norm": 9.04951000213623, "learning_rate": 9.94251921815637e-05, "loss": 2.7782, "step": 360 }, { "epoch": 0.1562165632356088, "grad_norm": 9.873688697814941, "learning_rate": 9.939869896051613e-05, "loss": 2.7622, "step": 365 }, { "epoch": 0.15835651615664456, "grad_norm": 8.37444019317627, "learning_rate": 9.93716125637129e-05, "loss": 2.7327, "step": 370 }, { "epoch": 0.16049646907768028, "grad_norm": 8.676753997802734, "learning_rate": 9.934393331640536e-05, "loss": 2.6887, "step": 375 }, { "epoch": 0.16263642199871603, "grad_norm": 8.401046752929688, "learning_rate": 9.931566155096378e-05, "loss": 2.6337, "step": 380 }, { "epoch": 0.16477637491975178, "grad_norm": 9.715431213378906, "learning_rate": 9.928679760687333e-05, "loss": 2.7182, "step": 385 }, { "epoch": 0.1669163278407875, "grad_norm": 7.727181434631348, "learning_rate": 9.925734183073001e-05, "loss": 2.7088, "step": 390 }, { "epoch": 0.16905628076182325, "grad_norm": 8.17239761352539, "learning_rate": 9.92272945762365e-05, "loss": 2.6574, "step": 395 }, { "epoch": 0.17119623368285897, "grad_norm": 9.848095893859863, "learning_rate": 9.919665620419792e-05, "loss": 2.6521, "step": 400 }, { "epoch": 0.17333618660389472, "grad_norm": 6.269837856292725, "learning_rate": 9.916542708251745e-05, "loss": 2.6658, "step": 405 }, { "epoch": 0.17547613952493046, "grad_norm": 9.290328979492188, "learning_rate": 9.913360758619199e-05, "loss": 2.6467, "step": 410 }, { "epoch": 0.17761609244596618, "grad_norm": 8.005675315856934, "learning_rate": 9.910119809730759e-05, "loss": 2.7059, "step": 415 }, { "epoch": 0.17975604536700193, "grad_norm": 7.254702091217041, "learning_rate": 9.906819900503486e-05, "loss": 2.6869, "step": 420 }, { "epoch": 0.18189599828803765, "grad_norm": 8.958318710327148, "learning_rate": 9.903461070562436e-05, "loss": 2.6647, "step": 425 }, { "epoch": 0.1840359512090734, "grad_norm": 10.019506454467773, "learning_rate": 9.900043360240181e-05, "loss": 2.6647, "step": 430 }, { "epoch": 0.18617590413010915, "grad_norm": 9.011824607849121, "learning_rate": 9.89656681057632e-05, "loss": 2.6462, "step": 435 }, { "epoch": 0.18831585705114487, "grad_norm": 8.521088600158691, "learning_rate": 9.893031463316996e-05, "loss": 2.6276, "step": 440 }, { "epoch": 0.19045580997218062, "grad_norm": 9.127740859985352, "learning_rate": 9.889437360914379e-05, "loss": 2.5787, "step": 445 }, { "epoch": 0.19259576289321634, "grad_norm": 7.98406457901001, "learning_rate": 9.885784546526177e-05, "loss": 2.6082, "step": 450 }, { "epoch": 0.19473571581425209, "grad_norm": 8.42713451385498, "learning_rate": 9.882073064015102e-05, "loss": 2.5825, "step": 455 }, { "epoch": 0.19687566873528783, "grad_norm": 10.31495189666748, "learning_rate": 9.87830295794835e-05, "loss": 2.5767, "step": 460 }, { "epoch": 0.19901562165632355, "grad_norm": 8.118849754333496, "learning_rate": 9.874474273597059e-05, "loss": 2.5494, "step": 465 }, { "epoch": 0.2011555745773593, "grad_norm": 7.514265537261963, "learning_rate": 9.870587056935777e-05, "loss": 2.5824, "step": 470 }, { "epoch": 0.20329552749839502, "grad_norm": 9.503680229187012, "learning_rate": 9.866641354641901e-05, "loss": 2.5055, "step": 475 }, { "epoch": 0.20543548041943077, "grad_norm": 6.921200275421143, "learning_rate": 9.862637214095121e-05, "loss": 2.628, "step": 480 }, { "epoch": 0.20757543334046652, "grad_norm": 6.561913013458252, "learning_rate": 9.858574683376844e-05, "loss": 2.5504, "step": 485 }, { "epoch": 0.20971538626150224, "grad_norm": 8.711860656738281, "learning_rate": 9.854453811269625e-05, "loss": 2.6222, "step": 490 }, { "epoch": 0.211855339182538, "grad_norm": 7.749279975891113, "learning_rate": 9.85027464725658e-05, "loss": 2.4887, "step": 495 }, { "epoch": 0.2139952921035737, "grad_norm": 9.228084564208984, "learning_rate": 9.846037241520782e-05, "loss": 2.5844, "step": 500 }, { "epoch": 0.21613524502460946, "grad_norm": 10.207733154296875, "learning_rate": 9.841741644944675e-05, "loss": 2.5684, "step": 505 }, { "epoch": 0.2182751979456452, "grad_norm": 6.548727512359619, "learning_rate": 9.837387909109452e-05, "loss": 2.5237, "step": 510 }, { "epoch": 0.22041515086668093, "grad_norm": 9.262266159057617, "learning_rate": 9.832976086294432e-05, "loss": 2.5004, "step": 515 }, { "epoch": 0.22255510378771667, "grad_norm": 8.879256248474121, "learning_rate": 9.828506229476444e-05, "loss": 2.5217, "step": 520 }, { "epoch": 0.2246950567087524, "grad_norm": 9.322160720825195, "learning_rate": 9.823978392329183e-05, "loss": 2.5006, "step": 525 }, { "epoch": 0.22683500962978814, "grad_norm": 11.270200729370117, "learning_rate": 9.819392629222568e-05, "loss": 2.5517, "step": 530 }, { "epoch": 0.2289749625508239, "grad_norm": 7.910391330718994, "learning_rate": 9.814748995222085e-05, "loss": 2.4775, "step": 535 }, { "epoch": 0.2311149154718596, "grad_norm": 11.13558578491211, "learning_rate": 9.810047546088133e-05, "loss": 2.5364, "step": 540 }, { "epoch": 0.23325486839289536, "grad_norm": 8.814207077026367, "learning_rate": 9.805288338275352e-05, "loss": 2.4367, "step": 545 }, { "epoch": 0.23539482131393108, "grad_norm": 8.249216079711914, "learning_rate": 9.800471428931939e-05, "loss": 2.4926, "step": 550 }, { "epoch": 0.23753477423496683, "grad_norm": 8.129894256591797, "learning_rate": 9.795596875898967e-05, "loss": 2.5597, "step": 555 }, { "epoch": 0.23967472715600258, "grad_norm": 7.8779683113098145, "learning_rate": 9.790664737709696e-05, "loss": 2.4822, "step": 560 }, { "epoch": 0.2418146800770383, "grad_norm": 9.75910472869873, "learning_rate": 9.785675073588855e-05, "loss": 2.459, "step": 565 }, { "epoch": 0.24395463299807404, "grad_norm": 10.04244613647461, "learning_rate": 9.78062794345195e-05, "loss": 2.4718, "step": 570 }, { "epoch": 0.2460945859191098, "grad_norm": 11.239801406860352, "learning_rate": 9.775523407904525e-05, "loss": 2.4605, "step": 575 }, { "epoch": 0.2482345388401455, "grad_norm": 11.514714241027832, "learning_rate": 9.770361528241452e-05, "loss": 2.4807, "step": 580 }, { "epoch": 0.25037449176118126, "grad_norm": 11.684077262878418, "learning_rate": 9.765142366446178e-05, "loss": 2.4928, "step": 585 }, { "epoch": 0.252514444682217, "grad_norm": 9.043937683105469, "learning_rate": 9.759865985189995e-05, "loss": 2.3884, "step": 590 }, { "epoch": 0.2546543976032527, "grad_norm": 8.806846618652344, "learning_rate": 9.754532447831285e-05, "loss": 2.4352, "step": 595 }, { "epoch": 0.25679435052428845, "grad_norm": 11.088272094726562, "learning_rate": 9.749141818414749e-05, "loss": 2.4701, "step": 600 }, { "epoch": 0.2589343034453242, "grad_norm": 10.800878524780273, "learning_rate": 9.743694161670646e-05, "loss": 2.4473, "step": 605 }, { "epoch": 0.26107425636635995, "grad_norm": 8.226394653320312, "learning_rate": 9.73818954301402e-05, "loss": 2.3953, "step": 610 }, { "epoch": 0.2632142092873957, "grad_norm": 7.874807834625244, "learning_rate": 9.732628028543906e-05, "loss": 2.4438, "step": 615 }, { "epoch": 0.2653541622084314, "grad_norm": 9.975316047668457, "learning_rate": 9.727009685042538e-05, "loss": 2.407, "step": 620 }, { "epoch": 0.26749411512946714, "grad_norm": 7.762071132659912, "learning_rate": 9.72133457997455e-05, "loss": 2.4803, "step": 625 }, { "epoch": 0.2696340680505029, "grad_norm": 8.840380668640137, "learning_rate": 9.715602781486166e-05, "loss": 2.393, "step": 630 }, { "epoch": 0.27177402097153863, "grad_norm": 9.467190742492676, "learning_rate": 9.709814358404378e-05, "loss": 2.3655, "step": 635 }, { "epoch": 0.2739139738925744, "grad_norm": 7.103944778442383, "learning_rate": 9.703969380236123e-05, "loss": 2.3414, "step": 640 }, { "epoch": 0.2760539268136101, "grad_norm": 7.247581958770752, "learning_rate": 9.698067917167446e-05, "loss": 2.4331, "step": 645 }, { "epoch": 0.2781938797346458, "grad_norm": 8.669685363769531, "learning_rate": 9.692110040062659e-05, "loss": 2.4395, "step": 650 }, { "epoch": 0.28033383265568157, "grad_norm": 8.114962577819824, "learning_rate": 9.68609582046349e-05, "loss": 2.3229, "step": 655 }, { "epoch": 0.2824737855767173, "grad_norm": 7.252189636230469, "learning_rate": 9.680025330588223e-05, "loss": 2.3969, "step": 660 }, { "epoch": 0.28461373849775307, "grad_norm": 9.382356643676758, "learning_rate": 9.67389864333083e-05, "loss": 2.3446, "step": 665 }, { "epoch": 0.2867536914187888, "grad_norm": 7.623274803161621, "learning_rate": 9.667715832260098e-05, "loss": 2.3587, "step": 670 }, { "epoch": 0.2888936443398245, "grad_norm": 7.448448657989502, "learning_rate": 9.661476971618744e-05, "loss": 2.3671, "step": 675 }, { "epoch": 0.29103359726086026, "grad_norm": 8.833525657653809, "learning_rate": 9.655182136322524e-05, "loss": 2.3739, "step": 680 }, { "epoch": 0.293173550181896, "grad_norm": 10.259654998779297, "learning_rate": 9.648831401959333e-05, "loss": 2.3668, "step": 685 }, { "epoch": 0.29531350310293175, "grad_norm": 10.228246688842773, "learning_rate": 9.642424844788298e-05, "loss": 2.3793, "step": 690 }, { "epoch": 0.2974534560239675, "grad_norm": 8.286998748779297, "learning_rate": 9.635962541738862e-05, "loss": 2.3696, "step": 695 }, { "epoch": 0.2995934089450032, "grad_norm": 8.957596778869629, "learning_rate": 9.62944457040986e-05, "loss": 2.3829, "step": 700 }, { "epoch": 0.30173336186603894, "grad_norm": 8.238673210144043, "learning_rate": 9.622871009068588e-05, "loss": 2.3366, "step": 705 }, { "epoch": 0.3038733147870747, "grad_norm": 8.315441131591797, "learning_rate": 9.616241936649862e-05, "loss": 2.3475, "step": 710 }, { "epoch": 0.30601326770811044, "grad_norm": 8.352023124694824, "learning_rate": 9.609557432755068e-05, "loss": 2.2942, "step": 715 }, { "epoch": 0.3081532206291462, "grad_norm": 10.10500717163086, "learning_rate": 9.602817577651217e-05, "loss": 2.3405, "step": 720 }, { "epoch": 0.3102931735501819, "grad_norm": 7.825136184692383, "learning_rate": 9.596022452269962e-05, "loss": 2.2845, "step": 725 }, { "epoch": 0.3124331264712176, "grad_norm": 10.414852142333984, "learning_rate": 9.589172138206648e-05, "loss": 2.3508, "step": 730 }, { "epoch": 0.3145730793922534, "grad_norm": 8.343073844909668, "learning_rate": 9.582266717719314e-05, "loss": 2.3429, "step": 735 }, { "epoch": 0.3167130323132891, "grad_norm": 8.42280101776123, "learning_rate": 9.575306273727713e-05, "loss": 2.2881, "step": 740 }, { "epoch": 0.31885298523432487, "grad_norm": 7.317922592163086, "learning_rate": 9.568290889812322e-05, "loss": 2.3279, "step": 745 }, { "epoch": 0.32099293815536056, "grad_norm": 8.232301712036133, "learning_rate": 9.561220650213326e-05, "loss": 2.2888, "step": 750 }, { "epoch": 0.3231328910763963, "grad_norm": 10.112414360046387, "learning_rate": 9.554095639829615e-05, "loss": 2.3637, "step": 755 }, { "epoch": 0.32527284399743206, "grad_norm": 10.274619102478027, "learning_rate": 9.546915944217764e-05, "loss": 2.373, "step": 760 }, { "epoch": 0.3274127969184678, "grad_norm": 10.280336380004883, "learning_rate": 9.539681649591002e-05, "loss": 2.3361, "step": 765 }, { "epoch": 0.32955274983950356, "grad_norm": 8.281644821166992, "learning_rate": 9.532392842818177e-05, "loss": 2.3094, "step": 770 }, { "epoch": 0.33169270276053925, "grad_norm": 9.085750579833984, "learning_rate": 9.52504961142272e-05, "loss": 2.3396, "step": 775 }, { "epoch": 0.333832655681575, "grad_norm": 8.183152198791504, "learning_rate": 9.517652043581583e-05, "loss": 2.2929, "step": 780 }, { "epoch": 0.33597260860261074, "grad_norm": 10.052477836608887, "learning_rate": 9.510200228124191e-05, "loss": 2.3608, "step": 785 }, { "epoch": 0.3381125615236465, "grad_norm": 8.700220108032227, "learning_rate": 9.502694254531364e-05, "loss": 2.2485, "step": 790 }, { "epoch": 0.34025251444468224, "grad_norm": 7.829571723937988, "learning_rate": 9.495134212934256e-05, "loss": 2.3197, "step": 795 }, { "epoch": 0.34239246736571793, "grad_norm": 7.233550071716309, "learning_rate": 9.48752019411326e-05, "loss": 2.2776, "step": 800 }, { "epoch": 0.3445324202867537, "grad_norm": 8.803876876831055, "learning_rate": 9.479852289496925e-05, "loss": 2.3447, "step": 805 }, { "epoch": 0.34667237320778943, "grad_norm": 7.353137016296387, "learning_rate": 9.472130591160855e-05, "loss": 2.3355, "step": 810 }, { "epoch": 0.3488123261288252, "grad_norm": 8.22417163848877, "learning_rate": 9.464355191826608e-05, "loss": 2.2921, "step": 815 }, { "epoch": 0.3509522790498609, "grad_norm": 9.752290725708008, "learning_rate": 9.456526184860579e-05, "loss": 2.31, "step": 820 }, { "epoch": 0.3530922319708966, "grad_norm": 9.317963600158691, "learning_rate": 9.448643664272876e-05, "loss": 2.2816, "step": 825 }, { "epoch": 0.35523218489193237, "grad_norm": 7.882107734680176, "learning_rate": 9.440707724716196e-05, "loss": 2.2285, "step": 830 }, { "epoch": 0.3573721378129681, "grad_norm": 7.263378620147705, "learning_rate": 9.432718461484688e-05, "loss": 2.2865, "step": 835 }, { "epoch": 0.35951209073400386, "grad_norm": 8.264389038085938, "learning_rate": 9.424675970512808e-05, "loss": 2.2278, "step": 840 }, { "epoch": 0.3616520436550396, "grad_norm": 7.658196926116943, "learning_rate": 9.416580348374163e-05, "loss": 2.2733, "step": 845 }, { "epoch": 0.3637919965760753, "grad_norm": 7.531430721282959, "learning_rate": 9.40843169228036e-05, "loss": 2.2451, "step": 850 }, { "epoch": 0.36593194949711105, "grad_norm": 6.57159423828125, "learning_rate": 9.400230100079829e-05, "loss": 2.2292, "step": 855 }, { "epoch": 0.3680719024181468, "grad_norm": 7.315071105957031, "learning_rate": 9.391975670256657e-05, "loss": 2.269, "step": 860 }, { "epoch": 0.37021185533918255, "grad_norm": 6.916749000549316, "learning_rate": 9.383668501929395e-05, "loss": 2.2789, "step": 865 }, { "epoch": 0.3723518082602183, "grad_norm": 6.972721576690674, "learning_rate": 9.37530869484988e-05, "loss": 2.2388, "step": 870 }, { "epoch": 0.374491761181254, "grad_norm": 7.823705673217773, "learning_rate": 9.36689634940203e-05, "loss": 2.2251, "step": 875 }, { "epoch": 0.37663171410228974, "grad_norm": 10.453600883483887, "learning_rate": 9.358431566600636e-05, "loss": 2.2375, "step": 880 }, { "epoch": 0.3787716670233255, "grad_norm": 10.327241897583008, "learning_rate": 9.349914448090156e-05, "loss": 2.2713, "step": 885 }, { "epoch": 0.38091161994436123, "grad_norm": 9.396978378295898, "learning_rate": 9.34134509614349e-05, "loss": 2.2676, "step": 890 }, { "epoch": 0.383051572865397, "grad_norm": 7.874500751495361, "learning_rate": 9.332723613660754e-05, "loss": 2.2645, "step": 895 }, { "epoch": 0.3851915257864327, "grad_norm": 9.300860404968262, "learning_rate": 9.32405010416804e-05, "loss": 2.2894, "step": 900 }, { "epoch": 0.3873314787074684, "grad_norm": 7.320137977600098, "learning_rate": 9.315324671816183e-05, "loss": 2.2072, "step": 905 }, { "epoch": 0.38947143162850417, "grad_norm": 7.579836368560791, "learning_rate": 9.306547421379497e-05, "loss": 2.2244, "step": 910 }, { "epoch": 0.3916113845495399, "grad_norm": 8.637717247009277, "learning_rate": 9.297718458254528e-05, "loss": 2.1833, "step": 915 }, { "epoch": 0.39375133747057567, "grad_norm": 6.750846862792969, "learning_rate": 9.288837888458782e-05, "loss": 2.2157, "step": 920 }, { "epoch": 0.39589129039161136, "grad_norm": 8.550662994384766, "learning_rate": 9.27990581862945e-05, "loss": 2.224, "step": 925 }, { "epoch": 0.3980312433126471, "grad_norm": 7.079535961151123, "learning_rate": 9.270922356022142e-05, "loss": 2.1862, "step": 930 }, { "epoch": 0.40017119623368286, "grad_norm": 6.939311981201172, "learning_rate": 9.261887608509579e-05, "loss": 2.2017, "step": 935 }, { "epoch": 0.4023111491547186, "grad_norm": 8.532953262329102, "learning_rate": 9.252801684580308e-05, "loss": 2.1758, "step": 940 }, { "epoch": 0.40445110207575435, "grad_norm": 7.4575018882751465, "learning_rate": 9.243664693337404e-05, "loss": 2.1896, "step": 945 }, { "epoch": 0.40659105499679005, "grad_norm": 7.295337677001953, "learning_rate": 9.234476744497149e-05, "loss": 2.252, "step": 950 }, { "epoch": 0.4087310079178258, "grad_norm": 6.507411479949951, "learning_rate": 9.225237948387722e-05, "loss": 2.2629, "step": 955 }, { "epoch": 0.41087096083886154, "grad_norm": 8.774213790893555, "learning_rate": 9.215948415947875e-05, "loss": 2.1715, "step": 960 }, { "epoch": 0.4130109137598973, "grad_norm": 9.428667068481445, "learning_rate": 9.20660825872559e-05, "loss": 2.1916, "step": 965 }, { "epoch": 0.41515086668093304, "grad_norm": 7.89529275894165, "learning_rate": 9.197217588876756e-05, "loss": 2.2247, "step": 970 }, { "epoch": 0.41729081960196873, "grad_norm": 7.525676727294922, "learning_rate": 9.187776519163811e-05, "loss": 2.1985, "step": 975 }, { "epoch": 0.4194307725230045, "grad_norm": 7.482941627502441, "learning_rate": 9.178285162954386e-05, "loss": 2.1853, "step": 980 }, { "epoch": 0.42157072544404023, "grad_norm": 6.92722225189209, "learning_rate": 9.168743634219955e-05, "loss": 2.1629, "step": 985 }, { "epoch": 0.423710678365076, "grad_norm": 7.211160659790039, "learning_rate": 9.159152047534454e-05, "loss": 2.1817, "step": 990 }, { "epoch": 0.4258506312861117, "grad_norm": 7.496613025665283, "learning_rate": 9.149510518072916e-05, "loss": 2.1736, "step": 995 }, { "epoch": 0.4279905842071474, "grad_norm": 7.383609294891357, "learning_rate": 9.139819161610082e-05, "loss": 2.2043, "step": 1000 }, { "epoch": 0.43013053712818317, "grad_norm": 7.53545618057251, "learning_rate": 9.130078094519008e-05, "loss": 2.232, "step": 1005 }, { "epoch": 0.4322704900492189, "grad_norm": 6.151658535003662, "learning_rate": 9.120287433769674e-05, "loss": 2.1939, "step": 1010 }, { "epoch": 0.43441044297025466, "grad_norm": 7.977633953094482, "learning_rate": 9.11044729692758e-05, "loss": 2.2096, "step": 1015 }, { "epoch": 0.4365503958912904, "grad_norm": 8.515596389770508, "learning_rate": 9.100557802152328e-05, "loss": 2.2042, "step": 1020 }, { "epoch": 0.4386903488123261, "grad_norm": 6.877818584442139, "learning_rate": 9.090619068196203e-05, "loss": 2.1006, "step": 1025 }, { "epoch": 0.44083030173336185, "grad_norm": 7.46975040435791, "learning_rate": 9.080631214402754e-05, "loss": 2.0974, "step": 1030 }, { "epoch": 0.4429702546543976, "grad_norm": 7.358278751373291, "learning_rate": 9.070594360705358e-05, "loss": 2.1188, "step": 1035 }, { "epoch": 0.44511020757543335, "grad_norm": 7.501420974731445, "learning_rate": 9.060508627625779e-05, "loss": 2.134, "step": 1040 }, { "epoch": 0.4472501604964691, "grad_norm": 7.990818500518799, "learning_rate": 9.050374136272717e-05, "loss": 2.1787, "step": 1045 }, { "epoch": 0.4493901134175048, "grad_norm": 6.6185760498046875, "learning_rate": 9.04019100834036e-05, "loss": 2.14, "step": 1050 }, { "epoch": 0.45153006633854054, "grad_norm": 9.137276649475098, "learning_rate": 9.029959366106923e-05, "loss": 2.1632, "step": 1055 }, { "epoch": 0.4536700192595763, "grad_norm": 8.287654876708984, "learning_rate": 9.019679332433173e-05, "loss": 2.1734, "step": 1060 }, { "epoch": 0.45580997218061203, "grad_norm": 8.167359352111816, "learning_rate": 9.009351030760958e-05, "loss": 2.116, "step": 1065 }, { "epoch": 0.4579499251016478, "grad_norm": 7.944402694702148, "learning_rate": 8.998974585111729e-05, "loss": 2.1126, "step": 1070 }, { "epoch": 0.4600898780226835, "grad_norm": 8.373970985412598, "learning_rate": 8.988550120085038e-05, "loss": 2.1179, "step": 1075 }, { "epoch": 0.4622298309437192, "grad_norm": 7.862525463104248, "learning_rate": 8.978077760857058e-05, "loss": 2.1599, "step": 1080 }, { "epoch": 0.46436978386475497, "grad_norm": 6.35279655456543, "learning_rate": 8.967557633179067e-05, "loss": 2.1058, "step": 1085 }, { "epoch": 0.4665097367857907, "grad_norm": 6.809432506561279, "learning_rate": 8.956989863375944e-05, "loss": 2.1049, "step": 1090 }, { "epoch": 0.46864968970682647, "grad_norm": 10.011709213256836, "learning_rate": 8.946374578344653e-05, "loss": 2.1654, "step": 1095 }, { "epoch": 0.47078964262786216, "grad_norm": 8.779292106628418, "learning_rate": 8.935711905552713e-05, "loss": 2.089, "step": 1100 }, { "epoch": 0.4729295955488979, "grad_norm": 8.868700981140137, "learning_rate": 8.925001973036677e-05, "loss": 2.16, "step": 1105 }, { "epoch": 0.47506954846993366, "grad_norm": 8.026627540588379, "learning_rate": 8.914244909400585e-05, "loss": 2.1683, "step": 1110 }, { "epoch": 0.4772095013909694, "grad_norm": 6.766802787780762, "learning_rate": 8.903440843814423e-05, "loss": 2.1475, "step": 1115 }, { "epoch": 0.47934945431200515, "grad_norm": 6.383385181427002, "learning_rate": 8.892589906012577e-05, "loss": 2.151, "step": 1120 }, { "epoch": 0.4814894072330409, "grad_norm": 7.344244956970215, "learning_rate": 8.881692226292269e-05, "loss": 2.1037, "step": 1125 }, { "epoch": 0.4836293601540766, "grad_norm": 8.650083541870117, "learning_rate": 8.870747935511992e-05, "loss": 2.0868, "step": 1130 }, { "epoch": 0.48576931307511234, "grad_norm": 7.451263427734375, "learning_rate": 8.859757165089943e-05, "loss": 2.1287, "step": 1135 }, { "epoch": 0.4879092659961481, "grad_norm": 7.021806716918945, "learning_rate": 8.848720047002446e-05, "loss": 2.1256, "step": 1140 }, { "epoch": 0.49004921891718384, "grad_norm": 7.580230712890625, "learning_rate": 8.837636713782358e-05, "loss": 2.058, "step": 1145 }, { "epoch": 0.4921891718382196, "grad_norm": 6.413537979125977, "learning_rate": 8.826507298517489e-05, "loss": 2.0742, "step": 1150 }, { "epoch": 0.4943291247592553, "grad_norm": 7.502536296844482, "learning_rate": 8.815331934848996e-05, "loss": 2.1497, "step": 1155 }, { "epoch": 0.496469077680291, "grad_norm": 6.689949035644531, "learning_rate": 8.804110756969781e-05, "loss": 2.112, "step": 1160 }, { "epoch": 0.4986090306013268, "grad_norm": 8.030281066894531, "learning_rate": 8.792843899622879e-05, "loss": 2.1844, "step": 1165 }, { "epoch": 0.5007489835223625, "grad_norm": 6.825489044189453, "learning_rate": 8.781531498099844e-05, "loss": 2.1259, "step": 1170 }, { "epoch": 0.5028889364433983, "grad_norm": 7.533500671386719, "learning_rate": 8.770173688239116e-05, "loss": 2.0923, "step": 1175 }, { "epoch": 0.505028889364434, "grad_norm": 6.412694454193115, "learning_rate": 8.758770606424398e-05, "loss": 2.1432, "step": 1180 }, { "epoch": 0.5071688422854698, "grad_norm": 6.011632442474365, "learning_rate": 8.747322389583013e-05, "loss": 2.0918, "step": 1185 }, { "epoch": 0.5093087952065054, "grad_norm": 7.687342166900635, "learning_rate": 8.735829175184267e-05, "loss": 2.1275, "step": 1190 }, { "epoch": 0.5114487481275412, "grad_norm": 7.902970314025879, "learning_rate": 8.724291101237784e-05, "loss": 2.0638, "step": 1195 }, { "epoch": 0.5135887010485769, "grad_norm": 6.375851631164551, "learning_rate": 8.71270830629187e-05, "loss": 2.1268, "step": 1200 }, { "epoch": 0.5157286539696126, "grad_norm": 11.488832473754883, "learning_rate": 8.701080929431824e-05, "loss": 2.1133, "step": 1205 }, { "epoch": 0.5178686068906484, "grad_norm": 7.46975564956665, "learning_rate": 8.689409110278292e-05, "loss": 2.0586, "step": 1210 }, { "epoch": 0.5200085598116841, "grad_norm": 10.648994445800781, "learning_rate": 8.677692988985575e-05, "loss": 2.0896, "step": 1215 }, { "epoch": 0.5221485127327199, "grad_norm": 8.54841136932373, "learning_rate": 8.665932706239949e-05, "loss": 2.0725, "step": 1220 }, { "epoch": 0.5242884656537556, "grad_norm": 6.593012809753418, "learning_rate": 8.654128403257982e-05, "loss": 2.0339, "step": 1225 }, { "epoch": 0.5264284185747914, "grad_norm": 8.725300788879395, "learning_rate": 8.642280221784828e-05, "loss": 2.0586, "step": 1230 }, { "epoch": 0.5285683714958271, "grad_norm": 8.611812591552734, "learning_rate": 8.630388304092536e-05, "loss": 2.049, "step": 1235 }, { "epoch": 0.5307083244168628, "grad_norm": 8.650274276733398, "learning_rate": 8.618452792978336e-05, "loss": 2.0758, "step": 1240 }, { "epoch": 0.5328482773378985, "grad_norm": 5.806057453155518, "learning_rate": 8.606473831762916e-05, "loss": 2.075, "step": 1245 }, { "epoch": 0.5349882302589343, "grad_norm": 6.725940227508545, "learning_rate": 8.59445156428872e-05, "loss": 2.1084, "step": 1250 }, { "epoch": 0.53712818317997, "grad_norm": 7.8081889152526855, "learning_rate": 8.582386134918204e-05, "loss": 2.076, "step": 1255 }, { "epoch": 0.5392681361010058, "grad_norm": 7.350069046020508, "learning_rate": 8.570277688532112e-05, "loss": 2.0705, "step": 1260 }, { "epoch": 0.5414080890220415, "grad_norm": 6.163577079772949, "learning_rate": 8.55812637052773e-05, "loss": 2.0214, "step": 1265 }, { "epoch": 0.5435480419430773, "grad_norm": 8.294384956359863, "learning_rate": 8.545932326817145e-05, "loss": 2.0657, "step": 1270 }, { "epoch": 0.545687994864113, "grad_norm": 7.981808185577393, "learning_rate": 8.533695703825493e-05, "loss": 2.0762, "step": 1275 }, { "epoch": 0.5478279477851488, "grad_norm": 7.62240743637085, "learning_rate": 8.521416648489193e-05, "loss": 2.0571, "step": 1280 }, { "epoch": 0.5499679007061845, "grad_norm": 7.337045192718506, "learning_rate": 8.509095308254191e-05, "loss": 2.0734, "step": 1285 }, { "epoch": 0.5521078536272201, "grad_norm": 6.613148212432861, "learning_rate": 8.496731831074189e-05, "loss": 2.0913, "step": 1290 }, { "epoch": 0.5542478065482559, "grad_norm": 6.551748275756836, "learning_rate": 8.484326365408866e-05, "loss": 2.0183, "step": 1295 }, { "epoch": 0.5563877594692916, "grad_norm": 6.191518783569336, "learning_rate": 8.471879060222094e-05, "loss": 2.0514, "step": 1300 }, { "epoch": 0.5585277123903274, "grad_norm": 6.579594135284424, "learning_rate": 8.459390064980146e-05, "loss": 2.083, "step": 1305 }, { "epoch": 0.5606676653113631, "grad_norm": 8.821534156799316, "learning_rate": 8.446859529649917e-05, "loss": 2.0494, "step": 1310 }, { "epoch": 0.5628076182323989, "grad_norm": 8.188867568969727, "learning_rate": 8.434287604697101e-05, "loss": 2.0842, "step": 1315 }, { "epoch": 0.5649475711534346, "grad_norm": 6.688650608062744, "learning_rate": 8.421674441084404e-05, "loss": 2.059, "step": 1320 }, { "epoch": 0.5670875240744704, "grad_norm": 9.34494400024414, "learning_rate": 8.409020190269716e-05, "loss": 2.0587, "step": 1325 }, { "epoch": 0.5692274769955061, "grad_norm": 9.420138359069824, "learning_rate": 8.396325004204303e-05, "loss": 2.0278, "step": 1330 }, { "epoch": 0.5713674299165419, "grad_norm": 8.35175609588623, "learning_rate": 8.383589035330977e-05, "loss": 1.9988, "step": 1335 }, { "epoch": 0.5735073828375776, "grad_norm": 7.808613300323486, "learning_rate": 8.370812436582267e-05, "loss": 2.0572, "step": 1340 }, { "epoch": 0.5756473357586133, "grad_norm": 6.553249359130859, "learning_rate": 8.357995361378583e-05, "loss": 2.0234, "step": 1345 }, { "epoch": 0.577787288679649, "grad_norm": 6.828055381774902, "learning_rate": 8.345137963626372e-05, "loss": 2.0428, "step": 1350 }, { "epoch": 0.5799272416006848, "grad_norm": 8.460224151611328, "learning_rate": 8.332240397716272e-05, "loss": 2.05, "step": 1355 }, { "epoch": 0.5820671945217205, "grad_norm": 6.507852554321289, "learning_rate": 8.319302818521255e-05, "loss": 2.0947, "step": 1360 }, { "epoch": 0.5842071474427563, "grad_norm": 6.2819085121154785, "learning_rate": 8.306325381394774e-05, "loss": 2.0366, "step": 1365 }, { "epoch": 0.586347100363792, "grad_norm": 7.773951530456543, "learning_rate": 8.293308242168889e-05, "loss": 2.0078, "step": 1370 }, { "epoch": 0.5884870532848278, "grad_norm": 6.938888072967529, "learning_rate": 8.280251557152399e-05, "loss": 2.0739, "step": 1375 }, { "epoch": 0.5906270062058635, "grad_norm": 7.499892234802246, "learning_rate": 8.26715548312897e-05, "loss": 2.011, "step": 1380 }, { "epoch": 0.5927669591268993, "grad_norm": 6.491375923156738, "learning_rate": 8.254020177355243e-05, "loss": 1.9787, "step": 1385 }, { "epoch": 0.594906912047935, "grad_norm": 7.310795307159424, "learning_rate": 8.240845797558958e-05, "loss": 2.0306, "step": 1390 }, { "epoch": 0.5970468649689706, "grad_norm": 6.266601085662842, "learning_rate": 8.227632501937045e-05, "loss": 1.9754, "step": 1395 }, { "epoch": 0.5991868178900064, "grad_norm": 7.139545440673828, "learning_rate": 8.214380449153735e-05, "loss": 2.0119, "step": 1400 }, { "epoch": 0.6013267708110421, "grad_norm": 7.089097499847412, "learning_rate": 8.201089798338655e-05, "loss": 2.0456, "step": 1405 }, { "epoch": 0.6034667237320779, "grad_norm": 6.723423004150391, "learning_rate": 8.187760709084911e-05, "loss": 2.0233, "step": 1410 }, { "epoch": 0.6056066766531136, "grad_norm": 6.559820175170898, "learning_rate": 8.174393341447177e-05, "loss": 1.995, "step": 1415 }, { "epoch": 0.6077466295741494, "grad_norm": 7.716728687286377, "learning_rate": 8.160987855939766e-05, "loss": 2.0308, "step": 1420 }, { "epoch": 0.6098865824951851, "grad_norm": 5.852180480957031, "learning_rate": 8.147544413534714e-05, "loss": 2.0049, "step": 1425 }, { "epoch": 0.6120265354162209, "grad_norm": 9.875506401062012, "learning_rate": 8.134063175659836e-05, "loss": 2.0243, "step": 1430 }, { "epoch": 0.6141664883372566, "grad_norm": 7.915645122528076, "learning_rate": 8.120544304196793e-05, "loss": 2.0137, "step": 1435 }, { "epoch": 0.6163064412582924, "grad_norm": 6.7806077003479, "learning_rate": 8.10698796147915e-05, "loss": 2.0122, "step": 1440 }, { "epoch": 0.618446394179328, "grad_norm": 6.318131446838379, "learning_rate": 8.093394310290421e-05, "loss": 2.0041, "step": 1445 }, { "epoch": 0.6205863471003638, "grad_norm": 6.302967071533203, "learning_rate": 8.079763513862116e-05, "loss": 2.0402, "step": 1450 }, { "epoch": 0.6227263000213995, "grad_norm": 8.546146392822266, "learning_rate": 8.066095735871786e-05, "loss": 2.036, "step": 1455 }, { "epoch": 0.6248662529424353, "grad_norm": 7.018831729888916, "learning_rate": 8.052391140441051e-05, "loss": 1.9625, "step": 1460 }, { "epoch": 0.627006205863471, "grad_norm": 6.960865497589111, "learning_rate": 8.038649892133632e-05, "loss": 2.013, "step": 1465 }, { "epoch": 0.6291461587845067, "grad_norm": 7.643087863922119, "learning_rate": 8.024872155953376e-05, "loss": 2.0234, "step": 1470 }, { "epoch": 0.6312861117055425, "grad_norm": 6.25241231918335, "learning_rate": 8.011058097342275e-05, "loss": 1.9427, "step": 1475 }, { "epoch": 0.6334260646265782, "grad_norm": 7.6290283203125, "learning_rate": 7.997207882178474e-05, "loss": 1.9874, "step": 1480 }, { "epoch": 0.635566017547614, "grad_norm": 10.151625633239746, "learning_rate": 7.983321676774285e-05, "loss": 1.9609, "step": 1485 }, { "epoch": 0.6377059704686497, "grad_norm": 6.939437389373779, "learning_rate": 7.969399647874191e-05, "loss": 2.0095, "step": 1490 }, { "epoch": 0.6398459233896854, "grad_norm": 7.24613618850708, "learning_rate": 7.955441962652835e-05, "loss": 1.9874, "step": 1495 }, { "epoch": 0.6419858763107211, "grad_norm": 8.051011085510254, "learning_rate": 7.941448788713024e-05, "loss": 2.0215, "step": 1500 }, { "epoch": 0.6441258292317569, "grad_norm": 7.338986873626709, "learning_rate": 7.927420294083705e-05, "loss": 1.941, "step": 1505 }, { "epoch": 0.6462657821527926, "grad_norm": 7.442719459533691, "learning_rate": 7.91335664721796e-05, "loss": 2.0215, "step": 1510 }, { "epoch": 0.6484057350738284, "grad_norm": 7.5543341636657715, "learning_rate": 7.899258016990969e-05, "loss": 1.9694, "step": 1515 }, { "epoch": 0.6505456879948641, "grad_norm": 7.134289264678955, "learning_rate": 7.885124572697998e-05, "loss": 2.0004, "step": 1520 }, { "epoch": 0.6526856409158999, "grad_norm": 6.567378520965576, "learning_rate": 7.870956484052346e-05, "loss": 1.9725, "step": 1525 }, { "epoch": 0.6548255938369356, "grad_norm": 7.110774040222168, "learning_rate": 7.856753921183331e-05, "loss": 2.0402, "step": 1530 }, { "epoch": 0.6569655467579714, "grad_norm": 6.285329818725586, "learning_rate": 7.842517054634226e-05, "loss": 1.974, "step": 1535 }, { "epoch": 0.6591054996790071, "grad_norm": 5.5663676261901855, "learning_rate": 7.828246055360226e-05, "loss": 2.0189, "step": 1540 }, { "epoch": 0.6612454526000427, "grad_norm": 6.485752582550049, "learning_rate": 7.813941094726384e-05, "loss": 1.9699, "step": 1545 }, { "epoch": 0.6633854055210785, "grad_norm": 5.899412155151367, "learning_rate": 7.79960234450556e-05, "loss": 1.9966, "step": 1550 }, { "epoch": 0.6655253584421142, "grad_norm": 7.0582051277160645, "learning_rate": 7.78522997687636e-05, "loss": 1.9811, "step": 1555 }, { "epoch": 0.66766531136315, "grad_norm": 6.355223178863525, "learning_rate": 7.770824164421062e-05, "loss": 1.9238, "step": 1560 }, { "epoch": 0.6698052642841857, "grad_norm": 7.558047771453857, "learning_rate": 7.756385080123546e-05, "loss": 1.9915, "step": 1565 }, { "epoch": 0.6719452172052215, "grad_norm": 6.747286319732666, "learning_rate": 7.741912897367221e-05, "loss": 1.9746, "step": 1570 }, { "epoch": 0.6740851701262572, "grad_norm": 6.1656293869018555, "learning_rate": 7.727407789932935e-05, "loss": 1.9615, "step": 1575 }, { "epoch": 0.676225123047293, "grad_norm": 7.076664447784424, "learning_rate": 7.7128699319969e-05, "loss": 2.0052, "step": 1580 }, { "epoch": 0.6783650759683287, "grad_norm": 7.908239841461182, "learning_rate": 7.698299498128587e-05, "loss": 2.0223, "step": 1585 }, { "epoch": 0.6805050288893645, "grad_norm": 6.080158233642578, "learning_rate": 7.68369666328864e-05, "loss": 1.9561, "step": 1590 }, { "epoch": 0.6826449818104001, "grad_norm": 6.817996025085449, "learning_rate": 7.669061602826768e-05, "loss": 1.9811, "step": 1595 }, { "epoch": 0.6847849347314359, "grad_norm": 5.7741475105285645, "learning_rate": 7.654394492479648e-05, "loss": 1.9011, "step": 1600 }, { "epoch": 0.6869248876524716, "grad_norm": 5.87952995300293, "learning_rate": 7.639695508368803e-05, "loss": 1.9695, "step": 1605 }, { "epoch": 0.6890648405735074, "grad_norm": 6.916929721832275, "learning_rate": 7.6249648269985e-05, "loss": 1.9536, "step": 1610 }, { "epoch": 0.6912047934945431, "grad_norm": 8.327000617980957, "learning_rate": 7.61020262525362e-05, "loss": 1.961, "step": 1615 }, { "epoch": 0.6933447464155789, "grad_norm": 7.954501628875732, "learning_rate": 7.59540908039754e-05, "loss": 2.0029, "step": 1620 }, { "epoch": 0.6954846993366146, "grad_norm": 6.015987873077393, "learning_rate": 7.580584370070001e-05, "loss": 1.9753, "step": 1625 }, { "epoch": 0.6976246522576504, "grad_norm": 7.0654520988464355, "learning_rate": 7.565728672284979e-05, "loss": 1.9378, "step": 1630 }, { "epoch": 0.6997646051786861, "grad_norm": 7.740458965301514, "learning_rate": 7.550842165428543e-05, "loss": 1.9464, "step": 1635 }, { "epoch": 0.7019045580997219, "grad_norm": 7.246245861053467, "learning_rate": 7.535925028256717e-05, "loss": 2.0064, "step": 1640 }, { "epoch": 0.7040445110207575, "grad_norm": 6.526742458343506, "learning_rate": 7.520977439893329e-05, "loss": 1.9545, "step": 1645 }, { "epoch": 0.7061844639417932, "grad_norm": 7.05415153503418, "learning_rate": 7.505999579827863e-05, "loss": 1.979, "step": 1650 }, { "epoch": 0.708324416862829, "grad_norm": 6.696120738983154, "learning_rate": 7.490991627913306e-05, "loss": 1.9026, "step": 1655 }, { "epoch": 0.7104643697838647, "grad_norm": 7.282532215118408, "learning_rate": 7.475953764363983e-05, "loss": 1.9877, "step": 1660 }, { "epoch": 0.7126043227049005, "grad_norm": 5.767569541931152, "learning_rate": 7.460886169753397e-05, "loss": 2.005, "step": 1665 }, { "epoch": 0.7147442756259362, "grad_norm": 6.214768409729004, "learning_rate": 7.445789025012055e-05, "loss": 1.949, "step": 1670 }, { "epoch": 0.716884228546972, "grad_norm": 6.871894836425781, "learning_rate": 7.430662511425308e-05, "loss": 1.9725, "step": 1675 }, { "epoch": 0.7190241814680077, "grad_norm": 7.443453311920166, "learning_rate": 7.415506810631155e-05, "loss": 1.9316, "step": 1680 }, { "epoch": 0.7211641343890435, "grad_norm": 6.8776140213012695, "learning_rate": 7.400322104618085e-05, "loss": 1.9556, "step": 1685 }, { "epoch": 0.7233040873100792, "grad_norm": 6.675306797027588, "learning_rate": 7.385108575722868e-05, "loss": 1.9833, "step": 1690 }, { "epoch": 0.7254440402311149, "grad_norm": 6.732053756713867, "learning_rate": 7.369866406628385e-05, "loss": 1.9036, "step": 1695 }, { "epoch": 0.7275839931521506, "grad_norm": 7.516349792480469, "learning_rate": 7.354595780361423e-05, "loss": 1.9876, "step": 1700 }, { "epoch": 0.7297239460731864, "grad_norm": 6.245758056640625, "learning_rate": 7.339296880290481e-05, "loss": 1.9403, "step": 1705 }, { "epoch": 0.7318638989942221, "grad_norm": 6.886528015136719, "learning_rate": 7.323969890123565e-05, "loss": 1.9668, "step": 1710 }, { "epoch": 0.7340038519152579, "grad_norm": 7.249610900878906, "learning_rate": 7.308614993905992e-05, "loss": 2.0097, "step": 1715 }, { "epoch": 0.7361438048362936, "grad_norm": 6.901829719543457, "learning_rate": 7.293232376018164e-05, "loss": 1.9935, "step": 1720 }, { "epoch": 0.7382837577573294, "grad_norm": 7.697319984436035, "learning_rate": 7.277822221173367e-05, "loss": 1.9295, "step": 1725 }, { "epoch": 0.7404237106783651, "grad_norm": 6.5634565353393555, "learning_rate": 7.262384714415551e-05, "loss": 1.9149, "step": 1730 }, { "epoch": 0.7425636635994008, "grad_norm": 6.889857769012451, "learning_rate": 7.2469200411171e-05, "loss": 1.947, "step": 1735 }, { "epoch": 0.7447036165204366, "grad_norm": 7.622896194458008, "learning_rate": 7.231428386976618e-05, "loss": 1.8894, "step": 1740 }, { "epoch": 0.7468435694414722, "grad_norm": 7.235098838806152, "learning_rate": 7.215909938016684e-05, "loss": 1.9483, "step": 1745 }, { "epoch": 0.748983522362508, "grad_norm": 5.822272300720215, "learning_rate": 7.200364880581637e-05, "loss": 1.9301, "step": 1750 }, { "epoch": 0.7511234752835437, "grad_norm": 6.663504123687744, "learning_rate": 7.184793401335322e-05, "loss": 1.9482, "step": 1755 }, { "epoch": 0.7532634282045795, "grad_norm": 7.109114170074463, "learning_rate": 7.169195687258859e-05, "loss": 1.9659, "step": 1760 }, { "epoch": 0.7554033811256152, "grad_norm": 7.012261867523193, "learning_rate": 7.15357192564839e-05, "loss": 1.9284, "step": 1765 }, { "epoch": 0.757543334046651, "grad_norm": 7.140257835388184, "learning_rate": 7.137922304112838e-05, "loss": 1.9185, "step": 1770 }, { "epoch": 0.7596832869676867, "grad_norm": 5.866714954376221, "learning_rate": 7.122247010571647e-05, "loss": 1.9535, "step": 1775 }, { "epoch": 0.7618232398887225, "grad_norm": 6.297852516174316, "learning_rate": 7.106546233252528e-05, "loss": 1.9455, "step": 1780 }, { "epoch": 0.7639631928097582, "grad_norm": 5.564944267272949, "learning_rate": 7.090820160689201e-05, "loss": 1.9153, "step": 1785 }, { "epoch": 0.766103145730794, "grad_norm": 6.030523777008057, "learning_rate": 7.07506898171913e-05, "loss": 1.9386, "step": 1790 }, { "epoch": 0.7682430986518297, "grad_norm": 7.070890426635742, "learning_rate": 7.059292885481253e-05, "loss": 1.9576, "step": 1795 }, { "epoch": 0.7703830515728654, "grad_norm": 7.497870922088623, "learning_rate": 7.04349206141371e-05, "loss": 1.8669, "step": 1800 }, { "epoch": 0.7725230044939011, "grad_norm": 5.917843818664551, "learning_rate": 7.02766669925158e-05, "loss": 1.9455, "step": 1805 }, { "epoch": 0.7746629574149368, "grad_norm": 6.862101078033447, "learning_rate": 7.011816989024583e-05, "loss": 1.9464, "step": 1810 }, { "epoch": 0.7768029103359726, "grad_norm": 6.385555267333984, "learning_rate": 6.995943121054816e-05, "loss": 1.9477, "step": 1815 }, { "epoch": 0.7789428632570083, "grad_norm": 6.692836761474609, "learning_rate": 6.980045285954456e-05, "loss": 1.916, "step": 1820 }, { "epoch": 0.7810828161780441, "grad_norm": 7.266310214996338, "learning_rate": 6.964123674623475e-05, "loss": 1.929, "step": 1825 }, { "epoch": 0.7832227690990798, "grad_norm": 6.483401775360107, "learning_rate": 6.948178478247355e-05, "loss": 1.9205, "step": 1830 }, { "epoch": 0.7853627220201156, "grad_norm": 7.336221694946289, "learning_rate": 6.932209888294777e-05, "loss": 1.9402, "step": 1835 }, { "epoch": 0.7875026749411513, "grad_norm": 7.22684383392334, "learning_rate": 6.916218096515336e-05, "loss": 1.8924, "step": 1840 }, { "epoch": 0.7896426278621871, "grad_norm": 6.436882972717285, "learning_rate": 6.900203294937229e-05, "loss": 1.9378, "step": 1845 }, { "epoch": 0.7917825807832227, "grad_norm": 6.200204849243164, "learning_rate": 6.884165675864954e-05, "loss": 1.864, "step": 1850 }, { "epoch": 0.7939225337042585, "grad_norm": 6.617162227630615, "learning_rate": 6.868105431877002e-05, "loss": 1.9279, "step": 1855 }, { "epoch": 0.7960624866252942, "grad_norm": 5.673040866851807, "learning_rate": 6.85202275582354e-05, "loss": 1.8682, "step": 1860 }, { "epoch": 0.79820243954633, "grad_norm": 6.891549110412598, "learning_rate": 6.835917840824097e-05, "loss": 1.8942, "step": 1865 }, { "epoch": 0.8003423924673657, "grad_norm": 6.0316386222839355, "learning_rate": 6.819790880265246e-05, "loss": 1.906, "step": 1870 }, { "epoch": 0.8024823453884015, "grad_norm": 7.065370559692383, "learning_rate": 6.803642067798284e-05, "loss": 1.9118, "step": 1875 }, { "epoch": 0.8046222983094372, "grad_norm": 6.079885959625244, "learning_rate": 6.7874715973369e-05, "loss": 1.934, "step": 1880 }, { "epoch": 0.806762251230473, "grad_norm": 6.481017112731934, "learning_rate": 6.771279663054853e-05, "loss": 1.9257, "step": 1885 }, { "epoch": 0.8089022041515087, "grad_norm": 6.038636207580566, "learning_rate": 6.755066459383637e-05, "loss": 1.8815, "step": 1890 }, { "epoch": 0.8110421570725445, "grad_norm": 6.196339130401611, "learning_rate": 6.738832181010149e-05, "loss": 1.9115, "step": 1895 }, { "epoch": 0.8131821099935801, "grad_norm": 6.638113498687744, "learning_rate": 6.722577022874345e-05, "loss": 1.9637, "step": 1900 }, { "epoch": 0.8153220629146158, "grad_norm": 6.1240973472595215, "learning_rate": 6.706301180166909e-05, "loss": 1.9453, "step": 1905 }, { "epoch": 0.8174620158356516, "grad_norm": 6.026691436767578, "learning_rate": 6.690004848326898e-05, "loss": 1.8705, "step": 1910 }, { "epoch": 0.8196019687566873, "grad_norm": 6.312042713165283, "learning_rate": 6.673688223039406e-05, "loss": 1.8869, "step": 1915 }, { "epoch": 0.8217419216777231, "grad_norm": 6.738498687744141, "learning_rate": 6.65735150023321e-05, "loss": 1.9406, "step": 1920 }, { "epoch": 0.8238818745987588, "grad_norm": 5.581632614135742, "learning_rate": 6.64099487607841e-05, "loss": 1.8928, "step": 1925 }, { "epoch": 0.8260218275197946, "grad_norm": 6.734776973724365, "learning_rate": 6.624618546984082e-05, "loss": 1.8833, "step": 1930 }, { "epoch": 0.8281617804408303, "grad_norm": 6.365813732147217, "learning_rate": 6.608222709595925e-05, "loss": 1.8952, "step": 1935 }, { "epoch": 0.8303017333618661, "grad_norm": 6.569833278656006, "learning_rate": 6.59180756079388e-05, "loss": 1.9004, "step": 1940 }, { "epoch": 0.8324416862829018, "grad_norm": 5.505297660827637, "learning_rate": 6.575373297689786e-05, "loss": 1.8726, "step": 1945 }, { "epoch": 0.8345816392039375, "grad_norm": 6.422504901885986, "learning_rate": 6.558920117625005e-05, "loss": 1.9392, "step": 1950 }, { "epoch": 0.8367215921249732, "grad_norm": 5.929808139801025, "learning_rate": 6.542448218168049e-05, "loss": 1.8829, "step": 1955 }, { "epoch": 0.838861545046009, "grad_norm": 5.6545796394348145, "learning_rate": 6.525957797112211e-05, "loss": 1.8625, "step": 1960 }, { "epoch": 0.8410014979670447, "grad_norm": 5.610157489776611, "learning_rate": 6.509449052473193e-05, "loss": 1.9184, "step": 1965 }, { "epoch": 0.8431414508880805, "grad_norm": 6.630683898925781, "learning_rate": 6.492922182486722e-05, "loss": 1.9024, "step": 1970 }, { "epoch": 0.8452814038091162, "grad_norm": 5.480852127075195, "learning_rate": 6.476377385606175e-05, "loss": 1.8924, "step": 1975 }, { "epoch": 0.847421356730152, "grad_norm": 6.156518459320068, "learning_rate": 6.459814860500194e-05, "loss": 1.8499, "step": 1980 }, { "epoch": 0.8495613096511877, "grad_norm": 6.882572174072266, "learning_rate": 6.443234806050298e-05, "loss": 1.8714, "step": 1985 }, { "epoch": 0.8517012625722234, "grad_norm": 6.394567489624023, "learning_rate": 6.4266374213485e-05, "loss": 1.889, "step": 1990 }, { "epoch": 0.8538412154932592, "grad_norm": 5.844759941101074, "learning_rate": 6.41002290569491e-05, "loss": 1.919, "step": 1995 }, { "epoch": 0.8559811684142948, "grad_norm": 5.819794178009033, "learning_rate": 6.393391458595345e-05, "loss": 1.8827, "step": 2000 }, { "epoch": 0.8581211213353306, "grad_norm": 7.27986478805542, "learning_rate": 6.37674327975894e-05, "loss": 1.8559, "step": 2005 }, { "epoch": 0.8602610742563663, "grad_norm": 6.1043195724487305, "learning_rate": 6.360078569095734e-05, "loss": 1.8751, "step": 2010 }, { "epoch": 0.8624010271774021, "grad_norm": 6.8420090675354, "learning_rate": 6.343397526714284e-05, "loss": 1.8557, "step": 2015 }, { "epoch": 0.8645409800984378, "grad_norm": 6.02893590927124, "learning_rate": 6.326700352919259e-05, "loss": 1.9407, "step": 2020 }, { "epoch": 0.8666809330194736, "grad_norm": 6.999791145324707, "learning_rate": 6.309987248209029e-05, "loss": 1.8652, "step": 2025 }, { "epoch": 0.8688208859405093, "grad_norm": 6.686702728271484, "learning_rate": 6.293258413273262e-05, "loss": 1.9068, "step": 2030 }, { "epoch": 0.8709608388615451, "grad_norm": 6.408459186553955, "learning_rate": 6.276514048990515e-05, "loss": 1.8797, "step": 2035 }, { "epoch": 0.8731007917825808, "grad_norm": 6.277412414550781, "learning_rate": 6.259754356425818e-05, "loss": 1.8417, "step": 2040 }, { "epoch": 0.8752407447036166, "grad_norm": 5.6100287437438965, "learning_rate": 6.242979536828262e-05, "loss": 1.87, "step": 2045 }, { "epoch": 0.8773806976246522, "grad_norm": 6.132706165313721, "learning_rate": 6.226189791628583e-05, "loss": 1.9009, "step": 2050 }, { "epoch": 0.879520650545688, "grad_norm": 5.885615825653076, "learning_rate": 6.209385322436746e-05, "loss": 1.8624, "step": 2055 }, { "epoch": 0.8816606034667237, "grad_norm": 6.174984455108643, "learning_rate": 6.192566331039514e-05, "loss": 1.8723, "step": 2060 }, { "epoch": 0.8838005563877595, "grad_norm": 5.804656982421875, "learning_rate": 6.175733019398034e-05, "loss": 1.8639, "step": 2065 }, { "epoch": 0.8859405093087952, "grad_norm": 7.446183204650879, "learning_rate": 6.15888558964541e-05, "loss": 1.8703, "step": 2070 }, { "epoch": 0.888080462229831, "grad_norm": 6.513533115386963, "learning_rate": 6.142024244084278e-05, "loss": 1.8334, "step": 2075 }, { "epoch": 0.8902204151508667, "grad_norm": 6.254002571105957, "learning_rate": 6.125149185184369e-05, "loss": 1.8476, "step": 2080 }, { "epoch": 0.8923603680719024, "grad_norm": 5.820918560028076, "learning_rate": 6.108260615580086e-05, "loss": 1.9445, "step": 2085 }, { "epoch": 0.8945003209929382, "grad_norm": 7.240386962890625, "learning_rate": 6.091358738068064e-05, "loss": 1.8563, "step": 2090 }, { "epoch": 0.8966402739139739, "grad_norm": 6.337396621704102, "learning_rate": 6.074443755604744e-05, "loss": 1.8851, "step": 2095 }, { "epoch": 0.8987802268350096, "grad_norm": 7.422278881072998, "learning_rate": 6.0575158713039234e-05, "loss": 1.8112, "step": 2100 }, { "epoch": 0.9009201797560453, "grad_norm": 6.84944486618042, "learning_rate": 6.040575288434329e-05, "loss": 1.865, "step": 2105 }, { "epoch": 0.9030601326770811, "grad_norm": 6.545510768890381, "learning_rate": 6.02362221041717e-05, "loss": 1.8829, "step": 2110 }, { "epoch": 0.9052000855981168, "grad_norm": 6.097507476806641, "learning_rate": 6.006656840823696e-05, "loss": 1.8363, "step": 2115 }, { "epoch": 0.9073400385191526, "grad_norm": 7.34000301361084, "learning_rate": 5.9896793833727496e-05, "loss": 1.8738, "step": 2120 }, { "epoch": 0.9094799914401883, "grad_norm": 5.860642433166504, "learning_rate": 5.972690041928331e-05, "loss": 1.8792, "step": 2125 }, { "epoch": 0.9116199443612241, "grad_norm": 6.345127582550049, "learning_rate": 5.9556890204971326e-05, "loss": 1.838, "step": 2130 }, { "epoch": 0.9137598972822598, "grad_norm": 6.226019382476807, "learning_rate": 5.93867652322611e-05, "loss": 1.881, "step": 2135 }, { "epoch": 0.9158998502032956, "grad_norm": 6.527356147766113, "learning_rate": 5.921652754400011e-05, "loss": 1.8406, "step": 2140 }, { "epoch": 0.9180398031243313, "grad_norm": 5.739786148071289, "learning_rate": 5.904617918438936e-05, "loss": 1.8496, "step": 2145 }, { "epoch": 0.920179756045367, "grad_norm": 7.399641990661621, "learning_rate": 5.887572219895877e-05, "loss": 1.8212, "step": 2150 }, { "epoch": 0.9223197089664027, "grad_norm": 5.731332302093506, "learning_rate": 5.8705158634542644e-05, "loss": 1.9046, "step": 2155 }, { "epoch": 0.9244596618874384, "grad_norm": 7.024575710296631, "learning_rate": 5.853449053925505e-05, "loss": 1.8799, "step": 2160 }, { "epoch": 0.9265996148084742, "grad_norm": 5.317823886871338, "learning_rate": 5.836371996246528e-05, "loss": 1.8679, "step": 2165 }, { "epoch": 0.9287395677295099, "grad_norm": 5.5442795753479, "learning_rate": 5.819284895477323e-05, "loss": 1.8641, "step": 2170 }, { "epoch": 0.9308795206505457, "grad_norm": 5.924790382385254, "learning_rate": 5.80218795679847e-05, "loss": 1.8456, "step": 2175 }, { "epoch": 0.9330194735715814, "grad_norm": 5.661588668823242, "learning_rate": 5.78508138550869e-05, "loss": 1.848, "step": 2180 }, { "epoch": 0.9351594264926172, "grad_norm": 5.443496227264404, "learning_rate": 5.7679653870223673e-05, "loss": 1.8579, "step": 2185 }, { "epoch": 0.9372993794136529, "grad_norm": 5.579583644866943, "learning_rate": 5.750840166867085e-05, "loss": 1.8477, "step": 2190 }, { "epoch": 0.9394393323346887, "grad_norm": 6.413700580596924, "learning_rate": 5.733705930681165e-05, "loss": 1.8889, "step": 2195 }, { "epoch": 0.9415792852557243, "grad_norm": 6.018016338348389, "learning_rate": 5.7165628842111866e-05, "loss": 1.8509, "step": 2200 }, { "epoch": 0.9437192381767601, "grad_norm": 6.163970470428467, "learning_rate": 5.699411233309528e-05, "loss": 1.8492, "step": 2205 }, { "epoch": 0.9458591910977958, "grad_norm": 7.421061992645264, "learning_rate": 5.682251183931886e-05, "loss": 1.8493, "step": 2210 }, { "epoch": 0.9479991440188316, "grad_norm": 6.061490058898926, "learning_rate": 5.6650829421348065e-05, "loss": 1.8128, "step": 2215 }, { "epoch": 0.9501390969398673, "grad_norm": 6.558000087738037, "learning_rate": 5.647906714073208e-05, "loss": 1.8476, "step": 2220 }, { "epoch": 0.9522790498609031, "grad_norm": 7.766129970550537, "learning_rate": 5.630722705997908e-05, "loss": 1.8357, "step": 2225 }, { "epoch": 0.9544190027819388, "grad_norm": 6.785080909729004, "learning_rate": 5.6135311242531473e-05, "loss": 1.845, "step": 2230 }, { "epoch": 0.9565589557029746, "grad_norm": 5.932000637054443, "learning_rate": 5.59633217527411e-05, "loss": 1.8277, "step": 2235 }, { "epoch": 0.9586989086240103, "grad_norm": 6.414392948150635, "learning_rate": 5.5791260655844424e-05, "loss": 1.9083, "step": 2240 }, { "epoch": 0.960838861545046, "grad_norm": 5.878129482269287, "learning_rate": 5.5619130017937806e-05, "loss": 1.8427, "step": 2245 }, { "epoch": 0.9629788144660818, "grad_norm": 5.418539524078369, "learning_rate": 5.5446931905952624e-05, "loss": 1.8509, "step": 2250 }, { "epoch": 0.9651187673871174, "grad_norm": 5.333776473999023, "learning_rate": 5.527466838763049e-05, "loss": 1.8308, "step": 2255 }, { "epoch": 0.9672587203081532, "grad_norm": 5.526288032531738, "learning_rate": 5.510234153149839e-05, "loss": 1.8443, "step": 2260 }, { "epoch": 0.9693986732291889, "grad_norm": 5.585949897766113, "learning_rate": 5.4929953406843906e-05, "loss": 1.8008, "step": 2265 }, { "epoch": 0.9715386261502247, "grad_norm": 6.400721549987793, "learning_rate": 5.475750608369029e-05, "loss": 1.8238, "step": 2270 }, { "epoch": 0.9736785790712604, "grad_norm": 6.028564453125, "learning_rate": 5.4585001632771656e-05, "loss": 1.8195, "step": 2275 }, { "epoch": 0.9758185319922962, "grad_norm": 6.440245628356934, "learning_rate": 5.4412442125508113e-05, "loss": 1.8225, "step": 2280 }, { "epoch": 0.9779584849133319, "grad_norm": 6.663422584533691, "learning_rate": 5.4239829633980866e-05, "loss": 1.8487, "step": 2285 }, { "epoch": 0.9800984378343677, "grad_norm": 6.021454811096191, "learning_rate": 5.4067166230907365e-05, "loss": 1.8593, "step": 2290 }, { "epoch": 0.9822383907554034, "grad_norm": 6.139098644256592, "learning_rate": 5.389445398961639e-05, "loss": 1.8337, "step": 2295 }, { "epoch": 0.9843783436764392, "grad_norm": 5.165621280670166, "learning_rate": 5.3721694984023194e-05, "loss": 1.899, "step": 2300 }, { "epoch": 0.9865182965974748, "grad_norm": 6.382976055145264, "learning_rate": 5.354889128860454e-05, "loss": 1.8559, "step": 2305 }, { "epoch": 0.9886582495185106, "grad_norm": 6.370285511016846, "learning_rate": 5.337604497837383e-05, "loss": 1.7831, "step": 2310 }, { "epoch": 0.9907982024395463, "grad_norm": 5.373252868652344, "learning_rate": 5.320315812885618e-05, "loss": 1.8073, "step": 2315 }, { "epoch": 0.992938155360582, "grad_norm": 6.694858551025391, "learning_rate": 5.3030232816063505e-05, "loss": 1.7992, "step": 2320 }, { "epoch": 0.9950781082816178, "grad_norm": 6.6283440589904785, "learning_rate": 5.28572711164696e-05, "loss": 1.8422, "step": 2325 }, { "epoch": 0.9972180612026535, "grad_norm": 6.046857833862305, "learning_rate": 5.268427510698517e-05, "loss": 1.8237, "step": 2330 }, { "epoch": 0.9993580141236893, "grad_norm": 6.323689937591553, "learning_rate": 5.251124686493291e-05, "loss": 1.8662, "step": 2335 }, { "epoch": 1.0012839717526214, "grad_norm": 6.170103549957275, "learning_rate": 5.233818846802255e-05, "loss": 1.7427, "step": 2340 }, { "epoch": 1.0034239246736572, "grad_norm": 5.78736686706543, "learning_rate": 5.216510199432596e-05, "loss": 1.83, "step": 2345 }, { "epoch": 1.005563877594693, "grad_norm": 5.895803451538086, "learning_rate": 5.199198952225212e-05, "loss": 1.8002, "step": 2350 }, { "epoch": 1.0077038305157286, "grad_norm": 6.435295581817627, "learning_rate": 5.1818853130522184e-05, "loss": 1.7659, "step": 2355 }, { "epoch": 1.0098437834367644, "grad_norm": 6.010552406311035, "learning_rate": 5.164569489814456e-05, "loss": 1.7948, "step": 2360 }, { "epoch": 1.0119837363578001, "grad_norm": 6.7105255126953125, "learning_rate": 5.147251690438992e-05, "loss": 1.797, "step": 2365 }, { "epoch": 1.014123689278836, "grad_norm": 6.607806205749512, "learning_rate": 5.1299321228766194e-05, "loss": 1.7717, "step": 2370 }, { "epoch": 1.0162636421998716, "grad_norm": 8.161765098571777, "learning_rate": 5.112610995099368e-05, "loss": 1.7855, "step": 2375 }, { "epoch": 1.0184035951209074, "grad_norm": 6.3959150314331055, "learning_rate": 5.095288515097999e-05, "loss": 1.8212, "step": 2380 }, { "epoch": 1.0205435480419431, "grad_norm": 5.896310806274414, "learning_rate": 5.0779648908795116e-05, "loss": 1.7695, "step": 2385 }, { "epoch": 1.0226835009629789, "grad_norm": 5.285399913787842, "learning_rate": 5.060640330464646e-05, "loss": 1.7961, "step": 2390 }, { "epoch": 1.0248234538840146, "grad_norm": 6.583270072937012, "learning_rate": 5.043315041885383e-05, "loss": 1.8068, "step": 2395 }, { "epoch": 1.0269634068050504, "grad_norm": 6.104979991912842, "learning_rate": 5.0259892331824474e-05, "loss": 1.7806, "step": 2400 }, { "epoch": 1.0291033597260861, "grad_norm": 5.723150730133057, "learning_rate": 5.008663112402811e-05, "loss": 1.7961, "step": 2405 }, { "epoch": 1.0312433126471217, "grad_norm": 5.977392673492432, "learning_rate": 4.991336887597192e-05, "loss": 1.7362, "step": 2410 }, { "epoch": 1.0333832655681574, "grad_norm": 6.383233070373535, "learning_rate": 4.974010766817555e-05, "loss": 1.7895, "step": 2415 }, { "epoch": 1.0355232184891932, "grad_norm": 6.362678527832031, "learning_rate": 4.95668495811462e-05, "loss": 1.7526, "step": 2420 }, { "epoch": 1.037663171410229, "grad_norm": 5.289202690124512, "learning_rate": 4.939359669535357e-05, "loss": 1.8182, "step": 2425 }, { "epoch": 1.0398031243312646, "grad_norm": 6.294672012329102, "learning_rate": 4.922035109120491e-05, "loss": 1.7909, "step": 2430 }, { "epoch": 1.0419430772523004, "grad_norm": 6.827216625213623, "learning_rate": 4.904711484902003e-05, "loss": 1.8041, "step": 2435 }, { "epoch": 1.0440830301733361, "grad_norm": 6.308746337890625, "learning_rate": 4.887389004900633e-05, "loss": 1.8017, "step": 2440 }, { "epoch": 1.046222983094372, "grad_norm": 5.453704357147217, "learning_rate": 4.870067877123382e-05, "loss": 1.791, "step": 2445 }, { "epoch": 1.0483629360154076, "grad_norm": 5.51895809173584, "learning_rate": 4.852748309561009e-05, "loss": 1.7959, "step": 2450 }, { "epoch": 1.0505028889364434, "grad_norm": 5.434797286987305, "learning_rate": 4.835430510185545e-05, "loss": 1.8243, "step": 2455 }, { "epoch": 1.0526428418574791, "grad_norm": 5.3560590744018555, "learning_rate": 4.818114686947783e-05, "loss": 1.8104, "step": 2460 }, { "epoch": 1.0547827947785149, "grad_norm": 6.535407543182373, "learning_rate": 4.800801047774789e-05, "loss": 1.7754, "step": 2465 }, { "epoch": 1.0569227476995506, "grad_norm": 5.90179967880249, "learning_rate": 4.783489800567405e-05, "loss": 1.8041, "step": 2470 }, { "epoch": 1.0590627006205864, "grad_norm": 6.035419464111328, "learning_rate": 4.766181153197746e-05, "loss": 1.8154, "step": 2475 }, { "epoch": 1.0612026535416221, "grad_norm": 5.992060661315918, "learning_rate": 4.748875313506711e-05, "loss": 1.8101, "step": 2480 }, { "epoch": 1.0633426064626579, "grad_norm": 6.405609130859375, "learning_rate": 4.7315724893014846e-05, "loss": 1.7881, "step": 2485 }, { "epoch": 1.0654825593836936, "grad_norm": 5.741700649261475, "learning_rate": 4.714272888353041e-05, "loss": 1.7986, "step": 2490 }, { "epoch": 1.0676225123047294, "grad_norm": 5.35413122177124, "learning_rate": 4.69697671839365e-05, "loss": 1.7832, "step": 2495 }, { "epoch": 1.0697624652257651, "grad_norm": 5.857509613037109, "learning_rate": 4.679684187114384e-05, "loss": 1.7965, "step": 2500 }, { "epoch": 1.0719024181468009, "grad_norm": 6.7070698738098145, "learning_rate": 4.6623955021626184e-05, "loss": 1.7924, "step": 2505 }, { "epoch": 1.0740423710678364, "grad_norm": 6.540157318115234, "learning_rate": 4.645110871139547e-05, "loss": 1.7951, "step": 2510 }, { "epoch": 1.0761823239888721, "grad_norm": 6.704575538635254, "learning_rate": 4.627830501597681e-05, "loss": 1.7831, "step": 2515 }, { "epoch": 1.078322276909908, "grad_norm": 6.622875213623047, "learning_rate": 4.610554601038361e-05, "loss": 1.7869, "step": 2520 }, { "epoch": 1.0804622298309436, "grad_norm": 5.487006187438965, "learning_rate": 4.593283376909264e-05, "loss": 1.7893, "step": 2525 }, { "epoch": 1.0826021827519794, "grad_norm": 6.443029403686523, "learning_rate": 4.576017036601914e-05, "loss": 1.7392, "step": 2530 }, { "epoch": 1.0847421356730151, "grad_norm": 5.588938236236572, "learning_rate": 4.558755787449189e-05, "loss": 1.7468, "step": 2535 }, { "epoch": 1.0868820885940509, "grad_norm": 5.544746398925781, "learning_rate": 4.541499836722835e-05, "loss": 1.7849, "step": 2540 }, { "epoch": 1.0890220415150866, "grad_norm": 5.714095592498779, "learning_rate": 4.5242493916309705e-05, "loss": 1.7508, "step": 2545 }, { "epoch": 1.0911619944361224, "grad_norm": 5.488661289215088, "learning_rate": 4.507004659315611e-05, "loss": 1.8118, "step": 2550 }, { "epoch": 1.0933019473571581, "grad_norm": 6.51017951965332, "learning_rate": 4.489765846850162e-05, "loss": 1.7632, "step": 2555 }, { "epoch": 1.0954419002781939, "grad_norm": 5.363729476928711, "learning_rate": 4.472533161236954e-05, "loss": 1.7984, "step": 2560 }, { "epoch": 1.0975818531992296, "grad_norm": 5.888598918914795, "learning_rate": 4.4553068094047394e-05, "loss": 1.7983, "step": 2565 }, { "epoch": 1.0997218061202654, "grad_norm": 6.100613594055176, "learning_rate": 4.438086998206221e-05, "loss": 1.7707, "step": 2570 }, { "epoch": 1.1018617590413011, "grad_norm": 5.885534286499023, "learning_rate": 4.4208739344155594e-05, "loss": 1.8299, "step": 2575 }, { "epoch": 1.1040017119623369, "grad_norm": 6.399609088897705, "learning_rate": 4.4036678247258924e-05, "loss": 1.7365, "step": 2580 }, { "epoch": 1.1061416648833726, "grad_norm": 7.783679008483887, "learning_rate": 4.386468875746854e-05, "loss": 1.8312, "step": 2585 }, { "epoch": 1.1082816178044084, "grad_norm": 5.783539772033691, "learning_rate": 4.369277294002093e-05, "loss": 1.7481, "step": 2590 }, { "epoch": 1.1104215707254441, "grad_norm": 5.326051235198975, "learning_rate": 4.352093285926793e-05, "loss": 1.7541, "step": 2595 }, { "epoch": 1.1125615236464799, "grad_norm": 5.783180236816406, "learning_rate": 4.334917057865194e-05, "loss": 1.7801, "step": 2600 }, { "epoch": 1.1147014765675156, "grad_norm": 5.667355537414551, "learning_rate": 4.3177488160681146e-05, "loss": 1.7967, "step": 2605 }, { "epoch": 1.1168414294885514, "grad_norm": 5.82685661315918, "learning_rate": 4.300588766690473e-05, "loss": 1.7543, "step": 2610 }, { "epoch": 1.1189813824095869, "grad_norm": 6.610193252563477, "learning_rate": 4.283437115788814e-05, "loss": 1.7785, "step": 2615 }, { "epoch": 1.1211213353306226, "grad_norm": 7.57377815246582, "learning_rate": 4.266294069318837e-05, "loss": 1.7679, "step": 2620 }, { "epoch": 1.1232612882516584, "grad_norm": 6.202375411987305, "learning_rate": 4.2491598331329154e-05, "loss": 1.8045, "step": 2625 }, { "epoch": 1.1254012411726941, "grad_norm": 6.378594875335693, "learning_rate": 4.232034612977634e-05, "loss": 1.7454, "step": 2630 }, { "epoch": 1.1275411940937299, "grad_norm": 5.158632755279541, "learning_rate": 4.21491861449131e-05, "loss": 1.7993, "step": 2635 }, { "epoch": 1.1296811470147656, "grad_norm": 6.776663780212402, "learning_rate": 4.19781204320153e-05, "loss": 1.7253, "step": 2640 }, { "epoch": 1.1318210999358014, "grad_norm": 5.978639125823975, "learning_rate": 4.180715104522679e-05, "loss": 1.7127, "step": 2645 }, { "epoch": 1.1339610528568371, "grad_norm": 5.210892200469971, "learning_rate": 4.1636280037534725e-05, "loss": 1.7764, "step": 2650 }, { "epoch": 1.1361010057778729, "grad_norm": 6.2656121253967285, "learning_rate": 4.1465509460744963e-05, "loss": 1.7425, "step": 2655 }, { "epoch": 1.1382409586989086, "grad_norm": 5.737490177154541, "learning_rate": 4.129484136545737e-05, "loss": 1.7626, "step": 2660 }, { "epoch": 1.1403809116199444, "grad_norm": 5.1411261558532715, "learning_rate": 4.112427780104124e-05, "loss": 1.7971, "step": 2665 }, { "epoch": 1.1425208645409801, "grad_norm": 5.617253303527832, "learning_rate": 4.0953820815610636e-05, "loss": 1.7431, "step": 2670 }, { "epoch": 1.1446608174620159, "grad_norm": 6.250400543212891, "learning_rate": 4.0783472455999886e-05, "loss": 1.7756, "step": 2675 }, { "epoch": 1.1468007703830516, "grad_norm": 7.598665714263916, "learning_rate": 4.06132347677389e-05, "loss": 1.7462, "step": 2680 }, { "epoch": 1.1489407233040874, "grad_norm": 5.751972675323486, "learning_rate": 4.0443109795028665e-05, "loss": 1.7599, "step": 2685 }, { "epoch": 1.1510806762251231, "grad_norm": 6.453872203826904, "learning_rate": 4.0273099580716725e-05, "loss": 1.7637, "step": 2690 }, { "epoch": 1.1532206291461589, "grad_norm": 6.619513034820557, "learning_rate": 4.010320616627252e-05, "loss": 1.7396, "step": 2695 }, { "epoch": 1.1553605820671946, "grad_norm": 6.701980113983154, "learning_rate": 3.993343159176307e-05, "loss": 1.6959, "step": 2700 }, { "epoch": 1.1575005349882304, "grad_norm": 5.974419116973877, "learning_rate": 3.976377789582832e-05, "loss": 1.8088, "step": 2705 }, { "epoch": 1.1596404879092659, "grad_norm": 5.028608798980713, "learning_rate": 3.959424711565672e-05, "loss": 1.7439, "step": 2710 }, { "epoch": 1.1617804408303019, "grad_norm": 5.743880748748779, "learning_rate": 3.9424841286960784e-05, "loss": 1.748, "step": 2715 }, { "epoch": 1.1639203937513374, "grad_norm": 5.568974018096924, "learning_rate": 3.925556244395259e-05, "loss": 1.7883, "step": 2720 }, { "epoch": 1.1660603466723731, "grad_norm": 5.259049892425537, "learning_rate": 3.908641261931937e-05, "loss": 1.7304, "step": 2725 }, { "epoch": 1.1682002995934089, "grad_norm": 5.7519850730896, "learning_rate": 3.8917393844199156e-05, "loss": 1.7241, "step": 2730 }, { "epoch": 1.1703402525144446, "grad_norm": 5.750739097595215, "learning_rate": 3.874850814815632e-05, "loss": 1.6955, "step": 2735 }, { "epoch": 1.1724802054354804, "grad_norm": 6.248133182525635, "learning_rate": 3.857975755915723e-05, "loss": 1.7689, "step": 2740 }, { "epoch": 1.1746201583565161, "grad_norm": 5.397186756134033, "learning_rate": 3.8411144103545904e-05, "loss": 1.7588, "step": 2745 }, { "epoch": 1.1767601112775519, "grad_norm": 5.462392330169678, "learning_rate": 3.824266980601968e-05, "loss": 1.7879, "step": 2750 }, { "epoch": 1.1789000641985876, "grad_norm": 6.177926063537598, "learning_rate": 3.807433668960488e-05, "loss": 1.7516, "step": 2755 }, { "epoch": 1.1810400171196234, "grad_norm": 5.820185661315918, "learning_rate": 3.7906146775632554e-05, "loss": 1.7227, "step": 2760 }, { "epoch": 1.1831799700406591, "grad_norm": 5.526188850402832, "learning_rate": 3.773810208371417e-05, "loss": 1.7781, "step": 2765 }, { "epoch": 1.1853199229616949, "grad_norm": 5.152137756347656, "learning_rate": 3.7570204631717395e-05, "loss": 1.775, "step": 2770 }, { "epoch": 1.1874598758827306, "grad_norm": 6.030661106109619, "learning_rate": 3.740245643574184e-05, "loss": 1.8056, "step": 2775 }, { "epoch": 1.1895998288037664, "grad_norm": 5.847061634063721, "learning_rate": 3.723485951009486e-05, "loss": 1.7345, "step": 2780 }, { "epoch": 1.191739781724802, "grad_norm": 5.8102126121521, "learning_rate": 3.706741586726738e-05, "loss": 1.7519, "step": 2785 }, { "epoch": 1.1938797346458379, "grad_norm": 6.639953136444092, "learning_rate": 3.690012751790972e-05, "loss": 1.7485, "step": 2790 }, { "epoch": 1.1960196875668736, "grad_norm": 5.996955871582031, "learning_rate": 3.673299647080742e-05, "loss": 1.7072, "step": 2795 }, { "epoch": 1.1981596404879093, "grad_norm": 5.289943218231201, "learning_rate": 3.656602473285717e-05, "loss": 1.7831, "step": 2800 }, { "epoch": 1.200299593408945, "grad_norm": 5.406320571899414, "learning_rate": 3.639921430904268e-05, "loss": 1.7397, "step": 2805 }, { "epoch": 1.2024395463299808, "grad_norm": 6.86167573928833, "learning_rate": 3.6232567202410624e-05, "loss": 1.7408, "step": 2810 }, { "epoch": 1.2045794992510164, "grad_norm": 5.778091907501221, "learning_rate": 3.606608541404656e-05, "loss": 1.765, "step": 2815 }, { "epoch": 1.2067194521720521, "grad_norm": 6.24983024597168, "learning_rate": 3.5899770943050924e-05, "loss": 1.7543, "step": 2820 }, { "epoch": 1.2088594050930879, "grad_norm": 6.323977470397949, "learning_rate": 3.573362578651501e-05, "loss": 1.7168, "step": 2825 }, { "epoch": 1.2109993580141236, "grad_norm": 6.535098075866699, "learning_rate": 3.556765193949704e-05, "loss": 1.7583, "step": 2830 }, { "epoch": 1.2131393109351594, "grad_norm": 5.957965850830078, "learning_rate": 3.5401851394998084e-05, "loss": 1.7272, "step": 2835 }, { "epoch": 1.2152792638561951, "grad_norm": 5.54447078704834, "learning_rate": 3.523622614393827e-05, "loss": 1.8009, "step": 2840 }, { "epoch": 1.2174192167772309, "grad_norm": 5.613691329956055, "learning_rate": 3.5070778175132806e-05, "loss": 1.7559, "step": 2845 }, { "epoch": 1.2195591696982666, "grad_norm": 5.480210781097412, "learning_rate": 3.4905509475268104e-05, "loss": 1.74, "step": 2850 }, { "epoch": 1.2216991226193024, "grad_norm": 6.216781139373779, "learning_rate": 3.474042202887792e-05, "loss": 1.7272, "step": 2855 }, { "epoch": 1.223839075540338, "grad_norm": 6.038024425506592, "learning_rate": 3.4575517818319534e-05, "loss": 1.7408, "step": 2860 }, { "epoch": 1.2259790284613739, "grad_norm": 5.762340068817139, "learning_rate": 3.4410798823749964e-05, "loss": 1.7576, "step": 2865 }, { "epoch": 1.2281189813824096, "grad_norm": 6.565494060516357, "learning_rate": 3.424626702310214e-05, "loss": 1.7008, "step": 2870 }, { "epoch": 1.2302589343034454, "grad_norm": 6.1579508781433105, "learning_rate": 3.408192439206121e-05, "loss": 1.7218, "step": 2875 }, { "epoch": 1.232398887224481, "grad_norm": 5.965446949005127, "learning_rate": 3.391777290404077e-05, "loss": 1.6989, "step": 2880 }, { "epoch": 1.2345388401455168, "grad_norm": 6.2155070304870605, "learning_rate": 3.3753814530159176e-05, "loss": 1.7461, "step": 2885 }, { "epoch": 1.2366787930665526, "grad_norm": 6.247200965881348, "learning_rate": 3.3590051239215916e-05, "loss": 1.7542, "step": 2890 }, { "epoch": 1.2388187459875883, "grad_norm": 4.975124359130859, "learning_rate": 3.342648499766791e-05, "loss": 1.7299, "step": 2895 }, { "epoch": 1.240958698908624, "grad_norm": 5.148104190826416, "learning_rate": 3.326311776960593e-05, "loss": 1.7252, "step": 2900 }, { "epoch": 1.2430986518296598, "grad_norm": 6.518229961395264, "learning_rate": 3.309995151673103e-05, "loss": 1.754, "step": 2905 }, { "epoch": 1.2452386047506954, "grad_norm": 5.220431804656982, "learning_rate": 3.293698819833093e-05, "loss": 1.7349, "step": 2910 }, { "epoch": 1.2473785576717313, "grad_norm": 5.202417373657227, "learning_rate": 3.277422977125656e-05, "loss": 1.7521, "step": 2915 }, { "epoch": 1.2495185105927669, "grad_norm": 5.129791736602783, "learning_rate": 3.2611678189898523e-05, "loss": 1.7653, "step": 2920 }, { "epoch": 1.2516584635138028, "grad_norm": 6.944059371948242, "learning_rate": 3.244933540616363e-05, "loss": 1.7411, "step": 2925 }, { "epoch": 1.2537984164348384, "grad_norm": 5.993988513946533, "learning_rate": 3.228720336945148e-05, "loss": 1.7544, "step": 2930 }, { "epoch": 1.255938369355874, "grad_norm": 5.735143184661865, "learning_rate": 3.2125284026631006e-05, "loss": 1.7374, "step": 2935 }, { "epoch": 1.2580783222769099, "grad_norm": 5.272975444793701, "learning_rate": 3.196357932201717e-05, "loss": 1.7362, "step": 2940 }, { "epoch": 1.2602182751979456, "grad_norm": 6.259987831115723, "learning_rate": 3.1802091197347544e-05, "loss": 1.745, "step": 2945 }, { "epoch": 1.2623582281189814, "grad_norm": 5.6015400886535645, "learning_rate": 3.164082159175904e-05, "loss": 1.7478, "step": 2950 }, { "epoch": 1.264498181040017, "grad_norm": 6.1617536544799805, "learning_rate": 3.147977244176461e-05, "loss": 1.757, "step": 2955 }, { "epoch": 1.2666381339610528, "grad_norm": 4.88234806060791, "learning_rate": 3.131894568122999e-05, "loss": 1.73, "step": 2960 }, { "epoch": 1.2687780868820886, "grad_norm": 5.751287460327148, "learning_rate": 3.115834324135047e-05, "loss": 1.7807, "step": 2965 }, { "epoch": 1.2709180398031243, "grad_norm": 5.901682376861572, "learning_rate": 3.099796705062773e-05, "loss": 1.7535, "step": 2970 }, { "epoch": 1.27305799272416, "grad_norm": 5.5405144691467285, "learning_rate": 3.083781903484667e-05, "loss": 1.7742, "step": 2975 }, { "epoch": 1.2751979456451958, "grad_norm": 5.621067523956299, "learning_rate": 3.067790111705225e-05, "loss": 1.6996, "step": 2980 }, { "epoch": 1.2773378985662316, "grad_norm": 5.193610191345215, "learning_rate": 3.051821521752647e-05, "loss": 1.7568, "step": 2985 }, { "epoch": 1.2794778514872673, "grad_norm": 5.218043804168701, "learning_rate": 3.0358763253765264e-05, "loss": 1.7299, "step": 2990 }, { "epoch": 1.281617804408303, "grad_norm": 6.297114849090576, "learning_rate": 3.0199547140455474e-05, "loss": 1.7814, "step": 2995 }, { "epoch": 1.2837577573293388, "grad_norm": 6.044188976287842, "learning_rate": 3.0040568789451862e-05, "loss": 1.7381, "step": 3000 }, { "epoch": 1.2858977102503744, "grad_norm": 5.990639686584473, "learning_rate": 2.9881830109754184e-05, "loss": 1.7458, "step": 3005 }, { "epoch": 1.2880376631714103, "grad_norm": 5.563923358917236, "learning_rate": 2.9723333007484218e-05, "loss": 1.7403, "step": 3010 }, { "epoch": 1.2901776160924459, "grad_norm": 5.493687152862549, "learning_rate": 2.9565079385862903e-05, "loss": 1.7141, "step": 3015 }, { "epoch": 1.2923175690134818, "grad_norm": 5.307217121124268, "learning_rate": 2.9407071145187494e-05, "loss": 1.7546, "step": 3020 }, { "epoch": 1.2944575219345174, "grad_norm": 5.565853118896484, "learning_rate": 2.924931018280871e-05, "loss": 1.7392, "step": 3025 }, { "epoch": 1.296597474855553, "grad_norm": 5.918313503265381, "learning_rate": 2.9091798393107994e-05, "loss": 1.7598, "step": 3030 }, { "epoch": 1.2987374277765888, "grad_norm": 5.450323104858398, "learning_rate": 2.8934537667474732e-05, "loss": 1.7204, "step": 3035 }, { "epoch": 1.3008773806976246, "grad_norm": 5.59390115737915, "learning_rate": 2.8777529894283538e-05, "loss": 1.7345, "step": 3040 }, { "epoch": 1.3030173336186603, "grad_norm": 5.590770721435547, "learning_rate": 2.8620776958871627e-05, "loss": 1.7814, "step": 3045 }, { "epoch": 1.305157286539696, "grad_norm": 5.129935264587402, "learning_rate": 2.8464280743516102e-05, "loss": 1.7527, "step": 3050 }, { "epoch": 1.3072972394607318, "grad_norm": 6.99540376663208, "learning_rate": 2.8308043127411423e-05, "loss": 1.7197, "step": 3055 }, { "epoch": 1.3094371923817676, "grad_norm": 6.13936710357666, "learning_rate": 2.8152065986646788e-05, "loss": 1.7153, "step": 3060 }, { "epoch": 1.3115771453028033, "grad_norm": 5.489097595214844, "learning_rate": 2.7996351194183646e-05, "loss": 1.7562, "step": 3065 }, { "epoch": 1.313717098223839, "grad_norm": 6.8664398193359375, "learning_rate": 2.784090061983317e-05, "loss": 1.7034, "step": 3070 }, { "epoch": 1.3158570511448748, "grad_norm": 5.367057800292969, "learning_rate": 2.7685716130233842e-05, "loss": 1.6691, "step": 3075 }, { "epoch": 1.3179970040659106, "grad_norm": 6.975956916809082, "learning_rate": 2.7530799588829005e-05, "loss": 1.7107, "step": 3080 }, { "epoch": 1.3201369569869463, "grad_norm": 6.1998443603515625, "learning_rate": 2.7376152855844495e-05, "loss": 1.7955, "step": 3085 }, { "epoch": 1.322276909907982, "grad_norm": 6.424160957336426, "learning_rate": 2.7221777788266324e-05, "loss": 1.7312, "step": 3090 }, { "epoch": 1.3244168628290178, "grad_norm": 5.118837356567383, "learning_rate": 2.7067676239818364e-05, "loss": 1.7105, "step": 3095 }, { "epoch": 1.3265568157500536, "grad_norm": 5.895681858062744, "learning_rate": 2.6913850060940083e-05, "loss": 1.7077, "step": 3100 }, { "epoch": 1.3286967686710893, "grad_norm": 6.501992702484131, "learning_rate": 2.676030109876434e-05, "loss": 1.728, "step": 3105 }, { "epoch": 1.3308367215921248, "grad_norm": 5.911489486694336, "learning_rate": 2.66070311970952e-05, "loss": 1.7214, "step": 3110 }, { "epoch": 1.3329766745131608, "grad_norm": 5.91454553604126, "learning_rate": 2.6454042196385798e-05, "loss": 1.6895, "step": 3115 }, { "epoch": 1.3351166274341963, "grad_norm": 5.649311065673828, "learning_rate": 2.6301335933716176e-05, "loss": 1.7488, "step": 3120 }, { "epoch": 1.3372565803552323, "grad_norm": 7.105709075927734, "learning_rate": 2.614891424277135e-05, "loss": 1.6687, "step": 3125 }, { "epoch": 1.3393965332762678, "grad_norm": 5.769228458404541, "learning_rate": 2.599677895381919e-05, "loss": 1.7175, "step": 3130 }, { "epoch": 1.3415364861973036, "grad_norm": 7.222846984863281, "learning_rate": 2.5844931893688473e-05, "loss": 1.721, "step": 3135 }, { "epoch": 1.3436764391183393, "grad_norm": 5.122998237609863, "learning_rate": 2.5693374885746957e-05, "loss": 1.7325, "step": 3140 }, { "epoch": 1.345816392039375, "grad_norm": 5.9646148681640625, "learning_rate": 2.5542109749879473e-05, "loss": 1.7016, "step": 3145 }, { "epoch": 1.3479563449604108, "grad_norm": 5.405865669250488, "learning_rate": 2.5391138302466062e-05, "loss": 1.7666, "step": 3150 }, { "epoch": 1.3500962978814466, "grad_norm": 6.047122478485107, "learning_rate": 2.5240462356360184e-05, "loss": 1.7438, "step": 3155 }, { "epoch": 1.3522362508024823, "grad_norm": 6.545495986938477, "learning_rate": 2.5090083720866952e-05, "loss": 1.7322, "step": 3160 }, { "epoch": 1.354376203723518, "grad_norm": 6.075828552246094, "learning_rate": 2.4940004201721384e-05, "loss": 1.7089, "step": 3165 }, { "epoch": 1.3565161566445538, "grad_norm": 5.580131530761719, "learning_rate": 2.479022560106673e-05, "loss": 1.7456, "step": 3170 }, { "epoch": 1.3586561095655896, "grad_norm": 5.123158931732178, "learning_rate": 2.4640749717432854e-05, "loss": 1.7118, "step": 3175 }, { "epoch": 1.3607960624866253, "grad_norm": 5.250161170959473, "learning_rate": 2.4491578345714587e-05, "loss": 1.7271, "step": 3180 }, { "epoch": 1.362936015407661, "grad_norm": 5.588467121124268, "learning_rate": 2.434271327715023e-05, "loss": 1.7026, "step": 3185 }, { "epoch": 1.3650759683286968, "grad_norm": 5.4725518226623535, "learning_rate": 2.419415629930001e-05, "loss": 1.7128, "step": 3190 }, { "epoch": 1.3672159212497326, "grad_norm": 5.664406776428223, "learning_rate": 2.4045909196024624e-05, "loss": 1.7244, "step": 3195 }, { "epoch": 1.3693558741707683, "grad_norm": 4.935891151428223, "learning_rate": 2.389797374746382e-05, "loss": 1.7123, "step": 3200 }, { "epoch": 1.3714958270918038, "grad_norm": 5.0147881507873535, "learning_rate": 2.3750351730015015e-05, "loss": 1.6921, "step": 3205 }, { "epoch": 1.3736357800128398, "grad_norm": 5.382151126861572, "learning_rate": 2.3603044916311963e-05, "loss": 1.78, "step": 3210 }, { "epoch": 1.3757757329338753, "grad_norm": 5.000919818878174, "learning_rate": 2.345605507520352e-05, "loss": 1.699, "step": 3215 }, { "epoch": 1.3779156858549113, "grad_norm": 4.845434665679932, "learning_rate": 2.3309383971732312e-05, "loss": 1.7508, "step": 3220 }, { "epoch": 1.3800556387759468, "grad_norm": 4.90060567855835, "learning_rate": 2.3163033367113602e-05, "loss": 1.6904, "step": 3225 }, { "epoch": 1.3821955916969826, "grad_norm": 5.377330303192139, "learning_rate": 2.3017005018714126e-05, "loss": 1.7236, "step": 3230 }, { "epoch": 1.3843355446180183, "grad_norm": 5.329347133636475, "learning_rate": 2.2871300680030995e-05, "loss": 1.6982, "step": 3235 }, { "epoch": 1.386475497539054, "grad_norm": 4.9429931640625, "learning_rate": 2.2725922100670644e-05, "loss": 1.6973, "step": 3240 }, { "epoch": 1.3886154504600898, "grad_norm": 5.456453323364258, "learning_rate": 2.25808710263278e-05, "loss": 1.7033, "step": 3245 }, { "epoch": 1.3907554033811256, "grad_norm": 5.3230791091918945, "learning_rate": 2.243614919876454e-05, "loss": 1.7447, "step": 3250 }, { "epoch": 1.3928953563021613, "grad_norm": 5.888358116149902, "learning_rate": 2.22917583557894e-05, "loss": 1.7347, "step": 3255 }, { "epoch": 1.395035309223197, "grad_norm": 5.590538024902344, "learning_rate": 2.214770023123641e-05, "loss": 1.7566, "step": 3260 }, { "epoch": 1.3971752621442328, "grad_norm": 5.970736503601074, "learning_rate": 2.2003976554944405e-05, "loss": 1.6941, "step": 3265 }, { "epoch": 1.3993152150652686, "grad_norm": 5.476358413696289, "learning_rate": 2.186058905273618e-05, "loss": 1.7016, "step": 3270 }, { "epoch": 1.4014551679863043, "grad_norm": 5.488466262817383, "learning_rate": 2.1717539446397754e-05, "loss": 1.7063, "step": 3275 }, { "epoch": 1.40359512090734, "grad_norm": 5.273731708526611, "learning_rate": 2.1574829453657746e-05, "loss": 1.7218, "step": 3280 }, { "epoch": 1.4057350738283758, "grad_norm": 5.290144920349121, "learning_rate": 2.1432460788166704e-05, "loss": 1.6931, "step": 3285 }, { "epoch": 1.4078750267494116, "grad_norm": 6.16803503036499, "learning_rate": 2.1290435159476545e-05, "loss": 1.6872, "step": 3290 }, { "epoch": 1.4100149796704473, "grad_norm": 5.351411819458008, "learning_rate": 2.1148754273020038e-05, "loss": 1.678, "step": 3295 }, { "epoch": 1.412154932591483, "grad_norm": 6.350516319274902, "learning_rate": 2.1007419830090306e-05, "loss": 1.728, "step": 3300 }, { "epoch": 1.4142948855125188, "grad_norm": 5.602326393127441, "learning_rate": 2.0866433527820406e-05, "loss": 1.6969, "step": 3305 }, { "epoch": 1.4164348384335543, "grad_norm": 7.531297206878662, "learning_rate": 2.072579705916295e-05, "loss": 1.738, "step": 3310 }, { "epoch": 1.4185747913545903, "grad_norm": 5.831987380981445, "learning_rate": 2.058551211286977e-05, "loss": 1.7518, "step": 3315 }, { "epoch": 1.4207147442756258, "grad_norm": 6.611841678619385, "learning_rate": 2.0445580373471658e-05, "loss": 1.6811, "step": 3320 }, { "epoch": 1.4228546971966618, "grad_norm": 4.91347599029541, "learning_rate": 2.03060035212581e-05, "loss": 1.6897, "step": 3325 }, { "epoch": 1.4249946501176973, "grad_norm": 6.171419620513916, "learning_rate": 2.0166783232257154e-05, "loss": 1.739, "step": 3330 }, { "epoch": 1.427134603038733, "grad_norm": 5.1369524002075195, "learning_rate": 2.0027921178215274e-05, "loss": 1.7164, "step": 3335 }, { "epoch": 1.4292745559597688, "grad_norm": 6.778637409210205, "learning_rate": 1.9889419026577266e-05, "loss": 1.7224, "step": 3340 }, { "epoch": 1.4314145088808046, "grad_norm": 4.898647785186768, "learning_rate": 1.9751278440466248e-05, "loss": 1.6844, "step": 3345 }, { "epoch": 1.4335544618018403, "grad_norm": 5.825972080230713, "learning_rate": 1.9613501078663693e-05, "loss": 1.7451, "step": 3350 }, { "epoch": 1.435694414722876, "grad_norm": 5.397356986999512, "learning_rate": 1.9476088595589504e-05, "loss": 1.715, "step": 3355 }, { "epoch": 1.4378343676439118, "grad_norm": 5.8002519607543945, "learning_rate": 1.9339042641282146e-05, "loss": 1.7006, "step": 3360 }, { "epoch": 1.4399743205649476, "grad_norm": 5.10746431350708, "learning_rate": 1.9202364861378842e-05, "loss": 1.6758, "step": 3365 }, { "epoch": 1.4421142734859833, "grad_norm": 5.267380237579346, "learning_rate": 1.9066056897095796e-05, "loss": 1.7395, "step": 3370 }, { "epoch": 1.444254226407019, "grad_norm": 5.395501136779785, "learning_rate": 1.8930120385208495e-05, "loss": 1.7082, "step": 3375 }, { "epoch": 1.4463941793280548, "grad_norm": 5.383699893951416, "learning_rate": 1.8794556958032062e-05, "loss": 1.7182, "step": 3380 }, { "epoch": 1.4485341322490906, "grad_norm": 5.4255290031433105, "learning_rate": 1.865936824340164e-05, "loss": 1.6796, "step": 3385 }, { "epoch": 1.4506740851701263, "grad_norm": 5.540951728820801, "learning_rate": 1.8524555864652865e-05, "loss": 1.7312, "step": 3390 }, { "epoch": 1.452814038091162, "grad_norm": 6.032477855682373, "learning_rate": 1.839012144060236e-05, "loss": 1.6458, "step": 3395 }, { "epoch": 1.4549539910121978, "grad_norm": 5.043034553527832, "learning_rate": 1.825606658552826e-05, "loss": 1.6969, "step": 3400 }, { "epoch": 1.4570939439332335, "grad_norm": 5.323840618133545, "learning_rate": 1.8122392909150904e-05, "loss": 1.7225, "step": 3405 }, { "epoch": 1.4592338968542693, "grad_norm": 5.682520389556885, "learning_rate": 1.7989102016613463e-05, "loss": 1.6943, "step": 3410 }, { "epoch": 1.4613738497753048, "grad_norm": 5.428680419921875, "learning_rate": 1.7856195508462663e-05, "loss": 1.6992, "step": 3415 }, { "epoch": 1.4635138026963408, "grad_norm": 5.1605024337768555, "learning_rate": 1.7723674980629572e-05, "loss": 1.6569, "step": 3420 }, { "epoch": 1.4656537556173763, "grad_norm": 5.788964748382568, "learning_rate": 1.759154202441044e-05, "loss": 1.7081, "step": 3425 }, { "epoch": 1.4677937085384123, "grad_norm": 5.63525390625, "learning_rate": 1.7459798226447577e-05, "loss": 1.7133, "step": 3430 }, { "epoch": 1.4699336614594478, "grad_norm": 5.6762003898620605, "learning_rate": 1.7328445168710323e-05, "loss": 1.6922, "step": 3435 }, { "epoch": 1.4720736143804836, "grad_norm": 5.393453121185303, "learning_rate": 1.7197484428476023e-05, "loss": 1.6721, "step": 3440 }, { "epoch": 1.4742135673015193, "grad_norm": 4.811567783355713, "learning_rate": 1.7066917578311132e-05, "loss": 1.7071, "step": 3445 }, { "epoch": 1.476353520222555, "grad_norm": 5.325284004211426, "learning_rate": 1.693674618605227e-05, "loss": 1.7109, "step": 3450 }, { "epoch": 1.4784934731435908, "grad_norm": 5.2245774269104, "learning_rate": 1.6806971814787458e-05, "loss": 1.7103, "step": 3455 }, { "epoch": 1.4806334260646266, "grad_norm": 5.08640718460083, "learning_rate": 1.66775960228373e-05, "loss": 1.7094, "step": 3460 }, { "epoch": 1.4827733789856623, "grad_norm": 4.874035835266113, "learning_rate": 1.6548620363736294e-05, "loss": 1.7177, "step": 3465 }, { "epoch": 1.484913331906698, "grad_norm": 5.959578037261963, "learning_rate": 1.6420046386214184e-05, "loss": 1.6586, "step": 3470 }, { "epoch": 1.4870532848277338, "grad_norm": 5.682189464569092, "learning_rate": 1.6291875634177346e-05, "loss": 1.6731, "step": 3475 }, { "epoch": 1.4891932377487695, "grad_norm": 5.864963054656982, "learning_rate": 1.616410964669025e-05, "loss": 1.6708, "step": 3480 }, { "epoch": 1.4913331906698053, "grad_norm": 5.3527140617370605, "learning_rate": 1.6036749957956993e-05, "loss": 1.6596, "step": 3485 }, { "epoch": 1.493473143590841, "grad_norm": 4.715349197387695, "learning_rate": 1.5909798097302865e-05, "loss": 1.6889, "step": 3490 }, { "epoch": 1.4956130965118768, "grad_norm": 5.459346771240234, "learning_rate": 1.578325558915598e-05, "loss": 1.6716, "step": 3495 }, { "epoch": 1.4977530494329125, "grad_norm": 5.306508541107178, "learning_rate": 1.5657123953029003e-05, "loss": 1.673, "step": 3500 }, { "epoch": 1.4998930023539483, "grad_norm": 5.33494234085083, "learning_rate": 1.5531404703500845e-05, "loss": 1.7047, "step": 3505 }, { "epoch": 1.5020329552749838, "grad_norm": 5.494200229644775, "learning_rate": 1.5406099350198544e-05, "loss": 1.7061, "step": 3510 }, { "epoch": 1.5041729081960198, "grad_norm": 4.832003116607666, "learning_rate": 1.528120939777908e-05, "loss": 1.6623, "step": 3515 }, { "epoch": 1.5063128611170553, "grad_norm": 5.2969970703125, "learning_rate": 1.5156736345911342e-05, "loss": 1.6838, "step": 3520 }, { "epoch": 1.5084528140380913, "grad_norm": 5.263791084289551, "learning_rate": 1.5032681689258105e-05, "loss": 1.691, "step": 3525 }, { "epoch": 1.5105927669591268, "grad_norm": 5.589069843292236, "learning_rate": 1.4909046917458097e-05, "loss": 1.7456, "step": 3530 }, { "epoch": 1.5127327198801628, "grad_norm": 5.666341781616211, "learning_rate": 1.4785833515108088e-05, "loss": 1.7284, "step": 3535 }, { "epoch": 1.5148726728011983, "grad_norm": 5.460354328155518, "learning_rate": 1.4663042961745083e-05, "loss": 1.6767, "step": 3540 }, { "epoch": 1.5170126257222343, "grad_norm": 5.311362266540527, "learning_rate": 1.4540676731828546e-05, "loss": 1.662, "step": 3545 }, { "epoch": 1.5191525786432698, "grad_norm": 5.237598419189453, "learning_rate": 1.4418736294722701e-05, "loss": 1.6841, "step": 3550 }, { "epoch": 1.5212925315643056, "grad_norm": 5.252871513366699, "learning_rate": 1.4297223114678887e-05, "loss": 1.7057, "step": 3555 }, { "epoch": 1.5234324844853413, "grad_norm": 4.881004333496094, "learning_rate": 1.4176138650817967e-05, "loss": 1.699, "step": 3560 }, { "epoch": 1.525572437406377, "grad_norm": 5.0006103515625, "learning_rate": 1.4055484357112808e-05, "loss": 1.7438, "step": 3565 }, { "epoch": 1.5277123903274128, "grad_norm": 4.589049339294434, "learning_rate": 1.3935261682370849e-05, "loss": 1.7316, "step": 3570 }, { "epoch": 1.5298523432484485, "grad_norm": 5.060586929321289, "learning_rate": 1.3815472070216656e-05, "loss": 1.7279, "step": 3575 }, { "epoch": 1.5319922961694843, "grad_norm": 5.7035322189331055, "learning_rate": 1.3696116959074635e-05, "loss": 1.7232, "step": 3580 }, { "epoch": 1.53413224909052, "grad_norm": 5.462157726287842, "learning_rate": 1.3577197782151724e-05, "loss": 1.674, "step": 3585 }, { "epoch": 1.5362722020115558, "grad_norm": 5.079245567321777, "learning_rate": 1.3458715967420193e-05, "loss": 1.7135, "step": 3590 }, { "epoch": 1.5384121549325915, "grad_norm": 5.461935043334961, "learning_rate": 1.3340672937600518e-05, "loss": 1.6261, "step": 3595 }, { "epoch": 1.5405521078536273, "grad_norm": 5.3971076011657715, "learning_rate": 1.3223070110144265e-05, "loss": 1.7194, "step": 3600 }, { "epoch": 1.5426920607746628, "grad_norm": 5.163456916809082, "learning_rate": 1.3105908897217084e-05, "loss": 1.6877, "step": 3605 }, { "epoch": 1.5448320136956988, "grad_norm": 6.106560707092285, "learning_rate": 1.2989190705681758e-05, "loss": 1.6893, "step": 3610 }, { "epoch": 1.5469719666167343, "grad_norm": 5.627121925354004, "learning_rate": 1.2872916937081308e-05, "loss": 1.6939, "step": 3615 }, { "epoch": 1.5491119195377703, "grad_norm": 5.5462117195129395, "learning_rate": 1.2757088987622152e-05, "loss": 1.6919, "step": 3620 }, { "epoch": 1.5512518724588058, "grad_norm": 5.333662986755371, "learning_rate": 1.2641708248157341e-05, "loss": 1.6444, "step": 3625 }, { "epoch": 1.5533918253798418, "grad_norm": 6.698999881744385, "learning_rate": 1.2526776104169868e-05, "loss": 1.7085, "step": 3630 }, { "epoch": 1.5555317783008773, "grad_norm": 5.591616630554199, "learning_rate": 1.241229393575603e-05, "loss": 1.6643, "step": 3635 }, { "epoch": 1.5576717312219133, "grad_norm": 5.279683589935303, "learning_rate": 1.2298263117608855e-05, "loss": 1.624, "step": 3640 }, { "epoch": 1.5598116841429488, "grad_norm": 5.046063423156738, "learning_rate": 1.2184685019001574e-05, "loss": 1.6992, "step": 3645 }, { "epoch": 1.5619516370639845, "grad_norm": 5.270909786224365, "learning_rate": 1.2071561003771214e-05, "loss": 1.6567, "step": 3650 }, { "epoch": 1.5640915899850203, "grad_norm": 5.26682186126709, "learning_rate": 1.1958892430302198e-05, "loss": 1.7055, "step": 3655 }, { "epoch": 1.566231542906056, "grad_norm": 5.099318504333496, "learning_rate": 1.184668065151005e-05, "loss": 1.7185, "step": 3660 }, { "epoch": 1.5683714958270918, "grad_norm": 5.019136905670166, "learning_rate": 1.1734927014825115e-05, "loss": 1.7033, "step": 3665 }, { "epoch": 1.5705114487481275, "grad_norm": 5.4806437492370605, "learning_rate": 1.162363286217642e-05, "loss": 1.7044, "step": 3670 }, { "epoch": 1.5726514016691633, "grad_norm": 4.8851118087768555, "learning_rate": 1.151279952997556e-05, "loss": 1.6831, "step": 3675 }, { "epoch": 1.574791354590199, "grad_norm": 5.282357692718506, "learning_rate": 1.1402428349100585e-05, "loss": 1.6689, "step": 3680 }, { "epoch": 1.5769313075112348, "grad_norm": 5.409157752990723, "learning_rate": 1.1292520644880105e-05, "loss": 1.6455, "step": 3685 }, { "epoch": 1.5790712604322705, "grad_norm": 5.711396217346191, "learning_rate": 1.1183077737077336e-05, "loss": 1.6363, "step": 3690 }, { "epoch": 1.5812112133533063, "grad_norm": 5.7200541496276855, "learning_rate": 1.107410093987425e-05, "loss": 1.6737, "step": 3695 }, { "epoch": 1.5833511662743418, "grad_norm": 5.475973606109619, "learning_rate": 1.0965591561855788e-05, "loss": 1.6393, "step": 3700 }, { "epoch": 1.5854911191953778, "grad_norm": 5.419764995574951, "learning_rate": 1.0857550905994175e-05, "loss": 1.6862, "step": 3705 }, { "epoch": 1.5876310721164133, "grad_norm": 4.915116786956787, "learning_rate": 1.0749980269633243e-05, "loss": 1.6878, "step": 3710 }, { "epoch": 1.5897710250374493, "grad_norm": 5.841743469238281, "learning_rate": 1.0642880944472878e-05, "loss": 1.6474, "step": 3715 }, { "epoch": 1.5919109779584848, "grad_norm": 5.388934135437012, "learning_rate": 1.0536254216553487e-05, "loss": 1.6828, "step": 3720 }, { "epoch": 1.5940509308795208, "grad_norm": 4.492440223693848, "learning_rate": 1.0430101366240575e-05, "loss": 1.6947, "step": 3725 }, { "epoch": 1.5961908838005563, "grad_norm": 4.8876495361328125, "learning_rate": 1.0324423668209349e-05, "loss": 1.7048, "step": 3730 }, { "epoch": 1.5983308367215923, "grad_norm": 5.363291263580322, "learning_rate": 1.021922239142944e-05, "loss": 1.6837, "step": 3735 }, { "epoch": 1.6004707896426278, "grad_norm": 5.814448356628418, "learning_rate": 1.0114498799149635e-05, "loss": 1.6784, "step": 3740 }, { "epoch": 1.6026107425636638, "grad_norm": 5.522982120513916, "learning_rate": 1.0010254148882731e-05, "loss": 1.6786, "step": 3745 }, { "epoch": 1.6047506954846993, "grad_norm": 5.147754192352295, "learning_rate": 9.906489692390426e-06, "loss": 1.7548, "step": 3750 }, { "epoch": 1.606890648405735, "grad_norm": 5.689370632171631, "learning_rate": 9.803206675668286e-06, "loss": 1.6861, "step": 3755 }, { "epoch": 1.6090306013267708, "grad_norm": 5.778447151184082, "learning_rate": 9.700406338930778e-06, "loss": 1.7062, "step": 3760 }, { "epoch": 1.6111705542478065, "grad_norm": 4.873246669769287, "learning_rate": 9.59808991659641e-06, "loss": 1.6678, "step": 3765 }, { "epoch": 1.6133105071688423, "grad_norm": 5.669099807739258, "learning_rate": 9.496258637272849e-06, "loss": 1.6874, "step": 3770 }, { "epoch": 1.615450460089878, "grad_norm": 5.129676818847656, "learning_rate": 9.394913723742227e-06, "loss": 1.6908, "step": 3775 }, { "epoch": 1.6175904130109138, "grad_norm": 5.257515907287598, "learning_rate": 9.294056392946427e-06, "loss": 1.673, "step": 3780 }, { "epoch": 1.6197303659319495, "grad_norm": 4.824296474456787, "learning_rate": 9.193687855972466e-06, "loss": 1.673, "step": 3785 }, { "epoch": 1.6218703188529853, "grad_norm": 5.123586654663086, "learning_rate": 9.093809318037989e-06, "loss": 1.7033, "step": 3790 }, { "epoch": 1.624010271774021, "grad_norm": 4.633101463317871, "learning_rate": 8.994421978476735e-06, "loss": 1.693, "step": 3795 }, { "epoch": 1.6261502246950568, "grad_norm": 5.188746452331543, "learning_rate": 8.8955270307242e-06, "loss": 1.6694, "step": 3800 }, { "epoch": 1.6282901776160923, "grad_norm": 5.478397369384766, "learning_rate": 8.797125662303257e-06, "loss": 1.6574, "step": 3805 }, { "epoch": 1.6304301305371283, "grad_norm": 5.444027900695801, "learning_rate": 8.699219054809937e-06, "loss": 1.6467, "step": 3810 }, { "epoch": 1.6325700834581638, "grad_norm": 4.998203754425049, "learning_rate": 8.6018083838992e-06, "loss": 1.6715, "step": 3815 }, { "epoch": 1.6347100363791998, "grad_norm": 5.098468780517578, "learning_rate": 8.504894819270854e-06, "loss": 1.6857, "step": 3820 }, { "epoch": 1.6368499893002353, "grad_norm": 5.074070453643799, "learning_rate": 8.408479524655477e-06, "loss": 1.728, "step": 3825 }, { "epoch": 1.6389899422212713, "grad_norm": 5.682558536529541, "learning_rate": 8.312563657800475e-06, "loss": 1.6509, "step": 3830 }, { "epoch": 1.6411298951423068, "grad_norm": 5.497644424438477, "learning_rate": 8.217148370456152e-06, "loss": 1.6649, "step": 3835 }, { "epoch": 1.6432698480633428, "grad_norm": 4.901436805725098, "learning_rate": 8.122234808361907e-06, "loss": 1.6662, "step": 3840 }, { "epoch": 1.6454098009843783, "grad_norm": 4.8296732902526855, "learning_rate": 8.027824111232435e-06, "loss": 1.6546, "step": 3845 }, { "epoch": 1.647549753905414, "grad_norm": 5.374551773071289, "learning_rate": 7.933917412744097e-06, "loss": 1.6182, "step": 3850 }, { "epoch": 1.6496897068264498, "grad_norm": 4.431797504425049, "learning_rate": 7.840515840521263e-06, "loss": 1.6917, "step": 3855 }, { "epoch": 1.6518296597474855, "grad_norm": 4.973811626434326, "learning_rate": 7.747620516122777e-06, "loss": 1.625, "step": 3860 }, { "epoch": 1.6539696126685213, "grad_norm": 5.412280082702637, "learning_rate": 7.655232555028518e-06, "loss": 1.6722, "step": 3865 }, { "epoch": 1.656109565589557, "grad_norm": 5.874164581298828, "learning_rate": 7.563353066625972e-06, "loss": 1.6448, "step": 3870 }, { "epoch": 1.6582495185105928, "grad_norm": 5.387955188751221, "learning_rate": 7.471983154196932e-06, "loss": 1.6627, "step": 3875 }, { "epoch": 1.6603894714316285, "grad_norm": 4.935237407684326, "learning_rate": 7.381123914904231e-06, "loss": 1.6915, "step": 3880 }, { "epoch": 1.6625294243526643, "grad_norm": 4.828623294830322, "learning_rate": 7.2907764397785845e-06, "loss": 1.6893, "step": 3885 }, { "epoch": 1.6646693772737, "grad_norm": 5.0376434326171875, "learning_rate": 7.200941813705497e-06, "loss": 1.6789, "step": 3890 }, { "epoch": 1.6668093301947358, "grad_norm": 5.704905986785889, "learning_rate": 7.111621115412193e-06, "loss": 1.6262, "step": 3895 }, { "epoch": 1.6689492831157713, "grad_norm": 5.178145885467529, "learning_rate": 7.02281541745472e-06, "loss": 1.7011, "step": 3900 }, { "epoch": 1.6710892360368073, "grad_norm": 5.51984167098999, "learning_rate": 6.9345257862050264e-06, "loss": 1.6325, "step": 3905 }, { "epoch": 1.6732291889578428, "grad_norm": 5.094444274902344, "learning_rate": 6.846753281838169e-06, "loss": 1.687, "step": 3910 }, { "epoch": 1.6753691418788788, "grad_norm": 4.961489200592041, "learning_rate": 6.759498958319599e-06, "loss": 1.7111, "step": 3915 }, { "epoch": 1.6775090947999143, "grad_norm": 5.067066669464111, "learning_rate": 6.6727638633924725e-06, "loss": 1.6547, "step": 3920 }, { "epoch": 1.6796490477209503, "grad_norm": 4.8792033195495605, "learning_rate": 6.58654903856511e-06, "loss": 1.645, "step": 3925 }, { "epoch": 1.6817890006419858, "grad_norm": 5.298194408416748, "learning_rate": 6.500855519098448e-06, "loss": 1.6642, "step": 3930 }, { "epoch": 1.6839289535630217, "grad_norm": 4.876554489135742, "learning_rate": 6.415684333993649e-06, "loss": 1.7346, "step": 3935 }, { "epoch": 1.6860689064840573, "grad_norm": 5.350904941558838, "learning_rate": 6.3310365059797094e-06, "loss": 1.7127, "step": 3940 }, { "epoch": 1.6882088594050932, "grad_norm": 5.282578945159912, "learning_rate": 6.246913051501202e-06, "loss": 1.6912, "step": 3945 }, { "epoch": 1.6903488123261288, "grad_norm": 5.434889793395996, "learning_rate": 6.163314980706058e-06, "loss": 1.6402, "step": 3950 }, { "epoch": 1.6924887652471645, "grad_norm": 5.1130571365356445, "learning_rate": 6.080243297433447e-06, "loss": 1.6301, "step": 3955 }, { "epoch": 1.6946287181682003, "grad_norm": 5.187545299530029, "learning_rate": 5.997698999201723e-06, "loss": 1.6238, "step": 3960 }, { "epoch": 1.696768671089236, "grad_norm": 4.982557773590088, "learning_rate": 5.915683077196415e-06, "loss": 1.6592, "step": 3965 }, { "epoch": 1.6989086240102718, "grad_norm": 4.824088096618652, "learning_rate": 5.834196516258378e-06, "loss": 1.6593, "step": 3970 }, { "epoch": 1.7010485769313075, "grad_norm": 4.828253746032715, "learning_rate": 5.753240294871937e-06, "loss": 1.6426, "step": 3975 }, { "epoch": 1.7031885298523433, "grad_norm": 5.136155605316162, "learning_rate": 5.6728153851531295e-06, "loss": 1.6446, "step": 3980 }, { "epoch": 1.705328482773379, "grad_norm": 5.751850128173828, "learning_rate": 5.592922752838053e-06, "loss": 1.6836, "step": 3985 }, { "epoch": 1.7074684356944148, "grad_norm": 5.147610664367676, "learning_rate": 5.513563357271256e-06, "loss": 1.6492, "step": 3990 }, { "epoch": 1.7096083886154505, "grad_norm": 5.698842525482178, "learning_rate": 5.43473815139422e-06, "loss": 1.643, "step": 3995 }, { "epoch": 1.7117483415364863, "grad_norm": 4.993765354156494, "learning_rate": 5.356448081733922e-06, "loss": 1.6808, "step": 4000 }, { "epoch": 1.7138882944575218, "grad_norm": 5.209753036499023, "learning_rate": 5.278694088391462e-06, "loss": 1.6976, "step": 4005 }, { "epoch": 1.7160282473785577, "grad_norm": 4.791884422302246, "learning_rate": 5.201477105030766e-06, "loss": 1.6394, "step": 4010 }, { "epoch": 1.7181682002995933, "grad_norm": 5.030315399169922, "learning_rate": 5.124798058867414e-06, "loss": 1.5982, "step": 4015 }, { "epoch": 1.7203081532206292, "grad_norm": 5.747689723968506, "learning_rate": 5.048657870657447e-06, "loss": 1.6619, "step": 4020 }, { "epoch": 1.7224481061416648, "grad_norm": 4.98396110534668, "learning_rate": 4.973057454686364e-06, "loss": 1.7076, "step": 4025 }, { "epoch": 1.7245880590627007, "grad_norm": 4.828854560852051, "learning_rate": 4.897997718758107e-06, "loss": 1.6687, "step": 4030 }, { "epoch": 1.7267280119837363, "grad_norm": 4.917362689971924, "learning_rate": 4.82347956418418e-06, "loss": 1.6901, "step": 4035 }, { "epoch": 1.7288679649047722, "grad_norm": 5.416051387786865, "learning_rate": 4.7495038857728155e-06, "loss": 1.6381, "step": 4040 }, { "epoch": 1.7310079178258078, "grad_norm": 5.832434177398682, "learning_rate": 4.676071571818236e-06, "loss": 1.6695, "step": 4045 }, { "epoch": 1.7331478707468437, "grad_norm": 5.534262657165527, "learning_rate": 4.603183504089997e-06, "loss": 1.7054, "step": 4050 }, { "epoch": 1.7352878236678793, "grad_norm": 5.344274997711182, "learning_rate": 4.5308405578223635e-06, "loss": 1.6645, "step": 4055 }, { "epoch": 1.737427776588915, "grad_norm": 5.08225679397583, "learning_rate": 4.45904360170385e-06, "loss": 1.6623, "step": 4060 }, { "epoch": 1.7395677295099508, "grad_norm": 5.531522274017334, "learning_rate": 4.387793497866744e-06, "loss": 1.6633, "step": 4065 }, { "epoch": 1.7417076824309865, "grad_norm": 4.792389869689941, "learning_rate": 4.31709110187678e-06, "loss": 1.6761, "step": 4070 }, { "epoch": 1.7438476353520223, "grad_norm": 5.555240631103516, "learning_rate": 4.246937262722866e-06, "loss": 1.6628, "step": 4075 }, { "epoch": 1.745987588273058, "grad_norm": 5.5509796142578125, "learning_rate": 4.177332822806873e-06, "loss": 1.6529, "step": 4080 }, { "epoch": 1.7481275411940937, "grad_norm": 4.76316499710083, "learning_rate": 4.108278617933525e-06, "loss": 1.6898, "step": 4085 }, { "epoch": 1.7502674941151295, "grad_norm": 4.821498870849609, "learning_rate": 4.039775477300378e-06, "loss": 1.6823, "step": 4090 }, { "epoch": 1.7524074470361652, "grad_norm": 5.2127299308776855, "learning_rate": 3.971824223487841e-06, "loss": 1.7028, "step": 4095 }, { "epoch": 1.754547399957201, "grad_norm": 4.9137701988220215, "learning_rate": 3.90442567244933e-06, "loss": 1.6858, "step": 4100 }, { "epoch": 1.7566873528782367, "grad_norm": 5.513342380523682, "learning_rate": 3.837580633501409e-06, "loss": 1.6466, "step": 4105 }, { "epoch": 1.7588273057992723, "grad_norm": 4.806028842926025, "learning_rate": 3.7712899093141407e-06, "loss": 1.7288, "step": 4110 }, { "epoch": 1.7609672587203082, "grad_norm": 5.0646796226501465, "learning_rate": 3.7055542959014087e-06, "loss": 1.6827, "step": 4115 }, { "epoch": 1.7631072116413438, "grad_norm": 5.064296245574951, "learning_rate": 3.640374582611389e-06, "loss": 1.6403, "step": 4120 }, { "epoch": 1.7652471645623797, "grad_norm": 6.094568252563477, "learning_rate": 3.575751552117029e-06, "loss": 1.625, "step": 4125 }, { "epoch": 1.7673871174834153, "grad_norm": 4.940478324890137, "learning_rate": 3.511685980406676e-06, "loss": 1.6847, "step": 4130 }, { "epoch": 1.7695270704044512, "grad_norm": 4.668292999267578, "learning_rate": 3.4481786367747627e-06, "loss": 1.6886, "step": 4135 }, { "epoch": 1.7716670233254868, "grad_norm": 5.22495174407959, "learning_rate": 3.3852302838125626e-06, "loss": 1.6492, "step": 4140 }, { "epoch": 1.7738069762465227, "grad_norm": 5.06997013092041, "learning_rate": 3.322841677399019e-06, "loss": 1.6441, "step": 4145 }, { "epoch": 1.7759469291675583, "grad_norm": 4.730724811553955, "learning_rate": 3.2610135666917007e-06, "loss": 1.6843, "step": 4150 }, { "epoch": 1.778086882088594, "grad_norm": 5.150337219238281, "learning_rate": 3.1997466941177666e-06, "loss": 1.6854, "step": 4155 }, { "epoch": 1.7802268350096297, "grad_norm": 4.789018630981445, "learning_rate": 3.139041795365094e-06, "loss": 1.634, "step": 4160 }, { "epoch": 1.7823667879306655, "grad_norm": 4.8880462646484375, "learning_rate": 3.0788995993734083e-06, "loss": 1.6258, "step": 4165 }, { "epoch": 1.7845067408517012, "grad_norm": 4.961301803588867, "learning_rate": 3.019320828325539e-06, "loss": 1.6463, "step": 4170 }, { "epoch": 1.786646693772737, "grad_norm": 4.813554286956787, "learning_rate": 2.9603061976387736e-06, "loss": 1.6903, "step": 4175 }, { "epoch": 1.7887866466937727, "grad_norm": 4.724278450012207, "learning_rate": 2.9018564159562224e-06, "loss": 1.646, "step": 4180 }, { "epoch": 1.7909265996148085, "grad_norm": 5.349635601043701, "learning_rate": 2.8439721851383383e-06, "loss": 1.6847, "step": 4185 }, { "epoch": 1.7930665525358442, "grad_norm": 4.801799297332764, "learning_rate": 2.786654200254496e-06, "loss": 1.6871, "step": 4190 }, { "epoch": 1.79520650545688, "grad_norm": 4.788339138031006, "learning_rate": 2.7299031495746252e-06, "loss": 1.7072, "step": 4195 }, { "epoch": 1.7973464583779157, "grad_norm": 4.700235843658447, "learning_rate": 2.6737197145609404e-06, "loss": 1.6723, "step": 4200 }, { "epoch": 1.7994864112989513, "grad_norm": 4.791804313659668, "learning_rate": 2.6181045698597972e-06, "loss": 1.674, "step": 4205 }, { "epoch": 1.8016263642199872, "grad_norm": 5.270465850830078, "learning_rate": 2.563058383293537e-06, "loss": 1.6434, "step": 4210 }, { "epoch": 1.8037663171410228, "grad_norm": 5.29375696182251, "learning_rate": 2.508581815852523e-06, "loss": 1.6719, "step": 4215 }, { "epoch": 1.8059062700620587, "grad_norm": 4.851658821105957, "learning_rate": 2.4546755216871496e-06, "loss": 1.6652, "step": 4220 }, { "epoch": 1.8080462229830943, "grad_norm": 5.476074695587158, "learning_rate": 2.40134014810004e-06, "loss": 1.7037, "step": 4225 }, { "epoch": 1.8101861759041302, "grad_norm": 5.279761791229248, "learning_rate": 2.3485763355382273e-06, "loss": 1.6697, "step": 4230 }, { "epoch": 1.8123261288251657, "grad_norm": 4.817234039306641, "learning_rate": 2.296384717585487e-06, "loss": 1.6609, "step": 4235 }, { "epoch": 1.8144660817462017, "grad_norm": 5.009280681610107, "learning_rate": 2.2447659209547443e-06, "loss": 1.6959, "step": 4240 }, { "epoch": 1.8166060346672372, "grad_norm": 5.085817813873291, "learning_rate": 2.1937205654805004e-06, "loss": 1.6254, "step": 4245 }, { "epoch": 1.8187459875882732, "grad_norm": 4.714953899383545, "learning_rate": 2.143249264111441e-06, "loss": 1.678, "step": 4250 }, { "epoch": 1.8208859405093087, "grad_norm": 5.392242908477783, "learning_rate": 2.09335262290305e-06, "loss": 1.6443, "step": 4255 }, { "epoch": 1.8230258934303445, "grad_norm": 4.728943347930908, "learning_rate": 2.04403124101033e-06, "loss": 1.6737, "step": 4260 }, { "epoch": 1.8251658463513802, "grad_norm": 4.887874603271484, "learning_rate": 1.995285710680622e-06, "loss": 1.6345, "step": 4265 }, { "epoch": 1.827305799272416, "grad_norm": 4.753354549407959, "learning_rate": 1.9471166172464917e-06, "loss": 1.7165, "step": 4270 }, { "epoch": 1.8294457521934517, "grad_norm": 5.000302314758301, "learning_rate": 1.8995245391186688e-06, "loss": 1.6339, "step": 4275 }, { "epoch": 1.8315857051144875, "grad_norm": 4.688614368438721, "learning_rate": 1.8525100477791602e-06, "loss": 1.6857, "step": 4280 }, { "epoch": 1.8337256580355232, "grad_norm": 4.900327205657959, "learning_rate": 1.8060737077743416e-06, "loss": 1.671, "step": 4285 }, { "epoch": 1.835865610956559, "grad_norm": 4.996348857879639, "learning_rate": 1.7602160767081822e-06, "loss": 1.6386, "step": 4290 }, { "epoch": 1.8380055638775947, "grad_norm": 4.658572673797607, "learning_rate": 1.7149377052355698e-06, "loss": 1.6028, "step": 4295 }, { "epoch": 1.8401455167986305, "grad_norm": 4.834623336791992, "learning_rate": 1.6702391370556957e-06, "loss": 1.6937, "step": 4300 }, { "epoch": 1.8422854697196662, "grad_norm": 4.5502028465271, "learning_rate": 1.6261209089054986e-06, "loss": 1.6596, "step": 4305 }, { "epoch": 1.8444254226407018, "grad_norm": 4.947803497314453, "learning_rate": 1.5825835505532516e-06, "loss": 1.6538, "step": 4310 }, { "epoch": 1.8465653755617377, "grad_norm": 5.3801774978637695, "learning_rate": 1.539627584792186e-06, "loss": 1.661, "step": 4315 }, { "epoch": 1.8487053284827732, "grad_norm": 5.104175090789795, "learning_rate": 1.4972535274342225e-06, "loss": 1.6419, "step": 4320 }, { "epoch": 1.8508452814038092, "grad_norm": 5.043144702911377, "learning_rate": 1.4554618873037551e-06, "loss": 1.6123, "step": 4325 }, { "epoch": 1.8529852343248447, "grad_norm": 4.843111038208008, "learning_rate": 1.4142531662315662e-06, "loss": 1.6482, "step": 4330 }, { "epoch": 1.8551251872458807, "grad_norm": 4.794622898101807, "learning_rate": 1.3736278590487927e-06, "loss": 1.7072, "step": 4335 }, { "epoch": 1.8572651401669162, "grad_norm": 4.912827491760254, "learning_rate": 1.3335864535809872e-06, "loss": 1.6744, "step": 4340 }, { "epoch": 1.8594050930879522, "grad_norm": 5.024925708770752, "learning_rate": 1.294129430642238e-06, "loss": 1.6379, "step": 4345 }, { "epoch": 1.8615450460089877, "grad_norm": 4.962955951690674, "learning_rate": 1.2552572640294247e-06, "loss": 1.6736, "step": 4350 }, { "epoch": 1.8636849989300235, "grad_norm": 4.8275275230407715, "learning_rate": 1.2169704205165277e-06, "loss": 1.6271, "step": 4355 }, { "epoch": 1.8658249518510592, "grad_norm": 4.821237087249756, "learning_rate": 1.179269359848989e-06, "loss": 1.6363, "step": 4360 }, { "epoch": 1.867964904772095, "grad_norm": 4.859674453735352, "learning_rate": 1.1421545347382378e-06, "loss": 1.6276, "step": 4365 }, { "epoch": 1.8701048576931307, "grad_norm": 4.546759605407715, "learning_rate": 1.105626390856218e-06, "loss": 1.6892, "step": 4370 }, { "epoch": 1.8722448106141665, "grad_norm": 4.87622594833374, "learning_rate": 1.0696853668300588e-06, "loss": 1.6633, "step": 4375 }, { "epoch": 1.8743847635352022, "grad_norm": 5.357635021209717, "learning_rate": 1.0343318942367951e-06, "loss": 1.6551, "step": 4380 }, { "epoch": 1.876524716456238, "grad_norm": 4.706364631652832, "learning_rate": 9.995663975981894e-07, "loss": 1.6708, "step": 4385 }, { "epoch": 1.8786646693772737, "grad_norm": 5.037484645843506, "learning_rate": 9.653892943756405e-07, "loss": 1.6359, "step": 4390 }, { "epoch": 1.8808046222983095, "grad_norm": 4.868020534515381, "learning_rate": 9.31800994965154e-07, "loss": 1.717, "step": 4395 }, { "epoch": 1.8829445752193452, "grad_norm": 5.029200077056885, "learning_rate": 8.98801902692431e-07, "loss": 1.6388, "step": 4400 }, { "epoch": 1.885084528140381, "grad_norm": 5.490426063537598, "learning_rate": 8.663924138080204e-07, "loss": 1.6471, "step": 4405 }, { "epoch": 1.8872244810614167, "grad_norm": 4.652840614318848, "learning_rate": 8.345729174825623e-07, "loss": 1.684, "step": 4410 }, { "epoch": 1.8893644339824522, "grad_norm": 4.7486252784729, "learning_rate": 8.033437958020973e-07, "loss": 1.701, "step": 4415 }, { "epoch": 1.8915043869034882, "grad_norm": 4.830610275268555, "learning_rate": 7.727054237635146e-07, "loss": 1.6273, "step": 4420 }, { "epoch": 1.8936443398245237, "grad_norm": 4.977087497711182, "learning_rate": 7.426581692700052e-07, "loss": 1.6923, "step": 4425 }, { "epoch": 1.8957842927455597, "grad_norm": 4.950675010681152, "learning_rate": 7.132023931266829e-07, "loss": 1.6456, "step": 4430 }, { "epoch": 1.8979242456665952, "grad_norm": 4.89157247543335, "learning_rate": 6.843384490362259e-07, "loss": 1.6838, "step": 4435 }, { "epoch": 1.9000641985876312, "grad_norm": 4.860950946807861, "learning_rate": 6.560666835946416e-07, "loss": 1.6309, "step": 4440 }, { "epoch": 1.9022041515086667, "grad_norm": 4.52632999420166, "learning_rate": 6.283874362871033e-07, "loss": 1.6225, "step": 4445 }, { "epoch": 1.9043441044297027, "grad_norm": 4.779524803161621, "learning_rate": 6.013010394838702e-07, "loss": 1.7016, "step": 4450 }, { "epoch": 1.9064840573507382, "grad_norm": 4.958658695220947, "learning_rate": 5.748078184363015e-07, "loss": 1.6794, "step": 4455 }, { "epoch": 1.908624010271774, "grad_norm": 5.345273017883301, "learning_rate": 5.489080912729428e-07, "loss": 1.6796, "step": 4460 }, { "epoch": 1.9107639631928097, "grad_norm": 5.009010314941406, "learning_rate": 5.236021689957237e-07, "loss": 1.6552, "step": 4465 }, { "epoch": 1.9129039161138455, "grad_norm": 4.576427936553955, "learning_rate": 4.988903554761948e-07, "loss": 1.64, "step": 4470 }, { "epoch": 1.9150438690348812, "grad_norm": 4.975499153137207, "learning_rate": 4.7477294745189584e-07, "loss": 1.6782, "step": 4475 }, { "epoch": 1.917183821955917, "grad_norm": 4.649482250213623, "learning_rate": 4.512502345228042e-07, "loss": 1.6728, "step": 4480 }, { "epoch": 1.9193237748769527, "grad_norm": 4.831498146057129, "learning_rate": 4.283224991478374e-07, "loss": 1.6622, "step": 4485 }, { "epoch": 1.9214637277979885, "grad_norm": 4.552825927734375, "learning_rate": 4.0599001664146097e-07, "loss": 1.6806, "step": 4490 }, { "epoch": 1.9236036807190242, "grad_norm": 4.854389190673828, "learning_rate": 3.842530551704027e-07, "loss": 1.5925, "step": 4495 }, { "epoch": 1.92574363364006, "grad_norm": 5.051529884338379, "learning_rate": 3.631118757504159e-07, "loss": 1.7039, "step": 4500 }, { "epoch": 1.9278835865610957, "grad_norm": 4.859341144561768, "learning_rate": 3.425667322431436e-07, "loss": 1.6972, "step": 4505 }, { "epoch": 1.9300235394821312, "grad_norm": 4.893909454345703, "learning_rate": 3.226178713530814e-07, "loss": 1.635, "step": 4510 }, { "epoch": 1.9321634924031672, "grad_norm": 4.657041549682617, "learning_rate": 3.0326553262460255e-07, "loss": 1.6971, "step": 4515 }, { "epoch": 1.9343034453242027, "grad_norm": 4.919985294342041, "learning_rate": 2.845099484391045e-07, "loss": 1.6954, "step": 4520 }, { "epoch": 1.9364433982452387, "grad_norm": 5.278502464294434, "learning_rate": 2.663513440121834e-07, "loss": 1.6607, "step": 4525 }, { "epoch": 1.9385833511662742, "grad_norm": 4.634570598602295, "learning_rate": 2.4878993739095857e-07, "loss": 1.6459, "step": 4530 }, { "epoch": 1.9407233040873102, "grad_norm": 4.695330619812012, "learning_rate": 2.3182593945144105e-07, "loss": 1.6391, "step": 4535 }, { "epoch": 1.9428632570083457, "grad_norm": 4.747430324554443, "learning_rate": 2.1545955389600248e-07, "loss": 1.6485, "step": 4540 }, { "epoch": 1.9450032099293817, "grad_norm": 5.009528636932373, "learning_rate": 1.9969097725094366e-07, "loss": 1.686, "step": 4545 }, { "epoch": 1.9471431628504172, "grad_norm": 4.869572639465332, "learning_rate": 1.8452039886410199e-07, "loss": 1.6684, "step": 4550 }, { "epoch": 1.9492831157714532, "grad_norm": 4.405606269836426, "learning_rate": 1.6994800090261997e-07, "loss": 1.6607, "step": 4555 }, { "epoch": 1.9514230686924887, "grad_norm": 4.575326919555664, "learning_rate": 1.5597395835071915e-07, "loss": 1.6839, "step": 4560 }, { "epoch": 1.9535630216135245, "grad_norm": 4.905466079711914, "learning_rate": 1.425984390076185e-07, "loss": 1.6075, "step": 4565 }, { "epoch": 1.9557029745345602, "grad_norm": 4.696757793426514, "learning_rate": 1.298216034855304e-07, "loss": 1.5987, "step": 4570 }, { "epoch": 1.957842927455596, "grad_norm": 5.084411144256592, "learning_rate": 1.1764360520769568e-07, "loss": 1.6828, "step": 4575 }, { "epoch": 1.9599828803766317, "grad_norm": 4.676747798919678, "learning_rate": 1.0606459040657935e-07, "loss": 1.7003, "step": 4580 }, { "epoch": 1.9621228332976675, "grad_norm": 4.811861991882324, "learning_rate": 9.508469812209986e-08, "loss": 1.6344, "step": 4585 }, { "epoch": 1.9642627862187032, "grad_norm": 4.689334392547607, "learning_rate": 8.470406019994714e-08, "loss": 1.673, "step": 4590 }, { "epoch": 1.966402739139739, "grad_norm": 4.811148166656494, "learning_rate": 7.492280129002271e-08, "loss": 1.6674, "step": 4595 }, { "epoch": 1.9685426920607747, "grad_norm": 4.551917552947998, "learning_rate": 6.574103884492422e-08, "loss": 1.6488, "step": 4600 }, { "epoch": 1.9706826449818104, "grad_norm": 4.65984582901001, "learning_rate": 5.715888311855211e-08, "loss": 1.6826, "step": 4605 }, { "epoch": 1.9728225979028462, "grad_norm": 4.9454240798950195, "learning_rate": 4.9176437164760726e-08, "loss": 1.6969, "step": 4610 }, { "epoch": 1.9749625508238817, "grad_norm": 4.670926570892334, "learning_rate": 4.1793796836142596e-08, "loss": 1.6635, "step": 4615 }, { "epoch": 1.9771025037449177, "grad_norm": 4.9202399253845215, "learning_rate": 3.5011050782879364e-08, "loss": 1.6676, "step": 4620 }, { "epoch": 1.9792424566659532, "grad_norm": 4.523751258850098, "learning_rate": 2.8828280451653755e-08, "loss": 1.6235, "step": 4625 }, { "epoch": 1.9813824095869892, "grad_norm": 4.875797748565674, "learning_rate": 2.3245560084700357e-08, "loss": 1.7377, "step": 4630 }, { "epoch": 1.9835223625080247, "grad_norm": 4.714092254638672, "learning_rate": 1.8262956718884117e-08, "loss": 1.6771, "step": 4635 }, { "epoch": 1.9856623154290607, "grad_norm": 4.643798351287842, "learning_rate": 1.3880530184934293e-08, "loss": 1.6668, "step": 4640 }, { "epoch": 1.9878022683500962, "grad_norm": 4.464611530303955, "learning_rate": 1.0098333106672852e-08, "loss": 1.6362, "step": 4645 }, { "epoch": 1.9899422212711322, "grad_norm": 4.848086357116699, "learning_rate": 6.9164109004427046e-09, "loss": 1.6405, "step": 4650 }, { "epoch": 1.9920821741921677, "grad_norm": 4.318144798278809, "learning_rate": 4.33480177451373e-09, "loss": 1.6798, "step": 4655 }, { "epoch": 1.9942221271132035, "grad_norm": 5.152481555938721, "learning_rate": 2.3535367286497966e-09, "loss": 1.6778, "step": 4660 }, { "epoch": 1.9963620800342392, "grad_norm": 4.668585777282715, "learning_rate": 9.726395537312806e-10, "loss": 1.6754, "step": 4665 }, { "epoch": 1.998502032955275, "grad_norm": 4.8391313552856445, "learning_rate": 1.921268314608593e-10, "loss": 1.6921, "step": 4670 }, { "epoch": 2.0, "step": 4674, "total_flos": 7.23540271133465e+18, "train_loss": 2.1290212889537883, "train_runtime": 7155.5853, "train_samples_per_second": 104.472, "train_steps_per_second": 0.653 } ], "logging_steps": 5, "max_steps": 4674, "num_input_tokens_seen": 0, "num_train_epochs": 2, "save_steps": 468, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 7.23540271133465e+18, "train_batch_size": 2, "trial_name": null, "trial_params": null }