diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,7730 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.9958960328317374, + "eval_steps": 250, + "global_step": 1095, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0027359781121751026, + "grad_norm": 0.24075667560100555, + "learning_rate": 2.2727272727272728e-06, + "loss": 1.9078, + "step": 1 + }, + { + "epoch": 0.005471956224350205, + "grad_norm": 0.2584983706474304, + "learning_rate": 4.5454545454545455e-06, + "loss": 2.1152, + "step": 2 + }, + { + "epoch": 0.008207934336525308, + "grad_norm": 0.24978317320346832, + "learning_rate": 6.818181818181818e-06, + "loss": 1.9649, + "step": 3 + }, + { + "epoch": 0.01094391244870041, + "grad_norm": 0.2477930635213852, + "learning_rate": 9.090909090909091e-06, + "loss": 1.9163, + "step": 4 + }, + { + "epoch": 0.013679890560875513, + "grad_norm": 0.2425757646560669, + "learning_rate": 1.1363636363636365e-05, + "loss": 2.1254, + "step": 5 + }, + { + "epoch": 0.016415868673050615, + "grad_norm": 0.251596599817276, + "learning_rate": 1.3636363636363637e-05, + "loss": 2.0832, + "step": 6 + }, + { + "epoch": 0.019151846785225718, + "grad_norm": 0.2359585464000702, + "learning_rate": 1.590909090909091e-05, + "loss": 1.8204, + "step": 7 + }, + { + "epoch": 0.02188782489740082, + "grad_norm": 0.25108030438423157, + "learning_rate": 1.8181818181818182e-05, + "loss": 1.9224, + "step": 8 + }, + { + "epoch": 0.024623803009575923, + "grad_norm": 0.25240400433540344, + "learning_rate": 2.0454545454545457e-05, + "loss": 1.8371, + "step": 9 + }, + { + "epoch": 0.027359781121751026, + "grad_norm": 0.22698192298412323, + "learning_rate": 2.272727272727273e-05, + "loss": 1.9962, + "step": 10 + }, + { + "epoch": 0.030095759233926128, + "grad_norm": 0.24213118851184845, + "learning_rate": 2.5e-05, + "loss": 1.9918, + "step": 11 + }, + { + "epoch": 0.03283173734610123, + "grad_norm": 0.20904198288917542, + "learning_rate": 2.7272727272727273e-05, + "loss": 1.9463, + "step": 12 + }, + { + "epoch": 0.03556771545827633, + "grad_norm": 0.21374110877513885, + "learning_rate": 2.954545454545455e-05, + "loss": 1.8628, + "step": 13 + }, + { + "epoch": 0.038303693570451436, + "grad_norm": 0.27983948588371277, + "learning_rate": 3.181818181818182e-05, + "loss": 2.1597, + "step": 14 + }, + { + "epoch": 0.04103967168262654, + "grad_norm": 0.2309175282716751, + "learning_rate": 3.409090909090909e-05, + "loss": 2.0305, + "step": 15 + }, + { + "epoch": 0.04377564979480164, + "grad_norm": 0.21376170217990875, + "learning_rate": 3.6363636363636364e-05, + "loss": 1.8866, + "step": 16 + }, + { + "epoch": 0.046511627906976744, + "grad_norm": 0.23382064700126648, + "learning_rate": 3.8636363636363636e-05, + "loss": 1.9912, + "step": 17 + }, + { + "epoch": 0.049247606019151846, + "grad_norm": 0.20445913076400757, + "learning_rate": 4.0909090909090915e-05, + "loss": 1.9773, + "step": 18 + }, + { + "epoch": 0.05198358413132695, + "grad_norm": 0.21313798427581787, + "learning_rate": 4.318181818181819e-05, + "loss": 1.9965, + "step": 19 + }, + { + "epoch": 0.05471956224350205, + "grad_norm": 0.22150328755378723, + "learning_rate": 4.545454545454546e-05, + "loss": 1.9538, + "step": 20 + }, + { + "epoch": 0.057455540355677154, + "grad_norm": 0.17602485418319702, + "learning_rate": 4.772727272727273e-05, + "loss": 1.9113, + "step": 21 + }, + { + "epoch": 0.060191518467852256, + "grad_norm": 0.20136825740337372, + "learning_rate": 5e-05, + "loss": 1.9837, + "step": 22 + }, + { + "epoch": 0.06292749658002736, + "grad_norm": 0.1977069228887558, + "learning_rate": 4.999989284560115e-05, + "loss": 1.8121, + "step": 23 + }, + { + "epoch": 0.06566347469220246, + "grad_norm": 0.18849150836467743, + "learning_rate": 4.9999571383323136e-05, + "loss": 1.8013, + "step": 24 + }, + { + "epoch": 0.06839945280437756, + "grad_norm": 0.17495211958885193, + "learning_rate": 4.9999035615921664e-05, + "loss": 1.8375, + "step": 25 + }, + { + "epoch": 0.07113543091655267, + "grad_norm": 0.28524237871170044, + "learning_rate": 4.999828554798952e-05, + "loss": 2.026, + "step": 26 + }, + { + "epoch": 0.07387140902872777, + "grad_norm": 0.2135252058506012, + "learning_rate": 4.999732118595654e-05, + "loss": 1.7874, + "step": 27 + }, + { + "epoch": 0.07660738714090287, + "grad_norm": 0.213677778840065, + "learning_rate": 4.999614253808959e-05, + "loss": 1.7844, + "step": 28 + }, + { + "epoch": 0.07934336525307797, + "grad_norm": 0.20930789411067963, + "learning_rate": 4.999474961449243e-05, + "loss": 1.857, + "step": 29 + }, + { + "epoch": 0.08207934336525308, + "grad_norm": 0.18232876062393188, + "learning_rate": 4.999314242710572e-05, + "loss": 1.7145, + "step": 30 + }, + { + "epoch": 0.08481532147742818, + "grad_norm": 0.2061523050069809, + "learning_rate": 4.9991320989706816e-05, + "loss": 1.8137, + "step": 31 + }, + { + "epoch": 0.08755129958960328, + "grad_norm": 0.19810044765472412, + "learning_rate": 4.9989285317909725e-05, + "loss": 1.8154, + "step": 32 + }, + { + "epoch": 0.09028727770177838, + "grad_norm": 0.19258971512317657, + "learning_rate": 4.9987035429164954e-05, + "loss": 1.8612, + "step": 33 + }, + { + "epoch": 0.09302325581395349, + "grad_norm": 0.21273574233055115, + "learning_rate": 4.9984571342759326e-05, + "loss": 1.6282, + "step": 34 + }, + { + "epoch": 0.09575923392612859, + "grad_norm": 0.18505199253559113, + "learning_rate": 4.998189307981586e-05, + "loss": 1.7134, + "step": 35 + }, + { + "epoch": 0.09849521203830369, + "grad_norm": 0.20874658226966858, + "learning_rate": 4.997900066329357e-05, + "loss": 1.7648, + "step": 36 + }, + { + "epoch": 0.1012311901504788, + "grad_norm": 0.18423666059970856, + "learning_rate": 4.9975894117987265e-05, + "loss": 1.7784, + "step": 37 + }, + { + "epoch": 0.1039671682626539, + "grad_norm": 0.16870878636837006, + "learning_rate": 4.9972573470527354e-05, + "loss": 1.7352, + "step": 38 + }, + { + "epoch": 0.106703146374829, + "grad_norm": 0.20023159682750702, + "learning_rate": 4.996903874937959e-05, + "loss": 1.7784, + "step": 39 + }, + { + "epoch": 0.1094391244870041, + "grad_norm": 0.22696524858474731, + "learning_rate": 4.9965289984844846e-05, + "loss": 1.7722, + "step": 40 + }, + { + "epoch": 0.1121751025991792, + "grad_norm": 0.18610844016075134, + "learning_rate": 4.9961327209058844e-05, + "loss": 1.761, + "step": 41 + }, + { + "epoch": 0.11491108071135431, + "grad_norm": 0.17688068747520447, + "learning_rate": 4.99571504559919e-05, + "loss": 1.6316, + "step": 42 + }, + { + "epoch": 0.11764705882352941, + "grad_norm": 0.20090024173259735, + "learning_rate": 4.995275976144862e-05, + "loss": 1.8064, + "step": 43 + }, + { + "epoch": 0.12038303693570451, + "grad_norm": 0.17762398719787598, + "learning_rate": 4.9948155163067565e-05, + "loss": 1.8417, + "step": 44 + }, + { + "epoch": 0.12311901504787962, + "grad_norm": 0.1874111294746399, + "learning_rate": 4.9943336700320985e-05, + "loss": 1.7245, + "step": 45 + }, + { + "epoch": 0.12585499316005472, + "grad_norm": 0.16139264404773712, + "learning_rate": 4.993830441451444e-05, + "loss": 1.8852, + "step": 46 + }, + { + "epoch": 0.12859097127222982, + "grad_norm": 0.16358217597007751, + "learning_rate": 4.993305834878643e-05, + "loss": 1.7314, + "step": 47 + }, + { + "epoch": 0.13132694938440492, + "grad_norm": 0.16465303301811218, + "learning_rate": 4.992759854810811e-05, + "loss": 1.6362, + "step": 48 + }, + { + "epoch": 0.13406292749658003, + "grad_norm": 0.1597519814968109, + "learning_rate": 4.99219250592828e-05, + "loss": 1.5936, + "step": 49 + }, + { + "epoch": 0.13679890560875513, + "grad_norm": 0.16433416306972504, + "learning_rate": 4.991603793094564e-05, + "loss": 1.5236, + "step": 50 + }, + { + "epoch": 0.13953488372093023, + "grad_norm": 0.16490432620048523, + "learning_rate": 4.9909937213563165e-05, + "loss": 1.7191, + "step": 51 + }, + { + "epoch": 0.14227086183310533, + "grad_norm": 0.1576634645462036, + "learning_rate": 4.990362295943287e-05, + "loss": 1.5944, + "step": 52 + }, + { + "epoch": 0.14500683994528044, + "grad_norm": 0.16687043011188507, + "learning_rate": 4.989709522268278e-05, + "loss": 1.7204, + "step": 53 + }, + { + "epoch": 0.14774281805745554, + "grad_norm": 0.18680374324321747, + "learning_rate": 4.989035405927093e-05, + "loss": 1.8231, + "step": 54 + }, + { + "epoch": 0.15047879616963064, + "grad_norm": 0.16919435560703278, + "learning_rate": 4.9883399526984944e-05, + "loss": 1.7527, + "step": 55 + }, + { + "epoch": 0.15321477428180574, + "grad_norm": 0.16811755299568176, + "learning_rate": 4.9876231685441545e-05, + "loss": 1.6824, + "step": 56 + }, + { + "epoch": 0.15595075239398085, + "grad_norm": 0.15195395052433014, + "learning_rate": 4.986885059608597e-05, + "loss": 1.6515, + "step": 57 + }, + { + "epoch": 0.15868673050615595, + "grad_norm": 0.17854413390159607, + "learning_rate": 4.986125632219152e-05, + "loss": 1.613, + "step": 58 + }, + { + "epoch": 0.16142270861833105, + "grad_norm": 0.168907031416893, + "learning_rate": 4.985344892885899e-05, + "loss": 1.798, + "step": 59 + }, + { + "epoch": 0.16415868673050615, + "grad_norm": 0.173956036567688, + "learning_rate": 4.98454284830161e-05, + "loss": 1.4941, + "step": 60 + }, + { + "epoch": 0.16689466484268126, + "grad_norm": 0.19879266619682312, + "learning_rate": 4.983719505341693e-05, + "loss": 1.7941, + "step": 61 + }, + { + "epoch": 0.16963064295485636, + "grad_norm": 0.16308054327964783, + "learning_rate": 4.982874871064134e-05, + "loss": 1.6984, + "step": 62 + }, + { + "epoch": 0.17236662106703146, + "grad_norm": 0.16789625585079193, + "learning_rate": 4.982008952709435e-05, + "loss": 1.6375, + "step": 63 + }, + { + "epoch": 0.17510259917920656, + "grad_norm": 0.16414505243301392, + "learning_rate": 4.9811217577005533e-05, + "loss": 1.5667, + "step": 64 + }, + { + "epoch": 0.17783857729138167, + "grad_norm": 0.16586610674858093, + "learning_rate": 4.980213293642837e-05, + "loss": 1.6688, + "step": 65 + }, + { + "epoch": 0.18057455540355677, + "grad_norm": 0.21007002890110016, + "learning_rate": 4.979283568323958e-05, + "loss": 1.7249, + "step": 66 + }, + { + "epoch": 0.18331053351573187, + "grad_norm": 0.15806783735752106, + "learning_rate": 4.97833258971385e-05, + "loss": 1.6918, + "step": 67 + }, + { + "epoch": 0.18604651162790697, + "grad_norm": 0.17978119850158691, + "learning_rate": 4.977360365964637e-05, + "loss": 1.6815, + "step": 68 + }, + { + "epoch": 0.18878248974008208, + "grad_norm": 0.16976381838321686, + "learning_rate": 4.976366905410562e-05, + "loss": 1.7212, + "step": 69 + }, + { + "epoch": 0.19151846785225718, + "grad_norm": 0.18810401856899261, + "learning_rate": 4.9753522165679195e-05, + "loss": 1.707, + "step": 70 + }, + { + "epoch": 0.19425444596443228, + "grad_norm": 0.17066176235675812, + "learning_rate": 4.974316308134978e-05, + "loss": 1.7765, + "step": 71 + }, + { + "epoch": 0.19699042407660738, + "grad_norm": 0.17143549025058746, + "learning_rate": 4.9732591889919114e-05, + "loss": 1.7317, + "step": 72 + }, + { + "epoch": 0.1997264021887825, + "grad_norm": 0.15364663302898407, + "learning_rate": 4.972180868200714e-05, + "loss": 1.6406, + "step": 73 + }, + { + "epoch": 0.2024623803009576, + "grad_norm": 0.1750328093767166, + "learning_rate": 4.971081355005133e-05, + "loss": 1.8737, + "step": 74 + }, + { + "epoch": 0.2051983584131327, + "grad_norm": 0.17950010299682617, + "learning_rate": 4.969960658830582e-05, + "loss": 1.7941, + "step": 75 + }, + { + "epoch": 0.2079343365253078, + "grad_norm": 0.1876641809940338, + "learning_rate": 4.968818789284063e-05, + "loss": 1.5792, + "step": 76 + }, + { + "epoch": 0.2106703146374829, + "grad_norm": 0.17048972845077515, + "learning_rate": 4.9676557561540845e-05, + "loss": 1.607, + "step": 77 + }, + { + "epoch": 0.213406292749658, + "grad_norm": 0.18965713679790497, + "learning_rate": 4.966471569410574e-05, + "loss": 1.6142, + "step": 78 + }, + { + "epoch": 0.2161422708618331, + "grad_norm": 0.16774089634418488, + "learning_rate": 4.965266239204798e-05, + "loss": 1.6446, + "step": 79 + }, + { + "epoch": 0.2188782489740082, + "grad_norm": 0.16337555646896362, + "learning_rate": 4.9640397758692715e-05, + "loss": 1.6319, + "step": 80 + }, + { + "epoch": 0.2216142270861833, + "grad_norm": 0.17532555758953094, + "learning_rate": 4.9627921899176686e-05, + "loss": 1.554, + "step": 81 + }, + { + "epoch": 0.2243502051983584, + "grad_norm": 0.16369390487670898, + "learning_rate": 4.9615234920447365e-05, + "loss": 1.7499, + "step": 82 + }, + { + "epoch": 0.2270861833105335, + "grad_norm": 0.170010045170784, + "learning_rate": 4.9602336931261995e-05, + "loss": 1.7786, + "step": 83 + }, + { + "epoch": 0.22982216142270862, + "grad_norm": 0.18399612605571747, + "learning_rate": 4.958922804218667e-05, + "loss": 1.8338, + "step": 84 + }, + { + "epoch": 0.23255813953488372, + "grad_norm": 0.16553553938865662, + "learning_rate": 4.957590836559541e-05, + "loss": 1.8186, + "step": 85 + }, + { + "epoch": 0.23529411764705882, + "grad_norm": 0.17777277529239655, + "learning_rate": 4.9562378015669166e-05, + "loss": 1.7326, + "step": 86 + }, + { + "epoch": 0.23803009575923392, + "grad_norm": 0.18170584738254547, + "learning_rate": 4.954863710839485e-05, + "loss": 1.7293, + "step": 87 + }, + { + "epoch": 0.24076607387140903, + "grad_norm": 0.17840476334095, + "learning_rate": 4.953468576156437e-05, + "loss": 1.6941, + "step": 88 + }, + { + "epoch": 0.24350205198358413, + "grad_norm": 0.16968044638633728, + "learning_rate": 4.9520524094773575e-05, + "loss": 1.7643, + "step": 89 + }, + { + "epoch": 0.24623803009575923, + "grad_norm": 0.17756614089012146, + "learning_rate": 4.950615222942125e-05, + "loss": 1.7599, + "step": 90 + }, + { + "epoch": 0.24897400820793433, + "grad_norm": 0.162063866853714, + "learning_rate": 4.94915702887081e-05, + "loss": 1.7365, + "step": 91 + }, + { + "epoch": 0.25170998632010944, + "grad_norm": 0.18559984862804413, + "learning_rate": 4.947677839763563e-05, + "loss": 1.4424, + "step": 92 + }, + { + "epoch": 0.25444596443228457, + "grad_norm": 0.17151783406734467, + "learning_rate": 4.946177668300515e-05, + "loss": 1.5231, + "step": 93 + }, + { + "epoch": 0.25718194254445964, + "grad_norm": 0.20623093843460083, + "learning_rate": 4.944656527341663e-05, + "loss": 1.8024, + "step": 94 + }, + { + "epoch": 0.25991792065663477, + "grad_norm": 0.19318287074565887, + "learning_rate": 4.943114429926763e-05, + "loss": 1.6366, + "step": 95 + }, + { + "epoch": 0.26265389876880985, + "grad_norm": 0.1877240240573883, + "learning_rate": 4.941551389275217e-05, + "loss": 1.5424, + "step": 96 + }, + { + "epoch": 0.265389876880985, + "grad_norm": 0.17889100313186646, + "learning_rate": 4.939967418785959e-05, + "loss": 1.6603, + "step": 97 + }, + { + "epoch": 0.26812585499316005, + "grad_norm": 0.18533948063850403, + "learning_rate": 4.938362532037341e-05, + "loss": 1.7546, + "step": 98 + }, + { + "epoch": 0.2708618331053352, + "grad_norm": 0.19442051649093628, + "learning_rate": 4.936736742787018e-05, + "loss": 1.6938, + "step": 99 + }, + { + "epoch": 0.27359781121751026, + "grad_norm": 0.20610813796520233, + "learning_rate": 4.935090064971827e-05, + "loss": 1.7104, + "step": 100 + }, + { + "epoch": 0.2763337893296854, + "grad_norm": 0.19741714000701904, + "learning_rate": 4.933422512707671e-05, + "loss": 1.6205, + "step": 101 + }, + { + "epoch": 0.27906976744186046, + "grad_norm": 0.20613451302051544, + "learning_rate": 4.931734100289393e-05, + "loss": 1.7549, + "step": 102 + }, + { + "epoch": 0.2818057455540356, + "grad_norm": 0.19265608489513397, + "learning_rate": 4.930024842190658e-05, + "loss": 1.8493, + "step": 103 + }, + { + "epoch": 0.28454172366621067, + "grad_norm": 0.1976325511932373, + "learning_rate": 4.9282947530638295e-05, + "loss": 1.8059, + "step": 104 + }, + { + "epoch": 0.2872777017783858, + "grad_norm": 0.20444822311401367, + "learning_rate": 4.926543847739841e-05, + "loss": 1.5852, + "step": 105 + }, + { + "epoch": 0.29001367989056087, + "grad_norm": 0.18967504799365997, + "learning_rate": 4.9247721412280667e-05, + "loss": 1.6282, + "step": 106 + }, + { + "epoch": 0.292749658002736, + "grad_norm": 0.206316277384758, + "learning_rate": 4.9229796487162e-05, + "loss": 1.6186, + "step": 107 + }, + { + "epoch": 0.2954856361149111, + "grad_norm": 0.17524588108062744, + "learning_rate": 4.9211663855701165e-05, + "loss": 1.6383, + "step": 108 + }, + { + "epoch": 0.2982216142270862, + "grad_norm": 0.2032768577337265, + "learning_rate": 4.9193323673337476e-05, + "loss": 1.785, + "step": 109 + }, + { + "epoch": 0.3009575923392613, + "grad_norm": 0.17396964132785797, + "learning_rate": 4.917477609728941e-05, + "loss": 1.4461, + "step": 110 + }, + { + "epoch": 0.3036935704514364, + "grad_norm": 0.19011397659778595, + "learning_rate": 4.915602128655333e-05, + "loss": 1.7313, + "step": 111 + }, + { + "epoch": 0.3064295485636115, + "grad_norm": 0.19239680469036102, + "learning_rate": 4.913705940190207e-05, + "loss": 1.6015, + "step": 112 + }, + { + "epoch": 0.3091655266757866, + "grad_norm": 0.19917477667331696, + "learning_rate": 4.9117890605883574e-05, + "loss": 1.6411, + "step": 113 + }, + { + "epoch": 0.3119015047879617, + "grad_norm": 0.19318480789661407, + "learning_rate": 4.909851506281952e-05, + "loss": 1.6737, + "step": 114 + }, + { + "epoch": 0.3146374829001368, + "grad_norm": 0.1994866579771042, + "learning_rate": 4.907893293880387e-05, + "loss": 1.518, + "step": 115 + }, + { + "epoch": 0.3173734610123119, + "grad_norm": 0.1941988468170166, + "learning_rate": 4.905914440170147e-05, + "loss": 1.7589, + "step": 116 + }, + { + "epoch": 0.320109439124487, + "grad_norm": 0.20112977921962738, + "learning_rate": 4.903914962114665e-05, + "loss": 1.6671, + "step": 117 + }, + { + "epoch": 0.3228454172366621, + "grad_norm": 0.21141856908798218, + "learning_rate": 4.9018948768541694e-05, + "loss": 1.6731, + "step": 118 + }, + { + "epoch": 0.32558139534883723, + "grad_norm": 0.18585005402565002, + "learning_rate": 4.899854201705542e-05, + "loss": 1.6276, + "step": 119 + }, + { + "epoch": 0.3283173734610123, + "grad_norm": 0.1975102424621582, + "learning_rate": 4.8977929541621676e-05, + "loss": 1.6103, + "step": 120 + }, + { + "epoch": 0.33105335157318744, + "grad_norm": 0.19365696609020233, + "learning_rate": 4.895711151893786e-05, + "loss": 1.5401, + "step": 121 + }, + { + "epoch": 0.3337893296853625, + "grad_norm": 0.20823922753334045, + "learning_rate": 4.8936088127463396e-05, + "loss": 1.5811, + "step": 122 + }, + { + "epoch": 0.33652530779753764, + "grad_norm": 0.2523277997970581, + "learning_rate": 4.8914859547418186e-05, + "loss": 1.5561, + "step": 123 + }, + { + "epoch": 0.3392612859097127, + "grad_norm": 0.19290074706077576, + "learning_rate": 4.889342596078109e-05, + "loss": 1.6078, + "step": 124 + }, + { + "epoch": 0.34199726402188785, + "grad_norm": 0.17994068562984467, + "learning_rate": 4.887178755128835e-05, + "loss": 1.7232, + "step": 125 + }, + { + "epoch": 0.3447332421340629, + "grad_norm": 0.18888920545578003, + "learning_rate": 4.8849944504432035e-05, + "loss": 1.5538, + "step": 126 + }, + { + "epoch": 0.34746922024623805, + "grad_norm": 0.18798014521598816, + "learning_rate": 4.882789700745843e-05, + "loss": 1.6081, + "step": 127 + }, + { + "epoch": 0.35020519835841313, + "grad_norm": 0.19080984592437744, + "learning_rate": 4.880564524936643e-05, + "loss": 1.4809, + "step": 128 + }, + { + "epoch": 0.35294117647058826, + "grad_norm": 0.19421504437923431, + "learning_rate": 4.878318942090594e-05, + "loss": 1.6523, + "step": 129 + }, + { + "epoch": 0.35567715458276333, + "grad_norm": 0.1982857584953308, + "learning_rate": 4.876052971457623e-05, + "loss": 1.6948, + "step": 130 + }, + { + "epoch": 0.35841313269493846, + "grad_norm": 0.20162495970726013, + "learning_rate": 4.8737666324624265e-05, + "loss": 1.7504, + "step": 131 + }, + { + "epoch": 0.36114911080711354, + "grad_norm": 0.20931164920330048, + "learning_rate": 4.8714599447043085e-05, + "loss": 1.4829, + "step": 132 + }, + { + "epoch": 0.36388508891928867, + "grad_norm": 0.21997657418251038, + "learning_rate": 4.869132927957007e-05, + "loss": 1.5546, + "step": 133 + }, + { + "epoch": 0.36662106703146374, + "grad_norm": 0.20648987591266632, + "learning_rate": 4.866785602168528e-05, + "loss": 1.6716, + "step": 134 + }, + { + "epoch": 0.3693570451436389, + "grad_norm": 0.20821620523929596, + "learning_rate": 4.864417987460975e-05, + "loss": 1.6197, + "step": 135 + }, + { + "epoch": 0.37209302325581395, + "grad_norm": 0.21084287762641907, + "learning_rate": 4.862030104130375e-05, + "loss": 1.6454, + "step": 136 + }, + { + "epoch": 0.3748290013679891, + "grad_norm": 0.21337808668613434, + "learning_rate": 4.859621972646503e-05, + "loss": 1.5733, + "step": 137 + }, + { + "epoch": 0.37756497948016415, + "grad_norm": 0.21093016862869263, + "learning_rate": 4.857193613652711e-05, + "loss": 1.6716, + "step": 138 + }, + { + "epoch": 0.3803009575923393, + "grad_norm": 0.20863549411296844, + "learning_rate": 4.854745047965745e-05, + "loss": 1.6117, + "step": 139 + }, + { + "epoch": 0.38303693570451436, + "grad_norm": 0.20271432399749756, + "learning_rate": 4.852276296575573e-05, + "loss": 1.5765, + "step": 140 + }, + { + "epoch": 0.3857729138166895, + "grad_norm": 0.20297355949878693, + "learning_rate": 4.849787380645201e-05, + "loss": 1.7296, + "step": 141 + }, + { + "epoch": 0.38850889192886456, + "grad_norm": 0.20244932174682617, + "learning_rate": 4.847278321510491e-05, + "loss": 1.7651, + "step": 142 + }, + { + "epoch": 0.3912448700410397, + "grad_norm": 0.20252294838428497, + "learning_rate": 4.844749140679982e-05, + "loss": 1.6427, + "step": 143 + }, + { + "epoch": 0.39398084815321477, + "grad_norm": 0.20163066685199738, + "learning_rate": 4.842199859834702e-05, + "loss": 1.6398, + "step": 144 + }, + { + "epoch": 0.3967168262653899, + "grad_norm": 0.21187250316143036, + "learning_rate": 4.839630500827982e-05, + "loss": 1.6173, + "step": 145 + }, + { + "epoch": 0.399452804377565, + "grad_norm": 0.18655481934547424, + "learning_rate": 4.837041085685273e-05, + "loss": 1.5184, + "step": 146 + }, + { + "epoch": 0.4021887824897401, + "grad_norm": 0.21056880056858063, + "learning_rate": 4.834431636603953e-05, + "loss": 1.5687, + "step": 147 + }, + { + "epoch": 0.4049247606019152, + "grad_norm": 0.2020365595817566, + "learning_rate": 4.831802175953138e-05, + "loss": 1.6445, + "step": 148 + }, + { + "epoch": 0.4076607387140903, + "grad_norm": 0.18952035903930664, + "learning_rate": 4.8291527262734886e-05, + "loss": 1.7945, + "step": 149 + }, + { + "epoch": 0.4103967168262654, + "grad_norm": 0.1896515041589737, + "learning_rate": 4.826483310277021e-05, + "loss": 1.546, + "step": 150 + }, + { + "epoch": 0.4131326949384405, + "grad_norm": 0.20328114926815033, + "learning_rate": 4.8237939508469075e-05, + "loss": 1.5621, + "step": 151 + }, + { + "epoch": 0.4158686730506156, + "grad_norm": 0.18790675699710846, + "learning_rate": 4.821084671037285e-05, + "loss": 1.6609, + "step": 152 + }, + { + "epoch": 0.4186046511627907, + "grad_norm": 0.20515787601470947, + "learning_rate": 4.8183554940730524e-05, + "loss": 1.671, + "step": 153 + }, + { + "epoch": 0.4213406292749658, + "grad_norm": 0.19960254430770874, + "learning_rate": 4.815606443349675e-05, + "loss": 1.4732, + "step": 154 + }, + { + "epoch": 0.4240766073871409, + "grad_norm": 0.21313494443893433, + "learning_rate": 4.812837542432984e-05, + "loss": 1.7094, + "step": 155 + }, + { + "epoch": 0.426812585499316, + "grad_norm": 0.1988086998462677, + "learning_rate": 4.8100488150589704e-05, + "loss": 1.6241, + "step": 156 + }, + { + "epoch": 0.42954856361149113, + "grad_norm": 0.1885240077972412, + "learning_rate": 4.80724028513359e-05, + "loss": 1.64, + "step": 157 + }, + { + "epoch": 0.4322845417236662, + "grad_norm": 0.19793474674224854, + "learning_rate": 4.8044119767325455e-05, + "loss": 1.6615, + "step": 158 + }, + { + "epoch": 0.43502051983584133, + "grad_norm": 0.2065788209438324, + "learning_rate": 4.801563914101096e-05, + "loss": 1.5414, + "step": 159 + }, + { + "epoch": 0.4377564979480164, + "grad_norm": 0.19855208694934845, + "learning_rate": 4.798696121653833e-05, + "loss": 1.7804, + "step": 160 + }, + { + "epoch": 0.44049247606019154, + "grad_norm": 0.21048115193843842, + "learning_rate": 4.795808623974485e-05, + "loss": 1.6356, + "step": 161 + }, + { + "epoch": 0.4432284541723666, + "grad_norm": 0.18900255858898163, + "learning_rate": 4.792901445815698e-05, + "loss": 1.7154, + "step": 162 + }, + { + "epoch": 0.44596443228454175, + "grad_norm": 0.21627336740493774, + "learning_rate": 4.7899746120988245e-05, + "loss": 1.5644, + "step": 163 + }, + { + "epoch": 0.4487004103967168, + "grad_norm": 0.22913017868995667, + "learning_rate": 4.7870281479137146e-05, + "loss": 1.5764, + "step": 164 + }, + { + "epoch": 0.45143638850889195, + "grad_norm": 0.1932600438594818, + "learning_rate": 4.784062078518496e-05, + "loss": 1.7484, + "step": 165 + }, + { + "epoch": 0.454172366621067, + "grad_norm": 0.2302398383617401, + "learning_rate": 4.781076429339359e-05, + "loss": 1.5941, + "step": 166 + }, + { + "epoch": 0.45690834473324216, + "grad_norm": 0.2317684292793274, + "learning_rate": 4.77807122597034e-05, + "loss": 1.6844, + "step": 167 + }, + { + "epoch": 0.45964432284541723, + "grad_norm": 0.2009151577949524, + "learning_rate": 4.775046494173098e-05, + "loss": 1.735, + "step": 168 + }, + { + "epoch": 0.46238030095759236, + "grad_norm": 0.23902402818202972, + "learning_rate": 4.7720022598767e-05, + "loss": 1.6606, + "step": 169 + }, + { + "epoch": 0.46511627906976744, + "grad_norm": 0.22881416976451874, + "learning_rate": 4.768938549177393e-05, + "loss": 1.6456, + "step": 170 + }, + { + "epoch": 0.46785225718194257, + "grad_norm": 0.1795547604560852, + "learning_rate": 4.765855388338383e-05, + "loss": 1.6853, + "step": 171 + }, + { + "epoch": 0.47058823529411764, + "grad_norm": 0.1893799602985382, + "learning_rate": 4.7627528037896106e-05, + "loss": 1.5587, + "step": 172 + }, + { + "epoch": 0.47332421340629277, + "grad_norm": 0.21993231773376465, + "learning_rate": 4.759630822127522e-05, + "loss": 1.8, + "step": 173 + }, + { + "epoch": 0.47606019151846785, + "grad_norm": 0.2114606648683548, + "learning_rate": 4.756489470114842e-05, + "loss": 1.543, + "step": 174 + }, + { + "epoch": 0.478796169630643, + "grad_norm": 0.24682380259037018, + "learning_rate": 4.753328774680347e-05, + "loss": 1.6537, + "step": 175 + }, + { + "epoch": 0.48153214774281805, + "grad_norm": 0.21504569053649902, + "learning_rate": 4.750148762918629e-05, + "loss": 1.7841, + "step": 176 + }, + { + "epoch": 0.4842681258549932, + "grad_norm": 0.2285577356815338, + "learning_rate": 4.746949462089868e-05, + "loss": 1.5911, + "step": 177 + }, + { + "epoch": 0.48700410396716826, + "grad_norm": 0.20859524607658386, + "learning_rate": 4.743730899619598e-05, + "loss": 1.5816, + "step": 178 + }, + { + "epoch": 0.4897400820793434, + "grad_norm": 0.2234918773174286, + "learning_rate": 4.740493103098468e-05, + "loss": 1.5763, + "step": 179 + }, + { + "epoch": 0.49247606019151846, + "grad_norm": 0.20745696127414703, + "learning_rate": 4.7372361002820085e-05, + "loss": 1.4441, + "step": 180 + }, + { + "epoch": 0.4952120383036936, + "grad_norm": 0.18486277759075165, + "learning_rate": 4.733959919090396e-05, + "loss": 1.6772, + "step": 181 + }, + { + "epoch": 0.49794801641586867, + "grad_norm": 0.21360483765602112, + "learning_rate": 4.7306645876082066e-05, + "loss": 1.5955, + "step": 182 + }, + { + "epoch": 0.5006839945280438, + "grad_norm": 0.2167551964521408, + "learning_rate": 4.727350134084182e-05, + "loss": 1.6222, + "step": 183 + }, + { + "epoch": 0.5034199726402189, + "grad_norm": 0.2134450525045395, + "learning_rate": 4.7240165869309846e-05, + "loss": 1.6059, + "step": 184 + }, + { + "epoch": 0.506155950752394, + "grad_norm": 0.21351881325244904, + "learning_rate": 4.720663974724953e-05, + "loss": 1.7258, + "step": 185 + }, + { + "epoch": 0.5088919288645691, + "grad_norm": 0.1952323615550995, + "learning_rate": 4.71729232620586e-05, + "loss": 1.4591, + "step": 186 + }, + { + "epoch": 0.5116279069767442, + "grad_norm": 0.2462124228477478, + "learning_rate": 4.7139016702766615e-05, + "loss": 1.5708, + "step": 187 + }, + { + "epoch": 0.5143638850889193, + "grad_norm": 0.24391762912273407, + "learning_rate": 4.7104920360032545e-05, + "loss": 1.7813, + "step": 188 + }, + { + "epoch": 0.5170998632010944, + "grad_norm": 0.21117964386940002, + "learning_rate": 4.707063452614224e-05, + "loss": 1.6948, + "step": 189 + }, + { + "epoch": 0.5198358413132695, + "grad_norm": 0.20722931623458862, + "learning_rate": 4.703615949500593e-05, + "loss": 1.5925, + "step": 190 + }, + { + "epoch": 0.5225718194254446, + "grad_norm": 0.22132059931755066, + "learning_rate": 4.700149556215571e-05, + "loss": 1.6448, + "step": 191 + }, + { + "epoch": 0.5253077975376197, + "grad_norm": 0.22742822766304016, + "learning_rate": 4.696664302474302e-05, + "loss": 1.5866, + "step": 192 + }, + { + "epoch": 0.5280437756497948, + "grad_norm": 0.2043454349040985, + "learning_rate": 4.693160218153607e-05, + "loss": 1.704, + "step": 193 + }, + { + "epoch": 0.53077975376197, + "grad_norm": 0.21721500158309937, + "learning_rate": 4.68963733329173e-05, + "loss": 1.6768, + "step": 194 + }, + { + "epoch": 0.533515731874145, + "grad_norm": 0.2336382120847702, + "learning_rate": 4.6860956780880796e-05, + "loss": 1.5906, + "step": 195 + }, + { + "epoch": 0.5362517099863201, + "grad_norm": 0.2269752472639084, + "learning_rate": 4.6825352829029705e-05, + "loss": 1.5308, + "step": 196 + }, + { + "epoch": 0.5389876880984952, + "grad_norm": 0.21256719529628754, + "learning_rate": 4.6789561782573635e-05, + "loss": 1.4821, + "step": 197 + }, + { + "epoch": 0.5417236662106704, + "grad_norm": 0.21226170659065247, + "learning_rate": 4.675358394832603e-05, + "loss": 1.7818, + "step": 198 + }, + { + "epoch": 0.5444596443228454, + "grad_norm": 0.21833762526512146, + "learning_rate": 4.671741963470155e-05, + "loss": 1.6556, + "step": 199 + }, + { + "epoch": 0.5471956224350205, + "grad_norm": 0.21411390602588654, + "learning_rate": 4.668106915171341e-05, + "loss": 1.5892, + "step": 200 + }, + { + "epoch": 0.5499316005471956, + "grad_norm": 0.22407633066177368, + "learning_rate": 4.664453281097075e-05, + "loss": 1.6843, + "step": 201 + }, + { + "epoch": 0.5526675786593708, + "grad_norm": 0.20786328613758087, + "learning_rate": 4.660781092567593e-05, + "loss": 1.6561, + "step": 202 + }, + { + "epoch": 0.5554035567715458, + "grad_norm": 0.2268221080303192, + "learning_rate": 4.657090381062189e-05, + "loss": 1.585, + "step": 203 + }, + { + "epoch": 0.5581395348837209, + "grad_norm": 0.2261151373386383, + "learning_rate": 4.6533811782189385e-05, + "loss": 1.6033, + "step": 204 + }, + { + "epoch": 0.560875512995896, + "grad_norm": 0.19799037277698517, + "learning_rate": 4.649653515834436e-05, + "loss": 1.6294, + "step": 205 + }, + { + "epoch": 0.5636114911080712, + "grad_norm": 0.20902524888515472, + "learning_rate": 4.645907425863514e-05, + "loss": 1.7012, + "step": 206 + }, + { + "epoch": 0.5663474692202463, + "grad_norm": 0.2055850476026535, + "learning_rate": 4.642142940418973e-05, + "loss": 1.5587, + "step": 207 + }, + { + "epoch": 0.5690834473324213, + "grad_norm": 0.2205619066953659, + "learning_rate": 4.638360091771309e-05, + "loss": 1.7727, + "step": 208 + }, + { + "epoch": 0.5718194254445964, + "grad_norm": 0.25723960995674133, + "learning_rate": 4.6345589123484314e-05, + "loss": 1.6037, + "step": 209 + }, + { + "epoch": 0.5745554035567716, + "grad_norm": 0.24420645833015442, + "learning_rate": 4.630739434735387e-05, + "loss": 1.7479, + "step": 210 + }, + { + "epoch": 0.5772913816689467, + "grad_norm": 0.24823623895645142, + "learning_rate": 4.626901691674083e-05, + "loss": 1.7085, + "step": 211 + }, + { + "epoch": 0.5800273597811217, + "grad_norm": 0.1966015100479126, + "learning_rate": 4.623045716063002e-05, + "loss": 1.544, + "step": 212 + }, + { + "epoch": 0.5827633378932968, + "grad_norm": 0.22418932616710663, + "learning_rate": 4.6191715409569244e-05, + "loss": 1.6949, + "step": 213 + }, + { + "epoch": 0.585499316005472, + "grad_norm": 0.24894970655441284, + "learning_rate": 4.6152791995666445e-05, + "loss": 1.5838, + "step": 214 + }, + { + "epoch": 0.5882352941176471, + "grad_norm": 0.230922132730484, + "learning_rate": 4.61136872525868e-05, + "loss": 1.6267, + "step": 215 + }, + { + "epoch": 0.5909712722298222, + "grad_norm": 0.22752411663532257, + "learning_rate": 4.6074401515549934e-05, + "loss": 1.4752, + "step": 216 + }, + { + "epoch": 0.5937072503419972, + "grad_norm": 0.22325444221496582, + "learning_rate": 4.6034935121327025e-05, + "loss": 1.7513, + "step": 217 + }, + { + "epoch": 0.5964432284541724, + "grad_norm": 0.21255116164684296, + "learning_rate": 4.599528840823787e-05, + "loss": 1.642, + "step": 218 + }, + { + "epoch": 0.5991792065663475, + "grad_norm": 0.24686329066753387, + "learning_rate": 4.5955461716148065e-05, + "loss": 1.6138, + "step": 219 + }, + { + "epoch": 0.6019151846785226, + "grad_norm": 0.20993772149085999, + "learning_rate": 4.5915455386466014e-05, + "loss": 1.7554, + "step": 220 + }, + { + "epoch": 0.6046511627906976, + "grad_norm": 0.2285137176513672, + "learning_rate": 4.587526976214006e-05, + "loss": 1.6073, + "step": 221 + }, + { + "epoch": 0.6073871409028728, + "grad_norm": 0.22348423302173615, + "learning_rate": 4.5834905187655526e-05, + "loss": 1.6707, + "step": 222 + }, + { + "epoch": 0.6101231190150479, + "grad_norm": 0.23014281690120697, + "learning_rate": 4.5794362009031735e-05, + "loss": 1.4968, + "step": 223 + }, + { + "epoch": 0.612859097127223, + "grad_norm": 0.23900607228279114, + "learning_rate": 4.575364057381909e-05, + "loss": 1.6317, + "step": 224 + }, + { + "epoch": 0.615595075239398, + "grad_norm": 0.24715475738048553, + "learning_rate": 4.571274123109606e-05, + "loss": 1.5805, + "step": 225 + }, + { + "epoch": 0.6183310533515732, + "grad_norm": 0.22451570630073547, + "learning_rate": 4.5671664331466205e-05, + "loss": 1.5956, + "step": 226 + }, + { + "epoch": 0.6210670314637483, + "grad_norm": 0.22505944967269897, + "learning_rate": 4.5630410227055154e-05, + "loss": 1.64, + "step": 227 + }, + { + "epoch": 0.6238030095759234, + "grad_norm": 0.24184450507164001, + "learning_rate": 4.5588979271507625e-05, + "loss": 1.6055, + "step": 228 + }, + { + "epoch": 0.6265389876880985, + "grad_norm": 0.22158478200435638, + "learning_rate": 4.5547371819984344e-05, + "loss": 1.5691, + "step": 229 + }, + { + "epoch": 0.6292749658002736, + "grad_norm": 0.23034054040908813, + "learning_rate": 4.5505588229159025e-05, + "loss": 1.6157, + "step": 230 + }, + { + "epoch": 0.6320109439124487, + "grad_norm": 0.240454763174057, + "learning_rate": 4.5463628857215306e-05, + "loss": 1.6337, + "step": 231 + }, + { + "epoch": 0.6347469220246238, + "grad_norm": 0.2200327068567276, + "learning_rate": 4.5421494063843695e-05, + "loss": 1.4941, + "step": 232 + }, + { + "epoch": 0.6374829001367989, + "grad_norm": 0.23234665393829346, + "learning_rate": 4.537918421023848e-05, + "loss": 1.5992, + "step": 233 + }, + { + "epoch": 0.640218878248974, + "grad_norm": 0.23742978274822235, + "learning_rate": 4.53366996590946e-05, + "loss": 1.481, + "step": 234 + }, + { + "epoch": 0.6429548563611491, + "grad_norm": 0.22323279082775116, + "learning_rate": 4.5294040774604576e-05, + "loss": 1.5291, + "step": 235 + }, + { + "epoch": 0.6456908344733242, + "grad_norm": 0.21182994544506073, + "learning_rate": 4.525120792245538e-05, + "loss": 1.5151, + "step": 236 + }, + { + "epoch": 0.6484268125854993, + "grad_norm": 0.24394573271274567, + "learning_rate": 4.5208201469825304e-05, + "loss": 1.6732, + "step": 237 + }, + { + "epoch": 0.6511627906976745, + "grad_norm": 0.20500163733959198, + "learning_rate": 4.516502178538079e-05, + "loss": 1.7066, + "step": 238 + }, + { + "epoch": 0.6538987688098495, + "grad_norm": 0.2329857349395752, + "learning_rate": 4.512166923927329e-05, + "loss": 1.4977, + "step": 239 + }, + { + "epoch": 0.6566347469220246, + "grad_norm": 0.23349729180335999, + "learning_rate": 4.507814420313608e-05, + "loss": 1.5837, + "step": 240 + }, + { + "epoch": 0.6593707250341997, + "grad_norm": 0.24870243668556213, + "learning_rate": 4.503444705008107e-05, + "loss": 1.6073, + "step": 241 + }, + { + "epoch": 0.6621067031463749, + "grad_norm": 0.2594373822212219, + "learning_rate": 4.4990578154695676e-05, + "loss": 1.6752, + "step": 242 + }, + { + "epoch": 0.66484268125855, + "grad_norm": 0.2545240819454193, + "learning_rate": 4.4946537893039476e-05, + "loss": 1.5185, + "step": 243 + }, + { + "epoch": 0.667578659370725, + "grad_norm": 0.24223731458187103, + "learning_rate": 4.4902326642641095e-05, + "loss": 1.7599, + "step": 244 + }, + { + "epoch": 0.6703146374829001, + "grad_norm": 0.23293475806713104, + "learning_rate": 4.485794478249493e-05, + "loss": 1.6919, + "step": 245 + }, + { + "epoch": 0.6730506155950753, + "grad_norm": 0.2452557235956192, + "learning_rate": 4.481339269305792e-05, + "loss": 1.4946, + "step": 246 + }, + { + "epoch": 0.6757865937072504, + "grad_norm": 0.23280148208141327, + "learning_rate": 4.4768670756246237e-05, + "loss": 1.548, + "step": 247 + }, + { + "epoch": 0.6785225718194254, + "grad_norm": 0.254730224609375, + "learning_rate": 4.4723779355432056e-05, + "loss": 1.6227, + "step": 248 + }, + { + "epoch": 0.6812585499316005, + "grad_norm": 0.25781765580177307, + "learning_rate": 4.4678718875440276e-05, + "loss": 1.6403, + "step": 249 + }, + { + "epoch": 0.6839945280437757, + "grad_norm": 0.2513294816017151, + "learning_rate": 4.463348970254518e-05, + "loss": 1.5416, + "step": 250 + }, + { + "epoch": 0.6839945280437757, + "eval_loss": 1.6065112352371216, + "eval_runtime": 13.3947, + "eval_samples_per_second": 11.497, + "eval_steps_per_second": 11.497, + "step": 250 + }, + { + "epoch": 0.6867305061559508, + "grad_norm": 0.26413506269454956, + "learning_rate": 4.458809222446717e-05, + "loss": 1.4294, + "step": 251 + }, + { + "epoch": 0.6894664842681258, + "grad_norm": 0.24414490163326263, + "learning_rate": 4.454252683036939e-05, + "loss": 1.6371, + "step": 252 + }, + { + "epoch": 0.6922024623803009, + "grad_norm": 0.23543910682201385, + "learning_rate": 4.449679391085444e-05, + "loss": 1.6226, + "step": 253 + }, + { + "epoch": 0.6949384404924761, + "grad_norm": 0.24758242070674896, + "learning_rate": 4.445089385796099e-05, + "loss": 1.7229, + "step": 254 + }, + { + "epoch": 0.6976744186046512, + "grad_norm": 0.2550957500934601, + "learning_rate": 4.4404827065160455e-05, + "loss": 1.6636, + "step": 255 + }, + { + "epoch": 0.7004103967168263, + "grad_norm": 0.22393615543842316, + "learning_rate": 4.4358593927353585e-05, + "loss": 1.5795, + "step": 256 + }, + { + "epoch": 0.7031463748290013, + "grad_norm": 0.22478732466697693, + "learning_rate": 4.431219484086712e-05, + "loss": 1.6236, + "step": 257 + }, + { + "epoch": 0.7058823529411765, + "grad_norm": 0.21901032328605652, + "learning_rate": 4.4265630203450355e-05, + "loss": 1.5929, + "step": 258 + }, + { + "epoch": 0.7086183310533516, + "grad_norm": 0.2183765321969986, + "learning_rate": 4.421890041427174e-05, + "loss": 1.6372, + "step": 259 + }, + { + "epoch": 0.7113543091655267, + "grad_norm": 0.23780032992362976, + "learning_rate": 4.417200587391547e-05, + "loss": 1.4413, + "step": 260 + }, + { + "epoch": 0.7140902872777017, + "grad_norm": 0.22141356766223907, + "learning_rate": 4.412494698437806e-05, + "loss": 1.6594, + "step": 261 + }, + { + "epoch": 0.7168262653898769, + "grad_norm": 0.2270331084728241, + "learning_rate": 4.407772414906487e-05, + "loss": 1.6701, + "step": 262 + }, + { + "epoch": 0.719562243502052, + "grad_norm": 0.24480143189430237, + "learning_rate": 4.403033777278666e-05, + "loss": 1.4914, + "step": 263 + }, + { + "epoch": 0.7222982216142271, + "grad_norm": 0.23042482137680054, + "learning_rate": 4.3982788261756114e-05, + "loss": 1.6845, + "step": 264 + }, + { + "epoch": 0.7250341997264022, + "grad_norm": 0.23995645344257355, + "learning_rate": 4.393507602358438e-05, + "loss": 1.658, + "step": 265 + }, + { + "epoch": 0.7277701778385773, + "grad_norm": 0.25735434889793396, + "learning_rate": 4.388720146727756e-05, + "loss": 1.4613, + "step": 266 + }, + { + "epoch": 0.7305061559507524, + "grad_norm": 0.27036193013191223, + "learning_rate": 4.3839165003233195e-05, + "loss": 1.5341, + "step": 267 + }, + { + "epoch": 0.7332421340629275, + "grad_norm": 0.2285039722919464, + "learning_rate": 4.3790967043236754e-05, + "loss": 1.453, + "step": 268 + }, + { + "epoch": 0.7359781121751026, + "grad_norm": 0.2619383931159973, + "learning_rate": 4.374260800045812e-05, + "loss": 1.4931, + "step": 269 + }, + { + "epoch": 0.7387140902872777, + "grad_norm": 0.22855600714683533, + "learning_rate": 4.369408828944801e-05, + "loss": 1.7458, + "step": 270 + }, + { + "epoch": 0.7414500683994528, + "grad_norm": 0.23299537599086761, + "learning_rate": 4.364540832613449e-05, + "loss": 1.5653, + "step": 271 + }, + { + "epoch": 0.7441860465116279, + "grad_norm": 0.25914594531059265, + "learning_rate": 4.35965685278193e-05, + "loss": 1.6876, + "step": 272 + }, + { + "epoch": 0.746922024623803, + "grad_norm": 0.2430303692817688, + "learning_rate": 4.3547569313174404e-05, + "loss": 1.5761, + "step": 273 + }, + { + "epoch": 0.7496580027359782, + "grad_norm": 0.23968341946601868, + "learning_rate": 4.34984111022383e-05, + "loss": 1.5577, + "step": 274 + }, + { + "epoch": 0.7523939808481532, + "grad_norm": 0.25374993681907654, + "learning_rate": 4.3449094316412485e-05, + "loss": 1.6936, + "step": 275 + }, + { + "epoch": 0.7551299589603283, + "grad_norm": 0.25317367911338806, + "learning_rate": 4.339961937845779e-05, + "loss": 1.6957, + "step": 276 + }, + { + "epoch": 0.7578659370725034, + "grad_norm": 0.231050044298172, + "learning_rate": 4.3349986712490795e-05, + "loss": 1.5386, + "step": 277 + }, + { + "epoch": 0.7606019151846786, + "grad_norm": 0.23644736409187317, + "learning_rate": 4.330019674398019e-05, + "loss": 1.6485, + "step": 278 + }, + { + "epoch": 0.7633378932968536, + "grad_norm": 0.24971966445446014, + "learning_rate": 4.32502498997431e-05, + "loss": 1.6196, + "step": 279 + }, + { + "epoch": 0.7660738714090287, + "grad_norm": 0.22572147846221924, + "learning_rate": 4.3200146607941444e-05, + "loss": 1.6654, + "step": 280 + }, + { + "epoch": 0.7688098495212038, + "grad_norm": 0.256597101688385, + "learning_rate": 4.3149887298078276e-05, + "loss": 1.6552, + "step": 281 + }, + { + "epoch": 0.771545827633379, + "grad_norm": 0.226451575756073, + "learning_rate": 4.3099472400994084e-05, + "loss": 1.6164, + "step": 282 + }, + { + "epoch": 0.774281805745554, + "grad_norm": 0.2507132291793823, + "learning_rate": 4.3048902348863116e-05, + "loss": 1.555, + "step": 283 + }, + { + "epoch": 0.7770177838577291, + "grad_norm": 0.23009182512760162, + "learning_rate": 4.2998177575189644e-05, + "loss": 1.4741, + "step": 284 + }, + { + "epoch": 0.7797537619699042, + "grad_norm": 0.24451418220996857, + "learning_rate": 4.294729851480429e-05, + "loss": 1.5704, + "step": 285 + }, + { + "epoch": 0.7824897400820794, + "grad_norm": 0.2464141994714737, + "learning_rate": 4.2896265603860246e-05, + "loss": 1.6357, + "step": 286 + }, + { + "epoch": 0.7852257181942545, + "grad_norm": 0.24296295642852783, + "learning_rate": 4.28450792798296e-05, + "loss": 1.5492, + "step": 287 + }, + { + "epoch": 0.7879616963064295, + "grad_norm": 0.2681036591529846, + "learning_rate": 4.279373998149954e-05, + "loss": 1.6973, + "step": 288 + }, + { + "epoch": 0.7906976744186046, + "grad_norm": 0.23149123787879944, + "learning_rate": 4.2742248148968576e-05, + "loss": 1.6626, + "step": 289 + }, + { + "epoch": 0.7934336525307798, + "grad_norm": 0.22642168402671814, + "learning_rate": 4.269060422364284e-05, + "loss": 1.7018, + "step": 290 + }, + { + "epoch": 0.7961696306429549, + "grad_norm": 0.23424457013607025, + "learning_rate": 4.263880864823221e-05, + "loss": 1.6973, + "step": 291 + }, + { + "epoch": 0.79890560875513, + "grad_norm": 0.24797867238521576, + "learning_rate": 4.2586861866746606e-05, + "loss": 1.6631, + "step": 292 + }, + { + "epoch": 0.801641586867305, + "grad_norm": 0.24017377197742462, + "learning_rate": 4.2534764324492115e-05, + "loss": 1.692, + "step": 293 + }, + { + "epoch": 0.8043775649794802, + "grad_norm": 0.22952371835708618, + "learning_rate": 4.248251646806719e-05, + "loss": 1.3692, + "step": 294 + }, + { + "epoch": 0.8071135430916553, + "grad_norm": 0.2534586787223816, + "learning_rate": 4.243011874535886e-05, + "loss": 1.7223, + "step": 295 + }, + { + "epoch": 0.8098495212038304, + "grad_norm": 0.23179998993873596, + "learning_rate": 4.237757160553883e-05, + "loss": 1.6688, + "step": 296 + }, + { + "epoch": 0.8125854993160054, + "grad_norm": 0.21562454104423523, + "learning_rate": 4.2324875499059693e-05, + "loss": 1.5706, + "step": 297 + }, + { + "epoch": 0.8153214774281806, + "grad_norm": 0.30047473311424255, + "learning_rate": 4.227203087765099e-05, + "loss": 1.5275, + "step": 298 + }, + { + "epoch": 0.8180574555403557, + "grad_norm": 0.23342545330524445, + "learning_rate": 4.221903819431543e-05, + "loss": 1.6569, + "step": 299 + }, + { + "epoch": 0.8207934336525308, + "grad_norm": 0.2696247100830078, + "learning_rate": 4.216589790332495e-05, + "loss": 1.5803, + "step": 300 + }, + { + "epoch": 0.8235294117647058, + "grad_norm": 0.24961510300636292, + "learning_rate": 4.21126104602168e-05, + "loss": 1.4535, + "step": 301 + }, + { + "epoch": 0.826265389876881, + "grad_norm": 0.24784626066684723, + "learning_rate": 4.205917632178972e-05, + "loss": 1.6035, + "step": 302 + }, + { + "epoch": 0.8290013679890561, + "grad_norm": 0.22941957414150238, + "learning_rate": 4.200559594609994e-05, + "loss": 1.6502, + "step": 303 + }, + { + "epoch": 0.8317373461012312, + "grad_norm": 0.23505181074142456, + "learning_rate": 4.195186979245728e-05, + "loss": 1.7577, + "step": 304 + }, + { + "epoch": 0.8344733242134063, + "grad_norm": 0.2516261041164398, + "learning_rate": 4.189799832142126e-05, + "loss": 1.4947, + "step": 305 + }, + { + "epoch": 0.8372093023255814, + "grad_norm": 0.28713610768318176, + "learning_rate": 4.1843981994797075e-05, + "loss": 1.8651, + "step": 306 + }, + { + "epoch": 0.8399452804377565, + "grad_norm": 0.2537023425102234, + "learning_rate": 4.178982127563169e-05, + "loss": 1.3971, + "step": 307 + }, + { + "epoch": 0.8426812585499316, + "grad_norm": 0.2575843930244446, + "learning_rate": 4.173551662820985e-05, + "loss": 1.7023, + "step": 308 + }, + { + "epoch": 0.8454172366621067, + "grad_norm": 0.2560933828353882, + "learning_rate": 4.168106851805009e-05, + "loss": 1.5333, + "step": 309 + }, + { + "epoch": 0.8481532147742818, + "grad_norm": 0.22800202667713165, + "learning_rate": 4.16264774119008e-05, + "loss": 1.5548, + "step": 310 + }, + { + "epoch": 0.8508891928864569, + "grad_norm": 0.2175893485546112, + "learning_rate": 4.1571743777736116e-05, + "loss": 1.7173, + "step": 311 + }, + { + "epoch": 0.853625170998632, + "grad_norm": 0.23807293176651, + "learning_rate": 4.151686808475204e-05, + "loss": 1.5548, + "step": 312 + }, + { + "epoch": 0.8563611491108071, + "grad_norm": 0.26670193672180176, + "learning_rate": 4.1461850803362314e-05, + "loss": 1.5273, + "step": 313 + }, + { + "epoch": 0.8590971272229823, + "grad_norm": 0.2579667866230011, + "learning_rate": 4.140669240519442e-05, + "loss": 1.6407, + "step": 314 + }, + { + "epoch": 0.8618331053351573, + "grad_norm": 0.23136496543884277, + "learning_rate": 4.135139336308559e-05, + "loss": 1.5134, + "step": 315 + }, + { + "epoch": 0.8645690834473324, + "grad_norm": 0.2494489699602127, + "learning_rate": 4.129595415107864e-05, + "loss": 1.5429, + "step": 316 + }, + { + "epoch": 0.8673050615595075, + "grad_norm": 0.2552199363708496, + "learning_rate": 4.124037524441803e-05, + "loss": 1.5348, + "step": 317 + }, + { + "epoch": 0.8700410396716827, + "grad_norm": 0.2449236959218979, + "learning_rate": 4.118465711954569e-05, + "loss": 1.7689, + "step": 318 + }, + { + "epoch": 0.8727770177838577, + "grad_norm": 0.2517363429069519, + "learning_rate": 4.112880025409701e-05, + "loss": 1.6258, + "step": 319 + }, + { + "epoch": 0.8755129958960328, + "grad_norm": 0.27004796266555786, + "learning_rate": 4.107280512689668e-05, + "loss": 1.5855, + "step": 320 + }, + { + "epoch": 0.8782489740082079, + "grad_norm": 0.27461642026901245, + "learning_rate": 4.101667221795465e-05, + "loss": 1.57, + "step": 321 + }, + { + "epoch": 0.8809849521203831, + "grad_norm": 0.2513730525970459, + "learning_rate": 4.0960402008461955e-05, + "loss": 1.5509, + "step": 322 + }, + { + "epoch": 0.8837209302325582, + "grad_norm": 0.2727186977863312, + "learning_rate": 4.090399498078664e-05, + "loss": 1.5905, + "step": 323 + }, + { + "epoch": 0.8864569083447332, + "grad_norm": 0.24590528011322021, + "learning_rate": 4.084745161846961e-05, + "loss": 1.5459, + "step": 324 + }, + { + "epoch": 0.8891928864569083, + "grad_norm": 0.22626182436943054, + "learning_rate": 4.079077240622043e-05, + "loss": 1.6127, + "step": 325 + }, + { + "epoch": 0.8919288645690835, + "grad_norm": 0.24243368208408356, + "learning_rate": 4.0733957829913296e-05, + "loss": 1.5687, + "step": 326 + }, + { + "epoch": 0.8946648426812586, + "grad_norm": 0.2647876739501953, + "learning_rate": 4.0677008376582716e-05, + "loss": 1.6761, + "step": 327 + }, + { + "epoch": 0.8974008207934336, + "grad_norm": 0.2398100048303604, + "learning_rate": 4.061992453441946e-05, + "loss": 1.5666, + "step": 328 + }, + { + "epoch": 0.9001367989056087, + "grad_norm": 0.2463439702987671, + "learning_rate": 4.056270679276631e-05, + "loss": 1.7682, + "step": 329 + }, + { + "epoch": 0.9028727770177839, + "grad_norm": 0.23695454001426697, + "learning_rate": 4.050535564211388e-05, + "loss": 1.4589, + "step": 330 + }, + { + "epoch": 0.905608755129959, + "grad_norm": 0.24376732110977173, + "learning_rate": 4.044787157409642e-05, + "loss": 1.4797, + "step": 331 + }, + { + "epoch": 0.908344733242134, + "grad_norm": 0.24585182964801788, + "learning_rate": 4.039025508148758e-05, + "loss": 1.5843, + "step": 332 + }, + { + "epoch": 0.9110807113543091, + "grad_norm": 0.266337126493454, + "learning_rate": 4.0332506658196225e-05, + "loss": 1.7075, + "step": 333 + }, + { + "epoch": 0.9138166894664843, + "grad_norm": 0.2664279043674469, + "learning_rate": 4.027462679926215e-05, + "loss": 1.6063, + "step": 334 + }, + { + "epoch": 0.9165526675786594, + "grad_norm": 0.2796095013618469, + "learning_rate": 4.021661600085189e-05, + "loss": 1.7072, + "step": 335 + }, + { + "epoch": 0.9192886456908345, + "grad_norm": 0.26439887285232544, + "learning_rate": 4.01584747602544e-05, + "loss": 1.5579, + "step": 336 + }, + { + "epoch": 0.9220246238030095, + "grad_norm": 0.2440110445022583, + "learning_rate": 4.010020357587687e-05, + "loss": 1.5581, + "step": 337 + }, + { + "epoch": 0.9247606019151847, + "grad_norm": 0.2567618787288666, + "learning_rate": 4.00418029472404e-05, + "loss": 1.7241, + "step": 338 + }, + { + "epoch": 0.9274965800273598, + "grad_norm": 0.26454323530197144, + "learning_rate": 3.9983273374975726e-05, + "loss": 1.5839, + "step": 339 + }, + { + "epoch": 0.9302325581395349, + "grad_norm": 0.2620810866355896, + "learning_rate": 3.9924615360818934e-05, + "loss": 1.7599, + "step": 340 + }, + { + "epoch": 0.93296853625171, + "grad_norm": 0.2683962285518646, + "learning_rate": 3.986582940760717e-05, + "loss": 1.6423, + "step": 341 + }, + { + "epoch": 0.9357045143638851, + "grad_norm": 0.22277259826660156, + "learning_rate": 3.9806916019274306e-05, + "loss": 1.546, + "step": 342 + }, + { + "epoch": 0.9384404924760602, + "grad_norm": 0.26338738203048706, + "learning_rate": 3.9747875700846646e-05, + "loss": 1.5739, + "step": 343 + }, + { + "epoch": 0.9411764705882353, + "grad_norm": 0.27578940987586975, + "learning_rate": 3.968870895843858e-05, + "loss": 1.6281, + "step": 344 + }, + { + "epoch": 0.9439124487004104, + "grad_norm": 0.24065442383289337, + "learning_rate": 3.962941629924824e-05, + "loss": 1.4456, + "step": 345 + }, + { + "epoch": 0.9466484268125855, + "grad_norm": 0.25176844000816345, + "learning_rate": 3.956999823155315e-05, + "loss": 1.6577, + "step": 346 + }, + { + "epoch": 0.9493844049247606, + "grad_norm": 0.2668481767177582, + "learning_rate": 3.951045526470592e-05, + "loss": 1.5013, + "step": 347 + }, + { + "epoch": 0.9521203830369357, + "grad_norm": 0.27450481057167053, + "learning_rate": 3.94507879091298e-05, + "loss": 1.5217, + "step": 348 + }, + { + "epoch": 0.9548563611491108, + "grad_norm": 0.2651226818561554, + "learning_rate": 3.939099667631438e-05, + "loss": 1.5877, + "step": 349 + }, + { + "epoch": 0.957592339261286, + "grad_norm": 0.27575191855430603, + "learning_rate": 3.933108207881112e-05, + "loss": 1.5836, + "step": 350 + }, + { + "epoch": 0.960328317373461, + "grad_norm": 0.248700350522995, + "learning_rate": 3.927104463022906e-05, + "loss": 1.6258, + "step": 351 + }, + { + "epoch": 0.9630642954856361, + "grad_norm": 0.27306076884269714, + "learning_rate": 3.921088484523032e-05, + "loss": 1.5255, + "step": 352 + }, + { + "epoch": 0.9658002735978112, + "grad_norm": 0.2685610055923462, + "learning_rate": 3.9150603239525765e-05, + "loss": 1.607, + "step": 353 + }, + { + "epoch": 0.9685362517099864, + "grad_norm": 0.2934316396713257, + "learning_rate": 3.909020032987051e-05, + "loss": 1.5252, + "step": 354 + }, + { + "epoch": 0.9712722298221614, + "grad_norm": 0.2768978774547577, + "learning_rate": 3.902967663405956e-05, + "loss": 1.6291, + "step": 355 + }, + { + "epoch": 0.9740082079343365, + "grad_norm": 0.29954272508621216, + "learning_rate": 3.8969032670923346e-05, + "loss": 1.481, + "step": 356 + }, + { + "epoch": 0.9767441860465116, + "grad_norm": 0.26432037353515625, + "learning_rate": 3.890826896032326e-05, + "loss": 1.8236, + "step": 357 + }, + { + "epoch": 0.9794801641586868, + "grad_norm": 0.2719402611255646, + "learning_rate": 3.884738602314719e-05, + "loss": 1.626, + "step": 358 + }, + { + "epoch": 0.9822161422708618, + "grad_norm": 0.2573629915714264, + "learning_rate": 3.8786384381305123e-05, + "loss": 1.4863, + "step": 359 + }, + { + "epoch": 0.9849521203830369, + "grad_norm": 0.27962571382522583, + "learning_rate": 3.87252645577246e-05, + "loss": 1.567, + "step": 360 + }, + { + "epoch": 0.987688098495212, + "grad_norm": 0.3235297203063965, + "learning_rate": 3.866402707634624e-05, + "loss": 1.5815, + "step": 361 + }, + { + "epoch": 0.9904240766073872, + "grad_norm": 0.24766239523887634, + "learning_rate": 3.8602672462119294e-05, + "loss": 1.6327, + "step": 362 + }, + { + "epoch": 0.9931600547195623, + "grad_norm": 0.2470933198928833, + "learning_rate": 3.85412012409971e-05, + "loss": 1.626, + "step": 363 + }, + { + "epoch": 0.9958960328317373, + "grad_norm": 0.2527347207069397, + "learning_rate": 3.847961393993261e-05, + "loss": 1.5373, + "step": 364 + }, + { + "epoch": 0.9986320109439124, + "grad_norm": 0.24633722007274628, + "learning_rate": 3.8417911086873834e-05, + "loss": 1.5994, + "step": 365 + }, + { + "epoch": 1.0013679890560876, + "grad_norm": 0.2634814381599426, + "learning_rate": 3.835609321075934e-05, + "loss": 1.5801, + "step": 366 + }, + { + "epoch": 1.0041039671682626, + "grad_norm": 0.2554261386394501, + "learning_rate": 3.8294160841513715e-05, + "loss": 1.6, + "step": 367 + }, + { + "epoch": 1.0068399452804377, + "grad_norm": 0.23322933912277222, + "learning_rate": 3.823211451004304e-05, + "loss": 1.4199, + "step": 368 + }, + { + "epoch": 1.009575923392613, + "grad_norm": 0.23709669709205627, + "learning_rate": 3.816995474823028e-05, + "loss": 1.5409, + "step": 369 + }, + { + "epoch": 1.012311901504788, + "grad_norm": 0.2459351122379303, + "learning_rate": 3.8107682088930794e-05, + "loss": 1.5137, + "step": 370 + }, + { + "epoch": 1.015047879616963, + "grad_norm": 0.2492617666721344, + "learning_rate": 3.8045297065967744e-05, + "loss": 1.6054, + "step": 371 + }, + { + "epoch": 1.0177838577291383, + "grad_norm": 0.2558859884738922, + "learning_rate": 3.798280021412749e-05, + "loss": 1.6189, + "step": 372 + }, + { + "epoch": 1.0205198358413132, + "grad_norm": 0.23855186998844147, + "learning_rate": 3.792019206915504e-05, + "loss": 1.7069, + "step": 373 + }, + { + "epoch": 1.0232558139534884, + "grad_norm": 0.2665524482727051, + "learning_rate": 3.7857473167749435e-05, + "loss": 1.5505, + "step": 374 + }, + { + "epoch": 1.0259917920656634, + "grad_norm": 0.24079938232898712, + "learning_rate": 3.779464404755919e-05, + "loss": 1.4712, + "step": 375 + }, + { + "epoch": 1.0287277701778386, + "grad_norm": 0.2564024329185486, + "learning_rate": 3.773170524717763e-05, + "loss": 1.5464, + "step": 376 + }, + { + "epoch": 1.0314637482900137, + "grad_norm": 0.23460035026073456, + "learning_rate": 3.766865730613828e-05, + "loss": 1.6146, + "step": 377 + }, + { + "epoch": 1.0341997264021887, + "grad_norm": 0.24386173486709595, + "learning_rate": 3.760550076491031e-05, + "loss": 1.4695, + "step": 378 + }, + { + "epoch": 1.036935704514364, + "grad_norm": 0.29760661721229553, + "learning_rate": 3.754223616489379e-05, + "loss": 1.3848, + "step": 379 + }, + { + "epoch": 1.039671682626539, + "grad_norm": 0.243236243724823, + "learning_rate": 3.7478864048415136e-05, + "loss": 1.4215, + "step": 380 + }, + { + "epoch": 1.042407660738714, + "grad_norm": 0.259069561958313, + "learning_rate": 3.7415384958722445e-05, + "loss": 1.5352, + "step": 381 + }, + { + "epoch": 1.0451436388508892, + "grad_norm": 0.281246542930603, + "learning_rate": 3.735179943998081e-05, + "loss": 1.6266, + "step": 382 + }, + { + "epoch": 1.0478796169630642, + "grad_norm": 0.2552565634250641, + "learning_rate": 3.728810803726767e-05, + "loss": 1.6849, + "step": 383 + }, + { + "epoch": 1.0506155950752394, + "grad_norm": 0.25851351022720337, + "learning_rate": 3.7224311296568134e-05, + "loss": 1.4284, + "step": 384 + }, + { + "epoch": 1.0533515731874146, + "grad_norm": 0.2506335973739624, + "learning_rate": 3.716040976477033e-05, + "loss": 1.4498, + "step": 385 + }, + { + "epoch": 1.0560875512995895, + "grad_norm": 0.3301680088043213, + "learning_rate": 3.709640398966067e-05, + "loss": 1.5649, + "step": 386 + }, + { + "epoch": 1.0588235294117647, + "grad_norm": 0.2851655185222626, + "learning_rate": 3.703229451991918e-05, + "loss": 1.6324, + "step": 387 + }, + { + "epoch": 1.06155950752394, + "grad_norm": 0.27497461438179016, + "learning_rate": 3.69680819051148e-05, + "loss": 1.5628, + "step": 388 + }, + { + "epoch": 1.0642954856361149, + "grad_norm": 0.2582625448703766, + "learning_rate": 3.6903766695700656e-05, + "loss": 1.5979, + "step": 389 + }, + { + "epoch": 1.06703146374829, + "grad_norm": 0.2679811418056488, + "learning_rate": 3.6839349443009364e-05, + "loss": 1.51, + "step": 390 + }, + { + "epoch": 1.069767441860465, + "grad_norm": 0.2523844242095947, + "learning_rate": 3.677483069924827e-05, + "loss": 1.6217, + "step": 391 + }, + { + "epoch": 1.0725034199726402, + "grad_norm": 0.23916080594062805, + "learning_rate": 3.671021101749476e-05, + "loss": 1.6128, + "step": 392 + }, + { + "epoch": 1.0752393980848154, + "grad_norm": 0.2704644203186035, + "learning_rate": 3.664549095169148e-05, + "loss": 1.5524, + "step": 393 + }, + { + "epoch": 1.0779753761969904, + "grad_norm": 0.27870067954063416, + "learning_rate": 3.6580671056641616e-05, + "loss": 1.6693, + "step": 394 + }, + { + "epoch": 1.0807113543091655, + "grad_norm": 0.2483988106250763, + "learning_rate": 3.6515751888004113e-05, + "loss": 1.4771, + "step": 395 + }, + { + "epoch": 1.0834473324213407, + "grad_norm": 0.2584916949272156, + "learning_rate": 3.645073400228895e-05, + "loss": 1.5099, + "step": 396 + }, + { + "epoch": 1.0861833105335157, + "grad_norm": 0.24778081476688385, + "learning_rate": 3.6385617956852286e-05, + "loss": 1.6118, + "step": 397 + }, + { + "epoch": 1.0889192886456909, + "grad_norm": 0.2260492891073227, + "learning_rate": 3.632040430989181e-05, + "loss": 1.5512, + "step": 398 + }, + { + "epoch": 1.0916552667578658, + "grad_norm": 0.2751753330230713, + "learning_rate": 3.6255093620441834e-05, + "loss": 1.559, + "step": 399 + }, + { + "epoch": 1.094391244870041, + "grad_norm": 0.2648756206035614, + "learning_rate": 3.618968644836859e-05, + "loss": 1.5666, + "step": 400 + }, + { + "epoch": 1.0971272229822162, + "grad_norm": 0.24056974053382874, + "learning_rate": 3.612418335436536e-05, + "loss": 1.5169, + "step": 401 + }, + { + "epoch": 1.0998632010943912, + "grad_norm": 0.23621565103530884, + "learning_rate": 3.605858489994771e-05, + "loss": 1.7854, + "step": 402 + }, + { + "epoch": 1.1025991792065664, + "grad_norm": 0.25893279910087585, + "learning_rate": 3.5992891647448696e-05, + "loss": 1.6353, + "step": 403 + }, + { + "epoch": 1.1053351573187415, + "grad_norm": 0.2806542217731476, + "learning_rate": 3.592710416001398e-05, + "loss": 1.7623, + "step": 404 + }, + { + "epoch": 1.1080711354309165, + "grad_norm": 0.29038006067276, + "learning_rate": 3.586122300159707e-05, + "loss": 1.4457, + "step": 405 + }, + { + "epoch": 1.1108071135430917, + "grad_norm": 0.27859950065612793, + "learning_rate": 3.5795248736954426e-05, + "loss": 1.4407, + "step": 406 + }, + { + "epoch": 1.1135430916552667, + "grad_norm": 0.2583613991737366, + "learning_rate": 3.5729181931640674e-05, + "loss": 1.7205, + "step": 407 + }, + { + "epoch": 1.1162790697674418, + "grad_norm": 0.25556913018226624, + "learning_rate": 3.5663023152003705e-05, + "loss": 1.5662, + "step": 408 + }, + { + "epoch": 1.119015047879617, + "grad_norm": 0.26569557189941406, + "learning_rate": 3.559677296517987e-05, + "loss": 1.5788, + "step": 409 + }, + { + "epoch": 1.121751025991792, + "grad_norm": 0.2778777778148651, + "learning_rate": 3.5530431939089084e-05, + "loss": 1.747, + "step": 410 + }, + { + "epoch": 1.1244870041039672, + "grad_norm": 0.29031357169151306, + "learning_rate": 3.546400064242997e-05, + "loss": 1.6342, + "step": 411 + }, + { + "epoch": 1.1272229822161424, + "grad_norm": 0.26415035128593445, + "learning_rate": 3.5397479644674964e-05, + "loss": 1.5624, + "step": 412 + }, + { + "epoch": 1.1299589603283173, + "grad_norm": 0.2662270963191986, + "learning_rate": 3.533086951606549e-05, + "loss": 1.6745, + "step": 413 + }, + { + "epoch": 1.1326949384404925, + "grad_norm": 0.2567162811756134, + "learning_rate": 3.5264170827607004e-05, + "loss": 1.6727, + "step": 414 + }, + { + "epoch": 1.1354309165526675, + "grad_norm": 0.28177884221076965, + "learning_rate": 3.519738415106413e-05, + "loss": 1.4554, + "step": 415 + }, + { + "epoch": 1.1381668946648427, + "grad_norm": 0.277425616979599, + "learning_rate": 3.513051005895576e-05, + "loss": 1.5096, + "step": 416 + }, + { + "epoch": 1.1409028727770179, + "grad_norm": 0.2886553108692169, + "learning_rate": 3.506354912455016e-05, + "loss": 1.6723, + "step": 417 + }, + { + "epoch": 1.1436388508891928, + "grad_norm": 0.245221808552742, + "learning_rate": 3.499650192186001e-05, + "loss": 1.4508, + "step": 418 + }, + { + "epoch": 1.146374829001368, + "grad_norm": 0.31830281019210815, + "learning_rate": 3.492936902563754e-05, + "loss": 1.4235, + "step": 419 + }, + { + "epoch": 1.1491108071135432, + "grad_norm": 0.2477853149175644, + "learning_rate": 3.486215101136954e-05, + "loss": 1.4991, + "step": 420 + }, + { + "epoch": 1.1518467852257182, + "grad_norm": 0.2735631465911865, + "learning_rate": 3.47948484552725e-05, + "loss": 1.5141, + "step": 421 + }, + { + "epoch": 1.1545827633378933, + "grad_norm": 0.263138085603714, + "learning_rate": 3.47274619342876e-05, + "loss": 1.7006, + "step": 422 + }, + { + "epoch": 1.1573187414500685, + "grad_norm": 0.2716856598854065, + "learning_rate": 3.465999202607583e-05, + "loss": 1.6627, + "step": 423 + }, + { + "epoch": 1.1600547195622435, + "grad_norm": 0.26602792739868164, + "learning_rate": 3.459243930901297e-05, + "loss": 1.5501, + "step": 424 + }, + { + "epoch": 1.1627906976744187, + "grad_norm": 0.27663204073905945, + "learning_rate": 3.452480436218471e-05, + "loss": 1.7229, + "step": 425 + }, + { + "epoch": 1.1655266757865936, + "grad_norm": 0.3033023178577423, + "learning_rate": 3.4457087765381584e-05, + "loss": 1.5789, + "step": 426 + }, + { + "epoch": 1.1682626538987688, + "grad_norm": 0.2735162675380707, + "learning_rate": 3.43892900990941e-05, + "loss": 1.6559, + "step": 427 + }, + { + "epoch": 1.170998632010944, + "grad_norm": 0.26795685291290283, + "learning_rate": 3.432141194450772e-05, + "loss": 1.4425, + "step": 428 + }, + { + "epoch": 1.173734610123119, + "grad_norm": 0.2699303925037384, + "learning_rate": 3.425345388349786e-05, + "loss": 1.5429, + "step": 429 + }, + { + "epoch": 1.1764705882352942, + "grad_norm": 0.2614208459854126, + "learning_rate": 3.418541649862494e-05, + "loss": 1.5916, + "step": 430 + }, + { + "epoch": 1.1792065663474691, + "grad_norm": 0.30162835121154785, + "learning_rate": 3.4117300373129376e-05, + "loss": 1.4801, + "step": 431 + }, + { + "epoch": 1.1819425444596443, + "grad_norm": 0.2973942458629608, + "learning_rate": 3.404910609092655e-05, + "loss": 1.5476, + "step": 432 + }, + { + "epoch": 1.1846785225718195, + "grad_norm": 0.26698416471481323, + "learning_rate": 3.3980834236601853e-05, + "loss": 1.6382, + "step": 433 + }, + { + "epoch": 1.1874145006839945, + "grad_norm": 0.3085285723209381, + "learning_rate": 3.391248539540565e-05, + "loss": 1.5059, + "step": 434 + }, + { + "epoch": 1.1901504787961696, + "grad_norm": 0.25455278158187866, + "learning_rate": 3.384406015324826e-05, + "loss": 1.6612, + "step": 435 + }, + { + "epoch": 1.1928864569083448, + "grad_norm": 0.2725888192653656, + "learning_rate": 3.3775559096694933e-05, + "loss": 1.5598, + "step": 436 + }, + { + "epoch": 1.1956224350205198, + "grad_norm": 0.34251880645751953, + "learning_rate": 3.370698281296083e-05, + "loss": 1.5973, + "step": 437 + }, + { + "epoch": 1.198358413132695, + "grad_norm": 0.2710118889808655, + "learning_rate": 3.363833188990599e-05, + "loss": 1.4116, + "step": 438 + }, + { + "epoch": 1.2010943912448702, + "grad_norm": 0.2622835636138916, + "learning_rate": 3.3569606916030294e-05, + "loss": 1.4546, + "step": 439 + }, + { + "epoch": 1.2038303693570451, + "grad_norm": 0.2685026228427887, + "learning_rate": 3.350080848046839e-05, + "loss": 1.5789, + "step": 440 + }, + { + "epoch": 1.2065663474692203, + "grad_norm": 0.2792280912399292, + "learning_rate": 3.343193717298469e-05, + "loss": 1.5287, + "step": 441 + }, + { + "epoch": 1.2093023255813953, + "grad_norm": 0.3190619647502899, + "learning_rate": 3.3362993583968264e-05, + "loss": 1.5803, + "step": 442 + }, + { + "epoch": 1.2120383036935705, + "grad_norm": 0.27128350734710693, + "learning_rate": 3.329397830442784e-05, + "loss": 1.5371, + "step": 443 + }, + { + "epoch": 1.2147742818057456, + "grad_norm": 0.2853277325630188, + "learning_rate": 3.322489192598665e-05, + "loss": 1.5739, + "step": 444 + }, + { + "epoch": 1.2175102599179206, + "grad_norm": 0.2518289387226105, + "learning_rate": 3.3155735040877465e-05, + "loss": 1.6408, + "step": 445 + }, + { + "epoch": 1.2202462380300958, + "grad_norm": 0.297750324010849, + "learning_rate": 3.308650824193744e-05, + "loss": 1.6255, + "step": 446 + }, + { + "epoch": 1.2229822161422708, + "grad_norm": 0.2988694906234741, + "learning_rate": 3.301721212260306e-05, + "loss": 1.5215, + "step": 447 + }, + { + "epoch": 1.225718194254446, + "grad_norm": 0.2906099557876587, + "learning_rate": 3.294784727690503e-05, + "loss": 1.5124, + "step": 448 + }, + { + "epoch": 1.2284541723666211, + "grad_norm": 0.28081414103507996, + "learning_rate": 3.2878414299463225e-05, + "loss": 1.6691, + "step": 449 + }, + { + "epoch": 1.231190150478796, + "grad_norm": 0.26118987798690796, + "learning_rate": 3.280891378548156e-05, + "loss": 1.524, + "step": 450 + }, + { + "epoch": 1.2339261285909713, + "grad_norm": 0.2828025817871094, + "learning_rate": 3.273934633074291e-05, + "loss": 1.7315, + "step": 451 + }, + { + "epoch": 1.2366621067031465, + "grad_norm": 0.27713146805763245, + "learning_rate": 3.2669712531603966e-05, + "loss": 1.534, + "step": 452 + }, + { + "epoch": 1.2393980848153214, + "grad_norm": 0.30562645196914673, + "learning_rate": 3.2600012984990165e-05, + "loss": 1.4404, + "step": 453 + }, + { + "epoch": 1.2421340629274966, + "grad_norm": 0.30407342314720154, + "learning_rate": 3.2530248288390555e-05, + "loss": 1.5327, + "step": 454 + }, + { + "epoch": 1.2448700410396718, + "grad_norm": 0.3026122450828552, + "learning_rate": 3.246041903985264e-05, + "loss": 1.4441, + "step": 455 + }, + { + "epoch": 1.2476060191518468, + "grad_norm": 0.3250129520893097, + "learning_rate": 3.2390525837977334e-05, + "loss": 1.661, + "step": 456 + }, + { + "epoch": 1.250341997264022, + "grad_norm": 0.29971036314964294, + "learning_rate": 3.232056928191376e-05, + "loss": 1.5989, + "step": 457 + }, + { + "epoch": 1.253077975376197, + "grad_norm": 0.296903133392334, + "learning_rate": 3.225054997135413e-05, + "loss": 1.5724, + "step": 458 + }, + { + "epoch": 1.255813953488372, + "grad_norm": 0.31808045506477356, + "learning_rate": 3.218046850652862e-05, + "loss": 1.6791, + "step": 459 + }, + { + "epoch": 1.2585499316005473, + "grad_norm": 0.2797161340713501, + "learning_rate": 3.211032548820019e-05, + "loss": 1.4345, + "step": 460 + }, + { + "epoch": 1.2612859097127223, + "grad_norm": 0.3660871386528015, + "learning_rate": 3.20401215176595e-05, + "loss": 1.5669, + "step": 461 + }, + { + "epoch": 1.2640218878248974, + "grad_norm": 0.30877435207366943, + "learning_rate": 3.196985719671968e-05, + "loss": 1.5406, + "step": 462 + }, + { + "epoch": 1.2667578659370724, + "grad_norm": 0.29524242877960205, + "learning_rate": 3.1899533127711215e-05, + "loss": 1.3333, + "step": 463 + }, + { + "epoch": 1.2694938440492476, + "grad_norm": 0.3187881112098694, + "learning_rate": 3.182914991347677e-05, + "loss": 1.5237, + "step": 464 + }, + { + "epoch": 1.2722298221614228, + "grad_norm": 0.2810283303260803, + "learning_rate": 3.1758708157366036e-05, + "loss": 1.4334, + "step": 465 + }, + { + "epoch": 1.2749658002735977, + "grad_norm": 0.27640587091445923, + "learning_rate": 3.168820846323053e-05, + "loss": 1.5078, + "step": 466 + }, + { + "epoch": 1.277701778385773, + "grad_norm": 0.2678714692592621, + "learning_rate": 3.161765143541843e-05, + "loss": 1.439, + "step": 467 + }, + { + "epoch": 1.2804377564979479, + "grad_norm": 0.315186470746994, + "learning_rate": 3.154703767876942e-05, + "loss": 1.4901, + "step": 468 + }, + { + "epoch": 1.283173734610123, + "grad_norm": 0.32562458515167236, + "learning_rate": 3.1476367798609475e-05, + "loss": 1.4822, + "step": 469 + }, + { + "epoch": 1.2859097127222983, + "grad_norm": 0.31798624992370605, + "learning_rate": 3.1405642400745664e-05, + "loss": 1.5145, + "step": 470 + }, + { + "epoch": 1.2886456908344734, + "grad_norm": 0.29922667145729065, + "learning_rate": 3.133486209146099e-05, + "loss": 1.4984, + "step": 471 + }, + { + "epoch": 1.2913816689466484, + "grad_norm": 0.3259302079677582, + "learning_rate": 3.12640274775092e-05, + "loss": 1.5048, + "step": 472 + }, + { + "epoch": 1.2941176470588236, + "grad_norm": 0.2870180010795593, + "learning_rate": 3.119313916610948e-05, + "loss": 1.52, + "step": 473 + }, + { + "epoch": 1.2968536251709986, + "grad_norm": 0.3135444223880768, + "learning_rate": 3.112219776494142e-05, + "loss": 1.494, + "step": 474 + }, + { + "epoch": 1.2995896032831737, + "grad_norm": 0.29159173369407654, + "learning_rate": 3.105120388213966e-05, + "loss": 1.7019, + "step": 475 + }, + { + "epoch": 1.302325581395349, + "grad_norm": 0.31548362970352173, + "learning_rate": 3.098015812628875e-05, + "loss": 1.6465, + "step": 476 + }, + { + "epoch": 1.305061559507524, + "grad_norm": 0.2820277214050293, + "learning_rate": 3.090906110641791e-05, + "loss": 1.5705, + "step": 477 + }, + { + "epoch": 1.307797537619699, + "grad_norm": 0.2773056924343109, + "learning_rate": 3.083791343199582e-05, + "loss": 1.785, + "step": 478 + }, + { + "epoch": 1.310533515731874, + "grad_norm": 0.2650894522666931, + "learning_rate": 3.0766715712925384e-05, + "loss": 1.6026, + "step": 479 + }, + { + "epoch": 1.3132694938440492, + "grad_norm": 0.2996635138988495, + "learning_rate": 3.06954685595385e-05, + "loss": 1.489, + "step": 480 + }, + { + "epoch": 1.3160054719562244, + "grad_norm": 0.3386262059211731, + "learning_rate": 3.062417258259084e-05, + "loss": 1.538, + "step": 481 + }, + { + "epoch": 1.3187414500683994, + "grad_norm": 0.30783751606941223, + "learning_rate": 3.055282839325661e-05, + "loss": 1.42, + "step": 482 + }, + { + "epoch": 1.3214774281805746, + "grad_norm": 0.28174683451652527, + "learning_rate": 3.0481436603123292e-05, + "loss": 1.5068, + "step": 483 + }, + { + "epoch": 1.3242134062927495, + "grad_norm": 0.2959563732147217, + "learning_rate": 3.0409997824186453e-05, + "loss": 1.4343, + "step": 484 + }, + { + "epoch": 1.3269493844049247, + "grad_norm": 0.24625587463378906, + "learning_rate": 3.0338512668844443e-05, + "loss": 1.5942, + "step": 485 + }, + { + "epoch": 1.3296853625171, + "grad_norm": 0.3001210689544678, + "learning_rate": 3.0266981749893157e-05, + "loss": 1.5715, + "step": 486 + }, + { + "epoch": 1.332421340629275, + "grad_norm": 0.2976323962211609, + "learning_rate": 3.0195405680520828e-05, + "loss": 1.4471, + "step": 487 + }, + { + "epoch": 1.33515731874145, + "grad_norm": 0.3303026854991913, + "learning_rate": 3.012378507430269e-05, + "loss": 1.6357, + "step": 488 + }, + { + "epoch": 1.3378932968536252, + "grad_norm": 0.308095246553421, + "learning_rate": 3.005212054519579e-05, + "loss": 1.6527, + "step": 489 + }, + { + "epoch": 1.3406292749658002, + "grad_norm": 0.2990988790988922, + "learning_rate": 2.99804127075337e-05, + "loss": 1.5795, + "step": 490 + }, + { + "epoch": 1.3433652530779754, + "grad_norm": 0.29910552501678467, + "learning_rate": 2.9908662176021225e-05, + "loss": 1.6597, + "step": 491 + }, + { + "epoch": 1.3461012311901506, + "grad_norm": 0.28705233335494995, + "learning_rate": 2.9836869565729176e-05, + "loss": 1.4978, + "step": 492 + }, + { + "epoch": 1.3488372093023255, + "grad_norm": 0.33388686180114746, + "learning_rate": 2.9765035492089072e-05, + "loss": 1.4049, + "step": 493 + }, + { + "epoch": 1.3515731874145007, + "grad_norm": 0.28467613458633423, + "learning_rate": 2.9693160570887873e-05, + "loss": 1.4809, + "step": 494 + }, + { + "epoch": 1.3543091655266757, + "grad_norm": 0.2524206340312958, + "learning_rate": 2.9621245418262694e-05, + "loss": 1.5561, + "step": 495 + }, + { + "epoch": 1.3570451436388509, + "grad_norm": 0.27135151624679565, + "learning_rate": 2.954929065069554e-05, + "loss": 1.5686, + "step": 496 + }, + { + "epoch": 1.359781121751026, + "grad_norm": 0.31753405928611755, + "learning_rate": 2.9477296885007988e-05, + "loss": 1.4994, + "step": 497 + }, + { + "epoch": 1.362517099863201, + "grad_norm": 0.274442583322525, + "learning_rate": 2.9405264738355946e-05, + "loss": 1.5835, + "step": 498 + }, + { + "epoch": 1.3652530779753762, + "grad_norm": 0.3164355754852295, + "learning_rate": 2.9333194828224316e-05, + "loss": 1.4077, + "step": 499 + }, + { + "epoch": 1.3679890560875512, + "grad_norm": 0.3447605073451996, + "learning_rate": 2.926108777242172e-05, + "loss": 1.4768, + "step": 500 + }, + { + "epoch": 1.3679890560875512, + "eval_loss": 1.573856234550476, + "eval_runtime": 13.7722, + "eval_samples_per_second": 11.182, + "eval_steps_per_second": 11.182, + "step": 500 + }, + { + "epoch": 1.3707250341997264, + "grad_norm": 0.31595364212989807, + "learning_rate": 2.9188944189075235e-05, + "loss": 1.6381, + "step": 501 + }, + { + "epoch": 1.3734610123119015, + "grad_norm": 0.28893882036209106, + "learning_rate": 2.9116764696625033e-05, + "loss": 1.6677, + "step": 502 + }, + { + "epoch": 1.3761969904240767, + "grad_norm": 0.27313435077667236, + "learning_rate": 2.9044549913819124e-05, + "loss": 1.5746, + "step": 503 + }, + { + "epoch": 1.3789329685362517, + "grad_norm": 0.2701267600059509, + "learning_rate": 2.897230045970804e-05, + "loss": 1.2618, + "step": 504 + }, + { + "epoch": 1.3816689466484269, + "grad_norm": 0.30766165256500244, + "learning_rate": 2.890001695363953e-05, + "loss": 1.4354, + "step": 505 + }, + { + "epoch": 1.3844049247606018, + "grad_norm": 0.34688156843185425, + "learning_rate": 2.8827700015253246e-05, + "loss": 1.4171, + "step": 506 + }, + { + "epoch": 1.387140902872777, + "grad_norm": 0.28550222516059875, + "learning_rate": 2.875535026447543e-05, + "loss": 1.6484, + "step": 507 + }, + { + "epoch": 1.3898768809849522, + "grad_norm": 0.2944300174713135, + "learning_rate": 2.868296832151361e-05, + "loss": 1.6701, + "step": 508 + }, + { + "epoch": 1.3926128590971272, + "grad_norm": 0.2670783996582031, + "learning_rate": 2.8610554806851264e-05, + "loss": 1.5159, + "step": 509 + }, + { + "epoch": 1.3953488372093024, + "grad_norm": 0.2730109989643097, + "learning_rate": 2.853811034124253e-05, + "loss": 1.463, + "step": 510 + }, + { + "epoch": 1.3980848153214773, + "grad_norm": 0.2924908995628357, + "learning_rate": 2.8465635545706858e-05, + "loss": 1.6673, + "step": 511 + }, + { + "epoch": 1.4008207934336525, + "grad_norm": 0.2832358777523041, + "learning_rate": 2.8393131041523702e-05, + "loss": 1.4224, + "step": 512 + }, + { + "epoch": 1.4035567715458277, + "grad_norm": 0.3116619884967804, + "learning_rate": 2.8320597450227186e-05, + "loss": 1.4651, + "step": 513 + }, + { + "epoch": 1.4062927496580027, + "grad_norm": 0.2656640410423279, + "learning_rate": 2.824803539360078e-05, + "loss": 1.5117, + "step": 514 + }, + { + "epoch": 1.4090287277701778, + "grad_norm": 0.321814626455307, + "learning_rate": 2.8175445493671972e-05, + "loss": 1.5859, + "step": 515 + }, + { + "epoch": 1.4117647058823528, + "grad_norm": 0.3401627540588379, + "learning_rate": 2.8102828372706926e-05, + "loss": 1.6612, + "step": 516 + }, + { + "epoch": 1.414500683994528, + "grad_norm": 0.3001568019390106, + "learning_rate": 2.803018465320515e-05, + "loss": 1.655, + "step": 517 + }, + { + "epoch": 1.4172366621067032, + "grad_norm": 0.27258241176605225, + "learning_rate": 2.795751495789418e-05, + "loss": 1.6612, + "step": 518 + }, + { + "epoch": 1.4199726402188784, + "grad_norm": 0.2985587418079376, + "learning_rate": 2.7884819909724224e-05, + "loss": 1.452, + "step": 519 + }, + { + "epoch": 1.4227086183310533, + "grad_norm": 0.2565891146659851, + "learning_rate": 2.78121001318628e-05, + "loss": 1.6304, + "step": 520 + }, + { + "epoch": 1.4254445964432285, + "grad_norm": 0.29893383383750916, + "learning_rate": 2.7739356247689446e-05, + "loss": 1.4655, + "step": 521 + }, + { + "epoch": 1.4281805745554035, + "grad_norm": 0.28050941228866577, + "learning_rate": 2.7666588880790335e-05, + "loss": 1.5134, + "step": 522 + }, + { + "epoch": 1.4309165526675787, + "grad_norm": 0.29267653822898865, + "learning_rate": 2.759379865495294e-05, + "loss": 1.4355, + "step": 523 + }, + { + "epoch": 1.4336525307797539, + "grad_norm": 0.279461145401001, + "learning_rate": 2.752098619416069e-05, + "loss": 1.5606, + "step": 524 + }, + { + "epoch": 1.4363885088919288, + "grad_norm": 0.260345995426178, + "learning_rate": 2.7448152122587634e-05, + "loss": 1.702, + "step": 525 + }, + { + "epoch": 1.439124487004104, + "grad_norm": 0.2932511270046234, + "learning_rate": 2.7375297064593063e-05, + "loss": 1.3368, + "step": 526 + }, + { + "epoch": 1.441860465116279, + "grad_norm": 0.3288111686706543, + "learning_rate": 2.730242164471616e-05, + "loss": 1.4919, + "step": 527 + }, + { + "epoch": 1.4445964432284542, + "grad_norm": 0.27260783314704895, + "learning_rate": 2.7229526487670676e-05, + "loss": 1.5306, + "step": 528 + }, + { + "epoch": 1.4473324213406293, + "grad_norm": 0.2748279273509979, + "learning_rate": 2.7156612218339544e-05, + "loss": 1.5238, + "step": 529 + }, + { + "epoch": 1.4500683994528043, + "grad_norm": 0.26095664501190186, + "learning_rate": 2.708367946176956e-05, + "loss": 1.5456, + "step": 530 + }, + { + "epoch": 1.4528043775649795, + "grad_norm": 0.3169284164905548, + "learning_rate": 2.701072884316595e-05, + "loss": 1.6298, + "step": 531 + }, + { + "epoch": 1.4555403556771545, + "grad_norm": 0.2977512776851654, + "learning_rate": 2.6937760987887112e-05, + "loss": 1.4394, + "step": 532 + }, + { + "epoch": 1.4582763337893296, + "grad_norm": 0.2777084708213806, + "learning_rate": 2.6864776521439166e-05, + "loss": 1.5062, + "step": 533 + }, + { + "epoch": 1.4610123119015048, + "grad_norm": 0.29830804467201233, + "learning_rate": 2.6791776069470658e-05, + "loss": 1.5056, + "step": 534 + }, + { + "epoch": 1.46374829001368, + "grad_norm": 0.35224923491477966, + "learning_rate": 2.6718760257767135e-05, + "loss": 1.4127, + "step": 535 + }, + { + "epoch": 1.466484268125855, + "grad_norm": 0.31385019421577454, + "learning_rate": 2.6645729712245832e-05, + "loss": 1.6469, + "step": 536 + }, + { + "epoch": 1.4692202462380302, + "grad_norm": 0.3666173815727234, + "learning_rate": 2.6572685058950298e-05, + "loss": 1.6698, + "step": 537 + }, + { + "epoch": 1.4719562243502051, + "grad_norm": 0.29228919744491577, + "learning_rate": 2.649962692404499e-05, + "loss": 1.5971, + "step": 538 + }, + { + "epoch": 1.4746922024623803, + "grad_norm": 0.2931617796421051, + "learning_rate": 2.6426555933809954e-05, + "loss": 1.4914, + "step": 539 + }, + { + "epoch": 1.4774281805745555, + "grad_norm": 0.2743561863899231, + "learning_rate": 2.635347271463544e-05, + "loss": 1.5475, + "step": 540 + }, + { + "epoch": 1.4801641586867305, + "grad_norm": 0.3290240466594696, + "learning_rate": 2.6280377893016516e-05, + "loss": 1.4155, + "step": 541 + }, + { + "epoch": 1.4829001367989056, + "grad_norm": 0.30161550641059875, + "learning_rate": 2.6207272095547718e-05, + "loss": 1.4564, + "step": 542 + }, + { + "epoch": 1.4856361149110806, + "grad_norm": 0.3175734877586365, + "learning_rate": 2.613415594891767e-05, + "loss": 1.5312, + "step": 543 + }, + { + "epoch": 1.4883720930232558, + "grad_norm": 0.3023424744606018, + "learning_rate": 2.606103007990371e-05, + "loss": 1.4688, + "step": 544 + }, + { + "epoch": 1.491108071135431, + "grad_norm": 0.3181714415550232, + "learning_rate": 2.5987895115366516e-05, + "loss": 1.738, + "step": 545 + }, + { + "epoch": 1.493844049247606, + "grad_norm": 0.2678375244140625, + "learning_rate": 2.5914751682244748e-05, + "loss": 1.5491, + "step": 546 + }, + { + "epoch": 1.4965800273597811, + "grad_norm": 0.3041532039642334, + "learning_rate": 2.5841600407549642e-05, + "loss": 1.6654, + "step": 547 + }, + { + "epoch": 1.499316005471956, + "grad_norm": 0.30127188563346863, + "learning_rate": 2.5768441918359692e-05, + "loss": 1.5671, + "step": 548 + }, + { + "epoch": 1.5020519835841313, + "grad_norm": 0.34145042300224304, + "learning_rate": 2.5695276841815186e-05, + "loss": 1.6131, + "step": 549 + }, + { + "epoch": 1.5047879616963065, + "grad_norm": 0.2628389000892639, + "learning_rate": 2.562210580511291e-05, + "loss": 1.7636, + "step": 550 + }, + { + "epoch": 1.5075239398084817, + "grad_norm": 0.2784722149372101, + "learning_rate": 2.5548929435500758e-05, + "loss": 1.4967, + "step": 551 + }, + { + "epoch": 1.5102599179206566, + "grad_norm": 0.29418709874153137, + "learning_rate": 2.547574836027231e-05, + "loss": 1.5771, + "step": 552 + }, + { + "epoch": 1.5129958960328316, + "grad_norm": 0.29850900173187256, + "learning_rate": 2.54025632067615e-05, + "loss": 1.4134, + "step": 553 + }, + { + "epoch": 1.5157318741450068, + "grad_norm": 0.29349079728126526, + "learning_rate": 2.5329374602337215e-05, + "loss": 1.4498, + "step": 554 + }, + { + "epoch": 1.518467852257182, + "grad_norm": 0.27603679895401, + "learning_rate": 2.525618317439793e-05, + "loss": 1.5103, + "step": 555 + }, + { + "epoch": 1.5212038303693571, + "grad_norm": 0.29642152786254883, + "learning_rate": 2.518298955036632e-05, + "loss": 1.5597, + "step": 556 + }, + { + "epoch": 1.523939808481532, + "grad_norm": 0.3032752573490143, + "learning_rate": 2.5109794357683885e-05, + "loss": 1.5697, + "step": 557 + }, + { + "epoch": 1.5266757865937073, + "grad_norm": 0.3295765519142151, + "learning_rate": 2.503659822380558e-05, + "loss": 1.6343, + "step": 558 + }, + { + "epoch": 1.5294117647058822, + "grad_norm": 0.28324607014656067, + "learning_rate": 2.496340177619442e-05, + "loss": 1.551, + "step": 559 + }, + { + "epoch": 1.5321477428180574, + "grad_norm": 0.2895617187023163, + "learning_rate": 2.489020564231612e-05, + "loss": 1.6563, + "step": 560 + }, + { + "epoch": 1.5348837209302326, + "grad_norm": 0.3268161714076996, + "learning_rate": 2.4817010449633688e-05, + "loss": 1.6045, + "step": 561 + }, + { + "epoch": 1.5376196990424078, + "grad_norm": 0.27146685123443604, + "learning_rate": 2.474381682560208e-05, + "loss": 1.5727, + "step": 562 + }, + { + "epoch": 1.5403556771545828, + "grad_norm": 0.27629193663597107, + "learning_rate": 2.4670625397662787e-05, + "loss": 1.4352, + "step": 563 + }, + { + "epoch": 1.5430916552667577, + "grad_norm": 0.27017953991889954, + "learning_rate": 2.4597436793238506e-05, + "loss": 1.646, + "step": 564 + }, + { + "epoch": 1.545827633378933, + "grad_norm": 0.289787232875824, + "learning_rate": 2.45242516397277e-05, + "loss": 1.7107, + "step": 565 + }, + { + "epoch": 1.548563611491108, + "grad_norm": 0.2762577533721924, + "learning_rate": 2.4451070564499245e-05, + "loss": 1.5694, + "step": 566 + }, + { + "epoch": 1.5512995896032833, + "grad_norm": 0.2959858179092407, + "learning_rate": 2.4377894194887095e-05, + "loss": 1.639, + "step": 567 + }, + { + "epoch": 1.5540355677154583, + "grad_norm": 0.28595441579818726, + "learning_rate": 2.4304723158184827e-05, + "loss": 1.3091, + "step": 568 + }, + { + "epoch": 1.5567715458276332, + "grad_norm": 0.3006284236907959, + "learning_rate": 2.4231558081640314e-05, + "loss": 1.569, + "step": 569 + }, + { + "epoch": 1.5595075239398084, + "grad_norm": 0.2800249457359314, + "learning_rate": 2.415839959245036e-05, + "loss": 1.5124, + "step": 570 + }, + { + "epoch": 1.5622435020519836, + "grad_norm": 0.3035907745361328, + "learning_rate": 2.4085248317755254e-05, + "loss": 1.5785, + "step": 571 + }, + { + "epoch": 1.5649794801641588, + "grad_norm": 0.3111293315887451, + "learning_rate": 2.4012104884633486e-05, + "loss": 1.5544, + "step": 572 + }, + { + "epoch": 1.5677154582763337, + "grad_norm": 0.31877052783966064, + "learning_rate": 2.39389699200963e-05, + "loss": 1.482, + "step": 573 + }, + { + "epoch": 1.570451436388509, + "grad_norm": 0.3056713342666626, + "learning_rate": 2.386584405108233e-05, + "loss": 1.5234, + "step": 574 + }, + { + "epoch": 1.573187414500684, + "grad_norm": 0.289650022983551, + "learning_rate": 2.3792727904452285e-05, + "loss": 1.398, + "step": 575 + }, + { + "epoch": 1.575923392612859, + "grad_norm": 0.28617268800735474, + "learning_rate": 2.3719622106983486e-05, + "loss": 1.5721, + "step": 576 + }, + { + "epoch": 1.5786593707250343, + "grad_norm": 0.29906001687049866, + "learning_rate": 2.3646527285364565e-05, + "loss": 1.3543, + "step": 577 + }, + { + "epoch": 1.5813953488372094, + "grad_norm": 0.2876405417919159, + "learning_rate": 2.3573444066190052e-05, + "loss": 1.5365, + "step": 578 + }, + { + "epoch": 1.5841313269493844, + "grad_norm": 0.3173394799232483, + "learning_rate": 2.3500373075955022e-05, + "loss": 1.4437, + "step": 579 + }, + { + "epoch": 1.5868673050615594, + "grad_norm": 0.2820620834827423, + "learning_rate": 2.342731494104971e-05, + "loss": 1.5992, + "step": 580 + }, + { + "epoch": 1.5896032831737346, + "grad_norm": 0.3694467842578888, + "learning_rate": 2.3354270287754174e-05, + "loss": 1.5443, + "step": 581 + }, + { + "epoch": 1.5923392612859097, + "grad_norm": 0.2960888147354126, + "learning_rate": 2.328123974223288e-05, + "loss": 1.5173, + "step": 582 + }, + { + "epoch": 1.595075239398085, + "grad_norm": 0.3080427348613739, + "learning_rate": 2.3208223930529347e-05, + "loss": 1.749, + "step": 583 + }, + { + "epoch": 1.59781121751026, + "grad_norm": 0.29521963000297546, + "learning_rate": 2.313522347856084e-05, + "loss": 1.522, + "step": 584 + }, + { + "epoch": 1.6005471956224349, + "grad_norm": 0.32432520389556885, + "learning_rate": 2.306223901211289e-05, + "loss": 1.6394, + "step": 585 + }, + { + "epoch": 1.60328317373461, + "grad_norm": 0.326926052570343, + "learning_rate": 2.2989271156834057e-05, + "loss": 1.5092, + "step": 586 + }, + { + "epoch": 1.6060191518467852, + "grad_norm": 0.2793523967266083, + "learning_rate": 2.291632053823045e-05, + "loss": 1.5266, + "step": 587 + }, + { + "epoch": 1.6087551299589604, + "grad_norm": 0.3044006824493408, + "learning_rate": 2.2843387781660452e-05, + "loss": 1.7137, + "step": 588 + }, + { + "epoch": 1.6114911080711354, + "grad_norm": 0.33133581280708313, + "learning_rate": 2.2770473512329333e-05, + "loss": 1.5494, + "step": 589 + }, + { + "epoch": 1.6142270861833106, + "grad_norm": 0.30066078901290894, + "learning_rate": 2.269757835528385e-05, + "loss": 1.387, + "step": 590 + }, + { + "epoch": 1.6169630642954855, + "grad_norm": 0.29674389958381653, + "learning_rate": 2.2624702935406943e-05, + "loss": 1.6985, + "step": 591 + }, + { + "epoch": 1.6196990424076607, + "grad_norm": 0.2950797975063324, + "learning_rate": 2.255184787741237e-05, + "loss": 1.558, + "step": 592 + }, + { + "epoch": 1.622435020519836, + "grad_norm": 0.32414302229881287, + "learning_rate": 2.2479013805839318e-05, + "loss": 1.4547, + "step": 593 + }, + { + "epoch": 1.625170998632011, + "grad_norm": 0.2893849015235901, + "learning_rate": 2.240620134504707e-05, + "loss": 1.6205, + "step": 594 + }, + { + "epoch": 1.627906976744186, + "grad_norm": 0.30795803666114807, + "learning_rate": 2.233341111920967e-05, + "loss": 1.4725, + "step": 595 + }, + { + "epoch": 1.630642954856361, + "grad_norm": 0.28829425573349, + "learning_rate": 2.226064375231056e-05, + "loss": 1.5676, + "step": 596 + }, + { + "epoch": 1.6333789329685362, + "grad_norm": 0.32631853222846985, + "learning_rate": 2.2187899868137206e-05, + "loss": 1.447, + "step": 597 + }, + { + "epoch": 1.6361149110807114, + "grad_norm": 0.2999882996082306, + "learning_rate": 2.211518009027579e-05, + "loss": 1.4184, + "step": 598 + }, + { + "epoch": 1.6388508891928866, + "grad_norm": 0.302869975566864, + "learning_rate": 2.204248504210582e-05, + "loss": 1.5356, + "step": 599 + }, + { + "epoch": 1.6415868673050615, + "grad_norm": 0.32066816091537476, + "learning_rate": 2.1969815346794857e-05, + "loss": 1.507, + "step": 600 + }, + { + "epoch": 1.6443228454172365, + "grad_norm": 0.26394781470298767, + "learning_rate": 2.189717162729309e-05, + "loss": 1.519, + "step": 601 + }, + { + "epoch": 1.6470588235294117, + "grad_norm": 0.28778427839279175, + "learning_rate": 2.182455450632803e-05, + "loss": 1.6243, + "step": 602 + }, + { + "epoch": 1.6497948016415869, + "grad_norm": 0.28757721185684204, + "learning_rate": 2.1751964606399224e-05, + "loss": 1.6119, + "step": 603 + }, + { + "epoch": 1.652530779753762, + "grad_norm": 0.32660529017448425, + "learning_rate": 2.167940254977282e-05, + "loss": 1.481, + "step": 604 + }, + { + "epoch": 1.655266757865937, + "grad_norm": 0.2857491075992584, + "learning_rate": 2.1606868958476304e-05, + "loss": 1.4287, + "step": 605 + }, + { + "epoch": 1.6580027359781122, + "grad_norm": 0.3346758782863617, + "learning_rate": 2.1534364454293148e-05, + "loss": 1.6915, + "step": 606 + }, + { + "epoch": 1.6607387140902872, + "grad_norm": 0.2858891189098358, + "learning_rate": 2.146188965875747e-05, + "loss": 1.4857, + "step": 607 + }, + { + "epoch": 1.6634746922024624, + "grad_norm": 0.2926919758319855, + "learning_rate": 2.1389445193148742e-05, + "loss": 1.5348, + "step": 608 + }, + { + "epoch": 1.6662106703146375, + "grad_norm": 0.3735974133014679, + "learning_rate": 2.1317031678486402e-05, + "loss": 1.697, + "step": 609 + }, + { + "epoch": 1.6689466484268127, + "grad_norm": 0.3130210340023041, + "learning_rate": 2.124464973552457e-05, + "loss": 1.6426, + "step": 610 + }, + { + "epoch": 1.6716826265389877, + "grad_norm": 0.36122313141822815, + "learning_rate": 2.117229998474676e-05, + "loss": 1.7516, + "step": 611 + }, + { + "epoch": 1.6744186046511627, + "grad_norm": 0.2955617904663086, + "learning_rate": 2.109998304636048e-05, + "loss": 1.4681, + "step": 612 + }, + { + "epoch": 1.6771545827633378, + "grad_norm": 0.3331190347671509, + "learning_rate": 2.1027699540291965e-05, + "loss": 1.5334, + "step": 613 + }, + { + "epoch": 1.679890560875513, + "grad_norm": 0.39551764726638794, + "learning_rate": 2.0955450086180882e-05, + "loss": 1.497, + "step": 614 + }, + { + "epoch": 1.6826265389876882, + "grad_norm": 0.2910989224910736, + "learning_rate": 2.088323530337498e-05, + "loss": 1.4593, + "step": 615 + }, + { + "epoch": 1.6853625170998632, + "grad_norm": 0.2927062511444092, + "learning_rate": 2.0811055810924768e-05, + "loss": 1.5019, + "step": 616 + }, + { + "epoch": 1.6880984952120381, + "grad_norm": 0.2986646592617035, + "learning_rate": 2.0738912227578283e-05, + "loss": 1.5167, + "step": 617 + }, + { + "epoch": 1.6908344733242133, + "grad_norm": 0.2710409462451935, + "learning_rate": 2.0666805171775687e-05, + "loss": 1.5491, + "step": 618 + }, + { + "epoch": 1.6935704514363885, + "grad_norm": 0.32555902004241943, + "learning_rate": 2.0594735261644056e-05, + "loss": 1.5945, + "step": 619 + }, + { + "epoch": 1.6963064295485637, + "grad_norm": 0.3129582703113556, + "learning_rate": 2.0522703114992018e-05, + "loss": 1.5388, + "step": 620 + }, + { + "epoch": 1.6990424076607387, + "grad_norm": 0.29060521721839905, + "learning_rate": 2.0450709349304463e-05, + "loss": 1.5335, + "step": 621 + }, + { + "epoch": 1.7017783857729138, + "grad_norm": 0.2991231381893158, + "learning_rate": 2.0378754581737308e-05, + "loss": 1.5857, + "step": 622 + }, + { + "epoch": 1.7045143638850888, + "grad_norm": 0.3089081346988678, + "learning_rate": 2.0306839429112136e-05, + "loss": 1.509, + "step": 623 + }, + { + "epoch": 1.707250341997264, + "grad_norm": 0.30709123611450195, + "learning_rate": 2.023496450791093e-05, + "loss": 1.5935, + "step": 624 + }, + { + "epoch": 1.7099863201094392, + "grad_norm": 0.29340359568595886, + "learning_rate": 2.0163130434270833e-05, + "loss": 1.547, + "step": 625 + }, + { + "epoch": 1.7127222982216144, + "grad_norm": 0.3518681824207306, + "learning_rate": 2.009133782397879e-05, + "loss": 1.513, + "step": 626 + }, + { + "epoch": 1.7154582763337893, + "grad_norm": 0.31759360432624817, + "learning_rate": 2.0019587292466306e-05, + "loss": 1.5252, + "step": 627 + }, + { + "epoch": 1.7181942544459643, + "grad_norm": 0.2931315302848816, + "learning_rate": 1.9947879454804216e-05, + "loss": 1.7602, + "step": 628 + }, + { + "epoch": 1.7209302325581395, + "grad_norm": 0.2938210368156433, + "learning_rate": 1.9876214925697323e-05, + "loss": 1.5154, + "step": 629 + }, + { + "epoch": 1.7236662106703147, + "grad_norm": 0.28195831179618835, + "learning_rate": 1.980459431947918e-05, + "loss": 1.4129, + "step": 630 + }, + { + "epoch": 1.7264021887824899, + "grad_norm": 0.31024137139320374, + "learning_rate": 1.973301825010685e-05, + "loss": 1.5008, + "step": 631 + }, + { + "epoch": 1.7291381668946648, + "grad_norm": 0.3060285747051239, + "learning_rate": 1.9661487331155563e-05, + "loss": 1.5469, + "step": 632 + }, + { + "epoch": 1.7318741450068398, + "grad_norm": 0.38896477222442627, + "learning_rate": 1.9590002175813553e-05, + "loss": 1.6018, + "step": 633 + }, + { + "epoch": 1.734610123119015, + "grad_norm": 0.32051512598991394, + "learning_rate": 1.9518563396876717e-05, + "loss": 1.5078, + "step": 634 + }, + { + "epoch": 1.7373461012311902, + "grad_norm": 0.2726178467273712, + "learning_rate": 1.9447171606743398e-05, + "loss": 1.4586, + "step": 635 + }, + { + "epoch": 1.7400820793433653, + "grad_norm": 0.2772533595561981, + "learning_rate": 1.9375827417409165e-05, + "loss": 1.5022, + "step": 636 + }, + { + "epoch": 1.7428180574555403, + "grad_norm": 0.30202561616897583, + "learning_rate": 1.9304531440461506e-05, + "loss": 1.5036, + "step": 637 + }, + { + "epoch": 1.7455540355677155, + "grad_norm": 0.2932077944278717, + "learning_rate": 1.923328428707461e-05, + "loss": 1.5541, + "step": 638 + }, + { + "epoch": 1.7482900136798905, + "grad_norm": 0.3096188008785248, + "learning_rate": 1.916208656800418e-05, + "loss": 1.3744, + "step": 639 + }, + { + "epoch": 1.7510259917920656, + "grad_norm": 0.282226026058197, + "learning_rate": 1.9090938893582088e-05, + "loss": 1.6306, + "step": 640 + }, + { + "epoch": 1.7537619699042408, + "grad_norm": 0.27213621139526367, + "learning_rate": 1.9019841873711255e-05, + "loss": 1.4953, + "step": 641 + }, + { + "epoch": 1.756497948016416, + "grad_norm": 0.30166521668434143, + "learning_rate": 1.8948796117860348e-05, + "loss": 1.5158, + "step": 642 + }, + { + "epoch": 1.759233926128591, + "grad_norm": 0.27089768648147583, + "learning_rate": 1.8877802235058585e-05, + "loss": 1.4954, + "step": 643 + }, + { + "epoch": 1.761969904240766, + "grad_norm": 0.31477153301239014, + "learning_rate": 1.8806860833890528e-05, + "loss": 1.5449, + "step": 644 + }, + { + "epoch": 1.7647058823529411, + "grad_norm": 0.2903262674808502, + "learning_rate": 1.8735972522490818e-05, + "loss": 1.4129, + "step": 645 + }, + { + "epoch": 1.7674418604651163, + "grad_norm": 0.3140583038330078, + "learning_rate": 1.8665137908539004e-05, + "loss": 1.6374, + "step": 646 + }, + { + "epoch": 1.7701778385772915, + "grad_norm": 0.29897943139076233, + "learning_rate": 1.859435759925434e-05, + "loss": 1.605, + "step": 647 + }, + { + "epoch": 1.7729138166894665, + "grad_norm": 0.2993355989456177, + "learning_rate": 1.8523632201390537e-05, + "loss": 1.495, + "step": 648 + }, + { + "epoch": 1.7756497948016414, + "grad_norm": 0.3301627039909363, + "learning_rate": 1.845296232123058e-05, + "loss": 1.6064, + "step": 649 + }, + { + "epoch": 1.7783857729138166, + "grad_norm": 0.30024030804634094, + "learning_rate": 1.8382348564581574e-05, + "loss": 1.3258, + "step": 650 + }, + { + "epoch": 1.7811217510259918, + "grad_norm": 0.3279036283493042, + "learning_rate": 1.8311791536769483e-05, + "loss": 1.4348, + "step": 651 + }, + { + "epoch": 1.783857729138167, + "grad_norm": 0.3231453001499176, + "learning_rate": 1.8241291842633966e-05, + "loss": 1.4327, + "step": 652 + }, + { + "epoch": 1.786593707250342, + "grad_norm": 0.3033272624015808, + "learning_rate": 1.817085008652324e-05, + "loss": 1.4668, + "step": 653 + }, + { + "epoch": 1.7893296853625171, + "grad_norm": 0.28523799777030945, + "learning_rate": 1.810046687228879e-05, + "loss": 1.5058, + "step": 654 + }, + { + "epoch": 1.792065663474692, + "grad_norm": 0.35666465759277344, + "learning_rate": 1.803014280328033e-05, + "loss": 1.4551, + "step": 655 + }, + { + "epoch": 1.7948016415868673, + "grad_norm": 0.33736082911491394, + "learning_rate": 1.795987848234051e-05, + "loss": 1.2991, + "step": 656 + }, + { + "epoch": 1.7975376196990425, + "grad_norm": 0.3194032311439514, + "learning_rate": 1.7889674511799812e-05, + "loss": 1.4764, + "step": 657 + }, + { + "epoch": 1.8002735978112177, + "grad_norm": 0.28460192680358887, + "learning_rate": 1.7819531493471392e-05, + "loss": 1.4962, + "step": 658 + }, + { + "epoch": 1.8030095759233926, + "grad_norm": 0.3380921483039856, + "learning_rate": 1.7749450028645875e-05, + "loss": 1.4772, + "step": 659 + }, + { + "epoch": 1.8057455540355676, + "grad_norm": 0.32514598965644836, + "learning_rate": 1.7679430718086243e-05, + "loss": 1.6421, + "step": 660 + }, + { + "epoch": 1.8084815321477428, + "grad_norm": 0.293314129114151, + "learning_rate": 1.7609474162022665e-05, + "loss": 1.5151, + "step": 661 + }, + { + "epoch": 1.811217510259918, + "grad_norm": 0.3001095950603485, + "learning_rate": 1.753958096014737e-05, + "loss": 1.4422, + "step": 662 + }, + { + "epoch": 1.8139534883720931, + "grad_norm": 0.3403547704219818, + "learning_rate": 1.7469751711609454e-05, + "loss": 1.6122, + "step": 663 + }, + { + "epoch": 1.816689466484268, + "grad_norm": 0.3033615052700043, + "learning_rate": 1.739998701500984e-05, + "loss": 1.596, + "step": 664 + }, + { + "epoch": 1.819425444596443, + "grad_norm": 0.34662926197052, + "learning_rate": 1.7330287468396033e-05, + "loss": 1.5458, + "step": 665 + }, + { + "epoch": 1.8221614227086183, + "grad_norm": 0.32165685296058655, + "learning_rate": 1.7260653669257093e-05, + "loss": 1.5433, + "step": 666 + }, + { + "epoch": 1.8248974008207934, + "grad_norm": 0.32157987356185913, + "learning_rate": 1.7191086214518447e-05, + "loss": 1.6236, + "step": 667 + }, + { + "epoch": 1.8276333789329686, + "grad_norm": 0.3016352355480194, + "learning_rate": 1.712158570053678e-05, + "loss": 1.3893, + "step": 668 + }, + { + "epoch": 1.8303693570451436, + "grad_norm": 0.40459707379341125, + "learning_rate": 1.7052152723094976e-05, + "loss": 1.3767, + "step": 669 + }, + { + "epoch": 1.8331053351573188, + "grad_norm": 0.34724652767181396, + "learning_rate": 1.698278787739695e-05, + "loss": 1.6067, + "step": 670 + }, + { + "epoch": 1.8358413132694937, + "grad_norm": 0.3703734874725342, + "learning_rate": 1.6913491758062557e-05, + "loss": 1.4538, + "step": 671 + }, + { + "epoch": 1.838577291381669, + "grad_norm": 0.31886354088783264, + "learning_rate": 1.6844264959122534e-05, + "loss": 1.6776, + "step": 672 + }, + { + "epoch": 1.841313269493844, + "grad_norm": 0.32411274313926697, + "learning_rate": 1.6775108074013356e-05, + "loss": 1.5405, + "step": 673 + }, + { + "epoch": 1.8440492476060193, + "grad_norm": 0.32398101687431335, + "learning_rate": 1.670602169557217e-05, + "loss": 1.5429, + "step": 674 + }, + { + "epoch": 1.8467852257181943, + "grad_norm": 0.3220042288303375, + "learning_rate": 1.663700641603174e-05, + "loss": 1.5339, + "step": 675 + }, + { + "epoch": 1.8495212038303692, + "grad_norm": 0.2694465219974518, + "learning_rate": 1.6568062827015317e-05, + "loss": 1.5641, + "step": 676 + }, + { + "epoch": 1.8522571819425444, + "grad_norm": 0.3089315593242645, + "learning_rate": 1.6499191519531614e-05, + "loss": 1.3319, + "step": 677 + }, + { + "epoch": 1.8549931600547196, + "grad_norm": 0.3302357792854309, + "learning_rate": 1.643039308396971e-05, + "loss": 1.478, + "step": 678 + }, + { + "epoch": 1.8577291381668948, + "grad_norm": 0.30908071994781494, + "learning_rate": 1.6361668110094007e-05, + "loss": 1.5557, + "step": 679 + }, + { + "epoch": 1.8604651162790697, + "grad_norm": 0.2804611027240753, + "learning_rate": 1.6293017187039174e-05, + "loss": 1.6756, + "step": 680 + }, + { + "epoch": 1.8632010943912447, + "grad_norm": 0.3693443238735199, + "learning_rate": 1.6224440903305076e-05, + "loss": 1.5825, + "step": 681 + }, + { + "epoch": 1.86593707250342, + "grad_norm": 0.33380305767059326, + "learning_rate": 1.615593984675174e-05, + "loss": 1.5312, + "step": 682 + }, + { + "epoch": 1.868673050615595, + "grad_norm": 0.3250941038131714, + "learning_rate": 1.6087514604594353e-05, + "loss": 1.5642, + "step": 683 + }, + { + "epoch": 1.8714090287277703, + "grad_norm": 0.3749421238899231, + "learning_rate": 1.6019165763398152e-05, + "loss": 1.6141, + "step": 684 + }, + { + "epoch": 1.8741450068399452, + "grad_norm": 0.32209932804107666, + "learning_rate": 1.5950893909073453e-05, + "loss": 1.4791, + "step": 685 + }, + { + "epoch": 1.8768809849521204, + "grad_norm": 0.34071722626686096, + "learning_rate": 1.5882699626870633e-05, + "loss": 1.6126, + "step": 686 + }, + { + "epoch": 1.8796169630642954, + "grad_norm": 0.2984413802623749, + "learning_rate": 1.5814583501375064e-05, + "loss": 1.5191, + "step": 687 + }, + { + "epoch": 1.8823529411764706, + "grad_norm": 0.33248278498649597, + "learning_rate": 1.574654611650214e-05, + "loss": 1.6346, + "step": 688 + }, + { + "epoch": 1.8850889192886457, + "grad_norm": 0.308098703622818, + "learning_rate": 1.567858805549229e-05, + "loss": 1.5521, + "step": 689 + }, + { + "epoch": 1.887824897400821, + "grad_norm": 0.2899650037288666, + "learning_rate": 1.56107099009059e-05, + "loss": 1.3675, + "step": 690 + }, + { + "epoch": 1.890560875512996, + "grad_norm": 0.33524468541145325, + "learning_rate": 1.5542912234618422e-05, + "loss": 1.6501, + "step": 691 + }, + { + "epoch": 1.8932968536251709, + "grad_norm": 0.3770296275615692, + "learning_rate": 1.54751956378153e-05, + "loss": 1.5518, + "step": 692 + }, + { + "epoch": 1.896032831737346, + "grad_norm": 0.3348696529865265, + "learning_rate": 1.540756069098702e-05, + "loss": 1.4388, + "step": 693 + }, + { + "epoch": 1.8987688098495212, + "grad_norm": 0.37193214893341064, + "learning_rate": 1.5340007973924176e-05, + "loss": 1.499, + "step": 694 + }, + { + "epoch": 1.9015047879616964, + "grad_norm": 0.33301231265068054, + "learning_rate": 1.5272538065712403e-05, + "loss": 1.6687, + "step": 695 + }, + { + "epoch": 1.9042407660738714, + "grad_norm": 0.33597227931022644, + "learning_rate": 1.5205151544727509e-05, + "loss": 1.589, + "step": 696 + }, + { + "epoch": 1.9069767441860463, + "grad_norm": 0.3002086281776428, + "learning_rate": 1.5137848988630465e-05, + "loss": 1.7462, + "step": 697 + }, + { + "epoch": 1.9097127222982215, + "grad_norm": 0.30558449029922485, + "learning_rate": 1.5070630974362473e-05, + "loss": 1.6179, + "step": 698 + }, + { + "epoch": 1.9124487004103967, + "grad_norm": 0.314813494682312, + "learning_rate": 1.5003498078139988e-05, + "loss": 1.6029, + "step": 699 + }, + { + "epoch": 1.915184678522572, + "grad_norm": 0.31106236577033997, + "learning_rate": 1.4936450875449845e-05, + "loss": 1.5462, + "step": 700 + }, + { + "epoch": 1.9179206566347469, + "grad_norm": 0.32419320940971375, + "learning_rate": 1.4869489941044235e-05, + "loss": 1.632, + "step": 701 + }, + { + "epoch": 1.920656634746922, + "grad_norm": 0.28728801012039185, + "learning_rate": 1.4802615848935875e-05, + "loss": 1.5625, + "step": 702 + }, + { + "epoch": 1.923392612859097, + "grad_norm": 0.35819682478904724, + "learning_rate": 1.4735829172393007e-05, + "loss": 1.416, + "step": 703 + }, + { + "epoch": 1.9261285909712722, + "grad_norm": 0.3097285032272339, + "learning_rate": 1.4669130483934512e-05, + "loss": 1.4333, + "step": 704 + }, + { + "epoch": 1.9288645690834474, + "grad_norm": 0.31058430671691895, + "learning_rate": 1.4602520355325039e-05, + "loss": 1.5175, + "step": 705 + }, + { + "epoch": 1.9316005471956226, + "grad_norm": 0.3093342185020447, + "learning_rate": 1.4535999357570046e-05, + "loss": 1.4105, + "step": 706 + }, + { + "epoch": 1.9343365253077975, + "grad_norm": 0.3064606189727783, + "learning_rate": 1.4469568060910915e-05, + "loss": 1.6593, + "step": 707 + }, + { + "epoch": 1.9370725034199725, + "grad_norm": 0.3192756772041321, + "learning_rate": 1.4403227034820139e-05, + "loss": 1.5919, + "step": 708 + }, + { + "epoch": 1.9398084815321477, + "grad_norm": 0.30623388290405273, + "learning_rate": 1.4336976847996303e-05, + "loss": 1.5053, + "step": 709 + }, + { + "epoch": 1.9425444596443229, + "grad_norm": 0.3014221787452698, + "learning_rate": 1.4270818068359336e-05, + "loss": 1.4876, + "step": 710 + }, + { + "epoch": 1.945280437756498, + "grad_norm": 0.34191763401031494, + "learning_rate": 1.420475126304558e-05, + "loss": 1.435, + "step": 711 + }, + { + "epoch": 1.948016415868673, + "grad_norm": 0.33173930644989014, + "learning_rate": 1.4138776998402927e-05, + "loss": 1.5467, + "step": 712 + }, + { + "epoch": 1.950752393980848, + "grad_norm": 0.3360328674316406, + "learning_rate": 1.4072895839986023e-05, + "loss": 1.4525, + "step": 713 + }, + { + "epoch": 1.9534883720930232, + "grad_norm": 0.2906378209590912, + "learning_rate": 1.4007108352551313e-05, + "loss": 1.5788, + "step": 714 + }, + { + "epoch": 1.9562243502051984, + "grad_norm": 0.3077907860279083, + "learning_rate": 1.3941415100052293e-05, + "loss": 1.4087, + "step": 715 + }, + { + "epoch": 1.9589603283173735, + "grad_norm": 0.2951776385307312, + "learning_rate": 1.387581664563465e-05, + "loss": 1.4699, + "step": 716 + }, + { + "epoch": 1.9616963064295485, + "grad_norm": 0.29453444480895996, + "learning_rate": 1.3810313551631426e-05, + "loss": 1.6203, + "step": 717 + }, + { + "epoch": 1.9644322845417237, + "grad_norm": 0.3803625702857971, + "learning_rate": 1.3744906379558165e-05, + "loss": 1.6514, + "step": 718 + }, + { + "epoch": 1.9671682626538987, + "grad_norm": 0.3245275318622589, + "learning_rate": 1.3679595690108193e-05, + "loss": 1.5904, + "step": 719 + }, + { + "epoch": 1.9699042407660738, + "grad_norm": 0.3239131271839142, + "learning_rate": 1.3614382043147725e-05, + "loss": 1.5117, + "step": 720 + }, + { + "epoch": 1.972640218878249, + "grad_norm": 0.3001211881637573, + "learning_rate": 1.3549265997711057e-05, + "loss": 1.5985, + "step": 721 + }, + { + "epoch": 1.9753761969904242, + "grad_norm": 0.29681622982025146, + "learning_rate": 1.3484248111995892e-05, + "loss": 1.3762, + "step": 722 + }, + { + "epoch": 1.9781121751025992, + "grad_norm": 0.38813284039497375, + "learning_rate": 1.3419328943358392e-05, + "loss": 1.548, + "step": 723 + }, + { + "epoch": 1.9808481532147741, + "grad_norm": 0.29337620735168457, + "learning_rate": 1.3354509048308527e-05, + "loss": 1.4466, + "step": 724 + }, + { + "epoch": 1.9835841313269493, + "grad_norm": 0.3134080171585083, + "learning_rate": 1.328978898250525e-05, + "loss": 1.523, + "step": 725 + }, + { + "epoch": 1.9863201094391245, + "grad_norm": 0.32457372546195984, + "learning_rate": 1.3225169300751738e-05, + "loss": 1.4018, + "step": 726 + }, + { + "epoch": 1.9890560875512997, + "grad_norm": 0.30359819531440735, + "learning_rate": 1.3160650556990644e-05, + "loss": 1.4209, + "step": 727 + }, + { + "epoch": 1.9917920656634747, + "grad_norm": 0.3376854360103607, + "learning_rate": 1.3096233304299346e-05, + "loss": 1.4914, + "step": 728 + }, + { + "epoch": 1.9945280437756496, + "grad_norm": 0.3038989305496216, + "learning_rate": 1.30319180948852e-05, + "loss": 1.4995, + "step": 729 + }, + { + "epoch": 1.9972640218878248, + "grad_norm": 0.2987303137779236, + "learning_rate": 1.2967705480080819e-05, + "loss": 1.5415, + "step": 730 + }, + { + "epoch": 2.0, + "grad_norm": 0.36076483130455017, + "learning_rate": 1.2903596010339338e-05, + "loss": 1.3103, + "step": 731 + }, + { + "epoch": 2.002735978112175, + "grad_norm": 0.329522043466568, + "learning_rate": 1.2839590235229668e-05, + "loss": 1.7937, + "step": 732 + }, + { + "epoch": 2.0054719562243504, + "grad_norm": 0.33667078614234924, + "learning_rate": 1.2775688703431871e-05, + "loss": 1.333, + "step": 733 + }, + { + "epoch": 2.008207934336525, + "grad_norm": 0.30493631958961487, + "learning_rate": 1.2711891962732342e-05, + "loss": 1.3868, + "step": 734 + }, + { + "epoch": 2.0109439124487003, + "grad_norm": 0.3139844536781311, + "learning_rate": 1.26482005600192e-05, + "loss": 1.4361, + "step": 735 + }, + { + "epoch": 2.0136798905608755, + "grad_norm": 0.30529195070266724, + "learning_rate": 1.258461504127756e-05, + "loss": 1.5679, + "step": 736 + }, + { + "epoch": 2.0164158686730507, + "grad_norm": 0.3701570928096771, + "learning_rate": 1.252113595158487e-05, + "loss": 1.7087, + "step": 737 + }, + { + "epoch": 2.019151846785226, + "grad_norm": 0.3016231060028076, + "learning_rate": 1.245776383510622e-05, + "loss": 1.5592, + "step": 738 + }, + { + "epoch": 2.0218878248974006, + "grad_norm": 0.29768821597099304, + "learning_rate": 1.2394499235089699e-05, + "loss": 1.4926, + "step": 739 + }, + { + "epoch": 2.024623803009576, + "grad_norm": 0.3342791497707367, + "learning_rate": 1.2331342693861716e-05, + "loss": 1.4944, + "step": 740 + }, + { + "epoch": 2.027359781121751, + "grad_norm": 0.30851346254348755, + "learning_rate": 1.2268294752822376e-05, + "loss": 1.7049, + "step": 741 + }, + { + "epoch": 2.030095759233926, + "grad_norm": 0.3023451864719391, + "learning_rate": 1.2205355952440817e-05, + "loss": 1.5705, + "step": 742 + }, + { + "epoch": 2.0328317373461013, + "grad_norm": 0.3253107964992523, + "learning_rate": 1.2142526832250561e-05, + "loss": 1.4404, + "step": 743 + }, + { + "epoch": 2.0355677154582765, + "grad_norm": 0.33139538764953613, + "learning_rate": 1.2079807930844977e-05, + "loss": 1.4355, + "step": 744 + }, + { + "epoch": 2.0383036935704513, + "grad_norm": 0.3191182613372803, + "learning_rate": 1.2017199785872523e-05, + "loss": 1.428, + "step": 745 + }, + { + "epoch": 2.0410396716826265, + "grad_norm": 0.33677810430526733, + "learning_rate": 1.1954702934032267e-05, + "loss": 1.4214, + "step": 746 + }, + { + "epoch": 2.0437756497948016, + "grad_norm": 0.31797662377357483, + "learning_rate": 1.1892317911069212e-05, + "loss": 1.5024, + "step": 747 + }, + { + "epoch": 2.046511627906977, + "grad_norm": 0.30894604325294495, + "learning_rate": 1.183004525176973e-05, + "loss": 1.5084, + "step": 748 + }, + { + "epoch": 2.049247606019152, + "grad_norm": 0.30327004194259644, + "learning_rate": 1.176788548995697e-05, + "loss": 1.5586, + "step": 749 + }, + { + "epoch": 2.0519835841313268, + "grad_norm": 0.31576216220855713, + "learning_rate": 1.1705839158486284e-05, + "loss": 1.5315, + "step": 750 + }, + { + "epoch": 2.0519835841313268, + "eval_loss": 1.5589618682861328, + "eval_runtime": 12.9752, + "eval_samples_per_second": 11.869, + "eval_steps_per_second": 11.869, + "step": 750 + }, + { + "epoch": 2.054719562243502, + "grad_norm": 0.3079953193664551, + "learning_rate": 1.1643906789240664e-05, + "loss": 1.6127, + "step": 751 + }, + { + "epoch": 2.057455540355677, + "grad_norm": 0.30269429087638855, + "learning_rate": 1.1582088913126172e-05, + "loss": 1.5077, + "step": 752 + }, + { + "epoch": 2.0601915184678523, + "grad_norm": 0.34395742416381836, + "learning_rate": 1.1520386060067401e-05, + "loss": 1.4675, + "step": 753 + }, + { + "epoch": 2.0629274965800275, + "grad_norm": 0.3478207588195801, + "learning_rate": 1.1458798759002897e-05, + "loss": 1.5474, + "step": 754 + }, + { + "epoch": 2.0656634746922027, + "grad_norm": 0.3087264895439148, + "learning_rate": 1.139732753788072e-05, + "loss": 1.4078, + "step": 755 + }, + { + "epoch": 2.0683994528043774, + "grad_norm": 0.2971005141735077, + "learning_rate": 1.1335972923653774e-05, + "loss": 1.6057, + "step": 756 + }, + { + "epoch": 2.0711354309165526, + "grad_norm": 0.30974239110946655, + "learning_rate": 1.1274735442275402e-05, + "loss": 1.4899, + "step": 757 + }, + { + "epoch": 2.073871409028728, + "grad_norm": 0.3504452705383301, + "learning_rate": 1.121361561869488e-05, + "loss": 1.5413, + "step": 758 + }, + { + "epoch": 2.076607387140903, + "grad_norm": 0.3584803640842438, + "learning_rate": 1.1152613976852804e-05, + "loss": 1.5634, + "step": 759 + }, + { + "epoch": 2.079343365253078, + "grad_norm": 0.3279834985733032, + "learning_rate": 1.1091731039676754e-05, + "loss": 1.4428, + "step": 760 + }, + { + "epoch": 2.082079343365253, + "grad_norm": 0.2989148199558258, + "learning_rate": 1.1030967329076658e-05, + "loss": 1.3392, + "step": 761 + }, + { + "epoch": 2.084815321477428, + "grad_norm": 0.3270657956600189, + "learning_rate": 1.0970323365940444e-05, + "loss": 1.4114, + "step": 762 + }, + { + "epoch": 2.0875512995896033, + "grad_norm": 0.31774088740348816, + "learning_rate": 1.0909799670129497e-05, + "loss": 1.4251, + "step": 763 + }, + { + "epoch": 2.0902872777017785, + "grad_norm": 0.3060950040817261, + "learning_rate": 1.0849396760474246e-05, + "loss": 1.4517, + "step": 764 + }, + { + "epoch": 2.0930232558139537, + "grad_norm": 0.2948378026485443, + "learning_rate": 1.078911515476968e-05, + "loss": 1.4872, + "step": 765 + }, + { + "epoch": 2.0957592339261284, + "grad_norm": 0.2836145758628845, + "learning_rate": 1.0728955369770941e-05, + "loss": 1.4987, + "step": 766 + }, + { + "epoch": 2.0984952120383036, + "grad_norm": 0.33373308181762695, + "learning_rate": 1.0668917921188885e-05, + "loss": 1.4806, + "step": 767 + }, + { + "epoch": 2.1012311901504788, + "grad_norm": 0.29727426171302795, + "learning_rate": 1.060900332368562e-05, + "loss": 1.506, + "step": 768 + }, + { + "epoch": 2.103967168262654, + "grad_norm": 0.3088816702365875, + "learning_rate": 1.0549212090870203e-05, + "loss": 1.7262, + "step": 769 + }, + { + "epoch": 2.106703146374829, + "grad_norm": 0.37396734952926636, + "learning_rate": 1.0489544735294088e-05, + "loss": 1.4311, + "step": 770 + }, + { + "epoch": 2.109439124487004, + "grad_norm": 0.3265590965747833, + "learning_rate": 1.0430001768446856e-05, + "loss": 1.3106, + "step": 771 + }, + { + "epoch": 2.112175102599179, + "grad_norm": 0.3493903577327728, + "learning_rate": 1.0370583700751774e-05, + "loss": 1.6644, + "step": 772 + }, + { + "epoch": 2.1149110807113543, + "grad_norm": 0.32931625843048096, + "learning_rate": 1.0311291041561428e-05, + "loss": 1.448, + "step": 773 + }, + { + "epoch": 2.1176470588235294, + "grad_norm": 0.3126463294029236, + "learning_rate": 1.0252124299153353e-05, + "loss": 1.5439, + "step": 774 + }, + { + "epoch": 2.1203830369357046, + "grad_norm": 0.3198404014110565, + "learning_rate": 1.0193083980725696e-05, + "loss": 1.4068, + "step": 775 + }, + { + "epoch": 2.12311901504788, + "grad_norm": 0.3446153998374939, + "learning_rate": 1.0134170592392836e-05, + "loss": 1.4597, + "step": 776 + }, + { + "epoch": 2.1258549931600546, + "grad_norm": 0.34225985407829285, + "learning_rate": 1.007538463918107e-05, + "loss": 1.5849, + "step": 777 + }, + { + "epoch": 2.1285909712722297, + "grad_norm": 0.3094581663608551, + "learning_rate": 1.0016726625024287e-05, + "loss": 1.5853, + "step": 778 + }, + { + "epoch": 2.131326949384405, + "grad_norm": 0.27853038907051086, + "learning_rate": 9.9581970527596e-06, + "loss": 1.5799, + "step": 779 + }, + { + "epoch": 2.13406292749658, + "grad_norm": 0.33034032583236694, + "learning_rate": 9.899796424123136e-06, + "loss": 1.456, + "step": 780 + }, + { + "epoch": 2.1367989056087553, + "grad_norm": 0.3268153965473175, + "learning_rate": 9.841525239745605e-06, + "loss": 1.3883, + "step": 781 + }, + { + "epoch": 2.13953488372093, + "grad_norm": 0.32485464215278625, + "learning_rate": 9.783383999148118e-06, + "loss": 1.4159, + "step": 782 + }, + { + "epoch": 2.1422708618331052, + "grad_norm": 0.29737138748168945, + "learning_rate": 9.72537320073785e-06, + "loss": 1.5254, + "step": 783 + }, + { + "epoch": 2.1450068399452804, + "grad_norm": 0.2851496934890747, + "learning_rate": 9.667493341803777e-06, + "loss": 1.3903, + "step": 784 + }, + { + "epoch": 2.1477428180574556, + "grad_norm": 0.32795658707618713, + "learning_rate": 9.60974491851242e-06, + "loss": 1.5147, + "step": 785 + }, + { + "epoch": 2.150478796169631, + "grad_norm": 0.33503639698028564, + "learning_rate": 9.552128425903586e-06, + "loss": 1.4495, + "step": 786 + }, + { + "epoch": 2.153214774281806, + "grad_norm": 0.31696316599845886, + "learning_rate": 9.494644357886124e-06, + "loss": 1.5787, + "step": 787 + }, + { + "epoch": 2.1559507523939807, + "grad_norm": 0.321809321641922, + "learning_rate": 9.437293207233695e-06, + "loss": 1.5817, + "step": 788 + }, + { + "epoch": 2.158686730506156, + "grad_norm": 0.29343757033348083, + "learning_rate": 9.380075465580552e-06, + "loss": 1.5604, + "step": 789 + }, + { + "epoch": 2.161422708618331, + "grad_norm": 0.4285193383693695, + "learning_rate": 9.322991623417285e-06, + "loss": 1.4387, + "step": 790 + }, + { + "epoch": 2.1641586867305063, + "grad_norm": 0.29187867045402527, + "learning_rate": 9.266042170086717e-06, + "loss": 1.4871, + "step": 791 + }, + { + "epoch": 2.1668946648426815, + "grad_norm": 0.2987869679927826, + "learning_rate": 9.209227593779573e-06, + "loss": 1.3972, + "step": 792 + }, + { + "epoch": 2.169630642954856, + "grad_norm": 0.29556939005851746, + "learning_rate": 9.152548381530405e-06, + "loss": 1.4057, + "step": 793 + }, + { + "epoch": 2.1723666210670314, + "grad_norm": 0.32454240322113037, + "learning_rate": 9.096005019213363e-06, + "loss": 1.5509, + "step": 794 + }, + { + "epoch": 2.1751025991792066, + "grad_norm": 0.30358558893203735, + "learning_rate": 9.039597991538043e-06, + "loss": 1.5466, + "step": 795 + }, + { + "epoch": 2.1778385772913817, + "grad_norm": 0.3079793155193329, + "learning_rate": 8.983327782045359e-06, + "loss": 1.8167, + "step": 796 + }, + { + "epoch": 2.180574555403557, + "grad_norm": 0.3283805847167969, + "learning_rate": 8.927194873103322e-06, + "loss": 1.5464, + "step": 797 + }, + { + "epoch": 2.1833105335157317, + "grad_norm": 0.3089551031589508, + "learning_rate": 8.871199745902997e-06, + "loss": 1.4067, + "step": 798 + }, + { + "epoch": 2.186046511627907, + "grad_norm": 0.3624851405620575, + "learning_rate": 8.815342880454311e-06, + "loss": 1.4652, + "step": 799 + }, + { + "epoch": 2.188782489740082, + "grad_norm": 0.31375181674957275, + "learning_rate": 8.75962475558198e-06, + "loss": 1.4584, + "step": 800 + }, + { + "epoch": 2.1915184678522572, + "grad_norm": 0.36298835277557373, + "learning_rate": 8.704045848921358e-06, + "loss": 1.6755, + "step": 801 + }, + { + "epoch": 2.1942544459644324, + "grad_norm": 0.3472815155982971, + "learning_rate": 8.648606636914416e-06, + "loss": 1.4749, + "step": 802 + }, + { + "epoch": 2.196990424076607, + "grad_norm": 0.30575889348983765, + "learning_rate": 8.593307594805586e-06, + "loss": 1.616, + "step": 803 + }, + { + "epoch": 2.1997264021887823, + "grad_norm": 0.33051928877830505, + "learning_rate": 8.53814919663769e-06, + "loss": 1.4458, + "step": 804 + }, + { + "epoch": 2.2024623803009575, + "grad_norm": 0.31675076484680176, + "learning_rate": 8.483131915247968e-06, + "loss": 1.4981, + "step": 805 + }, + { + "epoch": 2.2051983584131327, + "grad_norm": 0.33154037594795227, + "learning_rate": 8.428256222263888e-06, + "loss": 1.5108, + "step": 806 + }, + { + "epoch": 2.207934336525308, + "grad_norm": 0.3389985263347626, + "learning_rate": 8.373522588099211e-06, + "loss": 1.4522, + "step": 807 + }, + { + "epoch": 2.210670314637483, + "grad_norm": 0.35865381360054016, + "learning_rate": 8.318931481949906e-06, + "loss": 1.5059, + "step": 808 + }, + { + "epoch": 2.213406292749658, + "grad_norm": 0.3100070357322693, + "learning_rate": 8.264483371790156e-06, + "loss": 1.441, + "step": 809 + }, + { + "epoch": 2.216142270861833, + "grad_norm": 0.3514600694179535, + "learning_rate": 8.210178724368312e-06, + "loss": 1.6657, + "step": 810 + }, + { + "epoch": 2.218878248974008, + "grad_norm": 0.2982495427131653, + "learning_rate": 8.156018005202926e-06, + "loss": 1.5945, + "step": 811 + }, + { + "epoch": 2.2216142270861834, + "grad_norm": 0.2904994487762451, + "learning_rate": 8.10200167857874e-06, + "loss": 1.4298, + "step": 812 + }, + { + "epoch": 2.2243502051983586, + "grad_norm": 0.3352917432785034, + "learning_rate": 8.04813020754272e-06, + "loss": 1.5512, + "step": 813 + }, + { + "epoch": 2.2270861833105333, + "grad_norm": 0.31548169255256653, + "learning_rate": 7.994404053900076e-06, + "loss": 1.5735, + "step": 814 + }, + { + "epoch": 2.2298221614227085, + "grad_norm": 0.31222277879714966, + "learning_rate": 7.940823678210282e-06, + "loss": 1.4862, + "step": 815 + }, + { + "epoch": 2.2325581395348837, + "grad_norm": 0.3492494225502014, + "learning_rate": 7.887389539783208e-06, + "loss": 1.6087, + "step": 816 + }, + { + "epoch": 2.235294117647059, + "grad_norm": 0.32157906889915466, + "learning_rate": 7.834102096675064e-06, + "loss": 1.5771, + "step": 817 + }, + { + "epoch": 2.238030095759234, + "grad_norm": 0.3391493260860443, + "learning_rate": 7.780961805684575e-06, + "loss": 1.5289, + "step": 818 + }, + { + "epoch": 2.2407660738714092, + "grad_norm": 0.32275089621543884, + "learning_rate": 7.727969122349017e-06, + "loss": 1.407, + "step": 819 + }, + { + "epoch": 2.243502051983584, + "grad_norm": 0.34808090329170227, + "learning_rate": 7.675124500940317e-06, + "loss": 1.4914, + "step": 820 + }, + { + "epoch": 2.246238030095759, + "grad_norm": 0.35853490233421326, + "learning_rate": 7.62242839446117e-06, + "loss": 1.4587, + "step": 821 + }, + { + "epoch": 2.2489740082079344, + "grad_norm": 0.298554390668869, + "learning_rate": 7.569881254641142e-06, + "loss": 1.52, + "step": 822 + }, + { + "epoch": 2.2517099863201095, + "grad_norm": 0.3177081048488617, + "learning_rate": 7.517483531932809e-06, + "loss": 1.3685, + "step": 823 + }, + { + "epoch": 2.2544459644322847, + "grad_norm": 0.30928516387939453, + "learning_rate": 7.4652356755078895e-06, + "loss": 1.4636, + "step": 824 + }, + { + "epoch": 2.2571819425444595, + "grad_norm": 0.3332304060459137, + "learning_rate": 7.4131381332534e-06, + "loss": 1.4894, + "step": 825 + }, + { + "epoch": 2.2599179206566347, + "grad_norm": 0.331177681684494, + "learning_rate": 7.3611913517677875e-06, + "loss": 1.5236, + "step": 826 + }, + { + "epoch": 2.26265389876881, + "grad_norm": 0.3152174949645996, + "learning_rate": 7.30939577635717e-06, + "loss": 1.4508, + "step": 827 + }, + { + "epoch": 2.265389876880985, + "grad_norm": 0.33167538046836853, + "learning_rate": 7.257751851031428e-06, + "loss": 1.4507, + "step": 828 + }, + { + "epoch": 2.26812585499316, + "grad_norm": 0.26186057925224304, + "learning_rate": 7.206260018500469e-06, + "loss": 1.581, + "step": 829 + }, + { + "epoch": 2.270861833105335, + "grad_norm": 0.31686386466026306, + "learning_rate": 7.154920720170399e-06, + "loss": 1.5677, + "step": 830 + }, + { + "epoch": 2.27359781121751, + "grad_norm": 0.32669878005981445, + "learning_rate": 7.1037343961397495e-06, + "loss": 1.5512, + "step": 831 + }, + { + "epoch": 2.2763337893296853, + "grad_norm": 0.3603561520576477, + "learning_rate": 7.05270148519572e-06, + "loss": 1.6478, + "step": 832 + }, + { + "epoch": 2.2790697674418605, + "grad_norm": 0.3097962737083435, + "learning_rate": 7.001822424810359e-06, + "loss": 1.5819, + "step": 833 + }, + { + "epoch": 2.2818057455540357, + "grad_norm": 0.32818931341171265, + "learning_rate": 6.951097651136889e-06, + "loss": 1.5053, + "step": 834 + }, + { + "epoch": 2.2845417236662104, + "grad_norm": 0.41288748383522034, + "learning_rate": 6.900527599005918e-06, + "loss": 1.5391, + "step": 835 + }, + { + "epoch": 2.2872777017783856, + "grad_norm": 0.3428272008895874, + "learning_rate": 6.8501127019217346e-06, + "loss": 1.4097, + "step": 836 + }, + { + "epoch": 2.290013679890561, + "grad_norm": 0.34903377294540405, + "learning_rate": 6.799853392058561e-06, + "loss": 1.3627, + "step": 837 + }, + { + "epoch": 2.292749658002736, + "grad_norm": 0.3197937309741974, + "learning_rate": 6.749750100256902e-06, + "loss": 1.5362, + "step": 838 + }, + { + "epoch": 2.295485636114911, + "grad_norm": 0.3300299644470215, + "learning_rate": 6.699803256019815e-06, + "loss": 1.6251, + "step": 839 + }, + { + "epoch": 2.2982216142270864, + "grad_norm": 0.33497533202171326, + "learning_rate": 6.650013287509199e-06, + "loss": 1.5388, + "step": 840 + }, + { + "epoch": 2.300957592339261, + "grad_norm": 0.36092567443847656, + "learning_rate": 6.600380621542216e-06, + "loss": 1.4969, + "step": 841 + }, + { + "epoch": 2.3036935704514363, + "grad_norm": 0.3416774868965149, + "learning_rate": 6.550905683587513e-06, + "loss": 1.5842, + "step": 842 + }, + { + "epoch": 2.3064295485636115, + "grad_norm": 0.30258703231811523, + "learning_rate": 6.5015888977617016e-06, + "loss": 1.3653, + "step": 843 + }, + { + "epoch": 2.3091655266757867, + "grad_norm": 0.2986971139907837, + "learning_rate": 6.452430686825603e-06, + "loss": 1.5223, + "step": 844 + }, + { + "epoch": 2.311901504787962, + "grad_norm": 0.32431551814079285, + "learning_rate": 6.403431472180707e-06, + "loss": 1.4931, + "step": 845 + }, + { + "epoch": 2.314637482900137, + "grad_norm": 0.32962799072265625, + "learning_rate": 6.354591673865523e-06, + "loss": 1.4833, + "step": 846 + }, + { + "epoch": 2.317373461012312, + "grad_norm": 0.36730074882507324, + "learning_rate": 6.30591171055199e-06, + "loss": 1.4419, + "step": 847 + }, + { + "epoch": 2.320109439124487, + "grad_norm": 0.3103657066822052, + "learning_rate": 6.257391999541887e-06, + "loss": 1.4914, + "step": 848 + }, + { + "epoch": 2.322845417236662, + "grad_norm": 0.3909322917461395, + "learning_rate": 6.209032956763247e-06, + "loss": 1.4929, + "step": 849 + }, + { + "epoch": 2.3255813953488373, + "grad_norm": 0.346003919839859, + "learning_rate": 6.160834996766815e-06, + "loss": 1.3601, + "step": 850 + }, + { + "epoch": 2.3283173734610125, + "grad_norm": 0.3200609087944031, + "learning_rate": 6.112798532722438e-06, + "loss": 1.6096, + "step": 851 + }, + { + "epoch": 2.3310533515731873, + "grad_norm": 0.31736400723457336, + "learning_rate": 6.064923976415626e-06, + "loss": 1.4462, + "step": 852 + }, + { + "epoch": 2.3337893296853625, + "grad_norm": 0.2982243299484253, + "learning_rate": 6.0172117382438944e-06, + "loss": 1.6742, + "step": 853 + }, + { + "epoch": 2.3365253077975376, + "grad_norm": 0.36678165197372437, + "learning_rate": 5.969662227213352e-06, + "loss": 1.429, + "step": 854 + }, + { + "epoch": 2.339261285909713, + "grad_norm": 0.29900214076042175, + "learning_rate": 5.922275850935136e-06, + "loss": 1.5513, + "step": 855 + }, + { + "epoch": 2.341997264021888, + "grad_norm": 0.3283601403236389, + "learning_rate": 5.875053015621943e-06, + "loss": 1.7077, + "step": 856 + }, + { + "epoch": 2.3447332421340628, + "grad_norm": 0.33298107981681824, + "learning_rate": 5.827994126084532e-06, + "loss": 1.5916, + "step": 857 + }, + { + "epoch": 2.347469220246238, + "grad_norm": 0.3546065390110016, + "learning_rate": 5.781099585728267e-06, + "loss": 1.4382, + "step": 858 + }, + { + "epoch": 2.350205198358413, + "grad_norm": 0.30142465233802795, + "learning_rate": 5.734369796549652e-06, + "loss": 1.4578, + "step": 859 + }, + { + "epoch": 2.3529411764705883, + "grad_norm": 0.3024732768535614, + "learning_rate": 5.687805159132881e-06, + "loss": 1.4102, + "step": 860 + }, + { + "epoch": 2.3556771545827635, + "grad_norm": 0.3300219774246216, + "learning_rate": 5.641406072646418e-06, + "loss": 1.5036, + "step": 861 + }, + { + "epoch": 2.3584131326949382, + "grad_norm": 0.37552839517593384, + "learning_rate": 5.595172934839546e-06, + "loss": 1.4401, + "step": 862 + }, + { + "epoch": 2.3611491108071134, + "grad_norm": 0.3124082386493683, + "learning_rate": 5.549106142039018e-06, + "loss": 1.3749, + "step": 863 + }, + { + "epoch": 2.3638850889192886, + "grad_norm": 0.34462398290634155, + "learning_rate": 5.503206089145568e-06, + "loss": 1.578, + "step": 864 + }, + { + "epoch": 2.366621067031464, + "grad_norm": 0.3301493227481842, + "learning_rate": 5.4574731696306146e-06, + "loss": 1.7064, + "step": 865 + }, + { + "epoch": 2.369357045143639, + "grad_norm": 0.33227112889289856, + "learning_rate": 5.411907775532832e-06, + "loss": 1.4069, + "step": 866 + }, + { + "epoch": 2.3720930232558137, + "grad_norm": 0.3299727439880371, + "learning_rate": 5.366510297454816e-06, + "loss": 1.4481, + "step": 867 + }, + { + "epoch": 2.374829001367989, + "grad_norm": 0.3437407314777374, + "learning_rate": 5.321281124559727e-06, + "loss": 1.3406, + "step": 868 + }, + { + "epoch": 2.377564979480164, + "grad_norm": 0.3354208469390869, + "learning_rate": 5.276220644567948e-06, + "loss": 1.3835, + "step": 869 + }, + { + "epoch": 2.3803009575923393, + "grad_norm": 0.29964345693588257, + "learning_rate": 5.231329243753772e-06, + "loss": 1.3945, + "step": 870 + }, + { + "epoch": 2.3830369357045145, + "grad_norm": 0.3642790615558624, + "learning_rate": 5.186607306942085e-06, + "loss": 1.3748, + "step": 871 + }, + { + "epoch": 2.3857729138166897, + "grad_norm": 0.29215675592422485, + "learning_rate": 5.142055217505074e-06, + "loss": 1.5209, + "step": 872 + }, + { + "epoch": 2.3885088919288644, + "grad_norm": 0.35326606035232544, + "learning_rate": 5.097673357358907e-06, + "loss": 1.5554, + "step": 873 + }, + { + "epoch": 2.3912448700410396, + "grad_norm": 0.3523205816745758, + "learning_rate": 5.053462106960532e-06, + "loss": 1.5203, + "step": 874 + }, + { + "epoch": 2.3939808481532148, + "grad_norm": 0.3213190734386444, + "learning_rate": 5.00942184530433e-06, + "loss": 1.4171, + "step": 875 + }, + { + "epoch": 2.39671682626539, + "grad_norm": 0.3184296488761902, + "learning_rate": 4.96555294991892e-06, + "loss": 1.3045, + "step": 876 + }, + { + "epoch": 2.399452804377565, + "grad_norm": 0.3270394802093506, + "learning_rate": 4.921855796863933e-06, + "loss": 1.5365, + "step": 877 + }, + { + "epoch": 2.4021887824897403, + "grad_norm": 0.36887189745903015, + "learning_rate": 4.878330760726713e-06, + "loss": 1.2823, + "step": 878 + }, + { + "epoch": 2.404924760601915, + "grad_norm": 0.3077680766582489, + "learning_rate": 4.834978214619215e-06, + "loss": 1.5222, + "step": 879 + }, + { + "epoch": 2.4076607387140903, + "grad_norm": 0.3177037835121155, + "learning_rate": 4.791798530174699e-06, + "loss": 1.4489, + "step": 880 + }, + { + "epoch": 2.4103967168262654, + "grad_norm": 0.3479926586151123, + "learning_rate": 4.748792077544623e-06, + "loss": 1.5376, + "step": 881 + }, + { + "epoch": 2.4131326949384406, + "grad_norm": 0.32515963912010193, + "learning_rate": 4.7059592253954315e-06, + "loss": 1.5803, + "step": 882 + }, + { + "epoch": 2.415868673050616, + "grad_norm": 0.33568742871284485, + "learning_rate": 4.66330034090541e-06, + "loss": 1.5066, + "step": 883 + }, + { + "epoch": 2.4186046511627906, + "grad_norm": 0.328945130109787, + "learning_rate": 4.620815789761526e-06, + "loss": 1.4279, + "step": 884 + }, + { + "epoch": 2.4213406292749657, + "grad_norm": 0.3277706503868103, + "learning_rate": 4.578505936156302e-06, + "loss": 1.5832, + "step": 885 + }, + { + "epoch": 2.424076607387141, + "grad_norm": 0.35384321212768555, + "learning_rate": 4.5363711427847015e-06, + "loss": 1.4504, + "step": 886 + }, + { + "epoch": 2.426812585499316, + "grad_norm": 0.3523906469345093, + "learning_rate": 4.494411770840978e-06, + "loss": 1.5453, + "step": 887 + }, + { + "epoch": 2.4295485636114913, + "grad_norm": 0.30951249599456787, + "learning_rate": 4.452628180015664e-06, + "loss": 1.5488, + "step": 888 + }, + { + "epoch": 2.432284541723666, + "grad_norm": 0.3258694112300873, + "learning_rate": 4.411020728492374e-06, + "loss": 1.4878, + "step": 889 + }, + { + "epoch": 2.4350205198358412, + "grad_norm": 0.33756595849990845, + "learning_rate": 4.3695897729448485e-06, + "loss": 1.4415, + "step": 890 + }, + { + "epoch": 2.4377564979480164, + "grad_norm": 0.3134118318557739, + "learning_rate": 4.328335668533806e-06, + "loss": 1.3717, + "step": 891 + }, + { + "epoch": 2.4404924760601916, + "grad_norm": 0.2975947856903076, + "learning_rate": 4.2872587689039484e-06, + "loss": 1.5299, + "step": 892 + }, + { + "epoch": 2.443228454172367, + "grad_norm": 0.3447892963886261, + "learning_rate": 4.246359426180918e-06, + "loss": 1.4769, + "step": 893 + }, + { + "epoch": 2.4459644322845415, + "grad_norm": 0.32196173071861267, + "learning_rate": 4.20563799096827e-06, + "loss": 1.4507, + "step": 894 + }, + { + "epoch": 2.4487004103967167, + "grad_norm": 0.30560824275016785, + "learning_rate": 4.165094812344478e-06, + "loss": 1.6074, + "step": 895 + }, + { + "epoch": 2.451436388508892, + "grad_norm": 0.3304429352283478, + "learning_rate": 4.124730237859939e-06, + "loss": 1.5368, + "step": 896 + }, + { + "epoch": 2.454172366621067, + "grad_norm": 0.33104217052459717, + "learning_rate": 4.0845446135339945e-06, + "loss": 1.4162, + "step": 897 + }, + { + "epoch": 2.4569083447332423, + "grad_norm": 0.3179687261581421, + "learning_rate": 4.0445382838519365e-06, + "loss": 1.6278, + "step": 898 + }, + { + "epoch": 2.459644322845417, + "grad_norm": 0.3108776807785034, + "learning_rate": 4.004711591762133e-06, + "loss": 1.6218, + "step": 899 + }, + { + "epoch": 2.462380300957592, + "grad_norm": 0.3062944710254669, + "learning_rate": 3.965064878672983e-06, + "loss": 1.4816, + "step": 900 + }, + { + "epoch": 2.4651162790697674, + "grad_norm": 0.3266476094722748, + "learning_rate": 3.925598484450066e-06, + "loss": 1.846, + "step": 901 + }, + { + "epoch": 2.4678522571819426, + "grad_norm": 0.344200998544693, + "learning_rate": 3.886312747413204e-06, + "loss": 1.3894, + "step": 902 + }, + { + "epoch": 2.4705882352941178, + "grad_norm": 0.315949410200119, + "learning_rate": 3.847208004333561e-06, + "loss": 1.4174, + "step": 903 + }, + { + "epoch": 2.473324213406293, + "grad_norm": 0.3364209234714508, + "learning_rate": 3.8082845904307525e-06, + "loss": 1.4093, + "step": 904 + }, + { + "epoch": 2.4760601915184677, + "grad_norm": 0.3172418475151062, + "learning_rate": 3.7695428393699854e-06, + "loss": 1.3446, + "step": 905 + }, + { + "epoch": 2.478796169630643, + "grad_norm": 0.3134661316871643, + "learning_rate": 3.730983083259179e-06, + "loss": 1.507, + "step": 906 + }, + { + "epoch": 2.481532147742818, + "grad_norm": 0.3210635185241699, + "learning_rate": 3.6926056526461334e-06, + "loss": 1.3344, + "step": 907 + }, + { + "epoch": 2.4842681258549932, + "grad_norm": 0.32122674584388733, + "learning_rate": 3.6544108765156933e-06, + "loss": 1.5514, + "step": 908 + }, + { + "epoch": 2.4870041039671684, + "grad_norm": 0.36177971959114075, + "learning_rate": 3.6163990822869088e-06, + "loss": 1.5981, + "step": 909 + }, + { + "epoch": 2.4897400820793436, + "grad_norm": 0.3207845389842987, + "learning_rate": 3.578570595810274e-06, + "loss": 1.5989, + "step": 910 + }, + { + "epoch": 2.4924760601915183, + "grad_norm": 0.32466524839401245, + "learning_rate": 3.540925741364873e-06, + "loss": 1.483, + "step": 911 + }, + { + "epoch": 2.4952120383036935, + "grad_norm": 0.31727364659309387, + "learning_rate": 3.5034648416556486e-06, + "loss": 1.5308, + "step": 912 + }, + { + "epoch": 2.4979480164158687, + "grad_norm": 0.3401485085487366, + "learning_rate": 3.4661882178106176e-06, + "loss": 1.6316, + "step": 913 + }, + { + "epoch": 2.500683994528044, + "grad_norm": 0.3355841338634491, + "learning_rate": 3.429096189378114e-06, + "loss": 1.5952, + "step": 914 + }, + { + "epoch": 2.503419972640219, + "grad_norm": 0.3085014522075653, + "learning_rate": 3.392189074324073e-06, + "loss": 1.6774, + "step": 915 + }, + { + "epoch": 2.506155950752394, + "grad_norm": 0.3420393466949463, + "learning_rate": 3.355467189029257e-06, + "loss": 1.4805, + "step": 916 + }, + { + "epoch": 2.508891928864569, + "grad_norm": 0.3142971098423004, + "learning_rate": 3.3189308482865917e-06, + "loss": 1.4958, + "step": 917 + }, + { + "epoch": 2.511627906976744, + "grad_norm": 0.37307271361351013, + "learning_rate": 3.2825803652984516e-06, + "loss": 1.4541, + "step": 918 + }, + { + "epoch": 2.5143638850889194, + "grad_norm": 0.33256828784942627, + "learning_rate": 3.2464160516739755e-06, + "loss": 1.4565, + "step": 919 + }, + { + "epoch": 2.5170998632010946, + "grad_norm": 0.29588747024536133, + "learning_rate": 3.210438217426365e-06, + "loss": 1.4852, + "step": 920 + }, + { + "epoch": 2.5198358413132693, + "grad_norm": 0.314485102891922, + "learning_rate": 3.1746471709702964e-06, + "loss": 1.3962, + "step": 921 + }, + { + "epoch": 2.5225718194254445, + "grad_norm": 0.30730655789375305, + "learning_rate": 3.1390432191192115e-06, + "loss": 1.5221, + "step": 922 + }, + { + "epoch": 2.5253077975376197, + "grad_norm": 0.34515947103500366, + "learning_rate": 3.1036266670827014e-06, + "loss": 1.4745, + "step": 923 + }, + { + "epoch": 2.528043775649795, + "grad_norm": 0.3266810476779938, + "learning_rate": 3.068397818463936e-06, + "loss": 1.4541, + "step": 924 + }, + { + "epoch": 2.53077975376197, + "grad_norm": 0.2986440062522888, + "learning_rate": 3.033356975256979e-06, + "loss": 1.5795, + "step": 925 + }, + { + "epoch": 2.533515731874145, + "grad_norm": 0.337341845035553, + "learning_rate": 2.9985044378442933e-06, + "loss": 1.5815, + "step": 926 + }, + { + "epoch": 2.53625170998632, + "grad_norm": 0.3632700741291046, + "learning_rate": 2.963840504994075e-06, + "loss": 1.54, + "step": 927 + }, + { + "epoch": 2.538987688098495, + "grad_norm": 0.3714698553085327, + "learning_rate": 2.9293654738577647e-06, + "loss": 1.5019, + "step": 928 + }, + { + "epoch": 2.5417236662106704, + "grad_norm": 0.3200991153717041, + "learning_rate": 2.8950796399674573e-06, + "loss": 1.4587, + "step": 929 + }, + { + "epoch": 2.5444596443228455, + "grad_norm": 0.3523624539375305, + "learning_rate": 2.860983297233388e-06, + "loss": 1.4111, + "step": 930 + }, + { + "epoch": 2.5471956224350203, + "grad_norm": 0.3558140993118286, + "learning_rate": 2.8270767379414087e-06, + "loss": 1.3241, + "step": 931 + }, + { + "epoch": 2.5499316005471955, + "grad_norm": 0.3104929029941559, + "learning_rate": 2.7933602527504738e-06, + "loss": 1.6253, + "step": 932 + }, + { + "epoch": 2.5526675786593707, + "grad_norm": 0.33675897121429443, + "learning_rate": 2.7598341306901643e-06, + "loss": 1.5365, + "step": 933 + }, + { + "epoch": 2.555403556771546, + "grad_norm": 0.39364495873451233, + "learning_rate": 2.726498659158183e-06, + "loss": 1.5522, + "step": 934 + }, + { + "epoch": 2.558139534883721, + "grad_norm": 0.31920477747917175, + "learning_rate": 2.693354123917943e-06, + "loss": 1.5633, + "step": 935 + }, + { + "epoch": 2.5608755129958958, + "grad_norm": 0.37531107664108276, + "learning_rate": 2.660400809096045e-06, + "loss": 1.3603, + "step": 936 + }, + { + "epoch": 2.5636114911080714, + "grad_norm": 0.31020471453666687, + "learning_rate": 2.6276389971799153e-06, + "loss": 1.3862, + "step": 937 + }, + { + "epoch": 2.566347469220246, + "grad_norm": 0.3524358570575714, + "learning_rate": 2.595068969015327e-06, + "loss": 1.5923, + "step": 938 + }, + { + "epoch": 2.5690834473324213, + "grad_norm": 0.3156653940677643, + "learning_rate": 2.562691003804024e-06, + "loss": 1.6537, + "step": 939 + }, + { + "epoch": 2.5718194254445965, + "grad_norm": 0.3592538833618164, + "learning_rate": 2.5305053791013194e-06, + "loss": 1.3821, + "step": 940 + }, + { + "epoch": 2.5745554035567717, + "grad_norm": 0.3512822091579437, + "learning_rate": 2.498512370813716e-06, + "loss": 1.5147, + "step": 941 + }, + { + "epoch": 2.577291381668947, + "grad_norm": 0.31111329793930054, + "learning_rate": 2.466712253196535e-06, + "loss": 1.443, + "step": 942 + }, + { + "epoch": 2.5800273597811216, + "grad_norm": 0.31163808703422546, + "learning_rate": 2.4351052988515783e-06, + "loss": 1.5938, + "step": 943 + }, + { + "epoch": 2.582763337893297, + "grad_norm": 0.34021854400634766, + "learning_rate": 2.403691778724787e-06, + "loss": 1.4879, + "step": 944 + }, + { + "epoch": 2.585499316005472, + "grad_norm": 0.30090874433517456, + "learning_rate": 2.3724719621038923e-06, + "loss": 1.5188, + "step": 945 + }, + { + "epoch": 2.588235294117647, + "grad_norm": 0.2972038984298706, + "learning_rate": 2.341446116616172e-06, + "loss": 1.4422, + "step": 946 + }, + { + "epoch": 2.5909712722298224, + "grad_norm": 0.360759973526001, + "learning_rate": 2.310614508226078e-06, + "loss": 1.4662, + "step": 947 + }, + { + "epoch": 2.593707250341997, + "grad_norm": 0.33597201108932495, + "learning_rate": 2.2799774012330076e-06, + "loss": 1.4008, + "step": 948 + }, + { + "epoch": 2.5964432284541723, + "grad_norm": 0.3385523855686188, + "learning_rate": 2.2495350582690254e-06, + "loss": 1.4652, + "step": 949 + }, + { + "epoch": 2.5991792065663475, + "grad_norm": 0.3732840418815613, + "learning_rate": 2.219287740296605e-06, + "loss": 1.3399, + "step": 950 + }, + { + "epoch": 2.6019151846785227, + "grad_norm": 0.37210366129875183, + "learning_rate": 2.1892357066064128e-06, + "loss": 1.3226, + "step": 951 + }, + { + "epoch": 2.604651162790698, + "grad_norm": 0.33239394426345825, + "learning_rate": 2.1593792148150437e-06, + "loss": 1.3563, + "step": 952 + }, + { + "epoch": 2.6073871409028726, + "grad_norm": 0.3127799332141876, + "learning_rate": 2.1297185208628585e-06, + "loss": 1.5005, + "step": 953 + }, + { + "epoch": 2.610123119015048, + "grad_norm": 0.3832361698150635, + "learning_rate": 2.10025387901176e-06, + "loss": 1.3859, + "step": 954 + }, + { + "epoch": 2.612859097127223, + "grad_norm": 0.3399605453014374, + "learning_rate": 2.0709855418430317e-06, + "loss": 1.559, + "step": 955 + }, + { + "epoch": 2.615595075239398, + "grad_norm": 0.36757224798202515, + "learning_rate": 2.0419137602551516e-06, + "loss": 1.475, + "step": 956 + }, + { + "epoch": 2.6183310533515733, + "grad_norm": 0.3125869929790497, + "learning_rate": 2.013038783461674e-06, + "loss": 1.4586, + "step": 957 + }, + { + "epoch": 2.621067031463748, + "grad_norm": 0.31474483013153076, + "learning_rate": 1.9843608589890513e-06, + "loss": 1.4757, + "step": 958 + }, + { + "epoch": 2.6238030095759233, + "grad_norm": 0.3160763382911682, + "learning_rate": 1.9558802326745424e-06, + "loss": 1.5655, + "step": 959 + }, + { + "epoch": 2.6265389876880985, + "grad_norm": 0.41808021068573, + "learning_rate": 1.927597148664112e-06, + "loss": 1.4522, + "step": 960 + }, + { + "epoch": 2.6292749658002736, + "grad_norm": 0.38478416204452515, + "learning_rate": 1.8995118494102903e-06, + "loss": 1.5056, + "step": 961 + }, + { + "epoch": 2.632010943912449, + "grad_norm": 0.33798423409461975, + "learning_rate": 1.8716245756701694e-06, + "loss": 1.5478, + "step": 962 + }, + { + "epoch": 2.6347469220246236, + "grad_norm": 0.34396031498908997, + "learning_rate": 1.843935566503252e-06, + "loss": 1.438, + "step": 963 + }, + { + "epoch": 2.6374829001367988, + "grad_norm": 0.317111998796463, + "learning_rate": 1.816445059269481e-06, + "loss": 1.4886, + "step": 964 + }, + { + "epoch": 2.640218878248974, + "grad_norm": 0.314860075712204, + "learning_rate": 1.7891532896271547e-06, + "loss": 1.5923, + "step": 965 + }, + { + "epoch": 2.642954856361149, + "grad_norm": 0.29904037714004517, + "learning_rate": 1.7620604915309257e-06, + "loss": 1.3853, + "step": 966 + }, + { + "epoch": 2.6456908344733243, + "grad_norm": 0.34991443157196045, + "learning_rate": 1.7351668972297924e-06, + "loss": 1.5588, + "step": 967 + }, + { + "epoch": 2.648426812585499, + "grad_norm": 0.3276471197605133, + "learning_rate": 1.7084727372651155e-06, + "loss": 1.4885, + "step": 968 + }, + { + "epoch": 2.6511627906976747, + "grad_norm": 0.3197418749332428, + "learning_rate": 1.6819782404686263e-06, + "loss": 1.5063, + "step": 969 + }, + { + "epoch": 2.6538987688098494, + "grad_norm": 0.3408433198928833, + "learning_rate": 1.655683633960467e-06, + "loss": 1.4564, + "step": 970 + }, + { + "epoch": 2.6566347469220246, + "grad_norm": 0.33302435278892517, + "learning_rate": 1.62958914314727e-06, + "loss": 1.3688, + "step": 971 + }, + { + "epoch": 2.6593707250342, + "grad_norm": 0.34161651134490967, + "learning_rate": 1.6036949917201783e-06, + "loss": 1.5104, + "step": 972 + }, + { + "epoch": 2.662106703146375, + "grad_norm": 0.33324992656707764, + "learning_rate": 1.578001401652987e-06, + "loss": 1.4552, + "step": 973 + }, + { + "epoch": 2.66484268125855, + "grad_norm": 0.30797359347343445, + "learning_rate": 1.5525085932001832e-06, + "loss": 1.5157, + "step": 974 + }, + { + "epoch": 2.667578659370725, + "grad_norm": 0.3368433117866516, + "learning_rate": 1.5272167848950913e-06, + "loss": 1.435, + "step": 975 + }, + { + "epoch": 2.6703146374829, + "grad_norm": 0.3299857974052429, + "learning_rate": 1.5021261935479924e-06, + "loss": 1.5031, + "step": 976 + }, + { + "epoch": 2.6730506155950753, + "grad_norm": 0.29794323444366455, + "learning_rate": 1.4772370342442694e-06, + "loss": 1.4679, + "step": 977 + }, + { + "epoch": 2.6757865937072505, + "grad_norm": 0.31903737783432007, + "learning_rate": 1.4525495203425498e-06, + "loss": 1.6308, + "step": 978 + }, + { + "epoch": 2.6785225718194257, + "grad_norm": 0.3376205861568451, + "learning_rate": 1.428063863472895e-06, + "loss": 1.5566, + "step": 979 + }, + { + "epoch": 2.6812585499316004, + "grad_norm": 0.3032618761062622, + "learning_rate": 1.4037802735349743e-06, + "loss": 1.4632, + "step": 980 + }, + { + "epoch": 2.6839945280437756, + "grad_norm": 0.33116862177848816, + "learning_rate": 1.3796989586962544e-06, + "loss": 1.5811, + "step": 981 + }, + { + "epoch": 2.6867305061559508, + "grad_norm": 0.34618982672691345, + "learning_rate": 1.3558201253902531e-06, + "loss": 1.3734, + "step": 982 + }, + { + "epoch": 2.689466484268126, + "grad_norm": 0.3325488567352295, + "learning_rate": 1.3321439783147233e-06, + "loss": 1.4946, + "step": 983 + }, + { + "epoch": 2.692202462380301, + "grad_norm": 0.32722416520118713, + "learning_rate": 1.3086707204299414e-06, + "loss": 1.5425, + "step": 984 + }, + { + "epoch": 2.694938440492476, + "grad_norm": 0.3376059830188751, + "learning_rate": 1.2854005529569224e-06, + "loss": 1.4031, + "step": 985 + }, + { + "epoch": 2.697674418604651, + "grad_norm": 0.3256185054779053, + "learning_rate": 1.2623336753757348e-06, + "loss": 1.493, + "step": 986 + }, + { + "epoch": 2.7004103967168263, + "grad_norm": 0.3338906168937683, + "learning_rate": 1.2394702854237744e-06, + "loss": 1.5289, + "step": 987 + }, + { + "epoch": 2.7031463748290014, + "grad_norm": 0.31372472643852234, + "learning_rate": 1.2168105790940603e-06, + "loss": 1.5584, + "step": 988 + }, + { + "epoch": 2.7058823529411766, + "grad_norm": 0.3929591774940491, + "learning_rate": 1.1943547506335723e-06, + "loss": 1.482, + "step": 989 + }, + { + "epoch": 2.7086183310533514, + "grad_norm": 0.3206530809402466, + "learning_rate": 1.1721029925415738e-06, + "loss": 1.5406, + "step": 990 + }, + { + "epoch": 2.7113543091655266, + "grad_norm": 0.32885223627090454, + "learning_rate": 1.15005549556797e-06, + "loss": 1.3643, + "step": 991 + }, + { + "epoch": 2.7140902872777017, + "grad_norm": 0.3618469536304474, + "learning_rate": 1.1282124487116518e-06, + "loss": 1.5104, + "step": 992 + }, + { + "epoch": 2.716826265389877, + "grad_norm": 0.31617462635040283, + "learning_rate": 1.106574039218919e-06, + "loss": 1.4661, + "step": 993 + }, + { + "epoch": 2.719562243502052, + "grad_norm": 0.31909239292144775, + "learning_rate": 1.0851404525818204e-06, + "loss": 1.5708, + "step": 994 + }, + { + "epoch": 2.722298221614227, + "grad_norm": 0.3482041358947754, + "learning_rate": 1.0639118725366049e-06, + "loss": 1.4782, + "step": 995 + }, + { + "epoch": 2.725034199726402, + "grad_norm": 0.3372204601764679, + "learning_rate": 1.0428884810621403e-06, + "loss": 1.4833, + "step": 996 + }, + { + "epoch": 2.7277701778385772, + "grad_norm": 0.32383978366851807, + "learning_rate": 1.0220704583783242e-06, + "loss": 1.5989, + "step": 997 + }, + { + "epoch": 2.7305061559507524, + "grad_norm": 0.3657885491847992, + "learning_rate": 1.0014579829445842e-06, + "loss": 1.6592, + "step": 998 + }, + { + "epoch": 2.7332421340629276, + "grad_norm": 0.3399549722671509, + "learning_rate": 9.810512314583075e-07, + "loss": 1.3902, + "step": 999 + }, + { + "epoch": 2.7359781121751023, + "grad_norm": 0.3272568881511688, + "learning_rate": 9.60850378853348e-07, + "loss": 1.3684, + "step": 1000 + }, + { + "epoch": 2.7359781121751023, + "eval_loss": 1.5569868087768555, + "eval_runtime": 13.8259, + "eval_samples_per_second": 11.138, + "eval_steps_per_second": 11.138, + "step": 1000 + }, + { + "epoch": 2.738714090287278, + "grad_norm": 0.3006269037723541, + "learning_rate": 9.408555982985262e-07, + "loss": 1.4081, + "step": 1001 + }, + { + "epoch": 2.7414500683994527, + "grad_norm": 0.32424142956733704, + "learning_rate": 9.210670611961375e-07, + "loss": 1.47, + "step": 1002 + }, + { + "epoch": 2.744186046511628, + "grad_norm": 0.3788749873638153, + "learning_rate": 9.014849371804829e-07, + "loss": 1.4683, + "step": 1003 + }, + { + "epoch": 2.746922024623803, + "grad_norm": 0.3551234006881714, + "learning_rate": 8.821093941164233e-07, + "loss": 1.6791, + "step": 1004 + }, + { + "epoch": 2.7496580027359783, + "grad_norm": 0.38576632738113403, + "learning_rate": 8.629405980979332e-07, + "loss": 1.49, + "step": 1005 + }, + { + "epoch": 2.7523939808481535, + "grad_norm": 0.3681055009365082, + "learning_rate": 8.439787134466692e-07, + "loss": 1.4691, + "step": 1006 + }, + { + "epoch": 2.755129958960328, + "grad_norm": 0.3119199872016907, + "learning_rate": 8.252239027105924e-07, + "loss": 1.5351, + "step": 1007 + }, + { + "epoch": 2.7578659370725034, + "grad_norm": 0.31998804211616516, + "learning_rate": 8.066763266625282e-07, + "loss": 1.468, + "step": 1008 + }, + { + "epoch": 2.7606019151846786, + "grad_norm": 0.3693805932998657, + "learning_rate": 7.883361442988374e-07, + "loss": 1.5278, + "step": 1009 + }, + { + "epoch": 2.7633378932968538, + "grad_norm": 0.3538669943809509, + "learning_rate": 7.702035128380053e-07, + "loss": 1.5411, + "step": 1010 + }, + { + "epoch": 2.766073871409029, + "grad_norm": 0.3684946596622467, + "learning_rate": 7.522785877193378e-07, + "loss": 1.4696, + "step": 1011 + }, + { + "epoch": 2.7688098495212037, + "grad_norm": 0.28718748688697815, + "learning_rate": 7.345615226016011e-07, + "loss": 1.5301, + "step": 1012 + }, + { + "epoch": 2.771545827633379, + "grad_norm": 0.33110928535461426, + "learning_rate": 7.170524693617064e-07, + "loss": 1.6021, + "step": 1013 + }, + { + "epoch": 2.774281805745554, + "grad_norm": 0.31040212512016296, + "learning_rate": 6.997515780934244e-07, + "loss": 1.4569, + "step": 1014 + }, + { + "epoch": 2.7770177838577292, + "grad_norm": 0.3069436848163605, + "learning_rate": 6.826589971060837e-07, + "loss": 1.4963, + "step": 1015 + }, + { + "epoch": 2.7797537619699044, + "grad_norm": 0.3085753321647644, + "learning_rate": 6.657748729233026e-07, + "loss": 1.6278, + "step": 1016 + }, + { + "epoch": 2.782489740082079, + "grad_norm": 0.31628352403640747, + "learning_rate": 6.490993502817289e-07, + "loss": 1.4166, + "step": 1017 + }, + { + "epoch": 2.7852257181942544, + "grad_norm": 0.32792040705680847, + "learning_rate": 6.32632572129821e-07, + "loss": 1.4757, + "step": 1018 + }, + { + "epoch": 2.7879616963064295, + "grad_norm": 0.34112995862960815, + "learning_rate": 6.163746796265885e-07, + "loss": 1.4505, + "step": 1019 + }, + { + "epoch": 2.7906976744186047, + "grad_norm": 0.3526972830295563, + "learning_rate": 6.003258121404148e-07, + "loss": 1.4417, + "step": 1020 + }, + { + "epoch": 2.79343365253078, + "grad_norm": 0.3196633458137512, + "learning_rate": 5.844861072478336e-07, + "loss": 1.5024, + "step": 1021 + }, + { + "epoch": 2.7961696306429547, + "grad_norm": 0.3368687033653259, + "learning_rate": 5.688557007323708e-07, + "loss": 1.6036, + "step": 1022 + }, + { + "epoch": 2.79890560875513, + "grad_norm": 0.33089619874954224, + "learning_rate": 5.534347265833711e-07, + "loss": 1.5716, + "step": 1023 + }, + { + "epoch": 2.801641586867305, + "grad_norm": 0.35482943058013916, + "learning_rate": 5.382233169948541e-07, + "loss": 1.5263, + "step": 1024 + }, + { + "epoch": 2.80437756497948, + "grad_norm": 0.3278489112854004, + "learning_rate": 5.232216023643711e-07, + "loss": 1.4273, + "step": 1025 + }, + { + "epoch": 2.8071135430916554, + "grad_norm": 0.3070161044597626, + "learning_rate": 5.084297112919051e-07, + "loss": 1.7237, + "step": 1026 + }, + { + "epoch": 2.80984952120383, + "grad_norm": 0.3581262528896332, + "learning_rate": 4.938477705787508e-07, + "loss": 1.4228, + "step": 1027 + }, + { + "epoch": 2.8125854993160053, + "grad_norm": 0.3500869572162628, + "learning_rate": 4.794759052264259e-07, + "loss": 1.6744, + "step": 1028 + }, + { + "epoch": 2.8153214774281805, + "grad_norm": 0.3327549397945404, + "learning_rate": 4.653142384356324e-07, + "loss": 1.3299, + "step": 1029 + }, + { + "epoch": 2.8180574555403557, + "grad_norm": 0.32055315375328064, + "learning_rate": 4.513628916051532e-07, + "loss": 1.5822, + "step": 1030 + }, + { + "epoch": 2.820793433652531, + "grad_norm": 0.35483717918395996, + "learning_rate": 4.376219843308438e-07, + "loss": 1.4478, + "step": 1031 + }, + { + "epoch": 2.8235294117647056, + "grad_norm": 0.3212404251098633, + "learning_rate": 4.2409163440459697e-07, + "loss": 1.6226, + "step": 1032 + }, + { + "epoch": 2.8262653898768813, + "grad_norm": 0.3329022526741028, + "learning_rate": 4.107719578133301e-07, + "loss": 1.5192, + "step": 1033 + }, + { + "epoch": 2.829001367989056, + "grad_norm": 0.3651573061943054, + "learning_rate": 3.9766306873801064e-07, + "loss": 1.5169, + "step": 1034 + }, + { + "epoch": 2.831737346101231, + "grad_norm": 0.3816221356391907, + "learning_rate": 3.8476507955263475e-07, + "loss": 1.39, + "step": 1035 + }, + { + "epoch": 2.8344733242134064, + "grad_norm": 0.35796451568603516, + "learning_rate": 3.720781008233143e-07, + "loss": 1.4573, + "step": 1036 + }, + { + "epoch": 2.8372093023255816, + "grad_norm": 0.3537936210632324, + "learning_rate": 3.5960224130728857e-07, + "loss": 1.5344, + "step": 1037 + }, + { + "epoch": 2.8399452804377567, + "grad_norm": 0.35659074783325195, + "learning_rate": 3.473376079520224e-07, + "loss": 1.3814, + "step": 1038 + }, + { + "epoch": 2.8426812585499315, + "grad_norm": 0.31420376896858215, + "learning_rate": 3.352843058942623e-07, + "loss": 1.4956, + "step": 1039 + }, + { + "epoch": 2.8454172366621067, + "grad_norm": 0.31859591603279114, + "learning_rate": 3.2344243845915945e-07, + "loss": 1.6171, + "step": 1040 + }, + { + "epoch": 2.848153214774282, + "grad_norm": 0.30240893363952637, + "learning_rate": 3.118121071593677e-07, + "loss": 1.4719, + "step": 1041 + }, + { + "epoch": 2.850889192886457, + "grad_norm": 0.3180094361305237, + "learning_rate": 3.003934116941776e-07, + "loss": 1.4596, + "step": 1042 + }, + { + "epoch": 2.853625170998632, + "grad_norm": 0.3070809841156006, + "learning_rate": 2.891864499486724e-07, + "loss": 1.5858, + "step": 1043 + }, + { + "epoch": 2.856361149110807, + "grad_norm": 0.3325824737548828, + "learning_rate": 2.7819131799285967e-07, + "loss": 1.5769, + "step": 1044 + }, + { + "epoch": 2.859097127222982, + "grad_norm": 0.3298075199127197, + "learning_rate": 2.6740811008089393e-07, + "loss": 1.5655, + "step": 1045 + }, + { + "epoch": 2.8618331053351573, + "grad_norm": 0.35839757323265076, + "learning_rate": 2.568369186502162e-07, + "loss": 1.3463, + "step": 1046 + }, + { + "epoch": 2.8645690834473325, + "grad_norm": 0.3779064118862152, + "learning_rate": 2.464778343208074e-07, + "loss": 1.4612, + "step": 1047 + }, + { + "epoch": 2.8673050615595077, + "grad_norm": 0.32933345437049866, + "learning_rate": 2.36330945894378e-07, + "loss": 1.4131, + "step": 1048 + }, + { + "epoch": 2.8700410396716824, + "grad_norm": 0.3202325999736786, + "learning_rate": 2.2639634035363234e-07, + "loss": 1.5189, + "step": 1049 + }, + { + "epoch": 2.8727770177838576, + "grad_norm": 0.3435043394565582, + "learning_rate": 2.1667410286149992e-07, + "loss": 1.4026, + "step": 1050 + }, + { + "epoch": 2.875512995896033, + "grad_norm": 0.33565640449523926, + "learning_rate": 2.071643167604248e-07, + "loss": 1.4982, + "step": 1051 + }, + { + "epoch": 2.878248974008208, + "grad_norm": 0.31919005513191223, + "learning_rate": 1.9786706357163842e-07, + "loss": 1.6241, + "step": 1052 + }, + { + "epoch": 2.880984952120383, + "grad_norm": 0.33048737049102783, + "learning_rate": 1.8878242299446848e-07, + "loss": 1.586, + "step": 1053 + }, + { + "epoch": 2.883720930232558, + "grad_norm": 0.31295284628868103, + "learning_rate": 1.7991047290565056e-07, + "loss": 1.5296, + "step": 1054 + }, + { + "epoch": 2.886456908344733, + "grad_norm": 0.30597612261772156, + "learning_rate": 1.712512893586593e-07, + "loss": 1.4865, + "step": 1055 + }, + { + "epoch": 2.8891928864569083, + "grad_norm": 0.3064374327659607, + "learning_rate": 1.6280494658307e-07, + "loss": 1.5211, + "step": 1056 + }, + { + "epoch": 2.8919288645690835, + "grad_norm": 0.3949945867061615, + "learning_rate": 1.5457151698390347e-07, + "loss": 1.5316, + "step": 1057 + }, + { + "epoch": 2.8946648426812587, + "grad_norm": 0.4148292541503906, + "learning_rate": 1.4655107114101007e-07, + "loss": 1.3986, + "step": 1058 + }, + { + "epoch": 2.8974008207934334, + "grad_norm": 0.32118478417396545, + "learning_rate": 1.387436778084783e-07, + "loss": 1.7486, + "step": 1059 + }, + { + "epoch": 2.9001367989056086, + "grad_norm": 0.34837549924850464, + "learning_rate": 1.3114940391403263e-07, + "loss": 1.3545, + "step": 1060 + }, + { + "epoch": 2.902872777017784, + "grad_norm": 0.331442266702652, + "learning_rate": 1.2376831455845893e-07, + "loss": 1.5355, + "step": 1061 + }, + { + "epoch": 2.905608755129959, + "grad_norm": 0.3617335855960846, + "learning_rate": 1.1660047301505495e-07, + "loss": 1.5907, + "step": 1062 + }, + { + "epoch": 2.908344733242134, + "grad_norm": 0.32971206307411194, + "learning_rate": 1.0964594072907786e-07, + "loss": 1.6133, + "step": 1063 + }, + { + "epoch": 2.911080711354309, + "grad_norm": 0.34654051065444946, + "learning_rate": 1.0290477731722814e-07, + "loss": 1.7007, + "step": 1064 + }, + { + "epoch": 2.9138166894664845, + "grad_norm": 0.3612649738788605, + "learning_rate": 9.637704056713049e-08, + "loss": 1.2852, + "step": 1065 + }, + { + "epoch": 2.9165526675786593, + "grad_norm": 0.2997490465641022, + "learning_rate": 9.006278643683696e-08, + "loss": 1.5689, + "step": 1066 + }, + { + "epoch": 2.9192886456908345, + "grad_norm": 0.3305412828922272, + "learning_rate": 8.396206905436355e-08, + "loss": 1.5476, + "step": 1067 + }, + { + "epoch": 2.9220246238030096, + "grad_norm": 0.36031877994537354, + "learning_rate": 7.807494071720156e-08, + "loss": 1.6777, + "step": 1068 + }, + { + "epoch": 2.924760601915185, + "grad_norm": 0.3095710873603821, + "learning_rate": 7.24014518918903e-08, + "loss": 1.589, + "step": 1069 + }, + { + "epoch": 2.92749658002736, + "grad_norm": 0.30806005001068115, + "learning_rate": 6.694165121356732e-08, + "loss": 1.4103, + "step": 1070 + }, + { + "epoch": 2.9302325581395348, + "grad_norm": 0.36173051595687866, + "learning_rate": 6.169558548556887e-08, + "loss": 1.3676, + "step": 1071 + }, + { + "epoch": 2.93296853625171, + "grad_norm": 0.2915642559528351, + "learning_rate": 5.6663299679018955e-08, + "loss": 1.4523, + "step": 1072 + }, + { + "epoch": 2.935704514363885, + "grad_norm": 0.3511941730976105, + "learning_rate": 5.1844836932438134e-08, + "loss": 1.5194, + "step": 1073 + }, + { + "epoch": 2.9384404924760603, + "grad_norm": 0.3125559389591217, + "learning_rate": 4.7240238551385376e-08, + "loss": 1.4604, + "step": 1074 + }, + { + "epoch": 2.9411764705882355, + "grad_norm": 0.3312918245792389, + "learning_rate": 4.284954400810004e-08, + "loss": 1.4807, + "step": 1075 + }, + { + "epoch": 2.9439124487004102, + "grad_norm": 0.3593026101589203, + "learning_rate": 3.86727909411605e-08, + "loss": 1.7197, + "step": 1076 + }, + { + "epoch": 2.9466484268125854, + "grad_norm": 0.3443048298358917, + "learning_rate": 3.471001515516215e-08, + "loss": 1.6237, + "step": 1077 + }, + { + "epoch": 2.9493844049247606, + "grad_norm": 0.32331955432891846, + "learning_rate": 3.096125062041488e-08, + "loss": 1.5434, + "step": 1078 + }, + { + "epoch": 2.952120383036936, + "grad_norm": 0.3328414261341095, + "learning_rate": 2.7426529472648898e-08, + "loss": 1.6007, + "step": 1079 + }, + { + "epoch": 2.954856361149111, + "grad_norm": 0.31277886033058167, + "learning_rate": 2.4105882012734336e-08, + "loss": 1.6117, + "step": 1080 + }, + { + "epoch": 2.9575923392612857, + "grad_norm": 0.32368844747543335, + "learning_rate": 2.0999336706434282e-08, + "loss": 1.4795, + "step": 1081 + }, + { + "epoch": 2.960328317373461, + "grad_norm": 0.3160129487514496, + "learning_rate": 1.810692018414384e-08, + "loss": 1.5325, + "step": 1082 + }, + { + "epoch": 2.963064295485636, + "grad_norm": 0.3517950773239136, + "learning_rate": 1.5428657240676437e-08, + "loss": 1.4103, + "step": 1083 + }, + { + "epoch": 2.9658002735978113, + "grad_norm": 0.36211562156677246, + "learning_rate": 1.2964570835047318e-08, + "loss": 1.6729, + "step": 1084 + }, + { + "epoch": 2.9685362517099865, + "grad_norm": 0.3121832609176636, + "learning_rate": 1.0714682090270933e-08, + "loss": 1.3272, + "step": 1085 + }, + { + "epoch": 2.971272229822161, + "grad_norm": 0.33123496174812317, + "learning_rate": 8.679010293183298e-09, + "loss": 1.506, + "step": 1086 + }, + { + "epoch": 2.9740082079343364, + "grad_norm": 0.32456517219543457, + "learning_rate": 6.857572894281017e-09, + "loss": 1.5943, + "step": 1087 + }, + { + "epoch": 2.9767441860465116, + "grad_norm": 0.3651670813560486, + "learning_rate": 5.250385507568623e-09, + "loss": 1.4488, + "step": 1088 + }, + { + "epoch": 2.9794801641586868, + "grad_norm": 0.3096288740634918, + "learning_rate": 3.8574619104170265e-09, + "loss": 1.4775, + "step": 1089 + }, + { + "epoch": 2.982216142270862, + "grad_norm": 0.33430325984954834, + "learning_rate": 2.67881404346082e-09, + "loss": 1.5276, + "step": 1090 + }, + { + "epoch": 2.9849521203830367, + "grad_norm": 0.2988739609718323, + "learning_rate": 1.7144520104844798e-09, + "loss": 1.4893, + "step": 1091 + }, + { + "epoch": 2.987688098495212, + "grad_norm": 0.35142961144447327, + "learning_rate": 9.643840783363224e-10, + "loss": 1.4864, + "step": 1092 + }, + { + "epoch": 2.990424076607387, + "grad_norm": 0.2971309721469879, + "learning_rate": 4.2861667686744377e-10, + "loss": 1.6171, + "step": 1093 + }, + { + "epoch": 2.9931600547195623, + "grad_norm": 0.36940792202949524, + "learning_rate": 1.0715439885677913e-10, + "loss": 1.5264, + "step": 1094 + }, + { + "epoch": 2.9958960328317374, + "grad_norm": 0.31959155201911926, + "learning_rate": 0.0, + "loss": 1.585, + "step": 1095 + } + ], + "logging_steps": 1, + "max_steps": 1095, + "num_input_tokens_seen": 0, + "num_train_epochs": 3, + "save_steps": 250, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 1.0313245522312704e+16, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +}