diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,36513 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 5.0, + "eval_steps": 500, + "global_step": 5210, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.0009601536245799327, + "grad_norm": 4.608575884356107, + "learning_rate": 0.0, + "loss": 0.9304, + "step": 1 + }, + { + "epoch": 0.0019203072491598655, + "grad_norm": 5.267697768316677, + "learning_rate": 1.9193857965451055e-08, + "loss": 0.8959, + "step": 2 + }, + { + "epoch": 0.0028804608737397984, + "grad_norm": 3.674166203239673, + "learning_rate": 3.838771593090211e-08, + "loss": 0.9946, + "step": 3 + }, + { + "epoch": 0.003840614498319731, + "grad_norm": 4.212680693162968, + "learning_rate": 5.758157389635317e-08, + "loss": 1.0958, + "step": 4 + }, + { + "epoch": 0.004800768122899664, + "grad_norm": 4.966685183363455, + "learning_rate": 7.677543186180422e-08, + "loss": 1.1641, + "step": 5 + }, + { + "epoch": 0.005760921747479597, + "grad_norm": 3.778226764973662, + "learning_rate": 9.596928982725528e-08, + "loss": 0.9914, + "step": 6 + }, + { + "epoch": 0.00672107537205953, + "grad_norm": 4.083675255841295, + "learning_rate": 1.1516314779270635e-07, + "loss": 0.9815, + "step": 7 + }, + { + "epoch": 0.007681228996639462, + "grad_norm": 4.457244690216626, + "learning_rate": 1.343570057581574e-07, + "loss": 1.1257, + "step": 8 + }, + { + "epoch": 0.008641382621219395, + "grad_norm": 4.297331061465192, + "learning_rate": 1.5355086372360844e-07, + "loss": 1.1114, + "step": 9 + }, + { + "epoch": 0.009601536245799328, + "grad_norm": 3.475381879941915, + "learning_rate": 1.7274472168905953e-07, + "loss": 0.9588, + "step": 10 + }, + { + "epoch": 0.01056168987037926, + "grad_norm": 4.019411039718911, + "learning_rate": 1.9193857965451055e-07, + "loss": 1.082, + "step": 11 + }, + { + "epoch": 0.011521843494959194, + "grad_norm": 4.298656013398512, + "learning_rate": 2.1113243761996164e-07, + "loss": 0.9894, + "step": 12 + }, + { + "epoch": 0.012481997119539127, + "grad_norm": 4.298841455109918, + "learning_rate": 2.303262955854127e-07, + "loss": 1.0283, + "step": 13 + }, + { + "epoch": 0.01344215074411906, + "grad_norm": 3.8125619523001975, + "learning_rate": 2.4952015355086375e-07, + "loss": 0.9738, + "step": 14 + }, + { + "epoch": 0.014402304368698993, + "grad_norm": 5.948867639867879, + "learning_rate": 2.687140115163148e-07, + "loss": 1.1343, + "step": 15 + }, + { + "epoch": 0.015362457993278924, + "grad_norm": 4.516786504025045, + "learning_rate": 2.8790786948176586e-07, + "loss": 1.1204, + "step": 16 + }, + { + "epoch": 0.01632261161785886, + "grad_norm": 3.6831244206580913, + "learning_rate": 3.071017274472169e-07, + "loss": 0.9958, + "step": 17 + }, + { + "epoch": 0.01728276524243879, + "grad_norm": 3.9360004704440303, + "learning_rate": 3.262955854126679e-07, + "loss": 1.0483, + "step": 18 + }, + { + "epoch": 0.018242918867018725, + "grad_norm": 3.901791783080527, + "learning_rate": 3.4548944337811905e-07, + "loss": 1.0408, + "step": 19 + }, + { + "epoch": 0.019203072491598656, + "grad_norm": 4.3095324049321935, + "learning_rate": 3.646833013435701e-07, + "loss": 0.965, + "step": 20 + }, + { + "epoch": 0.020163226116178587, + "grad_norm": 4.148314319671607, + "learning_rate": 3.838771593090211e-07, + "loss": 1.1593, + "step": 21 + }, + { + "epoch": 0.02112337974075852, + "grad_norm": 3.535029565945762, + "learning_rate": 4.0307101727447224e-07, + "loss": 0.9418, + "step": 22 + }, + { + "epoch": 0.022083533365338453, + "grad_norm": 3.243808885930317, + "learning_rate": 4.2226487523992327e-07, + "loss": 0.9671, + "step": 23 + }, + { + "epoch": 0.023043686989918388, + "grad_norm": 3.1901459239600882, + "learning_rate": 4.414587332053743e-07, + "loss": 0.9802, + "step": 24 + }, + { + "epoch": 0.02400384061449832, + "grad_norm": 3.2819870328486016, + "learning_rate": 4.606525911708254e-07, + "loss": 1.0308, + "step": 25 + }, + { + "epoch": 0.024963994239078253, + "grad_norm": 3.5977491618398236, + "learning_rate": 4.798464491362765e-07, + "loss": 0.987, + "step": 26 + }, + { + "epoch": 0.025924147863658185, + "grad_norm": 3.3874834219166745, + "learning_rate": 4.990403071017275e-07, + "loss": 0.941, + "step": 27 + }, + { + "epoch": 0.02688430148823812, + "grad_norm": 3.240612452815504, + "learning_rate": 5.182341650671785e-07, + "loss": 1.072, + "step": 28 + }, + { + "epoch": 0.02784445511281805, + "grad_norm": 4.48751661697212, + "learning_rate": 5.374280230326296e-07, + "loss": 1.1463, + "step": 29 + }, + { + "epoch": 0.028804608737397985, + "grad_norm": 3.3988821883857896, + "learning_rate": 5.566218809980807e-07, + "loss": 1.0261, + "step": 30 + }, + { + "epoch": 0.029764762361977917, + "grad_norm": 3.2127956160200215, + "learning_rate": 5.758157389635317e-07, + "loss": 0.8477, + "step": 31 + }, + { + "epoch": 0.030724915986557848, + "grad_norm": 3.3352057499764154, + "learning_rate": 5.950095969289827e-07, + "loss": 0.9437, + "step": 32 + }, + { + "epoch": 0.03168506961113778, + "grad_norm": 3.060516597467567, + "learning_rate": 6.142034548944338e-07, + "loss": 1.0675, + "step": 33 + }, + { + "epoch": 0.03264522323571772, + "grad_norm": 2.308625467912329, + "learning_rate": 6.333973128598848e-07, + "loss": 0.9703, + "step": 34 + }, + { + "epoch": 0.033605376860297645, + "grad_norm": 2.187053407523522, + "learning_rate": 6.525911708253358e-07, + "loss": 1.006, + "step": 35 + }, + { + "epoch": 0.03456553048487758, + "grad_norm": 2.8405778838587903, + "learning_rate": 6.717850287907871e-07, + "loss": 0.8908, + "step": 36 + }, + { + "epoch": 0.035525684109457514, + "grad_norm": 2.566314021468516, + "learning_rate": 6.909788867562381e-07, + "loss": 0.9259, + "step": 37 + }, + { + "epoch": 0.03648583773403745, + "grad_norm": 3.0130066930691988, + "learning_rate": 7.101727447216891e-07, + "loss": 0.9417, + "step": 38 + }, + { + "epoch": 0.03744599135861738, + "grad_norm": 2.21081996209867, + "learning_rate": 7.293666026871402e-07, + "loss": 0.9622, + "step": 39 + }, + { + "epoch": 0.03840614498319731, + "grad_norm": 1.8475861673862382, + "learning_rate": 7.485604606525912e-07, + "loss": 0.7887, + "step": 40 + }, + { + "epoch": 0.039366298607777246, + "grad_norm": 2.381750801892437, + "learning_rate": 7.677543186180422e-07, + "loss": 0.8993, + "step": 41 + }, + { + "epoch": 0.040326452232357174, + "grad_norm": 2.295035679157881, + "learning_rate": 7.869481765834934e-07, + "loss": 1.0287, + "step": 42 + }, + { + "epoch": 0.04128660585693711, + "grad_norm": 2.3462274641458474, + "learning_rate": 8.061420345489445e-07, + "loss": 0.9755, + "step": 43 + }, + { + "epoch": 0.04224675948151704, + "grad_norm": 1.6515479761903826, + "learning_rate": 8.253358925143955e-07, + "loss": 0.8174, + "step": 44 + }, + { + "epoch": 0.04320691310609698, + "grad_norm": 1.7674375018904431, + "learning_rate": 8.445297504798465e-07, + "loss": 0.9022, + "step": 45 + }, + { + "epoch": 0.044167066730676906, + "grad_norm": 1.9577421293495636, + "learning_rate": 8.637236084452976e-07, + "loss": 1.0111, + "step": 46 + }, + { + "epoch": 0.04512722035525684, + "grad_norm": 1.7020469055537244, + "learning_rate": 8.829174664107486e-07, + "loss": 0.819, + "step": 47 + }, + { + "epoch": 0.046087373979836775, + "grad_norm": 2.01572674952168, + "learning_rate": 9.021113243761997e-07, + "loss": 0.9874, + "step": 48 + }, + { + "epoch": 0.04704752760441671, + "grad_norm": 2.3600284799271622, + "learning_rate": 9.213051823416508e-07, + "loss": 1.0881, + "step": 49 + }, + { + "epoch": 0.04800768122899664, + "grad_norm": 2.1334878781214934, + "learning_rate": 9.404990403071018e-07, + "loss": 1.0728, + "step": 50 + }, + { + "epoch": 0.04896783485357657, + "grad_norm": 2.081635795467802, + "learning_rate": 9.59692898272553e-07, + "loss": 1.0563, + "step": 51 + }, + { + "epoch": 0.04992798847815651, + "grad_norm": 1.872979323941484, + "learning_rate": 9.788867562380039e-07, + "loss": 0.7869, + "step": 52 + }, + { + "epoch": 0.050888142102736435, + "grad_norm": 2.253948697980142, + "learning_rate": 9.98080614203455e-07, + "loss": 1.0892, + "step": 53 + }, + { + "epoch": 0.05184829572731637, + "grad_norm": 1.7260286729835796, + "learning_rate": 1.0172744721689061e-06, + "loss": 0.922, + "step": 54 + }, + { + "epoch": 0.052808449351896304, + "grad_norm": 1.7645361159227275, + "learning_rate": 1.036468330134357e-06, + "loss": 0.963, + "step": 55 + }, + { + "epoch": 0.05376860297647624, + "grad_norm": 1.8586055165595463, + "learning_rate": 1.0556621880998082e-06, + "loss": 0.999, + "step": 56 + }, + { + "epoch": 0.05472875660105617, + "grad_norm": 1.843097634523443, + "learning_rate": 1.074856046065259e-06, + "loss": 1.009, + "step": 57 + }, + { + "epoch": 0.0556889102256361, + "grad_norm": 1.7610200133615168, + "learning_rate": 1.0940499040307102e-06, + "loss": 0.9579, + "step": 58 + }, + { + "epoch": 0.056649063850216036, + "grad_norm": 1.7138745961014983, + "learning_rate": 1.1132437619961614e-06, + "loss": 0.9182, + "step": 59 + }, + { + "epoch": 0.05760921747479597, + "grad_norm": 2.134408662198101, + "learning_rate": 1.1324376199616125e-06, + "loss": 0.8898, + "step": 60 + }, + { + "epoch": 0.0585693710993759, + "grad_norm": 1.2519878171090029, + "learning_rate": 1.1516314779270634e-06, + "loss": 0.8702, + "step": 61 + }, + { + "epoch": 0.05952952472395583, + "grad_norm": 1.1554566479823363, + "learning_rate": 1.1708253358925146e-06, + "loss": 0.7871, + "step": 62 + }, + { + "epoch": 0.06048967834853577, + "grad_norm": 1.6615875454598308, + "learning_rate": 1.1900191938579655e-06, + "loss": 0.9018, + "step": 63 + }, + { + "epoch": 0.061449831973115696, + "grad_norm": 1.7505284195298252, + "learning_rate": 1.2092130518234166e-06, + "loss": 1.0219, + "step": 64 + }, + { + "epoch": 0.06240998559769563, + "grad_norm": 1.5611445163160693, + "learning_rate": 1.2284069097888675e-06, + "loss": 0.9937, + "step": 65 + }, + { + "epoch": 0.06337013922227556, + "grad_norm": 1.0975898539730802, + "learning_rate": 1.2476007677543187e-06, + "loss": 0.7717, + "step": 66 + }, + { + "epoch": 0.0643302928468555, + "grad_norm": 1.2643294950307615, + "learning_rate": 1.2667946257197696e-06, + "loss": 0.9343, + "step": 67 + }, + { + "epoch": 0.06529044647143543, + "grad_norm": 1.2926314844705193, + "learning_rate": 1.2859884836852207e-06, + "loss": 0.9389, + "step": 68 + }, + { + "epoch": 0.06625060009601537, + "grad_norm": 1.5008851206841063, + "learning_rate": 1.3051823416506717e-06, + "loss": 0.9279, + "step": 69 + }, + { + "epoch": 0.06721075372059529, + "grad_norm": 1.4376243421140087, + "learning_rate": 1.324376199616123e-06, + "loss": 1.0277, + "step": 70 + }, + { + "epoch": 0.06817090734517522, + "grad_norm": 1.6332295422438075, + "learning_rate": 1.3435700575815741e-06, + "loss": 0.9154, + "step": 71 + }, + { + "epoch": 0.06913106096975516, + "grad_norm": 1.1937192619764858, + "learning_rate": 1.362763915547025e-06, + "loss": 0.865, + "step": 72 + }, + { + "epoch": 0.0700912145943351, + "grad_norm": 1.6160555706501898, + "learning_rate": 1.3819577735124762e-06, + "loss": 0.8731, + "step": 73 + }, + { + "epoch": 0.07105136821891503, + "grad_norm": 1.2828558000342634, + "learning_rate": 1.4011516314779271e-06, + "loss": 0.8508, + "step": 74 + }, + { + "epoch": 0.07201152184349496, + "grad_norm": 1.1433690942334571, + "learning_rate": 1.4203454894433783e-06, + "loss": 0.7113, + "step": 75 + }, + { + "epoch": 0.0729716754680749, + "grad_norm": 1.3498602336942411, + "learning_rate": 1.4395393474088292e-06, + "loss": 0.8771, + "step": 76 + }, + { + "epoch": 0.07393182909265482, + "grad_norm": 1.1090404504798739, + "learning_rate": 1.4587332053742803e-06, + "loss": 0.8389, + "step": 77 + }, + { + "epoch": 0.07489198271723475, + "grad_norm": 1.3371889492769014, + "learning_rate": 1.4779270633397312e-06, + "loss": 0.9674, + "step": 78 + }, + { + "epoch": 0.07585213634181469, + "grad_norm": 1.004920469127281, + "learning_rate": 1.4971209213051824e-06, + "loss": 0.8359, + "step": 79 + }, + { + "epoch": 0.07681228996639462, + "grad_norm": 1.1309480698032008, + "learning_rate": 1.5163147792706335e-06, + "loss": 0.871, + "step": 80 + }, + { + "epoch": 0.07777244359097456, + "grad_norm": 1.0926935444754795, + "learning_rate": 1.5355086372360844e-06, + "loss": 0.831, + "step": 81 + }, + { + "epoch": 0.07873259721555449, + "grad_norm": 1.3243569741514198, + "learning_rate": 1.5547024952015358e-06, + "loss": 0.929, + "step": 82 + }, + { + "epoch": 0.07969275084013443, + "grad_norm": 1.1459157332069554, + "learning_rate": 1.5738963531669867e-06, + "loss": 0.8642, + "step": 83 + }, + { + "epoch": 0.08065290446471435, + "grad_norm": 1.6430972393724526, + "learning_rate": 1.5930902111324378e-06, + "loss": 0.8614, + "step": 84 + }, + { + "epoch": 0.08161305808929428, + "grad_norm": 0.9635181271227865, + "learning_rate": 1.612284069097889e-06, + "loss": 0.8327, + "step": 85 + }, + { + "epoch": 0.08257321171387422, + "grad_norm": 1.0793717741613553, + "learning_rate": 1.63147792706334e-06, + "loss": 0.8913, + "step": 86 + }, + { + "epoch": 0.08353336533845415, + "grad_norm": 1.0314798311669489, + "learning_rate": 1.650671785028791e-06, + "loss": 0.7473, + "step": 87 + }, + { + "epoch": 0.08449351896303409, + "grad_norm": 1.135066098383753, + "learning_rate": 1.669865642994242e-06, + "loss": 0.8715, + "step": 88 + }, + { + "epoch": 0.08545367258761402, + "grad_norm": 1.1920535091940676, + "learning_rate": 1.689059500959693e-06, + "loss": 0.9432, + "step": 89 + }, + { + "epoch": 0.08641382621219396, + "grad_norm": 1.5782636211463792, + "learning_rate": 1.708253358925144e-06, + "loss": 1.0285, + "step": 90 + }, + { + "epoch": 0.08737397983677389, + "grad_norm": 1.3099591913322186, + "learning_rate": 1.7274472168905951e-06, + "loss": 0.9003, + "step": 91 + }, + { + "epoch": 0.08833413346135381, + "grad_norm": 1.1075263821827226, + "learning_rate": 1.746641074856046e-06, + "loss": 0.9024, + "step": 92 + }, + { + "epoch": 0.08929428708593375, + "grad_norm": 1.0596908140580912, + "learning_rate": 1.7658349328214972e-06, + "loss": 0.8997, + "step": 93 + }, + { + "epoch": 0.09025444071051368, + "grad_norm": 1.6824209174093476, + "learning_rate": 1.7850287907869481e-06, + "loss": 0.943, + "step": 94 + }, + { + "epoch": 0.09121459433509362, + "grad_norm": 1.4178621073994107, + "learning_rate": 1.8042226487523995e-06, + "loss": 0.9556, + "step": 95 + }, + { + "epoch": 0.09217474795967355, + "grad_norm": 1.9323428454897225, + "learning_rate": 1.8234165067178506e-06, + "loss": 0.9296, + "step": 96 + }, + { + "epoch": 0.09313490158425348, + "grad_norm": 1.113187686000337, + "learning_rate": 1.8426103646833015e-06, + "loss": 0.8704, + "step": 97 + }, + { + "epoch": 0.09409505520883342, + "grad_norm": 1.2127175436399928, + "learning_rate": 1.8618042226487527e-06, + "loss": 0.9178, + "step": 98 + }, + { + "epoch": 0.09505520883341334, + "grad_norm": 1.4205983901354309, + "learning_rate": 1.8809980806142036e-06, + "loss": 0.7902, + "step": 99 + }, + { + "epoch": 0.09601536245799328, + "grad_norm": 1.2336160656741093, + "learning_rate": 1.9001919385796547e-06, + "loss": 0.9751, + "step": 100 + }, + { + "epoch": 0.09697551608257321, + "grad_norm": 1.32218061984364, + "learning_rate": 1.919385796545106e-06, + "loss": 0.9422, + "step": 101 + }, + { + "epoch": 0.09793566970715314, + "grad_norm": 1.4349730066378146, + "learning_rate": 1.9385796545105568e-06, + "loss": 0.9295, + "step": 102 + }, + { + "epoch": 0.09889582333173308, + "grad_norm": 0.9816231563605403, + "learning_rate": 1.9577735124760077e-06, + "loss": 0.7518, + "step": 103 + }, + { + "epoch": 0.09985597695631301, + "grad_norm": 1.1071092246727834, + "learning_rate": 1.9769673704414586e-06, + "loss": 0.8828, + "step": 104 + }, + { + "epoch": 0.10081613058089295, + "grad_norm": 1.0429061975034752, + "learning_rate": 1.99616122840691e-06, + "loss": 0.7532, + "step": 105 + }, + { + "epoch": 0.10177628420547287, + "grad_norm": 1.3352134212660092, + "learning_rate": 2.015355086372361e-06, + "loss": 0.9469, + "step": 106 + }, + { + "epoch": 0.1027364378300528, + "grad_norm": 0.9859105142366051, + "learning_rate": 2.0345489443378122e-06, + "loss": 0.7679, + "step": 107 + }, + { + "epoch": 0.10369659145463274, + "grad_norm": 0.92010241449416, + "learning_rate": 2.053742802303263e-06, + "loss": 0.7839, + "step": 108 + }, + { + "epoch": 0.10465674507921267, + "grad_norm": 0.959546321803882, + "learning_rate": 2.072936660268714e-06, + "loss": 0.8704, + "step": 109 + }, + { + "epoch": 0.10561689870379261, + "grad_norm": 1.0121164230279138, + "learning_rate": 2.0921305182341654e-06, + "loss": 0.7748, + "step": 110 + }, + { + "epoch": 0.10657705232837254, + "grad_norm": 0.8983272418888552, + "learning_rate": 2.1113243761996164e-06, + "loss": 0.8037, + "step": 111 + }, + { + "epoch": 0.10753720595295248, + "grad_norm": 1.2574271355202917, + "learning_rate": 2.1305182341650673e-06, + "loss": 0.9004, + "step": 112 + }, + { + "epoch": 0.1084973595775324, + "grad_norm": 1.093188744338786, + "learning_rate": 2.149712092130518e-06, + "loss": 0.7215, + "step": 113 + }, + { + "epoch": 0.10945751320211233, + "grad_norm": 1.0677246649240915, + "learning_rate": 2.1689059500959696e-06, + "loss": 0.7169, + "step": 114 + }, + { + "epoch": 0.11041766682669227, + "grad_norm": 1.2362607538290646, + "learning_rate": 2.1880998080614205e-06, + "loss": 0.8536, + "step": 115 + }, + { + "epoch": 0.1113778204512722, + "grad_norm": 1.7753777515012845, + "learning_rate": 2.2072936660268714e-06, + "loss": 0.9374, + "step": 116 + }, + { + "epoch": 0.11233797407585214, + "grad_norm": 0.8545104864205748, + "learning_rate": 2.2264875239923228e-06, + "loss": 0.7226, + "step": 117 + }, + { + "epoch": 0.11329812770043207, + "grad_norm": 0.8855990202815591, + "learning_rate": 2.2456813819577737e-06, + "loss": 0.8334, + "step": 118 + }, + { + "epoch": 0.114258281325012, + "grad_norm": 1.1549401776774397, + "learning_rate": 2.264875239923225e-06, + "loss": 0.8969, + "step": 119 + }, + { + "epoch": 0.11521843494959194, + "grad_norm": 0.9811642711182177, + "learning_rate": 2.284069097888676e-06, + "loss": 0.9129, + "step": 120 + }, + { + "epoch": 0.11617858857417186, + "grad_norm": 1.533573497571935, + "learning_rate": 2.303262955854127e-06, + "loss": 0.9064, + "step": 121 + }, + { + "epoch": 0.1171387421987518, + "grad_norm": 0.9402456230334737, + "learning_rate": 2.322456813819578e-06, + "loss": 0.7372, + "step": 122 + }, + { + "epoch": 0.11809889582333173, + "grad_norm": 1.1086864304645334, + "learning_rate": 2.341650671785029e-06, + "loss": 0.875, + "step": 123 + }, + { + "epoch": 0.11905904944791167, + "grad_norm": 1.2002234675956465, + "learning_rate": 2.36084452975048e-06, + "loss": 0.7128, + "step": 124 + }, + { + "epoch": 0.1200192030724916, + "grad_norm": 0.8158104333736143, + "learning_rate": 2.380038387715931e-06, + "loss": 0.6694, + "step": 125 + }, + { + "epoch": 0.12097935669707154, + "grad_norm": 1.0877193848670241, + "learning_rate": 2.3992322456813823e-06, + "loss": 0.7247, + "step": 126 + }, + { + "epoch": 0.12193951032165147, + "grad_norm": 1.719738648966655, + "learning_rate": 2.4184261036468333e-06, + "loss": 0.7878, + "step": 127 + }, + { + "epoch": 0.12289966394623139, + "grad_norm": 1.0309225906761206, + "learning_rate": 2.437619961612284e-06, + "loss": 0.9741, + "step": 128 + }, + { + "epoch": 0.12385981757081133, + "grad_norm": 1.2089797111899467, + "learning_rate": 2.456813819577735e-06, + "loss": 0.9194, + "step": 129 + }, + { + "epoch": 0.12481997119539126, + "grad_norm": 0.9567385863393127, + "learning_rate": 2.4760076775431864e-06, + "loss": 0.7369, + "step": 130 + }, + { + "epoch": 0.1257801248199712, + "grad_norm": 1.142430713832724, + "learning_rate": 2.4952015355086374e-06, + "loss": 0.8294, + "step": 131 + }, + { + "epoch": 0.12674027844455113, + "grad_norm": 1.3090923807049384, + "learning_rate": 2.5143953934740883e-06, + "loss": 0.8456, + "step": 132 + }, + { + "epoch": 0.12770043206913106, + "grad_norm": 1.2503457366705912, + "learning_rate": 2.5335892514395392e-06, + "loss": 0.8744, + "step": 133 + }, + { + "epoch": 0.128660585693711, + "grad_norm": 1.0711467403091606, + "learning_rate": 2.5527831094049906e-06, + "loss": 0.8338, + "step": 134 + }, + { + "epoch": 0.12962073931829093, + "grad_norm": 1.6084156917236838, + "learning_rate": 2.5719769673704415e-06, + "loss": 0.8592, + "step": 135 + }, + { + "epoch": 0.13058089294287087, + "grad_norm": 1.3308971115078625, + "learning_rate": 2.5911708253358924e-06, + "loss": 0.7858, + "step": 136 + }, + { + "epoch": 0.1315410465674508, + "grad_norm": 1.0828775067399747, + "learning_rate": 2.6103646833013433e-06, + "loss": 0.9026, + "step": 137 + }, + { + "epoch": 0.13250120019203074, + "grad_norm": 1.2085159913760772, + "learning_rate": 2.629558541266795e-06, + "loss": 0.8446, + "step": 138 + }, + { + "epoch": 0.13346135381661065, + "grad_norm": 1.3699534639223174, + "learning_rate": 2.648752399232246e-06, + "loss": 0.8861, + "step": 139 + }, + { + "epoch": 0.13442150744119058, + "grad_norm": 1.1738872763207, + "learning_rate": 2.6679462571976974e-06, + "loss": 0.7103, + "step": 140 + }, + { + "epoch": 0.13538166106577051, + "grad_norm": 1.0481635182090046, + "learning_rate": 2.6871401151631483e-06, + "loss": 0.9215, + "step": 141 + }, + { + "epoch": 0.13634181469035045, + "grad_norm": 0.8045116032807893, + "learning_rate": 2.7063339731285992e-06, + "loss": 0.7805, + "step": 142 + }, + { + "epoch": 0.13730196831493038, + "grad_norm": 0.9818798002309104, + "learning_rate": 2.72552783109405e-06, + "loss": 0.7844, + "step": 143 + }, + { + "epoch": 0.13826212193951032, + "grad_norm": 1.1680069006683926, + "learning_rate": 2.7447216890595015e-06, + "loss": 0.7744, + "step": 144 + }, + { + "epoch": 0.13922227556409025, + "grad_norm": 1.2164677300673201, + "learning_rate": 2.7639155470249524e-06, + "loss": 0.825, + "step": 145 + }, + { + "epoch": 0.1401824291886702, + "grad_norm": 1.2465951929177923, + "learning_rate": 2.7831094049904033e-06, + "loss": 0.6901, + "step": 146 + }, + { + "epoch": 0.14114258281325012, + "grad_norm": 1.100866715811699, + "learning_rate": 2.8023032629558543e-06, + "loss": 0.8722, + "step": 147 + }, + { + "epoch": 0.14210273643783006, + "grad_norm": 1.6066181507941228, + "learning_rate": 2.8214971209213056e-06, + "loss": 0.8334, + "step": 148 + }, + { + "epoch": 0.14306289006241, + "grad_norm": 1.3751884963973906, + "learning_rate": 2.8406909788867565e-06, + "loss": 0.7968, + "step": 149 + }, + { + "epoch": 0.14402304368698993, + "grad_norm": 1.1715432702710724, + "learning_rate": 2.8598848368522074e-06, + "loss": 0.8436, + "step": 150 + }, + { + "epoch": 0.14498319731156986, + "grad_norm": 1.0127103261881165, + "learning_rate": 2.8790786948176584e-06, + "loss": 0.7558, + "step": 151 + }, + { + "epoch": 0.1459433509361498, + "grad_norm": 0.9868103566572486, + "learning_rate": 2.8982725527831097e-06, + "loss": 0.7393, + "step": 152 + }, + { + "epoch": 0.1469035045607297, + "grad_norm": 1.2398536101083277, + "learning_rate": 2.9174664107485606e-06, + "loss": 0.873, + "step": 153 + }, + { + "epoch": 0.14786365818530964, + "grad_norm": 1.1999787944229654, + "learning_rate": 2.9366602687140116e-06, + "loss": 0.9016, + "step": 154 + }, + { + "epoch": 0.14882381180988957, + "grad_norm": 1.1324449796486242, + "learning_rate": 2.9558541266794625e-06, + "loss": 0.7667, + "step": 155 + }, + { + "epoch": 0.1497839654344695, + "grad_norm": 1.3120704929861124, + "learning_rate": 2.975047984644914e-06, + "loss": 0.8744, + "step": 156 + }, + { + "epoch": 0.15074411905904944, + "grad_norm": 0.9062163071328126, + "learning_rate": 2.9942418426103648e-06, + "loss": 0.7798, + "step": 157 + }, + { + "epoch": 0.15170427268362938, + "grad_norm": 1.4076582801284412, + "learning_rate": 3.0134357005758157e-06, + "loss": 0.8565, + "step": 158 + }, + { + "epoch": 0.1526644263082093, + "grad_norm": 1.4673082584603003, + "learning_rate": 3.032629558541267e-06, + "loss": 0.8813, + "step": 159 + }, + { + "epoch": 0.15362457993278925, + "grad_norm": 1.374671366145908, + "learning_rate": 3.051823416506718e-06, + "loss": 0.8445, + "step": 160 + }, + { + "epoch": 0.15458473355736918, + "grad_norm": 1.2868007929715295, + "learning_rate": 3.071017274472169e-06, + "loss": 0.8934, + "step": 161 + }, + { + "epoch": 0.15554488718194912, + "grad_norm": 1.4464722202223055, + "learning_rate": 3.09021113243762e-06, + "loss": 0.7582, + "step": 162 + }, + { + "epoch": 0.15650504080652905, + "grad_norm": 1.105469935141118, + "learning_rate": 3.1094049904030716e-06, + "loss": 0.7808, + "step": 163 + }, + { + "epoch": 0.15746519443110898, + "grad_norm": 1.0911900597803557, + "learning_rate": 3.1285988483685225e-06, + "loss": 0.796, + "step": 164 + }, + { + "epoch": 0.15842534805568892, + "grad_norm": 1.3309231366723955, + "learning_rate": 3.1477927063339734e-06, + "loss": 0.8544, + "step": 165 + }, + { + "epoch": 0.15938550168026885, + "grad_norm": 1.2548221881193011, + "learning_rate": 3.1669865642994248e-06, + "loss": 0.8361, + "step": 166 + }, + { + "epoch": 0.1603456553048488, + "grad_norm": 1.7659278560067024, + "learning_rate": 3.1861804222648757e-06, + "loss": 0.8906, + "step": 167 + }, + { + "epoch": 0.1613058089294287, + "grad_norm": 1.1655867571269192, + "learning_rate": 3.2053742802303266e-06, + "loss": 0.894, + "step": 168 + }, + { + "epoch": 0.16226596255400863, + "grad_norm": 1.2433543508842448, + "learning_rate": 3.224568138195778e-06, + "loss": 0.8004, + "step": 169 + }, + { + "epoch": 0.16322611617858857, + "grad_norm": 0.9872268083143309, + "learning_rate": 3.243761996161229e-06, + "loss": 0.822, + "step": 170 + }, + { + "epoch": 0.1641862698031685, + "grad_norm": 1.3753973529861458, + "learning_rate": 3.26295585412668e-06, + "loss": 0.9343, + "step": 171 + }, + { + "epoch": 0.16514642342774843, + "grad_norm": 1.0586321759599753, + "learning_rate": 3.2821497120921307e-06, + "loss": 0.844, + "step": 172 + }, + { + "epoch": 0.16610657705232837, + "grad_norm": 0.9715401565614465, + "learning_rate": 3.301343570057582e-06, + "loss": 0.7735, + "step": 173 + }, + { + "epoch": 0.1670667306769083, + "grad_norm": 1.2748668842297577, + "learning_rate": 3.320537428023033e-06, + "loss": 0.7567, + "step": 174 + }, + { + "epoch": 0.16802688430148824, + "grad_norm": 0.9412191459151686, + "learning_rate": 3.339731285988484e-06, + "loss": 0.6285, + "step": 175 + }, + { + "epoch": 0.16898703792606817, + "grad_norm": 1.1052323344585788, + "learning_rate": 3.358925143953935e-06, + "loss": 0.9069, + "step": 176 + }, + { + "epoch": 0.1699471915506481, + "grad_norm": 1.08861582464332, + "learning_rate": 3.378119001919386e-06, + "loss": 0.7559, + "step": 177 + }, + { + "epoch": 0.17090734517522804, + "grad_norm": 0.9319805506601659, + "learning_rate": 3.397312859884837e-06, + "loss": 0.7105, + "step": 178 + }, + { + "epoch": 0.17186749879980798, + "grad_norm": 0.9771730792010799, + "learning_rate": 3.416506717850288e-06, + "loss": 0.8552, + "step": 179 + }, + { + "epoch": 0.1728276524243879, + "grad_norm": 0.8470050375728845, + "learning_rate": 3.435700575815739e-06, + "loss": 0.7615, + "step": 180 + }, + { + "epoch": 0.17378780604896785, + "grad_norm": 1.0959154608915553, + "learning_rate": 3.4548944337811903e-06, + "loss": 0.7262, + "step": 181 + }, + { + "epoch": 0.17474795967354778, + "grad_norm": 1.4625375688138138, + "learning_rate": 3.4740882917466412e-06, + "loss": 0.8587, + "step": 182 + }, + { + "epoch": 0.1757081132981277, + "grad_norm": 1.0000628986768312, + "learning_rate": 3.493282149712092e-06, + "loss": 0.7761, + "step": 183 + }, + { + "epoch": 0.17666826692270762, + "grad_norm": 1.0005373628571768, + "learning_rate": 3.512476007677543e-06, + "loss": 0.6796, + "step": 184 + }, + { + "epoch": 0.17762842054728756, + "grad_norm": 0.9968604487993922, + "learning_rate": 3.5316698656429944e-06, + "loss": 0.8201, + "step": 185 + }, + { + "epoch": 0.1785885741718675, + "grad_norm": 1.4993901938369656, + "learning_rate": 3.5508637236084453e-06, + "loss": 0.9407, + "step": 186 + }, + { + "epoch": 0.17954872779644743, + "grad_norm": 0.8704017460553644, + "learning_rate": 3.5700575815738963e-06, + "loss": 0.6974, + "step": 187 + }, + { + "epoch": 0.18050888142102736, + "grad_norm": 0.9761668674958587, + "learning_rate": 3.589251439539348e-06, + "loss": 0.8016, + "step": 188 + }, + { + "epoch": 0.1814690350456073, + "grad_norm": 1.0515756767474949, + "learning_rate": 3.608445297504799e-06, + "loss": 0.7232, + "step": 189 + }, + { + "epoch": 0.18242918867018723, + "grad_norm": 1.555490418358551, + "learning_rate": 3.62763915547025e-06, + "loss": 0.6891, + "step": 190 + }, + { + "epoch": 0.18338934229476717, + "grad_norm": 0.9343252294840784, + "learning_rate": 3.6468330134357012e-06, + "loss": 0.7632, + "step": 191 + }, + { + "epoch": 0.1843494959193471, + "grad_norm": 1.0717238045522897, + "learning_rate": 3.666026871401152e-06, + "loss": 0.7967, + "step": 192 + }, + { + "epoch": 0.18530964954392704, + "grad_norm": 0.9023425167215238, + "learning_rate": 3.685220729366603e-06, + "loss": 0.6851, + "step": 193 + }, + { + "epoch": 0.18626980316850697, + "grad_norm": 1.0354539206924438, + "learning_rate": 3.704414587332054e-06, + "loss": 0.6541, + "step": 194 + }, + { + "epoch": 0.1872299567930869, + "grad_norm": 0.9154002359123815, + "learning_rate": 3.7236084452975053e-06, + "loss": 0.6118, + "step": 195 + }, + { + "epoch": 0.18819011041766684, + "grad_norm": 1.4834012311015687, + "learning_rate": 3.7428023032629563e-06, + "loss": 0.7373, + "step": 196 + }, + { + "epoch": 0.18915026404224675, + "grad_norm": 0.9395875480988187, + "learning_rate": 3.761996161228407e-06, + "loss": 0.7418, + "step": 197 + }, + { + "epoch": 0.19011041766682668, + "grad_norm": 0.8228848151380725, + "learning_rate": 3.781190019193858e-06, + "loss": 0.639, + "step": 198 + }, + { + "epoch": 0.19107057129140662, + "grad_norm": 1.0222902437157457, + "learning_rate": 3.8003838771593095e-06, + "loss": 0.7843, + "step": 199 + }, + { + "epoch": 0.19203072491598655, + "grad_norm": 1.1971998828421855, + "learning_rate": 3.81957773512476e-06, + "loss": 0.8042, + "step": 200 + }, + { + "epoch": 0.19299087854056649, + "grad_norm": 1.3436354605563, + "learning_rate": 3.838771593090212e-06, + "loss": 0.6823, + "step": 201 + }, + { + "epoch": 0.19395103216514642, + "grad_norm": 1.0432691374129013, + "learning_rate": 3.857965451055662e-06, + "loss": 0.8763, + "step": 202 + }, + { + "epoch": 0.19491118578972635, + "grad_norm": 1.4598651306676347, + "learning_rate": 3.8771593090211136e-06, + "loss": 0.7037, + "step": 203 + }, + { + "epoch": 0.1958713394143063, + "grad_norm": 1.2815645347506386, + "learning_rate": 3.896353166986565e-06, + "loss": 0.7283, + "step": 204 + }, + { + "epoch": 0.19683149303888622, + "grad_norm": 1.0833856307385523, + "learning_rate": 3.915547024952015e-06, + "loss": 0.8133, + "step": 205 + }, + { + "epoch": 0.19779164666346616, + "grad_norm": 1.3786574048365472, + "learning_rate": 3.934740882917467e-06, + "loss": 0.9112, + "step": 206 + }, + { + "epoch": 0.1987518002880461, + "grad_norm": 1.2875346340348794, + "learning_rate": 3.953934740882917e-06, + "loss": 0.6257, + "step": 207 + }, + { + "epoch": 0.19971195391262603, + "grad_norm": 1.0242596133728246, + "learning_rate": 3.973128598848369e-06, + "loss": 0.8232, + "step": 208 + }, + { + "epoch": 0.20067210753720596, + "grad_norm": 0.9870530206057417, + "learning_rate": 3.99232245681382e-06, + "loss": 0.8147, + "step": 209 + }, + { + "epoch": 0.2016322611617859, + "grad_norm": 1.4004219647275036, + "learning_rate": 4.0115163147792705e-06, + "loss": 0.7407, + "step": 210 + }, + { + "epoch": 0.20259241478636583, + "grad_norm": 1.9647525624689024, + "learning_rate": 4.030710172744722e-06, + "loss": 0.9065, + "step": 211 + }, + { + "epoch": 0.20355256841094574, + "grad_norm": 1.1758284793178084, + "learning_rate": 4.049904030710173e-06, + "loss": 0.7354, + "step": 212 + }, + { + "epoch": 0.20451272203552567, + "grad_norm": 0.9399628887884863, + "learning_rate": 4.0690978886756245e-06, + "loss": 0.7094, + "step": 213 + }, + { + "epoch": 0.2054728756601056, + "grad_norm": 1.229532902404846, + "learning_rate": 4.088291746641076e-06, + "loss": 0.7294, + "step": 214 + }, + { + "epoch": 0.20643302928468554, + "grad_norm": 1.285481349569688, + "learning_rate": 4.107485604606526e-06, + "loss": 0.7874, + "step": 215 + }, + { + "epoch": 0.20739318290926548, + "grad_norm": 1.7048926937913806, + "learning_rate": 4.126679462571978e-06, + "loss": 0.9107, + "step": 216 + }, + { + "epoch": 0.2083533365338454, + "grad_norm": 1.2789559622430384, + "learning_rate": 4.145873320537428e-06, + "loss": 0.7613, + "step": 217 + }, + { + "epoch": 0.20931349015842535, + "grad_norm": 1.099423101490443, + "learning_rate": 4.1650671785028795e-06, + "loss": 0.8418, + "step": 218 + }, + { + "epoch": 0.21027364378300528, + "grad_norm": 1.0902575563764023, + "learning_rate": 4.184261036468331e-06, + "loss": 0.6956, + "step": 219 + }, + { + "epoch": 0.21123379740758522, + "grad_norm": 1.272728305370189, + "learning_rate": 4.203454894433781e-06, + "loss": 0.9046, + "step": 220 + }, + { + "epoch": 0.21219395103216515, + "grad_norm": 0.9179036321827532, + "learning_rate": 4.222648752399233e-06, + "loss": 0.8334, + "step": 221 + }, + { + "epoch": 0.21315410465674509, + "grad_norm": 1.1608689862055543, + "learning_rate": 4.241842610364684e-06, + "loss": 0.7988, + "step": 222 + }, + { + "epoch": 0.21411425828132502, + "grad_norm": 0.9424134482088689, + "learning_rate": 4.2610364683301346e-06, + "loss": 0.7275, + "step": 223 + }, + { + "epoch": 0.21507441190590496, + "grad_norm": 1.0056322925801187, + "learning_rate": 4.280230326295586e-06, + "loss": 0.8358, + "step": 224 + }, + { + "epoch": 0.2160345655304849, + "grad_norm": 1.358420329516284, + "learning_rate": 4.299424184261036e-06, + "loss": 0.789, + "step": 225 + }, + { + "epoch": 0.2169947191550648, + "grad_norm": 1.0604793977460594, + "learning_rate": 4.318618042226488e-06, + "loss": 0.7716, + "step": 226 + }, + { + "epoch": 0.21795487277964473, + "grad_norm": 1.3579766865243843, + "learning_rate": 4.337811900191939e-06, + "loss": 0.7773, + "step": 227 + }, + { + "epoch": 0.21891502640422467, + "grad_norm": 1.1969285333538193, + "learning_rate": 4.35700575815739e-06, + "loss": 0.7988, + "step": 228 + }, + { + "epoch": 0.2198751800288046, + "grad_norm": 1.0977349545103885, + "learning_rate": 4.376199616122841e-06, + "loss": 0.8328, + "step": 229 + }, + { + "epoch": 0.22083533365338454, + "grad_norm": 1.1029551255090795, + "learning_rate": 4.395393474088292e-06, + "loss": 0.686, + "step": 230 + }, + { + "epoch": 0.22179548727796447, + "grad_norm": 1.4552381896353486, + "learning_rate": 4.414587332053743e-06, + "loss": 0.8195, + "step": 231 + }, + { + "epoch": 0.2227556409025444, + "grad_norm": 0.894667659418859, + "learning_rate": 4.433781190019194e-06, + "loss": 0.7275, + "step": 232 + }, + { + "epoch": 0.22371579452712434, + "grad_norm": 0.8202068548985902, + "learning_rate": 4.4529750479846455e-06, + "loss": 0.5861, + "step": 233 + }, + { + "epoch": 0.22467594815170427, + "grad_norm": 1.029707146111314, + "learning_rate": 4.472168905950096e-06, + "loss": 0.6985, + "step": 234 + }, + { + "epoch": 0.2256361017762842, + "grad_norm": 1.0990681024322204, + "learning_rate": 4.491362763915547e-06, + "loss": 0.68, + "step": 235 + }, + { + "epoch": 0.22659625540086414, + "grad_norm": 1.4012799413122012, + "learning_rate": 4.510556621880998e-06, + "loss": 0.8869, + "step": 236 + }, + { + "epoch": 0.22755640902544408, + "grad_norm": 1.1973821399483648, + "learning_rate": 4.52975047984645e-06, + "loss": 0.7971, + "step": 237 + }, + { + "epoch": 0.228516562650024, + "grad_norm": 1.4686234982760031, + "learning_rate": 4.5489443378119005e-06, + "loss": 0.7288, + "step": 238 + }, + { + "epoch": 0.22947671627460395, + "grad_norm": 1.032604350539258, + "learning_rate": 4.568138195777352e-06, + "loss": 0.7507, + "step": 239 + }, + { + "epoch": 0.23043686989918388, + "grad_norm": 1.0822190269808825, + "learning_rate": 4.587332053742803e-06, + "loss": 0.8673, + "step": 240 + }, + { + "epoch": 0.2313970235237638, + "grad_norm": 1.0856558538512318, + "learning_rate": 4.606525911708254e-06, + "loss": 0.7822, + "step": 241 + }, + { + "epoch": 0.23235717714834372, + "grad_norm": 1.253257076933008, + "learning_rate": 4.625719769673705e-06, + "loss": 0.8852, + "step": 242 + }, + { + "epoch": 0.23331733077292366, + "grad_norm": 1.0533352620131817, + "learning_rate": 4.644913627639156e-06, + "loss": 0.8494, + "step": 243 + }, + { + "epoch": 0.2342774843975036, + "grad_norm": 1.2104402492231694, + "learning_rate": 4.664107485604607e-06, + "loss": 0.8848, + "step": 244 + }, + { + "epoch": 0.23523763802208353, + "grad_norm": 1.0309254759547493, + "learning_rate": 4.683301343570058e-06, + "loss": 0.7899, + "step": 245 + }, + { + "epoch": 0.23619779164666346, + "grad_norm": 0.9228274298271464, + "learning_rate": 4.702495201535509e-06, + "loss": 0.7684, + "step": 246 + }, + { + "epoch": 0.2371579452712434, + "grad_norm": 1.082494978756675, + "learning_rate": 4.72168905950096e-06, + "loss": 0.8015, + "step": 247 + }, + { + "epoch": 0.23811809889582333, + "grad_norm": 1.1404265500842226, + "learning_rate": 4.7408829174664115e-06, + "loss": 0.8995, + "step": 248 + }, + { + "epoch": 0.23907825252040327, + "grad_norm": 1.1383637841992147, + "learning_rate": 4.760076775431862e-06, + "loss": 0.7914, + "step": 249 + }, + { + "epoch": 0.2400384061449832, + "grad_norm": 1.2275079047335584, + "learning_rate": 4.779270633397313e-06, + "loss": 0.8338, + "step": 250 + }, + { + "epoch": 0.24099855976956314, + "grad_norm": 1.1011103569744343, + "learning_rate": 4.798464491362765e-06, + "loss": 0.7431, + "step": 251 + }, + { + "epoch": 0.24195871339414307, + "grad_norm": 1.0587489444236706, + "learning_rate": 4.817658349328215e-06, + "loss": 0.7682, + "step": 252 + }, + { + "epoch": 0.242918867018723, + "grad_norm": 1.26897029416747, + "learning_rate": 4.8368522072936665e-06, + "loss": 0.7562, + "step": 253 + }, + { + "epoch": 0.24387902064330294, + "grad_norm": 0.8705948431481193, + "learning_rate": 4.856046065259117e-06, + "loss": 0.7452, + "step": 254 + }, + { + "epoch": 0.24483917426788285, + "grad_norm": 1.092814530275756, + "learning_rate": 4.875239923224568e-06, + "loss": 0.7531, + "step": 255 + }, + { + "epoch": 0.24579932789246278, + "grad_norm": 1.297369303442484, + "learning_rate": 4.89443378119002e-06, + "loss": 0.7553, + "step": 256 + }, + { + "epoch": 0.24675948151704272, + "grad_norm": 1.0620363460937197, + "learning_rate": 4.91362763915547e-06, + "loss": 0.778, + "step": 257 + }, + { + "epoch": 0.24771963514162265, + "grad_norm": 1.007077263281022, + "learning_rate": 4.9328214971209215e-06, + "loss": 0.8044, + "step": 258 + }, + { + "epoch": 0.2486797887662026, + "grad_norm": 1.2092919265067725, + "learning_rate": 4.952015355086373e-06, + "loss": 0.7579, + "step": 259 + }, + { + "epoch": 0.24963994239078252, + "grad_norm": 1.0123125970450146, + "learning_rate": 4.971209213051823e-06, + "loss": 0.7101, + "step": 260 + }, + { + "epoch": 0.2506000960153625, + "grad_norm": 0.8793511419440708, + "learning_rate": 4.990403071017275e-06, + "loss": 0.6262, + "step": 261 + }, + { + "epoch": 0.2515602496399424, + "grad_norm": 1.078809487237499, + "learning_rate": 5.009596928982726e-06, + "loss": 0.6783, + "step": 262 + }, + { + "epoch": 0.2525204032645223, + "grad_norm": 0.941420750891124, + "learning_rate": 5.028790786948177e-06, + "loss": 0.6403, + "step": 263 + }, + { + "epoch": 0.25348055688910226, + "grad_norm": 1.3575222071854605, + "learning_rate": 5.047984644913628e-06, + "loss": 0.6863, + "step": 264 + }, + { + "epoch": 0.25444071051368217, + "grad_norm": 0.9902478107580348, + "learning_rate": 5.0671785028790784e-06, + "loss": 0.6048, + "step": 265 + }, + { + "epoch": 0.25540086413826213, + "grad_norm": 0.9383297421101107, + "learning_rate": 5.086372360844531e-06, + "loss": 0.7624, + "step": 266 + }, + { + "epoch": 0.25636101776284204, + "grad_norm": 0.9745420306969944, + "learning_rate": 5.105566218809981e-06, + "loss": 0.726, + "step": 267 + }, + { + "epoch": 0.257321171387422, + "grad_norm": 0.8823186085065814, + "learning_rate": 5.1247600767754325e-06, + "loss": 0.6667, + "step": 268 + }, + { + "epoch": 0.2582813250120019, + "grad_norm": 1.3077109124843154, + "learning_rate": 5.143953934740883e-06, + "loss": 0.7825, + "step": 269 + }, + { + "epoch": 0.25924147863658187, + "grad_norm": 0.8701718980494662, + "learning_rate": 5.163147792706334e-06, + "loss": 0.7284, + "step": 270 + }, + { + "epoch": 0.2602016322611618, + "grad_norm": 1.0947520139625904, + "learning_rate": 5.182341650671785e-06, + "loss": 0.7205, + "step": 271 + }, + { + "epoch": 0.26116178588574174, + "grad_norm": 0.9035421633997119, + "learning_rate": 5.201535508637236e-06, + "loss": 0.707, + "step": 272 + }, + { + "epoch": 0.26212193951032164, + "grad_norm": 1.2393989616692116, + "learning_rate": 5.220729366602687e-06, + "loss": 0.8318, + "step": 273 + }, + { + "epoch": 0.2630820931349016, + "grad_norm": 1.0820647264342638, + "learning_rate": 5.239923224568139e-06, + "loss": 0.7086, + "step": 274 + }, + { + "epoch": 0.2640422467594815, + "grad_norm": 0.9102334415716998, + "learning_rate": 5.25911708253359e-06, + "loss": 0.686, + "step": 275 + }, + { + "epoch": 0.2650024003840615, + "grad_norm": 0.9891112292775399, + "learning_rate": 5.278310940499041e-06, + "loss": 0.6594, + "step": 276 + }, + { + "epoch": 0.2659625540086414, + "grad_norm": 1.6096713844116006, + "learning_rate": 5.297504798464492e-06, + "loss": 0.7862, + "step": 277 + }, + { + "epoch": 0.2669227076332213, + "grad_norm": 1.1225227418118746, + "learning_rate": 5.3166986564299425e-06, + "loss": 0.7328, + "step": 278 + }, + { + "epoch": 0.26788286125780125, + "grad_norm": 1.0231886402817059, + "learning_rate": 5.335892514395395e-06, + "loss": 0.6849, + "step": 279 + }, + { + "epoch": 0.26884301488238116, + "grad_norm": 1.2155240110282657, + "learning_rate": 5.355086372360845e-06, + "loss": 0.7907, + "step": 280 + }, + { + "epoch": 0.2698031685069611, + "grad_norm": 1.292857617189349, + "learning_rate": 5.374280230326297e-06, + "loss": 0.726, + "step": 281 + }, + { + "epoch": 0.27076332213154103, + "grad_norm": 1.272186258694057, + "learning_rate": 5.393474088291747e-06, + "loss": 0.6674, + "step": 282 + }, + { + "epoch": 0.271723475756121, + "grad_norm": 1.3078425645435416, + "learning_rate": 5.4126679462571984e-06, + "loss": 0.7705, + "step": 283 + }, + { + "epoch": 0.2726836293807009, + "grad_norm": 1.0078428876763987, + "learning_rate": 5.431861804222649e-06, + "loss": 0.7072, + "step": 284 + }, + { + "epoch": 0.27364378300528086, + "grad_norm": 1.2618657788714938, + "learning_rate": 5.4510556621881e-06, + "loss": 0.8385, + "step": 285 + }, + { + "epoch": 0.27460393662986077, + "grad_norm": 1.1965521394945369, + "learning_rate": 5.470249520153551e-06, + "loss": 0.7379, + "step": 286 + }, + { + "epoch": 0.27556409025444073, + "grad_norm": 1.0040579235952913, + "learning_rate": 5.489443378119003e-06, + "loss": 0.6312, + "step": 287 + }, + { + "epoch": 0.27652424387902064, + "grad_norm": 1.1226365966223, + "learning_rate": 5.5086372360844535e-06, + "loss": 0.7333, + "step": 288 + }, + { + "epoch": 0.2774843975036006, + "grad_norm": 1.0426235237494172, + "learning_rate": 5.527831094049905e-06, + "loss": 0.8043, + "step": 289 + }, + { + "epoch": 0.2784445511281805, + "grad_norm": 1.1282354419614937, + "learning_rate": 5.547024952015355e-06, + "loss": 0.6533, + "step": 290 + }, + { + "epoch": 0.27940470475276047, + "grad_norm": 1.0876477193458896, + "learning_rate": 5.566218809980807e-06, + "loss": 0.8533, + "step": 291 + }, + { + "epoch": 0.2803648583773404, + "grad_norm": 1.2267395738659088, + "learning_rate": 5.585412667946257e-06, + "loss": 0.7215, + "step": 292 + }, + { + "epoch": 0.2813250120019203, + "grad_norm": 1.3886867771152398, + "learning_rate": 5.6046065259117085e-06, + "loss": 0.7892, + "step": 293 + }, + { + "epoch": 0.28228516562650025, + "grad_norm": 1.2328491859061401, + "learning_rate": 5.623800383877159e-06, + "loss": 0.8432, + "step": 294 + }, + { + "epoch": 0.28324531925108015, + "grad_norm": 1.518901224201875, + "learning_rate": 5.642994241842611e-06, + "loss": 0.8301, + "step": 295 + }, + { + "epoch": 0.2842054728756601, + "grad_norm": 1.456338581338527, + "learning_rate": 5.662188099808062e-06, + "loss": 0.8818, + "step": 296 + }, + { + "epoch": 0.28516562650024, + "grad_norm": 1.3027020402970058, + "learning_rate": 5.681381957773513e-06, + "loss": 0.8254, + "step": 297 + }, + { + "epoch": 0.28612578012482, + "grad_norm": 0.9080934518090018, + "learning_rate": 5.7005758157389635e-06, + "loss": 0.6895, + "step": 298 + }, + { + "epoch": 0.2870859337493999, + "grad_norm": 1.2130812480748785, + "learning_rate": 5.719769673704415e-06, + "loss": 0.7949, + "step": 299 + }, + { + "epoch": 0.28804608737397985, + "grad_norm": 1.1160052811250392, + "learning_rate": 5.738963531669866e-06, + "loss": 0.8164, + "step": 300 + }, + { + "epoch": 0.28900624099855976, + "grad_norm": 1.1210266919814806, + "learning_rate": 5.758157389635317e-06, + "loss": 0.649, + "step": 301 + }, + { + "epoch": 0.2899663946231397, + "grad_norm": 1.0879864278375668, + "learning_rate": 5.777351247600769e-06, + "loss": 0.8905, + "step": 302 + }, + { + "epoch": 0.29092654824771963, + "grad_norm": 1.0213016500599061, + "learning_rate": 5.7965451055662194e-06, + "loss": 0.7535, + "step": 303 + }, + { + "epoch": 0.2918867018722996, + "grad_norm": 0.865511806914633, + "learning_rate": 5.815738963531671e-06, + "loss": 0.7009, + "step": 304 + }, + { + "epoch": 0.2928468554968795, + "grad_norm": 1.0762956326446182, + "learning_rate": 5.834932821497121e-06, + "loss": 0.8394, + "step": 305 + }, + { + "epoch": 0.2938070091214594, + "grad_norm": 1.1011468772663233, + "learning_rate": 5.854126679462573e-06, + "loss": 0.7994, + "step": 306 + }, + { + "epoch": 0.29476716274603937, + "grad_norm": 0.9358929440958041, + "learning_rate": 5.873320537428023e-06, + "loss": 0.6915, + "step": 307 + }, + { + "epoch": 0.2957273163706193, + "grad_norm": 1.1361983938783253, + "learning_rate": 5.892514395393475e-06, + "loss": 0.7465, + "step": 308 + }, + { + "epoch": 0.29668746999519924, + "grad_norm": 0.9706557310474103, + "learning_rate": 5.911708253358925e-06, + "loss": 0.7263, + "step": 309 + }, + { + "epoch": 0.29764762361977914, + "grad_norm": 0.9957406689560145, + "learning_rate": 5.930902111324377e-06, + "loss": 0.6731, + "step": 310 + }, + { + "epoch": 0.2986077772443591, + "grad_norm": 0.8297692761648142, + "learning_rate": 5.950095969289828e-06, + "loss": 0.6627, + "step": 311 + }, + { + "epoch": 0.299567930868939, + "grad_norm": 1.0823838214362094, + "learning_rate": 5.969289827255279e-06, + "loss": 0.7685, + "step": 312 + }, + { + "epoch": 0.300528084493519, + "grad_norm": 1.075826853937253, + "learning_rate": 5.9884836852207295e-06, + "loss": 0.7684, + "step": 313 + }, + { + "epoch": 0.3014882381180989, + "grad_norm": 1.3059389423523045, + "learning_rate": 6.007677543186181e-06, + "loss": 0.8944, + "step": 314 + }, + { + "epoch": 0.30244839174267885, + "grad_norm": 1.083104462062558, + "learning_rate": 6.026871401151631e-06, + "loss": 0.7052, + "step": 315 + }, + { + "epoch": 0.30340854536725875, + "grad_norm": 0.9209470027209099, + "learning_rate": 6.0460652591170836e-06, + "loss": 0.6925, + "step": 316 + }, + { + "epoch": 0.3043686989918387, + "grad_norm": 1.5506192977662627, + "learning_rate": 6.065259117082534e-06, + "loss": 0.8079, + "step": 317 + }, + { + "epoch": 0.3053288526164186, + "grad_norm": 1.1809708521141837, + "learning_rate": 6.084452975047985e-06, + "loss": 0.8492, + "step": 318 + }, + { + "epoch": 0.3062890062409986, + "grad_norm": 0.888021044899069, + "learning_rate": 6.103646833013436e-06, + "loss": 0.6798, + "step": 319 + }, + { + "epoch": 0.3072491598655785, + "grad_norm": 1.059234718818979, + "learning_rate": 6.122840690978887e-06, + "loss": 0.7214, + "step": 320 + }, + { + "epoch": 0.3082093134901584, + "grad_norm": 1.2972122688066008, + "learning_rate": 6.142034548944338e-06, + "loss": 0.8314, + "step": 321 + }, + { + "epoch": 0.30916946711473836, + "grad_norm": 1.026669162294941, + "learning_rate": 6.161228406909789e-06, + "loss": 0.6907, + "step": 322 + }, + { + "epoch": 0.31012962073931827, + "grad_norm": 1.0317826337977176, + "learning_rate": 6.18042226487524e-06, + "loss": 0.6827, + "step": 323 + }, + { + "epoch": 0.31108977436389823, + "grad_norm": 1.1500376036788722, + "learning_rate": 6.199616122840692e-06, + "loss": 0.8111, + "step": 324 + }, + { + "epoch": 0.31204992798847814, + "grad_norm": 0.9380007217093529, + "learning_rate": 6.218809980806143e-06, + "loss": 0.5933, + "step": 325 + }, + { + "epoch": 0.3130100816130581, + "grad_norm": 1.5376009293846107, + "learning_rate": 6.238003838771594e-06, + "loss": 0.8007, + "step": 326 + }, + { + "epoch": 0.313970235237638, + "grad_norm": 1.4599272669739034, + "learning_rate": 6.257197696737045e-06, + "loss": 0.7693, + "step": 327 + }, + { + "epoch": 0.31493038886221797, + "grad_norm": 1.5064515194175732, + "learning_rate": 6.2763915547024955e-06, + "loss": 0.9682, + "step": 328 + }, + { + "epoch": 0.3158905424867979, + "grad_norm": 0.9818886506201384, + "learning_rate": 6.295585412667947e-06, + "loss": 0.6798, + "step": 329 + }, + { + "epoch": 0.31685069611137784, + "grad_norm": 0.9712226276725909, + "learning_rate": 6.314779270633397e-06, + "loss": 0.743, + "step": 330 + }, + { + "epoch": 0.31781084973595775, + "grad_norm": 1.401672652996209, + "learning_rate": 6.3339731285988495e-06, + "loss": 0.7874, + "step": 331 + }, + { + "epoch": 0.3187710033605377, + "grad_norm": 1.0783950280605588, + "learning_rate": 6.3531669865643e-06, + "loss": 0.8842, + "step": 332 + }, + { + "epoch": 0.3197311569851176, + "grad_norm": 1.066363631638222, + "learning_rate": 6.372360844529751e-06, + "loss": 0.7388, + "step": 333 + }, + { + "epoch": 0.3206913106096976, + "grad_norm": 1.2209253548266827, + "learning_rate": 6.391554702495202e-06, + "loss": 0.7856, + "step": 334 + }, + { + "epoch": 0.3216514642342775, + "grad_norm": 1.2441561987844856, + "learning_rate": 6.410748560460653e-06, + "loss": 0.746, + "step": 335 + }, + { + "epoch": 0.3226116178588574, + "grad_norm": 1.2247304882054608, + "learning_rate": 6.429942418426104e-06, + "loss": 0.6776, + "step": 336 + }, + { + "epoch": 0.32357177148343735, + "grad_norm": 1.0856409119088077, + "learning_rate": 6.449136276391556e-06, + "loss": 0.721, + "step": 337 + }, + { + "epoch": 0.32453192510801726, + "grad_norm": 1.0796733699008674, + "learning_rate": 6.4683301343570056e-06, + "loss": 0.6634, + "step": 338 + }, + { + "epoch": 0.3254920787325972, + "grad_norm": 0.953864425810896, + "learning_rate": 6.487523992322458e-06, + "loss": 0.658, + "step": 339 + }, + { + "epoch": 0.32645223235717713, + "grad_norm": 1.071183113729068, + "learning_rate": 6.506717850287908e-06, + "loss": 0.8166, + "step": 340 + }, + { + "epoch": 0.3274123859817571, + "grad_norm": 0.877389486714033, + "learning_rate": 6.52591170825336e-06, + "loss": 0.5952, + "step": 341 + }, + { + "epoch": 0.328372539606337, + "grad_norm": 1.0020588864869746, + "learning_rate": 6.54510556621881e-06, + "loss": 0.6962, + "step": 342 + }, + { + "epoch": 0.32933269323091696, + "grad_norm": 1.3953195365139277, + "learning_rate": 6.5642994241842614e-06, + "loss": 0.9079, + "step": 343 + }, + { + "epoch": 0.33029284685549687, + "grad_norm": 1.1863169477349746, + "learning_rate": 6.583493282149712e-06, + "loss": 0.7766, + "step": 344 + }, + { + "epoch": 0.33125300048007683, + "grad_norm": 1.199300644800143, + "learning_rate": 6.602687140115164e-06, + "loss": 0.8119, + "step": 345 + }, + { + "epoch": 0.33221315410465674, + "grad_norm": 1.0315407252952944, + "learning_rate": 6.621880998080615e-06, + "loss": 0.6961, + "step": 346 + }, + { + "epoch": 0.3331733077292367, + "grad_norm": 0.8332104483111213, + "learning_rate": 6.641074856046066e-06, + "loss": 0.6525, + "step": 347 + }, + { + "epoch": 0.3341334613538166, + "grad_norm": 1.0502384293543738, + "learning_rate": 6.6602687140115165e-06, + "loss": 0.7915, + "step": 348 + }, + { + "epoch": 0.33509361497839657, + "grad_norm": 1.75292168897863, + "learning_rate": 6.679462571976968e-06, + "loss": 0.9489, + "step": 349 + }, + { + "epoch": 0.3360537686029765, + "grad_norm": 1.0394537445812293, + "learning_rate": 6.698656429942419e-06, + "loss": 0.7422, + "step": 350 + }, + { + "epoch": 0.3370139222275564, + "grad_norm": 1.3457781986708817, + "learning_rate": 6.71785028790787e-06, + "loss": 0.7588, + "step": 351 + }, + { + "epoch": 0.33797407585213635, + "grad_norm": 1.1385632722570405, + "learning_rate": 6.737044145873322e-06, + "loss": 0.7602, + "step": 352 + }, + { + "epoch": 0.33893422947671625, + "grad_norm": 1.6067517847274677, + "learning_rate": 6.756238003838772e-06, + "loss": 0.8005, + "step": 353 + }, + { + "epoch": 0.3398943831012962, + "grad_norm": 1.233975400494269, + "learning_rate": 6.775431861804224e-06, + "loss": 0.7682, + "step": 354 + }, + { + "epoch": 0.3408545367258761, + "grad_norm": 1.4526697439532525, + "learning_rate": 6.794625719769674e-06, + "loss": 0.75, + "step": 355 + }, + { + "epoch": 0.3418146903504561, + "grad_norm": 0.9323361856360292, + "learning_rate": 6.8138195777351256e-06, + "loss": 0.6922, + "step": 356 + }, + { + "epoch": 0.342774843975036, + "grad_norm": 1.0609558520672935, + "learning_rate": 6.833013435700576e-06, + "loss": 0.8037, + "step": 357 + }, + { + "epoch": 0.34373499759961595, + "grad_norm": 0.9283068391125257, + "learning_rate": 6.852207293666027e-06, + "loss": 0.8207, + "step": 358 + }, + { + "epoch": 0.34469515122419586, + "grad_norm": 1.2741384759296661, + "learning_rate": 6.871401151631478e-06, + "loss": 0.7734, + "step": 359 + }, + { + "epoch": 0.3456553048487758, + "grad_norm": 1.2226624613325408, + "learning_rate": 6.89059500959693e-06, + "loss": 0.7153, + "step": 360 + }, + { + "epoch": 0.34661545847335573, + "grad_norm": 0.9392211836640727, + "learning_rate": 6.909788867562381e-06, + "loss": 0.6772, + "step": 361 + }, + { + "epoch": 0.3475756120979357, + "grad_norm": 1.4176149408001522, + "learning_rate": 6.928982725527832e-06, + "loss": 0.7478, + "step": 362 + }, + { + "epoch": 0.3485357657225156, + "grad_norm": 1.055027957209198, + "learning_rate": 6.9481765834932824e-06, + "loss": 0.7025, + "step": 363 + }, + { + "epoch": 0.34949591934709556, + "grad_norm": 0.9237996976063603, + "learning_rate": 6.967370441458734e-06, + "loss": 0.7641, + "step": 364 + }, + { + "epoch": 0.35045607297167547, + "grad_norm": 1.0023494135672188, + "learning_rate": 6.986564299424184e-06, + "loss": 0.773, + "step": 365 + }, + { + "epoch": 0.3514162265962554, + "grad_norm": 1.1170315829753363, + "learning_rate": 7.005758157389636e-06, + "loss": 0.7737, + "step": 366 + }, + { + "epoch": 0.35237638022083534, + "grad_norm": 1.6571294664693028, + "learning_rate": 7.024952015355086e-06, + "loss": 0.7678, + "step": 367 + }, + { + "epoch": 0.35333653384541525, + "grad_norm": 1.1463208677096601, + "learning_rate": 7.044145873320538e-06, + "loss": 0.7038, + "step": 368 + }, + { + "epoch": 0.3542966874699952, + "grad_norm": 1.2041496401683789, + "learning_rate": 7.063339731285989e-06, + "loss": 0.8187, + "step": 369 + }, + { + "epoch": 0.3552568410945751, + "grad_norm": 1.3123729805107278, + "learning_rate": 7.08253358925144e-06, + "loss": 0.7559, + "step": 370 + }, + { + "epoch": 0.3562169947191551, + "grad_norm": 1.2531666737811658, + "learning_rate": 7.101727447216891e-06, + "loss": 0.8967, + "step": 371 + }, + { + "epoch": 0.357177148343735, + "grad_norm": 1.6044073246344652, + "learning_rate": 7.120921305182342e-06, + "loss": 0.8522, + "step": 372 + }, + { + "epoch": 0.35813730196831495, + "grad_norm": 1.46386764159694, + "learning_rate": 7.1401151631477925e-06, + "loss": 0.8204, + "step": 373 + }, + { + "epoch": 0.35909745559289485, + "grad_norm": 1.124678220450824, + "learning_rate": 7.159309021113245e-06, + "loss": 0.7096, + "step": 374 + }, + { + "epoch": 0.3600576092174748, + "grad_norm": 1.0907290673814356, + "learning_rate": 7.178502879078696e-06, + "loss": 0.6543, + "step": 375 + }, + { + "epoch": 0.3610177628420547, + "grad_norm": 1.2365841743837318, + "learning_rate": 7.1976967370441466e-06, + "loss": 0.7116, + "step": 376 + }, + { + "epoch": 0.3619779164666347, + "grad_norm": 1.1952106366739865, + "learning_rate": 7.216890595009598e-06, + "loss": 0.6948, + "step": 377 + }, + { + "epoch": 0.3629380700912146, + "grad_norm": 1.156231992977645, + "learning_rate": 7.236084452975048e-06, + "loss": 0.7459, + "step": 378 + }, + { + "epoch": 0.3638982237157945, + "grad_norm": 1.0787123026665089, + "learning_rate": 7.2552783109405e-06, + "loss": 0.7374, + "step": 379 + }, + { + "epoch": 0.36485837734037446, + "grad_norm": 1.161408465298755, + "learning_rate": 7.27447216890595e-06, + "loss": 0.7205, + "step": 380 + }, + { + "epoch": 0.36581853096495437, + "grad_norm": 1.155974667611427, + "learning_rate": 7.2936660268714024e-06, + "loss": 0.8232, + "step": 381 + }, + { + "epoch": 0.36677868458953433, + "grad_norm": 1.2962147768859738, + "learning_rate": 7.312859884836853e-06, + "loss": 0.6994, + "step": 382 + }, + { + "epoch": 0.36773883821411424, + "grad_norm": 1.0432828620281644, + "learning_rate": 7.332053742802304e-06, + "loss": 0.7713, + "step": 383 + }, + { + "epoch": 0.3686989918386942, + "grad_norm": 1.6400306383833076, + "learning_rate": 7.351247600767755e-06, + "loss": 0.7455, + "step": 384 + }, + { + "epoch": 0.3696591454632741, + "grad_norm": 1.8687163253976649, + "learning_rate": 7.370441458733206e-06, + "loss": 0.8921, + "step": 385 + }, + { + "epoch": 0.37061929908785407, + "grad_norm": 1.15552226222218, + "learning_rate": 7.389635316698657e-06, + "loss": 0.8523, + "step": 386 + }, + { + "epoch": 0.371579452712434, + "grad_norm": 1.2197160624499277, + "learning_rate": 7.408829174664108e-06, + "loss": 0.8379, + "step": 387 + }, + { + "epoch": 0.37253960633701394, + "grad_norm": 0.9473852646791292, + "learning_rate": 7.4280230326295585e-06, + "loss": 0.6417, + "step": 388 + }, + { + "epoch": 0.37349975996159385, + "grad_norm": 1.0627377340766202, + "learning_rate": 7.447216890595011e-06, + "loss": 0.7168, + "step": 389 + }, + { + "epoch": 0.3744599135861738, + "grad_norm": 1.9613488950786884, + "learning_rate": 7.466410748560461e-06, + "loss": 0.832, + "step": 390 + }, + { + "epoch": 0.3754200672107537, + "grad_norm": 0.9768112116462052, + "learning_rate": 7.4856046065259125e-06, + "loss": 0.7722, + "step": 391 + }, + { + "epoch": 0.3763802208353337, + "grad_norm": 1.1088139902732483, + "learning_rate": 7.504798464491363e-06, + "loss": 0.7741, + "step": 392 + }, + { + "epoch": 0.3773403744599136, + "grad_norm": 1.1277644988362703, + "learning_rate": 7.523992322456814e-06, + "loss": 0.7548, + "step": 393 + }, + { + "epoch": 0.3783005280844935, + "grad_norm": 1.3490775794268282, + "learning_rate": 7.543186180422265e-06, + "loss": 0.7274, + "step": 394 + }, + { + "epoch": 0.37926068170907346, + "grad_norm": 1.2752102918651995, + "learning_rate": 7.562380038387716e-06, + "loss": 0.7634, + "step": 395 + }, + { + "epoch": 0.38022083533365336, + "grad_norm": 1.3248095204390724, + "learning_rate": 7.581573896353167e-06, + "loss": 0.7598, + "step": 396 + }, + { + "epoch": 0.3811809889582333, + "grad_norm": 1.4388469294151665, + "learning_rate": 7.600767754318619e-06, + "loss": 0.8855, + "step": 397 + }, + { + "epoch": 0.38214114258281323, + "grad_norm": 1.0323243137009235, + "learning_rate": 7.61996161228407e-06, + "loss": 0.6274, + "step": 398 + }, + { + "epoch": 0.3831012962073932, + "grad_norm": 1.4550500195070324, + "learning_rate": 7.63915547024952e-06, + "loss": 0.7161, + "step": 399 + }, + { + "epoch": 0.3840614498319731, + "grad_norm": 1.0927734198673136, + "learning_rate": 7.658349328214972e-06, + "loss": 0.6599, + "step": 400 + }, + { + "epoch": 0.38502160345655306, + "grad_norm": 1.0386436633896041, + "learning_rate": 7.677543186180423e-06, + "loss": 0.6468, + "step": 401 + }, + { + "epoch": 0.38598175708113297, + "grad_norm": 1.2385527094382942, + "learning_rate": 7.696737044145875e-06, + "loss": 0.6902, + "step": 402 + }, + { + "epoch": 0.38694191070571293, + "grad_norm": 1.128496898995166, + "learning_rate": 7.715930902111324e-06, + "loss": 0.8124, + "step": 403 + }, + { + "epoch": 0.38790206433029284, + "grad_norm": 0.8982281997977789, + "learning_rate": 7.735124760076776e-06, + "loss": 0.6971, + "step": 404 + }, + { + "epoch": 0.3888622179548728, + "grad_norm": 1.3087047484167935, + "learning_rate": 7.754318618042227e-06, + "loss": 0.7527, + "step": 405 + }, + { + "epoch": 0.3898223715794527, + "grad_norm": 1.0148715452035035, + "learning_rate": 7.773512476007678e-06, + "loss": 0.6074, + "step": 406 + }, + { + "epoch": 0.39078252520403267, + "grad_norm": 1.2181305573923982, + "learning_rate": 7.79270633397313e-06, + "loss": 0.7455, + "step": 407 + }, + { + "epoch": 0.3917426788286126, + "grad_norm": 1.257923950398552, + "learning_rate": 7.811900191938581e-06, + "loss": 0.7506, + "step": 408 + }, + { + "epoch": 0.3927028324531925, + "grad_norm": 0.9385542220228643, + "learning_rate": 7.83109404990403e-06, + "loss": 0.7203, + "step": 409 + }, + { + "epoch": 0.39366298607777245, + "grad_norm": 0.9489668803662427, + "learning_rate": 7.850287907869482e-06, + "loss": 0.6797, + "step": 410 + }, + { + "epoch": 0.39462313970235235, + "grad_norm": 1.138399536824701, + "learning_rate": 7.869481765834934e-06, + "loss": 0.721, + "step": 411 + }, + { + "epoch": 0.3955832933269323, + "grad_norm": 0.8761756089098698, + "learning_rate": 7.888675623800385e-06, + "loss": 0.7422, + "step": 412 + }, + { + "epoch": 0.3965434469515122, + "grad_norm": 1.136293201008876, + "learning_rate": 7.907869481765835e-06, + "loss": 0.7577, + "step": 413 + }, + { + "epoch": 0.3975036005760922, + "grad_norm": 0.9088338251669745, + "learning_rate": 7.927063339731288e-06, + "loss": 0.6826, + "step": 414 + }, + { + "epoch": 0.3984637542006721, + "grad_norm": 1.4226305549508294, + "learning_rate": 7.946257197696737e-06, + "loss": 0.7768, + "step": 415 + }, + { + "epoch": 0.39942390782525206, + "grad_norm": 1.6031042617246214, + "learning_rate": 7.965451055662189e-06, + "loss": 0.9425, + "step": 416 + }, + { + "epoch": 0.40038406144983196, + "grad_norm": 1.1182897641253295, + "learning_rate": 7.98464491362764e-06, + "loss": 0.744, + "step": 417 + }, + { + "epoch": 0.4013442150744119, + "grad_norm": 1.1624749434686374, + "learning_rate": 8.003838771593091e-06, + "loss": 0.8097, + "step": 418 + }, + { + "epoch": 0.40230436869899183, + "grad_norm": 1.0640000938896768, + "learning_rate": 8.023032629558541e-06, + "loss": 0.7217, + "step": 419 + }, + { + "epoch": 0.4032645223235718, + "grad_norm": 1.029486325121184, + "learning_rate": 8.042226487523992e-06, + "loss": 0.734, + "step": 420 + }, + { + "epoch": 0.4042246759481517, + "grad_norm": 1.4704604374496304, + "learning_rate": 8.061420345489444e-06, + "loss": 0.8084, + "step": 421 + }, + { + "epoch": 0.40518482957273166, + "grad_norm": 0.93094639130799, + "learning_rate": 8.080614203454895e-06, + "loss": 0.6201, + "step": 422 + }, + { + "epoch": 0.40614498319731157, + "grad_norm": 1.4281187881326085, + "learning_rate": 8.099808061420346e-06, + "loss": 0.8938, + "step": 423 + }, + { + "epoch": 0.4071051368218915, + "grad_norm": 1.147134853007333, + "learning_rate": 8.119001919385798e-06, + "loss": 0.7301, + "step": 424 + }, + { + "epoch": 0.40806529044647144, + "grad_norm": 1.0973175759148972, + "learning_rate": 8.138195777351249e-06, + "loss": 0.8135, + "step": 425 + }, + { + "epoch": 0.40902544407105135, + "grad_norm": 1.4733627401673446, + "learning_rate": 8.157389635316699e-06, + "loss": 0.9395, + "step": 426 + }, + { + "epoch": 0.4099855976956313, + "grad_norm": 0.9007463265785779, + "learning_rate": 8.176583493282152e-06, + "loss": 0.718, + "step": 427 + }, + { + "epoch": 0.4109457513202112, + "grad_norm": 0.9380859442585787, + "learning_rate": 8.195777351247601e-06, + "loss": 0.7025, + "step": 428 + }, + { + "epoch": 0.4119059049447912, + "grad_norm": 1.1911460339975852, + "learning_rate": 8.214971209213053e-06, + "loss": 0.8301, + "step": 429 + }, + { + "epoch": 0.4128660585693711, + "grad_norm": 1.12943551602526, + "learning_rate": 8.234165067178504e-06, + "loss": 0.8313, + "step": 430 + }, + { + "epoch": 0.41382621219395105, + "grad_norm": 1.185153963997797, + "learning_rate": 8.253358925143955e-06, + "loss": 0.7322, + "step": 431 + }, + { + "epoch": 0.41478636581853096, + "grad_norm": 0.9500772402083794, + "learning_rate": 8.272552783109405e-06, + "loss": 0.7539, + "step": 432 + }, + { + "epoch": 0.4157465194431109, + "grad_norm": 0.9685064250353801, + "learning_rate": 8.291746641074856e-06, + "loss": 0.7116, + "step": 433 + }, + { + "epoch": 0.4167066730676908, + "grad_norm": 1.439697940069654, + "learning_rate": 8.310940499040308e-06, + "loss": 0.8389, + "step": 434 + }, + { + "epoch": 0.4176668266922708, + "grad_norm": 1.650060931842286, + "learning_rate": 8.330134357005759e-06, + "loss": 0.8147, + "step": 435 + }, + { + "epoch": 0.4186269803168507, + "grad_norm": 1.5097188148620866, + "learning_rate": 8.34932821497121e-06, + "loss": 0.8671, + "step": 436 + }, + { + "epoch": 0.4195871339414306, + "grad_norm": 1.0491157820672372, + "learning_rate": 8.368522072936662e-06, + "loss": 0.7125, + "step": 437 + }, + { + "epoch": 0.42054728756601056, + "grad_norm": 1.2715868974375322, + "learning_rate": 8.387715930902111e-06, + "loss": 0.7865, + "step": 438 + }, + { + "epoch": 0.42150744119059047, + "grad_norm": 1.091876813620556, + "learning_rate": 8.406909788867563e-06, + "loss": 0.7577, + "step": 439 + }, + { + "epoch": 0.42246759481517043, + "grad_norm": 1.2597352342543193, + "learning_rate": 8.426103646833014e-06, + "loss": 0.823, + "step": 440 + }, + { + "epoch": 0.42342774843975034, + "grad_norm": 1.2434732628786342, + "learning_rate": 8.445297504798465e-06, + "loss": 0.7093, + "step": 441 + }, + { + "epoch": 0.4243879020643303, + "grad_norm": 1.1444923505918632, + "learning_rate": 8.464491362763915e-06, + "loss": 0.6886, + "step": 442 + }, + { + "epoch": 0.4253480556889102, + "grad_norm": 1.0409174548837277, + "learning_rate": 8.483685220729368e-06, + "loss": 0.7463, + "step": 443 + }, + { + "epoch": 0.42630820931349017, + "grad_norm": 1.2007743151913164, + "learning_rate": 8.502879078694818e-06, + "loss": 0.7989, + "step": 444 + }, + { + "epoch": 0.4272683629380701, + "grad_norm": 1.2718837493354382, + "learning_rate": 8.522072936660269e-06, + "loss": 0.7426, + "step": 445 + }, + { + "epoch": 0.42822851656265004, + "grad_norm": 1.2568901583053014, + "learning_rate": 8.54126679462572e-06, + "loss": 0.8255, + "step": 446 + }, + { + "epoch": 0.42918867018722995, + "grad_norm": 0.9969486622008054, + "learning_rate": 8.560460652591172e-06, + "loss": 0.6967, + "step": 447 + }, + { + "epoch": 0.4301488238118099, + "grad_norm": 1.1899514801382536, + "learning_rate": 8.579654510556623e-06, + "loss": 0.6996, + "step": 448 + }, + { + "epoch": 0.4311089774363898, + "grad_norm": 1.0152504780424239, + "learning_rate": 8.598848368522073e-06, + "loss": 0.7519, + "step": 449 + }, + { + "epoch": 0.4320691310609698, + "grad_norm": 1.2901043367604972, + "learning_rate": 8.618042226487526e-06, + "loss": 0.7782, + "step": 450 + }, + { + "epoch": 0.4330292846855497, + "grad_norm": 0.9785594955317184, + "learning_rate": 8.637236084452976e-06, + "loss": 0.6279, + "step": 451 + }, + { + "epoch": 0.4339894383101296, + "grad_norm": 0.9454490728083296, + "learning_rate": 8.656429942418427e-06, + "loss": 0.6849, + "step": 452 + }, + { + "epoch": 0.43494959193470956, + "grad_norm": 1.4515223195648956, + "learning_rate": 8.675623800383878e-06, + "loss": 0.8921, + "step": 453 + }, + { + "epoch": 0.43590974555928946, + "grad_norm": 1.1925663852839914, + "learning_rate": 8.69481765834933e-06, + "loss": 0.7893, + "step": 454 + }, + { + "epoch": 0.4368698991838694, + "grad_norm": 0.9817277941974228, + "learning_rate": 8.71401151631478e-06, + "loss": 0.7494, + "step": 455 + }, + { + "epoch": 0.43783005280844933, + "grad_norm": 1.073831518730329, + "learning_rate": 8.73320537428023e-06, + "loss": 0.6855, + "step": 456 + }, + { + "epoch": 0.4387902064330293, + "grad_norm": 1.1640819618236569, + "learning_rate": 8.752399232245682e-06, + "loss": 0.8602, + "step": 457 + }, + { + "epoch": 0.4397503600576092, + "grad_norm": 0.9732047167295669, + "learning_rate": 8.771593090211133e-06, + "loss": 0.7405, + "step": 458 + }, + { + "epoch": 0.44071051368218916, + "grad_norm": 1.4253504559694403, + "learning_rate": 8.790786948176585e-06, + "loss": 0.8094, + "step": 459 + }, + { + "epoch": 0.44167066730676907, + "grad_norm": 1.1120566176209299, + "learning_rate": 8.809980806142036e-06, + "loss": 0.8219, + "step": 460 + }, + { + "epoch": 0.44263082093134903, + "grad_norm": 1.1476137020908006, + "learning_rate": 8.829174664107486e-06, + "loss": 0.7989, + "step": 461 + }, + { + "epoch": 0.44359097455592894, + "grad_norm": 1.0376270729267838, + "learning_rate": 8.848368522072937e-06, + "loss": 0.7844, + "step": 462 + }, + { + "epoch": 0.4445511281805089, + "grad_norm": 1.88847392184355, + "learning_rate": 8.867562380038388e-06, + "loss": 0.8719, + "step": 463 + }, + { + "epoch": 0.4455112818050888, + "grad_norm": 1.1570205261545414, + "learning_rate": 8.88675623800384e-06, + "loss": 0.7337, + "step": 464 + }, + { + "epoch": 0.4464714354296688, + "grad_norm": 1.2953565641587041, + "learning_rate": 8.905950095969291e-06, + "loss": 0.6254, + "step": 465 + }, + { + "epoch": 0.4474315890542487, + "grad_norm": 1.1688420038847154, + "learning_rate": 8.925143953934742e-06, + "loss": 0.8725, + "step": 466 + }, + { + "epoch": 0.4483917426788286, + "grad_norm": 1.0335620781982393, + "learning_rate": 8.944337811900192e-06, + "loss": 0.7754, + "step": 467 + }, + { + "epoch": 0.44935189630340855, + "grad_norm": 0.8619939885192579, + "learning_rate": 8.963531669865643e-06, + "loss": 0.6979, + "step": 468 + }, + { + "epoch": 0.45031204992798846, + "grad_norm": 1.2976131274666975, + "learning_rate": 8.982725527831095e-06, + "loss": 0.667, + "step": 469 + }, + { + "epoch": 0.4512722035525684, + "grad_norm": 1.4021578605304392, + "learning_rate": 9.001919385796546e-06, + "loss": 0.8609, + "step": 470 + }, + { + "epoch": 0.4522323571771483, + "grad_norm": 1.2024029118089836, + "learning_rate": 9.021113243761996e-06, + "loss": 0.8361, + "step": 471 + }, + { + "epoch": 0.4531925108017283, + "grad_norm": 1.0460626952391077, + "learning_rate": 9.040307101727449e-06, + "loss": 0.7147, + "step": 472 + }, + { + "epoch": 0.4541526644263082, + "grad_norm": 1.2908116152032034, + "learning_rate": 9.0595009596929e-06, + "loss": 0.8197, + "step": 473 + }, + { + "epoch": 0.45511281805088816, + "grad_norm": 1.03318946346981, + "learning_rate": 9.07869481765835e-06, + "loss": 0.768, + "step": 474 + }, + { + "epoch": 0.45607297167546806, + "grad_norm": 1.3220638446805137, + "learning_rate": 9.097888675623801e-06, + "loss": 0.8007, + "step": 475 + }, + { + "epoch": 0.457033125300048, + "grad_norm": 1.1554417914733812, + "learning_rate": 9.117082533589252e-06, + "loss": 0.7636, + "step": 476 + }, + { + "epoch": 0.45799327892462793, + "grad_norm": 1.1310593331314387, + "learning_rate": 9.136276391554704e-06, + "loss": 0.6219, + "step": 477 + }, + { + "epoch": 0.4589534325492079, + "grad_norm": 1.0875448265190137, + "learning_rate": 9.155470249520153e-06, + "loss": 0.6831, + "step": 478 + }, + { + "epoch": 0.4599135861737878, + "grad_norm": 1.4031599853280303, + "learning_rate": 9.174664107485606e-06, + "loss": 0.7808, + "step": 479 + }, + { + "epoch": 0.46087373979836777, + "grad_norm": 1.391170940885926, + "learning_rate": 9.193857965451056e-06, + "loss": 0.7529, + "step": 480 + }, + { + "epoch": 0.46183389342294767, + "grad_norm": 1.0087271205949113, + "learning_rate": 9.213051823416507e-06, + "loss": 0.6919, + "step": 481 + }, + { + "epoch": 0.4627940470475276, + "grad_norm": 0.9830620697275546, + "learning_rate": 9.232245681381959e-06, + "loss": 0.6466, + "step": 482 + }, + { + "epoch": 0.46375420067210754, + "grad_norm": 1.1083503739334748, + "learning_rate": 9.25143953934741e-06, + "loss": 0.6824, + "step": 483 + }, + { + "epoch": 0.46471435429668745, + "grad_norm": 1.290594688103715, + "learning_rate": 9.27063339731286e-06, + "loss": 0.7947, + "step": 484 + }, + { + "epoch": 0.4656745079212674, + "grad_norm": 1.0041716569519412, + "learning_rate": 9.289827255278311e-06, + "loss": 0.6783, + "step": 485 + }, + { + "epoch": 0.4666346615458473, + "grad_norm": 1.1598686934846796, + "learning_rate": 9.309021113243763e-06, + "loss": 0.6613, + "step": 486 + }, + { + "epoch": 0.4675948151704273, + "grad_norm": 1.260111175387368, + "learning_rate": 9.328214971209214e-06, + "loss": 0.7378, + "step": 487 + }, + { + "epoch": 0.4685549687950072, + "grad_norm": 0.9393424317838525, + "learning_rate": 9.347408829174665e-06, + "loss": 0.8005, + "step": 488 + }, + { + "epoch": 0.46951512241958715, + "grad_norm": 1.1210853185169674, + "learning_rate": 9.366602687140117e-06, + "loss": 0.6807, + "step": 489 + }, + { + "epoch": 0.47047527604416706, + "grad_norm": 1.4438230017220477, + "learning_rate": 9.385796545105566e-06, + "loss": 0.7997, + "step": 490 + }, + { + "epoch": 0.471435429668747, + "grad_norm": 1.3277569402674447, + "learning_rate": 9.404990403071018e-06, + "loss": 0.7416, + "step": 491 + }, + { + "epoch": 0.4723955832933269, + "grad_norm": 1.2432681770756384, + "learning_rate": 9.424184261036469e-06, + "loss": 0.7542, + "step": 492 + }, + { + "epoch": 0.4733557369179069, + "grad_norm": 0.9683846009176716, + "learning_rate": 9.44337811900192e-06, + "loss": 0.6364, + "step": 493 + }, + { + "epoch": 0.4743158905424868, + "grad_norm": 1.0629872575186827, + "learning_rate": 9.46257197696737e-06, + "loss": 0.7682, + "step": 494 + }, + { + "epoch": 0.47527604416706676, + "grad_norm": 1.0658891130678316, + "learning_rate": 9.481765834932823e-06, + "loss": 0.7502, + "step": 495 + }, + { + "epoch": 0.47623619779164666, + "grad_norm": 1.3363058386178244, + "learning_rate": 9.500959692898273e-06, + "loss": 0.8112, + "step": 496 + }, + { + "epoch": 0.47719635141622657, + "grad_norm": 1.0265023768250465, + "learning_rate": 9.520153550863724e-06, + "loss": 0.7149, + "step": 497 + }, + { + "epoch": 0.47815650504080653, + "grad_norm": 1.2889831472177289, + "learning_rate": 9.539347408829175e-06, + "loss": 0.7936, + "step": 498 + }, + { + "epoch": 0.47911665866538644, + "grad_norm": 1.250136499572121, + "learning_rate": 9.558541266794627e-06, + "loss": 0.6048, + "step": 499 + }, + { + "epoch": 0.4800768122899664, + "grad_norm": 1.3111050762625585, + "learning_rate": 9.577735124760078e-06, + "loss": 0.728, + "step": 500 + }, + { + "epoch": 0.4810369659145463, + "grad_norm": 1.7608381448641053, + "learning_rate": 9.59692898272553e-06, + "loss": 0.7994, + "step": 501 + }, + { + "epoch": 0.4819971195391263, + "grad_norm": 1.2290182303499815, + "learning_rate": 9.61612284069098e-06, + "loss": 0.8003, + "step": 502 + }, + { + "epoch": 0.4829572731637062, + "grad_norm": 1.6135633512832115, + "learning_rate": 9.63531669865643e-06, + "loss": 0.838, + "step": 503 + }, + { + "epoch": 0.48391742678828614, + "grad_norm": 1.164478249652597, + "learning_rate": 9.654510556621882e-06, + "loss": 0.7259, + "step": 504 + }, + { + "epoch": 0.48487758041286605, + "grad_norm": 0.9993561066591795, + "learning_rate": 9.673704414587333e-06, + "loss": 0.7748, + "step": 505 + }, + { + "epoch": 0.485837734037446, + "grad_norm": 0.8358295488869283, + "learning_rate": 9.692898272552784e-06, + "loss": 0.6693, + "step": 506 + }, + { + "epoch": 0.4867978876620259, + "grad_norm": 1.1106122084032795, + "learning_rate": 9.712092130518234e-06, + "loss": 0.721, + "step": 507 + }, + { + "epoch": 0.4877580412866059, + "grad_norm": 1.487188285181112, + "learning_rate": 9.731285988483687e-06, + "loss": 0.7509, + "step": 508 + }, + { + "epoch": 0.4887181949111858, + "grad_norm": 1.1402411926952243, + "learning_rate": 9.750479846449137e-06, + "loss": 0.6647, + "step": 509 + }, + { + "epoch": 0.4896783485357657, + "grad_norm": 1.0367107322127078, + "learning_rate": 9.769673704414588e-06, + "loss": 0.7887, + "step": 510 + }, + { + "epoch": 0.49063850216034566, + "grad_norm": 1.3987993604220361, + "learning_rate": 9.78886756238004e-06, + "loss": 0.7736, + "step": 511 + }, + { + "epoch": 0.49159865578492556, + "grad_norm": 1.259632180608351, + "learning_rate": 9.80806142034549e-06, + "loss": 0.7329, + "step": 512 + }, + { + "epoch": 0.4925588094095055, + "grad_norm": 1.1142094683577137, + "learning_rate": 9.82725527831094e-06, + "loss": 0.6632, + "step": 513 + }, + { + "epoch": 0.49351896303408543, + "grad_norm": 1.0929712059339454, + "learning_rate": 9.846449136276392e-06, + "loss": 0.7179, + "step": 514 + }, + { + "epoch": 0.4944791166586654, + "grad_norm": 1.306135215088198, + "learning_rate": 9.865642994241843e-06, + "loss": 0.6585, + "step": 515 + }, + { + "epoch": 0.4954392702832453, + "grad_norm": 1.1252208319095651, + "learning_rate": 9.884836852207294e-06, + "loss": 0.8159, + "step": 516 + }, + { + "epoch": 0.49639942390782527, + "grad_norm": 1.486805087876587, + "learning_rate": 9.904030710172746e-06, + "loss": 0.8046, + "step": 517 + }, + { + "epoch": 0.4973595775324052, + "grad_norm": 1.1393464745618724, + "learning_rate": 9.923224568138197e-06, + "loss": 0.7898, + "step": 518 + }, + { + "epoch": 0.49831973115698514, + "grad_norm": 1.224267542983065, + "learning_rate": 9.942418426103647e-06, + "loss": 0.7973, + "step": 519 + }, + { + "epoch": 0.49927988478156504, + "grad_norm": 1.1859907199010784, + "learning_rate": 9.961612284069098e-06, + "loss": 0.8521, + "step": 520 + }, + { + "epoch": 0.500240038406145, + "grad_norm": 1.5259121980424533, + "learning_rate": 9.98080614203455e-06, + "loss": 0.8218, + "step": 521 + }, + { + "epoch": 0.501200192030725, + "grad_norm": 1.2072523535698836, + "learning_rate": 1e-05, + "loss": 0.7709, + "step": 522 + }, + { + "epoch": 0.5021603456553049, + "grad_norm": 1.2171256402748378, + "learning_rate": 9.999998877776725e-06, + "loss": 0.652, + "step": 523 + }, + { + "epoch": 0.5031204992798848, + "grad_norm": 1.049515981094851, + "learning_rate": 9.999995511107399e-06, + "loss": 0.7062, + "step": 524 + }, + { + "epoch": 0.5040806529044647, + "grad_norm": 1.0998316431712438, + "learning_rate": 9.999989899993536e-06, + "loss": 0.7867, + "step": 525 + }, + { + "epoch": 0.5050408065290446, + "grad_norm": 1.1011145838187344, + "learning_rate": 9.999982044437652e-06, + "loss": 0.6948, + "step": 526 + }, + { + "epoch": 0.5060009601536246, + "grad_norm": 1.0336696003692087, + "learning_rate": 9.999971944443276e-06, + "loss": 0.702, + "step": 527 + }, + { + "epoch": 0.5069611137782045, + "grad_norm": 1.1680220201586808, + "learning_rate": 9.999959600014942e-06, + "loss": 0.7674, + "step": 528 + }, + { + "epoch": 0.5079212674027844, + "grad_norm": 0.9829065057909755, + "learning_rate": 9.999945011158189e-06, + "loss": 0.7185, + "step": 529 + }, + { + "epoch": 0.5088814210273643, + "grad_norm": 1.0062085732737889, + "learning_rate": 9.999928177879567e-06, + "loss": 0.7118, + "step": 530 + }, + { + "epoch": 0.5098415746519444, + "grad_norm": 1.399248665460084, + "learning_rate": 9.999909100186634e-06, + "loss": 0.8042, + "step": 531 + }, + { + "epoch": 0.5108017282765243, + "grad_norm": 0.9818054618777152, + "learning_rate": 9.999887778087948e-06, + "loss": 0.7058, + "step": 532 + }, + { + "epoch": 0.5117618819011042, + "grad_norm": 1.165173517671421, + "learning_rate": 9.999864211593088e-06, + "loss": 0.7926, + "step": 533 + }, + { + "epoch": 0.5127220355256841, + "grad_norm": 1.1505524949085641, + "learning_rate": 9.999838400712627e-06, + "loss": 0.8351, + "step": 534 + }, + { + "epoch": 0.5136821891502641, + "grad_norm": 1.1208313391537508, + "learning_rate": 9.999810345458153e-06, + "loss": 0.605, + "step": 535 + }, + { + "epoch": 0.514642342774844, + "grad_norm": 1.1968339780644133, + "learning_rate": 9.999780045842262e-06, + "loss": 0.644, + "step": 536 + }, + { + "epoch": 0.5156024963994239, + "grad_norm": 1.0693441645105062, + "learning_rate": 9.999747501878551e-06, + "loss": 0.6703, + "step": 537 + }, + { + "epoch": 0.5165626500240038, + "grad_norm": 1.2133540246571193, + "learning_rate": 9.999712713581634e-06, + "loss": 0.8123, + "step": 538 + }, + { + "epoch": 0.5175228036485837, + "grad_norm": 1.0453334308646136, + "learning_rate": 9.99967568096712e-06, + "loss": 0.6976, + "step": 539 + }, + { + "epoch": 0.5184829572731637, + "grad_norm": 1.3264475966595732, + "learning_rate": 9.999636404051638e-06, + "loss": 0.7837, + "step": 540 + }, + { + "epoch": 0.5194431108977436, + "grad_norm": 1.0861707447698705, + "learning_rate": 9.999594882852817e-06, + "loss": 0.7799, + "step": 541 + }, + { + "epoch": 0.5204032645223235, + "grad_norm": 1.2946964966740842, + "learning_rate": 9.999551117389298e-06, + "loss": 0.6334, + "step": 542 + }, + { + "epoch": 0.5213634181469035, + "grad_norm": 0.929657847384208, + "learning_rate": 9.999505107680722e-06, + "loss": 0.8083, + "step": 543 + }, + { + "epoch": 0.5223235717714835, + "grad_norm": 1.1401440990502336, + "learning_rate": 9.999456853747744e-06, + "loss": 0.7132, + "step": 544 + }, + { + "epoch": 0.5232837253960634, + "grad_norm": 1.0340277545375338, + "learning_rate": 9.999406355612029e-06, + "loss": 0.7124, + "step": 545 + }, + { + "epoch": 0.5242438790206433, + "grad_norm": 1.247007016370199, + "learning_rate": 9.99935361329624e-06, + "loss": 0.7065, + "step": 546 + }, + { + "epoch": 0.5252040326452232, + "grad_norm": 0.9998755107014989, + "learning_rate": 9.999298626824053e-06, + "loss": 0.7451, + "step": 547 + }, + { + "epoch": 0.5261641862698032, + "grad_norm": 1.2332553479270054, + "learning_rate": 9.999241396220152e-06, + "loss": 0.7216, + "step": 548 + }, + { + "epoch": 0.5271243398943831, + "grad_norm": 1.2155641200696106, + "learning_rate": 9.999181921510229e-06, + "loss": 0.782, + "step": 549 + }, + { + "epoch": 0.528084493518963, + "grad_norm": 1.51168636775896, + "learning_rate": 9.999120202720978e-06, + "loss": 0.9307, + "step": 550 + }, + { + "epoch": 0.5290446471435429, + "grad_norm": 1.0359215075770465, + "learning_rate": 9.999056239880107e-06, + "loss": 0.6104, + "step": 551 + }, + { + "epoch": 0.530004800768123, + "grad_norm": 1.0464957447369558, + "learning_rate": 9.998990033016327e-06, + "loss": 0.6045, + "step": 552 + }, + { + "epoch": 0.5309649543927029, + "grad_norm": 1.1114456620613888, + "learning_rate": 9.998921582159357e-06, + "loss": 0.8023, + "step": 553 + }, + { + "epoch": 0.5319251080172828, + "grad_norm": 1.5494709814113228, + "learning_rate": 9.998850887339923e-06, + "loss": 0.7696, + "step": 554 + }, + { + "epoch": 0.5328852616418627, + "grad_norm": 1.4890090358538963, + "learning_rate": 9.998777948589762e-06, + "loss": 0.7569, + "step": 555 + }, + { + "epoch": 0.5338454152664426, + "grad_norm": 0.8865139886047781, + "learning_rate": 9.998702765941613e-06, + "loss": 0.7251, + "step": 556 + }, + { + "epoch": 0.5348055688910226, + "grad_norm": 1.3058829182513438, + "learning_rate": 9.998625339429227e-06, + "loss": 0.7223, + "step": 557 + }, + { + "epoch": 0.5357657225156025, + "grad_norm": 1.276478715271317, + "learning_rate": 9.998545669087357e-06, + "loss": 0.7673, + "step": 558 + }, + { + "epoch": 0.5367258761401824, + "grad_norm": 1.1676751331206674, + "learning_rate": 9.998463754951767e-06, + "loss": 0.7617, + "step": 559 + }, + { + "epoch": 0.5376860297647623, + "grad_norm": 1.2123803522040177, + "learning_rate": 9.998379597059228e-06, + "loss": 0.7129, + "step": 560 + }, + { + "epoch": 0.5386461833893423, + "grad_norm": 1.3814639588806619, + "learning_rate": 9.99829319544752e-06, + "loss": 0.8845, + "step": 561 + }, + { + "epoch": 0.5396063370139222, + "grad_norm": 1.11055749603962, + "learning_rate": 9.998204550155424e-06, + "loss": 0.7557, + "step": 562 + }, + { + "epoch": 0.5405664906385022, + "grad_norm": 1.0674759443075517, + "learning_rate": 9.998113661222732e-06, + "loss": 0.7595, + "step": 563 + }, + { + "epoch": 0.5415266442630821, + "grad_norm": 0.994763687652783, + "learning_rate": 9.998020528690247e-06, + "loss": 0.814, + "step": 564 + }, + { + "epoch": 0.5424867978876621, + "grad_norm": 1.0417008849188154, + "learning_rate": 9.99792515259977e-06, + "loss": 0.7379, + "step": 565 + }, + { + "epoch": 0.543446951512242, + "grad_norm": 1.1478987486458458, + "learning_rate": 9.997827532994118e-06, + "loss": 0.7426, + "step": 566 + }, + { + "epoch": 0.5444071051368219, + "grad_norm": 1.1565190817285003, + "learning_rate": 9.99772766991711e-06, + "loss": 0.7203, + "step": 567 + }, + { + "epoch": 0.5453672587614018, + "grad_norm": 1.1192363371512206, + "learning_rate": 9.997625563413574e-06, + "loss": 0.7739, + "step": 568 + }, + { + "epoch": 0.5463274123859817, + "grad_norm": 0.9567002043036419, + "learning_rate": 9.997521213529345e-06, + "loss": 0.705, + "step": 569 + }, + { + "epoch": 0.5472875660105617, + "grad_norm": 1.0897891589405941, + "learning_rate": 9.997414620311262e-06, + "loss": 0.6637, + "step": 570 + }, + { + "epoch": 0.5482477196351416, + "grad_norm": 0.9563438824899532, + "learning_rate": 9.997305783807176e-06, + "loss": 0.7579, + "step": 571 + }, + { + "epoch": 0.5492078732597215, + "grad_norm": 1.1554010195538116, + "learning_rate": 9.997194704065942e-06, + "loss": 0.8239, + "step": 572 + }, + { + "epoch": 0.5501680268843014, + "grad_norm": 1.2004353682200677, + "learning_rate": 9.997081381137424e-06, + "loss": 0.8628, + "step": 573 + }, + { + "epoch": 0.5511281805088815, + "grad_norm": 1.0220163262467359, + "learning_rate": 9.996965815072489e-06, + "loss": 0.6323, + "step": 574 + }, + { + "epoch": 0.5520883341334614, + "grad_norm": 1.6768057570791945, + "learning_rate": 9.996848005923012e-06, + "loss": 0.7965, + "step": 575 + }, + { + "epoch": 0.5530484877580413, + "grad_norm": 1.1602600016134357, + "learning_rate": 9.996727953741881e-06, + "loss": 0.6791, + "step": 576 + }, + { + "epoch": 0.5540086413826212, + "grad_norm": 0.9931168407930752, + "learning_rate": 9.996605658582982e-06, + "loss": 0.7462, + "step": 577 + }, + { + "epoch": 0.5549687950072012, + "grad_norm": 1.4796154459879218, + "learning_rate": 9.996481120501214e-06, + "loss": 0.8238, + "step": 578 + }, + { + "epoch": 0.5559289486317811, + "grad_norm": 1.381992426279988, + "learning_rate": 9.996354339552482e-06, + "loss": 0.7427, + "step": 579 + }, + { + "epoch": 0.556889102256361, + "grad_norm": 1.000051907716779, + "learning_rate": 9.996225315793692e-06, + "loss": 0.6611, + "step": 580 + }, + { + "epoch": 0.5578492558809409, + "grad_norm": 1.8016378676915354, + "learning_rate": 9.996094049282768e-06, + "loss": 0.7868, + "step": 581 + }, + { + "epoch": 0.5588094095055209, + "grad_norm": 1.1278980971599617, + "learning_rate": 9.995960540078627e-06, + "loss": 0.7274, + "step": 582 + }, + { + "epoch": 0.5597695631301008, + "grad_norm": 1.7484961964443118, + "learning_rate": 9.995824788241206e-06, + "loss": 0.9602, + "step": 583 + }, + { + "epoch": 0.5607297167546808, + "grad_norm": 1.145929824155346, + "learning_rate": 9.995686793831438e-06, + "loss": 0.7076, + "step": 584 + }, + { + "epoch": 0.5616898703792607, + "grad_norm": 1.1004359314897234, + "learning_rate": 9.995546556911271e-06, + "loss": 0.681, + "step": 585 + }, + { + "epoch": 0.5626500240038406, + "grad_norm": 1.0748063299411916, + "learning_rate": 9.995404077543654e-06, + "loss": 0.7148, + "step": 586 + }, + { + "epoch": 0.5636101776284206, + "grad_norm": 0.821838615755815, + "learning_rate": 9.995259355792544e-06, + "loss": 0.6614, + "step": 587 + }, + { + "epoch": 0.5645703312530005, + "grad_norm": 1.2133270777619678, + "learning_rate": 9.995112391722905e-06, + "loss": 0.7483, + "step": 588 + }, + { + "epoch": 0.5655304848775804, + "grad_norm": 1.2328883410395755, + "learning_rate": 9.994963185400708e-06, + "loss": 0.8195, + "step": 589 + }, + { + "epoch": 0.5664906385021603, + "grad_norm": 1.0076087642529332, + "learning_rate": 9.994811736892931e-06, + "loss": 0.7355, + "step": 590 + }, + { + "epoch": 0.5674507921267403, + "grad_norm": 1.2919954902291888, + "learning_rate": 9.994658046267556e-06, + "loss": 0.833, + "step": 591 + }, + { + "epoch": 0.5684109457513202, + "grad_norm": 1.0294781538999418, + "learning_rate": 9.994502113593575e-06, + "loss": 0.8137, + "step": 592 + }, + { + "epoch": 0.5693710993759001, + "grad_norm": 1.2539612701534901, + "learning_rate": 9.994343938940983e-06, + "loss": 0.6718, + "step": 593 + }, + { + "epoch": 0.57033125300048, + "grad_norm": 1.0280669880358662, + "learning_rate": 9.994183522380784e-06, + "loss": 0.7441, + "step": 594 + }, + { + "epoch": 0.5712914066250601, + "grad_norm": 0.9019541853420437, + "learning_rate": 9.994020863984986e-06, + "loss": 0.6743, + "step": 595 + }, + { + "epoch": 0.57225156024964, + "grad_norm": 1.058143387523447, + "learning_rate": 9.993855963826605e-06, + "loss": 0.6711, + "step": 596 + }, + { + "epoch": 0.5732117138742199, + "grad_norm": 1.0307534340163422, + "learning_rate": 9.993688821979664e-06, + "loss": 0.6419, + "step": 597 + }, + { + "epoch": 0.5741718674987998, + "grad_norm": 1.0157479655127266, + "learning_rate": 9.993519438519191e-06, + "loss": 0.7381, + "step": 598 + }, + { + "epoch": 0.5751320211233797, + "grad_norm": 1.5360923531404511, + "learning_rate": 9.993347813521219e-06, + "loss": 0.8652, + "step": 599 + }, + { + "epoch": 0.5760921747479597, + "grad_norm": 0.8794359882129259, + "learning_rate": 9.993173947062789e-06, + "loss": 0.6855, + "step": 600 + }, + { + "epoch": 0.5770523283725396, + "grad_norm": 1.2191476435446476, + "learning_rate": 9.992997839221947e-06, + "loss": 0.7825, + "step": 601 + }, + { + "epoch": 0.5780124819971195, + "grad_norm": 1.0797985506718768, + "learning_rate": 9.99281949007775e-06, + "loss": 0.8174, + "step": 602 + }, + { + "epoch": 0.5789726356216994, + "grad_norm": 1.397720342441156, + "learning_rate": 9.992638899710252e-06, + "loss": 0.8272, + "step": 603 + }, + { + "epoch": 0.5799327892462794, + "grad_norm": 0.9830573117461331, + "learning_rate": 9.99245606820052e-06, + "loss": 0.6114, + "step": 604 + }, + { + "epoch": 0.5808929428708594, + "grad_norm": 1.0176480970254405, + "learning_rate": 9.992270995630628e-06, + "loss": 0.7211, + "step": 605 + }, + { + "epoch": 0.5818530964954393, + "grad_norm": 1.2533849123137955, + "learning_rate": 9.992083682083649e-06, + "loss": 0.8221, + "step": 606 + }, + { + "epoch": 0.5828132501200192, + "grad_norm": 1.004562598726377, + "learning_rate": 9.991894127643666e-06, + "loss": 0.6359, + "step": 607 + }, + { + "epoch": 0.5837734037445992, + "grad_norm": 1.3104279757416193, + "learning_rate": 9.991702332395771e-06, + "loss": 0.9004, + "step": 608 + }, + { + "epoch": 0.5847335573691791, + "grad_norm": 1.1907424868300012, + "learning_rate": 9.991508296426057e-06, + "loss": 0.639, + "step": 609 + }, + { + "epoch": 0.585693710993759, + "grad_norm": 1.3716403450449266, + "learning_rate": 9.991312019821626e-06, + "loss": 0.8258, + "step": 610 + }, + { + "epoch": 0.5866538646183389, + "grad_norm": 0.9895375215410548, + "learning_rate": 9.991113502670583e-06, + "loss": 0.6116, + "step": 611 + }, + { + "epoch": 0.5876140182429188, + "grad_norm": 0.9266557576594376, + "learning_rate": 9.99091274506204e-06, + "loss": 0.6317, + "step": 612 + }, + { + "epoch": 0.5885741718674988, + "grad_norm": 0.9225656836771144, + "learning_rate": 9.990709747086118e-06, + "loss": 0.6896, + "step": 613 + }, + { + "epoch": 0.5895343254920787, + "grad_norm": 0.9279076070689715, + "learning_rate": 9.990504508833936e-06, + "loss": 0.6593, + "step": 614 + }, + { + "epoch": 0.5904944791166586, + "grad_norm": 1.3615841081527988, + "learning_rate": 9.990297030397626e-06, + "loss": 0.9277, + "step": 615 + }, + { + "epoch": 0.5914546327412386, + "grad_norm": 1.0888009181578198, + "learning_rate": 9.990087311870322e-06, + "loss": 0.6918, + "step": 616 + }, + { + "epoch": 0.5924147863658186, + "grad_norm": 1.2019128759256583, + "learning_rate": 9.989875353346164e-06, + "loss": 0.7467, + "step": 617 + }, + { + "epoch": 0.5933749399903985, + "grad_norm": 1.1162187027509973, + "learning_rate": 9.9896611549203e-06, + "loss": 0.7803, + "step": 618 + }, + { + "epoch": 0.5943350936149784, + "grad_norm": 1.0995846179986166, + "learning_rate": 9.98944471668888e-06, + "loss": 0.6535, + "step": 619 + }, + { + "epoch": 0.5952952472395583, + "grad_norm": 1.2874007059819845, + "learning_rate": 9.98922603874906e-06, + "loss": 0.8015, + "step": 620 + }, + { + "epoch": 0.5962554008641383, + "grad_norm": 1.1709057648541077, + "learning_rate": 9.989005121199004e-06, + "loss": 0.8659, + "step": 621 + }, + { + "epoch": 0.5972155544887182, + "grad_norm": 1.1298017335305068, + "learning_rate": 9.98878196413788e-06, + "loss": 0.8084, + "step": 622 + }, + { + "epoch": 0.5981757081132981, + "grad_norm": 1.1273397470795112, + "learning_rate": 9.988556567665858e-06, + "loss": 0.6616, + "step": 623 + }, + { + "epoch": 0.599135861737878, + "grad_norm": 1.1287215761694498, + "learning_rate": 9.988328931884116e-06, + "loss": 0.8314, + "step": 624 + }, + { + "epoch": 0.600096015362458, + "grad_norm": 1.2200338008797798, + "learning_rate": 9.98809905689484e-06, + "loss": 0.8232, + "step": 625 + }, + { + "epoch": 0.601056168987038, + "grad_norm": 1.0124851725821313, + "learning_rate": 9.987866942801216e-06, + "loss": 0.6169, + "step": 626 + }, + { + "epoch": 0.6020163226116179, + "grad_norm": 0.8272113332478752, + "learning_rate": 9.987632589707441e-06, + "loss": 0.6627, + "step": 627 + }, + { + "epoch": 0.6029764762361978, + "grad_norm": 0.9571583436346156, + "learning_rate": 9.98739599771871e-06, + "loss": 0.7432, + "step": 628 + }, + { + "epoch": 0.6039366298607777, + "grad_norm": 1.3672229056284988, + "learning_rate": 9.987157166941228e-06, + "loss": 0.8827, + "step": 629 + }, + { + "epoch": 0.6048967834853577, + "grad_norm": 1.174360049598899, + "learning_rate": 9.986916097482204e-06, + "loss": 0.7367, + "step": 630 + }, + { + "epoch": 0.6058569371099376, + "grad_norm": 0.8517725810532072, + "learning_rate": 9.98667278944985e-06, + "loss": 0.5979, + "step": 631 + }, + { + "epoch": 0.6068170907345175, + "grad_norm": 1.1119512062132852, + "learning_rate": 9.986427242953387e-06, + "loss": 0.7079, + "step": 632 + }, + { + "epoch": 0.6077772443590974, + "grad_norm": 1.6624527590685678, + "learning_rate": 9.986179458103036e-06, + "loss": 0.8425, + "step": 633 + }, + { + "epoch": 0.6087373979836774, + "grad_norm": 0.8868536609149565, + "learning_rate": 9.985929435010024e-06, + "loss": 0.6338, + "step": 634 + }, + { + "epoch": 0.6096975516082573, + "grad_norm": 1.062767851898318, + "learning_rate": 9.985677173786587e-06, + "loss": 0.6638, + "step": 635 + }, + { + "epoch": 0.6106577052328372, + "grad_norm": 1.130431234420582, + "learning_rate": 9.98542267454596e-06, + "loss": 0.7227, + "step": 636 + }, + { + "epoch": 0.6116178588574172, + "grad_norm": 1.0969498581452322, + "learning_rate": 9.985165937402386e-06, + "loss": 0.7648, + "step": 637 + }, + { + "epoch": 0.6125780124819972, + "grad_norm": 1.3792237075028597, + "learning_rate": 9.98490696247111e-06, + "loss": 0.8187, + "step": 638 + }, + { + "epoch": 0.6135381661065771, + "grad_norm": 1.60062571530598, + "learning_rate": 9.984645749868386e-06, + "loss": 0.7237, + "step": 639 + }, + { + "epoch": 0.614498319731157, + "grad_norm": 1.0548834013904989, + "learning_rate": 9.984382299711466e-06, + "loss": 0.7315, + "step": 640 + }, + { + "epoch": 0.6154584733557369, + "grad_norm": 1.6070075540077156, + "learning_rate": 9.984116612118612e-06, + "loss": 0.793, + "step": 641 + }, + { + "epoch": 0.6164186269803168, + "grad_norm": 1.027726907615115, + "learning_rate": 9.98384868720909e-06, + "loss": 0.7115, + "step": 642 + }, + { + "epoch": 0.6173787806048968, + "grad_norm": 0.9533472183683033, + "learning_rate": 9.983578525103166e-06, + "loss": 0.7205, + "step": 643 + }, + { + "epoch": 0.6183389342294767, + "grad_norm": 1.278072786735959, + "learning_rate": 9.983306125922111e-06, + "loss": 0.7248, + "step": 644 + }, + { + "epoch": 0.6192990878540566, + "grad_norm": 1.30698682005401, + "learning_rate": 9.983031489788208e-06, + "loss": 0.7187, + "step": 645 + }, + { + "epoch": 0.6202592414786365, + "grad_norm": 1.1294142037988335, + "learning_rate": 9.982754616824733e-06, + "loss": 0.7404, + "step": 646 + }, + { + "epoch": 0.6212193951032166, + "grad_norm": 1.1339707486746415, + "learning_rate": 9.982475507155974e-06, + "loss": 0.7069, + "step": 647 + }, + { + "epoch": 0.6221795487277965, + "grad_norm": 1.045473010633134, + "learning_rate": 9.98219416090722e-06, + "loss": 0.6399, + "step": 648 + }, + { + "epoch": 0.6231397023523764, + "grad_norm": 1.1491372204255086, + "learning_rate": 9.981910578204764e-06, + "loss": 0.6836, + "step": 649 + }, + { + "epoch": 0.6240998559769563, + "grad_norm": 1.3019640496272942, + "learning_rate": 9.981624759175902e-06, + "loss": 0.7818, + "step": 650 + }, + { + "epoch": 0.6250600096015363, + "grad_norm": 0.9641995848565628, + "learning_rate": 9.981336703948936e-06, + "loss": 0.7384, + "step": 651 + }, + { + "epoch": 0.6260201632261162, + "grad_norm": 1.0424494813372942, + "learning_rate": 9.98104641265317e-06, + "loss": 0.7581, + "step": 652 + }, + { + "epoch": 0.6269803168506961, + "grad_norm": 1.1749410105746338, + "learning_rate": 9.980753885418916e-06, + "loss": 0.8092, + "step": 653 + }, + { + "epoch": 0.627940470475276, + "grad_norm": 1.0608257188772605, + "learning_rate": 9.980459122377484e-06, + "loss": 0.7052, + "step": 654 + }, + { + "epoch": 0.628900624099856, + "grad_norm": 0.9851354208777588, + "learning_rate": 9.98016212366119e-06, + "loss": 0.7334, + "step": 655 + }, + { + "epoch": 0.6298607777244359, + "grad_norm": 1.1309322468621306, + "learning_rate": 9.979862889403353e-06, + "loss": 0.7993, + "step": 656 + }, + { + "epoch": 0.6308209313490158, + "grad_norm": 1.4423042663996517, + "learning_rate": 9.979561419738297e-06, + "loss": 0.8021, + "step": 657 + }, + { + "epoch": 0.6317810849735958, + "grad_norm": 1.1726897326686694, + "learning_rate": 9.979257714801348e-06, + "loss": 0.7577, + "step": 658 + }, + { + "epoch": 0.6327412385981757, + "grad_norm": 0.9727137981998019, + "learning_rate": 9.978951774728837e-06, + "loss": 0.728, + "step": 659 + }, + { + "epoch": 0.6337013922227557, + "grad_norm": 0.9433976828393242, + "learning_rate": 9.978643599658095e-06, + "loss": 0.705, + "step": 660 + }, + { + "epoch": 0.6346615458473356, + "grad_norm": 0.958260167898449, + "learning_rate": 9.97833318972746e-06, + "loss": 0.7253, + "step": 661 + }, + { + "epoch": 0.6356216994719155, + "grad_norm": 1.1728067197573666, + "learning_rate": 9.978020545076274e-06, + "loss": 0.7876, + "step": 662 + }, + { + "epoch": 0.6365818530964954, + "grad_norm": 1.0887980881688628, + "learning_rate": 9.977705665844875e-06, + "loss": 0.7163, + "step": 663 + }, + { + "epoch": 0.6375420067210754, + "grad_norm": 1.1071907542436588, + "learning_rate": 9.977388552174611e-06, + "loss": 0.7235, + "step": 664 + }, + { + "epoch": 0.6385021603456553, + "grad_norm": 1.1929947896189024, + "learning_rate": 9.977069204207832e-06, + "loss": 0.7432, + "step": 665 + }, + { + "epoch": 0.6394623139702352, + "grad_norm": 1.6705352459935405, + "learning_rate": 9.97674762208789e-06, + "loss": 0.7442, + "step": 666 + }, + { + "epoch": 0.6404224675948151, + "grad_norm": 0.9169215619256575, + "learning_rate": 9.976423805959138e-06, + "loss": 0.74, + "step": 667 + }, + { + "epoch": 0.6413826212193952, + "grad_norm": 1.6167894713612192, + "learning_rate": 9.976097755966935e-06, + "loss": 0.7734, + "step": 668 + }, + { + "epoch": 0.6423427748439751, + "grad_norm": 0.8516747743106268, + "learning_rate": 9.975769472257642e-06, + "loss": 0.6725, + "step": 669 + }, + { + "epoch": 0.643302928468555, + "grad_norm": 1.1960384862782092, + "learning_rate": 9.97543895497862e-06, + "loss": 0.7296, + "step": 670 + }, + { + "epoch": 0.6442630820931349, + "grad_norm": 1.151368014455306, + "learning_rate": 9.975106204278235e-06, + "loss": 0.7847, + "step": 671 + }, + { + "epoch": 0.6452232357177148, + "grad_norm": 1.4878206628916213, + "learning_rate": 9.974771220305855e-06, + "loss": 0.7559, + "step": 672 + }, + { + "epoch": 0.6461833893422948, + "grad_norm": 1.2396185589531643, + "learning_rate": 9.974434003211853e-06, + "loss": 0.7136, + "step": 673 + }, + { + "epoch": 0.6471435429668747, + "grad_norm": 1.1392041183252792, + "learning_rate": 9.974094553147599e-06, + "loss": 0.7088, + "step": 674 + }, + { + "epoch": 0.6481036965914546, + "grad_norm": 1.0042368852303203, + "learning_rate": 9.973752870265473e-06, + "loss": 0.7617, + "step": 675 + }, + { + "epoch": 0.6490638502160345, + "grad_norm": 1.0109453482022366, + "learning_rate": 9.973408954718848e-06, + "loss": 0.6427, + "step": 676 + }, + { + "epoch": 0.6500240038406145, + "grad_norm": 1.4767642145792161, + "learning_rate": 9.973062806662106e-06, + "loss": 0.7787, + "step": 677 + }, + { + "epoch": 0.6509841574651944, + "grad_norm": 1.149817371556, + "learning_rate": 9.97271442625063e-06, + "loss": 0.7017, + "step": 678 + }, + { + "epoch": 0.6519443110897744, + "grad_norm": 1.0927365492730468, + "learning_rate": 9.972363813640804e-06, + "loss": 0.6969, + "step": 679 + }, + { + "epoch": 0.6529044647143543, + "grad_norm": 1.1557208626189814, + "learning_rate": 9.972010968990012e-06, + "loss": 0.777, + "step": 680 + }, + { + "epoch": 0.6538646183389343, + "grad_norm": 1.0242126820972108, + "learning_rate": 9.971655892456646e-06, + "loss": 0.5974, + "step": 681 + }, + { + "epoch": 0.6548247719635142, + "grad_norm": 1.4722262917903646, + "learning_rate": 9.971298584200093e-06, + "loss": 0.7496, + "step": 682 + }, + { + "epoch": 0.6557849255880941, + "grad_norm": 1.228142897886962, + "learning_rate": 9.970939044380743e-06, + "loss": 0.7637, + "step": 683 + }, + { + "epoch": 0.656745079212674, + "grad_norm": 1.0246575470997343, + "learning_rate": 9.970577273159995e-06, + "loss": 0.73, + "step": 684 + }, + { + "epoch": 0.6577052328372539, + "grad_norm": 1.1064980469670687, + "learning_rate": 9.970213270700241e-06, + "loss": 0.7003, + "step": 685 + }, + { + "epoch": 0.6586653864618339, + "grad_norm": 0.9485775501229478, + "learning_rate": 9.969847037164881e-06, + "loss": 0.7478, + "step": 686 + }, + { + "epoch": 0.6596255400864138, + "grad_norm": 1.0209641451595586, + "learning_rate": 9.969478572718307e-06, + "loss": 0.7107, + "step": 687 + }, + { + "epoch": 0.6605856937109937, + "grad_norm": 1.2777317659195842, + "learning_rate": 9.969107877525925e-06, + "loss": 0.6346, + "step": 688 + }, + { + "epoch": 0.6615458473355736, + "grad_norm": 1.1304395876489817, + "learning_rate": 9.968734951754134e-06, + "loss": 0.716, + "step": 689 + }, + { + "epoch": 0.6625060009601537, + "grad_norm": 1.0589643707841647, + "learning_rate": 9.968359795570333e-06, + "loss": 0.7335, + "step": 690 + }, + { + "epoch": 0.6634661545847336, + "grad_norm": 1.312163556334331, + "learning_rate": 9.96798240914293e-06, + "loss": 0.7644, + "step": 691 + }, + { + "epoch": 0.6644263082093135, + "grad_norm": 1.4654703740824706, + "learning_rate": 9.96760279264133e-06, + "loss": 0.7075, + "step": 692 + }, + { + "epoch": 0.6653864618338934, + "grad_norm": 1.2308530665209227, + "learning_rate": 9.967220946235933e-06, + "loss": 0.6366, + "step": 693 + }, + { + "epoch": 0.6663466154584734, + "grad_norm": 1.4689991042569908, + "learning_rate": 9.966836870098153e-06, + "loss": 0.8619, + "step": 694 + }, + { + "epoch": 0.6673067690830533, + "grad_norm": 1.1544376204402063, + "learning_rate": 9.966450564400396e-06, + "loss": 0.5347, + "step": 695 + }, + { + "epoch": 0.6682669227076332, + "grad_norm": 1.018684998631377, + "learning_rate": 9.966062029316067e-06, + "loss": 0.7297, + "step": 696 + }, + { + "epoch": 0.6692270763322131, + "grad_norm": 1.2731963424973747, + "learning_rate": 9.965671265019576e-06, + "loss": 0.7425, + "step": 697 + }, + { + "epoch": 0.6701872299567931, + "grad_norm": 1.2094633115207627, + "learning_rate": 9.965278271686335e-06, + "loss": 0.834, + "step": 698 + }, + { + "epoch": 0.671147383581373, + "grad_norm": 1.298156294985173, + "learning_rate": 9.964883049492755e-06, + "loss": 0.7799, + "step": 699 + }, + { + "epoch": 0.672107537205953, + "grad_norm": 0.932403307235914, + "learning_rate": 9.964485598616244e-06, + "loss": 0.6305, + "step": 700 + }, + { + "epoch": 0.6730676908305329, + "grad_norm": 1.348840185082681, + "learning_rate": 9.964085919235217e-06, + "loss": 0.7148, + "step": 701 + }, + { + "epoch": 0.6740278444551128, + "grad_norm": 1.0730348632411175, + "learning_rate": 9.963684011529084e-06, + "loss": 0.7093, + "step": 702 + }, + { + "epoch": 0.6749879980796928, + "grad_norm": 1.7379797117287121, + "learning_rate": 9.963279875678256e-06, + "loss": 0.7236, + "step": 703 + }, + { + "epoch": 0.6759481517042727, + "grad_norm": 1.0101795335218668, + "learning_rate": 9.962873511864146e-06, + "loss": 0.7541, + "step": 704 + }, + { + "epoch": 0.6769083053288526, + "grad_norm": 1.095236103584098, + "learning_rate": 9.962464920269167e-06, + "loss": 0.6931, + "step": 705 + }, + { + "epoch": 0.6778684589534325, + "grad_norm": 0.9888949939493659, + "learning_rate": 9.962054101076733e-06, + "loss": 0.7213, + "step": 706 + }, + { + "epoch": 0.6788286125780125, + "grad_norm": 1.1704736246244622, + "learning_rate": 9.961641054471252e-06, + "loss": 0.7301, + "step": 707 + }, + { + "epoch": 0.6797887662025924, + "grad_norm": 1.1322459449428894, + "learning_rate": 9.96122578063814e-06, + "loss": 0.7355, + "step": 708 + }, + { + "epoch": 0.6807489198271723, + "grad_norm": 1.3295227832959442, + "learning_rate": 9.960808279763807e-06, + "loss": 0.7722, + "step": 709 + }, + { + "epoch": 0.6817090734517522, + "grad_norm": 1.2022742602302552, + "learning_rate": 9.960388552035666e-06, + "loss": 0.6774, + "step": 710 + }, + { + "epoch": 0.6826692270763323, + "grad_norm": 1.088527690096997, + "learning_rate": 9.959966597642127e-06, + "loss": 0.7082, + "step": 711 + }, + { + "epoch": 0.6836293807009122, + "grad_norm": 1.4359245485883494, + "learning_rate": 9.9595424167726e-06, + "loss": 0.7906, + "step": 712 + }, + { + "epoch": 0.6845895343254921, + "grad_norm": 1.0712636033705696, + "learning_rate": 9.9591160096175e-06, + "loss": 0.6228, + "step": 713 + }, + { + "epoch": 0.685549687950072, + "grad_norm": 1.1286122698085486, + "learning_rate": 9.95868737636823e-06, + "loss": 0.6131, + "step": 714 + }, + { + "epoch": 0.6865098415746519, + "grad_norm": 1.754001739111913, + "learning_rate": 9.958256517217204e-06, + "loss": 0.8415, + "step": 715 + }, + { + "epoch": 0.6874699951992319, + "grad_norm": 1.2415521559789686, + "learning_rate": 9.95782343235783e-06, + "loss": 0.6967, + "step": 716 + }, + { + "epoch": 0.6884301488238118, + "grad_norm": 0.9746405707963949, + "learning_rate": 9.95738812198451e-06, + "loss": 0.6819, + "step": 717 + }, + { + "epoch": 0.6893903024483917, + "grad_norm": 1.1467656093166074, + "learning_rate": 9.956950586292655e-06, + "loss": 0.7994, + "step": 718 + }, + { + "epoch": 0.6903504560729716, + "grad_norm": 1.1143672934666407, + "learning_rate": 9.956510825478669e-06, + "loss": 0.6832, + "step": 719 + }, + { + "epoch": 0.6913106096975516, + "grad_norm": 1.3721570448044658, + "learning_rate": 9.956068839739955e-06, + "loss": 0.8298, + "step": 720 + }, + { + "epoch": 0.6922707633221316, + "grad_norm": 1.6265371445701913, + "learning_rate": 9.955624629274916e-06, + "loss": 0.9974, + "step": 721 + }, + { + "epoch": 0.6932309169467115, + "grad_norm": 1.4662804140862407, + "learning_rate": 9.955178194282953e-06, + "loss": 0.7463, + "step": 722 + }, + { + "epoch": 0.6941910705712914, + "grad_norm": 0.8255635339080193, + "learning_rate": 9.954729534964469e-06, + "loss": 0.5698, + "step": 723 + }, + { + "epoch": 0.6951512241958714, + "grad_norm": 1.3585911943208118, + "learning_rate": 9.954278651520859e-06, + "loss": 0.7732, + "step": 724 + }, + { + "epoch": 0.6961113778204513, + "grad_norm": 1.060688981132975, + "learning_rate": 9.953825544154519e-06, + "loss": 0.7416, + "step": 725 + }, + { + "epoch": 0.6970715314450312, + "grad_norm": 1.2736780912673817, + "learning_rate": 9.953370213068847e-06, + "loss": 0.6838, + "step": 726 + }, + { + "epoch": 0.6980316850696111, + "grad_norm": 1.3360461372150172, + "learning_rate": 9.952912658468235e-06, + "loss": 0.7093, + "step": 727 + }, + { + "epoch": 0.6989918386941911, + "grad_norm": 1.0287316815152034, + "learning_rate": 9.952452880558073e-06, + "loss": 0.8231, + "step": 728 + }, + { + "epoch": 0.699951992318771, + "grad_norm": 1.1300889405704038, + "learning_rate": 9.951990879544753e-06, + "loss": 0.729, + "step": 729 + }, + { + "epoch": 0.7009121459433509, + "grad_norm": 0.9505504150767853, + "learning_rate": 9.95152665563566e-06, + "loss": 0.7065, + "step": 730 + }, + { + "epoch": 0.7018722995679308, + "grad_norm": 1.038257585330584, + "learning_rate": 9.951060209039182e-06, + "loss": 0.7397, + "step": 731 + }, + { + "epoch": 0.7028324531925108, + "grad_norm": 0.8443922390769166, + "learning_rate": 9.950591539964698e-06, + "loss": 0.706, + "step": 732 + }, + { + "epoch": 0.7037926068170908, + "grad_norm": 1.2080688483001045, + "learning_rate": 9.950120648622592e-06, + "loss": 0.7449, + "step": 733 + }, + { + "epoch": 0.7047527604416707, + "grad_norm": 0.9072170255903059, + "learning_rate": 9.94964753522424e-06, + "loss": 0.7127, + "step": 734 + }, + { + "epoch": 0.7057129140662506, + "grad_norm": 1.2648229256791117, + "learning_rate": 9.949172199982019e-06, + "loss": 0.7527, + "step": 735 + }, + { + "epoch": 0.7066730676908305, + "grad_norm": 1.3483477493599871, + "learning_rate": 9.9486946431093e-06, + "loss": 0.955, + "step": 736 + }, + { + "epoch": 0.7076332213154105, + "grad_norm": 1.1352257922072784, + "learning_rate": 9.948214864820455e-06, + "loss": 0.7398, + "step": 737 + }, + { + "epoch": 0.7085933749399904, + "grad_norm": 1.3783392370323646, + "learning_rate": 9.94773286533085e-06, + "loss": 0.8405, + "step": 738 + }, + { + "epoch": 0.7095535285645703, + "grad_norm": 1.0389249982585742, + "learning_rate": 9.947248644856852e-06, + "loss": 0.6672, + "step": 739 + }, + { + "epoch": 0.7105136821891502, + "grad_norm": 1.3151653766211417, + "learning_rate": 9.946762203615818e-06, + "loss": 0.764, + "step": 740 + }, + { + "epoch": 0.7114738358137302, + "grad_norm": 0.9917853380663585, + "learning_rate": 9.94627354182611e-06, + "loss": 0.7454, + "step": 741 + }, + { + "epoch": 0.7124339894383102, + "grad_norm": 1.1532058497358957, + "learning_rate": 9.945782659707082e-06, + "loss": 0.6812, + "step": 742 + }, + { + "epoch": 0.7133941430628901, + "grad_norm": 1.1572527961716201, + "learning_rate": 9.945289557479086e-06, + "loss": 0.7281, + "step": 743 + }, + { + "epoch": 0.71435429668747, + "grad_norm": 1.176236714703303, + "learning_rate": 9.944794235363468e-06, + "loss": 0.6686, + "step": 744 + }, + { + "epoch": 0.7153144503120499, + "grad_norm": 1.004038515625951, + "learning_rate": 9.944296693582576e-06, + "loss": 0.7013, + "step": 745 + }, + { + "epoch": 0.7162746039366299, + "grad_norm": 1.43231333048898, + "learning_rate": 9.94379693235975e-06, + "loss": 0.6745, + "step": 746 + }, + { + "epoch": 0.7172347575612098, + "grad_norm": 0.8278900895756431, + "learning_rate": 9.943294951919327e-06, + "loss": 0.5708, + "step": 747 + }, + { + "epoch": 0.7181949111857897, + "grad_norm": 1.4659914938897542, + "learning_rate": 9.94279075248664e-06, + "loss": 0.8242, + "step": 748 + }, + { + "epoch": 0.7191550648103696, + "grad_norm": 1.4706142820300785, + "learning_rate": 9.942284334288021e-06, + "loss": 0.8457, + "step": 749 + }, + { + "epoch": 0.7201152184349496, + "grad_norm": 1.1984466642743257, + "learning_rate": 9.941775697550796e-06, + "loss": 0.7123, + "step": 750 + }, + { + "epoch": 0.7210753720595295, + "grad_norm": 1.1050073348933755, + "learning_rate": 9.941264842503282e-06, + "loss": 0.8425, + "step": 751 + }, + { + "epoch": 0.7220355256841094, + "grad_norm": 0.8687768913712395, + "learning_rate": 9.9407517693748e-06, + "loss": 0.6265, + "step": 752 + }, + { + "epoch": 0.7229956793086894, + "grad_norm": 0.9769568173569657, + "learning_rate": 9.940236478395662e-06, + "loss": 0.724, + "step": 753 + }, + { + "epoch": 0.7239558329332694, + "grad_norm": 1.6207874366444537, + "learning_rate": 9.939718969797178e-06, + "loss": 0.8284, + "step": 754 + }, + { + "epoch": 0.7249159865578493, + "grad_norm": 0.9805328249455393, + "learning_rate": 9.939199243811653e-06, + "loss": 0.6822, + "step": 755 + }, + { + "epoch": 0.7258761401824292, + "grad_norm": 1.246024930996379, + "learning_rate": 9.938677300672382e-06, + "loss": 0.7775, + "step": 756 + }, + { + "epoch": 0.7268362938070091, + "grad_norm": 1.198028879321762, + "learning_rate": 9.938153140613662e-06, + "loss": 0.7028, + "step": 757 + }, + { + "epoch": 0.727796447431589, + "grad_norm": 0.9973811279361954, + "learning_rate": 9.937626763870784e-06, + "loss": 0.6884, + "step": 758 + }, + { + "epoch": 0.728756601056169, + "grad_norm": 1.03169869259323, + "learning_rate": 9.937098170680034e-06, + "loss": 0.6303, + "step": 759 + }, + { + "epoch": 0.7297167546807489, + "grad_norm": 0.9247880187978105, + "learning_rate": 9.936567361278687e-06, + "loss": 0.6576, + "step": 760 + }, + { + "epoch": 0.7306769083053288, + "grad_norm": 0.8417058658300991, + "learning_rate": 9.93603433590502e-06, + "loss": 0.6361, + "step": 761 + }, + { + "epoch": 0.7316370619299087, + "grad_norm": 1.03716134734401, + "learning_rate": 9.935499094798305e-06, + "loss": 0.7885, + "step": 762 + }, + { + "epoch": 0.7325972155544888, + "grad_norm": 1.2074321143830193, + "learning_rate": 9.934961638198804e-06, + "loss": 0.6944, + "step": 763 + }, + { + "epoch": 0.7335573691790687, + "grad_norm": 1.1848604524508257, + "learning_rate": 9.934421966347773e-06, + "loss": 0.8124, + "step": 764 + }, + { + "epoch": 0.7345175228036486, + "grad_norm": 1.0429348427785348, + "learning_rate": 9.93388007948747e-06, + "loss": 0.7377, + "step": 765 + }, + { + "epoch": 0.7354776764282285, + "grad_norm": 0.8888707066106554, + "learning_rate": 9.933335977861138e-06, + "loss": 0.6764, + "step": 766 + }, + { + "epoch": 0.7364378300528085, + "grad_norm": 1.0626278519468342, + "learning_rate": 9.93278966171302e-06, + "loss": 0.6276, + "step": 767 + }, + { + "epoch": 0.7373979836773884, + "grad_norm": 0.8962912358229971, + "learning_rate": 9.932241131288352e-06, + "loss": 0.6641, + "step": 768 + }, + { + "epoch": 0.7383581373019683, + "grad_norm": 0.8941146837643744, + "learning_rate": 9.931690386833363e-06, + "loss": 0.7262, + "step": 769 + }, + { + "epoch": 0.7393182909265482, + "grad_norm": 1.0424749755401177, + "learning_rate": 9.931137428595276e-06, + "loss": 0.7582, + "step": 770 + }, + { + "epoch": 0.7402784445511282, + "grad_norm": 1.5239003387542835, + "learning_rate": 9.930582256822307e-06, + "loss": 0.7967, + "step": 771 + }, + { + "epoch": 0.7412385981757081, + "grad_norm": 0.9789285534367823, + "learning_rate": 9.93002487176367e-06, + "loss": 0.7711, + "step": 772 + }, + { + "epoch": 0.742198751800288, + "grad_norm": 0.9798122544351332, + "learning_rate": 9.929465273669567e-06, + "loss": 0.7138, + "step": 773 + }, + { + "epoch": 0.743158905424868, + "grad_norm": 1.0089252998813754, + "learning_rate": 9.928903462791195e-06, + "loss": 0.7679, + "step": 774 + }, + { + "epoch": 0.7441190590494479, + "grad_norm": 1.129686771537608, + "learning_rate": 9.928339439380745e-06, + "loss": 0.7867, + "step": 775 + }, + { + "epoch": 0.7450792126740279, + "grad_norm": 1.0680489943573477, + "learning_rate": 9.927773203691403e-06, + "loss": 0.7233, + "step": 776 + }, + { + "epoch": 0.7460393662986078, + "grad_norm": 1.1991253915669609, + "learning_rate": 9.927204755977344e-06, + "loss": 0.8214, + "step": 777 + }, + { + "epoch": 0.7469995199231877, + "grad_norm": 1.1810464281071666, + "learning_rate": 9.926634096493738e-06, + "loss": 0.7266, + "step": 778 + }, + { + "epoch": 0.7479596735477676, + "grad_norm": 1.1918237303571164, + "learning_rate": 9.92606122549675e-06, + "loss": 0.7091, + "step": 779 + }, + { + "epoch": 0.7489198271723476, + "grad_norm": 1.4809878657655684, + "learning_rate": 9.925486143243534e-06, + "loss": 0.8235, + "step": 780 + }, + { + "epoch": 0.7498799807969275, + "grad_norm": 1.0623502206306763, + "learning_rate": 9.92490884999224e-06, + "loss": 0.6431, + "step": 781 + }, + { + "epoch": 0.7508401344215074, + "grad_norm": 1.1728978803988328, + "learning_rate": 9.924329346002007e-06, + "loss": 0.7072, + "step": 782 + }, + { + "epoch": 0.7518002880460873, + "grad_norm": 1.2661200193104183, + "learning_rate": 9.923747631532968e-06, + "loss": 0.7299, + "step": 783 + }, + { + "epoch": 0.7527604416706674, + "grad_norm": 1.369779764076495, + "learning_rate": 9.92316370684625e-06, + "loss": 0.6906, + "step": 784 + }, + { + "epoch": 0.7537205952952473, + "grad_norm": 0.9954497341903171, + "learning_rate": 9.922577572203969e-06, + "loss": 0.6794, + "step": 785 + }, + { + "epoch": 0.7546807489198272, + "grad_norm": 1.4017643770239192, + "learning_rate": 9.921989227869236e-06, + "loss": 0.8379, + "step": 786 + }, + { + "epoch": 0.7556409025444071, + "grad_norm": 1.3015683095250705, + "learning_rate": 9.921398674106152e-06, + "loss": 0.826, + "step": 787 + }, + { + "epoch": 0.756601056168987, + "grad_norm": 1.2155557789621108, + "learning_rate": 9.92080591117981e-06, + "loss": 0.8202, + "step": 788 + }, + { + "epoch": 0.757561209793567, + "grad_norm": 1.0881907984132828, + "learning_rate": 9.920210939356294e-06, + "loss": 0.7415, + "step": 789 + }, + { + "epoch": 0.7585213634181469, + "grad_norm": 1.2935427695919515, + "learning_rate": 9.919613758902682e-06, + "loss": 0.7854, + "step": 790 + }, + { + "epoch": 0.7594815170427268, + "grad_norm": 1.2385800002831058, + "learning_rate": 9.91901437008704e-06, + "loss": 0.7114, + "step": 791 + }, + { + "epoch": 0.7604416706673067, + "grad_norm": 1.001176807057794, + "learning_rate": 9.918412773178431e-06, + "loss": 0.7033, + "step": 792 + }, + { + "epoch": 0.7614018242918867, + "grad_norm": 1.1181435549505114, + "learning_rate": 9.917808968446903e-06, + "loss": 0.7477, + "step": 793 + }, + { + "epoch": 0.7623619779164666, + "grad_norm": 1.265201274814283, + "learning_rate": 9.917202956163496e-06, + "loss": 0.7502, + "step": 794 + }, + { + "epoch": 0.7633221315410466, + "grad_norm": 1.0814941759105932, + "learning_rate": 9.916594736600244e-06, + "loss": 0.7752, + "step": 795 + }, + { + "epoch": 0.7642822851656265, + "grad_norm": 1.015411031942487, + "learning_rate": 9.915984310030172e-06, + "loss": 0.8252, + "step": 796 + }, + { + "epoch": 0.7652424387902065, + "grad_norm": 1.0752951651724534, + "learning_rate": 9.91537167672729e-06, + "loss": 0.6898, + "step": 797 + }, + { + "epoch": 0.7662025924147864, + "grad_norm": 0.8978312152634739, + "learning_rate": 9.914756836966604e-06, + "loss": 0.6568, + "step": 798 + }, + { + "epoch": 0.7671627460393663, + "grad_norm": 1.4174184006325772, + "learning_rate": 9.914139791024112e-06, + "loss": 0.7441, + "step": 799 + }, + { + "epoch": 0.7681228996639462, + "grad_norm": 1.0309137281263778, + "learning_rate": 9.913520539176797e-06, + "loss": 0.8065, + "step": 800 + }, + { + "epoch": 0.7690830532885262, + "grad_norm": 1.1056726761800266, + "learning_rate": 9.912899081702633e-06, + "loss": 0.7709, + "step": 801 + }, + { + "epoch": 0.7700432069131061, + "grad_norm": 1.1237710277127422, + "learning_rate": 9.912275418880588e-06, + "loss": 0.7291, + "step": 802 + }, + { + "epoch": 0.771003360537686, + "grad_norm": 0.8081645629455623, + "learning_rate": 9.911649550990618e-06, + "loss": 0.6145, + "step": 803 + }, + { + "epoch": 0.7719635141622659, + "grad_norm": 0.8747329876189975, + "learning_rate": 9.911021478313665e-06, + "loss": 0.6091, + "step": 804 + }, + { + "epoch": 0.7729236677868458, + "grad_norm": 0.8873911870516857, + "learning_rate": 9.910391201131668e-06, + "loss": 0.6622, + "step": 805 + }, + { + "epoch": 0.7738838214114259, + "grad_norm": 0.9666313834432418, + "learning_rate": 9.909758719727549e-06, + "loss": 0.5906, + "step": 806 + }, + { + "epoch": 0.7748439750360058, + "grad_norm": 0.9382478033157138, + "learning_rate": 9.909124034385225e-06, + "loss": 0.6475, + "step": 807 + }, + { + "epoch": 0.7758041286605857, + "grad_norm": 1.4249896826413047, + "learning_rate": 9.908487145389595e-06, + "loss": 0.9343, + "step": 808 + }, + { + "epoch": 0.7767642822851656, + "grad_norm": 1.1015378414410202, + "learning_rate": 9.907848053026554e-06, + "loss": 0.6708, + "step": 809 + }, + { + "epoch": 0.7777244359097456, + "grad_norm": 0.8335101639979812, + "learning_rate": 9.907206757582987e-06, + "loss": 0.5111, + "step": 810 + }, + { + "epoch": 0.7786845895343255, + "grad_norm": 1.0698637825517112, + "learning_rate": 9.90656325934676e-06, + "loss": 0.7194, + "step": 811 + }, + { + "epoch": 0.7796447431589054, + "grad_norm": 1.3467599291721108, + "learning_rate": 9.905917558606734e-06, + "loss": 0.8593, + "step": 812 + }, + { + "epoch": 0.7806048967834853, + "grad_norm": 0.9265083094260931, + "learning_rate": 9.905269655652758e-06, + "loss": 0.691, + "step": 813 + }, + { + "epoch": 0.7815650504080653, + "grad_norm": 0.9569516294394528, + "learning_rate": 9.904619550775668e-06, + "loss": 0.6904, + "step": 814 + }, + { + "epoch": 0.7825252040326452, + "grad_norm": 1.2033247575360706, + "learning_rate": 9.903967244267287e-06, + "loss": 0.7906, + "step": 815 + }, + { + "epoch": 0.7834853576572252, + "grad_norm": 1.2045148381420874, + "learning_rate": 9.903312736420433e-06, + "loss": 0.7269, + "step": 816 + }, + { + "epoch": 0.7844455112818051, + "grad_norm": 0.9815529060942425, + "learning_rate": 9.902656027528903e-06, + "loss": 0.7521, + "step": 817 + }, + { + "epoch": 0.785405664906385, + "grad_norm": 1.0169730117521143, + "learning_rate": 9.90199711788749e-06, + "loss": 0.6917, + "step": 818 + }, + { + "epoch": 0.786365818530965, + "grad_norm": 1.279845677472938, + "learning_rate": 9.901336007791971e-06, + "loss": 0.8229, + "step": 819 + }, + { + "epoch": 0.7873259721555449, + "grad_norm": 1.1504766635141583, + "learning_rate": 9.900672697539108e-06, + "loss": 0.7362, + "step": 820 + }, + { + "epoch": 0.7882861257801248, + "grad_norm": 1.0363323444139323, + "learning_rate": 9.900007187426659e-06, + "loss": 0.6574, + "step": 821 + }, + { + "epoch": 0.7892462794047047, + "grad_norm": 1.0099042824557933, + "learning_rate": 9.89933947775336e-06, + "loss": 0.6754, + "step": 822 + }, + { + "epoch": 0.7902064330292847, + "grad_norm": 0.9730742596836239, + "learning_rate": 9.898669568818942e-06, + "loss": 0.6871, + "step": 823 + }, + { + "epoch": 0.7911665866538646, + "grad_norm": 0.7856584341577615, + "learning_rate": 9.897997460924118e-06, + "loss": 0.5745, + "step": 824 + }, + { + "epoch": 0.7921267402784445, + "grad_norm": 1.2340542727190076, + "learning_rate": 9.89732315437059e-06, + "loss": 0.7881, + "step": 825 + }, + { + "epoch": 0.7930868939030244, + "grad_norm": 1.0505728583822727, + "learning_rate": 9.896646649461047e-06, + "loss": 0.7125, + "step": 826 + }, + { + "epoch": 0.7940470475276045, + "grad_norm": 0.9100796308195849, + "learning_rate": 9.895967946499165e-06, + "loss": 0.6839, + "step": 827 + }, + { + "epoch": 0.7950072011521844, + "grad_norm": 0.9586535381174625, + "learning_rate": 9.895287045789608e-06, + "loss": 0.7303, + "step": 828 + }, + { + "epoch": 0.7959673547767643, + "grad_norm": 1.3749908323684445, + "learning_rate": 9.894603947638023e-06, + "loss": 0.7061, + "step": 829 + }, + { + "epoch": 0.7969275084013442, + "grad_norm": 1.1785565342144533, + "learning_rate": 9.893918652351048e-06, + "loss": 0.7537, + "step": 830 + }, + { + "epoch": 0.7978876620259241, + "grad_norm": 1.0812103342300967, + "learning_rate": 9.893231160236303e-06, + "loss": 0.8364, + "step": 831 + }, + { + "epoch": 0.7988478156505041, + "grad_norm": 0.9867997910803941, + "learning_rate": 9.892541471602395e-06, + "loss": 0.6845, + "step": 832 + }, + { + "epoch": 0.799807969275084, + "grad_norm": 0.9732673403408759, + "learning_rate": 9.891849586758918e-06, + "loss": 0.7633, + "step": 833 + }, + { + "epoch": 0.8007681228996639, + "grad_norm": 0.9872640710831527, + "learning_rate": 9.891155506016452e-06, + "loss": 0.7817, + "step": 834 + }, + { + "epoch": 0.8017282765242438, + "grad_norm": 1.055306679841214, + "learning_rate": 9.890459229686564e-06, + "loss": 0.6603, + "step": 835 + }, + { + "epoch": 0.8026884301488239, + "grad_norm": 0.9568869069835517, + "learning_rate": 9.889760758081803e-06, + "loss": 0.6943, + "step": 836 + }, + { + "epoch": 0.8036485837734038, + "grad_norm": 1.0072131187344215, + "learning_rate": 9.889060091515707e-06, + "loss": 0.6164, + "step": 837 + }, + { + "epoch": 0.8046087373979837, + "grad_norm": 0.9978428928688824, + "learning_rate": 9.8883572303028e-06, + "loss": 0.7224, + "step": 838 + }, + { + "epoch": 0.8055688910225636, + "grad_norm": 1.2332002879202255, + "learning_rate": 9.887652174758581e-06, + "loss": 0.6889, + "step": 839 + }, + { + "epoch": 0.8065290446471436, + "grad_norm": 1.0961807544454492, + "learning_rate": 9.88694492519955e-06, + "loss": 0.6717, + "step": 840 + }, + { + "epoch": 0.8074891982717235, + "grad_norm": 1.1307172390388518, + "learning_rate": 9.88623548194318e-06, + "loss": 0.8255, + "step": 841 + }, + { + "epoch": 0.8084493518963034, + "grad_norm": 0.9297760721920874, + "learning_rate": 9.88552384530793e-06, + "loss": 0.666, + "step": 842 + }, + { + "epoch": 0.8094095055208833, + "grad_norm": 1.1095655791690846, + "learning_rate": 9.884810015613255e-06, + "loss": 0.7067, + "step": 843 + }, + { + "epoch": 0.8103696591454633, + "grad_norm": 1.1551923464569889, + "learning_rate": 9.884093993179575e-06, + "loss": 0.7585, + "step": 844 + }, + { + "epoch": 0.8113298127700432, + "grad_norm": 0.9387136764430609, + "learning_rate": 9.883375778328312e-06, + "loss": 0.7272, + "step": 845 + }, + { + "epoch": 0.8122899663946231, + "grad_norm": 1.0708716761811787, + "learning_rate": 9.882655371381861e-06, + "loss": 0.7268, + "step": 846 + }, + { + "epoch": 0.813250120019203, + "grad_norm": 0.948634004314031, + "learning_rate": 9.881932772663609e-06, + "loss": 0.7238, + "step": 847 + }, + { + "epoch": 0.814210273643783, + "grad_norm": 1.185852375482035, + "learning_rate": 9.881207982497917e-06, + "loss": 0.7256, + "step": 848 + }, + { + "epoch": 0.815170427268363, + "grad_norm": 1.0054289883736454, + "learning_rate": 9.880481001210142e-06, + "loss": 0.6653, + "step": 849 + }, + { + "epoch": 0.8161305808929429, + "grad_norm": 0.9993636485340736, + "learning_rate": 9.879751829126612e-06, + "loss": 0.6556, + "step": 850 + }, + { + "epoch": 0.8170907345175228, + "grad_norm": 0.9645142898485913, + "learning_rate": 9.87902046657465e-06, + "loss": 0.6754, + "step": 851 + }, + { + "epoch": 0.8180508881421027, + "grad_norm": 0.9891246316672525, + "learning_rate": 9.878286913882553e-06, + "loss": 0.6649, + "step": 852 + }, + { + "epoch": 0.8190110417666827, + "grad_norm": 1.3063650699400922, + "learning_rate": 9.877551171379606e-06, + "loss": 0.8185, + "step": 853 + }, + { + "epoch": 0.8199711953912626, + "grad_norm": 1.1918128804072938, + "learning_rate": 9.876813239396079e-06, + "loss": 0.7668, + "step": 854 + }, + { + "epoch": 0.8209313490158425, + "grad_norm": 0.8694013244785461, + "learning_rate": 9.876073118263216e-06, + "loss": 0.6434, + "step": 855 + }, + { + "epoch": 0.8218915026404224, + "grad_norm": 1.092408433099053, + "learning_rate": 9.875330808313254e-06, + "loss": 0.7972, + "step": 856 + }, + { + "epoch": 0.8228516562650025, + "grad_norm": 1.412526926967191, + "learning_rate": 9.874586309879406e-06, + "loss": 0.656, + "step": 857 + }, + { + "epoch": 0.8238118098895824, + "grad_norm": 1.1217090893050323, + "learning_rate": 9.87383962329587e-06, + "loss": 0.8015, + "step": 858 + }, + { + "epoch": 0.8247719635141623, + "grad_norm": 1.1657462651565507, + "learning_rate": 9.873090748897826e-06, + "loss": 0.7275, + "step": 859 + }, + { + "epoch": 0.8257321171387422, + "grad_norm": 1.1787793782521532, + "learning_rate": 9.872339687021434e-06, + "loss": 0.7413, + "step": 860 + }, + { + "epoch": 0.8266922707633221, + "grad_norm": 1.340616231597411, + "learning_rate": 9.871586438003841e-06, + "loss": 0.7786, + "step": 861 + }, + { + "epoch": 0.8276524243879021, + "grad_norm": 1.179602094426698, + "learning_rate": 9.870831002183168e-06, + "loss": 0.7823, + "step": 862 + }, + { + "epoch": 0.828612578012482, + "grad_norm": 1.1223593173529196, + "learning_rate": 9.870073379898526e-06, + "loss": 0.7104, + "step": 863 + }, + { + "epoch": 0.8295727316370619, + "grad_norm": 1.400705044676731, + "learning_rate": 9.869313571490002e-06, + "loss": 0.6805, + "step": 864 + }, + { + "epoch": 0.8305328852616418, + "grad_norm": 1.3950024166572066, + "learning_rate": 9.868551577298665e-06, + "loss": 0.7823, + "step": 865 + }, + { + "epoch": 0.8314930388862218, + "grad_norm": 1.256792972347036, + "learning_rate": 9.867787397666568e-06, + "loss": 0.6954, + "step": 866 + }, + { + "epoch": 0.8324531925108017, + "grad_norm": 0.8067378155239507, + "learning_rate": 9.86702103293674e-06, + "loss": 0.6488, + "step": 867 + }, + { + "epoch": 0.8334133461353816, + "grad_norm": 0.9597200041625673, + "learning_rate": 9.866252483453199e-06, + "loss": 0.7559, + "step": 868 + }, + { + "epoch": 0.8343734997599616, + "grad_norm": 1.3440961886731113, + "learning_rate": 9.865481749560933e-06, + "loss": 0.8602, + "step": 869 + }, + { + "epoch": 0.8353336533845416, + "grad_norm": 0.8939860482760397, + "learning_rate": 9.864708831605918e-06, + "loss": 0.6142, + "step": 870 + }, + { + "epoch": 0.8362938070091215, + "grad_norm": 1.3127632405498024, + "learning_rate": 9.863933729935112e-06, + "loss": 0.779, + "step": 871 + }, + { + "epoch": 0.8372539606337014, + "grad_norm": 1.4737052784454916, + "learning_rate": 9.863156444896446e-06, + "loss": 0.7681, + "step": 872 + }, + { + "epoch": 0.8382141142582813, + "grad_norm": 1.2860252668236365, + "learning_rate": 9.862376976838835e-06, + "loss": 0.6934, + "step": 873 + }, + { + "epoch": 0.8391742678828612, + "grad_norm": 1.1395463857271173, + "learning_rate": 9.861595326112176e-06, + "loss": 0.7219, + "step": 874 + }, + { + "epoch": 0.8401344215074412, + "grad_norm": 1.009070130534368, + "learning_rate": 9.860811493067344e-06, + "loss": 0.7276, + "step": 875 + }, + { + "epoch": 0.8410945751320211, + "grad_norm": 1.0647868874105912, + "learning_rate": 9.86002547805619e-06, + "loss": 0.6948, + "step": 876 + }, + { + "epoch": 0.842054728756601, + "grad_norm": 0.8266180898239189, + "learning_rate": 9.85923728143155e-06, + "loss": 0.7082, + "step": 877 + }, + { + "epoch": 0.8430148823811809, + "grad_norm": 0.9831565698075821, + "learning_rate": 9.858446903547239e-06, + "loss": 0.7596, + "step": 878 + }, + { + "epoch": 0.843975036005761, + "grad_norm": 1.4438296502868566, + "learning_rate": 9.857654344758044e-06, + "loss": 0.8311, + "step": 879 + }, + { + "epoch": 0.8449351896303409, + "grad_norm": 1.563787214147985, + "learning_rate": 9.85685960541974e-06, + "loss": 0.8639, + "step": 880 + }, + { + "epoch": 0.8458953432549208, + "grad_norm": 1.0946544146378487, + "learning_rate": 9.856062685889076e-06, + "loss": 0.7273, + "step": 881 + }, + { + "epoch": 0.8468554968795007, + "grad_norm": 1.1508085335490117, + "learning_rate": 9.855263586523782e-06, + "loss": 0.763, + "step": 882 + }, + { + "epoch": 0.8478156505040807, + "grad_norm": 0.7477159491211194, + "learning_rate": 9.854462307682561e-06, + "loss": 0.627, + "step": 883 + }, + { + "epoch": 0.8487758041286606, + "grad_norm": 1.3831521429558586, + "learning_rate": 9.853658849725105e-06, + "loss": 0.8919, + "step": 884 + }, + { + "epoch": 0.8497359577532405, + "grad_norm": 0.9523410225588861, + "learning_rate": 9.852853213012072e-06, + "loss": 0.6652, + "step": 885 + }, + { + "epoch": 0.8506961113778204, + "grad_norm": 0.8440621024279631, + "learning_rate": 9.852045397905106e-06, + "loss": 0.6793, + "step": 886 + }, + { + "epoch": 0.8516562650024004, + "grad_norm": 0.9628712617820975, + "learning_rate": 9.851235404766827e-06, + "loss": 0.7682, + "step": 887 + }, + { + "epoch": 0.8526164186269803, + "grad_norm": 1.10112303741811, + "learning_rate": 9.850423233960831e-06, + "loss": 0.764, + "step": 888 + }, + { + "epoch": 0.8535765722515603, + "grad_norm": 0.8290764739831752, + "learning_rate": 9.849608885851693e-06, + "loss": 0.6713, + "step": 889 + }, + { + "epoch": 0.8545367258761402, + "grad_norm": 1.187481809310174, + "learning_rate": 9.848792360804966e-06, + "loss": 0.7793, + "step": 890 + }, + { + "epoch": 0.8554968795007201, + "grad_norm": 1.2014014665986943, + "learning_rate": 9.84797365918718e-06, + "loss": 0.7957, + "step": 891 + }, + { + "epoch": 0.8564570331253001, + "grad_norm": 1.5352983734519552, + "learning_rate": 9.847152781365839e-06, + "loss": 0.8209, + "step": 892 + }, + { + "epoch": 0.85741718674988, + "grad_norm": 1.0656341787458772, + "learning_rate": 9.846329727709429e-06, + "loss": 0.7326, + "step": 893 + }, + { + "epoch": 0.8583773403744599, + "grad_norm": 0.89403431681965, + "learning_rate": 9.845504498587406e-06, + "loss": 0.709, + "step": 894 + }, + { + "epoch": 0.8593374939990398, + "grad_norm": 1.015401224155412, + "learning_rate": 9.844677094370212e-06, + "loss": 0.6282, + "step": 895 + }, + { + "epoch": 0.8602976476236198, + "grad_norm": 0.938644230385865, + "learning_rate": 9.843847515429256e-06, + "loss": 0.7081, + "step": 896 + }, + { + "epoch": 0.8612578012481997, + "grad_norm": 1.1323586524295224, + "learning_rate": 9.843015762136927e-06, + "loss": 0.7606, + "step": 897 + }, + { + "epoch": 0.8622179548727796, + "grad_norm": 1.1903531592349523, + "learning_rate": 9.842181834866593e-06, + "loss": 0.8757, + "step": 898 + }, + { + "epoch": 0.8631781084973595, + "grad_norm": 1.1879468386238274, + "learning_rate": 9.841345733992594e-06, + "loss": 0.6575, + "step": 899 + }, + { + "epoch": 0.8641382621219396, + "grad_norm": 1.087325624822077, + "learning_rate": 9.840507459890244e-06, + "loss": 0.6507, + "step": 900 + }, + { + "epoch": 0.8650984157465195, + "grad_norm": 0.9657237185910814, + "learning_rate": 9.839667012935838e-06, + "loss": 0.7462, + "step": 901 + }, + { + "epoch": 0.8660585693710994, + "grad_norm": 1.0414975352895643, + "learning_rate": 9.838824393506644e-06, + "loss": 0.7267, + "step": 902 + }, + { + "epoch": 0.8670187229956793, + "grad_norm": 0.9438593248708991, + "learning_rate": 9.837979601980902e-06, + "loss": 0.6806, + "step": 903 + }, + { + "epoch": 0.8679788766202592, + "grad_norm": 1.2800688399586553, + "learning_rate": 9.837132638737834e-06, + "loss": 0.7785, + "step": 904 + }, + { + "epoch": 0.8689390302448392, + "grad_norm": 1.0178473453974097, + "learning_rate": 9.83628350415763e-06, + "loss": 0.6255, + "step": 905 + }, + { + "epoch": 0.8698991838694191, + "grad_norm": 1.2880937767052882, + "learning_rate": 9.835432198621457e-06, + "loss": 0.6661, + "step": 906 + }, + { + "epoch": 0.870859337493999, + "grad_norm": 0.9608735784060488, + "learning_rate": 9.83457872251146e-06, + "loss": 0.6812, + "step": 907 + }, + { + "epoch": 0.8718194911185789, + "grad_norm": 1.0969764334445848, + "learning_rate": 9.833723076210751e-06, + "loss": 0.7732, + "step": 908 + }, + { + "epoch": 0.8727796447431589, + "grad_norm": 0.7722051052853002, + "learning_rate": 9.832865260103423e-06, + "loss": 0.5588, + "step": 909 + }, + { + "epoch": 0.8737397983677389, + "grad_norm": 1.5185930389007303, + "learning_rate": 9.832005274574541e-06, + "loss": 0.6865, + "step": 910 + }, + { + "epoch": 0.8746999519923188, + "grad_norm": 1.0466221007331493, + "learning_rate": 9.831143120010143e-06, + "loss": 0.7048, + "step": 911 + }, + { + "epoch": 0.8756601056168987, + "grad_norm": 1.2581603579966176, + "learning_rate": 9.830278796797239e-06, + "loss": 0.6606, + "step": 912 + }, + { + "epoch": 0.8766202592414787, + "grad_norm": 1.0031861202706405, + "learning_rate": 9.829412305323818e-06, + "loss": 0.7588, + "step": 913 + }, + { + "epoch": 0.8775804128660586, + "grad_norm": 1.5109231034827575, + "learning_rate": 9.828543645978835e-06, + "loss": 0.7686, + "step": 914 + }, + { + "epoch": 0.8785405664906385, + "grad_norm": 0.9541242793362736, + "learning_rate": 9.827672819152223e-06, + "loss": 0.7057, + "step": 915 + }, + { + "epoch": 0.8795007201152184, + "grad_norm": 0.930848836653686, + "learning_rate": 9.82679982523489e-06, + "loss": 0.8373, + "step": 916 + }, + { + "epoch": 0.8804608737397984, + "grad_norm": 0.8351213660877761, + "learning_rate": 9.82592466461871e-06, + "loss": 0.6362, + "step": 917 + }, + { + "epoch": 0.8814210273643783, + "grad_norm": 1.6665042808958377, + "learning_rate": 9.825047337696532e-06, + "loss": 0.8145, + "step": 918 + }, + { + "epoch": 0.8823811809889582, + "grad_norm": 1.3889228139305345, + "learning_rate": 9.824167844862184e-06, + "loss": 0.7571, + "step": 919 + }, + { + "epoch": 0.8833413346135381, + "grad_norm": 1.0552246147514572, + "learning_rate": 9.823286186510456e-06, + "loss": 0.6295, + "step": 920 + }, + { + "epoch": 0.884301488238118, + "grad_norm": 1.0503134578542208, + "learning_rate": 9.822402363037118e-06, + "loss": 0.7932, + "step": 921 + }, + { + "epoch": 0.8852616418626981, + "grad_norm": 1.1169129482469298, + "learning_rate": 9.82151637483891e-06, + "loss": 0.8022, + "step": 922 + }, + { + "epoch": 0.886221795487278, + "grad_norm": 1.0773362308593635, + "learning_rate": 9.820628222313535e-06, + "loss": 0.7533, + "step": 923 + }, + { + "epoch": 0.8871819491118579, + "grad_norm": 0.9450074031059397, + "learning_rate": 9.819737905859685e-06, + "loss": 0.6949, + "step": 924 + }, + { + "epoch": 0.8881421027364378, + "grad_norm": 0.9031119730881773, + "learning_rate": 9.818845425877008e-06, + "loss": 0.7283, + "step": 925 + }, + { + "epoch": 0.8891022563610178, + "grad_norm": 0.865666045566144, + "learning_rate": 9.81795078276613e-06, + "loss": 0.5821, + "step": 926 + }, + { + "epoch": 0.8900624099855977, + "grad_norm": 1.0100469478753122, + "learning_rate": 9.817053976928646e-06, + "loss": 0.7533, + "step": 927 + }, + { + "epoch": 0.8910225636101776, + "grad_norm": 1.0527936122369437, + "learning_rate": 9.816155008767123e-06, + "loss": 0.6932, + "step": 928 + }, + { + "epoch": 0.8919827172347575, + "grad_norm": 0.8333309807679432, + "learning_rate": 9.8152538786851e-06, + "loss": 0.6628, + "step": 929 + }, + { + "epoch": 0.8929428708593375, + "grad_norm": 1.366332958709172, + "learning_rate": 9.814350587087083e-06, + "loss": 0.7792, + "step": 930 + }, + { + "epoch": 0.8939030244839175, + "grad_norm": 0.9127210005336927, + "learning_rate": 9.813445134378548e-06, + "loss": 0.5963, + "step": 931 + }, + { + "epoch": 0.8948631781084974, + "grad_norm": 0.9688156315477393, + "learning_rate": 9.812537520965946e-06, + "loss": 0.722, + "step": 932 + }, + { + "epoch": 0.8958233317330773, + "grad_norm": 0.8734856874686121, + "learning_rate": 9.811627747256694e-06, + "loss": 0.6533, + "step": 933 + }, + { + "epoch": 0.8967834853576572, + "grad_norm": 1.1996538417789966, + "learning_rate": 9.81071581365918e-06, + "loss": 0.8021, + "step": 934 + }, + { + "epoch": 0.8977436389822372, + "grad_norm": 0.7782880837157595, + "learning_rate": 9.809801720582761e-06, + "loss": 0.6337, + "step": 935 + }, + { + "epoch": 0.8987037926068171, + "grad_norm": 1.241375510903179, + "learning_rate": 9.808885468437764e-06, + "loss": 0.8058, + "step": 936 + }, + { + "epoch": 0.899663946231397, + "grad_norm": 0.9984549357600859, + "learning_rate": 9.807967057635484e-06, + "loss": 0.6715, + "step": 937 + }, + { + "epoch": 0.9006240998559769, + "grad_norm": 1.612263573576596, + "learning_rate": 9.807046488588186e-06, + "loss": 0.7684, + "step": 938 + }, + { + "epoch": 0.9015842534805569, + "grad_norm": 0.9560953567318755, + "learning_rate": 9.806123761709103e-06, + "loss": 0.7007, + "step": 939 + }, + { + "epoch": 0.9025444071051368, + "grad_norm": 1.1696828477556798, + "learning_rate": 9.80519887741244e-06, + "loss": 0.7824, + "step": 940 + }, + { + "epoch": 0.9035045607297167, + "grad_norm": 1.2451480990025774, + "learning_rate": 9.804271836113364e-06, + "loss": 0.816, + "step": 941 + }, + { + "epoch": 0.9044647143542967, + "grad_norm": 0.9536640280225801, + "learning_rate": 9.803342638228014e-06, + "loss": 0.6372, + "step": 942 + }, + { + "epoch": 0.9054248679788767, + "grad_norm": 0.7391240801814135, + "learning_rate": 9.802411284173502e-06, + "loss": 0.612, + "step": 943 + }, + { + "epoch": 0.9063850216034566, + "grad_norm": 1.089277700491067, + "learning_rate": 9.801477774367898e-06, + "loss": 0.7988, + "step": 944 + }, + { + "epoch": 0.9073451752280365, + "grad_norm": 0.8675997079996093, + "learning_rate": 9.800542109230247e-06, + "loss": 0.6562, + "step": 945 + }, + { + "epoch": 0.9083053288526164, + "grad_norm": 1.1584731501395555, + "learning_rate": 9.799604289180556e-06, + "loss": 0.8249, + "step": 946 + }, + { + "epoch": 0.9092654824771963, + "grad_norm": 0.8931861272908558, + "learning_rate": 9.798664314639805e-06, + "loss": 0.6226, + "step": 947 + }, + { + "epoch": 0.9102256361017763, + "grad_norm": 0.7958676768875341, + "learning_rate": 9.79772218602994e-06, + "loss": 0.5626, + "step": 948 + }, + { + "epoch": 0.9111857897263562, + "grad_norm": 1.1274446320695066, + "learning_rate": 9.796777903773868e-06, + "loss": 0.7325, + "step": 949 + }, + { + "epoch": 0.9121459433509361, + "grad_norm": 1.6436918793587711, + "learning_rate": 9.795831468295472e-06, + "loss": 0.7576, + "step": 950 + }, + { + "epoch": 0.913106096975516, + "grad_norm": 0.9519199622538829, + "learning_rate": 9.794882880019594e-06, + "loss": 0.6777, + "step": 951 + }, + { + "epoch": 0.914066250600096, + "grad_norm": 1.3511811414293768, + "learning_rate": 9.793932139372045e-06, + "loss": 0.6944, + "step": 952 + }, + { + "epoch": 0.915026404224676, + "grad_norm": 1.6853823596747193, + "learning_rate": 9.792979246779602e-06, + "loss": 0.8167, + "step": 953 + }, + { + "epoch": 0.9159865578492559, + "grad_norm": 0.9679278304656862, + "learning_rate": 9.79202420267001e-06, + "loss": 0.6716, + "step": 954 + }, + { + "epoch": 0.9169467114738358, + "grad_norm": 0.8875846722638174, + "learning_rate": 9.791067007471977e-06, + "loss": 0.6776, + "step": 955 + }, + { + "epoch": 0.9179068650984158, + "grad_norm": 1.1772177263021604, + "learning_rate": 9.790107661615179e-06, + "loss": 0.7093, + "step": 956 + }, + { + "epoch": 0.9188670187229957, + "grad_norm": 1.1549620029922816, + "learning_rate": 9.789146165530255e-06, + "loss": 0.6411, + "step": 957 + }, + { + "epoch": 0.9198271723475756, + "grad_norm": 1.1434851514172506, + "learning_rate": 9.78818251964881e-06, + "loss": 0.7028, + "step": 958 + }, + { + "epoch": 0.9207873259721555, + "grad_norm": 0.8836920295059776, + "learning_rate": 9.787216724403414e-06, + "loss": 0.5954, + "step": 959 + }, + { + "epoch": 0.9217474795967355, + "grad_norm": 1.0899293586144565, + "learning_rate": 9.786248780227604e-06, + "loss": 0.7355, + "step": 960 + }, + { + "epoch": 0.9227076332213154, + "grad_norm": 1.190062067333841, + "learning_rate": 9.785278687555879e-06, + "loss": 0.6962, + "step": 961 + }, + { + "epoch": 0.9236677868458953, + "grad_norm": 1.0814165951280394, + "learning_rate": 9.784306446823702e-06, + "loss": 0.769, + "step": 962 + }, + { + "epoch": 0.9246279404704753, + "grad_norm": 1.1042095834041685, + "learning_rate": 9.783332058467502e-06, + "loss": 0.6473, + "step": 963 + }, + { + "epoch": 0.9255880940950552, + "grad_norm": 1.0550430585383388, + "learning_rate": 9.782355522924674e-06, + "loss": 0.7857, + "step": 964 + }, + { + "epoch": 0.9265482477196352, + "grad_norm": 1.0585925318826632, + "learning_rate": 9.78137684063357e-06, + "loss": 0.7238, + "step": 965 + }, + { + "epoch": 0.9275084013442151, + "grad_norm": 1.3454075899995044, + "learning_rate": 9.780396012033514e-06, + "loss": 0.7132, + "step": 966 + }, + { + "epoch": 0.928468554968795, + "grad_norm": 1.574673131762867, + "learning_rate": 9.779413037564787e-06, + "loss": 0.769, + "step": 967 + }, + { + "epoch": 0.9294287085933749, + "grad_norm": 0.8652122236817356, + "learning_rate": 9.778427917668636e-06, + "loss": 0.6542, + "step": 968 + }, + { + "epoch": 0.9303888622179549, + "grad_norm": 0.8921018355811567, + "learning_rate": 9.777440652787273e-06, + "loss": 0.6678, + "step": 969 + }, + { + "epoch": 0.9313490158425348, + "grad_norm": 1.1528185965641355, + "learning_rate": 9.776451243363866e-06, + "loss": 0.787, + "step": 970 + }, + { + "epoch": 0.9323091694671147, + "grad_norm": 1.1132653970572794, + "learning_rate": 9.775459689842558e-06, + "loss": 0.7405, + "step": 971 + }, + { + "epoch": 0.9332693230916946, + "grad_norm": 1.2393363136160531, + "learning_rate": 9.774465992668438e-06, + "loss": 0.7021, + "step": 972 + }, + { + "epoch": 0.9342294767162747, + "grad_norm": 0.9355066306206749, + "learning_rate": 9.773470152287573e-06, + "loss": 0.7447, + "step": 973 + }, + { + "epoch": 0.9351896303408546, + "grad_norm": 0.9415751218932975, + "learning_rate": 9.772472169146977e-06, + "loss": 0.6681, + "step": 974 + }, + { + "epoch": 0.9361497839654345, + "grad_norm": 1.0843407626427144, + "learning_rate": 9.771472043694645e-06, + "loss": 0.7294, + "step": 975 + }, + { + "epoch": 0.9371099375900144, + "grad_norm": 1.0410943714404997, + "learning_rate": 9.770469776379514e-06, + "loss": 0.7667, + "step": 976 + }, + { + "epoch": 0.9380700912145943, + "grad_norm": 0.8336942772176519, + "learning_rate": 9.769465367651494e-06, + "loss": 0.6097, + "step": 977 + }, + { + "epoch": 0.9390302448391743, + "grad_norm": 1.2244204622837596, + "learning_rate": 9.768458817961455e-06, + "loss": 0.6858, + "step": 978 + }, + { + "epoch": 0.9399903984637542, + "grad_norm": 0.8684284584573115, + "learning_rate": 9.767450127761223e-06, + "loss": 0.6282, + "step": 979 + }, + { + "epoch": 0.9409505520883341, + "grad_norm": 0.917142452064741, + "learning_rate": 9.766439297503589e-06, + "loss": 0.6797, + "step": 980 + }, + { + "epoch": 0.941910705712914, + "grad_norm": 1.0197470571452965, + "learning_rate": 9.765426327642306e-06, + "loss": 0.7103, + "step": 981 + }, + { + "epoch": 0.942870859337494, + "grad_norm": 1.3988824865834113, + "learning_rate": 9.764411218632085e-06, + "loss": 0.7972, + "step": 982 + }, + { + "epoch": 0.943831012962074, + "grad_norm": 1.1531592311801235, + "learning_rate": 9.763393970928593e-06, + "loss": 0.6764, + "step": 983 + }, + { + "epoch": 0.9447911665866539, + "grad_norm": 1.225196826877668, + "learning_rate": 9.76237458498847e-06, + "loss": 0.6423, + "step": 984 + }, + { + "epoch": 0.9457513202112338, + "grad_norm": 0.9255053798709998, + "learning_rate": 9.7613530612693e-06, + "loss": 0.6171, + "step": 985 + }, + { + "epoch": 0.9467114738358138, + "grad_norm": 1.1231221742640263, + "learning_rate": 9.760329400229639e-06, + "loss": 0.7753, + "step": 986 + }, + { + "epoch": 0.9476716274603937, + "grad_norm": 0.9184948985050901, + "learning_rate": 9.759303602328993e-06, + "loss": 0.7141, + "step": 987 + }, + { + "epoch": 0.9486317810849736, + "grad_norm": 0.9349566038732365, + "learning_rate": 9.758275668027834e-06, + "loss": 0.7295, + "step": 988 + }, + { + "epoch": 0.9495919347095535, + "grad_norm": 0.9080066148908356, + "learning_rate": 9.757245597787592e-06, + "loss": 0.7392, + "step": 989 + }, + { + "epoch": 0.9505520883341335, + "grad_norm": 1.269783362818564, + "learning_rate": 9.756213392070654e-06, + "loss": 0.7644, + "step": 990 + }, + { + "epoch": 0.9515122419587134, + "grad_norm": 0.8583944516713555, + "learning_rate": 9.755179051340365e-06, + "loss": 0.6627, + "step": 991 + }, + { + "epoch": 0.9524723955832933, + "grad_norm": 0.9155158800596387, + "learning_rate": 9.75414257606103e-06, + "loss": 0.6376, + "step": 992 + }, + { + "epoch": 0.9534325492078732, + "grad_norm": 1.208151950924738, + "learning_rate": 9.75310396669791e-06, + "loss": 0.7503, + "step": 993 + }, + { + "epoch": 0.9543927028324531, + "grad_norm": 1.033989677497992, + "learning_rate": 9.752063223717227e-06, + "loss": 0.7704, + "step": 994 + }, + { + "epoch": 0.9553528564570332, + "grad_norm": 1.210710566475916, + "learning_rate": 9.75102034758616e-06, + "loss": 0.5415, + "step": 995 + }, + { + "epoch": 0.9563130100816131, + "grad_norm": 1.7618543655335772, + "learning_rate": 9.749975338772846e-06, + "loss": 0.8394, + "step": 996 + }, + { + "epoch": 0.957273163706193, + "grad_norm": 0.9979885638324807, + "learning_rate": 9.748928197746376e-06, + "loss": 0.5887, + "step": 997 + }, + { + "epoch": 0.9582333173307729, + "grad_norm": 1.4603713511847616, + "learning_rate": 9.7478789249768e-06, + "loss": 0.7873, + "step": 998 + }, + { + "epoch": 0.9591934709553529, + "grad_norm": 0.8390123183988712, + "learning_rate": 9.746827520935129e-06, + "loss": 0.6434, + "step": 999 + }, + { + "epoch": 0.9601536245799328, + "grad_norm": 1.0913093884651996, + "learning_rate": 9.745773986093325e-06, + "loss": 0.6331, + "step": 1000 + }, + { + "epoch": 0.9611137782045127, + "grad_norm": 1.7215798127009005, + "learning_rate": 9.744718320924307e-06, + "loss": 0.8343, + "step": 1001 + }, + { + "epoch": 0.9620739318290926, + "grad_norm": 1.0625049746786475, + "learning_rate": 9.743660525901952e-06, + "loss": 0.6828, + "step": 1002 + }, + { + "epoch": 0.9630340854536726, + "grad_norm": 1.0133495556448957, + "learning_rate": 9.742600601501095e-06, + "loss": 0.6493, + "step": 1003 + }, + { + "epoch": 0.9639942390782525, + "grad_norm": 0.9451001954187663, + "learning_rate": 9.741538548197522e-06, + "loss": 0.7463, + "step": 1004 + }, + { + "epoch": 0.9649543927028325, + "grad_norm": 1.3277949374129938, + "learning_rate": 9.74047436646798e-06, + "loss": 0.7634, + "step": 1005 + }, + { + "epoch": 0.9659145463274124, + "grad_norm": 1.3966822263033054, + "learning_rate": 9.739408056790168e-06, + "loss": 0.7821, + "step": 1006 + }, + { + "epoch": 0.9668746999519923, + "grad_norm": 1.5078577294435678, + "learning_rate": 9.73833961964274e-06, + "loss": 0.8854, + "step": 1007 + }, + { + "epoch": 0.9678348535765723, + "grad_norm": 1.4087987293027775, + "learning_rate": 9.737269055505308e-06, + "loss": 0.7581, + "step": 1008 + }, + { + "epoch": 0.9687950072011522, + "grad_norm": 1.0960917211859005, + "learning_rate": 9.736196364858435e-06, + "loss": 0.6522, + "step": 1009 + }, + { + "epoch": 0.9697551608257321, + "grad_norm": 1.3430973961906536, + "learning_rate": 9.735121548183642e-06, + "loss": 0.797, + "step": 1010 + }, + { + "epoch": 0.970715314450312, + "grad_norm": 0.9293505165863168, + "learning_rate": 9.7340446059634e-06, + "loss": 0.5644, + "step": 1011 + }, + { + "epoch": 0.971675468074892, + "grad_norm": 1.030867247805389, + "learning_rate": 9.732965538681139e-06, + "loss": 0.7227, + "step": 1012 + }, + { + "epoch": 0.9726356216994719, + "grad_norm": 0.9622507208004649, + "learning_rate": 9.73188434682124e-06, + "loss": 0.6653, + "step": 1013 + }, + { + "epoch": 0.9735957753240518, + "grad_norm": 1.2255800005686928, + "learning_rate": 9.73080103086904e-06, + "loss": 0.8339, + "step": 1014 + }, + { + "epoch": 0.9745559289486317, + "grad_norm": 1.0563989893236743, + "learning_rate": 9.729715591310824e-06, + "loss": 0.645, + "step": 1015 + }, + { + "epoch": 0.9755160825732118, + "grad_norm": 0.9398333861628091, + "learning_rate": 9.728628028633839e-06, + "loss": 0.7346, + "step": 1016 + }, + { + "epoch": 0.9764762361977917, + "grad_norm": 1.0391255896214462, + "learning_rate": 9.727538343326277e-06, + "loss": 0.8132, + "step": 1017 + }, + { + "epoch": 0.9774363898223716, + "grad_norm": 1.469070299585559, + "learning_rate": 9.72644653587729e-06, + "loss": 0.9106, + "step": 1018 + }, + { + "epoch": 0.9783965434469515, + "grad_norm": 0.9981697225057418, + "learning_rate": 9.725352606776973e-06, + "loss": 0.7238, + "step": 1019 + }, + { + "epoch": 0.9793566970715314, + "grad_norm": 0.7857074918295711, + "learning_rate": 9.724256556516383e-06, + "loss": 0.6378, + "step": 1020 + }, + { + "epoch": 0.9803168506961114, + "grad_norm": 1.3651912217413142, + "learning_rate": 9.723158385587525e-06, + "loss": 0.7986, + "step": 1021 + }, + { + "epoch": 0.9812770043206913, + "grad_norm": 1.255367872502659, + "learning_rate": 9.722058094483355e-06, + "loss": 0.7143, + "step": 1022 + }, + { + "epoch": 0.9822371579452712, + "grad_norm": 0.9567562529683361, + "learning_rate": 9.72095568369778e-06, + "loss": 0.6067, + "step": 1023 + }, + { + "epoch": 0.9831973115698511, + "grad_norm": 1.135525880037657, + "learning_rate": 9.719851153725668e-06, + "loss": 0.7057, + "step": 1024 + }, + { + "epoch": 0.9841574651944311, + "grad_norm": 1.0612664802665412, + "learning_rate": 9.718744505062822e-06, + "loss": 0.6853, + "step": 1025 + }, + { + "epoch": 0.985117618819011, + "grad_norm": 1.2485057106455786, + "learning_rate": 9.717635738206007e-06, + "loss": 0.6469, + "step": 1026 + }, + { + "epoch": 0.986077772443591, + "grad_norm": 1.0333640641052955, + "learning_rate": 9.716524853652942e-06, + "loss": 0.7068, + "step": 1027 + }, + { + "epoch": 0.9870379260681709, + "grad_norm": 1.1254910643496672, + "learning_rate": 9.715411851902284e-06, + "loss": 0.8391, + "step": 1028 + }, + { + "epoch": 0.9879980796927509, + "grad_norm": 1.1589838064442606, + "learning_rate": 9.71429673345365e-06, + "loss": 0.7025, + "step": 1029 + }, + { + "epoch": 0.9889582333173308, + "grad_norm": 1.3250429014521803, + "learning_rate": 9.713179498807607e-06, + "loss": 0.7709, + "step": 1030 + }, + { + "epoch": 0.9899183869419107, + "grad_norm": 1.4416705150751818, + "learning_rate": 9.712060148465669e-06, + "loss": 0.6959, + "step": 1031 + }, + { + "epoch": 0.9908785405664906, + "grad_norm": 1.1092477155984324, + "learning_rate": 9.710938682930298e-06, + "loss": 0.6921, + "step": 1032 + }, + { + "epoch": 0.9918386941910706, + "grad_norm": 1.142382995678008, + "learning_rate": 9.709815102704909e-06, + "loss": 0.6984, + "step": 1033 + }, + { + "epoch": 0.9927988478156505, + "grad_norm": 1.2806339587976485, + "learning_rate": 9.708689408293866e-06, + "loss": 0.8489, + "step": 1034 + }, + { + "epoch": 0.9937590014402304, + "grad_norm": 1.5699760322221297, + "learning_rate": 9.707561600202481e-06, + "loss": 0.7757, + "step": 1035 + }, + { + "epoch": 0.9947191550648103, + "grad_norm": 0.9806876764853485, + "learning_rate": 9.706431678937015e-06, + "loss": 0.6846, + "step": 1036 + }, + { + "epoch": 0.9956793086893903, + "grad_norm": 1.0528504760370094, + "learning_rate": 9.705299645004679e-06, + "loss": 0.6947, + "step": 1037 + }, + { + "epoch": 0.9966394623139703, + "grad_norm": 0.9003852775733797, + "learning_rate": 9.704165498913624e-06, + "loss": 0.6969, + "step": 1038 + }, + { + "epoch": 0.9975996159385502, + "grad_norm": 1.0114923779312535, + "learning_rate": 9.703029241172965e-06, + "loss": 0.6475, + "step": 1039 + }, + { + "epoch": 0.9985597695631301, + "grad_norm": 1.3534167661452765, + "learning_rate": 9.701890872292753e-06, + "loss": 0.7962, + "step": 1040 + }, + { + "epoch": 0.99951992318771, + "grad_norm": 1.2702910771854647, + "learning_rate": 9.700750392783986e-06, + "loss": 0.6732, + "step": 1041 + }, + { + "epoch": 1.0, + "grad_norm": 1.2702910771854647, + "learning_rate": 9.699607803158618e-06, + "loss": 0.532, + "step": 1042 + }, + { + "epoch": 1.00096015362458, + "grad_norm": 1.5110487943252837, + "learning_rate": 9.698463103929542e-06, + "loss": 0.5891, + "step": 1043 + }, + { + "epoch": 1.0019203072491598, + "grad_norm": 0.7969303706215065, + "learning_rate": 9.697316295610604e-06, + "loss": 0.5777, + "step": 1044 + }, + { + "epoch": 1.0028804608737398, + "grad_norm": 1.5293806184918348, + "learning_rate": 9.696167378716591e-06, + "loss": 0.7407, + "step": 1045 + }, + { + "epoch": 1.0038406144983196, + "grad_norm": 1.1236794712179317, + "learning_rate": 9.695016353763241e-06, + "loss": 0.632, + "step": 1046 + }, + { + "epoch": 1.0048007681228996, + "grad_norm": 1.1510769348363972, + "learning_rate": 9.693863221267238e-06, + "loss": 0.601, + "step": 1047 + }, + { + "epoch": 1.0057609217474797, + "grad_norm": 1.0897212700058942, + "learning_rate": 9.692707981746208e-06, + "loss": 0.6952, + "step": 1048 + }, + { + "epoch": 1.0067210753720595, + "grad_norm": 1.0611847600479067, + "learning_rate": 9.691550635718729e-06, + "loss": 0.6969, + "step": 1049 + }, + { + "epoch": 1.0076812289966395, + "grad_norm": 1.207102458712394, + "learning_rate": 9.690391183704318e-06, + "loss": 0.6293, + "step": 1050 + }, + { + "epoch": 1.0086413826212195, + "grad_norm": 0.9751133947510949, + "learning_rate": 9.689229626223441e-06, + "loss": 0.7174, + "step": 1051 + }, + { + "epoch": 1.0096015362457993, + "grad_norm": 1.000000656715612, + "learning_rate": 9.688065963797512e-06, + "loss": 0.5724, + "step": 1052 + }, + { + "epoch": 1.0105616898703793, + "grad_norm": 1.4651037042965558, + "learning_rate": 9.686900196948886e-06, + "loss": 0.6057, + "step": 1053 + }, + { + "epoch": 1.011521843494959, + "grad_norm": 1.0289676237018328, + "learning_rate": 9.685732326200859e-06, + "loss": 0.6361, + "step": 1054 + }, + { + "epoch": 1.0124819971195391, + "grad_norm": 1.3618297408799271, + "learning_rate": 9.684562352077679e-06, + "loss": 0.7034, + "step": 1055 + }, + { + "epoch": 1.0134421507441191, + "grad_norm": 1.0337606076511678, + "learning_rate": 9.683390275104533e-06, + "loss": 0.6149, + "step": 1056 + }, + { + "epoch": 1.014402304368699, + "grad_norm": 1.3210540444281758, + "learning_rate": 9.682216095807559e-06, + "loss": 0.7235, + "step": 1057 + }, + { + "epoch": 1.015362457993279, + "grad_norm": 1.2917468489810855, + "learning_rate": 9.681039814713827e-06, + "loss": 0.635, + "step": 1058 + }, + { + "epoch": 1.0163226116178588, + "grad_norm": 1.3111956888965486, + "learning_rate": 9.67986143235136e-06, + "loss": 0.7257, + "step": 1059 + }, + { + "epoch": 1.0172827652424388, + "grad_norm": 1.149137628716981, + "learning_rate": 9.67868094924912e-06, + "loss": 0.6367, + "step": 1060 + }, + { + "epoch": 1.0182429188670188, + "grad_norm": 1.0565001299585075, + "learning_rate": 9.677498365937017e-06, + "loss": 0.5431, + "step": 1061 + }, + { + "epoch": 1.0192030724915986, + "grad_norm": 1.000820884997687, + "learning_rate": 9.676313682945896e-06, + "loss": 0.6542, + "step": 1062 + }, + { + "epoch": 1.0201632261161786, + "grad_norm": 1.3601671794282044, + "learning_rate": 9.675126900807548e-06, + "loss": 0.6957, + "step": 1063 + }, + { + "epoch": 1.0211233797407586, + "grad_norm": 1.084951527919392, + "learning_rate": 9.673938020054709e-06, + "loss": 0.5765, + "step": 1064 + }, + { + "epoch": 1.0220835333653384, + "grad_norm": 1.166456668075392, + "learning_rate": 9.672747041221055e-06, + "loss": 0.6789, + "step": 1065 + }, + { + "epoch": 1.0230436869899184, + "grad_norm": 0.887458603305072, + "learning_rate": 9.671553964841206e-06, + "loss": 0.5923, + "step": 1066 + }, + { + "epoch": 1.0240038406144982, + "grad_norm": 0.6994294691373585, + "learning_rate": 9.670358791450713e-06, + "loss": 0.6856, + "step": 1067 + }, + { + "epoch": 1.0249639942390782, + "grad_norm": 1.2050996724454637, + "learning_rate": 9.669161521586086e-06, + "loss": 0.7988, + "step": 1068 + }, + { + "epoch": 1.0259241478636583, + "grad_norm": 1.3281032244939457, + "learning_rate": 9.66796215578476e-06, + "loss": 0.7569, + "step": 1069 + }, + { + "epoch": 1.026884301488238, + "grad_norm": 1.3492642159780655, + "learning_rate": 9.666760694585122e-06, + "loss": 0.6256, + "step": 1070 + }, + { + "epoch": 1.027844455112818, + "grad_norm": 1.3745429981764834, + "learning_rate": 9.665557138526492e-06, + "loss": 0.7721, + "step": 1071 + }, + { + "epoch": 1.0288046087373979, + "grad_norm": 0.9822068570711285, + "learning_rate": 9.664351488149136e-06, + "loss": 0.7684, + "step": 1072 + }, + { + "epoch": 1.029764762361978, + "grad_norm": 1.2721803352495713, + "learning_rate": 9.663143743994254e-06, + "loss": 0.6653, + "step": 1073 + }, + { + "epoch": 1.030724915986558, + "grad_norm": 1.041552222503309, + "learning_rate": 9.661933906603994e-06, + "loss": 0.7034, + "step": 1074 + }, + { + "epoch": 1.0316850696111377, + "grad_norm": 1.1837751802635352, + "learning_rate": 9.660721976521433e-06, + "loss": 0.617, + "step": 1075 + }, + { + "epoch": 1.0326452232357177, + "grad_norm": 1.4451863025436755, + "learning_rate": 9.659507954290599e-06, + "loss": 0.5942, + "step": 1076 + }, + { + "epoch": 1.0336053768602977, + "grad_norm": 0.8662904432223953, + "learning_rate": 9.658291840456454e-06, + "loss": 0.4988, + "step": 1077 + }, + { + "epoch": 1.0345655304848775, + "grad_norm": 0.9639625129611855, + "learning_rate": 9.657073635564893e-06, + "loss": 0.5872, + "step": 1078 + }, + { + "epoch": 1.0355256841094576, + "grad_norm": 0.98202749070384, + "learning_rate": 9.65585334016276e-06, + "loss": 0.6139, + "step": 1079 + }, + { + "epoch": 1.0364858377340374, + "grad_norm": 1.0033748612039537, + "learning_rate": 9.65463095479783e-06, + "loss": 0.526, + "step": 1080 + }, + { + "epoch": 1.0374459913586174, + "grad_norm": 1.0286083707592373, + "learning_rate": 9.653406480018819e-06, + "loss": 0.6296, + "step": 1081 + }, + { + "epoch": 1.0384061449831974, + "grad_norm": 1.4186018864785659, + "learning_rate": 9.652179916375382e-06, + "loss": 0.6323, + "step": 1082 + }, + { + "epoch": 1.0393662986077772, + "grad_norm": 1.048987295491667, + "learning_rate": 9.65095126441811e-06, + "loss": 0.5984, + "step": 1083 + }, + { + "epoch": 1.0403264522323572, + "grad_norm": 1.1790098582257498, + "learning_rate": 9.649720524698532e-06, + "loss": 0.6916, + "step": 1084 + }, + { + "epoch": 1.041286605856937, + "grad_norm": 0.9556042829892353, + "learning_rate": 9.648487697769112e-06, + "loss": 0.643, + "step": 1085 + }, + { + "epoch": 1.042246759481517, + "grad_norm": 0.8133611348102179, + "learning_rate": 9.647252784183254e-06, + "loss": 0.6535, + "step": 1086 + }, + { + "epoch": 1.043206913106097, + "grad_norm": 0.9897981422575274, + "learning_rate": 9.646015784495297e-06, + "loss": 0.6047, + "step": 1087 + }, + { + "epoch": 1.0441670667306768, + "grad_norm": 1.2267752645305483, + "learning_rate": 9.64477669926052e-06, + "loss": 0.6075, + "step": 1088 + }, + { + "epoch": 1.0451272203552568, + "grad_norm": 1.3238292218908252, + "learning_rate": 9.643535529035131e-06, + "loss": 0.7051, + "step": 1089 + }, + { + "epoch": 1.0460873739798369, + "grad_norm": 1.5632993144952576, + "learning_rate": 9.64229227437628e-06, + "loss": 0.6622, + "step": 1090 + }, + { + "epoch": 1.0470475276044167, + "grad_norm": 1.2863923805489401, + "learning_rate": 9.641046935842048e-06, + "loss": 0.5829, + "step": 1091 + }, + { + "epoch": 1.0480076812289967, + "grad_norm": 0.9758855494658143, + "learning_rate": 9.63979951399146e-06, + "loss": 0.671, + "step": 1092 + }, + { + "epoch": 1.0489678348535765, + "grad_norm": 1.22400635676477, + "learning_rate": 9.638550009384467e-06, + "loss": 0.7383, + "step": 1093 + }, + { + "epoch": 1.0499279884781565, + "grad_norm": 0.8371531266770208, + "learning_rate": 9.637298422581955e-06, + "loss": 0.5952, + "step": 1094 + }, + { + "epoch": 1.0508881421027365, + "grad_norm": 0.7397192551218993, + "learning_rate": 9.636044754145753e-06, + "loss": 0.6145, + "step": 1095 + }, + { + "epoch": 1.0518482957273163, + "grad_norm": 1.2804590644356055, + "learning_rate": 9.634789004638618e-06, + "loss": 0.6465, + "step": 1096 + }, + { + "epoch": 1.0528084493518963, + "grad_norm": 0.7206744162417552, + "learning_rate": 9.63353117462424e-06, + "loss": 0.6666, + "step": 1097 + }, + { + "epoch": 1.0537686029764763, + "grad_norm": 1.1434195814155435, + "learning_rate": 9.632271264667249e-06, + "loss": 0.587, + "step": 1098 + }, + { + "epoch": 1.0547287566010561, + "grad_norm": 1.0290623304729012, + "learning_rate": 9.631009275333204e-06, + "loss": 0.6467, + "step": 1099 + }, + { + "epoch": 1.0556889102256362, + "grad_norm": 1.1360460171994695, + "learning_rate": 9.629745207188596e-06, + "loss": 0.8371, + "step": 1100 + }, + { + "epoch": 1.056649063850216, + "grad_norm": 1.1453988893096245, + "learning_rate": 9.628479060800855e-06, + "loss": 0.6032, + "step": 1101 + }, + { + "epoch": 1.057609217474796, + "grad_norm": 0.7128857923854224, + "learning_rate": 9.627210836738338e-06, + "loss": 0.6731, + "step": 1102 + }, + { + "epoch": 1.058569371099376, + "grad_norm": 1.466527768079591, + "learning_rate": 9.62594053557034e-06, + "loss": 0.7217, + "step": 1103 + }, + { + "epoch": 1.0595295247239558, + "grad_norm": 1.416984961465442, + "learning_rate": 9.624668157867084e-06, + "loss": 0.6269, + "step": 1104 + }, + { + "epoch": 1.0604896783485358, + "grad_norm": 1.3810371757039088, + "learning_rate": 9.623393704199727e-06, + "loss": 0.5632, + "step": 1105 + }, + { + "epoch": 1.0614498319731156, + "grad_norm": 0.9934662594978854, + "learning_rate": 9.622117175140357e-06, + "loss": 0.7161, + "step": 1106 + }, + { + "epoch": 1.0624099855976956, + "grad_norm": 1.2094727838376098, + "learning_rate": 9.620838571261995e-06, + "loss": 0.7962, + "step": 1107 + }, + { + "epoch": 1.0633701392222756, + "grad_norm": 1.2298114570943566, + "learning_rate": 9.619557893138593e-06, + "loss": 0.6097, + "step": 1108 + }, + { + "epoch": 1.0643302928468554, + "grad_norm": 1.6030139988398944, + "learning_rate": 9.618275141345032e-06, + "loss": 0.6341, + "step": 1109 + }, + { + "epoch": 1.0652904464714354, + "grad_norm": 1.4528068307728832, + "learning_rate": 9.616990316457127e-06, + "loss": 0.6742, + "step": 1110 + }, + { + "epoch": 1.0662506000960155, + "grad_norm": 1.4810340725441336, + "learning_rate": 9.615703419051622e-06, + "loss": 0.6646, + "step": 1111 + }, + { + "epoch": 1.0672107537205953, + "grad_norm": 1.2063301052611641, + "learning_rate": 9.614414449706191e-06, + "loss": 0.5886, + "step": 1112 + }, + { + "epoch": 1.0681709073451753, + "grad_norm": 0.9906898992808743, + "learning_rate": 9.61312340899944e-06, + "loss": 0.6759, + "step": 1113 + }, + { + "epoch": 1.069131060969755, + "grad_norm": 0.8954179234598504, + "learning_rate": 9.6118302975109e-06, + "loss": 0.6811, + "step": 1114 + }, + { + "epoch": 1.070091214594335, + "grad_norm": 1.0562686078833756, + "learning_rate": 9.61053511582104e-06, + "loss": 0.6808, + "step": 1115 + }, + { + "epoch": 1.0710513682189151, + "grad_norm": 1.151123613333362, + "learning_rate": 9.609237864511249e-06, + "loss": 0.7211, + "step": 1116 + }, + { + "epoch": 1.072011521843495, + "grad_norm": 1.0924995827667974, + "learning_rate": 9.607938544163851e-06, + "loss": 0.751, + "step": 1117 + }, + { + "epoch": 1.072971675468075, + "grad_norm": 1.3204474803168247, + "learning_rate": 9.606637155362095e-06, + "loss": 0.6516, + "step": 1118 + }, + { + "epoch": 1.0739318290926547, + "grad_norm": 1.2692506872886318, + "learning_rate": 9.605333698690164e-06, + "loss": 0.6404, + "step": 1119 + }, + { + "epoch": 1.0748919827172347, + "grad_norm": 1.1176340089741394, + "learning_rate": 9.604028174733164e-06, + "loss": 0.6189, + "step": 1120 + }, + { + "epoch": 1.0758521363418148, + "grad_norm": 1.001771172149186, + "learning_rate": 9.60272058407713e-06, + "loss": 0.6097, + "step": 1121 + }, + { + "epoch": 1.0768122899663946, + "grad_norm": 0.9915205510414012, + "learning_rate": 9.601410927309026e-06, + "loss": 0.5683, + "step": 1122 + }, + { + "epoch": 1.0777724435909746, + "grad_norm": 1.2719858441510843, + "learning_rate": 9.600099205016744e-06, + "loss": 0.6787, + "step": 1123 + }, + { + "epoch": 1.0787325972155546, + "grad_norm": 1.3448172903889624, + "learning_rate": 9.5987854177891e-06, + "loss": 0.6033, + "step": 1124 + }, + { + "epoch": 1.0796927508401344, + "grad_norm": 1.3101282348556873, + "learning_rate": 9.597469566215841e-06, + "loss": 0.6452, + "step": 1125 + }, + { + "epoch": 1.0806529044647144, + "grad_norm": 0.9126392619552218, + "learning_rate": 9.596151650887637e-06, + "loss": 0.744, + "step": 1126 + }, + { + "epoch": 1.0816130580892942, + "grad_norm": 1.096678883600679, + "learning_rate": 9.594831672396088e-06, + "loss": 0.5663, + "step": 1127 + }, + { + "epoch": 1.0825732117138742, + "grad_norm": 1.2189680502188556, + "learning_rate": 9.593509631333717e-06, + "loss": 0.547, + "step": 1128 + }, + { + "epoch": 1.0835333653384542, + "grad_norm": 1.0132906792885186, + "learning_rate": 9.592185528293975e-06, + "loss": 0.5943, + "step": 1129 + }, + { + "epoch": 1.084493518963034, + "grad_norm": 1.4576556473526279, + "learning_rate": 9.590859363871237e-06, + "loss": 0.6347, + "step": 1130 + }, + { + "epoch": 1.085453672587614, + "grad_norm": 1.401688560160359, + "learning_rate": 9.589531138660803e-06, + "loss": 0.6151, + "step": 1131 + }, + { + "epoch": 1.086413826212194, + "grad_norm": 0.9310716344825233, + "learning_rate": 9.588200853258901e-06, + "loss": 0.6459, + "step": 1132 + }, + { + "epoch": 1.0873739798367739, + "grad_norm": 1.4211724125100058, + "learning_rate": 9.586868508262681e-06, + "loss": 0.6429, + "step": 1133 + }, + { + "epoch": 1.0883341334613539, + "grad_norm": 1.3769271746505671, + "learning_rate": 9.58553410427022e-06, + "loss": 0.597, + "step": 1134 + }, + { + "epoch": 1.0892942870859337, + "grad_norm": 0.8611070201311517, + "learning_rate": 9.584197641880513e-06, + "loss": 0.6302, + "step": 1135 + }, + { + "epoch": 1.0902544407105137, + "grad_norm": 0.8510601621483593, + "learning_rate": 9.582859121693488e-06, + "loss": 0.6772, + "step": 1136 + }, + { + "epoch": 1.0912145943350937, + "grad_norm": 0.8837900327645241, + "learning_rate": 9.581518544309993e-06, + "loss": 0.6534, + "step": 1137 + }, + { + "epoch": 1.0921747479596735, + "grad_norm": 1.294948526970718, + "learning_rate": 9.580175910331797e-06, + "loss": 0.6335, + "step": 1138 + }, + { + "epoch": 1.0931349015842535, + "grad_norm": 0.8466550562239562, + "learning_rate": 9.578831220361593e-06, + "loss": 0.7251, + "step": 1139 + }, + { + "epoch": 1.0940950552088333, + "grad_norm": 1.1271084988915, + "learning_rate": 9.577484475003e-06, + "loss": 0.6197, + "step": 1140 + }, + { + "epoch": 1.0950552088334133, + "grad_norm": 1.0209448673116264, + "learning_rate": 9.576135674860557e-06, + "loss": 0.5503, + "step": 1141 + }, + { + "epoch": 1.0960153624579934, + "grad_norm": 1.0666726258807853, + "learning_rate": 9.574784820539725e-06, + "loss": 0.7134, + "step": 1142 + }, + { + "epoch": 1.0969755160825732, + "grad_norm": 1.3085202162459826, + "learning_rate": 9.573431912646888e-06, + "loss": 0.5638, + "step": 1143 + }, + { + "epoch": 1.0979356697071532, + "grad_norm": 1.1263614260396178, + "learning_rate": 9.572076951789353e-06, + "loss": 0.6075, + "step": 1144 + }, + { + "epoch": 1.0988958233317332, + "grad_norm": 1.1508304501793614, + "learning_rate": 9.570719938575349e-06, + "loss": 0.7034, + "step": 1145 + }, + { + "epoch": 1.099855976956313, + "grad_norm": 1.2457271466884658, + "learning_rate": 9.56936087361402e-06, + "loss": 0.664, + "step": 1146 + }, + { + "epoch": 1.100816130580893, + "grad_norm": 1.0148341390353592, + "learning_rate": 9.56799975751544e-06, + "loss": 0.5465, + "step": 1147 + }, + { + "epoch": 1.1017762842054728, + "grad_norm": 0.972231722233613, + "learning_rate": 9.566636590890598e-06, + "loss": 0.7838, + "step": 1148 + }, + { + "epoch": 1.1027364378300528, + "grad_norm": 1.118725517314477, + "learning_rate": 9.565271374351405e-06, + "loss": 0.6724, + "step": 1149 + }, + { + "epoch": 1.1036965914546328, + "grad_norm": 1.1280201861194867, + "learning_rate": 9.563904108510693e-06, + "loss": 0.6967, + "step": 1150 + }, + { + "epoch": 1.1046567450792126, + "grad_norm": 0.9010003252714744, + "learning_rate": 9.562534793982212e-06, + "loss": 0.6369, + "step": 1151 + }, + { + "epoch": 1.1056168987037926, + "grad_norm": 1.0858740201525623, + "learning_rate": 9.56116343138063e-06, + "loss": 0.6405, + "step": 1152 + }, + { + "epoch": 1.1065770523283724, + "grad_norm": 1.4044943630930733, + "learning_rate": 9.559790021321544e-06, + "loss": 0.7943, + "step": 1153 + }, + { + "epoch": 1.1075372059529525, + "grad_norm": 1.4113371109108261, + "learning_rate": 9.558414564421456e-06, + "loss": 0.6222, + "step": 1154 + }, + { + "epoch": 1.1084973595775325, + "grad_norm": 1.0207132568930994, + "learning_rate": 9.557037061297799e-06, + "loss": 0.594, + "step": 1155 + }, + { + "epoch": 1.1094575132021123, + "grad_norm": 1.0170877698543306, + "learning_rate": 9.555657512568916e-06, + "loss": 0.62, + "step": 1156 + }, + { + "epoch": 1.1104176668266923, + "grad_norm": 1.1261083897157445, + "learning_rate": 9.554275918854073e-06, + "loss": 0.6561, + "step": 1157 + }, + { + "epoch": 1.1113778204512723, + "grad_norm": 1.105312343576896, + "learning_rate": 9.55289228077345e-06, + "loss": 0.6724, + "step": 1158 + }, + { + "epoch": 1.112337974075852, + "grad_norm": 1.1372175467826011, + "learning_rate": 9.551506598948152e-06, + "loss": 0.6643, + "step": 1159 + }, + { + "epoch": 1.1132981277004321, + "grad_norm": 1.2311915822855724, + "learning_rate": 9.550118874000196e-06, + "loss": 0.5947, + "step": 1160 + }, + { + "epoch": 1.114258281325012, + "grad_norm": 0.8672811136561209, + "learning_rate": 9.548729106552515e-06, + "loss": 0.6416, + "step": 1161 + }, + { + "epoch": 1.115218434949592, + "grad_norm": 1.0203158135277313, + "learning_rate": 9.54733729722896e-06, + "loss": 0.6898, + "step": 1162 + }, + { + "epoch": 1.116178588574172, + "grad_norm": 1.4296343327952086, + "learning_rate": 9.5459434466543e-06, + "loss": 0.5914, + "step": 1163 + }, + { + "epoch": 1.1171387421987518, + "grad_norm": 1.0449614731472696, + "learning_rate": 9.544547555454222e-06, + "loss": 0.6339, + "step": 1164 + }, + { + "epoch": 1.1180988958233318, + "grad_norm": 1.1184734718295897, + "learning_rate": 9.543149624255326e-06, + "loss": 0.6805, + "step": 1165 + }, + { + "epoch": 1.1190590494479116, + "grad_norm": 1.5050757753662851, + "learning_rate": 9.541749653685124e-06, + "loss": 0.6398, + "step": 1166 + }, + { + "epoch": 1.1200192030724916, + "grad_norm": 1.0627861251187731, + "learning_rate": 9.540347644372054e-06, + "loss": 0.6408, + "step": 1167 + }, + { + "epoch": 1.1209793566970716, + "grad_norm": 1.1010833442250243, + "learning_rate": 9.538943596945459e-06, + "loss": 0.7185, + "step": 1168 + }, + { + "epoch": 1.1219395103216514, + "grad_norm": 1.2756547962854146, + "learning_rate": 9.5375375120356e-06, + "loss": 0.5925, + "step": 1169 + }, + { + "epoch": 1.1228996639462314, + "grad_norm": 0.8752539321638535, + "learning_rate": 9.536129390273659e-06, + "loss": 0.6565, + "step": 1170 + }, + { + "epoch": 1.1238598175708114, + "grad_norm": 1.0638578949842945, + "learning_rate": 9.53471923229172e-06, + "loss": 0.5382, + "step": 1171 + }, + { + "epoch": 1.1248199711953912, + "grad_norm": 1.2180757484115217, + "learning_rate": 9.533307038722792e-06, + "loss": 0.7057, + "step": 1172 + }, + { + "epoch": 1.1257801248199713, + "grad_norm": 0.8609033731653876, + "learning_rate": 9.531892810200793e-06, + "loss": 0.6092, + "step": 1173 + }, + { + "epoch": 1.126740278444551, + "grad_norm": 0.8704390821138364, + "learning_rate": 9.530476547360555e-06, + "loss": 0.576, + "step": 1174 + }, + { + "epoch": 1.127700432069131, + "grad_norm": 1.1661236177930903, + "learning_rate": 9.529058250837823e-06, + "loss": 0.6458, + "step": 1175 + }, + { + "epoch": 1.128660585693711, + "grad_norm": 0.9868914515322151, + "learning_rate": 9.527637921269256e-06, + "loss": 0.616, + "step": 1176 + }, + { + "epoch": 1.1296207393182909, + "grad_norm": 1.1547561709579266, + "learning_rate": 9.526215559292423e-06, + "loss": 0.6516, + "step": 1177 + }, + { + "epoch": 1.130580892942871, + "grad_norm": 1.2761874113530798, + "learning_rate": 9.524791165545808e-06, + "loss": 0.6422, + "step": 1178 + }, + { + "epoch": 1.1315410465674507, + "grad_norm": 0.7061153931590968, + "learning_rate": 9.523364740668805e-06, + "loss": 0.6609, + "step": 1179 + }, + { + "epoch": 1.1325012001920307, + "grad_norm": 0.9106969571888363, + "learning_rate": 9.521936285301723e-06, + "loss": 0.6723, + "step": 1180 + }, + { + "epoch": 1.1334613538166107, + "grad_norm": 1.2465677960183414, + "learning_rate": 9.52050580008578e-06, + "loss": 0.6674, + "step": 1181 + }, + { + "epoch": 1.1344215074411905, + "grad_norm": 1.1124485532076975, + "learning_rate": 9.519073285663103e-06, + "loss": 0.6539, + "step": 1182 + }, + { + "epoch": 1.1353816610657705, + "grad_norm": 1.0718994744889718, + "learning_rate": 9.517638742676736e-06, + "loss": 0.6634, + "step": 1183 + }, + { + "epoch": 1.1363418146903506, + "grad_norm": 1.123650272367286, + "learning_rate": 9.516202171770627e-06, + "loss": 0.7377, + "step": 1184 + }, + { + "epoch": 1.1373019683149304, + "grad_norm": 1.2325714599866064, + "learning_rate": 9.51476357358964e-06, + "loss": 0.763, + "step": 1185 + }, + { + "epoch": 1.1382621219395104, + "grad_norm": 0.7080124445418774, + "learning_rate": 9.513322948779544e-06, + "loss": 0.6175, + "step": 1186 + }, + { + "epoch": 1.1392222755640902, + "grad_norm": 1.1380266027166621, + "learning_rate": 9.511880297987022e-06, + "loss": 0.6257, + "step": 1187 + }, + { + "epoch": 1.1401824291886702, + "grad_norm": 1.0476252524937764, + "learning_rate": 9.510435621859662e-06, + "loss": 0.6239, + "step": 1188 + }, + { + "epoch": 1.1411425828132502, + "grad_norm": 0.9012962777039913, + "learning_rate": 9.508988921045967e-06, + "loss": 0.6829, + "step": 1189 + }, + { + "epoch": 1.14210273643783, + "grad_norm": 1.401581170885142, + "learning_rate": 9.507540196195342e-06, + "loss": 0.5707, + "step": 1190 + }, + { + "epoch": 1.14306289006241, + "grad_norm": 0.8515026664947627, + "learning_rate": 9.506089447958108e-06, + "loss": 0.6342, + "step": 1191 + }, + { + "epoch": 1.1440230436869898, + "grad_norm": 1.4240129750591695, + "learning_rate": 9.504636676985486e-06, + "loss": 0.7272, + "step": 1192 + }, + { + "epoch": 1.1449831973115698, + "grad_norm": 1.0806415361440416, + "learning_rate": 9.503181883929613e-06, + "loss": 0.6605, + "step": 1193 + }, + { + "epoch": 1.1459433509361499, + "grad_norm": 1.0537884629020036, + "learning_rate": 9.501725069443528e-06, + "loss": 0.5705, + "step": 1194 + }, + { + "epoch": 1.1469035045607296, + "grad_norm": 0.9072811020520438, + "learning_rate": 9.50026623418118e-06, + "loss": 0.6128, + "step": 1195 + }, + { + "epoch": 1.1478636581853097, + "grad_norm": 1.2138946145038918, + "learning_rate": 9.498805378797427e-06, + "loss": 0.6345, + "step": 1196 + }, + { + "epoch": 1.1488238118098897, + "grad_norm": 1.2551533599064106, + "learning_rate": 9.497342503948027e-06, + "loss": 0.6823, + "step": 1197 + }, + { + "epoch": 1.1497839654344695, + "grad_norm": 1.3521081824818293, + "learning_rate": 9.49587761028965e-06, + "loss": 0.5514, + "step": 1198 + }, + { + "epoch": 1.1507441190590495, + "grad_norm": 1.0771335823289896, + "learning_rate": 9.494410698479873e-06, + "loss": 0.7498, + "step": 1199 + }, + { + "epoch": 1.1517042726836293, + "grad_norm": 0.9826695036864329, + "learning_rate": 9.492941769177176e-06, + "loss": 0.6652, + "step": 1200 + }, + { + "epoch": 1.1526644263082093, + "grad_norm": 1.0275626147254378, + "learning_rate": 9.491470823040944e-06, + "loss": 0.6806, + "step": 1201 + }, + { + "epoch": 1.1536245799327893, + "grad_norm": 0.9593951529609802, + "learning_rate": 9.489997860731475e-06, + "loss": 0.6689, + "step": 1202 + }, + { + "epoch": 1.1545847335573691, + "grad_norm": 1.02339546311265, + "learning_rate": 9.488522882909959e-06, + "loss": 0.598, + "step": 1203 + }, + { + "epoch": 1.1555448871819491, + "grad_norm": 1.2636604557566715, + "learning_rate": 9.4870458902385e-06, + "loss": 0.6347, + "step": 1204 + }, + { + "epoch": 1.156505040806529, + "grad_norm": 1.40261618222338, + "learning_rate": 9.485566883380105e-06, + "loss": 0.7028, + "step": 1205 + }, + { + "epoch": 1.157465194431109, + "grad_norm": 1.0663856930850502, + "learning_rate": 9.484085862998685e-06, + "loss": 0.5947, + "step": 1206 + }, + { + "epoch": 1.158425348055689, + "grad_norm": 1.4723561780039622, + "learning_rate": 9.482602829759052e-06, + "loss": 0.6393, + "step": 1207 + }, + { + "epoch": 1.1593855016802688, + "grad_norm": 1.1637330649963575, + "learning_rate": 9.481117784326927e-06, + "loss": 0.5944, + "step": 1208 + }, + { + "epoch": 1.1603456553048488, + "grad_norm": 1.032629931610833, + "learning_rate": 9.479630727368927e-06, + "loss": 0.5966, + "step": 1209 + }, + { + "epoch": 1.1613058089294288, + "grad_norm": 0.8392640016288045, + "learning_rate": 9.47814165955258e-06, + "loss": 0.6257, + "step": 1210 + }, + { + "epoch": 1.1622659625540086, + "grad_norm": 1.2493910091463432, + "learning_rate": 9.47665058154631e-06, + "loss": 0.681, + "step": 1211 + }, + { + "epoch": 1.1632261161785886, + "grad_norm": 0.98863249817308, + "learning_rate": 9.475157494019445e-06, + "loss": 0.4885, + "step": 1212 + }, + { + "epoch": 1.1641862698031684, + "grad_norm": 1.2539998735969364, + "learning_rate": 9.473662397642219e-06, + "loss": 0.7908, + "step": 1213 + }, + { + "epoch": 1.1651464234277484, + "grad_norm": 0.9570118437597045, + "learning_rate": 9.472165293085765e-06, + "loss": 0.5632, + "step": 1214 + }, + { + "epoch": 1.1661065770523285, + "grad_norm": 1.3070034260563663, + "learning_rate": 9.470666181022114e-06, + "loss": 0.8193, + "step": 1215 + }, + { + "epoch": 1.1670667306769082, + "grad_norm": 0.9907941564598948, + "learning_rate": 9.469165062124203e-06, + "loss": 0.6063, + "step": 1216 + }, + { + "epoch": 1.1680268843014883, + "grad_norm": 1.2502899623231871, + "learning_rate": 9.467661937065869e-06, + "loss": 0.6428, + "step": 1217 + }, + { + "epoch": 1.168987037926068, + "grad_norm": 0.9611806461190867, + "learning_rate": 9.466156806521847e-06, + "loss": 0.6312, + "step": 1218 + }, + { + "epoch": 1.169947191550648, + "grad_norm": 0.9348007304355214, + "learning_rate": 9.464649671167777e-06, + "loss": 0.6423, + "step": 1219 + }, + { + "epoch": 1.170907345175228, + "grad_norm": 1.0777819064389393, + "learning_rate": 9.463140531680193e-06, + "loss": 0.6771, + "step": 1220 + }, + { + "epoch": 1.171867498799808, + "grad_norm": 1.180486328750565, + "learning_rate": 9.461629388736534e-06, + "loss": 0.687, + "step": 1221 + }, + { + "epoch": 1.172827652424388, + "grad_norm": 0.9790522632115352, + "learning_rate": 9.460116243015131e-06, + "loss": 0.6485, + "step": 1222 + }, + { + "epoch": 1.173787806048968, + "grad_norm": 1.4658160863700345, + "learning_rate": 9.458601095195225e-06, + "loss": 0.7031, + "step": 1223 + }, + { + "epoch": 1.1747479596735477, + "grad_norm": 0.9591722122695792, + "learning_rate": 9.45708394595695e-06, + "loss": 0.5725, + "step": 1224 + }, + { + "epoch": 1.1757081132981277, + "grad_norm": 0.9493091678959932, + "learning_rate": 9.455564795981331e-06, + "loss": 0.5762, + "step": 1225 + }, + { + "epoch": 1.1766682669227075, + "grad_norm": 0.9288315445755897, + "learning_rate": 9.454043645950303e-06, + "loss": 0.7323, + "step": 1226 + }, + { + "epoch": 1.1776284205472876, + "grad_norm": 1.3989016722710288, + "learning_rate": 9.452520496546693e-06, + "loss": 0.742, + "step": 1227 + }, + { + "epoch": 1.1785885741718676, + "grad_norm": 1.7939405129425552, + "learning_rate": 9.450995348454228e-06, + "loss": 0.7528, + "step": 1228 + }, + { + "epoch": 1.1795487277964474, + "grad_norm": 1.4912406316983478, + "learning_rate": 9.44946820235753e-06, + "loss": 0.7062, + "step": 1229 + }, + { + "epoch": 1.1805088814210274, + "grad_norm": 1.4089671456132038, + "learning_rate": 9.447939058942119e-06, + "loss": 0.7029, + "step": 1230 + }, + { + "epoch": 1.1814690350456072, + "grad_norm": 1.130975861916411, + "learning_rate": 9.446407918894408e-06, + "loss": 0.7265, + "step": 1231 + }, + { + "epoch": 1.1824291886701872, + "grad_norm": 1.4660041943418791, + "learning_rate": 9.444874782901713e-06, + "loss": 0.6933, + "step": 1232 + }, + { + "epoch": 1.1833893422947672, + "grad_norm": 1.1556454893411412, + "learning_rate": 9.443339651652241e-06, + "loss": 0.5965, + "step": 1233 + }, + { + "epoch": 1.184349495919347, + "grad_norm": 0.9125541225468079, + "learning_rate": 9.441802525835095e-06, + "loss": 0.663, + "step": 1234 + }, + { + "epoch": 1.185309649543927, + "grad_norm": 0.9050441635898483, + "learning_rate": 9.440263406140276e-06, + "loss": 0.6862, + "step": 1235 + }, + { + "epoch": 1.186269803168507, + "grad_norm": 1.0466940814511771, + "learning_rate": 9.438722293258678e-06, + "loss": 0.5805, + "step": 1236 + }, + { + "epoch": 1.1872299567930868, + "grad_norm": 1.0008127872163366, + "learning_rate": 9.43717918788209e-06, + "loss": 0.6455, + "step": 1237 + }, + { + "epoch": 1.1881901104176669, + "grad_norm": 1.113676512163136, + "learning_rate": 9.435634090703195e-06, + "loss": 0.6278, + "step": 1238 + }, + { + "epoch": 1.1891502640422467, + "grad_norm": 1.2375996678281171, + "learning_rate": 9.434087002415571e-06, + "loss": 0.6556, + "step": 1239 + }, + { + "epoch": 1.1901104176668267, + "grad_norm": 0.8065055434730838, + "learning_rate": 9.432537923713689e-06, + "loss": 0.5956, + "step": 1240 + }, + { + "epoch": 1.1910705712914067, + "grad_norm": 1.1982682927770063, + "learning_rate": 9.430986855292916e-06, + "loss": 0.6755, + "step": 1241 + }, + { + "epoch": 1.1920307249159865, + "grad_norm": 0.9063500599021055, + "learning_rate": 9.429433797849508e-06, + "loss": 0.6875, + "step": 1242 + }, + { + "epoch": 1.1929908785405665, + "grad_norm": 1.1804643784727347, + "learning_rate": 9.427878752080613e-06, + "loss": 0.7355, + "step": 1243 + }, + { + "epoch": 1.1939510321651463, + "grad_norm": 1.0823916368736504, + "learning_rate": 9.426321718684282e-06, + "loss": 0.6471, + "step": 1244 + }, + { + "epoch": 1.1949111857897263, + "grad_norm": 1.3597785702813805, + "learning_rate": 9.424762698359442e-06, + "loss": 0.6655, + "step": 1245 + }, + { + "epoch": 1.1958713394143063, + "grad_norm": 0.8598556636072228, + "learning_rate": 9.423201691805927e-06, + "loss": 0.5505, + "step": 1246 + }, + { + "epoch": 1.1968314930388861, + "grad_norm": 0.995912244969561, + "learning_rate": 9.421638699724453e-06, + "loss": 0.4913, + "step": 1247 + }, + { + "epoch": 1.1977916466634662, + "grad_norm": 0.8356833278020595, + "learning_rate": 9.420073722816634e-06, + "loss": 0.7199, + "step": 1248 + }, + { + "epoch": 1.1987518002880462, + "grad_norm": 1.5643960156153773, + "learning_rate": 9.418506761784967e-06, + "loss": 0.7136, + "step": 1249 + }, + { + "epoch": 1.199711953912626, + "grad_norm": 0.9077061515511775, + "learning_rate": 9.416937817332846e-06, + "loss": 0.6886, + "step": 1250 + }, + { + "epoch": 1.200672107537206, + "grad_norm": 0.9405685661726997, + "learning_rate": 9.415366890164553e-06, + "loss": 0.5351, + "step": 1251 + }, + { + "epoch": 1.201632261161786, + "grad_norm": 1.1406637017982817, + "learning_rate": 9.41379398098526e-06, + "loss": 0.74, + "step": 1252 + }, + { + "epoch": 1.2025924147863658, + "grad_norm": 1.0823585567677922, + "learning_rate": 9.41221909050103e-06, + "loss": 0.6435, + "step": 1253 + }, + { + "epoch": 1.2035525684109458, + "grad_norm": 0.935137312913638, + "learning_rate": 9.410642219418816e-06, + "loss": 0.645, + "step": 1254 + }, + { + "epoch": 1.2045127220355256, + "grad_norm": 1.1245325934238484, + "learning_rate": 9.409063368446454e-06, + "loss": 0.6627, + "step": 1255 + }, + { + "epoch": 1.2054728756601056, + "grad_norm": 0.8878362745856924, + "learning_rate": 9.407482538292679e-06, + "loss": 0.78, + "step": 1256 + }, + { + "epoch": 1.2064330292846854, + "grad_norm": 1.2618439111980329, + "learning_rate": 9.405899729667105e-06, + "loss": 0.7082, + "step": 1257 + }, + { + "epoch": 1.2073931829092654, + "grad_norm": 1.0146361786520939, + "learning_rate": 9.404314943280238e-06, + "loss": 0.6167, + "step": 1258 + }, + { + "epoch": 1.2083533365338455, + "grad_norm": 1.089613875702965, + "learning_rate": 9.402728179843472e-06, + "loss": 0.732, + "step": 1259 + }, + { + "epoch": 1.2093134901584253, + "grad_norm": 0.8991377642980493, + "learning_rate": 9.40113944006909e-06, + "loss": 0.5676, + "step": 1260 + }, + { + "epoch": 1.2102736437830053, + "grad_norm": 1.0111428725977558, + "learning_rate": 9.39954872467026e-06, + "loss": 0.6712, + "step": 1261 + }, + { + "epoch": 1.2112337974075853, + "grad_norm": 0.9909065293109987, + "learning_rate": 9.397956034361033e-06, + "loss": 0.6683, + "step": 1262 + }, + { + "epoch": 1.212193951032165, + "grad_norm": 1.1665268239010138, + "learning_rate": 9.396361369856356e-06, + "loss": 0.6287, + "step": 1263 + }, + { + "epoch": 1.2131541046567451, + "grad_norm": 1.054290739854059, + "learning_rate": 9.394764731872054e-06, + "loss": 0.6621, + "step": 1264 + }, + { + "epoch": 1.2141142582813251, + "grad_norm": 1.0898386457870217, + "learning_rate": 9.393166121124841e-06, + "loss": 0.5746, + "step": 1265 + }, + { + "epoch": 1.215074411905905, + "grad_norm": 0.9566676778756585, + "learning_rate": 9.391565538332317e-06, + "loss": 0.614, + "step": 1266 + }, + { + "epoch": 1.216034565530485, + "grad_norm": 0.8400299196696328, + "learning_rate": 9.389962984212966e-06, + "loss": 0.6769, + "step": 1267 + }, + { + "epoch": 1.2169947191550647, + "grad_norm": 1.2005798867695259, + "learning_rate": 9.388358459486156e-06, + "loss": 0.7829, + "step": 1268 + }, + { + "epoch": 1.2179548727796448, + "grad_norm": 1.2021657494576357, + "learning_rate": 9.386751964872144e-06, + "loss": 0.5903, + "step": 1269 + }, + { + "epoch": 1.2189150264042246, + "grad_norm": 1.1025463800635396, + "learning_rate": 9.385143501092066e-06, + "loss": 0.667, + "step": 1270 + }, + { + "epoch": 1.2198751800288046, + "grad_norm": 1.2542456165397613, + "learning_rate": 9.383533068867947e-06, + "loss": 0.5988, + "step": 1271 + }, + { + "epoch": 1.2208353336533846, + "grad_norm": 1.0585888442495783, + "learning_rate": 9.381920668922688e-06, + "loss": 0.707, + "step": 1272 + }, + { + "epoch": 1.2217954872779644, + "grad_norm": 0.9638408484966245, + "learning_rate": 9.380306301980082e-06, + "loss": 0.6781, + "step": 1273 + }, + { + "epoch": 1.2227556409025444, + "grad_norm": 1.0296720167271556, + "learning_rate": 9.378689968764801e-06, + "loss": 0.5824, + "step": 1274 + }, + { + "epoch": 1.2237157945271244, + "grad_norm": 0.8171310165074823, + "learning_rate": 9.377071670002398e-06, + "loss": 0.6445, + "step": 1275 + }, + { + "epoch": 1.2246759481517042, + "grad_norm": 1.1465404921282132, + "learning_rate": 9.37545140641931e-06, + "loss": 0.7128, + "step": 1276 + }, + { + "epoch": 1.2256361017762842, + "grad_norm": 1.2379480031020065, + "learning_rate": 9.373829178742857e-06, + "loss": 0.6726, + "step": 1277 + }, + { + "epoch": 1.2265962554008643, + "grad_norm": 0.9153325601276985, + "learning_rate": 9.37220498770124e-06, + "loss": 0.5812, + "step": 1278 + }, + { + "epoch": 1.227556409025444, + "grad_norm": 1.0918854791705137, + "learning_rate": 9.370578834023539e-06, + "loss": 0.5427, + "step": 1279 + }, + { + "epoch": 1.228516562650024, + "grad_norm": 1.107125474420226, + "learning_rate": 9.368950718439718e-06, + "loss": 0.7187, + "step": 1280 + }, + { + "epoch": 1.2294767162746039, + "grad_norm": 0.8644454521860808, + "learning_rate": 9.367320641680621e-06, + "loss": 0.6019, + "step": 1281 + }, + { + "epoch": 1.2304368698991839, + "grad_norm": 1.0501231561275055, + "learning_rate": 9.365688604477974e-06, + "loss": 0.6184, + "step": 1282 + }, + { + "epoch": 1.2313970235237637, + "grad_norm": 1.0060824341294845, + "learning_rate": 9.364054607564376e-06, + "loss": 0.6439, + "step": 1283 + }, + { + "epoch": 1.2323571771483437, + "grad_norm": 0.7443161665525969, + "learning_rate": 9.362418651673317e-06, + "loss": 0.5219, + "step": 1284 + }, + { + "epoch": 1.2333173307729237, + "grad_norm": 0.9703118618007911, + "learning_rate": 9.360780737539155e-06, + "loss": 0.6886, + "step": 1285 + }, + { + "epoch": 1.2342774843975035, + "grad_norm": 1.1677659082945249, + "learning_rate": 9.359140865897136e-06, + "loss": 0.6069, + "step": 1286 + }, + { + "epoch": 1.2352376380220835, + "grad_norm": 1.0592077878875672, + "learning_rate": 9.357499037483377e-06, + "loss": 0.544, + "step": 1287 + }, + { + "epoch": 1.2361977916466635, + "grad_norm": 0.8785899027528711, + "learning_rate": 9.35585525303488e-06, + "loss": 0.6578, + "step": 1288 + }, + { + "epoch": 1.2371579452712433, + "grad_norm": 1.4517995476557288, + "learning_rate": 9.35420951328952e-06, + "loss": 0.7605, + "step": 1289 + }, + { + "epoch": 1.2381180988958234, + "grad_norm": 1.2961008752247485, + "learning_rate": 9.352561818986056e-06, + "loss": 0.6525, + "step": 1290 + }, + { + "epoch": 1.2390782525204034, + "grad_norm": 1.4183167203059521, + "learning_rate": 9.350912170864116e-06, + "loss": 0.6921, + "step": 1291 + }, + { + "epoch": 1.2400384061449832, + "grad_norm": 1.6474252803532177, + "learning_rate": 9.349260569664212e-06, + "loss": 0.6892, + "step": 1292 + }, + { + "epoch": 1.2409985597695632, + "grad_norm": 0.7534610842432165, + "learning_rate": 9.347607016127728e-06, + "loss": 0.6579, + "step": 1293 + }, + { + "epoch": 1.241958713394143, + "grad_norm": 1.3560586177855898, + "learning_rate": 9.34595151099693e-06, + "loss": 0.5546, + "step": 1294 + }, + { + "epoch": 1.242918867018723, + "grad_norm": 1.2039054130735463, + "learning_rate": 9.344294055014952e-06, + "loss": 0.6263, + "step": 1295 + }, + { + "epoch": 1.243879020643303, + "grad_norm": 1.2559966060931804, + "learning_rate": 9.342634648925813e-06, + "loss": 0.7259, + "step": 1296 + }, + { + "epoch": 1.2448391742678828, + "grad_norm": 1.126033627440921, + "learning_rate": 9.340973293474397e-06, + "loss": 0.6782, + "step": 1297 + }, + { + "epoch": 1.2457993278924628, + "grad_norm": 1.2534282014031837, + "learning_rate": 9.339309989406474e-06, + "loss": 0.6615, + "step": 1298 + }, + { + "epoch": 1.2467594815170426, + "grad_norm": 1.0497702541081584, + "learning_rate": 9.337644737468682e-06, + "loss": 0.5481, + "step": 1299 + }, + { + "epoch": 1.2477196351416227, + "grad_norm": 0.6864879499147802, + "learning_rate": 9.335977538408533e-06, + "loss": 0.6377, + "step": 1300 + }, + { + "epoch": 1.2486797887662027, + "grad_norm": 1.004307893115873, + "learning_rate": 9.334308392974417e-06, + "loss": 0.5459, + "step": 1301 + }, + { + "epoch": 1.2496399423907825, + "grad_norm": 1.0347046897868124, + "learning_rate": 9.332637301915594e-06, + "loss": 0.6139, + "step": 1302 + }, + { + "epoch": 1.2506000960153625, + "grad_norm": 1.0273603906823152, + "learning_rate": 9.3309642659822e-06, + "loss": 0.5295, + "step": 1303 + }, + { + "epoch": 1.2515602496399425, + "grad_norm": 0.9256038563076083, + "learning_rate": 9.329289285925242e-06, + "loss": 0.6627, + "step": 1304 + }, + { + "epoch": 1.2525204032645223, + "grad_norm": 1.0665551164613651, + "learning_rate": 9.327612362496601e-06, + "loss": 0.5403, + "step": 1305 + }, + { + "epoch": 1.2534805568891023, + "grad_norm": 1.0496992791473883, + "learning_rate": 9.325933496449031e-06, + "loss": 0.6226, + "step": 1306 + }, + { + "epoch": 1.2544407105136821, + "grad_norm": 1.0038462694094532, + "learning_rate": 9.324252688536156e-06, + "loss": 0.6301, + "step": 1307 + }, + { + "epoch": 1.2554008641382621, + "grad_norm": 0.990212670969131, + "learning_rate": 9.322569939512472e-06, + "loss": 0.6669, + "step": 1308 + }, + { + "epoch": 1.256361017762842, + "grad_norm": 1.0100661698775886, + "learning_rate": 9.320885250133347e-06, + "loss": 0.6576, + "step": 1309 + }, + { + "epoch": 1.257321171387422, + "grad_norm": 0.9595990366799756, + "learning_rate": 9.319198621155022e-06, + "loss": 0.6108, + "step": 1310 + }, + { + "epoch": 1.258281325012002, + "grad_norm": 0.9932039061329729, + "learning_rate": 9.317510053334604e-06, + "loss": 0.7106, + "step": 1311 + }, + { + "epoch": 1.2592414786365818, + "grad_norm": 0.827379518177105, + "learning_rate": 9.315819547430077e-06, + "loss": 0.7086, + "step": 1312 + }, + { + "epoch": 1.2602016322611618, + "grad_norm": 0.9826599022429947, + "learning_rate": 9.314127104200287e-06, + "loss": 0.6515, + "step": 1313 + }, + { + "epoch": 1.2611617858857418, + "grad_norm": 0.88254688167717, + "learning_rate": 9.312432724404957e-06, + "loss": 0.5874, + "step": 1314 + }, + { + "epoch": 1.2621219395103216, + "grad_norm": 0.9292760810081079, + "learning_rate": 9.310736408804674e-06, + "loss": 0.621, + "step": 1315 + }, + { + "epoch": 1.2630820931349016, + "grad_norm": 1.1642100883411828, + "learning_rate": 9.309038158160896e-06, + "loss": 0.486, + "step": 1316 + }, + { + "epoch": 1.2640422467594816, + "grad_norm": 0.9674174348083485, + "learning_rate": 9.30733797323595e-06, + "loss": 0.6359, + "step": 1317 + }, + { + "epoch": 1.2650024003840614, + "grad_norm": 1.12103017950127, + "learning_rate": 9.305635854793031e-06, + "loss": 0.6539, + "step": 1318 + }, + { + "epoch": 1.2659625540086414, + "grad_norm": 0.9958390684899106, + "learning_rate": 9.3039318035962e-06, + "loss": 0.6644, + "step": 1319 + }, + { + "epoch": 1.2669227076332212, + "grad_norm": 1.7022252513712437, + "learning_rate": 9.30222582041039e-06, + "loss": 0.5404, + "step": 1320 + }, + { + "epoch": 1.2678828612578013, + "grad_norm": 1.134706650262642, + "learning_rate": 9.300517906001399e-06, + "loss": 0.7979, + "step": 1321 + }, + { + "epoch": 1.268843014882381, + "grad_norm": 1.2286866546646331, + "learning_rate": 9.298808061135887e-06, + "loss": 0.5412, + "step": 1322 + }, + { + "epoch": 1.269803168506961, + "grad_norm": 0.964610599153338, + "learning_rate": 9.29709628658139e-06, + "loss": 0.5561, + "step": 1323 + }, + { + "epoch": 1.270763322131541, + "grad_norm": 1.0421977021558635, + "learning_rate": 9.295382583106303e-06, + "loss": 0.596, + "step": 1324 + }, + { + "epoch": 1.2717234757561209, + "grad_norm": 1.1132921252971018, + "learning_rate": 9.29366695147989e-06, + "loss": 0.5767, + "step": 1325 + }, + { + "epoch": 1.272683629380701, + "grad_norm": 1.3239230868033052, + "learning_rate": 9.291949392472276e-06, + "loss": 0.6072, + "step": 1326 + }, + { + "epoch": 1.273643783005281, + "grad_norm": 1.576167318967835, + "learning_rate": 9.290229906854462e-06, + "loss": 0.7313, + "step": 1327 + }, + { + "epoch": 1.2746039366298607, + "grad_norm": 1.3610779506237412, + "learning_rate": 9.288508495398301e-06, + "loss": 0.5892, + "step": 1328 + }, + { + "epoch": 1.2755640902544407, + "grad_norm": 1.0519195317352252, + "learning_rate": 9.28678515887652e-06, + "loss": 0.6665, + "step": 1329 + }, + { + "epoch": 1.2765242438790207, + "grad_norm": 0.8333354204993307, + "learning_rate": 9.285059898062701e-06, + "loss": 0.5669, + "step": 1330 + }, + { + "epoch": 1.2774843975036005, + "grad_norm": 1.1052555150098762, + "learning_rate": 9.2833327137313e-06, + "loss": 0.7173, + "step": 1331 + }, + { + "epoch": 1.2784445511281806, + "grad_norm": 1.2239098417947563, + "learning_rate": 9.281603606657632e-06, + "loss": 0.5458, + "step": 1332 + }, + { + "epoch": 1.2794047047527606, + "grad_norm": 0.9706516508907063, + "learning_rate": 9.27987257761787e-06, + "loss": 0.6695, + "step": 1333 + }, + { + "epoch": 1.2803648583773404, + "grad_norm": 1.043908515727729, + "learning_rate": 9.27813962738906e-06, + "loss": 0.6605, + "step": 1334 + }, + { + "epoch": 1.2813250120019202, + "grad_norm": 0.9817932598491614, + "learning_rate": 9.2764047567491e-06, + "loss": 0.5666, + "step": 1335 + }, + { + "epoch": 1.2822851656265002, + "grad_norm": 1.3581727939152737, + "learning_rate": 9.274667966476757e-06, + "loss": 0.5455, + "step": 1336 + }, + { + "epoch": 1.2832453192510802, + "grad_norm": 0.9419152885129211, + "learning_rate": 9.272929257351659e-06, + "loss": 0.5651, + "step": 1337 + }, + { + "epoch": 1.28420547287566, + "grad_norm": 1.3104273037342988, + "learning_rate": 9.27118863015429e-06, + "loss": 0.6209, + "step": 1338 + }, + { + "epoch": 1.28516562650024, + "grad_norm": 1.0614631606828682, + "learning_rate": 9.269446085666002e-06, + "loss": 0.6588, + "step": 1339 + }, + { + "epoch": 1.28612578012482, + "grad_norm": 1.2821567442640949, + "learning_rate": 9.267701624669006e-06, + "loss": 0.6519, + "step": 1340 + }, + { + "epoch": 1.2870859337493998, + "grad_norm": 1.111204239556066, + "learning_rate": 9.265955247946369e-06, + "loss": 0.6144, + "step": 1341 + }, + { + "epoch": 1.2880460873739799, + "grad_norm": 1.095390947221734, + "learning_rate": 9.26420695628202e-06, + "loss": 0.7597, + "step": 1342 + }, + { + "epoch": 1.2890062409985599, + "grad_norm": 1.2377076600006347, + "learning_rate": 9.262456750460754e-06, + "loss": 0.6322, + "step": 1343 + }, + { + "epoch": 1.2899663946231397, + "grad_norm": 0.9512879954462602, + "learning_rate": 9.260704631268211e-06, + "loss": 0.6935, + "step": 1344 + }, + { + "epoch": 1.2909265482477197, + "grad_norm": 1.3026950771394807, + "learning_rate": 9.258950599490906e-06, + "loss": 0.6922, + "step": 1345 + }, + { + "epoch": 1.2918867018722997, + "grad_norm": 1.4224545115172562, + "learning_rate": 9.257194655916202e-06, + "loss": 0.6217, + "step": 1346 + }, + { + "epoch": 1.2928468554968795, + "grad_norm": 0.9423072453719439, + "learning_rate": 9.255436801332324e-06, + "loss": 0.6512, + "step": 1347 + }, + { + "epoch": 1.2938070091214593, + "grad_norm": 1.2651014767302182, + "learning_rate": 9.253677036528356e-06, + "loss": 0.7078, + "step": 1348 + }, + { + "epoch": 1.2947671627460393, + "grad_norm": 1.2056034429549998, + "learning_rate": 9.251915362294233e-06, + "loss": 0.6946, + "step": 1349 + }, + { + "epoch": 1.2957273163706193, + "grad_norm": 1.1626763350371463, + "learning_rate": 9.250151779420756e-06, + "loss": 0.6759, + "step": 1350 + }, + { + "epoch": 1.2966874699951991, + "grad_norm": 1.0954103915987892, + "learning_rate": 9.248386288699575e-06, + "loss": 0.712, + "step": 1351 + }, + { + "epoch": 1.2976476236197791, + "grad_norm": 1.308612743368172, + "learning_rate": 9.246618890923203e-06, + "loss": 0.6832, + "step": 1352 + }, + { + "epoch": 1.2986077772443592, + "grad_norm": 1.039110620513189, + "learning_rate": 9.244849586885005e-06, + "loss": 0.6366, + "step": 1353 + }, + { + "epoch": 1.299567930868939, + "grad_norm": 0.9978311166737291, + "learning_rate": 9.243078377379202e-06, + "loss": 0.6312, + "step": 1354 + }, + { + "epoch": 1.300528084493519, + "grad_norm": 0.9165498261707806, + "learning_rate": 9.24130526320087e-06, + "loss": 0.6052, + "step": 1355 + }, + { + "epoch": 1.301488238118099, + "grad_norm": 0.735063514985535, + "learning_rate": 9.239530245145945e-06, + "loss": 0.6612, + "step": 1356 + }, + { + "epoch": 1.3024483917426788, + "grad_norm": 1.0692782405428205, + "learning_rate": 9.237753324011208e-06, + "loss": 0.6754, + "step": 1357 + }, + { + "epoch": 1.3034085453672588, + "grad_norm": 1.5840209243711638, + "learning_rate": 9.235974500594305e-06, + "loss": 0.6874, + "step": 1358 + }, + { + "epoch": 1.3043686989918388, + "grad_norm": 1.3359743683261467, + "learning_rate": 9.234193775693728e-06, + "loss": 0.7557, + "step": 1359 + }, + { + "epoch": 1.3053288526164186, + "grad_norm": 1.3974632602774053, + "learning_rate": 9.232411150108826e-06, + "loss": 0.6611, + "step": 1360 + }, + { + "epoch": 1.3062890062409986, + "grad_norm": 1.0801888072428878, + "learning_rate": 9.2306266246398e-06, + "loss": 0.6115, + "step": 1361 + }, + { + "epoch": 1.3072491598655784, + "grad_norm": 1.1295783496536176, + "learning_rate": 9.228840200087708e-06, + "loss": 0.6626, + "step": 1362 + }, + { + "epoch": 1.3082093134901585, + "grad_norm": 1.261770127027789, + "learning_rate": 9.227051877254452e-06, + "loss": 0.6509, + "step": 1363 + }, + { + "epoch": 1.3091694671147383, + "grad_norm": 1.219139014858422, + "learning_rate": 9.225261656942792e-06, + "loss": 0.6737, + "step": 1364 + }, + { + "epoch": 1.3101296207393183, + "grad_norm": 1.3184165602906317, + "learning_rate": 9.22346953995634e-06, + "loss": 0.6225, + "step": 1365 + }, + { + "epoch": 1.3110897743638983, + "grad_norm": 1.002937899496408, + "learning_rate": 9.22167552709956e-06, + "loss": 0.6944, + "step": 1366 + }, + { + "epoch": 1.312049927988478, + "grad_norm": 1.4336954896809986, + "learning_rate": 9.219879619177761e-06, + "loss": 0.7621, + "step": 1367 + }, + { + "epoch": 1.313010081613058, + "grad_norm": 0.8527614630941144, + "learning_rate": 9.21808181699711e-06, + "loss": 0.5507, + "step": 1368 + }, + { + "epoch": 1.3139702352376381, + "grad_norm": 0.8727290588806392, + "learning_rate": 9.21628212136462e-06, + "loss": 0.632, + "step": 1369 + }, + { + "epoch": 1.314930388862218, + "grad_norm": 1.4499825499908416, + "learning_rate": 9.214480533088154e-06, + "loss": 0.7411, + "step": 1370 + }, + { + "epoch": 1.315890542486798, + "grad_norm": 1.3665233860507873, + "learning_rate": 9.212677052976428e-06, + "loss": 0.7025, + "step": 1371 + }, + { + "epoch": 1.316850696111378, + "grad_norm": 1.3370211780233987, + "learning_rate": 9.210871681839006e-06, + "loss": 0.6297, + "step": 1372 + }, + { + "epoch": 1.3178108497359577, + "grad_norm": 0.8810290539270131, + "learning_rate": 9.209064420486296e-06, + "loss": 0.5714, + "step": 1373 + }, + { + "epoch": 1.3187710033605378, + "grad_norm": 0.8972394920514724, + "learning_rate": 9.207255269729559e-06, + "loss": 0.5948, + "step": 1374 + }, + { + "epoch": 1.3197311569851176, + "grad_norm": 1.4565885202363469, + "learning_rate": 9.205444230380904e-06, + "loss": 0.7471, + "step": 1375 + }, + { + "epoch": 1.3206913106096976, + "grad_norm": 1.3602857157125754, + "learning_rate": 9.203631303253292e-06, + "loss": 0.6143, + "step": 1376 + }, + { + "epoch": 1.3216514642342774, + "grad_norm": 1.2348121608011493, + "learning_rate": 9.201816489160518e-06, + "loss": 0.7407, + "step": 1377 + }, + { + "epoch": 1.3226116178588574, + "grad_norm": 1.1677390321780012, + "learning_rate": 9.199999788917238e-06, + "loss": 0.7074, + "step": 1378 + }, + { + "epoch": 1.3235717714834374, + "grad_norm": 1.6220671969702665, + "learning_rate": 9.198181203338949e-06, + "loss": 0.672, + "step": 1379 + }, + { + "epoch": 1.3245319251080172, + "grad_norm": 1.55816953971552, + "learning_rate": 9.196360733241992e-06, + "loss": 0.5813, + "step": 1380 + }, + { + "epoch": 1.3254920787325972, + "grad_norm": 0.6031797398946008, + "learning_rate": 9.194538379443561e-06, + "loss": 0.5942, + "step": 1381 + }, + { + "epoch": 1.3264522323571772, + "grad_norm": 1.0209739299888865, + "learning_rate": 9.192714142761687e-06, + "loss": 0.5466, + "step": 1382 + }, + { + "epoch": 1.327412385981757, + "grad_norm": 1.0685918514092403, + "learning_rate": 9.190888024015252e-06, + "loss": 0.7386, + "step": 1383 + }, + { + "epoch": 1.328372539606337, + "grad_norm": 2.4654106901615123, + "learning_rate": 9.189060024023981e-06, + "loss": 0.6167, + "step": 1384 + }, + { + "epoch": 1.329332693230917, + "grad_norm": 0.9254052338620629, + "learning_rate": 9.187230143608445e-06, + "loss": 0.7293, + "step": 1385 + }, + { + "epoch": 1.3302928468554969, + "grad_norm": 0.8997768367015149, + "learning_rate": 9.185398383590056e-06, + "loss": 0.6735, + "step": 1386 + }, + { + "epoch": 1.3312530004800769, + "grad_norm": 0.8765280760171718, + "learning_rate": 9.18356474479107e-06, + "loss": 0.7112, + "step": 1387 + }, + { + "epoch": 1.3322131541046567, + "grad_norm": 1.0199429321891622, + "learning_rate": 9.181729228034593e-06, + "loss": 0.6859, + "step": 1388 + }, + { + "epoch": 1.3331733077292367, + "grad_norm": 1.156182337546054, + "learning_rate": 9.179891834144565e-06, + "loss": 0.5132, + "step": 1389 + }, + { + "epoch": 1.3341334613538165, + "grad_norm": 1.1127871788424046, + "learning_rate": 9.178052563945773e-06, + "loss": 0.6018, + "step": 1390 + }, + { + "epoch": 1.3350936149783965, + "grad_norm": 1.0718387568033312, + "learning_rate": 9.176211418263845e-06, + "loss": 0.6811, + "step": 1391 + }, + { + "epoch": 1.3360537686029765, + "grad_norm": 1.6370091625480498, + "learning_rate": 9.174368397925254e-06, + "loss": 0.5976, + "step": 1392 + }, + { + "epoch": 1.3370139222275563, + "grad_norm": 1.3393311738083604, + "learning_rate": 9.17252350375731e-06, + "loss": 0.5573, + "step": 1393 + }, + { + "epoch": 1.3379740758521363, + "grad_norm": 0.9440285509468719, + "learning_rate": 9.170676736588167e-06, + "loss": 0.6261, + "step": 1394 + }, + { + "epoch": 1.3389342294767164, + "grad_norm": 0.8518148709205811, + "learning_rate": 9.168828097246819e-06, + "loss": 0.5317, + "step": 1395 + }, + { + "epoch": 1.3398943831012962, + "grad_norm": 1.001207334694378, + "learning_rate": 9.1669775865631e-06, + "loss": 0.647, + "step": 1396 + }, + { + "epoch": 1.3408545367258762, + "grad_norm": 0.8855183613979828, + "learning_rate": 9.165125205367685e-06, + "loss": 0.5526, + "step": 1397 + }, + { + "epoch": 1.3418146903504562, + "grad_norm": 0.8646748349945449, + "learning_rate": 9.16327095449209e-06, + "loss": 0.6218, + "step": 1398 + }, + { + "epoch": 1.342774843975036, + "grad_norm": 0.9207173830398178, + "learning_rate": 9.161414834768662e-06, + "loss": 0.679, + "step": 1399 + }, + { + "epoch": 1.343734997599616, + "grad_norm": 1.0246811568814804, + "learning_rate": 9.159556847030602e-06, + "loss": 0.6315, + "step": 1400 + }, + { + "epoch": 1.3446951512241958, + "grad_norm": 1.0399275691034016, + "learning_rate": 9.157696992111935e-06, + "loss": 0.5397, + "step": 1401 + }, + { + "epoch": 1.3456553048487758, + "grad_norm": 1.0881315389498556, + "learning_rate": 9.15583527084753e-06, + "loss": 0.7734, + "step": 1402 + }, + { + "epoch": 1.3466154584733556, + "grad_norm": 1.1227693560178058, + "learning_rate": 9.153971684073098e-06, + "loss": 0.6873, + "step": 1403 + }, + { + "epoch": 1.3475756120979356, + "grad_norm": 1.1593023174123618, + "learning_rate": 9.15210623262518e-06, + "loss": 0.7518, + "step": 1404 + }, + { + "epoch": 1.3485357657225157, + "grad_norm": 1.3285923577202956, + "learning_rate": 9.150238917341156e-06, + "loss": 0.6111, + "step": 1405 + }, + { + "epoch": 1.3494959193470955, + "grad_norm": 1.3982348325270315, + "learning_rate": 9.148369739059247e-06, + "loss": 0.6599, + "step": 1406 + }, + { + "epoch": 1.3504560729716755, + "grad_norm": 1.2849899368598579, + "learning_rate": 9.146498698618507e-06, + "loss": 0.6613, + "step": 1407 + }, + { + "epoch": 1.3514162265962555, + "grad_norm": 1.335510666991283, + "learning_rate": 9.14462579685882e-06, + "loss": 0.5636, + "step": 1408 + }, + { + "epoch": 1.3523763802208353, + "grad_norm": 1.1259159335740894, + "learning_rate": 9.142751034620923e-06, + "loss": 0.539, + "step": 1409 + }, + { + "epoch": 1.3533365338454153, + "grad_norm": 1.423562344991987, + "learning_rate": 9.140874412746365e-06, + "loss": 0.6782, + "step": 1410 + }, + { + "epoch": 1.3542966874699953, + "grad_norm": 0.9815155519301481, + "learning_rate": 9.138995932077549e-06, + "loss": 0.5981, + "step": 1411 + }, + { + "epoch": 1.3552568410945751, + "grad_norm": 1.5764550806390785, + "learning_rate": 9.137115593457702e-06, + "loss": 0.7785, + "step": 1412 + }, + { + "epoch": 1.3562169947191551, + "grad_norm": 1.3578567713502319, + "learning_rate": 9.135233397730889e-06, + "loss": 0.6807, + "step": 1413 + }, + { + "epoch": 1.357177148343735, + "grad_norm": 1.4123462218723584, + "learning_rate": 9.133349345742007e-06, + "loss": 0.8146, + "step": 1414 + }, + { + "epoch": 1.358137301968315, + "grad_norm": 1.3656080836677802, + "learning_rate": 9.131463438336785e-06, + "loss": 0.6289, + "step": 1415 + }, + { + "epoch": 1.3590974555928947, + "grad_norm": 0.984971772562819, + "learning_rate": 9.12957567636179e-06, + "loss": 0.6799, + "step": 1416 + }, + { + "epoch": 1.3600576092174748, + "grad_norm": 1.0615120744089255, + "learning_rate": 9.127686060664414e-06, + "loss": 0.645, + "step": 1417 + }, + { + "epoch": 1.3610177628420548, + "grad_norm": 0.7357874833677214, + "learning_rate": 9.125794592092891e-06, + "loss": 0.6548, + "step": 1418 + }, + { + "epoch": 1.3619779164666346, + "grad_norm": 0.8780975777003137, + "learning_rate": 9.123901271496276e-06, + "loss": 0.5611, + "step": 1419 + }, + { + "epoch": 1.3629380700912146, + "grad_norm": 0.8112244447711271, + "learning_rate": 9.122006099724463e-06, + "loss": 0.5917, + "step": 1420 + }, + { + "epoch": 1.3638982237157946, + "grad_norm": 1.221144802300544, + "learning_rate": 9.120109077628175e-06, + "loss": 0.6999, + "step": 1421 + }, + { + "epoch": 1.3648583773403744, + "grad_norm": 1.1403038947834057, + "learning_rate": 9.118210206058961e-06, + "loss": 0.6721, + "step": 1422 + }, + { + "epoch": 1.3658185309649544, + "grad_norm": 0.9730116515747882, + "learning_rate": 9.116309485869207e-06, + "loss": 0.6044, + "step": 1423 + }, + { + "epoch": 1.3667786845895344, + "grad_norm": 0.9493687999626979, + "learning_rate": 9.114406917912129e-06, + "loss": 0.6536, + "step": 1424 + }, + { + "epoch": 1.3677388382141142, + "grad_norm": 1.1547885016150892, + "learning_rate": 9.112502503041763e-06, + "loss": 0.5292, + "step": 1425 + }, + { + "epoch": 1.3686989918386943, + "grad_norm": 1.1910868300541848, + "learning_rate": 9.110596242112985e-06, + "loss": 0.6713, + "step": 1426 + }, + { + "epoch": 1.369659145463274, + "grad_norm": 0.9915731479060182, + "learning_rate": 9.108688135981494e-06, + "loss": 0.5338, + "step": 1427 + }, + { + "epoch": 1.370619299087854, + "grad_norm": 1.1633320943761714, + "learning_rate": 9.106778185503816e-06, + "loss": 0.6207, + "step": 1428 + }, + { + "epoch": 1.3715794527124339, + "grad_norm": 1.1229228094327428, + "learning_rate": 9.10486639153731e-06, + "loss": 0.6682, + "step": 1429 + }, + { + "epoch": 1.3725396063370139, + "grad_norm": 1.1359696559124128, + "learning_rate": 9.102952754940162e-06, + "loss": 0.6433, + "step": 1430 + }, + { + "epoch": 1.373499759961594, + "grad_norm": 1.1452071263196368, + "learning_rate": 9.101037276571378e-06, + "loss": 0.7461, + "step": 1431 + }, + { + "epoch": 1.3744599135861737, + "grad_norm": 1.0217787726374161, + "learning_rate": 9.099119957290798e-06, + "loss": 0.6545, + "step": 1432 + }, + { + "epoch": 1.3754200672107537, + "grad_norm": 1.1024321646787556, + "learning_rate": 9.097200797959089e-06, + "loss": 0.6808, + "step": 1433 + }, + { + "epoch": 1.3763802208353337, + "grad_norm": 1.2041112611388614, + "learning_rate": 9.095279799437737e-06, + "loss": 0.6022, + "step": 1434 + }, + { + "epoch": 1.3773403744599135, + "grad_norm": 1.245185695040958, + "learning_rate": 9.09335696258906e-06, + "loss": 0.6416, + "step": 1435 + }, + { + "epoch": 1.3783005280844935, + "grad_norm": 1.5245819538936007, + "learning_rate": 9.091432288276196e-06, + "loss": 0.6677, + "step": 1436 + }, + { + "epoch": 1.3792606817090736, + "grad_norm": 1.1315279970015388, + "learning_rate": 9.089505777363114e-06, + "loss": 0.5676, + "step": 1437 + }, + { + "epoch": 1.3802208353336534, + "grad_norm": 1.0880311134579044, + "learning_rate": 9.087577430714603e-06, + "loss": 0.5323, + "step": 1438 + }, + { + "epoch": 1.3811809889582334, + "grad_norm": 0.7997114854628721, + "learning_rate": 9.085647249196275e-06, + "loss": 0.6705, + "step": 1439 + }, + { + "epoch": 1.3821411425828132, + "grad_norm": 1.1168457432733025, + "learning_rate": 9.083715233674572e-06, + "loss": 0.5536, + "step": 1440 + }, + { + "epoch": 1.3831012962073932, + "grad_norm": 1.1030453093536399, + "learning_rate": 9.081781385016754e-06, + "loss": 0.6965, + "step": 1441 + }, + { + "epoch": 1.384061449831973, + "grad_norm": 1.112953806581517, + "learning_rate": 9.079845704090903e-06, + "loss": 0.6036, + "step": 1442 + }, + { + "epoch": 1.385021603456553, + "grad_norm": 1.089864089498288, + "learning_rate": 9.077908191765925e-06, + "loss": 0.676, + "step": 1443 + }, + { + "epoch": 1.385981757081133, + "grad_norm": 0.7941147593055814, + "learning_rate": 9.075968848911553e-06, + "loss": 0.6765, + "step": 1444 + }, + { + "epoch": 1.3869419107057128, + "grad_norm": 1.1538336794588986, + "learning_rate": 9.074027676398333e-06, + "loss": 0.5535, + "step": 1445 + }, + { + "epoch": 1.3879020643302928, + "grad_norm": 0.899873624460216, + "learning_rate": 9.07208467509764e-06, + "loss": 0.5086, + "step": 1446 + }, + { + "epoch": 1.3888622179548729, + "grad_norm": 1.2441712252685813, + "learning_rate": 9.07013984588166e-06, + "loss": 0.8172, + "step": 1447 + }, + { + "epoch": 1.3898223715794527, + "grad_norm": 1.1567695863868372, + "learning_rate": 9.068193189623412e-06, + "loss": 0.6214, + "step": 1448 + }, + { + "epoch": 1.3907825252040327, + "grad_norm": 1.1708998754473148, + "learning_rate": 9.06624470719673e-06, + "loss": 0.5282, + "step": 1449 + }, + { + "epoch": 1.3917426788286127, + "grad_norm": 0.9026111959152648, + "learning_rate": 9.064294399476265e-06, + "loss": 0.5809, + "step": 1450 + }, + { + "epoch": 1.3927028324531925, + "grad_norm": 1.2040360577373892, + "learning_rate": 9.062342267337487e-06, + "loss": 0.5999, + "step": 1451 + }, + { + "epoch": 1.3936629860777725, + "grad_norm": 1.1096935773411545, + "learning_rate": 9.060388311656689e-06, + "loss": 0.6106, + "step": 1452 + }, + { + "epoch": 1.3946231397023523, + "grad_norm": 1.0629999388500475, + "learning_rate": 9.058432533310982e-06, + "loss": 0.6841, + "step": 1453 + }, + { + "epoch": 1.3955832933269323, + "grad_norm": 1.2674360481064615, + "learning_rate": 9.056474933178293e-06, + "loss": 0.7552, + "step": 1454 + }, + { + "epoch": 1.3965434469515121, + "grad_norm": 1.3704949314290888, + "learning_rate": 9.054515512137367e-06, + "loss": 0.5866, + "step": 1455 + }, + { + "epoch": 1.3975036005760921, + "grad_norm": 0.9070807201679557, + "learning_rate": 9.052554271067771e-06, + "loss": 0.5786, + "step": 1456 + }, + { + "epoch": 1.3984637542006721, + "grad_norm": 1.0367508024077463, + "learning_rate": 9.05059121084988e-06, + "loss": 0.5443, + "step": 1457 + }, + { + "epoch": 1.399423907825252, + "grad_norm": 0.8228263048244029, + "learning_rate": 9.048626332364891e-06, + "loss": 0.665, + "step": 1458 + }, + { + "epoch": 1.400384061449832, + "grad_norm": 1.0817115873933376, + "learning_rate": 9.046659636494821e-06, + "loss": 0.5638, + "step": 1459 + }, + { + "epoch": 1.401344215074412, + "grad_norm": 1.1415393174650825, + "learning_rate": 9.044691124122496e-06, + "loss": 0.7456, + "step": 1460 + }, + { + "epoch": 1.4023043686989918, + "grad_norm": 1.042080496976787, + "learning_rate": 9.042720796131561e-06, + "loss": 0.6686, + "step": 1461 + }, + { + "epoch": 1.4032645223235718, + "grad_norm": 1.1222340889332822, + "learning_rate": 9.040748653406475e-06, + "loss": 0.752, + "step": 1462 + }, + { + "epoch": 1.4042246759481518, + "grad_norm": 1.3848173780245217, + "learning_rate": 9.038774696832511e-06, + "loss": 0.807, + "step": 1463 + }, + { + "epoch": 1.4051848295727316, + "grad_norm": 1.1288431218857278, + "learning_rate": 9.036798927295757e-06, + "loss": 0.7102, + "step": 1464 + }, + { + "epoch": 1.4061449831973116, + "grad_norm": 1.01517480041356, + "learning_rate": 9.034821345683117e-06, + "loss": 0.5792, + "step": 1465 + }, + { + "epoch": 1.4071051368218914, + "grad_norm": 1.0674838336963253, + "learning_rate": 9.032841952882306e-06, + "loss": 0.6886, + "step": 1466 + }, + { + "epoch": 1.4080652904464714, + "grad_norm": 1.2617241953266387, + "learning_rate": 9.030860749781848e-06, + "loss": 0.6348, + "step": 1467 + }, + { + "epoch": 1.4090254440710512, + "grad_norm": 1.034571044333529, + "learning_rate": 9.028877737271089e-06, + "loss": 0.5829, + "step": 1468 + }, + { + "epoch": 1.4099855976956313, + "grad_norm": 1.0033035253032856, + "learning_rate": 9.02689291624018e-06, + "loss": 0.7316, + "step": 1469 + }, + { + "epoch": 1.4109457513202113, + "grad_norm": 1.4016124783738984, + "learning_rate": 9.024906287580085e-06, + "loss": 0.6421, + "step": 1470 + }, + { + "epoch": 1.411905904944791, + "grad_norm": 0.9863253720673301, + "learning_rate": 9.022917852182582e-06, + "loss": 0.5401, + "step": 1471 + }, + { + "epoch": 1.412866058569371, + "grad_norm": 1.754467437274913, + "learning_rate": 9.020927610940256e-06, + "loss": 0.7443, + "step": 1472 + }, + { + "epoch": 1.413826212193951, + "grad_norm": 1.2353265052336058, + "learning_rate": 9.018935564746509e-06, + "loss": 0.6328, + "step": 1473 + }, + { + "epoch": 1.414786365818531, + "grad_norm": 1.0963681138104033, + "learning_rate": 9.016941714495546e-06, + "loss": 0.6851, + "step": 1474 + }, + { + "epoch": 1.415746519443111, + "grad_norm": 1.3878482660616152, + "learning_rate": 9.014946061082385e-06, + "loss": 0.747, + "step": 1475 + }, + { + "epoch": 1.416706673067691, + "grad_norm": 1.0494981175572002, + "learning_rate": 9.012948605402856e-06, + "loss": 0.6898, + "step": 1476 + }, + { + "epoch": 1.4176668266922707, + "grad_norm": 1.2387679013385462, + "learning_rate": 9.010949348353595e-06, + "loss": 0.6006, + "step": 1477 + }, + { + "epoch": 1.4186269803168507, + "grad_norm": 1.2177631267117774, + "learning_rate": 9.008948290832046e-06, + "loss": 0.8044, + "step": 1478 + }, + { + "epoch": 1.4195871339414305, + "grad_norm": 1.1236670782980807, + "learning_rate": 9.006945433736461e-06, + "loss": 0.4909, + "step": 1479 + }, + { + "epoch": 1.4205472875660106, + "grad_norm": 1.1811146156755659, + "learning_rate": 9.004940777965903e-06, + "loss": 0.6744, + "step": 1480 + }, + { + "epoch": 1.4215074411905904, + "grad_norm": 1.2952915847830806, + "learning_rate": 9.002934324420242e-06, + "loss": 0.8284, + "step": 1481 + }, + { + "epoch": 1.4224675948151704, + "grad_norm": 1.3186626780304642, + "learning_rate": 9.000926074000149e-06, + "loss": 0.6409, + "step": 1482 + }, + { + "epoch": 1.4234277484397504, + "grad_norm": 1.1083243673963934, + "learning_rate": 8.998916027607111e-06, + "loss": 0.6272, + "step": 1483 + }, + { + "epoch": 1.4243879020643302, + "grad_norm": 1.1486576286510621, + "learning_rate": 8.996904186143414e-06, + "loss": 0.6471, + "step": 1484 + }, + { + "epoch": 1.4253480556889102, + "grad_norm": 1.427612957312725, + "learning_rate": 8.994890550512152e-06, + "loss": 0.6782, + "step": 1485 + }, + { + "epoch": 1.4263082093134902, + "grad_norm": 1.2421048458874155, + "learning_rate": 8.992875121617224e-06, + "loss": 0.6714, + "step": 1486 + }, + { + "epoch": 1.42726836293807, + "grad_norm": 1.1475684942018722, + "learning_rate": 8.990857900363337e-06, + "loss": 0.5341, + "step": 1487 + }, + { + "epoch": 1.42822851656265, + "grad_norm": 0.7789219831356208, + "learning_rate": 8.988838887655997e-06, + "loss": 0.6485, + "step": 1488 + }, + { + "epoch": 1.42918867018723, + "grad_norm": 0.9022544631477947, + "learning_rate": 8.98681808440152e-06, + "loss": 0.4919, + "step": 1489 + }, + { + "epoch": 1.4301488238118099, + "grad_norm": 1.0487182980001104, + "learning_rate": 8.984795491507021e-06, + "loss": 0.5778, + "step": 1490 + }, + { + "epoch": 1.4311089774363899, + "grad_norm": 0.7765690777775737, + "learning_rate": 8.982771109880421e-06, + "loss": 0.6641, + "step": 1491 + }, + { + "epoch": 1.43206913106097, + "grad_norm": 0.9770150771741994, + "learning_rate": 8.980744940430443e-06, + "loss": 0.5373, + "step": 1492 + }, + { + "epoch": 1.4330292846855497, + "grad_norm": 1.0842856980753282, + "learning_rate": 8.978716984066613e-06, + "loss": 0.6931, + "step": 1493 + }, + { + "epoch": 1.4339894383101295, + "grad_norm": 1.5905291871795488, + "learning_rate": 8.976687241699258e-06, + "loss": 0.6763, + "step": 1494 + }, + { + "epoch": 1.4349495919347095, + "grad_norm": 1.1291160885943465, + "learning_rate": 8.97465571423951e-06, + "loss": 0.7141, + "step": 1495 + }, + { + "epoch": 1.4359097455592895, + "grad_norm": 0.8847319400098621, + "learning_rate": 8.972622402599298e-06, + "loss": 0.6199, + "step": 1496 + }, + { + "epoch": 1.4368698991838693, + "grad_norm": 1.0193428575785228, + "learning_rate": 8.970587307691357e-06, + "loss": 0.5403, + "step": 1497 + }, + { + "epoch": 1.4378300528084493, + "grad_norm": 1.0872663796559396, + "learning_rate": 8.968550430429212e-06, + "loss": 0.5434, + "step": 1498 + }, + { + "epoch": 1.4387902064330294, + "grad_norm": 1.2406773363844592, + "learning_rate": 8.966511771727205e-06, + "loss": 0.5879, + "step": 1499 + }, + { + "epoch": 1.4397503600576091, + "grad_norm": 0.8156024523464949, + "learning_rate": 8.96447133250046e-06, + "loss": 0.6312, + "step": 1500 + }, + { + "epoch": 1.4407105136821892, + "grad_norm": 1.257440132329098, + "learning_rate": 8.962429113664913e-06, + "loss": 0.6966, + "step": 1501 + }, + { + "epoch": 1.4416706673067692, + "grad_norm": 1.0628200522245184, + "learning_rate": 8.960385116137293e-06, + "loss": 0.6327, + "step": 1502 + }, + { + "epoch": 1.442630820931349, + "grad_norm": 1.1596260122624005, + "learning_rate": 8.958339340835128e-06, + "loss": 0.5703, + "step": 1503 + }, + { + "epoch": 1.443590974555929, + "grad_norm": 1.1298747586240705, + "learning_rate": 8.956291788676746e-06, + "loss": 0.5657, + "step": 1504 + }, + { + "epoch": 1.444551128180509, + "grad_norm": 0.8397355051213713, + "learning_rate": 8.95424246058127e-06, + "loss": 0.6697, + "step": 1505 + }, + { + "epoch": 1.4455112818050888, + "grad_norm": 1.0161743613809866, + "learning_rate": 8.952191357468621e-06, + "loss": 0.5619, + "step": 1506 + }, + { + "epoch": 1.4464714354296688, + "grad_norm": 1.2106464122559442, + "learning_rate": 8.950138480259519e-06, + "loss": 0.57, + "step": 1507 + }, + { + "epoch": 1.4474315890542486, + "grad_norm": 1.1265109245805303, + "learning_rate": 8.948083829875478e-06, + "loss": 0.5467, + "step": 1508 + }, + { + "epoch": 1.4483917426788286, + "grad_norm": 1.0933628766801322, + "learning_rate": 8.946027407238809e-06, + "loss": 0.588, + "step": 1509 + }, + { + "epoch": 1.4493518963034084, + "grad_norm": 1.104099886204487, + "learning_rate": 8.943969213272616e-06, + "loss": 0.6103, + "step": 1510 + }, + { + "epoch": 1.4503120499279885, + "grad_norm": 0.9561757908407287, + "learning_rate": 8.941909248900804e-06, + "loss": 0.5385, + "step": 1511 + }, + { + "epoch": 1.4512722035525685, + "grad_norm": 1.1169073689054023, + "learning_rate": 8.939847515048066e-06, + "loss": 0.6164, + "step": 1512 + }, + { + "epoch": 1.4522323571771483, + "grad_norm": 0.8551643979571723, + "learning_rate": 8.937784012639892e-06, + "loss": 0.676, + "step": 1513 + }, + { + "epoch": 1.4531925108017283, + "grad_norm": 1.0042682995889094, + "learning_rate": 8.935718742602568e-06, + "loss": 0.6287, + "step": 1514 + }, + { + "epoch": 1.4541526644263083, + "grad_norm": 1.0614362268942157, + "learning_rate": 8.933651705863172e-06, + "loss": 0.7373, + "step": 1515 + }, + { + "epoch": 1.455112818050888, + "grad_norm": 1.3401565218668898, + "learning_rate": 8.931582903349571e-06, + "loss": 0.5156, + "step": 1516 + }, + { + "epoch": 1.4560729716754681, + "grad_norm": 1.2341123257068967, + "learning_rate": 8.929512335990434e-06, + "loss": 0.7821, + "step": 1517 + }, + { + "epoch": 1.4570331253000481, + "grad_norm": 1.7078266511969085, + "learning_rate": 8.92744000471521e-06, + "loss": 0.6796, + "step": 1518 + }, + { + "epoch": 1.457993278924628, + "grad_norm": 1.4104645213806324, + "learning_rate": 8.925365910454153e-06, + "loss": 0.5957, + "step": 1519 + }, + { + "epoch": 1.458953432549208, + "grad_norm": 1.172908486399178, + "learning_rate": 8.923290054138298e-06, + "loss": 0.5026, + "step": 1520 + }, + { + "epoch": 1.4599135861737877, + "grad_norm": 0.9886504176722553, + "learning_rate": 8.921212436699476e-06, + "loss": 0.5457, + "step": 1521 + }, + { + "epoch": 1.4608737397983678, + "grad_norm": 0.7259194922521678, + "learning_rate": 8.919133059070307e-06, + "loss": 0.5767, + "step": 1522 + }, + { + "epoch": 1.4618338934229476, + "grad_norm": 1.1329123474845246, + "learning_rate": 8.9170519221842e-06, + "loss": 0.6652, + "step": 1523 + }, + { + "epoch": 1.4627940470475276, + "grad_norm": 1.3825900560877011, + "learning_rate": 8.914969026975354e-06, + "loss": 0.65, + "step": 1524 + }, + { + "epoch": 1.4637542006721076, + "grad_norm": 1.6252305372983962, + "learning_rate": 8.912884374378763e-06, + "loss": 0.6646, + "step": 1525 + }, + { + "epoch": 1.4647143542966874, + "grad_norm": 1.1724088736710376, + "learning_rate": 8.910797965330202e-06, + "loss": 0.5923, + "step": 1526 + }, + { + "epoch": 1.4656745079212674, + "grad_norm": 1.3940002698365594, + "learning_rate": 8.908709800766237e-06, + "loss": 0.73, + "step": 1527 + }, + { + "epoch": 1.4666346615458474, + "grad_norm": 1.2639384312093593, + "learning_rate": 8.906619881624225e-06, + "loss": 0.6578, + "step": 1528 + }, + { + "epoch": 1.4675948151704272, + "grad_norm": 1.0940793340621946, + "learning_rate": 8.90452820884231e-06, + "loss": 0.607, + "step": 1529 + }, + { + "epoch": 1.4685549687950072, + "grad_norm": 1.0542543562205764, + "learning_rate": 8.902434783359417e-06, + "loss": 0.6936, + "step": 1530 + }, + { + "epoch": 1.4695151224195873, + "grad_norm": 1.1879881235521863, + "learning_rate": 8.900339606115264e-06, + "loss": 0.6887, + "step": 1531 + }, + { + "epoch": 1.470475276044167, + "grad_norm": 0.8757796270384739, + "learning_rate": 8.898242678050354e-06, + "loss": 0.7135, + "step": 1532 + }, + { + "epoch": 1.471435429668747, + "grad_norm": 1.0409790443247378, + "learning_rate": 8.896144000105979e-06, + "loss": 0.5364, + "step": 1533 + }, + { + "epoch": 1.4723955832933269, + "grad_norm": 0.8076482116173296, + "learning_rate": 8.894043573224207e-06, + "loss": 0.6703, + "step": 1534 + }, + { + "epoch": 1.4733557369179069, + "grad_norm": 1.0625925178859534, + "learning_rate": 8.8919413983479e-06, + "loss": 0.5318, + "step": 1535 + }, + { + "epoch": 1.4743158905424867, + "grad_norm": 1.049571449075552, + "learning_rate": 8.889837476420703e-06, + "loss": 0.6438, + "step": 1536 + }, + { + "epoch": 1.4752760441670667, + "grad_norm": 0.8197834071354191, + "learning_rate": 8.887731808387043e-06, + "loss": 0.7928, + "step": 1537 + }, + { + "epoch": 1.4762361977916467, + "grad_norm": 1.0166060432466146, + "learning_rate": 8.885624395192131e-06, + "loss": 0.7061, + "step": 1538 + }, + { + "epoch": 1.4771963514162265, + "grad_norm": 1.129472818971423, + "learning_rate": 8.883515237781964e-06, + "loss": 0.6651, + "step": 1539 + }, + { + "epoch": 1.4781565050408065, + "grad_norm": 0.9573464426762843, + "learning_rate": 8.881404337103316e-06, + "loss": 0.7077, + "step": 1540 + }, + { + "epoch": 1.4791166586653866, + "grad_norm": 1.1004021355840239, + "learning_rate": 8.879291694103754e-06, + "loss": 0.5878, + "step": 1541 + }, + { + "epoch": 1.4800768122899663, + "grad_norm": 1.178159407699785, + "learning_rate": 8.877177309731618e-06, + "loss": 0.5575, + "step": 1542 + }, + { + "epoch": 1.4810369659145464, + "grad_norm": 1.0159381485089358, + "learning_rate": 8.875061184936031e-06, + "loss": 0.6366, + "step": 1543 + }, + { + "epoch": 1.4819971195391264, + "grad_norm": 0.8997652197523106, + "learning_rate": 8.872943320666901e-06, + "loss": 0.6194, + "step": 1544 + }, + { + "epoch": 1.4829572731637062, + "grad_norm": 0.8337447335571384, + "learning_rate": 8.870823717874912e-06, + "loss": 0.4955, + "step": 1545 + }, + { + "epoch": 1.4839174267882862, + "grad_norm": 1.3692228718413766, + "learning_rate": 8.868702377511535e-06, + "loss": 0.7163, + "step": 1546 + }, + { + "epoch": 1.484877580412866, + "grad_norm": 1.4991294593929196, + "learning_rate": 8.866579300529014e-06, + "loss": 0.4715, + "step": 1547 + }, + { + "epoch": 1.485837734037446, + "grad_norm": 1.1141143253404242, + "learning_rate": 8.864454487880377e-06, + "loss": 0.6575, + "step": 1548 + }, + { + "epoch": 1.4867978876620258, + "grad_norm": 1.356163506582758, + "learning_rate": 8.862327940519429e-06, + "loss": 0.6527, + "step": 1549 + }, + { + "epoch": 1.4877580412866058, + "grad_norm": 0.7735328461625155, + "learning_rate": 8.860199659400754e-06, + "loss": 0.6868, + "step": 1550 + }, + { + "epoch": 1.4887181949111858, + "grad_norm": 1.0901216355933054, + "learning_rate": 8.858069645479715e-06, + "loss": 0.6636, + "step": 1551 + }, + { + "epoch": 1.4896783485357656, + "grad_norm": 0.9660448042567134, + "learning_rate": 8.855937899712454e-06, + "loss": 0.5188, + "step": 1552 + }, + { + "epoch": 1.4906385021603457, + "grad_norm": 1.0414626727490273, + "learning_rate": 8.853804423055887e-06, + "loss": 0.6597, + "step": 1553 + }, + { + "epoch": 1.4915986557849257, + "grad_norm": 1.1575737580267278, + "learning_rate": 8.851669216467708e-06, + "loss": 0.6408, + "step": 1554 + }, + { + "epoch": 1.4925588094095055, + "grad_norm": 0.9906915429431391, + "learning_rate": 8.849532280906391e-06, + "loss": 0.5295, + "step": 1555 + }, + { + "epoch": 1.4935189630340855, + "grad_norm": 1.3905846032117894, + "learning_rate": 8.847393617331182e-06, + "loss": 0.5678, + "step": 1556 + }, + { + "epoch": 1.4944791166586655, + "grad_norm": 1.2418429156286361, + "learning_rate": 8.845253226702103e-06, + "loss": 0.5972, + "step": 1557 + }, + { + "epoch": 1.4954392702832453, + "grad_norm": 1.1443760223671664, + "learning_rate": 8.843111109979957e-06, + "loss": 0.6809, + "step": 1558 + }, + { + "epoch": 1.4963994239078253, + "grad_norm": 1.2741808025232688, + "learning_rate": 8.840967268126313e-06, + "loss": 0.6569, + "step": 1559 + }, + { + "epoch": 1.4973595775324051, + "grad_norm": 1.410259091742139, + "learning_rate": 8.83882170210352e-06, + "loss": 0.6657, + "step": 1560 + }, + { + "epoch": 1.4983197311569851, + "grad_norm": 1.228627765664867, + "learning_rate": 8.836674412874699e-06, + "loss": 0.6022, + "step": 1561 + }, + { + "epoch": 1.499279884781565, + "grad_norm": 1.1911640645100297, + "learning_rate": 8.834525401403747e-06, + "loss": 0.6037, + "step": 1562 + }, + { + "epoch": 1.500240038406145, + "grad_norm": 1.193801561305446, + "learning_rate": 8.83237466865533e-06, + "loss": 0.6554, + "step": 1563 + }, + { + "epoch": 1.501200192030725, + "grad_norm": 1.1818358561874878, + "learning_rate": 8.83022221559489e-06, + "loss": 0.6398, + "step": 1564 + }, + { + "epoch": 1.5021603456553048, + "grad_norm": 0.8670851130447548, + "learning_rate": 8.828068043188641e-06, + "loss": 0.6182, + "step": 1565 + }, + { + "epoch": 1.5031204992798848, + "grad_norm": 0.9407499345905682, + "learning_rate": 8.825912152403568e-06, + "loss": 0.6049, + "step": 1566 + }, + { + "epoch": 1.5040806529044648, + "grad_norm": 1.0151430781400177, + "learning_rate": 8.823754544207426e-06, + "loss": 0.6779, + "step": 1567 + }, + { + "epoch": 1.5050408065290446, + "grad_norm": 2.185154987340034, + "learning_rate": 8.821595219568741e-06, + "loss": 0.7579, + "step": 1568 + }, + { + "epoch": 1.5060009601536246, + "grad_norm": 1.6014949715673676, + "learning_rate": 8.819434179456814e-06, + "loss": 0.6781, + "step": 1569 + }, + { + "epoch": 1.5069611137782046, + "grad_norm": 1.1293765394233934, + "learning_rate": 8.817271424841711e-06, + "loss": 0.7799, + "step": 1570 + }, + { + "epoch": 1.5079212674027844, + "grad_norm": 1.4630651688147316, + "learning_rate": 8.81510695669427e-06, + "loss": 0.6272, + "step": 1571 + }, + { + "epoch": 1.5088814210273642, + "grad_norm": 1.0806000078130642, + "learning_rate": 8.812940775986098e-06, + "loss": 0.5766, + "step": 1572 + }, + { + "epoch": 1.5098415746519445, + "grad_norm": 1.0636465241656676, + "learning_rate": 8.810772883689567e-06, + "loss": 0.5887, + "step": 1573 + }, + { + "epoch": 1.5108017282765243, + "grad_norm": 1.0826398042376562, + "learning_rate": 8.808603280777826e-06, + "loss": 0.6777, + "step": 1574 + }, + { + "epoch": 1.511761881901104, + "grad_norm": 1.0211975863239149, + "learning_rate": 8.806431968224784e-06, + "loss": 0.4892, + "step": 1575 + }, + { + "epoch": 1.512722035525684, + "grad_norm": 0.9734221712990038, + "learning_rate": 8.80425894700512e-06, + "loss": 0.676, + "step": 1576 + }, + { + "epoch": 1.513682189150264, + "grad_norm": 1.1483419353910045, + "learning_rate": 8.802084218094278e-06, + "loss": 0.6756, + "step": 1577 + }, + { + "epoch": 1.5146423427748439, + "grad_norm": 0.9317491709337571, + "learning_rate": 8.799907782468473e-06, + "loss": 0.5977, + "step": 1578 + }, + { + "epoch": 1.515602496399424, + "grad_norm": 0.960517889530444, + "learning_rate": 8.797729641104687e-06, + "loss": 0.5713, + "step": 1579 + }, + { + "epoch": 1.516562650024004, + "grad_norm": 0.9671139854015403, + "learning_rate": 8.795549794980655e-06, + "loss": 0.7245, + "step": 1580 + }, + { + "epoch": 1.5175228036485837, + "grad_norm": 1.3137409695984128, + "learning_rate": 8.793368245074896e-06, + "loss": 0.7528, + "step": 1581 + }, + { + "epoch": 1.5184829572731637, + "grad_norm": 1.0429450014620933, + "learning_rate": 8.791184992366678e-06, + "loss": 0.6675, + "step": 1582 + }, + { + "epoch": 1.5194431108977438, + "grad_norm": 1.2760113502583366, + "learning_rate": 8.789000037836045e-06, + "loss": 0.6094, + "step": 1583 + }, + { + "epoch": 1.5204032645223235, + "grad_norm": 1.1925925207183954, + "learning_rate": 8.786813382463796e-06, + "loss": 0.5015, + "step": 1584 + }, + { + "epoch": 1.5213634181469033, + "grad_norm": 0.8226805823705247, + "learning_rate": 8.7846250272315e-06, + "loss": 0.6676, + "step": 1585 + }, + { + "epoch": 1.5223235717714836, + "grad_norm": 0.8228915234369152, + "learning_rate": 8.782434973121481e-06, + "loss": 0.5947, + "step": 1586 + }, + { + "epoch": 1.5232837253960634, + "grad_norm": 0.9574007111776064, + "learning_rate": 8.780243221116838e-06, + "loss": 0.7298, + "step": 1587 + }, + { + "epoch": 1.5242438790206432, + "grad_norm": 1.0767250417525245, + "learning_rate": 8.77804977220142e-06, + "loss": 0.7061, + "step": 1588 + }, + { + "epoch": 1.5252040326452232, + "grad_norm": 1.0975411528156236, + "learning_rate": 8.775854627359844e-06, + "loss": 0.5675, + "step": 1589 + }, + { + "epoch": 1.5261641862698032, + "grad_norm": 1.0705241678902788, + "learning_rate": 8.773657787577488e-06, + "loss": 0.6927, + "step": 1590 + }, + { + "epoch": 1.527124339894383, + "grad_norm": 1.2543988040625176, + "learning_rate": 8.771459253840489e-06, + "loss": 0.6857, + "step": 1591 + }, + { + "epoch": 1.528084493518963, + "grad_norm": 1.0609016838340781, + "learning_rate": 8.769259027135746e-06, + "loss": 0.6519, + "step": 1592 + }, + { + "epoch": 1.529044647143543, + "grad_norm": 0.8704882959786878, + "learning_rate": 8.767057108450919e-06, + "loss": 0.531, + "step": 1593 + }, + { + "epoch": 1.5300048007681228, + "grad_norm": 0.9282481560389485, + "learning_rate": 8.76485349877442e-06, + "loss": 0.7366, + "step": 1594 + }, + { + "epoch": 1.5309649543927029, + "grad_norm": 0.9511335341437126, + "learning_rate": 8.76264819909543e-06, + "loss": 0.6666, + "step": 1595 + }, + { + "epoch": 1.5319251080172829, + "grad_norm": 1.1111787893102145, + "learning_rate": 8.760441210403886e-06, + "loss": 0.6375, + "step": 1596 + }, + { + "epoch": 1.5328852616418627, + "grad_norm": 1.092816915711456, + "learning_rate": 8.758232533690479e-06, + "loss": 0.6397, + "step": 1597 + }, + { + "epoch": 1.5338454152664425, + "grad_norm": 1.3155672256646667, + "learning_rate": 8.75602216994666e-06, + "loss": 0.6269, + "step": 1598 + }, + { + "epoch": 1.5348055688910227, + "grad_norm": 0.741803062736798, + "learning_rate": 8.753810120164639e-06, + "loss": 0.6418, + "step": 1599 + }, + { + "epoch": 1.5357657225156025, + "grad_norm": 1.1873303304018221, + "learning_rate": 8.75159638533738e-06, + "loss": 0.5432, + "step": 1600 + }, + { + "epoch": 1.5367258761401823, + "grad_norm": 0.7668290863461923, + "learning_rate": 8.749380966458608e-06, + "loss": 0.7538, + "step": 1601 + }, + { + "epoch": 1.5376860297647623, + "grad_norm": 1.3027080236914383, + "learning_rate": 8.747163864522796e-06, + "loss": 0.6736, + "step": 1602 + }, + { + "epoch": 1.5386461833893423, + "grad_norm": 0.646484719298794, + "learning_rate": 8.744945080525182e-06, + "loss": 0.6115, + "step": 1603 + }, + { + "epoch": 1.5396063370139221, + "grad_norm": 1.5003456905289576, + "learning_rate": 8.742724615461753e-06, + "loss": 0.6875, + "step": 1604 + }, + { + "epoch": 1.5405664906385022, + "grad_norm": 1.4529237102280468, + "learning_rate": 8.740502470329251e-06, + "loss": 0.7543, + "step": 1605 + }, + { + "epoch": 1.5415266442630822, + "grad_norm": 1.264249597177811, + "learning_rate": 8.738278646125174e-06, + "loss": 0.7174, + "step": 1606 + }, + { + "epoch": 1.542486797887662, + "grad_norm": 0.8006931681927969, + "learning_rate": 8.736053143847774e-06, + "loss": 0.5752, + "step": 1607 + }, + { + "epoch": 1.543446951512242, + "grad_norm": 0.9612503521794223, + "learning_rate": 8.733825964496052e-06, + "loss": 0.6369, + "step": 1608 + }, + { + "epoch": 1.544407105136822, + "grad_norm": 1.2263827521809219, + "learning_rate": 8.731597109069768e-06, + "loss": 0.6944, + "step": 1609 + }, + { + "epoch": 1.5453672587614018, + "grad_norm": 0.8117605700550553, + "learning_rate": 8.72936657856943e-06, + "loss": 0.6205, + "step": 1610 + }, + { + "epoch": 1.5463274123859816, + "grad_norm": 1.0844304296657585, + "learning_rate": 8.727134373996298e-06, + "loss": 0.6762, + "step": 1611 + }, + { + "epoch": 1.5472875660105618, + "grad_norm": 1.0356413997768932, + "learning_rate": 8.724900496352388e-06, + "loss": 0.5406, + "step": 1612 + }, + { + "epoch": 1.5482477196351416, + "grad_norm": 1.2889216426140757, + "learning_rate": 8.722664946640463e-06, + "loss": 0.5896, + "step": 1613 + }, + { + "epoch": 1.5492078732597214, + "grad_norm": 1.1650345873037744, + "learning_rate": 8.720427725864035e-06, + "loss": 0.5814, + "step": 1614 + }, + { + "epoch": 1.5501680268843014, + "grad_norm": 0.7366014480030797, + "learning_rate": 8.71818883502737e-06, + "loss": 0.6081, + "step": 1615 + }, + { + "epoch": 1.5511281805088815, + "grad_norm": 1.1318867172011045, + "learning_rate": 8.715948275135482e-06, + "loss": 0.6559, + "step": 1616 + }, + { + "epoch": 1.5520883341334613, + "grad_norm": 0.9487112907304925, + "learning_rate": 8.713706047194135e-06, + "loss": 0.6027, + "step": 1617 + }, + { + "epoch": 1.5530484877580413, + "grad_norm": 1.0802687230047274, + "learning_rate": 8.711462152209843e-06, + "loss": 0.8259, + "step": 1618 + }, + { + "epoch": 1.5540086413826213, + "grad_norm": 1.6731252047990353, + "learning_rate": 8.709216591189861e-06, + "loss": 0.75, + "step": 1619 + }, + { + "epoch": 1.554968795007201, + "grad_norm": 1.1383448741414959, + "learning_rate": 8.706969365142202e-06, + "loss": 0.6404, + "step": 1620 + }, + { + "epoch": 1.555928948631781, + "grad_norm": 1.1778185492412427, + "learning_rate": 8.70472047507562e-06, + "loss": 0.6958, + "step": 1621 + }, + { + "epoch": 1.5568891022563611, + "grad_norm": 0.9686264191189811, + "learning_rate": 8.702469921999617e-06, + "loss": 0.6271, + "step": 1622 + }, + { + "epoch": 1.557849255880941, + "grad_norm": 1.397611275414117, + "learning_rate": 8.700217706924445e-06, + "loss": 0.7544, + "step": 1623 + }, + { + "epoch": 1.558809409505521, + "grad_norm": 1.4513596363263108, + "learning_rate": 8.697963830861095e-06, + "loss": 0.7178, + "step": 1624 + }, + { + "epoch": 1.559769563130101, + "grad_norm": 1.1690765029827235, + "learning_rate": 8.695708294821314e-06, + "loss": 0.659, + "step": 1625 + }, + { + "epoch": 1.5607297167546808, + "grad_norm": 1.0258868567164887, + "learning_rate": 8.693451099817583e-06, + "loss": 0.693, + "step": 1626 + }, + { + "epoch": 1.5616898703792605, + "grad_norm": 1.2127126511703938, + "learning_rate": 8.691192246863133e-06, + "loss": 0.676, + "step": 1627 + }, + { + "epoch": 1.5626500240038406, + "grad_norm": 1.4851939462454826, + "learning_rate": 8.688931736971941e-06, + "loss": 0.5758, + "step": 1628 + }, + { + "epoch": 1.5636101776284206, + "grad_norm": 1.0634247458009611, + "learning_rate": 8.686669571158724e-06, + "loss": 0.4995, + "step": 1629 + }, + { + "epoch": 1.5645703312530004, + "grad_norm": 0.8394074207994456, + "learning_rate": 8.684405750438944e-06, + "loss": 0.5563, + "step": 1630 + }, + { + "epoch": 1.5655304848775804, + "grad_norm": 1.1468296996400023, + "learning_rate": 8.682140275828809e-06, + "loss": 0.7286, + "step": 1631 + }, + { + "epoch": 1.5664906385021604, + "grad_norm": 1.0289285466404867, + "learning_rate": 8.679873148345262e-06, + "loss": 0.5079, + "step": 1632 + }, + { + "epoch": 1.5674507921267402, + "grad_norm": 0.9468257845941952, + "learning_rate": 8.677604369005996e-06, + "loss": 0.6046, + "step": 1633 + }, + { + "epoch": 1.5684109457513202, + "grad_norm": 1.1697711097966919, + "learning_rate": 8.67533393882944e-06, + "loss": 0.7152, + "step": 1634 + }, + { + "epoch": 1.5693710993759002, + "grad_norm": 1.4658046671031377, + "learning_rate": 8.673061858834766e-06, + "loss": 0.7187, + "step": 1635 + }, + { + "epoch": 1.57033125300048, + "grad_norm": 1.3430348470894165, + "learning_rate": 8.670788130041886e-06, + "loss": 0.7524, + "step": 1636 + }, + { + "epoch": 1.57129140662506, + "grad_norm": 1.0160045975668386, + "learning_rate": 8.668512753471453e-06, + "loss": 0.5755, + "step": 1637 + }, + { + "epoch": 1.57225156024964, + "grad_norm": 1.1038876177114614, + "learning_rate": 8.666235730144858e-06, + "loss": 0.7345, + "step": 1638 + }, + { + "epoch": 1.5732117138742199, + "grad_norm": 1.1054129476807606, + "learning_rate": 8.663957061084234e-06, + "loss": 0.6143, + "step": 1639 + }, + { + "epoch": 1.5741718674987997, + "grad_norm": 0.9453810347799845, + "learning_rate": 8.661676747312453e-06, + "loss": 0.569, + "step": 1640 + }, + { + "epoch": 1.5751320211233797, + "grad_norm": 0.9163203128093458, + "learning_rate": 8.65939478985312e-06, + "loss": 0.6403, + "step": 1641 + }, + { + "epoch": 1.5760921747479597, + "grad_norm": 0.9042041854556812, + "learning_rate": 8.657111189730583e-06, + "loss": 0.5737, + "step": 1642 + }, + { + "epoch": 1.5770523283725395, + "grad_norm": 0.9211637353686093, + "learning_rate": 8.654825947969924e-06, + "loss": 0.6265, + "step": 1643 + }, + { + "epoch": 1.5780124819971195, + "grad_norm": 0.8703238211166408, + "learning_rate": 8.652539065596966e-06, + "loss": 0.5899, + "step": 1644 + }, + { + "epoch": 1.5789726356216995, + "grad_norm": 1.2076903241189283, + "learning_rate": 8.650250543638264e-06, + "loss": 0.7573, + "step": 1645 + }, + { + "epoch": 1.5799327892462793, + "grad_norm": 1.2989796355785819, + "learning_rate": 8.647960383121113e-06, + "loss": 0.4868, + "step": 1646 + }, + { + "epoch": 1.5808929428708594, + "grad_norm": 0.8735979859576927, + "learning_rate": 8.64566858507354e-06, + "loss": 0.5502, + "step": 1647 + }, + { + "epoch": 1.5818530964954394, + "grad_norm": 0.8546304127362417, + "learning_rate": 8.643375150524308e-06, + "loss": 0.5565, + "step": 1648 + }, + { + "epoch": 1.5828132501200192, + "grad_norm": 1.0235829348311989, + "learning_rate": 8.641080080502919e-06, + "loss": 0.6458, + "step": 1649 + }, + { + "epoch": 1.5837734037445992, + "grad_norm": 1.1038066712654366, + "learning_rate": 8.6387833760396e-06, + "loss": 0.6618, + "step": 1650 + }, + { + "epoch": 1.5847335573691792, + "grad_norm": 1.2922891293982486, + "learning_rate": 8.636485038165323e-06, + "loss": 0.6367, + "step": 1651 + }, + { + "epoch": 1.585693710993759, + "grad_norm": 1.03589005937051, + "learning_rate": 8.634185067911781e-06, + "loss": 0.6252, + "step": 1652 + }, + { + "epoch": 1.5866538646183388, + "grad_norm": 1.1252711021383222, + "learning_rate": 8.631883466311412e-06, + "loss": 0.6921, + "step": 1653 + }, + { + "epoch": 1.5876140182429188, + "grad_norm": 0.8074113897518285, + "learning_rate": 8.629580234397377e-06, + "loss": 0.6712, + "step": 1654 + }, + { + "epoch": 1.5885741718674988, + "grad_norm": 1.098319269114691, + "learning_rate": 8.627275373203572e-06, + "loss": 0.6199, + "step": 1655 + }, + { + "epoch": 1.5895343254920786, + "grad_norm": 0.902340765050979, + "learning_rate": 8.624968883764626e-06, + "loss": 0.6737, + "step": 1656 + }, + { + "epoch": 1.5904944791166586, + "grad_norm": 0.9334608234803287, + "learning_rate": 8.622660767115897e-06, + "loss": 0.6133, + "step": 1657 + }, + { + "epoch": 1.5914546327412387, + "grad_norm": 1.4876461904454708, + "learning_rate": 8.620351024293475e-06, + "loss": 0.7382, + "step": 1658 + }, + { + "epoch": 1.5924147863658185, + "grad_norm": 1.207322463040773, + "learning_rate": 8.618039656334173e-06, + "loss": 0.5967, + "step": 1659 + }, + { + "epoch": 1.5933749399903985, + "grad_norm": 0.9301430109977181, + "learning_rate": 8.615726664275547e-06, + "loss": 0.5669, + "step": 1660 + }, + { + "epoch": 1.5943350936149785, + "grad_norm": 0.9535514138575241, + "learning_rate": 8.613412049155872e-06, + "loss": 0.6078, + "step": 1661 + }, + { + "epoch": 1.5952952472395583, + "grad_norm": 1.4079098575610178, + "learning_rate": 8.611095812014155e-06, + "loss": 0.7434, + "step": 1662 + }, + { + "epoch": 1.5962554008641383, + "grad_norm": 1.1057978602184975, + "learning_rate": 8.608777953890125e-06, + "loss": 0.5356, + "step": 1663 + }, + { + "epoch": 1.5972155544887183, + "grad_norm": 0.7716019001281486, + "learning_rate": 8.60645847582425e-06, + "loss": 0.5902, + "step": 1664 + }, + { + "epoch": 1.5981757081132981, + "grad_norm": 0.9718399381375142, + "learning_rate": 8.604137378857713e-06, + "loss": 0.6267, + "step": 1665 + }, + { + "epoch": 1.599135861737878, + "grad_norm": 0.9107669288703225, + "learning_rate": 8.601814664032434e-06, + "loss": 0.6825, + "step": 1666 + }, + { + "epoch": 1.6000960153624582, + "grad_norm": 1.0621926560828503, + "learning_rate": 8.599490332391054e-06, + "loss": 0.7198, + "step": 1667 + }, + { + "epoch": 1.601056168987038, + "grad_norm": 1.1535861334047675, + "learning_rate": 8.59716438497694e-06, + "loss": 0.6236, + "step": 1668 + }, + { + "epoch": 1.6020163226116177, + "grad_norm": 0.7516776372352658, + "learning_rate": 8.594836822834185e-06, + "loss": 0.6028, + "step": 1669 + }, + { + "epoch": 1.6029764762361978, + "grad_norm": 1.1685643580367917, + "learning_rate": 8.592507647007606e-06, + "loss": 0.6472, + "step": 1670 + }, + { + "epoch": 1.6039366298607778, + "grad_norm": 1.0980991426900863, + "learning_rate": 8.590176858542748e-06, + "loss": 0.6724, + "step": 1671 + }, + { + "epoch": 1.6048967834853576, + "grad_norm": 1.0716280689989421, + "learning_rate": 8.587844458485874e-06, + "loss": 0.7001, + "step": 1672 + }, + { + "epoch": 1.6058569371099376, + "grad_norm": 1.2080004720156137, + "learning_rate": 8.585510447883975e-06, + "loss": 0.657, + "step": 1673 + }, + { + "epoch": 1.6068170907345176, + "grad_norm": 1.089672816112207, + "learning_rate": 8.583174827784762e-06, + "loss": 0.5745, + "step": 1674 + }, + { + "epoch": 1.6077772443590974, + "grad_norm": 0.8344681189241184, + "learning_rate": 8.580837599236673e-06, + "loss": 0.5936, + "step": 1675 + }, + { + "epoch": 1.6087373979836774, + "grad_norm": 1.0357626272752878, + "learning_rate": 8.57849876328886e-06, + "loss": 0.6799, + "step": 1676 + }, + { + "epoch": 1.6096975516082574, + "grad_norm": 1.005742651907251, + "learning_rate": 8.576158320991205e-06, + "loss": 0.62, + "step": 1677 + }, + { + "epoch": 1.6106577052328372, + "grad_norm": 0.9722385404833888, + "learning_rate": 8.573816273394308e-06, + "loss": 0.6483, + "step": 1678 + }, + { + "epoch": 1.611617858857417, + "grad_norm": 1.2115310207437309, + "learning_rate": 8.571472621549488e-06, + "loss": 0.6395, + "step": 1679 + }, + { + "epoch": 1.6125780124819973, + "grad_norm": 1.0063716252451547, + "learning_rate": 8.569127366508782e-06, + "loss": 0.6235, + "step": 1680 + }, + { + "epoch": 1.613538166106577, + "grad_norm": 0.9894620552857323, + "learning_rate": 8.566780509324956e-06, + "loss": 0.5872, + "step": 1681 + }, + { + "epoch": 1.6144983197311569, + "grad_norm": 0.9067993442019877, + "learning_rate": 8.564432051051485e-06, + "loss": 0.6886, + "step": 1682 + }, + { + "epoch": 1.615458473355737, + "grad_norm": 1.361695050418203, + "learning_rate": 8.562081992742569e-06, + "loss": 0.64, + "step": 1683 + }, + { + "epoch": 1.616418626980317, + "grad_norm": 0.9111982814820835, + "learning_rate": 8.559730335453122e-06, + "loss": 0.5832, + "step": 1684 + }, + { + "epoch": 1.6173787806048967, + "grad_norm": 1.1180812700141665, + "learning_rate": 8.557377080238778e-06, + "loss": 0.5716, + "step": 1685 + }, + { + "epoch": 1.6183389342294767, + "grad_norm": 1.0120829591301022, + "learning_rate": 8.55502222815589e-06, + "loss": 0.4687, + "step": 1686 + }, + { + "epoch": 1.6192990878540567, + "grad_norm": 0.9703184027207251, + "learning_rate": 8.552665780261526e-06, + "loss": 0.6809, + "step": 1687 + }, + { + "epoch": 1.6202592414786365, + "grad_norm": 1.1399371500972935, + "learning_rate": 8.550307737613468e-06, + "loss": 0.65, + "step": 1688 + }, + { + "epoch": 1.6212193951032166, + "grad_norm": 1.2564973647208353, + "learning_rate": 8.547948101270215e-06, + "loss": 0.6402, + "step": 1689 + }, + { + "epoch": 1.6221795487277966, + "grad_norm": 0.810702579081814, + "learning_rate": 8.545586872290987e-06, + "loss": 0.5705, + "step": 1690 + }, + { + "epoch": 1.6231397023523764, + "grad_norm": 1.3279957089503416, + "learning_rate": 8.543224051735714e-06, + "loss": 0.7017, + "step": 1691 + }, + { + "epoch": 1.6240998559769562, + "grad_norm": 1.1568430995470347, + "learning_rate": 8.540859640665036e-06, + "loss": 0.7889, + "step": 1692 + }, + { + "epoch": 1.6250600096015364, + "grad_norm": 1.1895042285769468, + "learning_rate": 8.538493640140317e-06, + "loss": 0.649, + "step": 1693 + }, + { + "epoch": 1.6260201632261162, + "grad_norm": 0.9908405450035387, + "learning_rate": 8.536126051223627e-06, + "loss": 0.5093, + "step": 1694 + }, + { + "epoch": 1.626980316850696, + "grad_norm": 1.1436806207978791, + "learning_rate": 8.533756874977751e-06, + "loss": 0.6795, + "step": 1695 + }, + { + "epoch": 1.627940470475276, + "grad_norm": 1.0001606823595681, + "learning_rate": 8.53138611246619e-06, + "loss": 0.7585, + "step": 1696 + }, + { + "epoch": 1.628900624099856, + "grad_norm": 1.1612341599585667, + "learning_rate": 8.529013764753147e-06, + "loss": 0.6908, + "step": 1697 + }, + { + "epoch": 1.6298607777244358, + "grad_norm": 1.0257237359423448, + "learning_rate": 8.526639832903552e-06, + "loss": 0.6067, + "step": 1698 + }, + { + "epoch": 1.6308209313490158, + "grad_norm": 0.943695610556328, + "learning_rate": 8.524264317983032e-06, + "loss": 0.5724, + "step": 1699 + }, + { + "epoch": 1.6317810849735959, + "grad_norm": 1.1001793125055708, + "learning_rate": 8.521887221057932e-06, + "loss": 0.5564, + "step": 1700 + }, + { + "epoch": 1.6327412385981757, + "grad_norm": 0.7967117986230224, + "learning_rate": 8.519508543195304e-06, + "loss": 0.4922, + "step": 1701 + }, + { + "epoch": 1.6337013922227557, + "grad_norm": 0.9404742155404143, + "learning_rate": 8.517128285462914e-06, + "loss": 0.6039, + "step": 1702 + }, + { + "epoch": 1.6346615458473357, + "grad_norm": 1.1292581101586436, + "learning_rate": 8.514746448929235e-06, + "loss": 0.6944, + "step": 1703 + }, + { + "epoch": 1.6356216994719155, + "grad_norm": 1.4004869256437205, + "learning_rate": 8.512363034663441e-06, + "loss": 0.7079, + "step": 1704 + }, + { + "epoch": 1.6365818530964953, + "grad_norm": 1.3431801226281752, + "learning_rate": 8.50997804373543e-06, + "loss": 0.7172, + "step": 1705 + }, + { + "epoch": 1.6375420067210755, + "grad_norm": 1.3314862190720576, + "learning_rate": 8.507591477215793e-06, + "loss": 0.6085, + "step": 1706 + }, + { + "epoch": 1.6385021603456553, + "grad_norm": 1.209629508540758, + "learning_rate": 8.505203336175836e-06, + "loss": 0.7252, + "step": 1707 + }, + { + "epoch": 1.6394623139702351, + "grad_norm": 1.274399903410045, + "learning_rate": 8.502813621687568e-06, + "loss": 0.4603, + "step": 1708 + }, + { + "epoch": 1.6404224675948151, + "grad_norm": 1.408366215341856, + "learning_rate": 8.50042233482371e-06, + "loss": 0.6636, + "step": 1709 + }, + { + "epoch": 1.6413826212193952, + "grad_norm": 0.8719579855297336, + "learning_rate": 8.498029476657686e-06, + "loss": 0.5953, + "step": 1710 + }, + { + "epoch": 1.642342774843975, + "grad_norm": 1.049429410346072, + "learning_rate": 8.495635048263617e-06, + "loss": 0.5743, + "step": 1711 + }, + { + "epoch": 1.643302928468555, + "grad_norm": 1.0268285567030226, + "learning_rate": 8.493239050716344e-06, + "loss": 0.6271, + "step": 1712 + }, + { + "epoch": 1.644263082093135, + "grad_norm": 1.3780276128217048, + "learning_rate": 8.4908414850914e-06, + "loss": 0.5648, + "step": 1713 + }, + { + "epoch": 1.6452232357177148, + "grad_norm": 1.1025995482917144, + "learning_rate": 8.488442352465029e-06, + "loss": 0.5872, + "step": 1714 + }, + { + "epoch": 1.6461833893422948, + "grad_norm": 1.2177032533446264, + "learning_rate": 8.486041653914177e-06, + "loss": 0.6434, + "step": 1715 + }, + { + "epoch": 1.6471435429668748, + "grad_norm": 1.318212060341539, + "learning_rate": 8.483639390516488e-06, + "loss": 0.7588, + "step": 1716 + }, + { + "epoch": 1.6481036965914546, + "grad_norm": 1.4281094927704476, + "learning_rate": 8.481235563350316e-06, + "loss": 0.6115, + "step": 1717 + }, + { + "epoch": 1.6490638502160344, + "grad_norm": 1.1792419894081392, + "learning_rate": 8.478830173494712e-06, + "loss": 0.6358, + "step": 1718 + }, + { + "epoch": 1.6500240038406147, + "grad_norm": 0.9943172186490088, + "learning_rate": 8.47642322202943e-06, + "loss": 0.5389, + "step": 1719 + }, + { + "epoch": 1.6509841574651944, + "grad_norm": 1.0148518234658925, + "learning_rate": 8.474014710034923e-06, + "loss": 0.7093, + "step": 1720 + }, + { + "epoch": 1.6519443110897742, + "grad_norm": 1.1948035124293634, + "learning_rate": 8.471604638592348e-06, + "loss": 0.7758, + "step": 1721 + }, + { + "epoch": 1.6529044647143543, + "grad_norm": 0.9681052074623742, + "learning_rate": 8.469193008783562e-06, + "loss": 0.6116, + "step": 1722 + }, + { + "epoch": 1.6538646183389343, + "grad_norm": 1.3128560663557998, + "learning_rate": 8.466779821691117e-06, + "loss": 0.6443, + "step": 1723 + }, + { + "epoch": 1.654824771963514, + "grad_norm": 0.8640570945968897, + "learning_rate": 8.464365078398269e-06, + "loss": 0.6419, + "step": 1724 + }, + { + "epoch": 1.655784925588094, + "grad_norm": 1.4007162676256382, + "learning_rate": 8.461948779988967e-06, + "loss": 0.6014, + "step": 1725 + }, + { + "epoch": 1.656745079212674, + "grad_norm": 1.294619488237341, + "learning_rate": 8.459530927547864e-06, + "loss": 0.6103, + "step": 1726 + }, + { + "epoch": 1.657705232837254, + "grad_norm": 0.9582533270010303, + "learning_rate": 8.457111522160309e-06, + "loss": 0.679, + "step": 1727 + }, + { + "epoch": 1.658665386461834, + "grad_norm": 1.2081739113840666, + "learning_rate": 8.454690564912347e-06, + "loss": 0.6485, + "step": 1728 + }, + { + "epoch": 1.659625540086414, + "grad_norm": 0.9562788530236775, + "learning_rate": 8.452268056890717e-06, + "loss": 0.7311, + "step": 1729 + }, + { + "epoch": 1.6605856937109937, + "grad_norm": 1.0530110294051391, + "learning_rate": 8.44984399918286e-06, + "loss": 0.6905, + "step": 1730 + }, + { + "epoch": 1.6615458473355735, + "grad_norm": 0.8980922663700222, + "learning_rate": 8.447418392876907e-06, + "loss": 0.7196, + "step": 1731 + }, + { + "epoch": 1.6625060009601538, + "grad_norm": 1.2779398677085188, + "learning_rate": 8.44499123906169e-06, + "loss": 0.699, + "step": 1732 + }, + { + "epoch": 1.6634661545847336, + "grad_norm": 1.3388477077082197, + "learning_rate": 8.44256253882673e-06, + "loss": 0.5993, + "step": 1733 + }, + { + "epoch": 1.6644263082093134, + "grad_norm": 0.9604175271967059, + "learning_rate": 8.440132293262246e-06, + "loss": 0.6443, + "step": 1734 + }, + { + "epoch": 1.6653864618338934, + "grad_norm": 1.059165325720594, + "learning_rate": 8.437700503459149e-06, + "loss": 0.6622, + "step": 1735 + }, + { + "epoch": 1.6663466154584734, + "grad_norm": 1.2847355817366322, + "learning_rate": 8.435267170509044e-06, + "loss": 0.6376, + "step": 1736 + }, + { + "epoch": 1.6673067690830532, + "grad_norm": 0.9486013461835123, + "learning_rate": 8.432832295504224e-06, + "loss": 0.6135, + "step": 1737 + }, + { + "epoch": 1.6682669227076332, + "grad_norm": 1.1217559786960296, + "learning_rate": 8.430395879537684e-06, + "loss": 0.7307, + "step": 1738 + }, + { + "epoch": 1.6692270763322132, + "grad_norm": 1.166539193180166, + "learning_rate": 8.4279579237031e-06, + "loss": 0.6745, + "step": 1739 + }, + { + "epoch": 1.670187229956793, + "grad_norm": 1.2944973539291245, + "learning_rate": 8.425518429094848e-06, + "loss": 0.6738, + "step": 1740 + }, + { + "epoch": 1.671147383581373, + "grad_norm": 1.7769472382713831, + "learning_rate": 8.423077396807991e-06, + "loss": 0.72, + "step": 1741 + }, + { + "epoch": 1.672107537205953, + "grad_norm": 0.9923753385483782, + "learning_rate": 8.42063482793828e-06, + "loss": 0.6447, + "step": 1742 + }, + { + "epoch": 1.6730676908305329, + "grad_norm": 1.5206413904288867, + "learning_rate": 8.41819072358216e-06, + "loss": 0.7335, + "step": 1743 + }, + { + "epoch": 1.6740278444551127, + "grad_norm": 1.2603132686679719, + "learning_rate": 8.415745084836763e-06, + "loss": 0.7357, + "step": 1744 + }, + { + "epoch": 1.674987998079693, + "grad_norm": 1.3263524233400652, + "learning_rate": 8.413297912799909e-06, + "loss": 0.603, + "step": 1745 + }, + { + "epoch": 1.6759481517042727, + "grad_norm": 1.034102796940537, + "learning_rate": 8.410849208570108e-06, + "loss": 0.5277, + "step": 1746 + }, + { + "epoch": 1.6769083053288525, + "grad_norm": 0.9802629714566773, + "learning_rate": 8.408398973246557e-06, + "loss": 0.6366, + "step": 1747 + }, + { + "epoch": 1.6778684589534325, + "grad_norm": 1.7367971970710725, + "learning_rate": 8.405947207929142e-06, + "loss": 0.6344, + "step": 1748 + }, + { + "epoch": 1.6788286125780125, + "grad_norm": 1.0468229139336518, + "learning_rate": 8.40349391371843e-06, + "loss": 0.6003, + "step": 1749 + }, + { + "epoch": 1.6797887662025923, + "grad_norm": 0.8799505709158508, + "learning_rate": 8.401039091715685e-06, + "loss": 0.7298, + "step": 1750 + }, + { + "epoch": 1.6807489198271723, + "grad_norm": 1.0134659723179875, + "learning_rate": 8.398582743022845e-06, + "loss": 0.6389, + "step": 1751 + }, + { + "epoch": 1.6817090734517524, + "grad_norm": 0.9872492043726284, + "learning_rate": 8.39612486874254e-06, + "loss": 0.5881, + "step": 1752 + }, + { + "epoch": 1.6826692270763322, + "grad_norm": 1.1250605266987366, + "learning_rate": 8.393665469978086e-06, + "loss": 0.6378, + "step": 1753 + }, + { + "epoch": 1.6836293807009122, + "grad_norm": 1.0626430971346574, + "learning_rate": 8.391204547833478e-06, + "loss": 0.7311, + "step": 1754 + }, + { + "epoch": 1.6845895343254922, + "grad_norm": 1.1428411752670604, + "learning_rate": 8.388742103413397e-06, + "loss": 0.613, + "step": 1755 + }, + { + "epoch": 1.685549687950072, + "grad_norm": 1.064563945141614, + "learning_rate": 8.38627813782321e-06, + "loss": 0.6215, + "step": 1756 + }, + { + "epoch": 1.6865098415746518, + "grad_norm": 1.3895186464336546, + "learning_rate": 8.383812652168966e-06, + "loss": 0.6832, + "step": 1757 + }, + { + "epoch": 1.687469995199232, + "grad_norm": 1.0500015031401468, + "learning_rate": 8.38134564755739e-06, + "loss": 0.6395, + "step": 1758 + }, + { + "epoch": 1.6884301488238118, + "grad_norm": 1.0701915658229737, + "learning_rate": 8.378877125095901e-06, + "loss": 0.6869, + "step": 1759 + }, + { + "epoch": 1.6893903024483916, + "grad_norm": 0.9883890260744455, + "learning_rate": 8.376407085892586e-06, + "loss": 0.627, + "step": 1760 + }, + { + "epoch": 1.6903504560729716, + "grad_norm": 1.204954413250014, + "learning_rate": 8.373935531056222e-06, + "loss": 0.6609, + "step": 1761 + }, + { + "epoch": 1.6913106096975516, + "grad_norm": 0.9014119447216241, + "learning_rate": 8.371462461696267e-06, + "loss": 0.6924, + "step": 1762 + }, + { + "epoch": 1.6922707633221314, + "grad_norm": 0.9454956924843029, + "learning_rate": 8.36898787892285e-06, + "loss": 0.4615, + "step": 1763 + }, + { + "epoch": 1.6932309169467115, + "grad_norm": 1.016954677498344, + "learning_rate": 8.366511783846785e-06, + "loss": 0.5375, + "step": 1764 + }, + { + "epoch": 1.6941910705712915, + "grad_norm": 1.4344533155048056, + "learning_rate": 8.364034177579568e-06, + "loss": 0.7729, + "step": 1765 + }, + { + "epoch": 1.6951512241958713, + "grad_norm": 0.9825308632016129, + "learning_rate": 8.361555061233367e-06, + "loss": 0.5923, + "step": 1766 + }, + { + "epoch": 1.6961113778204513, + "grad_norm": 1.044828001221576, + "learning_rate": 8.359074435921032e-06, + "loss": 0.581, + "step": 1767 + }, + { + "epoch": 1.6970715314450313, + "grad_norm": 1.162410608452558, + "learning_rate": 8.35659230275609e-06, + "loss": 0.5913, + "step": 1768 + }, + { + "epoch": 1.698031685069611, + "grad_norm": 1.235270867502264, + "learning_rate": 8.354108662852743e-06, + "loss": 0.6416, + "step": 1769 + }, + { + "epoch": 1.6989918386941911, + "grad_norm": 1.166953971280625, + "learning_rate": 8.351623517325872e-06, + "loss": 0.5994, + "step": 1770 + }, + { + "epoch": 1.6999519923187711, + "grad_norm": 1.0380648883740493, + "learning_rate": 8.349136867291029e-06, + "loss": 0.6621, + "step": 1771 + }, + { + "epoch": 1.700912145943351, + "grad_norm": 1.1119353863991024, + "learning_rate": 8.346648713864447e-06, + "loss": 0.6469, + "step": 1772 + }, + { + "epoch": 1.7018722995679307, + "grad_norm": 0.896899098530902, + "learning_rate": 8.344159058163032e-06, + "loss": 0.6084, + "step": 1773 + }, + { + "epoch": 1.7028324531925108, + "grad_norm": 0.8633910130054333, + "learning_rate": 8.341667901304362e-06, + "loss": 0.5014, + "step": 1774 + }, + { + "epoch": 1.7037926068170908, + "grad_norm": 0.8323822273701, + "learning_rate": 8.339175244406693e-06, + "loss": 0.5515, + "step": 1775 + }, + { + "epoch": 1.7047527604416706, + "grad_norm": 1.0272487976042384, + "learning_rate": 8.336681088588949e-06, + "loss": 0.6712, + "step": 1776 + }, + { + "epoch": 1.7057129140662506, + "grad_norm": 1.3287844681034346, + "learning_rate": 8.33418543497073e-06, + "loss": 0.7486, + "step": 1777 + }, + { + "epoch": 1.7066730676908306, + "grad_norm": 1.071009262146182, + "learning_rate": 8.331688284672313e-06, + "loss": 0.5426, + "step": 1778 + }, + { + "epoch": 1.7076332213154104, + "grad_norm": 0.9979771404488532, + "learning_rate": 8.329189638814637e-06, + "loss": 0.5924, + "step": 1779 + }, + { + "epoch": 1.7085933749399904, + "grad_norm": 0.9845955447724891, + "learning_rate": 8.32668949851932e-06, + "loss": 0.5748, + "step": 1780 + }, + { + "epoch": 1.7095535285645704, + "grad_norm": 1.038187251919438, + "learning_rate": 8.324187864908646e-06, + "loss": 0.5973, + "step": 1781 + }, + { + "epoch": 1.7105136821891502, + "grad_norm": 0.9283230295023462, + "learning_rate": 8.321684739105573e-06, + "loss": 0.5842, + "step": 1782 + }, + { + "epoch": 1.7114738358137302, + "grad_norm": 0.9610776326476227, + "learning_rate": 8.319180122233729e-06, + "loss": 0.5744, + "step": 1783 + }, + { + "epoch": 1.7124339894383103, + "grad_norm": 1.1848026680157284, + "learning_rate": 8.316674015417407e-06, + "loss": 0.5415, + "step": 1784 + }, + { + "epoch": 1.71339414306289, + "grad_norm": 0.8383489208310458, + "learning_rate": 8.314166419781574e-06, + "loss": 0.5809, + "step": 1785 + }, + { + "epoch": 1.7143542966874699, + "grad_norm": 0.8902371551760121, + "learning_rate": 8.311657336451862e-06, + "loss": 0.6163, + "step": 1786 + }, + { + "epoch": 1.7153144503120499, + "grad_norm": 1.0486580496373912, + "learning_rate": 8.30914676655457e-06, + "loss": 0.6104, + "step": 1787 + }, + { + "epoch": 1.71627460393663, + "grad_norm": 1.171465563574018, + "learning_rate": 8.306634711216668e-06, + "loss": 0.702, + "step": 1788 + }, + { + "epoch": 1.7172347575612097, + "grad_norm": 1.1521729735887885, + "learning_rate": 8.304121171565788e-06, + "loss": 0.572, + "step": 1789 + }, + { + "epoch": 1.7181949111857897, + "grad_norm": 0.8221298578072059, + "learning_rate": 8.301606148730237e-06, + "loss": 0.5707, + "step": 1790 + }, + { + "epoch": 1.7191550648103697, + "grad_norm": 1.1029215524544493, + "learning_rate": 8.299089643838976e-06, + "loss": 0.6609, + "step": 1791 + }, + { + "epoch": 1.7201152184349495, + "grad_norm": 1.4561379954869085, + "learning_rate": 8.296571658021638e-06, + "loss": 0.6613, + "step": 1792 + }, + { + "epoch": 1.7210753720595295, + "grad_norm": 1.2738669594132566, + "learning_rate": 8.294052192408522e-06, + "loss": 0.792, + "step": 1793 + }, + { + "epoch": 1.7220355256841096, + "grad_norm": 1.4798740905642442, + "learning_rate": 8.291531248130589e-06, + "loss": 0.6698, + "step": 1794 + }, + { + "epoch": 1.7229956793086894, + "grad_norm": 1.134409326137547, + "learning_rate": 8.289008826319463e-06, + "loss": 0.6577, + "step": 1795 + }, + { + "epoch": 1.7239558329332694, + "grad_norm": 0.9551706414818583, + "learning_rate": 8.286484928107431e-06, + "loss": 0.6724, + "step": 1796 + }, + { + "epoch": 1.7249159865578494, + "grad_norm": 1.2374840247458698, + "learning_rate": 8.283959554627448e-06, + "loss": 0.7348, + "step": 1797 + }, + { + "epoch": 1.7258761401824292, + "grad_norm": 1.3416378545918735, + "learning_rate": 8.281432707013123e-06, + "loss": 0.7342, + "step": 1798 + }, + { + "epoch": 1.726836293807009, + "grad_norm": 1.227198660378129, + "learning_rate": 8.278904386398733e-06, + "loss": 0.697, + "step": 1799 + }, + { + "epoch": 1.727796447431589, + "grad_norm": 0.8602654088414831, + "learning_rate": 8.276374593919213e-06, + "loss": 0.5971, + "step": 1800 + }, + { + "epoch": 1.728756601056169, + "grad_norm": 0.8383919774027291, + "learning_rate": 8.27384333071016e-06, + "loss": 0.6112, + "step": 1801 + }, + { + "epoch": 1.7297167546807488, + "grad_norm": 1.0016028936180106, + "learning_rate": 8.27131059790783e-06, + "loss": 0.5832, + "step": 1802 + }, + { + "epoch": 1.7306769083053288, + "grad_norm": 1.1733238954383638, + "learning_rate": 8.268776396649145e-06, + "loss": 0.793, + "step": 1803 + }, + { + "epoch": 1.7316370619299088, + "grad_norm": 1.2035517212362503, + "learning_rate": 8.266240728071673e-06, + "loss": 0.6498, + "step": 1804 + }, + { + "epoch": 1.7325972155544886, + "grad_norm": 1.0897666129294943, + "learning_rate": 8.263703593313654e-06, + "loss": 0.6667, + "step": 1805 + }, + { + "epoch": 1.7335573691790687, + "grad_norm": 0.9073661107039916, + "learning_rate": 8.261164993513978e-06, + "loss": 0.6347, + "step": 1806 + }, + { + "epoch": 1.7345175228036487, + "grad_norm": 1.5648320656747918, + "learning_rate": 8.258624929812198e-06, + "loss": 0.7912, + "step": 1807 + }, + { + "epoch": 1.7354776764282285, + "grad_norm": 1.3532627183223036, + "learning_rate": 8.256083403348518e-06, + "loss": 0.6484, + "step": 1808 + }, + { + "epoch": 1.7364378300528085, + "grad_norm": 0.9287022732318587, + "learning_rate": 8.253540415263805e-06, + "loss": 0.5194, + "step": 1809 + }, + { + "epoch": 1.7373979836773885, + "grad_norm": 1.0796822687805636, + "learning_rate": 8.250995966699577e-06, + "loss": 0.5494, + "step": 1810 + }, + { + "epoch": 1.7383581373019683, + "grad_norm": 1.1107009166158903, + "learning_rate": 8.248450058798012e-06, + "loss": 0.5783, + "step": 1811 + }, + { + "epoch": 1.739318290926548, + "grad_norm": 0.9604962537986486, + "learning_rate": 8.245902692701939e-06, + "loss": 0.5204, + "step": 1812 + }, + { + "epoch": 1.7402784445511283, + "grad_norm": 0.8985142428195022, + "learning_rate": 8.243353869554845e-06, + "loss": 0.6438, + "step": 1813 + }, + { + "epoch": 1.7412385981757081, + "grad_norm": 0.900420965577512, + "learning_rate": 8.240803590500865e-06, + "loss": 0.6015, + "step": 1814 + }, + { + "epoch": 1.742198751800288, + "grad_norm": 1.229040739833296, + "learning_rate": 8.238251856684801e-06, + "loss": 0.6532, + "step": 1815 + }, + { + "epoch": 1.743158905424868, + "grad_norm": 0.9596600579734443, + "learning_rate": 8.23569866925209e-06, + "loss": 0.6131, + "step": 1816 + }, + { + "epoch": 1.744119059049448, + "grad_norm": 0.9098868199427298, + "learning_rate": 8.233144029348834e-06, + "loss": 0.5849, + "step": 1817 + }, + { + "epoch": 1.7450792126740278, + "grad_norm": 0.7450651835506069, + "learning_rate": 8.230587938121783e-06, + "loss": 0.6278, + "step": 1818 + }, + { + "epoch": 1.7460393662986078, + "grad_norm": 1.3584641927417642, + "learning_rate": 8.228030396718342e-06, + "loss": 0.8177, + "step": 1819 + }, + { + "epoch": 1.7469995199231878, + "grad_norm": 1.1138215385227221, + "learning_rate": 8.22547140628656e-06, + "loss": 0.6034, + "step": 1820 + }, + { + "epoch": 1.7479596735477676, + "grad_norm": 1.0849675016586429, + "learning_rate": 8.222910967975143e-06, + "loss": 0.6782, + "step": 1821 + }, + { + "epoch": 1.7489198271723476, + "grad_norm": 1.0631193133434942, + "learning_rate": 8.220349082933444e-06, + "loss": 0.6223, + "step": 1822 + }, + { + "epoch": 1.7498799807969276, + "grad_norm": 1.0135727168930355, + "learning_rate": 8.217785752311464e-06, + "loss": 0.5909, + "step": 1823 + }, + { + "epoch": 1.7508401344215074, + "grad_norm": 0.7561661716159543, + "learning_rate": 8.215220977259855e-06, + "loss": 0.6372, + "step": 1824 + }, + { + "epoch": 1.7518002880460872, + "grad_norm": 0.9250441764475402, + "learning_rate": 8.21265475892992e-06, + "loss": 0.6655, + "step": 1825 + }, + { + "epoch": 1.7527604416706675, + "grad_norm": 1.2282270838391987, + "learning_rate": 8.210087098473605e-06, + "loss": 0.653, + "step": 1826 + }, + { + "epoch": 1.7537205952952473, + "grad_norm": 1.1097200783241623, + "learning_rate": 8.207517997043504e-06, + "loss": 0.689, + "step": 1827 + }, + { + "epoch": 1.754680748919827, + "grad_norm": 0.9590947969068119, + "learning_rate": 8.204947455792863e-06, + "loss": 0.6778, + "step": 1828 + }, + { + "epoch": 1.755640902544407, + "grad_norm": 1.0631355802999023, + "learning_rate": 8.202375475875566e-06, + "loss": 0.5967, + "step": 1829 + }, + { + "epoch": 1.756601056168987, + "grad_norm": 1.0972744735808178, + "learning_rate": 8.19980205844615e-06, + "loss": 0.6279, + "step": 1830 + }, + { + "epoch": 1.757561209793567, + "grad_norm": 0.9125316595551967, + "learning_rate": 8.197227204659795e-06, + "loss": 0.6852, + "step": 1831 + }, + { + "epoch": 1.758521363418147, + "grad_norm": 1.0202579993835212, + "learning_rate": 8.194650915672323e-06, + "loss": 0.5599, + "step": 1832 + }, + { + "epoch": 1.759481517042727, + "grad_norm": 0.9641028651047512, + "learning_rate": 8.192073192640204e-06, + "loss": 0.6607, + "step": 1833 + }, + { + "epoch": 1.7604416706673067, + "grad_norm": 1.0321264396840382, + "learning_rate": 8.18949403672055e-06, + "loss": 0.685, + "step": 1834 + }, + { + "epoch": 1.7614018242918867, + "grad_norm": 1.1142882750756766, + "learning_rate": 8.186913449071117e-06, + "loss": 0.6774, + "step": 1835 + }, + { + "epoch": 1.7623619779164668, + "grad_norm": 1.2196631969025054, + "learning_rate": 8.184331430850304e-06, + "loss": 0.5104, + "step": 1836 + }, + { + "epoch": 1.7633221315410466, + "grad_norm": 1.021990399884995, + "learning_rate": 8.181747983217148e-06, + "loss": 0.591, + "step": 1837 + }, + { + "epoch": 1.7642822851656264, + "grad_norm": 0.9859026254377197, + "learning_rate": 8.179163107331336e-06, + "loss": 0.5823, + "step": 1838 + }, + { + "epoch": 1.7652424387902066, + "grad_norm": 0.8971928754930129, + "learning_rate": 8.176576804353186e-06, + "loss": 0.5824, + "step": 1839 + }, + { + "epoch": 1.7662025924147864, + "grad_norm": 0.8651992267585531, + "learning_rate": 8.173989075443666e-06, + "loss": 0.575, + "step": 1840 + }, + { + "epoch": 1.7671627460393662, + "grad_norm": 1.12415934868087, + "learning_rate": 8.171399921764377e-06, + "loss": 0.502, + "step": 1841 + }, + { + "epoch": 1.7681228996639462, + "grad_norm": 1.2742387716529824, + "learning_rate": 8.168809344477564e-06, + "loss": 0.7826, + "step": 1842 + }, + { + "epoch": 1.7690830532885262, + "grad_norm": 1.4484109927599673, + "learning_rate": 8.16621734474611e-06, + "loss": 0.5709, + "step": 1843 + }, + { + "epoch": 1.770043206913106, + "grad_norm": 1.2696285065207322, + "learning_rate": 8.163623923733533e-06, + "loss": 0.6374, + "step": 1844 + }, + { + "epoch": 1.771003360537686, + "grad_norm": 1.2535412969858437, + "learning_rate": 8.161029082603994e-06, + "loss": 0.6868, + "step": 1845 + }, + { + "epoch": 1.771963514162266, + "grad_norm": 1.3855775261414205, + "learning_rate": 8.158432822522291e-06, + "loss": 0.6984, + "step": 1846 + }, + { + "epoch": 1.7729236677868458, + "grad_norm": 0.8498398072770822, + "learning_rate": 8.155835144653856e-06, + "loss": 0.626, + "step": 1847 + }, + { + "epoch": 1.7738838214114259, + "grad_norm": 1.219824573506507, + "learning_rate": 8.153236050164757e-06, + "loss": 0.5783, + "step": 1848 + }, + { + "epoch": 1.7748439750360059, + "grad_norm": 0.9608903408185807, + "learning_rate": 8.1506355402217e-06, + "loss": 0.7243, + "step": 1849 + }, + { + "epoch": 1.7758041286605857, + "grad_norm": 1.0144995350899024, + "learning_rate": 8.14803361599203e-06, + "loss": 0.6871, + "step": 1850 + }, + { + "epoch": 1.7767642822851655, + "grad_norm": 1.3313755262778169, + "learning_rate": 8.14543027864372e-06, + "loss": 0.586, + "step": 1851 + }, + { + "epoch": 1.7777244359097457, + "grad_norm": 0.9957132992143335, + "learning_rate": 8.14282552934538e-06, + "loss": 0.6559, + "step": 1852 + }, + { + "epoch": 1.7786845895343255, + "grad_norm": 1.2923986777784746, + "learning_rate": 8.140219369266255e-06, + "loss": 0.6073, + "step": 1853 + }, + { + "epoch": 1.7796447431589053, + "grad_norm": 0.9127845535992563, + "learning_rate": 8.137611799576222e-06, + "loss": 0.5435, + "step": 1854 + }, + { + "epoch": 1.7806048967834853, + "grad_norm": 0.929215519990613, + "learning_rate": 8.135002821445793e-06, + "loss": 0.6305, + "step": 1855 + }, + { + "epoch": 1.7815650504080653, + "grad_norm": 1.0143603358268427, + "learning_rate": 8.132392436046107e-06, + "loss": 0.6905, + "step": 1856 + }, + { + "epoch": 1.7825252040326451, + "grad_norm": 0.9559284383420343, + "learning_rate": 8.129780644548938e-06, + "loss": 0.6128, + "step": 1857 + }, + { + "epoch": 1.7834853576572252, + "grad_norm": 0.9519699589571047, + "learning_rate": 8.127167448126698e-06, + "loss": 0.635, + "step": 1858 + }, + { + "epoch": 1.7844455112818052, + "grad_norm": 1.2331502386379996, + "learning_rate": 8.124552847952416e-06, + "loss": 0.5966, + "step": 1859 + }, + { + "epoch": 1.785405664906385, + "grad_norm": 1.1994027600373456, + "learning_rate": 8.121936845199759e-06, + "loss": 0.7028, + "step": 1860 + }, + { + "epoch": 1.786365818530965, + "grad_norm": 1.4147779911257676, + "learning_rate": 8.119319441043024e-06, + "loss": 0.6477, + "step": 1861 + }, + { + "epoch": 1.787325972155545, + "grad_norm": 0.8490520314350598, + "learning_rate": 8.116700636657137e-06, + "loss": 0.5516, + "step": 1862 + }, + { + "epoch": 1.7882861257801248, + "grad_norm": 1.1696054113032124, + "learning_rate": 8.11408043321765e-06, + "loss": 0.7455, + "step": 1863 + }, + { + "epoch": 1.7892462794047046, + "grad_norm": 1.225511448354425, + "learning_rate": 8.111458831900743e-06, + "loss": 0.5417, + "step": 1864 + }, + { + "epoch": 1.7902064330292848, + "grad_norm": 0.8514978364567153, + "learning_rate": 8.108835833883227e-06, + "loss": 0.6946, + "step": 1865 + }, + { + "epoch": 1.7911665866538646, + "grad_norm": 1.0733158336064204, + "learning_rate": 8.106211440342535e-06, + "loss": 0.6104, + "step": 1866 + }, + { + "epoch": 1.7921267402784444, + "grad_norm": 0.859932043516501, + "learning_rate": 8.103585652456732e-06, + "loss": 0.6925, + "step": 1867 + }, + { + "epoch": 1.7930868939030244, + "grad_norm": 0.7924816855317768, + "learning_rate": 8.100958471404506e-06, + "loss": 0.4791, + "step": 1868 + }, + { + "epoch": 1.7940470475276045, + "grad_norm": 0.9026770361495339, + "learning_rate": 8.098329898365168e-06, + "loss": 0.5668, + "step": 1869 + }, + { + "epoch": 1.7950072011521843, + "grad_norm": 0.7783217855735335, + "learning_rate": 8.095699934518659e-06, + "loss": 0.6691, + "step": 1870 + }, + { + "epoch": 1.7959673547767643, + "grad_norm": 1.3245388261385187, + "learning_rate": 8.093068581045538e-06, + "loss": 0.6839, + "step": 1871 + }, + { + "epoch": 1.7969275084013443, + "grad_norm": 0.9264206387795253, + "learning_rate": 8.090435839126998e-06, + "loss": 0.6947, + "step": 1872 + }, + { + "epoch": 1.797887662025924, + "grad_norm": 1.015192823602121, + "learning_rate": 8.087801709944842e-06, + "loss": 0.63, + "step": 1873 + }, + { + "epoch": 1.7988478156505041, + "grad_norm": 1.2166516771098983, + "learning_rate": 8.085166194681508e-06, + "loss": 0.6834, + "step": 1874 + }, + { + "epoch": 1.7998079692750841, + "grad_norm": 1.4029832180085575, + "learning_rate": 8.082529294520046e-06, + "loss": 0.7395, + "step": 1875 + }, + { + "epoch": 1.800768122899664, + "grad_norm": 1.0670024630500639, + "learning_rate": 8.079891010644134e-06, + "loss": 0.7066, + "step": 1876 + }, + { + "epoch": 1.8017282765242437, + "grad_norm": 0.9591006276036653, + "learning_rate": 8.077251344238068e-06, + "loss": 0.5438, + "step": 1877 + }, + { + "epoch": 1.802688430148824, + "grad_norm": 0.807560347256751, + "learning_rate": 8.074610296486771e-06, + "loss": 0.5772, + "step": 1878 + }, + { + "epoch": 1.8036485837734038, + "grad_norm": 1.1301704136208068, + "learning_rate": 8.071967868575775e-06, + "loss": 0.5224, + "step": 1879 + }, + { + "epoch": 1.8046087373979836, + "grad_norm": 0.8090738491171423, + "learning_rate": 8.069324061691242e-06, + "loss": 0.5858, + "step": 1880 + }, + { + "epoch": 1.8055688910225636, + "grad_norm": 1.0707391627369216, + "learning_rate": 8.066678877019945e-06, + "loss": 0.7289, + "step": 1881 + }, + { + "epoch": 1.8065290446471436, + "grad_norm": 0.9828589858485237, + "learning_rate": 8.064032315749283e-06, + "loss": 0.6256, + "step": 1882 + }, + { + "epoch": 1.8074891982717234, + "grad_norm": 0.9962401337718808, + "learning_rate": 8.061384379067266e-06, + "loss": 0.6766, + "step": 1883 + }, + { + "epoch": 1.8084493518963034, + "grad_norm": 0.8138116443056042, + "learning_rate": 8.058735068162527e-06, + "loss": 0.6071, + "step": 1884 + }, + { + "epoch": 1.8094095055208834, + "grad_norm": 0.8179160915657997, + "learning_rate": 8.05608438422431e-06, + "loss": 0.7092, + "step": 1885 + }, + { + "epoch": 1.8103696591454632, + "grad_norm": 1.3163333231081986, + "learning_rate": 8.053432328442483e-06, + "loss": 0.6915, + "step": 1886 + }, + { + "epoch": 1.8113298127700432, + "grad_norm": 1.3579706080157559, + "learning_rate": 8.05077890200752e-06, + "loss": 0.6606, + "step": 1887 + }, + { + "epoch": 1.8122899663946233, + "grad_norm": 1.1776867295102547, + "learning_rate": 8.04812410611052e-06, + "loss": 0.6988, + "step": 1888 + }, + { + "epoch": 1.813250120019203, + "grad_norm": 1.2589174346369656, + "learning_rate": 8.045467941943193e-06, + "loss": 0.5862, + "step": 1889 + }, + { + "epoch": 1.8142102736437828, + "grad_norm": 1.2265270873184735, + "learning_rate": 8.042810410697861e-06, + "loss": 0.6179, + "step": 1890 + }, + { + "epoch": 1.815170427268363, + "grad_norm": 1.0133059812027188, + "learning_rate": 8.04015151356746e-06, + "loss": 0.6033, + "step": 1891 + }, + { + "epoch": 1.8161305808929429, + "grad_norm": 0.8791820381095383, + "learning_rate": 8.037491251745543e-06, + "loss": 0.7375, + "step": 1892 + }, + { + "epoch": 1.8170907345175227, + "grad_norm": 1.040871201272267, + "learning_rate": 8.034829626426273e-06, + "loss": 0.6059, + "step": 1893 + }, + { + "epoch": 1.8180508881421027, + "grad_norm": 1.2459478903863541, + "learning_rate": 8.032166638804422e-06, + "loss": 0.7627, + "step": 1894 + }, + { + "epoch": 1.8190110417666827, + "grad_norm": 1.1669011277696746, + "learning_rate": 8.02950229007538e-06, + "loss": 0.6767, + "step": 1895 + }, + { + "epoch": 1.8199711953912625, + "grad_norm": 1.116550200147148, + "learning_rate": 8.026836581435142e-06, + "loss": 0.6242, + "step": 1896 + }, + { + "epoch": 1.8209313490158425, + "grad_norm": 1.005814348522918, + "learning_rate": 8.02416951408032e-06, + "loss": 0.586, + "step": 1897 + }, + { + "epoch": 1.8218915026404225, + "grad_norm": 1.3295399752112438, + "learning_rate": 8.02150108920813e-06, + "loss": 0.6137, + "step": 1898 + }, + { + "epoch": 1.8228516562650023, + "grad_norm": 1.3126930311763862, + "learning_rate": 8.018831308016398e-06, + "loss": 0.7176, + "step": 1899 + }, + { + "epoch": 1.8238118098895824, + "grad_norm": 1.4757208629612801, + "learning_rate": 8.01616017170356e-06, + "loss": 0.6723, + "step": 1900 + }, + { + "epoch": 1.8247719635141624, + "grad_norm": 0.7801099357841741, + "learning_rate": 8.013487681468663e-06, + "loss": 0.5786, + "step": 1901 + }, + { + "epoch": 1.8257321171387422, + "grad_norm": 1.0134564029272144, + "learning_rate": 8.010813838511357e-06, + "loss": 0.6652, + "step": 1902 + }, + { + "epoch": 1.826692270763322, + "grad_norm": 0.8964154794897466, + "learning_rate": 8.008138644031906e-06, + "loss": 0.4801, + "step": 1903 + }, + { + "epoch": 1.8276524243879022, + "grad_norm": 0.9900836513820425, + "learning_rate": 8.005462099231171e-06, + "loss": 0.6187, + "step": 1904 + }, + { + "epoch": 1.828612578012482, + "grad_norm": 1.0219218374611196, + "learning_rate": 8.002784205310626e-06, + "loss": 0.4844, + "step": 1905 + }, + { + "epoch": 1.8295727316370618, + "grad_norm": 0.9937049664117276, + "learning_rate": 8.00010496347235e-06, + "loss": 0.81, + "step": 1906 + }, + { + "epoch": 1.8305328852616418, + "grad_norm": 1.3696912850171175, + "learning_rate": 7.997424374919024e-06, + "loss": 0.6047, + "step": 1907 + }, + { + "epoch": 1.8314930388862218, + "grad_norm": 0.8695888682231789, + "learning_rate": 7.994742440853937e-06, + "loss": 0.5507, + "step": 1908 + }, + { + "epoch": 1.8324531925108016, + "grad_norm": 0.8029912452351989, + "learning_rate": 7.992059162480983e-06, + "loss": 0.7144, + "step": 1909 + }, + { + "epoch": 1.8334133461353816, + "grad_norm": 1.024031315206989, + "learning_rate": 7.98937454100465e-06, + "loss": 0.7038, + "step": 1910 + }, + { + "epoch": 1.8343734997599617, + "grad_norm": 0.9379701292046752, + "learning_rate": 7.986688577630042e-06, + "loss": 0.6368, + "step": 1911 + }, + { + "epoch": 1.8353336533845415, + "grad_norm": 1.1626921746350636, + "learning_rate": 7.984001273562858e-06, + "loss": 0.4876, + "step": 1912 + }, + { + "epoch": 1.8362938070091215, + "grad_norm": 0.765093029375882, + "learning_rate": 7.981312630009398e-06, + "loss": 0.5535, + "step": 1913 + }, + { + "epoch": 1.8372539606337015, + "grad_norm": 1.493058303836218, + "learning_rate": 7.97862264817657e-06, + "loss": 0.6126, + "step": 1914 + }, + { + "epoch": 1.8382141142582813, + "grad_norm": 1.0474886899563316, + "learning_rate": 7.975931329271874e-06, + "loss": 0.627, + "step": 1915 + }, + { + "epoch": 1.839174267882861, + "grad_norm": 1.0060827978459788, + "learning_rate": 7.973238674503412e-06, + "loss": 0.5583, + "step": 1916 + }, + { + "epoch": 1.8401344215074413, + "grad_norm": 0.8485255878794656, + "learning_rate": 7.970544685079895e-06, + "loss": 0.6075, + "step": 1917 + }, + { + "epoch": 1.8410945751320211, + "grad_norm": 1.0948252496214814, + "learning_rate": 7.96784936221062e-06, + "loss": 0.5305, + "step": 1918 + }, + { + "epoch": 1.842054728756601, + "grad_norm": 1.1354876526540547, + "learning_rate": 7.965152707105491e-06, + "loss": 0.5774, + "step": 1919 + }, + { + "epoch": 1.843014882381181, + "grad_norm": 1.2082984729614092, + "learning_rate": 7.962454720975008e-06, + "loss": 0.694, + "step": 1920 + }, + { + "epoch": 1.843975036005761, + "grad_norm": 1.1873456218694367, + "learning_rate": 7.959755405030269e-06, + "loss": 0.6698, + "step": 1921 + }, + { + "epoch": 1.8449351896303408, + "grad_norm": 1.1345266713800688, + "learning_rate": 7.957054760482964e-06, + "loss": 0.6462, + "step": 1922 + }, + { + "epoch": 1.8458953432549208, + "grad_norm": 0.9554105867510718, + "learning_rate": 7.954352788545388e-06, + "loss": 0.5843, + "step": 1923 + }, + { + "epoch": 1.8468554968795008, + "grad_norm": 0.9468802876993183, + "learning_rate": 7.951649490430424e-06, + "loss": 0.6401, + "step": 1924 + }, + { + "epoch": 1.8478156505040806, + "grad_norm": 1.0926269329377654, + "learning_rate": 7.948944867351556e-06, + "loss": 0.6098, + "step": 1925 + }, + { + "epoch": 1.8487758041286606, + "grad_norm": 0.8654462228441753, + "learning_rate": 7.946238920522862e-06, + "loss": 0.7336, + "step": 1926 + }, + { + "epoch": 1.8497359577532406, + "grad_norm": 0.8394409652705445, + "learning_rate": 7.943531651159007e-06, + "loss": 0.6228, + "step": 1927 + }, + { + "epoch": 1.8506961113778204, + "grad_norm": 0.8727248717468625, + "learning_rate": 7.940823060475257e-06, + "loss": 0.6318, + "step": 1928 + }, + { + "epoch": 1.8516562650024004, + "grad_norm": 1.0237836414075832, + "learning_rate": 7.938113149687474e-06, + "loss": 0.6455, + "step": 1929 + }, + { + "epoch": 1.8526164186269805, + "grad_norm": 1.598106761224662, + "learning_rate": 7.935401920012105e-06, + "loss": 0.6924, + "step": 1930 + }, + { + "epoch": 1.8535765722515603, + "grad_norm": 1.154091511674192, + "learning_rate": 7.932689372666192e-06, + "loss": 0.6076, + "step": 1931 + }, + { + "epoch": 1.85453672587614, + "grad_norm": 0.9933962172027189, + "learning_rate": 7.929975508867364e-06, + "loss": 0.6653, + "step": 1932 + }, + { + "epoch": 1.85549687950072, + "grad_norm": 1.0289186292742585, + "learning_rate": 7.927260329833854e-06, + "loss": 0.6823, + "step": 1933 + }, + { + "epoch": 1.8564570331253, + "grad_norm": 0.9757375555786784, + "learning_rate": 7.924543836784471e-06, + "loss": 0.5639, + "step": 1934 + }, + { + "epoch": 1.8574171867498799, + "grad_norm": 0.9313246628917996, + "learning_rate": 7.921826030938623e-06, + "loss": 0.6496, + "step": 1935 + }, + { + "epoch": 1.85837734037446, + "grad_norm": 0.9904979867772578, + "learning_rate": 7.919106913516299e-06, + "loss": 0.6132, + "step": 1936 + }, + { + "epoch": 1.85933749399904, + "grad_norm": 1.3040673337594413, + "learning_rate": 7.916386485738089e-06, + "loss": 0.6787, + "step": 1937 + }, + { + "epoch": 1.8602976476236197, + "grad_norm": 1.0202838458263193, + "learning_rate": 7.913664748825156e-06, + "loss": 0.6105, + "step": 1938 + }, + { + "epoch": 1.8612578012481997, + "grad_norm": 1.137407655299372, + "learning_rate": 7.910941703999265e-06, + "loss": 0.6515, + "step": 1939 + }, + { + "epoch": 1.8622179548727797, + "grad_norm": 1.1799016075204445, + "learning_rate": 7.908217352482759e-06, + "loss": 0.7167, + "step": 1940 + }, + { + "epoch": 1.8631781084973595, + "grad_norm": 1.376572934679239, + "learning_rate": 7.905491695498571e-06, + "loss": 0.6602, + "step": 1941 + }, + { + "epoch": 1.8641382621219396, + "grad_norm": 0.9024551367490704, + "learning_rate": 7.902764734270219e-06, + "loss": 0.6101, + "step": 1942 + }, + { + "epoch": 1.8650984157465196, + "grad_norm": 1.2623165248007306, + "learning_rate": 7.900036470021805e-06, + "loss": 0.613, + "step": 1943 + }, + { + "epoch": 1.8660585693710994, + "grad_norm": 0.8908534019026104, + "learning_rate": 7.89730690397802e-06, + "loss": 0.7345, + "step": 1944 + }, + { + "epoch": 1.8670187229956792, + "grad_norm": 1.2219964995820927, + "learning_rate": 7.894576037364136e-06, + "loss": 0.6142, + "step": 1945 + }, + { + "epoch": 1.8679788766202592, + "grad_norm": 1.1701283331410564, + "learning_rate": 7.89184387140601e-06, + "loss": 0.7137, + "step": 1946 + }, + { + "epoch": 1.8689390302448392, + "grad_norm": 1.1780638161310644, + "learning_rate": 7.889110407330083e-06, + "loss": 0.6452, + "step": 1947 + }, + { + "epoch": 1.869899183869419, + "grad_norm": 0.8909626695365555, + "learning_rate": 7.886375646363378e-06, + "loss": 0.6545, + "step": 1948 + }, + { + "epoch": 1.870859337493999, + "grad_norm": 1.1229937869782511, + "learning_rate": 7.883639589733498e-06, + "loss": 0.7138, + "step": 1949 + }, + { + "epoch": 1.871819491118579, + "grad_norm": 0.7759493102845842, + "learning_rate": 7.880902238668631e-06, + "loss": 0.6015, + "step": 1950 + }, + { + "epoch": 1.8727796447431588, + "grad_norm": 1.098330681816144, + "learning_rate": 7.878163594397543e-06, + "loss": 0.7063, + "step": 1951 + }, + { + "epoch": 1.8737397983677389, + "grad_norm": 1.1044481021724823, + "learning_rate": 7.875423658149583e-06, + "loss": 0.6902, + "step": 1952 + }, + { + "epoch": 1.8746999519923189, + "grad_norm": 1.2663156630113597, + "learning_rate": 7.87268243115468e-06, + "loss": 0.7063, + "step": 1953 + }, + { + "epoch": 1.8756601056168987, + "grad_norm": 1.4520521258078687, + "learning_rate": 7.869939914643343e-06, + "loss": 0.7229, + "step": 1954 + }, + { + "epoch": 1.8766202592414787, + "grad_norm": 1.0617459167629775, + "learning_rate": 7.867196109846653e-06, + "loss": 0.6476, + "step": 1955 + }, + { + "epoch": 1.8775804128660587, + "grad_norm": 1.3958866985961658, + "learning_rate": 7.86445101799628e-06, + "loss": 0.6608, + "step": 1956 + }, + { + "epoch": 1.8785405664906385, + "grad_norm": 1.2206399228833131, + "learning_rate": 7.861704640324466e-06, + "loss": 0.5386, + "step": 1957 + }, + { + "epoch": 1.8795007201152183, + "grad_norm": 1.0046968552784716, + "learning_rate": 7.858956978064025e-06, + "loss": 0.5919, + "step": 1958 + }, + { + "epoch": 1.8804608737397985, + "grad_norm": 1.0151619535695873, + "learning_rate": 7.85620803244836e-06, + "loss": 0.6173, + "step": 1959 + }, + { + "epoch": 1.8814210273643783, + "grad_norm": 1.0207747206779128, + "learning_rate": 7.85345780471144e-06, + "loss": 0.6329, + "step": 1960 + }, + { + "epoch": 1.8823811809889581, + "grad_norm": 0.9326902851316377, + "learning_rate": 7.850706296087811e-06, + "loss": 0.6369, + "step": 1961 + }, + { + "epoch": 1.8833413346135381, + "grad_norm": 1.0361948400957086, + "learning_rate": 7.8479535078126e-06, + "loss": 0.5498, + "step": 1962 + }, + { + "epoch": 1.8843014882381182, + "grad_norm": 0.957960449142528, + "learning_rate": 7.8451994411215e-06, + "loss": 0.6353, + "step": 1963 + }, + { + "epoch": 1.885261641862698, + "grad_norm": 0.9019753967112145, + "learning_rate": 7.842444097250787e-06, + "loss": 0.6288, + "step": 1964 + }, + { + "epoch": 1.886221795487278, + "grad_norm": 1.0750655440090873, + "learning_rate": 7.839687477437304e-06, + "loss": 0.5878, + "step": 1965 + }, + { + "epoch": 1.887181949111858, + "grad_norm": 0.9259874258322465, + "learning_rate": 7.836929582918464e-06, + "loss": 0.6749, + "step": 1966 + }, + { + "epoch": 1.8881421027364378, + "grad_norm": 1.147776615043257, + "learning_rate": 7.834170414932259e-06, + "loss": 0.5769, + "step": 1967 + }, + { + "epoch": 1.8891022563610178, + "grad_norm": 0.8621825258564246, + "learning_rate": 7.831409974717252e-06, + "loss": 0.5565, + "step": 1968 + }, + { + "epoch": 1.8900624099855978, + "grad_norm": 1.2248913855444168, + "learning_rate": 7.828648263512573e-06, + "loss": 0.6344, + "step": 1969 + }, + { + "epoch": 1.8910225636101776, + "grad_norm": 1.2907510720676452, + "learning_rate": 7.825885282557926e-06, + "loss": 0.6224, + "step": 1970 + }, + { + "epoch": 1.8919827172347574, + "grad_norm": 1.1862880222849024, + "learning_rate": 7.823121033093581e-06, + "loss": 0.6892, + "step": 1971 + }, + { + "epoch": 1.8929428708593377, + "grad_norm": 1.002890596821769, + "learning_rate": 7.820355516360383e-06, + "loss": 0.6472, + "step": 1972 + }, + { + "epoch": 1.8939030244839175, + "grad_norm": 1.2047448111603778, + "learning_rate": 7.817588733599742e-06, + "loss": 0.6551, + "step": 1973 + }, + { + "epoch": 1.8948631781084972, + "grad_norm": 0.9525275129416088, + "learning_rate": 7.814820686053634e-06, + "loss": 0.6269, + "step": 1974 + }, + { + "epoch": 1.8958233317330773, + "grad_norm": 0.6923440757128753, + "learning_rate": 7.812051374964611e-06, + "loss": 0.5931, + "step": 1975 + }, + { + "epoch": 1.8967834853576573, + "grad_norm": 0.8917493964077245, + "learning_rate": 7.809280801575784e-06, + "loss": 0.6596, + "step": 1976 + }, + { + "epoch": 1.897743638982237, + "grad_norm": 1.044122606482162, + "learning_rate": 7.806508967130838e-06, + "loss": 0.6629, + "step": 1977 + }, + { + "epoch": 1.898703792606817, + "grad_norm": 0.8740264354924745, + "learning_rate": 7.803735872874012e-06, + "loss": 0.6139, + "step": 1978 + }, + { + "epoch": 1.8996639462313971, + "grad_norm": 1.0016841226101536, + "learning_rate": 7.800961520050125e-06, + "loss": 0.6834, + "step": 1979 + }, + { + "epoch": 1.900624099855977, + "grad_norm": 0.8857095140329992, + "learning_rate": 7.798185909904552e-06, + "loss": 0.6602, + "step": 1980 + }, + { + "epoch": 1.901584253480557, + "grad_norm": 0.9550896040999265, + "learning_rate": 7.795409043683237e-06, + "loss": 0.6896, + "step": 1981 + }, + { + "epoch": 1.902544407105137, + "grad_norm": 0.8851191160717644, + "learning_rate": 7.792630922632682e-06, + "loss": 0.5623, + "step": 1982 + }, + { + "epoch": 1.9035045607297167, + "grad_norm": 1.1502009947200451, + "learning_rate": 7.789851547999957e-06, + "loss": 0.7963, + "step": 1983 + }, + { + "epoch": 1.9044647143542965, + "grad_norm": 1.1733042609983582, + "learning_rate": 7.787070921032693e-06, + "loss": 0.647, + "step": 1984 + }, + { + "epoch": 1.9054248679788768, + "grad_norm": 0.9294392569689884, + "learning_rate": 7.784289042979085e-06, + "loss": 0.6825, + "step": 1985 + }, + { + "epoch": 1.9063850216034566, + "grad_norm": 0.9934211403486944, + "learning_rate": 7.781505915087891e-06, + "loss": 0.6155, + "step": 1986 + }, + { + "epoch": 1.9073451752280364, + "grad_norm": 1.1918964161217966, + "learning_rate": 7.778721538608422e-06, + "loss": 0.6664, + "step": 1987 + }, + { + "epoch": 1.9083053288526164, + "grad_norm": 1.0426428053082084, + "learning_rate": 7.775935914790557e-06, + "loss": 0.6234, + "step": 1988 + }, + { + "epoch": 1.9092654824771964, + "grad_norm": 1.0510493862237833, + "learning_rate": 7.773149044884732e-06, + "loss": 0.7403, + "step": 1989 + }, + { + "epoch": 1.9102256361017762, + "grad_norm": 1.5115286509769619, + "learning_rate": 7.770360930141946e-06, + "loss": 0.6588, + "step": 1990 + }, + { + "epoch": 1.9111857897263562, + "grad_norm": 1.1302363335567547, + "learning_rate": 7.767571571813752e-06, + "loss": 0.5889, + "step": 1991 + }, + { + "epoch": 1.9121459433509362, + "grad_norm": 0.8397706236392392, + "learning_rate": 7.764780971152262e-06, + "loss": 0.5496, + "step": 1992 + }, + { + "epoch": 1.913106096975516, + "grad_norm": 1.0688534481923766, + "learning_rate": 7.761989129410149e-06, + "loss": 0.6942, + "step": 1993 + }, + { + "epoch": 1.914066250600096, + "grad_norm": 1.1145804131889687, + "learning_rate": 7.759196047840639e-06, + "loss": 0.7286, + "step": 1994 + }, + { + "epoch": 1.915026404224676, + "grad_norm": 1.238021302805448, + "learning_rate": 7.756401727697518e-06, + "loss": 0.6105, + "step": 1995 + }, + { + "epoch": 1.9159865578492559, + "grad_norm": 1.1728378204701007, + "learning_rate": 7.753606170235124e-06, + "loss": 0.6221, + "step": 1996 + }, + { + "epoch": 1.9169467114738357, + "grad_norm": 1.083706528150401, + "learning_rate": 7.750809376708357e-06, + "loss": 0.6819, + "step": 1997 + }, + { + "epoch": 1.917906865098416, + "grad_norm": 1.1110884491237523, + "learning_rate": 7.748011348372663e-06, + "loss": 0.6413, + "step": 1998 + }, + { + "epoch": 1.9188670187229957, + "grad_norm": 1.0215471235052087, + "learning_rate": 7.74521208648405e-06, + "loss": 0.719, + "step": 1999 + }, + { + "epoch": 1.9198271723475755, + "grad_norm": 0.9261412164580689, + "learning_rate": 7.742411592299077e-06, + "loss": 0.6255, + "step": 2000 + }, + { + "epoch": 1.9207873259721555, + "grad_norm": 0.872463961667252, + "learning_rate": 7.739609867074856e-06, + "loss": 0.634, + "step": 2001 + }, + { + "epoch": 1.9217474795967355, + "grad_norm": 1.0863303089243082, + "learning_rate": 7.736806912069048e-06, + "loss": 0.6633, + "step": 2002 + }, + { + "epoch": 1.9227076332213153, + "grad_norm": 0.8139746210348665, + "learning_rate": 7.734002728539873e-06, + "loss": 0.6113, + "step": 2003 + }, + { + "epoch": 1.9236677868458953, + "grad_norm": 0.8817300854256722, + "learning_rate": 7.731197317746099e-06, + "loss": 0.5994, + "step": 2004 + }, + { + "epoch": 1.9246279404704754, + "grad_norm": 1.136222412651614, + "learning_rate": 7.728390680947043e-06, + "loss": 0.655, + "step": 2005 + }, + { + "epoch": 1.9255880940950552, + "grad_norm": 1.2325785352151544, + "learning_rate": 7.725582819402575e-06, + "loss": 0.6905, + "step": 2006 + }, + { + "epoch": 1.9265482477196352, + "grad_norm": 1.0641373207977778, + "learning_rate": 7.722773734373114e-06, + "loss": 0.587, + "step": 2007 + }, + { + "epoch": 1.9275084013442152, + "grad_norm": 1.0629082498817422, + "learning_rate": 7.71996342711963e-06, + "loss": 0.6946, + "step": 2008 + }, + { + "epoch": 1.928468554968795, + "grad_norm": 1.5321764093757542, + "learning_rate": 7.717151898903636e-06, + "loss": 0.596, + "step": 2009 + }, + { + "epoch": 1.9294287085933748, + "grad_norm": 1.2130629515284264, + "learning_rate": 7.714339150987202e-06, + "loss": 0.6919, + "step": 2010 + }, + { + "epoch": 1.930388862217955, + "grad_norm": 1.1116261332670805, + "learning_rate": 7.711525184632936e-06, + "loss": 0.7311, + "step": 2011 + }, + { + "epoch": 1.9313490158425348, + "grad_norm": 1.1953661542991707, + "learning_rate": 7.708710001103998e-06, + "loss": 0.6029, + "step": 2012 + }, + { + "epoch": 1.9323091694671146, + "grad_norm": 1.497513151133857, + "learning_rate": 7.705893601664099e-06, + "loss": 0.6562, + "step": 2013 + }, + { + "epoch": 1.9332693230916946, + "grad_norm": 1.2799348499546876, + "learning_rate": 7.703075987577483e-06, + "loss": 0.7839, + "step": 2014 + }, + { + "epoch": 1.9342294767162747, + "grad_norm": 1.525526033893843, + "learning_rate": 7.700257160108953e-06, + "loss": 0.6587, + "step": 2015 + }, + { + "epoch": 1.9351896303408544, + "grad_norm": 1.1700523940769612, + "learning_rate": 7.697437120523845e-06, + "loss": 0.7259, + "step": 2016 + }, + { + "epoch": 1.9361497839654345, + "grad_norm": 1.0531375077014808, + "learning_rate": 7.694615870088051e-06, + "loss": 0.7215, + "step": 2017 + }, + { + "epoch": 1.9371099375900145, + "grad_norm": 1.3969434971840897, + "learning_rate": 7.691793410067993e-06, + "loss": 0.5855, + "step": 2018 + }, + { + "epoch": 1.9380700912145943, + "grad_norm": 1.3604813865106966, + "learning_rate": 7.68896974173065e-06, + "loss": 0.6194, + "step": 2019 + }, + { + "epoch": 1.9390302448391743, + "grad_norm": 0.9952031977741248, + "learning_rate": 7.686144866343533e-06, + "loss": 0.5427, + "step": 2020 + }, + { + "epoch": 1.9399903984637543, + "grad_norm": 0.812314858579823, + "learning_rate": 7.683318785174698e-06, + "loss": 0.5197, + "step": 2021 + }, + { + "epoch": 1.9409505520883341, + "grad_norm": 0.95987119699267, + "learning_rate": 7.680491499492743e-06, + "loss": 0.6596, + "step": 2022 + }, + { + "epoch": 1.941910705712914, + "grad_norm": 1.2319589747848858, + "learning_rate": 7.677663010566807e-06, + "loss": 0.6821, + "step": 2023 + }, + { + "epoch": 1.9428708593374941, + "grad_norm": 0.6606150963033354, + "learning_rate": 7.674833319666568e-06, + "loss": 0.6038, + "step": 2024 + }, + { + "epoch": 1.943831012962074, + "grad_norm": 1.200520352178802, + "learning_rate": 7.672002428062245e-06, + "loss": 0.653, + "step": 2025 + }, + { + "epoch": 1.9447911665866537, + "grad_norm": 0.7192438014804379, + "learning_rate": 7.669170337024593e-06, + "loss": 0.4693, + "step": 2026 + }, + { + "epoch": 1.9457513202112338, + "grad_norm": 1.4486619176506674, + "learning_rate": 7.66633704782491e-06, + "loss": 0.7564, + "step": 2027 + }, + { + "epoch": 1.9467114738358138, + "grad_norm": 1.2242450439191879, + "learning_rate": 7.663502561735028e-06, + "loss": 0.733, + "step": 2028 + }, + { + "epoch": 1.9476716274603936, + "grad_norm": 1.1407736193961824, + "learning_rate": 7.660666880027316e-06, + "loss": 0.593, + "step": 2029 + }, + { + "epoch": 1.9486317810849736, + "grad_norm": 1.1271456340334174, + "learning_rate": 7.657830003974684e-06, + "loss": 0.644, + "step": 2030 + }, + { + "epoch": 1.9495919347095536, + "grad_norm": 0.846049078907581, + "learning_rate": 7.654991934850573e-06, + "loss": 0.5331, + "step": 2031 + }, + { + "epoch": 1.9505520883341334, + "grad_norm": 0.7872449106149692, + "learning_rate": 7.652152673928964e-06, + "loss": 0.6157, + "step": 2032 + }, + { + "epoch": 1.9515122419587134, + "grad_norm": 0.9912770009275691, + "learning_rate": 7.649312222484367e-06, + "loss": 0.6688, + "step": 2033 + }, + { + "epoch": 1.9524723955832934, + "grad_norm": 1.0527641311770262, + "learning_rate": 7.646470581791834e-06, + "loss": 0.7184, + "step": 2034 + }, + { + "epoch": 1.9534325492078732, + "grad_norm": 0.9104283901825249, + "learning_rate": 7.643627753126946e-06, + "loss": 0.6, + "step": 2035 + }, + { + "epoch": 1.954392702832453, + "grad_norm": 1.2765136056418498, + "learning_rate": 7.640783737765817e-06, + "loss": 0.7672, + "step": 2036 + }, + { + "epoch": 1.9553528564570333, + "grad_norm": 1.4204717911379219, + "learning_rate": 7.6379385369851e-06, + "loss": 0.6158, + "step": 2037 + }, + { + "epoch": 1.956313010081613, + "grad_norm": 1.9379913067353052, + "learning_rate": 7.63509215206197e-06, + "loss": 0.7316, + "step": 2038 + }, + { + "epoch": 1.9572731637061929, + "grad_norm": 1.1268648862052064, + "learning_rate": 7.63224458427414e-06, + "loss": 0.4365, + "step": 2039 + }, + { + "epoch": 1.9582333173307729, + "grad_norm": 0.7974068444813239, + "learning_rate": 7.629395834899852e-06, + "loss": 0.477, + "step": 2040 + }, + { + "epoch": 1.959193470955353, + "grad_norm": 1.1109805277309888, + "learning_rate": 7.626545905217884e-06, + "loss": 0.6397, + "step": 2041 + }, + { + "epoch": 1.9601536245799327, + "grad_norm": 1.2398850876268863, + "learning_rate": 7.6236947965075325e-06, + "loss": 0.6679, + "step": 2042 + }, + { + "epoch": 1.9611137782045127, + "grad_norm": 1.1737119231817035, + "learning_rate": 7.6208425100486334e-06, + "loss": 0.6392, + "step": 2043 + }, + { + "epoch": 1.9620739318290927, + "grad_norm": 0.8854080472389517, + "learning_rate": 7.617989047121548e-06, + "loss": 0.6596, + "step": 2044 + }, + { + "epoch": 1.9630340854536725, + "grad_norm": 0.7676659103912914, + "learning_rate": 7.6151344090071624e-06, + "loss": 0.5967, + "step": 2045 + }, + { + "epoch": 1.9639942390782525, + "grad_norm": 1.0291718099826375, + "learning_rate": 7.612278596986895e-06, + "loss": 0.7092, + "step": 2046 + }, + { + "epoch": 1.9649543927028326, + "grad_norm": 0.7766136208698822, + "learning_rate": 7.609421612342689e-06, + "loss": 0.6606, + "step": 2047 + }, + { + "epoch": 1.9659145463274124, + "grad_norm": 1.1861269837590105, + "learning_rate": 7.606563456357016e-06, + "loss": 0.7037, + "step": 2048 + }, + { + "epoch": 1.9668746999519922, + "grad_norm": 1.0408178029162052, + "learning_rate": 7.603704130312868e-06, + "loss": 0.7429, + "step": 2049 + }, + { + "epoch": 1.9678348535765724, + "grad_norm": 1.1869576752680744, + "learning_rate": 7.6008436354937695e-06, + "loss": 0.7273, + "step": 2050 + }, + { + "epoch": 1.9687950072011522, + "grad_norm": 1.3051495633501524, + "learning_rate": 7.5979819731837635e-06, + "loss": 0.6918, + "step": 2051 + }, + { + "epoch": 1.969755160825732, + "grad_norm": 0.7596205378022328, + "learning_rate": 7.595119144667421e-06, + "loss": 0.5797, + "step": 2052 + }, + { + "epoch": 1.970715314450312, + "grad_norm": 0.9586448051063569, + "learning_rate": 7.5922551512298345e-06, + "loss": 0.689, + "step": 2053 + }, + { + "epoch": 1.971675468074892, + "grad_norm": 0.8801070575115794, + "learning_rate": 7.589389994156622e-06, + "loss": 0.6686, + "step": 2054 + }, + { + "epoch": 1.9726356216994718, + "grad_norm": 1.3509221888327454, + "learning_rate": 7.586523674733918e-06, + "loss": 0.6191, + "step": 2055 + }, + { + "epoch": 1.9735957753240518, + "grad_norm": 1.1120138619462479, + "learning_rate": 7.583656194248386e-06, + "loss": 0.5759, + "step": 2056 + }, + { + "epoch": 1.9745559289486319, + "grad_norm": 1.0233499758935727, + "learning_rate": 7.5807875539872054e-06, + "loss": 0.5934, + "step": 2057 + }, + { + "epoch": 1.9755160825732117, + "grad_norm": 1.3666746387565665, + "learning_rate": 7.577917755238079e-06, + "loss": 0.7105, + "step": 2058 + }, + { + "epoch": 1.9764762361977917, + "grad_norm": 1.1636696265207904, + "learning_rate": 7.57504679928923e-06, + "loss": 0.7389, + "step": 2059 + }, + { + "epoch": 1.9774363898223717, + "grad_norm": 1.1833938285444674, + "learning_rate": 7.572174687429395e-06, + "loss": 0.6764, + "step": 2060 + }, + { + "epoch": 1.9783965434469515, + "grad_norm": 1.0399387517184737, + "learning_rate": 7.569301420947841e-06, + "loss": 0.5481, + "step": 2061 + }, + { + "epoch": 1.9793566970715313, + "grad_norm": 0.9375153941817806, + "learning_rate": 7.566427001134342e-06, + "loss": 0.7324, + "step": 2062 + }, + { + "epoch": 1.9803168506961115, + "grad_norm": 1.19747434705813, + "learning_rate": 7.563551429279198e-06, + "loss": 0.6257, + "step": 2063 + }, + { + "epoch": 1.9812770043206913, + "grad_norm": 1.3612075834492465, + "learning_rate": 7.560674706673219e-06, + "loss": 0.5541, + "step": 2064 + }, + { + "epoch": 1.982237157945271, + "grad_norm": 1.2460446469104085, + "learning_rate": 7.557796834607736e-06, + "loss": 0.7148, + "step": 2065 + }, + { + "epoch": 1.9831973115698511, + "grad_norm": 1.2505377581381347, + "learning_rate": 7.554917814374595e-06, + "loss": 0.6051, + "step": 2066 + }, + { + "epoch": 1.9841574651944311, + "grad_norm": 1.3178231047232203, + "learning_rate": 7.552037647266157e-06, + "loss": 0.666, + "step": 2067 + }, + { + "epoch": 1.985117618819011, + "grad_norm": 1.2053322598268592, + "learning_rate": 7.5491563345753e-06, + "loss": 0.6736, + "step": 2068 + }, + { + "epoch": 1.986077772443591, + "grad_norm": 1.5118119390957352, + "learning_rate": 7.546273877595413e-06, + "loss": 0.6847, + "step": 2069 + }, + { + "epoch": 1.987037926068171, + "grad_norm": 1.0329897029908128, + "learning_rate": 7.5433902776204015e-06, + "loss": 0.6205, + "step": 2070 + }, + { + "epoch": 1.9879980796927508, + "grad_norm": 0.9974569099985744, + "learning_rate": 7.54050553594468e-06, + "loss": 0.5332, + "step": 2071 + }, + { + "epoch": 1.9889582333173308, + "grad_norm": 1.0690421060675295, + "learning_rate": 7.537619653863182e-06, + "loss": 0.7159, + "step": 2072 + }, + { + "epoch": 1.9899183869419108, + "grad_norm": 1.1632058889940493, + "learning_rate": 7.5347326326713454e-06, + "loss": 0.5751, + "step": 2073 + }, + { + "epoch": 1.9908785405664906, + "grad_norm": 1.422020334434976, + "learning_rate": 7.531844473665125e-06, + "loss": 0.7107, + "step": 2074 + }, + { + "epoch": 1.9918386941910706, + "grad_norm": 1.0079638983911854, + "learning_rate": 7.528955178140985e-06, + "loss": 0.5734, + "step": 2075 + }, + { + "epoch": 1.9927988478156506, + "grad_norm": 0.9123186568870082, + "learning_rate": 7.526064747395898e-06, + "loss": 0.6097, + "step": 2076 + }, + { + "epoch": 1.9937590014402304, + "grad_norm": 1.0309543923066218, + "learning_rate": 7.523173182727348e-06, + "loss": 0.5464, + "step": 2077 + }, + { + "epoch": 1.9947191550648102, + "grad_norm": 1.0854712934998054, + "learning_rate": 7.520280485433327e-06, + "loss": 0.7086, + "step": 2078 + }, + { + "epoch": 1.9956793086893903, + "grad_norm": 1.006296613540927, + "learning_rate": 7.517386656812338e-06, + "loss": 0.6084, + "step": 2079 + }, + { + "epoch": 1.9966394623139703, + "grad_norm": 0.915483611749352, + "learning_rate": 7.514491698163387e-06, + "loss": 0.6173, + "step": 2080 + }, + { + "epoch": 1.99759961593855, + "grad_norm": 1.0737464838300075, + "learning_rate": 7.511595610785991e-06, + "loss": 0.6215, + "step": 2081 + }, + { + "epoch": 1.99855976956313, + "grad_norm": 1.0588984534223536, + "learning_rate": 7.508698395980173e-06, + "loss": 0.7455, + "step": 2082 + }, + { + "epoch": 1.99951992318771, + "grad_norm": 1.0259188689456935, + "learning_rate": 7.5058000550464615e-06, + "loss": 0.5611, + "step": 2083 + }, + { + "epoch": 2.0, + "grad_norm": 1.3486689052941767, + "learning_rate": 7.502900589285891e-06, + "loss": 0.5107, + "step": 2084 + }, + { + "epoch": 2.00096015362458, + "grad_norm": 0.9242603330168586, + "learning_rate": 7.500000000000001e-06, + "loss": 0.54, + "step": 2085 + }, + { + "epoch": 2.00192030724916, + "grad_norm": 0.9543728385159, + "learning_rate": 7.497098288490834e-06, + "loss": 0.4942, + "step": 2086 + }, + { + "epoch": 2.00288046087374, + "grad_norm": 1.4659028048986025, + "learning_rate": 7.494195456060938e-06, + "loss": 0.5756, + "step": 2087 + }, + { + "epoch": 2.0038406144983196, + "grad_norm": 1.2257178724297157, + "learning_rate": 7.491291504013363e-06, + "loss": 0.4658, + "step": 2088 + }, + { + "epoch": 2.0048007681229, + "grad_norm": 0.8867977842650157, + "learning_rate": 7.488386433651662e-06, + "loss": 0.5053, + "step": 2089 + }, + { + "epoch": 2.0057609217474797, + "grad_norm": 1.192674325910438, + "learning_rate": 7.4854802462798905e-06, + "loss": 0.5525, + "step": 2090 + }, + { + "epoch": 2.0067210753720595, + "grad_norm": 1.0367888174874489, + "learning_rate": 7.482572943202604e-06, + "loss": 0.6436, + "step": 2091 + }, + { + "epoch": 2.0076812289966393, + "grad_norm": 1.0340421819235337, + "learning_rate": 7.4796645257248625e-06, + "loss": 0.5068, + "step": 2092 + }, + { + "epoch": 2.0086413826212195, + "grad_norm": 0.8516771088470378, + "learning_rate": 7.47675499515222e-06, + "loss": 0.4932, + "step": 2093 + }, + { + "epoch": 2.0096015362457993, + "grad_norm": 1.9015516331041107, + "learning_rate": 7.473844352790735e-06, + "loss": 0.6114, + "step": 2094 + }, + { + "epoch": 2.010561689870379, + "grad_norm": 1.2434236142855897, + "learning_rate": 7.470932599946965e-06, + "loss": 0.5542, + "step": 2095 + }, + { + "epoch": 2.0115218434949593, + "grad_norm": 1.1912469569660604, + "learning_rate": 7.468019737927963e-06, + "loss": 0.6125, + "step": 2096 + }, + { + "epoch": 2.012481997119539, + "grad_norm": 1.1075309095334798, + "learning_rate": 7.465105768041283e-06, + "loss": 0.5807, + "step": 2097 + }, + { + "epoch": 2.013442150744119, + "grad_norm": 1.3920065240518684, + "learning_rate": 7.462190691594973e-06, + "loss": 0.644, + "step": 2098 + }, + { + "epoch": 2.014402304368699, + "grad_norm": 1.0387681847196268, + "learning_rate": 7.459274509897583e-06, + "loss": 0.5815, + "step": 2099 + }, + { + "epoch": 2.015362457993279, + "grad_norm": 1.0086737416782097, + "learning_rate": 7.45635722425815e-06, + "loss": 0.6361, + "step": 2100 + }, + { + "epoch": 2.0163226116178588, + "grad_norm": 1.5131229782622566, + "learning_rate": 7.45343883598622e-06, + "loss": 0.4398, + "step": 2101 + }, + { + "epoch": 2.017282765242439, + "grad_norm": 0.9217517932823873, + "learning_rate": 7.45051934639182e-06, + "loss": 0.4969, + "step": 2102 + }, + { + "epoch": 2.018242918867019, + "grad_norm": 1.0006289807747306, + "learning_rate": 7.447598756785482e-06, + "loss": 0.5403, + "step": 2103 + }, + { + "epoch": 2.0192030724915986, + "grad_norm": 1.034626184859905, + "learning_rate": 7.444677068478223e-06, + "loss": 0.5091, + "step": 2104 + }, + { + "epoch": 2.0201632261161784, + "grad_norm": 0.9478952916646003, + "learning_rate": 7.441754282781561e-06, + "loss": 0.6067, + "step": 2105 + }, + { + "epoch": 2.0211233797407586, + "grad_norm": 1.1611452501373463, + "learning_rate": 7.438830401007504e-06, + "loss": 0.5924, + "step": 2106 + }, + { + "epoch": 2.0220835333653384, + "grad_norm": 1.316256258593302, + "learning_rate": 7.435905424468548e-06, + "loss": 0.5202, + "step": 2107 + }, + { + "epoch": 2.023043686989918, + "grad_norm": 1.23591737148139, + "learning_rate": 7.432979354477686e-06, + "loss": 0.5226, + "step": 2108 + }, + { + "epoch": 2.0240038406144985, + "grad_norm": 1.134488568665678, + "learning_rate": 7.430052192348398e-06, + "loss": 0.5593, + "step": 2109 + }, + { + "epoch": 2.0249639942390782, + "grad_norm": 0.873863500954897, + "learning_rate": 7.427123939394658e-06, + "loss": 0.5202, + "step": 2110 + }, + { + "epoch": 2.025924147863658, + "grad_norm": 1.5403396675611922, + "learning_rate": 7.424194596930925e-06, + "loss": 0.5958, + "step": 2111 + }, + { + "epoch": 2.0268843014882383, + "grad_norm": 1.4791655228119358, + "learning_rate": 7.421264166272152e-06, + "loss": 0.5972, + "step": 2112 + }, + { + "epoch": 2.027844455112818, + "grad_norm": 1.3053642943908599, + "learning_rate": 7.418332648733774e-06, + "loss": 0.5313, + "step": 2113 + }, + { + "epoch": 2.028804608737398, + "grad_norm": 1.3286714063074645, + "learning_rate": 7.415400045631723e-06, + "loss": 0.612, + "step": 2114 + }, + { + "epoch": 2.029764762361978, + "grad_norm": 0.9569340901030102, + "learning_rate": 7.41246635828241e-06, + "loss": 0.4132, + "step": 2115 + }, + { + "epoch": 2.030724915986558, + "grad_norm": 1.1873003684933803, + "learning_rate": 7.4095315880027365e-06, + "loss": 0.5841, + "step": 2116 + }, + { + "epoch": 2.0316850696111377, + "grad_norm": 1.5116393108346937, + "learning_rate": 7.406595736110089e-06, + "loss": 0.5357, + "step": 2117 + }, + { + "epoch": 2.0326452232357175, + "grad_norm": 1.0182481052961394, + "learning_rate": 7.4036588039223415e-06, + "loss": 0.6686, + "step": 2118 + }, + { + "epoch": 2.0336053768602977, + "grad_norm": 0.8425290443373554, + "learning_rate": 7.400720792757851e-06, + "loss": 0.5016, + "step": 2119 + }, + { + "epoch": 2.0345655304848775, + "grad_norm": 1.477362353438116, + "learning_rate": 7.397781703935458e-06, + "loss": 0.5692, + "step": 2120 + }, + { + "epoch": 2.0355256841094573, + "grad_norm": 1.4325155153609799, + "learning_rate": 7.39484153877449e-06, + "loss": 0.5293, + "step": 2121 + }, + { + "epoch": 2.0364858377340376, + "grad_norm": 0.9282735950305444, + "learning_rate": 7.391900298594752e-06, + "loss": 0.4984, + "step": 2122 + }, + { + "epoch": 2.0374459913586174, + "grad_norm": 0.9532270139954646, + "learning_rate": 7.388957984716541e-06, + "loss": 0.5398, + "step": 2123 + }, + { + "epoch": 2.038406144983197, + "grad_norm": 0.8762355601311758, + "learning_rate": 7.386014598460626e-06, + "loss": 0.5812, + "step": 2124 + }, + { + "epoch": 2.0393662986077774, + "grad_norm": 1.0739822627421551, + "learning_rate": 7.383070141148263e-06, + "loss": 0.5636, + "step": 2125 + }, + { + "epoch": 2.040326452232357, + "grad_norm": 1.2489651649480997, + "learning_rate": 7.380124614101185e-06, + "loss": 0.5317, + "step": 2126 + }, + { + "epoch": 2.041286605856937, + "grad_norm": 1.2927524784687077, + "learning_rate": 7.377178018641613e-06, + "loss": 0.5812, + "step": 2127 + }, + { + "epoch": 2.0422467594815172, + "grad_norm": 1.3946462494710277, + "learning_rate": 7.374230356092238e-06, + "loss": 0.5615, + "step": 2128 + }, + { + "epoch": 2.043206913106097, + "grad_norm": 0.9822202806600208, + "learning_rate": 7.371281627776237e-06, + "loss": 0.6693, + "step": 2129 + }, + { + "epoch": 2.044167066730677, + "grad_norm": 1.2864927131454644, + "learning_rate": 7.368331835017257e-06, + "loss": 0.5295, + "step": 2130 + }, + { + "epoch": 2.0451272203552566, + "grad_norm": 1.2331212956816664, + "learning_rate": 7.365380979139435e-06, + "loss": 0.6415, + "step": 2131 + }, + { + "epoch": 2.046087373979837, + "grad_norm": 1.1555702611744292, + "learning_rate": 7.362429061467377e-06, + "loss": 0.6135, + "step": 2132 + }, + { + "epoch": 2.0470475276044167, + "grad_norm": 1.0250907289031153, + "learning_rate": 7.359476083326164e-06, + "loss": 0.6169, + "step": 2133 + }, + { + "epoch": 2.0480076812289965, + "grad_norm": 1.0358471386936456, + "learning_rate": 7.35652204604136e-06, + "loss": 0.5128, + "step": 2134 + }, + { + "epoch": 2.0489678348535767, + "grad_norm": 1.0637155576382715, + "learning_rate": 7.353566950938997e-06, + "loss": 0.4894, + "step": 2135 + }, + { + "epoch": 2.0499279884781565, + "grad_norm": 1.5028422968427364, + "learning_rate": 7.35061079934559e-06, + "loss": 0.573, + "step": 2136 + }, + { + "epoch": 2.0508881421027363, + "grad_norm": 1.1093624220431402, + "learning_rate": 7.347653592588122e-06, + "loss": 0.5491, + "step": 2137 + }, + { + "epoch": 2.0518482957273165, + "grad_norm": 1.2646723948358842, + "learning_rate": 7.344695331994052e-06, + "loss": 0.4999, + "step": 2138 + }, + { + "epoch": 2.0528084493518963, + "grad_norm": 1.030349237091044, + "learning_rate": 7.341736018891308e-06, + "loss": 0.511, + "step": 2139 + }, + { + "epoch": 2.053768602976476, + "grad_norm": 1.1360884059780991, + "learning_rate": 7.338775654608299e-06, + "loss": 0.544, + "step": 2140 + }, + { + "epoch": 2.0547287566010564, + "grad_norm": 1.048952709164682, + "learning_rate": 7.335814240473899e-06, + "loss": 0.6217, + "step": 2141 + }, + { + "epoch": 2.055688910225636, + "grad_norm": 1.2728010363199347, + "learning_rate": 7.332851777817453e-06, + "loss": 0.4647, + "step": 2142 + }, + { + "epoch": 2.056649063850216, + "grad_norm": 1.157030872007594, + "learning_rate": 7.3298882679687836e-06, + "loss": 0.4886, + "step": 2143 + }, + { + "epoch": 2.0576092174747957, + "grad_norm": 0.9473979824746046, + "learning_rate": 7.326923712258173e-06, + "loss": 0.5894, + "step": 2144 + }, + { + "epoch": 2.058569371099376, + "grad_norm": 0.91389585244967, + "learning_rate": 7.323958112016383e-06, + "loss": 0.482, + "step": 2145 + }, + { + "epoch": 2.059529524723956, + "grad_norm": 0.9679054569476475, + "learning_rate": 7.320991468574637e-06, + "loss": 0.5391, + "step": 2146 + }, + { + "epoch": 2.0604896783485356, + "grad_norm": 1.3249677389529415, + "learning_rate": 7.318023783264633e-06, + "loss": 0.5262, + "step": 2147 + }, + { + "epoch": 2.061449831973116, + "grad_norm": 0.9183488536355273, + "learning_rate": 7.315055057418528e-06, + "loss": 0.4921, + "step": 2148 + }, + { + "epoch": 2.0624099855976956, + "grad_norm": 1.4264562988423015, + "learning_rate": 7.312085292368957e-06, + "loss": 0.4443, + "step": 2149 + }, + { + "epoch": 2.0633701392222754, + "grad_norm": 1.1406047321067216, + "learning_rate": 7.309114489449012e-06, + "loss": 0.5625, + "step": 2150 + }, + { + "epoch": 2.0643302928468557, + "grad_norm": 1.422593860911789, + "learning_rate": 7.306142649992256e-06, + "loss": 0.5264, + "step": 2151 + }, + { + "epoch": 2.0652904464714354, + "grad_norm": 1.0953180106941673, + "learning_rate": 7.3031697753327145e-06, + "loss": 0.5759, + "step": 2152 + }, + { + "epoch": 2.0662506000960152, + "grad_norm": 0.8941745426562419, + "learning_rate": 7.300195866804881e-06, + "loss": 0.4539, + "step": 2153 + }, + { + "epoch": 2.0672107537205955, + "grad_norm": 1.2100475289024843, + "learning_rate": 7.2972209257437125e-06, + "loss": 0.6848, + "step": 2154 + }, + { + "epoch": 2.0681709073451753, + "grad_norm": 1.3674083245206652, + "learning_rate": 7.2942449534846234e-06, + "loss": 0.5534, + "step": 2155 + }, + { + "epoch": 2.069131060969755, + "grad_norm": 0.9586407848894423, + "learning_rate": 7.291267951363501e-06, + "loss": 0.49, + "step": 2156 + }, + { + "epoch": 2.070091214594335, + "grad_norm": 1.0061110896808698, + "learning_rate": 7.288289920716686e-06, + "loss": 0.4975, + "step": 2157 + }, + { + "epoch": 2.071051368218915, + "grad_norm": 1.0432840568946489, + "learning_rate": 7.285310862880987e-06, + "loss": 0.575, + "step": 2158 + }, + { + "epoch": 2.072011521843495, + "grad_norm": 0.9872238049121338, + "learning_rate": 7.28233077919367e-06, + "loss": 0.5424, + "step": 2159 + }, + { + "epoch": 2.0729716754680747, + "grad_norm": 1.790221412169585, + "learning_rate": 7.279349670992464e-06, + "loss": 0.541, + "step": 2160 + }, + { + "epoch": 2.073931829092655, + "grad_norm": 1.032360706897286, + "learning_rate": 7.2763675396155545e-06, + "loss": 0.56, + "step": 2161 + }, + { + "epoch": 2.0748919827172347, + "grad_norm": 0.824803012243617, + "learning_rate": 7.27338438640159e-06, + "loss": 0.525, + "step": 2162 + }, + { + "epoch": 2.0758521363418145, + "grad_norm": 1.1790315047925064, + "learning_rate": 7.270400212689675e-06, + "loss": 0.577, + "step": 2163 + }, + { + "epoch": 2.0768122899663948, + "grad_norm": 1.2404278132966462, + "learning_rate": 7.2674150198193735e-06, + "loss": 0.5341, + "step": 2164 + }, + { + "epoch": 2.0777724435909746, + "grad_norm": 1.022275468924025, + "learning_rate": 7.264428809130708e-06, + "loss": 0.6559, + "step": 2165 + }, + { + "epoch": 2.0787325972155544, + "grad_norm": 1.2774007022863516, + "learning_rate": 7.2614415819641535e-06, + "loss": 0.5772, + "step": 2166 + }, + { + "epoch": 2.0796927508401346, + "grad_norm": 1.2454958358304937, + "learning_rate": 7.258453339660648e-06, + "loss": 0.6805, + "step": 2167 + }, + { + "epoch": 2.0806529044647144, + "grad_norm": 0.9254926320094997, + "learning_rate": 7.25546408356158e-06, + "loss": 0.5528, + "step": 2168 + }, + { + "epoch": 2.081613058089294, + "grad_norm": 1.1372835025449755, + "learning_rate": 7.252473815008794e-06, + "loss": 0.4636, + "step": 2169 + }, + { + "epoch": 2.082573211713874, + "grad_norm": 0.9132542974941551, + "learning_rate": 7.24948253534459e-06, + "loss": 0.4813, + "step": 2170 + }, + { + "epoch": 2.0835333653384542, + "grad_norm": 1.1585164013087945, + "learning_rate": 7.246490245911722e-06, + "loss": 0.4817, + "step": 2171 + }, + { + "epoch": 2.084493518963034, + "grad_norm": 0.9130120744811105, + "learning_rate": 7.243496948053394e-06, + "loss": 0.5988, + "step": 2172 + }, + { + "epoch": 2.085453672587614, + "grad_norm": 1.1587018142316117, + "learning_rate": 7.240502643113271e-06, + "loss": 0.664, + "step": 2173 + }, + { + "epoch": 2.086413826212194, + "grad_norm": 0.9006150195164618, + "learning_rate": 7.237507332435459e-06, + "loss": 0.5423, + "step": 2174 + }, + { + "epoch": 2.087373979836774, + "grad_norm": 0.946251970852323, + "learning_rate": 7.234511017364524e-06, + "loss": 0.6016, + "step": 2175 + }, + { + "epoch": 2.0883341334613537, + "grad_norm": 1.242912712579443, + "learning_rate": 7.2315136992454794e-06, + "loss": 0.6251, + "step": 2176 + }, + { + "epoch": 2.089294287085934, + "grad_norm": 1.1224268015584455, + "learning_rate": 7.228515379423787e-06, + "loss": 0.6026, + "step": 2177 + }, + { + "epoch": 2.0902544407105137, + "grad_norm": 0.8180908380728011, + "learning_rate": 7.225516059245365e-06, + "loss": 0.5581, + "step": 2178 + }, + { + "epoch": 2.0912145943350935, + "grad_norm": 0.9144789064207813, + "learning_rate": 7.222515740056571e-06, + "loss": 0.5401, + "step": 2179 + }, + { + "epoch": 2.0921747479596737, + "grad_norm": 1.2387776357857734, + "learning_rate": 7.219514423204219e-06, + "loss": 0.5254, + "step": 2180 + }, + { + "epoch": 2.0931349015842535, + "grad_norm": 1.5793470381166037, + "learning_rate": 7.216512110035567e-06, + "loss": 0.5215, + "step": 2181 + }, + { + "epoch": 2.0940950552088333, + "grad_norm": 1.0212283069024395, + "learning_rate": 7.213508801898325e-06, + "loss": 0.4813, + "step": 2182 + }, + { + "epoch": 2.095055208833413, + "grad_norm": 1.1890346255264337, + "learning_rate": 7.21050450014064e-06, + "loss": 0.5749, + "step": 2183 + }, + { + "epoch": 2.0960153624579934, + "grad_norm": 1.5029802892697517, + "learning_rate": 7.2074992061111145e-06, + "loss": 0.5671, + "step": 2184 + }, + { + "epoch": 2.096975516082573, + "grad_norm": 0.929036317951188, + "learning_rate": 7.204492921158794e-06, + "loss": 0.5416, + "step": 2185 + }, + { + "epoch": 2.097935669707153, + "grad_norm": 1.1339110148519138, + "learning_rate": 7.201485646633164e-06, + "loss": 0.6285, + "step": 2186 + }, + { + "epoch": 2.098895823331733, + "grad_norm": 0.8938875483158, + "learning_rate": 7.198477383884161e-06, + "loss": 0.4489, + "step": 2187 + }, + { + "epoch": 2.099855976956313, + "grad_norm": 1.1838577139933362, + "learning_rate": 7.195468134262159e-06, + "loss": 0.5581, + "step": 2188 + }, + { + "epoch": 2.100816130580893, + "grad_norm": 0.8025569757175272, + "learning_rate": 7.1924578991179815e-06, + "loss": 0.5391, + "step": 2189 + }, + { + "epoch": 2.101776284205473, + "grad_norm": 1.5927228117090462, + "learning_rate": 7.189446679802889e-06, + "loss": 0.5613, + "step": 2190 + }, + { + "epoch": 2.102736437830053, + "grad_norm": 1.0842887995533879, + "learning_rate": 7.186434477668586e-06, + "loss": 0.5342, + "step": 2191 + }, + { + "epoch": 2.1036965914546326, + "grad_norm": 1.3835446087699197, + "learning_rate": 7.1834212940672155e-06, + "loss": 0.6185, + "step": 2192 + }, + { + "epoch": 2.104656745079213, + "grad_norm": 1.1413603502415013, + "learning_rate": 7.180407130351368e-06, + "loss": 0.4431, + "step": 2193 + }, + { + "epoch": 2.1056168987037926, + "grad_norm": 1.0203641308253522, + "learning_rate": 7.177391987874065e-06, + "loss": 0.548, + "step": 2194 + }, + { + "epoch": 2.1065770523283724, + "grad_norm": 1.1363311509920815, + "learning_rate": 7.174375867988776e-06, + "loss": 0.5308, + "step": 2195 + }, + { + "epoch": 2.1075372059529527, + "grad_norm": 0.9567673445544885, + "learning_rate": 7.171358772049402e-06, + "loss": 0.5575, + "step": 2196 + }, + { + "epoch": 2.1084973595775325, + "grad_norm": 0.9923700778138494, + "learning_rate": 7.168340701410284e-06, + "loss": 0.6697, + "step": 2197 + }, + { + "epoch": 2.1094575132021123, + "grad_norm": 1.0364082103638248, + "learning_rate": 7.165321657426205e-06, + "loss": 0.5625, + "step": 2198 + }, + { + "epoch": 2.110417666826692, + "grad_norm": 0.9024784229244153, + "learning_rate": 7.1623016414523795e-06, + "loss": 0.6424, + "step": 2199 + }, + { + "epoch": 2.1113778204512723, + "grad_norm": 1.367149463303392, + "learning_rate": 7.159280654844462e-06, + "loss": 0.5617, + "step": 2200 + }, + { + "epoch": 2.112337974075852, + "grad_norm": 1.1617838082001823, + "learning_rate": 7.156258698958537e-06, + "loss": 0.554, + "step": 2201 + }, + { + "epoch": 2.113298127700432, + "grad_norm": 1.021449634098613, + "learning_rate": 7.153235775151134e-06, + "loss": 0.5342, + "step": 2202 + }, + { + "epoch": 2.114258281325012, + "grad_norm": 0.8308406626044873, + "learning_rate": 7.150211884779208e-06, + "loss": 0.5278, + "step": 2203 + }, + { + "epoch": 2.115218434949592, + "grad_norm": 0.9307394949034711, + "learning_rate": 7.147187029200151e-06, + "loss": 0.6127, + "step": 2204 + }, + { + "epoch": 2.1161785885741717, + "grad_norm": 1.8052285647276194, + "learning_rate": 7.144161209771788e-06, + "loss": 0.4938, + "step": 2205 + }, + { + "epoch": 2.117138742198752, + "grad_norm": 0.9771174668616823, + "learning_rate": 7.141134427852378e-06, + "loss": 0.5277, + "step": 2206 + }, + { + "epoch": 2.1180988958233318, + "grad_norm": 1.0292956633743262, + "learning_rate": 7.138106684800612e-06, + "loss": 0.5094, + "step": 2207 + }, + { + "epoch": 2.1190590494479116, + "grad_norm": 1.2451525712883664, + "learning_rate": 7.1350779819756085e-06, + "loss": 0.5448, + "step": 2208 + }, + { + "epoch": 2.120019203072492, + "grad_norm": 0.9661193881293042, + "learning_rate": 7.132048320736924e-06, + "loss": 0.5537, + "step": 2209 + }, + { + "epoch": 2.1209793566970716, + "grad_norm": 1.3789163777407878, + "learning_rate": 7.129017702444536e-06, + "loss": 0.611, + "step": 2210 + }, + { + "epoch": 2.1219395103216514, + "grad_norm": 1.5546841461614513, + "learning_rate": 7.12598612845886e-06, + "loss": 0.6078, + "step": 2211 + }, + { + "epoch": 2.122899663946231, + "grad_norm": 1.0926028952817535, + "learning_rate": 7.1229536001407376e-06, + "loss": 0.4314, + "step": 2212 + }, + { + "epoch": 2.1238598175708114, + "grad_norm": 1.0646968709677729, + "learning_rate": 7.119920118851438e-06, + "loss": 0.4824, + "step": 2213 + }, + { + "epoch": 2.1248199711953912, + "grad_norm": 1.3513451104794474, + "learning_rate": 7.116885685952655e-06, + "loss": 0.469, + "step": 2214 + }, + { + "epoch": 2.125780124819971, + "grad_norm": 0.9639076857555183, + "learning_rate": 7.113850302806518e-06, + "loss": 0.5212, + "step": 2215 + }, + { + "epoch": 2.1267402784445513, + "grad_norm": 1.0092878405418004, + "learning_rate": 7.110813970775574e-06, + "loss": 0.5365, + "step": 2216 + }, + { + "epoch": 2.127700432069131, + "grad_norm": 1.2257022771483324, + "learning_rate": 7.107776691222803e-06, + "loss": 0.4947, + "step": 2217 + }, + { + "epoch": 2.128660585693711, + "grad_norm": 1.0281895426486987, + "learning_rate": 7.104738465511606e-06, + "loss": 0.6361, + "step": 2218 + }, + { + "epoch": 2.129620739318291, + "grad_norm": 1.268479130840763, + "learning_rate": 7.1016992950058104e-06, + "loss": 0.5749, + "step": 2219 + }, + { + "epoch": 2.130580892942871, + "grad_norm": 1.0817101074711388, + "learning_rate": 7.098659181069667e-06, + "loss": 0.5998, + "step": 2220 + }, + { + "epoch": 2.1315410465674507, + "grad_norm": 0.9419880894216737, + "learning_rate": 7.0956181250678514e-06, + "loss": 0.5361, + "step": 2221 + }, + { + "epoch": 2.132501200192031, + "grad_norm": 1.1560205008626543, + "learning_rate": 7.09257612836546e-06, + "loss": 0.5673, + "step": 2222 + }, + { + "epoch": 2.1334613538166107, + "grad_norm": 1.2057235149267305, + "learning_rate": 7.089533192328011e-06, + "loss": 0.4456, + "step": 2223 + }, + { + "epoch": 2.1344215074411905, + "grad_norm": 1.2052989405818355, + "learning_rate": 7.08648931832145e-06, + "loss": 0.6141, + "step": 2224 + }, + { + "epoch": 2.1353816610657703, + "grad_norm": 1.0478638695612665, + "learning_rate": 7.0834445077121374e-06, + "loss": 0.6443, + "step": 2225 + }, + { + "epoch": 2.1363418146903506, + "grad_norm": 1.0239613188518975, + "learning_rate": 7.080398761866856e-06, + "loss": 0.5032, + "step": 2226 + }, + { + "epoch": 2.1373019683149304, + "grad_norm": 1.018760226884325, + "learning_rate": 7.077352082152809e-06, + "loss": 0.607, + "step": 2227 + }, + { + "epoch": 2.13826212193951, + "grad_norm": 1.148025927195372, + "learning_rate": 7.0743044699376165e-06, + "loss": 0.5498, + "step": 2228 + }, + { + "epoch": 2.1392222755640904, + "grad_norm": 0.8200781721139206, + "learning_rate": 7.0712559265893214e-06, + "loss": 0.5648, + "step": 2229 + }, + { + "epoch": 2.14018242918867, + "grad_norm": 1.1389520733011873, + "learning_rate": 7.068206453476381e-06, + "loss": 0.5307, + "step": 2230 + }, + { + "epoch": 2.14114258281325, + "grad_norm": 0.9912886676294732, + "learning_rate": 7.065156051967672e-06, + "loss": 0.4595, + "step": 2231 + }, + { + "epoch": 2.1421027364378302, + "grad_norm": 1.458899560172927, + "learning_rate": 7.062104723432485e-06, + "loss": 0.6631, + "step": 2232 + }, + { + "epoch": 2.14306289006241, + "grad_norm": 0.8813816859162074, + "learning_rate": 7.059052469240533e-06, + "loss": 0.4875, + "step": 2233 + }, + { + "epoch": 2.14402304368699, + "grad_norm": 0.9422304069243908, + "learning_rate": 7.0559992907619346e-06, + "loss": 0.4818, + "step": 2234 + }, + { + "epoch": 2.14498319731157, + "grad_norm": 1.0433863270739594, + "learning_rate": 7.0529451893672356e-06, + "loss": 0.4883, + "step": 2235 + }, + { + "epoch": 2.14594335093615, + "grad_norm": 1.083463354491641, + "learning_rate": 7.0498901664273825e-06, + "loss": 0.5466, + "step": 2236 + }, + { + "epoch": 2.1469035045607296, + "grad_norm": 0.9168061171424753, + "learning_rate": 7.046834223313746e-06, + "loss": 0.416, + "step": 2237 + }, + { + "epoch": 2.1478636581853094, + "grad_norm": 1.0983573939844062, + "learning_rate": 7.043777361398108e-06, + "loss": 0.5345, + "step": 2238 + }, + { + "epoch": 2.1488238118098897, + "grad_norm": 1.1026006301523767, + "learning_rate": 7.040719582052659e-06, + "loss": 0.5849, + "step": 2239 + }, + { + "epoch": 2.1497839654344695, + "grad_norm": 1.1229041477293433, + "learning_rate": 7.037660886650003e-06, + "loss": 0.608, + "step": 2240 + }, + { + "epoch": 2.1507441190590493, + "grad_norm": 1.0282597731998386, + "learning_rate": 7.034601276563155e-06, + "loss": 0.5216, + "step": 2241 + }, + { + "epoch": 2.1517042726836295, + "grad_norm": 1.2262513647641144, + "learning_rate": 7.031540753165545e-06, + "loss": 0.5375, + "step": 2242 + }, + { + "epoch": 2.1526644263082093, + "grad_norm": 1.3657286880174617, + "learning_rate": 7.028479317831006e-06, + "loss": 0.5714, + "step": 2243 + }, + { + "epoch": 2.153624579932789, + "grad_norm": 1.0830473500612943, + "learning_rate": 7.0254169719337864e-06, + "loss": 0.552, + "step": 2244 + }, + { + "epoch": 2.1545847335573693, + "grad_norm": 1.4461561056588648, + "learning_rate": 7.022353716848535e-06, + "loss": 0.4827, + "step": 2245 + }, + { + "epoch": 2.155544887181949, + "grad_norm": 1.057096879420475, + "learning_rate": 7.0192895539503205e-06, + "loss": 0.4658, + "step": 2246 + }, + { + "epoch": 2.156505040806529, + "grad_norm": 1.3215393080024684, + "learning_rate": 7.016224484614609e-06, + "loss": 0.6065, + "step": 2247 + }, + { + "epoch": 2.157465194431109, + "grad_norm": 1.2754535852244573, + "learning_rate": 7.01315851021728e-06, + "loss": 0.4785, + "step": 2248 + }, + { + "epoch": 2.158425348055689, + "grad_norm": 0.8719319849964379, + "learning_rate": 7.010091632134612e-06, + "loss": 0.5544, + "step": 2249 + }, + { + "epoch": 2.1593855016802688, + "grad_norm": 1.2593155434928438, + "learning_rate": 7.0070238517433e-06, + "loss": 0.5334, + "step": 2250 + }, + { + "epoch": 2.160345655304849, + "grad_norm": 1.1330654046107738, + "learning_rate": 7.003955170420434e-06, + "loss": 0.5208, + "step": 2251 + }, + { + "epoch": 2.161305808929429, + "grad_norm": 1.2484710985903849, + "learning_rate": 7.000885589543513e-06, + "loss": 0.5937, + "step": 2252 + }, + { + "epoch": 2.1622659625540086, + "grad_norm": 0.9836315272716349, + "learning_rate": 6.99781511049044e-06, + "loss": 0.5836, + "step": 2253 + }, + { + "epoch": 2.1632261161785884, + "grad_norm": 0.8452358278737337, + "learning_rate": 6.994743734639517e-06, + "loss": 0.5031, + "step": 2254 + }, + { + "epoch": 2.1641862698031686, + "grad_norm": 1.1668919313717263, + "learning_rate": 6.991671463369457e-06, + "loss": 0.6261, + "step": 2255 + }, + { + "epoch": 2.1651464234277484, + "grad_norm": 0.9413696122477315, + "learning_rate": 6.988598298059366e-06, + "loss": 0.531, + "step": 2256 + }, + { + "epoch": 2.1661065770523282, + "grad_norm": 0.9420458908509529, + "learning_rate": 6.985524240088756e-06, + "loss": 0.4476, + "step": 2257 + }, + { + "epoch": 2.1670667306769085, + "grad_norm": 1.1758684457243866, + "learning_rate": 6.982449290837538e-06, + "loss": 0.517, + "step": 2258 + }, + { + "epoch": 2.1680268843014883, + "grad_norm": 1.7800386777188943, + "learning_rate": 6.979373451686025e-06, + "loss": 0.6417, + "step": 2259 + }, + { + "epoch": 2.168987037926068, + "grad_norm": 1.4337750025305862, + "learning_rate": 6.976296724014927e-06, + "loss": 0.5546, + "step": 2260 + }, + { + "epoch": 2.1699471915506483, + "grad_norm": 1.3786218860218389, + "learning_rate": 6.973219109205356e-06, + "loss": 0.5388, + "step": 2261 + }, + { + "epoch": 2.170907345175228, + "grad_norm": 1.0618258194646955, + "learning_rate": 6.970140608638818e-06, + "loss": 0.5076, + "step": 2262 + }, + { + "epoch": 2.171867498799808, + "grad_norm": 1.7338047296555672, + "learning_rate": 6.967061223697223e-06, + "loss": 0.4912, + "step": 2263 + }, + { + "epoch": 2.172827652424388, + "grad_norm": 0.8916519090513646, + "learning_rate": 6.963980955762869e-06, + "loss": 0.5516, + "step": 2264 + }, + { + "epoch": 2.173787806048968, + "grad_norm": 0.9422127017638235, + "learning_rate": 6.960899806218458e-06, + "loss": 0.4112, + "step": 2265 + }, + { + "epoch": 2.1747479596735477, + "grad_norm": 1.2627279572143573, + "learning_rate": 6.957817776447087e-06, + "loss": 0.6057, + "step": 2266 + }, + { + "epoch": 2.1757081132981275, + "grad_norm": 0.9280773130916784, + "learning_rate": 6.954734867832241e-06, + "loss": 0.5171, + "step": 2267 + }, + { + "epoch": 2.1766682669227078, + "grad_norm": 0.9905577683589822, + "learning_rate": 6.9516510817578105e-06, + "loss": 0.5717, + "step": 2268 + }, + { + "epoch": 2.1776284205472876, + "grad_norm": 1.1751454090366942, + "learning_rate": 6.948566419608069e-06, + "loss": 0.5614, + "step": 2269 + }, + { + "epoch": 2.1785885741718674, + "grad_norm": 1.2412026855284037, + "learning_rate": 6.945480882767694e-06, + "loss": 0.6331, + "step": 2270 + }, + { + "epoch": 2.1795487277964476, + "grad_norm": 0.9185685217686413, + "learning_rate": 6.942394472621744e-06, + "loss": 0.5177, + "step": 2271 + }, + { + "epoch": 2.1805088814210274, + "grad_norm": 1.434092355643127, + "learning_rate": 6.939307190555679e-06, + "loss": 0.6101, + "step": 2272 + }, + { + "epoch": 2.181469035045607, + "grad_norm": 1.1546664904412804, + "learning_rate": 6.936219037955347e-06, + "loss": 0.5022, + "step": 2273 + }, + { + "epoch": 2.1824291886701874, + "grad_norm": 1.3881411890858537, + "learning_rate": 6.9331300162069846e-06, + "loss": 0.6449, + "step": 2274 + }, + { + "epoch": 2.183389342294767, + "grad_norm": 1.4231626563775166, + "learning_rate": 6.930040126697224e-06, + "loss": 0.5352, + "step": 2275 + }, + { + "epoch": 2.184349495919347, + "grad_norm": 0.9643207536001666, + "learning_rate": 6.926949370813079e-06, + "loss": 0.5814, + "step": 2276 + }, + { + "epoch": 2.1853096495439273, + "grad_norm": 0.8894548405312629, + "learning_rate": 6.92385774994196e-06, + "loss": 0.5097, + "step": 2277 + }, + { + "epoch": 2.186269803168507, + "grad_norm": 1.3618572655327714, + "learning_rate": 6.920765265471663e-06, + "loss": 0.5298, + "step": 2278 + }, + { + "epoch": 2.187229956793087, + "grad_norm": 1.077037347090665, + "learning_rate": 6.91767191879037e-06, + "loss": 0.4841, + "step": 2279 + }, + { + "epoch": 2.1881901104176666, + "grad_norm": 1.187289731607155, + "learning_rate": 6.9145777112866495e-06, + "loss": 0.5638, + "step": 2280 + }, + { + "epoch": 2.189150264042247, + "grad_norm": 1.3509930182156265, + "learning_rate": 6.911482644349462e-06, + "loss": 0.5707, + "step": 2281 + }, + { + "epoch": 2.1901104176668267, + "grad_norm": 0.9079890206437792, + "learning_rate": 6.908386719368147e-06, + "loss": 0.4809, + "step": 2282 + }, + { + "epoch": 2.1910705712914065, + "grad_norm": 1.3158411640553533, + "learning_rate": 6.9052899377324345e-06, + "loss": 0.552, + "step": 2283 + }, + { + "epoch": 2.1920307249159867, + "grad_norm": 1.1945920307842048, + "learning_rate": 6.902192300832435e-06, + "loss": 0.5206, + "step": 2284 + }, + { + "epoch": 2.1929908785405665, + "grad_norm": 1.1161698542709764, + "learning_rate": 6.899093810058643e-06, + "loss": 0.5349, + "step": 2285 + }, + { + "epoch": 2.1939510321651463, + "grad_norm": 1.0129646662385823, + "learning_rate": 6.895994466801943e-06, + "loss": 0.4909, + "step": 2286 + }, + { + "epoch": 2.1949111857897265, + "grad_norm": 0.785720546066982, + "learning_rate": 6.892894272453592e-06, + "loss": 0.5562, + "step": 2287 + }, + { + "epoch": 2.1958713394143063, + "grad_norm": 1.176481379742249, + "learning_rate": 6.889793228405237e-06, + "loss": 0.487, + "step": 2288 + }, + { + "epoch": 2.196831493038886, + "grad_norm": 1.2659944494299649, + "learning_rate": 6.8866913360489005e-06, + "loss": 0.5603, + "step": 2289 + }, + { + "epoch": 2.1977916466634664, + "grad_norm": 0.891824396482085, + "learning_rate": 6.883588596776994e-06, + "loss": 0.4518, + "step": 2290 + }, + { + "epoch": 2.198751800288046, + "grad_norm": 0.9718378046907263, + "learning_rate": 6.8804850119822986e-06, + "loss": 0.5012, + "step": 2291 + }, + { + "epoch": 2.199711953912626, + "grad_norm": 1.1515253226218582, + "learning_rate": 6.8773805830579845e-06, + "loss": 0.6426, + "step": 2292 + }, + { + "epoch": 2.2006721075372058, + "grad_norm": 1.0808812915933304, + "learning_rate": 6.8742753113975925e-06, + "loss": 0.4578, + "step": 2293 + }, + { + "epoch": 2.201632261161786, + "grad_norm": 1.1587103593406312, + "learning_rate": 6.871169198395049e-06, + "loss": 0.5647, + "step": 2294 + }, + { + "epoch": 2.202592414786366, + "grad_norm": 1.2830368134483428, + "learning_rate": 6.868062245444655e-06, + "loss": 0.5116, + "step": 2295 + }, + { + "epoch": 2.2035525684109456, + "grad_norm": 0.9888273438380815, + "learning_rate": 6.864954453941086e-06, + "loss": 0.5891, + "step": 2296 + }, + { + "epoch": 2.204512722035526, + "grad_norm": 1.0910465773902722, + "learning_rate": 6.861845825279399e-06, + "loss": 0.5684, + "step": 2297 + }, + { + "epoch": 2.2054728756601056, + "grad_norm": 0.9836998262220984, + "learning_rate": 6.85873636085502e-06, + "loss": 0.6049, + "step": 2298 + }, + { + "epoch": 2.2064330292846854, + "grad_norm": 1.143683722460904, + "learning_rate": 6.855626062063761e-06, + "loss": 0.5415, + "step": 2299 + }, + { + "epoch": 2.2073931829092657, + "grad_norm": 1.3627628991871574, + "learning_rate": 6.852514930301797e-06, + "loss": 0.4894, + "step": 2300 + }, + { + "epoch": 2.2083533365338455, + "grad_norm": 0.9646215360960887, + "learning_rate": 6.8494029669656845e-06, + "loss": 0.5391, + "step": 2301 + }, + { + "epoch": 2.2093134901584253, + "grad_norm": 1.6614106845749312, + "learning_rate": 6.846290173452348e-06, + "loss": 0.5892, + "step": 2302 + }, + { + "epoch": 2.2102736437830055, + "grad_norm": 0.7681870339356103, + "learning_rate": 6.843176551159087e-06, + "loss": 0.551, + "step": 2303 + }, + { + "epoch": 2.2112337974075853, + "grad_norm": 1.688322695234949, + "learning_rate": 6.8400621014835775e-06, + "loss": 0.6805, + "step": 2304 + }, + { + "epoch": 2.212193951032165, + "grad_norm": 1.0924474130289121, + "learning_rate": 6.836946825823858e-06, + "loss": 0.3479, + "step": 2305 + }, + { + "epoch": 2.213154104656745, + "grad_norm": 1.0545213214412013, + "learning_rate": 6.833830725578346e-06, + "loss": 0.5024, + "step": 2306 + }, + { + "epoch": 2.214114258281325, + "grad_norm": 0.9771886643473012, + "learning_rate": 6.83071380214582e-06, + "loss": 0.5355, + "step": 2307 + }, + { + "epoch": 2.215074411905905, + "grad_norm": 1.3549067705037805, + "learning_rate": 6.827596056925441e-06, + "loss": 0.641, + "step": 2308 + }, + { + "epoch": 2.2160345655304847, + "grad_norm": 0.9694903981259282, + "learning_rate": 6.824477491316725e-06, + "loss": 0.6004, + "step": 2309 + }, + { + "epoch": 2.216994719155065, + "grad_norm": 0.9874739474800195, + "learning_rate": 6.821358106719569e-06, + "loss": 0.5204, + "step": 2310 + }, + { + "epoch": 2.2179548727796448, + "grad_norm": 1.510713052530886, + "learning_rate": 6.818237904534226e-06, + "loss": 0.5127, + "step": 2311 + }, + { + "epoch": 2.2189150264042246, + "grad_norm": 1.2595518403245647, + "learning_rate": 6.815116886161324e-06, + "loss": 0.6615, + "step": 2312 + }, + { + "epoch": 2.219875180028805, + "grad_norm": 1.2551386659866655, + "learning_rate": 6.811995053001854e-06, + "loss": 0.5522, + "step": 2313 + }, + { + "epoch": 2.2208353336533846, + "grad_norm": 1.1456539086225281, + "learning_rate": 6.808872406457175e-06, + "loss": 0.4721, + "step": 2314 + }, + { + "epoch": 2.2217954872779644, + "grad_norm": 0.8045166359873522, + "learning_rate": 6.8057489479290085e-06, + "loss": 0.5141, + "step": 2315 + }, + { + "epoch": 2.2227556409025446, + "grad_norm": 1.3517950602434772, + "learning_rate": 6.802624678819441e-06, + "loss": 0.6937, + "step": 2316 + }, + { + "epoch": 2.2237157945271244, + "grad_norm": 1.1857114818615182, + "learning_rate": 6.799499600530925e-06, + "loss": 0.56, + "step": 2317 + }, + { + "epoch": 2.224675948151704, + "grad_norm": 1.2652105327532868, + "learning_rate": 6.796373714466272e-06, + "loss": 0.4745, + "step": 2318 + }, + { + "epoch": 2.225636101776284, + "grad_norm": 1.0637397227409495, + "learning_rate": 6.793247022028663e-06, + "loss": 0.5657, + "step": 2319 + }, + { + "epoch": 2.2265962554008643, + "grad_norm": 1.2460238834022297, + "learning_rate": 6.790119524621632e-06, + "loss": 0.4977, + "step": 2320 + }, + { + "epoch": 2.227556409025444, + "grad_norm": 1.2740604894255427, + "learning_rate": 6.786991223649084e-06, + "loss": 0.5146, + "step": 2321 + }, + { + "epoch": 2.228516562650024, + "grad_norm": 0.926237153900135, + "learning_rate": 6.783862120515277e-06, + "loss": 0.5681, + "step": 2322 + }, + { + "epoch": 2.229476716274604, + "grad_norm": 0.9988330432333267, + "learning_rate": 6.780732216624833e-06, + "loss": 0.3972, + "step": 2323 + }, + { + "epoch": 2.230436869899184, + "grad_norm": 1.095885002537761, + "learning_rate": 6.77760151338273e-06, + "loss": 0.6618, + "step": 2324 + }, + { + "epoch": 2.2313970235237637, + "grad_norm": 0.955222234964677, + "learning_rate": 6.7744700121943095e-06, + "loss": 0.548, + "step": 2325 + }, + { + "epoch": 2.232357177148344, + "grad_norm": 1.1476624773336968, + "learning_rate": 6.77133771446527e-06, + "loss": 0.5159, + "step": 2326 + }, + { + "epoch": 2.2333173307729237, + "grad_norm": 0.8123273760209445, + "learning_rate": 6.768204621601665e-06, + "loss": 0.528, + "step": 2327 + }, + { + "epoch": 2.2342774843975035, + "grad_norm": 1.6314501191901696, + "learning_rate": 6.765070735009906e-06, + "loss": 0.6033, + "step": 2328 + }, + { + "epoch": 2.2352376380220838, + "grad_norm": 1.5105126997235994, + "learning_rate": 6.761936056096761e-06, + "loss": 0.5298, + "step": 2329 + }, + { + "epoch": 2.2361977916466635, + "grad_norm": 1.193808886296744, + "learning_rate": 6.758800586269355e-06, + "loss": 0.5199, + "step": 2330 + }, + { + "epoch": 2.2371579452712433, + "grad_norm": 1.0487225146728896, + "learning_rate": 6.755664326935167e-06, + "loss": 0.6136, + "step": 2331 + }, + { + "epoch": 2.238118098895823, + "grad_norm": 1.362824827408439, + "learning_rate": 6.752527279502029e-06, + "loss": 0.6011, + "step": 2332 + }, + { + "epoch": 2.2390782525204034, + "grad_norm": 0.9389073331635173, + "learning_rate": 6.749389445378127e-06, + "loss": 0.5052, + "step": 2333 + }, + { + "epoch": 2.240038406144983, + "grad_norm": 1.0526327734300798, + "learning_rate": 6.746250825972005e-06, + "loss": 0.5006, + "step": 2334 + }, + { + "epoch": 2.240998559769563, + "grad_norm": 1.1827824553700668, + "learning_rate": 6.743111422692553e-06, + "loss": 0.6073, + "step": 2335 + }, + { + "epoch": 2.241958713394143, + "grad_norm": 1.51578979380233, + "learning_rate": 6.739971236949017e-06, + "loss": 0.5195, + "step": 2336 + }, + { + "epoch": 2.242918867018723, + "grad_norm": 1.125227461717384, + "learning_rate": 6.736830270150992e-06, + "loss": 0.4413, + "step": 2337 + }, + { + "epoch": 2.243879020643303, + "grad_norm": 1.042673510287957, + "learning_rate": 6.733688523708422e-06, + "loss": 0.6148, + "step": 2338 + }, + { + "epoch": 2.244839174267883, + "grad_norm": 1.4466183633336012, + "learning_rate": 6.7305459990316094e-06, + "loss": 0.6003, + "step": 2339 + }, + { + "epoch": 2.245799327892463, + "grad_norm": 1.0357649916251936, + "learning_rate": 6.727402697531193e-06, + "loss": 0.5981, + "step": 2340 + }, + { + "epoch": 2.2467594815170426, + "grad_norm": 0.9929925479732605, + "learning_rate": 6.7242586206181735e-06, + "loss": 0.6478, + "step": 2341 + }, + { + "epoch": 2.247719635141623, + "grad_norm": 0.8004720229019504, + "learning_rate": 6.721113769703887e-06, + "loss": 0.5089, + "step": 2342 + }, + { + "epoch": 2.2486797887662027, + "grad_norm": 1.0056538203059768, + "learning_rate": 6.71796814620003e-06, + "loss": 0.5418, + "step": 2343 + }, + { + "epoch": 2.2496399423907825, + "grad_norm": 1.1260843247375578, + "learning_rate": 6.714821751518635e-06, + "loss": 0.5194, + "step": 2344 + }, + { + "epoch": 2.2506000960153623, + "grad_norm": 0.9516273323340717, + "learning_rate": 6.7116745870720875e-06, + "loss": 0.5228, + "step": 2345 + }, + { + "epoch": 2.2515602496399425, + "grad_norm": 1.456476808458998, + "learning_rate": 6.708526654273113e-06, + "loss": 0.6069, + "step": 2346 + }, + { + "epoch": 2.2525204032645223, + "grad_norm": 1.3072688806574018, + "learning_rate": 6.7053779545347874e-06, + "loss": 0.5497, + "step": 2347 + }, + { + "epoch": 2.253480556889102, + "grad_norm": 1.048250496505981, + "learning_rate": 6.702228489270528e-06, + "loss": 0.4946, + "step": 2348 + }, + { + "epoch": 2.2544407105136823, + "grad_norm": 1.0496690320472093, + "learning_rate": 6.699078259894094e-06, + "loss": 0.5388, + "step": 2349 + }, + { + "epoch": 2.255400864138262, + "grad_norm": 0.9688487220489557, + "learning_rate": 6.695927267819593e-06, + "loss": 0.5984, + "step": 2350 + }, + { + "epoch": 2.256361017762842, + "grad_norm": 1.2185141354022164, + "learning_rate": 6.692775514461468e-06, + "loss": 0.5056, + "step": 2351 + }, + { + "epoch": 2.257321171387422, + "grad_norm": 1.0614292358503847, + "learning_rate": 6.689623001234511e-06, + "loss": 0.571, + "step": 2352 + }, + { + "epoch": 2.258281325012002, + "grad_norm": 1.1296621276111845, + "learning_rate": 6.686469729553849e-06, + "loss": 0.5724, + "step": 2353 + }, + { + "epoch": 2.2592414786365818, + "grad_norm": 0.8582511322406426, + "learning_rate": 6.683315700834953e-06, + "loss": 0.4999, + "step": 2354 + }, + { + "epoch": 2.260201632261162, + "grad_norm": 0.8613287945064796, + "learning_rate": 6.680160916493631e-06, + "loss": 0.4628, + "step": 2355 + }, + { + "epoch": 2.261161785885742, + "grad_norm": 1.483542890527988, + "learning_rate": 6.677005377946037e-06, + "loss": 0.4807, + "step": 2356 + }, + { + "epoch": 2.2621219395103216, + "grad_norm": 1.0556765585600885, + "learning_rate": 6.6738490866086515e-06, + "loss": 0.63, + "step": 2357 + }, + { + "epoch": 2.2630820931349014, + "grad_norm": 1.0456996598851618, + "learning_rate": 6.670692043898306e-06, + "loss": 0.5289, + "step": 2358 + }, + { + "epoch": 2.2640422467594816, + "grad_norm": 1.3551152692604027, + "learning_rate": 6.66753425123216e-06, + "loss": 0.5914, + "step": 2359 + }, + { + "epoch": 2.2650024003840614, + "grad_norm": 0.9504059413020357, + "learning_rate": 6.664375710027713e-06, + "loss": 0.5807, + "step": 2360 + }, + { + "epoch": 2.265962554008641, + "grad_norm": 0.9733497390231736, + "learning_rate": 6.661216421702801e-06, + "loss": 0.5759, + "step": 2361 + }, + { + "epoch": 2.2669227076332215, + "grad_norm": 1.239651371353343, + "learning_rate": 6.6580563876755945e-06, + "loss": 0.5163, + "step": 2362 + }, + { + "epoch": 2.2678828612578013, + "grad_norm": 0.9195396850869483, + "learning_rate": 6.6548956093646e-06, + "loss": 0.5418, + "step": 2363 + }, + { + "epoch": 2.268843014882381, + "grad_norm": 1.1034369651300087, + "learning_rate": 6.651734088188655e-06, + "loss": 0.5811, + "step": 2364 + }, + { + "epoch": 2.2698031685069613, + "grad_norm": 1.3877249971980943, + "learning_rate": 6.648571825566936e-06, + "loss": 0.5501, + "step": 2365 + }, + { + "epoch": 2.270763322131541, + "grad_norm": 1.1901368203660332, + "learning_rate": 6.645408822918945e-06, + "loss": 0.5802, + "step": 2366 + }, + { + "epoch": 2.271723475756121, + "grad_norm": 1.7551834561249287, + "learning_rate": 6.642245081664524e-06, + "loss": 0.6671, + "step": 2367 + }, + { + "epoch": 2.272683629380701, + "grad_norm": 1.2420794293486392, + "learning_rate": 6.6390806032238375e-06, + "loss": 0.5046, + "step": 2368 + }, + { + "epoch": 2.273643783005281, + "grad_norm": 0.9272166116224568, + "learning_rate": 6.63591538901739e-06, + "loss": 0.5252, + "step": 2369 + }, + { + "epoch": 2.2746039366298607, + "grad_norm": 1.0236145872195985, + "learning_rate": 6.632749440466013e-06, + "loss": 0.6363, + "step": 2370 + }, + { + "epoch": 2.2755640902544405, + "grad_norm": 0.9268058857786272, + "learning_rate": 6.629582758990864e-06, + "loss": 0.558, + "step": 2371 + }, + { + "epoch": 2.2765242438790207, + "grad_norm": 0.8661224125161839, + "learning_rate": 6.626415346013434e-06, + "loss": 0.5373, + "step": 2372 + }, + { + "epoch": 2.2774843975036005, + "grad_norm": 1.1503416267820565, + "learning_rate": 6.623247202955538e-06, + "loss": 0.6417, + "step": 2373 + }, + { + "epoch": 2.2784445511281803, + "grad_norm": 0.9631774310954262, + "learning_rate": 6.620078331239328e-06, + "loss": 0.6008, + "step": 2374 + }, + { + "epoch": 2.2794047047527606, + "grad_norm": 1.0572988947020006, + "learning_rate": 6.61690873228727e-06, + "loss": 0.544, + "step": 2375 + }, + { + "epoch": 2.2803648583773404, + "grad_norm": 1.6109399609121728, + "learning_rate": 6.613738407522169e-06, + "loss": 0.6017, + "step": 2376 + }, + { + "epoch": 2.28132501200192, + "grad_norm": 0.9386008298486445, + "learning_rate": 6.610567358367143e-06, + "loss": 0.5316, + "step": 2377 + }, + { + "epoch": 2.2822851656265004, + "grad_norm": 1.2291257971075966, + "learning_rate": 6.607395586245648e-06, + "loss": 0.5553, + "step": 2378 + }, + { + "epoch": 2.28324531925108, + "grad_norm": 0.9880000748680476, + "learning_rate": 6.6042230925814565e-06, + "loss": 0.5706, + "step": 2379 + }, + { + "epoch": 2.28420547287566, + "grad_norm": 1.3882384375731867, + "learning_rate": 6.601049878798667e-06, + "loss": 0.6662, + "step": 2380 + }, + { + "epoch": 2.2851656265002402, + "grad_norm": 1.0024878661703003, + "learning_rate": 6.597875946321701e-06, + "loss": 0.5747, + "step": 2381 + }, + { + "epoch": 2.28612578012482, + "grad_norm": 0.9871022506826898, + "learning_rate": 6.594701296575303e-06, + "loss": 0.5533, + "step": 2382 + }, + { + "epoch": 2.2870859337494, + "grad_norm": 1.0405673527467023, + "learning_rate": 6.5915259309845404e-06, + "loss": 0.5858, + "step": 2383 + }, + { + "epoch": 2.2880460873739796, + "grad_norm": 1.1935101241961101, + "learning_rate": 6.588349850974798e-06, + "loss": 0.4727, + "step": 2384 + }, + { + "epoch": 2.28900624099856, + "grad_norm": 1.6937651826408868, + "learning_rate": 6.585173057971787e-06, + "loss": 0.6207, + "step": 2385 + }, + { + "epoch": 2.2899663946231397, + "grad_norm": 0.9273325222250683, + "learning_rate": 6.5819955534015335e-06, + "loss": 0.5046, + "step": 2386 + }, + { + "epoch": 2.2909265482477195, + "grad_norm": 1.0349028239541116, + "learning_rate": 6.578817338690389e-06, + "loss": 0.5695, + "step": 2387 + }, + { + "epoch": 2.2918867018722997, + "grad_norm": 1.0067078089408212, + "learning_rate": 6.575638415265016e-06, + "loss": 0.5692, + "step": 2388 + }, + { + "epoch": 2.2928468554968795, + "grad_norm": 1.4122462213753166, + "learning_rate": 6.5724587845524024e-06, + "loss": 0.648, + "step": 2389 + }, + { + "epoch": 2.2938070091214593, + "grad_norm": 0.9882915938991353, + "learning_rate": 6.569278447979848e-06, + "loss": 0.6179, + "step": 2390 + }, + { + "epoch": 2.2947671627460395, + "grad_norm": 1.3776070344364486, + "learning_rate": 6.566097406974972e-06, + "loss": 0.5681, + "step": 2391 + }, + { + "epoch": 2.2957273163706193, + "grad_norm": 1.0032261079729023, + "learning_rate": 6.562915662965713e-06, + "loss": 0.567, + "step": 2392 + }, + { + "epoch": 2.296687469995199, + "grad_norm": 0.837628254593809, + "learning_rate": 6.5597332173803175e-06, + "loss": 0.4813, + "step": 2393 + }, + { + "epoch": 2.2976476236197794, + "grad_norm": 0.9072512534340048, + "learning_rate": 6.556550071647355e-06, + "loss": 0.5525, + "step": 2394 + }, + { + "epoch": 2.298607777244359, + "grad_norm": 1.1699080465870773, + "learning_rate": 6.553366227195702e-06, + "loss": 0.5854, + "step": 2395 + }, + { + "epoch": 2.299567930868939, + "grad_norm": 0.9777031836326662, + "learning_rate": 6.550181685454556e-06, + "loss": 0.5771, + "step": 2396 + }, + { + "epoch": 2.3005280844935188, + "grad_norm": 0.8933149048160616, + "learning_rate": 6.546996447853421e-06, + "loss": 0.5101, + "step": 2397 + }, + { + "epoch": 2.301488238118099, + "grad_norm": 1.3469317902735265, + "learning_rate": 6.5438105158221175e-06, + "loss": 0.5216, + "step": 2398 + }, + { + "epoch": 2.302448391742679, + "grad_norm": 0.9440025069528666, + "learning_rate": 6.5406238907907745e-06, + "loss": 0.3806, + "step": 2399 + }, + { + "epoch": 2.3034085453672586, + "grad_norm": 1.1338584353480947, + "learning_rate": 6.537436574189837e-06, + "loss": 0.496, + "step": 2400 + }, + { + "epoch": 2.304368698991839, + "grad_norm": 0.8664060336273369, + "learning_rate": 6.5342485674500554e-06, + "loss": 0.5588, + "step": 2401 + }, + { + "epoch": 2.3053288526164186, + "grad_norm": 0.844386738936581, + "learning_rate": 6.531059872002492e-06, + "loss": 0.5238, + "step": 2402 + }, + { + "epoch": 2.3062890062409984, + "grad_norm": 0.97008817337292, + "learning_rate": 6.527870489278519e-06, + "loss": 0.4965, + "step": 2403 + }, + { + "epoch": 2.3072491598655787, + "grad_norm": 1.2113093411636833, + "learning_rate": 6.524680420709813e-06, + "loss": 0.5232, + "step": 2404 + }, + { + "epoch": 2.3082093134901585, + "grad_norm": 0.7933762415126152, + "learning_rate": 6.521489667728365e-06, + "loss": 0.5103, + "step": 2405 + }, + { + "epoch": 2.3091694671147383, + "grad_norm": 1.6040011514959773, + "learning_rate": 6.51829823176647e-06, + "loss": 0.4945, + "step": 2406 + }, + { + "epoch": 2.3101296207393185, + "grad_norm": 1.1872680952204704, + "learning_rate": 6.5151061142567285e-06, + "loss": 0.5818, + "step": 2407 + }, + { + "epoch": 2.3110897743638983, + "grad_norm": 1.1867275559034063, + "learning_rate": 6.511913316632046e-06, + "loss": 0.4382, + "step": 2408 + }, + { + "epoch": 2.312049927988478, + "grad_norm": 1.0241798518777094, + "learning_rate": 6.508719840325639e-06, + "loss": 0.5529, + "step": 2409 + }, + { + "epoch": 2.313010081613058, + "grad_norm": 0.932597269090835, + "learning_rate": 6.505525686771021e-06, + "loss": 0.4931, + "step": 2410 + }, + { + "epoch": 2.313970235237638, + "grad_norm": 2.0244549254616926, + "learning_rate": 6.502330857402018e-06, + "loss": 0.4747, + "step": 2411 + }, + { + "epoch": 2.314930388862218, + "grad_norm": 0.9489041896092459, + "learning_rate": 6.49913535365275e-06, + "loss": 0.5403, + "step": 2412 + }, + { + "epoch": 2.3158905424867977, + "grad_norm": 0.9132995165306347, + "learning_rate": 6.495939176957648e-06, + "loss": 0.524, + "step": 2413 + }, + { + "epoch": 2.316850696111378, + "grad_norm": 1.1110988605125405, + "learning_rate": 6.49274232875144e-06, + "loss": 0.4239, + "step": 2414 + }, + { + "epoch": 2.3178108497359577, + "grad_norm": 1.1042601415143272, + "learning_rate": 6.489544810469157e-06, + "loss": 0.4548, + "step": 2415 + }, + { + "epoch": 2.3187710033605375, + "grad_norm": 0.7788382029560041, + "learning_rate": 6.4863466235461316e-06, + "loss": 0.5014, + "step": 2416 + }, + { + "epoch": 2.319731156985118, + "grad_norm": 0.8844795604785618, + "learning_rate": 6.483147769417992e-06, + "loss": 0.5911, + "step": 2417 + }, + { + "epoch": 2.3206913106096976, + "grad_norm": 1.155745030904022, + "learning_rate": 6.479948249520676e-06, + "loss": 0.5965, + "step": 2418 + }, + { + "epoch": 2.3216514642342774, + "grad_norm": 0.8708622921604449, + "learning_rate": 6.476748065290409e-06, + "loss": 0.5494, + "step": 2419 + }, + { + "epoch": 2.3226116178588576, + "grad_norm": 1.0298916457963645, + "learning_rate": 6.473547218163723e-06, + "loss": 0.5817, + "step": 2420 + }, + { + "epoch": 2.3235717714834374, + "grad_norm": 0.9776650358768887, + "learning_rate": 6.4703457095774395e-06, + "loss": 0.5358, + "step": 2421 + }, + { + "epoch": 2.324531925108017, + "grad_norm": 1.080675144348416, + "learning_rate": 6.467143540968684e-06, + "loss": 0.4904, + "step": 2422 + }, + { + "epoch": 2.325492078732597, + "grad_norm": 1.230767202332983, + "learning_rate": 6.4639407137748774e-06, + "loss": 0.5198, + "step": 2423 + }, + { + "epoch": 2.3264522323571772, + "grad_norm": 1.030541708041631, + "learning_rate": 6.4607372294337335e-06, + "loss": 0.4811, + "step": 2424 + }, + { + "epoch": 2.327412385981757, + "grad_norm": 1.0798731215135828, + "learning_rate": 6.457533089383261e-06, + "loss": 0.6193, + "step": 2425 + }, + { + "epoch": 2.328372539606337, + "grad_norm": 1.1990530071734347, + "learning_rate": 6.454328295061764e-06, + "loss": 0.5993, + "step": 2426 + }, + { + "epoch": 2.329332693230917, + "grad_norm": 1.0319319277157921, + "learning_rate": 6.451122847907843e-06, + "loss": 0.5686, + "step": 2427 + }, + { + "epoch": 2.330292846855497, + "grad_norm": 0.929150769455459, + "learning_rate": 6.447916749360386e-06, + "loss": 0.5257, + "step": 2428 + }, + { + "epoch": 2.3312530004800767, + "grad_norm": 0.8378201535255333, + "learning_rate": 6.444710000858578e-06, + "loss": 0.6194, + "step": 2429 + }, + { + "epoch": 2.332213154104657, + "grad_norm": 0.9679996392554938, + "learning_rate": 6.441502603841892e-06, + "loss": 0.5243, + "step": 2430 + }, + { + "epoch": 2.3331733077292367, + "grad_norm": 1.132744381814498, + "learning_rate": 6.438294559750098e-06, + "loss": 0.5697, + "step": 2431 + }, + { + "epoch": 2.3341334613538165, + "grad_norm": 0.9314276331579016, + "learning_rate": 6.43508587002325e-06, + "loss": 0.4477, + "step": 2432 + }, + { + "epoch": 2.3350936149783967, + "grad_norm": 1.5226059113858328, + "learning_rate": 6.431876536101695e-06, + "loss": 0.6278, + "step": 2433 + }, + { + "epoch": 2.3360537686029765, + "grad_norm": 1.2881405764305072, + "learning_rate": 6.42866655942607e-06, + "loss": 0.5379, + "step": 2434 + }, + { + "epoch": 2.3370139222275563, + "grad_norm": 1.12727054271157, + "learning_rate": 6.425455941437297e-06, + "loss": 0.5785, + "step": 2435 + }, + { + "epoch": 2.337974075852136, + "grad_norm": 1.723016354144707, + "learning_rate": 6.422244683576589e-06, + "loss": 0.4627, + "step": 2436 + }, + { + "epoch": 2.3389342294767164, + "grad_norm": 1.142417894256119, + "learning_rate": 6.419032787285446e-06, + "loss": 0.5908, + "step": 2437 + }, + { + "epoch": 2.339894383101296, + "grad_norm": 1.8797402056376855, + "learning_rate": 6.4158202540056536e-06, + "loss": 0.6451, + "step": 2438 + }, + { + "epoch": 2.340854536725876, + "grad_norm": 1.2870211657494477, + "learning_rate": 6.412607085179282e-06, + "loss": 0.5784, + "step": 2439 + }, + { + "epoch": 2.341814690350456, + "grad_norm": 1.159389786483229, + "learning_rate": 6.409393282248692e-06, + "loss": 0.5527, + "step": 2440 + }, + { + "epoch": 2.342774843975036, + "grad_norm": 1.0038200173284817, + "learning_rate": 6.406178846656521e-06, + "loss": 0.4619, + "step": 2441 + }, + { + "epoch": 2.343734997599616, + "grad_norm": 1.2176155016950183, + "learning_rate": 6.402963779845699e-06, + "loss": 0.5757, + "step": 2442 + }, + { + "epoch": 2.344695151224196, + "grad_norm": 1.442166780259645, + "learning_rate": 6.399748083259432e-06, + "loss": 0.6336, + "step": 2443 + }, + { + "epoch": 2.345655304848776, + "grad_norm": 0.9343353686932061, + "learning_rate": 6.3965317583412135e-06, + "loss": 0.4759, + "step": 2444 + }, + { + "epoch": 2.3466154584733556, + "grad_norm": 1.2702910517169863, + "learning_rate": 6.393314806534815e-06, + "loss": 0.5686, + "step": 2445 + }, + { + "epoch": 2.347575612097936, + "grad_norm": 1.2536723128102305, + "learning_rate": 6.390097229284296e-06, + "loss": 0.4995, + "step": 2446 + }, + { + "epoch": 2.3485357657225157, + "grad_norm": 1.0470844232843697, + "learning_rate": 6.386879028033988e-06, + "loss": 0.5053, + "step": 2447 + }, + { + "epoch": 2.3494959193470955, + "grad_norm": 1.1722093956745143, + "learning_rate": 6.3836602042285105e-06, + "loss": 0.4668, + "step": 2448 + }, + { + "epoch": 2.3504560729716752, + "grad_norm": 0.944788504592577, + "learning_rate": 6.3804407593127585e-06, + "loss": 0.5594, + "step": 2449 + }, + { + "epoch": 2.3514162265962555, + "grad_norm": 1.090347727493859, + "learning_rate": 6.377220694731904e-06, + "loss": 0.5469, + "step": 2450 + }, + { + "epoch": 2.3523763802208353, + "grad_norm": 1.1260538228003005, + "learning_rate": 6.374000011931403e-06, + "loss": 0.5433, + "step": 2451 + }, + { + "epoch": 2.353336533845415, + "grad_norm": 1.0542951287411089, + "learning_rate": 6.370778712356981e-06, + "loss": 0.4382, + "step": 2452 + }, + { + "epoch": 2.3542966874699953, + "grad_norm": 0.9019927253450573, + "learning_rate": 6.3675567974546506e-06, + "loss": 0.546, + "step": 2453 + }, + { + "epoch": 2.355256841094575, + "grad_norm": 1.4738986979211055, + "learning_rate": 6.364334268670691e-06, + "loss": 0.6841, + "step": 2454 + }, + { + "epoch": 2.356216994719155, + "grad_norm": 1.3022796833262742, + "learning_rate": 6.361111127451664e-06, + "loss": 0.5596, + "step": 2455 + }, + { + "epoch": 2.357177148343735, + "grad_norm": 1.0840341756830607, + "learning_rate": 6.357887375244398e-06, + "loss": 0.4968, + "step": 2456 + }, + { + "epoch": 2.358137301968315, + "grad_norm": 1.2951952187839153, + "learning_rate": 6.354663013496006e-06, + "loss": 0.4638, + "step": 2457 + }, + { + "epoch": 2.3590974555928947, + "grad_norm": 1.1111086015730918, + "learning_rate": 6.3514380436538695e-06, + "loss": 0.5432, + "step": 2458 + }, + { + "epoch": 2.360057609217475, + "grad_norm": 1.188571308975247, + "learning_rate": 6.348212467165639e-06, + "loss": 0.4349, + "step": 2459 + }, + { + "epoch": 2.361017762842055, + "grad_norm": 0.9545985278672484, + "learning_rate": 6.344986285479244e-06, + "loss": 0.5683, + "step": 2460 + }, + { + "epoch": 2.3619779164666346, + "grad_norm": 1.6278258187837187, + "learning_rate": 6.341759500042882e-06, + "loss": 0.6329, + "step": 2461 + }, + { + "epoch": 2.3629380700912144, + "grad_norm": 1.022302684447469, + "learning_rate": 6.338532112305025e-06, + "loss": 0.4961, + "step": 2462 + }, + { + "epoch": 2.3638982237157946, + "grad_norm": 1.107779000074782, + "learning_rate": 6.3353041237144096e-06, + "loss": 0.6454, + "step": 2463 + }, + { + "epoch": 2.3648583773403744, + "grad_norm": 1.1517936843117076, + "learning_rate": 6.332075535720047e-06, + "loss": 0.5464, + "step": 2464 + }, + { + "epoch": 2.365818530964954, + "grad_norm": 1.0677799092338685, + "learning_rate": 6.328846349771215e-06, + "loss": 0.4648, + "step": 2465 + }, + { + "epoch": 2.3667786845895344, + "grad_norm": 1.6285500755462932, + "learning_rate": 6.325616567317461e-06, + "loss": 0.5713, + "step": 2466 + }, + { + "epoch": 2.3677388382141142, + "grad_norm": 1.136262753647289, + "learning_rate": 6.3223861898086e-06, + "loss": 0.6101, + "step": 2467 + }, + { + "epoch": 2.368698991838694, + "grad_norm": 0.90489576314916, + "learning_rate": 6.319155218694716e-06, + "loss": 0.5386, + "step": 2468 + }, + { + "epoch": 2.3696591454632743, + "grad_norm": 1.878512432031467, + "learning_rate": 6.315923655426152e-06, + "loss": 0.6031, + "step": 2469 + }, + { + "epoch": 2.370619299087854, + "grad_norm": 1.1158148731899769, + "learning_rate": 6.312691501453527e-06, + "loss": 0.462, + "step": 2470 + }, + { + "epoch": 2.371579452712434, + "grad_norm": 0.9969182900346529, + "learning_rate": 6.30945875822772e-06, + "loss": 0.6176, + "step": 2471 + }, + { + "epoch": 2.372539606337014, + "grad_norm": 0.8883371107710113, + "learning_rate": 6.306225427199873e-06, + "loss": 0.5047, + "step": 2472 + }, + { + "epoch": 2.373499759961594, + "grad_norm": 1.011686097060432, + "learning_rate": 6.302991509821396e-06, + "loss": 0.5202, + "step": 2473 + }, + { + "epoch": 2.3744599135861737, + "grad_norm": 0.9917933465194929, + "learning_rate": 6.299757007543954e-06, + "loss": 0.5423, + "step": 2474 + }, + { + "epoch": 2.3754200672107535, + "grad_norm": 0.867445761455369, + "learning_rate": 6.296521921819489e-06, + "loss": 0.4968, + "step": 2475 + }, + { + "epoch": 2.3763802208353337, + "grad_norm": 1.226080037659591, + "learning_rate": 6.29328625410019e-06, + "loss": 0.6016, + "step": 2476 + }, + { + "epoch": 2.3773403744599135, + "grad_norm": 0.8714179357262549, + "learning_rate": 6.290050005838519e-06, + "loss": 0.5684, + "step": 2477 + }, + { + "epoch": 2.3783005280844933, + "grad_norm": 0.9008505419061984, + "learning_rate": 6.286813178487188e-06, + "loss": 0.6102, + "step": 2478 + }, + { + "epoch": 2.3792606817090736, + "grad_norm": 1.1543272859182996, + "learning_rate": 6.283575773499176e-06, + "loss": 0.4259, + "step": 2479 + }, + { + "epoch": 2.3802208353336534, + "grad_norm": 1.250841848975828, + "learning_rate": 6.280337792327722e-06, + "loss": 0.5412, + "step": 2480 + }, + { + "epoch": 2.381180988958233, + "grad_norm": 1.1754961010375546, + "learning_rate": 6.277099236426317e-06, + "loss": 0.4733, + "step": 2481 + }, + { + "epoch": 2.3821411425828134, + "grad_norm": 1.5148238383477088, + "learning_rate": 6.273860107248718e-06, + "loss": 0.5905, + "step": 2482 + }, + { + "epoch": 2.383101296207393, + "grad_norm": 1.1762920142876214, + "learning_rate": 6.270620406248931e-06, + "loss": 0.5957, + "step": 2483 + }, + { + "epoch": 2.384061449831973, + "grad_norm": 1.049323319918193, + "learning_rate": 6.267380134881229e-06, + "loss": 0.603, + "step": 2484 + }, + { + "epoch": 2.3850216034565532, + "grad_norm": 0.9600595231048848, + "learning_rate": 6.26413929460013e-06, + "loss": 0.5481, + "step": 2485 + }, + { + "epoch": 2.385981757081133, + "grad_norm": 0.8949485585188345, + "learning_rate": 6.260897886860415e-06, + "loss": 0.5511, + "step": 2486 + }, + { + "epoch": 2.386941910705713, + "grad_norm": 0.8341641848405127, + "learning_rate": 6.2576559131171174e-06, + "loss": 0.561, + "step": 2487 + }, + { + "epoch": 2.3879020643302926, + "grad_norm": 0.8356952968329642, + "learning_rate": 6.254413374825523e-06, + "loss": 0.512, + "step": 2488 + }, + { + "epoch": 2.388862217954873, + "grad_norm": 1.2912356680160988, + "learning_rate": 6.251170273441174e-06, + "loss": 0.501, + "step": 2489 + }, + { + "epoch": 2.3898223715794527, + "grad_norm": 1.3687753415351693, + "learning_rate": 6.247926610419865e-06, + "loss": 0.5907, + "step": 2490 + }, + { + "epoch": 2.390782525204033, + "grad_norm": 1.741973792817936, + "learning_rate": 6.244682387217638e-06, + "loss": 0.5909, + "step": 2491 + }, + { + "epoch": 2.3917426788286127, + "grad_norm": 1.733414899449277, + "learning_rate": 6.241437605290792e-06, + "loss": 0.638, + "step": 2492 + }, + { + "epoch": 2.3927028324531925, + "grad_norm": 1.098017920644056, + "learning_rate": 6.238192266095877e-06, + "loss": 0.5966, + "step": 2493 + }, + { + "epoch": 2.3936629860777723, + "grad_norm": 1.858505879300705, + "learning_rate": 6.234946371089688e-06, + "loss": 0.5663, + "step": 2494 + }, + { + "epoch": 2.3946231397023525, + "grad_norm": 0.9586710525031326, + "learning_rate": 6.231699921729275e-06, + "loss": 0.5443, + "step": 2495 + }, + { + "epoch": 2.3955832933269323, + "grad_norm": 1.2256198791441315, + "learning_rate": 6.228452919471933e-06, + "loss": 0.62, + "step": 2496 + }, + { + "epoch": 2.396543446951512, + "grad_norm": 0.9852276786729444, + "learning_rate": 6.225205365775206e-06, + "loss": 0.4704, + "step": 2497 + }, + { + "epoch": 2.3975036005760924, + "grad_norm": 0.8635032986917627, + "learning_rate": 6.221957262096887e-06, + "loss": 0.5254, + "step": 2498 + }, + { + "epoch": 2.398463754200672, + "grad_norm": 1.1244557413706344, + "learning_rate": 6.218708609895017e-06, + "loss": 0.6046, + "step": 2499 + }, + { + "epoch": 2.399423907825252, + "grad_norm": 1.146043243064041, + "learning_rate": 6.2154594106278775e-06, + "loss": 0.5711, + "step": 2500 + }, + { + "epoch": 2.4003840614498317, + "grad_norm": 0.9653797913238615, + "learning_rate": 6.212209665754001e-06, + "loss": 0.54, + "step": 2501 + }, + { + "epoch": 2.401344215074412, + "grad_norm": 0.9576143723527571, + "learning_rate": 6.2089593767321645e-06, + "loss": 0.4741, + "step": 2502 + }, + { + "epoch": 2.4023043686989918, + "grad_norm": 1.1649218757253375, + "learning_rate": 6.2057085450213864e-06, + "loss": 0.5723, + "step": 2503 + }, + { + "epoch": 2.403264522323572, + "grad_norm": 0.9175682966509715, + "learning_rate": 6.202457172080931e-06, + "loss": 0.6017, + "step": 2504 + }, + { + "epoch": 2.404224675948152, + "grad_norm": 1.2241928486652216, + "learning_rate": 6.199205259370304e-06, + "loss": 0.5167, + "step": 2505 + }, + { + "epoch": 2.4051848295727316, + "grad_norm": 0.93584799869271, + "learning_rate": 6.195952808349255e-06, + "loss": 0.4923, + "step": 2506 + }, + { + "epoch": 2.4061449831973114, + "grad_norm": 0.7834795850296479, + "learning_rate": 6.192699820477775e-06, + "loss": 0.4509, + "step": 2507 + }, + { + "epoch": 2.4071051368218916, + "grad_norm": 1.9248511223100664, + "learning_rate": 6.189446297216094e-06, + "loss": 0.7094, + "step": 2508 + }, + { + "epoch": 2.4080652904464714, + "grad_norm": 1.055263058174764, + "learning_rate": 6.186192240024684e-06, + "loss": 0.5797, + "step": 2509 + }, + { + "epoch": 2.4090254440710512, + "grad_norm": 1.451596012988708, + "learning_rate": 6.182937650364258e-06, + "loss": 0.5341, + "step": 2510 + }, + { + "epoch": 2.4099855976956315, + "grad_norm": 1.0779510051470163, + "learning_rate": 6.179682529695766e-06, + "loss": 0.5495, + "step": 2511 + }, + { + "epoch": 2.4109457513202113, + "grad_norm": 1.0889028245993762, + "learning_rate": 6.176426879480396e-06, + "loss": 0.5927, + "step": 2512 + }, + { + "epoch": 2.411905904944791, + "grad_norm": 1.4630636365034198, + "learning_rate": 6.173170701179576e-06, + "loss": 0.6391, + "step": 2513 + }, + { + "epoch": 2.412866058569371, + "grad_norm": 1.110996876012569, + "learning_rate": 6.169913996254967e-06, + "loss": 0.4694, + "step": 2514 + }, + { + "epoch": 2.413826212193951, + "grad_norm": 1.3124809791345493, + "learning_rate": 6.166656766168473e-06, + "loss": 0.5996, + "step": 2515 + }, + { + "epoch": 2.414786365818531, + "grad_norm": 0.9981468821518614, + "learning_rate": 6.1633990123822265e-06, + "loss": 0.4888, + "step": 2516 + }, + { + "epoch": 2.415746519443111, + "grad_norm": 0.8868689938749357, + "learning_rate": 6.1601407363586e-06, + "loss": 0.4886, + "step": 2517 + }, + { + "epoch": 2.416706673067691, + "grad_norm": 0.9276247128822482, + "learning_rate": 6.156881939560196e-06, + "loss": 0.479, + "step": 2518 + }, + { + "epoch": 2.4176668266922707, + "grad_norm": 1.4470991332967298, + "learning_rate": 6.153622623449858e-06, + "loss": 0.5096, + "step": 2519 + }, + { + "epoch": 2.4186269803168505, + "grad_norm": 1.3237570866236188, + "learning_rate": 6.150362789490654e-06, + "loss": 0.473, + "step": 2520 + }, + { + "epoch": 2.4195871339414308, + "grad_norm": 1.7682334211933788, + "learning_rate": 6.147102439145892e-06, + "loss": 0.6328, + "step": 2521 + }, + { + "epoch": 2.4205472875660106, + "grad_norm": 0.7815698215631511, + "learning_rate": 6.143841573879107e-06, + "loss": 0.4841, + "step": 2522 + }, + { + "epoch": 2.4215074411905904, + "grad_norm": 1.19672195307083, + "learning_rate": 6.140580195154065e-06, + "loss": 0.5615, + "step": 2523 + }, + { + "epoch": 2.4224675948151706, + "grad_norm": 0.9681133678481987, + "learning_rate": 6.137318304434768e-06, + "loss": 0.5036, + "step": 2524 + }, + { + "epoch": 2.4234277484397504, + "grad_norm": 1.0484371141077236, + "learning_rate": 6.1340559031854385e-06, + "loss": 0.6357, + "step": 2525 + }, + { + "epoch": 2.42438790206433, + "grad_norm": 0.9071284605464971, + "learning_rate": 6.130792992870539e-06, + "loss": 0.6447, + "step": 2526 + }, + { + "epoch": 2.42534805568891, + "grad_norm": 1.4903328940675329, + "learning_rate": 6.127529574954749e-06, + "loss": 0.5664, + "step": 2527 + }, + { + "epoch": 2.4263082093134902, + "grad_norm": 0.7112508735080384, + "learning_rate": 6.124265650902989e-06, + "loss": 0.3958, + "step": 2528 + }, + { + "epoch": 2.42726836293807, + "grad_norm": 0.876598457400479, + "learning_rate": 6.121001222180394e-06, + "loss": 0.5359, + "step": 2529 + }, + { + "epoch": 2.4282285165626503, + "grad_norm": 1.838248783503213, + "learning_rate": 6.117736290252335e-06, + "loss": 0.4881, + "step": 2530 + }, + { + "epoch": 2.42918867018723, + "grad_norm": 0.8244178851212894, + "learning_rate": 6.1144708565844026e-06, + "loss": 0.5553, + "step": 2531 + }, + { + "epoch": 2.43014882381181, + "grad_norm": 1.370833039667388, + "learning_rate": 6.111204922642414e-06, + "loss": 0.5652, + "step": 2532 + }, + { + "epoch": 2.4311089774363897, + "grad_norm": 1.0224639269077314, + "learning_rate": 6.107938489892417e-06, + "loss": 0.6292, + "step": 2533 + }, + { + "epoch": 2.43206913106097, + "grad_norm": 1.4056638544422844, + "learning_rate": 6.104671559800672e-06, + "loss": 0.6494, + "step": 2534 + }, + { + "epoch": 2.4330292846855497, + "grad_norm": 1.279030589080096, + "learning_rate": 6.101404133833674e-06, + "loss": 0.5567, + "step": 2535 + }, + { + "epoch": 2.4339894383101295, + "grad_norm": 0.965369013712244, + "learning_rate": 6.098136213458131e-06, + "loss": 0.5479, + "step": 2536 + }, + { + "epoch": 2.4349495919347097, + "grad_norm": 0.811936788515548, + "learning_rate": 6.0948678001409835e-06, + "loss": 0.4838, + "step": 2537 + }, + { + "epoch": 2.4359097455592895, + "grad_norm": 1.4508632995724344, + "learning_rate": 6.091598895349382e-06, + "loss": 0.6178, + "step": 2538 + }, + { + "epoch": 2.4368698991838693, + "grad_norm": 1.315776079609651, + "learning_rate": 6.088329500550706e-06, + "loss": 0.551, + "step": 2539 + }, + { + "epoch": 2.437830052808449, + "grad_norm": 1.0309376843616966, + "learning_rate": 6.08505961721255e-06, + "loss": 0.5907, + "step": 2540 + }, + { + "epoch": 2.4387902064330294, + "grad_norm": 0.9774921339196672, + "learning_rate": 6.081789246802731e-06, + "loss": 0.5253, + "step": 2541 + }, + { + "epoch": 2.439750360057609, + "grad_norm": 1.8226000020033482, + "learning_rate": 6.078518390789282e-06, + "loss": 0.4373, + "step": 2542 + }, + { + "epoch": 2.4407105136821894, + "grad_norm": 1.0582251492749066, + "learning_rate": 6.075247050640458e-06, + "loss": 0.5234, + "step": 2543 + }, + { + "epoch": 2.441670667306769, + "grad_norm": 1.7260864444229784, + "learning_rate": 6.071975227824724e-06, + "loss": 0.6115, + "step": 2544 + }, + { + "epoch": 2.442630820931349, + "grad_norm": 1.7230141368781013, + "learning_rate": 6.068702923810771e-06, + "loss": 0.547, + "step": 2545 + }, + { + "epoch": 2.4435909745559288, + "grad_norm": 1.0337078967488165, + "learning_rate": 6.0654301400675e-06, + "loss": 0.5322, + "step": 2546 + }, + { + "epoch": 2.444551128180509, + "grad_norm": 1.077894068675427, + "learning_rate": 6.062156878064026e-06, + "loss": 0.5091, + "step": 2547 + }, + { + "epoch": 2.445511281805089, + "grad_norm": 1.3279128101668733, + "learning_rate": 6.058883139269685e-06, + "loss": 0.5047, + "step": 2548 + }, + { + "epoch": 2.4464714354296686, + "grad_norm": 2.147066360747897, + "learning_rate": 6.055608925154019e-06, + "loss": 0.4498, + "step": 2549 + }, + { + "epoch": 2.447431589054249, + "grad_norm": 1.3374049080761576, + "learning_rate": 6.052334237186792e-06, + "loss": 0.5532, + "step": 2550 + }, + { + "epoch": 2.4483917426788286, + "grad_norm": 1.0663473328404744, + "learning_rate": 6.049059076837974e-06, + "loss": 0.7161, + "step": 2551 + }, + { + "epoch": 2.4493518963034084, + "grad_norm": 1.0044549382114532, + "learning_rate": 6.045783445577752e-06, + "loss": 0.522, + "step": 2552 + }, + { + "epoch": 2.4503120499279882, + "grad_norm": 1.3742670890197568, + "learning_rate": 6.042507344876518e-06, + "loss": 0.5938, + "step": 2553 + }, + { + "epoch": 2.4512722035525685, + "grad_norm": 0.910650313970339, + "learning_rate": 6.0392307762048816e-06, + "loss": 0.5379, + "step": 2554 + }, + { + "epoch": 2.4522323571771483, + "grad_norm": 0.9627230310579845, + "learning_rate": 6.03595374103366e-06, + "loss": 0.5392, + "step": 2555 + }, + { + "epoch": 2.4531925108017285, + "grad_norm": 1.145458285959235, + "learning_rate": 6.032676240833876e-06, + "loss": 0.5202, + "step": 2556 + }, + { + "epoch": 2.4541526644263083, + "grad_norm": 0.8140750529777154, + "learning_rate": 6.029398277076766e-06, + "loss": 0.5469, + "step": 2557 + }, + { + "epoch": 2.455112818050888, + "grad_norm": 1.0405936049835796, + "learning_rate": 6.0261198512337725e-06, + "loss": 0.4157, + "step": 2558 + }, + { + "epoch": 2.456072971675468, + "grad_norm": 0.8788759787462748, + "learning_rate": 6.022840964776548e-06, + "loss": 0.5172, + "step": 2559 + }, + { + "epoch": 2.457033125300048, + "grad_norm": 1.0732883769248467, + "learning_rate": 6.019561619176948e-06, + "loss": 0.5356, + "step": 2560 + }, + { + "epoch": 2.457993278924628, + "grad_norm": 0.8906641941924691, + "learning_rate": 6.016281815907036e-06, + "loss": 0.5221, + "step": 2561 + }, + { + "epoch": 2.4589534325492077, + "grad_norm": 1.2324033430189187, + "learning_rate": 6.013001556439079e-06, + "loss": 0.4163, + "step": 2562 + }, + { + "epoch": 2.459913586173788, + "grad_norm": 1.2455497806231217, + "learning_rate": 6.0097208422455526e-06, + "loss": 0.5585, + "step": 2563 + }, + { + "epoch": 2.4608737397983678, + "grad_norm": 1.0533865252806114, + "learning_rate": 6.006439674799132e-06, + "loss": 0.5587, + "step": 2564 + }, + { + "epoch": 2.4618338934229476, + "grad_norm": 0.8556710253638442, + "learning_rate": 6.0031580555727005e-06, + "loss": 0.505, + "step": 2565 + }, + { + "epoch": 2.4627940470475274, + "grad_norm": 1.0939080771517722, + "learning_rate": 5.999875986039341e-06, + "loss": 0.5696, + "step": 2566 + }, + { + "epoch": 2.4637542006721076, + "grad_norm": 0.938244140579834, + "learning_rate": 5.996593467672339e-06, + "loss": 0.5682, + "step": 2567 + }, + { + "epoch": 2.4647143542966874, + "grad_norm": 1.0243723639883222, + "learning_rate": 5.993310501945184e-06, + "loss": 0.5133, + "step": 2568 + }, + { + "epoch": 2.4656745079212676, + "grad_norm": 1.039600891886981, + "learning_rate": 5.990027090331562e-06, + "loss": 0.6911, + "step": 2569 + }, + { + "epoch": 2.4666346615458474, + "grad_norm": 0.8806020071668524, + "learning_rate": 5.986743234305361e-06, + "loss": 0.5951, + "step": 2570 + }, + { + "epoch": 2.4675948151704272, + "grad_norm": 0.9458631120569398, + "learning_rate": 5.983458935340669e-06, + "loss": 0.5176, + "step": 2571 + }, + { + "epoch": 2.468554968795007, + "grad_norm": 1.6442454513402067, + "learning_rate": 5.980174194911776e-06, + "loss": 0.5958, + "step": 2572 + }, + { + "epoch": 2.4695151224195873, + "grad_norm": 1.4319293390390937, + "learning_rate": 5.9768890144931626e-06, + "loss": 0.5127, + "step": 2573 + }, + { + "epoch": 2.470475276044167, + "grad_norm": 1.8969459507515172, + "learning_rate": 5.973603395559514e-06, + "loss": 0.5942, + "step": 2574 + }, + { + "epoch": 2.471435429668747, + "grad_norm": 0.9224176575825475, + "learning_rate": 5.970317339585707e-06, + "loss": 0.5498, + "step": 2575 + }, + { + "epoch": 2.472395583293327, + "grad_norm": 1.1947498550400801, + "learning_rate": 5.96703084804682e-06, + "loss": 0.4773, + "step": 2576 + }, + { + "epoch": 2.473355736917907, + "grad_norm": 1.004578430059494, + "learning_rate": 5.963743922418122e-06, + "loss": 0.583, + "step": 2577 + }, + { + "epoch": 2.4743158905424867, + "grad_norm": 1.357725894295687, + "learning_rate": 5.960456564175079e-06, + "loss": 0.5136, + "step": 2578 + }, + { + "epoch": 2.475276044167067, + "grad_norm": 0.966777258485813, + "learning_rate": 5.9571687747933505e-06, + "loss": 0.6178, + "step": 2579 + }, + { + "epoch": 2.4762361977916467, + "grad_norm": 0.971768856681797, + "learning_rate": 5.9538805557487896e-06, + "loss": 0.6749, + "step": 2580 + }, + { + "epoch": 2.4771963514162265, + "grad_norm": 1.194876344310327, + "learning_rate": 5.950591908517446e-06, + "loss": 0.5473, + "step": 2581 + }, + { + "epoch": 2.4781565050408068, + "grad_norm": 1.3229983395220837, + "learning_rate": 5.9473028345755535e-06, + "loss": 0.5612, + "step": 2582 + }, + { + "epoch": 2.4791166586653866, + "grad_norm": 1.0317895502097822, + "learning_rate": 5.944013335399546e-06, + "loss": 0.6091, + "step": 2583 + }, + { + "epoch": 2.4800768122899663, + "grad_norm": 1.0414457476888133, + "learning_rate": 5.940723412466044e-06, + "loss": 0.5305, + "step": 2584 + }, + { + "epoch": 2.481036965914546, + "grad_norm": 1.3840832496866207, + "learning_rate": 5.937433067251857e-06, + "loss": 0.5803, + "step": 2585 + }, + { + "epoch": 2.4819971195391264, + "grad_norm": 0.9928133197922491, + "learning_rate": 5.934142301233986e-06, + "loss": 0.618, + "step": 2586 + }, + { + "epoch": 2.482957273163706, + "grad_norm": 1.0766586909207183, + "learning_rate": 5.9308511158896235e-06, + "loss": 0.5838, + "step": 2587 + }, + { + "epoch": 2.483917426788286, + "grad_norm": 1.1825859995947656, + "learning_rate": 5.927559512696143e-06, + "loss": 0.6283, + "step": 2588 + }, + { + "epoch": 2.484877580412866, + "grad_norm": 1.032415138872871, + "learning_rate": 5.924267493131113e-06, + "loss": 0.6022, + "step": 2589 + }, + { + "epoch": 2.485837734037446, + "grad_norm": 0.9628596424680305, + "learning_rate": 5.9209750586722855e-06, + "loss": 0.6158, + "step": 2590 + }, + { + "epoch": 2.486797887662026, + "grad_norm": 1.2887726983995444, + "learning_rate": 5.917682210797598e-06, + "loss": 0.6021, + "step": 2591 + }, + { + "epoch": 2.487758041286606, + "grad_norm": 1.0204898747641036, + "learning_rate": 5.9143889509851785e-06, + "loss": 0.4879, + "step": 2592 + }, + { + "epoch": 2.488718194911186, + "grad_norm": 1.151434491288258, + "learning_rate": 5.91109528071333e-06, + "loss": 0.5356, + "step": 2593 + }, + { + "epoch": 2.4896783485357656, + "grad_norm": 0.9842095217042718, + "learning_rate": 5.9078012014605505e-06, + "loss": 0.4471, + "step": 2594 + }, + { + "epoch": 2.490638502160346, + "grad_norm": 1.33150273835374, + "learning_rate": 5.904506714705514e-06, + "loss": 0.6534, + "step": 2595 + }, + { + "epoch": 2.4915986557849257, + "grad_norm": 1.7258705689035156, + "learning_rate": 5.901211821927084e-06, + "loss": 0.6171, + "step": 2596 + }, + { + "epoch": 2.4925588094095055, + "grad_norm": 1.2626862822791978, + "learning_rate": 5.897916524604299e-06, + "loss": 0.5844, + "step": 2597 + }, + { + "epoch": 2.4935189630340853, + "grad_norm": 0.9250204182091345, + "learning_rate": 5.894620824216385e-06, + "loss": 0.5661, + "step": 2598 + }, + { + "epoch": 2.4944791166586655, + "grad_norm": 1.1826866287673776, + "learning_rate": 5.8913247222427464e-06, + "loss": 0.5676, + "step": 2599 + }, + { + "epoch": 2.4954392702832453, + "grad_norm": 0.9766514675342373, + "learning_rate": 5.888028220162967e-06, + "loss": 0.5587, + "step": 2600 + }, + { + "epoch": 2.496399423907825, + "grad_norm": 0.8732797696155599, + "learning_rate": 5.884731319456812e-06, + "loss": 0.5236, + "step": 2601 + }, + { + "epoch": 2.4973595775324053, + "grad_norm": 0.8430315332928399, + "learning_rate": 5.881434021604224e-06, + "loss": 0.3869, + "step": 2602 + }, + { + "epoch": 2.498319731156985, + "grad_norm": 1.274877961681524, + "learning_rate": 5.878136328085328e-06, + "loss": 0.6309, + "step": 2603 + }, + { + "epoch": 2.499279884781565, + "grad_norm": 1.8930945096618759, + "learning_rate": 5.874838240380419e-06, + "loss": 0.5417, + "step": 2604 + }, + { + "epoch": 2.5002400384061447, + "grad_norm": 1.1010418480785087, + "learning_rate": 5.871539759969978e-06, + "loss": 0.4826, + "step": 2605 + }, + { + "epoch": 2.501200192030725, + "grad_norm": 1.88062980130338, + "learning_rate": 5.8682408883346535e-06, + "loss": 0.585, + "step": 2606 + }, + { + "epoch": 2.5021603456553048, + "grad_norm": 1.212461006479849, + "learning_rate": 5.864941626955274e-06, + "loss": 0.5101, + "step": 2607 + }, + { + "epoch": 2.503120499279885, + "grad_norm": 0.8668789311085789, + "learning_rate": 5.861641977312844e-06, + "loss": 0.5702, + "step": 2608 + }, + { + "epoch": 2.504080652904465, + "grad_norm": 0.8667080794531593, + "learning_rate": 5.858341940888543e-06, + "loss": 0.5367, + "step": 2609 + }, + { + "epoch": 2.5050408065290446, + "grad_norm": 1.07778753805212, + "learning_rate": 5.855041519163718e-06, + "loss": 0.5217, + "step": 2610 + }, + { + "epoch": 2.5060009601536244, + "grad_norm": 1.1078478028218475, + "learning_rate": 5.851740713619896e-06, + "loss": 0.521, + "step": 2611 + }, + { + "epoch": 2.5069611137782046, + "grad_norm": 1.179715917210379, + "learning_rate": 5.848439525738772e-06, + "loss": 0.5676, + "step": 2612 + }, + { + "epoch": 2.5079212674027844, + "grad_norm": 1.1097176015372312, + "learning_rate": 5.845137957002214e-06, + "loss": 0.6292, + "step": 2613 + }, + { + "epoch": 2.5088814210273642, + "grad_norm": 1.2984048738403013, + "learning_rate": 5.841836008892262e-06, + "loss": 0.5074, + "step": 2614 + }, + { + "epoch": 2.5098415746519445, + "grad_norm": 1.0559915309308197, + "learning_rate": 5.838533682891123e-06, + "loss": 0.5888, + "step": 2615 + }, + { + "epoch": 2.5108017282765243, + "grad_norm": 0.9964536318540524, + "learning_rate": 5.835230980481177e-06, + "loss": 0.5694, + "step": 2616 + }, + { + "epoch": 2.511761881901104, + "grad_norm": 1.1724431371098636, + "learning_rate": 5.831927903144972e-06, + "loss": 0.507, + "step": 2617 + }, + { + "epoch": 2.512722035525684, + "grad_norm": 0.9083953835098307, + "learning_rate": 5.828624452365226e-06, + "loss": 0.5548, + "step": 2618 + }, + { + "epoch": 2.513682189150264, + "grad_norm": 1.5918065837595812, + "learning_rate": 5.82532062962482e-06, + "loss": 0.5391, + "step": 2619 + }, + { + "epoch": 2.514642342774844, + "grad_norm": 1.010645399404615, + "learning_rate": 5.822016436406805e-06, + "loss": 0.5103, + "step": 2620 + }, + { + "epoch": 2.515602496399424, + "grad_norm": 1.3509602518688228, + "learning_rate": 5.818711874194398e-06, + "loss": 0.5193, + "step": 2621 + }, + { + "epoch": 2.516562650024004, + "grad_norm": 1.2127030501075124, + "learning_rate": 5.8154069444709825e-06, + "loss": 0.5876, + "step": 2622 + }, + { + "epoch": 2.5175228036485837, + "grad_norm": 1.0025237235571185, + "learning_rate": 5.8121016487201074e-06, + "loss": 0.4904, + "step": 2623 + }, + { + "epoch": 2.5184829572731635, + "grad_norm": 1.0354864302247444, + "learning_rate": 5.80879598842548e-06, + "loss": 0.6199, + "step": 2624 + }, + { + "epoch": 2.5194431108977438, + "grad_norm": 1.5814691353189114, + "learning_rate": 5.805489965070983e-06, + "loss": 0.5325, + "step": 2625 + }, + { + "epoch": 2.5204032645223235, + "grad_norm": 0.8799344319638772, + "learning_rate": 5.8021835801406475e-06, + "loss": 0.5243, + "step": 2626 + }, + { + "epoch": 2.5213634181469033, + "grad_norm": 0.909422512011236, + "learning_rate": 5.798876835118681e-06, + "loss": 0.5372, + "step": 2627 + }, + { + "epoch": 2.5223235717714836, + "grad_norm": 1.1460977214328045, + "learning_rate": 5.79556973148944e-06, + "loss": 0.481, + "step": 2628 + }, + { + "epoch": 2.5232837253960634, + "grad_norm": 1.06360120832438, + "learning_rate": 5.792262270737451e-06, + "loss": 0.5626, + "step": 2629 + }, + { + "epoch": 2.524243879020643, + "grad_norm": 1.0066875885568904, + "learning_rate": 5.788954454347397e-06, + "loss": 0.5557, + "step": 2630 + }, + { + "epoch": 2.525204032645223, + "grad_norm": 1.2447816180455138, + "learning_rate": 5.7856462838041235e-06, + "loss": 0.5445, + "step": 2631 + }, + { + "epoch": 2.526164186269803, + "grad_norm": 1.4941631381271145, + "learning_rate": 5.782337760592631e-06, + "loss": 0.5912, + "step": 2632 + }, + { + "epoch": 2.527124339894383, + "grad_norm": 0.8837403230313842, + "learning_rate": 5.7790288861980795e-06, + "loss": 0.5031, + "step": 2633 + }, + { + "epoch": 2.5280844935189632, + "grad_norm": 1.0322148290442799, + "learning_rate": 5.77571966210579e-06, + "loss": 0.5291, + "step": 2634 + }, + { + "epoch": 2.529044647143543, + "grad_norm": 1.074461969605371, + "learning_rate": 5.772410089801234e-06, + "loss": 0.5288, + "step": 2635 + }, + { + "epoch": 2.530004800768123, + "grad_norm": 1.2540822198794377, + "learning_rate": 5.769100170770048e-06, + "loss": 0.3869, + "step": 2636 + }, + { + "epoch": 2.5309649543927026, + "grad_norm": 1.2917790473998079, + "learning_rate": 5.765789906498015e-06, + "loss": 0.4972, + "step": 2637 + }, + { + "epoch": 2.531925108017283, + "grad_norm": 1.185457765621514, + "learning_rate": 5.762479298471078e-06, + "loss": 0.6066, + "step": 2638 + }, + { + "epoch": 2.5328852616418627, + "grad_norm": 1.615534221090781, + "learning_rate": 5.759168348175335e-06, + "loss": 0.6056, + "step": 2639 + }, + { + "epoch": 2.5338454152664425, + "grad_norm": 1.039914366142861, + "learning_rate": 5.755857057097036e-06, + "loss": 0.624, + "step": 2640 + }, + { + "epoch": 2.5348055688910227, + "grad_norm": 1.2310248807706055, + "learning_rate": 5.752545426722583e-06, + "loss": 0.5919, + "step": 2641 + }, + { + "epoch": 2.5357657225156025, + "grad_norm": 0.963716619513092, + "learning_rate": 5.7492334585385325e-06, + "loss": 0.5855, + "step": 2642 + }, + { + "epoch": 2.5367258761401823, + "grad_norm": 0.8762076343673909, + "learning_rate": 5.745921154031593e-06, + "loss": 0.5786, + "step": 2643 + }, + { + "epoch": 2.537686029764762, + "grad_norm": 0.7901722509181774, + "learning_rate": 5.742608514688619e-06, + "loss": 0.4986, + "step": 2644 + }, + { + "epoch": 2.5386461833893423, + "grad_norm": 0.9899570633713458, + "learning_rate": 5.739295541996622e-06, + "loss": 0.5513, + "step": 2645 + }, + { + "epoch": 2.539606337013922, + "grad_norm": 1.1860857708819632, + "learning_rate": 5.7359822374427565e-06, + "loss": 0.4888, + "step": 2646 + }, + { + "epoch": 2.5405664906385024, + "grad_norm": 0.9829635047820741, + "learning_rate": 5.732668602514334e-06, + "loss": 0.4516, + "step": 2647 + }, + { + "epoch": 2.541526644263082, + "grad_norm": 0.857473761716564, + "learning_rate": 5.729354638698806e-06, + "loss": 0.4112, + "step": 2648 + }, + { + "epoch": 2.542486797887662, + "grad_norm": 0.8539632466434419, + "learning_rate": 5.726040347483779e-06, + "loss": 0.5202, + "step": 2649 + }, + { + "epoch": 2.5434469515122418, + "grad_norm": 1.349597198191044, + "learning_rate": 5.722725730357001e-06, + "loss": 0.6785, + "step": 2650 + }, + { + "epoch": 2.544407105136822, + "grad_norm": 1.476308254477948, + "learning_rate": 5.719410788806366e-06, + "loss": 0.4867, + "step": 2651 + }, + { + "epoch": 2.545367258761402, + "grad_norm": 1.1364375560304625, + "learning_rate": 5.716095524319919e-06, + "loss": 0.4572, + "step": 2652 + }, + { + "epoch": 2.5463274123859816, + "grad_norm": 0.9292151349949764, + "learning_rate": 5.712779938385846e-06, + "loss": 0.5275, + "step": 2653 + }, + { + "epoch": 2.547287566010562, + "grad_norm": 1.0603683681510738, + "learning_rate": 5.709464032492478e-06, + "loss": 0.6118, + "step": 2654 + }, + { + "epoch": 2.5482477196351416, + "grad_norm": 1.0695585278526674, + "learning_rate": 5.706147808128288e-06, + "loss": 0.541, + "step": 2655 + }, + { + "epoch": 2.5492078732597214, + "grad_norm": 1.1713880900693274, + "learning_rate": 5.702831266781898e-06, + "loss": 0.5174, + "step": 2656 + }, + { + "epoch": 2.550168026884301, + "grad_norm": 1.0340443847712604, + "learning_rate": 5.699514409942064e-06, + "loss": 0.4663, + "step": 2657 + }, + { + "epoch": 2.5511281805088815, + "grad_norm": 1.6840399739400633, + "learning_rate": 5.69619723909769e-06, + "loss": 0.5886, + "step": 2658 + }, + { + "epoch": 2.5520883341334613, + "grad_norm": 1.5517188707837135, + "learning_rate": 5.6928797557378145e-06, + "loss": 0.493, + "step": 2659 + }, + { + "epoch": 2.5530484877580415, + "grad_norm": 1.01998223291034, + "learning_rate": 5.689561961351624e-06, + "loss": 0.506, + "step": 2660 + }, + { + "epoch": 2.5540086413826213, + "grad_norm": 1.2733355383158487, + "learning_rate": 5.68624385742844e-06, + "loss": 0.568, + "step": 2661 + }, + { + "epoch": 2.554968795007201, + "grad_norm": 1.3483186301470194, + "learning_rate": 5.682925445457726e-06, + "loss": 0.4972, + "step": 2662 + }, + { + "epoch": 2.555928948631781, + "grad_norm": 0.9104335531564698, + "learning_rate": 5.679606726929076e-06, + "loss": 0.5291, + "step": 2663 + }, + { + "epoch": 2.556889102256361, + "grad_norm": 1.1588795293210066, + "learning_rate": 5.676287703332234e-06, + "loss": 0.553, + "step": 2664 + }, + { + "epoch": 2.557849255880941, + "grad_norm": 1.0427251508783315, + "learning_rate": 5.672968376157071e-06, + "loss": 0.4908, + "step": 2665 + }, + { + "epoch": 2.558809409505521, + "grad_norm": 0.943102809941431, + "learning_rate": 5.6696487468935955e-06, + "loss": 0.5077, + "step": 2666 + }, + { + "epoch": 2.559769563130101, + "grad_norm": 1.0191375722161495, + "learning_rate": 5.666328817031958e-06, + "loss": 0.5781, + "step": 2667 + }, + { + "epoch": 2.5607297167546808, + "grad_norm": 1.0860080564510013, + "learning_rate": 5.663008588062435e-06, + "loss": 0.5606, + "step": 2668 + }, + { + "epoch": 2.5616898703792605, + "grad_norm": 1.6069535272566735, + "learning_rate": 5.659688061475445e-06, + "loss": 0.6415, + "step": 2669 + }, + { + "epoch": 2.5626500240038403, + "grad_norm": 0.9241312991066435, + "learning_rate": 5.656367238761536e-06, + "loss": 0.5017, + "step": 2670 + }, + { + "epoch": 2.5636101776284206, + "grad_norm": 1.1079602817747738, + "learning_rate": 5.653046121411389e-06, + "loss": 0.5436, + "step": 2671 + }, + { + "epoch": 2.5645703312530004, + "grad_norm": 1.2295516632519532, + "learning_rate": 5.649724710915819e-06, + "loss": 0.5775, + "step": 2672 + }, + { + "epoch": 2.5655304848775806, + "grad_norm": 0.8842835928651863, + "learning_rate": 5.646403008765772e-06, + "loss": 0.5541, + "step": 2673 + }, + { + "epoch": 2.5664906385021604, + "grad_norm": 0.815281761560292, + "learning_rate": 5.643081016452323e-06, + "loss": 0.482, + "step": 2674 + }, + { + "epoch": 2.56745079212674, + "grad_norm": 1.029634729715191, + "learning_rate": 5.639758735466682e-06, + "loss": 0.5659, + "step": 2675 + }, + { + "epoch": 2.56841094575132, + "grad_norm": 1.1812805553945471, + "learning_rate": 5.636436167300181e-06, + "loss": 0.6244, + "step": 2676 + }, + { + "epoch": 2.5693710993759002, + "grad_norm": 0.9321973409793013, + "learning_rate": 5.633113313444288e-06, + "loss": 0.6212, + "step": 2677 + }, + { + "epoch": 2.57033125300048, + "grad_norm": 1.1961374002922658, + "learning_rate": 5.629790175390598e-06, + "loss": 0.5985, + "step": 2678 + }, + { + "epoch": 2.5712914066250603, + "grad_norm": 1.5530231817597748, + "learning_rate": 5.626466754630829e-06, + "loss": 0.5804, + "step": 2679 + }, + { + "epoch": 2.57225156024964, + "grad_norm": 1.2801498988049118, + "learning_rate": 5.623143052656831e-06, + "loss": 0.593, + "step": 2680 + }, + { + "epoch": 2.57321171387422, + "grad_norm": 0.8825201031543539, + "learning_rate": 5.619819070960577e-06, + "loss": 0.5012, + "step": 2681 + }, + { + "epoch": 2.5741718674987997, + "grad_norm": 1.1200609022324564, + "learning_rate": 5.616494811034168e-06, + "loss": 0.5668, + "step": 2682 + }, + { + "epoch": 2.5751320211233795, + "grad_norm": 1.004337461449322, + "learning_rate": 5.613170274369827e-06, + "loss": 0.4849, + "step": 2683 + }, + { + "epoch": 2.5760921747479597, + "grad_norm": 0.8727360467170426, + "learning_rate": 5.609845462459907e-06, + "loss": 0.5333, + "step": 2684 + }, + { + "epoch": 2.5770523283725395, + "grad_norm": 1.0406545366923414, + "learning_rate": 5.606520376796875e-06, + "loss": 0.5216, + "step": 2685 + }, + { + "epoch": 2.5780124819971197, + "grad_norm": 0.919124011318109, + "learning_rate": 5.603195018873329e-06, + "loss": 0.6206, + "step": 2686 + }, + { + "epoch": 2.5789726356216995, + "grad_norm": 1.4852216859352825, + "learning_rate": 5.59986939018199e-06, + "loss": 0.5242, + "step": 2687 + }, + { + "epoch": 2.5799327892462793, + "grad_norm": 1.2000340248357038, + "learning_rate": 5.596543492215691e-06, + "loss": 0.6142, + "step": 2688 + }, + { + "epoch": 2.580892942870859, + "grad_norm": 1.2157078489152757, + "learning_rate": 5.5932173264673975e-06, + "loss": 0.5207, + "step": 2689 + }, + { + "epoch": 2.5818530964954394, + "grad_norm": 1.226909166384669, + "learning_rate": 5.589890894430186e-06, + "loss": 0.5099, + "step": 2690 + }, + { + "epoch": 2.582813250120019, + "grad_norm": 1.38801630395136, + "learning_rate": 5.586564197597256e-06, + "loss": 0.5845, + "step": 2691 + }, + { + "epoch": 2.5837734037445994, + "grad_norm": 1.2007598474081262, + "learning_rate": 5.583237237461929e-06, + "loss": 0.5644, + "step": 2692 + }, + { + "epoch": 2.584733557369179, + "grad_norm": 1.3318445225092201, + "learning_rate": 5.579910015517642e-06, + "loss": 0.5989, + "step": 2693 + }, + { + "epoch": 2.585693710993759, + "grad_norm": 1.336662008737483, + "learning_rate": 5.576582533257946e-06, + "loss": 0.5732, + "step": 2694 + }, + { + "epoch": 2.586653864618339, + "grad_norm": 0.9636159574045997, + "learning_rate": 5.573254792176515e-06, + "loss": 0.6207, + "step": 2695 + }, + { + "epoch": 2.5876140182429186, + "grad_norm": 1.214235939611955, + "learning_rate": 5.569926793767136e-06, + "loss": 0.5635, + "step": 2696 + }, + { + "epoch": 2.588574171867499, + "grad_norm": 1.061638704288488, + "learning_rate": 5.566598539523711e-06, + "loss": 0.6396, + "step": 2697 + }, + { + "epoch": 2.5895343254920786, + "grad_norm": 1.477231542756541, + "learning_rate": 5.563270030940258e-06, + "loss": 0.4549, + "step": 2698 + }, + { + "epoch": 2.590494479116659, + "grad_norm": 1.6126428115809528, + "learning_rate": 5.559941269510908e-06, + "loss": 0.6009, + "step": 2699 + }, + { + "epoch": 2.5914546327412387, + "grad_norm": 0.863195874077329, + "learning_rate": 5.556612256729909e-06, + "loss": 0.5583, + "step": 2700 + }, + { + "epoch": 2.5924147863658185, + "grad_norm": 0.96510297442427, + "learning_rate": 5.553282994091617e-06, + "loss": 0.612, + "step": 2701 + }, + { + "epoch": 2.5933749399903983, + "grad_norm": 1.3476599241059, + "learning_rate": 5.549953483090506e-06, + "loss": 0.5237, + "step": 2702 + }, + { + "epoch": 2.5943350936149785, + "grad_norm": 0.8232905487050111, + "learning_rate": 5.546623725221152e-06, + "loss": 0.4715, + "step": 2703 + }, + { + "epoch": 2.5952952472395583, + "grad_norm": 1.817626592285635, + "learning_rate": 5.543293721978251e-06, + "loss": 0.6081, + "step": 2704 + }, + { + "epoch": 2.5962554008641385, + "grad_norm": 0.9796305099231157, + "learning_rate": 5.539963474856607e-06, + "loss": 0.502, + "step": 2705 + }, + { + "epoch": 2.5972155544887183, + "grad_norm": 1.1016208610029354, + "learning_rate": 5.536632985351132e-06, + "loss": 0.5806, + "step": 2706 + }, + { + "epoch": 2.598175708113298, + "grad_norm": 1.1112838847468167, + "learning_rate": 5.533302254956843e-06, + "loss": 0.5873, + "step": 2707 + }, + { + "epoch": 2.599135861737878, + "grad_norm": 1.3974448758581155, + "learning_rate": 5.529971285168873e-06, + "loss": 0.5397, + "step": 2708 + }, + { + "epoch": 2.600096015362458, + "grad_norm": 1.102305687285218, + "learning_rate": 5.526640077482462e-06, + "loss": 0.5855, + "step": 2709 + }, + { + "epoch": 2.601056168987038, + "grad_norm": 1.0834105086173802, + "learning_rate": 5.523308633392945e-06, + "loss": 0.5582, + "step": 2710 + }, + { + "epoch": 2.6020163226116177, + "grad_norm": 1.1138895731472624, + "learning_rate": 5.519976954395779e-06, + "loss": 0.6421, + "step": 2711 + }, + { + "epoch": 2.602976476236198, + "grad_norm": 1.0725641102844308, + "learning_rate": 5.516645041986516e-06, + "loss": 0.4871, + "step": 2712 + }, + { + "epoch": 2.603936629860778, + "grad_norm": 1.1949801296467149, + "learning_rate": 5.513312897660815e-06, + "loss": 0.5309, + "step": 2713 + }, + { + "epoch": 2.6048967834853576, + "grad_norm": 0.7610137344554979, + "learning_rate": 5.509980522914442e-06, + "loss": 0.5354, + "step": 2714 + }, + { + "epoch": 2.6058569371099374, + "grad_norm": 1.1752312404192111, + "learning_rate": 5.506647919243265e-06, + "loss": 0.6004, + "step": 2715 + }, + { + "epoch": 2.6068170907345176, + "grad_norm": 1.617690143178605, + "learning_rate": 5.503315088143251e-06, + "loss": 0.5618, + "step": 2716 + }, + { + "epoch": 2.6077772443590974, + "grad_norm": 1.057282458640301, + "learning_rate": 5.499982031110475e-06, + "loss": 0.6788, + "step": 2717 + }, + { + "epoch": 2.6087373979836777, + "grad_norm": 1.0834313330038319, + "learning_rate": 5.496648749641109e-06, + "loss": 0.5725, + "step": 2718 + }, + { + "epoch": 2.6096975516082574, + "grad_norm": 0.8414741095327913, + "learning_rate": 5.493315245231429e-06, + "loss": 0.5333, + "step": 2719 + }, + { + "epoch": 2.6106577052328372, + "grad_norm": 2.0680047470240686, + "learning_rate": 5.489981519377808e-06, + "loss": 0.6605, + "step": 2720 + }, + { + "epoch": 2.611617858857417, + "grad_norm": 1.375141648361915, + "learning_rate": 5.486647573576721e-06, + "loss": 0.5187, + "step": 2721 + }, + { + "epoch": 2.6125780124819973, + "grad_norm": 1.0374484919564513, + "learning_rate": 5.483313409324741e-06, + "loss": 0.5978, + "step": 2722 + }, + { + "epoch": 2.613538166106577, + "grad_norm": 0.8794906694859634, + "learning_rate": 5.479979028118536e-06, + "loss": 0.4158, + "step": 2723 + }, + { + "epoch": 2.614498319731157, + "grad_norm": 0.9556415312439337, + "learning_rate": 5.476644431454878e-06, + "loss": 0.5909, + "step": 2724 + }, + { + "epoch": 2.615458473355737, + "grad_norm": 1.042097717657121, + "learning_rate": 5.473309620830628e-06, + "loss": 0.4337, + "step": 2725 + }, + { + "epoch": 2.616418626980317, + "grad_norm": 1.0235818012601186, + "learning_rate": 5.469974597742749e-06, + "loss": 0.5137, + "step": 2726 + }, + { + "epoch": 2.6173787806048967, + "grad_norm": 0.8434336986130978, + "learning_rate": 5.4666393636882965e-06, + "loss": 0.5856, + "step": 2727 + }, + { + "epoch": 2.6183389342294765, + "grad_norm": 1.0289269282223525, + "learning_rate": 5.463303920164423e-06, + "loss": 0.5726, + "step": 2728 + }, + { + "epoch": 2.6192990878540567, + "grad_norm": 0.876350909586272, + "learning_rate": 5.459968268668372e-06, + "loss": 0.4906, + "step": 2729 + }, + { + "epoch": 2.6202592414786365, + "grad_norm": 1.2840493491555898, + "learning_rate": 5.456632410697481e-06, + "loss": 0.5594, + "step": 2730 + }, + { + "epoch": 2.6212193951032168, + "grad_norm": 1.0127773694122186, + "learning_rate": 5.453296347749183e-06, + "loss": 0.5075, + "step": 2731 + }, + { + "epoch": 2.6221795487277966, + "grad_norm": 1.3860567146301412, + "learning_rate": 5.449960081321e-06, + "loss": 0.4575, + "step": 2732 + }, + { + "epoch": 2.6231397023523764, + "grad_norm": 1.0199533453696517, + "learning_rate": 5.446623612910545e-06, + "loss": 0.5225, + "step": 2733 + }, + { + "epoch": 2.624099855976956, + "grad_norm": 0.8269796783584771, + "learning_rate": 5.443286944015525e-06, + "loss": 0.5097, + "step": 2734 + }, + { + "epoch": 2.6250600096015364, + "grad_norm": 1.0664631979932955, + "learning_rate": 5.439950076133734e-06, + "loss": 0.5606, + "step": 2735 + }, + { + "epoch": 2.626020163226116, + "grad_norm": 0.8895570701010829, + "learning_rate": 5.4366130107630566e-06, + "loss": 0.4576, + "step": 2736 + }, + { + "epoch": 2.626980316850696, + "grad_norm": 0.9217817865861233, + "learning_rate": 5.4332757494014675e-06, + "loss": 0.5129, + "step": 2737 + }, + { + "epoch": 2.6279404704752762, + "grad_norm": 1.066177295205498, + "learning_rate": 5.429938293547024e-06, + "loss": 0.5681, + "step": 2738 + }, + { + "epoch": 2.628900624099856, + "grad_norm": 1.0606628802778508, + "learning_rate": 5.426600644697877e-06, + "loss": 0.5056, + "step": 2739 + }, + { + "epoch": 2.629860777724436, + "grad_norm": 1.0503053541421365, + "learning_rate": 5.423262804352262e-06, + "loss": 0.4709, + "step": 2740 + }, + { + "epoch": 2.6308209313490156, + "grad_norm": 1.8386042944937735, + "learning_rate": 5.419924774008496e-06, + "loss": 0.601, + "step": 2741 + }, + { + "epoch": 2.631781084973596, + "grad_norm": 1.082547521247347, + "learning_rate": 5.416586555164989e-06, + "loss": 0.5412, + "step": 2742 + }, + { + "epoch": 2.6327412385981757, + "grad_norm": 0.8814343852429971, + "learning_rate": 5.4132481493202295e-06, + "loss": 0.586, + "step": 2743 + }, + { + "epoch": 2.633701392222756, + "grad_norm": 1.1250651363324953, + "learning_rate": 5.409909557972796e-06, + "loss": 0.5744, + "step": 2744 + }, + { + "epoch": 2.6346615458473357, + "grad_norm": 0.8687291175550564, + "learning_rate": 5.406570782621341e-06, + "loss": 0.558, + "step": 2745 + }, + { + "epoch": 2.6356216994719155, + "grad_norm": 1.030358641675026, + "learning_rate": 5.403231824764609e-06, + "loss": 0.5893, + "step": 2746 + }, + { + "epoch": 2.6365818530964953, + "grad_norm": 1.4855280963727258, + "learning_rate": 5.39989268590142e-06, + "loss": 0.6443, + "step": 2747 + }, + { + "epoch": 2.6375420067210755, + "grad_norm": 1.353291787264611, + "learning_rate": 5.396553367530679e-06, + "loss": 0.5846, + "step": 2748 + }, + { + "epoch": 2.6385021603456553, + "grad_norm": 0.8241599883739923, + "learning_rate": 5.39321387115137e-06, + "loss": 0.6277, + "step": 2749 + }, + { + "epoch": 2.639462313970235, + "grad_norm": 0.9429568535593399, + "learning_rate": 5.389874198262559e-06, + "loss": 0.4257, + "step": 2750 + }, + { + "epoch": 2.6404224675948154, + "grad_norm": 1.2544147684571567, + "learning_rate": 5.386534350363387e-06, + "loss": 0.4699, + "step": 2751 + }, + { + "epoch": 2.641382621219395, + "grad_norm": 0.9451996316895992, + "learning_rate": 5.3831943289530765e-06, + "loss": 0.4887, + "step": 2752 + }, + { + "epoch": 2.642342774843975, + "grad_norm": 1.1899392309777594, + "learning_rate": 5.379854135530929e-06, + "loss": 0.5271, + "step": 2753 + }, + { + "epoch": 2.6433029284685547, + "grad_norm": 0.8425542526503982, + "learning_rate": 5.376513771596319e-06, + "loss": 0.5593, + "step": 2754 + }, + { + "epoch": 2.644263082093135, + "grad_norm": 1.3338829011787532, + "learning_rate": 5.373173238648704e-06, + "loss": 0.5741, + "step": 2755 + }, + { + "epoch": 2.645223235717715, + "grad_norm": 0.9471815196604269, + "learning_rate": 5.369832538187609e-06, + "loss": 0.5136, + "step": 2756 + }, + { + "epoch": 2.646183389342295, + "grad_norm": 1.1035308295953958, + "learning_rate": 5.366491671712641e-06, + "loss": 0.54, + "step": 2757 + }, + { + "epoch": 2.647143542966875, + "grad_norm": 1.290015711595849, + "learning_rate": 5.363150640723479e-06, + "loss": 0.4995, + "step": 2758 + }, + { + "epoch": 2.6481036965914546, + "grad_norm": 1.4030864072237978, + "learning_rate": 5.359809446719878e-06, + "loss": 0.6412, + "step": 2759 + }, + { + "epoch": 2.6490638502160344, + "grad_norm": 1.3374177423017348, + "learning_rate": 5.356468091201661e-06, + "loss": 0.6014, + "step": 2760 + }, + { + "epoch": 2.6500240038406147, + "grad_norm": 1.011465602607862, + "learning_rate": 5.353126575668728e-06, + "loss": 0.4744, + "step": 2761 + }, + { + "epoch": 2.6509841574651944, + "grad_norm": 0.9607418937883007, + "learning_rate": 5.3497849016210515e-06, + "loss": 0.6151, + "step": 2762 + }, + { + "epoch": 2.6519443110897742, + "grad_norm": 0.7651995220803354, + "learning_rate": 5.3464430705586715e-06, + "loss": 0.5129, + "step": 2763 + }, + { + "epoch": 2.6529044647143545, + "grad_norm": 1.0016321728616067, + "learning_rate": 5.3431010839817e-06, + "loss": 0.5879, + "step": 2764 + }, + { + "epoch": 2.6538646183389343, + "grad_norm": 1.0984388087436279, + "learning_rate": 5.339758943390319e-06, + "loss": 0.6534, + "step": 2765 + }, + { + "epoch": 2.654824771963514, + "grad_norm": 0.9621442488908168, + "learning_rate": 5.3364166502847796e-06, + "loss": 0.5664, + "step": 2766 + }, + { + "epoch": 2.655784925588094, + "grad_norm": 0.8468291580428136, + "learning_rate": 5.3330742061654014e-06, + "loss": 0.4852, + "step": 2767 + }, + { + "epoch": 2.656745079212674, + "grad_norm": 1.2985751049055718, + "learning_rate": 5.329731612532575e-06, + "loss": 0.5849, + "step": 2768 + }, + { + "epoch": 2.657705232837254, + "grad_norm": 0.8763111580394592, + "learning_rate": 5.32638887088675e-06, + "loss": 0.5138, + "step": 2769 + }, + { + "epoch": 2.658665386461834, + "grad_norm": 0.7339367242929759, + "learning_rate": 5.323045982728451e-06, + "loss": 0.5476, + "step": 2770 + }, + { + "epoch": 2.659625540086414, + "grad_norm": 1.2811757059202413, + "learning_rate": 5.319702949558263e-06, + "loss": 0.5656, + "step": 2771 + }, + { + "epoch": 2.6605856937109937, + "grad_norm": 1.5681436780587878, + "learning_rate": 5.316359772876839e-06, + "loss": 0.493, + "step": 2772 + }, + { + "epoch": 2.6615458473355735, + "grad_norm": 1.347679937660324, + "learning_rate": 5.313016454184894e-06, + "loss": 0.585, + "step": 2773 + }, + { + "epoch": 2.6625060009601538, + "grad_norm": 0.8893616509634205, + "learning_rate": 5.309672994983209e-06, + "loss": 0.5743, + "step": 2774 + }, + { + "epoch": 2.6634661545847336, + "grad_norm": 1.5476080539515944, + "learning_rate": 5.306329396772627e-06, + "loss": 0.5673, + "step": 2775 + }, + { + "epoch": 2.6644263082093134, + "grad_norm": 1.3915506222380327, + "learning_rate": 5.302985661054053e-06, + "loss": 0.5582, + "step": 2776 + }, + { + "epoch": 2.6653864618338936, + "grad_norm": 0.8927995632650635, + "learning_rate": 5.2996417893284556e-06, + "loss": 0.519, + "step": 2777 + }, + { + "epoch": 2.6663466154584734, + "grad_norm": 1.3392126472571715, + "learning_rate": 5.296297783096861e-06, + "loss": 0.515, + "step": 2778 + }, + { + "epoch": 2.667306769083053, + "grad_norm": 1.0587509975921972, + "learning_rate": 5.292953643860358e-06, + "loss": 0.6577, + "step": 2779 + }, + { + "epoch": 2.668266922707633, + "grad_norm": 0.8777896031068066, + "learning_rate": 5.289609373120096e-06, + "loss": 0.6479, + "step": 2780 + }, + { + "epoch": 2.6692270763322132, + "grad_norm": 1.3657735621927407, + "learning_rate": 5.286264972377283e-06, + "loss": 0.5274, + "step": 2781 + }, + { + "epoch": 2.670187229956793, + "grad_norm": 1.1016616615171386, + "learning_rate": 5.282920443133184e-06, + "loss": 0.545, + "step": 2782 + }, + { + "epoch": 2.6711473835813733, + "grad_norm": 1.4640582077517341, + "learning_rate": 5.279575786889121e-06, + "loss": 0.6541, + "step": 2783 + }, + { + "epoch": 2.672107537205953, + "grad_norm": 0.9447135546130015, + "learning_rate": 5.2762310051464775e-06, + "loss": 0.6143, + "step": 2784 + }, + { + "epoch": 2.673067690830533, + "grad_norm": 0.7081120383251607, + "learning_rate": 5.272886099406688e-06, + "loss": 0.4984, + "step": 2785 + }, + { + "epoch": 2.6740278444551127, + "grad_norm": 1.2125439642491755, + "learning_rate": 5.269541071171246e-06, + "loss": 0.6023, + "step": 2786 + }, + { + "epoch": 2.674987998079693, + "grad_norm": 0.8922467247793752, + "learning_rate": 5.266195921941697e-06, + "loss": 0.4071, + "step": 2787 + }, + { + "epoch": 2.6759481517042727, + "grad_norm": 0.9275700548723981, + "learning_rate": 5.262850653219643e-06, + "loss": 0.5377, + "step": 2788 + }, + { + "epoch": 2.6769083053288525, + "grad_norm": 1.726446147407602, + "learning_rate": 5.259505266506742e-06, + "loss": 0.5053, + "step": 2789 + }, + { + "epoch": 2.6778684589534327, + "grad_norm": 1.3977026317966972, + "learning_rate": 5.256159763304703e-06, + "loss": 0.5421, + "step": 2790 + }, + { + "epoch": 2.6788286125780125, + "grad_norm": 1.1183935336200055, + "learning_rate": 5.252814145115281e-06, + "loss": 0.4734, + "step": 2791 + }, + { + "epoch": 2.6797887662025923, + "grad_norm": 1.0158485746568071, + "learning_rate": 5.249468413440293e-06, + "loss": 0.4863, + "step": 2792 + }, + { + "epoch": 2.680748919827172, + "grad_norm": 0.8616929124320747, + "learning_rate": 5.246122569781601e-06, + "loss": 0.5978, + "step": 2793 + }, + { + "epoch": 2.6817090734517524, + "grad_norm": 0.9340691499609243, + "learning_rate": 5.242776615641119e-06, + "loss": 0.5256, + "step": 2794 + }, + { + "epoch": 2.682669227076332, + "grad_norm": 1.0959314695399902, + "learning_rate": 5.2394305525208076e-06, + "loss": 0.6211, + "step": 2795 + }, + { + "epoch": 2.6836293807009124, + "grad_norm": 0.9167429208775514, + "learning_rate": 5.236084381922681e-06, + "loss": 0.5164, + "step": 2796 + }, + { + "epoch": 2.684589534325492, + "grad_norm": 1.26657483456402, + "learning_rate": 5.232738105348801e-06, + "loss": 0.6207, + "step": 2797 + }, + { + "epoch": 2.685549687950072, + "grad_norm": 1.302400802244244, + "learning_rate": 5.229391724301271e-06, + "loss": 0.6118, + "step": 2798 + }, + { + "epoch": 2.686509841574652, + "grad_norm": 0.9671848101889771, + "learning_rate": 5.22604524028225e-06, + "loss": 0.5496, + "step": 2799 + }, + { + "epoch": 2.687469995199232, + "grad_norm": 1.3124969202474654, + "learning_rate": 5.222698654793936e-06, + "loss": 0.4981, + "step": 2800 + }, + { + "epoch": 2.688430148823812, + "grad_norm": 1.0602400209247174, + "learning_rate": 5.219351969338576e-06, + "loss": 0.5206, + "step": 2801 + }, + { + "epoch": 2.6893903024483916, + "grad_norm": 1.266528249280877, + "learning_rate": 5.2160051854184616e-06, + "loss": 0.4813, + "step": 2802 + }, + { + "epoch": 2.690350456072972, + "grad_norm": 1.6180762661837926, + "learning_rate": 5.212658304535929e-06, + "loss": 0.5331, + "step": 2803 + }, + { + "epoch": 2.6913106096975516, + "grad_norm": 0.9858838394310673, + "learning_rate": 5.209311328193356e-06, + "loss": 0.5161, + "step": 2804 + }, + { + "epoch": 2.6922707633221314, + "grad_norm": 0.8941238520646599, + "learning_rate": 5.205964257893166e-06, + "loss": 0.5927, + "step": 2805 + }, + { + "epoch": 2.6932309169467112, + "grad_norm": 1.4707676253801614, + "learning_rate": 5.202617095137823e-06, + "loss": 0.5158, + "step": 2806 + }, + { + "epoch": 2.6941910705712915, + "grad_norm": 0.9200661719503983, + "learning_rate": 5.199269841429831e-06, + "loss": 0.5481, + "step": 2807 + }, + { + "epoch": 2.6951512241958713, + "grad_norm": 0.907416561880925, + "learning_rate": 5.195922498271737e-06, + "loss": 0.5722, + "step": 2808 + }, + { + "epoch": 2.6961113778204515, + "grad_norm": 1.8798059389301354, + "learning_rate": 5.192575067166126e-06, + "loss": 0.6103, + "step": 2809 + }, + { + "epoch": 2.6970715314450313, + "grad_norm": 1.423299903141147, + "learning_rate": 5.189227549615629e-06, + "loss": 0.6826, + "step": 2810 + }, + { + "epoch": 2.698031685069611, + "grad_norm": 0.851764037938656, + "learning_rate": 5.185879947122906e-06, + "loss": 0.6435, + "step": 2811 + }, + { + "epoch": 2.698991838694191, + "grad_norm": 1.1148566470561174, + "learning_rate": 5.182532261190663e-06, + "loss": 0.6071, + "step": 2812 + }, + { + "epoch": 2.699951992318771, + "grad_norm": 0.8721542840393476, + "learning_rate": 5.179184493321637e-06, + "loss": 0.5941, + "step": 2813 + }, + { + "epoch": 2.700912145943351, + "grad_norm": 0.9319223293987254, + "learning_rate": 5.175836645018608e-06, + "loss": 0.6197, + "step": 2814 + }, + { + "epoch": 2.7018722995679307, + "grad_norm": 0.9165388015037748, + "learning_rate": 5.1724887177843894e-06, + "loss": 0.6133, + "step": 2815 + }, + { + "epoch": 2.702832453192511, + "grad_norm": 1.0465203044552438, + "learning_rate": 5.1691407131218296e-06, + "loss": 0.5232, + "step": 2816 + }, + { + "epoch": 2.7037926068170908, + "grad_norm": 1.1803840783202921, + "learning_rate": 5.165792632533811e-06, + "loss": 0.5574, + "step": 2817 + }, + { + "epoch": 2.7047527604416706, + "grad_norm": 0.8704643622364252, + "learning_rate": 5.162444477523251e-06, + "loss": 0.4348, + "step": 2818 + }, + { + "epoch": 2.7057129140662504, + "grad_norm": 0.9190152242703277, + "learning_rate": 5.159096249593104e-06, + "loss": 0.5118, + "step": 2819 + }, + { + "epoch": 2.7066730676908306, + "grad_norm": 1.0719406102416777, + "learning_rate": 5.1557479502463495e-06, + "loss": 0.5721, + "step": 2820 + }, + { + "epoch": 2.7076332213154104, + "grad_norm": 0.9439114402850619, + "learning_rate": 5.152399580986005e-06, + "loss": 0.5997, + "step": 2821 + }, + { + "epoch": 2.7085933749399906, + "grad_norm": 0.9071815706578216, + "learning_rate": 5.149051143315118e-06, + "loss": 0.4983, + "step": 2822 + }, + { + "epoch": 2.7095535285645704, + "grad_norm": 0.9696598987887864, + "learning_rate": 5.1457026387367655e-06, + "loss": 0.5385, + "step": 2823 + }, + { + "epoch": 2.7105136821891502, + "grad_norm": 0.9638273619431147, + "learning_rate": 5.142354068754057e-06, + "loss": 0.5157, + "step": 2824 + }, + { + "epoch": 2.71147383581373, + "grad_norm": 0.8736657470189363, + "learning_rate": 5.139005434870128e-06, + "loss": 0.543, + "step": 2825 + }, + { + "epoch": 2.7124339894383103, + "grad_norm": 0.8985746559178751, + "learning_rate": 5.135656738588144e-06, + "loss": 0.5126, + "step": 2826 + }, + { + "epoch": 2.71339414306289, + "grad_norm": 0.8339644668159699, + "learning_rate": 5.132307981411301e-06, + "loss": 0.5605, + "step": 2827 + }, + { + "epoch": 2.71435429668747, + "grad_norm": 0.9584927474740805, + "learning_rate": 5.128959164842821e-06, + "loss": 0.6624, + "step": 2828 + }, + { + "epoch": 2.71531445031205, + "grad_norm": 0.9832834120811093, + "learning_rate": 5.125610290385948e-06, + "loss": 0.5626, + "step": 2829 + }, + { + "epoch": 2.71627460393663, + "grad_norm": 1.4531130506929584, + "learning_rate": 5.1222613595439605e-06, + "loss": 0.5188, + "step": 2830 + }, + { + "epoch": 2.7172347575612097, + "grad_norm": 1.3354914155952573, + "learning_rate": 5.118912373820154e-06, + "loss": 0.5278, + "step": 2831 + }, + { + "epoch": 2.7181949111857895, + "grad_norm": 1.5115030214565903, + "learning_rate": 5.1155633347178545e-06, + "loss": 0.6736, + "step": 2832 + }, + { + "epoch": 2.7191550648103697, + "grad_norm": 1.108017215305917, + "learning_rate": 5.1122142437404085e-06, + "loss": 0.5684, + "step": 2833 + }, + { + "epoch": 2.7201152184349495, + "grad_norm": 0.8892713533041173, + "learning_rate": 5.1088651023911905e-06, + "loss": 0.441, + "step": 2834 + }, + { + "epoch": 2.7210753720595298, + "grad_norm": 1.1567874326646286, + "learning_rate": 5.10551591217359e-06, + "loss": 0.5614, + "step": 2835 + }, + { + "epoch": 2.7220355256841096, + "grad_norm": 1.2315998773842656, + "learning_rate": 5.102166674591024e-06, + "loss": 0.5762, + "step": 2836 + }, + { + "epoch": 2.7229956793086894, + "grad_norm": 1.0734260610627955, + "learning_rate": 5.09881739114693e-06, + "loss": 0.4065, + "step": 2837 + }, + { + "epoch": 2.723955832933269, + "grad_norm": 1.1504831323701001, + "learning_rate": 5.095468063344767e-06, + "loss": 0.4678, + "step": 2838 + }, + { + "epoch": 2.7249159865578494, + "grad_norm": 1.339344806209807, + "learning_rate": 5.09211869268801e-06, + "loss": 0.5219, + "step": 2839 + }, + { + "epoch": 2.725876140182429, + "grad_norm": 1.1804555966763617, + "learning_rate": 5.088769280680156e-06, + "loss": 0.4341, + "step": 2840 + }, + { + "epoch": 2.726836293807009, + "grad_norm": 1.296624678563868, + "learning_rate": 5.08541982882472e-06, + "loss": 0.6106, + "step": 2841 + }, + { + "epoch": 2.727796447431589, + "grad_norm": 1.2549957277373123, + "learning_rate": 5.082070338625238e-06, + "loss": 0.5008, + "step": 2842 + }, + { + "epoch": 2.728756601056169, + "grad_norm": 0.9794688150434735, + "learning_rate": 5.078720811585257e-06, + "loss": 0.5464, + "step": 2843 + }, + { + "epoch": 2.729716754680749, + "grad_norm": 0.7783284828697912, + "learning_rate": 5.0753712492083455e-06, + "loss": 0.5024, + "step": 2844 + }, + { + "epoch": 2.7306769083053286, + "grad_norm": 1.3035420271573768, + "learning_rate": 5.072021652998086e-06, + "loss": 0.6407, + "step": 2845 + }, + { + "epoch": 2.731637061929909, + "grad_norm": 1.0125724172958173, + "learning_rate": 5.068672024458076e-06, + "loss": 0.5838, + "step": 2846 + }, + { + "epoch": 2.7325972155544886, + "grad_norm": 1.2426701036049863, + "learning_rate": 5.065322365091928e-06, + "loss": 0.6571, + "step": 2847 + }, + { + "epoch": 2.733557369179069, + "grad_norm": 1.3012295328835128, + "learning_rate": 5.061972676403269e-06, + "loss": 0.4668, + "step": 2848 + }, + { + "epoch": 2.7345175228036487, + "grad_norm": 0.8865773452230334, + "learning_rate": 5.058622959895738e-06, + "loss": 0.6179, + "step": 2849 + }, + { + "epoch": 2.7354776764282285, + "grad_norm": 1.0885644191112207, + "learning_rate": 5.055273217072988e-06, + "loss": 0.4369, + "step": 2850 + }, + { + "epoch": 2.7364378300528083, + "grad_norm": 0.9165856036911711, + "learning_rate": 5.051923449438679e-06, + "loss": 0.5733, + "step": 2851 + }, + { + "epoch": 2.7373979836773885, + "grad_norm": 0.9540132292687737, + "learning_rate": 5.048573658496492e-06, + "loss": 0.5949, + "step": 2852 + }, + { + "epoch": 2.7383581373019683, + "grad_norm": 0.8366834093479427, + "learning_rate": 5.0452238457501065e-06, + "loss": 0.4407, + "step": 2853 + }, + { + "epoch": 2.739318290926548, + "grad_norm": 0.8971465876763657, + "learning_rate": 5.0418740127032204e-06, + "loss": 0.5203, + "step": 2854 + }, + { + "epoch": 2.7402784445511283, + "grad_norm": 1.3695778579771662, + "learning_rate": 5.038524160859538e-06, + "loss": 0.5721, + "step": 2855 + }, + { + "epoch": 2.741238598175708, + "grad_norm": 1.3018734808593488, + "learning_rate": 5.035174291722772e-06, + "loss": 0.5383, + "step": 2856 + }, + { + "epoch": 2.742198751800288, + "grad_norm": 0.9550679487013927, + "learning_rate": 5.031824406796642e-06, + "loss": 0.5882, + "step": 2857 + }, + { + "epoch": 2.7431589054248677, + "grad_norm": 0.8927485043459621, + "learning_rate": 5.028474507584875e-06, + "loss": 0.6261, + "step": 2858 + }, + { + "epoch": 2.744119059049448, + "grad_norm": 0.7893565117198948, + "learning_rate": 5.025124595591206e-06, + "loss": 0.5053, + "step": 2859 + }, + { + "epoch": 2.7450792126740278, + "grad_norm": 1.0459122033902766, + "learning_rate": 5.0217746723193775e-06, + "loss": 0.5781, + "step": 2860 + }, + { + "epoch": 2.746039366298608, + "grad_norm": 1.0907240317149534, + "learning_rate": 5.018424739273129e-06, + "loss": 0.6002, + "step": 2861 + }, + { + "epoch": 2.746999519923188, + "grad_norm": 1.0419964460467377, + "learning_rate": 5.015074797956212e-06, + "loss": 0.4895, + "step": 2862 + }, + { + "epoch": 2.7479596735477676, + "grad_norm": 1.216878499598488, + "learning_rate": 5.01172484987238e-06, + "loss": 0.454, + "step": 2863 + }, + { + "epoch": 2.7489198271723474, + "grad_norm": 1.0576687649963499, + "learning_rate": 5.008374896525387e-06, + "loss": 0.509, + "step": 2864 + }, + { + "epoch": 2.7498799807969276, + "grad_norm": 1.8347390514096837, + "learning_rate": 5.0050249394189945e-06, + "loss": 0.6098, + "step": 2865 + }, + { + "epoch": 2.7508401344215074, + "grad_norm": 1.405774031281213, + "learning_rate": 5.001674980056959e-06, + "loss": 0.5591, + "step": 2866 + }, + { + "epoch": 2.7518002880460872, + "grad_norm": 1.1151900662766834, + "learning_rate": 4.998325019943044e-06, + "loss": 0.6112, + "step": 2867 + }, + { + "epoch": 2.7527604416706675, + "grad_norm": 0.9580892640928749, + "learning_rate": 4.994975060581007e-06, + "loss": 0.5234, + "step": 2868 + }, + { + "epoch": 2.7537205952952473, + "grad_norm": 0.9864450350398929, + "learning_rate": 4.991625103474614e-06, + "loss": 0.6642, + "step": 2869 + }, + { + "epoch": 2.754680748919827, + "grad_norm": 1.3753482467077893, + "learning_rate": 4.988275150127622e-06, + "loss": 0.6233, + "step": 2870 + }, + { + "epoch": 2.755640902544407, + "grad_norm": 1.0859931168811305, + "learning_rate": 4.984925202043789e-06, + "loss": 0.521, + "step": 2871 + }, + { + "epoch": 2.756601056168987, + "grad_norm": 1.0229236000477513, + "learning_rate": 4.981575260726873e-06, + "loss": 0.5044, + "step": 2872 + }, + { + "epoch": 2.757561209793567, + "grad_norm": 0.7735495482332387, + "learning_rate": 4.978225327680625e-06, + "loss": 0.4529, + "step": 2873 + }, + { + "epoch": 2.758521363418147, + "grad_norm": 0.9809142385638872, + "learning_rate": 4.974875404408794e-06, + "loss": 0.6003, + "step": 2874 + }, + { + "epoch": 2.759481517042727, + "grad_norm": 0.9247342259426234, + "learning_rate": 4.971525492415126e-06, + "loss": 0.5366, + "step": 2875 + }, + { + "epoch": 2.7604416706673067, + "grad_norm": 1.9299835428669425, + "learning_rate": 4.96817559320336e-06, + "loss": 0.4561, + "step": 2876 + }, + { + "epoch": 2.7614018242918865, + "grad_norm": 0.8949534381152232, + "learning_rate": 4.9648257082772305e-06, + "loss": 0.5095, + "step": 2877 + }, + { + "epoch": 2.7623619779164668, + "grad_norm": 1.1198384601344433, + "learning_rate": 4.9614758391404634e-06, + "loss": 0.5142, + "step": 2878 + }, + { + "epoch": 2.7633221315410466, + "grad_norm": 0.9360147241694083, + "learning_rate": 4.9581259872967795e-06, + "loss": 0.6007, + "step": 2879 + }, + { + "epoch": 2.7642822851656264, + "grad_norm": 0.9027492001719692, + "learning_rate": 4.954776154249896e-06, + "loss": 0.5609, + "step": 2880 + }, + { + "epoch": 2.7652424387902066, + "grad_norm": 1.267557537425477, + "learning_rate": 4.95142634150351e-06, + "loss": 0.6081, + "step": 2881 + }, + { + "epoch": 2.7662025924147864, + "grad_norm": 0.902476589047353, + "learning_rate": 4.94807655056132e-06, + "loss": 0.4766, + "step": 2882 + }, + { + "epoch": 2.767162746039366, + "grad_norm": 1.136589337895067, + "learning_rate": 4.944726782927014e-06, + "loss": 0.4529, + "step": 2883 + }, + { + "epoch": 2.768122899663946, + "grad_norm": 1.327656014793884, + "learning_rate": 4.9413770401042625e-06, + "loss": 0.6217, + "step": 2884 + }, + { + "epoch": 2.769083053288526, + "grad_norm": 1.5131516110491452, + "learning_rate": 4.938027323596732e-06, + "loss": 0.5914, + "step": 2885 + }, + { + "epoch": 2.770043206913106, + "grad_norm": 1.1522831309033787, + "learning_rate": 4.934677634908073e-06, + "loss": 0.5679, + "step": 2886 + }, + { + "epoch": 2.7710033605376863, + "grad_norm": 1.4848204252131276, + "learning_rate": 4.931327975541925e-06, + "loss": 0.4961, + "step": 2887 + }, + { + "epoch": 2.771963514162266, + "grad_norm": 1.0322928053227625, + "learning_rate": 4.927978347001916e-06, + "loss": 0.5455, + "step": 2888 + }, + { + "epoch": 2.772923667786846, + "grad_norm": 0.9299388285419378, + "learning_rate": 4.924628750791656e-06, + "loss": 0.5697, + "step": 2889 + }, + { + "epoch": 2.7738838214114256, + "grad_norm": 1.3715596322966592, + "learning_rate": 4.9212791884147456e-06, + "loss": 0.6172, + "step": 2890 + }, + { + "epoch": 2.774843975036006, + "grad_norm": 1.021732103832636, + "learning_rate": 4.917929661374764e-06, + "loss": 0.5135, + "step": 2891 + }, + { + "epoch": 2.7758041286605857, + "grad_norm": 1.2142376265131427, + "learning_rate": 4.9145801711752804e-06, + "loss": 0.5315, + "step": 2892 + }, + { + "epoch": 2.7767642822851655, + "grad_norm": 0.982717420233173, + "learning_rate": 4.911230719319846e-06, + "loss": 0.5403, + "step": 2893 + }, + { + "epoch": 2.7777244359097457, + "grad_norm": 1.1441271424914692, + "learning_rate": 4.907881307311992e-06, + "loss": 0.5888, + "step": 2894 + }, + { + "epoch": 2.7786845895343255, + "grad_norm": 1.109638211799611, + "learning_rate": 4.904531936655236e-06, + "loss": 0.513, + "step": 2895 + }, + { + "epoch": 2.7796447431589053, + "grad_norm": 1.1796956457561363, + "learning_rate": 4.901182608853071e-06, + "loss": 0.5994, + "step": 2896 + }, + { + "epoch": 2.780604896783485, + "grad_norm": 0.9698122948469821, + "learning_rate": 4.897833325408977e-06, + "loss": 0.4911, + "step": 2897 + }, + { + "epoch": 2.7815650504080653, + "grad_norm": 1.1345128740035557, + "learning_rate": 4.894484087826413e-06, + "loss": 0.5344, + "step": 2898 + }, + { + "epoch": 2.782525204032645, + "grad_norm": 1.1458155224917117, + "learning_rate": 4.891134897608811e-06, + "loss": 0.5364, + "step": 2899 + }, + { + "epoch": 2.7834853576572254, + "grad_norm": 0.9397418319561159, + "learning_rate": 4.887785756259591e-06, + "loss": 0.5709, + "step": 2900 + }, + { + "epoch": 2.784445511281805, + "grad_norm": 1.0170311583518241, + "learning_rate": 4.884436665282146e-06, + "loss": 0.5443, + "step": 2901 + }, + { + "epoch": 2.785405664906385, + "grad_norm": 1.2170631253268918, + "learning_rate": 4.881087626179847e-06, + "loss": 0.6374, + "step": 2902 + }, + { + "epoch": 2.7863658185309648, + "grad_norm": 1.5775108093947912, + "learning_rate": 4.877738640456042e-06, + "loss": 0.6124, + "step": 2903 + }, + { + "epoch": 2.787325972155545, + "grad_norm": 1.3087798900178886, + "learning_rate": 4.874389709614053e-06, + "loss": 0.5946, + "step": 2904 + }, + { + "epoch": 2.788286125780125, + "grad_norm": 0.9347091809549669, + "learning_rate": 4.871040835157181e-06, + "loss": 0.6385, + "step": 2905 + }, + { + "epoch": 2.7892462794047046, + "grad_norm": 0.7220209203380794, + "learning_rate": 4.8676920185887e-06, + "loss": 0.508, + "step": 2906 + }, + { + "epoch": 2.790206433029285, + "grad_norm": 0.9498971025305757, + "learning_rate": 4.8643432614118565e-06, + "loss": 0.4853, + "step": 2907 + }, + { + "epoch": 2.7911665866538646, + "grad_norm": 1.1940499656509447, + "learning_rate": 4.860994565129875e-06, + "loss": 0.6684, + "step": 2908 + }, + { + "epoch": 2.7921267402784444, + "grad_norm": 0.9148733237251847, + "learning_rate": 4.857645931245946e-06, + "loss": 0.5631, + "step": 2909 + }, + { + "epoch": 2.7930868939030242, + "grad_norm": 1.0164189280076747, + "learning_rate": 4.854297361263235e-06, + "loss": 0.5831, + "step": 2910 + }, + { + "epoch": 2.7940470475276045, + "grad_norm": 1.028809477737421, + "learning_rate": 4.850948856684884e-06, + "loss": 0.6655, + "step": 2911 + }, + { + "epoch": 2.7950072011521843, + "grad_norm": 1.3339558334636454, + "learning_rate": 4.8476004190139954e-06, + "loss": 0.5883, + "step": 2912 + }, + { + "epoch": 2.7959673547767645, + "grad_norm": 1.216316236366803, + "learning_rate": 4.844252049753651e-06, + "loss": 0.5467, + "step": 2913 + }, + { + "epoch": 2.7969275084013443, + "grad_norm": 1.0088969301176822, + "learning_rate": 4.840903750406898e-06, + "loss": 0.559, + "step": 2914 + }, + { + "epoch": 2.797887662025924, + "grad_norm": 1.0206529363082588, + "learning_rate": 4.8375555224767486e-06, + "loss": 0.5089, + "step": 2915 + }, + { + "epoch": 2.798847815650504, + "grad_norm": 1.4111941947397157, + "learning_rate": 4.83420736746619e-06, + "loss": 0.5706, + "step": 2916 + }, + { + "epoch": 2.799807969275084, + "grad_norm": 0.8486448778179049, + "learning_rate": 4.830859286878172e-06, + "loss": 0.4695, + "step": 2917 + }, + { + "epoch": 2.800768122899664, + "grad_norm": 1.2717958003470367, + "learning_rate": 4.8275112822156105e-06, + "loss": 0.4842, + "step": 2918 + }, + { + "epoch": 2.8017282765242437, + "grad_norm": 1.2693866916210295, + "learning_rate": 4.824163354981393e-06, + "loss": 0.6731, + "step": 2919 + }, + { + "epoch": 2.802688430148824, + "grad_norm": 1.639470803075458, + "learning_rate": 4.820815506678364e-06, + "loss": 0.5737, + "step": 2920 + }, + { + "epoch": 2.8036485837734038, + "grad_norm": 0.9229804157692403, + "learning_rate": 4.81746773880934e-06, + "loss": 0.576, + "step": 2921 + }, + { + "epoch": 2.8046087373979836, + "grad_norm": 0.9251750930089446, + "learning_rate": 4.8141200528770955e-06, + "loss": 0.6212, + "step": 2922 + }, + { + "epoch": 2.8055688910225633, + "grad_norm": 1.1494616993252966, + "learning_rate": 4.810772450384372e-06, + "loss": 0.4777, + "step": 2923 + }, + { + "epoch": 2.8065290446471436, + "grad_norm": 0.8265258950149946, + "learning_rate": 4.807424932833875e-06, + "loss": 0.6193, + "step": 2924 + }, + { + "epoch": 2.8074891982717234, + "grad_norm": 0.9044863619052229, + "learning_rate": 4.8040775017282644e-06, + "loss": 0.5613, + "step": 2925 + }, + { + "epoch": 2.8084493518963036, + "grad_norm": 1.158329265336177, + "learning_rate": 4.8007301585701726e-06, + "loss": 0.4908, + "step": 2926 + }, + { + "epoch": 2.8094095055208834, + "grad_norm": 1.0484283022409993, + "learning_rate": 4.7973829048621786e-06, + "loss": 0.6873, + "step": 2927 + }, + { + "epoch": 2.810369659145463, + "grad_norm": 1.0352659403810793, + "learning_rate": 4.794035742106834e-06, + "loss": 0.5028, + "step": 2928 + }, + { + "epoch": 2.811329812770043, + "grad_norm": 0.8197048707055922, + "learning_rate": 4.7906886718066445e-06, + "loss": 0.5401, + "step": 2929 + }, + { + "epoch": 2.8122899663946233, + "grad_norm": 1.5350399626221585, + "learning_rate": 4.7873416954640725e-06, + "loss": 0.5613, + "step": 2930 + }, + { + "epoch": 2.813250120019203, + "grad_norm": 0.887983419444355, + "learning_rate": 4.783994814581539e-06, + "loss": 0.54, + "step": 2931 + }, + { + "epoch": 2.814210273643783, + "grad_norm": 1.0429987583112454, + "learning_rate": 4.7806480306614256e-06, + "loss": 0.5531, + "step": 2932 + }, + { + "epoch": 2.815170427268363, + "grad_norm": 1.1935995907940316, + "learning_rate": 4.7773013452060654e-06, + "loss": 0.539, + "step": 2933 + }, + { + "epoch": 2.816130580892943, + "grad_norm": 1.2647882862564, + "learning_rate": 4.773954759717752e-06, + "loss": 0.5965, + "step": 2934 + }, + { + "epoch": 2.8170907345175227, + "grad_norm": 0.8387385243105556, + "learning_rate": 4.7706082756987295e-06, + "loss": 0.6562, + "step": 2935 + }, + { + "epoch": 2.8180508881421025, + "grad_norm": 1.2832883921377154, + "learning_rate": 4.7672618946512e-06, + "loss": 0.502, + "step": 2936 + }, + { + "epoch": 2.8190110417666827, + "grad_norm": 1.092179843546104, + "learning_rate": 4.76391561807732e-06, + "loss": 0.6711, + "step": 2937 + }, + { + "epoch": 2.8199711953912625, + "grad_norm": 1.0055939141975967, + "learning_rate": 4.760569447479193e-06, + "loss": 0.4059, + "step": 2938 + }, + { + "epoch": 2.8209313490158427, + "grad_norm": 0.9786410157357517, + "learning_rate": 4.757223384358884e-06, + "loss": 0.5426, + "step": 2939 + }, + { + "epoch": 2.8218915026404225, + "grad_norm": 1.0152212772507851, + "learning_rate": 4.753877430218401e-06, + "loss": 0.4667, + "step": 2940 + }, + { + "epoch": 2.8228516562650023, + "grad_norm": 1.0236248565204924, + "learning_rate": 4.750531586559707e-06, + "loss": 0.5683, + "step": 2941 + }, + { + "epoch": 2.823811809889582, + "grad_norm": 0.9447469720712466, + "learning_rate": 4.7471858548847214e-06, + "loss": 0.547, + "step": 2942 + }, + { + "epoch": 2.8247719635141624, + "grad_norm": 0.9256791194856672, + "learning_rate": 4.743840236695299e-06, + "loss": 0.4015, + "step": 2943 + }, + { + "epoch": 2.825732117138742, + "grad_norm": 1.0494974554288614, + "learning_rate": 4.740494733493258e-06, + "loss": 0.4664, + "step": 2944 + }, + { + "epoch": 2.826692270763322, + "grad_norm": 1.9042579098918375, + "learning_rate": 4.737149346780358e-06, + "loss": 0.6442, + "step": 2945 + }, + { + "epoch": 2.827652424387902, + "grad_norm": 1.2222045147265035, + "learning_rate": 4.733804078058305e-06, + "loss": 0.5163, + "step": 2946 + }, + { + "epoch": 2.828612578012482, + "grad_norm": 1.0208873395442013, + "learning_rate": 4.730458928828757e-06, + "loss": 0.4993, + "step": 2947 + }, + { + "epoch": 2.829572731637062, + "grad_norm": 1.0085328539903105, + "learning_rate": 4.7271139005933144e-06, + "loss": 0.511, + "step": 2948 + }, + { + "epoch": 2.8305328852616416, + "grad_norm": 0.9530941990881852, + "learning_rate": 4.723768994853523e-06, + "loss": 0.5053, + "step": 2949 + }, + { + "epoch": 2.831493038886222, + "grad_norm": 1.0544682630381126, + "learning_rate": 4.72042421311088e-06, + "loss": 0.6032, + "step": 2950 + }, + { + "epoch": 2.8324531925108016, + "grad_norm": 2.039841179674737, + "learning_rate": 4.717079556866818e-06, + "loss": 0.6091, + "step": 2951 + }, + { + "epoch": 2.833413346135382, + "grad_norm": 0.7630898576165969, + "learning_rate": 4.713735027622719e-06, + "loss": 0.4976, + "step": 2952 + }, + { + "epoch": 2.8343734997599617, + "grad_norm": 1.1800668687363773, + "learning_rate": 4.710390626879906e-06, + "loss": 0.4927, + "step": 2953 + }, + { + "epoch": 2.8353336533845415, + "grad_norm": 1.1137877328928956, + "learning_rate": 4.707046356139643e-06, + "loss": 0.5815, + "step": 2954 + }, + { + "epoch": 2.8362938070091213, + "grad_norm": 1.324999776891313, + "learning_rate": 4.703702216903142e-06, + "loss": 0.5604, + "step": 2955 + }, + { + "epoch": 2.8372539606337015, + "grad_norm": 1.2021393651999435, + "learning_rate": 4.700358210671545e-06, + "loss": 0.5078, + "step": 2956 + }, + { + "epoch": 2.8382141142582813, + "grad_norm": 0.9841798111047575, + "learning_rate": 4.697014338945947e-06, + "loss": 0.6838, + "step": 2957 + }, + { + "epoch": 2.839174267882861, + "grad_norm": 0.8593640999692991, + "learning_rate": 4.693670603227374e-06, + "loss": 0.5313, + "step": 2958 + }, + { + "epoch": 2.8401344215074413, + "grad_norm": 0.839383051904467, + "learning_rate": 4.690327005016792e-06, + "loss": 0.5929, + "step": 2959 + }, + { + "epoch": 2.841094575132021, + "grad_norm": 1.249979030318789, + "learning_rate": 4.686983545815107e-06, + "loss": 0.5601, + "step": 2960 + }, + { + "epoch": 2.842054728756601, + "grad_norm": 1.1991651360082194, + "learning_rate": 4.683640227123162e-06, + "loss": 0.5548, + "step": 2961 + }, + { + "epoch": 2.8430148823811807, + "grad_norm": 0.9230837658500113, + "learning_rate": 4.680297050441737e-06, + "loss": 0.5189, + "step": 2962 + }, + { + "epoch": 2.843975036005761, + "grad_norm": 1.0786500524276668, + "learning_rate": 4.67695401727155e-06, + "loss": 0.546, + "step": 2963 + }, + { + "epoch": 2.8449351896303408, + "grad_norm": 1.0300845286097802, + "learning_rate": 4.6736111291132505e-06, + "loss": 0.5902, + "step": 2964 + }, + { + "epoch": 2.845895343254921, + "grad_norm": 0.9355354955667121, + "learning_rate": 4.670268387467427e-06, + "loss": 0.458, + "step": 2965 + }, + { + "epoch": 2.846855496879501, + "grad_norm": 1.117434615366541, + "learning_rate": 4.666925793834599e-06, + "loss": 0.5796, + "step": 2966 + }, + { + "epoch": 2.8478156505040806, + "grad_norm": 0.9974172974227485, + "learning_rate": 4.663583349715222e-06, + "loss": 0.6005, + "step": 2967 + }, + { + "epoch": 2.8487758041286604, + "grad_norm": 1.112334304697473, + "learning_rate": 4.660241056609685e-06, + "loss": 0.4449, + "step": 2968 + }, + { + "epoch": 2.8497359577532406, + "grad_norm": 0.8208724186800234, + "learning_rate": 4.656898916018302e-06, + "loss": 0.5611, + "step": 2969 + }, + { + "epoch": 2.8506961113778204, + "grad_norm": 1.0722641318898372, + "learning_rate": 4.653556929441332e-06, + "loss": 0.6135, + "step": 2970 + }, + { + "epoch": 2.8516562650024007, + "grad_norm": 1.0965869899942944, + "learning_rate": 4.650215098378949e-06, + "loss": 0.609, + "step": 2971 + }, + { + "epoch": 2.8526164186269805, + "grad_norm": 1.3223276111201334, + "learning_rate": 4.646873424331272e-06, + "loss": 0.4943, + "step": 2972 + }, + { + "epoch": 2.8535765722515603, + "grad_norm": 0.8175684508979174, + "learning_rate": 4.64353190879834e-06, + "loss": 0.4866, + "step": 2973 + }, + { + "epoch": 2.85453672587614, + "grad_norm": 1.9024552987874916, + "learning_rate": 4.640190553280123e-06, + "loss": 0.5274, + "step": 2974 + }, + { + "epoch": 2.85549687950072, + "grad_norm": 1.5474165271644353, + "learning_rate": 4.636849359276521e-06, + "loss": 0.6289, + "step": 2975 + }, + { + "epoch": 2.8564570331253, + "grad_norm": 1.3727133407513454, + "learning_rate": 4.6335083282873605e-06, + "loss": 0.4485, + "step": 2976 + }, + { + "epoch": 2.85741718674988, + "grad_norm": 1.2338334468601542, + "learning_rate": 4.6301674618123924e-06, + "loss": 0.577, + "step": 2977 + }, + { + "epoch": 2.85837734037446, + "grad_norm": 1.1130501420510492, + "learning_rate": 4.626826761351299e-06, + "loss": 0.4546, + "step": 2978 + }, + { + "epoch": 2.85933749399904, + "grad_norm": 0.7632792686659663, + "learning_rate": 4.623486228403682e-06, + "loss": 0.4242, + "step": 2979 + }, + { + "epoch": 2.8602976476236197, + "grad_norm": 0.8560025752215499, + "learning_rate": 4.620145864469073e-06, + "loss": 0.4676, + "step": 2980 + }, + { + "epoch": 2.8612578012481995, + "grad_norm": 0.8304956239781156, + "learning_rate": 4.616805671046925e-06, + "loss": 0.5169, + "step": 2981 + }, + { + "epoch": 2.8622179548727797, + "grad_norm": 0.8369784073571794, + "learning_rate": 4.613465649636615e-06, + "loss": 0.6583, + "step": 2982 + }, + { + "epoch": 2.8631781084973595, + "grad_norm": 0.9018829233854131, + "learning_rate": 4.610125801737443e-06, + "loss": 0.4335, + "step": 2983 + }, + { + "epoch": 2.86413826212194, + "grad_norm": 1.077563234475322, + "learning_rate": 4.6067861288486315e-06, + "loss": 0.5521, + "step": 2984 + }, + { + "epoch": 2.8650984157465196, + "grad_norm": 0.9451248894082102, + "learning_rate": 4.603446632469322e-06, + "loss": 0.4489, + "step": 2985 + }, + { + "epoch": 2.8660585693710994, + "grad_norm": 0.8653059661890927, + "learning_rate": 4.600107314098583e-06, + "loss": 0.5393, + "step": 2986 + }, + { + "epoch": 2.867018722995679, + "grad_norm": 1.0572417546328456, + "learning_rate": 4.596768175235393e-06, + "loss": 0.5382, + "step": 2987 + }, + { + "epoch": 2.867978876620259, + "grad_norm": 1.019972356032487, + "learning_rate": 4.59342921737866e-06, + "loss": 0.5814, + "step": 2988 + }, + { + "epoch": 2.868939030244839, + "grad_norm": 1.3033043712718195, + "learning_rate": 4.590090442027206e-06, + "loss": 0.5162, + "step": 2989 + }, + { + "epoch": 2.869899183869419, + "grad_norm": 0.8925748328223191, + "learning_rate": 4.5867518506797705e-06, + "loss": 0.5998, + "step": 2990 + }, + { + "epoch": 2.8708593374939992, + "grad_norm": 1.3113807118490572, + "learning_rate": 4.583413444835013e-06, + "loss": 0.5451, + "step": 2991 + }, + { + "epoch": 2.871819491118579, + "grad_norm": 1.2727824784003843, + "learning_rate": 4.580075225991505e-06, + "loss": 0.5719, + "step": 2992 + }, + { + "epoch": 2.872779644743159, + "grad_norm": 1.3164380952076393, + "learning_rate": 4.576737195647741e-06, + "loss": 0.5968, + "step": 2993 + }, + { + "epoch": 2.8737397983677386, + "grad_norm": 0.9221847639444891, + "learning_rate": 4.573399355302124e-06, + "loss": 0.663, + "step": 2994 + }, + { + "epoch": 2.874699951992319, + "grad_norm": 0.916116575610588, + "learning_rate": 4.5700617064529774e-06, + "loss": 0.5737, + "step": 2995 + }, + { + "epoch": 2.8756601056168987, + "grad_norm": 1.4319594670206892, + "learning_rate": 4.566724250598535e-06, + "loss": 0.6928, + "step": 2996 + }, + { + "epoch": 2.876620259241479, + "grad_norm": 1.191313441910245, + "learning_rate": 4.563386989236944e-06, + "loss": 0.4899, + "step": 2997 + }, + { + "epoch": 2.8775804128660587, + "grad_norm": 1.0149931134472947, + "learning_rate": 4.560049923866266e-06, + "loss": 0.5079, + "step": 2998 + }, + { + "epoch": 2.8785405664906385, + "grad_norm": 1.5373731386660563, + "learning_rate": 4.556713055984477e-06, + "loss": 0.432, + "step": 2999 + }, + { + "epoch": 2.8795007201152183, + "grad_norm": 1.7074343779552237, + "learning_rate": 4.5533763870894555e-06, + "loss": 0.6384, + "step": 3000 + }, + { + "epoch": 2.8804608737397985, + "grad_norm": 0.8375219710394645, + "learning_rate": 4.5500399186790034e-06, + "loss": 0.6383, + "step": 3001 + }, + { + "epoch": 2.8814210273643783, + "grad_norm": 1.0632428704884742, + "learning_rate": 4.5467036522508185e-06, + "loss": 0.5493, + "step": 3002 + }, + { + "epoch": 2.882381180988958, + "grad_norm": 1.207052947052255, + "learning_rate": 4.54336758930252e-06, + "loss": 0.5615, + "step": 3003 + }, + { + "epoch": 2.8833413346135384, + "grad_norm": 0.7678019396949012, + "learning_rate": 4.54003173133163e-06, + "loss": 0.4637, + "step": 3004 + }, + { + "epoch": 2.884301488238118, + "grad_norm": 1.1681088860376656, + "learning_rate": 4.536696079835578e-06, + "loss": 0.6318, + "step": 3005 + }, + { + "epoch": 2.885261641862698, + "grad_norm": 1.23080410397212, + "learning_rate": 4.5333606363117035e-06, + "loss": 0.5864, + "step": 3006 + }, + { + "epoch": 2.8862217954872778, + "grad_norm": 1.0049753863707054, + "learning_rate": 4.530025402257252e-06, + "loss": 0.5366, + "step": 3007 + }, + { + "epoch": 2.887181949111858, + "grad_norm": 1.1876107138311136, + "learning_rate": 4.526690379169374e-06, + "loss": 0.5576, + "step": 3008 + }, + { + "epoch": 2.888142102736438, + "grad_norm": 0.9531136818959197, + "learning_rate": 4.5233555685451245e-06, + "loss": 0.4856, + "step": 3009 + }, + { + "epoch": 2.889102256361018, + "grad_norm": 1.2586919211655028, + "learning_rate": 4.5200209718814656e-06, + "loss": 0.5349, + "step": 3010 + }, + { + "epoch": 2.890062409985598, + "grad_norm": 1.87157955692469, + "learning_rate": 4.516686590675261e-06, + "loss": 0.636, + "step": 3011 + }, + { + "epoch": 2.8910225636101776, + "grad_norm": 1.1001353549011164, + "learning_rate": 4.513352426423281e-06, + "loss": 0.5301, + "step": 3012 + }, + { + "epoch": 2.8919827172347574, + "grad_norm": 0.8672854790851329, + "learning_rate": 4.510018480622193e-06, + "loss": 0.479, + "step": 3013 + }, + { + "epoch": 2.8929428708593377, + "grad_norm": 1.069144089238986, + "learning_rate": 4.506684754768573e-06, + "loss": 0.6274, + "step": 3014 + }, + { + "epoch": 2.8939030244839175, + "grad_norm": 1.0047158763513129, + "learning_rate": 4.503351250358893e-06, + "loss": 0.594, + "step": 3015 + }, + { + "epoch": 2.8948631781084972, + "grad_norm": 0.9946312004958189, + "learning_rate": 4.500017968889525e-06, + "loss": 0.4962, + "step": 3016 + }, + { + "epoch": 2.8958233317330775, + "grad_norm": 1.3986911433989682, + "learning_rate": 4.4966849118567515e-06, + "loss": 0.5117, + "step": 3017 + }, + { + "epoch": 2.8967834853576573, + "grad_norm": 0.7769163481046195, + "learning_rate": 4.4933520807567365e-06, + "loss": 0.5894, + "step": 3018 + }, + { + "epoch": 2.897743638982237, + "grad_norm": 1.0431629023014957, + "learning_rate": 4.490019477085558e-06, + "loss": 0.4432, + "step": 3019 + }, + { + "epoch": 2.898703792606817, + "grad_norm": 1.0246612106470556, + "learning_rate": 4.4866871023391865e-06, + "loss": 0.5405, + "step": 3020 + }, + { + "epoch": 2.899663946231397, + "grad_norm": 1.0507082713031515, + "learning_rate": 4.483354958013485e-06, + "loss": 0.5078, + "step": 3021 + }, + { + "epoch": 2.900624099855977, + "grad_norm": 1.8483135265392954, + "learning_rate": 4.480023045604223e-06, + "loss": 0.5474, + "step": 3022 + }, + { + "epoch": 2.901584253480557, + "grad_norm": 0.8890285573880536, + "learning_rate": 4.476691366607056e-06, + "loss": 0.5069, + "step": 3023 + }, + { + "epoch": 2.902544407105137, + "grad_norm": 1.4544176744902257, + "learning_rate": 4.473359922517541e-06, + "loss": 0.6014, + "step": 3024 + }, + { + "epoch": 2.9035045607297167, + "grad_norm": 0.9310882341777574, + "learning_rate": 4.4700287148311275e-06, + "loss": 0.5207, + "step": 3025 + }, + { + "epoch": 2.9044647143542965, + "grad_norm": 0.974213768892102, + "learning_rate": 4.466697745043158e-06, + "loss": 0.544, + "step": 3026 + }, + { + "epoch": 2.905424867978877, + "grad_norm": 1.4873506603586337, + "learning_rate": 4.463367014648872e-06, + "loss": 0.6402, + "step": 3027 + }, + { + "epoch": 2.9063850216034566, + "grad_norm": 1.134407923930257, + "learning_rate": 4.460036525143395e-06, + "loss": 0.6281, + "step": 3028 + }, + { + "epoch": 2.9073451752280364, + "grad_norm": 1.0120977178590904, + "learning_rate": 4.4567062780217486e-06, + "loss": 0.4844, + "step": 3029 + }, + { + "epoch": 2.9083053288526166, + "grad_norm": 1.2791076500871876, + "learning_rate": 4.45337627477885e-06, + "loss": 0.4925, + "step": 3030 + }, + { + "epoch": 2.9092654824771964, + "grad_norm": 1.0668621725583047, + "learning_rate": 4.450046516909496e-06, + "loss": 0.5593, + "step": 3031 + }, + { + "epoch": 2.910225636101776, + "grad_norm": 0.9021891829977514, + "learning_rate": 4.446717005908382e-06, + "loss": 0.5559, + "step": 3032 + }, + { + "epoch": 2.911185789726356, + "grad_norm": 1.0815019586591574, + "learning_rate": 4.4433877432700915e-06, + "loss": 0.5933, + "step": 3033 + }, + { + "epoch": 2.9121459433509362, + "grad_norm": 0.9222059200044922, + "learning_rate": 4.440058730489092e-06, + "loss": 0.5702, + "step": 3034 + }, + { + "epoch": 2.913106096975516, + "grad_norm": 0.9742529311574974, + "learning_rate": 4.436729969059745e-06, + "loss": 0.4932, + "step": 3035 + }, + { + "epoch": 2.9140662506000963, + "grad_norm": 1.2060995834914687, + "learning_rate": 4.433401460476291e-06, + "loss": 0.5679, + "step": 3036 + }, + { + "epoch": 2.915026404224676, + "grad_norm": 1.3623989808235863, + "learning_rate": 4.430073206232865e-06, + "loss": 0.6511, + "step": 3037 + }, + { + "epoch": 2.915986557849256, + "grad_norm": 0.910803855126238, + "learning_rate": 4.426745207823486e-06, + "loss": 0.4692, + "step": 3038 + }, + { + "epoch": 2.9169467114738357, + "grad_norm": 1.0472135165762646, + "learning_rate": 4.4234174667420546e-06, + "loss": 0.5371, + "step": 3039 + }, + { + "epoch": 2.917906865098416, + "grad_norm": 0.8765347699511459, + "learning_rate": 4.42008998448236e-06, + "loss": 0.5694, + "step": 3040 + }, + { + "epoch": 2.9188670187229957, + "grad_norm": 0.8886301221762489, + "learning_rate": 4.416762762538072e-06, + "loss": 0.562, + "step": 3041 + }, + { + "epoch": 2.9198271723475755, + "grad_norm": 0.9065298103819833, + "learning_rate": 4.413435802402744e-06, + "loss": 0.5103, + "step": 3042 + }, + { + "epoch": 2.9207873259721557, + "grad_norm": 1.051560868451231, + "learning_rate": 4.410109105569817e-06, + "loss": 0.5767, + "step": 3043 + }, + { + "epoch": 2.9217474795967355, + "grad_norm": 1.3605211310629468, + "learning_rate": 4.406782673532604e-06, + "loss": 0.5673, + "step": 3044 + }, + { + "epoch": 2.9227076332213153, + "grad_norm": 0.9631342778600814, + "learning_rate": 4.403456507784311e-06, + "loss": 0.6016, + "step": 3045 + }, + { + "epoch": 2.923667786845895, + "grad_norm": 1.5267605914221638, + "learning_rate": 4.400130609818012e-06, + "loss": 0.6287, + "step": 3046 + }, + { + "epoch": 2.9246279404704754, + "grad_norm": 1.2832581809596535, + "learning_rate": 4.3968049811266705e-06, + "loss": 0.5855, + "step": 3047 + }, + { + "epoch": 2.925588094095055, + "grad_norm": 1.1133356273353832, + "learning_rate": 4.393479623203126e-06, + "loss": 0.5513, + "step": 3048 + }, + { + "epoch": 2.9265482477196354, + "grad_norm": 1.3472110867020108, + "learning_rate": 4.390154537540095e-06, + "loss": 0.5396, + "step": 3049 + }, + { + "epoch": 2.927508401344215, + "grad_norm": 0.9741477875318071, + "learning_rate": 4.386829725630173e-06, + "loss": 0.5727, + "step": 3050 + }, + { + "epoch": 2.928468554968795, + "grad_norm": 1.0691734973697014, + "learning_rate": 4.383505188965835e-06, + "loss": 0.5953, + "step": 3051 + }, + { + "epoch": 2.929428708593375, + "grad_norm": 0.9946935210461184, + "learning_rate": 4.380180929039424e-06, + "loss": 0.6349, + "step": 3052 + }, + { + "epoch": 2.930388862217955, + "grad_norm": 0.9572456779873351, + "learning_rate": 4.376856947343171e-06, + "loss": 0.6243, + "step": 3053 + }, + { + "epoch": 2.931349015842535, + "grad_norm": 1.2367312077386956, + "learning_rate": 4.3735332453691725e-06, + "loss": 0.472, + "step": 3054 + }, + { + "epoch": 2.9323091694671146, + "grad_norm": 1.8085708065543822, + "learning_rate": 4.370209824609403e-06, + "loss": 0.635, + "step": 3055 + }, + { + "epoch": 2.933269323091695, + "grad_norm": 0.9167614826946134, + "learning_rate": 4.366886686555713e-06, + "loss": 0.4998, + "step": 3056 + }, + { + "epoch": 2.9342294767162747, + "grad_norm": 1.1843108961133864, + "learning_rate": 4.36356383269982e-06, + "loss": 0.5833, + "step": 3057 + }, + { + "epoch": 2.9351896303408544, + "grad_norm": 1.1558700054712718, + "learning_rate": 4.360241264533321e-06, + "loss": 0.542, + "step": 3058 + }, + { + "epoch": 2.9361497839654342, + "grad_norm": 0.9249557810139207, + "learning_rate": 4.356918983547678e-06, + "loss": 0.5136, + "step": 3059 + }, + { + "epoch": 2.9371099375900145, + "grad_norm": 0.9239812370800711, + "learning_rate": 4.353596991234228e-06, + "loss": 0.5449, + "step": 3060 + }, + { + "epoch": 2.9380700912145943, + "grad_norm": 1.8903715342860052, + "learning_rate": 4.3502752890841825e-06, + "loss": 0.5592, + "step": 3061 + }, + { + "epoch": 2.9390302448391745, + "grad_norm": 1.1373301173933708, + "learning_rate": 4.3469538785886115e-06, + "loss": 0.4991, + "step": 3062 + }, + { + "epoch": 2.9399903984637543, + "grad_norm": 1.3397196120331643, + "learning_rate": 4.343632761238465e-06, + "loss": 0.4994, + "step": 3063 + }, + { + "epoch": 2.940950552088334, + "grad_norm": 1.2104743324760188, + "learning_rate": 4.340311938524556e-06, + "loss": 0.4654, + "step": 3064 + }, + { + "epoch": 2.941910705712914, + "grad_norm": 1.1082009729444584, + "learning_rate": 4.336991411937566e-06, + "loss": 0.5203, + "step": 3065 + }, + { + "epoch": 2.942870859337494, + "grad_norm": 0.9814982846809901, + "learning_rate": 4.333671182968045e-06, + "loss": 0.5168, + "step": 3066 + }, + { + "epoch": 2.943831012962074, + "grad_norm": 0.9877106038691462, + "learning_rate": 4.330351253106405e-06, + "loss": 0.5403, + "step": 3067 + }, + { + "epoch": 2.9447911665866537, + "grad_norm": 1.1985681621184443, + "learning_rate": 4.327031623842931e-06, + "loss": 0.5349, + "step": 3068 + }, + { + "epoch": 2.945751320211234, + "grad_norm": 1.0081926647653574, + "learning_rate": 4.323712296667768e-06, + "loss": 0.5492, + "step": 3069 + }, + { + "epoch": 2.9467114738358138, + "grad_norm": 1.1225174524717747, + "learning_rate": 4.320393273070924e-06, + "loss": 0.472, + "step": 3070 + }, + { + "epoch": 2.9476716274603936, + "grad_norm": 0.9103503434742389, + "learning_rate": 4.317074554542277e-06, + "loss": 0.5209, + "step": 3071 + }, + { + "epoch": 2.9486317810849734, + "grad_norm": 1.0713157935747046, + "learning_rate": 4.313756142571561e-06, + "loss": 0.5833, + "step": 3072 + }, + { + "epoch": 2.9495919347095536, + "grad_norm": 1.4210267935329106, + "learning_rate": 4.310438038648376e-06, + "loss": 0.5986, + "step": 3073 + }, + { + "epoch": 2.9505520883341334, + "grad_norm": 1.0818317997358697, + "learning_rate": 4.307120244262189e-06, + "loss": 0.5342, + "step": 3074 + }, + { + "epoch": 2.9515122419587136, + "grad_norm": 1.038315223312765, + "learning_rate": 4.303802760902313e-06, + "loss": 0.5476, + "step": 3075 + }, + { + "epoch": 2.9524723955832934, + "grad_norm": 0.9581055780089218, + "learning_rate": 4.300485590057939e-06, + "loss": 0.5294, + "step": 3076 + }, + { + "epoch": 2.9534325492078732, + "grad_norm": 0.9610387983102177, + "learning_rate": 4.2971687332181035e-06, + "loss": 0.5661, + "step": 3077 + }, + { + "epoch": 2.954392702832453, + "grad_norm": 1.450956072149878, + "learning_rate": 4.293852191871712e-06, + "loss": 0.5403, + "step": 3078 + }, + { + "epoch": 2.9553528564570333, + "grad_norm": 0.9737547138290833, + "learning_rate": 4.290535967507524e-06, + "loss": 0.6408, + "step": 3079 + }, + { + "epoch": 2.956313010081613, + "grad_norm": 0.9700309566318754, + "learning_rate": 4.287220061614155e-06, + "loss": 0.6155, + "step": 3080 + }, + { + "epoch": 2.957273163706193, + "grad_norm": 1.468628724969963, + "learning_rate": 4.283904475680082e-06, + "loss": 0.5536, + "step": 3081 + }, + { + "epoch": 2.958233317330773, + "grad_norm": 1.2590260462192808, + "learning_rate": 4.280589211193635e-06, + "loss": 0.5509, + "step": 3082 + }, + { + "epoch": 2.959193470955353, + "grad_norm": 1.3289292495566223, + "learning_rate": 4.277274269643002e-06, + "loss": 0.6351, + "step": 3083 + }, + { + "epoch": 2.9601536245799327, + "grad_norm": 1.009392083833032, + "learning_rate": 4.273959652516223e-06, + "loss": 0.6123, + "step": 3084 + }, + { + "epoch": 2.9611137782045125, + "grad_norm": 0.881239926721038, + "learning_rate": 4.270645361301195e-06, + "loss": 0.629, + "step": 3085 + }, + { + "epoch": 2.9620739318290927, + "grad_norm": 1.2398684285711008, + "learning_rate": 4.267331397485666e-06, + "loss": 0.6218, + "step": 3086 + }, + { + "epoch": 2.9630340854536725, + "grad_norm": 1.03551343661382, + "learning_rate": 4.264017762557245e-06, + "loss": 0.5348, + "step": 3087 + }, + { + "epoch": 2.9639942390782528, + "grad_norm": 1.2133383321685893, + "learning_rate": 4.260704458003379e-06, + "loss": 0.5282, + "step": 3088 + }, + { + "epoch": 2.9649543927028326, + "grad_norm": 1.2295496409916837, + "learning_rate": 4.257391485311384e-06, + "loss": 0.5988, + "step": 3089 + }, + { + "epoch": 2.9659145463274124, + "grad_norm": 0.8450511701456784, + "learning_rate": 4.254078845968409e-06, + "loss": 0.4831, + "step": 3090 + }, + { + "epoch": 2.966874699951992, + "grad_norm": 1.1926041598670738, + "learning_rate": 4.2507665414614675e-06, + "loss": 0.5523, + "step": 3091 + }, + { + "epoch": 2.9678348535765724, + "grad_norm": 1.0802042541618795, + "learning_rate": 4.247454573277418e-06, + "loss": 0.5847, + "step": 3092 + }, + { + "epoch": 2.968795007201152, + "grad_norm": 0.8101417404705086, + "learning_rate": 4.244142942902965e-06, + "loss": 0.557, + "step": 3093 + }, + { + "epoch": 2.969755160825732, + "grad_norm": 0.9978787381280483, + "learning_rate": 4.240831651824665e-06, + "loss": 0.5475, + "step": 3094 + }, + { + "epoch": 2.9707153144503122, + "grad_norm": 0.7973446788429128, + "learning_rate": 4.237520701528923e-06, + "loss": 0.4878, + "step": 3095 + }, + { + "epoch": 2.971675468074892, + "grad_norm": 1.5089364654362984, + "learning_rate": 4.234210093501987e-06, + "loss": 0.5532, + "step": 3096 + }, + { + "epoch": 2.972635621699472, + "grad_norm": 1.2666030325705535, + "learning_rate": 4.230899829229955e-06, + "loss": 0.52, + "step": 3097 + }, + { + "epoch": 2.9735957753240516, + "grad_norm": 1.8606649883668172, + "learning_rate": 4.2275899101987665e-06, + "loss": 0.6, + "step": 3098 + }, + { + "epoch": 2.974555928948632, + "grad_norm": 0.9830532061304403, + "learning_rate": 4.2242803378942115e-06, + "loss": 0.4973, + "step": 3099 + }, + { + "epoch": 2.9755160825732117, + "grad_norm": 0.8637325283470042, + "learning_rate": 4.220971113801921e-06, + "loss": 0.5082, + "step": 3100 + }, + { + "epoch": 2.976476236197792, + "grad_norm": 1.0487416355119943, + "learning_rate": 4.2176622394073705e-06, + "loss": 0.5563, + "step": 3101 + }, + { + "epoch": 2.9774363898223717, + "grad_norm": 1.1094095686360002, + "learning_rate": 4.214353716195879e-06, + "loss": 0.5545, + "step": 3102 + }, + { + "epoch": 2.9783965434469515, + "grad_norm": 1.4765371168803285, + "learning_rate": 4.211045545652604e-06, + "loss": 0.6027, + "step": 3103 + }, + { + "epoch": 2.9793566970715313, + "grad_norm": 1.1411382826006238, + "learning_rate": 4.207737729262549e-06, + "loss": 0.5777, + "step": 3104 + }, + { + "epoch": 2.9803168506961115, + "grad_norm": 1.0429061032865632, + "learning_rate": 4.2044302685105635e-06, + "loss": 0.599, + "step": 3105 + }, + { + "epoch": 2.9812770043206913, + "grad_norm": 0.9688763849898634, + "learning_rate": 4.201123164881321e-06, + "loss": 0.54, + "step": 3106 + }, + { + "epoch": 2.982237157945271, + "grad_norm": 0.8287155643391799, + "learning_rate": 4.1978164198593524e-06, + "loss": 0.5797, + "step": 3107 + }, + { + "epoch": 2.9831973115698514, + "grad_norm": 1.0522580344235575, + "learning_rate": 4.194510034929019e-06, + "loss": 0.6053, + "step": 3108 + }, + { + "epoch": 2.984157465194431, + "grad_norm": 1.1055882647196837, + "learning_rate": 4.19120401157452e-06, + "loss": 0.4796, + "step": 3109 + }, + { + "epoch": 2.985117618819011, + "grad_norm": 0.79377984148004, + "learning_rate": 4.187898351279895e-06, + "loss": 0.3955, + "step": 3110 + }, + { + "epoch": 2.9860777724435907, + "grad_norm": 0.9330418861567719, + "learning_rate": 4.184593055529018e-06, + "loss": 0.4637, + "step": 3111 + }, + { + "epoch": 2.987037926068171, + "grad_norm": 1.0066222445936424, + "learning_rate": 4.181288125805603e-06, + "loss": 0.5016, + "step": 3112 + }, + { + "epoch": 2.9879980796927508, + "grad_norm": 0.7760807705242411, + "learning_rate": 4.177983563593198e-06, + "loss": 0.4345, + "step": 3113 + }, + { + "epoch": 2.988958233317331, + "grad_norm": 1.0514174207579965, + "learning_rate": 4.174679370375182e-06, + "loss": 0.4325, + "step": 3114 + }, + { + "epoch": 2.989918386941911, + "grad_norm": 0.9141277819725362, + "learning_rate": 4.1713755476347766e-06, + "loss": 0.5234, + "step": 3115 + }, + { + "epoch": 2.9908785405664906, + "grad_norm": 1.7132038904312563, + "learning_rate": 4.168072096855029e-06, + "loss": 0.5674, + "step": 3116 + }, + { + "epoch": 2.9918386941910704, + "grad_norm": 1.1024169808053594, + "learning_rate": 4.164769019518823e-06, + "loss": 0.5336, + "step": 3117 + }, + { + "epoch": 2.9927988478156506, + "grad_norm": 1.2565038017923282, + "learning_rate": 4.16146631710888e-06, + "loss": 0.5238, + "step": 3118 + }, + { + "epoch": 2.9937590014402304, + "grad_norm": 1.7202871269275235, + "learning_rate": 4.15816399110774e-06, + "loss": 0.6322, + "step": 3119 + }, + { + "epoch": 2.9947191550648102, + "grad_norm": 1.30347203310255, + "learning_rate": 4.154862042997789e-06, + "loss": 0.6312, + "step": 3120 + }, + { + "epoch": 2.9956793086893905, + "grad_norm": 1.2767992360688722, + "learning_rate": 4.151560474261229e-06, + "loss": 0.4833, + "step": 3121 + }, + { + "epoch": 2.9966394623139703, + "grad_norm": 1.5121299471074614, + "learning_rate": 4.148259286380105e-06, + "loss": 0.6428, + "step": 3122 + }, + { + "epoch": 2.99759961593855, + "grad_norm": 0.9268698627588248, + "learning_rate": 4.144958480836283e-06, + "loss": 0.5043, + "step": 3123 + }, + { + "epoch": 2.99855976956313, + "grad_norm": 1.2804185748689254, + "learning_rate": 4.141658059111459e-06, + "loss": 0.4551, + "step": 3124 + }, + { + "epoch": 2.99951992318771, + "grad_norm": 1.406455168534148, + "learning_rate": 4.138358022687156e-06, + "loss": 0.5526, + "step": 3125 + }, + { + "epoch": 3.0, + "grad_norm": 1.406455168534148, + "learning_rate": 4.135058373044728e-06, + "loss": 0.4663, + "step": 3126 + }, + { + "epoch": 3.00096015362458, + "grad_norm": 1.9239636981940516, + "learning_rate": 4.131759111665349e-06, + "loss": 0.5365, + "step": 3127 + }, + { + "epoch": 3.00192030724916, + "grad_norm": 1.3228733602616092, + "learning_rate": 4.128460240030025e-06, + "loss": 0.4984, + "step": 3128 + }, + { + "epoch": 3.00288046087374, + "grad_norm": 1.0793428323285756, + "learning_rate": 4.125161759619582e-06, + "loss": 0.4681, + "step": 3129 + }, + { + "epoch": 3.0038406144983196, + "grad_norm": 1.3004225045164435, + "learning_rate": 4.121863671914672e-06, + "loss": 0.4962, + "step": 3130 + }, + { + "epoch": 3.0048007681229, + "grad_norm": 1.063124342865363, + "learning_rate": 4.118565978395777e-06, + "loss": 0.4961, + "step": 3131 + }, + { + "epoch": 3.0057609217474797, + "grad_norm": 1.2375039499986231, + "learning_rate": 4.115268680543189e-06, + "loss": 0.4713, + "step": 3132 + }, + { + "epoch": 3.0067210753720595, + "grad_norm": 1.0197191376472794, + "learning_rate": 4.111971779837036e-06, + "loss": 0.4491, + "step": 3133 + }, + { + "epoch": 3.0076812289966393, + "grad_norm": 0.8492726664650959, + "learning_rate": 4.108675277757255e-06, + "loss": 0.5794, + "step": 3134 + }, + { + "epoch": 3.0086413826212195, + "grad_norm": 1.0663777392940654, + "learning_rate": 4.105379175783616e-06, + "loss": 0.519, + "step": 3135 + }, + { + "epoch": 3.0096015362457993, + "grad_norm": 1.3478638059709642, + "learning_rate": 4.102083475395702e-06, + "loss": 0.4668, + "step": 3136 + }, + { + "epoch": 3.010561689870379, + "grad_norm": 1.0797695538189431, + "learning_rate": 4.098788178072917e-06, + "loss": 0.4843, + "step": 3137 + }, + { + "epoch": 3.0115218434949593, + "grad_norm": 1.8100865280275094, + "learning_rate": 4.095493285294486e-06, + "loss": 0.5063, + "step": 3138 + }, + { + "epoch": 3.012481997119539, + "grad_norm": 1.1855532300263525, + "learning_rate": 4.092198798539452e-06, + "loss": 0.5044, + "step": 3139 + }, + { + "epoch": 3.013442150744119, + "grad_norm": 1.1934692504859403, + "learning_rate": 4.088904719286671e-06, + "loss": 0.3971, + "step": 3140 + }, + { + "epoch": 3.014402304368699, + "grad_norm": 1.3325929130702021, + "learning_rate": 4.085611049014825e-06, + "loss": 0.4535, + "step": 3141 + }, + { + "epoch": 3.015362457993279, + "grad_norm": 1.1200829101249894, + "learning_rate": 4.082317789202403e-06, + "loss": 0.4413, + "step": 3142 + }, + { + "epoch": 3.0163226116178588, + "grad_norm": 1.3120401515246576, + "learning_rate": 4.079024941327715e-06, + "loss": 0.4709, + "step": 3143 + }, + { + "epoch": 3.017282765242439, + "grad_norm": 0.9877878500932993, + "learning_rate": 4.075732506868889e-06, + "loss": 0.5058, + "step": 3144 + }, + { + "epoch": 3.018242918867019, + "grad_norm": 1.4300568342369686, + "learning_rate": 4.072440487303859e-06, + "loss": 0.5296, + "step": 3145 + }, + { + "epoch": 3.0192030724915986, + "grad_norm": 1.1336881826643528, + "learning_rate": 4.06914888411038e-06, + "loss": 0.4567, + "step": 3146 + }, + { + "epoch": 3.0201632261161784, + "grad_norm": 1.2940453158406124, + "learning_rate": 4.065857698766016e-06, + "loss": 0.4519, + "step": 3147 + }, + { + "epoch": 3.0211233797407586, + "grad_norm": 1.7038628093050756, + "learning_rate": 4.062566932748143e-06, + "loss": 0.4981, + "step": 3148 + }, + { + "epoch": 3.0220835333653384, + "grad_norm": 1.6262079457565777, + "learning_rate": 4.059276587533959e-06, + "loss": 0.3239, + "step": 3149 + }, + { + "epoch": 3.023043686989918, + "grad_norm": 1.3827984010683527, + "learning_rate": 4.0559866646004546e-06, + "loss": 0.3996, + "step": 3150 + }, + { + "epoch": 3.0240038406144985, + "grad_norm": 1.3727816316241965, + "learning_rate": 4.052697165424448e-06, + "loss": 0.4812, + "step": 3151 + }, + { + "epoch": 3.0249639942390782, + "grad_norm": 1.8550271569220913, + "learning_rate": 4.049408091482556e-06, + "loss": 0.4482, + "step": 3152 + }, + { + "epoch": 3.025924147863658, + "grad_norm": 1.6445361829073784, + "learning_rate": 4.046119444251211e-06, + "loss": 0.537, + "step": 3153 + }, + { + "epoch": 3.0268843014882383, + "grad_norm": 1.1083472073002156, + "learning_rate": 4.042831225206651e-06, + "loss": 0.5213, + "step": 3154 + }, + { + "epoch": 3.027844455112818, + "grad_norm": 0.9762787260620396, + "learning_rate": 4.039543435824923e-06, + "loss": 0.4877, + "step": 3155 + }, + { + "epoch": 3.028804608737398, + "grad_norm": 0.696652041806107, + "learning_rate": 4.036256077581879e-06, + "loss": 0.4781, + "step": 3156 + }, + { + "epoch": 3.029764762361978, + "grad_norm": 1.4782959603246209, + "learning_rate": 4.032969151953182e-06, + "loss": 0.4845, + "step": 3157 + }, + { + "epoch": 3.030724915986558, + "grad_norm": 1.2575536480360427, + "learning_rate": 4.029682660414294e-06, + "loss": 0.553, + "step": 3158 + }, + { + "epoch": 3.0316850696111377, + "grad_norm": 1.0250506698631832, + "learning_rate": 4.0263966044404885e-06, + "loss": 0.4167, + "step": 3159 + }, + { + "epoch": 3.0326452232357175, + "grad_norm": 1.1007246132458646, + "learning_rate": 4.023110985506839e-06, + "loss": 0.4504, + "step": 3160 + }, + { + "epoch": 3.0336053768602977, + "grad_norm": 0.9604262341570069, + "learning_rate": 4.019825805088225e-06, + "loss": 0.6067, + "step": 3161 + }, + { + "epoch": 3.0345655304848775, + "grad_norm": 0.736231017264886, + "learning_rate": 4.0165410646593324e-06, + "loss": 0.6075, + "step": 3162 + }, + { + "epoch": 3.0355256841094573, + "grad_norm": 1.1126325326648259, + "learning_rate": 4.01325676569464e-06, + "loss": 0.5049, + "step": 3163 + }, + { + "epoch": 3.0364858377340376, + "grad_norm": 1.277468939327407, + "learning_rate": 4.009972909668442e-06, + "loss": 0.366, + "step": 3164 + }, + { + "epoch": 3.0374459913586174, + "grad_norm": 1.2387999787478754, + "learning_rate": 4.006689498054817e-06, + "loss": 0.576, + "step": 3165 + }, + { + "epoch": 3.038406144983197, + "grad_norm": 1.0536507323725626, + "learning_rate": 4.00340653232766e-06, + "loss": 0.5671, + "step": 3166 + }, + { + "epoch": 3.0393662986077774, + "grad_norm": 1.124388549333207, + "learning_rate": 4.00012401396066e-06, + "loss": 0.5529, + "step": 3167 + }, + { + "epoch": 3.040326452232357, + "grad_norm": 1.5898668256174617, + "learning_rate": 3.9968419444273e-06, + "loss": 0.4886, + "step": 3168 + }, + { + "epoch": 3.041286605856937, + "grad_norm": 1.3596397519775179, + "learning_rate": 3.993560325200869e-06, + "loss": 0.4261, + "step": 3169 + }, + { + "epoch": 3.0422467594815172, + "grad_norm": 1.173627102680041, + "learning_rate": 3.99027915775445e-06, + "loss": 0.3541, + "step": 3170 + }, + { + "epoch": 3.043206913106097, + "grad_norm": 1.5247708167999623, + "learning_rate": 3.986998443560923e-06, + "loss": 0.5323, + "step": 3171 + }, + { + "epoch": 3.044167066730677, + "grad_norm": 1.3288642811110243, + "learning_rate": 3.9837181840929665e-06, + "loss": 0.4968, + "step": 3172 + }, + { + "epoch": 3.0451272203552566, + "grad_norm": 1.1244635357268893, + "learning_rate": 3.980438380823053e-06, + "loss": 0.5472, + "step": 3173 + }, + { + "epoch": 3.046087373979837, + "grad_norm": 1.1514397821116547, + "learning_rate": 3.977159035223451e-06, + "loss": 0.4639, + "step": 3174 + }, + { + "epoch": 3.0470475276044167, + "grad_norm": 1.3764671192085696, + "learning_rate": 3.973880148766228e-06, + "loss": 0.5002, + "step": 3175 + }, + { + "epoch": 3.0480076812289965, + "grad_norm": 0.9376345105827826, + "learning_rate": 3.970601722923234e-06, + "loss": 0.4932, + "step": 3176 + }, + { + "epoch": 3.0489678348535767, + "grad_norm": 0.9781571016286641, + "learning_rate": 3.9673237591661265e-06, + "loss": 0.5122, + "step": 3177 + }, + { + "epoch": 3.0499279884781565, + "grad_norm": 1.1548070316901065, + "learning_rate": 3.964046258966342e-06, + "loss": 0.3892, + "step": 3178 + }, + { + "epoch": 3.0508881421027363, + "grad_norm": 1.1524801035715257, + "learning_rate": 3.960769223795118e-06, + "loss": 0.5279, + "step": 3179 + }, + { + "epoch": 3.0518482957273165, + "grad_norm": 0.9123047804535923, + "learning_rate": 3.957492655123483e-06, + "loss": 0.4769, + "step": 3180 + }, + { + "epoch": 3.0528084493518963, + "grad_norm": 1.0365703704266618, + "learning_rate": 3.954216554422249e-06, + "loss": 0.61, + "step": 3181 + }, + { + "epoch": 3.053768602976476, + "grad_norm": 1.0829501655121776, + "learning_rate": 3.9509409231620264e-06, + "loss": 0.5291, + "step": 3182 + }, + { + "epoch": 3.0547287566010564, + "grad_norm": 0.9651497491775098, + "learning_rate": 3.947665762813209e-06, + "loss": 0.4728, + "step": 3183 + }, + { + "epoch": 3.055688910225636, + "grad_norm": 1.1138336240051303, + "learning_rate": 3.944391074845982e-06, + "loss": 0.4506, + "step": 3184 + }, + { + "epoch": 3.056649063850216, + "grad_norm": 1.0769846623869541, + "learning_rate": 3.941116860730319e-06, + "loss": 0.5404, + "step": 3185 + }, + { + "epoch": 3.0576092174747957, + "grad_norm": 1.2759802897492079, + "learning_rate": 3.9378431219359755e-06, + "loss": 0.5389, + "step": 3186 + }, + { + "epoch": 3.058569371099376, + "grad_norm": 0.8355257768799682, + "learning_rate": 3.934569859932502e-06, + "loss": 0.4113, + "step": 3187 + }, + { + "epoch": 3.059529524723956, + "grad_norm": 0.7681107528899316, + "learning_rate": 3.9312970761892305e-06, + "loss": 0.4384, + "step": 3188 + }, + { + "epoch": 3.0604896783485356, + "grad_norm": 1.0274639611804495, + "learning_rate": 3.9280247721752765e-06, + "loss": 0.5275, + "step": 3189 + }, + { + "epoch": 3.061449831973116, + "grad_norm": 0.9168806243794024, + "learning_rate": 3.924752949359545e-06, + "loss": 0.4163, + "step": 3190 + }, + { + "epoch": 3.0624099855976956, + "grad_norm": 1.116107818033291, + "learning_rate": 3.921481609210719e-06, + "loss": 0.3245, + "step": 3191 + }, + { + "epoch": 3.0633701392222754, + "grad_norm": 1.3137393963694208, + "learning_rate": 3.918210753197269e-06, + "loss": 0.3701, + "step": 3192 + }, + { + "epoch": 3.0643302928468557, + "grad_norm": 0.8407706817305267, + "learning_rate": 3.914940382787452e-06, + "loss": 0.4979, + "step": 3193 + }, + { + "epoch": 3.0652904464714354, + "grad_norm": 1.598056165872704, + "learning_rate": 3.911670499449295e-06, + "loss": 0.413, + "step": 3194 + }, + { + "epoch": 3.0662506000960152, + "grad_norm": 1.4350338721525306, + "learning_rate": 3.908401104650621e-06, + "loss": 0.5327, + "step": 3195 + }, + { + "epoch": 3.0672107537205955, + "grad_norm": 1.2740972841552856, + "learning_rate": 3.905132199859018e-06, + "loss": 0.5272, + "step": 3196 + }, + { + "epoch": 3.0681709073451753, + "grad_norm": 1.0656897951599038, + "learning_rate": 3.901863786541869e-06, + "loss": 0.4698, + "step": 3197 + }, + { + "epoch": 3.069131060969755, + "grad_norm": 1.2564910897357613, + "learning_rate": 3.898595866166329e-06, + "loss": 0.4034, + "step": 3198 + }, + { + "epoch": 3.070091214594335, + "grad_norm": 0.9616820602484117, + "learning_rate": 3.895328440199329e-06, + "loss": 0.5009, + "step": 3199 + }, + { + "epoch": 3.071051368218915, + "grad_norm": 1.0296404974142745, + "learning_rate": 3.892061510107586e-06, + "loss": 0.4821, + "step": 3200 + }, + { + "epoch": 3.072011521843495, + "grad_norm": 1.1667899184325679, + "learning_rate": 3.888795077357587e-06, + "loss": 0.4823, + "step": 3201 + }, + { + "epoch": 3.0729716754680747, + "grad_norm": 0.9112201502963381, + "learning_rate": 3.885529143415599e-06, + "loss": 0.4716, + "step": 3202 + }, + { + "epoch": 3.073931829092655, + "grad_norm": 1.048266883447334, + "learning_rate": 3.882263709747667e-06, + "loss": 0.4892, + "step": 3203 + }, + { + "epoch": 3.0748919827172347, + "grad_norm": 1.3039531108592672, + "learning_rate": 3.878998777819607e-06, + "loss": 0.4388, + "step": 3204 + }, + { + "epoch": 3.0758521363418145, + "grad_norm": 1.4831555191361303, + "learning_rate": 3.8757343490970115e-06, + "loss": 0.4733, + "step": 3205 + }, + { + "epoch": 3.0768122899663948, + "grad_norm": 1.158334615855496, + "learning_rate": 3.872470425045251e-06, + "loss": 0.4721, + "step": 3206 + }, + { + "epoch": 3.0777724435909746, + "grad_norm": 1.287705739501043, + "learning_rate": 3.869207007129463e-06, + "loss": 0.4967, + "step": 3207 + }, + { + "epoch": 3.0787325972155544, + "grad_norm": 1.232998826458165, + "learning_rate": 3.865944096814564e-06, + "loss": 0.5472, + "step": 3208 + }, + { + "epoch": 3.0796927508401346, + "grad_norm": 1.0185919178748002, + "learning_rate": 3.862681695565234e-06, + "loss": 0.4131, + "step": 3209 + }, + { + "epoch": 3.0806529044647144, + "grad_norm": 0.8653549775039087, + "learning_rate": 3.859419804845935e-06, + "loss": 0.5035, + "step": 3210 + }, + { + "epoch": 3.081613058089294, + "grad_norm": 1.219744652512554, + "learning_rate": 3.856158426120895e-06, + "loss": 0.47, + "step": 3211 + }, + { + "epoch": 3.082573211713874, + "grad_norm": 1.2564983811371668, + "learning_rate": 3.8528975608541095e-06, + "loss": 0.4151, + "step": 3212 + }, + { + "epoch": 3.0835333653384542, + "grad_norm": 0.7138516531906572, + "learning_rate": 3.849637210509346e-06, + "loss": 0.4206, + "step": 3213 + }, + { + "epoch": 3.084493518963034, + "grad_norm": 1.048381836978345, + "learning_rate": 3.846377376550144e-06, + "loss": 0.4641, + "step": 3214 + }, + { + "epoch": 3.085453672587614, + "grad_norm": 1.3334048711381585, + "learning_rate": 3.843118060439805e-06, + "loss": 0.3623, + "step": 3215 + }, + { + "epoch": 3.086413826212194, + "grad_norm": 1.138159930191426, + "learning_rate": 3.839859263641403e-06, + "loss": 0.4043, + "step": 3216 + }, + { + "epoch": 3.087373979836774, + "grad_norm": 1.0229895039801273, + "learning_rate": 3.836600987617776e-06, + "loss": 0.4506, + "step": 3217 + }, + { + "epoch": 3.0883341334613537, + "grad_norm": 0.8673136493237804, + "learning_rate": 3.833343233831529e-06, + "loss": 0.4152, + "step": 3218 + }, + { + "epoch": 3.089294287085934, + "grad_norm": 0.9357362030273287, + "learning_rate": 3.830086003745034e-06, + "loss": 0.5096, + "step": 3219 + }, + { + "epoch": 3.0902544407105137, + "grad_norm": 1.4772522514161999, + "learning_rate": 3.826829298820425e-06, + "loss": 0.4703, + "step": 3220 + }, + { + "epoch": 3.0912145943350935, + "grad_norm": 1.0072406751468546, + "learning_rate": 3.823573120519606e-06, + "loss": 0.4959, + "step": 3221 + }, + { + "epoch": 3.0921747479596737, + "grad_norm": 1.2202622857467345, + "learning_rate": 3.8203174703042345e-06, + "loss": 0.4395, + "step": 3222 + }, + { + "epoch": 3.0931349015842535, + "grad_norm": 0.861688638601738, + "learning_rate": 3.817062349635742e-06, + "loss": 0.4745, + "step": 3223 + }, + { + "epoch": 3.0940950552088333, + "grad_norm": 1.1104091449681204, + "learning_rate": 3.813807759975317e-06, + "loss": 0.4227, + "step": 3224 + }, + { + "epoch": 3.095055208833413, + "grad_norm": 1.1075278256066776, + "learning_rate": 3.8105537027839078e-06, + "loss": 0.4286, + "step": 3225 + }, + { + "epoch": 3.0960153624579934, + "grad_norm": 0.8914863475256624, + "learning_rate": 3.8073001795222274e-06, + "loss": 0.3922, + "step": 3226 + }, + { + "epoch": 3.096975516082573, + "grad_norm": 1.3212083607993594, + "learning_rate": 3.8040471916507467e-06, + "loss": 0.5374, + "step": 3227 + }, + { + "epoch": 3.097935669707153, + "grad_norm": 0.9732801603490799, + "learning_rate": 3.8007947406296975e-06, + "loss": 0.4555, + "step": 3228 + }, + { + "epoch": 3.098895823331733, + "grad_norm": 1.0735942347910503, + "learning_rate": 3.797542827919071e-06, + "loss": 0.3659, + "step": 3229 + }, + { + "epoch": 3.099855976956313, + "grad_norm": 1.3439091960514233, + "learning_rate": 3.7942914549786152e-06, + "loss": 0.4808, + "step": 3230 + }, + { + "epoch": 3.100816130580893, + "grad_norm": 1.450476722211154, + "learning_rate": 3.7910406232678363e-06, + "loss": 0.4025, + "step": 3231 + }, + { + "epoch": 3.101776284205473, + "grad_norm": 1.3777931479697445, + "learning_rate": 3.7877903342460002e-06, + "loss": 0.4184, + "step": 3232 + }, + { + "epoch": 3.102736437830053, + "grad_norm": 1.4940735049135367, + "learning_rate": 3.7845405893721237e-06, + "loss": 0.4585, + "step": 3233 + }, + { + "epoch": 3.1036965914546326, + "grad_norm": 0.9654084159876766, + "learning_rate": 3.781291390104986e-06, + "loss": 0.472, + "step": 3234 + }, + { + "epoch": 3.104656745079213, + "grad_norm": 1.6243116592159903, + "learning_rate": 3.7780427379031137e-06, + "loss": 0.4137, + "step": 3235 + }, + { + "epoch": 3.1056168987037926, + "grad_norm": 1.1887888248526208, + "learning_rate": 3.774794634224794e-06, + "loss": 0.6594, + "step": 3236 + }, + { + "epoch": 3.1065770523283724, + "grad_norm": 1.4731057093868527, + "learning_rate": 3.7715470805280695e-06, + "loss": 0.4132, + "step": 3237 + }, + { + "epoch": 3.1075372059529527, + "grad_norm": 1.7175492898097233, + "learning_rate": 3.7683000782707258e-06, + "loss": 0.5084, + "step": 3238 + }, + { + "epoch": 3.1084973595775325, + "grad_norm": 1.0061148986911161, + "learning_rate": 3.765053628910314e-06, + "loss": 0.4858, + "step": 3239 + }, + { + "epoch": 3.1094575132021123, + "grad_norm": 1.1488406903906454, + "learning_rate": 3.7618077339041244e-06, + "loss": 0.4184, + "step": 3240 + }, + { + "epoch": 3.110417666826692, + "grad_norm": 1.4276915640649592, + "learning_rate": 3.758562394709208e-06, + "loss": 0.4356, + "step": 3241 + }, + { + "epoch": 3.1113778204512723, + "grad_norm": 1.289728246681116, + "learning_rate": 3.7553176127823643e-06, + "loss": 0.5564, + "step": 3242 + }, + { + "epoch": 3.112337974075852, + "grad_norm": 1.0279051391307596, + "learning_rate": 3.7520733895801375e-06, + "loss": 0.5863, + "step": 3243 + }, + { + "epoch": 3.113298127700432, + "grad_norm": 1.0665447788363311, + "learning_rate": 3.748829726558827e-06, + "loss": 0.504, + "step": 3244 + }, + { + "epoch": 3.114258281325012, + "grad_norm": 0.9704432657322218, + "learning_rate": 3.745586625174479e-06, + "loss": 0.3969, + "step": 3245 + }, + { + "epoch": 3.115218434949592, + "grad_norm": 1.070127672460268, + "learning_rate": 3.7423440868828842e-06, + "loss": 0.5753, + "step": 3246 + }, + { + "epoch": 3.1161785885741717, + "grad_norm": 1.1184348251680445, + "learning_rate": 3.7391021131395866e-06, + "loss": 0.4954, + "step": 3247 + }, + { + "epoch": 3.117138742198752, + "grad_norm": 1.0837417056276069, + "learning_rate": 3.7358607053998717e-06, + "loss": 0.5256, + "step": 3248 + }, + { + "epoch": 3.1180988958233318, + "grad_norm": 1.2362034284400025, + "learning_rate": 3.732619865118772e-06, + "loss": 0.6156, + "step": 3249 + }, + { + "epoch": 3.1190590494479116, + "grad_norm": 0.9943377433158054, + "learning_rate": 3.7293795937510703e-06, + "loss": 0.5776, + "step": 3250 + }, + { + "epoch": 3.120019203072492, + "grad_norm": 1.0699939660758246, + "learning_rate": 3.7261398927512836e-06, + "loss": 0.5344, + "step": 3251 + }, + { + "epoch": 3.1209793566970716, + "grad_norm": 1.0025947534822273, + "learning_rate": 3.722900763573685e-06, + "loss": 0.5069, + "step": 3252 + }, + { + "epoch": 3.1219395103216514, + "grad_norm": 1.140528324525897, + "learning_rate": 3.71966220767228e-06, + "loss": 0.4209, + "step": 3253 + }, + { + "epoch": 3.122899663946231, + "grad_norm": 1.2978734786116022, + "learning_rate": 3.7164242265008244e-06, + "loss": 0.491, + "step": 3254 + }, + { + "epoch": 3.1238598175708114, + "grad_norm": 0.9999203946788728, + "learning_rate": 3.7131868215128135e-06, + "loss": 0.5186, + "step": 3255 + }, + { + "epoch": 3.1248199711953912, + "grad_norm": 1.057205022313916, + "learning_rate": 3.7099499941614823e-06, + "loss": 0.4692, + "step": 3256 + }, + { + "epoch": 3.125780124819971, + "grad_norm": 1.2980050492898259, + "learning_rate": 3.7067137458998092e-06, + "loss": 0.5775, + "step": 3257 + }, + { + "epoch": 3.1267402784445513, + "grad_norm": 1.1190426964775122, + "learning_rate": 3.7034780781805126e-06, + "loss": 0.4075, + "step": 3258 + }, + { + "epoch": 3.127700432069131, + "grad_norm": 1.152389270233681, + "learning_rate": 3.7002429924560458e-06, + "loss": 0.4699, + "step": 3259 + }, + { + "epoch": 3.128660585693711, + "grad_norm": 1.216668100461093, + "learning_rate": 3.6970084901786075e-06, + "loss": 0.4619, + "step": 3260 + }, + { + "epoch": 3.129620739318291, + "grad_norm": 1.0513476971501425, + "learning_rate": 3.6937745728001283e-06, + "loss": 0.4733, + "step": 3261 + }, + { + "epoch": 3.130580892942871, + "grad_norm": 0.7897814078083351, + "learning_rate": 3.6905412417722807e-06, + "loss": 0.4976, + "step": 3262 + }, + { + "epoch": 3.1315410465674507, + "grad_norm": 1.1588697994784707, + "learning_rate": 3.6873084985464734e-06, + "loss": 0.5196, + "step": 3263 + }, + { + "epoch": 3.132501200192031, + "grad_norm": 1.0634528742230984, + "learning_rate": 3.6840763445738485e-06, + "loss": 0.5228, + "step": 3264 + }, + { + "epoch": 3.1334613538166107, + "grad_norm": 1.0666700384191108, + "learning_rate": 3.6808447813052873e-06, + "loss": 0.5539, + "step": 3265 + }, + { + "epoch": 3.1344215074411905, + "grad_norm": 1.0515651543556919, + "learning_rate": 3.677613810191401e-06, + "loss": 0.4234, + "step": 3266 + }, + { + "epoch": 3.1353816610657703, + "grad_norm": 1.1951511818551965, + "learning_rate": 3.674383432682539e-06, + "loss": 0.4887, + "step": 3267 + }, + { + "epoch": 3.1363418146903506, + "grad_norm": 0.9813920067817109, + "learning_rate": 3.671153650228787e-06, + "loss": 0.4065, + "step": 3268 + }, + { + "epoch": 3.1373019683149304, + "grad_norm": 0.9676268553436997, + "learning_rate": 3.6679244642799544e-06, + "loss": 0.4306, + "step": 3269 + }, + { + "epoch": 3.13826212193951, + "grad_norm": 1.1881893731950244, + "learning_rate": 3.664695876285592e-06, + "loss": 0.5699, + "step": 3270 + }, + { + "epoch": 3.1392222755640904, + "grad_norm": 1.461591051048004, + "learning_rate": 3.6614678876949767e-06, + "loss": 0.4189, + "step": 3271 + }, + { + "epoch": 3.14018242918867, + "grad_norm": 1.0026465226135486, + "learning_rate": 3.658240499957118e-06, + "loss": 0.3928, + "step": 3272 + }, + { + "epoch": 3.14114258281325, + "grad_norm": 1.0357425119454975, + "learning_rate": 3.6550137145207577e-06, + "loss": 0.4824, + "step": 3273 + }, + { + "epoch": 3.1421027364378302, + "grad_norm": 1.2120623370860484, + "learning_rate": 3.651787532834362e-06, + "loss": 0.4207, + "step": 3274 + }, + { + "epoch": 3.14306289006241, + "grad_norm": 1.0440246953335104, + "learning_rate": 3.648561956346132e-06, + "loss": 0.5504, + "step": 3275 + }, + { + "epoch": 3.14402304368699, + "grad_norm": 1.0017968909914516, + "learning_rate": 3.645336986503995e-06, + "loss": 0.4758, + "step": 3276 + }, + { + "epoch": 3.14498319731157, + "grad_norm": 1.1240838343915827, + "learning_rate": 3.642112624755603e-06, + "loss": 0.3923, + "step": 3277 + }, + { + "epoch": 3.14594335093615, + "grad_norm": 1.107268652506849, + "learning_rate": 3.6388888725483394e-06, + "loss": 0.4472, + "step": 3278 + }, + { + "epoch": 3.1469035045607296, + "grad_norm": 1.2371282603532567, + "learning_rate": 3.6356657313293104e-06, + "loss": 0.4953, + "step": 3279 + }, + { + "epoch": 3.1478636581853094, + "grad_norm": 1.0116247415883994, + "learning_rate": 3.6324432025453494e-06, + "loss": 0.5046, + "step": 3280 + }, + { + "epoch": 3.1488238118098897, + "grad_norm": 1.0068164058837563, + "learning_rate": 3.62922128764302e-06, + "loss": 0.5694, + "step": 3281 + }, + { + "epoch": 3.1497839654344695, + "grad_norm": 1.4933537670353163, + "learning_rate": 3.6259999880685987e-06, + "loss": 0.4549, + "step": 3282 + }, + { + "epoch": 3.1507441190590493, + "grad_norm": 0.7830821285727417, + "learning_rate": 3.6227793052680983e-06, + "loss": 0.5753, + "step": 3283 + }, + { + "epoch": 3.1517042726836295, + "grad_norm": 0.9154062087447861, + "learning_rate": 3.6195592406872436e-06, + "loss": 0.5167, + "step": 3284 + }, + { + "epoch": 3.1526644263082093, + "grad_norm": 1.0552162978473834, + "learning_rate": 3.6163397957714895e-06, + "loss": 0.5033, + "step": 3285 + }, + { + "epoch": 3.153624579932789, + "grad_norm": 1.315451938632114, + "learning_rate": 3.6131209719660133e-06, + "loss": 0.5488, + "step": 3286 + }, + { + "epoch": 3.1545847335573693, + "grad_norm": 1.3608193150497787, + "learning_rate": 3.609902770715705e-06, + "loss": 0.5327, + "step": 3287 + }, + { + "epoch": 3.155544887181949, + "grad_norm": 1.2519325582184433, + "learning_rate": 3.6066851934651847e-06, + "loss": 0.4471, + "step": 3288 + }, + { + "epoch": 3.156505040806529, + "grad_norm": 1.1592147615083936, + "learning_rate": 3.6034682416587885e-06, + "loss": 0.4497, + "step": 3289 + }, + { + "epoch": 3.157465194431109, + "grad_norm": 0.9694913809238275, + "learning_rate": 3.600251916740569e-06, + "loss": 0.4233, + "step": 3290 + }, + { + "epoch": 3.158425348055689, + "grad_norm": 0.9701111295164058, + "learning_rate": 3.597036220154303e-06, + "loss": 0.4099, + "step": 3291 + }, + { + "epoch": 3.1593855016802688, + "grad_norm": 1.0199823864484145, + "learning_rate": 3.5938211533434797e-06, + "loss": 0.4941, + "step": 3292 + }, + { + "epoch": 3.160345655304849, + "grad_norm": 1.0669644487761487, + "learning_rate": 3.590606717751308e-06, + "loss": 0.4531, + "step": 3293 + }, + { + "epoch": 3.161305808929429, + "grad_norm": 0.771997878277345, + "learning_rate": 3.5873929148207187e-06, + "loss": 0.4119, + "step": 3294 + }, + { + "epoch": 3.1622659625540086, + "grad_norm": 1.0522984652052074, + "learning_rate": 3.5841797459943473e-06, + "loss": 0.426, + "step": 3295 + }, + { + "epoch": 3.1632261161785884, + "grad_norm": 1.0066860826212287, + "learning_rate": 3.5809672127145563e-06, + "loss": 0.4972, + "step": 3296 + }, + { + "epoch": 3.1641862698031686, + "grad_norm": 1.0884200144630383, + "learning_rate": 3.5777553164234114e-06, + "loss": 0.4472, + "step": 3297 + }, + { + "epoch": 3.1651464234277484, + "grad_norm": 1.5140542399282748, + "learning_rate": 3.5745440585627032e-06, + "loss": 0.4027, + "step": 3298 + }, + { + "epoch": 3.1661065770523282, + "grad_norm": 1.1461248235172055, + "learning_rate": 3.571333440573932e-06, + "loss": 0.3802, + "step": 3299 + }, + { + "epoch": 3.1670667306769085, + "grad_norm": 0.9912716650082614, + "learning_rate": 3.5681234638983054e-06, + "loss": 0.5616, + "step": 3300 + }, + { + "epoch": 3.1680268843014883, + "grad_norm": 1.186252350558142, + "learning_rate": 3.5649141299767504e-06, + "loss": 0.443, + "step": 3301 + }, + { + "epoch": 3.168987037926068, + "grad_norm": 0.730051744463476, + "learning_rate": 3.561705440249903e-06, + "loss": 0.5011, + "step": 3302 + }, + { + "epoch": 3.1699471915506483, + "grad_norm": 1.18465251140844, + "learning_rate": 3.558497396158108e-06, + "loss": 0.5261, + "step": 3303 + }, + { + "epoch": 3.170907345175228, + "grad_norm": 1.1108408003802177, + "learning_rate": 3.5552899991414248e-06, + "loss": 0.5155, + "step": 3304 + }, + { + "epoch": 3.171867498799808, + "grad_norm": 0.8408718415219331, + "learning_rate": 3.5520832506396156e-06, + "loss": 0.5019, + "step": 3305 + }, + { + "epoch": 3.172827652424388, + "grad_norm": 1.1925771987881413, + "learning_rate": 3.548877152092158e-06, + "loss": 0.4683, + "step": 3306 + }, + { + "epoch": 3.173787806048968, + "grad_norm": 1.2642793493238076, + "learning_rate": 3.5456717049382373e-06, + "loss": 0.5608, + "step": 3307 + }, + { + "epoch": 3.1747479596735477, + "grad_norm": 0.970020546344962, + "learning_rate": 3.5424669106167408e-06, + "loss": 0.4829, + "step": 3308 + }, + { + "epoch": 3.1757081132981275, + "grad_norm": 0.9906885100727404, + "learning_rate": 3.5392627705662686e-06, + "loss": 0.4891, + "step": 3309 + }, + { + "epoch": 3.1766682669227078, + "grad_norm": 1.0518499699182908, + "learning_rate": 3.5360592862251242e-06, + "loss": 0.4512, + "step": 3310 + }, + { + "epoch": 3.1776284205472876, + "grad_norm": 1.0474757515728943, + "learning_rate": 3.5328564590313155e-06, + "loss": 0.5234, + "step": 3311 + }, + { + "epoch": 3.1785885741718674, + "grad_norm": 1.2201900305719278, + "learning_rate": 3.529654290422563e-06, + "loss": 0.5018, + "step": 3312 + }, + { + "epoch": 3.1795487277964476, + "grad_norm": 1.6103658165194548, + "learning_rate": 3.526452781836279e-06, + "loss": 0.5324, + "step": 3313 + }, + { + "epoch": 3.1805088814210274, + "grad_norm": 1.2746683664480682, + "learning_rate": 3.5232519347095935e-06, + "loss": 0.464, + "step": 3314 + }, + { + "epoch": 3.181469035045607, + "grad_norm": 1.0960185852689768, + "learning_rate": 3.5200517504793254e-06, + "loss": 0.4887, + "step": 3315 + }, + { + "epoch": 3.1824291886701874, + "grad_norm": 0.9771584574866977, + "learning_rate": 3.516852230582008e-06, + "loss": 0.3827, + "step": 3316 + }, + { + "epoch": 3.183389342294767, + "grad_norm": 0.8402776598601022, + "learning_rate": 3.513653376453872e-06, + "loss": 0.533, + "step": 3317 + }, + { + "epoch": 3.184349495919347, + "grad_norm": 1.218000048801814, + "learning_rate": 3.510455189530845e-06, + "loss": 0.477, + "step": 3318 + }, + { + "epoch": 3.1853096495439273, + "grad_norm": 1.339773296808066, + "learning_rate": 3.507257671248561e-06, + "loss": 0.4487, + "step": 3319 + }, + { + "epoch": 3.186269803168507, + "grad_norm": 1.6925606341338868, + "learning_rate": 3.5040608230423533e-06, + "loss": 0.4381, + "step": 3320 + }, + { + "epoch": 3.187229956793087, + "grad_norm": 1.6901513748600399, + "learning_rate": 3.5008646463472506e-06, + "loss": 0.4492, + "step": 3321 + }, + { + "epoch": 3.1881901104176666, + "grad_norm": 1.4941852465466554, + "learning_rate": 3.497669142597984e-06, + "loss": 0.4977, + "step": 3322 + }, + { + "epoch": 3.189150264042247, + "grad_norm": 1.7854948637874957, + "learning_rate": 3.4944743132289793e-06, + "loss": 0.5432, + "step": 3323 + }, + { + "epoch": 3.1901104176668267, + "grad_norm": 1.4064662071931047, + "learning_rate": 3.4912801596743615e-06, + "loss": 0.5089, + "step": 3324 + }, + { + "epoch": 3.1910705712914065, + "grad_norm": 1.005166713978039, + "learning_rate": 3.4880866833679555e-06, + "loss": 0.556, + "step": 3325 + }, + { + "epoch": 3.1920307249159867, + "grad_norm": 1.162207357434463, + "learning_rate": 3.4848938857432727e-06, + "loss": 0.5408, + "step": 3326 + }, + { + "epoch": 3.1929908785405665, + "grad_norm": 1.3742301676493038, + "learning_rate": 3.4817017682335326e-06, + "loss": 0.3645, + "step": 3327 + }, + { + "epoch": 3.1939510321651463, + "grad_norm": 1.1408811361356186, + "learning_rate": 3.478510332271635e-06, + "loss": 0.4802, + "step": 3328 + }, + { + "epoch": 3.1949111857897265, + "grad_norm": 1.3935825313256787, + "learning_rate": 3.475319579290187e-06, + "loss": 0.4794, + "step": 3329 + }, + { + "epoch": 3.1958713394143063, + "grad_norm": 1.5822717757408087, + "learning_rate": 3.4721295107214835e-06, + "loss": 0.4593, + "step": 3330 + }, + { + "epoch": 3.196831493038886, + "grad_norm": 0.8524738384891366, + "learning_rate": 3.4689401279975087e-06, + "loss": 0.4832, + "step": 3331 + }, + { + "epoch": 3.1977916466634664, + "grad_norm": 1.0663359354385602, + "learning_rate": 3.4657514325499454e-06, + "loss": 0.5063, + "step": 3332 + }, + { + "epoch": 3.198751800288046, + "grad_norm": 0.9170238183896795, + "learning_rate": 3.4625634258101636e-06, + "loss": 0.4861, + "step": 3333 + }, + { + "epoch": 3.199711953912626, + "grad_norm": 1.001647895389685, + "learning_rate": 3.459376109209226e-06, + "loss": 0.5152, + "step": 3334 + }, + { + "epoch": 3.2006721075372058, + "grad_norm": 1.3001995421432293, + "learning_rate": 3.456189484177884e-06, + "loss": 0.4323, + "step": 3335 + }, + { + "epoch": 3.201632261161786, + "grad_norm": 1.012916159155316, + "learning_rate": 3.4530035521465803e-06, + "loss": 0.5205, + "step": 3336 + }, + { + "epoch": 3.202592414786366, + "grad_norm": 1.1970758320366057, + "learning_rate": 3.4498183145454445e-06, + "loss": 0.4571, + "step": 3337 + }, + { + "epoch": 3.2035525684109456, + "grad_norm": 1.180635387707229, + "learning_rate": 3.446633772804299e-06, + "loss": 0.4831, + "step": 3338 + }, + { + "epoch": 3.204512722035526, + "grad_norm": 0.9896897038481279, + "learning_rate": 3.443449928352646e-06, + "loss": 0.388, + "step": 3339 + }, + { + "epoch": 3.2054728756601056, + "grad_norm": 1.0014371263730026, + "learning_rate": 3.4402667826196845e-06, + "loss": 0.5363, + "step": 3340 + }, + { + "epoch": 3.2064330292846854, + "grad_norm": 1.1995139113871789, + "learning_rate": 3.4370843370342887e-06, + "loss": 0.5185, + "step": 3341 + }, + { + "epoch": 3.2073931829092657, + "grad_norm": 1.0838767809520797, + "learning_rate": 3.433902593025028e-06, + "loss": 0.4338, + "step": 3342 + }, + { + "epoch": 3.2083533365338455, + "grad_norm": 1.0049377683233403, + "learning_rate": 3.430721552020154e-06, + "loss": 0.5724, + "step": 3343 + }, + { + "epoch": 3.2093134901584253, + "grad_norm": 1.202614872907095, + "learning_rate": 3.4275412154475996e-06, + "loss": 0.4236, + "step": 3344 + }, + { + "epoch": 3.2102736437830055, + "grad_norm": 1.362368374088749, + "learning_rate": 3.4243615847349855e-06, + "loss": 0.4111, + "step": 3345 + }, + { + "epoch": 3.2112337974075853, + "grad_norm": 1.0263026967459457, + "learning_rate": 3.4211826613096135e-06, + "loss": 0.4428, + "step": 3346 + }, + { + "epoch": 3.212193951032165, + "grad_norm": 1.1244581590935658, + "learning_rate": 3.418004446598467e-06, + "loss": 0.5376, + "step": 3347 + }, + { + "epoch": 3.213154104656745, + "grad_norm": 0.6812781864592568, + "learning_rate": 3.4148269420282155e-06, + "loss": 0.5687, + "step": 3348 + }, + { + "epoch": 3.214114258281325, + "grad_norm": 1.4033588235401515, + "learning_rate": 3.4116501490252037e-06, + "loss": 0.5465, + "step": 3349 + }, + { + "epoch": 3.215074411905905, + "grad_norm": 0.8564337119299268, + "learning_rate": 3.4084740690154617e-06, + "loss": 0.5893, + "step": 3350 + }, + { + "epoch": 3.2160345655304847, + "grad_norm": 1.0117580984207832, + "learning_rate": 3.405298703424699e-06, + "loss": 0.4107, + "step": 3351 + }, + { + "epoch": 3.216994719155065, + "grad_norm": 1.1463898157504062, + "learning_rate": 3.4021240536783e-06, + "loss": 0.4721, + "step": 3352 + }, + { + "epoch": 3.2179548727796448, + "grad_norm": 1.181012503463394, + "learning_rate": 3.3989501212013344e-06, + "loss": 0.4971, + "step": 3353 + }, + { + "epoch": 3.2189150264042246, + "grad_norm": 1.3756866362828968, + "learning_rate": 3.3957769074185443e-06, + "loss": 0.4897, + "step": 3354 + }, + { + "epoch": 3.219875180028805, + "grad_norm": 1.2784531705669089, + "learning_rate": 3.392604413754351e-06, + "loss": 0.3867, + "step": 3355 + }, + { + "epoch": 3.2208353336533846, + "grad_norm": 0.917163876562131, + "learning_rate": 3.3894326416328575e-06, + "loss": 0.4702, + "step": 3356 + }, + { + "epoch": 3.2217954872779644, + "grad_norm": 1.0948257174927034, + "learning_rate": 3.3862615924778326e-06, + "loss": 0.4988, + "step": 3357 + }, + { + "epoch": 3.2227556409025446, + "grad_norm": 1.0310106773910908, + "learning_rate": 3.3830912677127313e-06, + "loss": 0.4795, + "step": 3358 + }, + { + "epoch": 3.2237157945271244, + "grad_norm": 1.3671893523526126, + "learning_rate": 3.3799216687606727e-06, + "loss": 0.4901, + "step": 3359 + }, + { + "epoch": 3.224675948151704, + "grad_norm": 0.9566398085583736, + "learning_rate": 3.3767527970444614e-06, + "loss": 0.5341, + "step": 3360 + }, + { + "epoch": 3.225636101776284, + "grad_norm": 1.1932589789270014, + "learning_rate": 3.3735846539865684e-06, + "loss": 0.4825, + "step": 3361 + }, + { + "epoch": 3.2265962554008643, + "grad_norm": 1.2833169196774847, + "learning_rate": 3.370417241009138e-06, + "loss": 0.485, + "step": 3362 + }, + { + "epoch": 3.227556409025444, + "grad_norm": 1.6946584550643682, + "learning_rate": 3.367250559533989e-06, + "loss": 0.337, + "step": 3363 + }, + { + "epoch": 3.228516562650024, + "grad_norm": 1.208036184721561, + "learning_rate": 3.3640846109826107e-06, + "loss": 0.5515, + "step": 3364 + }, + { + "epoch": 3.229476716274604, + "grad_norm": 1.5140749661296737, + "learning_rate": 3.3609193967761637e-06, + "loss": 0.5919, + "step": 3365 + }, + { + "epoch": 3.230436869899184, + "grad_norm": 1.0667318856557129, + "learning_rate": 3.3577549183354797e-06, + "loss": 0.5171, + "step": 3366 + }, + { + "epoch": 3.2313970235237637, + "grad_norm": 1.0646065738217612, + "learning_rate": 3.3545911770810565e-06, + "loss": 0.4703, + "step": 3367 + }, + { + "epoch": 3.232357177148344, + "grad_norm": 1.1498645585646654, + "learning_rate": 3.3514281744330644e-06, + "loss": 0.4772, + "step": 3368 + }, + { + "epoch": 3.2333173307729237, + "grad_norm": 1.2698844006251553, + "learning_rate": 3.348265911811346e-06, + "loss": 0.4839, + "step": 3369 + }, + { + "epoch": 3.2342774843975035, + "grad_norm": 1.1693955675875376, + "learning_rate": 3.345104390635401e-06, + "loss": 0.4786, + "step": 3370 + }, + { + "epoch": 3.2352376380220838, + "grad_norm": 1.1242271232796532, + "learning_rate": 3.3419436123244076e-06, + "loss": 0.5496, + "step": 3371 + }, + { + "epoch": 3.2361977916466635, + "grad_norm": 0.9245939122961161, + "learning_rate": 3.3387835782972e-06, + "loss": 0.4833, + "step": 3372 + }, + { + "epoch": 3.2371579452712433, + "grad_norm": 1.1901852269025686, + "learning_rate": 3.3356242899722877e-06, + "loss": 0.4112, + "step": 3373 + }, + { + "epoch": 3.238118098895823, + "grad_norm": 0.8720786666862682, + "learning_rate": 3.3324657487678423e-06, + "loss": 0.5028, + "step": 3374 + }, + { + "epoch": 3.2390782525204034, + "grad_norm": 1.101206910697223, + "learning_rate": 3.3293079561016957e-06, + "loss": 0.5186, + "step": 3375 + }, + { + "epoch": 3.240038406144983, + "grad_norm": 1.2504122943737073, + "learning_rate": 3.3261509133913484e-06, + "loss": 0.5274, + "step": 3376 + }, + { + "epoch": 3.240998559769563, + "grad_norm": 1.3041076369141953, + "learning_rate": 3.3229946220539657e-06, + "loss": 0.4301, + "step": 3377 + }, + { + "epoch": 3.241958713394143, + "grad_norm": 1.1048596462101312, + "learning_rate": 3.3198390835063685e-06, + "loss": 0.4669, + "step": 3378 + }, + { + "epoch": 3.242918867018723, + "grad_norm": 1.2118225930095048, + "learning_rate": 3.3166842991650493e-06, + "loss": 0.3906, + "step": 3379 + }, + { + "epoch": 3.243879020643303, + "grad_norm": 1.1081278091583657, + "learning_rate": 3.3135302704461526e-06, + "loss": 0.4835, + "step": 3380 + }, + { + "epoch": 3.244839174267883, + "grad_norm": 0.9183897912325273, + "learning_rate": 3.31037699876549e-06, + "loss": 0.4727, + "step": 3381 + }, + { + "epoch": 3.245799327892463, + "grad_norm": 1.2383478227332478, + "learning_rate": 3.307224485538534e-06, + "loss": 0.5473, + "step": 3382 + }, + { + "epoch": 3.2467594815170426, + "grad_norm": 0.9179539109617296, + "learning_rate": 3.3040727321804083e-06, + "loss": 0.4235, + "step": 3383 + }, + { + "epoch": 3.247719635141623, + "grad_norm": 1.0470684611668648, + "learning_rate": 3.300921740105908e-06, + "loss": 0.5502, + "step": 3384 + }, + { + "epoch": 3.2486797887662027, + "grad_norm": 1.1523109663948459, + "learning_rate": 3.297771510729474e-06, + "loss": 0.4698, + "step": 3385 + }, + { + "epoch": 3.2496399423907825, + "grad_norm": 0.7895437107468081, + "learning_rate": 3.2946220454652134e-06, + "loss": 0.4404, + "step": 3386 + }, + { + "epoch": 3.2506000960153623, + "grad_norm": 1.3301520489344474, + "learning_rate": 3.2914733457268876e-06, + "loss": 0.4792, + "step": 3387 + }, + { + "epoch": 3.2515602496399425, + "grad_norm": 1.1369329810424718, + "learning_rate": 3.288325412927914e-06, + "loss": 0.4875, + "step": 3388 + }, + { + "epoch": 3.2525204032645223, + "grad_norm": 1.0372480042346035, + "learning_rate": 3.285178248481366e-06, + "loss": 0.5181, + "step": 3389 + }, + { + "epoch": 3.253480556889102, + "grad_norm": 1.4600363522590416, + "learning_rate": 3.282031853799971e-06, + "loss": 0.5002, + "step": 3390 + }, + { + "epoch": 3.2544407105136823, + "grad_norm": 1.8264208969823579, + "learning_rate": 3.278886230296113e-06, + "loss": 0.4852, + "step": 3391 + }, + { + "epoch": 3.255400864138262, + "grad_norm": 1.2567708763509158, + "learning_rate": 3.2757413793818294e-06, + "loss": 0.362, + "step": 3392 + }, + { + "epoch": 3.256361017762842, + "grad_norm": 1.0573259497847434, + "learning_rate": 3.2725973024688073e-06, + "loss": 0.3959, + "step": 3393 + }, + { + "epoch": 3.257321171387422, + "grad_norm": 1.3673160118755916, + "learning_rate": 3.2694540009683926e-06, + "loss": 0.4424, + "step": 3394 + }, + { + "epoch": 3.258281325012002, + "grad_norm": 1.2415590684023268, + "learning_rate": 3.266311476291579e-06, + "loss": 0.4529, + "step": 3395 + }, + { + "epoch": 3.2592414786365818, + "grad_norm": 1.316246835233954, + "learning_rate": 3.2631697298490105e-06, + "loss": 0.5188, + "step": 3396 + }, + { + "epoch": 3.260201632261162, + "grad_norm": 1.8505419386055224, + "learning_rate": 3.2600287630509854e-06, + "loss": 0.4197, + "step": 3397 + }, + { + "epoch": 3.261161785885742, + "grad_norm": 1.0858363045616548, + "learning_rate": 3.256888577307448e-06, + "loss": 0.4654, + "step": 3398 + }, + { + "epoch": 3.2621219395103216, + "grad_norm": 1.149974975061263, + "learning_rate": 3.2537491740279946e-06, + "loss": 0.4222, + "step": 3399 + }, + { + "epoch": 3.2630820931349014, + "grad_norm": 1.1122434305885254, + "learning_rate": 3.250610554621875e-06, + "loss": 0.481, + "step": 3400 + }, + { + "epoch": 3.2640422467594816, + "grad_norm": 1.0259082741978534, + "learning_rate": 3.2474727204979728e-06, + "loss": 0.5001, + "step": 3401 + }, + { + "epoch": 3.2650024003840614, + "grad_norm": 0.9838965282016903, + "learning_rate": 3.2443356730648363e-06, + "loss": 0.5344, + "step": 3402 + }, + { + "epoch": 3.265962554008641, + "grad_norm": 1.0617414570144292, + "learning_rate": 3.2411994137306463e-06, + "loss": 0.5246, + "step": 3403 + }, + { + "epoch": 3.2669227076332215, + "grad_norm": 1.1538887232016177, + "learning_rate": 3.23806394390324e-06, + "loss": 0.4985, + "step": 3404 + }, + { + "epoch": 3.2678828612578013, + "grad_norm": 0.9960359606115007, + "learning_rate": 3.234929264990096e-06, + "loss": 0.4315, + "step": 3405 + }, + { + "epoch": 3.268843014882381, + "grad_norm": 1.327478953197582, + "learning_rate": 3.231795378398337e-06, + "loss": 0.4122, + "step": 3406 + }, + { + "epoch": 3.2698031685069613, + "grad_norm": 1.0052665100987026, + "learning_rate": 3.2286622855347306e-06, + "loss": 0.4576, + "step": 3407 + }, + { + "epoch": 3.270763322131541, + "grad_norm": 0.9796713931822297, + "learning_rate": 3.2255299878056913e-06, + "loss": 0.427, + "step": 3408 + }, + { + "epoch": 3.271723475756121, + "grad_norm": 1.1915662604324562, + "learning_rate": 3.2223984866172715e-06, + "loss": 0.4556, + "step": 3409 + }, + { + "epoch": 3.272683629380701, + "grad_norm": 1.1870264380788835, + "learning_rate": 3.21926778337517e-06, + "loss": 0.5556, + "step": 3410 + }, + { + "epoch": 3.273643783005281, + "grad_norm": 1.5499519945152198, + "learning_rate": 3.2161378794847255e-06, + "loss": 0.5996, + "step": 3411 + }, + { + "epoch": 3.2746039366298607, + "grad_norm": 1.0989856023449642, + "learning_rate": 3.2130087763509165e-06, + "loss": 0.4208, + "step": 3412 + }, + { + "epoch": 3.2755640902544405, + "grad_norm": 1.1789745802119442, + "learning_rate": 3.2098804753783696e-06, + "loss": 0.4553, + "step": 3413 + }, + { + "epoch": 3.2765242438790207, + "grad_norm": 1.0881776174621691, + "learning_rate": 3.2067529779713384e-06, + "loss": 0.4282, + "step": 3414 + }, + { + "epoch": 3.2774843975036005, + "grad_norm": 0.9339671894649936, + "learning_rate": 3.2036262855337304e-06, + "loss": 0.5593, + "step": 3415 + }, + { + "epoch": 3.2784445511281803, + "grad_norm": 1.086597740555038, + "learning_rate": 3.200500399469077e-06, + "loss": 0.4424, + "step": 3416 + }, + { + "epoch": 3.2794047047527606, + "grad_norm": 1.0465027172551948, + "learning_rate": 3.1973753211805597e-06, + "loss": 0.4828, + "step": 3417 + }, + { + "epoch": 3.2803648583773404, + "grad_norm": 1.0325905508363895, + "learning_rate": 3.1942510520709928e-06, + "loss": 0.5094, + "step": 3418 + }, + { + "epoch": 3.28132501200192, + "grad_norm": 1.15513350758389, + "learning_rate": 3.1911275935428256e-06, + "loss": 0.4294, + "step": 3419 + }, + { + "epoch": 3.2822851656265004, + "grad_norm": 1.518021291501259, + "learning_rate": 3.1880049469981468e-06, + "loss": 0.4575, + "step": 3420 + }, + { + "epoch": 3.28324531925108, + "grad_norm": 1.4587076038083908, + "learning_rate": 3.1848831138386772e-06, + "loss": 0.5337, + "step": 3421 + }, + { + "epoch": 3.28420547287566, + "grad_norm": 0.8795338185889969, + "learning_rate": 3.1817620954657745e-06, + "loss": 0.4557, + "step": 3422 + }, + { + "epoch": 3.2851656265002402, + "grad_norm": 1.2431144860536827, + "learning_rate": 3.178641893280433e-06, + "loss": 0.4621, + "step": 3423 + }, + { + "epoch": 3.28612578012482, + "grad_norm": 1.5686525243364142, + "learning_rate": 3.1755225086832755e-06, + "loss": 0.4119, + "step": 3424 + }, + { + "epoch": 3.2870859337494, + "grad_norm": 0.7486346018277268, + "learning_rate": 3.17240394307456e-06, + "loss": 0.4562, + "step": 3425 + }, + { + "epoch": 3.2880460873739796, + "grad_norm": 1.2865116519965656, + "learning_rate": 3.169286197854181e-06, + "loss": 0.4699, + "step": 3426 + }, + { + "epoch": 3.28900624099856, + "grad_norm": 1.2580008937051055, + "learning_rate": 3.1661692744216564e-06, + "loss": 0.4202, + "step": 3427 + }, + { + "epoch": 3.2899663946231397, + "grad_norm": 1.3041173598167417, + "learning_rate": 3.163053174176145e-06, + "loss": 0.4516, + "step": 3428 + }, + { + "epoch": 3.2909265482477195, + "grad_norm": 1.484892422938102, + "learning_rate": 3.159937898516424e-06, + "loss": 0.4718, + "step": 3429 + }, + { + "epoch": 3.2918867018722997, + "grad_norm": 1.1966360423356948, + "learning_rate": 3.1568234488409126e-06, + "loss": 0.6373, + "step": 3430 + }, + { + "epoch": 3.2928468554968795, + "grad_norm": 1.0375136993167255, + "learning_rate": 3.153709826547654e-06, + "loss": 0.6153, + "step": 3431 + }, + { + "epoch": 3.2938070091214593, + "grad_norm": 1.3851694711256348, + "learning_rate": 3.1505970330343172e-06, + "loss": 0.5655, + "step": 3432 + }, + { + "epoch": 3.2947671627460395, + "grad_norm": 1.1507859670549394, + "learning_rate": 3.147485069698204e-06, + "loss": 0.3919, + "step": 3433 + }, + { + "epoch": 3.2957273163706193, + "grad_norm": 1.3178751563655322, + "learning_rate": 3.1443739379362396e-06, + "loss": 0.4346, + "step": 3434 + }, + { + "epoch": 3.296687469995199, + "grad_norm": 1.051420712945425, + "learning_rate": 3.1412636391449792e-06, + "loss": 0.5403, + "step": 3435 + }, + { + "epoch": 3.2976476236197794, + "grad_norm": 1.223455075404112, + "learning_rate": 3.1381541747206034e-06, + "loss": 0.4533, + "step": 3436 + }, + { + "epoch": 3.298607777244359, + "grad_norm": 0.7936865086870284, + "learning_rate": 3.1350455460589157e-06, + "loss": 0.3837, + "step": 3437 + }, + { + "epoch": 3.299567930868939, + "grad_norm": 1.0857151502458677, + "learning_rate": 3.131937754555347e-06, + "loss": 0.4601, + "step": 3438 + }, + { + "epoch": 3.3005280844935188, + "grad_norm": 0.9741083327453095, + "learning_rate": 3.128830801604953e-06, + "loss": 0.3819, + "step": 3439 + }, + { + "epoch": 3.301488238118099, + "grad_norm": 1.041916953774406, + "learning_rate": 3.125724688602409e-06, + "loss": 0.4653, + "step": 3440 + }, + { + "epoch": 3.302448391742679, + "grad_norm": 1.1966069536100719, + "learning_rate": 3.122619416942019e-06, + "loss": 0.3917, + "step": 3441 + }, + { + "epoch": 3.3034085453672586, + "grad_norm": 1.0193960101948056, + "learning_rate": 3.1195149880177027e-06, + "loss": 0.4986, + "step": 3442 + }, + { + "epoch": 3.304368698991839, + "grad_norm": 0.7881438157390115, + "learning_rate": 3.1164114032230063e-06, + "loss": 0.4653, + "step": 3443 + }, + { + "epoch": 3.3053288526164186, + "grad_norm": 0.8632803495885336, + "learning_rate": 3.1133086639511e-06, + "loss": 0.5546, + "step": 3444 + }, + { + "epoch": 3.3062890062409984, + "grad_norm": 1.1017526483484688, + "learning_rate": 3.1102067715947637e-06, + "loss": 0.626, + "step": 3445 + }, + { + "epoch": 3.3072491598655787, + "grad_norm": 1.2786833885885354, + "learning_rate": 3.10710572754641e-06, + "loss": 0.3558, + "step": 3446 + }, + { + "epoch": 3.3082093134901585, + "grad_norm": 0.9570696118206218, + "learning_rate": 3.104005533198058e-06, + "loss": 0.5335, + "step": 3447 + }, + { + "epoch": 3.3091694671147383, + "grad_norm": 1.3751514725149228, + "learning_rate": 3.100906189941356e-06, + "loss": 0.4759, + "step": 3448 + }, + { + "epoch": 3.3101296207393185, + "grad_norm": 1.2906503133724956, + "learning_rate": 3.0978076991675667e-06, + "loss": 0.461, + "step": 3449 + }, + { + "epoch": 3.3110897743638983, + "grad_norm": 1.4476935118879004, + "learning_rate": 3.0947100622675663e-06, + "loss": 0.5398, + "step": 3450 + }, + { + "epoch": 3.312049927988478, + "grad_norm": 0.9318079668790964, + "learning_rate": 3.091613280631853e-06, + "loss": 0.45, + "step": 3451 + }, + { + "epoch": 3.313010081613058, + "grad_norm": 1.1702058935334203, + "learning_rate": 3.08851735565054e-06, + "loss": 0.4344, + "step": 3452 + }, + { + "epoch": 3.313970235237638, + "grad_norm": 1.2461461360558348, + "learning_rate": 3.0854222887133513e-06, + "loss": 0.4286, + "step": 3453 + }, + { + "epoch": 3.314930388862218, + "grad_norm": 0.9763357955033047, + "learning_rate": 3.082328081209633e-06, + "loss": 0.5813, + "step": 3454 + }, + { + "epoch": 3.3158905424867977, + "grad_norm": 1.1330988220710863, + "learning_rate": 3.079234734528339e-06, + "loss": 0.4451, + "step": 3455 + }, + { + "epoch": 3.316850696111378, + "grad_norm": 1.1369137673997254, + "learning_rate": 3.07614225005804e-06, + "loss": 0.5123, + "step": 3456 + }, + { + "epoch": 3.3178108497359577, + "grad_norm": 0.9528213068630654, + "learning_rate": 3.0730506291869235e-06, + "loss": 0.4762, + "step": 3457 + }, + { + "epoch": 3.3187710033605375, + "grad_norm": 1.0022913952498307, + "learning_rate": 3.069959873302778e-06, + "loss": 0.4013, + "step": 3458 + }, + { + "epoch": 3.319731156985118, + "grad_norm": 1.1282189510727791, + "learning_rate": 3.0668699837930175e-06, + "loss": 0.5274, + "step": 3459 + }, + { + "epoch": 3.3206913106096976, + "grad_norm": 0.9659139619938799, + "learning_rate": 3.0637809620446547e-06, + "loss": 0.4892, + "step": 3460 + }, + { + "epoch": 3.3216514642342774, + "grad_norm": 0.8951935658245836, + "learning_rate": 3.0606928094443216e-06, + "loss": 0.4825, + "step": 3461 + }, + { + "epoch": 3.3226116178588576, + "grad_norm": 1.3330222605731297, + "learning_rate": 3.0576055273782574e-06, + "loss": 0.4465, + "step": 3462 + }, + { + "epoch": 3.3235717714834374, + "grad_norm": 1.0080412211279601, + "learning_rate": 3.0545191172323076e-06, + "loss": 0.4193, + "step": 3463 + }, + { + "epoch": 3.324531925108017, + "grad_norm": 1.0969758704723247, + "learning_rate": 3.0514335803919314e-06, + "loss": 0.4456, + "step": 3464 + }, + { + "epoch": 3.325492078732597, + "grad_norm": 1.3671975807837193, + "learning_rate": 3.048348918242191e-06, + "loss": 0.5039, + "step": 3465 + }, + { + "epoch": 3.3264522323571772, + "grad_norm": 1.1750093089391647, + "learning_rate": 3.045265132167759e-06, + "loss": 0.4663, + "step": 3466 + }, + { + "epoch": 3.327412385981757, + "grad_norm": 1.1575758311513453, + "learning_rate": 3.0421822235529154e-06, + "loss": 0.5039, + "step": 3467 + }, + { + "epoch": 3.328372539606337, + "grad_norm": 1.035873531954432, + "learning_rate": 3.0391001937815425e-06, + "loss": 0.4767, + "step": 3468 + }, + { + "epoch": 3.329332693230917, + "grad_norm": 1.4534583618470323, + "learning_rate": 3.036019044237132e-06, + "loss": 0.5261, + "step": 3469 + }, + { + "epoch": 3.330292846855497, + "grad_norm": 0.8341856090733533, + "learning_rate": 3.0329387763027794e-06, + "loss": 0.5003, + "step": 3470 + }, + { + "epoch": 3.3312530004800767, + "grad_norm": 1.0478543324000291, + "learning_rate": 3.029859391361183e-06, + "loss": 0.3969, + "step": 3471 + }, + { + "epoch": 3.332213154104657, + "grad_norm": 1.219496309750024, + "learning_rate": 3.026780890794647e-06, + "loss": 0.3882, + "step": 3472 + }, + { + "epoch": 3.3331733077292367, + "grad_norm": 0.9239846362761154, + "learning_rate": 3.0237032759850736e-06, + "loss": 0.5066, + "step": 3473 + }, + { + "epoch": 3.3341334613538165, + "grad_norm": 1.157510897432906, + "learning_rate": 3.020626548313976e-06, + "loss": 0.5152, + "step": 3474 + }, + { + "epoch": 3.3350936149783967, + "grad_norm": 1.2597328595495847, + "learning_rate": 3.017550709162464e-06, + "loss": 0.4714, + "step": 3475 + }, + { + "epoch": 3.3360537686029765, + "grad_norm": 0.9380508511247198, + "learning_rate": 3.014475759911246e-06, + "loss": 0.5032, + "step": 3476 + }, + { + "epoch": 3.3370139222275563, + "grad_norm": 1.0435923128503564, + "learning_rate": 3.0114017019406362e-06, + "loss": 0.4262, + "step": 3477 + }, + { + "epoch": 3.337974075852136, + "grad_norm": 1.541404588932945, + "learning_rate": 3.0083285366305447e-06, + "loss": 0.5179, + "step": 3478 + }, + { + "epoch": 3.3389342294767164, + "grad_norm": 1.3649451627519333, + "learning_rate": 3.0052562653604827e-06, + "loss": 0.4553, + "step": 3479 + }, + { + "epoch": 3.339894383101296, + "grad_norm": 1.2231930622706915, + "learning_rate": 3.002184889509562e-06, + "loss": 0.4557, + "step": 3480 + }, + { + "epoch": 3.340854536725876, + "grad_norm": 1.1887514312279568, + "learning_rate": 2.9991144104564883e-06, + "loss": 0.4144, + "step": 3481 + }, + { + "epoch": 3.341814690350456, + "grad_norm": 1.4304855176373867, + "learning_rate": 2.9960448295795667e-06, + "loss": 0.573, + "step": 3482 + }, + { + "epoch": 3.342774843975036, + "grad_norm": 1.1859827630645317, + "learning_rate": 2.9929761482567017e-06, + "loss": 0.5282, + "step": 3483 + }, + { + "epoch": 3.343734997599616, + "grad_norm": 1.1700470288724143, + "learning_rate": 2.9899083678653885e-06, + "loss": 0.4676, + "step": 3484 + }, + { + "epoch": 3.344695151224196, + "grad_norm": 1.3223092752673435, + "learning_rate": 2.9868414897827236e-06, + "loss": 0.3924, + "step": 3485 + }, + { + "epoch": 3.345655304848776, + "grad_norm": 0.993114079383793, + "learning_rate": 2.983775515385393e-06, + "loss": 0.5712, + "step": 3486 + }, + { + "epoch": 3.3466154584733556, + "grad_norm": 1.1411752022316926, + "learning_rate": 2.9807104460496807e-06, + "loss": 0.4888, + "step": 3487 + }, + { + "epoch": 3.347575612097936, + "grad_norm": 1.0269578683448715, + "learning_rate": 2.9776462831514664e-06, + "loss": 0.5026, + "step": 3488 + }, + { + "epoch": 3.3485357657225157, + "grad_norm": 1.0339014426150241, + "learning_rate": 2.9745830280662157e-06, + "loss": 0.4463, + "step": 3489 + }, + { + "epoch": 3.3494959193470955, + "grad_norm": 0.8775852931129492, + "learning_rate": 2.971520682168996e-06, + "loss": 0.4852, + "step": 3490 + }, + { + "epoch": 3.3504560729716752, + "grad_norm": 1.0932945047678582, + "learning_rate": 2.968459246834455e-06, + "loss": 0.4676, + "step": 3491 + }, + { + "epoch": 3.3514162265962555, + "grad_norm": 1.1250239389917458, + "learning_rate": 2.9653987234368443e-06, + "loss": 0.5186, + "step": 3492 + }, + { + "epoch": 3.3523763802208353, + "grad_norm": 0.9436454989420195, + "learning_rate": 2.9623391133499986e-06, + "loss": 0.4263, + "step": 3493 + }, + { + "epoch": 3.353336533845415, + "grad_norm": 1.0609472334448762, + "learning_rate": 2.959280417947342e-06, + "loss": 0.5188, + "step": 3494 + }, + { + "epoch": 3.3542966874699953, + "grad_norm": 1.202275563516378, + "learning_rate": 2.9562226386018935e-06, + "loss": 0.2793, + "step": 3495 + }, + { + "epoch": 3.355256841094575, + "grad_norm": 1.7083387156092242, + "learning_rate": 2.953165776686254e-06, + "loss": 0.5163, + "step": 3496 + }, + { + "epoch": 3.356216994719155, + "grad_norm": 1.1409952406855968, + "learning_rate": 2.9501098335726188e-06, + "loss": 0.5052, + "step": 3497 + }, + { + "epoch": 3.357177148343735, + "grad_norm": 1.186020755033467, + "learning_rate": 2.947054810632768e-06, + "loss": 0.4582, + "step": 3498 + }, + { + "epoch": 3.358137301968315, + "grad_norm": 1.4150439410912465, + "learning_rate": 2.9440007092380663e-06, + "loss": 0.4449, + "step": 3499 + }, + { + "epoch": 3.3590974555928947, + "grad_norm": 1.356711673122678, + "learning_rate": 2.9409475307594683e-06, + "loss": 0.491, + "step": 3500 + }, + { + "epoch": 3.360057609217475, + "grad_norm": 1.0576636480136845, + "learning_rate": 2.937895276567516e-06, + "loss": 0.5146, + "step": 3501 + }, + { + "epoch": 3.361017762842055, + "grad_norm": 0.9253348342532148, + "learning_rate": 2.934843948032329e-06, + "loss": 0.3877, + "step": 3502 + }, + { + "epoch": 3.3619779164666346, + "grad_norm": 1.0595590531486225, + "learning_rate": 2.931793546523622e-06, + "loss": 0.4221, + "step": 3503 + }, + { + "epoch": 3.3629380700912144, + "grad_norm": 1.5780005221790405, + "learning_rate": 2.9287440734106802e-06, + "loss": 0.3656, + "step": 3504 + }, + { + "epoch": 3.3638982237157946, + "grad_norm": 1.2094603353904014, + "learning_rate": 2.9256955300623847e-06, + "loss": 0.5271, + "step": 3505 + }, + { + "epoch": 3.3648583773403744, + "grad_norm": 1.1825058315142765, + "learning_rate": 2.9226479178471933e-06, + "loss": 0.3477, + "step": 3506 + }, + { + "epoch": 3.365818530964954, + "grad_norm": 1.3214146414963868, + "learning_rate": 2.919601238133145e-06, + "loss": 0.4159, + "step": 3507 + }, + { + "epoch": 3.3667786845895344, + "grad_norm": 1.1719105500897151, + "learning_rate": 2.916555492287865e-06, + "loss": 0.4692, + "step": 3508 + }, + { + "epoch": 3.3677388382141142, + "grad_norm": 1.8800779319989285, + "learning_rate": 2.913510681678552e-06, + "loss": 0.4837, + "step": 3509 + }, + { + "epoch": 3.368698991838694, + "grad_norm": 1.2531946213367224, + "learning_rate": 2.9104668076719876e-06, + "loss": 0.4717, + "step": 3510 + }, + { + "epoch": 3.3696591454632743, + "grad_norm": 1.1105983974295732, + "learning_rate": 2.9074238716345426e-06, + "loss": 0.3971, + "step": 3511 + }, + { + "epoch": 3.370619299087854, + "grad_norm": 0.8628587269532157, + "learning_rate": 2.9043818749321506e-06, + "loss": 0.4435, + "step": 3512 + }, + { + "epoch": 3.371579452712434, + "grad_norm": 1.2458667397479701, + "learning_rate": 2.9013408189303337e-06, + "loss": 0.4599, + "step": 3513 + }, + { + "epoch": 3.372539606337014, + "grad_norm": 1.2264588368349105, + "learning_rate": 2.8983007049941904e-06, + "loss": 0.5525, + "step": 3514 + }, + { + "epoch": 3.373499759961594, + "grad_norm": 1.4314881554786696, + "learning_rate": 2.8952615344883938e-06, + "loss": 0.4697, + "step": 3515 + }, + { + "epoch": 3.3744599135861737, + "grad_norm": 0.9575411876956582, + "learning_rate": 2.8922233087771988e-06, + "loss": 0.4158, + "step": 3516 + }, + { + "epoch": 3.3754200672107535, + "grad_norm": 1.2465922949189534, + "learning_rate": 2.889186029224428e-06, + "loss": 0.5241, + "step": 3517 + }, + { + "epoch": 3.3763802208353337, + "grad_norm": 1.0781670235659189, + "learning_rate": 2.8861496971934843e-06, + "loss": 0.4696, + "step": 3518 + }, + { + "epoch": 3.3773403744599135, + "grad_norm": 1.1216142983619832, + "learning_rate": 2.883114314047346e-06, + "loss": 0.5838, + "step": 3519 + }, + { + "epoch": 3.3783005280844933, + "grad_norm": 1.1020393649253364, + "learning_rate": 2.880079881148564e-06, + "loss": 0.4021, + "step": 3520 + }, + { + "epoch": 3.3792606817090736, + "grad_norm": 1.1408753452661533, + "learning_rate": 2.8770463998592645e-06, + "loss": 0.4518, + "step": 3521 + }, + { + "epoch": 3.3802208353336534, + "grad_norm": 1.4476366523641173, + "learning_rate": 2.8740138715411413e-06, + "loss": 0.535, + "step": 3522 + }, + { + "epoch": 3.381180988958233, + "grad_norm": 1.0947910518321444, + "learning_rate": 2.8709822975554636e-06, + "loss": 0.4504, + "step": 3523 + }, + { + "epoch": 3.3821411425828134, + "grad_norm": 1.083311343722618, + "learning_rate": 2.8679516792630783e-06, + "loss": 0.4013, + "step": 3524 + }, + { + "epoch": 3.383101296207393, + "grad_norm": 1.1621997871934076, + "learning_rate": 2.864922018024392e-06, + "loss": 0.4416, + "step": 3525 + }, + { + "epoch": 3.384061449831973, + "grad_norm": 0.8587905880392465, + "learning_rate": 2.8618933151993895e-06, + "loss": 0.4537, + "step": 3526 + }, + { + "epoch": 3.3850216034565532, + "grad_norm": 0.8986686761045902, + "learning_rate": 2.858865572147622e-06, + "loss": 0.5098, + "step": 3527 + }, + { + "epoch": 3.385981757081133, + "grad_norm": 1.5340809367773565, + "learning_rate": 2.8558387902282124e-06, + "loss": 0.5056, + "step": 3528 + }, + { + "epoch": 3.386941910705713, + "grad_norm": 0.9800974216612418, + "learning_rate": 2.852812970799852e-06, + "loss": 0.4697, + "step": 3529 + }, + { + "epoch": 3.3879020643302926, + "grad_norm": 1.3399154574672762, + "learning_rate": 2.849788115220794e-06, + "loss": 0.5324, + "step": 3530 + }, + { + "epoch": 3.388862217954873, + "grad_norm": 1.2917476442777003, + "learning_rate": 2.846764224848867e-06, + "loss": 0.4676, + "step": 3531 + }, + { + "epoch": 3.3898223715794527, + "grad_norm": 1.5463920595501937, + "learning_rate": 2.843741301041463e-06, + "loss": 0.4506, + "step": 3532 + }, + { + "epoch": 3.390782525204033, + "grad_norm": 1.0781539145409702, + "learning_rate": 2.8407193451555398e-06, + "loss": 0.4419, + "step": 3533 + }, + { + "epoch": 3.3917426788286127, + "grad_norm": 1.0027066731044258, + "learning_rate": 2.8376983585476226e-06, + "loss": 0.4063, + "step": 3534 + }, + { + "epoch": 3.3927028324531925, + "grad_norm": 0.8821407393548042, + "learning_rate": 2.834678342573797e-06, + "loss": 0.5158, + "step": 3535 + }, + { + "epoch": 3.3936629860777723, + "grad_norm": 1.1939673584672799, + "learning_rate": 2.8316592985897173e-06, + "loss": 0.463, + "step": 3536 + }, + { + "epoch": 3.3946231397023525, + "grad_norm": 1.1480599497638293, + "learning_rate": 2.8286412279506006e-06, + "loss": 0.445, + "step": 3537 + }, + { + "epoch": 3.3955832933269323, + "grad_norm": 0.9313188281494226, + "learning_rate": 2.8256241320112253e-06, + "loss": 0.5166, + "step": 3538 + }, + { + "epoch": 3.396543446951512, + "grad_norm": 1.006547046161028, + "learning_rate": 2.822608012125937e-06, + "loss": 0.5434, + "step": 3539 + }, + { + "epoch": 3.3975036005760924, + "grad_norm": 1.201997757291289, + "learning_rate": 2.8195928696486345e-06, + "loss": 0.4984, + "step": 3540 + }, + { + "epoch": 3.398463754200672, + "grad_norm": 0.9038008372857832, + "learning_rate": 2.8165787059327844e-06, + "loss": 0.5657, + "step": 3541 + }, + { + "epoch": 3.399423907825252, + "grad_norm": 1.1774393610346174, + "learning_rate": 2.8135655223314166e-06, + "loss": 0.3767, + "step": 3542 + }, + { + "epoch": 3.4003840614498317, + "grad_norm": 1.12599849774403, + "learning_rate": 2.810553320197113e-06, + "loss": 0.5037, + "step": 3543 + }, + { + "epoch": 3.401344215074412, + "grad_norm": 1.2062143158323828, + "learning_rate": 2.8075421008820197e-06, + "loss": 0.4462, + "step": 3544 + }, + { + "epoch": 3.4023043686989918, + "grad_norm": 1.2403997832602278, + "learning_rate": 2.804531865737842e-06, + "loss": 0.4407, + "step": 3545 + }, + { + "epoch": 3.403264522323572, + "grad_norm": 1.850016546542534, + "learning_rate": 2.8015226161158403e-06, + "loss": 0.4886, + "step": 3546 + }, + { + "epoch": 3.404224675948152, + "grad_norm": 1.1404929419636376, + "learning_rate": 2.798514353366838e-06, + "loss": 0.4044, + "step": 3547 + }, + { + "epoch": 3.4051848295727316, + "grad_norm": 1.3213709708007022, + "learning_rate": 2.7955070788412085e-06, + "loss": 0.5141, + "step": 3548 + }, + { + "epoch": 3.4061449831973114, + "grad_norm": 0.8699083415858886, + "learning_rate": 2.7925007938888867e-06, + "loss": 0.4399, + "step": 3549 + }, + { + "epoch": 3.4071051368218916, + "grad_norm": 1.3754971584958042, + "learning_rate": 2.789495499859361e-06, + "loss": 0.4446, + "step": 3550 + }, + { + "epoch": 3.4080652904464714, + "grad_norm": 1.5106588377195465, + "learning_rate": 2.786491198101677e-06, + "loss": 0.5024, + "step": 3551 + }, + { + "epoch": 3.4090254440710512, + "grad_norm": 1.5740256115254103, + "learning_rate": 2.783487889964435e-06, + "loss": 0.5005, + "step": 3552 + }, + { + "epoch": 3.4099855976956315, + "grad_norm": 1.4239322372170375, + "learning_rate": 2.7804855767957833e-06, + "loss": 0.4261, + "step": 3553 + }, + { + "epoch": 3.4109457513202113, + "grad_norm": 0.7441920668705134, + "learning_rate": 2.7774842599434294e-06, + "loss": 0.5355, + "step": 3554 + }, + { + "epoch": 3.411905904944791, + "grad_norm": 1.1964170644097079, + "learning_rate": 2.7744839407546374e-06, + "loss": 0.5028, + "step": 3555 + }, + { + "epoch": 3.412866058569371, + "grad_norm": 0.8917016245174008, + "learning_rate": 2.7714846205762134e-06, + "loss": 0.4768, + "step": 3556 + }, + { + "epoch": 3.413826212193951, + "grad_norm": 1.0888707464088432, + "learning_rate": 2.7684863007545227e-06, + "loss": 0.5231, + "step": 3557 + }, + { + "epoch": 3.414786365818531, + "grad_norm": 1.0868139743790344, + "learning_rate": 2.765488982635477e-06, + "loss": 0.5151, + "step": 3558 + }, + { + "epoch": 3.415746519443111, + "grad_norm": 0.9436745273446261, + "learning_rate": 2.7624926675645415e-06, + "loss": 0.4133, + "step": 3559 + }, + { + "epoch": 3.416706673067691, + "grad_norm": 1.0352478830995882, + "learning_rate": 2.759497356886731e-06, + "loss": 0.5861, + "step": 3560 + }, + { + "epoch": 3.4176668266922707, + "grad_norm": 1.0762675806754036, + "learning_rate": 2.7565030519466063e-06, + "loss": 0.4028, + "step": 3561 + }, + { + "epoch": 3.4186269803168505, + "grad_norm": 0.8724198562506894, + "learning_rate": 2.75350975408828e-06, + "loss": 0.4266, + "step": 3562 + }, + { + "epoch": 3.4195871339414308, + "grad_norm": 1.2616615183996187, + "learning_rate": 2.7505174646554112e-06, + "loss": 0.4518, + "step": 3563 + }, + { + "epoch": 3.4205472875660106, + "grad_norm": 1.324485618575122, + "learning_rate": 2.7475261849912073e-06, + "loss": 0.4463, + "step": 3564 + }, + { + "epoch": 3.4215074411905904, + "grad_norm": 1.0485211087613273, + "learning_rate": 2.744535916438423e-06, + "loss": 0.539, + "step": 3565 + }, + { + "epoch": 3.4224675948151706, + "grad_norm": 1.3669013683457083, + "learning_rate": 2.7415466603393538e-06, + "loss": 0.6324, + "step": 3566 + }, + { + "epoch": 3.4234277484397504, + "grad_norm": 0.9702366821080235, + "learning_rate": 2.738558418035846e-06, + "loss": 0.517, + "step": 3567 + }, + { + "epoch": 3.42438790206433, + "grad_norm": 1.0843548366284068, + "learning_rate": 2.7355711908692942e-06, + "loss": 0.3979, + "step": 3568 + }, + { + "epoch": 3.42534805568891, + "grad_norm": 1.0300795464968182, + "learning_rate": 2.732584980180628e-06, + "loss": 0.3735, + "step": 3569 + }, + { + "epoch": 3.4263082093134902, + "grad_norm": 1.1337504956207816, + "learning_rate": 2.7295997873103286e-06, + "loss": 0.3978, + "step": 3570 + }, + { + "epoch": 3.42726836293807, + "grad_norm": 1.0188041814822386, + "learning_rate": 2.726615613598411e-06, + "loss": 0.4734, + "step": 3571 + }, + { + "epoch": 3.4282285165626503, + "grad_norm": 1.8133667201452484, + "learning_rate": 2.7236324603844467e-06, + "loss": 0.4625, + "step": 3572 + }, + { + "epoch": 3.42918867018723, + "grad_norm": 2.04669729109423, + "learning_rate": 2.7206503290075388e-06, + "loss": 0.5092, + "step": 3573 + }, + { + "epoch": 3.43014882381181, + "grad_norm": 1.285179302856492, + "learning_rate": 2.7176692208063317e-06, + "loss": 0.4538, + "step": 3574 + }, + { + "epoch": 3.4311089774363897, + "grad_norm": 1.0888419094057504, + "learning_rate": 2.7146891371190144e-06, + "loss": 0.4493, + "step": 3575 + }, + { + "epoch": 3.43206913106097, + "grad_norm": 1.1711896679652405, + "learning_rate": 2.7117100792833144e-06, + "loss": 0.493, + "step": 3576 + }, + { + "epoch": 3.4330292846855497, + "grad_norm": 1.140479659389972, + "learning_rate": 2.7087320486364997e-06, + "loss": 0.3899, + "step": 3577 + }, + { + "epoch": 3.4339894383101295, + "grad_norm": 1.1330798335546821, + "learning_rate": 2.7057550465153782e-06, + "loss": 0.4961, + "step": 3578 + }, + { + "epoch": 3.4349495919347097, + "grad_norm": 1.4964408783536312, + "learning_rate": 2.7027790742562905e-06, + "loss": 0.511, + "step": 3579 + }, + { + "epoch": 3.4359097455592895, + "grad_norm": 1.2967733590915969, + "learning_rate": 2.6998041331951196e-06, + "loss": 0.4862, + "step": 3580 + }, + { + "epoch": 3.4368698991838693, + "grad_norm": 1.2047896517623808, + "learning_rate": 2.6968302246672863e-06, + "loss": 0.4532, + "step": 3581 + }, + { + "epoch": 3.437830052808449, + "grad_norm": 1.1207257158205917, + "learning_rate": 2.693857350007746e-06, + "loss": 0.455, + "step": 3582 + }, + { + "epoch": 3.4387902064330294, + "grad_norm": 1.217969904538037, + "learning_rate": 2.6908855105509912e-06, + "loss": 0.4445, + "step": 3583 + }, + { + "epoch": 3.439750360057609, + "grad_norm": 1.2049534130896251, + "learning_rate": 2.6879147076310457e-06, + "loss": 0.4064, + "step": 3584 + }, + { + "epoch": 3.4407105136821894, + "grad_norm": 1.3761988292840837, + "learning_rate": 2.6849449425814713e-06, + "loss": 0.4026, + "step": 3585 + }, + { + "epoch": 3.441670667306769, + "grad_norm": 0.9415572000589764, + "learning_rate": 2.6819762167353695e-06, + "loss": 0.383, + "step": 3586 + }, + { + "epoch": 3.442630820931349, + "grad_norm": 0.9470684110040503, + "learning_rate": 2.6790085314253645e-06, + "loss": 0.43, + "step": 3587 + }, + { + "epoch": 3.4435909745559288, + "grad_norm": 0.894013486586558, + "learning_rate": 2.6760418879836186e-06, + "loss": 0.4578, + "step": 3588 + }, + { + "epoch": 3.444551128180509, + "grad_norm": 0.8156437597110491, + "learning_rate": 2.673076287741828e-06, + "loss": 0.4178, + "step": 3589 + }, + { + "epoch": 3.445511281805089, + "grad_norm": 1.0825128113462217, + "learning_rate": 2.670111732031218e-06, + "loss": 0.4715, + "step": 3590 + }, + { + "epoch": 3.4464714354296686, + "grad_norm": 1.0144452624565126, + "learning_rate": 2.6671482221825484e-06, + "loss": 0.5427, + "step": 3591 + }, + { + "epoch": 3.447431589054249, + "grad_norm": 1.0665060436734815, + "learning_rate": 2.6641857595261035e-06, + "loss": 0.4447, + "step": 3592 + }, + { + "epoch": 3.4483917426788286, + "grad_norm": 0.9795414216047997, + "learning_rate": 2.661224345391703e-06, + "loss": 0.4444, + "step": 3593 + }, + { + "epoch": 3.4493518963034084, + "grad_norm": 0.8353413950195184, + "learning_rate": 2.658263981108693e-06, + "loss": 0.5798, + "step": 3594 + }, + { + "epoch": 3.4503120499279882, + "grad_norm": 0.8687455116406125, + "learning_rate": 2.65530466800595e-06, + "loss": 0.5348, + "step": 3595 + }, + { + "epoch": 3.4512722035525685, + "grad_norm": 1.148995894038912, + "learning_rate": 2.652346407411881e-06, + "loss": 0.3702, + "step": 3596 + }, + { + "epoch": 3.4522323571771483, + "grad_norm": 1.2989800722974758, + "learning_rate": 2.649389200654412e-06, + "loss": 0.4112, + "step": 3597 + }, + { + "epoch": 3.4531925108017285, + "grad_norm": 0.9463725991275451, + "learning_rate": 2.6464330490610026e-06, + "loss": 0.4448, + "step": 3598 + }, + { + "epoch": 3.4541526644263083, + "grad_norm": 1.318494074442644, + "learning_rate": 2.6434779539586432e-06, + "loss": 0.5126, + "step": 3599 + }, + { + "epoch": 3.455112818050888, + "grad_norm": 1.3358485591347307, + "learning_rate": 2.640523916673838e-06, + "loss": 0.4431, + "step": 3600 + }, + { + "epoch": 3.456072971675468, + "grad_norm": 1.4512534259842509, + "learning_rate": 2.6375709385326255e-06, + "loss": 0.4406, + "step": 3601 + }, + { + "epoch": 3.457033125300048, + "grad_norm": 1.1449355325344546, + "learning_rate": 2.6346190208605653e-06, + "loss": 0.4695, + "step": 3602 + }, + { + "epoch": 3.457993278924628, + "grad_norm": 0.9559669934907598, + "learning_rate": 2.6316681649827423e-06, + "loss": 0.4964, + "step": 3603 + }, + { + "epoch": 3.4589534325492077, + "grad_norm": 1.201264073736653, + "learning_rate": 2.6287183722237663e-06, + "loss": 0.3984, + "step": 3604 + }, + { + "epoch": 3.459913586173788, + "grad_norm": 1.096836226946159, + "learning_rate": 2.625769643907763e-06, + "loss": 0.5035, + "step": 3605 + }, + { + "epoch": 3.4608737397983678, + "grad_norm": 0.7278545545970976, + "learning_rate": 2.622821981358388e-06, + "loss": 0.5073, + "step": 3606 + }, + { + "epoch": 3.4618338934229476, + "grad_norm": 1.3266279163313695, + "learning_rate": 2.6198753858988147e-06, + "loss": 0.4638, + "step": 3607 + }, + { + "epoch": 3.4627940470475274, + "grad_norm": 0.8010194816184371, + "learning_rate": 2.616929858851739e-06, + "loss": 0.3906, + "step": 3608 + }, + { + "epoch": 3.4637542006721076, + "grad_norm": 1.160469371575079, + "learning_rate": 2.613985401539377e-06, + "loss": 0.4674, + "step": 3609 + }, + { + "epoch": 3.4647143542966874, + "grad_norm": 1.385240494381929, + "learning_rate": 2.6110420152834614e-06, + "loss": 0.6116, + "step": 3610 + }, + { + "epoch": 3.4656745079212676, + "grad_norm": 1.4254500913885508, + "learning_rate": 2.6080997014052474e-06, + "loss": 0.4868, + "step": 3611 + }, + { + "epoch": 3.4666346615458474, + "grad_norm": 1.0313280287612518, + "learning_rate": 2.6051584612255132e-06, + "loss": 0.4166, + "step": 3612 + }, + { + "epoch": 3.4675948151704272, + "grad_norm": 0.9944099859028148, + "learning_rate": 2.6022182960645436e-06, + "loss": 0.4879, + "step": 3613 + }, + { + "epoch": 3.468554968795007, + "grad_norm": 0.9995198484098955, + "learning_rate": 2.5992792072421526e-06, + "loss": 0.4761, + "step": 3614 + }, + { + "epoch": 3.4695151224195873, + "grad_norm": 1.216734458299888, + "learning_rate": 2.596341196077659e-06, + "loss": 0.5055, + "step": 3615 + }, + { + "epoch": 3.470475276044167, + "grad_norm": 1.246247200010548, + "learning_rate": 2.593404263889911e-06, + "loss": 0.5008, + "step": 3616 + }, + { + "epoch": 3.471435429668747, + "grad_norm": 0.9976808037461659, + "learning_rate": 2.5904684119972656e-06, + "loss": 0.5077, + "step": 3617 + }, + { + "epoch": 3.472395583293327, + "grad_norm": 0.8269283769717857, + "learning_rate": 2.587533641717592e-06, + "loss": 0.4821, + "step": 3618 + }, + { + "epoch": 3.473355736917907, + "grad_norm": 1.2400549875225575, + "learning_rate": 2.5845999543682787e-06, + "loss": 0.5868, + "step": 3619 + }, + { + "epoch": 3.4743158905424867, + "grad_norm": 0.9329814175183276, + "learning_rate": 2.581667351266226e-06, + "loss": 0.6003, + "step": 3620 + }, + { + "epoch": 3.475276044167067, + "grad_norm": 1.3338791789490658, + "learning_rate": 2.5787358337278493e-06, + "loss": 0.4371, + "step": 3621 + }, + { + "epoch": 3.4762361977916467, + "grad_norm": 1.1511633734554696, + "learning_rate": 2.575805403069077e-06, + "loss": 0.4362, + "step": 3622 + }, + { + "epoch": 3.4771963514162265, + "grad_norm": 1.1202526138333042, + "learning_rate": 2.5728760606053443e-06, + "loss": 0.4908, + "step": 3623 + }, + { + "epoch": 3.4781565050408068, + "grad_norm": 1.249918071885655, + "learning_rate": 2.5699478076516034e-06, + "loss": 0.4459, + "step": 3624 + }, + { + "epoch": 3.4791166586653866, + "grad_norm": 0.9026035841529765, + "learning_rate": 2.5670206455223156e-06, + "loss": 0.536, + "step": 3625 + }, + { + "epoch": 3.4800768122899663, + "grad_norm": 1.1902872385852585, + "learning_rate": 2.5640945755314527e-06, + "loss": 0.4952, + "step": 3626 + }, + { + "epoch": 3.481036965914546, + "grad_norm": 1.0066852241962887, + "learning_rate": 2.5611695989924988e-06, + "loss": 0.4986, + "step": 3627 + }, + { + "epoch": 3.4819971195391264, + "grad_norm": 0.8493448562993611, + "learning_rate": 2.5582457172184405e-06, + "loss": 0.4542, + "step": 3628 + }, + { + "epoch": 3.482957273163706, + "grad_norm": 0.9521082941092199, + "learning_rate": 2.5553229315217763e-06, + "loss": 0.3476, + "step": 3629 + }, + { + "epoch": 3.483917426788286, + "grad_norm": 1.140189369962596, + "learning_rate": 2.5524012432145203e-06, + "loss": 0.5449, + "step": 3630 + }, + { + "epoch": 3.484877580412866, + "grad_norm": 0.9634567981887172, + "learning_rate": 2.5494806536081806e-06, + "loss": 0.4575, + "step": 3631 + }, + { + "epoch": 3.485837734037446, + "grad_norm": 0.8307670440375307, + "learning_rate": 2.5465611640137813e-06, + "loss": 0.4691, + "step": 3632 + }, + { + "epoch": 3.486797887662026, + "grad_norm": 1.2732339891636295, + "learning_rate": 2.543642775741849e-06, + "loss": 0.4743, + "step": 3633 + }, + { + "epoch": 3.487758041286606, + "grad_norm": 1.397784544861311, + "learning_rate": 2.540725490102418e-06, + "loss": 0.5432, + "step": 3634 + }, + { + "epoch": 3.488718194911186, + "grad_norm": 0.9880987144603872, + "learning_rate": 2.5378093084050287e-06, + "loss": 0.5647, + "step": 3635 + }, + { + "epoch": 3.4896783485357656, + "grad_norm": 0.8861042098478457, + "learning_rate": 2.534894231958719e-06, + "loss": 0.5291, + "step": 3636 + }, + { + "epoch": 3.490638502160346, + "grad_norm": 0.9065219243847623, + "learning_rate": 2.531980262072038e-06, + "loss": 0.4777, + "step": 3637 + }, + { + "epoch": 3.4915986557849257, + "grad_norm": 1.1297295787094745, + "learning_rate": 2.5290674000530365e-06, + "loss": 0.5404, + "step": 3638 + }, + { + "epoch": 3.4925588094095055, + "grad_norm": 1.646709559764397, + "learning_rate": 2.526155647209265e-06, + "loss": 0.4869, + "step": 3639 + }, + { + "epoch": 3.4935189630340853, + "grad_norm": 0.7720437600850697, + "learning_rate": 2.523245004847783e-06, + "loss": 0.5778, + "step": 3640 + }, + { + "epoch": 3.4944791166586655, + "grad_norm": 0.9884900889150412, + "learning_rate": 2.520335474275141e-06, + "loss": 0.4142, + "step": 3641 + }, + { + "epoch": 3.4954392702832453, + "grad_norm": 1.124525370980486, + "learning_rate": 2.5174270567973957e-06, + "loss": 0.5587, + "step": 3642 + }, + { + "epoch": 3.496399423907825, + "grad_norm": 0.9719887183464196, + "learning_rate": 2.5145197537201116e-06, + "loss": 0.5514, + "step": 3643 + }, + { + "epoch": 3.4973595775324053, + "grad_norm": 1.2104387401344738, + "learning_rate": 2.51161356634834e-06, + "loss": 0.5069, + "step": 3644 + }, + { + "epoch": 3.498319731156985, + "grad_norm": 1.2346000448649872, + "learning_rate": 2.5087084959866403e-06, + "loss": 0.4351, + "step": 3645 + }, + { + "epoch": 3.499279884781565, + "grad_norm": 1.099044269053787, + "learning_rate": 2.5058045439390634e-06, + "loss": 0.491, + "step": 3646 + }, + { + "epoch": 3.5002400384061447, + "grad_norm": 1.4946661581040868, + "learning_rate": 2.502901711509167e-06, + "loss": 0.4149, + "step": 3647 + }, + { + "epoch": 3.501200192030725, + "grad_norm": 1.227057661512799, + "learning_rate": 2.5000000000000015e-06, + "loss": 0.5229, + "step": 3648 + }, + { + "epoch": 3.5021603456553048, + "grad_norm": 1.2370509801878777, + "learning_rate": 2.4970994107141106e-06, + "loss": 0.4831, + "step": 3649 + }, + { + "epoch": 3.503120499279885, + "grad_norm": 1.2958315162178096, + "learning_rate": 2.4941999449535393e-06, + "loss": 0.3693, + "step": 3650 + }, + { + "epoch": 3.504080652904465, + "grad_norm": 1.0522251409633594, + "learning_rate": 2.491301604019828e-06, + "loss": 0.5309, + "step": 3651 + }, + { + "epoch": 3.5050408065290446, + "grad_norm": 1.2251693069715273, + "learning_rate": 2.4884043892140096e-06, + "loss": 0.4234, + "step": 3652 + }, + { + "epoch": 3.5060009601536244, + "grad_norm": 1.056679063466086, + "learning_rate": 2.4855083018366154e-06, + "loss": 0.4025, + "step": 3653 + }, + { + "epoch": 3.5069611137782046, + "grad_norm": 1.0109892189050866, + "learning_rate": 2.4826133431876644e-06, + "loss": 0.4746, + "step": 3654 + }, + { + "epoch": 3.5079212674027844, + "grad_norm": 0.8900546374984372, + "learning_rate": 2.479719514566672e-06, + "loss": 0.4223, + "step": 3655 + }, + { + "epoch": 3.5088814210273642, + "grad_norm": 1.1033470328591162, + "learning_rate": 2.4768268172726535e-06, + "loss": 0.472, + "step": 3656 + }, + { + "epoch": 3.5098415746519445, + "grad_norm": 1.270969298574359, + "learning_rate": 2.473935252604103e-06, + "loss": 0.5337, + "step": 3657 + }, + { + "epoch": 3.5108017282765243, + "grad_norm": 1.479006407731475, + "learning_rate": 2.4710448218590176e-06, + "loss": 0.5184, + "step": 3658 + }, + { + "epoch": 3.511761881901104, + "grad_norm": 1.1883063504227078, + "learning_rate": 2.468155526334875e-06, + "loss": 0.5456, + "step": 3659 + }, + { + "epoch": 3.512722035525684, + "grad_norm": 1.0164127178929956, + "learning_rate": 2.465267367328655e-06, + "loss": 0.4789, + "step": 3660 + }, + { + "epoch": 3.513682189150264, + "grad_norm": 1.1188398770666201, + "learning_rate": 2.46238034613682e-06, + "loss": 0.4018, + "step": 3661 + }, + { + "epoch": 3.514642342774844, + "grad_norm": 1.0569981650165516, + "learning_rate": 2.4594944640553207e-06, + "loss": 0.4344, + "step": 3662 + }, + { + "epoch": 3.515602496399424, + "grad_norm": 1.3233370416968462, + "learning_rate": 2.4566097223795997e-06, + "loss": 0.527, + "step": 3663 + }, + { + "epoch": 3.516562650024004, + "grad_norm": 1.1378472162358964, + "learning_rate": 2.453726122404587e-06, + "loss": 0.5237, + "step": 3664 + }, + { + "epoch": 3.5175228036485837, + "grad_norm": 1.0450132913587404, + "learning_rate": 2.4508436654247004e-06, + "loss": 0.4229, + "step": 3665 + }, + { + "epoch": 3.5184829572731635, + "grad_norm": 0.9468624461124091, + "learning_rate": 2.4479623527338447e-06, + "loss": 0.5138, + "step": 3666 + }, + { + "epoch": 3.5194431108977438, + "grad_norm": 1.1454995223310345, + "learning_rate": 2.4450821856254074e-06, + "loss": 0.4455, + "step": 3667 + }, + { + "epoch": 3.5204032645223235, + "grad_norm": 1.2594858516678467, + "learning_rate": 2.442203165392266e-06, + "loss": 0.3976, + "step": 3668 + }, + { + "epoch": 3.5213634181469033, + "grad_norm": 1.6642733147102062, + "learning_rate": 2.439325293326783e-06, + "loss": 0.3881, + "step": 3669 + }, + { + "epoch": 3.5223235717714836, + "grad_norm": 1.1145416415512275, + "learning_rate": 2.4364485707208037e-06, + "loss": 0.5226, + "step": 3670 + }, + { + "epoch": 3.5232837253960634, + "grad_norm": 1.3938203305124055, + "learning_rate": 2.433572998865659e-06, + "loss": 0.4091, + "step": 3671 + }, + { + "epoch": 3.524243879020643, + "grad_norm": 0.9522141730008076, + "learning_rate": 2.430698579052161e-06, + "loss": 0.5209, + "step": 3672 + }, + { + "epoch": 3.525204032645223, + "grad_norm": 1.0579883556910796, + "learning_rate": 2.4278253125706044e-06, + "loss": 0.5121, + "step": 3673 + }, + { + "epoch": 3.526164186269803, + "grad_norm": 1.0666069425645939, + "learning_rate": 2.424953200710773e-06, + "loss": 0.468, + "step": 3674 + }, + { + "epoch": 3.527124339894383, + "grad_norm": 1.124981702042489, + "learning_rate": 2.4220822447619223e-06, + "loss": 0.4579, + "step": 3675 + }, + { + "epoch": 3.5280844935189632, + "grad_norm": 1.1505524021429214, + "learning_rate": 2.419212446012796e-06, + "loss": 0.4537, + "step": 3676 + }, + { + "epoch": 3.529044647143543, + "grad_norm": 0.9687320319338322, + "learning_rate": 2.416343805751615e-06, + "loss": 0.4503, + "step": 3677 + }, + { + "epoch": 3.530004800768123, + "grad_norm": 1.3101841403266583, + "learning_rate": 2.4134763252660825e-06, + "loss": 0.386, + "step": 3678 + }, + { + "epoch": 3.5309649543927026, + "grad_norm": 1.2122305542938274, + "learning_rate": 2.4106100058433814e-06, + "loss": 0.4864, + "step": 3679 + }, + { + "epoch": 3.531925108017283, + "grad_norm": 1.4978240894195236, + "learning_rate": 2.4077448487701667e-06, + "loss": 0.5496, + "step": 3680 + }, + { + "epoch": 3.5328852616418627, + "grad_norm": 1.3428200095632699, + "learning_rate": 2.4048808553325804e-06, + "loss": 0.5052, + "step": 3681 + }, + { + "epoch": 3.5338454152664425, + "grad_norm": 1.039762695895337, + "learning_rate": 2.4020180268162374e-06, + "loss": 0.565, + "step": 3682 + }, + { + "epoch": 3.5348055688910227, + "grad_norm": 1.0547541209262872, + "learning_rate": 2.3991563645062317e-06, + "loss": 0.4556, + "step": 3683 + }, + { + "epoch": 3.5357657225156025, + "grad_norm": 1.1631940207140978, + "learning_rate": 2.3962958696871343e-06, + "loss": 0.4225, + "step": 3684 + }, + { + "epoch": 3.5367258761401823, + "grad_norm": 1.2427090021569067, + "learning_rate": 2.393436543642987e-06, + "loss": 0.3593, + "step": 3685 + }, + { + "epoch": 3.537686029764762, + "grad_norm": 1.2190538515419043, + "learning_rate": 2.3905783876573108e-06, + "loss": 0.508, + "step": 3686 + }, + { + "epoch": 3.5386461833893423, + "grad_norm": 1.1922326644684484, + "learning_rate": 2.387721403013107e-06, + "loss": 0.5042, + "step": 3687 + }, + { + "epoch": 3.539606337013922, + "grad_norm": 1.1304810096194196, + "learning_rate": 2.3848655909928392e-06, + "loss": 0.5237, + "step": 3688 + }, + { + "epoch": 3.5405664906385024, + "grad_norm": 1.019042176877277, + "learning_rate": 2.382010952878456e-06, + "loss": 0.4779, + "step": 3689 + }, + { + "epoch": 3.541526644263082, + "grad_norm": 1.039931316432492, + "learning_rate": 2.379157489951367e-06, + "loss": 0.5413, + "step": 3690 + }, + { + "epoch": 3.542486797887662, + "grad_norm": 1.1195582239210486, + "learning_rate": 2.376305203492468e-06, + "loss": 0.4932, + "step": 3691 + }, + { + "epoch": 3.5434469515122418, + "grad_norm": 1.00174933850564, + "learning_rate": 2.373454094782119e-06, + "loss": 0.4171, + "step": 3692 + }, + { + "epoch": 3.544407105136822, + "grad_norm": 1.0735340158142546, + "learning_rate": 2.370604165100148e-06, + "loss": 0.4101, + "step": 3693 + }, + { + "epoch": 3.545367258761402, + "grad_norm": 1.1214307020282934, + "learning_rate": 2.367755415725862e-06, + "loss": 0.4242, + "step": 3694 + }, + { + "epoch": 3.5463274123859816, + "grad_norm": 1.1645266080572896, + "learning_rate": 2.3649078479380316e-06, + "loss": 0.4577, + "step": 3695 + }, + { + "epoch": 3.547287566010562, + "grad_norm": 1.5721307352224687, + "learning_rate": 2.3620614630149013e-06, + "loss": 0.4848, + "step": 3696 + }, + { + "epoch": 3.5482477196351416, + "grad_norm": 1.3104680164652156, + "learning_rate": 2.359216262234184e-06, + "loss": 0.438, + "step": 3697 + }, + { + "epoch": 3.5492078732597214, + "grad_norm": 1.0779816651842788, + "learning_rate": 2.3563722468730567e-06, + "loss": 0.5184, + "step": 3698 + }, + { + "epoch": 3.550168026884301, + "grad_norm": 0.837558838380594, + "learning_rate": 2.3535294182081666e-06, + "loss": 0.4584, + "step": 3699 + }, + { + "epoch": 3.5511281805088815, + "grad_norm": 0.9892452703556603, + "learning_rate": 2.350687777515635e-06, + "loss": 0.5487, + "step": 3700 + }, + { + "epoch": 3.5520883341334613, + "grad_norm": 1.0156020785944047, + "learning_rate": 2.3478473260710387e-06, + "loss": 0.5219, + "step": 3701 + }, + { + "epoch": 3.5530484877580415, + "grad_norm": 1.0201654504271127, + "learning_rate": 2.34500806514943e-06, + "loss": 0.5055, + "step": 3702 + }, + { + "epoch": 3.5540086413826213, + "grad_norm": 0.7851140133339441, + "learning_rate": 2.3421699960253168e-06, + "loss": 0.4465, + "step": 3703 + }, + { + "epoch": 3.554968795007201, + "grad_norm": 0.9613923870294074, + "learning_rate": 2.3393331199726837e-06, + "loss": 0.5031, + "step": 3704 + }, + { + "epoch": 3.555928948631781, + "grad_norm": 1.0917314279342007, + "learning_rate": 2.3364974382649737e-06, + "loss": 0.4907, + "step": 3705 + }, + { + "epoch": 3.556889102256361, + "grad_norm": 1.1549333470707257, + "learning_rate": 2.3336629521750902e-06, + "loss": 0.4393, + "step": 3706 + }, + { + "epoch": 3.557849255880941, + "grad_norm": 1.1844655097667611, + "learning_rate": 2.3308296629754067e-06, + "loss": 0.398, + "step": 3707 + }, + { + "epoch": 3.558809409505521, + "grad_norm": 1.4718673944191167, + "learning_rate": 2.327997571937755e-06, + "loss": 0.5234, + "step": 3708 + }, + { + "epoch": 3.559769563130101, + "grad_norm": 0.7899354720749071, + "learning_rate": 2.3251666803334316e-06, + "loss": 0.3849, + "step": 3709 + }, + { + "epoch": 3.5607297167546808, + "grad_norm": 0.9801400054494465, + "learning_rate": 2.3223369894331947e-06, + "loss": 0.4092, + "step": 3710 + }, + { + "epoch": 3.5616898703792605, + "grad_norm": 1.2336551635147301, + "learning_rate": 2.3195085005072587e-06, + "loss": 0.4218, + "step": 3711 + }, + { + "epoch": 3.5626500240038403, + "grad_norm": 0.8657123624681815, + "learning_rate": 2.3166812148253045e-06, + "loss": 0.5174, + "step": 3712 + }, + { + "epoch": 3.5636101776284206, + "grad_norm": 1.1406124090395062, + "learning_rate": 2.313855133656469e-06, + "loss": 0.3652, + "step": 3713 + }, + { + "epoch": 3.5645703312530004, + "grad_norm": 1.0904647275948893, + "learning_rate": 2.3110302582693512e-06, + "loss": 0.4873, + "step": 3714 + }, + { + "epoch": 3.5655304848775806, + "grad_norm": 0.969364953364631, + "learning_rate": 2.3082065899320084e-06, + "loss": 0.4624, + "step": 3715 + }, + { + "epoch": 3.5664906385021604, + "grad_norm": 0.9765307121391327, + "learning_rate": 2.3053841299119524e-06, + "loss": 0.5346, + "step": 3716 + }, + { + "epoch": 3.56745079212674, + "grad_norm": 1.0690333161157577, + "learning_rate": 2.3025628794761544e-06, + "loss": 0.503, + "step": 3717 + }, + { + "epoch": 3.56841094575132, + "grad_norm": 0.862838816483104, + "learning_rate": 2.2997428398910497e-06, + "loss": 0.4878, + "step": 3718 + }, + { + "epoch": 3.5693710993759002, + "grad_norm": 1.2005659050029838, + "learning_rate": 2.2969240124225177e-06, + "loss": 0.4877, + "step": 3719 + }, + { + "epoch": 3.57033125300048, + "grad_norm": 0.919234353397464, + "learning_rate": 2.294106398335905e-06, + "loss": 0.3685, + "step": 3720 + }, + { + "epoch": 3.5712914066250603, + "grad_norm": 1.3125460441962709, + "learning_rate": 2.2912899988960015e-06, + "loss": 0.5297, + "step": 3721 + }, + { + "epoch": 3.57225156024964, + "grad_norm": 1.253709953333357, + "learning_rate": 2.288474815367065e-06, + "loss": 0.4826, + "step": 3722 + }, + { + "epoch": 3.57321171387422, + "grad_norm": 1.129404042598159, + "learning_rate": 2.2856608490128005e-06, + "loss": 0.4338, + "step": 3723 + }, + { + "epoch": 3.5741718674987997, + "grad_norm": 1.1505810636583158, + "learning_rate": 2.2828481010963647e-06, + "loss": 0.4836, + "step": 3724 + }, + { + "epoch": 3.5751320211233795, + "grad_norm": 1.246252816796723, + "learning_rate": 2.280036572880372e-06, + "loss": 0.5039, + "step": 3725 + }, + { + "epoch": 3.5760921747479597, + "grad_norm": 1.2784790157010364, + "learning_rate": 2.277226265626887e-06, + "loss": 0.4429, + "step": 3726 + }, + { + "epoch": 3.5770523283725395, + "grad_norm": 1.0481861440733469, + "learning_rate": 2.2744171805974257e-06, + "loss": 0.4873, + "step": 3727 + }, + { + "epoch": 3.5780124819971197, + "grad_norm": 1.2778175546190231, + "learning_rate": 2.27160931905296e-06, + "loss": 0.4806, + "step": 3728 + }, + { + "epoch": 3.5789726356216995, + "grad_norm": 1.098165453489468, + "learning_rate": 2.268802682253903e-06, + "loss": 0.4883, + "step": 3729 + }, + { + "epoch": 3.5799327892462793, + "grad_norm": 1.2114603753626112, + "learning_rate": 2.2659972714601264e-06, + "loss": 0.4976, + "step": 3730 + }, + { + "epoch": 3.580892942870859, + "grad_norm": 0.9533509049683406, + "learning_rate": 2.2631930879309525e-06, + "loss": 0.5182, + "step": 3731 + }, + { + "epoch": 3.5818530964954394, + "grad_norm": 1.4678657246619216, + "learning_rate": 2.2603901329251452e-06, + "loss": 0.4437, + "step": 3732 + }, + { + "epoch": 3.582813250120019, + "grad_norm": 1.1679398597192725, + "learning_rate": 2.2575884077009246e-06, + "loss": 0.5174, + "step": 3733 + }, + { + "epoch": 3.5837734037445994, + "grad_norm": 1.0503972593308595, + "learning_rate": 2.2547879135159495e-06, + "loss": 0.483, + "step": 3734 + }, + { + "epoch": 3.584733557369179, + "grad_norm": 0.9289524071299289, + "learning_rate": 2.2519886516273365e-06, + "loss": 0.4431, + "step": 3735 + }, + { + "epoch": 3.585693710993759, + "grad_norm": 1.209828943543199, + "learning_rate": 2.2491906232916457e-06, + "loss": 0.4678, + "step": 3736 + }, + { + "epoch": 3.586653864618339, + "grad_norm": 1.5376189296878735, + "learning_rate": 2.246393829764877e-06, + "loss": 0.4692, + "step": 3737 + }, + { + "epoch": 3.5876140182429186, + "grad_norm": 1.1288844498173878, + "learning_rate": 2.243598272302484e-06, + "loss": 0.4994, + "step": 3738 + }, + { + "epoch": 3.588574171867499, + "grad_norm": 1.3065944926194302, + "learning_rate": 2.240803952159362e-06, + "loss": 0.4572, + "step": 3739 + }, + { + "epoch": 3.5895343254920786, + "grad_norm": 1.265342252364368, + "learning_rate": 2.238010870589852e-06, + "loss": 0.4564, + "step": 3740 + }, + { + "epoch": 3.590494479116659, + "grad_norm": 0.9576001472932502, + "learning_rate": 2.23521902884774e-06, + "loss": 0.4557, + "step": 3741 + }, + { + "epoch": 3.5914546327412387, + "grad_norm": 1.0903177339210735, + "learning_rate": 2.23242842818625e-06, + "loss": 0.4634, + "step": 3742 + }, + { + "epoch": 3.5924147863658185, + "grad_norm": 1.4148767249512473, + "learning_rate": 2.2296390698580558e-06, + "loss": 0.4254, + "step": 3743 + }, + { + "epoch": 3.5933749399903983, + "grad_norm": 1.1502034934304621, + "learning_rate": 2.2268509551152684e-06, + "loss": 0.4472, + "step": 3744 + }, + { + "epoch": 3.5943350936149785, + "grad_norm": 1.0443676514435147, + "learning_rate": 2.2240640852094446e-06, + "loss": 0.5179, + "step": 3745 + }, + { + "epoch": 3.5952952472395583, + "grad_norm": 1.0334295753671674, + "learning_rate": 2.2212784613915816e-06, + "loss": 0.4259, + "step": 3746 + }, + { + "epoch": 3.5962554008641385, + "grad_norm": 1.335848923966106, + "learning_rate": 2.2184940849121123e-06, + "loss": 0.5216, + "step": 3747 + }, + { + "epoch": 3.5972155544887183, + "grad_norm": 0.9548860517089505, + "learning_rate": 2.215710957020914e-06, + "loss": 0.5389, + "step": 3748 + }, + { + "epoch": 3.598175708113298, + "grad_norm": 1.041411201420578, + "learning_rate": 2.2129290789673085e-06, + "loss": 0.5052, + "step": 3749 + }, + { + "epoch": 3.599135861737878, + "grad_norm": 1.1026477327746016, + "learning_rate": 2.2101484520000456e-06, + "loss": 0.4706, + "step": 3750 + }, + { + "epoch": 3.600096015362458, + "grad_norm": 1.0318213135724337, + "learning_rate": 2.2073690773673203e-06, + "loss": 0.4997, + "step": 3751 + }, + { + "epoch": 3.601056168987038, + "grad_norm": 0.8807303009517201, + "learning_rate": 2.2045909563167643e-06, + "loss": 0.4689, + "step": 3752 + }, + { + "epoch": 3.6020163226116177, + "grad_norm": 0.9186984880227923, + "learning_rate": 2.201814090095448e-06, + "loss": 0.3609, + "step": 3753 + }, + { + "epoch": 3.602976476236198, + "grad_norm": 0.8214188000362983, + "learning_rate": 2.1990384799498767e-06, + "loss": 0.5032, + "step": 3754 + }, + { + "epoch": 3.603936629860778, + "grad_norm": 1.182496994616656, + "learning_rate": 2.196264127125989e-06, + "loss": 0.3753, + "step": 3755 + }, + { + "epoch": 3.6048967834853576, + "grad_norm": 1.1228670731642691, + "learning_rate": 2.193491032869165e-06, + "loss": 0.4307, + "step": 3756 + }, + { + "epoch": 3.6058569371099374, + "grad_norm": 1.137019141890346, + "learning_rate": 2.190719198424216e-06, + "loss": 0.5454, + "step": 3757 + }, + { + "epoch": 3.6068170907345176, + "grad_norm": 1.0369007609946983, + "learning_rate": 2.1879486250353894e-06, + "loss": 0.5174, + "step": 3758 + }, + { + "epoch": 3.6077772443590974, + "grad_norm": 1.3671917447731123, + "learning_rate": 2.185179313946368e-06, + "loss": 0.4332, + "step": 3759 + }, + { + "epoch": 3.6087373979836777, + "grad_norm": 1.3303459448543085, + "learning_rate": 2.1824112664002612e-06, + "loss": 0.475, + "step": 3760 + }, + { + "epoch": 3.6096975516082574, + "grad_norm": 1.151641920586827, + "learning_rate": 2.1796444836396174e-06, + "loss": 0.4152, + "step": 3761 + }, + { + "epoch": 3.6106577052328372, + "grad_norm": 1.2268694170133585, + "learning_rate": 2.1768789669064205e-06, + "loss": 0.4331, + "step": 3762 + }, + { + "epoch": 3.611617858857417, + "grad_norm": 1.1768682100365933, + "learning_rate": 2.174114717442076e-06, + "loss": 0.508, + "step": 3763 + }, + { + "epoch": 3.6125780124819973, + "grad_norm": 1.3976781563151333, + "learning_rate": 2.1713517364874297e-06, + "loss": 0.5216, + "step": 3764 + }, + { + "epoch": 3.613538166106577, + "grad_norm": 1.131551678961866, + "learning_rate": 2.1685900252827485e-06, + "loss": 0.5117, + "step": 3765 + }, + { + "epoch": 3.614498319731157, + "grad_norm": 0.9706726329158502, + "learning_rate": 2.1658295850677408e-06, + "loss": 0.4293, + "step": 3766 + }, + { + "epoch": 3.615458473355737, + "grad_norm": 1.2978353247982275, + "learning_rate": 2.1630704170815386e-06, + "loss": 0.5624, + "step": 3767 + }, + { + "epoch": 3.616418626980317, + "grad_norm": 1.116083473577353, + "learning_rate": 2.160312522562699e-06, + "loss": 0.4809, + "step": 3768 + }, + { + "epoch": 3.6173787806048967, + "grad_norm": 1.3468108530134666, + "learning_rate": 2.1575559027492136e-06, + "loss": 0.4204, + "step": 3769 + }, + { + "epoch": 3.6183389342294765, + "grad_norm": 1.5594933534085889, + "learning_rate": 2.1548005588785e-06, + "loss": 0.45, + "step": 3770 + }, + { + "epoch": 3.6192990878540567, + "grad_norm": 0.8602918576594968, + "learning_rate": 2.1520464921874014e-06, + "loss": 0.4982, + "step": 3771 + }, + { + "epoch": 3.6202592414786365, + "grad_norm": 1.2003644889457366, + "learning_rate": 2.1492937039121914e-06, + "loss": 0.466, + "step": 3772 + }, + { + "epoch": 3.6212193951032168, + "grad_norm": 1.223810743019791, + "learning_rate": 2.1465421952885633e-06, + "loss": 0.4404, + "step": 3773 + }, + { + "epoch": 3.6221795487277966, + "grad_norm": 1.1282777272302784, + "learning_rate": 2.1437919675516407e-06, + "loss": 0.4875, + "step": 3774 + }, + { + "epoch": 3.6231397023523764, + "grad_norm": 1.4238496666103069, + "learning_rate": 2.1410430219359757e-06, + "loss": 0.4999, + "step": 3775 + }, + { + "epoch": 3.624099855976956, + "grad_norm": 1.1399952075571127, + "learning_rate": 2.1382953596755364e-06, + "loss": 0.516, + "step": 3776 + }, + { + "epoch": 3.6250600096015364, + "grad_norm": 1.1675105494430775, + "learning_rate": 2.1355489820037216e-06, + "loss": 0.484, + "step": 3777 + }, + { + "epoch": 3.626020163226116, + "grad_norm": 1.5866934246571585, + "learning_rate": 2.132803890153346e-06, + "loss": 0.4361, + "step": 3778 + }, + { + "epoch": 3.626980316850696, + "grad_norm": 1.2903098040377423, + "learning_rate": 2.1300600853566577e-06, + "loss": 0.4428, + "step": 3779 + }, + { + "epoch": 3.6279404704752762, + "grad_norm": 1.1407984683048416, + "learning_rate": 2.12731756884532e-06, + "loss": 0.4933, + "step": 3780 + }, + { + "epoch": 3.628900624099856, + "grad_norm": 1.027459534244755, + "learning_rate": 2.1245763418504173e-06, + "loss": 0.4886, + "step": 3781 + }, + { + "epoch": 3.629860777724436, + "grad_norm": 1.0942181619953522, + "learning_rate": 2.121836405602458e-06, + "loss": 0.51, + "step": 3782 + }, + { + "epoch": 3.6308209313490156, + "grad_norm": 0.7417101969205426, + "learning_rate": 2.1190977613313707e-06, + "loss": 0.5201, + "step": 3783 + }, + { + "epoch": 3.631781084973596, + "grad_norm": 1.0582103684843327, + "learning_rate": 2.1163604102665023e-06, + "loss": 0.6581, + "step": 3784 + }, + { + "epoch": 3.6327412385981757, + "grad_norm": 1.6004720545643263, + "learning_rate": 2.113624353636624e-06, + "loss": 0.4421, + "step": 3785 + }, + { + "epoch": 3.633701392222756, + "grad_norm": 1.2380137171182835, + "learning_rate": 2.110889592669918e-06, + "loss": 0.4854, + "step": 3786 + }, + { + "epoch": 3.6346615458473357, + "grad_norm": 1.0655190265487184, + "learning_rate": 2.1081561285939904e-06, + "loss": 0.447, + "step": 3787 + }, + { + "epoch": 3.6356216994719155, + "grad_norm": 1.0190173251848118, + "learning_rate": 2.1054239626358648e-06, + "loss": 0.5404, + "step": 3788 + }, + { + "epoch": 3.6365818530964953, + "grad_norm": 0.9016277543333088, + "learning_rate": 2.102693096021981e-06, + "loss": 0.4645, + "step": 3789 + }, + { + "epoch": 3.6375420067210755, + "grad_norm": 1.0905249306293494, + "learning_rate": 2.0999635299781976e-06, + "loss": 0.529, + "step": 3790 + }, + { + "epoch": 3.6385021603456553, + "grad_norm": 1.0432333904590128, + "learning_rate": 2.0972352657297845e-06, + "loss": 0.411, + "step": 3791 + }, + { + "epoch": 3.639462313970235, + "grad_norm": 1.0182766520315567, + "learning_rate": 2.0945083045014296e-06, + "loss": 0.5278, + "step": 3792 + }, + { + "epoch": 3.6404224675948154, + "grad_norm": 1.0280915175157022, + "learning_rate": 2.0917826475172427e-06, + "loss": 0.5295, + "step": 3793 + }, + { + "epoch": 3.641382621219395, + "grad_norm": 1.3129083329449196, + "learning_rate": 2.089058296000736e-06, + "loss": 0.4705, + "step": 3794 + }, + { + "epoch": 3.642342774843975, + "grad_norm": 1.4819794363441172, + "learning_rate": 2.0863352511748466e-06, + "loss": 0.5643, + "step": 3795 + }, + { + "epoch": 3.6433029284685547, + "grad_norm": 1.048438135848867, + "learning_rate": 2.083613514261913e-06, + "loss": 0.4837, + "step": 3796 + }, + { + "epoch": 3.644263082093135, + "grad_norm": 1.0991255701955989, + "learning_rate": 2.0808930864837007e-06, + "loss": 0.4736, + "step": 3797 + }, + { + "epoch": 3.645223235717715, + "grad_norm": 1.060065825731226, + "learning_rate": 2.07817396906138e-06, + "loss": 0.506, + "step": 3798 + }, + { + "epoch": 3.646183389342295, + "grad_norm": 1.0305012563717666, + "learning_rate": 2.0754561632155307e-06, + "loss": 0.426, + "step": 3799 + }, + { + "epoch": 3.647143542966875, + "grad_norm": 0.931002501810837, + "learning_rate": 2.072739670166148e-06, + "loss": 0.463, + "step": 3800 + }, + { + "epoch": 3.6481036965914546, + "grad_norm": 1.0176314728629507, + "learning_rate": 2.070024491132636e-06, + "loss": 0.5337, + "step": 3801 + }, + { + "epoch": 3.6490638502160344, + "grad_norm": 0.8311913086495, + "learning_rate": 2.067310627333811e-06, + "loss": 0.4432, + "step": 3802 + }, + { + "epoch": 3.6500240038406147, + "grad_norm": 1.014117237799301, + "learning_rate": 2.064598079987898e-06, + "loss": 0.4402, + "step": 3803 + }, + { + "epoch": 3.6509841574651944, + "grad_norm": 0.6819680264762884, + "learning_rate": 2.0618868503125273e-06, + "loss": 0.5504, + "step": 3804 + }, + { + "epoch": 3.6519443110897742, + "grad_norm": 1.1872475867322425, + "learning_rate": 2.059176939524741e-06, + "loss": 0.4089, + "step": 3805 + }, + { + "epoch": 3.6529044647143545, + "grad_norm": 1.1212236818671708, + "learning_rate": 2.056468348840995e-06, + "loss": 0.4413, + "step": 3806 + }, + { + "epoch": 3.6538646183389343, + "grad_norm": 1.0698093019264092, + "learning_rate": 2.0537610794771405e-06, + "loss": 0.4692, + "step": 3807 + }, + { + "epoch": 3.654824771963514, + "grad_norm": 0.8909055328097167, + "learning_rate": 2.0510551326484457e-06, + "loss": 0.418, + "step": 3808 + }, + { + "epoch": 3.655784925588094, + "grad_norm": 1.3020564706671507, + "learning_rate": 2.0483505095695755e-06, + "loss": 0.4975, + "step": 3809 + }, + { + "epoch": 3.656745079212674, + "grad_norm": 1.003057278453496, + "learning_rate": 2.045647211454613e-06, + "loss": 0.5627, + "step": 3810 + }, + { + "epoch": 3.657705232837254, + "grad_norm": 1.0211388788908284, + "learning_rate": 2.0429452395170372e-06, + "loss": 0.4647, + "step": 3811 + }, + { + "epoch": 3.658665386461834, + "grad_norm": 1.504892499000519, + "learning_rate": 2.0402445949697335e-06, + "loss": 0.4165, + "step": 3812 + }, + { + "epoch": 3.659625540086414, + "grad_norm": 1.0857535171455013, + "learning_rate": 2.037545279024993e-06, + "loss": 0.4049, + "step": 3813 + }, + { + "epoch": 3.6605856937109937, + "grad_norm": 1.1869996238194045, + "learning_rate": 2.0348472928945096e-06, + "loss": 0.3666, + "step": 3814 + }, + { + "epoch": 3.6615458473355735, + "grad_norm": 1.168459723293314, + "learning_rate": 2.0321506377893807e-06, + "loss": 0.4781, + "step": 3815 + }, + { + "epoch": 3.6625060009601538, + "grad_norm": 1.0862836544926067, + "learning_rate": 2.029455314920108e-06, + "loss": 0.4357, + "step": 3816 + }, + { + "epoch": 3.6634661545847336, + "grad_norm": 1.056344464364202, + "learning_rate": 2.026761325496589e-06, + "loss": 0.3488, + "step": 3817 + }, + { + "epoch": 3.6644263082093134, + "grad_norm": 1.6004071202908687, + "learning_rate": 2.024068670728127e-06, + "loss": 0.6168, + "step": 3818 + }, + { + "epoch": 3.6653864618338936, + "grad_norm": 2.1330535107668678, + "learning_rate": 2.021377351823432e-06, + "loss": 0.527, + "step": 3819 + }, + { + "epoch": 3.6663466154584734, + "grad_norm": 1.1470659153663634, + "learning_rate": 2.0186873699906017e-06, + "loss": 0.4922, + "step": 3820 + }, + { + "epoch": 3.667306769083053, + "grad_norm": 1.0598011721769491, + "learning_rate": 2.0159987264371445e-06, + "loss": 0.519, + "step": 3821 + }, + { + "epoch": 3.668266922707633, + "grad_norm": 0.9138033948010229, + "learning_rate": 2.0133114223699577e-06, + "loss": 0.4927, + "step": 3822 + }, + { + "epoch": 3.6692270763322132, + "grad_norm": 0.9235137477631645, + "learning_rate": 2.01062545899535e-06, + "loss": 0.566, + "step": 3823 + }, + { + "epoch": 3.670187229956793, + "grad_norm": 1.1116614261931967, + "learning_rate": 2.007940837519021e-06, + "loss": 0.4614, + "step": 3824 + }, + { + "epoch": 3.6711473835813733, + "grad_norm": 0.9962640331207523, + "learning_rate": 2.0052575591460636e-06, + "loss": 0.4634, + "step": 3825 + }, + { + "epoch": 3.672107537205953, + "grad_norm": 1.001008425099837, + "learning_rate": 2.002575625080977e-06, + "loss": 0.5174, + "step": 3826 + }, + { + "epoch": 3.673067690830533, + "grad_norm": 1.0279895057906892, + "learning_rate": 1.9998950365276514e-06, + "loss": 0.5231, + "step": 3827 + }, + { + "epoch": 3.6740278444551127, + "grad_norm": 1.2631330092758803, + "learning_rate": 1.997215794689375e-06, + "loss": 0.5177, + "step": 3828 + }, + { + "epoch": 3.674987998079693, + "grad_norm": 1.0452875649365183, + "learning_rate": 1.9945379007688314e-06, + "loss": 0.5364, + "step": 3829 + }, + { + "epoch": 3.6759481517042727, + "grad_norm": 1.1260108577949433, + "learning_rate": 1.991861355968096e-06, + "loss": 0.5284, + "step": 3830 + }, + { + "epoch": 3.6769083053288525, + "grad_norm": 0.7148469060758215, + "learning_rate": 1.989186161488643e-06, + "loss": 0.4806, + "step": 3831 + }, + { + "epoch": 3.6778684589534327, + "grad_norm": 1.063952880342718, + "learning_rate": 1.986512318531338e-06, + "loss": 0.6073, + "step": 3832 + }, + { + "epoch": 3.6788286125780125, + "grad_norm": 1.137127963521302, + "learning_rate": 1.9838398282964412e-06, + "loss": 0.4586, + "step": 3833 + }, + { + "epoch": 3.6797887662025923, + "grad_norm": 1.1803226032299732, + "learning_rate": 1.9811686919836054e-06, + "loss": 0.389, + "step": 3834 + }, + { + "epoch": 3.680748919827172, + "grad_norm": 0.9911285489470352, + "learning_rate": 1.9784989107918736e-06, + "loss": 0.5239, + "step": 3835 + }, + { + "epoch": 3.6817090734517524, + "grad_norm": 0.8276777288490931, + "learning_rate": 1.97583048591968e-06, + "loss": 0.4406, + "step": 3836 + }, + { + "epoch": 3.682669227076332, + "grad_norm": 0.947486413230033, + "learning_rate": 1.9731634185648584e-06, + "loss": 0.4647, + "step": 3837 + }, + { + "epoch": 3.6836293807009124, + "grad_norm": 1.1926616009097457, + "learning_rate": 1.970497709924622e-06, + "loss": 0.552, + "step": 3838 + }, + { + "epoch": 3.684589534325492, + "grad_norm": 1.3684420975143503, + "learning_rate": 1.9678333611955814e-06, + "loss": 0.3903, + "step": 3839 + }, + { + "epoch": 3.685549687950072, + "grad_norm": 1.023355141836186, + "learning_rate": 1.965170373573729e-06, + "loss": 0.5019, + "step": 3840 + }, + { + "epoch": 3.686509841574652, + "grad_norm": 1.6823560265523865, + "learning_rate": 1.9625087482544574e-06, + "loss": 0.5438, + "step": 3841 + }, + { + "epoch": 3.687469995199232, + "grad_norm": 1.1916669996347646, + "learning_rate": 1.959848486432542e-06, + "loss": 0.4666, + "step": 3842 + }, + { + "epoch": 3.688430148823812, + "grad_norm": 0.9046535451881969, + "learning_rate": 1.957189589302141e-06, + "loss": 0.5544, + "step": 3843 + }, + { + "epoch": 3.6893903024483916, + "grad_norm": 1.3042539268476607, + "learning_rate": 1.9545320580568077e-06, + "loss": 0.3806, + "step": 3844 + }, + { + "epoch": 3.690350456072972, + "grad_norm": 1.4284881962942269, + "learning_rate": 1.95187589388948e-06, + "loss": 0.4429, + "step": 3845 + }, + { + "epoch": 3.6913106096975516, + "grad_norm": 1.1642344349877685, + "learning_rate": 1.9492210979924805e-06, + "loss": 0.4918, + "step": 3846 + }, + { + "epoch": 3.6922707633221314, + "grad_norm": 0.799241305967219, + "learning_rate": 1.9465676715575203e-06, + "loss": 0.4752, + "step": 3847 + }, + { + "epoch": 3.6932309169467112, + "grad_norm": 1.0028058610874377, + "learning_rate": 1.9439156157756917e-06, + "loss": 0.5389, + "step": 3848 + }, + { + "epoch": 3.6941910705712915, + "grad_norm": 1.093868951632084, + "learning_rate": 1.9412649318374737e-06, + "loss": 0.5107, + "step": 3849 + }, + { + "epoch": 3.6951512241958713, + "grad_norm": 1.2293346312824376, + "learning_rate": 1.938615620932735e-06, + "loss": 0.4363, + "step": 3850 + }, + { + "epoch": 3.6961113778204515, + "grad_norm": 1.5130933913838918, + "learning_rate": 1.9359676842507186e-06, + "loss": 0.5077, + "step": 3851 + }, + { + "epoch": 3.6970715314450313, + "grad_norm": 1.1238119096822885, + "learning_rate": 1.933321122980057e-06, + "loss": 0.528, + "step": 3852 + }, + { + "epoch": 3.698031685069611, + "grad_norm": 1.2140115401556606, + "learning_rate": 1.9306759383087593e-06, + "loss": 0.4128, + "step": 3853 + }, + { + "epoch": 3.698991838694191, + "grad_norm": 1.8592799022004682, + "learning_rate": 1.9280321314242252e-06, + "loss": 0.3679, + "step": 3854 + }, + { + "epoch": 3.699951992318771, + "grad_norm": 1.2246118458849933, + "learning_rate": 1.925389703513232e-06, + "loss": 0.4778, + "step": 3855 + }, + { + "epoch": 3.700912145943351, + "grad_norm": 1.1750832993551101, + "learning_rate": 1.922748655761932e-06, + "loss": 0.5096, + "step": 3856 + }, + { + "epoch": 3.7018722995679307, + "grad_norm": 1.0519279990408164, + "learning_rate": 1.9201089893558677e-06, + "loss": 0.4687, + "step": 3857 + }, + { + "epoch": 3.702832453192511, + "grad_norm": 1.0062094032320303, + "learning_rate": 1.9174707054799558e-06, + "loss": 0.445, + "step": 3858 + }, + { + "epoch": 3.7037926068170908, + "grad_norm": 0.9081150690693638, + "learning_rate": 1.9148338053184933e-06, + "loss": 0.5094, + "step": 3859 + }, + { + "epoch": 3.7047527604416706, + "grad_norm": 0.9917559328970503, + "learning_rate": 1.9121982900551592e-06, + "loss": 0.535, + "step": 3860 + }, + { + "epoch": 3.7057129140662504, + "grad_norm": 1.5002787408890847, + "learning_rate": 1.909564160873004e-06, + "loss": 0.4836, + "step": 3861 + }, + { + "epoch": 3.7066730676908306, + "grad_norm": 1.1177091307108697, + "learning_rate": 1.9069314189544602e-06, + "loss": 0.4152, + "step": 3862 + }, + { + "epoch": 3.7076332213154104, + "grad_norm": 1.0614380295231434, + "learning_rate": 1.9043000654813433e-06, + "loss": 0.4936, + "step": 3863 + }, + { + "epoch": 3.7085933749399906, + "grad_norm": 1.1500221965488515, + "learning_rate": 1.9016701016348338e-06, + "loss": 0.4418, + "step": 3864 + }, + { + "epoch": 3.7095535285645704, + "grad_norm": 1.3080713530254886, + "learning_rate": 1.8990415285954976e-06, + "loss": 0.4278, + "step": 3865 + }, + { + "epoch": 3.7105136821891502, + "grad_norm": 1.000435917697426, + "learning_rate": 1.896414347543269e-06, + "loss": 0.5226, + "step": 3866 + }, + { + "epoch": 3.71147383581373, + "grad_norm": 1.0253677138621886, + "learning_rate": 1.8937885596574656e-06, + "loss": 0.5902, + "step": 3867 + }, + { + "epoch": 3.7124339894383103, + "grad_norm": 1.7138951431289862, + "learning_rate": 1.8911641661167762e-06, + "loss": 0.4834, + "step": 3868 + }, + { + "epoch": 3.71339414306289, + "grad_norm": 0.9001297497384394, + "learning_rate": 1.8885411680992594e-06, + "loss": 0.472, + "step": 3869 + }, + { + "epoch": 3.71435429668747, + "grad_norm": 1.2134955052246705, + "learning_rate": 1.885919566782352e-06, + "loss": 0.3747, + "step": 3870 + }, + { + "epoch": 3.71531445031205, + "grad_norm": 1.2486109383089363, + "learning_rate": 1.8832993633428643e-06, + "loss": 0.5055, + "step": 3871 + }, + { + "epoch": 3.71627460393663, + "grad_norm": 0.9578237580606398, + "learning_rate": 1.880680558956976e-06, + "loss": 0.4621, + "step": 3872 + }, + { + "epoch": 3.7172347575612097, + "grad_norm": 1.1294383846309326, + "learning_rate": 1.878063154800243e-06, + "loss": 0.4249, + "step": 3873 + }, + { + "epoch": 3.7181949111857895, + "grad_norm": 1.1396561743373135, + "learning_rate": 1.8754471520475863e-06, + "loss": 0.5239, + "step": 3874 + }, + { + "epoch": 3.7191550648103697, + "grad_norm": 1.4387812119997676, + "learning_rate": 1.8728325518733038e-06, + "loss": 0.4629, + "step": 3875 + }, + { + "epoch": 3.7201152184349495, + "grad_norm": 0.9601518279654647, + "learning_rate": 1.870219355451061e-06, + "loss": 0.4521, + "step": 3876 + }, + { + "epoch": 3.7210753720595298, + "grad_norm": 1.1369061406321572, + "learning_rate": 1.8676075639538943e-06, + "loss": 0.3902, + "step": 3877 + }, + { + "epoch": 3.7220355256841096, + "grad_norm": 1.264290690564499, + "learning_rate": 1.8649971785542097e-06, + "loss": 0.4139, + "step": 3878 + }, + { + "epoch": 3.7229956793086894, + "grad_norm": 0.9841744706554263, + "learning_rate": 1.8623882004237792e-06, + "loss": 0.4981, + "step": 3879 + }, + { + "epoch": 3.723955832933269, + "grad_norm": 1.2140239395794963, + "learning_rate": 1.8597806307337446e-06, + "loss": 0.4605, + "step": 3880 + }, + { + "epoch": 3.7249159865578494, + "grad_norm": 1.0122809883079258, + "learning_rate": 1.8571744706546213e-06, + "loss": 0.5375, + "step": 3881 + }, + { + "epoch": 3.725876140182429, + "grad_norm": 0.9299945433671742, + "learning_rate": 1.8545697213562813e-06, + "loss": 0.358, + "step": 3882 + }, + { + "epoch": 3.726836293807009, + "grad_norm": 1.188804231317955, + "learning_rate": 1.8519663840079721e-06, + "loss": 0.5174, + "step": 3883 + }, + { + "epoch": 3.727796447431589, + "grad_norm": 1.1092629350185208, + "learning_rate": 1.8493644597782996e-06, + "loss": 0.4704, + "step": 3884 + }, + { + "epoch": 3.728756601056169, + "grad_norm": 1.2262609049279667, + "learning_rate": 1.8467639498352442e-06, + "loss": 0.38, + "step": 3885 + }, + { + "epoch": 3.729716754680749, + "grad_norm": 0.9560647319306025, + "learning_rate": 1.8441648553461467e-06, + "loss": 0.5391, + "step": 3886 + }, + { + "epoch": 3.7306769083053286, + "grad_norm": 1.086947314689461, + "learning_rate": 1.8415671774777106e-06, + "loss": 0.4768, + "step": 3887 + }, + { + "epoch": 3.731637061929909, + "grad_norm": 1.0511073997157976, + "learning_rate": 1.8389709173960063e-06, + "loss": 0.4915, + "step": 3888 + }, + { + "epoch": 3.7325972155544886, + "grad_norm": 0.9677959792317167, + "learning_rate": 1.8363760762664678e-06, + "loss": 0.4843, + "step": 3889 + }, + { + "epoch": 3.733557369179069, + "grad_norm": 1.0713756992235126, + "learning_rate": 1.8337826552538911e-06, + "loss": 0.4454, + "step": 3890 + }, + { + "epoch": 3.7345175228036487, + "grad_norm": 0.9423399407475826, + "learning_rate": 1.8311906555224379e-06, + "loss": 0.5349, + "step": 3891 + }, + { + "epoch": 3.7354776764282285, + "grad_norm": 0.9525167268798441, + "learning_rate": 1.8286000782356245e-06, + "loss": 0.5551, + "step": 3892 + }, + { + "epoch": 3.7364378300528083, + "grad_norm": 0.904942346769752, + "learning_rate": 1.826010924556334e-06, + "loss": 0.5342, + "step": 3893 + }, + { + "epoch": 3.7373979836773885, + "grad_norm": 0.8432772139184227, + "learning_rate": 1.823423195646815e-06, + "loss": 0.4401, + "step": 3894 + }, + { + "epoch": 3.7383581373019683, + "grad_norm": 1.211176320508855, + "learning_rate": 1.820836892668666e-06, + "loss": 0.5373, + "step": 3895 + }, + { + "epoch": 3.739318290926548, + "grad_norm": 0.980186133683255, + "learning_rate": 1.8182520167828537e-06, + "loss": 0.467, + "step": 3896 + }, + { + "epoch": 3.7402784445511283, + "grad_norm": 1.1333613881232587, + "learning_rate": 1.8156685691496972e-06, + "loss": 0.4619, + "step": 3897 + }, + { + "epoch": 3.741238598175708, + "grad_norm": 1.079537447614277, + "learning_rate": 1.813086550928883e-06, + "loss": 0.4547, + "step": 3898 + }, + { + "epoch": 3.742198751800288, + "grad_norm": 1.6859175540977833, + "learning_rate": 1.8105059632794514e-06, + "loss": 0.5553, + "step": 3899 + }, + { + "epoch": 3.7431589054248677, + "grad_norm": 1.0163050132852784, + "learning_rate": 1.8079268073597978e-06, + "loss": 0.6202, + "step": 3900 + }, + { + "epoch": 3.744119059049448, + "grad_norm": 1.4034299593983564, + "learning_rate": 1.8053490843276788e-06, + "loss": 0.4142, + "step": 3901 + }, + { + "epoch": 3.7450792126740278, + "grad_norm": 0.8635569537928273, + "learning_rate": 1.802772795340207e-06, + "loss": 0.5466, + "step": 3902 + }, + { + "epoch": 3.746039366298608, + "grad_norm": 1.1086508319040311, + "learning_rate": 1.8001979415538507e-06, + "loss": 0.4346, + "step": 3903 + }, + { + "epoch": 3.746999519923188, + "grad_norm": 0.8020501928764563, + "learning_rate": 1.797624524124436e-06, + "loss": 0.5652, + "step": 3904 + }, + { + "epoch": 3.7479596735477676, + "grad_norm": 1.205370198851128, + "learning_rate": 1.7950525442071394e-06, + "loss": 0.4764, + "step": 3905 + }, + { + "epoch": 3.7489198271723474, + "grad_norm": 1.176167883572754, + "learning_rate": 1.792482002956495e-06, + "loss": 0.5206, + "step": 3906 + }, + { + "epoch": 3.7498799807969276, + "grad_norm": 1.5349338978715579, + "learning_rate": 1.7899129015263966e-06, + "loss": 0.4927, + "step": 3907 + }, + { + "epoch": 3.7508401344215074, + "grad_norm": 1.0906294491057715, + "learning_rate": 1.7873452410700809e-06, + "loss": 0.3882, + "step": 3908 + }, + { + "epoch": 3.7518002880460872, + "grad_norm": 0.9240064371691598, + "learning_rate": 1.7847790227401469e-06, + "loss": 0.5617, + "step": 3909 + }, + { + "epoch": 3.7527604416706675, + "grad_norm": 1.116016545720751, + "learning_rate": 1.7822142476885374e-06, + "loss": 0.4073, + "step": 3910 + }, + { + "epoch": 3.7537205952952473, + "grad_norm": 1.2140127137387613, + "learning_rate": 1.7796509170665571e-06, + "loss": 0.5582, + "step": 3911 + }, + { + "epoch": 3.754680748919827, + "grad_norm": 1.2331357315877847, + "learning_rate": 1.777089032024859e-06, + "loss": 0.4364, + "step": 3912 + }, + { + "epoch": 3.755640902544407, + "grad_norm": 1.4016512918660973, + "learning_rate": 1.774528593713441e-06, + "loss": 0.4629, + "step": 3913 + }, + { + "epoch": 3.756601056168987, + "grad_norm": 1.1567842910412773, + "learning_rate": 1.771969603281659e-06, + "loss": 0.5094, + "step": 3914 + }, + { + "epoch": 3.757561209793567, + "grad_norm": 0.9150352630096552, + "learning_rate": 1.7694120618782169e-06, + "loss": 0.4956, + "step": 3915 + }, + { + "epoch": 3.758521363418147, + "grad_norm": 1.2874225313145111, + "learning_rate": 1.7668559706511667e-06, + "loss": 0.4894, + "step": 3916 + }, + { + "epoch": 3.759481517042727, + "grad_norm": 0.8103318851271448, + "learning_rate": 1.7643013307479129e-06, + "loss": 0.4292, + "step": 3917 + }, + { + "epoch": 3.7604416706673067, + "grad_norm": 1.3081346179824862, + "learning_rate": 1.7617481433152022e-06, + "loss": 0.4678, + "step": 3918 + }, + { + "epoch": 3.7614018242918865, + "grad_norm": 1.600637321529164, + "learning_rate": 1.759196409499135e-06, + "loss": 0.4621, + "step": 3919 + }, + { + "epoch": 3.7623619779164668, + "grad_norm": 1.7248563076898713, + "learning_rate": 1.756646130445157e-06, + "loss": 0.4591, + "step": 3920 + }, + { + "epoch": 3.7633221315410466, + "grad_norm": 1.407207687240018, + "learning_rate": 1.7540973072980621e-06, + "loss": 0.5207, + "step": 3921 + }, + { + "epoch": 3.7642822851656264, + "grad_norm": 0.9463713629373397, + "learning_rate": 1.7515499412019903e-06, + "loss": 0.4003, + "step": 3922 + }, + { + "epoch": 3.7652424387902066, + "grad_norm": 1.1207872040845508, + "learning_rate": 1.7490040333004248e-06, + "loss": 0.4715, + "step": 3923 + }, + { + "epoch": 3.7662025924147864, + "grad_norm": 0.8120050174698782, + "learning_rate": 1.7464595847361954e-06, + "loss": 0.5019, + "step": 3924 + }, + { + "epoch": 3.767162746039366, + "grad_norm": 1.1402370705923843, + "learning_rate": 1.7439165966514831e-06, + "loss": 0.5053, + "step": 3925 + }, + { + "epoch": 3.768122899663946, + "grad_norm": 1.2563562511011135, + "learning_rate": 1.741375070187804e-06, + "loss": 0.4028, + "step": 3926 + }, + { + "epoch": 3.769083053288526, + "grad_norm": 2.0028719378457343, + "learning_rate": 1.7388350064860242e-06, + "loss": 0.4408, + "step": 3927 + }, + { + "epoch": 3.770043206913106, + "grad_norm": 1.5404120053122512, + "learning_rate": 1.7362964066863469e-06, + "loss": 0.4039, + "step": 3928 + }, + { + "epoch": 3.7710033605376863, + "grad_norm": 0.915029989906004, + "learning_rate": 1.7337592719283276e-06, + "loss": 0.4721, + "step": 3929 + }, + { + "epoch": 3.771963514162266, + "grad_norm": 1.1669503945125583, + "learning_rate": 1.7312236033508584e-06, + "loss": 0.4725, + "step": 3930 + }, + { + "epoch": 3.772923667786846, + "grad_norm": 1.0322905530871467, + "learning_rate": 1.7286894020921707e-06, + "loss": 0.3973, + "step": 3931 + }, + { + "epoch": 3.7738838214114256, + "grad_norm": 1.2102443686153808, + "learning_rate": 1.726156669289842e-06, + "loss": 0.4912, + "step": 3932 + }, + { + "epoch": 3.774843975036006, + "grad_norm": 0.9706334814101453, + "learning_rate": 1.7236254060807889e-06, + "loss": 0.4616, + "step": 3933 + }, + { + "epoch": 3.7758041286605857, + "grad_norm": 0.983761787230528, + "learning_rate": 1.7210956136012684e-06, + "loss": 0.5281, + "step": 3934 + }, + { + "epoch": 3.7767642822851655, + "grad_norm": 1.1635859888182039, + "learning_rate": 1.718567292986879e-06, + "loss": 0.6095, + "step": 3935 + }, + { + "epoch": 3.7777244359097457, + "grad_norm": 0.8723998414425806, + "learning_rate": 1.7160404453725543e-06, + "loss": 0.4562, + "step": 3936 + }, + { + "epoch": 3.7786845895343255, + "grad_norm": 1.1293150166627424, + "learning_rate": 1.7135150718925674e-06, + "loss": 0.5679, + "step": 3937 + }, + { + "epoch": 3.7796447431589053, + "grad_norm": 1.1096584802561167, + "learning_rate": 1.7109911736805385e-06, + "loss": 0.5861, + "step": 3938 + }, + { + "epoch": 3.780604896783485, + "grad_norm": 1.2762288013138434, + "learning_rate": 1.708468751869412e-06, + "loss": 0.5473, + "step": 3939 + }, + { + "epoch": 3.7815650504080653, + "grad_norm": 1.2029755140026448, + "learning_rate": 1.7059478075914799e-06, + "loss": 0.4715, + "step": 3940 + }, + { + "epoch": 3.782525204032645, + "grad_norm": 1.1678231158767183, + "learning_rate": 1.703428341978362e-06, + "loss": 0.4385, + "step": 3941 + }, + { + "epoch": 3.7834853576572254, + "grad_norm": 1.4223721118460644, + "learning_rate": 1.7009103561610252e-06, + "loss": 0.4775, + "step": 3942 + }, + { + "epoch": 3.784445511281805, + "grad_norm": 0.8085978878365441, + "learning_rate": 1.6983938512697655e-06, + "loss": 0.4579, + "step": 3943 + }, + { + "epoch": 3.785405664906385, + "grad_norm": 1.0128151559011735, + "learning_rate": 1.695878828434212e-06, + "loss": 0.5393, + "step": 3944 + }, + { + "epoch": 3.7863658185309648, + "grad_norm": 1.233514523118038, + "learning_rate": 1.6933652887833336e-06, + "loss": 0.4682, + "step": 3945 + }, + { + "epoch": 3.787325972155545, + "grad_norm": 0.7304627487638629, + "learning_rate": 1.6908532334454308e-06, + "loss": 0.4831, + "step": 3946 + }, + { + "epoch": 3.788286125780125, + "grad_norm": 1.16232494135482, + "learning_rate": 1.6883426635481392e-06, + "loss": 0.5213, + "step": 3947 + }, + { + "epoch": 3.7892462794047046, + "grad_norm": 0.7646794081219642, + "learning_rate": 1.6858335802184278e-06, + "loss": 0.4139, + "step": 3948 + }, + { + "epoch": 3.790206433029285, + "grad_norm": 1.3028968325156647, + "learning_rate": 1.683325984582594e-06, + "loss": 0.4575, + "step": 3949 + }, + { + "epoch": 3.7911665866538646, + "grad_norm": 1.4783428596196397, + "learning_rate": 1.680819877766272e-06, + "loss": 0.4673, + "step": 3950 + }, + { + "epoch": 3.7921267402784444, + "grad_norm": 1.498146910428719, + "learning_rate": 1.678315260894427e-06, + "loss": 0.5331, + "step": 3951 + }, + { + "epoch": 3.7930868939030242, + "grad_norm": 1.599070554279104, + "learning_rate": 1.6758121350913548e-06, + "loss": 0.4335, + "step": 3952 + }, + { + "epoch": 3.7940470475276045, + "grad_norm": 1.0546319831143243, + "learning_rate": 1.6733105014806834e-06, + "loss": 0.4151, + "step": 3953 + }, + { + "epoch": 3.7950072011521843, + "grad_norm": 1.0004624288599213, + "learning_rate": 1.6708103611853638e-06, + "loss": 0.3596, + "step": 3954 + }, + { + "epoch": 3.7959673547767645, + "grad_norm": 0.8841693193064396, + "learning_rate": 1.6683117153276878e-06, + "loss": 0.5316, + "step": 3955 + }, + { + "epoch": 3.7969275084013443, + "grad_norm": 1.1553617748917355, + "learning_rate": 1.6658145650292707e-06, + "loss": 0.5447, + "step": 3956 + }, + { + "epoch": 3.797887662025924, + "grad_norm": 1.0770525668556672, + "learning_rate": 1.6633189114110531e-06, + "loss": 0.4996, + "step": 3957 + }, + { + "epoch": 3.798847815650504, + "grad_norm": 1.2620691453511903, + "learning_rate": 1.6608247555933093e-06, + "loss": 0.4501, + "step": 3958 + }, + { + "epoch": 3.799807969275084, + "grad_norm": 1.4312504220509992, + "learning_rate": 1.6583320986956386e-06, + "loss": 0.4568, + "step": 3959 + }, + { + "epoch": 3.800768122899664, + "grad_norm": 1.3747786124916295, + "learning_rate": 1.6558409418369686e-06, + "loss": 0.6672, + "step": 3960 + }, + { + "epoch": 3.8017282765242437, + "grad_norm": 1.1357784839797793, + "learning_rate": 1.6533512861355543e-06, + "loss": 0.4539, + "step": 3961 + }, + { + "epoch": 3.802688430148824, + "grad_norm": 1.4784324229639423, + "learning_rate": 1.6508631327089725e-06, + "loss": 0.4767, + "step": 3962 + }, + { + "epoch": 3.8036485837734038, + "grad_norm": 1.749657398940357, + "learning_rate": 1.64837648267413e-06, + "loss": 0.528, + "step": 3963 + }, + { + "epoch": 3.8046087373979836, + "grad_norm": 1.4001479052399646, + "learning_rate": 1.6458913371472573e-06, + "loss": 0.4379, + "step": 3964 + }, + { + "epoch": 3.8055688910225633, + "grad_norm": 0.9724248120840239, + "learning_rate": 1.6434076972439106e-06, + "loss": 0.4493, + "step": 3965 + }, + { + "epoch": 3.8065290446471436, + "grad_norm": 1.2089777865100662, + "learning_rate": 1.6409255640789695e-06, + "loss": 0.3999, + "step": 3966 + }, + { + "epoch": 3.8074891982717234, + "grad_norm": 1.1890139752612174, + "learning_rate": 1.6384449387666351e-06, + "loss": 0.3957, + "step": 3967 + }, + { + "epoch": 3.8084493518963036, + "grad_norm": 1.0918089262696387, + "learning_rate": 1.6359658224204328e-06, + "loss": 0.5489, + "step": 3968 + }, + { + "epoch": 3.8094095055208834, + "grad_norm": 1.2735572997914448, + "learning_rate": 1.6334882161532161e-06, + "loss": 0.4587, + "step": 3969 + }, + { + "epoch": 3.810369659145463, + "grad_norm": 1.203926288656431, + "learning_rate": 1.6310121210771524e-06, + "loss": 0.4519, + "step": 3970 + }, + { + "epoch": 3.811329812770043, + "grad_norm": 1.2969788010262875, + "learning_rate": 1.628537538303736e-06, + "loss": 0.5325, + "step": 3971 + }, + { + "epoch": 3.8122899663946233, + "grad_norm": 1.1989039386120783, + "learning_rate": 1.6260644689437767e-06, + "loss": 0.4702, + "step": 3972 + }, + { + "epoch": 3.813250120019203, + "grad_norm": 1.0865728623294701, + "learning_rate": 1.6235929141074137e-06, + "loss": 0.419, + "step": 3973 + }, + { + "epoch": 3.814210273643783, + "grad_norm": 1.1972929334935527, + "learning_rate": 1.6211228749041007e-06, + "loss": 0.4221, + "step": 3974 + }, + { + "epoch": 3.815170427268363, + "grad_norm": 1.2058629895169104, + "learning_rate": 1.6186543524426101e-06, + "loss": 0.4712, + "step": 3975 + }, + { + "epoch": 3.816130580892943, + "grad_norm": 0.7150639392966994, + "learning_rate": 1.616187347831036e-06, + "loss": 0.3704, + "step": 3976 + }, + { + "epoch": 3.8170907345175227, + "grad_norm": 1.4134564371765639, + "learning_rate": 1.6137218621767902e-06, + "loss": 0.4629, + "step": 3977 + }, + { + "epoch": 3.8180508881421025, + "grad_norm": 1.0285428509236896, + "learning_rate": 1.6112578965866039e-06, + "loss": 0.4958, + "step": 3978 + }, + { + "epoch": 3.8190110417666827, + "grad_norm": 1.1106919237603665, + "learning_rate": 1.608795452166525e-06, + "loss": 0.5155, + "step": 3979 + }, + { + "epoch": 3.8199711953912625, + "grad_norm": 0.9984189443064169, + "learning_rate": 1.6063345300219162e-06, + "loss": 0.4263, + "step": 3980 + }, + { + "epoch": 3.8209313490158427, + "grad_norm": 1.0060149288925493, + "learning_rate": 1.6038751312574592e-06, + "loss": 0.5008, + "step": 3981 + }, + { + "epoch": 3.8218915026404225, + "grad_norm": 0.9101782682192767, + "learning_rate": 1.6014172569771564e-06, + "loss": 0.4947, + "step": 3982 + }, + { + "epoch": 3.8228516562650023, + "grad_norm": 1.1453386082332564, + "learning_rate": 1.598960908284316e-06, + "loss": 0.4514, + "step": 3983 + }, + { + "epoch": 3.823811809889582, + "grad_norm": 1.5280230530018748, + "learning_rate": 1.5965060862815713e-06, + "loss": 0.5467, + "step": 3984 + }, + { + "epoch": 3.8247719635141624, + "grad_norm": 1.5398681237936402, + "learning_rate": 1.5940527920708594e-06, + "loss": 0.4875, + "step": 3985 + }, + { + "epoch": 3.825732117138742, + "grad_norm": 1.3715231941699455, + "learning_rate": 1.5916010267534433e-06, + "loss": 0.5491, + "step": 3986 + }, + { + "epoch": 3.826692270763322, + "grad_norm": 1.132827396312227, + "learning_rate": 1.5891507914298936e-06, + "loss": 0.3903, + "step": 3987 + }, + { + "epoch": 3.827652424387902, + "grad_norm": 1.11042335309336, + "learning_rate": 1.5867020872000927e-06, + "loss": 0.4469, + "step": 3988 + }, + { + "epoch": 3.828612578012482, + "grad_norm": 1.2977517517389103, + "learning_rate": 1.5842549151632386e-06, + "loss": 0.4846, + "step": 3989 + }, + { + "epoch": 3.829572731637062, + "grad_norm": 1.195303421024155, + "learning_rate": 1.5818092764178405e-06, + "loss": 0.48, + "step": 3990 + }, + { + "epoch": 3.8305328852616416, + "grad_norm": 1.0073063884953934, + "learning_rate": 1.5793651720617203e-06, + "loss": 0.5019, + "step": 3991 + }, + { + "epoch": 3.831493038886222, + "grad_norm": 0.9365326092207242, + "learning_rate": 1.576922603192011e-06, + "loss": 0.5142, + "step": 3992 + }, + { + "epoch": 3.8324531925108016, + "grad_norm": 0.8995860933097729, + "learning_rate": 1.5744815709051532e-06, + "loss": 0.5226, + "step": 3993 + }, + { + "epoch": 3.833413346135382, + "grad_norm": 1.3349663616771277, + "learning_rate": 1.5720420762969012e-06, + "loss": 0.4431, + "step": 3994 + }, + { + "epoch": 3.8343734997599617, + "grad_norm": 1.5557802515426742, + "learning_rate": 1.5696041204623185e-06, + "loss": 0.483, + "step": 3995 + }, + { + "epoch": 3.8353336533845415, + "grad_norm": 1.0883915020823676, + "learning_rate": 1.5671677044957772e-06, + "loss": 0.4781, + "step": 3996 + }, + { + "epoch": 3.8362938070091213, + "grad_norm": 1.2148251581833054, + "learning_rate": 1.5647328294909598e-06, + "loss": 0.4284, + "step": 3997 + }, + { + "epoch": 3.8372539606337015, + "grad_norm": 1.245980954092881, + "learning_rate": 1.5622994965408527e-06, + "loss": 0.5778, + "step": 3998 + }, + { + "epoch": 3.8382141142582813, + "grad_norm": 1.2141459957545226, + "learning_rate": 1.5598677067377533e-06, + "loss": 0.4211, + "step": 3999 + }, + { + "epoch": 3.839174267882861, + "grad_norm": 1.2168386206877675, + "learning_rate": 1.55743746117327e-06, + "loss": 0.5021, + "step": 4000 + }, + { + "epoch": 3.8401344215074413, + "grad_norm": 0.9708078696586657, + "learning_rate": 1.5550087609383102e-06, + "loss": 0.5214, + "step": 4001 + }, + { + "epoch": 3.841094575132021, + "grad_norm": 1.0115662460484836, + "learning_rate": 1.5525816071230932e-06, + "loss": 0.4427, + "step": 4002 + }, + { + "epoch": 3.842054728756601, + "grad_norm": 1.1549305352553938, + "learning_rate": 1.5501560008171412e-06, + "loss": 0.5273, + "step": 4003 + }, + { + "epoch": 3.8430148823811807, + "grad_norm": 0.9841200403606181, + "learning_rate": 1.5477319431092835e-06, + "loss": 0.4897, + "step": 4004 + }, + { + "epoch": 3.843975036005761, + "grad_norm": 1.3534935188894743, + "learning_rate": 1.5453094350876563e-06, + "loss": 0.4788, + "step": 4005 + }, + { + "epoch": 3.8449351896303408, + "grad_norm": 0.9454490671210521, + "learning_rate": 1.5428884778396929e-06, + "loss": 0.4947, + "step": 4006 + }, + { + "epoch": 3.845895343254921, + "grad_norm": 0.982085650772257, + "learning_rate": 1.540469072452137e-06, + "loss": 0.5073, + "step": 4007 + }, + { + "epoch": 3.846855496879501, + "grad_norm": 1.1183833419766704, + "learning_rate": 1.5380512200110347e-06, + "loss": 0.4687, + "step": 4008 + }, + { + "epoch": 3.8478156505040806, + "grad_norm": 1.2972102500875846, + "learning_rate": 1.5356349216017336e-06, + "loss": 0.4551, + "step": 4009 + }, + { + "epoch": 3.8487758041286604, + "grad_norm": 1.7060004299228904, + "learning_rate": 1.5332201783088857e-06, + "loss": 0.4385, + "step": 4010 + }, + { + "epoch": 3.8497359577532406, + "grad_norm": 1.0363932977147206, + "learning_rate": 1.5308069912164403e-06, + "loss": 0.4592, + "step": 4011 + }, + { + "epoch": 3.8506961113778204, + "grad_norm": 0.9486440871438606, + "learning_rate": 1.5283953614076514e-06, + "loss": 0.5234, + "step": 4012 + }, + { + "epoch": 3.8516562650024007, + "grad_norm": 1.1287504733298028, + "learning_rate": 1.5259852899650784e-06, + "loss": 0.443, + "step": 4013 + }, + { + "epoch": 3.8526164186269805, + "grad_norm": 1.7121590389272714, + "learning_rate": 1.5235767779705717e-06, + "loss": 0.4277, + "step": 4014 + }, + { + "epoch": 3.8535765722515603, + "grad_norm": 1.1436615771183936, + "learning_rate": 1.5211698265052905e-06, + "loss": 0.5004, + "step": 4015 + }, + { + "epoch": 3.85453672587614, + "grad_norm": 0.9259865424761593, + "learning_rate": 1.518764436649684e-06, + "loss": 0.4672, + "step": 4016 + }, + { + "epoch": 3.85549687950072, + "grad_norm": 0.8042351206260997, + "learning_rate": 1.5163606094835115e-06, + "loss": 0.4571, + "step": 4017 + }, + { + "epoch": 3.8564570331253, + "grad_norm": 1.0242287750735588, + "learning_rate": 1.5139583460858249e-06, + "loss": 0.5616, + "step": 4018 + }, + { + "epoch": 3.85741718674988, + "grad_norm": 1.3296288625498407, + "learning_rate": 1.5115576475349715e-06, + "loss": 0.4741, + "step": 4019 + }, + { + "epoch": 3.85837734037446, + "grad_norm": 1.383565930716489, + "learning_rate": 1.5091585149086008e-06, + "loss": 0.4244, + "step": 4020 + }, + { + "epoch": 3.85933749399904, + "grad_norm": 1.118198473425921, + "learning_rate": 1.5067609492836572e-06, + "loss": 0.5082, + "step": 4021 + }, + { + "epoch": 3.8602976476236197, + "grad_norm": 1.3476105788891708, + "learning_rate": 1.5043649517363835e-06, + "loss": 0.4172, + "step": 4022 + }, + { + "epoch": 3.8612578012481995, + "grad_norm": 1.862910169367962, + "learning_rate": 1.5019705233423177e-06, + "loss": 0.5179, + "step": 4023 + }, + { + "epoch": 3.8622179548727797, + "grad_norm": 0.9927431150646678, + "learning_rate": 1.4995776651762906e-06, + "loss": 0.521, + "step": 4024 + }, + { + "epoch": 3.8631781084973595, + "grad_norm": 0.8802457449186407, + "learning_rate": 1.4971863783124308e-06, + "loss": 0.5404, + "step": 4025 + }, + { + "epoch": 3.86413826212194, + "grad_norm": 1.0088620252724212, + "learning_rate": 1.4947966638241662e-06, + "loss": 0.4715, + "step": 4026 + }, + { + "epoch": 3.8650984157465196, + "grad_norm": 1.1755045780960445, + "learning_rate": 1.4924085227842084e-06, + "loss": 0.4483, + "step": 4027 + }, + { + "epoch": 3.8660585693710994, + "grad_norm": 1.0702719556324969, + "learning_rate": 1.490021956264573e-06, + "loss": 0.4258, + "step": 4028 + }, + { + "epoch": 3.867018722995679, + "grad_norm": 0.8775672589847451, + "learning_rate": 1.487636965336558e-06, + "loss": 0.3576, + "step": 4029 + }, + { + "epoch": 3.867978876620259, + "grad_norm": 1.0924174648049145, + "learning_rate": 1.4852535510707662e-06, + "loss": 0.4449, + "step": 4030 + }, + { + "epoch": 3.868939030244839, + "grad_norm": 1.0990688153958843, + "learning_rate": 1.4828717145370863e-06, + "loss": 0.4891, + "step": 4031 + }, + { + "epoch": 3.869899183869419, + "grad_norm": 1.121158881487208, + "learning_rate": 1.4804914568046957e-06, + "loss": 0.4821, + "step": 4032 + }, + { + "epoch": 3.8708593374939992, + "grad_norm": 1.0268608560470882, + "learning_rate": 1.478112778942069e-06, + "loss": 0.4799, + "step": 4033 + }, + { + "epoch": 3.871819491118579, + "grad_norm": 1.4615768454751372, + "learning_rate": 1.4757356820169694e-06, + "loss": 0.4493, + "step": 4034 + }, + { + "epoch": 3.872779644743159, + "grad_norm": 1.7542196610056557, + "learning_rate": 1.4733601670964493e-06, + "loss": 0.4186, + "step": 4035 + }, + { + "epoch": 3.8737397983677386, + "grad_norm": 1.0868446926637694, + "learning_rate": 1.4709862352468541e-06, + "loss": 0.506, + "step": 4036 + }, + { + "epoch": 3.874699951992319, + "grad_norm": 1.1091506558753939, + "learning_rate": 1.4686138875338136e-06, + "loss": 0.475, + "step": 4037 + }, + { + "epoch": 3.8756601056168987, + "grad_norm": 1.4347493014909358, + "learning_rate": 1.4662431250222504e-06, + "loss": 0.4325, + "step": 4038 + }, + { + "epoch": 3.876620259241479, + "grad_norm": 0.8404888487332063, + "learning_rate": 1.4638739487763743e-06, + "loss": 0.3987, + "step": 4039 + }, + { + "epoch": 3.8775804128660587, + "grad_norm": 1.183921069828481, + "learning_rate": 1.4615063598596834e-06, + "loss": 0.5731, + "step": 4040 + }, + { + "epoch": 3.8785405664906385, + "grad_norm": 1.0639979381050415, + "learning_rate": 1.4591403593349652e-06, + "loss": 0.5338, + "step": 4041 + }, + { + "epoch": 3.8795007201152183, + "grad_norm": 1.1468457294566843, + "learning_rate": 1.4567759482642884e-06, + "loss": 0.4024, + "step": 4042 + }, + { + "epoch": 3.8804608737397985, + "grad_norm": 1.146199396866093, + "learning_rate": 1.4544131277090118e-06, + "loss": 0.4707, + "step": 4043 + }, + { + "epoch": 3.8814210273643783, + "grad_norm": 1.2054451240412933, + "learning_rate": 1.4520518987297854e-06, + "loss": 0.4779, + "step": 4044 + }, + { + "epoch": 3.882381180988958, + "grad_norm": 1.2773333772625617, + "learning_rate": 1.449692262386534e-06, + "loss": 0.4464, + "step": 4045 + }, + { + "epoch": 3.8833413346135384, + "grad_norm": 0.9227110359686093, + "learning_rate": 1.447334219738477e-06, + "loss": 0.5324, + "step": 4046 + }, + { + "epoch": 3.884301488238118, + "grad_norm": 1.3731221175775805, + "learning_rate": 1.4449777718441106e-06, + "loss": 0.5106, + "step": 4047 + }, + { + "epoch": 3.885261641862698, + "grad_norm": 1.2453865008405958, + "learning_rate": 1.4426229197612217e-06, + "loss": 0.5369, + "step": 4048 + }, + { + "epoch": 3.8862217954872778, + "grad_norm": 1.1758688469859149, + "learning_rate": 1.44026966454688e-06, + "loss": 0.5168, + "step": 4049 + }, + { + "epoch": 3.887181949111858, + "grad_norm": 1.1989467853498457, + "learning_rate": 1.4379180072574335e-06, + "loss": 0.5195, + "step": 4050 + }, + { + "epoch": 3.888142102736438, + "grad_norm": 1.1939247580457146, + "learning_rate": 1.4355679489485158e-06, + "loss": 0.4305, + "step": 4051 + }, + { + "epoch": 3.889102256361018, + "grad_norm": 0.6023502428972107, + "learning_rate": 1.433219490675045e-06, + "loss": 0.4727, + "step": 4052 + }, + { + "epoch": 3.890062409985598, + "grad_norm": 0.8724516734586564, + "learning_rate": 1.4308726334912176e-06, + "loss": 0.4893, + "step": 4053 + }, + { + "epoch": 3.8910225636101776, + "grad_norm": 1.2127240783439697, + "learning_rate": 1.4285273784505154e-06, + "loss": 0.4933, + "step": 4054 + }, + { + "epoch": 3.8919827172347574, + "grad_norm": 0.9695097636780937, + "learning_rate": 1.4261837266056938e-06, + "loss": 0.5214, + "step": 4055 + }, + { + "epoch": 3.8929428708593377, + "grad_norm": 1.160597549486381, + "learning_rate": 1.4238416790087945e-06, + "loss": 0.4482, + "step": 4056 + }, + { + "epoch": 3.8939030244839175, + "grad_norm": 0.8305336831110708, + "learning_rate": 1.4215012367111408e-06, + "loss": 0.4585, + "step": 4057 + }, + { + "epoch": 3.8948631781084972, + "grad_norm": 0.9272193865571398, + "learning_rate": 1.4191624007633286e-06, + "loss": 0.5662, + "step": 4058 + }, + { + "epoch": 3.8958233317330775, + "grad_norm": 1.0105506991936097, + "learning_rate": 1.41682517221524e-06, + "loss": 0.4465, + "step": 4059 + }, + { + "epoch": 3.8967834853576573, + "grad_norm": 1.1225112503216266, + "learning_rate": 1.4144895521160257e-06, + "loss": 0.5328, + "step": 4060 + }, + { + "epoch": 3.897743638982237, + "grad_norm": 1.279152969448898, + "learning_rate": 1.412155541514126e-06, + "loss": 0.5269, + "step": 4061 + }, + { + "epoch": 3.898703792606817, + "grad_norm": 1.8388135926587599, + "learning_rate": 1.4098231414572534e-06, + "loss": 0.4074, + "step": 4062 + }, + { + "epoch": 3.899663946231397, + "grad_norm": 1.6045605805300485, + "learning_rate": 1.4074923529923945e-06, + "loss": 0.566, + "step": 4063 + }, + { + "epoch": 3.900624099855977, + "grad_norm": 0.8785585405375341, + "learning_rate": 1.4051631771658159e-06, + "loss": 0.4144, + "step": 4064 + }, + { + "epoch": 3.901584253480557, + "grad_norm": 1.032110270713385, + "learning_rate": 1.402835615023061e-06, + "loss": 0.4382, + "step": 4065 + }, + { + "epoch": 3.902544407105137, + "grad_norm": 1.395893515938224, + "learning_rate": 1.400509667608947e-06, + "loss": 0.397, + "step": 4066 + }, + { + "epoch": 3.9035045607297167, + "grad_norm": 1.320399818859656, + "learning_rate": 1.398185335967568e-06, + "loss": 0.4619, + "step": 4067 + }, + { + "epoch": 3.9044647143542965, + "grad_norm": 1.1526727439294775, + "learning_rate": 1.3958626211422887e-06, + "loss": 0.3535, + "step": 4068 + }, + { + "epoch": 3.905424867978877, + "grad_norm": 1.4135671362045452, + "learning_rate": 1.3935415241757517e-06, + "loss": 0.5228, + "step": 4069 + }, + { + "epoch": 3.9063850216034566, + "grad_norm": 0.8189534748681332, + "learning_rate": 1.3912220461098763e-06, + "loss": 0.4531, + "step": 4070 + }, + { + "epoch": 3.9073451752280364, + "grad_norm": 1.1263794949554329, + "learning_rate": 1.3889041879858472e-06, + "loss": 0.4646, + "step": 4071 + }, + { + "epoch": 3.9083053288526166, + "grad_norm": 1.5344478431354853, + "learning_rate": 1.3865879508441294e-06, + "loss": 0.4896, + "step": 4072 + }, + { + "epoch": 3.9092654824771964, + "grad_norm": 1.392489823734356, + "learning_rate": 1.384273335724452e-06, + "loss": 0.5474, + "step": 4073 + }, + { + "epoch": 3.910225636101776, + "grad_norm": 1.170714564580904, + "learning_rate": 1.3819603436658258e-06, + "loss": 0.4303, + "step": 4074 + }, + { + "epoch": 3.911185789726356, + "grad_norm": 1.1368734226893547, + "learning_rate": 1.3796489757065284e-06, + "loss": 0.5351, + "step": 4075 + }, + { + "epoch": 3.9121459433509362, + "grad_norm": 1.0116935997285494, + "learning_rate": 1.3773392328841046e-06, + "loss": 0.5334, + "step": 4076 + }, + { + "epoch": 3.913106096975516, + "grad_norm": 1.2343182848929284, + "learning_rate": 1.3750311162353752e-06, + "loss": 0.4632, + "step": 4077 + }, + { + "epoch": 3.9140662506000963, + "grad_norm": 1.0308702206135525, + "learning_rate": 1.3727246267964288e-06, + "loss": 0.4387, + "step": 4078 + }, + { + "epoch": 3.915026404224676, + "grad_norm": 1.0340175430057703, + "learning_rate": 1.3704197656026242e-06, + "loss": 0.3849, + "step": 4079 + }, + { + "epoch": 3.915986557849256, + "grad_norm": 1.3134001215269133, + "learning_rate": 1.36811653368859e-06, + "loss": 0.5156, + "step": 4080 + }, + { + "epoch": 3.9169467114738357, + "grad_norm": 0.7877048705715364, + "learning_rate": 1.3658149320882196e-06, + "loss": 0.5337, + "step": 4081 + }, + { + "epoch": 3.917906865098416, + "grad_norm": 1.0372241944209186, + "learning_rate": 1.3635149618346793e-06, + "loss": 0.4775, + "step": 4082 + }, + { + "epoch": 3.9188670187229957, + "grad_norm": 1.4545514153391248, + "learning_rate": 1.3612166239604003e-06, + "loss": 0.4189, + "step": 4083 + }, + { + "epoch": 3.9198271723475755, + "grad_norm": 1.6400807256631689, + "learning_rate": 1.3589199194970825e-06, + "loss": 0.38, + "step": 4084 + }, + { + "epoch": 3.9207873259721557, + "grad_norm": 1.1161324377796293, + "learning_rate": 1.3566248494756934e-06, + "loss": 0.4868, + "step": 4085 + }, + { + "epoch": 3.9217474795967355, + "grad_norm": 0.8883548552685723, + "learning_rate": 1.3543314149264625e-06, + "loss": 0.5214, + "step": 4086 + }, + { + "epoch": 3.9227076332213153, + "grad_norm": 1.2708272295752705, + "learning_rate": 1.352039616878888e-06, + "loss": 0.3808, + "step": 4087 + }, + { + "epoch": 3.923667786845895, + "grad_norm": 0.8954602689209706, + "learning_rate": 1.3497494563617375e-06, + "loss": 0.488, + "step": 4088 + }, + { + "epoch": 3.9246279404704754, + "grad_norm": 1.4156612884362025, + "learning_rate": 1.347460934403036e-06, + "loss": 0.45, + "step": 4089 + }, + { + "epoch": 3.925588094095055, + "grad_norm": 1.18479486203872, + "learning_rate": 1.3451740520300787e-06, + "loss": 0.4398, + "step": 4090 + }, + { + "epoch": 3.9265482477196354, + "grad_norm": 1.0831389027518687, + "learning_rate": 1.3428888102694187e-06, + "loss": 0.4874, + "step": 4091 + }, + { + "epoch": 3.927508401344215, + "grad_norm": 0.9954881545826475, + "learning_rate": 1.3406052101468808e-06, + "loss": 0.4879, + "step": 4092 + }, + { + "epoch": 3.928468554968795, + "grad_norm": 0.9124216707052788, + "learning_rate": 1.338323252687549e-06, + "loss": 0.4502, + "step": 4093 + }, + { + "epoch": 3.929428708593375, + "grad_norm": 1.2429916730888522, + "learning_rate": 1.3360429389157664e-06, + "loss": 0.3754, + "step": 4094 + }, + { + "epoch": 3.930388862217955, + "grad_norm": 1.7210578064205935, + "learning_rate": 1.3337642698551428e-06, + "loss": 0.3892, + "step": 4095 + }, + { + "epoch": 3.931349015842535, + "grad_norm": 1.2540445183477316, + "learning_rate": 1.331487246528549e-06, + "loss": 0.4671, + "step": 4096 + }, + { + "epoch": 3.9323091694671146, + "grad_norm": 1.1752729101137955, + "learning_rate": 1.3292118699581153e-06, + "loss": 0.4323, + "step": 4097 + }, + { + "epoch": 3.933269323091695, + "grad_norm": 0.9954437339603511, + "learning_rate": 1.3269381411652366e-06, + "loss": 0.4047, + "step": 4098 + }, + { + "epoch": 3.9342294767162747, + "grad_norm": 0.9524777868909293, + "learning_rate": 1.324666061170562e-06, + "loss": 0.6096, + "step": 4099 + }, + { + "epoch": 3.9351896303408544, + "grad_norm": 1.0750406642574184, + "learning_rate": 1.3223956309940038e-06, + "loss": 0.4719, + "step": 4100 + }, + { + "epoch": 3.9361497839654342, + "grad_norm": 1.1347251012040827, + "learning_rate": 1.3201268516547382e-06, + "loss": 0.5203, + "step": 4101 + }, + { + "epoch": 3.9371099375900145, + "grad_norm": 1.0944097306950127, + "learning_rate": 1.317859724171192e-06, + "loss": 0.4772, + "step": 4102 + }, + { + "epoch": 3.9380700912145943, + "grad_norm": 1.3944724725377775, + "learning_rate": 1.3155942495610574e-06, + "loss": 0.4865, + "step": 4103 + }, + { + "epoch": 3.9390302448391745, + "grad_norm": 1.393872376217508, + "learning_rate": 1.3133304288412768e-06, + "loss": 0.4211, + "step": 4104 + }, + { + "epoch": 3.9399903984637543, + "grad_norm": 1.3494932929508532, + "learning_rate": 1.3110682630280603e-06, + "loss": 0.5309, + "step": 4105 + }, + { + "epoch": 3.940950552088334, + "grad_norm": 1.17868788523819, + "learning_rate": 1.3088077531368686e-06, + "loss": 0.4527, + "step": 4106 + }, + { + "epoch": 3.941910705712914, + "grad_norm": 1.2245565349535994, + "learning_rate": 1.3065489001824194e-06, + "loss": 0.5102, + "step": 4107 + }, + { + "epoch": 3.942870859337494, + "grad_norm": 0.9162536687842219, + "learning_rate": 1.3042917051786873e-06, + "loss": 0.5234, + "step": 4108 + }, + { + "epoch": 3.943831012962074, + "grad_norm": 1.2520030427754436, + "learning_rate": 1.3020361691389043e-06, + "loss": 0.3865, + "step": 4109 + }, + { + "epoch": 3.9447911665866537, + "grad_norm": 1.0770822645392424, + "learning_rate": 1.299782293075556e-06, + "loss": 0.4241, + "step": 4110 + }, + { + "epoch": 3.945751320211234, + "grad_norm": 0.9285368301633085, + "learning_rate": 1.297530078000384e-06, + "loss": 0.3585, + "step": 4111 + }, + { + "epoch": 3.9467114738358138, + "grad_norm": 1.0344233599670831, + "learning_rate": 1.295279524924382e-06, + "loss": 0.4348, + "step": 4112 + }, + { + "epoch": 3.9476716274603936, + "grad_norm": 0.9897457390323293, + "learning_rate": 1.2930306348577982e-06, + "loss": 0.4411, + "step": 4113 + }, + { + "epoch": 3.9486317810849734, + "grad_norm": 1.008796582147279, + "learning_rate": 1.2907834088101402e-06, + "loss": 0.468, + "step": 4114 + }, + { + "epoch": 3.9495919347095536, + "grad_norm": 1.0919656417194719, + "learning_rate": 1.2885378477901595e-06, + "loss": 0.4164, + "step": 4115 + }, + { + "epoch": 3.9505520883341334, + "grad_norm": 1.059155080492062, + "learning_rate": 1.2862939528058665e-06, + "loss": 0.4024, + "step": 4116 + }, + { + "epoch": 3.9515122419587136, + "grad_norm": 0.9145561364466925, + "learning_rate": 1.284051724864518e-06, + "loss": 0.4815, + "step": 4117 + }, + { + "epoch": 3.9524723955832934, + "grad_norm": 1.2025386746578184, + "learning_rate": 1.2818111649726306e-06, + "loss": 0.3387, + "step": 4118 + }, + { + "epoch": 3.9534325492078732, + "grad_norm": 1.3173067493707236, + "learning_rate": 1.2795722741359672e-06, + "loss": 0.3966, + "step": 4119 + }, + { + "epoch": 3.954392702832453, + "grad_norm": 0.820376797118698, + "learning_rate": 1.277335053359539e-06, + "loss": 0.5225, + "step": 4120 + }, + { + "epoch": 3.9553528564570333, + "grad_norm": 1.194067571446378, + "learning_rate": 1.2750995036476133e-06, + "loss": 0.4774, + "step": 4121 + }, + { + "epoch": 3.956313010081613, + "grad_norm": 0.8784531337871889, + "learning_rate": 1.2728656260037025e-06, + "loss": 0.5147, + "step": 4122 + }, + { + "epoch": 3.957273163706193, + "grad_norm": 1.1533340960198126, + "learning_rate": 1.2706334214305716e-06, + "loss": 0.506, + "step": 4123 + }, + { + "epoch": 3.958233317330773, + "grad_norm": 1.6351270381674599, + "learning_rate": 1.2684028909302348e-06, + "loss": 0.351, + "step": 4124 + }, + { + "epoch": 3.959193470955353, + "grad_norm": 0.731829491187216, + "learning_rate": 1.2661740355039499e-06, + "loss": 0.4465, + "step": 4125 + }, + { + "epoch": 3.9601536245799327, + "grad_norm": 1.0640778289239052, + "learning_rate": 1.2639468561522284e-06, + "loss": 0.474, + "step": 4126 + }, + { + "epoch": 3.9611137782045125, + "grad_norm": 2.016771247252845, + "learning_rate": 1.2617213538748268e-06, + "loss": 0.5078, + "step": 4127 + }, + { + "epoch": 3.9620739318290927, + "grad_norm": 1.1841314710691146, + "learning_rate": 1.2594975296707495e-06, + "loss": 0.5035, + "step": 4128 + }, + { + "epoch": 3.9630340854536725, + "grad_norm": 1.1036671376508644, + "learning_rate": 1.2572753845382484e-06, + "loss": 0.4209, + "step": 4129 + }, + { + "epoch": 3.9639942390782528, + "grad_norm": 1.0417370270306738, + "learning_rate": 1.255054919474819e-06, + "loss": 0.5141, + "step": 4130 + }, + { + "epoch": 3.9649543927028326, + "grad_norm": 1.1148626570006415, + "learning_rate": 1.252836135477203e-06, + "loss": 0.534, + "step": 4131 + }, + { + "epoch": 3.9659145463274124, + "grad_norm": 0.9617271472038451, + "learning_rate": 1.2506190335413943e-06, + "loss": 0.4632, + "step": 4132 + }, + { + "epoch": 3.966874699951992, + "grad_norm": 0.7577777482038071, + "learning_rate": 1.2484036146626205e-06, + "loss": 0.5134, + "step": 4133 + }, + { + "epoch": 3.9678348535765724, + "grad_norm": 1.1175602462362233, + "learning_rate": 1.2461898798353639e-06, + "loss": 0.4669, + "step": 4134 + }, + { + "epoch": 3.968795007201152, + "grad_norm": 0.9213626342830816, + "learning_rate": 1.2439778300533412e-06, + "loss": 0.4953, + "step": 4135 + }, + { + "epoch": 3.969755160825732, + "grad_norm": 1.1904391217338572, + "learning_rate": 1.241767466309522e-06, + "loss": 0.3834, + "step": 4136 + }, + { + "epoch": 3.9707153144503122, + "grad_norm": 1.0471316329462108, + "learning_rate": 1.2395587895961158e-06, + "loss": 0.4629, + "step": 4137 + }, + { + "epoch": 3.971675468074892, + "grad_norm": 0.9616781362666617, + "learning_rate": 1.2373518009045705e-06, + "loss": 0.5947, + "step": 4138 + }, + { + "epoch": 3.972635621699472, + "grad_norm": 1.4420661170373539, + "learning_rate": 1.2351465012255808e-06, + "loss": 0.4961, + "step": 4139 + }, + { + "epoch": 3.9735957753240516, + "grad_norm": 1.1630442478403853, + "learning_rate": 1.232942891549083e-06, + "loss": 0.4432, + "step": 4140 + }, + { + "epoch": 3.974555928948632, + "grad_norm": 1.15212310608021, + "learning_rate": 1.2307409728642539e-06, + "loss": 0.5117, + "step": 4141 + }, + { + "epoch": 3.9755160825732117, + "grad_norm": 1.265953078329887, + "learning_rate": 1.2285407461595122e-06, + "loss": 0.5225, + "step": 4142 + }, + { + "epoch": 3.976476236197792, + "grad_norm": 1.2350500785784575, + "learning_rate": 1.2263422124225132e-06, + "loss": 0.4487, + "step": 4143 + }, + { + "epoch": 3.9774363898223717, + "grad_norm": 1.1117089141723477, + "learning_rate": 1.2241453726401559e-06, + "loss": 0.4174, + "step": 4144 + }, + { + "epoch": 3.9783965434469515, + "grad_norm": 0.8972382030907151, + "learning_rate": 1.2219502277985817e-06, + "loss": 0.4392, + "step": 4145 + }, + { + "epoch": 3.9793566970715313, + "grad_norm": 1.1884767201653634, + "learning_rate": 1.2197567788831638e-06, + "loss": 0.482, + "step": 4146 + }, + { + "epoch": 3.9803168506961115, + "grad_norm": 1.0956299108243066, + "learning_rate": 1.2175650268785205e-06, + "loss": 0.4247, + "step": 4147 + }, + { + "epoch": 3.9812770043206913, + "grad_norm": 0.9047473318465206, + "learning_rate": 1.215374972768502e-06, + "loss": 0.3904, + "step": 4148 + }, + { + "epoch": 3.982237157945271, + "grad_norm": 1.4375665503159347, + "learning_rate": 1.213186617536204e-06, + "loss": 0.5551, + "step": 4149 + }, + { + "epoch": 3.9831973115698514, + "grad_norm": 1.050766956265714, + "learning_rate": 1.2109999621639561e-06, + "loss": 0.4392, + "step": 4150 + }, + { + "epoch": 3.984157465194431, + "grad_norm": 1.1837777696740561, + "learning_rate": 1.2088150076333222e-06, + "loss": 0.5144, + "step": 4151 + }, + { + "epoch": 3.985117618819011, + "grad_norm": 1.4306616871611946, + "learning_rate": 1.2066317549251055e-06, + "loss": 0.4162, + "step": 4152 + }, + { + "epoch": 3.9860777724435907, + "grad_norm": 1.7089999247541572, + "learning_rate": 1.2044502050193451e-06, + "loss": 0.4145, + "step": 4153 + }, + { + "epoch": 3.987037926068171, + "grad_norm": 1.2206210523280627, + "learning_rate": 1.2022703588953155e-06, + "loss": 0.4053, + "step": 4154 + }, + { + "epoch": 3.9879980796927508, + "grad_norm": 1.0254775921148231, + "learning_rate": 1.2000922175315278e-06, + "loss": 0.5259, + "step": 4155 + }, + { + "epoch": 3.988958233317331, + "grad_norm": 1.1665630930484832, + "learning_rate": 1.197915781905723e-06, + "loss": 0.4422, + "step": 4156 + }, + { + "epoch": 3.989918386941911, + "grad_norm": 1.1001199001856836, + "learning_rate": 1.1957410529948803e-06, + "loss": 0.5368, + "step": 4157 + }, + { + "epoch": 3.9908785405664906, + "grad_norm": 1.075018529315739, + "learning_rate": 1.1935680317752173e-06, + "loss": 0.4422, + "step": 4158 + }, + { + "epoch": 3.9918386941910704, + "grad_norm": 0.7857597868944044, + "learning_rate": 1.1913967192221742e-06, + "loss": 0.529, + "step": 4159 + }, + { + "epoch": 3.9927988478156506, + "grad_norm": 0.9411861184100021, + "learning_rate": 1.189227116310434e-06, + "loss": 0.5995, + "step": 4160 + }, + { + "epoch": 3.9937590014402304, + "grad_norm": 1.2649531390346653, + "learning_rate": 1.1870592240139033e-06, + "loss": 0.4276, + "step": 4161 + }, + { + "epoch": 3.9947191550648102, + "grad_norm": 1.0633526496414132, + "learning_rate": 1.18489304330573e-06, + "loss": 0.429, + "step": 4162 + }, + { + "epoch": 3.9956793086893905, + "grad_norm": 1.0737598014237413, + "learning_rate": 1.18272857515829e-06, + "loss": 0.3883, + "step": 4163 + }, + { + "epoch": 3.9966394623139703, + "grad_norm": 1.4406649557413433, + "learning_rate": 1.1805658205431868e-06, + "loss": 0.4952, + "step": 4164 + }, + { + "epoch": 3.99759961593855, + "grad_norm": 0.9857827185154254, + "learning_rate": 1.1784047804312598e-06, + "loss": 0.4536, + "step": 4165 + }, + { + "epoch": 3.99855976956313, + "grad_norm": 0.9085969015255203, + "learning_rate": 1.1762454557925756e-06, + "loss": 0.492, + "step": 4166 + }, + { + "epoch": 3.99951992318771, + "grad_norm": 1.2522481515344492, + "learning_rate": 1.174087847596433e-06, + "loss": 0.3796, + "step": 4167 + }, + { + "epoch": 4.0, + "grad_norm": 2.3617478918866177, + "learning_rate": 1.1719319568113602e-06, + "loss": 0.4645, + "step": 4168 + }, + { + "epoch": 4.00096015362458, + "grad_norm": 1.0228089331127097, + "learning_rate": 1.1697777844051105e-06, + "loss": 0.4895, + "step": 4169 + }, + { + "epoch": 4.00192030724916, + "grad_norm": 1.4808354983085061, + "learning_rate": 1.1676253313446706e-06, + "loss": 0.4577, + "step": 4170 + }, + { + "epoch": 4.002880460873739, + "grad_norm": 1.1169265009334828, + "learning_rate": 1.165474598596254e-06, + "loss": 0.3596, + "step": 4171 + }, + { + "epoch": 4.00384061449832, + "grad_norm": 0.9610152744839893, + "learning_rate": 1.1633255871253013e-06, + "loss": 0.4821, + "step": 4172 + }, + { + "epoch": 4.0048007681229, + "grad_norm": 1.100819168823591, + "learning_rate": 1.161178297896482e-06, + "loss": 0.3886, + "step": 4173 + }, + { + "epoch": 4.00576092174748, + "grad_norm": 1.560344140574424, + "learning_rate": 1.1590327318736887e-06, + "loss": 0.3792, + "step": 4174 + }, + { + "epoch": 4.0067210753720595, + "grad_norm": 1.1638371153671265, + "learning_rate": 1.1568888900200432e-06, + "loss": 0.3656, + "step": 4175 + }, + { + "epoch": 4.007681228996639, + "grad_norm": 0.9152074067317886, + "learning_rate": 1.1547467732978967e-06, + "loss": 0.5235, + "step": 4176 + }, + { + "epoch": 4.008641382621219, + "grad_norm": 1.0009183461893372, + "learning_rate": 1.152606382668819e-06, + "loss": 0.5299, + "step": 4177 + }, + { + "epoch": 4.0096015362458, + "grad_norm": 1.3534478469832056, + "learning_rate": 1.1504677190936115e-06, + "loss": 0.4403, + "step": 4178 + }, + { + "epoch": 4.0105616898703795, + "grad_norm": 1.1330371387665013, + "learning_rate": 1.148330783532292e-06, + "loss": 0.4855, + "step": 4179 + }, + { + "epoch": 4.011521843494959, + "grad_norm": 1.2623186318649344, + "learning_rate": 1.1461955769441141e-06, + "loss": 0.4975, + "step": 4180 + }, + { + "epoch": 4.012481997119539, + "grad_norm": 1.2575709816030505, + "learning_rate": 1.144062100287548e-06, + "loss": 0.459, + "step": 4181 + }, + { + "epoch": 4.013442150744119, + "grad_norm": 0.9310782530766164, + "learning_rate": 1.1419303545202854e-06, + "loss": 0.3379, + "step": 4182 + }, + { + "epoch": 4.014402304368699, + "grad_norm": 1.483922328937811, + "learning_rate": 1.1398003405992468e-06, + "loss": 0.3825, + "step": 4183 + }, + { + "epoch": 4.0153624579932785, + "grad_norm": 1.2249750724630295, + "learning_rate": 1.1376720594805719e-06, + "loss": 0.428, + "step": 4184 + }, + { + "epoch": 4.016322611617859, + "grad_norm": 1.2657921422968323, + "learning_rate": 1.1355455121196234e-06, + "loss": 0.3395, + "step": 4185 + }, + { + "epoch": 4.017282765242439, + "grad_norm": 1.2012124135035076, + "learning_rate": 1.1334206994709874e-06, + "loss": 0.4152, + "step": 4186 + }, + { + "epoch": 4.018242918867019, + "grad_norm": 1.0569575118521102, + "learning_rate": 1.1312976224884665e-06, + "loss": 0.4954, + "step": 4187 + }, + { + "epoch": 4.019203072491599, + "grad_norm": 1.1067383421150563, + "learning_rate": 1.1291762821250867e-06, + "loss": 0.3141, + "step": 4188 + }, + { + "epoch": 4.020163226116178, + "grad_norm": 0.9944493363007124, + "learning_rate": 1.1270566793331006e-06, + "loss": 0.4416, + "step": 4189 + }, + { + "epoch": 4.021123379740758, + "grad_norm": 1.082323844191331, + "learning_rate": 1.1249388150639702e-06, + "loss": 0.4461, + "step": 4190 + }, + { + "epoch": 4.022083533365339, + "grad_norm": 1.43474408959, + "learning_rate": 1.1228226902683847e-06, + "loss": 0.4098, + "step": 4191 + }, + { + "epoch": 4.023043686989919, + "grad_norm": 0.844896093678417, + "learning_rate": 1.1207083058962465e-06, + "loss": 0.4262, + "step": 4192 + }, + { + "epoch": 4.0240038406144985, + "grad_norm": 0.8683584109714455, + "learning_rate": 1.1185956628966832e-06, + "loss": 0.4131, + "step": 4193 + }, + { + "epoch": 4.024963994239078, + "grad_norm": 0.9444029751557702, + "learning_rate": 1.1164847622180392e-06, + "loss": 0.5157, + "step": 4194 + }, + { + "epoch": 4.025924147863658, + "grad_norm": 0.9856182532147272, + "learning_rate": 1.1143756048078707e-06, + "loss": 0.5141, + "step": 4195 + }, + { + "epoch": 4.026884301488238, + "grad_norm": 1.0951961014372649, + "learning_rate": 1.1122681916129586e-06, + "loss": 0.4607, + "step": 4196 + }, + { + "epoch": 4.027844455112818, + "grad_norm": 0.9031442935433569, + "learning_rate": 1.1101625235792978e-06, + "loss": 0.5384, + "step": 4197 + }, + { + "epoch": 4.028804608737398, + "grad_norm": 1.4404063615658798, + "learning_rate": 1.1080586016521e-06, + "loss": 0.3734, + "step": 4198 + }, + { + "epoch": 4.029764762361978, + "grad_norm": 1.0883888695253605, + "learning_rate": 1.1059564267757945e-06, + "loss": 0.4364, + "step": 4199 + }, + { + "epoch": 4.030724915986558, + "grad_norm": 1.4977311605432297, + "learning_rate": 1.1038559998940228e-06, + "loss": 0.4881, + "step": 4200 + }, + { + "epoch": 4.031685069611138, + "grad_norm": 0.8467897334018155, + "learning_rate": 1.1017573219496453e-06, + "loss": 0.5446, + "step": 4201 + }, + { + "epoch": 4.0326452232357175, + "grad_norm": 1.494093787775511, + "learning_rate": 1.0996603938847361e-06, + "loss": 0.4865, + "step": 4202 + }, + { + "epoch": 4.033605376860297, + "grad_norm": 1.1172721868839053, + "learning_rate": 1.0975652166405836e-06, + "loss": 0.4379, + "step": 4203 + }, + { + "epoch": 4.034565530484878, + "grad_norm": 1.3357464579781297, + "learning_rate": 1.0954717911576923e-06, + "loss": 0.3648, + "step": 4204 + }, + { + "epoch": 4.035525684109458, + "grad_norm": 1.0494931367800207, + "learning_rate": 1.0933801183757742e-06, + "loss": 0.4397, + "step": 4205 + }, + { + "epoch": 4.036485837734038, + "grad_norm": 1.093847679981737, + "learning_rate": 1.0912901992337626e-06, + "loss": 0.5484, + "step": 4206 + }, + { + "epoch": 4.037445991358617, + "grad_norm": 1.0098251961439262, + "learning_rate": 1.0892020346698e-06, + "loss": 0.2999, + "step": 4207 + }, + { + "epoch": 4.038406144983197, + "grad_norm": 0.8950132279843459, + "learning_rate": 1.087115625621239e-06, + "loss": 0.4694, + "step": 4208 + }, + { + "epoch": 4.039366298607777, + "grad_norm": 1.4797419378563925, + "learning_rate": 1.0850309730246472e-06, + "loss": 0.3842, + "step": 4209 + }, + { + "epoch": 4.040326452232357, + "grad_norm": 1.0555006824431175, + "learning_rate": 1.0829480778158025e-06, + "loss": 0.486, + "step": 4210 + }, + { + "epoch": 4.041286605856937, + "grad_norm": 1.29843019699927, + "learning_rate": 1.0808669409296951e-06, + "loss": 0.3899, + "step": 4211 + }, + { + "epoch": 4.042246759481517, + "grad_norm": 1.2177826969482468, + "learning_rate": 1.0787875633005256e-06, + "loss": 0.4925, + "step": 4212 + }, + { + "epoch": 4.043206913106097, + "grad_norm": 1.6704436579707684, + "learning_rate": 1.0767099458617025e-06, + "loss": 0.5365, + "step": 4213 + }, + { + "epoch": 4.044167066730677, + "grad_norm": 1.0136148725987781, + "learning_rate": 1.0746340895458475e-06, + "loss": 0.4631, + "step": 4214 + }, + { + "epoch": 4.045127220355257, + "grad_norm": 0.8765041498810147, + "learning_rate": 1.0725599952847894e-06, + "loss": 0.4409, + "step": 4215 + }, + { + "epoch": 4.046087373979836, + "grad_norm": 0.9427533038272904, + "learning_rate": 1.070487664009567e-06, + "loss": 0.4685, + "step": 4216 + }, + { + "epoch": 4.047047527604417, + "grad_norm": 0.9933754655194416, + "learning_rate": 1.0684170966504298e-06, + "loss": 0.5583, + "step": 4217 + }, + { + "epoch": 4.048007681228997, + "grad_norm": 1.158250101522896, + "learning_rate": 1.0663482941368303e-06, + "loss": 0.4586, + "step": 4218 + }, + { + "epoch": 4.048967834853577, + "grad_norm": 0.7532692687341203, + "learning_rate": 1.064281257397432e-06, + "loss": 0.4246, + "step": 4219 + }, + { + "epoch": 4.0499279884781565, + "grad_norm": 0.8551367076440082, + "learning_rate": 1.062215987360109e-06, + "loss": 0.4799, + "step": 4220 + }, + { + "epoch": 4.050888142102736, + "grad_norm": 1.2141855585896926, + "learning_rate": 1.0601524849519357e-06, + "loss": 0.4289, + "step": 4221 + }, + { + "epoch": 4.051848295727316, + "grad_norm": 1.0090877809711172, + "learning_rate": 1.0580907510991982e-06, + "loss": 0.4573, + "step": 4222 + }, + { + "epoch": 4.052808449351896, + "grad_norm": 1.0433680921569715, + "learning_rate": 1.0560307867273838e-06, + "loss": 0.445, + "step": 4223 + }, + { + "epoch": 4.053768602976477, + "grad_norm": 0.9502679652663969, + "learning_rate": 1.0539725927611915e-06, + "loss": 0.4869, + "step": 4224 + }, + { + "epoch": 4.054728756601056, + "grad_norm": 0.9674777898369518, + "learning_rate": 1.0519161701245234e-06, + "loss": 0.3871, + "step": 4225 + }, + { + "epoch": 4.055688910225636, + "grad_norm": 1.3676812062287418, + "learning_rate": 1.049861519740482e-06, + "loss": 0.4337, + "step": 4226 + }, + { + "epoch": 4.056649063850216, + "grad_norm": 1.7586505787627171, + "learning_rate": 1.04780864253138e-06, + "loss": 0.4283, + "step": 4227 + }, + { + "epoch": 4.057609217474796, + "grad_norm": 1.073864155625145, + "learning_rate": 1.0457575394187314e-06, + "loss": 0.4644, + "step": 4228 + }, + { + "epoch": 4.0585693710993755, + "grad_norm": 1.1607167995579928, + "learning_rate": 1.043708211323255e-06, + "loss": 0.362, + "step": 4229 + }, + { + "epoch": 4.059529524723956, + "grad_norm": 1.491228533906927, + "learning_rate": 1.0416606591648737e-06, + "loss": 0.3021, + "step": 4230 + }, + { + "epoch": 4.060489678348536, + "grad_norm": 1.0147464090050025, + "learning_rate": 1.039614883862709e-06, + "loss": 0.4277, + "step": 4231 + }, + { + "epoch": 4.061449831973116, + "grad_norm": 1.3055224272493493, + "learning_rate": 1.037570886335087e-06, + "loss": 0.3747, + "step": 4232 + }, + { + "epoch": 4.062409985597696, + "grad_norm": 0.8366133380077349, + "learning_rate": 1.0355286674995413e-06, + "loss": 0.4936, + "step": 4233 + }, + { + "epoch": 4.063370139222275, + "grad_norm": 0.9246694172987252, + "learning_rate": 1.0334882282727971e-06, + "loss": 0.4579, + "step": 4234 + }, + { + "epoch": 4.064330292846855, + "grad_norm": 1.4034925818936077, + "learning_rate": 1.0314495695707887e-06, + "loss": 0.4711, + "step": 4235 + }, + { + "epoch": 4.065290446471435, + "grad_norm": 1.0809037197410558, + "learning_rate": 1.029412692308645e-06, + "loss": 0.4059, + "step": 4236 + }, + { + "epoch": 4.066250600096016, + "grad_norm": 1.3469432718341132, + "learning_rate": 1.0273775974007017e-06, + "loss": 0.3732, + "step": 4237 + }, + { + "epoch": 4.0672107537205955, + "grad_norm": 0.8597826676320224, + "learning_rate": 1.0253442857604911e-06, + "loss": 0.4137, + "step": 4238 + }, + { + "epoch": 4.068170907345175, + "grad_norm": 1.524769301533565, + "learning_rate": 1.0233127583007424e-06, + "loss": 0.3453, + "step": 4239 + }, + { + "epoch": 4.069131060969755, + "grad_norm": 1.1619159943005153, + "learning_rate": 1.0212830159333886e-06, + "loss": 0.4099, + "step": 4240 + }, + { + "epoch": 4.070091214594335, + "grad_norm": 1.0543571257877578, + "learning_rate": 1.0192550595695584e-06, + "loss": 0.4851, + "step": 4241 + }, + { + "epoch": 4.071051368218915, + "grad_norm": 0.9272472552411923, + "learning_rate": 1.0172288901195805e-06, + "loss": 0.3227, + "step": 4242 + }, + { + "epoch": 4.072011521843495, + "grad_norm": 1.1430402130845945, + "learning_rate": 1.0152045084929813e-06, + "loss": 0.4766, + "step": 4243 + }, + { + "epoch": 4.072971675468075, + "grad_norm": 0.9113893300393455, + "learning_rate": 1.0131819155984818e-06, + "loss": 0.3991, + "step": 4244 + }, + { + "epoch": 4.073931829092655, + "grad_norm": 1.1436498993759299, + "learning_rate": 1.0111611123440041e-06, + "loss": 0.4557, + "step": 4245 + }, + { + "epoch": 4.074891982717235, + "grad_norm": 1.416402183830115, + "learning_rate": 1.0091420996366646e-06, + "loss": 0.4001, + "step": 4246 + }, + { + "epoch": 4.0758521363418145, + "grad_norm": 1.1313431713196778, + "learning_rate": 1.0071248783827764e-06, + "loss": 0.4447, + "step": 4247 + }, + { + "epoch": 4.076812289966394, + "grad_norm": 0.9859324415971613, + "learning_rate": 1.0051094494878505e-06, + "loss": 0.348, + "step": 4248 + }, + { + "epoch": 4.077772443590974, + "grad_norm": 0.917928864218637, + "learning_rate": 1.0030958138565883e-06, + "loss": 0.4204, + "step": 4249 + }, + { + "epoch": 4.078732597215555, + "grad_norm": 1.0581029548257181, + "learning_rate": 1.001083972392889e-06, + "loss": 0.4339, + "step": 4250 + }, + { + "epoch": 4.079692750840135, + "grad_norm": 1.178454031735879, + "learning_rate": 9.990739259998517e-07, + "loss": 0.4464, + "step": 4251 + }, + { + "epoch": 4.080652904464714, + "grad_norm": 0.7987610911287786, + "learning_rate": 9.970656755797598e-07, + "loss": 0.5038, + "step": 4252 + }, + { + "epoch": 4.081613058089294, + "grad_norm": 0.9193583831389514, + "learning_rate": 9.950592220340972e-07, + "loss": 0.4981, + "step": 4253 + }, + { + "epoch": 4.082573211713874, + "grad_norm": 0.8787063023269425, + "learning_rate": 9.930545662635393e-07, + "loss": 0.4153, + "step": 4254 + }, + { + "epoch": 4.083533365338454, + "grad_norm": 1.3368317566526455, + "learning_rate": 9.91051709167955e-07, + "loss": 0.3974, + "step": 4255 + }, + { + "epoch": 4.0844935189630345, + "grad_norm": 1.2647051227807544, + "learning_rate": 9.890506516464066e-07, + "loss": 0.425, + "step": 4256 + }, + { + "epoch": 4.085453672587614, + "grad_norm": 1.083904419924026, + "learning_rate": 9.870513945971444e-07, + "loss": 0.4483, + "step": 4257 + }, + { + "epoch": 4.086413826212194, + "grad_norm": 0.8488700025127257, + "learning_rate": 9.85053938917615e-07, + "loss": 0.4878, + "step": 4258 + }, + { + "epoch": 4.087373979836774, + "grad_norm": 0.8706164855245921, + "learning_rate": 9.83058285504455e-07, + "loss": 0.4989, + "step": 4259 + }, + { + "epoch": 4.088334133461354, + "grad_norm": 0.8575262207094165, + "learning_rate": 9.81064435253492e-07, + "loss": 0.4784, + "step": 4260 + }, + { + "epoch": 4.0892942870859335, + "grad_norm": 0.9887605308451878, + "learning_rate": 9.790723890597452e-07, + "loss": 0.4881, + "step": 4261 + }, + { + "epoch": 4.090254440710513, + "grad_norm": 1.1387912018631177, + "learning_rate": 9.770821478174207e-07, + "loss": 0.4661, + "step": 4262 + }, + { + "epoch": 4.091214594335094, + "grad_norm": 1.1655639448301596, + "learning_rate": 9.750937124199162e-07, + "loss": 0.4145, + "step": 4263 + }, + { + "epoch": 4.092174747959674, + "grad_norm": 1.0484353735066698, + "learning_rate": 9.731070837598227e-07, + "loss": 0.5036, + "step": 4264 + }, + { + "epoch": 4.0931349015842535, + "grad_norm": 1.4390833232501383, + "learning_rate": 9.71122262728913e-07, + "loss": 0.4491, + "step": 4265 + }, + { + "epoch": 4.094095055208833, + "grad_norm": 0.7749814621177304, + "learning_rate": 9.69139250218154e-07, + "loss": 0.4536, + "step": 4266 + }, + { + "epoch": 4.095055208833413, + "grad_norm": 1.4884168779464448, + "learning_rate": 9.67158047117696e-07, + "loss": 0.3593, + "step": 4267 + }, + { + "epoch": 4.096015362457993, + "grad_norm": 1.117155150474689, + "learning_rate": 9.651786543168834e-07, + "loss": 0.5245, + "step": 4268 + }, + { + "epoch": 4.096975516082574, + "grad_norm": 1.6872999701713718, + "learning_rate": 9.63201072704244e-07, + "loss": 0.4035, + "step": 4269 + }, + { + "epoch": 4.097935669707153, + "grad_norm": 0.954975072196789, + "learning_rate": 9.612253031674906e-07, + "loss": 0.3821, + "step": 4270 + }, + { + "epoch": 4.098895823331733, + "grad_norm": 0.951304544108571, + "learning_rate": 9.592513465935272e-07, + "loss": 0.5126, + "step": 4271 + }, + { + "epoch": 4.099855976956313, + "grad_norm": 0.921490814052639, + "learning_rate": 9.572792038684404e-07, + "loss": 0.4418, + "step": 4272 + }, + { + "epoch": 4.100816130580893, + "grad_norm": 1.7228555575088056, + "learning_rate": 9.553088758775048e-07, + "loss": 0.3493, + "step": 4273 + }, + { + "epoch": 4.101776284205473, + "grad_norm": 1.006455267047783, + "learning_rate": 9.533403635051813e-07, + "loss": 0.4744, + "step": 4274 + }, + { + "epoch": 4.102736437830052, + "grad_norm": 0.8733910351772793, + "learning_rate": 9.513736676351104e-07, + "loss": 0.4198, + "step": 4275 + }, + { + "epoch": 4.103696591454633, + "grad_norm": 1.4203884190388014, + "learning_rate": 9.494087891501213e-07, + "loss": 0.4427, + "step": 4276 + }, + { + "epoch": 4.104656745079213, + "grad_norm": 0.9757000809136535, + "learning_rate": 9.474457289322314e-07, + "loss": 0.5144, + "step": 4277 + }, + { + "epoch": 4.105616898703793, + "grad_norm": 1.1060393473010894, + "learning_rate": 9.454844878626329e-07, + "loss": 0.4511, + "step": 4278 + }, + { + "epoch": 4.1065770523283724, + "grad_norm": 1.3837663291815168, + "learning_rate": 9.435250668217088e-07, + "loss": 0.4659, + "step": 4279 + }, + { + "epoch": 4.107537205952952, + "grad_norm": 1.2887214163616825, + "learning_rate": 9.415674666890178e-07, + "loss": 0.4424, + "step": 4280 + }, + { + "epoch": 4.108497359577532, + "grad_norm": 1.1799751136383634, + "learning_rate": 9.396116883433104e-07, + "loss": 0.4207, + "step": 4281 + }, + { + "epoch": 4.109457513202113, + "grad_norm": 0.9030769312588787, + "learning_rate": 9.376577326625147e-07, + "loss": 0.5047, + "step": 4282 + }, + { + "epoch": 4.1104176668266925, + "grad_norm": 1.192159261759407, + "learning_rate": 9.35705600523737e-07, + "loss": 0.4577, + "step": 4283 + }, + { + "epoch": 4.111377820451272, + "grad_norm": 1.1049190893986787, + "learning_rate": 9.337552928032706e-07, + "loss": 0.4833, + "step": 4284 + }, + { + "epoch": 4.112337974075852, + "grad_norm": 0.8603276291123754, + "learning_rate": 9.318068103765871e-07, + "loss": 0.48, + "step": 4285 + }, + { + "epoch": 4.113298127700432, + "grad_norm": 0.95385485131521, + "learning_rate": 9.2986015411834e-07, + "loss": 0.4076, + "step": 4286 + }, + { + "epoch": 4.114258281325012, + "grad_norm": 1.0983527364540822, + "learning_rate": 9.279153249023637e-07, + "loss": 0.3698, + "step": 4287 + }, + { + "epoch": 4.1152184349495915, + "grad_norm": 0.7963206415542619, + "learning_rate": 9.259723236016688e-07, + "loss": 0.4433, + "step": 4288 + }, + { + "epoch": 4.116178588574172, + "grad_norm": 1.3193320038250071, + "learning_rate": 9.240311510884487e-07, + "loss": 0.3464, + "step": 4289 + }, + { + "epoch": 4.117138742198752, + "grad_norm": 1.058292953859479, + "learning_rate": 9.220918082340752e-07, + "loss": 0.4701, + "step": 4290 + }, + { + "epoch": 4.118098895823332, + "grad_norm": 0.8658520148648889, + "learning_rate": 9.201542959090987e-07, + "loss": 0.4166, + "step": 4291 + }, + { + "epoch": 4.119059049447912, + "grad_norm": 1.0574364621218113, + "learning_rate": 9.182186149832484e-07, + "loss": 0.5533, + "step": 4292 + }, + { + "epoch": 4.120019203072491, + "grad_norm": 0.94841809193125, + "learning_rate": 9.162847663254292e-07, + "loss": 0.4965, + "step": 4293 + }, + { + "epoch": 4.120979356697071, + "grad_norm": 1.2195490538076916, + "learning_rate": 9.143527508037243e-07, + "loss": 0.408, + "step": 4294 + }, + { + "epoch": 4.121939510321652, + "grad_norm": 0.9151466119180599, + "learning_rate": 9.124225692853995e-07, + "loss": 0.4738, + "step": 4295 + }, + { + "epoch": 4.122899663946232, + "grad_norm": 0.970946276289556, + "learning_rate": 9.104942226368879e-07, + "loss": 0.3946, + "step": 4296 + }, + { + "epoch": 4.123859817570811, + "grad_norm": 1.1728049596209218, + "learning_rate": 9.085677117238068e-07, + "loss": 0.3516, + "step": 4297 + }, + { + "epoch": 4.124819971195391, + "grad_norm": 0.9065424553789335, + "learning_rate": 9.066430374109419e-07, + "loss": 0.4066, + "step": 4298 + }, + { + "epoch": 4.125780124819971, + "grad_norm": 1.1154419366105508, + "learning_rate": 9.047202005622641e-07, + "loss": 0.4886, + "step": 4299 + }, + { + "epoch": 4.126740278444551, + "grad_norm": 0.908548924060711, + "learning_rate": 9.027992020409127e-07, + "loss": 0.5112, + "step": 4300 + }, + { + "epoch": 4.1277004320691315, + "grad_norm": 1.1014847745552856, + "learning_rate": 9.008800427092018e-07, + "loss": 0.4825, + "step": 4301 + }, + { + "epoch": 4.128660585693711, + "grad_norm": 1.0762917609699958, + "learning_rate": 8.989627234286225e-07, + "loss": 0.4954, + "step": 4302 + }, + { + "epoch": 4.129620739318291, + "grad_norm": 0.9915935095231558, + "learning_rate": 8.970472450598394e-07, + "loss": 0.4679, + "step": 4303 + }, + { + "epoch": 4.130580892942871, + "grad_norm": 1.2361143931951928, + "learning_rate": 8.951336084626894e-07, + "loss": 0.4544, + "step": 4304 + }, + { + "epoch": 4.131541046567451, + "grad_norm": 0.9749138566149839, + "learning_rate": 8.932218144961858e-07, + "loss": 0.407, + "step": 4305 + }, + { + "epoch": 4.1325012001920305, + "grad_norm": 1.1217454966773088, + "learning_rate": 8.913118640185086e-07, + "loss": 0.4004, + "step": 4306 + }, + { + "epoch": 4.13346135381661, + "grad_norm": 1.1038976678078036, + "learning_rate": 8.894037578870152e-07, + "loss": 0.3665, + "step": 4307 + }, + { + "epoch": 4.134421507441191, + "grad_norm": 0.9700248788900517, + "learning_rate": 8.874974969582379e-07, + "loss": 0.4415, + "step": 4308 + }, + { + "epoch": 4.135381661065771, + "grad_norm": 1.4278006870827293, + "learning_rate": 8.855930820878727e-07, + "loss": 0.4433, + "step": 4309 + }, + { + "epoch": 4.136341814690351, + "grad_norm": 1.1091730316071229, + "learning_rate": 8.836905141307933e-07, + "loss": 0.4636, + "step": 4310 + }, + { + "epoch": 4.13730196831493, + "grad_norm": 1.1142769292697934, + "learning_rate": 8.817897939410391e-07, + "loss": 0.4611, + "step": 4311 + }, + { + "epoch": 4.13826212193951, + "grad_norm": 0.9639399232985066, + "learning_rate": 8.79890922371826e-07, + "loss": 0.4674, + "step": 4312 + }, + { + "epoch": 4.13922227556409, + "grad_norm": 0.9840393238800248, + "learning_rate": 8.779939002755378e-07, + "loss": 0.4239, + "step": 4313 + }, + { + "epoch": 4.14018242918867, + "grad_norm": 1.3571010902250684, + "learning_rate": 8.760987285037248e-07, + "loss": 0.4589, + "step": 4314 + }, + { + "epoch": 4.14114258281325, + "grad_norm": 0.9979987934238568, + "learning_rate": 8.742054079071105e-07, + "loss": 0.4836, + "step": 4315 + }, + { + "epoch": 4.14210273643783, + "grad_norm": 1.104191561242044, + "learning_rate": 8.723139393355861e-07, + "loss": 0.3678, + "step": 4316 + }, + { + "epoch": 4.14306289006241, + "grad_norm": 0.7732049241607881, + "learning_rate": 8.704243236382115e-07, + "loss": 0.5113, + "step": 4317 + }, + { + "epoch": 4.14402304368699, + "grad_norm": 1.0667449949896786, + "learning_rate": 8.685365616632175e-07, + "loss": 0.4554, + "step": 4318 + }, + { + "epoch": 4.14498319731157, + "grad_norm": 1.1287351593004549, + "learning_rate": 8.666506542579961e-07, + "loss": 0.5003, + "step": 4319 + }, + { + "epoch": 4.145943350936149, + "grad_norm": 1.3527933770086522, + "learning_rate": 8.64766602269112e-07, + "loss": 0.3664, + "step": 4320 + }, + { + "epoch": 4.14690350456073, + "grad_norm": 0.9243479456992848, + "learning_rate": 8.628844065422992e-07, + "loss": 0.402, + "step": 4321 + }, + { + "epoch": 4.14786365818531, + "grad_norm": 1.0853048962497471, + "learning_rate": 8.610040679224518e-07, + "loss": 0.3226, + "step": 4322 + }, + { + "epoch": 4.14882381180989, + "grad_norm": 1.8494142670327685, + "learning_rate": 8.591255872536364e-07, + "loss": 0.4033, + "step": 4323 + }, + { + "epoch": 4.1497839654344695, + "grad_norm": 1.1547235391359174, + "learning_rate": 8.572489653790789e-07, + "loss": 0.4074, + "step": 4324 + }, + { + "epoch": 4.150744119059049, + "grad_norm": 0.8525740470029572, + "learning_rate": 8.553742031411783e-07, + "loss": 0.4835, + "step": 4325 + }, + { + "epoch": 4.151704272683629, + "grad_norm": 1.0738258050384755, + "learning_rate": 8.535013013814952e-07, + "loss": 0.401, + "step": 4326 + }, + { + "epoch": 4.15266442630821, + "grad_norm": 1.0421905543463854, + "learning_rate": 8.51630260940754e-07, + "loss": 0.4901, + "step": 4327 + }, + { + "epoch": 4.1536245799327896, + "grad_norm": 1.9839771489313465, + "learning_rate": 8.497610826588443e-07, + "loss": 0.5864, + "step": 4328 + }, + { + "epoch": 4.154584733557369, + "grad_norm": 0.8454376940347059, + "learning_rate": 8.478937673748211e-07, + "loss": 0.437, + "step": 4329 + }, + { + "epoch": 4.155544887181949, + "grad_norm": 0.9683212298990983, + "learning_rate": 8.460283159269023e-07, + "loss": 0.3862, + "step": 4330 + }, + { + "epoch": 4.156505040806529, + "grad_norm": 1.0770426460280165, + "learning_rate": 8.441647291524701e-07, + "loss": 0.4002, + "step": 4331 + }, + { + "epoch": 4.157465194431109, + "grad_norm": 0.8926379888513267, + "learning_rate": 8.423030078880667e-07, + "loss": 0.4265, + "step": 4332 + }, + { + "epoch": 4.1584253480556885, + "grad_norm": 0.8766685012161926, + "learning_rate": 8.404431529693996e-07, + "loss": 0.4383, + "step": 4333 + }, + { + "epoch": 4.159385501680269, + "grad_norm": 0.9426447039942522, + "learning_rate": 8.385851652313376e-07, + "loss": 0.3939, + "step": 4334 + }, + { + "epoch": 4.160345655304849, + "grad_norm": 1.2529311935400318, + "learning_rate": 8.367290455079125e-07, + "loss": 0.3699, + "step": 4335 + }, + { + "epoch": 4.161305808929429, + "grad_norm": 0.9787205617363444, + "learning_rate": 8.348747946323165e-07, + "loss": 0.3142, + "step": 4336 + }, + { + "epoch": 4.162265962554009, + "grad_norm": 1.3210305280646293, + "learning_rate": 8.330224134369014e-07, + "loss": 0.4638, + "step": 4337 + }, + { + "epoch": 4.163226116178588, + "grad_norm": 0.9801531970254048, + "learning_rate": 8.311719027531817e-07, + "loss": 0.4524, + "step": 4338 + }, + { + "epoch": 4.164186269803168, + "grad_norm": 1.031215780297556, + "learning_rate": 8.293232634118342e-07, + "loss": 0.3935, + "step": 4339 + }, + { + "epoch": 4.165146423427748, + "grad_norm": 1.1366654773442557, + "learning_rate": 8.274764962426912e-07, + "loss": 0.4191, + "step": 4340 + }, + { + "epoch": 4.166106577052329, + "grad_norm": 0.7775203896986472, + "learning_rate": 8.256316020747479e-07, + "loss": 0.4779, + "step": 4341 + }, + { + "epoch": 4.1670667306769085, + "grad_norm": 0.8980317212159564, + "learning_rate": 8.237885817361552e-07, + "loss": 0.4229, + "step": 4342 + }, + { + "epoch": 4.168026884301488, + "grad_norm": 0.9484593723923103, + "learning_rate": 8.219474360542273e-07, + "loss": 0.4587, + "step": 4343 + }, + { + "epoch": 4.168987037926068, + "grad_norm": 1.1054808758696812, + "learning_rate": 8.201081658554361e-07, + "loss": 0.3986, + "step": 4344 + }, + { + "epoch": 4.169947191550648, + "grad_norm": 1.527882440838069, + "learning_rate": 8.182707719654076e-07, + "loss": 0.347, + "step": 4345 + }, + { + "epoch": 4.170907345175228, + "grad_norm": 1.3505240312263833, + "learning_rate": 8.164352552089288e-07, + "loss": 0.4489, + "step": 4346 + }, + { + "epoch": 4.171867498799808, + "grad_norm": 1.6916111421535276, + "learning_rate": 8.146016164099446e-07, + "loss": 0.3571, + "step": 4347 + }, + { + "epoch": 4.172827652424388, + "grad_norm": 1.0646221367630506, + "learning_rate": 8.127698563915554e-07, + "loss": 0.5274, + "step": 4348 + }, + { + "epoch": 4.173787806048968, + "grad_norm": 1.2484292715766703, + "learning_rate": 8.109399759760194e-07, + "loss": 0.444, + "step": 4349 + }, + { + "epoch": 4.174747959673548, + "grad_norm": 1.1668223099744186, + "learning_rate": 8.091119759847493e-07, + "loss": 0.3771, + "step": 4350 + }, + { + "epoch": 4.1757081132981275, + "grad_norm": 0.9682240709956186, + "learning_rate": 8.07285857238313e-07, + "loss": 0.4213, + "step": 4351 + }, + { + "epoch": 4.176668266922707, + "grad_norm": 1.2349362734220277, + "learning_rate": 8.054616205564403e-07, + "loss": 0.4149, + "step": 4352 + }, + { + "epoch": 4.177628420547288, + "grad_norm": 1.4933877123150152, + "learning_rate": 8.036392667580084e-07, + "loss": 0.4818, + "step": 4353 + }, + { + "epoch": 4.178588574171868, + "grad_norm": 0.9192447638962328, + "learning_rate": 8.018187966610535e-07, + "loss": 0.4631, + "step": 4354 + }, + { + "epoch": 4.179548727796448, + "grad_norm": 0.980379117822449, + "learning_rate": 8.000002110827626e-07, + "loss": 0.4105, + "step": 4355 + }, + { + "epoch": 4.180508881421027, + "grad_norm": 1.319253235429286, + "learning_rate": 7.981835108394825e-07, + "loss": 0.4223, + "step": 4356 + }, + { + "epoch": 4.181469035045607, + "grad_norm": 0.9872882315648225, + "learning_rate": 7.963686967467111e-07, + "loss": 0.4224, + "step": 4357 + }, + { + "epoch": 4.182429188670187, + "grad_norm": 2.1465017034470275, + "learning_rate": 7.945557696190959e-07, + "loss": 0.3166, + "step": 4358 + }, + { + "epoch": 4.183389342294767, + "grad_norm": 0.8976886876840247, + "learning_rate": 7.927447302704422e-07, + "loss": 0.5547, + "step": 4359 + }, + { + "epoch": 4.1843494959193475, + "grad_norm": 1.2758054935655816, + "learning_rate": 7.909355795137058e-07, + "loss": 0.3662, + "step": 4360 + }, + { + "epoch": 4.185309649543927, + "grad_norm": 0.9514119430123684, + "learning_rate": 7.891283181609955e-07, + "loss": 0.3741, + "step": 4361 + }, + { + "epoch": 4.186269803168507, + "grad_norm": 0.9818243875674629, + "learning_rate": 7.873229470235722e-07, + "loss": 0.5299, + "step": 4362 + }, + { + "epoch": 4.187229956793087, + "grad_norm": 1.0348778856103753, + "learning_rate": 7.855194669118466e-07, + "loss": 0.3903, + "step": 4363 + }, + { + "epoch": 4.188190110417667, + "grad_norm": 0.9601524852459529, + "learning_rate": 7.837178786353811e-07, + "loss": 0.4132, + "step": 4364 + }, + { + "epoch": 4.189150264042246, + "grad_norm": 1.1093554204911817, + "learning_rate": 7.81918183002891e-07, + "loss": 0.4645, + "step": 4365 + }, + { + "epoch": 4.190110417666826, + "grad_norm": 1.2027994506192603, + "learning_rate": 7.801203808222396e-07, + "loss": 0.4558, + "step": 4366 + }, + { + "epoch": 4.191070571291407, + "grad_norm": 0.9758616955709369, + "learning_rate": 7.783244729004425e-07, + "loss": 0.3789, + "step": 4367 + }, + { + "epoch": 4.192030724915987, + "grad_norm": 1.067144813873389, + "learning_rate": 7.765304600436596e-07, + "loss": 0.4328, + "step": 4368 + }, + { + "epoch": 4.1929908785405665, + "grad_norm": 1.329729032766879, + "learning_rate": 7.747383430572081e-07, + "loss": 0.3728, + "step": 4369 + }, + { + "epoch": 4.193951032165146, + "grad_norm": 1.327894271246118, + "learning_rate": 7.729481227455498e-07, + "loss": 0.3563, + "step": 4370 + }, + { + "epoch": 4.194911185789726, + "grad_norm": 0.9759885002648725, + "learning_rate": 7.711597999122939e-07, + "loss": 0.4585, + "step": 4371 + }, + { + "epoch": 4.195871339414306, + "grad_norm": 1.0770740023102026, + "learning_rate": 7.693733753601995e-07, + "loss": 0.4242, + "step": 4372 + }, + { + "epoch": 4.196831493038887, + "grad_norm": 1.0916757540878639, + "learning_rate": 7.675888498911743e-07, + "loss": 0.3555, + "step": 4373 + }, + { + "epoch": 4.197791646663466, + "grad_norm": 0.9751146380302486, + "learning_rate": 7.658062243062725e-07, + "loss": 0.315, + "step": 4374 + }, + { + "epoch": 4.198751800288046, + "grad_norm": 0.7201333438208256, + "learning_rate": 7.640254994056967e-07, + "loss": 0.4189, + "step": 4375 + }, + { + "epoch": 4.199711953912626, + "grad_norm": 1.0943495629175872, + "learning_rate": 7.622466759887931e-07, + "loss": 0.3943, + "step": 4376 + }, + { + "epoch": 4.200672107537206, + "grad_norm": 1.097804637006829, + "learning_rate": 7.604697548540574e-07, + "loss": 0.4712, + "step": 4377 + }, + { + "epoch": 4.201632261161786, + "grad_norm": 1.134117033885051, + "learning_rate": 7.58694736799131e-07, + "loss": 0.4902, + "step": 4378 + }, + { + "epoch": 4.202592414786366, + "grad_norm": 1.2867175110808418, + "learning_rate": 7.569216226207999e-07, + "loss": 0.4477, + "step": 4379 + }, + { + "epoch": 4.203552568410946, + "grad_norm": 0.9635986344536933, + "learning_rate": 7.551504131149972e-07, + "loss": 0.4581, + "step": 4380 + }, + { + "epoch": 4.204512722035526, + "grad_norm": 1.2549734489048616, + "learning_rate": 7.533811090767984e-07, + "loss": 0.4083, + "step": 4381 + }, + { + "epoch": 4.205472875660106, + "grad_norm": 1.0346341836574662, + "learning_rate": 7.516137113004246e-07, + "loss": 0.4214, + "step": 4382 + }, + { + "epoch": 4.206433029284685, + "grad_norm": 1.1758488932610325, + "learning_rate": 7.498482205792457e-07, + "loss": 0.4676, + "step": 4383 + }, + { + "epoch": 4.207393182909265, + "grad_norm": 0.9125108956172018, + "learning_rate": 7.480846377057677e-07, + "loss": 0.4673, + "step": 4384 + }, + { + "epoch": 4.208353336533845, + "grad_norm": 1.4564810659659302, + "learning_rate": 7.463229634716462e-07, + "loss": 0.3907, + "step": 4385 + }, + { + "epoch": 4.209313490158426, + "grad_norm": 1.5349426567758955, + "learning_rate": 7.445631986676755e-07, + "loss": 0.4176, + "step": 4386 + }, + { + "epoch": 4.2102736437830055, + "grad_norm": 0.8116893514756358, + "learning_rate": 7.428053440837973e-07, + "loss": 0.3901, + "step": 4387 + }, + { + "epoch": 4.211233797407585, + "grad_norm": 1.0618448948486043, + "learning_rate": 7.410494005090946e-07, + "loss": 0.3491, + "step": 4388 + }, + { + "epoch": 4.212193951032165, + "grad_norm": 1.0709972050922476, + "learning_rate": 7.392953687317894e-07, + "loss": 0.4423, + "step": 4389 + }, + { + "epoch": 4.213154104656745, + "grad_norm": 1.375030001306386, + "learning_rate": 7.375432495392487e-07, + "loss": 0.3047, + "step": 4390 + }, + { + "epoch": 4.214114258281325, + "grad_norm": 0.952324512104103, + "learning_rate": 7.3579304371798e-07, + "loss": 0.463, + "step": 4391 + }, + { + "epoch": 4.215074411905905, + "grad_norm": 1.2558651253150959, + "learning_rate": 7.34044752053632e-07, + "loss": 0.3193, + "step": 4392 + }, + { + "epoch": 4.216034565530485, + "grad_norm": 0.8595265997554533, + "learning_rate": 7.322983753309959e-07, + "loss": 0.4808, + "step": 4393 + }, + { + "epoch": 4.216994719155065, + "grad_norm": 1.1045760086354, + "learning_rate": 7.305539143339985e-07, + "loss": 0.4097, + "step": 4394 + }, + { + "epoch": 4.217954872779645, + "grad_norm": 0.7915374479752565, + "learning_rate": 7.2881136984571e-07, + "loss": 0.4936, + "step": 4395 + }, + { + "epoch": 4.2189150264042246, + "grad_norm": 1.0908526339530273, + "learning_rate": 7.270707426483431e-07, + "loss": 0.6003, + "step": 4396 + }, + { + "epoch": 4.219875180028804, + "grad_norm": 0.9341987903260882, + "learning_rate": 7.253320335232439e-07, + "loss": 0.3379, + "step": 4397 + }, + { + "epoch": 4.220835333653384, + "grad_norm": 1.5674855072164864, + "learning_rate": 7.235952432509024e-07, + "loss": 0.4439, + "step": 4398 + }, + { + "epoch": 4.221795487277965, + "grad_norm": 0.9120035518167652, + "learning_rate": 7.218603726109413e-07, + "loss": 0.4495, + "step": 4399 + }, + { + "epoch": 4.222755640902545, + "grad_norm": 0.8944366659130141, + "learning_rate": 7.201274223821298e-07, + "loss": 0.4331, + "step": 4400 + }, + { + "epoch": 4.223715794527124, + "grad_norm": 0.9514219119648922, + "learning_rate": 7.183963933423699e-07, + "loss": 0.4905, + "step": 4401 + }, + { + "epoch": 4.224675948151704, + "grad_norm": 1.3943489187171438, + "learning_rate": 7.166672862687002e-07, + "loss": 0.3928, + "step": 4402 + }, + { + "epoch": 4.225636101776284, + "grad_norm": 1.214565473547559, + "learning_rate": 7.149401019372993e-07, + "loss": 0.3932, + "step": 4403 + }, + { + "epoch": 4.226596255400864, + "grad_norm": 1.5195610739463175, + "learning_rate": 7.132148411234818e-07, + "loss": 0.3896, + "step": 4404 + }, + { + "epoch": 4.2275564090254445, + "grad_norm": 1.0302111293927776, + "learning_rate": 7.11491504601699e-07, + "loss": 0.5536, + "step": 4405 + }, + { + "epoch": 4.228516562650024, + "grad_norm": 1.0380122600665802, + "learning_rate": 7.097700931455397e-07, + "loss": 0.4166, + "step": 4406 + }, + { + "epoch": 4.229476716274604, + "grad_norm": 1.0912850823669555, + "learning_rate": 7.080506075277243e-07, + "loss": 0.3937, + "step": 4407 + }, + { + "epoch": 4.230436869899184, + "grad_norm": 1.1944135216477008, + "learning_rate": 7.063330485201126e-07, + "loss": 0.488, + "step": 4408 + }, + { + "epoch": 4.231397023523764, + "grad_norm": 0.8459026612189624, + "learning_rate": 7.046174168936987e-07, + "loss": 0.3833, + "step": 4409 + }, + { + "epoch": 4.2323571771483435, + "grad_norm": 1.0128430552676349, + "learning_rate": 7.029037134186112e-07, + "loss": 0.5029, + "step": 4410 + }, + { + "epoch": 4.233317330772923, + "grad_norm": 1.3581601119845752, + "learning_rate": 7.011919388641148e-07, + "loss": 0.4926, + "step": 4411 + }, + { + "epoch": 4.234277484397504, + "grad_norm": 1.089810644774874, + "learning_rate": 6.994820939986025e-07, + "loss": 0.3769, + "step": 4412 + }, + { + "epoch": 4.235237638022084, + "grad_norm": 1.0701855565316365, + "learning_rate": 6.977741795896098e-07, + "loss": 0.5343, + "step": 4413 + }, + { + "epoch": 4.2361977916466635, + "grad_norm": 0.9992298903575676, + "learning_rate": 6.960681964038008e-07, + "loss": 0.4292, + "step": 4414 + }, + { + "epoch": 4.237157945271243, + "grad_norm": 1.1421889722876242, + "learning_rate": 6.943641452069705e-07, + "loss": 0.5045, + "step": 4415 + }, + { + "epoch": 4.238118098895823, + "grad_norm": 1.1337211290098332, + "learning_rate": 6.92662026764051e-07, + "loss": 0.4609, + "step": 4416 + }, + { + "epoch": 4.239078252520403, + "grad_norm": 1.3972395737905647, + "learning_rate": 6.909618418391045e-07, + "loss": 0.4463, + "step": 4417 + }, + { + "epoch": 4.240038406144984, + "grad_norm": 1.1080615797570958, + "learning_rate": 6.892635911953261e-07, + "loss": 0.5192, + "step": 4418 + }, + { + "epoch": 4.240998559769563, + "grad_norm": 1.3289900672497104, + "learning_rate": 6.875672755950441e-07, + "loss": 0.4912, + "step": 4419 + }, + { + "epoch": 4.241958713394143, + "grad_norm": 1.0115555082204035, + "learning_rate": 6.858728957997129e-07, + "loss": 0.4726, + "step": 4420 + }, + { + "epoch": 4.242918867018723, + "grad_norm": 1.3820913815000928, + "learning_rate": 6.841804525699236e-07, + "loss": 0.3195, + "step": 4421 + }, + { + "epoch": 4.243879020643303, + "grad_norm": 0.9404980902608489, + "learning_rate": 6.824899466653961e-07, + "loss": 0.4956, + "step": 4422 + }, + { + "epoch": 4.244839174267883, + "grad_norm": 1.0839776582644958, + "learning_rate": 6.808013788449791e-07, + "loss": 0.5658, + "step": 4423 + }, + { + "epoch": 4.245799327892462, + "grad_norm": 1.1769699753328842, + "learning_rate": 6.79114749866655e-07, + "loss": 0.451, + "step": 4424 + }, + { + "epoch": 4.246759481517043, + "grad_norm": 1.0114614360470604, + "learning_rate": 6.774300604875312e-07, + "loss": 0.5437, + "step": 4425 + }, + { + "epoch": 4.247719635141623, + "grad_norm": 1.1704788199467249, + "learning_rate": 6.757473114638458e-07, + "loss": 0.394, + "step": 4426 + }, + { + "epoch": 4.248679788766203, + "grad_norm": 1.015827273715958, + "learning_rate": 6.740665035509714e-07, + "loss": 0.4339, + "step": 4427 + }, + { + "epoch": 4.2496399423907825, + "grad_norm": 1.5050906563570037, + "learning_rate": 6.723876375034e-07, + "loss": 0.3998, + "step": 4428 + }, + { + "epoch": 4.250600096015362, + "grad_norm": 0.9893318347609339, + "learning_rate": 6.707107140747604e-07, + "loss": 0.477, + "step": 4429 + }, + { + "epoch": 4.251560249639942, + "grad_norm": 1.175298329064842, + "learning_rate": 6.690357340178011e-07, + "loss": 0.4447, + "step": 4430 + }, + { + "epoch": 4.252520403264523, + "grad_norm": 1.0576490355024406, + "learning_rate": 6.673626980844067e-07, + "loss": 0.4476, + "step": 4431 + }, + { + "epoch": 4.2534805568891025, + "grad_norm": 1.208711748055826, + "learning_rate": 6.656916070255842e-07, + "loss": 0.4743, + "step": 4432 + }, + { + "epoch": 4.254440710513682, + "grad_norm": 1.0782587346243588, + "learning_rate": 6.640224615914681e-07, + "loss": 0.4515, + "step": 4433 + }, + { + "epoch": 4.255400864138262, + "grad_norm": 1.0909404011006474, + "learning_rate": 6.62355262531319e-07, + "loss": 0.4656, + "step": 4434 + }, + { + "epoch": 4.256361017762842, + "grad_norm": 1.2066058971451301, + "learning_rate": 6.606900105935265e-07, + "loss": 0.4056, + "step": 4435 + }, + { + "epoch": 4.257321171387422, + "grad_norm": 1.365534030628649, + "learning_rate": 6.590267065256029e-07, + "loss": 0.3418, + "step": 4436 + }, + { + "epoch": 4.2582813250120015, + "grad_norm": 0.9844552332348545, + "learning_rate": 6.573653510741901e-07, + "loss": 0.4561, + "step": 4437 + }, + { + "epoch": 4.259241478636582, + "grad_norm": 1.0442577595603217, + "learning_rate": 6.557059449850494e-07, + "loss": 0.4649, + "step": 4438 + }, + { + "epoch": 4.260201632261162, + "grad_norm": 1.5771119655670944, + "learning_rate": 6.540484890030724e-07, + "loss": 0.3592, + "step": 4439 + }, + { + "epoch": 4.261161785885742, + "grad_norm": 1.1659416282352184, + "learning_rate": 6.523929838722726e-07, + "loss": 0.3645, + "step": 4440 + }, + { + "epoch": 4.262121939510322, + "grad_norm": 1.078049494468062, + "learning_rate": 6.507394303357894e-07, + "loss": 0.4677, + "step": 4441 + }, + { + "epoch": 4.263082093134901, + "grad_norm": 1.0508118717897696, + "learning_rate": 6.49087829135886e-07, + "loss": 0.4356, + "step": 4442 + }, + { + "epoch": 4.264042246759481, + "grad_norm": 1.4498334722820503, + "learning_rate": 6.474381810139446e-07, + "loss": 0.4595, + "step": 4443 + }, + { + "epoch": 4.265002400384062, + "grad_norm": 1.1094692463594185, + "learning_rate": 6.457904867104792e-07, + "loss": 0.4981, + "step": 4444 + }, + { + "epoch": 4.265962554008642, + "grad_norm": 0.9098243322565156, + "learning_rate": 6.441447469651213e-07, + "loss": 0.5511, + "step": 4445 + }, + { + "epoch": 4.2669227076332215, + "grad_norm": 0.9410277880142008, + "learning_rate": 6.42500962516624e-07, + "loss": 0.3342, + "step": 4446 + }, + { + "epoch": 4.267882861257801, + "grad_norm": 1.0298469402066288, + "learning_rate": 6.408591341028652e-07, + "loss": 0.3979, + "step": 4447 + }, + { + "epoch": 4.268843014882381, + "grad_norm": 1.035142320696121, + "learning_rate": 6.392192624608451e-07, + "loss": 0.5436, + "step": 4448 + }, + { + "epoch": 4.269803168506961, + "grad_norm": 1.0262178512812177, + "learning_rate": 6.375813483266835e-07, + "loss": 0.4625, + "step": 4449 + }, + { + "epoch": 4.270763322131541, + "grad_norm": 1.2681507987219254, + "learning_rate": 6.35945392435624e-07, + "loss": 0.3438, + "step": 4450 + }, + { + "epoch": 4.271723475756121, + "grad_norm": 1.0898167700551433, + "learning_rate": 6.343113955220276e-07, + "loss": 0.4123, + "step": 4451 + }, + { + "epoch": 4.272683629380701, + "grad_norm": 1.0630260987417366, + "learning_rate": 6.326793583193797e-07, + "loss": 0.4529, + "step": 4452 + }, + { + "epoch": 4.273643783005281, + "grad_norm": 1.0182487962488627, + "learning_rate": 6.310492815602831e-07, + "loss": 0.5655, + "step": 4453 + }, + { + "epoch": 4.274603936629861, + "grad_norm": 1.2327303090634718, + "learning_rate": 6.294211659764626e-07, + "loss": 0.3857, + "step": 4454 + }, + { + "epoch": 4.2755640902544405, + "grad_norm": 1.009455204999046, + "learning_rate": 6.277950122987631e-07, + "loss": 0.467, + "step": 4455 + }, + { + "epoch": 4.27652424387902, + "grad_norm": 1.3042464656689299, + "learning_rate": 6.261708212571438e-07, + "loss": 0.3489, + "step": 4456 + }, + { + "epoch": 4.277484397503601, + "grad_norm": 1.1822483571429399, + "learning_rate": 6.245485935806905e-07, + "loss": 0.3783, + "step": 4457 + }, + { + "epoch": 4.278444551128181, + "grad_norm": 0.9538121742017903, + "learning_rate": 6.229283299976041e-07, + "loss": 0.5196, + "step": 4458 + }, + { + "epoch": 4.279404704752761, + "grad_norm": 1.3927130292859722, + "learning_rate": 6.213100312352005e-07, + "loss": 0.4262, + "step": 4459 + }, + { + "epoch": 4.28036485837734, + "grad_norm": 1.3292202438041856, + "learning_rate": 6.196936980199181e-07, + "loss": 0.4451, + "step": 4460 + }, + { + "epoch": 4.28132501200192, + "grad_norm": 1.1446926937061679, + "learning_rate": 6.18079331077312e-07, + "loss": 0.4448, + "step": 4461 + }, + { + "epoch": 4.2822851656265, + "grad_norm": 0.774846484331113, + "learning_rate": 6.164669311320543e-07, + "loss": 0.4745, + "step": 4462 + }, + { + "epoch": 4.28324531925108, + "grad_norm": 1.4484757261162085, + "learning_rate": 6.148564989079347e-07, + "loss": 0.3511, + "step": 4463 + }, + { + "epoch": 4.2842054728756604, + "grad_norm": 0.9140691682864316, + "learning_rate": 6.13248035127857e-07, + "loss": 0.4921, + "step": 4464 + }, + { + "epoch": 4.28516562650024, + "grad_norm": 0.9016712253267545, + "learning_rate": 6.116415405138443e-07, + "loss": 0.4195, + "step": 4465 + }, + { + "epoch": 4.28612578012482, + "grad_norm": 1.035987517546343, + "learning_rate": 6.100370157870356e-07, + "loss": 0.4645, + "step": 4466 + }, + { + "epoch": 4.2870859337494, + "grad_norm": 1.1887084816991396, + "learning_rate": 6.08434461667684e-07, + "loss": 0.5482, + "step": 4467 + }, + { + "epoch": 4.28804608737398, + "grad_norm": 1.1201728937136024, + "learning_rate": 6.06833878875161e-07, + "loss": 0.405, + "step": 4468 + }, + { + "epoch": 4.289006240998559, + "grad_norm": 1.0066174407345565, + "learning_rate": 6.052352681279478e-07, + "loss": 0.4164, + "step": 4469 + }, + { + "epoch": 4.28996639462314, + "grad_norm": 1.0481544696809473, + "learning_rate": 6.036386301436448e-07, + "loss": 0.419, + "step": 4470 + }, + { + "epoch": 4.29092654824772, + "grad_norm": 1.0618489948336933, + "learning_rate": 6.020439656389676e-07, + "loss": 0.5291, + "step": 4471 + }, + { + "epoch": 4.2918867018723, + "grad_norm": 0.9405865659105404, + "learning_rate": 6.004512753297421e-07, + "loss": 0.4382, + "step": 4472 + }, + { + "epoch": 4.2928468554968795, + "grad_norm": 1.1299975350075617, + "learning_rate": 5.988605599309116e-07, + "loss": 0.4626, + "step": 4473 + }, + { + "epoch": 4.293807009121459, + "grad_norm": 1.0377816138637095, + "learning_rate": 5.97271820156528e-07, + "loss": 0.4099, + "step": 4474 + }, + { + "epoch": 4.294767162746039, + "grad_norm": 1.2302999158890702, + "learning_rate": 5.95685056719763e-07, + "loss": 0.3626, + "step": 4475 + }, + { + "epoch": 4.295727316370619, + "grad_norm": 0.9961375213603376, + "learning_rate": 5.941002703328974e-07, + "loss": 0.3926, + "step": 4476 + }, + { + "epoch": 4.2966874699952, + "grad_norm": 2.2479451107992428, + "learning_rate": 5.925174617073226e-07, + "loss": 0.434, + "step": 4477 + }, + { + "epoch": 4.297647623619779, + "grad_norm": 1.4175075649198223, + "learning_rate": 5.909366315535464e-07, + "loss": 0.4914, + "step": 4478 + }, + { + "epoch": 4.298607777244359, + "grad_norm": 0.8534080948985644, + "learning_rate": 5.893577805811856e-07, + "loss": 0.4018, + "step": 4479 + }, + { + "epoch": 4.299567930868939, + "grad_norm": 0.6793077800812086, + "learning_rate": 5.877809094989706e-07, + "loss": 0.5068, + "step": 4480 + }, + { + "epoch": 4.300528084493519, + "grad_norm": 1.1049764695956181, + "learning_rate": 5.862060190147417e-07, + "loss": 0.4824, + "step": 4481 + }, + { + "epoch": 4.3014882381180986, + "grad_norm": 0.9954385533117368, + "learning_rate": 5.84633109835449e-07, + "loss": 0.4763, + "step": 4482 + }, + { + "epoch": 4.302448391742679, + "grad_norm": 0.9487563599927328, + "learning_rate": 5.830621826671562e-07, + "loss": 0.4327, + "step": 4483 + }, + { + "epoch": 4.303408545367259, + "grad_norm": 1.5416099471982594, + "learning_rate": 5.814932382150346e-07, + "loss": 0.3276, + "step": 4484 + }, + { + "epoch": 4.304368698991839, + "grad_norm": 0.907646344471589, + "learning_rate": 5.799262771833675e-07, + "loss": 0.4563, + "step": 4485 + }, + { + "epoch": 4.305328852616419, + "grad_norm": 0.7292641239404976, + "learning_rate": 5.783613002755478e-07, + "loss": 0.4809, + "step": 4486 + }, + { + "epoch": 4.306289006240998, + "grad_norm": 1.4488767928829671, + "learning_rate": 5.767983081940731e-07, + "loss": 0.5004, + "step": 4487 + }, + { + "epoch": 4.307249159865578, + "grad_norm": 0.8757980682368414, + "learning_rate": 5.752373016405577e-07, + "loss": 0.4347, + "step": 4488 + }, + { + "epoch": 4.308209313490158, + "grad_norm": 1.1784110948522728, + "learning_rate": 5.736782813157204e-07, + "loss": 0.3357, + "step": 4489 + }, + { + "epoch": 4.309169467114739, + "grad_norm": 1.2082889439231683, + "learning_rate": 5.721212479193871e-07, + "loss": 0.47, + "step": 4490 + }, + { + "epoch": 4.3101296207393185, + "grad_norm": 0.9932145983115908, + "learning_rate": 5.705662021504943e-07, + "loss": 0.4699, + "step": 4491 + }, + { + "epoch": 4.311089774363898, + "grad_norm": 0.9426793040185881, + "learning_rate": 5.690131447070851e-07, + "loss": 0.4547, + "step": 4492 + }, + { + "epoch": 4.312049927988478, + "grad_norm": 0.7503710689496849, + "learning_rate": 5.674620762863109e-07, + "loss": 0.3817, + "step": 4493 + }, + { + "epoch": 4.313010081613058, + "grad_norm": 1.3958488505849693, + "learning_rate": 5.659129975844302e-07, + "loss": 0.3035, + "step": 4494 + }, + { + "epoch": 4.313970235237638, + "grad_norm": 1.2066119515708529, + "learning_rate": 5.643659092968062e-07, + "loss": 0.5303, + "step": 4495 + }, + { + "epoch": 4.314930388862218, + "grad_norm": 1.020308109195774, + "learning_rate": 5.628208121179107e-07, + "loss": 0.3165, + "step": 4496 + }, + { + "epoch": 4.315890542486798, + "grad_norm": 1.1745863963208425, + "learning_rate": 5.612777067413228e-07, + "loss": 0.4169, + "step": 4497 + }, + { + "epoch": 4.316850696111378, + "grad_norm": 1.1150202642573368, + "learning_rate": 5.597365938597243e-07, + "loss": 0.395, + "step": 4498 + }, + { + "epoch": 4.317810849735958, + "grad_norm": 1.1696750726671203, + "learning_rate": 5.581974741649066e-07, + "loss": 0.4384, + "step": 4499 + }, + { + "epoch": 4.3187710033605375, + "grad_norm": 1.3783733766796513, + "learning_rate": 5.566603483477607e-07, + "loss": 0.3204, + "step": 4500 + }, + { + "epoch": 4.319731156985117, + "grad_norm": 0.9876079093983445, + "learning_rate": 5.551252170982874e-07, + "loss": 0.4937, + "step": 4501 + }, + { + "epoch": 4.320691310609698, + "grad_norm": 0.9296670988101725, + "learning_rate": 5.535920811055928e-07, + "loss": 0.4967, + "step": 4502 + }, + { + "epoch": 4.321651464234278, + "grad_norm": 1.0871976192602064, + "learning_rate": 5.520609410578826e-07, + "loss": 0.5041, + "step": 4503 + }, + { + "epoch": 4.322611617858858, + "grad_norm": 0.9435435305415898, + "learning_rate": 5.505317976424712e-07, + "loss": 0.4403, + "step": 4504 + }, + { + "epoch": 4.323571771483437, + "grad_norm": 1.0117052877528614, + "learning_rate": 5.49004651545772e-07, + "loss": 0.4853, + "step": 4505 + }, + { + "epoch": 4.324531925108017, + "grad_norm": 1.2471165167903087, + "learning_rate": 5.474795034533064e-07, + "loss": 0.4208, + "step": 4506 + }, + { + "epoch": 4.325492078732597, + "grad_norm": 1.163381589307012, + "learning_rate": 5.459563540496987e-07, + "loss": 0.3361, + "step": 4507 + }, + { + "epoch": 4.326452232357177, + "grad_norm": 0.9948369298622194, + "learning_rate": 5.444352040186712e-07, + "loss": 0.4151, + "step": 4508 + }, + { + "epoch": 4.3274123859817575, + "grad_norm": 1.2134082766312317, + "learning_rate": 5.429160540430533e-07, + "loss": 0.4682, + "step": 4509 + }, + { + "epoch": 4.328372539606337, + "grad_norm": 1.4998570790890307, + "learning_rate": 5.413989048047752e-07, + "loss": 0.3206, + "step": 4510 + }, + { + "epoch": 4.329332693230917, + "grad_norm": 1.3959124111395018, + "learning_rate": 5.398837569848686e-07, + "loss": 0.322, + "step": 4511 + }, + { + "epoch": 4.330292846855497, + "grad_norm": 1.016386897809876, + "learning_rate": 5.38370611263469e-07, + "loss": 0.4268, + "step": 4512 + }, + { + "epoch": 4.331253000480077, + "grad_norm": 1.718250742759537, + "learning_rate": 5.368594683198087e-07, + "loss": 0.45, + "step": 4513 + }, + { + "epoch": 4.3322131541046565, + "grad_norm": 1.1866762837708889, + "learning_rate": 5.353503288322243e-07, + "loss": 0.3966, + "step": 4514 + }, + { + "epoch": 4.333173307729236, + "grad_norm": 0.9253850877437652, + "learning_rate": 5.338431934781535e-07, + "loss": 0.451, + "step": 4515 + }, + { + "epoch": 4.334133461353817, + "grad_norm": 1.2561321092435203, + "learning_rate": 5.323380629341324e-07, + "loss": 0.4467, + "step": 4516 + }, + { + "epoch": 4.335093614978397, + "grad_norm": 1.3632641866979487, + "learning_rate": 5.308349378757988e-07, + "loss": 0.4173, + "step": 4517 + }, + { + "epoch": 4.3360537686029765, + "grad_norm": 1.407036831917338, + "learning_rate": 5.293338189778874e-07, + "loss": 0.2751, + "step": 4518 + }, + { + "epoch": 4.337013922227556, + "grad_norm": 1.2780212410441019, + "learning_rate": 5.278347069142364e-07, + "loss": 0.4011, + "step": 4519 + }, + { + "epoch": 4.337974075852136, + "grad_norm": 1.1422680900779343, + "learning_rate": 5.263376023577821e-07, + "loss": 0.4181, + "step": 4520 + }, + { + "epoch": 4.338934229476716, + "grad_norm": 1.123026904967085, + "learning_rate": 5.248425059805557e-07, + "loss": 0.5086, + "step": 4521 + }, + { + "epoch": 4.339894383101297, + "grad_norm": 1.060183465329235, + "learning_rate": 5.23349418453692e-07, + "loss": 0.4731, + "step": 4522 + }, + { + "epoch": 4.340854536725876, + "grad_norm": 1.2510167414786553, + "learning_rate": 5.218583404474214e-07, + "loss": 0.4983, + "step": 4523 + }, + { + "epoch": 4.341814690350456, + "grad_norm": 1.2345514418247994, + "learning_rate": 5.203692726310738e-07, + "loss": 0.4174, + "step": 4524 + }, + { + "epoch": 4.342774843975036, + "grad_norm": 1.2788479147916623, + "learning_rate": 5.188822156730756e-07, + "loss": 0.4548, + "step": 4525 + }, + { + "epoch": 4.343734997599616, + "grad_norm": 0.8732961751624204, + "learning_rate": 5.173971702409491e-07, + "loss": 0.5299, + "step": 4526 + }, + { + "epoch": 4.344695151224196, + "grad_norm": 1.232851885938405, + "learning_rate": 5.159141370013165e-07, + "loss": 0.4358, + "step": 4527 + }, + { + "epoch": 4.345655304848776, + "grad_norm": 1.9851579327873001, + "learning_rate": 5.144331166198963e-07, + "loss": 0.3698, + "step": 4528 + }, + { + "epoch": 4.346615458473356, + "grad_norm": 1.1441699347447987, + "learning_rate": 5.129541097615009e-07, + "loss": 0.4816, + "step": 4529 + }, + { + "epoch": 4.347575612097936, + "grad_norm": 1.1372585439565508, + "learning_rate": 5.114771170900434e-07, + "loss": 0.3523, + "step": 4530 + }, + { + "epoch": 4.348535765722516, + "grad_norm": 1.1655901493154417, + "learning_rate": 5.100021392685267e-07, + "loss": 0.4205, + "step": 4531 + }, + { + "epoch": 4.3494959193470955, + "grad_norm": 1.5084867324248776, + "learning_rate": 5.08529176959055e-07, + "loss": 0.4414, + "step": 4532 + }, + { + "epoch": 4.350456072971675, + "grad_norm": 1.3280365995889196, + "learning_rate": 5.070582308228256e-07, + "loss": 0.3429, + "step": 4533 + }, + { + "epoch": 4.351416226596255, + "grad_norm": 1.074301942160981, + "learning_rate": 5.05589301520128e-07, + "loss": 0.4907, + "step": 4534 + }, + { + "epoch": 4.352376380220836, + "grad_norm": 1.2007583916027926, + "learning_rate": 5.041223897103509e-07, + "loss": 0.4296, + "step": 4535 + }, + { + "epoch": 4.3533365338454155, + "grad_norm": 0.8454902264937025, + "learning_rate": 5.026574960519747e-07, + "loss": 0.4284, + "step": 4536 + }, + { + "epoch": 4.354296687469995, + "grad_norm": 1.5634155969774313, + "learning_rate": 5.011946212025747e-07, + "loss": 0.4287, + "step": 4537 + }, + { + "epoch": 4.355256841094575, + "grad_norm": 1.2109126206771892, + "learning_rate": 4.997337658188207e-07, + "loss": 0.3732, + "step": 4538 + }, + { + "epoch": 4.356216994719155, + "grad_norm": 0.9904431828162874, + "learning_rate": 4.982749305564727e-07, + "loss": 0.4272, + "step": 4539 + }, + { + "epoch": 4.357177148343735, + "grad_norm": 0.9979274217998053, + "learning_rate": 4.968181160703877e-07, + "loss": 0.4326, + "step": 4540 + }, + { + "epoch": 4.3581373019683145, + "grad_norm": 1.0312499466279428, + "learning_rate": 4.953633230145144e-07, + "loss": 0.2995, + "step": 4541 + }, + { + "epoch": 4.359097455592895, + "grad_norm": 0.7224938097594399, + "learning_rate": 4.93910552041893e-07, + "loss": 0.372, + "step": 4542 + }, + { + "epoch": 4.360057609217475, + "grad_norm": 1.0709144019009873, + "learning_rate": 4.924598038046591e-07, + "loss": 0.397, + "step": 4543 + }, + { + "epoch": 4.361017762842055, + "grad_norm": 2.151572348900691, + "learning_rate": 4.910110789540346e-07, + "loss": 0.4301, + "step": 4544 + }, + { + "epoch": 4.361977916466635, + "grad_norm": 1.1224985143528168, + "learning_rate": 4.895643781403375e-07, + "loss": 0.4491, + "step": 4545 + }, + { + "epoch": 4.362938070091214, + "grad_norm": 1.2376849815327193, + "learning_rate": 4.881197020129797e-07, + "loss": 0.4396, + "step": 4546 + }, + { + "epoch": 4.363898223715794, + "grad_norm": 1.1241381462737696, + "learning_rate": 4.866770512204566e-07, + "loss": 0.4283, + "step": 4547 + }, + { + "epoch": 4.364858377340375, + "grad_norm": 1.0927577584920967, + "learning_rate": 4.852364264103621e-07, + "loss": 0.386, + "step": 4548 + }, + { + "epoch": 4.365818530964955, + "grad_norm": 0.9447907262807417, + "learning_rate": 4.837978282293731e-07, + "loss": 0.4059, + "step": 4549 + }, + { + "epoch": 4.366778684589534, + "grad_norm": 1.0765828788806624, + "learning_rate": 4.823612573232644e-07, + "loss": 0.4517, + "step": 4550 + }, + { + "epoch": 4.367738838214114, + "grad_norm": 1.1052584134893906, + "learning_rate": 4.809267143368979e-07, + "loss": 0.5089, + "step": 4551 + }, + { + "epoch": 4.368698991838694, + "grad_norm": 1.376233074781487, + "learning_rate": 4.794941999142222e-07, + "loss": 0.4837, + "step": 4552 + }, + { + "epoch": 4.369659145463274, + "grad_norm": 1.1806096422823795, + "learning_rate": 4.780637146982781e-07, + "loss": 0.3811, + "step": 4553 + }, + { + "epoch": 4.3706192990878545, + "grad_norm": 1.1241669745360072, + "learning_rate": 4.7663525933119625e-07, + "loss": 0.4013, + "step": 4554 + }, + { + "epoch": 4.371579452712434, + "grad_norm": 1.3671134807528549, + "learning_rate": 4.7520883445419387e-07, + "loss": 0.4703, + "step": 4555 + }, + { + "epoch": 4.372539606337014, + "grad_norm": 0.9831392593229334, + "learning_rate": 4.737844407075798e-07, + "loss": 0.4268, + "step": 4556 + }, + { + "epoch": 4.373499759961594, + "grad_norm": 1.2104135913434775, + "learning_rate": 4.723620787307465e-07, + "loss": 0.4959, + "step": 4557 + }, + { + "epoch": 4.374459913586174, + "grad_norm": 1.4830952674130176, + "learning_rate": 4.7094174916217817e-07, + "loss": 0.4121, + "step": 4558 + }, + { + "epoch": 4.3754200672107535, + "grad_norm": 1.2314396640248406, + "learning_rate": 4.69523452639446e-07, + "loss": 0.4283, + "step": 4559 + }, + { + "epoch": 4.376380220835333, + "grad_norm": 1.2594440126269677, + "learning_rate": 4.6810718979920734e-07, + "loss": 0.551, + "step": 4560 + }, + { + "epoch": 4.377340374459914, + "grad_norm": 1.093121773581581, + "learning_rate": 4.666929612772092e-07, + "loss": 0.4493, + "step": 4561 + }, + { + "epoch": 4.378300528084494, + "grad_norm": 1.137856047787528, + "learning_rate": 4.6528076770828045e-07, + "loss": 0.5164, + "step": 4562 + }, + { + "epoch": 4.379260681709074, + "grad_norm": 0.9180310069138364, + "learning_rate": 4.638706097263429e-07, + "loss": 0.3896, + "step": 4563 + }, + { + "epoch": 4.380220835333653, + "grad_norm": 1.0180404701188983, + "learning_rate": 4.6246248796440074e-07, + "loss": 0.382, + "step": 4564 + }, + { + "epoch": 4.381180988958233, + "grad_norm": 1.5199235971437013, + "learning_rate": 4.6105640305454337e-07, + "loss": 0.4492, + "step": 4565 + }, + { + "epoch": 4.382141142582813, + "grad_norm": 1.0314532935913057, + "learning_rate": 4.5965235562794754e-07, + "loss": 0.4477, + "step": 4566 + }, + { + "epoch": 4.383101296207393, + "grad_norm": 1.064903726380619, + "learning_rate": 4.582503463148763e-07, + "loss": 0.5554, + "step": 4567 + }, + { + "epoch": 4.384061449831973, + "grad_norm": 1.5702707118705925, + "learning_rate": 4.5685037574467574e-07, + "loss": 0.3958, + "step": 4568 + }, + { + "epoch": 4.385021603456553, + "grad_norm": 1.5379656344648542, + "learning_rate": 4.5545244454577863e-07, + "loss": 0.3598, + "step": 4569 + }, + { + "epoch": 4.385981757081133, + "grad_norm": 0.8654657383084474, + "learning_rate": 4.540565533457003e-07, + "loss": 0.4933, + "step": 4570 + }, + { + "epoch": 4.386941910705713, + "grad_norm": 1.7283784967091642, + "learning_rate": 4.5266270277104116e-07, + "loss": 0.3767, + "step": 4571 + }, + { + "epoch": 4.387902064330293, + "grad_norm": 1.2237399318454918, + "learning_rate": 4.512708934474863e-07, + "loss": 0.4854, + "step": 4572 + }, + { + "epoch": 4.388862217954872, + "grad_norm": 0.9246663696882009, + "learning_rate": 4.498811259998048e-07, + "loss": 0.5199, + "step": 4573 + }, + { + "epoch": 4.389822371579453, + "grad_norm": 0.9660510146192445, + "learning_rate": 4.4849340105184823e-07, + "loss": 0.5644, + "step": 4574 + }, + { + "epoch": 4.390782525204033, + "grad_norm": 0.9728257023773027, + "learning_rate": 4.4710771922654936e-07, + "loss": 0.4811, + "step": 4575 + }, + { + "epoch": 4.391742678828613, + "grad_norm": 0.9199310882092127, + "learning_rate": 4.4572408114592845e-07, + "loss": 0.4305, + "step": 4576 + }, + { + "epoch": 4.3927028324531925, + "grad_norm": 0.9779943800547182, + "learning_rate": 4.4434248743108587e-07, + "loss": 0.4469, + "step": 4577 + }, + { + "epoch": 4.393662986077772, + "grad_norm": 1.1789354200193212, + "learning_rate": 4.4296293870220276e-07, + "loss": 0.462, + "step": 4578 + }, + { + "epoch": 4.394623139702352, + "grad_norm": 0.9951559345660517, + "learning_rate": 4.415854355785443e-07, + "loss": 0.399, + "step": 4579 + }, + { + "epoch": 4.395583293326933, + "grad_norm": 0.9347679612202262, + "learning_rate": 4.4020997867845695e-07, + "loss": 0.4685, + "step": 4580 + }, + { + "epoch": 4.396543446951513, + "grad_norm": 1.40442396082108, + "learning_rate": 4.388365686193685e-07, + "loss": 0.4198, + "step": 4581 + }, + { + "epoch": 4.397503600576092, + "grad_norm": 1.1920837031436664, + "learning_rate": 4.3746520601778974e-07, + "loss": 0.452, + "step": 4582 + }, + { + "epoch": 4.398463754200672, + "grad_norm": 0.8465765408725517, + "learning_rate": 4.3609589148930766e-07, + "loss": 0.3935, + "step": 4583 + }, + { + "epoch": 4.399423907825252, + "grad_norm": 1.0887085174734115, + "learning_rate": 4.3472862564859496e-07, + "loss": 0.5193, + "step": 4584 + }, + { + "epoch": 4.400384061449832, + "grad_norm": 0.9376187730327918, + "learning_rate": 4.3336340910940186e-07, + "loss": 0.3983, + "step": 4585 + }, + { + "epoch": 4.4013442150744115, + "grad_norm": 1.0778029589212026, + "learning_rate": 4.320002424845604e-07, + "loss": 0.407, + "step": 4586 + }, + { + "epoch": 4.402304368698992, + "grad_norm": 1.2944301989546776, + "learning_rate": 4.3063912638598106e-07, + "loss": 0.3974, + "step": 4587 + }, + { + "epoch": 4.403264522323572, + "grad_norm": 1.031807254545532, + "learning_rate": 4.292800614246534e-07, + "loss": 0.4133, + "step": 4588 + }, + { + "epoch": 4.404224675948152, + "grad_norm": 0.8019802099081091, + "learning_rate": 4.2792304821064833e-07, + "loss": 0.448, + "step": 4589 + }, + { + "epoch": 4.405184829572732, + "grad_norm": 1.170700206301413, + "learning_rate": 4.265680873531136e-07, + "loss": 0.3703, + "step": 4590 + }, + { + "epoch": 4.406144983197311, + "grad_norm": 1.059936086357254, + "learning_rate": 4.2521517946027715e-07, + "loss": 0.4626, + "step": 4591 + }, + { + "epoch": 4.407105136821891, + "grad_norm": 1.0063863096673182, + "learning_rate": 4.2386432513944587e-07, + "loss": 0.5266, + "step": 4592 + }, + { + "epoch": 4.408065290446471, + "grad_norm": 0.9924005972054886, + "learning_rate": 4.225155249970014e-07, + "loss": 0.4086, + "step": 4593 + }, + { + "epoch": 4.409025444071052, + "grad_norm": 1.0385329428389618, + "learning_rate": 4.2116877963840786e-07, + "loss": 0.4435, + "step": 4594 + }, + { + "epoch": 4.4099855976956315, + "grad_norm": 1.037936602005678, + "learning_rate": 4.198240896682049e-07, + "loss": 0.4398, + "step": 4595 + }, + { + "epoch": 4.410945751320211, + "grad_norm": 1.6847855823495415, + "learning_rate": 4.18481455690008e-07, + "loss": 0.3883, + "step": 4596 + }, + { + "epoch": 4.411905904944791, + "grad_norm": 1.2269550769225495, + "learning_rate": 4.1714087830651226e-07, + "loss": 0.3814, + "step": 4597 + }, + { + "epoch": 4.412866058569371, + "grad_norm": 1.4977485360990013, + "learning_rate": 4.1580235811948753e-07, + "loss": 0.452, + "step": 4598 + }, + { + "epoch": 4.413826212193951, + "grad_norm": 1.1336189339111895, + "learning_rate": 4.14465895729782e-07, + "loss": 0.3192, + "step": 4599 + }, + { + "epoch": 4.414786365818531, + "grad_norm": 1.0160900366957863, + "learning_rate": 4.131314917373208e-07, + "loss": 0.4128, + "step": 4600 + }, + { + "epoch": 4.415746519443111, + "grad_norm": 0.9084506981875303, + "learning_rate": 4.1179914674110035e-07, + "loss": 0.5587, + "step": 4601 + }, + { + "epoch": 4.416706673067691, + "grad_norm": 1.0202889906469483, + "learning_rate": 4.1046886133919825e-07, + "loss": 0.3186, + "step": 4602 + }, + { + "epoch": 4.417666826692271, + "grad_norm": 1.184635343110861, + "learning_rate": 4.0914063612876473e-07, + "loss": 0.2861, + "step": 4603 + }, + { + "epoch": 4.4186269803168505, + "grad_norm": 1.0105819774998916, + "learning_rate": 4.0781447170602617e-07, + "loss": 0.4878, + "step": 4604 + }, + { + "epoch": 4.41958713394143, + "grad_norm": 0.916800302405071, + "learning_rate": 4.064903686662841e-07, + "loss": 0.4673, + "step": 4605 + }, + { + "epoch": 4.420547287566011, + "grad_norm": 1.241109346428586, + "learning_rate": 4.0516832760391256e-07, + "loss": 0.4678, + "step": 4606 + }, + { + "epoch": 4.421507441190591, + "grad_norm": 1.07599400529569, + "learning_rate": 4.0384834911236294e-07, + "loss": 0.3828, + "step": 4607 + }, + { + "epoch": 4.422467594815171, + "grad_norm": 1.1164020728168926, + "learning_rate": 4.025304337841607e-07, + "loss": 0.4583, + "step": 4608 + }, + { + "epoch": 4.42342774843975, + "grad_norm": 1.198347211801293, + "learning_rate": 4.012145822109015e-07, + "loss": 0.35, + "step": 4609 + }, + { + "epoch": 4.42438790206433, + "grad_norm": 0.9597224092712009, + "learning_rate": 3.9990079498325786e-07, + "loss": 0.4111, + "step": 4610 + }, + { + "epoch": 4.42534805568891, + "grad_norm": 1.0985588341273547, + "learning_rate": 3.985890726909747e-07, + "loss": 0.5562, + "step": 4611 + }, + { + "epoch": 4.42630820931349, + "grad_norm": 1.1466011011349508, + "learning_rate": 3.97279415922871e-07, + "loss": 0.4417, + "step": 4612 + }, + { + "epoch": 4.4272683629380705, + "grad_norm": 1.2435141211321377, + "learning_rate": 3.9597182526683763e-07, + "loss": 0.3718, + "step": 4613 + }, + { + "epoch": 4.42822851656265, + "grad_norm": 0.9479447933179591, + "learning_rate": 3.946663013098373e-07, + "loss": 0.3907, + "step": 4614 + }, + { + "epoch": 4.42918867018723, + "grad_norm": 1.2414848856718508, + "learning_rate": 3.9336284463790566e-07, + "loss": 0.3018, + "step": 4615 + }, + { + "epoch": 4.43014882381181, + "grad_norm": 1.2030546961756063, + "learning_rate": 3.9206145583615076e-07, + "loss": 0.3973, + "step": 4616 + }, + { + "epoch": 4.43110897743639, + "grad_norm": 1.081223155803227, + "learning_rate": 3.90762135488752e-07, + "loss": 0.5062, + "step": 4617 + }, + { + "epoch": 4.4320691310609694, + "grad_norm": 1.0381226103351264, + "learning_rate": 3.894648841789617e-07, + "loss": 0.5091, + "step": 4618 + }, + { + "epoch": 4.433029284685549, + "grad_norm": 2.100859562932593, + "learning_rate": 3.881697024890996e-07, + "loss": 0.4231, + "step": 4619 + }, + { + "epoch": 4.43398943831013, + "grad_norm": 1.0535313496527898, + "learning_rate": 3.8687659100056063e-07, + "loss": 0.537, + "step": 4620 + }, + { + "epoch": 4.43494959193471, + "grad_norm": 0.9168189758471965, + "learning_rate": 3.8558555029380937e-07, + "loss": 0.4552, + "step": 4621 + }, + { + "epoch": 4.4359097455592895, + "grad_norm": 0.8307751269320371, + "learning_rate": 3.8429658094837896e-07, + "loss": 0.3949, + "step": 4622 + }, + { + "epoch": 4.436869899183869, + "grad_norm": 1.0567023738650025, + "learning_rate": 3.830096835428737e-07, + "loss": 0.4792, + "step": 4623 + }, + { + "epoch": 4.437830052808449, + "grad_norm": 1.0118897659015862, + "learning_rate": 3.8172485865496876e-07, + "loss": 0.4811, + "step": 4624 + }, + { + "epoch": 4.438790206433029, + "grad_norm": 1.0944763064896719, + "learning_rate": 3.8044210686140784e-07, + "loss": 0.3227, + "step": 4625 + }, + { + "epoch": 4.43975036005761, + "grad_norm": 0.8597578977294977, + "learning_rate": 3.791614287380063e-07, + "loss": 0.4486, + "step": 4626 + }, + { + "epoch": 4.440710513682189, + "grad_norm": 1.3280114692241722, + "learning_rate": 3.778828248596439e-07, + "loss": 0.3981, + "step": 4627 + }, + { + "epoch": 4.441670667306769, + "grad_norm": 0.9824963134424053, + "learning_rate": 3.766062958002742e-07, + "loss": 0.4785, + "step": 4628 + }, + { + "epoch": 4.442630820931349, + "grad_norm": 1.0269967677537446, + "learning_rate": 3.7533184213291665e-07, + "loss": 0.5201, + "step": 4629 + }, + { + "epoch": 4.443590974555929, + "grad_norm": 1.2035620958243223, + "learning_rate": 3.7405946442966023e-07, + "loss": 0.4753, + "step": 4630 + }, + { + "epoch": 4.444551128180509, + "grad_norm": 0.9811362135399778, + "learning_rate": 3.727891632616626e-07, + "loss": 0.4886, + "step": 4631 + }, + { + "epoch": 4.445511281805089, + "grad_norm": 1.1164479040761177, + "learning_rate": 3.7152093919914644e-07, + "loss": 0.3473, + "step": 4632 + }, + { + "epoch": 4.446471435429669, + "grad_norm": 1.303352400684884, + "learning_rate": 3.702547928114053e-07, + "loss": 0.468, + "step": 4633 + }, + { + "epoch": 4.447431589054249, + "grad_norm": 0.8755270873828765, + "learning_rate": 3.689907246667984e-07, + "loss": 0.3892, + "step": 4634 + }, + { + "epoch": 4.448391742678829, + "grad_norm": 0.9283241839457869, + "learning_rate": 3.677287353327519e-07, + "loss": 0.4845, + "step": 4635 + }, + { + "epoch": 4.449351896303408, + "grad_norm": 1.049123002332921, + "learning_rate": 3.6646882537576144e-07, + "loss": 0.3652, + "step": 4636 + }, + { + "epoch": 4.450312049927988, + "grad_norm": 0.9071035520101988, + "learning_rate": 3.652109953613836e-07, + "loss": 0.401, + "step": 4637 + }, + { + "epoch": 4.451272203552568, + "grad_norm": 0.9180496356241947, + "learning_rate": 3.639552458542478e-07, + "loss": 0.4459, + "step": 4638 + }, + { + "epoch": 4.452232357177149, + "grad_norm": 1.1325922770177368, + "learning_rate": 3.627015774180459e-07, + "loss": 0.4246, + "step": 4639 + }, + { + "epoch": 4.4531925108017285, + "grad_norm": 1.1073477887363432, + "learning_rate": 3.614499906155355e-07, + "loss": 0.4852, + "step": 4640 + }, + { + "epoch": 4.454152664426308, + "grad_norm": 0.9385430408233402, + "learning_rate": 3.602004860085406e-07, + "loss": 0.4651, + "step": 4641 + }, + { + "epoch": 4.455112818050888, + "grad_norm": 1.0857351597359235, + "learning_rate": 3.589530641579514e-07, + "loss": 0.4507, + "step": 4642 + }, + { + "epoch": 4.456072971675468, + "grad_norm": 0.8128311834316073, + "learning_rate": 3.577077256237216e-07, + "loss": 0.4441, + "step": 4643 + }, + { + "epoch": 4.457033125300048, + "grad_norm": 1.0808085416470794, + "learning_rate": 3.564644709648707e-07, + "loss": 0.4415, + "step": 4644 + }, + { + "epoch": 4.4579932789246275, + "grad_norm": 1.1883011786961952, + "learning_rate": 3.5522330073948177e-07, + "loss": 0.3925, + "step": 4645 + }, + { + "epoch": 4.458953432549208, + "grad_norm": 1.0792365759954712, + "learning_rate": 3.53984215504703e-07, + "loss": 0.5069, + "step": 4646 + }, + { + "epoch": 4.459913586173788, + "grad_norm": 1.7224546507341325, + "learning_rate": 3.5274721581674664e-07, + "loss": 0.4679, + "step": 4647 + }, + { + "epoch": 4.460873739798368, + "grad_norm": 1.0270505359466444, + "learning_rate": 3.5151230223088915e-07, + "loss": 0.3891, + "step": 4648 + }, + { + "epoch": 4.461833893422948, + "grad_norm": 1.1936722087015554, + "learning_rate": 3.5027947530147045e-07, + "loss": 0.4193, + "step": 4649 + }, + { + "epoch": 4.462794047047527, + "grad_norm": 1.2134318744599935, + "learning_rate": 3.4904873558189055e-07, + "loss": 0.3959, + "step": 4650 + }, + { + "epoch": 4.463754200672107, + "grad_norm": 1.3284386265098087, + "learning_rate": 3.4782008362461817e-07, + "loss": 0.3829, + "step": 4651 + }, + { + "epoch": 4.464714354296688, + "grad_norm": 1.0788159580593186, + "learning_rate": 3.4659351998118204e-07, + "loss": 0.3785, + "step": 4652 + }, + { + "epoch": 4.465674507921268, + "grad_norm": 1.34088185550518, + "learning_rate": 3.4536904520217184e-07, + "loss": 0.3601, + "step": 4653 + }, + { + "epoch": 4.466634661545847, + "grad_norm": 0.9369895407189601, + "learning_rate": 3.441466598372423e-07, + "loss": 0.4827, + "step": 4654 + }, + { + "epoch": 4.467594815170427, + "grad_norm": 1.3212086612814424, + "learning_rate": 3.429263644351083e-07, + "loss": 0.3879, + "step": 4655 + }, + { + "epoch": 4.468554968795007, + "grad_norm": 1.5011069475026733, + "learning_rate": 3.417081595435484e-07, + "loss": 0.4307, + "step": 4656 + }, + { + "epoch": 4.469515122419587, + "grad_norm": 1.0083799795685993, + "learning_rate": 3.4049204570940185e-07, + "loss": 0.4076, + "step": 4657 + }, + { + "epoch": 4.4704752760441675, + "grad_norm": 1.241800592198575, + "learning_rate": 3.392780234785681e-07, + "loss": 0.4908, + "step": 4658 + }, + { + "epoch": 4.471435429668747, + "grad_norm": 0.9537669496623953, + "learning_rate": 3.3806609339600894e-07, + "loss": 0.4485, + "step": 4659 + }, + { + "epoch": 4.472395583293327, + "grad_norm": 1.2162317603348596, + "learning_rate": 3.36856256005747e-07, + "loss": 0.3165, + "step": 4660 + }, + { + "epoch": 4.473355736917907, + "grad_norm": 1.027184108405054, + "learning_rate": 3.356485118508657e-07, + "loss": 0.507, + "step": 4661 + }, + { + "epoch": 4.474315890542487, + "grad_norm": 0.935849857099973, + "learning_rate": 3.3444286147350955e-07, + "loss": 0.5405, + "step": 4662 + }, + { + "epoch": 4.4752760441670665, + "grad_norm": 1.3835524173254776, + "learning_rate": 3.332393054148786e-07, + "loss": 0.4296, + "step": 4663 + }, + { + "epoch": 4.476236197791646, + "grad_norm": 0.8944764702806797, + "learning_rate": 3.3203784421524056e-07, + "loss": 0.3917, + "step": 4664 + }, + { + "epoch": 4.477196351416227, + "grad_norm": 0.9643049038902026, + "learning_rate": 3.3083847841391514e-07, + "loss": 0.4942, + "step": 4665 + }, + { + "epoch": 4.478156505040807, + "grad_norm": 1.1445453507300092, + "learning_rate": 3.296412085492867e-07, + "loss": 0.3323, + "step": 4666 + }, + { + "epoch": 4.4791166586653866, + "grad_norm": 0.9241544707478011, + "learning_rate": 3.2844603515879647e-07, + "loss": 0.4916, + "step": 4667 + }, + { + "epoch": 4.480076812289966, + "grad_norm": 0.9935863549035892, + "learning_rate": 3.272529587789447e-07, + "loss": 0.5704, + "step": 4668 + }, + { + "epoch": 4.481036965914546, + "grad_norm": 0.8865031791214463, + "learning_rate": 3.2606197994529076e-07, + "loss": 0.4223, + "step": 4669 + }, + { + "epoch": 4.481997119539126, + "grad_norm": 1.3736724837829202, + "learning_rate": 3.2487309919245313e-07, + "loss": 0.2838, + "step": 4670 + }, + { + "epoch": 4.482957273163706, + "grad_norm": 1.3502526512112645, + "learning_rate": 3.236863170541066e-07, + "loss": 0.4459, + "step": 4671 + }, + { + "epoch": 4.483917426788286, + "grad_norm": 1.0425731557530016, + "learning_rate": 3.22501634062985e-07, + "loss": 0.372, + "step": 4672 + }, + { + "epoch": 4.484877580412866, + "grad_norm": 1.0981722570117114, + "learning_rate": 3.213190507508801e-07, + "loss": 0.4103, + "step": 4673 + }, + { + "epoch": 4.485837734037446, + "grad_norm": 0.9064090652531295, + "learning_rate": 3.2013856764864126e-07, + "loss": 0.423, + "step": 4674 + }, + { + "epoch": 4.486797887662026, + "grad_norm": 1.3331524416886484, + "learning_rate": 3.1896018528617456e-07, + "loss": 0.4736, + "step": 4675 + }, + { + "epoch": 4.487758041286606, + "grad_norm": 1.3950619923604852, + "learning_rate": 3.17783904192443e-07, + "loss": 0.4697, + "step": 4676 + }, + { + "epoch": 4.488718194911185, + "grad_norm": 1.356830743593381, + "learning_rate": 3.166097248954669e-07, + "loss": 0.3645, + "step": 4677 + }, + { + "epoch": 4.489678348535766, + "grad_norm": 0.9845654489031277, + "learning_rate": 3.15437647922322e-07, + "loss": 0.4774, + "step": 4678 + }, + { + "epoch": 4.490638502160346, + "grad_norm": 1.502083952953334, + "learning_rate": 3.1426767379914234e-07, + "loss": 0.3804, + "step": 4679 + }, + { + "epoch": 4.491598655784926, + "grad_norm": 0.933278485317963, + "learning_rate": 3.1309980305111674e-07, + "loss": 0.4455, + "step": 4680 + }, + { + "epoch": 4.4925588094095055, + "grad_norm": 0.9574803410164137, + "learning_rate": 3.119340362024875e-07, + "loss": 0.3957, + "step": 4681 + }, + { + "epoch": 4.493518963034085, + "grad_norm": 1.1913148882079458, + "learning_rate": 3.1077037377655827e-07, + "loss": 0.4653, + "step": 4682 + }, + { + "epoch": 4.494479116658665, + "grad_norm": 0.9520078309381338, + "learning_rate": 3.096088162956834e-07, + "loss": 0.4846, + "step": 4683 + }, + { + "epoch": 4.495439270283246, + "grad_norm": 1.1173728527292974, + "learning_rate": 3.084493642812725e-07, + "loss": 0.3963, + "step": 4684 + }, + { + "epoch": 4.4963994239078255, + "grad_norm": 1.051503282047858, + "learning_rate": 3.072920182537925e-07, + "loss": 0.3211, + "step": 4685 + }, + { + "epoch": 4.497359577532405, + "grad_norm": 1.0697288583757516, + "learning_rate": 3.0613677873276293e-07, + "loss": 0.4574, + "step": 4686 + }, + { + "epoch": 4.498319731156985, + "grad_norm": 0.9153817193755083, + "learning_rate": 3.049836462367589e-07, + "loss": 0.508, + "step": 4687 + }, + { + "epoch": 4.499279884781565, + "grad_norm": 1.0656737460462218, + "learning_rate": 3.038326212834103e-07, + "loss": 0.4427, + "step": 4688 + }, + { + "epoch": 4.500240038406145, + "grad_norm": 1.5554230690225583, + "learning_rate": 3.0268370438939767e-07, + "loss": 0.4265, + "step": 4689 + }, + { + "epoch": 4.5012001920307245, + "grad_norm": 0.9826281861393134, + "learning_rate": 3.015368960704584e-07, + "loss": 0.3888, + "step": 4690 + }, + { + "epoch": 4.502160345655305, + "grad_norm": 0.7502096687581634, + "learning_rate": 3.003921968413831e-07, + "loss": 0.4591, + "step": 4691 + }, + { + "epoch": 4.503120499279885, + "grad_norm": 1.3632152610718085, + "learning_rate": 2.9924960721601393e-07, + "loss": 0.3862, + "step": 4692 + }, + { + "epoch": 4.504080652904465, + "grad_norm": 1.283898241038968, + "learning_rate": 2.98109127707249e-07, + "loss": 0.3637, + "step": 4693 + }, + { + "epoch": 4.505040806529045, + "grad_norm": 1.1141174564795169, + "learning_rate": 2.9697075882703486e-07, + "loss": 0.4876, + "step": 4694 + }, + { + "epoch": 4.506000960153624, + "grad_norm": 1.4629851634358693, + "learning_rate": 2.95834501086375e-07, + "loss": 0.4729, + "step": 4695 + }, + { + "epoch": 4.506961113778204, + "grad_norm": 1.1765119852314543, + "learning_rate": 2.9470035499532425e-07, + "loss": 0.5148, + "step": 4696 + }, + { + "epoch": 4.507921267402784, + "grad_norm": 1.2243181961508938, + "learning_rate": 2.9356832106298584e-07, + "loss": 0.3447, + "step": 4697 + }, + { + "epoch": 4.508881421027365, + "grad_norm": 1.170493090676355, + "learning_rate": 2.9243839979751943e-07, + "loss": 0.4483, + "step": 4698 + }, + { + "epoch": 4.5098415746519445, + "grad_norm": 1.0775753225959048, + "learning_rate": 2.913105917061343e-07, + "loss": 0.4165, + "step": 4699 + }, + { + "epoch": 4.510801728276524, + "grad_norm": 0.9510803244993737, + "learning_rate": 2.9018489729509103e-07, + "loss": 0.4929, + "step": 4700 + }, + { + "epoch": 4.511761881901104, + "grad_norm": 0.8319880229367609, + "learning_rate": 2.8906131706970375e-07, + "loss": 0.5482, + "step": 4701 + }, + { + "epoch": 4.512722035525684, + "grad_norm": 0.8943703652143729, + "learning_rate": 2.879398515343329e-07, + "loss": 0.4644, + "step": 4702 + }, + { + "epoch": 4.5136821891502645, + "grad_norm": 1.5975482621957606, + "learning_rate": 2.8682050119239357e-07, + "loss": 0.4139, + "step": 4703 + }, + { + "epoch": 4.514642342774844, + "grad_norm": 1.3617928405004596, + "learning_rate": 2.857032665463499e-07, + "loss": 0.4734, + "step": 4704 + }, + { + "epoch": 4.515602496399424, + "grad_norm": 1.5159592096243613, + "learning_rate": 2.8458814809771694e-07, + "loss": 0.4307, + "step": 4705 + }, + { + "epoch": 4.516562650024004, + "grad_norm": 0.7069287483087185, + "learning_rate": 2.834751463470603e-07, + "loss": 0.4155, + "step": 4706 + }, + { + "epoch": 4.517522803648584, + "grad_norm": 1.1466902052701584, + "learning_rate": 2.823642617939926e-07, + "loss": 0.3466, + "step": 4707 + }, + { + "epoch": 4.5184829572731635, + "grad_norm": 1.14303972503422, + "learning_rate": 2.8125549493717984e-07, + "loss": 0.4133, + "step": 4708 + }, + { + "epoch": 4.519443110897743, + "grad_norm": 0.9522121915821085, + "learning_rate": 2.8014884627433404e-07, + "loss": 0.5797, + "step": 4709 + }, + { + "epoch": 4.520403264522324, + "grad_norm": 1.5584119740143225, + "learning_rate": 2.7904431630221937e-07, + "loss": 0.3904, + "step": 4710 + }, + { + "epoch": 4.521363418146904, + "grad_norm": 0.9210416961243966, + "learning_rate": 2.7794190551664604e-07, + "loss": 0.3868, + "step": 4711 + }, + { + "epoch": 4.522323571771484, + "grad_norm": 1.3442494699352647, + "learning_rate": 2.7684161441247584e-07, + "loss": 0.4929, + "step": 4712 + }, + { + "epoch": 4.523283725396063, + "grad_norm": 0.8237913680261635, + "learning_rate": 2.75743443483617e-07, + "loss": 0.3893, + "step": 4713 + }, + { + "epoch": 4.524243879020643, + "grad_norm": 0.9804219338583083, + "learning_rate": 2.746473932230281e-07, + "loss": 0.4672, + "step": 4714 + }, + { + "epoch": 4.525204032645223, + "grad_norm": 0.9146319451455552, + "learning_rate": 2.7355346412271223e-07, + "loss": 0.4603, + "step": 4715 + }, + { + "epoch": 4.526164186269803, + "grad_norm": 1.180443914481596, + "learning_rate": 2.7246165667372314e-07, + "loss": 0.4453, + "step": 4716 + }, + { + "epoch": 4.5271243398943835, + "grad_norm": 1.0442443022965664, + "learning_rate": 2.713719713661617e-07, + "loss": 0.3597, + "step": 4717 + }, + { + "epoch": 4.528084493518963, + "grad_norm": 0.9389450189934472, + "learning_rate": 2.702844086891765e-07, + "loss": 0.473, + "step": 4718 + }, + { + "epoch": 4.529044647143543, + "grad_norm": 1.1429502823703277, + "learning_rate": 2.6919896913096264e-07, + "loss": 0.4946, + "step": 4719 + }, + { + "epoch": 4.530004800768123, + "grad_norm": 1.285427504509259, + "learning_rate": 2.6811565317876154e-07, + "loss": 0.3764, + "step": 4720 + }, + { + "epoch": 4.530964954392703, + "grad_norm": 1.0897933396641855, + "learning_rate": 2.67034461318863e-07, + "loss": 0.4039, + "step": 4721 + }, + { + "epoch": 4.531925108017282, + "grad_norm": 1.0615716704044484, + "learning_rate": 2.659553940366016e-07, + "loss": 0.4663, + "step": 4722 + }, + { + "epoch": 4.532885261641862, + "grad_norm": 1.4840624933597613, + "learning_rate": 2.6487845181635994e-07, + "loss": 0.4214, + "step": 4723 + }, + { + "epoch": 4.533845415266443, + "grad_norm": 1.33164983121705, + "learning_rate": 2.638036351415668e-07, + "loss": 0.4708, + "step": 4724 + }, + { + "epoch": 4.534805568891023, + "grad_norm": 1.1187586323612524, + "learning_rate": 2.627309444946929e-07, + "loss": 0.4748, + "step": 4725 + }, + { + "epoch": 4.5357657225156025, + "grad_norm": 1.31466900113043, + "learning_rate": 2.6166038035725984e-07, + "loss": 0.3578, + "step": 4726 + }, + { + "epoch": 4.536725876140182, + "grad_norm": 1.4360734524150511, + "learning_rate": 2.605919432098331e-07, + "loss": 0.3739, + "step": 4727 + }, + { + "epoch": 4.537686029764762, + "grad_norm": 1.1550705092419071, + "learning_rate": 2.5952563353202064e-07, + "loss": 0.483, + "step": 4728 + }, + { + "epoch": 4.538646183389343, + "grad_norm": 0.8372637074357283, + "learning_rate": 2.584614518024786e-07, + "loss": 0.4724, + "step": 4729 + }, + { + "epoch": 4.539606337013923, + "grad_norm": 1.6567639736734052, + "learning_rate": 2.573993984989065e-07, + "loss": 0.3116, + "step": 4730 + }, + { + "epoch": 4.540566490638502, + "grad_norm": 1.4289361488618597, + "learning_rate": 2.5633947409804907e-07, + "loss": 0.3932, + "step": 4731 + }, + { + "epoch": 4.541526644263082, + "grad_norm": 1.2129696504945664, + "learning_rate": 2.552816790756951e-07, + "loss": 0.441, + "step": 4732 + }, + { + "epoch": 4.542486797887662, + "grad_norm": 1.2536367999467943, + "learning_rate": 2.542260139066771e-07, + "loss": 0.4802, + "step": 4733 + }, + { + "epoch": 4.543446951512242, + "grad_norm": 1.3447927634295767, + "learning_rate": 2.5317247906487153e-07, + "loss": 0.3472, + "step": 4734 + }, + { + "epoch": 4.544407105136822, + "grad_norm": 1.0619723245560864, + "learning_rate": 2.521210750231995e-07, + "loss": 0.4028, + "step": 4735 + }, + { + "epoch": 4.545367258761402, + "grad_norm": 0.7734107090656136, + "learning_rate": 2.510718022536246e-07, + "loss": 0.384, + "step": 4736 + }, + { + "epoch": 4.546327412385982, + "grad_norm": 1.1036974134222308, + "learning_rate": 2.5002466122715505e-07, + "loss": 0.4598, + "step": 4737 + }, + { + "epoch": 4.547287566010562, + "grad_norm": 1.2462577188622184, + "learning_rate": 2.489796524138394e-07, + "loss": 0.4144, + "step": 4738 + }, + { + "epoch": 4.548247719635142, + "grad_norm": 0.971740232758064, + "learning_rate": 2.479367762827739e-07, + "loss": 0.4539, + "step": 4739 + }, + { + "epoch": 4.549207873259721, + "grad_norm": 1.2436283674341448, + "learning_rate": 2.46896033302092e-07, + "loss": 0.4322, + "step": 4740 + }, + { + "epoch": 4.550168026884301, + "grad_norm": 0.9956978727370558, + "learning_rate": 2.4585742393897274e-07, + "loss": 0.4363, + "step": 4741 + }, + { + "epoch": 4.551128180508881, + "grad_norm": 1.1418167817812876, + "learning_rate": 2.44820948659637e-07, + "loss": 0.5482, + "step": 4742 + }, + { + "epoch": 4.552088334133462, + "grad_norm": 0.9323243316709064, + "learning_rate": 2.4378660792934715e-07, + "loss": 0.5052, + "step": 4743 + }, + { + "epoch": 4.5530484877580415, + "grad_norm": 1.0959351490461227, + "learning_rate": 2.4275440221240786e-07, + "loss": 0.3751, + "step": 4744 + }, + { + "epoch": 4.554008641382621, + "grad_norm": 1.0397603007711043, + "learning_rate": 2.4172433197216625e-07, + "loss": 0.4415, + "step": 4745 + }, + { + "epoch": 4.554968795007201, + "grad_norm": 1.1910560483330195, + "learning_rate": 2.4069639767100805e-07, + "loss": 0.4517, + "step": 4746 + }, + { + "epoch": 4.555928948631781, + "grad_norm": 0.9953672139473719, + "learning_rate": 2.39670599770363e-07, + "loss": 0.4191, + "step": 4747 + }, + { + "epoch": 4.556889102256361, + "grad_norm": 1.0480721588379325, + "learning_rate": 2.3864693873070054e-07, + "loss": 0.4109, + "step": 4748 + }, + { + "epoch": 4.5578492558809405, + "grad_norm": 1.2714793075569117, + "learning_rate": 2.376254150115309e-07, + "loss": 0.3871, + "step": 4749 + }, + { + "epoch": 4.558809409505521, + "grad_norm": 1.1473014772326129, + "learning_rate": 2.3660602907140672e-07, + "loss": 0.4092, + "step": 4750 + }, + { + "epoch": 4.559769563130101, + "grad_norm": 0.9050063687795113, + "learning_rate": 2.3558878136791697e-07, + "loss": 0.4069, + "step": 4751 + }, + { + "epoch": 4.560729716754681, + "grad_norm": 1.1588239489815373, + "learning_rate": 2.3457367235769468e-07, + "loss": 0.4068, + "step": 4752 + }, + { + "epoch": 4.5616898703792605, + "grad_norm": 1.4439982559960651, + "learning_rate": 2.335607024964115e-07, + "loss": 0.3089, + "step": 4753 + }, + { + "epoch": 4.56265002400384, + "grad_norm": 0.8533362222895106, + "learning_rate": 2.3254987223877868e-07, + "loss": 0.3971, + "step": 4754 + }, + { + "epoch": 4.563610177628421, + "grad_norm": 0.8538714424911488, + "learning_rate": 2.315411820385477e-07, + "loss": 0.4697, + "step": 4755 + }, + { + "epoch": 4.564570331253001, + "grad_norm": 0.8850294273129156, + "learning_rate": 2.3053463234850637e-07, + "loss": 0.4832, + "step": 4756 + }, + { + "epoch": 4.565530484877581, + "grad_norm": 1.6876140915223485, + "learning_rate": 2.295302236204866e-07, + "loss": 0.4635, + "step": 4757 + }, + { + "epoch": 4.56649063850216, + "grad_norm": 1.7123006347826433, + "learning_rate": 2.2852795630535661e-07, + "loss": 0.423, + "step": 4758 + }, + { + "epoch": 4.56745079212674, + "grad_norm": 1.2815384269336088, + "learning_rate": 2.2752783085302265e-07, + "loss": 0.4572, + "step": 4759 + }, + { + "epoch": 4.56841094575132, + "grad_norm": 1.2177641284463616, + "learning_rate": 2.2652984771243002e-07, + "loss": 0.3956, + "step": 4760 + }, + { + "epoch": 4.5693710993759, + "grad_norm": 1.0489753068784278, + "learning_rate": 2.2553400733156317e-07, + "loss": 0.3929, + "step": 4761 + }, + { + "epoch": 4.5703312530004805, + "grad_norm": 0.8411894507928341, + "learning_rate": 2.2454031015744448e-07, + "loss": 0.4813, + "step": 4762 + }, + { + "epoch": 4.57129140662506, + "grad_norm": 1.072517352365485, + "learning_rate": 2.2354875663613384e-07, + "loss": 0.3862, + "step": 4763 + }, + { + "epoch": 4.57225156024964, + "grad_norm": 1.0333875763883018, + "learning_rate": 2.2255934721272908e-07, + "loss": 0.4784, + "step": 4764 + }, + { + "epoch": 4.57321171387422, + "grad_norm": 0.7960202222219466, + "learning_rate": 2.2157208233136495e-07, + "loss": 0.3792, + "step": 4765 + }, + { + "epoch": 4.5741718674988, + "grad_norm": 1.1073194282711976, + "learning_rate": 2.2058696243521417e-07, + "loss": 0.5568, + "step": 4766 + }, + { + "epoch": 4.5751320211233795, + "grad_norm": 1.3864135854882582, + "learning_rate": 2.196039879664874e-07, + "loss": 0.4284, + "step": 4767 + }, + { + "epoch": 4.576092174747959, + "grad_norm": 1.3626765182526506, + "learning_rate": 2.186231593664312e-07, + "loss": 0.5581, + "step": 4768 + }, + { + "epoch": 4.57705232837254, + "grad_norm": 1.0512069642900155, + "learning_rate": 2.1764447707532722e-07, + "loss": 0.4252, + "step": 4769 + }, + { + "epoch": 4.57801248199712, + "grad_norm": 0.9769423191254444, + "learning_rate": 2.1666794153249792e-07, + "loss": 0.4284, + "step": 4770 + }, + { + "epoch": 4.5789726356216995, + "grad_norm": 0.8878212987644085, + "learning_rate": 2.1569355317629937e-07, + "loss": 0.3714, + "step": 4771 + }, + { + "epoch": 4.579932789246279, + "grad_norm": 1.2888709486320158, + "learning_rate": 2.1472131244412275e-07, + "loss": 0.3304, + "step": 4772 + }, + { + "epoch": 4.580892942870859, + "grad_norm": 0.8345795881164167, + "learning_rate": 2.1375121977239676e-07, + "loss": 0.4859, + "step": 4773 + }, + { + "epoch": 4.581853096495439, + "grad_norm": 1.20415974038489, + "learning_rate": 2.127832755965864e-07, + "loss": 0.4474, + "step": 4774 + }, + { + "epoch": 4.582813250120019, + "grad_norm": 0.9241071701133369, + "learning_rate": 2.118174803511913e-07, + "loss": 0.4495, + "step": 4775 + }, + { + "epoch": 4.583773403744599, + "grad_norm": 1.3261983159135524, + "learning_rate": 2.108538344697464e-07, + "loss": 0.4409, + "step": 4776 + }, + { + "epoch": 4.584733557369179, + "grad_norm": 1.0490455310698863, + "learning_rate": 2.0989233838482225e-07, + "loss": 0.386, + "step": 4777 + }, + { + "epoch": 4.585693710993759, + "grad_norm": 0.9553392408932545, + "learning_rate": 2.0893299252802367e-07, + "loss": 0.5054, + "step": 4778 + }, + { + "epoch": 4.586653864618339, + "grad_norm": 1.0530642504120253, + "learning_rate": 2.0797579732999063e-07, + "loss": 0.3779, + "step": 4779 + }, + { + "epoch": 4.587614018242919, + "grad_norm": 1.898932999851529, + "learning_rate": 2.0702075322039884e-07, + "loss": 0.375, + "step": 4780 + }, + { + "epoch": 4.588574171867499, + "grad_norm": 1.9203080044725702, + "learning_rate": 2.0606786062795702e-07, + "loss": 0.3602, + "step": 4781 + }, + { + "epoch": 4.589534325492079, + "grad_norm": 0.9058597459243907, + "learning_rate": 2.051171199804075e-07, + "loss": 0.4355, + "step": 4782 + }, + { + "epoch": 4.590494479116659, + "grad_norm": 1.2168556019858547, + "learning_rate": 2.0416853170452945e-07, + "loss": 0.4622, + "step": 4783 + }, + { + "epoch": 4.591454632741239, + "grad_norm": 1.4650957281350552, + "learning_rate": 2.032220962261322e-07, + "loss": 0.4145, + "step": 4784 + }, + { + "epoch": 4.5924147863658185, + "grad_norm": 1.30389333164178, + "learning_rate": 2.022778139700615e-07, + "loss": 0.4011, + "step": 4785 + }, + { + "epoch": 4.593374939990398, + "grad_norm": 1.9758087557470718, + "learning_rate": 2.0133568536019498e-07, + "loss": 0.4424, + "step": 4786 + }, + { + "epoch": 4.594335093614978, + "grad_norm": 1.0121911250520825, + "learning_rate": 2.0039571081944431e-07, + "loss": 0.4405, + "step": 4787 + }, + { + "epoch": 4.595295247239559, + "grad_norm": 1.1254257779385113, + "learning_rate": 1.994578907697542e-07, + "loss": 0.3679, + "step": 4788 + }, + { + "epoch": 4.5962554008641385, + "grad_norm": 1.157821565415064, + "learning_rate": 1.985222256321029e-07, + "loss": 0.4925, + "step": 4789 + }, + { + "epoch": 4.597215554488718, + "grad_norm": 1.1266668040253127, + "learning_rate": 1.9758871582649887e-07, + "loss": 0.4615, + "step": 4790 + }, + { + "epoch": 4.598175708113298, + "grad_norm": 1.1725380655090194, + "learning_rate": 1.9665736177198524e-07, + "loss": 0.4579, + "step": 4791 + }, + { + "epoch": 4.599135861737878, + "grad_norm": 1.0331884953100048, + "learning_rate": 1.9572816388663706e-07, + "loss": 0.3595, + "step": 4792 + }, + { + "epoch": 4.600096015362458, + "grad_norm": 0.8329083570637528, + "learning_rate": 1.9480112258756123e-07, + "loss": 0.3751, + "step": 4793 + }, + { + "epoch": 4.6010561689870375, + "grad_norm": 1.1219679308003967, + "learning_rate": 1.9387623829089764e-07, + "loss": 0.3678, + "step": 4794 + }, + { + "epoch": 4.602016322611618, + "grad_norm": 1.1426578543725208, + "learning_rate": 1.929535114118153e-07, + "loss": 0.4643, + "step": 4795 + }, + { + "epoch": 4.602976476236198, + "grad_norm": 1.839951916892278, + "learning_rate": 1.920329423645173e-07, + "loss": 0.3925, + "step": 4796 + }, + { + "epoch": 4.603936629860778, + "grad_norm": 1.3150107292337425, + "learning_rate": 1.9111453156223704e-07, + "loss": 0.3704, + "step": 4797 + }, + { + "epoch": 4.604896783485358, + "grad_norm": 1.709667630194013, + "learning_rate": 1.901982794172391e-07, + "loss": 0.3657, + "step": 4798 + }, + { + "epoch": 4.605856937109937, + "grad_norm": 1.23638543550137, + "learning_rate": 1.8928418634082068e-07, + "loss": 0.4846, + "step": 4799 + }, + { + "epoch": 4.606817090734517, + "grad_norm": 0.6429848154002672, + "learning_rate": 1.8837225274330572e-07, + "loss": 0.3755, + "step": 4800 + }, + { + "epoch": 4.607777244359097, + "grad_norm": 1.178232230479067, + "learning_rate": 1.87462479034054e-07, + "loss": 0.5208, + "step": 4801 + }, + { + "epoch": 4.608737397983678, + "grad_norm": 1.0899899828529767, + "learning_rate": 1.8655486562145275e-07, + "loss": 0.476, + "step": 4802 + }, + { + "epoch": 4.6096975516082574, + "grad_norm": 1.2574117619853196, + "learning_rate": 1.8564941291291883e-07, + "loss": 0.4262, + "step": 4803 + }, + { + "epoch": 4.610657705232837, + "grad_norm": 0.9061955696481128, + "learning_rate": 1.8474612131490098e-07, + "loss": 0.4249, + "step": 4804 + }, + { + "epoch": 4.611617858857417, + "grad_norm": 1.041083942578629, + "learning_rate": 1.8384499123287713e-07, + "loss": 0.4676, + "step": 4805 + }, + { + "epoch": 4.612578012481997, + "grad_norm": 1.2693179431520776, + "learning_rate": 1.8294602307135478e-07, + "loss": 0.3909, + "step": 4806 + }, + { + "epoch": 4.6135381661065775, + "grad_norm": 1.0226813030762867, + "learning_rate": 1.820492172338717e-07, + "loss": 0.4818, + "step": 4807 + }, + { + "epoch": 4.614498319731157, + "grad_norm": 1.870649909267636, + "learning_rate": 1.8115457412299363e-07, + "loss": 0.381, + "step": 4808 + }, + { + "epoch": 4.615458473355737, + "grad_norm": 0.8616460848106126, + "learning_rate": 1.802620941403166e-07, + "loss": 0.3836, + "step": 4809 + }, + { + "epoch": 4.616418626980317, + "grad_norm": 1.2083732396771987, + "learning_rate": 1.7937177768646508e-07, + "loss": 0.4336, + "step": 4810 + }, + { + "epoch": 4.617378780604897, + "grad_norm": 0.938800339545672, + "learning_rate": 1.7848362516109274e-07, + "loss": 0.4703, + "step": 4811 + }, + { + "epoch": 4.6183389342294765, + "grad_norm": 1.313447666563615, + "learning_rate": 1.775976369628829e-07, + "loss": 0.4097, + "step": 4812 + }, + { + "epoch": 4.619299087854056, + "grad_norm": 0.9220032408795396, + "learning_rate": 1.7671381348954408e-07, + "loss": 0.3955, + "step": 4813 + }, + { + "epoch": 4.620259241478637, + "grad_norm": 1.5912875171764935, + "learning_rate": 1.7583215513781725e-07, + "loss": 0.3773, + "step": 4814 + }, + { + "epoch": 4.621219395103217, + "grad_norm": 0.9212555151022918, + "learning_rate": 1.749526623034681e-07, + "loss": 0.474, + "step": 4815 + }, + { + "epoch": 4.622179548727797, + "grad_norm": 1.315833523722032, + "learning_rate": 1.7407533538129185e-07, + "loss": 0.3756, + "step": 4816 + }, + { + "epoch": 4.623139702352376, + "grad_norm": 1.2615289578029076, + "learning_rate": 1.732001747651113e-07, + "loss": 0.4456, + "step": 4817 + }, + { + "epoch": 4.624099855976956, + "grad_norm": 1.4158536370413721, + "learning_rate": 1.7232718084777666e-07, + "loss": 0.3716, + "step": 4818 + }, + { + "epoch": 4.625060009601536, + "grad_norm": 1.445241579427424, + "learning_rate": 1.7145635402116556e-07, + "loss": 0.4631, + "step": 4819 + }, + { + "epoch": 4.626020163226116, + "grad_norm": 0.9557225557144389, + "learning_rate": 1.7058769467618364e-07, + "loss": 0.4896, + "step": 4820 + }, + { + "epoch": 4.626980316850696, + "grad_norm": 1.3519843082438203, + "learning_rate": 1.697212032027612e-07, + "loss": 0.2894, + "step": 4821 + }, + { + "epoch": 4.627940470475276, + "grad_norm": 0.9224256638249925, + "learning_rate": 1.688568799898582e-07, + "loss": 0.3217, + "step": 4822 + }, + { + "epoch": 4.628900624099856, + "grad_norm": 0.9216992108440527, + "learning_rate": 1.6799472542545935e-07, + "loss": 0.3999, + "step": 4823 + }, + { + "epoch": 4.629860777724436, + "grad_norm": 0.9231941365283417, + "learning_rate": 1.671347398965767e-07, + "loss": 0.3665, + "step": 4824 + }, + { + "epoch": 4.630820931349016, + "grad_norm": 1.3808693258581342, + "learning_rate": 1.6627692378924975e-07, + "loss": 0.4512, + "step": 4825 + }, + { + "epoch": 4.631781084973595, + "grad_norm": 1.3485510873758093, + "learning_rate": 1.6542127748854108e-07, + "loss": 0.3721, + "step": 4826 + }, + { + "epoch": 4.632741238598175, + "grad_norm": 0.9918658002473777, + "learning_rate": 1.6456780137854346e-07, + "loss": 0.5436, + "step": 4827 + }, + { + "epoch": 4.633701392222756, + "grad_norm": 1.0361195140022446, + "learning_rate": 1.6371649584237093e-07, + "loss": 0.4174, + "step": 4828 + }, + { + "epoch": 4.634661545847336, + "grad_norm": 1.182004901649788, + "learning_rate": 1.6286736126216674e-07, + "loss": 0.3823, + "step": 4829 + }, + { + "epoch": 4.6356216994719155, + "grad_norm": 0.8701419208967126, + "learning_rate": 1.6202039801909763e-07, + "loss": 0.424, + "step": 4830 + }, + { + "epoch": 4.636581853096495, + "grad_norm": 1.1504272685963877, + "learning_rate": 1.6117560649335674e-07, + "loss": 0.4092, + "step": 4831 + }, + { + "epoch": 4.637542006721075, + "grad_norm": 0.9936816498598194, + "learning_rate": 1.6033298706416234e-07, + "loss": 0.6072, + "step": 4832 + }, + { + "epoch": 4.638502160345656, + "grad_norm": 0.91588079466234, + "learning_rate": 1.594925401097569e-07, + "loss": 0.416, + "step": 4833 + }, + { + "epoch": 4.639462313970236, + "grad_norm": 1.161219323651602, + "learning_rate": 1.5865426600740807e-07, + "loss": 0.462, + "step": 4834 + }, + { + "epoch": 4.640422467594815, + "grad_norm": 1.5677838114702125, + "learning_rate": 1.5781816513340764e-07, + "loss": 0.3762, + "step": 4835 + }, + { + "epoch": 4.641382621219395, + "grad_norm": 0.9077167818557987, + "learning_rate": 1.5698423786307314e-07, + "loss": 0.4921, + "step": 4836 + }, + { + "epoch": 4.642342774843975, + "grad_norm": 1.2432151317729871, + "learning_rate": 1.5615248457074516e-07, + "loss": 0.4517, + "step": 4837 + }, + { + "epoch": 4.643302928468555, + "grad_norm": 1.1282115531345462, + "learning_rate": 1.5532290562978948e-07, + "loss": 0.3731, + "step": 4838 + }, + { + "epoch": 4.6442630820931345, + "grad_norm": 1.0135208505894553, + "learning_rate": 1.5449550141259427e-07, + "loss": 0.4225, + "step": 4839 + }, + { + "epoch": 4.645223235717715, + "grad_norm": 1.2155129043556023, + "learning_rate": 1.5367027229057307e-07, + "loss": 0.5577, + "step": 4840 + }, + { + "epoch": 4.646183389342295, + "grad_norm": 1.2834642587348466, + "learning_rate": 1.5284721863416175e-07, + "loss": 0.503, + "step": 4841 + }, + { + "epoch": 4.647143542966875, + "grad_norm": 0.875103063905967, + "learning_rate": 1.5202634081282142e-07, + "loss": 0.5263, + "step": 4842 + }, + { + "epoch": 4.648103696591455, + "grad_norm": 0.8083863279715064, + "learning_rate": 1.512076391950351e-07, + "loss": 0.371, + "step": 4843 + }, + { + "epoch": 4.649063850216034, + "grad_norm": 0.9753449309385723, + "learning_rate": 1.5039111414830721e-07, + "loss": 0.4616, + "step": 4844 + }, + { + "epoch": 4.650024003840614, + "grad_norm": 1.3312206754583908, + "learning_rate": 1.4957676603917005e-07, + "loss": 0.3697, + "step": 4845 + }, + { + "epoch": 4.650984157465194, + "grad_norm": 0.8037315397522173, + "learning_rate": 1.4876459523317455e-07, + "loss": 0.4577, + "step": 4846 + }, + { + "epoch": 4.651944311089775, + "grad_norm": 1.1583517652034765, + "learning_rate": 1.479546020948952e-07, + "loss": 0.337, + "step": 4847 + }, + { + "epoch": 4.6529044647143545, + "grad_norm": 1.1034917280152576, + "learning_rate": 1.4714678698792894e-07, + "loss": 0.4576, + "step": 4848 + }, + { + "epoch": 4.653864618338934, + "grad_norm": 0.8404487653635586, + "learning_rate": 1.4634115027489627e-07, + "loss": 0.5042, + "step": 4849 + }, + { + "epoch": 4.654824771963514, + "grad_norm": 1.18548657659862, + "learning_rate": 1.4553769231743842e-07, + "loss": 0.3966, + "step": 4850 + }, + { + "epoch": 4.655784925588094, + "grad_norm": 1.055970806542243, + "learning_rate": 1.447364134762197e-07, + "loss": 0.4327, + "step": 4851 + }, + { + "epoch": 4.656745079212674, + "grad_norm": 0.9167219764458515, + "learning_rate": 1.439373141109246e-07, + "loss": 0.4236, + "step": 4852 + }, + { + "epoch": 4.6577052328372535, + "grad_norm": 1.4952176776863857, + "learning_rate": 1.4314039458026063e-07, + "loss": 0.3892, + "step": 4853 + }, + { + "epoch": 4.658665386461834, + "grad_norm": 1.4585786530087195, + "learning_rate": 1.4234565524195664e-07, + "loss": 0.4627, + "step": 4854 + }, + { + "epoch": 4.659625540086414, + "grad_norm": 1.3997699800947045, + "learning_rate": 1.4155309645276283e-07, + "loss": 0.4162, + "step": 4855 + }, + { + "epoch": 4.660585693710994, + "grad_norm": 0.9842333753736631, + "learning_rate": 1.407627185684507e-07, + "loss": 0.4124, + "step": 4856 + }, + { + "epoch": 4.6615458473355735, + "grad_norm": 0.863316922583818, + "learning_rate": 1.3997452194381034e-07, + "loss": 0.5363, + "step": 4857 + }, + { + "epoch": 4.662506000960153, + "grad_norm": 0.9762782180460298, + "learning_rate": 1.3918850693265763e-07, + "loss": 0.5241, + "step": 4858 + }, + { + "epoch": 4.663466154584734, + "grad_norm": 1.6979556539963534, + "learning_rate": 1.3840467388782418e-07, + "loss": 0.3888, + "step": 4859 + }, + { + "epoch": 4.664426308209314, + "grad_norm": 1.0081865531573144, + "learning_rate": 1.3762302316116527e-07, + "loss": 0.4033, + "step": 4860 + }, + { + "epoch": 4.665386461833894, + "grad_norm": 1.091767746541218, + "learning_rate": 1.3684355510355518e-07, + "loss": 0.3666, + "step": 4861 + }, + { + "epoch": 4.666346615458473, + "grad_norm": 0.8008406660157712, + "learning_rate": 1.3606627006488905e-07, + "loss": 0.5543, + "step": 4862 + }, + { + "epoch": 4.667306769083053, + "grad_norm": 1.3614175946932625, + "learning_rate": 1.3529116839408162e-07, + "loss": 0.373, + "step": 4863 + }, + { + "epoch": 4.668266922707633, + "grad_norm": 1.201104244046454, + "learning_rate": 1.3451825043906852e-07, + "loss": 0.4642, + "step": 4864 + }, + { + "epoch": 4.669227076332213, + "grad_norm": 1.0786220466963068, + "learning_rate": 1.337475165468033e-07, + "loss": 0.4308, + "step": 4865 + }, + { + "epoch": 4.6701872299567935, + "grad_norm": 1.1344337058104614, + "learning_rate": 1.3297896706326031e-07, + "loss": 0.4036, + "step": 4866 + }, + { + "epoch": 4.671147383581373, + "grad_norm": 1.2101706576117013, + "learning_rate": 1.3221260233343357e-07, + "loss": 0.46, + "step": 4867 + }, + { + "epoch": 4.672107537205953, + "grad_norm": 1.6304586067035027, + "learning_rate": 1.3144842270133573e-07, + "loss": 0.3969, + "step": 4868 + }, + { + "epoch": 4.673067690830533, + "grad_norm": 1.0671618880075804, + "learning_rate": 1.3068642850999957e-07, + "loss": 0.4138, + "step": 4869 + }, + { + "epoch": 4.674027844455113, + "grad_norm": 1.0223444659112357, + "learning_rate": 1.2992662010147428e-07, + "loss": 0.4489, + "step": 4870 + }, + { + "epoch": 4.6749879980796925, + "grad_norm": 1.2642173548965616, + "learning_rate": 1.2916899781683257e-07, + "loss": 0.3875, + "step": 4871 + }, + { + "epoch": 4.675948151704272, + "grad_norm": 1.1412850621951813, + "learning_rate": 1.2841356199616072e-07, + "loss": 0.4745, + "step": 4872 + }, + { + "epoch": 4.676908305328853, + "grad_norm": 1.3415299598720207, + "learning_rate": 1.276603129785664e-07, + "loss": 0.4558, + "step": 4873 + }, + { + "epoch": 4.677868458953433, + "grad_norm": 0.8547068664536178, + "learning_rate": 1.2690925110217523e-07, + "loss": 0.4189, + "step": 4874 + }, + { + "epoch": 4.6788286125780125, + "grad_norm": 0.970876090683359, + "learning_rate": 1.2616037670413084e-07, + "loss": 0.4404, + "step": 4875 + }, + { + "epoch": 4.679788766202592, + "grad_norm": 1.3348221156706659, + "learning_rate": 1.254136901205949e-07, + "loss": 0.3674, + "step": 4876 + }, + { + "epoch": 4.680748919827172, + "grad_norm": 1.0607047447901858, + "learning_rate": 1.2466919168674763e-07, + "loss": 0.4402, + "step": 4877 + }, + { + "epoch": 4.681709073451752, + "grad_norm": 1.3850927877470836, + "learning_rate": 1.2392688173678503e-07, + "loss": 0.377, + "step": 4878 + }, + { + "epoch": 4.682669227076333, + "grad_norm": 1.146457920670871, + "learning_rate": 1.2318676060392333e-07, + "loss": 0.4765, + "step": 4879 + }, + { + "epoch": 4.683629380700912, + "grad_norm": 1.3187711202669654, + "learning_rate": 1.2244882862039398e-07, + "loss": 0.3573, + "step": 4880 + }, + { + "epoch": 4.684589534325492, + "grad_norm": 1.0748003644300437, + "learning_rate": 1.217130861174476e-07, + "loss": 0.5518, + "step": 4881 + }, + { + "epoch": 4.685549687950072, + "grad_norm": 0.9285725553402598, + "learning_rate": 1.209795334253516e-07, + "loss": 0.478, + "step": 4882 + }, + { + "epoch": 4.686509841574652, + "grad_norm": 1.3821853568376987, + "learning_rate": 1.2024817087338826e-07, + "loss": 0.3444, + "step": 4883 + }, + { + "epoch": 4.687469995199232, + "grad_norm": 1.1346742359195487, + "learning_rate": 1.1951899878985983e-07, + "loss": 0.3314, + "step": 4884 + }, + { + "epoch": 4.688430148823812, + "grad_norm": 1.1054396444704837, + "learning_rate": 1.1879201750208291e-07, + "loss": 0.4511, + "step": 4885 + }, + { + "epoch": 4.689390302448392, + "grad_norm": 0.8909019999832792, + "learning_rate": 1.1806722733639253e-07, + "loss": 0.5304, + "step": 4886 + }, + { + "epoch": 4.690350456072972, + "grad_norm": 1.1991086945341605, + "learning_rate": 1.17344628618139e-07, + "loss": 0.4055, + "step": 4887 + }, + { + "epoch": 4.691310609697552, + "grad_norm": 0.888618511391377, + "learning_rate": 1.1662422167168841e-07, + "loss": 0.4966, + "step": 4888 + }, + { + "epoch": 4.692270763322131, + "grad_norm": 1.03938945352285, + "learning_rate": 1.159060068204243e-07, + "loss": 0.4609, + "step": 4889 + }, + { + "epoch": 4.693230916946711, + "grad_norm": 1.5886976054096131, + "learning_rate": 1.1518998438674656e-07, + "loss": 0.3648, + "step": 4890 + }, + { + "epoch": 4.694191070571291, + "grad_norm": 1.0111931849689184, + "learning_rate": 1.1447615469206863e-07, + "loss": 0.5052, + "step": 4891 + }, + { + "epoch": 4.695151224195872, + "grad_norm": 0.9451034878617085, + "learning_rate": 1.1376451805682143e-07, + "loss": 0.4494, + "step": 4892 + }, + { + "epoch": 4.6961113778204515, + "grad_norm": 1.0684130717439, + "learning_rate": 1.1305507480045108e-07, + "loss": 0.4941, + "step": 4893 + }, + { + "epoch": 4.697071531445031, + "grad_norm": 1.0331287204273547, + "learning_rate": 1.1234782524141896e-07, + "loss": 0.5403, + "step": 4894 + }, + { + "epoch": 4.698031685069611, + "grad_norm": 0.9787382958557056, + "learning_rate": 1.1164276969720278e-07, + "loss": 0.4327, + "step": 4895 + }, + { + "epoch": 4.698991838694191, + "grad_norm": 0.7971038099269354, + "learning_rate": 1.109399084842927e-07, + "loss": 0.4174, + "step": 4896 + }, + { + "epoch": 4.699951992318771, + "grad_norm": 1.0189260596115977, + "learning_rate": 1.1023924191819691e-07, + "loss": 0.4767, + "step": 4897 + }, + { + "epoch": 4.7009121459433505, + "grad_norm": 0.8119306713496375, + "learning_rate": 1.0954077031343658e-07, + "loss": 0.4478, + "step": 4898 + }, + { + "epoch": 4.701872299567931, + "grad_norm": 0.9838075867737415, + "learning_rate": 1.0884449398354868e-07, + "loss": 0.4201, + "step": 4899 + }, + { + "epoch": 4.702832453192511, + "grad_norm": 1.606158635230087, + "learning_rate": 1.0815041324108378e-07, + "loss": 0.4645, + "step": 4900 + }, + { + "epoch": 4.703792606817091, + "grad_norm": 1.2063985629173526, + "learning_rate": 1.0745852839760706e-07, + "loss": 0.3579, + "step": 4901 + }, + { + "epoch": 4.704752760441671, + "grad_norm": 0.8553193293134254, + "learning_rate": 1.0676883976369956e-07, + "loss": 0.4549, + "step": 4902 + }, + { + "epoch": 4.70571291406625, + "grad_norm": 1.002553399336827, + "learning_rate": 1.0608134764895306e-07, + "loss": 0.4566, + "step": 4903 + }, + { + "epoch": 4.70667306769083, + "grad_norm": 1.1243789394374941, + "learning_rate": 1.053960523619768e-07, + "loss": 0.3638, + "step": 4904 + }, + { + "epoch": 4.707633221315411, + "grad_norm": 1.2881083832451525, + "learning_rate": 1.0471295421039251e-07, + "loss": 0.4376, + "step": 4905 + }, + { + "epoch": 4.708593374939991, + "grad_norm": 0.9839190906016867, + "learning_rate": 1.040320535008349e-07, + "loss": 0.4521, + "step": 4906 + }, + { + "epoch": 4.70955352856457, + "grad_norm": 1.5004863605052488, + "learning_rate": 1.0335335053895391e-07, + "loss": 0.4402, + "step": 4907 + }, + { + "epoch": 4.71051368218915, + "grad_norm": 1.006124507240295, + "learning_rate": 1.0267684562941194e-07, + "loss": 0.3408, + "step": 4908 + }, + { + "epoch": 4.71147383581373, + "grad_norm": 1.1130247517133098, + "learning_rate": 1.0200253907588387e-07, + "loss": 0.4047, + "step": 4909 + }, + { + "epoch": 4.71243398943831, + "grad_norm": 1.2325609428027149, + "learning_rate": 1.0133043118105923e-07, + "loss": 0.4371, + "step": 4910 + }, + { + "epoch": 4.7133941430628905, + "grad_norm": 1.4056592768793215, + "learning_rate": 1.0066052224664058e-07, + "loss": 0.4301, + "step": 4911 + }, + { + "epoch": 4.71435429668747, + "grad_norm": 1.3159180364916698, + "learning_rate": 9.999281257334181e-08, + "loss": 0.4366, + "step": 4912 + }, + { + "epoch": 4.71531445031205, + "grad_norm": 1.4518304434272327, + "learning_rate": 9.932730246089262e-08, + "loss": 0.4812, + "step": 4913 + }, + { + "epoch": 4.71627460393663, + "grad_norm": 1.6672745203186197, + "learning_rate": 9.86639922080307e-08, + "loss": 0.4855, + "step": 4914 + }, + { + "epoch": 4.71723475756121, + "grad_norm": 0.9039540229330392, + "learning_rate": 9.800288211251119e-08, + "loss": 0.462, + "step": 4915 + }, + { + "epoch": 4.7181949111857895, + "grad_norm": 1.1369334787339889, + "learning_rate": 9.734397247109784e-08, + "loss": 0.4026, + "step": 4916 + }, + { + "epoch": 4.719155064810369, + "grad_norm": 0.8818340451726859, + "learning_rate": 9.668726357956903e-08, + "loss": 0.4413, + "step": 4917 + }, + { + "epoch": 4.72011521843495, + "grad_norm": 1.4896273099509099, + "learning_rate": 9.603275573271397e-08, + "loss": 0.4953, + "step": 4918 + }, + { + "epoch": 4.72107537205953, + "grad_norm": 1.0628416430946594, + "learning_rate": 9.538044922433432e-08, + "loss": 0.4605, + "step": 4919 + }, + { + "epoch": 4.72203552568411, + "grad_norm": 1.0993876630712076, + "learning_rate": 9.47303443472436e-08, + "loss": 0.3838, + "step": 4920 + }, + { + "epoch": 4.722995679308689, + "grad_norm": 1.168472568639651, + "learning_rate": 9.408244139326728e-08, + "loss": 0.4395, + "step": 4921 + }, + { + "epoch": 4.723955832933269, + "grad_norm": 1.1976832403588442, + "learning_rate": 9.34367406532416e-08, + "loss": 0.3265, + "step": 4922 + }, + { + "epoch": 4.724915986557849, + "grad_norm": 0.9147656733632449, + "learning_rate": 9.279324241701415e-08, + "loss": 0.5055, + "step": 4923 + }, + { + "epoch": 4.725876140182429, + "grad_norm": 1.1868320539821027, + "learning_rate": 9.215194697344554e-08, + "loss": 0.4473, + "step": 4924 + }, + { + "epoch": 4.726836293807009, + "grad_norm": 1.8420799450510248, + "learning_rate": 9.151285461040604e-08, + "loss": 0.3488, + "step": 4925 + }, + { + "epoch": 4.727796447431589, + "grad_norm": 1.2304430967818216, + "learning_rate": 9.08759656147773e-08, + "loss": 0.3462, + "step": 4926 + }, + { + "epoch": 4.728756601056169, + "grad_norm": 0.9520488195701418, + "learning_rate": 9.024128027245171e-08, + "loss": 0.5115, + "step": 4927 + }, + { + "epoch": 4.729716754680749, + "grad_norm": 0.7815379300936841, + "learning_rate": 8.960879886833307e-08, + "loss": 0.4298, + "step": 4928 + }, + { + "epoch": 4.730676908305329, + "grad_norm": 1.0394315489433514, + "learning_rate": 8.897852168633537e-08, + "loss": 0.6247, + "step": 4929 + }, + { + "epoch": 4.731637061929908, + "grad_norm": 1.1020238704377705, + "learning_rate": 8.835044900938339e-08, + "loss": 0.3987, + "step": 4930 + }, + { + "epoch": 4.732597215554489, + "grad_norm": 1.2263463836720976, + "learning_rate": 8.772458111941273e-08, + "loss": 0.4371, + "step": 4931 + }, + { + "epoch": 4.733557369179069, + "grad_norm": 1.3328171761452732, + "learning_rate": 8.7100918297367e-08, + "loss": 0.3241, + "step": 4932 + }, + { + "epoch": 4.734517522803649, + "grad_norm": 1.2551802685050564, + "learning_rate": 8.647946082320446e-08, + "loss": 0.5018, + "step": 4933 + }, + { + "epoch": 4.7354776764282285, + "grad_norm": 0.9142369627181196, + "learning_rate": 8.586020897588864e-08, + "loss": 0.3876, + "step": 4934 + }, + { + "epoch": 4.736437830052808, + "grad_norm": 1.1762289843339626, + "learning_rate": 8.524316303339609e-08, + "loss": 0.4208, + "step": 4935 + }, + { + "epoch": 4.737397983677388, + "grad_norm": 1.0550815448101691, + "learning_rate": 8.462832327271131e-08, + "loss": 0.4866, + "step": 4936 + }, + { + "epoch": 4.738358137301969, + "grad_norm": 1.058668993144802, + "learning_rate": 8.401568996983022e-08, + "loss": 0.4792, + "step": 4937 + }, + { + "epoch": 4.7393182909265485, + "grad_norm": 1.2542659226887585, + "learning_rate": 8.340526339975674e-08, + "loss": 0.3754, + "step": 4938 + }, + { + "epoch": 4.740278444551128, + "grad_norm": 0.9819182622733525, + "learning_rate": 8.279704383650555e-08, + "loss": 0.4479, + "step": 4939 + }, + { + "epoch": 4.741238598175708, + "grad_norm": 1.3028202373874764, + "learning_rate": 8.219103155309938e-08, + "loss": 0.3555, + "step": 4940 + }, + { + "epoch": 4.742198751800288, + "grad_norm": 0.8660422289707396, + "learning_rate": 8.158722682157005e-08, + "loss": 0.4085, + "step": 4941 + }, + { + "epoch": 4.743158905424868, + "grad_norm": 1.3875041639611745, + "learning_rate": 8.098562991296022e-08, + "loss": 0.4473, + "step": 4942 + }, + { + "epoch": 4.7441190590494475, + "grad_norm": 0.9151113514826871, + "learning_rate": 8.03862410973194e-08, + "loss": 0.484, + "step": 4943 + }, + { + "epoch": 4.745079212674028, + "grad_norm": 0.7596448931217988, + "learning_rate": 7.978906064370739e-08, + "loss": 0.4297, + "step": 4944 + }, + { + "epoch": 4.746039366298608, + "grad_norm": 1.2258444792057024, + "learning_rate": 7.919408882019142e-08, + "loss": 0.4487, + "step": 4945 + }, + { + "epoch": 4.746999519923188, + "grad_norm": 0.793912596275803, + "learning_rate": 7.86013258938495e-08, + "loss": 0.4838, + "step": 4946 + }, + { + "epoch": 4.747959673547768, + "grad_norm": 1.4583786481432341, + "learning_rate": 7.801077213076491e-08, + "loss": 0.4148, + "step": 4947 + }, + { + "epoch": 4.748919827172347, + "grad_norm": 1.113940011281248, + "learning_rate": 7.742242779603115e-08, + "loss": 0.4067, + "step": 4948 + }, + { + "epoch": 4.749879980796927, + "grad_norm": 1.4092380218749927, + "learning_rate": 7.683629315375086e-08, + "loss": 0.4295, + "step": 4949 + }, + { + "epoch": 4.750840134421507, + "grad_norm": 0.9293678014901214, + "learning_rate": 7.625236846703243e-08, + "loss": 0.43, + "step": 4950 + }, + { + "epoch": 4.751800288046088, + "grad_norm": 1.6534193883886423, + "learning_rate": 7.5670653997994e-08, + "loss": 0.3444, + "step": 4951 + }, + { + "epoch": 4.7527604416706675, + "grad_norm": 0.8622578441674419, + "learning_rate": 7.509115000776112e-08, + "loss": 0.4508, + "step": 4952 + }, + { + "epoch": 4.753720595295247, + "grad_norm": 1.4874994288456962, + "learning_rate": 7.451385675646627e-08, + "loss": 0.4145, + "step": 4953 + }, + { + "epoch": 4.754680748919827, + "grad_norm": 1.171764515099023, + "learning_rate": 7.393877450325105e-08, + "loss": 0.3923, + "step": 4954 + }, + { + "epoch": 4.755640902544407, + "grad_norm": 1.2786274251313396, + "learning_rate": 7.336590350626282e-08, + "loss": 0.4632, + "step": 4955 + }, + { + "epoch": 4.756601056168987, + "grad_norm": 1.1696736479222372, + "learning_rate": 7.279524402265758e-08, + "loss": 0.3443, + "step": 4956 + }, + { + "epoch": 4.757561209793567, + "grad_norm": 0.9854702719230389, + "learning_rate": 7.222679630859875e-08, + "loss": 0.4123, + "step": 4957 + }, + { + "epoch": 4.758521363418147, + "grad_norm": 0.7923822962355402, + "learning_rate": 7.166056061925609e-08, + "loss": 0.5323, + "step": 4958 + }, + { + "epoch": 4.759481517042727, + "grad_norm": 0.7356812366740414, + "learning_rate": 7.109653720880682e-08, + "loss": 0.3986, + "step": 4959 + }, + { + "epoch": 4.760441670667307, + "grad_norm": 1.1674277613161905, + "learning_rate": 7.053472633043457e-08, + "loss": 0.437, + "step": 4960 + }, + { + "epoch": 4.7614018242918865, + "grad_norm": 1.3706100197867934, + "learning_rate": 6.99751282363309e-08, + "loss": 0.3233, + "step": 4961 + }, + { + "epoch": 4.762361977916466, + "grad_norm": 1.1604441130532235, + "learning_rate": 6.941774317769268e-08, + "loss": 0.4747, + "step": 4962 + }, + { + "epoch": 4.763322131541047, + "grad_norm": 1.3250108388330313, + "learning_rate": 6.886257140472475e-08, + "loss": 0.5501, + "step": 4963 + }, + { + "epoch": 4.764282285165627, + "grad_norm": 1.0212396586151098, + "learning_rate": 6.830961316663775e-08, + "loss": 0.4548, + "step": 4964 + }, + { + "epoch": 4.765242438790207, + "grad_norm": 1.1639294654175862, + "learning_rate": 6.775886871164872e-08, + "loss": 0.3857, + "step": 4965 + }, + { + "epoch": 4.766202592414786, + "grad_norm": 0.8051939909618316, + "learning_rate": 6.721033828698098e-08, + "loss": 0.3928, + "step": 4966 + }, + { + "epoch": 4.767162746039366, + "grad_norm": 1.459218130958606, + "learning_rate": 6.666402213886313e-08, + "loss": 0.3408, + "step": 4967 + }, + { + "epoch": 4.768122899663946, + "grad_norm": 1.3492964232221831, + "learning_rate": 6.61199205125318e-08, + "loss": 0.3791, + "step": 4968 + }, + { + "epoch": 4.769083053288526, + "grad_norm": 0.8929234512957442, + "learning_rate": 6.557803365222714e-08, + "loss": 0.4892, + "step": 4969 + }, + { + "epoch": 4.7700432069131065, + "grad_norm": 0.8843561755935987, + "learning_rate": 6.503836180119849e-08, + "loss": 0.4556, + "step": 4970 + }, + { + "epoch": 4.771003360537686, + "grad_norm": 0.9240795086726811, + "learning_rate": 6.45009052016965e-08, + "loss": 0.4414, + "step": 4971 + }, + { + "epoch": 4.771963514162266, + "grad_norm": 0.6979346576133246, + "learning_rate": 6.396566409498039e-08, + "loss": 0.4472, + "step": 4972 + }, + { + "epoch": 4.772923667786846, + "grad_norm": 1.0614362995003612, + "learning_rate": 6.343263872131467e-08, + "loss": 0.3583, + "step": 4973 + }, + { + "epoch": 4.773883821411426, + "grad_norm": 0.9563491793609955, + "learning_rate": 6.290182931996846e-08, + "loss": 0.466, + "step": 4974 + }, + { + "epoch": 4.774843975036005, + "grad_norm": 0.8391276852505679, + "learning_rate": 6.237323612921675e-08, + "loss": 0.4898, + "step": 4975 + }, + { + "epoch": 4.775804128660585, + "grad_norm": 0.9597090155992283, + "learning_rate": 6.184685938633861e-08, + "loss": 0.398, + "step": 4976 + }, + { + "epoch": 4.776764282285166, + "grad_norm": 0.971137486645396, + "learning_rate": 6.132269932761947e-08, + "loss": 0.4431, + "step": 4977 + }, + { + "epoch": 4.777724435909746, + "grad_norm": 1.267122668625548, + "learning_rate": 6.080075618834891e-08, + "loss": 0.5065, + "step": 4978 + }, + { + "epoch": 4.7786845895343255, + "grad_norm": 0.9507539850267844, + "learning_rate": 6.028103020282228e-08, + "loss": 0.4919, + "step": 4979 + }, + { + "epoch": 4.779644743158905, + "grad_norm": 1.0701923975702317, + "learning_rate": 5.976352160433796e-08, + "loss": 0.4885, + "step": 4980 + }, + { + "epoch": 4.780604896783485, + "grad_norm": 1.1551069810306696, + "learning_rate": 5.9248230625200686e-08, + "loss": 0.5171, + "step": 4981 + }, + { + "epoch": 4.781565050408066, + "grad_norm": 1.412241643869421, + "learning_rate": 5.873515749671932e-08, + "loss": 0.3582, + "step": 4982 + }, + { + "epoch": 4.782525204032646, + "grad_norm": 0.8081650775943622, + "learning_rate": 5.822430244920685e-08, + "loss": 0.4219, + "step": 4983 + }, + { + "epoch": 4.783485357657225, + "grad_norm": 1.1116610362427617, + "learning_rate": 5.771566571197984e-08, + "loss": 0.4564, + "step": 4984 + }, + { + "epoch": 4.784445511281805, + "grad_norm": 1.0911799731899194, + "learning_rate": 5.72092475133601e-08, + "loss": 0.3598, + "step": 4985 + }, + { + "epoch": 4.785405664906385, + "grad_norm": 1.2795568684332475, + "learning_rate": 5.670504808067412e-08, + "loss": 0.4882, + "step": 4986 + }, + { + "epoch": 4.786365818530965, + "grad_norm": 1.2157900267444886, + "learning_rate": 5.620306764025085e-08, + "loss": 0.378, + "step": 4987 + }, + { + "epoch": 4.787325972155545, + "grad_norm": 1.1693247557414115, + "learning_rate": 5.57033064174245e-08, + "loss": 0.5115, + "step": 4988 + }, + { + "epoch": 4.788286125780125, + "grad_norm": 1.3610464976675363, + "learning_rate": 5.5205764636532266e-08, + "loss": 0.441, + "step": 4989 + }, + { + "epoch": 4.789246279404705, + "grad_norm": 1.1057349733443953, + "learning_rate": 5.471044252091551e-08, + "loss": 0.4532, + "step": 4990 + }, + { + "epoch": 4.790206433029285, + "grad_norm": 1.380890432768807, + "learning_rate": 5.421734029291858e-08, + "loss": 0.4394, + "step": 4991 + }, + { + "epoch": 4.791166586653865, + "grad_norm": 1.6010994351481194, + "learning_rate": 5.372645817389055e-08, + "loss": 0.3936, + "step": 4992 + }, + { + "epoch": 4.792126740278444, + "grad_norm": 1.0701981759846952, + "learning_rate": 5.323779638418236e-08, + "loss": 0.4799, + "step": 4993 + }, + { + "epoch": 4.793086893903024, + "grad_norm": 1.455181412551221, + "learning_rate": 5.275135514314911e-08, + "loss": 0.555, + "step": 4994 + }, + { + "epoch": 4.794047047527604, + "grad_norm": 1.27645281125826, + "learning_rate": 5.226713466915001e-08, + "loss": 0.3923, + "step": 4995 + }, + { + "epoch": 4.795007201152185, + "grad_norm": 1.4609538739748678, + "learning_rate": 5.1785135179545643e-08, + "loss": 0.355, + "step": 4996 + }, + { + "epoch": 4.7959673547767645, + "grad_norm": 1.0796528565548074, + "learning_rate": 5.130535689070071e-08, + "loss": 0.5151, + "step": 4997 + }, + { + "epoch": 4.796927508401344, + "grad_norm": 1.0709646652401668, + "learning_rate": 5.0827800017982376e-08, + "loss": 0.4795, + "step": 4998 + }, + { + "epoch": 4.797887662025924, + "grad_norm": 2.156807557723183, + "learning_rate": 5.035246477576083e-08, + "loss": 0.3752, + "step": 4999 + }, + { + "epoch": 4.798847815650504, + "grad_norm": 0.9900267881433429, + "learning_rate": 4.9879351377408734e-08, + "loss": 0.4411, + "step": 5000 + }, + { + "epoch": 4.799807969275084, + "grad_norm": 1.3097281496901851, + "learning_rate": 4.940846003530231e-08, + "loss": 0.3953, + "step": 5001 + }, + { + "epoch": 4.8007681228996635, + "grad_norm": 1.5642871992519123, + "learning_rate": 4.893979096081969e-08, + "loss": 0.4368, + "step": 5002 + }, + { + "epoch": 4.801728276524244, + "grad_norm": 1.0188868463034577, + "learning_rate": 4.847334436434037e-08, + "loss": 0.4169, + "step": 5003 + }, + { + "epoch": 4.802688430148824, + "grad_norm": 1.0283832936470796, + "learning_rate": 4.800912045524797e-08, + "loss": 0.5007, + "step": 5004 + }, + { + "epoch": 4.803648583773404, + "grad_norm": 1.5029285350762907, + "learning_rate": 4.7547119441927466e-08, + "loss": 0.5229, + "step": 5005 + }, + { + "epoch": 4.8046087373979836, + "grad_norm": 1.0538901984147298, + "learning_rate": 4.7087341531766864e-08, + "loss": 0.3801, + "step": 5006 + }, + { + "epoch": 4.805568891022563, + "grad_norm": 1.324369478183412, + "learning_rate": 4.662978693115439e-08, + "loss": 0.4654, + "step": 5007 + }, + { + "epoch": 4.806529044647144, + "grad_norm": 1.333077486173929, + "learning_rate": 4.617445584548186e-08, + "loss": 0.4119, + "step": 5008 + }, + { + "epoch": 4.807489198271724, + "grad_norm": 1.0225880236638505, + "learning_rate": 4.5721348479143e-08, + "loss": 0.3842, + "step": 5009 + }, + { + "epoch": 4.808449351896304, + "grad_norm": 1.4060556577339138, + "learning_rate": 4.527046503553234e-08, + "loss": 0.4361, + "step": 5010 + }, + { + "epoch": 4.809409505520883, + "grad_norm": 1.1957036171980913, + "learning_rate": 4.4821805717046864e-08, + "loss": 0.501, + "step": 5011 + }, + { + "epoch": 4.810369659145463, + "grad_norm": 0.9787724410074183, + "learning_rate": 4.437537072508491e-08, + "loss": 0.3446, + "step": 5012 + }, + { + "epoch": 4.811329812770043, + "grad_norm": 0.996299467897215, + "learning_rate": 4.393116026004618e-08, + "loss": 0.4115, + "step": 5013 + }, + { + "epoch": 4.812289966394623, + "grad_norm": 1.0115549019320391, + "learning_rate": 4.3489174521332833e-08, + "loss": 0.3722, + "step": 5014 + }, + { + "epoch": 4.8132501200192035, + "grad_norm": 0.9126117651097011, + "learning_rate": 4.304941370734616e-08, + "loss": 0.5471, + "step": 5015 + }, + { + "epoch": 4.814210273643783, + "grad_norm": 0.7846085957859918, + "learning_rate": 4.261187801549105e-08, + "loss": 0.4737, + "step": 5016 + }, + { + "epoch": 4.815170427268363, + "grad_norm": 1.5608139318157948, + "learning_rate": 4.217656764217204e-08, + "loss": 0.3668, + "step": 5017 + }, + { + "epoch": 4.816130580892943, + "grad_norm": 1.099803488631457, + "learning_rate": 4.1743482782796166e-08, + "loss": 0.4644, + "step": 5018 + }, + { + "epoch": 4.817090734517523, + "grad_norm": 0.8158778432417666, + "learning_rate": 4.1312623631770134e-08, + "loss": 0.4825, + "step": 5019 + }, + { + "epoch": 4.8180508881421025, + "grad_norm": 0.9106789335231578, + "learning_rate": 4.08839903825009e-08, + "loss": 0.4384, + "step": 5020 + }, + { + "epoch": 4.819011041766682, + "grad_norm": 1.0926509474000448, + "learning_rate": 4.045758322739957e-08, + "loss": 0.4491, + "step": 5021 + }, + { + "epoch": 4.819971195391263, + "grad_norm": 1.5911219887566954, + "learning_rate": 4.003340235787412e-08, + "loss": 0.4435, + "step": 5022 + }, + { + "epoch": 4.820931349015843, + "grad_norm": 1.1856930480347105, + "learning_rate": 3.961144796433502e-08, + "loss": 0.4514, + "step": 5023 + }, + { + "epoch": 4.8218915026404225, + "grad_norm": 1.1726696751660066, + "learning_rate": 3.919172023619355e-08, + "loss": 0.4326, + "step": 5024 + }, + { + "epoch": 4.822851656265002, + "grad_norm": 1.734404149191527, + "learning_rate": 3.8774219361860655e-08, + "loss": 0.3781, + "step": 5025 + }, + { + "epoch": 4.823811809889582, + "grad_norm": 0.9805679935983692, + "learning_rate": 3.835894552874808e-08, + "loss": 0.4952, + "step": 5026 + }, + { + "epoch": 4.824771963514162, + "grad_norm": 0.7514170971102834, + "learning_rate": 3.794589892326839e-08, + "loss": 0.3744, + "step": 5027 + }, + { + "epoch": 4.825732117138742, + "grad_norm": 1.445732296687345, + "learning_rate": 3.75350797308327e-08, + "loss": 0.3805, + "step": 5028 + }, + { + "epoch": 4.826692270763322, + "grad_norm": 1.0839632046033663, + "learning_rate": 3.712648813585407e-08, + "loss": 0.4165, + "step": 5029 + }, + { + "epoch": 4.827652424387902, + "grad_norm": 0.9701534195938427, + "learning_rate": 3.672012432174465e-08, + "loss": 0.3739, + "step": 5030 + }, + { + "epoch": 4.828612578012482, + "grad_norm": 0.8296847653170805, + "learning_rate": 3.631598847091688e-08, + "loss": 0.4379, + "step": 5031 + }, + { + "epoch": 4.829572731637062, + "grad_norm": 1.0767524351925464, + "learning_rate": 3.591408076478342e-08, + "loss": 0.357, + "step": 5032 + }, + { + "epoch": 4.830532885261642, + "grad_norm": 1.1096053501471805, + "learning_rate": 3.5514401383755505e-08, + "loss": 0.4687, + "step": 5033 + }, + { + "epoch": 4.831493038886222, + "grad_norm": 1.4866482743973553, + "learning_rate": 3.511695050724573e-08, + "loss": 0.4369, + "step": 5034 + }, + { + "epoch": 4.832453192510802, + "grad_norm": 0.8958708445370497, + "learning_rate": 3.4721728313664715e-08, + "loss": 0.4407, + "step": 5035 + }, + { + "epoch": 4.833413346135382, + "grad_norm": 0.9294298654858368, + "learning_rate": 3.432873498042444e-08, + "loss": 0.5196, + "step": 5036 + }, + { + "epoch": 4.834373499759962, + "grad_norm": 0.9348920356925896, + "learning_rate": 3.3937970683934893e-08, + "loss": 0.4301, + "step": 5037 + }, + { + "epoch": 4.8353336533845415, + "grad_norm": 1.2536392714464282, + "learning_rate": 3.354943559960577e-08, + "loss": 0.4297, + "step": 5038 + }, + { + "epoch": 4.836293807009121, + "grad_norm": 0.8043557599899609, + "learning_rate": 3.3163129901847e-08, + "loss": 0.4234, + "step": 5039 + }, + { + "epoch": 4.837253960633701, + "grad_norm": 1.4866701324310767, + "learning_rate": 3.277905376406654e-08, + "loss": 0.4822, + "step": 5040 + }, + { + "epoch": 4.838214114258282, + "grad_norm": 1.0122077616772496, + "learning_rate": 3.2397207358672024e-08, + "loss": 0.4176, + "step": 5041 + }, + { + "epoch": 4.8391742678828615, + "grad_norm": 1.2445365938766917, + "learning_rate": 3.201759085707079e-08, + "loss": 0.4098, + "step": 5042 + }, + { + "epoch": 4.840134421507441, + "grad_norm": 0.9607029819505497, + "learning_rate": 3.164020442966764e-08, + "loss": 0.449, + "step": 5043 + }, + { + "epoch": 4.841094575132021, + "grad_norm": 1.0185484393961937, + "learning_rate": 3.126504824586818e-08, + "loss": 0.3862, + "step": 5044 + }, + { + "epoch": 4.842054728756601, + "grad_norm": 0.8549733541522867, + "learning_rate": 3.0892122474076025e-08, + "loss": 0.4063, + "step": 5045 + }, + { + "epoch": 4.843014882381181, + "grad_norm": 1.0335421384351278, + "learning_rate": 3.0521427281693386e-08, + "loss": 0.3933, + "step": 5046 + }, + { + "epoch": 4.8439750360057605, + "grad_norm": 1.2391368799898677, + "learning_rate": 3.015296283512048e-08, + "loss": 0.3339, + "step": 5047 + }, + { + "epoch": 4.844935189630341, + "grad_norm": 1.2509800578794419, + "learning_rate": 2.9786729299758343e-08, + "loss": 0.4457, + "step": 5048 + }, + { + "epoch": 4.845895343254921, + "grad_norm": 0.9671288407106493, + "learning_rate": 2.942272684000491e-08, + "loss": 0.54, + "step": 5049 + }, + { + "epoch": 4.846855496879501, + "grad_norm": 1.1520318135178587, + "learning_rate": 2.9060955619256702e-08, + "loss": 0.372, + "step": 5050 + }, + { + "epoch": 4.847815650504081, + "grad_norm": 0.9319144448116563, + "learning_rate": 2.8701415799908815e-08, + "loss": 0.4941, + "step": 5051 + }, + { + "epoch": 4.84877580412866, + "grad_norm": 1.070882565350098, + "learning_rate": 2.8344107543356035e-08, + "loss": 0.565, + "step": 5052 + }, + { + "epoch": 4.84973595775324, + "grad_norm": 1.164358503633835, + "learning_rate": 2.7989031009988398e-08, + "loss": 0.4455, + "step": 5053 + }, + { + "epoch": 4.85069611137782, + "grad_norm": 0.9576045972050365, + "learning_rate": 2.763618635919729e-08, + "loss": 0.4493, + "step": 5054 + }, + { + "epoch": 4.851656265002401, + "grad_norm": 0.9859052075934536, + "learning_rate": 2.7285573749370465e-08, + "loss": 0.4165, + "step": 5055 + }, + { + "epoch": 4.8526164186269805, + "grad_norm": 0.9759591952474944, + "learning_rate": 2.693719333789424e-08, + "loss": 0.3959, + "step": 5056 + }, + { + "epoch": 4.85357657225156, + "grad_norm": 1.6916228334762529, + "learning_rate": 2.6591045281152418e-08, + "loss": 0.4815, + "step": 5057 + }, + { + "epoch": 4.85453672587614, + "grad_norm": 1.1637623283854825, + "learning_rate": 2.6247129734528474e-08, + "loss": 0.4088, + "step": 5058 + }, + { + "epoch": 4.85549687950072, + "grad_norm": 0.9643874496992656, + "learning_rate": 2.5905446852400594e-08, + "loss": 0.5052, + "step": 5059 + }, + { + "epoch": 4.8564570331253005, + "grad_norm": 1.4048550230246748, + "learning_rate": 2.5565996788148306e-08, + "loss": 0.427, + "step": 5060 + }, + { + "epoch": 4.85741718674988, + "grad_norm": 0.9317953985958425, + "learning_rate": 2.5228779694145832e-08, + "loss": 0.3779, + "step": 5061 + }, + { + "epoch": 4.85837734037446, + "grad_norm": 0.8975106174930008, + "learning_rate": 2.4893795721766534e-08, + "loss": 0.4195, + "step": 5062 + }, + { + "epoch": 4.85933749399904, + "grad_norm": 0.7927814892674755, + "learning_rate": 2.4561045021382346e-08, + "loss": 0.485, + "step": 5063 + }, + { + "epoch": 4.86029764762362, + "grad_norm": 0.8720161667507453, + "learning_rate": 2.4230527742359344e-08, + "loss": 0.4973, + "step": 5064 + }, + { + "epoch": 4.8612578012481995, + "grad_norm": 1.2585093263141438, + "learning_rate": 2.3902244033065512e-08, + "loss": 0.4572, + "step": 5065 + }, + { + "epoch": 4.862217954872779, + "grad_norm": 1.6788789791069185, + "learning_rate": 2.3576194040861866e-08, + "loss": 0.4782, + "step": 5066 + }, + { + "epoch": 4.86317810849736, + "grad_norm": 1.414839687090021, + "learning_rate": 2.3252377912110214e-08, + "loss": 0.4326, + "step": 5067 + }, + { + "epoch": 4.86413826212194, + "grad_norm": 0.977068520548313, + "learning_rate": 2.2930795792167615e-08, + "loss": 0.3233, + "step": 5068 + }, + { + "epoch": 4.86509841574652, + "grad_norm": 0.868506054639685, + "learning_rate": 2.2611447825389154e-08, + "loss": 0.4162, + "step": 5069 + }, + { + "epoch": 4.866058569371099, + "grad_norm": 0.7750212166806832, + "learning_rate": 2.2294334155125718e-08, + "loss": 0.4173, + "step": 5070 + }, + { + "epoch": 4.867018722995679, + "grad_norm": 0.9064544188794017, + "learning_rate": 2.1979454923727327e-08, + "loss": 0.368, + "step": 5071 + }, + { + "epoch": 4.867978876620259, + "grad_norm": 0.992579494216837, + "learning_rate": 2.1666810272539806e-08, + "loss": 0.3901, + "step": 5072 + }, + { + "epoch": 4.868939030244839, + "grad_norm": 1.2247131772624031, + "learning_rate": 2.1356400341905337e-08, + "loss": 0.4961, + "step": 5073 + }, + { + "epoch": 4.8698991838694194, + "grad_norm": 1.2712259325875028, + "learning_rate": 2.1048225271164125e-08, + "loss": 0.4253, + "step": 5074 + }, + { + "epoch": 4.870859337493999, + "grad_norm": 1.1591275819276787, + "learning_rate": 2.0742285198652734e-08, + "loss": 0.4923, + "step": 5075 + }, + { + "epoch": 4.871819491118579, + "grad_norm": 0.8658085315099535, + "learning_rate": 2.0438580261704087e-08, + "loss": 0.4324, + "step": 5076 + }, + { + "epoch": 4.872779644743159, + "grad_norm": 1.039298014186302, + "learning_rate": 2.0137110596648025e-08, + "loss": 0.4165, + "step": 5077 + }, + { + "epoch": 4.873739798367739, + "grad_norm": 1.44535186697573, + "learning_rate": 1.9837876338811292e-08, + "loss": 0.3649, + "step": 5078 + }, + { + "epoch": 4.874699951992318, + "grad_norm": 1.245908551410283, + "learning_rate": 1.9540877622516995e-08, + "loss": 0.4106, + "step": 5079 + }, + { + "epoch": 4.875660105616898, + "grad_norm": 1.0781166639004742, + "learning_rate": 1.924611458108461e-08, + "loss": 0.3914, + "step": 5080 + }, + { + "epoch": 4.876620259241479, + "grad_norm": 1.4211230254085798, + "learning_rate": 1.8953587346829948e-08, + "loss": 0.383, + "step": 5081 + }, + { + "epoch": 4.877580412866059, + "grad_norm": 1.5517652180259598, + "learning_rate": 1.8663296051065762e-08, + "loss": 0.4071, + "step": 5082 + }, + { + "epoch": 4.8785405664906385, + "grad_norm": 1.2114114580023057, + "learning_rate": 1.8375240824100028e-08, + "loss": 0.378, + "step": 5083 + }, + { + "epoch": 4.879500720115218, + "grad_norm": 1.0513556944457456, + "learning_rate": 1.8089421795238205e-08, + "loss": 0.3535, + "step": 5084 + }, + { + "epoch": 4.880460873739798, + "grad_norm": 1.4968238903736093, + "learning_rate": 1.7805839092781553e-08, + "loss": 0.465, + "step": 5085 + }, + { + "epoch": 4.881421027364379, + "grad_norm": 1.1405007768093178, + "learning_rate": 1.7524492844026576e-08, + "loss": 0.4803, + "step": 5086 + }, + { + "epoch": 4.882381180988959, + "grad_norm": 1.3866638610378546, + "learning_rate": 1.724538317526725e-08, + "loss": 0.3496, + "step": 5087 + }, + { + "epoch": 4.883341334613538, + "grad_norm": 1.0187832194723374, + "learning_rate": 1.69685102117928e-08, + "loss": 0.3701, + "step": 5088 + }, + { + "epoch": 4.884301488238118, + "grad_norm": 1.4084212742864863, + "learning_rate": 1.6693874077888805e-08, + "loss": 0.438, + "step": 5089 + }, + { + "epoch": 4.885261641862698, + "grad_norm": 1.0536901913621108, + "learning_rate": 1.6421474896835544e-08, + "loss": 0.4685, + "step": 5090 + }, + { + "epoch": 4.886221795487278, + "grad_norm": 1.5334535275759493, + "learning_rate": 1.615131279091131e-08, + "loss": 0.3611, + "step": 5091 + }, + { + "epoch": 4.8871819491118575, + "grad_norm": 1.2611632826678283, + "learning_rate": 1.588338788138799e-08, + "loss": 0.577, + "step": 5092 + }, + { + "epoch": 4.888142102736438, + "grad_norm": 0.9408967358501868, + "learning_rate": 1.5617700288534377e-08, + "loss": 0.4033, + "step": 5093 + }, + { + "epoch": 4.889102256361018, + "grad_norm": 0.9182097698873428, + "learning_rate": 1.5354250131615622e-08, + "loss": 0.5375, + "step": 5094 + }, + { + "epoch": 4.890062409985598, + "grad_norm": 1.240463356630294, + "learning_rate": 1.5093037528890463e-08, + "loss": 0.3847, + "step": 5095 + }, + { + "epoch": 4.891022563610178, + "grad_norm": 1.0300359990394057, + "learning_rate": 1.4834062597615107e-08, + "loss": 0.3564, + "step": 5096 + }, + { + "epoch": 4.891982717234757, + "grad_norm": 1.1013158862073584, + "learning_rate": 1.4577325454041003e-08, + "loss": 0.4245, + "step": 5097 + }, + { + "epoch": 4.892942870859337, + "grad_norm": 1.106545797520422, + "learning_rate": 1.4322826213414298e-08, + "loss": 0.5749, + "step": 5098 + }, + { + "epoch": 4.893903024483917, + "grad_norm": 0.960040175368356, + "learning_rate": 1.4070564989976387e-08, + "loss": 0.3848, + "step": 5099 + }, + { + "epoch": 4.894863178108498, + "grad_norm": 1.3938887097504613, + "learning_rate": 1.3820541896965579e-08, + "loss": 0.4164, + "step": 5100 + }, + { + "epoch": 4.8958233317330775, + "grad_norm": 1.6583964106689468, + "learning_rate": 1.3572757046614315e-08, + "loss": 0.388, + "step": 5101 + }, + { + "epoch": 4.896783485357657, + "grad_norm": 1.3875353619009965, + "learning_rate": 1.3327210550150293e-08, + "loss": 0.4067, + "step": 5102 + }, + { + "epoch": 4.897743638982237, + "grad_norm": 0.9864938493522539, + "learning_rate": 1.308390251779701e-08, + "loss": 0.3546, + "step": 5103 + }, + { + "epoch": 4.898703792606817, + "grad_norm": 1.4634855282024979, + "learning_rate": 1.2842833058772653e-08, + "loss": 0.4595, + "step": 5104 + }, + { + "epoch": 4.899663946231397, + "grad_norm": 1.3928005048454455, + "learning_rate": 1.260400228129066e-08, + "loss": 0.32, + "step": 5105 + }, + { + "epoch": 4.9006240998559765, + "grad_norm": 1.3315193818393363, + "learning_rate": 1.2367410292560277e-08, + "loss": 0.4512, + "step": 5106 + }, + { + "epoch": 4.901584253480557, + "grad_norm": 1.402112709594031, + "learning_rate": 1.2133057198784326e-08, + "loss": 0.4613, + "step": 5107 + }, + { + "epoch": 4.902544407105137, + "grad_norm": 0.9662704664685008, + "learning_rate": 1.1900943105160878e-08, + "loss": 0.5237, + "step": 5108 + }, + { + "epoch": 4.903504560729717, + "grad_norm": 0.9720094231775382, + "learning_rate": 1.1671068115884921e-08, + "loss": 0.3733, + "step": 5109 + }, + { + "epoch": 4.9044647143542965, + "grad_norm": 1.328747938924517, + "learning_rate": 1.1443432334144466e-08, + "loss": 0.3991, + "step": 5110 + }, + { + "epoch": 4.905424867978876, + "grad_norm": 0.9644265096027246, + "learning_rate": 1.1218035862122223e-08, + "loss": 0.5053, + "step": 5111 + }, + { + "epoch": 4.906385021603457, + "grad_norm": 1.322136554202964, + "learning_rate": 1.09948788009967e-08, + "loss": 0.3923, + "step": 5112 + }, + { + "epoch": 4.907345175228037, + "grad_norm": 1.428478812006693, + "learning_rate": 1.0773961250939991e-08, + "loss": 0.4183, + "step": 5113 + }, + { + "epoch": 4.908305328852617, + "grad_norm": 1.5965621060221347, + "learning_rate": 1.0555283311121101e-08, + "loss": 0.4922, + "step": 5114 + }, + { + "epoch": 4.909265482477196, + "grad_norm": 1.1610655346822698, + "learning_rate": 1.0338845079700954e-08, + "loss": 0.4809, + "step": 5115 + }, + { + "epoch": 4.910225636101776, + "grad_norm": 0.8931306602718819, + "learning_rate": 1.0124646653836279e-08, + "loss": 0.5179, + "step": 5116 + }, + { + "epoch": 4.911185789726356, + "grad_norm": 1.1999844882912658, + "learning_rate": 9.912688129679603e-09, + "loss": 0.34, + "step": 5117 + }, + { + "epoch": 4.912145943350936, + "grad_norm": 1.524911140052869, + "learning_rate": 9.702969602375378e-09, + "loss": 0.3102, + "step": 5118 + }, + { + "epoch": 4.9131060969755165, + "grad_norm": 1.1888212322198293, + "learning_rate": 9.495491166065519e-09, + "loss": 0.4304, + "step": 5119 + }, + { + "epoch": 4.914066250600096, + "grad_norm": 0.8370339215857046, + "learning_rate": 9.290252913883856e-09, + "loss": 0.5062, + "step": 5120 + }, + { + "epoch": 4.915026404224676, + "grad_norm": 0.8838455024416435, + "learning_rate": 9.087254937960032e-09, + "loss": 0.415, + "step": 5121 + }, + { + "epoch": 4.915986557849256, + "grad_norm": 1.4903560415545014, + "learning_rate": 8.886497329417265e-09, + "loss": 0.3926, + "step": 5122 + }, + { + "epoch": 4.916946711473836, + "grad_norm": 0.9053606944330268, + "learning_rate": 8.687980178374023e-09, + "loss": 0.415, + "step": 5123 + }, + { + "epoch": 4.9179068650984155, + "grad_norm": 0.9647380161249303, + "learning_rate": 8.491703573942356e-09, + "loss": 0.4446, + "step": 5124 + }, + { + "epoch": 4.918867018722995, + "grad_norm": 1.3693617755150729, + "learning_rate": 8.29766760422901e-09, + "loss": 0.4558, + "step": 5125 + }, + { + "epoch": 4.919827172347576, + "grad_norm": 1.2179187424465863, + "learning_rate": 8.105872356333755e-09, + "loss": 0.3484, + "step": 5126 + }, + { + "epoch": 4.920787325972156, + "grad_norm": 1.5332175038998923, + "learning_rate": 7.916317916352168e-09, + "loss": 0.4337, + "step": 5127 + }, + { + "epoch": 4.9217474795967355, + "grad_norm": 1.0003918034790997, + "learning_rate": 7.729004369373405e-09, + "loss": 0.4873, + "step": 5128 + }, + { + "epoch": 4.922707633221315, + "grad_norm": 0.9205647155475724, + "learning_rate": 7.543931799479653e-09, + "loss": 0.4052, + "step": 5129 + }, + { + "epoch": 4.923667786845895, + "grad_norm": 1.4386377949692875, + "learning_rate": 7.3611002897489015e-09, + "loss": 0.4085, + "step": 5130 + }, + { + "epoch": 4.924627940470475, + "grad_norm": 1.3599067463748153, + "learning_rate": 7.180509922251611e-09, + "loss": 0.4257, + "step": 5131 + }, + { + "epoch": 4.925588094095055, + "grad_norm": 1.139163702921157, + "learning_rate": 7.002160778053491e-09, + "loss": 0.4469, + "step": 5132 + }, + { + "epoch": 4.926548247719635, + "grad_norm": 1.3645010341460393, + "learning_rate": 6.826052937212724e-09, + "loss": 0.417, + "step": 5133 + }, + { + "epoch": 4.927508401344215, + "grad_norm": 1.3643477934842472, + "learning_rate": 6.652186478783296e-09, + "loss": 0.4737, + "step": 5134 + }, + { + "epoch": 4.928468554968795, + "grad_norm": 0.8448587410954269, + "learning_rate": 6.480561480811109e-09, + "loss": 0.4151, + "step": 5135 + }, + { + "epoch": 4.929428708593375, + "grad_norm": 1.0354494543652244, + "learning_rate": 6.311178020336761e-09, + "loss": 0.4958, + "step": 5136 + }, + { + "epoch": 4.930388862217955, + "grad_norm": 1.2147326495868704, + "learning_rate": 6.14403617339554e-09, + "loss": 0.5032, + "step": 5137 + }, + { + "epoch": 4.931349015842535, + "grad_norm": 0.9696535144120022, + "learning_rate": 5.979136015015208e-09, + "loss": 0.5898, + "step": 5138 + }, + { + "epoch": 4.932309169467115, + "grad_norm": 0.9746702442415063, + "learning_rate": 5.8164776192171094e-09, + "loss": 0.398, + "step": 5139 + }, + { + "epoch": 4.933269323091695, + "grad_norm": 1.3285842986567158, + "learning_rate": 5.656061059017837e-09, + "loss": 0.4384, + "step": 5140 + }, + { + "epoch": 4.934229476716275, + "grad_norm": 1.3033312799211494, + "learning_rate": 5.4978864064259005e-09, + "loss": 0.3526, + "step": 5141 + }, + { + "epoch": 4.9351896303408544, + "grad_norm": 0.9126781389699786, + "learning_rate": 5.3419537324445044e-09, + "loss": 0.454, + "step": 5142 + }, + { + "epoch": 4.936149783965434, + "grad_norm": 1.0040080868664603, + "learning_rate": 5.18826310706988e-09, + "loss": 0.4419, + "step": 5143 + }, + { + "epoch": 4.937109937590014, + "grad_norm": 0.9796725198047096, + "learning_rate": 5.0368145992929495e-09, + "loss": 0.4626, + "step": 5144 + }, + { + "epoch": 4.938070091214595, + "grad_norm": 1.3805326857551758, + "learning_rate": 4.887608277096001e-09, + "loss": 0.3561, + "step": 5145 + }, + { + "epoch": 4.9390302448391745, + "grad_norm": 1.0428810022184316, + "learning_rate": 4.740644207457678e-09, + "loss": 0.5836, + "step": 5146 + }, + { + "epoch": 4.939990398463754, + "grad_norm": 1.238303417596389, + "learning_rate": 4.5959224563474345e-09, + "loss": 0.4241, + "step": 5147 + }, + { + "epoch": 4.940950552088334, + "grad_norm": 1.235297425078616, + "learning_rate": 4.4534430887299694e-09, + "loss": 0.4415, + "step": 5148 + }, + { + "epoch": 4.941910705712914, + "grad_norm": 1.5714181493805923, + "learning_rate": 4.3132061685619005e-09, + "loss": 0.4448, + "step": 5149 + }, + { + "epoch": 4.942870859337494, + "grad_norm": 1.4210360472425325, + "learning_rate": 4.1752117587950945e-09, + "loss": 0.3297, + "step": 5150 + }, + { + "epoch": 4.9438310129620735, + "grad_norm": 0.906595718766946, + "learning_rate": 4.0394599213733346e-09, + "loss": 0.4804, + "step": 5151 + }, + { + "epoch": 4.944791166586654, + "grad_norm": 1.6793102175034818, + "learning_rate": 3.905950717233986e-09, + "loss": 0.3784, + "step": 5152 + }, + { + "epoch": 4.945751320211234, + "grad_norm": 1.0401983747434733, + "learning_rate": 3.774684206308554e-09, + "loss": 0.3924, + "step": 5153 + }, + { + "epoch": 4.946711473835814, + "grad_norm": 1.0420523529139116, + "learning_rate": 3.6456604475199055e-09, + "loss": 0.4338, + "step": 5154 + }, + { + "epoch": 4.947671627460394, + "grad_norm": 1.0285533862088443, + "learning_rate": 3.5188794987867093e-09, + "loss": 0.5015, + "step": 5155 + }, + { + "epoch": 4.948631781084973, + "grad_norm": 1.0652557416318498, + "learning_rate": 3.394341417018998e-09, + "loss": 0.417, + "step": 5156 + }, + { + "epoch": 4.949591934709553, + "grad_norm": 0.8211314402017954, + "learning_rate": 3.272046258120942e-09, + "loss": 0.4742, + "step": 5157 + }, + { + "epoch": 4.950552088334134, + "grad_norm": 0.9730651546065674, + "learning_rate": 3.151994076988629e-09, + "loss": 0.4185, + "step": 5158 + }, + { + "epoch": 4.951512241958714, + "grad_norm": 1.550655918576398, + "learning_rate": 3.034184927513395e-09, + "loss": 0.353, + "step": 5159 + }, + { + "epoch": 4.952472395583293, + "grad_norm": 1.0009479773329741, + "learning_rate": 2.9186188625779376e-09, + "loss": 0.3225, + "step": 5160 + }, + { + "epoch": 4.953432549207873, + "grad_norm": 1.1117566628566564, + "learning_rate": 2.805295934058538e-09, + "loss": 0.3916, + "step": 5161 + }, + { + "epoch": 4.954392702832453, + "grad_norm": 0.9949207272477235, + "learning_rate": 2.694216192824506e-09, + "loss": 0.4388, + "step": 5162 + }, + { + "epoch": 4.955352856457033, + "grad_norm": 0.8870075449913225, + "learning_rate": 2.5853796887387316e-09, + "loss": 0.4247, + "step": 5163 + }, + { + "epoch": 4.9563130100816135, + "grad_norm": 1.1345176979813556, + "learning_rate": 2.47878647065658e-09, + "loss": 0.3433, + "step": 5164 + }, + { + "epoch": 4.957273163706193, + "grad_norm": 1.0651711888218474, + "learning_rate": 2.374436586426443e-09, + "loss": 0.4421, + "step": 5165 + }, + { + "epoch": 4.958233317330773, + "grad_norm": 1.2129388831427408, + "learning_rate": 2.27233008289085e-09, + "loss": 0.4443, + "step": 5166 + }, + { + "epoch": 4.959193470955353, + "grad_norm": 1.3664212214230247, + "learning_rate": 2.1724670058825838e-09, + "loss": 0.4382, + "step": 5167 + }, + { + "epoch": 4.960153624579933, + "grad_norm": 1.04263073981288, + "learning_rate": 2.074847400230784e-09, + "loss": 0.398, + "step": 5168 + }, + { + "epoch": 4.9611137782045125, + "grad_norm": 1.1722104802613906, + "learning_rate": 1.9794713097548433e-09, + "loss": 0.5069, + "step": 5169 + }, + { + "epoch": 4.962073931829092, + "grad_norm": 0.8243771578752006, + "learning_rate": 1.886338777268293e-09, + "loss": 0.4242, + "step": 5170 + }, + { + "epoch": 4.963034085453673, + "grad_norm": 1.1138825072860248, + "learning_rate": 1.7954498445771352e-09, + "loss": 0.489, + "step": 5171 + }, + { + "epoch": 4.963994239078253, + "grad_norm": 1.1685493389282025, + "learning_rate": 1.7068045524815114e-09, + "loss": 0.4924, + "step": 5172 + }, + { + "epoch": 4.964954392702833, + "grad_norm": 0.8628981710697552, + "learning_rate": 1.6204029407718147e-09, + "loss": 0.4389, + "step": 5173 + }, + { + "epoch": 4.965914546327412, + "grad_norm": 0.9752931832801603, + "learning_rate": 1.5362450482336865e-09, + "loss": 0.5359, + "step": 5174 + }, + { + "epoch": 4.966874699951992, + "grad_norm": 0.9944015147203397, + "learning_rate": 1.4543309126446858e-09, + "loss": 0.4481, + "step": 5175 + }, + { + "epoch": 4.967834853576572, + "grad_norm": 1.6993954156604336, + "learning_rate": 1.374660570774844e-09, + "loss": 0.3301, + "step": 5176 + }, + { + "epoch": 4.968795007201152, + "grad_norm": 0.7550334465591608, + "learning_rate": 1.2972340583877751e-09, + "loss": 0.4681, + "step": 5177 + }, + { + "epoch": 4.969755160825732, + "grad_norm": 1.4877848652477177, + "learning_rate": 1.2220514102390114e-09, + "loss": 0.468, + "step": 5178 + }, + { + "epoch": 4.970715314450312, + "grad_norm": 1.2603219991752135, + "learning_rate": 1.149112660077667e-09, + "loss": 0.3782, + "step": 5179 + }, + { + "epoch": 4.971675468074892, + "grad_norm": 1.5932806416619052, + "learning_rate": 1.0784178406442192e-09, + "loss": 0.3616, + "step": 5180 + }, + { + "epoch": 4.972635621699472, + "grad_norm": 0.9456974079383582, + "learning_rate": 1.0099669836743931e-09, + "loss": 0.4605, + "step": 5181 + }, + { + "epoch": 4.973595775324052, + "grad_norm": 1.0605197578624472, + "learning_rate": 9.43760119893611e-10, + "loss": 0.3115, + "step": 5182 + }, + { + "epoch": 4.974555928948631, + "grad_norm": 0.9515697992755997, + "learning_rate": 8.797972790219877e-10, + "loss": 0.3714, + "step": 5183 + }, + { + "epoch": 4.975516082573212, + "grad_norm": 1.4573860961194787, + "learning_rate": 8.180784897715565e-10, + "loss": 0.457, + "step": 5184 + }, + { + "epoch": 4.976476236197792, + "grad_norm": 1.015072706107275, + "learning_rate": 7.586037798479329e-10, + "loss": 0.4341, + "step": 5185 + }, + { + "epoch": 4.977436389822372, + "grad_norm": 0.8891637134337853, + "learning_rate": 7.013731759475395e-10, + "loss": 0.4277, + "step": 5186 + }, + { + "epoch": 4.9783965434469515, + "grad_norm": 1.232368610611868, + "learning_rate": 6.463867037614923e-10, + "loss": 0.3846, + "step": 5187 + }, + { + "epoch": 4.979356697071531, + "grad_norm": 0.8693382528761674, + "learning_rate": 5.936443879728249e-10, + "loss": 0.4826, + "step": 5188 + }, + { + "epoch": 4.980316850696111, + "grad_norm": 0.8860195787172938, + "learning_rate": 5.431462522559327e-10, + "loss": 0.4861, + "step": 5189 + }, + { + "epoch": 4.981277004320692, + "grad_norm": 0.9964801950407475, + "learning_rate": 4.948923192793497e-10, + "loss": 0.5072, + "step": 5190 + }, + { + "epoch": 4.9822371579452716, + "grad_norm": 0.9280557523722671, + "learning_rate": 4.4888261070408223e-10, + "loss": 0.5056, + "step": 5191 + }, + { + "epoch": 4.983197311569851, + "grad_norm": 1.129209496152681, + "learning_rate": 4.0511714718305396e-10, + "loss": 0.53, + "step": 5192 + }, + { + "epoch": 4.984157465194431, + "grad_norm": 1.2597070459144213, + "learning_rate": 3.6359594836277156e-10, + "loss": 0.4663, + "step": 5193 + }, + { + "epoch": 4.985117618819011, + "grad_norm": 0.9751585238349448, + "learning_rate": 3.24319032880549e-10, + "loss": 0.382, + "step": 5194 + }, + { + "epoch": 4.986077772443591, + "grad_norm": 1.481484926785769, + "learning_rate": 2.872864183678381e-10, + "loss": 0.3733, + "step": 5195 + }, + { + "epoch": 4.9870379260681705, + "grad_norm": 1.4032654679426264, + "learning_rate": 2.5249812144856335e-10, + "loss": 0.4836, + "step": 5196 + }, + { + "epoch": 4.987998079692751, + "grad_norm": 0.9277705194905103, + "learning_rate": 2.1995415773856666e-10, + "loss": 0.5167, + "step": 5197 + }, + { + "epoch": 4.988958233317331, + "grad_norm": 1.2258570435859233, + "learning_rate": 1.8965454184671773e-10, + "loss": 0.387, + "step": 5198 + }, + { + "epoch": 4.989918386941911, + "grad_norm": 0.9924078353603917, + "learning_rate": 1.615992873732486e-10, + "loss": 0.4717, + "step": 5199 + }, + { + "epoch": 4.990878540566491, + "grad_norm": 0.8285429982259878, + "learning_rate": 1.3578840691308438e-10, + "loss": 0.424, + "step": 5200 + }, + { + "epoch": 4.99183869419107, + "grad_norm": 0.9218774212055307, + "learning_rate": 1.1222191205195743e-10, + "loss": 0.4128, + "step": 5201 + }, + { + "epoch": 4.99279884781565, + "grad_norm": 1.2712175336651776, + "learning_rate": 9.089981336807275e-11, + "loss": 0.3499, + "step": 5202 + }, + { + "epoch": 4.99375900144023, + "grad_norm": 0.9820496496723373, + "learning_rate": 7.182212043377323e-11, + "loss": 0.3324, + "step": 5203 + }, + { + "epoch": 4.994719155064811, + "grad_norm": 1.2192713071695829, + "learning_rate": 5.4988841812209095e-11, + "loss": 0.3896, + "step": 5204 + }, + { + "epoch": 4.9956793086893905, + "grad_norm": 1.1385866933018771, + "learning_rate": 4.039998505900311e-11, + "loss": 0.405, + "step": 5205 + }, + { + "epoch": 4.99663946231397, + "grad_norm": 1.0288482489490556, + "learning_rate": 2.8055556724471132e-11, + "loss": 0.5846, + "step": 5206 + }, + { + "epoch": 4.99759961593855, + "grad_norm": 1.2764357875772903, + "learning_rate": 1.7955562348626054e-11, + "loss": 0.5251, + "step": 5207 + }, + { + "epoch": 4.99855976956313, + "grad_norm": 1.2640086373758026, + "learning_rate": 1.0100006465618706e-11, + "loss": 0.3624, + "step": 5208 + }, + { + "epoch": 4.99951992318771, + "grad_norm": 1.049982002681759, + "learning_rate": 4.4888926020725254e-12, + "loss": 0.4065, + "step": 5209 + }, + { + "epoch": 5.0, + "grad_norm": 1.049982002681759, + "learning_rate": 1.1222232765284447e-12, + "loss": 0.362, + "step": 5210 + }, + { + "epoch": 5.0, + "step": 5210, + "total_flos": 432224790970368.0, + "train_loss": 0.5731080526353759, + "train_runtime": 83358.552, + "train_samples_per_second": 0.5, + "train_steps_per_second": 0.063 + } + ], + "logging_steps": 1, + "max_steps": 5210, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 432224790970368.0, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}